apiVersion: v1
items:
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Flux project
      provider-url: ""
    name: flux
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.toolkit.fluxcd.io/v2beta1",
                "kind": "HelmRelease",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "5m",
                    "chart": {
                      "spec": {
                          "chart": "podinfo",
                          "version": "4.0.x",
                          "sourceRef": {
                            "kind": "HelmRepository",
                            "name": "podinfo",
                            "namespace": "flux-system"
                          },
                          "interval": "1m"
                      }
                    },
                    "values": {
                      "replicaCount": 2
                    }
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "GitRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "1m",
                    "ref": {
                      "branch": "master"
                    },
                    "url": "https://github.com/stefanprodan/podinfo"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmRepository",
                "metadata": {
                    "name": "stable",
                    "namespace": "default"
                },
                "spec": {
                    "url": "https://kubernetes-charts.storage.googleapis.com/",
                    "interval": "10m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmChart",
                "metadata": {
                    "name": "redis",
                    "namespace": "default"
                },
                "spec": {
                    "chart": "redis",
                    "version": "10.5.7",
                    "sourceRef": {
                      "name": "stable",
                      "kind": "HelmRepository"
                    },
                    "interval": "5m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "Bucket",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "1m",
                    "provider": "generic",
                    "bucketName": "podinfo",
                    "endpoint": "minio.minio.svc.cluster.local:9000",
                    "insecure": true,
                    "secretRef": {
                      "name": "minio-credentials"
                    }
                }
              },
              {
                "apiVersion": "kustomize.toolkit.fluxcd.io/v1beta2",
                "kind": "Kustomization",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "5m",
                    "path": "./deploy/overlays/dev",
                    "prune": true,
                    "sourceRef": {
                      "kind": "GitRepository",
                      "name": "podinfo"
                    },
                    "timeout": "2m",
                    "wait": true
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "address": "https://github.com/stefanprodan/podinfo",
                    "secretRef": {
                      "name": "api-token"
                    }
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Alert",
                "metadata": {
                    "name": "webapp",
                    "namespace": "default"
                },
                "spec": {
                    "providerRef": {
                      "name": "on-call-slack"
                    },
                    "eventSeverity": "info",
                    "eventSources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "Bucket",
                          "name": "secrets"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-backend"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-frontend"
                      }
                    ]
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Receiver",
                "metadata": {
                    "name": "github-receiver",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "events": [
                      "ping",
                      "push"
                    ],
                    "secretRef": {
                      "name": "webhook-token"
                    },
                    "resources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "HelmRepository",
                          "name": "webapp"
                      }
                    ]
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "image": "ghcr.io/stefanprodan/podinfo",
                    "interval": "1m0s"
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImagePolicy",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "imageRepositoryRef": {
                      "name": "podinfo"
                    },
                    "policy": {
                      "semver": {
                          "range": "5.0.x"
                      }
                    }
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageUpdateAutomation",
                "metadata": {
                  "name": "flux-system",
                  "namespace": "flux-system"
                },
                "spec": {
                  "interval": "1m0s",
                  "sourceRef": {
                    "kind": "GitRepository",
                    "name": "flux-system"
                  },
                  "git": {
                    "checkout": {
                      "ref": {
                        "branch": "main"
                      }
                    },
                    "commit": {
                      "author": {
                        "email": "fluxcdbot@users.noreply.github.com",
                        "name": "fluxcdbot"
                      },
                      "messageTemplate": "{{range .Updated.Images}}{{println .}}{{end}}"
                    },
                    "push": {
                      "branch": "main"
                    }
                  },
                  "update": {
                    "path": "./clusters/my-cluster",
                    "strategy": "Setters"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/fluxcd/source-controller:v1.3.0
          description: Flux is a Continuous Delivery solution for Kubernetes.
          operatorframework.io/suggested-namespace: flux-system
          support: The Flux community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta3
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta1
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta2
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta1
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta2
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta2
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta2
          - description: OCIRepository
            displayName: OCIRepository
            kind: OCIRepository
            name: ocirepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta3
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta2
        description: |-
          [Flux](https://fluxcd.io/) is a Continuous Delivery solution for Kubernetes.
          Flux is constructed with the [GitOps Toolkit](https://toolkit.fluxcd.io/components/), a set of composable APIs and specialized tools for keeping Kubernetes clusters in sync with sources of configuration (like Git & Helm repositories), and automating updates to configuration when there is new code to deploy.

          Flux is a CNCF incubation project made for:

          - **cluster operators** who automate provision and configuration of clusters
          - **platform engineers** who build continuous delivery for developer teams
          - **app developers** who rely on continuous delivery to get their code live

          Features:

          - Source configuration from Git and Helm repositories, and S3-compatible buckets (e.g., Minio)
          - Kustomize and Helm support
          - Event-triggered and periodic reconciliation
          - Integration with Kubernetes RBAC
          - Health assessment (clusters and workloads)
          - Dependency management (infrastructure and workloads)
          - Secrets management (builtin support for Mozilla SOPS and Cloud KMS)
          - Alerting to external systems (webhook senders)
          - External events handling (webhook receivers)
          - Automated container image updates to Git (image scanning and patching)
          - Policy-driven validation (OPA, Kyverno, admission controllers)
          - Seamless integration with Git providers (GitHub, GitLab, Bitbucket)
          - Interoperability with Cluster API (CAPI) providers

          ### Steps to support bootstraping with Flux CLI

          SecComp Profile should be removed as it requires a special privilege to get the access.
          Also, it requires to enforce `runUserAs` to the same UID provided by the images to prevent OpenShift
          to alter the user id expected by our controllers.

          To do so, please specifiy a patch using the following Kustomization file, for example.

          ```yaml
          apiVersion: kustomize.config.k8s.io/v1beta1
          kind: Kustomization
          resources:
            - gotk-components.yaml
            - gotk-sync.yaml
          patches:
            - patch: |
                apiVersion: apps/v1
                kind: Deployment
                metadata:
                  name: all
                spec:
                  template:
                    spec:
                      containers:
                        - name: manager
                          securityContext:
                            runAsUser: 65534
                            seccompProfile:
                              $patch: delete
              target:
                kind: Deployment
                labelSelector: app.kubernetes.io/part-of=flux
          ```

          Then before installing Flux, please also perform the following steps to bind Flux to the `nonroot` SecurityContextConstraint (SCC). This applies to installation via OC's OLM or via Flux CLI.

          ```sh
          # replace default "flux-system" with the project/namespace name for your flux installation
          namespace="flux-system"
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:source-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:kustomize-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:helm-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:notification-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-automation-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-reflector-controller

          ```

          To see complete steps please refer to this [documentation](https://github.com/fluxcd/website/blob/main/content/en/docs/use-cases/openshift.md).

          If you've already installed Flux, you might need to uninstall it after applying the above pre-requisite steps.

          ### Network Policies

          This version of Flux is shipped without the normally provided network Policies.
          By default all traffic is enabled, and anyone can access the source-controller
          artifacts.

          It is not recommended to run it in production as is.

          Please visit [Flux's policies](https://github.com/fluxcd/flux2/tree/main/manifests/policies)
          to deploy the needed network policies and secure your Flux setup.

          ### Breaking Changes

          - Flux 0.18.x ships with v1beta2 of the kustomize.toolkit.fluxcd.io/Kustomization CRD.
          - Flux 0.25.x is the last Flux release where Kubernetes 1.19 is supported.

          ### Flux on OpenShift Use Cases

          - [The GitOps with Flux + OpenShift](https://www.youtube.com/watch?v=8ybCtNVP5To) talk by Andrew Block (Red Hat) showing how to setup multi-tenacy on OpenShift using Flux.
          - [Flux OpenShift Use Case](https://fluxcd.io/docs/use-cases/openshift/) on FluxCD.io
        displayName: Flux
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - gitops
        - cd
        - continuous delivery
        - fluxcd
        - helm
        - kustomize
        links:
        - name: Homepage
          url: https://fluxcd.io/
        - name: Documentation
          url: https://fluxcd.io/docs/
        - name: Slack
          url: https://cloud-native.slack.com/messages/flux
        - name: Mailing List
          url: https://lists.cncf.io/g/cncf-flux-dev
        - name: Source Code
          url: https://github.com/fluxcd/flux2
        maintainers:
        - email: support@weave.works
          name: The Flux maintainers
        maturity: stable
        provider:
          name: The Flux project
        relatedImages:
        - ghcr.io/fluxcd/helm-controller:v1.0.1
        - ghcr.io/fluxcd/image-automation-controller:v0.38.0
        - ghcr.io/fluxcd/image-reflector-controller:v0.32.0
        - ghcr.io/fluxcd/kustomize-controller:v1.3.0
        - ghcr.io/fluxcd/notification-controller:v1.3.0
        - ghcr.io/fluxcd/source-controller:v1.3.0
        version: 2.3.0
      entries:
      - name: flux.v2.3.0
        version: 2.3.0
      - name: flux.v2.2.2
        version: 2.2.2
      - name: flux.v2.2.0
        version: 2.2.0
      - name: flux.v2.0.1
        version: 2.0.1
      - name: flux.v0.41.2
        version: 0.41.2
      - name: flux.v0.40.0
        version: 0.40.0
      - name: flux.v0.39.0
        version: 0.39.0
      - name: flux.v0.38.3
        version: 0.38.3
      - name: flux.v0.37.0
        version: 0.37.0
      - name: flux.v0.36.0
        version: 0.36.0
      - name: flux.v0.35.0
        version: 0.35.0
      - name: flux.v0.31.4
        version: 0.31.4
      - name: flux.v0.31.3
        version: 0.31.3
      - name: flux.v0.30.2
        version: 0.30.2
      - name: flux.v0.28.5
        version: 0.28.5
      - name: flux.v0.28.4
        version: 0.28.4
      - name: flux.v0.28.3
        version: 0.28.3
      - name: flux.v0.27.4
        version: 0.27.4
      - name: flux.v0.27.3
        version: 0.27.3
      - name: flux.v0.27.2
        version: 0.27.2
      - name: flux.v0.27.0
        version: 0.27.0
      - name: flux.v0.26.1
        version: 0.26.1
      - name: flux.v0.26.0
        version: 0.26.0
      - name: flux.v0.25.3
        version: 0.25.3
      - name: flux.v0.25.2
        version: 0.25.2
      - name: flux.v0.24.1
        version: 0.24.1
      - name: flux.v0.24.0
        version: 0.24.0
      - name: flux.v0.23.0
        version: 0.23.0
      - name: flux.v0.22.1
        version: 0.22.1
      - name: flux.v0.22.0
        version: 0.22.0
      - name: flux.v0.21.1
        version: 0.21.1
      - name: flux.v0.21.0
        version: 0.21.0
      - name: flux.v0.20.1
        version: 0.20.1
      - name: flux.v0.20.0
        version: 0.20.0
      - name: flux.v0.19.1
        version: 0.19.1
      - name: flux.v0.18.3
        version: 0.18.3
      - name: flux.v0.18.1
        version: 0.18.1
      - name: flux.v0.17.2
        version: 0.17.2
      - name: flux.v0.17.1
        version: 0.17.1
      - name: flux.v0.17.0
        version: 0.17.0
      - name: flux.v0.16.2
        version: 0.16.2
      - name: flux.v0.16.1
        version: 0.16.1
      - name: flux.v0.16.0
        version: 0.16.0
      - name: flux.v0.15.3
        version: 0.15.3
      - name: flux.v0.15.2
        version: 0.15.2
      - name: flux.v0.14.2
        version: 0.14.2
      - name: flux.v0.14.1
        version: 0.14.1
      - name: flux.v0.14.0
        version: 0.14.0
      - name: flux.v0.13.4
        version: 0.13.4
      name: stable
    defaultChannel: stable
    packageName: flux
    provider:
      name: The Flux project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          olm.skipRange: <1.0.4
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
        version: 1.0.4
      entries:
      - name: datadog-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      name: stable
    defaultChannel: stable
    packageName: datadog-operator-certified
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kiali-ossm
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kiali-operator.v2.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "Kiali",
                "metadata": {
                  "name": "kiali"
                },
                "spec": {
                  "version": "default",
                  "istio_namespace": "istio-system",
                  "deployment": {
                    "logger": {
                      "log_level": "info"
                    },
                    "view_only_mode": false
                  }
                }
              },
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "OSSMConsole",
                "metadata": {
                  "name": "ossmconsole"
                },
                "spec": {
                  "version": "default",
                  "kiali": {
                    "serviceName": "",
                    "serviceNamespace": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9-operator@sha256:0612dfa52f2bf39a1d87e640409313585fedc7cfcff6d00e77bb14809d27d646
          createdAt: "2024-12-12T10:44:15Z"
          description: This productized operator provides Kiali and OSSMC. Kiali is
            the Istio observability and management Console. OSSMC is the OpenShift
            Service Mesh Console plugin, powered by Kiali.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.1.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kiali/kiali
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A configuration file for a Kiali installation.
            displayName: Kiali
            kind: Kiali
            name: kialis.kiali.io
            version: v1alpha1
          - description: A configuration file for a OpenShift Service Mesh Console
              installation.
            displayName: OpenShift Service Mesh Console
            kind: OSSMConsole
            name: ossmconsoles.kiali.io
            version: v1alpha1
        description: |-
          ## About the managed application

          A Microservice Architecture breaks up the monolith into many smaller pieces
          that are composed together. Patterns to secure the communication between
          services like fault tolerance (via timeout, retry, circuit breaking, etc.)
          have come up as well as distributed tracing to be able to see where calls
          are going.

          A service mesh can now provide these services on a platform level and frees
          the application writers from those tasks. Routing decisions are done at the
          mesh level.

          Kiali works with OpenShift Service Mesh to visualize the service
          mesh topology, to provide visibility into features like circuit breakers,
          request rates and more. It offers insights about the mesh components at
          different levels, from abstract Applications to Services and Workloads.

          See [https://www.kiali.io](https://www.kiali.io) to read more.

          ### Accessing the UI

          By default, the Kiali operator exposes the Kiali UI as an OpenShift Route.

          If on OpenShift, you can create an OSSMConsole CR to have the operator
          install the OpenShift ServiceMesh Console plugin to the OpenShift Console
          thus providing an interface directly integrated with the OpenShift Console.

          ## About this Operator

          ### Kiali Custom Resource Configuration Settings

          For quick descriptions of all the settings you can configure in the Kiali
          Custom Resource (CR), see
          [the kiali.io docs](https://kiali.io/docs/configuration/kialis.kiali.io/).

          ## Prerequisites for enabling this Operator

          Kiali is a companion tool for OpenShift Service Mesh. So before you install Kiali, you must have
          already installed Service Mesh.
        displayName: Kiali Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - service-mesh
        - observability
        - monitoring
        - istio
        links:
        - name: Installation Guide
          url: https://www.kiali.io/docs/installation
        - name: Features
          url: https://www.kiali.io/docs/features
        - name: Documentation Home
          url: https://www.kiali.io
        - name: Blogs and Articles
          url: https://medium.com/kialiproject
        - name: Server and UI Source Code
          url: https://github.com/kiali/kiali
        - name: Operator Source Code
          url: https://github.com/kiali/kiali-operator
        - name: OpenShift Service Mesh Console Code
          url: https://github.com/kiali/openshift-servicemesh-plugin
        maintainers:
        - email: kiali-dev@googlegroups.com
          name: Kiali Developers Google Group
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel8@sha256:e37d76dd7c7a2db79ff1e96afef19b07c3beaa8ebf0c7bd85b808908b8f0ce77
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9-operator@sha256:0612dfa52f2bf39a1d87e640409313585fedc7cfcff6d00e77bb14809d27d646
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9@sha256:06fde6c696a33d17b0d2f7620ec46422c82277fd17a3fe6d7319de05c732aace
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:976a31ca8cf17ac0887b8a586edfcc29eb30175802461b3114758af525e7b800
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:66e4d933a9d69158bf87dd45835183bf2308b77c48908ab64496c5d22c816293
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:27a6aebffa3d049ee48050126e6107db2e07c49e9ce9297175facb6368e1e792
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-ossmc-rhel9@sha256:99643f294c3f92a6fc82f321cab7c756e4879fefb48fa824f1bf49ff7182658d
        version: 2.1.2
      entries:
      - name: kiali-operator.v2.1.2
        version: 2.1.2
      - name: kiali-operator.v1.89.8
        version: 1.89.8
      - name: kiali-operator.v1.89.7
        version: 1.89.7
      - name: kiali-operator.v1.89.6
        version: 1.89.6
      - name: kiali-operator.v1.89.1
        version: 1.89.1
      - name: kiali-operator.v1.73.10
        version: 1.73.10
      - name: kiali-operator.v1.73.8
        version: 1.73.8
      - name: kiali-operator.v1.73.7
        version: 1.73.7
      - name: kiali-operator.v1.73.4
        version: 1.73.4
      - name: kiali-operator.v1.65.11
        version: 1.65.11
      - name: kiali-operator.v1.65.10
        version: 1.65.10
      - name: kiali-operator.v1.65.9
        version: 1.65.9
      - name: kiali-operator.v1.65.8
        version: 1.65.8
      - name: kiali-operator.v1.65.7
        version: 1.65.7
      - name: kiali-operator.v1.65.6
        version: 1.65.6
      - name: kiali-operator.v1.57.7
        version: 1.57.7
      name: candidate
    - currentCSV: kiali-operator.v2.11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "Kiali",
                "metadata": {
                  "name": "kiali"
                },
                "spec": {
                  "version": "default",
                  "istio_namespace": "istio-system",
                  "deployment": {
                    "logger": {
                      "log_level": "info"
                    },
                    "view_only_mode": false
                  }
                }
              },
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "OSSMConsole",
                "metadata": {
                  "name": "ossmconsole"
                },
                "spec": {
                  "version": "default",
                  "kiali": {
                    "serviceName": "",
                    "serviceNamespace": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/kiali-rhel9-operator@sha256:2970861bd8e65e36d5080dc4d56340d08a52dbb1217be352998d135eb854732f
          createdAt: "2025-09-10T00:28:49Z"
          description: This productized operator provides Kiali and OSSMC. Kiali is
            the Istio observability and management Console. OSSMC is the OpenShift
            Service Mesh Console plugin, powered by Kiali.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kiali/kiali
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A configuration file for a Kiali installation.
            displayName: Kiali
            kind: Kiali
            name: kialis.kiali.io
            version: v1alpha1
          - description: A configuration file for a OpenShift Service Mesh Console
              installation.
            displayName: OpenShift Service Mesh Console
            kind: OSSMConsole
            name: ossmconsoles.kiali.io
            version: v1alpha1
        description: |-
          ## About the managed application

          A Microservice Architecture breaks up the monolith into many smaller pieces
          that are composed together. Patterns to secure the communication between
          services like fault tolerance (via timeout, retry, circuit breaking, etc.)
          have come up as well as distributed tracing to be able to see where calls
          are going.

          A service mesh can now provide these services on a platform level and frees
          the application writers from those tasks. Routing decisions are done at the
          mesh level.

          Kiali works with OpenShift Service Mesh to visualize the service
          mesh topology, to provide visibility into features like circuit breakers,
          request rates and more. It offers insights about the mesh components at
          different levels, from abstract Applications to Services and Workloads.

          See [https://www.kiali.io](https://www.kiali.io) to read more.

          ### Accessing the UI

          By default, the Kiali operator exposes the Kiali UI as an OpenShift Route.

          If on OpenShift, you can create an OSSMConsole CR to have the operator
          install the OpenShift ServiceMesh Console plugin to the OpenShift Console
          thus providing an interface directly integrated with the OpenShift Console.

          ## About this Operator

          ### Kiali Custom Resource Configuration Settings

          For quick descriptions of all the settings you can configure in the Kiali
          Custom Resource (CR), see
          [the kiali.io docs](https://kiali.io/docs/configuration/kialis.kiali.io/).

          ## Prerequisites for enabling this Operator

          Kiali is a companion tool for OpenShift Service Mesh. So before you install Kiali, you must have
          already installed Service Mesh.
        displayName: Kiali Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - service-mesh
        - observability
        - monitoring
        - istio
        links:
        - name: Installation Guide
          url: https://www.kiali.io/docs/installation
        - name: Features
          url: https://www.kiali.io/docs/features
        - name: Documentation Home
          url: https://www.kiali.io
        - name: Blogs and Articles
          url: https://medium.com/kialiproject
        - name: Server and UI Source Code
          url: https://github.com/kiali/kiali
        - name: Operator Source Code
          url: https://github.com/kiali/kiali-operator
        - name: OpenShift Service Mesh Console Code
          url: https://github.com/kiali/openshift-servicemesh-plugin
        maintainers:
        - email: kiali-dev@googlegroups.com
          name: Kiali Developers Google Group
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9-operator@sha256:2970861bd8e65e36d5080dc4d56340d08a52dbb1217be352998d135eb854732f
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:7487fbb4b776dfca28ad09311ba857e07ad98bc91ca5ee6fcaaa5b7783d04fbd
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel9@sha256:9a0d4862a25c2999a3f85b9a7a0332dab4ec1744548191d7a00c838b06ba26d2
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel9@sha256:fe25be563d50e5ad4e5475f153e03d4b24c7db094d7e824c51ebaae79418b1f7
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel8@sha256:81db94df331799f4087596378914b712d85d117dde3df2d4902079a82d4a9c55
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9@sha256:a00ec00e466e080592dca5c1cc153698b1412a54e1979647f74281f79b5691cf
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9@sha256:85eeb972e46f26ff30e04991e76ec3050d98b334e7b7f201a094f8d50b06854e
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:d729ef8d74c2f6e02d8c68ca761b1fb3c7c3706f1290b684468f0acb0ef313e0
        version: 2.11.3
      entries:
      - name: kiali-operator.v2.11.3
        version: 2.11.3
      - name: kiali-operator.v2.11.2
        version: 2.11.2
      - name: kiali-operator.v2.11.1
        version: 2.11.1
      - name: kiali-operator.v2.4.7
        version: 2.4.7
      - name: kiali-operator.v2.4.6
        version: 2.4.6
      - name: kiali-operator.v2.4.5
        version: 2.4.5
      - name: kiali-operator.v2.4.4
        version: 2.4.4
      - name: kiali-operator.v2.4.3
        version: 2.4.3
      - name: kiali-operator.v1.89.10
        version: 1.89.10
      - name: kiali-operator.v1.89.9
        version: 1.89.9
      - name: kiali-operator.v1.89.8
        version: 1.89.8
      - name: kiali-operator.v1.89.7
        version: 1.89.7
      - name: kiali-operator.v1.89.6
        version: 1.89.6
      - name: kiali-operator.v1.89.1
        version: 1.89.1
      - name: kiali-operator.v1.73.10
        version: 1.73.10
      - name: kiali-operator.v1.73.8
        version: 1.73.8
      - name: kiali-operator.v1.73.7
        version: 1.73.7
      - name: kiali-operator.v1.73.4
        version: 1.73.4
      - name: kiali-operator.v1.65.11
        version: 1.65.11
      - name: kiali-operator.v1.65.10
        version: 1.65.10
      - name: kiali-operator.v1.65.9
        version: 1.65.9
      - name: kiali-operator.v1.65.8
        version: 1.65.8
      - name: kiali-operator.v1.65.7
        version: 1.65.7
      - name: kiali-operator.v1.65.6
        version: 1.65.6
      - name: kiali-operator.v1.57.7
        version: 1.57.7
      name: stable
    defaultChannel: stable
    packageName: kiali-ossm
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: test-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: test-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "test.openstack.org/v1beta1",
                "kind": "HorizonTest",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "test-operator",
                    "app.kubernetes.io/instance": "horizontest-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "horizontest",
                    "app.kubernetes.io/part-of": "test-operator"
                  },
                  "name": "horizontest-sample"
                },
                "spec": {
                  "adminPassword": "12345678",
                  "adminUsername": "admin",
                  "authUrl": "https://keystone-public-openstack.apps.ocp.openstack.lab",
                  "backoffLimit": 0,
                  "containerImage": "",
                  "dashboardUrl": "https://horizon-openstack.apps.ocp.openstack.lab/",
                  "flavorName": "m1.tiny",
                  "horizonRepoBranch": "master",
                  "horizonTestDir": "/var/lib/horizontest",
                  "imageUrl": "http://download.cirros-cloud.net/0.6.2/cirros-0.6.2-x86_64-disk.img",
                  "logsDirectoryName": "horizon",
                  "parallel": false,
                  "password": "horizontest",
                  "projectName": "horizontest",
                  "repoUrl": "https://review.opendev.org/openstack/horizon",
                  "storageClass": "local-storage",
                  "user": "horizontest"
                }
              },
              {
                "apiVersion": "test.openstack.org/v1beta1",
                "kind": "Tempest",
                "metadata": {
                  "name": "tempest-tests",
                  "namespace": "openstack"
                },
                "spec": {
                  "containerImage": "",
                  "tempestRun": {
                    "concurrency": 8,
                    "includeList": "tempest.api.identity.v3.*\n"
                  },
                  "tempestconfRun": null
                }
              },
              {
                "apiVersion": "test.openstack.org/v1beta1",
                "kind": "Tobiko",
                "metadata": {
                  "name": "tobiko-tests",
                  "namespace": "openstack"
                },
                "spec": {
                  "config": "[DEFAULT]\nlog_file = value\n\n[testcase]\ntimeout = value\ntest_runner_timeout = value\n\n[ubuntu]\ninterface_name = value\n\n[keystone]\ninterface_name = value\n",
                  "containerImage": "",
                  "testenv": "py3",
                  "version": "master",
                  "workflow": [
                    {
                      "stepName": "unit-tests",
                      "testenv": "py3"
                    },
                    {
                      "stepName": "functional-tests",
                      "testenv": "functional"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:43Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: Ansible Test
            kind: AnsibleTest
            name: ansibletests.test.openstack.org
            version: v1beta1
          - displayName: Horizon Test
            kind: HorizonTest
            name: horizontests.test.openstack.org
            version: v1beta1
          - displayName: Tempest
            kind: Tempest
            name: tempests.test.openstack.org
            version: v1beta1
          - kind: Tobiko
            name: tobikoes.test.openstack.org
            version: v1beta1
        description: Test Operator
        displayName: OpenStack Test Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Tempest
        - Tobiko
        - Tests
        links:
        - name: Test Operator
          url: https://github.com/openstack-k8s-operators/test-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/test-rhel9-operator@sha256:17949ad6a69a14ea38870d471fd4f5a51122927f795976ac7e2b0189e2d0da5a
        - registry.redhat.io/rhoso/openstack-tempest-all-rhel9@sha256:2f07239d932a67a6589844f4380f613cc16614ff4de7699f64c17359bbdc5534
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: test-operator.v1.0.6
        version: 1.0.6
      - name: test-operator.v1.0.5
        version: 1.0.5
      - name: test-operator.v1.0.4
        version: 1.0.4
      - name: test-operator.v1.0.3
        version: 1.0.3
      - name: test-operator.v1.0.2
        version: 1.0.2
      - name: test-operator.v1.0.1
        version: 1.0.1
      - name: test-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: test-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhsso-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhsso-operator.7.5.1-opr-005
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm"
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "False"
          containerImage: registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:66577ec32a9a975f304e52b8ab49b9f7602158dcab1ef5d8c28a5d3455491336
          createdAt: "2021-08-09 00:00:00"
          description: An Operator for installing and managing Red Hat Single Sign-On
          olm.skipRange: <7.5.0
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Red Hat Single Sign-On Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Red Hat Single Sign-On.

          Red Hat Single Sign-On lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
        displayName: Red Hat Single Sign-On Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://access.redhat.com/products/red-hat-single-sign-on
        - name: Keycloak
          url: https://www.keycloak.org/
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat Customer Service
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso7-rhel8-init-container@sha256:614a9b1d7975769e8559e0b8c1578d16f6e3f7fee0778283797f4d62f5fc2100
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:5f0840a69642d13016430151892ec987fd67e711563b1dfb1a57353037c3bf79
        - registry.redhat.io/rh-sso-7/sso75-openj9-openshift-rhel8@sha256:2d4ea7226f43e7e8e8d3fcef6d425eb5104a15acf07791355033300b80db2901
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:66577ec32a9a975f304e52b8ab49b9f7602158dcab1ef5d8c28a5d3455491336
        version: 7.5.1-opr-005
      entries:
      - name: rhsso-operator.7.5.1-opr-005
        version: 7.5.1-opr-005
      - name: rhsso-operator.7.5.1-opr-004
        version: 7.5.1-opr-004
      - name: rhsso-operator.7.5.1-opr-003
        version: 7.5.1-opr-003
      - name: rhsso-operator.7.5.1-opr-002
        version: 7.5.1-opr-002
      - name: rhsso-operator.7.5.1-opr-001
        version: 7.5.1-opr-001
      - name: rhsso-operator.7.5.0
        version: 7.5.0
      name: alpha
    - currentCSV: rhsso-operator.7.6.12-opr-004
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm"
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "False"
          containerImage: registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:229265579ec77f2bd18c936ba19e1f9dbc6b2f0a13d963d5eae5b52c260010d1
          createdAt: "2021-08-09 00:00:00"
          description: An Operator for installing and managing Red Hat Single Sign-On
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Red Hat Single Sign-On Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Red Hat Single Sign-On.

          Red Hat Single Sign-On lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
        displayName: Red Hat Single Sign-On Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://access.redhat.com/products/red-hat-single-sign-on
        - name: Keycloak
          url: https://www.keycloak.org/
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat Customer Service
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:229265579ec77f2bd18c936ba19e1f9dbc6b2f0a13d963d5eae5b52c260010d1
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:923206112ed0873ee58edd0817dbce7bb4c9a89dc75d9b3ba2affe9aed4811de
        - registry.redhat.io/rh-sso-7/sso7-rhel8-init-container@sha256:282daf78934ca495ad4051371a58e3af0917bb9241d043a66e23fc8b12fbd1ce
        version: 7.6.12-opr-004
      entries:
      - name: rhsso-operator.7.6.12-opr-004
        version: 7.6.12-opr-004
      - name: rhsso-operator.7.6.12-opr-003
        version: 7.6.12-opr-003
      - name: rhsso-operator.7.6.12-opr-002
        version: 7.6.12-opr-002
      - name: rhsso-operator.7.6.12-opr-001
        version: 7.6.12-opr-001
      - name: rhsso-operator.7.6.11-opr-006
        version: 7.6.11-opr-006
      - name: rhsso-operator.7.6.11-opr-005
        version: 7.6.11-opr-005
      - name: rhsso-operator.7.6.11-opr-004
        version: 7.6.11-opr-004
      - name: rhsso-operator.7.6.11-opr-003
        version: 7.6.11-opr-003
      - name: rhsso-operator.7.6.11-opr-002
        version: 7.6.11-opr-002
      - name: rhsso-operator.7.6.11-opr-001
        version: 7.6.11-opr-001
      - name: rhsso-operator.7.6.10-opr-001
        version: 7.6.10-opr-001
      - name: rhsso-operator.7.6.9-opr-003
        version: 7.6.9-opr-003
      - name: rhsso-operator.7.6.9-opr-002
        version: 7.6.9-opr-002
      - name: rhsso-operator.7.6.9-opr-001
        version: 7.6.9-opr-001
      - name: rhsso-operator.7.6.8-opr-002
        version: 7.6.8-opr-002
      - name: rhsso-operator.7.6.8-opr-001
        version: 7.6.8-opr-001
      - name: rhsso-operator.7.6.7-opr-002
        version: 7.6.7-opr-002
      - name: rhsso-operator.7.6.7-opr-001
        version: 7.6.7-opr-001
      - name: rhsso-operator.7.6.6-opr-004
        version: 7.6.6-opr-004
      - name: rhsso-operator.7.6.6-opr-003
        version: 7.6.6-opr-003
      - name: rhsso-operator.7.6.6-opr-002
        version: 7.6.6-opr-002
      - name: rhsso-operator.7.6.6-opr-001
        version: 7.6.6-opr-001
      - name: rhsso-operator.7.6.5-opr-004
        version: 7.6.5-opr-004
      - name: rhsso-operator.7.6.5-opr-003
        version: 7.6.5-opr-003
      - name: rhsso-operator.7.6.5-opr-002
        version: 7.6.5-opr-002
      - name: rhsso-operator.7.6.5-opr-001
        version: 7.6.5-opr-001
      - name: rhsso-operator.7.6.4-opr-003
        version: 7.6.4-opr-003
      - name: rhsso-operator.7.6.4-opr-002
        version: 7.6.4-opr-002
      - name: rhsso-operator.7.6.4-opr-001
        version: 7.6.4-opr-001
      - name: rhsso-operator.7.6.3-opr-002
        version: 7.6.3-opr-002
      - name: rhsso-operator.7.6.3-opr-001
        version: 7.6.3-opr-001
      - name: rhsso-operator.7.6.2-opr-003
        version: 7.6.2-opr-003
      - name: rhsso-operator.7.6.2-opr-002
        version: 7.6.2-opr-002
      - name: rhsso-operator.7.6.2-opr-001
        version: 7.6.2-opr-001
      - name: rhsso-operator.7.6.1-opr-005
        version: 7.6.1-opr-005
      - name: rhsso-operator.7.6.1-opr-004
        version: 7.6.1-opr-004
      - name: rhsso-operator.7.6.1-opr-003
        version: 7.6.1-opr-003
      - name: rhsso-operator.7.6.1-opr-002
        version: 7.6.1-opr-002
      - name: rhsso-operator.7.6.1-opr-001
        version: 7.6.1-opr-001
      - name: rhsso-operator.7.6.0-opr-005
        version: 7.6.0-opr-005
      - name: rhsso-operator.7.6.0-opr-004
        version: 7.6.0-opr-004
      - name: rhsso-operator.7.6.0-opr-003
        version: 7.6.0-opr-003
      - name: rhsso-operator.7.6.0-opr-002
        version: 7.6.0-opr-002
      - name: rhsso-operator.7.6.0-opr-001
        version: 7.6.0-opr-001
      - name: rhsso-operator.7.5.2-opr-003
        version: 7.5.2-opr-003
      - name: rhsso-operator.7.5.2-opr-002
        version: 7.5.2-opr-002
      - name: rhsso-operator.7.5.2-opr-001
        version: 7.5.2-opr-001
      - name: rhsso-operator.7.5.1-opr-011
        version: 7.5.1-opr-011
      - name: rhsso-operator.7.5.1-opr-010
        version: 7.5.1-opr-010
      - name: rhsso-operator.7.5.1-opr-009
        version: 7.5.1-opr-009
      - name: rhsso-operator.7.5.1-opr-009-0.1646836600.p
        version: 7.5.1-opr-009+0.1646836600.p
      - name: rhsso-operator.7.5.1-opr-008
        version: 7.5.1-opr-008
      - name: rhsso-operator.7.5.1-opr-007
        version: 7.5.1-opr-007
      - name: rhsso-operator.7.5.1-opr-006
        version: 7.5.1-opr-006
      - name: rhsso-operator.7.5.1-opr-005
        version: 7.5.1-opr-005
      - name: rhsso-operator.7.5.1-opr-004
        version: 7.5.1-opr-004
      - name: rhsso-operator.7.5.1-opr-003
        version: 7.5.1-opr-003
      - name: rhsso-operator.7.5.1-opr-002
        version: 7.5.1-opr-002
      - name: rhsso-operator.7.5.1-opr-001
        version: 7.5.1-opr-001
      - name: rhsso-operator.7.5.0
        version: 7.5.0
      name: stable
    defaultChannel: stable
    packageName: rhsso-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-athena-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-athena-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "athena.services.k8s.aws/v1alpha1",
                "kind": "WorkGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/athena-controller:1.1.1
          createdAt: "2025-09-19T21:06:02Z"
          description: AWS Athena controller is a service controller for managing
            Athena resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PreparedStatement represents the state of an AWS athena PreparedStatement
              resource.
            displayName: PreparedStatement
            kind: PreparedStatement
            name: preparedstatements.athena.services.k8s.aws
            version: v1alpha1
          - description: WorkGroup represents the state of an AWS athena WorkGroup
              resource.
            displayName: WorkGroup
            kind: WorkGroup
            name: workgroups.athena.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Athena resources in AWS from within your Kubernetes cluster.

          **About Amazon Athena**

          Amazon Athena is an interactive query service that makes it easy to analyze data directly in Amazon Simple Storage Service (Amazon S3) using standard SQL. With a few actions in the AWS Management Console, you can point Athena at your data stored in Amazon S3 and begin using standard SQL to run ad-hoc queries and get results in seconds.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Athena
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - athena
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Athena Developer Resources
          url: https://aws.amazon.com/athena/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: athena maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/athena-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-athena-controller.v1.1.1
        version: 1.1.1
      - name: ack-athena-controller.v1.1.0
        version: 1.1.0
      - name: ack-athena-controller.v1.0.14
        version: 1.0.14
      - name: ack-athena-controller.v1.0.13
        version: 1.0.13
      - name: ack-athena-controller.v1.0.12
        version: 1.0.12
      - name: ack-athena-controller.v1.0.11
        version: 1.0.11
      - name: ack-athena-controller.v1.0.10
        version: 1.0.10
      - name: ack-athena-controller.v1.0.9
        version: 1.0.9
      - name: ack-athena-controller.v1.0.8
        version: 1.0.8
      - name: ack-athena-controller.v1.0.7
        version: 1.0.7
      - name: ack-athena-controller.v1.0.6
        version: 1.0.6
      - name: ack-athena-controller.v1.0.5
        version: 1.0.5
      - name: ack-athena-controller.v1.0.4
        version: 1.0.4
      - name: ack-athena-controller.v1.0.1
        version: 1.0.1
      - name: ack-athena-controller.v1.0.0
        version: 1.0.0
      - name: ack-athena-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-athena-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Protocol Labs
      provider-url: ""
    name: ipfs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ipfs-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.ipfs.io/v1alpha1",
                "kind": "CircuitRelay",
                "metadata": {
                  "name": "circuitrelay-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cluster.ipfs.io/v1alpha1",
                "kind": "IpfsCluster",
                "metadata": {
                  "name": "ipfscluster-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Storage, Networking, Big Data
          containerImage: quay.io/redhat-et-ipfs/ipfs-operator:v0.0.1
          createdAt: 2022-04-05T15:50:37-0400
          operatorframework.io/suggested-namespace: ipfs-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ipfs-cluster/ipfs-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CircuitRelay is the Schema for the circuitrelays API.
            displayName: Circuit Relay
            kind: CircuitRelay
            name: circuitrelays.cluster.ipfs.io
            version: v1alpha1
          - description: IpfsCluster is the Schema for the ipfs API.
            displayName: Ipfs Cluster
            kind: IpfsCluster
            name: ipfsclusters.cluster.ipfs.io
            version: v1alpha1
        description: "The IPFS Operator enables users to easily create IPFS Cluster
          instances\nwithin their Kubernetes or OpenShift clusters. \n\n### Supported
          Features\n* Horizontal scaling\n* Joining collaborative clusters\n* Controlling
          IPFS node resource usage\n* Specifying reprovider settings  \n\n### Documentation
          \nDocumentation can be found on our [read-the-docs page](https://ipfs-operator.readthedocs.io/en/latest/),
          as well as our [GitHub repository](https://github.com/ipfs-cluster/ipfs-operator).\n\n###
          Getting Help\nIn case you run into issues while using the operator, please
          create an issue in our [GitHub\nrepository](https://github.com/ipfs-cluster/ipfs-operator).\n\n###
          Contributing\nWe welcome contributions of all types and sizes! You can find
          open issues\nin our [GitHub Repository](https://github.com/ipfs-cluster/ipfs-operator/issues),\nas
          well as make your own issue if you feel that something is missing.\n\n###
          License\nIPFS Operator is licensed under the [Apache 2.0 License](https://github.com/ipfs-cluster/ipfs-operator/blob/main/LICENSE)"
        displayName: IPFS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipfs
        - ipfs cluster
        - decentralized storage
        - peer-to-peer
        - distributed hash table
        - content-addressed
        links:
        - name: IPFS Operator GitHub page
          url: https://github.com/ipfs-cluster/ipfs-operator
        - name: IPFS Cluster
          url: https://ipfscluster.io/
        - name: IPFS
          url: https://ipfs.io/
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Protocol Labs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/redhat-et-ipfs/ipfs-operator:v0.0.1
        version: 0.0.2
      entries:
      - name: ipfs-operator.v0.0.2
        version: 0.0.2
      - name: ipfs-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ipfs-operator
    provider:
      name: Protocol Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: gaudi-ai-sw-tools-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gaudi-ai-sw-tools-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aitools.intel/v1",
                "kind": "GaudiAIToolsContainer",
                "metadata": {
                  "name": "intel-gaudi-1.17.0-495"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagespec": {
                    "registry": "registry.connect.redhat.com",
                    "repo": "intel/gaudi-notebooks",
                    "tags": [
                      {
                        "gaudi_software": "1.17.0-495",
                        "rhel_os": "9.2"
                      }
                    ]
                  },
                  "nameOverride": "",
                  "namespace": "redhat-ods-applications"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: registry.connect.redhat.com/intel/gaudi-ai-sw-tools-operator@sha256:0c286f1e4cec57a189d749501f82345afbe0a1dbbbf92b906ad3c549e47b89cd
          createdAt: "2024-09-11T21:27:02Z"
          description: Operator to enable Gaudi AI SW Tools into OpenShift AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: GaudiAIToolsContainer
            name: gaudiaitoolscontainers.aitools.intel
            version: v1
        description: Operator to enable Gaudi AI SW Tools into OpenShift AI
        displayName: Intel® Gaudi AI SW Tools Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gaudi
        - opendatahub
        - openshiftai
        - notebooks
        links:
        - name: Intel® Gaudi AI SW Tools Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/66d7aa630eb66a02febc8103
        maintainers:
        - email: sharvil.shah@intel.com
          name: sharvil
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.connect.redhat.com/intel/gaudi-ai-sw-tools-operator@sha256:0c286f1e4cec57a189d749501f82345afbe0a1dbbbf92b906ad3c549e47b89cd
        version: 0.0.2
      entries:
      - name: gaudi-ai-sw-tools-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: gaudi-ai-sw-tools-operator
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-certified-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "psp": false,
                  "serviceAccount": "default",
                  "controller": {
                    "enabled": true,
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": "registry.connect.redhat.com/neuvector/controller",
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": ""
                      }
                    },
                    "pvc": {
                      "enabled": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "path": "/",
                          "annotations": {
                            "ingress.kubernetes.io/protocol": "https"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "path": "/",
                          "annotations": {
                            "ingress.kubernetes.io/protocol": "https"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "path": "/",
                      "annotations": {
                        "ingress.kubernetes.io/protocol": "https"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {}
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/enforcer",
                    "priorityClassName": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      }
                    ],
                    "resources": {}
                  },
                  "manager": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/manager",
                    "priorityClassName": "",
                    "env": {
                      "ssl": true
                    },
                    "svc": {
                      "type": "NodePort",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "path": "/",
                      "annotations": {},
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "affinity": {},
                    "tolerations": [],
                    "nodeSelector": {}
                  },
                  "cve": {
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "image": "registry.access.redhat.com/ubi8@sha256:228824aa581f3b31bf79411f8448b798291c667a37155bdea61cfa128b2833f2",
                      "schedule": "0 0 * * *",
                      "priorityClassName": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": "registry.connect.redhat.com/neuvector/scanner@sha256:42004e13b348bb43b58ae8538c983a70b237e3aca870758ac629fb4fc43e8130",
                      "priorityClassName": "",
                      "resources": {},
                      "affinity": {},
                      "tolerations": [],
                      "nodeSelector": {}
                    }
                  },
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "resources": {},
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "true"
          containerImage: registry.connect.redhat.com/neuvector/neuvector-operator@sha256:6155ac3c95aa6471c9857e0a6f5e5d19e6fc1d8cc7d014108d7d4a10d14f96bb
          createdAt: "2020-12-02 14:19:21"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/neuvector-certified-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/neuvector-certified-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: Neuvector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm/tree/master/charts/core) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart.  NeuVector Operator versions are tied to NeuVector product versions. Version 1.3.4 of the NeuVector Certified Operator deploys version 4.4.3 of NeuVector.

          **Complete below steps to Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Grant Service Account Access to the Privileged SCC

                   oc -n neuvector adm policy add-scc-to-user privileged -z default

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s                    neuvector/default


          **Add NeuVector license from NeuVector WebUI->setting**


          #Deploying the NeuVector Operator#


          Please refer to the instructions [here](https://github.com/neuvector/neuvector-operator/blob/master/README.md)


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Networking
        - Security
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - registry.connect.redhat.com/neuvector/scanner@sha256:42004e13b348bb43b58ae8538c983a70b237e3aca870758ac629fb4fc43e8130
        - registry.access.redhat.com/ubi8@sha256:228824aa581f3b31bf79411f8448b798291c667a37155bdea61cfa128b2833f2
        - registry.connect.redhat.com/neuvector/neuvector-operator@sha256:6155ac3c95aa6471c9857e0a6f5e5d19e6fc1d8cc7d014108d7d4a10d14f96bb
        - registry.connect.redhat.com/neuvector/controller@sha256:acc943568520282061547b310c50080dc32924a531bbb543d959b281eb6664f4
        - registry.connect.redhat.com/neuvector/enforcer@sha256:be1d299ec3532d55444681a882e9fd1238949da6598987308e2800ed26291b7a
        - registry.connect.redhat.com/neuvector/manager@sha256:8245cd179d11fff279690c5a422f46b4ff252ed72b9b998f0c9b5453bc7d8fc0
        version: 1.3.4
      entries:
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-certified-operator-rhmp
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dpu-network-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dpu-network-operator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dpu.openshift.io/v1alpha1",
                "kind": "OVNKubeConfig",
                "metadata": {
                  "name": "ovnkubeconfig-sample"
                },
                "spec": {
                  "kubeConfigFile": "tenant-cluster-1-kubeconf",
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/dpu-worker": ""
                    }
                  },
                  "poolName": "dpu"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional, Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/dpu-network-rhel8-operator@sha256:d07b8ecb5d1625599878b5c1e34b0de94188b81d90c829787a2d13969f6d5971
          createdAt: "2023-07-13T18:03:11Z"
          description: The operator is responsible for the life-cycle management of
            the ovn-kube components and the necessary host network initialization
            on DPU cards.
          olm.skipRange: '>=4.10.0-0 <4.13.0-202309181427'
          operatorframework.io/suggested-namespace: openshift-dpu-network-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/dpu-network-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OVNKubeConfig is the Schema for the ovnkubeconfigs API
            displayName: OVNKube Config
            kind: OVNKubeConfig
            name: ovnkubeconfigs.dpu.openshift.io
            version: v1alpha1
        description: The operator to be responsible for the life-cycle management
          of the ovn-kube components and the necessary host network initialization
          on DPU cards.
        displayName: DPU Network Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dpu
        - network
        links:
        - name: Dpu Network Operator
          url: https://dpu-network-operator.domain
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:767682dd3bd65651a8204c9fb732b9c48b99127189992b43abc6ecce027f4589
        - registry.redhat.io/openshift4/dpu-network-rhel8-operator@sha256:d07b8ecb5d1625599878b5c1e34b0de94188b81d90c829787a2d13969f6d5971
        version: 4.13.0-202309181427
      entries:
      - name: dpu-network-operator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: stable
    defaultChannel: stable
    packageName: dpu-network-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-glue-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-glue-controller.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/glue-controller:0.2.1
          createdAt: "2025-09-19T20:55:49Z"
          description: AWS Glue controller is a service controller for managing Glue
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Job represents the state of an AWS glue Job resource.
            displayName: Job
            kind: Job
            name: jobs.glue.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Glue resources in AWS from within your Kubernetes cluster.

          **About Amazon Glue**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Glue
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - glue
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Glue Developer Resources
          url: https://aws.amazon.com/Glue/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: glue maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/glue-controller:0.2.1
        version: 0.2.1
      entries:
      - name: ack-glue-controller.v0.2.1
        version: 0.2.1
      - name: ack-glue-controller.v0.2.0
        version: 0.2.0
      - name: ack-glue-controller.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ack-glue-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fujitsu
      provider-url: https://www.postgresql.fastware.com/
    name: fujitsu-enterprise-postgres-operator-with-cg-module
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPAction",
                "metadata": {
                  "name": "new-fep-action"
                },
                "spec": {
                  "fepAction": {
                    "args": [
                      "new-fep-sts-0",
                      "new-fep-sts-1"
                    ],
                    "type": "reload"
                  },
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "targetClusterName": "new-fep"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPExporter",
                "metadata": {
                  "name": "new-fep-exporter"
                },
                "spec": {
                  "fepExporter": {
                    "exporterLogLevel": "error",
                    "fepClusterList": [
                      "new-fep1"
                    ],
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "userCustomQueries": "usr_example:\n  query: \"SELECT EXTRACT(EPOCH FROM (now() - pg_last_xact_replay_timestamp())) as lag\"\n  master: true\n  metrics:\n    - lag:\n        usage: \"GAUGE\"\n        description: \"Replication lag behind master in seconds\""
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPLogging",
                "metadata": {
                  "name": "new-fep-logging"
                },
                "spec": {
                  "fepLogging": {
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPPgpool2",
                "metadata": {
                  "name": "new-fep-pgpool2"
                },
                "spec": {
                  "clientAuthMethod": "scram",
                  "count": 2,
                  "customhba": "local   all         all                               trust\nhost    all         all         127.0.0.1/32          trust\nhost    all         all         ::1/128               trust\n",
                  "customlogsize": "128Mi",
                  "customparams": "listen_addresses = '*'\npcp_listen_addresses = '*'\nnum_init_children = 32\nreserved_connections = 0\nenable_pool_hba = off\nallow_clear_text_frontend_auth = off\nauthentication_timeout = 80\nbackend_weight0 = 1\nbackend_weight1 = 1\nbackend_flag0 = 'ALWAYS_PRIMARY'\nbackend_flag1 = 'DISALLOW_TO_FAILOVER'\nconnection_cache = on\nmax_pool = 4\nlisten_backlog_multiplier = 2\nserialize_accept = off\nchild_life_time = 300\nclient_idle_limit = 0\nchild_max_connections = 0\nconnection_life_time = 0\nreset_query_list = 'ABORT; DISCARD ALL'\nclient_min_messages = info\nlog_min_messages = debug1\nlog_statement = on\nlog_per_node_statement = on\nlog_client_messages = on\nlog_hostname = on\nlog_connections = on\nlog_line_prefix = '%t: pid %p: '\nload_balance_mode = on\nignore_leading_white_space = on\nwhite_function_list = ''\nblack_function_list = 'currval,lastval,nextval,setval'\nblack_query_pattern_list = ''\ndatabase_redirect_preference_list = ''\napp_name_redirect_preference_list = ''\nallow_sql_comments = off\ndisable_load_balance_on_write = 'transaction'\nstatement_level_load_balance = on\nsr_check_period = 0\nsr_check_user = 'postgres'\ndelay_threshold = 0\nlog_standby_delay = 'none'\nssl = on\nssl_ciphers = 'HIGH:MEDIUM:+3DES:!aNULL'\nssl_prefer_server_ciphers = off\nssl_ecdh_curve = 'prime256v1'\nssl_dh_params_file = ''\nrelcache_expire = 0\nrelcache_size = 256\ncheck_temp_table = catalog\ncheck_unlogged_table = on\nenable_shared_relcache = off\nrelcache_query_target = primary\nwd_port0 = 9000\nfailover_on_backend_error = off\n",
                  "custompcp": "none",
                  "customsslcacert": "none",
                  "customsslcert": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                  "customsslkey": "none",
                  "fepclustername": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "limits": {
                    "cpu": "400m",
                    "memory": "512Mi"
                  },
                  "requests": {
                    "cpu": "200m",
                    "memory": "256Mi"
                  },
                  "serviceport": 9999,
                  "statusport": 9898
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPRestore",
                "metadata": {
                  "name": "new-fep-restore"
                },
                "spec": {
                  "fromFEPcluster": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "mcSpec": {
                    "limits": {
                      "cpu": "200m",
                      "memory": "300Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "restoretype": "latest",
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "toFEPcluster": "new-fep-2"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v2",
                "kind": "FEPCluster",
                "metadata": {
                  "name": "new-fep"
                },
                "spec": {
                  "fep": {
                    "customAnnotations": {
                      "allDeployments": {}
                    },
                    "forceSsl": true,
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "instances": 1,
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "podAntiAffinity": false,
                    "podDisruptionBudget": false,
                    "servicePort": 27500,
                    "syncMode": "off",
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "usePodName": true
                  },
                  "fepChildCrVal": {
                    "backup": {
                      "image": {
                        "pullPolicy": "IfNotPresent"
                      },
                      "mcSpec": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      },
                      "pgbackrestParams": "# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[global]\nrepo1-retention-full=7\nrepo1-retention-full-type=time\nlog-path=/database/log/backup\n",
                      "postScript": " ",
                      "preScript": " ",
                      "schedule": {
                        "num": 2
                      },
                      "schedule1": {
                        "schedule": "15 0 * * 0",
                        "type": "full"
                      },
                      "schedule2": {
                        "schedule": "15 0 * * 1-6",
                        "type": "incr"
                      },
                      "schedule3": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule4": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule5": {
                        "schedule": " ",
                        "type": " "
                      }
                    },
                    "customPgAudit": "# define pg audit custom params here to override defaults.\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[output]\nlogger = 'auditlog'\nlog_directory = '/database/log/audit'\nlog_truncate_on_rotation = on\nlog_filename = 'pgaudit-%a.log'\nlog_rotation_age = 1d\nlog_rotation_size = 0\n[rule]\n",
                    "customPgHba": "# define pg_hba custom rules here to be merged with default rules.\n# TYPE     DATABASE        USER        ADDRESS        METHOD\n",
                    "customPgParams": "# define custom postgresql.conf parameters below to override defaults.\n# Current values are as per default FEP deployment\n# If you add a library to shared_preload_libraries, add it after the default value.\n# fsep_operator_security is available in FEP15 and later container images.\nshared_preload_libraries='pgx_datamasking,pg_prewarm,pg_stat_statements,fsep_operator_security'\nsession_preload_libraries='pg_prewarm'\nmax_prepared_transactions = 100\nmax_worker_processes = 30\nmax_connections = 100\nwork_mem = 1MB\nmaintenance_work_mem = 12MB\nshared_buffers = 128MB\neffective_cache_size = 384MB\ncheckpoint_completion_target = 0.8\n# tcp parameters\ntcp_keepalives_idle = 30\ntcp_keepalives_interval = 10\ntcp_keepalives_count = 3\n# logging parameters in default fep installation\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\nlog_directory = '/database/log'\nlog_filename = 'logfile-%a.log'\nlog_file_mode = 0600\nlog_truncate_on_rotation = on\nlog_rotation_age = 1d\nlog_rotation_size = 0\nlog_checkpoints = on\nlog_line_prefix = '%e %t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h'\nlog_lock_waits = on\nlog_autovacuum_min_duration = 60s\nlogging_collector = on\npgaudit.config_file='/opt/app-root/src/pgaudit-cfg/pgaudit.conf'\nlog_replication_commands = on\nlog_min_messages = WARNING\nlog_destination = csvlog\n# wal_archive parameters in default fep installation\narchive_mode = on\narchive_command = 'pgbackrest --stanza=backupstanza --config=/database/userdata/pgbackrest.conf archive-push %p'\nwal_level = replica\nmax_wal_senders = 12\nwal_keep_segments = 64\ntrack_activities = on\ntrack_counts = on\npassword_encryption = 'scram-sha-256'\n",
                    "storage": {
                      "archivewalVol": {
                        "size": "1Gi"
                      },
                      "backupVol": {
                        "size": "2Gi"
                      },
                      "dataVol": {
                        "size": "2Gi"
                      },
                      "logVol": {
                        "size": "1Gi"
                      },
                      "tablespaceVol": {
                        "size": "512Mi"
                      },
                      "walVol": {
                        "size": "1200Mi"
                      }
                    },
                    "sysUsers": {
                      "pgRewindPassword": "rewind_password",
                      "pgRewindUser": "rewind_user",
                      "pgdb": "mydb",
                      "pgpassword": "mydbpassword",
                      "pgreplpassword": "repluserpwd",
                      "pgrepluser": "repluser",
                      "pguser": "mydbuser",
                      "tdepassphrase": "tde-passphrase"
                    },
                    "systemCertificates": {
                      "cacrt": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                      "crt": "-----BEGIN CERTIFICATE-----\nMIIDUTCCAjmgAwIBAgIRAMocW3qMoHrD6qRvMPppMkMwDQYJKoZIhvcNAQELBQAw\nNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9yIEt1\nYmVybmV0ZXMwHhcNMjEwMjA2MDQzMjM2WhcNMjYwMjA1MDQzMjM2WjA/MRAwDgYD\nVQQKEwdGdWppdHN1MSswKQYDVQQDEyJGVUpJVFNVIEVudGVycHJpc2UgUG9zdGdy\nZXMgU2VydmVyMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA4AI33yvH\nZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I2e4SceTKi6O3C/I1XuvWlpng\n5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4ANB5JWWqDOjrRT3o7nRPGXfila\nbP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYpmjdbfxabTz07ig0+6/cwKoRR\nxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTFYvmAH7gcdssSFBt8NPlUATHE\nsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6Wdgmu5H2pDml8CDNLDv98Aj7i\n+I5SRKKcVPlnuQIDAQABo1AwTjAdBgNVHSUEFjAUBggrBgEFBQcDAQYIKwYBBQUH\nAwIwDAYDVR0TAQH/BAIwADAfBgNVHSMEGDAWgBQcwrrUO0u+FhIUuVdrDRCQRsi6\nZjANBgkqhkiG9w0BAQsFAAOCAQEAm5dxBoI9pScOCvRAchg4CprdRDSJb9K6yB3O\nnCAxnM47iHeXnY3WlnI388kHu8DU7O4ba1tJbGs3KY9KzioPk43pU12jWkO1onoF\n+mTDjx/Ef1cYWA9r5q/LtgTa6Q2sxV4O2x67QW82aAnaxO34dV5zWCPIvAoovZBV\nHRT+BgCg3r2vD1RGKK2nl1aYJtWhO1SZubam+VttdZ/vbM9oOJctxmImsEtBXjkY\nKteePdQtLL5o03JhyXWyRshCq+HMmKf2KgyY8gvydGcP4eLQdBWcW40LcnVq6UjT\n0kJycJEKngMVademq1ZWHGaiYB7hyT6GhgIcHUJ2cKrPgbEh1Q==\n-----END CERTIFICATE-----",
                      "key": "-----BEGIN RSA PRIVATE KEY-----\nMIIEowIBAAKCAQEA4AI33yvHZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I\n2e4SceTKi6O3C/I1XuvWlpng5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4AN\nB5JWWqDOjrRT3o7nRPGXfilabP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYp\nmjdbfxabTz07ig0+6/cwKoRRxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTF\nYvmAH7gcdssSFBt8NPlUATHEsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6W\ndgmu5H2pDml8CDNLDv98Aj7i+I5SRKKcVPlnuQIDAQABAoIBAFPQYKlOzw/+BA0b\nyMIUpdctIMb/54CR/xR0mVw1DbSjigNVPjHUQvB8Y1B2FAITQObgJO06bAv0QdWN\nRb0/v/yYiNJDFjaLjaIAHlO/2+oWrXbFaZqgpVDJhB+e1xaZr2x7XGxm+p925k30\nl6pvIRY+I8JRKvZiV1VZHwL/R3JOtPr++xMZtLVjVOI+f+ySqJ+TZHuAjm49EKxj\ncEmmJ28b7QcziXsvKy00f+zbqLIBKXQdZAFU5eEr1BsDRXdRW+Kf0XIvftuy4BJZ\nvoKT+VGhEvF/qysswL4+6IAO6tpuYnnM0Y2d3sOGoWPkTcQK0MekYKzL/WmtCjNs\n9hodJtECgYEA5EWyhEOf4uOKe5TDp697UCUvXLoOR58FDe/S8XNvScn29jjOkqIg\nOMoqo9xAkJTNTzqn5UUdt1x/pgM2NxlPLFijrc0zQlX3SoOO2ryDd9WNi7YKtN16\nKJqa536WeZu2OEbuAZ+S3GALVy1RPeTNPnUOmKnF06DjDUGzLNCZy10CgYEA+zfw\n952DWuz1U0Z4wvAEqqcgUKXPKrkTXV/iUnjkDkrLYVr0ZofDNTXrdHl+UedFmaOC\ncieZn6DNhcdz5tKtyysGMH3g/qs9PfoGUngvcXsy0Egk04l3x1jc8TTCLqXZXYaQ\nHMsx51n+R58oncPtzYSUOr9qQ6PbC2CstTbFJA0CgYEAjGEsUliAB/jknfEzjXjG\nPdhQUxb8VyE864Az2lah9t/kJzFyIAziAeqZ5GE7t247AGFTBRTHHI8e1Qoemi3P\nWbc9GVIbFs1lIYbcIDpUIyrKPEP8O5QEXtoNLxXTFgAjRGKiVY87spjCAJ+W2ZhO\ne/1it5GYXfgQCYQA2yuBmOUCgYANRkR2YR1axaCk+NlSu6oTdmdPu6M5x7PNQE7O\nOtMaKjua9lppvIzFGAdMDUtueoEEAE7ZR1xnwfB6PDLUpJdIYAqgr1YfPt8qkjaZ\nTv56yZ7CwL0pbF8m6nwqRrZoDp1wwraEvvvxFKFKGY/k3kCHlpTakdjEoDjn3gDi\nRnWeVQKBgCEneMSzucei5LRppRtRaJw/Btll8qlPMlX3W7dxQ3cLwpmLOn0m51Fp\nPIZ44zYK8R6fu4+/sSrlfaIg86Ugeufp6YNxyNROKxUGza5vDIu5OftwWtBeg+UK\nZ8lLWNdX6pp7WMujmF3H1DrkBbauYMUKZ4UxUYtelgHERMePIxwb\n-----END RSA PRIVATE KEY-----"
                    }
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "false"
          containerImage: quay.io/fujitsu/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:eaf601e7aef525ebcd4b769c859ed74c7a8077f843b3a75d76713eabf0295745
          createdAt: "2025-09-11T09:32:52Z"
          description: OpenShift Operator for Fujitsu Enterprise Postgres
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.1.0 <7.1.6'
          operators.openshift.io/valid-subscription: '["Fujitsu Enterprise Postgres
            Advanced Edition with Cryptographic Module Annual Subscription License
            per Core 17 Operator for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/internal-objects: '["fepautoscales.fep.fujitsu.io","fepconfigs.fep.fujitsu.io","fepusers.fep.fujitsu.io","fepcerts.fep.fujitsu.io","fepvolumes.fep.fujitsu.io","fepbackups.fep.fujitsu.io"]'
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: fj-dbaas-tec@dl.jp.fujitsu.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: FEPAction
            name: fepactions.fep.fujitsu.io
            version: v1
          - kind: FEPAutoscale
            name: fepautoscales.fep.fujitsu.io
            version: v1
          - kind: FEPBackup
            name: fepbackups.fep.fujitsu.io
            version: v1
          - kind: FEPCert
            name: fepcerts.fep.fujitsu.io
            version: v1
          - kind: FEPCluster
            name: fepclusters.fep.fujitsu.io
            version: v2
          - kind: FEPConfig
            name: fepconfigs.fep.fujitsu.io
            version: v1
          - kind: FEPExporter
            name: fepexporters.fep.fujitsu.io
            version: v1
          - kind: FEPLogging
            name: feploggings.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2Cert
            name: feppgpool2certs.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2
            name: feppgpool2s.fep.fujitsu.io
            version: v1
          - kind: FEPRestore
            name: feprestores.fep.fujitsu.io
            version: v1
          - kind: FEPUpgrade
            name: fepupgrades.fep.fujitsu.io
            version: v1
          - kind: FEPUser
            name: fepusers.fep.fujitsu.io
            version: v1
          - kind: FEPVolume
            name: fepvolumes.fep.fujitsu.io
            version: v1
        description: "Fujitsu Enterprise Postgres 17 delivers an enterprise-grade
          PostgreSQL on OpenShift Container Platform.\n\nThis solution provides the
          flexibility of a hybrid cloud solution while delivering an enhanced distribution
          of PostgreSQL to support enterprise-level workloads and provide improved
          deployment and management, availability, performance, data governance and
          security.  \n\nAvailable as a multi-architecture container built for both
          amd64.\n\nThe download and Use of the Product is strictly subject to the
          terms of the End User License Agreement with Fujitsu Limited found at https://www.fast.fujitsu.com/fujitsu-enterprise-postgres-license-agreements.
          Where the Product that has been embedded as a whole or part into a third
          party program, only Authorised Customers may download and use the Product.\n"
        displayName: Fujitsu Enterprise Postgres Operator with Cryptographic Module
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - Fujitsu
        - fep
        links:
        - name: Fujitsu Enterprise Postgres
          url: https://www.postgresql.fastware.com/
        maintainers:
        - email: fj-dbaas-tec@dl.jp.fujitsu.com
          name: Fujitsu
        maturity: stable
        provider:
          name: Fujitsu
          url: https://www.postgresql.fastware.com/
        relatedImages:
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-restore-with-cryptographic-module@sha256:edeb7b72a89e05995661e6a17e75b5b477a3ca351e0db146a939a33e88e16c97
        - quay.io/fujitsu/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:eaf601e7aef525ebcd4b769c859ed74c7a8077f843b3a75d76713eabf0295745
        - quay.io/fujitsu/fujitsu-enterprise-postgres-utils-with-cryptographic-module@sha256:7d169df0b8fe47000891afde79be8dd7978da3f893d1e92f01f57fd8c0b3b4c6
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-pgpool2-with-cryptographic-module@sha256:e510bc2d30df33e4c9d3babb898ff9cbd13704fa3fa69a284bf672765fb03083
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-server-with-cryptographic-module@sha256:7778f4b03861b56371d18f7726b68ad21d0faa7c1a25cb38fe7cac483d81c095
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-restore-with-cryptographic-module@sha256:6c0d6e10cfe1868be2986fdde14c6f55f86b12cea01a82199b1e2e0c3154ef75
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-upgrade-with-cryptographic-module@sha256:b8be0344b6cf99b2c5de2b16f04f6f2b05bde861d9b38afd319bbbc4dd72a620
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-upgrade-with-cryptographic-module@sha256:caf6e1d2cc31d01958e4adcb4c3c307d761f350d5b0963ce19214f041c14184a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-restore-with-cryptographic-module@sha256:98a3b0bfa6ec7a9a4a08fb524c331c3de6a9b580d6386a7b5430955ed84b2fc5
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-backup-with-cryptographic-module@sha256:dd15cf3b6b4a5c959efef8b602abe5d64b4f5a424fb6a3e55079a2040e67cd7a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-backup-with-cryptographic-module@sha256:03b7f3c76e3eb775e4ec7f66f7506ee1feeab56dca96e868ac3f38b9c1ca5567
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentd-with-cryptographic-module@sha256:f743f35012e6fd3d54a6b54baeeee98bd7c65fe52923acd87546066fa9b54d7e
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-upgrade-with-cryptographic-module@sha256:1fe690a9c04a55213062ad3bef8c0cff73849f8777c3945b9a8ae2dcff6330fa
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-utils-with-cryptographic-module@sha256:8b2b6a981c863f7075ced1235c614317b6526626b42405295a9c664736f0a80f
        - quay.io/fujitsu/fujitsu-enterprise-postgres-cronjob-with-cryptographic-module@sha256:809a64600396378ff8c825595d40f72aedbc2b57f394420548262b085a3f2ea5
        - quay.io/fujitsu/fujitsu-enterprise-postgres-exporter-with-cryptographic-module@sha256:71e1d1e715e2d89757e80d90b8d5b11baa729110cd5f7898df920a4c85f2bf8a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentbit-with-cryptographic-module@sha256:bb661b3aad0b825d56a9696d6d2d9f5c04add2a83e793fc3d4a73980f4a7a0ac
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-utils-with-cryptographic-module@sha256:4caf1dc5211ac03893ce9a62b417f412ffd411f13569f3ac75e12ca5128af9e4
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-pgpool2-with-cryptographic-module@sha256:a4c5a59ca18f6fa295b78d570f03991f1a4b4bd6df52f8526a1254ac1a567990
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-pgpool2-with-cryptographic-module@sha256:e8aa27cbb57503ec5f4f37beeb2e833ace08d4c583d4361b7ce945c7461a904c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-server-with-cryptographic-module@sha256:cad4e36dc2b1e5ca6bf9308d9f93958d4c8f2bbdc14e000ca0b5700fa335df44
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-backup-with-cryptographic-module@sha256:5b49aa46386ca954ebcbea28fbabd15929cb64fc2a4d2dd2796c9d2e15b212c6
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-server-with-cryptographic-module@sha256:a012f21f63729ba667af097375e08874156e2cf1b94ae203985d45f11031fea9
        version: 7.1.6
      entries:
      - name: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.6
        version: 7.1.6
      - name: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.5
        version: 7.1.5
      name: stable
    defaultChannel: stable
    packageName: fujitsu-enterprise-postgres-operator-with-cg-module
    provider:
      name: Fujitsu
      url: https://www.postgresql.fastware.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-05-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx8-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx8-operator-rhmp/support?utm_source=openshift_console
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:154f85cc6721a70909dee755ddd88897ad51038afe2538b6dee4d26e7d8b9afb
        version: 0.0.10
      entries:
      - name: joget-dx8-openshift-operator.v0.0.10
        version: 0.0.10
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-operator-rhmp
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://www.redhat.com/
    name: rhdh
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhdh-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
          createdAt: "2025-09-12T06:25:11Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.7.1'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.7/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:62f32e50727c5006766a34daa5b6472b649cd9894f3fe9543b8ecc67e6760e8e
        version: 1.7.1
      entries:
      - name: rhdh-operator.v1.7.1
        version: 1.7.1
      - name: rhdh-operator.v1.7.0
        version: 1.7.0
      - name: rhdh-operator.v1.6.5
        version: 1.6.5
      - name: rhdh-operator.v1.6.4
        version: 1.6.4
      - name: rhdh-operator.v1.6.3
        version: 1.6.3
      - name: rhdh-operator.v1.6.2
        version: 1.6.2
      - name: rhdh-operator.v1.6.1
        version: 1.6.1
      - name: rhdh-operator.v1.6.0
        version: 1.6.0
      - name: rhdh-operator.v1.5.3
        version: 1.5.3
      - name: rhdh-operator.v1.5.2
        version: 1.5.2
      - name: rhdh-operator.v1.5.1
        version: 1.5.1
      - name: rhdh-operator.v1.4.3
        version: 1.4.3
      - name: rhdh-operator.v1.4.2
        version: 1.4.2
      - name: rhdh-operator.v1.4.1
        version: 1.4.1
      - name: rhdh-operator.v1.4.0
        version: 1.4.0
      - name: rhdh-operator.v1.3.5
        version: 1.3.5
      - name: rhdh-operator.v1.3.4
        version: 1.3.4
      - name: rhdh-operator.v1.3.3
        version: 1.3.3
      - name: rhdh-operator.v1.3.1
        version: 1.3.1
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast
    - currentCSV: rhdh-operator.v1.1.2-0.1714688890.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:9bea7eabdea44342248dfd6091a8f5c3b6e65884c916b3c2a073a4a64481aa6f
          createdAt: "2023-03-20T16:11:34Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins, configuration settings, and deployment
            mechanisms, which can help streamline the process of setting up a self-managed
            internal developer portal for adopters who are just starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: rhdh-operator.v1.1.2
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.1.0 <1.2.0'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/janus-idp/operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: asoro@redhat.com
          name: Armel Soro
        - email: gazarenk@redhat.com
          name: Gennady Azarenkov
        - email: jianrzha@redhat.com
          name: Jianrong Zhang
        - email: nboldt@redhat.com
          name: Nick Boldt
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:95ed239eb427c7b1e422ece28c5277b5018efb751c2b472bf502b8216219c047
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:9bea7eabdea44342248dfd6091a8f5c3b6e65884c916b3c2a073a4a64481aa6f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:d7ae2456857cab09a9e9acf69c9990ef18deb7f920046b8c2483fbfd0d42bbfc
        - registry.redhat.io/rhel9/postgresql-15@sha256:d5f3d496ecd55688e384cc0627700a97af6924582123bcf00749c375724af3de
        version: 1.1.2+0.1714688890.p
      entries:
      - name: rhdh-operator.v1.1.2
        version: 1.1.2
      - name: rhdh-operator.v1.1.2-0.1714608881.p
        version: 1.1.2+0.1714608881.p
      - name: rhdh-operator.v1.1.2-0.1714688618.p
        version: 1.1.2+0.1714688618.p
      - name: rhdh-operator.v1.1.2-0.1714688890.p
        version: 1.1.2+0.1714688890.p
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.1
    - currentCSV: rhdh-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:c9051a6c339cd40cbc7474a33b32a6b94913af964e12ca5da0c0b7e1c6f444d5
          createdAt: "2024-11-28T17:55:19Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.2.6'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.2/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.2/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: asoro@redhat.com
          name: Armel Soro
        - email: cdaley@redhat.com
          name: Corey Daley
        - email: gazarenk@redhat.com
          name: Gennady Azarenkov
        - email: nboldt@redhat.com
          name: Nick Boldt
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:ef29f2b340a6dc52c5014239876d54a54f7c0f0bf4ea917391438a69c94a8532
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:c9051a6c339cd40cbc7474a33b32a6b94913af964e12ca5da0c0b7e1c6f444d5
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:025eb58c69b9cf5992ecf2fccb29a3d36061cbcf62652a1c6878f0c4d767c1fb
        - registry.redhat.io/rhel9/postgresql-15@sha256:24fb4e7914a6e1464d015be9e5582cc4b9da224137408bd429e7ea4f391aa198
        version: 1.2.6
      entries:
      - name: rhdh-operator.v1.2.6
        version: 1.2.6
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.2
    - currentCSV: rhdh-operator.v1.3.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:5ce579edc10bde5c1c400104f7e0d9f86ca277cd565aaf3dd84b8636f5b949a8
          createdAt: "2025-02-10T18:12:52Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.3.5'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.3/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:5ce579edc10bde5c1c400104f7e0d9f86ca277cd565aaf3dd84b8636f5b949a8
        - registry.redhat.io/rhel9/postgresql-15@sha256:24fb4e7914a6e1464d015be9e5582cc4b9da224137408bd429e7ea4f391aa198
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:c044fed8485a62a5051ac726525da546936be22f268db05491bf81582e0e6194
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b4934c45ba595b36e65c73b734c58d1517e7f43b7cfc309ca8bc853733582f76
        version: 1.3.5
      entries:
      - name: rhdh-operator.v1.3.5
        version: 1.3.5
      - name: rhdh-operator.v1.3.4
        version: 1.3.4
      - name: rhdh-operator.v1.3.3
        version: 1.3.3
      - name: rhdh-operator.v1.3.1
        version: 1.3.1
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.3
    - currentCSV: rhdh-operator.v1.4.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:ae5b52c1b35e82f3c564402e367a2a2028e16806867bba0868643089ba7ed223
          createdAt: "2025-04-03T05:42:32Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.4.3'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.4/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:ae5b52c1b35e82f3c564402e367a2a2028e16806867bba0868643089ba7ed223
        version: 1.4.3
      entries:
      - name: rhdh-operator.v1.4.3
        version: 1.4.3
      - name: rhdh-operator.v1.4.2
        version: 1.4.2
      - name: rhdh-operator.v1.4.1
        version: 1.4.1
      - name: rhdh-operator.v1.4.0
        version: 1.4.0
      name: fast-1.4
    - currentCSV: rhdh-operator.v1.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:8fe167cc649a1c53994c376f3cf3eed8c6e318741a0c08e5aaf07cb0b74e7f8a
          createdAt: "2025-06-05T14:31:21Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.5.3'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.5/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:8fe167cc649a1c53994c376f3cf3eed8c6e318741a0c08e5aaf07cb0b74e7f8a
        - registry.redhat.io/rhel9/postgresql-15@sha256:74377f5a61dd7e3e0ba5d2fa880b7107025433a026396567c79eac927b4212bc
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:2ad8c87e82df27c2b6ceba94179d1f0a40fe775368663c1b5572611888e0e02b
        version: 1.5.3
      entries:
      - name: rhdh-operator.v1.5.3
        version: 1.5.3
      - name: rhdh-operator.v1.5.2
        version: 1.5.2
      - name: rhdh-operator.v1.5.1
        version: 1.5.1
      name: fast-1.5
    - currentCSV: rhdh-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:50f20300c81621221260d03d22b1442750e3f1d2e62f3e511d63925d84c4d1f9
          createdAt: "2025-08-27T18:43:12Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.6.5'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.6/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:50f20300c81621221260d03d22b1442750e3f1d2e62f3e511d63925d84c4d1f9
        - registry.redhat.io/rhel9/postgresql-15@sha256:5e225c1646251d2509b3a6e45fad23abfb11dd5c45d3224b59178f90deb38a70
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:177a7308f61a6b1ba21bc0f7bf550a93d4478e52479e2c94a2437f2152e3d8ff
        version: 1.6.5
      entries:
      - name: rhdh-operator.v1.6.5
        version: 1.6.5
      - name: rhdh-operator.v1.6.4
        version: 1.6.4
      - name: rhdh-operator.v1.6.3
        version: 1.6.3
      - name: rhdh-operator.v1.6.2
        version: 1.6.2
      - name: rhdh-operator.v1.6.1
        version: 1.6.1
      - name: rhdh-operator.v1.6.0
        version: 1.6.0
      name: fast-1.6
    - currentCSV: rhdh-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
          createdAt: "2025-09-12T06:25:11Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.7.1'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.7/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:62f32e50727c5006766a34daa5b6472b649cd9894f3fe9543b8ecc67e6760e8e
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
        version: 1.7.1
      entries:
      - name: rhdh-operator.v1.7.1
        version: 1.7.1
      - name: rhdh-operator.v1.7.0
        version: 1.7.0
      name: fast-1.7
    defaultChannel: fast
    packageName: rhdh
    provider:
      name: Red Hat Inc.
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-odlm
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operand-deployment-lifecycle-manager.v4.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandBindInfo",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "bindings": {
                    "public": {
                      "configmap": "mongodb-configmap",
                      "secret": "mongodb-secret"
                    }
                  },
                  "description": "Binding information that should be accessible to MongoDB adopters",
                  "operand": "mongodb-atlas-kubernetes",
                  "registry": "example-service"
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "services": [
                    {
                      "name": "jaeger",
                      "spec": {
                        "jaeger": {
                          "strategy": "allinone"
                        }
                      }
                    },
                    {
                      "name": "mongodb-atlas-kubernetes",
                      "spec": {
                        "atlasDeployment": {
                          "deploymentSpec": {
                            "name": "test-deployment"
                          },
                          "projectRef": {
                            "name": "my-project"
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "operators": [
                    {
                      "channel": "stable",
                      "installMode": "cluster",
                      "name": "jaeger",
                      "namespace": "default",
                      "packageName": "jaeger",
                      "sourceName": "community-operators",
                      "sourceNamespace": "openshift-marketplace"
                    },
                    {
                      "channel": "stable",
                      "name": "mongodb-atlas-kubernetes",
                      "namespace": "default",
                      "packageName": "mongodb-atlas-kubernetes",
                      "sourceName": "community-operators",
                      "sourceNamespace": "openshift-marketplace"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandRequest",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "requests": [
                    {
                      "operands": [
                        {
                          "name": "jaeger"
                        },
                        {
                          "name": "mongodb-atlas-kubernetes"
                        }
                      ],
                      "registry": "example-service"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, Monitoring, Logging & Tracing, Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
          createdAt: "2025-02-14T21:16:49Z"
          description: The Operand Deployment Lifecycle Manager provides a Kubernetes
            CRD-based API to manage the lifecycle of operands.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-odlm
          olm.skipRange: '>=1.2.0 <4.5.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/operand-deployment-lifecycle-manager
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OperandBindInfo is the Schema for the operandbindinfoes API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandBindInfo
            kind: OperandBindInfo
            name: operandbindinfos.operator.ibm.com
            version: v1alpha1
          - description: OperandConfig is the Schema for the operandconfigs API. Documentation
              For additional details regarding install parameters check https://ibm.biz/icpfs39install.
              License By installing this product you accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandConfig
            kind: OperandConfig
            name: operandconfigs.operator.ibm.com
            version: v1alpha1
          - description: OperandRegistry is the Schema for the operandregistries API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandRegistry
            kind: OperandRegistry
            name: operandregistries.operator.ibm.com
            version: v1alpha1
          - description: OperandRequest is the Schema for the operandrequests API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandRequest
            kind: OperandRequest
            name: operandrequests.operator.ibm.com
            version: v1alpha1
          - description: OperatorConfig is the Schema for the operatorconfigs API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.ibm.com
            version: v1alpha1
        description: |-
          # Introduction

          The operand is the instance managed by the operator. ODLM is used to manage the lifecycle of for a group of operands, compared with operator lifecycle manager, ODLM focus on the management of operands but not operators.

          # Details

          ### What are operands?

          Operator is a method of packaging, deploying and managing a Kubernetes application.
          Operands are the services and applications that Operator manage.

          For example, cert-manager operator deploys a cert-manager deployment, then the cert-manager-operator is an operator and the cert-manager deployment is an operand.

          The ODLM will have four CRDs:

          | Resource                 | Short Name | Description                                                                                |
          |--------------------------|------------|--------------------------------------------------------------------------------------------|
          | OperandRequest | opreq | It defines which operator/operand want to be installed in the cluster |
          | OperandRegistry | opreg | It defines the OLM information, like channel and catalog source, for each operator |
          | OperandConfig | opcon | It defines the parameters that should be used to install the operator's operand |
          | OperandBindInfo | opbi | It identifies secrets and/or configmaps that should be shared with requests |

          ### Goal

          1. A single entry point to manage a group of operands
          1. User can select a set of operands to install
          1. The install can be invoked through either OCP UI or CLI

          ## Prerequisites
          - [operator-sdk] version v1.3.0.
          - [go] version 1.15.7+
          - [oc] version v3.11+ or [kubectl] v1.11.3+
          - Access to an Openshift v4.3+ cluster

          ### Resources Required

          #### Minimum scheduling capacity

          | Software                   | Memory (MB) | CPU (cores) |
          | -------------------------- | ----------- | ----------- |
          | Operand Deployment Lifecycle Manager | 200          | 0.2        |
          | **Total**                  | 200         | 0.2         |

          ## Installing

          ### Create Operator NS, Group, Subscription

          ```yaml
          apiVersion: v1
          kind: Namespace
          metadata:
            name: odlm

          ---
          apiVersion: operators.coreos.com/v1alpha2
          kind: OperatorGroup
          metadata:
            name: operatorgroup
            namespace: odlm
          spec:
            targetNamespaces:
            - odlm

          ---
          apiVersion: v1
          data:
            namespaces: odlm
          kind: ConfigMap
          metadata:
            name: namespace-scope
            namespace: odlm

          ---
          apiVersion: operators.coreos.com/v1alpha1
          kind: Subscription
          metadata:
            name: operand-deployment-lifecycle-manager
            namespace: odlm
          spec:
            channel: v3.20
            name: ibm-odlm
            source: opencloud-operators
            sourceNamespace: olm
            config:
              env:
              - name: INSTALL_SCOPE
                value: namespaced
          END
          ```

          ## Configuration

          OperandRegistry defines the OLM information used for installation, like package name and catalog source, for each operator.

          Following is an example of the OperandRegistry CR:

          **NOTE:** The "spec.operators[*].name" parameter must be unique for each entry.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRegistry
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            operators:
            - name: jenkins [3]
              namespace: default [4]
              channel: alpha [5]
              packageName: jenkins-operator [6]
              scope: public [7]
              sourceName: community-operators [8]
              sourceNamespace: openshift-marketplace [9]
              installMode: cluster [10]
              installPlanApproval: Manual [11]
          ```

          The OperandRegistry Custom Resource (CR) lists OLM Operator information for operands that may be requested for installation and/or access by an application that runs in a namespace. The registry CR specifies:

          1. `name` of the OperandRegistry
          2. `namespace` of the OperandRegistry
          3. `name` is the name of the operator, which should be the same as the services name in the OperandConfig and OperandRequest.
          4. `namespace` defines the namespace where the operator and its CR will be deployed. (1) When InstallMode is `cluster`, the operator will be deployed into the `openshift-operators` namespace and the operator CRs will be deployed into the namespace this parameter defines. (2) When InstallMode is empty or set to `namespace`, it is the namespace where both operator and operator CR will be deployed.
          5. `channel` is the name of OLM channel that is subscribed for the operator.
          6. `packageName` is the name of the package in CatalogSource that is subscribed for the operator.
          7. (optional) `scope` is an indicator, either public or private, that dictates whether deployment can be requested from other namespaces (public) or only from the namespace of this OperandRegistry (private). The default value is private.
          8. `sourceName` is the name of the CatalogSource.
          9. `sourceNamespace` is the namespace of the CatalogSource.
          10. (optional) `installMode` is the install mode of the operator, can be either `namespace` (OLM one namespace) or `cluster` (OLM all namespaces). The default value is `namespace`. Operator is deployed in `openshift-operators` namespace when InstallMode is set to `cluster`.
          11. (optional) `installPlanApproval` is the approval mode for emitted installplan. The default value is `Automatic`.

          ### OperandConfig Spec

          OperandConfig defines the individual operand configuration. The OperandConfig Custom Resource (CR) defines the parameters for each operator that is listed in the OperandRegistry that should be used to install the operator instance by specifying an installation CR.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          Kind: OperandConfigs
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            services:
            - name: jenkins [3]
              spec: [4]
                jenkins:
                  port: 8081
          ```

          OperandConfig defines the individual operand deployment config:

          1. `name` of the OperandConfig
          2. `namespace` of the OperandConfig
          3. `name` is the name of the operator, which should be the same as the services name in the OperandRegistry and OperandRequest.
          4. `spec` defines a map. Its key is the kind name of the custom resource. Its value is merged to the spec field of custom resource. For more details, you can check the following topic **How does ODLM create the individual operator CR?**

          #### How does Operator create the individual operator CR

          Jenkins Operator has one CRD: Jenkins:

          The OperandConfig CR has

          ```yaml
          - name: jenkins
            spec:
              jenkins:
                service:
                  port: 8081
          ```

          The Jenkins Operator CSV has

          ```yaml
          apiVersion: operators.coreos.com/v1alpha1
          kind: ClusterServiceVersion
          metadata:
            annotations:
            alm-examples: |-
              [
              {
                "apiVersion":"jenkins.io/v1alpha2",
                "kind":"Jenkins",
                "metadata": {
                  "name":"example"
                },
                "spec":{
                  ...
                  "service":{"port":8080,"type":"ClusterIP"},
                  ...
                }
              }
            ]
          ```

          The ODLM will deep merge the OperandConfig CR spec and Jenkins Operator CSV alm-examples to create the Jenkins CR.

          ```yaml
          apiVersion: jenkins.io/v1alpha2
          kind: Jenkins
          metadata:
            name: example
          spec:
            ...
            service:
              port: 8081
              type: ClusterIP
            ...
          ```

          For day2 operations, the ODLM will patch the OperandConfigs CR spec to the existing Jenkins CR.

          ### OperandRequest Spec

          OperandRequest defines which operator/operand you want to install in the cluster.

          **NOTE:** OperandRequest custom resource is used to trigger a deployment for Operators and Operands.

          #### OperandRequest sample to create custom resource via OperandConfig

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRequest
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            requests:
            - registry: example-service [3]
              registryNamespace: example-service-ns [4]
              operands: [5]
              - name: jenkins [6]
                bindings: [7]
                  public:
                    secret: jenkins-operator-credential [8]
                    configmap: jenkins-operator-base-configuration [9]
          ```

          1. `name` of the OperandRequest
          2. `namespace` of the OperandRequest
          3. `registry` identifies the name of the OperandRegistry CR from which this operand deployment is being requested.
          4. (optional) `registryNamespace` identifies the namespace in which the OperandRegistry CR is defined. **Note:** If the `registryNamespace` is not specified then it is assumed that the OperandRegistry CR is in the current (OperandRequest's) namespace.
          5. `operands` in the CR is a list of operands.
          6. `name` of operands in the CR must match a name specification in an OperandRegistry's CR.
          7. (optional) The `bindings` of the operands is a map to get and rename the secret and/or configmap from the provider and create them in the requester's namespace. If the requester wants to rename the secret and/or configmap, they need to know the key of the binding in the OperandBindInfo. If the key of the bindings map is prefixed with public, it means the secret and/or configmap can be shared with the requester in the other namespace. If the key of the bindings map is prefixed with private, it means the secret and/or configmap can only be shared within its own namespace.
          8. (optional) `secret` names a secret that should be created in the requester's namespace with formatted data that can be used to interact with the service.
          9. (optional) `configmap` names a configmap that should be created in the requester's namespace with formatted data that can be used to interact with the service.

          #### OperandRequest sample to create custom resource via OperandRequest

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRequest
          metadata:
            name: example-service
            namespace: example-service-ns
          spec:
            requests:
            - registry: example-service
              registryNamespace: example-service-ns
              operands:
              - name: jenkins
                kind: Jenkins [1]
                apiVersion: "jenkins.io/v1alpha2" [2]
                instanceName: "example" [3]
                spec: [4]
                  service:
                    port: 8081
          ```

          1. `kind` of the target custom resource. If it is set in the operand item, ODLM will create CR according to OperandRequest only and get rid of the alm-example and OperandConfig.
          2. `apiVersion` of the target custom resource.
          3. `instanceName` is the name of the custom resource. If `instanceName` is not set, the name of the custom resource will be created with the name of the OperandRequest as a prefix.
          4. `spec` is the spec field of the target CR.

          ### OperandBindInfo Spec

          The ODLM will use the OperandBindInfo to copy the generated secret and/or configmap to a requester's namespace when a service is requested with the OperandRequest CR. An example specification for an OperandBindInfo CR is shown below.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandBindInfo
          metadata:
            name: publicjenkinsbinding [1]
            namespace: example-service-ns [2]
          spec:
            operand: jenkins [3]
            registry: example-service [4]
            description: "Binding information that should be accessible to jenkins adopters" [5]
            bindings: [6]
              public:
                secret: jenkins-operator-credentials-example [7]
                configmap: jenkins-operator-base-configuration-example [8]
          ```

          Fields in this CR are described below.

          1. `name` of the OperandBindInfo
          2. `namespace` of the OperandBindInfo
          3. The `operand` should be the the individual operator name.
          4. The `registry` section must match the name in the OperandRegistry CR in the current namespace.
          5. `description` is used to add a detailed description of a service.
          6. The `bindings` section is used to specify information about the access/configuration data that is to be shared. If the key of the bindings map is prefixed with public, it means the secret and/or configmap can be shared with the requester in the other namespace. If the key of the bindings map is prefixed with private, it means the secret and/or configmap can only be shared within its own namespace. If the key of the bindings map is prefixed with protected, it means the secret and/or configmap can only be shared if it is explicitly declared in the OperandRequest.
          7. The `secret` field names an existing secret, if any, that has been created and holds information that is to be shared with the requester.
          8. The `configmap` field identifies a configmap object, if any, that should be shared with the requester

          ODLM will use the OperandBindInfo CR to pass information to an adopter when they create a OperandRequest to access the service, assuming that both have compatible scopes. ODLM will copy the information from the shared service's "OperandBindInfo.bindinfo[].secret" and/or "OperandBindInfo.bindinfo[].configmap" to the requester namespace.

          **NOTE:** If in the OperandRequest, there is no secret and/or configmap name specified in the bindings or no bindings field in the element of operands, ODLM will copy the secret and/or configmap to the requester's namespace and rename them to the name of the OperandBindInfo + secret/configmap name.

          ### E2E Use Case

          1. User installs ODLM from OLM

              The ODLM will automatically generate two default CRD CRs, since OperandRegistry and OperandConfigs CRs don't define the state, so it should be fine.

          2. **Optional:** Users can update the OperandConfigs CR with their own parameter values

              Users can skip this step if the default configurations match the requirement.

          3. User creates the OperandRequest CR from OLM

              This tells ODLM that users want to install some of the individual operator/operand.

          4. The rest works will be done by OLM and ODLM

              Finally, users will get what they want.

          ### Operator/Operand Upgrade

          - For operator/operand upgrade, you only need to publish your operator OLM to your operator channel, and OLM will handle the upgrade automatically.
          - If there are major version, then you may want to update `channel` in `OperandRegistry` to trigger upgrade.

          ### Limitations

          ODLM depends on NamespaceScope Operator

          ### SecurityContextConstraints Requirements

          The Operand Deployment Lifecycle Manager supports running under the OpenShift Container Platform default restricted security context constraints.
        displayName: Operand Deployment Lifecycle Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Operand Deployment Lifecycle Manager Project
          url: https://github.com/IBM/operand-deployment-lifecycle-manager
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d4dff1a8968c18d984e16c4c9f1517c3a10e23a11bf166c64d734773324d4c9d
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
        - registry.marketplace.redhat.com/cpopen/operand-deployment-lifecycle-manager-bundle@sha256:68f71f2767a40ca9333f0a670e07493e8a8c13a5c7ca877ea5bd8a544b352d9a
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:27195e9758afdf8521971866fd60ea3dafed91a146c3c4165876e5ea476a4622
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bb074bdf89cbc7c703fadb17639608599ab4f66db9be5617e5f79a0550676790
        version: 4.5.0
      entries:
      - name: operand-deployment-lifecycle-manager.v4.5.0
        version: 4.5.0
      name: v4.5
    defaultChannel: v4.5
    packageName: ibm-odlm
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: egressip-ipam-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: egressip-ipam-operator.v1.2.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "EgressIPAM",
                "metadata": {
                  "name": "egressipam-aws"
                },
                "spec": {
                  "cidrAssignments": [
                    {
                      "CIDR": "10.0.128.0/19",
                      "labelValue": "us-west-2a"
                    },
                    {
                      "CIDR": "10.0.160.0/19",
                      "labelValue": "us-west-2b",
                      "reservedIPs": [
                        "10.0.160.126"
                      ]
                    },
                    {
                      "CIDR": "10.0.192.0/19",
                      "labelValue": "us-west-2c"
                    }
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "topologyLabel": "failure-domain.beta.kubernetes.io/zone"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "EgressIPAM",
                "metadata": {
                  "name": "egressipam-baremetal"
                },
                "spec": {
                  "cidrAssignments": [
                    {
                      "CIDR": "192.169.0.0/24",
                      "labelValue": "true"
                    }
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "topologyLabel": "egressGateway"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Networking, Cloud Provider, Security
          certified: "false"
          containerImage: quay.io/redhat-cop/egressip-ipam-operator@sha256:d6129bd287b87696a23e0da64dcd598958f779dc88ae33b46b3c909dbab80adf
          createdAt: "2022-07-25T15:07:22Z"
          description: This operator provides automation around managing egressIPs.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: egressip-ipam-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/egressip-ipam-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgressIPAM is the Schema for the egressipams API
            displayName: Egress IPAM
            kind: EgressIPAM
            name: egressipams.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator automates the assignment of egressIPs to namespaces.
          Namespaces can opt in to receiving one or more egressIPs with the following annotation `egressip-ipam-operator.redhat-cop.io/egressipam:<egressIPAM>`, where `egressIPAM` is the CRD that controls how egressIPs are assigned.
          IPs assigned to the namespace can be looked up in the following annotation: `egressip-ipam-operator.redhat-cop.io/egressips`.

          EgressIP assignments is managed by the EgressIPAM CRD. Here is an example of it:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: EgressIPAM
          metadata:
            name: example-egressipam
          spec:
            cidrAssignments:
              - labelValue: "true"
                CIDR: 192.169.0.0/24
                reservedIPs:
                - "192.159.0.5"
            topologyLabel: egressGateway
            nodeSelector:
              node-role.kubernetes.io/worker: ""
          ```

          This EgressIPAM specifies that all nodes that comply with the specified node selector and that also have labels `egressGateway: "true"` will be assigned egressIP from the specified CIDR.

          Note that the `cidrAssigments` field is an array and therefore, multiple groups of nodes can be identified with the labelValue and different CIDRs can be assigned to them. This is usually not necessary on a bare metal deployment.

          In the bare metal scenario, when this egressCRD is created, all the `hostsubnet` relative to the nodes selected by this EgressIPAM will be update to have the EgressCIDRs field equal to the specified CIDR (see below for cloud provider scenarios).

          When a namespace is created with the opt-in annotation: `egressip-ipam-operator.redhat-cop.io/egressipam=<egressIPAM>`, an available egressIP is selected from the CIDR and assigned to the namespace.The `netnamespace` associated with this namespace is updated to use that egressIP.

          It is possible to specify a set of reserved IPs. These IPs must belong to the CIDR and will never be assigned.

          ## Passing EgressIPs as input

          The normal mode of operation of this operator is to pick a random IP from the configured CIDR. However, it also supports a scenario where egressIPs are picked by an external process and passed as input.

          In this case IPs must me passed as an annotation to the namespace, like this: `egressip-ipam-operator.redhat-cop.io/egressips=IP1,IP2...`. The value of the annotation is a comma separated array of ip with no spaces.

          There must be exactly one IP per CIDR defined in the referenced egressIPAM. Moreover, each IP must belong to the corresponding CIDR. Because this situation can lead to inconsistencies, failing to have correspondence between IPs in the namespace annotation and CIDRs in the egressIPAM CR will cause the operator to error out and stop processing all the namespaces associated with the given EgressIPAM CR.

          It also also a responsibility of the progress picking the IPs to ensure that those IPs are actually available.

          ## Assumptions

          The following assumptions apply when using this operator:

          1. If multiple EgressIPAMs are defined, their selected nodes MUST NOT overlap.
          2. When using a cloud provider the topology label MUST be `failure-domain.beta.kubernetes.io/zone`.

          ## Support for AWS

          In AWS as well as other cloud providers, one cannot freely assign IPs to machines. Additional steps need to be performed in this case. Considering this EgressIPAM

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: EgressIPAM
          metadata:
            name: egressipam-aws
          spec:
            cidrAssignments:
              - labelValue: "eu-central-1a"
                CIDR: 10.0.128.0/20
              - labelValue: "eu-central-1b"
                CIDR: 10.0.144.0/20
              - labelValue: "eu-central-1c"
                CIDR: 10.0.160.0/20
            topologyLabel: failure-domain.beta.kubernetes.io/zone
            nodeSelector:
              matchLabels:
                node-role.kubernetes.io/worker: ""
          ```

          When a namespace with the opt-in annotation is created, the following happens:

          1. for each of the CIDRs, an available IP is selected and assigned to the namespace
          2. the relative `netnamespace` is update to reflect the assignment (multiple IPs will be assigned in this case).
          3. one node per zone is selected to carry the egressIP
          4. the relative aws machines are assigned the additional IP on the main interface (support for secondary interfaces in not available)
          5. the relative `hostsubnets` are updated to reflect the assigned IP, the `egressIP` field is updated.
        displayName: Egressip Ipam Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - egressIPs
        - high availability
        - firewall support
        - traffic management
        - ipam
        links:
        - name: Egressip Ipam Operator
          url: https://github.com/redhat-cop/egressip-ipam-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/egressip-ipam-operator:v0.1.2
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/egressip-ipam-operator@sha256:d6129bd287b87696a23e0da64dcd598958f779dc88ae33b46b3c909dbab80adf
        version: 1.2.4
      entries:
      - name: egressip-ipam-operator.v1.2.4
        version: 1.2.4
      - name: egressip-ipam-operator.v1.2.3
        version: 1.2.3
      - name: egressip-ipam-operator.v1.2.2
        version: 1.2.2
      - name: egressip-ipam-operator.v1.2.1
        version: 1.2.1
      - name: egressip-ipam-operator.v1.2.0
        version: 1.2.0
      - name: egressip-ipam-operator.v1.1.0
        version: 1.1.0
      - name: egressip-ipam-operator.v1.0.8
        version: 1.0.8
      - name: egressip-ipam-operator.v1.0.7
        version: 1.0.7
      - name: egressip-ipam-operator.v1.0.6
        version: 1.0.6
      - name: egressip-ipam-operator.v1.0.5
        version: 1.0.5
      - name: egressip-ipam-operator.v1.0.4
        version: 1.0.4
      - name: egressip-ipam-operator.v1.0.3
        version: 1.0.3
      - name: egressip-ipam-operator.v1.0.2
        version: 1.0.2
      - name: egressip-ipam-operator.v1.0.1
        version: 1.0.1
      - name: egressip-ipam-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: egressip-ipam-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: hive-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hive-operator.v1.2.4953-c248a51
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:c248a51ae1
          createdAt: "2025-09-22T13:41:19Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeploymentCustomization is the Schema for clusterdeploymentcustomizations
              API.
            displayName: ClusterDeploymentCustomization
            kind: ClusterDeploymentCustomization
            name: clusterdeploymentcustomizations.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: |-
              ClusterPool represents a pool of clusters that should be kept ready to be given out to users. Clusters are removed
              from the pool once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: |-
              MachinePoolNameLease is the Schema for the MachinePoolNameLeases API. This resource is mostly empty
              as we're primarily relying on the name to determine if a lease is available.
              Note that not all cloud providers require the use of a lease for naming, at present this
              is only required for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * IBM Cloud
          * Red Hat OpenStack
          * vSphere

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Troubleshooting](https://github.com/openshift/hive/blob/master/docs/troubleshooting.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
              logLevel: debug
              targetNamespace: hive
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:c248a51ae1
        version: 1.2.4953-c248a51
      entries:
      - name: hive-operator.v1.2.4953-c248a51
        version: 1.2.4953-c248a51
      - name: hive-operator.v1.2.4928-76a65b6
        version: 1.2.4928-76a65b6
      - name: hive-operator.v1.2.4902-4929f16
        version: 1.2.4902-4929f16
      - name: hive-operator.v1.2.4896-a29a238
        version: 1.2.4896-a29a238
      - name: hive-operator.v1.2.4868-45ae912
        version: 1.2.4868-45ae912
      - name: hive-operator.v1.2.4863-e9ad0fa
        version: 1.2.4863-e9ad0fa
      - name: hive-operator.v1.2.4774-87bff59
        version: 1.2.4774-87bff59
      - name: hive-operator.v1.2.4766-6a82857
        version: 1.2.4766-6a82857
      - name: hive-operator.v1.2.4746-d25e0db
        version: 1.2.4746-d25e0db
      - name: hive-operator.v1.2.4738-b19773e
        version: 1.2.4738-b19773e
      - name: hive-operator.v1.2.4731-41f153f
        version: 1.2.4731-41f153f
      - name: hive-operator.v1.2.4703-02f3f00
        version: 1.2.4703-02f3f00
      - name: hive-operator.v1.2.4693-ec3bc72
        version: 1.2.4693-ec3bc72
      - name: hive-operator.v1.2.4686-553c6c2
        version: 1.2.4686-553c6c2
      - name: hive-operator.v1.2.4682-406d6ca
        version: 1.2.4682-406d6ca
      - name: hive-operator.v1.2.4672-af54e2f
        version: 1.2.4672-af54e2f
      - name: hive-operator.v1.2.4668-3ccd44d
        version: 1.2.4668-3ccd44d
      - name: hive-operator.v1.2.4664-00f84cc
        version: 1.2.4664-00f84cc
      - name: hive-operator.v1.2.4658-c13e207
        version: 1.2.4658-c13e207
      - name: hive-operator.v1.2.4654-3268a18
        version: 1.2.4654-3268a18
      - name: hive-operator.v1.2.4650-531d549
        version: 1.2.4650-531d549
      - name: hive-operator.v1.2.4644-365f074
        version: 1.2.4644-365f074
      - name: hive-operator.v1.2.4638-56d6d64
        version: 1.2.4638-56d6d64
      - name: hive-operator.v1.2.4625-389d717
        version: 1.2.4625-389d717
      - name: hive-operator.v1.2.4615-52100dd
        version: 1.2.4615-52100dd
      - name: hive-operator.v1.2.4605-b41ca3f
        version: 1.2.4605-b41ca3f
      - name: hive-operator.v1.2.4584-95efcb7
        version: 1.2.4584-95efcb7
      - name: hive-operator.v1.2.4574-1ec27ad
        version: 1.2.4574-1ec27ad
      - name: hive-operator.v1.2.4568-b3eca52
        version: 1.2.4568-b3eca52
      - name: hive-operator.v1.2.4564-bdf9d08
        version: 1.2.4564-bdf9d08
      - name: hive-operator.v1.2.4552-bf2d689
        version: 1.2.4552-bf2d689
      - name: hive-operator.v1.2.4532-cddd4f9
        version: 1.2.4532-cddd4f9
      - name: hive-operator.v1.2.4511-1fce770
        version: 1.2.4511-1fce770
      - name: hive-operator.v1.2.4500-b1ac27b
        version: 1.2.4500-b1ac27b
      - name: hive-operator.v1.2.4485-47df0d9
        version: 1.2.4485-47df0d9
      - name: hive-operator.v1.2.4469-7be6433
        version: 1.2.4469-7be6433
      - name: hive-operator.v1.2.4463-d7ead60
        version: 1.2.4463-d7ead60
      - name: hive-operator.v1.2.4461-97686bb
        version: 1.2.4461-97686bb
      - name: hive-operator.v1.2.4459-f910c9b
        version: 1.2.4459-f910c9b
      - name: hive-operator.v1.2.4451-83aedb9
        version: 1.2.4451-83aedb9
      - name: hive-operator.v1.2.4446-e11a23b
        version: 1.2.4446-e11a23b
      - name: hive-operator.v1.2.4441-fdbfaf0
        version: 1.2.4441-fdbfaf0
      - name: hive-operator.v1.2.4436-44323bc
        version: 1.2.4436-44323bc
      - name: hive-operator.v1.2.4428-ceaafa8
        version: 1.2.4428-ceaafa8
      - name: hive-operator.v1.2.4426-fb27114
        version: 1.2.4426-fb27114
      - name: hive-operator.v1.2.4420-fe0e302
        version: 1.2.4420-fe0e302
      - name: hive-operator.v1.2.4418-eb5403f
        version: 1.2.4418-eb5403f
      - name: hive-operator.v1.2.4416-fba8d52
        version: 1.2.4416-fba8d52
      - name: hive-operator.v1.2.4408-9c5806a
        version: 1.2.4408-9c5806a
      - name: hive-operator.v1.2.4404-410fed9
        version: 1.2.4404-410fed9
      - name: hive-operator.v1.2.4400-8c54fc9
        version: 1.2.4400-8c54fc9
      - name: hive-operator.v1.2.4386-6f1fd3e
        version: 1.2.4386-6f1fd3e
      - name: hive-operator.v1.2.4382-0e95d5b
        version: 1.2.4382-0e95d5b
      - name: hive-operator.v1.2.4366-df65e7e
        version: 1.2.4366-df65e7e
      - name: hive-operator.v1.2.4360-de169dd
        version: 1.2.4360-de169dd
      - name: hive-operator.v1.2.4355-e5f809f
        version: 1.2.4355-e5f809f
      name: alpha
    - currentCSV: hive-operator.v2.5.3516-a2ed9b3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.5.3516-a2ed9b3
          createdAt: "2022-06-03T13:24:12Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:v2.5.3516-a2ed9b3
        version: 2.5.3516-a2ed9b3
      entries:
      - name: hive-operator.v2.5.3516-a2ed9b3
        version: 2.5.3516-a2ed9b3
      - name: hive-operator.v2.5.3514-4052426
        version: 2.5.3514-4052426
      - name: hive-operator.v2.5.3508-6cb94c6
        version: 2.5.3508-6cb94c6
      - name: hive-operator.v2.5.3506-f42617c
        version: 2.5.3506-f42617c
      - name: hive-operator.v2.5.3495-5cfdd44
        version: 2.5.3495-5cfdd44
      - name: hive-operator.v2.5.3491-6feded1
        version: 2.5.3491-6feded1
      - name: hive-operator.v2.5.3489-6dec9c3
        version: 2.5.3489-6dec9c3
      name: mce-2.0
    - currentCSV: hive-operator.v2.3.3039-60b8a9a
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.3.3039-60b8a9a
          createdAt: "2022-08-15T17:23:39Z"
          description: OpenShift cluster provisioning and management at scale.
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:v2.3.3039-60b8a9a
        version: 2.3.3039-60b8a9a
      entries:
      - name: hive-operator.v2.3.3039-60b8a9a
        version: 2.3.3039-60b8a9a
      name: ocm-2.3
    - currentCSV: hive-operator.v2.4.3231-ba51985
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.4.3231-ba51985
          createdAt: "2022-08-15T11:53:28Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:v2.4.3231-ba51985
        version: 2.4.3231-ba51985
      entries:
      - name: hive-operator.v2.4.3231-ba51985
        version: 2.4.3231-ba51985
      - name: hive-operator.v2.4.3227-74831db
        version: 2.4.3227-74831db
      - name: hive-operator.v2.4.3222-d68346c
        version: 2.4.3222-d68346c
      - name: hive-operator.v2.4.3206-b0714a7
        version: 2.4.3206-b0714a7
      - name: hive-operator.v2.4.3203-672cfe7
        version: 2.4.3203-672cfe7
      - name: hive-operator.v2.4.3201-8c50440
        version: 2.4.3201-8c50440
      - name: hive-operator.v2.4.3199-cce4cb9
        version: 2.4.3199-cce4cb9
      - name: hive-operator.v2.4.3197-6c9c8e4
        version: 2.4.3197-6c9c8e4
      - name: hive-operator.v2.4.3194-f82b7d8
        version: 2.4.3194-f82b7d8
      - name: hive-operator.v2.4.3192-fae0111
        version: 2.4.3192-fae0111
      - name: hive-operator.v2.4.3189-026033b
        version: 2.4.3189-026033b
      - name: hive-operator.v2.4.3180-7299056
        version: 2.4.3180-7299056
      - name: hive-operator.v1.2.3179-0756b70e6
        version: 1.2.3179-0756b70e6
      - name: hive-operator.v1.1.16
        version: 1.1.16
      - name: hive-operator.v1.1.15
        version: 1.1.15
      - name: hive-operator.v1.1.14
        version: 1.1.14
      - name: hive-operator.v1.1.13
        version: 1.1.13
      - name: hive-operator.v1.1.12
        version: 1.1.12
      - name: hive-operator.v1.1.11
        version: 1.1.11
      - name: hive-operator.v1.1.10
        version: 1.1.10
      - name: hive-operator.v1.1.9
        version: 1.1.9
      - name: hive-operator.v1.1.8
        version: 1.1.8
      - name: hive-operator.v1.1.7
        version: 1.1.7
      - name: hive-operator.v1.1.6
        version: 1.1.6
      name: ocm-2.4
    defaultChannel: alpha
    packageName: hive-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-admin-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-admin-operator.v2024.10.2-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": null
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:ed279ff5416cce855118f66a73d55385aba26af19735eed052010ba653f64b75
          createdAt: "2024-10-25T01:47:18Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2024.10.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2023.9.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:7c4868db021beaffa762eb0de43739bbc865a4bf5900f1d17f78caed8612c073
        - registry.connect.redhat.com/ca/uma-operator@sha256:ed279ff5416cce855118f66a73d55385aba26af19735eed052010ba653f64b75
        version: 2024.10.2-12
      entries:
      - name: uma-admin-operator.v2024.10.2-12
        version: 2024.10.2-12
      name: "2024.10"
    - currentCSV: uma-admin-operator.v2025.1.3-33
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.1.1.33,990033",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": null
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:6a932a065add9f36f9615a93037315603f2d9cd8a2e3c64b1b1fc390bb4f1dcf
          createdAt: "2025-02-18T05:34:34Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.1.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2023.9.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.1.1.26,990026"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:6a932a065add9f36f9615a93037315603f2d9cd8a2e3c64b1b1fc390bb4f1dcf
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:ae91a28d57c8a359f81757e4bf79472f2ac189ac0ae3ee0d87e8ff9c433f1ccc
        version: 2025.1.3-33
      entries:
      - name: uma-admin-operator.v2025.1.3-33
        version: 2025.1.3-33
      name: "2025.1"
    - currentCSV: uma-admin-operator.v2025.2.1-25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.2.1.25,990025",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverCollectorCert": "",
                    "serverCollectorKey": "",
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:d201790c2d9ded6bc7a1c5f68a267574e7b5d39c39ce2f98e20ef630b0f7365f
          createdAt: "2025-03-05T05:51:02Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.2.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2023.9.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.1.1.26,990026"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:d201790c2d9ded6bc7a1c5f68a267574e7b5d39c39ce2f98e20ef630b0f7365f
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:8bcf0dac889f38a0bf77eb9babe3f726e3b2564d5988c05bdc5c2d91ae189c61
        version: 2025.2.1-25
      entries:
      - name: uma-admin-operator.v2025.2.1-25
        version: 2025.2.1-25
      name: "2025.2"
    - currentCSV: uma-admin-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:4c2a7f68a47428f3e386107254f74fa8c34e477eb40fc17c173df92f3c66c559
          createdAt: "2025-04-16T06:55:39Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.4.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.4.2.5,990005"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        - registry.connect.redhat.com/ca/uma-operator@sha256:4c2a7f68a47428f3e386107254f74fa8c34e477eb40fc17c173df92f3c66c559
        version: 2025.4.5-5
      entries:
      - name: uma-admin-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-admin-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:514b953eaf0959bc120f80bdc0a2e19d8f059d3fe2883a4f6843a398470a56b1
          createdAt: "2025-07-24T14:54:21Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.5.2.12,990012"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:514b953eaf0959bc120f80bdc0a2e19d8f059d3fe2883a4f6843a398470a56b1
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        version: 2025.5.4-12
      entries:
      - name: uma-admin-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-admin-operator.v2025.7.3-21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:188a2b8d2255ded6141a4b795f1b6d7bd98017b4fd57c9c1a1d38fc701358057
          createdAt: "2025-08-23T09:00:37Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.3
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.7.1.23,990023"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:188a2b8d2255ded6141a4b795f1b6d7bd98017b4fd57c9c1a1d38fc701358057
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.3-21
      entries:
      - name: uma-admin-operator.v2025.7.3-21
        version: 2025.7.3-21
      name: "2025.7"
    - currentCSV: uma-admin-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:0cfb63c62258e54d67e7baca874003a04477f30bcf7f56861f565ad068070580
          createdAt: "2025-08-15T03:20:28Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-admin-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.8.1.9,990009"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:0cfb63c62258e54d67e7baca874003a04477f30bcf7f56861f565ad068070580
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        version: 2025.8.2-9
      entries:
      - name: uma-admin-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-admin-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
          createdAt: "2025-09-24T12:23:56Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.12,990012"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-admin-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-admin-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:7b93de818fa5160e72f878b0350d13fc1d5943eac95b0794d3d4f930cf21b59a
          createdAt: "2025-07-30T08:07:39Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-admin-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "99.99.0.serp_F152813,19"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:7b93de818fa5160e72f878b0350d13fc1d5943eac95b0794d3d4f930cf21b59a
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        version: 2025.7.2-19
      entries:
      - name: uma-admin-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-admin-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1cf3909dd2caeabc53209d6f8565dcb6c3562b9c6eb715dda8ba1953f264668e
          createdAt: "2025-09-11T11:16:06Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.8,990008"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:1cf3909dd2caeabc53209d6f8565dcb6c3562b9c6eb715dda8ba1953f264668e
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-admin-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-admin-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
          createdAt: "2025-09-24T12:23:56Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.12,990012"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-admin-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-admin-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: octavia-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: octavia-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "octavia.openstack.org/v1beta1",
                "kind": "Octavia",
                "metadata": {
                  "name": "octavia"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "octavia",
                  "databaseInstance": "openstack",
                  "lbMgmtNetwork": {
                    "availabilityZones": [
                      "zone-1"
                    ]
                  },
                  "octaviaAPI": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "octavia",
                    "databaseInstance": "openstack",
                    "persistenceDatabaseAccount": "octavia-persistence",
                    "preserveJobs": false,
                    "secret": "osp-secret",
                    "serviceAccount": "octavia",
                    "serviceUser": "octavia"
                  },
                  "octaviaHealthManager": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "octavia",
                    "databaseInstance": "openstack",
                    "networkAttachments": [
                      "octavia"
                    ],
                    "persistenceDatabaseAccount": "octavia-persistence",
                    "preserveJobs": false,
                    "role": "healthmanager",
                    "secret": "osp-secret",
                    "serviceAccount": "octavia",
                    "serviceUser": "octavia"
                  },
                  "octaviaHousekeeping": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "octavia",
                    "databaseInstance": "openstack",
                    "networkAttachments": [
                      "octavia"
                    ],
                    "persistenceDatabaseAccount": "octavia-persistence",
                    "preserveJobs": false,
                    "role": "housekeeping",
                    "secret": "osp-secret",
                    "serviceAccount": "octavia",
                    "serviceUser": "octavia"
                  },
                  "octaviaWorker": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "octavia",
                    "databaseInstance": "openstack",
                    "networkAttachments": [
                      "octavia"
                    ],
                    "persistenceDatabaseAccount": "octavia-persistence",
                    "preserveJobs": false,
                    "role": "worker",
                    "secret": "osp-secret",
                    "serviceAccount": "octavia",
                    "serviceUser": "octavia"
                  },
                  "persistenceDatabaseAccount": "octavia-persistence",
                  "preserveJobs": false,
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "osp-secret",
                  "serviceUser": "octavia"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:01Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: Octavia Amphora Controller
            kind: OctaviaAmphoraController
            name: octaviaamphoracontrollers.octavia.openstack.org
            version: v1beta1
          - description: OctaviaAPI is the Schema for the octaviaapis API
            displayName: Octavia API
            kind: OctaviaAPI
            name: octaviaapis.octavia.openstack.org
            version: v1beta1
          - kind: OctaviaRsyslog
            name: octaviarsyslogs.octavia.openstack.org
            version: v1beta1
          - description: Octavia is the Schema for the octavia API
            displayName: Octavia
            kind: Octavia
            name: octavias.octavia.openstack.org
            version: v1beta1
        description: Octavia Operator
        displayName: Octavia Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Load-Balancer
        - Octavia
        links:
        - name: Octavia Operator
          url: https://github.com/openstack-k8s-operators/octavia-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/octavia-rhel9-operator@sha256:ddd280ec099cde63e38bd964d7f0608c1c34e5c9f6b140f86dde2126b0357ef5
        - registry.redhat.io/rhoso/openstack-octavia-health-manager-rhel9@sha256:7d305af6e9e6c636510186dd80037849d19bb1e6d2b2507c025d1d0d9254874a
        - registry.redhat.io/rhoso/openstack-octavia-worker-rhel9@sha256:0919ddff6421430f8ad24cb21a89c65f02f98df424267f7e436c97d555112e2a
        - registry.redhat.io/ubi9/httpd-24@sha256:dbfbe3383e12d63df7178e4cdd5b1e6e51a84c781041b9192026ea6edb72fbf3
        - registry.redhat.io/rhoso/openstack-octavia-api-rhel9@sha256:ae9aae94c2523933a346bdade16158b4302b3a0bee93dec89bae5d15606c8dfb
        - registry.redhat.io/rhoso/openstack-octavia-housekeeping-rhel9@sha256:d54868077feb205199b76e83fc9220d9b4215dbeba037cf13fb68dc6577a07dc
        version: 1.0.6
      entries:
      - name: octavia-operator.v1.0.6
        version: 1.0.6
      - name: octavia-operator.v1.0.5
        version: 1.0.5
      - name: octavia-operator.v1.0.4
        version: 1.0.4
      - name: octavia-operator.v1.0.3
        version: 1.0.3
      - name: octavia-operator.v1.0.2
        version: 1.0.2
      - name: octavia-operator.v1.0.1
        version: 1.0.1
      - name: octavia-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: octavia-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: gatekeeper-operator-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gatekeeper-operator-product.v3.11.2-0.1725401426.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel8-operator@sha256:c8894e4d02fb2c822ab2e02e489dd9b67c953d96bd3b96a8305219a2245dfd99
          createdAt: 21 Mar 2024, 01:59
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.11.0
          olm.substitutesFor: gatekeeper-operator-product.v3.11.2-0.1721233953.p
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/gatekeeper/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the Schema for the gatekeepers API
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |-
          Gatekeeper allows administrators to detect and reject non-compliant commits to an infrastructure-as-code system's source-of-truth. This strengthens compliance efforts and prevents a bad state from slowing down the organization.

          ## Support & Troubleshooting
          Gatekeeper is available and supported as part of a valid Red Hat software subscription when it is used with Red Hat Advanced Cluster Management for Kubernetes. Troubleshooting information is available in the Gatekeeper documentation.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/gatekeeper/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - registry.redhat.io/gatekeeper/gatekeeper-rhel8-operator@sha256:c8894e4d02fb2c822ab2e02e489dd9b67c953d96bd3b96a8305219a2245dfd99
        - registry.redhat.io/gatekeeper/gatekeeper-rhel8@sha256:701b165def9e4b92d91b8eac4917721e76e634d2159b8efc69bef23e10b51b9d
        version: 3.11.2+0.1725401426.p
      entries:
      - name: gatekeeper-operator-product.v3.11.2-0.1725401426.p
        version: 3.11.2+0.1725401426.p
      - name: gatekeeper-operator-product.v3.11.2-0.1721233953.p
        version: 3.11.2+0.1721233953.p
      - name: gatekeeper-operator-product.v3.11.2-0.1718224960.p
        version: 3.11.2+0.1718224960.p
      - name: gatekeeper-operator-product.v3.11.2
        version: 3.11.2
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.11"
    - currentCSV: gatekeeper-operator-product.v3.14.3-0.1746550072.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e47f0e9da800eb0cf4a0da6af98986508a35798a41581158ee7dbc977cb651dc
          createdAt: 16 Jan 2025, 06:57
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.14.3
          olm.substitutesFor: gatekeeper-operator-product.v3.14.3-0.1744033158.p
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the Schema for the gatekeepers API
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |-
          Gatekeeper allows administrators to detect and reject non-compliant commits to an infrastructure-as-code system's source-of-truth. This strengthens compliance efforts and prevents a bad state from slowing down the organization.

          ## Support & Troubleshooting
          Gatekeeper is available and supported as part of a valid Red Hat software subscription when it is used with Red Hat Advanced Cluster Management for Kubernetes. Troubleshooting information is available in the Gatekeeper documentation.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e47f0e9da800eb0cf4a0da6af98986508a35798a41581158ee7dbc977cb651dc
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:6255f7a673e6989617176842d8f0cc897e4630b5283e812a7a2d52dbb2a1f335
        version: 3.14.3+0.1746550072.p
      entries:
      - name: gatekeeper-operator-product.v3.14.3-0.1746550072.p
        version: 3.14.3+0.1746550072.p
      - name: gatekeeper-operator-product.v3.14.3-0.1744033158.p
        version: 3.14.3+0.1744033158.p
      - name: gatekeeper-operator-product.v3.14.3-0.1742934403.p
        version: 3.14.3+0.1742934403.p
      - name: gatekeeper-operator-product.v3.14.3-0.1740676608.p
        version: 3.14.3+0.1740676608.p
      - name: gatekeeper-operator-product.v3.14.3
        version: 3.14.3
      - name: gatekeeper-operator-product.v3.14.2
        version: 3.14.2
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.14"
    - currentCSV: gatekeeper-operator-product.v3.15.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:8b43248aed06563e06f7a0347ff1c45fa5892e3a9adb2b5c363962a0b152eff0
          createdAt: 13 Mar 2025, 17:07
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.15.4
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: Gatekeeper allows administrators to detect and reject non-compliant
          commits to an infrastructure-as-code system's source-of-truth. This strengthens
          compliance efforts and prevents a bad state from slowing down the organization.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:8b43248aed06563e06f7a0347ff1c45fa5892e3a9adb2b5c363962a0b152eff0
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:29690144667b3753a7e76095e136d6fc42325270e7b9f22f97dfd411439e9f54
        version: 3.15.4
      entries:
      - name: gatekeeper-operator-product.v3.15.4
        version: 3.15.4
      - name: gatekeeper-operator-product.v3.15.3
        version: 3.15.3
      - name: gatekeeper-operator-product.v3.15.2
        version: 3.15.2
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.15"
    - currentCSV: gatekeeper-operator-product.v3.17.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:3c027ec78398df0a8fc09eb79e222a9508b4f18154d1e1b477d4f156181ae51a
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.17.3
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e44078fc6073a79168ed52922b370773b451d69bf335aeb91d26a1462df21a9b
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:3c027ec78398df0a8fc09eb79e222a9508b4f18154d1e1b477d4f156181ae51a
        version: 3.17.3
      entries:
      - name: gatekeeper-operator-product.v3.17.3
        version: 3.17.3
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.17"
    - currentCSV: gatekeeper-operator-product.v3.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:0803b7219ceff7aa676488eeba58607a19c9c18a2bb8eae8b3b16cfec411cb6e
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.18.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:49726cb2702441305e8047b24054575170fb595cfe5ebf815cdbfdaa957b02f4
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:0803b7219ceff7aa676488eeba58607a19c9c18a2bb8eae8b3b16cfec411cb6e
        version: 3.18.1
      entries:
      - name: gatekeeper-operator-product.v3.18.1
        version: 3.18.1
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.18"
    - currentCSV: gatekeeper-operator-product.v3.19.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.19.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:60ee347425c6cfb36f0e1e834b33cbb3d77aa2d6c89ac513216d0b9fc55d4b52
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
        version: 3.19.1
      entries:
      - name: gatekeeper-operator-product.v3.19.1
        version: 3.19.1
      - name: gatekeeper-operator-product.v3.19.0
        version: 3.19.0
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.19"
    - currentCSV: gatekeeper-operator-product.v3.19.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.19.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:60ee347425c6cfb36f0e1e834b33cbb3d77aa2d6c89ac513216d0b9fc55d4b52
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
        version: 3.19.1
      entries:
      - name: gatekeeper-operator-product.v3.19.1
        version: 3.19.1
      - name: gatekeeper-operator-product.v3.19.0
        version: 3.19.0
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: stable
    defaultChannel: stable
    packageName: gatekeeper-operator-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-activemqartemis: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/
    name: amq-broker-rhel8
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-broker-operator.v7.10.7-opr-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                    "name": "ex-aao",
                    "application": "ex-aao-app"
                },
                "spec": {
                    "deploymentPlan": {
                        "image": "placeholder",
                        "size": 2,
                        "requireLogin": false,
                        "persistenceEnabled": false,
                        "journalType": "nio",
                        "messageMigration": false,
                        "jolokiaAgentEnabled": false,
                        "managementRBACEnabled": true
                    }
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                    "name": "ex-aaoaddress"
                },
                "spec": {
                    "addressName": "myAddress0",
                    "queueName": "myQueue0",
                    "routingType": "anycast"
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                    "name": "ex-aaoscaledown"
                },
                "spec": {
                    "localOnly": "true"
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                    "name": "ex-keycloak"
                },
                "spec": {
                    "loginModules": {
                        "propertiesLoginModules": [
                            {
                                "name": "prop-module",
                                "users": [
                                    {
                                        "name": "superman",
                                        "roles": [
                                            "root"
                                        ]
                                    }
                                ]
                            }
                        ]
                    },
                    "securityDomains": {
                        "brokerDomain": {
                            "name": "activemq",
                            "loginModules": [
                                {
                                    "name": "prop-module",
                                    "flag": "sufficient"
                                }
                            ]
                        }
                    }
                }
            }]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:1a7aa54d2799d238eb5f49f7a95a78a896f6bf8d222567e9118e0e3963cc9aad
          createdAt: "2021-06-18T00:02:04Z"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A stateful deployment of one or more brokers
            displayName: AMQ Broker
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: Adding and removing addresses via custom resource definitions
            displayName: AMQ Broker Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: Provides message migration on clustered broker scaledown
            displayName: AMQ Broker Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: Security configuration for the broker
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
          to deploy and manage stateful AMQ Broker broker clusters
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: rkieley@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:1a7aa54d2799d238eb5f49f7a95a78a896f6bf8d222567e9118e0e3963cc9aad
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ba21779c29b431a22c824e91820705c9501fb0b7d6f7e1ae22fcd3becc8c8a2f
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:d4010af53429502f41d722c39d69a7bb147bb06d19e3d4a69de70dbe88e3990e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b575d6eebf44dc9049445e357488efb17a690953852419b534df30e0105b9ca4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ff6fefad518a6c997d4c5a6e475ba89640260167f0bc27715daf3cc30116fad1
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:56282c6a461e945a83bffb13d3f3d8d9621800f0fc9cd755a75d21eacd8f5fbf
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:979b59325aa0f34eb05625201beba53fccbb83bd5eb80a89dcb5261ae358138f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:83faefa67b33d93836a1be3cb438bb147242d602565dcbe59b4ed42b26072191
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:967d840b831b310abd3a468a8652d1fa48b2ce28b342bfab53999245deb004f1
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:2cb14c917246ffcbd5d861ca6f3d58cb1146a4e55715ee08160e115d3dfc1c3e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f37f98c809c6f29a83e3d5a3ac4494e28efe9b25d33c54f533c6a08662244622
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:4e09ee3bd8ff3479fdd00e221b7b1505ae04ca6c64fd6bc9e2f1a34b8e98d137
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0d857b95084e39c011651be89482d4dcc77c2d1c3ac3c6a819f3c9d5a57a425f
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:9465e398a55c89c121b3ae50374414190a991839fa5e1018fc02551f6aa3b9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3c00b524722b500ff0922fe7a8722868557512183a8f69248f37a68c45c6a770
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a94fa8135d3765363e442bda45df6e02a3e22bee7ea198bcb0d3584331a4320c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c5f4c08e068b9721967cf7c7cbd9a9e93fb5e39b264dd13b653e99d8f3fa9e0e
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:aa7af7e038a12b5202108b362c8af243954603e84034a16c0192684c86d5d28b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b74d03ed852a3731467ffda95266ce49f2065972f1c37bf254f3d52b34c11991
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:7abfdcfd7b2d3d12b9234f7b057bd199bceefa47bdf02f62236f82a13fdbf5b7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:869cda365447b0b7b53ed21d2496884c15dc9c07295a9dfa0337c2c9974fe018
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:5ff28e3062d3eabb9f52d8271b837fdab06cafbd8b8f732b1bc00de49ad69aad
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f58678f03a250dbb296ebeb112c9150340155799ba4dad9e90bf347b9ed0ca7b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3219f90789690a3c5cfcc86e5f602c443e75f1fdb3a19660f43ee3996bfa8f62
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:f2651cb097eb631fc6af90e4a8404d36516aa2205c546125bf0c37739ab525fa
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:35b7d96d5065f280a528626abd2c3df84e79ccea93ab6bd8b8837af66c330f30
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:45a5e054546a2c1ee338518f8477f728f5f7108c9dd8b516b4c79d4a7d6f5ba7
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:1a9f13fd565113055b2aa2022bd0709a63c1dc058a8570e0fd7eba5c7b101334
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:97c3aa5673610875a53571449f7d5210ae8d675072d044f17b416725ed0dba88
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:ba0f5df7da08bc030ff5e0e68ecb52a931c25f6b67c6556aa8a9661c82d69d42
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:83ed89fa942ccbf4451954956db42138c83f9b015e7de07fc159ddc4bb6dc843
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9a02decbc460ddfc501d8a4a9b41cd084776090951906b027838c0829e1fc15a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ff819e913406911735da39c51699405f747bd697116a273a00897ff176bf05ff
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:5d39f26f06769fe5c989ea0e0ee8f9166120e152ce2c9bc3336fc8d7cab2ff92
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:e6f71001dccfbbe1cd7cf4fded69e181c10f46cc5c8bd3a8d3045a3f3ac41e2f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:61e5a8c1e46f1c18648e0406e2a57c52f911979079020cc57c7a4b06a56940bb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:e03945756d86bd189e8c33027fa668d2065c168510085dfd06defd9f40d3e0ab
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:34739366153b825dcb113e0ed6115ef5656085cf8d47492789ff17464042b576
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:34ab1f66d8daa544e3823464383e9788e307665c875499838fbed95f3a74a4dd
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0b2126cfb6054fdf428c1f43b69e36e93a09a49ce15350e9273c98cc08c6598b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a8b2a9364fd06c8e1c29b8f8f28d3e4844bf7840a62b67137b663c558982d4a3
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:6e9e0e3fddc44fbf87591d8904176407f1ec409d578a7d108c13cdd48d85131b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:67ae74e72927557c55611d0248347b2099b117ac1aa245dac05856fe782fb59c
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:c3ac4602b48867f3d4f7cc0efd5c9f133098e409626cfb402f80386ff460dbf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c9eba0d1f9fba588711774f4db75221f057578407065127ed60c33eef564f499
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:86ff145bd68e797a0507495570063b54cbe4186fcb4acd83c900406619ef9524
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:d4a9176f49ffd9db6051fbda15d47a1aa32c282fb5010d0caf5b9243e80bf055
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:4440b20349032bb5c0bb2a8deb89832b37767af6f0b3208b067867a85641913e
        version: 7.10.7-opr-1
      entries:
      - name: amq-broker-operator.v7.10.7-opr-1
        version: 7.10.7-opr-1
      - name: amq-broker-operator.v7.10.6-opr-1
        version: 7.10.6-opr-1
      - name: amq-broker-operator.v7.10.5-opr-1
        version: 7.10.5-opr-1
      - name: amq-broker-operator.v7.10.4-opr-1
        version: 7.10.4-opr-1
      - name: amq-broker-operator.v7.10.3-opr-3
        version: 7.10.3-opr-3
      - name: amq-broker-operator.v7.10.3-opr-2
        version: 7.10.3-opr-2
      - name: amq-broker-operator.v7.10.3-opr-1
        version: 7.10.3-opr-1
      - name: amq-broker-operator.v7.10.2-opr-2-0.1676475747.p
        version: 7.10.2-opr-2+0.1676475747.p
      - name: amq-broker-operator.v7.10.2-opr-2-0.1680622941.p
        version: 7.10.2-opr-2+0.1680622941.p
      - name: amq-broker-operator.v7.10.2-opr-2
        version: 7.10.2-opr-2
      - name: amq-broker-operator.v7.10.2-opr-1
        version: 7.10.2-opr-1
      - name: amq-broker-operator.v7.10.1-opr-2
        version: 7.10.1-opr-2
      - name: amq-broker-operator.v7.10.1-opr-1
        version: 7.10.1-opr-1
      - name: amq-broker-operator.v7.10.0-opr-4
        version: 7.10.0-opr-4
      - name: amq-broker-operator.v7.10.0-opr-3
        version: 7.10.0-opr-3
      - name: amq-broker-operator.v7.10.0-opr-2
        version: 7.10.0-opr-2
      - name: amq-broker-operator.v7.10.0-opr-1
        version: 7.10.0-opr-1
      name: 7.10.x
    - currentCSV: amq-broker-operator.v7.11.8-opr-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisSecurity is the Schema for the activemqartemissecurities
              API
            displayName: Active MQArtemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:b9aecae6b0ce0b241c7a5b9e6e096395940bb6a384373a4d121822cf10c26d99
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b74d03ed852a3731467ffda95266ce49f2065972f1c37bf254f3d52b34c11991
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:442b6638133d42fa7faa41d86e47be95cdca56c2e16efcd6e7e22991c3812a36
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f37f98c809c6f29a83e3d5a3ac4494e28efe9b25d33c54f533c6a08662244622
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0b2126cfb6054fdf428c1f43b69e36e93a09a49ce15350e9273c98cc08c6598b
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:4440b20349032bb5c0bb2a8deb89832b37767af6f0b3208b067867a85641913e
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:7abfdcfd7b2d3d12b9234f7b057bd199bceefa47bdf02f62236f82a13fdbf5b7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b9f93c243ead118651083880786a81c624f390fcc3596c7ec5ebd55f34f95ea3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:35b7d96d5065f280a528626abd2c3df84e79ccea93ab6bd8b8837af66c330f30
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b575d6eebf44dc9049445e357488efb17a690953852419b534df30e0105b9ca4
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:f2651cb097eb631fc6af90e4a8404d36516aa2205c546125bf0c37739ab525fa
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:83ed89fa942ccbf4451954956db42138c83f9b015e7de07fc159ddc4bb6dc843
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:967d840b831b310abd3a468a8652d1fa48b2ce28b342bfab53999245deb004f1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0d857b95084e39c011651be89482d4dcc77c2d1c3ac3c6a819f3c9d5a57a425f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:e03945756d86bd189e8c33027fa668d2065c168510085dfd06defd9f40d3e0ab
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:56282c6a461e945a83bffb13d3f3d8d9621800f0fc9cd755a75d21eacd8f5fbf
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5e5d8b551ab8ca377e0bc939037970a2f31ee14e9f555853ebea2759d673895a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9a02decbc460ddfc501d8a4a9b41cd084776090951906b027838c0829e1fc15a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0f443f5aa818dd85bfa93e51e37bdac3909628a283aa52a89125b527dd75d9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:5d39f26f06769fe5c989ea0e0ee8f9166120e152ce2c9bc3336fc8d7cab2ff92
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:9465e398a55c89c121b3ae50374414190a991839fa5e1018fc02551f6aa3b9aa
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:1a9f13fd565113055b2aa2022bd0709a63c1dc058a8570e0fd7eba5c7b101334
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:45a5e054546a2c1ee338518f8477f728f5f7108c9dd8b516b4c79d4a7d6f5ba7
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c9eba0d1f9fba588711774f4db75221f057578407065127ed60c33eef564f499
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a94fa8135d3765363e442bda45df6e02a3e22bee7ea198bcb0d3584331a4320c
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:34ab1f66d8daa544e3823464383e9788e307665c875499838fbed95f3a74a4dd
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:86ff145bd68e797a0507495570063b54cbe4186fcb4acd83c900406619ef9524
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c5f4c08e068b9721967cf7c7cbd9a9e93fb5e39b264dd13b653e99d8f3fa9e0e
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:d4a9176f49ffd9db6051fbda15d47a1aa32c282fb5010d0caf5b9243e80bf055
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:29f07006ad62023c20d34971cb3ae900349efae5ab2918aaa8e5647d788032be
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ff819e913406911735da39c51699405f747bd697116a273a00897ff176bf05ff
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b0b3684ef1ebf8b44d3a4c46391984173792dde1bd7363ea06d43a8f95a717d0
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:cc438b4bc2b14f6bff4e5d32e7dc0fad6d2ca391c4e84b5ae925b7b40af76729
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ba21779c29b431a22c824e91820705c9501fb0b7d6f7e1ae22fcd3becc8c8a2f
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:ba0f5df7da08bc030ff5e0e68ecb52a931c25f6b67c6556aa8a9661c82d69d42
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:e6f71001dccfbbe1cd7cf4fded69e181c10f46cc5c8bd3a8d3045a3f3ac41e2f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:67ae74e72927557c55611d0248347b2099b117ac1aa245dac05856fe782fb59c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:979b59325aa0f34eb05625201beba53fccbb83bd5eb80a89dcb5261ae358138f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ade37271426ab8f543d72f8da3e9f663d665f9427379a2539fc9323f6897cc44
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:34739366153b825dcb113e0ed6115ef5656085cf8d47492789ff17464042b576
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:869cda365447b0b7b53ed21d2496884c15dc9c07295a9dfa0337c2c9974fe018
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ff6fefad518a6c997d4c5a6e475ba89640260167f0bc27715daf3cc30116fad1
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:61e5a8c1e46f1c18648e0406e2a57c52f911979079020cc57c7a4b06a56940bb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0dde839c2dce7cb684094bf26523c8e16677de03149a0fff468b8c3f106e1f4f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5f842ebe9a6a4294a056ad10189a5f277cfd3fb4e6e0b3225e194bd5c131151d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3c00b524722b500ff0922fe7a8722868557512183a8f69248f37a68c45c6a770
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:4e09ee3bd8ff3479fdd00e221b7b1505ae04ca6c64fd6bc9e2f1a34b8e98d137
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:d4010af53429502f41d722c39d69a7bb147bb06d19e3d4a69de70dbe88e3990e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a8b2a9364fd06c8e1c29b8f8f28d3e4844bf7840a62b67137b663c558982d4a3
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:5ff28e3062d3eabb9f52d8271b837fdab06cafbd8b8f732b1bc00de49ad69aad
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:d2ae3defac36f00112c6b773e13626336ba15d6f50f905ad1c33f122bfe1a6a9
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:97c3aa5673610875a53571449f7d5210ae8d675072d044f17b416725ed0dba88
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f58678f03a250dbb296ebeb112c9150340155799ba4dad9e90bf347b9ed0ca7b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:40ccaf2bb2e0f8132f2541b39cb835f17b38050b3ca9232f0861fd5fcaf4324a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:6e9e0e3fddc44fbf87591d8904176407f1ec409d578a7d108c13cdd48d85131b
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:c3ac4602b48867f3d4f7cc0efd5c9f133098e409626cfb402f80386ff460dbf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:137b893ad15e0cd5353bae134c8fae833257a12641b12d0f774f1565ce66f63c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:89bb769b4c33eb931cc7866eeabec9a315b913851cfa84bc6e11ae8d1e8b2cbb
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:2cb14c917246ffcbd5d861ca6f3d58cb1146a4e55715ee08160e115d3dfc1c3e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:25c3732e945b4bc3e9aca721ce709d0217339b065d9dac9afdb6202e9fd8a799
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:8b4c4400a02c6cacfd978147b158e2ccbe5da955578b4b970a5ee7edac865326
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3219f90789690a3c5cfcc86e5f602c443e75f1fdb3a19660f43ee3996bfa8f62
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:7d53744ae8c77aea09c59165486c9ed0eaa131140f6b346042293bce0bd8a745
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:aa7af7e038a12b5202108b362c8af243954603e84034a16c0192684c86d5d28b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9c1ac38bf8eec67a3da57c3cc61105513776a749408cb287fe6448a17889de4e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:83faefa67b33d93836a1be3cb438bb147242d602565dcbe59b4ed42b26072191
        version: 7.11.8-opr-1
      entries:
      - name: amq-broker-operator.v7.11.8-opr-1
        version: 7.11.8-opr-1
      - name: amq-broker-operator.v7.11.7-opr-2-1742200979
        version: 7.11.7-opr-2-1742200979
      - name: amq-broker-operator.v7.11.7-opr-2-1740583644
        version: 7.11.7-opr-2-1740583644
      - name: amq-broker-operator.v7.11.7-opr-2-1736956309
        version: 7.11.7-opr-2-1736956309
      - name: amq-broker-operator.v7.11.7-opr-2-1731330819
        version: 7.11.7-opr-2-1731330819
      - name: amq-broker-operator.v7.11.7-opr-2
        version: 7.11.7-opr-2
      - name: amq-broker-operator.v7.11.7-opr-1-1726157430
        version: 7.11.7-opr-1-1726157430
      - name: amq-broker-operator.v7.11.7-opr-1-1723204800
        version: 7.11.7-opr-1-1723204800
      - name: amq-broker-operator.v7.11.7-opr-1
        version: 7.11.7-opr-1
      - name: amq-broker-operator.v7.11.6-opr-2
        version: 7.11.6-opr-2
      - name: amq-broker-operator.v7.11.5-opr-1
        version: 7.11.5-opr-1
      - name: amq-broker-operator.v7.11.4-opr-1
        version: 7.11.4-opr-1
      - name: amq-broker-operator.v7.11.3-opr-1
        version: 7.11.3-opr-1
      - name: amq-broker-operator.v7.11.2-opr-1
        version: 7.11.2-opr-1
      - name: amq-broker-operator.v7.11.1-opr-2
        version: 7.11.1-opr-2
      - name: amq-broker-operator.v7.11.1-opr-1
        version: 7.11.1-opr-1
      - name: amq-broker-operator.v7.11.0-opr-3
        version: 7.11.0-opr-3
      - name: amq-broker-operator.v7.11.0-opr-2
        version: 7.11.0-opr-2
      name: 7.11.x
    - currentCSV: amq-broker-operator.v7.12.5-opr-1-0.1757928550.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-broker-operator.v7.12.5-opr-1-0.1756216162.p
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:7d53744ae8c77aea09c59165486c9ed0eaa131140f6b346042293bce0bd8a745
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9c1ac38bf8eec67a3da57c3cc61105513776a749408cb287fe6448a17889de4e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:29f07006ad62023c20d34971cb3ae900349efae5ab2918aaa8e5647d788032be
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:dc13871dfed52402c40fd192900fc3cb5a49da4fa08e4ec4d0e23a879e83db25
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5e5d8b551ab8ca377e0bc939037970a2f31ee14e9f555853ebea2759d673895a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:cc438b4bc2b14f6bff4e5d32e7dc0fad6d2ca391c4e84b5ae925b7b40af76729
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d71506472636d940b5dbd3e708d3e0448d777eeecfd38de61ba49e281d2979c6
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ffffa9875f0379e9373f89f05eb06e5a193273bb04bc3aa5f85b044357b79098
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:b9aecae6b0ce0b241c7a5b9e6e096395940bb6a384373a4d121822cf10c26d99
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:442b6638133d42fa7faa41d86e47be95cdca56c2e16efcd6e7e22991c3812a36
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:25c3732e945b4bc3e9aca721ce709d0217339b065d9dac9afdb6202e9fd8a799
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ade37271426ab8f543d72f8da3e9f663d665f9427379a2539fc9323f6897cc44
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:16253895bc113f395dc6ab5cd511901e43b6ee676369ec80d598a8741c3c9bf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:89bb769b4c33eb931cc7866eeabec9a315b913851cfa84bc6e11ae8d1e8b2cbb
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:bb6fbd68475a7852b4d99eea6c4ab313f9267da7963162f0d75375d7063409e7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b9f93c243ead118651083880786a81c624f390fcc3596c7ec5ebd55f34f95ea3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9cc48eecf1442ae04b8543fa5d4381a13bc2831390850828834d387006d1342b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:47fd4ce2533496828aba37bd1f9715e2164d5c90bd0fc6b25e7e0786d723bf01
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:023478b773a0c42739b86e822a4d435e01ef341b8136566ea52463f1b7e447ed
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:eacc4fb55e8c4c30894a51944960d536508001a49fda62782c6cb2ca8767961d
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0dde839c2dce7cb684094bf26523c8e16677de03149a0fff468b8c3f106e1f4f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:8b4c4400a02c6cacfd978147b158e2ccbe5da955578b4b970a5ee7edac865326
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d42d713da0ce6806fdc6492b6342586783e6865a82a8647d3c4288439b1751ee
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:7f9b921d89b6d4d91b365659afff127d99f1455b94b48a10a6a7f4c38dd8b700
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:90f1678c19c74454b77aed14fa6b3074a1c6f80fe884f1c6136e7e907f402aec
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b2402179b1d0ff9c950defe5d8808195ef9a4c3e13e3c469cae31ab51d5f5014
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:8329acdff170c5a8ae6ae16f6c37d3f7ab68d106a6e8be50c3fe131ec5ecceeb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5f842ebe9a6a4294a056ad10189a5f277cfd3fb4e6e0b3225e194bd5c131151d
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:40920687402fd747ca02d384e63692a5e1db625723a2bf3056a607c92ee8eee1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0f443f5aa818dd85bfa93e51e37bdac3909628a283aa52a89125b527dd75d9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:40ccaf2bb2e0f8132f2541b39cb835f17b38050b3ca9232f0861fd5fcaf4324a
        version: 7.12.5-opr-1+0.1757928550.p
      entries:
      - name: amq-broker-operator.v7.12.5-opr-1-0.1757928550.p
        version: 7.12.5-opr-1+0.1757928550.p
      - name: amq-broker-operator.v7.12.5-opr-1-0.1756216162.p
        version: 7.12.5-opr-1+0.1756216162.p
      - name: amq-broker-operator.v7.12.5-opr-1-0.1754272113.p
        version: 7.12.5-opr-1+0.1754272113.p
      - name: amq-broker-operator.v7.12.5-opr-1
        version: 7.12.5-opr-1
      - name: amq-broker-operator.v7.12.4-opr-1
        version: 7.12.4-opr-1
      - name: amq-broker-operator.v7.12.4-opr-1-0.1747217191.p
        version: 7.12.4-opr-1+0.1747217191.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741881887.p
        version: 7.12.3-opr-1+0.1741881887.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739991235.p
        version: 7.12.3-opr-1+0.1739991235.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1730142550.p
        version: 7.12.3-opr-1+0.1730142550.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1730813578.p
        version: 7.12.3-opr-1+0.1730813578.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1731483798.p
        version: 7.12.3-opr-1+0.1731483798.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1732654016.p
        version: 7.12.3-opr-1+0.1732654016.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1733315503.p
        version: 7.12.3-opr-1+0.1733315503.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1737643631.p
        version: 7.12.3-opr-1+0.1737643631.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1738596517.p
        version: 7.12.3-opr-1+0.1738596517.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739390944.p
        version: 7.12.3-opr-1+0.1739390944.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739889884.p
        version: 7.12.3-opr-1+0.1739889884.p
      - name: amq-broker-operator.v7.12.3-opr-1
        version: 7.12.3-opr-1
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741884937.p
        version: 7.12.3-opr-1+0.1741884937.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741883943.p
        version: 7.12.3-opr-1+0.1741883943.p
      - name: amq-broker-operator.v7.12.2-opr-1
        version: 7.12.2-opr-1
      - name: amq-broker-operator.v7.12.1-opr-1-0.1724054375.p
        version: 7.12.1-opr-1+0.1724054375.p
      - name: amq-broker-operator.v7.12.1-opr-1-0.1723669176.p
        version: 7.12.1-opr-1+0.1723669176.p
      - name: amq-broker-operator.v7.12.1-opr-1-0.1721247907.p
        version: 7.12.1-opr-1+0.1721247907.p
      - name: amq-broker-operator.v7.12.1-opr-1
        version: 7.12.1-opr-1
      - name: amq-broker-operator.v7.12.0-opr-2
        version: 7.12.0-opr-2
      - name: amq-broker-operator.v7.12.0-opr-1
        version: 7.12.0-opr-1
      name: 7.12.x
    defaultChannel: 7.12.x
    packageName: amq-broker-rhel8
    provider:
      name: Red Hat
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecrpublic-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecrpublic-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecrpublic-controller:1.1.1
          createdAt: "2025-09-20T01:21:52Z"
          description: AWS ECR PUBLIC controller is a service controller for managing
            ECR PUBLIC resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Repository represents the state of an AWS ecrpublic Repository
              resource.
            displayName: Repository
            kind: Repository
            name: repositories.ecrpublic.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ECR PUBLIC resources in AWS from within your Kubernetes cluster.

          **About Amazon ECR PUBLIC**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECR PUBLIC
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecrpublic
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECR PUBLIC Developer Resources
          url: https://aws.amazon.com/ECR PUBLIC/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecrpublic maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecrpublic-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ecrpublic-controller.v1.1.1
        version: 1.1.1
      - name: ack-ecrpublic-controller.v1.1.0
        version: 1.1.0
      - name: ack-ecrpublic-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecrpublic-controller.v1.0.12
        version: 1.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecrpublic-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-s3-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-s3-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "s3.services.k8s.aws/v1alpha1",
                "kind": "Bucket",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/s3-controller:1.1.1
          createdAt: "2025-09-19T19:23:45Z"
          description: AWS S3 controller is a service controller for managing S3 resources
            in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Bucket represents the state of an AWS s3 Bucket resource.
            displayName: Bucket
            kind: Bucket
            name: buckets.s3.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Simple Storage Service (S3) resources in AWS from within your Kubernetes cluster.

          **About Amazon S3**

          Amazon Simple Storage Service (Amazon S3) is an object storage service that offers industry-leading scalability, data availability, security, and performance. This means customers of all sizes and industries can use it to store and protect any amount of data for a range of use cases, such as data lakes, websites, mobile applications, backup and restore, archive, enterprise applications, IoT devices, and big data analytics. Amazon S3 provides easy-to-use management features so you can organize your data and configure finely-tuned access controls to meet your specific business, organizational, and compliance requirements. Amazon S3 is designed for 99.999999999% (11 9s) of durability, and stores data for millions of applications for companies all around the world.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon S3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - s3
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon S3 Developer Resources
          url: https://aws.amazon.com/s3/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: s3 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/s3-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-s3-controller.v1.1.1
        version: 1.1.1
      - name: ack-s3-controller.v1.1.0
        version: 1.1.0
      - name: ack-s3-controller.v1.0.35
        version: 1.0.35
      - name: ack-s3-controller.v1.0.33
        version: 1.0.33
      - name: ack-s3-controller.v1.0.32
        version: 1.0.32
      - name: ack-s3-controller.v1.0.31
        version: 1.0.31
      - name: ack-s3-controller.v1.0.30
        version: 1.0.30
      - name: ack-s3-controller.v1.0.29
        version: 1.0.29
      - name: ack-s3-controller.v1.0.28
        version: 1.0.28
      - name: ack-s3-controller.v1.0.27
        version: 1.0.27
      - name: ack-s3-controller.v1.0.26
        version: 1.0.26
      - name: ack-s3-controller.v1.0.25
        version: 1.0.25
      - name: ack-s3-controller.v1.0.24
        version: 1.0.24
      - name: ack-s3-controller.v1.0.23
        version: 1.0.23
      - name: ack-s3-controller.v1.0.21
        version: 1.0.21
      - name: ack-s3-controller.v1.0.20
        version: 1.0.20
      - name: ack-s3-controller.v1.0.19
        version: 1.0.19
      - name: ack-s3-controller.v1.0.18
        version: 1.0.18
      - name: ack-s3-controller.v1.0.17
        version: 1.0.17
      - name: ack-s3-controller.v1.0.16
        version: 1.0.16
      - name: ack-s3-controller.v1.0.15
        version: 1.0.15
      - name: ack-s3-controller.v1.0.14
        version: 1.0.14
      - name: ack-s3-controller.v1.0.13
        version: 1.0.13
      - name: ack-s3-controller.v1.0.12
        version: 1.0.12
      - name: ack-s3-controller.v1.0.11
        version: 1.0.11
      - name: ack-s3-controller.v1.0.9
        version: 1.0.9
      - name: ack-s3-controller.v1.0.7
        version: 1.0.7
      - name: ack-s3-controller.v1.0.6
        version: 1.0.6
      - name: ack-s3-controller.v1.0.5
        version: 1.0.5
      - name: ack-s3-controller.v1.0.4
        version: 1.0.4
      - name: ack-s3-controller.v1.0.3
        version: 1.0.3
      - name: ack-s3-controller.v1.0.2
        version: 1.0.2
      - name: ack-s3-controller.v1.0.1
        version: 1.0.1
      - name: ack-s3-controller.v1.0.0
        version: 1.0.0
      - name: ack-s3-controller.v0.1.8
        version: 0.1.8
      - name: ack-s3-controller.v0.1.7
        version: 0.1.7
      - name: ack-s3-controller.v0.1.6
        version: 0.1.6
      - name: ack-s3-controller.v0.1.5
        version: 0.1.5
      - name: ack-s3-controller.v0.1.4
        version: 0.1.4
      - name: ack-s3-controller.v0.1.3
        version: 0.1.3
      - name: ack-s3-controller.v0.1.2
        version: 0.1.2
      - name: ack-s3-controller.v0.1.1
        version: 0.1.1
      - name: ack-s3-controller.v0.1.0
        version: 0.1.0
      - name: ack-s3-controller.v0.0.20
        version: 0.0.20
      - name: ack-s3-controller.v0.0.19
        version: 0.0.19
      - name: ack-s3-controller.v0.0.18
        version: 0.0.18
      - name: ack-s3-controller.v0.0.17
        version: 0.0.17
      - name: ack-s3-controller.v0.0.16
        version: 0.0.16
      - name: ack-s3-controller.v0.0.15
        version: 0.0.15
      - name: ack-s3-controller.v0.0.14
        version: 0.0.14
      - name: ack-s3-controller.v0.0.13
        version: 0.0.13
      - name: ack-s3-controller.v0.0.12
        version: 0.0.12
      - name: ack-s3-controller.v0.0.11
        version: 0.0.11
      - name: ack-s3-controller.v0.0.10
        version: 0.0.10
      - name: ack-s3-controller.v0.0.9
        version: 0.0.9
      - name: ack-s3-controller.v0.0.8
        version: 0.0.8
      - name: ack-s3-controller.v0.0.7
        version: 0.0.7
      name: alpha
    defaultChannel: alpha
    packageName: ack-s3-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: numaresources-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: numaresources-operator.v4.16.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodetopology.openshift.io/v1alpha1",
                "kind": "NUMAResourcesOperator",
                "metadata": {
                  "name": "numaresourcesoperator"
                },
                "spec": {
                  "nodeGroups": [
                    {
                      "machineConfigPoolSelector": {
                        "matchLabels": {
                          "pools.operator.machineconfiguration.openshift.io/worker": ""
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "nodetopology.openshift.io/v1alpha1",
                "kind": "NUMAResourcesScheduler",
                "metadata": {
                  "name": "numaresourcesscheduler"
                },
                "spec": {
                  "imageSpec": "URL_OF_SCHEDULER_IMAGE_FROM_REDHAT_REGISTRY",
                  "logLevel": "Normal",
                  "schedulerName": "topo-aware-scheduler"
                }
              }
            ]
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "true"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.15.0 <4.16.6'
          operators.openshift.io/valid-subscription: "[\n\t\"OpenShift Kubernetes
            Engine\",\n\t\"OpenShift Container Platform\",\n\t\"OpenShift Platform
            Plus\"\n]"
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NUMAResourcesOperator is the Schema for the numaresourcesoperators
              API
            displayName: NUMA Resources Operator
            kind: NUMAResourcesOperator
            name: numaresourcesoperators.nodetopology.openshift.io
            version: v1
          - description: NUMAResourcesOperator is the Schema for the numaresourcesoperators
              API
            displayName: NUMA Resources Operator
            kind: NUMAResourcesOperator
            name: numaresourcesoperators.nodetopology.openshift.io
            version: v1alpha1
          - description: NUMAResourcesScheduler is the Schema for the numaresourcesschedulers
              API
            displayName: NUMA Aware Scheduler
            kind: NUMAResourcesScheduler
            name: numaresourcesschedulers.nodetopology.openshift.io
            version: v1
          - description: NUMAResourcesScheduler is the Schema for the numaresourcesschedulers
              API
            displayName: NUMA Aware Scheduler
            kind: NUMAResourcesScheduler
            name: numaresourcesschedulers.nodetopology.openshift.io
            version: v1alpha1
        description: NUMA resources exporter operator
        displayName: numaresources-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - RTE
        - NUMA
        links:
        - name: Numaresources Operator
          url: https://github.com/openshift-kni/numaresources-operator
        maintainers:
        - email: fromani@redhat.com
          name: fromani
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/numaresources-rhel9-operator@sha256:aca8232e89172ca9fd9ce4a66cae9e8bb027b6d16482ce6b8e27fb7779c398e2
        version: 4.16.6
      entries:
      - name: numaresources-operator.v4.16.6
        version: 4.16.6
      - name: numaresources-operator.v4.16.5
        version: 4.16.5
      - name: numaresources-operator.v4.16.4
        version: 4.16.4
      - name: numaresources-operator.v4.16.3
        version: 4.16.3
      - name: numaresources-operator.v4.16.2
        version: 4.16.2
      - name: numaresources-operator.v4.16.1
        version: 4.16.1
      - name: numaresources-operator.v4.16.0
        version: 4.16.0
      name: "4.16"
    defaultChannel: "4.16"
    packageName: numaresources-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.9-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lvms-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lvms-operator.v4.15.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "my-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "default": true,
                        "fstype": "xfs",
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:696edeb4c168646b49b8b0ca5b3d06071ab66fdb010e7cd46888d8b2f3a0c7a8
          description: Logical volume manager storage provides dynamically provisioned
            local storage for container workloads
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.11'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "test-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
          operatorframework.io/suggested-namespace: openshift-storage
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-storage",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/internal-objects: '["logicalvolumes.topolvm.io",
            "lvmvolumegroups.lvm.topolvm.io", "lvmvolumegroupnodestatuses.lvm.topolvm.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/lvm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LogicalVolume
            name: logicalvolumes.topolvm.io
            version: v1
          - description: LVMCluster is the Schema for the lvmclusters API
            displayName: LVMCluster
            kind: LVMCluster
            name: lvmclusters.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroupNodeStatus
            name: lvmvolumegroupnodestatuses.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroup
            name: lvmvolumegroups.lvm.topolvm.io
            version: v1alpha1
        description: Logical volume manager storage provides dynamically provisioned
          local storage.
        displayName: LVM Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - local storage
        - operator
        - LVM
        links:
        - name: Source Repository
          url: https://github.com/openshift/lvm-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:771d513c4d4e5d77d797b4c18f86e23162a20c6d39955e93d24f62b94e62e785
        - registry.redhat.io/lvms4/lvms-must-gather-rhel9@sha256:382ceea1151da56d0931403207c9f85721d58a66b39b1cd79790eca6219807d0
        - registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:696edeb4c168646b49b8b0ca5b3d06071ab66fdb010e7cd46888d8b2f3a0c7a8
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:82c01635b32821f04a7cfcf80da8b3f7e097786f220e0c85aab47bda4536a1d6
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:2db4f32824a8f9f6780dc3d236f73684cb20db41466ffa8f9a89fd3f5eb62fdd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:7ce50398cd29df6e07ccd89790a53a711a7e5b484698c48f6581f5d4e9efb05a
        - registry.redhat.io/lvms4/topolvm-rhel9@sha256:074a206d36c9e4198d90baed86f3046abd2ec12f9b36a715b60900011366b3d4
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:05dceaff7f0b2b769ddfd9261cc66d42aab9158ce99ed3dbd6c229e784072635
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:479454d0d4de6609d8a4696b9c7f0097115afb45db3bdd46631d58afd77dcc58
        version: 4.15.11
      entries:
      - name: lvms-operator.v4.15.11
        version: 4.15.11
      - name: lvms-operator.v4.15.10
        version: 4.15.10
      - name: lvms-operator.v4.15.9
        version: 4.15.9
      - name: lvms-operator.v4.15.8
        version: 4.15.8
      - name: lvms-operator.v4.15.7
        version: 4.15.7
      - name: lvms-operator.v4.15.6
        version: 4.15.6
      - name: lvms-operator.v4.15.5
        version: 4.15.5
      - name: lvms-operator.v4.15.4
        version: 4.15.4
      - name: lvms-operator.v4.15.3
        version: 4.15.3
      - name: lvms-operator.v4.15.2
        version: 4.15.2
      - name: lvms-operator.v4.15.1
        version: 4.15.1
      name: stable-4.15
    - currentCSV: lvms-operator.v4.16.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "my-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "default": true,
                        "fstype": "xfs",
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:b9d80bb263a54192501d5ba9366a4f0c19b34875b51bad4fa75ef22c60bb3ee7
          description: Logical volume manager storage provides dynamically provisioned
            local storage for container workloads
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.9'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "test-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
          operatorframework.io/suggested-namespace: openshift-storage
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-storage",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["logicalvolumes.topolvm.io",
            "lvmvolumegroups.lvm.topolvm.io", "lvmvolumegroupnodestatuses.lvm.topolvm.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lvm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LogicalVolume
            name: logicalvolumes.topolvm.io
            version: v1
          - description: LVMCluster is the Schema for the lvmclusters API
            displayName: LVMCluster
            kind: LVMCluster
            name: lvmclusters.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroupNodeStatus
            name: lvmvolumegroupnodestatuses.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroup
            name: lvmvolumegroups.lvm.topolvm.io
            version: v1alpha1
        description: Logical volume manager storage provides dynamically provisioned
          local storage.
        displayName: LVM Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - local storage
        - operator
        - LVM
        links:
        - name: Source Repository
          url: https://github.com/openshift/lvm-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/lvms4/topolvm-rhel9@sha256:227748103049d2cced921ba9fc4e2d3671f08c897cc2a5627d755fe5f9d2767d
        - registry.redhat.io/lvms4/lvms-must-gather-rhel9@sha256:20b28b87b3fe6d6bf739181376231d6f8b7454c34b59eaf2e06c04a9fca8f41d
        - registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:b9d80bb263a54192501d5ba9366a4f0c19b34875b51bad4fa75ef22c60bb3ee7
        version: 4.16.9
      entries:
      - name: lvms-operator.v4.16.9
        version: 4.16.9
      - name: lvms-operator.v4.16.8
        version: 4.16.8
      - name: lvms-operator.v4.16.7
        version: 4.16.7
      - name: lvms-operator.v4.16.6
        version: 4.16.6
      - name: lvms-operator.v4.16.5
        version: 4.16.5
      - name: lvms-operator.v4.16.4
        version: 4.16.4
      - name: lvms-operator.v4.16.3
        version: 4.16.3
      - name: lvms-operator.v4.16.2
        version: 4.16.2
      - name: lvms-operator.v4.16.1
        version: 4.16.1
      - name: lvms-operator.v4.16.0
        version: 4.16.0
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: lvms-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-custom-metrics-autoscaler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: custom-metrics-autoscaler.v2.17.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eventing.keda.sh/v1alpha1",
                "kind": "CloudEventSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keda",
                    "app.kubernetes.io/instance": "cloudeventsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cloudeventsource",
                    "app.kubernetes.io/part-of": "keda"
                  },
                  "name": "cloudeventsource-sample"
                },
                "spec": {
                  "clusterName": "cluster-sample",
                  "destination": {
                    "http": {
                      "uri": "http://foo.bar"
                    }
                  }
                }
              },
              {
                "apiVersion": "eventing.keda.sh/v1alpha1",
                "kind": "ClusterCloudEventSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keda",
                    "app.kubernetes.io/instance": "clustercloudeventsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clustercloudeventsource",
                    "app.kubernetes.io/part-of": "keda"
                  },
                  "name": "clustercloudeventsource-sample"
                },
                "spec": {
                  "clusterName": "clustercluster-sample",
                  "destination": {
                    "http": {
                      "uri": "http://foo.bar"
                    }
                  }
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ClusterTriggerAuthentication",
                "metadata": {
                  "name": "example-clustertriggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "example-secret-parameter"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "KedaController",
                "metadata": {
                  "name": "keda",
                  "namespace": "openshift-keda"
                },
                "spec": {
                  "admissionWebhooks": {
                    "logEncoder": "console",
                    "logLevel": "info"
                  },
                  "metricsServer": {
                    "logLevel": "0"
                  },
                  "operator": {
                    "logEncoder": "console",
                    "logLevel": "info"
                  },
                  "serviceAccount": null,
                  "watchNamespace": ""
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledJob",
                "metadata": {
                  "name": "scaledjob-sample"
                },
                "spec": {
                  "failedJobsHistoryLimit": 5,
                  "jobTargetRef": {
                    "activeDeadlineSeconds": 600,
                    "backoffLimit": 6,
                    "completions": 1,
                    "parallelism": 1,
                    "template": null
                  },
                  "maxReplicaCount": 100,
                  "pollingInterval": 30,
                  "successfulJobsHistoryLimit": 5,
                  "triggers": [
                    {
                      "metadata": {
                        "property": "example-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledObject",
                "metadata": {
                  "name": "example-scaledobject"
                },
                "spec": {
                  "cooldownPeriod": 300,
                  "maxReplicaCount": 100,
                  "minReplicaCount": 0,
                  "pollingInterval": 30,
                  "scaleTargetRef": {
                    "name": "example-deployment"
                  },
                  "triggers": [
                    {
                      "metadata": {
                        "property": "example-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "TriggerAuthentication",
                "metadata": {
                  "name": "example-triggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "example-secret-parameter"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Cloud Provider
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9-operator@sha256:9a359ae6612d2f22fe983653418143873cb9939f83dca6f008ec3cb8f3d53990
          createdAt: 20 Aug 2025, 21:31
          description: Custom Metrics Autoscaler Operator, an event-driven autoscaler
            based upon KEDA
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.1 <2.17.2-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-keda
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/custom-metrics-autoscaler-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudEventSource
            name: cloudeventsources.eventing.keda.sh
            version: v1alpha1
          - kind: ClusterCloudEventSource
            name: clustercloudeventsources.eventing.keda.sh
            version: v1alpha1
          - description: Describes cluster wide authentication parameters
            displayName: ClusterTriggerAuthentication
            kind: ClusterTriggerAuthentication
            name: clustertriggerauthentications.keda.sh
            version: v1alpha1
          - description: |
              Represents an installation of a particular version of KEDA Controller.
            displayName: KedaController
            kind: KedaController
            name: kedacontrollers.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your jobs and what the triggers
            displayName: ScaledJob
            kind: ScaledJob
            name: scaledjobs.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your application and what the
              triggers are
            displayName: ScaledObject
            kind: ScaledObject
            name: scaledobjects.keda.sh
            version: v1alpha1
          - description: Describes authentication parameters
            displayName: TriggerAuthentication
            kind: TriggerAuthentication
            name: triggerauthentications.keda.sh
            version: v1alpha1
        description: |
          ## About the managed application
          Custom Metrics Autoscaler for OpenShift is an event driven autoscaler based upon KEDA.  Custom Metrics Autoscaler can monitor event sources like Kafka, RabbitMQ, or cloud event sources and feed the metrics from those sources into the Kubernetes horizontal pod autoscaler.  With Custom Metrics Autoscaler, you can have event driven and serverless scale of deployments within any Kubernetes cluster.
          ## About this Operator
          The Custom Metrics Autoscaler Operator deploys and manages installation of KEDA Controller in the cluster. Install this operator and follow installation instructions on how to install Custom Metrics Autoscaler in your cluster.

          ## Prerequisites for enabling this Operator
          ## How to install Custom Metrics Autoscaler in the cluster
          The installation of Custom Metrics Autoscaler is triggered by the creation of `KedaController` resource. Please refer to the [KedaController Spec](https://github.com/openshift/custom-metrics-autoscaler-operator/blob/main/README.md#the-kedacontroller-custom-resource) for more details on available options.

          Only a resource named `keda` in the namespace where the operator is installed will trigger the installation, reconfiguration or removal of the KEDA Controller resource.

          There should be only one KEDA Controller in the cluster.

          The operator will behave in this manner whether it is installed with the `AllNamespaces` or `OwnNamespace` install mode. While the operator more closely matches the `OwnNamespace` semantics, `AllNamespaces` is a supported installation mode to allow it to be installed to namespaces with existing `OperatorGroups` which require that installation mode, such as `openshift-operators`.
        displayName: Custom Metrics Autoscaler
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keda
        - scaling
        - event
        links:
        - name: Custom Metrics Autoscaler Documentation
          url: https://docs.openshift.com/container-platform/latest/nodes/cma/nodes-cma-autoscaling-custom.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9-operator@sha256:9a359ae6612d2f22fe983653418143873cb9939f83dca6f008ec3cb8f3d53990
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9@sha256:ace9358fbd2d81a6e16bfa049c763ca16acf6cc2a59cdc0b6282e5e0dfc902cc
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-adapter-rhel9@sha256:0c8ae9323f927346e168ba829f6c7679af6b2dd359c8d0a020dd0ae3954481ce
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-admission-webhooks-rhel9@sha256:75631a21de057844e9ae45ae2496a22e16f3a32a6fd5f6cb5f90e173e3be2180
        version: 2.17.2-1
      entries:
      - name: custom-metrics-autoscaler.v2.17.2-1
        version: 2.17.2-1
      - name: custom-metrics-autoscaler.v2.15.1-6
        version: 2.15.1-6
      - name: custom-metrics-autoscaler.v2.15.1-4
        version: 2.15.1-4
      - name: custom-metrics-autoscaler.v2.14.1-467
        version: 2.14.1-467
      - name: custom-metrics-autoscaler.v2.14.1-454
        version: 2.14.1-454
      - name: custom-metrics-autoscaler.v2.13.1-427
        version: 2.13.1-427
      - name: custom-metrics-autoscaler.v2.12.1-394
        version: 2.12.1-394
      - name: custom-metrics-autoscaler.v2.11.2-322
        version: 2.11.2-322
      - name: custom-metrics-autoscaler.v2.11.2-311
        version: 2.11.2-311
      - name: custom-metrics-autoscaler.v2.10.1-267
        version: 2.10.1-267
      name: stable
    defaultChannel: stable
    packageName: openshift-custom-metrics-autoscaler-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com
    name: aws-load-balancer-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-load-balancer-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
          createdAt: 2022-06-28T17:56:41
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:d7bc364512178c36671d8a4b5a76cf7cb10f8e56997106187b0fe1f032670ece
        version: 0.0.1
      entries:
      - name: aws-load-balancer-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: aws-load-balancer-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
          createdAt: 2022-12-09T15:32:32
          olm.skipRange: <0.2.0
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:cfd210a2cb705156883fda126d052e05eb84345cc27ba04688067f8ed9a802ea
        version: 0.2.0
      entries:
      - name: aws-load-balancer-operator.v0.2.0
        version: 0.2.0
      name: stable-v0
    - currentCSV: aws-load-balancer-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
          createdAt: 2022-06-28T17:56:41
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:d7bc364512178c36671d8a4b5a76cf7cb10f8e56997106187b0fe1f032670ece
        version: 0.0.1
      entries:
      - name: aws-load-balancer-operator.v0.0.1
        version: 0.0.1
      name: stable-v0.1
    - currentCSV: aws-load-balancer-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
          createdAt: 2022-12-09T15:32:32
          olm.skipRange: <0.2.0
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:cfd210a2cb705156883fda126d052e05eb84345cc27ba04688067f8ed9a802ea
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        version: 0.2.0
      entries:
      - name: aws-load-balancer-operator.v0.2.0
        version: 0.2.0
      name: stable-v0.2
    - currentCSV: aws-load-balancer-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
          createdAt: 2024-01-23T11:21:00
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:2e0b9332a44d8d9c23e19c7accab0813a651f39210257820db508cac28876595
        version: 1.1.1
      entries:
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1
    - currentCSV: aws-load-balancer-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:6e15df39a3d25755a79888f739e1af8b052c3fba57bf8f6b3434ef58fd4288cf
          createdAt: 2024-01-24T09:13:21
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat <<EOF | oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
          EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note about UPI:__ For UPI based installs, additional documentation can be found in
          [VPC and subnets](https://github.com/openshift/aws-load-balancer-operator/blob/release-1.0/docs/prerequisites.md#vpc-and-subnets)

          __Note about STS:__ For clusters using STS, use the [STS documentation](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#for-sts-clusters) instead.
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9f9f8f69ab98c1f25cc301572339e0cc5c5b76f2e29ed26e10ac91ad20e0090c
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:6e15df39a3d25755a79888f739e1af8b052c3fba57bf8f6b3434ef58fd4288cf
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:73db97d20ee9415e9a8941cc6a36ed81d901d760d9949e95e10c90866ef8bf64
        version: 1.0.1
      entries:
      - name: aws-load-balancer-operator.v1.0.1
        version: 1.0.1
      - name: aws-load-balancer-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: aws-load-balancer-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
          createdAt: 2024-01-23T11:21:00
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:2e0b9332a44d8d9c23e19c7accab0813a651f39210257820db508cac28876595
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
        version: 1.1.1
      entries:
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    defaultChannel: stable-v1
    packageName: aws-load-balancer-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://github.com/redhat-developer/app-services-operator/
    name: rhoas-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhoas-operator.0.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "ServiceRegistryConnection",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace",
                  "labels": {
                    "app.kubernetes.io/component": "external-service",
                    "app.kubernetes.io/managed-by": "rhoas"
                  }
                },
                "spec": {
                  "accessTokenSecretName": "rh-managed-services-api-accesstoken",
                  "serviceRegistryId": "exampleId",
                  "credentials": {
                    "serviceAccountSecretName": "service-account-secret"
                  }
                }
              },
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "CloudServiceAccountRequest",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace"
                },
                "spec": {
                  "serviceAccountName": "rhoas-sa",
                  "serviceAccountDescription": "Operator created service account",
                  "serviceAccountSecretName": "service-account-credentials",
                  "accessTokenSecretName": "rh-managed-services-api-accesstoken"
                }
              },
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "CloudServicesRequest",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace",
                  "labels": {
                    "app.kubernetes.io/component": "external-service",
                    "app.kubernetes.io/managed-by": "rhoas"
                  }
                },
                "spec": {
                  "accessTokenSecretName": "rh-cloud-services-api-accesstoken"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: quay.io/rhoas/service-operator:0.13.0
          createdAt: "2023-01-20 16:09:37"
          description: Red Hat OpenShift Application Services Operator
          olm.skipRange: '>=0.0.1 <0.13.0'
          repository: https://github.com/redhat-developer/app-services-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Creates RHOAS Service Accounts
            displayName: Cloud Service Account Request
            kind: CloudServiceAccountRequest
            name: cloudserviceaccountrequests.rhoas.redhat.com
            version: v1alpha1
          - description: Cloud Services Request
            displayName: Cloud Services Request
            kind: CloudServicesRequest
            name: cloudservicesrequests.rhoas.redhat.com
            version: v1alpha1
          - description: Binds Kafka Instance
            displayName: Kafka Connection
            kind: KafkaConnection
            name: kafkaconnections.rhoas.redhat.com
            version: v1alpha1
          - description: Binds Service Registry Instance
            displayName: Service Registry Connection
            kind: ServiceRegistryConnection
            name: serviceregistryconnections.rhoas.redhat.com
            version: v1alpha1
        description: "**OpenShift Application Services Operator (RHOAS-Operator)**
          creates instances of Red Hat OpenShift Application Services in your cluster.
          Operator is currently available as limited Beta and it cannot be consumed
          without invitation.\n\nRed Hat OpenShift Application Services, when added
          to Red Hat OpenShift Dedicated,  provide a streamlined developer experience
          for building, deploying,  and scaling cloud-native applications in open
          hybrid-cloud environments.   \n\nRHOAS-Operator enables developers to connect
          with their Application Services directly in their own cluster by using [RHOAS
          cli](https://github.com/redhat-developer/app-services-cli) and OpenShift
          Console UI.  \n\nRHOAS-Operator requires developers to connect with their
          Red Hat OpenShift Streams for Apache Kafka (RHOSAK) instances.  "
        displayName: OpenShift Application Services (RHOAS)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - rhoas
        - kafka
        - serviceregistry
        - app-services
        - application
        - services
        links:
        - name: RHOAS Operator Documentation
          url: https://github.com/redhat-developer/app-services-operator/tree/main/docs
        maintainers:
        - email: rhosak-eval-support@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://github.com/redhat-developer/app-services-operator/
        relatedImages:
        - quay.io/rhoas/service-operator:0.13.0
        version: 0.13.0
      entries:
      - name: rhoas-operator.0.13.0
        version: 0.13.0
      - name: rhoas-operator.0.12.0
        version: 0.12.0
      - name: rhoas-operator.0.11.0
        version: 0.11.0
      - name: rhoas-operator.0.9.6
        version: 0.9.6
      - name: rhoas-operator.0.9.5
        version: 0.9.5
      - name: rhoas-operator.0.9.3
        version: 0.9.3
      - name: rhoas-operator.0.9.2
        version: 0.9.2
      - name: rhoas-operator.0.9.1
        version: 0.9.1
      - name: rhoas-operator.0.9.0
        version: 0.9.0
      - name: rhoas-operator.0.8.2
        version: 0.8.2
      - name: rhoas-operator.0.7.10
        version: 0.7.10
      - name: rhoas-operator.0.7.9
        version: 0.7.9
      - name: rhoas-operator.0.7.8
        version: 0.7.8
      - name: rhoas-operator.0.7.6
        version: 0.7.6
      - name: rhoas-operator.0.7.2
        version: 0.7.2
      - name: rhoas-operator.0.7.1
        version: 0.7.1
      - name: rhoas-operator.0.6.8
        version: 0.6.8
      name: beta
    defaultChannel: beta
    packageName: rhoas-operator
    provider:
      name: Red Hat Inc.
      url: https://github.com/redhat-developer/app-services-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Rebaca Technologies Pvt Ltd
      provider-url: https://www.rebaca.com/telecom/solutions
    name: abot-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: abot-operator-rhmp.3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "abot.com/v3",
                "kind": "Abot",
                "metadata": {
                  "name": "abot"
                },
                "spec": {
                  "abotanalytics": {
                    "image": "registry.connect.redhat.com/rebaca/abot-analytics@sha256:c854d9eafbe26f834e10f5a97a587b2a2bf05558554897817bb8e2556124b75d",
                    "resources": {
                      "limits": {
                        "cpu": "4000m",
                        "ephemeral-storage": "8Gi",
                        "memory": "6000Mi"
                      },
                      "requests": {
                        "cpu": "2000m",
                        "ephemeral-storage": "4Gi",
                        "memory": "4000Mi"
                      }
                    }
                  },
                  "abotanalyticsbase": {
                    "image": "registry.connect.redhat.com/rebaca/abot-analyticsbase@sha256:dcc2a417a889acd6ec63538a6dfa393d4c1942dac178a04e5d65189a1ae054dc",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotbaseimage": "registry.connect.redhat.com/rebaca/abot-baseimage@sha256:2cbe38298d26be5896e0dc071aae111dd921d0dbd05c3525290f6ff31c73e926",
                  "abotbdd": {
                    "image": "registry.connect.redhat.com/rebaca/abot-bdd@sha256:d5c6bfe12a548cbdb4e0d83d5f18e5ab0c17c3af5340e437b2d0c23740c9c9e6",
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "ephemeral-storage": "8Gi",
                        "memory": "2000Mi"
                      },
                      "requests": {
                        "cpu": "1000m",
                        "ephemeral-storage": "4Gi",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotgrafana": {
                    "image": "registry.redhat.io/rhel8/grafana@sha256:d894cd34f2d8cac30e0fab69bd869f4b0f258b5086a8fd7ead0c5886ce2c5f7a",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "1500Mi"
                      },
                      "requests": {
                        "cpu": "300m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotgui": {
                    "image": "registry.connect.redhat.com/rebaca/abot-gui@sha256:78b512d6bd223a1f5d35f4cbd221f6637e2c9267726ce8899ee5d86aaac82b6f",
                    "resources": {
                      "limits": {
                        "cpu": "4000m",
                        "ephemeral-storage": "8Gi",
                        "memory": "6000Mi"
                      },
                      "requests": {
                        "cpu": "2000m",
                        "ephemeral-storage": "4Gi",
                        "memory": "4000Mi"
                      }
                    }
                  },
                  "abotkibana": {
                    "image": "registry.connect.redhat.com/elastic/kibana@sha256:2be3542d50f51ad786ead958a399f26852339e9417279688aad194c6794336ce",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "abotlogstash": {
                    "image": "registry.connect.redhat.com/elastic/logstash@sha256:dbc25dc5c31dfee08e86221af9f549b50075b2ccf7f255779c5f022c54cd892e",
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "2000Mi"
                      }
                    }
                  },
                  "abotmongo": {
                    "image": "registry.connect.redhat.com/rebaca/abot-mongodb@sha256:be9059e7e8bc1636a84575416b823c20463228785598662ee81c27e4c3c987dc",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "abotnginx": {
                    "image": "registry.connect.redhat.com/rebaca/abot-nginx@sha256:3edc22b40cd30230768115efdceec3d054d71f78214b4100b7425d8484154378",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "abotrabbitmq": {
                    "image": "registry.connect.redhat.com/rebaca/abot-rabbitmq@sha256:e609bec067fe7740590bb272db27d1a7770db47a0fc802e2288c022e101c9e25",
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "abotsidecar": {
                    "alertmanagerimage": "registry.redhat.io/rhacm2/prometheus-alertmanager-rhel8@sha256:ecaee8ad24a9c7cb3b945ee6d2c1275b9de2a676e1868abc2687e687385b3f6a",
                    "kubestatemetricsimage": "registry.redhat.io/rhacm2/kube-state-metrics-rhel8@sha256:e6bb6fa0c155bebf64749368dd951c15f1dbfa3559e2ba6734d3661d3a95c00c",
                    "prometheusimage": "registry.connect.redhat.com/kasten/prometheus@sha256:1e490fcccdd50c1026150ba4cad7c29297d2b73b3d5c80240ff31509e666363d",
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "8000Mi"
                      },
                      "requests": {
                        "cpu": "1000m",
                        "memory": "6000Mi"
                      }
                    }
                  },
                  "elasticsearch": {
                    "image": "registry.connect.redhat.com/elastic/elasticsearch@sha256:f47fd7df80ab05635e31d733f6969d38205f9f81cf95da8594a047f3d07c59ae",
                    "replicas": 3,
                    "resources": {
                      "limits": {
                        "cpu": "1200m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "800m",
                        "memory": "1000Mi"
                      }
                    },
                    "volumeClaim": {
                      "storage": "50Gi"
                    }
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "initContainers": {
                    "image": "registry.access.redhat.com/ubi8/ubi@sha256:a965f33ee4ee57dc8e40a1f9350ddf28ed0727b6cf80db46cdad0486a7580f9d",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "protocol": "https",
                  "pvc": {
                    "abotvol": "100Gi",
                    "analyticsvol": "100Gi",
                    "databak": "10Gi",
                    "mongovol": "10Gi",
                    "prometheusvol": "10Gi"
                  },
                  "replicas": 1,
                  "security": {
                    "fsGroup": 1018450000,
                    "seLinuxOptions": {
                      "level": "s0:c123,c456"
                    }
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "name": "abot"
                  },
                  "storageclass": "ocs-storagecluster-cephfs"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/rebaca/abot-operator@sha256:d3e8e1b6988827b3c636bdfcb57cda26870089d90cf2a732be522f5474c0a734
          createdAt: "2024-10-03T20:32:24Z"
          description: The ABot Test Bench is a test automation tool for 4G, 5G and
            ORAN Telecom networks that follow specifications and use standards driven
            design and compliance as specified by the 3GPP consortium.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/abot-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/abot-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=1.0.0 <3.0.0'
          operatorframework.io/suggested-namespace: abot
          operators.openshift.io/valid-subscription: '["ABot Basic License for 15
            days"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: https://abot.rebaca.com/support/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Abot
            name: abots.abot.com
            version: v3
        description: A Helm based operator for deploying Abot application in Openshift
          cluster. The ABot Test Bench is a test automation tool for 4G, 5G and ORAN
          Telecom networks that follow specifications and use standards driven design
          and compliance as specified by the 3GPP consortium. It provides a platform
          for a user to quickly check for protocol compliance and test for load characteristics
          of a particular system under test.
        displayName: Abot Operator-v3.0.0
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 5G
        - 4G
        - ORAN
        - test tool
        - compliance
        - interop
        links:
        - name: Abot Operator
          url: https://abot-operator.domain
        maintainers:
        - email: suchismita.nag@rebaca.com
          name: Suchismita Nag
        - email: samir@rebaca.com
          name: Samir Chatterjee
        maturity: alpha
        provider:
          name: Rebaca Technologies Pvt Ltd
          url: https://www.rebaca.com/telecom/solutions
        relatedImages:
        - registry.connect.redhat.com/rebaca/abot-rabbitmq@sha256:e609bec067fe7740590bb272db27d1a7770db47a0fc802e2288c022e101c9e25
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel8@sha256:e6bb6fa0c155bebf64749368dd951c15f1dbfa3559e2ba6734d3661d3a95c00c
        - registry.connect.redhat.com/rebaca/abot-bdd@sha256:d5c6bfe12a548cbdb4e0d83d5f18e5ab0c17c3af5340e437b2d0c23740c9c9e6
        - registry.connect.redhat.com/rebaca/abot-baseimage@sha256:2cbe38298d26be5896e0dc071aae111dd921d0dbd05c3525290f6ff31c73e926
        - registry.connect.redhat.com/elastic/elasticsearch@sha256:f47fd7df80ab05635e31d733f6969d38205f9f81cf95da8594a047f3d07c59ae
        - registry.connect.redhat.com/rebaca/abot-analyticsbase@sha256:dcc2a417a889acd6ec63538a6dfa393d4c1942dac178a04e5d65189a1ae054dc
        - registry.connect.redhat.com/rebaca/abot-mongodb@sha256:be9059e7e8bc1636a84575416b823c20463228785598662ee81c27e4c3c987dc
        - registry.connect.redhat.com/rebaca/abot-nginx@sha256:3edc22b40cd30230768115efdceec3d054d71f78214b4100b7425d8484154378
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e34f912158985525f6efb3c4b675f20854fc677c867d12fe3cd2e971309428dc
        - registry.connect.redhat.com/rebaca/abot-operator@sha256:d3e8e1b6988827b3c636bdfcb57cda26870089d90cf2a732be522f5474c0a734
        - registry.redhat.io/rhel8/grafana@sha256:d894cd34f2d8cac30e0fab69bd869f4b0f258b5086a8fd7ead0c5886ce2c5f7a
        - registry.connect.redhat.com/elastic/logstash@sha256:dbc25dc5c31dfee08e86221af9f549b50075b2ccf7f255779c5f022c54cd892e
        - registry.connect.redhat.com/kasten/prometheus@sha256:1e490fcccdd50c1026150ba4cad7c29297d2b73b3d5c80240ff31509e666363d
        - registry.connect.redhat.com/rebaca/abot-analytics@sha256:c854d9eafbe26f834e10f5a97a587b2a2bf05558554897817bb8e2556124b75d
        - registry.connect.redhat.com/rebaca/abot-gui@sha256:78b512d6bd223a1f5d35f4cbd221f6637e2c9267726ce8899ee5d86aaac82b6f
        - registry.connect.redhat.com/elastic/kibana@sha256:2be3542d50f51ad786ead958a399f26852339e9417279688aad194c6794336ce
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel8@sha256:ecaee8ad24a9c7cb3b945ee6d2c1275b9de2a676e1868abc2687e687385b3f6a
        - registry.access.redhat.com/ubi8/ubi@sha256:a965f33ee4ee57dc8e40a1f9350ddf28ed0727b6cf80db46cdad0486a7580f9d
        version: 3.0.0
      entries:
      - name: abot-operator-rhmp.3.0.0
        version: 3.0.0
      - name: abot-operator.2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: abot-operator-rhmp
    provider:
      name: Rebaca Technologies Pvt Ltd
      url: https://www.rebaca.com/telecom/solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: secrets-store-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secrets-store-csi-driver-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9-operator@sha256:96f3a8df9ac5ff3828b0add052af00f7d9df0ebb622a6adfd9772658d55baeb8
          createdAt: "2023-06-12T00:00:00Z"
          description: Install and configure Secrets Store CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.13.0-0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/secrets-store-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SecretProviderClass is the Schema for the secretproviderclasses
              API
            displayName: Secret Provider Class
            kind: SecretProviderClass
            name: secretproviderclasses.secrets-store.csi.x-k8s.io
            version: v1
          - description: SecretProviderClassPodStatus is the Schema for the secretproviderclassespodstatus
              API
            displayName: Secret Provider Class Pod Status
            kind: SecretProviderClassPodStatus
            name: secretproviderclasspodstatuses.secrets-store.csi.x-k8s.io
            version: v1
        description: |
          Operator that installs and configures the CSI driver for Secrets Store. The Secrets Store CSI Driver Operator is a Technology Preview feature. For more information on Red Hat Technology Preview features support scope, see https://access.redhat.com/support/offerings/techpreview/.
        displayName: Secrets Store CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: Documentation
          url: https://github.com/openshift/secrets-store-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/secrets-store-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: preview
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0627c09fe3f5a015f117fb30af1f4d17340aefa48b23994737d3901a804b63d3
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:a5704d26d0b562dba9633d1c107f34e05c750c0e3a64dbcace4dab7b964fff8d
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9-operator@sha256:96f3a8df9ac5ff3828b0add052af00f7d9df0ebb622a6adfd9772658d55baeb8
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9@sha256:148672ccd882a43c5c1d52763da465558b5ed54310418ec40c7ca0b56e408b1a
        version: 4.16.0-202509240123
      entries:
      - name: secrets-store-csi-driver-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: secrets-store-csi-driver-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: secrets-store-csi-driver-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: secrets-store-csi-driver-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: secrets-store-csi-driver-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: secrets-store-csi-driver-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: secrets-store-csi-driver-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: secrets-store-csi-driver-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: secrets-store-csi-driver-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: secrets-store-csi-driver-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: secrets-store-csi-driver-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: secrets-store-csi-driver-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: secrets-store-csi-driver-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: secrets-store-csi-driver-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: secrets-store-csi-driver-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: secrets-store-csi-driver-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: secrets-store-csi-driver-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: secrets-store-csi-driver-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: secrets-store-csi-driver-operator.v4.16.0-202410031007
        version: 4.16.0-202410031007
      - name: secrets-store-csi-driver-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: secrets-store-csi-driver-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: secrets-store-csi-driver-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: secrets-store-csi-driver-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: secrets-store-csi-driver-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: secrets-store-csi-driver-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: secrets-store-csi-driver-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: secrets-store-csi-driver-operator.v4.16.0-202408021139
        version: 4.16.0-202408021139
      - name: secrets-store-csi-driver-operator.v4.16.0-202407190036
        version: 4.16.0-202407190036
      - name: secrets-store-csi-driver-operator.v4.16.0-202407111006
        version: 4.16.0-202407111006
      - name: secrets-store-csi-driver-operator.v4.16.0-202407081937
        version: 4.16.0-202407081937
      - name: secrets-store-csi-driver-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: secrets-store-csi-driver-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: secrets-store-csi-driver-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: secrets-store-csi-driver-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: preview
    defaultChannel: preview
    packageName: secrets-store-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/limitador-operator
    name: limitador-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: limitador-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "name": "toy_get_route",
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/limitador-rhel9-operator@sha256:edb28f112a0d3536f2f52467f38df685d953f46f1999cbd61f4fd4fa4805cd34
          createdAt: "2025-04-07T15:52:04Z"
          description: Enables rate limiting for Gateways and applications in a Gateway
            API network
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Limitador is the Schema for the limitadors API
            displayName: Limitador
            kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: Enables rate limiting for Gateways and applications in a Gateway
          API network. - Part of Red Hat Connectivity Link
        displayName: Red Hat - Limitador Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        - kuadrant
        links:
        - name: GitHub
          url: https://github.com/kuadrant/limitador-operator
        - name: Documentation
          url: https://github.com/Kuadrant/limitador/tree/main/docs
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - registry.redhat.io/rhcl-1/limitador-rhel9-operator@sha256:edb28f112a0d3536f2f52467f38df685d953f46f1999cbd61f4fd4fa4805cd34
        - registry.redhat.io/rhcl-1/limitador-rhel9@sha256:98f85a05da53fbf7f20845d1969441620407cc13a185fd52714308679f116425
        version: 1.1.0
      entries:
      - name: limitador-operator.v1.1.0
        version: 1.1.0
      - name: limitador-operator.v1.0.2
        version: 1.0.2
      - name: limitador-operator.v1.0.1
        version: 1.0.1
      - name: limitador-operator.v0.12.1
        version: 0.12.1
      name: stable
    defaultChannel: stable
    packageName: limitador-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/limitador-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres-community
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.17.0-rc1
        version: 1.17.0-rc1
      - name: stackgres-community.v1.16.3
        version: 1.16.3
      - name: stackgres-community.v1.16.2
        version: 1.16.2
      - name: stackgres-community.v1.16.1
        version: 1.16.1
      - name: stackgres-community.v1.16.0
        version: 1.16.0
      - name: stackgres-community.v1.16.0-rc3
        version: 1.16.0-rc3
      - name: stackgres-community.v1.16.0-rc2
        version: 1.16.0-rc2
      - name: stackgres-community.v1.16.0-rc1
        version: 1.16.0-rc1
      - name: stackgres-community.v1.15.2
        version: 1.15.2
      - name: stackgres-community.v1.15.1
        version: 1.15.1
      - name: stackgres-community.v1.15.0
        version: 1.15.0
      - name: stackgres-community.v1.15.0-rc4
        version: 1.15.0-rc4
      - name: stackgres-community.v1.15.0-rc3
        version: 1.15.0-rc3
      - name: stackgres-community.v1.15.0-rc2
        version: 1.15.0-rc2
      - name: stackgres-community.v1.15.0-rc1
        version: 1.15.0-rc1
      - name: stackgres-community.v1.14.3
        version: 1.14.3
      - name: stackgres-community.v1.14.2
        version: 1.14.2
      - name: stackgres-community.v1.14.1
        version: 1.14.1
      - name: stackgres-community.v1.14.0
        version: 1.14.0
      - name: stackgres-community.v1.14.0-rc2
        version: 1.14.0-rc2
      - name: stackgres-community.v1.14.0-rc1
        version: 1.14.0-rc1
      - name: stackgres-community.v1.13.0
        version: 1.13.0
      - name: stackgres-community.v1.13.0-rc1
        version: 1.13.0-rc1
      - name: stackgres-community.v1.12.0
        version: 1.12.0
      - name: stackgres-community.v1.12.0-rc1
        version: 1.12.0-rc1
      name: candidate
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.17.0-rc1
        version: 1.17.0-rc1
      - name: stackgres-community.v1.16.3
        version: 1.16.3
      - name: stackgres-community.v1.16.2
        version: 1.16.2
      - name: stackgres-community.v1.16.1
        version: 1.16.1
      - name: stackgres-community.v1.16.0
        version: 1.16.0
      - name: stackgres-community.v1.16.0-rc3
        version: 1.16.0-rc3
      - name: stackgres-community.v1.16.0-rc2
        version: 1.16.0-rc2
      - name: stackgres-community.v1.16.0-rc1
        version: 1.16.0-rc1
      - name: stackgres-community.v1.15.2
        version: 1.15.2
      - name: stackgres-community.v1.15.1
        version: 1.15.1
      - name: stackgres-community.v1.15.0
        version: 1.15.0
      - name: stackgres-community.v1.15.0-rc4
        version: 1.15.0-rc4
      - name: stackgres-community.v1.15.0-rc3
        version: 1.15.0-rc3
      - name: stackgres-community.v1.15.0-rc2
        version: 1.15.0-rc2
      - name: stackgres-community.v1.15.0-rc1
        version: 1.15.0-rc1
      - name: stackgres-community.v1.14.3
        version: 1.14.3
      - name: stackgres-community.v1.14.2
        version: 1.14.2
      - name: stackgres-community.v1.14.1
        version: 1.14.1
      - name: stackgres-community.v1.14.0
        version: 1.14.0
      - name: stackgres-community.v1.14.0-rc2
        version: 1.14.0-rc2
      - name: stackgres-community.v1.14.0-rc1
        version: 1.14.0-rc1
      - name: stackgres-community.v1.13.0
        version: 1.13.0
      - name: stackgres-community.v1.13.0-rc1
        version: 1.13.0-rc1
      - name: stackgres-community.v1.12.0
        version: 1.12.0
      - name: stackgres-community.v1.12.0-rc1
        version: 1.12.0-rc1
      name: fast
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.16.3
        version: 1.16.3
      - name: stackgres-community.v1.16.2
        version: 1.16.2
      - name: stackgres-community.v1.16.1
        version: 1.16.1
      - name: stackgres-community.v1.16.0
        version: 1.16.0
      - name: stackgres-community.v1.15.2
        version: 1.15.2
      - name: stackgres-community.v1.15.1
        version: 1.15.1
      - name: stackgres-community.v1.15.0
        version: 1.15.0
      - name: stackgres-community.v1.14.3
        version: 1.14.3
      - name: stackgres-community.v1.14.2
        version: 1.14.2
      - name: stackgres-community.v1.14.1
        version: 1.14.1
      - name: stackgres-community.v1.14.0
        version: 1.14.0
      - name: stackgres-community.v1.13.0
        version: 1.13.0
      - name: stackgres-community.v1.12.0
        version: 1.12.0
      name: stable
    defaultChannel: stable
    packageName: stackgres-community
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: stolostron-engine
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stolostron-engine.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:b7d0c8413a80c251779fe3c1c9200f9fcfbbe085ccee3ab968ebd4a39209946c
          createdAt: "2024-06-28T02:42:26Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/work@sha256:4ef2518aa107ebd28d5fb1a98d63d82a8e0610544600169e0dd5f345599e518b
        - quay.io/stolostron/assisted-installer@sha256:a31ed7b8e4ff2f9e5ade2e61191cc759e1d47ddbba95a2af41bfa53fa6ddf855
        - quay.io/stolostron/cluster-api-provider-agent@sha256:c074ca1f28d03e86efe4f09200a38af4ac947247a19c46c5300eefcd4f337006
        - quay.io/stolostron/placement@sha256:0dbcc7c62569920afcad1443529f10e38d651a5fe5ae14ade89d6e42fb522e60
        - quay.io/stolostron/managed-serviceaccount@sha256:4c3f1e0b9c4f2937ceb46e6b3f63e3a07061f9b29fdda7dcfdf9ebd74c3bb30b
        - quay.io/stolostron/backplane-operator@sha256:b7d0c8413a80c251779fe3c1c9200f9fcfbbe085ccee3ab968ebd4a39209946c
        - quay.io/stolostron/assisted-installer-controller@sha256:213dc6fde247a5b9862f5076d40394164dd596cde7ca2061e58b73f4d2dbc89e
        - quay.io/stolostron/hive@sha256:fd39b8340cfaeb9849e83bd910e8a3575791fcc58dbbdeb3e66528f009b33d65
        - quay.io/stolostron/cluster-image-set-controller@sha256:632661c6cbfce7df0545bc2fccf1ef87257abb3cc92dcede054806613a3540e3
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:f0d6d46d3a4d65fba6b74bdc9eab99e8149133f83a5c8486f22a73ab92f78ed6
        - quay.io/stolostron/assisted-service@sha256:653728f209070c6adad846a4780449f56985b19006024cfe8137a3a1854fdc85
        - quay.io/stolostron/registration@sha256:8dca354dd63739d7578c24d57b2a72c0676c979344e05c6f2634a8a9fae3218d
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:26f8bb3c9b531e96e6538a5bff96569628f45c06eb3b991ebb421d4e90d7c7e3
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/registration-operator@sha256:4b6c31592c405db11375f4694a7f15bb13c5c9a770786a1607ea0be615396435
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/discovery-testserver@sha256:83d7d6cfaef629a4e3d5d0e9fe35456aa220f1eb68f6fe611d85232acf46bc73
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/discovery-operator-tests@sha256:c98c42b0c16306de86d7b365096cb809a0c797b0a4b7dc04a7de877d0ba5733a
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:b76a201edcad4805e32c40d3f0ad3f61584a879aa00c9e49aae3b2985bfb5fbc
        - quay.io/stolostron/addon-manager@sha256:87c829911438061a53aff2b297416bf43080fdd8d93558caefc198e6f9cb382f
        - quay.io/stolostron/backplane-operator-test@sha256:961d8c40aa3526ca368fd6833202800ba2eb344c97788a646d6fb350a5338397
        - quay.io/stolostron/hypershift-operator@sha256:7dd5eaf195e6b1ca4cecd50514f5b7045948184f9ee617341bc90edc5f9914f2
        - quay.io/stolostron/backplane-must-gather@sha256:2676ebb7438ca42f0d852100612715c429fe9fbe072297fc178d51c34378d1af
        - quay.io/stolostron/discovery-operator@sha256:acb2c4033da62055feb49ff3023b2850a2768996e1fa80f52588b3e8e458e8f5
        - quay.io/stolostron/assisted-image-service@sha256:0c29a74e0f82d1782d5170ff674f5bfcd89ecc5d180fe5b583a5e0f388b9504b
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:a39c52097faf7a21cc717768401851466f8c9341f08103e4d1fe8459ae7f5ea2
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/helloworld@sha256:abc0d8002c5fe6d95a1f24fce1b7899ca24fac576c894ea7443cf7e15e6a42c2
        - quay.io/stolostron/image-based-install-operator@sha256:b9af0ddb7ae8e6fd29b953dfb917546dfd676206e2ad143e892172b6e1cea5a1
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:b269d88e30551636e024aa1472ea4dd71b873b5a2778a199b32732488fd776fb
        - quay.io/stolostron/multicloud-manager@sha256:05da692b2780453115ef6c195658cadb2f28eb0f3a86dc0a586e459dd7c55afe
        - quay.io/stolostron/cluster-curator-controller@sha256:e3a91a4b06dd73fc4e401ddf034e61d4d5b2d324d2c5cea92941e1a56c4023c1
        - quay.io/stolostron/hypershift-addon-operator@sha256:47ed6a7c00036599051718d6015bee35818ea5bbaa14f1c01bedb338dce90668
        - quay.io/stolostron/assisted-installer-agent@sha256:a2efc62ca0080779acfea2e413e524608bc0e8f72b2b917e6c7e76864d08b2b6
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/provider-credential-controller@sha256:d3bd513c06591786216f2baab46c9f57e373d1939f038edacecb2180f673d888
        - quay.io/stolostron/cluster-proxy@sha256:6cd53c588bb9c5064ba0eb8eab03cd12c86976b353209102059f7d76ec59d7c3
        - quay.io/stolostron/cluster-proxy-addon@sha256:f70a26b276c67ad791d337e35cd102afa82f1005bc6ecbc60309932d2d163a6b
        - quay.io/stolostron/managedcluster-import-controller@sha256:925bbc6173501bdb60d329a1f5a134ea3e9230034509320f93382f1c570be275
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/console-mce@sha256:4e49f2b2fdb0491c225e4d5dc143a6d7c0a25cd855057b3792c7ee94a868ac85
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/assisted-service-el8@sha256:31422ffce8b94d662131986bd084fcba503809cb939389096f9c9dbb27cb58db
        - quay.io/stolostron/clusterclaims-controller@sha256:a0613147b0ec010b8a4442a3c515f7ffc48ec7d8137436a0e8b935443cac2f96
        version: 0.4.0
      entries:
      - name: stolostron-engine.v0.4.0
        version: 0.4.0
      name: community-0.4
    - currentCSV: stolostron-engine.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:30d5bcccc542988ac788dbca305e6f29fb678470b19e48adada9ca6f5803f04a
          createdAt: "2024-10-02T05:51:59Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.5.0 <0.5.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/assisted-installer@sha256:dcc4e28c7eea55dd2ac22dd75c2b7b96ee3552fcbeba8128ad8528fda235d879
        - quay.io/stolostron/cluster-curator-controller@sha256:fc2ad8e0b6701638ab89bb62a49700fdaadfe1149228cd0b3cd4a41810020c6d
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/work@sha256:4467a4bd00586ab837cb7e8bf56ea1faff82ed5ce2fffdeb7e51ee5fafb433b0
        - quay.io/stolostron/cluster-image-set-controller@sha256:654e218ecf0a3f67520c0521b544eea9d4552e92f54bf424e88ccab24e7c151e
        - quay.io/stolostron/assisted-installer-agent@sha256:16b040dd0690b62bc051aa6369272d004d2157a8e5230d318ba827adba8e8d63
        - quay.io/stolostron/hypershift-addon-operator@sha256:ba199ebfe0eea3ef6b07123f51116831e7b0a73860b0410ece723bcc9a3230af
        - quay.io/stolostron/placement@sha256:dc0ad451b085b3a747ee648475f8e8102b645df69b898fc88f1cb190d1b95a83
        - quay.io/stolostron/cluster-proxy@sha256:be09bb293a990600ec48aba2f9bf8b457d5bdbc6b136be319e5feebadb232a8c
        - quay.io/stolostron/discovery-operator@sha256:db56e2a78ed27a46b47b519e29db66636b9abb499fd507d99451bae47438e32b
        - quay.io/stolostron/hive@sha256:707ac9a2ad18dbc4c4686f1b72946077c7232afd05505d5e50e13f806a452392
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:4b5b2bea2aad6c7e16fe1520a2fb8454b0312f2749d2f398ca54e644e813055c
        - quay.io/stolostron/cluster-api-provider-agent@sha256:1458de75311675a5acec7361e507edb24b970ffa3403e916de3413f7061bfe36
        - quay.io/stolostron/clusterclaims-controller@sha256:1c0b3c4730ef1ab628bf2c1b7cf12b1b93b149abdefc4d908dbdc90e4b6c1c19
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:2fe53994fa5bde93112a2cc75758e0f68be013a34c8206066d0b315c2e6a6f47
        - quay.io/stolostron/assisted-service-el8@sha256:c8248f9a7ac02f47373d097953cee38b66125e98b98922f18c578a591f04bcca
        - quay.io/stolostron/console-mce@sha256:10a54c5012ef6439539297919c71d9b53aaaddc6342d90677113d6ce7b42e3fe
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/backplane-operator-test@sha256:048f21cf89fb5ed2557ffc90ae5d7104cdd5ce0e6a2a865ce6f4230d09400115
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:b76a201edcad4805e32c40d3f0ad3f61584a879aa00c9e49aae3b2985bfb5fbc
        - quay.io/stolostron/addon-manager@sha256:9384ae0e9bcc9f85d3fe36e44ab7b3a61777197846559b5add0c56f513e35333
        - quay.io/stolostron/assisted-image-service@sha256:703c12a918d3c66b62c29a5abd0e8c5032299289cfed459671d6ca82c4f35c73
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/assisted-service@sha256:6383ba508a0cac030a46df96ac37daaeddc27ad1fb41d8b4dca5a87cf4ab9c31
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:14a0648b76301de182154a8d7d6b7e7eca0bfbed226e6dcebddb75af5152451f
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/registration-operator@sha256:b42341143b1e1693748f4b7d69e97c56c8a5f9408abc5db1de96ba246b7b98a9
        - quay.io/stolostron/backplane-operator@sha256:30d5bcccc542988ac788dbca305e6f29fb678470b19e48adada9ca6f5803f04a
        - quay.io/stolostron/discovery-operator-tests@sha256:e8d3d1675242272cc139b0dba15605d00e25b17b4b1cc1b5b4f3ea8913ffb2da
        - quay.io/stolostron/discovery-testserver@sha256:0b23021d6c0c7d6c7aa97a92fd0c40f31fdff6363a5b10a1edc864b5d8d54108
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/multicloud-manager@sha256:6dd212840ea7fadd427b503ce69e6e32464881a0047e2af86f9f2c7562bd4703
        - quay.io/stolostron/managedcluster-import-controller@sha256:01ab0c8b664692d69c2a9b3363e007a5611c332ac0e3491234dd0cbfba31c88a
        - quay.io/stolostron/registration@sha256:4779cf430d479806efd2f3cb7e2a9ed3dce749287609c39ad45f1b06e220d023
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/managed-serviceaccount@sha256:667f823799984404c0c1589df7b857a6a35218c2653254905dc22ca90ce1542b
        - quay.io/stolostron/backplane-must-gather@sha256:92e992e23855375a8bae542480d789d1bb25a48959896bfb75f659df2a37dd98
        - quay.io/stolostron/image-based-install-operator@sha256:bd76503406e6d20693e636461411e158c8dde3d03c4904ecb8a1850deeefa4f5
        - quay.io/stolostron/hypershift-operator@sha256:a30fb6cf4d5e103820813019ce452d8134b0fee8ecd3c2166d5681199d4a0cad
        - quay.io/stolostron/cluster-proxy-addon@sha256:7aa09b0d53fdc0f5cec9c8a13c160277a9274466025f11b5b04da95c9a10e59c
        - quay.io/stolostron/provider-credential-controller@sha256:d240ad43c23de822f91a33849448ae70a4cf27570da3387bdecd101ca2f0b65d
        - quay.io/stolostron/assisted-installer-controller@sha256:01dd7c7fb9cb335215afefde204309d3127847b900658ee9744b9d31df847565
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/helloworld@sha256:96a9b8855fb8800a85aed6fdc4911295dc2c12629bbdcc34ed8dc8061827b17e
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        version: 0.5.3
      entries:
      - name: stolostron-engine.v0.5.3
        version: 0.5.3
      - name: stolostron-engine.v0.5.2
        version: 0.5.2
      - name: stolostron-engine.v0.5.1
        version: 0.5.1
      - name: stolostron-engine.v0.5.0
        version: 0.5.0
      name: community-0.5
    - currentCSV: stolostron-engine.v0.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:fa1fa6ef3958fd9d50bfb38197f033a635a05b47c32e653822246e708d8598cc
          createdAt: "2025-02-24T01:37:53Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.6.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/discovery-operator-tests@sha256:21449ecd291cf7203cc8e27a2a00ac60e74f5993708f9e7465e8efdfdca1c89f
        - quay.io/stolostron/assisted-installer@sha256:c88976ab4b35d458a39044e522d2d5ab1916d8485af07bbef373af1bcbbd6af0
        - quay.io/stolostron/assisted-service-el8@sha256:f50f4dece91100f45ef9e20b1e52557a55db45b17a380f32d365ee013d695d5c
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:b016a001a53bc37ae534b74cc832552759800746c27e16adcf0626ff9ad2e4c0
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:2781d87c1568f2e40409f5431bfd1a9a4c9ca35655537e7d4db65d9b881edd08
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        - quay.io/stolostron/placement@sha256:543f09446652bff8c4149945d94d715eff7a76d82783f2bad89e712798542c73
        - quay.io/stolostron/provider-credential-controller@sha256:a5ac3999d0aefb1dd9a239f9394485f93e4252a45eb4a43a43e603aaff80b3bb
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:44f8bb72b975456d4151a88a560fc525ad6f4da7968a38e94a86b6dd77d3c543
        - quay.io/stolostron/registration@sha256:9c221a40ffa119ac154d76f578e0919b5a63137c908afb0583519a9b8a399b2b
        - quay.io/stolostron/hypershift-operator@sha256:2d1809223ce5e75b4133a39b78a269703c6b5234f7ec2a7c0072124c345fae0a
        - quay.io/stolostron/clusterclaims-controller@sha256:b796cb2edd3d4c22b29fc6d2b45e5328a59b7ec1bdc52d5500e1a4ee74725d6c
        - quay.io/stolostron/cluster-api-provider-agent@sha256:dda3e9aef611c744c4c89c2689a12d0649a173c1a64954c3037de55ec8eacec9
        - quay.io/stolostron/multicloud-manager@sha256:0293f0852206562b891375b167cfe17843e2e17585f04c9bb6482d1fe78089d0
        - quay.io/stolostron/backplane-operator@sha256:fa1fa6ef3958fd9d50bfb38197f033a635a05b47c32e653822246e708d8598cc
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/discovery-testserver@sha256:f757a7a11108a88de4bc0ecc5e16f74b3bd9b5110ec94bfeef2074121ec1e47f
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/cluster-proxy-addon@sha256:7da9da3e4ecd1e0df636c9f72f9f30cc7ff24b81e247ba9825cef39f18e86182
        - quay.io/stolostron/image-based-install-operator@sha256:fecbff73eb97b28bedd7c17d5adbef4826f181a6838ad4cc72e7478acf9b4254
        - quay.io/stolostron/hive@sha256:1e1f60492c6a3b43b69fe8a0640f9b81f5a1cfafc0afcd6854c3cfe362ed063a
        - quay.io/stolostron/cluster-proxy@sha256:b341dd7058cdffccb00a026908031fa123fec5ba58ba9307dbb873b75d250e4d
        - quay.io/stolostron/cluster-image-set-controller@sha256:629361c1c0ec50fe6410fc6fe6417cba89957c49ab89201c6470c0f97c8fc91d
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/assisted-image-service@sha256:6846719f91dd954ce4b4c6e7c9654c7d79db72990270acd3c34bdd2a89af34fc
        - quay.io/stolostron/console-mce@sha256:bfaf9180cd1e5af4d34daf4f1dcdee89951067a605f21bfe6e1ce6bda85b0334
        - quay.io/stolostron/helloworld@sha256:4838f699ea2461f8358dcda32b3ac96ed05b55d4e1958eac2bdacae4807501d1
        - quay.io/stolostron/ose_cluster_api_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/assisted-installer-controller@sha256:3b0b43fdbbbbd5c1e7588fc449b002e1ed9ad0ea6daa774925a962bd499a4195
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/work@sha256:6ec081826f32467ea1e1a91f6624df349a382fb8a1cbbda906141281ac2bad5c
        - quay.io/stolostron/ose_aws_cluster_api_controllers_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/backplane-operator-test@sha256:b04661ad658423c9c7557b722adb2159d988c10445de1c59f39c44583262cb42
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/assisted-installer-agent@sha256:cd74c8b8888c100bf567dd07cfff40097c458d18fcc29a11ed3830f697156267
        - quay.io/stolostron/backplane-must-gather@sha256:99b558d687e1c17036b164d819cb1c9b96a814be79e3c87960ef7cdb521eee7d
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/discovery-operator@sha256:43f40b9c20b112248fcdb42d114fe4af73fdd45b05d9ff7855d827cdd5e971b4
        - quay.io/stolostron/managed-serviceaccount@sha256:0e4edd9ab95c1558fba40a0c41c5cc326996e9e19e296324ea10b5a442a7ef63
        - quay.io/stolostron/hypershift-addon-operator@sha256:65f2b7a356c22413307ce7b7de38ce00c31e4c43b6ee3b4b484c0bc61dc483e9
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:65180432218d035ffc7996a7643ddb5300eb8b09e47e09cb270cb9b961aecf7a
        - quay.io/stolostron/managedcluster-import-controller@sha256:ecfe7c1f2d24a75bfb8db2a0a2765fe5e9136f81f341882de896431570fd10e9
        - quay.io/stolostron/addon-manager@sha256:7e347a177d8c5b80a78411a3e8dbf04e49d157a4ca67e7fd5170970c2ba62da9
        - quay.io/stolostron/registration-operator@sha256:52289e6cf5f159d344717b49e6226ff9d4ee33942f63cc543d1b820d7ee24fc5
        - quay.io/stolostron/assisted-service@sha256:58271e54eda979c43fb474c3bb255f7328e380a7af9ba833e5f7558f323bd239
        - quay.io/stolostron/cluster-curator-controller@sha256:5d05462c2c9d93b21656a436c01fbda4256757bd057e5de7014eb6f320310a1e
        version: 0.6.3
      entries:
      - name: stolostron-engine.v0.6.3
        version: 0.6.3
      - name: stolostron-engine.v0.6.2
        version: 0.6.2
      - name: stolostron-engine.v0.6.1
        version: 0.6.1
      - name: stolostron-engine.v0.6.0
        version: 0.6.0
      name: community-0.6
    - currentCSV: stolostron-engine.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:9b39a5b6a502778d7884e276b276a9b44e6dbc88e7e82470e26d91ded3cb3839
          createdAt: "2025-06-09T15:04:58Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.7.1'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational components for managing multiple Kubernetes-based clusters. The deployment of the multicluster engine components is
              determined based on the configuration that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/cluster-image-set-controller@sha256:7f05ad13ffaa11daec066c5889b87df1b631e6eb3dc1c2913a2766c8e65218aa
        - quay.io/stolostron/cluster-proxy@sha256:1528e36dfa4fa2c9063d84c0f2b9aa8d2b549a750e812d4a2e2b8c7515bfe5cb
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:1e95f55d6918263e79de1c1e2be8a96cd3d35a346dde6d3e2fb484f48e567811
        - quay.io/stolostron/hive@sha256:86ff94d3e7c174ac4d6818183c66791edc7638965dd09e2d41172609c81f451f
        - quay.io/stolostron/cluster-api-provider-openshift-assisted-bootstrap@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/console-mce@sha256:7f538a21eb22ccd57765f47692da23be9cecfd3b7b65986f1413234bdf709530
        - quay.io/stolostron/assisted-installer-agent@sha256:ffceb06ce038013291e092324d3c49091cd1b2ff1569abe77c29484075e28107
        - quay.io/stolostron/hypershift-addon-operator@sha256:ed1c1580fee28310fc1b137c5c70271448e58ed29af4371515fe9c74512a2c7e
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/placement@sha256:6276e411c9f7cf7072d3d364569b6c3cd472742ab4ad48912f18cb4091d27a6f
        - quay.io/stolostron/ose_cluster_api_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/cluster-proxy-addon@sha256:c776405c4b1b623500c69a12b7d76563646b27cf23bab74dc8f1361534cfb81e
        - quay.io/stolostron/addon-manager@sha256:a245847f135f96c4b4fcaa673fc32013c8e407252b3259175c796e3c16111c71
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/discovery-testserver@sha256:faef24208a65283b4157521079a7bf23bc4ebd553d70a83a78c324b5cb0e24e4
        - quay.io/stolostron/discovery-operator@sha256:a6a37237bc06f0635a936b803d3c53f0f0caf5426c76b37951c4e157b181ecb3
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:1d4221cfc364743b58e15bad3afe9548f4151f06b0dbf0aec7c00715ba75abba
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/ip-address-manager@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/assisted-installer@sha256:17b78d73775a93b31ca872db79520ddc49d2be042fdf1599f419fbcbfd9a98ac
        - quay.io/stolostron/assisted-service-el8@sha256:f52c55691850a1f2f254b0e9377ba97493be337e012223e5a2b3190f03e1a349
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:44f8bb72b975456d4151a88a560fc525ad6f4da7968a38e94a86b6dd77d3c543
        - quay.io/stolostron/discovery-operator-tests@sha256:d96fe1f4f49ecfd4a27bedd87d7070a3c7f3f82b6ee8bb59e98a262d8cedcfc0
        - quay.io/stolostron/clusterclaims-controller@sha256:afab145b12a52f81e1f2eaa189731e3a6287b54f2f1f1ad2976ad378fa7dfa3a
        - quay.io/stolostron/ose_aws_cluster_api_controllers_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/backplane-operator@sha256:9b39a5b6a502778d7884e276b276a9b44e6dbc88e7e82470e26d91ded3cb3839
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        - quay.io/stolostron/assisted-image-service@sha256:b6df9c9fc8bbfdc58e2169b9e199e1eed188d3d847511eba01eb4920fabe3992
        - quay.io/stolostron/multicloud-manager@sha256:5a95a5d1dbb17456d7ce62f1a517ae49ca780e25f3ebb74b32346de1df23b4c9
        - quay.io/stolostron/helloworld@sha256:a790795409f6e781f506b6a63f39f7d344f01efa30c3de87b1727dc2010d5bbf
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/backplane-must-gather@sha256:f4f9596eade78333628199f62f8e71e4c2b3957c20da5bad807c12a2a02560db
        - quay.io/stolostron/managedcluster-import-controller@sha256:4d0be693db0576eb2e9e3d409bf1479cc60784f5747ed1b25b6a2443c5e91e36
        - quay.io/stolostron/cluster-api-provider-agent@sha256:2b22890b948ad1d597c424d73d4fe34c7c4d265ad6e54fa0ed87a7dec920ac57
        - quay.io/stolostron/ose-baremetal-cluster-api-controllers-rhel9@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/cluster-curator-controller@sha256:d667bd0892812eb4e1f724679759469db335db4ab801f9d57ea8818d8cd13188
        - quay.io/stolostron/image-based-install-operator@sha256:9d6796007b616de60560f55f14546295c4e48b3768d059098fce9367674a3368
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/registration@sha256:2db2c28af25007c77979cc456d929f5f077cd671cc4e0562c43efce2c42c4cd8
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/provider-credential-controller@sha256:3484cb0c2102f4a522efe31bc4c77910a202b42e1ad082a0a8709cc1a7698649
        - quay.io/stolostron/work@sha256:709c41d1a7a1eaf10b334dd0c0105a261d6db811693d5ae67eb55b9a26daed87
        - quay.io/stolostron/assisted-installer-controller@sha256:8ea9d12fa30825fe4f965a7d506af687e8de709f8d07dd75204747da2a6e09a8
        - quay.io/stolostron/assisted-service@sha256:af3532d487e03a077ce23e4178abbcb41cf45726b646d7660840c036677715e8
        - quay.io/stolostron/hypershift-operator@sha256:c9a8fc943ef35126434cc767f152abd09943a4d2d194a9966495fdc0870b417c
        - quay.io/stolostron/registration-operator@sha256:a9fdf837e149c8659fc99deab6a284975c8eb8785549381fa082303cdb172e44
        - quay.io/stolostron/managed-serviceaccount@sha256:4bcd5c9b2044254c28215fd78bf1e3fcd919578238986831fcea8055b2764696
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:1757b9ad635839087c4bef198d5e81f4abb7aee38bc74690021f9e8887366bcf
        - quay.io/stolostron/backplane-operator-test@sha256:8d6514fc0db5ee79c346ddc31bb094ad87fa375519da67e32398300d70d72f6e
        - quay.io/stolostron/cluster-api-provider-openshift-assisted-control-plane@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        version: 0.7.1
      entries:
      - name: stolostron-engine.v0.7.1
        version: 0.7.1
      - name: stolostron-engine.v0.7.0
        version: 0.7.0
      name: community-0.7
    defaultChannel: community-0.7
    packageName: stolostron-engine
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pulp Community
      provider-url: https://github.com/pulp/pulp-operator
    name: pulp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pulp-operator.v1.0.0-alpha.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache_enabled": true,
                  "content": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "ingress_type": "Route",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/"
                  },
                  "storage_type": "File",
                  "worker": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example-pulp"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true,
                    "redis_storage_class": "standard"
                  },
                  "content": {
                    "replicas": 1
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "pulp",
                  "file_storage_access_mode": "ReadWriteOnce",
                  "file_storage_size": "2Gi",
                  "file_storage_storage_class": "standard",
                  "image_version": "nightly",
                  "image_web_version": "nightly",
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/",
                    "telemetry": false
                  },
                  "web": {
                    "replicas": 1
                  },
                  "worker": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "galaxy-example"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "api": {
                    "replicas": 1
                  },
                  "content": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "galaxy",
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "file_storage_storage_class": "standard",
                  "image": "quay.io/pulp/galaxy-minimal",
                  "image_version": "nightly",
                  "image_web": "quay.io/pulp/galaxy-web",
                  "image_web_version": "nightly",
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/api/galaxy/pulp/",
                    "telemetry": false
                  },
                  "signing_scripts_configmap": "signing-scripts",
                  "signing_secret": "signing-galaxy",
                  "storage_type": "File",
                  "web": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "worker": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "galaxy-example",
                  "deployment_type": "galaxy",
                  "instance_name": "pulp",
                  "postgres_configuration_secret": "galaxy-example-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "galaxy-example"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.0.0-alpha.1
          createdAt: "2021-03-26 16:57:40"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1alpha1
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1alpha1
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1alpha1
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://docs.pulpproject.org/](https://docs.pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          Note that Pulp operator works with three different types of service containers (the operator itself, the main service and the web service):

          |           | Operator | Main | Web |
          | --------- | -------- | ---- | --- |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[pulp-minimal](https://quay.io/repository/pulp/pulp-minimal?tab=tags) | [pulp-web](https://quay.io/repository/pulp/pulp-web?tab=tags) |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[galaxy-minimal](https://quay.io/repository/pulp/galaxy-minimal?tab=tags) | [galaxy-web](https://quay.io/repository/pulp/galaxy-web?tab=tags) |

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://docs.pulpproject.org/pulp_operator/container/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://docs.pulpproject.org/pulp_operator/](https://docs.pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://docs.pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: beta
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - docker.io/library/postgres:13
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/pulp/pulp-operator:v1.0.0-alpha.1
        - quay.io/pulp/pulp-minimal:stable
        - quay.io/pulp/pulp-web:stable
        - docker.io/library/redis:latest
        version: 1.0.0-alpha.1
      entries:
      - name: pulp-operator.v1.0.0-alpha.1
        version: 1.0.0-alpha.1
      name: alpha
    - currentCSV: pulp-operator.v1.0.0-beta.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "Pulp",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true
                  },
                  "content": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/"
                  },
                  "worker": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "Pulp",
                "metadata": {
                  "name": "galaxy-example"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "api": {
                    "replicas": 1
                  },
                  "content": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "galaxy",
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "file_storage_storage_class": "standard",
                  "image": "quay.io/pulp/galaxy-minimal",
                  "image_version": "4.6.3",
                  "image_web": "quay.io/pulp/galaxy-web",
                  "image_web_version": "4.6.3",
                  "ingress_type": "nodeport",
                  "nodeport_port": 30000,
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "analytics": false,
                    "ansible_api_hostname": "http://nodeport.local:30000",
                    "api_root": "/api/galaxy/pulp/",
                    "content_origin": "http://nodeport.local:30000",
                    "galaxy_feature_flags": {
                      "ai_deny_index": true,
                      "execution_environments": true
                    },
                    "token_server": "http://nodeport.local:30000/token/"
                  },
                  "signing_scripts": "signing-scripts",
                  "signing_secret": "signing-galaxy",
                  "web": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "worker": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "galaxy-example",
                  "deployment_type": "galaxy",
                  "postgres_configuration_secret": "galaxy-example-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "galaxy-example"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.0.0-beta.4
          createdAt: "2024-01-18T20:33:37Z"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1beta2
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1beta2
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1beta2
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://docs.pulpproject.org/](https://docs.pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          Note that Pulp operator works with three different types of service containers (the operator itself, the main service and the web service):

          |           | Operator | Main | Web |
          | --------- | -------- | ---- | --- |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[pulp-minimal](https://quay.io/repository/pulp/pulp-minimal?tab=tags) | [pulp-web](https://quay.io/repository/pulp/pulp-web?tab=tags) |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[galaxy-minimal](https://quay.io/repository/pulp/galaxy-minimal?tab=tags) | [galaxy-web](https://quay.io/repository/pulp/galaxy-web?tab=tags) |

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://docs.pulpproject.org/pulp_operator/container/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://docs.pulpproject.org/pulp_operator/](https://docs.pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://docs.pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: beta
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.4
        - quay.io/pulp/pulp-operator:v1.0.0-beta.4
        - quay.io/pulp/pulp-minimal:stable
        - quay.io/pulp/pulp-web:stable
        - docker.io/library/redis:latest
        - docker.io/library/postgres:13
        version: 1.0.0-beta.4
      entries:
      - name: pulp-operator.v1.0.0-beta.4
        version: 1.0.0-beta.4
      - name: pulp-operator.v1.0.0-beta.3
        version: 1.0.0-beta.3
      - name: pulp-operator.v1.0.0-beta.2
        version: 1.0.0-beta.2
      - name: pulp-operator.v1.0.0-beta.1
        version: 1.0.0-beta.1
      - name: pulp-operator.v1.0.0-alpha.9
        version: 1.0.0-alpha.9
      - name: pulp-operator.v1.0.0-alpha.8
        version: 1.0.0-alpha.8
      - name: pulp-operator.v1.0.0-alpha.7
        version: 1.0.0-alpha.7
      - name: pulp-operator.v1.0.0-alpha.6
        version: 1.0.0-alpha.6
      - name: pulp-operator.v1.0.0-alpha.5
        version: 1.0.0-alpha.5
      - name: pulp-operator.v1.0.0-alpha.4
        version: 1.0.0-alpha.4
      - name: pulp-operator.v1.0.0-alpha.3
        version: 1.0.0-alpha.3
      - name: pulp-operator.v1.0.0-alpha.2
        version: 1.0.0-alpha.2
      - name: pulp-operator.v1.0.0-alpha.1
        version: 1.0.0-alpha.1
      name: beta
    defaultChannel: beta
    packageName: pulp-operator
    provider:
      name: Pulp Community
      url: https://github.com/pulp/pulp-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-common-service-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-common-service-operator.v4.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v3",
                "kind": "CommonService",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-common-service-operator",
                    "app.kubernetes.io/managed-by": "ibm-common-service-operator",
                    "app.kubernetes.io/name": "ibm-common-service-operator"
                  },
                  "name": "example-commonservice"
                },
                "spec": {
                  "size": "starterset"
                }
              }
            ]
          capabilities: Seamless Upgrades
          cloudPakThemesVersion: styles4100.css
          containerImage: registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
          createdAt: "2025-03-18T21:27:31Z"
          description: The IBM Cloud Pak foundational services operator is used to
            deploy IBM foundational services.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-common-service-operator
          nss.operator.ibm.com/managed-webhooks: ""
          olm.skipRange: '>=3.3.0 <4.12.0'
          operatorChannel: v4.12
          operatorVersion: 4.12.0
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/ibm-common-service-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CommonService is the Schema for the commonservices API. Documentation
              For additional details regarding install parameters check https://ibm.biz/icpfs39install.
              License By installing this product you accept the license terms https://ibm.biz/icpfs39license
            displayName: CommonService
            kind: CommonService
            name: commonservices.operator.ibm.com
            version: v3
        description: |-
          # Introduction

          The `IBM Cloud Pak foundational services` is a bridge to connect IBM Cloud Paks and Operand Deployment Lifecycle Manager (ODLM) with IBM Cloud Pak foundational services. You can also install the `IBM Cloud Pak foundational services` in stand-alone mode.

          When you install this operator, the operator completes the following tasks:

            - Installs ODLM operator namespace
            - Creates the Common Services `OperandRegistry` and `OperandConfig` in the service namespace

          # Details
          For more information about installing this operator and other Common Services operators, see [Installer documentation](http://ibm.biz/cpcs_opinstall). If you are using this operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For more information about IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).

          For more information about the available IBM Cloud Pak foundational services, see the [IBM Knowledge Center](http://ibm.biz/cpcsdocs).

          ## Supported platforms

          Red Hat OpenShift Container Platform 4.5 or newer installed on one of the following platforms:

           - Linux x86_64
           - Linux on Power (ppc64le)
           - Linux on IBM Z and LinuxONE

          ## Operator versions

           - 4.12.0

          ## Prerequisites

          Before you install this operator, you need to first install the operator prerequisites:

          - For the list of prerequisites for installing the operator, see the IBM Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).

          ### Resources Required

          #### Minimum scheduling capacity

          | Software                   | Memory (MB) | CPU (cores) | Disk (GB) | Nodes  |
          | -------------------------- | ----------- | ----------- | --------- | ------ |
          | ibm common service operator | 200          | 0.2        | 1          | worker |
          | **Total**                  | 200         | 0.2         | 1         |        |

          ## Documentation

          - If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak. For a list of IBM Cloud Paks, see [IBM Cloud Paks that use IBM Cloud Pak foundational services](http://ibm.biz/cpcs_cloudpaks).
          - If you are using the operator in stand-alone mode or with an IBM Containerized Software, see the IBM Cloud Pak foundational services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall).

          ## Installing

          For installation, see the [IBM Cloud Platform Common Services documentation](http://ibm.biz/cpcsdocs).

          ## Configuration

          For configuration, see the [IBM Cloud Platform Common Services documentation](http://ibm.biz/cpcsdocs).

          ### SecurityContextConstraints Requirements

          The IBM Common Service Operator supports running with the OpenShift Container Platform 4.3 default restricted Security Context Constraints (SCCs) and IBM Cloud Pak Security Context Constraints (SCCs).

          For more information about the OpenShift Container Platform Security Context Constraints, see [Managing Security Context Constraints](https://docs.openshift.com/container-platform/4.3/authentication/managing-security-context-constraints.html).

          For more information about the IBM Cloud Pak Security Context Constraints, see [Managing Security Context Constraints](https://ibm.biz/cpkspec-scc).

          ### Limitations

          No
        displayName: IBM Cloud Pak foundational services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Common Service Operator
          url: https://github.com/IBM/ibm-common-service-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90046eb6170e12439ccc5aa343f19a4d4fc99a10d98b383cf1dc0e020ebeed55
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:18f44ab15289625fa2c6b72ece6e06afb51058dcba30b8f38b4fdeedf8c9e51b
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:0fdf1beeb4888f648f6dbd17a29dcc53a715d2c4301bc714a4172b2daf8d13be
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:4209e78f12120b9c9f320496b780c3993a9317825ea17dc7d554452b28c99014
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:9a537eae7e581322e2f0764ed38a797ba7cd260724a285f24d5095058b7b47a0
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-operator-bundle@sha256:086eaeaaaf440a1d4a254984afacd8d04583aafb9ec11ac072feb42a87b986f7
        version: 4.12.0
      entries:
      - name: ibm-common-service-operator.v4.12.0
        version: 4.12.0
      name: v4.12
    defaultChannel: v4.12
    packageName: ibm-common-service-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: rebellions
      provider-url: https://rebellions.ai/
    name: rbln-npu-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: candidate
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: fast
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: rbln-npu-operator
    provider:
      name: rebellions
      url: https://rebellions.ai/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/component: operator
      app.kubernetes.io/managed-by: kustomize
      app.kubernetes.io/name: zabbix-operator-certified
      app.kubernetes.io/part-of: zabbix-operator-certified
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zabbix LLC
      provider-url: ""
    name: zabbix-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zabbix-operator-certified-rhmp.v6.0.40
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent",
                "metadata": {
                  "labels": {
                    "app": "agent",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "log_remote_commands": true,
                  "max_lines_per_second": 20,
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent2",
                "metadata": {
                  "labels": {
                    "app": "agent2",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent2"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "enable_persistent_buffer": false,
                  "enable_status_port": false,
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "persistent_buffer_claim": "",
                  "persistent_buffer_period": "1h",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixFull",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-full"
                },
                "spec": {
                  "builtin_report_service": true,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysql_volumeclaim": "mysql-full-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxyMysql",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-mysql"
                },
                "spec": {
                  "db_server_port": 3306,
                  "enable_snmp_traps": false,
                  "internal_db": true,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix_proxy",
                  "proxy": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "zabbix_mysql_volumeclaim": "mysql-proxy-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxySqlite",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-sqlite"
                },
                "spec": {
                  "db_data_volumeclaim": "zabbix-proxy-sqlite",
                  "enable_snmp_traps": false,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "proxy": {
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "node_name_as_db_name": true,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "proxy_size": 1,
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  }
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixServer",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-server"
                },
                "spec": {
                  "builtin_report_service": true,
                  "db_server_host": "mysql-server",
                  "db_server_port": 3306,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix",
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              }
            ]
          ansible.operator-sdk/reconcile-period: 5s
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          certified: "true"
          containerImage: registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
          createdAt: "2025-05-26T07:41:22Z"
          description: Zabbix operator with multiple deployment variants and different
            components
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/zabbix-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/zabbix-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <6.0.40
          operatorframework.io/suggested-namespace: zabbix
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: ""
          support: Zabbix
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Zabbix agent 2 is a new generation of Zabbix agent and may
              be used in place of Zabbix agent
            displayName: Zabbix agent 2
            kind: ZabbixAgent2
            name: zabbixagent2s.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix agent is deployed on a monitoring nodes to actively
              monitor local resources and applications
            displayName: Zabbix agent
            kind: ZabbixAgent
            name: zabbixagents.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database server, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Full
            kind: ZabbixFull
            name: zabbixfulls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with MySQL database server
            displayName: Zabbix proxy (MySQL)
            kind: ZabbixProxyMysql
            name: zabbixproxymysqls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with SQLite database
            displayName: Zabbix proxy (SQLite3)
            kind: ZabbixProxySqlite
            name: zabbixproxysqlites.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database support, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Server
            kind: ZabbixServer
            name: zabbixservers.kubernetes.zabbix.com
            version: v1alpha1
        description: "## About this Operator\n\nZabbix helps you to real-time monitoring
          of millions of metrics collected from tens of thousands of servers, virtual
          machines and network devices.\nThe Zabbix Operator allows users to easily
          deploy, manage, and maintain Zabbix deployments on OpenShift. By installing
          this integration you will be able to deploy Zabbix server / proxies and
          other components with a single command.\n## Supported Features\n* **Zabbix
          Server** - Simple Zabbix installation with included Zabbix server, Zabbix
          web-interface and Zabbix Java Gateway with MySQL database support. The feature
          does not provide MySQL service and requires an external MySQL database.
          \n* **Zabbix Server (Full)** - Zabbix installation with included Zabbix
          server, Zabbix web-interface, Zabbix Java Gateway and MySQL server instance.\n*
          **Zabbix proxy (SQLite3)** - Very simple way to gain power of Zabbix proxy.
          The feature has  SQLite3 support for Zabbix proxies and allow to specify
          amount of proxies. \n* **Zabbix proxy (MySQL)** - Another option of Zabbix
          proxy. The option support and deliver MySQL database.\n* **Zabbix agent**
          - Zabbix agent can be deployed on each available node for stability and
          performance monitoring on remote nodes. It allows to gather metrics with
          full automation!* \n## Prerequisites\nAll deployment options are require
          additional information during deployment. Please, check the following instructions
          and provide required configuration:\n* **Zabbix Server** - MySQL database
          host information and MySQL database credentials in specially formatted *Secret*.
          Additionally it is possible to specify SSL certificates for HTTPS support
          in *Secret*.\n* **Zabbix Server (Full)** - MySQL database credentials in
          specially formatted *Secret*. MySQL database volume name information. Additionally
          it is possible to specify SSL certificates for HTTPS support in *Secret*.\n*
          **Zabbix proxy (SQLite3)** - Zabbix server host information only.\n* **Zabbix
          proxy (MySQL)** - MySQL database credentials in specially formatted *Secrets*
          and Zabbix server host information.\n* **Zabbix agent** - Zabbix server
          host information only for outgoing and incoming connections. Zabbix agent
          uses \"privileged\" mode to monitor node resources! For example, running
          processes.* \n### MySQL credentials\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-full-secrets\ndata:\n  mysql_root_password: emFiYml4X3N1cGVyX3Jvb3Q=
          [1]\n  mysql_zabbix_password: emFiYml4X3VzZXJfcGFzc3dvcmQ= [2]\n  mysql_zabbix_username:
          emFiYml4 [3]\ntype: Opaque\n```\nUsing MySQL root password (*mysqlrootpassword*
          [1]) Zabbix server / proxy will try to create MySQL database schema with
          grant permissions to *mysqlzabbixusername* [2] and *mysqlzabbixpassword*
          [3].\n### SSL certificates for HTTPS\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-web-sslsecret\ndata:\n  ssl.crt: >-\n   < ssl.crt data>\n
          \ ssl.key: >-\n\t < ssl.key data >\n  dhparam.pem: >-\n   <  dhparam.pem
          data >\n```\nFiles *ssl.crt*, *ssl.key* and *dhparam.perm* are required
          for Zabbix web-interface for SSL support.\n"
        displayName: Zabbix Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - zabbix
        - appliance
        - agent
        - metrics
        links:
        - name: Zabbix
          url: https://www.zabbix.com
        - name: Zabbix Official Documentation
          url: https://www.zabbix.com/documentation/6.0/en/manual/quickstart
        - name: Zabbix Operator Installation
          url: https://www.zabbix.com/documentation/6.0/en/manual/installation/containers/openshift
        - name: Downloads
          url: https://www.zabbix.com/download
        maintainers:
        - email: alexey.pustovalov@zabbix.com
          name: Alexey Pustovalov
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Zabbix LLC
        relatedImages:
        - registry.connect.redhat.com/zabbix/zabbix-agent2-60@sha256:424003d52f3cb4360adf2b3619d8b9f8de9196c630ee7e4ceb5d177c1c5a36e4
        - registry.connect.redhat.com/zabbix/zabbix-server-mysql-60@sha256:637484faec6e9a55049c13dbd1f100c03d757c3a21749388c0605040f02d05ce
        - registry.connect.redhat.com/zabbix/zabbix-proxy-mysql-60@sha256:c2ae74d97c083ff348dc9c55932a33da4779cd9efdd9d7c5a79acfffae5b9ed0
        - registry.connect.redhat.com/zabbix/zabbix-java-gateway-60@sha256:a4feb725c12c4ac90eb9244c6fafdb6708679f2e0e7f6aa683c765127306412e
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
        - registry.connect.redhat.com/zabbix/zabbix-web-mysql-60@sha256:611d2670e1bf95b81226976c0a1f64434b2a7766f196dadbd646a7d8319f4cb1
        - registry.connect.redhat.com/zabbix/zabbix-web-service-60@sha256:4a333094e110713470a897b6406c5e340e962caf32a429dddf19865041fcc448
        - registry.redhat.io/rhel9/mysql-80@sha256:c4b5246d0c0d089a766ea07768c22d9d6be8049ac1379bd345db99390ffe2a36
        - registry.connect.redhat.com/zabbix/zabbix-snmptraps-60@sha256:e61e6b1c551dec8488cce0998f94d83d6d10b5d89a465d7119b088313032cb81
        - registry.connect.redhat.com/zabbix/zabbix-proxy-sqlite-60@sha256:62e62ebc8c1bc13fb610f7a83412e65d852f74b3c2c212baa75ea27fc7fa03eb
        - registry.connect.redhat.com/zabbix/zabbix-agent-60@sha256:2895b5e45785aaf528d4a6f78c8af0a0d494dbe9362403fdd470673d911d5004
        - quay.io/zabbix/zabbix-operator@sha256:43d26488a19967bdb2a8aafde5f43820be95011811da9cc5ec623c9029801675
        version: 6.0.40
      entries:
      - name: zabbix-operator-certified-rhmp.v6.0.40
        version: 6.0.40
      - name: zabbix-operator-certified-rhmp.v6.0.38
        version: 6.0.38
      - name: zabbix-operator-certified-rhmp.v6.0.36
        version: 6.0.36
      - name: zabbix-operator-certified-rhmp.v6.0.33
        version: 6.0.33
      - name: zabbix-operator-certified-rhmp.v6.0.29
        version: 6.0.29
      - name: zabbix-operator-certified-rhmp.v6.0.27
        version: 6.0.27
      - name: zabbix-operator-certified-rhmp.v6.0.26
        version: 6.0.26
      - name: zabbix-operator-certified-rhmp.v6.0.25
        version: 6.0.25
      - name: zabbix-operator-certified-rhmp.v6.0.17
        version: 6.0.17
      - name: zabbix-operator-certified-rhmp.v6.0.14
        version: 6.0.14
      - name: zabbix-operator-certified-rhmp.v6.0.13
        version: 6.0.13
      name: lts
    defaultChannel: lts
    packageName: zabbix-operator-certified-rhmp
    provider:
      name: Zabbix LLC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: run-once-duration-override-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runoncedurationoverrideoperator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "RunOnceDurationOverride",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "runOnceDurationOverride": {
                    "spec": {
                      "activeDeadlineSeconds": 3600
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certifiedLevel: "false"
          containerImage: registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9-operator@sha256:54993506a3a2fc1e88c9b5aac97ae90d2a04a06ae6e3a97bf487b5e8996b95fb
          createdAt: 2024/07/01
          description: An operator to manage the OpenShift RunOnceDurationOverride
            Mutating Admission Webhook Server
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.1.2'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/run-once-duration-override-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RunOnceDurationOverride is the schema for the runoncedurationoverrides
              API
            displayName: Run Once Duration Override
            kind: RunOnceDurationOverride
            name: runoncedurationoverrides.operator.openshift.io
            version: v1
        displayName: Run Once Duration Override Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mutating
        - webhook
        - workload
        - run-once
        links:
        - name: Source Code
          url: https://github.com/openshift/run-once-duration-override
        - name: Source Code
          url: https://github.com/openshift/run-once-duration-override-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9-operator@sha256:54993506a3a2fc1e88c9b5aac97ae90d2a04a06ae6e3a97bf487b5e8996b95fb
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9@sha256:3dc6e2a5900472e3e083abcb659a13fea3d08f7a9ff167df524b6f890a16dc8e
        version: 1.1.2
      entries:
      - name: runoncedurationoverrideoperator.v1.1.2
        version: 1.1.2
      - name: runoncedurationoverrideoperator.v1.1.1
        version: 1.1.1
      - name: runoncedurationoverrideoperator.v1.1.0
        version: 1.1.0
      - name: runoncedurationoverrideoperator.v1.0.2
        version: 1.0.2
      - name: runoncedurationoverrideoperator.v1.0.1
        version: 1.0.1
      - name: runoncedurationoverrideoperator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: run-once-duration-override-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicast-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicast-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:ccd64090b85536ddaf1b81ca07d40279c9f277b1c5911628f11d678b98dc3bad
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:ccd64090b85536ddaf1b81ca07d40279c9f277b1c5911628f11d678b98dc3bad
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:f0e9f64c353fdbee3bf7ec0eaf070c3b75330ef659a881ec4ebc874ba584d085
        version: 0.5.3
      entries:
      - name: apicast-operator.v0.5.3
        version: 0.5.3
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      name: threescale-2.11
    - currentCSV: apicast-operator.v0.6.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:8c7fe66b8d4044ad965e7b3d149e9d8e3846ac4e55465873f7095a25b297b3df
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:8c7fe66b8d4044ad965e7b3d149e9d8e3846ac4e55465873f7095a25b297b3df
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:12f7a8e69a1369a51bbcd2b34d4ec8c66dd1fc9fb515349a04e7b202953423e3
        version: 0.6.4
      entries:
      - name: apicast-operator.v0.6.4
        version: 0.6.4
      - name: apicast-operator.v0.6.3
        version: 0.6.3
      - name: apicast-operator.v0.6.2
        version: 0.6.2
      - name: apicast-operator.v0.6.1
        version: 0.6.1
      - name: apicast-operator.v0.6.0
        version: 0.6.0
      - name: apicast-operator.v0.5.3
        version: 0.5.3
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      name: threescale-2.12
    - currentCSV: apicast-operator.v0.7.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:bde44c0fb9f13f212184b24e0bef45db734a8439a92179a2089e5aebd6b87b63
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          olm.skipRange: '>=0.6.2 <0.7.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.13/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:bde44c0fb9f13f212184b24e0bef45db734a8439a92179a2089e5aebd6b87b63
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:88dc7a289c803248efbe151999707325b1bc0aac2d26827ea1707e46e5c71652
        version: 0.7.5
      entries:
      - name: apicast-operator.v0.7.5
        version: 0.7.5
      name: threescale-2.13
    - currentCSV: apicast-operator.v0.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:c5e49be7b43a0058f974a5496976255ab3fbd0677bb7bcf9c3438341cf55965a
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.2 <0.8.3'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.14/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:c5e49be7b43a0058f974a5496976255ab3fbd0677bb7bcf9c3438341cf55965a
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:f8cd25658fa6a154e7ea6e5adf0f9dae3fc596ec9989021de6a021ec94eef226
        version: 0.8.3
      entries:
      - name: apicast-operator.v0.8.3
        version: 0.8.3
      - name: apicast-operator.v0.8.2
        version: 0.8.2
      - name: apicast-operator.v0.8.1
        version: 0.8.1
      name: threescale-2.14
    - currentCSV: apicast-operator.v0.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel9-operator@sha256:c4d6acae4e4cc3236c5b8d8dcfb339b072688ab114785d8598bbe68b1f3571a9
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=0.8.0 <0.12.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.14/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel9-operator@sha256:c4d6acae4e4cc3236c5b8d8dcfb339b072688ab114785d8598bbe68b1f3571a9
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        version: 0.12.5
      entries:
      - name: apicast-operator.v0.12.5
        version: 0.12.5
      - name: apicast-operator.v0.12.4
        version: 0.12.4
      - name: apicast-operator.v0.12.3
        version: 0.12.3
      - name: apicast-operator.v0.12.2
        version: 0.12.2
      - name: apicast-operator.v0.12.1
        version: 0.12.1
      - name: apicast-operator.v0.12.0
        version: 0.12.0
      name: threescale-2.15
    - currentCSV: apicast-operator.v0.12.1-mas
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-mas/apicast-rhel9-operator@sha256:e9f22a8af53f881f1948f5d6a033c35a530895633f2eae522b2fb67f9dc6088b
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.17"}]'
          olm.skipRange: '>=0.7.9 <0.12.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-mas/apicast-gateway-rhel8@sha256:f174636a9ad43a2419fff5a9e6797e3185c540a9509188fd1606aef4acea7bdf
        - registry.redhat.io/3scale-mas/apicast-rhel9-operator@sha256:e9f22a8af53f881f1948f5d6a033c35a530895633f2eae522b2fb67f9dc6088b
        version: 0.12.1-mas
      entries:
      - name: apicast-operator.v0.12.1-mas
        version: 0.12.1-mas
      - name: apicast-operator.v0.7.9-mas
        version: 0.7.9-mas
      - name: apicast-operator.v0.7.8-mas
        version: 0.7.8-mas
      - name: apicast-operator.v0.7.7-mas
        version: 0.7.7-mas
      - name: apicast-operator.v0.7.6-mas
        version: 0.7.6-mas
      - name: apicast-operator.v0.7.5-mas
        version: 0.7.5-mas
      - name: apicast-operator.v0.7.4-mas
        version: 0.7.4-mas
      - name: apicast-operator.v0.7.3-mas
        version: 0.7.3-mas
      - name: apicast-operator.v0.7.2-mas
        version: 0.7.2-mas
      - name: apicast-operator.v0.7.1-mas
        version: 0.7.1-mas
      - name: apicast-operator.v0.7.0-mas
        version: 0.7.0-mas
      - name: apicast-operator.v0.6.0-mas
        version: 0.6.0-mas
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      name: threescale-mas
    defaultChannel: threescale-2.15
    packageName: apicast-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: cloud-bulldozer
      provider-url: ""
    name: ripsaw
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ripsaw.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery, Cloud Provider
          containerImage: quay.io/cloud-bulldozer/benchmark-operator:latest
          description: '[Benchmark-operator](https://github.com/cloud-bulldozer/ripsaw/)
            is a benchmarking tool to benchmark your kubernetes cluster. The intent
            of this Operator is to deploy common workloads to establish a performance
            baseline of Kubernetes cluster'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/cloud-bulldozer/benchmark-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Benchmark
            name: benchmarks.ripsaw.cloudbulldozer.io
            version: v1alpha1
        description: '[Benchmark-operator](https://github.com/cloud-bulldozer/benchmark-operator/)
          is a benchmarking tool to benchmark your kubernetes cluster. The intent
          of this Operator is to deploy common workloads to establish a performance
          baseline of Kubernetes cluster'
        displayName: benchmark-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - performance
        - benchmark
        links:
        - name: Site
          url: https://github.com/cloud-bulldozer/benchmark-operator
        - name: Documentation
          url: https://github.com/cloud-bulldozer/benchmark-operator/tree/master/docs
        maintainers:
        - email: rsevilla@redhat.com
          name: rsevilla87
        - email: jtaleric@redhat.com
          name: jtaleric
        - email: rzaleski@redhat.com
          name: dry939
        - email: kwhitley@redhat.com
          name: whitleykeith
        maturity: alpha
        provider:
          name: cloud-bulldozer
        relatedImages:
        - quay.io/cloud-bulldozer/benchmark-operator:latest
        - k8s.gcr.io/redis:v1
        version: 0.1.0
      entries:
      - name: ripsaw.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ripsaw
    provider:
      name: cloud-bulldozer
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: 3scale-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: 3scale-operator.v0.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "ProductCRName": "product1-sample",
                  "Production": true
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/3scale-rhel7-operator@sha256:0751f9f5343a1264224c36db3db292e5a9e28c40cd96649ef48b5399c34668dd
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.16"}]'
          olm.skipRange: '>=0.10.0 <0.11.13'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/3scale-rhel7-operator@sha256:0751f9f5343a1264224c36db3db292e5a9e28c40cd96649ef48b5399c34668dd
        - registry.redhat.io/3scale-amp2/backend-rhel8@sha256:cdc53799ff9d022462a9a4afcbcd8043497cd40ccdc6fafdcb9ccf7ce69f43d1
        - registry.redhat.io/3scale-amp2/system-rhel7@sha256:61dcc91e5d3c91a0ea85c02f2599cadcb5496b1c09a8838b4e1c8e4efe126e8f
        - registry.redhat.io/3scale-amp2/memcached-rhel7@sha256:7e2ab1e24646fbfef878eb2b5a74ed865ae812b60aa0ba505045f838f26ebbfc
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:a95d09fc3b224f550dec3de3d23fd2dbfc0a220dc869b4ad9559ee2f85327daa
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:f8cd25658fa6a154e7ea6e5adf0f9dae3fc596ec9989021de6a021ec94eef226
        - registry.redhat.io/rhel8/redis-6@sha256:5250d644e80aea794b0f732e9285883a4348dace83557f177f64d78fe4dcabd0
        - registry.redhat.io/openshift4/ose-cli@sha256:1a8caae40be21f23ba3cb48b75809a73333907033a6041927aa568b5c9290f3c
        - registry.redhat.io/3scale-amp2/searchd-rhel7@sha256:8e2001c021fad554cb8a96bb34e9973010fe4206526f19aa040fa196709fb688
        - registry.redhat.io/3scale-amp2/zync-rhel9@sha256:0d2205d9f61a58414c4ba472ce8f19b12d2df722772847d5fb7442e664759627
        - registry.redhat.io/rhel8/mysql-80@sha256:08547a93659ad804d637b91e8e1eb3464b75783d494e6c40e1c921bd74f31645
        version: 0.11.13
      entries:
      - name: 3scale-operator.v0.11.13
        version: 0.11.13
      - name: 3scale-operator.v0.11.12
        version: 0.11.12
      name: threescale-2.14
    - currentCSV: 3scale-operator.v0.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                },
                "status": {
                  "deployments": {
                    "ready": [
                      "deployment x ready"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                },
                "status": {
                  "adminId": 1,
                  "tenantId": 2
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ApplicationAuth",
                "metadata": {
                  "name": "applicationauth-sample"
                },
                "spec": {
                  "applicationCRName": "application-cr-name",
                  "authSecretRef": {
                    "name": "auth-secret-reference"
                  },
                  "generateSecret": true
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "productCRName": "product1-sample",
                  "production": true
                },
                "status": {}
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/3scale-rhel9-operator@sha256:fcea6a77d0a48f35f4c3ce63e72dde104aac6b580f340950b311eb418fff5ec9
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=0.11.0 <0.12.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          rht_backend_redis_requirements: 6.2.0
          rht_mysql_requirements: 8.0.0
          rht_postgres_requirements: 10.0.0
          rht_system_redis_requirements: 6.2.0
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: ApplicationAuth is the Schema for the applicationauths API
            displayName: Application Auth
            kind: ApplicationAuth
            name: applicationauths.capabilities.3scale.net
            version: v1beta1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/3scale-rhel9-operator@sha256:fcea6a77d0a48f35f4c3ce63e72dde104aac6b580f340950b311eb418fff5ec9
        - registry.redhat.io/3scale-amp2/manticore-rhel9@sha256:ccf6baf5f2e9494a34739a79abe0f46e36d8ab9dc81c426f7a3a8964e8f39859
        - registry.redhat.io/3scale-amp2/backend-rhel8@sha256:eb2288de900dc42ead755b28cdb2509b25ecc8dd256781beaa4a2a86e7cbe532
        - registry.redhat.io/rhel9/memcached@sha256:8738110ef16c9e35deeb07c0577ffe09d6ef029761b504a8d20253fa715da394
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:a95d09fc3b224f550dec3de3d23fd2dbfc0a220dc869b4ad9559ee2f85327daa
        - registry.redhat.io/3scale-amp2/zync-rhel9@sha256:df9d62897ff286165772150a8b5e6952d2ce3b52a8afa07bd693a2d95c2c67ff
        - registry.redhat.io/openshift4/ose-cli@sha256:1a8caae40be21f23ba3cb48b75809a73333907033a6041927aa568b5c9290f3c
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/rhel8/mysql-80@sha256:ee9aca9096473da004f9140305742eb7b9a77ed81ce55d5a6503388cccab030f
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        - registry.redhat.io/3scale-amp2/system-rhel8@sha256:4c72dfeb1919a1de4fbc47f4611d0c830ef89cf9f17255ad55a9b370618e0e8f
        version: 0.12.5
      entries:
      - name: 3scale-operator.v0.12.5
        version: 0.12.5
      - name: 3scale-operator.v0.12.4
        version: 0.12.4
      - name: 3scale-operator.v0.12.3
        version: 0.12.3
      - name: 3scale-operator.v0.12.2
        version: 0.12.2
      - name: 3scale-operator.v0.12.1
        version: 0.12.1
      - name: 3scale-operator.v0.12.0
        version: 0.12.0
      name: threescale-2.15
    - currentCSV: 3scale-operator.v0.12.1-mas
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                },
                "status": {
                  "deployments": {
                    "ready": [
                      "deployment x ready"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                },
                "status": {
                  "adminId": 1,
                  "tenantId": 2
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ApplicationAuth",
                "metadata": {
                  "name": "applicationauth-sample"
                },
                "spec": {
                  "applicationCRName": "application-cr-name",
                  "authSecretRef": {
                    "name": "auth-secret-reference"
                  },
                  "generateSecret": true
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "productCRName": "product1-sample",
                  "production": true
                },
                "status": {}
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-mas/3scale-rhel9-operator@sha256:8f41ab6027616e4bb65b1fddbcf44cc8a25a79cdda19d5cc354ac4228a15141e
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.17"}]'
          olm.skipRange: '>=0.11.8 <0.12.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          rht_backend_redis_requirements: 6.2.0
          rht_mysql_requirements: 8.0.0
          rht_postgres_requirements: 10.0.0
          rht_system_redis_requirements: 6.2.0
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: ApplicationAuth is the Schema for the applicationauths API
            displayName: Application Auth
            kind: ApplicationAuth
            name: applicationauths.capabilities.3scale.net
            version: v1beta1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-mas/3scale-manticore-search-rhel9@sha256:7379186175c9c0f0a6d922cd2a0754b429c9728dfbe8979a9d6025789ebae0da
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:a95d09fc3b224f550dec3de3d23fd2dbfc0a220dc869b4ad9559ee2f85327daa
        - registry.redhat.io/3scale-mas/backend-rhel8@sha256:d0c561c6b9dfe9741722ba4a257fbd2df4d2a709457a0b0657ef7bcca1f2a5ca
        - registry.redhat.io/3scale-mas/zync-rhel9@sha256:78e5223a1a878f6527bd88d382948c125b7cb4a81540648244553af426d1a944
        - registry.redhat.io/3scale-mas/memcached-rhel7@sha256:7a1abee11b4120f399485b438bda1e5262f04de28d3e788b207eecc3a9de9ed2
        - registry.redhat.io/3scale-mas/3scale-rhel9-operator@sha256:8f41ab6027616e4bb65b1fddbcf44cc8a25a79cdda19d5cc354ac4228a15141e
        - registry.redhat.io/rhel8/mysql-80@sha256:754f6830e0bc4aadb973b1f613f2bc7278ec13ad0f0e1e87655a2bb2f76cb9b3
        - registry.redhat.io/openshift4/ose-cli@sha256:a688994ea404ce6ce4bb5ec1249f0e779de0d6c187b5078e2ff067e60b50937d
        - registry.redhat.io/3scale-mas/apicast-gateway-rhel8@sha256:f174636a9ad43a2419fff5a9e6797e3185c540a9509188fd1606aef4acea7bdf
        - registry.redhat.io/rhel8/redis-6@sha256:218f7fa2458c274cc34f79190423e0aba03027f55ef6b3334d916951a668b7b7
        - registry.redhat.io/3scale-mas/system-rhel8@sha256:0fa28aaf3467fbd22fc1335dc6a712043e4dead694c3f14b60e89740bb7eb13e
        version: 0.12.1-mas
      entries:
      - name: 3scale-operator.v0.12.1-mas
        version: 0.12.1-mas
      - name: 3scale-operator.v0.11.8-mas
        version: 0.11.8-mas
      - name: 3scale-operator.v0.11.7-mas
        version: 0.11.7-mas
      - name: 3scale-operator.v0.11.6-mas
        version: 0.11.6-mas
      - name: 3scale-operator.v0.11.5-mas
        version: 0.11.5-mas
      - name: 3scale-operator.v0.11.3-mas
        version: 0.11.3-mas
      - name: 3scale-operator.v0.11.2-mas
        version: 0.11.2-mas
      - name: 3scale-operator.v0.11.1-mas
        version: 0.11.1-mas
      - name: 3scale-operator.v0.11.0-mas
        version: 0.11.0-mas
      - name: 3scale-operator.v0.10.0-mas
        version: 0.10.0-mas
      name: threescale-mas
    defaultChannel: threescale-2.15
    packageName: 3scale-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/streamshub
    name: amq-streams-console
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: alpha
    - currentCSV: amq-streams-console.v2.8.0-13-0.1738265629.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:1b1b80710885b43600e22ff2a24930b5cab06cd91fad439ad8a282c00b4564b9
          createdAt: "2024-11-12T15:34:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-streams-console.v2.8.0-13
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.8
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.23.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:1b1b80710885b43600e22ff2a24930b5cab06cd91fad439ad8a282c00b4564b9
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:38b80d81e54d59667c1f7dfa4da4125d5c7430816da130ddc84678a4e3cff6cf
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:366d60ff48fb9ba32dc73d92a775046d7b1eb1df21c93e31605372c7858f16ca
        version: 2.8.0-13+0.1738265629.p
      entries:
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.8.x
    - currentCSV: amq-streams-console.v2.9.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
          createdAt: "2025-09-22T10:41:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0-0 <2.9.2-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.9
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:839368767bec735cd12f002874917c23f58550fe21df976858bdc92946985ec6
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:b7444b70a5dc4543a9976c08a619ff6978c9ff86ac7085d300b4870b7c6b7851
        version: 2.9.2-2
      entries:
      - name: amq-streams-console.v2.9.2-2
        version: 2.9.2-2
      - name: amq-streams-console.v2.9.2-1
        version: 2.9.2-1
      - name: amq-streams-console.v2.9.1-7
        version: 2.9.1-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.9.x
    - currentCSV: amq-streams-console.v2.9.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
          createdAt: "2025-09-22T10:41:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0-0 <2.9.2-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.9
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:839368767bec735cd12f002874917c23f58550fe21df976858bdc92946985ec6
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:b7444b70a5dc4543a9976c08a619ff6978c9ff86ac7085d300b4870b7c6b7851
        version: 2.9.2-2
      entries:
      - name: amq-streams-console.v2.9.2-2
        version: 2.9.2-2
      - name: amq-streams-console.v2.9.2-1
        version: 2.9.2-1
      - name: amq-streams-console.v2.9.1-7
        version: 2.9.1-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.x
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-3.0.x
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-3.x
    defaultChannel: alpha
    packageName: amq-streams-console
    provider:
      name: Red Hat
      url: https://github.com/streamshub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat CCIT
      provider-url: ""
    name: reportportal-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: reportportal-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ReportPortal",
                "metadata": {
                  "name": "reportportal-sample"
                },
                "spec": {
                  "accesskey": "AKIAIOSFODNN7EXAMPLE",
                  "analyzer_image": "quay.io/waynesun09/service-auto-analyzer:5.3.5",
                  "api_image": "quay.io/waynesun09/service-api:5.3.5-rootless",
                  "api_replicas": 1,
                  "app_domain": "apps.test-example.com",
                  "bucket_prefix": null,
                  "default_bucket_name": null,
                  "enable_pg_restic_backup": "yes",
                  "es_backup_schedule": "@daily",
                  "es_s3_backup_dir": "s3_backup",
                  "es_snapshot_bucket": "es-snapshot-123123",
                  "gateway_replicas": 1,
                  "index_image": "quay.io/waynesun09/service-index:5.0.10",
                  "index_replicas": 1,
                  "integration_salt_path": null,
                  "migration_image": "quay.io/waynesun09/migrations:5.3.5",
                  "pg_restic_password": "rp_user",
                  "pg_restic_s3_bucket_init": "yes",
                  "pg_s3_bucket_name": "pgbackup-123123",
                  "s3_enabled": false,
                  "s3_endpoint": null,
                  "secretkey": "wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY",
                  "uat_image": "quay.io/waynesun09/service-authorization:5.3.5",
                  "ui_image": "quay.io/waynesun09/service-ui:5.3.5",
                  "ui_replicas": 1
                }
              },
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ReportPortalRestore",
                "metadata": {
                  "name": "reportportalrestore-sample"
                },
                "spec": {
                  "accesskey": "AKIAIOSFODNN7EXAMPLE",
                  "bucket_prefix": null,
                  "es_restore": "yes",
                  "es_restore_date": "2020-08-07",
                  "es_s3_backup_dir": "s3_backup",
                  "es_snapshot_bucket": "es-snapshot-123123",
                  "pg_recovery": "yes",
                  "pg_recovery_target_time": "2020-08-07 00:00:00",
                  "pg_restic_password": "rp_user",
                  "pg_s3_bucket_name": "pgbackup-123123",
                  "s3_endpoint": null,
                  "secretkey": "wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY"
                }
              },
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ServiceApiServiceMonitor",
                "metadata": {
                  "name": "serviceapiservicemonitor-sample"
                },
                "spec": {
                  "bearer_token_secret_key": "token",
                  "bearer_token_secret_name": "rp-user-auth"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/waynesun09/rp5-operator:v0.0.8
          createdAt: "2021-04-22 14:39:00"
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/waynesun09/reportportal-operator
          support: waynesun09
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ReportPortal Restore from s3 backup
            kind: ReportPortalRestore
            name: reportportalrestores.rp5.reportportal.io
            version: v1
          - description: ReportPortal v5 Deployment with ReportPortal Operator
            kind: ReportPortal
            name: reportportals.rp5.reportportal.io
            version: v1
          - description: ReportPortal service-api ServiceMonitor for metrics
            kind: ServiceApiServiceMonitor
            name: serviceapiservicemonitors.rp5.reportportal.io
            version: v1
        description: |-
          ReportPortal is a test automation analytics platform. ReportPortal Operator support instantiating the ReportPortal services and required dependencies.

          What services are included:

              - Report Portal v5.3
              - PostgreSQL
              - Elasticsearch
              - RabbitMQ
              - Minio

          The Operator also support data backup & restore and metrics monitoring.

          ## Install

          Install the Operator to a namespace, then create ReportPortal v5 instance under Provided APIs on the UI, all Report Portal v5 services and dependencies will be installed.

          ## Data backup and restore

          The Report Portal v5 data backup and restore in the operator include two parts, PostgreSQL database data and Elasticsearch index data.

          The Operator support automatically data backup and restore for both.

          ### PostgreSQL database WAL archiving and recovery

          The Operator supports PostgreSQL WAL archiving and recovery point-in-time via Minio S3 using restic.

          #### WAL archiving

          When deploy Report Portal v5 instance, make sure `enable_pg_restic_backup` option is set as `yes`, then the WAL archiving will be enabled when initiate the database, the WAL archive data will be automatically backup to S3 storage via Minio using restic command.

          Parameter `pg_restic_s3_bucket_init` is for init the bucket if not exist, `pg_s3_bucket_name` is the bucket name and `pg_restic_password` is the restic password for encryption.

          #### WAL recovery

          WAL archive recovery will download all WAL archive files to the Postgres container and restore point-in-time.

          To restore PostgreSQL database point-in-time, on Operator UI select `ReportPortalRestore` under Provided APIs to create ReportPortalRestore operand.

          After the ReportPortalRestore operand created, the PG WAL restore will get started, as restore time will base on the database size, the process will last in minutes, hours or days.

          **Note:** Current Postgres container liveless probe timeout is set as one hour in the postgresql statefulset, it might cause restore fail if restore time is longer than one hour, so it might need be adjusted to be longer.

          ### Elasticsearch snapshot and restore

          The operator supports elasticsearch snapshot and restore with data saved in S3 bucket via Minio, the elasticsearch container have installed the s3 repository plugin.

          #### Snapshot

          When deploy Report Portal v5 instance, make sure following elasticsearch parameters are set:

              es_s3_backup_dir: s3_backup
              es_snapshot_bucket: es-snapshot-123123
              es_backup_schedule: '@daily'

          #### Restore

          Elasticsearch restore support restore to certain date as default snapshot schedule is daily.

          To restore elasticsearch from snapshot, on Operator UI select `ReportPortalRestore` under Provided APIs to create ReportPortalRestore operand.

          The backup dir and bucket name should be same as when create the RP instance, `es_restore` need set to `yes` to enable the restore and `es_restore_date` is the restore target date.

          The restore time will also be depend on the snapshot size.

          ## Metrics

          The operator by default have enabled metrics, the default ansible operator port 8080 is used for the operator container.

          All Report Portal v5 depended services, rabbitmq, minio, postgresql and elasticsearch have metrics enabled in this operator.

          The Report Portal gatway service have metrics enabled.

          By default several ServiceMonitors have been defined when create ReportPortal v5 operand:

              $ oc get servicemonitor
              NAME                    AGE
              elasticsearch-metrics   1d21h
              gateway-metrics         1d21h
              minio-metrics           1d21h
              postgres-metrics        1d21h
              rabbitmq-metrics        1d21h

          To export Report Portal v5 service-api metrics with ServiceMonitor, create a `secret` with RP user api token. The api access token could be found with login to Report Portal -> Click user icon -> Profile.

          Then select `ServiceApiServiceMonitor` under provided APIs to Create Instance on operator UI.

          Input the secret name and key name that have been created, and click create, a api-metrics servicemonitor will be created.

          To check the metrics, if [user workload monitoring](https://docs.openshift.com/container-platform/4.5/monitoring/monitoring-your-own-services.html) is enabled on the cluster, click to Monitoring on OpenShift UI and navigate to Metrics, where you could query the metrics.

          You could also deploy Promethues Operator in the current project and query via it.

          ## Doc

          More detail with installation, data backup & restore and metrics could be found at https://github.com/waynesun09/reportportal-operator#reportportal-operator
        displayName: reportportal-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - reportportal
        - failure analysis
        - test automation
        links:
        - name: Documentation
          url: https://github.com/waynesun09/reportportal-operator/blob/master/README.md
        - name: ReportPortal.io
          url: https://reportportal.io
        maintainers:
        - email: gsun@redhat.com
          name: Wayne Sun
        maturity: alpha
        provider:
          name: Red Hat CCIT
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.7.0
        - quay.io/waynesun09/rp5-operator:v0.0.8
        version: 0.0.8
      entries:
      - name: reportportal-operator.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: reportportal-operator
    provider:
      name: Red Hat CCIT
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://www.redhat.com/
    name: analytics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: analytics-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyData",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalydata-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalydata",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalydata-sample"
                },
                "spec": {
                  "anomalyname": "etcd_object_namespaces",
                  "config": {
                    "max": 100,
                    "min": 60,
                    "query": "max(apiserver_storage_objects{resource=~\"namespaces\"}) by (resource)"
                  },
                  "method": "min_max",
                  "metricdata": {
                    "groupeddata": "{ \"resource\": \"namespaces\" }",
                    "latestvalue": 121,
                    "timestamp": 1695804362
                  }
                }
              },
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalyengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalyengine",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalyengine-sample"
                },
                "spec": {
                  "anomalyqueryconfiguration": "etcd_objects_all:\n  method: percentage_change\n  query: max(apiserver_storage_objects{resource!~\"certificatesigningrequests.certificates.k8s.io|installplans.operators.coreos.com|operators.operators.coreos.com|subscriptions.operators.coreos.com|anomalydata.observability-analytics.redhat.com|anomalyengines.observability-analytics.redhat.com\"}) by (resource)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: True\nfailed_pods:\n  method: min_max\n  query: count(kube_pod_status_phase{phase!~\"Succeeded|Pending|Running|Unknown\"}) by (phase, namespace)\n  min: 0\n  max: 10\n  have_multi_result_data: True\nworkload_memory_usage_bytes:\n  method: percentage_change\n  query: sum(container_memory_working_set_bytes{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nworkload_cpu_usage_cores:\n  method: percentage_change\n  query: sum(container_cpu_usage_seconds_total{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nrest_client_requests_total:\n  method: percentage_change\n  query: sum(rest_client_requests_total)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\n",
                  "cronjobconfig": {
                    "anomalyqueries": "etcd_objects_all,failed_pods,workload_memory_usage_bytes,workload_cpu_usage_cores,rest_client_requests_total",
                    "name": "osa-anomaly-detection",
                    "resource": {
                      "cpulimit": "256m",
                      "cpurequest": "128m",
                      "memorylimit": "512Mi",
                      "memoryrequest": "256Mi"
                    },
                    "schedule": "*/5 * * * *"
                  },
                  "namespace": "osa-anomaly-detection",
                  "serviceaccountrolebinding": {
                    "clusterrolename": "cluster-monitoring-view",
                    "name": "osa-in-cluster-thanos-querier"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
          createdAt: "2024-03-12T08:02:25Z"
          description: Deploys and Manages Observability Analytics Operator
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4-alpha
          repository: https://github.com/openshift/analytics-operator
          support: RedHat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnomalyData is the Schema for the anomalydata API
            displayName: Anomaly Data
            kind: AnomalyData
            name: anomalydata.observability-analytics.redhat.com
            version: v1alpha1
          - description: AnomalyEngine is the Schema for the anomalyengines API
            displayName: Anomaly Engine
            kind: AnomalyEngine
            name: anomalyengines.observability-analytics.redhat.com
            version: v1alpha1
        description: "### Introduction\nObservability Analytics for Red Hat OpenShift
          runs OpenShift analytics processes like anomaly detection.\n\n### API Backward
          Compatibility\n**NOTE:** Until the Operator supports **seamless upgrades**,
          a new version of the operator may introduce a change that is **NOT** backward-compatible
          with the previous version. Thus, to upgrade the operator, uninstall the
          already installed version completely (including CRDs) and install the new
          version.\n\n### Installtion Steps for the Anomaly Engine\n Create namespace
          if you would like to install this operator into separate namespace (You
          can utilize existing namespace as well). ```bash kubectl create ns <Namespace>
          ``` Install the Observability Analytics operator to specific `<Namespace>`,
          wait until you see the operator in the **Installed Operator** list. Click
          on the Operator name and go to Operator details.\n\nHere you should see
          a tab called **Anomaly Engine**. Click on it, and then click on the **Create
          AnomalyEngine** button, which will redirect you to a form where you need
          to provide the required inputs for the Anomaly Engine. Once successfully
          created, you should see a cronjob in the given namespace. This cronjob will
          identify anomalies and store them into **anomalydata**.\n\n### How to See
          Anomaly Data Captured by Anomaly Engine?\nWe are storing anomaly data into
          CR.\n ####Option 1:Through `kubectl`` command\nYou can check detected anomalies
          using the following command from your terminal. \n\n```bash\nkubectl get
          anomalydata -n <Namespace>\nkubectl describe anomalydata <NAME of the anomalydata>
          -n <Namespace> \n```\n #### Option 2: From Installed Operator UI Go to installed
          operator and select **Observability Analytics** operator, now go to **Anomaly
          Data** tab where you can able to see the detected anomaly, click on specific
          anomaly and you should be able to see details of the same. ### Documentation\nDocumentation
          around how to test Operator codebase as a developer can be found [here](https://github.com/openshift/analytics-operator#readme)\n\n###
          Contributing\nYou can contribute by:\n  * Raising [issues](https://github.com/openshift/analytics-operator/issues)\n
          \   related to observability-analytics-operator\n  * Fixing issues by opening
          [Pull Requests](https://github.com/openshift/analytics-operator/pulls)\n
          \ * Discussing observability-analytics-operator\n\n### License\nobservability-analytics-operator
          is licensed under the Apache 2.0 license.\n"
        displayName: Observability Analytics
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - observability-analytics-operator
        - analytics-operator
        - analytics
        - anomaly
        links:
        - name: Analytics Operator
          url: https://analytics-operator.domain
        maintainers:
        - email: rzalavad@redhat.com
          name: Raj
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
        version: 0.1.1
      entries:
      - name: analytics-operator.v0.1.1
        version: 0.1.1
      - name: analytics-operator.v0.1.0
        version: 0.1.0
      name: alpha
    - currentCSV: analytics-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyData",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalydata-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalydata",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalydata-sample"
                },
                "spec": {
                  "anomalyname": "etcd_object_namespaces",
                  "config": {
                    "max": 100,
                    "min": 60,
                    "query": "max(apiserver_storage_objects{resource=~\"namespaces\"}) by (resource)"
                  },
                  "method": "min_max",
                  "metricdata": {
                    "groupeddata": "{ \"resource\": \"namespaces\" }",
                    "latestvalue": 121,
                    "timestamp": 1695804362
                  }
                }
              },
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalyengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalyengine",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalyengine-sample"
                },
                "spec": {
                  "anomalyqueryconfiguration": "etcd_objects_all:\n  method: percentage_change\n  query: max(apiserver_storage_objects{resource!~\"certificatesigningrequests.certificates.k8s.io|installplans.operators.coreos.com|operators.operators.coreos.com|subscriptions.operators.coreos.com|anomalydata.observability-analytics.redhat.com|anomalyengines.observability-analytics.redhat.com\"}) by (resource)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: True\nfailed_pods:\n  method: min_max\n  query: count(kube_pod_status_phase{phase!~\"Succeeded|Pending|Running|Unknown\"}) by (phase, namespace)\n  min: 0\n  max: 10\n  have_multi_result_data: True\nworkload_memory_usage_bytes:\n  method: percentage_change\n  query: sum(container_memory_working_set_bytes{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nworkload_cpu_usage_cores:\n  method: percentage_change\n  query: sum(container_cpu_usage_seconds_total{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nrest_client_requests_total:\n  method: percentage_change\n  query: sum(rest_client_requests_total)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\n",
                  "cronjobconfig": {
                    "anomalyqueries": "etcd_objects_all,failed_pods,workload_memory_usage_bytes,workload_cpu_usage_cores,rest_client_requests_total",
                    "name": "osa-anomaly-detection",
                    "resource": {
                      "cpulimit": "256m",
                      "cpurequest": "128m",
                      "memorylimit": "512Mi",
                      "memoryrequest": "256Mi"
                    },
                    "schedule": "*/5 * * * *"
                  },
                  "namespace": "osa-anomaly-detection",
                  "serviceaccountrolebinding": {
                    "clusterrolename": "cluster-monitoring-view",
                    "name": "osa-in-cluster-thanos-querier"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
          createdAt: "2024-03-12T08:02:25Z"
          description: Deploys and Manages Observability Analytics Operator
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4-alpha
          repository: https://github.com/openshift/analytics-operator
          support: RedHat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnomalyData is the Schema for the anomalydata API
            displayName: Anomaly Data
            kind: AnomalyData
            name: anomalydata.observability-analytics.redhat.com
            version: v1alpha1
          - description: AnomalyEngine is the Schema for the anomalyengines API
            displayName: Anomaly Engine
            kind: AnomalyEngine
            name: anomalyengines.observability-analytics.redhat.com
            version: v1alpha1
        description: "### Introduction\nObservability Analytics for Red Hat OpenShift
          runs OpenShift analytics processes like anomaly detection.\n\n### API Backward
          Compatibility\n**NOTE:** Until the Operator supports **seamless upgrades**,
          a new version of the operator may introduce a change that is **NOT** backward-compatible
          with the previous version. Thus, to upgrade the operator, uninstall the
          already installed version completely (including CRDs) and install the new
          version.\n\n### Installtion Steps for the Anomaly Engine\n Create namespace
          if you would like to install this operator into separate namespace (You
          can utilize existing namespace as well). ```bash kubectl create ns <Namespace>
          ``` Install the Observability Analytics operator to specific `<Namespace>`,
          wait until you see the operator in the **Installed Operator** list. Click
          on the Operator name and go to Operator details.\n\nHere you should see
          a tab called **Anomaly Engine**. Click on it, and then click on the **Create
          AnomalyEngine** button, which will redirect you to a form where you need
          to provide the required inputs for the Anomaly Engine. Once successfully
          created, you should see a cronjob in the given namespace. This cronjob will
          identify anomalies and store them into **anomalydata**.\n\n### How to See
          Anomaly Data Captured by Anomaly Engine?\nWe are storing anomaly data into
          CR.\n ####Option 1:Through `kubectl`` command\nYou can check detected anomalies
          using the following command from your terminal. \n\n```bash\nkubectl get
          anomalydata -n <Namespace>\nkubectl describe anomalydata <NAME of the anomalydata>
          -n <Namespace> \n```\n #### Option 2: From Installed Operator UI Go to installed
          operator and select **Observability Analytics** operator, now go to **Anomaly
          Data** tab where you can able to see the detected anomaly, click on specific
          anomaly and you should be able to see details of the same. ### Documentation\nDocumentation
          around how to test Operator codebase as a developer can be found [here](https://github.com/openshift/analytics-operator#readme)\n\n###
          Contributing\nYou can contribute by:\n  * Raising [issues](https://github.com/openshift/analytics-operator/issues)\n
          \   related to observability-analytics-operator\n  * Fixing issues by opening
          [Pull Requests](https://github.com/openshift/analytics-operator/pulls)\n
          \ * Discussing observability-analytics-operator\n\n### License\nobservability-analytics-operator
          is licensed under the Apache 2.0 license.\n"
        displayName: Observability Analytics
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - observability-analytics-operator
        - analytics-operator
        - analytics
        - anomaly
        links:
        - name: Analytics Operator
          url: https://analytics-operator.domain
        maintainers:
        - email: rzalavad@redhat.com
          name: Raj
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
        version: 0.1.1
      entries:
      - name: analytics-operator.v0.1.1
        version: 0.1.1
      - name: analytics-operator.v0.1.0
        version: 0.1.0
      name: dev-preview
    defaultChannel: alpha
    packageName: analytics-operator
    provider:
      name: Red Hat Inc.
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-atlas-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-atlas-kubernetes.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupCompliancePolicy",
                "metadata": {
                  "name": "my-backup-compliance-policy"
                },
                "spec": {
                  "authorizedEmail": "john.doe@example.com",
                  "authorizedUserFirstName": "John",
                  "authorizedUserLastName": "Doe",
                  "copyProtectionEnabled": false,
                  "encryptionAtRestEnabled": false,
                  "onDemandPolicy": {
                    "retentionUnit": "weeks",
                    "retentionValue": 3
                  },
                  "overwriteBackupPolicies": false,
                  "pointInTimeEnabled": true,
                  "restoreWindowDays": 42,
                  "scheduledPolicyItems": [
                    {
                      "frequencyInterval": 2,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "hourly",
                      "retentionUnit": "days",
                      "retentionValue": 2
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "weekly",
                      "retentionUnit": "weeks",
                      "retentionValue": 4
                    },
                    {
                      "frequencyInterval": 15,
                      "frequencyType": "monthly",
                      "retentionUnit": "months",
                      "retentionValue": 12
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": false,
                  "copySettings": [
                    {
                      "cloudProvider": "AWS",
                      "frequencies": [
                        "WEEKLY"
                      ],
                      "regionName": "US_WEST_1",
                      "shouldCopyOplogs": false
                    }
                  ],
                  "policy": {
                    "name": "atlasbackuppolicy-sample"
                  },
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2,
                  "useOrgAndGroupNamesInExportPrefix": true
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasCustomRole",
                "metadata": {
                  "name": "shard-operator-role"
                },
                "spec": {
                  "projectRef": {
                    "name": "my-project"
                  },
                  "role": {
                    "actions": [
                      {
                        "name": "getShardMap",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "shardingState",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "connPoolStats",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "getLog",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      }
                    ],
                    "inheritedRoles": [
                      {
                        "database": "admin",
                        "name": "operator-role-1"
                      }
                    ],
                    "name": "my-role"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDataFederation",
                "metadata": {
                  "name": "my-data-federation",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "name": "my-data-federation",
                  "privateEndpoints": [
                    {
                      "endpointId": "vpce-03f9eeaa764e32454",
                      "provider": "AWS",
                      "type": "DATA_LAKE"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project",
                    "namespace": "mongodb-atlas-system"
                  },
                  "storage": {
                    "databases": [
                      {
                        "collections": [
                          {
                            "dataSources": [
                              {
                                "storeName": "http-test",
                                "urls": [
                                  "https://data.cityofnewyork.us/api/views/vfnx-vebw/rows.csv"
                                ]
                              }
                            ],
                            "name": "test-collection-1"
                          }
                        ],
                        "name": "test-db-1"
                      }
                    ],
                    "stores": [
                      {
                        "name": "http-test",
                        "provider": "http"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "the-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "theuser"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-deployment"
                },
                "spec": {
                  "backupRef": {
                    "name": "atlasbackupschedule-sample"
                  },
                  "deploymentSpec": {
                    "backupEnabled": true,
                    "clusterType": "REPLICASET",
                    "mongoDBMajorVersion": "7.0",
                    "name": "test-deployment",
                    "replicationSpecs": [
                      {
                        "regionConfigs": [
                          {
                            "electableSpecs": {
                              "instanceSize": "M10",
                              "nodeCount": 3
                            },
                            "priority": 7,
                            "providerName": "AWS",
                            "regionName": "US_EAST_1"
                          }
                        ],
                        "zoneName": "Zone 1"
                      }
                    ],
                    "tags": [
                      {
                        "key": "team",
                        "value": "atlas-operator"
                      }
                    ]
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasFederatedAuth",
                "metadata": {
                  "name": "atlas-default-federated-auth",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "connectionSecretRef": {
                    "name": "my-org-secret",
                    "namespace": "mongodb-atlas-system"
                  },
                  "domainAllowList": [
                    "my-org-domain.com"
                  ],
                  "domainRestrictionEnabled": true,
                  "enabled": true,
                  "postAuthRoleGrants": [
                    "ORG_MEMBER"
                  ],
                  "roleMappings": [
                    {
                      "externalGroupName": "org-admin",
                      "roleAssignments": [
                        {
                          "role": "ORG_OWNER"
                        }
                      ]
                    },
                    {
                      "externalGroupName": "dev-team",
                      "roleAssignments": [
                        {
                          "role": "ORG_GROUP_CREATOR"
                        },
                        {
                          "projectName": "dev-project",
                          "role": "GROUP_OWNER"
                        }
                      ]
                    }
                  ],
                  "ssoDebugEnabled": false
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasIPAccessList",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasipaccesslist-sample"
                },
                "spec": {
                  "entries": [
                    {
                      "cidrBlock": "192.168.1.0/24"
                    },
                    {
                      "deleteAfterDate": "2025-01-20T23:59:59+02:00",
                      "ipAddress": "10.0.0.1"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkContainer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasnetworkcontainer-sample"
                },
                "spec": {
                  "cidrBlock": "10.11.0.0/16",
                  "connectionSecret": {
                    "name": "atlas-connection-secret"
                  },
                  "externalProjectRef": {
                    "projectId": "66e2f2b621571b7e69a89b66"
                  },
                  "provider": "AWS",
                  "region": "US_EAST_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkPeering",
                "metadata": {
                  "name": "atlasnetworkpeering-sample"
                },
                "spec": {
                  "atlasCidrBlock": "10.12.0.0/16",
                  "awsConfiguration": {
                    "accepterRegionName": "US_EAST_1",
                    "awsAccountId": "23214235145",
                    "routeTableCidrBlock": "10.11.0.0/16",
                    "vpcId": "vpcid-7123782134"
                  },
                  "containerId": "623412394512350",
                  "containerRegion": "US_EAST_1",
                  "projectRef": {
                    "name": "atlas-project",
                    "namespace": "namespace"
                  },
                  "provider": "AWS"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasPrivateEndpoint",
                "metadata": {
                  "name": "atlasprivateendpoint-sample"
                },
                "spec": {
                  "awsConfiguration": {
                    "id": "vpce-f4k34w51d"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "provider": "AWS",
                  "region": "EU_CENTRAL_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    },
                    {
                      "cidrBlock": "203.0.113.0/24",
                      "comment": "CIDR block for Application Server B - D"
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasSearchIndexConfig",
                "metadata": {
                  "name": "atlassearchindexconfig-sample"
                },
                "spec": {
                  "analyzer": "lucene.standard",
                  "analyzers": [
                    {
                      "name": "test-analyzer",
                      "tokenFilters": [
                        {
                          "asciiFolding": {
                            "originalTokens": "omit"
                          },
                          "type": "asciiFolding"
                        }
                      ],
                      "tokenizer": {
                        "type": "whiteSpace",
                        "whitespace": {
                          "maxTokenLenght": 200
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamConnection",
                "metadata": {
                  "name": "my-streamconnection-kafka"
                },
                "spec": {
                  "kafkaConfig": {
                    "authentication": {
                      "credentials": {
                        "name": "kafka-userpass"
                      },
                      "mechanism": "SCRAM-512"
                    },
                    "bootstrapServers": "kafka.server1:9001,kafka.server2:9002,kafka.server3:9003",
                    "security": {
                      "certificate": {
                        "name": "kafka-cert"
                      },
                      "protocol": "SSL"
                    }
                  },
                  "name": "kafka-config",
                  "type": "Kafka"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamInstance",
                "metadata": {
                  "name": "my-streaminstance-sample"
                },
                "spec": {
                  "clusterConfig": {
                    "provider": "AWS",
                    "region": "VIRGINIA_USA",
                    "tier": "SP30"
                  },
                  "connectionRegistry": [
                    {
                      "name": "my-streamconnection-sample"
                    }
                  ],
                  "name": "my-instance",
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasTeam",
                "metadata": {
                  "name": "atlasteam-sample"
                },
                "spec": {
                  "name": "my-team",
                  "usernames": [
                    "username@mongodb.com"
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasThirdPartyIntegration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasintegration-sample"
                },
                "spec": {
                  "datadog": {
                    "apiKeySecretRef": {
                      "name": "secret-ref"
                    },
                    "region": "string",
                    "sendCollectionLatencyMetrics": "string",
                    "sendDatabaseMetrics": "string"
                  },
                  "projectRef": {
                    "name": "my-atlas-project"
                  },
                  "type": "DATADOG"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/mongodb/mongodb-atlas-kubernetes-operator:2.11.0
          createdAt: "2025-09-25T14:34:34Z"
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/mongodb/mongodb-atlas-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The AtlasBackupCompliancePolicy is a configuration that enforces
              specific backup and retention requirements
            displayName: Atlas Backup Compliance Policy
            kind: AtlasBackupCompliancePolicy
            name: atlasbackupcompliancepolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasCustomRole is the Schema for the AtlasCustomRole API
            displayName: Atlas Custom Role
            kind: AtlasCustomRole
            name: atlascustomroles.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDataFederation is the Schema for the Atlas Data Federation
              API
            displayName: Atlas Data Federation
            kind: AtlasDataFederation
            name: atlasdatafederations.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasdeployments API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasFederatedAuth is the Schema for the Atlasfederatedauth
              API
            displayName: Atlas Federated Auth
            kind: AtlasFederatedAuth
            name: atlasfederatedauths.atlas.mongodb.com
            version: v1
          - description: AtlasIPAccessList is the Schema for the atlasipaccesslists
              API.
            displayName: Atlas IPAccess List
            kind: AtlasIPAccessList
            name: atlasipaccesslists.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkContainer is the Schema for the AtlasNetworkContainer
              API
            displayName: Atlas Network Container
            kind: AtlasNetworkContainer
            name: atlasnetworkcontainers.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkPeering is the Schema for the AtlasNetworkPeering
              API
            displayName: Atlas Network Peering
            kind: AtlasNetworkPeering
            name: atlasnetworkpeerings.atlas.mongodb.com
            version: v1
          - displayName: Atlas Org Settings
            kind: AtlasOrgSettings
            name: atlasorgsettings.atlas.mongodb.com
            version: v1
          - description: "The AtlasPrivateEndpoint custom resource definition (CRD)
              defines a desired [Private Endpoint](https://www.mongodb.com/docs/atlas/security-private-endpoint/#std-label-private-endpoint-overview)
              configuration for an Atlas project. It allows a private connection between
              your cloud provider and Atlas that doesn't send information through
              a public network. \n You can use private endpoints to create a unidirectional
              connection to Atlas clusters from your virtual network."
            displayName: Atlas Private Endpoint
            kind: AtlasPrivateEndpoint
            name: atlasprivateendpoints.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
          - description: Atlas Search Index Config is the Schema for the Atlas Search
              API
            displayName: Atlas Search Index Config
            kind: AtlasSearchIndexConfig
            name: atlassearchindexconfigs.atlas.mongodb.com
            version: v1
          - description: AtlasStreamConnection is the Schema for the atlasstreamconnections
              API
            displayName: Atlas Stream Connection
            kind: AtlasStreamConnection
            name: atlasstreamconnections.atlas.mongodb.com
            version: v1
          - description: AtlasStreamInstance is the Schema for the atlasstreaminstances
              API
            displayName: Atlas Stream Instance
            kind: AtlasStreamInstance
            name: atlasstreaminstances.atlas.mongodb.com
            version: v1
          - description: AtlasTeam is the Schema for the Atlas Teams API
            displayName: Atlas Team
            kind: AtlasTeam
            name: atlasteams.atlas.mongodb.com
            version: v1
          - description: AtlasThirdPartyIntegration is the Schema for the atlas 3rd
              party integrations API.
            displayName: Atlas Third Party Integration
            kind: AtlasThirdPartyIntegration
            name: atlasthirdpartyintegrations.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB
          Atlas — the only multi-cloud document database service that gives you the versatility you need to build sophisticated
          and resilient applications that can adapt to changing customer demands and market trends.

          The full documentation for the Operator can be found [here](https://docs.atlas.mongodb.com/atlas-operator/)

          ## Getting Started

          ### Supported features

          * Create and configure a project, or connect to an existing one.
          * Deploy, manage, scale, and tear down clusters.
          * Support for serverless instances.
          * Create and edit database users.
          * Manage IP Access Lists, network peering and private endpoints.
          * Configure and control Atlas’s fully managed cloud backup.
          * Configure federated authentication for your Atlas organization
          * Integrate Atlas monitoring with Prometheus.

          ... and more.

          To view the list of custom resources and each CR's schema, visit our [reference](https://www.mongodb.com/docs/atlas/operator/stable/custom-resources/)
          documentation. See the [Quickstart](https://www.mongodb.com/docs/atlas/operator/stable/ak8so-quick-start/) to get
          started with Atlas Kubernetes Operator.
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator:2.11.0
        version: 2.11.0
      entries:
      - name: mongodb-atlas-kubernetes.v2.11.0
        version: 2.11.0
      - name: mongodb-atlas-kubernetes.v2.10.0
        version: 2.10.0
      - name: mongodb-atlas-kubernetes.v2.9.1
        version: 2.9.1
      - name: mongodb-atlas-kubernetes.v2.9.0
        version: 2.9.0
      - name: mongodb-atlas-kubernetes.v2.8.2
        version: 2.8.2
      - name: mongodb-atlas-kubernetes.v2.8.1
        version: 2.8.1
      - name: mongodb-atlas-kubernetes.v2.8.0
        version: 2.8.0
      - name: mongodb-atlas-kubernetes.v2.7.1
        version: 2.7.1
      - name: mongodb-atlas-kubernetes.v2.7.0
        version: 2.7.0
      - name: mongodb-atlas-kubernetes.v2.6.1
        version: 2.6.1
      - name: mongodb-atlas-kubernetes.v2.6.0
        version: 2.6.0
      - name: mongodb-atlas-kubernetes.v2.5.0
        version: 2.5.0
      - name: mongodb-atlas-kubernetes.v2.4.1
        version: 2.4.1
      - name: mongodb-atlas-kubernetes.v2.4.0
        version: 2.4.0
      - name: mongodb-atlas-kubernetes.v2.3.1
        version: 2.3.1
      - name: mongodb-atlas-kubernetes.v2.3.0
        version: 2.3.0
      - name: mongodb-atlas-kubernetes.v2.2.2
        version: 2.2.2
      - name: mongodb-atlas-kubernetes.v2.2.1
        version: 2.2.1
      - name: mongodb-atlas-kubernetes.v2.2.0
        version: 2.2.0
      - name: mongodb-atlas-kubernetes.v2.1.0
        version: 2.1.0
      - name: mongodb-atlas-kubernetes.v2.0.1
        version: 2.0.1
      - name: mongodb-atlas-kubernetes.v2.0.0
        version: 2.0.0
      - name: mongodb-atlas-kubernetes.v1.9.1
        version: 1.9.1
      - name: mongodb-atlas-kubernetes.v1.9.0
        version: 1.9.0
      - name: mongodb-atlas-kubernetes.v1.8.2
        version: 1.8.2
      - name: mongodb-atlas-kubernetes.v1.8.1
        version: 1.8.1
      - name: mongodb-atlas-kubernetes.v1.8.0
        version: 1.8.0
      - name: mongodb-atlas-kubernetes.v1.7.3
        version: 1.7.3
      - name: mongodb-atlas-kubernetes.v1.7.2
        version: 1.7.2
      - name: mongodb-atlas-kubernetes.v1.7.1
        version: 1.7.1
      - name: mongodb-atlas-kubernetes.v1.7.0
        version: 1.7.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-atlas-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OpenTelemetry Community
      provider-url: ""
    name: opentelemetry-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opentelemetry-operator.v0.136.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "Instrumentation",
                "metadata": {
                  "name": "instrumentation"
                },
                "spec": {
                  "exporter": {
                    "endpoint": "http://otel-collector-headless:4317"
                  }
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpAMPBridge",
                "metadata": {
                  "name": "opampbridge-sample"
                },
                "spec": {
                  "capabilities": {
                    "AcceptsOpAMPConnectionSettings": true,
                    "AcceptsOtherConnectionSettings": true,
                    "AcceptsRemoteConfig": true,
                    "AcceptsRestartCommand": true,
                    "ReportsEffectiveConfig": true,
                    "ReportsHealth": true,
                    "ReportsOwnLogs": true,
                    "ReportsOwnMetrics": true,
                    "ReportsOwnTraces": true,
                    "ReportsRemoteConfig": true,
                    "ReportsStatus": true
                  },
                  "componentsAllowed": {
                    "exporters": [
                      "debug"
                    ],
                    "processors": [
                      "memory_limiter"
                    ],
                    "receivers": [
                      "otlp"
                    ]
                  },
                  "endpoint": "ws://opamp-server:4320/v1/opamp"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": "receivers:\n  otlp:\n    protocols: \n      grpc: {}\n      http: {}\n\nexporters:\n  debug: {}\n\nservice:\n  pipelines:\n    traces:\n      receivers: [otlp]\n      exporters: [debug]\n"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1beta1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": {
                    "exporters": {
                      "debug": {}
                    },
                    "receivers": {
                      "otlp": {
                        "protocols": {
                          "grpc": {},
                          "http": {}
                        }
                      }
                    },
                    "service": {
                      "pipelines": {
                        "traces": {
                          "exporters": [
                            "debug"
                          ],
                          "receivers": [
                            "otlp"
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          certified: "false"
          containerImage: ghcr.io/open-telemetry/opentelemetry-operator/opentelemetry-operator
          createdAt: "2025-09-29T13:00:19Z"
          description: Provides the OpenTelemetry components, including the Collector
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/open-telemetry/opentelemetry-operator
          support: OpenTelemetry Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instrumentation is the spec for OpenTelemetry instrumentation.
            displayName: OpenTelemetry Instrumentation
            kind: Instrumentation
            name: instrumentations.opentelemetry.io
            version: v1alpha1
          - description: OpAMPBridge is the Schema for the opampbridges API.
            displayName: OpAMP Bridge
            kind: OpAMPBridge
            name: opampbridges.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1beta1
          - description: TargetAllocator is the Schema for the targetallocators API.
            displayName: Target Allocator
            kind: TargetAllocator
            name: targetallocators.opentelemetry.io
            version: v1alpha1
        description: |-
          OpenTelemetry is a collection of tools, APIs, and SDKs. You use it to instrument, generate, collect, and export telemetry data (metrics, logs, and traces) for analysis in order to understand your software's performance and behavior.

          ### Operator features

          * **Sidecar injection** - annotate your pods and let the operator inject a sidecar.
          * **Managed upgrades** - updating the operator will automatically update your OpenTelemetry collectors.
          * **Deployment modes** - your collector can be deployed as sidecar, daemon set, or regular deployment.
          * **Service port management** - the operator detects which ports need to be exposed based on the provided configuration.
        displayName: Community OpenTelemetry Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opentelemetry
        - tracing
        - logging
        - metrics
        - monitoring
        - troubleshooting
        links:
        - name: OpenTelemetry Operator
          url: https://github.com/open-telemetry/opentelemetry-operator
        maintainers:
        - email: jpkroehling@redhat.com
          name: Juraci Paixão Kröhling
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: OpenTelemetry Community
        relatedImages:
        - ghcr.io/open-telemetry/opentelemetry-operator/opentelemetry-operator:0.136.0
        - quay.io/brancz/kube-rbac-proxy:v0.13.1
        version: 0.136.0
      entries:
      - name: opentelemetry-operator.v0.136.0
        version: 0.136.0
      - name: opentelemetry-operator.v0.135.0
        version: 0.135.0
      - name: opentelemetry-operator.v0.134.0
        version: 0.134.0
      - name: opentelemetry-operator.v0.132.0
        version: 0.132.0
      - name: opentelemetry-operator.v0.131.0
        version: 0.131.0
      - name: opentelemetry-operator.v0.129.1
        version: 0.129.1
      - name: opentelemetry-operator.v0.127.0
        version: 0.127.0
      - name: opentelemetry-operator.v0.126.0
        version: 0.126.0
      - name: opentelemetry-operator.v0.125.0
        version: 0.125.0
      - name: opentelemetry-operator.v0.124.0
        version: 0.124.0
      - name: opentelemetry-operator.v0.123.0
        version: 0.123.0
      - name: opentelemetry-operator.v0.122.0
        version: 0.122.0
      - name: opentelemetry-operator.v0.121.0
        version: 0.121.0
      - name: opentelemetry-operator.v0.120.0
        version: 0.120.0
      - name: opentelemetry-operator.v0.119.0
        version: 0.119.0
      - name: opentelemetry-operator.v0.118.0
        version: 0.118.0
      - name: opentelemetry-operator.v0.117.0
        version: 0.117.0
      - name: opentelemetry-operator.v0.116.0
        version: 0.116.0
      - name: opentelemetry-operator.v0.115.0
        version: 0.115.0
      - name: opentelemetry-operator.v0.114.1
        version: 0.114.1
      - name: opentelemetry-operator.v0.112.0
        version: 0.112.0
      - name: opentelemetry-operator.v0.111.0
        version: 0.111.0
      - name: opentelemetry-operator.v0.110.0
        version: 0.110.0
      - name: opentelemetry-operator.v0.109.0
        version: 0.109.0
      - name: opentelemetry-operator.v0.108.0
        version: 0.108.0
      - name: opentelemetry-operator.v0.107.0
        version: 0.107.0
      - name: opentelemetry-operator.v0.106.0
        version: 0.106.0
      - name: opentelemetry-operator.v0.105.0
        version: 0.105.0
      name: alpha
    defaultChannel: alpha
    packageName: opentelemetry-operator
    provider:
      name: OpenTelemetry Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: StreamNative
      provider-url: https://streamnative.io
    name: sn-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc:latest",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway:0.10.0"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator:v0.10.0
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator:v0.10.0
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      - name: sn-operator.v0.9.0
        version: 0.9.0
      - name: sn-operator.v0.8.5
        version: 0.8.5
      - name: sn-operator.v0.7.10
        version: 0.7.10
      name: alpha
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc:latest",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway:0.10.0"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator:v0.10.0
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator:v0.10.0
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      - name: sn-operator.v0.9.0
        version: 0.9.0
      - name: sn-operator.v0.8.5
        version: 0.8.5
      - name: sn-operator.v0.7.10
        version: 0.7.10
      name: beta
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc:latest",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway:0.10.0"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator:v0.10.0
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator:v0.10.0
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      - name: sn-operator.v0.9.0
        version: 0.9.0
      - name: sn-operator.v0.8.5
        version: 0.8.5
      - name: sn-operator.v0.7.10
        version: 0.7.10
      name: stable
    defaultChannel: alpha
    packageName: sn-operator
    provider:
      name: StreamNative
      url: https://streamnative.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: operator-certification-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operator-certification-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "certification.redhat.com/v1alpha1",
                "kind": "OperatorPipeline",
                "metadata": {
                  "name": "operatorpipeline-sample"
                },
                "spec": {
                  "applyCIPipeline": true,
                  "applyHostedPipeline": false,
                  "applyReleasePipeline": false,
                  "gitHubSecretName": "github-api-token",
                  "kubeconfigSecretName": "kubeconfig",
                  "operatorPipelinesRelease": "main",
                  "pyxisSecretName": "pyxis-api-secret"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-06-23T16:02:00Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io.bundle.channel.default.v1: alpha
          operators.operatorframework.io.bundle.channels: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-openshift-ecosystem/operator-certification-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OperatorPipeline is the Schema for the operatorpipelines
              API
            displayName: Operator Pipeline
            kind: OperatorPipeline
            name: operatorpipelines.certification.redhat.com
            version: v1alpha1
        description: |-
          A Kubernetes operator to provision resources for the operator certification pipeline. This operator is installed in all namespaces which can support multi-tenant scenarios. **Note:** This operator should only be used by Red Hat partners attempting to certify their operator(s).


          **Requirements**


          The certification operator requires that you have the following tools installed, functional, and in your path.
          - [Install](https://docs.openshift.com/container-platform/4.8/cli_reference/openshift_cli/getting-started-cli.html#installing-openshift-cli) oc, the OpenShift CLI tool (tested with version 4.7.13)
          - [Install](https://tekton.dev/docs/cli/) tkn, the Tekton CLI tool (tested with version 0.19.1)
          - [Install](https://git-scm.com/downloads) git, the Git CLI tool (tested with 2.32.0)
          - The certification pipeline expects you to have the source files of your Operator bundle


          **Pre - Installation**


          The below steps exist for simplicity and to make the installation clearer. The operator watches all namespaces, so if secrets/configs/etc already exist in another namespace, feel free to use the existing namespace when following the operator installation steps.


          **Create a new namespace where the following secrets will be applied.**
          ```
          oc new-project oco
          ```


          **Add the kubeconfig secret which will be used to deploy the operator under test and run the certification checks.**
          * Open a terminal window
          * Set the `KUBECONFIG` environment variable
          ```
          export KUBECONFIG=/path/to/your/cluster/kubeconfig
          ```
          > *This kubeconfig will be used to deploy the Operator under test and run the certification checks.*
          ```
          oc create secret generic kubeconfig --from-file=kubeconfig=$KUBECONFIG
          ```


          **Configuring steps for submitting the results**
          - Add the github API token to the repo where the PR will be created
          ```
          oc create secret generic github-api-token --from-literal GITHUB_TOKEN=<github token>
          ```
          - Add RedHat Container API access key

            This API access key is specifically related to your unique partner account for Red Hat Connect portal. Instructions to obtain your API key can be found: [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/operator-cert-workflow.md#step-b---get-api-key)
          ```
          oc create secret generic pyxis-api-secret --from-literal pyxis_api_key=< API KEY >
          ```

          - Optional pipeline configurations can be found [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md#optional-configuration)


          **Execute the Pipeline (Development Iterations)**


          A pre-requisite to running a pipeline is that a `workspace-template.yaml` exists in the directory you want to execute the `tkn` commands from.

          To create a workspace-template.yaml
          ```
          cat <<EOF > workspace-template.yaml
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 5Gi
          EOF
          ```

          There are multiple ways to execute the Pipeline which can be found [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md#execute-the-pipeline-development-iterations)
        displayName: Operator Certification Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - certification
        - pipeline
        - openshift
        links:
        - name: Operator Certification CI Pipeline
          url: https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md
        - name: Red Hat OpenShift Pipelines
          url: https://github.com/openshift/tektoncd-operator
        maintainers:
        - email: certification-operator-devel@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/opdev/operator-certification-operator@sha256:7c84772f96650ff757c2552ba421355d54da41a5ed2155f72a14911b8a3f18ff
        version: 1.3.1
      entries:
      - name: operator-certification-operator.v1.3.1
        version: 1.3.1
      name: alpha
    defaultChannel: alpha
    packageName: operator-certification-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: LINBIT
      provider-url: https://linbit.com
    name: linstor-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: linstor-operator.v2.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorCluster",
                "metadata": {
                  "name": "linstorcluster"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorNodeConnection",
                "metadata": {
                  "name": "example-connection"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatellite",
                "metadata": {
                  "name": "example-satellite"
                },
                "spec": {
                  "clusterRef": {
                    "name": "linstorcluster"
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatelliteConfiguration",
                "metadata": {
                  "name": "all-satellites"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          com.redhat.openshift.versions: v4.7
          containerImage: drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
          createdAt: "2025-09-25T13:44:22Z"
          description: LINSTOR Operator deploys LINBIT SDS in the cluster.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: linbit-sds
          operators.openshift.io/valid-subscription: '[''LINBIT SDS Subscription'']'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/internal-objects: '["linstorsatellites.piraeus.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster-wide configuration for LINSTOR.
            displayName: Linstor Cluster
            kind: LinstorCluster
            name: linstorclusters.piraeus.io
            version: v1
          - description: Configuration for connections between specific nodes.
            displayName: Linstor Node Connection
            kind: LinstorNodeConnection
            name: linstornodeconnections.piraeus.io
            version: v1
          - description: Configuration for specific Satellites/Nodes in the LINSTOR
              Cluster.
            displayName: Linstor Satellite Configuration
            kind: LinstorSatelliteConfiguration
            name: linstorsatelliteconfigurations.piraeus.io
            version: v1
          - description: Resource to configure a single satellite, managed by the
              Operator itself.
            displayName: Linstor Satellite
            kind: LinstorSatellite
            name: linstorsatellites.piraeus.io
            version: v1
          required:
          - kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
        description: |
          The LINSTOR Operator manages [LINSTOR](https://github.com/LINBIT/linstor-server) clusters in Kubernetes.

          All components of the LINSTOR software stack can be managed by the operator:
          * DRBD
          * LINSTOR
          * LINSTOR CSI driver
          * LINSTOR HA Controller
        displayName: LINSTOR Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: User Guide
          url: https://linbit.com/drbd-user-guide/linstor-guide-1_0-en/
        - name: LINSTOR Operator
          url: https://charts.linstor.io/
        - name: Source Code
          url: https://github.com/LINBIT/linstor-operator-builder
        maintainers:
        - email: support@linbit.com
          name: LINBIT
        maturity: stable
        minKubeVersion: 1.20.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: DaemonSet
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: storage.k8s.io
          kind: CSIDriver
          version: v1
        provider:
          name: LINBIT
          url: https://linbit.com
        relatedImages:
        - drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:f8cde5b5c289f7b3d233773f4e79990ba392fe6e14e0177de80714e7d3e98862
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ac3963bfd4246da386b82d810a348e9d42a9e4b746acaf4dbcbe0766643cc2fa
        - drbd.io/drbd-shutdown-guard@sha256:7ba049ecb205a0086e425d2ff37285b00449549d89e179b4b943ace6e59a9156
        - drbd.io/drbd9-rhel8@sha256:7cddfa770de66cefcb42c2b33da57d0f1adfcc5f1cc12522977aa6993722bb2a
        - drbd.io/linstor-k8s-ha-controller@sha256:ed0f0f86dad88feadf8c3c32d7d143eaa52b1130af3576315a562909b39d2afa
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:2669cde3db2268d37bc671f494eaa06480478a90b8943054af14cbffc050debf
        - drbd.io/drbd-reactor@sha256:77c51aeeefdcd788cb1583738f93778032fcc03c36fb289b692d3cf195c8b184
        - drbd.io/linstor-csi@sha256:9c7ef452b0997d3373803201614c3b8075687e75c397359738a9e4ef5530cab3
        - drbd.io/linstor-controller@sha256:841b3aa84cdf13aaaf1243bc5da8ac2205ca3fe0186fdfc7a04b69397f0020d6
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:624f61b1ae9d32f5a074ee8c12839d975fba9efcd3bfa7981cfe38febf5c4495
        - drbd.io/drbd9-rhel9@sha256:4eb553336ed64e127068a4566370b040166bf11301e21ac47ac5b92237ec9d2a
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:5a650dc21a6a038de6267f117d01adfd904ff3f5c94f428f6a13735f18bf94d6
        - drbd.io/drbd9-rhel7@sha256:0c96b216e55a128a9944276af9080aced8bd72fdf1fd424f024ed9ace2c1feb5
        - drbd.io/ktls-utils@sha256:3351a6e0fc048fe1544499095cdb8fdd221e832cecfa8a9a01f046c271991138
        - drbd.io/linstor-satellite@sha256:ce0a0b05494c6e6d5f16121e5762bf58e7b78f40c00b84780e2abba2a690af1b
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:700be620923fc99ed03f9f1c4a911e4084beb30685361270c49165b27ad1d374
        - drbd.io/drbd9-rhel10@sha256:da0a7015cf15724a7c73100d505891e62262a726141347df7f2d51ec79a0a1e1
        - drbd.io/linstor-affinity-controller@sha256:78e5b299ef1702114528678387b5e74821d7658bc02f05f0bd7edd9a176fba39
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:cd58f1789027e32eae08f00dac7b2275496c41020fb7cdebc10e9986c625d4f4
        version: 2.9.1
      entries:
      - name: linstor-operator.v2.9.1
        version: 2.9.1
      name: dev
    - currentCSV: linstor-operator.v2.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorCluster",
                "metadata": {
                  "name": "linstorcluster"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorNodeConnection",
                "metadata": {
                  "name": "example-connection"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatellite",
                "metadata": {
                  "name": "example-satellite"
                },
                "spec": {
                  "clusterRef": {
                    "name": "linstorcluster"
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatelliteConfiguration",
                "metadata": {
                  "name": "all-satellites"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          com.redhat.openshift.versions: v4.7
          containerImage: drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
          createdAt: "2025-09-25T13:44:22Z"
          description: LINSTOR Operator deploys LINBIT SDS in the cluster.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: linbit-sds
          operators.openshift.io/valid-subscription: '[''LINBIT SDS Subscription'']'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/internal-objects: '["linstorsatellites.piraeus.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster-wide configuration for LINSTOR.
            displayName: Linstor Cluster
            kind: LinstorCluster
            name: linstorclusters.piraeus.io
            version: v1
          - description: Configuration for connections between specific nodes.
            displayName: Linstor Node Connection
            kind: LinstorNodeConnection
            name: linstornodeconnections.piraeus.io
            version: v1
          - description: Configuration for specific Satellites/Nodes in the LINSTOR
              Cluster.
            displayName: Linstor Satellite Configuration
            kind: LinstorSatelliteConfiguration
            name: linstorsatelliteconfigurations.piraeus.io
            version: v1
          - description: Resource to configure a single satellite, managed by the
              Operator itself.
            displayName: Linstor Satellite
            kind: LinstorSatellite
            name: linstorsatellites.piraeus.io
            version: v1
          required:
          - kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
        description: |
          The LINSTOR Operator manages [LINSTOR](https://github.com/LINBIT/linstor-server) clusters in Kubernetes.

          All components of the LINSTOR software stack can be managed by the operator:
          * DRBD
          * LINSTOR
          * LINSTOR CSI driver
          * LINSTOR HA Controller
        displayName: LINSTOR Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: User Guide
          url: https://linbit.com/drbd-user-guide/linstor-guide-1_0-en/
        - name: LINSTOR Operator
          url: https://charts.linstor.io/
        - name: Source Code
          url: https://github.com/LINBIT/linstor-operator-builder
        maintainers:
        - email: support@linbit.com
          name: LINBIT
        maturity: stable
        minKubeVersion: 1.20.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: DaemonSet
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: storage.k8s.io
          kind: CSIDriver
          version: v1
        provider:
          name: LINBIT
          url: https://linbit.com
        relatedImages:
        - drbd.io/drbd9-rhel7@sha256:0c96b216e55a128a9944276af9080aced8bd72fdf1fd424f024ed9ace2c1feb5
        - drbd.io/linstor-controller@sha256:841b3aa84cdf13aaaf1243bc5da8ac2205ca3fe0186fdfc7a04b69397f0020d6
        - drbd.io/drbd9-rhel9@sha256:4eb553336ed64e127068a4566370b040166bf11301e21ac47ac5b92237ec9d2a
        - drbd.io/linstor-k8s-ha-controller@sha256:ed0f0f86dad88feadf8c3c32d7d143eaa52b1130af3576315a562909b39d2afa
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:f8cde5b5c289f7b3d233773f4e79990ba392fe6e14e0177de80714e7d3e98862
        - drbd.io/drbd-shutdown-guard@sha256:7ba049ecb205a0086e425d2ff37285b00449549d89e179b4b943ace6e59a9156
        - drbd.io/linstor-affinity-controller@sha256:78e5b299ef1702114528678387b5e74821d7658bc02f05f0bd7edd9a176fba39
        - drbd.io/linstor-satellite@sha256:ce0a0b05494c6e6d5f16121e5762bf58e7b78f40c00b84780e2abba2a690af1b
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:5a650dc21a6a038de6267f117d01adfd904ff3f5c94f428f6a13735f18bf94d6
        - drbd.io/linstor-csi@sha256:9c7ef452b0997d3373803201614c3b8075687e75c397359738a9e4ef5530cab3
        - drbd.io/ktls-utils@sha256:3351a6e0fc048fe1544499095cdb8fdd221e832cecfa8a9a01f046c271991138
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ac3963bfd4246da386b82d810a348e9d42a9e4b746acaf4dbcbe0766643cc2fa
        - drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:cd58f1789027e32eae08f00dac7b2275496c41020fb7cdebc10e9986c625d4f4
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:624f61b1ae9d32f5a074ee8c12839d975fba9efcd3bfa7981cfe38febf5c4495
        - drbd.io/drbd9-rhel8@sha256:7cddfa770de66cefcb42c2b33da57d0f1adfcc5f1cc12522977aa6993722bb2a
        - drbd.io/drbd-reactor@sha256:77c51aeeefdcd788cb1583738f93778032fcc03c36fb289b692d3cf195c8b184
        - drbd.io/drbd9-rhel10@sha256:da0a7015cf15724a7c73100d505891e62262a726141347df7f2d51ec79a0a1e1
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:700be620923fc99ed03f9f1c4a911e4084beb30685361270c49165b27ad1d374
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:2669cde3db2268d37bc671f494eaa06480478a90b8943054af14cbffc050debf
        version: 2.9.1
      entries:
      - name: linstor-operator.v2.9.1
        version: 2.9.1
      name: stable
    defaultChannel: dev
    packageName: linstor-operator
    provider:
      name: LINBIT
      url: https://linbit.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: https://opdev.github.io/
    name: loki-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.example.com/v1alpha1",
                "kind": "Loki",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/install-disable-crds": "true"
                  },
                  "name": "loki-sample"
                },
                "spec": {
                  "global": {
                    "clusterDomain": "cluster.local",
                    "dnsNamespace": "openshift-dns",
                    "dnsService": "dns-default"
                  },
                  "loki": {
                    "auth_enabled": false,
                    "commonConfig": {
                      "replication_factor": 1
                    },
                    "storage": {
                      "type": "filesystem"
                    }
                  },
                  "monitoring": {
                    "dashboards": {
                      "enabled": false
                    },
                    "lokiCanary": {
                      "enabled": false
                    },
                    "rules": {
                      "alerting": false,
                      "enabled": false
                    },
                    "selfMonitoring": {
                      "enabled": false,
                      "grafanaAgent": {
                        "installOperator": false
                      }
                    },
                    "serviceMonitor": {
                      "enabled": false,
                      "metricsInstance": {
                        "enabled": false
                      }
                    }
                  },
                  "rbac": {
                    "sccEnabled": true
                  },
                  "singleBinary": {
                    "replicas": 1
                  },
                  "test": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/loki-helm-operator-controller:0.0.1
          createdAt: "2023-07-14T15:06:17Z"
          description: An experimental operator that installs loki from its official
            helmchart.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys the loki operator helm chart
            displayName: Loki
            kind: Loki
            name: lokis.charts.example.com
            version: v1alpha1
        description: Loki operator based on the helm chart with single binary mode
          deployment optionality
        displayName: Loki Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loki
        - single-binary
        - demo
        - helm
        links:
        - name: Loki Helm Operator
          url: https://github.com/skattoju/loki-helm-operator
        maintainers:
        - email: skattoju@redhat.com
          name: Sid Kattoju
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: https://opdev.github.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/opdev/loki-helm-operator-controller:0.0.1
        version: 0.0.1
      entries:
      - name: loki-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: loki-helm-operator
    provider:
      name: opdev
      url: https://opdev.github.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hdfs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hdfs-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:e8b4ad2c3b127537239b003ae19a3f58b541d599fe3d5a74a6718612fe645d50
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:e8b4ad2c3b127537239b003ae19a3f58b541d599fe3d5a74a6718612fe645d50
        version: 24.11.1
      entries:
      - name: hdfs-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: hdfs-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:00c79036eed6dd6749736b964787acb638f1a6e0156c0bba5651b93738534377
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:00c79036eed6dd6749736b964787acb638f1a6e0156c0bba5651b93738534377
        version: 25.3.0
      entries:
      - name: hdfs-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: hdfs-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
        version: 25.7.0
      entries:
      - name: hdfs-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hdfs-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
        version: 25.7.0
      entries:
      - name: hdfs-operator.v25.7.0
        version: 25.7.0
      - name: hdfs-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hdfs-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: seldon
      provider-url: ""
    name: seldon-deploy-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: seldon-deploy-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1alpha1",
                "kind": "SeldonDeploy",
                "metadata": {
                  "name": "seldondeploy-sample"
                },
                "spec": {
                  "affinity": {},
                  "alibidetect": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306"
                  },
                  "batchjobs": {
                    "mc": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48"
                    },
                    "processor": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d"
                    },
                    "pvc": {
                      "defaultSize": "1Gi"
                    },
                    "serviceAccount": "workflow"
                  },
                  "defaultUserID": "",
                  "docker": {
                    "user": "unknown"
                  },
                  "elasticsearch": {
                    "basicAuth": true,
                    "secret": {
                      "name": "elastic-credentials",
                      "passwordKey": "password",
                      "userKey": "username"
                    },
                    "url": "https://elasticsearch-seldon-es-http.seldon-logs:9200"
                  },
                  "enableAppAnalytics": false,
                  "enableAppAuth": false,
                  "env": {
                    "USERID_CLAIM_KEY": "name"
                  },
                  "external": {
                    "protocol": "http"
                  },
                  "fullnameOverride": "",
                  "gitops": {
                    "argocd": {
                      "enabled": false,
                      "namespace": "argocd"
                    },
                    "fileFormat": "json",
                    "git": {
                      "email": "",
                      "secret": "git-creds",
                      "skipVerifyGit": true,
                      "token": "",
                      "user": "",
                      "webhook": {
                        "service": {
                          "create": true,
                          "loadBalancerSourceRanges": {}
                        }
                      }
                    }
                  },
                  "image": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e",
                    "pullPolicy": "Always"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": []
                      }
                    ],
                    "tls": []
                  },
                  "ingressGateway": {
                    "ingressNamespace": "istio-system",
                    "kfServingIngressService": "istio-ingressgateway",
                    "seldonIngressService": "istio-ingressgateway"
                  },
                  "kfserving": {
                    "curlForm": "MODEL_NAME={{ .ModelName }}\u003cbr\u003e\nCLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\nSERVICE_HOSTNAME=$(kubectl get inferenceservice {{ .ModelName }} -o jsonpath='{.status.url}' | cut -d \"/\" -f 3)\u003cbr\u003e\ncurl -v -H \"Host: ${SERVICE_HOSTNAME}\" {{ .KfServingProtocol }}://$CLUSTER_IP/v1/models/$MODEL_NAME:predict -d '{{ .Payload }}'\n",
                    "enabled": false,
                    "explainForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:explain",
                    "protocol": "http",
                    "requestForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:predict"
                  },
                  "loadtest": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c"
                  },
                  "metadata": {
                    "pg": {
                      "enabled": false,
                      "secret": "metadata-postgres"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "openshiftMarketplace": {
                    "cleanupClusterServiceVersions": false,
                    "kubectlCleanupImage": "registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382",
                    "prometheus": {
                      "monitorSpecs": {
                        "create": true
                      }
                    },
                    "seldonCore": {
                      "istioGateway": {
                        "create": true,
                        "name": "seldon-gateway",
                        "namespace": "istio-system"
                      },
                      "subscription": {
                        "apiVersion": "operators.coreos.com/v1alpha1",
                        "channel": "stable",
                        "create": true,
                        "istioEnabled": true,
                        "metricsPath": "/metrics",
                        "requestLoggerEndpoint": "http://broker-ingress.knative-eventing.svc.cluster.local/seldon-logs/default"
                      }
                    }
                  },
                  "prometheus": {
                    "knative": {
                      "url": "http://prometheus-system-np.knative-monitoring.svc.cluster.local:8080/api/v1/"
                    },
                    "seldon": {
                      "jwtSecretKey": "jwt-seldon.txt",
                      "jwtSecretName": "jwt-seldon",
                      "namespaceMetricName": "namespace",
                      "resourceMetricsUrl": "https://prometheus-k8s.openshift-monitoring:9091/api/v1/",
                      "serviceMetricName": "exported_service",
                      "url": "https://thanos-querier.openshift-monitoring.svc:9091/api/v1/"
                    }
                  },
                  "rbac": {
                    "clusterWide": true,
                    "create": true,
                    "readNamespaces": true
                  },
                  "replicaCount": 1,
                  "requestLogger": {
                    "create": true,
                    "elasticsearch": {
                      "host": "elasticsearch-seldon-es-http.seldon-logs",
                      "port": "9200",
                      "protocol": "https"
                    },
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398",
                    "imagePullPolicy": "IfNotPresent",
                    "namespace": {
                      "create": false,
                      "name": "seldon-logs"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "600m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "200Mi"
                      }
                    },
                    "trigger": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "broker": "default"
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "800m",
                      "memory": "800Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "seldon": {
                    "curlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions -d '{{ .Payload }}'\n",
                    "enabled": true,
                    "explainForm": "http://{{ .ModelName }}-{{ .Predictor }}-explainer.{{ .Namespace }}:9000/v1/models/{{ .ModelName }}:explain",
                    "kfservingV2CurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .GraphModelName }}/infer -d '{{ .Payload }}'\n",
                    "protocol": "http",
                    "seldonRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions",
                    "tensorFlowCurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict -d '{{ .Payload }}'\n",
                    "tensorflowRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict",
                    "v2RequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .ModelName }}/infer"
                  },
                  "service": {
                    "port": 80,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "create": true
                  },
                  "serviceAccountName": "seldon-deploy",
                  "skipVerifyHttpCalls": true,
                  "tolerations": [],
                  "virtualService": {
                    "create": true,
                    "gateways": [
                      "istio-system/seldon-gateway"
                    ],
                    "prefix": "/seldon-deploy/"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
          createdAt: "2020-12-11 15:00:00"
          description: The Seldon deploy operator allows the creation of Seldon Deploy
            applications to provide management, monitoring and updating of machine
            learning models from a range of technologies. It includes the ability
            to run explanations as well as outlier, drift and adversarial detection
            against running models. There is a full metrics and logging integration
            via Prometheus and ELK stacks as well as full GitOps operations to allow
            full auditing and recovery of updated models.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SeldonIO/seldon-deploy-operator
          support: Seldon Dev
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: SeldonDeploy
            name: seldondeploys.machinelearning.seldon.io
            version: v1alpha1
        description: The Seldon deploy operator allows the creation of Seldon Deploy
          applications to provide management, monitoring and updating of machine learning
          models from a range of technologies. It includes the ability to run explanations
          as well as outlier, drift and adversarial detection against running models.
          There is a full metrics and logging integration via Prometheus and ELK stacks
          as well as full GitOps operations to allow full auditing and recovery of
          updated models.
        displayName: Seldon Deploy Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - machine learning
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://deploy.seldon.io
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        provider:
          name: seldon
        relatedImages:
        - registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398
        - registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
        - registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c
        version: 1.2.1
      entries:
      - name: seldon-deploy-operator.v1.2.1
        version: 1.2.1
      - name: seldon-deploy-operator.v1.2.0
        version: 1.2.0
      - name: seldon-deploy-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: seldon-deploy-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1alpha1",
                "kind": "SeldonDeploy",
                "metadata": {
                  "name": "seldondeploy-sample"
                },
                "spec": {
                  "affinity": {},
                  "alibidetect": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306"
                  },
                  "batchjobs": {
                    "mc": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48"
                    },
                    "processor": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d"
                    },
                    "pvc": {
                      "defaultSize": "1Gi"
                    },
                    "serviceAccount": "workflow"
                  },
                  "defaultUserID": "",
                  "docker": {
                    "user": "unknown"
                  },
                  "elasticsearch": {
                    "basicAuth": true,
                    "secret": {
                      "name": "elastic-credentials",
                      "passwordKey": "password",
                      "userKey": "username"
                    },
                    "url": "https://elasticsearch-seldon-es-http.seldon-logs:9200"
                  },
                  "enableAppAnalytics": false,
                  "enableAppAuth": false,
                  "env": {
                    "USERID_CLAIM_KEY": "name"
                  },
                  "external": {
                    "protocol": "http"
                  },
                  "fullnameOverride": "",
                  "gitops": {
                    "argocd": {
                      "enabled": false,
                      "namespace": "argocd"
                    },
                    "fileFormat": "json",
                    "git": {
                      "email": "",
                      "secret": "git-creds",
                      "skipVerifyGit": true,
                      "token": "",
                      "user": "",
                      "webhook": {
                        "service": {
                          "create": true,
                          "loadBalancerSourceRanges": {}
                        }
                      }
                    }
                  },
                  "image": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e",
                    "pullPolicy": "Always"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": []
                      }
                    ],
                    "tls": []
                  },
                  "ingressGateway": {
                    "ingressNamespace": "istio-system",
                    "kfServingIngressService": "istio-ingressgateway",
                    "seldonIngressService": "istio-ingressgateway"
                  },
                  "kfserving": {
                    "curlForm": "MODEL_NAME={{ .ModelName }}\u003cbr\u003e\nCLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\nSERVICE_HOSTNAME=$(kubectl get inferenceservice {{ .ModelName }} -o jsonpath='{.status.url}' | cut -d \"/\" -f 3)\u003cbr\u003e\ncurl -v -H \"Host: ${SERVICE_HOSTNAME}\" {{ .KfServingProtocol }}://$CLUSTER_IP/v1/models/$MODEL_NAME:predict -d '{{ .Payload }}'\n",
                    "enabled": false,
                    "explainForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:explain",
                    "protocol": "http",
                    "requestForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:predict"
                  },
                  "loadtest": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c"
                  },
                  "metadata": {
                    "pg": {
                      "enabled": false,
                      "secret": "metadata-postgres"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "openshiftMarketplace": {
                    "cleanupClusterServiceVersions": false,
                    "kubectlCleanupImage": "registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382",
                    "prometheus": {
                      "monitorSpecs": {
                        "create": true
                      }
                    },
                    "seldonCore": {
                      "istioGateway": {
                        "create": true,
                        "name": "seldon-gateway",
                        "namespace": "istio-system"
                      },
                      "subscription": {
                        "apiVersion": "operators.coreos.com/v1alpha1",
                        "channel": "stable",
                        "create": true,
                        "istioEnabled": true,
                        "metricsPath": "/metrics",
                        "requestLoggerEndpoint": "http://broker-ingress.knative-eventing.svc.cluster.local/seldon-logs/default"
                      }
                    }
                  },
                  "prometheus": {
                    "knative": {
                      "url": "http://prometheus-system-np.knative-monitoring.svc.cluster.local:8080/api/v1/"
                    },
                    "seldon": {
                      "jwtSecretKey": "jwt-seldon.txt",
                      "jwtSecretName": "jwt-seldon",
                      "namespaceMetricName": "namespace",
                      "resourceMetricsUrl": "https://prometheus-k8s.openshift-monitoring:9091/api/v1/",
                      "serviceMetricName": "exported_service",
                      "url": "https://thanos-querier.openshift-monitoring.svc:9091/api/v1/"
                    }
                  },
                  "rbac": {
                    "clusterWide": true,
                    "create": true,
                    "readNamespaces": true
                  },
                  "replicaCount": 1,
                  "requestLogger": {
                    "create": true,
                    "elasticsearch": {
                      "host": "elasticsearch-seldon-es-http.seldon-logs",
                      "port": "9200",
                      "protocol": "https"
                    },
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398",
                    "imagePullPolicy": "IfNotPresent",
                    "namespace": {
                      "create": false,
                      "name": "seldon-logs"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "600m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "200Mi"
                      }
                    },
                    "trigger": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "broker": "default"
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "800m",
                      "memory": "800Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "seldon": {
                    "curlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions -d '{{ .Payload }}'\n",
                    "enabled": true,
                    "explainForm": "http://{{ .ModelName }}-{{ .Predictor }}-explainer.{{ .Namespace }}:9000/v1/models/{{ .ModelName }}:explain",
                    "kfservingV2CurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .GraphModelName }}/infer -d '{{ .Payload }}'\n",
                    "protocol": "http",
                    "seldonRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions",
                    "tensorFlowCurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict -d '{{ .Payload }}'\n",
                    "tensorflowRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict",
                    "v2RequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .ModelName }}/infer"
                  },
                  "service": {
                    "port": 80,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "create": true
                  },
                  "serviceAccountName": "seldon-deploy",
                  "skipVerifyHttpCalls": true,
                  "tolerations": [],
                  "virtualService": {
                    "create": true,
                    "gateways": [
                      "istio-system/seldon-gateway"
                    ],
                    "prefix": "/seldon-deploy/"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
          createdAt: "2020-12-11 15:00:00"
          description: The Seldon deploy operator allows the creation of Seldon Deploy
            applications to provide management, monitoring and updating of machine
            learning models from a range of technologies. It includes the ability
            to run explanations as well as outlier, drift and adversarial detection
            against running models. There is a full metrics and logging integration
            via Prometheus and ELK stacks as well as full GitOps operations to allow
            full auditing and recovery of updated models.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SeldonIO/seldon-deploy-operator
          support: Seldon Dev
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: SeldonDeploy
            name: seldondeploys.machinelearning.seldon.io
            version: v1alpha1
        description: The Seldon deploy operator allows the creation of Seldon Deploy
          applications to provide management, monitoring and updating of machine learning
          models from a range of technologies. It includes the ability to run explanations
          as well as outlier, drift and adversarial detection against running models.
          There is a full metrics and logging integration via Prometheus and ELK stacks
          as well as full GitOps operations to allow full auditing and recovery of
          updated models.
        displayName: Seldon Deploy Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - machine learning
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://deploy.seldon.io
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        provider:
          name: seldon
        relatedImages:
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d
        - registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48
        - registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c
        - registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e
        version: 1.2.1
      entries:
      - name: seldon-deploy-operator.v1.2.1
        version: 1.2.1
      - name: seldon-deploy-operator.v1.2.0
        version: 1.2.0
      - name: seldon-deploy-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: seldon-deploy-operator-rhmp
    provider:
      name: seldon
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: rhods-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhods-operator.2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
          createdAt: "2025-08-12T18:04:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "llamastackoperator": {
                      "managementState": "Removed"
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (including PyTorch, TensorFlow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack – unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.23
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fbe10282ada2fa6bf07ae97b497f46ebe735de130b4ce0182bf34301429999cb
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:2d2e7803b2b1e15017ca04961f9dcd25e89c47b07b101ca54cb9e793e6027202
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:e50adae13e9d773e7be8bf81b7525dac3f5e3864c9cdf4c287342bd568ee1d6b
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:bb7c352ccc10a8c60ef8d573368d0ada59e73cd131653d11bcebdf9f564acc10
        - registry.redhat.io/rhel9/mariadb-105@sha256:050dd5a7a32395b73b8680570e967e55050b152727412fdd73a25d8816e62d53
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:ccfc84028a95f1cc6a16e06ff6e9ed2ace74f74e5e9963a4fb33544c2ce005c2
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:2886e40f94fde72eb5959ad564bbdd466657208c09bac731b3860b0b54f104c6
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:b038d20def504c140e4365f2eb117a979534629332bec75a2ba750ebafdefe3d
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5fb7e6c3ee1e3b170cd146ea9cd172fa953326b08163897605ca74a3644f2819
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:795ab1842ef675e3d8b47deb8a7485a69064428e823dfeb930a6a7e7017cde0f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:80d284cac2ef6638fe028102b71a70f26041576c11a49414e77430954a3c1b99
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f23f8051886d0f0240937507471c818f440bca8703b8209a836d5dbe3ea2f7c8
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:16ff53f4f1a1e1e66826c8355c654038da83df545427a931485ec7a2d3764b16
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:23c098f7e791441cfcfa6bf37de51763511bd8bf4d543747b7d37ed591d68886
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:edcce7cddfa4c250108532e56541ddb95099612c180b2297e5934bc89d2bb46e
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:ff740fc3615edaa90dae976da1aa6455d64811e239eb66b422836a3aae791b41
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:240692b86bc8a4fe0b545e6e4cd7cf4ff0ca990406bc55f0446d565fd67e37f6
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:a4649e8ad4970bc3d5e967811c9694fa82d78ee9f8f025b8b2fbd37237eb14fc
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:68e74bdf518fc624cc25a59a18b2a87d0aa2a9b8aad30750ace7115d297a21ca
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:ca21e218e26c46e3c63d926241846f8f307fd4a586cc4b04147da49af6018ef5
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:857560240bb45bfc2db6d7743abf11a47f929a0d3035940256a6527afd313412
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:84c3213e11299ec032aa0e5a4dfa9b3055c4246b04b029e888d8d9f1caf5c30c
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:4847ac1be55284815670c2048c1edfc56c7227494280d3c6094aa23f1ed1a9d2
        - registry.redhat.io/ubi9/toolbox@sha256:4213516b6fe7957a01072720efd9ecbc7d4443238e151b5fe0263bd0c36792b1
        - registry.redhat.io/rhoai/odh-llama-stack-core-rhel9@sha256:43b60b1ee6f66fec38fe2ffbbe08dca8541ef162332e4bd8e422ecd24ee02646
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:467b04f4ee4118bbe0934d1ab8b80f72df8703dcdac61c2b53fdb5bd1b404c14
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:a1497e64349634b2567d28f685052de118f64f6b32e7d3ea469212e2004d7eb3
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:e0c9ac73891781ef421e0a020cfd5c962f7632d38020897183cf048b1819326a
        - registry.redhat.io/rhoai/odh-llama-stack-k8s-operator-rhel9@sha256:626b9ef09ee82c8d425ddb7ed3f2ffbb3d1019f541e31604db891787c881d5c9
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:3243caa4c8aac75069d19225dcee361bf9c012782328a6266b7016b3026f6d56
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:756b7bdaa7f6e7c4a18488e690188dd91153a33913d225bab001b0085382e514
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:da8debeb2b84a42aba0b98785ccc20e02ed520e125428fe012df3ca473cbb755
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:208005da389b7d4176ad8bffc703312011a1299954df151e2abbe8f3dec01d19
        version: 2.23.0
      entries:
      - name: rhods-operator.2.23.0
        version: 2.23.0
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.20.0
        version: 2.20.0
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.18.0
        version: 2.18.0
      - name: rhods-operator.2.17.0
        version: 2.17.0
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.15.0
        version: 2.15.0
      - name: rhods-operator.2.14.0
        version: 2.14.0
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: alpha
    - currentCSV: rhods-operator.2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhods/odh-rhel8-operator@sha256:a4d7f020509b9720063e87e1c2f19af04e4c9b4986b199be4dc617831b971434
          createdAt: "2023-11-21T08:37:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            Data Science
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:08c90a8f133b4cdb1477a742bc44be4d8fc4e88153b8bc2a2a148cbdf19f4e7d
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:6a56010de933f172b195a1a575855d37b70a4968be8edb35157f6ca193969ad2
          olm.skipRange: '>=1.33.0 <2.4.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "rhods",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "rhods",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Data Science"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:eb96f4b753606528750b3c28f7f9256582f81fa030450b05afdef684634903be
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift Data Science
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift Data Science (RHODS) is a complete platform for the entire lifecycle of your AI/ML projects. It is the flagship product of OpenShift AI.

          When using RHODS, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. RHODS can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * RHODS dashboard
          * Curated Notebook Images (incl CUDA, PyTorch, Tensorflow)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh (and Kserve) with a provided OpenVino Model Serving Runtime
          * Ability to use other runtimes for serving
          * Model Monitoring
        displayName: Red Hat OpenShift Data Science
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift Data Science
        - RHODS
        - rhods
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift Data Science
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-data-science
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhods/odh-ml-pipelines-api-server-rhel8@sha256:be57f237df156474a709667fa4fb4aa142118e3459963fed56648c1cb2e38cd6
        - registry.redhat.io/rhods/odh-kuberay-operator-controller-rhel8@sha256:bfc805aa1689a4bd6f7e617acf9eb1c2c307624dd49956e9e7442afac328151a
        - registry.redhat.io/rhods/odh-data-science-pipelines-operator-controller-rhel8@sha256:250db47cb873207cfb75c7e8a2f0de08b75b056a72db73362d2d79c08cb7b185
        - registry.redhat.io/rhods/odh-trustyai-service-operator-rhel8@sha256:68471555d0b2bb8150ebfae8d4f3d1dbdff9c5b961ff2305ef23ab7514cbbea1
        - registry.redhat.io/rhods/odh-codeflare-operator-rhel8@sha256:2841fd3051a50a1b0d9e3caafa20a8084b5c472d49393b372828960dbf156674
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhods/odh-model-controller-rhel8@sha256:fde989817cd1a6ec8eb408a7f6c73e28ba043ab2d88035586b1b7e55fb3405cb
        - registry.redhat.io/rhods/odh-modelmesh-runtime-adapter-rhel8@sha256:08478a1647d0413783bcbcd966a73b444b580464de46efedb39504420f22053e
        - registry.redhat.io/rhel8/mariadb-103@sha256:08c90a8f133b4cdb1477a742bc44be4d8fc4e88153b8bc2a2a148cbdf19f4e7d
        - registry.redhat.io/rhods/odh-trustyai-service-rhel8@sha256:fe17d6e07ebaa088f274dd8b5e83892dacc11cb336065b06582a3c055095dba5
        - registry.redhat.io/rhods/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:c5758102609821a6fdabec0277e03c4c728f82176093d1ade0570aba5f3cd8f9
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhods/odh-ml-pipelines-persistenceagent-rhel8@sha256:857a8eb444375505ccb34a11cce69c1794ceabbfb00df06933d7f12cd187319b
        - registry.redhat.io/ubi8/ubi-micro@sha256:6a56010de933f172b195a1a575855d37b70a4968be8edb35157f6ca193969ad2
        - registry.redhat.io/rhods/odh-rhel8-operator@sha256:a4d7f020509b9720063e87e1c2f19af04e4c9b4986b199be4dc617831b971434
        - registry.redhat.io/rhods/odh-mm-rest-proxy-rhel8@sha256:8d989f54009484d7f43fe79d7c37757418fc3c15cd4d7e80a2076677a9645bdf
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhods/odh-ml-pipelines-cache-rhel8@sha256:33aae32e3d348d382906f6e926281852f54006a94d7bdac188bb33debadec059
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhods/odh-modelmesh-rhel8@sha256:3827221f84d3aa22440d6b364b990a4ec0666dfb5963f174f7ed518b3650d223
        - registry.redhat.io/rhods/odh-kf-notebook-controller-rhel8@sha256:cd0ab8d70559be6074fbaf1caa1d50f8eb75e1bff1a22d6d69380f19190df0b6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:eb96f4b753606528750b3c28f7f9256582f81fa030450b05afdef684634903be
        - registry.redhat.io/rhods/odh-ml-pipelines-artifact-manager-rhel8@sha256:76c67d7015a2db6965219db6533e0cddb5b7cf699c88df3ef81155f9c192a3cc
        - registry.redhat.io/rhods/odh-modelmesh-serving-controller-rhel8@sha256:d94395fb251e45aaa73d09bc6106a2947c16b90d87fc3b64bfe4f2f600be5841
        - registry.redhat.io/rhods/odh-dashboard-rhel8@sha256:b98171c5543815abf7c095fba92796ceed9ea44bd8b78db143d2815c6bcfdcf6
        - registry.redhat.io/rhods/odh-notebook-controller-rhel8@sha256:753994fce63802f1e1121990a5c1dd722216446abee54cc1a7c50ab195134414
        version: 2.4.0
      entries:
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: beta
    - currentCSV: rhods-operator.2.21.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:7d0b8d69b6defb5f232da066573b0067ec810a8ccfb2f3cb91d472d2979ab7a7
          createdAt: "2025-09-24T14:42:33Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.21
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:687c8eeed55f021ecaab1307f0e88b5b16d91f72d63b3d7168d7bbee90e8947b
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:04859c878ffcc8c42550a98be625ac52212f9b79a6fd5b3baeafb3592926cbe9
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:2a16c353fd13dc1f3f5c4266f1b9786a8e7a100dc05f2d554bdaef082271063f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:239f226328e5ac77d6dee498217d55103a6d562f4829e38abf65ba55fda0aca2
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:7da004e158006dcbdcf0c3b78c40411b82c9d56ebf54b1c30c245b6550b8ee7d
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:2b4136569e10ad7d47c82bedf55c1705a833dc67c8f833656996bfe1fddbef3e
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:52e1ba8ebf4d55c64c94c9076185c5612b537e6e5ed4944fcd6ea87ccb9cd182
        - registry.redhat.io/ubi9/toolbox@sha256:7656f815a0d13ed37f56d11635559a87a3d00840253755c71b644b858f67d556
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:7d0b8d69b6defb5f232da066573b0067ec810a8ccfb2f3cb91d472d2979ab7a7
        - registry.redhat.io/rhel9/mariadb-105@sha256:c390b12f5c2c3ea7cd897b59eba7127ce22f474130cd8c1bbbbaadae870baca4
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:1b5452b87b36de2b62febc777efdc6233a168ffcc09e2067e482cc444463a57d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:cb47e165935f0d9d3c88088689c37e68a69519357a20d3c52ab6f71240dab507
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:003fdc6cd1821d620e1ce4050ee608e2d2a32f2b08c8fa9c147e2e14418a8d91
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:80fd53a580ce7bc4d02023d38a04912e84c2cd50679fc5ea034afba935623157
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:fe6593bc73e77d315d82b7921bb1975c19fda1e7630ef5b30b69f5901ce30178
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:e68492a6a75cb83be0a968b832c1478b97f14e661376328085fd5416a08cbc31
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:c3e283c0536266195dd461042ca0036ed98cc33650d806be132db4fba45239e4
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:152b425725228953f43b57a4a034a6725993a5ca556af328ae5e06aa3f1acbfc
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:a88696cff8cd2cc2e295c9ec082b7f25f61dfc9ac12f1c208529ff7036d5af29
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fc98ed817a4e9ab84e2a66063987304f0cdbe6e5620a58a7f25cccd61b34ebf9
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:5c0d56f434298f6e8bf6cf7061cd1ae6df9915225cbf04b010b6b59a64775cc9
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:cfaf4eb86ddb9c717340ab17fa4b83323e5b40977ac70265376c87329a3f6f20
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:4c57dcf4df09fe0d03578cc8b9a3ca2ea485b6af0ca0203bfb85e640058b7daf
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:4c192208c618b49f31cf31c9de1a33c622ce3f120204c2eec8ff7c439ea6785c
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:a936d87646843edea2a84eb429bb44eb730bf3e04e8900fecef66721426dba89
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:9c2d6ce05775ac34ee5695b3367d12bf2a9c14e318a9544d00a56653ea5f2007
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:dc0db686b7688f98b5f8595661022cdffa5d26a078289cc7aed528f0ee1d08a4
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:6a388ee3cbe3813ac955d319f0ab65a81da321d982c9881c6c81fe135311fe1d
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5f581957e4f8deebf6104a394272e6fbb1c54d813bf069e95875d865eac209ad
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:2638e3a90bbb242dd2483fefe9fc8b472f334c510d0139b9b7c875c30dab7984
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:85202cb75250a1ba2691d65b52b306fd77566660dd29cdf56bf8962e8bb4939c
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:1b7befcd29c9c8f783bdd3c5654b0412ea45cd88cd18eb5f8f5a411f1323c6bd
        version: 2.21.1
      entries:
      - name: rhods-operator.2.21.1
        version: 2.21.1
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: embedded
    - currentCSV: rhods-operator.2.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
          createdAt: "2025-09-24T13:27:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.16
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:22557a6fa52d2f311750a9ba253860f423ba697d26efa02ef8524a8258d2a909
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:d7d4fa406e0fcf0507894a7676532b27f45be742467e603a86f98ea5d2615df8
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:2ba2b2c4db8bb334c50f4dfb54059f060361186900a44c06eed00b7a3c43977e
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:7b70af8847b0806d8b43c399e2b3109f016fd864f5e9d30c44e2baca5d1359dc
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a0f72ffefb2b74b488dd949493f5d295a39bb9c97f578bf219d0138601f65468
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:6cf74044ae8d5308a2dfe03fa5d81086c89302db7f4cdbe2f4174a1c48b77869
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:de5f91180ead7d73a1825fe8b032fde9b8b01392569c9789f5ce1b4b9c08a98f
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:ee01e89f98feb185f6cd59c564e590a13e4d8d9ea760cca8de51426eb71b83a1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:04f305c44413ae7dcb7017e53570ee49a509701792c5f50efadd64f47395730b
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:c8456d98e90b6505957ab3686e9fd2f156e29f123c5558e581c206daf1e7d93a
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:f738aab1eab25854c93e1d8d4d98100a8ae7bb45a6b83f0326774e4220b1183b
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:5dcdcc2424602a69451f16d31dbfa1d43cb72c095ba561eb9076f0cd1e8182ed
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:efd039012559786deb3c521a130886c265e88d635e08baace2e573a9df270134
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:27b113d96453c2054d4c965a963d2badba6daac235eddaf23234c76e87bcb069
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:4c5ff3496b2a2a739939d94ee9dafc02b682100785d228dc2fde480fb597b7a5
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:4a7599f8a866eb169c9a62885906adbf6df0417c0be15857df1eef20cd9b1be2
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:633984b538f027c93c5886ea1045dcbe81eeda74acf80001d5fc5f765bdbe0be
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:23d307a36b69e0df04f72a7d3b35e28d8417a8bbe23dba31e8e977569785c078
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:11a301728573adf2b64ea72d0cb2d83ea5d4dbebea759f346e99f18c3d368c6e
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:5192f1dbd2a9ab92ae390c4ae506efbed0970545b6122e95b014728ac937e777
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:2eedd1e2465d2e78fafb39e234dbb413a82713f187c64d70a80340ec94807d30
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f37e4048f3a152798286793f3abfc6ed814453fcbe2667255a7e78eee483c5a3
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:97e2bd9b587f08e135a9aeb9b3e0dc6eafa1a9bdacbb5ecb681ce9bd5aa37fc9
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:65482864055021272a18b5b26792ab00cfa5fc9cc005d8d3a884cc82991506f3
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:4a212ec634225c14beac09be24ddf336e562f2aa9a13555fb1196f366ddae23c
        version: 2.16.3
      entries:
      - name: rhods-operator.2.16.3
        version: 2.16.3
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.8.5
        version: 2.8.5
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: eus-2.16
    - currentCSV: rhods-operator.2.8.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:3b9e8a71ba795a2ea91f4129253ccb4be3a79c7c1c6a87082f96ccc67eabbbac
          createdAt: "2024-11-07T07:17:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.8.0 <2.8.5'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:eeb8bdc94e7fa2b454fedf72eb75756052f8bfb9d92b116a8267a5584e2528c1
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:f9c474d454d0b006e06859601ec9a64b5e5c91e9a47bfab6ad6d8cce5aa6b89d
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:45498198fadfbc1457bf2a6084ae7ea515d454d86d148783aae3abdaca6e999d
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:60507a3dddbd88560fd85204699147facd2b6371fb25e0bf6701dad02c0b4dd4
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:df34e9d2d121a3def2dbb86f108b3fb6948bc7ba7cb44c70801dfd2e81d3fbb1
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:13dff890ec67fa56864c244a40fc404e9eac05f33573dd16d088b37f9340ec41
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:28ec01e514c15cae145499840597d1be1b750123bb6014de2ecc3ef0d60fa134
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:9aa5a93831992588766ac8594ae3c4dc7111d08a90fdbc8751c40484339e46d4
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:038ce76571217d6c624d614652e8b96ef6484562cf52ec1be88392b8ef8a3bbd
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:1a1074263ae26e4c8b0e3863228c35e50133169f377ece6c47b58a12e0fcf453
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:81ae2b95f2da740cb5264b9ba8b85f0f58bb4ab80d6431f66d52848ed0c21afb
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:1dbf09cb1624c9f961f8ac8e371c97aa7c3fbfcad4c4d02ab4fad0eb86d2aa95
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:61a506fbf263cd72951ac8bd151127e0c74fc74f0a05f7aaecaa54fde47ee490
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:3b9e8a71ba795a2ea91f4129253ccb4be3a79c7c1c6a87082f96ccc67eabbbac
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:13b850a5a3f8776c6c8db887c666d88df72901895b5ef1be9a8bd372cc02fcb8
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:8ffc5da798e783c65d5be7f344219b2951d8c2233d6494a4ae2bb5c86eccd699
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:d87e4b1cba72eb4f50b3f2d59863799929c22ca37b01a1ff4828a9ce0be31f25
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:2f6db789ceb3f90f9c9911c6392e62d0a8e4d61be68a7f7cbd781f47faab6389
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:4b1b9f05a94713292a7a2f6782fe1d3eea6f36214c36b38f402aa91a554ab18a
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:5c48474563d98df32e7243e33302428ea8836be5e1a2e30d20b4d796a62cbf6b
        version: 2.8.5
      entries:
      - name: rhods-operator.2.8.5
        version: 2.8.5
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: eus-2.8
    - currentCSV: rhods-operator.2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
          createdAt: "2025-08-12T18:04:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "llamastackoperator": {
                      "managementState": "Removed"
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (including PyTorch, TensorFlow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack – unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.23
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:edcce7cddfa4c250108532e56541ddb95099612c180b2297e5934bc89d2bb46e
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-llama-stack-k8s-operator-rhel9@sha256:626b9ef09ee82c8d425ddb7ed3f2ffbb3d1019f541e31604db891787c881d5c9
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:ff740fc3615edaa90dae976da1aa6455d64811e239eb66b422836a3aae791b41
        - registry.redhat.io/ubi9/toolbox@sha256:4213516b6fe7957a01072720efd9ecbc7d4443238e151b5fe0263bd0c36792b1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhel9/mariadb-105@sha256:050dd5a7a32395b73b8680570e967e55050b152727412fdd73a25d8816e62d53
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:68e74bdf518fc624cc25a59a18b2a87d0aa2a9b8aad30750ace7115d297a21ca
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:a1497e64349634b2567d28f685052de118f64f6b32e7d3ea469212e2004d7eb3
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:e50adae13e9d773e7be8bf81b7525dac3f5e3864c9cdf4c287342bd568ee1d6b
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:84c3213e11299ec032aa0e5a4dfa9b3055c4246b04b029e888d8d9f1caf5c30c
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:ccfc84028a95f1cc6a16e06ff6e9ed2ace74f74e5e9963a4fb33544c2ce005c2
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:bb7c352ccc10a8c60ef8d573368d0ada59e73cd131653d11bcebdf9f564acc10
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:2d2e7803b2b1e15017ca04961f9dcd25e89c47b07b101ca54cb9e793e6027202
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:b038d20def504c140e4365f2eb117a979534629332bec75a2ba750ebafdefe3d
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:3243caa4c8aac75069d19225dcee361bf9c012782328a6266b7016b3026f6d56
        - registry.redhat.io/rhoai/odh-llama-stack-core-rhel9@sha256:43b60b1ee6f66fec38fe2ffbbe08dca8541ef162332e4bd8e422ecd24ee02646
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:23c098f7e791441cfcfa6bf37de51763511bd8bf4d543747b7d37ed591d68886
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5fb7e6c3ee1e3b170cd146ea9cd172fa953326b08163897605ca74a3644f2819
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:857560240bb45bfc2db6d7743abf11a47f929a0d3035940256a6527afd313412
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:795ab1842ef675e3d8b47deb8a7485a69064428e823dfeb930a6a7e7017cde0f
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f23f8051886d0f0240937507471c818f440bca8703b8209a836d5dbe3ea2f7c8
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fbe10282ada2fa6bf07ae97b497f46ebe735de130b4ce0182bf34301429999cb
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:80d284cac2ef6638fe028102b71a70f26041576c11a49414e77430954a3c1b99
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:467b04f4ee4118bbe0934d1ab8b80f72df8703dcdac61c2b53fdb5bd1b404c14
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:ca21e218e26c46e3c63d926241846f8f307fd4a586cc4b04147da49af6018ef5
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:e0c9ac73891781ef421e0a020cfd5c962f7632d38020897183cf048b1819326a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:a4649e8ad4970bc3d5e967811c9694fa82d78ee9f8f025b8b2fbd37237eb14fc
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:208005da389b7d4176ad8bffc703312011a1299954df151e2abbe8f3dec01d19
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:240692b86bc8a4fe0b545e6e4cd7cf4ff0ca990406bc55f0446d565fd67e37f6
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:756b7bdaa7f6e7c4a18488e690188dd91153a33913d225bab001b0085382e514
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:2886e40f94fde72eb5959ad564bbdd466657208c09bac731b3860b0b54f104c6
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:16ff53f4f1a1e1e66826c8355c654038da83df545427a931485ec7a2d3764b16
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:4847ac1be55284815670c2048c1edfc56c7227494280d3c6094aa23f1ed1a9d2
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:da8debeb2b84a42aba0b98785ccc20e02ed520e125428fe012df3ca473cbb755
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        version: 2.23.0
      entries:
      - name: rhods-operator.2.23.0
        version: 2.23.0
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.20.0
        version: 2.20.0
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.18.0
        version: 2.18.0
      - name: rhods-operator.2.17.0
        version: 2.17.0
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.15.0
        version: 2.15.0
      - name: rhods-operator.2.14.0
        version: 2.14.0
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: fast
    - currentCSV: rhods-operator.2.22.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
          createdAt: "2025-09-25T11:28:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.22
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:7b3d7787ba95156de9c284ad78f2d63a0d3bcb525f4d8d11ebf0203d4c5c1fe0
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:7fd540586a1ca111370100a473923b3dc6cec4255e9e4f3654bfbabd9158a2bb
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:a44d32a18b1e906c1411cafc82102881573fcd04e040a4e2ba6bd1bd617b70fa
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:6234e15f001122ab222481321153fb0eba7f4d86bf3dfd64edf3f6cecbe7d62f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:abe1f9dd11c7313fdd55b678be511a5cdc3546654a68e2b53d42c1013fec8d00
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:b2ca229bb4c4dbf04b96766b8af8d9f55f2edef59ec4110de0d20f8c857e7d27
        - registry.redhat.io/rhel9/mariadb-105@sha256:5e77b33ce04c1eafc2e8840586ad6a5b51148dea8a46dda151071e28858331f0
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:2f3f305c97e50057781553416fb4026cde306ead0fbe0213627b9b5bbcc8a0c3
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:e6fbe6723b4801f6854303b3fd7d4a4a36727bdd28dc28002b7de4929739ccc0
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:2f1a0a0803b2f5118b167e9d9c2e86660add5f79ee6f47044f63b0e8abe0b75f
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:cb5be3dd7098f4c8d66f9e85ad04656460df6102652deddc6cccb884a12be733
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:9104cbc2d595933f0a2acb7313e41ed69905e76a9fae6bd9820e7717b56a632b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e3b90530c6e38f6940d7f9408e9fa2c81eb06a857ec2b0bab5d006b1e9a63a4f
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:32c9511846077ef6035eb19be7653f6a78e0f207eb420ab93b18f742ab5779b6
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:27415b1c763cf89b79391f0b039936c3e078a078bda4420f13bf10ef9be31d5e
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:463e9be9c4409ba1e5a42713bd65844a023e7a3e9032d4e862edbaa5fe35300a
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:1709fa3c79aad4ba7eb9be8299949396092c8e20210124e0c0936385bc04e839
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:b2c2ea9d35edbcd16e1ae559c88a453129b877fc049ab7a7cd6c2ebb8f0bcb59
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:bd6d1599b4ff9eb890be9808014715b28ca19247222d65f40df024db4dd3525a
        - registry.redhat.io/ubi9/toolbox@sha256:ea4ce8cb8f59cf1b0709f1606d7cab074c3806d4016964eedc0555410cb5dfeb
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:d59ed40de694bc0e0564bab159a5b586cffa0721b6d761fb2d81dd42e6225d1c
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:74d4e0174655cf6f6273857325a1e714a1531c5a06ad38bf139bee9c7d8f5f19
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:5d79f0bcca0a213b53aee6d88834e503d343821596b857b13d3a6362de8ebeab
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:5914f71e37c815ce2567826896e051303bdef52fae1acf8d02e672fc24c7eb9d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:2065b96bd6f0a830eab66e36217a927448cfc5a7dabf2a5e7d907aa6c40657a4
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:2b3aace0d80474470e71c2c8291b25738e096109229fe6d5d3e06e9885ee58b6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:13f3679ab289eea9de54fe129a7c83f6c8ec5d26b61b430637a7c4cfcd31a74a
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:ccba67a6a5aa3403e0018f29dcdfef483220dea8065d8b572e80c653ca992add
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:2fe9473cd1d6c445d2ca7cc357ee4bb009aa1838f9cb69ea418296b2a044816b
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:d7c4ac937add511460a36375d9b6b3c475a3a9a7cc96c00df2fbad6219aee0e9
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f9965177de24f3165ec4c001c14cf1e2c2c01ffe847839af837c442e3c96beae
        version: 2.22.2
      entries:
      - name: rhods-operator.2.22.2
        version: 2.22.2
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable
    - currentCSV: rhods-operator.2.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:8643f7ee8afacfe6ae08cbbf81da76ca97032dc3ce3da4fddf16e330e9e33a08
          createdAt: "2024-11-14T21:18:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.10.0 <2.10.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:a1d590d0e2b3fae9ecbd843a664dab083061e797f66a9f0a021b0c2c487d7e45
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:0af6a91553f40ea40e86d85edc204658af3e5a45e8ab94cdb6b767e545df59b9
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:214ca0608fcbc0fd585bf01768d28c7ba17082d1043349ec8445afd663095383
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:d2514bf4b4b2d9abc19facac3f1a55405eb3ad0d9e9203e8816bfcfa08c9a6c5
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:4876f14ca9cff0e2b998a56f2f56d0775edf3963de610f4b810b4d94fb12d719
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:24238abd46633709feb75f2a3e41cf6c9e8528d2c0e4be2fac6f56de44e3fb9b
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:d9b55e8f25b81050380b4aab040243dc18a7432c48912c9b660c97dcc5eea008
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:8643f7ee8afacfe6ae08cbbf81da76ca97032dc3ce3da4fddf16e330e9e33a08
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:9f7b1cb8623de8b0550b0ba124d3b50f5e960ed011b33a3262c6346dbe2dff95
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:807ec05e9e107afc969e718d0c2c980959e2fd462e777a5ff63e55156ea150a4
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:fdc2bc06b0dfa79ebc6451ac3793b5e8d9f165171633831dafe46ebc18c10c35
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:4da77afafde99b15fd357a4ab6609644bf6330a1d9c0c03e64726ed1b9633230
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:cba9f25ac6bb2734bb990a622b515c838b881edb688bf44ff196ebe7684ab74b
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:8911afbb65c8c4f10c3597624485be54b9103e87a5020b718b35ec3df2c9bf00
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:d709f65e1194bc877fa4f9ec681b598467703d94139d57b58c24490d3a59fb82
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:220cad7a22d40998063c6b4fb2f63e69330e64ca7118490a6680b05fc99b7733
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:8fed8f58732bff512289fff738d0a94be67286c41b24eaf5928bbcb30d88c88a
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:232ebac25084db056fdde49477b5c1e6a03a0228fea3758bcf6234b3d0eaa582
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:2007da5abf840712cd2edb50462b2d8d571a0193c3ae1550d4a279eddf2b335f
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:a4b9553fde02d9c67a9c98f063eaf82fb1d2689f4b1269e57bf5f12c8d4adecf
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:5d15175f934017c04d6a6057200ff55b2172a1f4bcbeb0f2cd9e9e545631d93f
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:f4d7d3ba4a0d3322bc9f327e095e9ec56073758f6dcd8da8c7bb185cf952ead5
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:7df2f8bc2c5c0f4f347d009b62a058b5107413a2abd789f6010eb4d6c9e4f8ee
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:a9c6e3afdf605465766fa93334cc0f6bb5e46a5fdfd7ec64d648e2ed72b05ccf
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:267d6795bf53f9a1955d73445776a428dd042af0cd3cb8bb5b91692a1bc1e5d2
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:65db5e8cb2fa46d020ed4837d4479d133b52167c373e65f81cb590eec5391af2
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:0ff5664c2eaadf3d3073b3d7fb711da47fb78d21f67710d8b8aad5452cb0355d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:12741379ab24d2286b5fae77f4383fb65713af7b3356fb43d8254ea85d4b0ee3
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:52e8deeeeaa21b17370a5d5cd644745da745b6e1cee8c6295d6df75c8d2aec04
        version: 2.10.2
      entries:
      - name: rhods-operator.2.10.2
        version: 2.10.2
      - name: rhods-operator.2.10.1
        version: 2.10.1
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.10
    - currentCSV: rhods-operator.2.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:28b5fff602ad6fb36d462e1905f453509ec22fc251a1e0cf64fedba86ff420d7
          createdAt: "2024-10-17T10:53:18Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.13.0 <2.13.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:cd6b535d95eb1b44fe261a5336270fd898b966f431c14e173944484381c26d25
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:c2bd493e9345eccd124bbe0440e45cce8cf57588827d8eb9533f8bbadaa3ebbe
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:74e6cfa0bc481588cb8dc61c206ad567fbd2b28f43c2a86bee4a8e0eff55447c
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:28b5fff602ad6fb36d462e1905f453509ec22fc251a1e0cf64fedba86ff420d7
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:77bca19180f30ef08f4d1598ac6c365e190bb7909ac4a1fb35d9ba9ba8d6ff7a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a37529eff3656a570d796083404e5a3f63603afecfe9e738da1a243c0cb3d7c2
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:5d6354e65106460ab9285b42124faf288298387a8b16317565bd591f009c39a6
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:fb630d6f5817512733b8ea05f7147121928917a24ac8d98239f43d920880c359
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:6b2c0d1dd40e2c276684e17818b14375bc778e6c88df0bdc3cc24bfc9294804f
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:8883bf71abf39b4a6985940c287eb6ac7dc137d491264805d48bb112b65bd008
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:9ccdd7fe8223057a2234a3a2eb3b492e7b69e6ce714cc26a805fe2dd82935195
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:850901511bde065cf73e4c18d787dc10b6e91ca6139fc3ba68da28c41dd28236
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:688389195aebf3f43e74e40d591d2fd5f9ae41994d8ac2db661f292f6eda012a
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:2566b73e98b783871fb500bb1a8cc4e3d58a513c1bb18abfb525da0818e6138b
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:fd4ce524acc1769475079c7231f5ea79ac29551c513855ef34e5694d2c9fbbd3
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:a55f73d72ad2b07f01c49949cdb00969bfe4789375b3d7cdf9814b969fcc808d
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:7d085942e4da73c2ff4885c62f10b904dc70fd08e819e62f925eb1b4ddfa1554
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:8faab013568ee023968a5ec5b3a79977d7ddae3deafff6328e6fd56a27198e24
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:71abac4fb3d613edef7f3ba117a73bbeea61f341d6182e8310360a76576c789e
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:030c53690fe25e15bc9907f51c3f04fab39b4d085ff2098d26316ba32cc8b09f
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:8c19c7fa535f36bdce3b73cf3300e321909d44954cbc63a8aef6321c5f2fa693
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:59e9cc9cd57ba48ca7e87ae1e797a902f826bb890cb198ccf549f76ea581ba6e
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:bf25fce160dd7b8cce11e68a501006e4fa81a7858645038cdefdbba03df7ecdb
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:a3048eaa8924ce3571c6e78014438183fae70d9e8c833a7c07774c675104e3d4
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:5c2ee4136ee0014d1ef395b5e3bcab125c846f5e258ece3631d912f431c8d208
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:7c895490d49beb6bd724f47ad1b17ce501650557b0889314799c4072632f2b99
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f5b17f44c5ab72d6cf658f72f24f15e06a5ac9be1790336f7228ab8e7cfc15c3
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:cf3779c64bbaef1988886ed629c516e7d6ef7bdc73b9c95ad4c087b62e75750d
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:363a35c332fc80b068157951d5e50e6b7209947649cc64803ffc02726e7581ab
        version: 2.13.1
      entries:
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.13
    - currentCSV: rhods-operator.2.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
          createdAt: "2025-09-24T13:27:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.16
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:97e2bd9b587f08e135a9aeb9b3e0dc6eafa1a9bdacbb5ecb681ce9bd5aa37fc9
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:efd039012559786deb3c521a130886c265e88d635e08baace2e573a9df270134
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:d7d4fa406e0fcf0507894a7676532b27f45be742467e603a86f98ea5d2615df8
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:7b70af8847b0806d8b43c399e2b3109f016fd864f5e9d30c44e2baca5d1359dc
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f37e4048f3a152798286793f3abfc6ed814453fcbe2667255a7e78eee483c5a3
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:4a7599f8a866eb169c9a62885906adbf6df0417c0be15857df1eef20cd9b1be2
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:4c5ff3496b2a2a739939d94ee9dafc02b682100785d228dc2fde480fb597b7a5
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:4a212ec634225c14beac09be24ddf336e562f2aa9a13555fb1196f366ddae23c
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:2ba2b2c4db8bb334c50f4dfb54059f060361186900a44c06eed00b7a3c43977e
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:f738aab1eab25854c93e1d8d4d98100a8ae7bb45a6b83f0326774e4220b1183b
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:23d307a36b69e0df04f72a7d3b35e28d8417a8bbe23dba31e8e977569785c078
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:04f305c44413ae7dcb7017e53570ee49a509701792c5f50efadd64f47395730b
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:2eedd1e2465d2e78fafb39e234dbb413a82713f187c64d70a80340ec94807d30
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:5dcdcc2424602a69451f16d31dbfa1d43cb72c095ba561eb9076f0cd1e8182ed
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:11a301728573adf2b64ea72d0cb2d83ea5d4dbebea759f346e99f18c3d368c6e
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:27b113d96453c2054d4c965a963d2badba6daac235eddaf23234c76e87bcb069
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:c8456d98e90b6505957ab3686e9fd2f156e29f123c5558e581c206daf1e7d93a
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:de5f91180ead7d73a1825fe8b032fde9b8b01392569c9789f5ce1b4b9c08a98f
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a0f72ffefb2b74b488dd949493f5d295a39bb9c97f578bf219d0138601f65468
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:6cf74044ae8d5308a2dfe03fa5d81086c89302db7f4cdbe2f4174a1c48b77869
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:633984b538f027c93c5886ea1045dcbe81eeda74acf80001d5fc5f765bdbe0be
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:5192f1dbd2a9ab92ae390c4ae506efbed0970545b6122e95b014728ac937e777
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:65482864055021272a18b5b26792ab00cfa5fc9cc005d8d3a884cc82991506f3
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:22557a6fa52d2f311750a9ba253860f423ba697d26efa02ef8524a8258d2a909
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:ee01e89f98feb185f6cd59c564e590a13e4d8d9ea760cca8de51426eb71b83a1
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        version: 2.16.3
      entries:
      - name: rhods-operator.2.16.3
        version: 2.16.3
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.16
    - currentCSV: rhods-operator.2.19.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:c605e5fdb3d66302781146347df6da5a91dee97e4f8f5d9ac33f105b8cdee383
          createdAt: "2025-09-24T13:27:42Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:b30d60cd458133430d4c92bf84911e03cecd02f60e88a58d1c6c003543cf833a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:f0ee0d27bb784e289f7d88cc8ee0e085ca70e88a5d126562105542f259a1ac01
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:8ce44de8c683f198bf24ba36cd17e89708153d11f5b42c0a27e77f8fdb233551
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.19
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:8a6732ac3ca60f13dcb1fd8f186d2fc5dc601b9b5981a58f039af9e237f20044
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:b20ee16b776871d7c49b23f2bc91726812998561129f5dbd14eddbfe57964c93
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:c605e5fdb3d66302781146347df6da5a91dee97e4f8f5d9ac33f105b8cdee383
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:b30d60cd458133430d4c92bf84911e03cecd02f60e88a58d1c6c003543cf833a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:f49483df1d342d8e39f9a8085920aad79313a475df02a2292a136ecb784402b5
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:c8c23c07038f59db848dea0d4247b62840dbe58e1c581dc0722fb9d7f9be3b75
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:4e8f6c6ca0ed516b8310af0332637683b5bfc8f2946eb935c647698dffbbcc79
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhel8/mariadb-103@sha256:f0ee0d27bb784e289f7d88cc8ee0e085ca70e88a5d126562105542f259a1ac01
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:d24e277f468b0a48f38801e8e0bcfeb8264c3f7c128073ddaa17934b8572cc3c
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:520ed21b80e5487b3a0a13b7e5c7d791d4dd28e0b6799d57abb65d53fca5a03f
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:6275a19ef06114b33d79833a94ef10cc8203368eee98e5d415e73d20cae1db85
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:87685c5cab105d876ff0d14b0e24d8dc2a1ff5b55bebe36028c5f3376b2a5271
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:6ad35cf5c4c62ed5ad4e192afe8cebead9796cb24a964401c76ff8656c47742f
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:b6039c3c33a5dc345ceb8e757437bb13023d2bf85cc2b4c10b9ab3887e78e0c0
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:6f01ac7cc62ab961b9c8a630a558338a61f6d7902301c4a1ad337495ac7ce85a
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:4313cae03c041d461bc276abc23d706b582b5d16feab316796ca43478b9dc705
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:b24c25dfdcefb4ec26e7f311c00798df108f7685def2b1762b7624fdfb160ebf
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:bbef4839f24facce111d21aca43898dcf4e6cf57adb16200e3f1590bd5f89eae
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:49c02285fcef6949892cecca20eeee49d03a9bb4d17bb9ceff7c7a0dbbb8ab40
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel8@sha256:6166f525948a018c41d735014d97fdfa52024f7be31e18976d85d1a0d88901a0
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:baefbe4e7dd5aa74c71cb81da4a7a2b888a02d3eb1398ec4419bd8003b77f0c2
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:8ce44de8c683f198bf24ba36cd17e89708153d11f5b42c0a27e77f8fdb233551
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:af4721d61dfb61c75488a1bd6a5f62b23e20d7b7586a9cf51ba062ece2837715
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:bf6231bfdbe52b597d86e03e4730ec105ebe7a639fcd529268b7c0423f2a890c
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:71cd15f0b1acba3a48ebedff037bf8a846b426c18363a4de8896a376b6b7b3d5
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:53ac36baa374159b9065c718a9ede821bbb61d9ebe9502b2243e0a9f7aca0d16
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:d47514cbceb3cc96e9d33b7db6fc1df60893e5c4118bf6086cdd27f04c472773
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:876ca7ce7b0b975688fb4a77d6cd1a6f87cf9eb40c2a9a5b84d8cffdc80eeebd
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:de0e4b7fb822101198817e83cd7e9f4f7406e75ec91d07a833a27a417cffe66b
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:c1c97b6beb759d3f6f2dd53c4258e4ccbb3bf35cf11873bd446b135defaa69d6
        version: 2.19.3
      entries:
      - name: rhods-operator.2.19.3
        version: 2.19.3
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.19
    - currentCSV: rhods-operator.2.22.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
          createdAt: "2025-09-25T11:28:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.22
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:13f3679ab289eea9de54fe129a7c83f6c8ec5d26b61b430637a7c4cfcd31a74a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:d59ed40de694bc0e0564bab159a5b586cffa0721b6d761fb2d81dd42e6225d1c
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:d7c4ac937add511460a36375d9b6b3c475a3a9a7cc96c00df2fbad6219aee0e9
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:b2ca229bb4c4dbf04b96766b8af8d9f55f2edef59ec4110de0d20f8c857e7d27
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:9104cbc2d595933f0a2acb7313e41ed69905e76a9fae6bd9820e7717b56a632b
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f9965177de24f3165ec4c001c14cf1e2c2c01ffe847839af837c442e3c96beae
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:2f3f305c97e50057781553416fb4026cde306ead0fbe0213627b9b5bbcc8a0c3
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:b2c2ea9d35edbcd16e1ae559c88a453129b877fc049ab7a7cd6c2ebb8f0bcb59
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:6234e15f001122ab222481321153fb0eba7f4d86bf3dfd64edf3f6cecbe7d62f
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:5d79f0bcca0a213b53aee6d88834e503d343821596b857b13d3a6362de8ebeab
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:a44d32a18b1e906c1411cafc82102881573fcd04e040a4e2ba6bd1bd617b70fa
        - registry.redhat.io/rhel9/mariadb-105@sha256:5e77b33ce04c1eafc2e8840586ad6a5b51148dea8a46dda151071e28858331f0
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:7fd540586a1ca111370100a473923b3dc6cec4255e9e4f3654bfbabd9158a2bb
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:2fe9473cd1d6c445d2ca7cc357ee4bb009aa1838f9cb69ea418296b2a044816b
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:27415b1c763cf89b79391f0b039936c3e078a078bda4420f13bf10ef9be31d5e
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:7b3d7787ba95156de9c284ad78f2d63a0d3bcb525f4d8d11ebf0203d4c5c1fe0
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:2065b96bd6f0a830eab66e36217a927448cfc5a7dabf2a5e7d907aa6c40657a4
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e3b90530c6e38f6940d7f9408e9fa2c81eb06a857ec2b0bab5d006b1e9a63a4f
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:e6fbe6723b4801f6854303b3fd7d4a4a36727bdd28dc28002b7de4929739ccc0
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/ubi9/toolbox@sha256:ea4ce8cb8f59cf1b0709f1606d7cab074c3806d4016964eedc0555410cb5dfeb
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:5914f71e37c815ce2567826896e051303bdef52fae1acf8d02e672fc24c7eb9d
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:1709fa3c79aad4ba7eb9be8299949396092c8e20210124e0c0936385bc04e839
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:cb5be3dd7098f4c8d66f9e85ad04656460df6102652deddc6cccb884a12be733
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:2b3aace0d80474470e71c2c8291b25738e096109229fe6d5d3e06e9885ee58b6
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:2f1a0a0803b2f5118b167e9d9c2e86660add5f79ee6f47044f63b0e8abe0b75f
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:ccba67a6a5aa3403e0018f29dcdfef483220dea8065d8b572e80c653ca992add
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:463e9be9c4409ba1e5a42713bd65844a023e7a3e9032d4e862edbaa5fe35300a
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:abe1f9dd11c7313fdd55b678be511a5cdc3546654a68e2b53d42c1013fec8d00
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:bd6d1599b4ff9eb890be9808014715b28ca19247222d65f40df024db4dd3525a
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:74d4e0174655cf6f6273857325a1e714a1531c5a06ad38bf139bee9c7d8f5f19
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:32c9511846077ef6035eb19be7653f6a78e0f207eb420ab93b18f742ab5779b6
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        version: 2.22.2
      entries:
      - name: rhods-operator.2.22.2
        version: 2.22.2
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.22
    - currentCSV: rhods-operator.2.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:b3563e201bf2720f60e73c0b04cb70566ea52eda7d09a28753f3a1487993c2e8
          createdAt: "2024-08-01T06:23:03Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.8.0 <2.8.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:eb4da3a645b53acfd6ce8758c043ae19b4126e62bcca09165114a141c2c8d014
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:92d9e4b53d580df6cc1f14a849b477749a1306941a2e8f9d1b6160f4911c3cff
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:b3563e201bf2720f60e73c0b04cb70566ea52eda7d09a28753f3a1487993c2e8
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:5f39b11f0cd6db2d7d6e51ee666c030e2643aee92a090b49656c1e7dfd8a5430
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:90833e85e0de3d7de870074c8b04319cd11b913b9652af6a5727aae41a91458c
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:91525fac5fadece75fbdafc40b94445fb92c88cfc8c2f06b596f8c6377e59c9e
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:ab8792fcd16b1aa0065444ba54335cb3e747379740c3ec2d56beb46ff80cb044
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:987c768c0688431a53e48a80591f2a41e074994169f9e691b28b89528d8e736f
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:1a6e29214e06a2665baadbed598ba758d6141bfa522984eb05f05a80558a403c
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:850b78a201effcafe33705dcd2670f1c470d8b46ba6e307032403627b55a7cfe
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:080377d42d782bf525b4072384d0cf8409b48019b3748ca7c855f340bcb94251
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:b6ecb3858c201498024e3bfc4a93e630699be4b1fdc3a65fd2d70e4b375eb9c2
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:df2edcaf4402b11ce24439263956e3b0cc43efbac2fdf152afacb1a398dcc59a
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:889f046b255b13d99801108588d3abf577611b828657b5240a18ea21f0e3cbbe
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:1c49888ba2e9449c04d5588040c099290fce4793c2ba8874c1037316e550d6e1
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:0b8665e8978deb73be2cc65830f08a74a5a48f3e3ee7b2b817f7bbacba257d22
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:68e1331f13dfaf6641a1e7d54f4cf29c0ef42663abc98b6dc69167892d2d8adb
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:06a6d5cbbe80d49bca9e4db9929571a69e35ae7951fcae84dcd6460f7ddf622c
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:9799003449bfb61ae8f59efd2a611b0da3bccadd9b20cc1519f6c9279d085337
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:9c90873fbd5af8fac8efc3c7c27fba742a06fad2d47f336319dd05d2b116867b
        version: 2.8.4
      entries:
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.8
    defaultChannel: stable
    packageName: rhods-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: ansible-automation-platform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aap-operator.v2.4.0-0.1758079821
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.1-0.1635279521 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the hub, including all hosted Ansible
              content, secrets and the database. By default, a persistent volume claim
              will be created using the default StorageClass on your cluster to store
              the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous hub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Automation Hub
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-24

          ## How to Upgrade

          Starting with AAP 2.4 in the May 2025 async release, the managed database was upgraded to PostgreSQL 15. It is recommended to take a backup of your deployment before upgrading.

          To take a backup, create AutomationControllerBackup and AutomationHubBackup objects for your deployments.
          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel.

          Note: Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8-operator@sha256:5b957031609c8aabf7575e675bc636c1152babc9564d503802301aff4642a404
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-runner-rhel8@sha256:e14286dc9605be3a8c6cb144d0b2e3f3bb3211129005009e7566ea6c16c91f52
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8@sha256:7ae8db2b5d12b619f40db1e581db1e982a695bc1a29a2628222fff09e7c25249
        - registry.redhat.io/ansible-automation-platform-24/ee-minimal-rhel8@sha256:7670e41037f023e3fc6ee95d2342787a80077273ab8d30ba6804029ff05b4edc
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-ui-rhel8@sha256:d229d4e09b2b788d1ba3413ea7ec15374bd1613d507c285ac1ddc10570b9473f
        - registry.redhat.io/ansible-automation-platform-24/hub-web-rhel8@sha256:6ff1f602f0d9cad201b3addb65acc570a9d527de005ca85fc861d3cb98138710
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8@sha256:bfd13d6b39c5ec3a67adbafb05a0482328f0b9b97a517b45c093f6c35a3a92d3
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8@sha256:04fb5cb6b0d7c79cf5d4b754fe917224ddfcb2b9a4994462db404c833ed69483
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-rhel8-operator@sha256:b120e6a5ef376662b4324f0458eae31627087dc91fc613788143bc3954a15318
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-24/ee-supported-rhel8@sha256:f1384eb558f7da1e4a35d51408b3cf5a5a9be08f4d50503f473b26909bc246fc
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8@sha256:cccb4db08735551c697a7b2cbf75812f34500e691f9cdbaea02b30edea292e47
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8-operator@sha256:5253c9fd587ff2ca94d251f2b5fc7f284080adc5eb64643df5299beb2015592f
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8-operator@sha256:3c28a1326ecb7495952a5bb8be13b0972a4378af82611b0b9d990c5fa9fa82c3
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8-operator@sha256:e20106a8a04abd568a4c56861cea39ae444c0bd204bc807ac01f1705e12937d3
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        version: 2.4.0+0.1758079821
      entries:
      - name: aap-operator.v2.4.0-0.1702352195
        version: 2.4.0+0.1702352195
      - name: aap-operator.v2.4.0-0.1707506910
        version: 2.4.0+0.1707506910
      - name: aap-operator.v2.4.0-0.1708974895
        version: 2.4.0+0.1708974895
      - name: aap-operator.v2.4.0-0.1711590540
        version: 2.4.0+0.1711590540
      - name: aap-operator.v2.4.0-0.1713466670
        version: 2.4.0+0.1713466670
      - name: aap-operator.v2.4.0-0.1715898994
        version: 2.4.0+0.1715898994
      - name: aap-operator.v2.4.0-0.1717558474
        version: 2.4.0+0.1717558474
      - name: aap-operator.v2.4.0-0.1718152210
        version: 2.4.0+0.1718152210
      - name: aap-operator.v2.4.0-0.1720482179
        version: 2.4.0+0.1720482179
      - name: aap-operator.v2.4.0-0.1725256739
        version: 2.4.0+0.1725256739
      - name: aap-operator.v2.4.0-0.1726111413
        version: 2.4.0+0.1726111413
      - name: aap-operator.v2.4.0-0.1727257740
        version: 2.4.0+0.1727257740
      - name: aap-operator.v2.4.0-0.1730152858
        version: 2.4.0+0.1730152858
      - name: aap-operator.v2.4.0-0.1733185647
        version: 2.4.0+0.1733185647
      - name: aap-operator.v2.4.0-0.1733943951
        version: 2.4.0+0.1733943951
      - name: aap-operator.v2.4.0-0.1737661496
        version: 2.4.0+0.1737661496
      - name: aap-operator.v2.4.0-0.1741194076
        version: 2.4.0+0.1741194076
      - name: aap-operator.v2.4.0-0.1742444794
        version: 2.4.0+0.1742444794
      - name: aap-operator.v2.4.0-0.1743629049
        version: 2.4.0+0.1743629049
      - name: aap-operator.v2.4.0-0.1746132523
        version: 2.4.0+0.1746132523
      - name: aap-operator.v2.4.0-0.1749069319
        version: 2.4.0+0.1749069319
      - name: aap-operator.v2.4.0-0.1750877268
        version: 2.4.0+0.1750877268
      - name: aap-operator.v2.4.0-0.1753232084
        version: 2.4.0+0.1753232084
      - name: aap-operator.v2.4.0-0.1755833431
        version: 2.4.0+0.1755833431
      - name: aap-operator.v2.4.0-0.1758079821
        version: 2.4.0+0.1758079821
      name: stable-2.4
    - currentCSV: aap-operator.v2.4.0-0.1758080293
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.1-0.1635283332 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the hub, including all hosted Ansible
              content, secrets and the database. By default, a persistent volume claim
              will be created using the default StorageClass on your cluster to store
              the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous hub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Automation Hub
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-24

          ## How to Upgrade

          Starting with AAP 2.4 in the May 2025 async release, the managed database was upgraded to PostgreSQL 15. It is recommended to take a backup of your deployment before upgrading.

          To take a backup, create AutomationControllerBackup and AutomationHubBackup objects for your deployments.
          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel.

          Note: Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8@sha256:bfd13d6b39c5ec3a67adbafb05a0482328f0b9b97a517b45c093f6c35a3a92d3
        - registry.redhat.io/ansible-automation-platform-24/ee-minimal-rhel8@sha256:7670e41037f023e3fc6ee95d2342787a80077273ab8d30ba6804029ff05b4edc
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8@sha256:04fb5cb6b0d7c79cf5d4b754fe917224ddfcb2b9a4994462db404c833ed69483
        - registry.redhat.io/ansible-automation-platform-24/ee-supported-rhel8@sha256:f1384eb558f7da1e4a35d51408b3cf5a5a9be08f4d50503f473b26909bc246fc
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-rhel8-operator@sha256:b120e6a5ef376662b4324f0458eae31627087dc91fc613788143bc3954a15318
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8-operator@sha256:5253c9fd587ff2ca94d251f2b5fc7f284080adc5eb64643df5299beb2015592f
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8-operator@sha256:3c28a1326ecb7495952a5bb8be13b0972a4378af82611b0b9d990c5fa9fa82c3
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8@sha256:7ae8db2b5d12b619f40db1e581db1e982a695bc1a29a2628222fff09e7c25249
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8@sha256:cccb4db08735551c697a7b2cbf75812f34500e691f9cdbaea02b30edea292e47
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8-operator@sha256:5b957031609c8aabf7575e675bc636c1152babc9564d503802301aff4642a404
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8-operator@sha256:e20106a8a04abd568a4c56861cea39ae444c0bd204bc807ac01f1705e12937d3
        - registry.redhat.io/ansible-automation-platform-24/hub-web-rhel8@sha256:6ff1f602f0d9cad201b3addb65acc570a9d527de005ca85fc861d3cb98138710
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-runner-rhel8@sha256:e14286dc9605be3a8c6cb144d0b2e3f3bb3211129005009e7566ea6c16c91f52
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-ui-rhel8@sha256:d229d4e09b2b788d1ba3413ea7ec15374bd1613d507c285ac1ddc10570b9473f
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        version: 2.4.0+0.1758080293
      entries:
      - name: aap-operator.v2.4.0-0.1702352614
        version: 2.4.0+0.1702352614
      - name: aap-operator.v2.4.0-0.1707507304
        version: 2.4.0+0.1707507304
      - name: aap-operator.v2.4.0-0.1708975279
        version: 2.4.0+0.1708975279
      - name: aap-operator.v2.4.0-0.1711591330
        version: 2.4.0+0.1711591330
      - name: aap-operator.v2.4.0-0.1713467078
        version: 2.4.0+0.1713467078
      - name: aap-operator.v2.4.0-0.1715899379
        version: 2.4.0+0.1715899379
      - name: aap-operator.v2.4.0-0.1717558885
        version: 2.4.0+0.1717558885
      - name: aap-operator.v2.4.0-0.1718152680
        version: 2.4.0+0.1718152680
      - name: aap-operator.v2.4.0-0.1720482740
        version: 2.4.0+0.1720482740
      - name: aap-operator.v2.4.0-0.1725257213
        version: 2.4.0+0.1725257213
      - name: aap-operator.v2.4.0-0.1726111804
        version: 2.4.0+0.1726111804
      - name: aap-operator.v2.4.0-0.1727259094
        version: 2.4.0+0.1727259094
      - name: aap-operator.v2.4.0-0.1730153340
        version: 2.4.0+0.1730153340
      - name: aap-operator.v2.4.0-0.1733186325
        version: 2.4.0+0.1733186325
      - name: aap-operator.v2.4.0-0.1733945743
        version: 2.4.0+0.1733945743
      - name: aap-operator.v2.4.0-0.1737662040
        version: 2.4.0+0.1737662040
      - name: aap-operator.v2.4.0-0.1741194557
        version: 2.4.0+0.1741194557
      - name: aap-operator.v2.4.0-0.1742445409
        version: 2.4.0+0.1742445409
      - name: aap-operator.v2.4.0-0.1743630487
        version: 2.4.0+0.1743630487
      - name: aap-operator.v2.4.0-0.1746133932
        version: 2.4.0+0.1746133932
      - name: aap-operator.v2.4.0-0.1749069693
        version: 2.4.0+0.1749069693
      - name: aap-operator.v2.4.0-0.1750878254
        version: 2.4.0+0.1750878254
      - name: aap-operator.v2.4.0-0.1753232791
        version: 2.4.0+0.1753232791
      - name: aap-operator.v2.4.0-0.1755833968
        version: 2.4.0+0.1755833968
      - name: aap-operator.v2.4.0-0.1758080293
        version: 2.4.0+0.1758080293
      name: stable-2.4-cluster-scoped
    - currentCSV: aap-operator.v2.5.0-0.1758147230
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727257740 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation gateway - Unify the management of AAP from a central location.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner


          ## New in AAP 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-25

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.5

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console. For example, change the channel from `stable-2.4` to `stable-2.5`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.

          After changing to the new 2.5 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place.

          #### Create the AnsibleAutomationPlatform custom resource

          Once that is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: myaap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8@sha256:3a989f148648a5a62e687058ec3ca772fa22aa10faec89453626d2b17b3fd9c2
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8-operator@sha256:43e731b9a435b3a412108170b4b39e6b6215be51d362a9f03c33fa412e97da53
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8-operator@sha256:8d8182e3a87c55eacc222e71df4a9bf7a2e51102d913a5824cc91e36056d4249
        - registry.redhat.io/ansible-automation-platform-25/hub-web-rhel8@sha256:e96832efcc03c6754031cb70f56c808135f62d1b80a882dd6d6fc9e1a4cfc69b
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8@sha256:697343aaea153ab22d7efa8f1bb5aed37a55cbb353fc206f4a355e8b878ae418
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8@sha256:ac0de2b7d354eb0c15a6c12861a4de3577a982c84af661fbab25517d8105c085
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8@sha256:aed3973eb953dda3571260dcf1b33bc7dcbc18c41e3350b8c361b7e635fb9527
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-runner-rhel8@sha256:ace1d2ab49ba5ff9f11ab18e8edce27c9e73ae7298bf7335569fa1740a44899b
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8-operator@sha256:acc0fcd509eeac46db307ab744dd0a0c26770a93e495cf2ec0ccb61a91c6fbc5
        - registry.redhat.io/ansible-automation-platform-25/ee-minimal-rhel8@sha256:41f944eb1450db56c3b14fc7baece8f148d5f10886a57bad730bec60fb28ac04
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8-operator@sha256:07aa7e257d04ae0eda6fcaa4af9121f1aece439231fce4f924e6467da91f0022
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-ui-rhel8@sha256:9ab35e80a6172018ed3f5cd179e0bcb6befc1378583d9e73c3a64a7cf9e8050f
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8@sha256:9f2867603b9747c2fe37862c2c5efac4a4530902f323a8bf7cc0717de4cdb5aa
        - registry.redhat.io/ansible-automation-platform-25/gateway-proxy-rhel8@sha256:789d55f355fa0d90f9509a912d199123bfc5a4041930a72a8fd4ed1d819f0d84
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8-operator@sha256:503fc7a20e2dd7de2cb55b9be5f97efa0d2b31f054ce326f23f3a11b4d683e64
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-rhel8-operator@sha256:74be9182723b41f87b3d13b3d86b6a9049bba88e2902cc9453264aa8a0f584e4
        - registry.redhat.io/ansible-automation-platform-25/ee-supported-rhel8@sha256:52d2c73fa2015af5f1b9a88ee48085c7c7bab44330fc9f95914f48b58e4b48bc
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-chatbot-rhel8@sha256:9a37c9739c488ccf211ec5efc6debb045223d508c9cb3fa5a8256eca98ff2f0c
        version: 2.5.0+0.1758147230
      entries:
      - name: aap-operator.v2.5.0-0.1727240614
        version: 2.5.0+0.1727240614
      - name: aap-operator.v2.5.0-0.1727875185
        version: 2.5.0+0.1727875185
      - name: aap-operator.v2.5.0-0.1728520175
        version: 2.5.0+0.1728520175
      - name: aap-operator.v2.5.0-0.1729741547
        version: 2.5.0+0.1729741547
      - name: aap-operator.v2.5.0-0.1731456196
        version: 2.5.0+0.1731456196
      - name: aap-operator.v2.5.0-0.1733193164
        version: 2.5.0+0.1733193164
      - name: aap-operator.v2.5.0-0.1733969526
        version: 2.5.0+0.1733969526
      - name: aap-operator.v2.5.0-0.1736893446
        version: 2.5.0+0.1736893446
      - name: aap-operator.v2.5.0-0.1737101025
        version: 2.5.0+0.1737101025
      - name: aap-operator.v2.5.0-0.1737675968
        version: 2.5.0+0.1737675968
      - name: aap-operator.v2.5.0-0.1739248042
        version: 2.5.0+0.1739248042
      - name: aap-operator.v2.5.0-0.1740093573
        version: 2.5.0+0.1740093573
      - name: aap-operator.v2.5.0-0.1740773472
        version: 2.5.0+0.1740773472
      - name: aap-operator.v2.5.0-0.1741368605
        version: 2.5.0+0.1741368605
      - name: aap-operator.v2.5.0-0.1742434024
        version: 2.5.0+0.1742434024
      - name: aap-operator.v2.5.0-0.1743660124
        version: 2.5.0+0.1743660124
      - name: aap-operator.v2.5.0-0.1746137767
        version: 2.5.0+0.1746137767
      - name: aap-operator.v2.5.0-0.1747343762
        version: 2.5.0+0.1747343762
      - name: aap-operator.v2.5.0-0.1749074128
        version: 2.5.0+0.1749074128
      - name: aap-operator.v2.5.0-0.1749604727
        version: 2.5.0+0.1749604727
      - name: aap-operator.v2.5.0-0.1750901111
        version: 2.5.0+0.1750901111
      - name: aap-operator.v2.5.0-0.1753900156
        version: 2.5.0+0.1753900156
      - name: aap-operator.v2.5.0-0.1755835086
        version: 2.5.0+0.1755835086
      - name: aap-operator.v2.5.0-0.1758147230
        version: 2.5.0+0.1758147230
      name: stable-2.5
    - currentCSV: aap-operator.v2.5.0-0.1758147817
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727259094 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation gateway - Unify the management of AAP from a central location.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner


          ## New in AAP 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-25

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.5

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console. For example, change the channel from `stable-2.4` to `stable-2.5`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.

          After changing to the new 2.5 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place.

          #### Create the AnsibleAutomationPlatform custom resource

          Once that is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: myaap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8-operator@sha256:acc0fcd509eeac46db307ab744dd0a0c26770a93e495cf2ec0ccb61a91c6fbc5
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-ui-rhel8@sha256:9ab35e80a6172018ed3f5cd179e0bcb6befc1378583d9e73c3a64a7cf9e8050f
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8-operator@sha256:43e731b9a435b3a412108170b4b39e6b6215be51d362a9f03c33fa412e97da53
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8-operator@sha256:8d8182e3a87c55eacc222e71df4a9bf7a2e51102d913a5824cc91e36056d4249
        - registry.redhat.io/ansible-automation-platform-25/hub-web-rhel8@sha256:e96832efcc03c6754031cb70f56c808135f62d1b80a882dd6d6fc9e1a4cfc69b
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-chatbot-rhel8@sha256:9a37c9739c488ccf211ec5efc6debb045223d508c9cb3fa5a8256eca98ff2f0c
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8@sha256:3a989f148648a5a62e687058ec3ca772fa22aa10faec89453626d2b17b3fd9c2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/ee-minimal-rhel8@sha256:41f944eb1450db56c3b14fc7baece8f148d5f10886a57bad730bec60fb28ac04
        - registry.redhat.io/ansible-automation-platform-25/gateway-proxy-rhel8@sha256:789d55f355fa0d90f9509a912d199123bfc5a4041930a72a8fd4ed1d819f0d84
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8@sha256:aed3973eb953dda3571260dcf1b33bc7dcbc18c41e3350b8c361b7e635fb9527
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8@sha256:697343aaea153ab22d7efa8f1bb5aed37a55cbb353fc206f4a355e8b878ae418
        - registry.redhat.io/ansible-automation-platform-25/ee-supported-rhel8@sha256:52d2c73fa2015af5f1b9a88ee48085c7c7bab44330fc9f95914f48b58e4b48bc
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8@sha256:ac0de2b7d354eb0c15a6c12861a4de3577a982c84af661fbab25517d8105c085
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-rhel8-operator@sha256:74be9182723b41f87b3d13b3d86b6a9049bba88e2902cc9453264aa8a0f584e4
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-runner-rhel8@sha256:ace1d2ab49ba5ff9f11ab18e8edce27c9e73ae7298bf7335569fa1740a44899b
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8@sha256:9f2867603b9747c2fe37862c2c5efac4a4530902f323a8bf7cc0717de4cdb5aa
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8-operator@sha256:07aa7e257d04ae0eda6fcaa4af9121f1aece439231fce4f924e6467da91f0022
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8-operator@sha256:503fc7a20e2dd7de2cb55b9be5f97efa0d2b31f054ce326f23f3a11b4d683e64
        version: 2.5.0+0.1758147817
      entries:
      - name: aap-operator.v2.5.0-0.1727241223
        version: 2.5.0+0.1727241223
      - name: aap-operator.v2.5.0-0.1727875896
        version: 2.5.0+0.1727875896
      - name: aap-operator.v2.5.0-0.1728520687
        version: 2.5.0+0.1728520687
      - name: aap-operator.v2.5.0-0.1729742145
        version: 2.5.0+0.1729742145
      - name: aap-operator.v2.5.0-0.1731456854
        version: 2.5.0+0.1731456854
      - name: aap-operator.v2.5.0-0.1733193761
        version: 2.5.0+0.1733193761
      - name: aap-operator.v2.5.0-0.1733970210
        version: 2.5.0+0.1733970210
      - name: aap-operator.v2.5.0-0.1736894220
        version: 2.5.0+0.1736894220
      - name: aap-operator.v2.5.0-0.1737101781
        version: 2.5.0+0.1737101781
      - name: aap-operator.v2.5.0-0.1737676792
        version: 2.5.0+0.1737676792
      - name: aap-operator.v2.5.0-0.1739248649
        version: 2.5.0+0.1739248649
      - name: aap-operator.v2.5.0-0.1740094176
        version: 2.5.0+0.1740094176
      - name: aap-operator.v2.5.0-0.1740774104
        version: 2.5.0+0.1740774104
      - name: aap-operator.v2.5.0-0.1741369251
        version: 2.5.0+0.1741369251
      - name: aap-operator.v2.5.0-0.1742434756
        version: 2.5.0+0.1742434756
      - name: aap-operator.v2.5.0-0.1743660958
        version: 2.5.0+0.1743660958
      - name: aap-operator.v2.5.0-0.1746138413
        version: 2.5.0+0.1746138413
      - name: aap-operator.v2.5.0-0.1747345055
        version: 2.5.0+0.1747345055
      - name: aap-operator.v2.5.0-0.1749074612
        version: 2.5.0+0.1749074612
      - name: aap-operator.v2.5.0-0.1749607543
        version: 2.5.0+0.1749607543
      - name: aap-operator.v2.5.0-0.1750901870
        version: 2.5.0+0.1750901870
      - name: aap-operator.v2.5.0-0.1753900674
        version: 2.5.0+0.1753900674
      - name: aap-operator.v2.5.0-0.1755835623
        version: 2.5.0+0.1755835623
      - name: aap-operator.v2.5.0-0.1758147817
        version: 2.5.0+0.1758147817
      name: stable-2.5-cluster-scoped
    - currentCSV: aap-operator.v2.6.0-0.1758684783
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727257740 <2.6.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Automation Gateway - Unify the management of AAP from a central location.
            * Automation Controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private Automation Hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Ansible Lightspeed: Deploy and manage the lifecycle of an Ansible Lightspeed application.
            * Platform Resource Operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## New in AAP 2.6 and 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.


          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-26

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5 and above, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.6

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console.

          **For customers on AAP 2.4:** Change the channel directly from `stable-2.4` to `stable-2.6` to upgrade directly to AAP 2.6, which will be released in October.

          **For customers on AAP 2.5:** Change the channel from `stable-2.5` to `stable-2.6`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.

          After changing to the new 2.6 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place. When upgrading from AAP 2.4 to 2.6, your external authentication configuration in Automation Controller will automatically be migrated to the new unified Platform UI.

          **Database:** AAP 2.6 continues to use PostgreSQL 15.

          #### Create the AnsibleAutomationPlatform custom resource (Required for upgrades from 2.4 only)

          **If upgrading from AAP 2.4:** Once the operator upgrade is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created.

          **If upgrading from AAP 2.5:** This step is not required as your AnsibleAutomationPlatform custom resource already exists.

          Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          Next, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda (lightspeed is disabled by default):


          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: aap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-26/gateway-rhel9-operator@sha256:fa1b102a0690767a1e08935a0e164c1eabb2d4cf9afccd310262ee9185544de7
        - registry.redhat.io/ansible-automation-platform-26/hub-rhel9@sha256:8041b141cd8921aaaa20cce15eab4c6189da9f449d99057189284891c071ac3b
        - registry.redhat.io/ansible-automation-platform-26/gateway-proxy-rhel9@sha256:13284d20c6fb4f2de24ec1715e7abdb5c4b0e4298c5034a7795f494b93634a02
        - registry.redhat.io/rhel9/postgresql-15@sha256:5e225c1646251d2509b3a6e45fad23abfb11dd5c45d3224b59178f90deb38a70
        - registry.redhat.io/ansible-automation-platform-26/controller-rhel9-operator@sha256:863e77e52ab01cec2a71672a224080a59d963d4fc08b9ed6f3cc7950729b1746
        - registry.redhat.io/ansible-automation-platform-26/hub-rhel9-operator@sha256:18114aab6db0d00780fdeaf9efa74ee44cab94824c2fd35f4b9998b6378ad8b3
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-rhel9-operator@sha256:cfbb98e17297134c187673f355342ac5ad60a56d73f4b0c918be2ff244286cc0
        - registry.redhat.io/ansible-automation-platform-26/platform-resource-rhel9-operator@sha256:969bbd033fbba33624329ee9347b3b89fb5a1f64e71892f5c6c7d74ed14a6d39
        - registry.redhat.io/ansible-automation-platform-26/controller-rhel9@sha256:64f5e77a7842965e9c6f7bc7990354b71bb19ffb06ab8ef7793dde956cb6c0f0
        - registry.redhat.io/ansible-automation-platform-26/ee-minimal-rhel9@sha256:261c324cac081e47d559c5fccbd48e7ec8f5bf64e3f27287fbe4fb8aff6b7d99
        - registry.redhat.io/ansible-automation-platform-26/hub-web-rhel9@sha256:aa4041672542a7f828d8a51c79e9c4dcea2a185388e4f20f828104db37e03a9b
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-rhel9@sha256:01e56bee772d15b120d6e4b02a59024b21eb8b266127b1a9b7f2c535ecdb2855
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-ui-rhel9@sha256:5b5f38a5f76781a042eec9cbf70b8be026878e4fa61c7fb90d1b0ad64bd21592
        - registry.redhat.io/ubi9/ubi-minimal@sha256:7c5495d5fad59aaee12abc3cbbd2b283818ee1e814b00dbc7f25bf2d14fa4f0c
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-chatbot-rhel9@sha256:8901fba14b21569b02108d2ab9050d558dee29f8177fae9d1bfcd5de3f4ce7c1
        - registry.redhat.io/ansible-automation-platform-26/mcp-tools-rhel9@sha256:0c1f6f6958f2570c8cd073287555d28c70ef9dded702ec014d3ead8ce572463e
        - registry.redhat.io/ansible-automation-platform-26/platform-resource-runner-rhel9@sha256:9983043386df38a17ff6fffef64dc187691e0b1fefec01eef3af39aa08e532ef
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-rhel9-operator@sha256:bfd55bffd8239e224d4ebba500d8517e8e66915078d6b59244f5e910fc5e10c8
        - registry.redhat.io/ansible-automation-platform-26/gateway-rhel9@sha256:c85dbbf6beaa264afec480dc1f06576aada531cdfd15f247f4db37dc9702be33
        - registry.redhat.io/rhel9/redis-6@sha256:0962d18bb5292c0e524354f319b63287c2298a4a2fd34556c68fa285df381427
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-rhel9@sha256:c5d4e2e76944e0991729e5c3b99f43b7f29ea08ad112e5e6513feb8fb1d899a1
        - registry.redhat.io/ansible-automation-platform-26/ee-supported-rhel9@sha256:2534d07f1ae41af32adb681e283cbaea465cbe0f9d234ed5478a7bfad7676e7f
        version: 2.6.0+0.1758684783
      entries:
      - name: aap-operator.v2.6.0-0.1758684783
        version: 2.6.0+0.1758684783
      name: stable-2.6
    - currentCSV: aap-operator.v2.6.0-0.1758685241
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727259094 <2.6.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Automation Gateway - Unify the management of AAP from a central location.
            * Automation Controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private Automation Hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Ansible Lightspeed: Deploy and manage the lifecycle of an Ansible Lightspeed application.
            * Platform Resource Operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## New in AAP 2.6 and 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.


          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-26

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5 and above, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.6

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console.

          **For customers on AAP 2.4:** Change the channel directly from `stable-2.4` to `stable-2.6` to upgrade directly to AAP 2.6, which will be released in October.

          **For customers on AAP 2.5:** Change the channel from `stable-2.5` to `stable-2.6`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.

          After changing to the new 2.6 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place. When upgrading from AAP 2.4 to 2.6, your external authentication configuration in Automation Controller will automatically be migrated to the new unified Platform UI.

          **Database:** AAP 2.6 continues to use PostgreSQL 15.

          #### Create the AnsibleAutomationPlatform custom resource (Required for upgrades from 2.4 only)

          **If upgrading from AAP 2.4:** Once the operator upgrade is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created.

          **If upgrading from AAP 2.5:** This step is not required as your AnsibleAutomationPlatform custom resource already exists.

          Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          Next, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda (lightspeed is disabled by default):


          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: aap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-26/gateway-rhel9@sha256:c85dbbf6beaa264afec480dc1f06576aada531cdfd15f247f4db37dc9702be33
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-ui-rhel9@sha256:5b5f38a5f76781a042eec9cbf70b8be026878e4fa61c7fb90d1b0ad64bd21592
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-rhel9-operator@sha256:cfbb98e17297134c187673f355342ac5ad60a56d73f4b0c918be2ff244286cc0
        - registry.redhat.io/ansible-automation-platform-26/controller-rhel9@sha256:64f5e77a7842965e9c6f7bc7990354b71bb19ffb06ab8ef7793dde956cb6c0f0
        - registry.redhat.io/ansible-automation-platform-26/ee-supported-rhel9@sha256:2534d07f1ae41af32adb681e283cbaea465cbe0f9d234ed5478a7bfad7676e7f
        - registry.redhat.io/ansible-automation-platform-26/ee-minimal-rhel9@sha256:261c324cac081e47d559c5fccbd48e7ec8f5bf64e3f27287fbe4fb8aff6b7d99
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-rhel9@sha256:c5d4e2e76944e0991729e5c3b99f43b7f29ea08ad112e5e6513feb8fb1d899a1
        - registry.redhat.io/ansible-automation-platform-26/hub-web-rhel9@sha256:aa4041672542a7f828d8a51c79e9c4dcea2a185388e4f20f828104db37e03a9b
        - registry.redhat.io/ansible-automation-platform-26/platform-resource-runner-rhel9@sha256:9983043386df38a17ff6fffef64dc187691e0b1fefec01eef3af39aa08e532ef
        - registry.redhat.io/ansible-automation-platform-26/gateway-rhel9-operator@sha256:fa1b102a0690767a1e08935a0e164c1eabb2d4cf9afccd310262ee9185544de7
        - registry.redhat.io/ansible-automation-platform-26/platform-resource-rhel9-operator@sha256:969bbd033fbba33624329ee9347b3b89fb5a1f64e71892f5c6c7d74ed14a6d39
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-chatbot-rhel9@sha256:8901fba14b21569b02108d2ab9050d558dee29f8177fae9d1bfcd5de3f4ce7c1
        - registry.redhat.io/ansible-automation-platform-26/mcp-tools-rhel9@sha256:0c1f6f6958f2570c8cd073287555d28c70ef9dded702ec014d3ead8ce572463e
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-rhel9-operator@sha256:bfd55bffd8239e224d4ebba500d8517e8e66915078d6b59244f5e910fc5e10c8
        - registry.redhat.io/ansible-automation-platform-26/hub-rhel9-operator@sha256:18114aab6db0d00780fdeaf9efa74ee44cab94824c2fd35f4b9998b6378ad8b3
        - registry.redhat.io/rhel9/postgresql-15@sha256:5e225c1646251d2509b3a6e45fad23abfb11dd5c45d3224b59178f90deb38a70
        - registry.redhat.io/ubi9/ubi-minimal@sha256:7c5495d5fad59aaee12abc3cbbd2b283818ee1e814b00dbc7f25bf2d14fa4f0c
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-rhel9@sha256:01e56bee772d15b120d6e4b02a59024b21eb8b266127b1a9b7f2c535ecdb2855
        - registry.redhat.io/ansible-automation-platform-26/gateway-proxy-rhel9@sha256:13284d20c6fb4f2de24ec1715e7abdb5c4b0e4298c5034a7795f494b93634a02
        - registry.redhat.io/ansible-automation-platform-26/hub-rhel9@sha256:8041b141cd8921aaaa20cce15eab4c6189da9f449d99057189284891c071ac3b
        - registry.redhat.io/rhel9/redis-6@sha256:0962d18bb5292c0e524354f319b63287c2298a4a2fd34556c68fa285df381427
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/ansible-automation-platform-26/controller-rhel9-operator@sha256:863e77e52ab01cec2a71672a224080a59d963d4fc08b9ed6f3cc7950729b1746
        version: 2.6.0+0.1758685241
      entries:
      - name: aap-operator.v2.6.0-0.1758685241
        version: 2.6.0+0.1758685241
      name: stable-2.6-cluster-scoped
    defaultChannel: stable-2.6-cluster-scoped
    packageName: ansible-automation-platform-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Instana
      provider-url: ""
    name: instana-agent-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: instana-agent-operator.v2.1.35
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instana.io/v1",
                "kind": "InstanaAgent",
                "metadata": {
                  "name": "instana-agent",
                  "namespace": "instana-agent"
                },
                "spec": {
                  "agent": {
                    "configuration_yaml": "# You can leave this empty, or use this to configure your instana agent.\n# See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.\n",
                    "endpointHost": "ingress-red-saas.instana.io",
                    "endpointPort": "443",
                    "env": {},
                    "key": "replace-key"
                  },
                  "cluster": {
                    "name": "my-cluster"
                  },
                  "opentelemetry": {
                    "grpc": {
                      "enabled": true
                    },
                    "http": {
                      "enabled": true
                    }
                  },
                  "zone": {
                    "name": "edited-zone"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring,Logging & Tracing,OpenShift Optional
          certified: "false"
          containerImage: icr.io/instana/instana-agent-operator@sha256:fe949ae0ca9cc05ffe3569cd7a6f4fde48f5e3ee64f27e20e0b3102733a6e809
          description: Fully automated Application Performance Monitoring (APM) for
            microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/instana/instana-agent-operator
          support: Instana
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InstanaAgent is the Schema for the agents API
            displayName: Instana Agent
            kind: InstanaAgent
            name: agents.instana.io
            version: v1
          - description: InstanaAgentRemote is the Schema for the agents API
            displayName: Instana Agent Remote
            kind: InstanaAgentRemote
            name: agentsremote.instana.io
            version: v1
        description: |
          # Instana

          Instana is an [APM solution](https://www.instana.com/product-overview/) built for microservices that enables IT Ops to build applications faster and deliver higher quality services by automating monitoring, tracing and root cause analysis. The solution is optimized for [Kubernetes](https://www.instana.com/automatic-kubernetes-monitoring/) and [OpenShift](https://www.instana.com/blog/automatic-root-cause-analysis-for-openshift-applications/).

          ## Instana Agent Operator

          This is the Kubernetes Operator for installing the Instana Agent on Kubernetes or OpenShift.

          ## Prerequisites for OpenShift

          Before the agent will be able to run in OpenShift, you need to perform a couple of extra configuration steps.

          You need to set up a project for the Instana Agent and configure it's permissions.

          The project you create here needs to be the namespace where you create the Instana Agent custom resource that the operator will use to deploy the agent.

          For example, create the `instana-agent` project:

              oc new-project instana-agent

          Then, ensure the `instana-agent` service account is in the privileged security context:

              oc adm policy add-scc-to-user privileged -z instana-agent

          The `instana-agent-remote` service account must be allowed to use any userid:

              oc adm policy add-scc-to-user anyuid -z instana-agent-remote -n instana-agent

          This service account will be created by the operator.

          Now you can proceed with installing the operator for the Instana agent.

          ## Installation and Configuration

          First, install this operator from [OperatorHub.io](https://operatorhub.io/), [OpenShift Container Platform](https://www.openshift.com/), or [OKD](https://www.okd.io/).

          Second, create a custom resource with the agent configuration in the target namespace (for now, this MUST always be the `instana-agent` namespace). The operator will pick up the custom resource and install the Instana agent accordingly.

          The following is a minimal template of the custom resource:

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgent
          metadata:
            name: instana-agent
            namespace: instana-agent
          spec:
            zone:
              name: my-zone # (optional) name of the zone of the host
            cluster:
              name: replace-me # replace with the name of your Kubernetes cluster
            agent:
              key: replace-me # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io # the monitoring ingress endpoint
              endpointPort: "443" # the monitoring ingress endpoint port, wrapped in quotes
              env:
                INSTANA_AGENT_TAGS: example
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://ibm.biz/monitoring-k8s
          ```

          Save the template in a file `instana-agent.yaml` and edit the following values:

          * If your target namespace is not `instana-agent`, replace the `namespace:` accordingly.
          * `agent.key` must be set with your Instana agent key.
          * `agent.endpointHost` must be set with the monitoring ingress endpoint, generally either `saas-us-west-2.instana.io` or `saas-eu-west-1.instana.io`.
          * `agent.endpointPort` must be set with the monitoring ingress port, generally "443" (wrapped in quotes).
          * `zone.name` should be set with the name of the Kubernetes cluster that is be displayed in Instana.

          For advanced configuration, you can edit the contents of the `configuration.yaml` file. View documentation [here](https://ibm.biz/monitoring-k8s).

          Apply the custom resource with `kubectl apply -f instana-agent.yaml`. After some time, you should see `instana-agent` Pods being created on each node of your cluster, and your cluster should show on the infrastructure map on your Instana Web interface.

          In addition to the InstanaAgent custom resource, a remote agent can be deployed. This agent will not monitor workload running inside the cluster, but can deploy a single replica of an agent which can get an independent configuration for running remote monitoring for certain sensors.

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgentRemote
          metadata:
            name: remote-agent
            namespace: "instana-agent"
          spec:
            zone:
              name: edited-zone # name of the zone of the host
            hostname:
              name: use-custom-hostname-for-remote-agent # (optional) give your remote agent pod a custom hostname
            agent:
              key: replace-key # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io
              endpointPort: "443"
              env: {}
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.
          ```

          ## Uninstalling

          In order to uninstall the Instana agent, simply remove the custom resource with `kubectl delete -f instana-agent.yaml`.

          ## Source Code

          The Instana agent operator is an open source project hosted on [https://github.com/instana/instana-agent-operator](https://github.com/instana/instana-agent-operator/).
        displayName: Instana Agent Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - APM
        - Instana
        links:
        - name: GitHub Repository
          url: https://github.com/instana/instana-agent-operator
        maintainers:
        - email: support@instana.com
          name: Instana
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: Instana
        relatedImages:
        - icr.io/instana/instana-agent-operator@sha256:fe949ae0ca9cc05ffe3569cd7a6f4fde48f5e3ee64f27e20e0b3102733a6e809
        - icr.io/instana/agent@sha256:ed71e6c3e5bdb155853a2db8c916460e6a867fb02627e3a332186188cbbd98dd
        version: 2.1.35
      entries:
      - name: instana-agent-operator.v2.1.35
        version: 2.1.35
      - name: instana-agent-operator.v2.1.34
        version: 2.1.34
      - name: instana-agent-operator.v2.1.33
        version: 2.1.33
      - name: instana-agent-operator.v2.1.32
        version: 2.1.32
      - name: instana-agent-operator.v2.1.31
        version: 2.1.31
      - name: instana-agent-operator.v2.1.30
        version: 2.1.30
      - name: instana-agent-operator.v2.1.29
        version: 2.1.29
      - name: instana-agent-operator.v2.1.28
        version: 2.1.28
      - name: instana-agent-operator.v2.1.27
        version: 2.1.27
      - name: instana-agent-operator.v2.1.25
        version: 2.1.25
      - name: instana-agent-operator.v2.1.24
        version: 2.1.24
      - name: instana-agent-operator.v2.1.23
        version: 2.1.23
      - name: instana-agent-operator.v2.1.22
        version: 2.1.22
      - name: instana-agent-operator.v2.1.21
        version: 2.1.21
      - name: instana-agent-operator.v2.1.20
        version: 2.1.20
      - name: instana-agent-operator.v2.1.19
        version: 2.1.19
      - name: instana-agent-operator.v2.1.18
        version: 2.1.18
      - name: instana-agent-operator.v2.1.17
        version: 2.1.17
      - name: instana-agent-operator.v2.1.16
        version: 2.1.16
      - name: instana-agent-operator.v2.1.15
        version: 2.1.15
      - name: instana-agent-operator.v2.1.14
        version: 2.1.14
      - name: instana-agent-operator.v2.1.13
        version: 2.1.13
      - name: instana-agent-operator.v2.1.12
        version: 2.1.12
      - name: instana-agent-operator.v2.1.11
        version: 2.1.11
      - name: instana-agent-operator.v2.1.10
        version: 2.1.10
      - name: instana-agent-operator.v2.1.9
        version: 2.1.9
      - name: instana-agent-operator.v2.1.8
        version: 2.1.8
      - name: instana-agent-operator.v2.1.7
        version: 2.1.7
      - name: instana-agent-operator.v2.1.6
        version: 2.1.6
      - name: instana-agent-operator.v2.1.5
        version: 2.1.5
      - name: instana-agent-operator.v2.1.4
        version: 2.1.4
      - name: instana-agent-operator.v2.1.3
        version: 2.1.3
      - name: instana-agent-operator.v2.1.2
        version: 2.1.2
      - name: instana-agent-operator.v2.1.1
        version: 2.1.1
      - name: instana-agent-operator.v2.1.0
        version: 2.1.0
      - name: instana-agent-operator.v2.0.32
        version: 2.0.32
      - name: instana-agent-operator.v2.0.31
        version: 2.0.31
      - name: instana-agent-operator.v2.0.30
        version: 2.0.30
      - name: instana-agent-operator.v2.0.27
        version: 2.0.27
      - name: instana-agent-operator.v2.0.19
        version: 2.0.19
      - name: instana-agent-operator.v2.0.18
        version: 2.0.18
      - name: instana-agent-operator.v2.0.17
        version: 2.0.17
      - name: instana-agent-operator.v2.0.13
        version: 2.0.13
      - name: instana-agent-operator.v2.0.12
        version: 2.0.12
      - name: instana-agent-operator.v2.0.11
        version: 2.0.11
      - name: instana-agent-operator.v2.0.10
        version: 2.0.10
      - name: instana-agent-operator.v2.0.9
        version: 2.0.9
      - name: instana-agent-operator.v2.0.8
        version: 2.0.8
      - name: instana-agent-operator.v2.0.7
        version: 2.0.7
      - name: instana-agent-operator.v2.0.5
        version: 2.0.5
      - name: instana-agent-operator.v2.0.4
        version: 2.0.4
      name: stable
    defaultChannel: stable
    packageName: instana-agent-operator
    provider:
      name: Instana
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ses-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ses-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ses-controller:1.1.1
          createdAt: "2025-09-19T19:37:25Z"
          description: AWS SES controller is a service controller for managing SES
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConfigurationSet represents the state of an AWS ses ConfigurationSet
              resource.
            displayName: ConfigurationSet
            kind: ConfigurationSet
            name: configurationsets.ses.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Simple Email Service (SES) resources in AWS from within your Kubernetes cluster.

          **About Amazon SES**

          Amazon Simple Email Service (SES) is an email platform that provides an easy, cost-effective way for you to send and receive email using your own email addresses and domains.
          For example, you can send marketing emails such as special offers, transactional emails such as order confirmations, and other types of correspondence such as newsletters. When you use Amazon SES to receive mail, you can develop software solutions such as email autoresponders, email unsubscribe systems, and applications that generate customer support tickets from incoming emails.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SES
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ses
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SES Developer Resources
          url: https://aws.amazon.com/ses/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ses maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ses-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ses-controller.v1.1.1
        version: 1.1.1
      - name: ack-ses-controller.v1.1.0
        version: 1.1.0
      - name: ack-ses-controller.v1.0.13
        version: 1.0.13
      - name: ack-ses-controller.v1.0.12
        version: 1.0.12
      - name: ack-ses-controller.v1.0.11
        version: 1.0.11
      - name: ack-ses-controller.v1.0.10
        version: 1.0.10
      - name: ack-ses-controller.v1.0.9
        version: 1.0.9
      - name: ack-ses-controller.v1.0.8
        version: 1.0.8
      - name: ack-ses-controller.v1.0.7
        version: 1.0.7
      - name: ack-ses-controller.v1.0.6
        version: 1.0.6
      - name: ack-ses-controller.v1.0.5
        version: 1.0.5
      - name: ack-ses-controller.v1.0.3
        version: 1.0.3
      - name: ack-ses-controller.v1.0.2
        version: 1.0.2
      - name: ack-ses-controller.v1.0.1
        version: 1.0.1
      - name: ack-ses-controller.v1.0.0
        version: 1.0.0
      - name: ack-ses-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-ses-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco Networking - Compute Solutions
      provider-url: ""
    name: ucs-ci-solutions-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ucs-ci-solutions-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intersight.cisco.com/v1alpha1",
                "kind": "UCSCISolutions",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ucs-ci-solutions-operator",
                    "app.kubernetes.io/instance": "ucscisolutions-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ucscisolutions",
                    "app.kubernetes.io/part-of": "ucs-ci-solutions-operator"
                  },
                  "name": "ucscisolutions-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: docker.io/pkoppa/ucs-ci-solutions-operator:latest
          createdAt: "2023-07-15T15:19:39Z"
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: UCSCISolutions
            name: ucscisolutions.intersight.cisco.com
            version: v1alpha1
        description: The operator is used to automate the creation of  Pools, Policies
          and Server Profile Templates required to deploy Cisco UCS Converged Infrastructure
          Solution in Intersight.
        displayName: UCS-CI-Solutions-Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ucscisolutions
        - intersight
        - ucs
        - ci
        - flexpod
        - flashstack
        links:
        - name: Ucs Ci Solutions Operator
          url: https://ucs-ci-solutions-operator.domain
        maintainers:
        - email: pkoppa@cisco.com
          name: Paniraj Koppa
        - email: chrobrie@cisco.com
          name: Chris O'Brien
        maturity: alpha
        provider:
          name: Cisco Networking - Compute Solutions
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - docker.io/pkoppa/ucs-ci-solutions-operator:latest
        version: 0.0.1
      entries:
      - name: ucs-ci-solutions-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ucs-ci-solutions-operator
    provider:
      name: Cisco Networking - Compute Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-diagnostics
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-diagnostics.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Diagnostics",
                "metadata": {
                  "name": "diagnostics-sample"
                },
                "spec": {
                  "agentDiagnosticsService": {
                    "port": 8080
                  },
                  "auth": {
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "wapdbDiagnosticsService": {
                    "dumpedTables": "[ \"--regexpr cluster_logs_week_*\", \"current_week_cluster_logs\", \"config_metadata_dbs\", \"config_metadata_dbs_id_seq\", \"config_metadata_tags\", \"config_tags\", \"csv_reports_meta\", \"custom_import_handler_dbs\", \"data_point_db_models\", \"db_test_entries\", \"license_server_metadata\", \"operations\", \"result_file_metadata_dbs\", \"result_metadata_dbs\", \"result_metadata_tags\", \"result_tags\", \"session_meta\", \"session_types\", \"sessions\", \"stat_view_query_data\", \"test_run_configs\", \"test_run_infos\", \"test_runs\", \"wireless_globalsettings\", \"wireless_sessionconfigs\", \"wireless_templates\"]",
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:37:10Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Diagnostics
            name: diagnostics.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Diagnostics Operator
        displayName: Keysight WAP Diagnostics Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Diagnostics
          url: https://keysight-wap-diagnostics.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-diagnostics@sha256:97bf8affa7bf1f3a391f5f036313e7cd8196be2eb8e95debeab06c1b6c964c5e
        version: 0.1.21
      entries:
      - name: keysight-wap-diagnostics.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-diagnostics
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JFrog
      provider-url: ""
    name: openshiftartifactoryha-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: artifactory-ha-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftArtifactoryHa",
                "metadata": {
                  "name": "openshiftartifactoryha"
                },
                "spec": {
                  "artifactory-ha": {
                    "artifactory": {
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE",
                      "node": {
                        "replicaCount": 2,
                        "waitForPrimaryStartup": {
                          "enabled": false
                        }
                      }
                    },
                    "databaseUpgradeReady": true,
                    "database": {
                      "driver": "OVERRIDE",
                      "password": "OVERRIDE",
                      "type": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "nginx": {
                      "http": {
                        "externalPort": 80,
                        "internalPort": 8080
                      },
                      "https": {
                        "externalPort": 443,
                        "internalPort": 8443
                      },
                      "service": {
                        "ssloffload": false
                      },
                      "tlsSecretName": "OVERRIDE"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "waitForDatabase": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/artifactory-operator@sha256:de69bb16052d5ac6fed9882eb65d49446b6fa1cc5ba830dd2a39e882712b1347
          createdAt: 2024-03-23 00:00:00+00:00
          description: JFrog Artifactory Enterprise deploys Artifactory in a high
            availability environment across multiple pods
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Artifactory HA Instances
            displayName: Artifactory HA
            kind: OpenshiftArtifactoryHa
            name: openshiftartifactoryhas.charts.helm.k8s.io
            version: v1
        description: |2+
           ## Breaking change
          Please update to version 1.1.10 of the operator.
          Version 1.1.8 and 1.1.9 have issues. Please skip these versions.
          If you are using operator version < 1.2.0 there is a issue with upgrade where artifactory version is not updated if you update operator version. As workaround you can update the Custom Resource OpenshiftArtifactoryHa which has hardcored image value.After upgrading to version 1.2.0 you can remove the below parameters from your CR so that future upgrade works without workaround. ``` artifactory-ha.artifactory.image.registry artifactory-ha.artifactory.image.repository artifactory-ha.artifactory.image.tag ```
          ## Overview
          Openshift Operator to deploy JFrog Artifactory Enterprise into your Openshift cluster.
          ## Security Context Constraints
          To deploy this helm chart you will need to be a cluster admin w/ access to the anyuid scc and add the operator service account to the anyuid scc.
          ```
          oc adm policy add-scc-to-user anyuid -z artifactory-ha-operator -n openshift-operators
          ```
          Add the service account for the Artifactory chart to deploy successfully:
          ``` oc adm policy add-scc-to-user anyuid -z openshiftartifactoryha-artifactory-ha -n artifactory ```
          ### Usage

          An external DB is required. The operator will not deploy a DB but will require you to specify the configuration values to connect to it.

          Search for JFrog and click JFrog Artifactory Enterprise Operator to install.

          Go to the Installed Operators.

          Wait for the JFrog Artifactory Enterprise Operator to complete the installation.

          Open the Operator and click on the provided API: Artifactory HA.

          Click Create New Instance and provide the following parameters for your DB configuration:

          ```
          DATABASE_TYPE
          DATABASE_DRIVER
          DATABASE_URL
          DATABASE_USER
          DATABASE_PASSWORD
          ```
          Master key and Join key must be supplied. To generate a new key for each run the command below:

          ```
          # Create a key
          export JOIN_KEY=$(openssl rand -hex 32)
          echo ${JOIN_KEY}
          ```

          To use TLS you will need to first create a k8s tls secret to store your .crt and .key file into.
          Then supply the value of this k8s secret into the TLS_SECRET field.
          ``` oc create secret tls my_tls_secret --cert=tls.crt --key=tls.key --namespace=my_namespace ```
          Click Create for Artifactory Enterprise to deploy into OpenShift and connect to it on the external IP exposed by the load balancer.
          ### Create a route
          To expose Artifactory from Openshift we recommend you create a new route in Openshift.
          You can either use the oc command line tool or the Openshift web console to generate a new route.
          Depending upon where you terminate TLS you will need to either specify pass through or edge.
          Command Line (Edge):
          ``` oc create route edge --service=openshiftartifactory-ha --cert=tls.crt --key=tls.key --ca-cert=ca.crt --hostname=www.example.com ```
          Or for more information visit the official Openshift documentation on routes here:
          https://docs.openshift.com/container-platform/4.6/networking/routes/route-configuration.html

        displayName: JFrog Artifactory Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Artifact Repository
          url: https://jfrog.com/artifactory/
        - name: Artifactory Video
          url: https://www.youtube.com/watch?v=r2_A5CPo43U
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: alpha
        provider:
          name: JFrog
        relatedImages:
        - registry.redhat.io/rhel8/nginx-120@sha256:87f658744af695273c434fac8246b716728b691a17d5d0c5b89247362e61e45c
        - registry.connect.redhat.com/jfrog/artifactory-operator@sha256:de69bb16052d5ac6fed9882eb65d49446b6fa1cc5ba830dd2a39e882712b1347
        - registry.connect.redhat.com/jfrog/artifactory-pro@sha256:0986ca71419ae628892db11e7ca7da5ac1c60219c144ce9aeb2de6e683a45262
        version: 1.3.3
      entries:
      - name: artifactory-ha-operator.v1.3.3
        version: 1.3.3
      name: alpha
    - currentCSV: artifactory-ha-operator.v1.1.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftArtifactoryHa",
                "metadata": {
                  "name": "openshiftartifactoryha"
                },
                "spec": {
                  "artifactory-ha": {
                    "artifactory": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/artifactory-pro",
                        "tag": "7.27.3-1"
                      },
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE",
                      "uid": "1000721030",
                      "node": {
                        "replicaCount": 2,
                        "waitForPrimaryStartup": {
                          "enabled": false
                        }
                      }
                    },
                    "databaseUpgradeReady": true,
                    "database": {
                      "driver": "OVERRIDE",
                      "password": "OVERRIDE",
                      "type": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "initContainerImage": "registry.connect.redhat.com/jfrog/init@sha256:164b41720a37f75c6129c8a28d05cc56f821022313d44ff918cce4a2146d93e8",
                    "nginx": {
                      "uid": "1000720104",
                      "gid": "1000720107",
                      "http": {
                        "externalPort": 80,
                        "internalPort": 8080
                      },
                      "https": {
                        "externalPort": 443,
                        "internalPort": 8443
                      },
                      "image": {
                        "registry": "registry.redhat.io",
                        "repository": "rhel8/nginx-116",
                        "tag": "latest"
                      },
                      "service": {
                        "ssloffload": false
                      },
                      "tlsSecretName": "OVERRIDE"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "waitForDatabase": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/artifactory-operator@sha256:e292553e1b6e792ff19276c7f6565d7ed61dd95289eb176066329954bc93e371
          createdAt: 2020-03-25 00:00:00+00:00
          description: JFrog Artifactory Enterprise deploys Artifactory in a high
            availability environment across multiple pods
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Artifactory HA Instances
            displayName: Artifactory HA
            kind: OpenshiftArtifactoryHa
            name: openshiftartifactoryhas.charts.helm.k8s.io
            version: v1
        description: |2+
           ## Breaking change
          Please update to version 1.1.10 of the operator.
          Version 1.1.8 and 1.1.9 have issues. Please skip these versions.

          ## Overview
          Openshift Operator to deploy JFrog Artifactory Enterprise into your Openshift cluster.
          ## Security Context Constraints
          To deploy this helm chart you will need to be a cluster admin w/ access to the anyuid scc and add the operator service account to the anyuid scc.
          ```
          oc adm policy add-scc-to-user anyuid -z artifactory-ha-operator -n openshift-operators
          ```
          Add the service account for the Artifactory chart to deploy successfully:
          ``` oc adm policy add-scc-to-user anyuid -z openshiftartifactoryha-artifactory-ha -n artifactory ```
          ### Usage

          An external DB is required. The operator will not deploy a DB but will require you to specify the configuration values to connect to it.

          Search for JFrog and click JFrog Artifactory Enterprise Operator to install.

          Go to the Installed Operators.

          Wait for the JFrog Artifactory Enterprise Operator to complete the installation.

          Open the Operator and click on the provided API: Artifactory HA.

          Click Create New Instance and provide the following parameters for your DB configuration:

          ```
          DATABASE_TYPE
          DATABASE_DRIVER
          DATABASE_URL
          DATABASE_USER
          DATABASE_PASSWORD
          ```
          Master key and Join key must be supplied. To generate a new key for each run the command below:

          ```
          # Create a key
          export JOIN_KEY=$(openssl rand -hex 32)
          echo ${JOIN_KEY}
          ```

          To use TLS you will need to first create a k8s tls secret to store your .crt and .key file into.
          Then supply the value of this k8s secret into the TLS_SECRET field.
          ``` oc create secret tls my_tls_secret --cert=tls.crt --key=tls.key --namespace=my_namespace ```
          Click Create for Artifactory Enterprise to deploy into OpenShift and connect to it on the external IP exposed by the load balancer.
          ### Create a route
          To expose Artifactory from Openshift we recommend you create a new route in Openshift.
          You can either use the oc command line tool or the Openshift web console to generate a new route.
          Depending upon where you terminate TLS you will need to either specify pass through or edge.
          Command Line (Edge):
          ``` oc create route edge --service=openshiftartifactory-ha --cert=tls.crt --key=tls.key --ca-cert=ca.crt --hostname=www.example.com ```
          Or for more information visit the official Openshift documentation on routes here:
          https://docs.openshift.com/container-platform/4.6/networking/routes/route-configuration.html

        displayName: JFrog Artifactory Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Artifact Repository
          url: https://jfrog.com/artifactory/
        - name: Artifactory Video
          url: https://www.youtube.com/watch?v=r2_A5CPo43U
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: stable
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/init@sha256:164b41720a37f75c6129c8a28d05cc56f821022313d44ff918cce4a2146d93e8
        - registry.connect.redhat.com/jfrog/artifactory-operator@sha256:e292553e1b6e792ff19276c7f6565d7ed61dd95289eb176066329954bc93e371
        - registry.connect.redhat.com/jfrog/artifactory-pro@sha256:7289dde95d1e4d8e161a28cffd9c4a0cc96f92f7536d069ab63ec304a92254fd
        - registry.redhat.io/rhel8/nginx-116@sha256:d71d368d6094a5777000a8d99164959ccb97fca355e855e24bd92c13a8df04f6
        version: 1.1.18
      entries:
      - name: artifactory-ha-operator.v1.1.18
        version: 1.1.18
      name: stable
    defaultChannel: alpha
    packageName: openshiftartifactoryha-operator
    provider:
      name: JFrog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: advanced-cluster-management
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: advanced-cluster-management.v2.10.9
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-08-06T03:17:11Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0 <2.10.9'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/internal-objects: '[]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterHub defines the configuration for an instance
              of the MultiCluster Hub
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:95df18f4479b7f555a41508486624bde714778730921c20c665e9f23baadbabf
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:8f1bb3a8422d29960d8b90dd24a813a56875abf1f0393f78b862eaa48ddda783
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:50331cdd61332c8f4128b3dab237255463b307c5c40b0607730a7d3b33b9c6b5
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:5b162aeab4aa51185c880696d1472e152a20ada926a3baea0fb8494793ad238e
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:0ba2471b7a67532bdb90e02904671f0e724cdb39f74a45392ae6c693c25b354d
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:4b16da3070e30d879421aaa0d4a73aa2a1b0b852f62b8faabd801c03f22d5319
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:ced03229f208739867a43b9ed9bd158f19f4242dd1f5f045d3e5db81b2b1d171
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:4ad5ccf9ce682024b6e3025a9bcadd6f32692aa18d892afd670eacb7b4031860
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:1c2dbcf179508ea1c602f7817f85b0d3f58d6d85ae93ccc523e88a3936c9c690
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:23713ebb12d510d37de7e060d5ce4ff2198990f83cebbd2302e74bd83a55cecd
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:1aaba9ca65e363c7f9c07ef2f66f23cb8fff677204a56373634d319113d2c84b
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:e3c80e8e3af5c02c202dfb288c63b20c44c08f2a0474d2568cfb0b878f7fe888
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:9e78db48ff0b0fc0ac0347ca055a4b66e2f917073499adc6e2db2a9c16a70e01
        - registry.redhat.io/rhacm2/console-rhel9@sha256:1538a3a935aa01d8800d8145c45037eaab6987ad6e6e927017ca1cfea4bb111c
        - registry.redhat.io/rhel9/postgresql-13@sha256:31c77c19480ab82cb33525d32419b68cf59daccf0b9936e88e8238e48a2d47e5
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:fbc5b1456a37951c9f70f64357919222739f4a7e533fb6dde5b93672535fc87e
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:95890bb1aad544d97ae1b80e801a800c95ef51e1034b60e680b67bdb2f5aeb69
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:9e931cab5fcbf3df0ffe6fff8d12303800423ace3d749ee0e33b55300c816b19
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:dcd6a420467929b9af8c52c77c8717d22f3977e6b6b392b5dc90cbb04005b9f1
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:14094f09dfb166a6b342bd91e2db9162f0de88d55655ccce44d345047c700187
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:23b4d7fbc776c63dcd7007442a557124581939ce110257822e8233261d31fb7d
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:bbe57064cb1c40c35929526a082f11b2da6fd3746b6eef8af51ca63a4fd7d98b
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:46e3b8ccf3e7069b44eb3ad407329c03048b5219ef2886e237e1ab0dc3f7bae4
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:bb4b72c1510da84aad2d67516c58f37495d7682c51c8f2088e2c4a350fa36c50
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:dcbc4c5e2d46ec2b4911d8cd9f96a4c141bf7dd18b79aa4c25b77141e4f60b48
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:69a8faa2f4da2281c6c124508634628510d91ad40f725a9b41790e0821d64c14
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:500718f834e05c7a4531eb4f9d8ce560053c5125f76f6921c38c9cf150046a23
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:9db39230c528bfc675bc422a7acd40c394169b9632b088ca16358751994c9cac
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:335c903dde6f21e9c54f025c45666f6b4b6e7d455dfd5f85a6373b3947069f0a
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:3ccc45b0e9483f33be0d3b32a6d46fa794976131ad2f9861d96e319c8e35b9be
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:2f68a16a14f192b2cceaa07ed4840c08fbf413718a39bd584cacde866be7f2b1
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:9c4eb0ef8758d943796e13f65bc92b190db614eabd0fe7375ebfecc90e96e32c
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:da09087ba7d558611d733551b6ebc27ccf441581fc264efe7a5933db3816ef89
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:e02992cec8edcc9374c4cd30dfd7bccbda77d6d5cdd309862ab5265494f1278e
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:f4474ed21a71ad490778a6a378a856bec1884b4b9dc2c0e6c6140cc6c9d03b61
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:805e443fbfca4f816dc2dcea71f4cf723fd753ccea6d13f04d25945f29cb47dc
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:d8b8d130f75647cfeb9036db50d66fc4f1fca56fead6add4ba24c52f9213543e
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:13c33a997be0ddef59d1fb31835751367d4f901a1a6eac41284931bf7988035c
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:814f34f9e55b59d96021d3160b1dad7a68ffe208b3f4a999840b228aa7b3300b
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:47a8fa9541275f849a24e2ef1df87b7e9051dedd0ee8bd833d60defbd8724b46
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:3803d82d2b829c0e0883c0504176a96d2012cb7096cfcace316831a2b55bca3f
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:de68d598587a967d4b3f5e0e28c4675f74fd79cae7a5bb197c1b4b44bb4b31d9
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:35ff861165fccd848b8d328fae46cb8b3ac0c853c63062ec8672a6d41a23f333
        - registry.redhat.io/rhacm2/iam-policy-controller-rhel9@sha256:8c2b356bd9ff4ebfba9059a5f20cad8bd336af237d84cdc21f5c0aa35fabeabe
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:d989e9d2078f8a1e4b18de64f917e6d5e290c5cac769899f829b5f6bbfe6f0f2
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:87093f85c6235c99d4cca9d503b0e4198e3731e5fc3171506485091570e70bb5
        version: 2.10.9
      entries:
      - name: advanced-cluster-management.v2.10.9
        version: 2.10.9
      - name: advanced-cluster-management.v2.10.8
        version: 2.10.8
      - name: advanced-cluster-management.v2.10.7
        version: 2.10.7
      - name: advanced-cluster-management.v2.10.6
        version: 2.10.6
      - name: advanced-cluster-management.v2.10.5
        version: 2.10.5
      - name: advanced-cluster-management.v2.10.4
        version: 2.10.4
      - name: advanced-cluster-management.v2.10.3
        version: 2.10.3
      - name: advanced-cluster-management.v2.10.2
        version: 2.10.2
      - name: advanced-cluster-management.v2.10.1
        version: 2.10.1
      - name: advanced-cluster-management.v2.10.0
        version: 2.10.0
      name: release-2.10
    - currentCSV: advanced-cluster-management.v2.11.8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-08-01T05:25:37Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.10.0 <2.11.8'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:956d549e64c412f86f2083c8ddf97fea269d7806d8826aab93178bec9ed5d7af
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:9d821f4373722d9cb2b679cf5793ef22cb0b9ebcbb5a38060504c93c05c492c4
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:aef5ffbaa578d006bc7caec8fb35446dc263a51813f84c756aa57903b428329b
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:d248a5fad4ae3ee3210eb6dc4f58406d2f1e8f5c390f3c491040b437fe1adeba
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:447b1f6067a82476e14dfbb74a8f324fb86956420e55575f0017b91784f303a5
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:a081f531efb5b54d7d14d9ba0519552ecfc87c3a80d9d3ec4e598d2b445fc689
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:ffd3d9c363d8e15232bd6d12c5f1a77f7cb1ca18f94b7aea79d226568be4586e
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:95890bb1aad544d97ae1b80e801a800c95ef51e1034b60e680b67bdb2f5aeb69
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:f5c56a7381eb80c02a7c140f7c3f0ed08bb93697cd51732ee57498c080d8d3e5
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:8d15be5f40d135fd7069ac768dcf9188d8d8299c65db56df04e66f8a28941966
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:94c89ebd9257797ddc264511399689feb70d56e53332d9894c6d95fbdc51de40
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:6d4e3114494e051b3b875eeeae4bad4d48665fbcbbd3574a98f87b5a5742b41d
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:9be2d97fa3c1140db7d86157503a96ca2d3bdfb57125ab2230c8cf0649ea1e90
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:d4feeb3d3322210765d916ef1f49b47a1527f1c61041f94834c4395319a9ec32
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:49f0ec1aced72e6a2f0c862b0c6c0cd430d997a09ceb9eb0a7d25aaa1d9798b7
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:a9e0533af92ebadca27c9fe93fe4db9c68a1a8bc0060f4d13f67f26eb5691cb3
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:b3e0427341e3b1cbb1fe9166c56184eb0b0ec526d8a3992306e4861f65e4f0c0
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:ded47519fcf30e60cf38d44a101494b0318757969675782365de88b51cab24f0
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:968eb25c5cfcbdcea87b92dbff734d01b5954c13c13f5e794ad1db7bd886e2b4
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:d7552b2a3c864a3fc0d279a2519535e170fd8608195fe1f9abb4b538dfe2a3df
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:d6666c572294c7c85990304d21eaae3040b9c4e47fcb7b446289cb09aabc2d26
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:4a62d8706772ede443cffa005b984c26d137c35a33aafedaa7f01050b618c2a2
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:d4bfef313ff9044ae36aa5f832651a4dd09228d4b9b186803009aafe9774ee2c
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:1e8a26ebfcc33e60f60da4ba8d58cf7c152a99752eac6ebc9e114187e4cc2267
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:4dce3cebeb4ea0f1d3fc1c7e036dc6a3fd09e02f8f2ae4f3a753d9f4f97defe7
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:b713264826124df2c2b3a99dc268a3f3bd7eb24a9010b93c9876d83656e70e1a
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:e36832aa4852ea807d44f2b1f2c402bdaac6b7e2246bf9dd811b41f0b5e8e647
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:eaa719f9d9db8f91eced81e0e164bbc12bb502bf4760fc8c673cc542b4414771
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:45d4404abe02596611b29316bd10985f7a0684b1d305a57d3b24068ac776cf6e
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:251d3ee881c3e902e87f1752648ab084e8780aef79eb88a253ba79f984a01bcf
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:e347cd0f57e0691c61263e6554495c76ce56c7443bce3fd010cf76d068083fee
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:ad10d15efb6833b2daaf153210a080420395a7debfcb873dd757bb924197a654
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:36c90a52b12af24faaf28301d0d1a0c522723bd6cadd7391e5ee1f43f3019eef
        - registry.redhat.io/rhacm2/console-rhel9@sha256:2ed2f4add387caf790b2b41a65a3986e0046e08d0b8406ad17e4f0440ba0ac4a
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:5ba07381111651906ac3c93344e705152eb006826cb36bf5c05bd88723c4c999
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:7d39d0c29b02c399ffdcd83278693a8dd9c5ca3931e660af056030b126d10055
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:eec0e9bd6ea6f146939516da6c7ca9c3755888d96e81954f2f282fae63c1847c
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:827ca1beb707c02908914c517c8ab14b77b1c65ca7690dbbe1bcff600b7551d5
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:4e0c2cf5229fd68102a9e6d0f3155c651b1159b7af7b057151ce52990f747b08
        - registry.redhat.io/rhel9/postgresql-13@sha256:31c77c19480ab82cb33525d32419b68cf59daccf0b9936e88e8238e48a2d47e5
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:ef66846a3f178234618fd1e265067f863bbb56f14cbca0dfc96b0d741885dc9b
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:4b16da3070e30d879421aaa0d4a73aa2a1b0b852f62b8faabd801c03f22d5319
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:028b7258ec88d04397798552ed19a459110660ddbdf3ab375a7a1f74049c455a
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:147c567ce95673ba7151bcd290474756e13018ebc9f7275d61677ffe586a8a76
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:2878b5c41f8d58e986e86803bfd7eb3a05ccf23f7e49ebba56ad30245609b23d
        version: 2.11.8
      entries:
      - name: advanced-cluster-management.v2.11.8
        version: 2.11.8
      - name: advanced-cluster-management.v2.11.7
        version: 2.11.7
      - name: advanced-cluster-management.v2.11.6
        version: 2.11.6
      - name: advanced-cluster-management.v2.11.5
        version: 2.11.5
      - name: advanced-cluster-management.v2.11.4
        version: 2.11.4
      - name: advanced-cluster-management.v2.11.3
        version: 2.11.3
      - name: advanced-cluster-management.v2.11.2
        version: 2.11.2
      - name: advanced-cluster-management.v2.11.1
        version: 2.11.1
      - name: advanced-cluster-management.v2.11.0
        version: 2.11.0
      name: release-2.11
    - currentCSV: advanced-cluster-management.v2.12.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-21T21:16:17Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <2.12.4'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:b83e630e77bd011c374d73d801f10a3230f6798ede10d7f39bf2e8ff6437cf2c
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:3d3260ff206c5ee249443b1f8da49562584f7ee1abd284691b42f1728b6ed5dc
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:b85f3de5231f2f1bc3b4e9c95917f562f39e09dc9b7db35690024163e3a5afe5
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:6f2c4f1abc4388e76414544e65f08238c96fc4d6fbf71fa79330d9dca117d221
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:ec65a5744945ca20f804d94999a4002efab9ddad360a58f9498800bffa9fa0f4
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:187879345174e515863bc1b29a961a0a0d9ad44006a6e12de715b4508dec38a8
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:93df0586f9281fa2879eed0fdf2fb115be355d572cc24dc5c5c9a92137044877
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:58f08a93cd35bd2ac19f94a1779205c81a2fc3a552d1e77b7c659fd51c9dd34f
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:bd24cfabb81b464a31c3d8c21fa6371e46b83094d5c750acc33346203fe15e84
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:c2259b90decf24f8fb1cb058d4751c75fd1558e3c9fd1079857ad9c47ca9e8d2
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:22e262c0353c6f0d5d1b916074836c53437e22697f67c85a43e6d717daf651fa
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:c402a4ac61e11dc9e45fdc7bc7fcb5bf5ca3475a66d782aea60d9f6524e51a24
        - registry.redhat.io/rhel8/postgresql-13@sha256:826a1a75f4186f36217b5c1f1aa270838c4b55b1021d33a4876566bfc3b8e629
        - registry.redhat.io/rhacm2/console-rhel9@sha256:d239df44dafac3bde3eb7ae8c80cb342fa4bb4eabe22f27d2c92d283324e6e10
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:9c297b4c1f9ba17908594f1843ea6de6c5a5e0f3500258c23fbb8b5684249c9e
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:cdaf7be69a5d66cfa4777feb8266421dafd55353d6703e7e591bb9fb23017abc
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:7c548dd869f3458fb70a536708306cb615b9166c7e866cc1215e82f183b39529
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:1419e274b4e2e4c37985208c3b1b9ecd0cdef493427ace3c71ff35ed8fcc394b
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:a5c65dee1f7e4177050244109c10e384899e01d1924b41a540373aa7daf4f274
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:c46657831915ed5f42f4a55e725bc9d17e0671191f4509e610468fda160710c0
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:1e5c12ed0de090198e097cc5deb133ad075fba224f239b8b892e6e1672ec9883
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:fff4fb0924f866b305ad0ecc5d2682da6cb20de15448a2a8c640df182719b4e2
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:28730c8f2344e51b0b9ca603ac63b60885b39be86582943bd1cc200779eee0c5
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:1ccd17b482df844ec42936993ff162f665b5a473ad344270ff2425cd63d428fc
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:6a7c54cb7e6c3cc506ffb8a4f3ec7443ef80798e2c917e20b5f06d39bde02bc4
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:7375991ce3b9e64cb7542dc176414f0974047667d08beac184e2adadd9f83532
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:3ac71b363aef8db127b88e0a63ee616d0d06342142f83522e34c9cac4dd77b2e
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:c4c3bc4f45ae194b884d89319625f2a0970dada823877a35feecd8c733cb2c8d
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:7050f71c3a038fec16db8d0eee69d7d5a4294c043c13e9cf69c3cb2e28897f4c
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:70ddf96707d398d5444bb7bbde6c684275cff62b8a83c7cf6143ff529daf0dee
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:f3a00f55dee152e0345e2616ac14f652dcb4f37f3b9dfca6f049ac5522874c6e
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:f1c371fef66e96199c4a96e21192b47bca423b282d6f99e927644840aaedfd80
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:8ccbbc058d324429aef0c31c99de8c7b3a5abc02ec6f7a14cc60eef63a8f5c69
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:7b64803ddd3b7886ad3ec3de943a38cec5aeb9604732ab7a5ea5061748bf1411
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:5156129064dbe4e9989b41018b1bb250b2eb3f02e0c0afd75a874da9b54ad3c2
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:875e4bf42ad192cfb954e38c8474a529aa938855d69dc0d83c4ee4a4777c1028
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:3faa2318f159d264a9194fc01239d3a2098c184b365b153e3c56e555d5d4992e
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:7968e96b48b2ed63e7657793cf6cc2564055809df19cb43b7e284ac03f2cc26c
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:dca845ad05b02619b81f231eb57dd826720c221d22282fe7422ceebe57191371
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:794ef15591f217be629d6c680c56e84962923993f6dd77514ca5d2bdbd96217b
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:c10a72bf8c0d76b9b4d9d7c0bfca4d83a3f7dd7f6a37b4cc24ee7591638a2482
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:08dd3f6805fa02478671927982a2c77bdec0a5b7e7fd1c9faf20c8af2bb7b428
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:fbfade039739179995828b2ce4f4411b1315ac9a0e4a98e96a115012676a859d
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:b782190b3a285003d319321be09a25ee1376c951cb0cd146cbb634c9b1ca97b3
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:8e62fa4decf01a3beae7080f19469f33822f5010a688ada5db8d2d23b008e498
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:3f286e3c24fe73b4b42e6cf4b6769f91e333014450e1ab135b1e5b4a271cbaec
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:c2ef90b4704251dc9f7ef5288f5bad7aeb4cd2b3706ec531c60f6f4346073a77
        version: 2.12.4
      entries:
      - name: advanced-cluster-management.v2.12.4
        version: 2.12.4
      - name: advanced-cluster-management.v2.12.3
        version: 2.12.3
      - name: advanced-cluster-management.v2.12.2
        version: 2.12.2
      - name: advanced-cluster-management.v2.12.1
        version: 2.12.1
      - name: advanced-cluster-management.v2.12.0
        version: 2.12.0
      name: release-2.12
    - currentCSV: advanced-cluster-management.v2.13.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-09-08T18:37:07Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <2.13.4'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:dc271885425ab74e0554b531a2078133ce02006d0bb9688f3b54a850cfefd8f8
        - registry.redhat.io/rhel8/postgresql-16@sha256:89dfeaefc7e9e974e47a3365e22f354314f8a5e906e93a0152afe4e42f6a82d0
        - registry.redhat.io/rhacm2/console-rhel9@sha256:5526edf70b784bddfc2b7e1c154a44201c346fc20117635500bb54d325921c25
        - registry.redhat.io/rhacm2/acm-flightctl-worker-rhel9@sha256:57f69874b2df77c185374ddcea530af6783410a0f35528ee1a44cc2dee6f7fa2
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:0040791500337f88af7d922895487549915faa3b69dacba1a23540e7d6f2e580
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:ba51f7d077858a0abe32a254970390b50cd7d75433b5661e17f887c48c56a30f
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:65e1f0eb4e6c797c6dc6518674b28370db3fdaa6d7e17f4233c07207224dcb48
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:3028d011e7b88dd9597821128275f606fdba639e5b83e4e6a8783fa2616cba6f
        - registry.redhat.io/rhacm2/acm-flightctl-api-rhel9@sha256:8302b536c896290533f3f50084cb9047cb55af3a4404887b70704d23862619fa
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:d29ceda561f36faa581d76d8b889f7e01d253d33c88709078855ae26cada4917
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:0272215ab43b5688295d94cce0de4490a2fbabc1dd09ed41645be7acef1e3c9b
        - registry.redhat.io/rhacm2/acm-flightctl-ocp-ui-rhel9@sha256:89863ce0eb5fb73272bda8762d74461552f8502fbe262b16cbde01064cc533b3
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:84109be1814719c343b8b4676981cdbcea0356f012b6c854221c271bdb518fea
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:fc16203de469dd1ba16e428a9d7efbb16a5f511d866ef6fcc253530de1f032a6
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:ab1e4f8685896e2016a0b4c4479378058314f365e4f09ef42cba67f3649bb306
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:b3535378919524df259b85587ca83f5620433914418c9941374951c36e0530ae
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:1b7e6e6a4ae36209f98b1539c37c5975a82363b33fe5ac1e3a320e58b99be239
        - registry.redhat.io/rhacm2/acm-flightctl-periodic-rhel9@sha256:2c9b3c3d6cd7e435da337447580dfbfde8c4932308405eeae5d97ed5ec57fbfc
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:588678bea97237d87cd2e55a0e8000e73461cfc421cc91d554a2f1e352b68d14
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:7f742f024c787932932b8a4100a5329e70aad39537e0c20c2a4e245fc6245e25
        - registry.redhat.io/rhel9/redis-7@sha256:848f4298a9465dafb7ce9790e991bd8a11de2558e3a6685e1d7c4a6e0fc5f371
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:ccc9c0524149820d701c39355af89287ad749832053ea55d1585e77347fa0ec0
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:7b15c6b492e5cb2725b9b345aa3bfae4108e2815bc233ac063b81a67c6550fb6
        - registry.redhat.io/rhel8/postgresql-13@sha256:826a1a75f4186f36217b5c1f1aa270838c4b55b1021d33a4876566bfc3b8e629
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:b1620cebc62240f93b12e6937d15760440be119c421c127214974bdfa9d08ebd
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:0d8202cd10170bb0a138d52a1d71e630ff74df466541897ba86ed324a75ea1d8
        - registry.redhat.io/openshift4/ose-cli-artifacts-rhel9@sha256:67ff9f19edac8d623ce9080a8c05cde3da73afd68e7352e4ee5b22a9ccb596bc
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:3490daccc203fade26e7df348c74a907ab0ffc8901f4965606c8227849087c44
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:f74fb6b4d42f1d7431bfc9d2424f953732fed06d2abd8d1fc27b878b4d1322d5
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:ce342242c50c103fc5e4b0126551231977b35bd477eb627fb8f5d2a6510ddaa0
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:e23115b10a92c323b4cdaf0412919c273dbca33f19910120498bd49cb5810ff5
        - registry.redhat.io/rhacm2/acm-flightctl-ui-rhel9@sha256:86800d67ec8ed63e5408947ba6a441e5afc1050b72a86487b9357674ed495a7e
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:f1db9b10481319dc142ca6046e4e3141addb4c0afab482eeaf80643fad9ad039
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:96bdb2e25ae0f381de0c94a09a3cd11f3c0bdc1d8cfb4fd5ced2cd7b24fbe2c2
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:4d068813e3eb704ea7b2d342d5a2ce8fa2ac71793b1b37634c18704ef0a5e8d7
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:5fc35a944895cf7ea6afa5929c45304aaa9f17397b4f1e42cb8c799d9aead9d5
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:f9ba60c2a415a863c6bd6030fb4663913368a9c271bb40b8acfdc294dcf3a7eb
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:8b152051ef2c4c077379e72aa510c0bd613b219254cad66fab0cd3ec0d1d4bf1
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:04bcddae590e2debf35090055dd52ae4bcaac4b369727266be53ac64670203d0
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:49e99f41cbea3faefc3ae93385a10b544ce91df0d6639cfee65760be6ae0df00
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:81cc71faaf20a986d233fee799a7d2b954cd3f43875a2689f76c53920ccf4df0
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:67939d6e4342372820d7af6777f0ffd123e86e4e8186132c3d4892cb717a5e7c
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:6f95308a508499d2d995653d3fd2def94ac5170b4c3f391d32a1e0594250c33a
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:669e9e2178476bc7a4688f3ebf4c20402608bb4b10e4df6863ec821c7e572299
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:ae1f175ffbd3698f2b3036bf9b62b5c9b24729cfb8248345b8bdd1525ccc5bb4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:ab0192a51ee37b7508f3e4db66ffa53b904dcb1f76d9aa1ac0e80447d8ff9e6a
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:5aec572fc18c8ed533620553aba2c676453297ad2440336a5228a5a4f50f5782
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:a0218b9d5f1803f60d9b295d1504a870780895cffe8f1c5b3ce4c3553da5e4f9
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:a2ab651af1c97e54d2c14de78ff554c0f25e8d000af93a74fc7b7ed913f4e338
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:72c6896a5a4dbb7717d76d9eeade96130d97909eee07e8d3cc17af0658b306ab
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:9ef9f1e2010d77211126482c03e8c72dfb7458940dc95ac952ee8dd1dedd2991
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:6ed6f62f5c05a24938335287b785b80c253424e1ee20d83311cd87c6a2090895
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:2da264ca59278f63990e579939371d45afec3bc7d9922a608520e7647a0377ac
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:1393744757874e79d1fb058cb5d3fe7dcbbcd2f187bdcdcdf34e67227c61fbb9
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:296a31e1aad9f6288bdf80fb3191642e3dd404e9e8fcad198016bedbd3abe9b4
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:3cbb9d271c2f23d430a623c83606bbc8a80379c24d13e610886ad51fd21ce1e4
        version: 2.13.4
      entries:
      - name: advanced-cluster-management.v2.13.4
        version: 2.13.4
      - name: advanced-cluster-management.v2.13.3
        version: 2.13.3
      - name: advanced-cluster-management.v2.13.2
        version: 2.13.2
      - name: advanced-cluster-management.v2.13.1
        version: 2.13.1
      - name: advanced-cluster-management.v2.13.0
        version: 2.13.0
      name: release-2.13
    - currentCSV: advanced-cluster-management.v2.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-25T02:10:33Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.13.0 <2.14.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          repository: https://github.com/open-cluster-management/multiclusterhub-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MulticlusterHub defines the configuration
              for an instance of a multicluster hub, a central point for managing multiple
              Kubernetes-based clusters. The deployment of multicluster hub components
              is determined based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n### Where to find more installation
          information\nYou can find additional installation guidance in the [install
          documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:364c7416a9642dfb37f05e2eac4427918959b12b159ac90a9ca310c584d6434a
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:69f0b9a7fb494572adf2795be865d6d6ba4b67db0eed0e9d61f5dcb5e9bf6efe
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:c2ef90b4704251dc9f7ef5288f5bad7aeb4cd2b3706ec531c60f6f4346073a77
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:0e1eba23e0e8454f2dec240704e7ccc922351d58880f72514d3b2bc30c9e995e
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:c6ff6cae5bc6447d0ed24e458fc5d5fc709b2106fe93b143f9546bd962ea7f87
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:ad7f30c61bb11c0d32686e75b01f3b286a1fb06edd8d2ef6a9e2d924fba2b565
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:b489229190f9a3da66bd154e65490c3a35d1999d2db6707759f06eadb02b968d
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:8e6ee423acfafe28c002911e832d47dffc00a5227325bd6a5b1ab359c7eee5f4
        - registry.redhat.io/rhacm2/console-rhel9@sha256:d4c42a203110050a3834be8d5a0eee384ad7e9fbb991da589d75acd15055b8b1
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:0cecc3d87afcd0d63bdeae3ecae37f8a38534fbdc8bad5c548e0d4f4fb0156a4
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:8f4f59e237f366f08367cabbea3eb51e766ad52db0d35c5efc56182c13fc5c09
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:8253a9225f30b0f0a405e2c3293924acbfbf81f3d9c50bc4431fe4052a9fe908
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:55ef2c163b0c0deb3402e39acf37ff6bb05e2a5d268dcae5f8a400128d67c584
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:b7b7581e7743148d19283250d3300d292b96c48ecae96f6e6202a7ca8204e2ab
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:c329cfd004c58a7fcc40f4fbb986399dc2fe08d7922d2b4f210664f58751b1a6
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:143f49d2ce1f58282e276396722ba37165cd1bf2051d537bb4e6208786e5aed6
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:53deacec64238607afda29c27463d72ac3b582c9670e2e77a9c7e48851214795
        - registry.redhat.io/openshift4/ose-cli-rhel9@sha256:a48758b69b9a7451b7e0caea465a181f854b41f1e74ce86c15216ee1bb31fcb9
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:f9a9e10fde9065d650a7dc8eba83349e5db0a21727027464cf144724d0de8cfa
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:935baf67627d5e76c0766d5d2a9e36b398a12535f097749a987dd0c5de788570
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:0252cf95019717f03c7ac72fb034b5bd9ad7aef2da0a92dd21c56cf47f03b812
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:d56b997a9084d56c63f539147052e64ba30258afe91a9ff832f3673cbccbe467
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ee742e7baa8bcbbfccf79abcf15d22347d14f280eafaf22c42cef71051b10fba
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:0ab21d4189669901fe085c7f3a4b4c9778c853f8abd706bbfffcb59942fbc8e6
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:9c79b1153fc927e7582c5346ea0408e03f794eb983597f02a5e816ae961ba86b
        - registry.redhat.io/rhem/flightctl-periodic-rhel9@sha256:ee59f307e468e2b7b38f7c040898d587d441f8fb3d59cb022ae23e4717739e3f
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:2b1b1c707e0b5217c66a96fb82070de47fc0342747081d4be0049c4542944c8a
        - registry.redhat.io/rhel8/postgresql-16@sha256:d4abbd9492347e0aa7e1b0ff967e74d4b6a4edd6fd20d896f2f2ae03c623181c
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:46c5835a4bfc0e4767d0959bc47e939c195be197e9715aae17788d4d0e799d56
        - registry.redhat.io/rhel9/redis-7@sha256:ff31822cba556e40b6b00720839cc56f474fdb9257fb39b7604fafc11761d379
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:0ea80a1039d217e899ea012b77e99a7c5832e5b7e292dca473759c449bf2d1ea
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:e063c54c7182b70d78d24b5026f174436885383980c9b1e3e7664fa2adfad883
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:d550013d385552d6242b56c5b60f15b3ebb488c2033f808de54bcc639abc2f30
        - registry.redhat.io/rhem/flightctl-ui-ocp-rhel9@sha256:973eb3ca3da32c87ba308928a21fbd25514b2f6a6662300d4928479cb554cff3
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:bcb5949ebc31bbf6b45dfd34aae9c80c82a08531e8a3d9f6eef3354f942a5a5d
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:cee29bdb03927dc1866c267232d391ea5ee762b23c6b04a777e7d61d14a61bba
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:0cc23b5063c0d3264ed345dad8453901e4db18cc9bed1656f2799f72de75bd53
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:8151a89ebfebf2e6d336849fdd329b123b7f486938808ee80bf6bd8d44f6985c
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:7ee0aea84bf00b7b42027fa3e970c1f4b279e345a698b4050c9f413eb0383bb1
        - registry.redhat.io/rhem/flightctl-ui-rhel9@sha256:940c0564ced75b7912d3ffc923a1f6bab9aed0f4a52962f7f894d08937b7d37e
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:ce5ff191023b5e14c339b2a03aa1be05a12d024efa5302652065597e630996b6
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:90d399736ca81bb244415efaf05738beff806912a6ac7444a7d4839089b7a081
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:556d737195a4878877c650a93619af1528e52694f2a8e47e13463a603d2f4900
        - registry.redhat.io/rhem/flightctl-cli-artifacts-rhel9@sha256:3491ec932d90112e89f3a4cebe5bdd9ed6a7811f116df43bab8bb5f827129361
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:d9fa60b5b1d6b6c94c5d8b610c9a7b979202de8b46a344e0a65d06f6ef21a7ef
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:40b3e063435eea6e526a57d6bd1ae489ca703e81eeca82dda1c2e6aedec1262c
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:95085fc33a0150a39b6ef4e616fd23f0bddd9dcad0e75c4a89ed27113626e6eb
        - registry.redhat.io/rhel9/memcached@sha256:955138db9a7a23ee31198aabebd0811438fee11bb0b624c38ddb90efe52bd07c
        - registry.redhat.io/rhem/flightctl-worker-rhel9@sha256:1b1f6f85af0910f1acd029f54a608b25d3a2977c3d8b323266391543324305b4
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:f3e34b200e9480363c56da1cd88b2de37fc6a490691e4596a132217f8ea286d8
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:c70242a037e36690d8dd1030254dde778917356b4f1f61db2cecb03b6c2d7380
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:cb5346afe1c7b58eb758cdced94f2c1161c1621033b52f29757efe3a50320a65
        - registry.redhat.io/rhem/flightctl-api-rhel9@sha256:83c6c7819e04aa58d961518367ce6ad3159a28300ca097967f630044d181be34
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:8e57af77cc5dee3aaa83f6ae6e62e232c5fc4fbe4dca8eb755226351963058e3
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:44e305ca15e7c9b50bd7fa8bb391f5ee66906aa26ecc5f612d029aaee5bb4cb8
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:d65aef370a7ed01c0c7cc0ad7f0800359db7863db4ec8845de5b450992f26933
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:ad534f7b0efea45b9f4a0a026c76a2a030cb608acdb835dfea3852332acfcae1
        version: 2.14.0
      entries:
      - name: advanced-cluster-management.v2.14.0
        version: 2.14.0
      name: release-2.14
    defaultChannel: release-2.14
    packageName: advanced-cluster-management
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/component: kmm-hub
      app.kubernetes.io/name: kmm-hub
      app.kubernetes.io/part-of: kmm
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: kernel-module-management-hub
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kernel-module-management-hub.v2.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:888a3acad339cf38a324c718e0cbe1237ca5bf2f4636b01d001752658e4cca1b
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.0.2'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:888a3acad339cf38a324c718e0cbe1237ca5bf2f4636b01d001752658e4cca1b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62c44057a27abd41cf3daf60b8736da59a82dadb65fa6c32bef65d19bd49ea49
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:192e08b121a04ae69aea60c35199e8f5e3c42305e1ebd024a605a371de4ab472
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:94b4ddc45cb34d4ffe794a51b99c868cb709f90cc4eb8b28284d70c526a0a226
        version: 2.0.2
      entries:
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.0
    - currentCSV: kernel-module-management-hub.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:d1c1490d4c20dc2ec31effa1af067676c4d264e3b7d4670fc11c2baf8a400001
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.1.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:14e100e5853444114029116255ebfb7a54e5ed51503ba969740ad48b5ca80f70
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:1a9a19df17b464d1708b2905d03b8726e4480ca1f11aa97a15ff4c049a24ea4b
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:d1c1490d4c20dc2ec31effa1af067676c4d264e3b7d4670fc11c2baf8a400001
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:611556d999ea97f9dbb78fec5963baa02de7a26626d7096ba63b177e3bcf9707
        version: 2.1.1
      entries:
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.1
    - currentCSV: kernel-module-management-hub.v2.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:90c1946b3f577cde0435402b04baa52dc54e5aab3bb84bd834ac026a57453296
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.2.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:b1d98c0f14448ae8eb80821cbec236150c338674aae2a2b37f8d0ab0d5f1430c
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:53fddf0b226eb579c71c86112a8d9b6b1d90a1498d6f6985932b16b8be532f35
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:c3dc9ad4ebc68cd85a074b520903a9310e72a90a73bdf4aa7c42b7ed66b7bcba
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:90c1946b3f577cde0435402b04baa52dc54e5aab3bb84bd834ac026a57453296
        version: 2.2.1
      entries:
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.2
    - currentCSV: kernel-module-management-hub.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:f1d8f3c95f621c3faf41867c5dd4bc13fa2b9b291422d6184e6e42f1d1460259
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.3.0'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:f1d8f3c95f621c3faf41867c5dd4bc13fa2b9b291422d6184e6e42f1d1460259
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:c5005b74402f680d11c9ea682a9a0d3f74b6ab0d01d25298d232f910f1a29c34
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:f9fde334af68220943017a7491130e6a82f15e63522f5bb3931aa7d9b00ccc40
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:06ac089cc89b303e525fd71d6439b958f34ab944a72209f1cc8cb562ecfd1112
        version: 2.3.0
      entries:
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.3
    - currentCSV: kernel-module-management-hub.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm-hub",
                    "app.kubernetes.io/name": "kmm-hub",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:2efb81137ad2981b994c62d460c917749308133a7dcad5b9a2b251e18f3052ec
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        version: 2.4.1
      entries:
      - name: kernel-module-management-hub.v2.4.1
        version: 2.4.1
      - name: kernel-module-management-hub.v2.4.0
        version: 2.4.0
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.4
    - currentCSV: kernel-module-management-hub.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm-hub",
                    "app.kubernetes.io/name": "kmm-hub",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:2efb81137ad2981b994c62d460c917749308133a7dcad5b9a2b251e18f3052ec
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        version: 2.4.1
      entries:
      - name: kernel-module-management-hub.v2.4.1
        version: 2.4.1
      - name: kernel-module-management-hub.v2.4.0
        version: 2.4.0
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: kernel-module-management-hub
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: cert-utils-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-utils-operator.v1.3.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/cert-utils-operator@sha256:0290e7b2800a5aa8eb57b0268908f2a20a93b16403c76df4c61021048161d6ff
          createdAt: "2024-01-23T12:29:04Z"
          description: Set of utilities for TLS certificates
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-utils-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/cert-utils-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Cert utils operator is a set of functionalities around certificates packaged in a [Kubernetes operator](https://github.com/operator-framework/operator-sdk).

          Certificates are assumed to be available in a [secret](https://kubernetes.io/docs/concepts/configuration/secret/) of type `kubernetes.io/tls` (other types of secrets are *ignored* by this operator).
          By convention this type of secrets have three optional entries:

          1. `tls.key`: the private key of the certificate.
          2. `tls.crt`: the actual certificate.
          3. `ca.crt`: the CA bundle that validates the certificate.

          The functionalities are the following:

          1. [Ability to populate route certificates](https://github.com/redhat-cop/cert-utils-operator#Populating-route-certificates)
          2. [Ability to create java keystore and truststore from the certificates](https://github.com/redhat-cop/cert-utils-operator#Creating-java-keystore-and-truststore)
          3. [Ability to show info regarding the certificates](https://github.com/redhat-cop/cert-utils-operator#Showing-info-on-the-certificates)
          4. [Ability to alert when a certificate is about to expire](https://github.com/redhat-cop/cert-utils-operator#Alerting-when-a-certificate-is-about-to-expire)
          5. [Ability to inject ca bundles in ValidatingWebhookConfiguration, MutatingWebhookConfiguration, CustomResourceDefinition object](https://github.com/redhat-cop/cert-utils-operator#CA-injection)

          All these feature are activated via opt-in annotations.
        displayName: Cert Utils Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Cert Utils Operator
          url: https://github.com/redhat-cop/cert-utils-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/cert-utils-operator
        maintainers:
        - email: rspazzoli@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/cert-utils-operator@sha256:0290e7b2800a5aa8eb57b0268908f2a20a93b16403c76df4c61021048161d6ff
        version: 1.3.12
      entries:
      - name: cert-utils-operator.v1.3.12
        version: 1.3.12
      - name: cert-utils-operator.v1.3.11
        version: 1.3.11
      - name: cert-utils-operator.v1.3.10
        version: 1.3.10
      - name: cert-utils-operator.v1.3.9
        version: 1.3.9
      - name: cert-utils-operator.v1.3.8
        version: 1.3.8
      - name: cert-utils-operator.v1.3.7
        version: 1.3.7
      - name: cert-utils-operator.v1.3.6
        version: 1.3.6
      - name: cert-utils-operator.v1.3.5
        version: 1.3.5
      - name: cert-utils-operator.v1.3.4
        version: 1.3.4
      - name: cert-utils-operator.v1.3.3
        version: 1.3.3
      - name: cert-utils-operator.v1.3.2
        version: 1.3.2
      - name: cert-utils-operator.v1.3.1
        version: 1.3.1
      - name: cert-utils-operator.v1.3.0
        version: 1.3.0
      - name: cert-utils-operator.v1.2.0
        version: 1.2.0
      - name: cert-utils-operator.v1.1.0
        version: 1.1.0
      - name: cert-utils-operator.v1.0.7
        version: 1.0.7
      - name: cert-utils-operator.v1.0.6
        version: 1.0.6
      - name: cert-utils-operator.v1.0.5
        version: 1.0.5
      - name: cert-utils-operator.v1.0.4
        version: 1.0.4
      - name: cert-utils-operator.v1.0.3
        version: 1.0.3
      - name: cert-utils-operator.v1.0.2
        version: 1.0.2
      - name: cert-utils-operator.v1.0.1
        version: 1.0.1
      - name: cert-utils-operator.v1.0.0
        version: 1.0.0
      - name: cert-utils-operator.v0.2.2
        version: 0.2.2
      - name: cert-utils-operator.v0.2.1
        version: 0.2.1
      - name: cert-utils-operator.v0.2.0
        version: 0.2.0
      - name: cert-utils-operator.v0.1.1
        version: 0.1.1
      - name: cert-utils-operator.v0.1.0
        version: 0.1.0
      - name: cert-utils-operator.v0.0.3
        version: 0.0.3
      - name: cert-utils-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cert-utils-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://access.redhat.com/products/red-hat-integration
    name: red-hat-camel-k
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: red-hat-camel-k-operator.v1.10.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
          createdAt: "2025-01-08T17:00:43Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
        version: 1.10.10
      entries:
      - name: red-hat-camel-k-operator.v1.10.10
        version: 1.10.10
      - name: red-hat-camel-k-operator.v1.10.9
        version: 1.10.9
      - name: red-hat-camel-k-operator.v1.10.9-0.1743522834.p
        version: 1.10.9+0.1743522834.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1742828111.p
        version: 1.10.9+0.1742828111.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741884247.p
        version: 1.10.9+0.1741884247.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741875993.p
        version: 1.10.9+0.1741875993.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741580860.p
        version: 1.10.9+0.1741580860.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739988175.p
        version: 1.10.9+0.1739988175.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739388437.p
        version: 1.10.9+0.1739388437.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739387844.p
        version: 1.10.9+0.1739387844.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.10.x
    - currentCSV: red-hat-camel-k-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1beta1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:9e4fb5369e0479d4a68f60725e7d12abdc66c552babe59f1358f6e6b0f26357d
          createdAt: "2020-10-06T02:45:00Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:9e4fb5369e0479d4a68f60725e7d12abdc66c552babe59f1358f6e6b0f26357d
        version: 1.4.2
      entries:
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.4.x
    - currentCSV: red-hat-camel-k-operator.v1.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
          createdAt: "2022-08-15T20:17:25Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <=1.6.8'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
        version: 1.6.10
      entries:
      - name: red-hat-camel-k-operator.v1.6.10
        version: 1.6.10
      - name: red-hat-camel-k-operator.v1.6.9
        version: 1.6.9
      name: 1.6.x
    - currentCSV: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:e916f9040d3f4feb26cc1534d3ab2722890b34c7fe146cdce8b6085b47206e22
          createdAt: "2022-11-28T10:43:18Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <1.8.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:e916f9040d3f4feb26cc1534d3ab2722890b34c7fe146cdce8b6085b47206e22
        version: 1.8.2+0.1675913507.p
      entries:
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.8.x
    - currentCSV: red-hat-camel-k-operator.v1.10.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
          createdAt: "2025-01-08T17:00:43Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
        version: 1.10.10
      entries:
      - name: red-hat-camel-k-operator.v1.10.10
        version: 1.10.10
      - name: red-hat-camel-k-operator.v1.10.9
        version: 1.10.9
      - name: red-hat-camel-k-operator.v1.10.9-0.1743522834.p
        version: 1.10.9+0.1743522834.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1742828111.p
        version: 1.10.9+0.1742828111.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741884247.p
        version: 1.10.9+0.1741884247.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741875993.p
        version: 1.10.9+0.1741875993.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741580860.p
        version: 1.10.9+0.1741580860.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739988175.p
        version: 1.10.9+0.1739988175.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739388437.p
        version: 1.10.9+0.1739388437.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739387844.p
        version: 1.10.9+0.1739387844.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: candidate
    - currentCSV: red-hat-camel-k-operator.v1.10.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
          createdAt: "2025-01-08T17:00:43Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
        version: 1.10.10
      entries:
      - name: red-hat-camel-k-operator.v1.10.10
        version: 1.10.10
      - name: red-hat-camel-k-operator.v1.10.9
        version: 1.10.9
      - name: red-hat-camel-k-operator.v1.10.9-0.1743522834.p
        version: 1.10.9+0.1743522834.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1742828111.p
        version: 1.10.9+0.1742828111.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741884247.p
        version: 1.10.9+0.1741884247.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741875993.p
        version: 1.10.9+0.1741875993.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741580860.p
        version: 1.10.9+0.1741580860.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739988175.p
        version: 1.10.9+0.1739988175.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739388437.p
        version: 1.10.9+0.1739388437.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739387844.p
        version: 1.10.9+0.1739387844.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: latest
    - currentCSV: red-hat-camel-k-operator.v1.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
          createdAt: "2022-08-15T20:17:25Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <=1.6.8'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
        version: 1.6.10
      entries:
      - name: red-hat-camel-k-operator.v1.6.10
        version: 1.6.10
      - name: red-hat-camel-k-operator.v1.6.9
        version: 1.6.9
      name: stable
    defaultChannel: 1.10.x
    packageName: red-hat-camel-k
    provider:
      name: Red Hat
      url: https://access.redhat.com/products/red-hat-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.17.40-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5-beta.1 <8.17.40-beta.1'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBM Turbonomic software automates and simultaneously optimizes
              performance, compliance, and cost in real time. Turbonomic discovers
              and automates resource optimization for the complete application stack.
              Applications are continually resourced to perform while satisfying technical
              and business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
          - kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Optimization for Kubernetes
          IBM Turbonomic software automates and simultaneously optimizes performance, compliance, and cost in real time.
          Turbonomic discovers and automates resource optimization for the complete application stack. Applications are continually resourced to perform while satisfying technical and business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (proactive rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          Turbonomic assures application performance by giving workloads the resources they need when they need them.

          ### How Does It Work?
          Turbonomic uses a client agent called — KubeTurbo — that runs in your cluster to discover, monitor and optimize resource usage within your environment.
          KubeTurbo runs together with the default Kubernetes scheduler and sends data back to the Turbonomic for analysis.
          Turbonomic determines the right actions that drive continuous health for your workloads and your cluster.

          Turbonomic discovers Red Hat OpenShift Virtualization resources through a new or existing target configuration with Red Hat OpenShift. For more information refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=targets-red-hat-openshift-virtualization)

          ### Get Started
          Please follow the [Kubeturbo Documentation Page](https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets) for all deployment options and an overview of typical use cases.

          ### Known Issues
          For known issues refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=notes-known-issues)
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Kubeturbo Documentation Page
          url: https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets
        - name: Kubeturbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
        - name: IBM Turbonomic Application Resource Management
          url: https://www.ibm.com/products/turbonomic
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/kubeturbo-operator@sha256:7111c7c4002a447dc6977dbb73e96507bae56766fc0552fb2d036b20583722ba
        version: 8.17.40-beta.1
      entries:
      - name: kubeturbo-operator.v8.17.40-beta.1
        version: 8.17.40-beta.1
      name: beta
    - currentCSV: kubeturbo-operator.v8.17.40
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5 <8.17.40'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBM Turbonomic software automates and simultaneously optimizes
              performance, compliance, and cost in real time. Turbonomic discovers
              and automates resource optimization for the complete application stack.
              Applications are continually resourced to perform while satisfying technical
              and business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
          - kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Optimization for Kubernetes
          IBM Turbonomic software automates and simultaneously optimizes performance, compliance, and cost in real time.
          Turbonomic discovers and automates resource optimization for the complete application stack. Applications are continually resourced to perform while satisfying technical and business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (proactive rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          Turbonomic assures application performance by giving workloads the resources they need when they need them.

          ### How Does It Work?
          Turbonomic uses a client agent called — KubeTurbo — that runs in your cluster to discover, monitor and optimize resource usage within your environment.
          KubeTurbo runs together with the default Kubernetes scheduler and sends data back to the Turbonomic for analysis.
          Turbonomic determines the right actions that drive continuous health for your workloads and your cluster.

          Turbonomic discovers Red Hat OpenShift Virtualization resources through a new or existing target configuration with Red Hat OpenShift. For more information refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=targets-red-hat-openshift-virtualization)

          ### Get Started
          Please follow the [Kubeturbo Documentation Page](https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets) for all deployment options and an overview of typical use cases.

          ### Known Issues
          For known issues refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=notes-known-issues)
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Kubeturbo Documentation Page
          url: https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets
        - name: Kubeturbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
        - name: IBM Turbonomic Application Resource Management
          url: https://www.ibm.com/products/turbonomic
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/kubeturbo-operator@sha256:829dc6b9800efd3fd77d07f955a53610c7b2888b93495b2db32f1f48e93a05a6
        version: 8.17.40
      entries:
      - name: kubeturbo-operator.v8.17.40
        version: 8.17.40
      name: stable
    defaultChannel: stable
    packageName: kubeturbo-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: VASTData
      provider-url: https://www.vastdata.com
    name: vast-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vast-csi-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"storage.vastdata.com/v1","kind":"VastCSIDriver","metadata":{"name":"csi.vastdata.com"},"spec":{"applySecurityContextConstraints":true,"attachRequired":true,"controller":{"dnsPolicy":"Default","extraArgs":{"csiAttacher":[],"csiProvisioner":[],"csiResizer":[],"csiSnapshotter":[]},"nodeAffinity":{},"nodeSelector":{},"podAffinity":{},"podAntiAffinity":{},"resources":{"csiAttacher":{"limits":{"memory":"200Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiProvisioner":{"limits":{"memory":"400Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiResizer":{"limits":{"memory":"300Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiSnapshotter":{"limits":{"memory":"200Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiVastPlugin":{"limits":{"memory":"400Mi"},"requests":{"cpu":"100m","memory":"50Mi"}}},"runOnControlPlane":false,"runOnMaster":false,"tolerations":[]},"deletionViewPolicy":null,"deletionVipPool":null,"dontUseTrashApi":false,"driverType":"nfs","endpoint":null,"image":{"csiAttacher":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-attacher@sha256:b4d611100ece2f9bc980d1cb19c2285b8868da261e3b1ee8f45448ab5512ab94"},"csiNodeDriverRegistrar":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac"},"csiProvisioner":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-provisioner@sha256:405a14e1aa702f7ea133cea459e8395fe40a6125c088c55569e696d48e1bd385"},"csiResizer":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-resizer@sha256:a541e6cc2d8b011bb21b1d4ffec6b090e85270cce6276ee302d86153eec0af43"},"csiSnapshotter":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-snapshotter@sha256:2e04046334baf9be425bb0fa1d04c2d1720d770825eedbdbcdb10d430da4ad8c"},"csiVastPlugin":{"imagePullPolicy":"IfNotPresent","repository":"registry.connect.redhat.com/vastdataorg/csi@sha256:7169f279d0dd7130535d74df5b938ed6b7944bfd4fe52d242ceee5178f2ea156"}},"imagePullSecrets":[],"kubeletPath":"/var/lib/kubelet","logLevel":5,"node":{"dnsPolicy":"Default","nodeAffinity":{},"nodeSelector":{},"podAffinity":{},"podAntiAffinity":{},"propagateHostMountOptions":true,"resources":{"csiVastPlugin":{"limits":{"memory":"400Mi"},"requests":{"cpu":"100m","memory":"50Mi"}},"nodeDriverRegistrar":{"limits":{"memory":"100Mi"},"requests":{"cpu":"10m","memory":"20Mi"}}},"tolerations":[]},"numWorkers":10,"operationRetryIntervalMax":60,"operationRetryIntervalStart":10,"operationTimeout":15,"secretName":null,"sslCertsSecretName":null,"truncateVolumeName":64,"useLocalIpForMount":null,"verifySsl":false}},{"apiVersion":"storage.vastdata.com/v1","kind":"VastStorage","metadata":{"name":"vastdata-filesystem"},"spec":{"allowVolumeExpansion":true,"blockingClones":false,"clusterName":"","createSnapshotClass":true,"driverType":"nfs","ephemeralVolumeNameFormat":"csi:{namespace}:{name}:{id}","fsType":"","hostEncryption":null,"mountOptions":[],"provisioner":"","qosPolicy":"","reclaimPolicy":"Delete","secretName":"","secretNamespace":"","setDefaultStorageClass":false,"snapshotClass":{"deletionPolicy":"Delete","setDefaultSnapshotClass":false,"snapshotNameFormat":"csi:{namespace}:{name}:{id}"},"storagePath":"","subsystem":"","tenantName":"","viewPolicy":"","vipPool":"","vipPoolFQDN":null,"volumeGroup":"","volumeNameFormat":"csi:{namespace}:{name}:{id}"}},{"apiVersion":"storage.vastdata.com/v1","kind":"VastCluster","metadata":{"name":"vastcluster-sample"},"spec":{"endpoint":null,"passphrase":null,"password":null,"sslCert":null,"tenant":null,"token":null,"username":null}}]'
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/vastdataorg/csi-operator@sha256:ba7f30a42e327004ecbc81ff5ccc453bf3046a29bab5785b9f8a9746e5c6a3fd
          createdAt: "2024-07-01T11:59:59Z"
          description: |-
            VAST CSI Operator supports seamless installation and upgrade of VAST Data's CSI Driver.
            The VAST CSI Driver allows container orchestration frameworks such as Kubernetes to dynamically provision, manage and scale volumes on the [VAST Data Platform](https://www.vastdata.com/platform/overview). VAST provides scalable, reliable, and fast persistent storage that can be accessed remotely by any Kubernetes application container.
            ### Installation
            Refer to the VASTData Operator for Kubernetes [official documentation](https://support.vastdata.com/s/document-item?bundleId=vast-csi-driver-2.6-administrator-s-guide&topicId=deploying-vast-csi-driver/deploying-vast-csi-driver-on-openshift-with-vast-csi-operator.html&_LANG=enus).
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: vast-csi
          operators.operatorframework.io/builder: operator-sdk-v1.3.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/vast-data/vast-csi
          support: VastData Infra Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VastCSIDriver is a common specification for VAST CSI Controller
              and VAST CSI Node plugins, typically intended for creation in a single
              instance.
            displayName: VastCSIDriver
            kind: VastCSIDriver
            name: vastcsidrivers.storage.vastdata.com
            version: v1
          - description: VastCluster encapsulates the connection details required
              for accessing a VAST cluster. This information is ultimately stored
              in a Kubernetes Secret.
            displayName: VastCluster
            kind: VastCluster
            name: vastclusters.storage.vastdata.com
            version: v1
          - description: Represents VAST storage. Each VastStorage instance generates
              a corresponding storage class that can be used in PVCs to create workloads.
            displayName: VastStorage
            kind: VastStorage
            name: vaststorages.storage.vastdata.com
            version: v1
        description: |-
          VAST CSI Operator supports seamless installation and upgrade of VAST Data's CSI Driver.
          The VAST CSI Driver allows container orchestration frameworks such as Kubernetes to dynamically provision, manage and scale volumes on the [VAST Data Platform](https://www.vastdata.com/platform/overview). VAST provides scalable, reliable, and fast persistent storage that can be accessed remotely by any Kubernetes application container.
          ### Installation
          Refer to the VASTData Operator for Kubernetes [official documentation](https://support.vastdata.com/s/document-item?bundleId=vast-csi-driver-2.6-administrator-s-guide&topicId=deploying-vast-csi-driver/deploying-vast-csi-driver-on-openshift-with-vast-csi-operator.html&_LANG=enus).
        displayName: VAST CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - VAST
        - VASTData
        - CSI
        - Multi Cluster
        - AI
        - High-Performance Computing
        links:
        - name: VAST CSI Plugin
          url: https://github.com/vast-data/vast-csi
        maintainers:
        - email: volodymyr.boiko@vastdata.com
          name: VAST Infrastructure Team
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: VASTData
          url: https://www.vastdata.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.connect.redhat.com/vastdataorg/csi-operator@sha256:ba7f30a42e327004ecbc81ff5ccc453bf3046a29bab5785b9f8a9746e5c6a3fd
        - registry.connect.redhat.com/vastdataorg/csi@sha256:7169f279d0dd7130535d74df5b938ed6b7944bfd4fe52d242ceee5178f2ea156
        - registry.k8s.io/sig-storage/csi-resizer@sha256:a541e6cc2d8b011bb21b1d4ffec6b090e85270cce6276ee302d86153eec0af43
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:2e04046334baf9be425bb0fa1d04c2d1720d770825eedbdbcdb10d430da4ad8c
        - registry.k8s.io/sig-storage/csi-attacher@sha256:b4d611100ece2f9bc980d1cb19c2285b8868da261e3b1ee8f45448ab5512ab94
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:405a14e1aa702f7ea133cea459e8395fe40a6125c088c55569e696d48e1bd385
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac
        version: 2.6.3
      entries:
      - name: vast-csi-operator.v2.6.3
        version: 2.6.3
      name: stable
    defaultChannel: stable
    packageName: vast-csi-operator
    provider:
      name: VASTData
      url: https://www.vastdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MySQL
      provider-url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
    name: ndb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ndb-operator.v1.6.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "mysql.oracle.com/v1","kind": "NdbCluster","metadata":
            {"name": "example-ndb"},"spec": {"redundancyLevel": 2,"dataNode": {"nodeCount":
            2},"mysqlNode": {"nodeCount":2}}}]'
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: container-registry.oracle.com/mysql/community-ndb-operator@sha256:b562dc3303ffae9f4dab1d73fd0fe05833bdaa8b731af4227da81af617c75c6b
          createdAt: "2024-12-18T15:05:46.000Z"
          description: The MySQL NDB Operator is a Kubernetes operator for managing
            a MySQL NDB Cluster setup inside a Kubernetes Cluster.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Ndb Cluster
            displayName: NdbCluster
            kind: NdbCluster
            name: ndbclusters.mysql.oracle.com
            version: v1
        description: |
          ## NdbCluster Custom Resource
          The NdbCluster custom resource is defined with custom fields that map to the configuration of a MySQL Cluster. To deploy a MySQL Cluster inside the K8s Cluster, an NdbCluster object with the desired configuration of the MySQL Cluster has to be created in the K8s Cluster. The NDB Operator watches for any changes made to any NdbCluster objects and will receive an event when the NdbCluster object is created in the K8s Cluster. It will then start deploying the MySQL Cluster, with the specified configuration, inside the K8s Cluster using native Kubernetes workloads and resources. Once the MySQL Cluster is ready, the NdbCluster resource can be further modified to make changes to the MySQL Cluster configuration. The Operator will receive this event as well and will update the configuration of the already running MySQL Cluster.

          The documentation at [NdbCluster-CRD.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md) explains the various custom fields of the NdbCluster CRD and the effect they have on the MySQL Cluster configuration. The [examples](https://github.com/mysql/mysql-ndb-operator/tree/main/docs/examples) folder has a few examples of the NdbCluster resource objects which can be used to deploy MySQL Cluster inside the K8s Cluster.
          ## Deploy a simple MySQL Cluster

          The [examples/example-ndb.yaml](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/examples/example-ndb.yaml) file has a NdbCluster object example-ndb that defines a simple MySQL Cluster with 2 data nodes and 2 MySQL Servers. To create this object in the default namespace of the K8s Cluster, run this command from the root of the repo :
          ```
          kubectl apply -f docs/examples/example-ndb.yaml
          ```

          Various NdbCluster status fields report back the actual status of the MySQL Cluster running inside the K8s Cluster. Wait for the MySQL Cluster nodes to be started and become ready using the [UpToDate](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md#ndbclusterconditiontypestring-alias) Condition defined in the NdbCluster object status :
          ```
          kubectl wait --for=condition=UpToDate ndb example-ndb --timeout=10m
          ```
          The MySQL Cluster is ready for transactions once the NdbCluster is UpToDate, i.e. the UpToDate condition becomes true.

          At any point, a brief status of an NdbCluster object can be viewed using kubectl get ndb command.
          ```
          kubectl get ndb example-ndb
          ```

          The output of the command gives a short summary the MySQL Cluster nodes controlled by the respective NdbCluster object.
          ```
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE   UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:2/2    Ready:2/2 3m    True
          ```

          To list all the pods created by the NDB Operator, run :
          ```
          kubectl get pods -l mysql.oracle.com/v1=example-ndb
          ```

          For example-ndb NdbCluster object, two management nodes (*-mgmd-* pods), two multi-threaded data nodes (*-ndbmtd-* pods) and two MySQL Servers (*-mysqld-*) would have been created by the NDB operator :
          ```
          NAME                   READY   STATUS    RESTARTS   AGE
          example-ndb-mgmd-0     1/1     Running   0          3m53s
          example-ndb-mgmd-1     1/1     Running   0          2m5s
          example-ndb-mysqld-0   1/1     Running   0          78s
          example-ndb-mysqld-1   1/1     Running   0          78s
          example-ndb-ndbmtd-0   1/1     Running   0          3m53s
          example-ndb-ndbmtd-1   1/1     Running   0          3m53s
          ```
          If running inside a openshift cluster, pods from statefulset "example-ndb-mysqld" will not start up. By default openshift runs all the pod with "restricted" scc, but the statefulset "example-ndb-mysqld" required higher privilege. Use the below command to check the exact scc required
          ```
          oc get pods example-ndb-mysqld-0 -o yaml | oc adm policy scc-subject-review -f -
          ```
          Create a user account with the scc and set the user to the statefulset "example-ndb-mysqld". Below commands creates an user with "anyuid" and assign it to the statefulset.
          ```
          oc create sa any-uid-user
          oc adm policy add-scc-to-user anyuid -z any-uid-user
          oc set sa sts example-ndb-mysqld any-uid-user
          ```
          ## Connect to the MySQL Cluster

          The NDB Operator, by default, creates few Services to expose the services offered by the MySQL Cluster nodes within the Kubernetes Cluster.
          ```
          kubectl get services -l mysql.oracle.com/v1=example-ndb
          ```

          The above command will generate an output similar to :
          ```
          NAME                 TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S) AGE
          example-ndb-mgmd     ClusterIP   10.100.207.255   <none>        1186/TCP 5m
          example-ndb-mysqld   ClusterIP   10.103.182.157   <none>        3306/TCP 5m
          example-ndb-ndbmtd   ClusterIP   None             <none>        1186/TCP 5m
          ```

          The NDB Operator will also create a root user with a random password on all the MySQL Servers. The root password will be stored in a K8s Secret which can be retrieved from the NdbCluster object status :
          ```
          kubectl get ndb example-ndb -o jsonpath={.status.generatedRootPasswordSecretName}
          ```

          To extract and decode the password from the Secret :
          ```
          base64 -d <<< \
          $(kubectl get secret example-ndb-mysqld-root-password \
          -o jsonpath={.data.password})
          ```

          One can also specify the password and host to be used with the root account via the NdbCluster spec. See the [CRD documentation](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md#ndbmysqldspec) for more details on this.
          ## Access MySQL Cluster from inside K8s

          To connect to the Management and MySQL Servers from within the Kubernetes Cluster, you can use the example-ndb-mgmd and example-ndb-mysqld services as the MySQL Cluster connectstring and MySQL host respectively. You can also use the mgmd and mysqld pods' hostnames to directly connect to that particular pod. The mysql, ndb_mgm clients and any other ndb tools will work as expected.

          A demonstration is available at [connect-from-inside-k8s-demo.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/connect-from-inside-k8s-demo.md).
          ## Access MySQL Cluster from outside K8s

          By default, the Management and MySQL services created by the NDB Operator are of type [ClusterIP](https://kubernetes.io/docs/concepts/services-networking/service/#publishing-services-service-types) and are accessible only from within the K8s Cluster. To expose them outside the K8s Cluster, the services have to be upgraded into a LoadBalancer type. This can be done by setting the spec.managementNode.enableLoadBalancer and spec.mysqlNode.enableLoadBalancer fields of the NdbCluster resource to true. These options can be enabled when the NdbCluster resource object is created or via an update to the object when the MySQL Cluster is already running.

          Once enabled, the previously ClusterIP type example-ndb-mgmd and example-ndb-mysqld services will be upgraded to LoadBalancer type, and they will be available at the external IP assigned to them by the cloud provider.

          Another way to access these services without enabling the LoadBalancers support is to use the kubectl port-forward command.

          In both ways, only the mysql and ndb_mgm clients work. Any NDB tool which uses the NDBAPI to connect to the MySQL data nodes will not work as expected from outside the K8s Cluster.

          A demonstration is available at [connect-from-outside-k8s-demo.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/connect-from-outside-k8s-demo.md).
          ## Updating MySQL Cluster Configuration

          Once a MySQL Cluster has been deployed by the NDB Operator inside the K8s Cluster, its configuration can be updated by updating the spec of the NdbCluster resource object that represents it. The NDB Operator will pick up any changes made to the NdbCluster resource object and will apply the updated configuration to the respective MySQL Cluster. The NDB Operator takes care of updating the MySQL Cluster config file and restarting all the management/data nodes if required. When the update is being handled by the NdbOperator, it will set the UpToDate condition of the NdbCluster resource object to false and will not accept any further updates until it completes the current one.

          The kubectl get ndb example-ndb command will report the UpToDate status of an NdbCluster. It can be used to check if the current update has been completed.
          ```
          $ kubectl get ndb example-ndb
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE   UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:1/2    Ready:2/2 10m   False
          ```

          The UpToDate condition can also be used to wait for the operator to complete the update.
          ```
          kubectl wait --for=condition=UpToDate ndb example-ndb --timeout=10m
          ```

          Once the update is complete, the UpToDate condition will be set back to true by the operator.
          ```
          $ kubectl get ndb example-ndb
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE      UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:2/2    Ready:2/2 10m50s   True
          ```

          ## Delete a MySQL Cluster

          To stop and remove the MySQL Cluster running inside the K8s Cluster, delete the NdbCluster resource object.
          ```
          kubectl delete -f docs/examples/example-ndb.yaml
          ```

          (or)
          ```
          kubectl delete ndb example-ndb
          ```

          This will delete all the data, the pods running the MySQL Cluster nodes and also delete all other associated K8s resources created by the NDB Operator.
        displayName: Ndb Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - ndb
        - in-memory
        links:
        - name: Document
          url: https://dev.mysql.com/doc/ndb-operator/en/
        maintainers:
        - email: paulo.p.silva@oracle.com
          name: paulofasilva
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: MySQL
          url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
        relatedImages:
        - container-registry.oracle.com/mysql/community-ndb-operator@sha256:b562dc3303ffae9f4dab1d73fd0fe05833bdaa8b731af4227da81af617c75c6b
        version: 1.6.2
      entries:
      - name: ndb-operator.v1.6.2
        version: 1.6.2
      - name: ndb-operator.v1.6.1
        version: 1.6.1
      - name: ndb-operator.v1.6.0
        version: 1.6.0
      name: alpha
    defaultChannel: alpha
    packageName: ndb-operator
    provider:
      name: MySQL
      url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: mercury-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mercury-operator.1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:1.0.1",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CustomerCreditRating",
                  "expose": ["http"]
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:1.0.1
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:1.0.1
        version: 1.0.1
      entries:
      - name: mercury-operator.1.0.1
        version: 1.0.1
      name: 1.0.x
    - currentCSV: mercury-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_CREDIT_RATING",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-offer",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-offer-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_OFFER",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-party-routing-profile",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/party-routing-profile-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "PARTY_ROUTING_PROFILE",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-example-sdi": {
                "description": "Example Service Domain Infrastructure"
            },
            {
             "example-customer-credit-rating": {
               "description": "Example Customer Credit Rating Service Domain"
            },
            {
             "example-customer-offer": {
               "description": "Example Customer Offer Service Domain"
            },
            {
             "example-party-routing-profile": {
               "description": "Example Party Routing Profile Service Domain"
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:v1.0.2
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - description: Represents a Kafka cluster
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Integration is the Schema for the Camel K integrations API
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:v1.0.2
        version: 1.0.2
      entries:
      - name: mercury-operator.v1.0.2
        version: 1.0.2
      name: dev
    - currentCSV: mercury-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_CREDIT_RATING",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-offer",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-offer-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_OFFER",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-party-routing-profile",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/party-routing-profile-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "PARTY_ROUTING_PROFILE",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-example-sdi": {
                "description": "Example Service Domain Infrastructure"
            },
            {
             "example-customer-credit-rating": {
               "description": "Example Customer Credit Rating Service Domain"
            },
            {
             "example-customer-offer": {
               "description": "Example Customer Offer Service Domain"
            },
            {
             "example-party-routing-profile": {
               "description": "Example Party Routing Profile Service Domain"
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:v1.0.2
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - description: Represents a Kafka cluster
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Integration is the Schema for the Camel K integrations API
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:v1.0.2
        version: 1.0.2
      entries:
      - name: mercury-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: 1.0.x
    packageName: mercury-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-ati-updates
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-ati-updates.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ati.loadcore.keysight.com/v1alpha1",
                "kind": "Updates",
                "metadata": {
                  "name": "updates-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "installWait": {
                    "enabled": true,
                    "image": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:40:21Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Updates
            name: updates.ati.loadcore.keysight.com
            version: v1alpha1
        description: Keysight ATI Updates Operator
        displayName: Keysight ATI Updates Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Ati Updates
          url: https://keysight-ati-updates.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-ati-updates@sha256:abda76eb50d77e5a95191c9dbb21eefe64af5b794511ceafdfda007c15449c6e
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        version: 0.1.21
      entries:
      - name: keysight-ati-updates.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-ati-updates
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: CyberArk
      provider-url: https://www.cyberark.com/
    name: vcp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vcp-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "installer.venafi.com/v1alpha1",
                "kind": "VenafiInstall",
                "metadata": {
                  "name": "install-venafi-enhanced-issuer"
                },
                "spec": {
                  "certManager": {
                    "install": true
                  },
                  "globals": {
                    "imagePullSecretNames": [
                      "venafi"
                    ]
                  },
                  "venafiConnection": {
                    "install": true
                  },
                  "venafiEnhancedIssuer": {
                    "install": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:9da79fd0941c21e433f22aaaccc43ebbd4646e685aea0050194954d8b4019c05
          createdAt: "2024-05-30T11:35:41Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              VenafiInstall is an object for installing Venafi components onto a Kubernetes
              cluster
            displayName: VenafiInstall
            kind: VenafiInstall
            name: venafiinstalls.installer.venafi.com
            version: v1alpha1
        description: |
          Venafi Control Plane Operator for Red Hat OpenShift simplifies deployment of Venafi's enterprise components and dependencies.

          The enterprise components include:
          - Enterprise cert-manager: An enhanced version of cert-manager designed for application certificate management within your clusters.
          - Venafi Enhanced Issuer: Enables seamless certificate enrollment for your clusters directly from the Venafi Control Plane.
          - Venafi Firefly: A high-performance, lightweight issuer for certificates, centrally managed by the Venafi Control Plane.
          - Venafi Kubernetes Agent: Provides real-time visibility into certificates and Kubernetes resources directly within the Venafi Control Plane.
        displayName: Venafi Control Plane Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - venafi
        - vcp-operator
        links:
        - name: Documentation
          url: https://docs.venafi.cloud/
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Venafi
          url: https://venafi.com
        relatedImages:
        - registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:9da79fd0941c21e433f22aaaccc43ebbd4646e685aea0050194954d8b4019c05
        version: 1.0.0
      entries:
      - name: vcp-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: vcp-operator.v1.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "installer.venafi.com/v1alpha1",
                "kind": "VenafiInstall",
                "metadata": {
                  "name": "install-venafi-enhanced-issuer"
                },
                "spec": {
                  "certManager": {
                    "install": true
                  },
                  "globals": {
                    "imagePullSecretNames": [
                      "venafi"
                    ]
                  },
                  "venafiConnection": {
                    "install": true
                  },
                  "venafiEnhancedIssuer": {
                    "install": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:eb2b06d191dbfbb02332375a7bca6bc010d5410e53b3bc5fb8c320101292408c
          createdAt: "2025-07-22T06:44:30Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              VenafiInstall is an object for installing Venafi components onto a Kubernetes
              cluster
            displayName: VenafiInstall
            kind: VenafiInstall
            name: venafiinstalls.installer.venafi.com
            version: v1alpha1
        description: |
          The operator, formerly known as the Venafi Control Plane Operator, simplifies the deployment of enterprise components part of the CyberArk Certificate Manager for Kubernetes.

          The components include:
          - cert-manager: An enterprise distribution of cert-manager designed for application certificate management within your clusters.
          - Enterprise Issuer: Enables seamless certificate enrollment directly from the Certificate Manager.
          - Workload Identity Manager: A high-performance issuer for certificates, centrally managed by the Certificate Manager.
          - Agent: Provides real-time visibility into cluster resources directly within the Certificate Manager.
        displayName: CyberArk Certificate Manager Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cyberark
        - vcp-operator
        links:
        - name: Documentation
          url: https://docs.venafi.cloud/vaas/k8s-components/c-vcpo-overview/
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:eb2b06d191dbfbb02332375a7bca6bc010d5410e53b3bc5fb8c320101292408c
        version: 1.8.1
      entries:
      - name: vcp-operator.v1.8.1
        version: 1.8.1
      - name: vcp-operator.v1.8.0
        version: 1.8.0
      - name: vcp-operator.v1.7.0
        version: 1.7.0
      - name: vcp-operator.v1.6.0
        version: 1.6.0
      - name: vcp-operator.v1.5.0
        version: 1.5.0
      - name: vcp-operator.v1.4.0
        version: 1.4.0
      - name: vcp-operator.v1.3.0
        version: 1.3.0
      - name: vcp-operator.v1.2.1
        version: 1.2.1
      - name: vcp-operator.v1.2.0
        version: 1.2.0
      - name: vcp-operator.v1.1.1
        version: 1.1.1
      name: stable
    defaultChannel: stable
    packageName: vcp-operator
    provider:
      name: CyberArk
      url: https://www.cyberark.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE Ezmeral
      provider-url: ""
    name: hpe-ezmeral-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-ezmeral-csi-operator.v1.0.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ezmeral.hpe.com/v1",
                "kind": "HPEEzmeralCSIDriver",
                "metadata": {
                  "name": "hpeezmeralcsidriver"
                },
                "spec": {
                  "controllerImage": "registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03",
                  "nodeImage": "registry.connect.redhat.com/maprtech/csi-kdfplugin@sha256:ffd273ee2e608d2ef64fd3b2bcbb02d5f0a6c173ed96a15eaf5549ea45fea8d1",
                  "pullPolicy": "Always"
                }
              },
              {
                "apiVersion": "ezmeral.hpe.com/v1",
                "kind": "HPEEzmeralNFSCSIDriver",
                "metadata": {
                  "name": "hpeezmeralnfscsidriver"
                },
                "spec": {
                  "controllerImage": "registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03",
                  "nodeImage": "registry.connect.redhat.com/maprtech/csi-nfsplugin@sha256:a5c954f62f1395bafe2d68454ac604bca4be97573a3482fcbd8e76b6361cc08a",
                  "pullPolicy": "Always"
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator@sha256:c9c243c51cd760c2c2a66b489807a5e22053fcbeb94fa29afd10d9c91e2d3908
          createdAt: "2025-06-25T11:51:05Z"
          description: A Container Storage Interface (CSI) driver for HPE Ezmeral
            Data Fabric platform. The CSI driver allows you to use HPE Ezmeral Data
            Fabric with your preferred container orchestrator.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mapr/mapr-csi
          support: HPE Ezmeral Data Fabric
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE Ezmeral CSI Driver (FUSE)
            displayName: HPEEzmeralCSIDriver
            kind: HPEEzmeralCSIDriver
            name: hpeezmeralcsidrivers.ezmeral.hpe.com
            version: v1
          - description: HPE Ezmeral CSI Driver (NFS)
            displayName: HPEEzmeralNFSCSIDriver
            kind: HPEEzmeralNFSCSIDriver
            name: hpeezmeralnfscsidrivers.ezmeral.hpe.com
            version: v1
        description: |
          The HPE Ezmeral CSI Operator for Kubernetes packages, deploys, manages, upgrades the HPE Ezmeral CSI Driver on Kubernetes and OpenShift for dynamic provisioning of persistent volumes on HPE Ezmeral Data Fabric platform.
          The HPE Ezmeral CSI Driver for Kubernetes leverages HPE Ezmeral Data Fabric platform to provide scalable and persistent storage for stateful applications. Please refer to our CSI driver [documentation](https://docs.datafabric.hpe.com/home/CSIdriver/csi_overview.html) for full list of supported CSI features.
          ## Installation
          Refer to the HPE Ezmeral CSI Operator for Kubernetes [documentation](https://docs.datafabric.hpe.com/home/CSIdriver/csi_installing_operator.html)
        displayName: HPE Ezmeral Data Fabric CSI Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE Ezmeral
        - HPE Ezmeral Data Fabric
        - Storage
        - FileSystem
        - BlockVolume
        links:
        - name: Documentation
          url: https://docs.datafabric.hpe.com/home/CSIdriver/csi_overview.html
        maintainers:
        - email: ezmeral-csi-support@hpe.com
          name: HPE Ezmeral Data Fabric
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: HPE Ezmeral
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6ba961c2c2a29750c0132fe6dd6fa9f6001010afbc5f19b98add87b31b54bcf6
        - registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator@sha256:c9c243c51cd760c2c2a66b489807a5e22053fcbeb94fa29afd10d9c91e2d3908
        - registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03
        - registry.connect.redhat.com/maprtech/csi-kdfplugin@sha256:ffd273ee2e608d2ef64fd3b2bcbb02d5f0a6c173ed96a15eaf5549ea45fea8d1
        - registry.connect.redhat.com/maprtech/csi-nfsplugin@sha256:a5c954f62f1395bafe2d68454ac604bca4be97573a3482fcbd8e76b6361cc08a
        version: 1.0.18
      entries:
      - name: hpe-ezmeral-csi-operator.v1.0.18
        version: 1.0.18
      name: stable
    defaultChannel: stable
    packageName: hpe-ezmeral-csi-operator
    provider:
      name: HPE Ezmeral
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzo-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzo-operator.v3.1.200
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzo.cambridgesemantics.com/v2",
                "kind": "Anzo",
                "metadata": {
                  "name": "no-persistence"
                },
                "spec": {
                  "jsonActivation": true,
                  "jvmMemory": "4096m",
                  "nodeConfig": {
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "app": "anzo"
                        }
                      },
                      "serviceName": "anzo-no-persistence",
                      "template": {
                        "metadata": {
                          "annotations": {
                            "cluster-autoscaler.kubernetes.io/safe-to-evict": "false"
                          },
                          "labels": {
                            "app": "anzo"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "registry.connect.redhat.com/cambridgesemantics/anzo@sha256:5b192604338f602fdabe55fecefd9eb82400a77c5aa7f04bd8e54aead8644c6d",
                              "imagePullPolicy": "Always",
                              "name": "anzo",
                              "resources": {
                                "limits": {
                                  "cpu": "2.0",
                                  "memory": "8.0Gi"
                                },
                                "requests": {
                                  "cpu": "2.0",
                                  "memory": "8.0Gi"
                                }
                              }
                            }
                          ],
                          "initContainers": [
                            {
                              "name": "anzo-init",
                              "resources": {
                                "limits": {
                                  "cpu": "0.5",
                                  "memory": "2Gi"
                                },
                                "requests": {
                                  "cpu": "0.5",
                                  "memory": "2Gi"
                                }
                              }
                            }
                          ],
                          "securityContext": {
                            "fsGroup": 1001
                          },
                          "serviceAccountName": "anzo-operator"
                        }
                      }
                    }
                  },
                  "role": "AnzoServer"
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/anzo-operator@sha256:4c0c8bf7acbfbb96f15459a20364608e5d0a0933c6326a89d219666ee0d88ca2
          createdAt: "2025-07-09T06:52:12Z"
          description: kubernetes operator for Altair Graph Studio
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Altair Customer Support
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoBackup is the Schema for the anzobackups API
            displayName: Altair Graph Studio Backup
            kind: AnzoBackup
            name: anzobackups.anzobackup.cambridgesemantics.com
            version: v1beta1
          - description: Altair Graph Studio is the Schema for the anzos API
            displayName: Altair Graph Studio Operator
            kind: Anzo
            name: anzos.anzo.cambridgesemantics.com
            version: v2
        description: |-
          The Altair Graph Studio Operator provides the way to install and configure an altair graph studio server and agent setup on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v3.1.2/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/deploy-k8s.htm )

          ### Support

          We offer Support to our customers [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Studio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        links:
        - name: Altair Graph Studio Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v3.1.2/README_openshift.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzo-operator@sha256:4c0c8bf7acbfbb96f15459a20364608e5d0a0933c6326a89d219666ee0d88ca2
        - registry.connect.redhat.com/cambridgesemantics/anzo@sha256:5b192604338f602fdabe55fecefd9eb82400a77c5aa7f04bd8e54aead8644c6d
        version: 3.1.200
      entries:
      - name: anzo-operator.v3.1.200
        version: 3.1.200
      - name: anzo-operator.v3.1.106
        version: 3.1.106
      - name: anzo-operator.v3.1.105
        version: 3.1.105
      name: stable
    defaultChannel: stable
    packageName: anzo-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: ""
    name: power-monitoring-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: power-monitoring-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "Kepler",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "kepler",
                    "app.kubernetes.io/name": "kepler",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "kepler"
                },
                "spec": {
                  "exporter": {
                    "deployment": {
                      "port": 9103
                    },
                    "redfish": null
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: registry.redhat.io/openshift-power-monitoring/power-monitoring-rhel9-operator@sha256:5eec15f6fe9bd4bdb1f5fb8057dc67c448c3711bcc7a0a4760f0d90c07688d1e
          createdAt: "2025-02-14T06:23:56Z"
          description: Deploys and Manages Kepler on OpenShift
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/internal-objects: |-
            [
              "keplerinternals.kepler.system.sustainable.computing.io"
            ]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift-power-monitoring/power-monitoring-operator
          support: Team Power Monitoring (team-powermon@redhat.com)
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KeplerInternal is the Schema for internal/unsupported API
            displayName: KeplerInternal
            kind: KeplerInternal
            name: keplerinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: Kepler is the Schema for the keplers API
            displayName: Kepler
            kind: Kepler
            name: keplers.kepler.system.sustainable.computing.io
            version: v1alpha1
        description: |
          ### Introduction
          Kepler (Kubernetes Efficient Power Level Exporter) uses eBPF to probe
          energy-related system stats and exports them as Prometheus metrics.

          ### NOTE
          To view the metrics exposed by Kepler, ensure that
          [User Project Monitoring](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/monitoring/configuring-user-workload-monitoring)
          is enabled in your cluster.

          ### Documentation
          Documentation and installation guide can be found below:
            * [Installation Guide](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/power_monitoring/installing-power-monitoring)
        displayName: Power monitoring for Red Hat OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kepler
        - power
        - energy
        - monitoring
        links:
        - name: Installation Guide
          url: https://docs.redhat.com/documentation/openshift_container_platform/latest/html/power_monitoring/installing-power-monitoring
        maintainers:
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: vimalkum@redhat.com
          name: Vimal Kumar
        - email: vprashar@redhat.com
          name: Vibhu Prashar
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift-power-monitoring/power-monitoring-rhel9-operator@sha256:5eec15f6fe9bd4bdb1f5fb8057dc67c448c3711bcc7a0a4760f0d90c07688d1e
        - registry.redhat.io/openshift-power-monitoring/kepler-rhel9@sha256:27d3cb0a98f1cffd649745e3886ad92995d83a0de6823ddc2d770bf036e2eeaf
        version: 0.4.0
      entries:
      - name: power-monitoring-operator.v0.4.0
        version: 0.4.0
      - name: power-monitoring-operator.v0.3.0
        version: 0.3.0
      - name: power-monitoring-operator.v0.2.0
        version: 0.2.0
      - name: power-monitoring-operator.v0.1.0
        version: 0.1.0
      name: tech-preview
    defaultChannel: tech-preview
    packageName: power-monitoring-operator
    provider:
      name: Red Hat, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtc-operator.v1.7.19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigrationController",
                  "metadata": {
                    "name": "migration-controller",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "azure_resource_group": "",
                    "cluster_name": "host",
                    "migration_velero": true,
                    "migration_controller": true,
                    "migration_ui": true,
                    "migration_log_reader": true,
                    "olm_managed": true,
                    "restic_timeout": "1h",
                    "mig_pv_limit": "100",
                    "mig_pod_limit": "100",
                    "mig_namespace_limit": "10"
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageMigration",
                  "metadata": {
                    "name": "directimagemigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageStreamMigration",
                  "metadata": {
                    "name": "directimagestreammigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigAnalytic",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "plan": {
                      "name": "foo",
                      "namespace": "openshift-migration"
                    },
                    "analyzeK8SResources": true,
                    "analyzeImageCount": true,
                    "analyzePVCapacity": true
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigCluster",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectVolumeMigration",
                  "metadata": {
                    "name": "direct",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                      "kind": "DirectVolumeMigrationProgress",
                      "metadata": {
                        "name": "direct-progress",
                        "namespace": "openshift-migration"
                      },
                      "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigPlan",
                  "metadata": {
                    "name": "migplan",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigMigration",
                  "metadata": {
                    "name": "migmigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigStorage",
                  "metadata": {
                    "name": "migstorage",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:89ba5ccb1cd9c44b8c8d2189f1fc6b04e780d46b19ee8a22e20c0b8c9d94d91f
          createdAt: "2025-04-28T19:55:38Z"
          description: Facilitates migration of container workloads from OpenShift
            3.x to OpenShift 4.x
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.5.0 <1.7.19'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-migration
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:3b6168c6df0d86d76ae4de91755c631b7f6a1f84286257d18c97d5a6af312291
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/mig-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: a request to migrate PVCs from a set of namespaces directly
              to a target
            displayName: DirectVolumeMigration
            kind: DirectVolumeMigration
            name: directvolumemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate watch the progress of rsync pod migrating
              a PVC to a target
            displayName: DirectVolumeMigrationProgress
            kind: DirectVolumeMigrationProgress
            name: directvolumemigrationprogresses.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a set of namespaces
              directly to a target
            displayName: DirectImageMigration
            kind: DirectImageMigration
            name: directimagemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a single imagestream
              directly to a target
            displayName: DirectImageStreamMigration
            kind: DirectImageStreamMigration
            name: directimagestreammigrations.migration.openshift.io
            version: v1alpha1
          - description: A request to Analyze a MigPlan and its namespaces
            displayName: MigAnalytic
            kind: MigAnalytic
            name: miganalytics.migration.openshift.io
            version: v1alpha1
          - description: A cluster defined for migration
            displayName: MigCluster
            kind: MigCluster
            name: migclusters.migration.openshift.io
            version: v1alpha1
          - description: A migration process
            displayName: MigMigration
            kind: MigMigration
            name: migmigrations.migration.openshift.io
            version: v1alpha1
          - description: A migration plan
            displayName: MigPlan
            kind: MigPlan
            name: migplans.migration.openshift.io
            version: v1alpha1
          - description: A migration contoller deployment
            displayName: MigrationController
            kind: MigrationController
            name: migrationcontrollers.migration.openshift.io
            version: v1alpha1
          - description: A migration hook
            displayName: MigHook
            kind: MigHook
            name: mighooks.migration.openshift.io
            version: v1alpha1
          - description: A migration storage location
            displayName: MigStorage
            kind: MigStorage
            name: migstorages.migration.openshift.io
            version: v1alpha1
        description: |-
          The Migration Toolkit for Containers Operator enables installation of the application migration tool components.

          After you have installed the Operator, you can install components by creating a MigrationController CR.

          By default, the Operator installs the following components on an OpenShift Container Platform 4 target cluster:

          * velero and restic, to back up data on the source cluster and restore it on the target cluster
          * mig-controller, to coordinate the backup and restore processes
          * mig-ui, the web console for managing migrations

          If you experience any issues or have feature requests, please file an [issue](https://github.com/konveyor/mig-operator/issues)

          For more information, see the [OpenShift Container Platform Migration Guide](https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html).
        displayName: Migration Toolkit for Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - crane
        - mtc
        links:
        - name: Migrations
          url: https://github.com/konveyor/mig-controller
        - name: Documentation
          url: https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html
        - name: Crane Operator
          url: https://github.com/konveyor/mig-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhmtc/openshift-migration-hook-runner-rhel8@sha256:c8c932a486d9b83e5a93290058d9e84691b87cd4161f8132a1bb6de9a360605e
        - registry.redhat.io/rhmtc/openshift-migration-velero-restic-restore-helper-rhel8@sha256:932216fd08d8ad97baf48ac7b8c989806d30a191337c650df37da537c153dc17
        - registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:3b6168c6df0d86d76ae4de91755c631b7f6a1f84286257d18c97d5a6af312291
        - registry.redhat.io/rhmtc/openshift-migration-velero-rhel8@sha256:bc010bc31e8b553f86aacbca39e28a95e74994c56e85c9fc1b8b80d0791ecb16
        - registry.redhat.io/rhmtc/openshift-migration-log-reader-rhel8@sha256:a0c8ab3a5bbc78c158c2ccccb1d06430a20f40358649196aa4e7311e54f9bdf8
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-aws-rhel8@sha256:023465788745625ad3b4dc0574e35b93d6ac8eea17be2ad1fc88959149a0df2c
        - registry.redhat.io/rhmtc/openshift-migration-ui-rhel8@sha256:66b288170dca8007382d13dff3f2e1868e409fb8002e04ec1fc5c480f868d9b6
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-gcp-rhel8@sha256:cc70c5a63af77948760bc42bf695a5e4eb0bce6285d6b350bba89a53f4b4e7e9
        - registry.redhat.io/rhmtc/openshift-migration-registry-rhel8@sha256:49485dfe708fe1294dec3f0bf34fedc38845a22f91915bba523d728a586b0c30
        - registry.redhat.io/rhmtc/openshift-migration-controller-rhel8@sha256:9b1490b04fc235191515fcd07d7af2e81f846d865016831e17872bb82bf4eaf5
        - registry.redhat.io/rhmtc/openshift-migration-rsync-transfer-rhel8@sha256:23a2bc5ad0f99fc0e7517a0172fdf3597fa619543b0192757983219a578f4ee2
        - registry.redhat.io/rhmtc/openshift-velero-plugin-rhel8@sha256:02698c655a01df2237aee6c570db96f13c0018cee2ae58b96b9570d93c354617
        - registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:89ba5ccb1cd9c44b8c8d2189f1fc6b04e780d46b19ee8a22e20c0b8c9d94d91f
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-microsoft-azure-rhel8@sha256:dad8fa4dbce3b4562bf06756fd74c2007812ec4edbc308e4517ff123c5519767
        version: 1.7.19
      entries:
      - name: mtc-operator.v1.7.19
        version: 1.7.19
      - name: mtc-operator.v1.7.18
        version: 1.7.18
      - name: mtc-operator.v1.7.17
        version: 1.7.17
      - name: mtc-operator.v1.7.16
        version: 1.7.16
      - name: mtc-operator.v1.7.15
        version: 1.7.15
      - name: mtc-operator.v1.7.14
        version: 1.7.14
      - name: mtc-operator.v1.7.13
        version: 1.7.13
      - name: mtc-operator.v1.7.12
        version: 1.7.12
      - name: mtc-operator.v1.7.11
        version: 1.7.11
      - name: mtc-operator.v1.7.10
        version: 1.7.10
      - name: mtc-operator.v1.7.9
        version: 1.7.9
      - name: mtc-operator.v1.7.8
        version: 1.7.8
      - name: mtc-operator.v1.7.7
        version: 1.7.7
      - name: mtc-operator.v1.7.6
        version: 1.7.6
      - name: mtc-operator.v1.7.5
        version: 1.7.5
      - name: mtc-operator.v1.7.4
        version: 1.7.4
      - name: mtc-operator.v1.7.3
        version: 1.7.3
      - name: mtc-operator.v1.7.2
        version: 1.7.2
      - name: mtc-operator.v1.7.1
        version: 1.7.1
      - name: mtc-operator.v1.7.0
        version: 1.7.0
      name: release-v1.7
    - currentCSV: mtc-operator.v1.8.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigrationController",
                  "metadata": {
                    "name": "migration-controller",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "azure_resource_group": "",
                    "cluster_name": "host",
                    "migration_velero": true,
                    "migration_controller": true,
                    "migration_ui": true,
                    "migration_log_reader": true,
                    "olm_managed": true,
                    "restic_timeout": "1h",
                    "mig_pv_limit": "100",
                    "mig_pod_limit": "100",
                    "mig_namespace_limit": "10"
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageMigration",
                  "metadata": {
                    "name": "directimagemigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageStreamMigration",
                  "metadata": {
                    "name": "directimagestreammigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigAnalytic",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "plan": {
                      "name": "foo",
                      "namespace": "openshift-migration"
                    },
                    "analyzeK8SResources": true,
                    "analyzeImageCount": true,
                    "analyzePVCapacity": true
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigCluster",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectVolumeMigration",
                  "metadata": {
                    "name": "direct",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                      "kind": "DirectVolumeMigrationProgress",
                      "metadata": {
                        "name": "direct-progress",
                        "namespace": "openshift-migration"
                      },
                      "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigPlan",
                  "metadata": {
                    "name": "migplan",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigMigration",
                  "metadata": {
                    "name": "migmigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigStorage",
                  "metadata": {
                    "name": "migstorage",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:8a44fe2aea3e842dcccb1e20df4818f4ca573fe13f51427ae8b603d0000a9411
          createdAt: "2025-08-22T15:27:29Z"
          description: Facilitates migration of container workloads from OpenShift
            3.x to OpenShift 4.x
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.5.0 <1.8.10'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-migration
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:f9492269b75608057fbc5edd53603807f1f80cd8581dea12032c88e04cf8f4cb
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/mig-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: a request to migrate PVCs from a set of namespaces directly
              to a target
            displayName: DirectVolumeMigration
            kind: DirectVolumeMigration
            name: directvolumemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate watch the progress of rsync pod migrating
              a PVC to a target
            displayName: DirectVolumeMigrationProgress
            kind: DirectVolumeMigrationProgress
            name: directvolumemigrationprogresses.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a set of namespaces
              directly to a target
            displayName: DirectImageMigration
            kind: DirectImageMigration
            name: directimagemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a single imagestream
              directly to a target
            displayName: DirectImageStreamMigration
            kind: DirectImageStreamMigration
            name: directimagestreammigrations.migration.openshift.io
            version: v1alpha1
          - description: A request to Analyze a MigPlan and its namespaces
            displayName: MigAnalytic
            kind: MigAnalytic
            name: miganalytics.migration.openshift.io
            version: v1alpha1
          - description: A cluster defined for migration
            displayName: MigCluster
            kind: MigCluster
            name: migclusters.migration.openshift.io
            version: v1alpha1
          - description: A migration process
            displayName: MigMigration
            kind: MigMigration
            name: migmigrations.migration.openshift.io
            version: v1alpha1
          - description: A migration plan
            displayName: MigPlan
            kind: MigPlan
            name: migplans.migration.openshift.io
            version: v1alpha1
          - description: A migration contoller deployment
            displayName: MigrationController
            kind: MigrationController
            name: migrationcontrollers.migration.openshift.io
            version: v1alpha1
          - description: A migration hook
            displayName: MigHook
            kind: MigHook
            name: mighooks.migration.openshift.io
            version: v1alpha1
          - description: A migration storage location
            displayName: MigStorage
            kind: MigStorage
            name: migstorages.migration.openshift.io
            version: v1alpha1
        description: |-
          The Migration Toolkit for Containers Operator enables installation of the application migration tool components.

          After you have installed the Operator, you can install components by creating a MigrationController CR.

          By default, the Operator installs the following components on an OpenShift Container Platform 4 target cluster:

          * velero and restic, to back up data on the source cluster and restore it on the target cluster
          * mig-controller, to coordinate the backup and restore processes
          * mig-ui, the web console for managing migrations

          If you experience any issues or have feature requests, please file an [issue](https://github.com/konveyor/mig-operator/issues)

          For more information, see the [OpenShift Container Platform Migration Guide](https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html).
        displayName: Migration Toolkit for Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - crane
        - mtc
        links:
        - name: Migrations
          url: https://github.com/konveyor/mig-controller
        - name: Documentation
          url: https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html
        - name: Crane Operator
          url: https://github.com/konveyor/mig-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhmtc/openshift-migration-registry-rhel8@sha256:accef2a0dd69b254bd120232e5dc86a56aa8f25bcfdba3f9bab7e4127c157762
        - registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:f9492269b75608057fbc5edd53603807f1f80cd8581dea12032c88e04cf8f4cb
        - registry.redhat.io/rhmtc/openshift-migration-controller-rhel8@sha256:c776387c662e21aa61378d0136b2495826e1c157548f9ca9c95d044af38ee169
        - registry.redhat.io/rhmtc/openshift-migration-hook-runner-rhel8@sha256:adaa4f1afefb4d5853713821cb6e7de0d9f5a640f0243f20b41a44be84707caf
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-mtc-rhel8@sha256:5f46ba04146438f2abd6905468d2ecbeb19374a83771d364904ac2ee6afdbefa
        - registry.redhat.io/rhmtc/openshift-migration-openvpn-rhel8@sha256:d21b99d8edefa7a1bab30c85969ec7d7560d06be308bd787e698e409d54cbc19
        - registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:8a44fe2aea3e842dcccb1e20df4818f4ca573fe13f51427ae8b603d0000a9411
        - registry.redhat.io/rhmtc/openshift-migration-log-reader-rhel8@sha256:4f76bcf3df3bb7073b8aabd3da072cacfec05ecf4bfb47e23340074756f7fc30
        - registry.redhat.io/rhmtc/openshift-migration-rsync-transfer-rhel8@sha256:442e7c7f2cc075bf468403ace8846c96ad2fe43eb7f6562d566caa24e4f86873
        - registry.redhat.io/rhmtc/openshift-migration-ui-rhel8@sha256:86e747e0330b0156379d273d0c29c13a65dafbf4f536656723e2bc746872286c
        version: 1.8.10
      entries:
      - name: mtc-operator.v1.8.10
        version: 1.8.10
      - name: mtc-operator.v1.8.9
        version: 1.8.9
      - name: mtc-operator.v1.8.8
        version: 1.8.8
      - name: mtc-operator.v1.8.7
        version: 1.8.7
      - name: mtc-operator.v1.8.6
        version: 1.8.6
      - name: mtc-operator.v1.8.5
        version: 1.8.5
      - name: mtc-operator.v1.8.4
        version: 1.8.4
      - name: mtc-operator.v1.8.3
        version: 1.8.3
      - name: mtc-operator.v1.8.2
        version: 1.8.2
      - name: mtc-operator.v1.8.1
        version: 1.8.1
      - name: mtc-operator.v1.8.0
        version: 1.8.0
      name: release-v1.8
    defaultChannel: release-v1.8
    packageName: mtc-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Performance
      provider-url: ""
    name: cluster-impairment-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-impairment-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "all-impairments"
                },
                "spec": {
                  "duration": 480,
                  "egress": {
                    "bandwidth": 10000,
                    "latency": 50,
                    "loss": 0.02
                  },
                  "ingress": {
                    "bandwidth": 10000,
                    "latency": 50,
                    "loss": 0.02
                  },
                  "interfaces": [
                    "ens2f0"
                  ],
                  "linkFlapping": {
                    "downTime": 30,
                    "enable": true,
                    "upTime": 120
                  },
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "duplication"
                },
                "spec": {
                  "duration": 60,
                  "egress": {
                    "duplication": 25
                  },
                  "ingress": {
                    "duplication": 25
                  },
                  "interfaces": [
                    "ens3"
                  ],
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "two-min-flap"
                },
                "spec": {
                  "duration": 480,
                  "interfaces": [
                    "ens2f0"
                  ],
                  "linkFlapping": {
                    "downTime": 120,
                    "enable": true,
                    "upTime": 120
                  },
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "typical-scenario"
                },
                "spec": {
                  "duration": 30,
                  "egress": {
                    "latency": 50
                  },
                  "ingress": {
                    "latency": 50
                  },
                  "interfaces": [
                    "ens2f0",
                    "eno1"
                  ],
                  "loss": 0.02,
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "uneven-latency"
                },
                "spec": {
                  "duration": 60,
                  "egress": {
                    "latency": 100
                  },
                  "ingress": {
                    "latency": 10
                  },
                  "interfaces": [
                    "ens2f0"
                  ],
                  "nodeSelector": {
                    "key": "node-role.kubernetes.io/worker",
                    "value": ""
                  },
                  "startDelay": 5
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Developer Tools
          containerImage: quay.io/redhat-performance/cluster-impairment-operator:1.0.4
          createdAt: "2021-08-11 12:15:00"
          description: Run Impairments on Kubernetes/Openshift Nodes
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/redhat-performance/cluster-impairment-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterImpairment
            name: clusterimpairments.apps.redhat.com
            version: v1alpha1
        description: "cluster-impairment-operator is an operator designed to automate
          the application of impairments to the nodes in a cluster.\n\n## Features\n\nTraffic
          Direction:\n\n* Egress impairments affect the traffic going out.    \n*
          Ingress impairments affect the traffic coming in. (Ingress impairments require
          the kernel module IFB.)\n\n### Impairments\n\n| Impairment    | Description
          \                               | Unit    | Uses Netem | Correlation Supported
          |\n|---------------|--------------------------------------------|---------|------------|-----------------------|\n|
          Bandwidth     | The bandwidth limit                        | kbit/s  | Yes
          \       | No                    |\n| Latency       | The delay of the packets
          \                  | ms      | Yes        | Yes                   |\n| Packet
          Loss   | The percent of packets that are dropped    | percent | Yes        |
          Yes                   |\n| Corruption    | The percent of packets that are
          corrupted  | percent | Yes        | Yes                   |\n| Duplication
          \  | The percent of packets that are duplicated | percent | Yes        |
          Yes                   |\n| Link Flapping | Turns the interface on and off
          \            | bool    | No         | No                    |\n\n\nOn the
          tested environment (RHEL CoreOS 48.84), the impairments can be used alongside
          link flapping.\n\n\n#### Latency Options\n\nIn addition to simply delaying
          the packets, there are advanced latency options.\n\n##### Jitter\n\nSemi-randomly
          adds or subtracts from the latency according to the distribution up to the
          amount specified.\nIf latency is 100ms, and jitter is 10ms, the actual latency
          will vary 100±10ms.\n\n\n##### Distribution\n\nThe distribution of the jitter.
          The options are:\n* Normal\n* Uniform\n* Pareto\n* Paretonormal\n\n#####
          Jitter correlation\n\nThe percent chance that the next latency's jitter
          value will correlate with the preceeding latency.\n\n##### Reorder\n\nThe
          percentage of packets that are not delayed, causing a reorder due to them
          being sent before the delayed ones.\nApplying jitter itself has the potential
          to also cause reordering.\n\n##### Reorder correlation\n\nThe percet chance
          that the value for the next reorder will correlate with the preceeding value.\n\n####
          Loss options\n\n##### Correlation\n\nThe percent chance that the previous
          loss value for a packet correlates with the loss value for the next packet.\n\n####
          Corruption options\n\n###### Corruption\n\nThe percent chance that each
          packet will be corrupted.\n\n##### Corruption correlation\n\nThe percet
          chance that the value for the next corruption will correlate with the preceeding
          value.\n\n#### Duplication options\n\n##### Duplication\n\nThe percent chance
          that each packet will be duplicated.\n\n##### Duplication correlation\n\nThe
          percet chance that the value for the next duplication will correlate with
          the preceeding value.\n\n\n\n## Configuration\n\nHere is an example of the
          ClusterImpairment custom resource.\n\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: test-impairment-cr\nspec:\n  duration:
          30 # seconds\n  startDelay: 5 # seconds. It typically takes about 2-3 seconds
          for the Daemonset to run\n  interfaces:\n  - \"ens2f0\"\n  ingress: # uses
          ifb\n    bandwidth: 0 # kbit\n    latency: 10 # ms\n    latencyOptions:\n
          \     jitter: 5 # ms\n      jitterCorrelation: 25 # percent\n      distribution:
          normal\n      reorder: 25 # percent of packets that will skip the delay\n
          \     reorderCorrelation: 25 # percent\n    loss: 0 # percent\n    lossOptions:\n
          \     correlation: 25 # percent\n    corruption: 0.1 # percent\n    corruptionOptions:\n
          \     correlation: 25 # percent\n    duplication: 1 # percent\n    duplicationOptions:\n
          \     correlation: 25 # percent\n  egress:\n    bandwidth: 0 # kbit\n    latency:
          100 # ms\n    loss: 0 # percent\n  linkFlapping:\n    enable: false\n    downTime:
          3 # Seconds\n    upTime: 3 # Seconds\n  nodeSelector:\n    key: \"node-role.kubernetes.io/worker\"\n
          \   value: \"\"\n    \n```\n\n#### Interfaces\n\nThe interfaces configuration
          option is a list of all interfaces that the impairments should be applied
          to. The current implementation will error out once it hits an invalid interface.\n\nIf
          an invalid interface is found, it will print out the list of interfaces.\n\n####
          Node Selector\n\nThe default node selector is all worker nodes, but you
          can change it to whatever node selector you want by setting the key and
          value.\nThere is a toleration to allow it to run on master nodes if the
          node selector includes master nodes.\n\n#### Duration\n\nThe number of seconds
          the impairments should run. It will try to sync the start and end time between
          all pods.\nIf set to -1, it will run indefinitely (a year), until you remove
          the ClusterImpairment custom resource. This is good for testing that requires
          steady impairments.\n\nIf the script is link flapping, set this value to
          be short enough so that if the link flap interrupts communication between
          the nodes, the daemonset will remove itself.\n\n#### Start Delay\n\nThe
          delay before starting the script. If you want the pods to be in sync, a
          minimum of a few seconds should be used because the pods take time to start
          up.\n\nYou can also utilize this feature to run an separate impairment after
          another. Just apply two resources at the same time, but add the duration
          and start delay of the first to the start delay of the second.\n\n#### Ingress
          and Egress\n\nThe configuration sections \"ingress\" and \"egress\" apply
          to each direction.\n\n##### Examples:\n\n##### Example 1\n\n\nIn this example,
          egress latency is set to 100ms, but the ingress latency is set to 10ms.
          So the latency to the interface will end up being 10ms, but 100ms going
          out. When pinging, this will result in 110ms of latency.\n\n```\n\napiVersion:
          apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n  name: uneven-latency\nspec:\n
          \ duration: 60\n  startDelay: 5\n  interfaces:\n  - \"ens2f0\"\n  ingress:\n
          \   latency: 10 # ms\n  egress:\n    latency: 100 # ms\n\n```\n\n##### Example
          2\n\n\nIn this example, link flapping will be enabled, and it will turn
          the interface on and off every 2 minutes. That is long enough for kubernetes
          to determine that a node is unavailable.\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: two-min-flap\nspec:\n  duration: 480\n
          \ startDelay: 5\n  interfaces:\n  - \"ens2f0\"\n  linkFlapping:\n    enable:
          true\n    downTime: 120 # Seconds\n    upTime: 120 # Seconds\n\n```\n\n#####
          Example 3\n\n\nIn this example, a realistic set of impairments is applied
          to `ens2f0` and `eno1` for 30 seconds:\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: typical-scenario\nspec:\n  duration:
          30 # seconds\n  startDelay: 5 # seconds\n  interfaces:\n  - \"ens2f0\"\n
          \ - \"eno1\"\n  egress:\n    latency: 50 # ms. Bidirectional, so total of
          100ms\n  ingress:\n    latency: 50 # ms. Bidirectional, so total of 100ms\n
          \ loss: 0.02 # percent\n\n```\n\n##### Example 4\n\n\nIn this example, a
          combination of latency, packet loss, bandwidth, and link flapping is applied.\n\n```\n\napiVersion:
          apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n  name: all-impairments\nspec:\n
          \ duration: 480 # seconds\n  startDelay: 5 # seconds\n  interfaces:\n  -
          \"ens2f0\"\n  egress:\n    latency: 50 # ms. Bidirectional, so total of
          100ms\n    loss: 0.02 # percent\n    bandwidth: 1000 # 1000 kbit/s, about
          1 mbit/s\n  ingress:\n    latency: 50 # ms. Bidirectional, so total of 100ms\n
          \   loss: 0.02 # percent\n    bandwidth: 1000 # 1000 kbit/s, about 1 mbit/s\n
          \ linkFlapping:\n    enable: true\n    downTime: 30 # Seconds\n    upTime:
          120 # Seconds\n\n```\n\n##### Example 5\n\n\nIn this example, packet duplicaiton
          is applied.\n\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n
          \ name: duplication\nspec:\n  duration: 60\n  startDelay: 5\n  interfaces:\n
          \ - \"ens3\"\n  ingress:\n    duplication: 25 # percent\n  egress:\n    duplication:
          25 # percent\n\n```\n\n## Setup\n\n### Requirements\n\n1. You need `make`
          installed.\n2. You need access to the kubernetes cluster with a kubeconfig.\n\n###
          Installation\n\nTo run using the current latest image:\n1. Clone the repository.\n2.
          Run `make deploy` with KUBECONFIG environment variable set to the location
          of the kubeconfig.\n\nTo run with your own code, there are more steps.\n1.
          Fork the repository.\n2. Clone to a machine that has access to the Kubernetes
          cluster and the kubeconfig.\n3. Modify the makefile to change the `IMG`
          variable to your image repository. If you do not have podman installed,
          also change podman to docker.\n4. Run `make docker-build` then `make docker-push`.\n5.
          Then run `make deploy`\n\n### Deploying from operator-hub\n\n*Important*
          You must manually apply the serviceaccount!\n\nFirst, Follow the instructions
          on this site to deploy from here.\n\nNext, apply the manifest \"serviceaccount.yaml\"
          from the manifests folder.\n\nThen you can apply your ClusterImpairment
          resources.\n\n## Running impairments\n\nFirst, configure a ClusterImpairment
          type's spec with the information for the impairment you want to run. Most
          importantly, set the interface(s). You can list the interfaces with `ifconfig`.
          If an invalid interface is set, the pod's logs will show `ifconfig` output.\n\n\nOnce
          the clusterimpairment type is set, apply it and it will work.\n\n\nTo re-run
          an impairment, delete the impairment with `oc delete -f path/to/manifest.yaml`,
          then apply it again.\n\n\n## Limitations\n\n\n### Multiple Impairments\n\n\nYou
          should avoid any configurations that apply impairments to the same interface
          on the same node. There are potential conflicts. The worker pod will attempt
          to remove all impairments on that interface before applying new ones.\n\n\nInstead,
          take advantage of the full control of both ingress and egress impairments
          from within the same ClusterImpairment resource.\n\n\nThere is currently
          a configured limit of 20 ClusterImpairment resources working at the same
          time. If that is a limitation for your use case, you can change it in `config/manager/manager.yaml`
          or you can try to find a way to make each ClusterImpairment do more.\n\n###
          Traffic Control (TC)\n\nTraffic control is how cluster-impairment-operator
          applies the latency, bandwidth, and packet loss impairments. The limitation
          is due to Linux not being realtime operating system, so the impairment will
          not be perfectly consistent.\n\n### Link Flapping\n\nWhen link flapping,
          if you flap the link that Kubernetes uses to communicate with the pods,
          you may be unable to remove the pod until the link is on long enough for
          Kubernetes to communicate with the impaired node.\n\n\nIn this case, it
          is helpful to set the duration properly instead of running for an indefinite
          or large amount of time, because the node will properly go back to the unimpaired
          state at that time.\n\n\nIf the cluster becomes offline due to the link
          flapping when you do not want it to be offline, soft restarting the nodes
          after removing the custom resource should remove all impairments.\n"
        displayName: cluster-impairment-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network
        - impairment
        - kubernetes
        - openshift
        - tc
        - trafficcontrol
        links:
        - name: Cluster Impairment Operator
          url: https://github.com/redhat-performance/cluster-impairment-operator
        maintainers:
        - email: joconnel@redhat.com
          name: Jared O'Connell
        - email: akrzos@redhat.com
          name: Alex Krzos
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat Performance
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/redhat-performance/cluster-impairment-operator:1.0.4
        version: 1.0.4
      entries:
      - name: cluster-impairment-operator.v1.0.4
        version: 1.0.4
      - name: cluster-impairment-operator.v1.0.3
        version: 1.0.3
      - name: cluster-impairment-operator.v1.0.2
        version: 1.0.2
      - name: cluster-impairment-operator.v1.0.1
        version: 1.0.1
      name: beta
    defaultChannel: beta
    packageName: cluster-impairment-operator
    provider:
      name: Red Hat Performance
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zettaset
      provider-url: ""
    name: xcrypt-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zts-xcrypt-operator.v0.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xcrypt.zettaset.com/v1alpha1",
                "kind": "XCrypt",
                "metadata": {
                  "name": "zts-masterset"
                },
                "spec": {
                  "status": {
                    "podNames": "zts-ca",
                    "replicas": "1,"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.connect.redhat.com/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
          createdAt: "2021-06-16T12:29:13.97Z"
          description: XCrypt Operator deploys Zettaset XCrypt Container Encryption  -
            the only software-defined solution that offers transparent high performance
            data-at-rest encryption of critical data in container environments.
          support: Zettaset
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XCrypt is the Schema for the xcrypts API
            kind: XCrypt
            name: xcrypts.xcrypt.zettaset.com
            version: v1alpha1
        description: "Zettaset’s RedHat Certified XCrypt Operator allows users to
          deploy multiple XCrypt components on OpenShift with just a few simple commands.
          The XCrypt Operator deployment provides software-defined encryption that
          transparently protects container data throughout your OpenShift environment.\n\n##
          Core Features\n* **Automated XCrypt provisioning** - Deploying a Zettaset
          XCrypt has never been easier. Adjust cluster specific configuration and
          let the Zettaset Operator take care of provisioning XCrypt services and
          setting up the XCrypt encryption to your exact specification. \n* **CSI
          Driver deployment**: Zettaset XCrypt Operator automatically deploys Zettaset
          XCrypt CSI Driver in order to provide automated and transparent data-at-rest
          \ encryption functionality compatible with Openshift and any other container
          orchestrator. Once the Operator is provisioned, the driver runs fine tuned
          and ready to work with other XCrypt components.\n* **Multi-node topology**:
          Encrypt as many nodes and volumes as you need to. XCrypt components can
          be installed on multiple nodes across the cluster. This is defined by two
          labels: zts-master and zts-worker. Normally zts-master node is a single
          node that runs Zettaset XCrypt major services required for the product operation.
          Multiple nodes can be marked as zts-worker and those are the nodes that
          can have encrypted volumes.\n* **Seamless Encrypted Volume mount**: Volumes
          are not only being encrypted, but also are mounted/unmounted to containers
          with Persistent Volume Claims. This happens as seamlessly as regular volumes
          provision and mount/unmount.\n* **Robust failover**: A node failure doesn’t
          affect XCrypt operations on other nodes of the cluster. Encryption on a
          restarted/recovered node gets back to normal functioning if the node stays
          healthy and data hasn’t been corrupted during the restart\n\n## About Zettaset
          XCrypt Kubernetes Encryption for OpenShift\nXCrypt Kubernetes Encryption
          for OpenShift is a software-only solution that enables encryption of data
          at rest stored in Red Hat OpenShift container environments without slowing
          down the DevSecOps process. In addition to providing a transparent and high-performance
          layer of security that flexibly protects container data across on-premise,
          in hybrid, and multi-cloud environments, it also helps your organization
          efficiently transition from DevOps to DevSecOps.  \n\nMore information about
          Zettaset XCrypt Kubernetes Encryption for OpenShift can be found in the
          [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)\n\n##
          RedHat OpenShift + Zettaset: The next generation of secure enterprise infrastructure
          for DevSecOps  \nBy combining RedHat’s Container Platform with Zettaset’s
          certified software-defined encryption, you can create the next generation
          of secure enterprise infrastructure that DevSecOps requires. Together they
          create the stable, consistent, and supported base that your organization
          needs to develop applications faster, simplify updates, drive innovation,
          and reduce the risk of potential security breaches and data theft.\n###
          Core Features\n* Software-only for simple and scalable deployments \n* Negligible
          performance impact\n* Transparent to developers and administrators \n* Works
          with AWS, Google, and Azure \n* Unique encryption key per container volume
          \n* Encrypted volumes are only available when in use\n* Direct integration
          with OpenShift \n* Automated encryption policy management Secure erase of
          volumes\n* Ability to securely decommission a node\n\n## How does XCrypt
          Kubernetes Encryption fit into OpenShift? \nXCrypt XCrypt Kubernetes Encryption
          for OpenShift Encryption makes it incredibly easy to protect container data
          in your OpenShift environment. \n### Use cases include:\n* Transitioning
          from DevOps to DevSecOps \n* Ensuring data protection in OpenShift container
          environments with negligible performance impact \n* Achieving compliance
          in regulated industries like healthcare and finance\n\n## Other Information\nRole-based
          access control needs to be configured prior to the Operator installation.
          Please find details in the [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)
          \   "
        displayName: Zettaset XCrypt Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - encryption
        - storage
        links:
        - name: Zettaset
          url: https://www.zettaset.com/
        maintainers:
        - email: info@zettaset.com
          name: Zettaset
        maturity: alpha
        provider:
          name: Zettaset
        relatedImages:
        - registry.connect.redhat.com/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
        - registry.connect.redhat.com/zts/xcrypt-license-server-1@sha256:bc5486892a41e70f5dc4c4e1d82fca8db2149598fb8b5950ba5da1ee2d9d245d
        - registry.connect.redhat.com/zts/xcrypt-host-manager-1@sha256:9faa4b597eb74df412761dffa057b64fb69dbf9fa016414962eb2099a343e734
        - registry.connect.redhat.com/zts/zts-csi-driver-1@sha256:c15fa9a7cc40891103f426206663dcebe1210c8ee3ab6037cbf64e1f86b9c7e5
        - registry.connect.redhat.com/zts/xcrypt-ca-1@sha256:9f1015dfb459b69f5dbf449efa0517452095a6cf4f20826ebae278f9f800ffe7
        - registry.connect.redhat.com/zts/xcrypt-kmip-server-1@sha256:9af5d21dd748dfbf4e107f27a5f10fa3c0ba928ade9b989a744b620ce34ec757
        version: 0.0.17
      entries:
      - name: zts-xcrypt-operator.v0.0.17
        version: 0.0.17
      name: alpha
    defaultChannel: alpha
    packageName: xcrypt-operator
    provider:
      name: Zettaset
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-spark-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spark-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:a6c82a6d6ebaa0d2572d644ac9013451e020f134c6fe46222fef37b1e9aa3cbb
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:a6c82a6d6ebaa0d2572d644ac9013451e020f134c6fe46222fef37b1e9aa3cbb
        version: 24.11.1
      entries:
      - name: spark-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: spark-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:8ad9948bf8264924a0eedd92c2e58269117487b9d57816384447bb6c999be681
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:8ad9948bf8264924a0eedd92c2e58269117487b9d57816384447bb6c999be681
        version: 25.3.0
      entries:
      - name: spark-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: spark-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkConnectServerSpec via
              `CustomResource`
            displayName: sparkconnectservers.spark.stackable.tech
            kind: SparkConnectServer
            name: sparkconnectservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
        version: 25.7.0
      entries:
      - name: spark-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: spark-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkConnectServerSpec via
              `CustomResource`
            displayName: sparkconnectservers.spark.stackable.tech
            kind: SparkConnectServer
            name: sparkconnectservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
        version: 25.7.0
      entries:
      - name: spark-operator.v25.7.0
        version: 25.7.0
      - name: spark-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-spark-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://skupper.io
    name: skupper-netobs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-netobs-operator.v2.1.1-rh-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability.skupper.io/v2alpha1",
                "kind": "NetworkObserver",
                "metadata": {
                  "name": "networkobserver-sample"
                },
                "spec": {
                  "auth": {
                    "basic": {
                      "create": true,
                      "secretName": ""
                    },
                    "openshift": {
                      "cookieSecretName": "",
                      "createCookieSecret": true,
                      "serviceAccount": {
                        "create": true,
                        "nameOverride": ""
                      }
                    },
                    "strategy": "openshift"
                  },
                  "extraArgs": null,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "registry.redhat.io/service-interconnect/skupper-network-observer-rhel9",
                    "tag": "2.1.1"
                  },
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nginx": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/ubi9/nginx-124",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": 9.6
                  },
                  "openshiftOauthProxy": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-oauth-proxy-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.19.0"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "prometheus": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-prometheus-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.16.0"
                  },
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "router": {
                    "certificate": {
                      "create": true,
                      "nameOverride": ""
                    },
                    "endpoint": "amqps://skupper-router-local"
                  },
                  "securityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    }
                  },
                  "service": {
                    "port": 443,
                    "type": "ClusterIP"
                  },
                  "skipManagementLabels": false,
                  "tls": {
                    "openshiftIssued": true,
                    "secretName": "",
                    "skupperIssued": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
          createdAt: "2025-08-27T20:21:18Z"
          description: Red Hat Service Interconnect Network Observer Operator provides
            the ability to monitor and observe a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkObserver
            name: networkobservers.observability.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect Network Observer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/ubi9/nginx-124@sha256:cbcca4743015b4ef167383a080f4711a6049e275e0b66f80a74f8b087b0dfc7f
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:d3056b35d9a205b9f2c48d924f199c5ac23904eb18d526e4bff229e7c7181415
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
        version: 2.1.1-rh-8
      entries:
      - name: skupper-netobs-operator.v2.1.1-rh-8
        version: 2.1.1-rh-8
      name: stable-2
    - currentCSV: skupper-netobs-operator.v2.1.1-rh-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability.skupper.io/v2alpha1",
                "kind": "NetworkObserver",
                "metadata": {
                  "name": "networkobserver-sample"
                },
                "spec": {
                  "auth": {
                    "basic": {
                      "create": true,
                      "secretName": ""
                    },
                    "openshift": {
                      "cookieSecretName": "",
                      "createCookieSecret": true,
                      "serviceAccount": {
                        "create": true,
                        "nameOverride": ""
                      }
                    },
                    "strategy": "openshift"
                  },
                  "extraArgs": null,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "registry.redhat.io/service-interconnect/skupper-network-observer-rhel9",
                    "tag": "2.1.1"
                  },
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nginx": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/ubi9/nginx-124",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": 9.6
                  },
                  "openshiftOauthProxy": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-oauth-proxy-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.19.0"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "prometheus": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-prometheus-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.16.0"
                  },
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "router": {
                    "certificate": {
                      "create": true,
                      "nameOverride": ""
                    },
                    "endpoint": "amqps://skupper-router-local"
                  },
                  "securityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    }
                  },
                  "service": {
                    "port": 443,
                    "type": "ClusterIP"
                  },
                  "skipManagementLabels": false,
                  "tls": {
                    "openshiftIssued": true,
                    "secretName": "",
                    "skupperIssued": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
          createdAt: "2025-08-27T20:21:18Z"
          description: Red Hat Service Interconnect Network Observer Operator provides
            the ability to monitor and observe a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkObserver
            name: networkobservers.observability.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect Network Observer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:d3056b35d9a205b9f2c48d924f199c5ac23904eb18d526e4bff229e7c7181415
        - registry.redhat.io/ubi9/nginx-124@sha256:cbcca4743015b4ef167383a080f4711a6049e275e0b66f80a74f8b087b0dfc7f
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        version: 2.1.1-rh-8
      entries:
      - name: skupper-netobs-operator.v2.1.1-rh-8
        version: 2.1.1-rh-8
      name: stable-2.1
    defaultChannel: stable-2
    packageName: skupper-netobs-operator
    provider:
      name: Red Hat
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/trustification/trusted-profile-analyzer-operator
    name: rhtpa-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhtpa-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtpa.io/v1",
                "kind": "TrustedProfileAnalyzer",
                "metadata": {
                  "name": "trustedprofileanalyzer-sample"
                },
                "spec": {
                  "appDomain": "change-me",
                  "collector": {},
                  "database": {},
                  "image": {
                    "fullName": "registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e",
                    "pullPolicy": "IfNotPresent"
                  },
                  "infrastructure": {
                    "port": 9010
                  },
                  "ingress": {},
                  "metrics": {
                    "enabled": false
                  },
                  "modules": {
                    "createDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "createImporters": {
                      "enabled": false,
                      "image": {},
                      "importers": {}
                    },
                    "importer": {
                      "concurrency": 1,
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {},
                      "workingDirectory": {
                        "size": "32Gi"
                      }
                    },
                    "migrateDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "server": {
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "ingress": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {}
                    }
                  },
                  "oidc": {
                    "clients": {
                      "cli": {
                        "clientSecret": null
                      },
                      "frontend": {}
                    }
                  },
                  "openshift": {
                    "useServiceCa": true
                  },
                  "partOf": "trustify",
                  "replicas": 1,
                  "rust": {},
                  "storage": {},
                  "tls": {},
                  "tracing": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-09-19T15:04:16Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Profile Analyzer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trustification/trusted-profile-analyzer-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: TrustedProfileAnalyzer
            name: trustedprofileanalyzers.rhtpa.io
            version: v1
        description: |
          [Tech Preview]
          Red Hat Trusted Profile Analyzer (RHTPA) provides a single source of truth for your software inventory across all your environments. You can verify the integrity of your in-house and third-party applications against their respective Software Bill of Materials (SBOM) document.  Trusted Profile Analyzer provides vulnerability, component and license transparency, strengthens your software supply chain security, and enables an accurate, up-to-date understanding of your software inventory.

          Features and Benefits:

          Enhanced Software Supply Chain Security: Offers critical transparency and integrity verification across the entire software lifecycle.

          Reduced Operational Risk: Proactively identifies and helps remediate vulnerabilities minimizing risk and compliance violations.

          Improved Compliance and Auditability: Simplifies the process of meeting regulatory requirements for software transparency and integrity by providing verifiable documentation and continuous monitoring.

          Faster Remediation and Incident Response: Provides targeted, actionable insights to prioritize and accelerate the patching and remediation of vulnerabilities.

          Increased Developer Productivity & Trust: Integrates security into existing workflows without significant overhead, allowing developers to build and deploy with greater confidence in the integrity of their code and components.

          Centralized Visibility and Control: Consolidates disparate security information into a single pane of glass, offering a holistic view of your application security posture.

          Red Hat Trusted Profile Analyzer is a crucial tool for any organization running applications on OpenShift that needs to confidently understand, verify, and secure the software they deploy, especially in an era of complex software supply chains and increasing regulatory scrutiny.

          [Documentation](https://docs.redhat.com/en/documentation/red_hat_trusted_profile_analyzer/2.1).
        displayName: Red Hat Trusted Profile Analyzer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rhtpa
        - trusted-profile-analyzer
        links:
        - name: Trusted Profile Analyzer Operator
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        maintainers:
        - email: rhtpa-support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        relatedImages:
        - registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
        - registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e
        version: 1.0.2
      entries:
      - name: rhtpa-operator.v1.0.2
        version: 1.0.2
      - name: rhtpa-operator.v1.0.1
        version: 1.0.1
      - name: rhtpa-operator.v1.0.0
        version: 1.0.0
      name: stable
    - currentCSV: rhtpa-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtpa.io/v1",
                "kind": "TrustedProfileAnalyzer",
                "metadata": {
                  "name": "trustedprofileanalyzer-sample"
                },
                "spec": {
                  "appDomain": "change-me",
                  "collector": {},
                  "database": {},
                  "image": {
                    "fullName": "registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e",
                    "pullPolicy": "IfNotPresent"
                  },
                  "infrastructure": {
                    "port": 9010
                  },
                  "ingress": {},
                  "metrics": {
                    "enabled": false
                  },
                  "modules": {
                    "createDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "createImporters": {
                      "enabled": false,
                      "image": {},
                      "importers": {}
                    },
                    "importer": {
                      "concurrency": 1,
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {},
                      "workingDirectory": {
                        "size": "32Gi"
                      }
                    },
                    "migrateDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "server": {
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "ingress": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {}
                    }
                  },
                  "oidc": {
                    "clients": {
                      "cli": {
                        "clientSecret": null
                      },
                      "frontend": {}
                    }
                  },
                  "openshift": {
                    "useServiceCa": true
                  },
                  "partOf": "trustify",
                  "replicas": 1,
                  "rust": {},
                  "storage": {},
                  "tls": {},
                  "tracing": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-09-19T15:04:16Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Profile Analyzer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trustification/trusted-profile-analyzer-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: TrustedProfileAnalyzer
            name: trustedprofileanalyzers.rhtpa.io
            version: v1
        description: |
          [Tech Preview]
          Red Hat Trusted Profile Analyzer (RHTPA) provides a single source of truth for your software inventory across all your environments. You can verify the integrity of your in-house and third-party applications against their respective Software Bill of Materials (SBOM) document.  Trusted Profile Analyzer provides vulnerability, component and license transparency, strengthens your software supply chain security, and enables an accurate, up-to-date understanding of your software inventory.

          Features and Benefits:

          Enhanced Software Supply Chain Security: Offers critical transparency and integrity verification across the entire software lifecycle.

          Reduced Operational Risk: Proactively identifies and helps remediate vulnerabilities minimizing risk and compliance violations.

          Improved Compliance and Auditability: Simplifies the process of meeting regulatory requirements for software transparency and integrity by providing verifiable documentation and continuous monitoring.

          Faster Remediation and Incident Response: Provides targeted, actionable insights to prioritize and accelerate the patching and remediation of vulnerabilities.

          Increased Developer Productivity & Trust: Integrates security into existing workflows without significant overhead, allowing developers to build and deploy with greater confidence in the integrity of their code and components.

          Centralized Visibility and Control: Consolidates disparate security information into a single pane of glass, offering a holistic view of your application security posture.

          Red Hat Trusted Profile Analyzer is a crucial tool for any organization running applications on OpenShift that needs to confidently understand, verify, and secure the software they deploy, especially in an era of complex software supply chains and increasing regulatory scrutiny.

          [Documentation](https://docs.redhat.com/en/documentation/red_hat_trusted_profile_analyzer/2.1).
        displayName: Red Hat Trusted Profile Analyzer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rhtpa
        - trusted-profile-analyzer
        links:
        - name: Trusted Profile Analyzer Operator
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        maintainers:
        - email: rhtpa-support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        relatedImages:
        - registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
        - registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e
        version: 1.0.2
      entries:
      - name: rhtpa-operator.v1.0.2
        version: 1.0.2
      - name: rhtpa-operator.v1.0.1
        version: 1.0.1
      - name: rhtpa-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable
    packageName: rhtpa-operator
    provider:
      name: Red Hat
      url: https://github.com/trustification/trusted-profile-analyzer-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: nova-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nova-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "Nova",
                "metadata": {
                  "name": "nova"
                },
                "spec": {
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaAPI",
                "metadata": {
                  "name": "novaapi-sample"
                },
                "spec": {
                  "apiDatabaseHostname": "openstack",
                  "cell0DatabaseHostname": "openstack",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "registeredCells": {},
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaCell",
                "metadata": {
                  "name": "novacell1-with-upcall-sample"
                },
                "spec": {
                  "apiDatabaseHostname": "openstack",
                  "cellDatabaseHostname": "openstack",
                  "cellName": "cell1",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaCompute",
                "metadata": {
                  "name": "nova-compute-ironic"
                },
                "spec": {
                  "novaComputeTemplates": {
                    "compute-ironic": {
                      "computeDriver": "ironic.IronicDriver",
                      "customServiceConfig": "",
                      "networkAttachments": [
                        "internalapi"
                      ],
                      "replicas": 1,
                      "resources": {}
                    }
                  }
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaConductor",
                "metadata": {
                  "name": "novaconductor-sample"
                },
                "spec": {
                  "cellName": "cell1",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaMetadata",
                "metadata": {
                  "name": "novametadata-sample"
                },
                "spec": {
                  "apiDatabaseAccount": "nova-api",
                  "apiDatabaseHostname": "openstack",
                  "cellDatabaseAccount": "nova-cell0",
                  "cellDatabaseHostname": "openstack",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "registeredCells": {},
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaNoVNCProxy",
                "metadata": {
                  "name": "novanovncproxy-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaScheduler",
                "metadata": {
                  "name": "novascheduler-sample"
                },
                "spec": {
                  "apiDatabaseHostname": "openstack",
                  "cell0DatabaseHostname": "openstack",
                  "customServiceConfig": "[DEFAULT]\ndebug=true\n",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "registeredCells": {},
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:21Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nova
            name: nova.nova.openstack.org
            version: v1beta1
          - description: NovaAPI is the Schema for the novaapis API
            displayName: Nova API
            kind: NovaAPI
            name: novaapis.nova.openstack.org
            version: v1beta1
          - description: NovaCell is the Schema for the novacells API
            displayName: Nova Cell
            kind: NovaCell
            name: novacells.nova.openstack.org
            version: v1beta1
          - description: NovaCompute is the Schema for the NovaCompute
            displayName: Nova Compute
            kind: NovaCompute
            name: novacomputes.nova.openstack.org
            version: v1beta1
          - description: NovaConductor is the Schema for the novaconductors API
            displayName: Nova Conductor
            kind: NovaConductor
            name: novaconductors.nova.openstack.org
            version: v1beta1
          - description: NovaMetadata is the Schema for the novametadata API
            displayName: Nova Metadata
            kind: NovaMetadata
            name: novametadata.nova.openstack.org
            version: v1beta1
          - description: NovaNoVNCProxy is the Schema for the novanovncproxies API
            displayName: Nova No VNCProxy
            kind: NovaNoVNCProxy
            name: novanovncproxies.nova.openstack.org
            version: v1beta1
          - description: NovaScheduler is the Schema for the novaschedulers API
            displayName: Nova Scheduler
            kind: NovaScheduler
            name: novaschedulers.nova.openstack.org
            version: v1beta1
        description: Nova Operator
        displayName: Nova Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Nova
        links:
        - name: Nova Operator
          url: https://github.com/openstack-k8s-operators/nova-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-nova-scheduler-rhel9@sha256:ba91196cd4f9b6e6f49868c91343b8eca7cc0d5016c36e83e2cc76f45cbe777b
        - registry.redhat.io/rhoso/openstack-nova-compute-rhel9@sha256:43cbe46ab263403fc303b6bc2388863029aad33465b237db18da7cd6d42078e1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/nova-rhel9-operator@sha256:f058c56df02a108505d51b9d19973894c2ee59055776c9f3fd7f7afe23a5106a
        - registry.redhat.io/rhoso/openstack-nova-api-rhel9@sha256:bbb8ebceb6f30b6c8beb120429e69f323bd231c09aaaddc5be9f15a527e928ed
        - registry.redhat.io/rhoso/openstack-nova-conductor-rhel9@sha256:500c36fd519611be8662e97ca31b1e59269c14d8aa94286dad9925e957462155
        - registry.redhat.io/rhoso/openstack-nova-novncproxy-rhel9@sha256:bf534f0c72f188d5f51e40b0392ace182db2055dff9ad14b01eeef13056febbf
        version: 1.0.6
      entries:
      - name: nova-operator.v1.0.6
        version: 1.0.6
      - name: nova-operator.v1.0.5
        version: 1.0.5
      - name: nova-operator.v1.0.4
        version: 1.0.4
      - name: nova-operator.v1.0.3
        version: 1.0.3
      - name: nova-operator.v1.0.2
        version: 1.0.2
      - name: nova-operator.v1.0.1
        version: 1.0.1
      - name: nova-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: nova-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq7-interconnect-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: 1.10.x
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: 1.2.0
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: latest
    defaultChannel: 1.10.x
    packageName: amq7-interconnect-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-registry
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-operator.v1.1.3-v2.6.4.final
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-mem"
                },
                "spec": {
                  "configuration": {
                    "persistence": "mem"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-operator:1.1.3
          createdAt: "2024-09-17"
          description: Deploy and manage Apicurio Registry on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents an Apicurio Registry instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Apicurio Registry

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Provides a REST API and web UI to manage the artifacts and artifact meta-data
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, in-memory
          - Compatible with Confluent and IBM APIs
          - Runs on a lightweight Quarkus platform
          - Includes Maven plugin to integrate with Maven based builds

          ## Apicurio Registry Operator

          Provides a quick and easy way to deploy and manage Apicurio Registry on Kubernetes.

          **Features**
          - Supports installation and configuration of the Registry for all storage options
          - Easily perform a rolling upgrade of Apicurio Registry

          ## Prerequisites

          This operator does not deploy storage for Apicurio Registry. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/main/LICENSE)
        displayName: Apicurio Registry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-operator:1.1.3
        version: 1.1.3-v2.6.4.final
      entries:
      - name: apicurio-registry-operator.v1.1.3-v2.6.4.final
        version: 1.1.3-v2.6.4.final
      - name: apicurio-registry-operator.v1.1.2-v2.5.11.final
        version: 1.1.2-v2.5.11.final
      - name: apicurio-registry-operator.v1.1.1-v2.5.9.final
        version: 1.1.1-v2.5.9.final
      - name: apicurio-registry-operator.v1.1.0-v2.4.12.final
        version: 1.1.0-v2.4.12.final
      - name: apicurio-registry-operator.v1.0.0-v2.0.0.final
        version: 1.0.0-v2.0.0.final
      name: 2.x
    defaultChannel: 2.x
    packageName: apicurio-registry
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: volume-expander-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: volume-expander-operator.v0.3.6
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage,Monitoring
          certified: "false"
          containerImage: quay.io/redhat-cop/volume-expander-operator@sha256:e4059a66a9af7c786c8c651abb8a57a55a0e528342e3a59f8c264b40f87d6e25
          createdAt: "2022-05-05T13:25:53Z"
          description: This operator provides automation expanding pod's volumes when
            they are about to run out of space.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: volume-expander-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/volume-expander-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The purpose of the volume-expander-operator is to expand volumes when they are running out of space.
          This is achieved by using the [volume expansion feature](https://kubernetes.io/docs/concepts/storage/persistent-volumes/#expanding-persistent-volumes-claims).

          The operator periodically checks the `kubelet_volume_stats_used_bytes` and `kubelet_volume_stats_capacity_bytes published` by the kubelets to decide when to expand a volume.
          Notice that these metrics are generated only when a volume is mounted to a pod. Also the kubelet takes a minute or two to start generating accurate values for these metrics. The operator accounts for that.

          This operator works based on the following annotations to PersistentVolumeClaim resources:

          | Annotation | Default &nbsp;  &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;  | Description  |
          | :--- | :--- | :--- |
          | `volume-expander-operator.redhat-cop.io/autoexpand`  | N/A  | if set to "true" enables the volume-expander-operator to watch on this PVC  |
          | `volume-expander-operator.redhat-cop.io/polling-frequency`  | `"30s"` | How frequently to poll the volume metrics. Express this value as a valid golang [Duration](https://golang.org/pkg/time/#ParseDuration)  |
          | `volume-expander-operator.redhat-cop.io/expand-threshold-percent` | `"80"` | the percentage of used storage after which the volume will be expanded. This must be a positive integer. |
          | `volume-expander-operator.redhat-cop.io/expand-by-percent` | `"25"` | the percentage by which the volume will be expanded, relative to the current size. This must be an integer between 0 and 100 |
          | `volume-expander-operator.redhat-cop.io/expand-up-to` | MaxInt64 | the upper bound for this volume to be expanded to. The default value is the largest quantity representable and is intended to be interpreted as infinite. If the default is used it is recommend to ensure the namespace has a quota on the used storage class. |

          Note that not all of the storage driver implementations support volume expansion. It is a responsibility of the user/platform administrator to ensure that storage class and the persistent volume claim meet all the requirements needed for the volume expansion feature to work properly.

          This operator was tested with [OCS](https://www.redhat.com/en/technologies/cloud-computing/openshift-container-storage), but should work with any other storage driver that supports volume expansion.
        displayName: Volume Expander Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - volume
        - storage
        - csi
        - expansion
        links:
        - name: Volume Expander Operator
          url: https://github.com/redhat-cop/volume-expander-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/volume-expander-operator@sha256:e4059a66a9af7c786c8c651abb8a57a55a0e528342e3a59f8c264b40f87d6e25
        version: 0.3.6
      entries:
      - name: volume-expander-operator.v0.3.6
        version: 0.3.6
      - name: volume-expander-operator.v0.3.5
        version: 0.3.5
      - name: volume-expander-operator.v0.3.4
        version: 0.3.4
      - name: volume-expander-operator.v0.3.3
        version: 0.3.3
      - name: volume-expander-operator.v0.3.2
        version: 0.3.2
      - name: volume-expander-operator.v0.3.1
        version: 0.3.1
      - name: volume-expander-operator.v0.3.0
        version: 0.3.0
      - name: volume-expander-operator.v0.2.0
        version: 0.2.0
      - name: volume-expander-operator.v0.1.8
        version: 0.1.8
      - name: volume-expander-operator.v0.1.7
        version: 0.1.7
      - name: volume-expander-operator.v0.1.6
        version: 0.1.6
      - name: volume-expander-operator.v0.1.5
        version: 0.1.5
      - name: volume-expander-operator.v0.1.4
        version: 0.1.4
      - name: volume-expander-operator.v0.1.3
        version: 0.1.3
      - name: volume-expander-operator.v0.1.2
        version: 0.1.2
      - name: volume-expander-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: volume-expander-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-eventbridge-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-eventbridge-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Archive",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Endpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "EventBus",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/eventbridge-controller:1.1.1
          createdAt: "2025-09-19T20:51:55Z"
          description: AWS EventBridge controller is a service controller for managing
            EventBridge resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Archive represents the state of an AWS eventbridge Archive
              resource.
            displayName: Archive
            kind: Archive
            name: archives.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: Endpoint represents the state of an AWS eventbridge Endpoint
              resource.
            displayName: Endpoint
            kind: Endpoint
            name: endpoints.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: EventBus represents the state of an AWS eventbridge EventBus
              resource.
            displayName: EventBus
            kind: EventBus
            name: eventbuses.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: Rule represents the state of an AWS eventbridge Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.eventbridge.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EventBridge resources in AWS from within your Kubernetes cluster.

          **About Amazon EventBridge**

          EventBridge is a serverless service that uses events to connect application components together, making it easier for you to build scalable event-driven applications. Use it to route events from sources such as home-grown applications, AWS services, and third- party software to consumer applications across your organization. EventBridge provides a simple and consistent way to ingest, filter, transform, and deliver events so you can build new applications quickly.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EventBridge
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eventbridge
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EventBridge Developer Resources
          url: https://aws.amazon.com/eventbridge/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eventbridge maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/eventbridge-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-eventbridge-controller.v1.1.1
        version: 1.1.1
      - name: ack-eventbridge-controller.v1.1.0
        version: 1.1.0
      - name: ack-eventbridge-controller.v1.0.28
        version: 1.0.28
      - name: ack-eventbridge-controller.v1.0.27
        version: 1.0.27
      - name: ack-eventbridge-controller.v1.0.26
        version: 1.0.26
      - name: ack-eventbridge-controller.v1.0.25
        version: 1.0.25
      - name: ack-eventbridge-controller.v1.0.24
        version: 1.0.24
      - name: ack-eventbridge-controller.v1.0.23
        version: 1.0.23
      - name: ack-eventbridge-controller.v1.0.22
        version: 1.0.22
      - name: ack-eventbridge-controller.v1.0.21
        version: 1.0.21
      - name: ack-eventbridge-controller.v1.0.20
        version: 1.0.20
      - name: ack-eventbridge-controller.v1.0.18
        version: 1.0.18
      - name: ack-eventbridge-controller.v1.0.17
        version: 1.0.17
      - name: ack-eventbridge-controller.v1.0.16
        version: 1.0.16
      - name: ack-eventbridge-controller.v1.0.15
        version: 1.0.15
      - name: ack-eventbridge-controller.v1.0.14
        version: 1.0.14
      - name: ack-eventbridge-controller.v1.0.13
        version: 1.0.13
      - name: ack-eventbridge-controller.v1.0.12
        version: 1.0.12
      - name: ack-eventbridge-controller.v1.0.11
        version: 1.0.11
      - name: ack-eventbridge-controller.v1.0.6
        version: 1.0.6
      - name: ack-eventbridge-controller.v1.0.5
        version: 1.0.5
      - name: ack-eventbridge-controller.v1.0.4
        version: 1.0.4
      - name: ack-eventbridge-controller.v1.0.3
        version: 1.0.3
      - name: ack-eventbridge-controller.v1.0.1
        version: 1.0.1
      - name: ack-eventbridge-controller.v1.0.0
        version: 1.0.0
      - name: ack-eventbridge-controller.v0.0.7
        version: 0.0.7
      - name: ack-eventbridge-controller.v0.0.6
        version: 0.0.6
      - name: ack-eventbridge-controller.v0.0.5
        version: 0.0.5
      - name: ack-eventbridge-controller.v0.0.4
        version: 0.0.4
      - name: ack-eventbridge-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-eventbridge-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Oracle Corporation
      provider-url: https://github.com/oracle/coherence-operator
    name: coherence-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: coherence-operator.v3.5.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "coherence.oracle.com/v1",
                "kind": "Coherence",
                "metadata": {
                  "name": "coherence-sample"
                },
                "spec": {
                  "replicas": 3
                }
              },
              {
                "apiVersion": "coherence.oracle.com/v1",
                "kind": "CoherenceJob",
                "metadata": {
                  "name": "coherence-job-sample"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "false"
          containerImage: container-registry.oracle.com/middleware/coherence-operator:3.5.4
          createdAt: "2025-08-28T16:15:09Z"
          description: The Oracle Coherence Kubernetes Operator enables easy management
            of Coherence clusters in a Kubernetes environment.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/arch.amd64: supported
          operatorframework.io/arch.arm64: supported
          operatorframework.io/os.linux: supported
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/oracle/coherence-operator
          support: Oracle Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Coherence is the top level schema for the Coherence API and
              custom resource definition (CRD).
            displayName: Coherence
            kind: Coherence
            name: coherence.coherence.oracle.com
            version: v1
          - description: |-
              CoherenceJob is the top level schema for the CoherenceJob API and custom resource definition (CRD)
              for configuring Coherence Job workloads.
            displayName: Coherence Job
            kind: CoherenceJob
            name: coherencejob.coherence.oracle.com
            version: v1
        description: |
          The Oracle Coherence Kubernetes Operator enables easy management of Coherence clusters in a Kubernetes environment.

          ### Installation Guide & Configuration

          For the complete installation guide and all configuration options please refer to
          [Coherence Operator Documentation](https://oracle.github.io/coherence-operator/docs/latest/#/about/01_overview).
        displayName: Coherence Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Coherence
        - Oracle
        - In Memory Data
        - Caching
        links:
        - name: Coherence Kubernetes Operator
          url: https://oracle.github.io/coherence-operator/docs/latest
        - name: Oracle Coherence
          url: https://coherence.community
        maintainers:
        - email: jonathan.knight@oracle.com
          name: Jonathan Knight
        maturity: mature
        minKubeVersion: 1.26.0
        provider:
          name: Oracle Corporation
          url: https://github.com/oracle/coherence-operator
        relatedImages:
        - container-registry.oracle.com/middleware/coherence-operator:3.5.4
        version: 3.5.4
      entries:
      - name: coherence-operator.v3.5.4
        version: 3.5.4
      - name: coherence-operator.v3.5.2
        version: 3.5.2
      - name: coherence-operator.v3.5.0
        version: 3.5.0
      - name: coherence-operator.v3.4.3
        version: 3.4.3
      name: stable
    defaultChannel: stable
    packageName: coherence-operator
    provider:
      name: Oracle Corporation
      url: https://github.com/oracle/coherence-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nxrm-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nxrm-operator-certified.v3.58.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusRepo",
                "metadata": {
                  "name": "example-nexusrepo"
                },
                "spec": {
                  "config": {
                    "data": null,
                    "enabled": false,
                    "mountPath": "/sonatype-nexus-conf"
                  },
                  "deployment": {
                    "additionalContainers": null,
                    "additionalVolumeMounts": null,
                    "additionalVolumes": null,
                    "annotations": {},
                    "initContainers": null,
                    "postStart": {
                      "command": null
                    },
                    "preStart": {
                      "command": null
                    },
                    "terminationGracePeriodSeconds": 120
                  },
                  "deploymentStrategy": {
                    "type": "Recreate"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "path": "/",
                    "rules": null,
                    "tls": {
                      "enabled": true,
                      "secretName": "nexus-tls"
                    }
                  },
                  "nexus": {
                    "dockerPort": 5003,
                    "env": [
                      {
                        "name": "INSTALL4J_ADD_VM_PARAMS",
                        "value": "-Xms2703M -Xmx2703M -XX:MaxDirectMemorySize=2703M -XX:+UnlockExperimentalVMOptions -XX:+UseCGroupMemoryLimitForHeap"
                      }
                    ],
                    "hostAliases": [],
                    "imageName": "registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:06fe31f40ff3fd7556fc7b8f6b2efe80ada1d9079ec9aaad60d1fdc0ee0677f0",
                    "imagePullPolicy": "IfNotPresent",
                    "imagePullSecret": "",
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "path": "/",
                      "periodSeconds": 30
                    },
                    "nexusPort": 8081,
                    "podAnnotations": {},
                    "readinessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "path": "/",
                      "periodSeconds": 30
                    },
                    "resources": {},
                    "securityContext": {
                    },
                    "service": {
                      "type": "NodePort"
                    }
                  },
                  "nexusProxyRoute": {
                    "annotations": null,
                    "enabled": false,
                    "labels": null
                  },
                  "persistence": {
                    "accessMode": "ReadWriteOnce",
                    "enabled": true,
                    "storageSize": "8Gi"
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": null,
                    "enabled": false,
                    "labels": null,
                    "name": "docker",
                    "portName": "docker"
                  },
                  "secret": {
                    "data": null,
                    "enabled": false,
                    "mountPath": "/etc/secret-volume",
                    "readOnly": true
                  },
                  "service": {
                    "annotations": {},
                    "enabled": false,
                    "labels": {},
                    "ports": [
                      {
                        "name": "nexus-service",
                        "port": 80,
                        "targetPort": 80
                      }
                    ]
                  },
                  "statefulset": {
                    "enabled": false
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nxrm-operator-certified@sha256:4945ad993a10e27b358b2d9869d05f23ca761162ec088e5a833939022bc23192
          createdAt: "2023-07-14"
          description: |-
            Nexus Repository is the central source of control to efficiently manage all binaries
            and build artifacts across your DevOps pipeline.
          olm.skipRange: <3.58.1-2
          repository: https://github.com/sonatype/operator-nxrm3
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus Repository
            displayName: NexusRepo
            kind: NexusRepo
            name: nexusrepos.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Repository is the central source of control to efficiently manage all binaries
          and build artifacts across your DevOps pipeline.
          The flow of open source and third-party components into and through an organization
          creates a complex software supply chain.
          Nexus Repository delivers speed, efficiency, and quality to the governance
          and management of all dependencies, libraries, and applications for your DevOps teams.

          ## Core Capabilities

          * **Dependency Management**:
            Improves reliability with repeatable, fast access to secure dependencies
          * **Developer Productivity**:
            Streamline developer workflows by enabling the sharing of components and applications across teams
          * **Supply Chain Performance**:
            Improve speed-to-market and reduced build times with release advanced staging and component tagging
          * **CI/CD Integrations**:
            Increase DevOps scalability with integrations to the most popular build and deployment tools

          Version control systems and package registries do not scale when managing proprietary,
          open source, and third-party components.
          Organizations need a central binary and build artifact repository to manage dependencies
          across the entire software supply chain.

          ## Limitations

          High Availability Clustering (HA-C) is not supported for Nexus Repository Pro for OpenShift.

          This operator will be released on a quarterly basis.

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus Repository Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage

          Once the server instance is created by the operator and running,
          you'll want to expose the service as you see fit:
          1. Create a Route to that service for nexus.port (8081).

          By default, the Nexus Repository starts up in OSS mode until a license is installed.

          The Nexus Repository can be further configured via the NexusRepo custom resource definition:

          | Parameter                                   | Description                         | Default                                 |
          | ------------------------------------------  | ----------------------------------  | ----------------------------------------|
          | `statefulset.enabled`                       | Use statefulset instead of deployment | `false` |
          | `deploymentStrategy.type`                   | Deployment Strategy     |  `Recreate` |
          | `nexus.env`                                 | Nexus environment variables         | `See example.` |
          | `nexus.resources`                           | Nexus resource requests and limits  | `{}`                                    |
          | `nexus.dockerPort`                          | Port to access docker               | `5003`                                  |
          | `nexus.nexusPort`                           | Internal port for Nexus service     | `8081`                                  |
          | `nexus.service.type`                        | Service for Nexus                   |`NodePort`                                |
          | `nexus.service.clusterIp`                   | Specific cluster IP when service type is cluster IP. Use None for headless service |`nil`   |
          | `nexus.securityContext`                     | Security Context |
          | `nexus.labels`                              | Service labels                      | `{}`                                    |
          | `nexus.podAnnotations`                      | Pod Annotations                     | `{}`
          | `nexus.livenessProbe.initialDelaySeconds`   | LivenessProbe initial delay         | 30                                      |
          | `nexus.livenessProbe.periodSeconds`         | Seconds between polls               | 30                                      |
          | `nexus.livenessProbe.failureThreshold`      | Number of attempts before failure   | 6                                       |
          | `nexus.livenessProbe.timeoutSeconds`        | Time in seconds after liveness probe times out    | `nil`                     |
          | `nexus.livenessProbe.path`                  | Path for LivenessProbe              | /                                       |
          | `nexus.readinessProbe.initialDelaySeconds`  | ReadinessProbe initial delay        | 30                                      |
          | `nexus.readinessProbe.periodSeconds`        | Seconds between polls               | 30                                      |
          | `nexus.readinessProbe.failureThreshold`     | Number of attempts before failure   | 6                                       |
          | `nexus.readinessProbe.timeoutSeconds`       | Time in seconds after readiness probe times out    | `nil`                    |
          | `nexus.readinessProbe.path`                 | Path for ReadinessProbe             | /                                       |
          | `nexus.hostAliases`                         | Aliases for IPs in /etc/hosts       | []                                      |
          | `ingress.enabled`                           | Create an ingress for Nexus         | `true`                                  |
          | `ingress.annotations`                       | Annotations to enhance ingress configuration  | `{}`                          |
          | `ingress.tls.enabled`                       | Enable TLS                          | `true`                                 |
          | `ingress.tls.secretName`                    | Name of the secret storing TLS cert, `false` to use the Ingress' default certificate | `nexus-tls`                             |
          | `ingress.path`                              | Path for ingress rules. GCP users should set to `/*` | `/`                    |
          | `tolerations`                               | tolerations list                    | `[]`                                    |
          | `config.enabled`                            | Enable configmap                    | `false`                                 |
          | `config.mountPath`                          | Path to mount the config            | `/sonatype-nexus-conf`                  |
          | `config.data`                               | Configmap data                      | `nil`                                   |
          | `deployment.terminationGracePeriodSeconds`  | Time to allow for clean shutdown    | 120                                     |
          | `deployment.annotations`                    | Annotations to enhance deployment configuration  | `{}`                       |
          | `deployment.initContainers`                 | Init containers to run before main containers  | `nil`                        |
          | `deployment.postStart.command`              | Command to run after starting the nexus container  | `nil`                    |
          | `deployment.preStart.command`               | Command to run before starting the nexus container  | `nil`                   |
          | `deployment.additionalContainers`           | Add additional Container         | `nil`                                      |
          | `deployment.additionalVolumes`              | Add additional Volumes           | `nil`                                      |
          | `deployment.additionalVolumeMounts`         | Add additional Volume mounts     | `nil`                                      |
          | `secret.enabled`                            | Enable secret                    | `false`                                    |
          | `secret.mountPath`                          | Path to mount the secret         | `/etc/secret-volume`                       |
          | `secret.readOnly`                           | Secret readonly state            | `true`                                     |
          | `secret.data`                               | Secret data                      | `nil`                                      |
          | `service.enabled`                           | Enable additional service        | `nil`                                      |
          | `service.name`                              | Service name                     | `nil`                                      |
          | `service.portName`                          | Service port name                | `nil`                                      |
          | `service.labels`                            | Service labels                   | `nil`                                      |
          | `service.annotations`                       | Service annotations              | `nil`                                      |
          | `service.loadBalancerSourceRanges`          | Service LoadBalancer source IP whitelist | `nil`                              |
          | `service.targetPort`                        | Service port                     | `nil`                                      |
          | `service.port`                              | Port for exposing service        | `nil`                                      |
          | `route.enabled`         | Set to true to create route for additional service | `false` |
          | `route.name`            | Name of route                                      | `docker` |
          | `route.portName`        | Target port name of service                        | `docker` |
          | `route.labels`          | Labels to be added to route                        | `{}` |
          | `route.annotations`     | Annotations to be added to route                   | `{}` |
          | `route.path`            | Host name of Route e.g jenkins.example.com         | nil |
        displayName: Nexus Repository Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - repository
        - sonatype
        links:
        - name: Nexus Repository Manager
          url: https://www.sonatype.com/product-nexus-repository
        maintainers:
        - email: support@sonatype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:06fe31f40ff3fd7556fc7b8f6b2efe80ada1d9079ec9aaad60d1fdc0ee0677f0
        - registry.connect.redhat.com/sonatype/nxrm-operator-certified@sha256:4945ad993a10e27b358b2d9869d05f23ca761162ec088e5a833939022bc23192
        version: 3.58.1-2
      entries:
      - name: nxrm-operator-certified.v3.58.1-2
        version: 3.58.1-2
      name: stable
    defaultChannel: stable
    packageName: nxrm-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: AMD
      provider-url: ""
    name: amd-t2-january
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amd-t2-january.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovfect2.amd.com/v1",
                "kind": "SriovT2Card",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "amd-t2-january",
                    "app.kubernetes.io/instance": "sriovt2card-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sriovt2card",
                    "app.kubernetes.io/part-of": "amd-t2-january"
                  },
                  "name": "sriovt2card",
                  "namespace": "amd-t2"
                },
                "spec": {
                  "acceleratorSelector": {
                    "pciAddress": "0000:01:00.0"
                  },
                  "nodeSelector": {
                    "kubernetes.io/hostname": "7c-c2-55-85-ba-fe.amd.com"
                  },
                  "physicalFunction": {
                    "pfDriver": "qdma-pf",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-07-14T09:46:54Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovT2Card is the Schema for the sriovt2cards API
            displayName: Sriov T2 Card
            kind: SriovT2Card
            name: sriovt2cards.sriovfect2.amd.com
            version: v1
        description: The AMD-T2 Operator for accelerator enables the seamless integration
          and management of AMD T2 accelerator within Openshift/Kubernetes envirnoment.
        displayName: AMD-T2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - SRIOV
        - Xilinx
        - FPGA
        - Telco
        - T2
        - MultiArch
        - AMD
        - TelcoAccelerator
        links:
        - name: Amd T2 January
          url: https://amd-t2-january.domain
        maintainers:
        - email: dhirajs.lakhane@amd.com
          name: DhirajLakhane
        maturity: alpha
        provider:
          name: AMD
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:928e64203edad8f1bba23593c7be04f0f8410c6e4feb98d9e9c2d00a8ff59048
        - quay.io/amdt2operator/amd-t2-january@sha256:318f56e2670582c217b8973f69f0675e7fff7dc19d07875294f2ff2fc5403902
        version: 1.0.1
      entries:
      - name: amd-t2-january.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: amd-t2-january
    provider:
      name: AMD
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Cloudera
      provider-url: https://cloudera.com
    name: cfm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cfm-operator.v2.9.0-b96
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "Nifi",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifi-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifi",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifi-sample"
                },
                "spec": {
                  "hostName": "nifi.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifi-k8s",
                    "tag": "2.8.0-b94-nifi_1.25.0.2.3.13.0-36"
                  },
                  "replicas": 1,
                  "stateManagement": {
                    "clusterProvider": {
                      "class": "org.apache.nifi.kubernetes.state.provider.KubernetesConfigMapStateProvider",
                      "id": "kubernetes-provider"
                    }
                  },
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.8.0-b94"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              },
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "NifiRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifiregistry-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifiregistry",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifiregistry-sample"
                },
                "spec": {
                  "hostName": "registry.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s",
                    "tag": "2.8.0-b94-nifi_1.25.0.2.3.13.0-36"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.8.0-b94"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning,Big Data
          com.redhat.openshift.versions: v4.10
          containerImage: container.repository.cloudera.com/cloudera/cfm-operator@sha256:890d81188143ccf9ebe2c2fe5e039cdff3a2df53f7512555673f0468147e991e
          createdAt: "2024-12-06T19:03:24Z"
          description: Cloudera Flow Management - Kubernetes Operator for Apache NiFideploys
            and operates Apache NiFi
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: A valid Cloudera license is required.
            Contact a Cloudera account team to learn more.
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NifiRegistry is the Schema for the nifiregistries API.
            displayName: Nifi Registry
            kind: NifiRegistry
            name: nifiregistries.cfm.cloudera.com
            version: v1alpha1
          - description: Nifi is the Schema for the nifis API.
            displayName: Nifi
            kind: Nifi
            name: nifis.cfm.cloudera.com
            version: v1alpha1
        description: |-
          Cloudera Flow Management - Kubernetes Operator for Apache NiFi brings enterprise-grade NiFi and NiFi Registry deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of NiFi and NiFi Registry for enterprise deployments.
          To learn more about Cloudera Flow Management - Kubernetes Operator for Apache NiFi read this
          [deep dive blog post](https://community.cloudera.com/t5/Community-Articles/Cloudera-Flow-Management-Operator-A-technical-deep-dive/ta-p/390473).

          Cloudera Flow Management - Kubernetes Operator for Apache NiFi consists of and ships multiple components including Apache NiFi, Apache NiFi Registry,
          and a custom sidecar image for logging aggregation.

          ## Installation prerequisites

          Installing Cloudera Flow Management - Kubernetes Operator for Apache NiFi requires the a valid license as well as Cloudera credentials
          (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly.
          The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) from which installation artifacts are pulled.

          ### Obtaining a license
          Contact [Cloudera sales](https://cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets
          Before installation, create two `Secrets`. One containing your Cloudera credentials and one containing your license. Cloudera credentials are provided to you
          together with your license.

          1. Create a `Secret` containing your Cloudera credentials.
              `oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***]
               --docker-server=container.repository.cloudera.com`

          2. Create a `Secret` containing your license.
              `oc create secret generic cfm-operator-license --from-file=license.txt=[***PATH TO LICENSE FILE***]`

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces in which you
            plan to deploy NiFi clusters or NiFi Registry deployments. Cloudera recommends that you create the `Secret` in all required namespaces before installation
            if you know which namespaces you will be using to deploy NiFi clusters or NiFi Registry deployments.

          ## Learn more
          To learn more, visit the [Cloudera Flow Management - Kubernetes Operator for Apache NiFi documentation library](https://docs.cloudera.com/cfm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Flow Management - Kubernetes Operator for Apache NiFi, visit the following resources:
          * [Apache NiFi](https://nifi.apache.org/)
          * [Apache NiFi Registry](https://nifi.apache.org/registry.html)
        displayName: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        - nifiregistry
        - cloudera
        - cfm
        links:
        - name: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
          url: https://docs.cloudera.com/cfm-operator
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s@sha256:914dcde9ee97894fa6f5deec285e53bf81d196e071a997cabb40fedd52deccc3
        - container.repository.cloudera.com/cloudera/cfm-tini@sha256:6493d23e20fca8daf31823403fca3dd9b42da983464fbe788a6863e141e8b66b
        - container.repository.cloudera.com/cloudera/cfm-nifi-k8s@sha256:0d700b7eff096dfc362868345826c3a6460fa43a700273b1422e95015b7727ee
        - container.repository.cloudera.com/cloudera/cfm-operator@sha256:890d81188143ccf9ebe2c2fe5e039cdff3a2df53f7512555673f0468147e991e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86af06fb3b6c17b22be47118b0a8c797fa7fdd3c00dbe99a2d2a593ca8fc3d63
        version: 2.9.0-b96
      entries:
      - name: cfm-operator.v2.9.0-b96
        version: 2.9.0-b96
      name: alpha
    - currentCSV: cfm-operator.v2.10.0-b135
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "Nifi",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifi-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifi",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifi-sample"
                },
                "spec": {
                  "hostName": "nifi.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifi-k8s",
                    "tag": "2.9.0-b96-nifi_1.27.0.2.3.14.0-14"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.9.0-b96"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              },
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "NifiRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifiregistry-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifiregistry",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifiregistry-sample"
                },
                "spec": {
                  "hostName": "registry.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s",
                    "tag": "2.9.0-b96-nifi_1.27.0.2.3.14.0-14"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.9.0-b96"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning,Big Data
          com.redhat.openshift.versions: v4.10
          containerImage: container.repository.cloudera.com/cloudera/cfm-operator@sha256:0312b6dadd5904ea9507bb857ab89d7266bf244c93b86855e0528e72813b7a85
          createdAt: "2025-05-05T14:39:51Z"
          description: Cloudera Flow Management - Kubernetes Operator for Apache NiFideploys
            and operates Apache NiFi
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: A valid Cloudera license is required.
            Contact a Cloudera account team to learn more.
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NifiRegistry is the Schema for the nifiregistries API.
            displayName: Nifi Registry
            kind: NifiRegistry
            name: nifiregistries.cfm.cloudera.com
            version: v1alpha1
          - description: Nifi is the Schema for the nifis API.
            displayName: Nifi
            kind: Nifi
            name: nifis.cfm.cloudera.com
            version: v1alpha1
        description: |-
          Cloudera Flow Management - Kubernetes Operator for Apache NiFi brings enterprise-grade NiFi and NiFi Registry deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of NiFi and NiFi Registry for enterprise deployments.
          To learn more about Cloudera Flow Management - Kubernetes Operator for Apache NiFi read this
          [deep dive blog post](https://community.cloudera.com/t5/Community-Articles/Cloudera-Flow-Management-Operator-A-technical-deep-dive/ta-p/390473).

          Cloudera Flow Management - Kubernetes Operator for Apache NiFi consists of and ships multiple components including Apache NiFi, Apache NiFi Registry,
          and a custom sidecar image for logging aggregation.

          ## Installation prerequisites

          Installing Cloudera Flow Management - Kubernetes Operator for Apache NiFi requires the a valid license as well as Cloudera credentials
          (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly.
          The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) from which installation artifacts are pulled.

          ### Obtaining a license
          Contact [Cloudera sales](https://cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets
          Before installation, create two `Secrets`. One containing your Cloudera credentials and one containing your license. Cloudera credentials are provided to you
          together with your license.

          1. Create a `Secret` containing your Cloudera credentials.
              `oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***]
               --docker-server=container.repository.cloudera.com --namespace [***OPERATOR INSTALL NAMESPACE***]`

          2. Create a `Secret` containing your license.
              `oc create secret generic cfm-operator-license --from-file=license.txt=[***PATH TO LICENSE FILE***] --namespace [***OPERATOR INSTALL NAMESPACE***]`

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces in which you
            plan to deploy NiFi clusters or NiFi Registry deployments. Cloudera recommends that you create the `Secret` in all required namespaces before installation
            if you know which namespaces you will be using to deploy NiFi clusters or NiFi Registry deployments.

          ## Learn more
          To learn more, visit the [Cloudera Flow Management - Kubernetes Operator for Apache NiFi documentation library](https://docs.cloudera.com/cfm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Flow Management - Kubernetes Operator for Apache NiFi, visit the following resources:
          * [Apache NiFi](https://nifi.apache.org/)
          * [Apache NiFi Registry](https://nifi.apache.org/registry.html)
        displayName: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        - nifiregistry
        - cloudera
        - cfm
        links:
        - name: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
          url: https://docs.cloudera.com/cfm-operator
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/cfm-nifi-k8s@sha256:8607c81906422d65784c6dca74a37af3b7c87012bca65ce28a995366014aec86
        - container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s@sha256:52e00f102579b972500eb187e64a7662a6a46caa883df743e454b1d5111b34e5
        - container.repository.cloudera.com/cloudera/cfm-operator@sha256:0312b6dadd5904ea9507bb857ab89d7266bf244c93b86855e0528e72813b7a85
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86af06fb3b6c17b22be47118b0a8c797fa7fdd3c00dbe99a2d2a593ca8fc3d63
        - container.repository.cloudera.com/cloudera/cfm-tini@sha256:d42ad5f05a9bb283fed6dbbd820855714a31662f5b851e691f13703153e90d13
        version: 2.10.0-b135
      entries:
      - name: cfm-operator.v2.10.0-b135
        version: 2.10.0-b135
      name: stable
    defaultChannel: stable
    packageName: cfm-operator
    provider:
      name: Cloudera
      url: https://cloudera.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Splunk Inc.
      provider-url: www.splunk.com
    name: splunk-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: splunk-operator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "IndexerCluster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "indexercluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "LicenseMaster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "licensemaster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "MonitoringConsole",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "monitoringconsole-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "SearchHeadCluster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "searchheadcluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "Standalone",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "standalone-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "clustermanager-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "IndexerCluster",
                "metadata": {
                  "name": "indexercluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "LicenseManager",
                "metadata": {
                  "name": "licensemanager-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "MonitoringConsole",
                "metadata": {
                  "name": "monitoringconsole-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "SearchHeadCluster",
                "metadata": {
                  "name": "searchheadcluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "Standalone",
                "metadata": {
                  "name": "standalone-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Big Data, Logging & Tracing, Monitoring, Security, AI/Machine
            Learning
          containerImage: splunk/splunk-operator@sha256:ed02561995d46ff8eea143fa78f5da118ee6fdb500832e9a33b42d059a67e229
          createdAt: "2025-09-08T10:17:59Z"
          description: The Splunk Operator for Kubernetes enables you to quickly and
            easily deploy Splunk Enterprise on your choice of private or public cloud
            provider. The Operator simplifies scaling and management of Splunk Enterprise
            by automating administrative workflows using Kubernetes best practices.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=2.4.0 <3.0.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/splunk/splunk-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterManager is the Schema for the cluster manager API
            displayName: Cluster Manager
            kind: ClusterManager
            name: clustermanagers.enterprise.splunk.com
            version: v4
          - description: ClusterMaster is the Schema for the cluster manager API
            displayName: Cluster Master
            kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v3
          - kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v1
          - kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v2
          - description: IndexerCluster is the Schema for a Splunk Enterprise indexer
              cluster
            displayName: Indexer Cluster
            kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v3
          - description: IndexerCluster is the Schema for a Splunk Enterprise indexer
              cluster
            displayName: Indexer Cluster
            kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v4
          - kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v1
          - kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v2
          - description: LicenseManager is the Schema for a Splunk Enterprise license
              manager.
            displayName: License Manager
            kind: LicenseManager
            name: licensemanagers.enterprise.splunk.com
            version: v4
          - description: LicenseMaster is the Schema for a Splunk Enterprise license
              master.
            displayName: License Master
            kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v3
          - kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v1
          - kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v2
          - description: MonitoringConsole is the Schema for the monitoringconsole
              API
            displayName: Monitoring Console
            kind: MonitoringConsole
            name: monitoringconsoles.enterprise.splunk.com
            version: v3
          - description: MonitoringConsole is the Schema for the monitoringconsole
              API
            displayName: Monitoring Console
            kind: MonitoringConsole
            name: monitoringconsoles.enterprise.splunk.com
            version: v4
          - description: SearchHeadCluster is the Schema for a Splunk Enterprise search
              head cluster
            displayName: Search Head Cluster
            kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v3
          - description: SearchHeadCluster is the Schema for a Splunk Enterprise search
              head cluster
            displayName: Search Head Cluster
            kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v4
          - kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v1
          - kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v2
          - description: Standalone is the Schema for a Splunk Enterprise standalone
              instances.
            displayName: Standalone
            kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v3
          - description: Standalone is the Schema for a Splunk Enterprise standalone
              instances.
            displayName: Standalone
            kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v4
          - kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v1
          - kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v2
        description: The Splunk Operator for Kubernetes enables you to quickly and
          easily deploy Splunk Enterprise on your choice of private or public cloud
          provider. The Operator simplifies scaling and management of Splunk Enterprise
          by automating administrative workflows using Kubernetes best practices.
        displayName: Splunk Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Splunk
        - Enterprise
        - Operator
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://github.com/splunk/splunk-operator/tree/main/docs
        maintainers:
        - email: support@splunk.com
          name: Splunk Operator Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Splunk Inc.
          url: www.splunk.com
        relatedImages:
        - splunk/splunk-operator@sha256:ed02561995d46ff8eea143fa78f5da118ee6fdb500832e9a33b42d059a67e229
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - splunk/splunk@sha256:4a5ce5a7b4e84a87396f0664895095c6510a0f2bf52e8747a373c7357f3ee313
        version: 3.0.0
      entries:
      - name: splunk-operator.v3.0.0
        version: 3.0.0
      - name: splunk-operator.v2.8.1
        version: 2.8.1
      - name: splunk-operator.v2.8.0
        version: 2.8.0
      - name: splunk-operator.v2.7.1
        version: 2.7.1
      - name: splunk-operator.v2.7.0
        version: 2.7.0
      name: stable
    defaultChannel: stable
    packageName: splunk-operator
    provider:
      name: Splunk Inc.
      url: www.splunk.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-emrcontainers-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-emrcontainers-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "emrcontainers.services.k8s.aws/v1alpha1",
                "kind": "VirtualCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/emrcontainers-controller:1.1.1
          createdAt: "2025-09-19T19:36:26Z"
          description: AWS EMR on EKS controller is a service controller for managing
            EMR on EKS resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: JobRun represents the state of an AWS emrcontainers JobRun
              resource.
            displayName: JobRun
            kind: JobRun
            name: jobruns.emrcontainers.services.k8s.aws
            version: v1alpha1
          - description: VirtualCluster represents the state of an AWS emrcontainers
              VirtualCluster resource.
            displayName: VirtualCluster
            kind: VirtualCluster
            name: virtualclusters.emrcontainers.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EMR on EKS resources in AWS from within your Kubernetes cluster.

          **About Amazon EMR on EKS**

          Amazon EMR on EKS provides a deployment option for Amazon EMR that allows you to run open-source big data frameworks on Amazon Elastic Kubernetes Service (Amazon EKS). With this deployment option, you can focus on running analytics workloads while Amazon EMR on EKS builds, configures, and manages containers for open-source applications.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EMR on EKS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - emrcontainers
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EMR on EKS Developer Resources
          url: https://docs.aws.amazon.com/emr/latest/EMR-on-EKS-DevelopmentGuide/emr-eks.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: emr on eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/emrcontainers-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-emrcontainers-controller.v1.1.1
        version: 1.1.1
      - name: ack-emrcontainers-controller.v1.1.0
        version: 1.1.0
      - name: ack-emrcontainers-controller.v1.0.27
        version: 1.0.27
      - name: ack-emrcontainers-controller.v1.0.25
        version: 1.0.25
      - name: ack-emrcontainers-controller.v1.0.24
        version: 1.0.24
      - name: ack-emrcontainers-controller.v1.0.23
        version: 1.0.23
      - name: ack-emrcontainers-controller.v1.0.22
        version: 1.0.22
      - name: ack-emrcontainers-controller.v1.0.21
        version: 1.0.21
      - name: ack-emrcontainers-controller.v1.0.18
        version: 1.0.18
      - name: ack-emrcontainers-controller.v1.0.16
        version: 1.0.16
      - name: ack-emrcontainers-controller.v1.0.15
        version: 1.0.15
      - name: ack-emrcontainers-controller.v1.0.14
        version: 1.0.14
      - name: ack-emrcontainers-controller.v1.0.13
        version: 1.0.13
      - name: ack-emrcontainers-controller.v1.0.12
        version: 1.0.12
      - name: ack-emrcontainers-controller.v1.0.11
        version: 1.0.11
      - name: ack-emrcontainers-controller.v1.0.9
        version: 1.0.9
      - name: ack-emrcontainers-controller.v1.0.8
        version: 1.0.8
      - name: ack-emrcontainers-controller.v1.0.4
        version: 1.0.4
      - name: ack-emrcontainers-controller.v1.0.3
        version: 1.0.3
      - name: ack-emrcontainers-controller.v1.0.1
        version: 1.0.1
      - name: ack-emrcontainers-controller.v1.0.0
        version: 1.0.0
      - name: ack-emrcontainers-controller.v0.1.2
        version: 0.1.2
      - name: ack-emrcontainers-controller.v0.1.0
        version: 0.1.0
      - name: ack-emrcontainers-controller.v0.0.10
        version: 0.0.10
      - name: ack-emrcontainers-controller.v0.0.9
        version: 0.0.9
      - name: ack-emrcontainers-controller.v0.0.8
        version: 0.0.8
      - name: ack-emrcontainers-controller.v0.0.7
        version: 0.0.7
      - name: ack-emrcontainers-controller.v0.0.6
        version: 0.0.6
      - name: ack-emrcontainers-controller.v0.0.5
        version: 0.0.5
      - name: ack-emrcontainers-controller.v0.0.4
        version: 0.0.4
      - name: ack-emrcontainers-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-emrcontainers-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/cluster-nfd-operator
    name: nfd
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfd.4.16.0-202507211806
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nfd.openshift.io/v1",
                "kind": "NodeFeatureDiscovery",
                "metadata": {
                  "name": "nfd-instance",
                  "namespace": "openshift-nfd"
                },
                "spec": {
                  "customConfig": {
                    "configData": "#    - name: \"more.kernel.features\"\n#      matchOn:\n#      - loadedKMod: [\"example_kmod3\"]\n#    - name: \"more.features.by.nodename\"\n#      value: customValue\n#      matchOn:\n#      - nodename: [\"special-.*-node-.*\"]\n"
                  },
                  "operand": {
                    "image": "registry.redhat.io/openshift4/ose-node-feature-discovery-rhel9@sha256:bb65ec6a6f88688630305e081cb4bcb88aa538b5f6a3385ee28c61522dcd1662",
                    "imagePullPolicy": "IfNotPresent",
                    "servicePort": 12000
                  },
                  "workerConfig": {
                    "configData": "core:\n#  labelWhiteList:\n#  noPublish: false\n  sleepInterval: 60s\n#  sources: [all]\n#  klog:\n#    addDirHeader: false\n#    alsologtostderr: false\n#    logBacktraceAt:\n#    logtostderr: true\n#    skipHeaders: false\n#    stderrthreshold: 2\n#    v: 0\n#    vmodule:\n##   NOTE: the following options are not dynamically run-time \n##          configurable and require a nfd-worker restart to take effect\n##          after being changed\n#    logDir:\n#    logFile:\n#    logFileMaxSize: 1800\n#    skipLogHeaders: false\nsources:\n#  cpu:\n#    cpuid:\n##     NOTE: whitelist has priority over blacklist\n#      attributeBlacklist:\n#        - \"BMI1\"\n#        - \"BMI2\"\n#        - \"CLMUL\"\n#        - \"CMOV\"\n#        - \"CX16\"\n#        - \"ERMS\"\n#        - \"F16C\"\n#        - \"HTT\"\n#        - \"LZCNT\"\n#        - \"MMX\"\n#        - \"MMXEXT\"\n#        - \"NX\"\n#        - \"POPCNT\"\n#        - \"RDRAND\"\n#        - \"RDSEED\"\n#        - \"RDTSCP\"\n#        - \"SGX\"\n#        - \"SSE\"\n#        - \"SSE2\"\n#        - \"SSE3\"\n#        - \"SSE4.1\"\n#        - \"SSE4.2\"\n#        - \"SSSE3\"\n#      attributeWhitelist:\n#  kernel:\n#    kconfigFile: \"/path/to/kconfig\"\n#    configOpts:\n#      - \"NO_HZ\"\n#      - \"X86\"\n#      - \"DMI\"\n  pci:\n    deviceClassWhitelist:\n      - \"0200\"\n      - \"03\"\n      - \"12\"\n    deviceLabelFields:\n#      - \"class\"\n      - \"vendor\"\n#      - \"device\"\n#      - \"subsystem_vendor\"\n#      - \"subsystem_device\"\n#  usb:\n#    deviceClassWhitelist:\n#      - \"0e\"\n#      - \"ef\"\n#      - \"fe\"\n#      - \"ff\"\n#    deviceLabelFields:\n#      - \"class\"\n#      - \"vendor\"\n#      - \"device\"\n#  custom:\n#    - name: \"my.kernel.feature\"\n#      matchOn:\n#        - loadedKMod: [\"example_kmod1\", \"example_kmod2\"]\n#    - name: \"my.pci.feature\"\n#      matchOn:\n#        - pciId:\n#            class: [\"0200\"]\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#        - pciId :\n#            vendor: [\"8086\"]\n#            device: [\"1000\", \"1100\"]\n#    - name: \"my.usb.feature\"\n#      matchOn:\n#        - usbId:\n#          class: [\"ff\"]\n#          vendor: [\"03e7\"]\n#          device: [\"2485\"]\n#        - usbId:\n#          class: [\"fe\"]\n#          vendor: [\"1a6e\"]\n#          device: [\"089a\"]\n#    - name: \"my.combined.feature\"\n#      matchOn:\n#        - pciId:\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#          loadedKMod : [\"vendor_kmod1\", \"vendor_kmod2\"]\n"
                  }
                }
              },
              {
                "apiVersion": "nfd.openshift.io/v1alpha1",
                "kind": "NodeFeatureRule",
                "metadata": {
                  "name": "my-sample-rule-object"
                },
                "spec": {
                  "rules": [
                    {
                      "labels": {
                        "my-sample-feature": "true"
                      },
                      "matchFeatures": [
                        {
                          "feature": "kernel.loadedmodule",
                          "matchExpressions": {
                            "dummy": {
                              "op": "Exists"
                            }
                          }
                        },
                        {
                          "feature": "kernel.config",
                          "matchExpressions": {
                            "X86": {
                              "op": "In",
                              "value": [
                                "y"
                              ]
                            }
                          }
                        }
                      ],
                      "name": "my sample rule"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery,OpenShift Optional
          containerImage: registry.redhat.io/openshift4/ose-cluster-nfd-rhel9-operator@sha256:c1b6a5b8a59c08458d57c0fc2bb2bc9b056f006d15edb4d3161a3d1c0bc97ae6
          description: |-
            The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

            NFD consists  of the following software components:

            The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

            ##NFD-Master
            NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

            ##NFD-Worker
            NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

            ##NFD-Topology-Updater
            NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.0-202507211806'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-nfd
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/internal-objects: '["noderesourcetopologies.topology.node.k8s.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/cluster-nfd-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              The NodeFeatureDiscovery instance is the CustomResource being watched by the NFD-Operator, and holds all the needed information to setup the behaviour of the master and worker pods
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
          - description: |
              NodeFeatureRule resource specifies a configuration for feature-based customization of node objects, such as node labeling.
            kind: NodeFeatureRule
            name: nodefeaturerules.nfd.openshift.io
            version: v1alpha1
          - kind: NodeFeature
            name: nodefeatures.nfd.openshift.io
            version: v1alpha1
        description: |-
          The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

          NFD consists  of the following software components:

          The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

          ##NFD-Master
          NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

          ##NFD-Worker
          NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

          ##NFD-Topology-Updater
          NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
        displayName: Node Feature Discovery Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - feature-discovery
        - feature-detection
        - node-labels
        links:
        - name: Node Feature Discovery Operator
          url: https://docs.openshift.com/container-platform/4.16/hardware_enablement/psap-node-feature-discovery-operator.html
        - name: Node Feature Discovery Documentation
          url: https://kubernetes-sigs.github.io/node-feature-discovery/stable/get-started/index.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/cluster-nfd-operator
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:bcca89fb000b0fe8722ad7807b62a43ef3ae8fa76db1d790c285a975bfecb1aa
        - registry.redhat.io/openshift4/ose-cluster-nfd-rhel9-operator@sha256:c1b6a5b8a59c08458d57c0fc2bb2bc9b056f006d15edb4d3161a3d1c0bc97ae6
        - registry.redhat.io/openshift4/ose-node-feature-discovery-rhel9@sha256:bb65ec6a6f88688630305e081cb4bcb88aa538b5f6a3385ee28c61522dcd1662
        version: 4.16.0-202507211806
      entries:
      - name: nfd.4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: nfd.4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: nfd.4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: nfd.4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: nfd.4.16.0-202504271139
        version: 4.16.0-202504271139
      - name: nfd.4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: nfd.4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: nfd.4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: nfd.4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: nfd.4.16.0-202501271512
        version: 4.16.0-202501271512
      - name: nfd.4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: nfd.4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: nfd.4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: nfd.4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: nfd.4.16.0-202410251436
        version: 4.16.0-202410251436
      - name: nfd.4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: nfd.4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: nfd.4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: nfd.4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: nfd.4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: nfd.4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: nfd.4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: nfd.4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: nfd.4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: nfd.4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: nfd.4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: nfd.4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: nfd.4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: nfd.4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: nfd
    provider:
      name: Red Hat
      url: https://github.com/openshift/cluster-nfd-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhpam-kogito-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhpam-kogito-operator.7.13.5-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "ruleunit-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogitoinfra-sample"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "kafka.strimzi.io/v1beta2",
                    "kind": "Kafka",
                    "name": "kogito-kafka"
                  }
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator@sha256:19497d4aff0ea3ea114dccb0d0e265458bb14ba6d325353eb82b554547dbd2e4
          description: RHPAM Kogito Operator for deployment and management of Kogito
            services.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Process Automation
            Manager", "Red Hat Decision Manager"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/rhpam-kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.rhpam.kiegroup.org
            version: v1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.rhpam.kiegroup.org
            version: v1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.rhpam.kiegroup.org
            version: v1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.rhpam.kiegroup.org
            version: v1
        description: |-
          RHPAM Kogito Operator is designed for deployment and management of Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Builds a RHPAM Kogito service from a remote git repository and creates a custom image for the given runtime: Red Hat build of Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Deploy RHPAM Kogito service from user provided image.
          * Setup Grafana Dashboard when Grafana Operator is available in cluster.
          * Create Prometheus monitor object when Prometheus operator is available in cluster.

          **IMPORTANT!** RHPAM Kogito Operator may need Grafana and Prometheus Operator to configure monitoring in cases if you do not provide these services beforehand. If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the RHPAM Kogito Operator.
        displayName: RHPAM Kogito Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-process-automation-manager
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_process_automation_manager/7.13/html/getting_started_with_red_hat_build_of_kogito_in_red_hat_process_automation_manager/assembly-deploying-kogito-microservices-on-openshift
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator@sha256:19497d4aff0ea3ea114dccb0d0e265458bb14ba6d325353eb82b554547dbd2e4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9492c67eaa7ab37df6f5fa623c4aac955b7d7ffaab9ae2e57dea745e82907e40
        version: 7.13.5-4
      entries:
      - name: rhpam-kogito-operator.7.13.5-4
        version: 7.13.5-4
      - name: rhpam-kogito-operator.7.13.5-3
        version: 7.13.5-3
      - name: rhpam-kogito-operator.7.13.5-2
        version: 7.13.5-2
      - name: rhpam-kogito-operator.7.13.5-1
        version: 7.13.5-1
      - name: rhpam-kogito-operator.7.13.4-3
        version: 7.13.4-3
      - name: rhpam-kogito-operator.7.13.4-2
        version: 7.13.4-2
      - name: rhpam-kogito-operator.7.13.4-1
        version: 7.13.4-1
      - name: rhpam-kogito-operator.7.13.3-1
        version: 7.13.3-1
      - name: rhpam-kogito-operator.7.13.2-3
        version: 7.13.2-3
      - name: rhpam-kogito-operator.7.13.2-2
        version: 7.13.2-2
      - name: rhpam-kogito-operator.7.13.2-1
        version: 7.13.2-1
      - name: rhpam-kogito-operator.7.13.1-2
        version: 7.13.1-2
      - name: rhpam-kogito-operator.7.13.1-1
        version: 7.13.1-1
      - name: rhpam-kogito-operator.7.13.0-1
        version: 7.13.0-1
      - name: rhpam-kogito-operator.v7.12.0-4
        version: 7.12.0-4
      - name: rhpam-kogito-operator.v7.12.0-3
        version: 7.12.0-3
      - name: rhpam-kogito-operator.v7.12.0-2
        version: 7.12.0-2
      name: 7.x
    defaultChannel: 7.x
    packageName: rhpam-kogito-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://skupper.io
    name: skupper-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-operator.v1.4.8-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus@sha256:65602221cdff7f718d4149f4dead167d05dbbe0cea1eaa8bf4b98be588135583
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:c665677c639699d6a5ef24aed5a1a3d9c0a5ee25f077a6db42a9fc6247d93132
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:eedde9f2019d4e919612bcc18f427c0d88b4846e773751af231ba0e560d48869
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:1fd908e80ef21fe7d7051c400cb099f2f3ca51c8cdd698705ef9784d34688911
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:4c9cc9df83baa1ef37cfb429babe5b86a962b8bb096209b3fb74291c67c94000
        version: 1.4.8-rh-1
      entries:
      - name: skupper-operator.v1.4.8-rh-1
        version: 1.4.8-rh-1
      - name: skupper-operator.v1.4.7-rh-3
        version: 1.4.7-rh-3
      - name: skupper-operator.v1.4.7-rh-2
        version: 1.4.7-rh-2
      - name: skupper-operator.v1.4.7-rh-1
        version: 1.4.7-rh-1
      - name: skupper-operator.v1.4.5-rh-4
        version: 1.4.5-rh-4
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1
    - currentCSV: skupper-operator.v1.4.8-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:eedde9f2019d4e919612bcc18f427c0d88b4846e773751af231ba0e560d48869
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:1fd908e80ef21fe7d7051c400cb099f2f3ca51c8cdd698705ef9784d34688911
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:4c9cc9df83baa1ef37cfb429babe5b86a962b8bb096209b3fb74291c67c94000
        - registry.redhat.io/openshift4/ose-prometheus@sha256:65602221cdff7f718d4149f4dead167d05dbbe0cea1eaa8bf4b98be588135583
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:c665677c639699d6a5ef24aed5a1a3d9c0a5ee25f077a6db42a9fc6247d93132
        version: 1.4.8-rh-1
      entries:
      - name: skupper-operator.v1.4.8-rh-1
        version: 1.4.8-rh-1
      - name: skupper-operator.v1.4.7-rh-3
        version: 1.4.7-rh-3
      - name: skupper-operator.v1.4.7-rh-2
        version: 1.4.7-rh-2
      - name: skupper-operator.v1.4.7-rh-1
        version: 1.4.7-rh-1
      - name: skupper-operator.v1.4.5-rh-4
        version: 1.4.5-rh-4
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.4
    - currentCSV: skupper-operator.v1.5.5-rh-6
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:9e9e1f2288f0218e3ba2abf4bebac9731d7cf9f9c567e0f6de6767b0b66cbee4
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:a5d7c1b2b9db672548126c0873d062a55667d93291fd3e32809d3c9c900c548e
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:9e9e1f2288f0218e3ba2abf4bebac9731d7cf9f9c567e0f6de6767b0b66cbee4
        - registry.redhat.io/openshift4/ose-prometheus@sha256:53afaf53975807765fc3777314edd218c789c17680c1f57a8e33a8f9a76f38bf
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:3e78ec8cdb729a704d5a26e54697f285ba7c0b9afee6160b7ecc703876120dea
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:be9dbcd2a3a1e1e536921390dcfd283fc45dc4423d2705376ef37a9fb53f7cf1
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:1de071dabf620265252f3922c632f17f676ba716ce4c3df931a0ccb3871dbb6c
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:5f8621d0013febcecbed4d7c82db11a4dded99ad8ce6b6d83858b9bdbe79ee5a
        version: 1.5.5-rh-6
      entries:
      - name: skupper-operator.v1.5.5-rh-6
        version: 1.5.5-rh-6
      - name: skupper-operator.v1.5.5-rh-5
        version: 1.5.5-rh-5
      - name: skupper-operator.v1.5.5-rh-4
        version: 1.5.5-rh-4
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.5
    - currentCSV: skupper-operator.v1.8.7-rh-2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:b0d43628b779a101a24e4b271318a34c21f3b71bcba2bf20804f941a93d27455
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.8.7-rh-2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:56b8fde75a6e7df235c560a0f4b9114398da5ba8b20f75f1ea7338f48ce19506
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:d7d252cff08ba1589afd8707ed349f87d3b6214906a4fb49932bda0a4d19fee4
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:5ea36e96f878577104ab7601e57e7a2a233f11dbdc533454f95e2e14a066d21f
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:3adeef827eca0382cd410f7c3e99a0fee9b50791cb0e0c6fc130b72a2550684f
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:b0d43628b779a101a24e4b271318a34c21f3b71bcba2bf20804f941a93d27455
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:f798745e799515065af099a816057b9c98a741e951d30a0f45ed47ca50e6724e
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:61542c5f0989f51ac5d6c96f5d0e29aca75614b79e7626299c77c5a74636b564
        version: 1.8.7-rh-2
      entries:
      - name: skupper-operator.v1.8.7-rh-2
        version: 1.8.7-rh-2
      - name: skupper-operator.v1.8.7-rh-1
        version: 1.8.7-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.8
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.9
    - currentCSV: skupper-operator.v2.1.1-rh-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
          createdAt: "2025-08-19T20:43:19Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:36872037066fae806cbf88cab673229f25e592295972aa18227827233362eec5
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:c01d0eab0841b1a5fddb58a0ae110beb464b91390f02662590bebc7b2a7213b6
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e8df53454cf12a14338790af6ed209798010990ca3320096e8b44acf901c4123
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:1a290a5ac8b9d1b082e740054675fe432aa424c2f55dd662caf67c740044f229
        version: 2.1.1-rh-3
      entries:
      - name: skupper-operator.v2.1.1-rh-3
        version: 2.1.1-rh-3
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2
    - currentCSV: skupper-operator.v2.0.1-rh-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:b5c73dc59a9184cd120c2ee35256cc097c790b8ef488af0c732d5efe3192b768
          createdAt: "2025-05-29T13:35:40Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.0.1-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:92f44cff160874efe57a985609e599b13c49c0c4cef846dccc80833af2ae74f3
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:e239c859a1f49816ae3068455e703b8a12919f11158b0f6ee0e6bfb31e8a8082
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:f2a41674ed196c089748d6386c18c14580e502de33235ef23c37eddeb6350f41
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:95062ff1256135cfa17907e82b09b2a211371ea0e76ccc711b7c737e040d87e9
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:498b5bb648e5bbaad07663941600173abac7d60a10744305f619509ddf7563e0
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:bd3f72544841c91c2d573a65ac3c988756f3523e17e0988027793b0f5f3ec3d1
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:b5c73dc59a9184cd120c2ee35256cc097c790b8ef488af0c732d5efe3192b768
        version: 2.0.1-rh-2
      entries:
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2.0
    - currentCSV: skupper-operator.v2.1.1-rh-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
          createdAt: "2025-08-19T20:43:19Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:c01d0eab0841b1a5fddb58a0ae110beb464b91390f02662590bebc7b2a7213b6
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e8df53454cf12a14338790af6ed209798010990ca3320096e8b44acf901c4123
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:1a290a5ac8b9d1b082e740054675fe432aa424c2f55dd662caf67c740044f229
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:36872037066fae806cbf88cab673229f25e592295972aa18227827233362eec5
        version: 2.1.1-rh-3
      entries:
      - name: skupper-operator.v2.1.1-rh-3
        version: 2.1.1-rh-3
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2.1
    defaultChannel: stable-2
    packageName: skupper-operator
    provider:
      name: Red Hat
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: placement-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: placement-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "placement.openstack.org/v1beta1",
                "kind": "PlacementAPI",
                "metadata": {
                  "name": "placement"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "placement",
                  "databaseInstance": "openstack",
                  "preserveJobs": false,
                  "replicas": 1,
                  "secret": "placement-secret",
                  "serviceUser": "placement"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:43Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PlacementAPI is the Schema for the placementapis API
            displayName: Placement API
            kind: PlacementAPI
            name: placementapis.placement.openstack.org
            version: v1beta1
        description: Placement Operator
        displayName: Placement Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Placement
        links:
        - name: Placement Operator
          url: https://github.com/openstack-k8s-operators/placement-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/placement-rhel9-operator@sha256:d3be48618b1f0f717aea162ae41876137fee8b35329d2e3e66e9d4eaccf62376
        - registry.redhat.io/rhoso/openstack-placement-api-rhel9@sha256:3308f2eccfbe7416a85443696273417bfb816df36b810ebbefa108dda12a5b19
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: placement-operator.v1.0.6
        version: 1.0.6
      - name: placement-operator.v1.0.5
        version: 1.0.5
      - name: placement-operator.v1.0.4
        version: 1.0.4
      - name: placement-operator.v1.0.3
        version: 1.0.3
      - name: placement-operator.v1.0.2
        version: 1.0.2
      - name: placement-operator.v1.0.1
        version: 1.0.1
      - name: placement-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: placement-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/stolostron/multicluster-global-hub
    name: multicluster-global-hub-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-global-hub-operator.v1.4.0-dev.20250302
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2025-02-14T08:24:47Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: planning
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/postgresql-16:9.5-1732622748
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-inventory/inventory-api:latest
        - quay.io/prometheuscommunity/postgres-exporter:v0.15.0
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-relations/spicedb-operator:latest
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-relations/spicedb:latest
        - quay.io/stolostron/grafana:2.12.0-SNAPSHOT-2024-09-03-21-11-25
        - quay.io/stolostron/multicluster-global-hub-operator:v1.4.0
        - quay.io/stolostron/multicluster-global-hub-agent:v1.4.0
        - quay.io/stolostron/multicluster-global-hub-manager:v1.4.0
        version: 1.4.0-dev.20250302
      entries:
      - name: multicluster-global-hub-operator.v1.4.0-dev.20250302
        version: 1.4.0-dev.20250302
      - name: multicluster-global-hub-operator.v1.4.0-dev.20250225
        version: 1.4.0-dev.20250225
      - name: multicluster-global-hub-operator.v1.3.0-dev.20240911
        version: 1.3.0-dev.20240911
      - name: multicluster-global-hub-operator.v1.3.0-dev.20240724
        version: 1.3.0-dev.20240724
      - name: multicluster-global-hub-operator.v1.2.0-dev.20240516
        version: 1.2.0-dev.20240516
      - name: multicluster-global-hub-operator.v1.2.0-dev
        version: 1.2.0-dev
      - name: multicluster-global-hub-operator.v1.1.0-dev
        version: 1.1.0-dev
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: multicluster-global-hub-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-operator-globalhub-1-5@sha256:4db1251323bd64930461cb5025da41baffcf5c00c51ea4e49940b7130971b510
          createdAt: "2025-06-10T06:58:35Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-operator-globalhub-1-5@sha256:4db1251323bd64930461cb5025da41baffcf5c00c51ea4e49940b7130971b510
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-manager-globalhub-1-5@sha256:01253426d65f1297a1955a8393051e6a4b1df5e5249774fc5189a75f35e7d644
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-agent-globalhub-1-5@sha256:6f7520eac83abec7780f831d8473544d0b8db43fc6e77d59e59317b09eb1f6c1
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/glo-grafana-globalhub-1-5@sha256:73fdf00538e63595b7cc92d484e087ff815579a95d7259c579c6e7dc1beacff6
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/postgres-exporter-globalhub-1-5@sha256:1108572a8bdafde926254f1ff94c2b22e048706be660b07c793e9f9e038a0f4a
        version: 1.5.0
      entries:
      - name: multicluster-global-hub-operator.v1.5.0
        version: 1.5.0
      - name: multicluster-global-hub-operator.v1.4.0
        version: 1.4.0
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-1.5
    - currentCSV: multicluster-global-hub-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2024-03-05T02:13:54Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.32.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:latest
        version: 1.1.0
      entries:
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.10
    - currentCSV: multicluster-global-hub-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2024-07-03T06:52:02Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v1.2.0
        version: 1.2.1
      entries:
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.11
    - currentCSV: multicluster-global-hub-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator@sha256:636aea861da80fc36fb27a419a18c5b3fbb1bc072e8ae33c75e1cc47e2c05f75
          createdAt: "2024-10-10T05:42:48Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          kafka-zookeeper-0                                       1/1     Running   0          21m
          kafka-zookeeper-1                                       1/1     Running   0          21m
          kafka-zookeeper-2                                       1/1     Running   0          21m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgres-0                      2/2     Running   0          22m
          ```
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:636aea861da80fc36fb27a419a18c5b3fbb1bc072e8ae33c75e1cc47e2c05f75
        - quay.io/prometheuscommunity/postgres-exporter@sha256:386b12d19eab2a37d7cd8ca8b4c7491cc7a830d9581f49af6c98a393da9605e6
        - quay.io/stolostron/inventory-api@sha256:a15e62ee070dba5d74cf336b3b89b467593c5a33cfe3623d922f0af4fd7df392
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/release-globalhub-1-3/multicluster-global-hub-manager-globalhub-1-3@sha256:7656d01a4d7b39c61c3249695d68011d9e0df308627b9efd117686e901281ab1
        - quay.io/stolostron/multicluster-global-hub-manager@sha256:715cb5d0c6c407ec5c0303fc7e6d83fc77ce7e06775e169a475f7c075f1e7000
        - quay.io/stolostron/grafana@sha256:719b320a05d10158fadce5ae2fc6e4254a18957a6608b6f81e876cf46d3a6897
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        version: 1.3.0
      entries:
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.12
    - currentCSV: multicluster-global-hub-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator@sha256:085e9fc6cdc6a42df10a38856c8ccff25e5acffaba67ef95da72027e90e507fc
          createdAt: "2025-02-14T08:24:47Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:085e9fc6cdc6a42df10a38856c8ccff25e5acffaba67ef95da72027e90e507fc
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-inventory/inventory-api@sha256:6a55a1919ecd92d94e05c12be1b6450956f5fda2ce17af7027cac3806c4970a4
        - quay.io/prometheuscommunity/postgres-exporter@sha256:386b12d19eab2a37d7cd8ca8b4c7491cc7a830d9581f49af6c98a393da9605e6
        - quay.io/stolostron/multicluster-global-hub-manager@sha256:4cd602ff6f8dd951848dc8a95c856004295c9f4710f77d4dafcc0735c1492395
        - quay.io/stolostron/multicluster-global-hub-agent@sha256:8dfc33dfed33a85733ee5e02dcd4f05662c75263791e0961ba17cbb3d0c340c9
        - quay.io/stolostron/grafana@sha256:719b320a05d10158fadce5ae2fc6e4254a18957a6608b6f81e876cf46d3a6897
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        version: 1.4.0
      entries:
      - name: multicluster-global-hub-operator.v1.4.0
        version: 1.4.0
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.13
    - currentCSV: multicluster-global-hub-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "storage": {
                    "name": "storage-secret"
                  },
                  "transport": {
                    "name": "transport-secret"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.5.0
          createdAt: "2022-08-23T20:00:31Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha1
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.5 or later needs to be installed.\n- PostgreSQL needs to be installed
          and you must create database for the multicluster global hub. You must create
          a secret named postgresql-secret that contains the database access credential
          in `open-cluster-management` namespace. You can run the following command
          to create the secret: \n\n```\nkubectl create secret generic storage-secret
          -n open-cluster-management \\\n  --from-literal=database_uri=<postgresql-uri>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)
          to install postgres in `hoh-postgres` namespace and create the secret `storage-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n- Kafka needs to be installed,
          and you must create two sections, `spec` and `status`. You need to create
          a secret named kafka-secret that contains the kafka access information should
          be created in `open-cluster-management` namespace. You can run the following
          command to create the secret: \n\n```\nkubectl create secret generic transport-secret
          -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n  --from-literal=CA=<CA-for-kafka-server>\n```\n >_Note:_ There is a
          sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `transport-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n## How to Install\n\nInstall the
          Multicluster Global Hub Operator by following the instructions that are
          displayed after you select the Install button. The operator must be installed
          in the same namespace as Red Hat Advanced Cluster Management for Kubernetes
          (RHACM), by default, in `open-cluster-management` namespace.\n\nA pod will
          be created in `open-cluster-management` namespace\n\n```\n$ kubectl get
          pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          19s\n```\n\nThe operator is
          now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installing the operator,
          create an instance of the MulticlusterGlobalHub resource to install the
          Multicluster Global Hub.\n\nThe pods of Multicluster Global Hub are created
          in the `open-cluster-management` namespace:\n\n```\n$ kubectl get pods -n
          open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\nmulticluster-global-hub-rbac-77888b7dc4-z9vbp
          \                   1/1     Running   0          10s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.5.0
        version: 0.5.0
      entries:
      - name: multicluster-global-hub-operator.v0.5.0
        version: 0.5.0
      name: release-2.6
    - currentCSV: multicluster-global-hub-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha2",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "largeScale": {
                      "kafka": {
                        "name": "transport-secret"
                      },
                      "postgres": {
                        "name": "storage-secret"
                      }
                    },
                    "type": "largeScale"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.6.0
          createdAt: "2023-02-09T20:00:31Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha2
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.7 or later needs to be installed.\n- PostgreSQL needs to be installed
          and you must create database for the multicluster global hub. You must create
          a secret named postgresql-secret that contains the database access credential
          in `open-cluster-management` namespace. You can run the following command
          to create the secret: \n\n```\nkubectl create secret generic storage-secret
          -n open-cluster-management \\\n  --from-literal=database_uri=<postgresql-uri>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)
          to install postgres in `hoh-postgres` namespace and create the secret `storage-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n- Kafka needs to be installed,
          and you must create two sections, `spec` and `status`. You need to create
          a secret named kafka-secret that contains the kafka access information should
          be created in `open-cluster-management` namespace. You can run the following
          command to create the secret: \n\n```\nkubectl create secret generic transport-secret
          -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n  --from-literal=CA=<CA-for-kafka-server>\n```\n >_Note:_ There is a
          sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `transport-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n## How to Install\n\nInstall the
          Multicluster Global Hub Operator by following the instructions that are
          displayed after you select the Install button. The operator must be installed
          in the same namespace as Red Hat Advanced Cluster Management for Kubernetes
          (RHACM), by default, in `open-cluster-management` namespace.\n\nA pod will
          be created in `open-cluster-management` namespace\n\n```\n$ kubectl get
          pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          19s\n```\n\nThe operator is
          now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installing the operator,
          create an instance of the MulticlusterGlobalHub resource to install the
          Multicluster Global Hub.\n\nThe pods of Multicluster Global Hub are created
          in the `open-cluster-management` namespace:\n\n```\n$ kubectl get pods -n
          open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: multicluster-global-hub-operator.v0.6.0
        version: 0.6.0
      name: release-2.7
    - currentCSV: multicluster-global-hub-operator.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha3",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "largeScale": {
                      "kafka": {
                        "transportFormat": "cloudEvents"
                      }
                    },
                    "type": "largeScale"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.7.0
          createdAt: "2023-07-19T07:53:07Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha3
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.7 or later needs to be installed.\n## How to Install\n\nInstall the Multicluster
          Global Hub Operator by following the instructions that are displayed after
          you select the Install button. The operator must be installed in the same
          namespace as Red Hat Advanced Cluster Management for Kubernetes (RHACM),
          by default, in `open-cluster-management` namespace.\n\nA pod will be created
          in `open-cluster-management` namespace\n\n```\n$ kubectl get pods -n open-cluster-management\nNAME
          \                                                            READY   STATUS
          \   RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp                1/1
          \    Running   0          19s\n```\n\nThe operator is now providing new
          Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installed the operator,
          create an instance of the MulticlusterGlobalHub resource to instantiate
          the Multicluster Global Hub.\n\nYou need to prepare:\n\n- PostgreSQL needs
          to be installed and you must create database for the multicluster global
          hub. You must create a secret named multicluster-global-hub-storage that
          contains the database access credential in `open-cluster-management` namespace.
          You can run the following command to create the secret: \n\n```\nkubectl
          create secret generic multicluster-global-hub-storage -n open-cluster-management
          \\\n  --from-literal=database_uri=<postgresql-uri> \\\n --from-file=ca.crt=<CA-for-postgres-server>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)(Note:
          the client version of kubectl must be v1.21+) to install postgres in `hoh-postgres`
          namespace and create the secret `multicluster-global-hub-storage` in namespace
          `open-cluster-management` automatically. To override the secret namespace,
          set `TARGET_NAMESPACE` environment variable to the RHACM installation namespace
          before executing the script.\n\n- Kafka needs to be installed, and you must
          create three topics, `spec` `status` and `event`. You need to create a secret
          named multicluster-global-hub-transport that contains the kafka access information
          should be created in `open-cluster-management` namespace. You can run the
          following command to create the secret: \n\n```\nkubectl create secret generic
          multicluster-global-hub-transport -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n      --from-file=ca.crt=<CA-cert-for-kafka-server> --from-file=client.crt=<Client-cert-for-kafka-server>
          --from-file=client.key=<Client-key-for-kafka-server> \n```\n >_Note:_ There
          is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `multicluster-global-hub-transport`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\nOnce an instance of the MulticlusterGlobalHub
          is created, the following pods are created in the `open-cluster-management`
          namespace:\n\n```\n$ kubectl get pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-grafana-79f8d59f58-qwcjl
          \                1/1     Running   0          22s\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: beta
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.7.0
        version: 0.7.0
      entries:
      - name: multicluster-global-hub-operator.v0.7.0
        version: 0.7.0
      name: release-2.8
    - currentCSV: multicluster-global-hub-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v1.0.0
          createdAt: "2023-10-16T03:07:33Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the MultiCluster Global Hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.7 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```

          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v1.0.0
        version: 1.0.0
      entries:
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.9
    defaultChannel: release-1.5
    packageName: multicluster-global-hub-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/stolostron/multicluster-global-hub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhbk-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhbk-operator.v22.0.13-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
          createdAt: "2024-09-17T11:20:22Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:884afe0ca1698d85a02aa47b6d60b8222e50d70ad280db6f98e1a4bbdbe13c40
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
        version: 22.0.13-opr.1
      entries:
      - name: rhbk-operator.v22.0.13-opr.1
        version: 22.0.13-opr.1
      - name: rhbk-operator.v22.0.12-opr.1
        version: 22.0.12-opr.1
      - name: rhbk-operator.v22.0.11-opr.2
        version: 22.0.11-opr.2
      - name: rhbk-operator.v22.0.11-opr.1
        version: 22.0.11-opr.1
      - name: rhbk-operator.v22.0.10-opr.1
        version: 22.0.10-opr.1
      - name: rhbk-operator.v22.0.9-opr.1
        version: 22.0.9-opr.1
      - name: rhbk-operator.v22.0.8-opr.2
        version: 22.0.8-opr.2
      - name: rhbk-operator.v22.0.8-opr.1
        version: 22.0.8-opr.1
      - name: rhbk-operator.v22.0.7-opr.1
        version: 22.0.7-opr.1
      - name: rhbk-operator.v22.0.6-opr.1
        version: 22.0.6-opr.1
      name: stable-v22
    - currentCSV: rhbk-operator.v22.0.13-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
          createdAt: "2024-09-17T11:20:22Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:884afe0ca1698d85a02aa47b6d60b8222e50d70ad280db6f98e1a4bbdbe13c40
        version: 22.0.13-opr.1
      entries:
      - name: rhbk-operator.v22.0.13-opr.1
        version: 22.0.13-opr.1
      - name: rhbk-operator.v22.0.12-opr.1
        version: 22.0.12-opr.1
      - name: rhbk-operator.v22.0.11-opr.2
        version: 22.0.11-opr.2
      - name: rhbk-operator.v22.0.11-opr.1
        version: 22.0.11-opr.1
      - name: rhbk-operator.v22.0.10-opr.1
        version: 22.0.10-opr.1
      - name: rhbk-operator.v22.0.9-opr.1
        version: 22.0.9-opr.1
      - name: rhbk-operator.v22.0.8-opr.2
        version: 22.0.8-opr.2
      - name: rhbk-operator.v22.0.8-opr.1
        version: 22.0.8-opr.1
      - name: rhbk-operator.v22.0.7-opr.1
        version: 22.0.7-opr.1
      - name: rhbk-operator.v22.0.6-opr.1
        version: 22.0.6-opr.1
      name: stable-v22.0
    - currentCSV: rhbk-operator.v24.0.11-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
          createdAt: "2025-03-26T02:27:51Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:d339ecf788d1681ba9945a2ea63ca675caec863ec75f497ef916103f4c898003
        version: 24.0.11-opr.1
      entries:
      - name: rhbk-operator.v24.0.11-opr.1
        version: 24.0.11-opr.1
      - name: rhbk-operator.v24.0.10-opr.1
        version: 24.0.10-opr.1
      - name: rhbk-operator.v24.0.9-opr.1
        version: 24.0.9-opr.1
      - name: rhbk-operator.v24.0.8-opr.1
        version: 24.0.8-opr.1
      - name: rhbk-operator.v24.0.7-opr.1
        version: 24.0.7-opr.1
      - name: rhbk-operator.v24.0.6-opr.1
        version: 24.0.6-opr.1
      - name: rhbk-operator.v24.0.5-opr.1
        version: 24.0.5-opr.1
      - name: rhbk-operator.v24.0.4-opr.1
        version: 24.0.4-opr.1
      - name: rhbk-operator.v24.0.3-opr.1
        version: 24.0.3-opr.1
      name: stable-v24
    - currentCSV: rhbk-operator.v24.0.11-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
          createdAt: "2025-03-26T02:27:51Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:d339ecf788d1681ba9945a2ea63ca675caec863ec75f497ef916103f4c898003
        version: 24.0.11-opr.1
      entries:
      - name: rhbk-operator.v24.0.11-opr.1
        version: 24.0.11-opr.1
      - name: rhbk-operator.v24.0.10-opr.1
        version: 24.0.10-opr.1
      - name: rhbk-operator.v24.0.9-opr.1
        version: 24.0.9-opr.1
      - name: rhbk-operator.v24.0.8-opr.1
        version: 24.0.8-opr.1
      - name: rhbk-operator.v24.0.7-opr.1
        version: 24.0.7-opr.1
      - name: rhbk-operator.v24.0.6-opr.1
        version: 24.0.6-opr.1
      - name: rhbk-operator.v24.0.5-opr.1
        version: 24.0.5-opr.1
      - name: rhbk-operator.v24.0.4-opr.1
        version: 24.0.4-opr.1
      - name: rhbk-operator.v24.0.3-opr.1
        version: 24.0.3-opr.1
      name: stable-v24.0
    - currentCSV: rhbk-operator.v26.2.9-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
          createdAt: "2025-09-18T15:03:55Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:50a62b46af44ce0ff7e6c3a15e1bfb2a9611ba4a8b65b8d0ab65817064692c55
        version: 26.2.9-opr.1
      entries:
      - name: rhbk-operator.v26.2.9-opr.1
        version: 26.2.9-opr.1
      - name: rhbk-operator.v26.2.8-opr.1
        version: 26.2.8-opr.1
      - name: rhbk-operator.v26.2.7-opr.1
        version: 26.2.7-opr.1
      - name: rhbk-operator.v26.2.6-opr.1
        version: 26.2.6-opr.1
      - name: rhbk-operator.v26.2.5-opr.1
        version: 26.2.5-opr.1
      - name: rhbk-operator.v26.2.4-opr.2
        version: 26.2.4-opr.2
      - name: rhbk-operator.v26.2.4-opr.1
        version: 26.2.4-opr.1
      - name: rhbk-operator.v26.0.15-opr.1
        version: 26.0.15-opr.1
      - name: rhbk-operator.v26.0.14-opr.1
        version: 26.0.14-opr.1
      - name: rhbk-operator.v26.0.13-opr.1
        version: 26.0.13-opr.1
      - name: rhbk-operator.v26.0.12-opr.1
        version: 26.0.12-opr.1
      - name: rhbk-operator.v26.0.11-opr.1
        version: 26.0.11-opr.1
      - name: rhbk-operator.v26.0.10-opr.1
        version: 26.0.10-opr.1
      - name: rhbk-operator.v26.0.9-opr.1
        version: 26.0.9-opr.1
      - name: rhbk-operator.v26.0.8-opr.1
        version: 26.0.8-opr.1
      - name: rhbk-operator.v26.0.7-opr.1
        version: 26.0.7-opr.1
      - name: rhbk-operator.v26.0.6-opr.1
        version: 26.0.6-opr.1
      - name: rhbk-operator.v26.0.5-opr.1
        version: 26.0.5-opr.1
      name: stable-v26
    - currentCSV: rhbk-operator.v26.0.15-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:fc5fc8cfb2c100c5b3a33fc11a78dfca569f5f34f45d88439340392d2c60afb2
          createdAt: "2025-08-28T19:34:30Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:fc5fc8cfb2c100c5b3a33fc11a78dfca569f5f34f45d88439340392d2c60afb2
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:4bb4d3070fe118bdd7de24e8b68133b8b6e158cea48c9f1abad2b4297be0dfdb
        version: 26.0.15-opr.1
      entries:
      - name: rhbk-operator.v26.0.15-opr.1
        version: 26.0.15-opr.1
      - name: rhbk-operator.v26.0.14-opr.1
        version: 26.0.14-opr.1
      - name: rhbk-operator.v26.0.13-opr.1
        version: 26.0.13-opr.1
      - name: rhbk-operator.v26.0.12-opr.1
        version: 26.0.12-opr.1
      - name: rhbk-operator.v26.0.11-opr.1
        version: 26.0.11-opr.1
      - name: rhbk-operator.v26.0.10-opr.1
        version: 26.0.10-opr.1
      - name: rhbk-operator.v26.0.9-opr.1
        version: 26.0.9-opr.1
      - name: rhbk-operator.v26.0.8-opr.1
        version: 26.0.8-opr.1
      - name: rhbk-operator.v26.0.7-opr.1
        version: 26.0.7-opr.1
      - name: rhbk-operator.v26.0.6-opr.1
        version: 26.0.6-opr.1
      - name: rhbk-operator.v26.0.5-opr.1
        version: 26.0.5-opr.1
      name: stable-v26.0
    - currentCSV: rhbk-operator.v26.2.9-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
          createdAt: "2025-09-18T15:03:55Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:50a62b46af44ce0ff7e6c3a15e1bfb2a9611ba4a8b65b8d0ab65817064692c55
        version: 26.2.9-opr.1
      entries:
      - name: rhbk-operator.v26.2.9-opr.1
        version: 26.2.9-opr.1
      - name: rhbk-operator.v26.2.8-opr.1
        version: 26.2.8-opr.1
      - name: rhbk-operator.v26.2.7-opr.1
        version: 26.2.7-opr.1
      - name: rhbk-operator.v26.2.6-opr.1
        version: 26.2.6-opr.1
      - name: rhbk-operator.v26.2.5-opr.1
        version: 26.2.5-opr.1
      - name: rhbk-operator.v26.2.4-opr.2
        version: 26.2.4-opr.2
      - name: rhbk-operator.v26.2.4-opr.1
        version: 26.2.4-opr.1
      name: stable-v26.2
    defaultChannel: stable-v26.2
    packageName: rhbk-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://horreum.hyperfoil.io
    name: horreum-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: horreum-operator.v0.7.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hyperfoil.io/v1alpha1",
                "kind": "Horreum",
                "metadata": {
                  "name": "horreum"
                },
                "spec": {
                  "nodeHost": "127.0.0.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/hyperfoil/horreum-operator:0.7.9
          createdAt: "2023-06-13T10:42:39Z"
          description: Performance results repository
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/hyperfoil/horreum-operator
          support: Red Hat Application Runtimes Performance Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Horreum is the object configuring Horreum performance results
              repository
            displayName: Horreum
            kind: Horreum
            name: horreums.hyperfoil.io
            version: v1alpha1
        description: Performance results repository
        displayName: Horreum
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - performance
        - repository
        - database
        - results
        - hyperfoil
        - benchmark
        links:
        - name: Horreum Operator
          url: https://horreum-operator.domain
        maintainers:
        - email: rvansa@redhat.com
          name: Radim Vansa
        - email: spederse@redhat.com
          name: Ståle W. Pedersen
        - email: wreicher@redhat.com
          name: Will Reichert
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat, Inc.
          url: https://horreum.hyperfoil.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/hyperfoil/horreum-operator:0.7.9
        version: 0.7.9
      entries:
      - name: horreum-operator.v0.7.9
        version: 0.7.9
      - name: horreum-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: horreum-operator
    provider:
      name: Red Hat, Inc.
      url: https://horreum.hyperfoil.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Automotive SIG CentOS
      provider-url: ""
    name: automotive-infra
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: automotive-infra.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "ContainerBuild",
                "metadata": {
                  "name": "cb-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "ImageBuild",
                "metadata": {
                  "name": "imagebuild-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "OperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "automotive-infra"
                  },
                  "name": "operatorconfig-sample",
                  "namespace": "automotive-infra-operator"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/fedora-sig-robotics/infra-ci-operator:v0.0.10
          createdAt: "2025-03-31T14:06:33Z"
          description: Streamline CI/CD for RHIVOS and its ecosystem using OpenShift
            Pipelines
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: automotive-infra-operator
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator
          support: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Client
            name: clients.jumpstarter.dev
            version: v1alpha1
          - kind: ContainerBuild
            name: containerbuilds.automotive.sig.centos.org
            version: v1
          - kind: ExporterAccessPolicy
            name: exporteraccesspolicies.jumpstarter.dev
            version: v1alpha1
          - kind: Exporter
            name: exporters.jumpstarter.dev
            version: v1alpha1
          - kind: ImageBuild
            name: imagebuilds.automotive.sig.centos.org
            version: v1
          - kind: Lease
            name: leases.jumpstarter.dev
            version: v1alpha1
          - kind: OperatorConfig
            name: operatorconfigs.automotive.sig.centos.org
            version: v1
        description: This operator is designed to streamline CI/CD for RHIVOS and
          its ecosystem using OpenShift Pipelines
        displayName: Automotive-CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Automotive
        - RHIVOS
        - Pipelines
        - CI
        links:
        - name: Automotive infra Operator
          url: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator
        maintainers:
        - email: automotive-sig@lists.centos.org
          name: Automotive SIG CentOS Community
        maturity: alpha
        provider:
          name: Automotive SIG CentOS
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/fedora-sig-robotics/infra-ci-operator:v0.0.10
        version: 0.0.10
      entries:
      - name: automotive-infra.v0.0.10
        version: 0.0.10
      - name: automotive-infra.v0.0.9
        version: 0.0.9
      - name: automotive-infra.v0.0.8
        version: 0.0.8
      - name: automotive-infra.v0.0.7
        version: 0.0.7
      - name: automotive-infra.v0.0.6
        version: 0.0.6
      - name: automotive-infra.v0.0.5
        version: 0.0.5
      - name: automotive-infra.v0.0.4
        version: 0.0.4
      - name: automotive-infra.v0.0.3
        version: 0.0.3
      - name: automotive-infra.v0.0.2
        version: 0.0.2
      - name: automotive-infra.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: automotive-infra
    provider:
      name: Automotive SIG CentOS
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: ibm-licensing-operator
      app.kubernetes.io/managed-by: ibm-licensing-operator
      app.kubernetes.io/name: ibm-licensing
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      intent: projected
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-licensing-operator-app
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-licensing-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "IBMLicensing",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-licensing-operator",
                    "app.kubernetes.io/managed-by": "ibm-licensing-operator",
                    "app.kubernetes.io/name": "ibm-licensing"
                  },
                  "name": "instance"
                },
                "spec": {
                  "version": "4.2.14",
                  "apiSecretToken": "ibm-licensing-token",
                  "datasource": "datacollector",
                  "httpsEnable": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
          createdAt: "2025-02-25T12:16:11Z"
          description: The IBM Licensing Operator provides a Kubernetes CRD-Based
            API to monitor the license usage of products.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-licensing-operator-app
          olm.skipRange: '>=1.0.0 <4.2.14'
          operatorframework.io/suggested-namespace: ibm-licensing
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/internal-objects: '["ibmlicensingmetadatas.operator.ibm.com",
            "ibmlicensingdefinitions.operator.ibm.com", "ibmlicensingquerysources.operator.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'IBMLicensing custom resource is used to create an instance
              of the License Service, used to collect information about license usage
              of IBM containerized products and IBM Cloud Paks per cluster. You can
              retrieve license usage data through a dedicated API call and generate
              an audit snapshot on demand. Documentation: For additional details regarding
              install parameters check: https://ibm.biz/icpfs39install. License: Please
              refer to the IBM Terms website (ibm.biz/lsvc-lic) to find the license
              terms for the particular IBM product for which you are deploying this
              component.'
            displayName: IBM License Service
            kind: IBMLicensing
            name: ibmlicensings.operator.ibm.com
            version: v1alpha1
          - kind: IBMLicensingDefinition
            name: ibmlicensingdefinitions.operator.ibm.com
            version: v1
          - description: 'IBMLicensingMetadata is the schema for IBM License Service.
              Thanks to IBMLicensingMetadata, you can track the license usage of Virtual
              Processor Core (VPC) metric by Pods that are managed by automation,
              for which licensing annotations are not defined at the time of deployment,
              such as the open source OpenLiberty. / For more information, see documentation:
              https://ibm.biz/icpfs39install. By installing the operator for IBM License
              Service, you accept the license terms: https://ibm.biz/icpfs39license.'
            displayName: IBMLicensing Metadata
            kind: IBMLicensingMetadata
            name: ibmlicensingmetadatas.operator.ibm.com
            version: v1alpha1
          - kind: IBMLicensingQuerySource
            name: ibmlicensingquerysources.operator.ibm.com
            version: v1
        description: |-
          **Important:**
          - If you are using the IBM Licensing Operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For the link to your IBM Cloud Pak documentation, see [IBM Cloud Paks that use Common Services](https://ibm.biz/BdyGwb).
          - If you are using a stand-alone IBM Container Software, you can use the IBM Licensing Operator directly. For more information, see [ibm-licensing-operator for stand-alone IBM Containerized Software](https://ibm.biz/BdyGwh).

          **IBM Licensing Operator overview**

          IBM Licensing Operator installs License Service. You can use License Service to collect information about license usage of IBM Containerized products and IBM Cloud Paks per cluster. You can retrieve license usage data through a dedicated API call and generate an audit snapshot on demand.

          **Supported platforms**

           Red Hat OpenShift Container Platform 4.10 or newer installed on Linux x86_64, Linux on Power (ppc64le), Linux on IBM Z and LinuxONE.

          **Prerequisites**

           Prerequisites depend on the integration of the License Service with an IBM Cloud Pak or IBM Containerized Software. For more information, see the applicable IBM Cloud Pak documentation or [ibm-licensing-operator for stand-alone IBM Containerized Software](https://ibm.biz/BdyGwh).
        displayName: IBM Licensing
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBMLicensing
        - IBM
        - Cloud
        - License
        - Service
        - License Service
        links:
        - name: GitHub
          url: https://github.com/IBM/ibm-licensing-operator
        maintainers:
        - email: talk2sam@us.ibm.com
          name: talk2sam
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9186c350fa54e8b792d72e47be6b54a67b610f726175c9641795010d45ad51f9
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:ec137115210dae9b69cb2899ed8a5437d94c8a658ae1ae5eb295232b34b0925e
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:96e6db93b2d16ba590252d6568b68f254548aaed73ec17e4bdd8d530ea668be6
        version: 4.2.14
      entries:
      - name: ibm-licensing-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-licensing-operator-app
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator-rhmp.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/dynatrace-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/dynatrace-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator-rhmp.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator-rhmp.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator-rhmp.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator-rhmp.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator-rhmp.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator-rhmp.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator-rhmp.v1.5.0
        version: 1.5.0
      - name: dynatrace-operator-rhmp.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator-rhmp.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator-rhmp.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator-rhmp
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-opensearchserverless-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-opensearchserverless-controller.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/opensearchserverless-controller:0.2.1
          createdAt: "2025-09-20T21:43:17Z"
          description: AWS OpenSearchServerless controller is a service controller
            for managing OpenSearchServerless resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Collection represents the state of an AWS opensearchserverless
              Collection resource.
            displayName: Collection
            kind: Collection
            name: collections.opensearchserverless.services.k8s.aws
            version: v1alpha1
          - description: SecurityPolicy represents the state of an AWS opensearchserverless
              SecurityPolicy resource.
            displayName: SecurityPolicy
            kind: SecurityPolicy
            name: securitypolicies.opensearchserverless.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon OpenSearchServerless resources in AWS from within your Kubernetes cluster.

          **About Amazon OpenSearchServerless**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon OpenSearchServerless
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opensearchserverless
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon OpenSearchServerless Developer Resources
          url: https://aws.amazon.com/OpenSearchServerless/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: opensearchserverless maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/opensearchserverless-controller:0.2.1
        version: 0.2.1
      entries:
      - name: ack-opensearchserverless-controller.v0.2.1
        version: 0.2.1
      - name: ack-opensearchserverless-controller.v0.2.0
        version: 0.2.0
      - name: ack-opensearchserverless-controller.v0.1.2
        version: 0.1.2
      - name: ack-opensearchserverless-controller.v0.1.1
        version: 0.1.1
      - name: ack-opensearchserverless-controller.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ack-opensearchserverless-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-framework
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-framework.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Framework",
                "metadata": {
                  "name": "framework-sample"
                },
                "spec": {
                  "kcos-eula": {
                    "image": {
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    },
                    "service": {
                      "nodePort": 0,
                      "type": "ClusterIP"
                    }
                  },
                  "kcos-logging-framework": {
                    "image": {
                      "remoteRepository": "docker.io",
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    },
                    "ingress": {
                      "annotations": {
                        "nginx.ingress.kubernetes.io/auth-method": "POST",
                        "nginx.ingress.kubernetes.io/auth-response-headers": "X-AUTH-USER",
                        "nginx.ingress.kubernetes.io/auth-url": "https://127.0.0.1/token/validate",
                        "nginx.ingress.kubernetes.io/cors-allow-credentials": "true",
                        "nginx.ingress.kubernetes.io/cors-allow-methods": "PUT, GET, POST, OPTIONS, PATCH",
                        "nginx.ingress.kubernetes.io/cors-allow-origin": "https://admin.example.com,https://localhost:4200",
                        "nginx.ingress.kubernetes.io/enable-cors": "true",
                        "nginx.ingress.kubernetes.io/proxy-body-size": "0",
                        "nginx.ingress.kubernetes.io/ssl-redirect": "true"
                      },
                      "enabled": true
                    }
                  },
                  "kcos-postgresql": {
                    "jobCopySecret": {
                      "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                    },
                    "postgresql": {
                      "image": {
                        "registry": "prod.harbor.keysight.digital/loadcore"
                      },
                      "initdbScripts": {
                        "init_script.sh": "#!/bin/sh\nexport PGUSER=postgres\npsql -c \"SELECT 1 FROM pg_database WHERE datname = 'db'\" | grep -q 1 || psql -c \"CREATE DATABASE db\"\npsql -c \"SELECT 1 FROM pg_database WHERE datname = 'kcoslicensingdb'\" | grep -q 1 || psql -c \"CREATE DATABASE kcoslicensingdb\"\n"
                      },
                      "metrics": {
                        "image": {
                          "registry": "docker.io"
                        }
                      },
                      "persistence": {
                        "storageClass": "managed-loadcore-storage"
                      },
                      "pgHbaConfiguration": "local all all trust\nhost all all 0.0.0.0/0 trust",
                      "securityContext": {
                        "enabled": false
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": false
                        }
                      },
                      "volumePermissions": {
                        "image": {
                          "registry": "docker.io"
                        },
                        "securityContext": {
                          "runAsUser": "auto"
                        }
                      }
                    }
                  },
                  "kcos-system-diagnostics": {
                    "environment": {
                      "ELASTICSEARCH_ENDPOINT": "elasticsearch.keysight-wap.svc.cluster.local.",
                      "IGNORE_DATA_PREFIX": "wireless_stats"
                    },
                    "image": {
                      "initContainer": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf",
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    }
                  },
                  "subcharts": {
                    "enable": {
                      "kcos-coredns-fixup": false,
                      "kcos-eula": true,
                      "kcos-local-storage": false,
                      "kcos-logging-framework": true,
                      "kcos-logging-rotate": false,
                      "kcos-naas": false,
                      "kcos-postgresql": true,
                      "kcos-shell": false,
                      "kcos-system-diagnostics": true,
                      "kcos-vital-service": false,
                      "nfs-server-provisioner": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:27:28Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Framework
            name: frameworks.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Framework Operator
        displayName: Keysight KCOS Framework Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Framework
          url: https://keysight-kcos-framework.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-framework@sha256:531f278def85e33d453479fce74127e12e3ed62a251872eab421c6561bedd0c5
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        version: 0.1.21
      entries:
      - name: keysight-kcos-framework.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-framework
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx-openshift-operator.v0.0.36
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
          createdAt: "2019-07-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX on JBoss EAP 7
            displayName: Joget DX on JBoss EAP 7
            kind: JogetDX
            name: jogetdx.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX7/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
        - registry.connect.redhat.com/joget/joget-dx7-eap7@sha256:7245501f5fccf011d75c034b7d55542cfe3bd2c64cc7e48b3224a48b623af282
        version: 0.0.36
      entries:
      - name: joget-dx-openshift-operator.v0.0.36
        version: 0.0.36
      - name: joget-dx-openshift-operator.v0.0.35
        version: 0.0.35
      - name: joget-dx-openshift-operator.v0.0.34
        version: 0.0.34
      - name: joget-dx-openshift-operator.v0.0.33
        version: 0.0.33
      - name: joget-dx-openshift-operator.v0.0.32
        version: 0.0.32
      - name: joget-dx-openshift-operator.v0.0.31
        version: 0.0.31
      - name: joget-dx-openshift-operator.v0.0.30
        version: 0.0.30
      - name: joget-dx-openshift-operator.v0.0.29
        version: 0.0.29
      - name: joget-dx-openshift-operator.v0.0.28
        version: 0.0.28
      - name: joget-dx-openshift-operator.v0.0.27
        version: 0.0.27
      - name: joget-dx-openshift-operator.v0.0.26
        version: 0.0.26
      - name: joget-dx-openshift-operator.v0.0.25
        version: 0.0.25
      - name: joget-dx-openshift-operator.v0.0.24
        version: 0.0.24
      - name: joget-dx-openshift-operator.v0.0.23
        version: 0.0.23
      - name: joget-dx-openshift-operator.v0.0.22
        version: 0.0.22
      - name: joget-dx-openshift-operator.v0.0.21
        version: 0.0.21
      - name: joget-dx-openshift-operator.v0.0.20
        version: 0.0.20
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Habana Labs Ltd.
      provider-url: https://habana.ai
    name: habana-ai-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: habana-ai-operator.v1.19.2-32-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.19.2-32"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.19.2"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.19.2-32"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.19.2-32"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.19.2-32"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.19.2-32"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.19.2-32"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.19.2-32"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.19.2-32"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.19.2-32"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.19.2-32"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:975ed2135061d3bdc145c83abaa15f99e4c9f6efb15c708de41aa00fbcb72333
          createdAt: "2025-02-10T07:33:30Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.19.2-32-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi AI accelerator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:ac16752b53df69573da257d3a3a717a214b2eb1656779a7ef94084d74363fb46
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:b12802e7ef05273fceaef3a9987c3c4ae760714e6f333363f665eb26be9175e9
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:f05cd78d7593b877ec8a75175a9faa72d9b991d81c74af9ecd92208fde296eca
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:170a15041fc3e3901b4e8e0b9a32d53fde11026776c17a6560d500dacb976c56
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:d728a91a36503f07dfacd49fdbfefe8083607ca33babb5f2dfc1e9de05386454
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:e667695b845cd2be887f117dac66861d2c5e74def009d2939559ae7c30574dc0
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:975ed2135061d3bdc145c83abaa15f99e4c9f6efb15c708de41aa00fbcb72333
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:eeb4573de20d3341a085c75e669b7d37c424e348421d7916440fbbfff12ea482
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:71d4c7181930c71c0ca4e6b696ea416a687eec6d798a43e854fcd1bfdc224a0c
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:72cc75353615369d12a07f5488bcdc041a8435906ca6e16c53b001016108e099
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:373b9caa27a96bb7979f7a6aec246795142ce3a28c6b97a8902035ec20814fa2
        version: 1.19.2-32-1
      entries:
      - name: habana-ai-operator.v1.19.2-32-1
        version: 1.19.2-32-1
      name: "1.19"
    - currentCSV: habana-ai-operator.v1.20.0-543-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.20.0-543"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.20.0-4"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.20.0-543-2"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.20.0-543"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.20.0-543"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.20.0-543-4"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.20.0-543"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.20.0-543-4"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:d5b37eafe48446e1ffa492e195f4519b77d4a11f6750e5c71f27c1536bdfd2cc
          createdAt: "2025-02-26T09:03:44Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.20.0-543-5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:9ab54a84a00e8afc078eec2a7d643f09b1bd674e0e230733fbce7108e392e009
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:d5b37eafe48446e1ffa492e195f4519b77d4a11f6750e5c71f27c1536bdfd2cc
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:317765eac4e0332a0c4f6bead5563b343a8c22e6e0eaa62276ff66be08608c30
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:c5beac6209d0671d3a49b63f1cf95401055b042430eafaefc3859b5c633f0e84
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:68d0f44170cbf3ea15b793e0503fb07ce7e32488011fc85ccb6c9234b3b9302f
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:3ea80aa8994e8b19fbbdcd18ba2c35800774da3ca4e27a2514d0f5f9676786fa
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:957702b58ee66d7339a20b2a5ece783606d6ccebde627c58951d5ecd6bb1b13f
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:cba03259b33c4cb9bcd22622ab0dda962846ea0a80c727779365cf55129ad308
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:d14c457106b6133fc41ef727ac40a6bf2f0e5c97f21659969d1daebdc1310b43
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:c662de244c5900c89b918b16982298676d9a6d553360a5716174b28d4d64a635
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:a673900f08c38c9bb40c90b3cddea580afcc6a99195d816cd0b5cae9831ca5c9
        version: 1.20.0-543-5
      entries:
      - name: habana-ai-operator.v1.20.0-543-5
        version: 1.20.0-543-5
      name: "1.20"
    - currentCSV: habana-ai-operator.v1.21.5-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.21.5-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.21.5-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.21.5-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:0741da99140d137474b9e7edba921672eab8ff6277e1dd8bf3a4179e87f01dd4
          createdAt: "2025-09-17T11:55:32Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.21.5-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:0741da99140d137474b9e7edba921672eab8ff6277e1dd8bf3a4179e87f01dd4
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:a216ec6a6e7df50e067c557e9ed5871a3308b2ba42a4e41c9642f0c2c4503d88
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:9692d079ef18a364ed235c187482f329b3a5764a773e363acb1238044ad4a884
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:c95d3ffd6c91be3418415740258d33a8dbcedad6fe7d6bfe68d4dd122027447b
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:bc3749ff153fff71e252820cbf22fcd0753b195b2204f9731e42af29fb7ddaa0
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:bd57e7059467f0ca1667d19a70bd9a6ad96b83bbcb53c09fdad5a262fee76af9
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:c84e997c7d8c22c4d4e902a29e45e5364f0ad306f17d03a7d8ddb4391142e167
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:256cd8bebec3ed835e187e925a14a6b605ce9b3a17fe7d8886e0b32cde163d90
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:1e488c8ad31784dab0fe70144f5146d5809e2fc7c86424d1b512ab887787e28f
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:c7147359dcd6a2e5adee95259d37743ff302b1deaeb13b184044f4029d33f44a
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:d7f64d15c379524597aa2b2c5b4586c890442fa32356692013cf665b01b0f092
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:a0b45ccfb6af54b1043ccc2336ad3d7990cc447af5ba3621f3eb1953cd40ffc5
        version: 1.21.5-6
      entries:
      - name: habana-ai-operator.v1.21.5-6
        version: 1.21.5-6
      name: "1.21"
    - currentCSV: habana-ai-operator.v1.22.1-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.22.1-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
          createdAt: "2025-09-17T10:15:21Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.22.1-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:1e15f3de21c2b1e20f54d1f8a973290c6d940624506880c5bc3eb880011bcc02
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:d1974d72774c626e77707e8f44bcc631ac142bfe2a5395714969a964c963e36e
        - vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer@sha256:2cf151c2a518e3e1c7c09efb8301e469d266d1f030fff2e6180b086cc2e04ee7
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:704b4bebdfe6b85693df4b213f67862859a554c21ce4e614736172f64b541ea7
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:c9ff6f3fe24ef87c7bfc71bdb1cd7900941df52f9ad1293badd7e14783ce7ef6
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:e34a6820a58e7e8090b521781c76acc3867c94f8291634de7b138f928e0bdb74
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:53c1844b5e8eb08d0cc29865a732100208c5dfc0fe55df8b32e589e319bcab6c
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:38c21eae6249e5571921998660418e59870bc0880e7b8c9b95a4a8db30ef2e7d
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:1925586fda34c6fa444a59b1fa7be20b3cf16e845bf56566ad2a41c276b0cfe9
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:392fcd2a4730299e6e3bab9f7e96fc3b710de8adb374a2c64b7316e097ed05f9
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:7140877b23291a5db3e5e28def3c9d5e7cb7be62f046e2fb7ac45eff369cea84
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:8f413edfff1978b35ba3998ed5f1c6293e6764b1d2cd9faa08f0bd81dc6e3e10
        version: 1.22.1-6
      entries:
      - name: habana-ai-operator.v1.22.1-6
        version: 1.22.1-6
      name: "1.22"
    - currentCSV: habana-ai-operator.v1.22.1-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.22.1-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
          createdAt: "2025-09-17T10:15:21Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.22.1-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:c9ff6f3fe24ef87c7bfc71bdb1cd7900941df52f9ad1293badd7e14783ce7ef6
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:392fcd2a4730299e6e3bab9f7e96fc3b710de8adb374a2c64b7316e097ed05f9
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:38c21eae6249e5571921998660418e59870bc0880e7b8c9b95a4a8db30ef2e7d
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:1925586fda34c6fa444a59b1fa7be20b3cf16e845bf56566ad2a41c276b0cfe9
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:704b4bebdfe6b85693df4b213f67862859a554c21ce4e614736172f64b541ea7
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:7140877b23291a5db3e5e28def3c9d5e7cb7be62f046e2fb7ac45eff369cea84
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:8f413edfff1978b35ba3998ed5f1c6293e6764b1d2cd9faa08f0bd81dc6e3e10
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:1e15f3de21c2b1e20f54d1f8a973290c6d940624506880c5bc3eb880011bcc02
        - vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer@sha256:2cf151c2a518e3e1c7c09efb8301e469d266d1f030fff2e6180b086cc2e04ee7
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:e34a6820a58e7e8090b521781c76acc3867c94f8291634de7b138f928e0bdb74
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:53c1844b5e8eb08d0cc29865a732100208c5dfc0fe55df8b32e589e319bcab6c
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:d1974d72774c626e77707e8f44bcc631ac142bfe2a5395714969a964c963e36e
        version: 1.22.1-6
      entries:
      - name: habana-ai-operator.v1.22.1-6
        version: 1.22.1-6
      name: stable
    defaultChannel: stable
    packageName: habana-ai-operator
    provider:
      name: Habana Labs Ltd.
      url: https://habana.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-multicluster-orchestrator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-multicluster-orchestrator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "items": [
                    {
                      "clusterName": "local-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    },
                    {
                      "clusterName": "spoke-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    }
                  ],
                  "manageS3": false,
                  "schedulingIntervals": [
                    "5m"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["odf-multicluster-console"]'
          createdAt: "2025-09-12T07:33:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          Orchestrator for OpenShift Data Foundation clusters running across multiple OpenShift clusters.
          It uses Red Hat Advanced Cluster Management for Kubernetes as the multicluster control plane.
        displayName: ODF Multicluster Orchestrator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: ODF Multicluster Orchestrator
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: uchapaga@redhat.com
          name: Umanga Chapagain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-multicluster-rhel9-operator@sha256:09e2c7a20ab92ddc5d589072bb0ab9ce2ed8706ed8364f1b11192d7a7396e138
        - registry.redhat.io/odf4/odf-multicluster-console-rhel9@sha256:7527f4a14aed0386b8f160fa8a92b7ff1f58114970e07c3ec6c433d2a53459b5
        version: 4.15.20-rhodf
      entries:
      - name: odf-multicluster-orchestrator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odf-multicluster-orchestrator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odf-multicluster-orchestrator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odf-multicluster-orchestrator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odf-multicluster-orchestrator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odf-multicluster-orchestrator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odf-multicluster-orchestrator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odf-multicluster-orchestrator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odf-multicluster-orchestrator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odf-multicluster-orchestrator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odf-multicluster-orchestrator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odf-multicluster-orchestrator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odf-multicluster-orchestrator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odf-multicluster-orchestrator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odf-multicluster-orchestrator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odf-multicluster-orchestrator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odf-multicluster-orchestrator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odf-multicluster-orchestrator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odf-multicluster-orchestrator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odf-multicluster-orchestrator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odf-multicluster-orchestrator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "items": [
                    {
                      "clusterName": "local-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    },
                    {
                      "clusterName": "spoke-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    }
                  ],
                  "manageS3": false,
                  "schedulingIntervals": [
                    "5m"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["odf-multicluster-console"]'
          createdAt: "2025-09-12T06:32:35Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          Orchestrator for OpenShift Data Foundation clusters running across multiple OpenShift clusters.
          It uses Red Hat Advanced Cluster Management for Kubernetes as the multicluster control plane.
        displayName: ODF Multicluster Orchestrator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: ODF Multicluster Orchestrator
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: uchapaga@redhat.com
          name: Umanga Chapagain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-multicluster-rhel9-operator@sha256:78d6bc14b95d1ee91ed529422062ec191afc489d6ab4e75906ae3c7fe58749cb
        - registry.redhat.io/odf4/odf-multicluster-console-rhel9@sha256:97d3645560388f6ae08fc12cafb58b6c766e38d83de04b645b3e6fb03fbba231
        version: 4.16.16-rhodf
      entries:
      - name: odf-multicluster-orchestrator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odf-multicluster-orchestrator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odf-multicluster-orchestrator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odf-multicluster-orchestrator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odf-multicluster-orchestrator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odf-multicluster-orchestrator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odf-multicluster-orchestrator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odf-multicluster-orchestrator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odf-multicluster-orchestrator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odf-multicluster-orchestrator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odf-multicluster-orchestrator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odf-multicluster-orchestrator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odf-multicluster-orchestrator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odf-multicluster-orchestrator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odf-multicluster-orchestrator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odf-multicluster-orchestrator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odf-multicluster-orchestrator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odf-multicluster-orchestrator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Argo CD Community
      provider-url: ""
    name: argocd-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: argocd-operator.v0.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd-sample"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCDExport",
                "metadata": {
                  "name": "argocdexport-sample"
                },
                "spec": {
                  "argocd": "argocd-sample"
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd-sample"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/argoprojlabs/argocd-operator@sha256:223ad82fb697471ede79201c4a042115c9ca5c00a172d5bdcd5a780e58c46c79
          createdAt: "2025-09-11T06:29:33Z"
          description: Argo CD is a declarative, GitOps continuous delivery tool for
            Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/argoproj-labs/argocd-operator
          support: Argo CD
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: An ApplicationSet is a group or set of Application resources.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: ArgoCDExport is the Schema for the argocdexports API
            displayName: Argo CDExport
            kind: ArgoCDExport
            name: argocdexports.argoproj.io
            version: v1alpha1
          - description: ArgoCD is the Schema for the argocds API
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: ArgoCD is the Schema for the argocds API
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: NotificationsConfiguration is the Schema for the notificationsconfiguration
              API
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
        description: |
          ## Overview

          The Argo CD Operator manages the full lifecycle for [Argo CD](https://argoproj.github.io/argo-cd/) and it's
          components. The operator's goal is to automate the tasks required when operating an Argo CD cluster.

          Beyond installation, the operator helps to automate the process of upgrading, backing up and restoring as needed and
          remove the human as much as possible. In addition, the operator aims to provide deep insights into the Argo CD
          environment by configuring Prometheus and Grafana to aggregate, visualize and expose the metrics already exported by
          Argo CD.

          The operator aims to provide the following, and is a work in progress.

          * Easy configuration and installation of the Argo CD components with sane defaults to get up and running quickly.
          * Provide seamless upgrades to the Argo CD components.
          * Ability to back up and restore an Argo CD cluster from a point in time or on a recurring schedule.
          * Aggregate and expose the metrics for Argo CD and the operator itself using Prometheus and Grafana.
          * Autoscale the Argo CD components as necessary to handle variability in demand.

          ## Usage

          Deploy a basic Argo CD cluster by creating a new ArgoCD resource in the namespace where the operator is installed.

          ```
          apiVersion: argoproj.io/v1alpha1
          kind: ArgoCD
          metadata:
            name: example-argocd
          spec: {}
          ```

          ## Backup

          Backup the cluster above by creating a new ArgoCDExport resource in the namespace where the operator is installed.

          ```
          apiVersion: argoproj.io/v1alpha1
          kind: ArgoCDExport
          metadata:
            name: example-argocdexport
          spec:
            argocd: example-argocd
          ```

          See the [documentation](https://argocd-operator.readthedocs.io) and examples on
          [GitHub](https://github.com/argoproj-labs/argocd-operator) for more information.
        displayName: Argo CD
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - kubernetes
        links:
        - name: Argo CD Project
          url: https://argoproj.github.io/argo-cd/
        - name: Operator Documentation
          url: https://argocd-operator.readthedocs.io
        - name: Operator Source Code
          url: https://github.com/argoproj-labs/argocd-operator
        maintainers:
        - email: aveerama@redhat.com
          name: Abhishek Veeramalla
        maturity: alpha
        provider:
          name: Argo CD Community
        relatedImages:
        - quay.io/argoprojlabs/argocd-operator@sha256:223ad82fb697471ede79201c4a042115c9ca5c00a172d5bdcd5a780e58c46c79
        version: 0.15.0
      entries:
      - name: argocd-operator.v0.15.0
        version: 0.15.0
      - name: argocd-operator.v0.14.1
        version: 0.14.1
      - name: argocd-operator.v0.14.0
        version: 0.14.0
      - name: argocd-operator.v0.13.0
        version: 0.13.0
      - name: argocd-operator.v0.12.0
        version: 0.12.0
      - name: argocd-operator.v0.11.0
        version: 0.11.0
      - name: argocd-operator.v0.10.1
        version: 0.10.1
      - name: argocd-operator.v0.10.0
        version: 0.10.0
      - name: argocd-operator.v0.9.1
        version: 0.9.1
      - name: argocd-operator.v0.9.0
        version: 0.9.0
      - name: argocd-operator.v0.8.0
        version: 0.8.0
      - name: argocd-operator.v0.7.0
        version: 0.7.0
      - name: argocd-operator.v0.5.0
        version: 0.5.0
      - name: argocd-operator.v0.4.0
        version: 0.4.0
      - name: argocd-operator.v0.3.0
        version: 0.3.0
      - name: argocd-operator.v0.2.1
        version: 0.2.1
      - name: argocd-operator.v0.2.0
        version: 0.2.0
      - name: argocd-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: argocd-operator
    provider:
      name: Argo CD Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: oadp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oadp-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "Velero",
                "metadata": {
                  "name": "example-velero",
                  "namespace":"openshift-adp"
                },
                "spec": {
                  "backupStorageLocations": [
                    {
                      "config": {
                        "profile": "default",
                        "region": "us-east-1"
                      },
                      "name": "default",
                      "objectStorage": {
                        "bucket": "myBucket",
                        "prefix": "velero"
                      },
                      "provider": "aws"
                    }
                  ],
                  "defaultVeleroPlugins": [
                    "aws",
                    "csi",
                    "openshift"
                  ],
                  "veleroFeatureFlags": [
                    "EnableCSI"
                  ],
                  "enableRestic": true,
                  "volumeSnapshotLocations": [
                    {
                      "config": {
                        "profile": "default",
                        "region": "us-west-1"
                      },
                      "name": "default",
                      "provider": "aws"
                    }
                  ]
                }
               },
                {
                  "apiVersion": "velero.io/v1",
                  "kind": "Backup",
                  "metadata": {
                    "name": "backup",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "BackupStorageLocation",
                  "metadata": {
                    "name": "backupstoragelocation",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "DeleteBackupRequest",
                  "metadata": {
                    "name": "deletebackuprequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "DownloadRequest",
                  "metadata": {
                    "name": "downloadrequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "PodVolumeBackup",
                  "metadata": {
                    "name": "podvolumebackup",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "PodVolumeRestore",
                  "metadata": {
                    "name": "podvolumerestore",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "ResticRepository",
                  "metadata": {
                    "name": "resticrepository",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "Restore",
                  "metadata": {
                    "name": "restore",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "Schedule",
                  "metadata": {
                    "name": "schedule",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "ServerStatusRequest",
                  "metadata": {
                    "name": "serverstatusrequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "VolumeSnapshotLocation",
                  "metadata": {
                    "name": "volumesnapshotlocation",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/konveyor/oadp-operator@sha256:ecb8534fbc245e890b9c7766946d01e1a8b4ce9be359c4ada4c95fd908b679b3
          createdAt: "2020-09-08T12:21:00Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          olm.skipRange: '>=0.0.0 <0.4.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Velero is a tool to safely backup and restore, perform disaster
              recovery and migrate Kubernetes cluster resources and persistent volumes.
            displayName: Velero
            kind: Velero
            name: veleroes.oadp.openshift.io
            version: v1alpha1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - description: Backup is a Velero resource that respresents the capture
              of Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Velero on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://oadp.konveyor.io/
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/konveyor/velero-restic-restore-helper@sha256:c4d12829e42cabe70b662a659875900677dfb415f912f931c3bf4c09a26aed5b
        - quay.io/konveyor/openshift-velero-plugin@sha256:aec8b4fc66cfcebd5a35ce99d8c985453ebf25c11453d30256089a8b425990b5
        - quay.io/konveyor/velero-plugin-for-aws@sha256:2197c6fef32910747526d33c1462b75cb7940aeace65791d0a5ddaa038480e52
        - quay.io/konveyor/velero-plugin-for-csi@sha256:5c2261effc902d8baa16e38265f9ad1fbe5284a3e6abd2bbc97761c161fa92cf
        - quay.io/konveyor/registry@sha256:b74ab3278b79eb89f99b6970c6aa0f5306221a8495b5496faad6cab2f74ff2cd
        - quay.io/konveyor/oadp-operator@sha256:ecb8534fbc245e890b9c7766946d01e1a8b4ce9be359c4ada4c95fd908b679b3
        - quay.io/konveyor/velero-plugin-for-microsoft-azure@sha256:62a344e5cf196e55dae92e537648987d758ccc75985bc45e9be6d0646a9e64d0
        - quay.io/konveyor/velero-plugin-for-gcp@sha256:3300458b1d6c1514dca7d24353ad03bd92db396b2a29f7d3522563160a5f5219
        - quay.io/konveyor/velero@sha256:e7865cc707664eb6ee767e77ccd933420604c2562736b53e16c8043a51df9d86
        version: 0.4.2
      entries:
      - name: oadp-operator.v0.4.2
        version: 0.4.2
      - name: oadp-operator.v0.4.1
        version: 0.4.1
      - name: oadp-operator.v0.4.0
        version: 0.4.0
      - name: oadp-operator.v0.3.0
        version: 0.3.0
      name: beta
    - currentCSV: oadp-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "restic": {
                      "enable": true
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ResticRepository",
                "metadata": {
                  "name": "resticrepository",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/konveyor/oadp-operator@sha256:90609044d52173619ea1e63ddef0ef3e7c1c939ac737b64bd427cbed5374dd27
          createdAt: "2020-09-08T12:21:00Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          olm.skipRange: '>=0.0.0 <0.5.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is a Velero resource that respresents the capture
              of Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Data Protection Applications on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://oadp.konveyor.io/
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/konveyor/registry@sha256:1be970879206258335357e3cf630138c24b8a5ccab35aae05eccb260e5e8ffb4
        - quay.io/konveyor/openshift-velero-plugin@sha256:221ab0bcd1d95fc166ef7319d187e196b59d2fbf9eefd29d91bd6435d45b5c3c
        - quay.io/konveyor/oadp-operator@sha256:90609044d52173619ea1e63ddef0ef3e7c1c939ac737b64bd427cbed5374dd27
        - quay.io/konveyor/velero@sha256:0ea5336dcbf0a1fab561697b24da01a0e5b533ee23553fbfd5f8e002f5a31c3e
        - quay.io/konveyor/velero-restic-restore-helper@sha256:c16248d027c09cd5b7360ae38484df88ca1cc01f9e470e3eade95076ee1a6b84
        - quay.io/konveyor/velero-plugin-for-aws@sha256:1e8bc27d6b60000e1406085d582910cb530af3fb901ef9bc780628df6c766288
        - quay.io/konveyor/velero-plugin-for-microsoft-azure@sha256:2d0089cbe715b482fb42e1acd9add70b975a8575636d8b143e6f90cef317ea13
        - quay.io/konveyor/velero-plugin-for-gcp@sha256:ee463ed2a8019c560961b118f8a87b587ed302f24fd25480c5d4ddad2557e92f
        - quay.io/konveyor/velero-plugin-for-csi@sha256:8c36ccfd294dafc415a7f113770cbbd66dbf6ecf8fef3bc4e73adc9878c606b0
        version: 0.5.6
      entries:
      - name: oadp-operator.v0.5.6
        version: 0.5.6
      - name: oadp-operator.v0.5.5
        version: 0.5.5
      - name: oadp-operator.v0.5.4
        version: 0.5.4
      - name: oadp-operator.v0.5.3
        version: 0.5.3
      - name: oadp-operator.v0.5.2
        version: 0.5.2
      - name: oadp-operator.v0.5.1
        version: 0.5.1
      - name: oadp-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: oadp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-access-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-access-operator.v23.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyAccess",
                "metadata": {
                  "name": "isva-sample"
                },
                "spec": {
                  "image": "icr.io/isva/verify-access-wrp:10.0.5.0",
                  "instance": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isva/verify-access-operator:23.3.0
          createdAt: Mon Mar  20 00:59:34 UTC 2023
          description: The IBM Security Verify Access Operator manages the lifecycle
            of IBM Security Verify Access worker containers.
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM-Security/verify-access-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyAccess is the Schema for the ibmsecurityverifyaccesses
              API.
            displayName: IBMSecurity Verify Access
            kind: IBMSecurityVerifyAccess
            name: ibmsecurityverifyaccesses.ibm.com
            version: v1
        description: |+
          In a world of highly fragmented access management environments, [IBM Security Verify Access](https://www.ibm.com/au-en/products/verify-access) helps you simplify your users' access while more securely adopting web, mobile and cloud technologies. This solution helps you strike a balance between usability and security through the use of risk-based access, single sign-on, integrated access management control, identity federation and its mobile multi-factor authentication capability, IBM Verify. Take back control of your access management with IBM Security Verify Access.

          The IBM Security Verify Access operator provides lifecycle management of the lightweight containers which are used to protect an environment, namely:
          * [Web Reverse Proxy](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-web-reverse-proxy)
          * [Runtime](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-runtime)
          * [Distributed Session Cache](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-distributed-session-cache)

          The operator will manage the deployment of these lightweight IBM Security Verify Access worker containers, and also control the rolling restart of these containers when a configuration snapshot is updated.

          See the project [Readme](https://github.com/IBM-Security/verify-access-operator/blob/master/README.md) for further information and details.

        displayName: IBM Security Verify Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Access Product Information
          url: https://www.ibm.com/au-en/products/verify-access
        - name: Verify Access Documentation
          url: https://www.ibm.com/docs/en/sva
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Access Development Team
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - icr.io/isva/verify-access-operator:23.3.0
        version: 23.3.0
      entries:
      - name: ibm-security-verify-access-operator.v23.3.0
        version: 23.3.0
      - name: ibm-security-verify-access-operator.v22.10.0
        version: 22.10.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-access-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-directory-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-directory-operator.v23.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyDirectory",
                "metadata": {
                  "name": "ibmsecurityverifydirectory-sample"
                },
                "spec": {
                  "pods": {
                    "configMap": {
                      "proxy": {
                        "key": "config.yaml",
                        "name": "isvd-proxy-config"
                      },
                      "server": {
                        "key": "config.yaml",
                        "name": "isvd-server-config"
                      }
                    },
                    "image": {
                      "label": "latest",
                      "repo": "icr.io/isvd"
                    },
                    "proxy": {
                      "replicas": 1
                    },
                    "serviceAccountName": "isvd"
                  },
                  "replicas": {
                    "pvcs": [
                      "replica-1",
                      "replica-2"
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: icr.io/isvd/verify-directory-operator@sha256:d1dd7037584c6818dcf8d5aa69472278833bb8e37fba2aec5a69185ba88dcc34
          createdAt: "2023-04-18T22:10:32Z"
          description: The IBM Security Verify Directory operator is used to scale
            an IBM Security Verify Directory Server environment.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ibm-security/verify-directory-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyDirectory is the Schema for the ibmsecurityverifydirectories
              API
            displayName: IBMSecurity Verify Directory
            kind: IBMSecurityVerifyDirectory
            name: ibmsecurityverifydirectories.ibm.com
            version: v1
        description: "IBM Security Verify Directory is a scalable, standards-based
          identity directory that helps simplify identity and directory management.
          Verify Directory helps consolidate identity silos into a single identity
          source. Verify Directory is purpose-built to provide a directory foundation
          that can help provide a trusted identity data infrastructure that assists
          in enabling mission-critical security and authentication. It is designed
          to deliver a reliable, scalable, standards-based identity data platform
          that interoperates with a broad range of operating systems and applications.
          Verify Directory supports Lightweight Directory Access Protocol (LDAP) V3,
          offering a flexible and highly scalable LDAP infrastructure. \n\nThe operator
          can be used to manage deployed instances of the IBM Security Verify Directory
          server, providing an easy mechanism to scale the number of servers which
          are used to handle the workload.\n\nSee the project [Readme](https://github.com/ibm-security/verify-directory-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Security Verify Directory
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Security
        - Identity and Access
        links:
        - name: IBM Security Verify Directory Documentation
          url: https://www.ibm.com/docs/en/svd
        maintainers:
        - email: iamdev@us.ibm.com
          name: IBM Security Verify Directory Development Team
        maturity: stable
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - icr.io/isvd/verify-directory-operator@sha256:d1dd7037584c6818dcf8d5aa69472278833bb8e37fba2aec5a69185ba88dcc34
        version: 23.4.1
      entries:
      - name: ibm-security-verify-directory-operator.v23.4.1
        version: 23.4.1
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-directory-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Flux project
      provider-url: ""
    name: flux
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.toolkit.fluxcd.io/v2beta1",
                "kind": "HelmRelease",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "5m",
                    "chart": {
                      "spec": {
                          "chart": "podinfo",
                          "version": "4.0.x",
                          "sourceRef": {
                            "kind": "HelmRepository",
                            "name": "podinfo",
                            "namespace": "flux-system"
                          },
                          "interval": "1m"
                      }
                    },
                    "values": {
                      "replicaCount": 2
                    }
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "GitRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "1m",
                    "ref": {
                      "branch": "master"
                    },
                    "url": "https://github.com/stefanprodan/podinfo"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmRepository",
                "metadata": {
                    "name": "stable",
                    "namespace": "default"
                },
                "spec": {
                    "url": "https://kubernetes-charts.storage.googleapis.com/",
                    "interval": "10m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmChart",
                "metadata": {
                    "name": "redis",
                    "namespace": "default"
                },
                "spec": {
                    "chart": "redis",
                    "version": "10.5.7",
                    "sourceRef": {
                      "name": "stable",
                      "kind": "HelmRepository"
                    },
                    "interval": "5m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "Bucket",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "1m",
                    "provider": "generic",
                    "bucketName": "podinfo",
                    "endpoint": "minio.minio.svc.cluster.local:9000",
                    "insecure": true,
                    "secretRef": {
                      "name": "minio-credentials"
                    }
                }
              },
              {
                "apiVersion": "kustomize.toolkit.fluxcd.io/v1beta2",
                "kind": "Kustomization",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "5m",
                    "path": "./deploy/overlays/dev",
                    "prune": true,
                    "sourceRef": {
                      "kind": "GitRepository",
                      "name": "podinfo"
                    },
                    "timeout": "2m",
                    "wait": true
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "address": "https://github.com/stefanprodan/podinfo",
                    "secretRef": {
                      "name": "api-token"
                    }
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Alert",
                "metadata": {
                    "name": "webapp",
                    "namespace": "default"
                },
                "spec": {
                    "providerRef": {
                      "name": "on-call-slack"
                    },
                    "eventSeverity": "info",
                    "eventSources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "Bucket",
                          "name": "secrets"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-backend"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-frontend"
                      }
                    ]
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Receiver",
                "metadata": {
                    "name": "github-receiver",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "events": [
                      "ping",
                      "push"
                    ],
                    "secretRef": {
                      "name": "webhook-token"
                    },
                    "resources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "HelmRepository",
                          "name": "webapp"
                      }
                    ]
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "image": "ghcr.io/stefanprodan/podinfo",
                    "interval": "1m0s"
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImagePolicy",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "imageRepositoryRef": {
                      "name": "podinfo"
                    },
                    "policy": {
                      "semver": {
                          "range": "5.0.x"
                      }
                    }
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageUpdateAutomation",
                "metadata": {
                  "name": "flux-system",
                  "namespace": "flux-system"
                },
                "spec": {
                  "interval": "1m0s",
                  "sourceRef": {
                    "kind": "GitRepository",
                    "name": "flux-system"
                  },
                  "git": {
                    "checkout": {
                      "ref": {
                        "branch": "main"
                      }
                    },
                    "commit": {
                      "author": {
                        "email": "fluxcdbot@users.noreply.github.com",
                        "name": "fluxcdbot"
                      },
                      "messageTemplate": "{{range .Updated.Images}}{{println .}}{{end}}"
                    },
                    "push": {
                      "branch": "main"
                    }
                  },
                  "update": {
                    "path": "./clusters/my-cluster",
                    "strategy": "Setters"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.connect.redhat.com/flux/source-controller@sha256:b576e3f010e3ff07e716064ab0dad042aba93f0ea5c250535dd18bd678c45337
          description: Flux is a Continuous Delivery solution for Kubernetes.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/flux/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/flux/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: flux-system
          support: The Flux community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta1
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta2
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta2
          - description: OCIRepository
            displayName: OCIRepository
            kind: OCIRepository
            name: ocirepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta2
        description: |-
          [Flux](https://fluxcd.io/) is a Continuous Delivery solution for Kubernetes.
          Flux is constructed with the [GitOps Toolkit](https://toolkit.fluxcd.io/components/), a set of composable APIs and specialized tools for keeping Kubernetes clusters in sync with sources of configuration (like Git & Helm repositories), and automating updates to configuration when there is new code to deploy.

          Flux is a CNCF incubation project made for:

          - **cluster operators** who automate provision and configuration of clusters
          - **platform engineers** who build continuous delivery for developer teams
          - **app developers** who rely on continuous delivery to get their code live

          Features:

          - Source configuration from Git and Helm repositories, and S3-compatible buckets (e.g., Minio)
          - Kustomize and Helm support
          - Event-triggered and periodic reconciliation
          - Integration with Kubernetes RBAC
          - Health assessment (clusters and workloads)
          - Dependency management (infrastructure and workloads)
          - Secrets management (builtin support for Mozilla SOPS and Cloud KMS)
          - Alerting to external systems (webhook senders)
          - External events handling (webhook receivers)
          - Automated container image updates to Git (image scanning and patching)
          - Policy-driven validation (OPA, Kyverno, admission controllers)
          - Seamless integration with Git providers (GitHub, GitLab, Bitbucket)
          - Interoperability with Cluster API (CAPI) providers

          ### Steps to support bootstraping with Flux CLI

          SecComp Profile should be removed as it requires a special privilege to get the access.
          Also, it requires to enforce `runUserAs` to the same UID provided by the images to prevent OpenShift
          to alter the user id expected by our controllers.

          To do so, please specifiy a patch using the following Kustomization file, for example.

          ```yaml
          apiVersion: kustomize.config.k8s.io/v1beta1
          kind: Kustomization
          resources:
            - gotk-components.yaml
            - gotk-sync.yaml
          patches:
            - patch: |
                apiVersion: apps/v1
                kind: Deployment
                metadata:
                  name: all
                spec:
                  template:
                    spec:
                      containers:
                        - name: manager
                          securityContext:
                            runAsUser: 65534
                            seccompProfile:
                              $patch: delete
              target:
                kind: Deployment
                labelSelector: app.kubernetes.io/part-of=flux
          ```

          Then before installing Flux, please also perform the following steps to bind Flux to the `nonroot` SecurityContextConstraint (SCC). This applies to installation via OC's OLM or via Flux CLI.

          ```sh
          # replace default "flux-system" with the project/namespace name for your flux installation
          namespace="flux-system"
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:source-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:kustomize-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:helm-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:notification-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-automation-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-reflector-controller

          ```

          To see complete steps please refer to this [documentation](https://github.com/fluxcd/website/blob/main/content/en/docs/use-cases/openshift.md).

          If you've already installed Flux, you might need to uninstall it after applying the above pre-requisite steps.

          ### Network Policies

          This version of Flux is shipped without the normally provided network Policies.
          By default all traffic is enabled, and anyone can access the source-controller
          artifacts.

          It is not recommended to run it in production as is.

          Please visit [Flux's policies](https://github.com/fluxcd/flux2/tree/main/manifests/policies)
          to deploy the needed network policies and secure your Flux setup.

          ### Breaking Changes

          - Flux 0.18.x ships with v1beta2 of the kustomize.toolkit.fluxcd.io/Kustomization CRD.
          - Flux 0.25.x is the last Flux release where Kubernetes 1.19 is supported.

          ### Flux on OpenShift Use Cases

          - [The GitOps with Flux + OpenShift](https://www.youtube.com/watch?v=8ybCtNVP5To) talk by Andrew Block (Red Hat) showing how to setup multi-tenacy on OpenShift using Flux.
          - [Flux OpenShift Use Case](https://fluxcd.io/docs/use-cases/openshift/) on FluxCD.io
        displayName: Flux
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - gitops
        - cd
        - continuous delivery
        - fluxcd
        - helm
        - kustomize
        links:
        - name: Homepage
          url: https://fluxcd.io/
        - name: Documentation
          url: https://fluxcd.io/docs/
        - name: Slack
          url: https://cloud-native.slack.com/messages/flux
        - name: Mailing List
          url: https://lists.cncf.io/g/cncf-flux-dev
        - name: Source Code
          url: https://github.com/fluxcd/flux2
        maintainers:
        - email: support@weave.works
          name: The Flux maintainers
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: The Flux project
        relatedImages:
        - registry.connect.redhat.com/flux/helm-controller@sha256:cc24a7b73b023ad86a6c3a845088d153ea290dda7cab1ec03eb85aaf814e6049
        - registry.connect.redhat.com/flux/image-automation-controller@sha256:0c51595307e14acc15f22d50b414a87513ebb03530ef4a8f7d6cd9f8b757243e
        - registry.connect.redhat.com/flux/image-reflector-controller@sha256:3bef7fde22be641304e5ca69f77150e416eba5141f8aa13f68be7f810e55aa59
        - registry.connect.redhat.com/flux/kustomize-controller@sha256:246290908d899b9f6feb920f88913bd2d4673cf852a3020714fd4152c7f78237
        - registry.connect.redhat.com/flux/notification-controller@sha256:b3d81234ba3312e0106f2f7cabc73fbe1d7bf683eda36b9ed5bbdaa43789be40
        - registry.connect.redhat.com/flux/source-controller@sha256:b576e3f010e3ff07e716064ab0dad042aba93f0ea5c250535dd18bd678c45337
        version: 2.0.1
      entries:
      - name: flux.v2.0.1
        version: 2.0.1
      name: stable
    defaultChannel: stable
    packageName: flux
    provider:
      name: The Flux project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: telemetry-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: telemetry-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "Autoscaling",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "telemetry-operator",
                    "app.kubernetes.io/instance": "autoscaling-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "autoscaling",
                    "app.kubernetes.io/part-of": "telemetry-operator"
                  },
                  "name": "autoscaling"
                },
                "spec": {
                  "aodh": {
                    "databaseAccount": "aodh",
                    "databaseInstance": "openstack",
                    "memcachedInstance": "memcached",
                    "passwordSelectors": null,
                    "secret": "osp-secret"
                  }
                }
              },
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "Ceilometer",
                "metadata": {
                  "name": "ceilometer",
                  "namespace": "openstack"
                },
                "spec": {
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "Logging",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "telemetry-operator",
                    "app.kubernetes.io/instance": "logging-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "logging",
                    "app.kubernetes.io/part-of": "telemetry-operator"
                  },
                  "name": "logging"
                },
                "spec": {
                  "cloNamespace": "openshift-logging",
                  "port": 10514
                }
              },
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "MetricStorage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "telemetry-operator",
                    "app.kubernetes.io/instance": "metricstorage-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "metricstorage",
                    "app.kubernetes.io/part-of": "telemetry-operator"
                  },
                  "name": "metricstorage-sample"
                },
                "spec": {
                  "monitoringStack": {
                    "alertingEnabled": true,
                    "scrapeInterval": "30s",
                    "storage": {
                      "persistent": {
                        "pvcStorageRequest": "20G"
                      },
                      "retention": "24h",
                      "strategy": "persistent"
                    }
                  }
                }
              },
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "Telemetry",
                "metadata": {
                  "name": "telemetry"
                },
                "spec": {
                  "autoscaling": {
                    "aodh": {
                      "databaseAccount": "aodh",
                      "databaseInstance": "openstack",
                      "memcachedInstance": "memcached",
                      "passwordSelectors": null,
                      "secret": "osp-secret"
                    },
                    "enabled": false,
                    "heatInstance": "heat"
                  },
                  "ceilometer": {
                    "enabled": true,
                    "secret": "osp-secret"
                  },
                  "logging": {
                    "enabled": false,
                    "port": 10514
                  },
                  "metricStorage": {
                    "enabled": false,
                    "monitoringStack": {
                      "alertingEnabled": true,
                      "scrapeInterval": "30s",
                      "storage": {
                        "persistent": {
                          "pvcStorageRequest": "20G"
                        },
                        "retention": "24h",
                        "strategy": "persistent"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:47Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Autoscaling is the Schema for the autoscalings API
            displayName: Autoscaling
            kind: Autoscaling
            name: autoscalings.telemetry.openstack.org
            version: v1beta1
          - description: Ceilometer is the Schema for the ceilometers API
            displayName: Ceilometer
            kind: Ceilometer
            name: ceilometers.telemetry.openstack.org
            version: v1beta1
          - description: Logging is the Schema for the loggings API
            displayName: Logging
            kind: Logging
            name: loggings.telemetry.openstack.org
            version: v1beta1
          - description: MetricStorage is the Schema for the metricstorages API
            displayName: Metric Storage
            kind: MetricStorage
            name: metricstorages.telemetry.openstack.org
            version: v1beta1
          - description: Telemetry is the Schema for the telemetry API
            displayName: Telemetry
            kind: Telemetry
            name: telemetries.telemetry.openstack.org
            version: v1beta1
        description: Telemetry Operator
        displayName: Telemetry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Telemetry
        - Ceilometer
        links:
        - name: Telemetry Operator
          url: https://github.com/openstack-k8s-operators/telemetry-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-ceilometer-notification-rhel9@sha256:239e5376f65682fd644e8d9b10b27ad879747a056c8fd26ff19882e34c5bb056
        - registry.redhat.io/rhoso/openstack-ceilometer-ipmi-rhel9@sha256:86618d8f54df09210349e382188f3b6563ab96ee511fb4da6eb6dc80e4027aae
        - registry.redhat.io/rhoso/openstack-aodh-listener-rhel9@sha256:3599a25515a96b3ff4768f85fecb4bf955789fb9b5d8abc3181678062020ae64
        - registry.redhat.io/rhoso/openstack-ceilometer-compute-rhel9@sha256:b8a0412041c50e312bef39c17a9c2600e9dbb42ea380d55e58243cda641cef4f
        - registry.redhat.io/rhoso/openstack-aodh-evaluator-rhel9@sha256:16027b7d4f07385c87f97e1581d4454d4ce7ed654499720d9ba5bc6fc74fc246
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/telemetry-rhel9-operator@sha256:83962f19193f98b87c54136f02d18797a8047f65e0d75c11ce9518e50aabc050
        - registry.redhat.io/rhoso/openstack-ceilometer-central-rhel9@sha256:1c79d1edb2404a8614ea29f24e2a869695d75b1695ea0937eebc9c57250fc4ab
        - registry.redhat.io/rhoso/openstack-aodh-api-rhel9@sha256:5ab39a468ab014f2fec4cb9cd210b430ae1c1262f52b9e0f22881d6d8e9def97
        - registry.redhat.io/rhoso/openstack-aodh-notifier-rhel9@sha256:f729548ac3d718f2f5b6c85045e7c676e1656207ee54f43540d246bd97426894
        - registry.redhat.io/rhoso-operators/sg-core-rhel9@sha256:a6e4edc24409361cb770bb44fe5b475dc58748426ccc4003abad7e70f01f1b97
        - registry.redhat.io/ubi9/httpd-24@sha256:dbfbe3383e12d63df7178e4cdd5b1e6e51a84c781041b9192026ea6edb72fbf3
        version: 1.0.6
      entries:
      - name: telemetry-operator.v1.0.6
        version: 1.0.6
      - name: telemetry-operator.v1.0.5
        version: 1.0.5
      - name: telemetry-operator.v1.0.4
        version: 1.0.4
      - name: telemetry-operator.v1.0.3
        version: 1.0.3
      - name: telemetry-operator.v1.0.2
        version: 1.0.2
      - name: telemetry-operator.v1.0.1
        version: 1.0.1
      - name: telemetry-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: telemetry-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Trustification
      provider-url: ""
    name: trustify-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trustify-operator.v0.1.0-alpha.9
      currentCSVDesc:
        annotations:
          alm-examples: |
            [{
              "apiVersion": "org.trustify/v1alpha1",
              "kind": "Trustify",
              "metadata": {
                "name": "myapp"
              },
              "spec": { }
            }]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          containerImage: ghcr.io/trustification/trustify-operator:v0.1.0-alpha.9
          createdAt: "2025-03-07T11:57:13+00:00"
          description: An Operator for installing and managing Trustify
          repository: https://github.com/trustification/trustify-operator
          support: https://github.com/trustification/trustify-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Trustify instance
            displayName: Trustify
            kind: Trustify
            name: trustifies.org.trustify
            version: v1alpha1
        description: |+
          Trustify is vendor-neutral, thought-leadering, mostly informational collection of resources devoted to making Software Supply Chains easier to create, manage, consume and ultimately... to trust!


          You can use Trustify for:

          - Store and Manage all the SBOM (Software Bill of Materials) files of your company.
          - Understand which are the Vulnerabilities that affect each of your SBOM files
          - Understand exactly which Packages are included/shipped within each SBOM

          Trustify is a project within the [Trustification community](https://trustification.io/).


          ### Install
          Once you have successfully installed the Operator, proceed to deploy components by creating the required CR.

          By default, the Operator installs the following components on a target cluster:

          * Server
          * UI

          ### Documentation
          Documentation can be found on our [website](https://trustification.io/).

          ### Getting help
          If you encounter any issues while using Trustify, you can create an issue on our [Github repo](https://github.com/trustification/trustify/issues), for bugs, enhancements or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find
          * Fixing issues by opening [Pull Requests](https://github.com/trustification/trustify/pulls)

        displayName: Trustify Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trust
        links:
        - name: Website
          url: https://trustification.io/
        - name: Github
          url: https://github.com/trustification/trustify
        maintainers:
        - email: trustification@googlegroups.com
          name: Trustification
        maturity: alpha
        minKubeVersion: 1.23.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Trustification
        relatedImages:
        - ghcr.io/trustification/trustify-operator:v0.1.0-alpha.9
        version: 0.1.0-alpha.9
      entries:
      - name: trustify-operator.v0.1.0-alpha.9
        version: 0.1.0-alpha.9
      - name: trustify-operator.v0.1.0-alpha.8
        version: 0.1.0-alpha.8
      - name: trustify-operator.v0.1.0-alpha.7
        version: 0.1.0-alpha.7
      - name: trustify-operator.v0.1.0-alpha.6
        version: 0.1.0-alpha.6
      - name: trustify-operator.v0.1.0-alpha.5
        version: 0.1.0-alpha.5
      - name: trustify-operator.v0.1.0-alpha.4
        version: 0.1.0-alpha.4
      - name: trustify-operator.v0.1.0-alpha.3
        version: 0.1.0-alpha.3
      - name: trustify-operator.v0.1.0-alpha.2
        version: 0.1.0-alpha.2
      name: alpha
    defaultChannel: alpha
    packageName: trustify-operator
    provider:
      name: Trustification
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Aqua Security, Inc.
      provider-url: ""
    name: aqua
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aqua-operator.v2022.4.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaCsp",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "namespace": "aqua",
                    "version": "2022.4",
                    "requirements": true
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry",
                    "dbDiskSize": 10,
                    "databaseSecret": {
                      "key": "db-password",
                      "name": "aqua-database-password"
                    }
                  },
                  "database": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "database",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "gateway": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "gateway",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "server": {
                    "replicas": 1,
                    "service": "LoadBalancer",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "console",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "route": true,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaDatabase",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "database",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "diskSize": 10,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "image": {
                      "repository": "enforcer",
                      "registry": "registry.aquasec.com",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "IfNotPresent"
                    }
                  },
                  "runAsNonRoot": false,
                  "gateway": {
                    "host": "aqua-gateway",
                    "port": 8443
                  },
                  "token": "<<your-token>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaKubeEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "config": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "aqua-secure",
                    "imagePullSecret": "aqua-registry",
                    "kubebench_image": "docker.io/aquasec/kube-bench:v0.7.1"
                  },
                  "deploy": {
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "tag": "<<KUBE_ENFORCER_TAG>>",
                      "repository": "kube-enforcer",
                      "pullPolicy": "Always"
                    }
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "starboard-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  }
                }
              },
              {
              "apiVersion": "operator.aquasec.com/v1alpha1",
              "kind": "AquaLightning",
              "metadata": {
                "name": "aqua"
              },
              "spec": {
                "global": {
                  "gateway_address": "aqua-gateway.aqua:8443",
                  "cluster_name": "Default-cluster-name"
                },
                "common": {
                  "imagePullSecret": "aqua-registry"
                },
                "kubeEnforcer": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                  "allowAnyVersion": null,
                  "deploy": {
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "tag": "2022.4",
                      "repository": "kube-enforcer",
                      "pullPolicy": "Always"
                    }
                  },
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "starboard-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  },
                  "env": []
                },
                "enforcer": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "deploy": {
                    "image": {
                      "repository": "enforcer",
                      "registry": "registry.aquasec.com",
                      "tag": "2022.4",
                      "pullPolicy": "IfNotPresent"
                    }
                  },
                  "token": "<<ENFORCER_GROUP_TOKEN>>",
                  "env": [],
                  "secret": {
                    "name": null,
                    "key": null
                  },
                  "runAsNonRoot": null,
                  "rhcosVersion": null
                }
              }
            },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaGateway",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "gateway",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaScanner",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "replicas": 1,
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "scanner",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "runAsNonRoot": false,
                  "login": {
                    "username": "<<YOUR AQUA USER NAME>>",
                    "password": "<<YOUR AQUA USER PASSWORD>>",
                    "host": "http://aqua-server:8080",
                    "token": ""
                  }
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaServer",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "LoadBalancer",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "console",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "runAsNonRoot": false
                }
              },
              {
                  "apiVersion": "operator.aquasec.com/v1alpha1",
                  "kind": "AquaCloudConnector",
                  "metadata": {
                    "name": "aqua"
                  },
                  "spec": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "common": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "image": {
                        "repository": "aqua-cloud-connector",
                        "registry": "registry.aquasec.com",
                        "tag": "<<IMAGE TAG>>"
                      },
                      "replicas": 1
                    },
                    "runAsNonRoot": false,
                    "gateway": {
                      "host": "aqua-gateway",
                      "port": "8443"
                    },
                    "login": {
                      "host": "aqua-gateway:8443",
                      "password": "<<YOUR AQUA USER PASSWORD>>",
                      "token": "<<YOUR AQUA CLOUD CONNECTOR TOKEN>>",
                      "username": "<<YOUR AQUA USER NAME>>"
                    },
                    "tunnels": [
                      {
                        "host": "<account>.azurecr.io",
                        "port": "443"
                      },
                      {
                        "host": "gke.gcr.io",
                        "port": "443"
                      },
                      {
                        "host": "<account>.jfrog.io",
                        "port": "443"
                      },
                      {
                        "host": "1.1.1.1",
                        "port": "443"
                      },
                      {
                        "region": "us-east1",
                        "type": "ecr"
                      }
                    ]
                  }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: docker.io/aquasec/aqua-operator:2022.4.18
          createdAt: "2020-03-25T08:00:00Z"
          description: The Aqua Security Operator runs within a Openshift cluster
            and provides a means to deploy and manage Aqua Security cluster and components.
          operators.operatorframework.io/internal-objects: '["aquastarboards.aquasecurity.github.io","configauditreports.aquasecurity.github.io",
            "clusterconfigauditreports.aquasecurity.github.io"]'
          repository: https://github.com/aquasecurity/aqua-operator
          support: Aqua Security, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aqua Security CSP Deployment with Aqua Operator
            displayName: AquaCsp
            kind: AquaCsp
            name: aquacsps.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Lightning Enforcers Deployment with Aqua Operator
            displayName: AquaLightningEnforcers
            kind: AquaLightning
            name: aqualightnings.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Database Deployment with Aqua Operator
            displayName: AquaDatabase
            kind: AquaDatabase
            name: aquadatabases.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Enforcer Deployment with Aqua Operator
            displayName: AquaEnforcer
            kind: AquaEnforcer
            name: aquaenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security KubeEnforcer Deployment with Aqua Operator
            displayName: AquaKubeEnforcer
            kind: AquaKubeEnforcer
            name: aquakubeenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Gateway Deployment with Aqua Operator
            displayName: AquaGateway
            kind: AquaGateway
            name: aquagateways.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Scanner Deployment with Aqua Operator
            displayName: AquaScanner
            kind: AquaScanner
            name: aquascanners.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Server Deployment with Aqua Operator
            displayName: AquaServer
            kind: AquaServer
            name: aquaservers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Starboard for kube-enforcer.
            displayName: AquaStarboard
            kind: AquaStarboard
            name: aquastarboards.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ClusterConfigAuditReport
            kind: ClusterConfigAuditReport
            name: clusterconfigauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ConfigAuditReport
            kind: ConfigAuditReport
            name: configauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: Aqua Security Cloud Connector Deployment with Aqua Operator
            displayName: AquaCloudConnector
            kind: AquaCloudConnector
            name: aquacloudconnector.operator.aquasec.com
            version: v1alpha1
        description: |-
          The Aqua Security Operator runs within an OpenShift cluster and provides a means to deploy and manage the Aqua Security cluster and components :
          * Server (aka “console”)
          * Database (for production environments we recommend to use an external database and not the Aqua default database)
          * Gateway
          * Enforcer (aka “agent”)
          * KubeEnforcer
          * Scanner
          * CSP (package that contains the Server, Database, and Gateway)
          * LightningEnforcer (package that contains Enforcer and Kube Enforcer)

          Use the Aqua-Operator to
          * Deploy Aqua Security components on OpenShift
          * Scale up Aqua Security components with extra replicas
          * Assign metadata tags to Aqua Security components

          ## Before You Begin Using the Operator CRDs

          Obtain access to the Aqua registry - https://www.aquasec.com/about-us/contact-us/


          ## Instructions and exmaples

          Follow instructions and exmaples
          [here](https://github.com/aquasecurity/aqua-operator/blob/2022.4.0/docs/DeployOpenShiftOperator.md)
        displayName: Aqua Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aqua-security
        - scanning
        - security
        - runtime-security
        links:
        - name: Aqua Security
          url: https://www.aquasec.com/
        - name: Aqua Operator Github
          url: https://github.com/aquasecurity/aqua-operator
        maintainers:
        - email: support@aquasec.com
          name: Aqua Support, Aqua Security
        maturity: alpha
        provider:
          name: Aqua Security, Inc.
        relatedImages:
        - docker.io/aquasec/aqua-operator:2022.4.18
        version: 2022.4.18
      entries:
      - name: aqua-operator.v2022.4.18
        version: 2022.4.18
      - name: aqua-operator.v2022.4.17
        version: 2022.4.17
      name: 2022.4.0
    defaultChannel: 2022.4.0
    packageName: aqua
    provider:
      name: Aqua Security, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: The bpfman Community
      provider-url: https://bpfman.io/
    name: bpfman-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bpfman-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfapplication"
                  },
                  "name": "bpfapplication-sample"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:v0.5.6"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "kprobe": {
                        "bpffunctionname": "kprobe_counter",
                        "func_name": "try_to_wake_up",
                        "offset": 0,
                        "retprobe": false
                      },
                      "type": "Kprobe"
                    },
                    {
                      "tracepoint": {
                        "bpffunctionname": "tracepoint_kill_recorder",
                        "names": [
                          "syscalls/sys_enter_kill"
                        ]
                      },
                      "type": "Tracepoint"
                    },
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "containernames": [
                            "bpfman",
                            "bpfman-agent"
                          ],
                          "namespace": "bpfman",
                          "pods": {
                            "matchLabels": {
                              "name": "bpfman-daemon"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfNsApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfnsapplication"
                  },
                  "name": "bpfapplication-sample",
                  "namespace": "acme"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:v0.5.6"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FentryProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fentryprogram"
                  },
                  "name": "fentry-example"
                },
                "spec": {
                  "bpffunctionname": "test_fentry",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fentry:v0.5.6"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FexitProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fexitprogram"
                  },
                  "name": "fexit-example"
                },
                "spec": {
                  "bpffunctionname": "test_fexit",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fexit:v0.5.6"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "KprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kprobeprogram"
                  },
                  "name": "kprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_kprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/kprobe:v0.5.6"
                    }
                  },
                  "func_name": "try_to_wake_up",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "offset": 0,
                  "retprobe": false
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcnsprogram"
                  },
                  "name": "tc-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcprogram"
                  },
                  "name": "tc-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:v0.5.6"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxnsprogram"
                  },
                  "name": "tcx-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "tcx_next",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxprogram"
                  },
                  "name": "tcx-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "tcx_pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:v0.5.6"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TracepointProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tracepointprogram"
                  },
                  "name": "tracepoint-example"
                },
                "spec": {
                  "bpffunctionname": "enter_openat",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tracepoint:v0.5.6"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "names": [
                    "syscalls/sys_enter_openat"
                  ],
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobensprogram"
                  },
                  "name": "uprobe-example",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobeprogram"
                  },
                  "name": "uprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:v0.5.6"
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpnsprogram"
                  },
                  "name": "xdp-ns-pass-all-nodes",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpprogram"
                  },
                  "name": "xdp-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:v0.5.6"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/bpfman/bpfman-operator:v0.5.6
          createdAt: "2025-02-18T13:44:53Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: bpfman
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "bpfman",
                "labels": {
                  "pod-security.kubernetes.io/enforce": "privileged",
                  "pod-security.kubernetes.io/audit": "privileged",
                  "pod-security.kubernetes.io/warn": "privileged",
                },
                "annotations": {
                  "openshift.io/node-selector": ""
                },
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/bpfman/bpfman
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BpfApplication is the Schema for the bpfapplications API
            displayName: Bpf Application
            kind: BpfApplication
            name: bpfapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsApplication is the Schema for the BpfNsApplications
              API
            displayName: Bpf Namespaced Application
            kind: BpfNsApplication
            name: bpfnsapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsProgram is the Schema for the BpfNsProgram API
            displayName: Bpf Namespaced Program
            kind: BpfNsProgram
            name: bpfnsprograms.bpfman.io
            version: v1alpha1
          - description: BpfProgram is the Schema for the BpfProgram API
            displayName: Bpf Program
            kind: BpfProgram
            name: bpfprograms.bpfman.io
            version: v1alpha1
          - description: FentryProgram is the Schema for the Fentryprograms API
            displayName: Fentry Program
            kind: FentryProgram
            name: fentryprograms.bpfman.io
            version: v1alpha1
          - description: FexitProgram is the Schema for the Fexitprograms API
            displayName: Fexit Program
            kind: FexitProgram
            name: fexitprograms.bpfman.io
            version: v1alpha1
          - description: KprobeProgram is the Schema for the Kprobeprograms API
            displayName: Kprobe Program
            kind: KprobeProgram
            name: kprobeprograms.bpfman.io
            version: v1alpha1
          - description: TcNsProgram is the Schema for the Tcnsprograms API
            displayName: TC Namespaced Program
            kind: TcNsProgram
            name: tcnsprograms.bpfman.io
            version: v1alpha1
          - description: TcProgram is the Schema for the Tcprograms API
            displayName: TC Program
            kind: TcProgram
            name: tcprograms.bpfman.io
            version: v1alpha1
          - description: TcxNsProgram is the Schema for the Tcxnsprograms API
            displayName: TCX Namespaced Program
            kind: TcxNsProgram
            name: tcxnsprograms.bpfman.io
            version: v1alpha1
          - description: TcxProgram is the Schema for the Tcxprograms API
            displayName: TCX Program
            kind: TcxProgram
            name: tcxprograms.bpfman.io
            version: v1alpha1
          - description: TracepointProgram is the Schema for the Tracepointprograms
              API
            displayName: Tracepoint Program
            kind: TracepointProgram
            name: tracepointprograms.bpfman.io
            version: v1alpha1
          - description: UprobeNsProgram is the Schema for the Uprobensprograms API
            displayName: Uprobe Namespaced Program
            kind: UprobeNsProgram
            name: uprobensprograms.bpfman.io
            version: v1alpha1
          - description: UprobeProgram is the Schema for the Uprobeprograms API
            displayName: Uprobe Program
            kind: UprobeProgram
            name: uprobeprograms.bpfman.io
            version: v1alpha1
          - description: XdpNsProgram is the Schema for the Xdpnsprograms API
            displayName: Xdp Namespaced Program
            kind: XdpNsProgram
            name: xdpnsprograms.bpfman.io
            version: v1alpha1
          - description: XdpProgram is the Schema for the Xdpprograms API
            displayName: Xdp Program
            kind: XdpProgram
            name: xdpprograms.bpfman.io
            version: v1alpha1
        description: "The bpfman Operator is a Kubernetes Operator for deploying [bpfman](https://bpfman.netlify.app/),
          a system daemon\nfor managing eBPF programs. It deploys bpfman itself along
          with CRDs to make deploying\neBPF programs in Kubernetes much easier.\n\n##
          Quick Start\n\nTo get bpfman up and running quickly simply click 'install'
          to deploy the bpfman-operator in the bpfman namespace via operator-hub.\n##
          Configuration\n\nThe `bpfman-config` configmap is automatically created
          in the `bpfman` namespace and used to configure the bpfman deployment.\n\nTo
          edit the config simply run\n\n```bash\nkubectl edit cm bpfman-config\n```\n\nThe
          following fields are adjustable\n\n- `bpfman.agent.image`: The image used
          for the bpfman-agent`\n- `bpfman.image`: The image used for bpfman`\n- `bpfman.log.level`:
          the log level for bpfman, currently supports `debug`, `info`, `warn`, `error`,
          and `fatal`, defaults to `info`\n- `bpfman.agent.log.level`: the log level
          for the bpfman-agent currently supports `info`, `debug`, and `trace` \n\nThe
          bpfman operator deploys eBPF programs via CRDs. The following CRDs are currently
          available, \n\n- BpfApplication\n- FentryProgram\n- FexitProgram\n - KprobeProgram\n-
          TcProgram\n- TcxProgram\n- TracepointProgram\n- UprobeProgram\n- XdpProgram\n
          - BpfNsApplication\n- TcNsProgram\n- TcxNsProgram\n- UprobeNsProgram\n-
          XdpNsProgram\n- BpfProgram\n - BpfNsProgram\n\n ## More information\n\nPlease
          checkout the [bpfman community website](https://bpfman.io/) for more information."
        displayName: Bpfman Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ebpf
        - kubernetes
        links:
        - name: bpfman website
          url: https://bpfman.io/
        maintainers:
        - email: astoycos@redhat.com
          name: Andrew Stoycos
        - email: afredette@redhat.com
          name: Andre Fredette
        maturity: alpha
        provider:
          name: The bpfman Community
          url: https://bpfman.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/bpfman/bpfman-operator:v0.5.6
        version: 0.5.6
      entries:
      - name: bpfman-operator.v0.5.6
        version: 0.5.6
      - name: bpfman-operator.v0.5.5
        version: 0.5.5
      - name: bpfman-operator.v0.5.4
        version: 0.5.4
      - name: bpfman-operator.v0.5.2
        version: 0.5.2
      - name: bpfman-operator.v0.5.1
        version: 0.5.1
      - name: bpfman-operator.v0.5.0
        version: 0.5.0
      - name: bpfman-operator.v0.4.2
        version: 0.4.2
      - name: bpfman-operator.v0.4.1
        version: 0.4.1
      name: alpha
    defaultChannel: alpha
    packageName: bpfman-operator
    provider:
      name: The bpfman Community
      url: https://bpfman.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Snyk Ltd.
      provider-url: ""
    name: snyk-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: snyk-operator.v1.92.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1alpha1",
                "kind": "SnykMonitor",
                "metadata": {
                  "name": "snyk-monitor"
                },
                "spec": {
                  "clusterName": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "snyk/kubernetes-monitor",
                    "tag": "1.92.0-ubi8"
                  },
                  "integrationApi": "",
                  "monitorSecrets": "snyk-monitor",
                  "scope": "Cluster",
                  "temporaryStorageSize": "50Gi",
                  "pvc": {
                    "enabled": false,
                    "create": false,
                    "name": "snyk-monitor-pvc",
                    "storageClassName": null
                  },
                  "initContainerImage": {
                    "repository": "busybox",
                    "tag": "latest"
                  },
                  "requests": {
                    "memory": "400Mi"
                  },
                  "limits": {
                    "memory": "2Gi"
                  },
                  "nodeAffinity": {
                    "disableBetaArchNodeSelector": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Security
          certified: "false"
          containerImage: docker.io/snyk/kubernetes-operator:1.92.0
          createdAt: "2022-06-09T17:49:36Z"
          description: A Kubernetes Operator for creating and managing Snyk controller
            instances.
          olm.skipRange: <1.68.2
          repository: https://github.com/snyk/kubernetes-monitor
          support: Snyk Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a deployment of the Snyk controller, which scans
              container images in the cluster for vulnerabilities.
            displayName: Snyk Monitor
            kind: SnykMonitor
            name: snykmonitors.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          A Kubernetes Operator for creating and managing Snyk Kubernetes controller instances.

          Snyk integrates with Kubernetes, enabling you to import and test your running workloads and identify vulnerabilities in their associated images and configurations that might make those workloads less secure. Once imported, Snyk continues to monitor those workloads, identifying additional security issues as new images are deployed and the workload configuration changes.

          ## Prerequisites

          To start using the Operator and the Snyk controller that it manages, you will need to create a secret containing your Snyk integrationId and Docker config file:

          ```
          kubectl create namespace snyk-monitor
          kubectl create secret generic snyk-monitor -n snyk-monitor --from-literal=dockercfg.json={} --from-literal=integrationId=${SNYK_INTEGRATION_ID}
          ```

          Refer to the [Snyk documentation](https://support.snyk.io/hc/en-us/articles/360003916138-Kubernetes-integration-overview) for more details.
        displayName: Snyk Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - snyk
        - controller
        - monitor
        - operator
        - security
        - cluster
        - container
        - image
        - security
        - vulnerabilities
        - advisories
        - audit
        - scan
        - scanning
        links:
        - name: Website
          url: https://snyk.io
        - name: Documentation
          url: https://support.snyk.io/hc/en-us/articles/360003916138-Kubernetes-integration-overview
        maintainers:
        - email: support@snyk.io
          name: Snyk Ltd.
        maturity: stable
        provider:
          name: Snyk Ltd.
        relatedImages:
        - docker.io/snyk/kubernetes-operator:1.92.0
        version: 1.92.0
      entries:
      - name: snyk-operator.v1.92.0
        version: 1.92.0
      - name: snyk-operator.v1.90.2
        version: 1.90.2
      - name: snyk-operator.v1.90.1
        version: 1.90.1
      - name: snyk-operator.v1.88.4
        version: 1.88.4
      - name: snyk-operator.v1.85.2
        version: 1.85.2
      - name: snyk-operator.v1.82.1
        version: 1.82.1
      - name: snyk-operator.v1.81.1
        version: 1.81.1
      - name: snyk-operator.v1.80.0
        version: 1.80.0
      - name: snyk-operator.v1.73.0
        version: 1.73.0
      - name: snyk-operator.v1.72.0
        version: 1.72.0
      - name: snyk-operator.v1.70.4
        version: 1.70.4
      - name: snyk-operator.v1.70.1
        version: 1.70.1
      name: stable
    defaultChannel: stable
    packageName: snyk-operator
    provider:
      name: Snyk Ltd.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-sds-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-sds-operator.v8.0.230
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.230"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleStoragePool",
                "metadata": {
                  "name": "infoscalestoragepool-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:f4947eab2b31fccc86b5d5d0d3fd4eb1f7513fd8325e379cc259f5ccd1dfc3e4
          createdAt: "2023-10-26T20:36:59Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          olm.skipRange: <8.0.230
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleStoragePool is the Schema for the infoscalestoragepools
              API
            displayName: InfoScale Storage Pool
            kind: InfoScaleStoragePool
            name: infoscalestoragepools.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.230"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:f4947eab2b31fccc86b5d5d0d3fd4eb1f7513fd8325e379cc259f5ccd1dfc3e4
        version: 8.0.230
      entries:
      - name: infoscale-sds-operator.v8.0.230
        version: 8.0.230
      - name: infoscale-sds-operator.v8.0.220
        version: 8.0.220
      - name: infoscale-sds-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-sds-operator.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-sds-operator.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.330"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:dab501b99373c840c7e8e3ab479fa273b39f9bf8a896f3f1e4e945f4ab81f64e
          createdAt: "2024-06-24T09:03:32Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.330"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:dab501b99373c840c7e8e3ab479fa273b39f9bf8a896f3f1e4e945f4ab81f64e
        version: 8.0.330
      entries:
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-sds-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.410"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
          createdAt: "2024-07-30T09:00:57Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Arctera InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/168721626-168743109-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.410"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Arctera recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
        version: 8.0.410
      entries:
      - name: infoscale-sds-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-sds-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-sds-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-sds-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.410"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
          createdAt: "2024-07-30T09:00:57Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Arctera InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/168721626-168743109-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.410"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Arctera recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
        version: 8.0.410
      entries:
      - name: infoscale-sds-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-sds-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-sds-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-sds-operator.v8.0.210
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.210"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:26573b98743245c0b607abbe0d981abce6ce9fb1c5c0cd9fe735351b7136f4e8
          createdAt: "2022-11-26T16:36:59Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleStoragePool is the Schema for the infoscalestoragepools
              API
            displayName: InfoScale Storage Pool
            kind: InfoScaleStoragePool
            name: infoscalestoragepools.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.210"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:1667227b5111fe52ed304ab19ff9fadfc677239e56b2c93a3b89b1bfbc31354d
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:26573b98743245c0b607abbe0d981abce6ce9fb1c5c0cd9fe735351b7136f4e8
        version: 8.0.210
      entries:
      - name: infoscale-sds-operator.v8.0.210
        version: 8.0.210
      - name: infoscale-sds-operator.v8.0.201
        version: 8.0.201
      - name: infoscale-sds-operator.v8.0.200
        version: 8.0.200
      - name: infoscale-sds-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-sds-operator.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-sds-operator
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: ovn-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ovn-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ovn.openstack.org/v1beta1",
                "kind": "OVNController",
                "metadata": {
                  "name": "ovncontroller-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "ovn.openstack.org/v1beta1",
                "kind": "OVNDBCluster",
                "metadata": {
                  "name": "ovndbcluster-nb-sample"
                },
                "spec": {
                  "dbType": "NB",
                  "logLevel": "info",
                  "storageClass": "local-storage",
                  "storageRequest": "10G"
                }
              },
              {
                "apiVersion": "ovn.openstack.org/v1beta1",
                "kind": "OVNDBCluster",
                "metadata": {
                  "name": "ovndbcluster-sb-sample"
                },
                "spec": {
                  "dbType": "SB",
                  "logLevel": "info",
                  "storageClass": "local-storage",
                  "storageRequest": "10G"
                }
              },
              {
                "apiVersion": "ovn.openstack.org/v1beta1",
                "kind": "OVNNorthd",
                "metadata": {
                  "name": "ovnnorthd-sample"
                },
                "spec": {
                  "logLevel": "info"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:45Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OVNController is the Schema for the ovncontrollers API
            displayName: OVNController
            kind: OVNController
            name: ovncontrollers.ovn.openstack.org
            version: v1beta1
          - description: OVNDBCluster is the Schema for the ovndbclusters API
            displayName: OVNDBCluster
            kind: OVNDBCluster
            name: ovndbclusters.ovn.openstack.org
            version: v1beta1
          - description: OVNNorthd is the Schema for the ovnnorthds API
            displayName: OVNNorthd
            kind: OVNNorthd
            name: ovnnorthds.ovn.openstack.org
            version: v1beta1
        description: OVN Operator
        displayName: OVN Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Network
        - Ovn
        links:
        - name: Ovn Operator
          url: https://github.com/openstack-k8s-operators/ovn-operator
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/ovn-rhel9-operator@sha256:bce19011b3623c66a59efe6ea874e2403440813bdc970cd6170853a8093076f3
        - registry.redhat.io/rhoso/openstack-ovn-base-rhel9@sha256:b0d424f6ea6de8192382541a95aa56fa32b9b95e8d433c41c4a6dadbe4127fae
        - registry.redhat.io/rhoso/openstack-ovn-nb-db-server-rhel9@sha256:e71c8aac00a78e3d786332aaada3657276f69483df1e441b70589601ca038a5b
        - registry.redhat.io/rhoso/openstack-ovn-sb-db-server-rhel9@sha256:7aa2832f80846d5ac71733a1f186d39dce3654ebb661d32b8e4580d92be8e338
        - registry.redhat.io/rhoso/openstack-ovn-northd-rhel9@sha256:f1e4ce04b67d888fa950b83d1f00459c67c602397336da81de3ff71d4052a5d8
        - registry.redhat.io/rhoso/openstack-ovn-controller-rhel9@sha256:6c7827dbe25c168d28250f197d5ea6f3f60cfbab4c974267ec5ea0f8d631c3f5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: ovn-operator.v1.0.6
        version: 1.0.6
      - name: ovn-operator.v1.0.5
        version: 1.0.5
      - name: ovn-operator.v1.0.4
        version: 1.0.4
      - name: ovn-operator.v1.0.3
        version: 1.0.3
      - name: ovn-operator.v1.0.2
        version: 1.0.2
      - name: ovn-operator.v1.0.1
        version: 1.0.1
      - name: ovn-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: ovn-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: fuse-apicurito
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-apicurito.v7.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1alpha1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:8d19ad579f874d6b9a08f127c8900906913a5a7c1e69cbe4b5cb0333b8d86152
          createdAt: 2022-02-28 09:35:38 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.10.3'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:8d19ad579f874d6b9a08f127c8900906913a5a7c1e69cbe4b5cb0333b8d86152
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:8dbeeb8b04c81f41d896f71db90813de57e67493fa85389b63ac23f390785043
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:c02c218ccfbb385ea29af98409995a3d9c26cb0b4c76caf2e0779fa233b5ff8c
        version: 7.10.3
      entries:
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.10.x
    - currentCSV: fuse-apicurito.v7.11.1-0.1687357785.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:2f8d4450187839f0e0c761e5ddcf61b03f5996acfe3d75b60120e76e5b0167c1
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.11.2'
          olm.substitutesFor: fuse-apicurito.v7.11.1-0.1682509045.p
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:2f8d4450187839f0e0c761e5ddcf61b03f5996acfe3d75b60120e76e5b0167c1
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:df61aafc4023b9068a8636c35eabb0d5fda400f523b64cb1fc5d8c480272c435
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:4c493a2d690d0d7c1684c53d15c06b47d691aeffdde6dd471b6e75fed2fec6a9
        version: 7.11.1+0.1687357785.p
      entries:
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.11.x
    - currentCSV: fuse-apicurito.v7.12.1-0.1706187670.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:7ccefee63165e7eaf9a609e8a9f978a06f27deff69a318fe399a29292ad99106
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.12.1'
          olm.substitutesFor: fuse-apicurito.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:7ccefee63165e7eaf9a609e8a9f978a06f27deff69a318fe399a29292ad99106
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:4cd612283e8136839f1f48a3a220dac7c6c38adf14b61bc03a981e514baefe9b
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:bee0eeaa7898af835f80c5217059d73361daa42197fa6cbbaf416e1ba373eb54
        version: 7.12.1+0.1706187670.p
      entries:
      - name: fuse-apicurito.v7.12.1-0.1706187670.p
        version: 7.12.1+0.1706187670.p
      - name: fuse-apicurito.v7.12.1
        version: 7.12.1
      - name: fuse-apicurito.v7.12.0
        version: 7.12.0
      - name: fuse-apicurito.v7.12.0-0.1697748407.p
        version: 7.12.0+0.1697748407.p
      - name: fuse-apicurito.v7.12.0-0.1697419531.p
        version: 7.12.0+0.1697419531.p
      - name: fuse-apicurito.v7.12.0-0.1694087055.p
        version: 7.12.0+0.1694087055.p
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.12.x
    - currentCSV: fuse-apicurito.v7.13.0-0.1753785079.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:d58d345d74059242d57d5c1a75f31517769230d3014a42243fda20aa93ebd5c8
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.7.0 <7.13.0'
          olm.substitutesFor: fuse-apicurito.v7.13.0-0.1751291353.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:d58d345d74059242d57d5c1a75f31517769230d3014a42243fda20aa93ebd5c8
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:f5bf9240b28b3664bdb11b1f57f611e6194ff8643ae939b6ab3b2b176705f487
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:4e64a5930a10b76147ac128a216cd7c0346da2ff3e9f31ed86311941ed193cc1
        version: 7.13.0+0.1753785079.p
      entries:
      - name: fuse-apicurito.v7.13.0-0.1739885035.p
        version: 7.13.0+0.1739885035.p
      - name: fuse-apicurito.v7.13.0-0.1741884705.p
        version: 7.13.0+0.1741884705.p
      - name: fuse-apicurito.v7.13.0-0.1753785079.p
        version: 7.13.0+0.1753785079.p
      - name: fuse-apicurito.v7.13.0-0.1721224330.p
        version: 7.13.0+0.1721224330.p
      - name: fuse-apicurito.v7.13.0-0.1753778625.p
        version: 7.13.0+0.1753778625.p
      - name: fuse-apicurito.v7.13.0-0.1753702605.p
        version: 7.13.0+0.1753702605.p
      - name: fuse-apicurito.v7.13.0-0.1721607497.p
        version: 7.13.0+0.1721607497.p
      - name: fuse-apicurito.v7.13.0-0.1751291353.p
        version: 7.13.0+0.1751291353.p
      - name: fuse-apicurito.v7.13.0-0.1748960166.p
        version: 7.13.0+0.1748960166.p
      - name: fuse-apicurito.v7.13.0-0.1745523192.p
        version: 7.13.0+0.1745523192.p
      - name: fuse-apicurito.v7.13.0-0.1744737027.p
        version: 7.13.0+0.1744737027.p
      - name: fuse-apicurito.v7.13.0-0.1744280419.p
        version: 7.13.0+0.1744280419.p
      - name: fuse-apicurito.v7.13.0-0.1741875975.p
        version: 7.13.0+0.1741875975.p
      - name: fuse-apicurito.v7.13.0-0.1739473863.p
        version: 7.13.0+0.1739473863.p
      - name: fuse-apicurito.v7.13.0-0.1738056248.p
        version: 7.13.0+0.1738056248.p
      - name: fuse-apicurito.v7.13.0-0.1731460815.p
        version: 7.13.0+0.1731460815.p
      - name: fuse-apicurito.v7.13.0-0.1729189341.p
        version: 7.13.0+0.1729189341.p
      - name: fuse-apicurito.v7.13.0-0.1727957847.p
        version: 7.13.0+0.1727957847.p
      - name: fuse-apicurito.v7.13.0-0.1727193594.p
        version: 7.13.0+0.1727193594.p
      - name: fuse-apicurito.v7.13.0-0.1725022690.p
        version: 7.13.0+0.1725022690.p
      - name: fuse-apicurito.v7.13.0-0.1724659732.p
        version: 7.13.0+0.1724659732.p
      - name: fuse-apicurito.v7.13.0-0.1724053718.p
        version: 7.13.0+0.1724053718.p
      - name: fuse-apicurito.v7.13.0-0.1723709434.p
        version: 7.13.0+0.1723709434.p
      - name: fuse-apicurito.v7.13.0
        version: 7.13.0
      - name: fuse-apicurito.v7.13.0-0.1720707249.p
        version: 7.13.0+0.1720707249.p
      - name: fuse-apicurito.v7.13.0-0.1718285973.p
        version: 7.13.0+0.1718285973.p
      - name: fuse-apicurito.v7.13.0-0.1717672390.p
        version: 7.13.0+0.1717672390.p
      - name: fuse-apicurito.v7.13.0-0.1753778936.p
        version: 7.13.0+0.1753778936.p
      - name: fuse-apicurito.v7.13.0-0.1753701210.p
        version: 7.13.0+0.1753701210.p
      - name: fuse-apicurito.v7.12.1-0.1706187670.p
        version: 7.12.1+0.1706187670.p
      - name: fuse-apicurito.v7.12.1
        version: 7.12.1
      - name: fuse-apicurito.v7.12.0-0.1697419531.p
        version: 7.12.0+0.1697419531.p
      - name: fuse-apicurito.v7.12.0-0.1697748407.p
        version: 7.12.0+0.1697748407.p
      - name: fuse-apicurito.v7.12.0-0.1694087055.p
        version: 7.12.0+0.1694087055.p
      - name: fuse-apicurito.v7.12.0
        version: 7.12.0
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.13.x
    - currentCSV: fuse-apicurito.v7.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
                    "apiVersion": "apicur.io/v1alpha1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
            }]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:05c9676aada2d34a985a59abcf55105a3765d75b0d3764089d2b32e9ca5d367c
          createdAt: "2020-10-06 12:48:47"
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.9.3'
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito Project
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CRD for Apicurito
            displayName: Apicurito CRD
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:05c9676aada2d34a985a59abcf55105a3765d75b0d3764089d2b32e9ca5d367c
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:5535864ce393ae0ae5ee52ed0c7661e0351fb7df2f9235da70608245519530b2
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:dc39b79c360511ff999eba5b2c299db3e08c9c244a0a1f3da8c90c02695e506d
        version: 7.9.3
      entries:
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.9.x
    defaultChannel: fuse-apicurito-7.13.x
    packageName: fuse-apicurito
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-organizations-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-organizations-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "organizations.services.k8s.aws/v1alpha1",
                "kind": "OrganizationalUnit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/organizations-controller:1.1.1
          createdAt: "2025-09-19T19:35:20Z"
          description: AWS Organizations controller is a service controller for managing
            Organizations resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Account represents the state of an AWS organizations Account
              resource.
            displayName: Account
            kind: Account
            name: accounts.organizations.services.k8s.aws
            version: v1alpha1
          - description: OrganizationalUnit represents the state of an AWS organizations
              OrganizationalUnit resource.
            displayName: OrganizationalUnit
            kind: OrganizationalUnit
            name: organizationalunits.organizations.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Organizations resources in AWS from within your Kubernetes cluster.

          **About Amazon Organizations**

          With AWS Organizations you can perform account management activities at scale by consolidating multiple AWS accounts into a single organization. Consolidating accounts simplifies how you use other AWS services. You can leverage the multi-account management services available in AWS Organizations with select AWS services to perform tasks on all accounts that are members of your organization.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Organizations
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - organizations
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Organizations Developer Resources
          url: https://aws.amazon.com/organizations/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: organizations maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/organizations-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-organizations-controller.v1.1.1
        version: 1.1.1
      - name: ack-organizations-controller.v1.0.13
        version: 1.0.13
      - name: ack-organizations-controller.v1.0.11
        version: 1.0.11
      - name: ack-organizations-controller.v1.0.10
        version: 1.0.10
      - name: ack-organizations-controller.v1.0.9
        version: 1.0.9
      - name: ack-organizations-controller.v1.0.8
        version: 1.0.8
      - name: ack-organizations-controller.v1.0.7
        version: 1.0.7
      - name: ack-organizations-controller.v1.0.6
        version: 1.0.6
      - name: ack-organizations-controller.v1.0.5
        version: 1.0.5
      - name: ack-organizations-controller.v1.0.4
        version: 1.0.4
      - name: ack-organizations-controller.v1.0.1
        version: 1.0.1
      - name: ack-organizations-controller.v1.0.0
        version: 1.0.0
      - name: ack-organizations-controller.v0.0.10
        version: 0.0.10
      - name: ack-organizations-controller.v0.0.9
        version: 0.0.9
      - name: ack-organizations-controller.v0.0.8
        version: 0.0.8
      - name: ack-organizations-controller.v0.0.7
        version: 0.0.7
      - name: ack-organizations-controller.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: ack-organizations-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeRocketCI
      provider-url: https://docs.kuberocketci.io
    name: nexus-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-operator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexus-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexus",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexus-sample"
                },
                "spec": {
                  "secret": "nexus-secret",
                  "url": "https://nexus-example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRepository",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrepository-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrepository",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrepository-sample"
                },
                "spec": {
                  "go": {
                    "proxy": {
                      "cleanup": {
                        "policyNames": [
                          "cleanup-policy"
                        ]
                      },
                      "httpClient": {
                        "autoBlock": true,
                        "blocked": true,
                        "connection": {
                          "retries": 1440,
                          "timeout": 1440,
                          "userAgentSuffix": "user-agent-suffix"
                        }
                      },
                      "name": "go-proxy",
                      "negativeCache": {
                        "enabled": true,
                        "timeToLive": 1440
                      },
                      "online": true,
                      "proxy": {
                        "contentMaxAge": 1440,
                        "metadataMaxAge": 1440,
                        "remoteUrl": "https://test.com"
                      },
                      "routingRule": "routing-rule",
                      "storage": {
                        "blobStoreName": "blob-store-name",
                        "strictContentTypeValidation": true
                      }
                    }
                  },
                  "nexusRef": {
                    "name": "nexus-sample"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrole",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrole-sample"
                },
                "spec": {
                  "description": "test-role",
                  "id": "test-role",
                  "name": "test-role",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "privileges": [
                    "nx-blobstores-all"
                  ]
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexususer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexususer",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexususer-sample"
                },
                "spec": {
                  "email": "test@gmail.com",
                  "firstName": "John1",
                  "id": "test-user",
                  "lastName": "Doe",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "roles": [
                    "nx-admin"
                  ],
                  "secret": "$user-secret:password"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/nexus-operator:3.0.0
          createdAt: "2023-12-11T12:54:50Z"
          description: An Operator for managing Nexus entities.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-nexus-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus is the Schema for the nexus API.
            displayName: Nexus
            kind: Nexus
            name: nexuses.edp.epam.com
            version: v1alpha1
          - description: NexusRepository is the Schema for the nexusrepositories API.
            displayName: Nexus Repository
            kind: NexusRepository
            name: nexusrepositories.edp.epam.com
            version: v1alpha1
          - description: NexusRole is the Schema for the nexusroles API.
            displayName: Nexus Role
            kind: NexusRole
            name: nexusroles.edp.epam.com
            version: v1alpha1
          - description: NexusUser is the Schema for the nexususers API.
            displayName: Nexus User
            kind: NexusUser
            name: nexususers.edp.epam.com
            version: v1alpha1
        description: |
          Nexus Operator is an operator that is responsible for establishing
          a connection to provided NexusRepository Server, reconciling Nexus entities (groups,
          users, repositories, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Nexus Repository.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  nexus-secret
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Nexus` with Nexus instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1alpha1
              kind: Nexus
              metadata:
                name: nexus-sample
              spec:
                secret: nexus-secret                # Secret name
                url: https://nexus.example.com      # Nexus URL
              ```

              Wait for the `.status` field with  `status.connected: true`
        displayName: EDP Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - edp
        - nexus
        - operator
        - artifact
        - storage
        links:
        - name: Nexus Operator
          url: https://github.com/epam/edp-nexus-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/nexus-operator:3.0.0
        version: 3.0.0
      entries:
      - name: nexus-operator.v3.0.0
        version: 3.0.0
      name: alpha
    - currentCSV: nexus-operator.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexus-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexus",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexus-sample"
                },
                "spec": {
                  "secret": "nexus-secret",
                  "url": "https://nexus-example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusBlobStore",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusblobstore-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusblobstore",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusblobstore-sample"
                },
                "spec": {
                  "file": {
                    "path": "storage-blobstore/nexusblobstore-sample"
                  },
                  "name": "nexusblobstore-sample",
                  "nexusRef": {
                    "name": "nexus-sample"
                  },
                  "softQuota": {
                    "limit": 1000,
                    "type": "spaceUsedQuota"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusCleanupPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexuscleanuppolicy-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexuscleanuppolicy",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexuscleanuppolicy-sample"
                },
                "spec": {
                  "criteria": {
                    "lastBlobUpdated": 30
                  },
                  "description": "Cleanup policy for go",
                  "format": "go",
                  "name": "go-cleanup-policy",
                  "nexusRef": {
                    "name": "nexus-sample"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRepository",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrepository-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrepository",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrepository-sample"
                },
                "spec": {
                  "go": {
                    "proxy": {
                      "cleanup": {
                        "policyNames": [
                          "cleanup-policy"
                        ]
                      },
                      "httpClient": {
                        "autoBlock": true,
                        "blocked": true,
                        "connection": {
                          "retries": 1440,
                          "timeout": 1440,
                          "userAgentSuffix": "user-agent-suffix"
                        }
                      },
                      "name": "go-proxy",
                      "negativeCache": {
                        "enabled": true,
                        "timeToLive": 1440
                      },
                      "online": true,
                      "proxy": {
                        "contentMaxAge": 1440,
                        "metadataMaxAge": 1440,
                        "remoteUrl": "https://test.com"
                      },
                      "routingRule": "routing-rule",
                      "storage": {
                        "blobStoreName": "blob-store-name",
                        "strictContentTypeValidation": true
                      }
                    }
                  },
                  "nexusRef": {
                    "name": "nexus-sample"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrole",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrole-sample"
                },
                "spec": {
                  "description": "test-role",
                  "id": "test-role",
                  "name": "test-role",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "privileges": [
                    "nx-blobstores-all"
                  ]
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusScript",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusscript-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusscript",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusscript-sample"
                },
                "spec": {
                  "content": "security.setAnonymousAccess(Boolean.valueOf(args))\n",
                  "name": "anonymous",
                  "nexusRef": {
                    "name": "nexus-sample"
                  },
                  "payload": "true"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexususer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexususer",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexususer-sample"
                },
                "spec": {
                  "email": "test@gmail.com",
                  "firstName": "John1",
                  "id": "test-user",
                  "lastName": "Doe",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "roles": [
                    "nx-admin"
                  ],
                  "secret": "$user-secret:password"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/nexus-operator:3.5.0
          createdAt: "2025-04-22T11:21:24Z"
          description: An Operator for managing Nexus entities.
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/epam/edp-nexus-operator
          support: KubeRocketCI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NexusBlobStore is the Schema for the nexusblobstores API.
            displayName: Nexus Blob Store
            kind: NexusBlobStore
            name: nexusblobstores.edp.epam.com
            version: v1alpha1
          - description: NexusCleanupPolicy is the Schema for the cleanuppolicies
              API.
            displayName: Nexus Cleanup Policy
            kind: NexusCleanupPolicy
            name: nexuscleanuppolicies.edp.epam.com
            version: v1alpha1
          - description: Nexus is the Schema for the nexus API.
            displayName: Nexus
            kind: Nexus
            name: nexuses.edp.epam.com
            version: v1alpha1
          - description: NexusRepository is the Schema for the nexusrepositories API.
            displayName: Nexus Repository
            kind: NexusRepository
            name: nexusrepositories.edp.epam.com
            version: v1alpha1
          - description: NexusRole is the Schema for the nexusroles API.
            displayName: Nexus Role
            kind: NexusRole
            name: nexusroles.edp.epam.com
            version: v1alpha1
          - description: NexusScript is the Schema for the nexusscripts API.
            displayName: Nexus Script
            kind: NexusScript
            name: nexusscripts.edp.epam.com
            version: v1alpha1
          - description: NexusUser is the Schema for the nexususers API.
            displayName: Nexus User
            kind: NexusUser
            name: nexususers.edp.epam.com
            version: v1alpha1
        description: |
          Nexus Operator is an operator that is responsible for establishing
          a connection to provided NexusRepository Server, reconciling Nexus entities (groups,
          users, repositories, etc) according to the created CRs.

          ## Quick Start

          1. Login into Nexus and create user. Attach permissions to user such as scripts, rules, blobs etc. Insert user credentials into Kubernetes secret.

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name: nexus-admin-password
              data:
                password: cGFzcw==  # base64-encoded value of "pass"
                user:     dXNlcg==  # base64-encoded value of "user"
              ```

          2. Create Custom Resource `kind: Nexus` with Nexus instance URL and secret created on the previous step:

              ```yaml
              apiVersion: edp.epam.com/v1alpha1
              kind: Nexus
              metadata:
                name: nexus
              spec:
                secret: nexus-admin-password
                url: http://nexus.example.com
              ```

              Wait for the `.status` field with  `status.connected: true`

          3. Create Role using Custom Resources NexusRole:

              ```yaml
              apiVersion: edp.epam.com/v1alpha1
              kind: NexusRole
              metadata:
                name: edp-admin
              spec:
                description: Read and write access to all repos and scripts
                id: edp-admin
                name: edp-admin
                nexusRef:
                  kind: Nexus
                  name: nexus
                privileges:
                  - nx-apikey-all
                  - nx-repository-view-*-*-add
                  - nx-repository-view-*-*-browse
                  - nx-repository-view-*-*-edit
                  - nx-repository-view-*-*-read
                  - nx-script-*-add
                  - nx-script-*-delete
                  - nx-script-*-run
                  - nx-search-read
              ```
        displayName: EDP Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - artifact
        - edp
        - kuberocketci
        - nexus
        - operator
        - storage
        links:
        - name: Nexus Operator
          url: https://github.com/epam/edp-nexus-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: KubeRocketCI
          url: https://docs.kuberocketci.io
        relatedImages:
        - docker.io/epamedp/nexus-operator:3.5.0
        version: 3.5.0
      entries:
      - name: nexus-operator.v3.5.0
        version: 3.5.0
      - name: nexus-operator.v3.4.0
        version: 3.4.0
      - name: nexus-operator.v3.3.0
        version: 3.3.0
      name: stable
    defaultChannel: stable
    packageName: nexus-operator
    provider:
      name: KubeRocketCI
      url: https://docs.kuberocketci.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp, Inc.
      provider-url: https://www.netapp.com/
    name: trident-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trident-operator.v25.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "trident.netapp.io/v1",
                "kind": "TridentOrchestrator",
                "metadata": {
                  "name": "trident"
                },
                "spec": {
                  "IPv6": false,
                  "debug": true,
                  "nodePrep": [],
                  "imagePullSecrets": [],
                  "imageRegistry": "",
                  "namespace": "trident",
                  "silenceAutosupport": false
                }
              }
            ]
          categories: Storage
          containerImage: docker.io/netapp/trident-operator@sha256:c3448b6847550f45849a5cdbb9e0cedb35290d21131bc513d833d95eb42a43e0
          createdAt: "2025-09-03"
          description: Trident Operator, to manage NetApp Trident installations
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          operators.openshift.io/valid-subscription: No subscriptions are required
          repository: https://github.com/netapp/trident
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Used to deploy NetApp Trident.
            displayName: Trident Orchestrator
            kind: TridentOrchestrator
            name: tridentorchestrators.trident.netapp.io
            version: v1
          - description: Automates AWS FSxN backend configuration
            displayName: Trident Configurator
            kind: TridentConfigurator
            name: tridentconfigurators.trident.netapp.io
            version: v1
        description: |
          NetApp Trident is an open source storage provisioner and orchestrator maintained by NetApp. It enables you to create storage volumes for containerized applications managed by Docker and Kubernetes. For full release information, including patch release changes, see https://docs.netapp.com/us-en/trident/trident-rn.html.
        displayName: NetApp Trident
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - NetApp
        - ONTAP
        - Trident
        - storage
        - performance
        - open source
        - Astra
        - Cloud Native
        - AWS
        - Azure
        - GCP
        - Cloud Volumes
        links:
        - name: GitHub Repository
          url: https://github.com/NetApp/trident
        - name: Trident documentation
          url: https://docs.netapp.com/us-en/trident/index.html
        - name: Support policy
          url: https://mysupport.netapp.com/site/info/version-support
        - name: Release Notes
          url: https://docs.netapp.com/us-en/trident/trident-rn.html
        maintainers:
        - email: ng-trident-openshift-operator@netapp.com
          name: NetApp
        provider:
          name: NetApp, Inc.
          url: https://www.netapp.com/
        relatedImages:
        - docker.io/netapp/trident-operator@sha256:c3448b6847550f45849a5cdbb9e0cedb35290d21131bc513d833d95eb42a43e0
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:b88958c38883f5b1e3d307da90aac34a5ceafdb6bd3d989b062cb24ecafd1fbd
        - docker.io/netapp/trident-autosupport@sha256:21b3afbebfac3cc28c4eb319c2d385187bd8ebb728d8cfd82dbc437cc9b83b33
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:bdee8df520467b30312e7b14d7b6ef3963557ccd38269913de3fa51bb1bceec9
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:145ab76f9606e9d51dcd1e4e159c67021ecf53c24d8d80ed7bbeb2cc7fb5c42b
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6781d64bfb01e252595130eec814b9d2fe6426720a003cc61e82e9963767bfd4
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:b48bd5bc612c1bbe73d203c70e6caaae2b5360bd31eba20eb06cbd2102cc34fc
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:0ef3bf2a4dcbb5aa829ce9cc5f12efaccaf99b246b4b15abbb66b0135bec4cab
        - docker.io/netapp/trident@sha256:7925717979a84f93b62a2592682cde3443d6350ec4e88f8fd1a7d3ea812e8900
        version: 25.6.2
      entries:
      - name: trident-operator.v25.6.2
        version: 25.6.2
      - name: trident-operator.v25.6.1
        version: 25.6.1
      - name: trident-operator.v25.6.0
        version: 25.6.0
      - name: trident-operator.v25.2.1
        version: 25.2.1
      - name: trident-operator.v25.2.0
        version: 25.2.0
      name: stable
    defaultChannel: stable
    packageName: trident-operator
    provider:
      name: NetApp, Inc.
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: xspc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: xspc-operator.v1.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.hpe.com/v1",
                "kind": "XSPC",
                "metadata": {
                  "name": "xspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.connect.redhat.com/hpestorage/xspc-csi-driver-operator@sha256:8b8724998a8301057a3c4187c7b94bf4484aeb807065a12f84142cf844ea5534
          createdAt: "2024-12-17T03:53:30Z"
          description: An operator for managing HPE XP Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XSPC is the Schema for the xspcs API
            displayName: XSPC
            kind: XSPC
            name: xspcs.csi.hpe.com
            version: v1
        description: |-
          ## About
          HPE XP Storage Plug-in for Containers (XSPC) is a plugin that integrates HPE XP storage into Kubernetes based clusters.
          XSPC provides persistent volume provisioning capabilities from HPE's block storage arrays.
        displayName: HPE XP Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - XP7
        - XP8
        - HPE
        - Storage
        - Container Storage Interface
        - CSI
        links:
        - name: Knowledge Base
          url: https://www.hpe.com/us/en/storage/enterprise-xp-storage.html
        maintainers:
        - email: hpexpci@hpe.com
          name: HPE
        maturity: stable
        provider:
          name: HPE
        relatedImages:
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.connect.redhat.com/hpestorage/xspc-csi-driver@sha256:bdeac71c0b54c2779910a8d8faea03c1122000127704f17ceadba4938b685353
        - registry.connect.redhat.com/hpestorage/xspc-csi-driver-operator@sha256:8b8724998a8301057a3c4187c7b94bf4484aeb807065a12f84142cf844ea5534
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        version: 1.15.0
      entries:
      - name: xspc-operator.v1.15.0
        version: 1.15.0
      name: stable
    defaultChannel: stable
    packageName: xspc-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Entando, Inc
      provider-url: ""
    name: entando-k8s-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: entando-k8s-operator.v6.3.2-pr3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: beta
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - docker.io/entando/entando-k8s-service@sha256:775391f0005e64764288bfe007d9ea576f13618c02ea400d030a7502cfb551bd
        - docker.io/entando/entando-k8s-database-service-controller@sha256:58ca4ad2fca83a715aba9b68609f9056f633d556eb9845b3aff5c05769369539
        - docker.io/entando/app-builder@sha256:1ca90c09302b9cadf8950278d92ad4d7f3a14f02dea8704320794aa77ada8631
        - docker.io/entando/entando-k8s-app-controller@sha256:c68d3b129fc625f5d1ee10ed2a107e28c145049d4b513dcdb2c736555025e339
        - docker.io/entando/entando-k8s-plugin-controller@sha256:a20c753dda86e0f54f210e446b404733045ba426fc7186dc3936daed8224fbcb
        - docker.io/entando/entando-component-manager@sha256:668191fc86b090452baaba18ec7bbc3d07df5a99750e0c5b98759599d565cfbe
        - docker.io/entando/entando-redhat-sso@sha256:3692adb3694b3bf278bed7a678c9d0eaa0aba8e3e3ed4117ca04156525bcac72
        - registry.redhat.io/rhel8/mysql-80@sha256:570bce7dc649ac7ac85e7ecb196aabc281c77a94590aa7df7d9746869ce757ea
        - docker.io/entando/entando-k8s-dbjob@sha256:8ae47dbf93201965106663299f06daad544982fa244d4a0f365a1b717727e5ed
        - docker.io/entando/entando-k8s-composite-app-controller@sha256:4bc6797bb41b3e407b64530c3cff7a78f3fc2c22863caad9c606ccd650310b61
        - docker.io/entando/entando-k8s-app-plugin-link-controller@sha256:aa7ba0861705213434e610530c5bf5da1bdb33f1834d3c346174f5a28145c0bd
        - docker.io/entando/entando-de-app-eap@sha256:1f0e35d7f7ee8123721cf8e80ac96ac2f594f21a2b68cf9e720917d090c8095e
        - docker.io/entando/entando-k8s-keycloak-controller@sha256:f1ed42001260802353197433cea7df4c2d0e5fd9bcd926bb5ae7c654dfad2ffe
        - registry.redhat.io/rhel8/postgresql-12@sha256:f4e5c728b644bf1888ec8086424852ed74b5596a511be29e636fb10218fc9b6f
        - entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        - docker.io/entando/entando-k8s-cluster-infrastructure-controller@sha256:86dbea6696349d7b2393e12bacef4d1acec2dd3f8551e397fa26c92706a18dd3
        version: 6.3.2-pr3
      entries:
      - name: entando-k8s-operator.v6.3.2-pr3
        version: 6.3.2-pr3
      name: alpha
    - currentCSV: entando-k8s-operator.v6.3.2-pr3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: beta
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - docker.io/entando/entando-redhat-sso@sha256:3692adb3694b3bf278bed7a678c9d0eaa0aba8e3e3ed4117ca04156525bcac72
        - entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        - docker.io/entando/entando-de-app-eap@sha256:1f0e35d7f7ee8123721cf8e80ac96ac2f594f21a2b68cf9e720917d090c8095e
        - docker.io/entando/entando-k8s-composite-app-controller@sha256:4bc6797bb41b3e407b64530c3cff7a78f3fc2c22863caad9c606ccd650310b61
        - docker.io/entando/entando-k8s-app-controller@sha256:c68d3b129fc625f5d1ee10ed2a107e28c145049d4b513dcdb2c736555025e339
        - docker.io/entando/entando-k8s-plugin-controller@sha256:a20c753dda86e0f54f210e446b404733045ba426fc7186dc3936daed8224fbcb
        - docker.io/entando/entando-component-manager@sha256:668191fc86b090452baaba18ec7bbc3d07df5a99750e0c5b98759599d565cfbe
        - docker.io/entando/entando-k8s-app-plugin-link-controller@sha256:aa7ba0861705213434e610530c5bf5da1bdb33f1834d3c346174f5a28145c0bd
        - docker.io/entando/entando-k8s-database-service-controller@sha256:58ca4ad2fca83a715aba9b68609f9056f633d556eb9845b3aff5c05769369539
        - docker.io/entando/app-builder@sha256:1ca90c09302b9cadf8950278d92ad4d7f3a14f02dea8704320794aa77ada8631
        - docker.io/entando/entando-k8s-keycloak-controller@sha256:f1ed42001260802353197433cea7df4c2d0e5fd9bcd926bb5ae7c654dfad2ffe
        - registry.redhat.io/rhel8/mysql-80@sha256:570bce7dc649ac7ac85e7ecb196aabc281c77a94590aa7df7d9746869ce757ea
        - registry.redhat.io/rhel8/postgresql-12@sha256:f4e5c728b644bf1888ec8086424852ed74b5596a511be29e636fb10218fc9b6f
        - docker.io/entando/entando-k8s-cluster-infrastructure-controller@sha256:86dbea6696349d7b2393e12bacef4d1acec2dd3f8551e397fa26c92706a18dd3
        - docker.io/entando/entando-k8s-dbjob@sha256:8ae47dbf93201965106663299f06daad544982fa244d4a0f365a1b717727e5ed
        - docker.io/entando/entando-k8s-service@sha256:775391f0005e64764288bfe007d9ea576f13618c02ea400d030a7502cfb551bd
        version: 6.3.2-pr3
      entries:
      - name: entando-k8s-operator.v6.3.2-pr3
        version: 6.3.2-pr3
      name: beta
    - currentCSV: entando-k8s-operator.v6.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: registry.hub.docker.com/entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - registry.hub.docker.com/entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        version: 6.3.2
      entries:
      - name: entando-k8s-operator.v6.3.2
        version: 6.3.2
      name: stable
    defaultChannel: stable
    packageName: entando-k8s-operator-rhmp
    provider:
      name: Entando, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: submariner.io
      provider-url: ""
    name: submariner
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: submariner.v0.17.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:067ba6e58a6fe150a212881a29c90aa7cbd22f53a6877eebc413072f73f9179a
          createdAt: "2025-06-18 01:35:57"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.16.0 <0.17.6'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.17 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:d58214791a60cd9e4366c6ff534584eedbd37386da71b868867e5e212df270f9
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:a565d8c094f3755b6c9303c5b8ae03a5883796e0fea754b0202e5792f1e472c3
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:067ba6e58a6fe150a212881a29c90aa7cbd22f53a6877eebc413072f73f9179a
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:b76998e8bbef1d06aa8dbc49af4f42a3c5603d480a67e46baf11168514c7759a
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:56848aefe1d156147ab73e7353017b2663f63e2e01fe43ede57abc72cfc0c50f
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:e6dbc4bf3e8b1a3c1a7ed10eb463d134f9ed28a03c7bb523c82c5431c5c9ae90
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:5510cb1bb3e0ca5f2365f4a5fc147ca693d3c220f1a8e7835f8b13b66b957ae8
        version: 0.17.6
      entries:
      - name: submariner.v0.17.6
        version: 0.17.6
      - name: submariner.v0.17.5
        version: 0.17.5
      - name: submariner.v0.17.4
        version: 0.17.4
      - name: submariner.v0.17.2
        version: 0.17.2
      - name: submariner.v0.17.2-0.1725901206.p
        version: 0.17.2+0.1725901206.p
      - name: submariner.v0.17.1
        version: 0.17.1
      - name: submariner.v0.17.0
        version: 0.17.0
      name: stable-0.17
    - currentCSV: submariner.v0.18.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:d8d7f0147c7b65b41903e101240905703992a7bbadb33e302aa86b23000d2095
          createdAt: "2025-05-07 20:54:56"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.17.0 <0.18.5'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.18 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:e5c16f0d720e15d54781870a5e63c85540256355009584f7d1bd03ef20a2eae4
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:00072fea3fdb5e294a01da0ee44165fc52219a965a27bdf7d792276f6c3a36c3
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:e8366433d1f50e2062b86d7f749f17cd6566197560256e77f57b2c165f277733
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:a07ef1bfec7549b32ad94f2c44a168860d5f3b46fd13593c068c58bb137c77fc
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:d8d7f0147c7b65b41903e101240905703992a7bbadb33e302aa86b23000d2095
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:27f18015b8896b04457c356d03276d50c34b8c1d84142ec07342ff2c1c4e81fa
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:e913be7a440b6cc0c1dddbda93a5979fe28b99e63f12a7949966b69126fcee90
        version: 0.18.5
      entries:
      - name: submariner.v0.18.5
        version: 0.18.5
      - name: submariner.v0.18.4
        version: 0.18.4
      - name: submariner.v0.18.3
        version: 0.18.3
      - name: submariner.v0.18.2
        version: 0.18.2
      - name: submariner.v0.18.0
        version: 0.18.0
      name: stable-0.18
    - currentCSV: submariner.v0.19.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:4afaf6689e2225c5c7f6fbe002202a1cdf9444e23fd210cc2d86da71c9f78626
          createdAt: "2025-04-22 21:12:45"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.18.0 <0.19.4'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.19 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:d9d4d03013d65d11a4e8ecda411015f359e4215a3a53059a37a810acf2e3ea6c
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:4afaf6689e2225c5c7f6fbe002202a1cdf9444e23fd210cc2d86da71c9f78626
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:5b83c5889aa0d8f7838dd5afb8c901dfb0ce0fbe01175e8468ec50ddedef9223
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:70008af49a9d1ee516204715f66bd415c9caa7462f6b178dc1dc91e513605fcd
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:ef73b8964cd66ffb496191056594e0886afa9d17c03976b4159413c89ed569bf
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:5938d67a0a2653e92f0995440700512992ed82269820679f68483dcc1b2d98d9
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:1bbb624bbefbb75407071449d0144258732e219e3a7fc67ac709c3bcee560576
        version: 0.19.4
      entries:
      - name: submariner.v0.19.4
        version: 0.19.4
      - name: submariner.v0.19.3
        version: 0.19.3
      - name: submariner.v0.19.2
        version: 0.19.2
      - name: submariner.v0.19.0
        version: 0.19.0
      name: stable-0.19
    - currentCSV: submariner.v0.20.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "quay.io/submariner",
                  "version": "0.20.2"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "quay.io/submariner",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "0.20.2"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: $(SUBMARINER_OPERATOR_IMAGE)
          createdAt: "2025-09-30 14:21:54"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/submariner-io/submariner-operator
          support: Submariner
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.20 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: submariner.io
        relatedImages:
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:fd6c46b4a0570b5e7b6457c465e5a1b843ba6f9c209c1856dbebfee1068e35ea
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:70d563a6e6526b9a146a2973210b2abdae09d5acab403fa3f0b465ed5e3c9b11
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:77e0da2db6d5bb2831e6c51214e08685c9f7f3b8b1e6ac53b4fd1418644ad6a1
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:0eef6521c96b29722076c1e220770124c8ed214379abc4654d9567cff8b52d04
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:8ea02aba8f9dd901f1f0ccff97ec8d157bebe5a90a6c924affe5e01cf3454e72
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:ed521b241da60efabfe2e2b80c51639d1a82a43945945879decbfb192bbc6d23
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:4a4ac198e9d0f4a4aaed8f0ec0e1d79bd8fa712b671ee697f7ff9d5616a02d84
        version: 0.20.2
      entries:
      - name: submariner.v0.20.2
        version: 0.20.2
      - name: submariner.v0.20.1
        version: 0.20.1
      - name: submariner.v0.20.0
        version: 0.20.0
      name: stable-0.20
    - currentCSV: submariner.v0.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "quay.io/submariner",
                  "version": "v0.21.0"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "quay.io/submariner",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v0.21.0"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: $(SUBMARINER_OPERATOR_IMAGE)
          createdAt: "2025-08-13 21:57:14"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/submariner-io/submariner-operator
          support: Submariner
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.21 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: submariner.io
        relatedImages:
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:cd851eab4a2214ea7a699bd01d2abac1322d01562f27b35f58f3c0b8716f88fb
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:bcdeaa67664568c5e2a6928236ff69b175ce65ddf7a58c1664d091b926000052
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:b0ad338aed82367048cf024767b0b09e6bb09bbf7dd5128e6711e1feec6f696e
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:65765c7f08c5454b557fbda91c49efbd2cc0d4d973cf8acc7ff40ec8a381839d
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:f5909b1802986622ab169a82d5dd7877af72660903e961b3097cc29e30eb9799
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:b307f748125fd459c3474fdab645688675a06515a9924b47cedba80e038458db
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:0b74f058400f24a63cf9b57a388914cc6dda38e03d639196871f3fd76fb7f2ae
        version: 0.21.0
      entries:
      - name: submariner.v0.21.0
        version: 0.21.0
      name: stable-0.21
    defaultChannel: stable-0.21
    packageName: submariner
    provider:
      name: submariner.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app: kube-green
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: kube-green Community
      provider-url: https://kube-green.dev
    name: kube-green
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kube-green.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kube-green.com/v1alpha1",
                "kind": "SleepInfo",
                "metadata": {
                  "labels": {
                    "app": "kube-green"
                  },
                  "name": "sleepinfo-sample"
                },
                "spec": {
                  "excludeRef": [
                    {
                      "apiVersion": "apps/v1",
                      "kind": "Deployment",
                      "name": "api-gateway"
                    }
                  ],
                  "sleepAt": "20:00",
                  "suspendCronJobs": true,
                  "timeZone": "Europe/Rome",
                  "wakeUpAt": "08:00",
                  "weekdays": "1-5"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          containerImage: docker.io/kubegreen/kube-green:0.7.1
          createdAt: "2025-06-29T15:45:17Z"
          description: Suspend your pods when no-one's using them to save energy and
            restore when necessary
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kube-green/kube-green
          support: kube-green Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SleepInfo is the Schema for the sleepinfos API
            displayName: SleepInfo
            kind: SleepInfo
            name: sleepinfos.kube-green.com
            version: v1alpha1
        description: |
          ## About this Operator

          [kube-green](https://kube-green.dev) shuts down (some of) your resources when you don't need them, reducing the consumption of resources of the cluster.

          Install this operator and apply the SleepInfo CRD to stop your dev namespaces from running during weekends, nights or when it is not necessary.
          See the [docs](https://kube-green.dev/docs/getting-started/) for more information.
          This operator is responsible to suspend and restart pods at specific times, defined by the SleepInfo CRD.

          An example of SleepInfo:

          ```yaml
          apiVersion: kube-green.com/v1alpha1
          kind: SleepInfo
          metadata:
            name: working-hours
          spec:
            weekdays: "1-5"
            sleepAt: "20:00"
            wakeUpAt: "08:00"
            timeZone: "Europe/Rome"
          ```

          This configuration stop pods everyday from monday to friday at 20:00 and restore them at 8:00.
          In this way, your cluster nodes could downscale, producing less CO2 and saving money.
        displayName: kube-green
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - green
        - green software
        - resources
        - downscale
        - autoscale
        - climate change
        links:
        - name: Homepage
          url: https://kube-green.dev
        - name: Source Code
          url: https://github.com/kube-green/kube-green
        maintainers:
        - email: bianchidavide12+kubegreen@gmail.com
          name: Davide
        maturity: alpha
        minKubeVersion: 1.27.0
        provider:
          name: kube-green Community
          url: https://kube-green.dev
        relatedImages:
        - docker.io/kubegreen/kube-green:0.7.1
        version: 0.7.1
      entries:
      - name: kube-green.v0.7.1
        version: 0.7.1
      - name: kube-green.v0.7.0
        version: 0.7.0
      - name: kube-green.v0.6.0
        version: 0.6.0
      - name: kube-green.v0.5.2
        version: 0.5.2
      - name: kube-green.v0.5.1
        version: 0.5.1
      - name: kube-green.v0.5.0
        version: 0.5.0
      - name: kube-green.v0.4.1
        version: 0.4.1
      - name: kube-green.v0.4.0
        version: 0.4.0
      - name: kube-green.v0.3.1
        version: 0.3.1
      - name: kube-green.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: kube-green
    provider:
      name: kube-green Community
      url: https://kube-green.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-wafv2-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-wafv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wafv2.services.k8s.aws/v1alpha1",
                "kind": "IPSet",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/wafv2-controller:1.1.1
          createdAt: "2025-09-20T21:41:21Z"
          description: AWS WAFV2 controller is a service controller for managing WAFV2
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IPSet represents the state of an AWS wafv2 IPSet resource.
            displayName: IPSet
            kind: IPSet
            name: ipsets.wafv2.services.k8s.aws
            version: v1alpha1
          - description: RuleGroup represents the state of an AWS wafv2 RuleGroup
              resource.
            displayName: RuleGroup
            kind: RuleGroup
            name: rulegroups.wafv2.services.k8s.aws
            version: v1alpha1
          - description: WebACL represents the state of an AWS wafv2 WebACL resource.
            displayName: WebACL
            kind: WebACL
            name: webacls.wafv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Web Application Firewall (Amazon WAFV2) resources in AWS from within your Kubernetes cluster.

          **About Amazon WAFV2**

          AWS WAF is a web application firewall that lets you monitor the HTTP and HTTPS requests that are forwarded to your protected web application resources. You can protect the following resource types:
          - Amazon CloudFront distribution
          - Amazon API Gateway REST API
          - Application Load Balancer
          - AWS AppSync GraphQL API
          - Amazon Cognito user pool
          - AWS App Runner service
          - AWS Verified Access instance

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon WAFV2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - wafv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon WAFV2 Developer Resources
          url: https://aws.amazon.com/waf/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: wafv2 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/wafv2-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-wafv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-wafv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-wafv2-controller.v1.0.14
        version: 1.0.14
      - name: ack-wafv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-wafv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-wafv2-controller.v1.0.10
        version: 1.0.10
      - name: ack-wafv2-controller.v1.0.9
        version: 1.0.9
      - name: ack-wafv2-controller.v1.0.8
        version: 1.0.8
      - name: ack-wafv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-wafv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-wafv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-wafv2-controller.v1.0.1
        version: 1.0.1
      - name: ack-wafv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-wafv2-controller.v0.0.4
        version: 0.0.4
      - name: ack-wafv2-controller.v0.0.3
        version: 0.0.3
      - name: ack-wafv2-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-wafv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/in-en/products/cloud-pak-for-data
    name: ibm-zen-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-zen-operator.v6.1.3
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.3 <6.1.3'
          operators.operatorframework.io/builder: operator-sdk-v1.6.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: represents zen instance
            kind: ZenService
            name: zenservices.zen.cpd.ibm.com
            version: v1
          - description: represents zen extension instance
            kind: ZenExtension
            name: zenextensions.zen.cpd.ibm.com
            version: v1
        description: |
          The Zen operator is a component in IBM® Cloud Pak foundational services. It serves a framework for user access management, defining user roles and for powering persona driven user experiences for Cloud Paks and Service offerings
        displayName: IBM Zen Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cloud Pak
        - IBM
        - Access management
        - Platform UI
        - Zen framework
        maintainers:
        - email: cp4d@ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/in-en/products/cloud-pak-for-data
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:443b51f8b10acc85bdefde7193e0f45b1c423fa7fbdcaa28342805815c43db3d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:3d5c6573faf4ae00cbb21cbfe9b2ad5cd8ffb658df0c5ba09bd98b3bf54f9979
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:fa01c7d11c23a47bb84375389e69619455c2879b5ed6bb5b9b753b945dcb02c3
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:45194b56f9fcb3af83158eb0297248364db40e8164fb4fe4f5656c58aeac3fa5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:a99b1f5a4eb0306b83562de0184fc826c34835ffbf46d929ceb9e70e0d325361
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:5f48081ffa453ec740c2562c24603d78b481b5e236d6bcc73e96203cabe153f2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:0522fc5482404fa9be179c7225a5692c8d71d755fda0b2d3928e452adcf030fb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:45ca5370b726e31c55f9d7d249374189cb48dc05ab89d675c35225d53b90e7c9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:aaee5240471872a8f1a77d4440a696d45041c5ce06849d516daccb90f5341b6d
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:abd96661823553d623eb0ea1fac8271d64391243dd0f13eb4a4fcc858dfe10a5
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:82afac013ef9b6b9d0eacb4575c979b65abd6f2821971182b90c71bd488ef366
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:b87fa8374976ccd3dad90ddbf6b65cd6ae5a028ba896562f284f089380ebe44a
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:24dff64a80b910f0b4c5442d4f611f9df7602a16c5231eff2c9d08f07b2a9f9a
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:ad027c91c719fd1d931cbac8c76735474f676b228f9ab23d1f243201246b3238
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:c6862158a2a5ae0a4e4206bfdc7f73e1c2359544966c72cf4fc3a617acb13697
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:10ede3c4e3c7e0cbc26e7b6e797cad35e39face1a1b9d4a4e2440ab0feeeb814
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:deb66a64288a370c5da5c7cc0ba4e2ba4ab816d7c5a080a0c505b05f55fa6072
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:7dbff355db7739152961cb6a97887d2e43bc960ac58837c126e03ed1a4480a3a
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:a72c2b70a9f0a93047d47247cddc1616138cbc87337289ddb208565af6e4369d
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:183ee860bf48076ea155db08032eed7004c1204a363625f54218734b00175441
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:525924934f74b28783dacf630a6401ccd4c88d253aafa2d312b7993d8b9a14c1
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:a7c050ae29cb93754f59d8f83676275ee9a91ae9a9956395b97abf8c5e20e96e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8da8b240176c00f8f6eadf32e5614c663cc1d56d546c73996cca614b715acf4a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:f06a1f50cd18d6f48762fdd0394307b44c0573d0cc91f53724c5b3c8c5e2b8e4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:c4ff5dae4560e5a943c7218df3a5aa9ca1a9ee2d1f7dfac73db05892725c34a9
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-bundle@sha256:552fc53664722903eccaa7d3782a3e84413509fd357725fe0b0f4a077c48ab2c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:719b0871f0d32da894ec80dc2a23483623b7802a5c6f935cf3672468f99faa40
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:c8baecad9c97ddbbb3f9f02c90abbf483d46974b7bc9605622a302cbeba1de74
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:b5a8fb14c3bb7c89ff970ca85f06dfcdca832e2b45ad86e8bb1b6c0a0e7e0764
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:8444703b2a7822f82141afcf4f36bf255650d755321b9e9fe2cbd1bdea41d5c7
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:e9173d81ef2a75839c0e1d677b72bb00b8e6d7ef0511afab103f08a884a03738
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:a8d2e42757c6ccbfe4372465b1827c633aad012c5194bf8b08afcd5691caa6f5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:6bbcb64cf9adb4d3b6a4c259eb910e7da5c3c58732643e60d416399b2083cc0d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:ea57c7cb1ccd5bce17f6cb35e326ab974718be8cd19b0cd38ea8112e91c51939
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ee161954766211651d7cb13f3f5a3c025ac04094ba742be453ce789c6f0e24a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:ecbc65705de76804a3c7642339ecfb4cc0f5a61891058e07c20bf2b63f1b82b8
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:1df782d0d923544767696d719bd0ac10802b8e06508a104af5e023a519f32dee
        version: 6.1.3
      entries:
      - name: ibm-zen-operator.v6.1.3
        version: 6.1.3
      name: v6.1
    defaultChannel: v6.1
    packageName: ibm-zen-operator
    provider:
      name: IBM
      url: https://www.ibm.com/in-en/products/cloud-pak-for-data
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: hpe-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-csi-operator.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.hpe.com/v1",
                "kind": "HPECSIDriver",
                "metadata": {
                  "name": "hpecsidriver-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "csp": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "disable": {
                    "alletra6000": false,
                    "alletra9000": false,
                    "alletraStorageMP": false,
                    "b10000FileService": false,
                    "nimble": false,
                    "primera": false
                  },
                  "disableHostDeletion": false,
                  "disableNodeConfiguration": false,
                  "disableNodeConformance": false,
                  "disableNodeGetVolumeStats": false,
                  "disableNodeMonitor": false,
                  "disablePreInstallHooks": false,
                  "imagePullPolicy": "IfNotPresent",
                  "images": {
                    "b10000FileServiceCSP": "quay.io/hpestorage/alletrastoragemp-b10000-nfs-csp@sha256:6a1855a2ef1a83b886bb1f45b596e8e420e5e13c84bbf446c8c202eaf27a2013",
                    "csiAttacher": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                    "csiControllerDriver": "quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676",
                    "csiExtensions": "quay.io/hpestorage/csi-extensions@sha256:27074aaa8df3b867cfa3d12318612c45f8e541d750c27db5e080734592160c89",
                    "csiNodeDriver": "quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676",
                    "csiNodeDriverRegistrar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                    "csiProvisioner": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                    "csiResizer": "registry.k8s.io/sig-storage/csi-resizer@sha256:8ddd178ba5d08973f1607f9b84619b58320948de494b31c9d7cd5375b316d6d4",
                    "csiSnapshotter": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769",
                    "csiVolumeGroupProvisioner": "quay.io/hpestorage/volume-group-provisioner@sha256:0a5f39f52113b963978a695bb8e5a5a3c4d387a02eb4d4a9647b4bc804de7399",
                    "csiVolumeGroupSnapshotter": "quay.io/hpestorage/volume-group-snapshotter@sha256:177d9abf8fbd3ac26ca4c12ca4b4adde20ebab308c79f10b04632f0171746fa8",
                    "csiVolumeMutator": "quay.io/hpestorage/volume-mutator@sha256:bf6beac1af5edcd3c9774f0dcc9579b79ef9cc8e4b74a4238082ab9aa443b5ed",
                    "nfsProvisioner": "quay.io/hpestorage/nfs-provisioner@sha256:5bf5952453ecca28e26724df85ebd5c8f0b7479ae4cb683d4a09c415da00e002",
                    "nimbleCSP": "quay.io/hpestorage/alletra-6000-and-nimble-csp@sha256:5480ffdebd74fdc547bc4a11b4434abac7bbfee3cee349f1ac893fbad84bb8d1",
                    "primera3parCSP": "quay.io/hpestorage/alletra-9000-primera-and-3par-csp@sha256:8af6e87075c47ba8d9714d55848ab2a0fef62b0768681e30d9e63ed7f396b0bf"
                  },
                  "iscsi": {
                    "chapSecretName": ""
                  },
                  "kubeletRootDir": "/var/lib/kubelet",
                  "logLevel": "info",
                  "maxVolumesPerNode": 100,
                  "node": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/hpestorage/csi-driver-operator@sha256:29e92f68cf0d7691a6c56876e7dde0d69a8cb3ff6ae6c9371c604e61cba16898
          createdAt: "2025-09-25T16:28:16Z"
          description: A Container Storage Interface (CSI) driver for HPE storage
            platforms. The CSI driver allows you to use HPE storage with your preferred
            container orchestrator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <3.0.1'
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/hpe-storage/co-deployments
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE CSI Driver for Kubernetes
            displayName: HPECSIDriver
            kind: HPECSIDriver
            name: hpecsidrivers.storage.hpe.com
            version: v1
        description: |
          The HPE CSI Operator for OpenShift deploys the HPE CSI Driver on Kubernetes and OpenShift for dynamic provisioning of persistent volumes on HPE storage systems.
          The HPE CSI Driver for Kubernetes leverages HPE storage platforms to provide scalable and persistent storage for stateful applications.
          Please refer to our CSI driver [documentation](https://scod.hpedev.io/csi_driver/index.html) for supported HPE platforms and full list of supported CSI features.
          ## Installation
          Refer to the HPE CSI Operator for OpenShift [documentation](https://scod.hpedev.io/csi_driver/deployment.html#operator).
        displayName: HPE CSI Operator for OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE CSI
        - HPE Storage
        - CSI
        links:
        - name: Documentation
          url: https://scod.hpedev.io/csi_driver/index.html
        maintainers:
        - email: dev-hi-containers@hpe.com
          name: HPE Storage Containers Team
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: HPE
        relatedImages:
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - quay.io/hpestorage/volume-mutator@sha256:bf6beac1af5edcd3c9774f0dcc9579b79ef9cc8e4b74a4238082ab9aa443b5ed
        - registry.connect.redhat.com/hpestorage/csi-driver-operator@sha256:29e92f68cf0d7691a6c56876e7dde0d69a8cb3ff6ae6c9371c604e61cba16898
        - quay.io/hpestorage/alletrastoragemp-b10000-nfs-csp@sha256:6a1855a2ef1a83b886bb1f45b596e8e420e5e13c84bbf446c8c202eaf27a2013
        - quay.io/hpestorage/alletra-9000-primera-and-3par-csp@sha256:8af6e87075c47ba8d9714d55848ab2a0fef62b0768681e30d9e63ed7f396b0bf
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - quay.io/hpestorage/nfs-provisioner@sha256:5bf5952453ecca28e26724df85ebd5c8f0b7479ae4cb683d4a09c415da00e002
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8ddd178ba5d08973f1607f9b84619b58320948de494b31c9d7cd5375b316d6d4
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769
        - quay.io/hpestorage/volume-group-provisioner@sha256:0a5f39f52113b963978a695bb8e5a5a3c4d387a02eb4d4a9647b4bc804de7399
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - quay.io/hpestorage/volume-group-snapshotter@sha256:177d9abf8fbd3ac26ca4c12ca4b4adde20ebab308c79f10b04632f0171746fa8
        - quay.io/hpestorage/csi-extensions@sha256:27074aaa8df3b867cfa3d12318612c45f8e541d750c27db5e080734592160c89
        - quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676
        - quay.io/hpestorage/alletra-6000-and-nimble-csp@sha256:5480ffdebd74fdc547bc4a11b4434abac7bbfee3cee349f1ac893fbad84bb8d1
        version: 3.0.1
      entries:
      - name: hpe-csi-operator.v3.0.1
        version: 3.0.1
      name: stable
    defaultChannel: stable
    packageName: hpe-csi-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SmartX
      provider-url: https://www.smartx.com
    name: iomesh-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: iomesh-operator.0.10.1-rc4
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
          createdAt: 2021-08-17 00:00:00+00:00
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/iomesh-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/iomesh-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/iomesh/docs
          support: SmartX, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: iomesh
            displayName: iomesh
            kind: IOMeshCluster
            name: iomeshclusters.iomesh.com
            version: v1alpha1
        description: "IOMesh (https://www.iomesh.com/) is a cloud-native storage product
          specifically designed and developed for Kubernetes. It takes ZBS as its
          core, a “production-ready” distributed storage system developed independently
          by SmartX, to provide production-level high performance and reliable persistent
          storage for business-critical applications such as MySQL, Cassandra, MongoDB,
          etc., contributing to the containerized transformation of stateful applications.
          \     \n## Pre-Requisites\n* According to https://docs.iomesh.com/deploy/prerequisites,
          make sure that your openshift cluster meets the \"Installation Requirements\"
          and set up the worker nodes correctly according to the \"Setup Worker Node\"
          chapter.\n* Run IOMesh Operator pre-installation script in an environment
          where oc or kubectl can be used to access the openshift cluster, the script
          will install the dependencies of IOMesh Operator and specific IOMesh settings
          for the openshift cluster:\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-pre-install-openshift.sh
          | sh -\n```\n## Install IOMesh Operator and IOMesh Cluster\n* Click install
          button on the current web page to install IOMesh Operator\n* On the **Installed
          Operators** -> **IOMesh Operator** -> **Create instance** -> **YAML view**,
          fill in an IOMesh Custom Resources according https://iomesh.run/iomesh.yaml,
          change the `spec.*.dataCIDR` to your own data network CIDR:\n## Post Install\n*
          Run IOMesh Operator post-installation script in an environment where oc
          or kubectl can be used to access the openshift cluster, the script will
          install IOMesh CSI Driver\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-post-install-openshift.sh
          | sh -\n```\n## Getting Started\n* Try to quickstart setup and use IOMesh:
          https://docs.iomesh.com/deploy/setup-iomesh\n"
        displayName: IOMesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iomesh-operator
        links:
        - name: IOMesh Operator
          url: https://iomesh.com
        maintainers:
        - email: na.wang@smartx.com
          name: Na Wang
        maturity: alpha
        provider:
          name: SmartX
          url: https://www.smartx.com
        relatedImages:
        - iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
        version: 0.10.1-rc4
      entries:
      - name: iomesh-operator.0.10.1-rc4
        version: 0.10.1-rc4
      name: stable
    defaultChannel: stable
    packageName: iomesh-operator-rhmp
    provider:
      name: SmartX
      url: https://www.smartx.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: openstack-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openstack-operator.v1.0.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openstack.org/v1beta1",
                "kind": "OpenStack",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openstack-operator",
                    "app.kubernetes.io/instance": "openstack",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "openstack",
                    "app.kubernetes.io/part-of": "openstack-operator"
                  },
                  "name": "openstack",
                  "namespace": "openstack-operators"
                },
                "spec": {
                  "operatorOverrides": [
                    {
                      "controllerManager": {
                        "resources": {
                          "limits": {
                            "cpu": "600m",
                            "memory": "2Gi"
                          }
                        }
                      },
                      "name": "infra",
                      "replicas": 1
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-02T13:33:09Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.14
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.openstack.org/v1beta1","kind":"OpenStack","metadata":{"name":"openstack","namespace":"openstack-operators"},"spec":{}}'
          operatorframework.io/suggested-namespace: openstack-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenStack is the Schema for the openstacks API
            displayName: Open Stack
            kind: OpenStack
            name: openstacks.operator.openstack.org
            version: v1beta1
        description: Install and configure OpenStack
        displayName: OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        links:
        - name: Openstack Operator
          url: https://github.com/openstack-k8s-operators/
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/glance-rhel9-operator@sha256:ee22cf417b7771930a442c2302277e6d37f368f5ae551fee9a50dbbe41c8fe3a
        - registry.redhat.io/rhoso-operators/mysqld-exporter-rhel9@sha256:b49e966cdd588071e0e1f2791ec6c62310078a7d39ea77f83843d3708a07c2b5
        - registry.redhat.io/rhoso/openstack-swift-proxy-server-rhel9@sha256:858551fd1c519797cae0e146e9c05412ed6e66dac01d0e68e62dd2d4bd335391
        - registry.redhat.io/rhoso/openstack-netutils-rhel9@sha256:b02ab1617922a2560bf739172bc2ce0f6b1fb3ef55054bd17e4b16bdf9512b4b
        - registry.redhat.io/rhoso/openstack-cinder-backup-rhel9@sha256:910693233ee3a1f3832ebe744f67a5139249523456de4abbd731c9d2f544b7c3
        - registry.redhat.io/rhoso/ironic-python-agent-rhel9@sha256:2ccd8e22a242059ca5c4e2bce853c36a71c36fdf40e8e36ab183a8218b0cc87b
        - registry.redhat.io/rhoso/openstack-frr-rhel9@sha256:f6516277149d1560c0bf2ac74a99cfb4ac2c84f9f090ab231b6d12bdb824de3b
        - registry.redhat.io/rhoso/openstack-octavia-housekeeping-rhel9@sha256:0a79259e398af746d0e8a6dda1aae5ee976c6bf8991b17f4a5da0d07dca1e36c
        - registry.redhat.io/rhoso/openstack-swift-object-rhel9@sha256:8fc2ab8caa9c7c35d94ef51c303763ea71e40b9c795c4555c17e6194344419b3
        - registry.redhat.io/ubi9/httpd-24@sha256:29382518403a4361c48fccc057163dc4a9f43ee159432754fa63c76485b223f5
        - registry.redhat.io/rhoso-operators/heat-rhel9-operator@sha256:583abb81a49db85949f7a4b846591ad8532afa79440ad7e2937ee00eacd972b3
        - registry.redhat.io/rhoso/openstack-barbican-keystone-listener-rhel9@sha256:cd8d4f0a94b4fe080ebf8e459f091db6637ceec6d83ac23cf04d66bd03c71e76
        - registry.redhat.io/rhoso/openstack-neutron-metadata-agent-ovn-rhel9@sha256:20323b151474703df610b6c24ccb9c748265196bbc70a7b5ece1b801ff540cd9
        - registry.redhat.io/rhoso-operators/test-rhel9-operator@sha256:df7566483abb18aa3c45d8bc4cd91531c7eb5b63345e511151a7cdca2dbcef17
        - registry.redhat.io/rhoso/openstack-iscsid-rhel9@sha256:508bd67021c271a10bb55ab4af033dc4e931708674b9c5ba94f72c07cd0c8ee2
        - registry.redhat.io/rhoso-operators/openstack-baremetal-agent-rhel9@sha256:08675b16d89d7931b1ef7b96e1e884c88e8e7fbfbe79a944f9213c52aba3ba83
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/rhoso/openstack-ironic-neutron-agent-rhel9@sha256:0abaa8f437025d2e7006c59c546ff6935556d707b57993719faef94142e1c43d
        - registry.redhat.io/rhoso/openstack-aodh-evaluator-rhel9@sha256:48c9192acf157df5b8f0827e35f0380d50cd4562676aba3c0d7791f92328af02
        - registry.redhat.io/rhoso/openstack-manila-scheduler-rhel9@sha256:8ce47f7fd8660167b3b292659d4fc58cf34ed671d5779693d24b71a9fddc1e7c
        - registry.redhat.io/rhoso/openstack-neutron-dhcp-agent-rhel9@sha256:cc9223ef705bcd5d0dcabbf71ad9f6e540ae7f003a6afd13161301f842dfa5fa
        - registry.redhat.io/rhoso/edpm-hardened-uefi-rhel9@sha256:c18ec7023fbcc4ab5183f689adc1f01a626a4923c15c61a37388f99904db7bc3
        - registry.redhat.io/rhoso/openstack-ceilometer-central-rhel9@sha256:e6d9f47406b7e3ff2a7f2256088cc64f043b8e497382b9f569eb90182598bbe9
        - registry.redhat.io/rhoso/openstack-cinder-api-rhel9@sha256:72a674caeeeecce3b60217f3607ad5b00651fa3d52f2f5d905f6e6395638dbc8
        - registry.redhat.io/rhoso/openstack-glance-api-rhel9@sha256:b1c0b09bb9aafe3bc31f43f8a99a92dde8d78cf90e781dd14300580986d8c980
        - registry.redhat.io/rhoso-operators/designate-rhel9-operator@sha256:d446af622bd32666eb4eb5f9b7f38f0c33ac81251a2230db9d5baf71081b9d6a
        - registry.redhat.io/rhoso/openstack-designate-producer-rhel9@sha256:04af4230f815a6e3fde22003cf382eebd9cbc3c3ec4056795f4474abc1e65fea
        - registry.redhat.io/rhoso-operators/mariadb-rhel9-operator@sha256:907e3a9d15941440d5289d7e448d9f0351053fc20fcb8cb3b9db0a84debd876f
        - registry.redhat.io/rhoso/openstack-manila-api-rhel9@sha256:e4dd3d5b66c5216b45a1515a975b85fd2ac4ef698a9350dd16e01417da462b32
        - registry.redhat.io/rhoso/openstack-cinder-volume-rhel9@sha256:8a09e80dac94fb4b3d35033e91812f95f0acac66532d9abc88aa1fb715d21822
        - registry.redhat.io/rhoso/openstack-designate-worker-rhel9@sha256:a6324a8a72722822df7692b8df0420388a1b4aecfff3da7e91da4afe4d9e2260
        - registry.redhat.io/rhoso/openstack-ceilometer-compute-rhel9@sha256:296f7f6c6bdcd7f5d07b8470308a4c71055f29a17c48d2ebb03b7f54a0e927ff
        - registry.redhat.io/rhoso/openstack-heat-api-cfn-rhel9@sha256:2789929ae3edd7c68579f34457d3071273e49a66f462043cd9c46e75bdda282a
        - registry.redhat.io/openshift4/ose-kube-state-metrics-rhel9@sha256:db6074b8aeedfd7603ae3adbe67071115f54e7f01a24997f1cd8fa8958aeddce
        - registry.redhat.io/rhoso-operators/cinder-rhel9-operator@sha256:12b8dd8fd3683250365eff6f07339cb976ee29da80642e591c10e9fa5ba627c9
        - registry.redhat.io/rhoso/openstack-heat-engine-rhel9@sha256:151e04b3e5a408aadef3a17954255acc3733d09961878a135e177188457447f0
        - registry.redhat.io/rhoso/openstack-rsyslog-rhel9@sha256:272e729accc2b64b8fcc0ece4ae5aa9ce6d454e34a25a8a8e2b99ac56b73361e
        - registry.redhat.io/rhoso/openstack-tempest-all-rhel9@sha256:16b1fbfd36ce9a27106bf09c414eaa93f14c9a6b11839068976c2fdb537d2ba7
        - registry.redhat.io/rhoso/openstack-octavia-api-rhel9@sha256:7b8d5927162ddab512096beb02e837ba51f6353c5187f3de72eb3db6e56ab6e7
        - registry.redhat.io/rhoso/openstack-keystone-rhel9@sha256:ac076c81247d774a9aeee1b63540ca6e18058213cf6aad10a6fd1f535a394ef4
        - registry.redhat.io/rhoso/openstack-cinder-scheduler-rhel9@sha256:2079ffe9b620e9277573ac796bf7e393b520631130e5bbb6f5ab51d872c943f3
        - registry.redhat.io/rhoso/openstack-aodh-notifier-rhel9@sha256:d2e89f27aac9afa504411ae6a6163442cd717f8c560e505a7da8e3fd9e2cff95
        - registry.redhat.io/rhoso/openstack-manila-share-rhel9@sha256:277174c885998e6f5b867d099664f6fa28f974bffb524903ff75283035cd8dc9
        - registry.redhat.io/rhoso/openstack-neutron-server-rhel9@sha256:2d3afde3201aa18b192ad0fb079c1e4f4f673565604f1af2b79aba8a7bc88bf2
        - registry.redhat.io/rhoso/openstack-ovn-northd-rhel9@sha256:668f742b2a18579bc8cfc610a862e43c5b877413cf65eabbe0ee36b3c7869446
        - registry.redhat.io/rhoso-operators/openstack-baremetal-rhel9-operator@sha256:03916e507b88943c0e00cedeb3cf20fa0787ccc762e47f38d305ff4d1b125b12
        - registry.redhat.io/rhoso/openstack-ovn-nb-db-server-rhel9@sha256:8c2ce1027779dc66d67a0e0d03e48b226d3278208cae0b2b81b35ad1b1f3ae72
        - registry.redhat.io/rhoso/openstack-neutron-ovn-agent-rhel9@sha256:bed1956c66b0e70064b51e34faac64036f77c7ecb14f46ceac4bc9ef52f0b9e0
        - registry.redhat.io/rhoso/openstack-nova-scheduler-rhel9@sha256:f0e5d3a693c9998ca3839b3ec7dfd5da0f9cf139e0b5b6913e92a10d5a13e6bc
        - registry.redhat.io/rhoso/openstack-designate-central-rhel9@sha256:d28d8d91d8bf75a1b77a5d650c7f2650a1791e2a1b0341450fa46978c955b92f
        - registry.redhat.io/rhoso/openstack-barbican-api-rhel9@sha256:bc942204843c45c6cb7ea841af8611f04efb003de7eeccc6813934aad556da73
        - registry.redhat.io/rhoso/openstack-octavia-health-manager-rhel9@sha256:8065851058019757b56f86ca4f030d44f169100eb2e37e08e0587a7b1e4d2e72
        - registry.redhat.io/rhoso/openstack-nova-api-rhel9@sha256:4116206f1fbe60fdd9458a5edee848571c94e2b0512d9c2cfcfd6890d60d4395
        - registry.redhat.io/rhoso-operators/keystone-rhel9-operator@sha256:879d4b1f9e720f8581b3a55c788a97bbe7b52f9108df3d5dce2e9bcb4200df6d
        - registry.redhat.io/rhoso/openstack-horizon-rhel9@sha256:82d0e6d24733afd305996e0f4a118df3e48621c40c9669b01f48669dbf40d868
        - registry.redhat.io/rhoso/openstack-designate-api-rhel9@sha256:b76ca7820553285bb4ee2bfb08e39aedbfe8c16b4edb478c06418a1e335a64ab
        - registry.redhat.io/rhoso/openstack-swift-container-rhel9@sha256:95fec49643fc18f402b8ba8057f2c35f214bf6138c5f0d7bbe1b96cbd78ef8d6
        - registry.redhat.io/rhoso-operators/swift-rhel9-operator@sha256:c99ef14338290d63e2bcbf129e953b434aea382e47072812bae31b1ea838e7df
        - registry.redhat.io/rhoso/openstack-ovn-base-rhel9@sha256:a3b8ec7ad1073facb8876acc787549a1a68b9bf81bebdc127f13eb6385e2d1da
        - registry.redhat.io/rhoso/openstack-ovn-bgp-agent-rhel9@sha256:a404c3416ab65e7134f438362e887529be474639e725e18f3542e86973edac87
        - registry.redhat.io/rhoso/openstack-nova-novncproxy-rhel9@sha256:04881867f9b24403bf6ef051eee3902a17e8476eee894ef579c0d5d130d2510d
        - registry.redhat.io/rhoso/openstack-ironic-inspector-rhel9@sha256:88357e83e8a4bbc3c5a37f0420ad75556821834a975cd8a1cf5fee56de2a4f99
        - registry.redhat.io/rhoso/openstack-ovn-controller-rhel9@sha256:3c42b7e0305c508d74602b9378143deb07a569443bba35850cdb7b6e3f5202e2
        - registry.redhat.io/rhoso/openstack-nova-conductor-rhel9@sha256:c67ca0e6c8126f56c6a83632b6a96bd4a35006f5f414a8779467b81864b31d12
        - registry.redhat.io/rhoso-operators/ee-openstack-ansible-ee-rhel9@sha256:61e729fb89c4872837cbe9cb1e3cea402b61f0da240751302605df487721f0aa
        - registry.redhat.io/rhoso-operators/ovn-rhel9-operator@sha256:66cd8add6078b63a36879c974db9e9f358593bb0651aa39fe85153d25404ae27
        - registry.redhat.io/rhoso/openstack-designate-backend-bind9-rhel9@sha256:82cca999357b0ab25bcc081bf802fab831587cf661db806af4ca453e0a626743
        - registry.redhat.io/rhoso/openstack-rabbitmq-rhel9@sha256:56bc2d2b201225fb98db43ea9ca68cceb4842f8f0f1a96ab3ab62ab785651d12
        - registry.redhat.io/rhoso-operators/prometheus-podman-exporter-rhel9@sha256:f33e7506db6dfdf530d3e2442baaf360d212bb1a47a7db887dba967ed0655cc2
        - registry.redhat.io/rhoso/openstack-nova-compute-rhel9@sha256:6264075310231af358803b53d5a09250dab32d63fc5966aa8fb957e5f56096a8
        - registry.redhat.io/rhoso-operators/neutron-rhel9-operator@sha256:30b6a812c47db9e6691211185c304c3e8fc6a8b9ba7ca79acf59c5ec07fb9fb0
        - registry.redhat.io/rhoso-operators/openstack-must-gather-rhel9@sha256:60b4355f4dbef3ad64103357086b10a5c0af126dd23798d6a9d725646f1e13b4
        - registry.redhat.io/rhoso/openstack-ironic-api-rhel9@sha256:50ccfe7d9ba22a50e0c1c8159cade6f4b1ac13ad8f5770f20c7be18fba58c8c3
        - registry.redhat.io/rhoso/openstack-ironic-pxe-rhel9@sha256:42e35ded74033f1f07287441917829cc934930a5df696c33cd8957be3c10fa41
        - registry.redhat.io/rhoso/openstack-openstackclient-rhel9@sha256:2952521d4b1799fd32ff773ef2d457a3abc354b639cf7f1e0f7f6615ca35ac8f
        - registry.redhat.io/rhoso-operators/manila-rhel9-operator@sha256:d9e69eb140b8ff76d5bf49deab8d700639ccdfd1ab5c8c7667bc2bbb50f3d60b
        - registry.redhat.io/rhoso-operators/horizon-rhel9-operator@sha256:9a2a13635588c6cc9949b272e1a1708c5fb53f19b05b8310ace703df5958ac63
        - registry.redhat.io/rhoso-operators/nova-rhel9-operator@sha256:3156dde249e49b40bed71092fb5213b112d43b663ba8c556e355aed524bec822
        - registry.redhat.io/rhoso-operators/ironic-rhel9-operator@sha256:dd67ffe3578ce619748735c1a5cd2a57d704d0c36c1b841d513ea36479aa4682
        - registry.redhat.io/rhoso/openstack-aodh-api-rhel9@sha256:4adc09cc1d922b387320da467de89d3f41a9c49b46d9b728458fec36505ec940
        - registry.redhat.io/rhoso/openstack-cron-rhel9@sha256:f09e724c69cec2c3b685783f9263b26c280a57b5909a723341416b342d18f9a4
        - registry.redhat.io/rhoso/openstack-ironic-conductor-rhel9@sha256:538da9fdeec546d794d192361d0a76f8b4e7682b6eb6d9cea371d8a52d603800
        - registry.redhat.io/rhoso/openstack-swift-account-rhel9@sha256:31e727f37db92fcc21319c44396576a03a118ba25d403715650162f7c0ead31e
        - registry.redhat.io/rhoso/openstack-redis-rhel9@sha256:c6dbeacbbd5aed9d13e540f1c987a574ea45747983c280446e5afadcf7c33b37
        - registry.redhat.io/rhoso/openstack-ovn-sb-db-server-rhel9@sha256:8ba4f8ed49e214f421e94133da0f17b54ce99a52bbbd367d0aef357de844ea01
        - registry.redhat.io/openshift4/ose-prometheus-node-exporter-rhel9@sha256:479c51de2fc8a3b3858d71bbe4eeafbd5b43263472346f9dde348115082894cc
        - registry.redhat.io/rhoso-operators/octavia-rhel9-operator@sha256:aa0f00f8e03dfdb547e53cf611581e9cc091eaf049558121c2bc5d702e556bcb
        - registry.redhat.io/rhoso/openstack-mariadb-rhel9@sha256:6115a7458cfa412c1659242fcec64b63f0a5ca414ecf3c5f890f0afec0bca069
        - registry.redhat.io/rhoso/openstack-multipathd-rhel9@sha256:6ff8ae176f8a633bcd78b1f8407924c74bb2f317eba16b90e7f0f9e53af74e8e
        - registry.redhat.io/rhoso/openstack-neutron-sriov-agent-rhel9@sha256:118d5f9430192ab48cf159ee2f7912f08e935ad56ef3364f18b3680210bb3e52
        - registry.redhat.io/rhoso-operators/barbican-rhel9-operator@sha256:b3b769e0dc32c12c5cc9578c2cac2622b973177ca14697bc1042d21d473e4e84
        - registry.redhat.io/rhoso/openstack-heat-api-rhel9@sha256:ccbc41b552bb34d223c9615a4e3b56bb516bc5bd8a694f048c15063ba45fca9f
        - registry.redhat.io/rhoso-operators/telemetry-rhel9-operator@sha256:0635515a5fed12f17de4cc5ac5f9567e3b3546fc2f69730c7beab25cf722f538
        - registry.redhat.io/rhoso/openstack-aodh-listener-rhel9@sha256:037845be6787f610e479345206b0fcfb2f1dc430540816bc5a54a1d26da2acbd
        - registry.redhat.io/rhoso/openstack-ceilometer-notification-rhel9@sha256:bd2b5f50a2bef4e77e3daa89aca51df6e0e55c3784886296d92c97616cd566d1
        - registry.redhat.io/rhoso-operators/sg-core-rhel9@sha256:b56923f869ff57396e6822133dfc3b21007c3df652cfe962c7ad3028214d988c
        - registry.redhat.io/rhoso/openstack-barbican-worker-rhel9@sha256:485525116fb3c22d8c60330dd87ebd3854cd71ded36bf88bcf18cee5957e22a3
        - registry.redhat.io/rhoso-operators/openstack-network-exporter-rhel9@sha256:8e8e775e37c28cb11b7aae9199405b85d5862dce195d165ca598fa1dc89bf914
        - registry.redhat.io/rhoso/openstack-placement-api-rhel9@sha256:0aa5e138947bf236f31c81bfb6f641d95af63812c06dca43141721b383c7e08e
        - registry.redhat.io/openshift-power-monitoring/kepler-rhel9@sha256:8e9e5bfeed61aaeee85bf5eefc3390556b8d4cbfc797a82cabe95bc0367126eb
        - registry.redhat.io/rhoso/openstack-ceilometer-ipmi-rhel9@sha256:2fb5723c68d0f2c3ae28b86e606ccae4b80f3bd9b16c01dd6c9adcae629db7f1
        - registry.redhat.io/rhoso-operators/openstack-rhel9-operator@sha256:64e2ad312f757101f083eedcb17e0b6c05cf6ebf6ed9e7db2e33e84c6aca9898
        - registry.redhat.io/rhoso/openstack-unbound-rhel9@sha256:7f17a719e8e49d457ff16437709c2b4eca5b9c2f58972d1e351c5c1c1ac38962
        - registry.redhat.io/rhoso-operators/infra-rhel9-operator@sha256:a2dc031a6ad2726d155e77cafb676f8e35a86e963b97ccdceed0da137e67ba0d
        - registry.redhat.io/rhoso/openstack-designate-mdns-rhel9@sha256:ce43dd1a30afd4753aa11d32df2b24ed4bf3d67847044aaae8f3e5c8901caf05
        - registry.redhat.io/rhoso/openstack-octavia-worker-rhel9@sha256:303e35926e43fe403673e18a25c2e815547c28a5f8c0f09d470cd4148a21c528
        - registry.redhat.io/rhoso-operators/placement-rhel9-operator@sha256:a16bd4fb929eb5bcf96f2aefcb5a7abf4143f4f9ed24e53de93f4a3caa89a507
        - registry.redhat.io/rhoso-operators/rabbitmq-cluster-rhel9-operator@sha256:e095e6cbda841b9c726c1ec0839a98964b894cf25ae9228052d314b1ee2f161d
        - registry.redhat.io/rhoso/openstack-memcached-rhel9@sha256:3e704b9664c65be39e83c2ed802722ffeaf969d8e571147a2be128d5c56331c1
        version: 1.0.14
      entries:
      - name: openstack-operator.v1.0.14
        version: 1.0.14
      - name: openstack-operator.v1.0.13
        version: 1.0.13
      - name: openstack-operator.v1.0.12
        version: 1.0.12
      - name: openstack-operator.v1.0.11
        version: 1.0.11
      - name: openstack-operator.v1.0.10
        version: 1.0.10
      - name: openstack-operator.v1.0.9
        version: 1.0.9
      - name: openstack-operator.v1.0.8
        version: 1.0.8
      - name: openstack-operator.v1.0.7
        version: 1.0.7
      - name: openstack-operator.v1.0.6
        version: 1.0.6
      - name: openstack-operator.v1.0.5
        version: 1.0.5
      - name: openstack-operator.v1.0.4
        version: 1.0.4
      - name: openstack-operator.v1.0.3
        version: 1.0.3
      - name: openstack-operator.v1.0.2
        version: 1.0.2
      - name: openstack-operator.v1.0.1
        version: 1.0.1
      - name: openstack-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: openstack-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Devfile
      provider-url: https://devfile.io
    name: devworkspace-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devworkspace-operator.v0.36.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/internal-objects: '["devworkspaceroutings.controller.devfile.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DevWorkspaceOperatorConfig
            name: devworkspaceoperatorconfigs.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspaceRouting
            name: devworkspaceroutings.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha2
          - kind: DevWorkspaceTemplate
            name: devworkspacetemplates.workspace.devfile.io
            version: v1alpha1
          - kind: DevWorkspaceTemplate
            name: devworkspacetemplates.workspace.devfile.io
            version: v1alpha2
        description: |
          The DevWorkspace Operator enables cluster-level support for the
          [Devfile 2.0 spec](https://docs.devfile.io), enabling static, reproducible
          configurations for launching cloud-based editors and IDEs in OpenShift and
          Kubernetes clusters.

          Leveraging the language server protocol and the Eclipse Theia web IDE, the
          DevWorkspace operator provides easy configuration of full development
          environments on the cloud with support for a wide variety of languages and
          technologies, including Go, Java, Typescript/Javascript, Python, and more.

          The DevWorkspace Operator is also used in the Web Terminal Operator to
          automatically provision Web Terminal environments.

          ## Installing the operator
          The DevWorkspace Operator can be installed directly from the OperatorHub UI and
          will be available in all namespaces on the cluster. DevWorkspace creation is
          driven by the DevWorkspace custom resource, which can be created in any
          namespace to provision a full development environment. To get started, browse
          the DevWorkspace [spec](https://devfile.io/docs/2.3.0/devfile-schema).

          Once a DevWorkspace is started, it can be accessed via the URL in its
          `.status.mainUrl` field.

          It's recommended to install the DevWorkspace Operator to the
          `openshift-operators` namespace for compatibility.

          ## Uninstalling the operator
          The DevWorkspace Operator utilizes finalizers on resources it creates and
          webhooks to restrict access to development resources. As a result, manual steps
          are required in uninstalling the operator. See the
          [documentation](https://github.com/devfile/devworkspace-operator/blob/main/docs/uninstall.md)
          for details.
        displayName: DevWorkspace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Devworkspace Operator
        - DevWorkspaces
        - Devfile
        links:
        - name: Devworkspace Operator
          url: https://github.com/devfile/devworkspace-operator
        maintainers:
        - email: dakwon@redhat.com
          name: David Kwon
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        maturity: alpha
        provider:
          name: Devfile
          url: https://devfile.io
        relatedImages:
        - registry.redhat.io/devworkspace/devworkspace-rhel9-operator@sha256:83b6abf6e57703da1125e037ea9353c224602bddc3146e5b00df47f1b1976e9a
        - registry.redhat.io/ubi9/ubi-minimal@sha256:a50731d3397a4ee28583f1699842183d4d24fadcc565c4688487af9ee4e13a44
        - registry.redhat.io/devworkspace/devworkspace-project-clone-rhel9@sha256:312a013201877cea03956e3db2ba19a3d7f842de2c3a65927c65c3e8d22a57fe
        version: 0.36.0
      entries:
      - name: devworkspace-operator.v0.36.0
        version: 0.36.0
      - name: devworkspace-operator.v0.35.1
        version: 0.35.1
      - name: devworkspace-operator.v0.34.0
        version: 0.34.0
      - name: devworkspace-operator.v0.33.0
        version: 0.33.0
      - name: devworkspace-operator.v0.32.1
        version: 0.32.1
      - name: devworkspace-operator.v0.31.2
        version: 0.31.2
      - name: devworkspace-operator.v0.31.1
        version: 0.31.1
      - name: devworkspace-operator.v0.31.0
        version: 0.31.0
      - name: devworkspace-operator.v0.30.2
        version: 0.30.2
      - name: devworkspace-operator.v0.30.1
        version: 0.30.1
      - name: devworkspace-operator.v0.30.0
        version: 0.30.0
      - name: devworkspace-operator.v0.29.0
        version: 0.29.0
      - name: devworkspace-operator.v0.28.0-0.1719314743.p
        version: 0.28.0+0.1719314743.p
      - name: devworkspace-operator.v0.28.0
        version: 0.28.0
      - name: devworkspace-operator.v0.27.0-0.1714987663.p
        version: 0.27.0+0.1714987663.p
      - name: devworkspace-operator.v0.27.0
        version: 0.27.0
      - name: devworkspace-operator.v0.26.0
        version: 0.26.0
      - name: devworkspace-operator.v0.25.0
        version: 0.25.0
      - name: devworkspace-operator.v0.24.0
        version: 0.24.0
      - name: devworkspace-operator.v0.23.0
        version: 0.23.0
      - name: devworkspace-operator.v0.22.0
        version: 0.22.0
      - name: devworkspace-operator.v0.21.2
        version: 0.21.2
      - name: devworkspace-operator.v0.21.1
        version: 0.21.1
      - name: devworkspace-operator.v0.20.0
        version: 0.20.0
      - name: devworkspace-operator.v0.19.1
        version: 0.19.1
      - name: devworkspace-operator.v0.19.1-0.1682321189.p
        version: 0.19.1+0.1682321189.p
      - name: devworkspace-operator.v0.19.1-0.1679521112.p
        version: 0.19.1+0.1679521112.p
      - name: devworkspace-operator.v0.18.1
        version: 0.18.1
      - name: devworkspace-operator.v0.18.1-0.1675929565.p
        version: 0.18.1+0.1675929565.p
      - name: devworkspace-operator.v0.17.0
        version: 0.17.0
      - name: devworkspace-operator.v0.16.0-0.1666668361.p
        version: 0.16.0+0.1666668361.p
      - name: devworkspace-operator.v0.16.0
        version: 0.16.0
      - name: devworkspace-operator.v0.15.2-0.1661828401.p
        version: 0.15.2+0.1661828401.p
      - name: devworkspace-operator.v0.15.2
        version: 0.15.2
      - name: devworkspace-operator.v0.14.1
        version: 0.14.1
      - name: devworkspace-operator.v0.13.0
        version: 0.13.0
      - name: devworkspace-operator.v0.12.0
        version: 0.12.0
      - name: devworkspace-operator.v0.11.0
        version: 0.11.0
      - name: devworkspace-operator.v0.10.0
        version: 0.10.0
      - name: devworkspace-operator.v0.9.0
        version: 0.9.0
      name: fast
    defaultChannel: fast
    packageName: devworkspace-operator
    provider:
      name: Devfile
      url: https://devfile.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microcks.io
      provider-url: ""
    name: microcks
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: microcks-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "microcks.github.io/v1alpha1",
                "kind": "MicrocksInstall",
                "metadata": {
                  "name": "my-microcksinstall"
                },
                "spec": {
                  "name": "my-microcksinstall",
                  "version": "1.10.0",
                  "microcks": {
                    "replicas": 1
                  },
                  "postman": {
                    "replicas": 1
                  },
                  "keycloak": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "1Gi"
                  },
                  "mongodb": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "2Gi"
                  }
                }
              },
              {
                "apiVersion": "microcks.github.io/v1alpha1",
                "kind": "MicrocksInstall",
                "metadata": {
                  "name": "my-microcksinstall-minikube"
                },
                "spec": {
                  "name": "my-microcksinstall-minikube",
                  "version": "1.10.0",
                  "microcks": {
                    "replicas": 2,
                    "url": "microcks.192.168.99.100.nip.io"
                  },
                  "postman": {
                    "replicas": 2
                  },
                  "keycloak": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "1Gi",
                    "url": "keycloak.192.168.99.100.nip.io"
                  },
                  "mongodb": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "2Gi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/microcks/microcks-ansible-operator:1.10.0
          createdAt: "2024-08-01T09:14:55Z"
          description: Open Source mocking and testing platform for API and microservices
          repository: https://github.com/microcks/microcks-ansible-operator
          support: microcks.github.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Microcks installation
            displayName: MicrocksInstall
            kind: MicrocksInstall
            name: microcksinstalls.microcks.github.io
            version: v1alpha1
        description: |
          Microcks is an Open source tool for API and microservices mocking and testing. It provides a platform for referencing, deploying mocks and allow contract-testing of your APIs. It can also be considered as a Service Virtualization solution because it will allow you to provide fake API or Service implementation before development being actually done. It supports Async APIs, REST APIs, gRPC, GraphQL and SOAP WebServices. It perfectly integrates into an iterative, contract-first delivery process.
          Microcks tries not reinventing the wheel and let you capitalize on already known tools of standards. It supports [AsyncAPI Specifications](https://asyncapi.com), [OpenAPI Specification](https://www.openapis.org/) `3.x`, [gRPC Protobuf](https://grpc.io/), [GraphQL Schema](https://graphql.org) out-of-the-box and can also relies on popular tools like [SoapUI](http://www.soapui.com) or [Postman](http://www.getpostman.io) to edit your API request / response samples as well as your test scripts.
          ### Operator features
          * **Create/Destroy**: Easily launch Microcks installations (`MicrocksInstall` custom resource) for your Kubernetes namespace
          ### MicrocksInstall options
          * **keycloak** component can be deployed as part of the install or you can reuse an existing instance. You can reference an existing `Secret` for providing TLS certificates for securing the `Ingress`,
          * **mongodb** component can be deployed as part of the install or you can reuse an existing instance, You can reference an existing `Secret` for providing connection credentials,
          * **features** component allow to activate optional features like the mocking of AsyncAPIs. In case you turn `features.async.enabled` to `true` and `features.async.kafka.install` to `true`, you must have the [Strimzi Operator](https://operatorhub.io/operator/strimzi-kafka-operator) already installed on your cluster/namespace.
          ### Documentation
          Documentation to the current _master_ branch as well as all releases can be found on our [GitHub](https://github.com/microcks/microcks-ansible-operator).
          ### License
          Microcks is licensed under the [Apache License, Version 2.0](https://github.com/microcks/microcks/blob/master/LICENSE).
        displayName: Microcks Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mock
        - test
        - API
        - microservice
        links:
        - name: Product Page
          url: https://microcks.io
        - name: Microcks Operator Source Code
          url: https://github.com/microcks/microcks-ansible-operator
        maintainers:
        - email: laurent.broudoux@gmail.com
          name: lbroudoux
        maturity: alpha
        provider:
          name: Microcks.io
        relatedImages:
        - quay.io/microcks/microcks-ansible-operator:1.10.0
        version: 1.10.0
      entries:
      - name: microcks-operator.v1.10.0
        version: 1.10.0
      - name: microcks-operator.v1.9.0
        version: 1.9.0
      - name: microcks-operator.v1.8.1
        version: 1.8.1
      - name: microcks-operator.v1.8.0
        version: 1.8.0
      - name: microcks-operator.v1.7.1
        version: 1.7.1
      - name: microcks-operator.v1.7.0
        version: 1.7.0
      - name: microcks-operator.v1.6.1
        version: 1.6.1
      - name: microcks-operator.v1.6.0
        version: 1.6.0
      - name: microcks-operator.v1.5.2
        version: 1.5.2
      - name: microcks-operator.v1.5.1
        version: 1.5.1
      - name: microcks-operator.v1.5.0
        version: 1.5.0
      - name: microcks-operator.v1.4.1
        version: 1.4.1
      name: stable
    defaultChannel: stable
    packageName: microcks
    provider:
      name: Microcks.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: ""
    name: pmem-csi-operator-os
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pmem-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pmem-csi.intel.com/v1beta1",
                "kind": "PmemCSIDeployment",
                "metadata": {
                  "name": "pmem-csi.intel.com"
                },
                "spec": {
                  "deviceMode": "lvm",
                  "nodeSelector": {
                    "feature.node.kubernetes.io/memory-nv.dax": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
          createdAt: "2021-08-26T10:45:23Z"
          description: An operator for deploying and managing the PMEM-CSI driver
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: ""
          repository: https://github.com/intel/pmem-csi/tree/release-1.0
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a PMEM-CSI driver deployment in the cluster.
            displayName: PmemCSIDeployment
            kind: PmemCSIDeployment
            name: pmemcsideployments.pmem-csi.intel.com
            version: v1beta1
        description: |2

          Intel PMEM-CSI is a [CSI](https://github.com/container-storage-interface/spec)
          storage driver for container orchestrators like Kubernetes. It makes local
          persistent memory ([PMEM](https://pmem.io/)) available as a filesystem volume to
          container applications. This is the operator to deploy and manage the
          [PMEM-CSI](https://intel.github.io/pmem-csi/1.0/README.html)
          driver on a Kubernetes cluster.

          Refer to the [PMEM-CSI documentation](https://intel.github.io/pmem-csi/1.0/README.html)
          for more details on deploying and using the PMEM-CSI driver.
        displayName: Operator for PMEM-CSI driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - PMEM-CSI
        - PMEM-CSI operator
        - Persistent Memory
        - PMEM
        links:
        - name: PMEM-CSI
          url: https://intel.github.io/pmem-csi/1.0
        maintainers:
        - email: olev.kartau@intel.com
          name: Olev Kartau
        - email: patrick.ohly@intel.com
          name: Patrick Ohly
        - email: amarnath.valluri@intel.com
          name: Amarnath Valluri
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Intel® Corporation
        relatedImages:
        - registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
        version: 1.0.0
      entries:
      - name: pmem-csi-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: pmem-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pmem-csi.intel.com/v1beta1",
                "kind": "PmemCSIDeployment",
                "metadata": {
                  "name": "pmem-csi.intel.com"
                },
                "spec": {
                  "deviceMode": "lvm",
                  "nodeSelector": {
                    "feature.node.kubernetes.io/memory-nv.dax": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
          createdAt: "2021-08-26T10:45:23Z"
          description: An operator for deploying and managing the PMEM-CSI driver
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: ""
          repository: https://github.com/intel/pmem-csi/tree/release-1.0
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a PMEM-CSI driver deployment in the cluster.
            displayName: PmemCSIDeployment
            kind: PmemCSIDeployment
            name: pmemcsideployments.pmem-csi.intel.com
            version: v1beta1
        description: |2

          Intel PMEM-CSI is a [CSI](https://github.com/container-storage-interface/spec)
          storage driver for container orchestrators like Kubernetes. It makes local
          persistent memory ([PMEM](https://pmem.io/)) available as a filesystem volume to
          container applications. This is the operator to deploy and manage the
          [PMEM-CSI](https://intel.github.io/pmem-csi/1.0/README.html)
          driver on a Kubernetes cluster.

          Refer to the [PMEM-CSI documentation](https://intel.github.io/pmem-csi/1.0/README.html)
          for more details on deploying and using the PMEM-CSI driver.
        displayName: Operator for PMEM-CSI driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - PMEM-CSI
        - PMEM-CSI operator
        - Persistent Memory
        - PMEM
        links:
        - name: PMEM-CSI
          url: https://intel.github.io/pmem-csi/1.0
        maintainers:
        - email: olev.kartau@intel.com
          name: Olev Kartau
        - email: patrick.ohly@intel.com
          name: Patrick Ohly
        - email: amarnath.valluri@intel.com
          name: Amarnath Valluri
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Intel® Corporation
        relatedImages:
        - registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
        version: 1.0.0
      entries:
      - name: pmem-csi-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pmem-csi-operator-os
    provider:
      name: Intel® Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OceanBase
      provider-url: https://oceanbase.com
    name: oceanbase-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oceanbase-operator.v2.2.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Full Lifecycle
          createdAt: "2024-08-14T16:21:06Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: OBClusterOperation
            name: obclusteroperations.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBCluster
            name: obclusters.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBParameter
            name: obparameters.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBResourceRescue
            name: obresourcerescues.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBServer
            name: observers.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantBackupPolicy
            name: obtenantbackuppolicies.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantBackup
            name: obtenantbackups.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantOperation
            name: obtenantoperations.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantRestore
            name: obtenantrestores.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenant
            name: obtenants.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBZone
            name: obzones.oceanbase.oceanbase.com
            version: v1alpha1
        description: Ob Operator description. TODO.
        displayName: Ob Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - oceanbase-operator
        - ob-operator
        links:
        - name: OceanBase Operator
          url: https://oceanbase.github.io/ob-operator
        maintainers:
        - email: hongxin.sun.vip@gmail.com
          name: Hongxin Sun
        maturity: alpha
        provider:
          name: OceanBase
          url: https://oceanbase.com
        relatedImages:
        - quay.io/oceanbasedev/oceanbase-operator@sha256:b11fe60bb8de545d3ed105acc50f2d6c5d6aa8093f3ae4ca7ebce58b1e5e7b48
        - quay.io/brancz/kube-rbac-proxy@sha256:18f12fc3a0085bbc303cfbd2ad1151127c1c959c4641e36027547b6b66eb4c92
        version: 2.2.2
      entries:
      - name: oceanbase-operator.v2.2.2
        version: 2.2.2
      name: alpha
    - currentCSV: oceanbase-operator.v2.2.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Full Lifecycle
          createdAt: "2024-08-14T16:21:06Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: OBClusterOperation
            name: obclusteroperations.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBCluster
            name: obclusters.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBParameter
            name: obparameters.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBResourceRescue
            name: obresourcerescues.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBServer
            name: observers.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantBackupPolicy
            name: obtenantbackuppolicies.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantBackup
            name: obtenantbackups.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantOperation
            name: obtenantoperations.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantRestore
            name: obtenantrestores.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenant
            name: obtenants.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBZone
            name: obzones.oceanbase.oceanbase.com
            version: v1alpha1
        description: Ob Operator description. TODO.
        displayName: Ob Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - oceanbase-operator
        - ob-operator
        links:
        - name: OceanBase Operator
          url: https://oceanbase.github.io/ob-operator
        maintainers:
        - email: hongxin.sun.vip@gmail.com
          name: Hongxin Sun
        maturity: alpha
        provider:
          name: OceanBase
          url: https://oceanbase.com
        relatedImages:
        - quay.io/oceanbasedev/oceanbase-operator@sha256:b11fe60bb8de545d3ed105acc50f2d6c5d6aa8093f3ae4ca7ebce58b1e5e7b48
        - quay.io/brancz/kube-rbac-proxy@sha256:18f12fc3a0085bbc303cfbd2ad1151127c1c959c4641e36027547b6b66eb4c92
        version: 2.2.2
      entries:
      - name: oceanbase-operator.v2.2.2
        version: 2.2.2
      name: stable
    defaultChannel: stable
    packageName: oceanbase-operator
    provider:
      name: OceanBase
      url: https://oceanbase.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.2.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"customImageNames":
            false,"externalArangoDBName": "arangodb","repository": "registry.connect.redhat.com/turbonomic","tag":"8.3.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/turbonomic/t8c-operator@sha256:d5164e6dd24a0feff10cab62530cafff1bc92d0e8ec2d16c6f4d880c6e2f50bd
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/t8c-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/t8c-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/turbonomic/t8c-operator@sha256:d5164e6dd24a0feff10cab62530cafff1bc92d0e8ec2d16c6f4d880c6e2f50bd
        version: 42.2.0
      entries:
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: stable
    defaultChannel: stable
    packageName: t8c-certified-rhmp
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ptp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ptp-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "NodePtpDevice",
                "metadata": {
                  "name": "node.example.com"
                },
                "spec": {}
              },
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "PtpConfig",
                "metadata": {
                  "name": "example-ptpconfig"
                },
                "spec": {
                  "profile": [
                    {
                      "interface": "ens787f1",
                      "name": "profile1",
                      "phc2sysOpts": "-a -r",
                      "ptp4lOpts": "-s -2",
                      "ptpClockThreshold": {}
                    }
                  ],
                  "recommend": [
                    {
                      "match": [
                        {
                          "nodeLabel": "node-role.kubernetes.io/worker=",
                          "nodeName": "node.example.com"
                        }
                      ],
                      "priority": 4,
                      "profile": "profile1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "PtpOperatorConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "daemonNodeSelector": {
                    "feature.node.kubernetes.io/ptp-capable": "yes",
                    "node-role.kubernetes.io/worker": ""
                  },
                  "ptpEventConfig": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-ptp-rhel9-operator@sha256:012334e6624d5440439824454bda1d1dc79804938174232a3b470176bf462b5e
          createdAt: "2024-03-13"
          description: This software enables configuration of Precision Time Protocol(PTP)
            on Kubernetes. It detects hardware capable PTP devices on each node, and
            configures linuxptp processes such as ptp4l, phc2sys and timemaster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.16.0-202509240123'
          operatorframework.io/suggested-namespace: openshift-ptp
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/ptp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodePtpDevice is the Schema for the nodeptpdevices API
            displayName: Node Ptp Device
            kind: NodePtpDevice
            name: nodeptpdevices.ptp.openshift.io
            version: v1
          - description: PtpConfig is the Schema for the ptpconfigs API
            displayName: Ptp Config
            kind: PtpConfig
            name: ptpconfigs.ptp.openshift.io
            version: v1
          - description: PtpOperatorConfig is the Schema for the ptpoperatorconfigs
              API
            displayName: Ptp Operator Config
            kind: PtpOperatorConfig
            name: ptpoperatorconfigs.ptp.openshift.io
            version: v1
        description: |
          # Precision Time Protocol (PTP) Operator for Openshift

          ## Introdution
          PTP Operator manages cluster wide PTP configuration.
          This operator has to run in namespace 'openshift-ptp'. An Operator Group is also required to install this operator:
          ```
          $ oc create -f - <<EOF
          apiVersion: v1
          kind: Namespace
          metadata:
            name: openshift-ptp
            labels:
              name: openshift-ptp
              openshift.io/cluster-monitoring: "true"
          EOF

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: ptp-operators
            namespace: openshift-ptp
          spec:
            targetNamespaces:
            - openshift-ptp
          EOF
          ```
        displayName: PTP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ptp
        links:
        - name: Documentation operator
          url: https://github.com/openshift/ptp-operator
        - name: Documentation operand
          url: https://github.com/openshift/linuxptp-daemon
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-cloud-event-proxy-rhel9@sha256:3fa2aad8cf026eba586645c851f07a1d752c85b30795fdcf037b138ef3bbb6aa
        - registry.redhat.io/openshift4/ose-ptp-rhel9-operator@sha256:012334e6624d5440439824454bda1d1dc79804938174232a3b470176bf462b5e
        - registry.redhat.io/openshift4/ose-ptp-rhel9@sha256:93b4717337b551603cbb6147d66d7b70a5580332953ddf42304f7092d8739c83
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        version: 4.16.0-202509240123
      entries:
      - name: ptp-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: ptp-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: ptp-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: ptp-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: ptp-operator.v4.16.0-202506201205
        version: 4.16.0-202506201205
      - name: ptp-operator.v4.16.0-202505291140
        version: 4.16.0-202505291140
      - name: ptp-operator.v4.16.0-202505161704
        version: 4.16.0-202505161704
      - name: ptp-operator.v4.16.0-202505020146
        version: 4.16.0-202505020146
      - name: ptp-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: ptp-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: ptp-operator.v4.16.0-202502260004
        version: 4.16.0-202502260004
      - name: ptp-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: ptp-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: ptp-operator.v4.16.0-202501201505
        version: 4.16.0-202501201505
      - name: ptp-operator.v4.16.0-202501072135
        version: 4.16.0-202501072135
      - name: ptp-operator.v4.16.0-202412311307
        version: 4.16.0-202412311307
      - name: ptp-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: ptp-operator.v4.16.0-202412102004
        version: 4.16.0-202412102004
      - name: ptp-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: ptp-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: ptp-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: ptp-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: ptp-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: ptp-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: ptp-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: ptp-operator.v4.16.0-202409131735
        version: 4.16.0-202409131735
      - name: ptp-operator.v4.16.0-202409062036
        version: 4.16.0-202409062036
      - name: ptp-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: ptp-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: ptp-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: ptp-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: ptp-operator.v4.16.0-202407311708
        version: 4.16.0-202407311708
      - name: ptp-operator.v4.16.0-202407301837
        version: 4.16.0-202407301837
      - name: ptp-operator.v4.16.0-202407160236
        version: 4.16.0-202407160236
      - name: ptp-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: ptp-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: ptp-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: ptp-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: ptp-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: ptp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: IBM
      provider-url: ""
    name: ibm-block-csi-operator-community
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12
    - currentCSV: ibm-block-csi-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
        version: 1.12.0
      entries:
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.0
    - currentCSV: ibm-block-csi-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
        version: 1.12.1
      entries:
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.1
    - currentCSV: ibm-block-csi-operator.v1.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:3fc8df547138a2009225d63c2a3111a672340bdf9f3cb24a889cbb69f2c33c2d
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:3fc8df547138a2009225d63c2a3111a672340bdf9f3cb24a889cbb69f2c33c2d
        version: 1.12.2
      entries:
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.2
    - currentCSV: ibm-block-csi-operator.v1.12.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:2bd461a1e40a1366c871585aa7a4e8039e686a66bf7192efdeb1c215209a071d
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:2bd461a1e40a1366c871585aa7a4e8039e686a66bf7192efdeb1c215209a071d
        version: 1.12.3
      entries:
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.3
    - currentCSV: ibm-block-csi-operator.v1.12.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e29db476de3778222bf049926223dee20f6bc34e55a04df5f0f3df38741ff23e
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e29db476de3778222bf049926223dee20f6bc34e55a04df5f0f3df38741ff23e
        version: 1.12.4
      entries:
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.4
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.5
    defaultChannel: stable
    packageName: ibm-block-csi-operator-community
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Citrus Framework
      provider-url: ""
    name: yaks
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: yaks-operator.v0.17.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "yaks.citrusframework.org/v1alpha1",
                "kind": "Instance",
                "metadata": {
                  "labels": {
                    "app": "yaks"
                  },
                  "name": "example-instance"
                },
                "spec": {
                  "operator": {
                    "global": true
                  }
                }
              },
              {
                "apiVersion": "yaks.citrusframework.org/v1alpha1",
                "kind": "Test",
                "metadata": {
                  "labels": {
                    "app": "yaks"
                  },
                  "name": "example-test"
                },
                "spec": {
                  "source": {
                    "content": "Feature: hello world\n\n  Scenario: print slogan\n    Given YAKS does Cloud-Native BDD testing\n    Then YAKS rocks!",
                    "language": "feature",
                    "name": "example.feature"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/citrusframework/yaks:0.17.1
          createdAt: "2023-04-26T20:37:31Z"
          description: YAKS is a platform to enable Cloud Native BDD testing on Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["instances.yaks.citrusframework.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/citrusframework/yaks
          support: Citrus Framework
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance is the Schema for the yaks instance
            displayName: Instance
            kind: Instance
            name: instances.yaks.citrusframework.org
            version: v1alpha1
          - description: Test is the Schema for the tests API
            displayName: Test
            kind: Test
            name: tests.yaks.citrusframework.org
            version: v1alpha1
        description: |
          YAKS
          ====

          YAKS is a platform to enable Cloud Native BDD testing on Kubernetes.

          ## Running a YAKS test

          With the YAKS operator installed, you can run tests by creating a `Test` custom resource on the cluster.

          Tests are defined using [Gherkin](https://cucumber.io/docs/gherkin/) syntax. YAKS provides a set of predefined steps which
          help to connect with different messaging transports (Http REST, JMS, Kafka, Knative eventing) and verify responses with
          assertions on message header and body content.

          The example below defines a simple test resource.

          ```
          apiVersion: yaks.citrusframework.org/v1alpha1
          kind: Test
          metadata:
            name: example
          spec:
            source:
              name: example.feature
              content: |-
                Feature: hello world

                  Scenario: print slogan
                    Given YAKS does Cloud-Native BDD testing
                    Then YAKS rocks!
          ```

          Refer to the [YAKS repository](https://github.com/citrusframework/yaks) for more complex examples.
        displayName: YAKS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - yaks
        - testing
        - microservices
        - knative
        - kafka
        - serverless
        - camel
        - camel-k
        links:
        - name: Project page
          url: https://citrusframework.org/
        - name: YAKS source code repository
          url: https://github.com/citrusframework/yaks
        maintainers:
        - email: cdeppisch@redhat.com
          name: Christoph Deppisch
        - email: nferraro@redhat.com
          name: Nicola Ferraro
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Citrus Framework
        relatedImages:
        - docker.io/citrusframework/yaks:0.17.1
        version: 0.17.1
      entries:
      - name: yaks-operator.v0.17.1
        version: 0.17.1
      - name: yaks-operator.v0.17.0
        version: 0.17.0
      - name: yaks-operator.v0.16.0
        version: 0.16.0
      - name: yaks-operator.v0.15.1
        version: 0.15.1
      - name: yaks-operator.v0.15.0
        version: 0.15.0
      - name: yaks-operator.v0.14.3
        version: 0.14.3
      - name: yaks-operator.v0.14.2
        version: 0.14.2
      - name: yaks-operator.v0.14.1
        version: 0.14.1
      - name: yaks-operator.v0.13.0
        version: 0.13.0
      - name: yaks-operator.v0.12.2
        version: 0.12.2
      - name: yaks-operator.v0.11.0
        version: 0.11.0
      - name: yaks-operator.v0.10.0
        version: 0.10.0
      - name: yaks-operator.v0.9.0
        version: 0.9.0
      - name: yaks-operator.v0.8.0
        version: 0.8.0
      - name: yaks-operator.v0.7.0
        version: 0.7.0
      - name: yaks-operator.v0.6.0
        version: 0.6.0
      - name: yaks-operator.v0.5.0
        version: 0.5.0
      - name: yaks-operator.v0.4.0
        version: 0.4.0
      name: alpha
    defaultChannel: alpha
    packageName: yaks
    provider:
      name: Citrus Framework
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator.v1.5.0
        version: 1.5.0
      - name: dynatrace-operator.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetScaler
      provider-url: https://netscaler.com
    name: netscaler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netscaler-operator.v3.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCloudController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalercloudcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "azure": {
                    "clientsecret": "",
                    "image": "quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb",
                    "location": "",
                    "pullPolicy": "IfNotPresent",
                    "resourcegroupname": "",
                    "resources": {},
                    "subnetname": "",
                    "subscriptionid": "",
                    "tenantid": "",
                    "vnetname": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCpxWithIngressController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalercpxwithingresscontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "admSettings": {
                    "admIP": "",
                    "bandWidth": 1000,
                    "bandWidthLicense": false,
                    "cpxCores": 1,
                    "licenseEdition": "PLATINUM",
                    "licenseServerIP": "",
                    "licenseServerPort": 27000,
                    "loginSecret": "",
                    "platform": false,
                    "vCPULicense": false
                  },
                  "affinity": {},
                  "analyticsConfig": {
                    "distributedTracing": {
                      "enable": false,
                      "samplingRate": 100
                    },
                    "endpoint": {
                      "metrics": {
                        "service": ""
                      },
                      "transactions": {
                        "service": ""
                      }
                    },
                    "required": false,
                    "timeseries": {
                      "auditLogs": {
                        "enable": false
                      },
                      "events": {
                        "enable": false
                      },
                      "metrics": {
                        "enable": false,
                        "enableNativeScrape": false,
                        "exportFrequency": 30,
                        "mode": "avro",
                        "schemaFile": "schema.json"
                      },
                      "port": 5563
                    },
                    "transactions": {
                      "enable": false,
                      "port": 5557
                    }
                  },
                  "exporter": {
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4",
                    "ports": {
                      "containerPort": 8888
                    },
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {}
                  },
                  "extraVolumes": [],
                  "imagePullSecrets": [],
                  "ingressController": {
                    "defaultSSLCertSecret": "",
                    "defaultSSLSNICertSecret": "",
                    "disableAPIServerCertVerify": false,
                    "disableOpenshiftRoutes": true,
                    "enableLivenessProbe": true,
                    "enableReadinessProbe": true,
                    "entityPrefix": "k8s",
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "ingressClass": [],
                    "ipam": false,
                    "jsonLog": false,
                    "kubernetesURL": "",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "logLevel": "INFO",
                    "logProxy": "",
                    "namespaceLabels": "",
                    "openshift": true,
                    "optimizeEndpointBinding": false,
                    "profileHttpFrontend": {},
                    "profileSslFrontend": {},
                    "profileTcpFrontend": {},
                    "pullPolicy": "IfNotPresent",
                    "rbacRole": false,
                    "readinessProbe": {
                      "failureThreshold": 3,
                      "initialDelaySeconds": 50,
                      "periodSeconds": 60,
                      "successThreshold": 1
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "128m",
                        "memory": "500Mi"
                      }
                    },
                    "routeLabels": "",
                    "setAsDefaultIngressClass": false,
                    "updateIngressStatus": true
                  },
                  "netscalerCpx": {
                    "bgpSettings": {
                      "bgpConfig": [
                        {
                          "bgpRouter": {
                            "localAS": 100,
                            "neighbor": [
                              {
                                "ASOriginationInterval": 10,
                                "address": "",
                                "advertisementInterval": 10,
                                "remoteAS": 100
                              }
                            ]
                          }
                        }
                      ],
                      "bgpPort": 179
                    },
                    "cpxBgpRouter": false,
                    "cpxCommands": [],
                    "cpxLicenseAggregator": "",
                    "cpxShellCommands": [],
                    "daemonSet": false,
                    "enableLivenessProbe": true,
                    "enableStartupProbe": true,
                    "extraVolumeMounts": [],
                    "hostname": "",
                    "image": "quay.io/netscaler/netscaler-cpx@sha256:232596eaacda68c0e2e210f26b64cedc8d70ae12c5e8ae2e1a66896a479ae4f8",
                    "ingressIP": "",
                    "livenessProbe": {
                      "failureThreshold": 3,
                      "periodSeconds": 5,
                      "successThreshold": 1
                    },
                    "mgmtHttpPort": 9080,
                    "mgmtHttpsPort": 9443,
                    "nitroReadTimeout": 20,
                    "nsConfigDnsRec": false,
                    "nsCookieVersion": 0,
                    "nsDnsNameserver": "",
                    "nsEnableLabels": true,
                    "nsGateway": "192.168.1.1",
                    "nsHttp2ServerSide": "OFF",
                    "nsIP": "192.168.1.2",
                    "nsLbHashAlgo": {
                      "hashAlgorithm": "default",
                      "hashFingers": 500,
                      "required": false
                    },
                    "nsProtocol": "HTTP",
                    "nsSvcLbDnsRec": false,
                    "prometheusCredentialSecret": "",
                    "pullPolicy": "IfNotPresent",
                    "service": {
                      "annotations": {},
                      "spec": {}
                    },
                    "startupProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1
                    }
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "replicaCount": 1,
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerGslbController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalergslbcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "gslbController": {
                    "disableAPIServerCertVerify": false,
                    "entityPrefix": "",
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "kubernetesURL": "",
                    "localCluster": "",
                    "localRegion": "",
                    "logLevel": "Debug",
                    "openshift": true,
                    "pullPolicy": "IfNotPresent",
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "32m",
                        "memory": "128Mi"
                      }
                    },
                    "siteData": [
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteMask": "",
                        "siteName": "",
                        "sitePublicIp": "",
                        "siteRegion": ""
                      },
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteMask": "",
                        "siteName": "",
                        "sitePublicIp": "",
                        "siteRegion": ""
                      }
                    ]
                  },
                  "imagePullSecrets": [],
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsIP": "",
                    "nsPort": 443,
                    "nsProtocol": "HTTPS"
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerIngressController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscaleringresscontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "analyticsConfig": {
                    "distributedTracing": {
                      "enable": false,
                      "samplingRate": 100
                    },
                    "endpoint": {
                      "metrics": {
                        "service": ""
                      },
                      "transactions": {
                        "service": ""
                      }
                    },
                    "required": false,
                    "timeseries": {
                      "auditLogs": {
                        "enable": false
                      },
                      "events": {
                        "enable": false
                      },
                      "metrics": {
                        "enable": false,
                        "enableNativeScrape": false,
                        "exportFrequency": 50,
                        "mode": "avro",
                        "schemaFile": "schema.json"
                      },
                      "port": 50002
                    },
                    "transactions": {
                      "enable": false,
                      "port": 50005
                    }
                  },
                  "exporter": {
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4",
                    "ports": {
                      "containerPort": 8888
                    },
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {},
                    "serviceMonitorExtraLabels": {}
                  },
                  "extraVolumes": [],
                  "hostAliases": [],
                  "imagePullSecrets": [],
                  "ingressController": {
                    "clusterName": "",
                    "defaultSSLCertSecret": "",
                    "defaultSSLSNICertSecret": "",
                    "disableAPIServerCertVerify": false,
                    "disableOpenshiftRoutes": true,
                    "enableLivenessProbe": true,
                    "enableReadinessProbe": true,
                    "entityPrefix": "",
                    "extraVolumeMounts": [],
                    "ignoreNodeExternalIP": false,
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "ingressClass": [],
                    "ipam": false,
                    "jsonLog": false,
                    "kubernetesURL": "",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "logLevel": "INFO",
                    "logProxy": "",
                    "multiClusterPrefix": "",
                    "namespaceLabels": "",
                    "nodeLabels": "",
                    "nodeWatch": false,
                    "nsValidateCert": {},
                    "nsncPbr": false,
                    "openshift": true,
                    "optimizeEndpointBinding": false,
                    "podIPsforServiceGroupMembers": false,
                    "profileHttpFrontend": {},
                    "profileSslFrontend": {},
                    "profileTcpFrontend": {},
                    "pullPolicy": "IfNotPresent",
                    "rbacRole": false,
                    "readinessProbe": {
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 60,
                      "successThreshold": 1
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "32m",
                        "memory": "128Mi"
                      }
                    },
                    "routeLabels": "",
                    "serviceClass": [],
                    "setAsDefaultIngressClass": false,
                    "updateIngressStatus": true
                  },
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsConfigDnsRec": false,
                    "nsCookieVersion": 0,
                    "nsDnsNameserver": "",
                    "nsHttp2ServerSide": "OFF",
                    "nsIP": "",
                    "nsLbHashAlgo": {
                      "hashAlgorithm": "default",
                      "hashFingers": 256,
                      "required": false
                    },
                    "nsPort": 443,
                    "nsProtocol": "HTTPS",
                    "nsSNIPs": [],
                    "nsSvcLbDnsRec": false,
                    "nsVIP": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerIpamController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscaleripamcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "clusterName": "",
                  "image": "quay.io/netscaler/netscaler-ipam-controller@sha256:7c25b59a2e1df0e4f5615a779a9bea939c60a3eca065cb6c2295daf0d03831d8",
                  "localIPAM": {
                    "vipRange": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "reuseIngressVIP": true,
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerObservabilityExporter",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalerobservabilityexporter-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "elasticsearch": {
                    "enabled": false,
                    "server": "http://elasticsearch:9200"
                  },
                  "headless": false,
                  "image": "quay.io/netscaler/netscaler-observability-exporter@sha256:626e3a954c67bbb910477352dc740aaab21ff1c1cfb03354becf50beab17d4a4",
                  "jsonTransRateLimiting": {
                    "enabled": false,
                    "limit": 100,
                    "queueLimit": 1000,
                    "window": 5
                  },
                  "kafka": {
                    "auditLogs": "no",
                    "broker": "",
                    "dataFormat": "AVRO",
                    "enabled": false,
                    "events": "no",
                    "topic": "HTTP"
                  },
                  "nodePortRequired": false,
                  "nodeSelector": {},
                  "nsoeLogLevel": "INFO",
                  "nstracing": {
                    "enabled": false,
                    "server": "zipkin:9411/api/v1/spans"
                  },
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "replicas": 1,
                  "resources": {},
                  "splunk": {
                    "authToken": "",
                    "enabled": false,
                    "indexPrefix": "adc_nsoe",
                    "server": "http://splunk:8088"
                  },
                  "timeseries": {
                    "enabled": false
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-operator@sha256:e367c8c8ba6171e9ef5dd496592a5c62f4aa764f93d0cb78835aa29024c063d0
          createdAt: "2025-04-15T08:51:39Z"
          description: This is an operator to install Cloud Native portfolio provided
            by NetScaler
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler Cloud Controller for NetScaler VPX/MPX/SDX
              with Azure Redhat OpenShift in Azure.
            displayName: Netscaler Cloud Controller
            kind: NetscalerCloudController
            name: netscalercloudcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler CPX with Ingress Controller.
            displayName: Netscaler CPX With Ingress Controller
            kind: NetscalerCpxWithIngressController
            name: netscalercpxwithingresscontrollers.netscaler.com
            version: v1
          - description: To install NetScaler GSLB Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler GSLB Controller
            kind: NetscalerGslbController
            name: netscalergslbcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler Ingress Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler Ingress Controller
            kind: NetscalerIngressController
            name: netscaleringresscontrollers.netscaler.com
            version: v1
          - description: To install NetScaler IPAM Controller for NetScaler CPX/VPX/MPX/SDX.
            displayName: Netscaler IPAM Controller
            kind: NetscalerIpamController
            name: netscaleripamcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler Observability Exporter for NetScaler.
            displayName: Netscaler Observability Exporter
            kind: NetscalerObservabilityExporter
            name: netscalerobservabilityexporters.netscaler.com
            version: v1
        description: |
          NetScaler provides various products to empower Developer/DevOps managing your applications in OpenShift Cluster. Using this operator you can deploy below controllers for NetScaler appliance:
          1. [NetScaler Cloud Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. [NetScaler Ingress Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/unified-ingress-deployment)
          3. [NetScaler CPX with Ingress Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/two-tier-deployment)
          4. [NetScaler GSLB Controller ](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/gslb/gslb)
          5. [NetScaler Observability Exporter](https://docs.netscaler.com/en-us/netscaler-observability-exporter/deploy-coe)
          6. [NetScaler IPAM Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/ipam-installation.html)

          ### Installation
          Refer [installation instructions](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-operator-openshift).

          This operator version contains:
          1. NetScaler Ingress Controller version [2.2.10](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.2.10)
          2. NetScaler CPX version [14.1-25.111](https://docs.netscaler.com/en-us/updates?product=NetScaler-ADC-%28includes-NetScaler-Gateway%29&version=14.1&build=25.56)
          3. NetScaler GSLB Controller version [2.2.10](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.2.10)
          4. NetScaler Observability Exporter version [1.10.001](https://github.com/citrix/citrix-helm-charts/releases/tag/nsoe-v1.10.001)
          5. NetScaler Metrics Exporter [1.4.9](https://github.com/citrix/citrix-adc-metrics-exporter/releases/tag/1.4.9)
          6. NetScaler IPAM Controller [2.1.2](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.3.15)

          ### Support

          For debugging issues, please use [NetScaler Kubernetes Plugin](https://github.com/netscaler/modern-apps-toolkit/tree/main/netscaler-plugin).

          For questions and support, the following channels are available:
          - [NetScaler community portal](https://community.netscaler.com/s/)
          - [Contact us](https://www.netscaler.com/contact)
        displayName: NetScaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        - ADC
        - Kubernetes
        - OpenShift
        - Citrix
        - Controller
        - Observability
        - Metrices
        links:
        - name: Netscaler Operator
          url: https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-operator-openshift
        maintainers:
        - email: priyanka.sharma@cloud.com
          name: Priyanka Sharma
        - email: arijit.ray@cloud.com
          name: Arijit Ray
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: NetScaler
          url: https://netscaler.com
        relatedImages:
        - quay.io/netscaler/netscaler-cpx@sha256:232596eaacda68c0e2e210f26b64cedc8d70ae12c5e8ae2e1a66896a479ae4f8
        - quay.io/netscaler/netscaler-observability-exporter@sha256:626e3a954c67bbb910477352dc740aaab21ff1c1cfb03354becf50beab17d4a4
        - quay.io/netscaler/netscaler-ipam-controller@sha256:7c25b59a2e1df0e4f5615a779a9bea939c60a3eca065cb6c2295daf0d03831d8
        - quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f
        - quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:89f44c45373da73a51b0b9a02de7efe519dc2d5343b65fe382526cccc92f8f5c
        - registry.connect.redhat.com/citrix/netscaler-operator@sha256:e367c8c8ba6171e9ef5dd496592a5c62f4aa764f93d0cb78835aa29024c063d0
        - quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4
        version: 3.2.0
      entries:
      - name: netscaler-operator.v3.2.0
        version: 3.2.0
      - name: netscaler-operator.v3.1.2
        version: 3.1.2
      - name: netscaler-operator.v3.1.1
        version: 3.1.1
      - name: netscaler-operator.v3.0.1
        version: 3.0.1
      name: stable
    - currentCSV: netscaler-operator.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCloudController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netscaler-operator",
                    "app.kubernetes.io/instance": "nscc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscalercloudcontroller",
                    "app.kubernetes.io/part-of": "netscaler-operator"
                  },
                  "name": "nscc"
                },
                "spec": {
                  "azure": {
                    "clientsecret": "",
                    "image": "quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb",
                    "location": "",
                    "pullPolicy": "IfNotPresent",
                    "resourcegroupname": "",
                    "resources": {},
                    "subnetname": "",
                    "subscriptionid": "",
                    "tenantid": "",
                    "vnetname": ""
                  }
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerGslbController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netscaler-operator",
                    "app.kubernetes.io/instance": "nsgc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscalergslbcontroller",
                    "app.kubernetes.io/part-of": "netscaler-operator"
                  },
                  "name": "nsgc"
                },
                "spec": {
                  "gslbController": {
                    "acceptLicense": "Yes",
                    "cleanupGSLBSiteConfig": false,
                    "disableAPIServerCertVerify": false,
                    "entityPrefix": "",
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:88114c518ecceb54b000f38bfc330efe184bfeb5b5df0ee308950cf29e7a777f",
                    "imagePullSecrets": [],
                    "kubernetesURL": "",
                    "localCluster": "",
                    "localRegion": "",
                    "logLevel": "Debug",
                    "openshift": true,
                    "pullPolicy": "IfNotPresent",
                    "siteData": [
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteName": "",
                        "siteRegion": ""
                      },
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteName": "",
                        "siteRegion": ""
                      }
                    ]
                  },
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsIP": "",
                    "nsPort": 443,
                    "nsProtocol": "HTTPS"
                  },
                  "resources": {
                    "limits": {},
                    "requests": {
                      "cpu": "32m",
                      "memory": "128Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-operator@sha256:3f71f9a46d7b29e3da98a301965f635c03a022b5535b586d7d5c699341f68193
          createdAt: "2024-07-31T06:33:49Z"
          description: This is an operator to install Cloud Native portfolio provided
            by NetScaler
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler Cloud Controller for NetScaler VPX/MPX/SDX
              with Azure Redhat OpenShift in Azure.
            displayName: Netscaler Cloud Controller
            kind: NetscalerCloudController
            name: netscalercloudcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler GSLB Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler GSLB Controller
            kind: NetscalerGslbController
            name: netscalergslbcontrollers.netscaler.com
            version: v1
        description: |-
          NetScaler provides various products to empower Developer/DevOps managing your applications in OpenShift Cluster. Using this operator you can deploy below controllers for NetScaler appliance:
          1. [NetScaler Cloud Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. [NetScaler GSLB Controller ](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/gslb-openshift-operator)

          ### Installation
          Refer [installation instructions](https://github.com/netscaler/netscaler-k8s-ingress-controller/blob/master/docs/deploy/deploy-ns-operator.md).

          This operator version contains:
          1. NetScaler Cloud Controller version [1.2.0](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. NetScaler GSLB Controller version [1.43.7](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v1.43.7)

          ### Support

          For debugging issues, please use [NetScaler Kubernetes Plugin](https://github.com/netscaler/modern-apps-toolkit/tree/main/netscaler-plugin).

          For questions and support, the following channels are available:
          - [NetScaler community portal](https://community.netscaler.com/s/)
          - [Contact us](https://www.netscaler.com/contact)
        displayName: NetScaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        links:
        - name: Netscaler Operator
          url: https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/gslb-openshift-operator
        maintainers:
        - email: priyanka.sharma@cloud.com
          name: Priyanka Sharma
        - email: arijit.ray@cloud.com
          name: Arijit Ray
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: NetScaler
          url: https://netscaler.com
        relatedImages:
        - quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:88114c518ecceb54b000f38bfc330efe184bfeb5b5df0ee308950cf29e7a777f
        - quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:23405618050a4ceea8ea0c748cc7d9d16dbb80f10e4c39f7e82014a3ac80cec9
        - registry.connect.redhat.com/citrix/netscaler-operator@sha256:3f71f9a46d7b29e3da98a301965f635c03a022b5535b586d7d5c699341f68193
        version: 3.0.1
      entries:
      - name: netscaler-operator.v3.0.1
        version: 3.0.1
      name: tech-preview
    defaultChannel: stable
    packageName: netscaler-operator
    provider:
      name: NetScaler
      url: https://netscaler.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: open-liberty-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: open-liberty-operator.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4",
                  "expose": true,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/open-liberty-operator-controller@sha256:18f188e48725c785557c8edea46f69c464d6f81ab47860a3bb0ec978c20a26db
          createdAt: "2022-03-25T14:00:00Z"
          description: Deploy and manage applications running on Liberty
          olm.skipRange: <0.8.1
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          This advanced Operator can be used to deploy and manage Open Liberty applications with consistent, production-grade QoS. This operator is based on the [Runtime Component Operator](https://github.com/application-stacks/runtime-component-operator) and provides all of its capabilities in addition to Open Liberty specific features such as gathering traces and dumps (Day-2 operations) and easily configuring and managing the single sign-on information for your Open Liberty applications.

          Open Liberty Operator enables enterprise architects to govern the way their applications get deployed & managed in the cluster, while dramatically reducing the learning curve for developers to deploy into Kubernetes - allowing them to focus on writing the code!

          Here are some key features:

          #### Application Lifecyle
          You can deploy your Open Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the Operator will watch for any updates and will re-deploy the modified image.

          #### Custom RBAC
          This Operator is capable of using a custom ServiceAccount from the caller, allowing it to follow RBAC restrictions. By default it creates a ServiceAccount if one is not specified, which can also be bound with specific roles.

          #### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as: labels, annotations, and environment variables from a ConfigMap, a Secret or a value.

          #### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configuration, such as TLS settings, are also easily enabled.

          #### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto scaling to create (and delete) instances based on resource consumption.

          #### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you.
          This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart.
          We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          You can also easily configure and use a single storage for serviceability related Day-2 operations, such as gatherig server traces and dumps.

          #### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating kubernetes Secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to its required services without any intervention or interruption.

          #### Single Sign-On (SSO)
          Open Liberty provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows to easily configure and manage the single sign-on information for your applications.

          #### Exposing metrics to Prometheus
          Expose the Open Liberty application's metrics via the Prometheus Operator.
          You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `spec.monitoring` field to control configurations such as poll interval and security credentials.

          #### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply add the following configuration (to specify the volume size and the location to persist) to your OpenLibertyApplication CR:
          ``` storage: size: 2Gi mountPath: "/logs" ```

          #### Integration with OpenShift Serverless
          Deploy your serverless runtime component using a single toggle.  The Operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing your pod to automatically scale to 0 when it is idle.

          #### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and how they are connected.

          See our [**documentation**](https://github.com/OpenLiberty/open-liberty-operator/tree/main/doc/) for more information.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://github.com/OpenLiberty/open-liberty-operator/tree/main/doc
        - name: Open Liberty
          url: https://openliberty.io/
        - name: Open Liberty Container Images
          url: https://github.com/OpenLiberty/ci.docker
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: beta
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/open-liberty-operator-controller@sha256:18f188e48725c785557c8edea46f69c464d6f81ab47860a3bb0ec978c20a26db
        - registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4
        version: 0.8.1
      entries:
      - name: open-liberty-operator.v0.8.1
        version: 0.8.1
      - name: open-liberty-operator.v0.8.0
        version: 0.8.0
      name: beta2
    - currentCSV: open-liberty-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:d45c296d43dc4c4100f38bb736f073e04c8b4d72b68ddf1189e311aff8ad66e3
          createdAt: "2023-08-28T16:39:10Z"
          description: Deploy and manage containerized Liberty applications
          olm.skipRange: '>=0.8.0 <1.2.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:d45c296d43dc4c4100f38bb736f073e04c8b4d72b68ddf1189e311aff8ad66e3
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02
        version: 1.2.2
      entries:
      - name: open-liberty-operator.v1.2.2
        version: 1.2.2
      name: v1.2
    - currentCSV: open-liberty-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:d68bb66a8133702a198f7be655d7cff5f500bb620c460a81547971e00124ef73
          createdAt: "2024-08-01T20:19:13Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.3.3'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:d68bb66a8133702a198f7be655d7cff5f500bb620c460a81547971e00124ef73
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91
        version: 1.3.3
      entries:
      - name: open-liberty-operator.v1.3.3
        version: 1.3.3
      name: v1.3
    - currentCSV: open-liberty-operator.v1.4.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:67b5b46dfed008dab4ea7fd176d3bc31dae8940d72194a10aebb98b1dea9043e
          createdAt: "2025-07-18T20:45:06Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.4.4'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:67b5b46dfed008dab4ea7fd176d3bc31dae8940d72194a10aebb98b1dea9043e
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3
        version: 1.4.4
      entries:
      - name: open-liberty-operator.v1.4.4
        version: 1.4.4
      name: v1.4
    - currentCSV: open-liberty-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyPerformanceData",
                "metadata": {
                  "name": "openliberty-performance-data-sample"
                },
                "spec": {
                  "interval": 30,
                  "podName": "Specify_Pod_Name_Here",
                  "timespan": 240
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:c3b031fe14054f65d6612cc505b7681336873dc32ce1091451d8c7799034bc83
          createdAt: "2025-09-16T03:08:31Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.5.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server performance data
            displayName: OpenLibertyPerformanceData
            kind: OpenLibertyPerformanceData
            name: openlibertyperformancedata.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:c3b031fe14054f65d6612cc505b7681336873dc32ce1091451d8c7799034bc83
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8
        version: 1.5.0
      entries:
      - name: open-liberty-operator.v1.5.0
        version: 1.5.0
      name: v1.5
    defaultChannel: v1.5
    packageName: open-liberty-certified
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: RabbitMQ
      provider-url: https://rabbitmq.com
    name: rabbitmq-cluster-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-cluster-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rabbitmq.com/v1beta1",
                "kind": "RabbitmqCluster",
                "metadata": {
                  "name": "sample",
                  "namespace": "rabbitmq-system"
                },
                "spec": {
                  "replicas": 1,
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-07-10T12:52:23Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RabbitmqCluster is the Schema for the RabbitmqCluster API.
              Each instance of this object corresponds to a single RabbitMQ cluster.
            displayName: Rabbitmq Cluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: RabbitMQ Cluster Operator
        displayName: RabbitMQ Cluster Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        links:
        - name: Cluster Operator
          url: https://github.com/rabbitmq/cluster-operator
        maturity: alpha
        provider:
          name: RabbitMQ
          url: https://rabbitmq.com
        relatedImages:
        - registry.redhat.io/rhoso-operators/rabbitmq-cluster-rhel9-operator@sha256:e4b198236d70b7dfdd29d8ed45058dec3335b3530e0b676fcc862bfec2efce79
        version: 1.0.6
      entries:
      - name: rabbitmq-cluster-operator.v1.0.6
        version: 1.0.6
      - name: rabbitmq-cluster-operator.v1.0.5
        version: 1.0.5
      - name: rabbitmq-cluster-operator.v1.0.4
        version: 1.0.4
      - name: rabbitmq-cluster-operator.v1.0.3
        version: 1.0.3
      - name: rabbitmq-cluster-operator.v1.0.2
        version: 1.0.2
      - name: rabbitmq-cluster-operator.v1.0.1
        version: 1.0.1
      - name: rabbitmq-cluster-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: rabbitmq-cluster-operator
    provider:
      name: RabbitMQ
      url: https://rabbitmq.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fossul Framework
      provider-url: https://github.com/fossul/fossul
    name: fossul-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fossul-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n    \"kind\":
            \"Backup\",\n    \"metadata\": {\n      \"name\": \"mariadb-sample\"\n
            \   },\n    \"spec\": {\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\"\n    }\n  },\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n    \"kind\":
            \"Restore\",\n    \"metadata\": {\n      \"name\": \"mariadb-sample\"\n
            \   },\n    \"spec\": {\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\",\n      \"workflow_id\": \"xxxx\"\n    }\n  },        \n  {\n
            \   \"apiVersion\": \"fossul.io/v1\",\n    \"kind\": \"BackupConfig\",\n
            \   \"metadata\": {\n      \"name\": \"mariadb-sample\"\n    },\n    \"spec\":
            {\n      \"container_platform\": \"openshift\",\n      \"operator_controlled\":
            true,\n      \"auto_discovery\": false,\n      \"deployment_name\": \"mariadb\",\n
            \     \"deployment_type\": \"DeploymentConfig\",\n      \"job_retention\":
            50,\n      \"overwrite_pcv_on_restore\": true,\n      \"policies\": [\n
            \       {\n          \"policy\": \"hourly\",\n          \"retentionNumber\":
            3\n        },\n        {\n          \"policy\": \"daily\",\n          \"retentionNumber\":
            10\n        }\n      ],\n      \"pvc_deletion_timeout\": 300,\n      \"restore_to_new_pvc\":
            false,\n      \"snapshot_timeout\": 180,\n      \"storage_plugin\": \"csi.so\",\n
            \     \"app_plugin\": \"mariadb.so\"\n    }\n  },\n  {\n    \"apiVersion\":
            \"fossul.io/v1\",\n    \"kind\": \"BackupSchedule\",\n    \"metadata\":
            {\n      \"name\": \"mariadb-sample\"\n    },\n    \"spec\": {\n      \"cron_schedule\":
            \"59 23 * * *\",\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\"\n\n    }\n  },\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n
            \   \"kind\": \"Fossul\",\n    \"metadata\": {\n      \"name\": \"fossul-sample\"\n
            \   },\n    \"spec\": {\n      \"enable_routes\": false,\n      \"enable_virtualmachines\":
            false,\n      \"container_platform\": \"openshift\"\n    }\n  }\n]"
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/fossul/fossul-operator:v0.6.0
          createdAt: "2022-01-25T20:09:00Z"
          description: Fossul is a application centric container native backup and
            restore framework for databases or applications
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/fossul/fossul
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages application backup configurations
            kind: BackupConfig
            name: backupconfigs.fossul.io
            version: v1
          - description: Manages application backups
            kind: Backup
            name: backups.fossul.io
            version: v1
          - description: Manages application restores
            kind: Restore
            name: restores.fossul.io
            version: v1
          - description: Manages application backup scheduling
            kind: BackupSchedule
            name: backupschedules.fossul.io
            version: v1
          - description: Manages deployment of the Fossul Framework
            kind: Fossul
            name: fossuls.fossul.io
            version: v1
        description: "## Fossul Framework\nApplication backup and restore framework
          for container native applications on Kubernetes and OpenShift. Fossul provides
          application centric backup and recovery enabling integration with the application
          creating an application consistent backup. Since there are many applications
          and providers for storage or archive Fossul enables plugins that allow proper
          integration. Fossul provides plugins for postgresql, mariadb, mysql and
          mongo as well as plugin to integrate with CSI snapshots. Plugins can be
          written in any language. The Fossul operator deploys Fossul via the Fossul
          custom resource and additionally provides BackupConfig, Backup, BackupSchedule
          and Restore custom resources. Each user can configure and manage their own
          backups with simple custom resources that can be part of CI/CD or source
          control. Users can also opt to manage backups through Fossul CLI/API instead
          of the custom resources provided by the operator.\n\n## Installation\nIt
          is recommended to install the operator into the openshift-operators namespace
          which is the default location for operators. Once the operator is installed
          create a new project called ```fossul``` and create the Fossul custom resource
          under the fossul namespace. This will deploy the Fossul Framework. You can
          choose another namespace for the Fossul Framework but then also need to
          provide the optional 'fossul_namespace' parameter with the spec for all
          the additional custom resources. This is required so the Fossul operator
          can communicate with Fossul.\n\nOnce Fossul is deployed you can manage backups
          either through custom resources provided by the operator or the Fossul CLI/API.
          If using the customer resources, first create a BackupConfig custom resource
          for every application you want to backup. Once that is done, you can optionally
          create a backup by simply creating a Backup custom resource. You can also
          create a BackupSchedule custom resource using cron syntax which will schedule
          your backup to occur on a regular interval or do so via a Kubernetes job.
          Finally if a backup has been created you can perform a restore through a
          custom resource as well, providing the workflow id from the completed backup.
          If you are creating backups through the custom resource the workflow id
          will be appended to the spec once the backup completes. If you are using
          the Fossul CLI/API to create backups you will need to get the workflow id
          through the Fossul CLI/API directly using the backupList action.\n\nFor
          users of the cli, a cli pod is deployed with credentials and if desired
          you can create or manage backups from within that pod. To use the API you
          need the credentials stored in the fossul secret. The cli has already been
          pre-configured with these credentials. The API can also be optionally exposed
          via routes and API documentation is under ```/api/v1/index.html```.\n\n##
          Quick Setup\nWhen creating Fossul custom resources generally you will only
          need to update the ```deployment_name``` and ```deployment_type``` (DeploymentConfig,
          Deployment or VirtualMachine). All custom resources must be created in the
          namespace where the database or application exists. \n\n### Deploy Fossul
          in fossul namespace\n```$ oc new-project fossul```\n<pre>\n$ vi fossul.yaml\nkind:
          Fossul\napiVersion: fossul.io/v1\nmetadata:\n  name: fossul-sample\n  namespace:
          fossul\nspec: {}\n</pre>\n```$ oc create -f fossul.yaml```\n\n### Deploy
          Mariadb Database\nMake sure you create all Fossul custom resources in the
          same namespace as the database, application or virtual machine\n\n### Create
          MariaDB Fossul BackupConfig\nCreates a backup configuration for MariaDB
          or MySQL databases\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true\n  auto_discovery:
          false\n  deployment_name: mariadb\n  deployment_type: DeploymentConfig\n
          \ job_retention: 50\n  overwrite_pcv_on_restore: true\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc: false\n  snapshot_timeout:
          180\n  storage_plugin: csi.so\n  app_plugin: mariadb.so \n</pre>\n```$ oc
          create -f backupconfig.yaml```\n\n### Create PostgreSQL Fossul BackupConfig\nCreates
          a backup config for PostgreSQL databases, you need to ensure the user defined
          in secret has backup permissions\n\n```$ vi backupconfig.yaml```\n<pre>\nkind:
          BackupConfig\napiVersion: fossul.io/v1\nmetadata:\n  name: postgresql-sample\n
          \ namespace: databases\nspec:\n  container_platform: openshift\n  operator_controlled:
          true    \n  auto_discovery: false\n  deployment_name: postgresql \n  deployment_type:
          DeploymentConfig\n  job_retention: 50\n  overwrite_pcv_on_restore: true\n
          \ policies:\n    - policy: hourly\n      retentionNumber: 3\n    - policy:
          daily\n      retentionNumber: 10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc:
          false\n  snapshot_timeout: 180\n  storage_plugin: csi.so\n  app_plugin:
          postgres.so \n</pre>\n```$ oc create -f backupconfig.yaml```\n\n### Create
          Kubevirt Fossul BackupConfig\nCreates a backup configuration for virtual
          machines\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: rhel8-vm-sample\n  namespace: virtualmachines\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true    \n  auto_discovery:
          false\n  deployment_name: rhel8-vm\n  deployment_type: VirtualMachine\n
          \ job_retention: 50\n  overwrite_pcv_on_restore: true\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc: false\n  snapshot_timeout:
          180\n  storage_plugin: csi.so\n  app_plugin: kubevirt.so \n</pre>\n```$
          oc create -f backupconfig.yaml```\n\n### Create Tar Fossul BackupConfig\nCreates
          a backup config using tar to backup folders and files where csi snapshots
          are not desired\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: tar-sample\n  namespace: databases\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true    \n  auto_discovery:
          false\n  deployment_name: anyname\n  job_retention: 50\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  backup_source_paths: /opt/app-root/src/test\n  storage_plugin: tar.so\n</pre>\n```$
          oc create -f backupconfig.yaml```\n\n### Create BackupSchedule\nA backup
          can be scheduled per policy, defined by backup configuration using cron
          syntax\n\n```$ vi backupschedule.yaml```\n<pre>\nkind: BackupSchedule\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ cron_schedule: '59 23 * * *'\n  deployment_name: mariadb\n  policy: daily\n</pre>\n```$
          oc create -f backupschedule.yaml```\n\n### Create Backup\nA backup will
          be created according to policy and deleted either manually or as defined
          in retention policy\n\n```$ vi backup.yaml```\n<pre>\nkind: Backup\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ deployment_name: mariadb\n  policy: daily\n</pre>\n```$ oc create -f backup.yaml```\n\n###
          Perform Restore\nA restore requires a workflow_id located in the backup
          spec, upon a successful restore the custom resource will deleted automatically\n\n```$
          vi restore.yaml```\n<pre>\nkind: Restore\napiVersion: fossul.io/v1\nmetadata:\n
          \ name: mariadb-sample\n  namespace: databases\nspec:\n  deployment_name:
          mariadb\n  policy: daily\n  workflow_id: xxxx\n</pre>\n```$ oc create -f
          restore.yaml```\n"
        displayName: Fossul Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - application
        - consistent
        links:
        - name: Fossul Framework Community
          url: https://github.com/fossul/fossul
        maintainers:
        - email: ktenzer@redhat.com
          name: Keith Tenzer
        maturity: alpha
        provider:
          name: Fossul Framework
          url: https://github.com/fossul/fossul
        relatedImages:
        - quay.io/fossul/fossul-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: fossul-operator.v0.6.0
        version: 0.6.0
      - name: fossul-operator.v0.5.0
        version: 0.5.0
      - name: fossul-operator.v0.4.0
        version: 0.4.0
      - name: fossul-operator.v0.3.0
        version: 0.3.0
      - name: fossul-operator.v0.2.0
        version: 0.2.0
      - name: fossul-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: fossul-operator
    provider:
      name: Fossul Framework
      url: https://github.com/fossul/fossul
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Weaveworks
      provider-url: ""
    name: tf-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tf-controller.v0.9.0-rc.8
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"infra.contrib.fluxcd.io/v1alpha1\",\n
            \   \"kind\": \"Terraform\",\n    \"metadata\": {\n        \"name\": \"helloworld-tf\",\n
            \       \"namespace\": \"flux-system\"\n    },\n    \"spec\": {\n        \"interval\":
            \"5m\",\n        \"approvePlan\": \"auto\",\n        \"path\": \"./\",\n
            \       \"sourceRef\": {\n          \"kind\": \"GitRepository\",\n          \"name\":
            \"helloworld\",\n          \"namespace\": \"flux-system\"\n        }      \n
            \   }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/weaveworks/tf-controller:v0.9.0-rc.8
          description: Weave TF-controller is a Flux controller that brings GitOps
            to Terraform resources.
          operatorframework.io/suggested-namespace: flux-system
          support: Weaveworks
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Terraform
            displayName: Terraform
            kind: Terraform
            name: terraforms.infra.contrib.fluxcd.io
            version: v1alpha1
        description: "**Weave GitOps Terraform Controller** (aka Weave TF-controller)
          is a Flux controller for reconciling Terraform resources in the GitOps way.
          \nWith the power of Flux together with Terraform, TF-controller allows you
          to GitOps-ify infrastructure, \nand application resources, in the Kubernetes
          and Terraform universe, together at your own pace.\n\nTF-controller offers
          many GitOps models:\n  1. **GitOps Automation Model:** GitOps your Terraform
          resources from the provision steps to the enforcement steps, like a whole
          EKS cluster.\n  2. **Hybrid GitOps Automation Model:** GitOps parts of your
          existing infrastructure resources. For example, you have an existing EKS
          cluster. You can choose to GitOps only its nodegroup, or its security group.\n
          \ 3. **State Enforcement Model:** You have a TFSTATE file, and you'd like
          to use GitOps enforce it, without changing anything else.\n  4. **Drift
          Detection Model:** You have a TFSTATE file, and you'd like to use GitOps
          just for drift detection, so you can decide to do things later when a drift
          occurs.\n\nFeatures\n  * GitOps Automation for Terraform\n  * Drift detection\n
          \ * Plan and Manual Approve"
        displayName: Weave GitOps Terraform Controller
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - weave
        - flux
        - cd
        - continuous delivery
        - terraform
        - gitops
        links:
        - name: Weave GitOps Terraform Controller
          url: https://github.com/weaveworks/tf-controller
        - name: Documentation
          url: https://weaveworks.github.io/tf-controller/
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Weaveworks
        relatedImages:
        - ghcr.io/weaveworks/tf-controller:v0.9.0-rc.8
        version: 0.9.0-rc.8
      entries:
      - name: tf-controller.v0.9.0-rc.8
        version: 0.9.0-rc.8
      - name: tf-controller.v0.9.0-rc.3
        version: 0.9.0-rc.3
      name: stable
    defaultChannel: stable
    packageName: tf-controller
    provider:
      name: Weaveworks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Randoli
      provider-url: https://www.randoli.io
    name: app-insights-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: app-insights-operator.v1.17.5-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "appinsights.randoli/v1beta1",
                "kind": "RandoliAgent",
                "metadata": {
                  "name": "randoli",
                  "namespace": "randoli-agents"
                },
                "spec": {
                  "costManagement": {
                    "opencost": {
                      "exporter": {
                        "image": {
                          "tag": "1.112.1"
                        }
                      }
                    },
                    "vpaOperator": {
                      "enabled": false
                    }
                  },
                  "logs": {
                    "vector": {
                      "agent": {
                        "containerSecurityContext": {
                          "privileged": true
                        }
                      }
                    }
                  },
                  "network": {
                    "standaloneConsole": {
                      "enable": true
                    }
                  },
                  "observability": {
                    "netobserv": {
                      "enabled": false
                    },
                    "otel": {
                      "operator": {
                        "enabled": false
                      }
                    }
                  },
                  "resources": {
                    "limits": {
                      "memory": "2Gi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "512Mi"
                    }
                  },
                  "securityContext": {
                    "fsGroup": 1001
                  },
                  "tags": {
                    "costManagement": false,
                    "observability": false
                  }
                }
              }
            ]
          categories: Developer Tools
          containerImage: docker.io/randoli/app-insights-operator:1.17.5-4@sha256:22e7150c29557f2a71186389e5d7d60bfe82a4ea022e1403033c6ad61bc95f7a
          createdAt: "2025-09-23T19:45:43Z"
          description: Randoli Agent connects OpenShift clusters to Randoli Observability
            & Cost Management.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[]'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Randoli
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: RandoliAgent
            name: randoliagents.appinsights.randoli
            version: v1beta1
        description: |-
          Randoli Agent connects OpenShift clusters to Randoli Observability & Cost Management platforms.

          This operator provides a container-based agent to be installed in your OpenShift clusters for connecting to Randoli Observability & Cost Management platforms

          **Prerequisites**

          If you are using the Randoli Observability platform, you will need to install the following Operators:
          - Red Hat build of OpenTelemetry Operator
          - Network Observability

          If you are using the Randoli Cost Management platform, you will need to install the following Operator:
          - VerticalPodAutoscaler

          For additional information please use our Openshift installation guide here: [https://docs.insights.randoli.io/agent/openshift](https://docs.insights.randoli.io/agent/openshift)

           **Randoli Observability for Kubernetes**

          [Product Page](https://www.randoli.io/product/kubernetes-observability)

          **Randoli Cost Management for Kubernetes**

          [Product Page](https://www.randoli.io/product/kubernetes-cost-management)
        displayName: Randoli Agent
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - developer tools
        - insights
        - monitoring
        links:
        - name: Randoli Observability for Kubernetes
          url: https://www.randoli.io/product/kubernetes-observability
        - name: Randoli Cost Management for Kubernetes
          url: https://www.randoli.io/product/kubernetes-cost-management
        maintainers:
        - email: contact@randoli.ca
          name: Randoli Inc.
        maturity: beta
        minKubeVersion: 1.24.0
        provider:
          name: Randoli
          url: https://www.randoli.io
        relatedImages:
        - docker.io/randoli/app-insights-operator:1.17.5-4@sha256:22e7150c29557f2a71186389e5d7d60bfe82a4ea022e1403033c6ad61bc95f7a
        version: 1.17.5-4
      entries:
      - name: app-insights-operator.v1.17.5-4
        version: 1.17.5-4
      - name: app-insights-operator.v1.17.5-3
        version: 1.17.5-3
      - name: app-insights-operator.v1.17.5-2
        version: 1.17.5-2
      - name: app-insights-operator.v1.17.5-1
        version: 1.17.5-1
      - name: app-insights-operator.v1.17.3-1
        version: 1.17.3-1
      - name: app-insights-operator.v1.17.2-1
        version: 1.17.2-1
      - name: app-insights-operator.v1.17.1-1
        version: 1.17.1-1
      - name: app-insights-operator.v1.17.0-2
        version: 1.17.0-2
      - name: app-insights-operator.v1.17.0-1
        version: 1.17.0-1
      name: stable
    defaultChannel: stable
    packageName: app-insights-operator
    provider:
      name: Randoli
      url: https://www.randoli.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ProphetStor Data Services, Inc.
      provider-url: ""
    name: federatorai-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: federatorai.v4.7.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "federatorai.containers.ai/v1alpha1",
                "kind": "AlamedaService",
                "metadata": {
                  "name": "my-alamedaservice"
                },
                "spec": {
                  "alamedaAi": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "10Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "alamedaInfluxdb": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "100Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "fedemeterInfluxdb": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "10Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "imageLocation": "registry.connect.redhat.com/prophetstor",
                  "storages": [
                    {
                      "accessMode": ["ReadWriteOnce"],
                      "size": "2Gi",
                      "type": "pvc",
                      "usage": "log"
                    }
                  ],
                  "version": "v4.7.1-b1760"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: registry.connect.redhat.com/prophetstor/federatorai-operator@sha256:bdd87f6b7d7ba9d12ed8ddb9c5151e5a89781e88d7a02b73814b16a766968047
          createdAt: "2021-10-12T01:14:00Z"
          description: Federator.ai Operator provides easy configuration and management
            of AI-based Kubernetes resource orchestrator
          operatorframework.io/suggested-namespace: federatorai
          repository: https://access.redhat.com/containers/#/registry.connect.redhat.com/prophetstor/federatorai-operator
          support: ProphetStor Data Services, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of Federator.ai.
            displayName: AlamedaService
            kind: AlamedaService
            name: alamedaservices.federatorai.containers.ai
            version: v1alpha1
        description: |-
          Federator.ai helps enterprises optimize cloud resources, maximize application performance, and save significant cost without excessive over-provisioning or under-provisioning of resources, meeting the service-level requirements of their applications.

          Enterprises often lack understanding of the resources needed to support their applications. This leads to either excessive over-provisioning or under-provisioning of resources (CPU, memory, storage). Using machine learning, Federator.ai determines the optimal cloud resources needed to support any workload on OpenShift and helps users find the best-cost instances from cloud providers for their applications.


          **Multi-layer workload prediction**

          Using machine learning and math-based algorithms, Federator.ai predicts containerized application and cluster node resource usage as the basis for resource recommendations at application level as well as at cluster node level. Federator.ai supports prediction for both physical/virtual CPUs and memories.


          **Auto-scaling via resource recommendation**

          Federator.ai utilizes the predicted resource usage to recommend the right number and size of pods for applications. Integrated with Datadog's WPA, applications are automatically scaled to meet the predicted resource usage.


          **Application-aware recommendation execution**

          Optimizing the resource usage and performance goals, Federator.ai uses application specific metrics for workload prediction and pod capacity estimation to auto-scale the right number of pods for best performance without overprovisioning.


          **Multi-cloud Cost Analysis**

          With resource usage prediction, Federator.ai analyzes potential cost of a cluster on different public cloud providers. It also recommend appropriate cluster nodes and instance types based on resource usage.


          **Custom Datadog/Sysdig Dashboards**

          Predefined custom Datadog/Sysdig Dashboards for workload prediction/recommendation visualization for cluster nodes and applications.


          **Additional resources**

          Want more product information? Explore detailed information about using this product and where to find additional help.

          * [Federator.ai Datasheet](https://prophetstor.com/wp-content/uploads/datasheets/Federator.ai.pdf)
          * [Quick Start Guide](https://prophetstor.com/wp-content/uploads/documentation/Federator.ai/Latest%20Version/ProphetStor%20Federator.ai%20Quick%20Installation%20Guide.pdf)
          * [Installation Guide](https://prophetstor.com/wp-content/uploads/2021/08/ProphetStor-Federator.ai-v4.7.0-Installation-Guide-v1.0.pdf)
          * [User Guide](https://prophetstor.com/wp-content/uploads/2021/08/Federator.ai-4.7.0-User-Guide.pdf)
          * [Release Notes](https://prophetstor.com/wp-content/uploads/2021/08/Federator.ai-4.7.0-Release-Notes.pdf)
          * [Company Information](https://www.prophetstor.com/)


          ### Prerequisite
          *  The [Openshift](https://www.openshift.com) version 4.4.x or later if using OpenShift platform.
          *  The [Kubernetes](https://kubernetes.io/) version 1.16 or later if using Kubernetes environment.
        displayName: Federator.ai
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - AI
        - Resource Orchestration
        - NoOps
        links:
        - name: Website
          url: https://prophetstor.com/federator-ai-2/
        - name: Quickstart guide
          url: https://prophetstor.com/wp-content/uploads/documentation/Federator.ai/Latest%20Version/ProphetStor%20Federator.ai%20Quick%20Installation%20Guide.pdf
        maintainers:
        - email: support@prophetstor.com
          name: ProphetStor Data Services, Inc.
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: ProphetStor Data Services, Inc.
        relatedImages:
        - registry.connect.redhat.com/prophetstor/federatorai-operator@sha256:bdd87f6b7d7ba9d12ed8ddb9c5151e5a89781e88d7a02b73814b16a766968047
        - registry.connect.redhat.com/prophetstor/alameda-recommender-ubi@sha256:13e941df9252cea3024ecb5e2247a696b64928015c1a3a4a04ee4e362388b1e4
        - registry.connect.redhat.com/prophetstor/federatorai-data-adapter@sha256:adf5d655fa030dc05133516d45642d7282fdd66e0aa8e7a7331ed614a41d3fe0
        - registry.connect.redhat.com/prophetstor/alameda-operator-ubi@sha256:9a863394cb1c89228223b26addb0d12bb91a97233fe2d5d19ecd5998e0cfe0d6
        - registry.connect.redhat.com/prophetstor/federatorai-agent-ubi@sha256:cfadc84f31f3776ded15d4d8dda92e00b7d9b1bce0ff5440277176292f9304d3
        - registry.connect.redhat.com/prophetstor/federatorai-agent-preloader@sha256:66daea6193e399f8363df5ba42e1fcd643db6df3a9075dbcd8c3464fdc91e776
        - registry.connect.redhat.com/prophetstor/federatorai-rest-ubi@sha256:3e92528615543ea900bf38387c4eb8439b60759be5be2ac5311a9dbaafc7026b
        - registry.connect.redhat.com/prophetstor/alameda-rabbitmq@sha256:b396d9aa0470e593a36f5799c04fc3c9e59b5723f703f4b019f7ef92528e6741
        - registry.connect.redhat.com/prophetstor/alameda-notifier-ubi@sha256:1f75034fb78d62924a87cb4a5435661b4afa990d6960431e784029742fbae0da
        - registry.connect.redhat.com/prophetstor/alameda-ai-dispatcher@sha256:6e0dc1136b0f8e5971d92b95ce0d6dfac475044c073bada50f23b315f4478ae4
        - registry.connect.redhat.com/prophetstor/alameda-ai@sha256:9293298bb2a72d0b0764050b6f6d8da707b46a83363a3b6aa7be60f512e2ec61
        - registry.connect.redhat.com/prophetstor/alameda-analyzer-ubi@sha256:6117ac4692793d49bc0204f13091dec119bdbc7b169e9deb86c243aca00ded27
        - registry.connect.redhat.com/prophetstor/federatorai-dashboard-backend@sha256:022dbf6696edb8eaf389fdbd1dff14c40a51c237c01f8a02a7ea65c60cd59f38
        - registry.connect.redhat.com/prophetstor/alameda-datahub-ubi@sha256:3819d6acce8d376dbe3b12469be2528d0ce9ac95c2a86507aeb76b6e3a5f5f38
        - registry.connect.redhat.com/prophetstor/alameda-executor-ubi@sha256:4512e3be7a89d0ca9f120b68f6ab9766d000bf2b01b279cd55b2dbc433aacfb6
        - registry.connect.redhat.com/prophetstor/fedemeter-api-ubi@sha256:451456207fe8d16b9824b3f64d00645766dcccbc8d685b6942e6c00d0d0c627e
        - registry.connect.redhat.com/prophetstor/federatorai-dashboard-frontend@sha256:1c1147425ca6302103363a1a3bcf27baeef13b10ec6780ef266b7f52fe27ffeb
        - registry.connect.redhat.com/prophetstor/fedemeter-influxdb@sha256:36904f003e8e29d4a9fd738655f401fbe840d60158d7d820cda79a09858dc2c2
        - registry.connect.redhat.com/prophetstor/alameda-influxdb@sha256:485c2119a6a56807e28ad03b2087129950118b55531126a66bb414063052ae76
        version: 4.7.1-1
      entries:
      - name: federatorai.v4.7.1-1
        version: 4.7.1-1
      - name: federatorai.v4.6.1-4
        version: 4.6.1-4
      name: stable
    defaultChannel: stable
    packageName: federatorai-certified
    provider:
      name: ProphetStor Data Services, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-route53resolver-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-route53resolver-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "route53resolver.services.k8s.aws/v1alpha1",
                "kind": "ResolverEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "route53resolver.services.k8s.aws/v1alpha1",
                "kind": "ResolverRule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/route53resolver-controller:1.1.1
          createdAt: "2025-09-20T01:21:08Z"
          description: AWS Route53Resolver controller is a service controller for
            managing Route53Resolver resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ResolverEndpoint represents the state of an AWS route53resolver
              ResolverEndpoint resource.
            displayName: ResolverEndpoint
            kind: ResolverEndpoint
            name: resolverendpoints.route53resolver.services.k8s.aws
            version: v1alpha1
          - description: ResolverRule represents the state of an AWS route53resolver
              ResolverRule resource.
            displayName: ResolverRule
            kind: ResolverRule
            name: resolverrules.route53resolver.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Route53Resolver resources in AWS from within your Kubernetes cluster.

          **About Amazon Route53Resolver**

          Amazon Route 53 Resolver responds recursively to DNS queries from AWS resources for public records, Amazon VPC-specific DNS names, and Amazon Route 53 private hosted zones, and is available by default in all VPCs.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Route53Resolver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - route53resolver
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Route53Resolver Developer Resources
          url: https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/resolver.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: route53resolver maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/route53resolver-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-route53resolver-controller.v1.1.1
        version: 1.1.1
      - name: ack-route53resolver-controller.v1.0.15
        version: 1.0.15
      - name: ack-route53resolver-controller.v1.0.14
        version: 1.0.14
      - name: ack-route53resolver-controller.v1.0.13
        version: 1.0.13
      - name: ack-route53resolver-controller.v1.0.11
        version: 1.0.11
      - name: ack-route53resolver-controller.v1.0.10
        version: 1.0.10
      - name: ack-route53resolver-controller.v1.0.9
        version: 1.0.9
      - name: ack-route53resolver-controller.v1.0.7
        version: 1.0.7
      - name: ack-route53resolver-controller.v1.0.6
        version: 1.0.6
      - name: ack-route53resolver-controller.v1.0.5
        version: 1.0.5
      - name: ack-route53resolver-controller.v1.0.3
        version: 1.0.3
      - name: ack-route53resolver-controller.v1.0.2
        version: 1.0.2
      - name: ack-route53resolver-controller.v1.0.0
        version: 1.0.0
      - name: ack-route53resolver-controller.v0.0.11
        version: 0.0.11
      - name: ack-route53resolver-controller.v0.0.10
        version: 0.0.10
      - name: ack-route53resolver-controller.v0.0.9
        version: 0.0.9
      - name: ack-route53resolver-controller.v0.0.8
        version: 0.0.8
      - name: ack-route53resolver-controller.v0.0.7
        version: 0.0.7
      - name: ack-route53resolver-controller.v0.0.4
        version: 0.0.4
      - name: ack-route53resolver-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-route53resolver-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: eventing-kogito
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eventing-kogito.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kogito.knative.dev/v1alpha1",
                "kind": "KogitoSource",
                "metadata": {
                  "name": "kogito-order-processing"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "kind": "Broker",
                      "name": "default"
                    }
                  },
                  "subject": {
                    "apiVersion": "serving.knative.dev/v1",
                    "kind": "Service",
                    "name": "serverless-workflow-order-processing"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: gcr.io/knative-releases/knative.dev/eventing-kogito/cmd/controller:v1.2.0
          createdAt: "2021-08-25T12:00:00Z"
          olm.targetNamespaces: knative-kogito
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/knative-sandbox/eventing-kogito
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoSource is the representation of a Knative Eventing
              Source for a Kogito custom Service.
            displayName: Kogito Source
            kind: KogitoSource
            name: kogitosources.kogito.knative.dev
            version: v1alpha1
        description: |-
          Knative Eventing Source for Kogito Services. This controller will enable easy configuration of any Kogito custom Service
          with the Knative Eventing platform. Please find more information about using this source in the [Official Kogito Documentation](https://docs.jboss.org/kogito/release/latest/html_single/#_kogito_knative_eventing_add_on)

          **Note:** The admission webhooks are not installed via OLM at this time. Because of that, please ensure that you create the `KogitoSource` objects with the subbject and sink information with the appropriate namespace.
        displayName: Kogito Knative Eventing Source
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless workflow
        - kogito
        - knative
        - dmn
        - bpmn
        - eventing
        - cloudevents
        - messaging
        links:
        - name: Source Repository
          url: https://github.com/knative-sandbox/eventing-kogito
        - name: Kogito Knative Add-On Documentation
          url: https://docs.jboss.org/kogito/release/latest/html_single/#_kogito_knative_eventing_add_on
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - gcr.io/knative-releases/knative.dev/eventing-kogito/cmd/controller:v1.2.0
        version: 1.2.0
      entries:
      - name: eventing-kogito.v1.2.0
        version: 1.2.0
      - name: eventing-kogito.v1.1.0
        version: 1.1.0
      name: alpha
    defaultChannel: alpha
    packageName: eventing-kogito
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KEDA Community
      provider-url: ""
    name: keda
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keda.v2.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "KedaController",
                "metadata": {
                  "name": "keda",
                  "namespace": "keda"
                },
                "spec": {
                  "watchNamespace": "",
                  "operator": {
                    "logLevel": "info",
                    "logEncoder": "console"
                  },
                  "metricsServer": {
                    "logLevel": "0"
                  },
                  "serviceAccount": {}
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledJob",
                "metadata": {
                  "name": "scaledjob-sample"
                },
                "spec": {
                  "failedJobsHistoryLimit": 5,
                  "jobTargetRef": {
                    "activeDeadlineSeconds": 600,
                    "backoffLimit": 6,
                    "completions": 1,
                    "parallelism": 1,
                    "template": null
                  },
                  "maxReplicaCount": 100,
                  "pollingInterval": 30,
                  "successfulJobsHistoryLimit": 5,
                  "triggers": [
                    {
                      "metadata": {
                        "property": "examle-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledObject",
                "metadata": {
                  "name": "example-scaledobject"
                },
                "spec": {
                  "cooldownPeriod": 300,
                  "maxReplicaCount": 100,
                  "minReplicaCount": 0,
                  "pollingInterval": 30,
                  "scaleTargetRef": {
                    "name": "example-deployment"
                  },
                  "triggers": [
                    {
                      "metadata": {
                        "property": "examle-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "TriggerAuthentication",
                "metadata": {
                  "name": "example-triggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "exmaple-secret-parameter"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ClusterTriggerAuthentication",
                "metadata": {
                  "name": "example-clustertriggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "exmaple-secret-parameter"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Cloud Provider
          certified: "false"
          containerImage: ghcr.io/kedacore/keda-olm-operator:2.7.1
          createdAt: "2022-05-11T00:00:00.000Z"
          description: '[DEPRECATED] Use Custom Metrics Autoscaler Operator instead'
          operatorframework.io/suggested-namespace: keda
          operators.operatorframework.io/builder: operator-sdk-v1.1.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/kedacore/keda-olm-operator
          support: KEDA Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              Represents an installation of a particular version of KEDA Controller.
            displayName: KedaController
            kind: KedaController
            name: kedacontrollers.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your jobs and what the triggers
            displayName: ScaledJob
            kind: ScaledJob
            name: scaledjobs.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your application and what the
              triggers are
            displayName: ScaledObject
            kind: ScaledObject
            name: scaledobjects.keda.sh
            version: v1alpha1
          - description: Describes authentication parameters
            displayName: TriggerAuthentication
            kind: TriggerAuthentication
            name: triggerauthentications.keda.sh
            version: v1alpha1
          - description: Describes cluster wide authentication parameters
            displayName: ClusterTriggerAuthentication
            kind: ClusterTriggerAuthentication
            name: clustertriggerauthentications.keda.sh
            version: v1alpha1
        description: "## Deprecated\n This KEDA community operator is deprecated,
          please use `Custom Metrics Autoscaler` Operator instead.\n## About the managed
          application\nKEDA is a Kubernetes-based event driven autoscaler.  KEDA can
          monitor event sources like Kafka, RabbitMQ, or cloud event sources and feed
          the metrics from those sources into the Kubernetes horizontal pod autoscaler.
          \ With KEDA, you can have event driven and serverless scale of deployments
          within any Kubernetes cluster.\n## About this Operator\nThe KEDA Operator
          deploys and manages installation of KEDA Controller in the cluster. Install
          this operator and follow installation instructions on how to install KEDA
          in you cluster.\n\n## Prerequisites for enabling this Operator\nPlease note
          that you can not run both KEDA 1.x and 2.x on the same Kubernetes cluster.
          You need to uninstall KEDA 1.x first, in order to install and use KEDA 2.x.\n\nDon't
          forget to uninstall KEDA v1 CRDs as well, to ensure that, please run:\n\n```
          oc delete crd scaledobjects.keda.k8s.io ```\n\n``` oc delete crd triggerauthentications.keda.k8s.io
          \ ```\n## How to install KEDA in the cluster\nThe installation of KEDA is
          triggered by the creation of `KedaController` resource, please refer to
          [KedaController Example](https://github.com/kedacore/keda-olm-operator#kedacontroller-spec)
          for more deatils on available options.\n\nOnly resource named `keda` in
          namespace `keda` will trigger the installation, reconfiguration or removal
          of the KEDA Controller resource.\n\nThere could be only one KEDA Controller
          in the cluster. \n"
        displayName: '[DEPRECATED] KEDA'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keda
        - scaling
        - event
        links:
        - name: Homepage
          url: https://keda.sh
        - name: Source Code
          url: https://github.com/kedacore/keda
        - name: Slack Channel
          url: https://kubernetes.slack.com/messages/CKZJ36A5D
        maintainers:
        - email: cncf-keda-maintainers@lists.cncf.io
          name: KEDA Community
        maturity: deprecated
        minKubeVersion: 1.17.0
        provider:
          name: KEDA Community
        relatedImages:
        - ghcr.io/kedacore/keda-olm-operator:2.7.1
        version: 2.7.1
      entries:
      - name: keda.v2.7.1
        version: 2.7.1
      - name: keda.v2.6.1
        version: 2.6.1
      - name: keda.v2.6.0
        version: 2.6.0
      - name: keda.v2.5.0
        version: 2.5.0
      - name: keda.v2.4.0
        version: 2.4.0
      - name: keda.v2.3.0
        version: 2.3.0
      - name: keda.v2.2.0
        version: 2.2.0
      - name: keda.v2.1.0
        version: 2.1.0
      - name: keda.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: keda
    provider:
      name: KEDA Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: stolostron
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stolostron.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:3443da607b276d39ba28a8d6561d296f30e9bac5ab06d5c70715bddbde0a58b2
          createdAt: "2024-06-28T04:52:57Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/multiclusterhub-operator@sha256:3443da607b276d39ba28a8d6561d296f30e9bac5ab06d5c70715bddbde0a58b2
        - quay.io/stolostron/governance-policy-addon-controller@sha256:593dc0a689091ce7c916e6be764c222c1020629ccff7eafbf17524936e4f8262
        - quay.io/stolostron/observability-e2e-test@sha256:8d9341b1814cd085a20f9feceab181a8989d6684fd44d6c8e54a0c6df4dd7216
        - quay.io/stolostron/governance-policy-framework-addon@sha256:a1c0d40561befaf5e849060beaee4b201f42d77c5090823f23ef43a75106b00b
        - quay.io/stolostron/thanos-receive-controller@sha256:b1a191714cead24f66e8de7217db3c4d4a253ecf0238a912c8b88ba8b47c972d
        - quay.io/stolostron/console@sha256:c7900403ef8dba3aac52f999532d876bbfe2f66f00a255047850594b64021830
        - quay.io/stolostron/volsync-mover-rclone@sha256:b247b23944e20104f801c172cbbb620740a5b51d1777ecdbf98b4c128f1ef171
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/prometheus@sha256:8250af047c03c6b0c07b76420e0a5c2338f053d73b2dca1ecba4bb2f211715fe
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/search-collector@sha256:136ad44ff41a29dec48e87d10af5b608f335e4242c41ac17cba2c3c4c8205bd6
        - quay.io/stolostron/multicluster-operators-channel@sha256:d791aa2c6032c1de50cf4033c426646613b182e98c406a091b88fb9ffefc0c98
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/helloworld@sha256:81a0b06c7fcf915b1bb40d33d29d0bfd4d501404463c1f46bffc503e7d1cc45b
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:02e53b8a6529e3209809dada4113054e3398de18738948cb01ed1d24dcb8e8be
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/insights-metrics@sha256:49a177357e865d5f8347840fe695531d539d17e28d2b9971821e41ebbee014bd
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/rbac-query-proxy@sha256:24aafcba20ded0892209cfcc196bff88f877ed51eb928d5d350b17041d170893
        - quay.io/stolostron/volsync-mover-restic@sha256:284c3358f0f2158d599f27487d76a7fdb21842fc0e95bc639726dfdad14b7e7d
        - quay.io/stolostron/prometheus-config-reloader@sha256:bc93ca9f1b6d95e8b963fe1af4c17f87c8c7d938a348627c29cc3cf42b70ca3e
        - quay.io/stolostron/prometheus-alertmanager@sha256:8222bbdd37c246703de54ed3dc109fa8c2f6ca17c5bb90106948683858afdc65
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/multicloud-integrations@sha256:60be5b945ead8f845a29a8be18bd26922b83f009fde1412017f1ca985eb79b55
        - quay.io/stolostron/grafana@sha256:e40d66da749108ae1eecf909c95c87bf7a289328e8b373ae4402368c474f2da5
        - quay.io/stolostron/observatorium@sha256:af4e315e4c6da3607f84e24f31e0f37f8572d161736b4748b2e57e0a2c2dd36c
        - quay.io/stolostron/volsync-addon-controller@sha256:7774767ed20263bba4857a0d4be012e12215f640fce01fbd3a085bf860d087e7
        - quay.io/stolostron/volsync-mover-syncthing@sha256:b5fad266ca8d3cd8efcbf4e8241c446004747b551b2b8e06acae10900d7a0f1b
        - quay.io/stolostron/thanos@sha256:cc78cd786643d27ed85eb3670852722e915a20b20a21479e1477fb859dec6c5e
        - quay.io/stolostron/prometheus-operator@sha256:490b0056b9a23a1dc514a3036eac6593c7c8961de27d08ff24817d280325138c
        - quay.io/stolostron/cluster-backup-controller@sha256:9c77444bd42e6e99163f4dce9f9975276aea082100f53fd938cddf74ad3b1574
        - quay.io/stolostron/multicluster-operators-application@sha256:199b696dbd7589b77e2391b71504fbcb7d3ae785972dac3d001c4c1bffa342dd
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/kube-rbac-proxy@sha256:9626cb9b2ee8f00d2abdb14cd2ce56d77435859071c10d8394bc7ee2754ce87f
        - quay.io/stolostron/volsync-mover-rsync@sha256:f7affcc16ae9bf6d12f69511b7a269f5fd065d1d87392c4c38cb5d7af5b89e23
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/search-indexer@sha256:068c38489201316dc12d475de03c8abb11575338bd75e25c1262eca8aac49cb2
        - quay.io/stolostron/search-v2-api@sha256:6b97b2c7e64de7cd154fd83da3bc690a57b67f01c83170238fe9c2e764df91a8
        - quay.io/stolostron/cert-policy-controller@sha256:27673358610dc1ca308de472dbd794dcbb38250bb299b53c736549b1e8441a0f
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/metrics-collector@sha256:4101a2f2c4230bdc1420a58b546a241dfc5414cf92b57f3b73619c1be14dbc32
        - quay.io/stolostron/observatorium-operator@sha256:5da7f478ecf3609a3b86eb563e4dd22588f3d13efdc2e33f5b672fe95583ce68
        - quay.io/stolostron/grafana-dashboard-loader@sha256:cff97fe3e32dd14ed2f3b7b02cb205715cfee6cfd6dc7e7a630101ea5147e566
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:9d1d5e2b5313b898188d2c845dcda5d604b73f16e89c71915ea308281c1f56ff
        - quay.io/stolostron/klusterlet-addon-controller@sha256:01e5fca5e5c6caca6878bf8025471b7c77164ca7b4a2d764e20101fc79290df0
        - quay.io/stolostron/search-e2e@sha256:8dfbc899ba1ffaa8fb6eb2861f6db87a13d2565cf4551fd47df7fe8253217fb2
        - quay.io/stolostron/config-policy-controller@sha256:197d8491e7c889db8f2acb726e0146d384c979acc7f634c24a9d4b21f7d444f3
        - quay.io/stolostron/must-gather@sha256:1f7c129859bdaf6bd93dda77aec01f7cd03000fbfaccd809f26d33b9e3f66e51
        - quay.io/stolostron/search-v2-operator@sha256:c41517f372fc1874895b39e70cd1c07391110697a5e49ebcf56f4a5df499d810
        - quay.io/stolostron/kube-state-metrics@sha256:ef1d671ffa098ed95186de6f15ea91fe648d103c1c7a57871ed46af07c6f7881
        - quay.io/stolostron/submariner-addon@sha256:0f1b4ae59ff5c933542f9affed20bfa6694deb804f2e97f49ebd0f2e1b925eca
        - quay.io/stolostron/insights-client@sha256:aac5b9b66800c0b862e8bfa1fea533a4aec854d8c52baddbfa03977e1ed520cb
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:bf6bec399bc2072186e9d12ed865167fe940f3f619d69ffe6c178c14a3bdadbe
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/multicluster-observability-operator@sha256:245e1815a73b5c48f790ae5a5d85fac6642cce8a5a3bf122275809e97a4d519e
        - quay.io/stolostron/grc-policy-framework-tests@sha256:f2b01192fd900860428942c7d5b900604ffcab0d42731676a1acdeebd3716557
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/cluster-permission@sha256:716d2cbcb7f69b5640d04669b15884e0adf053597f278a0bb1d8ee77e4a7af4d
        - quay.io/stolostron/governance-policy-propagator@sha256:200558b5aa2366383bd298e159d9f7966a2c9b981038c79cde536edbfbe0092c
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/volsync@sha256:ba7399b316359118b1c1b915d667fca9a4a0d12a10beae69e6dee9c7236753be
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:7f6b6e45f0be0a14d7d9d79828572b9400f4d559f084028e3254cfd40948780d
        - quay.io/stolostron/multicluster-operators-subscription@sha256:8350df0a7d4f19c8e52eb58eecf5223bb2a0ccbe4d4aa00fcf8dff832cb43e5d
        version: 0.4.0
      entries:
      - name: stolostron.v0.4.0
        version: 0.4.0
      name: community-0.4
    - currentCSV: stolostron.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:781f6712369cd4b9bdc8ba0ee2f35a12db198ca190621e199dc10cfea3f5efc9
          createdAt: "2024-10-02T08:01:20Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.5.0 <0.5.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/governance-policy-framework-addon@sha256:5f2b326a279b7c42c233ea5f11b89ecaf4cfd7c549d021ddcc4e7b0c8926fc6d
        - quay.io/stolostron/rbac-query-proxy@sha256:2f1c78176d0ca0b21c237a9f7a3f77f7d857dd848f3a42254cee1e4e469b4c3c
        - quay.io/stolostron/multiclusterhub-operator@sha256:781f6712369cd4b9bdc8ba0ee2f35a12db198ca190621e199dc10cfea3f5efc9
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:d42ddd30570ccdf388ccd65cd3f7c72de2d857d328e05aef8e855508ccf91898
        - quay.io/stolostron/observatorium@sha256:8d0f02996b5bf02be02ff3743821909b6c2396f0f0519652dde0c5d1d76f8a05
        - quay.io/stolostron/prometheus@sha256:147f3a0cee8fa7b11c82e21ca95c28278e1ab53ccb5b70f3af0425bb8a74f61b
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/volsync-mover-syncthing@sha256:b5fad266ca8d3cd8efcbf4e8241c446004747b551b2b8e06acae10900d7a0f1b
        - quay.io/stolostron/search-collector@sha256:4082f7a37aa3f10c725bb9bebf94ee8dab97d22302cd5e59b35d0e3bcc888cef
        - quay.io/stolostron/volsync-mover-rclone@sha256:b247b23944e20104f801c172cbbb620740a5b51d1777ecdbf98b4c128f1ef171
        - quay.io/stolostron/insights-metrics@sha256:49a177357e865d5f8347840fe695531d539d17e28d2b9971821e41ebbee014bd
        - quay.io/stolostron/acm-cli@sha256:85a5c55dfa4708435807d1266b5607525e33f55c5b4672fd8ba78987a1c48ef3
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/thanos-receive-controller@sha256:58c97fe5ec2131f9dcfb70563b59ed0bcae94d8bddefa51cb4784e73b1d90e4f
        - quay.io/stolostron/cluster-backup-controller@sha256:688844c722db78f383d90986150bde93bb421068e7f5b4128d19043f3e01f93e
        - quay.io/stolostron/config-policy-controller@sha256:5af6a0e17da7b7261ac71bcb1e2d991856bbb5450686d188026798e05f0617ba
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/volsync-addon-controller@sha256:d0b15da92e0ff9792d750b7b48d72b38d6abb91488c20578c7be3382c561877f
        - quay.io/stolostron/observability-e2e-test@sha256:c361f5a0f806498c6c1150c0c0d5a42222efff13a18cb28dbfd5e4bd835cb4b7
        - quay.io/stolostron/siteconfig-operator@sha256:7f47f0c61d5b53411ed78046c5a02664c6693d940ef37d6db1853670196be334
        - quay.io/stolostron/insights-client@sha256:aac5b9b66800c0b862e8bfa1fea533a4aec854d8c52baddbfa03977e1ed520cb
        - quay.io/stolostron/kube-rbac-proxy@sha256:1b708bff7fb46391f2460d34be28d30add38501cdb084306df01faae75637232
        - quay.io/stolostron/multicluster-observability-operator@sha256:fb0519952ef6436cc66496409f00e89cd6e58a41776df96859a86176a7322d05
        - quay.io/stolostron/multicluster-operators-subscription@sha256:30b1b5bf6a7a12e525450905a474babdd7e38235d9be60f3cb04d60a745a1b96
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:a8d812a527902bd4d9dfd23f90ce90f415f5a6fabb007d1c41b4c998e1ca098a
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/observatorium-operator@sha256:38f22e30db7713e9f2e239463803cdb5cdfe7108a524ffa5d7e95201bae5300b
        - quay.io/stolostron/cluster-permission@sha256:716d2cbcb7f69b5640d04669b15884e0adf053597f278a0bb1d8ee77e4a7af4d
        - quay.io/stolostron/thanos@sha256:dd4d4f8fbb9a8f51134260fc728d19f61a93b511fd7889636089944ac46163bd
        - quay.io/stolostron/search-v2-operator@sha256:6df11a36ea19a1b3da191964f851dee95b87fd993e8f0ef8906b1e63dd450d17
        - quay.io/stolostron/metrics-collector@sha256:c580106b2c7173251afa6714b710fd109811023d88a8009fb741f3d00e486e76
        - quay.io/stolostron/volsync-mover-rsync@sha256:f7affcc16ae9bf6d12f69511b7a269f5fd065d1d87392c4c38cb5d7af5b89e23
        - quay.io/stolostron/search-v2-api@sha256:338634d767383716c7a88ed2bc9c58275e9d5953932293443725883d7efcffd5
        - quay.io/stolostron/multicluster-operators-application@sha256:329c82b92fe2448fe26923fbd82d2eadacb7f96c75689d96393c573f95e6d650
        - quay.io/stolostron/must-gather@sha256:362bbfd54059975369ce6d3869d34e346a0e76159163714cfcb3597022b7c88f
        - quay.io/stolostron/prometheus-config-reloader@sha256:337f7d2036c17459a41884e38c569a3601364b02660946b49d39ce40ef97bd48
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/governance-policy-addon-controller@sha256:9f387eb8b54d1c86fc20d205dbce199738849cb0ce15a47052ddbee8781d2d9b
        - quay.io/stolostron/volsync-mover-restic@sha256:284c3358f0f2158d599f27487d76a7fdb21842fc0e95bc639726dfdad14b7e7d
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/search-e2e@sha256:ddaa50ecfafcd7501903420a47495f733dce36fcb91c8ce9b8d94f4448ed501d
        - quay.io/stolostron/submariner-addon@sha256:07287987fb0041efc16557cb9d7403dc15b5959a5d0dd0d81cd72bc2658cba10
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/klusterlet-addon-controller@sha256:cf4cad3599c4ce538dc007cda9f5b06538799e8198912d907058ecdd4bbb2451
        - quay.io/stolostron/multicloud-integrations@sha256:d4cb3a0a50bc5d22f48c96ed3a10ea9843c6ef2f83daeae411a7ddb7ea332ea6
        - quay.io/stolostron/prometheus-operator@sha256:5ff1f7a4be0216f582f5cb40de3412cbe8126433144c8201750f6c3c0fee920e
        - quay.io/stolostron/governance-policy-propagator@sha256:69ab8f492fa418596db6da578b3b579fc2f1506c5bf780e6aba3ccd05d8b02b9
        - quay.io/stolostron/console@sha256:e42d2f7e39a93f13003193142a3d1f3f9e55966cbf55b94eac93c726373a2516
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/volsync@sha256:ba7399b316359118b1c1b915d667fca9a4a0d12a10beae69e6dee9c7236753be
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:dbdbec1600d57c440173aecf19c57709d69e945dcecb348bcf4a7c37b3700b70
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/search-indexer@sha256:7471f2808686d42c2eae9f0a537f9b520648f31e87072a376243f309fd044caa
        - quay.io/stolostron/grafana@sha256:c9869c7f590bc4c54151c7e4fa9ed76dfe35c24a0ea3c19a6a1b26e61924b6e1
        - quay.io/stolostron/cert-policy-controller@sha256:809cae1aef02dd3fb68143ebdba818b20c661bb628a7922ee1d705efe4947f35
        - quay.io/stolostron/helloworld@sha256:48724b53542a3fcaf342d6ae2056bdf846cbe2ee50879613953b045bfcbe1c0a
        - quay.io/stolostron/kube-state-metrics@sha256:dc534373feda042afa385026c7da674912ca310df2a4598d83451d4474bb3f44
        - quay.io/stolostron/grc-policy-framework-tests@sha256:57794e2b15949cedf3462df2d38b95e86ca1e140a5ae81e4a9b3e7e06072847d
        - quay.io/stolostron/prometheus-alertmanager@sha256:19f2e277cbdd2bd4814b8d8034167c1a7cfe9f1d698fb16f7604fa65cb8bcb1c
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/multicluster-operators-channel@sha256:260a11aae954299ce8fbb0f9c3102afd3b5cf3f1b972b5a71e56c273ba1d532f
        - quay.io/stolostron/grafana-dashboard-loader@sha256:85031d1a56514be239c585487aec05a8cbb08c17aeb60c0c84c685eb9d9f08b5
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:358ac0b0a3afdfac06f784f2e02a305cbef4e6f648e31a00981d3b097746412a
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        version: 0.5.3
      entries:
      - name: stolostron.v0.5.3
        version: 0.5.3
      - name: stolostron.v0.5.2
        version: 0.5.2
      - name: stolostron.v0.5.1
        version: 0.5.1
      - name: stolostron.v0.5.0
        version: 0.5.0
      name: community-0.5
    - currentCSV: stolostron.v0.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:09f1f4266938fba88376a50c3e2837c5517b4f3a947cf86638a17e52eb17c1f6
          createdAt: "2025-02-24T01:42:20Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.6.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/flightctl-worker@sha256:4dddded7d2d4cce63972c8ca81f213c2cedb551cc645ec85495a9692dc36bb4c
        - quay.io/stolostron/helloworld@sha256:56f0d641d4f07b922f1c28c536eacad747c56a564e0a65f084106c2ffc9bd99d
        - quay.io/stolostron/acm-cli@sha256:8c8c059ce2f6d131c410972bf8fd0b7e399cf17cd044c7ddc89a9668c0ceac4c
        - quay.io/stolostron/search-v2-api@sha256:694523242aabb8db1f27b5afcd8ab2f63a9cdb6056c33a79badf8e68a337738d
        - quay.io/stolostron/multicluster-observability-operator@sha256:bed0dd30646d7fce894a625e6cf52e7829c272f47dde0baf1f788e9f85f66e42
        - quay.io/stolostron/cert-policy-controller@sha256:98dd74d7375ca2b546075d5c8f06a9e2c4f2ce3b77247dccf36bcaa46bf564c7
        - quay.io/stolostron/governance-policy-propagator@sha256:c21574a2421fefe3fcabeef14640d6fe699d950106c9ed056b3ba9be3cbeabe0
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/flightctl-periodic@sha256:a3ed745dc65a5886c4cc05348cf4822bfbf2c6d28b85a39fa0b8f3e1f53fdf8a
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/search-indexer@sha256:4f9be902fe976fd88a77b737cc1f56f00d48740a6a62661468328cf3bb9da31a
        - quay.io/stolostron/search-v2-operator@sha256:85e53a97240f6fd1c4251101146cc2bcc51eb91d9b171efb5b58edade9e0244e
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/multicluster-operators-application@sha256:59802fcfae5b6fd0bebe58e98412885e71f32464b8dbbf85364ddaf090aa3e00
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:9c85fb91c4be3287cac2a376465836689df885902099096f0f76ffca50a6d311
        - quay.io/stolostron/search-e2e@sha256:43e1bbfb7259612ffb408108c71b825f6e61a268c7e405b8d855f6878b6f96dd
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/flightctl-api@sha256:64fb1b655bef6aeff03cf64b6450155d3cb570ed4713f084590fe515f4fbad02
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:51a8acc47098175db148ebc19a10cdba2803d4a9f5884dd38b1c839702398789
        - quay.io/stolostron/thanos@sha256:53eee303836070a7c7621bfb2a094e3809146497d1653dc65d16055aacea332c
        - quay.io/stolostron/siteconfig-operator@sha256:97dc128666780ec9a4c955f392ab98f81f0d6038ec2159e8b1d55aab393cf118
        - quay.io/stolostron/config-policy-controller@sha256:8cf54eb6f5b49a60ed165fa4b70e4984a1362f2f31e8304c3a22e8a6a953a606
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/governance-policy-addon-controller@sha256:fa0738abc0368f9dadd8b0845d868512d26be3525a1a07b9de8a9cefafaa42b5
        - quay.io/stolostron/klusterlet-addon-controller@sha256:39e6940eb956fe9640342e11ecbcbc7d6df0169e1b340a60afc19aff7dd713a5
        - quay.io/stolostron/thanos-receive-controller@sha256:d73f98e93a258adffc1a4bd875c79a9629166599e68ef777a790a1f3146a2929
        - quay.io/stolostron/cluster-permission@sha256:443d8b30e3bb83c139c64ddfc38b0d5e6d7a25a4fae783303ab2f2c27d015b26
        - quay.io/stolostron/grafana@sha256:51092f6754106a90aff28cdd80c2f0447c85c1bbc72db15878d0b65b93e1ead6
        - quay.io/stolostron/volsync-addon-controller@sha256:d9dad837f2f735674e5711c7fff1cfe2c0c92160fceb7d1548a62348dd43b81c
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:92369b6deb21ab3cb0a12ae5c74b455cfefb8698826bca39e26648e608e48a2b
        - quay.io/stolostron/observatorium@sha256:398ecee4d3f12e84e48b12009a4b02fc8b360dfc0c1607ffecc26e6bc04d0ea2
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/prometheus-config-reloader@sha256:2b26d2fb5c69e54aecfbcaab46e652d136d726f0604844031c8771874f1fd769
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/grc-policy-framework-tests@sha256:9b84dbafef35bd568b2bf79283b6850ba1ff57d70239a27cc0cc815fb1a55515
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/metrics-collector@sha256:c7433ee4ccdb56717802acf258530e2d6d237787d760a8fbcfd642790254d5ab
        - quay.io/stolostron/cluster-backup-controller@sha256:c85f88d42b42e4366f8ff7ad6c3e6b375e98c5b6a12a9c24a6333929e48f1811
        - quay.io/stolostron/observatorium-operator@sha256:5eb4c0da92c62b664aaf3bc187924870275e687440c3b0a3ad4f1c760c8b35e9
        - quay.io/stolostron/kube-rbac-proxy@sha256:5b521e4f0e5e54c53cea714020069aa796da23d1482d14e4b7c908098ac69ae9
        - quay.io/stolostron/observability-e2e-test@sha256:03403aeb59b364f578761ae4bcb302bd3589baff500d733775e8cb4f2f961236
        - quay.io/stolostron/insights-metrics@sha256:f2f4cf14090252af32832ff6df8a9f896e5314751199156c4e5ae0755e4515e6
        - quay.io/stolostron/search-collector@sha256:54ece9741641e1bb6e30f15d97a65c92b9c6408ba2572a2f6658f537134adff1
        - quay.io/stolostron/multicluster-operators-subscription@sha256:a056ce7d438f68d33ca3dfd42ec3d9a34b347cdf05499066412784109726e7ad
        - quay.io/stolostron/submariner-addon@sha256:9c72b5177de8ddc56dd3544997e9bd222895545b7c578e43e4051596fb1fa873
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/prometheus-alertmanager@sha256:b98480f8ba6ffd19fcee3b412757ad34428e6feab7e4196bb13f2f940306e2c9
        - quay.io/stolostron/insights-client@sha256:978d8b7c3e692272024b71b8c005b752c08dc643dde8258ad75961ddcdcce67f
        - quay.io/stolostron/prometheus@sha256:3d1810567f67ae1a7bb024a1c4e494ee510cfb14c7ed21a8b5e63e41fa7bcd68
        - quay.io/stolostron/flightctl-ocp-ui@sha256:6cff6425850d2b4a8b885a684afb01244d7cd9de7035d7ad71b1e10e5bab25e9
        - quay.io/stolostron/kube-state-metrics@sha256:cd24916e55ad905e386fa3df8a648272114353a0d75283224a4dddb1955b2687
        - quay.io/stolostron/flightctl-ui@sha256:fea717f3b8b4c343ec60ce7d74a691f1124b6847d55e30a73dc0d50ea610d370
        - quay.io/stolostron/multicluster-operators-channel@sha256:501a5053bd3fd5cc9de34e7e020890d731321b090a0e5bf11b8b3350d002f573
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/stolostron/multiclusterhub-operator@sha256:09f1f4266938fba88376a50c3e2837c5517b4f3a947cf86638a17e52eb17c1f6
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/prometheus-operator@sha256:f80e79ea87cde66955e1e46d264e414c072085da9db6876ee2ae68c87d2452e5
        - quay.io/stolostron/governance-policy-framework-addon@sha256:047124271d0b30176a94454b293c82c2eb67a6f06757fc22c54c455fd0913157
        - quay.io/stolostron/console@sha256:eed52f3ab860a3a224ce178d92765ca50d72d6261f04df040a39577881738252
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/multicloud-integrations@sha256:0081065f9e59c5ed41087344668ac86012f796cdbbf3575fd8491c7dfd24fbf8
        - quay.io/stolostron/must-gather@sha256:2bf9d3c15f32a6bfe0a7a8d39b9dc62ac6cbb6fe47b7cf1f6bc5e30b17e5e849
        - quay.io/stolostron/grafana-dashboard-loader@sha256:6c6c2aa528373536d5db66fffeb87a85b79f3cf723e5d9eeaefcb74f91c1a8a1
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:05da11b343938866c5afcfe844bcd15d9c513fdb10fd7ea0d3473b455b427c2a
        - quay.io/stolostron/rbac-query-proxy@sha256:096ad23c9b9c7e90451c09b48ded67a82575e75c8cc19f22c459d63ae8b5a786
        version: 0.6.3
      entries:
      - name: stolostron.v0.6.3
        version: 0.6.3
      - name: stolostron.v0.6.2
        version: 0.6.2
      - name: stolostron.v0.6.1
        version: 0.6.1
      - name: stolostron.v0.6.0
        version: 0.6.0
      name: community-0.6
    - currentCSV: stolostron.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:ab6660283a25bdf2ea9afe8c76bfbc32f4610d69af2bb13a7bca37a1f5961bdd
          createdAt: "2025-06-09T15:10:05Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.7.1'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MulticlusterHub defines the configuration
              for an instance of a multicluster hub, a central point for managing multiple
              Kubernetes-based clusters. The deployment of multicluster hub components
              is determined based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/rbac-query-proxy@sha256:1c198b6388a72ccfe2c5b68d57a2b22893fb65d977992ec44ee8cad6ee1ca976
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/stolostron/governance-policy-addon-controller@sha256:6ecc7557168177474423cea60eadc07d8382adb3e97dee669a8590859a41b08c
        - quay.io/stolostron/flightctl-api@sha256:64fb1b655bef6aeff03cf64b6450155d3cb570ed4713f084590fe515f4fbad02
        - quay.io/stolostron/cert-policy-controller@sha256:6deb41867f8781f4e34f69a54d4aeca190d588ba2991279d39cb5a82c2a5c5be
        - quay.io/stolostron/flightctl-ui@sha256:fea717f3b8b4c343ec60ce7d74a691f1124b6847d55e30a73dc0d50ea610d370
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/cluster-backup-controller@sha256:e576c2de1f931b7d33da2dbc03d71ce323766bd9158bf4b128221ddd1ab3fb9f
        - quay.io/stolostron/observatorium@sha256:18a06117ef114a1e68474fd9832adacd30b03f60ec1aa55374285a20e5311f13
        - quay.io/stolostron/insights-metrics@sha256:861dc03fff55c1d9aff96b5ed15b9d38d5932df1c00a4423e35ead19da6eda03
        - quay.io/stolostron/siteconfig-operator@sha256:16b00f6c4d81eea26e663ca5596526d85bf7777728d1f0ca7e4dd765c64d2bb5
        - quay.io/stolostron/search-indexer@sha256:58320b54847e3bb6baa558bda65e9d0ae2198df480ba27f46f96ba1454614508
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:4b29cc08c9e35054a4a820472ea19cd51e043ab1a90495418a36922df74b9540
        - quay.io/stolostron/multiclusterhub-operator@sha256:ab6660283a25bdf2ea9afe8c76bfbc32f4610d69af2bb13a7bca37a1f5961bdd
        - quay.io/stolostron/prometheus-alertmanager@sha256:a9da6f01cfa61fef803f09ef3f6b27e33a07b97fc1f1e11b6a6c8e13b8a4b90e
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/volsync-addon-controller@sha256:224a0a03a25e5ce61f4c50f6252370876ac0cbf4f9711e997e98a00d6a124fe9
        - quay.io/stolostron/origin-cli@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/search-v2-operator@sha256:2ae9ab48615966ccf91b2022ccda12913a3db3cfbad5351dd8a20648b01137d8
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/console@sha256:ba89196efe30c812a767d58a41170b56e2b04167a72e7f13180bf40474bdc824
        - quay.io/stolostron/prometheus-config-reloader@sha256:024ee6dda10c6d22ddeab57d7d6503b7aa9ac78981cfcec8ffae94b4914500e3
        - quay.io/stolostron/submariner-addon@sha256:51359ccdc0ea29bcaf7ec2b190878e9f3333b1530cb1128e142d893fb87eba8b
        - quay.io/stolostron/prometheus-operator@sha256:8c1c12562ec00da8eaedb7d428694ebdd3a749c0b1ce03fe2b5663ae40e1c701
        - quay.io/stolostron/flightctl-periodic@sha256:a3ed745dc65a5886c4cc05348cf4822bfbf2c6d28b85a39fa0b8f3e1f53fdf8a
        - quay.io/stolostron/governance-policy-framework-addon@sha256:b1b24c40d68415525e2bae21810fc439d3e96fcc5fa75c92e7f2eee284422bfd
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:05da11b343938866c5afcfe844bcd15d9c513fdb10fd7ea0d3473b455b427c2a
        - quay.io/stolostron/helloworld@sha256:4465550664bd210cf0df8edfd706de2fc51cacf01b66a962b5607935f9666ebc
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:54c6a2eec497eafdbf861fb95d78d961fb24906b197fc69b4e96d3e8b53e6496
        - quay.io/stolostron/multicluster-observability-operator@sha256:3d95297297a301fca1915268fec71d7a6ecef98c6c2ab8494d9befacb2e169fa
        - quay.io/stolostron/thanos@sha256:9d504e2aa1001135eb6e454c58c620d102f2b0a85016b7587452dd1595794c93
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/must-gather@sha256:ad8db72d559d06f4d467f78213d4f8f2cb00f8e76e942d8e75394eb7dadd29d2
        - quay.io/stolostron/acm-cli@sha256:2f0cec3bed4ad77f573e8cd51a7e2c62d68a59e4e261ea801d5ac203fd2724fc
        - quay.io/stolostron/grafana@sha256:cebaed7e5abe0064c65ea69bd4e14794d153de9172c4c3b6de4730f23fed7d24
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/metrics-collector@sha256:c7433ee4ccdb56717802acf258530e2d6d237787d760a8fbcfd642790254d5ab
        - quay.io/stolostron/observability-e2e-test@sha256:4e60d3e3e1b5dbab3bcc1c4d37a4940a2a42a065a9df539fb243fb174dde6716
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/search-v2-api@sha256:786866ca8c71a8263ddef45ca9bd9e48cfd1b3ad6215725fbf6f79967bda20ad
        - quay.io/stolostron/multicluster-operators-channel@sha256:318d885ab33d5afe17334e831c955f1c3cdf804281b7cc8523330820ac22eee2
        - quay.io/stolostron/multicloud-integrations@sha256:665e9385531674a5849aa661eb207a0c6e39ea563fc891c57678cddbf7878d8e
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/observatorium-operator@sha256:2f11444703addec163a00c99529c228594fb52df13de47f08faac2e446bec483
        - quay.io/stolostron/governance-policy-propagator@sha256:2b77422fb03177f647c04716e7242bc08f241a1ef9c8968e492ae57816f4c4fb
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/grafana-dashboard-loader@sha256:6c6c2aa528373536d5db66fffeb87a85b79f3cf723e5d9eeaefcb74f91c1a8a1
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/cluster-permission@sha256:e45139d459951b2e751018fe7ecabe1ff22b90ac87be81fb8f17a12f7b7f6a65
        - quay.io/stolostron/search-collector@sha256:2b3dfd21c0ba7b7b84fac390002e47fae927cd5fdf13e5d1d960703aa6c5bede
        - quay.io/stolostron/insights-client@sha256:2cbd20252f10a5a64f93af50f117f05a865aa9d578e5f6e55e4249470b761604
        - quay.io/stolostron/kube-rbac-proxy@sha256:e5555e256a518ca34bea01c959bbb5497582b53dbfaaf326e1852430c175637a
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:1efd8e470b88a992e2f35ce18bdf4970f91af7ba917b8d92627b47ce8c10eb40
        - quay.io/stolostron/flightctl-worker@sha256:4dddded7d2d4cce63972c8ca81f213c2cedb551cc645ec85495a9692dc36bb4c
        - quay.io/stolostron/config-policy-controller@sha256:b1718d5facb95c5572ac658c899f9fce0a7e8c08de93804de3327cd50cdb073b
        - quay.io/stolostron/redis-7-c9s@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/search-e2e@sha256:1c7dcc9f3588620b3570a493c2d744de059f425b572b9cb5725142b89414a904
        - quay.io/stolostron/flightctl-ocp-ui@sha256:6cff6425850d2b4a8b885a684afb01244d7cd9de7035d7ad71b1e10e5bab25e9
        - quay.io/stolostron/thanos-receive-controller@sha256:f445bbfc18bbc4646e0432842022d3be5305809e52ddcb9daf0ec000844d5854
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/klusterlet-addon-controller@sha256:4cb2f4f4e14cd67a4b2f9549cba714b635d9e1c2d8b869b0e0573fd7a41bc13e
        - quay.io/stolostron/multicluster-operators-application@sha256:9f804b4f2da910e0d7f122176f72c0b0ef4b30785fde0369a66f95f4bac58a82
        - quay.io/stolostron/prometheus@sha256:b959035ff754050ca092863bb74c35ba9622d9adc9e8eb7caf3b430cfe211730
        - quay.io/stolostron/grc-policy-framework-tests@sha256:327122b13ec79ea488d6f424806549e36ebfbabd055f9b40217ba541f4a118a9
        - quay.io/stolostron/multicluster-operators-subscription@sha256:d7c137379781ce6b697f73a464b89d04fbc20c3da00ce21a195ac79ab57a6dc7
        - quay.io/stolostron/kube-state-metrics@sha256:3503df61f3d9ebf3cbf66deef865450b32c7403839fdaca42504febaac00687c
        version: 0.7.1
      entries:
      - name: stolostron.v0.7.1
        version: 0.7.1
      - name: stolostron.v0.7.0
        version: 0.7.0
      name: community-0.7
    defaultChannel: community-0.7
    packageName: stolostron
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MinIO
      provider-url: https://min.io
    name: minio-object-store-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-object-store-operator.v2025.8.19175300
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aistor.min.io/v1",
                "kind": "ObjectStore",
                "metadata": {
                  "annotations": {
                    "prometheus.io/path": "/minio/v2/metrics/cluster",
                    "prometheus.io/port": "9000",
                    "prometheus.io/scrape": "true"
                  },
                  "labels": {
                    "app": "minio"
                  },
                  "name": "my-objectstore"
                },
                "spec": {
                  "cache": {
                    "enabled": false
                  },
                  "certificates": {
                    "disableAutoCert": false
                  },
                  "configuration": {
                    "name": "storage-configuration"
                  },
                  "firewall": {
                    "enabled": false
                  },
                  "mountPath": "/export",
                  "podManagementPolicy": "Parallel",
                  "pools": [
                    {
                      "name": "pool-0",
                      "securityContext": {
                        "fsGroupChangePolicy": "OnRootMismatch"
                      },
                      "servers": 4,
                      "volumeClaimTemplate": {
                        "apiVersion": "v1",
                        "kind": "persistentvolumeclaims",
                        "spec": {
                          "accessModes": [
                            "ReadWriteOnce"
                          ],
                          "resources": {
                            "requests": {
                              "storage": "1Ti"
                            }
                          },
                          "storageClassName": "standard"
                        }
                      },
                      "volumesPerServer": 4
                    }
                  ]
                }
              },
              {
                "apiVersion": "aistor.min.io/v1alpha1",
                "kind": "AdminJob",
                "metadata": {
                  "name": "minio-job"
                },
                "spec": {
                  "commands": [
                    {
                      "args": {
                        "name": "iac-bucket"
                      },
                      "op": "make-bucket"
                    }
                  ],
                  "containerSecurityContext": {},
                  "objectStore": {
                    "name": "my-objectstore",
                    "namespace": "my-objectstore"
                  },
                  "securityContext": {},
                  "serviceAccountName": "my-objectstore-sa"
                }
              },
              {
                "apiVersion": "sts.min.io/v1beta1",
                "kind": "PolicyBinding",
                "metadata": {
                  "name": "example-binding"
                },
                "spec": {
                  "application": {
                    "namespace": "my-objectstore",
                    "serviceaccount": "my-objectstore-sa"
                  },
                  "policies": [
                    "consoleAdmin"
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,AI/Machine Learning
          createdAt: "2025-08-27T14:50:56Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AdminJob
            name: adminjobs.aistor.min.io
            version: v1alpha1
          - kind: ObjectStore
            name: objectstores.aistor.min.io
            version: v1
          - kind: PolicyBinding
            name: policybindings.sts.min.io
            version: v1alpha1
          - kind: PolicyBinding
            name: policybindings.sts.min.io
            version: v1beta1
        description: "MinIO AIStor extends the most sophisticated object store on
          the market. It is designed for the exascale data \ninfrastructure challenges
          presented by modern AI workloads. AIStor provides new features, along with
          performance and\nscalability improvements, to enable enterprises to store
          all AI data in one infrastructure.\n\nMinIO is the standard for building
          <a href=\"https://min.io/solutions/object-storage-for-ai\">large scale AI
          data infrastructure</a>.\n\nThe MinIO AIStor is a software-defined, Amazon
          S3 compatible object store that will run anywhere - from the public cloud
          to the edge.<br><br>\n\nEnterprises use MinIO to deliver against artificial
          intelligence, machine learning, analytics, application, backup and archival
          workloads - all from a single platform.\n\nThe use cases include storage
          for HDFS replacements, AI datalakes, SIEM datalakes, analytics datalakes,
          object storage as a service, streaming and database warm/hot storage.<br><br>\n\nMinIO’s
          AIStor is the fastest object storage server, with production READ/WRITE
          speeds in excess of 2.2 TiB/s on commodity hardware.\n\nBuilt for the cloud
          operating model it is native to the technologies and architectures that
          define the cloud. These include containerization, orchestration with Kubernetes,
          microservices and multi-tenancy - making it the most widely integrated object
          store in the market today.<br><br>\n\nThe AIStor offers a host of enterprise
          features including inline erasure coding, bit-rot detection, state-of-the-art
          encryption, active-active replication, object locking, lifecycle management,
          observability, key-management, caching, firewall/load balancing, global
          console, search and identity + access management.<br><br>\n\nThe AIStor
          is priced on usable capacity and includes support through the MinIO Subscription
          Network.\n\nThe Subscription Network is designed to optimize deployments
          of MinIO and reduce the risk associated with data loss and data breach while
          providing 24/7/365 direct-to-engineering support.<br><br>\n\n<a href=\"https://resources.min.io/request-a-demo-marketplaces/\">Request
          a demo today</a>"
        displayName: Minio AIStor Object Store
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Storage
        - S3
        - MinIO
        - Object Storage
        links:
        - name: Minio AIStor
          url: https://min.io
        - name: Support
          url: https://subnet.min.io
        - name: Github
          url: https://github.com/minio/aistor
        maintainers:
        - email: dev@min.io
          name: MinIO
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: MinIO
          url: https://min.io
        relatedImages:
        - quay.io/minio/aistor/operator@sha256:f80139efa062fdd83f78318b20f371a827b68e854b03dd9706b9eba4c493aeb2
        version: 2025.8.19175300
      entries:
      - name: minio-object-store-operator.v2025.8.19175300
        version: 2025.8.19175300
      - name: minio-object-store-operator.v2025.8.1184927
        version: 2025.8.1184927
      - name: minio-object-store-operator.v2025.7.1011319
        version: 2025.7.1011319
      - name: minio-object-store-operator.v2025.6.3065135
        version: 2025.6.3065135
      - name: minio-object-store-operator.v2025.5.12190907
        version: 2025.5.12190907
      name: stable
    defaultChannel: stable
    packageName: minio-object-store-operator
    provider:
      name: MinIO
      url: https://min.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: https://stakater.com
    name: template-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: template-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-07-04T00:40:36Z"
          description: Operator to distribute plain kubernetes manifests, helm charts
            and, copy secrets and configmaps to multiple namespaces
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.0.1
          operatorframework.io/suggested-namespace: template-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterTemplateInstance is the Schema for the clustertemplateinstances
              API
            displayName: Cluster Template Instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateGroupInstance is the Schema for the TemplateGroupInstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.templates.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
        description: '"Operator to distribute plain kubernetes manifests, helm charts
          and, copy secrets and configmaps to multiple namespaces"'
        displayName: Template Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - templates
        - secrets
        - configmaps
        - distribution
        - stakater
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/template-operator@sha256:ad8312417e4e9621b74d3ad03020516840feef9c4b52c4f0d894dbb6dc6f906a
        version: 0.0.1
      entries:
      - name: template-operator.v0.0.1
        version: 0.0.1
      name: release-0.0
    - currentCSV: template-operator.v0.1.5
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-08-29T06:41:42Z"
          description: Operator to distribute plain kubernetes manifests, helm charts
            and, copy secrets and configmaps to multiple namespaces
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.1.5
          operatorframework.io/suggested-namespace: template-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterTemplateInstance is the Schema for the clustertemplateinstances
              API
            displayName: Cluster Template Instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateGroupInstance is the Schema for the TemplateGroupInstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.templates.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
        description: '"Operator to distribute plain kubernetes manifests, helm charts
          and, copy secrets and configmaps to multiple namespaces"'
        displayName: Template Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - templates
        - secrets
        - configmaps
        - distribution
        - stakater
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/template-operator@sha256:88f722baf22ac07704503f098f2b0ea7d76815a1ab0552474b1c9ba6639d8142
        version: 0.1.5
      entries:
      - name: template-operator.v0.1.5
        version: 0.1.5
      name: release-0.1
    defaultChannel: release-0.1
    packageName: template-operator
    provider:
      name: Stakater
      url: https://stakater.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Starburst Data, Inc.
      provider-url: https://www.starburstdata.com
    name: starburst-enterprise-helm-operator-paygo-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: starburst-enterprise-helm-operator-paygo.v402.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstEnterprise",
                "metadata": {
                  "name": "starburstenterprise-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "cache": {
                    "diskUsagePercentage": 80,
                    "enabled": false,
                    "ttl": "7d",
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "catalogs": {
                    "tpch": "connector.name=tpch\ntpch.splits-per-node=4"
                  },
                  "commonLabels": {},
                  "coordinator": {
                    "additionalProperties": "",
                    "affinity": {},
                    "deploymentAnnotations": {},
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=true\nnode-scheduler.include-coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://localhost:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "resources": {
                      "cpu": 16,
                      "memory": "60Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "tolerations": []
                  },
                  "environment": null,
                  "expose": {
                    "clusterIp": {
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      }
                    },
                    "ingress": {
                      "annotations": {},
                      "host": null,
                      "ingressClassName": null,
                      "ingressName": "coordinator-ingress",
                      "path": "/",
                      "pathType": "ImplementationSpecific",
                      "serviceName": "starburst",
                      "servicePort": 8080,
                      "tls": {
                        "enabled": true,
                        "secretName": null
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "nodePort": 30080,
                          "port": 8080
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "externalSecrets": {
                    "enabled": false,
                    "goDaddy": {
                      "backendType": "secretsManager"
                    },
                    "secretPrefix": "external0",
                    "type": "goDaddy"
                  },
                  "extraArguments": [],
                  "extraSecret": {
                    "file": null,
                    "name": null
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise",
                    "tag": "402-e.1"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "initImage": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise-init",
                    "tag": "1.3.2"
                  },
                  "internal": {
                    "ports": {
                      "http": {
                        "port": 8080
                      },
                      "https": {
                        "port": 8443
                      }
                    }
                  },
                  "internalTls": false,
                  "licenseManager": {
                    "cronjob": {
                      "affinity": {},
                      "nodeSelector": {},
                      "priorityClassName": null,
                      "securityContext": {},
                      "tolerations": []
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "registry.connect.redhat.com/starburst/license-manager",
                      "tag": "0.0.2"
                    },
                    "url": "https://license-api.dogfood.eng.starburstdata.net/test/license/checkout"
                  },
                  "livenessProbe": null,
                  "prometheus": {
                    "agent": {
                      "config": "/etc/starburst/telemetry/prometheus.yaml",
                      "port": 8081,
                      "version": "0.16.1"
                    },
                    "enabled": true,
                    "rules": [
                      {
                        "attrNameSnakeCase": true,
                        "name": "$1",
                        "pattern": "trino.execution<name=QueryManager><>(running_queries|queued_queries)",
                        "type": "GAUGE"
                      },
                      {
                        "name": "failed_queries",
                        "pattern": "trino.execution<name=QueryManager><>FailedQueries\\.TotalCount",
                        "type": "COUNTER"
                      }
                    ]
                  },
                  "query": {
                    "maxConcurrentQueries": 3
                  },
                  "readinessProbe": null,
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "securityContext": {},
                  "sharedSecret": null,
                  "siac": {
                    "additionalHeapSizePercentage": 15,
                    "enabled": false
                  },
                  "spilling": {
                    "enabled": false,
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "startupProbe": null,
                  "usageMetrics": {
                    "enabled": false,
                    "usageClient": {
                      "initialDelay": "1m",
                      "interval": "1m"
                    }
                  },
                  "userDatabase": {
                    "enabled": false,
                    "users": [
                      {
                        "password": "46991b33f7a75ff79213c0dc0e610610",
                        "username": "admin"
                      }
                    ]
                  },
                  "worker": {
                    "additionalProperties": "",
                    "affinity": {},
                    "autoscaling": {
                      "enabled": false,
                      "maxReplicas": 100,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "deploymentAnnotations": {},
                    "deploymentTerminationGracePeriodSeconds": 300,
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://coordinator:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "kedaScaler": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.connect.redhat.com/starburst/keda-trino-scaler",
                        "tag": "0.1.7"
                      },
                      "port": 8021,
                      "scaledObjectSpec": {
                        "advanced": {},
                        "cooldownPeriod": 300,
                        "idleReplicaCount": 0,
                        "maxReplicaCount": 100,
                        "minReplicaCount": 1,
                        "pollingInterval": 30,
                        "scaleTargetRef": {
                          "name": "worker"
                        },
                        "triggers": [
                          {
                            "metadata": {
                              "numberOfQueriesPerWorker": "10",
                              "scaleInToIdleReplicaCountIfNoQueuedQueriesLeft": "true",
                              "scaleMethod": "query_queue"
                            },
                            "type": "external"
                          }
                        ]
                      }
                    },
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "replicas": 2,
                    "resources": {
                      "cpu": 16,
                      "memory": "100Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "starburstWorkerShutdownGracePeriodSeconds": 120,
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstHive",
                "metadata": {
                  "name": "starbursthive-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "affinity": {},
                  "commonLabels": {},
                  "database": {
                    "external": {
                      "driver": null,
                      "jdbcUrl": null,
                      "password": null,
                      "user": null
                    },
                    "internal": {
                      "databaseName": "hive",
                      "databasePassword": "HivePass1234",
                      "databaseUser": "hive",
                      "driver": "org.postgresql.Driver",
                      "env": [],
                      "envFrom": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.redhat.io/rhel8/postgresql-10",
                        "tag": "1"
                      },
                      "port": 5432,
                      "resources": {
                        "limits": {
                          "cpu": 2,
                          "memory": "1Gi"
                        },
                        "requests": {
                          "cpu": 2,
                          "memory": "1Gi"
                        }
                      },
                      "volume": {
                        "emptyDir": {}
                      }
                    },
                    "type": "internal"
                  },
                  "deploymentAnnotations": {},
                  "env": [],
                  "envFrom": [],
                  "expose": {
                    "clusterIp": {
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "nodePort": 30083,
                          "port": 9083
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "extraArguments": [],
                  "hdfs": {
                    "hadoopUserName": null
                  },
                  "heapSizePercentage": 85,
                  "hiveMetastoreStorageAuthorization": true,
                  "hiveMetastoreWarehouseDir": null,
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/hive",
                    "tag": "3.1.2-e.22"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "nodeSelector": {},
                  "objectStorage": {
                    "adl": {
                      "oauth2": {
                        "clientId": null,
                        "credential": null,
                        "refreshUrl": null
                      }
                    },
                    "awsS3": {
                      "accessKey": null,
                      "endpoint": null,
                      "pathStyleAccess": false,
                      "region": null,
                      "secretKey": null
                    },
                    "azure": {
                      "abfs": {
                        "accessKey": {
                          "accessKey": null,
                          "storageAccount": null
                        },
                        "authType": "accessKey",
                        "oauth": {
                          "clientId": null,
                          "endpoint": null,
                          "secret": null
                        }
                      },
                      "wasb": {
                        "accessKey": null,
                        "storageAccount": null
                      }
                    },
                    "gs": {
                      "cloudKeyFileSecret": null
                    }
                  },
                  "podAnnotations": {},
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "resources": {
                    "limits": {
                      "cpu": 1,
                      "memory": "1Gi"
                    },
                    "requests": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "securityContext": {},
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
          description: Deploys StarburstEnterprise Helm Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-paygo-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-paygo-rhmp/support?utm_source=openshift_console
          marketplace.redhat.com/meterDefinition: |-
            {
              "apiVersion": "marketplace.redhat.com/v1beta1",
              "kind": "MeterDefinition",
              "metadata": {
                "name": "sep-vcpu-usage"
              },
              "spec": {
                "group": "charts.starburstdata.com",
                "kind": "StarburstEnterprise",
                "resourceFilters": [
                  {
                    "namespace": {
                      "useOperatorGroup": true
                    },
                    "ownerCRD": {
                      "apiVersion": "charts.starburstdata.com/v1alpha1",
                      "kind": "StarburstEnterprise"
                    },
                    "workloadType": "Pod"
                  }
                ],
                "meters": [
                  {
                    "aggregation": "sum",
                    "period": "1h",
                    "metricId": "container_cpu_usage_core_avg",
                    "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=~\"coordinator|worker\"}[5m])*100",
                    "workloadType": "Pod"
                  }
                ]
              }
            }
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Starburst Data
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: StarburstEnterprise
            name: starburstenterprises.charts.starburstdata.com
            version: v1alpha1
          - kind: StarburstHive
            name: starbursthives.charts.starburstdata.com
            version: v1alpha1
        description: CRs for deploying starburst enterprise and hive
        displayName: Starburst Enterprise Pay as you go Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - starburst
        - trino
        - sep
        - bigdata
        - data
        links:
        - name: Starburst Enterprise Helm Operator
          url: https://www.starburstdata.com
        maintainers:
        - email: hello@starburstdata.com
          name: Starburst Data, Inc.
        maturity: alpha
        provider:
          name: Starburst Data, Inc.
          url: https://www.starburstdata.com
        relatedImages:
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
        version: 402.1.0
      entries:
      - name: starburst-enterprise-helm-operator-paygo.v402.1.0
        version: 402.1.0
      name: alpha
    defaultChannel: alpha
    packageName: starburst-enterprise-helm-operator-paygo-rhmp
    provider:
      name: Starburst Data, Inc.
      url: https://www.starburstdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: redhat-oadp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oadp-operator.v1.0.19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "restic": {
                      "enable": true
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws",
                        "kubevirt"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ResticRepository",
                "metadata": {
                  "name": "resticrepository",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/oadp/oadp-rhel8-operator@sha256:caf87a9832c153833feb34affc2b797ad0e643362448742b0405d457e5f0ba41
          createdAt: "2025-04-14T13:25:43Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <1.0.19'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/oadp/oadp-mustgather-rhel8@sha256:af4c6af48f38510c7d52f678798a9affea2211dc488ed18f83a5d005b7ef4b6b
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is a Velero resource that represents the capture of
              Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Data Protection Applications on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://github.com/openshift/oadp-operator
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/oadp/oadp-velero-plugin-for-microsoft-azure-rhel8@sha256:3e81b5e55f96504ba76d9d47bbeec50c7b1fd90b7702fc31b1dba248050ce9b4
        - registry.redhat.io/oadp/oadp-velero-plugin-for-gcp-rhel8@sha256:05a164c27da94b374ce03451c851a607a3cd436b73b9e5b70723c6dc8cacb3d8
        - registry.redhat.io/oadp/oadp-mustgather-rhel8@sha256:af4c6af48f38510c7d52f678798a9affea2211dc488ed18f83a5d005b7ef4b6b
        - registry.redhat.io/oadp/oadp-rhel8-operator@sha256:caf87a9832c153833feb34affc2b797ad0e643362448742b0405d457e5f0ba41
        - registry.redhat.io/oadp/oadp-registry-rhel8@sha256:0886b3b4ce0295080bf2c5ee95f2b7b15a4483bcd94cb739e7e7154d21150397
        - registry.redhat.io/oadp/oadp-velero-restic-restore-helper-rhel8@sha256:5d288335b5ad114b9eb61410befb4a9d1f84f0ba0c01558098bf4567fda9a8c5
        - registry.redhat.io/oadp/oadp-velero-plugin-rhel8@sha256:694603b9c1fd5eec939b28f8727fd7c9c0d380620c5081006f2655e98358464f
        - registry.redhat.io/oadp/oadp-velero-plugin-for-csi-rhel8@sha256:05ec16bdf1f0fadc4ddad244b86934f73569d2e66cc19f03b3553e257441a662
        - registry.redhat.io/oadp/oadp-velero-rhel8@sha256:39dec8b9ccdcabeaab3e56e14b79cfed4d8574cb172a886d017813a437e4ffff
        - registry.redhat.io/oadp/oadp-velero-plugin-for-aws-rhel8@sha256:28045538fff9e5051cd01e56876971cf9401a448e1dd63bff17fbb900629f22f
        - registry.redhat.io/oadp/oadp-kubevirt-velero-plugin-rhel8@sha256:d862a1b0420fcfb91403f90047b00035efba112af6bc5c50adb723014a620767
        version: 1.0.19
      entries:
      - name: oadp-operator.v1.0.19
        version: 1.0.19
      - name: oadp-operator.v1.0.18
        version: 1.0.18
      - name: oadp-operator.v1.0.17
        version: 1.0.17
      - name: oadp-operator.v1.0.16
        version: 1.0.16
      - name: oadp-operator.v1.0.15
        version: 1.0.15
      - name: oadp-operator.v1.0.14
        version: 1.0.14
      - name: oadp-operator.v1.0.13
        version: 1.0.13
      - name: oadp-operator.v1.0.12
        version: 1.0.12
      - name: oadp-operator.v1.0.11
        version: 1.0.11
      - name: oadp-operator.v1.0.10
        version: 1.0.10
      - name: oadp-operator.v1.0.9
        version: 1.0.9
      - name: oadp-operator.v1.0.8
        version: 1.0.8
      - name: oadp-operator.v1.0.7
        version: 1.0.7
      - name: oadp-operator.v1.0.6
        version: 1.0.6
      - name: oadp-operator.v1.0.5
        version: 1.0.5
      - name: oadp-operator.v1.0.4
        version: 1.0.4
      - name: oadp-operator.v1.0.3
        version: 1.0.3
      - name: oadp-operator.v1.0.2
        version: 1.0.2
      - name: oadp-operator.v1.0.1
        version: 1.0.1
      - name: oadp-operator.v1.0.0
        version: 1.0.0
      name: stable-1.0
    - currentCSV: oadp-operator.v1.4.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "nodeAgent": {
                      "enable": true,
                      "uploaderType": "restic"
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws",
                        "kubevirt"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: |
            Cloud Provider,Developer Tools,Modernization & Migration,OpenShift Optional,Storage
          certified: "false"
          containerImage: registry.redhat.io/oadp/oadp-rhel9-operator@sha256:ee010fb34448b772690a8a98808aa25aa1578a81061618f948b1d16b5d63d333
          createdAt: "2025-07-10T04:59:11Z"
          description: OADP (OpenShift API for Data Protection) operator sets up and
            installs Data Protection Applications on the OpenShift platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <1.4.5'
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/oadp/oadp-mustgather-rhel9@sha256:0cee6fecf052bae3ede6bb6f032b25c0c7cd63ced1e26328f2b02144f24fbb12
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A backup repository is an indicator of a connection from
              the restic/kopia server to the backupstoragelocation.
            displayName: BackupRepository
            kind: BackupRepository
            name: backuprepositories.velero.io
            version: v1
          - description: Backup is a Velero resource that represents the capture of
              Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - description: The CloudStorage API automates the creation of a bucket for
              object storage.
            displayName: CloudStorage
            kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataDownload represents a data download of a volume snapshot.
              There is one DataDownload created per volume to be restored.
            displayName: DataDownload
            kind: DataDownload
            name: datadownloads.velero.io
            version: v2alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DataUpload acts as the protocol between data mover plugins
              and data mover controller for the datamover backup operation
            displayName: DataUpload
            kind: DataUpload
            name: datauploads.velero.io
            version: v2alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: |
          **OpenShift API for Data Protection (OADP)** operator sets up and installs
          Velero on the OpenShift platform, allowing users to backup and restore
          applications.

          Backup and restore Kubernetes resources and internal images, at the
          granularity of a namespace, using a version of Velero appropriate for
          the installed version of OADP.

          OADP backs up Kubernetes objects and internal images by saving them as
          an archive file on object storage. OADP backs up persistent volumes
          (PVs) by creating snapshots with the native cloud snapshot API or with
          the Container Storage Interface (CSI). For cloud providers that do not
          support snapshots, OADP backs up resources and PV data with Restic or Kopia.

          * [Installing OADP for application backup and restore][install-doc]

          * [Installing OADP on a ROSA cluster and using STS, please follow the Getting Started Steps 1-3 in order to obtain the role ARN needed for using the standardized STS configuration flow via OLM][rosa-sts-doc]

          * [Frequently Asked Questions][faq-doc]

          [install-doc]: https://docs.openshift.com/container-platform/latest/backup_and_restore/application_backup_and_restore/installing/about-installing-oadp.html
          [rosa-sts-doc]: https://mobb.ninja/docs/misc/oadp/rosa-sts/
          [faq-doc]: https://access.redhat.com/articles/5456281
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - disaster backup
        - disaster recovery
        - disaster restore
        - disaster
        - backup
        - restore
        - backup automation
        - recovery
        - data protection
        - data management
        - data
        - protection
        - management
        - application resilience
        - resilience
        - velero
        - openshift
        - oadp
        - cloud-native
        - replication
        - kopia
        - restic
        - network file system
        - nfs
        - simple cloud storage
        - s3
        - minio
        - cloud storage
        - data foundation
        - ceph
        - csi
        - container storage interface
        links:
        - name: OADP Operator
          url: https://github.com/openshift/oadp-operator
        maintainers:
        - email: oadp-maintainers@redhat.com
          name: OADP Operator maintainers
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/oadp/oadp-velero-plugin-for-legacy-aws-rhel9@sha256:ab5c8d33684a361fa3bb10a006ef489be92cc955ad3d173bbfa77949dcc42aa8
        - registry.redhat.io/oadp/oadp-kubevirt-velero-plugin-rhel9@sha256:c3b32cb9f5b82021c08cd70467e68aad66746ef211d73612ecfe0af625706267
        - registry.redhat.io/oadp/oadp-velero-plugin-for-aws-rhel9@sha256:79d39186cae5696949813a0ea6c46eeededad29e6bde171d7a9b0011c993bb26
        - registry.redhat.io/oadp/oadp-velero-plugin-for-microsoft-azure-rhel9@sha256:9f05592bf0e5b76322ca582bdede387f9f4aabbbdc436549156c05d652c951d0
        - registry.redhat.io/oadp/oadp-mustgather-rhel9@sha256:0cee6fecf052bae3ede6bb6f032b25c0c7cd63ced1e26328f2b02144f24fbb12
        - registry.redhat.io/oadp/oadp-rhel9-operator@sha256:ee010fb34448b772690a8a98808aa25aa1578a81061618f948b1d16b5d63d333
        - registry.redhat.io/oadp/oadp-velero-rhel9@sha256:17160f99940aa355dc91ab3d16ba6bf91faade6423434008cf8f6e5023b1e413
        - registry.redhat.io/oadp/oadp-velero-plugin-for-gcp-rhel9@sha256:00800ead11ea308794b1536d964f8a043e13ede9657cabb23afa42cba7cf2ce6
        - registry.redhat.io/oadp/oadp-velero-plugin-rhel9@sha256:3ee9d3e799cedc28c400d364b3a8790c7e88acd5f603b09ff512a22d91e265b3
        - registry.redhat.io/oadp/oadp-velero-restic-restore-helper-rhel9@sha256:e0647defc7c428f9a04a5f6d4532dc86e57b92f31b4f907253ae14af05c5993f
        version: 1.4.5
      entries:
      - name: oadp-operator.v1.4.5
        version: 1.4.5
      - name: oadp-operator.v1.4.4
        version: 1.4.4
      - name: oadp-operator.v1.4.3
        version: 1.4.3
      - name: oadp-operator.v1.4.2
        version: 1.4.2
      - name: oadp-operator.v1.4.1
        version: 1.4.1
      - name: oadp-operator.v1.4.0
        version: 1.4.0
      name: stable-1.4
    defaultChannel: stable-1.4
    packageName: redhat-oadp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Oracle
      provider-url: https://www.oracle.com/cloud/
    name: oci-ccm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oci-ccm-operator.v0.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ccm.oraclecloud.com/v1alpha1",
                "kind": "OCICloudManager",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "oci-ccm-operator",
                    "app.kubernetes.io/instance": "ocicloudmanager",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ocicloudmanager",
                    "app.kubernetes.io/part-of": "oci-ccm-operator"
                  },
                  "name": "ocicloudmanager"
                },
                "spec": {
                  "compartment": "CompartmentId",
                  "driverNamespace": "kube-system",
                  "loadBalancer": {
                    "securityList": "SecurityListId",
                    "securityListManagementMode": "Frontend",
                    "subnet": "SubnetId"
                  },
                  "vcn": "VcnId"
                }
              },
              {
                "apiVersion": "ccm.oraclecloud.com/v1alpha1",
                "kind": "OCICsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "oci-ccm-operator",
                    "app.kubernetes.io/instance": "ocicsidriver",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ocicsidriver",
                    "app.kubernetes.io/part-of": "oci-ccm-operator"
                  },
                  "name": "ocicsidriver"
                },
                "spec": {
                  "compartment": "CompartmentId",
                  "driverNamespace": "kube-system",
                  "vcn": "VcnId"
                }
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider, Storage
          containerImage: phx.ocir.io/ax1eoqvpvbtg/oci-ccm-operator:0.0.9
          createdAt: "2023-07-28T23:42:18Z"
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Oracle
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: OCICloudManager
            name: ocicloudmanagers.ccm.oraclecloud.com
            version: v1alpha1
          - kind: OCICsiDriver
            name: ocicsidrivers.ccm.oraclecloud.com
            version: v1alpha1
        description: OCI Kubernetes Cloud Controller Manager implementation (or out-of-tree
          cloud-provider) for Oracle Cloud Infrastructure (OCI).
        displayName: oci-ccm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OCI
        - CCM
        - OpenShift
        - Kubernetes
        links:
        - name: Oracle
          url: https://www.oracle.com/cloud/
        - name: OCI Cloud Controller Manager (CCM)
          url: https://github.com/oracle/oci-cloud-controller-manager
        maintainers:
        - email: david.d.campbell@oracle.com
          name: David Campbell
        - email: alex.blinov@oracle.com
          name: Alex Blinov
        maturity: alpha
        provider:
          name: Oracle
          url: https://www.oracle.com/cloud/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - phx.ocir.io/ax1eoqvpvbtg/oci-ccm-operator:0.0.9
        version: 0.0.9
      entries:
      - name: oci-ccm-operator.v0.0.9
        version: 0.0.9
      name: alpha
    defaultChannel: alpha
    packageName: oci-ccm-operator
    provider:
      name: Oracle
      url: https://www.oracle.com/cloud/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: loki-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-operator.v5.9.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:6728046841367066216b304ed06249b80c57c74151c6a75c9df04a6f7f76e03d
          createdAt: "2025-01-30T17:34:53Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.7.0-0 <5.9.15'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:0c3943fa009f0c649f6d592b0ac410653c8c44b4a90e41fabdfa5f3f7db1e43d
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:6728046841367066216b304ed06249b80c57c74151c6a75c9df04a6f7f76e03d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:2e164e377d4fba98e7f63f43f25b6b9dad441d2e2914c2bf02f3525d5c9aa483
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:542b28fd51194b46efc4e5af10c3241b52db4c71f0d5d1b1f184eb39764e6c53
        version: 5.9.15
      entries:
      - name: loki-operator.v5.9.15
        version: 5.9.15
      - name: loki-operator.v5.9.14
        version: 5.9.14
      - name: loki-operator.v5.9.13
        version: 5.9.13
      - name: loki-operator.v5.9.12
        version: 5.9.12
      - name: loki-operator.v5.9.11
        version: 5.9.11
      - name: loki-operator.v5.9.10
        version: 5.9.10
      - name: loki-operator.v5.9.9
        version: 5.9.9
      - name: loki-operator.v5.9.8
        version: 5.9.8
      - name: loki-operator.v5.9.7
        version: 5.9.7
      - name: loki-operator.v5.9.6
        version: 5.9.6
      - name: loki-operator.v5.9.5
        version: 5.9.5
      - name: loki-operator.v5.9.4
        version: 5.9.4
      - name: loki-operator.v5.9.3
        version: 5.9.3
      - name: loki-operator.v5.9.2
        version: 5.9.2
      - name: loki-operator.v5.9.1
        version: 5.9.1
      - name: loki-operator.v5.9.0
        version: 5.9.0
      name: stable
    - currentCSV: loki-operator.v5.9.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:6728046841367066216b304ed06249b80c57c74151c6a75c9df04a6f7f76e03d
          createdAt: "2025-01-30T17:34:53Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.7.0-0 <5.9.15'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:2e164e377d4fba98e7f63f43f25b6b9dad441d2e2914c2bf02f3525d5c9aa483
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:542b28fd51194b46efc4e5af10c3241b52db4c71f0d5d1b1f184eb39764e6c53
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:0c3943fa009f0c649f6d592b0ac410653c8c44b4a90e41fabdfa5f3f7db1e43d
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:6728046841367066216b304ed06249b80c57c74151c6a75c9df04a6f7f76e03d
        version: 5.9.15
      entries:
      - name: loki-operator.v5.9.15
        version: 5.9.15
      - name: loki-operator.v5.9.14
        version: 5.9.14
      - name: loki-operator.v5.9.13
        version: 5.9.13
      - name: loki-operator.v5.9.12
        version: 5.9.12
      - name: loki-operator.v5.9.11
        version: 5.9.11
      - name: loki-operator.v5.9.10
        version: 5.9.10
      - name: loki-operator.v5.9.9
        version: 5.9.9
      - name: loki-operator.v5.9.8
        version: 5.9.8
      - name: loki-operator.v5.9.7
        version: 5.9.7
      - name: loki-operator.v5.9.6
        version: 5.9.6
      - name: loki-operator.v5.9.5
        version: 5.9.5
      - name: loki-operator.v5.9.4
        version: 5.9.4
      - name: loki-operator.v5.9.3
        version: 5.9.3
      - name: loki-operator.v5.9.2
        version: 5.9.2
      - name: loki-operator.v5.9.1
        version: 5.9.1
      - name: loki-operator.v5.9.0
        version: 5.9.0
      name: stable-5.9
    - currentCSV: loki-operator.v6.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.14-v4.17
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:1cb1d2cad8632791529eb754938612f5e5ed373284e42f309d603059ab7ec7c5
          createdAt: "2025-03-12T12:05:37Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=5.8.0-0 <6.0.10'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.0
          operators.operatorframework.io.bundle.channels.v1: stable-6.0
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:9aa8d8e4a7fd7ddad5d842a22b09aecdbdfa31ca5eb8c4b17f0818636af5fbf8
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:69e69ab2628742774d12dd9dde50607625efdff50ea4671f2859692fbee5ecb4
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:98d2cf3437faa2400c01edb3c694816e0214c8f9dca0a0b0cdf5882a720248f3
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:1cb1d2cad8632791529eb754938612f5e5ed373284e42f309d603059ab7ec7c5
        version: 6.0.10
      entries:
      - name: loki-operator.v6.0.10
        version: 6.0.10
      - name: loki-operator.v6.0.9
        version: 6.0.9
      - name: loki-operator.v6.0.8
        version: 6.0.8
      - name: loki-operator.v6.0.7
        version: 6.0.7
      - name: loki-operator.v6.0.6
        version: 6.0.6
      - name: loki-operator.v6.0.5
        version: 6.0.5
      - name: loki-operator.v6.0.4
        version: 6.0.4
      - name: loki-operator.v6.0.3
        version: 6.0.3
      - name: loki-operator.v6.0.2
        version: 6.0.2
      - name: loki-operator.v6.0.1
        version: 6.0.1
      - name: loki-operator.v6.0.0
        version: 6.0.0
      name: stable-6.0
    - currentCSV: loki-operator.v6.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.15-v4.18
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:0bde3c1a0b667a693f443688ed3926eef1705c45ecf4cce3a3dedbddca7a2784
          createdAt: "2025-03-12T11:19:10Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=5.9.0-0 <6.1.8'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.1
          operators.operatorframework.io.bundle.channels.v1: stable-6.1
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:0bde3c1a0b667a693f443688ed3926eef1705c45ecf4cce3a3dedbddca7a2784
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:731f92c146f299bc3662b309a3068bde18fad51f65a7d29063e582b61e07042f
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:c34676932ddd14db213dbf42a1d341f51692f3f47f1b751d06187e292c99a170
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b94833f55644c289fe2c1297b68a4f9931baf977587c6a2c4c7a3e484c704f64
        version: 6.1.8
      entries:
      - name: loki-operator.v6.1.8
        version: 6.1.8
      - name: loki-operator.v6.1.7
        version: 6.1.7
      - name: loki-operator.v6.1.6
        version: 6.1.6
      - name: loki-operator.v6.1.5
        version: 6.1.5
      - name: loki-operator.v6.1.4
        version: 6.1.4
      - name: loki-operator.v6.1.3
        version: 6.1.3
      - name: loki-operator.v6.1.2
        version: 6.1.2
      - name: loki-operator.v6.1.1
        version: 6.1.1
      - name: loki-operator.v6.1.0
        version: 6.1.0
      name: stable-6.1
    - currentCSV: loki-operator.v6.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.16-v4.19
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f401be018c5ba35e8ea1e68831cc20067f0f4d1be031d05b222548c1bdfb5f7e
          createdAt: "2025-02-18T18:01:39Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=6.0.0-0 <6.2.5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.2
          operators.operatorframework.io.bundle.channels.v1: stable-6.2
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f401be018c5ba35e8ea1e68831cc20067f0f4d1be031d05b222548c1bdfb5f7e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:99d1e7c5c1f0609b33de576884782550f969b28dc622810e6661d4c60a664215
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:8955d8186ae98095623f2083b6d5389d3ac9f2612fcfcf695835219c3e1216c4
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b8c5b46a94a6c85c9d382eeeb65a1d2b4014a7dc765e9abfba706c92857462dc
        version: 6.2.5
      entries:
      - name: loki-operator.v6.2.5
        version: 6.2.5
      - name: loki-operator.v6.2.4
        version: 6.2.4
      - name: loki-operator.v6.2.3
        version: 6.2.3
      - name: loki-operator.v6.2.2
        version: 6.2.2
      - name: loki-operator.v6.2.1
        version: 6.2.1
      - name: loki-operator.v6.2.0
        version: 6.2.0
      name: stable-6.2
    defaultChannel: stable-6.2
    packageName: loki-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-route53-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-route53-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/route53-controller:1.1.1
          createdAt: "2025-09-19T19:15:31Z"
          description: AWS Route53 controller is a service controller for managing
            Route53 resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HealthCheck represents the state of an AWS route53 HealthCheck
              resource.
            displayName: HealthCheck
            kind: HealthCheck
            name: healthchecks.route53.services.k8s.aws
            version: v1alpha1
          - description: HostedZone represents the state of an AWS route53 HostedZone
              resource.
            displayName: HostedZone
            kind: HostedZone
            name: hostedzones.route53.services.k8s.aws
            version: v1alpha1
          - description: RecordSet represents the state of an AWS route53 RecordSet
              resource.
            displayName: RecordSet
            kind: RecordSet
            name: recordsets.route53.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Route53 resources in AWS from within your Kubernetes cluster.

          **About Amazon Route53**

          Amazon Route 53 is a highly available and scalable Domain Name System (DNS) web service. You can use Route 53 to perform three main functions in any combination: domain registration, DNS routing, and health checking.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Route53
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - route53
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Route53 Developer Resources
          url: https://docs.aws.amazon.com/route53/index.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/route53-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-route53-controller.v1.1.1
        version: 1.1.1
      - name: ack-route53-controller.v1.1.0
        version: 1.1.0
      - name: ack-route53-controller.v1.0.6
        version: 1.0.6
      - name: ack-route53-controller.v1.0.5
        version: 1.0.5
      - name: ack-route53-controller.v1.0.3
        version: 1.0.3
      - name: ack-route53-controller.v1.0.2
        version: 1.0.2
      - name: ack-route53-controller.v0.1.0
        version: 0.1.0
      - name: ack-route53-controller.v0.0.28
        version: 0.0.28
      - name: ack-route53-controller.v0.0.27
        version: 0.0.27
      - name: ack-route53-controller.v0.0.26
        version: 0.0.26
      - name: ack-route53-controller.v0.0.25
        version: 0.0.25
      - name: ack-route53-controller.v0.0.24
        version: 0.0.24
      - name: ack-route53-controller.v0.0.23
        version: 0.0.23
      - name: ack-route53-controller.v0.0.22
        version: 0.0.22
      - name: ack-route53-controller.v0.0.20
        version: 0.0.20
      - name: ack-route53-controller.v0.0.19
        version: 0.0.19
      - name: ack-route53-controller.v0.0.18
        version: 0.0.18
      - name: ack-route53-controller.v0.0.17
        version: 0.0.17
      - name: ack-route53-controller.v0.0.16
        version: 0.0.16
      - name: ack-route53-controller.v0.0.15
        version: 0.0.15
      - name: ack-route53-controller.v0.0.14
        version: 0.0.14
      - name: ack-route53-controller.v0.0.13
        version: 0.0.13
      - name: ack-route53-controller.v0.0.9
        version: 0.0.9
      - name: ack-route53-controller.v0.0.6
        version: 0.0.6
      - name: ack-route53-controller.v0.0.4
        version: 0.0.4
      - name: ack-route53-controller.v0.0.3
        version: 0.0.3
      - name: ack-route53-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-route53-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://www.medik8s.io/
    name: self-node-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: self-node-remediation.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/self-node-remediation-operator:v0.9.0
          createdAt: "2024-07-25 14:40:52"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.9.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Medik8s
          url: https://www.medik8s.io/
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/medik8s/self-node-remediation-operator:v0.9.0
        version: 0.9.0
      entries:
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      - name: self-node-remediation.v0.7.0
        version: 0.7.0
      - name: self-node-remediation.v0.6.1
        version: 0.6.1
      - name: self-node-remediation.v0.6.0
        version: 0.6.0
      - name: self-node-remediation.v0.5.0
        version: 0.5.0
      - name: self-node-remediation.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: self-node-remediation
    provider:
      name: Medik8s
      url: https://www.medik8s.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudfront-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudfront-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "CachePolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "Distribution",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudfront-controller:1.2.1
          createdAt: "2025-09-19T21:12:14Z"
          description: AWS CloudFront controller is a service controller for managing
            CloudFront resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CachePolicy represents the state of an AWS cloudfront CachePolicy
              resource.
            displayName: CachePolicy
            kind: CachePolicy
            name: cachepolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: Distribution represents the state of an AWS cloudfront Distribution
              resource.
            displayName: Distribution
            kind: Distribution
            name: distributions.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: Function represents the state of an AWS cloudfront Function
              resource.
            displayName: Function
            kind: Function
            name: functions.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: OriginAccessControl represents the state of an AWS cloudfront
              OriginAccessControl resource.
            displayName: OriginAccessControl
            kind: OriginAccessControl
            name: originaccesscontrols.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: OriginRequestPolicy represents the state of an AWS cloudfront
              OriginRequestPolicy resource.
            displayName: OriginRequestPolicy
            kind: OriginRequestPolicy
            name: originrequestpolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: ResponseHeadersPolicy represents the state of an AWS cloudfront
              ResponseHeadersPolicy resource.
            displayName: ResponseHeadersPolicy
            kind: ResponseHeadersPolicy
            name: responseheaderspolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: VPCOrigin represents the state of an AWS cloudfront VPCOrigin
              resource.
            displayName: VPCOrigin
            kind: VPCOrigin
            name: vpcorigins.cloudfront.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudFront resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudFront**

          Amazon CloudFront is a web service that speeds up distribution of your static and dynamic web content, such as .html, .css, .js, and image files, to your users. CloudFront delivers your content through a worldwide network of data centers called edge locations.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon CloudFront
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudfront
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudFront Developer Resources
          url: https://aws.amazon.com/cloudfront/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudfront maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudfront-controller:1.2.1
        version: 1.2.1
      entries:
      - name: ack-cloudfront-controller.v1.2.1
        version: 1.2.1
      - name: ack-cloudfront-controller.v1.2.0
        version: 1.2.0
      - name: ack-cloudfront-controller.v1.1.6
        version: 1.1.6
      - name: ack-cloudfront-controller.v1.1.4
        version: 1.1.4
      - name: ack-cloudfront-controller.v1.1.3
        version: 1.1.3
      - name: ack-cloudfront-controller.v1.1.2
        version: 1.1.2
      - name: ack-cloudfront-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudfront-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudfront-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudfront-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudfront-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudfront-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudfront-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudfront-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudfront-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudfront-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudfront-controller.v0.0.15
        version: 0.0.15
      - name: ack-cloudfront-controller.v0.0.14
        version: 0.0.14
      - name: ack-cloudfront-controller.v0.0.13
        version: 0.0.13
      - name: ack-cloudfront-controller.v0.0.12
        version: 0.0.12
      - name: ack-cloudfront-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudfront-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudfront-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudfront-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudfront-controller.v0.0.6
        version: 0.0.6
      - name: ack-cloudfront-controller.v0.0.5
        version: 0.0.5
      - name: ack-cloudfront-controller.v0.0.4
        version: 0.0.4
      - name: ack-cloudfront-controller.v0.0.3
        version: 0.0.3
      - name: ack-cloudfront-controller.v0.0.2
        version: 0.0.2
      - name: ack-cloudfront-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudfront-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/openshift/lightspeed-service
    name: lightspeed-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lightspeed-operator.v1.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ols.openshift.io/v1alpha1",
                "kind": "OLSConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightspeed-operator",
                    "app.kubernetes.io/instance": "olsconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "olsconfig",
                    "app.kubernetes.io/part-of": "lightspeed-operator"
                  },
                  "name": "cluster"
                },
                "spec": {
                  "llm": {
                    "providers": [
                      {
                        "credentialsSecretRef": {
                          "name": "credentials"
                        },
                        "models": [
                          {
                            "name": "gpt-3.5-turbo-1106"
                          }
                        ],
                        "name": "OpenAI"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-08-28T16:37:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lightspeed
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Kubernetes Engine", "OpenShift Virtualization
            Engine"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lightspeed-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Red Hat OpenShift Lightspeed instance. OLSConfig is the Schema
              for the olsconfigs API
            displayName: OLSConfig
            kind: OLSConfig
            name: olsconfigs.ols.openshift.io
            version: v1alpha1
        description: |-
          OpenShift Lightspeed Operator provides generative AI-based virtual assistant which integrates into the OpenShift web console. OpenShift Lightspeed can answer natural language questions related to OpenShift Container Platform.

          To configure the OpenShift Lightspeed Operator, you need a Large Language Model (LLM) . You can host the LLM using RHELAI or RHOAI or use one of the models from the following Software-as-a-Service providers:

          - OpenAI
          - Microsoft Azure OpenAI
          - IBM WatsonX

          For more information, see [About OpenShift Lightspeed](https://docs.redhat.com/en/documentation/red_hat_openshift_lightspeed/1.0tp1/html/about/ols-about-openshift-lightspeed) in the official product documentation.

          Questions, comments, concerns, or just want to talk to the OpenShift Lightspeed team? [Email us](mailto:openshift-lightspeed-contact-requests@redhat.com)!
        displayName: OpenShift Lightspeed Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ai
        - assistant
        - openshift
        - llm
        links:
        - name: Lightspeed Operator
          url: https://github.com/openshift/lightspeed-operator
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/lightspeed-service
        relatedImages:
        - registry.redhat.io/openshift-lightspeed/lightspeed-rhel9-operator@sha256:b12803642df4d9b3ad0cf800315021483f543c6a8827b3901e57bdb65f10dc48
        - registry.redhat.io/openshift-lightspeed/openshift-mcp-server-rhel9@sha256:f1e4521ff2cc076f361c04d8c9aeac09ffcedb89a2f1dbe71f85caa27472adab
        - registry.redhat.io/openshift-lightspeed/lightspeed-service-api-rhel9@sha256:ec359967a87bc7ee2c1d6a61e553acbdb0dc85d6f240a1865bf57d3ba5dc7193
        - registry.redhat.io/openshift-lightspeed/lightspeed-console-plugin-rhel9@sha256:dfc42c2651ac8793f61ff0e54334ecb006dc4a77487bc7c6878cdf47bd4ac080
        version: 1.0.5
      entries:
      - name: lightspeed-operator.v1.0.5
        version: 1.0.5
      name: alpha
    - currentCSV: lightspeed-operator.v1.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ols.openshift.io/v1alpha1",
                "kind": "OLSConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightspeed-operator",
                    "app.kubernetes.io/instance": "olsconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "olsconfig",
                    "app.kubernetes.io/part-of": "lightspeed-operator"
                  },
                  "name": "cluster"
                },
                "spec": {
                  "llm": {
                    "providers": [
                      {
                        "credentialsSecretRef": {
                          "name": "credentials"
                        },
                        "models": [
                          {
                            "name": "gpt-3.5-turbo-1106"
                          }
                        ],
                        "name": "OpenAI"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-08-28T16:37:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lightspeed
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Kubernetes Engine", "OpenShift Virtualization
            Engine"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lightspeed-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Red Hat OpenShift Lightspeed instance. OLSConfig is the Schema
              for the olsconfigs API
            displayName: OLSConfig
            kind: OLSConfig
            name: olsconfigs.ols.openshift.io
            version: v1alpha1
        description: |-
          OpenShift Lightspeed Operator provides generative AI-based virtual assistant which integrates into the OpenShift web console. OpenShift Lightspeed can answer natural language questions related to OpenShift Container Platform.

          To configure the OpenShift Lightspeed Operator, you need a Large Language Model (LLM) . You can host the LLM using RHELAI or RHOAI or use one of the models from the following Software-as-a-Service providers:

          - OpenAI
          - Microsoft Azure OpenAI
          - IBM WatsonX

          For more information, see [About OpenShift Lightspeed](https://docs.redhat.com/en/documentation/red_hat_openshift_lightspeed/1.0tp1/html/about/ols-about-openshift-lightspeed) in the official product documentation.

          Questions, comments, concerns, or just want to talk to the OpenShift Lightspeed team? [Email us](mailto:openshift-lightspeed-contact-requests@redhat.com)!
        displayName: OpenShift Lightspeed Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ai
        - assistant
        - openshift
        - llm
        links:
        - name: Lightspeed Operator
          url: https://github.com/openshift/lightspeed-operator
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/lightspeed-service
        relatedImages:
        - registry.redhat.io/openshift-lightspeed/lightspeed-service-api-rhel9@sha256:ec359967a87bc7ee2c1d6a61e553acbdb0dc85d6f240a1865bf57d3ba5dc7193
        - registry.redhat.io/openshift-lightspeed/lightspeed-console-plugin-rhel9@sha256:dfc42c2651ac8793f61ff0e54334ecb006dc4a77487bc7c6878cdf47bd4ac080
        - registry.redhat.io/openshift-lightspeed/openshift-mcp-server-rhel9@sha256:f1e4521ff2cc076f361c04d8c9aeac09ffcedb89a2f1dbe71f85caa27472adab
        - registry.redhat.io/openshift-lightspeed/lightspeed-rhel9-operator@sha256:b12803642df4d9b3ad0cf800315021483f543c6a8827b3901e57bdb65f10dc48
        version: 1.0.5
      entries:
      - name: lightspeed-operator.v1.0.5
        version: 1.0.5
      name: stable
    defaultChannel: stable
    packageName: lightspeed-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/openshift/lightspeed-service
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kubevirt-hyperconverged
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubevirt-hyperconverged-operator.v4.16.23
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:eaacb0d177fd64e2eefbcb2dc47633dadf4d54fcbe0f0f750b078ab3f4f189e1
          createdAt: "2025-09-29 21:23:00"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.14.1 <4.16.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:3e1c68d5695bab96c382661a33581e60597b59afd21660ec760d8fe5e20ff9e3
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.16.23-1,hyperconverged-cluster-operator-test-rhel9:v4.16.23-1,kubevirt-common-instancetypes-test-rhel9:v4.16.23-1,kubevirt-tekton-tasks-test-rhel9:v4.16.23-1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:e24191be577575fa9d462c87c6f51775a0cf0f5d1812d12c4c95a68ec34b34ff
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:6e4aa75bece3c2fe076e282b3f1ad8296299ec0551e379c7c88830a0ce8ed5bf
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:1be66eee321bc47dcc0e7ff4674a485b9a4552243080d796d5f3b0d417721dc4
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:28c0bdf60404fb435cfa0cfdc6d314f789417c44c3206ce8cd1e7b99031fb529
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:659abe1d740607b2cb3850c9ab29101fa5d5c821d2a09aab5994849697a504f3
        - registry.redhat.io/container-native-virtualization/mtq-operator-rhel9@sha256:9de5ded35edfbe98121bdd4404c84dcc72c8cba1742dc27df6382ef882db2183
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:0f0478e9357f8b6ace6d844e6e424db192398195a8cea2fa6398be8996d05be6
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:2843d199dc2748749d5f48fb9f90656962e011533d79dc7e0a4f3592b12aa69a
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:6b0c7cc17abd95fce45b6ec83ec291e843eedbc90dce0ecc449aec6e64562057
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:97d37927164f0ea986f209912e6247fdab37b680c214afc4bdbb7051caaf58b1
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:cc4ff488aa97e179c4df39ac5ae8085070017106adf1c5f2222b35d8f2234ea8
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:f8e2ea3988d4a77a4e809e1fd073ffa5f99393ca3eb69a2c1cb54dd4ab60e482
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:885670f2e9ad43224cbc96f34b23dd2642de00f688bade0f6b89dd3d3fb83dbd
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:251d6fb3197a7642d86193191aa69c85a45c5521d7fc509704edd46d5d4e90a7
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:fcbadecf01b81625d55f9219428baabc33ee7231b4ccf801c90dee3095907187
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:a0af0e5d4b142f28dc72c5636df5040c4ce92c6b12ab183d3a9bc7a9d2d2f418
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:c033ddd0ed30c0d25f4c2e9b30fca4e40959bb6d74029a1f452ff048b97de88a
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:7206ac7959c07f474f454ab48bcf3522ade95efb27fd9fbd3938d1387a1420af
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:468f117a1b55af59e72a9ff46499a8d0bb00884d450d57b692e2f7b557c9bf3d
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:e01871006b9e4cb1b6f477319de63b758eac672a70a298caed2032ec19986680
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:7587832b877f32621eb26e67a5c516728f293ecd98f74dd70d5f03474a5429d6
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:68b650d6779fc4d61f05d97e6d62b80fcd638721994153205eec68f8c1d999ce
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:19ca70dceb79b10ff458f0ea41faa6e5552a5f08828620ad7177efa34166a90e
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:84dbfa51755d60e34b194c1acf679ac083feb49038112212828fe8101dcec2a5
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:8462188ec7533d2105c0dc9154d8dcf1425b96d6945eef16ea774a5dff9bba1d
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:11030a8d8cd03efdf27fe3cfc9ebeadeb20851b6de1e9d8a93ba7f2bc6a77a96
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:8cf7993326ba220f650ae2fa5383d0d6cf3e2c7f3346f0a339a9d212b8df02f6
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:c18b6e315f5577cc18d46cd76595d48d4a49026a6b2f5763c3efb26919bde5eb
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:52b4f7c7a4f5c5c5d07a720c5eefa70709f520b1d939aec4cd971d8f3c9b12bc
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:ef8260d72239ad2126ca023250bfd976f5c050fe9c0727a0f7ef007446f5363b
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:18be4ef8b6c2cedab4a2e031939fc1f9cc23020bde7fccb7997e4222b673f57f
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:def80efdcc58d545aaf4b69b784a348cb36e97afd8e0124480521a2b980fa7c3
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:b031a12e73cb75e207203fd3785619e3a5a22d7d5a8549488c035f70374235be
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:14ea56d4149b72b98b6e7f5b33ba5bdd9115e7c643fd103c9df7a26b2a06700a
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:9e358366fd11dd167255abd930016aa4fc35b1b9dc69ce27316635488efc8ec4
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:6bf45534310768308958428692b29afa25f2740d0c62165d7182b0b3e9e944cf
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:eaacb0d177fd64e2eefbcb2dc47633dadf4d54fcbe0f0f750b078ab3f4f189e1
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:91b86228495f38d5528be1a8ffec96695ad72fafb76d2ead87f7bee25540d8b6
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:19b56bbb081db7198d570c3c691e142991aaa81857ee417d47903291ed0e7a1c
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:372652529b6a8705df581e7ab8acc2ec5c3cdf8fb273448052c59963abe4d984
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:7f89f2ba59309b26cf83f58ff3ba35569182097258fa543d5d88f3a8cd64ca00
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:bf1bb2fe09e028506fe7cf383368d76e358b8c59ecc74d3618bc250c1e97271c
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:a79579b18a7df9f23c584303593a9404c64df023458d55c715aeb71dcacc92a0
        - registry.redhat.io/container-native-virtualization/mtq-lock-server-rhel9@sha256:bb1dfc204e48a13985d6c2ec9cdf17f47330617725ee6dd854f0dae0f1901ba4
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:4005d165f1485eec98ca24ce41bef79cd8088002d43ffb709ec65f842a176bdc
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:3f3554ef043c80494db536b28eb03fa7043e53bb85ce426820c252fb999ce04e
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:531702869a34f76850d332efb84ccf571fafa281014afd1787497ca9de8aed37
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:a081e7e68def13b7de369d3f66469035d61fd1feef1bee4f95ba5f7b4041d960
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:e7672121156e050f05af771f75cfe26558699da8a074bf7051f710705a27d743
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:bd532e59edae5d2edac78b9249cd6a9f0779f44e25b019d824a1b9803f078bde
        - registry.redhat.io/container-native-virtualization/mtq-controller-rhel9@sha256:a774f0bd125c66b45e35e327c9c7e0afd7947d3c5ed222b3d5d223941366ba2c
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:52e76b4208ee4e7da8a0729b9bdb294729454d0f62580d64dd104b6921f464ed
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:3b57db7aaa3c61335044ec78cbd061f78ce99cdb4b75b3ce342e1a603da4bcef
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:7e19b42a648cee71c60cb69fbcf6ebca05c938942e16d23f1cd61a619cffdcb6
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:4046e8e862e0c55f09e9031b95e6a77af4d491593e39770c644287d44888b0bb
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:517c88a6be8842d4eb71ad30f99ce1afca954af1df2c1bf4a9584f4ec69fcf12
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:3e1c68d5695bab96c382661a33581e60597b59afd21660ec760d8fe5e20ff9e3
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:cca121312777d2727fe2c50a14583243ce69e6e356b499f133736787522095b2
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:ad24eb52bd65495224661d4894336d88fde342aeaab16b1084ceff6337f92059
        version: 4.16.23
      entries:
      - name: kubevirt-hyperconverged-operator.v4.16.23
        version: 4.16.23
      - name: kubevirt-hyperconverged-operator.v4.16.22
        version: 4.16.22
      - name: kubevirt-hyperconverged-operator.v4.16.21
        version: 4.16.21
      - name: kubevirt-hyperconverged-operator.v4.16.20
        version: 4.16.20
      - name: kubevirt-hyperconverged-operator.v4.16.19
        version: 4.16.19
      - name: kubevirt-hyperconverged-operator.v4.16.18
        version: 4.16.18
      - name: kubevirt-hyperconverged-operator.v4.16.17
        version: 4.16.17
      - name: kubevirt-hyperconverged-operator.v4.16.16
        version: 4.16.16
      - name: kubevirt-hyperconverged-operator.v4.16.15
        version: 4.16.15
      - name: kubevirt-hyperconverged-operator.v4.16.14
        version: 4.16.14
      - name: kubevirt-hyperconverged-operator.v4.16.13
        version: 4.16.13
      - name: kubevirt-hyperconverged-operator.v4.16.12
        version: 4.16.12
      - name: kubevirt-hyperconverged-operator.v4.16.11
        version: 4.16.11
      - name: kubevirt-hyperconverged-operator.v4.16.10
        version: 4.16.10
      - name: kubevirt-hyperconverged-operator.v4.16.9
        version: 4.16.9
      - name: kubevirt-hyperconverged-operator.v4.16.8
        version: 4.16.8
      name: candidate
    - currentCSV: kubevirt-hyperconverged-operator.v4.99.0-0.1723448771
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:5b2c64218342a5744182fc06f550bcdcf954e931275a9d1505ef0bbfbd0f8044
          createdAt: "2024-08-12 07:46:45"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.14.1 <4.16.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:d74c629e415d177069a0570fb8db9252d2b193d2ec7972c844686ffd14033f62
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.99.0-73,kubevirt-common-instancetypes-test-rhel9:v4.99.0-32,kubevirt-tekton-tasks-test-rhel9:v4.99.0-31
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:cc190d7c44b20f10ae61709fd2f5f7c0b248df73b768eb696d5a6cec90a34e59
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:41ee360ac43ee69ccee62d4ac3d2f07c8a8589e5b85c659b47018dd36c2b932f
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:b410508ecc71b0a3bf4825793efde4b07085e2243d8b3585da852fd1940c4be1
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:820a9e6bd0329df5c541ef0b67b3c7cdd9f81be33de87c5b5e5e43777bea197e
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:811e8e7e9872408794ef36229c42c6c437b1a4ace76c7134d91bb2331fe7815b
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:0769512df0fa1325d76a5f10a1a704cf60fc65f62f57b157ea53bc9ba682851f
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:8c64b526fd42a17c909bf45eb795121a81d502424a9cf11f8087118751333894
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:5f23fc8e9bc386622460c23b9644f58648d9f748f5670927a02888cd36c99a21
        - registry.redhat.io/container-native-virtualization/mtq-controller-rhel9@sha256:ad5938ebd5f07196ce5c35232a1c3b0a653301a83929ce84c3d1cb4dd1b91816
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:37e6b4fa5ff436572412ec0bdb1edc4a49e62c994e056c93578605f6b74adbcd
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:07435335fdbf6178bc00cc02b93cc0adb42784339151b2d9c31dc8da98b8bc6a
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:cffec516ab9ddc50cab77339d0cb3239ef4e7fd43bd2afdc73370aa487ea5704
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:dec430b74887e9b42d639f14c07d9f58a4881cac24e949dab98f96af150f1958
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:c61732be8024b4ede45eec8f786e51b198632ee5bfb0c2f5af84f37544584283
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:5ba79dd6bdf9e3778d3e301919b37e45cfc1eaf72fda77eba6588788829ba29b
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:c14cd037d14489321629908fc6b30248f7a20d866619177a165d72cfc1b6b7ef
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:83c03929cf11d7202e6c6cb9dd767bd4d7473f4fc1d562b79719ceb3d8382e68
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:6b7beed3cba9093b3aeb3109f23ec38c20f5568915dff774ad5e9e6bb5db3a58
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:e39732ebd3a4452fa7a678aeea5e64a7973fd72802eff193b16680c2f0129cb1
        - registry.redhat.io/container-native-virtualization/mtq-lock-server-rhel9@sha256:01cc6797800ccfb99453858ab36c0100c5ca343f1f71f975914b2437d1f757ba
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:9822f7a1590d830ccd64cc0729d2e30a2e04fb2b022f54498e1f1abff3748c2c
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:bf2ec130d1f3f32314d740037313f49f874a271e8b9a7f6cd2b93698a8423e8d
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:42a34caa18b5f5452b9d5603602ea74b1ee59a8bbb52bddf18901fecc27aed8a
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:18dfb38223647e11020e3ef1093a695454b90cc50922f80afb612212b7d13309
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:4c66a26b9f74d83a9bc60edb40ffd9d6ff2382d20500677d4567f8e141c3ddf0
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:758a9f6309a367d3f31cc03dd89356a8232a0fbd277570c01ebfa1fa30089a8d
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:2e5d5a2a70cda1f414f22b1c713d77d068c5fb16b1487a572019a9aa7130e6be
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:84b199e98e283c863d52a7b00044c20cb56ad563494adbec2229af265bf3f7c5
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:6baf67f3f410c071cc26f675d86edbbae5eaeb3410888cf622fad384b43b80ea
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:3a45cd2f3ea25f356167ce4fa20894a2a0706b4bcffa42c806f6017812053466
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:84679ffe8ce86c674818531897d122f5568d10c7e1e2ff919185d5771e5b4c39
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:ac33de246ee42bbfe735d7349d85c6df7e583e7599dea7dd7a2235bd0e59845f
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:e7145f016642b02bd651e21aa19edc7f93d0d1c2da3c6c7eed5d1174cf23619c
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:2564af054f52a8bb340e0270379c72da18658662a2a5bd26ac7bd979ddfccb56
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:4cff43cc145361c74c357055d51060ccc27e6bca58ccc3b5f5b983c7b8545114
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:09a30b59c906ddc4604d70f18210bcebef1f90f8bf965422b0caba363cd88989
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:63389195cf584a07ddbc2efec90cf6deaf737b91314b67db2b9f53dcc7acb57c
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:6193e569d4ab43806ac7b4870ef3634a5de3687527d6d190e59a9f735fa4778c
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:f4e7876fa8b27a8082efa99455d9bf7b0c479cff1e4c8a26136bf63092699c0b
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:d7e8e075a08b148ddfa91b44bb9cae1ef823756760279c4bb6d8489bc7e0eb2c
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:bf30b6c7173ebb6a4592246af16db87d7d391182f52286b559f8bad8f4b6cc5b
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:15179f0826831f051fdb867067be7c5528f01f442ae421b8782ae0cd1ed1b73d
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:4b3ac5ec2fafef0b7702a02847996f0903ec8d4262ccd03e8fa74885265ea4e9
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:d6dfd062df3d0d49f7efb928b835b9c8e0bcfb514a653ba2896a0152f24a0a21
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:185a83c994aced74546fe630cd71656f1e4112cb15be75f41dfcb1b9ed0966bd
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:b59e34cdb6214eacb9dc176a7b33125d7c1871f6b81696d59646294bfaf749e7
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:8e19d99232be4c37e857a4931627fb0aed721893e27b0d0e522fcf13bf67c1c9
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:06cc71822684c3ed094f327553451daf076d6bb97065c314da1a7d2adbbef3d2
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:4bdafc4404d5cfd844d05f32d88c262fba56dc931bfd5522b78c25d4ff253883
        - registry.redhat.io/container-native-virtualization/kubevirt-ipam-controller-rhel9@sha256:f33ac4d7477a730f273a462f171aed35dea38374d5ea7ef0bae2d7f74413a757
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:af5d2a2d4f0b6f0e1a13f1029073bb56980d8a70503b21628f18e8986e9b531c
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:5b2c64218342a5744182fc06f550bcdcf954e931275a9d1505ef0bbfbd0f8044
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:88f02cbebd6c3a2add6e0f8ca744e42bbd182f5a07d4fac59cf5cf07a4edfd45
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:fa0478fbf91248166f0eed8ec5f8dbb5b31d26031438376d9301836b6a069643
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:9bfdaf1cca6eac15b63a4ea6b7de28f2328516c5cf8bd5727935dd729e45657d
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:30c669770b0b992a3c84bb81c892e70a24fb14af187e8a4ee19cb3e0dd56ddbb
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:c5e8f44f647f593b00d3427172932d7aa694379d124209b9c66b4c1ffe3d4116
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:e6d9a6c41433360abd42d1118f5af3df9592d6d36af5f65082ae214cc8513c1b
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:d74c629e415d177069a0570fb8db9252d2b193d2ec7972c844686ffd14033f62
        - registry.redhat.io/container-native-virtualization/mtq-operator-rhel9@sha256:033bf7a6ef8ced1ac7275a9ddea8b16a10340a0fb0fb39619d89354d661f05b6
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:442e5cbcc63961642b662544032240bffb3863227324128a67f0bb6a05a1c364
        version: 4.99.0-0.1723448771
      entries:
      - name: kubevirt-hyperconverged-operator.v4.99.0-0.1723448771
        version: 4.99.0-0.1723448771
      name: dev-preview
    - currentCSV: kubevirt-hyperconverged-operator.v4.16.21
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:08f15657a1eee02927c5a6ef661e6e703d0064e091751bd2d8fbad5f118d9672
          createdAt: "2025-09-20 19:53:21"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.14.1 <4.16.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:88863a8dd2145bf17f7c5378c81e7c0be7a01dc47ee75bd7e5203a8fed8895b9
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.16.21-3,hyperconverged-cluster-operator-test-rhel9:v4.16.21-1,kubevirt-common-instancetypes-test-rhel9:v4.16.21-1,kubevirt-tekton-tasks-test-rhel9:v4.16.21-1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:9afe1ec9f776854caae38b89198a82e8e47a45d6c08f8dbd65a06c85f741c6a0
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:611d60b6ea4285fbc8a9f945fa5d06c09679a8f7ec1d70ec18edae7eec1a6bb3
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:299825fa33ab3f491cf954dd2e4ba23fdba2a96723b9ac3d98b3264b0b9fbb28
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:f6b075a56380143a7892e1fe5fdb212f8aa1c974f6a2d7e7e9b940789e44c010
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:d346688e02d03ddd9918f7a344d6eb3d5ddac3f1802dc82daf72bc56e7a9d496
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:406681aff7ea0a17941449593f839418edffdb073e2afffd53174f0df6292f94
        - registry.redhat.io/container-native-virtualization/mtq-controller-rhel9@sha256:253ef56b2867d33b3a475a456878b4a60f56473122eeed8c169568a9724ed30c
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:e54d87e6735151ed3c5295b23ca2f7ec666afb6421923636ef836bca1cf883d6
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:6385b346b3dd8f2bb3284871d326c851cf743a5e3fc3c49a1e51a1d43b0eafb4
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:8cd1052b9f69d0a0d8db41e90414756e83cbd07d7bd03c2faa85c41393d6539c
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:b80e68822fb0b5e78a995a97e02fe2ec448d67a5aa2224497fdf12b812182a0b
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:51a1fb0eb77e4f5c8bdfaf3b5074369b93ef63f5a7af62e22ed0789731f8dc73
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:361a48df213d52ced2a9963c2a8bde72a0981f1c497235fb918ea8c065580eeb
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:355e0946cfa963aeb5b6c3e8d476a9c3aee202baf77b95ce3f69c8e4b9e73836
        - registry.redhat.io/container-native-virtualization/mtq-lock-server-rhel9@sha256:c447251fea804340047bbaff885ad56be7a0d2612af1de26765f6cfc37da25e6
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:3daa460ed347e6eaab290b0a7fb0a08da6f210fb8be88af33f6e256511af80b6
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:dcc73f518ca7557cbd84b236367b3ecb7baab405268d54543783b4b772e14568
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:f7e5e04adb0467eb0aa4691a68b0d6485db9a52f99494097bf5b598e08f2edfb
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:2a109f94282b76ca0ff929adc35c93c0a98e34b43aba248d2975957fa149133f
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:cca121312777d2727fe2c50a14583243ce69e6e356b499f133736787522095b2
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:a0af0e5d4b142f28dc72c5636df5040c4ce92c6b12ab183d3a9bc7a9d2d2f418
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:88863a8dd2145bf17f7c5378c81e7c0be7a01dc47ee75bd7e5203a8fed8895b9
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:8cf7993326ba220f650ae2fa5383d0d6cf3e2c7f3346f0a339a9d212b8df02f6
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:edf12760e7fe4c8d37c22e172c0da270c219b7ff16d4ebb6b62729545309f6fa
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:a040696505e266e8f06a5f06b108f2db5ad3a6ed803d547d736455c1928a6a0e
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:d9560ed2758d3793139d3e13b7e5ce9d0f4ac0f8a6a8767449822c033cb2a456
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:9d997314bc7c6b30c2c3ff54b27bcbc1cdeac6dba300cab8a898b97fbce05ff7
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:dbff3d1592b192253a288fb3230a0ff125564540a30a6dd5ac881e79ed93b60a
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:9ae64f17191b98bb75f613969a9415103dad304e4fd20fb444f4fc4e57a57e0e
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:14952f38ab1d7f0abe4d54883e2c93d01ec1240ae809aefa9ca8cb4aaef080e1
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:15a971cef680f6f43e6b8ad7e0f5dab572ebacebc35cfbc6f7949a4ccd7789c2
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:d41ece8df9482ff860706edcdf27b53ac97d7c87407b3acfe8b1d9adfbf40e38
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:7581cd2330c20be2b77621e0e14da183d2b94f40fa5d3f448dede5f8c182759e
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:6fbdfdf3b9206058d5efc6653194bf6c42c07655817e3c97044da678e84b310a
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:3aa57a7ee9b342844ea4aa20bbbdf7a43b5d41f7866baa619f25473a230011f3
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:91b86228495f38d5528be1a8ffec96695ad72fafb76d2ead87f7bee25540d8b6
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:2e240514425d70039f5b48fe1908c56fae6efcc3eff93c63ec67f4aa3318a5fb
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:69388282df18b68c1d937ac98a5026b34221a54e5ce4b9cf4b89379b66623488
        - registry.redhat.io/container-native-virtualization/mtq-operator-rhel9@sha256:e29ffbe332d5da66ab4612a50ed73813ccbde92622474273b617cccd3bc24cd7
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:d253f1dbb54c70cf5874c2177bcd1cc3697381916cd4cf825bda934bfdb3d53c
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:7752f41a0f380fcd0d2652d286d4aac1ebb94a8a5fbb655b2e7058054f2eca4d
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:b20230f9a74652cf676bcf05dad81fc1ff64f9cff7232782c62cf585db48b115
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:0c5e1ee8ae0f19237a2c8b93466a7fc217a112cf1fb6dd55e2be27f95c9b66a3
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:bf7562a3bbc7655565b69fc96f2deae15729bbdef158cd90b97fc5ecc58d5bd8
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:fd2b4a337a8e45061cb6f455de3eefe27b004844c724fd821efc3b47215b74d3
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:77211068f3a8d331f1b21c64918861398f2cdbb2a3b760a5e8df51c89efea5b9
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:cddcdbf1d02b6e55cd93787991b5cf8704d0fb3a4a0bef5f58c918b57e26ee16
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:387fba10dc8f16fbfbc8c84274c5f1f5f6d01cf11f17f1b54fa0775a3300c6de
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:53ae4db672daad95d99f2955545ba335095c96630efc2a2ff19c5019a9413e87
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:63262574fa7d51d3b8c26b79e22d78f462c70239746aec6b8b3b951a02919e81
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:4046e8e862e0c55f09e9031b95e6a77af4d491593e39770c644287d44888b0bb
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:42fc0d1d112b30f65d6a48fa0442e2f41872658d2431ac8618dadf0aa7a41208
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:9e87e7d8679211a317d11e588868b40235da8c668bfb2ee10b564f6850fb91d1
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:38e767770ea0a74c4d594a382eed72bf9efa527e45ac6438acb79a1039f57fbd
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:75d76d7274f5169a414d8024125d4f0d8709888ac2b5d6121360635c7967d878
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:96092abe951f495f4f7dd0bdcb371b7efcb03a78a15e3ed8f1b37502203a6316
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:08f15657a1eee02927c5a6ef661e6e703d0064e091751bd2d8fbad5f118d9672
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:7e19b42a648cee71c60cb69fbcf6ebca05c938942e16d23f1cd61a619cffdcb6
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:878e1c8bd72cb3002457b6fbcddca99c0f4b00f6231fea52387d540b12e6329f
        version: 4.16.21
      entries:
      - name: kubevirt-hyperconverged-operator.v4.16.21
        version: 4.16.21
      - name: kubevirt-hyperconverged-operator.v4.16.17
        version: 4.16.17
      - name: kubevirt-hyperconverged-operator.v4.16.13
        version: 4.16.13
      - name: kubevirt-hyperconverged-operator.v4.16.7
        version: 4.16.7
      - name: kubevirt-hyperconverged-operator.v4.16.6
        version: 4.16.6
      - name: kubevirt-hyperconverged-operator.v4.16.5
        version: 4.16.5
      - name: kubevirt-hyperconverged-operator.v4.16.4
        version: 4.16.4
      - name: kubevirt-hyperconverged-operator.v4.16.3
        version: 4.16.3
      - name: kubevirt-hyperconverged-operator.v4.16.2
        version: 4.16.2
      - name: kubevirt-hyperconverged-operator.v4.16.1
        version: 4.16.1
      - name: kubevirt-hyperconverged-operator.v4.16.0
        version: 4.16.0
      - name: kubevirt-hyperconverged-operator.v4.15.2
        version: 4.15.2
      - name: kubevirt-hyperconverged-operator.v4.15.1
        version: 4.15.1
      - name: kubevirt-hyperconverged-operator.v4.15.0
        version: 4.15.0
      - name: kubevirt-hyperconverged-operator.v4.14.3
        version: 4.14.3
      - name: kubevirt-hyperconverged-operator.v4.14.2
        version: 4.14.2
      - name: kubevirt-hyperconverged-operator.v4.14.1
        version: 4.14.1
      - name: kubevirt-hyperconverged-operator.v4.14.0
        version: 4.14.0
      - name: kubevirt-hyperconverged-operator.v4.13.4
        version: 4.13.4
      - name: kubevirt-hyperconverged-operator.v4.13.3
        version: 4.13.3
      - name: kubevirt-hyperconverged-operator.v4.13.2
        version: 4.13.2
      - name: kubevirt-hyperconverged-operator.v4.13.1
        version: 4.13.1
      - name: kubevirt-hyperconverged-operator.v4.13.0
        version: 4.13.0
      - name: kubevirt-hyperconverged-operator.v4.12.2
        version: 4.12.2
      - name: kubevirt-hyperconverged-operator.v4.12.1
        version: 4.12.1
      - name: kubevirt-hyperconverged-operator.v4.12.0
        version: 4.12.0
      name: stable
    defaultChannel: stable
    packageName: kubevirt-hyperconverged
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudtrail-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudtrail-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudtrail.services.k8s.aws/v1alpha1",
                "kind": "EventDataStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudtrail.services.k8s.aws/v1alpha1",
                "kind": "Trail",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudtrail-controller:1.1.1
          createdAt: "2025-09-19T21:16:25Z"
          description: AWS CloudTrail controller is a service controller for managing
            CloudTrail resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventDataStore represents the state of an AWS cloudtrail
              EventDataStore resource.
            displayName: EventDataStore
            kind: EventDataStore
            name: eventdatastores.cloudtrail.services.k8s.aws
            version: v1alpha1
          - description: Trail represents the state of an AWS cloudtrail Trail resource.
            displayName: Trail
            kind: Trail
            name: trails.cloudtrail.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Compute Cloud (CloudTrail) resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudTrail**

          AWS CloudTrail is an AWS service that helps you enable operational and risk auditing, governance, and compliance of your AWS account. Actions taken by a user, role, or an AWS service are recorded as events in CloudTrail. Events include actions taken in the AWS Management Console, AWS Command Line Interface, and AWS SDKs and APIs.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudTrail
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudtrail
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudTrail Developer Resources
          url: https://aws.amazon.com/cloudtrail/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudtrail maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudtrail-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-cloudtrail-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudtrail-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudtrail-controller.v1.0.28
        version: 1.0.28
      - name: ack-cloudtrail-controller.v1.0.26
        version: 1.0.26
      - name: ack-cloudtrail-controller.v1.0.25
        version: 1.0.25
      - name: ack-cloudtrail-controller.v1.0.24
        version: 1.0.24
      - name: ack-cloudtrail-controller.v1.0.23
        version: 1.0.23
      - name: ack-cloudtrail-controller.v1.0.22
        version: 1.0.22
      - name: ack-cloudtrail-controller.v1.0.21
        version: 1.0.21
      - name: ack-cloudtrail-controller.v1.0.20
        version: 1.0.20
      - name: ack-cloudtrail-controller.v1.0.19
        version: 1.0.19
      - name: ack-cloudtrail-controller.v1.0.16
        version: 1.0.16
      - name: ack-cloudtrail-controller.v1.0.15
        version: 1.0.15
      - name: ack-cloudtrail-controller.v1.0.14
        version: 1.0.14
      - name: ack-cloudtrail-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudtrail-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudtrail-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudtrail-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudtrail-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudtrail-controller.v1.0.5
        version: 1.0.5
      - name: ack-cloudtrail-controller.v1.0.4
        version: 1.0.4
      - name: ack-cloudtrail-controller.v1.0.3
        version: 1.0.3
      - name: ack-cloudtrail-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudtrail-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudtrail-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudtrail-controller.v0.1.3
        version: 0.1.3
      - name: ack-cloudtrail-controller.v0.1.2
        version: 0.1.2
      - name: ack-cloudtrail-controller.v0.1.1
        version: 0.1.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudtrail-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Provider Name
      provider-url: https://your.domain
    name: smilecdr-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smilecdr-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.smilecdr.com/v1alpha1",
                "kind": "Smilecdr",
                "metadata": {
                  "name": "smilecdr"
                },
                "spec": {
                  "labels": {
                    "application": "smilecdr",
                    "client": "smilecdr",
                    "env": "dev",
                    "version": "one"
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Smilecdr
            name: smilecdrs.operator.smilecdr.com
            version: v1alpha1
        description: Smilecdr Operator description. TODO.
        displayName: Smilecdr Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - smilecdr-operator
        links:
        - name: Smilecdr Operator
          url: https://smilecdr-operator.domain
        maintainers:
        - email: your@email.com
          name: Maintainer Name
        maturity: alpha
        provider:
          name: Provider Name
          url: https://your.domain
        relatedImages:
        - quay.io/jibran_shaukat/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - quay.io/jibran_shaukat/smilecdr@sha256:cd6b2cec6f5b9ad9e0f0434532f8fb00b6919fcd840e686d96ce08874f2154db
        version: 0.0.1
      entries:
      - name: smilecdr-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: smilecdr-operator
    provider:
      name: Provider Name
      url: https://your.domain
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: model-builder-for-vision-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-modelbuilder-for-vision-rhmp.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "Modelbuilder",
                "metadata": {
                  "name": "modelbuilder-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "backup_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "metadata_storage": {
                    "storage_class": "ibmc-block-bronze",
                    "storage_size": "20G"
                  },
                  "photo_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "500G"
                  },
                  "in_memory_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              },
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "ModelbuilderAWS",
                "metadata": {
                  "name": "modelbuilder-aws-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "postgres": {
                    "storage_size": "20G",
                    "storage_class": "gp2",
                    "backup_storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
          createdAt: "2021-08-28T12:00:00Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/model-builder-for-vision-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/model-builder-for-vision-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.0.3
          operators.operatorframework.io/builder: operator-sdk-v1.11.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/ibm-modelbuilder-for-vision/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Model Builder for Vision on AWS
            displayName: Deploy-Modelbuilder on AWS
            kind: ModelbuilderAWS
            name: modelbuildersaws.modelbuilder.com
            version: v1alpha1
          - description: Deploy Model Builder for Vision on IBM
            displayName: Deploy-Modelbuilder on IBM
            kind: Modelbuilder
            name: modelbuilders.modelbuilder.com
            version: v1alpha1
        description: IBM Services Software Model Builder for Vision is an AI training
          platform that can use the capabilities of IBM Cloud and Amazon Web Services
          and their GPUs to quickly train computer vision models for compatible mobile
          apps. IBM Services Software Inspection Workbench iOS and iPadOS app will
          be used as the exclusive UI for labeling and training. Specifically, IBM
          Model Builder for Vision will train computer vision models that can then
          be deployed for use on the IBM Services Software Inspector Portable , IBM
          Services Software Inspector Wearable and the IBM Maximo Visual Inspection
          Mobile apps.
        displayName: IBM Model Builder for Vision
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Model Builder
        - AI
        - Computer Vision
        - Inspection
        - Mobile
        - Core ML
        - TensorFlow Lite
        links:
        - name: IBM Model Builder for Vision
          url: http://ibm.biz/model_builder
        maintainers:
        - email: isshelp@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.connect.redhat.com/ibm/modelbuilder-augmentor@sha256:ef5d0dde3eb334d919c2773af63007f1ef5e735ae040ef77fffe3440f854f068
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/rhel8/redis-6@sha256:2f65f46b25d84a3cb41a6a10f93786bcab47b7354ca9ac13a8e602ea833eba60
        - registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
        - registry.connect.redhat.com/ibm/modelbuilder-image-server@sha256:ecdb665af651417c50335e7259b89ef8096521abec431976bd7abe4c1edc49d6
        - registry.connect.redhat.com/ibm/modelbuilder-auth-service@sha256:f96b444fdcf7781775c9eba942bcc5e5ebdcf5f50820447f5eb26061c4c674cd
        - registry.connect.redhat.com/ibm/modelbuilder-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/modelbuilder-cv-training-service@sha256:c3c9080b08b87022c784fd98ac3e2774ec89ce29a2dd19d91dee039e3fdb2b36
        version: 1.0.3
      entries:
      - name: ibm-modelbuilder-for-vision-rhmp.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: model-builder-for-vision-certified-rhmp
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: fuse-console
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-console.v7.10.3-0.1655306919.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "replicas": 1,
                  "version": "1.10",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "100Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:37dc38375f5d553ac1d9f20f66ce005ed9495354b81d7915ae3bd7dd6af132f3
          createdAt: "2020-09-14T10:59:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.10.3'
          olm.substitutesFor: fuse-console.v7.10.3-0.1652362834.p
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.10/html/managing_fuse
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:37dc38375f5d553ac1d9f20f66ce005ed9495354b81d7915ae3bd7dd6af132f3
        version: 7.10.3+0.1655306919.p
      entries:
      - name: fuse-console.v7.10.3
        version: 7.10.3
      - name: fuse-console.v7.10.3-0.1652362834.p
        version: 7.10.3+0.1652362834.p
      - name: fuse-console.v7.10.3-0.1655306919.p
        version: 7.10.3+0.1655306919.p
      name: 7.10.x
    - currentCSV: fuse-console.v7.11.1-0.1681224695.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.11",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:518b5ace9178e2aafd7b6f247fd41d9392abd1a7bccb4cb53783531820611193
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.11.1'
          olm.substitutesFor: fuse-console.v7.11.1-0.1675972857.p
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.11/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:518b5ace9178e2aafd7b6f247fd41d9392abd1a7bccb4cb53783531820611193
        version: 7.11.1+0.1681224695.p
      entries:
      - name: fuse-console.v7.11.1
        version: 7.11.1
      - name: fuse-console.v7.11.1-0.1669922044.p
        version: 7.11.1+0.1669922044.p
      - name: fuse-console.v7.11.1-0.1675972857.p
        version: 7.11.1+0.1675972857.p
      - name: fuse-console.v7.11.1-0.1681224695.p
        version: 7.11.1+0.1681224695.p
      name: 7.11.x
    - currentCSV: fuse-console.v7.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.12",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:4e6889bf76d323956e0b594db7381e8162fea6b108f53f9e6dd0b38cf3596c8d
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.12.1'
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.12/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:4e6889bf76d323956e0b594db7381e8162fea6b108f53f9e6dd0b38cf3596c8d
        version: 7.12.1
      entries:
      - name: fuse-console.v7.12.1
        version: 7.12.1
      name: 7.12.x
    - currentCSV: fuse-console.v7.13.0-0.1757604404.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.13",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:72d98254a821bce237b3dde896f2825628cb7b6df564e23949337590586e8bea
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.7.0 <7.13.0'
          olm.substitutesFor: fuse-console.v7.13.0-0.1754559584.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.13/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:72d98254a821bce237b3dde896f2825628cb7b6df564e23949337590586e8bea
        version: 7.13.0+0.1757604404.p
      entries:
      - name: fuse-console.v7.13.0-0.1744281421.p
        version: 7.13.0+0.1744281421.p
      - name: fuse-console.v7.13.0-0.1723709432.p
        version: 7.13.0+0.1723709432.p
      - name: fuse-console.v7.13.0-0.1757604404.p
        version: 7.13.0+0.1757604404.p
      - name: fuse-console.v7.13.0-0.1721224332.p
        version: 7.13.0+0.1721224332.p
      - name: fuse-console.v7.13.0-0.1744737026.p
        version: 7.13.0+0.1744737026.p
      - name: fuse-console.v7.13.0-0.1724659726.p
        version: 7.13.0+0.1724659726.p
      - name: fuse-console.v7.13.0-0.1725026290.p
        version: 7.13.0+0.1725026290.p
      - name: fuse-console.v7.13.0-0.1748960166.p
        version: 7.13.0+0.1748960166.p
      - name: fuse-console.v7.13.0-0.1738056254.p
        version: 7.13.0+0.1738056254.p
      - name: fuse-console.v7.13.0-0.1739885034.p
        version: 7.13.0+0.1739885034.p
      - name: fuse-console.v7.13.0-0.1741875974.p
        version: 7.13.0+0.1741875974.p
      - name: fuse-console.v7.13.0-0.1741884704.p
        version: 7.13.0+0.1741884704.p
      - name: fuse-console.v7.13.0-0.1719822976.p
        version: 7.13.0+0.1719822976.p
      - name: fuse-console.v7.13.0
        version: 7.13.0
      - name: fuse-console.v7.13.0-0.1731460809.p
        version: 7.13.0+0.1731460809.p
      - name: fuse-console.v7.13.0-0.1749640277.p
        version: 7.13.0+0.1749640277.p
      - name: fuse-console.v7.13.0-0.1751291353.p
        version: 7.13.0+0.1751291353.p
      - name: fuse-console.v7.13.0-0.1753702606.p
        version: 7.13.0+0.1753702606.p
      - name: fuse-console.v7.13.0-0.1753778626.p
        version: 7.13.0+0.1753778626.p
      - name: fuse-console.v7.13.0-0.1753785080.p
        version: 7.13.0+0.1753785080.p
      - name: fuse-console.v7.13.0-0.1753860418.p
        version: 7.13.0+0.1753860418.p
      - name: fuse-console.v7.13.0-0.1754303647.p
        version: 7.13.0+0.1754303647.p
      - name: fuse-console.v7.13.0-0.1754559584.p
        version: 7.13.0+0.1754559584.p
      - name: fuse-console.v7.13.0-0.1754567932.p
        version: 7.13.0+0.1754567932.p
      - name: fuse-console.v7.12.1
        version: 7.12.1
      name: 7.13.x
    - currentCSV: fuse-console.v7.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "replicas": 1,
                  "version": "1.9",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "100Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:846d57f3fd9f307a98b0169b6733c26299209d99ed189f219c48f36ba771b198
          createdAt: "2020-09-14T10:59:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.9.0'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.9/html/managing_fuse
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:846d57f3fd9f307a98b0169b6733c26299209d99ed189f219c48f36ba771b198
        version: 7.9.0
      entries:
      - name: fuse-console.v7.9.0
        version: 7.9.0
      name: 7.9.x
    defaultChannel: 7.13.x
    packageName: fuse-console
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: opentelemetry-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opentelemetry-operator.v0.135.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "Instrumentation",
                "metadata": {
                  "name": "instrumentation"
                },
                "spec": {
                  "exporter": {
                    "endpoint": "http://otel-collector-headless:4317"
                  }
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpAMPBridge",
                "metadata": {
                  "name": "opampbridge-sample"
                },
                "spec": {
                  "capabilities": {
                    "AcceptsOpAMPConnectionSettings": true,
                    "AcceptsOtherConnectionSettings": true,
                    "AcceptsRemoteConfig": true,
                    "AcceptsRestartCommand": true,
                    "ReportsEffectiveConfig": true,
                    "ReportsHealth": true,
                    "ReportsOwnLogs": true,
                    "ReportsOwnMetrics": true,
                    "ReportsOwnTraces": true,
                    "ReportsRemoteConfig": true,
                    "ReportsStatus": true
                  },
                  "componentsAllowed": {
                    "exporters": [
                      "debug"
                    ],
                    "processors": [
                      "memory_limiter"
                    ],
                    "receivers": [
                      "otlp"
                    ]
                  },
                  "endpoint": "ws://opamp-server:4320/v1/opamp"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": "receivers:\n  otlp:\n    protocols: \n      grpc: {}\n      http: {}\n\nexporters:\n  debug: {}\n\nservice:\n  pipelines:\n    traces:\n      receivers: [otlp]\n      exporters: [debug]\n"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1beta1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": {
                    "exporters": {
                      "debug": {}
                    },
                    "receivers": {
                      "otlp": {
                        "protocols": {
                          "grpc": {},
                          "http": {}
                        }
                      }
                    },
                    "service": {
                      "pipelines": {
                        "traces": {
                          "exporters": [
                            "debug"
                          ],
                          "receivers": [
                            "otlp"
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/rhosdt/opentelemetry-rhel8-operator@sha256:45c6c52744e9a17ffd883a2864cb7722ce3f9580e522c976d4f55623118f3504
          createdAt: 24 Sep 2025, 10:19
          description: Provides the OpenTelemetry components, including the Collector
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.33.0 <0.135.0-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-opentelemetry-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-telemetry/opentelemetry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instrumentation is the spec for OpenTelemetry instrumentation.
            displayName: OpenTelemetry Instrumentation
            kind: Instrumentation
            name: instrumentations.opentelemetry.io
            version: v1alpha1
          - description: OpAMPBridge is the Schema for the opampbridges API.
            displayName: OpAMP Bridge
            kind: OpAMPBridge
            name: opampbridges.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1beta1
          - description: TargetAllocator is the Schema for the targetallocators API.
            displayName: Target Allocator
            kind: TargetAllocator
            name: targetallocators.opentelemetry.io
            version: v1alpha1
        description: |-
          Red Hat build of OpenTelemetry is a collection of tools, APIs, and SDKs. You use it to instrument, generate, collect, and export telemetry data (metrics, logs, and traces) for analysis in order to understand your software's performance and behavior.
          This operator was previously called Red Hat OpenShift distributed tracing data collection.

          ### Operator features

          * **Sidecar injection** - annotate your pods and let the operator inject a sidecar.
          * **Managed upgrades** - updating the operator will automatically update your OpenTelemetry collectors.
          * **Deployment modes** - your collector can be deployed as sidecar, daemon set, or regular deployment.
          * **Service port management** - the operator detects which ports need to be exposed based on the provided configuration.

          ## Examples

          There is a list of examples to help you deploy different configurations in [this GitHub repository](https://github.com/os-observability/redhat-rhosdt-samples)

          ### Support & Troubleshooting

          Red Hat build of OpenTelemetry is available as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/4.17/html/red_hat_build_of_opentelemetry/index) is recommended when installing, configuring, and managing the Operator and instances.
        displayName: Red Hat build of OpenTelemetry
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opentelemetry
        - tracing
        - logging
        - metrics
        - monitoring
        - troubleshooting
        links:
        - name: OpenTelemetry Operator
          url: https://github.com/open-telemetry/opentelemetry-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/opentelemetry-rhel8-operator@sha256:45c6c52744e9a17ffd883a2864cb7722ce3f9580e522c976d4f55623118f3504
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5a6edfecf1636805e6f0bcad02df10bceec764f0580f3ac8fe62aba978d50006
        - registry.redhat.io/rhosdt/opentelemetry-collector-rhel8@sha256:5f31caf1f4a12cf2832ac3e84877f4fb0c856073820498c8a5be13698348c38a
        - registry.redhat.io/rhosdt/opentelemetry-target-allocator-rhel8@sha256:f77a26edafba06a4fca1a07083e06a908664852cf228b1814455ffa8155cc979
        version: 0.135.0-1
      entries:
      - name: opentelemetry-operator.v0.135.0-1
        version: 0.135.0-1
      - name: opentelemetry-operator.v0.127.0-2
        version: 0.127.0-2
      - name: opentelemetry-operator.v0.127.0-1
        version: 0.127.0-1
      - name: opentelemetry-operator.v0.119.0-2
        version: 0.119.0-2
      - name: opentelemetry-operator.v0.119.0-1
        version: 0.119.0-1
      - name: opentelemetry-operator.v0.113.0-2
        version: 0.113.0-2
      - name: opentelemetry-operator.v0.113.0-1
        version: 0.113.0-1
      - name: opentelemetry-operator.v0.107.0-5
        version: 0.107.0-5
      - name: opentelemetry-operator.v0.107.0-4-0.1727168137.p
        version: 0.107.0-4+0.1727168137.p
      - name: opentelemetry-operator.v0.107.0-4
        version: 0.107.0-4
      - name: opentelemetry-operator.v0.102.0-3
        version: 0.102.0-3
      - name: opentelemetry-operator.v0.102.0-2
        version: 0.102.0-2
      - name: opentelemetry-operator.v0.100.1-3
        version: 0.100.1-3
      - name: opentelemetry-operator.v0.93.0-3
        version: 0.93.0-3
      - name: opentelemetry-operator.v0.93.0-2
        version: 0.93.0-2
      - name: opentelemetry-operator.v0.89.0-3
        version: 0.89.0-3
      - name: opentelemetry-operator.v0.81.1-5
        version: 0.81.1-5
      - name: opentelemetry-operator.v0.74.0-5-0.1687199949.p
        version: 0.74.0-5+0.1687199949.p
      - name: opentelemetry-operator.v0.74.0-5
        version: 0.74.0-5
      - name: opentelemetry-operator.v0.60.0-2
        version: 0.60.0-2
      - name: opentelemetry-operator.v0.56.0-1
        version: 0.56.0-1
      name: stable
    defaultChannel: stable
    packageName: opentelemetry-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: http://marketplace.redhat.com
    name: marketplace-games-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marketplace-games-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "game.marketplace.redhat.com/v1",
                "kind": "Arcade",
                "metadata": {
                  "name": "arcade-sample"
                },
                "spec": {
                  "size": 1
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Arcade is the Schema for the arcades API
            displayName: Arcade Instance
            kind: Arcade
            name: arcades.game.marketplace.redhat.com
            version: v1
        description: RHM Arcade Operator - a simple operator used to showcase how
          to build an operator around an existing web application that can can run
          on Openshift, Vanilla k8s, or on-boarded to Red Hat Marketplace.
        displayName: Arcade Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - rhm
        - red hat
        - marketplace
        - games
        - arcade
        links:
        - name: Marketplace Games Operator
          url: https://github.com/redhat-marketplace/marketplace-games-operator
        maturity: alpha
        provider:
          name: Red Hat Marketplace
          url: http://marketplace.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.connect.redhat.com/rhm-labs/arcade-operator@sha256:d6959be6693728d2dc4cd38b69251551d23c5002cb52db2684df31a0a8f7b0ab
        version: 1.0.0
      entries:
      - name: marketplace-games-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: marketplace-games-operator
    provider:
      name: Red Hat Marketplace
      url: http://marketplace.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-loadcore-agents
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-loadcore-agents.v0.2.25
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2024-11-20T18:10:53Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <=10.0.0'
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Agents
            name: agents.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight LoadCore Agents Operator
        displayName: Keysight LoadCore Agents Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Loadcore Agents
          url: https://keysight-loadcore-agents.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:7efeeb8b29872a6f0271f651d7ae02c91daea16d853c50e374c310f044d8c76c
        - prod.harbor.keysight.digital/loadcore/keysight-loadcore-agents@sha256:96e18b6270097c23bad5445a15613e467d06c6fbdbaf39545edb79915aacc04f
        version: 0.2.25
      entries:
      - name: keysight-loadcore-agents.v0.2.25
        version: 0.2.25
      name: alpha
    defaultChannel: alpha
    packageName: keysight-loadcore-agents
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-secret-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secret-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/tools@sha256:47204b2697dfbab5eaf9d6dcb4c160dd427a148d6b840727239b773810e74fc9
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:a6d6631d35c32b874c995b6af889636ac8dd1f39dcb68810d9f9b10a4e2c1987
        - quay.io/stackable/tools@sha256:47204b2697dfbab5eaf9d6dcb4c160dd427a148d6b840727239b773810e74fc9
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:161e754437428c68490333337303fb30b695a9ac677c3a00a1dc27565b855ab8
        - quay.io/stackable/secret-operator@sha256:1267311830649f928b528512a3cdc843efad6dd32b851d99c84639fbe70e35a3
        version: 24.11.1
      entries:
      - name: secret-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: secret-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:6eb08b99259d62514d87844e1b7470382d500696fecc9a1232f6bb536ed897fa
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:dd730457133f619d8759269abcfa79d7aeb817e01ba7af2d3aa1417df406ea56
        - quay.io/stackable/secret-operator@sha256:6eb08b99259d62514d87844e1b7470382d500696fecc9a1232f6bb536ed897fa
        version: 25.3.0
      entries:
      - name: secret-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: secret-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
          - description: Trust stores managed by the operator.
            displayName: Trust stores managed by the operator.
            kind: TrustStore
            name: truststores.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
        version: 25.7.0
      entries:
      - name: secret-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: secret-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
          - description: Trust stores managed by the operator.
            displayName: Trust stores managed by the operator.
            kind: TrustStore
            name: truststores.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        version: 25.7.0
      entries:
      - name: secret-operator.v25.7.0
        version: 25.7.0
      - name: secret-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-secret-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zettaset
      provider-url: ""
    name: xcrypt-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zts-xcrypt-operator.v0.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xcrypt.zettaset.com/v1alpha1",
                "kind": "XCrypt",
                "metadata": {
                  "name": "zts-masterset"
                },
                "spec": {
                  "status": {
                    "podNames": "zts-ca",
                    "replicas": "1,"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
          createdAt: "2021-06-16T12:29:13.97Z"
          description: XCrypt Operator deploys Zettaset XCrypt Container Encryption  -
            the only software-defined solution that offers transparent high performance
            data-at-rest encryption of critical data in container environments.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/xcrypt-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/xcrypt-operator-rhmp/support?utm_source=openshift_console
          support: Zettaset
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XCrypt is the Schema for the xcrypts API
            kind: XCrypt
            name: xcrypts.xcrypt.zettaset.com
            version: v1alpha1
        description: "Zettaset’s RedHat Certified XCrypt Operator allows users to
          deploy multiple XCrypt components on OpenShift with just a few simple commands.
          The XCrypt Operator deployment provides software-defined encryption that
          transparently protects container data throughout your OpenShift environment.\n\n##
          Core Features\n* **Automated XCrypt provisioning** - Deploying a Zettaset
          XCrypt has never been easier. Adjust cluster specific configuration and
          let the Zettaset Operator take care of provisioning XCrypt services and
          setting up the XCrypt encryption to your exact specification. \n* **CSI
          Driver deployment**: Zettaset XCrypt Operator automatically deploys Zettaset
          XCrypt CSI Driver in order to provide automated and transparent data-at-rest
          \ encryption functionality compatible with Openshift and any other container
          orchestrator. Once the Operator is provisioned, the driver runs fine tuned
          and ready to work with other XCrypt components.\n* **Multi-node topology**:
          Encrypt as many nodes and volumes as you need to. XCrypt components can
          be installed on multiple nodes across the cluster. This is defined by two
          labels: zts-master and zts-worker. Normally zts-master node is a single
          node that runs Zettaset XCrypt major services required for the product operation.
          Multiple nodes can be marked as zts-worker and those are the nodes that
          can have encrypted volumes.\n* **Seamless Encrypted Volume mount**: Volumes
          are not only being encrypted, but also are mounted/unmounted to containers
          with Persistent Volume Claims. This happens as seamlessly as regular volumes
          provision and mount/unmount.\n* **Robust failover**: A node failure doesn’t
          affect XCrypt operations on other nodes of the cluster. Encryption on a
          restarted/recovered node gets back to normal functioning if the node stays
          healthy and data hasn’t been corrupted during the restart\n\n## About Zettaset
          XCrypt Kubernetes Encryption for OpenShift\nXCrypt Kubernetes Encryption
          for OpenShift is a software-only solution that enables encryption of data
          at rest stored in Red Hat OpenShift container environments without slowing
          down the DevSecOps process. In addition to providing a transparent and high-performance
          layer of security that flexibly protects container data across on-premise,
          in hybrid, and multi-cloud environments, it also helps your organization
          efficiently transition from DevOps to DevSecOps.  \n\nMore information about
          Zettaset XCrypt Kubernetes Encryption for OpenShift can be found in the
          [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)\n\n##
          RedHat OpenShift + Zettaset: The next generation of secure enterprise infrastructure
          for DevSecOps  \nBy combining RedHat’s Container Platform with Zettaset’s
          certified software-defined encryption, you can create the next generation
          of secure enterprise infrastructure that DevSecOps requires. Together they
          create the stable, consistent, and supported base that your organization
          needs to develop applications faster, simplify updates, drive innovation,
          and reduce the risk of potential security breaches and data theft.\n###
          Core Features\n* Software-only for simple and scalable deployments \n* Negligible
          performance impact\n* Transparent to developers and administrators \n* Works
          with AWS, Google, and Azure \n* Unique encryption key per container volume
          \n* Encrypted volumes are only available when in use\n* Direct integration
          with OpenShift \n* Automated encryption policy management Secure erase of
          volumes\n* Ability to securely decommission a node\n\n## How does XCrypt
          Kubernetes Encryption fit into OpenShift? \nXCrypt XCrypt Kubernetes Encryption
          for OpenShift Encryption makes it incredibly easy to protect container data
          in your OpenShift environment. \n### Use cases include:\n* Transitioning
          from DevOps to DevSecOps \n* Ensuring data protection in OpenShift container
          environments with negligible performance impact \n* Achieving compliance
          in regulated industries like healthcare and finance\n\n## Other Information\nRole-based
          access control needs to be configured prior to the Operator installation.
          Please find details in the [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)
          \   "
        displayName: Zettaset XCrypt Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - encryption
        - storage
        links:
        - name: Zettaset
          url: https://www.zettaset.com/
        maintainers:
        - email: info@zettaset.com
          name: Zettaset
        maturity: alpha
        provider:
          name: Zettaset
        relatedImages:
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-ca-1@sha256:9f1015dfb459b69f5dbf449efa0517452095a6cf4f20826ebae278f9f800ffe7
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-kmip-server-1@sha256:9af5d21dd748dfbf4e107f27a5f10fa3c0ba928ade9b989a744b620ce34ec757
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-license-server-1@sha256:bc5486892a41e70f5dc4c4e1d82fca8db2149598fb8b5950ba5da1ee2d9d245d
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-host-manager-1@sha256:9faa4b597eb74df412761dffa057b64fb69dbf9fa016414962eb2099a343e734
        - registry.marketplace.redhat.com/rhm/zts/zts-csi-driver-1@sha256:c15fa9a7cc40891103f426206663dcebe1210c8ee3ab6037cbf64e1f86b9c7e5
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
        version: 0.0.17
      entries:
      - name: zts-xcrypt-operator.v0.0.17
        version: 0.0.17
      name: alpha
    defaultChannel: alpha
    packageName: xcrypt-operator-rhmp
    provider:
      name: Zettaset
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://www.medik8s.io/
    name: customized-user-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: customized-user-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediation",
                "metadata": {
                  "name": "customizeduserremediation-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediationConfig",
                "metadata": {
                  "name": "customizeduserremediationconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediationTemplate",
                "metadata": {
                  "name": "customizeduserremediationtemplate-sample",
                  "namespace": "openshift-workload-availability"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/customized-user-remediation-operator:v0.1.0
          createdAt: "2024-03-25 16:17:57"
          description: Customized User Remediation Operator will remediate unhealthy
            nodes by running a pre defined user script on the unhealthy node. Since
            the user has full control over the script content, is may be quite risky
            so apply discretion.
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/customized-user-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CustomizedUserRemediationConfig is the Schema for the customizeduserremediationconfigs
              API
            displayName: Customized User Remediation Config
            kind: CustomizedUserRemediationConfig
            name: customizeduserremediationconfigs.customized-user-remediation.medik8s.io
            version: v1alpha1
          - description: CustomizedUserRemediation is the Schema for the customizeduserremediations
              API
            displayName: Customized User Remediation
            kind: CustomizedUserRemediation
            name: customizeduserremediations.customized-user-remediation.medik8s.io
            version: v1alpha1
          - description: CustomizedUserRemediationTemplate is the Schema for the customizeduserremediationtemplates
              API
            displayName: Customized User Remediation Template
            kind: CustomizedUserRemediationTemplate
            name: customizeduserremediationtemplates.customized-user-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Customized User Remediation (CUR) works with Node Health Check to remediate unhealthy nodes by running a user customized script on the unhealthy node.
          This operator does not preform any fencing on the remediated node so users should be careful with the scripts that are being used.

          ### Compatibility
          CUR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Flexibility - CUR is quite flexible sine it runs in the remediation process a user customized script.
        displayName: Customized User Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - script
        - CUR
        links:
        - name: Customized User Remediation
          url: https://customized-user-remediation.domain
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Medik8s
          url: https://www.medik8s.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/medik8s/customized-user-remediation-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: customized-user-remediation.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: customized-user-remediation
    provider:
      name: Medik8s
      url: https://www.medik8s.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-acm-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-acm-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acm.services.k8s.aws/v1alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/acm-controller:1.1.1
          createdAt: "2025-09-20T01:19:56Z"
          description: AWS ACM controller is a service controller for managing ACM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Certificate represents the state of an AWS acm Certificate
              resource.
            displayName: Certificate
            kind: Certificate
            name: certificates.acm.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage AWS Certificate Manager (ACM) resources in AWS from within your Kubernetes cluster.

          **About Amazon ACM**

          AWS Certificate Manager (ACM) handles the complexity of creating, storing, and renewing public and private SSL/TLS X.509 certificates and keys that protect your AWS websites and applications. You can provide certificates for your [integrated AWS services](https://docs.aws.amazon.com/acm/latest/userguide/acm-services.html) either by issuing them directly with ACM or by [importing](https://docs.aws.amazon.com/acm/latest/userguide/import-certificate.html) third-party certificates into the ACM management system. ACM certificates can secure singular domain names, multiple specific domain names, wildcard domains, or combinations of these. ACM wildcard certificates can protect an unlimited number of subdomains. You can also [export](https://docs.aws.amazon.com/acm/latest/userguide/export-private.html) ACM certificates signed by AWS Private CA for use anywhere in your internal PKI.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon ACM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ACM Developer Resources
          url: https://aws.amazon.com/certificate-manager/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: acm maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/acm-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-acm-controller.v1.1.1
        version: 1.1.1
      - name: ack-acm-controller.v1.1.0
        version: 1.1.0
      - name: ack-acm-controller.v1.0.13
        version: 1.0.13
      - name: ack-acm-controller.v1.0.12
        version: 1.0.12
      - name: ack-acm-controller.v1.0.11
        version: 1.0.11
      - name: ack-acm-controller.v1.0.10
        version: 1.0.10
      - name: ack-acm-controller.v1.0.9
        version: 1.0.9
      - name: ack-acm-controller.v1.0.8
        version: 1.0.8
      - name: ack-acm-controller.v1.0.7
        version: 1.0.7
      - name: ack-acm-controller.v1.0.6
        version: 1.0.6
      - name: ack-acm-controller.v1.0.5
        version: 1.0.5
      - name: ack-acm-controller.v1.0.4
        version: 1.0.4
      - name: ack-acm-controller.v1.0.1
        version: 1.0.1
      - name: ack-acm-controller.v1.0.0
        version: 1.0.0
      - name: ack-acm-controller.v0.0.20
        version: 0.0.20
      - name: ack-acm-controller.v0.0.19
        version: 0.0.19
      - name: ack-acm-controller.v0.0.18
        version: 0.0.18
      - name: ack-acm-controller.v0.0.17
        version: 0.0.17
      - name: ack-acm-controller.v0.0.16
        version: 0.0.16
      - name: ack-acm-controller.v0.0.12
        version: 0.0.12
      - name: ack-acm-controller.v0.0.10
        version: 0.0.10
      - name: ack-acm-controller.v0.0.9
        version: 0.0.9
      - name: ack-acm-controller.v0.0.6
        version: 0.0.6
      - name: ack-acm-controller.v0.0.5
        version: 0.0.5
      - name: ack-acm-controller.v0.0.4
        version: 0.0.4
      - name: ack-acm-controller.v0.0.2
        version: 0.0.2
      - name: ack-acm-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-acm-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ricardo Maraschini
      provider-url: https://github.com/ricardomaraschini/tagger
    name: tagger
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tagger.v2.1.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tagger.dev/v1beta1",
                "kind": "Tag",
                "metadata": {
                  "name": "nginx"
                },
                "spec": {
                  "from": "docker.io/library/nginx:latest",
                  "generation": 0,
                  "mirror": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          containerImage: quay.io/tagger/operator:v2.1.17
          createdAt: "2021-08-02T00:00:00Z"
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/ricardomaraschini/tagger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tag is a Container Image representation as a Custom Resource.
            displayName: Tag
            kind: Tag
            name: tags.tagger.dev
            version: v1beta1
        description: |-
          A Container Image Orchestrator. Adds mirroring and Container Image manipulation capabilities
          to a Kubernetes cluster.

          Keeping track of all Container Images in use in a Kubernetes cluster is a complicated task.
          Container Images may come from numerous different Image Registries. In some cases, controlling
          how a stable version of a given Container Image looks escapes the user's authority. To add to
          this, Container Runtimes rely on remote registries (from the cluster's point of view) when
          obtaining Container Images, potentially making the process of pulling their blobs (manifests,
          config, and layers) slower.

          The notion of indexing Container Image versions by Tags is helpful. Still, it does not provide
          users with the right confidence to always use the intended Container Image – today's "latest"
          tag might not be tomorrow's "latest" tag. In addition to that, these Image Registries allow
          access to Container Images by their Manifest content's hash (i.e., usually sha256), which gives
          users the confidence at a cost in semantics.

          When releasing a new version of an application to Push and to Deploy are split into two
          distinct steps. Both the pusher and the puller need access to the same Image Registry, adding
          complexity.  Credentials are one example of the concerns. Other factors may pop up when
          running, for instance, in an air-gapped environment, where the cluster may not reach external
          Image Registries.

          Tagger aims to overcome these caveats by providing an image management abstraction layer. For
          instance, by providing a direct mapping between a Container Image tag (e.g., "latest") and its
          correspondent Manifest content's hash, users can then refer to the Container Image by its tag
          – and be sure to use that specific version. More than that, when allied with an Internal Image
          Registry, Tagger can also automatically mirror Container Images into the cluster.

          While using Tagger, Deployments can refer to Container Images by an arbitrarily defined name,
          such as "my-app", and Tagger will make sure that they use the right Container Image through its
          internal "tag to Manifest content's hash" mapping.

          For each new "release" of a given Container Image, Tagger creates a new Generation for it,
          making it easy to roll back to previously pushed "releases" of the same Container Image in case
          of problems.

          When integrated with an Internal Registry, Tagger allows users to push or pull Images directly
          without requiring an external Image Registry. It works as a layer between the user and the
          Internal Registry. Every time a new "release" of Container Images is pushed, all Deployments
          are updated automatically. Users don't need to know about the Internal Registry existence, if
          they are logged-in to the Kubernetes cluster, they can obtain old or provide new "Generations"
          for a Container Image.

          In summary, Tagger mirrors remote Container Images into a Kubernetes cluster indexing them in
          different Generations (allowing easy navigation through these multiple Generations), provides
          an interface allowing users to pull and push images directly to the Kubernetes cluster and
          provides full integration with Kubernetes Deployments (automatic triggers new rollouts on
          Container Image changes).

          By default Tagger won't be able to mirror until you provide it with a mirror registry config.
          After install you can configure the mirror by editing the Secret mirror-registry-config in the
          operator namespace. Follow below an example of a valid mirror-registry-config secret (you gonna
          have to provide your own address, username, password, etc).

          ```yaml
          apiVersion: v1
          kind: Secret
          metadata:
            name: mirror-registry-config
          data:
            address: cmVnaXN0cnkuaW8=
            username: YWRtaW4=
            password: d2hhdCB3ZXJlIHlvdSB0aGlua2luZz8K
            token: YW4gb3B0aW9uYWwgdG9rZW4gZm9yIHRva2VuIGJhc2VkIGF1dGg=
            insecure: dHJ1ZQ==
          ```

          To make pushing an pulling images easier you should go ahead and grab the `kubectl` plugin
          available [here](https://github.com/ricardomaraschini/tagger/releases). You can download a
          full version of the README with usage examples and configuration options
          [here](https://github.com/ricardomaraschini/tagger/releases/download/v2.1.17/README.pdf).
        displayName: Tagger
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tagger
        - mirroring
        - image
        links:
        - name: Documentation
          url: https://github.com/ricardomaraschini/tagger/releases/download/v2.1.17/README.pdf
        - name: Plugin (kubectl)
          url: https://github.com/ricardomaraschini/tagger/releases/tag/v2.1.17
        maintainers:
        - email: rmarasch@redhat.com
          name: Ricardo Maraschini
        maturity: alpha
        provider:
          name: Ricardo Maraschini
          url: https://github.com/ricardomaraschini/tagger
        relatedImages:
        - quay.io/tagger/operator:v2.1.17
        version: 2.1.17
      entries:
      - name: tagger.v2.1.17
        version: 2.1.17
      name: alpha
    defaultChannel: alpha
    packageName: tagger
    provider:
      name: Ricardo Maraschini
      url: https://github.com/ricardomaraschini/tagger
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecs-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "Service",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "TaskDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecs-controller:1.1.1
          createdAt: "2025-09-19T21:02:44Z"
          description: AWS ECS controller is a service controller for managing ECS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster represents the state of an AWS ecs Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.ecs.services.k8s.aws
            version: v1alpha1
          - description: Service represents the state of an AWS ecs Service resource.
            displayName: Service
            kind: Service
            name: services.ecs.services.k8s.aws
            version: v1alpha1
          - description: TaskDefinition represents the state of an AWS ecs TaskDefinition
              resource.
            displayName: TaskDefinition
            kind: TaskDefinition
            name: taskdefinitions.ecs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Container Service (ECS) resources in AWS from within your Kubernetes cluster.

          **About Amazon ECS**

          Amazon Elastic Container Service (Amazon ECS) is a fully managed container orchestration service that helps you easily deploy, manage, and scale containerized applications. As a fully managed service, Amazon ECS comes with AWS configuration and operational best practices built-in. It's integrated with both AWS and third-party tools, such as Amazon Elastic Container Registry and Docker. This integration makes it easier for teams to focus on building the applications, not the environment. You can run and scale your container workloads across AWS Regions in the cloud, and on-premises, without the complexity of managing a control plane.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECS Developer Resources
          url: https://aws.amazon.com/ecs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecs-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ecs-controller.v1.1.1
        version: 1.1.1
      - name: ack-ecs-controller.v1.0.15
        version: 1.0.15
      - name: ack-ecs-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecs-controller.v1.0.13
        version: 1.0.13
      - name: ack-ecs-controller.v1.0.12
        version: 1.0.12
      - name: ack-ecs-controller.v1.0.11
        version: 1.0.11
      - name: ack-ecs-controller.v1.0.10
        version: 1.0.10
      - name: ack-ecs-controller.v1.0.9
        version: 1.0.9
      - name: ack-ecs-controller.v1.0.8
        version: 1.0.8
      - name: ack-ecs-controller.v1.0.7
        version: 1.0.7
      - name: ack-ecs-controller.v1.0.6
        version: 1.0.6
      - name: ack-ecs-controller.v1.0.2
        version: 1.0.2
      - name: ack-ecs-controller.v1.0.1
        version: 1.0.1
      - name: ack-ecs-controller.v1.0.0
        version: 1.0.0
      - name: ack-ecs-controller.v0.0.8
        version: 0.0.8
      - name: ack-ecs-controller.v0.0.7
        version: 0.0.7
      - name: ack-ecs-controller.v0.0.6
        version: 0.0.6
      - name: ack-ecs-controller.v0.0.5
        version: 0.0.5
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseWeb
      provider-url: ""
    name: cnfv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cnfv-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eweb.enterpriseweb.com/v1",
                "kind": "CloudNFV",
                "metadata": {
                  "name": "cloudnfv-sample"
                },
                "spec": {
                  "size": 1
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CloudNFV is the Schema for the cloudnfvs API
            displayName: Cloud NFV
            kind: CloudNFV
            name: cloudnfvs.eweb.enterpriseweb.com
            version: v1
        description: Manages a cluster of CloudNFV Runtime Modules
        displayName: CloudNFV
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - automation
        - NFV
        - controller
        links:
        - name: Cnfv Operator
          url: https://cnfv-operator.domain
        maintainers:
        - email: bill@enterpriseweb.com
          name: Bill Malyk
        maturity: alpha
        provider:
          name: EnterpriseWeb
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61a164ad711478ac603242338f7b3c55986319479878b092be128581a823598e
        - quay.io/wmalyk28/cnfv-op@sha256:e5913ad5dd20cb1b2299d2387237ef15e645c9a8c9d4f5fe0044b86a6b3dc1df
        - quay.io/wmalyk28/cloudnfv-runtime@sha256:fcde8375576c9bff602d3a16ef1ba8ae88b780166649ba6ae355737b1d35a03f
        version: 0.0.1
      entries:
      - name: cnfv-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cnfv-operator
    provider:
      name: EnterpriseWeb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzounstructured-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzounstructured-operator.v3.1.200
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzounstructured.clusters.cambridgesemantics.com/v1",
                "kind": "AnzoUnstructured",
                "metadata": {
                  "name": "au01"
                },
                "spec": {
                  "auWorker": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-w",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-worker@sha256:964633d6b16ddc517a09ac86f3e3698f5a161a72cfa50ea9ba4e301fa631e347",
                                "name": "w",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  },
                  "msLeader": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-ms",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzo-microservices-leader@sha256:0637603c00a99533d57e16009516ec73d40703ae151e631fc281f51f1f0ca93b",
                                "name": "ms",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/unstructured-operator@sha256:633a8992e35601d99c0d690dd71af314605c223b57c2038fa7090be06977b2f9
          createdAt: "2025-07-08T09:47:55Z"
          description: Kubernetes Operator for Altair Graph Studio Unstructured
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoUnstructured is the Schema for the anzounstructureds
              API
            displayName: Altair Graph Studio Unstructured Operator
            kind: AnzoUnstructured
            name: anzounstructureds.anzounstructured.clusters.cambridgesemantics.com
            version: v1
        description: |-
          The Altair Graph Studio Unstructured Operator provides the way to install and configure an anzo unstructured setup on Red Hat K8S environment.
          Currently, this is possible only through existing Anzo installation.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v3.1.2/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Studio Unstructured Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Database
        - Big Data
        links:
        - name: Altair Graph Studio Unstructured Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v3.1.2/README.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/unstructured-operator@sha256:633a8992e35601d99c0d690dd71af314605c223b57c2038fa7090be06977b2f9
        - registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-worker@sha256:964633d6b16ddc517a09ac86f3e3698f5a161a72cfa50ea9ba4e301fa631e347
        - registry.connect.redhat.com/cambridgesemantics/anzo-microservices-leader@sha256:0637603c00a99533d57e16009516ec73d40703ae151e631fc281f51f1f0ca93b
        version: 3.1.200
      entries:
      - name: anzounstructured-operator.v3.1.200
        version: 3.1.200
      - name: anzounstructured-operator.v3.1.104
        version: 3.1.104
      - name: anzounstructured-operator.v3.1.103
        version: 3.1.103
      name: stable
    defaultChannel: stable
    packageName: anzounstructured-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: openshift-secondary-scheduler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secondaryscheduleroperator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "SecondaryScheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-secondary-scheduler-operator"
                },
                "spec": {
                  "managementState": "Managed",
                  "schedulerConfig": "secondary-scheduler-config"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-rhel9-operator@sha256:210cc51760c8d3a9e1e5ba2b41f575b667e0d541b995362426e5a28f3f666600
          createdAt: 2024/10/01
          description: Runs a secondary scheduler in an OpenShift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.2.0 <1.3.1'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/secondary-scheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SecondaryScheduler is the schema for the secondaryschedulers
              API
            displayName: Secondary Scheduler
            kind: SecondaryScheduler
            name: secondaryschedulers.operator.openshift.io
            version: v1
        description: |
          The Secondary Scheduler Operator provides the ability to use a customized scheduler image that was developed using the scheduler plugin framework as a secondary scheduler in OpenShift.

          ## Secondary Scheduler Configuration

          Once the Secondary Scheduler Operator is installed, you can configure the plugins to run for your secondary scheduler.

          The configuration of the secondary scheduler is defined through a config map that wraps the KubeSchedulerConfiguration YAML file in "config.yaml".

          ## Additional Parameters

          Additionally, the following parameters can be configured:

          * `schedulerConfig` - Set the config map configuration for the secondary scheduler.
          * `schedulerImage` - Set the default operand image for the secondary scheduler.
        displayName: Secondary Scheduler Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - secondary
        - scheduling
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/secondary-scheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-rhel9-operator@sha256:210cc51760c8d3a9e1e5ba2b41f575b667e0d541b995362426e5a28f3f666600
        version: 1.3.1
      entries:
      - name: secondaryscheduleroperator.v1.3.1
        version: 1.3.1
      - name: secondaryscheduleroperator.v1.3.0
        version: 1.3.0
      - name: secondaryscheduleroperator.v1.2.2
        version: 1.2.2
      - name: secondaryscheduleroperator.v1.2.1
        version: 1.2.1
      - name: secondaryscheduleroperator.v1.2.0
        version: 1.2.0
      - name: secondaryscheduleroperator.v1.1.4
        version: 1.1.4
      - name: secondaryscheduleroperator.v1.1.2
        version: 1.1.2
      name: stable
    defaultChannel: stable
    packageName: openshift-secondary-scheduler-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: costmanagement-metrics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: costmanagement-metrics-operator.4.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "true"
          containerImage: registry.redhat.io/costmanagement/costmanagement-metrics-rhel9-operator@sha256:c5bb5291052d57390b9f4bdc660c1926827072b6dcbe09d85f5b6fc375562182
          createdAt: "2025-08-18T12:10:15Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: costmanagement-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Cost Management Metrics Operator
          ## Introduction
          The `costmanagement-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Cost Management Metrics Operator (`costmanagement-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created CostManagementMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The CostManagementMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **FIPS Compliance:** Supports deployment in high-security environments when run on an OpenShift cluster with FIPS mode enabled.

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created CostManagementMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the CostManagementMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `CostManagementMetricsConfig`, creating a new `CostManagementMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `costmanagement-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: costmanagement-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the CostManagementMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the CostManagementMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon CostManagementMetricsConfig creation** (This parameter will start to appear in CostManagementMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the CostManagementMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the costmanagement-metrics-operator
          **Note** There are separate instructions for configuring the `costmanagement-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `costmanagement-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the CostManagementMetricsConfig
          Configure the costmanagement-metrics-operator by creating a `CostManagementMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `costmanagement-metrics-operator` -> `CostManagementMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the costmanagement-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `costmanagement-metrics-operator-data` with 10Gi of storage. To configure the costmanagement-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `costmanagement-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `costmanagement-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the costmanagement-metrics-operator for a restricted network
          ##### Create the CostManagementMetricsConfig
          Configure the costmanagement-metrics-operator by creating a `CostManagementMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `costmanagement-metrics-operator` -> `CostManagementMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `costmanagement-metrics-operator-data` with 10Gi of storage. To configure the costmanagement-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `costmanagement-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: costmanagement-metrics-operator
                labels:
                  app: costmanagement-metrics-operator
              spec:
                volumes:
                - name: costmanagement-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: costmanagement-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: costmanagement-metrics-operator-reports
                    mountPath: /tmp/costmanagement-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/costmanagement-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/costmanagement-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `costmanagement-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the CostManagementMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Cost Management Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Cost Management Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/costmanagement/costmanagement-metrics-rhel9-operator@sha256:c5bb5291052d57390b9f4bdc660c1926827072b6dcbe09d85f5b6fc375562182
        version: 4.1.0
      entries:
      - name: costmanagement-metrics-operator.4.1.0
        version: 4.1.0
      - name: costmanagement-metrics-operator.4.0.0
        version: 4.0.0
      - name: costmanagement-metrics-operator.3.3.2
        version: 3.3.2
      - name: costmanagement-metrics-operator.3.3.1
        version: 3.3.1
      - name: costmanagement-metrics-operator.3.3.0
        version: 3.3.0
      - name: costmanagement-metrics-operator.3.2.1
        version: 3.2.1
      - name: costmanagement-metrics-operator.3.2.0
        version: 3.2.0
      - name: costmanagement-metrics-operator.3.1.1
        version: 3.1.1
      - name: costmanagement-metrics-operator.3.1.0
        version: 3.1.0
      - name: costmanagement-metrics-operator.3.0.1
        version: 3.0.1
      - name: costmanagement-metrics-operator.3.0.0
        version: 3.0.0
      - name: costmanagement-metrics-operator.2.0.0
        version: 2.0.0
      - name: costmanagement-metrics-operator.1.2.0
        version: 1.2.0
      - name: costmanagement-metrics-operator.1.1.9
        version: 1.1.9
      - name: costmanagement-metrics-operator.1.1.8
        version: 1.1.8
      - name: costmanagement-metrics-operator.1.1.7
        version: 1.1.7
      - name: costmanagement-metrics-operator.1.1.6
        version: 1.1.6
      - name: costmanagement-metrics-operator.1.1.5
        version: 1.1.5
      - name: costmanagement-metrics-operator.1.1.4
        version: 1.1.4
      - name: costmanagement-metrics-operator.1.1.3
        version: 1.1.3
      - name: costmanagement-metrics-operator.1.1.0
        version: 1.1.0
      - name: costmanagement-metrics-operator.1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: costmanagement-metrics-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-applicationautoscaling-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-applicationautoscaling-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "applicationautoscaling.services.k8s.aws/v1alpha1",
                "kind": "ScalableTarget",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "applicationautoscaling.services.k8s.aws/v1alpha1",
                "kind": "ScalingPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/applicationautoscaling-controller:1.1.1
          createdAt: "2025-09-20T21:43:01Z"
          description: AWS Application Auto Scaling controller is a service controller
            for managing Application Auto Scaling resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ScalableTarget represents the state of an AWS applicationautoscaling
              ScalableTarget resource.
            displayName: ScalableTarget
            kind: ScalableTarget
            name: scalabletargets.applicationautoscaling.services.k8s.aws
            version: v1alpha1
          - description: ScalingPolicy represents the state of an AWS applicationautoscaling
              ScalingPolicy resource.
            displayName: ScalingPolicy
            kind: ScalingPolicy
            name: scalingpolicies.applicationautoscaling.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Application Auto Scaling resources in AWS from within your Kubernetes cluster.
          **About Amazon Application Auto Scaling**
          Application Auto Scaling is a web service for developers and system administrators who need a solution for automatically scaling their scalable resources for individual AWS services beyond Amazon EC2. Application Auto Scaling allows you to configure automatic scaling for the following resources:
          - AppStream 2.0 fleets
          - Aurora replicas
          - Amazon Comprehend document classification and entity recognizer endpoints
          - DynamoDB tables and global secondary indexes
          - Amazon Elastic Container Service (ECS) services
          - Amazon EMR clusters
          - Amazon Keyspaces (for Apache Cassandra) tables
          - Lambda function provisioned concurrency
          - Amazon Managed Streaming for Apache Kafka (MSK) broker storage
          - SageMaker endpoint variants
          - Spot Fleet requests
          - Custom resources provided by your own applications or services. For more information, see the [GitHub repository](https://github.com/aws/aws-auto-scaling-custom-resource)

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Application Auto Scaling
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - applicationautoscaling
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Application Auto Scaling User Guide
          url: https://docs.aws.amazon.com/autoscaling/application/userguide/what-is-application-auto-scaling.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: application auto scaling maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/applicationautoscaling-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-applicationautoscaling-controller.v1.1.1
        version: 1.1.1
      - name: ack-applicationautoscaling-controller.v1.1.0
        version: 1.1.0
      - name: ack-applicationautoscaling-controller.v1.0.30
        version: 1.0.30
      - name: ack-applicationautoscaling-controller.v1.0.29
        version: 1.0.29
      - name: ack-applicationautoscaling-controller.v1.0.27
        version: 1.0.27
      - name: ack-applicationautoscaling-controller.v1.0.26
        version: 1.0.26
      - name: ack-applicationautoscaling-controller.v1.0.25
        version: 1.0.25
      - name: ack-applicationautoscaling-controller.v1.0.24
        version: 1.0.24
      - name: ack-applicationautoscaling-controller.v1.0.23
        version: 1.0.23
      - name: ack-applicationautoscaling-controller.v1.0.22
        version: 1.0.22
      - name: ack-applicationautoscaling-controller.v1.0.21
        version: 1.0.21
      - name: ack-applicationautoscaling-controller.v1.0.19
        version: 1.0.19
      - name: ack-applicationautoscaling-controller.v1.0.18
        version: 1.0.18
      - name: ack-applicationautoscaling-controller.v1.0.17
        version: 1.0.17
      - name: ack-applicationautoscaling-controller.v1.0.16
        version: 1.0.16
      - name: ack-applicationautoscaling-controller.v1.0.15
        version: 1.0.15
      - name: ack-applicationautoscaling-controller.v1.0.11
        version: 1.0.11
      - name: ack-applicationautoscaling-controller.v1.0.9
        version: 1.0.9
      - name: ack-applicationautoscaling-controller.v1.0.8
        version: 1.0.8
      - name: ack-applicationautoscaling-controller.v1.0.7
        version: 1.0.7
      - name: ack-applicationautoscaling-controller.v1.0.6
        version: 1.0.6
      - name: ack-applicationautoscaling-controller.v1.0.5
        version: 1.0.5
      - name: ack-applicationautoscaling-controller.v1.0.4
        version: 1.0.4
      - name: ack-applicationautoscaling-controller.v1.0.3
        version: 1.0.3
      - name: ack-applicationautoscaling-controller.v1.0.2
        version: 1.0.2
      - name: ack-applicationautoscaling-controller.v1.0.0
        version: 1.0.0
      - name: ack-applicationautoscaling-controller.v0.2.16
        version: 0.2.16
      - name: ack-applicationautoscaling-controller.v0.2.15
        version: 0.2.15
      - name: ack-applicationautoscaling-controller.v0.2.14
        version: 0.2.14
      - name: ack-applicationautoscaling-controller.v0.2.13
        version: 0.2.13
      - name: ack-applicationautoscaling-controller.v0.2.12
        version: 0.2.12
      - name: ack-applicationautoscaling-controller.v0.2.10
        version: 0.2.10
      - name: ack-applicationautoscaling-controller.v0.2.9
        version: 0.2.9
      - name: ack-applicationautoscaling-controller.v0.2.8
        version: 0.2.8
      - name: ack-applicationautoscaling-controller.v0.2.7
        version: 0.2.7
      - name: ack-applicationautoscaling-controller.v0.2.6
        version: 0.2.6
      - name: ack-applicationautoscaling-controller.v0.2.5
        version: 0.2.5
      - name: ack-applicationautoscaling-controller.v0.2.4
        version: 0.2.4
      - name: ack-applicationautoscaling-controller.v0.2.3
        version: 0.2.3
      - name: ack-applicationautoscaling-controller.v0.2.2
        version: 0.2.2
      - name: ack-applicationautoscaling-controller.v0.2.1
        version: 0.2.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-applicationautoscaling-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-api-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-api-controller.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "example-registry-with-studio"
                },
                "spec": {
                  "app": {
                    "host": "example-registry-with-studio-app.apps.yourcluster.example"
                  },
                  "studioUi": {
                    "enabled": true,
                    "host": "example-registry-with-studio-studio-ui.apps.yourcluster.example"
                  },
                  "ui": {
                    "host": "example-registry-with-studio-ui.apps.yourcluster.example"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-operator:3.0.5
          createdAt: "2024-12-03T11:30:52Z"
          description: Deploy and manage Apicurio API Controller on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Apicurio API
              Controller.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Apicurio API Controller

          Apicurio API Controller consists of 2 components:
           - Apicurio Registry 3
           - Apicurio Studio

          Storage options: PostgreSQL, Kafka, MySQL, SQL Server, in-memory.

          ## Apicurio API Controller Operator

          Provides a quick and easy way to deploy and manage Apicurio API Controller on Kubernetes.

          **Features**
          - Supports installation and configuration of Apicurio API Controller
          - Perform a rolling upgrade of Apicurio API Controller

          ## Prerequisites

          This operator does not deploy storage for Apicurio API Controller. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio API Controller and Apicurio API Controller Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Apicurio API Controller
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - API Management
        - Apicurio Registry
        - Apicurio Studio
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-operator:3.0.5
        version: 0.0.1
      entries:
      - name: apicurio-api-controller.v0.0.1
        version: 0.0.1
      name: 0.x
    defaultChannel: 0.x
    packageName: apicurio-api-controller
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: unsupported
      operatorframework.io/arch.s390x: unsupported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: https://www.konveyor.io
    name: konveyor-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: konveyor-operator.v0.8.0-alpha.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:c1d1942b3bb3c4e99d9b5484493e27a432455608fee4548f1d352be34a8d6ded
          createdAt: "2025-08-13T23:43:34Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0-alpha.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-hub@sha256:6be8fe8307e6686b178ace42f6db5cab8f04c5e2e29e02d2edcb5dc28ee337e9
        - quay.io/konveyor/tackle2-addon-platform@sha256:12efa5b7fbf6d829879b19e2820dcf782e648059f51544ebd25251dfc67fc7ca
        - quay.io/konveyor/postgresql-15-c9s@sha256:d75c5a1a0fa686ccf701e5e55633aa73d0cbb2cd9b5d8d4d4e80d7c46d78c121
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-operator@sha256:c1d1942b3bb3c4e99d9b5484493e27a432455608fee4548f1d352be34a8d6ded
        - quay.io/konveyor/tackle2-addon-discovery@sha256:ba865a74287ff94375ed316f348ee2c1d962d8d350c8afae5dbb0aee9b64fdea
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:7a5750fb76935e55b7b59e8363f21f5f9b48b39492d7f3e58e2e6a84d07ab027
        - quay.io/konveyor/kai-solution-server@sha256:83987d12d8290e085b6b66de56d347a8bf262815a5f99796177cb9120513ee3a
        - quay.io/konveyor/java-external-provider@sha256:42a156f1d81352f5e09fa0ffe2b51a4d93a1f476755eb0bac9a1d24bd3a89631
        - quay.io/konveyor/kantra@sha256:06ae932d9c2f51035bc5a488dd61e8a31c344e72615b013419199168701ea3fb
        - quay.io/konveyor/tackle2-ui@sha256:867bf24dd1687a79f2a911c22c1059ca7bd025a307d290290265653a696b8e98
        - quay.io/konveyor/tackle-keycloak-init@sha256:07c2ca07922d273bd2a3d9156ad8af0e5bf973aeb316496ee65d7201c381c1b1
        - quay.io/konveyor/origin-oauth-proxy@sha256:2c811da843057119b246784f21dc7eec490669a8569e8295d8e914cdce555b24
        - quay.io/konveyor/generic-external-provider@sha256:d58e867b51d5e76cd3597cfb7f2478d221b966f84509dbdd44c609ff5d9169be
        version: 0.8.0-alpha.4
      entries:
      - name: konveyor-operator.v0.8.0-alpha.4
        version: 0.8.0-alpha.4
      - name: konveyor-operator.v0.8.0-alpha.2
        version: 0.8.0-alpha.2
      - name: konveyor-operator.v0.8.0-alpha.1
        version: 0.8.0-alpha.1
      - name: konveyor-operator.v0.7.0-alpha.2
        version: 0.7.0-alpha.2
      - name: konveyor-operator.v0.7.0-alpha.1
        version: 0.7.0-alpha.1
      - name: konveyor-operator.v0.6.0-alpha.2
        version: 0.6.0-alpha.2
      - name: konveyor-operator.v0.6.0-alpha.1
        version: 0.6.0-alpha.1
      - name: konveyor-operator.v0.5.0-alpha.1
        version: 0.5.0-alpha.1
      - name: konveyor-operator.v0.4.0-alpha.1
        version: 0.4.0-alpha.1
      - name: konveyor-operator.v0.3.0-alpha.6
        version: 0.3.0-alpha.6
      - name: konveyor-operator.v0.3.0-alpha.5
        version: 0.3.0-alpha.5
      - name: konveyor-operator.v0.3.0-alpha.4
        version: 0.3.0-alpha.4
      - name: konveyor-operator.v0.3.0-alpha.3
        version: 0.3.0-alpha.3
      - name: konveyor-operator.v0.3.0-alpha.2
        version: 0.3.0-alpha.2
      - name: konveyor-operator.v0.3.0-alpha.1
        version: 0.3.0-alpha.1
      - name: konveyor-operator.v0.2.0-alpha.3
        version: 0.2.0-alpha.3
      - name: konveyor-operator.v0.2.0-alpha.2
        version: 0.2.0-alpha.2
      - name: konveyor-operator.v0.2.0-alpha.1
        version: 0.2.0-alpha.1
      name: alpha
    - currentCSV: konveyor-operator.v0.8.0-beta.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:019a79dec40445b6289948e173e5bf2ebfa9b495a7bb077abd634c0667953a54
          createdAt: "2025-09-17T19:50:24Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0-beta.5'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/kai-solution-server@sha256:1e69cf8a342090a03a9dddb1b28ae3627052b6d636ffa290a11761b94fd10648
        - quay.io/konveyor/tackle2-ui@sha256:78398832e092557639531230cf6d023bb07b589f3587afeaacf8c68640abc454
        - quay.io/konveyor/tackle2-operator@sha256:019a79dec40445b6289948e173e5bf2ebfa9b495a7bb077abd634c0667953a54
        - quay.io/konveyor/tackle-keycloak-init@sha256:0b4bfc1fea756f10a0657861c4c422303a2349d8d9b04a23238a32ceaf106ac5
        - quay.io/konveyor/origin-oauth-proxy@sha256:8988f14db7dde3edcaed0541cecb0b4fd58fab9dbaffb79a6c1c4aee13b9bd73
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/kantra@sha256:ad6f5c05e5e950e33a745accd03ad933fb7bb4907f44455f88a2b5c27e6dd15e
        - quay.io/konveyor/tackle2-addon-platform@sha256:d0280ade44ba5e15210c07342608576b38e6993af2e479740675a4706cd2c875
        - quay.io/konveyor/tackle2-addon-discovery@sha256:a29a0248e222d9ac3225e9d3df603f1383fa8844856bf68ea618d303f0b79883
        - quay.io/konveyor/java-external-provider@sha256:0cbdd91cc5057889b8daa2f7611fa0c944fab951c73bdf88a3e28795e9a5fc33
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:fed49c5ff14bb232e63d2dc180d3b530da63f11bf97055fc9d86cbb9e47104a4
        - quay.io/konveyor/generic-external-provider@sha256:a6c550bad7a09a1c51e226247f11c00d8453987698832cb566582842219f1f3c
        - quay.io/konveyor/tackle2-hub@sha256:1611913558c9f1203a844b43969d7b1539dab3578b83460e8383ee6757c3c34c
        - quay.io/konveyor/postgresql-15-c9s@sha256:157a2941ae971d8f72b2f8af1d4a29deeb5548012a978fa986c1c6e52395da4d
        version: 0.8.0-beta.5
      entries:
      - name: konveyor-operator.v0.8.0-beta.5
        version: 0.8.0-beta.5
      - name: konveyor-operator.v0.8.0-beta.4
        version: 0.8.0-beta.4
      - name: konveyor-operator.v0.8.0-beta.2
        version: 0.8.0-beta.2
      - name: konveyor-operator.v0.8.0-beta.1
        version: 0.8.0-beta.1
      - name: konveyor-operator.v0.6.0-beta.2
        version: 0.6.0-beta.2
      - name: konveyor-operator.v0.6.0-beta.1
        version: 0.6.0-beta.1
      - name: konveyor-operator.v0.5.0-beta.3
        version: 0.5.0-beta.3
      - name: konveyor-operator.v0.5.0-beta.2
        version: 0.5.0-beta.2
      - name: konveyor-operator.v0.5.0-beta.1
        version: 0.5.0-beta.1
      - name: konveyor-operator.v0.3.0-beta.3
        version: 0.3.0-beta.3
      - name: konveyor-operator.v0.3.0-beta.2
        version: 0.3.0-beta.2
      - name: konveyor-operator.v0.3.0-beta.1.2
        version: 0.3.0-beta.1.2
      - name: konveyor-operator.v0.3.0-beta.1.1
        version: 0.3.0-beta.1.1
      - name: konveyor-operator.v0.3.0-beta.1
        version: 0.3.0-beta.1
      name: beta
    - currentCSV: konveyor-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "move2kube-plan",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "image": "quay.io/konveyor/tackle-addon-move2kube-plan:latest"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:9688d0763d8c7812beee715a061c3c198006448855d7b891d45833f62cf869fb
          createdAt: "2023-05-25T15:23:45Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.1.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)

        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:9688d0763d8c7812beee715a061c3c198006448855d7b891d45833f62cf869fb
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/postgresql-12-centos7@sha256:1287d947464fbf5275feb31d106d2feb856c6117f0eaec8d947f2e03bd8618c5
        - quay.io/konveyor/tackle2-addon@sha256:35be344bbd3393f5634f52aafd21e9c3964d1e0ba5011ecad1561600d476bbe9
        - quay.io/konveyor/tackle-keycloak-init@sha256:b0347e5a608352c4b7bb796b65a67d417e98ba4896db71a487fa26243214cfcf
        - quay.io/konveyor/tackle2-hub@sha256:3d15d158ffb6f1fc4ca5aaf29a12173fe344c289773902c9161056bdf69acc92
        - quay.io/konveyor/tackle2-ui@sha256:99650804e65e2b6cb15a37c00189d7f463e4d19a0f866a79e9c09c8420a4d028
        - quay.io/konveyor/tackle2-addon-windup@sha256:81f0625cb82eb24ded51c7258f229f57778820c62e3be6ec05fb2b12d2970825
        - quay.io/konveyor/tackle-pathfinder@sha256:f9c5aa49f1d27f477cdd82c21441ae0287acdcdc9b2b3fcf197884007dc5f7dd
        version: 0.1.2
      entries:
      - name: konveyor-operator.v0.1.2
        version: 0.1.2
      - name: konveyor-operator.v0.1.1
        version: 0.1.1
      - name: konveyor-operator.v0.1.0
        version: 0.1.0
      name: konveyor-0.1
    - currentCSV: konveyor-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "move2kube-plan",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "image": "quay.io/konveyor/tackle-addon-move2kube-plan:latest"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:1860bb099ac64d9a2af27692f0671b7240e4d19746fffa9f9ae2356ff2818e10
          createdAt: "2023-07-06T06:17:40Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.2.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)

        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:1860bb099ac64d9a2af27692f0671b7240e4d19746fffa9f9ae2356ff2818e10
        - quay.io/konveyor/postgresql-12-centos7@sha256:6016530ab1fce71d29421ce9a97b3a89fbbb126b87b1b079c6aadebb7ff5664c
        - quay.io/konveyor/tackle2-hub@sha256:d9c4f21f44098a2526474968b6d88094dca8b5eaeb9042716332c9cb67df47cd
        - quay.io/konveyor/tackle2-ui@sha256:d10c1e717d160c5eb2581649a8dd7cdf4922a74d92db014f72f42a48a3134341
        - quay.io/konveyor/tackle2-addon-windup@sha256:538dfdb134d53a876d126dbd16679638584a606ef2b9e8fbe69ed1f29733c03a
        - quay.io/konveyor/tackle-keycloak-init@sha256:ca6c85bdf8417d4be116c2098b8d8ad5ae8b7ab0b636c4cbe4a392aa424e2d47
        - quay.io/konveyor/tackle2-addon@sha256:11693e16363bcd98249209d63a29f0c6cfffab3a4071649cbeea2fd276bc2440
        - quay.io/konveyor/tackle-pathfinder@sha256:190269b061b2065ca8ba9357bf9b959a2f587a547a44769ae0afd1ac86b4b197
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        version: 0.2.1
      entries:
      - name: konveyor-operator.v0.2.1
        version: 0.2.1
      - name: konveyor-operator.v0.2.0
        version: 0.2.0
      name: konveyor-0.2
    - currentCSV: konveyor-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:cd2305373f50ce927393e43f5735915e46c32d6f8a76091ee2808218885bba72
          createdAt: "2024-05-02T22:12:42Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.3.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:cd2305373f50ce927393e43f5735915e46c32d6f8a76091ee2808218885bba72
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:51bda0c6577c15684583d607a4f0cebef760814102423f4b2025016093a19fc1
        - quay.io/konveyor/origin-oauth-proxy@sha256:05f6bd1216b921b1f47a3821f0e22b84ed6b141bb2ada474c12bd9fde613108c
        - quay.io/konveyor/tackle2-hub@sha256:bf8803721d06d1a39251e22892343a013d5ef0fb4c1883c95e1eb55aa683f486
        - quay.io/konveyor/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle-keycloak-init@sha256:a63702cc0cb36830a7830ed70ca344a9aacc201ba6f3c94806ca2e93160df4dd
        - quay.io/konveyor/tackle2-ui@sha256:bbf58d2ea613489f91efae450421afafb9052416b0d7971606e2729e89fd7de6
        version: 0.3.2
      entries:
      - name: konveyor-operator.v0.3.2
        version: 0.3.2
      - name: konveyor-operator.v0.3.1
        version: 0.3.1
      - name: konveyor-operator.v0.3.0
        version: 0.3.0
      name: konveyor-0.3
    - currentCSV: konveyor-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:496db8ca859608d04008ac36cf7bdc9cafd7214294d33a9c5646083c361ffcb6
          createdAt: "2024-06-05T03:08:51Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.4.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:496db8ca859608d04008ac36cf7bdc9cafd7214294d33a9c5646083c361ffcb6
        - quay.io/konveyor/tackle-keycloak-init@sha256:4c7e07b1fce05f7dca47edecc0ba0ce373cc38e48879cc7eed72510d2865a18e
        - quay.io/konveyor/tackle2-ui@sha256:cb1255fc5fa4608fc659c3455f3164664b960cc2d9a3b2208967798e3532df78
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:e84ea36d206a8c143794e3986940377e77c7618225cad594bdf9a4047d7823dc
        - quay.io/konveyor/origin-oauth-proxy@sha256:05f6bd1216b921b1f47a3821f0e22b84ed6b141bb2ada474c12bd9fde613108c
        - quay.io/konveyor/tackle2-hub@sha256:b87acd8e2ee4a7edc4369bc375f64e2b98bc479d09556439e2ee4d74b74970a9
        - quay.io/konveyor/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        version: 0.4.0
      entries:
      - name: konveyor-operator.v0.4.0
        version: 0.4.0
      name: konveyor-0.4
    - currentCSV: konveyor-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:474d739ed470701b628f1f9e971925ac7fbe0b543f3e36305d932b963b1225fd
          createdAt: "2024-08-27T15:48:02Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.5.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-hub@sha256:faf91ca68e24979642a44fa013b1fb9edfee415eee835ccef9dce2eb418e2909
        - quay.io/konveyor/java-external-provider@sha256:33fb5d24b998f01c481a0a03e3f1bf3f7a6c93f5b662885102a0c966026e3375
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle-keycloak-init@sha256:5f652be5914ae6e23537b9fe02ccc9cd638c08de5462d9f58135a405c81f5287
        - quay.io/konveyor/tackle2-ui@sha256:36f29fe01d3078f0e100b24d1ed126850c5a773f4b0ae922fd4c4487de1ac29d
        - quay.io/konveyor/generic-external-provider@sha256:334a631538d31efc8d82adae4f52303336800fa3037b5055c72400918d81955c
        - quay.io/konveyor/postgresql-15-c9s@sha256:039ce764cadd03b4cfea954700139b8ceff415a9a0970baa7f2cf60f7f929174
        - quay.io/konveyor/tackle2-operator@sha256:474d739ed470701b628f1f9e971925ac7fbe0b543f3e36305d932b963b1225fd
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:2afbe8bea09904e084e026cd0a294fbe1c7b05d0e387f6633b09c5eaf4665e02
        - quay.io/konveyor/tackle2-addon-discovery@sha256:461f960bb1d29e719fb472fe9be9cf3deca17d984687957f54546a09047b0d3c
        version: 0.5.3
      entries:
      - name: konveyor-operator.v0.5.3
        version: 0.5.3
      - name: konveyor-operator.v0.5.1
        version: 0.5.1
      - name: konveyor-operator.v0.5.0
        version: 0.5.0
      name: konveyor-0.5
    - currentCSV: konveyor-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:8ee143c473570efd27b6f93332dc272871e11c66cc20f381c3a1e6e5688abf68
          createdAt: "2025-03-05T21:14:52Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.6.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle-keycloak-init@sha256:a3741269aa15edd2f9fb7df853a03a942d16a88eec114f3e54f82ef63b82adc4
        - quay.io/konveyor/tackle2-ui@sha256:12b0eda13a63508b9c13d4f9309bbfbe19d564627cf0de672307af1a6ec993fb
        - quay.io/konveyor/tackle2-operator@sha256:8ee143c473570efd27b6f93332dc272871e11c66cc20f381c3a1e6e5688abf68
        - quay.io/konveyor/java-external-provider@sha256:15148cde9a6fedef2207fe36f0ac3bd1ecdc6e3ac76f1cbec9eb97ffcf247231
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle2-hub@sha256:52a8b9be038fbfbc4e22172a4a8c7191d2b06fd8670e5df2ec442e3ce00d8c8a
        - quay.io/konveyor/tackle2-addon-discovery@sha256:3eb6f395c451e2c9a867871b3984a25f9f8e6ddcfe388f37fbd6d46a61a70b1e
        - quay.io/konveyor/generic-external-provider@sha256:27c1731859a4a639d2750cff8d1246281f4b0466bb5b069169c3e170c3c19b30
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:0b5cbb322fbe96335522410376c2e8505a8302088c0e454f3e5e7e28ba7c96a6
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/postgresql-15-c9s@sha256:55f8014c1edafab8f9085228a0d207c4e98820c615be34b935ce98e942d388b1
        version: 0.6.1
      entries:
      - name: konveyor-operator.v0.6.1
        version: 0.6.1
      - name: konveyor-operator.v0.6.0
        version: 0.6.0
      name: konveyor-0.6
    - currentCSV: konveyor-operator.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:f5870bc26b2a07ddacf00ed5b5f7df7d1068489010a80bcc543686c6c9576fbb
          createdAt: "2025-07-10T15:16:17Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:f5870bc26b2a07ddacf00ed5b5f7df7d1068489010a80bcc543686c6c9576fbb
        - quay.io/konveyor/tackle-keycloak-init@sha256:dab4cb1229c8a729d14b4cecb4aa3b5c4d48e94c36e84d785c922fc08d120617
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:7ed52ea9b95650241b5c2c01914775e29be23f7ed75fabb6e7c71969f3f0728f
        - quay.io/konveyor/generic-external-provider@sha256:facca8861559a658710ad3aef8b325731595bf944cce1907715d6efd33887072
        - quay.io/konveyor/tackle2-hub@sha256:52f66607f3134a4165ce691c6e9ce992d93927e9591055daafdc1bfc1dd09dd9
        - quay.io/konveyor/tackle2-ui@sha256:f87b04762fd163f40c49f3bcffd473f0c3f80a643576c6a3e2c2b4a14395c09f
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/postgresql-15-c9s@sha256:41460103a0bab189cd77acb6e741d9c5a6d3506479f7e56df40eb156d9b2cadd
        - quay.io/konveyor/tackle2-addon-discovery@sha256:60951801aa539ee59cd43f3c5f42a2ca751ca70e995cbb258b7d55ebb24d8450
        - quay.io/konveyor/java-external-provider@sha256:a0ef733e9fce3845d02b5360440add85427acad697d2157ef92242dec06bd73b
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/origin-oauth-proxy@sha256:2c811da843057119b246784f21dc7eec490669a8569e8295d8e914cdce555b24
        version: 0.7.1
      entries:
      - name: konveyor-operator.v0.7.1
        version: 0.7.1
      - name: konveyor-operator.v0.7.0
        version: 0.7.0
      name: konveyor-0.7
    - currentCSV: konveyor-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:68b6672fccdf54c65cc83f4c82db023053d6b61409bea880f58b53ccfea7b1ed
          createdAt: "2025-09-29T20:41:09Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-addon-platform@sha256:d0280ade44ba5e15210c07342608576b38e6993af2e479740675a4706cd2c875
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:8c82ad880ca2a2d5cd4b08125feca16d13d014d15f4f77677306086d1067f9aa
        - quay.io/konveyor/generic-external-provider@sha256:c7b6618bfff81d73f1b455963107ffdaf19334e6552ce076c009476da3f37f21
        - quay.io/konveyor/postgresql-15-c9s@sha256:af275f35a307367dcee9a339a6903cf79e3a8a41cabf79fb7cea7015895f0b58
        - quay.io/konveyor/origin-oauth-proxy@sha256:8988f14db7dde3edcaed0541cecb0b4fd58fab9dbaffb79a6c1c4aee13b9bd73
        - quay.io/konveyor/kai-solution-server@sha256:bab6e224c50de5b44cc37ce0297a3b91255efc08c1ad0fccf14fef9e5182775f
        - quay.io/konveyor/tackle2-operator@sha256:68b6672fccdf54c65cc83f4c82db023053d6b61409bea880f58b53ccfea7b1ed
        - quay.io/konveyor/tackle-keycloak-init@sha256:c0051b8c3a3c891fe12012623aee775080ecb47af8a4e75b8be8b0b4855d9f2c
        - quay.io/konveyor/tackle2-ui@sha256:f3f7380be65ff1fb3459c37a3d4865354bab685dc836893f78221f16ba579f70
        - quay.io/konveyor/kantra@sha256:a5b5591a51143574a6ce6678b141b9cd3e14328c6fad773412dde21b442fe364
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-hub@sha256:b3e6db1ffc195143b8e5d54b60949256e154b0f8b667736512a084e015cdf2a7
        - quay.io/konveyor/tackle2-addon-discovery@sha256:96a5b37834fbed9506014bcbd909258081fd833ef9e16daa104effd0e3a3c6ea
        - quay.io/konveyor/java-external-provider@sha256:8eab08a04e98051f89352250d63523a5b591f35a6497f5d69a5a9d556a94abf9
        version: 0.8.0
      entries:
      - name: konveyor-operator.v0.8.0
        version: 0.8.0
      name: konveyor-0.8
    - currentCSV: konveyor-operator.v0.7.0-rc.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
          createdAt: "2025-05-21T17:12:48Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.0-rc.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-ui@sha256:afe9141366a23c072286517e9af5436d14fd1df775aaa2fa13c3911764515777
        - quay.io/konveyor/postgresql-15-c9s@sha256:012cd6611aa9d3f4e0a82636d6817f381c19e3b0809199216fbe7fa3396a2ef5
        - quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/generic-external-provider@sha256:de55f78939576c110eb6f084a7ca51338fedba699fbb5135f23392783a5d7b95
        - quay.io/konveyor/tackle2-addon-discovery@sha256:89d2b0f8636897eb4dc1957a74a5b5041c0f3267a747adcf2bc6ca56b17b9937
        - quay.io/konveyor/origin-oauth-proxy@sha256:6c2847fd9667938b1d1294d988038e885e43cd9fb378862dae3419ce4f21631f
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle-keycloak-init@sha256:c3124971dc31be9201de2a6a1f6b2f886adf4c96731188dd5e750798cde3cfe7
        - quay.io/konveyor/tackle2-hub@sha256:bed834eb5488eef23b0d1e8af194b0646cf74c39c197e0d57acc5aed2299d9f5
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:46150725a2adbde8dfdc2250614d437a50e64bbf0b4562dabc73715a287de67b
        - quay.io/konveyor/java-external-provider@sha256:61ab20e37916abfeeadde685336660f0e8bbf519403d1f7100d43c1e42a72fbd
        version: 0.7.0-rc.1
      entries:
      - name: konveyor-operator.v0.7.0-rc.1
        version: 0.7.0-rc.1
      - name: konveyor-operator.v0.6.0-rc.2
        version: 0.6.0-rc.2
      - name: konveyor-operator.v0.6.0-rc.1
        version: 0.6.0-rc.1
      - name: konveyor-operator.v0.5.0-rc.1
        version: 0.5.0-rc.1
      - name: konveyor-operator.v0.3.0-rc.4
        version: 0.3.0-rc.4
      - name: konveyor-operator.v0.3.0-rc.3
        version: 0.3.0-rc.3
      - name: konveyor-operator.v0.3.0-rc.2
        version: 0.3.0-rc.2
      - name: konveyor-operator.v0.3.0-rc.1
        version: 0.3.0-rc.1
      name: rc
    - currentCSV: konveyor-operator.v0.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:33cf01f19a7043bcc5bdb12a3ed6bde7a9094981aa2383554111ebec9b4ba498
          createdAt: "2025-03-26T20:25:41Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.6.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/generic-external-provider@sha256:bc1112352f1a2c3172c5e7d5e637b5c5a69bf7d98b78f494fa593e8872e5cfd1
        - quay.io/konveyor/tackle2-ui@sha256:3c568c6047a044095080b6013b643a4041deae7b45936a443d4c4d708a81b945
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:734752be111a97e4b7e0030d563c32883f419ad031b2f350cb8e7333f2f37de3
        - quay.io/konveyor/tackle2-hub@sha256:0df49087d0c9341c2120a7c3aa2ba7907c056f94e1602ffd0496ca7113c75382
        - quay.io/konveyor/postgresql-15-c9s@sha256:7957d3b272b9f4e469bdc0c514ec9e70e8fbdc34a7250b4f4c62f9bb73cbf08d
        - quay.io/konveyor/tackle2-addon-discovery@sha256:1b89266d6bdfa497ba8137e0d33ff53785ac476dd96bf4fd9a185ef0ddf513ce
        - quay.io/konveyor/java-external-provider@sha256:fe85689b57125d678df685012b93c8a567fcb9251777cd711e9b82bdb22a734c
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/tackle2-operator@sha256:33cf01f19a7043bcc5bdb12a3ed6bde7a9094981aa2383554111ebec9b4ba498
        - quay.io/konveyor/tackle-keycloak-init@sha256:e32fa48351f410361e8bc100303deeb40d8a66d7eefb0cf7b1bc16e067a9a795
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        version: 0.6.2
      entries:
      - name: konveyor-operator.v0.6.2
        version: 0.6.2
      name: release-0.6
    - currentCSV: konveyor-operator.v0.7.0-rc.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
          createdAt: "2025-05-21T17:12:48Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.0-rc.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/generic-external-provider@sha256:de55f78939576c110eb6f084a7ca51338fedba699fbb5135f23392783a5d7b95
        - quay.io/konveyor/origin-oauth-proxy@sha256:6c2847fd9667938b1d1294d988038e885e43cd9fb378862dae3419ce4f21631f
        - quay.io/konveyor/tackle2-addon-discovery@sha256:89d2b0f8636897eb4dc1957a74a5b5041c0f3267a747adcf2bc6ca56b17b9937
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-hub@sha256:bed834eb5488eef23b0d1e8af194b0646cf74c39c197e0d57acc5aed2299d9f5
        - quay.io/konveyor/postgresql-15-c9s@sha256:012cd6611aa9d3f4e0a82636d6817f381c19e3b0809199216fbe7fa3396a2ef5
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:46150725a2adbde8dfdc2250614d437a50e64bbf0b4562dabc73715a287de67b
        - quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
        - quay.io/konveyor/tackle-keycloak-init@sha256:c3124971dc31be9201de2a6a1f6b2f886adf4c96731188dd5e750798cde3cfe7
        - quay.io/konveyor/tackle2-ui@sha256:afe9141366a23c072286517e9af5436d14fd1df775aaa2fa13c3911764515777
        - quay.io/konveyor/java-external-provider@sha256:61ab20e37916abfeeadde685336660f0e8bbf519403d1f7100d43c1e42a72fbd
        version: 0.7.0-rc.1
      entries:
      - name: konveyor-operator.v0.7.0-rc.1
        version: 0.7.0-rc.1
      name: release-0.7
    defaultChannel: konveyor-0.8
    packageName: konveyor-operator
    provider:
      name: Konveyor
      url: https://www.konveyor.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: sriov-network-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sriov-network-operator.v4.13.0-202309190544
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovIBNetwork",
                "metadata": {
                  "name": "example-sriovibnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "linkState": "auto",
                  "networkNamespace": "default",
                  "resourceName": "mellanoxnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetwork",
                "metadata": {
                  "name": "example-sriovnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "intelnics",
                  "spoofChk": "on",
                  "trust": "off",
                  "vlan": 0
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodePolicy",
                "metadata": {
                  "name": "policy-1",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "deviceType": "vfio-pci",
                  "mtu": 9000,
                  "nicSelector": {
                    "pfNames": [
                      "eth1"
                    ],
                    "rootDevices": [
                      "0000:01:00.1"
                    ],
                    "vendor": "8086"
                  },
                  "nodeSelector": {
                    "feature.node.kubernetes.io/network-sriov.capable": "true"
                  },
                  "numVfs": 6,
                  "priority": 99,
                  "resourceName": "intelnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodeState",
                "metadata": {
                  "name": "minikube",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "interfaces": []
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkPoolConfig",
                "metadata": {
                  "name": "sriovnetworkpoolconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovOperatorConfig",
                "metadata": {
                  "name": "default",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "enableInjector": true,
                  "enableOperatorWebhook": true,
                  "logLevel": 2
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: quay.io/pliurh/sriov-network-operator-manager:bd1
          createdAt: "2023-08-02T09:22:34Z"
          description: An operator for configuring SR-IOV components and initializing
            SRIOV network devices in Openshift cluster.
          olm.skipRange: '>=4.3.0-0 <4.13.0-202309190544'
          operatorframework.io/suggested-namespace: openshift-sriov-network-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "cni"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/k8snetworkplumbingwg/sriov-network-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovIBNetwork is the Schema for the sriovibnetworks API
            displayName: Sriov IBNetwork
            kind: SriovIBNetwork
            name: sriovibnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodePolicy is the Schema for the sriovnetworknodepolicies
              API
            displayName: Sriov Network Node Policy
            kind: SriovNetworkNodePolicy
            name: sriovnetworknodepolicies.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodeState is the Schema for the sriovnetworknodestates
              API
            displayName: Sriov Network Node State
            kind: SriovNetworkNodeState
            name: sriovnetworknodestates.sriovnetwork.openshift.io
            version: v1
          - kind: SriovNetworkPoolConfig
            name: sriovnetworkpoolconfigs.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetwork is the Schema for the sriovnetworks API
            displayName: Sriov Network
            kind: SriovNetwork
            name: sriovnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovOperatorConfig is the Schema for the sriovoperatorconfigs
              API
            displayName: Sriov Operator Config
            kind: SriovOperatorConfig
            name: sriovoperatorconfigs.sriovnetwork.openshift.io
            version: v1
        description: |
          # SR-IOV Network Operator for Openshift

          ## Introduction
          The sriov-network-operator is generally responsible for configuring the sriov components in a openshift cluster.

          ### Supported Features
          * Initialize the SR-IOV NICs on nodes.
          * provision SR-IOV device plugin on selected node.
          * provision SR-IOV CNI plugin on selected nodes.
          * manage configuration of SR-IOV device plugin.
          * generate net-att-def CRs for SR-IOV CNI plugin.

          This operator has to run in namespace 'openshift-sriov-network-operator'. An Operator Group is also required to install this operator:

          ```
          $ oc create namespace openshift-sriov-network-operator

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: sriov-network-operators
            namespace: openshift-sriov-network-operator
          spec:
            targetNamespaces:
            - openshift-sriov-network-operator
          EOF
          ```
        displayName: SR-IOV Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - network
        - sr-iov
        - sriov
        links:
        - name: Source Code
          url: https://github.com/k8snetworkplumbingwg/sriov-network-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-sriov-network-config-daemon@sha256:a2053c5366eb89112dfbc7a9994685fc198cce6fa87fdd6d1c0a07727e6eee03
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin@sha256:82d4846f8dfb9e0d02e7ff3f532f8f2bf44c9aba63c37d5ce2efb36969e5c453
        - registry.redhat.io/openshift4/ose-sriov-network-operator@sha256:3314ee580113b985e275583585b1f2adf381c5a0885b51100b3a01872a77a555
        - registry.redhat.io/openshift4/ose-sriov-network-webhook@sha256:bfb84d00dbb6b18d596ca2720b52d922c589249d2f988679189ea0eea985e58a
        - registry.redhat.io/openshift4/sriov-cni-rhel9@sha256:a4f0ec210cd065cb7f77dd25257aadf8345f47819cc728c5b6229acb79a09bb3
        - registry.redhat.io/openshift4/ose-sriov-dp-admission-controller@sha256:8dfaa07a5775a7156d303019879e41b17833e5643794f6e1087881d2a4dd1781
        - registry.redhat.io/openshift4/ose-sriov-infiniband-cni@sha256:716aa0b9ecfddfea0a62615f424f4abeab5362899c695e9ecc9df4df05f1add5
        version: 4.13.0-202309190544
      entries:
      - name: sriov-network-operator.v4.13.0-202309190544
        version: 4.13.0-202309190544
      name: "4.13"
    - currentCSV: sriov-network-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovIBNetwork",
                "metadata": {
                  "name": "example-sriovibnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "linkState": "auto",
                  "networkNamespace": "default",
                  "resourceName": "mellanoxnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetwork",
                "metadata": {
                  "name": "example-sriovnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "intelnics",
                  "spoofChk": "on",
                  "trust": "off",
                  "vlan": 0
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodePolicy",
                "metadata": {
                  "name": "policy-1",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "deviceType": "vfio-pci",
                  "mtu": 9000,
                  "nicSelector": {
                    "pfNames": [
                      "eth1"
                    ],
                    "rootDevices": [
                      "0000:01:00.1"
                    ],
                    "vendor": "8086"
                  },
                  "nodeSelector": {
                    "feature.node.kubernetes.io/network-sriov.capable": "true"
                  },
                  "numVfs": 6,
                  "priority": 99,
                  "resourceName": "intelnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodeState",
                "metadata": {
                  "name": "minikube",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "interfaces": []
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkPoolConfig",
                "metadata": {
                  "name": "sriovnetworkpoolconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovOperatorConfig",
                "metadata": {
                  "name": "default",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "enableInjector": true,
                  "enableOperatorWebhook": true,
                  "logLevel": 2
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-sriov-network-rhel9-operator@sha256:8dcae2a77e904713fba9c5907789f9b22ffb1f941f793adc4b604798d29a44f2
          createdAt: "2024-05-07T23:50:52Z"
          description: An operator for configuring SR-IOV components and initializing
            SRIOV network devices in Openshift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.16.0-202509240123'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "sriovnetwork.openshift.io/v1",
              "kind": "SriovOperatorConfig",
              "metadata": {
                "name": "default"
              },
              "spec": {
                "configurationMode": "daemon",
                "disableDrain": false,
                "enableInjector": true,
                "enableOperatorWebhook": true,
                "logLevel": 2
              }
            }
          operatorframework.io/suggested-namespace: openshift-sriov-network-operator
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-sriov-network-operator",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["sriovnetworknodestates.sriovnetwork.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/k8snetworkplumbingwg/sriov-network-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovIBNetwork is the Schema for the sriovibnetworks API
            displayName: Sriov IBNetwork
            kind: SriovIBNetwork
            name: sriovibnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodePolicy is the Schema for the sriovnetworknodepolicies
              API
            displayName: Sriov Network Node Policy
            kind: SriovNetworkNodePolicy
            name: sriovnetworknodepolicies.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodeState is the Schema for the sriovnetworknodestates
              API
            displayName: Sriov Network Node State
            kind: SriovNetworkNodeState
            name: sriovnetworknodestates.sriovnetwork.openshift.io
            version: v1
          - kind: SriovNetworkPoolConfig
            name: sriovnetworkpoolconfigs.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetwork is the Schema for the sriovnetworks API
            displayName: Sriov Network
            kind: SriovNetwork
            name: sriovnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovOperatorConfig is the Schema for the sriovoperatorconfigs
              API
            displayName: Sriov Operator Config
            kind: SriovOperatorConfig
            name: sriovoperatorconfigs.sriovnetwork.openshift.io
            version: v1
        description: |
          # SR-IOV Network Operator for Openshift

          ## Introduction
          The sriov-network-operator is generally responsible for configuring the sriov components in a openshift cluster.

          ### Supported Features
          * Initialize the SR-IOV NICs on nodes.
          * provision SR-IOV device plugin on selected node.
          * provision SR-IOV CNI plugin on selected nodes.
          * manage configuration of SR-IOV device plugin.
          * generate net-att-def CRs for SR-IOV CNI plugin.

          This operator has to run in namespace 'openshift-sriov-network-operator'. An Operator Group is also required to install this operator:

          ```
          $ oc create namespace openshift-sriov-network-operator

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: sriov-network-operators
            namespace: openshift-sriov-network-operator
          spec:
            targetNamespaces:
            - openshift-sriov-network-operator
          EOF
          ```
        displayName: SR-IOV Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - network
        - sr-iov
        - sriov
        links:
        - name: Source Code
          url: https://github.com/k8snetworkplumbingwg/sriov-network-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-sriov-network-config-daemon-rhel9@sha256:61afbb271a08dc3ab3b5509df535d9456f9b2c1b267554386b4bf859b6bccbe4
        - registry.redhat.io/openshift4/ose-sriov-infiniband-cni-rhel9@sha256:f828e415db73bf2df3fa7b26238a2f5c8a53cda6b396c11a0337d688108e6647
        - registry.redhat.io/openshift4/sriov-cni-rhel9@sha256:cac67ac25a0d319b5ea68cb3f63eac058339f1981811c7a3466b9daa2534a38d
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin-rhel9@sha256:5b3bc02e31eecd0d3468db729b03239d518ef73112b0fdb5ec3f05a788baa0be
        - registry.redhat.io/openshift4/ose-sriov-network-rhel9-operator@sha256:8dcae2a77e904713fba9c5907789f9b22ffb1f941f793adc4b604798d29a44f2
        - registry.redhat.io/openshift4/ose-sriov-network-webhook-rhel9@sha256:a6a5cc76d88b6b7aa86ee24b882ce43be33c8c6d4bcfae5f1b5490ead71ce8b6
        - registry.redhat.io/openshift4/ose-sriov-network-metrics-exporter-rhel9@sha256:a0580d8156de0ca1e5a119441196e9babbeb4d7edc502659c7ac66a49b1928d5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-sriov-dp-admission-controller-rhel9@sha256:20fd8a0081ffdea05796c8f6a0539a0e246da24f739f7bbf0722c90acfdeed15
        version: 4.16.0-202509240123
      entries:
      - name: sriov-network-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: sriov-network-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: sriov-network-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: sriov-network-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: sriov-network-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: sriov-network-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: sriov-network-operator.v4.16.0-202504290837
        version: 4.16.0-202504290837
      - name: sriov-network-operator.v4.16.0-202504151638
        version: 4.16.0-202504151638
      - name: sriov-network-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: sriov-network-operator.v4.16.0-202502262005
        version: 4.16.0-202502262005
      - name: sriov-network-operator.v4.16.0-202502171404
        version: 4.16.0-202502171404
      - name: sriov-network-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: sriov-network-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: sriov-network-operator.v4.16.0-202501081506
        version: 4.16.0-202501081506
      - name: sriov-network-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: sriov-network-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: sriov-network-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: sriov-network-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: sriov-network-operator.v4.16.0-202410300306
        version: 4.16.0-202410300306
      - name: sriov-network-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: sriov-network-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: sriov-network-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: sriov-network-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: sriov-network-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: sriov-network-operator.v4.16.0-202409111535
        version: 4.16.0-202409111535
      - name: sriov-network-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: sriov-network-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: sriov-network-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: sriov-network-operator.v4.16.0-202408120337
        version: 4.16.0-202408120337
      - name: sriov-network-operator.v4.16.0-202407100107
        version: 4.16.0-202407100107
      - name: sriov-network-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: sriov-network-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: sriov-network-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: sriov-network-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: sriov-network-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: klusterlet
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: klusterlet.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "Klusterlet",
                "metadata": {
                  "name": "klusterlet"
                },
                "spec": {
                  "clusterName": "cluster1",
                  "deployOption": {
                    "mode": "Singleton"
                  },
                  "externalServerURLs": [
                    {
                      "url": "https://localhost"
                    }
                  ],
                  "imagePullSpec": "quay.io/open-cluster-management/registration-operator:v0.13.1",
                  "namespace": "open-cluster-management-agent",
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/registration-operator:v0.13.1
          createdAt: "2024-04-10T15:27:49Z"
          description: Manages the installation and upgrade of the Klusterlet.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management-io/ocm
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Klusterlet defines the configurations of agents running on
              the managed cluster
            displayName: Klusterlet
            kind: Klusterlet
            name: klusterlets.operator.open-cluster-management.io
            version: v1
        description: "Note: This operator will stop updating, please use the latest
          operator in OperatorHub.io https://operatorhub.io/operator/klusterlet. \n\nThe
          Klusterlet proivdes the registation to the Hub clusters as a managed cluster.
          This operator supports the installation and upgrade of Klusterlet.\n\n##
          Prerequisites\nYou need a Hub cluster which has installed ClusterManager.\n\n##
          How to Install\n* Install Klusterlet operator by following instructions
          in top right button `Install`.\n* Optional: Create an agent namespace used
          to deploy Klusterlet. The default namespace is `open-cluster-management-agent`.\n*
          Create a secret `bootstrap-hub-kubeconfig` with the kubeconfig file of hub
          cluster in the agent namespace.\n\n  ```\n  $ kubectl create secret generic
          bootstrap-hub-kubeconfig --from-file=kubeconfig=<kubeconfig file of hub
          clsuter> -n <agent namespace>\n  ```\n\n* Create an instance of the Klusterlet
          resource to install the Klusterlet.\n\nThe pods of Klusterlet are created
          in agent namespace.\n\n```\n$ kubectl get pods -n open-cluster-management-agent\n
          \ NAME                                READY   STATUS    RESTARTS   AGE\n
          \ klusterlet-agent-55c6574876-nfxm9   1/1     Running   0          8m38s\n```\n\nThere
          is a ManagedCluster created in the Hub cluster.\n\n```\n$ kubectl get managedclusters.cluster.open-cluster-management.io\nNAME
          \                  HUB ACCEPTED   MANAGED CLUSTER URLS     JOINED   AVAILABLE
          \  AGE\ncluster1               false                                                        27s\n```\n"
        displayName: Klusterlet
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open-cluster-management
        - Klusterlet
        links:
        - name: Registration Operator
          url: https://github.com/open-cluster-management-io/ocm
        maintainers:
        - email: zyin@redhat.com
          name: Zhiwei Yin
        - email: deads@redhat.com
          name: David Eads
        - email: jqiu@redhat.com
          name: Jian Qiu
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/open-cluster-management/registration-operator:v0.13.1
        version: 0.13.1
      entries:
      - name: klusterlet.v0.13.1
        version: 0.13.1
      - name: klusterlet.v0.13.0
        version: 0.13.0
      - name: klusterlet.v0.12.0
        version: 0.12.0
      - name: klusterlet.v0.11.0
        version: 0.11.0
      - name: klusterlet.v0.10.0
        version: 0.10.0
      - name: klusterlet.v0.9.1
        version: 0.9.1
      - name: klusterlet.v0.8.0
        version: 0.8.0
      - name: klusterlet.v0.7.0
        version: 0.7.0
      - name: klusterlet.v0.6.0
        version: 0.6.0
      - name: klusterlet.v0.5.0
        version: 0.5.0
      - name: klusterlet.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: klusterlet
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.12
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-06-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:9aed98a061b2c76f3c512c1d6453df5920f0c779b33d7168c1e7ab9d8bc928a4
        version: 0.0.12
      entries:
      - name: joget-dx8-openshift-operator.v0.0.12
        version: 0.0.12
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Dell Technologies
      provider-url: https://github.com/dell/csm-operator
    name: dell-csm-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dell-csm-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "authorization",
                  "namespace": "authorization"
                },
                "spec": {
                  "modules": [
                    {
                      "components": [
                        {
                          "enabled": true,
                          "name": "nginx"
                        },
                        {
                          "enabled": true,
                          "name": "cert-manager"
                        },
                        {
                          "authorizationController": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231",
                          "authorizationControllerReplicas": 1,
                          "certificate": "",
                          "controllerReconcileInterval": "5m",
                          "enabled": true,
                          "hostname": "csm-authorization.com",
                          "leaderElection": true,
                          "name": "proxy-server",
                          "opa": "docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095",
                          "opaKubeMgmt": "docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517",
                          "openTelemetryCollectorAddress": "",
                          "privateKey": "",
                          "proxyServerIngress": [
                            {
                              "annotations": {},
                              "hosts": [],
                              "ingressClassName": "nginx"
                            }
                          ],
                          "proxyService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682",
                          "proxyServiceReplicas": 1,
                          "roleService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c",
                          "roleServiceReplicas": 1,
                          "storageService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5",
                          "storageServiceReplicas": 1,
                          "tenantService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358",
                          "tenantServiceReplicas": 1
                        },
                        {
                          "commander": "docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a",
                          "name": "redis",
                          "redis": "docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f",
                          "redisCommander": "rediscommander",
                          "redisName": "redis-csm",
                          "redisReplicas": 5,
                          "redisSecretProviderClass": [
                            {
                              "redisPasswordKey": "",
                              "redisSecretName": "",
                              "redisUsernameKey": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "sentinel": "sentinel"
                        },
                        {
                          "configSecretProviderClass": [
                            {
                              "configSecretName": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "name": "config"
                        },
                        {
                          "name": "storage-system-credentials",
                          "secretProviderClasses": {
                            "vault": [
                              "secret-provider-class-1",
                              "secret-provider-class-2"
                            ]
                          }
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": true,
                      "forceRemoveModule": true,
                      "name": "authorization-proxy-server"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "isilon",
                  "namespace": "isilon"
                },
                "spec": {
                  "driver": {
                    "authSecret": "isilon-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VERBOSE",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_ISI_PORT",
                          "value": "8080"
                        },
                        {
                          "name": "X_CSI_ISI_PATH",
                          "value": "/ifs/data/csi"
                        },
                        {
                          "name": "X_CSI_ISI_NO_PROBE_ON_START",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_AUTOPROBE",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_AUTH_TYPE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_CUSTOM_TOPOLOGY_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOISILON_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "AZ_RECONCILE_INTERVAL",
                          "value": "1h"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_ISI_QUOTA_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_ACCESS_ZONE",
                          "value": "System"
                        },
                        {
                          "name": "X_CSI_ISI_VOLUME_PATH_PERMISSIONS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_IGNORE_UNRESOLVABLE_HOSTS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "isilon",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerscale"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSCALE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_PERFORMANCE_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSCALE_QUOTA_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "ISICLIENT_INSECURE",
                              "value": "true"
                            },
                            {
                              "name": "ISICLIENT_AUTH_TYPE",
                              "value": "1"
                            },
                            {
                              "name": "ISICLIENT_VERBOSE",
                              "value": "0"
                            },
                            {
                              "name": "POWERSCALE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSCALE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0",
                          "name": "metrics-powerscale"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-isilon/csi_sock",
                            "--mode=node",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powermax",
                  "namespace": "powermax"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powermax-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_MANAGED_ARRAYS",
                          "value": "000000000001,000000000002"
                        },
                        {
                          "name": "X_CSI_POWERMAX_PORTGROUPS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_TRANSPORT_PROTOCOL",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_K8S_CLUSTER_PREFIX",
                          "value": "CSM"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "X_CSI_VSPHERE_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VSPHERE_PORTGROUP",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VSPHERE_HOSTNAME",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VCENTER_HOST",
                          "value": ""
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "X_CSI_POWERMAX_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "X_CSI_POWERMAX_DEBUG",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_REVPROXY_USE_SECRET",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powermax",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERMAX_ISCSI_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_TOPOLOGY_CONTROL_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": [
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/memory-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/disk-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/network-unavailable",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REVPROXY_TLS_SECRET",
                              "value": "csirevproxy-tls-secret"
                            },
                            {
                              "name": "X_CSI_REVPROXY_PORT",
                              "value": "2222"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            },
                            {
                              "name": "DeployAsSidecar",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265",
                          "name": "csipowermax-reverseproxy"
                        }
                      ],
                      "configVersion": "v2.14.0",
                      "name": "csireverseproxy"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powermax"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERMAX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERMAX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414",
                          "name": "metrics-powermax"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/powermax.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powerstore",
                  "namespace": "powerstore"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powerstore-config",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_NODE_NAME_PREFIX",
                          "value": "csi-node"
                        },
                        {
                          "name": "X_CSI_FC_PORTS_FILTER_FILE_PATH",
                          "value": "/etc/fc-ports-filter"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOPOWERSTORE_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSM_AUTH_ENABLED",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_MULTI_NAS_FAILURE_THRESHOLD",
                          "value": "5"
                        },
                        {
                          "name": "X_CSI_MULTI_NAS_COOLDOWN_PERIOD",
                          "value": "5m"
                        },
                        {
                          "name": "X_CSI_NFS_ACLS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerstore",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-powerstore.dellemc.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerstore"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSTORE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_SPACE_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_ARRAY_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_FILE_SYSTEM_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSTORE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSTORE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "ZIPKIN_URI",
                              "value": ""
                            },
                            {
                              "name": "ZIPKIN_SERVICE_NAME",
                              "value": "metrics-powerstore"
                            },
                            {
                              "name": "ZIPKIN_PROBABILITY",
                              "value": "0.0"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60",
                          "name": "metrics-powerstore"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "unity",
                  "namespace": "unity"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_UNITY_ALLOW_MULTI_POD_ACCESS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_EPHEMERAL_STAGING_PATH",
                          "value": "/var/lib/kubelet/plugins/kubernetes.io/csi/pv/"
                        },
                        {
                          "name": "X_CSI_ISCSI_CHROOT",
                          "value": "/noderoot"
                        },
                        {
                          "name": "X_CSI_UNITY_SYNC_NODEINFO_INTERVAL",
                          "value": "15"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOUNITY_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOUNITY_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "TENANT_NAME",
                          "value": ""
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_UNITY_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "unity",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "vxflexos",
                  "namespace": "vxflexos"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLELISTVOLUMESNAPSHOT",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLESNAPSHOTCGDELETE",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "INFO"
                        },
                        {
                          "name": "GOSCALEIO_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOSCALEIO_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_QUOTA_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "INTERFACE_NAMES",
                          "value": null
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERFLEX_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "File",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerflex",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "initContainers": [
                      {
                        "envs": [
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          },
                          {
                            "name": "REPO_ADDRESS",
                            "value": "sftp://0.0.0.0"
                          },
                          {
                            "name": "REPO_USER",
                            "value": "sdcRepoUser"
                          },
                          {
                            "name": "MODULE_SIGCHECK",
                            "value": "0"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "imagePullPolicy": "IfNotPresent",
                        "name": "sdc"
                      }
                    ],
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_SDC_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_APPROVE_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_PREFIX",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_SDC_SFTP_REPO_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 1,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "enabled": false,
                        "envs": [
                          {
                            "name": "HOST_PID",
                            "value": "1"
                          },
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "name": "sdc-monitor"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "csi-external-health-monitor-controller"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERFLEX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_SDC_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_SDC_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERFLEX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d",
                          "name": "metrics-powerflex"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerflex"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--csisock=unix:/var/lib/kubelet/plugins/vxflexos.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
          createdAt: "2025-09-05T13:18:42Z"
          description: Easily install and manage Dell’s CSI Drivers and CSM
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/dell/csm-operator
          support: Dell Technologies
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ContainerStorageModule is the Schema for the containerstoragemodules
              API
            displayName: Container Storage Module
            kind: ContainerStorageModule
            name: containerstoragemodules.storage.dell.com
            version: v1
        description: "Dell Container Storage Modules (CSM) Operator is a Kubernetes
          Operator which can be \nused to install and manage Dell’s CSI drivers and
          CSM modules. \nBy using Dell CSM Operator, enterprises can quickly and easily
          deploy the \nCSM modules for CSI drivers making it easy for DevOps \nteams
          to build and optimize their Kubernetes workloads using enterprise grade
          storage \nwhile consistently aligning with the desired outcomes of the organization.\nThis
          operator is available as both community and certified operator for Openshift
          \nand as a community operator for upstream Kubernetes.\nOperator can be
          deployed via OperatorHub.io. using Operator Lifecycle Manager (OLM)\n\n###
          About Dell CSM\n\nDell CSM brings powerful enterprise storage features and
          functionality to Kubernetes for \neasier adoption of cloud native workloads,
          improved productivity and scalable operations. \n\nCSM delivers software
          modules for storage management that provide developers access to \nbuild
          automation for enhanced IT needs and other critical enterprise storage features.
          \n\nCurrently, the following modules are available as part of Dell CSM.
          \n*\tObservability - Delivers a high-level view of the storage capacity
          \n  and performance usage via Grafana dashboards to the Kubernetes users\n*\tReplication
          - Enables array replication capabilities for Kubernetes users\n*\tAuthorization
          - Provides both storage and Kubernetes administrators the ability \n  to
          apply RBAC for Dell CSI Drivers\n*\tResiliency - Designed to make Kubernetes
          applications more resilient to \n  node failures\n\nThese modules are available
          for various storage platforms in the Dell portfolio, \nsuch as Dell PowerStore,
          Dell PowerMax, Dell PowerFlex, Dell PowerScale, and Dell Unity XT.\n\nFor
          more information, visit [Dell’s Container Storage Module web page](https://www.dell.com/en-us/dt/solutions/kubernetes-containers/kubernetes-storage.htm)\n\nAdditional
          Materials\n*\t[Documentation](https://dell.github.io/csm-docs)\n*\t[CSM
          Solution Brief](https://www.delltechnologies.com/asset/en-us/products/storage/briefs-summaries/csm-solution-brief.pdf)\n"
        displayName: Dell Container Storage Modules
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dell Container Storage Modules
        - Dell CSI Driver
        - Dell CSM Modules
        - Powerflex
        - Powerscale
        - Powerstore
        - Unity
        - Authorization
        - Observability
        - Replication
        links:
        - name: Documentation
          url: https://dell.github.io/csm-docs
        maintainers:
        - email: container.storage.modules@dell.com
          name: Dell Container Storage Modules
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Dell Technologies
          url: https://github.com/dell/csm-operator
        relatedImages:
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0
        - registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa
        - docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265
        - registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea
        - docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682
        - registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334
        - docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f
        - quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:59335acaca515fceed5f0350c46cfcfd6037475f0cf35ae8ab41d37333039e2e
        - ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e
        - quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540
        - registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf
        - registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16
        - registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe
        version: 1.10.0
      entries:
      - name: dell-csm-operator-certified.v1.10.0
        version: 1.10.0
      - name: dell-csm-operator-certified.v1.9.1
        version: 1.9.1
      - name: dell-csm-operator-certified.v1.9.0
        version: 1.9.0
      - name: dell-csm-operator-certified.v1.8.1
        version: 1.8.1
      - name: dell-csm-operator-certified.v1.8.0
        version: 1.8.0
      - name: dell-csm-operator-certified.v1.7.0
        version: 1.7.0
      - name: dell-csm-operator-certified.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: dell-csm-operator-certified
    provider:
      name: Dell Technologies
      url: https://github.com/dell/csm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: external-dns-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: external-dns-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "provider": {
                    "type": "AWS",
                    "aws": {
                      "credentials": {
                        "name": "aws-access-key"
                      }
                    }
                  },
                  "source": {
                    "type": "Service"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:6be1403e418f9b63ec8ebcc8858e602409bc41fdee4586acc627623a78b50eb5
          createdAt: 2022-03-01T21:00:46
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## Prerequisites and Requirements
           The ExternalDNS Operator has minimal cluster level permissions. Unfortunately OLM doesn't allow the local role creation in the operand namespace yet.
           Therefore make sure the following commands are run against your cluster before the installation of the operator:
           * Create the operand namespace: `oc create ns external-dns`
           * Grant the operator access to manage operand resources: `oc apply -f https://raw.githubusercontent.com/openshift/external-dns-operator/release-0.1/config/rbac/extra-roles.yaml`
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b6fde43491c2e0733c8fe6ee8660f7f6435fe20c0b6d435cfc7f86eeab8b4748
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:6be1403e418f9b63ec8ebcc8858e602409bc41fdee4586acc627623a78b50eb5
        - registry.redhat.io/edo/external-dns-rhel8@sha256:c1134bb46172997ef7278b6cefbb0da44e72a9f808a7cd67b3c65d464754cab9
        version: 0.1.2
      entries:
      - name: external-dns-operator.v0.1.2
        version: 0.1.2
      name: alpha
    - currentCSV: external-dns-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
          createdAt: 2023-11-15T09:24:04
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e33874a51971a90917cf437fc6cbeea1da569ba07217f66f1bba96d594f58aed
        - registry.redhat.io/edo/external-dns-rhel8@sha256:6263a78b2a170e56dfca2173f14dbf14f0a5bddb6ea600bfdef248c7d5f20e4b
        version: 1.2.0
      entries:
      - name: external-dns-operator.v1.2.0
        version: 1.2.0
      name: stable-v1
    - currentCSV: external-dns-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:0877eb0a111efe0bff63173a2dfeb7a06a61afd0efbcdba053322a539e0dc667
          createdAt: 2024-01-22T09:26:37
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/edo/external-dns-rhel8@sha256:17f4127cbb9a9304769e9655c19cbea2260107c7e9371a99a9aa3e85d759c00b
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:0877eb0a111efe0bff63173a2dfeb7a06a61afd0efbcdba053322a539e0dc667
        version: 1.0.1
      entries:
      - name: external-dns-operator.v1.0.1
        version: 1.0.1
      - name: external-dns-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: external-dns-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:9ca751e84b8843cc62dc60550e5c4af741bf09a381ef62220e7eb3760ffe8524
          createdAt: 2024-01-22T09:56:14
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:9ca751e84b8843cc62dc60550e5c4af741bf09a381ef62220e7eb3760ffe8524
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/edo/external-dns-rhel8@sha256:ffcc3f2e9b9feb889e7882465f7f034aba6b9ee0699a1b6c24e83f501a63ecca
        version: 1.1.1
      entries:
      - name: external-dns-operator.v1.1.1
        version: 1.1.1
      - name: external-dns-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: external-dns-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
          createdAt: 2023-11-15T09:24:04
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e33874a51971a90917cf437fc6cbeea1da569ba07217f66f1bba96d594f58aed
        - registry.redhat.io/edo/external-dns-rhel8@sha256:6263a78b2a170e56dfca2173f14dbf14f0a5bddb6ea600bfdef248c7d5f20e4b
        version: 1.2.0
      entries:
      - name: external-dns-operator.v1.2.0
        version: 1.2.0
      name: stable-v1.2
    defaultChannel: stable-v1
    packageName: external-dns-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: rook-ceph-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rook-ceph-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephCluster",
                "metadata": {
                  "name": "my-rook-ceph",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "cephVersion": {
                    "image": "registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d"
                  },
                  "dataDirHostPath": "/var/lib/rook",
                  "mon": {
                    "count": 3
                  },
                  "dashboard": {
                    "enabled": true
                  },
                  "network": {
                    "hostNetwork": false
                  },
                  "rbdMirroring": {
                    "workers": 0
                  },
                  "storage": {
                    "useAllNodes": true,
                    "useAllDevices": true
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephBlockPool",
                "metadata": {
                  "name": "replicapool",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "failureDomain": "host",
                  "replicated": {
                    "size": 3
                  },
                  "annotations": null
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStore",
                "metadata": {
                  "name": "my-store",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "metadataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "dataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "gateway": {
                    "type": "s3",
                    "sslCertificateRef": null,
                    "port": 8080,
                    "securePort": null,
                    "instances": 1,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStoreUser",
                "metadata": {
                  "name": "my-user",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "store": "my-store",
                  "displayName": "my display name"
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephNFS",
                "metadata": {
                  "name": "my-nfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "rados": {
                    "pool": "myfs-data0",
                    "namespace": "nfs-ns"
                  },
                  "server": {
                    "active": 3,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephClient",
                "metadata": {
                  "name": "cinder",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "caps": {
                    "mon": "profile rbd",
                    "osd": "profile rbd pool=volumes, profile rbd pool=vms, profile rbd-read-only pool=images"
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephFilesystem",
                "metadata": {
                  "name": "myfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "dataPools": [
                    {
                      "compressionMode": "",
                      "crushRoot": "",
                      "deviceClass": "",
                      "erasureCoded": {
                        "algorithm": "",
                        "codingChunks": 0,
                        "dataChunks": 0
                      },
                      "failureDomain": "host",
                      "replicated": {
                        "requireSafeReplicaSize": false,
                        "size": 1,
                        "targetSizeRatio": 0.5
                      }
                    }
                  ],
                  "metadataPool": {
                    "compressionMode": "",
                    "crushRoot": "",
                    "deviceClass": "",
                    "erasureCoded": {
                      "algorithm": "",
                      "codingChunks": 0,
                      "dataChunks": 0
                    },
                    "failureDomain": "",
                    "replicated": {
                      "requireSafeReplicaSize": false,
                      "size": 1,
                      "targetSizeRatio": 0
                    }
                  },
                  "metadataServer": {
                    "activeCount": 1,
                    "activeStandby": true,
                    "placement": {},
                    "resources": {}
                  },
                  "preservePoolsOnDelete": false,
                  "preserveFilesystemOnDelete": false
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephRBDMirror",
                "metadata": {
                  "name": "my-rbd-mirror",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "annotations": null,
                  "count": 1,
                  "placement": {
                    "topologyKey": "kubernetes.io/hostname"
                  },
                  "resources": null
                }
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:cf7f72469d25fb37301a8159406613673f8968c82af1035fd45a470e0ae8acfa
          externalClusterScript: |-
            IiIiCkNvcHlyaWdodCAyMDIwIFRoZSBSb29rIEF1dGhvcnMuIEFsbCByaWdodHMgcmVzZXJ2ZWQu
            CgpMaWNlbnNlZCB1bmRlciB0aGUgQXBhY2hlIExpY2Vuc2UsIFZlcnNpb24gMi4wICh0aGUgIkxp
            Y2Vuc2UiKTsKeW91IG1heSBub3QgdXNlIHRoaXMgZmlsZSBleGNlcHQgaW4gY29tcGxpYW5jZSB3
            aXRoIHRoZSBMaWNlbnNlLgpZb3UgbWF5IG9idGFpbiBhIGNvcHkgb2YgdGhlIExpY2Vuc2UgYXQK
            CglodHRwOi8vd3d3LmFwYWNoZS5vcmcvbGljZW5zZXMvTElDRU5TRS0yLjAKClVubGVzcyByZXF1
            aXJlZCBieSBhcHBsaWNhYmxlIGxhdyBvciBhZ3JlZWQgdG8gaW4gd3JpdGluZywgc29mdHdhcmUK
            ZGlzdHJpYnV0ZWQgdW5kZXIgdGhlIExpY2Vuc2UgaXMgZGlzdHJpYnV0ZWQgb24gYW4gIkFTIElT
            IiBCQVNJUywKV0lUSE9VVCBXQVJSQU5USUVTIE9SIENPTkRJVElPTlMgT0YgQU5ZIEtJTkQsIGVp
            dGhlciBleHByZXNzIG9yIGltcGxpZWQuClNlZSB0aGUgTGljZW5zZSBmb3IgdGhlIHNwZWNpZmlj
            IGxhbmd1YWdlIGdvdmVybmluZyBwZXJtaXNzaW9ucyBhbmQKbGltaXRhdGlvbnMgdW5kZXIgdGhl
            IExpY2Vuc2UuCiIiIgoKaW1wb3J0IGVycm5vCmltcG9ydCBzeXMKaW1wb3J0IGpzb24KaW1wb3J0
            IGFyZ3BhcnNlCmltcG9ydCByZQppbXBvcnQgc3VicHJvY2VzcwppbXBvcnQgaG1hYwpmcm9tIGhh
            c2hsaWIgaW1wb3J0IHNoYTEgYXMgc2hhCmZyb20gb3MgaW1wb3J0IGxpbmVzZXAgYXMgTElORVNF
            UApmcm9tIG9zIGltcG9ydCBwYXRoCmZyb20gZW1haWwudXRpbHMgaW1wb3J0IGZvcm1hdGRhdGUK
            aW1wb3J0IHJlcXVlc3RzCmZyb20gcmVxdWVzdHMuYXV0aCBpbXBvcnQgQXV0aEJhc2UKCnB5M2sg
            PSBGYWxzZQppZiBzeXMudmVyc2lvbl9pbmZvLm1ham9yID49IDM6CiAgICBweTNrID0gVHJ1ZQog
            ICAgaW1wb3J0IHVybGxpYi5wYXJzZQogICAgZnJvbSBpcGFkZHJlc3MgaW1wb3J0IGlwX2FkZHJl
            c3MsIElQdjRBZGRyZXNzCgpNb2R1bGVOb3RGb3VuZEVycm9yID0gSW1wb3J0RXJyb3IKCnRyeToK
            ICAgIGltcG9ydCByYWRvcwpleGNlcHQgTW9kdWxlTm90Rm91bmRFcnJvciBhcyBub01vZEVycjoK
            ICAgIHByaW50KGYiRXJyb3I6IHtub01vZEVycn1cbkV4aXRpbmcgdGhlIHNjcmlwdC4uLiIpCiAg
            ICBzeXMuZXhpdCgxKQoKdHJ5OgogICAgaW1wb3J0IHJiZApleGNlcHQgTW9kdWxlTm90Rm91bmRF
            cnJvciBhcyBub01vZEVycjoKICAgIHByaW50KGYiRXJyb3I6IHtub01vZEVycn1cbkV4aXRpbmcg
            dGhlIHNjcmlwdC4uLiIpCiAgICBzeXMuZXhpdCgxKQoKdHJ5OgogICAgIyBmb3IgMi43LngKICAg
            IGZyb20gU3RyaW5nSU8gaW1wb3J0IFN0cmluZ0lPCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9y
            OgogICAgIyBmb3IgMy54CiAgICBmcm9tIGlvIGltcG9ydCBTdHJpbmdJTwoKdHJ5OgogICAgIyBm
            b3IgMi43LngKICAgIGZyb20gdXJscGFyc2UgaW1wb3J0IHVybHBhcnNlCiAgICBmcm9tIHVybGxp
            YiBpbXBvcnQgdXJsZW5jb2RlIGFzIHVybGVuY29kZQpleGNlcHQgTW9kdWxlTm90Rm91bmRFcnJv
            cjoKICAgICMgZm9yIDMueAogICAgZnJvbSB1cmxsaWIucGFyc2UgaW1wb3J0IHVybHBhcnNlCiAg
            ICBmcm9tIHVybGxpYi5wYXJzZSBpbXBvcnQgdXJsZW5jb2RlIGFzIHVybGVuY29kZQoKdHJ5Ogog
            ICAgZnJvbSBiYXNlNjQgaW1wb3J0IGVuY29kZXN0cmluZwpleGNlcHQ6CiAgICBmcm9tIGJhc2U2
            NCBpbXBvcnQgZW5jb2RlYnl0ZXMgYXMgZW5jb2Rlc3RyaW5nCgoKY2xhc3MgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbihFeGNlcHRpb24pOgogICAgcGFzcwoKCiMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMgRHVtbXlSYWRv
            cyAjIyMjIyMjIyMjIyMjIyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjCiMgdGhpcyBpcyBtYWlubHkgZm9yIHRlc3RpbmcgYW5kIGNvdWxkIGJlIHVz
            ZWQgd2hlcmUgJ3JhZG9zJyBpcyBub3QgYXZhaWxhYmxlCgoKY2xhc3MgRHVtbXlSYWRvcyhvYmpl
            Y3QpOgogICAgZGVmIF9faW5pdF9fKHNlbGYpOgogICAgICAgIHNlbGYucmV0dXJuX3ZhbCA9IDAK
            ICAgICAgICBzZWxmLmVycl9tZXNzYWdlID0gIiIKICAgICAgICBzZWxmLnN0YXRlID0gImNvbm5l
            Y3RlZCIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwID0ge30KICAgICAgICBzZWxmLmNtZF9u
            YW1lcyA9IHt9CiAgICAgICAgc2VsZi5faW5pdF9jbWRfb3V0cHV0X21hcCgpCiAgICAgICAgc2Vs
            Zi5kdW1teV9ob3N0X2lwX21hcCA9IHt9CgogICAgZGVmIF9pbml0X2NtZF9vdXRwdXRfbWFwKHNl
            bGYpOgogICAgICAgIGpzb25fZmlsZV9uYW1lID0gInRlc3QtZGF0YS9jZXBoLXN0YXR1cy1vdXQi
            CiAgICAgICAgc2NyaXB0X2RpciA9IHBhdGguYWJzcGF0aChwYXRoLmRpcm5hbWUoX19maWxlX18p
            KQogICAgICAgIGNlcGhfc3RhdHVzX3N0ciA9ICIiCiAgICAgICAgd2l0aCBvcGVuKAogICAgICAg
            ICAgICBwYXRoLmpvaW4oc2NyaXB0X2RpciwganNvbl9maWxlX25hbWUpLCBtb2RlPSJyIiwgZW5j
            b2Rpbmc9IlVURi04IgogICAgICAgICkgYXMganNvbl9maWxlOgogICAgICAgICAgICBjZXBoX3N0
            YXR1c19zdHIgPSBqc29uX2ZpbGUucmVhZCgpCiAgICAgICAgc2VsZi5jbWRfbmFtZXNbImZzIGxz
            Il0gPSAiIiJ7ImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJmcyBscyJ9IiIiCiAgICAgICAg
            c2VsZi5jbWRfbmFtZXNbInF1b3J1bV9zdGF0dXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3Jt
            YXQiOiAianNvbiIsICJwcmVmaXgiOiAicXVvcnVtX3N0YXR1cyJ9IiIiCiAgICAgICAgKQogICAg
            ICAgIHNlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXSA9ICgKICAgICAgICAgICAgIiIieyJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAibWdyIHNlcnZpY2VzIn0iIiIKICAgICAgICApCiAg
            ICAgICAgIyBhbGwgdGhlIGNvbW1hbmRzIGFuZCB0aGVpciBvdXRwdXQKICAgICAgICBzZWxmLmNt
            ZF9vdXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJmcyBscyJdXSA9ICgKICAgICAgICAgICAgIiIi
            W3sibmFtZSI6Im15ZnMiLCJtZXRhZGF0YV9wb29sIjoibXlmcy1tZXRhZGF0YSIsIm1ldGFkYXRh
            X3Bvb2xfaWQiOjIsImRhdGFfcG9vbF9pZHMiOlszXSwiZGF0YV9wb29scyI6WyJteWZzLXJlcGxp
            Y2F0ZWQiXX1dIiIiCiAgICAgICAgKQogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbc2VsZi5j
            bWRfbmFtZXNbInF1b3J1bV9zdGF0dXMiXV0gPSAoCiAgICAgICAgICAgICIiInsiZWxlY3Rpb25f
            ZXBvY2giOjMsInF1b3J1bSI6WzBdLCJxdW9ydW1fbmFtZXMiOlsiYSJdLCJxdW9ydW1fbGVhZGVy
            X25hbWUiOiJhIiwicXVvcnVtX2FnZSI6MTQzODUsImZlYXR1cmVzIjp7InF1b3J1bV9jb24iOiI0
            NTQwMTM4MjkyODM2Njk2MDYzIiwicXVvcnVtX21vbiI6WyJrcmFrZW4iLCJsdW1pbm91cyIsIm1p
            bWljIiwib3NkbWFwLXBydW5lIiwibmF1dGlsdXMiLCJvY3RvcHVzIl19LCJtb25tYXAiOnsiZXBv
            Y2giOjEsImZzaWQiOiJhZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5ZDBmMWMiLCJtb2Rp
            ZmllZCI6IjIwMjAtMDUtMDdUMDM6MzY6MzkuOTE4MDM1WiIsImNyZWF0ZWQiOiIyMDIwLTA1LTA3
            VDAzOjM2OjM5LjkxODAzNVoiLCJtaW5fbW9uX3JlbGVhc2UiOjE1LCJtaW5fbW9uX3JlbGVhc2Vf
            bmFtZSI6Im9jdG9wdXMiLCJmZWF0dXJlcyI6eyJwZXJzaXN0ZW50IjpbImtyYWtlbiIsImx1bWlu
            b3VzIiwibWltaWMiLCJvc2RtYXAtcHJ1bmUiLCJuYXV0aWx1cyIsIm9jdG9wdXMiXSwib3B0aW9u
            YWwiOltdfSwibW9ucyI6W3sicmFuayI6MCwibmFtZSI6ImEiLCJwdWJsaWNfYWRkcnMiOnsiYWRk
            cnZlYyI6W3sidHlwZSI6InYyIiwiYWRkciI6IjEwLjExMC4yMDUuMTc0OjMzMDAiLCJub25jZSI6
            MH0seyJ0eXBlIjoidjEiLCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OSIsIm5vbmNlIjowfV19
            LCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OS8wIiwicHVibGljX2FkZHIiOiIxMC4xMTAuMjA1
            LjE3NDo2Nzg5LzAiLCJwcmlvcml0eSI6MCwid2VpZ2h0IjowfV19fSIiIgogICAgICAgICkKICAg
            ICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXV0g
            PSAoCiAgICAgICAgICAgICIiInsiZGFzaGJvYXJkIjoiaHR0cHM6Ly9jZXBoLWRhc2hib2FyZDo4
            NDQzLyIsInByb21ldGhldXMiOiJodHRwOi8vY2VwaC1kYXNoYm9hcmQtZGI6OTI4My8ifSIiIgog
            ICAgICAgICkKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNh
            cHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMiLCAib3Nk
            IiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0
            YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250
            cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJdLCAiZW50aXR5Ijog
            ImNsaWVudC5oZWFsdGhjaGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRo
            IGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuaGVh
            bHRoY2hlY2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRB
            PT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cyIs
            Im9zZCI6InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5j
            b250cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJ9fV0iIiIKICAg
            ICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIs
            ICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAib3NkIiwgInBy
            b2ZpbGUgcmJkIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtbm9kZSIsICJmb3JtYXQiOiAi
            anNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIi
            W3siZW50aXR5IjoiY2xpZW50LmNzaS1yYmQtbm9kZSIsImtleSI6IkFRQk9nck5lSGJLMUF4QUF1
            YllCZVY4UzFVL0dQenE1U1ZlcTZnPT0iLCJjYXBzIjp7Im1vbiI6InByb2ZpbGUgcmJkLCBhbGxv
            dyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsIm9zZCI6InByb2ZpbGUgcmJkIn19XSIiIgogICAg
            ICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsibW9uIiwg
            InByb2ZpbGUgcmJkLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxs
            b3cgcnciLCAib3NkIiwgInByb2ZpbGUgcmJkIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQt
            cHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNy
            ZWF0ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktcmJkLXByb3Zp
            c2lvbmVyIiwia2V5IjoiQVFCTmdyTmUxZ2V5S3hBQThla1ZpUmRFK2hzczVPd2VZQmt3Tmc9PSIs
            ImNhcHMiOnsibWdyIjoiYWxsb3cgcnciLCJtb24iOiJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFu
            ZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJwcm9maWxlIHJiZCJ9fV0iIiIKICAgICAgICBzZWxm
            LmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyBy
            LCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3Nk
            IiwgImFsbG93IHJ3IHRhZyBjZXBoZnMgKj0qIiwgIm1kcyIsICJhbGxvdyBydyJdLCAiZW50aXR5
            IjogImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4Ijog
            ImF1dGggZ2V0LW9yLWNyZWF0ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVu
            dC5jc2ktY2VwaGZzLW5vZGUiLCJrZXkiOiJBUUJPZ3JOZUVOdW5LeEFBUENtZ0U3UjZHOERjWG5h
            SjFGMzJxZz09IiwiY2FwcyI6eyJtZHMiOiJhbGxvdyBydyIsIm1nciI6ImFsbG93IHJ3IiwibW9u
            IjoiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBy
            dyB0YWcgY2VwaGZzICo9KiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAg
            ICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kICdvc2Qg
            YmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3NkIiwgImFsbG93IHJ3IHRhZyBjZXBo
            ZnMgbWV0YWRhdGE9KiJdLCAiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVy
            IiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgog
            ICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIs
            ImtleSI6IkFRQk9nck5lQUZnY0dCQUF2R3FLT0FEMEQzeHhtVlkwUjkxMmRnPT0iLCJjYXBzIjp7
            Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2Nr
            bGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoifX1dIiIiCiAgICAg
            ICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAi
            YWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdyIiwgImFsbG93IHJ3
            IiwgIm9zZCIsICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiXSwgImVudGl0eSI6ICJj
            bGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVuc2hpZnQtc3RvcmFnZSIsICJmb3JtYXQi
            OiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0g
            IiIiW3siZW50aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0
            b3JhZ2UiLCJrZXkiOiJCUUJPZ3JOZUFGZ2NHQkFBdkdxS09BRDBEM3h4bVZZMFI5MTJkZz09Iiwi
            Y2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29z
            ZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIn19XSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsi
            bW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciIsICJh
            bGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT1teWZzIl0sICJl
            bnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0b3JhZ2Ut
            bXlmcyIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0i
            IiIKICAgICAgICBdID0gIiIiW3siZW50aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9u
            ZXItb3BlbnNoaWZ0LXN0b3JhZ2UtbXlmcyIsImtleSI6IkNRQk9nck5lQUZnY0dCQUF2R3FLT0FE
            MEQzeHhtVlkwUjkxMmRnPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cg
            ciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2Vw
            aGZzIG1ldGFkYXRhPW15ZnMifX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAg
            ICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCBxdW9y
            dW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIiwgImFsbG93IGNvbW1hbmQg
            Y29uZmlnIiwgIm9zZCIsICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFsbG93IHJ3eCBwb29sPWRl
            ZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9ZGVm
            YXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJndy5sb2csIGFsbG93IHgg
            cG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4Il0sICJlbnRpdHkiOiAiY2xpZW50LmhlYWx0
            aGNoZWNrZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0
            ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5oZWFsdGhjaGVja2VyIiwi
            a2V5IjoiQVFERmtiTmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsi
            bW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFu
            ZCB2ZXJzaW9uIiwgIm1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsICJvc2QiOiAicHJvZmls
            ZSByYmQtcmVhZC1vbmx5LCBhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhLCBhbGxvdyBy
            IHBvb2w9LnJndy5yb290LCBhbGxvdyBydyBwb29sPWRlZmF1bHQucmd3LmNvbnRyb2wsIGFsbG93
            IHJ4IHBvb2w9ZGVmYXVsdC5yZ3cubG9nLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0
            cy5pbmRleCJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAi
            IiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMs
            IGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cgY29tbWFuZCBjb25maWciLCAi
            b3NkIiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5j
            b250cm9sLCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1
            bHQucmd3LmJ1Y2tldHMuaW5kZXgiXSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIs
            ICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBjYXBzIn0iIiIKICAgICAgICBdID0g
            IiIiW3siZW50aXR5IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0NWJG
            UkFBVG5kTE5VU1JLcnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIjog
            ImFsbG93IGNvbW1hbmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFs
            bG93IHJ3eCBwb29sPWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFs
            bG93IHJ3IHBvb2w9ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJn
            dy5sb2csIGFsbG93IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4In19XSIiIgogICAg
            ICAgIHNlbGYuY21kX291dHB1dF9tYXBbIiIieyJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            bWdyIHNlcnZpY2VzIn0iIiJdID0gKAogICAgICAgICAgICAiIiJ7ImRhc2hib2FyZCI6ICJodHRw
            Oi8vcm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6NzAwMC8iLCAicHJvbWV0aGV1cyI6
            ICJodHRwOi8vcm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6OTI4My8ifSIiIgogICAg
            ICAgICkKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0
            eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            YXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJ7ImRhc2hib2FyZCI6ICJodHRwOi8vcm9vay1j
            ZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6NzAwMC8iLCAicHJvbWV0aGV1cyI6ICJodHRwOi8v
            cm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6OTI4My8ifSIiIgogICAgICAgIHNlbGYu
            Y21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNsaWVudC5oZWFsdGhj
            aGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldCJ9IiIiCiAgICAg
            ICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5oZWFsdGhjaGVja2VyIiwia2V5IjoiQVFERmti
            TmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjogImFsbG93
            IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFuZCB2ZXJzaW9uIiwg
            Im1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsICJvc2QiOiAicHJvZmlsZSByYmQtcmVhZC1v
            bmx5LCBhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhLCBhbGxvdyByIHBvb2w9LnJndy5y
            b290LCBhbGxvdyBydyBwb29sPWRlZmF1bHQucmd3LmNvbnRyb2wsIGFsbG93IHJ4IHBvb2w9ZGVm
            YXVsdC5yZ3cubG9nLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJ9fV0i
            IiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0eSI6
            ICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJh
            dXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0
            X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtbm9kZSIsICJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJb
            XSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5
            IjogImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRf
            b3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMt
            cHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIK
            ICAgICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAg
            ICAgICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVuc2hp
            ZnQtc3RvcmFnZSIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgog
            ICAgICAgIF0gPSAiIiJbXSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAg
            ICAgICIiInsiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlm
            dC1zdG9yYWdlLW15ZnMiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0i
            IiIKICAgICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAg
            ICAgICAgICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJb
            eyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsImtleSI6IkFRREZrYk5l
            ZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRBPT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIi
            LCAibWdyIjoiYWxsb3cgcnciLCAib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0q
            In19XSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2Fw
            cyI6IFsibW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1n
            ciIsICJhbGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIl0s
            ICJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpz
            b24iLCAicHJlZml4IjogImF1dGggY2FwcyJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmtiTmVmdDViRlJBQVRu
            ZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cgciwgIGFsbG93IGNv
            bW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciI6ImFsbG93IHJ3IiwgIm9zZCI6ImFsbG93IHJ3
            IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFw
            Wyd7ImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJzdGF0dXMifSddID0gY2VwaF9zdGF0dXNf
            c3RyCgogICAgZGVmIHNodXRkb3duKHNlbGYpOgogICAgICAgIHBhc3MKCiAgICBkZWYgZ2V0X2Zz
            aWQoc2VsZik6CiAgICAgICAgcmV0dXJuICJhZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5
            ZDBmMWMiCgogICAgZGVmIGNvbmZfcmVhZF9maWxlKHNlbGYpOgogICAgICAgIHBhc3MKCiAgICBk
            ZWYgY29ubmVjdChzZWxmKToKICAgICAgICBwYXNzCgogICAgZGVmIHBvb2xfZXhpc3RzKHNlbGYs
            IHBvb2xfbmFtZSk6CiAgICAgICAgcmV0dXJuIFRydWUKCiAgICBkZWYgbW9uX2NvbW1hbmQoc2Vs
            ZiwgY21kLCBvdXQpOgogICAgICAgIGpzb25fY21kID0ganNvbi5sb2FkcyhjbWQpCiAgICAgICAg
            anNvbl9jbWRfc3RyID0ganNvbi5kdW1wcyhqc29uX2NtZCwgc29ydF9rZXlzPVRydWUpCiAgICAg
            ICAgY21kX291dHB1dCA9IHNlbGYuY21kX291dHB1dF9tYXBbanNvbl9jbWRfc3RyXQogICAgICAg
            IHJldHVybiBzZWxmLnJldHVybl92YWwsIGNtZF9vdXRwdXQsIHN0cihzZWxmLmVycl9tZXNzYWdl
            LmVuY29kZSgidXRmLTgiKSkKCiAgICBkZWYgX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoc2VsZiwg
            aG9zdF9uYW1lKToKICAgICAgICBpcF9yZWdfeCA9IHJlLmNvbXBpbGUociJcZHsxLDN9LlxkezEs
            M30uXGR7MSwzfS5cZHsxLDN9IikKICAgICAgICAjIGlmIHByb3ZpZGVkIGhvc3QgaXMgZGlyZWN0
            bHkgYW4gSVAgYWRkcmVzcywgcmV0dXJuIHRoZSBzYW1lCiAgICAgICAgaWYgaXBfcmVnX3gubWF0
            Y2goaG9zdF9uYW1lKToKICAgICAgICAgICAgcmV0dXJuIGhvc3RfbmFtZQogICAgICAgIGltcG9y
            dCByYW5kb20KCiAgICAgICAgaG9zdF9pcCA9IHNlbGYuZHVtbXlfaG9zdF9pcF9tYXAuZ2V0KGhv
            c3RfbmFtZSwgIiIpCiAgICAgICAgaWYgbm90IGhvc3RfaXA6CiAgICAgICAgICAgIGhvc3RfaXAg
            PSBmIjE3Mi45LntyYW5kb20ucmFuZGludCgwLCAyNTQpfS57cmFuZG9tLnJhbmRpbnQoMCwgMjU0
            KX0iCiAgICAgICAgICAgIHNlbGYuZHVtbXlfaG9zdF9pcF9tYXBbaG9zdF9uYW1lXSA9IGhvc3Rf
            aXAKICAgICAgICBkZWwgcmFuZG9tCiAgICAgICAgcmV0dXJuIGhvc3RfaXAKCiAgICBAY2xhc3Nt
            ZXRob2QKICAgIGRlZiBSYWRvcyhjb25mZmlsZT1Ob25lKToKICAgICAgICByZXR1cm4gRHVtbXlS
            YWRvcygpCgoKY2xhc3MgUzNBdXRoKEF1dGhCYXNlKToKICAgICIiIkF0dGFjaGVzIEFXUyBBdXRo
            ZW50aWNhdGlvbiB0byB0aGUgZ2l2ZW4gUmVxdWVzdCBvYmplY3QuIiIiCgogICAgc2VydmljZV9i
            YXNlX3VybCA9ICJzMy5hbWF6b25hd3MuY29tIgoKICAgIGRlZiBfX2luaXRfXyhzZWxmLCBhY2Nl
            c3Nfa2V5LCBzZWNyZXRfa2V5LCBzZXJ2aWNlX3VybD1Ob25lKToKICAgICAgICBpZiBzZXJ2aWNl
            X3VybDoKICAgICAgICAgICAgc2VsZi5zZXJ2aWNlX2Jhc2VfdXJsID0gc2VydmljZV91cmwKICAg
            ICAgICBzZWxmLmFjY2Vzc19rZXkgPSBzdHIoYWNjZXNzX2tleSkKICAgICAgICBzZWxmLnNlY3Jl
            dF9rZXkgPSBzdHIoc2VjcmV0X2tleSkKCiAgICBkZWYgX19jYWxsX18oc2VsZiwgcik6CiAgICAg
            ICAgIyBDcmVhdGUgZGF0ZSBoZWFkZXIgaWYgaXQgaXMgbm90IGNyZWF0ZWQgeWV0LgogICAgICAg
            IGlmICJkYXRlIiBub3QgaW4gci5oZWFkZXJzIGFuZCAieC1hbXotZGF0ZSIgbm90IGluIHIuaGVh
            ZGVyczoKICAgICAgICAgICAgci5oZWFkZXJzWyJkYXRlIl0gPSBmb3JtYXRkYXRlKHRpbWV2YWw9
            Tm9uZSwgbG9jYWx0aW1lPUZhbHNlLCB1c2VnbXQ9VHJ1ZSkKICAgICAgICBzaWduYXR1cmUgPSBz
            ZWxmLmdldF9zaWduYXR1cmUocikKICAgICAgICBpZiBweTNrOgogICAgICAgICAgICBzaWduYXR1
            cmUgPSBzaWduYXR1cmUuZGVjb2RlKCJ1dGYtOCIpCiAgICAgICAgci5oZWFkZXJzWyJBdXRob3Jp
            emF0aW9uIl0gPSBmIkFXUyB7c2VsZi5hY2Nlc3Nfa2V5fTp7c2lnbmF0dXJlfSIKICAgICAgICBy
            ZXR1cm4gcgoKICAgIGRlZiBnZXRfc2lnbmF0dXJlKHNlbGYsIHIpOgogICAgICAgIGNhbm9uaWNh
            bF9zdHJpbmcgPSBzZWxmLmdldF9jYW5vbmljYWxfc3RyaW5nKHIudXJsLCByLmhlYWRlcnMsIHIu
            bWV0aG9kKQogICAgICAgIGlmIHB5M2s6CiAgICAgICAgICAgIGtleSA9IHNlbGYuc2VjcmV0X2tl
            eS5lbmNvZGUoInV0Zi04IikKICAgICAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZy5lbmNv
            ZGUoInV0Zi04IikKICAgICAgICBlbHNlOgogICAgICAgICAgICBrZXkgPSBzZWxmLnNlY3JldF9r
            ZXkKICAgICAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZwogICAgICAgIGggPSBobWFjLm5l
            dyhrZXksIG1zZywgZGlnZXN0bW9kPXNoYSkKICAgICAgICByZXR1cm4gZW5jb2Rlc3RyaW5nKGgu
            ZGlnZXN0KCkpLnN0cmlwKCkKCiAgICBkZWYgZ2V0X2Nhbm9uaWNhbF9zdHJpbmcoc2VsZiwgdXJs
            LCBoZWFkZXJzLCBtZXRob2QpOgogICAgICAgIHBhcnNlZHVybCA9IHVybHBhcnNlKHVybCkKICAg
            ICAgICBvYmplY3RrZXkgPSBwYXJzZWR1cmwucGF0aFsxOl0KCiAgICAgICAgYnVja2V0ID0gcGFy
            c2VkdXJsLm5ldGxvY1s6IC1sZW4oc2VsZi5zZXJ2aWNlX2Jhc2VfdXJsKV0KICAgICAgICBpZiBs
            ZW4oYnVja2V0KSA+IDE6CiAgICAgICAgICAgICMgcmVtb3ZlIGxhc3QgZG90CiAgICAgICAgICAg
            IGJ1Y2tldCA9IGJ1Y2tldFs6LTFdCgogICAgICAgIGludGVyZXN0aW5nX2hlYWRlcnMgPSB7ImNv
            bnRlbnQtbWQ1IjogIiIsICJjb250ZW50LXR5cGUiOiAiIiwgImRhdGUiOiAiIn0KICAgICAgICBm
            b3Iga2V5IGluIGhlYWRlcnM6CiAgICAgICAgICAgIGxrID0ga2V5Lmxvd2VyKCkKICAgICAgICAg
            ICAgdHJ5OgogICAgICAgICAgICAgICAgbGsgPSBsay5kZWNvZGUoInV0Zi04IikKICAgICAgICAg
            ICAgZXhjZXB0OgogICAgICAgICAgICAgICAgcGFzcwogICAgICAgICAgICBpZiBoZWFkZXJzW2tl
            eV0gYW5kICgKICAgICAgICAgICAgICAgIGxrIGluIGludGVyZXN0aW5nX2hlYWRlcnMua2V5cygp
            IG9yIGxrLnN0YXJ0c3dpdGgoIngtYW16LSIpCiAgICAgICAgICAgICk6CiAgICAgICAgICAgICAg
            ICBpbnRlcmVzdGluZ19oZWFkZXJzW2xrXSA9IGhlYWRlcnNba2V5XS5zdHJpcCgpCgogICAgICAg
            ICMgSWYgeC1hbXotZGF0ZSBpcyB1c2VkIGl0IHN1cGVyc2VkZXMgdGhlIGRhdGUgaGVhZGVyLgog
            ICAgICAgIGlmIG5vdCBweTNrOgogICAgICAgICAgICBpZiAieC1hbXotZGF0ZSIgaW4gaW50ZXJl
            c3RpbmdfaGVhZGVyczoKICAgICAgICAgICAgICAgIGludGVyZXN0aW5nX2hlYWRlcnNbImRhdGUi
            XSA9ICIiCiAgICAgICAgZWxzZToKICAgICAgICAgICAgaWYgIngtYW16LWRhdGUiIGluIGludGVy
            ZXN0aW5nX2hlYWRlcnM6CiAgICAgICAgICAgICAgICBpbnRlcmVzdGluZ19oZWFkZXJzWyJkYXRl
            Il0gPSAiIgoKICAgICAgICBidWYgPSBmInttZXRob2R9XG4iCiAgICAgICAgZm9yIGtleSBpbiBz
            b3J0ZWQoaW50ZXJlc3RpbmdfaGVhZGVycy5rZXlzKCkpOgogICAgICAgICAgICB2YWwgPSBpbnRl
            cmVzdGluZ19oZWFkZXJzW2tleV0KICAgICAgICAgICAgaWYga2V5LnN0YXJ0c3dpdGgoIngtYW16
            LSIpOgogICAgICAgICAgICAgICAgYnVmICs9IGYie2tleX06e3ZhbH1cbiIKICAgICAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgICAgIGJ1ZiArPSBmInt2YWx9XG4iCgogICAgICAgICMgYXBwZW5k
            IHRoZSBidWNrZXQgaWYgaXQgZXhpc3RzCiAgICAgICAgaWYgYnVja2V0ICE9ICIiOgogICAgICAg
            ICAgICBidWYgKz0gZiIve2J1Y2tldH0iCgogICAgICAgICMgYWRkIHRoZSBvYmplY3RrZXkuIGV2
            ZW4gaWYgaXQgZG9lc24ndCBleGlzdCwgYWRkIHRoZSBzbGFzaAogICAgICAgIGJ1ZiArPSBmIi97
            b2JqZWN0a2V5fSIKCiAgICAgICAgcmV0dXJuIGJ1ZgoKCmNsYXNzIFJhZG9zSlNPTjoKICAgIEVY
            VEVSTkFMX1VTRVJfTkFNRSA9ICJjbGllbnQuaGVhbHRoY2hlY2tlciIKICAgIEVYVEVSTkFMX1JH
            V19BRE1JTl9PUFNfVVNFUl9OQU1FID0gInJndy1hZG1pbi1vcHMtdXNlciIKICAgIEVNUFRZX09V
            VFBVVF9MSVNUID0gIkVtcHR5IG91dHB1dCBsaXN0IgogICAgREVGQVVMVF9SR1dfUE9PTF9QUkVG
            SVggPSAiZGVmYXVsdCIKICAgIERFRkFVTFRfTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUID0gIjky
            ODMiCgogICAgQGNsYXNzbWV0aG9kCiAgICBkZWYgZ2VuX2FyZ19wYXJzZXIoY2xzLCBhcmdzX3Rv
            X3BhcnNlPU5vbmUpOgogICAgICAgIGFyZ1AgPSBhcmdwYXJzZS5Bcmd1bWVudFBhcnNlcigpCgog
            ICAgICAgIGNvbW1vbl9ncm91cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJjb21tb24iKQog
            ICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoIi0tdmVyYm9zZSIsICItdiIsIGFjdGlv
            bj0ic3RvcmVfdHJ1ZSIsIGRlZmF1bHQ9RmFsc2UpCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9h
            cmd1bWVudCgKICAgICAgICAgICAgIi0tY2VwaC1jb25mIiwgIi1jIiwgaGVscD0iUHJvdmlkZSBh
            IGNlcGggY29uZiBmaWxlLiIsIHR5cGU9c3RyCiAgICAgICAgKQogICAgICAgIGNvbW1vbl9ncm91
            cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLWtleXJpbmciLCAiLWsiLCBoZWxwPSJQYXRo
            IHRvIGNlcGgga2V5cmluZyBmaWxlLiIsIHR5cGU9c3RyCiAgICAgICAgKQogICAgICAgIGNvbW1v
            bl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJ1bi1hcy11c2VyIiwKICAgICAg
            ICAgICAgIi11IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgdHlwZT1zdHIs
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIGEgdXNlciBuYW1lIHRvIGNoZWNrIHRoZSBjbHVz
            dGVyJ3MgaGVhbHRoIHN0YXR1cywgbXVzdCBiZSBwcmVmaXhlZCBieSAnY2xpZW50LiciLAogICAg
            ICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1j
            bHVzdGVyLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJL
            dWJlcm5ldGVzIGNsdXN0ZXIgbmFtZShsZWdhY3kgZmxhZyksIE5vdGU6IEVpdGhlciB1c2UgdGhp
            cyBvciAtLWs4cy1jbHVzdGVyLW5hbWUiLAogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAu
            YWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1rOHMtY2x1c3Rlci1uYW1lIiwgZGVmYXVsdD0i
            IiwgaGVscD0iS3ViZXJuZXRlcyBjbHVzdGVyIG5hbWUiCiAgICAgICAgKQogICAgICAgIGNvbW1v
            bl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLW5hbWVzcGFjZSIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIGhlbHA9Ik5hbWVzcGFjZSB3aGVyZSBDZXBoQ2x1
            c3RlciBpcyBydW5uaW5nIiwKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tcmd3LXBvb2wtcHJlZml4IiwgZGVmYXVsdD0iIiwgaGVscD0i
            UkdXIFBvb2wgcHJlZml4IgogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3Vt
            ZW50KAogICAgICAgICAgICAiLS1yZXN0cmljdGVkLWF1dGgtcGVybWlzc2lvbiIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlJlc3RyaWN0IGNlcGhDU0lLZXly
            aW5ncyBhdXRoIHBlcm1pc3Npb25zIHRvIHNwZWNpZmljIHBvb2xzLCBjbHVzdGVyLiIKICAgICAg
            ICAgICAgKyAiTWFuZGF0b3J5IGZsYWdzIHRoYXQgbmVlZCB0byBiZSBzZXQgYXJlIC0tcmJkLWRh
            dGEtcG9vbC1uYW1lLCBhbmQgLS1rOHMtY2x1c3Rlci1uYW1lLiIKICAgICAgICAgICAgKyAiLS1j
            ZXBoZnMtZmlsZXN5c3RlbS1uYW1lIGZsYWcgY2FuIGFsc28gYmUgcGFzc2VkIGluIGNhc2Ugb2Yg
            Y2VwaGZzIHVzZXIgcmVzdHJpY3Rpb24sIHNvIGl0IGNhbiByZXN0cmljdCB1c2VyIHRvIHBhcnRp
            Y3VsYXIgY2VwaGZzIGZpbGVzeXN0ZW0iCiAgICAgICAgICAgICsgInNhbXBsZSBydW46IGBweXRo
            b24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwtY2x1c3Rlci1yZXNvdXJjZXMucHkgLS1jZXBo
            ZnMtZmlsZXN5c3RlbS1uYW1lIG15ZnMgLS1yYmQtZGF0YS1wb29sLW5hbWUgcmVwbGljYXBvb2wg
            LS1rOHMtY2x1c3Rlci1uYW1lIHJvb2tzdG9yYWdlIC0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Np
            b24gdHJ1ZWAiCiAgICAgICAgICAgICsgIk5vdGU6IFJlc3RyaWN0aW5nIHRoZSBjc2ktdXNlcnMg
            cGVyIHBvb2wsIGFuZCBwZXIgY2x1c3RlciB3aWxsIHJlcXVpcmUgY3JlYXRpbmcgbmV3IGNzaS11
            c2VycyBhbmQgbmV3IHNlY3JldHMgZm9yIHRoYXQgY3NpLXVzZXJzLiIKICAgICAgICAgICAgKyAi
            U28gYXBwbHkgdGhlc2Ugc2VjcmV0cyBvbmx5IHRvIG5ldyBgQ29uc3VtZXIgY2x1c3RlcmAgZGVw
            bG95bWVudCB3aGlsZSB1c2luZyB0aGUgc2FtZSBgU291cmNlIGNsdXN0ZXJgLiIsCiAgICAgICAg
            KQogICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXYyLXBv
            cnQtZW5hYmxlIiwKICAgICAgICAgICAgYWN0aW9uPSJzdG9yZV90cnVlIiwKICAgICAgICAgICAg
            ZGVmYXVsdD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iRW5hYmxlIHYyIG1vbiBwb3J0KDMzMDAp
            IGZvciBtb25zIiwKICAgICAgICApCgogICAgICAgIG91dHB1dF9ncm91cCA9IGFyZ1AuYWRkX2Fy
            Z3VtZW50X2dyb3VwKCJvdXRwdXQiKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQo
            CiAgICAgICAgICAgICItLWZvcm1hdCIsCiAgICAgICAgICAgICItdCIsCiAgICAgICAgICAgIGNo
            b2ljZXM9WyJqc29uIiwgImJhc2giXSwKICAgICAgICAgICAgZGVmYXVsdD0ianNvbiIsCiAgICAg
            ICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBvdXRwdXQgZm9ybWF0IChqc29uIHwgYmFzaCkiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1vdXRwdXQiLAogICAgICAgICAgICAiLW8iLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAg
            ICAgICAgICBoZWxwPSJPdXRwdXQgd2lsbCBiZSBzdG9yZWQgaW50byB0aGUgcHJvdmlkZWQgZmls
            ZSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAog
            ICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgQ2VwaCBmaWxlc3lzdGVt
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tY2VwaGZzLW1ldGFkYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIs
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBjZXBoZnMgbWV0YWRh
            dGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAg
            ICAgICAgICAgICItLWNlcGhmcy1kYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9
            IiIsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBjZXBoZnMgZGF0
            YSBwb29sIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAg
            ICAgICAgICAgIi0tcmJkLWRhdGEtcG9vbC1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwK
            ICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRo
            ZSBuYW1lIG9mIHRoZSBSQkQgZGF0YXBvb2wiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3Jv
            dXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1hbGlhcy1yYmQtZGF0YS1wb29sLW5hbWUi
            LAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAg
            ICAgICAgICAgaGVscD0iUHJvdmlkZXMgYW4gYWxpYXMgZm9yIHRoZSAgUkJEIGRhdGEgcG9vbCBu
            YW1lLCBuZWNlc3NhcnkgaWYgYSBzcGVjaWFsIGNoYXJhY3RlciBpcyBwcmVzZW50IGluIHRoZSBw
            b29sIG5hbWUgc3VjaCBhcyBhIHBlcmlvZCBvciB1bmRlcnNjb3JlIiwKICAgICAgICApCiAgICAg
            ICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmd3LWVuZHBvaW50
            IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAg
            ICAgICAgICAgIGhlbHA9IlJBRE9TIEdhdGV3YXkgZW5kcG9pbnQgKGluIGA8SVB2ND46PFBPUlQ+
            YCBvciBgPFtJUHY2XT46PFBPUlQ+YCBvciBgPEZRRE4+OjxQT1JUPmAgZm9ybWF0KSIsCiAgICAg
            ICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJn
            dy10bHMtY2VydC1wYXRoIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVx
            dWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlJBRE9TIEdhdGV3YXkgZW5kcG9pbnQgVExT
            IGNlcnRpZmljYXRlIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVu
            dCgKICAgICAgICAgICAgIi0tcmd3LXNraXAtdGxzIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFs
            c2UsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9Iklnbm9yZSBU
            TFMgY2VydGlmaWNhdGlvbiB2YWxpZGF0aW9uIHdoZW4gYSBzZWxmLXNpZ25lZCBjZXJ0aWZpY2F0
            ZSBpcyBwcm92aWRlZCAoTk9UIFJFQ09NTUVOREVEIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tbW9uaXRvcmluZy1lbmRwb2ludCIs
            CiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAg
            ICAgICAgICBoZWxwPSJDZXBoIE1hbmFnZXIgcHJvbWV0aGV1cyBleHBvcnRlciBlbmRwb2ludHMg
            KGNvbW1hIHNlcGFyYXRlZCBsaXN0IG9mIChmb3JtYXQgYDxJUHY0PmAgb3IgYDxbSVB2Nl0+YCBv
            ciBgPEZRRE4+YCkgZW50cmllcyBvZiBhY3RpdmUgYW5kIHN0YW5kYnkgbWdycykiLAogICAgICAg
            ICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1tb25p
            dG9yaW5nLWVuZHBvaW50LXBvcnQiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAg
            ICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iQ2VwaCBNYW5hZ2VyIHByb21ldGhl
            dXMgZXhwb3J0ZXIgcG9ydCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJn
            dW1lbnQoCiAgICAgICAgICAgICItLXNraXAtbW9uaXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAg
            ICAgICAgIGhlbHA9IkRvIG5vdCBjaGVjayBmb3IgYSBtb25pdG9yaW5nIGVuZHBvaW50IGZvciB0
            aGUgQ2VwaCBjbHVzdGVyIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tcmJkLW1ldGFkYXRhLWVjLXBvb2wtbmFtZSIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBo
            ZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiBlcmFzdXJlIGNvZGVkIFJCRCBtZXRhZGF0YSBwb29s
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tZHJ5LXJ1biIsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFj
            dGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGhlbHA9IkRyeSBydW4gcHJpbnRzIHRoZSBl
            eGVjdXRlZCBjb21tYW5kcyB3aXRob3V0IHJ1bm5pbmcgdGhlbSIsCiAgICAgICAgKQogICAgICAg
            IG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJhZG9zLW5hbWVzcGFj
            ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAog
            ICAgICAgICAgICBoZWxwPSJEaXZpZGVzIGEgcG9vbCBpbnRvIHNlcGFyYXRlIGxvZ2ljYWwgbmFt
            ZXNwYWNlcywgdXNlZCBmb3IgY3JlYXRpbmcgUkJEIFBWQyBpbiBhIENlcGhCbG9ja1Bvb2xSYWRv
            c05hbWVzcGFjZSAoc2hvdWxkIGJlIGxvd2VyIGNhc2UpIiwKICAgICAgICApCiAgICAgICAgb3V0
            cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tc3Vidm9sdW1lLWdyb3VwIiwK
            ICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAg
            ICAgICAgIGhlbHA9InByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBzdWJ2b2x1bWUgZ3JvdXAiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1yZ3ctcmVhbG0tbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJl
            cXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUg
            cmd3LXJlYWxtIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgK
            ICAgICAgICAgICAgIi0tcmd3LXpvbmUtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAg
            ICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUg
            bmFtZSBvZiB0aGUgcmd3LXpvbmUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS1yZ3ctem9uZWdyb3VwLW5hbWUiLAogICAgICAgICAg
            ICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVs
            cD0icHJvdmlkZXMgdGhlIG5hbWUgb2YgdGhlIHJndy16b25lZ3JvdXAiLAogICAgICAgICkKICAg
            ICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS10b3BvbG9neS1w
            b29scyIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNl
            LAogICAgICAgICAgICBoZWxwPSJjb21tYS1zZXBhcmF0ZWQgbGlzdCBvZiB0b3BvbG9neS1jb25z
            dHJhaW5lZCByYmQgcG9vbHMiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2Fy
            Z3VtZW50KAogICAgICAgICAgICAiLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi1sYWJlbCIsCiAg
            ICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAg
            ICAgICBoZWxwPSJrOHMgY2x1c3RlciBmYWlsdXJlIGRvbWFpbiBsYWJlbCAoZXhhbXBsZTogem9u
            ZSwgcmFjaywgb3IgaG9zdCkgZm9yIHRoZSB0b3BvbG9neS1wb29scyB0aGF0IG1hdGNoIHRoZSBj
            ZXBoIGRvbWFpbiIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQo
            CiAgICAgICAgICAgICItLXRvcG9sb2d5LWZhaWx1cmUtZG9tYWluLXZhbHVlcyIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBo
            ZWxwPSJjb21tYS1zZXBhcmF0ZWQgbGlzdCBvZiB0aGUgazhzIGNsdXN0ZXIgZmFpbHVyZSBkb21h
            aW4gdmFsdWVzIGNvcnJlc3BvbmRpbmcgdG8gZWFjaCBvZiB0aGUgcG9vbHMgaW4gdGhlIGB0b3Bv
            bG9neS1wb29sc2AgbGlzdCIsCiAgICAgICAgKQoKICAgICAgICB1cGdyYWRlX2dyb3VwID0gYXJn
            UC5hZGRfYXJndW1lbnRfZ3JvdXAoInVwZ3JhZGUiKQogICAgICAgIHVwZ3JhZGVfZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS11cGdyYWRlIiwKICAgICAgICAgICAgYWN0aW9uPSJz
            dG9yZV90cnVlIiwKICAgICAgICAgICAgZGVmYXVsdD1GYWxzZSwKICAgICAgICAgICAgaGVscD0i
            VXBncmFkZXMgdGhlIGNlcGhDU0lLZXlyaW5ncyhGb3IgZXhhbXBsZTogY2xpZW50LmNzaS1jZXBo
            ZnMtcHJvdmlzaW9uZXIpIGFuZCBjbGllbnQuaGVhbHRoY2hlY2tlciBjZXBoIHVzZXJzIHdpdGgg
            bmV3IHBlcm1pc3Npb25zIG5lZWRlZCBmb3IgdGhlIG5ldyBjbHVzdGVyIHZlcnNpb24gYW5kIG9s
            ZGVyIHBlcm1pc3Npb24gd2lsbCBzdGlsbCBiZSBhcHBsaWVkLiIKICAgICAgICAgICAgKyAiU2Ft
            cGxlIHJ1bjogYHB5dGhvbjMgL2V0Yy9jZXBoL2NyZWF0ZS1leHRlcm5hbC1jbHVzdGVyLXJlc291
            cmNlcy5weSAtLXVwZ3JhZGVgLCB0aGlzIHdpbGwgdXBncmFkZSBhbGwgdGhlIGRlZmF1bHQgY3Np
            IHVzZXJzKG5vbi1yZXN0cmljdGVkKSIKICAgICAgICAgICAgKyAiRm9yIHJlc3RyaWN0ZWQgdXNl
            cnMoRm9yIGV4YW1wbGU6IGNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1z
            dG9yYWdlLW15ZnMpLCB1c2VycyBjcmVhdGVkIHVzaW5nIC0tcmVzdHJpY3RlZC1hdXRoLXBlcm1p
            c3Npb24gZmxhZyBuZWVkIHRvIHBhc3MgbWFuZGF0b3J5IGZsYWdzIgogICAgICAgICAgICArICJt
            YW5kYXRvcnkgZmxhZ3M6ICctLXJiZC1kYXRhLXBvb2wtbmFtZSwgLS1rOHMtY2x1c3Rlci1uYW1l
            IGFuZCAtLXJ1bi1hcy11c2VyJyBmbGFncyB3aGlsZSB1cGdyYWRpbmciCiAgICAgICAgICAgICsg
            ImluIGNhc2Ugb2YgY2VwaGZzIHVzZXJzIGlmIHlvdSBoYXZlIHBhc3NlZCAtLWNlcGhmcy1maWxl
            c3lzdGVtLW5hbWUgZmxhZyB3aGlsZSBjcmVhdGluZyB1c2VyIHRoZW4gd2hpbGUgdXBncmFkaW5n
            IGl0IHdpbGwgYmUgbWFuZGF0b3J5IHRvbyIKICAgICAgICAgICAgKyAiU2FtcGxlIHJ1bjogYHB5
            dGhvbjMgL2V0Yy9jZXBoL2NyZWF0ZS1leHRlcm5hbC1jbHVzdGVyLXJlc291cmNlcy5weSAtLXVw
            Z3JhZGUgLS1yYmQtZGF0YS1wb29sLW5hbWUgcmVwbGljYXBvb2wgLS1rOHMtY2x1c3Rlci1uYW1l
            IHJvb2tzdG9yYWdlICAtLXJ1bi1hcy11c2VyIGNsaWVudC5jc2ktcmJkLW5vZGUtcm9va3N0b3Jh
            Z2UtcmVwbGljYXBvb2xgIgogICAgICAgICAgICArICJQUzogQW4gZXhpc3Rpbmcgbm9uLXJlc3Ry
            aWN0ZWQgdXNlciBjYW5ub3QgYmUgY29udmVydGVkIHRvIGEgcmVzdHJpY3RlZCB1c2VyIGJ5IHVw
            Z3JhZGluZy4iCiAgICAgICAgICAgICsgIlVwZ3JhZGUgZmxhZyBzaG91bGQgb25seSBiZSB1c2Vk
            IHRvIGFwcGVuZCBuZXcgcGVybWlzc2lvbnMgdG8gdXNlcnMsIGl0IHNob3VsZG4ndCBiZSB1c2Vk
            IGZvciBjaGFuZ2luZyB1c2VyIGFscmVhZHkgYXBwbGllZCBwZXJtaXNzaW9uLCBmb3IgZXhhbXBs
            ZSB5b3Ugc2hvdWxkbid0IGNoYW5nZSBpbiB3aGljaCBwb29sIHVzZXIgaGFzIGFjY2VzcyIsCiAg
            ICAgICAgKQoKICAgICAgICBpZiBhcmdzX3RvX3BhcnNlOgogICAgICAgICAgICBhc3NlcnQgKAog
            ICAgICAgICAgICAgICAgdHlwZShhcmdzX3RvX3BhcnNlKSA9PSBsaXN0CiAgICAgICAgICAgICks
            ICJBcmd1bWVudCB0byAnZ2VuX2FyZ19wYXJzZXInIHNob3VsZCBiZSBhIGxpc3QiCiAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgYXJnc190b19wYXJzZSA9IHN5cy5hcmd2WzE6XQogICAgICAgIHJl
            dHVybiBhcmdQLnBhcnNlX2FyZ3MoYXJnc190b19wYXJzZSkKCiAgICBkZWYgdmFsaWRhdGVfcmJk
            X21ldGFkYXRhX2VjX3Bvb2xfbmFtZShzZWxmKToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2Vy
            LnJiZF9tZXRhZGF0YV9lY19wb29sX25hbWU6CiAgICAgICAgICAgIHJiZF9tZXRhZGF0YV9lY19w
            b29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUKICAg
            ICAgICAgICAgcmJkX3Bvb2xfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9u
            YW1lCgogICAgICAgICAgICBpZiByYmRfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiRmxh
            ZyAnLS1yYmQtZGF0YS1wb29sLW5hbWUnIHNob3VsZCBub3QgYmUgZW1wdHkiCiAgICAgICAgICAg
            ICAgICApCgogICAgICAgICAgICBpZiByYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lID09ICIiOgog
            ICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAg
            ICAgICAgICAgICAiRmxhZyAnLS1yYmQtbWV0YWRhdGEtZWMtcG9vbC1uYW1lJyBzaG91bGQgbm90
            IGJlIGVtcHR5IgogICAgICAgICAgICAgICAgKQoKICAgICAgICAgICAgY21kX2pzb24gPSB7InBy
            ZWZpeCI6ICJvc2QgZHVtcCIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgICAgIHJldF92YWws
            IGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikK
            ICAgICAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAg
            ICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAg
            ICAgICAgZiJ7Y21kX2pzb25bJ3ByZWZpeCddfSBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAg
            ICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVN
            UFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgbWV0YWRhdGFf
            cG9vbF9leGlzdCwgcG9vbF9leGlzdCA9IEZhbHNlLCBGYWxzZQoKICAgICAgICAgICAgZm9yIGtl
            eSBpbiBqc29uX291dFsicG9vbHMiXToKICAgICAgICAgICAgICAgICMgaWYgZXJhc3VyZV9jb2Rl
            X3Byb2ZpbGUgaXMgZW1wdHkgYW5kIHBvb2wgbmFtZSBleGlzdHMgdGhlbiBpdCByZXBsaWNhIHBv
            b2wKICAgICAgICAgICAgICAgIGlmICgKICAgICAgICAgICAgICAgICAgICBrZXlbImVyYXN1cmVf
            Y29kZV9wcm9maWxlIl0gPT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQga2V5WyJwb29sX25h
            bWUiXSA9PSByYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lCiAgICAgICAgICAgICAgICApOgogICAg
            ICAgICAgICAgICAgICAgIG1ldGFkYXRhX3Bvb2xfZXhpc3QgPSBUcnVlCiAgICAgICAgICAgICAg
            ICAjIGlmIGVyYXN1cmVfY29kZV9wcm9maWxlIGlzIG5vdCBlbXB0eSBhbmQgcG9vbCBuYW1lIGV4
            aXN0cyB0aGVuIGl0IGlzIGVjIHBvb2wKICAgICAgICAgICAgICAgIGlmIGtleVsiZXJhc3VyZV9j
            b2RlX3Byb2ZpbGUiXSBhbmQga2V5WyJwb29sX25hbWUiXSA9PSByYmRfcG9vbF9uYW1lOgogICAg
            ICAgICAgICAgICAgICAgIHBvb2xfZXhpc3QgPSBUcnVlCgogICAgICAgICAgICBpZiBub3QgbWV0
            YWRhdGFfcG9vbF9leGlzdDoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVF
            eGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIlByb3ZpZGVkIHJiZF9lY19tZXRhZGF0YV9w
            b29sIG5hbWUsIgogICAgICAgICAgICAgICAgICAgIGYiIHtyYmRfbWV0YWRhdGFfZWNfcG9vbF9u
            YW1lfSwgZG9lcyBub3QgZXhpc3QiCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmIG5v
            dCBwb29sX2V4aXN0OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICBmIlByb3ZpZGVkIHJiZF9kYXRhX3Bvb2wgbmFtZSwg
            e3JiZF9wb29sX25hbWV9LCBkb2VzIG5vdCBleGlzdCIKICAgICAgICAgICAgICAgICkKICAgICAg
            ICAgICAgcmV0dXJuIHJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUKCiAgICBkZWYgZHJ5X3J1bihz
            ZWxmLCBtc2cpOgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAg
            ICAgcHJpbnQoIkV4ZWN1dGU6ICIgKyAiJyIgKyBtc2cgKyAiJyIpCgogICAgZGVmIHZhbGlkYXRl
            X3Jnd19lbmRwb2ludF90bHNfY2VydChzZWxmKToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd190bHNfY2VydF9wYXRoOgogICAgICAgICAgICB3aXRoIG9wZW4oc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfdGxzX2NlcnRfcGF0aCwgZW5jb2Rpbmc9InV0ZjgiKSBhcyBmOgogICAgICAgICAgICAg
            ICAgY29udGVudHMgPSBmLnJlYWQoKQogICAgICAgICAgICAgICAgcmV0dXJuIGNvbnRlbnRzLnJz
            dHJpcCgpCgogICAgZGVmIF9jaGVja19jb25mbGljdGluZ19vcHRpb25zKHNlbGYpOgogICAgICAg
            IGlmIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnVwZ3JhZGUgYW5kIG5vdCBzZWxmLl9hcmdfcGFyc2Vy
            LnJiZF9kYXRhX3Bvb2xfbmFtZToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICJFaXRoZXIgJy0tdXBncmFkZScgb3IgJy0tcmJkLWRh
            dGEtcG9vbC1uYW1lIDxwb29sX25hbWU+JyBzaG91bGQgYmUgc3BlY2lmaWVkIgogICAgICAgICAg
            ICApCgogICAgZGVmIF9pbnZhbGlkX2VuZHBvaW50KHNlbGYsIGVuZHBvaW50X3N0cik6CiAgICAg
            ICAgIyBleHRyYWN0IHRoZSBwb3J0IGJ5IGdldHRpbmcgdGhlIGxhc3Qgc3BsaXQgb24gYDpgIGRl
            bGltaXRlcgogICAgICAgIHRyeToKICAgICAgICAgICAgZW5kcG9pbnRfc3RyX2lwLCBwb3J0ID0g
            ZW5kcG9pbnRfc3RyLnJzcGxpdCgiOiIsIDEpCiAgICAgICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAg
            ICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oZiJOb3QgYSBwcm9wZXIg
            ZW5kcG9pbnQ6IHtlbmRwb2ludF9zdHJ9IikKCiAgICAgICAgdHJ5OgogICAgICAgICAgICBpZiBl
            bmRwb2ludF9zdHJfaXBbMF0gPT0gIlsiOgogICAgICAgICAgICAgICAgZW5kcG9pbnRfc3RyX2lw
            ID0gZW5kcG9pbnRfc3RyX2lwWzEgOiBsZW4oZW5kcG9pbnRfc3RyX2lwKSAtIDFdCiAgICAgICAg
            ICAgIGlwX3R5cGUgPSAoCiAgICAgICAgICAgICAgICAiSVB2NCIgaWYgdHlwZShpcF9hZGRyZXNz
            KGVuZHBvaW50X3N0cl9pcCkpIGlzIElQdjRBZGRyZXNzIGVsc2UgIklQdjYiCiAgICAgICAgICAg
            ICkKICAgICAgICBleGNlcHQgVmFsdWVFcnJvcjoKICAgICAgICAgICAgaXBfdHlwZSA9ICJGUURO
            IgogICAgICAgIGlmIG5vdCBwb3J0LmlzZGlnaXQoKToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0
            aW9uRmFpbHVyZUV4Y2VwdGlvbihmIlBvcnQgbm90IHZhbGlkOiB7cG9ydH0iKQogICAgICAgIGlu
            dFBvcnQgPSBpbnQocG9ydCkKICAgICAgICBpZiBpbnRQb3J0IDwgMSBvciBpbnRQb3J0ID4gMioq
            MTYgLSAxOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKGYiT3V0
            IG9mIHJhbmdlIHBvcnQgbnVtYmVyOiB7cG9ydH0iKQoKICAgICAgICByZXR1cm4gaXBfdHlwZQoK
            ICAgIGRlZiBlbmRwb2ludF9kaWFsKHNlbGYsIGVuZHBvaW50X3N0ciwgaXBfdHlwZSwgdGltZW91
            dD0zLCBjZXJ0PU5vbmUpOgogICAgICAgICMgaWYgdGhlICdjbHVzdGVyJyBpbnN0YW5jZSBpcyBh
            IGR1bW15IG9uZSwKICAgICAgICAjIGRvbid0IHRyeSB0byByZWFjaCBvdXQgdG8gdGhlIGVuZHBv
            aW50CiAgICAgICAgaWYgaXNpbnN0YW5jZShzZWxmLmNsdXN0ZXIsIER1bW15UmFkb3MpOgogICAg
            ICAgICAgICByZXR1cm4gIiIsICIiLCAiIgogICAgICAgIGlmIGlwX3R5cGUgPT0gIklQdjYiOgog
            ICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBlbmRwb2ludF9zdHJfaXAsIGVuZHBvaW50
            X3N0cl9wb3J0ID0gZW5kcG9pbnRfc3RyLnJzcGxpdCgiOiIsIDEpCiAgICAgICAgICAgIGV4Y2Vw
            dCBWYWx1ZUVycm9yOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICBmIk5vdCBhIHByb3BlciBlbmRwb2ludDoge2VuZHBv
            aW50X3N0cn0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmIGVuZHBvaW50X3N0cl9p
            cFswXSAhPSAiWyI6CiAgICAgICAgICAgICAgICBlbmRwb2ludF9zdHJfaXAgPSAiWyIgKyBlbmRw
            b2ludF9zdHJfaXAgKyAiXSIKICAgICAgICAgICAgZW5kcG9pbnRfc3RyID0gIjoiLmpvaW4oW2Vu
            ZHBvaW50X3N0cl9pcCwgZW5kcG9pbnRfc3RyX3BvcnRdKQoKICAgICAgICBwcm90b2NvbHMgPSBb
            Imh0dHAiLCAiaHR0cHMiXQogICAgICAgIHJlc3BvbnNlX2Vycm9yID0gTm9uZQogICAgICAgIGZv
            ciBwcmVmaXggaW4gcHJvdG9jb2xzOgogICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBl
            cCA9IGYie3ByZWZpeH06Ly97ZW5kcG9pbnRfc3RyfSIKICAgICAgICAgICAgICAgIHZlcmlmeSA9
            IE5vbmUKICAgICAgICAgICAgICAgICMgSWYgdmVyaWZ5IGlzIHNldCB0byBhIHBhdGggdG8gYSBk
            aXJlY3RvcnksCiAgICAgICAgICAgICAgICAjIHRoZSBkaXJlY3RvcnkgbXVzdCBoYXZlIGJlZW4g
            cHJvY2Vzc2VkIHVzaW5nIHRoZSBjX3JlaGFzaCB1dGlsaXR5IHN1cHBsaWVkIHdpdGggT3BlblNT
            TC4KICAgICAgICAgICAgICAgIGlmIHByZWZpeCA9PSAiaHR0cHMiIGFuZCBzZWxmLl9hcmdfcGFy
            c2VyLnJnd19za2lwX3RsczoKICAgICAgICAgICAgICAgICAgICB2ZXJpZnkgPSBGYWxzZQogICAg
            ICAgICAgICAgICAgICAgIHIgPSByZXF1ZXN0cy5oZWFkKGVwLCB0aW1lb3V0PXRpbWVvdXQsIHZl
            cmlmeT1GYWxzZSkKICAgICAgICAgICAgICAgIGVsaWYgcHJlZml4ID09ICJodHRwcyIgYW5kIGNl
            cnQ6CiAgICAgICAgICAgICAgICAgICAgdmVyaWZ5ID0gY2VydAogICAgICAgICAgICAgICAgICAg
            IHIgPSByZXF1ZXN0cy5oZWFkKGVwLCB0aW1lb3V0PXRpbWVvdXQsIHZlcmlmeT1jZXJ0KQogICAg
            ICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgICAgICByID0gcmVxdWVzdHMuaGVhZChl
            cCwgdGltZW91dD10aW1lb3V0KQogICAgICAgICAgICAgICAgaWYgci5zdGF0dXNfY29kZSA9PSAy
            MDA6CiAgICAgICAgICAgICAgICAgICAgcmV0dXJuIHByZWZpeCwgdmVyaWZ5LCAiIgogICAgICAg
            ICAgICBleGNlcHQgRXhjZXB0aW9uIGFzIGVycjoKICAgICAgICAgICAgICAgIHJlc3BvbnNlX2Vy
            cm9yID0gZXJyCiAgICAgICAgICAgICAgICBjb250aW51ZQogICAgICAgIHN5cy5zdGRlcnIud3Jp
            dGUoCiAgICAgICAgICAgIGYidW5hYmxlIHRvIGNvbm5lY3QgdG8gZW5kcG9pbnQ6IHtlbmRwb2lu
            dF9zdHJ9LCBmYWlsZWQgZXJyb3I6IHtyZXNwb25zZV9lcnJvcn0iCiAgICAgICAgKQogICAgICAg
            IHJldHVybiAoCiAgICAgICAgICAgICIiLAogICAgICAgICAgICAiIiwKICAgICAgICAgICAgKCIt
            MSIpLAogICAgICAgICkKCiAgICBkZWYgX19pbml0X18oc2VsZiwgYXJnX2xpc3Q9Tm9uZSk6CiAg
            ICAgICAgc2VsZi5vdXRfbWFwID0ge30KICAgICAgICBzZWxmLl9leGNsdWRlZF9rZXlzID0gc2V0
            KCkKICAgICAgICBzZWxmLl9hcmdfcGFyc2VyID0gc2VsZi5nZW5fYXJnX3BhcnNlcihhcmdzX3Rv
            X3BhcnNlPWFyZ19saXN0KQogICAgICAgIHNlbGYuX2NoZWNrX2NvbmZsaWN0aW5nX29wdGlvbnMo
            KQogICAgICAgIHNlbGYucnVuX2FzX3VzZXIgPSBzZWxmLl9hcmdfcGFyc2VyLnJ1bl9hc191c2Vy
            CiAgICAgICAgc2VsZi5vdXRwdXRfZmlsZSA9IHNlbGYuX2FyZ19wYXJzZXIub3V0cHV0CiAgICAg
            ICAgc2VsZi5jZXBoX2NvbmYgPSBzZWxmLl9hcmdfcGFyc2VyLmNlcGhfY29uZgogICAgICAgIHNl
            bGYuY2VwaF9rZXlyaW5nID0gc2VsZi5fYXJnX3BhcnNlci5rZXlyaW5nCiAgICAgICAgIyBpZiB1
            c2VyIG5vdCBwcm92aWRlZCwgZ2l2ZSBhIGRlZmF1bHQgdXNlcgogICAgICAgIGlmIG5vdCBzZWxm
            LnJ1bl9hc191c2VyIGFuZCBub3Qgc2VsZi5fYXJnX3BhcnNlci51cGdyYWRlOgogICAgICAgICAg
            ICBzZWxmLnJ1bl9hc191c2VyID0gc2VsZi5FWFRFUk5BTF9VU0VSX05BTUUKICAgICAgICBpZiBu
            b3Qgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXggYW5kIG5vdCBzZWxmLl9hcmdfcGFy
            c2VyLnVwZ3JhZGU6CiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4
            ID0gc2VsZi5ERUZBVUxUX1JHV19QT09MX1BSRUZJWAogICAgICAgIGlmIHNlbGYuY2VwaF9jb25m
            OgogICAgICAgICAgICBrd2FyZ3MgPSB7fQogICAgICAgICAgICBpZiBzZWxmLmNlcGhfa2V5cmlu
            ZzoKICAgICAgICAgICAgICAgIGt3YXJnc1siY29uZiJdID0geyJrZXlyaW5nIjogc2VsZi5jZXBo
            X2tleXJpbmd9CiAgICAgICAgICAgIHNlbGYuY2x1c3RlciA9IHJhZG9zLlJhZG9zKGNvbmZmaWxl
            PXNlbGYuY2VwaF9jb25mLCAqKmt3YXJncykKICAgICAgICBlbHNlOgogICAgICAgICAgICBzZWxm
            LmNsdXN0ZXIgPSByYWRvcy5SYWRvcygpCiAgICAgICAgICAgIHNlbGYuY2x1c3Rlci5jb25mX3Jl
            YWRfZmlsZSgpCiAgICAgICAgc2VsZi5jbHVzdGVyLmNvbm5lY3QoKQoKICAgIGRlZiBzaHV0ZG93
            bihzZWxmKToKICAgICAgICBpZiBzZWxmLmNsdXN0ZXIuc3RhdGUgPT0gImNvbm5lY3RlZCI6CiAg
            ICAgICAgICAgIHNlbGYuY2x1c3Rlci5zaHV0ZG93bigpCgogICAgZGVmIGdldF9mc2lkKHNlbGYp
            OgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJu
            IHNlbGYuZHJ5X3J1bigiY2VwaCBmc2lkIikKICAgICAgICByZXR1cm4gc3RyKHNlbGYuY2x1c3Rl
            ci5nZXRfZnNpZCgpKQoKICAgIGRlZiBfY29tbW9uX2NtZF9qc29uX2dlbihzZWxmLCBjbWRfanNv
            bik6CiAgICAgICAgY21kID0ganNvbi5kdW1wcyhjbWRfanNvbiwgc29ydF9rZXlzPVRydWUpCiAg
            ICAgICAgcmV0X3ZhbCwgY21kX291dCwgZXJyX21zZyA9IHNlbGYuY2x1c3Rlci5tb25fY29tbWFu
            ZChjbWQsIGIiIikKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnZlcmJvc2U6CiAgICAgICAg
            ICAgIHByaW50KGYiQ29tbWFuZCBJbnB1dDoge2NtZH0iKQogICAgICAgICAgICBwcmludCgKICAg
            ICAgICAgICAgICAgIGYiUmV0dXJuIFZhbDoge3JldF92YWx9XG5Db21tYW5kIE91dHB1dDoge2Nt
            ZF9vdXR9XG4iCiAgICAgICAgICAgICAgICBmIkVycm9yIE1lc3NhZ2U6IHtlcnJfbXNnfVxuLS0t
            LS0tLS0tLVxuIgogICAgICAgICAgICApCiAgICAgICAganNvbl9vdXQgPSB7fQogICAgICAgICMg
            aWYgdGhlcmUgaXMgbm8gZXJyb3IgKGkuZTsgcmV0X3ZhbCBpcyBaRVJPKSBhbmQgJ2NtZF9vdXQn
            IGlzIG5vdCBlbXB0eQogICAgICAgICMgdGhlbiBjb252ZXJ0ICdjbWRfb3V0JyB0byBhIGpzb24g
            b3V0cHV0CiAgICAgICAgaWYgcmV0X3ZhbCA9PSAwIGFuZCBjbWRfb3V0OgogICAgICAgICAgICBq
            c29uX291dCA9IGpzb24ubG9hZHMoY21kX291dCkKICAgICAgICByZXR1cm4gcmV0X3ZhbCwganNv
            bl9vdXQsIGVycl9tc2cKCiAgICBkZWYgZ2V0X2NlcGhfZXh0ZXJuYWxfbW9uX2RhdGEoc2VsZik6
            CiAgICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJxdW9ydW1fc3RhdHVzIiwgImZvcm1hdCI6
            ICJqc29uIn0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAg
            IHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGggIiArIGNtZF9qc29uWyJwcmVmaXgiXSkKICAgICAg
            ICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4o
            Y21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwK
            ICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgIidxdW9y
            dW1fc3RhdHVzJyBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAgICAgIGYiRXJyb3I6IHtl
            cnJfbXNnIGlmIHJldF92YWwgIT0gMCBlbHNlIHNlbGYuRU1QVFlfT1VUUFVUX0xJU1R9IgogICAg
            ICAgICAgICApCiAgICAgICAgcV9sZWFkZXJfbmFtZSA9IGpzb25fb3V0WyJxdW9ydW1fbGVhZGVy
            X25hbWUiXQogICAgICAgIHFfbGVhZGVyX2RldGFpbHMgPSB7fQogICAgICAgIHFfbGVhZGVyX21h
            dGNoaW5nX2xpc3QgPSBbCiAgICAgICAgICAgIGwgZm9yIGwgaW4ganNvbl9vdXRbIm1vbm1hcCJd
            WyJtb25zIl0gaWYgbFsibmFtZSJdID09IHFfbGVhZGVyX25hbWUKICAgICAgICBdCiAgICAgICAg
            aWYgbGVuKHFfbGVhZGVyX21hdGNoaW5nX2xpc3QpID09IDA6CiAgICAgICAgICAgIHJhaXNlIEV4
            ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oIk5vIG1hdGNoaW5nICdtb24nIGRldGFpbHMgZm91bmQi
            KQogICAgICAgIHFfbGVhZGVyX2RldGFpbHMgPSBxX2xlYWRlcl9tYXRjaGluZ19saXN0WzBdCiAg
            ICAgICAgIyBnZXQgdGhlIGFkZHJlc3MgdmVjdG9yIG9mIHRoZSBxdW9ydW0tbGVhZGVyCiAgICAg
            ICAgcV9sZWFkZXJfYWRkcnZlYyA9IHFfbGVhZGVyX2RldGFpbHMuZ2V0KCJwdWJsaWNfYWRkcnMi
            LCB7fSkuZ2V0KCJhZGRydmVjIiwgW10pCiAgICAgICAgaXBfYWRkciA9IHN0cihxX2xlYWRlcl9k
            ZXRhaWxzWyJwdWJsaWNfYWRkciJdLnNwbGl0KCIvIilbMF0pCgogICAgICAgIGlmIHNlbGYuX2Fy
            Z19wYXJzZXIudjJfcG9ydF9lbmFibGU6CiAgICAgICAgICAgIGlmIHFfbGVhZGVyX2FkZHJ2ZWNb
            MF1bInR5cGUiXSA9PSAidjIiOgogICAgICAgICAgICAgICAgaXBfYWRkciA9IHFfbGVhZGVyX2Fk
            ZHJ2ZWNbMF1bImFkZHIiXQogICAgICAgICAgICBlbGlmIGxlbihxX2xlYWRlcl9hZGRydmVjKSA+
            IDEgYW5kIHFfbGVhZGVyX2FkZHJ2ZWNbMV1bInR5cGUiXSA9PSAidjIiOgogICAgICAgICAgICAg
            ICAgaXBfYWRkciA9IHFfbGVhZGVyX2FkZHJ2ZWNbMV1bImFkZHIiXQogICAgICAgICAgICBlbHNl
            OgogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZSgKICAgICAgICAgICAgICAgICAgICAi
            J3YyJyBhZGRyZXNzIHR5cGUgbm90IHByZXNlbnQsIGFuZCAndjItcG9ydC1lbmFibGUnIGZsYWcg
            aXMgcHJvdmlkZWQiCiAgICAgICAgICAgICAgICApCgogICAgICAgIHJldHVybiBmIntzdHIocV9s
            ZWFkZXJfbmFtZSl9PXtpcF9hZGRyfSIKCiAgICBkZWYgX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAo
            c2VsZiwgaG9zdF9uYW1lLCBwb3J0LCBpcF90eXBlKToKICAgICAgICAjIGlmICdjbHVzdGVyJyBp
            bnN0YW5jZSBpcyBhIGR1bW15IHR5cGUsCiAgICAgICAgIyBjYWxsIHRoZSBkdW1teSBpbnN0YW5j
            ZSdzICJjb252ZXJ0IiBtZXRob2QKICAgICAgICBpZiBub3QgaG9zdF9uYW1lOgogICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKCJFbXB0eSBob3N0bmFtZSBwcm92aWRl
            ZCIpCiAgICAgICAgaWYgaXNpbnN0YW5jZShzZWxmLmNsdXN0ZXIsIER1bW15UmFkb3MpOgogICAg
            ICAgICAgICByZXR1cm4gc2VsZi5jbHVzdGVyLl9jb252ZXJ0X2hvc3RuYW1lX3RvX2lwKGhvc3Rf
            bmFtZSkKCiAgICAgICAgaWYgaXBfdHlwZSA9PSAiRlFETiI6CiAgICAgICAgICAgICMgY2hlY2sg
            d2hpY2ggaXAgRlFETiBzaG91bGQgYmUgY29udmVydGVkIHRvLCBJUHY0IG9yIElQdjYKICAgICAg
            ICAgICAgIyBjaGVjayB0aGUgaG9zdCBpcCwgdGhlIGVuZHBvaW50IGlwIHR5cGUgd291bGQgYmUg
            c2ltaWxhciB0byBob3N0IGlwCiAgICAgICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAib3Jj
            aCBob3N0IGxzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICAgICAgcmV0X3ZhbCwganNvbl9v
            dXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAg
            ICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAogICAgICAgICAgICBp
            ZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICAgICAgcmFp
            c2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJ29yY2gg
            aG9zdCBscycgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAgICAgICAgICAgICAgICAgZiJFcnJvcjog
            e2Vycl9tc2cgaWYgcmV0X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAg
            ICAgICAgICAgICAgICApCiAgICAgICAgICAgIGhvc3RfYWRkciA9IGpzb25fb3V0WzBdWyJhZGRy
            Il0KICAgICAgICAgICAgIyBhZGQgOjgwIHNhbXBsZSBwb3J0IGluIGlwX3R5cGUsIGFzIF9pbnZh
            bGlkX2VuZHBvaW50IGFsc28gdmVyaWZ5IHBvcnQKICAgICAgICAgICAgaG9zdF9pcF90eXBlID0g
            c2VsZi5faW52YWxpZF9lbmRwb2ludChob3N0X2FkZHIgKyAiOjgwIikKICAgICAgICAgICAgaW1w
            b3J0IHNvY2tldAoKICAgICAgICAgICAgIyBleGFtcGxlIG91dHB1dCBbKDxBZGRyZXNzRmFtaWx5
            LkFGX0lORVQ6IDI+LCA8U29ja2V0S2luZC5TT0NLX1NUUkVBTTogMT4sIDYsICcnLCAoJzkzLjE4
            NC4yMTYuMzQnLCA4MCkpLCAuLi5dCiAgICAgICAgICAgICMgd2UgbmVlZCB0byBnZXQgOTMuMTg0
            LjIxNi4zNCBzbyBpdCB3b3VsZCBiZSBpcFswXVs0XVswXQogICAgICAgICAgICBpZiBob3N0X2lw
            X3R5cGUgPT0gIklQdjYiOgogICAgICAgICAgICAgICAgaXAgPSBzb2NrZXQuZ2V0YWRkcmluZm8o
            CiAgICAgICAgICAgICAgICAgICAgaG9zdF9uYW1lLCBwb3J0LCBmYW1pbHk9c29ja2V0LkFGX0lO
            RVQ2LCBwcm90bz1zb2NrZXQuSVBQUk9UT19UQ1AKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgZWxpZiBob3N0X2lwX3R5cGUgPT0gIklQdjQiOgogICAgICAgICAgICAgICAgaXAgPSBzb2Nr
            ZXQuZ2V0YWRkcmluZm8oCiAgICAgICAgICAgICAgICAgICAgaG9zdF9uYW1lLCBwb3J0LCBmYW1p
            bHk9c29ja2V0LkFGX0lORVQsIHByb3RvPXNvY2tldC5JUFBST1RPX1RDUAogICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICBkZWwgc29ja2V0CiAgICAgICAgICAgIHJldHVybiBpcFswXVs0XVsw
            XQogICAgICAgIHJldHVybiBob3N0X25hbWUKCiAgICBkZWYgZ2V0X2FjdGl2ZV9hbmRfc3RhbmRi
            eV9tZ3JzKHNlbGYpOgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAg
            ICAgICAgcmV0dXJuICIiLCBzZWxmLmRyeV9ydW4oImNlcGggc3RhdHVzIikKICAgICAgICBtb25p
            dG9yaW5nX2VuZHBvaW50X3BvcnQgPSBzZWxmLl9hcmdfcGFyc2VyLm1vbml0b3JpbmdfZW5kcG9p
            bnRfcG9ydAogICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdCA9IHNlbGYuX2FyZ19w
            YXJzZXIubW9uaXRvcmluZ19lbmRwb2ludAogICAgICAgIHN0YW5kYnlfbWdycyA9IFtdCiAgICAg
            ICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdDoKICAgICAgICAgICAgY21kX2pz
            b24gPSB7InByZWZpeCI6ICJzdGF0dXMiLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgICAgICBy
            ZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21k
            X2pzb24pCiAgICAgICAgICAgICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQs
            CiAgICAgICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAgICAg
            ICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAg
            ICAgICAgICInbWdyIHNlcnZpY2VzJyBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAgICAg
            ICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09V
            VFBVVF9MSVNUfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRw
            b2ludCA9ICgKICAgICAgICAgICAgICAgIGpzb25fb3V0LmdldCgibWdybWFwIiwge30pLmdldCgi
            c2VydmljZXMiLCB7fSkuZ2V0KCJwcm9tZXRoZXVzIiwgIiIpCiAgICAgICAgICAgICkKICAgICAg
            ICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnQ6CiAgICAgICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJjYW4ndCBmaW5k
            IG1vbml0b3JpbmdfZW5kcG9pbnQsIHByb21ldGhldXMgbW9kdWxlIG1pZ2h0IG5vdCBiZSBlbmFi
            bGVkLCAiCiAgICAgICAgICAgICAgICAgICAgImVuYWJsZSB0aGUgbW9kdWxlIGJ5IHJ1bm5pbmcg
            J2NlcGggbWdyIG1vZHVsZSBlbmFibGUgcHJvbWV0aGV1cyciCiAgICAgICAgICAgICAgICApCiAg
            ICAgICAgICAgICMgbm93IGNoZWNrIHRoZSBzdGFuZC1ieSBtZ3ItcwogICAgICAgICAgICBzdGFu
            ZGJ5X2FyciA9IGpzb25fb3V0LmdldCgibWdybWFwIiwge30pLmdldCgic3RhbmRieXMiLCBbXSkK
            ICAgICAgICAgICAgZm9yIGVhY2hfc3RhbmRieSBpbiBzdGFuZGJ5X2FycjoKICAgICAgICAgICAg
            ICAgIGlmICJuYW1lIiBpbiBlYWNoX3N0YW5kYnkua2V5cygpOgogICAgICAgICAgICAgICAgICAg
            IHN0YW5kYnlfbWdycy5hcHBlbmQoZWFjaF9zdGFuZGJ5WyJuYW1lIl0pCiAgICAgICAgICAgIHRy
            eToKICAgICAgICAgICAgICAgIHBhcnNlZF9lbmRwb2ludCA9IHVybHBhcnNlKG1vbml0b3Jpbmdf
            ZW5kcG9pbnQpCiAgICAgICAgICAgIGV4Y2VwdCBWYWx1ZUVycm9yOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBmImlu
            dmFsaWQgZW5kcG9pbnQ6IHttb25pdG9yaW5nX2VuZHBvaW50fSIKICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0ID0gcGFyc2VkX2VuZHBvaW50
            Lmhvc3RuYW1lCiAgICAgICAgICAgIGlmIG5vdCBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQ6CiAg
            ICAgICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQgPSBzdHIocGFyc2VkX2VuZHBv
            aW50LnBvcnQpCgogICAgICAgICMgaWYgbW9uaXRvcmluZyBlbmRwb2ludCBwb3J0IGlzIG5vdCBz
            ZXQsIHB1dCBhIGRlZmF1bHQgbW9uIHBvcnQKICAgICAgICBpZiBub3QgbW9uaXRvcmluZ19lbmRw
            b2ludF9wb3J0OgogICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQgPSBzZWxmLkRF
            RkFVTFRfTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUCgogICAgICAgICMgdXNlciBjb3VsZCBnaXZl
            IGNvbW1hIGFuZCBzcGFjZSBzZXBhcmF0ZWQgaW5wdXRzIChsaWtlIC0tbW9uaXRvcmluZy1lbmRw
            b2ludD0iPGlwMT4sIDxpcDI+IikKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Qg
            PSBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3QucmVwbGFjZSgiLCIsICIgIikKICAgICAgICBt
            b25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXQgPSBtb25pdG9yaW5nX2VuZHBvaW50X2lw
            X2xpc3Quc3BsaXQoKQogICAgICAgICMgaWYgbW9uaXRvcmluZy1lbmRwb2ludCBjb3VsZCBub3Qg
            YmUgZm91bmQsIHJhaXNlIGFuIGVycm9yCiAgICAgICAgaWYgbGVuKG1vbml0b3JpbmdfZW5kcG9p
            bnRfaXBfbGlzdF9zcGxpdCkgPT0gMDoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVy
            ZUV4Y2VwdGlvbigiTm8gJ21vbml0b3JpbmctZW5kcG9pbnQnIGZvdW5kIikKICAgICAgICAjIGZp
            cnN0IGlwIGlzIHRyZWF0ZWQgYXMgdGhlIG1haW4gbW9uaXRvcmluZy1lbmRwb2ludAogICAgICAg
            IG1vbml0b3JpbmdfZW5kcG9pbnRfaXAgPSBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3Bs
            aXRbMF0KICAgICAgICAjIHJlc3Qgb2YgdGhlIGlwLXMgYXJlIGFkZGVkIHRvIHRoZSAnc3RhbmRi
            eV9tZ3JzJyBsaXN0CiAgICAgICAgc3RhbmRieV9tZ3JzLmV4dGVuZChtb25pdG9yaW5nX2VuZHBv
            aW50X2lwX2xpc3Rfc3BsaXRbMTpdKQogICAgICAgIGZhaWxlZF9pcCA9IG1vbml0b3JpbmdfZW5k
            cG9pbnRfaXAKCiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludCA9ICI6Ii5qb2luKAogICAgICAg
            ICAgICBbbW9uaXRvcmluZ19lbmRwb2ludF9pcCwgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0XQog
            ICAgICAgICkKICAgICAgICBpcF90eXBlID0gc2VsZi5faW52YWxpZF9lbmRwb2ludChtb25pdG9y
            aW5nX2VuZHBvaW50KQogICAgICAgIHRyeToKICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2lu
            dF9pcCA9IHNlbGYuX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoCiAgICAgICAgICAgICAgICBtb25p
            dG9yaW5nX2VuZHBvaW50X2lwLCBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQsIGlwX3R5cGUKICAg
            ICAgICAgICAgKQogICAgICAgICAgICAjIGNvbGxlY3QgYWxsIHRoZSAnc3RhbmQtYnknIG1nciBp
            cHMKICAgICAgICAgICAgbWdyX2lwcyA9IFtdCiAgICAgICAgICAgIGZvciBlYWNoX3N0YW5kYnlf
            bWdyIGluIHN0YW5kYnlfbWdyczoKICAgICAgICAgICAgICAgIGZhaWxlZF9pcCA9IGVhY2hfc3Rh
            bmRieV9tZ3IKICAgICAgICAgICAgICAgIG1ncl9pcHMuYXBwZW5kKAogICAgICAgICAgICAgICAg
            ICAgIHNlbGYuX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoCiAgICAgICAgICAgICAgICAgICAgICAg
            IGVhY2hfc3RhbmRieV9tZ3IsIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydCwgaXBfdHlwZQogICAg
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgICkKICAgICAgICBleGNlcHQ6CiAgICAg
            ICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBm
            IkNvbnZlcnNpb24gb2YgaG9zdDoge2ZhaWxlZF9pcH0gdG8gSVAgZmFpbGVkLiAiCiAgICAgICAg
            ICAgICAgICAiUGxlYXNlIGVudGVyIHRoZSBJUCBhZGRyZXNzZXMgb2YgYWxsIHRoZSBjZXBoLW1n
            cnMgd2l0aCB0aGUgJy0tbW9uaXRvcmluZy1lbmRwb2ludCcgZmxhZyIKICAgICAgICAgICAgKQoK
            ICAgICAgICBfLCBfLCBlcnIgPSBzZWxmLmVuZHBvaW50X2RpYWwobW9uaXRvcmluZ19lbmRwb2lu
            dCwgaXBfdHlwZSkKICAgICAgICBpZiBlcnIgPT0gIi0xIjoKICAgICAgICAgICAgcmFpc2UgRXhl
            Y3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihlcnIpCiAgICAgICAgIyBhZGQgdGhlIHZhbGlkYXRlZCBh
            Y3RpdmUgbWdyIElQIGludG8gdGhlIGZpcnN0IGluZGV4CiAgICAgICAgbWdyX2lwcy5pbnNlcnQo
            MCwgbW9uaXRvcmluZ19lbmRwb2ludF9pcCkKICAgICAgICBhbGxfbWdyX2lwc19zdHIgPSAiLCIu
            am9pbihtZ3JfaXBzKQogICAgICAgIHJldHVybiBhbGxfbWdyX2lwc19zdHIsIG1vbml0b3Jpbmdf
            ZW5kcG9pbnRfcG9ydAoKICAgIGRlZiBjaGVja191c2VyX2V4aXN0KHNlbGYsIHVzZXIpOgogICAg
            ICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAiYXV0aCBnZXQiLCAiZW50aXR5IjogZiJ7dXNlcn0i
            LCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBfID0gc2VsZi5f
            Y29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3Ig
            bGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICByZXR1cm4gIiIKICAgICAgICByZXR1cm4g
            c3RyKGpzb25fb3V0WzBdWyJrZXkiXSkKCiAgICBkZWYgZ2V0X2NlcGhmc19wcm92aXNpb25lcl9j
            YXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gImNsaWVudC5jc2ktY2VwaGZz
            LXByb3Zpc2lvbmVyIgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAgICJtb24iOiAiYWxsb3cg
            ciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLAogICAgICAgICAgICAibWdyIjogImFs
            bG93IHJ3IiwKICAgICAgICAgICAgIm9zZCI6ICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRh
            PSoiLAogICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0
            aF9wZXJtaXNzaW9uOgogICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3Bh
            cnNlci5rOHNfY2x1c3Rlcl9uYW1lCiAgICAgICAgICAgIGlmIGs4c19jbHVzdGVyX25hbWUgPT0g
            IiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAg
            ICAgICAgICAgICAgICAgICJrOHNfY2x1c3Rlcl9uYW1lIG5vdCBmb3VuZCwgcGxlYXNlIHNldCB0
            aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgY2VwaGZzX2ZpbGVzeXN0ZW0gPSBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVt
            X25hbWUKICAgICAgICAgICAgaWYgY2VwaGZzX2ZpbGVzeXN0ZW0gPT0gIiI6CiAgICAgICAgICAg
            ICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9uYW1lfSIKICAgICAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0te2s4c19jbHVzdGVyX25h
            bWV9LXtjZXBoZnNfZmlsZXN5c3RlbX0iCiAgICAgICAgICAgICAgICBjYXBzWyJvc2QiXSA9IGYi
            YWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT17Y2VwaGZzX2ZpbGVzeXN0ZW19IgoKICAgICAg
            ICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVmIGdldF9jZXBoZnNfbm9kZV9jYXBzX2FuZF9l
            bnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiCiAg
            ICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJhbGxvdyByLCBhbGxvdyBjb21tYW5k
            ICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJtZ3IiOiAiYWxsb3cgcnciLAogICAgICAg
            ICAgICAib3NkIjogImFsbG93IHJ3IHRhZyBjZXBoZnMgKj0qIiwKICAgICAgICAgICAgIm1kcyI6
            ICJhbGxvdyBydyIsCiAgICAgICAgfQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJp
            Y3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAgIGs4c19jbHVzdGVyX25hbWUgPSBzZWxm
            Ll9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAgICAgICAgICAgaWYgazhzX2NsdXN0ZXJf
            bmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRp
            b24oCiAgICAgICAgICAgICAgICAgICAgIms4c19jbHVzdGVyX25hbWUgbm90IGZvdW5kLCBwbGVh
            c2Ugc2V0IHRoZSAnLS1rOHMtY2x1c3Rlci1uYW1lJyBmbGFnIgogICAgICAgICAgICAgICAgKQog
            ICAgICAgICAgICBjZXBoZnNfZmlsZXN5c3RlbSA9IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2Zp
            bGVzeXN0ZW1fbmFtZQogICAgICAgICAgICBpZiBjZXBoZnNfZmlsZXN5c3RlbSA9PSAiIjoKICAg
            ICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0te2s4c19jbHVzdGVyX25hbWV9IgogICAg
            ICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17azhzX2Ns
            dXN0ZXJfbmFtZX0te2NlcGhmc19maWxlc3lzdGVtfSIKICAgICAgICAgICAgICAgIGNhcHNbIm9z
            ZCJdID0gZiJhbGxvdyBydyB0YWcgY2VwaGZzICo9e2NlcGhmc19maWxlc3lzdGVtfSIKCiAgICAg
            ICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfZW50aXR5KAogICAgICAgIHNlbGYs
            CiAgICAgICAgZW50aXR5LAogICAgICAgIHJiZF9wb29sX25hbWUsCiAgICAgICAgYWxpYXNfcmJk
            X3Bvb2xfbmFtZSwKICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lLAogICAgICAgIHJhZG9zX25hbWVz
            cGFjZSwKICAgICk6CiAgICAgICAgaWYgKAogICAgICAgICAgICByYmRfcG9vbF9uYW1lLmNvdW50
            KCIuIikgIT0gMAogICAgICAgICAgICBvciByYmRfcG9vbF9uYW1lLmNvdW50KCJfIikgIT0gMAog
            ICAgICAgICAgICBvciBhbGlhc19yYmRfcG9vbF9uYW1lICE9ICIiCiAgICAgICAgICAgICMgY2hl
            Y2tpbmcgYWxpYXNfcmJkX3Bvb2xfbmFtZSBpcyBub3QgZW1wdHkgYXMgdGhlcmUgbWF5YmUgYSBz
            cGVjaWFsIGNoYXJhY3RlciB1c2VkIG90aGVyIHRoYW4gLiBvciBfCiAgICAgICAgKToKICAgICAg
            ICAgICAgaWYgYWxpYXNfcmJkX3Bvb2xfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgInBsZWFzZSBz
            ZXQgdGhlICctLWFsaWFzLXJiZC1kYXRhLXBvb2wtbmFtZScgZmxhZyBhcyB0aGUgcmJkIGRhdGEg
            cG9vbCBuYW1lIGNvbnRhaW5zICcuJyBvciAnXyciCiAgICAgICAgICAgICAgICApCiAgICAgICAg
            ICAgIGlmICgKICAgICAgICAgICAgICAgIGFsaWFzX3JiZF9wb29sX25hbWUuY291bnQoIi4iKSAh
            PSAwCiAgICAgICAgICAgICAgICBvciBhbGlhc19yYmRfcG9vbF9uYW1lLmNvdW50KCJfIikgIT0g
            MAogICAgICAgICAgICApOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJy0tYWxpYXMtcmJkLWRhdGEtcG9vbC1uYW1l
            JyBmbGFnIHZhbHVlIHNob3VsZCBub3QgY29udGFpbiAnLicgb3IgJ18nIgogICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9uYW1lfS17
            YWxpYXNfcmJkX3Bvb2xfbmFtZX0iCiAgICAgICAgZWxzZToKICAgICAgICAgICAgZW50aXR5ID0g
            ZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0te3JiZF9wb29sX25hbWV9IgoKICAgICAgICBp
            ZiByYWRvc19uYW1lc3BhY2U6CiAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0te3JhZG9z
            X25hbWVzcGFjZX0iCiAgICAgICAgcmV0dXJuIGVudGl0eQoKICAgIGRlZiBnZXRfcmJkX3Byb3Zp
            c2lvbmVyX2NhcHNfYW5kX2VudGl0eShzZWxmKToKICAgICAgICBlbnRpdHkgPSAiY2xpZW50LmNz
            aS1yYmQtcHJvdmlzaW9uZXIiCiAgICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJw
            cm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLAogICAgICAgICAgICAi
            bWdyIjogImFsbG93IHJ3IiwKICAgICAgICAgICAgIm9zZCI6ICJwcm9maWxlIHJiZCIsCiAgICAg
            ICAgfQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Np
            b246CiAgICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRh
            X3Bvb2xfbmFtZQogICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3Bh
            cnNlci5hbGlhc19yYmRfZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFt
            ZSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQogICAgICAgICAgICByYWRvc19u
            YW1lc3BhY2UgPSBzZWxmLl9hcmdfcGFyc2VyLnJhZG9zX25hbWVzcGFjZQogICAgICAgICAgICBp
            ZiByYmRfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAibWFuZGF0b3J5IGZsYWcgbm90IGZv
            dW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1yYmQtZGF0YS1wb29sLW5hbWUnIGZsYWciCiAgICAgICAg
            ICAgICAgICApCiAgICAgICAgICAgIGlmIGs4c19jbHVzdGVyX25hbWUgPT0gIiI6CiAgICAgICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLWs4cy1jbHVz
            dGVyLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVudGl0eSA9IHNl
            bGYuZ2V0X2VudGl0eSgKICAgICAgICAgICAgICAgIGVudGl0eSwKICAgICAgICAgICAgICAgIHJi
            ZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lLAogICAgICAg
            ICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAgICAgICAgICAgIHJhZG9zX25hbWVzcGFj
            ZSwKICAgICAgICAgICAgKQogICAgICAgICAgICBpZiByYWRvc19uYW1lc3BhY2UgIT0gIiI6CiAg
            ICAgICAgICAgICAgICBjYXBzWyJvc2QiXSA9ICgKICAgICAgICAgICAgICAgICAgICBmInByb2Zp
            bGUgcmJkIHBvb2w9e3JiZF9wb29sX25hbWV9IG5hbWVzcGFjZT17cmFkb3NfbmFtZXNwYWNlfSIK
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGNhcHNb
            Im9zZCJdID0gZiJwcm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9uYW1lfSIKCiAgICAgICAgcmV0
            dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfcmJkX25vZGVfY2Fwc19hbmRfZW50aXR5KHNl
            bGYpOgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3NpLXJiZC1ub2RlIgogICAgICAgIGNhcHMg
            PSB7CiAgICAgICAgICAgICJtb24iOiAicHJvZmlsZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBi
            bG9ja2xpc3QnIiwKICAgICAgICAgICAgIm9zZCI6ICJwcm9maWxlIHJiZCIsCiAgICAgICAgfQog
            ICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb246CiAg
            ICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xf
            bmFtZQogICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5h
            bGlhc19yYmRfZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNl
            bGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQogICAgICAgICAgICByYWRvc19uYW1lc3Bh
            Y2UgPSBzZWxmLl9hcmdfcGFyc2VyLnJhZG9zX25hbWVzcGFjZQogICAgICAgICAgICBpZiByYmRf
            cG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAibWFuZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBw
            bGVhc2Ugc2V0IHRoZSAnLS1yYmQtZGF0YS1wb29sLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAg
            ICApCiAgICAgICAgICAgIGlmIGs4c19jbHVzdGVyX25hbWUgPT0gIiI6CiAgICAgICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJt
            YW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLWs4cy1jbHVzdGVyLW5h
            bWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVudGl0eSA9IHNlbGYuZ2V0
            X2VudGl0eSgKICAgICAgICAgICAgICAgIGVudGl0eSwKICAgICAgICAgICAgICAgIHJiZF9wb29s
            X25hbWUsCiAgICAgICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lLAogICAgICAgICAgICAg
            ICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAgICAgICAgICAgIHJhZG9zX25hbWVzcGFjZSwKICAg
            ICAgICAgICAgKQogICAgICAgICAgICBpZiByYWRvc19uYW1lc3BhY2UgIT0gIiI6CiAgICAgICAg
            ICAgICAgICBjYXBzWyJvc2QiXSA9ICgKICAgICAgICAgICAgICAgICAgICBmInByb2ZpbGUgcmJk
            IHBvb2w9e3JiZF9wb29sX25hbWV9IG5hbWVzcGFjZT17cmFkb3NfbmFtZXNwYWNlfSIKICAgICAg
            ICAgICAgICAgICkKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJd
            ID0gZiJwcm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9uYW1lfSIKCiAgICAgICAgcmV0dXJuIGNh
            cHMsIGVudGl0eQoKICAgIGRlZiBnZXRfZGVmYXVsdFVzZXJfY2Fwc19hbmRfZW50aXR5KHNlbGYp
            OgogICAgICAgIGVudGl0eSA9IHNlbGYucnVuX2FzX3VzZXIKICAgICAgICBjYXBzID0gewogICAg
            ICAgICAgICAibW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxs
            b3cgY29tbWFuZCB2ZXJzaW9uIiwKICAgICAgICAgICAgIm1nciI6ICJhbGxvdyBjb21tYW5kIGNv
            bmZpZyIsCiAgICAgICAgICAgICJvc2QiOiBmInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cg
            cnd4IHBvb2w9e3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cubWV0YSwgYWxs
            b3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD17c2VsZi5fYXJnX3BhcnNlci5yZ3df
            cG9vbF9wcmVmaXh9LnJndy5jb250cm9sLCBhbGxvdyByeCBwb29sPXtzZWxmLl9hcmdfcGFyc2Vy
            LnJnd19wb29sX3ByZWZpeH0ucmd3LmxvZywgYWxsb3cgeCBwb29sPXtzZWxmLl9hcmdfcGFyc2Vy
            LnJnd19wb29sX3ByZWZpeH0ucmd3LmJ1Y2tldHMuaW5kZXgiLAogICAgICAgIH0KCiAgICAgICAg
            cmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfY2Fwc19hbmRfZW50aXR5KHNlbGYsIHVz
            ZXJfbmFtZSk6CiAgICAgICAgaWYgImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiBpbiB1
            c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIg
            IT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVk
            X2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZ2V0X2NlcGhm
            c19wcm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoKQogICAgICAgIGlmICJjbGllbnQuY3NpLWNl
            cGhmcy1ub2RlIiBpbiB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQuY3NpLWNlcGhm
            cy1ub2RlIiAhPSB1c2VyX25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJl
            c3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5n
            ZXRfY2VwaGZzX25vZGVfY2Fwc19hbmRfZW50aXR5KCkKICAgICAgICBpZiAiY2xpZW50LmNzaS1y
            YmQtcHJvdmlzaW9uZXIiIGluIHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVudC5jc2kt
            cmJkLXByb3Zpc2lvbmVyIiAhPSB1c2VyX25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdf
            cGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1
            cm4gc2VsZi5nZXRfcmJkX3Byb3Zpc2lvbmVyX2NhcHNfYW5kX2VudGl0eSgpCiAgICAgICAgaWYg
            ImNsaWVudC5jc2ktcmJkLW5vZGUiIGluIHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVu
            dC5jc2ktcmJkLW5vZGUiICE9IHVzZXJfbmFtZToKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19w
            YXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVlCiAgICAgICAgICAgIHJldHVy
            biBzZWxmLmdldF9yYmRfbm9kZV9jYXBzX2FuZF9lbnRpdHkoKQogICAgICAgIGlmICJjbGllbnQu
            aGVhbHRoY2hlY2tlciIgaW4gdXNlcl9uYW1lOgogICAgICAgICAgICBpZiAiY2xpZW50LmhlYWx0
            aGNoZWNrZXIiICE9IHVzZXJfbmFtZToKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIu
            cmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVlCiAgICAgICAgICAgIHJldHVybiBzZWxm
            LmdldF9kZWZhdWx0VXNlcl9jYXBzX2FuZF9lbnRpdHkoKQoKICAgICAgICByYWlzZSBFeGVjdXRp
            b25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICBmIm5vIHVzZXIgZm91bmQgd2l0aCB1c2Vy
            X25hbWU6IHt1c2VyX25hbWV9LCAiCiAgICAgICAgICAgICJnZXRfY2Fwc19hbmRfZW50aXR5IGNv
            bW1hbmQgZmFpbGVkLlxuIgogICAgICAgICkKCiAgICBkZWYgY3JlYXRlX2NlcGhDU0lLZXlyaW5n
            X3VzZXIoc2VsZiwgdXNlcik6CiAgICAgICAgIiIiCiAgICAgICAgY29tbWFuZDogY2VwaCBhdXRo
            IGdldC1vci1jcmVhdGUgY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIgbW9uICdhbGxvdyBy
            JyBtZ3IgJ2FsbG93IHJ3JyBvc2QgJ2FsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KicKICAg
            ICAgICAiIiIKICAgICAgICBjYXBzLCBlbnRpdHkgPSBzZWxmLmdldF9jYXBzX2FuZF9lbnRpdHko
            dXNlcikKICAgICAgICBjbWRfanNvbiA9IHsKICAgICAgICAgICAgInByZWZpeCI6ICJhdXRoIGdl
            dC1vci1jcmVhdGUiLAogICAgICAgICAgICAiZW50aXR5IjogZW50aXR5LAogICAgICAgICAgICAi
            Y2FwcyI6IFtjYXAgZm9yIGNhcF9saXN0IGluIGxpc3QoY2Fwcy5pdGVtcygpKSBmb3IgY2FwIGlu
            IGNhcF9saXN0XSwKICAgICAgICAgICAgImZvcm1hdCI6ICJqc29uIiwKICAgICAgICB9CgogICAg
            ICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuICgKICAg
            ICAgICAgICAgICAgIHNlbGYuZHJ5X3J1bigKICAgICAgICAgICAgICAgICAgICAiY2VwaCAiCiAg
            ICAgICAgICAgICAgICAgICAgKyBjbWRfanNvblsicHJlZml4Il0KICAgICAgICAgICAgICAgICAg
            ICArICIgIgogICAgICAgICAgICAgICAgICAgICsgY21kX2pzb25bImVudGl0eSJdCiAgICAgICAg
            ICAgICAgICAgICAgKyAiICIKICAgICAgICAgICAgICAgICAgICArICIgIi5qb2luKGNtZF9qc29u
            WyJjYXBzIl0pCiAgICAgICAgICAgICAgICApLAogICAgICAgICAgICAgICAgIiIsCiAgICAgICAg
            ICAgICkKICAgICAgICAjIGNoZWNrIGlmIHVzZXIgYWxyZWFkeSBleGlzdAogICAgICAgIHVzZXJf
            a2V5ID0gc2VsZi5jaGVja191c2VyX2V4aXN0KGVudGl0eSkKICAgICAgICBpZiB1c2VyX2tleSAh
            PSAiIjoKICAgICAgICAgICAgcmV0dXJuIHVzZXJfa2V5LCBmIntlbnRpdHkuc3BsaXQoJy4nLCAx
            KVsxXX0iCiAgICAgICAgICAgICMgZW50aXR5LnNwbGl0KCcuJywxKVsxXSB0byByZW5hbWUgZW50
            aXR5KGNsaWVudC5jc2ktcmJkLW5vZGUpIGFzIGNzaS1yYmQtbm9kZQoKICAgICAgICByZXRfdmFs
            LCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24p
            CiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwKICAgICAgICBp
            ZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiInYXV0aCBnZXQtb3It
            Y3JlYXRlIHt1c2VyfScgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAgICAgICAgICAgICBmIkVycm9y
            OiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIK
            ICAgICAgICAgICAgKQogICAgICAgIHJldHVybiBzdHIoanNvbl9vdXRbMF1bImtleSJdKSwgZiJ7
            ZW50aXR5LnNwbGl0KCcuJywgMSlbMV19IgogICAgICAgICMgZW50aXR5LnNwbGl0KCcuJywxKVsx
            XSB0byByZW5hbWUgZW50aXR5KGNsaWVudC5jc2ktcmJkLW5vZGUpIGFzIGNzaS1yYmQtbm9kZQoK
            ICAgIGRlZiBnZXRfY2VwaGZzX2RhdGFfcG9vbF9kZXRhaWxzKHNlbGYpOgogICAgICAgIGNtZF9q
            c29uID0geyJwcmVmaXgiOiAiZnMgbHMiLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIGlmIHNl
            bGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigi
            Y2VwaCAiICsgY21kX2pzb25bInByZWZpeCJdKQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBl
            cnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICAjIGlm
            IHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LCByZXBvcnQgYW4gZXJyb3IKICAgICAg
            ICBpZiByZXRfdmFsICE9IDA6CiAgICAgICAgICAgICMgaWYgZnMgYW5kIGRhdGFfcG9vbCBhcmd1
            bWVudHMgYXJlIG5vdCBzZXQsIHNpbGVudGx5IHJldHVybgogICAgICAgICAgICBpZiAoCiAgICAg
            ICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUgPT0gIiIK
            ICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFt
            ZSA9PSAiIgogICAgICAgICAgICApOgogICAgICAgICAgICAgICAgcmV0dXJuCiAgICAgICAgICAg
            ICMgaWYgdXNlciBoYXMgcHJvdmlkZWQgYW55IG9mIHRoZQogICAgICAgICAgICAjICctLWNlcGhm
            cy1maWxlc3lzdGVtLW5hbWUnIG9yICctLWNlcGhmcy1kYXRhLXBvb2wtbmFtZScgYXJndW1lbnRz
            LAogICAgICAgICAgICAjIHJhaXNlIGFuIGV4Y2VwdGlvbiBhcyB3ZSBhcmUgdW5hYmxlIHRvIHZl
            cmlmeSB0aGUgYXJncwogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KAogICAgICAgICAgICAgICAgZiInZnMgbHMnIGNlcGggY2FsbCBmYWlsZWQgd2l0aCBlcnJvcjog
            e2Vycl9tc2d9IgogICAgICAgICAgICApCgogICAgICAgIG1hdGNoaW5nX2pzb25fb3V0ID0ge30K
            ICAgICAgICAjIGlmICctLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUnIGFyZ3VtZW50IGlzIHByb3Zp
            ZGVkLAogICAgICAgICMgY2hlY2sgd2hldGhlciB0aGUgcHJvdmlkZWQgZmlsZXN5c3RlbS1uYW1l
            IGV4aXN0cyBvciBub3QKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lz
            dGVtX25hbWU6CiAgICAgICAgICAgICMgZ2V0IHRoZSBtYXRjaGluZyBsaXN0CiAgICAgICAgICAg
            IG1hdGNoaW5nX2pzb25fb3V0X2xpc3QgPSBbCiAgICAgICAgICAgICAgICBtYXRjaGVkCiAgICAg
            ICAgICAgICAgICBmb3IgbWF0Y2hlZCBpbiBqc29uX291dAogICAgICAgICAgICAgICAgaWYgc3Ry
            KG1hdGNoZWRbIm5hbWUiXSkgPT0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9u
            YW1lCiAgICAgICAgICAgIF0KICAgICAgICAgICAgIyB1bmFibGUgdG8gZmluZCBhIG1hdGNoaW5n
            IGZzLW5hbWUsIHJhaXNlIGFuIGVycm9yCiAgICAgICAgICAgIGlmIGxlbihtYXRjaGluZ19qc29u
            X291dF9saXN0KSA9PSAwOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBmIkZpbGVzeXN0ZW0gcHJvdmlkZWQsICd7c2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lfScsICIKICAgICAgICAgICAgICAg
            ICAgICBmImlzIG5vdCBmb3VuZCBpbiB0aGUgZnMtbGlzdDoge1tzdHIoeFsnbmFtZSddKSBmb3Ig
            eCBpbiBqc29uX291dF19IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBtYXRjaGluZ19q
            c29uX291dCA9IG1hdGNoaW5nX2pzb25fb3V0X2xpc3RbMF0KICAgICAgICAjIGlmIGNlcGhmcyBm
            aWxlc3lzdGVtIG5hbWUgaXMgbm90IHByb3ZpZGVkLAogICAgICAgICMgdHJ5IHRvIGdldCBhIGRl
            ZmF1bHQgZnMgbmFtZSBieSBkb2luZyB0aGUgZm9sbG93aW5nCiAgICAgICAgZWxzZToKICAgICAg
            ICAgICAgIyBhLiBjaGVjayBpZiB0aGVyZSBpcyBvbmx5IG9uZSBmaWxlc3lzdGVtIGlzIHByZXNl
            bnQKICAgICAgICAgICAgaWYgbGVuKGpzb25fb3V0KSA9PSAxOgogICAgICAgICAgICAgICAgbWF0
            Y2hpbmdfanNvbl9vdXQgPSBqc29uX291dFswXQogICAgICAgICAgICAjIGIuIG9yIGVsc2UsIGNo
            ZWNrIGlmIGRhdGFfcG9vbCBuYW1lIGlzIHByb3ZpZGVkCiAgICAgICAgICAgIGVsaWYgc2VsZi5f
            YXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWU6CiAgICAgICAgICAgICAgICAjIGFuZCBp
            ZiBwcmVzZW50LCBjaGVjayB3aGV0aGVyIHRoZXJlIGV4aXN0cyBhIGZzIHdoaWNoIGhhcyB0aGUg
            ZGF0YV9wb29sCiAgICAgICAgICAgICAgICBmb3IgZWFjaEogaW4ganNvbl9vdXQ6CiAgICAgICAg
            ICAgICAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUgaW4g
            ZWFjaEpbImRhdGFfcG9vbHMiXToKICAgICAgICAgICAgICAgICAgICAgICAgbWF0Y2hpbmdfanNv
            bl9vdXQgPSBlYWNoSgogICAgICAgICAgICAgICAgICAgICAgICBicmVhawogICAgICAgICAgICAg
            ICAgIyBpZiB0aGVyZSBpcyBubyBtYXRjaGluZyBmcyBleGlzdHMsIHRoYXQgbWVhbnMgcHJvdmlk
            ZWQgZGF0YV9wb29sIG5hbWUgaXMgaW52YWxpZAogICAgICAgICAgICAgICAgaWYgbm90IG1hdGNo
            aW5nX2pzb25fb3V0OgogICAgICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVF
            eGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgICAgIGYiUHJvdmlkZWQgZGF0YV9wb29sIG5h
            bWUsIHtzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZX0sIgogICAgICAgICAg
            ICAgICAgICAgICAgICAiIGRvZXMgbm90IGV4aXN0cyIKICAgICAgICAgICAgICAgICAgICApCiAg
            ICAgICAgICAgICMgYy4gaWYgbm90aGluZyBpcyBzZXQgYW5kIGNvdWxkbid0IGZpbmQgYSBkZWZh
            dWx0LAogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgIyBqdXN0IHJldHVybiBzaWxl
            bnRseQogICAgICAgICAgICAgICAgcmV0dXJuCgogICAgICAgIGlmIG1hdGNoaW5nX2pzb25fb3V0
            OgogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUgPSBz
            dHIobWF0Y2hpbmdfanNvbl9vdXRbIm5hbWUiXSkKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5jZXBoZnNfbWV0YWRhdGFfcG9vbF9uYW1lID0gc3RyKAogICAgICAgICAgICAgICAgbWF0Y2hp
            bmdfanNvbl9vdXRbIm1ldGFkYXRhX3Bvb2wiXQogICAgICAgICAgICApCgogICAgICAgIGlmIGlz
            aW5zdGFuY2UobWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXSwgbGlzdCk6CiAgICAgICAg
            ICAgICMgaWYgdGhlIHVzZXIgaGFzIGFscmVhZHkgcHJvdmlkZWQgZGF0YS1wb29sLW5hbWUsCiAg
            ICAgICAgICAgICMgdGhyb3VnaCAtLWNlcGhmcy1kYXRhLXBvb2wtbmFtZQogICAgICAgICAgICBp
            ZiBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZToKICAgICAgICAgICAgICAg
            ICMgaWYgdGhlIHByb3ZpZGVkIG5hbWUgaXMgbm90IG1hdGNoaW5nIHdpdGggdGhlIG9uZSBpbiB0
            aGUgbGlzdAogICAgICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2Fy
            Z19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgICAgICAgICAgbm90IGlu
            IG1hdGNoaW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0KICAgICAgICAgICAgICAgICk6CiAgICAg
            ICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAg
            ICAgICAgICAgICAgICAgZiJQcm92aWRlZCBkYXRhLXBvb2wtbmFtZTogJ3tzZWxmLl9hcmdfcGFy
            c2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZX0nLCAiCiAgICAgICAgICAgICAgICAgICAgICAgICJk
            b2Vzbid0IG1hdGNoIGZyb20gdGhlIGRhdGEtcG9vbHMgbGlzdDogIgogICAgICAgICAgICAgICAg
            ICAgICAgICBmIntbc3RyKHgpIGZvciB4IGluIG1hdGNoaW5nX2pzb25fb3V0WydkYXRhX3Bvb2xz
            J11dfSIKICAgICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICMgaWYgZGF0YV9wb29sIG5h
            bWUgaXMgbm90IHByb3ZpZGVkLAogICAgICAgICAgICAjIHRoZW4gdHJ5IHRvIGZpbmQgYSBkZWZh
            dWx0IGRhdGEgcG9vbCBuYW1lCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICAjIGlm
            IG5vIGRhdGFfcG9vbHMgZXhpc3QsIHNpbGVudGx5IHJldHVybgogICAgICAgICAgICAgICAgaWYg
            bGVuKG1hdGNoaW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0pID09IDA6CiAgICAgICAgICAgICAg
            ICAgICAgcmV0dXJuCiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRh
            X3Bvb2xfbmFtZSA9IHN0cigKICAgICAgICAgICAgICAgICAgICBtYXRjaGluZ19qc29uX291dFsi
            ZGF0YV9wb29scyJdWzBdCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICMgaWYgdGhlcmUg
            YXJlIG1vcmUgdGhhbiBvbmUgJ2RhdGFfcG9vbHMnIGV4aXN0LAogICAgICAgICAgICAjIHRoZW4g
            d2FybiB0aGUgdXNlciB0aGF0IHdlIGFyZSB1c2luZyB0aGUgc2VsZWN0ZWQgbmFtZQogICAgICAg
            ICAgICBpZiBsZW4obWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXSkgPiAxOgogICAgICAg
            ICAgICAgICAgcHJpbnQoCiAgICAgICAgICAgICAgICAgICAgIldBUk5JTkc6IE11bHRpcGxlIGRh
            dGEgcG9vbHMgZGV0ZWN0ZWQ6ICIKICAgICAgICAgICAgICAgICAgICBmIntbc3RyKHgpIGZvciB4
            IGluIG1hdGNoaW5nX2pzb25fb3V0WydkYXRhX3Bvb2xzJ11dfVxuIgogICAgICAgICAgICAgICAg
            ICAgIGYiVXNpbmcgdGhlIGRhdGEtcG9vbDogJ3tzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRh
            X3Bvb2xfbmFtZX0nXG4iCiAgICAgICAgICAgICAgICApCgogICAgZGVmIGNyZWF0ZV9jaGVja2Vy
            S2V5KHNlbGYsIHVzZXIpOgogICAgICAgIGNhcHMsIGVudGl0eSA9IHNlbGYuZ2V0X2NhcHNfYW5k
            X2VudGl0eSh1c2VyKQogICAgICAgIGNtZF9qc29uID0gewogICAgICAgICAgICAicHJlZml4Ijog
            ImF1dGggZ2V0LW9yLWNyZWF0ZSIsCiAgICAgICAgICAgICJlbnRpdHkiOiBlbnRpdHksCiAgICAg
            ICAgICAgICJjYXBzIjogW2NhcCBmb3IgY2FwX2xpc3QgaW4gbGlzdChjYXBzLml0ZW1zKCkpIGZv
            ciBjYXAgaW4gY2FwX2xpc3RdLAogICAgICAgICAgICAiZm9ybWF0IjogImpzb24iLAogICAgICAg
            IH0KCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1
            cm4gc2VsZi5kcnlfcnVuKAogICAgICAgICAgICAgICAgImNlcGggIgogICAgICAgICAgICAgICAg
            KyBjbWRfanNvblsicHJlZml4Il0KICAgICAgICAgICAgICAgICsgIiAiCiAgICAgICAgICAgICAg
            ICArIGNtZF9qc29uWyJlbnRpdHkiXQogICAgICAgICAgICAgICAgKyAiICIKICAgICAgICAgICAg
            ICAgICsgIiAiLmpvaW4oY21kX2pzb25bImNhcHMiXSkKICAgICAgICAgICAgKQogICAgICAgICMg
            Y2hlY2sgaWYgdXNlciBhbHJlYWR5IGV4aXN0CiAgICAgICAgdXNlcl9rZXkgPSBzZWxmLmNoZWNr
            X3VzZXJfZXhpc3QoZW50aXR5KQogICAgICAgIGlmIHVzZXJfa2V5ICE9ICIiOgogICAgICAgICAg
            ICByZXR1cm4gdXNlcl9rZXkKCiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBz
            ZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICMgaWYgdGhlcmUgaXMg
            YW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihq
            c29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgICAgIGYiJ2F1dGggZ2V0LW9yLWNyZWF0ZSB7c2VsZi5ydW5fYXNfdXNl
            cn0nIGNvbW1hbmQgZmFpbGVkXG4iCiAgICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBp
            ZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAg
            KQogICAgICAgIHJldHVybiBzdHIoanNvbl9vdXRbMF1bImtleSJdKQoKICAgIGRlZiBnZXRfY2Vw
            aF9kYXNoYm9hcmRfbGluayhzZWxmKToKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4IjogIm1n
            ciBzZXJ2aWNlcyIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNl
            ci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKCJjZXBoICIgKyBjbWRf
            anNvblsicHJlZml4Il0pCiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIF8gPSBzZWxmLl9jb21t
            b25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNj
            ZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkg
            PT0gMDoKICAgICAgICAgICAgcmV0dXJuIE5vbmUKICAgICAgICBpZiAiZGFzaGJvYXJkIiBub3Qg
            aW4ganNvbl9vdXQ6CiAgICAgICAgICAgIHJldHVybiBOb25lCiAgICAgICAgcmV0dXJuIGpzb25f
            b3V0WyJkYXNoYm9hcmQiXQoKICAgIGRlZiBjcmVhdGVfcmd3X2FkbWluX29wc191c2VyKHNlbGYp
            OgogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAgICAgICAg
            ICAidXNlciIsCiAgICAgICAgICAgICJjcmVhdGUiLAogICAgICAgICAgICAiLS11aWQiLAogICAg
            ICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9PUFNfVVNFUl9OQU1FLAogICAgICAgICAg
            ICAiLS1kaXNwbGF5LW5hbWUiLAogICAgICAgICAgICAiUm9vayBSR1cgQWRtaW4gT3BzIHVzZXIi
            LAogICAgICAgICAgICAiLS1jYXBzIiwKICAgICAgICAgICAgImJ1Y2tldHM9Kjt1c2Vycz0qO3Vz
            YWdlPXJlYWQ7bWV0YWRhdGE9cmVhZDt6b25lPXJlYWQiLAogICAgICAgICAgICAiLS1yZ3ctcmVh
            bG0iLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lLAogICAgICAg
            ICAgICAiLS1yZ3ctem9uZWdyb3VwIiwKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3df
            em9uZWdyb3VwX25hbWUsCiAgICAgICAgICAgICItLXJndy16b25lIiwKICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1lLAogICAgICAgIF0KICAgICAgICBpZiBzZWxmLl9h
            cmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGgg
            IiArICIgIi5qb2luKGNtZCkpCiAgICAgICAgdHJ5OgogICAgICAgICAgICBvdXRwdXQgPSBzdWJw
            cm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAg
            ZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yIGFzIGV4ZWNFcnI6CiAgICAgICAg
            ICAgICMgaWYgdGhlIHVzZXIgYWxyZWFkeSBleGlzdHMsIHdlIGp1c3QgcXVlcnkgaXQKICAgICAg
            ICAgICAgaWYgZXhlY0Vyci5yZXR1cm5jb2RlID09IGVycm5vLkVFWElTVDoKICAgICAgICAgICAg
            ICAgIGNtZCA9IFsKICAgICAgICAgICAgICAgICAgICAicmFkb3Nndy1hZG1pbiIsCiAgICAgICAg
            ICAgICAgICAgICAgInVzZXIiLAogICAgICAgICAgICAgICAgICAgICJpbmZvIiwKICAgICAgICAg
            ICAgICAgICAgICAiLS11aWQiLAogICAgICAgICAgICAgICAgICAgIHNlbGYuRVhURVJOQUxfUkdX
            X0FETUlOX09QU19VU0VSX05BTUUsCiAgICAgICAgICAgICAgICAgICAgIi0tcmd3LXJlYWxtIiwK
            ICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lLAogICAg
            ICAgICAgICAgICAgICAgICItLXJndy16b25lZ3JvdXAiLAogICAgICAgICAgICAgICAgICAgIHNl
            bGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lLAogICAgICAgICAgICAgICAgICAgICIt
            LXJndy16b25lIiwKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25l
            X25hbWUsCiAgICAgICAgICAgICAgICBdCiAgICAgICAgICAgICAgICB0cnk6CiAgICAgICAgICAg
            ICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3Vi
            cHJvY2Vzcy5QSVBFKQogICAgICAgICAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJv
            Y2Vzc0Vycm9yIGFzIGV4ZWNFcnI6CiAgICAgICAgICAgICAgICAgICAgZXJyX21zZyA9ICgKICAg
            ICAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5kIHtjbWR9LiBP
            dXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICAgICAgZiJDb2Rl
            OiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtleGVjRXJyLnN0ZGVycn0iCiAgICAgICAg
            ICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21z
            ZykKICAgICAgICAgICAgICAgICAgICByZXR1cm4gTm9uZSwgTm9uZSwgRmFsc2UsICItMSIKICAg
            ICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVycl9tc2cgPSAoCiAgICAgICAgICAgICAg
            ICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5kIHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJy
            Lm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICBmIkNvZGU6IHtleGVjRXJyLnJldHVybmNv
            ZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykKICAgICAgICAgICAgICAgIHJldHVybiBO
            b25lLCBOb25lLCBGYWxzZSwgIi0xIgoKICAgICAgICAjIGlmIGl0IGlzIHB5dGhvbjIsIGRvbid0
            IGNoZWNrIGZvciBjZXBoIHZlcnNpb24gZm9yIGFkZGluZyBgaW5mbz1yZWFkYCBjYXAocmd3X3Zh
            bGlkYXRpb24pCiAgICAgICAgaWYgc3lzLnZlcnNpb25faW5mby5tYWpvciA8IDM6CiAgICAgICAg
            ICAgIGpzb25vdXRwdXQgPSBqc29uLmxvYWRzKG91dHB1dCkKICAgICAgICAgICAgcmV0dXJuICgK
            ICAgICAgICAgICAgICAgIGpzb25vdXRwdXRbImtleXMiXVswXVsiYWNjZXNzX2tleSJdLAogICAg
            ICAgICAgICAgICAganNvbm91dHB1dFsia2V5cyJdWzBdWyJzZWNyZXRfa2V5Il0sCiAgICAgICAg
            ICAgICAgICBGYWxzZSwKICAgICAgICAgICAgICAgICIiLAogICAgICAgICAgICApCgogICAgICAg
            ICMgc2VwYXJhdGVseSBhZGQgaW5mbz1yZWFkIGNhcHMgZm9yIHJndy1lbmRwb2ludCBpcCB2YWxp
            ZGF0aW9uCiAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkID0gVHJ1ZQogICAgICAgIGNtZCA9IFsK
            ICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAgICAgICAgICAiY2FwcyIsCiAgICAgICAg
            ICAgICJhZGQiLAogICAgICAgICAgICAiLS11aWQiLAogICAgICAgICAgICBzZWxmLkVYVEVSTkFM
            X1JHV19BRE1JTl9PUFNfVVNFUl9OQU1FLAogICAgICAgICAgICAiLS1jYXBzIiwKICAgICAgICAg
            ICAgImluZm89cmVhZCIsCiAgICAgICAgICAgICItLXJndy1yZWFsbSIsCiAgICAgICAgICAgIHNl
            bGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUsCiAgICAgICAgICAgICItLXJndy16b25lZ3Jv
            dXAiLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSwKICAg
            ICAgICAgICAgIi0tcmd3LXpvbmUiLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196
            b25lX25hbWUsCiAgICAgICAgXQogICAgICAgIHRyeToKICAgICAgICAgICAgb3V0cHV0ID0gc3Vi
            cHJvY2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAg
            IGV4Y2VwdCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJvciBhcyBleGVjRXJyOgogICAgICAg
            ICAgICAjIGlmIHRoZSBjZXBoIHZlcnNpb24gbm90IHN1cHBvcnRlZCBmb3IgYWRkaW5nIGBpbmZv
            PXJlYWRgIGNhcChyZ3dfdmFsaWRhdGlvbikKICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAg
            ICAgImNvdWxkIG5vdCBhZGQgY2FwczogdW5hYmxlIHRvIGFkZCBjYXBzOiBpbmZvPXJlYWRcbiIK
            ICAgICAgICAgICAgICAgIGluIGV4ZWNFcnIuc3RkZXJyLmRlY29kZSgidXRmLTgiKQogICAgICAg
            ICAgICAgICAgYW5kIGV4ZWNFcnIucmV0dXJuY29kZSA9PSAyNDQKICAgICAgICAgICAgKToKICAg
            ICAgICAgICAgICAgIGluZm9fY2FwX3N1cHBvcnRlZCA9IEZhbHNlCiAgICAgICAgICAgIGVsc2U6
            CiAgICAgICAgICAgICAgICBlcnJfbXNnID0gKAogICAgICAgICAgICAgICAgICAgIGYiZmFpbGVk
            IHRvIGV4ZWN1dGUgY29tbWFuZCB7Y21kfS4gT3V0cHV0OiB7ZXhlY0Vyci5vdXRwdXR9LiAiCiAg
            ICAgICAgICAgICAgICAgICAgZiJDb2RlOiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtl
            eGVjRXJyLnN0ZGVycn0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICBzeXMuc3Rk
            ZXJyLndyaXRlKGVycl9tc2cpCiAgICAgICAgICAgICAgICByZXR1cm4gTm9uZSwgTm9uZSwgRmFs
            c2UsICItMSIKCiAgICAgICAganNvbm91dHB1dCA9IGpzb24ubG9hZHMob3V0cHV0KQogICAgICAg
            IHJldHVybiAoCiAgICAgICAgICAgIGpzb25vdXRwdXRbImtleXMiXVswXVsiYWNjZXNzX2tleSJd
            LAogICAgICAgICAgICBqc29ub3V0cHV0WyJrZXlzIl1bMF1bInNlY3JldF9rZXkiXSwKICAgICAg
            ICAgICAgaW5mb19jYXBfc3VwcG9ydGVkLAogICAgICAgICAgICAiIiwKICAgICAgICApCgogICAg
            ZGVmIHZhbGlkYXRlX3JiZF9wb29sKHNlbGYsIHBvb2xfbmFtZSk6CiAgICAgICAgaWYgbm90IHNl
            bGYuY2x1c3Rlci5wb29sX2V4aXN0cyhwb29sX25hbWUpOgogICAgICAgICAgICByYWlzZSBFeGVj
            dXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcG9v
            bCwgJ3twb29sX25hbWV9JywgZG9lcyBub3QgZXhpc3QiCiAgICAgICAgICAgICkKCiAgICBkZWYg
            aW5pdF9yYmRfcG9vbChzZWxmLCByYmRfcG9vbF9uYW1lKToKICAgICAgICBpZiBpc2luc3RhbmNl
            KHNlbGYuY2x1c3RlciwgRHVtbXlSYWRvcyk6CiAgICAgICAgICAgIHJldHVybgogICAgICAgIGlv
            Y3R4ID0gc2VsZi5jbHVzdGVyLm9wZW5faW9jdHgocmJkX3Bvb2xfbmFtZSkKICAgICAgICByYmRf
            aW5zdCA9IHJiZC5SQkQoKQogICAgICAgIHJiZF9pbnN0LnBvb2xfaW5pdChpb2N0eCwgVHJ1ZSkK
            CiAgICBkZWYgdmFsaWRhdGVfcmFkb3NfbmFtZXNwYWNlKHNlbGYpOgogICAgICAgIHJiZF9wb29s
            X25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgIHJhZG9z
            X25hbWVzcGFjZSA9IHNlbGYuX2FyZ19wYXJzZXIucmFkb3NfbmFtZXNwYWNlCiAgICAgICAgaWYg
            cmFkb3NfbmFtZXNwYWNlID09ICIiOgogICAgICAgICAgICByZXR1cm4KICAgICAgICBpZiByYWRv
            c19uYW1lc3BhY2UuaXNsb3dlcigpID09IEZhbHNlOgogICAgICAgICAgICByYWlzZSBFeGVjdXRp
            b25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmFkb3Mg
            TmFtZXNwYWNlLCAne3JhZG9zX25hbWVzcGFjZX0nLCAiCiAgICAgICAgICAgICAgICBmImNvbnRh
            aW5zIHVwcGVyIGNhc2UiCiAgICAgICAgICAgICkKICAgICAgICByYmRfaW5zdCA9IHJiZC5SQkQo
            KQogICAgICAgIGlvY3R4ID0gc2VsZi5jbHVzdGVyLm9wZW5faW9jdHgocmJkX3Bvb2xfbmFtZSkK
            ICAgICAgICBpZiByYmRfaW5zdC5uYW1lc3BhY2VfZXhpc3RzKGlvY3R4LCByYWRvc19uYW1lc3Bh
            Y2UpIGlzIEZhbHNlOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmFkb3MgTmFtZXNwYWNlLCAne3JhZG9z
            X25hbWVzcGFjZX0nLCAiCiAgICAgICAgICAgICAgICBmImlzIG5vdCBmb3VuZCBpbiB0aGUgcG9v
            bCAne3JiZF9wb29sX25hbWV9JyIKICAgICAgICAgICAgKQoKICAgIGRlZiBnZXRfb3JfY3JlYXRl
            X3N1YnZvbHVtZV9ncm91cChzZWxmLCBzdWJ2b2x1bWVfZ3JvdXAsIGNlcGhmc19maWxlc3lzdGVt
            X25hbWUpOgogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgImNlcGgiLAogICAgICAgICAgICAi
            ZnMiLAogICAgICAgICAgICAic3Vidm9sdW1lZ3JvdXAiLAogICAgICAgICAgICAiZ2V0cGF0aCIs
            CiAgICAgICAgICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgIHN1YnZvbHVt
            ZV9ncm91cCwKICAgICAgICBdCiAgICAgICAgdHJ5OgogICAgICAgICAgICBfID0gc3VicHJvY2Vz
            cy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2Vw
            dCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJvcjoKICAgICAgICAgICAgY21kID0gWwogICAg
            ICAgICAgICAgICAgImNlcGgiLAogICAgICAgICAgICAgICAgImZzIiwKICAgICAgICAgICAgICAg
            ICJzdWJ2b2x1bWVncm91cCIsCiAgICAgICAgICAgICAgICAiY3JlYXRlIiwKICAgICAgICAgICAg
            ICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgICAgICBzdWJ2b2x1bWVfZ3Jv
            dXAsCiAgICAgICAgICAgIF0KICAgICAgICAgICAgdHJ5OgogICAgICAgICAgICAgICAgXyA9IHN1
            YnByb2Nlc3MuY2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAgICAg
            ICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBm
            InN1YnZvbHVtZSBncm91cCB7c3Vidm9sdW1lX2dyb3VwfSBpcyBub3QgYWJsZSB0byBnZXQgY3Jl
            YXRlZCIKICAgICAgICAgICAgICAgICkKCiAgICBkZWYgcGluX3N1YnZvbHVtZSgKICAgICAgICBz
            ZWxmLCBzdWJ2b2x1bWVfZ3JvdXAsIGNlcGhmc19maWxlc3lzdGVtX25hbWUsIHBpbl90eXBlLCBw
            aW5fc2V0dGluZwogICAgKToKICAgICAgICBjbWQgPSBbCiAgICAgICAgICAgICJjZXBoIiwKICAg
            ICAgICAgICAgImZzIiwKICAgICAgICAgICAgInN1YnZvbHVtZWdyb3VwIiwKICAgICAgICAgICAg
            InBpbiIsCiAgICAgICAgICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgIHN1
            YnZvbHVtZV9ncm91cCwKICAgICAgICAgICAgcGluX3R5cGUsCiAgICAgICAgICAgIHBpbl9zZXR0
            aW5nLAogICAgICAgIF0KICAgICAgICB0cnk6CiAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNo
            ZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1
            YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJzdWJ2b2x1bWUgZ3JvdXAge3N1YnZv
            bHVtZV9ncm91cH0gaXMgbm90IGFibGUgdG8gZ2V0IHBpbm5lZCIKICAgICAgICAgICAgKQoKICAg
            IGRlZiBnZXRfcmd3X2ZzaWQoc2VsZiwgYmFzZV91cmwsIHZlcmlmeSk6CiAgICAgICAgYWNjZXNz
            X2tleSA9IHNlbGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX0FDQ0VTU19LRVkiXQogICAg
            ICAgIHNlY3JldF9rZXkgPSBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9TRUNSRVRf
            S0VZIl0KICAgICAgICByZ3dfZW5kcG9pbnQgPSBzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2lu
            dAogICAgICAgIGJhc2VfdXJsID0gYmFzZV91cmwgKyAiOi8vIiArIHJnd19lbmRwb2ludCArICIv
            YWRtaW4vaW5mbz8iCiAgICAgICAgcGFyYW1zID0geyJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAg
            cmVxdWVzdF91cmwgPSBiYXNlX3VybCArIHVybGVuY29kZShwYXJhbXMpCgogICAgICAgIHRyeToK
            ICAgICAgICAgICAgciA9IHJlcXVlc3RzLmdldCgKICAgICAgICAgICAgICAgIHJlcXVlc3RfdXJs
            LAogICAgICAgICAgICAgICAgYXV0aD1TM0F1dGgoYWNjZXNzX2tleSwgc2VjcmV0X2tleSwgcmd3
            X2VuZHBvaW50KSwKICAgICAgICAgICAgICAgIHZlcmlmeT12ZXJpZnksCiAgICAgICAgICAgICkK
            ICAgICAgICBleGNlcHQgcmVxdWVzdHMuZXhjZXB0aW9ucy5UaW1lb3V0OgogICAgICAgICAgICBz
            eXMuc3RkZXJyLndyaXRlKAogICAgICAgICAgICAgICAgZiJpbnZhbGlkIGVuZHBvaW50Oiwgbm90
            IGFibGUgdG8gY2FsbCBhZG1pbi1vcHMgYXBpe3Jnd19lbmRwb2ludH0iCiAgICAgICAgICAgICkK
            ICAgICAgICAgICAgcmV0dXJuICIiLCAiLTEiCiAgICAgICAgcjEgPSByLmpzb24oKQogICAgICAg
            IGlmIHIxIGlzIE5vbmUgb3IgcjEuZ2V0KCJpbmZvIikgaXMgTm9uZToKICAgICAgICAgICAgc3lz
            LnN0ZGVyci53cml0ZSgKICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHJndyBFbmRwb2lu
            dCwgJ3tzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludH0nLCBpcyBpbnZhbGlkLiIKICAgICAg
            ICAgICAgKQogICAgICAgICAgICByZXR1cm4gKAogICAgICAgICAgICAgICAgIiIsCiAgICAgICAg
            ICAgICAgICAiLTEiLAogICAgICAgICAgICApCgogICAgICAgIHJldHVybiByMVsiaW5mbyJdWyJz
            dG9yYWdlX2JhY2tlbmRzIl1bMF1bImNsdXN0ZXJfaWQiXSwgIiIKCiAgICBkZWYgdmFsaWRhdGVf
            cmd3X2VuZHBvaW50KHNlbGYsIGluZm9fY2FwX3N1cHBvcnRlZCk6CiAgICAgICAgIyBpZiB0aGUg
            J2NsdXN0ZXInIGluc3RhbmNlIGlzIGEgZHVtbXkgb25lLAogICAgICAgICMgZG9uJ3QgdHJ5IHRv
            IHJlYWNoIG91dCB0byB0aGUgZW5kcG9pbnQKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1
            c3RlciwgRHVtbXlSYWRvcyk6CiAgICAgICAgICAgIHJldHVybgoKICAgICAgICByZ3dfZW5kcG9p
            bnQgPSBzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludAoKICAgICAgICAjIHZhbGlkYXRlIHJn
            dyBlbmRwb2ludCBvbmx5IGlmIGlwIGFkZHJlc3MgaXMgcGFzc2VkCiAgICAgICAgaXBfdHlwZSA9
            IHNlbGYuX2ludmFsaWRfZW5kcG9pbnQocmd3X2VuZHBvaW50KQoKICAgICAgICAjIGNoZWNrIGlm
            IHRoZSByZ3cgZW5kcG9pbnQgaXMgcmVhY2hhYmxlCiAgICAgICAgY2VydCA9IE5vbmUKICAgICAg
            ICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNlci5yZ3dfc2tpcF90bHMgYW5kIHNlbGYudmFsaWRhdGVf
            cmd3X2VuZHBvaW50X3Rsc19jZXJ0KCk6CiAgICAgICAgICAgIGNlcnQgPSBzZWxmLl9hcmdfcGFy
            c2VyLnJnd190bHNfY2VydF9wYXRoCiAgICAgICAgYmFzZV91cmwsIHZlcmlmeSwgZXJyID0gc2Vs
            Zi5lbmRwb2ludF9kaWFsKHJnd19lbmRwb2ludCwgaXBfdHlwZSwgY2VydD1jZXJ0KQogICAgICAg
            IGlmIGVyciAhPSAiIjoKICAgICAgICAgICAgcmV0dXJuICItMSIKCiAgICAgICAgIyBjaGVjayBp
            ZiB0aGUgcmd3IGVuZHBvaW50IGJlbG9uZ3MgdG8gdGhlIHNhbWUgY2x1c3RlcgogICAgICAgICMg
            b25seSBjaGVjayBpZiBgaW5mb2AgY2FwIGlzIHN1cHBvcnRlZAogICAgICAgIGlmIGluZm9fY2Fw
            X3N1cHBvcnRlZDoKICAgICAgICAgICAgZnNpZCA9IHNlbGYuZ2V0X2ZzaWQoKQogICAgICAgICAg
            ICByZ3dfZnNpZCwgZXJyID0gc2VsZi5nZXRfcmd3X2ZzaWQoYmFzZV91cmwsIHZlcmlmeSkKICAg
            ICAgICAgICAgaWYgZXJyID09ICItMSI6CiAgICAgICAgICAgICAgICByZXR1cm4gIi0xIgogICAg
            ICAgICAgICBpZiBmc2lkICE9IHJnd19mc2lkOgogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53
            cml0ZSgKICAgICAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCByZ3cgRW5kcG9pbnQsICd7
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfZW5kcG9pbnR9JywgaXMgaW52YWxpZC4gV2UgYXJlIHZhbGlk
            YXRpbmcgYnkgY2FsbGluZyB0aGUgYWRtaW5vcHMgYXBpIHRocm91Z2ggcmd3LWVuZHBvaW50IGFu
            ZCB2YWxpZGF0aW5nIHRoZSBjbHVzdGVyX2lkICd7cmd3X2ZzaWR9JyBpcyBlcXVhbCB0byB0aGUg
            Y2VwaCBjbHVzdGVyIGZzaWQgJ3tmc2lkfSciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAg
            ICAgICByZXR1cm4gIi0xIgoKICAgICAgICAjIGNoZWNrIGlmIHRoZSByZ3cgZW5kcG9pbnQgcG9v
            bCBleGlzdAogICAgICAgICMgb25seSB2YWxpZGF0ZSBpZiByZ3dfcG9vbF9wcmVmaXggaXMgcGFz
            c2VkIGVsc2UgaXQgd2lsbCB0YWtlIGRlZmF1bHQgdmFsdWUgYW5kIHdlIGRvbid0IGNyZWF0ZSB0
            aGVzZSBkZWZhdWx0IHBvb2xzCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9w
            cmVmaXggIT0gImRlZmF1bHQiOgogICAgICAgICAgICByZ3dfcG9vbHNfdG9fdmFsaWRhdGUgPSBb
            CiAgICAgICAgICAgICAgICBmIntzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3
            Lm1ldGEiLAogICAgICAgICAgICAgICAgIi5yZ3cucm9vdCIsCiAgICAgICAgICAgICAgICBmIntz
            ZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmNvbnRyb2wiLAogICAgICAgICAg
            ICAgICAgZiJ7c2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5sb2ciLAogICAg
            ICAgICAgICBdCiAgICAgICAgICAgIGZvciBfcmd3X3Bvb2xfdG9fdmFsaWRhdGUgaW4gcmd3X3Bv
            b2xzX3RvX3ZhbGlkYXRlOgogICAgICAgICAgICAgICAgaWYgbm90IHNlbGYuY2x1c3Rlci5wb29s
            X2V4aXN0cyhfcmd3X3Bvb2xfdG9fdmFsaWRhdGUpOgogICAgICAgICAgICAgICAgICAgIHN5cy5z
            dGRlcnIud3JpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHBvb2ws
            ICd7X3Jnd19wb29sX3RvX3ZhbGlkYXRlfScsIGRvZXMgbm90IGV4aXN0IgogICAgICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgICAgICAgICByZXR1cm4gIi0xIgoKICAgICAgICByZXR1cm4g
            IiIKCiAgICBkZWYgdmFsaWRhdGVfcmd3X211bHRpc2l0ZShzZWxmLCByZ3dfbXVsdGlzaXRlX2Nv
            bmZpZ19uYW1lLCByZ3dfbXVsdGlzaXRlX2NvbmZpZyk6CiAgICAgICAgaWYgcmd3X211bHRpc2l0
            ZV9jb25maWcgIT0gIiI6CiAgICAgICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgICAgICJyYWRv
            c2d3LWFkbWluIiwKICAgICAgICAgICAgICAgIHJnd19tdWx0aXNpdGVfY29uZmlnLAogICAgICAg
            ICAgICAgICAgImdldCIsCiAgICAgICAgICAgICAgICAiLS1yZ3ctIiArIHJnd19tdWx0aXNpdGVf
            Y29uZmlnLAogICAgICAgICAgICAgICAgcmd3X211bHRpc2l0ZV9jb25maWdfbmFtZSwKICAgICAg
            ICAgICAgXQogICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBfID0gc3VicHJvY2Vzcy5j
            aGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAgICBleGNl
            cHQgc3VicHJvY2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3IgYXMgZXhlY0VycjoKICAgICAgICAgICAg
            ICAgIGVycl9tc2cgPSAoCiAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBj
            b21tYW5kIHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAgICAgICAgICAg
            ICAgICBmIkNvZGU6IHtleGVjRXJyLnJldHVybmNvZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJy
            fSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJy
            X21zZykKICAgICAgICAgICAgICAgIHJldHVybiAiLTEiCiAgICAgICAgcmV0dXJuICIiCgogICAg
            ZGVmIGNvbnZlcnRfY29tbWFfc2VwYXJhdGVkX3RvX2FycmF5KHNlbGYsIHZhbHVlKToKICAgICAg
            ICByZXR1cm4gdmFsdWUuc3BsaXQoIiwiKQoKICAgIGRlZiByYWlzZV9leGNlcHRpb25faWZfYW55
            X3RvcG9sb2d5X2ZsYWdfaXNfbWlzc2luZyhzZWxmKToKICAgICAgICBpZiAoCiAgICAgICAgICAg
            ICgKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMgIT0gIiIK
            ICAgICAgICAgICAgICAgIGFuZCAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCA9PSAiIgogICAgICAgICAgICAgICAgICAg
            IG9yIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfZmFpbHVyZV9kb21haW5fdmFsdWVzID09ICIi
            CiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICkKICAgICAgICAgICAgb3IgKAogICAgICAg
            ICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCAh
            PSAiIgogICAgICAgICAgICAgICAgYW5kICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdf
            cGFyc2VyLnRvcG9sb2d5X3Bvb2xzID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5f
            YXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMgPT0gIiIKICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgKQogICAgICAgICAgICBvciAoCiAgICAgICAgICAgICAgICBz
            ZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX3ZhbHVlcyAhPSAiIgogICAg
            ICAgICAgICAgICAgYW5kICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRv
            cG9sb2d5X3Bvb2xzID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5fYXJnX3BhcnNl
            ci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCA9PSAiIgogICAgICAgICAgICAgICAgKQog
            ICAgICAgICAgICApCiAgICAgICAgKToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVy
            ZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICJwcm92aWRlIGFsbCB0aGUgdG9wb2xvZ3kgZmxh
            Z3MgLS10b3BvbG9neS1wb29scywgLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi1sYWJlbCwgLS10
            b3BvbG9neS1mYWlsdXJlLWRvbWFpbi12YWx1ZXMiCiAgICAgICAgICAgICkKCiAgICBkZWYgdmFs
            aWRhdGVfdG9wb2xvZ3lfdmFsdWVzKHNlbGYsIHRvcG9sb2d5X3Bvb2xzLCB0b3BvbG9neV9mZCk6
            CiAgICAgICAgaWYgbGVuKHRvcG9sb2d5X3Bvb2xzKSAhPSBsZW4odG9wb2xvZ3lfZmQpOgogICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ZiJUaGUgcHJvdmlkZWQgdG9wb2xvZ3kgcG9vbHMsICd7dG9wb2xvZ3lfcG9vbHN9JywgYW5kICIK
            ICAgICAgICAgICAgICAgIGYidG9wb2xvZ3kgZmFpbHVyZSBkb21haW4sICd7dG9wb2xvZ3lfZmR9
            JywiCiAgICAgICAgICAgICAgICBmImFyZSBvZiBkaWZmZXJlbnQgbGVuZ3RoLCAne2xlbih0b3Bv
            bG9neV9wb29scyl9JyBhbmQgJ3tsZW4odG9wb2xvZ3lfZmQpfScgcmVzcGN0aXZlbHkiCiAgICAg
            ICAgICAgICkKICAgICAgICByZXR1cm4KCiAgICBkZWYgdmFsaWRhdGVfdG9wb2xvZ3lfcmJkX3Bv
            b2xzKHNlbGYsIHRvcG9sb2d5X3JiZF9wb29scyk6CiAgICAgICAgZm9yIHBvb2wgaW4gdG9wb2xv
            Z3lfcmJkX3Bvb2xzOgogICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3JiZF9wb29sKHBvb2wpCgog
            ICAgZGVmIGluaXRfdG9wb2xvZ3lfcmJkX3Bvb2xzKHNlbGYsIHRvcG9sb2d5X3JiZF9wb29scyk6
            CiAgICAgICAgZm9yIHBvb2wgaW4gdG9wb2xvZ3lfcmJkX3Bvb2xzOgogICAgICAgICAgICBzZWxm
            LmluaXRfcmJkX3Bvb2wocG9vbCkKCiAgICBkZWYgX2dlbl9vdXRwdXRfbWFwKHNlbGYpOgogICAg
            ICAgIGlmIHNlbGYub3V0X21hcDoKICAgICAgICAgICAgcmV0dXJuCiAgICAgICAgIyBzdXBwb3J0
            IGxlZ2FjeSBmbGFnIHdpdGggdXBncmFkZXMKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNs
            dXN0ZXJfbmFtZToKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1l
            ID0gc2VsZi5fYXJnX3BhcnNlci5jbHVzdGVyX25hbWUKICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            Lms4c19jbHVzdGVyX25hbWUgPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuazhzX2Ns
            dXN0ZXJfbmFtZS5sb3dlcigpCiAgICAgICAgKSAgIyBhbHdheXMgY29udmVydCBjbHVzdGVyIG5h
            bWUgdG8gbG93ZXJjYXNlIGNoYXJhY3RlcnMKICAgICAgICBzZWxmLnZhbGlkYXRlX3JiZF9wb29s
            KHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lKQogICAgICAgIHNlbGYuaW5pdF9y
            YmRfcG9vbChzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZSkKICAgICAgICBzZWxm
            LnZhbGlkYXRlX3JhZG9zX25hbWVzcGFjZSgpCiAgICAgICAgc2VsZi5fZXhjbHVkZWRfa2V5cy5h
            ZGQoIks4U19DTFVTVEVSX05BTUUiKQogICAgICAgIHNlbGYuZ2V0X2NlcGhmc19kYXRhX3Bvb2xf
            ZGV0YWlscygpCiAgICAgICAgc2VsZi5vdXRfbWFwWyJOQU1FU1BBQ0UiXSA9IHNlbGYuX2FyZ19w
            YXJzZXIubmFtZXNwYWNlCiAgICAgICAgc2VsZi5vdXRfbWFwWyJLOFNfQ0xVU1RFUl9OQU1FIl0g
            PSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBb
            IlJPT0tfRVhURVJOQUxfRlNJRCJdID0gc2VsZi5nZXRfZnNpZCgpCiAgICAgICAgc2VsZi5vdXRf
            bWFwWyJST09LX0VYVEVSTkFMX1VTRVJOQU1FIl0gPSBzZWxmLnJ1bl9hc191c2VyCiAgICAgICAg
            c2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0NFUEhfTU9OX0RBVEEiXSA9IHNlbGYuZ2V0X2Nl
            cGhfZXh0ZXJuYWxfbW9uX2RhdGEoKQogICAgICAgIHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5B
            TF9VU0VSX1NFQ1JFVCJdID0gc2VsZi5jcmVhdGVfY2hlY2tlcktleSgKICAgICAgICAgICAgImNs
            aWVudC5oZWFsdGhjaGVja2VyIgogICAgICAgICkKICAgICAgICBzZWxmLm91dF9tYXBbIlJPT0tf
            RVhURVJOQUxfREFTSEJPQVJEX0xJTksiXSA9IHNlbGYuZ2V0X2NlcGhfZGFzaGJvYXJkX2xpbmso
            KQogICAgICAgICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VDUkVU
            Il0sCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FIl0s
            CiAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3VzZXIoImNsaWVudC5jc2kt
            cmJkLW5vZGUiKQogICAgICAgICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BS
            T1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9SQkRfUFJP
            VklTSU9ORVJfU0VDUkVUX05BTUUiXSwKICAgICAgICApID0gc2VsZi5jcmVhdGVfY2VwaENTSUtl
            eXJpbmdfdXNlcigiY2xpZW50LmNzaS1yYmQtcHJvdmlzaW9uZXIiKQogICAgICAgIHNlbGYub3V0
            X21hcFsiQ0VQSEZTX1BPT0xfTkFNRSJdID0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9w
            b29sX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBbIkNFUEhGU19NRVRBREFUQV9QT09MX05BTUUi
            XSA9ICgKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfbWV0YWRhdGFfcG9vbF9u
            YW1lCiAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSA9IHNl
            bGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgIHNlbGYub3V0X21h
            cFsiUkVTVFJJQ1RFRF9BVVRIX1BFUk1JU1NJT04iXSA9ICgKICAgICAgICAgICAgc2VsZi5fYXJn
            X3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbgogICAgICAgICkKICAgICAgICBzZWxm
            Lm91dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdID0gc2VsZi5fYXJnX3BhcnNlci5yYWRvc19uYW1l
            c3BhY2UKICAgICAgICBzZWxmLm91dF9tYXBbIlNVQlZPTFVNRV9HUk9VUCJdID0gc2VsZi5fYXJn
            X3BhcnNlci5zdWJ2b2x1bWVfZ3JvdXAKICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNf
            Tk9ERV9TRUNSRVQiXSA9ICIiCiAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJ
            U0lPTkVSX1NFQ1JFVCJdID0gIiIKICAgICAgICAjIGNyZWF0ZSBDZXBoRlMgbm9kZSBhbmQgcHJv
            dmlzaW9uZXIga2V5cmluZyBvbmx5IHdoZW4gTURTIGV4aXN0cwogICAgICAgIGlmIHNlbGYub3V0
            X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSBhbmQgc2VsZi5vdXRfbWFwWyJDRVBIRlNfUE9PTF9OQU1F
            Il06CiAgICAgICAgICAgICgKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhG
            U19OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZT
            X05PREVfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lL
            ZXlyaW5nX3VzZXIoImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiKQogICAgICAgICAgICAoCiAgICAg
            ICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUIl0s
            CiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VD
            UkVUX05BTUUiXSwKICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3Vz
            ZXIoImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIikKICAgICAgICAgICAgIyBjcmVhdGUg
            dGhlIGRlZmF1bHQgImNzaSIgc3Vidm9sdW1lZ3JvdXAKICAgICAgICAgICAgc2VsZi5nZXRfb3Jf
            Y3JlYXRlX3N1YnZvbHVtZV9ncm91cCgKICAgICAgICAgICAgICAgICJjc2kiLCBzZWxmLl9hcmdf
            cGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICAgICAgKQogICAgICAgICAgICAj
            IHBpbiB0aGUgZGVmYXVsdCAiY3NpIiBzdWJ2b2x1bWVncm91cAogICAgICAgICAgICBzZWxmLnBp
            bl9zdWJ2b2x1bWUoCiAgICAgICAgICAgICAgICAiY3NpIiwgc2VsZi5fYXJnX3BhcnNlci5jZXBo
            ZnNfZmlsZXN5c3RlbV9uYW1lLCAiZGlzdHJpYnV0ZWQiLCAiMSIKICAgICAgICAgICAgKQogICAg
            ICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlNVQlZPTFVNRV9HUk9VUCJdOgogICAgICAgICAgICAg
            ICAgc2VsZi5nZXRfb3JfY3JlYXRlX3N1YnZvbHVtZV9ncm91cCgKICAgICAgICAgICAgICAgICAg
            ICBzZWxmLl9hcmdfcGFyc2VyLnN1YnZvbHVtZV9ncm91cCwKICAgICAgICAgICAgICAgICAgICBz
            ZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgICAgICBzZWxmLnBpbl9zdWJ2b2x1bWUoCiAgICAgICAgICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci5zdWJ2b2x1bWVfZ3JvdXAsCiAgICAgICAgICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lLAogICAgICAgICAgICAgICAgICAg
            ICJkaXN0cmlidXRlZCIsCiAgICAgICAgICAgICAgICAgICAgIjEiLAogICAgICAgICAgICAgICAg
            KQogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1RMU19DRVJUIl0gPSAiIgogICAgICAgIHNlbGYu
            b3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdID0gIiIKICAgICAgICBzZWxmLm91dF9tYXBb
            Ik1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJdID0gIiIKICAgICAgICBpZiBub3Qgc2VsZi5fYXJn
            X3BhcnNlci5za2lwX21vbml0b3JpbmdfZW5kcG9pbnQ6CiAgICAgICAgICAgICgKICAgICAgICAg
            ICAgICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdLAogICAgICAgICAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UX1BPUlQiXSwKICAgICAgICAgICAg
            KSA9IHNlbGYuZ2V0X2FjdGl2ZV9hbmRfc3RhbmRieV9tZ3JzKCkKICAgICAgICBzZWxmLm91dF9t
            YXBbIlJCRF9QT09MX05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1l
            CiAgICAgICAgc2VsZi5vdXRfbWFwWyJSQkRfTUVUQURBVEFfRUNfUE9PTF9OQU1FIl0gPSAoCiAg
            ICAgICAgICAgIHNlbGYudmFsaWRhdGVfcmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZSgpCiAgICAg
            ICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9PTFMiXSA9IHNlbGYuX2FyZ19w
            YXJzZXIudG9wb2xvZ3lfcG9vbHMKICAgICAgICBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxV
            UkVfRE9NQUlOX0xBQkVMIl0gPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xv
            Z3lfZmFpbHVyZV9kb21haW5fbGFiZWwKICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJU
            T1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiXSA9ICgKICAgICAgICAgICAgc2VsZi5fYXJn
            X3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMKICAgICAgICApCiAgICAgICAg
            aWYgKAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X3Bvb2xzICE9ICIiCiAg
            ICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX2xh
            YmVsICE9ICIiCiAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1
            cmVfZG9tYWluX3ZhbHVlcyAhPSAiIgogICAgICAgICk6CiAgICAgICAgICAgIHNlbGYudmFsaWRh
            dGVfdG9wb2xvZ3lfdmFsdWVzKAogICAgICAgICAgICAgICAgc2VsZi5jb252ZXJ0X2NvbW1hX3Nl
            cGFyYXRlZF90b19hcnJheShzZWxmLm91dF9tYXBbIlRPUE9MT0dZX1BPT0xTIl0pLAogICAgICAg
            ICAgICAgICAgc2VsZi5jb252ZXJ0X2NvbW1hX3NlcGFyYXRlZF90b19hcnJheSgKICAgICAgICAg
            ICAgICAgICAgICBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX1ZBTFVFUyJd
            CiAgICAgICAgICAgICAgICApLAogICAgICAgICAgICApCiAgICAgICAgICAgIHNlbGYudmFsaWRh
            dGVfdG9wb2xvZ3lfcmJkX3Bvb2xzKAogICAgICAgICAgICAgICAgc2VsZi5jb252ZXJ0X2NvbW1h
            X3NlcGFyYXRlZF90b19hcnJheShzZWxmLm91dF9tYXBbIlRPUE9MT0dZX1BPT0xTIl0pCiAgICAg
            ICAgICAgICkKICAgICAgICAgICAgc2VsZi5pbml0X3RvcG9sb2d5X3JiZF9wb29scygKICAgICAg
            ICAgICAgICAgIHNlbGYuY29udmVydF9jb21tYV9zZXBhcmF0ZWRfdG9fYXJyYXkoc2VsZi5vdXRf
            bWFwWyJUT1BPTE9HWV9QT09MUyJdKQogICAgICAgICAgICApCiAgICAgICAgZWxzZToKICAgICAg
            ICAgICAgc2VsZi5yYWlzZV9leGNlcHRpb25faWZfYW55X3RvcG9sb2d5X2ZsYWdfaXNfbWlzc2lu
            ZygpCgogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1BPT0xfUFJFRklYIl0gPSBzZWxmLl9hcmdf
            cGFyc2VyLnJnd19wb29sX3ByZWZpeAogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX0VORFBPSU5U
            Il0gPSAiIgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50OgogICAgICAg
            ICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgICAgICBzZWxmLmNy
            ZWF0ZV9yZ3dfYWRtaW5fb3BzX3VzZXIoKQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAg
            ICAgaWYgKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25h
            bWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9u
            ZWdyb3VwX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfem9uZV9uYW1lICE9ICIiCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAg
            ICAgIGVyciA9IHNlbGYudmFsaWRhdGVfcmd3X211bHRpc2l0ZSgKICAgICAgICAgICAgICAgICAg
            ICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwgInJlYWxtIgogICAgICAgICAg
            ICAgICAgICAgICkKICAgICAgICAgICAgICAgICAgICBlcnIgPSBzZWxmLnZhbGlkYXRlX3Jnd19t
            dWx0aXNpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pv
            bmVncm91cF9uYW1lLCAiem9uZWdyb3VwIgogICAgICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgICAgICAgICBlcnIgPSBzZWxmLnZhbGlkYXRlX3Jnd19tdWx0aXNpdGUoCiAgICAgICAgICAg
            ICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVfbmFtZSwgInpvbmUiCiAgICAg
            ICAgICAgICAgICAgICAgKQoKICAgICAgICAgICAgICAgIGlmICgKICAgICAgICAgICAgICAgICAg
            ICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lID09ICIiCiAgICAgICAgICAgICAgICAg
            ICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lID09ICIiCiAgICAgICAg
            ICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVfbmFtZSA9PSAiIgogICAg
            ICAgICAgICAgICAgKSBvciAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5y
            Z3dfcmVhbG1fbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFy
            c2VyLnJnd196b25lZ3JvdXBfbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd196b25lX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICk6CiAgICAg
            ICAgICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJH
            V19BRE1JTl9PUFNfVVNFUl9BQ0NFU1NfS0VZIl0sCiAgICAgICAgICAgICAgICAgICAgICAgIHNl
            bGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX1NFQ1JFVF9LRVkiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkLAogICAgICAgICAgICAgICAgICAgICAgICBl
            cnIsCiAgICAgICAgICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRlX3Jnd19hZG1pbl9vcHNfdXNl
            cigpCiAgICAgICAgICAgICAgICAgICAgZXJyID0gc2VsZi52YWxpZGF0ZV9yZ3dfZW5kcG9pbnQo
            aW5mb19jYXBfc3VwcG9ydGVkKQogICAgICAgICAgICAgICAgICAgIGlmIHNlbGYuX2FyZ19wYXJz
            ZXIucmd3X3Rsc19jZXJ0X3BhdGg6CiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYub3V0X21h
            cFsiUkdXX1RMU19DRVJUIl0gPSAoCiAgICAgICAgICAgICAgICAgICAgICAgICAgICBzZWxmLnZh
            bGlkYXRlX3Jnd19lbmRwb2ludF90bHNfY2VydCgpCiAgICAgICAgICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgICAgICAgICAjIGlmIHRoZXJlIGlzIG5vIGVycm9yLCBzZXQgdGhlIFJHV19F
            TkRQT0lOVAogICAgICAgICAgICAgICAgICAgIGlmIGVyciAhPSAiLTEiOgogICAgICAgICAgICAg
            ICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdID0gc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfZW5kcG9pbnQKICAgICAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICAgICAg
            ZXJyID0gIlBsZWFzZSBwcm92aWRlIGFsbCB0aGUgUkdXIG11bHRpc2l0ZSBwYXJhbWV0ZXJzIG9y
            IG5vbmUgb2YgdGhlbSIKICAgICAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKGVycikK
            CiAgICBkZWYgZ2VuX3NoZWxsX291dChzZWxmKToKICAgICAgICBzZWxmLl9nZW5fb3V0cHV0X21h
            cCgpCiAgICAgICAgc2hPdXRJTyA9IFN0cmluZ0lPKCkKICAgICAgICBmb3IgaywgdiBpbiBzZWxm
            Lm91dF9tYXAuaXRlbXMoKToKICAgICAgICAgICAgaWYgdiBhbmQgayBub3QgaW4gc2VsZi5fZXhj
            bHVkZWRfa2V5czoKICAgICAgICAgICAgICAgIHNoT3V0SU8ud3JpdGUoZiJleHBvcnQge2t9PXt2
            fXtMSU5FU0VQfSIpCiAgICAgICAgc2hPdXQgPSBzaE91dElPLmdldHZhbHVlKCkKICAgICAgICBz
            aE91dElPLmNsb3NlKCkKICAgICAgICByZXR1cm4gc2hPdXQKCiAgICBkZWYgZ2VuX2pzb25fb3V0
            KHNlbGYpOgogICAgICAgIHNlbGYuX2dlbl9vdXRwdXRfbWFwKCkKICAgICAgICBpZiBzZWxmLl9h
            cmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiAiIgogICAgICAgIGpzb25fb3V0
            ID0gWwogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFtZSI6ICJyb29rLWNlcGgtbW9u
            LWVuZHBvaW50cyIsCiAgICAgICAgICAgICAgICAia2luZCI6ICJDb25maWdNYXAiLAogICAgICAg
            ICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiBzZWxmLm91dF9t
            YXBbIlJPT0tfRVhURVJOQUxfQ0VQSF9NT05fREFUQSJdLAogICAgICAgICAgICAgICAgICAgICJt
            YXhNb25JZCI6ICIwIiwKICAgICAgICAgICAgICAgICAgICAibWFwcGluZyI6ICJ7fSIsCiAgICAg
            ICAgICAgICAgICB9LAogICAgICAgICAgICB9LAogICAgICAgICAgICB7CiAgICAgICAgICAgICAg
            ICAibmFtZSI6ICJyb29rLWNlcGgtbW9uIiwKICAgICAgICAgICAgICAgICJraW5kIjogIlNlY3Jl
            dCIsCiAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAiYWRtaW4t
            c2VjcmV0IjogImFkbWluLXNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImZzaWQiOiBzZWxm
            Lm91dF9tYXBbIlJPT0tfRVhURVJOQUxfRlNJRCJdLAogICAgICAgICAgICAgICAgICAgICJtb24t
            c2VjcmV0IjogIm1vbi1zZWNyZXQiLAogICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgfSwK
            ICAgICAgICAgICAgewogICAgICAgICAgICAgICAgIm5hbWUiOiAicm9vay1jZXBoLW9wZXJhdG9y
            LWNyZWRzIiwKICAgICAgICAgICAgICAgICJraW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAg
            ICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJS
            T09LX0VYVEVSTkFMX1VTRVJOQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgInVzZXJLZXkiOiBz
            ZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAg
            IH0sCiAgICAgICAgICAgIH0sCiAgICAgICAgXQoKICAgICAgICAjIGlmICdNT05JVE9SSU5HX0VO
            RFBPSU5UJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ21vbml0b3JpbmctZW5kcG9pbnQnIHRvIENs
            dXN0ZXIKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19F
            TkRQT0lOVCJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9J
            TlRfUE9SVCJdCiAgICAgICAgKToKICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAg
            ICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjogIm1vbml0b3JpbmctZW5kcG9p
            bnQiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIkNlcGhDbHVzdGVyIiwKICAgICAgICAg
            ICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgIk1vbml0b3JpbmdF
            bmRwb2ludCI6IHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdLAogICAgICAgICAg
            ICAgICAgICAgICAgICAiTW9uaXRvcmluZ1BvcnQiOiBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdf
            RU5EUE9JTlRfUE9SVCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9
            CiAgICAgICAgICAgICkKCiAgICAgICAgIyBpZiAnQ1NJX1JCRF9OT0RFX1NFQ1JFVCcgZXhpc3Rz
            LCB0aGVuIG9ubHkgYWRkICdyb29rLWNzaS1yYmQtcHJvdmlzaW9uZXInIFNlY3JldAogICAgICAg
            IGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VDUkVUIl0KICAg
            ICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FIl0KICAg
            ICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAg
            ICAgICAgICAgICAgICAgICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9O
            T0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQi
            LAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAi
            dXNlcklEIjogc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VDUkVUX05BTUUiXSwKICAgICAg
            ICAgICAgICAgICAgICAgICAgInVzZXJLZXkiOiBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9T
            RUNSRVQiXSwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAg
            ICAgICApCiAgICAgICAgIyBpZiAnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQnIGV4aXN0cywg
            dGhlbiBvbmx5IGFkZCAncm9vay1jc2ktcmJkLXByb3Zpc2lvbmVyJyBTZWNyZXQKICAgICAgICBp
            ZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQi
            XQogICAgICAgICAgICBhbmQgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JF
            VF9OQU1FIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAg
            ICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21h
            cFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAg
            ICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAg
            ICAgICAgICAgICAgICAgICAgInVzZXJJRCI6IHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJ
            T05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNl
            bGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAg
            ICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBpZiAnQ1NJ
            X0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAncm9vay1j
            c2ktY2VwaGZzLXByb3Zpc2lvbmVyJyBTZWNyZXQKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNl
            bGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVQiXQogICAgICAgICAgICBh
            bmQgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0KICAg
            ICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAg
            ICAgICAgICAgICAgICAgICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhG
            U19QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQi
            OiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAg
            ICAgICAgICAgImFkbWluSUQiOiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJf
            U0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgImFkbWluS2V5Ijogc2VsZi5v
            dXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAg
            ICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdDU0lf
            Q0VQSEZTX05PREVfU0VDUkVUJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ3Jvb2stY3NpLWNlcGhm
            cy1ub2RlJyBTZWNyZXQKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJ
            X0NFUEhGU19OT0RFX1NFQ1JFVCJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9D
            RVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJdCiAgICAgICAgKToKICAgICAgICAgICAganNvbl9vdXQu
            YXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjogZiJy
            b29rLXtzZWxmLm91dF9tYXBbJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAg
            ICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0
            YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImFkbWluSUQiOiBzZWxmLm91dF9tYXBbIkNT
            SV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAiYWRt
            aW5LZXkiOiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVQiXSwKICAgICAgICAg
            ICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBp
            ZiAnUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSycgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdy
            b29rLWNlcGgtZGFzaGJvYXJkLWxpbmsnIFNlY3JldAogICAgICAgIGlmIHNlbGYub3V0X21hcFsi
            Uk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSyJdOgogICAgICAgICAgICBqc29uX291dC5hcHBl
            bmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAicm9vay1j
            ZXBoLWRhc2hib2FyZC1saW5rIiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQi
            LAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAi
            dXNlcklEIjogImNlcGgtZGFzaGJvYXJkLWxpbmsiLAogICAgICAgICAgICAgICAgICAgICAgICAi
            dXNlcktleSI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSyJdLAog
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAg
            ICAgICAjIGlmICdSQURPU19OQU1FU1BBQ0UnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCB0aGUgIlJB
            RE9TX05BTUVTUEFDRSIgbmFtZXNwYWNlCiAgICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLm91
            dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIlJF
            U1RSSUNURURfQVVUSF9QRVJNSVNTSU9OIl0KICAgICAgICAgICAgYW5kIG5vdCBzZWxmLm91dF9t
            YXBbIlJCRF9NRVRBREFUQV9FQ19QT09MX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpz
            b25fb3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFt
            ZSI6ICJyYWRvcy1uYW1lc3BhY2UiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIkNlcGhC
            bG9ja1Bvb2xSYWRvc05hbWVzcGFjZSIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAg
            ICAgICAgICAgICAgICAgICAgICAgICJyYWRvc05hbWVzcGFjZU5hbWUiOiBzZWxmLm91dF9tYXBb
            IlJBRE9TX05BTUVTUEFDRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYu
            b3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAg
            ICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAg
            ICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjogImNlcGgtcmJkLXJhZG9z
            LW5hbWVzcGFjZSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwK
            ICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgInBv
            b2wiOiBzZWxmLm91dF9tYXBbIlJCRF9QT09MX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAg
            ICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNyZXQtbmFtZSI6IGYicm9vay17
            c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAg
            ICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250cm9sbGVyLWV4cGFuZC1z
            ZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NF
            Q1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5p
            by9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRf
            Tk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAg
            ICAgIH0KICAgICAgICAgICAgKQogICAgICAgIGVsc2U6CiAgICAgICAgICAgIGlmIHNlbGYub3V0
            X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdOgogICAgICAgICAgICAgICAganNvbl9v
            dXQuYXBwZW5kKAogICAgICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAgICAg
            Im5hbWUiOiAiY2VwaC1yYmQiLAogICAgICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9y
            YWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAg
            ICAgICAgICAgICAgICAgICJkYXRhUG9vbCI6IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJd
            LAogICAgICAgICAgICAgICAgICAgICAgICAgICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIlJCRF9N
            RVRBREFUQV9FQ19QT09MX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJjc2ku
            c3RvcmFnZS5rOHMuaW8vcHJvdmlzaW9uZXItc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0
            X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNlY3Jl
            dC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVU
            X05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5p
            by9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRf
            Tk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAg
            ICAgICAgICAgICAgfQogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBlbHNlOgogICAgICAg
            ICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgICAgIHsKICAgICAgICAg
            ICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaC1yYmQiLAogICAgICAgICAgICAgICAgICAgICAg
            ICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAgICAgICAiZGF0YSI6
            IHsKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJwb29sIjogc2VsZi5vdXRfbWFwWyJSQkRf
            UE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2Uuazhz
            LmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9S
            QkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAg
            ICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYi
            cm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwK
            ICAgICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vbm9kZS1zdGFn
            ZS1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX05PREVfU0VDUkVU
            X05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgICAg
            IH0KICAgICAgICAgICAgICAgICkKCiAgICAgICAgIyBpZiAnVE9QT0xPR1lfUE9PTFMnLCAnVE9Q
            T0xPR1lfRkFJTFVSRV9ET01BSU5fTEFCRUwnLCAnVE9QT0xPR1lfRkFJTFVSRV9ET01BSU5fVkFM
            VUVTJyAgZXhpc3RzLAogICAgICAgICMgdGhlbiBvbmx5IGFkZCAndG9wb2xvZ3knIFN0b3JhZ2VD
            bGFzcwogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09M
            UyJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlO
            X0xBQkVMIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfRkFJTFVSRV9E
            T01BSU5fVkFMVUVTIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAg
            ICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaC1yYmQtdG9w
            b2xvZ3kiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAg
            ICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJ0b3BvbG9n
            eUZhaWx1cmVEb21haW5MYWJlbCI6IHNlbGYub3V0X21hcFsKICAgICAgICAgICAgICAgICAgICAg
            ICAgICAgICJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9MQUJFTCIKICAgICAgICAgICAgICAgICAg
            ICAgICAgXSwKICAgICAgICAgICAgICAgICAgICAgICAgInRvcG9sb2d5RmFpbHVyZURvbWFpblZh
            bHVlcyI6IHNlbGYub3V0X21hcFsKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJUT1BPTE9H
            WV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiCiAgICAgICAgICAgICAgICAgICAgICAgIF0sCiAgICAg
            ICAgICAgICAgICAgICAgICAgICJ0b3BvbG9neVBvb2xzIjogc2VsZi5vdXRfbWFwWyJUT1BPTE9H
            WV9QT09MUyJdLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3By
            b3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJP
            VklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0
            b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxm
            Lm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUi
            OiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAg
            ICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCgogICAg
            ICAgICMgaWYgJ0NFUEhGU19GU19OQU1FJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ2NlcGhmcycg
            U3RvcmFnZUNsYXNzCiAgICAgICAgaWYgc2VsZi5vdXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdOgog
            ICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAg
            ICAgICAgICAgIm5hbWUiOiAiY2VwaGZzIiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJT
            dG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAg
            ICAgICAgICAgICAiZnNOYW1lIjogc2VsZi5vdXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdLAogICAg
            ICAgICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYub3V0X21hcFsiQ0VQSEZTX1BPT0xfTkFN
            RSJdLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lv
            bmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9DRVBIRlNfUFJPVklT
            SU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3Jh
            Z2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91
            dF9tYXBbJ0NTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUi
            OiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FJ119IiwK
            ICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAg
            ICAgICAgIyBpZiAnUkdXX0VORFBPSU5UJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ2NlcGgtcmd3
            JyBTdG9yYWdlQ2xhc3MKICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdOgog
            ICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAg
            ICAgICAgICAgIm5hbWUiOiAiY2VwaC1yZ3ciLAogICAgICAgICAgICAgICAgICAgICJraW5kIjog
            IlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAg
            ICAgICAgICAgICAgICJlbmRwb2ludCI6IHNlbGYub3V0X21hcFsiUkdXX0VORFBPSU5UIl0sCiAg
            ICAgICAgICAgICAgICAgICAgICAgICJwb29sUHJlZml4Ijogc2VsZi5vdXRfbWFwWyJSR1dfUE9P
            TF9QUkVGSVgiXSwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAg
            ICAgICAgICApCiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsK
            ICAgICAgICAgICAgICAgICAgICAibmFtZSI6ICJyZ3ctYWRtaW4tb3BzLXVzZXIiLAogICAgICAg
            ICAgICAgICAgICAgICJraW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEi
            OiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJhY2Nlc3NLZXkiOiBzZWxmLm91dF9tYXBbIlJH
            V19BRE1JTl9PUFNfVVNFUl9BQ0NFU1NfS0VZIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJz
            ZWNyZXRLZXkiOiBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9TRUNSRVRfS0VZIl0s
            CiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQog
            ICAgICAgICMgaWYgJ1JHV19UTFNfQ0VSVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkIHRoZSAiY2Vw
            aC1yZ3ctdGxzLWNlcnQiIHNlY3JldAogICAgICAgIGlmIHNlbGYub3V0X21hcFsiUkdXX1RMU19D
            RVJUIl06CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAg
            ICAgICAgICAgICAgICAgICAibmFtZSI6ICJjZXBoLXJndy10bHMtY2VydCIsCiAgICAgICAgICAg
            ICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsK
            ICAgICAgICAgICAgICAgICAgICAgICAgImNlcnQiOiBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VS
            VCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAg
            ICkKCiAgICAgICAgcmV0dXJuIGpzb24uZHVtcHMoanNvbl9vdXQpICsgTElORVNFUAoKICAgIGRl
            ZiB1cGdyYWRlX3VzZXJzX3Blcm1pc3Npb25zKHNlbGYpOgogICAgICAgIHVzZXJzID0gWwogICAg
            ICAgICAgICAiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIsCiAgICAgICAgICAgICJjbGllbnQuY3Np
            LWNlcGhmcy1wcm92aXNpb25lciIsCiAgICAgICAgICAgICJjbGllbnQuY3NpLXJiZC1ub2RlIiwK
            ICAgICAgICAgICAgImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwKICAgICAgICAgICAgImNs
            aWVudC5oZWFsdGhjaGVja2VyIiwKICAgICAgICBdCiAgICAgICAgaWYgc2VsZi5ydW5fYXNfdXNl
            ciAhPSAiIiBhbmQgc2VsZi5ydW5fYXNfdXNlciBub3QgaW4gdXNlcnM6CiAgICAgICAgICAgIHVz
            ZXJzLmFwcGVuZChzZWxmLnJ1bl9hc191c2VyKQogICAgICAgIGZvciB1c2VyIGluIHVzZXJzOgog
            ICAgICAgICAgICBzZWxmLnVwZ3JhZGVfdXNlcl9wZXJtaXNzaW9ucyh1c2VyKQoKICAgIGRlZiBn
            ZXRfcmd3X3Bvb2xfbmFtZV9kdXJpbmdfdXBncmFkZShzZWxmLCB1c2VyLCBjYXBzKToKICAgICAg
            ICBpZiB1c2VyID09ICJjbGllbnQuaGVhbHRoY2hlY2tlciI6CiAgICAgICAgICAgICMgd2hlbiBh
            ZG1pbiBoYXMgbm90IHByb3ZpZGVkIHJndyBwb29sIG5hbWUgZHVyaW5nIHVwZ3JhZGUsCiAgICAg
            ICAgICAgICMgZ2V0IHRoZSByZ3cgcG9vbCBuYW1lIGZyb20gY2xpZW50LmhlYWx0aGNoZWNrZXIg
            dXNlciB3aGljaCB3YXMgdXNlZCBkdXJpbmcgY29ubmVjdGlvbgogICAgICAgICAgICBpZiBub3Qg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXg6CiAgICAgICAgICAgICAgICAjIFRvIGdl
            dCB2YWx1ZSAnZGVmYXVsdCcgd2hpY2ggaXMgcmd3IHBvb2wgbmFtZSBmcm9tICdhbGxvdyByd3gg
            cG9vbD1kZWZhdWx0LnJndy5tZXRhJwogICAgICAgICAgICAgICAgcGF0dGVybiA9IHIicG9vbD0o
            Lio/KVwucmd3XC5tZXRhIgogICAgICAgICAgICAgICAgbWF0Y2ggPSByZS5zZWFyY2gocGF0dGVy
            biwgY2FwcykKICAgICAgICAgICAgICAgIGlmIG1hdGNoOgogICAgICAgICAgICAgICAgICAgIHNl
            bGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4ID0gbWF0Y2guZ3JvdXAoMSkKICAgICAgICAg
            ICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAgICAgImZhaWxlZCB0byBnZXQgcmd3IHBvb2wg
            bmFtZSBmb3IgdXBncmFkZSIKICAgICAgICAgICAgICAgICAgICApCgogICAgZGVmIHVwZ3JhZGVf
            dXNlcl9wZXJtaXNzaW9ucyhzZWxmLCB1c2VyKToKICAgICAgICAjIGNoZWNrIHdoZXRoZXIgdGhl
            IGdpdmVuIHVzZXIgZXhpc3RzIG9yIG5vdAogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAi
            YXV0aCBnZXQiLCAiZW50aXR5IjogZiJ7dXNlcn0iLCAiZm9ybWF0IjogImpzb24ifQogICAgICAg
            IHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihj
            bWRfanNvbikKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgog
            ICAgICAgICAgICBwcmludChmInVzZXIge3VzZXJ9IG5vdCBmb3VuZCBmb3IgdXBncmFkaW5nLiIp
            CiAgICAgICAgICAgIHJldHVybgogICAgICAgIGV4aXN0aW5nX2NhcHMgPSBqc29uX291dFswXVsi
            Y2FwcyJdCiAgICAgICAgc2VsZi5nZXRfcmd3X3Bvb2xfbmFtZV9kdXJpbmdfdXBncmFkZSh1c2Vy
            LCBzdHIoZXhpc3RpbmdfY2FwcykpCiAgICAgICAgbmV3X2NhcCwgXyA9IHNlbGYuZ2V0X2NhcHNf
            YW5kX2VudGl0eSh1c2VyKQogICAgICAgIGNhcF9rZXlzID0gWyJtb24iLCAibWdyIiwgIm9zZCIs
            ICJtZHMiXQogICAgICAgIGNhcHMgPSBbXQogICAgICAgIGZvciBlYWNoQ2FwIGluIGNhcF9rZXlz
            OgogICAgICAgICAgICBjdXJfY2FwX3ZhbHVlcyA9IGV4aXN0aW5nX2NhcHMuZ2V0KGVhY2hDYXAs
            ICIiKQogICAgICAgICAgICBuZXdfY2FwX3ZhbHVlcyA9IG5ld19jYXAuZ2V0KGVhY2hDYXAsICIi
            KQogICAgICAgICAgICBjdXJfY2FwX3Blcm1fbGlzdCA9IFsKICAgICAgICAgICAgICAgIHguc3Ry
            aXAoKSBmb3IgeCBpbiBjdXJfY2FwX3ZhbHVlcy5zcGxpdCgiLCIpIGlmIHguc3RyaXAoKQogICAg
            ICAgICAgICBdCiAgICAgICAgICAgIG5ld19jYXBfcGVybV9saXN0ID0gWwogICAgICAgICAgICAg
            ICAgeC5zdHJpcCgpIGZvciB4IGluIG5ld19jYXBfdmFsdWVzLnNwbGl0KCIsIikgaWYgeC5zdHJp
            cCgpCiAgICAgICAgICAgIF0KICAgICAgICAgICAgIyBhcHBlbmQgbmV3X2NhcF9saXN0IHRvIGN1
            cl9jYXBfbGlzdCB0byBtYWludGFpbiB0aGUgb3JkZXIgb2YgY2FwcwogICAgICAgICAgICBjdXJf
            Y2FwX3Blcm1fbGlzdC5leHRlbmQobmV3X2NhcF9wZXJtX2xpc3QpCiAgICAgICAgICAgICMgZWxp
            bWluYXRlIGR1cGxpY2F0ZXMgd2l0aG91dCB1c2luZyAnc2V0JwogICAgICAgICAgICAjIHNldCBy
            ZS1vcmRlcnMgaXRlbXMgaW4gdGhlIGxpc3QgYW5kIHdlIGhhdmUgdG8ga2VlcCB0aGUgb3JkZXIK
            ICAgICAgICAgICAgbmV3X2NhcF9saXN0ID0gW10KICAgICAgICAgICAgW25ld19jYXBfbGlzdC5h
            cHBlbmQoeCkgZm9yIHggaW4gY3VyX2NhcF9wZXJtX2xpc3QgaWYgeCBub3QgaW4gbmV3X2NhcF9s
            aXN0XQogICAgICAgICAgICBleGlzdGluZ19jYXBzW2VhY2hDYXBdID0gIiwgIi5qb2luKG5ld19j
            YXBfbGlzdCkKICAgICAgICAgICAgaWYgZXhpc3RpbmdfY2Fwc1tlYWNoQ2FwXToKICAgICAgICAg
            ICAgICAgIGNhcHMuYXBwZW5kKGVhY2hDYXApCiAgICAgICAgICAgICAgICBjYXBzLmFwcGVuZChl
            eGlzdGluZ19jYXBzW2VhY2hDYXBdKQogICAgICAgIGNtZF9qc29uID0gewogICAgICAgICAgICAi
            cHJlZml4IjogImF1dGggY2FwcyIsCiAgICAgICAgICAgICJlbnRpdHkiOiB1c2VyLAogICAgICAg
            ICAgICAiY2FwcyI6IGNhcHMsCiAgICAgICAgICAgICJmb3JtYXQiOiAianNvbiIsCiAgICAgICAg
            fQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9q
            c29uX2dlbihjbWRfanNvbikKICAgICAgICBpZiByZXRfdmFsICE9IDA6CiAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIidhdXRoIGNh
            cHMge3VzZXJ9JyBjb21tYW5kIGZhaWxlZC5cbiBFcnJvcjoge2Vycl9tc2d9IgogICAgICAgICAg
            ICApCiAgICAgICAgcHJpbnQoZiJVcGRhdGVkIHVzZXIge3VzZXJ9IHN1Y2Nlc3NmdWxseS4iKQoK
            ICAgIGRlZiBtYWluKHNlbGYpOgogICAgICAgIGdlbmVyYXRlZF9vdXRwdXQgPSAiIgogICAgICAg
            IGlmIHNlbGYuX2FyZ19wYXJzZXIudXBncmFkZToKICAgICAgICAgICAgc2VsZi51cGdyYWRlX3Vz
            ZXJzX3Blcm1pc3Npb25zKCkKICAgICAgICBlbGlmIHNlbGYuX2FyZ19wYXJzZXIuZm9ybWF0ID09
            ICJqc29uIjoKICAgICAgICAgICAgZ2VuZXJhdGVkX291dHB1dCA9IHNlbGYuZ2VuX2pzb25fb3V0
            KCkKICAgICAgICBlbGlmIHNlbGYuX2FyZ19wYXJzZXIuZm9ybWF0ID09ICJiYXNoIjoKICAgICAg
            ICAgICAgZ2VuZXJhdGVkX291dHB1dCA9IHNlbGYuZ2VuX3NoZWxsX291dCgpCiAgICAgICAgZWxz
            ZToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAg
            ICAgICAgIGYiVW5zdXBwb3J0ZWQgZm9ybWF0OiB7c2VsZi5fYXJnX3BhcnNlci5mb3JtYXR9Igog
            ICAgICAgICAgICApCiAgICAgICAgcHJpbnQoZ2VuZXJhdGVkX291dHB1dCkKICAgICAgICBpZiBz
            ZWxmLm91dHB1dF9maWxlIGFuZCBnZW5lcmF0ZWRfb3V0cHV0OgogICAgICAgICAgICBmT3V0ID0g
            b3BlbihzZWxmLm91dHB1dF9maWxlLCBtb2RlPSJ3IiwgZW5jb2Rpbmc9IlVURi04IikKICAgICAg
            ICAgICAgZk91dC53cml0ZShnZW5lcmF0ZWRfb3V0cHV0KQogICAgICAgICAgICBmT3V0LmNsb3Nl
            KCkKCgojIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMKIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIE1BSU4gIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIwppZiBfX25hbWVfXyA9PSAiX19t
            YWluX18iOgogICAgcmpPYmogPSBSYWRvc0pTT04oKQogICAgdHJ5OgogICAgICAgIHJqT2JqLm1h
            aW4oKQogICAgZXhjZXB0IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24gYXMgZXJyOgogICAgICAg
            IHByaW50KGYiRXhlY3V0aW9uIEZhaWxlZDoge2Vycn0iKQogICAgICAgIHJhaXNlIGVycgogICAg
            ZXhjZXB0IEtleUVycm9yIGFzIGtFcnI6CiAgICAgICAgcHJpbnQoZiJLZXlFcnJvcjoge2tFcnJ9
            IikKICAgIGV4Y2VwdCBPU0Vycm9yIGFzIG9zRXJyOgogICAgICAgIHByaW50KGYiRXJyb3Igd2hp
            bGUgdHJ5aW5nIHRvIG91dHB1dCB0aGUgZGF0YToge29zRXJyfSIpCiAgICBmaW5hbGx5OgogICAg
            ICAgIHJqT2JqLnNodXRkb3duKCkK
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/red-hat-storage/rook
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Ceph cluster.
            displayName: Ceph Cluster
            kind: CephCluster
            name: cephclusters.ceph.rook.io
            version: v1
          - description: Represents a Ceph Block Pool.
            displayName: Ceph Block Pool
            kind: CephBlockPool
            name: cephblockpools.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store.
            displayName: Ceph Object Store
            kind: CephObjectStore
            name: cephobjectstores.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store User.
            displayName: Ceph Object Store User
            kind: CephObjectStoreUser
            name: cephobjectstoreusers.ceph.rook.io
            version: v1
          - description: Represents a cluster of Ceph NFS ganesha gateways.
            displayName: Ceph NFS
            kind: CephNFS
            name: cephnfses.ceph.rook.io
            version: v1
          - description: Represents a Ceph User.
            displayName: Ceph Client
            kind: CephClient
            name: cephclients.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem.
            displayName: Ceph Filesystem
            kind: CephFilesystem
            name: cephfilesystems.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem Mirror.
            displayName: Ceph Filesystem Mirror
            kind: CephFilesystemMirror
            name: cephfilesystemmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph RBD Mirror.
            displayName: Ceph RBD Mirror
            kind: CephRBDMirror
            name: cephrbdmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Realm.
            displayName: Ceph Object Store Realm
            kind: CephObjectRealm
            name: cephobjectrealms.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone Group.
            displayName: Ceph Object Store Zone Group
            kind: CephObjectZoneGroup
            name: cephobjectzonegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone.
            displayName: Ceph Object Store Zone
            kind: CephObjectZone
            name: cephobjectzones.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Notification.
            displayName: Ceph Bucket Notification
            kind: CephBucketNotification
            name: cephbucketnotifications.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Topic.
            displayName: Ceph Bucket Topic
            kind: CephBucketTopic
            name: cephbuckettopics.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem SubVolumeGroup.
            displayName: Ceph Filesystem SubVolumeGroup
            kind: CephFilesystemSubVolumeGroup
            name: cephfilesystemsubvolumegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph BlockPool Rados Namespace.
            displayName: Ceph BlockPool Rados Namespace
            kind: CephBlockPoolRadosNamespace
            name: cephblockpoolradosnamespaces.ceph.rook.io
            version: v1
          - description: Represents a Ceph COSI Driver.
            displayName: Ceph COSI Driver
            kind: CephCOSIDriver
            name: cephcosidrivers.ceph.rook.io
            version: v1
        description: |2

          The Rook-Ceph storage operator packages, deploys, manages, upgrades and scales Ceph storage for providing persistent storage to infrastructure services (Logging, Metrics, Registry) as well as stateful applications in Kubernetes clusters.

          ## Rook-Ceph Storage Operator

          Rook runs as a cloud-native service in Kubernetes clusters for optimal integration with applications in need of storage, and handles the heavy-lifting behind the scenes such as provisioning and management.
          Rook orchestrates battle-tested open-source storage technology Ceph, which has years of production deployments and runs some of the worlds largest clusters.

          Ceph is a massively scalable, software-defined, cloud native storage platform that offers block, file and object storage services.
          Ceph can be used to back a wide variety of applications including relational databases, NoSQL databases, CI/CD tool-sets, messaging, AI/ML and analytics applications.
          Ceph is a proven storage platform that backs some of the world's largest storage deployments and has a large vibrant open source community backing the project.

          ## Supported features
          * **High Availability and resiliency** - Ceph has no single point of failures (SPOF) and all its components work natively in a highly available fashion
          * **Data Protection** - Ceph periodically scrub for inconsistent objects and repair them if necessary, making sure your replicas are always coherent
          * **Consistent storage platform across hybrid cloud** - Ceph can be deployed anywhere (on-premise or bare metal) and thus offers a similar experience regardless
          * **Block, File & Object storage service** - Ceph can expose your data through several storage interfaces, solving all the application use cases
          * **Scale up/down** - addition and removal of storage is fully covered by the operator.
          * **Dashboard** - The Operator deploys a dashboard for monitoring and introspecting your cluster.

          ## Before you start
          https://rook.io/docs/rook/v1.0/k8s-pre-reqs.html
        displayName: Rook-Ceph
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rook
        - ceph
        - storage
        - object storage
        - open source
        - block storage
        - shared filesystem
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/rook
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:cf7f72469d25fb37301a8159406613673f8968c82af1035fd45a470e0ae8acfa
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:80c01633a6bfca44df37ac38f764035be6998fce965c8709e4f6d54d17eeaa7b
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6fee56851da3a87b9f7f477fc59e3385e118dc0b398861a8f13459c998a5af77
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c11a599bfd09b9b22888e03eabdbe1054d4049718a591d2a45bf0ed89800111e
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:82a4b481028195cfd75471656fe40c1cbbade4c7b2d170d6bf9422d9d3b84fa7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:693a60d6a1162659213ae554b1c6ee200b0300a7da72c01415044cf507acab49
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:0afcb516eb528857887c6ee1a9cba081dc8f6b7d47d9f26a0dbb2dca0cefe07c
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:8d8326c74795c9c6bd8f33a9f9e81d2304a70b1f9be3c2f8d50e76be4b0292ff
        version: 4.16.16-rhodf
      entries:
      - name: rook-ceph-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: rook-ceph-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: rook-ceph-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: rook-ceph-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: rook-ceph-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: rook-ceph-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: rook-ceph-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: rook-ceph-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: rook-ceph-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: rook-ceph-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: rook-ceph-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: rook-ceph-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: rook-ceph-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: rook-ceph-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: rook-ceph-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: rook-ceph-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: rook-ceph-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: rook-ceph-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: logic-operator-rhel8
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: logic-operator-rhel8.v1.36.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "host/namespace",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          createdAt: "2025-05-26T14:10:15Z"
          description: OpenShift Serverless Logic Kubernetes Operator for deploying
            workflow applications based on the CNCF Serverless Workflow specification
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-serverless-logic
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kubesmarts/kie-tools
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: Sonata Flow Build
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: Sonata Flow Cluster Platform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: Sonata Flow Platform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: Sonata Flow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          OpenShift Serverless Logic Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification v0.8](https://serverlessworkflow.io/).
        displayName: OpenShift Serverless Logic Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        - workflows
        links:
        - name: Documentation
          url: https://docs.openshift.com/serverless/
        - name: Source Repository
          url: https://github.com/kiegroup/kogito-serverless-operator/
        maintainers:
        - email: serverless-logic-team@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/logic-rhel8-operator@sha256:8d3682448ebdac3aeabb2d23842b7e67a252b95f959c408af805037f9728fd3c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4564ca3dc5bac80d6faddaf94c817fbbc270698a9399d8a21ee1005d85ceda56
        version: 1.36.0
      entries:
      - name: logic-operator-rhel8.v1.36.0
        version: 1.36.0
      - name: logic-operator-rhel8.v1.35.0
        version: 1.35.0
      - name: logic-operator-rhel8.v1.34.0
        version: 1.34.0
      - name: logic-operator-rhel8.v1.33.0
        version: 1.33.0
      name: alpha
    defaultChannel: alpha
    packageName: logic-operator-rhel8
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: The cert-manager maintainers
      provider-url: https://cert-manager.io/
    name: cert-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-manager.v1.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "example-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "example-clusterissuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example-certificate"
                },
                "spec": {
                  "dnsNames": [
                    "example.com"
                  ],
                  "issuerRef": {
                    "name": "example-issuer"
                  },
                  "secretName": "example-certificate-tls"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "name": "example-certificaterequest"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "name": "example-order"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "example-challenge"
                },
                "spec": {}
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: quay.io/jetstack/cert-manager-controller:v1.16.5
          createdAt: 2025-05-02T09:11:31
          olm.skipRange: '>=1.16.0 <1.16.5'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/internal-objects: |-
            [
              "challenges.acme.cert-manager.io",
              "orders.acme.cert-manager.io"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/cert-manager/cert-manager
          support: The cert-manager maintainers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A CertificateRequest is used to request a signed certificate from one of the
              configured issuers.

              All fields within the CertificateRequest's `spec` are immutable after creation.
              A CertificateRequest will either succeed or fail, as denoted by its `Ready` status
              condition and its `status.failureTime` field.

              A CertificateRequest is a one-shot resource, meaning it represents a single
              point in time request for a certificate and cannot be re-used.
            displayName: CertificateRequest
            kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - description: |-
              A Certificate resource should be created to ensure an up to date and signed
              X.509 certificate is stored in the Kubernetes Secret resource named in `spec.secretName`.

              The stored certificate will be renewed before it expires (as configured by `spec.renewBefore`).
            displayName: Certificate
            kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: Challenge is a type to represent a Challenge request with
              an ACME server
            displayName: (Internal) Challenge
            kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - description: |-
              A ClusterIssuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is similar to an Issuer, however it is cluster-scoped and therefore can
              be referenced by resources that exist in *any* namespace, not just the same
              namespace as the referent.
            displayName: ClusterIssuer
            kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - description: |-
              An Issuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is scoped to a single namespace and can therefore only be referenced by
              resources within the same namespace.
            displayName: Issuer
            kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - description: Order is a type to represent an Order with an ACME server
            displayName: (Internal) Order
            kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          cert-manager is a Kubernetes add-on to automate the management and issuance of TLS certificates from various issuing sources.
          It will ensure certificates are valid and up to date periodically, and attempt to renew certificates at an appropriate time before expiry.

          cert-manager has the following [built in issuers][]:

          * ACME: sign certificates using Let's Encrypt (or any other [ACME][] compliant signer).
          * Vault: sign certificates using HashiCorp Vault.
          * Venafi: sign certificates using Venafi Cloud or Venafi Trust Protection Platform.
          * SelfSigned: the private key of the certificate will be used to sign the certificate.
          * CA: a private key stored as a Kubernetes Secret will be used to sign the certificate.

          [built in issuers]: https://cert-manager.io/docs/configuration/#supported-issuer-types
          [ACME]: https://datatracker.ietf.org/doc/html/rfc8555

          cert-manager also supports [external issuers][],
          which are out-of-tree components that provide support for other certificate authorities, such as:
          [Google Cloud Certificate Authority](https://github.com/jetstack/google-cas-issuer),
          [AWS Private Certificate Authority](https://github.com/cert-manager/aws-privateca-issuer), and
          [Cloudflare Origin CA](https://github.com/cloudflare/origin-ca-issuer)

          [external issuers]: https://cert-manager.io/docs/configuration/external/

          In order to configure cert-manager to issue certificates, you must first create an Issuer or ClusterIssuer.
          These resources represent a particular signing authority and tell cert-manager how to handle your certificate requests.
          The [Issuer Documentation][] has more detail.

          [configure cert-manager]: https://cert-manager.io/docs/configuration/
          [Issuer Documentation]: https://cert-manager.io/docs/concepts/issuer/

          Once an Issuer has been configured, you can create a Certificate resource,
          which is a definition of a certificate request that is to be honored by an issuer
          and which is to be kept up-to-date.
          This is the usual way that you will interact with cert-manager to request signed certificates.
          The [Certificate Documentation][] has more detail.

          [Certificate Documentation]: https://cert-manager.io/docs/usage/certificate/
        displayName: cert-manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Documentation
          url: https://cert-manager.io/
        - name: Release Notes
          url: https://cert-manager.io/docs/release-notes
        maintainers:
        - email: cert-manager-maintainers@googlegroups.com
          name: The cert-manager maintainers
        maturity: stable
        minKubeVersion: 1.19.0-0
        provider:
          name: The cert-manager maintainers
          url: https://cert-manager.io/
        relatedImages:
        - quay.io/jetstack/cert-manager-controller:v1.16.5
        - quay.io/jetstack/cert-manager-cainjector:v1.16.5
        - quay.io/jetstack/cert-manager-webhook:v1.16.5
        version: 1.16.5
      entries:
      - name: cert-manager.v1.16.5
        version: 1.16.5
      - name: cert-manager.v1.16.1
        version: 1.16.1
      - name: cert-manager.v1.16.0-beta.0
        version: 1.16.0-beta.0
      - name: cert-manager.v1.15.2
        version: 1.15.2
      - name: cert-manager.v1.15.0
        version: 1.15.0
      - name: cert-manager.v1.14.2
        version: 1.14.2
      - name: cert-manager.v1.14.2-rc1
        version: 1.14.2-rc1
      - name: cert-manager.v1.14.1-rc1
        version: 1.14.1-rc1
      - name: cert-manager.v1.13.3
        version: 1.13.3
      - name: cert-manager.v1.13.3-rc1
        version: 1.13.3-rc1
      - name: cert-manager.v1.13.1
        version: 1.13.1
      - name: cert-manager.v1.13.1-rc1
        version: 1.13.1-rc1
      - name: cert-manager.v1.12.2
        version: 1.12.2
      - name: cert-manager.v1.12.2-rc1
        version: 1.12.2-rc1
      - name: cert-manager.v1.11.4
        version: 1.11.4
      - name: cert-manager.v1.11.4-rc1
        version: 1.11.4-rc1
      - name: cert-manager.v1.11.3-rc1
        version: 1.11.3-rc1
      - name: cert-manager.v1.11.0
        version: 1.11.0
      - name: cert-manager.v1.11.0-rc1
        version: 1.11.0-rc1
      - name: cert-manager.v1.10.2
        version: 1.10.2
      - name: cert-manager.v1.10.2-rc1
        version: 1.10.2-rc1
      - name: cert-manager.v1.10.1
        version: 1.10.1
      - name: cert-manager.v1.10.1-rc1
        version: 1.10.1-rc1
      - name: cert-manager.v1.10.0
        version: 1.10.0
      - name: cert-manager.v1.10.0-rc1
        version: 1.10.0-rc1
      - name: cert-manager.v1.9.1
        version: 1.9.1
      - name: cert-manager.v1.9.1-rc1
        version: 1.9.1-rc1
      - name: cert-manager.v1.8.2
        version: 1.8.2
      - name: cert-manager.v1.8.2-rc1
        version: 1.8.2-rc1
      - name: cert-manager.v1.8.0
        version: 1.8.0
      - name: cert-manager.v1.8.0-rc1
        version: 1.8.0-rc1
      - name: cert-manager.v1.7.2
        version: 1.7.2
      - name: cert-manager.v1.7.1
        version: 1.7.1
      - name: cert-manager.v1.7.1-rc1
        version: 1.7.1-rc1
      - name: cert-manager.v1.6.2
        version: 1.6.2
      - name: cert-manager.v1.6.2-rc1
        version: 1.6.2-rc1
      name: candidate
    - currentCSV: cert-manager.v1.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "example-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "example-clusterissuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example-certificate"
                },
                "spec": {
                  "dnsNames": [
                    "example.com"
                  ],
                  "issuerRef": {
                    "name": "example-issuer"
                  },
                  "secretName": "example-certificate-tls"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "name": "example-certificaterequest"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "name": "example-order"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "example-challenge"
                },
                "spec": {}
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: quay.io/jetstack/cert-manager-controller:v1.16.5
          createdAt: 2025-05-02T09:11:31
          olm.skipRange: '>=1.16.0 <1.16.5'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/internal-objects: |-
            [
              "challenges.acme.cert-manager.io",
              "orders.acme.cert-manager.io"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/cert-manager/cert-manager
          support: The cert-manager maintainers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A CertificateRequest is used to request a signed certificate from one of the
              configured issuers.

              All fields within the CertificateRequest's `spec` are immutable after creation.
              A CertificateRequest will either succeed or fail, as denoted by its `Ready` status
              condition and its `status.failureTime` field.

              A CertificateRequest is a one-shot resource, meaning it represents a single
              point in time request for a certificate and cannot be re-used.
            displayName: CertificateRequest
            kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - description: |-
              A Certificate resource should be created to ensure an up to date and signed
              X.509 certificate is stored in the Kubernetes Secret resource named in `spec.secretName`.

              The stored certificate will be renewed before it expires (as configured by `spec.renewBefore`).
            displayName: Certificate
            kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: Challenge is a type to represent a Challenge request with
              an ACME server
            displayName: (Internal) Challenge
            kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - description: |-
              A ClusterIssuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is similar to an Issuer, however it is cluster-scoped and therefore can
              be referenced by resources that exist in *any* namespace, not just the same
              namespace as the referent.
            displayName: ClusterIssuer
            kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - description: |-
              An Issuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is scoped to a single namespace and can therefore only be referenced by
              resources within the same namespace.
            displayName: Issuer
            kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - description: Order is a type to represent an Order with an ACME server
            displayName: (Internal) Order
            kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          cert-manager is a Kubernetes add-on to automate the management and issuance of TLS certificates from various issuing sources.
          It will ensure certificates are valid and up to date periodically, and attempt to renew certificates at an appropriate time before expiry.

          cert-manager has the following [built in issuers][]:

          * ACME: sign certificates using Let's Encrypt (or any other [ACME][] compliant signer).
          * Vault: sign certificates using HashiCorp Vault.
          * Venafi: sign certificates using Venafi Cloud or Venafi Trust Protection Platform.
          * SelfSigned: the private key of the certificate will be used to sign the certificate.
          * CA: a private key stored as a Kubernetes Secret will be used to sign the certificate.

          [built in issuers]: https://cert-manager.io/docs/configuration/#supported-issuer-types
          [ACME]: https://datatracker.ietf.org/doc/html/rfc8555

          cert-manager also supports [external issuers][],
          which are out-of-tree components that provide support for other certificate authorities, such as:
          [Google Cloud Certificate Authority](https://github.com/jetstack/google-cas-issuer),
          [AWS Private Certificate Authority](https://github.com/cert-manager/aws-privateca-issuer), and
          [Cloudflare Origin CA](https://github.com/cloudflare/origin-ca-issuer)

          [external issuers]: https://cert-manager.io/docs/configuration/external/

          In order to configure cert-manager to issue certificates, you must first create an Issuer or ClusterIssuer.
          These resources represent a particular signing authority and tell cert-manager how to handle your certificate requests.
          The [Issuer Documentation][] has more detail.

          [configure cert-manager]: https://cert-manager.io/docs/configuration/
          [Issuer Documentation]: https://cert-manager.io/docs/concepts/issuer/

          Once an Issuer has been configured, you can create a Certificate resource,
          which is a definition of a certificate request that is to be honored by an issuer
          and which is to be kept up-to-date.
          This is the usual way that you will interact with cert-manager to request signed certificates.
          The [Certificate Documentation][] has more detail.

          [Certificate Documentation]: https://cert-manager.io/docs/usage/certificate/
        displayName: cert-manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Documentation
          url: https://cert-manager.io/
        - name: Release Notes
          url: https://cert-manager.io/docs/release-notes
        maintainers:
        - email: cert-manager-maintainers@googlegroups.com
          name: The cert-manager maintainers
        maturity: stable
        minKubeVersion: 1.19.0-0
        provider:
          name: The cert-manager maintainers
          url: https://cert-manager.io/
        relatedImages:
        - quay.io/jetstack/cert-manager-controller:v1.16.5
        - quay.io/jetstack/cert-manager-cainjector:v1.16.5
        - quay.io/jetstack/cert-manager-webhook:v1.16.5
        version: 1.16.5
      entries:
      - name: cert-manager.v1.16.5
        version: 1.16.5
      - name: cert-manager.v1.16.1
        version: 1.16.1
      - name: cert-manager.v1.15.2
        version: 1.15.2
      - name: cert-manager.v1.15.0
        version: 1.15.0
      - name: cert-manager.v1.14.2
        version: 1.14.2
      - name: cert-manager.v1.13.3
        version: 1.13.3
      - name: cert-manager.v1.13.1
        version: 1.13.1
      - name: cert-manager.v1.13.1-rc1
        version: 1.13.1-rc1
      - name: cert-manager.v1.12.2
        version: 1.12.2
      - name: cert-manager.v1.11.4
        version: 1.11.4
      - name: cert-manager.v1.11.0
        version: 1.11.0
      - name: cert-manager.v1.10.2
        version: 1.10.2
      - name: cert-manager.v1.10.1
        version: 1.10.1
      - name: cert-manager.v1.10.0
        version: 1.10.0
      - name: cert-manager.v1.9.1
        version: 1.9.1
      - name: cert-manager.v1.8.2
        version: 1.8.2
      - name: cert-manager.v1.8.0
        version: 1.8.0
      - name: cert-manager.v1.7.2
        version: 1.7.2
      - name: cert-manager.v1.7.1
        version: 1.7.1
      - name: cert-manager.v1.6.2
        version: 1.6.2
      - name: cert-manager.v1.6.1
        version: 1.6.1
      - name: cert-manager.v1.6.0
        version: 1.6.0
      - name: cert-manager.v1.5.4
        version: 1.5.4
      - name: cert-manager.v1.5.3
        version: 1.5.3
      - name: cert-manager.v1.4.4
        version: 1.4.4
      - name: cert-manager.v1.4.3
        version: 1.4.3
      - name: cert-manager.v1.4.2
        version: 1.4.2
      - name: cert-manager.v1.4.1
        version: 1.4.1
      - name: cert-manager.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: cert-manager
    provider:
      name: The cert-manager maintainers
      url: https://cert-manager.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: TrendMicro
      provider-url: trendmicro.com
    name: visionone-containersecurity
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: visionone-containersecurity.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "visionone-containersecurity-sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          com.redhat.openshift.versions: v4.12-v4.20
          containerImage: public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
          createdAt: "2025-09-18T22:43:41Z"
          description: Trend Vision One™ Container Security delivers full lifecycle
            protection for containers with real-time threat detection, policy enforcement,
            compliance assurance, and rapid incident response.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trendmicro-system
          operatorhub.io/ui-metadata-max-k8s-version: "1.26"
          operators.openshift.io/valid-subscription: VisionOneContainerSecurity Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trendmicro/visionone-container-security-helm
          support: TrendMicro Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trend Vision One™ Container Security delivers full lifecycle
              protection for containers with real-time threat detection, policy enforcement,
              compliance assurance, and rapid incident response.
            displayName: VisionOneContainerSecurity
            kind: VisionOneContainerSecurity
            name: visiononecontainersecurities.container-security.trendmicro.com
            version: v1alpha1
        description: "[Trend Vision One™ Container Security](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-container-security)
          delivers comprehensive, end-to-end\n protection for containerized workloads
          across OpenShift and Kubernetes environments.\n It continuously monitors
          for threats, vulnerabilities, policy violations, and\n exposed secrets ensuring
          compliance and enabling rapid, automated response. Purpose-built\n for hybrid
          cloud, it instantly isolates compromised containers to prevent lateral\n
          movement and minimize impact. Seamlessly integrating into OpenShift DevSecOps\n
          pipelines, it empowers security teams to enforce policies while enabling
          developers\n to innovate and deploy with confidence without disrupting workflows
          or compromising\n agility.\n\n **Installation**\n \n - Use the recommended
          `trendmicro-system` namespace.\n - Refer to the [Trend Vision One™ Documentation](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator)
          for registration and installation steps.\n"
        displayName: Trend Vision One™ Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trendmicro
        - visionone
        - container-security
        links:
        - name: Documentation
          url: https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator
        maintainers:
        - email: support@trendmicro.com
          name: TrendMicro VisionOne Support Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: TrendMicro
          url: trendmicro.com
        relatedImages:
        - public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
        version: 0.0.2
      entries:
      - name: visionone-containersecurity.v0.0.2
        version: 0.0.2
      - name: visionone-containersecurity.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: visionone-containersecurity.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "visionone-containersecurity-sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          com.redhat.openshift.versions: v4.12-v4.20
          containerImage: public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
          createdAt: "2025-09-18T22:43:41Z"
          description: Trend Vision One™ Container Security delivers full lifecycle
            protection for containers with real-time threat detection, policy enforcement,
            compliance assurance, and rapid incident response.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trendmicro-system
          operatorhub.io/ui-metadata-max-k8s-version: "1.26"
          operators.openshift.io/valid-subscription: VisionOneContainerSecurity Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trendmicro/visionone-container-security-helm
          support: TrendMicro Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trend Vision One™ Container Security delivers full lifecycle
              protection for containers with real-time threat detection, policy enforcement,
              compliance assurance, and rapid incident response.
            displayName: VisionOneContainerSecurity
            kind: VisionOneContainerSecurity
            name: visiononecontainersecurities.container-security.trendmicro.com
            version: v1alpha1
        description: "[Trend Vision One™ Container Security](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-container-security)
          delivers comprehensive, end-to-end\n protection for containerized workloads
          across OpenShift and Kubernetes environments.\n It continuously monitors
          for threats, vulnerabilities, policy violations, and\n exposed secrets ensuring
          compliance and enabling rapid, automated response. Purpose-built\n for hybrid
          cloud, it instantly isolates compromised containers to prevent lateral\n
          movement and minimize impact. Seamlessly integrating into OpenShift DevSecOps\n
          pipelines, it empowers security teams to enforce policies while enabling
          developers\n to innovate and deploy with confidence without disrupting workflows
          or compromising\n agility.\n\n **Installation**\n \n - Use the recommended
          `trendmicro-system` namespace.\n - Refer to the [Trend Vision One™ Documentation](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator)
          for registration and installation steps.\n"
        displayName: Trend Vision One™ Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trendmicro
        - visionone
        - container-security
        links:
        - name: Documentation
          url: https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator
        maintainers:
        - email: support@trendmicro.com
          name: TrendMicro VisionOne Support Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: TrendMicro
          url: trendmicro.com
        relatedImages:
        - public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
        version: 0.0.2
      entries:
      - name: visionone-containersecurity.v0.0.2
        version: 0.0.2
      name: stable
    defaultChannel: stable
    packageName: visionone-containersecurity
    provider:
      name: TrendMicro
      url: trendmicro.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Docling Maintainers
      provider-url: https://docling-project.github.io/docling/
    name: docling-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: docling-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2025-05-19T20:20:14Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DoclingServe is the Schema for the doclingserves API
            displayName: Docling Serve
            kind: DoclingServe
            name: doclingserves.docling.github.io
            version: v1alpha1
        description: |-
          **Overview**

          The Docling Operator is responsible for managing the full lifecycle of Docling Serve instances in your Kubernetes or OpenShift container platforms.


          **Docling Serve**

          The Docling Serve application deploys a REST API for consuming and integrating Docling in your workflows.

          The API offers endpoints for multiple types of interaction, i.e. both synchronous and asynchronous endpoints, as well as status subscriptions via websockets.

          The options for the Docling conversion can be customized at runtime by the users (e.g. which OCR engine, which enrichment models, etc). Users can also choose which of the many DoclingDocument export formats (JSON, Markdown, HTML, DocTags, etc) to receive in the result response.

          Docling Serve allows to choose between a local compute engine, i.e. running Docling in the same container as the API server, or a distributed compute engine which leverages Kubeflow Pipelines to parallelize large-scale ingestion jobs.

          **Docling**

          Docling is a powerful library which simplifies document processing, parsing diverse formats — including advanced PDF understanding — and providing seamless integrations with the gen AI ecosystem.

          Learn more about the Docling Project by visiting the GitHub organization at [https://github.com/docling-project](https://github.com/docling-project).


          **Sample DoclingServe Custom Resources**


          To deploy with the `local` compute engine with the UI enabled and Route exposed:
          ```yaml
          apiVersion: docling.github.io/v1alpha1
          kind: DoclingServe
          metadata:
            labels:
              app.kubernetes.io/name: docling-operator
              app.kubernetes.io/managed-by: kustomize
            name: doclingserve-local-sample
          spec:
            apiServer:
              image: "quay.io/docling-project/docling-serve:latest"
              enableUI: true
              instances: 1
            engine:
              local:
                numWorkers: 2
            route:
              enabled: true
          ```

          To deploy with the `KFP` compute engine with the UI enabled and Route exposed. **Note: Kubeflow Pipeline must be installed and setup in the cluster to use this option.**:
          ```yaml
          apiVersion: docling.github.io/v1alpha1
          kind: DoclingServe
          metadata:
            labels:
              app.kubernetes.io/name: docling-operator
              app.kubernetes.io/managed-by: kustomize
            name: doclingserve-kfp-sample
          spec:
            apiServer:
              image: "quay.io/docling-project/docling-serve:latest"
              enableUI: true
              instances: 1
            engine:
              kfp:
                endpoint: "https://NAME.NAMESPACE.svc.cluster.local:8888"
            route:
              enabled: true
          ```

          **Additional Configurations**


          The CRD exposes `apiServer.configMapName` to enable additional configurations of the `docling-serve` workload. For a full list of configurations can be found on docling-serve's
          [Configuration](https://github.com/docling-project/docling-serve/blob/main/docs/configuration.md) page.
        displayName: Docling Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - document
        - processing
        links:
        - name: Docling Operator
          url: https://docling-project.github.io/docling/
        maintainers:
        - email: docling-technical-discuss@lists.lfaidata.foundation
          name: Docling Maintainers
        maturity: alpha
        provider:
          name: Docling Maintainers
          url: https://docling-project.github.io/docling/
        relatedImages:
        - quay.io/docling-project/docling-operator@sha256:200d8db51874c4b1e3d20d8d1423e9a4807843fbb29f514cd5d8ba872e47f4af
        version: 0.0.1
      entries:
      - name: docling-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: docling-operator
    provider:
      name: Docling Maintainers
      url: https://docling-project.github.io/docling/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubemq.io
      provider-url: ""
    name: kubemq-operator-marketplace
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubemq-operator-marketplace.1.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
              "apiVersion": "core.k8s.kubemq.io/v1beta1",
              "kind": "KubemqCluster",
              "metadata": {
                "name": "kubemq-cluster"
              },
              "spec": {
                "key": "<Your-KubeMQ-License-Key>",
                "replicas": 3
              }
            },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqConnector",
                "metadata": {
                  "name": "kubemq-bridges"
                },
                "spec": {
                  "type": "bridges",
                  "config": "bindings: null"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.connect.redhat.com/kubemq/kubemq-operator@sha256:3a9c87e462eb1caa81925dee72407f73fca41f19a3f9bcaf4e76e569db1dfbab
          createdAt: 2020-03-11 12:00:00+00:00
          description: Kubernetes Messaging Platform
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          support: support@kubemq.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kubemq Cluster
            displayName: Kubemq Cluster
            kind: KubemqCluster
            name: kubemqclusters.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Connector
            displayName: Kubemq Connector
            kind: KubemqConnector
            name: kubemqconnectors.core.k8s.kubemq.io
            version: v1beta1
        description: |2-

          ## Kubernetes Messaging Platform
          A message-based architecture for microservices providing flexible connectivity for hybrid environments: multi-cloud, on-premises, and at the edge. Powered by enterprise-grade message broker and message queue with ready to use connectors, bridges, and control center.

          Messaging platform components:
            - **KubeMQ Cluster** - Enterprise-grade message broker and message queue, scalable, high available, and secured.
            - **KubeMQ Targets Connector** - connects a KubeMQ cluster with external systems and cloud services.
            - **KubeMQ Bridges Connector** - bridge, replicate, aggregate, and transform messages between KubeMQ clusters.
            - **KubeMQ Sources Connector** - transform external systems and cloud services data to KubeMQ cluster messages.
          \n### Kubernetes Native
          Innovative and modern message queue and message broker in a lightweight container developed to run in Kubernetes, certified in the CNCF landscape, and connect natively to the cloud-native ecosystem.

          ### Ease of use

          Simple deployment in Kubernetes in less than 1 minute. Developer friendly by simple to use SDKs and elimination of the many developers and DevOps-centered challenges to define exchanges, brokers, channels, routes, and predefined topics

          ### All messaging patterns

          KubeMQ is available with all messaging patterns such as Queue, Stream, Pub/Sub, and RPC, Saving the need to maintain multiple messaging platforms for different use cases.

          ### Hybrid Environments

          KubeMQ enables the microservices from multiple environments to communicate and build one hybrid infrastructure solution across clouds, on-prem, and at the edge.

          ### KubeMQ Cluster key features

          - All-batteries included Messaging Queue Broker for the Kubernetes environment
          - Deployed with Operator for full life cycle operation
          - Blazing fast (written in Go), small and lightweight Docker container
          - Asynchronous and Synchronous messaging with support for  `At Most Once Delivery` and `At Least Once Delivery` models
          - Supports durable FIFO based Queue, Publish-Subscribe Events, Publish-Subscribe with Persistence (Events Store), RPC Command and Query messaging patterns
          - Supports gRPC, Rest and WebSocket Transport protocols with TLS support (both RPC and Stream modes)
          - Supports Access control Authorization and Authentication
          - Supports message masticating and smart routing
          - No Message broker configuration needed (i.e., queues, exchanges)
          - .Net, Java, Python, Go, and NodeJS SDKs

          ### KubeMQ Targets Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major services types (databases, cache, messaging, serverless, HTTP, etc.)
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-targets)

          ### KubeMQ Bridges Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Build Any Topology - connects KubeMQ clusters in 1:1, 1:n , n:1, n:n
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-bridges)

          ### KubeMQ Sources Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major messaging services such as Kafka, ActiveMQ, SQS, and more.
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-sources)

          ## Getting Started

          [Start with OpenShift](https://docs.kubemq.io/getting-started/create-cluster/openshift)

          ## Get Enterprise Trial License

          [Register](https://account.kubemq.io/login/register)

          ## Request a Demo

          [See KubeMQ in Action](https://kubemq.io/request-a-demo/)
        displayName: Kubemq Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - messaging
        - queue
        - broker
        - pubsub
        - rpc
        links:
        - name: Learn more about Kubemq
          url: https://kubemq.io/
        - name: Register for free license
          url: https://account.kubemq.io/login/register
        - name: Contact Sales
          url: https://kubemq.io/contact-sales/
        - name: Documentation
          url: https://docs.kubemq.io/
        - name: Get Help
          url: https://kubemq.io/contact-support/
        maintainers:
        - email: support@kubemq.io
          name: kubemq
        maturity: alpha
        provider:
          name: Kubemq.io
        relatedImages:
        - registry.connect.redhat.com/kubemq/kubemq-operator@sha256:3a9c87e462eb1caa81925dee72407f73fca41f19a3f9bcaf4e76e569db1dfbab
        - registry.connect.redhat.com/kubemq/kubemq-uni@sha256:51154b2a112c5cdc17ee8d338524c02ccd995f9386180b1cc232b7a875a23e73
        - registry.connect.redhat.com/kubemq/kubemq-targets@sha256:2a09e1718a03729a121728e4e3683e7e1f0690789d7f9857d76bebf1cc01d436
        - registry.connect.redhat.com/kubemq/kubemq-sources@sha256:5736575b6cae39d129346f83fe2d8ab9cdbd1c1d2a99a66b4b9c01631fb5db21
        - registry.connect.redhat.com/kubemq/kubemq-sources@sha256:8ed531e58c60760888ffb18165883386bda9cc5624645353406c94750f5c3008
        version: 1.8.6
      entries:
      - name: kubemq-operator-marketplace.1.8.6
        version: 1.8.6
      - name: kubemq-operator-marketplace.1.8.5
        version: 1.8.5
      - name: kubemq-operator-marketplace.v1.8.4
        version: 1.8.4
      - name: kubemq-operator-marketplace.v1.8.3
        version: 1.8.3
      name: alpha
    defaultChannel: alpha
    packageName: kubemq-operator-marketplace
    provider:
      name: Kubemq.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: candidate
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: fast
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: stable
    defaultChannel: stable
    packageName: stackgres-certified
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzounstructured-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzounstructured-operator.v2.0.101
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzounstructured.clusters.cambridgesemantics.com/v1",
                "kind": "AnzoUnstructured",
                "metadata": {
                  "name": "au01"
                },
                "spec": {
                  "auWorker": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-w",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-unstructured-worker@sha256:4db61bdf6dc8b42824a06d2f8d8aeecddde765d123b40a63a7bbf87845719fb2",
                                "name": "w",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  },
                  "msLeader": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-ms",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-microservices-leader@sha256:0579fd97e81e436eaff3b58c5395f13aa924940a3d51f0bdea5163d327620286",
                                "name": "ms",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/unstructured-operator@sha256:a47af068a7aa55c12be84c80d263feeed5597ecc434986b4f2fff782b568bb04
          createdAt: "2021-08-02"
          description: Kubernetes Operator for Anzo Unstructured
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzounstructured-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzounstructured-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoUnstructured is the Schema for the anzounstructureds
              API
            displayName: AnzoUnstructured Operator
            kind: AnzoUnstructured
            name: anzounstructureds.anzounstructured.clusters.cambridgesemantics.com
            version: v1
        description: |-
          The Anzo Unstructured Operator provides the way to install and configure an anzo unstructured setup on Red Hat K8S environment.
          Currently, this is possible only through existing Anzo installation.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v2.0.0/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers through [ Cambridge Semantics Customer Center ]( https://customercenter.cambridgesemantics.com/ ).
        displayName: Anzo Unstructured Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Database
        - Big Data
        links:
        - name: Anzo Unstructured Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v2.0.0/README.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/unstructured-operator@sha256:a47af068a7aa55c12be84c80d263feeed5597ecc434986b4f2fff782b568bb04
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-microservices-leader@sha256:0579fd97e81e436eaff3b58c5395f13aa924940a3d51f0bdea5163d327620286
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-unstructured-worker@sha256:4db61bdf6dc8b42824a06d2f8d8aeecddde765d123b40a63a7bbf87845719fb2
        version: 2.0.0
      entries:
      - name: anzounstructured-operator.v2.0.101
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anzounstructured-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudwatchlogs-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudwatchlogs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudwatchlogs.services.k8s.aws/v1alpha1",
                "kind": "LogGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudwatchlogs-controller:1.1.1
          createdAt: "2025-09-19T20:51:16Z"
          description: AWS CloudWatch Logs controller is a service controller for
            managing CloudWatch Logs resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LogGroup represents the state of an AWS cloudwatchlogs LogGroup
              resource.
            displayName: LogGroup
            kind: LogGroup
            name: loggroups.cloudwatchlogs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudWatch Logs resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudWatch Logs**

          You can use Amazon CloudWatch Logs to monitor, store, and access your log files from Amazon Elastic Compute Cloud (Amazon EC2) instances, AWS CloudTrail, Route 53, and other sources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudWatch Logs
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudwatchlogs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudWatch Logs Developer Resources
          url: https://aws.amazon.com/cloudwatch/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudwatch maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudwatchlogs-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-cloudwatchlogs-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudwatchlogs-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudwatchlogs-controller.v1.0.14
        version: 1.0.14
      - name: ack-cloudwatchlogs-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudwatchlogs-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudwatchlogs-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudwatchlogs-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudwatchlogs-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudwatchlogs-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudwatchlogs-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudwatchlogs-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudwatchlogs-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudwatchlogs-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudwatchlogs-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudwatchlogs-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudwatchlogs-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudwatchlogs-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudwatchlogs-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudwatchlogs-controller.v0.0.7
        version: 0.0.7
      - name: ack-cloudwatchlogs-controller.v0.0.4
        version: 0.0.4
      - name: ack-cloudwatchlogs-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudwatchlogs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Palo Alto Networks
      provider-url: ""
    name: pcc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pcc-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "name": "pcc-console",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "ConsoleDefender",
                "metadata": {
                  "name": "pcc-consoledefender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "defenderConfig": {
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Defender",
                "metadata": {
                  "name": "pcc-defender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "defenderConfig": {
                    "clusterAddress": "twistlock-console.example.com",
                    "consoleAddress": "https://twistlock-console.example.com:8083",
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: quay.io/prismacloud/pcc-operator@sha256:b8fcfbd6c51286c874e00db1bd35523386cec406fa4050ef44c0a887730cf9b8
          createdAt: "2021-11-15"
          description: Deploy Prisma Cloud Compute for cloud-native security in your
            clusters
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator
          support: pcc-integrations@paloaltonetworks.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Install the Prisma Cloud Compute Console and Defenders
            displayName: Console and Defenders
            kind: ConsoleDefender
            name: consoledefenders.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install the Prisma Cloud Compute Console
            displayName: Console
            kind: Console
            name: consoles.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install Prisma Cloud Compute Defenders
            displayName: Defenders
            kind: Defender
            name: defenders.pcc.paloaltonetworks.com
            version: v1alpha1
        description: |
          ## Features
          - Deploy Console
              - Create initial local administrator user
              - Add license
          - Deploy Defenders
          - Support for offline deployments
          - Upgrade Console
          - Upgrade Defenders

          ## Installation
          See the [Prisma Cloud Compute Operator documentation](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator) for installation guides.

          ## Support
          This operator is available "as is," and should be seen as community-supported; however, Palo Alto Networks will continue contributing whenever possible.
          Palo Alto Networks does not provide technical support or help with using or troubleshooting the operator through normal support channels.
          We encourage you to open GitHub [issues](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/issues) and [pull requests](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/pulls) to track bugs and feature requests.

          ## Other links
          [Prisma Cloud Compute product documentation](https://docs.paloaltonetworks.com/prisma/prisma-cloud/prisma-cloud-admin-compute.html)

          [Prisma Cloud Compute API documentation](https://prisma.pan.dev/api/cloud/cwpp)
        displayName: Prisma Cloud Compute Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - palo
        - palo alto
        - palo alto networks
        - paloalto
        - twistlock
        - compute
        - prisma
        - prisma cloud
        - prisma cloud compute
        - security
        - image
        - images
        - container
        - containers
        - docker
        - podman
        - vulnerability
        - vulnerabilities
        - compliance
        - runtime
        - scan
        - scanning
        - monitor
        - monitoring
        - alert
        - alerting
        links:
        - name: Prisma Cloud
          url: https://www.paloaltonetworks.com/prisma/cloud
        maintainers:
        - email: pcc-integrations@paloaltonetworks.com
          name: Wyatt Gill
        maturity: alpha
        provider:
          name: Palo Alto Networks
        relatedImages:
        - quay.io/prismacloud/pcc-operator@sha256:b8fcfbd6c51286c874e00db1bd35523386cec406fa4050ef44c0a887730cf9b8
        version: 0.2.0
      entries:
      - name: pcc-operator.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: pcc-operator
    provider:
      name: Palo Alto Networks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: koku-metrics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: koku-metrics-operator.v4.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
          createdAt: "2025-09-25T19:04:35Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: koku-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Cost Management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Koku Metrics Operator
          ## Introduction
          The `koku-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Koku Metrics Operator (`koku-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created KokuMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The KokuMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.2.0:
          * Enabled gathering and reporting NVIDIA GPU metrics for resource optimization.
          * (Bugfix) Fixed custom CA certificate validation and configuration issues.
          * (Bugfix) Fixed issue where ROS queries were missing workload metadata for namespaces labeled only with the legacy `insights_cost_management_optimizations` label.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **Progress towards FIPS 140 Compliance:** The operator is using the Go Cryptographic Module v1.0.0 to progress toward FIPS 140 compliance. Although this module is not validated at the time of this release, the operator aims to meet stricter security standards when the module does successfully achieve FIPS validation.
          * **API Name Change:** The Custom Resource Definition (CRD) for configuring the upstream `Koku Metrics Operator` was renamed from `KokuMetricsConfig` to `CostManagementMetricsConfig`.

          ### NOTE:
            * The API name change impacts only users of the upstream (community) Koku Metrics Operator.
            * The downstream (Red Hat-supported) Cost Management Metrics Operator is not impacted by this change and users should continue using their existing configurations.

          **Important Upgrade Instructions:**

            If you are upgrading the upstream Koku Metrics Operator to version 4.0.0 or higher, you must manually migrate your configuration. The operator will no longer recognize existing `KokuMetricsConfig` resources.

            To successfully upgrade and retain your operator's configuration, complete the following steps. You can use the provided `oc` commands replacing the angle brackets (`< >`) with your specific values:

            1. Backup existing configuration (recommended):

               Before you upgrade, retrieve the details of your current `KokuMetricsConfig` instance to help you recreate it with the new API name.

               ```
               oc get kokumetricsconfig -n koku-metrics-operator <your-config-name> -o yaml > koku-metrics-config-backup.yaml
               ```

            2. Delete the previous `KokuMetricsConfig` instance (recommended):

                After the operator is upgraded to `version 4.0.0`, delete any existing `KokuMetricsConfig` resources.

                ```
                oc delete kokumetricsconfig -n koku-metrics-operator <your-config-name>
                ```

            3. Create the new `CostManagementMetricsConfig` instance:

                Using the information from your backup (or a new configuration), create a new `CostManagementMetricsConfig` instance. Adjust the spec section based on your specific needs. For more information about configurations, [refer to the configurable parameters](#configurable-parameters).

            4. Apply your new configuration:

                ```
                oc apply -f <your-costmanagementmetricsconfig-file>.yaml -n koku-metrics-operator
                ```

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created KokuMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the KokuMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `KokuMetricsConfig`, creating a new `KokuMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `koku-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: koku-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the KokuMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the KokuMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon KokuMetricsConfig creation** (This parameter will start to appear in KokuMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the KokuMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the koku-metrics-operator
          **Note** There are separate instructions for configuring the `koku-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `koku-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the koku-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `koku-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `koku-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the koku-metrics-operator for a restricted network
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `koku-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: koku-metrics-operator
                labels:
                  app: koku-metrics-operator
              spec:
                volumes:
                - name: koku-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: koku-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: koku-metrics-operator-reports
                    mountPath: /tmp/koku-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/koku-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/koku-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `koku-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the KokuMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Koku Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Koku Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
        version: 4.2.0
      entries:
      - name: koku-metrics-operator.v4.2.0
        version: 4.2.0
      - name: koku-metrics-operator.v4.1.0
        version: 4.1.0
      - name: koku-metrics-operator.v4.0.0
        version: 4.0.0
      - name: koku-metrics-operator.v3.3.2
        version: 3.3.2
      - name: koku-metrics-operator.v3.3.1
        version: 3.3.1
      - name: koku-metrics-operator.v3.3.0
        version: 3.3.0
      - name: koku-metrics-operator.v3.2.1
        version: 3.2.1
      - name: koku-metrics-operator.v3.2.0
        version: 3.2.0
      - name: koku-metrics-operator.v3.1.0
        version: 3.1.0
      - name: koku-metrics-operator.v3.0.1
        version: 3.0.1
      - name: koku-metrics-operator.v3.0.0
        version: 3.0.0
      - name: koku-metrics-operator.v2.0.0
        version: 2.0.0
      - name: koku-metrics-operator.v1.2.0
        version: 1.2.0
      - name: koku-metrics-operator.v1.1.9
        version: 1.1.9
      - name: koku-metrics-operator.v1.1.8
        version: 1.1.8
      - name: koku-metrics-operator.v1.1.7
        version: 1.1.7
      - name: koku-metrics-operator.v1.1.6
        version: 1.1.6
      - name: koku-metrics-operator.v1.1.5
        version: 1.1.5
      - name: koku-metrics-operator.v1.1.4
        version: 1.1.4
      - name: koku-metrics-operator.v1.1.3
        version: 1.1.3
      - name: koku-metrics-operator.v1.1.2
        version: 1.1.2
      - name: koku-metrics-operator.v1.1.1
        version: 1.1.1
      - name: koku-metrics-operator.v0.9.8
        version: 0.9.8
      - name: koku-metrics-operator.v0.9.7
        version: 0.9.7
      - name: koku-metrics-operator.v0.9.6
        version: 0.9.6
      - name: koku-metrics-operator.v0.9.5
        version: 0.9.5
      - name: koku-metrics-operator.v0.9.4
        version: 0.9.4
      - name: koku-metrics-operator.v0.9.3
        version: 0.9.3
      - name: koku-metrics-operator.v0.9.2
        version: 0.9.2
      - name: koku-metrics-operator.v0.9.1
        version: 0.9.1
      - name: koku-metrics-operator.v0.9.0
        version: 0.9.0
      name: alpha
    - currentCSV: koku-metrics-operator.v4.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
          createdAt: "2025-09-25T19:04:35Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: koku-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Cost Management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Koku Metrics Operator
          ## Introduction
          The `koku-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Koku Metrics Operator (`koku-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created KokuMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The KokuMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.2.0:
          * Enabled gathering and reporting NVIDIA GPU metrics for resource optimization.
          * (Bugfix) Fixed custom CA certificate validation and configuration issues.
          * (Bugfix) Fixed issue where ROS queries were missing workload metadata for namespaces labeled only with the legacy `insights_cost_management_optimizations` label.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **Progress towards FIPS 140 Compliance:** The operator is using the Go Cryptographic Module v1.0.0 to progress toward FIPS 140 compliance. Although this module is not validated at the time of this release, the operator aims to meet stricter security standards when the module does successfully achieve FIPS validation.
          * **API Name Change:** The Custom Resource Definition (CRD) for configuring the upstream `Koku Metrics Operator` was renamed from `KokuMetricsConfig` to `CostManagementMetricsConfig`.

          ### NOTE:
            * The API name change impacts only users of the upstream (community) Koku Metrics Operator.
            * The downstream (Red Hat-supported) Cost Management Metrics Operator is not impacted by this change and users should continue using their existing configurations.

          **Important Upgrade Instructions:**

            If you are upgrading the upstream Koku Metrics Operator to version 4.0.0 or higher, you must manually migrate your configuration. The operator will no longer recognize existing `KokuMetricsConfig` resources.

            To successfully upgrade and retain your operator's configuration, complete the following steps. You can use the provided `oc` commands replacing the angle brackets (`< >`) with your specific values:

            1. Backup existing configuration (recommended):

               Before you upgrade, retrieve the details of your current `KokuMetricsConfig` instance to help you recreate it with the new API name.

               ```
               oc get kokumetricsconfig -n koku-metrics-operator <your-config-name> -o yaml > koku-metrics-config-backup.yaml
               ```

            2. Delete the previous `KokuMetricsConfig` instance (recommended):

                After the operator is upgraded to `version 4.0.0`, delete any existing `KokuMetricsConfig` resources.

                ```
                oc delete kokumetricsconfig -n koku-metrics-operator <your-config-name>
                ```

            3. Create the new `CostManagementMetricsConfig` instance:

                Using the information from your backup (or a new configuration), create a new `CostManagementMetricsConfig` instance. Adjust the spec section based on your specific needs. For more information about configurations, [refer to the configurable parameters](#configurable-parameters).

            4. Apply your new configuration:

                ```
                oc apply -f <your-costmanagementmetricsconfig-file>.yaml -n koku-metrics-operator
                ```

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created KokuMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the KokuMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `KokuMetricsConfig`, creating a new `KokuMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `koku-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: koku-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the KokuMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the KokuMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon KokuMetricsConfig creation** (This parameter will start to appear in KokuMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the KokuMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the koku-metrics-operator
          **Note** There are separate instructions for configuring the `koku-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `koku-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the koku-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `koku-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `koku-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the koku-metrics-operator for a restricted network
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `koku-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: koku-metrics-operator
                labels:
                  app: koku-metrics-operator
              spec:
                volumes:
                - name: koku-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: koku-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: koku-metrics-operator-reports
                    mountPath: /tmp/koku-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/koku-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/koku-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `koku-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the KokuMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Koku Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Koku Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
        version: 4.2.0
      entries:
      - name: koku-metrics-operator.v4.2.0
        version: 4.2.0
      - name: koku-metrics-operator.v4.1.0
        version: 4.1.0
      - name: koku-metrics-operator.v4.0.0
        version: 4.0.0
      - name: koku-metrics-operator.v3.3.2
        version: 3.3.2
      - name: koku-metrics-operator.v3.3.1
        version: 3.3.1
      - name: koku-metrics-operator.v3.3.0
        version: 3.3.0
      - name: koku-metrics-operator.v3.2.1
        version: 3.2.1
      - name: koku-metrics-operator.v3.2.0
        version: 3.2.0
      - name: koku-metrics-operator.v3.1.0
        version: 3.1.0
      - name: koku-metrics-operator.v3.0.1
        version: 3.0.1
      - name: koku-metrics-operator.v3.0.0
        version: 3.0.0
      - name: koku-metrics-operator.v2.0.0
        version: 2.0.0
      - name: koku-metrics-operator.v1.2.0
        version: 1.2.0
      - name: koku-metrics-operator.v1.1.9
        version: 1.1.9
      - name: koku-metrics-operator.v1.1.8
        version: 1.1.8
      - name: koku-metrics-operator.v1.1.7
        version: 1.1.7
      - name: koku-metrics-operator.v1.1.6
        version: 1.1.6
      - name: koku-metrics-operator.v1.1.5
        version: 1.1.5
      - name: koku-metrics-operator.v1.1.4
        version: 1.1.4
      - name: koku-metrics-operator.v1.1.3
        version: 1.1.3
      - name: koku-metrics-operator.v1.1.2
        version: 1.1.2
      - name: koku-metrics-operator.v1.1.1
        version: 1.1.1
      - name: koku-metrics-operator.v0.9.8
        version: 0.9.8
      - name: koku-metrics-operator.v0.9.7
        version: 0.9.7
      - name: koku-metrics-operator.v0.9.6
        version: 0.9.6
      - name: koku-metrics-operator.v0.9.5
        version: 0.9.5
      - name: koku-metrics-operator.v0.9.4
        version: 0.9.4
      - name: koku-metrics-operator.v0.9.3
        version: 0.9.3
      - name: koku-metrics-operator.v0.9.2
        version: 0.9.2
      - name: koku-metrics-operator.v0.9.1
        version: 0.9.1
      - name: koku-metrics-operator.v0.9.0
        version: 0.9.0
      name: beta
    defaultChannel: beta
    packageName: koku-metrics-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Lenovo (Shanghai) Computer Technology Co., Ltd.
      provider-url: ""
    name: lenovo-locd-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lenovo-locd-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "locd.lenovo.com/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample",
                  "namespace": "locd-operator-system"
                },
                "spec": {
                  "affinity": {},
                  "argocd": {
                    "password": "xxxxxxxxxxx",
                    "server": "argocd.svc.cluster.local",
                    "user": "admin"
                  },
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "core": {
                    "image": "registry.connect.redhat.com/loc-devops/core@sha256:f771217b3f7849e1b63cc6abb93b8b0fc9636fd4a201e31807a15f773138c609",
                    "pullPolicy": "Always"
                  },
                  "fullnameOverride": "",
                  "global": {
                    "managementNamespace": "loc-devops",
                    "serviceEndpoint": "loc-devops.domain",
                    "serviceaccount": {
                      "annotations": {
                        "managedBy": "LOC DevOps"
                      },
                      "create": true,
                      "name": "loc-devops-admin"
                    }
                  },
                  "gui": {
                    "image": "registry.connect.redhat.com/loc-devops/gui@sha256:65142a36095188e879467038936b7b6b6ee39dcea6d42a437fa7066fb2bc0f44",
                    "pullPolicy": "Always"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "className": "nginx",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "loc-devops.domain",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "Prefix",
                            "servicePort": 8080
                          },
                          {
                            "path": "/api",
                            "pathType": "Prefix",
                            "servicePort": 8000
                          },
                          {
                            "path": "/swagger",
                            "pathType": "Prefix",
                            "servicePort": 8000
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "replicaCount": 1,
                  "resources": {},
                  "securityContext": {},
                  "service": {
                    "port": {
                      "core": 8000,
                      "gui": 8080
                    },
                    "type": "ClusterIP"
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Cluster
            name: clusters.locd.lenovo.com
            version: v1alpha1
        description: This is operator used to manage components of Lenovo Open Cloud
          DevOps
        displayName: lenovo-locd-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Application management
        links:
        - name: LOCD Operator
          url: https://locd-operator.domain
        maintainers:
        - email: xujt5@lenovo.com
          name: jintaoxu
        maturity: alpha
        provider:
          name: Lenovo (Shanghai) Computer Technology Co., Ltd.
        relatedImages:
        - registry.connect.redhat.com/loc-devops/gui@sha256:65142a36095188e879467038936b7b6b6ee39dcea6d42a437fa7066fb2bc0f44
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b5786bbbef725badf3dfcc2c2c7a86ead5ebb584c978c47aae8b9a62e241b80d
        - registry.connect.redhat.com/loc-devops/operator@sha256:4c17a3688ae3098e4b5c2c0530f49b1da8dedcc6fc1d3cd848513a1f189812a7
        - registry.connect.redhat.com/loc-devops/core@sha256:f771217b3f7849e1b63cc6abb93b8b0fc9636fd4a201e31807a15f773138c609
        version: 0.1.0
      entries:
      - name: lenovo-locd-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: lenovo-locd-operator
    provider:
      name: Lenovo (Shanghai) Computer Technology Co., Ltd.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtr-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtr-operator.1.2.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
             {
               "apiVersion": "windup.jboss.org/v1",
               "kind": "Windup",
               "metadata": {
                 "namespace": "mtr",
                 "name": "mtr",
                 "labels": {
                   "application": "windup"
                 }
               },
               "spec": {}
             }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mtr/mtr-rhel8-operator@sha256:4df7c4ad353339e8b958ffc651e811667767c8a0dcdd61f89352c926da52a513
          createdAt: "2024-09-03T17:22:23Z"
          description: Migration Toolkit for Runtimes is an analysis tool that supports
            the modernization and migration of Java applications.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          imagePullPolicy: Always
          operatorframework.io/suggested-namespace: mtr
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/windup/windup-operator/
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "Create an MTR Web Console application.  \n\nTo access the
              MTR Web Console application from the \"Developer\" perspective, please
              go to Topology and click on the \"Open URL\" icon on the web-console
              pod.  \n\nAlternatively from the \"Administrator\" perspective, please
              go to Networking->Routes and click on the Location hyperlink.  \n\nIn
              order to connect with default login credentials, please use \"migration\"
              as username and \"password\" as password.  \n\nKnown issue  \nIf you
              want to customize the MTR Web Console instance's parameters and you
              can not see any in the `Form view`, please switch to the `YAML view`
              and change them as needed.\n"
            displayName: Migration Toolkit for Runtimes
            kind: Windup
            name: windups.windup.jboss.org
            version: v1
        description: "The Migration Toolkit for Runtimes (MTR) is a web console application
          that supports large-scale Java application modernization and migration projects
          across a broad range of transformations and use cases.  \n\nIt analyzes
          application code, supports effort estimation, accelerates code migration,
          and enables users to move applications to containers.  \n\nFor more information
          please refer to the [https://developers.redhat.com/products/mtr/overview](https://developers.redhat.com/products/mtr/overview)
          page.\n"
        displayName: Migration Toolkit for Runtimes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - security
        - alerting
        - metric
        - troubleshooting
        - run-time
        - migration
        - modernization
        links:
        - name: Website
          url: https://developers.redhat.com/products/mtr/overview
        - name: GitHub
          url: https://github.com/windup/windup-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_runtimes
        maintainers:
        - email: migrate@redhat.com
          name: Migration Toolkit for Runtimes
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mtr/mtr-rhel8-operator@sha256:4df7c4ad353339e8b958ffc651e811667767c8a0dcdd61f89352c926da52a513
        - registry.redhat.io/mtr/mtr-web-container-rhel8@sha256:cb88a5453b506ca70563537ec6db2f44ff8cb4050c9652ec86eb5128aa313003
        - registry.redhat.io/mtr/mtr-web-executor-container-rhel8@sha256:2f0622774367f33efc321926196681322ab1535144e6a0785608c8050fc59f2e
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        version: 1.2.7
      entries:
      - name: mtr-operator.1.2.7
        version: 1.2.7
      - name: mtr-operator.1.2.6
        version: 1.2.6
      - name: mtr-operator.1.2.5
        version: 1.2.5
      - name: mtr-operator.1.2.4
        version: 1.2.4
      - name: mtr-operator.1.2.3
        version: 1.2.3
      - name: mtr-operator.1.2.2
        version: 1.2.2
      - name: mtr-operator.1.2.1
        version: 1.2.1
      - name: mtr-operator.1.2.0
        version: 1.2.0
      - name: mtr-operator.1.1.1
        version: 1.1.1
      - name: mtr-operator.1.1.0
        version: 1.1.0
      - name: mtr-operator.1.0.3
        version: 1.0.3
      - name: mtr-operator.1.0.2
        version: 1.0.2
      - name: mtr-operator.1.0.1
        version: 1.0.1
      - name: mtr-operator.1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: mtr-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://github.com/che-incubator/kubernetes-image-puller-operator
    name: kubernetes-imagepuller-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubernetes-imagepuller-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "che.eclipse.org/v1alpha1",
                "kind": "KubernetesImagePuller",
                "metadata": {
                  "name": "image-puller"
                },
                "spec": {
                  "daemonsetName": "k8s-image-puller",
                  "images": "java11-maven=quay.io/eclipse/che-java11-maven:next;che-theia=quay.io/eclipse/che-theia:next"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/eclipse/kubernetes-image-puller-operator:1.1.0
          createdAt: "2024-01-15T16:23:04Z"
          description: Create and manage kubernetes-image-puller instances.
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/che-incubator/kubernetes-image-puller-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubernetesImagePuller is the Schema for the kubernetesimagepullers
              API
            displayName: Kubernetes Image Puller
            kind: KubernetesImagePuller
            name: kubernetesimagepullers.che.eclipse.org
            version: v1alpha1
        description: "## About the managed application\n\nCreate a `KubernetesImagePuller`
          custom resource to automatically configure and run an instance of the [kubernetes-image-puller.](https://github.com/che-incubator/kubernetes-image-puller)\n##
          About this Operator\n\nThe `KubernetesImagePuller` custom resource understands
          the following fields in the `spec`:\n\n1. `configMapName` - The name of
          the `ConfigMap` to create\t\n2. `daemonsetName` - The name of the `DaemonSet`
          to be create.\n3. `deploymentName` - The name of the `kubernetes-image-puller`
          `Deployment` to create.\n4. `images` - A list of key-value pairs separated
          by semicolons of images to pull.  For example: `java=quay.io/eclipse/che-java8-maven:latest;theia=quay.io/eclipse/che-theia:next`\n5.
          `cachingIntervalHours` - The amount of time, in hours, between `DaemonSet`
          health checks.\n6. `cachingMemoryRequest` - The memory request for each
          cached image when the puller is running.\n7. `cachingMemoryLimit` - The
          memory limit for each cached image when the puller is running.\n8. `nodeSelector`
          - Node selector applied to pods created by the `DaemonSet`.\n## Prerequisites
          for enabling this Operator\n\nThe operator requires an existing namespace
          to be installed in.\nCheck the [README](https://github.com/che-incubator/kubernetes-image-puller-operator#installing-the-operator)
          for more details.\n"
        displayName: Kubernetes Image Puller Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - che
        - kubernetes-image-puller
        links:
        - name: Repository
          url: https://github.com/che-incubator/kubernetes-image-puller-operator
        maintainers:
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        provider:
          name: Red Hat, Inc.
          url: https://github.com/che-incubator/kubernetes-image-puller-operator
        relatedImages:
        - quay.io/eclipse/kubernetes-image-puller-operator:1.1.0
        - quay.io/brancz/kube-rbac-proxy:v0.11.0
        version: 1.1.0
      entries:
      - name: kubernetes-imagepuller-operator.v1.1.0
        version: 1.1.0
      - name: kubernetes-imagepuller-operator.v1.0.6
        version: 1.0.6
      - name: kubernetes-imagepuller-operator.v1.0.5
        version: 1.0.5
      - name: kubernetes-imagepuller-operator.v1.0.4
        version: 1.0.4
      - name: kubernetes-imagepuller-operator.v1.0.3
        version: 1.0.3
      - name: kubernetes-imagepuller-operator.v1.0.2
        version: 1.0.2
      - name: kubernetes-imagepuller-operator.v1.0.1
        version: 1.0.1
      - name: kubernetes-imagepuller-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: kubernetes-imagepuller-operator
    provider:
      name: Red Hat, Inc.
      url: https://github.com/che-incubator/kubernetes-image-puller-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Silicom Ltd.
      provider-url: http://www.silicom-usa.com
    name: silicom-sts-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: silicom-sts-operator.v0.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "namespace": "openshift-operators",
                  "nodeSelector": {
                    "sts.silicom.com/config": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1"
                },
                "spec": {
                  "aprLevel": 0,
                  "domainNumber": 24,
                  "esmcMode": 1,
                  "forwardable": 0,
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-TBC.8275.1",
                  "mode10MHz": 2,
                  "modePPS": 2,
                  "namespace": "openshift-operators",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  },
                  "phyLedsCtl": 0,
                  "priority2": 128,
                  "src10MHz": 1,
                  "srcPPS": 1,
                  "syncOption": 1,
                  "synceCpu": 10,
                  "synceRecClkPort": 0,
                  "tracePtpMsg": -1,
                  "twoStep": 0
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "worker2"
                },
                "spec": null,
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "images": null,
                  "sro": null
                }
              }
            ]
          capabilities: Basic Install
          categories: Drivers and plugins
          certified: "false"
          containerImage: quay.io/silicom/sts-operator:0.0.6
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicomdk/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicomDK/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/silicomDK/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - quay.io/silicom/sts-plugin@sha256:44339fd400d46d9dddb02c2da47c6f9f96ad2605a5c6de8a7dce2f8c3da3bb57
        - quay.io/silicom/phc2sys@sha256:0f3b63a3059b7d6158a916814281097f0d418b99a5fac9e8e5fd9bae58fb0b74
        - quay.io/silicom/tsyncd@sha256:78be78ec8939f025f9a8b899693ab77b3990116d1c7ffb2115c620b7c3382d9d
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/silicom/sts-operator:0.0.6
        - quay.io/silicom/grpc-tsyncd@sha256:aac7233423439e352d8eb39cef63e479c30c2f80b063c84ffc8797cc5a045117
        - quay.io/silicom/tsync_extts@sha256:be6a5950a5fdc6fdf8015767901856bf173d01bdc803be7141e4f9795ab1478a
        - quay.io/silicom/ice-driver-src@sha256:ca728f648769d31501cdedda92554bcfa890754405b0b06d41629cea47707da8
        - quay.io/silicom/gpsd@sha256:649e8a6c524920e1c564d8f797b8d64235b2527aa923f94bc9c7a0b7aba8ae7f
        version: 0.0.6
      entries:
      - name: silicom-sts-operator.v0.0.6
        version: 0.0.6
      - name: silicom-sts-operator.v0.0.5
        version: 0.0.5
      - name: silicom-sts-operator.v0.0.4
        version: 0.0.4
      - name: silicom-sts-operator.v0.0.3
        version: 0.0.3
      - name: silicom-sts-operator.v0.0.2
        version: 0.0.2
      - name: silicom-sts-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: silicom-sts-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "sts-silicom"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "namespace": "sts-silicom",
                  "nodeSelector": {
                    "sts.silicom.com/config": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1"
                },
                "spec": {
                  "aprLevel": 0,
                  "domainNumber": 24,
                  "esmcMode": 1,
                  "forwardable": 0,
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-TBC.8275.1",
                  "mode10MHz": 2,
                  "modePPS": 2,
                  "namespace": "sts-silicom",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  },
                  "phyLedsCtl": 0,
                  "priority2": 128,
                  "src10MHz": 1,
                  "srcPPS": 1,
                  "syncOption": 1,
                  "synceCpu": 10,
                  "synceRecClkPort": 0,
                  "tracePtpMsg": -1,
                  "twoStep": 0
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "worker2"
                },
                "spec": null,
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config"
                },
                "spec": {
                  "gpsSvcPort": 2947,
                  "grpcSvcPort": 50051,
                  "images": {
                    "gpsd": "quay.io/silicom/gpsd:3.23.1",
                    "grpcTsyncd": "quay.io/silicom/grpc-tsyncd:2.1.0.0",
                    "phcs2Sys": "quay.io/silicom/phcs2sys:3.1.1",
                    "stsPlugin": "quay.io/silicom/sts-plugin:0.0.2",
                    "ts2Phcs": "quay.io/silicom/ts2phcs:1.0.0",
                    "tsyncd": "quay.io/silicom/tsyncd:2.1.0.0"
                  },
                  "namespace": "sts-silicom",
                  "sro": {
                    "chart": {
                      "name": "ice-special-resource",
                      "repository": {
                        "name": "ice-special-resource",
                        "url": "http://ice-driver-src"
                      },
                      "version": "0.0.1"
                    },
                    "iceVersion": "1.7.16",
                    "namespace": "openshift-operators",
                    "srcImage": "quay.io/silicom/ice-driver-src:1.7.16",
                    "srcSvcPort": 3000
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Drivers and plugins
          certified: "false"
          containerImage: quay.io/silicom/sts-operator:0.0.2
          createdAt: 2021-11-17T10:21:24 CET
          description: This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of Silicom STS device
            plugins for Kubernetes.
          operators.operatorframework.io.bundle.channels.v1: stable,development
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicomdk/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicomDK/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/silicomDK/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/silicom/sts-operator:0.0.2
        version: 0.0.2
      entries:
      - name: silicom-sts-operator.v0.0.2
        version: 0.0.2
      - name: silicom-sts-operator.v0.0.1
        version: 0.0.1
      name: beta
    defaultChannel: alpha
    packageName: silicom-sts-operator
    provider:
      name: Silicom Ltd.
      url: http://www.silicom-usa.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-logging
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-logging.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Logging",
                "metadata": {
                  "name": "logging-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "elasticsearch": {
                    "storage": "16Gi",
                    "tag": "7.17.6-3747-35"
                  },
                  "environment": "openshift",
                  "fluentbit": {
                    "runOnMaster": false,
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "grafana": {
                    "dashboards": [
                      "wap",
                      "load-core"
                    ],
                    "dataSources": [
                      "wap",
                      "load-core"
                    ]
                  },
                  "image": {
                    "elasticInitImage": "rhel8/toolbox@sha256:a753c24cb8b03066d494f010ad654d23fb0559f14d7121542298da3e0d48340d",
                    "elasticInitRepository": "registry.access.redhat.com",
                    "fbImage": "fluent-bit:1.8.15-3747-20",
                    "fbRepository": "prod.harbor.keysight.digital/loadcore",
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "createClusterRoles": true,
                    "serviceAccountName": "load-core-service-account"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:33:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Logging
            name: loggings.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Logging Operator
        displayName: Keysight WAP Logging Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Logging
          url: https://keysight-wap-logging.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-logging@sha256:f1e3b3ad84cd0f96c92cdaf4609e29257035cb982a761e832dae2095e636c721
        version: 0.1.21
      entries:
      - name: keysight-wap-logging.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-logging
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hazelcast, Inc
      provider-url: ""
    name: hazelcast-platform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hazelcast-platform-operator.v5.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "cache-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "CronHotBackup",
                "metadata": {
                  "name": "cronhotbackup-sample"
                },
                "spec": {
                  "hotBackupTemplate": {
                    "metadata": {
                      "labels": {
                        "cron-hotbackup": "true"
                      }
                    },
                    "spec": {
                      "hazelcastResourceName": "hazelcast"
                    }
                  },
                  "schedule": "*/30 * * * *"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Flow",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hazelcast-platform-operator"
                  },
                  "name": "flow"
                },
                "spec": {
                  "database": {
                    "host": "flow-postgresql",
                    "secretName": "flow-db-secret"
                  },
                  "env": [
                    {
                      "name": "OPTIONS",
                      "value": "--flow.analytics.persistRemoteCallResponses=true --flow.analytics.persistResults=false --flow.config.custom.managementCenterUrl=http://34.140.160.158/mc\n"
                    }
                  ],
                  "externalConnectivity": {
                    "ingress": {
                      "annotations": {
                        "kubernetes.io/ingress.class": "nginx",
                        "nginx.ingress.kubernetes.io/enable-opentelemetry": "true",
                        "nginx.ingress.kubernetes.io/ssl-redirect": "false"
                      },
                      "hostname": "",
                      "ingressClassName": "nginx",
                      "path": "/"
                    }
                  },
                  "imagePullSecrets": [
                    {
                      "name": "hazelcast-cloud-hazelcast-connect-demo-pull-secret"
                    }
                  ],
                  "licenseKeySecretName": "flow-mc-license-key",
                  "repository": "quay.io/hazelcast_cloud/hazelcast-flow-internal",
                  "size": 3,
                  "version": "next"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast"
                },
                "spec": {
                  "clusterSize": 3,
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast-sample"
                },
                "spec": {
                  "clusterSize": 3,
                  "jet": {
                    "bucketConfig": {
                      "bucketURI": "gs://operator-user-code/jetJobs",
                      "secretName": "br-secret-gcp"
                    },
                    "enabled": true,
                    "resourceUploadEnabled": true
                  },
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HazelcastEndpoint",
                "metadata": {
                  "name": "hazelcastendpoint-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "port": 5701,
                  "type": "Discovery"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HotBackup",
                "metadata": {
                  "name": "hot-backup"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-longrun"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-longrun-2.0.0.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-1.0.2.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJobSnapshot",
                "metadata": {
                  "name": "jetjobsnapshot-sample"
                },
                "spec": {
                  "cancelJob": false,
                  "jetJobResourceName": "jet-job"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ManagementCenter",
                "metadata": {
                  "name": "managementcenter"
                },
                "spec": {
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "hazelcastClusters": [
                    {
                      "address": "hazelcast",
                      "name": "dev"
                    }
                  ],
                  "licenseKeySecretName": "hazelcast-license-key",
                  "persistence": {
                    "enabled": true,
                    "size": "10Gi"
                  },
                  "repository": "hazelcast/management-center"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Map",
                "metadata": {
                  "name": "map"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "MultiMap",
                "metadata": {
                  "name": "multimap-sample"
                },
                "spec": {
                  "backupCount": 3,
                  "binary": true,
                  "collectionType": "LIST",
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "queue-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ReplicatedMap",
                "metadata": {
                  "name": "replicatedmap-sample"
                },
                "spec": {
                  "asyncFillup": true,
                  "hazelcastResourceName": "hazelcast",
                  "inMemoryFormat": "OBJECT"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "topic-sample"
                },
                "spec": {
                  "globalOrderingEnabled": true,
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "UserCodeNamespace",
                "metadata": {
                  "name": "ucn-sample"
                },
                "spec": {
                  "bucketConfig": {
                    "bucketURI": "gs://operator-user-code/userCodeNamespace",
                    "secretName": "br-secret-gcp"
                  },
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "VectorCollection",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hazelcast-platform-operator"
                  },
                  "name": "vectorcollection-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "indexes": [
                    {
                      "dimension": 10,
                      "efConstruction": 256,
                      "maxDegree": 32,
                      "metric": "Dot",
                      "useDeduplication": false
                    }
                  ]
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanReplication",
                "metadata": {
                  "name": "wanreplication-sample"
                },
                "spec": {
                  "endpoints": "35.192.33.252",
                  "resources": [
                    {
                      "kind": "Map",
                      "name": "map-sample"
                    },
                    {
                      "kind": "Hazelcast",
                      "name": "hz-sample"
                    }
                  ],
                  "targetClusterName": "dev"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanSync",
                "metadata": {
                  "name": "wansync-sample"
                },
                "spec": {
                  "wanReplicationResourceName": "wan-replication-sample"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: docker.io/hazelcast/hazelcast-platform-operator@sha256:59328a4ea7a4e2110e0750364db792cf266f1e7231e3bb321e3de09d535dab23
          createdAt: "2025-03-13T11:59:59Z"
          description: Install Hazelcast clusters in Kubernetes environments.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hazelcast/hazelcast-platform-operator/
          support: Hazelcast, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.hazelcast.com
            version: v1alpha1
          - description: CronHotBackup is the Schema for the cronhotbackups API
            displayName: Cron Hot Backup
            kind: CronHotBackup
            name: cronhotbackups.hazelcast.com
            version: v1alpha1
          - description: Flow is the Schema for the flows API
            displayName: Flow
            kind: Flow
            name: flows.hazelcast.com
            version: v1alpha1
          - description: HazelcastEndpoint is the Schema for the hazelcastendpoints
              API
            displayName: Hazelcast Endpoint
            kind: HazelcastEndpoint
            name: hazelcastendpoints.hazelcast.com
            version: v1alpha1
          - description: Hazelcast is the Schema for the hazelcasts API
            displayName: Hazelcast
            kind: Hazelcast
            name: hazelcasts.hazelcast.com
            version: v1alpha1
          - description: HotBackup is the Schema for the hot backup API
            displayName: Hot Backup
            kind: HotBackup
            name: hotbackups.hazelcast.com
            version: v1alpha1
          - description: JetJob is the Schema for the jetjobs API
            displayName: Jet Job
            kind: JetJob
            name: jetjobs.hazelcast.com
            version: v1alpha1
          - description: JetJobSnapshot is the Schema for the jetjobsnapshots API
            displayName: Jet Job Snapshot
            kind: JetJobSnapshot
            name: jetjobsnapshots.hazelcast.com
            version: v1alpha1
          - description: ManagementCenter is the Schema for the managementcenters
              API
            displayName: Management Center
            kind: ManagementCenter
            name: managementcenters.hazelcast.com
            version: v1alpha1
          - description: Map is the Schema for the maps API
            displayName: Map
            kind: Map
            name: maps.hazelcast.com
            version: v1alpha1
          - description: MultiMap is the Schema for the multimaps API
            displayName: Multi Map
            kind: MultiMap
            name: multimaps.hazelcast.com
            version: v1alpha1
          - description: Queue is the Schema for the queues API
            displayName: Queue
            kind: Queue
            name: queues.hazelcast.com
            version: v1alpha1
          - description: ReplicatedMap is the Schema for the replicatedmaps API
            displayName: Replicated Map
            kind: ReplicatedMap
            name: replicatedmaps.hazelcast.com
            version: v1alpha1
          - description: Topic is the Schema for the topics API
            displayName: Topic
            kind: Topic
            name: topics.hazelcast.com
            version: v1alpha1
          - description: UserCodeNamespace is the Schema for the usercodenamespaces
              API
            displayName: User Code Namespace
            kind: UserCodeNamespace
            name: usercodenamespaces.hazelcast.com
            version: v1alpha1
          - description: VectorCollection is the Schema for the vectorcollections
              API
            displayName: Vector Collection
            kind: VectorCollection
            name: vectorcollections.hazelcast.com
            version: v1alpha1
          - description: WanReplication is the Schema for the wanreplications API
            displayName: Wan Replication
            kind: WanReplication
            name: wanreplications.hazelcast.com
            version: v1alpha1
          - description: WanSync is the Schema for the wansyncs API
            displayName: Wan Sync
            kind: WanSync
            name: wansyncs.hazelcast.com
            version: v1alpha1
        description: |
          Easily deploy Hazelcast clusters and Management Center into Kubernetes environments and manage their lifecycles.

          ## Before You Start

          If you are planning to create Hazelcast Platform Enterprise clusters, you need to [create a secret](https://docs.hazelcast.com/operator/latest/get-started#step-2-start-the-hazelcast-cluster) for the license. You can request a trial license key from [here](https://trialrequest.hazelcast.com).

          For Hazelcast Platform clusters, you can simply continue.

          ## Documentation

          1. [Get started](https://docs.hazelcast.com/operator/latest/get-started) with the Operator
          2. [Connect to the cluster from outside Kubernetes](https://guides.hazelcast.org/hazelcast-platform-operator-expose-externally)
          3. [Restore a Cluster from Cloud Storage with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-external-backup-restore)
          4. [Replicate Data between Two Hazelcast Clusters with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-wan-replication)

          ## Features

          Hazelcast Platform Operator supports the features below:

          * Custom resource for Hazelcast Platform (Enterprise) and Management Center
          * Observe status of Hazelcast clusters and Management Center instances
          * High Availability Mode configuration to create clusters that are resilient to node and zone failures
          * Support for TLS connections between members using self-signed certificates
          * Scale up and down Hazelcast clusters
          * Expose Hazelcast cluster to external clients ([Smart & Unisocket](https://docs.hazelcast.com/hazelcast/latest/clients/java#java-client-operation-modes))
          * Backup Hazelcast persistence data to cloud storage with the possibility of scheduling it and restoring the data accordingly
          * WAN Replication feature when you need to synchronize multiple Hazelcast clusters, which are connected by WANs
          * Full and Delta WAN Sync support
          * Tiered Storage support for Hazelcast and Map CRs
          * CP Subsystem configuration support for Hazelcast CR
          * User Code Deployment feature, which allows you to deploy custom and domain classes from cloud storages and URLs to Hazelcast members
          * The User Code Namespaces feature allows you to deploy custom and domain classes at runtime without a members restart
          * Support the configuration of advanced networking options
          * Support Multi-namespace configuration
          * ExecutorService and EntryProcessor support
          * Support several data structures like Map, Topic, MultiMap, ReplicatedMap, Queue and Cache which can be created dynamically via specific Custom Resources
          * MapStore, Near Cache and off-heap memory (HD memory and native memory) support for the Map CR
          * Native Memory support for the Cache CR
          * Support Jet configuration and Jet Job submission using the JetJob CR
          * Support for exporting the snapshots of JetJob CRs using JetJobSnapshot CR
          * Support for custom configurations using ConfigMap
        displayName: Hazelcast Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hazelcast
        - keyvalue
        - in-memory
        - database
        - caching
        links:
        - name: Documentation
          url: https://docs.hazelcast.com/operator/latest/
        - name: Source Code
          url: https://github.com/hazelcast/hazelcast-platform-operator
        maintainers:
        - email: cloudnative@hazelcast.com
          name: Hazelcast Cloud Native Team
        maturity: alpha
        minKubeVersion: 1.22.7
        provider:
          name: Hazelcast, Inc
        relatedImages:
        - docker.io/hazelcast/hazelcast-platform-operator@sha256:59328a4ea7a4e2110e0750364db792cf266f1e7231e3bb321e3de09d535dab23
        version: 5.15.0
      entries:
      - name: hazelcast-platform-operator.v5.15.0
        version: 5.15.0
      name: alpha
    defaultChannel: alpha
    packageName: hazelcast-platform-operator
    provider:
      name: Hazelcast, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: tang-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tang-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "daemons.redhat.com/v1alpha1",
                "kind": "TangServer",
                "metadata": {
                  "finalizers": [
                    "finalizer.daemons.tangserver.redhat.com"
                  ],
                  "name": "tangserver",
                  "namespace": "nbde"
                },
                "spec": {
                  "healthScript": "/usr/bin/tangd-health-check",
                  "image": "registry.redhat.io/rhel9/tang",
                  "keyPath": "/var/db/tang",
                  "persistentVolumeClaim": "tangserver-pvc",
                  "replicas": 1,
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:f43c96035f43040b3644e37119a3b35fc433f366d4b636459d1b9b02a05b5d3d
          createdAt: 2023-11-23T12:49:40
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TangServer is the Schema for the tangservers API
            displayName: Tang Server
            kind: TangServer
            name: tangservers.daemons.redhat.com
            version: v1alpha1
        description: NBDE Tang Server operator allows Tang Server deployment on OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nbde
        - tang
        links:
        - name: Tang Operator
          url: https://github.com/latchset/tang-operator
        maintainers:
        - email: sarroutb@redhat.com
          name: Sergio Arroutbi Braojos
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:97cade2c1ee468261aec5400728c8d44de387b459134aec7a4c3b5ec5a335d2c
        - registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:f43c96035f43040b3644e37119a3b35fc433f366d4b636459d1b9b02a05b5d3d
        version: 1.0.0
      entries:
      - name: tang-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: tang-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "daemons.redhat.com/v1alpha1",
                "kind": "TangServer",
                "metadata": {
                  "finalizers": [
                    "finalizer.daemons.tangserver.redhat.com"
                  ],
                  "name": "tangserver",
                  "namespace": "nbde"
                },
                "spec": {
                  "healthScript": "/usr/bin/tangd-health-check",
                  "image": "registry.redhat.io/rhel9/tang",
                  "keyPath": "/var/db/tang",
                  "persistentVolumeClaim": "tangserver-pvc",
                  "replicas": 1,
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:1cb2f794689817282ac570cacb106d75a2751a1926d89d3b83b619e02d869e5e
          createdAt: 2024-12-04T19:01:27
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TangServer is the Schema for the tangservers API
            displayName: Tang Server
            kind: TangServer
            name: tangservers.daemons.redhat.com
            version: v1alpha1
        description: NBDE Tang Server operator allows Tang Server deployment on OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nbde
        - tang
        links:
        - name: Tang Operator
          url: https://github.com/latchset/tang-operator
        maintainers:
        - email: sarroutb@redhat.com
          name: Sergio Arroutbi Braojos
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1172e150fff22c5eeab572f26961f3f53fbf896ee76d08c7503cfe2777c55458
        - registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:1cb2f794689817282ac570cacb106d75a2751a1926d89d3b83b619e02d869e5e
        version: 1.0.3
      entries:
      - name: tang-operator.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: alpha
    packageName: tang-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: www.redhat.com
    name: compliance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: compliance-operator.v0.1.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.17 <0.1.32'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          repository: https://github.com/openshift/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repo
          url: https://github.com/openshift/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/4.7/security/compliance_operator/compliance-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/compliance-rhel8-operator@sha256:ed68a0574ddeaf7d676403cd843f28cb586894f35dbb3ca5e06a5264bd42f030
        - registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c
        - registry.redhat.io/openshift4/compliance-openscap-rhel8@sha256:867c6d4205653d8a954d234cbaf470313009eb983d4f455fb45f49c4a7764422
        version: 0.1.32
      entries:
      - name: compliance-operator.v0.1.32
        version: 0.1.32
      name: "4.7"
    - currentCSV: compliance-operator.v0.1.61
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-node"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-platform"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "description": "Example of a tailoredProfile that extends OCP4 FedRAMP Moderate",
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.17 <0.1.61'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            displayName: Compliance Remediation
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            displayName: Compliance Scan
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            displayName: Compliance Suite
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            displayName: Profile Bundle
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            displayName: Scan Setting Binding
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            displayName: Tailored Profile
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repo
          url: https://github.com/openshift/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/compliance_operator/compliance-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b
        - registry.redhat.io/compliance/openshift-compliance-openscap-rhel8@sha256:83ad3113964ec48c5cf28315d923adb6edb95e45960e2756665d617dcd9b2e47
        - registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:5578e2225365ac1c15e4c189c5c7a8d14fd3fd26189414b36028f7f0879bb1dc
        version: 0.1.61
      entries:
      - name: compliance-operator.v0.1.61
        version: 0.1.61
      name: release-0.1
    - currentCSV: compliance-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-node"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-platform"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "description": "Example of a tailoredProfile that extends OCP4 FedRAMP Moderate",
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          must-gather-image: registry.redhat.io/compliance/openshift-compliance-must-gather-rhel8@sha256:e6468ccc55d72539b7fb980d0bce5c072e85879d4e9db39974d53eaab4d32b25
          olm.skipRange: '>=0.1.17 <1.7.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ComplianceAsCode/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            displayName: Compliance Remediation
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            displayName: Compliance Scan
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            displayName: Compliance Suite
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            displayName: Profile Bundle
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            displayName: Scan Setting Binding
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            displayName: Tailored Profile
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repository
          url: https://github.com/ComplianceAsCode/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/compliance_operator/co-overview.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4
        - registry.redhat.io/compliance/openshift-compliance-openscap-rhel8@sha256:d0f89d82845c6245a18e7db345ac3be92b2863d57f55f63e21652086128920c9
        - registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:8d3257bd6a020df53f8d7a400ae877ae029c78ffcd8ec9e35270ee3ff5c8e509
        - registry.redhat.io/compliance/openshift-compliance-must-gather-rhel8@sha256:e6468ccc55d72539b7fb980d0bce5c072e85879d4e9db39974d53eaab4d32b25
        version: 1.7.1
      entries:
      - name: compliance-operator.v1.7.1
        version: 1.7.1
      - name: compliance-operator.v1.7.0
        version: 1.7.0
      - name: compliance-operator.v1.6.2
        version: 1.6.2
      - name: compliance-operator.v1.6.1
        version: 1.6.1
      - name: compliance-operator.v1.6.0
        version: 1.6.0
      - name: compliance-operator.v1.5.1
        version: 1.5.1
      - name: compliance-operator.v1.5.0
        version: 1.5.0
      - name: compliance-operator.v1.4.1
        version: 1.4.1
      - name: compliance-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: compliance-operator
    provider:
      name: Red Hat Inc.
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: GitLab Inc
      provider-url: https://about.gitlab.com/
    name: gitlab-operator-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "certmanager": {
                        "install": false
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-26T07:16:35Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3281fe12aaae428fd7102332e7c0d7033515042d7830eba3953891ab0c3e3a29
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      - name: gitlab-operator-kubernetes.v2.1.1
        version: 2.1.1
      - name: gitlab-operator-kubernetes.v2.1.0
        version: 2.1.0
      - name: gitlab-operator-kubernetes.v2.0.2
        version: 2.0.2
      - name: gitlab-operator-kubernetes.v2.0.1
        version: 2.0.1
      - name: gitlab-operator-kubernetes.v2.0.0
        version: 2.0.0
      - name: gitlab-operator-kubernetes.v1.12.2
        version: 1.12.2
      - name: gitlab-operator-kubernetes.v1.12.1
        version: 1.12.1
      - name: gitlab-operator-kubernetes.v1.12.0
        version: 1.12.0
      - name: gitlab-operator-kubernetes.v1.11.3
        version: 1.11.3
      - name: gitlab-operator-kubernetes.v1.11.2
        version: 1.11.2
      - name: gitlab-operator-kubernetes.v1.11.1
        version: 1.11.1
      - name: gitlab-operator-kubernetes.v1.11.0
        version: 1.11.0
      - name: gitlab-operator-kubernetes.v1.10.2
        version: 1.10.2
      - name: gitlab-operator-kubernetes.v1.10.1
        version: 1.10.1
      - name: gitlab-operator-kubernetes.v1.10.0
        version: 1.10.0
      - name: gitlab-operator-kubernetes.v1.9.2
        version: 1.9.2
      - name: gitlab-operator-kubernetes.v1.9.1
        version: 1.9.1
      - name: gitlab-operator-kubernetes.v1.9.0
        version: 1.9.0
      - name: gitlab-operator-kubernetes.v1.8.3
        version: 1.8.3
      - name: gitlab-operator-kubernetes.v1.8.2
        version: 1.8.2
      - name: gitlab-operator-kubernetes.v1.8.1
        version: 1.8.1
      - name: gitlab-operator-kubernetes.v1.8.0
        version: 1.8.0
      - name: gitlab-operator-kubernetes.v1.7.2
        version: 1.7.2
      - name: gitlab-operator-kubernetes.v1.7.1
        version: 1.7.1
      - name: gitlab-operator-kubernetes.v1.7.0
        version: 1.7.0
      - name: gitlab-operator-kubernetes.v1.6.1
        version: 1.6.1
      - name: gitlab-operator-kubernetes.v1.6.0
        version: 1.6.0
      - name: gitlab-operator-kubernetes.v1.5.1
        version: 1.5.1
      - name: gitlab-operator-kubernetes.v1.4.2
        version: 1.4.2
      - name: gitlab-operator-kubernetes.v1.4.1
        version: 1.4.1
      - name: gitlab-operator-kubernetes.v1.4.0
        version: 1.4.0
      - name: gitlab-operator-kubernetes.v1.3.3
        version: 1.3.3
      - name: gitlab-operator-kubernetes.v1.3.2
        version: 1.3.2
      - name: gitlab-operator-kubernetes.v1.3.1
        version: 1.3.1
      - name: gitlab-operator-kubernetes.v1.3.0
        version: 1.3.0
      name: stable
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "certmanager": {
                        "install": false
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-26T07:16:35Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3281fe12aaae428fd7102332e7c0d7033515042d7830eba3953891ab0c3e3a29
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      - name: gitlab-operator-kubernetes.v2.1.1
        version: 2.1.1
      - name: gitlab-operator-kubernetes.v2.1.0
        version: 2.1.0
      - name: gitlab-operator-kubernetes.v2.0.2
        version: 2.0.2
      - name: gitlab-operator-kubernetes.v2.0.1
        version: 2.0.1
      - name: gitlab-operator-kubernetes.v2.0.0
        version: 2.0.0
      - name: gitlab-operator-kubernetes.v1.12.2
        version: 1.12.2
      - name: gitlab-operator-kubernetes.v1.12.1
        version: 1.12.1
      - name: gitlab-operator-kubernetes.v1.12.0
        version: 1.12.0
      - name: gitlab-operator-kubernetes.v1.11.3
        version: 1.11.3
      - name: gitlab-operator-kubernetes.v1.11.2
        version: 1.11.2
      - name: gitlab-operator-kubernetes.v1.11.1
        version: 1.11.1
      - name: gitlab-operator-kubernetes.v1.11.0
        version: 1.11.0
      - name: gitlab-operator-kubernetes.v1.10.2
        version: 1.10.2
      - name: gitlab-operator-kubernetes.v1.10.1
        version: 1.10.1
      - name: gitlab-operator-kubernetes.v1.10.0
        version: 1.10.0
      - name: gitlab-operator-kubernetes.v1.9.2
        version: 1.9.2
      - name: gitlab-operator-kubernetes.v1.9.1
        version: 1.9.1
      - name: gitlab-operator-kubernetes.v1.9.0
        version: 1.9.0
      - name: gitlab-operator-kubernetes.v1.8.3
        version: 1.8.3
      - name: gitlab-operator-kubernetes.v1.8.2
        version: 1.8.2
      - name: gitlab-operator-kubernetes.v1.8.1
        version: 1.8.1
      - name: gitlab-operator-kubernetes.v1.8.0
        version: 1.8.0
      - name: gitlab-operator-kubernetes.v1.7.2
        version: 1.7.2
      - name: gitlab-operator-kubernetes.v1.7.1
        version: 1.7.1
      - name: gitlab-operator-kubernetes.v1.7.0
        version: 1.7.0
      - name: gitlab-operator-kubernetes.v1.6.1
        version: 1.6.1
      - name: gitlab-operator-kubernetes.v1.6.0
        version: 1.6.0
      - name: gitlab-operator-kubernetes.v1.5.1
        version: 1.5.1
      - name: gitlab-operator-kubernetes.v1.4.2
        version: 1.4.2
      - name: gitlab-operator-kubernetes.v1.4.1
        version: 1.4.1
      - name: gitlab-operator-kubernetes.v1.4.0
        version: 1.4.0
      - name: gitlab-operator-kubernetes.v1.3.3
        version: 1.3.3
      - name: gitlab-operator-kubernetes.v1.3.2
        version: 1.3.2
      - name: gitlab-operator-kubernetes.v1.3.1
        version: 1.3.1
      - name: gitlab-operator-kubernetes.v1.3.0
        version: 1.3.0
      name: unstable
    defaultChannel: stable
    packageName: gitlab-operator-kubernetes
    provider:
      name: GitLab Inc
      url: https://about.gitlab.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Antline-Soft
      provider-url: ""
    name: cxl-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cxl-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-05-08T08:36:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
        description: memory extend operator
        displayName: extend-community-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - memory
        - extend
        - community
        - operator
        links:
        - name: Cxl Operator
          url: https://cxl-operator.domain
        maintainers:
        - email: tomboy@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        provider:
          name: Antline-Soft
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - rasen708/cxl-operator:latest
        version: 0.0.1
      entries:
      - name: cxl-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cxl-operator
    provider:
      name: Antline-Soft
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: hyperfoil-bundle
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hyperfoil-operator.v0.26.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hyperfoil.io/v1alpha2",
                "kind": "Hyperfoil",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "hyperfoil-operator",
                    "app.kubernetes.io/instance": "hyperfoil-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hyperfoil",
                    "app.kubernetes.io/part-of": "hyperfoil-operator"
                  },
                  "name": "hyperfoil-sample"
                },
                "spec": {
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/hyperfoil/hyperfoil-operator:0.26.0
          createdAt: "2024-06-11T17:39:20Z"
          description: Microservice-oriented distributed benchmark framework.
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Hyperfoil/hyperfoil-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hyperfoil is the Schema for the hyperfoils API
            displayName: Hyperfoil
            kind: Hyperfoil
            name: hyperfoils.hyperfoil.io
            version: v1alpha2
        description: |
          Hyperfoil is a modern benchmark framework oriented promising
          accurate results with more flexibility and distributed architecture.

          See more on [hyperfoil.io](http://hyperfoil.io).

          Hyperfoil is licensed under Apache License 2.0
        displayName: Hyperfoil
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hyperfoil
        - benchmark
        - test
        - load
        - driver
        links:
        - name: Website
          url: http://hyperfoil.io
        - name: GitHub
          url: http://github.com/Hyperfoil
        maintainers:
        - email: rvansa@redhat.com
          name: Radim Vansa
        - email: spederse@redhat.com
          name: Ståle W. Pedersen
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/hyperfoil/hyperfoil-operator:0.26.0
        version: 0.26.0
      entries:
      - name: hyperfoil-operator.v0.26.0
        version: 0.26.0
      - name: hyperfoil-operator.v0.24.2
        version: 0.24.2
      - name: hyperfoil-operator.v0.21.0
        version: 0.21.0
      - name: hyperfoil-operator.v0.15.1
        version: 0.15.1
      - name: hyperfoil-operator.v0.15.0
        version: 0.15.0
      name: alpha
    defaultChannel: alpha
    packageName: hyperfoil-bundle
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC
      provider-url: https://www.nec.com
    name: sanstoragecsi-operator-bundle
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sanstoragecsi-operator-bundle.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.nec.com/v1",
                "kind": "SANStorageCSI",
                "metadata": {
                  "name": "sanstoragecsi"
                },
                "spec": {
                  "driverName": "san.csi.nec.com",
                  "images": {
                    "csiDriver": "",
                    "externalAttacher": "registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:bd8262cc4b6c405e175c39623323dd98a74bbf70e6aa90851a72612056c184ff",
                    "externalProvisioner": "registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:85f1cc8f21568e7885e492b29660f29e09796fb23ebdf5fd526aa2e99d6b2fa8",
                    "livenessProbe": "registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:1a0d15012ba70955af754b03a7488b7c6ace813646ec7c3ce845f80597050c3c",
                    "nodeRegistrar": "registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:dc7f0992be6c2853dc0e798afda368e36005a94ad76b21fa2f09790621318864"
                  },
                  "parameters": {
                    "controller": {
                      "logLevel": 5
                    },
                    "node": {
                      "livenessProbePort": 9808,
                      "logLevel": 5,
                      "maxVolumesPerNode": 1024
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          createdAt: "2022-06-15T00:00:00Z"
          description: An operator for managing the NEC Storage M Series CSI Driver
          operatorframework.io/suggested-namespace: nec-san-csi-operator-system
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: NEC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SANStorageCSI is the Schema for the sanstoragecsis API
            displayName: NEC Storage M Series CSI Driver
            kind: SANStorageCSI
            name: sanstoragecsis.csi.nec.com
            version: v1
        description: |
          This operator deploys the NEC Storage M Series CSI Driver on OpenShift Container Platform.

          NEC Storage M Series CSI Driver provides persistent storage for stateful applications using NEC Storage M Series.
          Please refer to the documentation of the NEC Storage M Series CSI Driver for detail information about supported NEC Storage Platforms and supported CSI features.
        displayName: NEC Storage M Series CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - NEC
        - M Series
        - SAN Storage
        - CSI Driver
        links:
        - name: NEC Storage M Series
          url: https://www.nec.com/en/global/prod/storage/product/san/index.html
        maturity: stable
        provider:
          name: NEC
          url: https://www.nec.com
        relatedImages:
        - registry.connect.redhat.com/nec/sanstoragecsi-operator@sha256:6656c2b5036bdc013712ecb640bbbebe0e3e90cf625c73697dbb49e3e201e2ac
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:dc7f0992be6c2853dc0e798afda368e36005a94ad76b21fa2f09790621318864
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:bd8262cc4b6c405e175c39623323dd98a74bbf70e6aa90851a72612056c184ff
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:85f1cc8f21568e7885e492b29660f29e09796fb23ebdf5fd526aa2e99d6b2fa8
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:1a0d15012ba70955af754b03a7488b7c6ace813646ec7c3ce845f80597050c3c
        version: 1.1.0
      entries:
      - name: sanstoragecsi-operator-bundle.v1.1.0
        version: 1.1.0
      - name: sanstoragecsi-operator-bundle.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: sanstoragecsi-operator-bundle
    provider:
      name: NEC
      url: https://www.nec.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nxiq-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nxiq-operator-certified.v1.193.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusIQ",
                "metadata": {
                  "name": "example-nexusiq"
                },
                "spec": {
                  "deployment": {
                    "postStart": {
                      "command": null
                    }
                  },
                  "deploymentStrategy": {
                    "type": "Recreate"
                  },
                  "fullnameOverride": "",
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [],
                    "path": "/",
                    "tls": []
                  },
                  "iq": {
                    "adminPort": 8071,
                    "applicationPort": 8070,
                    "env": [
                      {
                        "name": "JAVA_OPTS",
                        "value": "-Djava.util.prefs.userRoot=$(SONATYPE_WORK)/javaprefs -Djavax.net.ssl.trustStore=/etc/nexus-iq-server/jks/keystore.jks"
                      }
                    ],
                    "configYaml": {
                      "sonatypeWork": "/sonatype-work",
                      "createSampleData": true,
                      "server": {
                        "adminConnectors": [
                          {
                            "port": 8071,
                            "type": "http"
                          }
                        ],
                        "applicationConnectors": [
                          {
                            "port": 8070,
                            "type": "http"
                          }
                        ],
                        "requestLog": {
                          "appenders": [
                            {
                              "type": "file",
                              "currentLogFilename": "/var/log/nexus-iq-server/request.log",
                              "logFormat": "%clientHost %l %user [%date] \"%requestURL\" %statusCode %bytesSent %elapsedTime \"%header{User-Agent}\"",
                              "archivedLogFilenamePattern": "/var/log/nexus-iq-server/request-%d.log.gz",
                              "archivedFileCount": 50
                            }
                          ]
                        }
                      },
                      "logging": {
                        "level": "DEBUG",
                        "loggers": {
                          "com.sonatype.insight.scan": "INFO",
                          "eu.medsea.mimeutil.MimeUtil2": "INFO",
                          "org.apache.http": "INFO",
                          "org.apache.http.wire": "ERROR",
                          "org.eclipse.birt.report.engine.layout.pdf.font.FontConfigReader": "WARN",
                          "org.eclipse.jetty": "INFO",
                          "org.apache.shiro.web.filter.authc.BasicHttpAuthenticationFilter": "INFO",
                          "com.sonatype.insight.audit": {
                            "appenders": [
                              {
                                "type": "file",
                                "currentLogFilename": "/var/log/nexus-iq-server/audit.log",
                                "archivedLogFilenamePattern": "/var/log/nexus-iq-server/audit-%d.log.gz",
                                "archivedFileCount": 50
                              }
                            ]
                          },
                          "com.sonatype.insight.policy.violation": {
                            "appenders": [
                              {
                                "type": "console",
                                "threshold": "INFO"
                              }
                            ]
                          }
                        },
                        "appenders": [
                          {
                            "type": "console",
                            "threshold": "INFO",
                            "logFormat": "%d{'yyyy-MM-dd HH:mm:ss,SSSZ'} %level [%thread] %X{username} %logger - %msg%n"
                          },
                          {
                            "type": "file",
                            "threshold": "ALL",
                            "logFormat": "%d{'yyyy-MM-dd HH:mm:ss,SSSZ'} %level [%thread] %X{username} %logger - %msg%n",
                            "currentLogFilename": "/var/log/nexus-iq-server/clm-server.log",
                            "archivedLogFilenamePattern": "/var/log/nexus-iq-server/clm-server-%d.log.gz",
                            "archivedFileCount": 50
                          }
                        ]
                      }
                    },
                    "imageName": "registry.connect.redhat.com/sonatype/nexus-iq-server@sha256:d65df6bdff1f32bf32a87e5cb38f5e9f98b75b81e9acccbffbd967e1f71691f0",
                    "imagePullPolicy": "IfNotPresent",
                    "imagePullSecret": "",
                    "licenseSecret": "",
                    "memory": "1Gi",
                    "name": "nxiq",
                    "certificates": {}
                  },
                  "nameOverride": "",
                  "persistence": {
                    "accessMode": "ReadWriteOnce",
                    "storageSize": "1Gi"
                  },
                  "service": {
                    "annotations": {},
                    "enabled": false,
                    "labels": {},
                    "ports": [
                      {
                        "name": "nexus-service",
                        "port": 80,
                        "targetPort": 80
                      }
                    ]
                  },
                  "serviceAccount": {
                    "create": true,
                    "name": null
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nxiq-operator-certified@sha256:59c52729189da87d9b5c3de8ef85326662d94c99835ab99d6e8d7fefdcf53ef0
          createdAt: "2020-07-20"
          description: |-
            Nexus Lifecycle is an open source governance platform that continuously identifies risk,
            automatically enforces policy, and provides visibility throughout the entire SDLC.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.193.0-1
          repository: https://github.com/sonatype/operator-iq
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus IQ Server
            displayName: NexusIQ
            kind: NexusIQ
            name: nexusiqs.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Lifecycle is an open source governance platform that continuously identifies risk,
          automatically enforces policy, and provides visibility throughout the entire SDLC.
          Developers leveraging open source require a solution that helps them make safer choices
          while still delivering at DevOps speed.
          Nexus Lifecycle empowers developers and security teams to choose the best components
          and govern open source usage, ensuring organizations continue to innovate with less risk.

          ## Core Capabilities

          * **Dependency Management**:
            Generate a precise software bill of materials for all applications
            to identify open source components and gather contextual direct
            and transitive dependency information for faster action.
          * **One-Click Remediation**:
            Upgrade to the best components and versions based on real-time intelligence.
            Available in your preferred IDE or a GitHub pull request.
          * **Automated Open Source Policy Enforcement**:
            Create custom security, license, and architectural policies based
            on application type or organization and contextually enforce those policies
            across every stage of the SDLC.
          * **Continuous Monitoring**:
            Real-time notifications of newly discovered defects,
            as well as vulnerabilities based on component,
            risk level or application/container affected.

          ## Extras

          * IDE Integrations: Eclipse, IntelliJ, and Visual Studio
          * Source Control Integrations: GitHub, GitLab, and Bitbucket
          * Chrome Browser Extension
          * Red Hat Clair for Container Scanning

          ## Limitations

          This operator will be released on a quarterly basis.

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus IQ Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage

          Once the server instance is created by the operator and running,
          you'll want to expose the service as you see fit:
          1. Create a Route to the new service for iq.applicationPort (8070).
          2. Visit the URL provided by the Route, login, and set new credentials.
            The default credentials are `admin`/`admin123`.

          The Nexus IQ Server can be further configured via the NexusIQ custom resource definition:

          | Parameter            | Description                                                  | Default           |
          | -------------------- | ------------------------------------------------------------ | ----------------- |
          | `iq.applicationPort` | Port of the application connector. Must match the value in the `configYaml` property | `8070`            |
          | `iq.adminPort`       | Port of the application connector. Must match the value in the `configYaml` property | `8071`            |
          | `iq.memory`          | The amount of RAM to allocate                                | `1Gi`             |
          | `iq.env`             | Customize the server environment, including `JAVA_OPTS`      | See example YAML show when ceating a NexusIQ. |
          | `iq.licenseSecret`   | The base-64 encoded license file to be installed at startup  | `""`              |
          | `iq.configYaml`      | A YAML block which will be used as a configuration block for IQ Server. | See example YAML shown when creating a NexusIQ. |
          | `iq.certificates`   | Add your own certificates for your own trusted Certificate Authorities with names and plain armored text. | `{}` |
          | `ingress.enabled`                           | Create an ingress for Nexus         | `true`                                  |
          | `ingress.annotations`                       | Annotations to enhance ingress configuration  | `{}`                          |
          | `ingress.tls.enabled`                       | Enable TLS                          | `true`                                 |
          | `ingress.tls.secretName`                    | Name of the secret storing TLS cert, `false` to use the Ingress' default certificate | `nexus-tls`                             |
          | `ingress.path`                              | Path for ingress rules. GCP users should set to `/*` | `/`                    |
          | `deploymentStrategy.type`                   | Deployment Strategy                 |  `Recreate` |
          | `initContainers`                            | Containers to run before starting the server | `nil` |
          | `deployment.postStart.command`              | Command to run after starting the IQ Server container  | `nil`                    |
          | `deployment.terminationGracePeriodSeconds` | Time to allow for clean shutdown                        | 120                      |
          | `persistence.storageClass` | The provisioner class                        | `-` (disables dynamic provisioning)            |
          | `persistence.storageSize` | The amount of drive space to allocate                        | `1Gi`             |
          | `persistence.accessMode` | Default access mode                        | `ReadWriteOnce`             |
          | `persistence.volumeConfiguration` | A YAML block to configure the persistent volume type. Defaults to `hostPath` which should not be used in production | `hostPath`             |

          ## Configuring IQ Server

          You can define the `config.yml` for IQ Server in your CRD on startup.
          It is the `iq.configYaml` property. For more details, see the [Configuring IQ Server](https://help.sonatype.com/iqserver/configuring) help page.

          ## Installing the License

          The license file can be installed via the UI when IQ server is running, or it can be done as a part of the deploy.
          If you leave the `licenseFile` field empty/commented, IQ Server will start and prompt you to manually install the license
          when you first enter the GUI.

          ### Installing the License Automatically
          To do it automatically, first encode your `.lic` file in Base 64 with no line breaks, eg:

          ```bash
          base64 --wrap=0 mylicense.lic > lic.base64
          ```

          Then add this value to your CRD file as `iq.licenseSecret`, eg:

          ```yaml
          iq:
            licenseSecret: bXkgc2FtcGxlIGxpY2Vuc2U=
          ```

          Specify the `licenseFile` path in `iq.configYaml` as:

          ```yaml
          iq:
            configYaml:
              sonatypeWork: /sonatype-work
              createSampleData: true
              # add this line and the `licenseSecret` above to autoconfigure licensing
              licenseFile: /etc/nexus-iq-license/license_lic
              server:
                applicationConnectors:
                  - type: http
                    port: 8070
                adminConnectors:
                  - type: http
                    port: 8071
          ```

          ## Adding Certificates for Your Own Trusted Certificate Authorities

          Specify a set of certificates to be imported into the Java trust store
          of the running IQ Server. More than one certificate can be provided.

          ```yaml
          iq:
            certificates:
              my-company-cert: |
                -----BEGIN CERTIFICATE-----
                ...
                -----END CERTIFICATE-----
          ```

          ## Running Commands Before Initialization of the Server

          If you need custom initialization, you can use standard Kubernetes
          [`initContainers`](https://kubernetes.io/docs/concepts/workloads/pods/init-containers/)
          to run any series of containers and commands you need.

          ```yaml
          initContainers:
            - name: an-init-container
              image: busybox:1.28
              command: ['/bin/sh', '-c', 'echo "initializing"']
          ```
        displayName: Nexus IQ Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - sonatype
        - licensing
        - policy
        - components
        links:
        - name: Nexus Lifecycle
          url: https://www.sonatype.com/product-nexus-lifecycle
        maintainers:
        - email: support@sonaype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nxiq-operator-certified@sha256:59c52729189da87d9b5c3de8ef85326662d94c99835ab99d6e8d7fefdcf53ef0
        - registry.connect.redhat.com/sonatype/nexus-iq-server@sha256:d65df6bdff1f32bf32a87e5cb38f5e9f98b75b81e9acccbffbd967e1f71691f0
        version: 1.193.0-1
      entries:
      - name: nxiq-operator-certified.v1.193.0-1
        version: 1.193.0-1
      name: stable
    defaultChannel: stable
    packageName: nxiq-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com/
    name: openstack-baremetal-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openstack-baremetal-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "baremetal.openstack.org/v1beta1",
                "kind": "OpenStackBaremetalSet",
                "metadata": {
                  "name": "openstackbaremetalset"
                },
                "spec": {
                  "baremetalHosts": {
                    "compute-0": "172.22.0.100/24",
                    "compute-1": "172.22.0.101/24"
                  },
                  "ctlplaneGateway": "172.22.0.3",
                  "ctlplaneInterface": "enp1s0",
                  "deploymentSSHSecret": "baremetalset-deploy-secret",
                  "domainName": "osptest.openstack.org",
                  "osImage": "edpm-hardened-uefi.qcow2",
                  "passwordSecret": "baremetalset-password-secret"
                }
              },
              {
                "apiVersion": "baremetal.openstack.org/v1beta1",
                "kind": "OpenStackProvisionServer",
                "metadata": {
                  "name": "openstackprovisionserver"
                },
                "spec": {
                  "interface": "enp1s0",
                  "osImage": "edpm-hardened-uefi.qcow2",
                  "port": 8080
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:07Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenStackBaremetalSet is the Schema for the openstackbaremetalsets
              API
            displayName: OpenStack BaremetalSet
            kind: OpenStackBaremetalSet
            name: openstackbaremetalsets.baremetal.openstack.org
            version: v1beta1
          - description: OpenStackProvisionServer used to serve custom images for
              baremetal provisioning with Metal3
            displayName: OpenStackProvisionServer
            kind: OpenStackProvisionServer
            name: openstackprovisionservers.baremetal.openstack.org
            version: v1beta1
        description: An operator for handling baremetal provisioning concerns for
          OpenStack
        displayName: OpenStack Baremetal Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cn-openstack
        - openstack
        links:
        - name: Openstack Baremetal Operator
          url: https://github.com/openstack-k8s-operators/openstack-baremetal-operator
        maturity: beta
        provider:
          name: Red Hat
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/ubi9/httpd-24@sha256:dbfbe3383e12d63df7178e4cdd5b1e6e51a84c781041b9192026ea6edb72fbf3
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/openstack-baremetal-rhel9-operator@sha256:a6c69c03e59faf50973d07ccea32b80cb9c46ceb60ed00ffe7311ca42d501296
        - registry.redhat.io/rhoso/edpm-hardened-uefi-rhel9@sha256:e4bfacb9d8de056a251bee16a66b69631b96fb19fb926d8e33acd379d40a92d6
        - registry.redhat.io/rhoso-operators/openstack-baremetal-agent-rhel9@sha256:8f6f8e99042710819aa639eae6513f36c4642190ecd85d58a2ed819f1b55e38f
        version: 1.0.6
      entries:
      - name: openstack-baremetal-operator.v1.0.6
        version: 1.0.6
      - name: openstack-baremetal-operator.v1.0.5
        version: 1.0.5
      - name: openstack-baremetal-operator.v1.0.4
        version: 1.0.4
      - name: openstack-baremetal-operator.v1.0.3
        version: 1.0.3
      - name: openstack-baremetal-operator.v1.0.2
        version: 1.0.2
      - name: openstack-baremetal-operator.v1.0.1
        version: 1.0.1
      - name: openstack-baremetal-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: openstack-baremetal-operator
    provider:
      name: Red Hat
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-msg-broker
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-msg-broker.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Messagebroker",
                "metadata": {
                  "name": "messagebroker-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "chiselServer": {
                    "enabled": false,
                    "port": 8080
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": false
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "natsCore": {
                    "reloader": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats-server-config-reloader",
                      "tag": "0.8.3-3747-30"
                    },
                    "server": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats",
                      "tag": "2.9.0-3747-21"
                    }
                  },
                  "natsHttpProxy": {
                    "enabled": false,
                    "port": 8080
                  },
                  "natsStreaming": {
                    "image": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats-streaming",
                      "tag": "0.23.2-3747-31"
                    }
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "tagSuffix": "-ocp"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:31:32Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Messagebroker
            name: messagebrokers.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Message Broker Operator
        displayName: Keysight WAP Message Broker Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Msg Broker
          url: https://keysight-wap-msg-broker.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-msg-broker@sha256:4701b6e456b099697bf1e25aeb4bc8e32b1892bb0956e36d84f9f80666b3a1fc
        version: 0.1.21
      entries:
      - name: keysight-wap-msg-broker.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-msg-broker
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: VictoriaMetrics
      provider-url: ""
    name: victoriametrics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: victoriametrics-operator.v0.63.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLAgent",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "persistentVolumeClaimRetentionPolicy": {
                    "whenDeleted": "Delete"
                  },
                  "remoteWrite": [
                    {
                      "maxDiskUsage": "10GB",
                      "url": "http://vlsingle-example-0.default.svc:9428/internal/insert"
                    }
                  ],
                  "remoteWriteSettings": {
                    "maxBlockSize": "30MB"
                  },
                  "replicaCount": 2,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "350Mi"
                    }
                  },
                  "storage": {
                    "volumeClaimTemplate": {
                      "spec": {
                        "resources": {
                          "requests": {
                            "storage": "50Gi"
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "victoriametrics-operator"
                  },
                  "name": "example"
                },
                "spec": {
                  "requestsLoadBalancer": {
                    "enabled": true,
                    "spec": {}
                  },
                  "vlinsert": {
                    "replicaCount": 1
                  },
                  "vlselect": {
                    "replicaCount": 1
                  },
                  "vlstorage": {
                    "replicaCount": 2,
                    "retentionPeriod": "1y",
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "50Gi"
                            }
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLSingle",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "victoriametrics-operator"
                  },
                  "name": "sample"
                },
                "spec": {
                  "retentionMaxDiskSpaceUsageBytes": "20GB",
                  "retentionPeriod": "1y",
                  "storage": {
                    "resources": {
                      "requests": {
                        "storage": "50Gi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VLogs",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "retentionPeriod": "2"
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAgent",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "extraArgs": {
                    "memory.allowedPercent": "40"
                  },
                  "remoteWrite": [
                    {
                      "maxDiskUsage": "10GB",
                      "url": "http://vmsingle-dc-1.default.svc:8429/api/v1/write"
                    },
                    {
                      "maxDiskUsage": "15GB",
                      "url": "http://vmsingle-dc-2.default.svc:8429/api/v1/write"
                    },
                    {
                      "url": "http://vminsert-main.default.svc:8480/insert/1:1/prometheus/api/v1/write"
                    }
                  ],
                  "remoteWriteSettings": {
                    "maxDiskUsagePerURL": "5GB"
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "350Mi"
                    }
                  },
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlert",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "datasource": {
                    "url": "http://vmsingle-example-pvc.default.svc:8429"
                  },
                  "evaluationInterval": "30s",
                  "notifier": {
                    "url": "http://vmalertmanager-example.default.svc:9093"
                  },
                  "replicaCount": 1,
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlertmanager",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "configSecret": "am-global-config",
                  "replicaCount": 1,
                  "selectAllByDefault": true,
                  "storage": {
                    "volumeClaimTemplate": {
                      "spec": {
                        "resources": {
                          "requests": {
                            "storage": "2Gi"
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlertmanagerConfig",
                "metadata": {
                  "name": "example-email-web"
                },
                "spec": {
                  "receivers": [
                    {
                      "email_configs": [
                        {
                          "from": "alerting@example.com",
                          "smarthost": "example.com:25",
                          "text": "ALARM",
                          "to": "some-email@example.com"
                        }
                      ],
                      "name": "email",
                      "webhook_configs": [
                        {
                          "url": "http://some-other-wh"
                        }
                      ]
                    }
                  ],
                  "route": {
                    "group_interval": "1m",
                    "receiver": "email",
                    "routes": [
                      {
                        "receiver": "email"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAuth",
                "metadata": {
                  "name": "vmauth-sample"
                },
                "spec": {
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "victoriametrics/vmauth"
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "250m",
                      "memory": "350Mi"
                    }
                  },
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMCluster",
                "metadata": {
                  "name": "example-persistent"
                },
                "spec": {
                  "replicationFactor": 2,
                  "retentionPeriod": "4",
                  "vminsert": {
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "0.5",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "vmselect": {
                    "cacheMountPath": "/select-cache",
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "0.5",
                        "memory": "500Mi"
                      }
                    },
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "2Gi"
                            }
                          }
                        }
                      }
                    }
                  },
                  "vmstorage": {
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1500Mi"
                      }
                    },
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "10Gi"
                            }
                          }
                        }
                      }
                    },
                    "storageDataPath": "/vm-data"
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMNodeScrape",
                "metadata": {
                  "name": "cadvisor-metrics"
                },
                "spec": {
                  "bearerTokenFile": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                  "interval": "10s",
                  "relabelConfigs": [
                    {
                      "action": "labelmap",
                      "regex": "__meta_kubernetes_node_label_(.+)"
                    },
                    {
                      "replacement": "kubernetes.default.svc:443",
                      "targetLabel": "__address__"
                    },
                    {
                      "regex": "(.+)",
                      "replacement": "/api/v1/nodes/$1/proxy/metrics/cadvisor",
                      "sourceLabels": [
                        "__meta_kubernetes_node_name"
                      ],
                      "targetLabel": "__metrics_path__"
                    }
                  ],
                  "scheme": "https",
                  "scrapeTimeout": "2s",
                  "selector": {},
                  "tlsConfig": {
                    "caFile": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                    "insecureSkipVerify": true
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMPodScrape",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "podMetricsEndpoints": [
                    {
                      "port": "metrics",
                      "scheme": "http"
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app.kubernetes.io/name": "example-application",
                      "monitored-by": "vm-operator"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMProbe",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "interval": "2s",
                  "jobName": "static-probe",
                  "module": "http_2xx",
                  "targets": {
                    "staticConfig": {
                      "targets": [
                        "vmagent-example.default.svc:8429/health"
                      ]
                    }
                  },
                  "vmProberSpec": {
                    "url": "prometheus-blackbox-exporter.default.svc:9115"
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMRule",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "groups": [
                    {
                      "name": "kafka",
                      "rules": [
                        {
                          "alert": "coordinator down",
                          "annotations": {
                            "description": "kafka coordinator is down",
                            "value": "{{ $value }}"
                          },
                          "expr": "ml_app_gauge{exec_context=\"consumer_group_state\"} == 0",
                          "for": "60s",
                          "labels": {
                            "job": "{{ $labels.job }}",
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMScrapeConfig",
                "metadata": {
                  "name": "mongodb"
                },
                "spec": {
                  "consulSDConfigs": [
                    {
                      "server": "https://consul-dns:8500",
                      "services": [
                        "mongodb"
                      ]
                    }
                  ],
                  "relabelConfigs": [
                    {
                      "action": "replace",
                      "sourceLabels": [
                        "__meta_consul_service"
                      ],
                      "targetLabel": "job"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMServiceScrape",
                "metadata": {
                  "name": "example-scrape"
                },
                "spec": {
                  "endpoints": [
                    {
                      "port": "http"
                    }
                  ],
                  "namespaceSelector": {},
                  "selector": {
                    "matchLabels": {
                      "app.kubernetes.io/name": "example-application",
                      "monitored-by": "vm-operator"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMSingle",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "retentionPeriod": "1"
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMSingle",
                "metadata": {
                  "name": "example-pvc"
                },
                "spec": {
                  "removePvcAfterDelete": true,
                  "retentionPeriod": "1",
                  "storage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMStaticScrape",
                "metadata": {
                  "name": "vmstaticscrape-sample"
                },
                "spec": {
                  "jobName": "static",
                  "targetEndpoints": [
                    {
                      "labels": {
                        "env": "dev",
                        "project": "operator"
                      },
                      "targets": [
                        "192.168.0.1:9100",
                        "196.168.0.50:9100"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMUser",
                "metadata": {
                  "name": "vmuser-tenant-1"
                },
                "spec": {
                  "bearerToken": "some-token",
                  "targetRefs": [
                    {
                      "crd": {
                        "kind": "VMCluster/vminsert",
                        "name": "test-persistent",
                        "namespace": "vm"
                      },
                      "target_path_suffix": "/insert/1"
                    },
                    {
                      "crd": {
                        "kind": "VMCluster/vmselect",
                        "name": "test-persistent",
                        "namespace": "vm"
                      },
                      "target_path_suffix": "/select/1"
                    },
                    {
                      "paths": [
                        "/internal/resetRollupResultCache"
                      ],
                      "static": {
                        "url": "http://vmselect-test-persistent.default.svc:8481/"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: docker.io/victoriametrics/operator:v0.63.0
          createdAt: "2025-09-12T12:58:29Z"
          description: Provides monitoring capabilities for kubernetes clusters and
            applications
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: "1.30"
          operators.openshift.io/infrastructure-features: '[fips]'
          operators.operatorframework.io.bundle.channel.default.v1: beta
          operators.operatorframework.io.bundle.channels.v1: beta
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/VictoriaMetrics/operator
          support: VictoriaMetrics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VLAgent - is a tiny but brave agent, which helps you collect
              logs from various sources and stores them in VictoriaLogs.
            displayName: VLAgent
            kind: VLAgent
            name: vlagents.operator.victoriametrics.com
            version: v1
          - description: VLCluster is fast, cost-effective and scalable logs database.
            displayName: VLCluster
            kind: VLCluster
            name: vlclusters.operator.victoriametrics.com
            version: v1
          - description: |-
              VLogs is fast, cost-effective and scalable logs database.
              VLogs is the Schema for the vlogs API
            displayName: VLogs
            kind: VLogs
            name: vlogs.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VLSingle is fast, cost-effective and scalable logs database.
              VLSingle is the Schema for the API
            displayName: VLSingle
            kind: VLSingle
            name: vlsingles.operator.victoriametrics.com
            version: v1
          - description: |-
              VMAgent - is a tiny but brave agent, which helps you collect metrics from various sources and stores them in VictoriaMetrics
              or any other Prometheus-compatible storage system that supports the remote_write protocol.
            displayName: VMAgent
            kind: VMAgent
            name: vmagents.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlertmanagerConfig is the Schema for the vmalertmanagerconfigs
              API
            displayName: VMAlertmanager Config
            kind: VMAlertmanagerConfig
            name: vmalertmanagerconfigs.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlertmanager represents Victoria-Metrics deployment for
              Alertmanager.
            displayName: VMAlertmanager
            kind: VMAlertmanager
            name: vmalertmanagers.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlert  executes a list of given alerting or recording rules
              against configured address.
            displayName: VMAlert
            kind: VMAlert
            name: vmalerts.operator.victoriametrics.com
            version: v1beta1
          - description: VMAnomaly is the Schema for the vmanomalies API.
            displayName: VMAnomaly
            kind: VMAnomaly
            name: vmanomalies.operator.victoriametrics.com
            version: v1
          - description: VMAuth is the Schema for the vmauths API
            displayName: VMAuth
            kind: VMAuth
            name: vmauths.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMCluster is fast, cost-effective and scalable time-series database.
              Cluster version with
            displayName: VMCluster
            kind: VMCluster
            name: vmclusters.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMNodeScrape defines discovery for targets placed on kubernetes nodes,
              usually its node-exporters and other host services.
              InternalIP is used as __address__ for scraping.
            displayName: VMNode Scrape
            kind: VMNodeScrape
            name: vmnodescrapes.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMPodScrape is scrape configuration for pods,
              it generates vmagent's config for scraping pod targets
              based on selectors.
            displayName: VMPod Scrape
            kind: VMPodScrape
            name: vmpodscrapes.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMProbe defines a probe for targets, that will be executed with prober,
              like blackbox exporter.
              It helps to monitor reachability of target with various checks.
            displayName: VMProbe
            kind: VMProbe
            name: vmprobes.operator.victoriametrics.com
            version: v1beta1
          - description: VMRule defines rule records for vmalert application
            displayName: VMRule
            kind: VMRule
            name: vmrules.operator.victoriametrics.com
            version: v1beta1
          - description: VMScrapeConfig specifies a set of targets and parameters
              describing how to scrape them.
            displayName: VMScrape Config
            kind: VMScrapeConfig
            name: vmscrapeconfigs.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMServiceScrape is scrape configuration for endpoints associated with
              kubernetes service,
              it generates scrape configuration for vmagent based on selectors.
              result config will scrape service endpoints
            displayName: VMService Scrape
            kind: VMServiceScrape
            name: vmservicescrapes.operator.victoriametrics.com
            version: v1beta1
          - description: VMSingle  is fast, cost-effective and scalable time-series
              database.
            displayName: VMSingle
            kind: VMSingle
            name: vmsingles.operator.victoriametrics.com
            version: v1beta1
          - description: VMStaticScrape  defines static targets configuration for
              scraping.
            displayName: VMStatic Scrape
            kind: VMStaticScrape
            name: vmstaticscrapes.operator.victoriametrics.com
            version: v1beta1
          - description: VMUser is the Schema for the vmusers API
            displayName: VMUser
            kind: VMUser
            name: vmusers.operator.victoriametrics.com
            version: v1beta1
          - description: VTCluster is fast, cost-effective and scalable traces database.
            displayName: VTCluster
            kind: VTCluster
            name: vtclusters.operator.victoriametrics.com
            version: v1
          - description: |-
              VTSingle is fast, cost-effective and scalable traces database.
              VTSingle is the Schema for the API
            displayName: VTSingle
            kind: VTSingle
            name: vtsingles.operator.victoriametrics.com
            version: v1
        description: |
          Operator manages VictoriaMetrics applications and provides monitoring features for applications running inside and outside kubernetes cluster. It has support for prometheus-operator objects and
          provides migration mechanism.

          Operator designed to use existing installation of VictoriaMetrics inside or outside kubernetes cluster or it could
          bring own resources and fully manages them.

           You can read more about operator at docs:
            - quick start [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/quick-start.md)
            - high availability [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/high-availability.md)
            - design and description of implementation [design](https://github.com/VictoriaMetrics/operator/blob/master/docs/design.md)
            - operator objects description [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/api.md)

          # VictoriaMetrics
          * VictoriaMetrics can be used as long-term storage for Prometheus or for [vmagent](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmagent/README.md).
          See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#prometheus-setup) for details.
          * Supports [Prometheus querying API](https://prometheus.io/docs/prometheus/latest/querying/api/), so it can be used as Prometheus drop-in replacement in Grafana.
          VictoriaMetrics implements [MetricsQL](https://github.com/VictoriaMetrics/VictoriaMetrics/wiki/MetricsQL) query language, which is inspired by PromQL.
          * Supports global query view. Multiple Prometheus instances may write data into VictoriaMetrics. Later this data may be used in a single query.
          * High performance and good scalability for both [inserts](https://medium.com/@valyala/high-cardinality-tsdb-benchmarks-victoriametrics-vs-timescaledb-vs-influxdb-13e6ee64dd6b)
          and [selects](https://medium.com/@valyala/when-size-matters-benchmarking-victoriametrics-vs-timescale-and-influxdb-6035811952d4).
          [Outperforms InfluxDB and TimescaleDB by up to 20x](https://medium.com/@valyala/measuring-vertical-scalability-for-time-series-databases-in-google-cloud-92550d78d8ae).
          * [Uses 10x less RAM than InfluxDB](https://medium.com/@valyala/insert-benchmarks-with-inch-influxdb-vs-victoriametrics-e31a41ae2893) when working with millions of unique time series (aka high cardinality).
          * Optimized for time series with high churn rate. Think about [prometheus-operator](https://github.com/coreos/prometheus-operator) metrics from frequent deployments in Kubernetes.
          * High data compression, so [up to 70x more data points](https://medium.com/@valyala/when-size-matters-benchmarking-victoriametrics-vs-timescale-and-influxdb-6035811952d4)
          may be crammed into limited storage comparing to TimescaleDB.
          * Optimized for storage with high-latency IO and low IOPS (HDD and network storage in AWS, Google Cloud, Microsoft Azure, etc). See [graphs from these benchmarks](https://medium.com/@valyala/high-cardinality-tsdb-benchmarks-victoriametrics-vs-timescaledb-vs-influxdb-13e6ee64dd6b).
          * A single-node VictoriaMetrics may substitute moderately sized clusters built with competing solutions such as Thanos, M3DB, Cortex, InfluxDB or TimescaleDB.
          See [vertical scalability benchmarks](https://medium.com/@valyala/measuring-vertical-scalability-for-time-series-databases-in-google-cloud-92550d78d8ae),
          [comparing Thanos to VictoriaMetrics cluster](https://medium.com/@valyala/comparing-thanos-to-victoriametrics-cluster-b193bea1683)
          and [Remote Write Storage Wars](https://promcon.io/2019-munich/talks/remote-write-storage-wars/) talk
          from [PromCon 2019](https://promcon.io/2019-munich/talks/remote-write-storage-wars/).
          * Easy operation:
              * VictoriaMetrics consists of a single [small executable](https://medium.com/@valyala/stripping-dependency-bloat-in-victoriametrics-docker-image-983fb5912b0d) without external dependencies.
              * All the configuration is done via explicit command-line flags with reasonable defaults.
              * All the data is stored in a single directory pointed by `-storageDataPath` flag.
              * Easy and fast backups from [instant snapshots](https://medium.com/@valyala/how-victoriametrics-makes-instant-snapshots-for-multi-terabyte-time-series-data-e1f3fb0e0282)
              to S3 or GCS with [vmbackup](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmbackup/README.md) / [vmrestore](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmrestore/README.md).
              See [this article](https://medium.com/@valyala/speeding-up-backups-for-big-time-series-databases-533c1a927883) for more details.
          * Storage is protected from corruption on unclean shutdown (i.e. OOM, hardware reset or `kill -9`) thanks to [the storage architecture](https://medium.com/@valyala/how-victoriametrics-makes-instant-snapshots-for-multi-terabyte-time-series-data-e1f3fb0e0282).
          * Supports metrics' scraping, ingestion and [backfilling](https://github.com/VictoriaMetrics/VictoriaMetrics#backfilling) via the following protocols:
            * [Metrics from Prometheus exporters](https://github.com/prometheus/docs/blob/master/content/docs/instrumenting/exposition_formats.md#text-based-format)
              such as [node_exporter](https://github.com/prometheus/node_exporter). See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-scrape-prometheus-exporters-such-as-node-exporter) for details.
              * [Prometheus remote write API](https://prometheus.io/docs/prometheus/latest/configuration/configuration/#remote_write)
              * [InfluxDB line protocol](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-send-data-from-influxdb-compatible-agents-such-as-telegraf) over HTTP, TCP and UDP.
              * [Graphite plaintext protocol](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-send-data-from-graphite-compatible-agents-such-as-statsd) with [tags](https://graphite.readthedocs.io/en/latest/tags.html#carbon)
              if `-graphiteListenAddr` is set.
              * [OpenTSDB put message](https://github.com/VictoriaMetrics/VictoriaMetrics#sending-data-via-telnet-put-protocol) if `-opentsdbListenAddr` is set.
              * [HTTP OpenTSDB /api/put requests](https://github.com/VictoriaMetrics/VictoriaMetrics#sending-opentsdb-data-via-http-apiput-requests) if `-opentsdbHTTPListenAddr` is set.
              * [How to import time series data](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-time-series-data).
              * [Prometheus exposition format](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-data-in-prometheus-exposition-format).
              * [Arbitrary CSV data](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-csv-data).
          * Supports metrics' relabeling. See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#relabeling) for details.
          * Ideally works with big amounts of time series data from Kubernetes, IoT sensors, connected cars, industrial telemetry, financial data and various Enterprise workloads.
        displayName: VictoriaMetrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - prometheus
        - victoriametrics
        links:
        - name: VictoriaMetrics
          url: https://victoriametrics.com
        maintainers:
        - email: nik@victoriametrics.com
          name: f41gh7
        - email: info@victoriametrics.com
          name: info
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: VictoriaMetrics
        relatedImages:
        - docker.io/victoriametrics/operator:v0.63.0
        version: 0.63.0
      entries:
      - name: victoriametrics-operator.v0.63.0
        version: 0.63.0
      - name: victoriametrics-operator.v0.62.0
        version: 0.62.0
      - name: victoriametrics-operator.v0.61.2
        version: 0.61.2
      - name: victoriametrics-operator.v0.61.1
        version: 0.61.1
      - name: victoriametrics-operator.v0.61.0
        version: 0.61.0
      - name: victoriametrics-operator.v0.60.2
        version: 0.60.2
      - name: victoriametrics-operator.v0.60.1
        version: 0.60.1
      - name: victoriametrics-operator.v0.60.0
        version: 0.60.0
      - name: victoriametrics-operator.v0.59.2
        version: 0.59.2
      - name: victoriametrics-operator.v0.59.1
        version: 0.59.1
      - name: victoriametrics-operator.v0.59.0
        version: 0.59.0
      - name: victoriametrics-operator.v0.58.1
        version: 0.58.1
      - name: victoriametrics-operator.v0.58.0
        version: 0.58.0
      - name: victoriametrics-operator.v0.57.0
        version: 0.57.0
      - name: victoriametrics-operator.v0.56.0
        version: 0.56.0
      - name: victoriametrics-operator.v0.55.0
        version: 0.55.0
      - name: victoriametrics-operator.v0.54.1
        version: 0.54.1
      - name: victoriametrics-operator.v0.53.0
        version: 0.53.0
      - name: victoriametrics-operator.v0.52.0
        version: 0.52.0
      - name: victoriametrics-operator.v0.51.1
        version: 0.51.1
      - name: victoriametrics-operator.v0.49.0
        version: 0.49.0
      - name: victoriametrics-operator.v0.48.4
        version: 0.48.4
      - name: victoriametrics-operator.v0.48.3
        version: 0.48.3
      - name: victoriametrics-operator.v0.48.1
        version: 0.48.1
      - name: victoriametrics-operator.v0.47.3
        version: 0.47.3
      - name: victoriametrics-operator.v0.47.1
        version: 0.47.1
      - name: victoriametrics-operator.v0.47.0
        version: 0.47.0
      - name: victoriametrics-operator.v0.46.4
        version: 0.46.4
      - name: victoriametrics-operator.v0.46.3
        version: 0.46.3
      - name: victoriametrics-operator.v0.39.4
        version: 0.39.4
      - name: victoriametrics-operator.v0.33.0
        version: 0.33.0
      - name: victoriametrics-operator.v0.32.1
        version: 0.32.1
      - name: victoriametrics-operator.v0.32.0
        version: 0.32.0
      - name: victoriametrics-operator.v0.29.1
        version: 0.29.1
      - name: victoriametrics-operator.v0.27.2
        version: 0.27.2
      - name: victoriametrics-operator.v0.26.2
        version: 0.26.2
      - name: victoriametrics-operator.v0.26.0
        version: 0.26.0
      - name: victoriametrics-operator.v0.25.1
        version: 0.25.1
      - name: victoriametrics-operator.v0.25.0
        version: 0.25.0
      - name: victoriametrics-operator.v0.24.0
        version: 0.24.0
      - name: victoriametrics-operator.v0.23.3
        version: 0.23.3
      - name: victoriametrics-operator.v0.22.1
        version: 0.22.1
      - name: victoriametrics-operator.v0.22.0
        version: 0.22.0
      - name: victoriametrics-operator.v0.21.0
        version: 0.21.0
      - name: victoriametrics-operator.v0.20.1
        version: 0.20.1
      - name: victoriametrics-operator.v0.19.1
        version: 0.19.1
      - name: victoriametrics-operator.v0.19.0
        version: 0.19.0
      - name: victoriametrics-operator.v0.18.2
        version: 0.18.2
      - name: victoriametrics-operator.v0.18.1
        version: 0.18.1
      - name: victoriametrics-operator.v0.18.0
        version: 0.18.0
      name: beta
    defaultChannel: beta
    packageName: victoriametrics-operator
    provider:
      name: VictoriaMetrics
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: F5 Networks Inc.
      provider-url: ""
    name: f5-bigip-ctlr-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: alpha
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: beta
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: stable
    defaultChannel: stable
    packageName: f5-bigip-ctlr-operator
    provider:
      name: F5 Networks Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dns-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dns-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSHealthCheckProbe",
                "metadata": {
                  "name": "$NAME"
                },
                "spec": {
                  "additionalHeadersRef": {
                    "name": "headers"
                  },
                  "address": "192.168.0.16",
                  "allowInsecureCertificate": true,
                  "failureThreshold": 5,
                  "hostname": "test.com",
                  "interval": "60s",
                  "path": "/healthz",
                  "port": 443,
                  "protocol": "HTTPS"
                }
              },
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSRecord",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dns-operator",
                    "app.kubernetes.io/instance": "dnsrecord-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dnsrecord",
                    "app.kubernetes.io/part-of": "dns-operator"
                  },
                  "name": "dnsrecord-sample"
                },
                "spec": {
                  "endpoints": [
                    {
                      "dnsName": "dnsrecord-simple.example.com",
                      "recordTTL": 60,
                      "recordType": "A",
                      "targets": [
                        "52.215.108.61",
                        "52.30.101.221"
                      ]
                    }
                  ],
                  "providerRef": {
                    "name": "dns-provider-creds"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/dns-rhel9-operator@sha256:8ad3876e36ba2f76355499d98a97672f54f26754fc9d219d9cb1c5b9f9874430
          createdAt: "2025-04-07T08:13:58Z"
          description: Configures how north-south traffic from outside the network
            is balanced and reaches Gateways
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kuadrant/dns-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSHealthCheckProbe is the Schema for the dnshealthcheckprobes
              API
            displayName: DNSHealthCheckProbe
            kind: DNSHealthCheckProbe
            name: dnshealthcheckprobes.kuadrant.io
            version: v1alpha1
          - description: DNSRecord is the Schema for the dnsrecords API
            displayName: DNSRecord
            kind: DNSRecord
            name: dnsrecords.kuadrant.io
            version: v1alpha1
        description: Configures how north-south traffic from outside the network is
          balanced and reaches Gateways - Part of Red Hat Connectivity Link
        displayName: Red Hat - DNS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dns
        - kuadrant
        links:
        - name: GitHub
          url: https://github.com/kuadrant/dns-operator
        - name: Documentation
          url: https://github.com/Kuadrant/dns-operator/tree/main/docs
        maintainers:
        - email: mnairn@redhat.com
          name: Michael Nairn
        - email: pbrookes@redhat.com
          name: Phil Brookes
        - email: cbrookes@redhat.com
          name: Craig Brookes
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhcl-1/dns-rhel9-operator@sha256:8ad3876e36ba2f76355499d98a97672f54f26754fc9d219d9cb1c5b9f9874430
        version: 1.1.0
      entries:
      - name: dns-operator.v1.1.0
        version: 1.1.0
      - name: dns-operator.v1.0.2
        version: 1.0.2
      - name: dns-operator.v1.0.1
        version: 1.0.1
      - name: dns-operator.v0.12.0
        version: 0.12.0
      name: stable
    defaultChannel: stable
    packageName: dns-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecr-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecr-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.services.k8s.aws/v1alpha1",
                "kind": "Repository",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecr-controller:1.3.1
          createdAt: "2025-09-20T01:21:24Z"
          description: AWS ECR controller is a service controller for managing ECR
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PullThroughCacheRule represents the state of an AWS ecr PullThroughCacheRule
              resource.
            displayName: PullThroughCacheRule
            kind: PullThroughCacheRule
            name: pullthroughcacherules.ecr.services.k8s.aws
            version: v1alpha1
          - description: Repository represents the state of an AWS ecr Repository
              resource.
            displayName: Repository
            kind: Repository
            name: repositories.ecr.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Container Registry (ECR) resources in AWS from within your Kubernetes cluster.

          **About Amazon ECR**

          Amazon Elastic Container Registry (Amazon ECR) is an AWS managed container image registry service that is secure, scalable, and reliable. Amazon ECR supports private repositories with resource-based permissions using AWS IAM. This is so that specified users or Amazon EC2 instances can access your container repositories and images. You can use your preferred CLI to push, pull, and manage Docker images, Open Container Initiative (OCI) images, and OCI compatible artifacts.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECR
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECR Developer Resources
          url: https://aws.amazon.com/ecr/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecr maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecr-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-ecr-controller.v1.3.1
        version: 1.3.1
      - name: ack-ecr-controller.v1.3.0
        version: 1.3.0
      - name: ack-ecr-controller.v1.2.0
        version: 1.2.0
      - name: ack-ecr-controller.v1.1.0
        version: 1.1.0
      - name: ack-ecr-controller.v1.0.33
        version: 1.0.33
      - name: ack-ecr-controller.v1.0.32
        version: 1.0.32
      - name: ack-ecr-controller.v1.0.31
        version: 1.0.31
      - name: ack-ecr-controller.v1.0.30
        version: 1.0.30
      - name: ack-ecr-controller.v1.0.29
        version: 1.0.29
      - name: ack-ecr-controller.v1.0.28
        version: 1.0.28
      - name: ack-ecr-controller.v1.0.27
        version: 1.0.27
      - name: ack-ecr-controller.v1.0.25
        version: 1.0.25
      - name: ack-ecr-controller.v1.0.24
        version: 1.0.24
      - name: ack-ecr-controller.v1.0.23
        version: 1.0.23
      - name: ack-ecr-controller.v1.0.22
        version: 1.0.22
      - name: ack-ecr-controller.v1.0.21
        version: 1.0.21
      - name: ack-ecr-controller.v1.0.20
        version: 1.0.20
      - name: ack-ecr-controller.v1.0.19
        version: 1.0.19
      - name: ack-ecr-controller.v1.0.17
        version: 1.0.17
      - name: ack-ecr-controller.v1.0.16
        version: 1.0.16
      - name: ack-ecr-controller.v1.0.15
        version: 1.0.15
      - name: ack-ecr-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecr-controller.v1.0.12
        version: 1.0.12
      - name: ack-ecr-controller.v1.0.10
        version: 1.0.10
      - name: ack-ecr-controller.v1.0.9
        version: 1.0.9
      - name: ack-ecr-controller.v1.0.8
        version: 1.0.8
      - name: ack-ecr-controller.v1.0.7
        version: 1.0.7
      - name: ack-ecr-controller.v1.0.6
        version: 1.0.6
      - name: ack-ecr-controller.v1.0.4
        version: 1.0.4
      - name: ack-ecr-controller.v1.0.3
        version: 1.0.3
      - name: ack-ecr-controller.v1.0.2
        version: 1.0.2
      - name: ack-ecr-controller.v1.0.0
        version: 1.0.0
      - name: ack-ecr-controller.v0.1.9
        version: 0.1.9
      - name: ack-ecr-controller.v0.1.7
        version: 0.1.7
      - name: ack-ecr-controller.v0.1.6
        version: 0.1.6
      - name: ack-ecr-controller.v0.1.5
        version: 0.1.5
      - name: ack-ecr-controller.v0.1.4
        version: 0.1.4
      - name: ack-ecr-controller.v0.1.3
        version: 0.1.3
      - name: ack-ecr-controller.v0.1.2
        version: 0.1.2
      - name: ack-ecr-controller.v0.1.1
        version: 0.1.1
      - name: ack-ecr-controller.v0.1.0
        version: 0.1.0
      - name: ack-ecr-controller.v0.0.27
        version: 0.0.27
      - name: ack-ecr-controller.v0.0.26
        version: 0.0.26
      - name: ack-ecr-controller.v0.0.25
        version: 0.0.25
      - name: ack-ecr-controller.v0.0.24
        version: 0.0.24
      - name: ack-ecr-controller.v0.0.23
        version: 0.0.23
      - name: ack-ecr-controller.v0.0.22
        version: 0.0.22
      - name: ack-ecr-controller.v0.0.21
        version: 0.0.21
      - name: ack-ecr-controller.v0.0.20
        version: 0.0.20
      - name: ack-ecr-controller.v0.0.19
        version: 0.0.19
      - name: ack-ecr-controller.v0.0.18
        version: 0.0.18
      - name: ack-ecr-controller.v0.0.17
        version: 0.0.17
      - name: ack-ecr-controller.v0.0.16
        version: 0.0.16
      - name: ack-ecr-controller.v0.0.15
        version: 0.0.15
      - name: ack-ecr-controller.v0.0.14
        version: 0.0.14
      - name: ack-ecr-controller.v0.0.13
        version: 0.0.13
      - name: ack-ecr-controller.v0.0.12
        version: 0.0.12
      - name: ack-ecr-controller.v0.0.11
        version: 0.0.11
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecr-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-elbv2-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-elbv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "LoadBalancer",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "TargetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/elbv2-controller:1.1.1
          createdAt: "2025-09-19T22:16:38Z"
          description: AWS ELB controller is a service controller for managing ELB
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Listener represents the state of an AWS elbv2 Listener resource.
            displayName: Listener
            kind: Listener
            name: listeners.elbv2.services.k8s.aws
            version: v1alpha1
          - description: LoadBalancer represents the state of an AWS elbv2 LoadBalancer
              resource.
            displayName: LoadBalancer
            kind: LoadBalancer
            name: loadbalancers.elbv2.services.k8s.aws
            version: v1alpha1
          - description: Rule represents the state of an AWS elbv2 Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.elbv2.services.k8s.aws
            version: v1alpha1
          - description: TargetGroup represents the state of an AWS elbv2 TargetGroup
              resource.
            displayName: TargetGroup
            kind: TargetGroup
            name: targetgroups.elbv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Load Balancing (ELB) resources in AWS from within your Kubernetes cluster.

          **About Amazon ELB**

          Elastic Load Balancing automatically distributes your incoming traffic across multiple targets, such as EC2 instances, containers, and IP addresses, in one or more Availability Zones. It monitors the health of its registered targets, and routes traffic only to the healthy targets. Elastic Load Balancing scales your load balancer capacity automatically in response to changes in incoming traffic.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ELB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elbv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ELB Developer Resources
          url: https://aws.amazon.com/elasticloadbalancing
        maintainers:
        - email: ack-maintainers@amazon.com
          name: elb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/elbv2-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-elbv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-elbv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-elbv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-elbv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-elbv2-controller.v1.0.10
        version: 1.0.10
      - name: ack-elbv2-controller.v1.0.9
        version: 1.0.9
      - name: ack-elbv2-controller.v1.0.8
        version: 1.0.8
      - name: ack-elbv2-controller.v1.0.7
        version: 1.0.7
      - name: ack-elbv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-elbv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-elbv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-elbv2-controller.v1.0.3
        version: 1.0.3
      - name: ack-elbv2-controller.v1.0.2
        version: 1.0.2
      - name: ack-elbv2-controller.v1.0.1
        version: 1.0.1
      - name: ack-elbv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-elbv2-controller.v0.0.11
        version: 0.0.11
      - name: ack-elbv2-controller.v0.0.10
        version: 0.0.10
      - name: ack-elbv2-controller.v0.0.9
        version: 0.0.9
      - name: ack-elbv2-controller.v0.0.8
        version: 0.0.8
      - name: ack-elbv2-controller.v0.0.6
        version: 0.0.6
      - name: ack-elbv2-controller.v0.0.5
        version: 0.0.5
      - name: ack-elbv2-controller.v0.0.4
        version: 0.0.4
      - name: ack-elbv2-controller.v0.0.3
        version: 0.0.3
      - name: ack-elbv2-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-elbv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hazelcast, Inc
      provider-url: ""
    name: hazelcast-platform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hazelcast-platform-operator.v5.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "cache-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "CronHotBackup",
                "metadata": {
                  "name": "cronhotbackup-sample"
                },
                "spec": {
                  "hotBackupTemplate": {
                    "metadata": {
                      "labels": {
                        "cron-hotbackup": "true"
                      }
                    },
                    "spec": {
                      "hazelcastResourceName": "hazelcast"
                    }
                  },
                  "schedule": "*/30 * * * *"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast"
                },
                "spec": {
                  "clusterSize": 3,
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast-sample"
                },
                "spec": {
                  "clusterSize": 3,
                  "jet": {
                    "bucketConfig": {
                      "bucketURI": "gs://operator-user-code/jetJobs",
                      "secretName": "br-secret-gcp"
                    },
                    "enabled": true,
                    "resourceUploadEnabled": true
                  },
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HazelcastEndpoint",
                "metadata": {
                  "name": "hazelcastendpoint-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "port": 5701,
                  "type": "Discovery"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HotBackup",
                "metadata": {
                  "name": "hot-backup"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-longrun"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-longrun-2.0.0.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-1.0.2.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJobSnapshot",
                "metadata": {
                  "name": "jetjobsnapshot-sample"
                },
                "spec": {
                  "cancelJob": false,
                  "jetJobResourceName": "jet-job"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ManagementCenter",
                "metadata": {
                  "name": "managementcenter"
                },
                "spec": {
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "hazelcastClusters": [
                    {
                      "address": "hazelcast",
                      "name": "dev"
                    }
                  ],
                  "licenseKeySecretName": "hazelcast-license-key",
                  "persistence": {
                    "enabled": true,
                    "size": "10Gi"
                  },
                  "repository": "hazelcast/management-center"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Map",
                "metadata": {
                  "name": "map"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "MultiMap",
                "metadata": {
                  "name": "multimap-sample"
                },
                "spec": {
                  "backupCount": 3,
                  "binary": true,
                  "collectionType": "LIST",
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "queue-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ReplicatedMap",
                "metadata": {
                  "name": "replicatedmap-sample"
                },
                "spec": {
                  "asyncFillup": true,
                  "hazelcastResourceName": "hazelcast",
                  "inMemoryFormat": "OBJECT"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "topic-sample"
                },
                "spec": {
                  "globalOrderingEnabled": true,
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "UserCodeNamespace",
                "metadata": {
                  "name": "ucn-sample"
                },
                "spec": {
                  "bucketConfig": {
                    "bucketURI": "gs://operator-user-code/userCodeNamespace",
                    "secretName": "br-secret-gcp"
                  },
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanReplication",
                "metadata": {
                  "name": "wanreplication-sample"
                },
                "spec": {
                  "endpoints": "35.192.33.252",
                  "resources": [
                    {
                      "kind": "Map",
                      "name": "map-sample"
                    },
                    {
                      "kind": "Hazelcast",
                      "name": "hz-sample"
                    }
                  ],
                  "targetClusterName": "dev"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanSync",
                "metadata": {
                  "name": "wansync-sample"
                },
                "spec": {
                  "wanReplicationResourceName": "wan-replication-sample"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: docker.io/hazelcast/hazelcast-platform-operator@sha256:99005afda5c71ceeac200928dfed4ea35480f871053101e3aeaa44005e3fda14
          createdAt: "2024-07-29T15:35:56Z"
          description: '[DEPRECATED] Use the certified Hazelcast Operator instead'
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hazelcast/hazelcast-platform-operator/
          support: Hazelcast, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.hazelcast.com
            version: v1alpha1
          - description: CronHotBackup is the Schema for the cronhotbackups API
            displayName: Cron Hot Backup
            kind: CronHotBackup
            name: cronhotbackups.hazelcast.com
            version: v1alpha1
          - description: HazelcastEndpoint is the Schema for the hazelcastendpoints
              API
            displayName: Hazelcast Endpoint
            kind: HazelcastEndpoint
            name: hazelcastendpoints.hazelcast.com
            version: v1alpha1
          - description: Hazelcast is the Schema for the hazelcasts API
            displayName: Hazelcast
            kind: Hazelcast
            name: hazelcasts.hazelcast.com
            version: v1alpha1
          - description: HotBackup is the Schema for the hot backup API
            displayName: Hot Backup
            kind: HotBackup
            name: hotbackups.hazelcast.com
            version: v1alpha1
          - description: JetJob is the Schema for the jetjobs API
            displayName: Jet Job
            kind: JetJob
            name: jetjobs.hazelcast.com
            version: v1alpha1
          - description: JetJobSnapshot is the Schema for the jetjobsnapshots API
            displayName: Jet Job Snapshot
            kind: JetJobSnapshot
            name: jetjobsnapshots.hazelcast.com
            version: v1alpha1
          - description: ManagementCenter is the Schema for the managementcenters
              API
            displayName: Management Center
            kind: ManagementCenter
            name: managementcenters.hazelcast.com
            version: v1alpha1
          - description: Map is the Schema for the maps API
            displayName: Map
            kind: Map
            name: maps.hazelcast.com
            version: v1alpha1
          - description: MultiMap is the Schema for the multimaps API
            displayName: Multi Map
            kind: MultiMap
            name: multimaps.hazelcast.com
            version: v1alpha1
          - description: Queue is the Schema for the queues API
            displayName: Queue
            kind: Queue
            name: queues.hazelcast.com
            version: v1alpha1
          - description: ReplicatedMap is the Schema for the replicatedmaps API
            displayName: Replicated Map
            kind: ReplicatedMap
            name: replicatedmaps.hazelcast.com
            version: v1alpha1
          - description: Topic is the Schema for the topics API
            displayName: Topic
            kind: Topic
            name: topics.hazelcast.com
            version: v1alpha1
          - description: UserCodeNamespace is the Schema for the usercodenamespaces
              API
            displayName: User Code Namespace
            kind: UserCodeNamespace
            name: usercodenamespaces.hazelcast.com
            version: v1alpha1
          - description: WanReplication is the Schema for the wanreplications API
            displayName: Wan Replication
            kind: WanReplication
            name: wanreplications.hazelcast.com
            version: v1alpha1
          - description: WanSync is the Schema for the wansyncs API
            displayName: Wan Sync
            kind: WanSync
            name: wansyncs.hazelcast.com
            version: v1alpha1
        description: |
          ## Deprecated
          Use the certified Hazelcast Operator instead.
          # Hazelcast Platform Operator #

          Easily deploy Hazelcast clusters and Management Center into Kubernetes environments and manage their lifecycles.

          ## Before You Start

          If you are planning to create Hazelcast Platform Enterprise clusters, you need to [create a secret](https://docs.hazelcast.com/operator/latest/get-started#step-2-start-the-hazelcast-cluster) for the license. You can request a trial license key from [here](https://trialrequest.hazelcast.com).

          For Hazelcast Platform clusters, you can simply continue.

          ## Documentation

          1. [Get started](https://docs.hazelcast.com/operator/latest/get-started) with the Operator
          2. [Connect to the cluster from outside Kubernetes](https://guides.hazelcast.org/hazelcast-platform-operator-expose-externally)
          3. [Restore a Cluster from Cloud Storage with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-external-backup-restore)
          4. [Replicate Data between Two Hazelcast Clusters with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-wan-replication)

          ## Features

          Hazelcast Platform Operator supports the features below:

          * Custom resource for Hazelcast Platform (Enterprise) and Management Center
          * Observe status of Hazelcast clusters and Management Center instances
          * High Availability Mode configuration to create clusters that are resilient to node and zone failures
          * Support for TLS connections between members using self-signed certificates
          * Scale up and down Hazelcast clusters
          * Expose Hazelcast cluster to external clients ([Smart & Unisocket](https://docs.hazelcast.com/hazelcast/latest/clients/java#java-client-operation-modes))
          * Backup Hazelcast persistence data to cloud storage with the possibility of scheduling it and restoring the data accordingly
          * WAN Replication feature when you need to synchronize multiple Hazelcast clusters, which are connected by WANs
          * Full and Delta WAN Sync support
          * Tiered Storage support for Hazelcast and Map CRs
          * CP Subsystem configuration support for Hazelcast CR
          * User Code Deployment feature, which allows you to deploy custom and domain classes from cloud storages and URLs to Hazelcast members
          * The User Code Namespaces feature allows you to deploy custom and domain classes at runtime without a members restart
          * Support the configuration of advanced networking options
          * Support Multi-namespace configuration
          * ExecutorService and EntryProcessor support
          * Support several data structures like Map, Topic, MultiMap, ReplicatedMap, Queue and Cache which can be created dynamically via specific Custom Resources
          * MapStore, Near Cache and off-heap memory (HD memory and native memory) support for the Map CR
          * Native Memory support for the Cache CR
          * Support Jet configuration and Jet Job submission using the JetJob CR
          * Support for exporting the snapshots of JetJob CRs using JetJobSnapshot CR
          * Support for custom configurations using ConfigMap
        displayName: '[DEPRECATED] Hazelcast Platform Operator'
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hazelcast
        - keyvalue
        - in-memory
        - database
        - caching
        links:
        - name: Documentation
          url: https://docs.hazelcast.com/operator/latest/
        - name: Source Code
          url: https://github.com/hazelcast/hazelcast-platform-operator
        maintainers:
        - email: cloudnative@hazelcast.com
          name: Hazelcast Cloud Native Team
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Hazelcast, Inc
        relatedImages:
        - docker.io/hazelcast/hazelcast-platform-operator@sha256:99005afda5c71ceeac200928dfed4ea35480f871053101e3aeaa44005e3fda14
        version: 5.13.1
      entries:
      - name: hazelcast-platform-operator.v5.13.1
        version: 5.13.1
      - name: hazelcast-platform-operator.v5.13.0
        version: 5.13.0
      - name: hazelcast-platform-operator.v5.12.0
        version: 5.12.0
      - name: hazelcast-platform-operator.v5.11.0
        version: 5.11.0
      - name: hazelcast-platform-operator.v5.10.0
        version: 5.10.0
      - name: hazelcast-platform-operator.v5.9.0
        version: 5.9.0
      - name: hazelcast-platform-operator.v5.8.0
        version: 5.8.0
      - name: hazelcast-platform-operator.v5.7.0
        version: 5.7.0
      - name: hazelcast-platform-operator.v5.6.0
        version: 5.6.0
      - name: hazelcast-platform-operator.v5.5.0
        version: 5.5.0
      - name: hazelcast-platform-operator.v5.4.0
        version: 5.4.0
      - name: hazelcast-platform-operator.v5.3.0
        version: 5.3.0
      - name: hazelcast-platform-operator.v5.2.0
        version: 5.2.0
      - name: hazelcast-platform-operator.v5.1.0
        version: 5.1.0
      - name: hazelcast-platform-operator.v5.0.0
        version: 5.0.0
      name: alpha
    defaultChannel: alpha
    packageName: hazelcast-platform-operator
    provider:
      name: Hazelcast, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-storage
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-storage.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Storage",
                "metadata": {
                  "name": "storage-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore"
                  },
                  "minio": {
                    "clusterDomain": "cluster.local",
                    "helmKubectlJqImage": {
                      "repository": "docker.io/bskim45/helm-kubectl-jq"
                    },
                    "image": {
                      "repository": "prod.harbor.keysight.digital/loadcore/wap-minio",
                      "tag": "RELEASE.2021-12-29T06-49-06Z-3747-37-ocp"
                    },
                    "mcImage": {
                      "repository": "prod.harbor.keysight.digital/loadcore/wap-minio-mc",
                      "tag": "RELEASE.2021-12-29T06-52-55Z-3747-37-ocp"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "securityContext": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:32:20Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Storage
            name: storages.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Storage Operator
        displayName: Keysight WAP Storage Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Storage
          url: https://keysight-wap-storage.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-storage@sha256:9f7710cfe72658cd5456fd54f86ca6be5c9a4de2a01114c0f7531b70712bd981
        version: 0.1.21
      entries:
      - name: keysight-wap-storage.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-storage
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: patch-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: patch-operator.v0.1.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Patch",
                "metadata": {
                  "name": "test-complex-patch"
                },
                "spec": {
                  "patches": {
                    "test-complex-patch": {
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 1).metadata.name }}: {{ (index . 2).metadata.name }}    \n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "kind": "Namespace",
                          "name": "test-patch-operator"
                        },
                        {
                          "apiVersion": "v1",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "test-patch-operator"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "test-patch-operator"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/patch-operator@sha256:030ade9b9428deca3ac6390618dc4d6e6d728577d1467a22dc4d4b68c9848886
          createdAt: "2024-05-10T14:07:19Z"
          description: An operator to apply patches to Kubernetes objects in a declarative
            way.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: patch-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/patch-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Patch
            name: patches.redhatcop.redhat.io
            version: v1alpha1
        description: "The patch operator helps with defining patches in a declarative
          way. This operator has two main features:\n\n1. [ability to patch an object
          at creation time via a mutating webhook](#creation-time-patch-injection)\n2.
          [ability to enforce patches on one or more objects via a controller](#runtime-patch-enforcement)\n
          \n## Creation-time patch injection\n\nWhy apply a patch at creation time
          when you could directly create the correct object? The reason is that sometime
          the correct value depends on configuration set on the specific cluster in
          which the object is being deployed. For example, an ingress/route hostname
          might depend on the specific cluster. Consider the following example based
          on cert-manager:\n \n```yaml\napiVersion: cert-manager.io/v1\nkind: ClusterIssuer\nmetadata:\n
          \ name: letsencrypt-issuer\nspec:\n  acme:\n    server: 'https://acme-v02.api.letsencrypt.org/directory'\n
          \   email: {{ .Values.letsencrypt.email }}\n    privateKeySecretRef:\n      name:
          letsencrypt-staging\n    solvers:  \n    - dns01:\n        route53:\n          accessKeyID:
          << access_key >>\n          secretAccessKeySecretRef:\n            name:
          cert-manager-dns-credentials\n            key: aws_secret_access_key\n          region:
          << region >>\n          hostedZoneID: << hosted_zone_id >>\n ```\n\nIn this
          example the fields: `<< access_key >>`, `<< region >>` and `<< hosted_zone_id
          >>` are dependent on the specific region in which the cluster is being deployed
          and in many cases they are discoverable from other configurations already
          present in the cluster. If you want to deploy the above Cluster Issuer object
          with a gitops approach, then there is no easy way to discover those values.
          The solution so far is to manually discover those values and create a different
          gitops configuration for each cluster. But consider if you could look up
          values at deploy time based on the cluster you are deploying to. Here is
          how this object might look:\n\n```yaml\napiVersion: cert-manager.io/v1\nkind:
          ClusterIssuer\nmetadata:\n  name: letsencrypt-issuer\n  namespace: {{ .Release.Namespace
          }}\n  annotations:\n    \"redhat-cop.redhat.io/patch\": |\n      spec:\n
          \       acme:\n        - dns01:\n            route53:\n              accessKeyID:
          {{ (lookup \"v1\" \"Secret\" .metadata.namespace \"cert-manager-dns-credentials\").data.aws_access_key_id
          | b64dec }}\n              secretAccessKeySecretRef:\n                name:
          cert-manager-dns-credentials\n                key: aws_secret_access_key\n
          \             region: {{ (lookup \"config.openshift.io/v1\" \"Infrastructure\"
          \"\" \"cluster\").status.platformStatus.aws.region }}\n              hostedZoneID:
          {{ (lookup \"config.openshift.io/v1\" \"DNS\" \"\" \"cluster\").spec.publicZone.id
          }} \nspec:\n  acme:\n    server: 'https://acme-v02.api.letsencrypt.org/directory'\n
          \   email: {{ .Values.letsencrypt.email }}\n    privateKeySecretRef:\n      name:
          letsencrypt-staging\n    solvers:  \n    - dns01:\n        route53:\n          accessKeyID:
          << access_key >>\n          secretAccessKeySecretRef:\n            name:
          cert-manager-dns-credentials\n            key: aws_secret_access_key\n          region:
          << region >>\n          hostedZoneID: << hosted_zone_id >>\n ```\n\nThe
          annotation specifies a patch that will be applied by a [MutatingWebhook](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/),
          as you can see the three values are now looked up from different configurations
          in the cluster.\n\nTwo annotations influences the behavior of this MutatingWebhook\n
          \n1. \"redhat-cop.redhat.io/patch\" : this is the patch itself. The patch
          is evaluated as a template with the object itself as it's only parameter.
          The template is expressed in golang template notation and supports the same
          functions as helm template including the [lookup](https://helm.sh/docs/chart_template_guide/functions_and_pipelines/#using-the-lookup-function)
          function which plays a major role here. The patch must be expressed in yaml
          for readability. It will be converted to json by the webhook logic.\n2.
          \"redhat-cop.redhat.io/patch-type\" : this is the type of json patch. The
          possible values are: `application/json-patch+json`, `application/merge-patch+json`
          and `application/strategic-merge-patch+json`. If this annotation is omitted
          it defaults to strategic merge.\n\n### Security Considerations\n \nThe lookup
          function, if used by the template, is executed with a client which impersonates
          the user issuing the object creation/update request. This should prevent
          security permission leakage.\n\n### Installing the creation time webhook\n
          \nThe creation time webhook is not installed by the operator. This is because
          there is no way to know which specific object type should be intercepted
          and intercepting all of the types would be too inefficient. It's up to the
          administrator then to install the webhook. Here is some guidance.\n\nIf
          you installed the operator via OLM, use the following webhook template:\n\n```yaml\napiVersion:
          admissionregistration.k8s.io/v1\n kind: MutatingWebhookConfiguration\nmetadata:\n
          \ name: patch-operator-inject\n  annotations:\n    service.beta.openshift.io/inject-cabundle:
          \"true\"\nwebhooks:\n - admissionReviewVersions:\n  - v1\n  clientConfig:\n
          \   service:\n      name: patch-operator-webhook-service\n      namespace:
          patch-operator\n      path: /inject\n  failurePolicy: Fail\n  name: patch-operator-inject.redhatcop.redhat.io\n
          \ rules:\n  - << add your intercepted objects here >>\n  sideEffects: None\n```\n\nIf
          you installed the operator via the Helm chart and are using cert-manager,
          use the following webhook template:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\n
          kind: MutatingWebhookConfiguration\nmetadata:\n  name: patch-operator-inject\n
          \ annotations:\n    cert-manager.io/inject-ca-from: '{{ .Release.Namespace
          }}/webhook-server-cert'\n webhooks:\n- admissionReviewVersions:\n  - v1\n
          \ clientConfig:\n    service:\n      name: patch-operator-webhook-service\n
          \     namespace: patch-operator\n      path: /inject\n  failurePolicy: Fail\n
          \ name: patch-operator-inject.redhatcop.redhat.io\n  rules:\n  - << add
          your intercepted objects here >>\n  sideEffects: None\n```  \n\nYou should
          need to enable the webhook only for `CREATE` operations. So for example
          to enable the webhook on configmaps:\n\n```yaml\n  rules:\n  - apiGroups:\n
          \   - \"\"\n    apiVersions:\n    - v1\n    operations:\n    - CREATE\n
          \   resources:\n    - configmaps\n```\n\n## Runtime patch enforcement\n\nThere
          are situations when we need to patch pre-existing objects. Again this is
          a use case that is hard to model with gitops operators which will work only
          on object that they own. Especially with sophisticated Kubernetes distributions,
          it is not uncommon that a Kubernetes instance, at installation time, is
          configured with some default settings. Changing those configurations means
          patching those objects. For example, let's take the case of OpenShift Oauth
          configuration. This object is present by default and it is expected to be
          patched with any newly enabled authentication mechanism. This is how it
          looks like after installation:\n\n```yaml\napiVersion: config.openshift.io/v1\n
          kind: OAuth\nmetadata:\n  name: cluster\n  ownerReferences:\n    - apiVersion:
          config.openshift.io/v1\n      kind: ClusterVersion\n      name: version\n
          \     uid: 9a9d450b-3076-4e30-ac05-a889d6341fc3\n  resourceVersion: '20405124'\n
          spec: []\n```\n\nIf we need to patch it we can use the patch controller
          and the `Patch` object as follows:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: Patch\nmetadata:\n  name: gitlab-ocp-oauth-provider\n  namespace:
          openshift-config\nspec:\n  serviceAccountRef:\n    name: default\n  patches:\n
          \   gitlab-ocp-oauth-provider:\n      targetObjectRef:\n        apiVersion:
          config.openshift.io/v1\n        kind: OAuth\n        name: cluster\n      patchTemplate:
          |\n        spec:\n          identityProviders:\n          - name: my-github
          \n            mappingMethod: claim \n            type: GitHub\n            github:\n
          \             clientID: \"{{ (index . 1).data.client_id | b64dec }}\" \n
          \             clientSecret: \n                name: ocp-github-app-credentials\n
          \             organizations: \n              - my-org\n              teams:
          []            \n      patchType: application/merge-patch+json\n      sourceObjectRefs:\n
          \     - apiVersion: v1\n        kind: Secret\n        name: ocp-github-app-credentials\n
          \       namespace: openshift-config\n```\n\nThis will cause the OAuth object
          to be patched and the patch to be enforced. That means that if anything
          changes on the secret that we use a parameter (which may be rotated) or
          in Oauth object itself, the patch will be reapplied. In this case we are
          adding a gitlab authentication provider.\n\n A `patch` has the following
          fields:\n\n`targetObjectRef` this refers to the object(s) receiving the
          patch. Mutliple object can be selected based on the following rules:\n \n|
          Namespaced Type | Namespace | Name | Selection type |\n| --- | --- | ---
          | --- |\n| yes | null | null | multiple selection across namespaces |\n|
          yes | null | not null | multiple selection across namespaces where the name
          corresponds to the passed name |\n| yes | not null | null | multiple selection
          within a namespace |\n| yes | not null | not nul | single selection |\n|
          no | N/A | null | multiple selection  |\n| no | N/A | not null | single
          selection |\n\nSelection can be further narrowed down by filtering by labels
          and/or annotations using the `labelSelector` and `annotationSelector` fields.
          The patch will be applied to all of the selected instances.\n\n`sourceObjectRefs`
          these are the objects that will be watched and become part of the parameters
          of the patch template. Name and Namespace of sourceRefObjects are interpreted
          as golang templates with the current target instance and the only parameter.
          This allows to select different source object for each target object.\n
          \nSo, for example, with this patch:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: Patch\nmetadata:\n  name: multiple-namespaced-targets-patch\nspec:\n
          \ serviceAccountRef:\n    name: default\n  patches:\n    multiple-namespaced-targets-patch:\n
          \     targetObjectRef:\n        apiVersion: v1\n        kind: ServiceAccount\n
          \       name: deployer\n      patchTemplate: |\n        metadata:\n          annotations:\n
          \           {{ (index . 1).metadata.uid }}: {{ (index . 1) }}\n      patchType:
          application/strategic-merge-patch+json\n      sourceObjectRefs:\n      -
          apiVersion: v1\n        kind: ServiceAccount\n        name: default\n        namespace:
          \"{{ .metadata.namespace }}\"\n        fieldPath: $.metadata.uid\n```\n\nThe
          `deployer` service accounts from all namespaces are selected as target of
          this patch, each patch template will receive a different parameter and that
          is the `default` service account of the same namespace as the namespace
          of the `deployer` service account being processed.\n\n`sourceObjectRefs`
          also have the `fieldPath` field which can contain a jsonpath expression.
          If a value is passed the jsonpath expression will be calculate for the current
          source object and the result will be passed as parameter of the template.\n\n`patchTemplate`
          This is the the template that will be evaluated. The result must be a valid
          patch compatible with the requested type and expressed in yaml for readability.
          The parameters passed to the template are the target object and then the
          all of the source object. So if you want to refer to the target object in
          the template you can use this expression `(index . 0)`. Higher indexes refer
          to the sourceObjectRef array. The template is expressed in golang template
          notation and supports the same functions as helm template.\n\n`patchType`
          is the type of the json patch. The possible values are: `application/json-patch+json`,
          `application/merge-patch+json` and `application/strategic-merge-patch+json`.
          If this annotation is omitted it defaults to strategic merge.\n\n### Patch
          Controller Security Considerations\n\nThe patch enforcement enacted by the
          patch controller is executed with a client which uses the service account
          referenced by the `serviceAccountRef` field. So before a patch object can
          actually work an administrator must have granted the needed permissions
          to a service account in the same namespace. The `serviceAccountRef` will
          default to the `default` service account if not specified.\n\n### Patch
          Controller Performance Considerations\n\nThe patch controller will create
          a controller-manager and per `Patch` object and a reconciler for each of
          the `PatchSpec` defined in the array on patches in the `Patch` object.\nThese
          reconcilers share the same cached client. In order to be able to watch changes
          on target and source objects of a `PatchSpec`, all of the target and source
          object type instances will be cached by the client. This is a normal behavior
          of a controller-manager client, but it implies that if you create patches
          on object types that have many instances in etcd (Secrets, ServiceAccounts,
          Namespaces for example), the patch operator instance will require a significant
          amount of memory. A way to contain this issue is to try to aggregate together
          `PatchSpec` that deal with the same object types. This will cause those
          object type instances to cached only once.  \n"
        displayName: Patch Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - patch
        - enforcing
        - declarative
        links:
        - name: Patch Operator
          url: https://github.com/redhat-cop/patch-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/patch-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/patch-operator@sha256:030ade9b9428deca3ac6390618dc4d6e6d728577d1467a22dc4d4b68c9848886
        version: 0.1.11
      entries:
      - name: patch-operator.v0.1.11
        version: 0.1.11
      - name: patch-operator.v0.1.10
        version: 0.1.10
      - name: patch-operator.v0.1.9
        version: 0.1.9
      - name: patch-operator.v0.1.8
        version: 0.1.8
      - name: patch-operator.v0.1.7
        version: 0.1.7
      - name: patch-operator.v0.1.6
        version: 0.1.6
      - name: patch-operator.v0.1.5
        version: 0.1.5
      - name: patch-operator.v0.1.4
        version: 0.1.4
      - name: patch-operator.v0.1.3
        version: 0.1.3
      - name: patch-operator.v0.1.2
        version: 0.1.2
      - name: patch-operator.v0.1.1
        version: 0.1.1
      - name: patch-operator.v0.1.0
        version: 0.1.0
      - name: patch-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: patch-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-bedrockagent-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-bedrockagent-controller.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bedrockagent.services.k8s.aws/v1alpha1",
                "kind": "Agent",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/bedrockagent-controller:0.1.1
          createdAt: "2025-09-20T01:18:56Z"
          description: AWS Bedrock Agent controller is a service controller for managing
            Bedrock Agent resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Agent represents the state of an AWS bedrockagent Agent resource.
            displayName: Agent
            kind: Agent
            name: agents.bedrockagent.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Bedrock Agent resources in AWS from within your Kubernetes cluster.

          **About Amazon Bedrock Agent**

          Amazon Bedrock is a fully managed service that makes high-performing foundation models (FMs) from leading AI companies and Amazon available for your use through a unified API. You can choose from a wide range of foundation models to find the model that is best suited for your use case. Amazon Bedrock also offers a broad set of capabilities to build generative AI applications with security, privacy, and responsible AI. Using Amazon Bedrock, you can easily experiment with and evaluate top foundation models for your use cases, privately customize them with your data using techniques such as fine-tuning and Retrieval Augmented Generation (RAG), and build agents that execute tasks using your enterprise systems and data sources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Bedrock Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bedrockagent
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Bedrock Agent Developer Resources
          url: https://aws.amazon.com/bedrock/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: bedrock-agent maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/bedrockagent-controller:0.1.1
        version: 0.1.1
      entries:
      - name: ack-bedrockagent-controller.v0.1.1
        version: 0.1.1
      - name: ack-bedrockagent-controller.v0.1.0
        version: 0.1.0
      - name: ack-bedrockagent-controller.v0.0.6
        version: 0.0.6
      - name: ack-bedrockagent-controller.v0.0.5
        version: 0.0.5
      - name: ack-bedrockagent-controller.v0.0.4
        version: 0.0.4
      - name: ack-bedrockagent-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-bedrockagent-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: devopstales
      provider-url: devopstales.github.io
    name: community-trivy-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: community-trivy-operator.v2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "trivy-operator.devopstales.io/v1",
                "kind": "NamespaceScanner",
                "metadata": {
                  "name": "trivy-operator-main-config"
                },
                "spec": {
                  "crontab": "*/5 * * * *",
                  "namespace_selector": "trivy-scan",
                  "clusterWide": "False",
                  "policyreport": "False",
                  "registry": [
                    {
                      "name": "docker.io",
                      "user": "",
                      "password": ""
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: devopstales/trivy-operator:2.4.1
          createdAt: "2022-07-11T15:00:00Z"
          description: Trivy Operator for scheduled imagescans and an Admission Control.
          repository: https://github.com/devopstales/trivy-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Confifuration for trivy-operator.
            displayName: NamespaceScanner
            kind: NamespaceScanner
            name: namespace-scanners.trivy-operator.devopstales.io
            version: v1
          - description: |
              VulnerabilityReport summarizes vulnerabilities in application dependencies and operating system packages
              built into container images.
            displayName: VulnerabilityReport
            kind: VulnerabilityReport
            name: vulnerabilityreports.trivy-operator.devopstales.io
            version: v1
        description: |
          Trivy Operator for scheduled imagescans and an Admission Control.

          ### Add Github token to extend your Github API Rate Limit
          The Operator use Github to download Trivy database. If your Github API Rate Limit Exceeded add your Github token to the operator deployment as GITHUB_TOKEN variable.

          ```yaml
          ...
          env:
          - name: GITHUB_TOKEN
            value: "githubToken"
          ```

          ### Scheduled Image scans
          Default every 5 minutes execute a scan script. It will get image list from all namespaces with the label `trivy-scan=true`, and then scan this images with trivy, finally we will get metrics on `http://[pod-ip]:9115/metrics`

          You can chane the schedule of the image scanning by setting the `crontab` variable in the namespace-scanner object.


          ### Cache Scan results
          From the 2.4 version you can use an external redis to store scanning results and speed up the scanning process.

          ```bash
          ...
          env:
          - name: REDIS_ENABLED
            value: "True"
          - name: REDIS_BACKEND
            value: redis://redis-cache:6379
          ```

          ### Persist Trivy database cache
          To Persist Trivy cache you need to create a pvc and mount to the Deployment under `/home/trivy-operator/trivy-cache`

          ```yaml
          ...
            volumeMounts:
            - name: cache
              mountPath: "/home/trivy-operator/trivy-cache"
          volumes:
          - name: cache
            persistentVolumeClaim:
              claimName: trivy-cache
          ```

          ### Trivy Image Validator
          The admission controller function can be configured as a ValidatingWebhook in a k8s cluster. Kubernetes will send requests to the admission server when a Pod creation is initiated. The admission controller checks the image using trivy if it is in a namespace with the lable `trivy-operator-validation=true`.

          To enable the admission controller you need to set the `clusterWide` variable to `true` in the namespace-scanner object.

          Enable the scheduled scan and the admission controller for a specific namespace:
          ```bash
          kubectl label namespaces test-apps trivy-scan=true
          # or
          kubectl label namespaces test-apps trivy-operator-validation=true
          ```

          You can define policy to the Admission Controller, by adding annotation to the pod trough the deployment:

          ```bash
          spec:
            ...
            template:
              metadata:
                annotations:
                  trivy.security.devopstales.io/medium: "5"
                  trivy.security.devopstales.io/low: "10"
                  trivy.security.devopstales.io/critical: "2"
          ...
          ```

          ### Monitorng
          Trivy operator has a prometheus endpoint wher yo can get the resoults of the namespace scans and the adission controller.

          ```bash
          curl -s http://10.43.179.39:9115/metrics | grep so_vulnerabilities

          # HELP so_vulnerabilities Container vulnerabilities
          # TYPE so_vulnerabilities gauge
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="UNKNOWN"} 0
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="LOW"} 23
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="MEDIUM"} 93
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="HIGH"} 76
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="CRITICAL"} 25
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="UNKNOWN"} 0
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="LOW"} 23
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="MEDIUM"} 88
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="HIGH"} 60
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="CRITICAL"} 8
          ```

          ```bash
          curl -s http://10.43.179.39:9115/metrics | grep ac_vulnerabilities

          # HELP ac_vulnerabilities Admission Controller vulnerabilities
          # TYPE ac_vulnerabilities gauge
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="UNKNOWN"} 0.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="LOW"} 83.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="MEDIUM"} 6.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="HIGH"} 6.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="CRITICAL"} 4.0
          ```

          ### Policy Report reation
          The PolicyReport object is a protopype object probosed by the Kubernetes policy work group. The Policy Report Custom Resource Definition (CRD) can be used as a common way to provide policy results to Kubernetes cluster administrators and users, using native tools.

          Add the PolicyReport CRDs to your cluster (v1alpha2):
          ```bash
          kubectl create -f https://github.com/kubernetes-sigs/wg-policy-prototypes/raw/master/policy-report/crd/v1alpha2/wgpolicyk8s.io_policyreports.yaml
          ```

          You can enable the Policy Report creatin by setting the `policyreport` to `True` in the NamespaceScanner object.

          This objects can be visualized by the Policy Reporter UI. The Policy Reporter UI is a monitoring and Observability Tool for the PolicyReport CRD with an optional UI. It is created by Kyverno. The main goal was a tool to visualize the resoluts of the Kyverno policies, but because it uses the PolicyReports CRD it can visualize the resoults of the trivy-operator scans.
        displayName: Community Trivy Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trivy
        - security
        links:
        - name: Documentation
          url: https://devopstales.github.io/trivy-operator
        - name: Blog
          url: https://devopstales.github.io
        maintainers:
        - email: devopstales@protonmail.com
          name: devopstales
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: devopstales
          url: devopstales.github.io
        relatedImages:
        - devopstales/trivy-operator:2.4.1
        version: 2.4.0
      entries:
      - name: community-trivy-operator.v2.4.0
        version: 2.4.0
      - name: community-trivy-operator.v2.3.0
        version: 2.3.0
      - name: community-trivy-operator.v2.2.0
        version: 2.2.0
      - name: community-trivy-operator.v2.1.1
        version: 2.1.1
      name: stable
    defaultChannel: stable
    packageName: community-trivy-operator
    provider:
      name: devopstales
      url: devopstales.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: www.ibm.com
    name: ibm-elasticsearch-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-elasticsearch-operator.v1.1.2667
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
          createdAt: "2025-01-30"
          description: Elasticsearch operator for use within Cloudpak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-elasticsearch-operator
          olm.skipRange: <0.0.0 >1.1.0
          operators.openshift.io/valid-subscription: '["IBM Elasticsearch"]'
          operators.operatorframework.io/internal-objects: '["elasticsearchclusters.elasticsearch.opencontent.ibm.com"]'
          operators.operatorframework.io/operator-type: non-standalone
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ElasticsearchCluster
            name: elasticsearchclusters.elasticsearch.opencontent.ibm.com
            version: v1
        description: |-
          # Introduction
          An Elasticsearch operator for use with IBM Cloudpak services.
          # Details
          ibm-elasticsearch-operator supports installing Elasticsearch major version 6 and 7. This operator supports two CRDs Elasticsearch and Elasticsearchdetails. Currently Elasticsearch is an abstraction of Elasticsearchdetails, but there is no difference between both the CRDs. Creating either of them provisions elasticsearch instances. This abstraction layer has been created to support any changes to the operator with out breaking the CRD/API for adopting products.

          ## Prerequisites
          - Openshift 3.11+
          ## Resources Required
          In short the answer is it depends. Especially if you are running Elasticsearch 6 you need 3 or a greater odd number of replicas to make a cluster and not have split brain. Each replica will have a sidecar that needs 0.1 cpu and 2gb of ram. You should size Elasticsearch based on your use-case. The default is a full cpu and 4 cpu and 64 gb ram. To assist sizing requirements we have included a copy of [esrally](https://esrally.readthedocs.io/en/stable/) on all the docker images.
          ## Installing
          ElasticsearchCluster is the strategic Kind moving forward. There are two flavors of CRD supported that will soon be deprecated. A simple Elasticseach "All in one" crd and a more detailed CRD for teams who need greater control. Upon deploy the elasticsearch is available from the service (not the headless) created by the operator. Only https will be available on port 443. the service will also be in the ElasticsearchDetails status.

          ### Case install
          To install the component using case you must either use the native install or catalog install. Upon installing the catalog you can then use either OLM or ODLM to install but must create the OperatorGroup & Subscription for either method to avoid conflict.

          # Configuration
          Architecture:
            - x86
            - s390x
            - ppc64le
          Storage: RWO Encrypted Block Storage with 120+ IOPS/GB
          Clouds:
          - IBM CLoud (ibm-block-gold, Gen2 custom iops storage class)
          - AWS (EBS)
          - Azure (For Azure please ensure you are using a higher tier of AzureDisk storage)
          - OCS (RBD not Ceph)
          - Portworx (Non-shared replacated)
          - Please do not use with NFS

          ### Limitations
          This product is not deployable standalone and it is part of the LICCR already accepted
          ## SecurityContextConstraints Requirements
          This component requires a restricted SCC.
        displayName: IBM OpenContent Elasticsearch
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - internal
        - Elasticsearch
        maintainers:
        - email: noreply@ibm.com
          name: IBM CloudpakOpen Community
        maturity: alpha
        provider:
          name: IBM
          url: www.ibm.com
        relatedImages:
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:47a3745c78280f98f2ee352bb9192cdcd2fcd6d4c1a3627e7aa42178a0f107fd
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:30a5b2c4e3c231373fa18d8aaab16d3505cc11a8a5d4be36eaf99d7c0dee35f2
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:2668a562eb8df1117c40acbc3600d6735469d210e5aae4aa03bf9f54d0ebfe46
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:2c39ff1850b7a9d32a11e99393039ef74ab363f3fffec1e4fd542f2ae148a6b4
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:130872119947fe91cbfb8dd82b3baaf78f7adcaca5933ca394eca3f9ee84c8cc
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
        version: 1.1.2667
      entries:
      - name: ibm-elasticsearch-operator.v1.1.2667
        version: 1.1.2667
      name: v1.1
    defaultChannel: v1.1
    packageName: ibm-elasticsearch-operator
    provider:
      name: IBM
      url: www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: ovms-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openvino-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "ModelServer",
                "metadata": {
                  "name": "model-server-sample"
                },
                "spec": {
                  "image_name": "registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0",
                  "deployment_parameters": {
                    "replicas": 1
                  },
                  "service_parameters": {
                    "grpc_port": 8080,
                    "rest_port": 8081
                  },
                  "models_settings": {
                    "single_model_mode": true,
                    "config_configmap_name": "",
                    "config_path": "",
                    "model_name": "resnet",
                    "model_path": "gs://<bucket_name>/<model_dir>",
                    "nireq": 0,
                    "plugin_config": "{\"PERFORMANCE_HINT\":\"LATENCY\"}",
                    "batch_size": "",
                    "shape": "",
                    "model_version_policy": "{\"latest\": { \"num_versions\":1 }}",
                    "layout": "",
                    "target_device": "CPU",
                    "is_stateful": false,
                    "idle_sequence_cleanup": false,
                    "low_latency_transformation": true,
                    "max_sequence_number": 0
                  },
                  "server_settings": {
                    "file_system_poll_wait_seconds": 0,
                    "sequence_cleaner_poll_wait_minutes": 0,
                    "log_level": "INFO",
                    "grpc_workers": 1,
                    "rest_workers": 0
                  },
                  "models_repository": {
                    "https_proxy": "",
                    "http_proxy": "",
                    "storage_type": "google",
                    "models_host_path": "",
                    "models_volume_claim": "",
                    "aws_secret_access_key": "",
                    "aws_access_key_id": "",
                    "aws_region": "",
                    "s3_compat_api_endpoint": "",
                    "gcp_creds_secret_name": "",
                    "azure_storage_connection_string": ""
                  }
                }
              },
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "Notebook",
                "metadata": {
                  "name": "v2024.1"
                },
                "spec": {
                  "git_ref": "2024.1",
                  "auto_update_image": false
                }
              }
            ]
          alm-examples-metadata: '{ "model-server-sample":{"description":"Example
            of the model server configuration"}, "v2024":{"description":"Default notebook
            configuration"} }'
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
          createdAt: "2024-05-14T09:29:35Z"
          description: OpenVINO Toolkit Operator manages OpenVINO components in OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.34.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Provides inference serving through gRPC & REST
            displayName: Model Server
            kind: ModelServer
            name: modelservers.intel.com
            version: v1alpha1
          - description: Adds a collection of ready-to-run Jupyter notebooks for learning
              and experimenting with the OpenVINO™ Toolkit to Red Hat® OpenShift®
              Data Science
            displayName: Notebook
            kind: Notebook
            name: notebooks.intel.com
            version: v1alpha1
        description: |
          OpenVINO Toolkit Operator manages OpenVINO components in OpenShift.

          Currently there available components are ModelServer and Notebook.

          # Model Server
          [OpenVINO™ Model Server](https://github.com/openvinotoolkit/model_server) (OVMS) is a scalable, high-performance solution for serving machine learning models optimized for Intel® architectures. The server provides an inference service via gRPC or REST API for any models trained in a framework that is supported by [OpenVINO](https://software.intel.com/en-us/openvino-toolki://docs.openvino.ai/latest/index.html).
          Model Server configuration parameters is explained [here](https://github.com/openvinotoolkit/operator/blob/main/docs/modelserver_params.md).
          ## Using the cluster
          OpenVINO Model Server can be consumed as a `Service`. It is called like with `ModelServer` resource with `-ovms` suffix.
          The suffix is omitted when `ovms` phrase is included in the name.
          The service exposes gRPC and REST API interfaces to run inference requests.
          ```
          oc get pods
          NAME                                        READY   STATUS    RESTARTS   AGE
          model-server-sample-ovms-586f6f76df-dpps4   1/1     Running   0          8h

          oc get services
          NAME                       TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)             AGE
          model-server-sample-ovms   ClusterIP   172.25.199.210   <none>        8080/TCP,8081/TCP   8h
          ```

          # Notebook
          The Notebook resource integrates JupyterHub from OpenShift Data Science or Open Data Hub with a container image that includes [developer
          tools](https://github.com/openvinotoolkit/openvino/blob/master/docs/install_guides/pypi-openvino-dev.md)
          from the OpenVINO toolkit and a set of Jupyter notebook tutorials. It enables selecting a defined image `OpenVINO™ Toolkit` from
          the Jupyter Spawner choice list.

          Create the `Notebook` resource in the same project with JupyterHub and RedHat OpenShift Data Science operator as seen in [documentation](https://github.com/openvinotoolkit/operator/blob/main/docs/notebook_in_rhods.md).
          It builds the image in the cluster based on Dockerfile from [openvino_notebooks](https://github.com/openvinotoolkit/openvino_notebooks).

          ## References
          OpenVINO Model Server on [Github](https://github.com/openvinotoolkit/model_server)

          OpenVINO Model Server Operator on [Github](https://github.com/openvinotoolkit/operator)
        displayName: OpenVINO Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Inference OpenVINO
        links:
        - name: OpenVINO Toolkit Operator
          url: https://github.com/openvinotoolkit/operator
        - name: OpenVINO Model Server
          url: https://github.com/openvinotoolkit/model_server
        - name: OpenVINO Notebooks
          url: https://github.com/openvinotoolkit/openvino_notebooks
        maintainers:
        - email: openvino-operator-support@intel.com
          name: Operator Support
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5714b93ac39277affdc10c3ba142d8c032ed0ddb2ba61c04cbddac710464c4
        - registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
        - registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0
        version: 1.2.0
      entries:
      - name: openvino-operator.v1.2.0
        version: 1.2.0
      - name: openvino-operator.v1.1.0
        version: 1.1.0
      - name: openvino-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: ovms-operator
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx-openshift-operator.v0.0.33
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
          createdAt: "2019-07-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx-operator-rhmp/support?utm_source=openshift_console
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX on JBoss EAP 7
            displayName: Joget DX on JBoss EAP 7
            kind: JogetDX
            name: jogetdx.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX7/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
        - registry.connect.redhat.com/joget/joget-dx7-eap7@sha256:7245501f5fccf011d75c034b7d55542cfe3bd2c64cc7e48b3224a48b623af282
        version: 0.0.33
      entries:
      - name: joget-dx-openshift-operator.v0.0.33
        version: 0.0.33
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx-operator-rhmp
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/cluster-nfd-operator
    name: openshift-nfd-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfd.v4.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nfd.openshift.io/v1",
                "kind": "NodeFeatureDiscovery",
                "metadata": {
                  "name": "nfd-instance",
                  "namespace": "openshift-nfd"
                },
                "spec": {
                  "customConfig": {
                    "configData": "#    - name: \"more.kernel.features\"\n#      matchOn:\n#      - loadedKMod: [\"example_kmod3\"]\n#    - name: \"more.features.by.nodename\"\n#      value: customValue\n#      matchOn:\n#      - nodename: [\"special-.*-node-.*\"]\n"
                  },
                  "operand": {
                    "image": "quay.io/openshift/origin-node-feature-discovery:4.10",
                    "servicePort": 12000
                  },
                  "workerConfig": {
                    "configData": "core:\n#  labelWhiteList:\n#  noPublish: false\n  sleepInterval: 60s\n#  sources: [all]\n#  klog:\n#    addDirHeader: false\n#    alsologtostderr: false\n#    logBacktraceAt:\n#    logtostderr: true\n#    skipHeaders: false\n#    stderrthreshold: 2\n#    v: 0\n#    vmodule:\n##   NOTE: the following options are not dynamically run-time \n##          configurable and require a nfd-worker restart to take effect\n##          after being changed\n#    logDir:\n#    logFile:\n#    logFileMaxSize: 1800\n#    skipLogHeaders: false\nsources:\n#  cpu:\n#    cpuid:\n##     NOTE: whitelist has priority over blacklist\n#      attributeBlacklist:\n#        - \"BMI1\"\n#        - \"BMI2\"\n#        - \"CLMUL\"\n#        - \"CMOV\"\n#        - \"CX16\"\n#        - \"ERMS\"\n#        - \"F16C\"\n#        - \"HTT\"\n#        - \"LZCNT\"\n#        - \"MMX\"\n#        - \"MMXEXT\"\n#        - \"NX\"\n#        - \"POPCNT\"\n#        - \"RDRAND\"\n#        - \"RDSEED\"\n#        - \"RDTSCP\"\n#        - \"SGX\"\n#        - \"SSE\"\n#        - \"SSE2\"\n#        - \"SSE3\"\n#        - \"SSE4.1\"\n#        - \"SSE4.2\"\n#        - \"SSSE3\"\n#      attributeWhitelist:\n#  kernel:\n#    kconfigFile: \"/path/to/kconfig\"\n#    configOpts:\n#      - \"NO_HZ\"\n#      - \"X86\"\n#      - \"DMI\"\n  pci:\n    deviceClassWhitelist:\n      - \"0200\"\n      - \"03\"\n      - \"12\"\n    deviceLabelFields:\n#      - \"class\"\n      - \"vendor\"\n#      - \"device\"\n#      - \"subsystem_vendor\"\n#      - \"subsystem_device\"\n#  usb:\n#    deviceClassWhitelist:\n#      - \"0e\"\n#      - \"ef\"\n#      - \"fe\"\n#      - \"ff\"\n#    deviceLabelFields:\n#      - \"class\"\n#      - \"vendor\"\n#      - \"device\"\n#  custom:\n#    - name: \"my.kernel.feature\"\n#      matchOn:\n#        - loadedKMod: [\"example_kmod1\", \"example_kmod2\"]\n#    - name: \"my.pci.feature\"\n#      matchOn:\n#        - pciId:\n#            class: [\"0200\"]\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#        - pciId :\n#            vendor: [\"8086\"]\n#            device: [\"1000\", \"1100\"]\n#    - name: \"my.usb.feature\"\n#      matchOn:\n#        - usbId:\n#          class: [\"ff\"]\n#          vendor: [\"03e7\"]\n#          device: [\"2485\"]\n#        - usbId:\n#          class: [\"fe\"]\n#          vendor: [\"1a6e\"]\n#          device: [\"089a\"]\n#    - name: \"my.combined.feature\"\n#      matchOn:\n#        - pciId:\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#          loadedKMod : [\"vendor_kmod1\", \"vendor_kmod2\"]\n"
                  }
                }
              },
              {
                "apiVersion": "nfd.openshift.io/v1alpha1",
                "kind": "NodeFeatureRule",
                "metadata": {
                  "name": "my-sample-rule-object"
                },
                "spec": {
                  "rules": [
                    {
                      "labels": {
                        "my-sample-feature": "true"
                      },
                      "matchFeatures": [
                        {
                          "feature": "kernel.loadedmodule",
                          "matchExpressions": {
                            "dummy": {
                              "op": "Exists"
                            }
                          }
                        },
                        {
                          "feature": "kernel.config",
                          "matchExpressions": {
                            "X86": {
                              "op": "In",
                              "value": [
                                "y"
                              ]
                            }
                          }
                        }
                      ],
                      "name": "my sample rule"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery,OpenShift Optional
          containerImage: quay.io/openshift/origin-cluster-nfd-operator:4.10
          description: |-
            The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

            NFD consists  of the following software components:

            The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

            ##NFD-Master
            NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

            ##NFD-Worker
            NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

            ##NFD-Topology-Updater
            NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
          olm.skipRange: '>=4.6.0 <4.11.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-nfd
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/internal-objects: '["noderesourcetopologies.topology.node.k8s.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/cluster-nfd-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              The NodeFeatureDiscovery instance is the CustomResource being watched by the NFD-Operator, and holds all the needed information to setup the behaviour of the master and worker pods
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
          - description: |
              NodeFeatureRule resource specifies a configuration for feature-based customization of node objects, such as node labeling.
            kind: NodeFeatureRule
            name: nodefeaturerules.nfd.openshift.io
            version: v1alpha1
          - description: |
              NodeResourceTopology resource describes node resources and their topology.
            kind: NodeResourceTopology
            name: noderesourcetopologies.topology.node.k8s.io
            version: v1alpha1
        description: |-
          The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

          NFD consists  of the following software components:

          The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

          ##NFD-Master
          NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

          ##NFD-Worker
          NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

          ##NFD-Topology-Updater
          NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
        displayName: Node Feature Discovery Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - feature-discovery
        - feature-detection
        - node-labels
        links:
        - name: Node Feature Discovery Operator
          url: https://docs.openshift.com/container-platform/4.9/hardware_enablement/psap-node-feature-discovery-operator.html
        - name: Node Feature Discovery Documentation
          url: https://kubernetes-sigs.github.io/node-feature-discovery/stable/get-started/index.html
        maintainers:
        - email: openshift-psap@redhat.com
          name: openshift-psap
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/cluster-nfd-operator
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy
        - quay.io/openshift/origin-cluster-nfd-operator:4.10
        version: 4.10.0
      entries:
      - name: nfd.v4.10.0
        version: 4.10.0
      - name: nfd.v4.9.0
        version: 4.9.0
      - name: nfd.v4.8.0
        version: 4.8.0
      name: stable
    defaultChannel: stable
    packageName: openshift-nfd-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/cluster-nfd-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kepler Operator Contributors
      provider-url: https://sustainable-computing.io/
    name: kepler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kepler-operator.v0.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "Kepler",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "kepler",
                    "app.kubernetes.io/name": "kepler",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "kepler"
                },
                "spec": {
                  "exporter": {
                    "deployment": {
                      "port": 9103
                    },
                    "redfish": null
                  }
                }
              },
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "PowerMonitor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "powermonitor",
                    "app.kubernetes.io/name": "powermonitor",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "power-monitor"
                },
                "spec": {
                  "kepler": {
                    "config": {
                      "logLevel": "info"
                    },
                    "deployment": {
                      "security": {
                        "mode": "none"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: quay.io/sustainable_computing_io/kepler-operator:0.20.0
          createdAt: "2025-07-18T06:13:30Z"
          description: Deploys and Manages Kepler on Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: |-
            [
              "keplerinternals.kepler.system.sustainable.computing.io",
              "powermonitorinternals.kepler.system.sustainable.computing.io"
            ]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/sustainable-computing-io/kepler-operator
          support: https://github.com/sustainable-computing-io/kepler-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KeplerInternal is the Schema for internal/unsupported API
            displayName: KeplerInternal
            kind: KeplerInternal
            name: keplerinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: Kepler is deprecated and will be removed in a future release.
              Please use PowerMonitor instead.
            displayName: Kepler
            kind: Kepler
            name: keplers.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitorInternal is the Schema for internal/unsupported
              API
            displayName: PowerMonitorInternal
            kind: PowerMonitorInternal
            name: powermonitorinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitor is the Schema for the powermonitors API
            displayName: PowerMonitor
            kind: PowerMonitor
            name: powermonitors.kepler.system.sustainable.computing.io
            version: v1alpha1
        description: |-
          Kepler (Kubernetes-based Efficient Power Level Exporter) is a Prometheus exporter that
          measures energy consumption metrics at the container, pod, and node level in Kubernetes clusters.

          ### Deprecation Notice
          **Important:** The Kepler CR (Custom Resource) is deprecated and will be removed in a future release. Please use the new PowerMonitor instead.

          ### Prerequisites
          To view the metrics exposed by Kepler, **[User Workload Monitoring](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/monitoring/configuring-user-workload-monitoring)**
          MUST be enabled in your cluster.

          ### Documentation
          * [User Documentation](https://github.com/sustainable-computing-io/kepler-operator/blob/v1alpha1/README.md)
          * [Developer Guide](https://github.com/sustainable-computing-io/kepler-operator/tree/v1alpha1/docs/developer)

          ### Contributing
          * [Report Bugs](https://github.com/sustainable-computing-io/kepler-operator/issues/new?template=bug_report.md) - Found a bug? Let us know!
          * [Browse Issues](https://github.com/sustainable-computing-io/kepler-operator/issues) - See what we're working on

          ### License
          Licensed under [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)
        displayName: Kepler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kepler
        - power
        - energy
        - monitoring
        links:
        - name: Installation Guide
          url: https://sustainable-computing.io/installation/community-operator/
        - name: Kepler Project
          url: https://sustainable-computing.io/
        maintainers:
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: vprashar@redhat.com
          name: Vibhu Prashar
        - email: vimalkum@redhat.com
          name: Vimal Kumar
        - email: kaliu@redhat.com
          name: Kaiyi Liu
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Kepler Operator Contributors
          url: https://sustainable-computing.io/
        relatedImages:
        - quay.io/sustainable_computing_io/kepler-operator:0.20.0
        - quay.io/sustainable_computing_io/kepler:v0.10.2
        - quay.io/brancz/kube-rbac-proxy:v0.19.0
        version: 0.20.0
      entries:
      - name: kepler-operator.v0.20.0
        version: 0.20.0
      - name: kepler-operator.v0.19.0
        version: 0.19.0
      - name: kepler-operator.v0.18.0
        version: 0.18.0
      - name: kepler-operator.v0.17.0
        version: 0.17.0
      - name: kepler-operator.v0.16.0
        version: 0.16.0
      - name: kepler-operator.v0.15.0
        version: 0.15.0
      name: alpha
    defaultChannel: alpha
    packageName: kepler-operator
    provider:
      name: Kepler Operator Contributors
      url: https://sustainable-computing.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MuleSoft
      provider-url: anypoint.mulesoft.com
    name: runtime-fabric-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runtime-fabric-operator.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rtf.operators.mulesoft.com/v1",
                "kind": "RuntimeFabric",
                "metadata": {
                  "name": "runtime-fabric"
                },
                "spec": {
                  "activationData": null,
                  "agent": {
                    "replicas": 1,
                    "app": {
                      "resources": {
                        "limits": {
                          "cpu": "1000m",
                          "memory": "500Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      }
                    },
                    "rtfd": {
                      "resources": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "150m",
                          "memory": "200Mi"
                        }
                      },
                      "watch": {
                        "enabled": true,
                        "components": {
                          "deployments": {
                            "enabled": true
                          },
                          "ingress": {
                            "enabled": true
                          },
                          "secrets": {
                            "enabled": true
                          },
                          "persistencegateways": {
                            "enabled": true
                          },
                          "nodes": {
                            "enabled": true
                          },
                          "jobs": {
                            "enabled": true
                          },
                          "routes": {
                            "enabled": true
                          }
                        }
                      }
                    }
                  },
                  "resource-cache": {
                    "replicas": 1,
                    "nginx": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "250Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    },
                    "fetcher": {
                      "resources": {
                        "limits": {
                          "cpu": "50m",
                          "memory": "100Mi"
                        },
                        "requests": {
                          "cpu": "50m",
                          "memory": "50Mi"
                        }
                      }
                    }
                  },
                  "mule-clusterip-service": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "global": {
                    "cluster": {
                      "flavor": "byok",
                      "type": "openshift"
                    },
                    "containerLogPaths": [
                      "/var/lib/docker/containers",
                      "/var/log/containers",
                      "/var/log/pods"
                    ],
                    "core": {
                      "affinity": {},
                      "tolerations": [],
                      "topologySpreadConstraints": [
                        {
                          "maxSkew": 1,
                          "topologyKey": "topology.kubernetes.io/zone",
                          "whenUnsatisfiable": "ScheduleAnyway"
                        }
                      ]
                    },
                    "worker": {
                      "affinity": {},
                      "tolerations": []
                    },
                    "deployment": {
                      "nodeSelector": ""
                    },
                    "image": {
                      "provider": "aws",
                      "pullSecretName": "rtf-pull-secret",
                      "rtfRegistry": "rtf-runtime-registry.kprod.msap.io",
                      "repositoryNamespace": "mulesoft"
                    },
                    "service": {
                      "clusterIp": null
                    },
                    "authorizedNamespaces": false,
                    "fipsEnabled": false,
                    "crds": {
                      "install": true,
                      "keep": true,
                      "flex": {
                        "enabled": false
                      }
                    },
                    "nodeWatcherEnabled": true,
                    "nodeReconciliation": true,
                    "deploymentRateLimitPerSecond": 1
                  },
                  "muleLicense": null,
                  "proxy": {
                    "http_no_proxy": null,
                    "http_proxy": null,
                    "monitoring_proxy": null
                  },
                  "customLog4jEnabled": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
          createdAt: "2025-07-31T21:12:41Z"
          description: Runtime Fabric Operator Application from Mulesoft
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mulesoft/rtf-ocp-operator
          support: MuleSoft
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Runtime Fabric Operator from Mulesoft
              **Activation data**
              Used to install Runtime Fabric on a Kubernetes service. Copy this value from the Runtime Fabric setup page in Runtime Manager.
              **Registry URL**
              The registry from which to retrieve the images needed to install and run Runtime Fabric. If you're using a local registry, specify the URL here.
                * Default registry URLs:
                    - US control plane: rtf-runtime-registry.kprod.msap.io
                    - EU control plane: rtf-runtime-registry.kprod-eu.msap.io
              **Registry pull secret**
              Used to retrieve images from the registry. Change this value only if you did so during the operator installation.
              **Mule license key**
              Your company’s Mule license key. The key is not required during install, but it is required to deploy applications to Runtime Fabric. Must be Base64 encoded.
            displayName: Runtime Fabric
            kind: RuntimeFabric
            name: runtimefabrics.rtf.operators.mulesoft.com
            version: v1
        description: |-
          Anypoint Runtime Fabric is a container service that you install in your existing Red Hat OpenShift cluster that enables you to run Mule applications, composite APIs, and API gateways across any environment with centralized management via Anypoint Platform.
          Capabilities of Runtime Fabric include:
          * Isolation between applications by running a separate Mule runtime server per application.
          * Ability to run multiple versions of Mule runtime server on the same set of resources.
          * Scaling applications across multiple replicas.
          * Automated application fail-over.
          * Application management with Anypoint Runtime Manager.
          Runtime Fabric supports the following Red Hat OpenShift deployment options:
          * Red Hat OpenShift Service on AWS
          * Microsoft Azure Red Hat OpenShift
          * Red Hat OpenShift Dedicated
          * Red Hat OpenShift on IBM Cloud
          * Self-managed Red Hat OpenShift editions (Performance Plus, OCP, Kubernetes engine)
          Runtime Fabric supports Red Hat OpenShift versions 4.16 to 4.19.
        displayName: Runtime Fabric Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - agent
        - rtf
        - runtime
        - fabric
        - mulesoft
        links:
        - name: Runtime Fabric Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/62901512fefb1b137ba23946
        - name: Documentation
          url: https://docs.mulesoft.com/release-notes/runtime-fabric/runtime-fabric
        maintainers:
        - email: support@mulesoft.com
          name: Mulesoft
        maturity: stable
        provider:
          name: MuleSoft
          url: anypoint.mulesoft.com
        relatedImages:
        - quay.io/rtf/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
        version: 2.11.0
      entries:
      - name: runtime-fabric-operator.v2.11.0
        version: 2.11.0
      - name: runtime-fabric-operator.v2.10.34
        version: 2.10.34
      - name: runtime-fabric-operator.v2.10.17
        version: 2.10.17
      - name: runtime-fabric-operator.v2.10.0
        version: 2.10.0
      - name: runtime-fabric-operator.v2.9.2
        version: 2.9.2
      - name: runtime-fabric-operator.v2.8.2
        version: 2.8.2
      - name: runtime-fabric-operator.v2.7.0
        version: 2.7.0
      - name: runtime-fabric-operator.v2.6.52
        version: 2.6.52
      name: alpha
    - currentCSV: runtime-fabric-operator.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rtf.operators.mulesoft.com/v1",
                "kind": "RuntimeFabric",
                "metadata": {
                  "name": "runtime-fabric"
                },
                "spec": {
                  "activationData": null,
                  "agent": {
                    "replicas": 1,
                    "app": {
                      "resources": {
                        "limits": {
                          "cpu": "1000m",
                          "memory": "500Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      }
                    },
                    "rtfd": {
                      "resources": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "150m",
                          "memory": "200Mi"
                        }
                      },
                      "watch": {
                        "enabled": true,
                        "components": {
                          "deployments": {
                            "enabled": true
                          },
                          "ingress": {
                            "enabled": true
                          },
                          "secrets": {
                            "enabled": true
                          },
                          "persistencegateways": {
                            "enabled": true
                          },
                          "nodes": {
                            "enabled": true
                          },
                          "jobs": {
                            "enabled": true
                          },
                          "routes": {
                            "enabled": true
                          }
                        }
                      }
                    }
                  },
                  "resource-cache": {
                    "replicas": 1,
                    "nginx": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "250Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    },
                    "fetcher": {
                      "resources": {
                        "limits": {
                          "cpu": "50m",
                          "memory": "100Mi"
                        },
                        "requests": {
                          "cpu": "50m",
                          "memory": "50Mi"
                        }
                      }
                    }
                  },
                  "mule-clusterip-service": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "global": {
                    "cluster": {
                      "flavor": "byok",
                      "type": "openshift"
                    },
                    "containerLogPaths": [
                      "/var/lib/docker/containers",
                      "/var/log/containers",
                      "/var/log/pods"
                    ],
                    "core": {
                      "affinity": {},
                      "tolerations": [],
                      "topologySpreadConstraints": [
                        {
                          "maxSkew": 1,
                          "topologyKey": "topology.kubernetes.io/zone",
                          "whenUnsatisfiable": "ScheduleAnyway"
                        }
                      ]
                    },
                    "worker": {
                      "affinity": {},
                      "tolerations": []
                    },
                    "deployment": {
                      "nodeSelector": ""
                    },
                    "image": {
                      "provider": "aws",
                      "pullSecretName": "rtf-pull-secret",
                      "rtfRegistry": "rtf-runtime-registry.kprod.msap.io",
                      "repositoryNamespace": "mulesoft"
                    },
                    "service": {
                      "clusterIp": null
                    },
                    "authorizedNamespaces": false,
                    "fipsEnabled": false,
                    "crds": {
                      "install": true,
                      "keep": true,
                      "flex": {
                        "enabled": false
                      }
                    },
                    "nodeWatcherEnabled": true,
                    "nodeReconciliation": true,
                    "deploymentRateLimitPerSecond": 1
                  },
                  "muleLicense": null,
                  "proxy": {
                    "http_no_proxy": null,
                    "http_proxy": null,
                    "monitoring_proxy": null
                  },
                  "customLog4jEnabled": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
          createdAt: "2025-07-31T21:12:41Z"
          description: Runtime Fabric Operator Application from Mulesoft
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mulesoft/rtf-ocp-operator
          support: MuleSoft
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Runtime Fabric Operator from Mulesoft
              **Activation data**
              Used to install Runtime Fabric on a Kubernetes service. Copy this value from the Runtime Fabric setup page in Runtime Manager.
              **Registry URL**
              The registry from which to retrieve the images needed to install and run Runtime Fabric. If you're using a local registry, specify the URL here.
                * Default registry URLs:
                    - US control plane: rtf-runtime-registry.kprod.msap.io
                    - EU control plane: rtf-runtime-registry.kprod-eu.msap.io
              **Registry pull secret**
              Used to retrieve images from the registry. Change this value only if you did so during the operator installation.
              **Mule license key**
              Your company’s Mule license key. The key is not required during install, but it is required to deploy applications to Runtime Fabric. Must be Base64 encoded.
            displayName: Runtime Fabric
            kind: RuntimeFabric
            name: runtimefabrics.rtf.operators.mulesoft.com
            version: v1
        description: |-
          Anypoint Runtime Fabric is a container service that you install in your existing Red Hat OpenShift cluster that enables you to run Mule applications, composite APIs, and API gateways across any environment with centralized management via Anypoint Platform.
          Capabilities of Runtime Fabric include:
          * Isolation between applications by running a separate Mule runtime server per application.
          * Ability to run multiple versions of Mule runtime server on the same set of resources.
          * Scaling applications across multiple replicas.
          * Automated application fail-over.
          * Application management with Anypoint Runtime Manager.
          Runtime Fabric supports the following Red Hat OpenShift deployment options:
          * Red Hat OpenShift Service on AWS
          * Microsoft Azure Red Hat OpenShift
          * Red Hat OpenShift Dedicated
          * Red Hat OpenShift on IBM Cloud
          * Self-managed Red Hat OpenShift editions (Performance Plus, OCP, Kubernetes engine)
          Runtime Fabric supports Red Hat OpenShift versions 4.16 to 4.19.
        displayName: Runtime Fabric Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - agent
        - rtf
        - runtime
        - fabric
        - mulesoft
        links:
        - name: Runtime Fabric Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/62901512fefb1b137ba23946
        - name: Documentation
          url: https://docs.mulesoft.com/release-notes/runtime-fabric/runtime-fabric
        maintainers:
        - email: support@mulesoft.com
          name: Mulesoft
        maturity: stable
        provider:
          name: MuleSoft
          url: anypoint.mulesoft.com
        relatedImages:
        - quay.io/rtf/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
        version: 2.11.0
      entries:
      - name: runtime-fabric-operator.v2.11.0
        version: 2.11.0
      - name: runtime-fabric-operator.v2.10.34
        version: 2.10.34
      - name: runtime-fabric-operator.v2.10.17
        version: 2.10.17
      - name: runtime-fabric-operator.v2.10.0
        version: 2.10.0
      - name: runtime-fabric-operator.v2.9.2
        version: 2.9.2
      - name: runtime-fabric-operator.v2.8.2
        version: 2.8.2
      - name: runtime-fabric-operator.v2.7.0
        version: 2.7.0
      - name: runtime-fabric-operator.v2.6.52
        version: 2.6.52
      name: stable
    defaultChannel: stable
    packageName: runtime-fabric-operator
    provider:
      name: MuleSoft
      url: anypoint.mulesoft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-codeartifact-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-codeartifact-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/codeartifact-controller:1.1.1
          createdAt: "2025-09-19T22:00:24Z"
          description: AWS codeartifact controller is a service controller for managing
            codeartifact resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Domain represents the state of an AWS codeartifact Domain
              resource.
            displayName: Domain
            kind: Domain
            name: domains.codeartifact.services.k8s.aws
            version: v1alpha1
          - description: PackageGroup represents the state of an AWS codeartifact
              PackageGroup resource.
            displayName: PackageGroup
            kind: PackageGroup
            name: packagegroups.codeartifact.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon codeartifact resources in AWS from within your Kubernetes cluster.

          **About Amazon codeartifact**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon codeartifact
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - codeartifact
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon codeartifact Developer Resources
          url: https://aws.amazon.com/codeartifact/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: codeartifact maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/codeartifact-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-codeartifact-controller.v1.1.1
        version: 1.1.1
      - name: ack-codeartifact-controller.v1.1.0
        version: 1.1.0
      - name: ack-codeartifact-controller.v1.0.13
        version: 1.0.13
      - name: ack-codeartifact-controller.v1.0.11
        version: 1.0.11
      name: alpha
    defaultChannel: alpha
    packageName: ack-codeartifact-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-elasticache-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-elasticache-controller.v1.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "CacheParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "CacheSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "ReplicationGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "Snapshot",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "UserGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/elasticache-controller:1.2.3
          createdAt: "2025-09-19T19:19:33Z"
          description: AWS ElastiCache controller is a service controller for managing
            ElastiCache resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CacheCluster represents the state of an AWS elasticache CacheCluster
              resource.
            displayName: CacheCluster
            kind: CacheCluster
            name: cacheclusters.elasticache.services.k8s.aws
            version: v1alpha1
          - description: CacheParameterGroup represents the state of an AWS elasticache
              CacheParameterGroup resource.
            displayName: CacheParameterGroup
            kind: CacheParameterGroup
            name: cacheparametergroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: CacheSubnetGroup represents the state of an AWS elasticache
              CacheSubnetGroup resource.
            displayName: CacheSubnetGroup
            kind: CacheSubnetGroup
            name: cachesubnetgroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ReplicationGroup represents the state of an AWS elasticache
              ReplicationGroup resource.
            displayName: ReplicationGroup
            kind: ReplicationGroup
            name: replicationgroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCache represents the state of an AWS elasticache
              ServerlessCache resource.
            displayName: ServerlessCache
            kind: ServerlessCache
            name: serverlesscaches.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCacheSnapshot represents the state of an AWS elasticache
              ServerlessCacheSnapshot resource.
            displayName: ServerlessCacheSnapshot
            kind: ServerlessCacheSnapshot
            name: serverlesscachesnapshots.elasticache.services.k8s.aws
            version: v1alpha1
          - description: Snapshot represents the state of an AWS elasticache Snapshot
              resource.
            displayName: Snapshot
            kind: Snapshot
            name: snapshots.elasticache.services.k8s.aws
            version: v1alpha1
          - description: UserGroup represents the state of an AWS elasticache UserGroup
              resource.
            displayName: UserGroup
            kind: UserGroup
            name: usergroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS elasticache User resource.
            displayName: User
            kind: User
            name: users.elasticache.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ElastiCache resources in AWS from within your Kubernetes cluster.

          **About Amazon ElastiCache**

          Amazon ElastiCache allows you to seamlessly set up, run, and scale popular open-source compatible in-memory data stores in the cloud. Build data-intensive apps or boost the performance of your existing databases by retrieving data from high throughput and low latency in-memory data stores. Amazon ElastiCache is a popular choice for real-time use cases like Caching, Session Stores, Gaming, Geospatial Services, Real-Time Analytics, and Queuing.

          Amazon ElastiCache offers fully managed Redis, voted the most loved database by developers in the Stack Overflow 2020 Developer Survey, and Memcached for your most demanding applications that require sub-millisecond response times.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ElastiCache
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticache
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ElastiCache Developer Resources
          url: https://aws.amazon.com/elasticache/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: elasticache maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/elasticache-controller:1.2.3
        version: 1.2.3
      entries:
      - name: ack-elasticache-controller.v1.2.3
        version: 1.2.3
      - name: ack-elasticache-controller.v1.2.2
        version: 1.2.2
      - name: ack-elasticache-controller.v1.2.1
        version: 1.2.1
      - name: ack-elasticache-controller.v1.2.0
        version: 1.2.0
      - name: ack-elasticache-controller.v1.1.0
        version: 1.1.0
      - name: ack-elasticache-controller.v1.0.7
        version: 1.0.7
      - name: ack-elasticache-controller.v1.0.6
        version: 1.0.6
      - name: ack-elasticache-controller.v1.0.5
        version: 1.0.5
      - name: ack-elasticache-controller.v1.0.4
        version: 1.0.4
      - name: ack-elasticache-controller.v1.0.3
        version: 1.0.3
      - name: ack-elasticache-controller.v1.0.2
        version: 1.0.2
      - name: ack-elasticache-controller.v1.0.1
        version: 1.0.1
      - name: ack-elasticache-controller.v1.0.0
        version: 1.0.0
      - name: ack-elasticache-controller.v0.2.6
        version: 0.2.6
      - name: ack-elasticache-controller.v0.2.5
        version: 0.2.5
      - name: ack-elasticache-controller.v0.2.4
        version: 0.2.4
      - name: ack-elasticache-controller.v0.2.3
        version: 0.2.3
      - name: ack-elasticache-controller.v0.2.2
        version: 0.2.2
      - name: ack-elasticache-controller.v0.2.1
        version: 0.2.1
      - name: ack-elasticache-controller.v0.2.0
        version: 0.2.0
      - name: ack-elasticache-controller.v0.1.1
        version: 0.1.1
      - name: ack-elasticache-controller.v0.1.0
        version: 0.1.0
      - name: ack-elasticache-controller.v0.0.31
        version: 0.0.31
      - name: ack-elasticache-controller.v0.0.30
        version: 0.0.30
      - name: ack-elasticache-controller.v0.0.29
        version: 0.0.29
      - name: ack-elasticache-controller.v0.0.28
        version: 0.0.28
      - name: ack-elasticache-controller.v0.0.27
        version: 0.0.27
      - name: ack-elasticache-controller.v0.0.26
        version: 0.0.26
      - name: ack-elasticache-controller.v0.0.25
        version: 0.0.25
      - name: ack-elasticache-controller.v0.0.24
        version: 0.0.24
      - name: ack-elasticache-controller.v0.0.23
        version: 0.0.23
      - name: ack-elasticache-controller.v0.0.22
        version: 0.0.22
      - name: ack-elasticache-controller.v0.0.21
        version: 0.0.21
      - name: ack-elasticache-controller.v0.0.20
        version: 0.0.20
      - name: ack-elasticache-controller.v0.0.19
        version: 0.0.19
      - name: ack-elasticache-controller.v0.0.18
        version: 0.0.18
      - name: ack-elasticache-controller.v0.0.17
        version: 0.0.17
      - name: ack-elasticache-controller.v0.0.16
        version: 0.0.16
      - name: ack-elasticache-controller.v0.0.15
        version: 0.0.15
      - name: ack-elasticache-controller.v0.0.14
        version: 0.0.14
      - name: ack-elasticache-controller.v0.0.13
        version: 0.0.13
      - name: ack-elasticache-controller.v0.0.12
        version: 0.0.12
      - name: ack-elasticache-controller.v0.0.11
        version: 0.0.11
      - name: ack-elasticache-controller.v0.0.10
        version: 0.0.10
      - name: ack-elasticache-controller.v0.0.9
        version: 0.0.9
      - name: ack-elasticache-controller.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: ack-elasticache-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ember-csi-community-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ember-csi-community-operator.v0.9.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ember-csi.io/v1alpha1",
                "kind": "EmberStorageBackend",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "config": {
                    "envVars": {
                      "X_CSI_PERSISTENCE_CONFIG": {
                        "storage": "crd"
                      },
                      "X_CSI_DEBUG_MODE": "",
                      "X_CSI_DEFAULT_MOUNT_FS": "ext4",
                      "X_CSI_EMBER_CONFIG": {
                        "grpc_workers": 30,
                        "disable_logs": false,
                        "plugin_name": "",
                        "slow_operations": true,
                        "disabled": [],
                        "enable_probe": false,
                        "debug": false,
                        "user_id": "ember_csi.io",
                        "project_id": "ember_csi.io"
                      },
                      "X_CSI_BACKEND_CONFIG": {
                        "enable_unsupported_driver": false,
                        "driver": "ACCESSIscsi",
                        "name": "",
                        "driver__ACCESSIscsi__vrts_lun_sparse": true,
                        "driver__ACCESSIscsi__vrts_target_config": "/etc/cinder/vrts_target.xml",
                        "driver__AS13000__as13000_ipsan_pools__transform_csv": "Pool0",
                        "driver__AS13000__as13000_meta_pool": "",
                        "driver__AS13000__as13000_token_available_time": 3300,
                        "driver__Acs5000ISCSI__acs5000_copy_interval": 5,
                        "driver__Acs5000ISCSI__acs5000_target": 0,
                        "driver__Acs5000ISCSI__acs5000_volpool_name__transform_csv": "pool01",
                        "driver__DPLFC__dpl_pool": "",
                        "driver__DPLFC__dpl_port": 8357,
                        "driver__DPLISCSI__dpl_pool": "",
                        "driver__DPLISCSI__dpl_port": 8357,
                        "driver__DSWARE__dsware_rest_url": "",
                        "driver__DSWARE__dsware_storage_pools": "",
                        "driver__DSWARE__manager_ips__transform_csv_kvs": "",
                        "driver__Datera__backend_availability_zone": "",
                        "driver__Datera__chap_password": "",
                        "driver__Datera__chap_username": "",
                        "driver__Datera__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__Datera__driver_client_cert": "",
                        "driver__Datera__driver_client_cert_key": "",
                        "driver__Datera__driver_data_namespace": "",
                        "driver__Datera__driver_ssl_cert_path": "",
                        "driver__Datera__driver_ssl_cert_verify": false,
                        "driver__Datera__driver_use_ssl": false,
                        "driver__Datera__filter_function": "",
                        "driver__Datera__goodness_function": "",
                        "driver__Datera__iscsi_iotype": "fileio",
                        "driver__Datera__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__Datera__iscsi_target_flags": "",
                        "driver__Datera__iscsi_write_cache": "on",
                        "driver__Datera__num_shell_tries": 3,
                        "driver__Datera__num_volume_device_scan_tries": 3,
                        "driver__Datera__report_discard_supported": false,
                        "driver__Datera__storage_protocol": "iscsi",
                        "driver__Datera__target_helper": "tgtadm",
                        "driver__Datera__target_ip_address": "$my_ip",
                        "driver__Datera__target_port": 3260,
                        "driver__Datera__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__Datera__target_protocol": "iscsi",
                        "driver__Datera__trace_flags__transform_csv": "",
                        "driver__Datera__use_chap_auth": false,
                        "driver__Datera__volume_backend_name": "",
                        "driver__Datera__volume_clear": "zero",
                        "driver__Datera__volume_clear_ionice": "",
                        "driver__Datera__volume_clear_size": 0,
                        "driver__Datera__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__Datera__volume_copy_bps_limit": 0,
                        "driver__Datera__volume_dd_blocksize": "1M",
                        "driver__Datera__volumes_dir": "$state_path/volumes",
                        "driver__FJDXFC__cinder_eternus_config_file": "/etc/cinder/cinder_fujitsu_eternus_dx.xml",
                        "driver__FJDXISCSI__cinder_eternus_config_file": "/etc/cinder/cinder_fujitsu_eternus_dx.xml",
                        "driver__FlashSystemFC__flashsystem_connection_protocol": "FC",
                        "driver__FlashSystemFC__flashsystem_multihostmap_enabled": true,
                        "driver__FlashSystemISCSI__flashsystem_connection_protocol": "FC",
                        "driver__FlashSystemISCSI__flashsystem_multihostmap_enabled": true,
                        "driver__GPFS__gpfs_images_dir": "",
                        "driver__GPFS__gpfs_images_share_mode__transform_empty_none": "",
                        "driver__GPFS__gpfs_max_clone_depth": 0,
                        "driver__GPFS__gpfs_mount_point_base": "",
                        "driver__GPFS__gpfs_sparse_volumes": true,
                        "driver__GPFS__gpfs_storage_pool": "system",
                        "driver__GPFSRemote__gpfs_hosts__transform_csv": "",
                        "driver__GPFSRemote__gpfs_hosts_key_file": "$state_path/ssh_known_hosts",
                        "driver__GPFSRemote__gpfs_images_dir": "",
                        "driver__GPFSRemote__gpfs_images_share_mode__transform_empty_none": "",
                        "driver__GPFSRemote__gpfs_max_clone_depth": 0,
                        "driver__GPFSRemote__gpfs_mount_point_base": "",
                        "driver__GPFSRemote__gpfs_private_key": "",
                        "driver__GPFSRemote__gpfs_sparse_volumes": true,
                        "driver__GPFSRemote__gpfs_ssh_port": 22,
                        "driver__GPFSRemote__gpfs_storage_pool": "system",
                        "driver__GPFSRemote__gpfs_strict_host_key_policy": false,
                        "driver__GPFSRemote__gpfs_user_login": "root",
                        "driver__GPFSRemote__gpfs_user_password": "",
                        "driver__HBSDFC__backend_availability_zone": "",
                        "driver__HBSDFC__chap_password": "",
                        "driver__HBSDFC__chap_username": "",
                        "driver__HBSDFC__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__HBSDFC__driver_client_cert": "",
                        "driver__HBSDFC__driver_client_cert_key": "",
                        "driver__HBSDFC__driver_data_namespace": "",
                        "driver__HBSDFC__driver_ssl_cert_path": "",
                        "driver__HBSDFC__driver_ssl_cert_verify": false,
                        "driver__HBSDFC__driver_use_ssl": false,
                        "driver__HBSDFC__filter_function": "",
                        "driver__HBSDFC__goodness_function": "",
                        "driver__HBSDFC__iscsi_iotype": "fileio",
                        "driver__HBSDFC__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__HBSDFC__iscsi_target_flags": "",
                        "driver__HBSDFC__iscsi_write_cache": "on",
                        "driver__HBSDFC__num_shell_tries": 3,
                        "driver__HBSDFC__num_volume_device_scan_tries": 3,
                        "driver__HBSDFC__report_discard_supported": false,
                        "driver__HBSDFC__storage_protocol": "iscsi",
                        "driver__HBSDFC__target_helper": "tgtadm",
                        "driver__HBSDFC__target_ip_address": "$my_ip",
                        "driver__HBSDFC__target_port": 3260,
                        "driver__HBSDFC__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__HBSDFC__target_protocol": "iscsi",
                        "driver__HBSDFC__trace_flags__transform_csv": "",
                        "driver__HBSDFC__use_chap_auth": false,
                        "driver__HBSDFC__volume_backend_name": "",
                        "driver__HBSDFC__volume_clear": "zero",
                        "driver__HBSDFC__volume_clear_ionice": "",
                        "driver__HBSDFC__volume_clear_size": 0,
                        "driver__HBSDFC__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__HBSDFC__volume_copy_bps_limit": 0,
                        "driver__HBSDFC__volume_dd_blocksize": "1M",
                        "driver__HBSDFC__volumes_dir": "$state_path/volumes",
                        "driver__HBSDISCSI__backend_availability_zone": "",
                        "driver__HBSDISCSI__chap_password": "",
                        "driver__HBSDISCSI__chap_username": "",
                        "driver__HBSDISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__HBSDISCSI__driver_client_cert": "",
                        "driver__HBSDISCSI__driver_client_cert_key": "",
                        "driver__HBSDISCSI__driver_data_namespace": "",
                        "driver__HBSDISCSI__driver_ssl_cert_path": "",
                        "driver__HBSDISCSI__driver_ssl_cert_verify": false,
                        "driver__HBSDISCSI__driver_use_ssl": false,
                        "driver__HBSDISCSI__filter_function": "",
                        "driver__HBSDISCSI__goodness_function": "",
                        "driver__HBSDISCSI__iscsi_iotype": "fileio",
                        "driver__HBSDISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__HBSDISCSI__iscsi_target_flags": "",
                        "driver__HBSDISCSI__iscsi_write_cache": "on",
                        "driver__HBSDISCSI__num_shell_tries": 3,
                        "driver__HBSDISCSI__num_volume_device_scan_tries": 3,
                        "driver__HBSDISCSI__report_discard_supported": false,
                        "driver__HBSDISCSI__storage_protocol": "iscsi",
                        "driver__HBSDISCSI__target_helper": "tgtadm",
                        "driver__HBSDISCSI__target_ip_address": "$my_ip",
                        "driver__HBSDISCSI__target_port": 3260,
                        "driver__HBSDISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__HBSDISCSI__target_protocol": "iscsi",
                        "driver__HBSDISCSI__trace_flags__transform_csv": "",
                        "driver__HBSDISCSI__use_chap_auth": false,
                        "driver__HBSDISCSI__volume_backend_name": "",
                        "driver__HBSDISCSI__volume_clear": "zero",
                        "driver__HBSDISCSI__volume_clear_ionice": "",
                        "driver__HBSDISCSI__volume_clear_size": 0,
                        "driver__HBSDISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__HBSDISCSI__volume_copy_bps_limit": 0,
                        "driver__HBSDISCSI__volume_dd_blocksize": "1M",
                        "driver__HBSDISCSI__volumes_dir": "$state_path/volumes",
                        "driver__HPE3PARFC__hpe3par_api_url": "",
                        "driver__HPE3PARFC__hpe3par_cpg__transform_csv": "OpenStack",
                        "driver__HPE3PARFC__hpe3par_cpg_snap": "",
                        "driver__HPE3PARFC__hpe3par_debug": false,
                        "driver__HPE3PARFC__hpe3par_iscsi_chap_enabled": false,
                        "driver__HPE3PARFC__hpe3par_iscsi_ips__transform_csv": "",
                        "driver__HPE3PARFC__hpe3par_password": "",
                        "driver__HPE3PARFC__hpe3par_snapshot_expiration": "",
                        "driver__HPE3PARFC__hpe3par_snapshot_retention": "",
                        "driver__HPE3PARFC__hpe3par_target_nsp": "",
                        "driver__HPE3PARFC__hpe3par_username": "",
                        "driver__HPE3PARFC__san_ip": "",
                        "driver__HPE3PARFC__san_login": "admin",
                        "driver__HPE3PARFC__san_password": "",
                        "driver__HPE3PARFC__san_private_key": "",
                        "driver__HPE3PARFC__san_ssh_port": 22,
                        "driver__HPE3PARFC__ssh_conn_timeout": 30,
                        "driver__HPE3PARFC__target_ip_address": "$my_ip",
                        "driver__HPE3PARFC__target_port": 3260,
                        "driver__HPE3PARFC__unique_fqdn_network": true,
                        "driver__HPE3PARISCSI__hpe3par_api_url": "",
                        "driver__HPE3PARISCSI__hpe3par_cpg__transform_csv": "OpenStack",
                        "driver__HPE3PARISCSI__hpe3par_cpg_snap": "",
                        "driver__HPE3PARISCSI__hpe3par_debug": false,
                        "driver__HPE3PARISCSI__hpe3par_iscsi_chap_enabled": false,
                        "driver__HPE3PARISCSI__hpe3par_iscsi_ips__transform_csv": "",
                        "driver__HPE3PARISCSI__hpe3par_password": "",
                        "driver__HPE3PARISCSI__hpe3par_snapshot_expiration": "",
                        "driver__HPE3PARISCSI__hpe3par_snapshot_retention": "",
                        "driver__HPE3PARISCSI__hpe3par_target_nsp": "",
                        "driver__HPE3PARISCSI__hpe3par_username": "",
                        "driver__HPE3PARISCSI__san_ip": "",
                        "driver__HPE3PARISCSI__san_login": "admin",
                        "driver__HPE3PARISCSI__san_password": "",
                        "driver__HPE3PARISCSI__san_private_key": "",
                        "driver__HPE3PARISCSI__san_ssh_port": 22,
                        "driver__HPE3PARISCSI__ssh_conn_timeout": 30,
                        "driver__HPE3PARISCSI__target_ip_address": "$my_ip",
                        "driver__HPE3PARISCSI__target_port": 3260,
                        "driver__HPE3PARISCSI__unique_fqdn_network": true,
                        "driver__HPMSAFC__driver_ssl_cert_path": "",
                        "driver__HPMSAFC__driver_ssl_cert_verify": false,
                        "driver__HPMSAFC__driver_use_ssl": false,
                        "driver__HPMSAFC__hpmsa_pool_name": "A",
                        "driver__HPMSAFC__hpmsa_pool_type": "virtual",
                        "driver__HPMSAFC__san_ip": "",
                        "driver__HPMSAFC__san_login": "admin",
                        "driver__HPMSAFC__san_password": "",
                        "driver__HPMSAISCSI__driver_ssl_cert_path": "",
                        "driver__HPMSAISCSI__driver_ssl_cert_verify": false,
                        "driver__HPMSAISCSI__driver_use_ssl": false,
                        "driver__HPMSAISCSI__hpmsa_iscsi_ips__transform_csv": "",
                        "driver__HPMSAISCSI__hpmsa_pool_name": "A",
                        "driver__HPMSAISCSI__hpmsa_pool_type": "virtual",
                        "driver__HPMSAISCSI__san_ip": "",
                        "driver__HPMSAISCSI__san_login": "admin",
                        "driver__HPMSAISCSI__san_password": "",
                        "driver__HuaweiFC__cinder_huawei_conf_file": "/etc/cinder/cinder_huawei_conf.xml",
                        "driver__HuaweiFC__hypermetro_devices": "",
                        "driver__HuaweiFC__metro_domain_name": "",
                        "driver__HuaweiFC__metro_san_address": "",
                        "driver__HuaweiFC__metro_san_password": "",
                        "driver__HuaweiFC__metro_san_user": "",
                        "driver__HuaweiFC__metro_storage_pools": "",
                        "driver__HuaweiISCSI__cinder_huawei_conf_file": "/etc/cinder/cinder_huawei_conf.xml",
                        "driver__HuaweiISCSI__hypermetro_devices": "",
                        "driver__HuaweiISCSI__metro_domain_name": "",
                        "driver__HuaweiISCSI__metro_san_address": "",
                        "driver__HuaweiISCSI__metro_san_password": "",
                        "driver__HuaweiISCSI__metro_san_user": "",
                        "driver__HuaweiISCSI__metro_storage_pools": "",
                        "driver__IBMStorage__chap": "disabled",
                        "driver__IBMStorage__connection_type": "iscsi",
                        "driver__IBMStorage__management_ips": "",
                        "driver__IBMStorage__proxy": "cinder.volume.drivers.ibm.ibm_storage.proxy.IBMStorageProxy",
                        "driver__InStorageMCSFC__instorage_mcs_allow_tenant_qos": false,
                        "driver__InStorageMCSFC__instorage_mcs_localcopy_rate": 50,
                        "driver__InStorageMCSFC__instorage_mcs_localcopy_timeout": 120,
                        "driver__InStorageMCSFC__instorage_mcs_vol_autoexpand": true,
                        "driver__InStorageMCSFC__instorage_mcs_vol_compression": false,
                        "driver__InStorageMCSFC__instorage_mcs_vol_grainsize": 256,
                        "driver__InStorageMCSFC__instorage_mcs_vol_intier": true,
                        "driver__InStorageMCSFC__instorage_mcs_vol_iogrp": "0",
                        "driver__InStorageMCSFC__instorage_mcs_vol_rsize": 2,
                        "driver__InStorageMCSFC__instorage_mcs_vol_warning": 0,
                        "driver__InStorageMCSFC__instorage_mcs_volpool_name__transform_csv": "volpool",
                        "driver__InStorageMCSFC__instorage_san_secondary_ip": "",
                        "driver__InStorageMCSISCSI__instorage_mcs_allow_tenant_qos": false,
                        "driver__InStorageMCSISCSI__instorage_mcs_localcopy_rate": 50,
                        "driver__InStorageMCSISCSI__instorage_mcs_localcopy_timeout": 120,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_autoexpand": true,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_compression": false,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_grainsize": 256,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_intier": true,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_iogrp": "0",
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_rsize": 2,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_warning": 0,
                        "driver__InStorageMCSISCSI__instorage_mcs_volpool_name__transform_csv": "volpool",
                        "driver__InStorageMCSISCSI__instorage_san_secondary_ip": "",
                        "driver__InfiniboxVolume__chap_password": "",
                        "driver__InfiniboxVolume__chap_username": "",
                        "driver__InfiniboxVolume__infinidat_iscsi_netspaces__transform_csv": "",
                        "driver__InfiniboxVolume__infinidat_pool_name": "",
                        "driver__InfiniboxVolume__infinidat_storage_protocol": "fc",
                        "driver__InfiniboxVolume__infinidat_use_compression": false,
                        "driver__InfiniboxVolume__num_volume_device_scan_tries": 3,
                        "driver__InfiniboxVolume__san_ip": "",
                        "driver__InfiniboxVolume__san_login": "admin",
                        "driver__InfiniboxVolume__san_password": "",
                        "driver__InfiniboxVolume__san_thin_provision": true,
                        "driver__InfiniboxVolume__use_chap_auth": false,
                        "driver__InfiniboxVolume__volume_dd_blocksize": "1M",
                        "driver__InfortrendCLIFC__infortrend_cli_cache": false,
                        "driver__InfortrendCLIFC__infortrend_cli_max_retries": 5,
                        "driver__InfortrendCLIFC__infortrend_cli_path": "/opt/bin/Infortrend/raidcmd_ESDS10.jar",
                        "driver__InfortrendCLIFC__infortrend_cli_timeout": 60,
                        "driver__InfortrendCLIFC__infortrend_iqn_prefix": "iqn.2002-10.com.infortrend",
                        "driver__InfortrendCLIFC__infortrend_pools_name__transform_csv": "",
                        "driver__InfortrendCLIFC__infortrend_slots_a_channels_id__transform_csv": "",
                        "driver__InfortrendCLIFC__infortrend_slots_b_channels_id__transform_csv": "",
                        "driver__InfortrendCLIFC__java_path": "/usr/bin/java",
                        "driver__InfortrendCLIISCSI__infortrend_cli_cache": false,
                        "driver__InfortrendCLIISCSI__infortrend_cli_max_retries": 5,
                        "driver__InfortrendCLIISCSI__infortrend_cli_path": "/opt/bin/Infortrend/raidcmd_ESDS10.jar",
                        "driver__InfortrendCLIISCSI__infortrend_cli_timeout": 60,
                        "driver__InfortrendCLIISCSI__infortrend_iqn_prefix": "iqn.2002-10.com.infortrend",
                        "driver__InfortrendCLIISCSI__infortrend_pools_name__transform_csv": "",
                        "driver__InfortrendCLIISCSI__infortrend_slots_a_channels_id__transform_csv": "",
                        "driver__InfortrendCLIISCSI__infortrend_slots_b_channels_id__transform_csv": "",
                        "driver__InfortrendCLIISCSI__java_path": "/usr/bin/java",
                        "driver__JovianISCSI__backend_availability_zone": "",
                        "driver__JovianISCSI__chap_password": "",
                        "driver__JovianISCSI__chap_username": "",
                        "driver__JovianISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__JovianISCSI__driver_client_cert": "",
                        "driver__JovianISCSI__driver_client_cert_key": "",
                        "driver__JovianISCSI__driver_data_namespace": "",
                        "driver__JovianISCSI__driver_ssl_cert_path": "",
                        "driver__JovianISCSI__driver_ssl_cert_verify": false,
                        "driver__JovianISCSI__driver_use_ssl": false,
                        "driver__JovianISCSI__filter_function": "",
                        "driver__JovianISCSI__goodness_function": "",
                        "driver__JovianISCSI__iscsi_iotype": "fileio",
                        "driver__JovianISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__JovianISCSI__iscsi_target_flags": "",
                        "driver__JovianISCSI__iscsi_write_cache": "on",
                        "driver__JovianISCSI__num_shell_tries": 3,
                        "driver__JovianISCSI__num_volume_device_scan_tries": 3,
                        "driver__JovianISCSI__report_discard_supported": false,
                        "driver__JovianISCSI__storage_protocol": "iscsi",
                        "driver__JovianISCSI__target_helper": "tgtadm",
                        "driver__JovianISCSI__target_ip_address": "$my_ip",
                        "driver__JovianISCSI__target_port": 3260,
                        "driver__JovianISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__JovianISCSI__target_protocol": "iscsi",
                        "driver__JovianISCSI__trace_flags__transform_csv": "",
                        "driver__JovianISCSI__use_chap_auth": false,
                        "driver__JovianISCSI__volume_backend_name": "",
                        "driver__JovianISCSI__volume_clear": "zero",
                        "driver__JovianISCSI__volume_clear_ionice": "",
                        "driver__JovianISCSI__volume_clear_size": 0,
                        "driver__JovianISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__JovianISCSI__volume_copy_bps_limit": 0,
                        "driver__JovianISCSI__volume_dd_blocksize": "1M",
                        "driver__JovianISCSI__volumes_dir": "$state_path/volumes",
                        "driver__KaminarioISCSI__disable_discovery": false,
                        "driver__KaminarioISCSI__san_ip": "",
                        "driver__KaminarioISCSI__san_login": "admin",
                        "driver__KaminarioISCSI__san_password": "",
                        "driver__KaminarioISCSI__unique_fqdn_network": true,
                        "driver__KaminarioISCSI__volume_dd_blocksize": "1M",
                        "driver__LVMVolume__iscsi_iotype": "fileio",
                        "driver__LVMVolume__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__LVMVolume__iscsi_target_flags": "",
                        "driver__LVMVolume__iscsi_write_cache": "on",
                        "driver__LVMVolume__lvm_conf_file": "/etc/cinder/lvm.conf",
                        "driver__LVMVolume__lvm_mirrors": 0,
                        "driver__LVMVolume__lvm_suppress_fd_warnings": false,
                        "driver__LVMVolume__lvm_type": "auto",
                        "driver__LVMVolume__nvmet_port_id": 1,
                        "driver__LVMVolume__scst_target_driver": "iscsi",
                        "driver__LVMVolume__scst_target_iqn_name": "",
                        "driver__LVMVolume__spdk_max_queue_depth": 64,
                        "driver__LVMVolume__spdk_rpc_ip": "",
                        "driver__LVMVolume__spdk_rpc_password": "",
                        "driver__LVMVolume__spdk_rpc_port": 8000,
                        "driver__LVMVolume__spdk_rpc_username": "",
                        "driver__LVMVolume__target_helper": "tgtadm",
                        "driver__LVMVolume__target_ip_address": "$my_ip",
                        "driver__LVMVolume__target_port": 3260,
                        "driver__LVMVolume__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__LVMVolume__target_protocol": "iscsi",
                        "driver__LVMVolume__volume_clear": "zero",
                        "driver__LVMVolume__volume_clear_size": 0,
                        "driver__LVMVolume__volume_dd_blocksize": "1M",
                        "driver__LVMVolume__volume_group": "cinder-volumes",
                        "driver__LVMVolume__volumes_dir": "$state_path/volumes",
                        "driver__LenovoFC__lenovo_pool_name": "A",
                        "driver__LenovoFC__lenovo_pool_type": "virtual",
                        "driver__LenovoISCSI__lenovo_iscsi_ips__transform_csv": "",
                        "driver__LenovoISCSI__lenovo_pool_name": "A",
                        "driver__LenovoISCSI__lenovo_pool_type": "virtual",
                        "driver__LinstorDrbd__linstor_autoplace_count": 0,
                        "driver__LinstorDrbd__linstor_controller_diskless": true,
                        "driver__LinstorDrbd__linstor_default_blocksize": 4096,
                        "driver__LinstorDrbd__linstor_default_storage_pool_name": "DfltStorPool",
                        "driver__LinstorDrbd__linstor_default_uri": "linstor://localhost",
                        "driver__LinstorDrbd__linstor_default_volume_group_name": "drbd-vg",
                        "driver__LinstorDrbd__linstor_volume_downsize_factor__transform_string_float": "4096",
                        "driver__LinstorIscsi__linstor_autoplace_count": 0,
                        "driver__LinstorIscsi__linstor_controller_diskless": true,
                        "driver__LinstorIscsi__linstor_default_blocksize": 4096,
                        "driver__LinstorIscsi__linstor_default_storage_pool_name": "DfltStorPool",
                        "driver__LinstorIscsi__linstor_default_uri": "linstor://localhost",
                        "driver__LinstorIscsi__linstor_default_volume_group_name": "drbd-vg",
                        "driver__LinstorIscsi__linstor_volume_downsize_factor__transform_string_float": "4096",
                        "driver__MStorageFC__nec_actual_free_capacity": false,
                        "driver__MStorageFC__nec_auto_accesscontrol": true,
                        "driver__MStorageFC__nec_backend_max_ld_count": 1024,
                        "driver__MStorageFC__nec_backup_ldname_format": "LX:%s",
                        "driver__MStorageFC__nec_backup_pools__transform_csv": "",
                        "driver__MStorageFC__nec_cv_ldname_format": "LX:__ControlVolume_%xh",
                        "driver__MStorageFC__nec_diskarray_name": "",
                        "driver__MStorageFC__nec_ismcli_fip": "",
                        "driver__MStorageFC__nec_ismcli_password": "",
                        "driver__MStorageFC__nec_ismcli_privkey": "",
                        "driver__MStorageFC__nec_ismcli_user": "",
                        "driver__MStorageFC__nec_ismview_alloptimize": false,
                        "driver__MStorageFC__nec_ismview_dir": "/tmp/nec/cinder",
                        "driver__MStorageFC__nec_ldname_format": "LX:%s",
                        "driver__MStorageFC__nec_ldset": "",
                        "driver__MStorageFC__nec_pools__transform_csv": "",
                        "driver__MStorageFC__nec_queryconfig_view": false,
                        "driver__MStorageFC__nec_ssh_pool_port_number": 22,
                        "driver__MStorageFC__nec_unpairthread_timeout": 3600,
                        "driver__MStorageISCSI__nec_actual_free_capacity": false,
                        "driver__MStorageISCSI__nec_auto_accesscontrol": true,
                        "driver__MStorageISCSI__nec_backend_max_ld_count": 1024,
                        "driver__MStorageISCSI__nec_backup_ldname_format": "LX:%s",
                        "driver__MStorageISCSI__nec_backup_pools__transform_csv": "",
                        "driver__MStorageISCSI__nec_cv_ldname_format": "LX:__ControlVolume_%xh",
                        "driver__MStorageISCSI__nec_diskarray_name": "",
                        "driver__MStorageISCSI__nec_ismcli_fip": "",
                        "driver__MStorageISCSI__nec_ismcli_password": "",
                        "driver__MStorageISCSI__nec_ismcli_privkey": "",
                        "driver__MStorageISCSI__nec_ismcli_user": "",
                        "driver__MStorageISCSI__nec_ismview_alloptimize": false,
                        "driver__MStorageISCSI__nec_ismview_dir": "/tmp/nec/cinder",
                        "driver__MStorageISCSI__nec_ldname_format": "LX:%s",
                        "driver__MStorageISCSI__nec_ldset": "",
                        "driver__MStorageISCSI__nec_pools__transform_csv": "",
                        "driver__MStorageISCSI__nec_queryconfig_view": false,
                        "driver__MStorageISCSI__nec_ssh_pool_port_number": 22,
                        "driver__MStorageISCSI__nec_unpairthread_timeout": 3600,
                        "driver__MacroSANFC__backend_availability_zone": "",
                        "driver__MacroSANFC__chap_password": "",
                        "driver__MacroSANFC__chap_username": "",
                        "driver__MacroSANFC__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__MacroSANFC__driver_client_cert": "",
                        "driver__MacroSANFC__driver_client_cert_key": "",
                        "driver__MacroSANFC__driver_data_namespace": "",
                        "driver__MacroSANFC__driver_ssl_cert_path": "",
                        "driver__MacroSANFC__driver_ssl_cert_verify": false,
                        "driver__MacroSANFC__driver_use_ssl": false,
                        "driver__MacroSANFC__filter_function": "",
                        "driver__MacroSANFC__goodness_function": "",
                        "driver__MacroSANFC__iscsi_iotype": "fileio",
                        "driver__MacroSANFC__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__MacroSANFC__iscsi_target_flags": "",
                        "driver__MacroSANFC__iscsi_write_cache": "on",
                        "driver__MacroSANFC__num_shell_tries": 3,
                        "driver__MacroSANFC__num_volume_device_scan_tries": 3,
                        "driver__MacroSANFC__report_discard_supported": false,
                        "driver__MacroSANFC__storage_protocol": "iscsi",
                        "driver__MacroSANFC__target_helper": "tgtadm",
                        "driver__MacroSANFC__target_ip_address": "$my_ip",
                        "driver__MacroSANFC__target_port": 3260,
                        "driver__MacroSANFC__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__MacroSANFC__target_protocol": "iscsi",
                        "driver__MacroSANFC__trace_flags__transform_csv": "",
                        "driver__MacroSANFC__use_chap_auth": false,
                        "driver__MacroSANFC__volume_backend_name": "",
                        "driver__MacroSANFC__volume_clear": "zero",
                        "driver__MacroSANFC__volume_clear_ionice": "",
                        "driver__MacroSANFC__volume_clear_size": 0,
                        "driver__MacroSANFC__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__MacroSANFC__volume_copy_bps_limit": 0,
                        "driver__MacroSANFC__volume_dd_blocksize": "1M",
                        "driver__MacroSANFC__volumes_dir": "$state_path/volumes",
                        "driver__MacroSANISCSI__backend_availability_zone": "",
                        "driver__MacroSANISCSI__chap_password": "",
                        "driver__MacroSANISCSI__chap_username": "",
                        "driver__MacroSANISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__MacroSANISCSI__driver_client_cert": "",
                        "driver__MacroSANISCSI__driver_client_cert_key": "",
                        "driver__MacroSANISCSI__driver_data_namespace": "",
                        "driver__MacroSANISCSI__driver_ssl_cert_path": "",
                        "driver__MacroSANISCSI__driver_ssl_cert_verify": false,
                        "driver__MacroSANISCSI__driver_use_ssl": false,
                        "driver__MacroSANISCSI__filter_function": "",
                        "driver__MacroSANISCSI__goodness_function": "",
                        "driver__MacroSANISCSI__iscsi_iotype": "fileio",
                        "driver__MacroSANISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__MacroSANISCSI__iscsi_target_flags": "",
                        "driver__MacroSANISCSI__iscsi_write_cache": "on",
                        "driver__MacroSANISCSI__num_shell_tries": 3,
                        "driver__MacroSANISCSI__num_volume_device_scan_tries": 3,
                        "driver__MacroSANISCSI__report_discard_supported": false,
                        "driver__MacroSANISCSI__storage_protocol": "iscsi",
                        "driver__MacroSANISCSI__target_helper": "tgtadm",
                        "driver__MacroSANISCSI__target_ip_address": "$my_ip",
                        "driver__MacroSANISCSI__target_port": 3260,
                        "driver__MacroSANISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__MacroSANISCSI__target_protocol": "iscsi",
                        "driver__MacroSANISCSI__trace_flags__transform_csv": "",
                        "driver__MacroSANISCSI__use_chap_auth": false,
                        "driver__MacroSANISCSI__volume_backend_name": "",
                        "driver__MacroSANISCSI__volume_clear": "zero",
                        "driver__MacroSANISCSI__volume_clear_ionice": "",
                        "driver__MacroSANISCSI__volume_clear_size": 0,
                        "driver__MacroSANISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__MacroSANISCSI__volume_copy_bps_limit": 0,
                        "driver__MacroSANISCSI__volume_dd_blocksize": "1M",
                        "driver__MacroSANISCSI__volumes_dir": "$state_path/volumes",
                        "driver__NetAppCmodeFibreChannel__netapp_vserver": "",
                        "driver__NetAppCmodeISCSI__netapp_vserver": "",
                        "driver__NexentaISCSI__nexenta_blocksize": 4096,
                        "driver__NexentaISCSI__nexenta_dataset_compression": "on",
                        "driver__NexentaISCSI__nexenta_dataset_dedup": "off",
                        "driver__NexentaISCSI__nexenta_dataset_description": "",
                        "driver__NexentaISCSI__nexenta_folder": "",
                        "driver__NexentaISCSI__nexenta_group_snapshot_template": "group-snapshot-%s",
                        "driver__NexentaISCSI__nexenta_host": "",
                        "driver__NexentaISCSI__nexenta_host_group_prefix": "cinder",
                        "driver__NexentaISCSI__nexenta_iscsi_target_host_group": "all",
                        "driver__NexentaISCSI__nexenta_iscsi_target_portal_groups": "",
                        "driver__NexentaISCSI__nexenta_iscsi_target_portal_port": 3260,
                        "driver__NexentaISCSI__nexenta_iscsi_target_portals": "",
                        "driver__NexentaISCSI__nexenta_lu_writebackcache_disabled": false,
                        "driver__NexentaISCSI__nexenta_luns_per_target": 100,
                        "driver__NexentaISCSI__nexenta_ns5_blocksize": 32,
                        "driver__NexentaISCSI__nexenta_origin_snapshot_template": "origin-snapshot-%s",
                        "driver__NexentaISCSI__nexenta_rest_backoff_factor__transform_string_float": "0.5",
                        "driver__NexentaISCSI__nexenta_rest_connect_timeout__transform_string_float": "30",
                        "driver__NexentaISCSI__nexenta_rest_protocol": "auto",
                        "driver__NexentaISCSI__nexenta_rest_read_timeout__transform_string_float": "300",
                        "driver__NexentaISCSI__nexenta_rest_retry_count": 3,
                        "driver__NexentaISCSI__nexenta_sparse": false,
                        "driver__NexentaISCSI__nexenta_target_group_prefix": "cinder",
                        "driver__NexentaISCSI__nexenta_target_prefix": "iqn.1986-03.com.sun:02:cinder",
                        "driver__NexentaISCSI__nexenta_use_https": true,
                        "driver__NexentaISCSI__nexenta_volume": "cinder",
                        "driver__NexentaISCSI__nexenta_volume_group": "iscsi",
                        "driver__NimbleFC__nimble_pool_name": "default",
                        "driver__NimbleFC__nimble_subnet_label": "*",
                        "driver__NimbleFC__nimble_verify_cert_path": "",
                        "driver__NimbleFC__nimble_verify_certificate": false,
                        "driver__NimbleISCSI__nimble_pool_name": "default",
                        "driver__NimbleISCSI__nimble_subnet_label": "*",
                        "driver__NimbleISCSI__nimble_verify_cert_path": "",
                        "driver__NimbleISCSI__nimble_verify_certificate": false,
                        "driver__PVMEFC__driver_ssl_cert_path": "",
                        "driver__PVMEFC__driver_ssl_cert_verify": false,
                        "driver__PVMEFC__driver_use_ssl": false,
                        "driver__PVMEFC__pvme_pool_name": "A",
                        "driver__PVMEFC__san_ip": "",
                        "driver__PVMEFC__san_login": "admin",
                        "driver__PVMEFC__san_password": "",
                        "driver__PVMEISCSI__driver_ssl_cert_path": "",
                        "driver__PVMEISCSI__driver_ssl_cert_verify": false,
                        "driver__PVMEISCSI__driver_use_ssl": false,
                        "driver__PVMEISCSI__pvme_iscsi_ips__transform_csv": "",
                        "driver__PVMEISCSI__pvme_pool_name": "A",
                        "driver__PVMEISCSI__san_ip": "",
                        "driver__PVMEISCSI__san_login": "admin",
                        "driver__PVMEISCSI__san_password": "",
                        "driver__PowerFlex__powerflex_allow_migration_during_rebuild": false,
                        "driver__PowerFlex__powerflex_allow_non_padded_volumes": false,
                        "driver__PowerFlex__powerflex_max_over_subscription_ratio__transform_string_float": "10.0",
                        "driver__PowerFlex__powerflex_rest_server_port": 443,
                        "driver__PowerFlex__powerflex_round_volume_capacity": true,
                        "driver__PowerFlex__powerflex_server_api_version": "",
                        "driver__PowerFlex__powerflex_storage_pools": "",
                        "driver__PowerFlex__powerflex_unmap_volume_before_deletion": false,
                        "driver__PowerMaxFC__driver_ssl_cert_verify": false,
                        "driver__PowerMaxFC__initiator_check": false,
                        "driver__PowerMaxFC__interval": 3,
                        "driver__PowerMaxFC__load_balance": false,
                        "driver__PowerMaxFC__load_balance_real_time": false,
                        "driver__PowerMaxFC__load_data_format": "Avg",
                        "driver__PowerMaxFC__load_look_back": 60,
                        "driver__PowerMaxFC__load_look_back_real_time": 1,
                        "driver__PowerMaxFC__port_group_load_metric": "PercentBusy",
                        "driver__PowerMaxFC__port_load_metric": "PercentBusy",
                        "driver__PowerMaxFC__powermax_array": "",
                        "driver__PowerMaxFC__powermax_array_tag_list__transform_csv": "",
                        "driver__PowerMaxFC__powermax_port_group_name_template": "portGroupName",
                        "driver__PowerMaxFC__powermax_port_groups__transform_csv": "",
                        "driver__PowerMaxFC__powermax_service_level": "",
                        "driver__PowerMaxFC__powermax_short_host_name_template": "shortHostName",
                        "driver__PowerMaxFC__powermax_snapvx_unlink_limit": 3,
                        "driver__PowerMaxFC__powermax_srp": "",
                        "driver__PowerMaxFC__retries": 200,
                        "driver__PowerMaxFC__san_ip": "",
                        "driver__PowerMaxFC__san_login": "admin",
                        "driver__PowerMaxFC__san_password": "",
                        "driver__PowerMaxFC__u4p_failover_autofailback": true,
                        "driver__PowerMaxFC__u4p_failover_backoff_factor": 1,
                        "driver__PowerMaxFC__u4p_failover_retries": 3,
                        "driver__PowerMaxFC__u4p_failover_target__transform_csv_kvs": "",
                        "driver__PowerMaxFC__u4p_failover_timeout": 20,
                        "driver__PowerMaxFC__vmax_workload": "",
                        "driver__PowerMaxISCSI__chap_password": "",
                        "driver__PowerMaxISCSI__chap_username": "",
                        "driver__PowerMaxISCSI__driver_ssl_cert_verify": false,
                        "driver__PowerMaxISCSI__initiator_check": false,
                        "driver__PowerMaxISCSI__interval": 3,
                        "driver__PowerMaxISCSI__load_balance": false,
                        "driver__PowerMaxISCSI__load_balance_real_time": false,
                        "driver__PowerMaxISCSI__load_data_format": "Avg",
                        "driver__PowerMaxISCSI__load_look_back": 60,
                        "driver__PowerMaxISCSI__load_look_back_real_time": 1,
                        "driver__PowerMaxISCSI__port_group_load_metric": "PercentBusy",
                        "driver__PowerMaxISCSI__port_load_metric": "PercentBusy",
                        "driver__PowerMaxISCSI__powermax_array": "",
                        "driver__PowerMaxISCSI__powermax_array_tag_list__transform_csv": "",
                        "driver__PowerMaxISCSI__powermax_port_group_name_template": "portGroupName",
                        "driver__PowerMaxISCSI__powermax_port_groups__transform_csv": "",
                        "driver__PowerMaxISCSI__powermax_service_level": "",
                        "driver__PowerMaxISCSI__powermax_short_host_name_template": "shortHostName",
                        "driver__PowerMaxISCSI__powermax_snapvx_unlink_limit": 3,
                        "driver__PowerMaxISCSI__powermax_srp": "",
                        "driver__PowerMaxISCSI__retries": 200,
                        "driver__PowerMaxISCSI__san_ip": "",
                        "driver__PowerMaxISCSI__san_login": "admin",
                        "driver__PowerMaxISCSI__san_password": "",
                        "driver__PowerMaxISCSI__u4p_failover_autofailback": true,
                        "driver__PowerMaxISCSI__u4p_failover_backoff_factor": 1,
                        "driver__PowerMaxISCSI__u4p_failover_retries": 3,
                        "driver__PowerMaxISCSI__u4p_failover_target__transform_csv_kvs": "",
                        "driver__PowerMaxISCSI__u4p_failover_timeout": 20,
                        "driver__PowerMaxISCSI__use_chap_auth": false,
                        "driver__PowerMaxISCSI__vmax_workload": "",
                        "driver__PowerStore__powerstore_appliances__transform_csv": "",
                        "driver__PowerStore__powerstore_ports__transform_csv": "",
                        "driver__PureFC__driver_ssl_cert_path": "",
                        "driver__PureFC__driver_ssl_cert_verify": false,
                        "driver__PureFC__pure_api_token": "",
                        "driver__PureFC__pure_automatic_max_oversubscription_ratio": true,
                        "driver__PureFC__pure_eradicate_on_delete": false,
                        "driver__PureFC__pure_host_personality__transform_empty_none": "",
                        "driver__PureFC__pure_iscsi_cidr": "0.0.0.0/0",
                        "driver__PureFC__pure_replica_interval_default": 3600,
                        "driver__PureFC__pure_replica_retention_long_term_default": 7,
                        "driver__PureFC__pure_replica_retention_long_term_per_day_default": 3,
                        "driver__PureFC__pure_replica_retention_short_term_default": 14400,
                        "driver__PureFC__pure_replication_pg_name": "cinder-group",
                        "driver__PureFC__pure_replication_pod_name": "cinder-pod",
                        "driver__PureFC__san_ip": "",
                        "driver__PureFC__use_chap_auth": false,
                        "driver__PureISCSI__driver_ssl_cert_path": "",
                        "driver__PureISCSI__driver_ssl_cert_verify": false,
                        "driver__PureISCSI__pure_api_token": "",
                        "driver__PureISCSI__pure_automatic_max_oversubscription_ratio": true,
                        "driver__PureISCSI__pure_eradicate_on_delete": false,
                        "driver__PureISCSI__pure_host_personality__transform_empty_none": "",
                        "driver__PureISCSI__pure_iscsi_cidr": "0.0.0.0/0",
                        "driver__PureISCSI__pure_replica_interval_default": 3600,
                        "driver__PureISCSI__pure_replica_retention_long_term_default": 7,
                        "driver__PureISCSI__pure_replica_retention_long_term_per_day_default": 3,
                        "driver__PureISCSI__pure_replica_retention_short_term_default": 14400,
                        "driver__PureISCSI__pure_replication_pg_name": "cinder-group",
                        "driver__PureISCSI__pure_replication_pod_name": "cinder-pod",
                        "driver__PureISCSI__san_ip": "",
                        "driver__PureISCSI__use_chap_auth": false,
                        "driver__QnapISCSI__chap_password": "",
                        "driver__QnapISCSI__chap_username": "",
                        "driver__QnapISCSI__driver_ssl_cert_verify": false,
                        "driver__QnapISCSI__qnap_management_url": "",
                        "driver__QnapISCSI__qnap_poolname": "",
                        "driver__QnapISCSI__qnap_storage_protocol": "iscsi",
                        "driver__QnapISCSI__san_login": "admin",
                        "driver__QnapISCSI__san_password": "",
                        "driver__QnapISCSI__target_ip_address": "$my_ip",
                        "driver__QnapISCSI__use_chap_auth": false,
                        "driver__Quobyte__quobyte_client_cfg": "",
                        "driver__Quobyte__quobyte_mount_point_base": "$state_path/mnt",
                        "driver__Quobyte__quobyte_overlay_volumes": false,
                        "driver__Quobyte__quobyte_qcow2_volumes": true,
                        "driver__Quobyte__quobyte_sparsed_volumes": true,
                        "driver__Quobyte__quobyte_volume_from_snapshot_cache": false,
                        "driver__Quobyte__quobyte_volume_url": "",
                        "driver__RBD__deferred_deletion_delay": 0,
                        "driver__RBD__deferred_deletion_purge_interval": 60,
                        "driver__RBD__enable_deferred_deletion": false,
                        "driver__RBD__rados_connect_timeout": -1,
                        "driver__RBD__rados_connection_interval": 5,
                        "driver__RBD__rados_connection_retries": 3,
                        "driver__RBD__rbd_ceph_conf": "",
                        "driver__RBD__rbd_cluster_name": "ceph",
                        "driver__RBD__rbd_exclusive_cinder_pool": false,
                        "driver__RBD__rbd_flatten_volume_from_snapshot": false,
                        "driver__RBD__rbd_keyring_conf": "",
                        "driver__RBD__rbd_max_clone_depth": 5,
                        "driver__RBD__rbd_pool": "rbd",
                        "driver__RBD__rbd_secret_uuid": "",
                        "driver__RBD__rbd_store_chunk_size": 4,
                        "driver__RBD__rbd_user": "",
                        "driver__RBD__report_dynamic_total_capacity": true,
                        "driver__RBD__volume_dd_blocksize": "1M",
                        "driver__RBDISCSI__deferred_deletion_delay": 0,
                        "driver__RBDISCSI__deferred_deletion_purge_interval": 60,
                        "driver__RBDISCSI__driver_ssl_cert_verify": false,
                        "driver__RBDISCSI__enable_deferred_deletion": false,
                        "driver__RBDISCSI__rados_connect_timeout": -1,
                        "driver__RBDISCSI__rados_connection_interval": 5,
                        "driver__RBDISCSI__rados_connection_retries": 3,
                        "driver__RBDISCSI__rbd_ceph_conf": "",
                        "driver__RBDISCSI__rbd_cluster_name": "ceph",
                        "driver__RBDISCSI__rbd_exclusive_cinder_pool": false,
                        "driver__RBDISCSI__rbd_flatten_volume_from_snapshot": false,
                        "driver__RBDISCSI__rbd_iscsi_api_debug": false,
                        "driver__RBDISCSI__rbd_iscsi_api_password": "",
                        "driver__RBDISCSI__rbd_iscsi_api_url": "",
                        "driver__RBDISCSI__rbd_iscsi_api_user": "",
                        "driver__RBDISCSI__rbd_iscsi_target_iqn": "",
                        "driver__RBDISCSI__rbd_max_clone_depth": 5,
                        "driver__RBDISCSI__rbd_pool": "rbd",
                        "driver__RBDISCSI__rbd_secret_uuid": "",
                        "driver__RBDISCSI__rbd_store_chunk_size": 4,
                        "driver__RBDISCSI__rbd_user": "",
                        "driver__RBDISCSI__report_dynamic_total_capacity": true,
                        "driver__RBDISCSI__suppress_requests_ssl_warnings": false,
                        "driver__RBDISCSI__volume_dd_blocksize": "1M",
                        "driver__RSD__podm_password": "",
                        "driver__RSD__podm_url": "",
                        "driver__RSD__podm_username": "",
                        "driver__SCFC__dell_api_async_rest_timeout": 15,
                        "driver__SCFC__dell_api_sync_rest_timeout": 30,
                        "driver__SCFC__dell_sc_api_port": 3033,
                        "driver__SCFC__dell_sc_server_folder": "openstack",
                        "driver__SCFC__dell_sc_ssn": 64702,
                        "driver__SCFC__dell_sc_verify_cert": false,
                        "driver__SCFC__dell_sc_volume_folder": "openstack",
                        "driver__SCFC__dell_server_os": "Red Hat Linux 6.x",
                        "driver__SCFC__excluded_domain_ips__transform_csv": "",
                        "driver__SCFC__included_domain_ips__transform_csv": "",
                        "driver__SCFC__secondary_san_ip": "",
                        "driver__SCFC__secondary_san_login": "Admin",
                        "driver__SCFC__secondary_san_password": "",
                        "driver__SCFC__secondary_sc_api_port": 3033,
                        "driver__SCISCSI__dell_api_async_rest_timeout": 15,
                        "driver__SCISCSI__dell_api_sync_rest_timeout": 30,
                        "driver__SCISCSI__dell_sc_api_port": 3033,
                        "driver__SCISCSI__dell_sc_server_folder": "openstack",
                        "driver__SCISCSI__dell_sc_ssn": 64702,
                        "driver__SCISCSI__dell_sc_verify_cert": false,
                        "driver__SCISCSI__dell_sc_volume_folder": "openstack",
                        "driver__SCISCSI__dell_server_os": "Red Hat Linux 6.x",
                        "driver__SCISCSI__excluded_domain_ips__transform_csv": "",
                        "driver__SCISCSI__included_domain_ips__transform_csv": "",
                        "driver__SCISCSI__secondary_san_ip": "",
                        "driver__SCISCSI__secondary_san_login": "Admin",
                        "driver__SCISCSI__secondary_san_password": "",
                        "driver__SCISCSI__secondary_sc_api_port": 3033,
                        "driver__SdsISCSI__backend_availability_zone": "",
                        "driver__SdsISCSI__chap_password": "",
                        "driver__SdsISCSI__chap_username": "",
                        "driver__SdsISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__SdsISCSI__driver_client_cert": "",
                        "driver__SdsISCSI__driver_client_cert_key": "",
                        "driver__SdsISCSI__driver_data_namespace": "",
                        "driver__SdsISCSI__driver_ssl_cert_path": "",
                        "driver__SdsISCSI__driver_ssl_cert_verify": false,
                        "driver__SdsISCSI__driver_use_ssl": false,
                        "driver__SdsISCSI__filter_function": "",
                        "driver__SdsISCSI__goodness_function": "",
                        "driver__SdsISCSI__iscsi_iotype": "fileio",
                        "driver__SdsISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__SdsISCSI__iscsi_target_flags": "",
                        "driver__SdsISCSI__iscsi_write_cache": "on",
                        "driver__SdsISCSI__num_shell_tries": 3,
                        "driver__SdsISCSI__num_volume_device_scan_tries": 3,
                        "driver__SdsISCSI__report_discard_supported": false,
                        "driver__SdsISCSI__storage_protocol": "iscsi",
                        "driver__SdsISCSI__target_helper": "tgtadm",
                        "driver__SdsISCSI__target_ip_address": "$my_ip",
                        "driver__SdsISCSI__target_port": 3260,
                        "driver__SdsISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__SdsISCSI__target_protocol": "iscsi",
                        "driver__SdsISCSI__trace_flags__transform_csv": "",
                        "driver__SdsISCSI__use_chap_auth": false,
                        "driver__SdsISCSI__volume_backend_name": "",
                        "driver__SdsISCSI__volume_clear": "zero",
                        "driver__SdsISCSI__volume_clear_ionice": "",
                        "driver__SdsISCSI__volume_clear_size": 0,
                        "driver__SdsISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__SdsISCSI__volume_copy_bps_limit": 0,
                        "driver__SdsISCSI__volume_dd_blocksize": "1M",
                        "driver__SdsISCSI__volumes_dir": "$state_path/volumes",
                        "driver__SolidFire__driver_ssl_cert_verify": false,
                        "driver__SolidFire__san_ip": "",
                        "driver__SolidFire__san_login": "admin",
                        "driver__SolidFire__san_password": "",
                        "driver__SolidFire__sf_account_prefix": "",
                        "driver__SolidFire__sf_allow_tenant_qos": false,
                        "driver__SolidFire__sf_api_port": 443,
                        "driver__SolidFire__sf_api_request_timeout": 30,
                        "driver__SolidFire__sf_cluster_pairing_timeout": 60,
                        "driver__SolidFire__sf_emulate_512": true,
                        "driver__SolidFire__sf_enable_vag": false,
                        "driver__SolidFire__sf_provisioning_calc": "maxProvisionedSpace",
                        "driver__SolidFire__sf_svip": "",
                        "driver__SolidFire__sf_volume_clone_timeout": 600,
                        "driver__SolidFire__sf_volume_create_timeout": 60,
                        "driver__SolidFire__sf_volume_pairing_timeout": 3600,
                        "driver__SolidFire__sf_volume_prefix": "UUID-",
                        "driver__StorPool__storpool_replication": 3,
                        "driver__StorPool__storpool_template": "",
                        "driver__StorwizeSVCFC__cycle_period_seconds": 300,
                        "driver__StorwizeSVCFC__storwize_peer_pool": "",
                        "driver__StorwizeSVCFC__storwize_preferred_host_site__transform_csv_kvs": "",
                        "driver__StorwizeSVCFC__storwize_san_secondary_ip": "",
                        "driver__StorwizeSVCFC__storwize_svc_allow_tenant_qos": false,
                        "driver__StorwizeSVCFC__storwize_svc_flashcopy_rate": 50,
                        "driver__StorwizeSVCFC__storwize_svc_flashcopy_timeout": 120,
                        "driver__StorwizeSVCFC__storwize_svc_mirror_pool": "",
                        "driver__StorwizeSVCFC__storwize_svc_multipath_enabled": false,
                        "driver__StorwizeSVCFC__storwize_svc_retain_aux_volume": false,
                        "driver__StorwizeSVCFC__storwize_svc_stretched_cluster_partner": "",
                        "driver__StorwizeSVCFC__storwize_svc_vol_autoexpand": true,
                        "driver__StorwizeSVCFC__storwize_svc_vol_compression": false,
                        "driver__StorwizeSVCFC__storwize_svc_vol_easytier": true,
                        "driver__StorwizeSVCFC__storwize_svc_vol_grainsize": 256,
                        "driver__StorwizeSVCFC__storwize_svc_vol_iogrp": "0",
                        "driver__StorwizeSVCFC__storwize_svc_vol_nofmtdisk": false,
                        "driver__StorwizeSVCFC__storwize_svc_vol_rsize": 2,
                        "driver__StorwizeSVCFC__storwize_svc_vol_warning": 0,
                        "driver__StorwizeSVCFC__storwize_svc_volpool_name__transform_csv": "volpool",
                        "driver__StorwizeSVCISCSI__cycle_period_seconds": 300,
                        "driver__StorwizeSVCISCSI__storwize_peer_pool": "",
                        "driver__StorwizeSVCISCSI__storwize_preferred_host_site__transform_csv_kvs": "",
                        "driver__StorwizeSVCISCSI__storwize_san_secondary_ip": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_allow_tenant_qos": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_flashcopy_rate": 50,
                        "driver__StorwizeSVCISCSI__storwize_svc_flashcopy_timeout": 120,
                        "driver__StorwizeSVCISCSI__storwize_svc_iscsi_chap_enabled": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_mirror_pool": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_retain_aux_volume": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_stretched_cluster_partner": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_autoexpand": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_compression": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_easytier": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_grainsize": 256,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_iogrp": "0",
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_nofmtdisk": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_rsize": 2,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_warning": 0,
                        "driver__StorwizeSVCISCSI__storwize_svc_volpool_name__transform_csv": "volpool",
                        "driver__SynoISCSI__chap_password": "",
                        "driver__SynoISCSI__chap_username": "",
                        "driver__SynoISCSI__driver_use_ssl": false,
                        "driver__SynoISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__SynoISCSI__synology_admin_port": 5000,
                        "driver__SynoISCSI__synology_device_id": "",
                        "driver__SynoISCSI__synology_one_time_pass": "",
                        "driver__SynoISCSI__synology_password": "",
                        "driver__SynoISCSI__synology_pool_name": "",
                        "driver__SynoISCSI__synology_ssl_verify": true,
                        "driver__SynoISCSI__synology_username": "admin",
                        "driver__SynoISCSI__target_ip_address": "$my_ip",
                        "driver__SynoISCSI__target_port": 3260,
                        "driver__SynoISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__SynoISCSI__target_protocol": "iscsi",
                        "driver__SynoISCSI__use_chap_auth": false,
                        "driver__Unity__remove_empty_host": false,
                        "driver__Unity__unity_io_ports__transform_csv": "",
                        "driver__Unity__unity_storage_pool_names__transform_csv": "",
                        "driver__VNX__default_timeout": 31536000,
                        "driver__VNX__destroy_empty_storage_group": false,
                        "driver__VNX__force_delete_lun_in_storagegroup": true,
                        "driver__VNX__ignore_pool_full_threshold": false,
                        "driver__VNX__initiator_auto_deregistration": false,
                        "driver__VNX__initiator_auto_registration": false,
                        "driver__VNX__io_port_list__transform_csv": "",
                        "driver__VNX__iscsi_initiators": "",
                        "driver__VNX__max_luns_per_storage_group": 255,
                        "driver__VNX__naviseccli_path": "",
                        "driver__VNX__storage_vnx_authentication_type": "global",
                        "driver__VNX__storage_vnx_pool_names__transform_csv": "",
                        "driver__VNX__storage_vnx_security_file_dir": "",
                        "driver__VNX__vnx_async_migrate": true,
                        "driver__VZStorage__vzstorage_default_volume_format": "raw",
                        "driver__VZStorage__vzstorage_mount_options__transform_csv": "",
                        "driver__VZStorage__vzstorage_mount_point_base": "$state_path/mnt",
                        "driver__VZStorage__vzstorage_shares_config": "/etc/cinder/vzstorage_shares",
                        "driver__VZStorage__vzstorage_sparsed_volumes": true,
                        "driver__VZStorage__vzstorage_used_ratio__transform_string_float": "0.95",
                        "driver__WindowsISCSI__windows_iscsi_lun_path": "C:\\iSCSIVirtualDisks",
                        "driver__WindowsSmbfs__smbfs_default_volume_format": "vhd",
                        "driver__WindowsSmbfs__smbfs_mount_point_base": "C:\\OpenStack\\_mnt",
                        "driver__WindowsSmbfs__smbfs_pool_mappings__transform_csv_kvs": "",
                        "driver__WindowsSmbfs__smbfs_shares_config": "C:\\OpenStack\\smbfs_shares.txt",
                        "driver__XtremIOFC__driver_ssl_cert_path": "",
                        "driver__XtremIOFC__driver_ssl_cert_verify": false,
                        "driver__XtremIOFC__san_ip": "",
                        "driver__XtremIOFC__san_login": "admin",
                        "driver__XtremIOFC__san_password": "",
                        "driver__XtremIOFC__xtremio_array_busy_retry_count": 5,
                        "driver__XtremIOFC__xtremio_array_busy_retry_interval": 5,
                        "driver__XtremIOFC__xtremio_clean_unused_ig": false,
                        "driver__XtremIOFC__xtremio_cluster_name": "",
                        "driver__XtremIOISCSI__driver_ssl_cert_path": "",
                        "driver__XtremIOISCSI__driver_ssl_cert_verify": false,
                        "driver__XtremIOISCSI__san_ip": "",
                        "driver__XtremIOISCSI__san_login": "admin",
                        "driver__XtremIOISCSI__san_password": "",
                        "driver__XtremIOISCSI__xtremio_array_busy_retry_count": 5,
                        "driver__XtremIOISCSI__xtremio_array_busy_retry_interval": 5,
                        "driver__XtremIOISCSI__xtremio_clean_unused_ig": false,
                        "driver__XtremIOISCSI__xtremio_cluster_name": "",
                        "driver__ZadaraVPSAISCSI__zadara_access_key": "",
                        "driver__ZadaraVPSAISCSI__zadara_default_snap_policy": false,
                        "driver__ZadaraVPSAISCSI__zadara_ssl_cert_verify": true,
                        "driver__ZadaraVPSAISCSI__zadara_use_iser": true,
                        "driver__ZadaraVPSAISCSI__zadara_vol_encrypt": false,
                        "driver__ZadaraVPSAISCSI__zadara_vol_name_template": "OS_%s",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_host": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_poolname": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_port": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_use_ssl": false
                      }
                    },
                    "sysFiles": {
                      "name": ""
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/embercsi/ember-csi-operator:latest
          createdAt: 2021-02-22:17:47:00
          description: Multi-vendor CSI plugin supporting 80+ storage drivers
          repository: https://github.com/embercsi/ember-csi-operator
          support: http://readthedocs.org/projects/ember-csi/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an Ember CSI storage backend
            displayName: Storage Backends
            kind: EmberStorageBackend
            name: embercsis.ember-csi.io
            version: v1alpha1
        description: |
          Ember-CSI is a multi-vendor CSI plugin driver supporting
          over 80 storage drivers in a single plugin to provide block and mount
          storage to container orchestration systems.

          The operator adds support for management of multiple Ember CSI instances
          within a cluster and simplifies storage backend deployments.

          ### Supported drivers
          Ember-CSI includes a good number of storage drivers, but due to limitation on hardware availability only a small number of them have been validated at one point or another. In alphabetical order they are:

          * HPE3PARFC
          * HPE3PARISCSI
          * KaminarioISCSI
          * LVMVolume
          * PowerMaxFC
          * PowerMaxISCSI
          * PureFC
          * PureISCSI
          * QnapISCSI
          * RBD
          * SolidFire
          * SynoISCSI
          * XtremIOFC
          * XtremIOISCSI

          ### Getting started
          Please have a look at our quick start documentation on how to get started:
          [github.com/embercsi/ember-csi-operator#quick-start](https://github.com/embercsi/ember-csi-operator#quick-start).
          We recommend to use the user friendly "Edit Form" when creating a new
          EmberStorageBackend. You'll find the link in the upper right corner of the
          editor.

          ### Documentation
          Documentation on Ember CSI itself can be found at [docs.ember-csi.io](http://docs.ember-csi.io/).

          ### Getting help
          Please contact us if you encounter any issues while using Ember CSI.
          Get in touch through the #ember-csi channel on Freenode, also accessible through
          [Kiwi IRC](https://kiwiirc.com/nextclient/irc.freenode.net/ember-csi).

          You can also create an issue on the [Github project](https://github.com/embercsi/ember-csi-operator)
          for bugs, enhancements, or other requests.
        displayName: Ember CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Ember-CSI
        - CSI
        links:
        - name: Learn more about the project
          url: http://ember-csi.io/
        - name: Documentation
          url: http://readthedocs.org/projects/ember-csi/
        - name: Ember-CSI Source Code
          url: https://github.com/embercsi/ember-csi
        - name: Ember-CSI Operator Source Code
          url: https://github.com/embercsi/ember-csi-operator
        maintainers:
        - email: geguileo@redhat.com
          name: Gorka Eguileor
        - email: cschwede@redhat.com
          name: Christian Schwede
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/embercsi/ember-csi-operator:latest
        version: 0.9.9
      entries:
      - name: ember-csi-community-operator.v0.9.9
        version: 0.9.9
      - name: ember-csi-community-operator.v0.9.8
        version: 0.9.8
      - name: ember-csi-community-operator.v0.9.7
        version: 0.9.7
      name: beta
    defaultChannel: beta
    packageName: ember-csi-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-dynamodb-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-dynamodb-controller.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "GlobalTable",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "Table",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/dynamodb-controller:1.5.2
          createdAt: "2025-09-19T19:19:47Z"
          description: AWS DynamoDB controller is a service controller for managing
            DynamoDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup represents the state of an AWS dynamodb Backup resource.
            displayName: Backup
            kind: Backup
            name: backups.dynamodb.services.k8s.aws
            version: v1alpha1
          - description: GlobalTable represents the state of an AWS dynamodb GlobalTable
              resource.
            displayName: GlobalTable
            kind: GlobalTable
            name: globaltables.dynamodb.services.k8s.aws
            version: v1alpha1
          - description: Table represents the state of an AWS dynamodb Table resource.
            displayName: Table
            kind: Table
            name: tables.dynamodb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon DynamoDB resources in AWS from within your Kubernetes cluster.

          **About Amazon DynamoDB**

          Amazon DynamoDB is a fully managed NoSQL database service that provides fast and predictable performance with seamless scalability. DynamoDB lets you offload the administrative burdens of operating and scaling a distributed database so that you don't have to worry about hardware provisioning, setup and configuration, replication, software patching, or cluster scaling. DynamoDB also offers encryption at rest, which eliminates the operational burden and complexity involved in protecting sensitive data. For more information, see [DynamoDB Encryption at Rest](https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/EncryptionAtRest.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon DynamoDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dynamodb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon DynamoDB Developer Resources
          url: https://aws.amazon.com/dynamodb/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: dynamodb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/dynamodb-controller:1.5.2
        version: 1.5.2
      entries:
      - name: ack-dynamodb-controller.v1.5.2
        version: 1.5.2
      - name: ack-dynamodb-controller.v1.5.1
        version: 1.5.1
      - name: ack-dynamodb-controller.v1.5.0
        version: 1.5.0
      - name: ack-dynamodb-controller.v1.4.10
        version: 1.4.10
      - name: ack-dynamodb-controller.v1.4.9
        version: 1.4.9
      - name: ack-dynamodb-controller.v1.4.7
        version: 1.4.7
      - name: ack-dynamodb-controller.v1.4.6
        version: 1.4.6
      - name: ack-dynamodb-controller.v1.4.5
        version: 1.4.5
      - name: ack-dynamodb-controller.v1.4.4
        version: 1.4.4
      - name: ack-dynamodb-controller.v1.4.3
        version: 1.4.3
      - name: ack-dynamodb-controller.v1.4.2
        version: 1.4.2
      - name: ack-dynamodb-controller.v1.4.1
        version: 1.4.1
      - name: ack-dynamodb-controller.v1.3.0
        version: 1.3.0
      - name: ack-dynamodb-controller.v1.2.23
        version: 1.2.23
      - name: ack-dynamodb-controller.v1.2.22
        version: 1.2.22
      - name: ack-dynamodb-controller.v1.2.21
        version: 1.2.21
      - name: ack-dynamodb-controller.v1.2.20
        version: 1.2.20
      - name: ack-dynamodb-controller.v1.2.15
        version: 1.2.15
      - name: ack-dynamodb-controller.v1.2.14
        version: 1.2.14
      - name: ack-dynamodb-controller.v1.2.13
        version: 1.2.13
      - name: ack-dynamodb-controller.v1.2.12
        version: 1.2.12
      - name: ack-dynamodb-controller.v1.2.11
        version: 1.2.11
      - name: ack-dynamodb-controller.v1.2.9
        version: 1.2.9
      - name: ack-dynamodb-controller.v1.2.8
        version: 1.2.8
      - name: ack-dynamodb-controller.v1.2.6
        version: 1.2.6
      - name: ack-dynamodb-controller.v1.2.5
        version: 1.2.5
      - name: ack-dynamodb-controller.v1.2.4
        version: 1.2.4
      - name: ack-dynamodb-controller.v1.2.2
        version: 1.2.2
      - name: ack-dynamodb-controller.v1.2.1
        version: 1.2.1
      - name: ack-dynamodb-controller.v1.2.0
        version: 1.2.0
      - name: ack-dynamodb-controller.v1.1.1
        version: 1.1.1
      - name: ack-dynamodb-controller.v1.1.0
        version: 1.1.0
      - name: ack-dynamodb-controller.v1.0.3
        version: 1.0.3
      - name: ack-dynamodb-controller.v1.0.2
        version: 1.0.2
      - name: ack-dynamodb-controller.v1.0.1
        version: 1.0.1
      - name: ack-dynamodb-controller.v1.0.0
        version: 1.0.0
      - name: ack-dynamodb-controller.v0.1.10
        version: 0.1.10
      - name: ack-dynamodb-controller.v0.1.9
        version: 0.1.9
      - name: ack-dynamodb-controller.v0.1.8
        version: 0.1.8
      - name: ack-dynamodb-controller.v0.1.7
        version: 0.1.7
      - name: ack-dynamodb-controller.v0.1.6
        version: 0.1.6
      - name: ack-dynamodb-controller.v0.1.5
        version: 0.1.5
      - name: ack-dynamodb-controller.v0.1.4
        version: 0.1.4
      - name: ack-dynamodb-controller.v0.1.3
        version: 0.1.3
      - name: ack-dynamodb-controller.v0.1.1
        version: 0.1.1
      - name: ack-dynamodb-controller.v0.1.0
        version: 0.1.0
      - name: ack-dynamodb-controller.v0.0.20
        version: 0.0.20
      - name: ack-dynamodb-controller.v0.0.19
        version: 0.0.19
      - name: ack-dynamodb-controller.v0.0.18
        version: 0.0.18
      - name: ack-dynamodb-controller.v0.0.17
        version: 0.0.17
      - name: ack-dynamodb-controller.v0.0.16
        version: 0.0.16
      - name: ack-dynamodb-controller.v0.0.15
        version: 0.0.15
      - name: ack-dynamodb-controller.v0.0.14
        version: 0.0.14
      - name: ack-dynamodb-controller.v0.0.13
        version: 0.0.13
      - name: ack-dynamodb-controller.v0.0.12
        version: 0.0.12
      - name: ack-dynamodb-controller.v0.0.11
        version: 0.0.11
      - name: ack-dynamodb-controller.v0.0.10
        version: 0.0.10
      - name: ack-dynamodb-controller.v0.0.9
        version: 0.0.9
      - name: ack-dynamodb-controller.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: ack-dynamodb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/limitador-operator
    name: limitador-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: limitador-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/limitador-operator:v0.5.0
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: The Limitador operator installs and maintains limitador instances
        displayName: Limitador
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        links:
        - name: Limitador Operator
          url: https://github.com/Kuadrant/limitador-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - quay.io/kuadrant/limitador-operator:v0.5.0
        - quay.io/kuadrant/limitador:v1.2.0
        version: 0.5.0
      entries:
      - name: limitador-operator.v0.5.0
        version: 0.5.0
      - name: limitador-operator.v0.4.0
        version: 0.4.0
      - name: limitador-operator.v0.3.0
        version: 0.3.0
      name: alpha
    - currentCSV: limitador-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "name": "toy_get_route",
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/limitador-operator:v0.11.0
          createdAt: "2024-10-02T10:46:31Z"
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: The Limitador operator installs and maintains limitador instances
        displayName: Limitador
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        links:
        - name: Limitador Operator
          url: https://github.com/Kuadrant/limitador-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - quay.io/kuadrant/limitador-operator:v0.11.0
        - quay.io/kuadrant/limitador:v1.6.0
        version: 0.11.0
      entries:
      - name: limitador-operator.v0.11.0
        version: 0.11.0
      - name: limitador-operator.v0.10.0
        version: 0.10.0
      - name: limitador-operator.v0.6.0
        version: 0.6.0
      - name: limitador-operator.v0.5.0
        version: 0.5.0
      - name: limitador-operator.v0.4.0
        version: 0.4.0
      - name: limitador-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: limitador-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/limitador-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Commvault
      provider-url: ""
    name: hedvig-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hedvig-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\":\"hedvig.io/v1\",\n    \"kind\":\"HedvigDeploy\",\n
            \   \"metadata\": {\n      \"name\": \"testdeploy\",\n      \"namespace\":
            \"hedvig\"\n    },\n    \"spec\": {\n      \"cluster\": {\n        \"name\":
            \"hedvigcl\",\n        \"seeds\": [\n          \"hedvigcl1.hedvig.io\",\n
            \         \"hedvigcl2.hedvig.io\",\n          \"hedvigcl3.hedvig.io\"\n
            \       ]\n      },\n      \"k8s\": {\n        \"name\": \"k8scl\"\n      },\n
            \     \"proxy\": {\n        \"repository\": \"hedvig/hedvig-proxy\",\n
            \       \"tag\": \"v-4.3.1\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"controller\": {\n        \"repository\": \"hedvig/hedvig-csi-volume-driver\",\n
            \       \"tag\": \"v-2.0\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"node\": {\n        \"repository\": \"hedvig/hedvig-csi-volume-driver\",\n
            \       \"tag\": \"v-2.0\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"sidecars\": [\n        {\n          \"name\": \"csi-provisioner\",\n
            \         \"repository\": \"quay.io/k8scsi/csi-provisioner\",\n          \"tag\":
            \"v1.6.0\",\n          \"imagePullPolicy\": \"IfNotPresent\"\n        },\n
            \       {\n          \"name\": \"csi-snapshotter\",\n          \"repository\":
            \"quay.io/k8scsi/csi-snapshotter\",\n          \"tag\": \"v2.1.0\",\n
            \         \"imagePullPolicy\": \"IfNotPresent\"\n        },              \n
            \       {\n          \"name\": \"csi-resizer\",\n          \"repository\":
            \"quay.io/k8scsi/csi-resizer\",\n          \"tag\": \"v0.5.0\",\n          \"imagePullPolicy\":
            \"IfNotPresent\"\n        },              \n        {\n          \"name\":
            \"csi-node-driver-registrar\",\n          \"repository\": \"quay.io/k8scsi/csi-node-driver-registrar\",\n
            \         \"tag\": \"v1.0-canary\",\n          \"imagePullPolicy\": \"IfNotPresent\"\n
            \       }\n      ]\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: hedvig/hedvig-operator:2.0.0-ubi
          createdAt: ""
          description: Hedvig Operator is the official operator to deploy and manage
            Hedvig components in container orchestrators such as Kubernetes and Openshift.
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hedvig Deploy
            displayName: HedvigDeploy
            kind: HedvigDeploy
            name: hedvigdeploys.hedvig.io
            version: v1
        description: "The Hedvig Distributed Storage Platform is a software-defined,
          scale-out storage solution. It distributes data simultaneously across multiple
          locations – from on-premises data centers to the cloud – and scales capacity
          on-demand by leveraging the storage of commodity x86/ARM servers. The distributed
          write maximizes availability and protects data from hardware failures ranging
          from a single disk to an entire site, improving disaster recovery planning.\n\nHedvig
          provides a CSI Driver for accelerating your journey into the container ecosystem.
          CSI Driver enables containerized applications to maintain state by dynamically
          provisioning and consuming Hedvig virtual disks as persistent volumes. \n###
          \n\nHedvig Operator deploys the Hedvig Storage Proxy and the Hedvig CSI
          Driver to manage the life cycle of persistent storage in Kubernetes and
          Openshift.\n\n**Supported Container Orchestrators:**\n* Kubernetes 1.16+\n*
          Red Hat OpenShift 4.6+\n\n**Supported Hedvig Releases:**\n* Hedvig 4.0+\nPlease
          refer to the [documentation](https://documentation.commvault.com/commvault/hedvig/others/pdf/Hedvig_Operator_User_Guide.pdf)
          for steps to install and use the operator.\n\n\n\n\n\n"
        displayName: Hedvig Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - CSI
        - Hedvig
        - Cloud
        - Block
        - NFS
        - iSCSI
        - Snapshot
        - Volume
        links:
        - name: Documentation
          url: https://documentation.commvault.com/commvault/hedvig/article
        - name: Hedvig CSI User Guide
          url: https://documentation.commvault.com/commvault/hedvig/others/pdf/Hedvig_CSI_User_Guide.pdf
        maintainers:
        - email: ashenoy@commvault.com
          name: Abhijith Shenoy
        maturity: stable
        provider:
          name: Commvault
        relatedImages:
        - hedvig/hedvig-operator:2.0.0-ubi
        version: 1.0.2
      entries:
      - name: hedvig-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: hedvig-operator
    provider:
      name: Commvault
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: External Secrets
      provider-url: https://external-secrets.io
    name: external-secrets-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: external-secrets-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": "1234"
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "key": "provider-key",
                      "property": "provider-key-property",
                      "version": "provider-key-version"
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterExternalSecret",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "externalSecretName": "hello-world-es",
                  "externalSecretSpec": {
                    "data": [
                      {
                        "remoteRef": {
                          "key": "provider-key",
                          "property": "provider-key-property",
                          "version": "provider-key-version"
                        },
                        "secretKey": "secret-key-to-be-managed"
                      }
                    ],
                    "dataFrom": [
                      {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    ],
                    "refreshInterval": "1h",
                    "secretStoreRef": {
                      "kind": "SecretStore",
                      "name": "secret-store-name"
                    },
                    "target": {
                      "creationPolicy": "Merge",
                      "name": "my-secret",
                      "template": {
                        "data": {
                          "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                        },
                        "metadata": {
                          "annotations": {},
                          "labels": {}
                        },
                        "templateFrom": [
                          {
                            "configMap": {
                              "items": [
                                {
                                  "key": "alertmanager.yaml"
                                }
                              ],
                              "name": "alertmanager"
                            }
                          }
                        ],
                        "type": "kubernetes.io/dockerconfigjson"
                      }
                    }
                  },
                  "namespaceSelector": {
                    "matchLabels": {
                      "cool": "label"
                    }
                  },
                  "refreshTime": "1m"
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": 1234
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "extract": {
                        "conversionStrategy": "Default",
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    },
                    {
                      "find": {
                        "conversionStrategy": "Unicode",
                        "name": {
                          "regexp": ".*foobar.*"
                        },
                        "path": "path-to-filter",
                        "tags": {
                          "foo": "bar"
                        }
                      }
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "deletionPolicy": "Retain",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.external-secrets.io/v1alpha1",
                "kind": "OperatorConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "requeueInterval": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  },
                  "concurrent": 1,
                  "controllerClass": "",
                  "crds": {
                    "createClusterExternalSecret": true,
                    "createClusterSecretStore": true
                  },
                  "createOperator": true,
                  "deploymentAnnotations": {},
                  "extraArgs": {},
                  "extraEnv": [],
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/external-secrets/external-secrets",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "installCRDs": false,
                  "leaderElect": false,
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "priorityClassName": "",
                  "processClusterExternalSecret": true,
                  "processClusterStore": true,
                  "prometheus": {
                    "enabled": false,
                    "service": {
                      "port": 8080
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "scopedNamespace": "",
                  "scopedRBAC": false,
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "certCheckInterval": "5m",
                    "certDir": "/tmp/certs",
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
          createdAt: "2024-12-03T10:44:19Z"
          description: Operator to configure external-secrets helm-chart based operator
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/external-secrets/external-secrets-helm-operator
          support: external-secrets
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ACRAccessToken
            name: acraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster scoped resource that allows you to manage ExternalSecrets
              across namespaces
            displayName: ClusterExternalSecret
            kind: ClusterExternalSecret
            name: clusterexternalsecrets.external-secrets.io
            version: v1beta1
          - kind: ClusterGenerator
            name: clustergenerators.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1alpha1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1beta1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1beta1
          - kind: ECRAuthorizationToken
            name: ecrauthorizationtokens.generators.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1alpha1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1beta1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1beta1
          - kind: Fake
            name: fakes.generators.external-secrets.io
            version: v1alpha1
          - kind: GCRAccessToken
            name: gcraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - kind: GithubAccessToken
            name: githubaccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Configures external-secrets helm-chart based operator
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.external-secrets.io
            version: v1alpha1
          - kind: Password
            name: passwords.generators.external-secrets.io
            version: v1alpha1
          - kind: PushSecret
            name: pushsecrets.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1alpha1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1beta1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1beta1
          - kind: STSSessionToken
            name: stssessiontokens.generators.external-secrets.io
            version: v1alpha1
          - kind: UUID
            name: uuids.generators.external-secrets.io
            version: v1alpha1
          - kind: VaultDynamicSecret
            name: vaultdynamicsecrets.generators.external-secrets.io
            version: v1alpha1
          - kind: Webhook
            name: webhooks.generators.external-secrets.io
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK (Helm version) to configure **[official external-secrets operator helm chart](https://github.com/external-secrets/external-secrets)**, so it can be installed via OLM without having to do any change on current Helm Charts.

          Before any other resources provided by this Operator can be deployed, it is essential to create an
          OperatorConfig resource.

          The usual Helm Chart file `values.yaml`, like:
          ```yaml
          prometheus:
            enabled: true
            service:
              port: 8080
          resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          need to be encapsulated into a new custom resource called `OperatorConfig`:
          ```yaml
          apiVersion: operator.external-secrets.io/v1alpha1
          kind: OperatorConfig
          metadata:
            name: cluster
          spec:
            prometheus:
              enabled: true
              service:
                port: 8080
            resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          Once the OperatorConfig resource is deployed, the operator will create all helm chart resources, using the custom resource name as a preffix for all resources names, like for example a `Deployment` called `cluster-external-secrets`.

          ## Documentation

          * [Upstream Documentation](https://external-secrets.io)
          * [Helm Operator Install](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/install.md)
          * [Helm Operator Development](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/development.md)
          * [Helm Operator Release](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/release.md)

          ## Contributing

          You can contribute by:

          * Raising any issues you find using External Secrets Operator
          * Fixing issues by opening [Pull Requests](https://github.com/external-secrets/external-secrets-helm-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving documentation
          * Talking about External Secrets Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/external-secrets/external-secrets-helm-operator/issues).

          ## License

          External Secrets Operator is under Apache 2.0 license. See the [LICENSE](LICENSE) file for details.
        displayName: External Secrets Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secrets
        - secrets-manager
        - vault
        - aws
        - azure
        - google
        - ibm
        - akeyless
        - yandex
        - gitlab
        - oracle
        links:
        - name: External Secrets Operator
          url: https://external-secrets.io
        - name: GitHub
          url: https://github.com/external-secrets/external-secrets
        - name: Slack
          url: https://kubernetes.slack.com/messages/external-secrets
        maintainers:
        - email: contact@external-secrets.io
          name: external-secrets community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: External Secrets
          url: https://external-secrets.io
        relatedImages:
        - ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
        version: 0.11.0
      entries:
      - name: external-secrets-operator.v0.11.0
        version: 0.11.0
      - name: external-secrets-operator.v0.10.7
        version: 0.10.7
      - name: external-secrets-operator.v0.10.6
        version: 0.10.6
      - name: external-secrets-operator.v0.10.5
        version: 0.10.5
      - name: external-secrets-operator.v0.10.4
        version: 0.10.4
      - name: external-secrets-operator.v0.10.3
        version: 0.10.3
      - name: external-secrets-operator.v0.10.2
        version: 0.10.2
      - name: external-secrets-operator.v0.10.1
        version: 0.10.1
      - name: external-secrets-operator.v0.10.0
        version: 0.10.0
      - name: external-secrets-operator.v0.9.20
        version: 0.9.20
      - name: external-secrets-operator.v0.9.19
        version: 0.9.19
      - name: external-secrets-operator.v0.9.18
        version: 0.9.18
      - name: external-secrets-operator.v0.9.17
        version: 0.9.17
      - name: external-secrets-operator.v0.9.16
        version: 0.9.16
      - name: external-secrets-operator.v0.9.15-2
        version: 0.9.15-2
      - name: external-secrets-operator.v0.9.14
        version: 0.9.14
      - name: external-secrets-operator.v0.9.13
        version: 0.9.13
      - name: external-secrets-operator.v0.9.12
        version: 0.9.12
      - name: external-secrets-operator.v0.9.11
        version: 0.9.11
      - name: external-secrets-operator.v0.9.10
        version: 0.9.10
      - name: external-secrets-operator.v0.9.9
        version: 0.9.9
      - name: external-secrets-operator.v0.9.8
        version: 0.9.8
      - name: external-secrets-operator.v0.9.7
        version: 0.9.7
      - name: external-secrets-operator.v0.9.6
        version: 0.9.6
      - name: external-secrets-operator.v0.9.5
        version: 0.9.5
      - name: external-secrets-operator.v0.9.4
        version: 0.9.4
      - name: external-secrets-operator.v0.9.3
        version: 0.9.3
      - name: external-secrets-operator.v0.9.2
        version: 0.9.2
      - name: external-secrets-operator.v0.9.1
        version: 0.9.1
      - name: external-secrets-operator.v0.9.0
        version: 0.9.0
      - name: external-secrets-operator.v0.8.7
        version: 0.8.7
      - name: external-secrets-operator.v0.8.6
        version: 0.8.6
      - name: external-secrets-operator.v0.8.5
        version: 0.8.5
      - name: external-secrets-operator.v0.8.4
        version: 0.8.4
      - name: external-secrets-operator.v0.8.3
        version: 0.8.3
      - name: external-secrets-operator.v0.8.2
        version: 0.8.2
      - name: external-secrets-operator.v0.8.1
        version: 0.8.1
      - name: external-secrets-operator.v0.8.0
        version: 0.8.0
      - name: external-secrets-operator.v0.7.2
        version: 0.7.2
      - name: external-secrets-operator.v0.7.1
        version: 0.7.1
      - name: external-secrets-operator.v0.7.0
        version: 0.7.0
      - name: external-secrets-operator.v0.7.0-rc1
        version: 0.7.0-rc1
      - name: external-secrets-operator.v0.6.1
        version: 0.6.1
      - name: external-secrets-operator.v0.6.0
        version: 0.6.0
      - name: external-secrets-operator.v0.6.0-rc1
        version: 0.6.0-rc1
      - name: external-secrets-operator.v0.5.9
        version: 0.5.9
      - name: external-secrets-operator.v0.5.8
        version: 0.5.8
      - name: external-secrets-operator.v0.5.7
        version: 0.5.7
      - name: external-secrets-operator.v0.5.6
        version: 0.5.6
      - name: external-secrets-operator.v0.5.5
        version: 0.5.5
      - name: external-secrets-operator.v0.5.4
        version: 0.5.4
      - name: external-secrets-operator.v0.5.3
        version: 0.5.3
      - name: external-secrets-operator.v0.5.2
        version: 0.5.2
      - name: external-secrets-operator.v0.5.1
        version: 0.5.1
      - name: external-secrets-operator.v0.5.0
        version: 0.5.0
      - name: external-secrets-operator.v0.4.4
        version: 0.4.4
      - name: external-secrets-operator.v0.4.3
        version: 0.4.3
      - name: external-secrets-operator.v0.4.2
        version: 0.4.2
      - name: external-secrets-operator.v0.4.1
        version: 0.4.1
      - name: external-secrets-operator.v0.3.11
        version: 0.3.11
      - name: external-secrets-operator.v0.3.10
        version: 0.3.10
      - name: external-secrets-operator.v0.3.9
        version: 0.3.9
      - name: external-secrets-operator.v0.3.8
        version: 0.3.8
      name: alpha
    - currentCSV: external-secrets-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": "1234"
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "key": "provider-key",
                      "property": "provider-key-property",
                      "version": "provider-key-version"
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterExternalSecret",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "externalSecretName": "hello-world-es",
                  "externalSecretSpec": {
                    "data": [
                      {
                        "remoteRef": {
                          "key": "provider-key",
                          "property": "provider-key-property",
                          "version": "provider-key-version"
                        },
                        "secretKey": "secret-key-to-be-managed"
                      }
                    ],
                    "dataFrom": [
                      {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    ],
                    "refreshInterval": "1h",
                    "secretStoreRef": {
                      "kind": "SecretStore",
                      "name": "secret-store-name"
                    },
                    "target": {
                      "creationPolicy": "Merge",
                      "name": "my-secret",
                      "template": {
                        "data": {
                          "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                        },
                        "metadata": {
                          "annotations": {},
                          "labels": {}
                        },
                        "templateFrom": [
                          {
                            "configMap": {
                              "items": [
                                {
                                  "key": "alertmanager.yaml"
                                }
                              ],
                              "name": "alertmanager"
                            }
                          }
                        ],
                        "type": "kubernetes.io/dockerconfigjson"
                      }
                    }
                  },
                  "namespaceSelector": {
                    "matchLabels": {
                      "cool": "label"
                    }
                  },
                  "refreshTime": "1m"
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": 1234
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "extract": {
                        "conversionStrategy": "Default",
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    },
                    {
                      "find": {
                        "conversionStrategy": "Unicode",
                        "name": {
                          "regexp": ".*foobar.*"
                        },
                        "path": "path-to-filter",
                        "tags": {
                          "foo": "bar"
                        }
                      }
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "deletionPolicy": "Retain",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.external-secrets.io/v1alpha1",
                "kind": "OperatorConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "requeueInterval": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  },
                  "concurrent": 1,
                  "controllerClass": "",
                  "crds": {
                    "createClusterExternalSecret": true,
                    "createClusterSecretStore": true
                  },
                  "createOperator": true,
                  "deploymentAnnotations": {},
                  "extraArgs": {},
                  "extraEnv": [],
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/external-secrets/external-secrets",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "installCRDs": false,
                  "leaderElect": false,
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "priorityClassName": "",
                  "processClusterExternalSecret": true,
                  "processClusterStore": true,
                  "prometheus": {
                    "enabled": false,
                    "service": {
                      "port": 8080
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "scopedNamespace": "",
                  "scopedRBAC": false,
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "certCheckInterval": "5m",
                    "certDir": "/tmp/certs",
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
          createdAt: "2024-12-03T10:44:19Z"
          description: Operator to configure external-secrets helm-chart based operator
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/external-secrets/external-secrets-helm-operator
          support: external-secrets
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ACRAccessToken
            name: acraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster scoped resource that allows you to manage ExternalSecrets
              across namespaces
            displayName: ClusterExternalSecret
            kind: ClusterExternalSecret
            name: clusterexternalsecrets.external-secrets.io
            version: v1beta1
          - kind: ClusterGenerator
            name: clustergenerators.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1alpha1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1beta1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1beta1
          - kind: ECRAuthorizationToken
            name: ecrauthorizationtokens.generators.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1alpha1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1beta1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1beta1
          - kind: Fake
            name: fakes.generators.external-secrets.io
            version: v1alpha1
          - kind: GCRAccessToken
            name: gcraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - kind: GithubAccessToken
            name: githubaccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Configures external-secrets helm-chart based operator
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.external-secrets.io
            version: v1alpha1
          - kind: Password
            name: passwords.generators.external-secrets.io
            version: v1alpha1
          - kind: PushSecret
            name: pushsecrets.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1alpha1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1beta1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1beta1
          - kind: STSSessionToken
            name: stssessiontokens.generators.external-secrets.io
            version: v1alpha1
          - kind: UUID
            name: uuids.generators.external-secrets.io
            version: v1alpha1
          - kind: VaultDynamicSecret
            name: vaultdynamicsecrets.generators.external-secrets.io
            version: v1alpha1
          - kind: Webhook
            name: webhooks.generators.external-secrets.io
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK (Helm version) to configure **[official external-secrets operator helm chart](https://github.com/external-secrets/external-secrets)**, so it can be installed via OLM without having to do any change on current Helm Charts.

          Before any other resources provided by this Operator can be deployed, it is essential to create an
          OperatorConfig resource.

          The usual Helm Chart file `values.yaml`, like:
          ```yaml
          prometheus:
            enabled: true
            service:
              port: 8080
          resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          need to be encapsulated into a new custom resource called `OperatorConfig`:
          ```yaml
          apiVersion: operator.external-secrets.io/v1alpha1
          kind: OperatorConfig
          metadata:
            name: cluster
          spec:
            prometheus:
              enabled: true
              service:
                port: 8080
            resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          Once the OperatorConfig resource is deployed, the operator will create all helm chart resources, using the custom resource name as a preffix for all resources names, like for example a `Deployment` called `cluster-external-secrets`.

          ## Documentation

          * [Upstream Documentation](https://external-secrets.io)
          * [Helm Operator Install](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/install.md)
          * [Helm Operator Development](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/development.md)
          * [Helm Operator Release](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/release.md)

          ## Contributing

          You can contribute by:

          * Raising any issues you find using External Secrets Operator
          * Fixing issues by opening [Pull Requests](https://github.com/external-secrets/external-secrets-helm-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving documentation
          * Talking about External Secrets Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/external-secrets/external-secrets-helm-operator/issues).

          ## License

          External Secrets Operator is under Apache 2.0 license. See the [LICENSE](LICENSE) file for details.
        displayName: External Secrets Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secrets
        - secrets-manager
        - vault
        - aws
        - azure
        - google
        - ibm
        - akeyless
        - yandex
        - gitlab
        - oracle
        links:
        - name: External Secrets Operator
          url: https://external-secrets.io
        - name: GitHub
          url: https://github.com/external-secrets/external-secrets
        - name: Slack
          url: https://kubernetes.slack.com/messages/external-secrets
        maintainers:
        - email: contact@external-secrets.io
          name: external-secrets community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: External Secrets
          url: https://external-secrets.io
        relatedImages:
        - ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
        version: 0.11.0
      entries:
      - name: external-secrets-operator.v0.11.0
        version: 0.11.0
      - name: external-secrets-operator.v0.10.7
        version: 0.10.7
      - name: external-secrets-operator.v0.10.6
        version: 0.10.6
      - name: external-secrets-operator.v0.10.5
        version: 0.10.5
      - name: external-secrets-operator.v0.10.4
        version: 0.10.4
      - name: external-secrets-operator.v0.10.3
        version: 0.10.3
      - name: external-secrets-operator.v0.10.2
        version: 0.10.2
      - name: external-secrets-operator.v0.10.1
        version: 0.10.1
      - name: external-secrets-operator.v0.10.0
        version: 0.10.0
      - name: external-secrets-operator.v0.9.20
        version: 0.9.20
      - name: external-secrets-operator.v0.9.19
        version: 0.9.19
      - name: external-secrets-operator.v0.9.18
        version: 0.9.18
      - name: external-secrets-operator.v0.9.17
        version: 0.9.17
      - name: external-secrets-operator.v0.9.16
        version: 0.9.16
      - name: external-secrets-operator.v0.9.15-2
        version: 0.9.15-2
      - name: external-secrets-operator.v0.9.14
        version: 0.9.14
      - name: external-secrets-operator.v0.9.13
        version: 0.9.13
      - name: external-secrets-operator.v0.9.12
        version: 0.9.12
      - name: external-secrets-operator.v0.9.11
        version: 0.9.11
      - name: external-secrets-operator.v0.9.10
        version: 0.9.10
      - name: external-secrets-operator.v0.9.9
        version: 0.9.9
      - name: external-secrets-operator.v0.9.8
        version: 0.9.8
      - name: external-secrets-operator.v0.9.7
        version: 0.9.7
      - name: external-secrets-operator.v0.9.6
        version: 0.9.6
      - name: external-secrets-operator.v0.9.5
        version: 0.9.5
      - name: external-secrets-operator.v0.9.4
        version: 0.9.4
      - name: external-secrets-operator.v0.9.3
        version: 0.9.3
      - name: external-secrets-operator.v0.9.2
        version: 0.9.2
      - name: external-secrets-operator.v0.9.1
        version: 0.9.1
      - name: external-secrets-operator.v0.9.0
        version: 0.9.0
      - name: external-secrets-operator.v0.8.7
        version: 0.8.7
      - name: external-secrets-operator.v0.8.6
        version: 0.8.6
      - name: external-secrets-operator.v0.8.5
        version: 0.8.5
      - name: external-secrets-operator.v0.8.4
        version: 0.8.4
      - name: external-secrets-operator.v0.8.3
        version: 0.8.3
      - name: external-secrets-operator.v0.8.2
        version: 0.8.2
      - name: external-secrets-operator.v0.8.1
        version: 0.8.1
      - name: external-secrets-operator.v0.8.0
        version: 0.8.0
      - name: external-secrets-operator.v0.7.2
        version: 0.7.2
      - name: external-secrets-operator.v0.7.1
        version: 0.7.1
      - name: external-secrets-operator.v0.7.0
        version: 0.7.0
      - name: external-secrets-operator.v0.7.0-rc1
        version: 0.7.0-rc1
      - name: external-secrets-operator.v0.6.1
        version: 0.6.1
      - name: external-secrets-operator.v0.6.0
        version: 0.6.0
      - name: external-secrets-operator.v0.5.9
        version: 0.5.9
      - name: external-secrets-operator.v0.5.8
        version: 0.5.8
      - name: external-secrets-operator.v0.5.7
        version: 0.5.7
      - name: external-secrets-operator.v0.5.6
        version: 0.5.6
      - name: external-secrets-operator.v0.5.5
        version: 0.5.5
      - name: external-secrets-operator.v0.5.4
        version: 0.5.4
      - name: external-secrets-operator.v0.5.3
        version: 0.5.3
      - name: external-secrets-operator.v0.5.2
        version: 0.5.2
      - name: external-secrets-operator.v0.5.1
        version: 0.5.1
      - name: external-secrets-operator.v0.5.0
        version: 0.5.0
      - name: external-secrets-operator.v0.4.4
        version: 0.4.4
      - name: external-secrets-operator.v0.4.2
        version: 0.4.2
      - name: external-secrets-operator.v0.4.1
        version: 0.4.1
      - name: external-secrets-operator.v0.3.11
        version: 0.3.11
      - name: external-secrets-operator.v0.3.10
        version: 0.3.10
      - name: external-secrets-operator.v0.3.9
        version: 0.3.9
      name: stable
    defaultChannel: alpha
    packageName: external-secrets-operator
    provider:
      name: External Secrets
      url: https://external-secrets.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/kuadrant-operator
    name: rhcl-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhcl-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "authpolicy-sample"
                },
                "spec": {
                  "rules": {
                    "authentication": {
                      "apikey": {
                        "apiKey": {
                          "selector": {}
                        },
                        "credentials": {
                          "authorizationHeader": {
                            "prefix": "APIKEY"
                          }
                        }
                      }
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "DNSPolicy",
                "metadata": {
                  "name": "dnspolicy-sample"
                },
                "spec": {
                  "healthCheck": {
                    "protocol": "HTTP"
                  },
                  "providerRefs": [
                    {
                      "name": "provider-ref"
                    }
                  ],
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "ratelimitpolicy-sample"
                },
                "spec": {
                  "limits": {
                    "toys": {
                      "rates": [
                        {
                          "limit": 50,
                          "window": "1m"
                        }
                      ]
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "TLSPolicy",
                "metadata": {
                  "name": "tlspolicy-sample"
                },
                "spec": {
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "self-signed-ca"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/rhcl-rhel9-operator@sha256:92f14fd8de74bc72c407947cf5fa4739a7b455e5d351baeb64521e4c07f02a9e
          createdAt: "2025-04-28T13:10:44Z"
          description: Red Hat Connectivity Link enables you to secure, protect, and
            connect your APIs and applications in multicluster, multicloud, and hybrid
            cloud environments
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthPolicy enables authentication and authorization for service
              workloads in a Gateway API network
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1
          - description: DNSPolicy configures how North-South based traffic should
              be balanced and reach the gateways
            displayName: DNSPolicy
            kind: DNSPolicy
            name: dnspolicies.kuadrant.io
            version: v1
          - description: Kuadrant configures installations of Kuadrant Service Protection
              components
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: RateLimitPolicy enables rate limiting for service workloads
              in a Gateway API network
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1
          - description: TLSPolicy provides tls for gateway listeners by managing
              the lifecycle of tls certificates
            displayName: TLSPolicy
            kind: TLSPolicy
            name: tlspolicies.kuadrant.io
            version: v1
        description: Red Hat Connectivity Link enables you to secure, protect, and
          connect your APIs and applications in multicluster, multicloud, and hybrid
          cloud environments
        displayName: Red Hat Connectivity Link
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: GitHub
          url: https://github.com/kuadrant/kuadrant-operator
        - name: Documentation
          url: https://github.com/Kuadrant/kuadrant-operator/tree/main/docs
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - registry.redhat.io/rhcl-1/rhcl-console-plugin-rhel9@sha256:ae92d638d993d782986e761d5aa3c7c5da2b8af0aabc1decb332a4f1487dacd3
        - registry.redhat.io/rhcl-1/rhcl-rhel9-operator@sha256:92f14fd8de74bc72c407947cf5fa4739a7b455e5d351baeb64521e4c07f02a9e
        - registry.redhat.io/rhcl-1/wasm-shim-rhel9@sha256:845bb8af57f3d219aa09b9c0bb20fa945306f3b413e34e196760c8cdf624a532
        version: 1.1.0
      entries:
      - name: rhcl-operator.v1.1.0
        version: 1.1.0
      - name: rhcl-operator.v1.0.2
        version: 1.0.2
      - name: rhcl-operator.v1.0.1
        version: 1.0.1
      - name: rhcl-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: rhcl-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/kuadrant-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sqs-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sqs-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sqs.services.k8s.aws/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sqs-controller:1.2.1
          createdAt: "2025-09-19T19:14:25Z"
          description: AWS SQS controller is a service controller for managing SQS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Queue represents the state of an AWS sqs Queue resource.
            displayName: Queue
            kind: Queue
            name: queues.sqs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Simple Queue Service (SQS) resources in AWS from within your Kubernetes cluster.

          **About Amazon SQS**

          Amazon Simple Queue Service (Amazon SQS) offers a secure, durable, and available hosted queue that lets you integrate and decouple distributed software systems and components. Amazon SQS offers common constructs such as [dead-letter queues](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGuide/sqs-dead-letter-queues.html) and [cost allocation tags](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGuide/sqs-queue-tags.html). It provides a generic web services API that you can access using any programming language that the AWS SDK supports.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SQS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sqs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SQS Developer Resources
          url: https://aws.amazon.com/sqs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sqs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sqs-controller:1.2.1
        version: 1.2.1
      entries:
      - name: ack-sqs-controller.v1.2.1
        version: 1.2.1
      - name: ack-sqs-controller.v1.2.0
        version: 1.2.0
      - name: ack-sqs-controller.v1.1.16
        version: 1.1.16
      - name: ack-sqs-controller.v1.1.14
        version: 1.1.14
      - name: ack-sqs-controller.v1.1.13
        version: 1.1.13
      - name: ack-sqs-controller.v1.1.12
        version: 1.1.12
      - name: ack-sqs-controller.v1.1.11
        version: 1.1.11
      - name: ack-sqs-controller.v1.1.10
        version: 1.1.10
      - name: ack-sqs-controller.v1.1.9
        version: 1.1.9
      - name: ack-sqs-controller.v1.1.8
        version: 1.1.8
      - name: ack-sqs-controller.v1.1.7
        version: 1.1.7
      - name: ack-sqs-controller.v1.1.6
        version: 1.1.6
      - name: ack-sqs-controller.v1.1.5
        version: 1.1.5
      - name: ack-sqs-controller.v1.1.4
        version: 1.1.4
      - name: ack-sqs-controller.v1.1.3
        version: 1.1.3
      - name: ack-sqs-controller.v1.1.2
        version: 1.1.2
      - name: ack-sqs-controller.v1.1.1
        version: 1.1.1
      - name: ack-sqs-controller.v1.1.0
        version: 1.1.0
      - name: ack-sqs-controller.v1.0.17
        version: 1.0.17
      - name: ack-sqs-controller.v1.0.16
        version: 1.0.16
      - name: ack-sqs-controller.v1.0.15
        version: 1.0.15
      - name: ack-sqs-controller.v1.0.14
        version: 1.0.14
      - name: ack-sqs-controller.v1.0.13
        version: 1.0.13
      - name: ack-sqs-controller.v1.0.12
        version: 1.0.12
      - name: ack-sqs-controller.v1.0.10
        version: 1.0.10
      - name: ack-sqs-controller.v1.0.7
        version: 1.0.7
      - name: ack-sqs-controller.v1.0.5
        version: 1.0.5
      - name: ack-sqs-controller.v1.0.3
        version: 1.0.3
      - name: ack-sqs-controller.v1.0.2
        version: 1.0.2
      - name: ack-sqs-controller.v1.0.1
        version: 1.0.1
      - name: ack-sqs-controller.v1.0.0
        version: 1.0.0
      - name: ack-sqs-controller.v0.0.7
        version: 0.0.7
      - name: ack-sqs-controller.v0.0.6
        version: 0.0.6
      - name: ack-sqs-controller.v0.0.5
        version: 0.0.5
      - name: ack-sqs-controller.v0.0.4
        version: 0.0.4
      - name: ack-sqs-controller.v0.0.2
        version: 0.0.2
      - name: ack-sqs-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-sqs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-rds-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-rds-controller.v1.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rds.services.k8s.aws/v1alpha1",
                "kind": "DBParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "rds.services.k8s.aws/v1alpha1",
                "kind": "DBSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/rds-controller:1.6.2
          createdAt: "2025-09-22T17:27:42Z"
          description: AWS RDS controller is a service controller for managing RDS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DBClusterEndpoint represents the state of an AWS rds DBClusterEndpoint
              resource.
            displayName: DBClusterEndpoint
            kind: DBClusterEndpoint
            name: dbclusterendpoints.rds.services.k8s.aws
            version: v1alpha1
          - description: DBClusterParameterGroup represents the state of an AWS rds
              DBClusterParameterGroup resource.
            displayName: DBClusterParameterGroup
            kind: DBClusterParameterGroup
            name: dbclusterparametergroups.rds.services.k8s.aws
            version: v1alpha1
          - description: DBCluster represents the state of an AWS rds DBCluster resource.
            displayName: DBCluster
            kind: DBCluster
            name: dbclusters.rds.services.k8s.aws
            version: v1alpha1
          - description: DBClusterSnapshot represents the state of an AWS rds DBClusterSnapshot
              resource.
            displayName: DBClusterSnapshot
            kind: DBClusterSnapshot
            name: dbclustersnapshots.rds.services.k8s.aws
            version: v1alpha1
          - description: DBInstance represents the state of an AWS rds DBInstance
              resource.
            displayName: DBInstance
            kind: DBInstance
            name: dbinstances.rds.services.k8s.aws
            version: v1alpha1
          - description: DBParameterGroup represents the state of an AWS rds DBParameterGroup
              resource.
            displayName: DBParameterGroup
            kind: DBParameterGroup
            name: dbparametergroups.rds.services.k8s.aws
            version: v1alpha1
          - description: DBProxy represents the state of an AWS rds DBProxy resource.
            displayName: DBProxy
            kind: DBProxy
            name: dbproxies.rds.services.k8s.aws
            version: v1alpha1
          - description: DBSnapshot represents the state of an AWS rds DBSnapshot
              resource.
            displayName: DBSnapshot
            kind: DBSnapshot
            name: dbsnapshots.rds.services.k8s.aws
            version: v1alpha1
          - description: DBSubnetGroup represents the state of an AWS rds DBSubnetGroup
              resource.
            displayName: DBSubnetGroup
            kind: DBSubnetGroup
            name: dbsubnetgroups.rds.services.k8s.aws
            version: v1alpha1
          - description: GlobalCluster represents the state of an AWS rds GlobalCluster
              resource.
            displayName: GlobalCluster
            kind: GlobalCluster
            name: globalclusters.rds.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Relational Database Service ("RDS") resources in AWS from within your Kubernetes cluster.

          **About Amazon RDS**

          Amazon Relational Database Service (Amazon RDS) makes it easy to set up, operate, and scale a relational database in the cloud. It provides cost-efficient and resizable capacity while automating time-consuming administration tasks such as hardware provisioning, database setup, patching and backups. It frees you to focus on your applications so you can give them the fast performance, high availability, security and compatibility they need.

          Amazon RDS is available on several database instance types - optimized for memory, performance or I/O -and provides you with six familiar database engines to choose from, including Amazon Aurora, PostgreSQL, MySQL, MariaDB, Oracle Database, and SQL Server. You can use the AWS Database Migration Service to easily migrate or replicate your existing databases to Amazon RDS.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon RDS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rds
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon RDS Developer Resources
          url: https://aws.amazon.com/rds/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: rds maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/rds-controller:1.6.2
        version: 1.6.2
      entries:
      - name: ack-rds-controller.v1.6.2
        version: 1.6.2
      - name: ack-rds-controller.v1.6.1
        version: 1.6.1
      - name: ack-rds-controller.v1.6.0
        version: 1.6.0
      - name: ack-rds-controller.v1.5.0
        version: 1.5.0
      - name: ack-rds-controller.v1.4.26
        version: 1.4.26
      - name: ack-rds-controller.v1.4.25
        version: 1.4.25
      - name: ack-rds-controller.v1.4.24
        version: 1.4.24
      - name: ack-rds-controller.v1.4.22
        version: 1.4.22
      - name: ack-rds-controller.v1.4.21
        version: 1.4.21
      - name: ack-rds-controller.v1.4.20
        version: 1.4.20
      - name: ack-rds-controller.v1.4.19
        version: 1.4.19
      - name: ack-rds-controller.v1.4.18
        version: 1.4.18
      - name: ack-rds-controller.v1.4.17
        version: 1.4.17
      - name: ack-rds-controller.v1.4.16
        version: 1.4.16
      - name: ack-rds-controller.v1.4.15
        version: 1.4.15
      - name: ack-rds-controller.v1.4.14
        version: 1.4.14
      - name: ack-rds-controller.v1.4.13
        version: 1.4.13
      - name: ack-rds-controller.v1.4.12
        version: 1.4.12
      - name: ack-rds-controller.v1.4.11
        version: 1.4.11
      - name: ack-rds-controller.v1.4.10
        version: 1.4.10
      - name: ack-rds-controller.v1.4.8
        version: 1.4.8
      - name: ack-rds-controller.v1.4.7
        version: 1.4.7
      - name: ack-rds-controller.v1.4.6
        version: 1.4.6
      - name: ack-rds-controller.v1.4.4
        version: 1.4.4
      - name: ack-rds-controller.v1.4.3
        version: 1.4.3
      - name: ack-rds-controller.v1.4.2
        version: 1.4.2
      - name: ack-rds-controller.v1.4.1
        version: 1.4.1
      - name: ack-rds-controller.v1.4.0
        version: 1.4.0
      - name: ack-rds-controller.v1.3.0
        version: 1.3.0
      - name: ack-rds-controller.v1.2.5
        version: 1.2.5
      - name: ack-rds-controller.v1.2.4
        version: 1.2.4
      - name: ack-rds-controller.v1.2.3
        version: 1.2.3
      - name: ack-rds-controller.v1.2.2
        version: 1.2.2
      - name: ack-rds-controller.v1.2.1
        version: 1.2.1
      - name: ack-rds-controller.v1.2.0
        version: 1.2.0
      - name: ack-rds-controller.v1.1.11
        version: 1.1.11
      - name: ack-rds-controller.v1.1.9
        version: 1.1.9
      - name: ack-rds-controller.v1.1.8
        version: 1.1.8
      - name: ack-rds-controller.v1.1.7
        version: 1.1.7
      - name: ack-rds-controller.v1.1.6
        version: 1.1.6
      - name: ack-rds-controller.v1.1.5
        version: 1.1.5
      - name: ack-rds-controller.v1.1.4
        version: 1.1.4
      - name: ack-rds-controller.v1.1.3
        version: 1.1.3
      - name: ack-rds-controller.v1.1.2
        version: 1.1.2
      - name: ack-rds-controller.v1.1.1
        version: 1.1.1
      - name: ack-rds-controller.v1.1.0
        version: 1.1.0
      - name: ack-rds-controller.v1.0.0
        version: 1.0.0
      - name: ack-rds-controller.v0.1.3
        version: 0.1.3
      - name: ack-rds-controller.v0.1.1
        version: 0.1.1
      - name: ack-rds-controller.v0.1.0
        version: 0.1.0
      - name: ack-rds-controller.v0.0.30
        version: 0.0.30
      - name: ack-rds-controller.v0.0.29
        version: 0.0.29
      - name: ack-rds-controller.v0.0.28
        version: 0.0.28
      - name: ack-rds-controller.v0.0.27
        version: 0.0.27
      - name: ack-rds-controller.v0.0.26
        version: 0.0.26
      - name: ack-rds-controller.v0.0.25
        version: 0.0.25
      - name: ack-rds-controller.v0.0.24
        version: 0.0.24
      - name: ack-rds-controller.v0.0.23
        version: 0.0.23
      - name: ack-rds-controller.v0.0.22
        version: 0.0.22
      - name: ack-rds-controller.v0.0.21
        version: 0.0.21
      - name: ack-rds-controller.v0.0.20
        version: 0.0.20
      - name: ack-rds-controller.v0.0.19
        version: 0.0.19
      - name: ack-rds-controller.v0.0.18
        version: 0.0.18
      - name: ack-rds-controller.v0.0.17
        version: 0.0.17
      - name: ack-rds-controller.v0.0.16
        version: 0.0.16
      - name: ack-rds-controller.v0.0.15
        version: 0.0.15
      - name: ack-rds-controller.v0.0.14
        version: 0.0.14
      - name: ack-rds-controller.v0.0.13
        version: 0.0.13
      - name: ack-rds-controller.v0.0.12
        version: 0.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-rds-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: integrity-shield-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: integrity-shield-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/open-cluster-management/integrity-shield-observer",
                    "imagePullPolicy": "Always",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "provenanece": true,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  resp.status_code == 200\n  result := json.unmarshal(resp.raw_body)\n  result.allow == false\n  not is_detect_mode\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"tls_insecure_skip_verify\": true\n  })\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "requestHandlerConfig": "defaultConstraintAction:\n  audit:\n    inform: true\n  admissionControl:\n    enforce: false\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/open-cluster-management/integrity-shield-api",
                    "imagePullPolicy": "Always",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/open-cluster-management/integrity-shield-operator:0.3.0
          createdAt: "2021-10-14T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.

          [K8s Integrity Shield](https://github.com/open-cluster-management/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.

          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Detect (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster

          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).

          ## Preparations before installation

          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).


          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.

          If you want to change the settings such as default run mode (detection/enforcement) or audit interval,  please check [here](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).

          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-7b7f768bf7-fhrpg                           1/1     Running   0          20s
          integrity-shield-observer-5bc66f75f7-tn8fw                      1/1     Running   0          25s
          integrity-shield-operator-controller-manager-65b7fb58f7-j25zd   2/2     Running   0          3h5m
          ```

          After installation, you can protect Kubernetes resources by following this [document](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).

          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:

          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)

          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/open-cluster-management/integrity-shield
        - name: Operator Documentation
          url: https://github.com/open-cluster-management/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/open-cluster-management/integrity-shield-operator:0.3.0
        version: 0.3.0
      entries:
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.0
    - currentCSV: integrity-shield-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/open-cluster-management/integrity-shield-observer",
                    "imagePullPolicy": "Always",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  resp.status_code == 200\n  result := json.unmarshal(resp.raw_body)\n  result.allow == false\n  not is_detect_mode\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"tls_insecure_skip_verify\": true\n  })\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: detect\n  admissionOnly: false\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/open-cluster-management/integrity-shield-api",
                    "imagePullPolicy": "Always",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/open-cluster-management/integrity-shield-operator:0.3.1
          createdAt: "2021-10-22T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.

          [K8s Integrity Shield](https://github.com/open-cluster-management/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.

          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Detect (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster

          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).

          ## Preparations before installation

          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).


          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.

          If you want to change the settings such as default run mode (detection/enforcement) or audit interval,  please check [here](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).

          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-7b7f768bf7-fhrpg                           1/1     Running   0          20s
          integrity-shield-observer-5bc66f75f7-tn8fw                      1/1     Running   0          25s
          integrity-shield-operator-controller-manager-65b7fb58f7-j25zd   2/2     Running   0          3h5m
          ```

          After installation, you can protect Kubernetes resources by following this [document](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).

          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:

          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)

          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/open-cluster-management/integrity-shield
        - name: Operator Documentation
          url: https://github.com/open-cluster-management/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/open-cluster-management/integrity-shield-operator:0.3.1
        version: 0.3.1
      entries:
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.1
    - currentCSV: integrity-shield-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/stolostron/integrity-shield-observer",
                    "imagePullPolicy": "IfNotPresent",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  should_be_denied(resp)\n  not is_detect_mode\n  result := get_result(resp)\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"timeout\": \"2500ms\",\n    \"raise_error\": false,\n    \"tls_ca_cert\": REPLACE_WITH_CA_CERT\n  })\n}\n\n# get decision result from API response\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  not is_null(raw_body)\n  result := json.unmarshal(raw_body)\n}\n# if error, return it as result\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  is_null(raw_body)\n  error_msg := object.get(resp, \"error\", \"\")\n  result := {\"http_error\":error_msg}\n}\n\n# if API response is null, deny\nshould_be_denied(resp) {\n  is_null(resp)\n}\n\n# if API response has status_code other than 200, deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  resp.status_code != 200\n}\n\n# if decision result is \"allow == false\" or not set, then deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  result := get_result(resp)\n  allow := object.get(result, \"allow\", false)\n  allow == false\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "reporter": {
                    "image": "quay.io/stolostron/integrity-shield-reporter",
                    "imagePullPolicy": "IfNotPresent",
                    "intervalSeconds": "10",
                    "name": "integrity-shield-reporter",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: inform\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\ndecisionReporterConfig:\n  enabled: true\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/stolostron/integrity-shield-api",
                    "imagePullPolicy": "IfNotPresent",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/stolostron/integrity-shield-operator:0.3.2
          createdAt: "2022-01-25T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.15.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.
          [K8s Integrity Shield](https://github.com/stolostron/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.
          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Inform (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster
          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).
          ## Preparations before installation
          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).
          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.
          If you want to change the settings such as default run mode (inform/enforce) or audit interval,  please check [here](https://github.com/stolostron/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).
          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-5799f69896-zbcjs                           2/2     Running   0          88s
          integrity-shield-observer-757f58f7d7-7pf5h                      1/1     Running   0          55s
          integrity-shield-operator-controller-manager-6f94d7d684-dg5bt   2/2     Running   0          4m27s
          ```
          After installation, you can protect Kubernetes resources by following this [document](https://github.com/stolostron/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).
          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:
          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)
          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/stolostron/integrity-shield
        - name: Operator Documentation
          url: https://github.com/stolostron/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/stolostron/integrity-shield-operator:0.3.2
        version: 0.3.2
      entries:
      - name: integrity-shield-operator.v0.3.2
        version: 0.3.2
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.2
    - currentCSV: integrity-shield-operator.v0.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": false,
                    "image": "quay.io/stolostron/integrity-shield-observer",
                    "imagePullPolicy": "IfNotPresent",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  should_be_denied(resp)\n  not is_detect_mode\n  result := get_result(resp)\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"timeout\": \"2500ms\",\n    \"raise_error\": false,\n    \"tls_ca_cert\": REPLACE_WITH_CA_CERT\n  })\n}\n\n# get decision result from API response\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  not is_null(raw_body)\n  result := json.unmarshal(raw_body)\n}\n# if error, return it as result\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  is_null(raw_body)\n  error_msg := object.get(resp, \"error\", \"\")\n  result := {\"http_error\":error_msg}\n}\n\n# if API response is null, deny\nshould_be_denied(resp) {\n  is_null(resp)\n}\n\n# if API response has status_code other than 200, deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  resp.status_code != 200\n}\n\n# if decision result is \"allow == false\" or not set, then deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  result := get_result(resp)\n  allow := object.get(result, \"allow\", false)\n  allow == false\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "reporter": {
                    "image": "quay.io/stolostron/integrity-shield-reporter",
                    "imagePullPolicy": "IfNotPresent",
                    "intervalSeconds": "10",
                    "name": "integrity-shield-reporter",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: inform\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\ndecisionReporterConfig:\n  enabled: true\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/stolostron/integrity-shield-api",
                    "imagePullPolicy": "IfNotPresent",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/stolostron/integrity-shield-operator:0.3.3
          createdAt: "2022-03-18T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.15.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.
          [K8s Integrity Shield](https://github.com/stolostron/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.
          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Inform (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster
          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).
          ## Preparations before installation
          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).
          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.
          If you want to change the settings such as default run mode (inform/enforce) or audit interval,  please check [here](https://github.com/stolostron/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).
          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-5799f69896-zbcjs                           2/2     Running   0          88s
          integrity-shield-observer-757f58f7d7-7pf5h                      1/1     Running   0          55s
          integrity-shield-operator-controller-manager-6f94d7d684-dg5bt   2/2     Running   0          4m27s
          ```
          After installation, you can protect Kubernetes resources by following this [document](https://github.com/stolostron/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).
          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:
          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)
          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/stolostron/integrity-shield
        - name: Operator Documentation
          url: https://github.com/stolostron/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/stolostron/integrity-shield-operator:0.3.3
        version: 0.3.3
      entries:
      - name: integrity-shield-operator.v0.3.3
        version: 0.3.3
      - name: integrity-shield-operator.v0.3.2
        version: 0.3.2
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.3
    defaultChannel: alpha-0.3.3
    packageName: integrity-shield-operator
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: akaris
      provider-url: https://github.com/andreaskaris/sosreport-operator
    name: sosreport-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sosreport-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "support.openshift.io/v1alpha1",
                "kind": "Sosreport",
                "metadata": {
                  "name": "sosreport-sample"
                },
                "spec": {
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Sosreport is the Schema for the sosreports API
            displayName: Sosreport
            kind: Sosreport
            name: sosreports.support.openshift.io
            version: v1alpha1
        description: Operator that facilitates the generation of sosreports
        displayName: sosreport-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sosreport
        - debugging
        - troubleshooting
        links:
        - name: Sosreport Operator
          url: https://github.com/andreaskaris/sosreport-operator
        maintainers:
        - email: ak.karis@gmail.com
          name: Andreas Karis
        maturity: alpha
        provider:
          name: akaris
          url: https://github.com/andreaskaris/sosreport-operator
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/akaris/sosreport-operator:0.0.2
        version: 0.0.2
      entries:
      - name: sosreport-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: sosreport-operator
    provider:
      name: akaris
      url: https://github.com/andreaskaris/sosreport-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: awss3-operator-registry
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: awss3operator.1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/screeley44/aws-s3-provisioner:v1.0.0
          createdAt: "2019-05-02 16:12:00"
          description: Manage the full lifecycle of installing, configuring and managing
            AWS S3 Provisioner.
          repository: https://github.com/yard-turkey/aws-s3-provisioner
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - description: instance of an AWS S3 Bucket
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
          - description: Request for an AWS S3 Bucket
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
        description: |
          AWS S3 Operator will deploy the AWS S3 Provisioner which will dynamically or statically
          provision AWS S3 Bucket storage and access. The operator deploys the ObjectBucket (OB) and ObjectBucketClaim (OBC)
          CustomResourceDefinitions. The OB/OBC model follows the traditional Kubernetes PV/PVC pattern, when an OBC is detected
          the operator will act on the OBC to either provision a brand new S3 Bucket in AWS or gain access to an existing
          S3 Bucket in AWS. The operator produces an OB and ConfigMap and Secret which can then be consumed by application pods.

          **Important Note**: Currently, while in preview, this operator does not
          support automatic upgrades. You must remove the old version of the operator
          manually before installing a new version.
          ## Using AWS S3 Operator

          ### Administrator Creates Secret
          This secret will contain the elevated/admin privileges needed by the provisioner
          to properly access and create S3 Buckets and IAM users and policies. The AWS Access ID
          and AWS Secret Key will be needed for this.
          1. Create the Kubernetes Secret for the Provisioner's Owner Access.
          ```yaml
          apiVersion: v1
          kind: Secret
          metadata:
            name: s3-bucket-owner [1]
            namespace: s3-provisioner [2]
          type: Opaque
          data:
            AWS_ACCESS_KEY_ID: *base64 encoded value* [3]
            AWS_SECRET_ACCESS_KEY: *base64 encoded value* [4]
          ```
          1. Name of the secret, this will be referenced in StorageClass.
          1. Namespace where the Secret will exist.
          1. Your AWS_ACCESS_KEY_ID base64 encoded.
          1. Your AWS_SECRET_ACCESS_KEY base64 encoded.
          ```
           # kubectl create -f creds.yaml
          secret/s3-bucket-owner created
          ```
          ### Administrator Creates StorageClass
          The StorageClass defines the name of the provisioner and holds other properties that are needed to provision a new bucket, including
          the Owner Secret and Namespace, and the AWS Region.
          #### Greenfield Example:
          For Greenfield, a new, dynamic bucket will be generated.
          1. Create the Kubernetes StorageClass for the Provisioner.
          ```yaml
          kind: StorageClass
          apiVersion: storage.k8s.io/v1
          metadata:
            name: s3-buckets [1]
          provisioner: aws-s3.io/bucket [2]
          parameters:
            region: us-west-1 [3]
            secretName: s3-bucket-owner [4]
            secretNamespace: s3-provisioner [5]
          reclaimPolicy: Delete [6]
          ```
          1. Name of the StorageClass, this will be referenced in the User ObjectBucketClaim.
          1. Provisioner name
          1. AWS Region that the StorageClass will serve
          1. Name of the bucket owner Secret created above
          1. Namespace where the Secret will exist
          1. reclaimPolicy (Delete or Retain) indicates if the bucket can be deleted when the OBC is deleted.
          **NOTE:** the absence of the `bucketName` Parameter key in the storage class indicates this is a new bucket and its name is based on the bucket name fields in the OBC.
          ```
           # kubectl create -f storageclass-greenfield.yaml
          storageclass.storage.k8s.io/s3-buckets created
          ```
          #### Brownfield Example:
          For brownfield, the StorageClass defines the name of the provisioner and the name of the existing bucket. It also includes other properties needed by the target
          provisioner, including: the Owner Secret and Namespace, and the AWS Region
          1. Create the Kubernetes StorageClass for the Provisioner.
          ```yaml
          kind: StorageClass
          apiVersion: storage.k8s.io/v1
          metadata:
            name: s3-existing-buckets [1]
          provisioner: aws-s3.io/bucket [2]
          parameters:
            bucketName: my-existing-bucket [3]
            region: us-west-1 [4]
            secretName: s3-bucket-owner [5]
            secretNamespace: s3-provisioner [6]
          ```
          1. Name of the StorageClass, this will be referenced in the User ObjectBucketClaim.
          1. Provisioner name
          1. Name of the existing bucket
          1. AWS Region that the StorageClass will serve
          1. Name of the bucket owner Secret created above
          1. Namespace for that bucket owner secret
          **NOTE:** the storage class's `reclaimPolicy` is ignored for existing buckets.
          ```
           # kubectl create -f storageclass-brownfield.yaml
          storageclass.storage.k8s.io/s3-buckets created
          ```
          ### User Creates ObjectBucketClaim
          An ObjectBucketClaim follows the same concept as a PVC, in that
          it is a request for Object Storage, the user doesn't need to
          concern him/herself with the underlying storage, just that
          they need access to it. The user will work with the cluster/storage
          administrator to get the proper StorageClass needed and will
          then request access via the OBC.
          #### Greenfield Request Example:
          1. Create the ObjectBucketClaim.
          ```yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucketClaim
          metadata:
            name: myobc [1]
            namespace: s3-provisioner [2]
          spec:
            generateBucketName: mybucket [3]
            bucketName: my-awesome-bucket [4]
            storageClassName: s3-buckets [5]
          ```
          1. Name of the OBC
          1. Namespace of the OBC
          1. Name prepended to a random string used to generate a bucket name. It is ignored if bucketName is defined
          1. Name of new bucket which must be unique across all AWS regions, otherwise an error occurs when creating the bucket. If present, this name overrides `generateName`
          1. StorageClass name
          **NOTE:** if both `generateBucketName` and `bucketName` are omitted, and the storage class does _not_ define a bucket name, then a new, random bucket name is generated with no prefix.
          ```
           # kubectl create -f obc-brownfield.yaml
          objectbucketclaim.objectbucket.io/myobc created
          ```
          #### Brownfield Request Example:
          1. Create the ObjectBucketClaim.
          ```yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucketClaim
          metadata:
            name: myobc [1]
            namespace: s3-provisioner [2]
          spec:
            storageClassName: s3-existing-buckets [3]
          ```
          1. Name of the OBC
          1. Namespace of the OBC
          1. StorageClass name
          **NOTE:** in the OBC here there is no reference to the bucket's name. This is defined in the storage class and is not a concern of the user creating the claim to this bucket.  An OBC does have fields for defining a bucket name for greenfield use only.
          ```
           # kubectl create -f obc-brownfield.yaml
          objectbucketclaim.objectbucket.io/myobc created
          ```
          ### Results and Recap
          Let's pause for a moment and digest what just happened.
          After creating the OBC, and assuming the S3 provisioner is running, we now have
          the following Kubernetes resources:
          .  a global ObjectBucket (OB) which contains: bucket endpoint info (including region and bucket name), a reference to the OBC, and a reference to the storage class. Unique to S3, the OB also contains the bucket Amazon Resource Name (ARN).Note: there is always a 1:1 relationship between an OBC and an OB.
          .  a ConfigMap in the same namespace as the OBC, which contains the same endpoint data found in the OB.
          .  a Secret in the same namespace as the OBC, which contains the AWS key-pairs needed to access the bucket.
          And of course, we have a *new* AWS S3 Bucket which you should be able to see via the AWS Console.
          *ObjectBucket*
          ```yaml
           # kubectl get ob obc-s3-provisioner-my-awesome-bucket -o yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucket
          metadata:
            creationTimestamp: "2019-04-03T15:42:22Z"
            generation: 1
            name: obc-s3-provisioner-my-awesome-bucket
            resourceVersion: "15057"
            selfLink: /apis/objectbucket.io/v1alpha1/objectbuckets/obc-s3-provisioner-my-awesome-bucket
            uid: 0bfe8e84-576d-4c4e-984b-f73c4460f736
          spec:
            Connection:
              additionalState:
                ARN: arn:aws:iam::<accountid>:policy/my-awesome-bucket-vSgD5 [1]
                UserName: my-awesome-bucket-vSgD5 [2]
              endpoint:
                additionalConfig: null
                bucketHost: s3-us-west-1.amazonaws.com
                bucketName: my-awesome-bucket [3]
                bucketPort: 443
                region: us-west-1
                ssl: true
                subRegion: ""
            claimRef: null [4]
            reclaimPolicy: null
            storageClassName: s3-buckets [5]
          ```
          1. The AWS Policy created for this user and bucket.
          1. The new user generated by the Provisioner to access this existing bucket.
          1. The bucket name.
          1. The reference to the OBC (not filled in yet).
          1. The reference to the StorageClass used.
          *ConfigMap*
          ```yaml
           # kubectl get cm myobc -n s3-provisioner -o yaml
          apiVersion: v1
          data:
            BUCKET_HOST: s3-us-west-1.amazonaws.com [1]
            BUCKET_NAME: my-awesome-bucket [2]
            BUCKET_PORT: "443"
            BUCKET_REGION: us-west-1
            BUCKET_SSL: "true"
            BUCKET_SUBREGION: ""
          kind: ConfigMap
          metadata:
            creationTimestamp: "2019-04-01T19:11:38Z"
            finalizers:
            - objectbucket.io/finalizer
            name: my-awesome-bucket
            namespace: s3-provisioner
            resourceVersion: "892"
            selfLink: /api/v1/namespaces/s3-provisioner/configmaps/my-awesome-bucket
            uid: 2edcc58a-aff8-4a29-814a-ffbb6439a9cd
          ```
          1. The AWS S3 host.
          1. The name of the new bucket we are gaining access to.
          *Secret*
          ```yaml
           # kubectl get secret my-awesome-bucket -n s3-provisioner -o yaml
          apiVersion: v1
          data:
            AWS_ACCESS_KEY_ID: *the_new_access_id* [1]
            AWS_SECRET_ACCESS_KEY: *the_new_access_key_value* [2]
          kind: Secret
          metadata:
            creationTimestamp: "2019-04-03T15:42:22Z"
            finalizers:
            - objectbucket.io/finalizer
            name: my-awesome-bucket
            namespace: s3-provisioner
            resourceVersion: "15058"
            selfLink: /api/v1/namespaces/s3-provisioner/secrets/screeley-provb-5
            uid: 225c71a5-9d75-4ccc-b41f-bfe91b272a13
          type: Opaque
          ```
          1. The new generated AWS Access Key ID.
          1. The new generated AWS Secret Access Key.
          What happened in AWS? The first thing we do on any OBC request is
          create a new IAM user and generate Access ID and Secret Keys.
          This allows us to also better control ACLs. We also create a policy
          in IAM which we then attach to the user and bucket. We also created a new bucket, called *my-awesome-bucket*.
          When the OBC is deleted all of its Kubernetes and AWS resources will also be deleted, which includes:
          the generated OB, Secret, ConfigMap, IAM user, and policy.
          If the _retainPolicy_ on the StorageClass for this bucket is *"Delete"*, then, in addition to the above cleanup, the physical bucket is also deleted.
          **NOTE:** The actual bucket is only deleted if the Storage Class's _reclaimPolicy_ is "Delete".
          ### User Creates Pod
          Now that we have our bucket and connection/access information, a pod
          can be used to access the bucket. This can be done in several different
          ways, but the key here is that the provisioner has provided the proper
          endpoints and keys to access the bucket. The user then simply references
          the keys.
          1. Create a Sample Pod to Access the Bucket.
          ```yaml
          apiVersion: v1
          kind: Pod
          metadata:
            name: photo1
            labels:
              name: photo1
          spec:
            containers:
            - name: photo1
              image: docker.io/screeley44/photo-gallery:latest
              imagePullPolicy: Always
              envFrom:
              - configMapRef:
                  name: my-awesome-bucket <1>
              - secretRef:
                  name: my-awesome-bucket <2>
              ports:
              - containerPort: 3000
                protocol: TCP
          ```
          1. Name of the generated configmap from the provisioning process
          1. Name of the generated secret from the provisioning process
          *[Note]* Generated ConfigMap and Secret are same name as the OBC!
          Lastly, expose the pod as a service so you can access the url from a browser. In this example,
          I exposed as a LoadBalancer
          ```
            # kubectl expose pod photo1 --type=LoadBalancer --name=photo1 -n your-namespace
          ```
          To access via a url use the EXTERNAL-IP
          ```
            # kubectl get svc photo1
            NAME                         TYPE           CLUSTER-IP       EXTERNAL-IP                                                               PORT(S)          AGE
            photo1                       LoadBalancer   100.66.124.105   a00c53ccb3c5411e9b6550a7c0e50a2a-2010797808.us-east-1.elb.amazonaws.com   3000:32344/TCP   6d
          ```
          **NOTE:** This is just one example of a Pod that can utilize the bucket information,
          there are several ways that these pod applications can be developed and therefore
          the method of getting the actual values needed from the Secrets and ConfigMaps
          will vary greatly, but the idea remains the same, that the pod consumes the generated
          ConfigMap and Secret created by the provisioner.
        displayName: AWS S3 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - awss3provisioner
        - aws
        - S3
        - provisioner
        - buckets
        links:
        - name: AWS S3 Provisioner
          url: https://github.com/yard-turkey/aws-s3-provisioner
        - name: Provisioner Examples
          url: https://github.com/yard-turkey/aws-s3-provisioner/tree/master/examples
        - name: Operator Documentation
          url: https://github.com/yard-turkey/awss3operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - quay.io/screeley44/aws-s3-provisioner:v1.0.0
        version: 1.0.1
      entries:
      - name: awss3operator.1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: awss3-operator-registry
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Managed OpenShift Black Belt
      provider-url: https://github.com/rh-mobb/ocm-operator
    name: ocm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocm-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "GitLabIdentityProvider",
                "metadata": {
                  "name": "gitlab-sample"
                },
                "spec": {
                  "clientID": "gitlab",
                  "clientSecret": {
                    "name": "gitlab"
                  },
                  "clusterName": "my-cluster",
                  "displayName": "gitlab-sample",
                  "mappingMethod": "claim",
                  "url": "https://gitlab.com"
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "LDAPIdentityProvider",
                "metadata": {
                  "name": "ldap-sample"
                },
                "spec": {
                  "bindDN": "CN=test,OU=Users,DC=example,DC=com",
                  "bindPassword": {
                    "name": "test"
                  },
                  "clusterName": "my-cluster",
                  "displayName": "ldap-sample",
                  "mappingMethod": "claim",
                  "url": "ldap://test.example.com:389"
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "MachinePool",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "aws": {
                    "spotInstances": {
                      "enabled": false
                    }
                  },
                  "clusterName": "my-cluster",
                  "instanceType": "m5.xlarge",
                  "labels": {
                    "owner": "my-user"
                  },
                  "maximumNodesPerZone": 1,
                  "minimumNodesPerZone": 1,
                  "wait": false
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "ROSACluster",
                "metadata": {
                  "name": "rosa-sample"
                },
                "spec": {
                  "accountID": "111111111111",
                  "defaultMachinePool": {
                    "instanceType": "m5.xlarge",
                    "labels": {
                      "owner": "my-owner"
                    },
                    "minimumNodesPerZone": 2
                  },
                  "displayName": "my-cluster",
                  "iam": {
                    "userRole": "arn:aws:iam::111111111111:role/ManagedOpenShift-User-my_user-Role"
                  },
                  "tags": {
                    "owner": "my-owner"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ocm-operator:v0.1.0
          createdAt: "2023-06-02T17:32:25Z"
          operatorframework.io/suggested-namespace: ocm-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ocm-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLabIdentityProvider is the Schema for the gitlabidentityproviders
              API.
            displayName: Git Lab Identity Provider
            kind: GitLabIdentityProvider
            name: gitlabidentityproviders.ocm.mobb.redhat.com
            version: v1alpha1
          - description: LDAPIdentityProvider is the Schema for the ldapidentityproviders
              API.
            displayName: LDAPIdentity Provider
            kind: LDAPIdentityProvider
            name: ldapidentityproviders.ocm.mobb.redhat.com
            version: v1alpha1
          - description: MachinePool is the Schema for the machinepools API.
            displayName: Machine Pool
            kind: MachinePool
            name: machinepools.ocm.mobb.redhat.com
            version: v1alpha1
          - description: ROSACluster is the Schema for the clusters API.
            displayName: ROSACluster
            kind: ROSACluster
            name: rosaclusters.ocm.mobb.redhat.com
            version: v1alpha1
        description: "# Quick Start\n\n## Prerequisites\n\n- [oc](https://docs.openshift.com/container-platform/4.8/cli_reference/openshift_cli/getting-started-cli.html)\n-
          [aws](https://docs.aws.amazon.com/cli/latest/userguide/getting-started-install.html)\n-
          [ROSA Cluster (4.12+)](https://mobb.ninja/docs/rosa/sts/)\n\n> **Warning:**\n>
          This is only tested on ROSA, but may work on other Kubernetes and OpenShift
          variants.  It requires\n> an AWS STS-enabled cluster.  Additionally, OpenShift
          versions 4.12+ are recommended, as they \n> enable [CRD CEL validation](https://kubernetes.io/blog/2022/09/23/crd-validation-rules-beta/).
          \ Versions\n> prior to 4.12 provide a lower version of Kubernetes that does
          not enable this feature.  They may \n> work, but provide no input validation
          when submitting custom resources to the cluster.\n\nBefore installing this
          operator, there are a couple secrets that must be created.\n\n1. OCM Token:
          used to authenticate against the OCM API.  The controller will not start
          without\nthis token as it used for all calls into OCM.\n2. AWS Credentials:
          uses an assume role to a role created in the instructions below in order\nto
          access AWS resources.  This is needed for cluster creation.  See \n[Create
          AWS IAM Policies and Roles](#create-aws-iam-policies-and-roles) for more
          details.\n\n> **Note:**\n> Certain custom resources have their own specific
          prerequisites.  Please use `oc explain` \n> or read the docs [here](https://github.com/rh-mobb/ocm-operator/tree/main/docs)
          for more details.\n\n### Create OCM Token Secret\n\n1. Create a namespace
          where you wish to install the operator:\n\n```bash\noc new-project ocm-operator\n```\n\n2.
          Create a secret containing the OCM_TOKEN.  This token can be obtained form
          \nhttps://console.redhat.com/openshift/token and is used by the operator
          to authenticate \nagainst the OCM API.  This token must exist in the same
          namespace that the operator \nis running and be named `ocm-token`.  It also
          expects the key to be called `OCM_TOKEN` \nas the operator is expecting
          this value as an environment variable.\n\n```bash\noc create secret generic
          ocm-token \\\n  --namespace=ocm-operator \\\n  --from-literal=OCM_TOKEN=${MY_OCM_TOKEN}\n```\n\n###
          Create AWS IAM Policies and Roles\n\nThe operator will need to elevate privileges
          in order to perform things like \ncreating the operator-roles for the clusters.
          \ Because of this, the operator \nmust have a specific role created to allow
          it these permissions.\n\n**NOTE:** please understand what you are doing
          if you deviate from the known good \npolicies.  If errors or more stringent
          security lockdowns are found, please submit a PR \nso that we can get this
          fixed.\n\n1. Set `ACCOUNT_ID` environment variable to define your AWS account
          ID:\n\n```bash\nexport ACCOUNT_ID=111111111111\n```\n\n2. Create the permissions
          boundary.  Because the operator needs to create policies and \nroles, the
          boundary ensures that the operator is not allowed to create additional\npermissions
          outside of the defined boundary.  The sample permission set is located \nat
          `test/aws/boundary.json` in this repository:\n\n```bash\nboundary=$(curl
          -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/boundary.json\")\naws
          iam create-policy \\\n  --policy-name \"OCMOperatorBoundary\" \\\n  --policy-document
          \"$boundary\"\n```\n\n3. Create the policy.  This policy sets what the operator
          is allowed to do.  For any \n`iam` permission, the boundary created above
          is used.  The sample permission set is \nlocated at `test/aws/iam_policy.json`
          in this repository:\n\n```bash\npolicy=$(curl -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/iam_policy.json\")\naws
          iam create-policy \\\n  --policy-name \"OCMOperator\" \\\n  --policy-document
          \"$policy\"\n```\n\n4. Create the role using a trust policy and attach the
          previously created role.  The trust \npolicy is located at `test/aws/trust_policy.json`
          in this repository.  Please note that \nthe trust policy requires an OIDC
          configuration.  The OIDC configuration refers to \n**where the operator
          is running, NOT what it is provisioning**:\n\n```bash\ntrust_policy=$(curl
          -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/trust_policy.json\")\naws
          iam create-role \\\n    --role-name OCMOperator \\\n    --assume-role-policy-document
          \"$trust_policy\"\naws iam attach-role-policy \\\n    --role-name OCMOperator
          \\\n    --policy-arn arn:aws:iam::$ACCOUNT_ID:policy/OCMOperator\n```\n\n5.
          Finally, create the secret containing the assume role credentials.  The
          previous steps allow \nthe operator to assume the role you created in the
          previous step with the permissions created \nvia the previous policies.\n\n```bash\ncat
          <<EOF > /tmp/credentials\n[default]\nrole_arn = arn:aws:iam::$ACCOUNT_ID:role/OCMOperator\nweb_identity_token_file
          = /var/run/secrets/openshift/serviceaccount/token\nEOF\n\noc create secret
          generic aws-credentials \\\n  --namespace=ocm-operator \\\n  --from-file=credentials=/tmp/credentials\n```\n\n##
          Install the Operator\n\n```bash\ncat <<EOF | oc apply -f -\napiVersion:
          operators.coreos.com/v1alpha1\nkind: Subscription\nmetadata:\n  name: ocm-operator\n
          \ namespace: ocm-operator\nspec:\n  channel: v0.1\n  installPlanApproval:
          Automatic\n  name: ocm-operator\n  source: community-operators\n  sourceNamespace:
          ocm-operator\n  startingCSV: ocm-operator.v0.1.0\nEOF\n```\n\n## Provision
          Resources\n\nOnce the operator is available and running, you can provision
          any of the \nresources that it manages.  Also note that documentation is
          always available \nonce the operator is installed as well by using the `oc
          explain` command.  For \nexample, `oc explain rosacluster.spec.clusterName`
          will give you detailed documentation \nabout what the field does.\n\nSee
          the following documentation for details:\n\n* [ROSA Clusters](https://github.com/rh-mobb/ocm-operator/blob/main/docs/clusters.md)\n*
          [Machine Pools](https://github.com/rh-mobb/ocm-operator/blob/main/docs/machinepools.md)\n*
          [Identity Providers](https://github.com/rh-mobb/ocm-operator/blob/main/docs/identityproviders.md)\n"
        displayName: OpenShift Cluster Manager Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ocm
        - rosa
        - redhat
        - machinepool
        - mobb
        links:
        - name: OCM Machine Pool Operator
          url: https://github.com/rh-mobb/ocm-operator
        maintainers:
        - email: dscott@redhat.com
          name: Dustin Scott
        maturity: alpha
        provider:
          name: Managed OpenShift Black Belt
          url: https://github.com/rh-mobb/ocm-operator
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/mobb/ocm-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: ocm-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ocm-operator
    provider:
      name: Managed OpenShift Black Belt
      url: https://github.com/rh-mobb/ocm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SAP LinuxLab - Red Hat
      provider-url: https://github.com/sap-linuxlab
    name: sap-hana-express-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sap-hana-express-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "db.sap-redhat.io/v1alpha1",
                "kind": "HanaExpress",
                "metadata": {
                  "name": "hanaexpress-sample"
                },
                "spec": {
                  "credential": {
                    "key": "hxepasswd.json",
                    "name": "hxepasswd"
                  },
                  "pvcSize": "1Gi"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/redhat-sap-cop/sap-hanaexpress-operator:0.1.4
          createdAt: "2023-07-16T20:53:02Z"
          description: Manage the lifecycle of SAP Hana Express
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HanaExpress is the Schema for the hanaexpresses API
            displayName: Hana Express
            kind: HanaExpress
            name: hanaexpresses.db.sap-redhat.io
            version: v1alpha1
        description: Operator for managing the lifecycle of SAP Hana Express on Kubernetes
          platform
        displayName: SAP Hana Express Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Database
        - SAP
        links:
        - name: Sap Hana Express Operator
          url: https://github.com/redhat-sap/sap-hana-express-operator
        maintainers:
        - email: mjiao@redhat.com
          name: RedHat@SAPLinuxLab
        maturity: alpha
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/redhat-sap-cop/sap-hanaexpress-operator:0.1.4
        version: 1.0.2
      entries:
      - name: sap-hana-express-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: sap-hana-express-operator
    provider:
      name: SAP LinuxLab - Red Hat
      url: https://github.com/sap-linuxlab
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-registry-3
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-3.v3.0.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "simple"
                },
                "spec": {
                  "app": {
                    "ingress": {
                      "host": "simple-app.apps.cluster.example"
                    }
                  },
                  "ui": {
                    "ingress": {
                      "host": "simple-ui.apps.cluster.example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-3-operator:3.0.15
          createdAt: "2025-09-22T13:04:40Z"
          description: Deploy and manage Apicurio Registry 3 on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Apicurio Registry
              version 3.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Apicurio Registry 3

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Allows you manage event schemas and API designs using the Apicurio Registry web console, REST API, Maven plug-in, or Java client
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, MySQL, SQL Server, in-memory
          - Compatible with Confluent APIs
          - Runs on a lightweight Quarkus platform

          ## Apicurio Registry 3 Operator

          Provides a quick and easy way to deploy and manage Apicurio Registry 3 on Kubernetes.

          **Features**
          - Supports installation and configuration of Apicurio Registry 3
          - Perform a rolling upgrade of Apicurio Registry 3

          ## Prerequisites

          This operator does not deploy storage for Apicurio Registry 3. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio Registry 3 and Apicurio Registry 3 Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Apicurio Registry 3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-3-operator:3.0.15
        version: 3.0.15
      entries:
      - name: apicurio-registry-3.v3.0.15
        version: 3.0.15
      - name: apicurio-registry-3.v3.0.14
        version: 3.0.14
      - name: apicurio-registry-3.v3.0.12
        version: 3.0.12
      - name: apicurio-registry-3.v3.0.9
        version: 3.0.9
      - name: apicurio-registry-3.v3.0.8
        version: 3.0.8
      - name: apicurio-registry-3.v3.0.7
        version: 3.0.7
      name: 3.x
    defaultChannel: 3.x
    packageName: apicurio-registry-3
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM Edge
      provider-url: ""
    name: infrastructure-asset-orchestrator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infrastructure-asset-orchestrator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infrastructure.asset.orchestrator.com/v1",
                "kind": "Generatekey",
                "metadata": {
                  "name": "modelbuilder-token"
                },
                "spec": {}
              },
              {
                "apiVersion": "infrastructure.asset.orchestrator.com/v1",
                "kind": "Mb-broker-service",
                "metadata": {
                  "name": "mb-broker-service-sample"
                },
                "spec": {
                  "env_type": "prod",
                  "cloud_type": "IBM",
                  "external_db": {
                    "enabled": false
                  },
                  "postgres": {
                    "enable_pg_tls": true,
                    "storage_class": "",
                    "storage_size": "20G",
                    "backup_type": "incremental",
                    "backup_frequency": "@daily"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/iao-operator@sha256:5cc395f22d86641e175339deb134af6c74de92b81470a65a5c736c8005d645fd
          createdAt: "2021-08-19T12:00:00Z"
          olm.skipRange: <1.0.2
          operators.operatorframework.io/builder: operator-sdk-v1.12.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/infrastructure-asset-orchestrator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Generate key CR
            displayName: GenerateKey
            kind: Generatekey
            name: generatekeys.infrastructure.asset.orchestrator.com
            version: v1
          - description: Deploy MB broker deployment
            displayName: MBBrokerService
            kind: Mb-broker-service
            name: mb-broker-services.infrastructure.asset.orchestrator.com
            version: v1
        description: Infrastructure Asset Orchestrator provides users access to cloud
          resources and services. Leverages the dynamic nature of the cloud platforms
          to provide significant cost savings and efficiencies. Using the Infrastructure
          Asset Orchestrator user can provision, configure, utilize and deprovision
          variety of services on the IBM Cloud Catalog.
        displayName: Infrastructure Asset Orchestrator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cloud-service
        - mbbroker-service
        links:
        - name: Infrastructure Asset Orchestrator
          url: https://infrastructure-asset-orchestrator.domain
        maintainers:
        - email: edge@us.ibm.com
          name: IBM-Edge
        maturity: alpha
        provider:
          name: IBM Edge
        relatedImages:
        - registry.connect.redhat.com/ibm/iao-operator@sha256:5cc395f22d86641e175339deb134af6c74de92b81470a65a5c736c8005d645fd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:609c19d1718947ee814b829334c37e1d720acf5fb642b0712b8ab1b14d57f53f
        - registry.connect.redhat.com/ibm/iao-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/iao-orchestrator-service@sha256:145928929fd0c2b7d9ed5fb374d7cc0e6b48269847e810c5472580a1538ad44c
        - registry.connect.redhat.com/ibm/iao-mb-broker-service@sha256:6531b4196c51246958027453c2d9c7dbcdb4a1f55c826d1da238d256cdbc8a10
        version: 1.0.2
      entries:
      - name: infrastructure-asset-orchestrator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: infrastructure-asset-orchestrator-certified
    provider:
      name: IBM Edge
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datasance
      provider-url: ""
    name: datasance-pot-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datasance-pot-operator.v3.4.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasance.com/v3",
                "kind": "ControlPlane",
                "metadata": {
                  "name": "pot"
                },
                "spec": {
                  "replicas": {
                    "controller": 2
                  },
                  "database": {
                    "provider": "postgres",
                    "user": "user",
                    "host": "postgre.host",
                    "port": 5432,
                    "password": "password",
                    "databaseName": "dbname"
                  },
                  "auth": {
                    "url": "auth.url",
                    "realm": "realm",
                    "realmKey": "realmKey",
                    "ssl": "none",
                    "controllerClient": "controllerClient",
                    "controllerSecret": "controllerClientSecret",
                    "viewerClient": "viewerClient"
                  },
                  "images": {
                    "pullSecret": "pullsecret",
                    "controller": "ghcr.io/datasance/controller@sha256:aa896bae6c3e6eda70938a03aaa649226356d9e4a1d03ba427d4e4ee3090ca6c",
                    "portManager": "ghcr.io/datasance/port-manager@sha256:4bf3793e9e3cd10df9b46c459e983631393825ca788ffbbb3ec3659158ad61fe",
                    "proxy": "ghcr.io/datasance/proxy@sha256:276dbed0c8ef4f5b03fdf29d5c863b6f3efcae50444ce8123220c5673272860f",
                    "router": "ghcr.io/datasance/router@sha256:82657fc8f39d33d8ef0161cf27d705cbdc0a97dca07b22a5fe78fd4213ef0d31"
                  },
                  "services": {
                    "controller": {
                      "type": "LoadBalancer",
                      "annotations": {
                        "service.beta.kubernetes.io/azure-load-balancer-internal": "false"
                      }
                    },
                    "proxy": {
                      "type": "LoadBalancer",
                      "annotations": {
                        "service.beta.kubernetes.io/azure-load-balancer-internal": "false"
                      }
                    },
                    "router": {
                      "type": "LoadBalancer",
                      "annotations": {
                        "service.beta.kubernetes.io/azure-load-balancer-internal": "false"
                      }
                    }
                  },
                  "controller": {
                    "ecnViewerUrl": "pot.demo.datasance.com",
                    "https": true,
                    "secretName": "secretName"
                  },
                  "ingresses": {
                    "controller": {
                      "annotations": {
                        "cert-manager.io/cluster-issuer": "letsencrypt"
                      },
                      "ingressClassName": "nginx",
                      "host": "pot.demo.datasance.com",
                      "secretName": "secretName"
                    },
                    "router": {
                      "address": "pot.demo.datasance.com",
                      "messagePort": 5672,
                      "interiorPort": 55672,
                      "edgePort": 45672
                    },
                    "httpProxy": {
                      "address": "pot.demo.datasance.com"
                    },
                    "tcpProxy": {
                      "address": "pot.demo.datasance.com"
                    }
                  }
                },
                "status": {
                  "conditions": [
                    {
                      "lastTransitionTime": "2022-04-10T22:44:09Z",
                      "message": "",
                      "reason": "initial_status",
                      "status": "True",
                      "type": "deploying"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          containerImage: ghcr.io/datasance/operator@sha256:143b9307c2520ef16b40c4c92e6faee8b24b7fb3a97682231e53236ddd532687
          createdAt: "2024-11-04T20:45:43Z"
          description: Datasance PoT is an Enteprise version of Eclipse ioFog Platform.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Datasance PoT Enterprise Open
            Source Subscriptions based on number of Edge device that ControlPlane
            manages"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/datasance/iofog-operator
          support: Datasance Teknoloji A.S.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Datasance PoT supported ioFog ControlPlane CRD for managing
              container workloads on Edge device
            kind: ControlPlane
            name: controlplanes.datasance.com
            version: v3
        description: ioFog Control Plane on Kubernetes
        displayName: iofog-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iofog
        - datasance
        - pot
        links:
        - name: Datasance PoT // Iofog Operator
          url: https://docs.datasance.com
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Datasance
        relatedImages:
        - ghcr.io/datasance/operator@sha256:143b9307c2520ef16b40c4c92e6faee8b24b7fb3a97682231e53236ddd532687
        - ghcr.io/datasance/controller@sha256:aa896bae6c3e6eda70938a03aaa649226356d9e4a1d03ba427d4e4ee3090ca6c
        - ghcr.io/datasance/router@sha256:82657fc8f39d33d8ef0161cf27d705cbdc0a97dca07b22a5fe78fd4213ef0d31
        - ghcr.io/datasance/proxy@sha256:276dbed0c8ef4f5b03fdf29d5c863b6f3efcae50444ce8123220c5673272860f
        - ghcr.io/datasance/port-manager@sha256:4bf3793e9e3cd10df9b46c459e983631393825ca788ffbbb3ec3659158ad61fe
        version: 3.4.9
      entries:
      - name: datasance-pot-operator.v3.4.9
        version: 3.4.9
      name: stable
    defaultChannel: stable
    packageName: datasance-pot-operator
    provider:
      name: Datasance
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: gcp-filestore-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gcp-filestore-csi-driver-operator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel8-operator@sha256:1ec61d2988b95a10260c7932e9dc6a9d1d5b3eb4148b22afd438aa4ee75dc9b3
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure GCP Filestore CSI driver.
          olm.skipRange: '>=4.11.0-0 <4.13.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/gcp-filestore-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Operator that installs and configures the CSI driver for GCP Filestore. The GCP Filestore CSI Driver Operator is a Technology Preview feature. For more information on Red Hat Technology Preview features support scope, see https://access.redhat.com/support/offerings/techpreview/.
        displayName: GCP Filestore CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - filestore
        links:
        - name: Documentation
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: preview
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:700c0e50f86a70f506403aba75e020a17530ada284af7eee77d4b4ce3dbf4102
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:4f1e873080754ae9aebc361cd77f965e1ade369403346aa0cf4cf502082d18e4
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:ee45e11a6fe8cf60f3ac285cba1b23459ca95068bd9769042ae052c778c3e080
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:537fbc8fe2b212dc798361099d2c6fd981cc846c1cf5454b703ba08d488b61f8
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel8@sha256:a1a4e7b93d6609561a7a2cd732ce772bf0047448e5019298d66c4b985baecfb2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:767682dd3bd65651a8204c9fb732b9c48b99127189992b43abc6ecce027f4589
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel8-operator@sha256:1ec61d2988b95a10260c7932e9dc6a9d1d5b3eb4148b22afd438aa4ee75dc9b3
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:6f1b497772b953468f0b816face8810ace8ea44fa91dd4a3b3e583da2cac350c
        version: 4.13.0-202309181427
      entries:
      - name: gcp-filestore-csi-driver-operator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: preview
    - currentCSV: gcp-filestore-csi-driver-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9-operator@sha256:2dc13de58b391237dc48b9335d88b99f5e26d0e5edc5a664c7d39f16ca9b297d
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure GCP Filestore CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.11.0-0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/gcp-filestore-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Operator that installs and configures the CSI driver for GCP Filestore.
        displayName: GCP Filestore CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - filestore
        links:
        - name: Documentation
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:a5704d26d0b562dba9633d1c107f34e05c750c0e3a64dbcace4dab7b964fff8d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9@sha256:8e801d924071c8f2513555450e688bff60d4100b2ca56d2ba9517038a1d48d4d
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:9c3c3d2b857ce289324aee56df1ba4ab5e5ea1e76cd86ac988233463915834b4
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:78e02d937740d742eb33b08c7317f211c25020c375069d5a97344107716608d3
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:b28def976fc155dff57ff3003cc462a8e9923b325af0b319b65220c66688699e
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0627c09fe3f5a015f117fb30af1f4d17340aefa48b23994737d3901a804b63d3
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9-operator@sha256:2dc13de58b391237dc48b9335d88b99f5e26d0e5edc5a664c7d39f16ca9b297d
        version: 4.16.0-202509240123
      entries:
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202508270905
        version: 4.16.0-202508270905
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202504121405
        version: 4.16.0-202504121405
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: gcp-filestore-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-prometheusservice-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-prometheusservice-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "prometheusservice.services.k8s.aws/v1alpha1",
                "kind": "RuleGroupsNamespace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "prometheusservice.services.k8s.aws/v1alpha1",
                "kind": "Workspace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/prometheusservice-controller:1.3.1
          createdAt: "2025-09-19T21:11:38Z"
          description: AWS Prometheus controller is a service controller for managing
            Prometheus resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertManagerDefinition represents the state of an AWS prometheusservice
              AlertManagerDefinition resource.
            displayName: AlertManagerDefinition
            kind: AlertManagerDefinition
            name: alertmanagerdefinitions.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: LoggingConfiguration represents the state of an AWS prometheusservice
              LoggingConfiguration resource.
            displayName: LoggingConfiguration
            kind: LoggingConfiguration
            name: loggingconfigurations.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: RuleGroupsNamespace represents the state of an AWS prometheusservice
              RuleGroupsNamespace resource.
            displayName: RuleGroupsNamespace
            kind: RuleGroupsNamespace
            name: rulegroupsnamespaces.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: Workspace represents the state of an AWS prometheusservice
              Workspace resource.
            displayName: Workspace
            kind: Workspace
            name: workspaces.prometheusservice.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Prometheus resources in AWS from within your Kubernetes cluster.

          **About Amazon Managed Service for Prometheus**

          Amazon Managed Service for Prometheus is a serverless, Prometheus-compatible monitoring service for container metrics that makes it easier to securely monitor container environments at scale. With Amazon Managed Service for Prometheus, you can use the same open-source Prometheus data model and query language that you use today to monitor the performance of your containerized workloads, and also enjoy improved scalability, availability, and security without having to manage the underlying infrastructure.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Prometheus
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - prometheusservice
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Prometheus Developer Resources
          url: https://aws.amazon.com/prometheus/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: prometheus maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/prometheusservice-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-prometheusservice-controller.v1.3.1
        version: 1.3.1
      - name: ack-prometheusservice-controller.v1.3.0
        version: 1.3.0
      - name: ack-prometheusservice-controller.v1.2.27
        version: 1.2.27
      - name: ack-prometheusservice-controller.v1.2.26
        version: 1.2.26
      - name: ack-prometheusservice-controller.v1.2.24
        version: 1.2.24
      - name: ack-prometheusservice-controller.v1.2.23
        version: 1.2.23
      - name: ack-prometheusservice-controller.v1.2.22
        version: 1.2.22
      - name: ack-prometheusservice-controller.v1.2.20
        version: 1.2.20
      - name: ack-prometheusservice-controller.v1.2.19
        version: 1.2.19
      - name: ack-prometheusservice-controller.v1.2.17
        version: 1.2.17
      - name: ack-prometheusservice-controller.v1.2.16
        version: 1.2.16
      - name: ack-prometheusservice-controller.v1.2.15
        version: 1.2.15
      - name: ack-prometheusservice-controller.v1.2.14
        version: 1.2.14
      - name: ack-prometheusservice-controller.v1.2.13
        version: 1.2.13
      - name: ack-prometheusservice-controller.v1.2.12
        version: 1.2.12
      - name: ack-prometheusservice-controller.v1.2.11
        version: 1.2.11
      - name: ack-prometheusservice-controller.v1.2.10
        version: 1.2.10
      - name: ack-prometheusservice-controller.v1.2.9
        version: 1.2.9
      - name: ack-prometheusservice-controller.v1.2.6
        version: 1.2.6
      - name: ack-prometheusservice-controller.v1.2.5
        version: 1.2.5
      - name: ack-prometheusservice-controller.v1.2.4
        version: 1.2.4
      - name: ack-prometheusservice-controller.v1.2.3
        version: 1.2.3
      - name: ack-prometheusservice-controller.v1.2.2
        version: 1.2.2
      - name: ack-prometheusservice-controller.v1.2.1
        version: 1.2.1
      - name: ack-prometheusservice-controller.v1.2.0
        version: 1.2.0
      - name: ack-prometheusservice-controller.v1.1.0
        version: 1.1.0
      - name: ack-prometheusservice-controller.v1.0.0
        version: 1.0.0
      - name: ack-prometheusservice-controller.v0.1.3
        version: 0.1.3
      - name: ack-prometheusservice-controller.v0.1.2
        version: 0.1.2
      - name: ack-prometheusservice-controller.v0.1.1
        version: 0.1.1
      - name: ack-prometheusservice-controller.v0.1.0
        version: 0.1.0
      - name: ack-prometheusservice-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-prometheusservice-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: networkservicemesh.io
      provider-url: networkservicemesh.io
    name: nsm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nsm-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nsm.networkservicemesh.io/v1alpha1",
                "kind": "NSM",
                "metadata": {
                  "name": "nsm-sample",
                  "namespace": "nsm"
                },
                "spec": {
                  "forwarders": [
                    {
                      "image": "ghcr.io/networkservicemesh/cmd-forwarder-vpp:v1.8.0",
                      "type": "vpp"
                    }
                  ],
                  "nsmLogLevel": "TRACE",
                  "nsmPullPolicy": "IfNotPresent",
                  "registry": {
                    "image": "ghcr.io/networkservicemesh/cmd-registry-memory:v1.8.0",
                    "type": "memory"
                  },
                  "version": "v1.8.0",
                  "webhook": {
                    "image": "ghcr.io/networkservicemesh/admission-webhook-k8s:v1.8.0"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: quay.io/acmenezes/nsm-operator:v1.8.0
          createdAt: "2023-03-02T19:36:46Z"
          operatorframework.io/suggested-namespace: nsm
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/networkservicemesh/nsm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkServiceEndpoint
            name: networkserviceendpoints.networkservicemesh.io
            version: v1
          - kind: NetworkService
            name: networkservices.networkservicemesh.io
            version: v1
          - kind: NSM
            name: nsms.nsm.networkservicemesh.io
            version: v1alpha1
        description: |-
          The Network Service Mesh Operator manages and install network service meshes.

          Check networkservicemesh.io

          Install requirements:

          Before installing the operator it's necessary to have installed [spire](https://github.com/spiffe/spire)

          We have a ready to install yaml file for spire [here](https://github.com/networkservicemesh/nsm-operator/blob/master/config/spire/spire.yaml)
        displayName: nsm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network
        - service mesh
        links:
        - name: Nsm Operator
          url: https://nsm-operator.domain
        maintainers:
        - email: adcmenezes@gmail.com
          name: Alexandre Menezes
        maturity: alpha
        provider:
          name: networkservicemesh.io
          url: networkservicemesh.io
        relatedImages:
        - quay.io/acmenezes/nsm-operator:v1.8.0
        version: 1.8.0
      entries:
      - name: nsm-operator.v1.8.0
        version: 1.8.0
      name: alpha
    defaultChannel: alpha
    packageName: nsm-operator
    provider:
      name: networkservicemesh.io
      url: networkservicemesh.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: observability-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      name: candidate
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      - name: observability-operator.v0.0.22
        version: 0.0.22
      name: development
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      name: stable
    defaultChannel: stable
    packageName: observability-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-atlas-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-atlas-kubernetes.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "id": "1",
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "WEEKLY",
                      "id": "2",
                      "retentionUnit": "DAYS",
                      "retentionValue": 6
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": true,
                  "policies": [
                    {
                      "name": "atlas-default-backuppolicy",
                      "namespace": "mongodb-atlas-system"
                    },
                    {
                      "name": "atlas-default-backuppolicy2",
                      "namespace": "mongodb-atlas-system"
                    }
                  ],
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "my-database-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "david"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-cluster"
                },
                "spec": {
                  "deploymentSpec": {
                    "name": "test-cluster",
                    "providerSettings": {
                      "instanceSizeName": "M10",
                      "providerName": "AWS",
                      "regionName": "US_EAST_1"
                    }
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          containerImage: registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator@sha256:3cfbd700e4771c2c7e75c2c41cf6de57c9ef71ecd286e96a21d2567e43424bc3
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasclusters API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB Atlas —
          the only multi-cloud document database service that gives you the versatility you need to build sophisticated and resilient applications that can adapt to changing customer demands and market trends.

          > Current Status: *Beta*. The Operator gives users the ability to provision
          > Atlas projects, clusters and database users using Kubernetes Specifications and bind connection information
          > into applications deployed to Kubernetes.

          ## Quick Start guide
          ### Step 1. Deploy Kubernetes operator by clicking Install button.

          ### Step 2. Create Atlas Cluster

          **1.** Create an Atlas API Key Secret
          In order to work with the Atlas Operator you need to provide [authentication information](https://docs.atlas.mongodb.com/configure-api-access)
           to allow the Atlas Operator to communicate with Atlas API. Once you have generated a Public and Private key in Atlas, you can create a Kuberentes Secret with:
          ```
          kubectl create secret generic mongodb-atlas-operator-api-key \
                   --from-literal="orgId=<the_atlas_organization_id>" \
                   --from-literal="publicApiKey=<the_atlas_api_public_key>" \
                   --from-literal="privateApiKey=<the_atlas_api_private_key>" \
                   -n openshift-operators
          ```
          (Note, that you should use the namespace where the Operator was installed - it's `openshift-operators` by default)

          **2.** Create an `AtlasProject` Custom Resource

          The `AtlasProject` CustomResource represents Atlas Projects in our Kubernetes cluster. You need to specify
          `projectIpAccessList` with the IP addresses or CIDR blocks of any hosts that will connect to the Atlas Cluster.
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasProject
          metadata:
            name: my-project
          spec:
            name: Test Atlas Operator Project
            projectIpAccessList:
              - ipAddress: "192.0.2.15"
                comment: "IP address for Application Server A"
              - ipAddress: "203.0.113.0/24"
                comment: "CIDR block for Application Server B - D"
          ```
          **3.** Create an `AtlasDeployment` Custom Resource.
          The example below is a minimal configuration to create an M10 Atlas cluster in the AWS US East region. For a full list of properties, check
          `atlasclusters.atlas.mongodb.com` [CRD specification](config/crd/bases/atlas.mongodb.com_atlasclusters.yaml)):
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasDeployment
          metadata:
            name: my-atlas-cluster
          spec:
            name: "Test-cluster"
            projectRef:
              name: my-project
            providerSettings:
              instanceSizeName: M10
              providerName: AWS
              regionName: US_EAST_1
          ```

          **4.** Create a database user password Kubernetes Secret
          The Secret must be created in the same namespace as the `AtlasDeployment` and `AtlasProject` were created.
          ```
          kubectl create secret generic the-user-password --from-literal="password=P@@sword%"
          ```

          **5.** Create an `AtlasDatabaseUser` Custom Resource

          In order to connect to an Atlas Cluster the database user needs to be created. `AtlasDatabaseUser` resource should reference
          the password Kubernetes Secret created in the previous step.
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasDatabaseUser
          metadata:
            name: my-database-user
          spec:
            roles:
              - roleName: "readWriteAnyDatabase"
                databaseName: "admin"
            projectRef:
              name: my-project
            username: theuser
            passwordSecretRef:
              name: the-user-password
          ```
          **6.** Wait for the `AtlasDatabaseUser` Custom Resource to be ready

          Wait until the AtlasDatabaseUser resource gets to "ready" status (it will wait until the cluster is created that may take around 10 minutes):
          ```
          kubectl get atlasdatabaseusers my-database-user -o=jsonpath='{.status.conditions[?(@.type=="Ready")].status}'
          True
          ```
          ### Step 3. Connect your application to the Atlas Cluster

          The Atlas Operator will create a Kubernetes Secret with the information necessary to connect to the Atlas Cluster created
          in the previous step. An application in the same Kubernetes Cluster can mount and use the Secret:

          ```
          ...
          containers:
                - name: test-app
                  env:
                   - name: "CONNECTION_STRING"
                     valueFrom:
                       secretKeyRef:
                         name: test-atlas-operator-project-test-cluster-theuser
                         key: connectionStringStandardSrv

          ```
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator@sha256:3cfbd700e4771c2c7e75c2c41cf6de57c9ef71ecd286e96a21d2567e43424bc3
        version: 1.0.0
      entries:
      - name: mongodb-atlas-kubernetes.v1.0.0
        version: 1.0.0
      name: beta
    - currentCSV: mongodb-atlas-kubernetes.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupCompliancePolicy",
                "metadata": {
                  "name": "my-backup-compliance-policy"
                },
                "spec": {
                  "authorizedEmail": "john.doe@example.com",
                  "authorizedUserFirstName": "John",
                  "authorizedUserLastName": "Doe",
                  "copyProtectionEnabled": false,
                  "encryptionAtRestEnabled": false,
                  "onDemandPolicy": {
                    "retentionUnit": "weeks",
                    "retentionValue": 3
                  },
                  "overwriteBackupPolicies": false,
                  "pointInTimeEnabled": true,
                  "restoreWindowDays": 42,
                  "scheduledPolicyItems": [
                    {
                      "frequencyInterval": 2,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "hourly",
                      "retentionUnit": "days",
                      "retentionValue": 2
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "weekly",
                      "retentionUnit": "weeks",
                      "retentionValue": 4
                    },
                    {
                      "frequencyInterval": 15,
                      "frequencyType": "monthly",
                      "retentionUnit": "months",
                      "retentionValue": 12
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": false,
                  "copySettings": [
                    {
                      "cloudProvider": "AWS",
                      "frequencies": [
                        "WEEKLY"
                      ],
                      "regionName": "US_WEST_1",
                      "shouldCopyOplogs": false
                    }
                  ],
                  "policy": {
                    "name": "atlasbackuppolicy-sample"
                  },
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2,
                  "useOrgAndGroupNamesInExportPrefix": true
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasCustomRole",
                "metadata": {
                  "name": "shard-operator-role"
                },
                "spec": {
                  "projectRef": {
                    "name": "my-project"
                  },
                  "role": {
                    "actions": [
                      {
                        "name": "getShardMap",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "shardingState",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "connPoolStats",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "getLog",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      }
                    ],
                    "inheritedRoles": [
                      {
                        "database": "admin",
                        "name": "operator-role-1"
                      }
                    ],
                    "name": "my-role"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDataFederation",
                "metadata": {
                  "name": "my-data-federation",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "name": "my-data-federation",
                  "privateEndpoints": [
                    {
                      "endpointId": "vpce-03f9eeaa764e32454",
                      "provider": "AWS",
                      "type": "DATA_LAKE"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project",
                    "namespace": "mongodb-atlas-system"
                  },
                  "storage": {
                    "databases": [
                      {
                        "collections": [
                          {
                            "dataSources": [
                              {
                                "storeName": "http-test",
                                "urls": [
                                  "https://data.cityofnewyork.us/api/views/vfnx-vebw/rows.csv"
                                ]
                              }
                            ],
                            "name": "test-collection-1"
                          }
                        ],
                        "name": "test-db-1"
                      }
                    ],
                    "stores": [
                      {
                        "name": "http-test",
                        "provider": "http"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "the-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "theuser"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-deployment"
                },
                "spec": {
                  "backupRef": {
                    "name": "atlasbackupschedule-sample"
                  },
                  "deploymentSpec": {
                    "backupEnabled": true,
                    "clusterType": "REPLICASET",
                    "mongoDBMajorVersion": "7.0",
                    "name": "test-deployment",
                    "replicationSpecs": [
                      {
                        "regionConfigs": [
                          {
                            "electableSpecs": {
                              "instanceSize": "M10",
                              "nodeCount": 3
                            },
                            "priority": 7,
                            "providerName": "AWS",
                            "regionName": "US_EAST_1"
                          }
                        ],
                        "zoneName": "Zone 1"
                      }
                    ],
                    "tags": [
                      {
                        "key": "team",
                        "value": "atlas-operator"
                      }
                    ]
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasFederatedAuth",
                "metadata": {
                  "name": "atlas-default-federated-auth",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "connectionSecretRef": {
                    "name": "my-org-secret",
                    "namespace": "mongodb-atlas-system"
                  },
                  "domainAllowList": [
                    "my-org-domain.com"
                  ],
                  "domainRestrictionEnabled": true,
                  "enabled": true,
                  "postAuthRoleGrants": [
                    "ORG_MEMBER"
                  ],
                  "roleMappings": [
                    {
                      "externalGroupName": "org-admin",
                      "roleAssignments": [
                        {
                          "role": "ORG_OWNER"
                        }
                      ]
                    },
                    {
                      "externalGroupName": "dev-team",
                      "roleAssignments": [
                        {
                          "role": "ORG_GROUP_CREATOR"
                        },
                        {
                          "projectName": "dev-project",
                          "role": "GROUP_OWNER"
                        }
                      ]
                    }
                  ],
                  "ssoDebugEnabled": false
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasIPAccessList",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasipaccesslist-sample"
                },
                "spec": {
                  "entries": [
                    {
                      "cidrBlock": "192.168.1.0/24"
                    },
                    {
                      "deleteAfterDate": "2025-01-20T23:59:59+02:00",
                      "ipAddress": "10.0.0.1"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkContainer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasnetworkcontainer-sample"
                },
                "spec": {
                  "cidrBlock": "10.11.0.0/16",
                  "connectionSecret": {
                    "name": "atlas-connection-secret"
                  },
                  "externalProjectRef": {
                    "projectId": "66e2f2b621571b7e69a89b66"
                  },
                  "provider": "AWS",
                  "region": "US_EAST_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkPeering",
                "metadata": {
                  "name": "atlasnetworkpeering-sample"
                },
                "spec": {
                  "atlasCidrBlock": "10.12.0.0/16",
                  "awsConfiguration": {
                    "accepterRegionName": "US_EAST_1",
                    "awsAccountId": "23214235145",
                    "routeTableCidrBlock": "10.11.0.0/16",
                    "vpcId": "vpcid-7123782134"
                  },
                  "containerId": "623412394512350",
                  "containerRegion": "US_EAST_1",
                  "projectRef": {
                    "name": "atlas-project",
                    "namespace": "namespace"
                  },
                  "provider": "AWS"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasPrivateEndpoint",
                "metadata": {
                  "name": "atlasprivateendpoint-sample"
                },
                "spec": {
                  "awsConfiguration": {
                    "id": "vpce-f4k34w51d"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "provider": "AWS",
                  "region": "EU_CENTRAL_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    },
                    {
                      "cidrBlock": "203.0.113.0/24",
                      "comment": "CIDR block for Application Server B - D"
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasSearchIndexConfig",
                "metadata": {
                  "name": "atlassearchindexconfig-sample"
                },
                "spec": {
                  "analyzer": "lucene.standard",
                  "analyzers": [
                    {
                      "name": "test-analyzer",
                      "tokenFilters": [
                        {
                          "asciiFolding": {
                            "originalTokens": "omit"
                          },
                          "type": "asciiFolding"
                        }
                      ],
                      "tokenizer": {
                        "type": "whiteSpace",
                        "whitespace": {
                          "maxTokenLenght": 200
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamConnection",
                "metadata": {
                  "name": "my-streamconnection-kafka"
                },
                "spec": {
                  "kafkaConfig": {
                    "authentication": {
                      "credentials": {
                        "name": "kafka-userpass"
                      },
                      "mechanism": "SCRAM-512"
                    },
                    "bootstrapServers": "kafka.server1:9001,kafka.server2:9002,kafka.server3:9003",
                    "security": {
                      "certificate": {
                        "name": "kafka-cert"
                      },
                      "protocol": "SSL"
                    }
                  },
                  "name": "kafka-config",
                  "type": "Kafka"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamInstance",
                "metadata": {
                  "name": "my-streaminstance-sample"
                },
                "spec": {
                  "clusterConfig": {
                    "provider": "AWS",
                    "region": "VIRGINIA_USA",
                    "tier": "SP30"
                  },
                  "connectionRegistry": [
                    {
                      "name": "my-streamconnection-sample"
                    }
                  ],
                  "name": "my-instance",
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasTeam",
                "metadata": {
                  "name": "atlasteam-sample"
                },
                "spec": {
                  "name": "my-team",
                  "usernames": [
                    "username@mongodb.com"
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasThirdPartyIntegration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasintegration-sample"
                },
                "spec": {
                  "datadog": {
                    "apiKeySecretRef": {
                      "name": "secret-ref"
                    },
                    "region": "string",
                    "sendCollectionLatencyMetrics": "string",
                    "sendDatabaseMetrics": "string"
                  },
                  "projectRef": {
                    "name": "my-atlas-project"
                  },
                  "type": "DATADOG"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:8ac3f6b1c0f31af5f32562ddeed37dd6586ac47e1e7e111947df3ed570c5bed9
          createdAt: "2025-09-25T14:34:34Z"
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/mongodb/mongodb-atlas-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The AtlasBackupCompliancePolicy is a configuration that enforces
              specific backup and retention requirements
            displayName: Atlas Backup Compliance Policy
            kind: AtlasBackupCompliancePolicy
            name: atlasbackupcompliancepolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasCustomRole is the Schema for the AtlasCustomRole API
            displayName: Atlas Custom Role
            kind: AtlasCustomRole
            name: atlascustomroles.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDataFederation is the Schema for the Atlas Data Federation
              API
            displayName: Atlas Data Federation
            kind: AtlasDataFederation
            name: atlasdatafederations.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasdeployments API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasFederatedAuth is the Schema for the Atlasfederatedauth
              API
            displayName: Atlas Federated Auth
            kind: AtlasFederatedAuth
            name: atlasfederatedauths.atlas.mongodb.com
            version: v1
          - description: AtlasIPAccessList is the Schema for the atlasipaccesslists
              API.
            displayName: Atlas IPAccess List
            kind: AtlasIPAccessList
            name: atlasipaccesslists.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkContainer is the Schema for the AtlasNetworkContainer
              API
            displayName: Atlas Network Container
            kind: AtlasNetworkContainer
            name: atlasnetworkcontainers.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkPeering is the Schema for the AtlasNetworkPeering
              API
            displayName: Atlas Network Peering
            kind: AtlasNetworkPeering
            name: atlasnetworkpeerings.atlas.mongodb.com
            version: v1
          - displayName: Atlas Org Settings
            kind: AtlasOrgSettings
            name: atlasorgsettings.atlas.mongodb.com
            version: v1
          - description: "The AtlasPrivateEndpoint custom resource definition (CRD)
              defines a desired [Private Endpoint](https://www.mongodb.com/docs/atlas/security-private-endpoint/#std-label-private-endpoint-overview)
              configuration for an Atlas project. It allows a private connection between
              your cloud provider and Atlas that doesn't send information through
              a public network. \n You can use private endpoints to create a unidirectional
              connection to Atlas clusters from your virtual network."
            displayName: Atlas Private Endpoint
            kind: AtlasPrivateEndpoint
            name: atlasprivateendpoints.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
          - description: Atlas Search Index Config is the Schema for the Atlas Search
              API
            displayName: Atlas Search Index Config
            kind: AtlasSearchIndexConfig
            name: atlassearchindexconfigs.atlas.mongodb.com
            version: v1
          - description: AtlasStreamConnection is the Schema for the atlasstreamconnections
              API
            displayName: Atlas Stream Connection
            kind: AtlasStreamConnection
            name: atlasstreamconnections.atlas.mongodb.com
            version: v1
          - description: AtlasStreamInstance is the Schema for the atlasstreaminstances
              API
            displayName: Atlas Stream Instance
            kind: AtlasStreamInstance
            name: atlasstreaminstances.atlas.mongodb.com
            version: v1
          - description: AtlasTeam is the Schema for the Atlas Teams API
            displayName: Atlas Team
            kind: AtlasTeam
            name: atlasteams.atlas.mongodb.com
            version: v1
          - description: AtlasThirdPartyIntegration is the Schema for the atlas 3rd
              party integrations API.
            displayName: Atlas Third Party Integration
            kind: AtlasThirdPartyIntegration
            name: atlasthirdpartyintegrations.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB
          Atlas — the only multi-cloud document database service that gives you the versatility you need to build sophisticated
          and resilient applications that can adapt to changing customer demands and market trends.

          The full documentation for the Operator can be found [here](https://docs.atlas.mongodb.com/atlas-operator/)

          ## Getting Started

          ### Supported features

          * Create and configure a project, or connect to an existing one.
          * Deploy, manage, scale, and tear down clusters.
          * Support for serverless instances.
          * Create and edit database users.
          * Manage IP Access Lists, network peering and private endpoints.
          * Configure and control Atlas’s fully managed cloud backup.
          * Configure federated authentication for your Atlas organization
          * Integrate Atlas monitoring with Prometheus.

          ... and more.

          To view the list of custom resources and each CR's schema, visit our [reference](https://www.mongodb.com/docs/atlas/operator/stable/custom-resources/)
          documentation. See the [Quickstart](https://www.mongodb.com/docs/atlas/operator/stable/ak8so-quick-start/) to get
          started with Atlas Kubernetes Operator.
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:8ac3f6b1c0f31af5f32562ddeed37dd6586ac47e1e7e111947df3ed570c5bed9
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:9eb6465e723ede43ac47eed6de22042d26c0efb276301a4424a6a3948a8f5b15
        version: 2.11.0
      entries:
      - name: mongodb-atlas-kubernetes.v2.11.0
        version: 2.11.0
      - name: mongodb-atlas-kubernetes.v2.10.0
        version: 2.10.0
      - name: mongodb-atlas-kubernetes.v2.9.1
        version: 2.9.1
      - name: mongodb-atlas-kubernetes.v2.9.0
        version: 2.9.0
      - name: mongodb-atlas-kubernetes.v2.8.2
        version: 2.8.2
      - name: mongodb-atlas-kubernetes.v2.8.1
        version: 2.8.1
      - name: mongodb-atlas-kubernetes.v2.8.0
        version: 2.8.0
      - name: mongodb-atlas-kubernetes.v2.7.1
        version: 2.7.1
      - name: mongodb-atlas-kubernetes.v2.7.0
        version: 2.7.0
      - name: mongodb-atlas-kubernetes.v2.6.1
        version: 2.6.1
      - name: mongodb-atlas-kubernetes.v2.6.0
        version: 2.6.0
      - name: mongodb-atlas-kubernetes.v2.5.0
        version: 2.5.0
      - name: mongodb-atlas-kubernetes.v2.4.1
        version: 2.4.1
      - name: mongodb-atlas-kubernetes.v2.4.0
        version: 2.4.0
      - name: mongodb-atlas-kubernetes.v2.3.1
        version: 2.3.1
      - name: mongodb-atlas-kubernetes.v2.3.0
        version: 2.3.0
      - name: mongodb-atlas-kubernetes.v2.2.2
        version: 2.2.2
      - name: mongodb-atlas-kubernetes.v2.2.1
        version: 2.2.1
      - name: mongodb-atlas-kubernetes.v2.2.0
        version: 2.2.0
      - name: mongodb-atlas-kubernetes.v2.1.0
        version: 2.1.0
      - name: mongodb-atlas-kubernetes.v2.0.1
        version: 2.0.1
      - name: mongodb-atlas-kubernetes.v2.0.0
        version: 2.0.0
      - name: mongodb-atlas-kubernetes.v1.9.1
        version: 1.9.1
      - name: mongodb-atlas-kubernetes.v1.9.0
        version: 1.9.0
      - name: mongodb-atlas-kubernetes.v1.8.2
        version: 1.8.2
      - name: mongodb-atlas-kubernetes.v1.8.1
        version: 1.8.1
      - name: mongodb-atlas-kubernetes.v1.8.0
        version: 1.8.0
      - name: mongodb-atlas-kubernetes.v1.7.3
        version: 1.7.3
      - name: mongodb-atlas-kubernetes.v1.7.2
        version: 1.7.2
      - name: mongodb-atlas-kubernetes.v1.7.1
        version: 1.7.1
      - name: mongodb-atlas-kubernetes.v1.7.0
        version: 1.7.0
      - name: mongodb-atlas-kubernetes.v1.6.1
        version: 1.6.1
      - name: mongodb-atlas-kubernetes.v1.6.0
        version: 1.6.0
      - name: mongodb-atlas-kubernetes.v1.5.1
        version: 1.5.1
      - name: mongodb-atlas-kubernetes.v1.5.0
        version: 1.5.0
      - name: mongodb-atlas-kubernetes.v1.4.1
        version: 1.4.1
      - name: mongodb-atlas-kubernetes.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-atlas-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco AppDynamics
      provider-url: https://www.appdynamics.com
    name: appdynamics-cloud-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appdynamics-cloud-operator.v24.7.0-133
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clustermon",
                "metadata": {
                  "name": "clustermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "clusterName": "test",
                  "image": "appdynamics/appdynamics-cloud-k8s-monitoring:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Containermon",
                "metadata": {
                  "name": "containermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "name": "ContainerMon"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "DbCollector",
                "metadata": {
                  "name": "dbcollector",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-db-collector:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "DbConfig",
                "metadata": {
                  "name": "dbconfig",
                  "namespace": "appdynamics"
                },
                "spec": null
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Inframon",
                "metadata": {
                  "name": "inframon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-k8s-monitoring:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "LogCollector",
                "metadata": {
                  "name": "logcollector-sample",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-log-collector-agent:latest",
                  "resources": {
                    "limits": {
                      "cpu": "400m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "150Mi"
                    }
                  }
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Servermon",
                "metadata": {
                  "name": "servermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "name": "ServerMon"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-08-06T08:33:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clustermon is the Schema for the clustermons API
            displayName: Clustermon
            kind: Clustermon
            name: clustermons.cluster.appdynamics.com
            version: v1alpha1
          - description: Containermon is the Schema for the containermons API
            displayName: Containermon
            kind: Containermon
            name: containermons.cluster.appdynamics.com
            version: v1alpha1
          - kind: DbCollector
            name: dbcollectors.cluster.appdynamics.com
            version: v1alpha1
          - kind: DbConfig
            name: dbconfigs.cluster.appdynamics.com
            version: v1alpha1
          - description: Inframon is the Schema for the inframons API
            displayName: Inframon
            kind: Inframon
            name: inframons.cluster.appdynamics.com
            version: v1alpha1
          - kind: LogCollector
            name: logcollectors.cluster.appdynamics.com
            version: v1alpha1
          - description: Servermon is the Schema for the servermons API
            displayName: Servermon
            kind: Servermon
            name: servermons.cluster.appdynamics.com
            version: v1alpha1
        description: The Appdynamics Cloud Operator is used to install Cluster Collector,
          Container Collector, Server Collector, Log Collector and Cluster Agent
        displayName: Cisco AppDynamics Cloud Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - appdynamics
        links:
        - name: Appdynamics Cloud Operator
          url: https://appdynamics-cloud-operator.domain
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: Cisco AppDynamics
          url: https://www.appdynamics.com
        relatedImages:
        - registry.connect.redhat.com/appdynamics/appdynamics-cloud-operator@sha256:2710d63425942f60af21f060a02eb3c40f5a5f0c952ea6593b3b9769c7ca1014
        version: 24.7.0-133
      entries:
      - name: appdynamics-cloud-operator.v24.7.0-133
        version: 24.7.0-133
      name: alpha
    defaultChannel: alpha
    packageName: appdynamics-cloud-operator
    provider:
      name: Cisco AppDynamics
      url: https://www.appdynamics.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: smb-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smb-csi-driver-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9-operator@sha256:1019812a3969f844a0f3fcd45091fccc7d0d816207e230a2f399556c41ebc705
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure CIFS/SMB CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.15.0-0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/csi-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          CIFS/SMB CSI Driver Operator provides Server Message Block (SMB) CSI Driver that enables you to create and mount CIFS/SMB PersistentVolumes.
        displayName: CIFS/SMB CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - smb
        - cifs
        links:
        - name: Documentation
          url: https://github.com/openshift/csi-operator
        - name: Source Repository
          url: https://github.com/openshift/csi-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: preview
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9-operator@sha256:1019812a3969f844a0f3fcd45091fccc7d0d816207e230a2f399556c41ebc705
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0627c09fe3f5a015f117fb30af1f4d17340aefa48b23994737d3901a804b63d3
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:a5704d26d0b562dba9633d1c107f34e05c750c0e3a64dbcace4dab7b964fff8d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9@sha256:9f516d06c279cf1f82e8948ce2290395eec5f6611abf5ea919e4d6548d81ec10
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:9c3c3d2b857ce289324aee56df1ba4ab5e5ea1e76cd86ac988233463915834b4
        version: 4.16.0-202509240123
      entries:
      - name: smb-csi-driver-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: smb-csi-driver-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: smb-csi-driver-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: smb-csi-driver-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: smb-csi-driver-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: smb-csi-driver-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: smb-csi-driver-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: smb-csi-driver-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: smb-csi-driver-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: smb-csi-driver-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: smb-csi-driver-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: smb-csi-driver-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: smb-csi-driver-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: smb-csi-driver-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: smb-csi-driver-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: smb-csi-driver-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: smb-csi-driver-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: smb-csi-driver-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: smb-csi-driver-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: smb-csi-driver-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: smb-csi-driver-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: smb-csi-driver-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: smb-csi-driver-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: smb-csi-driver-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: smb-csi-driver-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: smb-csi-driver-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: smb-csi-driver-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: smb-csi-driver-operator.v4.16.0-202406251906
        version: 4.16.0-202406251906
      - name: smb-csi-driver-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: smb-csi-driver-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: preview
    defaultChannel: preview
    packageName: smb-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: node-maintenance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-maintenance-operator.v5.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:58846ab62346b1004e11f36bd901c96f8d315958eb39fbb272d8b328faebcf04
          createdAt: "2023-10-20 15:36:51"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=4.10.0 <5.0.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-node-maintenance-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:58846ab62346b1004e11f36bd901c96f8d315958eb39fbb272d8b328faebcf04
        version: 5.0.1
      entries:
      - name: node-maintenance-operator.v5.0.1
        version: 5.0.1
      name: 4.12-eus
    - currentCSV: node-maintenance-operator.v5.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:2efd1976d2fb110391e0e5937b90476d3eeaacb4dd29cd52375f2b8566f17b41
          createdAt: "2023-10-20 15:50:49"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=5.0.0 <5.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:2efd1976d2fb110391e0e5937b90476d3eeaacb4dd29cd52375f2b8566f17b41
        version: 5.2.1
      entries:
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.14-eus
    - currentCSV: node-maintenance-operator.v5.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:7ca5853885534198544d23067087d80047b0e1421903085a2cb328bd00def147
          createdAt: "2024-08-10 19:55:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.2.0 <5.3.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:7ca5853885534198544d23067087d80047b0e1421903085a2cb328bd00def147
        version: 5.3.1
      entries:
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.16-eus
    - currentCSV: node-maintenance-operator.v5.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-10 10:59:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.3.0 <5.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.4/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel9-operator@sha256:752bff8bebcaf3df1f9d126dc4b07b622a44b02bfcf0f82b9100a7693627079a
        version: 5.4.1
      entries:
      - name: node-maintenance-operator.v5.4.1
        version: 5.4.1
      - name: node-maintenance-operator.v5.4.0
        version: 5.4.0
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.18-eus
    - currentCSV: node-maintenance-operator.v5.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-10 10:59:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.3.0 <5.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.4/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel9-operator@sha256:752bff8bebcaf3df1f9d126dc4b07b622a44b02bfcf0f82b9100a7693627079a
        version: 5.4.1
      entries:
      - name: node-maintenance-operator.v5.4.1
        version: 5.4.1
      - name: node-maintenance-operator.v5.4.0
        version: 5.4.0
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: stable
    defaultChannel: stable
    packageName: node-maintenance-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sagemaker-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sagemaker-controller.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "App",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "DataQualityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Domain",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Endpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "EndpointConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "FeatureGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "HyperParameterTuningJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Model",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelBiasJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelExplainabilityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelPackage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelPackageGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelQualityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "MonitoringSchedule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "NotebookInstance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "NotebookInstanceLifecycleConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ProcessingJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "TrainingJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "TransformJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "UserProfile",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sagemaker-controller:1.4.1
          createdAt: "2025-09-22T19:07:36Z"
          description: ACK service controller for Amazon SageMaker manages SageMaker
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: App represents the state of an AWS sagemaker App resource.
            displayName: App
            kind: App
            name: apps.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: DataQualityJobDefinition represents the state of an AWS sagemaker
              DataQualityJobDefinition resource.
            displayName: DataQualityJobDefinition
            kind: DataQualityJobDefinition
            name: dataqualityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Domain represents the state of an AWS sagemaker Domain resource.
            displayName: Domain
            kind: Domain
            name: domains.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: EndpointConfig represents the state of an AWS sagemaker EndpointConfig
              resource.
            displayName: EndpointConfig
            kind: EndpointConfig
            name: endpointconfigs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Endpoint represents the state of an AWS sagemaker Endpoint
              resource.
            displayName: Endpoint
            kind: Endpoint
            name: endpoints.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: FeatureGroup represents the state of an AWS sagemaker FeatureGroup
              resource.
            displayName: FeatureGroup
            kind: FeatureGroup
            name: featuregroups.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: HyperParameterTuningJob represents the state of an AWS sagemaker
              HyperParameterTuningJob resource.
            displayName: HyperParameterTuningJob
            kind: HyperParameterTuningJob
            name: hyperparametertuningjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: InferenceComponent represents the state of an AWS sagemaker
              InferenceComponent resource.
            displayName: InferenceComponent
            kind: InferenceComponent
            name: inferencecomponents.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: LabelingJob represents the state of an AWS sagemaker LabelingJob
              resource.
            displayName: LabelingJob
            kind: LabelingJob
            name: labelingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelBiasJobDefinition represents the state of an AWS sagemaker
              ModelBiasJobDefinition resource.
            displayName: ModelBiasJobDefinition
            kind: ModelBiasJobDefinition
            name: modelbiasjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelExplainabilityJobDefinition represents the state of
              an AWS sagemaker ModelExplainabilityJobDefinition resource.
            displayName: ModelExplainabilityJobDefinition
            kind: ModelExplainabilityJobDefinition
            name: modelexplainabilityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelPackageGroup represents the state of an AWS sagemaker
              ModelPackageGroup resource.
            displayName: ModelPackageGroup
            kind: ModelPackageGroup
            name: modelpackagegroups.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelPackage represents the state of an AWS sagemaker ModelPackage
              resource.
            displayName: ModelPackage
            kind: ModelPackage
            name: modelpackages.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelQualityJobDefinition represents the state of an AWS
              sagemaker ModelQualityJobDefinition resource.
            displayName: ModelQualityJobDefinition
            kind: ModelQualityJobDefinition
            name: modelqualityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Model represents the state of an AWS sagemaker Model resource.
            displayName: Model
            kind: Model
            name: models.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: MonitoringSchedule represents the state of an AWS sagemaker
              MonitoringSchedule resource.
            displayName: MonitoringSchedule
            kind: MonitoringSchedule
            name: monitoringschedules.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: NotebookInstanceLifecycleConfig represents the state of an
              AWS sagemaker NotebookInstanceLifecycleConfig resource.
            displayName: NotebookInstanceLifecycleConfig
            kind: NotebookInstanceLifecycleConfig
            name: notebookinstancelifecycleconfigs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: NotebookInstance represents the state of an AWS sagemaker
              NotebookInstance resource.
            displayName: NotebookInstance
            kind: NotebookInstance
            name: notebookinstances.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: PipelineExecution represents the state of an AWS sagemaker
              PipelineExecution resource.
            displayName: PipelineExecution
            kind: PipelineExecution
            name: pipelineexecutions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Pipeline represents the state of an AWS sagemaker Pipeline
              resource.
            displayName: Pipeline
            kind: Pipeline
            name: pipelines.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ProcessingJob represents the state of an AWS sagemaker ProcessingJob
              resource.
            displayName: ProcessingJob
            kind: ProcessingJob
            name: processingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: TrainingJob represents the state of an AWS sagemaker TrainingJob
              resource.
            displayName: TrainingJob
            kind: TrainingJob
            name: trainingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: TransformJob represents the state of an AWS sagemaker TransformJob
              resource.
            displayName: TransformJob
            kind: TransformJob
            name: transformjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: UserProfile represents the state of an AWS sagemaker UserProfile
              resource.
            displayName: UserProfile
            kind: UserProfile
            name: userprofiles.sagemaker.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon SageMaker resources in AWS from within your Kubernetes cluster.

          **About Amazon SageMaker**

          Amazon SageMaker helps data scientists and developers to prepare, build, train, and deploy high-quality machine learning (ML) models quickly by bringing together a broad set of capabilities purpose-built for ML.

          For more information on Amazon SageMaker, visit the [product page](https://aws.amazon.com/sagemaker/).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SageMaker
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sagemaker
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SageMaker Developer Resources
          url: https://aws.amazon.com/sagemaker/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sagemaker maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sagemaker-controller:1.4.1
        version: 1.4.1
      entries:
      - name: ack-sagemaker-controller.v1.4.1
        version: 1.4.1
      - name: ack-sagemaker-controller.v1.4.0
        version: 1.4.0
      - name: ack-sagemaker-controller.v1.3.3
        version: 1.3.3
      - name: ack-sagemaker-controller.v1.3.2
        version: 1.3.2
      - name: ack-sagemaker-controller.v1.3.1
        version: 1.3.1
      - name: ack-sagemaker-controller.v1.3.0
        version: 1.3.0
      - name: ack-sagemaker-controller.v1.2.20
        version: 1.2.20
      - name: ack-sagemaker-controller.v1.2.19
        version: 1.2.19
      - name: ack-sagemaker-controller.v1.2.18
        version: 1.2.18
      - name: ack-sagemaker-controller.v1.2.17
        version: 1.2.17
      - name: ack-sagemaker-controller.v1.2.16
        version: 1.2.16
      - name: ack-sagemaker-controller.v1.2.13
        version: 1.2.13
      - name: ack-sagemaker-controller.v1.2.12
        version: 1.2.12
      - name: ack-sagemaker-controller.v1.2.11
        version: 1.2.11
      - name: ack-sagemaker-controller.v1.2.9
        version: 1.2.9
      - name: ack-sagemaker-controller.v1.2.8
        version: 1.2.8
      - name: ack-sagemaker-controller.v1.2.5
        version: 1.2.5
      - name: ack-sagemaker-controller.v1.2.4
        version: 1.2.4
      - name: ack-sagemaker-controller.v1.2.3
        version: 1.2.3
      - name: ack-sagemaker-controller.v1.2.2
        version: 1.2.2
      - name: ack-sagemaker-controller.v1.2.1
        version: 1.2.1
      - name: ack-sagemaker-controller.v1.2.0
        version: 1.2.0
      - name: ack-sagemaker-controller.v1.1.0
        version: 1.1.0
      - name: ack-sagemaker-controller.v1.0.0
        version: 1.0.0
      - name: ack-sagemaker-controller.v0.6.0
        version: 0.6.0
      - name: ack-sagemaker-controller.v0.5.1
        version: 0.5.1
      - name: ack-sagemaker-controller.v0.5.0
        version: 0.5.0
      - name: ack-sagemaker-controller.v0.4.5
        version: 0.4.5
      - name: ack-sagemaker-controller.v0.4.4
        version: 0.4.4
      - name: ack-sagemaker-controller.v0.4.3
        version: 0.4.3
      - name: ack-sagemaker-controller.v0.4.2
        version: 0.4.2
      - name: ack-sagemaker-controller.v0.4.0
        version: 0.4.0
      - name: ack-sagemaker-controller.v0.3.4
        version: 0.3.4
      - name: ack-sagemaker-controller.v0.3.3
        version: 0.3.3
      - name: ack-sagemaker-controller.v0.3.2
        version: 0.3.2
      - name: ack-sagemaker-controller.v0.3.1
        version: 0.3.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-sagemaker-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudwatch-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudwatch-controller.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudwatch-controller:1.3.0
          createdAt: "2025-10-02T23:05:17Z"
          description: AWS CloudWatch controller is a service controller for managing
            CloudWatch resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Dashboard represents the state of an AWS cloudwatch Dashboard
              resource.
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.cloudwatch.services.k8s.aws
            version: v1alpha1
          - description: MetricAlarm represents the state of an AWS cloudwatch MetricAlarm
              resource.
            displayName: MetricAlarm
            kind: MetricAlarm
            name: metricalarms.cloudwatch.services.k8s.aws
            version: v1alpha1
          - description: MetricStream represents the state of an AWS cloudwatch MetricStream
              resource.
            displayName: MetricStream
            kind: MetricStream
            name: metricstreams.cloudwatch.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudWatch resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudWatch**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudWatch
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudwatch
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudWatch Developer Resources
          url: https://aws.amazon.com/CloudWatch/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudwatch maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudwatch-controller:1.3.0
        version: 1.3.0
      entries:
      - name: ack-cloudwatch-controller.v1.3.0
        version: 1.3.0
      - name: ack-cloudwatch-controller.v1.2.1
        version: 1.2.1
      - name: ack-cloudwatch-controller.v1.2.0
        version: 1.2.0
      - name: ack-cloudwatch-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudwatch-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudwatch-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudwatch-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudwatch-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudwatch-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudwatch-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudwatch-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudwatch-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudwatch-controller.v1.0.5
        version: 1.0.5
      - name: ack-cloudwatch-controller.v1.0.4
        version: 1.0.4
      - name: ack-cloudwatch-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudwatch-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudwatch-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudwatch-controller.v0.0.12
        version: 0.0.12
      - name: ack-cloudwatch-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudwatch-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudwatch-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudwatch-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudwatch-controller.v0.0.3
        version: 0.0.3
      - name: ack-cloudwatch-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudwatch-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OpenShift Database Access
      provider-url: ""
    name: dbaas-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dbaas-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSConnection",
                "metadata": {
                  "name": "dbaasconnection-example"
                },
                "spec": {
                  "databaseServiceID": "instance-example",
                  "databaseServiceType": "instance",
                  "inventoryRef": {
                    "name": "inventory-example",
                    "namespace": "openshift-dbaas-operator"
                  }
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSInstance",
                "metadata": {
                  "name": "dbaasinstance-sample"
                },
                "spec": {
                  "cloudProvider": "AWS",
                  "cloudRegion": "US_EAST_1",
                  "inventoryRef": {
                    "name": "inventory-example",
                    "namespace": "openshift-dbaas-operator"
                  },
                  "name": "my-cluster-free"
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSInventory",
                "metadata": {
                  "name": "inventory-example"
                },
                "spec": {
                  "credentialsRef": {
                    "name": "dbaas-vendor-credentials"
                  },
                  "providerRef": {
                    "name": "provider-example-registration"
                  }
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSPolicy",
                "metadata": {
                  "name": "dbaaspolicy-sample"
                },
                "spec": {
                  "disableProvisions": false
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSProvider",
                "metadata": {
                  "labels": {
                    "related-to": "dbaas-operator",
                    "type": "dbaas-provider-registration"
                  },
                  "name": "provider-example-registration"
                },
                "spec": {
                  "allowsFreeTrial": true,
                  "connectionKind": "ProviderConnection",
                  "credentialFields": [
                    {
                      "displayName": "ProviderAPIKeyD",
                      "helpText": "This is the API Key for the example provider to show on UI",
                      "key": "APIKey",
                      "required": true,
                      "type": "maskedstring"
                    }
                  ],
                  "externalProvisionDescription": "Follow the guide to start a free Provider Serverless (beta) cluster",
                  "externalProvisionURL": "https://www.exmample.com/docs/provider/quickstart.html",
                  "groupVersion": "dbaas.redhat.com/v1beta1",
                  "instanceKind": "ProviderInstance",
                  "inventoryKind": "ProviderInventory",
                  "provider": {
                    "displayDescription": "This is an example for providers on how to implement their operator to integrate with DBaaS.",
                    "displayName": "DBaaS Provider Example ",
                    "icon": {
                      "base64data": "SGVsbG8sIHdvcmxkLg==",
                      "mediatype": "image/png"
                    },
                    "name": "Provider Example"
                  },
                  "provisioningParameters": {
                    "cloudProvider": {
                      "conditionalData": [
                        {
                          "defaultValue": "GCP",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "FREETRIAL"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        },
                        {
                          "defaultValue": "AWS",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "SERVERLESS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Amazon Web Services",
                              "value": "AWS"
                            },
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        },
                        {
                          "defaultValue": "AWS",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Amazon Web Services",
                              "value": "AWS"
                            },
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        }
                      ],
                      "displayName": "Cloud Provider"
                    },
                    "dedicatedLocationLabel": {
                      "displayName": "Select regions \u0026 nodes",
                      "helpText": "Select the geographical region where you want the database instance to run, and set the number of nodes you want running in this dedicated cluster."
                    },
                    "hardwareLabel": {
                      "displayName": "Hardware per node",
                      "helpText": "Select the compute and storage requirements for this database instance."
                    },
                    "machineType": {
                      "conditionalData": [
                        {
                          "defaultValue": "m5.large",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "AWS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "2 vCPU, 8 GiB RAM",
                              "value": "m5.large"
                            },
                            {
                              "displayValue": "4 vCPU, 16 GiB RAM",
                              "value": "m5.xlarge"
                            },
                            {
                              "displayValue": "8 vCPU, 32 GiB RAM",
                              "value": "m5.2xlarge"
                            },
                            {
                              "displayValue": "16 vCPU, 64 GiB RAM",
                              "value": "m5.4xlarge"
                            },
                            {
                              "displayValue": "32 vCPU, 128 GiB RAM",
                              "value": "m5.8xlarge"
                            }
                          ]
                        },
                        {
                          "defaultValue": "n1-standard-2",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "GCP"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "2 vCPU, 7.5 GiB RAM",
                              "value": "n1-standard-2"
                            },
                            {
                              "displayValue": "4 vCPU, 15 GiB RAM",
                              "value": "n1-standard-4"
                            },
                            {
                              "displayValue": "8 vCPU, 30 GiB RAM",
                              "value": "n1-standard-8"
                            },
                            {
                              "displayValue": "16 vCPU, 60 GiB RAM",
                              "value": "n1-standard-16"
                            },
                            {
                              "displayValue": "32 vCPU, 120 GiB RAM",
                              "value": "n1-standard-32"
                            }
                          ]
                        }
                      ],
                      "displayName": "Compute"
                    },
                    "name": {
                      "displayName": "Cluster name"
                    },
                    "nodes": {
                      "displayName": "nodes"
                    },
                    "plan": {
                      "conditionalData": [
                        {
                          "defaultValue": "SERVERLESS",
                          "options": [
                            {
                              "displayValue": "Free trial",
                              "value": "FREETRIAL"
                            },
                            {
                              "displayValue": "Serverless",
                              "value": "SERVERLESS"
                            },
                            {
                              "displayValue": "Dedicated",
                              "value": "DEDICATED"
                            }
                          ]
                        }
                      ],
                      "displayName": "Hosting plan"
                    },
                    "planLabel": {
                      "displayName": "Select a plan"
                    },
                    "regions": {
                      "displayName": "us-west-2"
                    },
                    "serverlessLocationLabel": {
                      "displayName": "Select regions",
                      "helpText": "Select the geographical region where you want the database instance to run."
                    },
                    "spendLimit": {
                      "conditionalData": [
                        {
                          "defaultValue": "0",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "SERVERLESS"
                            }
                          ]
                        }
                      ],
                      "displayName": "Spend limit"
                    },
                    "spendLimitLabel": {
                      "displayName": "Spend limit",
                      "helpText": "Set a spending limit on resources for this database instance."
                    },
                    "storageGib": {
                      "conditionalData": [
                        {
                          "defaultValue": "15",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "AWS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "15 GiB",
                              "value": "15"
                            },
                            {
                              "displayValue": "35 GiB",
                              "value": "35"
                            },
                            {
                              "displayValue": "75 GiB",
                              "value": "75"
                            },
                            {
                              "displayValue": "150 GiB",
                              "value": "150"
                            },
                            {
                              "displayValue": "300 GiB",
                              "value": "300"
                            },
                            {
                              "displayValue": "600 GiB",
                              "value": "600"
                            }
                          ]
                        },
                        {
                          "defaultValue": "15",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "GCP"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "15 GiB",
                              "value": "15"
                            },
                            {
                              "displayValue": "35 GiB",
                              "value": "35"
                            },
                            {
                              "displayValue": "75 GiB",
                              "value": "75"
                            },
                            {
                              "displayValue": "150 GiB",
                              "value": "150"
                            },
                            {
                              "displayValue": "300 GiB",
                              "value": "300"
                            },
                            {
                              "displayValue": "600 GiB",
                              "value": "600"
                            }
                          ]
                        }
                      ],
                      "displayName": "Storage"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: quay.io/ecosystem-appeng/dbaas-operator:v0.5.0
          description: Discover & connect with database instances hosted on 3rd-party
            ISV cloud platforms.
          operatorframework.io/suggested-namespace: openshift-dbaas-operator
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/internal-objects: '["dbaasconnections.dbaas.redhat.com","dbaasproviders.dbaas.redhat.com","dbaasplatforms.dbaas.redhat.com","dbaasinstances.dbaas.redhat.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/RHEcosystemAppEng/dbaas-operator
          support: rhoda-requests@redhat.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DBaaSConnection
            name: dbaasconnections.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSConnection defines the schema for the DBaaSConnection
              API.
            displayName: DBaaSConnection
            kind: DBaaSConnection
            name: dbaasconnections.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSInstance
            name: dbaasinstances.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSInstance defines the schema for the DBaaSInstance API.
            displayName: DBaaSInstance
            kind: DBaaSInstance
            name: dbaasinstances.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSInventory
            name: dbaasinventories.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSInventory defines the schema for the DBaaSInventory
              API. Inventory objects must be created in a valid namespace, determined
              by the existence of a DBaaSPolicy object.
            displayName: Provider Account
            kind: DBaaSInventory
            name: dbaasinventories.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSPlatform
            name: dbaasplatforms.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSPlatform defines the schema for the DBaaSPlatform API.
            displayName: DBaaSPlatform
            kind: DBaaSPlatform
            name: dbaasplatforms.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSPolicy
            name: dbaaspolicies.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSPolicy enables administrative capabilities within a
              namespace, and sets a default inventory policy. Policy defaults can
              be overridden on a per-inventory basis.
            displayName: Provider Account Policy
            kind: DBaaSPolicy
            name: dbaaspolicies.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSProvider
            name: dbaasproviders.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSProvider defines the schema for the DBaaSProvider API.
            displayName: DBaaSProvider
            kind: DBaaSProvider
            name: dbaasproviders.dbaas.redhat.com
            version: v1beta1
        description: |
          The OpenShift Database Access Operator enables OpenShift users to discover & connect with database instances
          hosted on 3rd-party ISV cloud platforms such as CrunchyData Bridge & CockroachCloud.

          ## Feedback & Support

          Please contact us at [rhoda-requests@redhat.com](mailto:rhoda-requests@redhat.com)

          For more information on how to use our service, please refer to our [Quick Starts](quickstart?keyword=database).
        displayName: OpenShift Database Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rhoda
        - oda
        - dbaas
        - database
        - db
        - data
        - access
        - crunchy
        - bridge
        - cockroach
        - cockroachdb
        - cloud
        - ack
        - amazon
        - rds
        - relational
        links:
        - name: GitHub
          url: https://github.com/RHEcosystemAppEng/dbaas-operator
        maintainers:
        - email: ecosystem-appeng@redhat.com
          name: Red Hat Ecosystem App Engineering
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: OpenShift Database Access
        relatedImages:
        - quay.io/openshift/origin-kube-rbac-proxy:4.10.0
        - quay.io/ecosystem-appeng/dbaas-dynamic-plugin:0.5.0
        - registry.developers.crunchydata.com/crunchydata/crunchy-bridge-operator-catalog:v0.0.6
        - gcr.io/cockroach-shared/ccapi-k8s-operator-catalog:v0.0.5
        - quay.io/ecosystem-appeng/rds-dbaas-operator-catalog:v0.4.0
        - quay.io/ecosystem-appeng/dbaas-operator:v0.5.0
        version: 0.5.0
      entries:
      - name: dbaas-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: dbaas-operator
    provider:
      name: OpenShift Database Access
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ControlPlane Ltd.
      provider-url: https://control-plane.io
    name: flux-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux-operator.v0.31.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "FluxInstance",
                "metadata": {
                  "name": "flux",
                  "namespace": "flux-system",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcileEvery": "1h",
                    "fluxcd.controlplane.io/reconcileTimeout": "5m"
                  }
                },
                "spec": {
                  "distribution": {
                    "version": "2.x",
                    "registry": "ghcr.io/fluxcd"
                  },
                  "components": [
                    "source-controller",
                    "kustomize-controller",
                    "helm-controller",
                    "notification-controller",
                    "image-reflector-controller",
                    "image-automation-controller"
                  ],
                  "cluster": {
                    "type": "openshift",
                    "size": "medium",
                    "domain": "cluster.local",
                    "networkPolicy": true,
                    "multitenant": true,
                    "tenantDefaultServiceAccount": "flux"
                  }
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "FluxReport",
                "metadata": {
                  "name": "flux",
                  "namespace": "flux-system"
                },
                "spec": {
                  "distribution": {
                    "entitlement": "Issued by controlplane",
                    "managedBy": "flux-operator",
                    "status": "Installed",
                    "version": "v2.3.0"
                  },
                "components": [
                    {
                      "image": "ghcr.io/fluxcd/kustomize-controller:v1.3.0@sha256:48a032574dd45c39750ba0f1488e6f1ae36756a38f40976a6b7a588d83acefc1",
                      "name": "kustomize-controller",
                      "ready": true,
                      "status": "Current Deployment is available. Replicas: 1"
                    },
                    {
                      "image": "ghcr.io/fluxcd/source-controller:v1.3.0@sha256:161da425b16b64dda4b3cec2ba0f8d7442973aba29bb446db3b340626181a0bc",
                      "name": "source-controller",
                      "ready": true,
                      "status": "Current Deployment is available. Replicas: 1"
                    }
                  ],
                "reconcilers": [
                    {
                      "apiVersion": "kustomize.toolkit.fluxcd.io/v1",
                      "kind": "Kustomization",
                      "stats": {
                        "failing": 0,
                        "running": 1,
                        "suspended": 0
                      }
                    },
                    {
                      "apiVersion": "source.toolkit.fluxcd.io/v1",
                      "kind": "GitRepository",
                      "stats": {
                        "failing": 0,
                        "running": 1,
                        "suspended": 0,
                        "totalSize": "3.7 MiB"
                      }
                    }
                  ],
                  "sync": {
                    "ready": true,
                    "id": "kustomization/flux-system",
                    "path": "clusters/production",
                    "source": "https://github.com/my-org/my-fleet.git",
                    "status": "Applied revision: refs/heads/main@sha1:a90cd1ac35de01c175f7199315d3f4cd60195911"
                  }
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "ResourceSet",
                "metadata": {
                  "name": "podinfo",
                  "namespace": "default",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcile": "enabled",
                    "fluxcd.controlplane.io/reconcileEvery": "30m",
                    "fluxcd.controlplane.io/reconcileTimeout": "5m"
                  }
                },
                "spec": {
                  "commonMetadata": {
                    "labels": {
                      "app.kubernetes.io/name": "podinfo"
                    }
                  },
                  "inputs": [
                    {
                      "tenant": "team1",
                      "version": "6.7.x",
                      "replicas": "2"
                    },
                    {
                      "tenant": "team2",
                      "version": "6.6.x",
                      "replicas": "3"
                    }
                  ],
                  "resources": [
                    {
                      "apiVersion": "source.toolkit.fluxcd.io/v1beta2",
                      "kind": "OCIRepository",
                      "metadata": {
                        "name": "podinfo-<< inputs.tenant >>",
                        "namespace": "default"
                      },
                      "spec": {
                        "interval": "10m",
                        "url": "oci://ghcr.io/stefanprodan/charts/podinfo",
                        "ref": {
                          "semver": "<< inputs.version | quote >>"
                        }
                      }
                    },
                    {
                      "apiVersion": "helm.toolkit.fluxcd.io/v2",
                      "kind": "HelmRelease",
                      "metadata": {
                        "name": "podinfo-<< inputs.tenant >>",
                        "namespace": "default"
                      },
                      "spec": {
                        "interval": "1h",
                        "releaseName": "podinfo-<< inputs.tenant >>",
                        "chartRef": {
                          "kind": "OCIRepository",
                          "name": "podinfo-<< inputs.tenant >>"
                        },
                        "values": {
                          "replicaCount": "<< inputs.replicas | int >>"
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "ResourceSetInputProvider",
                "metadata": {
                  "name": "demo",
                  "namespace": "default",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcile": "enabled",
                    "fluxcd.controlplane.io/reconcileEvery": "10m",
                    "fluxcd.controlplane.io/reconcileTimeout": "1m"
                  }
                },
                "spec": {
                  "type": "GitHubPullRequest",
                  "url": "https://github.com/fluxcd-testing/pr-testing",
                  "filter": {
                    "limit": 10,
                    "includeBranch": "^stefanprodan-patch-.*$",
                    "labels": [
                      "enhancement"
                    ]
                  },
                  "skip": {
                    "labels": [
                        "deploy/flux-preview-pause",
                        "!test-build-push/passed"
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/controlplaneio-fluxcd/flux-operator:v0.31.0-ubi
          createdAt: "2025-10-06T20:33:03Z"
          description: The Flux Operator manages the lifecycle of the CNCF Flux project.
          operatorframework.io/suggested-namespace: flux-system
          repository: https://github.com/controlplaneio-fluxcd/flux-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Flux Instance
            displayName: FluxInstance
            kind: FluxInstance
            name: fluxinstances.fluxcd.controlplane.io
            version: v1
          - description: Flux Report (Autogenerated)
            displayName: FluxReport
            kind: FluxReport
            name: fluxreports.fluxcd.controlplane.io
            version: v1
          - description: Resource Set
            displayName: ResourceSet
            kind: ResourceSet
            name: resourcesets.fluxcd.controlplane.io
            version: v1
          - description: ResourceSet Input Provider
            displayName: ResourceSetInputProvider
            kind: ResourceSetInputProvider
            name: resourcesetinputproviders.fluxcd.controlplane.io
            version: v1
        description: |-
          [Flux](https://fluxcd.io) is a leading GitOps Continuous Delivery tool used to streamline and automate application deployments on Kubernetes. Flux enables teams to achieve a reliable and auditable delivery process while promoting collaboration and traceability across different environments.
          Flux is powered by the GitOps Toolkit, a set of composable APIs and specialized tools that enable a wide range of continuous delivery use-cases, from simple Kubernetes deployment pipelines to multi-tenant and multi-cluster progressive delivery rollouts.
          For more information about the CNCF-graduated Flux project, please see the [Flux architecture overview](https://fluxcd.control-plane.io/guides/flux-architecture/).
          ### Operator Capabilities
          The [Flux Operator](https://github.com/controlplaneio-fluxcd/flux-operator) provides a declarative API for the installation, configuration and upgrade of CNCF Flux and the ControlPlane [enterprise distribution](https://fluxcd.control-plane.io/).
          The operator allows the configuration of Flux multi-tenancy lockdown, network policies, persistent storage, sharding, vertical scaling, custom patches, and the synchronization of the cluster state from Git repositories, OCI artifacts and S3-compatible storage.
          The Flux Operator supervises the Flux controllers and provides a unified view of all the Flux resources that define the GitOps workflows for the target cluster. The operator generates reports, emits events, and exports Prometheus metrics to help with monitoring and troubleshooting Flux.
          ### Operator Features
          The operator extends Flux with self-service capabilities, deployment windows, and preview environments for GitHub, GitLab and Azure DevOps pull requests testing.
          The operator ResourceSet API enables platform teams to define their own application standard as a group of Flux and Kubernetes resources that can be templated, parameterized and deployed as a single unit on self-service environments. For more information, please see the [ResourceSet documentation](https://fluxcd.control-plane.io/operator/resourcesets/introduction/).
          ### OpenShift Support
          The Flux Operator should be installed in a dedicated namespace, e.g. `flux-system`.
          To deploy Flux on OpenShift clusters, create a `FluxInstance` custom resource named `flux` in the same namespace as the operator and set the `.spec.cluster.type` field to `openshift`.
          For more information on how to configure the Flux instance, please see the [Flux Operator documentation](https://fluxcd.control-plane.io/operator/flux-config/).
        displayName: Flux Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - fluxcd
        - gitops
        - continuous delivery
        links:
        - name: Source Code
          url: https://github.com/controlplaneio-fluxcd/flux-operator
        - name: Documentation
          url: https://fluxcd.control-plane.io/operator/
        - name: Enterprise Support
          url: https://fluxcd.control-plane.io/pricing/
        maintainers:
        - email: flux-enterprise@control-plane.io
          name: ControlPlane Flux Team
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: ControlPlane Ltd.
          url: https://control-plane.io
        relatedImages:
        - ghcr.io/controlplaneio-fluxcd/flux-operator:v0.31.0-ubi
        version: 0.31.0
      entries:
      - name: flux-operator.v0.31.0
        version: 0.31.0
      - name: flux-operator.v0.30.0
        version: 0.30.0
      - name: flux-operator.v0.28.0
        version: 0.28.0
      - name: flux-operator.v0.27.0
        version: 0.27.0
      - name: flux-operator.v0.26.0
        version: 0.26.0
      - name: flux-operator.v0.24.1
        version: 0.24.1
      - name: flux-operator.v0.24.0
        version: 0.24.0
      - name: flux-operator.v0.23.0
        version: 0.23.0
      - name: flux-operator.v0.22.0
        version: 0.22.0
      - name: flux-operator.v0.21.0
        version: 0.21.0
      - name: flux-operator.v0.20.0
        version: 0.20.0
      - name: flux-operator.v0.19.0
        version: 0.19.0
      - name: flux-operator.v0.18.0
        version: 0.18.0
      name: stable
    defaultChannel: stable
    packageName: flux-operator
    provider:
      name: ControlPlane Ltd.
      url: https://control-plane.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-load-core
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-load-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Loadcore",
                "metadata": {
                  "name": "loadcore-sample"
                },
                "spec": {
                  "accessRoute": {
                    "enabled": true,
                    "host": "loadcore.apps.cluster.perf.keysight.lab"
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "reportTemplates": {
                    "enabled": false
                  },
                  "wirelessDataModel": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:42:44Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Loadcore
            name: loadcores.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight LoadCore Operator
        displayName: Keysight LoadCore Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Load Core
          url: https://keysight-load-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-load-core@sha256:245d9be60ff6e9a6e5e42a026a4ee7b8168f391a0491de23ae827d1906af88e5
        version: 0.1.21
      entries:
      - name: keysight-load-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-load-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CloudBees
      provider-url: https://www.cloudbees.com/products/continuous-integration
    name: cloudbees-ci-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudbees-ci.v3.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.cloudbees.com/v1alpha1",
                "kind": "CloudBeesCI",
                "metadata": {
                  "name": "cloudbeesci-sample"
                },
                "spec": {
                  "Agents": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-agent:2.289.3.2"
                    },
                    "ImagePullSecrets": null,
                    "SeparateNamespace": {
                      "Create": false,
                      "Enabled": false,
                      "Name": null
                    }
                  },
                  "Hibernation": {
                    "Enabled": false,
                    "Image": {
                      "dockerImage": "cloudbees/managed-master-hibernation-monitor:247.c5dfce00a179",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "NodeSelector": {},
                    "Tolerations": []
                  },
                  "Master": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-mm:2.289.3.2"
                    },
                    "JavaOpts": null,
                    "OperationsCenterNamespace": null
                  },
                  "NetworkPolicy": {
                    "Enabled": false,
                    "JMXSelectors": [],
                    "ingressControllerSelector": []
                  },
                  "OperationsCenter": {
                    "AgentListenerPort": 50000,
                    "Annotations": {},
                    "CSRF": {
                      "ProxyCompatibility": false
                    },
                    "CasC": {
                      "ConfigMapName": "oc-casc-bundle",
                      "Enabled": false
                    },
                    "ContainerPort": 8080,
                    "ContextPath": null,
                    "Enabled": true,
                    "ExtraConfigMaps": [],
                    "ExtraContainers": [],
                    "ExtraGroovyConfiguration": {},
                    "ExtraVolumeMounts": [],
                    "ExtraVolumes": [],
                    "HealthProbeLivenessFailureThreshold": 12,
                    "HealthProbes": true,
                    "HostName": null,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-cloud-core-oc:2.289.3.2",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "Ingress": {
                      "Annotations": {
                        "kubernetes.io/tls-acme": "false"
                      },
                      "Class": "nginx",
                      "tls": {
                        "Enable": false,
                        "SecretName": null
                      }
                    },
                    "JavaOpts": null,
                    "JenkinsOpts": null,
                    "LoadBalancerIP": null,
                    "LoadBalancerSourceRanges": [
                      "0.0.0.0/0"
                    ],
                    "Name": "cjoc",
                    "NodeSelector": {},
                    "Platform": "standard",
                    "Protocol": "http",
                    "Resources": {
                      "Limits": {
                        "Cpu": 1,
                        "Memory": "2G"
                      },
                      "Requests": {
                        "Cpu": 1,
                        "Memory": "2G"
                      }
                    },
                    "Route": {
                      "tls": {
                        "CACertificate": null,
                        "Certificate": null,
                        "DestinationCACertificate": null,
                        "Enable": false,
                        "InsecureEdgeTerminationPolicy": "Redirect",
                        "Key": null,
                        "Termination": "edge"
                      }
                    },
                    "ServiceAgentListenerPort": 50000,
                    "ServiceAnnotations": {},
                    "ServicePort": 80,
                    "ServiceType": "ClusterIP",
                    "Tolerations": []
                  },
                  "Persistence": {
                    "AccessMode": "ReadWriteOnce",
                    "Annotations": {},
                    "Size": "20Gi",
                    "StorageClass": null
                  },
                  "PodSecurityPolicy": {
                    "Annotations": {},
                    "Enabled": false
                  },
                  "Subdomain": false,
                  "ingress-nginx": {
                    "Enabled": false,
                    "controller": {
                      "admissionWebhooks": {
                        "patch": {
                          "nodeSelector": {
                            "kubernetes.io/os": "linux"
                          }
                        }
                      },
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "nginx-ingress": {
                    "Enabled": false,
                    "controller": {
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "rbac": {
                    "agentsServiceAccountName": "jenkins-agents",
                    "hibernationMonitorServiceAccountName": "managed-master-hibernation-monitor",
                    "install": true,
                    "masterServiceAccountName": "jenkins",
                    "serviceAccountName": "cjoc"
                  },
                  "sidecarinjector": {
                    "Enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cloudbees-ci-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cloudbees-ci-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudBeesCI
            name: cloudbeescis.charts.cloudbees.com
            version: v1alpha1
        description: CloudBees CI
        displayName: CloudBees CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudbees
        - ci
        - enterprise
        - continuous
        - integration
        - jenkins
        - automation
        - pipeline
        - devops
        links:
        - name: CloudBees CI
          url: https://www.cloudbees.com/products/continuous-integration
        maintainers:
        - email: support@cloudbees.com
          name: Support
        - email: sales@cloudbees.com
          name: Sales
        maturity: alpha
        provider:
          name: CloudBees
          url: https://www.cloudbees.com/products/continuous-integration
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.marketplace.redhat.com/rhm/cloudbees/cloudbees-ci-operator@sha256:bd0bd72e9e7807b0fa307de8a93a56967221fe2dd181bac02383f0f9be9120e4
        version: 3.34.1
      entries:
      - name: cloudbees-ci.v3.34.1
        version: 3.34.1
      name: alpha
    defaultChannel: alpha
    packageName: cloudbees-ci-rhmp
    provider:
      name: CloudBees
      url: https://www.cloudbees.com/products/continuous-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: bamoe-kogito-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bamoe-kogito-operator.8.0.5-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "ruleunit-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogitoinfra-sample"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "kafka.strimzi.io/v1beta2",
                    "kind": "Kafka",
                    "name": "kogito-kafka"
                  }
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator@sha256:8a99fcd217502b7e850af1e1097de58b6cb19d1e786394811a8e075c4d871279
          description: IBM BAMOE Kogito Operator for deployment and management of
            Kogito services.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["IBM Business Automation Manager
            Open Edition"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: https://www.ibm.com/mysupport
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.rhpam.kiegroup.org
            version: v1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.rhpam.kiegroup.org
            version: v1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.rhpam.kiegroup.org
            version: v1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.rhpam.kiegroup.org
            version: v1
        description: |-
          IBM BAMOE Kogito Operator is designed for deployment and management of Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Builds a IBM BAMOE Kogito service from a remote git repository and creates a custom image for the given runtime: Red Hat build of Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Deploy IBM BAMOE Kogito service from user provided image.
          * Setup Grafana Dashboard when Grafana Operator is available in cluster.
          * Create Prometheus monitor object when Prometheus operator is available in cluster.

          **IMPORTANT!** IBM BAMOE Kogito Operator may need Grafana and Prometheus Operator to configure monitoring in cases if you do not provide these services beforehand. If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the IBM BAMOE Kogito Operator.
        displayName: IBM BAMOE Kogito Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://ibm.com
        - name: Documentation
          url: https://www.ibm.com/docs/en/ibamoe
        - name: Source
          url: https://github.com/kiegroup/kogito-operator
        maintainers:
        - email: mysphelp@us.ibm.com
          name: IBM
        maturity: beta
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9492c67eaa7ab37df6f5fa623c4aac955b7d7ffaab9ae2e57dea745e82907e40
        - registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator@sha256:8a99fcd217502b7e850af1e1097de58b6cb19d1e786394811a8e075c4d871279
        version: 8.0.5-3
      entries:
      - name: bamoe-kogito-operator.8.0.5-3
        version: 8.0.5-3
      - name: bamoe-kogito-operator.8.0.5-2
        version: 8.0.5-2
      - name: bamoe-kogito-operator.8.0.5-1
        version: 8.0.5-1
      - name: bamoe-kogito-operator.8.0.4-2
        version: 8.0.4-2
      - name: bamoe-kogito-operator.8.0.4-1
        version: 8.0.4-1
      - name: bamoe-kogito-operator.8.0.3-1
        version: 8.0.3-1
      - name: bamoe-kogito-operator.8.0.2-3
        version: 8.0.2-3
      - name: bamoe-kogito-operator.8.0.2-2
        version: 8.0.2-2
      - name: bamoe-kogito-operator.8.0.2-1
        version: 8.0.2-1
      - name: bamoe-kogito-operator.8.0.1-2
        version: 8.0.1-2
      - name: bamoe-kogito-operator.8.0.1-1
        version: 8.0.1-1
      - name: bamoe-kogito-operator.8.0.0-1
        version: 8.0.0-1
      name: 8.x
    defaultChannel: 8.x
    packageName: bamoe-kogito-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc
      provider-url: ""
    name: openshift-gitops-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-gitops-operator.v1.10.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:b0dc693854319ada341e2e3872b34d7db9823b470ed369edfde88d3ae73a40e1
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.8.19\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:a74e0a1fe3589ebe9f226a2bad20e9a5d5601a0e62c5f3290902f466f55b7160
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:2bec31b0c5407ffcecfd8b860d8afb1629eaab31a250f60191f94a824e0a52f4
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:c779ad87a7f71d851c2af818abf16e5151c43d3fd54e64eac2fc7117c00753f5
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:b0dc693854319ada341e2e3872b34d7db9823b470ed369edfde88d3ae73a40e1
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:b5c8a6c7438a45125fc7fd970946f715ab7193b7446d625fae80dde9d1f3f914
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:4c842f0ea8f0e9b53eb16d94488fc2f4298bd8e4be2f48713be9157da24549e6
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:ece2fee4227ad1becadca47028b78ceaf8717a98ec84e3f1b5188a135c4b093c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d35dff5d59e748ef0404e7e76b0237ea96b2bb19b86fa04a3a34df7a379277b2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5d5061dbc2ec5082cf14b6c600fb5400b83cf91d7ccebfa80680a238d275db
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:0905ba3f5cf87780a1e95b6a00ea7cd340d4ed1b5054e4b2c82406270baeaefa
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ec9f60018694dcc5d431ba47d5536b761b71cb3f66684978fe6bb74c157679ac
        version: 1.10.6
      entries:
      - name: openshift-gitops-operator.v1.10.6
        version: 1.10.6
      - name: openshift-gitops-operator.v1.10.5
        version: 1.10.5
      - name: openshift-gitops-operator.v1.10.4
        version: 1.10.4
      - name: openshift-gitops-operator.v1.10.3
        version: 1.10.3
      - name: openshift-gitops-operator.v1.10.2
        version: 1.10.2
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.10
    - currentCSV: openshift-gitops-operator.v1.11.7-0.1724840231.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d6ad2ee900bf99a3672049cc8ce9c94870d41176501ded1abbc1b229b9553401
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          olm.substitutesFor: openshift-gitops-operator.v1.11.7
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.9.21\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:4d22a7a6cc2f0442889512267ae3d9356f05ff0dfdbbe137c3d8ebebc24818c8
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:42f809f32f1d43186b849e3cc78063d3ebb2b54c55b2a1b77908af09a5ce044b
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:0b926fc9eae775425faa1ab9709a36fdac15afe4bfe523720e17ecf5df45bb28
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:0df37cf0a0e48efae02a0711c6da61c6165e7856db97097e98d56c8b6a137bb9
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:eade50b8321c1b2e0a8ac4dcb34596d43dea0a137b8b97b6eb68c6d469c8b200
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:20566173a8e907980e603d82fae4d96fcdfd5c43d9004c2e506cf0e09a47b91e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/rhel8/redis-6@sha256:c4a98010f7dd5faec2047bb9803061704233a8efc930be2a02386ac16d062d86
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c94c2167dc957856682a104b9a6187d3d321dec71ae794acc0dc37794d2cc373
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:7ea96000768b3cf7711150531c4c1c7edff2c45a2f29fe10bd9811ab6b281909
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d6ad2ee900bf99a3672049cc8ce9c94870d41176501ded1abbc1b229b9553401
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:959de6a95ef8978f66fdbe1033bcaa0ddeff9a376023e5d1d2698c5acb417305
        version: 1.11.7+0.1724840231.p
      entries:
      - name: openshift-gitops-operator.v1.11.7-0.1724840231.p
        version: 1.11.7+0.1724840231.p
      - name: openshift-gitops-operator.v1.11.7
        version: 1.11.7
      - name: openshift-gitops-operator.v1.11.6
        version: 1.11.6
      - name: openshift-gitops-operator.v1.11.5
        version: 1.11.5
      - name: openshift-gitops-operator.v1.11.4
        version: 1.11.4
      - name: openshift-gitops-operator.v1.11.3
        version: 1.11.3
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.11
    - currentCSV: openshift-gitops-operator.v1.12.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:48dd1254fd3a887aef2eae9239f68637eae18514a50f72ea504414d83dc727b8
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.10.18\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:dbf3038daf7c797123c511f9a609c3485cb1a5bcae53c614fad17e0a65fa7ee3
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a1d4e26ea755181818401697f18260b9e876aa441be14830f9142081ee22a24d
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:73918f3b3c507fc64006b7a500ecfa9d58151d694852f696c41a4412edffc78e
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:33bbb88343285a0510670384b87d0226dd6de8aa12c78b68921362c4d6f5a6aa
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:48dd1254fd3a887aef2eae9239f68637eae18514a50f72ea504414d83dc727b8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9e0fee30842a16d7b8a39d336123c80c5271cb2ceb50759fc7ee51a4ed09ceb3
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:06a6637c4776f4285bcb295d327ee60b1aed61b7adc5f7ab52898d79281ced70
        - registry.redhat.io/rhel8/redis-6@sha256:a848cc4edb4405846c673187dd68ce66c328dd1ad42f193f5737288c7971399d
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:b5736d1106db29fcdf42328ca20b96382d4339f90b97a07120ea73ee174ea1ad
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:7d82c719b1cef64607d30e8aaa5ab293955771c5b1dbd430e0befc988a43356d
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:1cf485a889c7586cb4944f000b53c0e97d08f97ba14e580bec57dad7cb13712e
        version: 1.12.6
      entries:
      - name: openshift-gitops-operator.v1.12.6
        version: 1.12.6
      - name: openshift-gitops-operator.v1.12.5-0.1727371523.p
        version: 1.12.5+0.1727371523.p
      - name: openshift-gitops-operator.v1.12.5
        version: 1.12.5
      - name: openshift-gitops-operator.v1.12.4
        version: 1.12.4
      - name: openshift-gitops-operator.v1.12.3
        version: 1.12.3
      - name: openshift-gitops-operator.v1.12.2
        version: 1.12.2
      - name: openshift-gitops-operator.v1.12.1
        version: 1.12.1
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.12
    - currentCSV: openshift-gitops-operator.v1.13.3-0.1741683398.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d41034bcf5c28bcb8afef08678793d0cc27a4032e2811a0c496c3bd0f10ecc17
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          olm.substitutesFor: openshift-gitops-operator.v1.13.3-0.1738142397.p
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.11.12\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_gitops/
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d41034bcf5c28bcb8afef08678793d0cc27a4032e2811a0c496c3bd0f10ecc17
        - registry.redhat.io/rhel8/redis-6@sha256:a848cc4edb4405846c673187dd68ce66c328dd1ad42f193f5737288c7971399d
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:b982df02ea7c45a7603d83fe82cac12e1faac303a1b72af6a727512bb97441df
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3f2804bc2af437ff1f06c214b48a1a780dfcfab995650d901c03270057eb9f3c
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:06a6637c4776f4285bcb295d327ee60b1aed61b7adc5f7ab52898d79281ced70
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a7c3e31ce414e6c6fecf1652343b22c52502902676bf4d7a75bf42bdabfca06
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:a366229e6ee26ebe2d7f765b2a2da6057a29e783835655b5d97545ac563b541a
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:d4d5724f67859d24c32f0f5735f5af3e325f77db97f58bc7f17ed4b831b7e534
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c02dc2c79ee24ce5450ea4e19493ac2a1997eeb5bfd359291642c529826d0cdd
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:e87515d016565bb69cb2c2ac40432e0127dd2cb20a490ac82019836032ba15e5
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:bde11fa83b7e708677b08ea995041f480b88edfb93e4a62c10380d101c0a349c
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        version: 1.13.3+0.1741683398.p
      entries:
      - name: openshift-gitops-operator.v1.13.3-0.1741683398.p
        version: 1.13.3+0.1741683398.p
      - name: openshift-gitops-operator.v1.13.3-0.1738142397.p
        version: 1.13.3+0.1738142397.p
      - name: openshift-gitops-operator.v1.13.3
        version: 1.13.3
      - name: openshift-gitops-operator.v1.13.2
        version: 1.13.2
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.13
    - currentCSV: openshift-gitops-operator.v1.14.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:7cf924af77d9c56e70710b99e080ed8aee162f05ec472f32085a76ef85cf9879
          createdAt: "2025-05-22 12:40:45"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.12.13-1\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_gitops/
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:2c45cfc01f9ec7577afc09cadc4c1a6f470836f6a0a7c24d66b791913f89bb05
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:42a79aad641e8607fc339e4ce7f9cf1d258be01f863dbc31a94d2c27edb645bd
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:97e32825f59c33cca7d8dba3580095205c20435ccd15b91cf0aff37db9841e0e
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:7cf924af77d9c56e70710b99e080ed8aee162f05ec472f32085a76ef85cf9879
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:b7d3c882fc25fbbea28951afb197d4d47ace253864889339396e20f0c6897a7f
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:bcf6b3136a5e71803cb518c8763c165bb46ee37ef7bf55503911d7fcb36b3a8f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c369df980874aae270ca7bc6455f1ca45467d75d25fe80d7bcb06a7826bdeaa3
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:cd4965d91c163160bcf43a02c54ca60b8704281c95de904f904df942bfd83bcd
        - registry.redhat.io/rhel8/redis-6@sha256:a9f30ed08abcab0122556a5485fbcf6e0fc03860f19954dcebdbf5e754876a45
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:7358787959af7e7e589edfc043a97d80ebab87536bde365d33be7a600071bf0e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1ed3b514bf9af6ac0d386e3739874d1d964e667d54e041c2d67edd769b0009fb
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:8096f824bd96e3148cf4b6936766c888d1fff0af9e9d0be440ae4feb6adc03f2
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:1112160243f227a5d393eeee165bd4fed0d3f42646a9280cbe9af60ef728aeb2
        version: 1.14.4
      entries:
      - name: openshift-gitops-operator.v1.14.4
        version: 1.14.4
      - name: openshift-gitops-operator.v1.14.3-0.1746016855.p
        version: 1.14.3+0.1746016855.p
      - name: openshift-gitops-operator.v1.14.3-0.1742836272.p
        version: 1.14.3+0.1742836272.p
      - name: openshift-gitops-operator.v1.14.3
        version: 1.14.3
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.14
    - currentCSV: openshift-gitops-operator.v1.15.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ad49a3d14f843bde0d2365093d00ffa6d3b9d59a75c81e0c56710ad37b53bd92
          createdAt: "2024-11-25T03:57:37Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v2.13.9\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:d41454824632475f0e88aa6e0031ec643bd37e1d575df36cd3ee63daf1403281
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:6cec76d7ea70d77923cf4285177e375d34e96e3aecbe77b68cd5ff769e6e540b
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:911733d78ae6f36f642b4d63f044cd102b7eee170508eaf518ac7acf0f339698
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ad49a3d14f843bde0d2365093d00ffa6d3b9d59a75c81e0c56710ad37b53bd92
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:41717c27915251bda8afeef651b609b8a9ed410e0fa09c4cf8fa7e91d6a94292
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:0c7f72546d0d918bfd21aff7157b983ff7c832b269f5efb84d2ed723a09bf612
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:5bda26e2c16beebf66dd8f3c3c55d48e2f00db870f2d9d4d44570d9d3b8bddb0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:a9e2360f00c5c9840a100a6748bd905b8f34169140a33fb0ca4af9cf6db71642
        version: 1.15.4
      entries:
      - name: openshift-gitops-operator.v1.15.4
        version: 1.15.4
      - name: openshift-gitops-operator.v1.15.3
        version: 1.15.3
      - name: openshift-gitops-operator.v1.15.2
        version: 1.15.2
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.15
    - currentCSV: openshift-gitops-operator.v1.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ead539d7cf54992cf1bfeae6a73a453059a7dffef44505991ff33319c8d40c50
          createdAt: "2025-03-07T13:13:07Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v2.14.16\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:58284a12174d19c2d04723fa1f551d96b490872e0fa2ad94af1b3e52dc6ad955
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a5e2a12ec5dea607a6651462d8e7920cc73f3272dba4041c397a14f2c73d7532
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:03d0d9093422396330aa8611adf77abdc9be759f3d2e23850e4abe778bd6002b
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:cab809fd1288d31c93f4f5c4facb7efaff9f844e21a859d1e4557831575a13c8
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c37aa141387c847eac825363dddda3eb59203d3694120dba439b00cf0c66f320
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:2598a3339bc920b91018d3f7acbe78ccbee684b5182c874e48296744d2373f9d
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ead539d7cf54992cf1bfeae6a73a453059a7dffef44505991ff33319c8d40c50
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:ceff1cde6d043daae599b578eec9759c6d49cd5cf770460b9d075fa879a16e81
        version: 1.16.3
      entries:
      - name: openshift-gitops-operator.v1.16.3
        version: 1.16.3
      - name: openshift-gitops-operator.v1.16.2
        version: 1.16.2
      - name: openshift-gitops-operator.v1.16.1
        version: 1.16.1
      - name: openshift-gitops-operator.v1.16.0-0.1746014725.p
        version: 1.16.0+0.1746014725.p
      - name: openshift-gitops-operator.v1.16.0
        version: 1.16.0
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.16
    - currentCSV: openshift-gitops-operator.v1.17.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:1b86eb4e071b1c526931b6d18c3d8f51522293699059c514488602c320271049
          createdAt: "2025-07-30T13:03:16Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.0.14\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:3fc4941fa4a6f2dbe9dd82622a0d3fb794ecd5a9dee286a38b60a7f6a947d685
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:cc45e212e7eb4e7e8daac6ea74071d5d98260de0b1f6d1cea0e7876111ccea8c
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:2f5f997bce924445de735ae0508dca1a7bba561bc4acdacf659928488233cb8a
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:5c9ea426cd60e7b8d1d8e4fe763909200612434c65596855334054e26cbfe3d0
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:70e5e3900d0c89999c69be1d58adf84c1b3664aacbffa9c5e0efc8ae1b6ef9c1
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:1b86eb4e071b1c526931b6d18c3d8f51522293699059c514488602c320271049
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:d73b4f5b355f556208751281923bb8b43edffc061a673c04d53b88f0e50ecb26
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:a2ae66772d90e734bb3385a59852e6016325a21b916a2580a33cf0da7cb0b527
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:c0a513ed54b631629efafe6e94b775d5bd0a943c44fbe96f02a0d5bb1f989099
        version: 1.17.1
      entries:
      - name: openshift-gitops-operator.v1.17.1
        version: 1.17.1
      - name: openshift-gitops-operator.v1.17.0
        version: 1.17.0
      name: gitops-1.17
    - currentCSV: openshift-gitops-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
          createdAt: "2025-08-21T01:20:45Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.18.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: NamespaceManagement
            name: namespacemanagements.argoproj.io
            version: v1beta1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.1.6\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/rhel9/redis-7@sha256:b4f58ad83ab8719a2e102f38ee526546cd1d56428901afc3add45c9a0e2a2f12
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:1eef9ecd58cf8391f29360ef74dd8aa5dcb48e1582d1500e445706d8302c951e
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a0544c14823492165550d83a6d8ba79dd632b46144d3fdcb543793726111d76
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:6284873bcf1593c210003e0292ac5583f03f03e80d937098d808d48fa0d01cf6
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:5a09fcba3d9e6fc462719e81a46af6022cf7f70a05bcb387ed959a49f6f37dbc
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:adca9c55b80d899d500bf0a7234cf61ae30ffed1b2c927ddea8d9fd07c9acb6a
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7f6ceae56c60fd432425c36f93a9e9bbbb96ef9422333ad6ea41ff3a2eaf8374
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:640e3298c7cd2faa84971b0d861446c4892ab8602108af05b8b5b049a4de624f
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c2283997d47298d49378e8200b149ea613c7735317ef12e74b7915a590f41ce2
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:08efd3e7a79c2945693c99ed31ec912223becff7238b1e28d1b1da5d63f93019
        version: 1.18.0
      entries:
      - name: openshift-gitops-operator.v1.18.0
        version: 1.18.0
      name: gitops-1.18
    - currentCSV: openshift-gitops-operator.v1.6.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:104cc7564fcc3261faa3fbf9c95720b4434e15103b876c080ed012b244ef1dae
          createdAt: "2021-09-15T22:31:21Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD v2.4.28\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces. \n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc.
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:104cc7564fcc3261faa3fbf9c95720b4434e15103b876c080ed012b244ef1dae
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:0cc4fa704e581cd5581c816668183f602fae6ef9ab3c04ce0d99d561287ef140
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:06f1150114d4c0bd236d2a5d4716c9c3e95a9733058297790389d6ba2a1bf264
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:e9e2f950cda8153cff112e43872b98e4977d3cbdf59713b8de92defef6d34b6d
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:dbcad68b70ab7918e944099aef4177c9bc3378e9ca49499dbb94a175927d3209
        - registry.redhat.io/rhel8/redis-6@sha256:53598a6effeb90e4f1b005b2521beffd2fa2b0c52d0e7f2347ee2abd2577cab3
        version: 1.6.6
      entries:
      - name: openshift-gitops-operator.v1.6.6
        version: 1.6.6
      - name: openshift-gitops-operator.v1.6.5
        version: 1.6.5
      - name: openshift-gitops-operator.v1.6.4
        version: 1.6.4
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.6
    - currentCSV: openshift-gitops-operator.v1.7.4-0.1690486082.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:e0967a4058fe5a9bfe7e998a3347529c99536cf3a44de9e40f9632035ff5a2fe
          createdAt: "2021-09-15T22:31:21Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.substitutesFor: openshift-gitops-operator.v1.7.4
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD v2.5.16\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficient privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces. \n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/rhel8/redis-6@sha256:53598a6effeb90e4f1b005b2521beffd2fa2b0c52d0e7f2347ee2abd2577cab3
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:e0967a4058fe5a9bfe7e998a3347529c99536cf3a44de9e40f9632035ff5a2fe
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c55aadf45397083bec4e42d6c0514ed79169a07fd3d9d75671904df2c0520e52
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:45789c49d9334e0ed062641313714c101bd0117dc5633316b1ec63e7a6c76c15
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:26050fb95978e90e8d00e895032045e77cb236008a0e1417459e1aeca8a48c69
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:5c11be8ec9db54331893f2f298aa27006e77b0eaded2d88d0531c77da91a421f
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:f1371d2012796e9aaca105112190ff89ec3d0be95968e05b19db101480f67975
        version: 1.7.4+0.1690486082.p
      entries:
      - name: openshift-gitops-operator.v1.7.4-0.1690486082.p
        version: 1.7.4+0.1690486082.p
      - name: openshift-gitops-operator.v1.7.4
        version: 1.7.4
      - name: openshift-gitops-operator.v1.7.3
        version: 1.7.3
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.7
    - currentCSV: openshift-gitops-operator.v1.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f32df77b3c23d6a833ff887386285ca7a5bd436641f0ec1ffc934c8247fa315f
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.skipRange: ""
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.6.153\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:8cc59901689858e06f5c2942f8c975449c17454fa8364da6153d9b5a3538a985
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:f77594bc053be144b33ff9603ee9675c7e82946ec0dbfb04d8f942c8d73155da
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f32df77b3c23d6a833ff887386285ca7a5bd436641f0ec1ffc934c8247fa315f
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:8aa68fddef88921aeeede24a849d3dabbceff01dc03185342c21e6ee58033eaf
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:16d9184345e6d5a19f326a7b52f76d8a59b96763f233c50546ff017a625e5ff5
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:0d13c03048b4435d9f5045a6fb7bd4b8808461b26aeca60c0047fa5937e1d45d
        version: 1.8.6
      entries:
      - name: openshift-gitops-operator.v1.8.6
        version: 1.8.6
      - name: openshift-gitops-operator.v1.8.5
        version: 1.8.5
      - name: openshift-gitops-operator.v1.8.4-0.1693214052.p
        version: 1.8.4+0.1693214052.p
      - name: openshift-gitops-operator.v1.8.4
        version: 1.8.4
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.8
    - currentCSV: openshift-gitops-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "argo-rollout"
                },
                "spec": null
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:11a6084e11aa32511a34bd6cfb4f4995aec690c1f5ed61861e99f6e42d62afb4
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.skipRange: ""
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.7.16\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:80f92842fe4cba008a873ad2ab6059a3a24e76e16671d8ece034dbbe666996bc
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:3d17786e2f4063154b9e3a40b96bf8ec0a14b86395cea5fe7ca18c5b43764478
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a568c8c647e2e2d6abdf12138eeef8f8bdde1fdcbf786e9ca8c8f93bc79acaf9
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:43bb961d847477902e66a92edd1539322cac8d5f74706b048a458ad037a0c7a3
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:11a6084e11aa32511a34bd6cfb4f4995aec690c1f5ed61861e99f6e42d62afb4
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ec9f60018694dcc5d431ba47d5536b761b71cb3f66684978fe6bb74c157679ac
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:3873d1f18b5643e1de2601419410270342650a21428bb26a4cbee20c0c63f98b
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:86012145d4b8798dc12f6f4f07a9316a193188aaac865e324b63f4e60e8ec584
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:136d3a936759e112d3ea2e78f7c02bd882b0fd3011f4c452f45c32c43822772d
        version: 1.9.4
      entries:
      - name: openshift-gitops-operator.v1.9.4
        version: 1.9.4
      - name: openshift-gitops-operator.v1.9.3
        version: 1.9.3
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.9
    - currentCSV: openshift-gitops-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
          createdAt: "2025-08-21T01:20:45Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.18.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: NamespaceManagement
            name: namespacemanagements.argoproj.io
            version: v1beta1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.1.6\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:adca9c55b80d899d500bf0a7234cf61ae30ffed1b2c927ddea8d9fd07c9acb6a
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7f6ceae56c60fd432425c36f93a9e9bbbb96ef9422333ad6ea41ff3a2eaf8374
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:6284873bcf1593c210003e0292ac5583f03f03e80d937098d808d48fa0d01cf6
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:08efd3e7a79c2945693c99ed31ec912223becff7238b1e28d1b1da5d63f93019
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:640e3298c7cd2faa84971b0d861446c4892ab8602108af05b8b5b049a4de624f
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c2283997d47298d49378e8200b149ea613c7735317ef12e74b7915a590f41ce2
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:5a09fcba3d9e6fc462719e81a46af6022cf7f70a05bcb387ed959a49f6f37dbc
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:1eef9ecd58cf8391f29360ef74dd8aa5dcb48e1582d1500e445706d8302c951e
        - registry.redhat.io/rhel9/redis-7@sha256:b4f58ad83ab8719a2e102f38ee526546cd1d56428901afc3add45c9a0e2a2f12
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a0544c14823492165550d83a6d8ba79dd632b46144d3fdcb543793726111d76
        version: 1.18.0
      entries:
      - name: openshift-gitops-operator.v1.18.0
        version: 1.18.0
      - name: openshift-gitops-operator.v1.17.1
        version: 1.17.1
      - name: openshift-gitops-operator.v1.17.0
        version: 1.17.0
      - name: openshift-gitops-operator.v1.16.2
        version: 1.16.2
      - name: openshift-gitops-operator.v1.16.1
        version: 1.16.1
      - name: openshift-gitops-operator.v1.16.0-0.1746014725.p
        version: 1.16.0+0.1746014725.p
      - name: openshift-gitops-operator.v1.16.0
        version: 1.16.0
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: latest
    defaultChannel: latest
    packageName: openshift-gitops-operator
    provider:
      name: Red Hat Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: node-maintenance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-maintenance-operator.v0.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-maintenance-operator:v0.17.0
          createdAt: "2024-03-13 18:56:47"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=0.12.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/medik8s/node-maintenance-operator:v0.17.0
        version: 0.17.0
      entries:
      - name: node-maintenance-operator.v0.17.0
        version: 0.17.0
      - name: node-maintenance-operator.v0.16.1
        version: 0.16.1
      - name: node-maintenance-operator.v0.16.0
        version: 0.16.0
      - name: node-maintenance-operator.v0.15.0
        version: 0.15.0
      - name: node-maintenance-operator.v0.14.0
        version: 0.14.0
      - name: node-maintenance-operator.v0.13.0
        version: 0.13.0
      - name: node-maintenance-operator.v0.12.1
        version: 0.12.1
      name: stable
    defaultChannel: stable
    packageName: node-maintenance-operator
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://www.redhat.com
    name: sap-commerce-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sap-commerce-operator.v0.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hybris.hybris.org/v1alpha1",
                "kind": "HybrisApp",
                "metadata": {
                  "name": "hybrisapp-sample"
                },
                "spec": {
                  "aJPServicePort": 30000,
                  "apachejvmRouteName": "",
                  "baseImageName": "hybris-base",
                  "baseImageTag": "v0.2",
                  "hybrisANTTaskNames": "",
                  "podHealthyProbePath": "/platform/init",
                  "sourceRepoLocalPropertiesOverride": "",
                  "sourceRepoRef": "main",
                  "sourceRepoURL": "https://github.com/redhat-sap/sap-commerce-app-example.git",
                  "startupProbeFailureThreshold": 100,
                  "startupProbePeriodSecond": 60
                }
              },
              {
                "apiVersion": "hybris.hybris.org/v1alpha1",
                "kind": "HybrisBase",
                "metadata": {
                  "name": "hybrisbase-sample"
                },
                "spec": {
                  "URL": "http://nexus.redhat.com/nexus/service/local/repositories/releases/content/org/hybris/hybris/1.0/hybris-1.0.tar.gz",
                  "buildSourceRepo": "https://github.com/redhat-sap/sap-commerce-operator",
                  "buildSourceRepoBranch": "main",
                  "imageName": "hybris-base",
                  "imageTag": "v0.2",
                  "jdkURL": "https://github.com/SAP/SapMachine/releases/download/sapmachine-11.0.5/sapmachine-jdk-11.0.5-1.x86_64.rpm",
                  "password": "admin123",
                  "username": "admin"
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          containerImage: quay.io/redhat-sap-cop/hybris-operator:0.0.4
          description: Builds and runs SAP Commerce in OpenShift
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-sap/sap-commerce-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HybrisApp is the Schema for the hybrisapps API
            displayName: Hybris Application
            kind: HybrisApp
            name: hybrisapps.hybris.hybris.org
            version: v1alpha1
          - description: HybrisBase is the Schema for the Hybris Base API
            displayName: Hybris Base
            kind: HybrisBase
            name: hybrisbases.hybris.hybris.org
            version: v1alpha1
        description: |-
          Builds and runs SAP Commerce in OpenShift

          * The operator provides APIs to initialize SAP Commerce with the provided SAP Commerce Server. Though there are no restrictions from the operator on the base package to be used from SAP, the default SAP base image provides a server preconfigured with an application server based on Apache Tomcat. Both HSQLDB and Tomcat are included in the SAP Commerce package, and are designed to be used for testing, development, and demonstration purposes. The custom classes created for various business needs along with an external RDBMS configuration can be used in a HybrisApp creation to execute SAP Commerce in a production setup.

          * The operator also enables the use of SAP Commerce in a clustered model. A SAP Commerce Cluster is a number of individual SAP Commerce installations using a common set of data on one database. A cluster configuration lets you balance the load between the nodes to scale SAP Commerce in a way that is transparent to the user.

          * The clustering offers few communication protocols as options (UDP: both multicast or unicast, or JGroups). In order to support multicast for SAP Commerce clustering the corresponding OpenShift namespace has to be enabled for multicast usage.

          For more details on SAP Commerce consult the product documentation: [https://help.sap.com/viewer/product/SAP_COMMERCE](https://help.sap.com/viewer/product/SAP_COMMERCE)
        displayName: SAP Commerce Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sap-commerce-operator
        links:
        - name: SAP Commerce Operator
          url: https://github.com/redhat-sap/sap-commerce-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
          url: https://www.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/redhat-sap-cop/hybris-operator:0.0.4
        version: 0.0.4
      entries:
      - name: sap-commerce-operator.v0.0.4
        version: 0.0.4
      - name: sap-commerce-operator.v0.0.3
        version: 0.0.3
      - name: sap-commerce-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: sap-commerce-operator
    provider:
      name: Red Hat, Inc.
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cognitoidentityprovider-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cognitoidentityprovider-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cognitoidentityprovider-controller:1.1.1
          createdAt: "2025-09-19T20:58:49Z"
          description: AWS Cognito Identity Provider controller is a service controller
            for managing Cognito Identity Provider resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: UserPool represents the state of an AWS cognitoidentityprovider
              UserPool resource.
            displayName: UserPool
            kind: UserPool
            name: userpools.cognitoidentityprovider.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Cognito Identity Provider resources in AWS from within your Kubernetes cluster.

          **About Amazon Cognito Identity Provider**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Cognito Identity Provider
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cognitoidentityprovider
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Cognito Identity Provider Developer Resources
          url: https://aws.amazon.com/Cognito Identity Provider/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cognitoidentityprovider maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cognitoidentityprovider-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-cognitoidentityprovider-controller.v1.1.1
        version: 1.1.1
      - name: ack-cognitoidentityprovider-controller.v1.1.0
        version: 1.1.0
      - name: ack-cognitoidentityprovider-controller.v1.0.12
        version: 1.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-cognitoidentityprovider-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sumo Logic
      provider-url: https://www.sumologic.com/
    name: sumologic-kubernetes-collection-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "helm-operator.sumologic.com/v1alpha1",
                    "kind": "SumologicCollection",
                    "metadata": {
                        "name": "collection"
                    },
                    "spec": {
                        "sumologic": {
                            "accessId": "",
                            "accessKey": "",
                            "clusterName": "kubernetes",
                            "scc": {
                                "create": true
                            }
                        },
                        "kube-prometheus-stack": {
                            "prometheus-node-exporter": {
                                "service": {
                                    "port": 9200,
                                    "targetPort": 9200
                                }
                            }
                        },
                        "otellogs": {
                            "daemonset": {
                                "containers": {
                                    "otelcol": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                },
                                "initContainers": {
                                    "changeowner": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                }
                            }
                        }
                    }
                }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing,Monitoring
          containerImage: registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
          createdAt: "2024-07-19T10:48:00Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator.git
          support: Sumo Logic
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configuration for Sumo Logic Kubernetes Collection
            displayName: SumologicCollection
            kind: SumologicCollection
            name: sumologiccollections.helm-operator.sumologic.com
            version: v1alpha1
        description: Helm Operator for the Sumo Logic Kubernetes Collection Helm Chart
          v4.13.0-2
        displayName: Sumo Logic Kubernetes Collection Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - collection
        - logging
        - metrics
        - monitoring
        links:
        - name: Sumo Logic Kubernetes Collection Helm Operator
          url: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator
        maintainers:
        - email: collection@sumologic.com
          name: Sumo Logic
        maturity: alpha
        provider:
          name: Sumo Logic
          url: https://www.sumologic.com/
        relatedImages:
        - registry.connect.redhat.com/sumologic/autoinstrumentation-java@sha256:242958b7d8905f4ac02f6ad38897ba58b9badf2a6213a389e67dcc71fee3630c
        - registry.connect.redhat.com/sumologic/thanos@sha256:16faf7191668e86d8b576bba055dd0cbdeb009635976b9e07bd00f926032f11c
        - registry.connect.redhat.com/sumologic/tailing-sidecar@sha256:738d653a78dc415243ab3627db9135768175fb7f51d0803566f5d9e824c8744b
        - registry.connect.redhat.com/sumologic/opentelemetry-operator@sha256:5587fe5b7ebfa57ad320ad2521103f908317388d1741c40e5f7be5ed9e5f9ad7
        - registry.connect.redhat.com/sumologic/kube-state-metrics@sha256:3551ededa327a35db798ec8286cd843580e3ae395d2231e041e8c4b7621127ee
        - registry.connect.redhat.com/sumologic/tailing-sidecar-operator@sha256:c16b8b18e9b263fc3eb24059bba512bcfc3201b32105af6e5518392826ddd66c
        - registry.connect.redhat.com/sumologic/telegraf-operator@sha256:ff958222be01c0c3966474463e43bbd5bcf348790dd34bb761718cb1cea2d793
        - registry.connect.redhat.com/sumologic/kubernetes-tools-kubectl@sha256:d3155b864cc74f327adedf7d013a8f0fcb11c2e73a5655b7cafa848824e80172
        - registry.connect.redhat.com/sumologic/kubernetes-setup@sha256:f07eaf68fa11ffc708e4485a83bd3763fc5712f7d73d29edfe5b9c9e08f05df6
        - registry.connect.redhat.com/sumologic/sumologic-otel-collector@sha256:4554d43dba204e3ec8e91b36442a438f10234fa0cb4db7fb7880b257a6485112
        - registry.connect.redhat.com/sumologic/autoinstrumentation-nodejs@sha256:c206eef4fdecaa58b4f01e5e7f0dbd74e9606eab2b4debf47461d36191e248d1
        - registry.connect.redhat.com/sumologic/autoinstrumentation-dotnet@sha256:f4f385c4489c74b5c43f2fe6b6620a9eb8d9ecab470567236b3c45f636106640
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
        - registry.connect.redhat.com/sumologic/metrics-server@sha256:65d2a3ddc4b4183f438fa18b655079c93c905cb9e3b8c54977336205dc1fff07
        - registry.connect.redhat.com/sumologic/nginx-unprivileged@sha256:7260555d16c94c4fe1e016626b4cb8a5c0de6b700e2eea28744331fb298f71ae
        - registry.connect.redhat.com/sumologic/node-exporter@sha256:11e9e604432a66d9b388cb1bf70eac077c7f0372d1264342ef927d76b8ef0272
        - registry.connect.redhat.com/sumologic/prometheus-operator@sha256:58a329e6582456b32abebb470086191f8366b927af4703741613993465f116b8
        - registry.connect.redhat.com/sumologic/prometheus@sha256:0ef6eaf570a40499dd5de92a79593e950bc771b014e0ee6ff7419b4b82a36fcc
        - registry.connect.redhat.com/sumologic/autoinstrumentation-python@sha256:f0fa45fa45e64c7df88f21181ae18ee96fe83cd3988258663d0806ea6f39e6bb
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:e1f49fcf5b7e86d9eb49c16dbbf9484cb486f55f999fd51d7c738c3b1f2666cf
        - registry.connect.redhat.com/sumologic/telegraf@sha256:75efd2e509c6017ea0fc5c9103ce068cbc89087b16506a4ef48cb5ff00eab84e
        - registry.connect.redhat.com/sumologic/busybox@sha256:209bc8e0b1d958fc699f3bb75b1248544e372ad35cdc46c991bd1698f641e1f9
        - registry.connect.redhat.com/sumologic/prometheus-config-reloader@sha256:0b605d36782afd8f50fd631d0e9c7b19d39ce962af89b7a737b2f09a8ceff291
        version: 4.13.0-2
      entries:
      - name: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
        version: 4.13.0-2
      name: alpha
    defaultChannel: alpha
    packageName: sumologic-kubernetes-collection-helm-operator
    provider:
      name: Sumo Logic
      url: https://www.sumologic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: rh-service-binding-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-binding-operator.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ServiceBinding expresses intent to bind an operator-backed
              service with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding for Kubernetes. Describes the connection
              between a Provisioned Service and an Application Projection. Implements
              Service Binding Specifcation for Kubernetes (https://github.com/k8s-service-bindings/spec).
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: alpha
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:fe6629401bc8447da95ec55949b5a27e620364bb6e02f86230262699ac8a5d55
        version: 0.9.1
      entries:
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: beta
    - currentCSV: service-binding-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindablekinds-sample"
                }
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:e799bd06043028d214baeeffc683710ad2eed7b83272fc653e3b89f69711319b
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:e799bd06043028d214baeeffc683710ad2eed7b83272fc653e3b89f69711319b
        version: 0.11.0
      entries:
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: candidate
    - currentCSV: service-binding-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:3d7a3648e86848402577dd19c1f8fd130a5556ebdfbc0f663a72f9e141024812
        version: 0.10.0
      entries:
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: preview
    - currentCSV: service-binding-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindable-kinds"
                },
                "spec": {}
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  },
                  "type": "postgres",
                  "workload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: registry.redhat.io/ocp-tools-4/service-binding-rhel8-operator@sha256:9c8fb0c8f9d2f54685561af21e595aa88bf309eaebebdf23fb6997b75d026b18
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'.
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1alpha3
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1beta1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
          - description: Service Binding implementing community specification (https://github.com/k8s-service-bindings/spec)
              describing the connection between a backing service and an application
              workload.
            displayName: Service Binding (Spec API)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1beta1
        description: |
          The Service Binding Operator manages the data plane for applications and backing services. It enables developers to connect their
          applications to backing services (REST API, databases, event buses and others) with a consistent and predictable
          experience. The operator reads data made available by the control plane of backing services and projects the bindings
          data to applications either as accessible files mounted into the application’s container or environment variables.

          In Kubernetes, each service suggests a different way to access their secrets and developers will consume them in their
          application in a custom way. While this provides a good deal of flexibility, it also creates discrepancies and
          complexity dealing with each specific services, which leads large development teams to lose overall velocity.

          Service Binding Operator removes the error-prone manual configuration of binding information, provides service operators
          a low-touch administrative experience to provision and manage access to services and enriches development lifecycle.

          For additional details refer to the project [documentation](https://redhat-developer.github.io/service-binding-operator).

          Get Started with Service Binding Operator
          * Find out about SBO’s [prerequisites](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/installing-service-binding.html).
          * Refer to the [Quick Start Guide](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/quick-start.html) to see SBO in action in a simple scenario.

          Need Help
          * Raise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)

          Licence
          * Service Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE)

          Service Binding Operator is providing and supporting two different APIs.
          * `binding.operators.coreos.com/v1alpha1`: This API is compliant with the Service Binding specification. It is fully supported and considered as the mature API. It’s the API we recommend in production environments.
          * `servicebinding.io/v1beta1`: This API is the one defined in the Service Binding specification. It is available in Service Binding Operator 1.x.

          We’ll consider supporting only the API from the specification, once the Service Binding specification will be officially GA. Of course, that will be a breaking change and handle in a safe way for you.
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Documentation
          url: https://redhat-developer.github.io/service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4/service-binding-rhel8-operator@sha256:9c8fb0c8f9d2f54685561af21e595aa88bf309eaebebdf23fb6997b75d026b18
        version: 1.3.3
      entries:
      - name: service-binding-operator.v1.3.3
        version: 1.3.3
      - name: service-binding-operator.v1.3.1
        version: 1.3.1
      - name: service-binding-operator.v1.3.0
        version: 1.3.0
      - name: service-binding-operator.v1.2.0
        version: 1.2.0
      - name: service-binding-operator.v1.1.1
        version: 1.1.1
      - name: service-binding-operator.v1.1.0
        version: 1.1.0
      - name: service-binding-operator.v1.0.1
        version: 1.0.1
      - name: service-binding-operator.v1.0.0
        version: 1.0.0
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: rh-service-binding-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cluster-logging
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-logging.v5.9.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "name": "remote-elasticsearch",
                      "secret": {
                        "name": "elasticsearch"
                      },
                      "type": "elasticsearch",
                      "url": "remote.example.org:9200"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "audit",
                        "infrastructure"
                      ],
                      "name": "enable-default-log-store",
                      "outputRefs": [
                        "default"
                      ]
                    },
                    {
                      "inputRefs": [
                        "application"
                      ],
                      "name": "forward-to-remote",
                      "outputRefs": [
                        "remote-elasticsearch"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogging",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "collection": {
                    "type": "vector"
                  },
                  "logStore": {
                    "elasticsearch": {
                      "nodeCount": 3,
                      "redundancyPolicy": "SingleRedundancy",
                      "resources": {
                        "requests": {
                          "memory": "2Gi"
                        }
                      },
                      "storage": {
                        "size": "200G"
                      }
                    },
                    "retentionPolicy": {
                      "application": {
                        "maxAge": "7d"
                      }
                    },
                    "type": "elasticsearch"
                  },
                  "managementState": "Managed",
                  "visualization": {
                    "kibana": {
                      "replicas": 1
                    },
                    "type": "kibana"
                  }
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          console.openshift.io/plugins: '["logging-view-plugin"]'
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:54054faf0d2f375dd6d7fc7929a2dbb529508e1b3c004a56fc97c2a682d16be2
          createdAt: "2025-03-14T20:19:21Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing your aggregated logging stack.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.7.0-0 <5.9.15'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Logging (team-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.


              There are built-in input names for common log categories, and you can
              define custom inputs to do additional filtering.


              There is a built-in output name for the default openshift log store, but
              you can define your own outputs with a URL and other connection information
              to forward logs to other stores or processors, inside or outside the cluster.


              For more details see the documentation on the API fields.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.logging.openshift.io
            version: v1
          - description: A Red Hat OpenShift Logging instance. ClusterLogging is the
              Schema for the clusterloggings API
            displayName: Cluster Logging
            kind: ClusterLogging
            name: clusterloggings.logging.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates and manages the aggregated logging stack as a cluster-wide service.

          ##Features
          * **Create/Destroy**: Launch and create an aggregated logging stack to support the entire OCP cluster.
          * **Simplified Configuration**: Configure your aggregated logging cluster's structure like components and end points easily.
          * Configure log forwarding to an in-cluster, Red Hat managed log storage solution based on LokiStack or Elasticsearch

          ## Prerequisites
          ### Red Hat OpenShift Logging Namespace
          Cluster logging and the Red Hat OpenShift Logging Operator is only deployable to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - fluentd
        - logging
        - aggregated
        - efk
        - vector
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Fluentd
          url: https://www.fluentd.org/
        - name: Vector
          url: https://vector.dev/
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:54054faf0d2f375dd6d7fc7929a2dbb529508e1b3c004a56fc97c2a682d16be2
        - registry.redhat.io/openshift-logging/logging-view-plugin-rhel9@sha256:8023b072b45e9a21ddef83a56f4760d476053250d5a920d8eb5f03eecab8b1b3
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:54e3a5efaae5ecc2a08db03a421aa79e323b48c764115ea84d83a37c21d1d997
        - registry.redhat.io/openshift-logging/fluentd-rhel9@sha256:a676c9164ec2523b2002dc5eba234ce774926dbf7fe87de6d72b53f99cbf1db5
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:865572275b79462f14e84c69438a61afe9465cfe45ec164caf30965730b276f7
        version: 5.9.15
      entries:
      - name: cluster-logging.v5.9.15
        version: 5.9.15
      - name: cluster-logging.v5.9.14
        version: 5.9.14
      - name: cluster-logging.v5.9.13
        version: 5.9.13
      - name: cluster-logging.v5.9.12
        version: 5.9.12
      - name: cluster-logging.v5.9.11
        version: 5.9.11
      - name: cluster-logging.v5.9.10
        version: 5.9.10
      - name: cluster-logging.v5.9.9
        version: 5.9.9
      - name: cluster-logging.v5.9.8
        version: 5.9.8
      - name: cluster-logging.v5.9.7
        version: 5.9.7
      - name: cluster-logging.v5.9.6
        version: 5.9.6
      - name: cluster-logging.v5.9.5
        version: 5.9.5
      - name: cluster-logging.v5.9.4
        version: 5.9.4
      - name: cluster-logging.v5.9.3
        version: 5.9.3
      - name: cluster-logging.v5.9.2
        version: 5.9.2
      - name: cluster-logging.v5.9.1
        version: 5.9.1
      - name: cluster-logging.v5.9.0
        version: 5.9.0
      name: stable
    - currentCSV: cluster-logging.v5.8.22
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "name": "remote-elasticsearch",
                      "secret": {
                        "name": "elasticsearch"
                      },
                      "type": "elasticsearch",
                      "url": "remote.example.org:9200"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "audit",
                        "infrastructure"
                      ],
                      "name": "enable-default-log-store",
                      "outputRefs": [
                        "default"
                      ]
                    },
                    {
                      "inputRefs": [
                        "application"
                      ],
                      "name": "forward-to-remote",
                      "outputRefs": [
                        "remote-elasticsearch"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogging",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "collection": {
                    "type": "vector"
                  },
                  "logStore": {
                    "elasticsearch": {
                      "nodeCount": 3,
                      "redundancyPolicy": "SingleRedundancy",
                      "resources": {
                        "requests": {
                          "memory": "2Gi"
                        }
                      },
                      "storage": {
                        "size": "200G"
                      }
                    },
                    "retentionPolicy": {
                      "application": {
                        "maxAge": "7d"
                      }
                    },
                    "type": "elasticsearch"
                  },
                  "managementState": "Managed",
                  "visualization": {
                    "kibana": {
                      "replicas": 1
                    },
                    "type": "kibana"
                  }
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.12-v4.17
          console.openshift.io/plugins: '["logging-view-plugin"]'
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:8e13bd3e1654fe26c1d4d8d4a1bfb12e04cee7899cf16b5fa07b62b8f7494af5
          createdAt: "2024-10-17T17:45:03Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing your aggregated logging stack.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.22'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-5.8
          operators.operatorframework.io.bundle.channels.v1: stable-5.8
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Logging (team-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "ClusterLogForwarder is an API to configure forwarding logs.
              \n You configure forwarding by specifying a list of `pipelines`, which
              forward from a set of named inputs to a set of named outputs. \n There
              are built-in input names for common log categories, and you can define
              custom inputs to do additional filtering. \n There is a built-in output
              name for the default openshift log store, but you can define your own
              outputs with a URL and other connection information to forward logs
              to other stores or processors, inside or outside the cluster. \n For
              more details see the documentation on the API fields."
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.logging.openshift.io
            version: v1
          - description: A Red Hat OpenShift Logging instance. ClusterLogging is the
              Schema for the clusterloggings API
            displayName: Cluster Logging
            kind: ClusterLogging
            name: clusterloggings.logging.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          ### Notice for OCP 4.16 and 4.17 Only

          Logging 5.8 is supported on OCP 4.16 and 4.17 to help customers upgrade and migrate to Logging 6.0. **Logging 5.8 will be End of Life in November 2025.**  At that time, it will no longer be available or supported.

          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates and manages the aggregated logging stack as a cluster-wide service.

          ##Features
          * **Create/Destroy**: Launch and create an aggregated logging stack to support the entire OCP cluster.
          * **Simplified Configuration**: Configure your aggregated logging cluster's structure like components and end points easily.
          * Configure log forwarding to an in-cluster, Red Hat managed log storage solution based on LokiStack or Elasticsearch

          ## Prerequisites
          ### Red Hat OpenShift Logging Namespace
          Cluster logging and the Red Hat OpenShift Logging Operator is only deployable to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - fluentd
        - logging
        - aggregated
        - efk
        - vector
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Fluentd
          url: https://www.fluentd.org/
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/logging-view-plugin-rhel9@sha256:f00d4e7ce6b37f1a053c5dbe554359d844989c2243574000ada0f5bc678ffff8
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:8e13bd3e1654fe26c1d4d8d4a1bfb12e04cee7899cf16b5fa07b62b8f7494af5
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:f26518dd3dbf6b6a9e9e20e40741687996b1160c6ad8ad417989197800c2b3fe
        - registry.redhat.io/openshift-logging/fluentd-rhel9@sha256:9296e50fccae60dea580186bc5ae2a387172dc90ece35663d96c4553a2446a01
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:8f29d6eaeb4a11c9a61e3d763605e3c9c8655df7fbb4e39e5b8311d18449f04a
        version: 5.8.22
      entries:
      - name: cluster-logging.v5.8.22
        version: 5.8.22
      - name: cluster-logging.v5.8.21
        version: 5.8.21
      - name: cluster-logging.v5.8.20
        version: 5.8.20
      - name: cluster-logging.v5.8.19
        version: 5.8.19
      - name: cluster-logging.v5.8.18
        version: 5.8.18
      - name: cluster-logging.v5.8.17
        version: 5.8.17
      - name: cluster-logging.v5.8.16
        version: 5.8.16
      - name: cluster-logging.v5.8.15
        version: 5.8.15
      - name: cluster-logging.v5.8.14
        version: 5.8.14
      - name: cluster-logging.v5.8.13
        version: 5.8.13
      - name: cluster-logging.v5.8.12
        version: 5.8.12
      - name: cluster-logging.v5.8.11
        version: 5.8.11
      - name: cluster-logging.v5.8.10
        version: 5.8.10
      - name: cluster-logging.v5.8.9
        version: 5.8.9
      - name: cluster-logging.v5.8.8
        version: 5.8.8
      - name: cluster-logging.v5.8.7
        version: 5.8.7
      - name: cluster-logging.v5.8.6
        version: 5.8.6
      - name: cluster-logging.v5.8.5
        version: 5.8.5
      - name: cluster-logging.v5.8.4
        version: 5.8.4
      - name: cluster-logging.v5.8.3
        version: 5.8.3
      - name: cluster-logging.v5.8.2
        version: 5.8.2
      - name: cluster-logging.v5.8.1
        version: 5.8.1
      - name: cluster-logging.v5.8.0
        version: 5.8.0
      name: stable-5.8
    - currentCSV: cluster-logging.v5.9.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "name": "remote-elasticsearch",
                      "secret": {
                        "name": "elasticsearch"
                      },
                      "type": "elasticsearch",
                      "url": "remote.example.org:9200"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "audit",
                        "infrastructure"
                      ],
                      "name": "enable-default-log-store",
                      "outputRefs": [
                        "default"
                      ]
                    },
                    {
                      "inputRefs": [
                        "application"
                      ],
                      "name": "forward-to-remote",
                      "outputRefs": [
                        "remote-elasticsearch"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogging",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "collection": {
                    "type": "vector"
                  },
                  "logStore": {
                    "elasticsearch": {
                      "nodeCount": 3,
                      "redundancyPolicy": "SingleRedundancy",
                      "resources": {
                        "requests": {
                          "memory": "2Gi"
                        }
                      },
                      "storage": {
                        "size": "200G"
                      }
                    },
                    "retentionPolicy": {
                      "application": {
                        "maxAge": "7d"
                      }
                    },
                    "type": "elasticsearch"
                  },
                  "managementState": "Managed",
                  "visualization": {
                    "kibana": {
                      "replicas": 1
                    },
                    "type": "kibana"
                  }
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          console.openshift.io/plugins: '["logging-view-plugin"]'
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:54054faf0d2f375dd6d7fc7929a2dbb529508e1b3c004a56fc97c2a682d16be2
          createdAt: "2025-03-14T20:19:21Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing your aggregated logging stack.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.7.0-0 <5.9.15'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Logging (team-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.


              There are built-in input names for common log categories, and you can
              define custom inputs to do additional filtering.


              There is a built-in output name for the default openshift log store, but
              you can define your own outputs with a URL and other connection information
              to forward logs to other stores or processors, inside or outside the cluster.


              For more details see the documentation on the API fields.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.logging.openshift.io
            version: v1
          - description: A Red Hat OpenShift Logging instance. ClusterLogging is the
              Schema for the clusterloggings API
            displayName: Cluster Logging
            kind: ClusterLogging
            name: clusterloggings.logging.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates and manages the aggregated logging stack as a cluster-wide service.

          ##Features
          * **Create/Destroy**: Launch and create an aggregated logging stack to support the entire OCP cluster.
          * **Simplified Configuration**: Configure your aggregated logging cluster's structure like components and end points easily.
          * Configure log forwarding to an in-cluster, Red Hat managed log storage solution based on LokiStack or Elasticsearch

          ## Prerequisites
          ### Red Hat OpenShift Logging Namespace
          Cluster logging and the Red Hat OpenShift Logging Operator is only deployable to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - fluentd
        - logging
        - aggregated
        - efk
        - vector
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Fluentd
          url: https://www.fluentd.org/
        - name: Vector
          url: https://vector.dev/
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:54054faf0d2f375dd6d7fc7929a2dbb529508e1b3c004a56fc97c2a682d16be2
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:54e3a5efaae5ecc2a08db03a421aa79e323b48c764115ea84d83a37c21d1d997
        - registry.redhat.io/openshift-logging/fluentd-rhel9@sha256:a676c9164ec2523b2002dc5eba234ce774926dbf7fe87de6d72b53f99cbf1db5
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:865572275b79462f14e84c69438a61afe9465cfe45ec164caf30965730b276f7
        - registry.redhat.io/openshift-logging/logging-view-plugin-rhel9@sha256:8023b072b45e9a21ddef83a56f4760d476053250d5a920d8eb5f03eecab8b1b3
        version: 5.9.15
      entries:
      - name: cluster-logging.v5.9.15
        version: 5.9.15
      - name: cluster-logging.v5.9.14
        version: 5.9.14
      - name: cluster-logging.v5.9.13
        version: 5.9.13
      - name: cluster-logging.v5.9.12
        version: 5.9.12
      - name: cluster-logging.v5.9.11
        version: 5.9.11
      - name: cluster-logging.v5.9.10
        version: 5.9.10
      - name: cluster-logging.v5.9.9
        version: 5.9.9
      - name: cluster-logging.v5.9.8
        version: 5.9.8
      - name: cluster-logging.v5.9.7
        version: 5.9.7
      - name: cluster-logging.v5.9.6
        version: 5.9.6
      - name: cluster-logging.v5.9.5
        version: 5.9.5
      - name: cluster-logging.v5.9.4
        version: 5.9.4
      - name: cluster-logging.v5.9.3
        version: 5.9.3
      - name: cluster-logging.v5.9.2
        version: 5.9.2
      - name: cluster-logging.v5.9.1
        version: 5.9.1
      - name: cluster-logging.v5.9.0
        version: 5.9.0
      name: stable-5.9
    - currentCSV: cluster-logging.v6.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.14-v4.17
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:3e36ea9a808945d3817d169f736b1758875f2fbc6a7ae845266e1545a1b713c7
          createdAt: "2025-05-14T12:32:39Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.8.0-0 <6.0.10'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.0
          operators.operatorframework.io.bundle.channels.v1: stable-6.0
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:3e36ea9a808945d3817d169f736b1758875f2fbc6a7ae845266e1545a1b713c7
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:4e9936c090818d1d8f812487a9a8218fda5b11f8c06167fffe35bae1bafcadf9
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:15487d7b9d8f83d0df0b6fbf380f20df3518dfb1b831fac4c3209e811494b436
        version: 6.0.10
      entries:
      - name: cluster-logging.v6.0.10
        version: 6.0.10
      - name: cluster-logging.v6.0.9
        version: 6.0.9
      - name: cluster-logging.v6.0.8
        version: 6.0.8
      - name: cluster-logging.v6.0.7
        version: 6.0.7
      - name: cluster-logging.v6.0.6
        version: 6.0.6
      - name: cluster-logging.v6.0.5
        version: 6.0.5
      - name: cluster-logging.v6.0.4
        version: 6.0.4
      - name: cluster-logging.v6.0.3
        version: 6.0.3
      - name: cluster-logging.v6.0.2
        version: 6.0.2
      - name: cluster-logging.v6.0.1
        version: 6.0.1
      - name: cluster-logging.v6.0.0
        version: 6.0.0
      name: stable-6.0
    - currentCSV: cluster-logging.v6.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.15-v4.18
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f7101c962de4e5643bd7f760104a41cc48f020889159a97e2d6711703fbbec64
          createdAt: "2025-05-14T12:38:55Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.9.0-0 <6.1.8'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.1
          operators.operatorframework.io.bundle.channels.v1: stable-6.1
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f7101c962de4e5643bd7f760104a41cc48f020889159a97e2d6711703fbbec64
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:96e438dd5c080cfa47080410f375dc71da24f6d046cf609db41c8b873d85404f
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:76037c744b34b1c4fccf18f17888813d9670bb6c061c2c3de843aa43c0497170
        version: 6.1.8
      entries:
      - name: cluster-logging.v6.1.8
        version: 6.1.8
      - name: cluster-logging.v6.1.7
        version: 6.1.7
      - name: cluster-logging.v6.1.6
        version: 6.1.6
      - name: cluster-logging.v6.1.5
        version: 6.1.5
      - name: cluster-logging.v6.1.4
        version: 6.1.4
      - name: cluster-logging.v6.1.3
        version: 6.1.3
      - name: cluster-logging.v6.1.2
        version: 6.1.2
      - name: cluster-logging.v6.1.1
        version: 6.1.1
      - name: cluster-logging.v6.1.0
        version: 6.1.0
      name: stable-6.1
    - currentCSV: cluster-logging.v6.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.16-v4.19
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
          createdAt: "2025-07-29T15:25:15Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0-0 <6.2.5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.2
          operators.operatorframework.io.bundle.channels.v1: stable-6.2
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Installing logging](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html-single/logging/index#installing-logging-6-2)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:c76c853565fc93da88ba0116425839cfb86cab63cbde9856d7cca78278d7758a
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:502a5d2891049bbc32115fda28086841351cd9e6a3110b4c3c412f683f2b2491
        version: 6.2.5
      entries:
      - name: cluster-logging.v6.2.5
        version: 6.2.5
      - name: cluster-logging.v6.2.4
        version: 6.2.4
      - name: cluster-logging.v6.2.3
        version: 6.2.3
      - name: cluster-logging.v6.2.2
        version: 6.2.2
      - name: cluster-logging.v6.2.1
        version: 6.2.1
      - name: cluster-logging.v6.2.0
        version: 6.2.0
      name: stable-6.2
    defaultChannel: stable-6.2
    packageName: cluster-logging
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhacs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhacs-operator.v3.74.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
          createdAt: "2024-02-06T23:06:13.271846+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.73.0 < 3.74.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:2c0840f4573373a19eb8c9a26501138ef0e76efbd508fd69a6260e78194d86c8
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:0482d4464d6d39c01abab8df882604dd227c52b5420d36fe7414aaff8067cbec
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:71b9fdcaf4cc761b1499be8e6ccff6c463cc6b9856e9deb1f93baad45ea59eb0
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:74624389ac3261ed699b1f64e5717353da68895709d30a011bc8813b61a4a249
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:120b9891ee9db839ab0b2029a62e67ec4747c47c6ffe15af38e9035160f57b78
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:22bf00e1457b4a9e63901892aa4375a1d4a9d1b2c067c1f054d918a24b222b10
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:a108e9a7967001e0bdfed7124b01e9a3dbf1dec308444762973030aec57e1d5a
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:87eb62fe8aa7952d0d69dc13c4867ac2f51e821f56d149a9b588c826caec4c16
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d546f2d3f668ce2f20e594394ade0a86d808f6c8c88cadbba227083d4abac940
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:264102d4bca7d9bd384dc04e5bbff1a7b43343ea8e59cc79047cdda50a3f299f
        version: 3.74.9
      entries:
      - name: rhacs-operator.v3.74.9
        version: 3.74.9
      - name: rhacs-operator.v3.74.8
        version: 3.74.8
      - name: rhacs-operator.v3.74.7
        version: 3.74.7
      - name: rhacs-operator.v3.74.6
        version: 3.74.6
      - name: rhacs-operator.v3.74.5
        version: 3.74.5
      - name: rhacs-operator.v3.74.4
        version: 3.74.4
      - name: rhacs-operator.v3.74.3
        version: 3.74.3
      - name: rhacs-operator.v3.74.2
        version: 3.74.2
      - name: rhacs-operator.v3.74.1
        version: 3.74.1
      - name: rhacs-operator.v3.74.0
        version: 3.74.0
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: latest
    - currentCSV: rhacs-operator.v3.62.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:56905c2e7df0e29ca2b6301cd81b4f9bbe02b483374708f7f7da219a4b29428d
          createdAt: "2021-08-13T17:10:01.410173+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.61.0 < 3.62.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.5.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main page before deploying, or consult the RHACS
              documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create an cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead. If you want to deploy Secured Cluster Service in another namespace, additionally pass the `-n <namespace>` flag.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e30f7835cf52dbce9afdbaf55e5a93d8baf9c4b3a3ae488d1d42c1adba7f46af
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:56905c2e7df0e29ca2b6301cd81b4f9bbe02b483374708f7f7da219a4b29428d
        - registry.redhat.io/rh-acs/collector@sha256:98cfab15684506f1d0978eb701e8de478d91fc5bda3013ffd3b436aa95f07cc0
        - registry.redhat.io/rh-acs/main@sha256:9932304bb62b084ee1ae7e75bb21d6dfb737260c851f2d8daf05082f79230c66
        - registry.redhat.io/rh-acs/scanner@sha256:266a575c2547a3223890e8b73db0f14d69dc5a22217c7e4e3dc79644e6ab0258
        - registry.redhat.io/rh-acs/scanner-db@sha256:ff877f503e08ecca0ce7a7d8db760b2e4004e8e8ae2f27fb9b9c00a05b58c93a
        - registry.redhat.io/rh-acs/collector@sha256:558c0dabdc52efaee7bc889ef397ac2ac30f7c54562c2cc0070d0ef991a8b31b
        version: 3.62.1
      entries:
      - name: rhacs-operator.v3.62.1
        version: 3.62.1
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.62
    - currentCSV: rhacs-operator.v3.64.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fa478167f2f738f244bed6a06450679be73f8464c40f86acde4ffd390505692b
          createdAt: "2021-09-22T05:10:01.286657+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.63.0 < 3.64.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0cacd00c0a147cc8295138565ef2dc2ddbd9ad1fb25f6be817719a75ea4c6448
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fa478167f2f738f244bed6a06450679be73f8464c40f86acde4ffd390505692b
        - registry.redhat.io/rh-acs/scanner@sha256:8b7121bce8f9c7133d29245ebb265a4b8f9831a66a57f11b2d68b9b8b5cf79cc
        - registry.redhat.io/rh-acs/scanner-db@sha256:67e81bf896267b7acab8cb6755cde552173de4f0881c38eaca8e491bd926cc53
        - registry.redhat.io/rh-acs/collector@sha256:2b3a8bd3af62516a6714fe612e5fdc0433b91d873182e290d3ddc2b32ce291f0
        - registry.redhat.io/rh-acs/collector@sha256:3eb099da081cf208d46be5ccfa7b5e9ac0e2cad49f127fa4db91e49e3f26db42
        - registry.redhat.io/rh-acs/main@sha256:6aeb2223124540ac217c38b016dea99d770ab897dde4906a821d8776d91ea03d
        version: 3.64.2
      entries:
      - name: rhacs-operator.v3.64.2
        version: 3.64.2
      - name: rhacs-operator.v3.64.1
        version: 3.64.1
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.64
    - currentCSV: rhacs-operator.v3.65.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:229c7e5f967c70dfbc3cf638d521267ea6acf875b055499dfd30cb19c0b1f69c
          createdAt: "2021-09-21T13:26:28.711610+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.64.0 < 3.65.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/collector@sha256:3eb099da081cf208d46be5ccfa7b5e9ac0e2cad49f127fa4db91e49e3f26db42
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0cacd00c0a147cc8295138565ef2dc2ddbd9ad1fb25f6be817719a75ea4c6448
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:229c7e5f967c70dfbc3cf638d521267ea6acf875b055499dfd30cb19c0b1f69c
        - registry.redhat.io/rh-acs/main@sha256:e672ef9a1239065f8cf0f331757861fea082e971a9e9760d1e65795f399cbde1
        - registry.redhat.io/rh-acs/scanner@sha256:bfe9b2205bfca6aebed14827b4bda098bfad574e67ffa967ad339543062200fe
        - registry.redhat.io/rh-acs/scanner-db@sha256:20f258500ac2ba4e626df92dbcd193cf7ab8687ac251591be2442b8a46871835
        - registry.redhat.io/rh-acs/collector@sha256:2b3a8bd3af62516a6714fe612e5fdc0433b91d873182e290d3ddc2b32ce291f0
        version: 3.65.1
      entries:
      - name: rhacs-operator.v3.65.1
        version: 3.65.1
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.65
    - currentCSV: rhacs-operator.v3.66.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:10cdb0ff3e67b65737cde420536494bd84c2b7ccf65192bbed2eb43eaed66a41
          createdAt: "2021-10-20T16:38:38.091764+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.65.0 < 3.66.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bd74f0b95ff2881faf6eb4520feecd067b69a759f09bbfecd467447ce2a2a25e
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:10cdb0ff3e67b65737cde420536494bd84c2b7ccf65192bbed2eb43eaed66a41
        - registry.redhat.io/rh-acs/main@sha256:97230c01e76d90c7fe35ffeaa591be3d9175f350a8fd88599aded09a1edf930c
        - registry.redhat.io/rh-acs/scanner@sha256:323ee86d6a9cbeb32270c659c1154f3a62c36535c6ea894e4e29e93eb1f4a84f
        - registry.redhat.io/rh-acs/scanner-db@sha256:490b0bf49161e4bc6de2504726aaf360e0ca9a214aea6b886fff13153bdb263d
        - registry.redhat.io/rh-acs/collector@sha256:5673c27554a71502e06e08945a4424c6ce5da27f489854f3aa452e6d31240aee
        - registry.redhat.io/rh-acs/collector@sha256:d93e13ee1d1cc4b7f82e644a7d43e560aac42bd1e06d25b476cea9a4338bf9ce
        version: 3.66.1
      entries:
      - name: rhacs-operator.v3.66.1
        version: 3.66.1
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.66
    - currentCSV: rhacs-operator.v3.67.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:893ebcb2d0c96647e56bce3b16fe32851df59cd71f2c0fc8f6e66ca3a669bb70
          createdAt: "2021-12-16T12:54:10.117625+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.66.0 < 3.67.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/scanner@sha256:2860ed271da6064149bb53568c5ad60e129918d55a2f244f9ade4a0bea4c4933
        - registry.redhat.io/rh-acs/scanner-db@sha256:6f8d47277732eb56317d576c2d06fb1f539d46d16520e6ca34abec290392cef9
        - registry.redhat.io/rh-acs/collector@sha256:57c2dfa3dbc8368076291fbc627baa01dcb5ebc50f838e02ce7ac38a128d027b
        - registry.redhat.io/rh-acs/collector@sha256:e4648723ccbd43fa5acf005d9734bfc06d41e19e1be0bde839880e28fa361e16
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86e5fa1fa294987114be200890c2e516501e424aee0fb98ece25c95e7716295b
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:893ebcb2d0c96647e56bce3b16fe32851df59cd71f2c0fc8f6e66ca3a669bb70
        - registry.redhat.io/rh-acs/main@sha256:8db7630d094b661411ebb3f35248848d3147c735677d0922f7a04de9a41eacc9
        version: 3.67.2
      entries:
      - name: rhacs-operator.v3.67.2
        version: 3.67.2
      - name: rhacs-operator.v3.67.1
        version: 3.67.1
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.67
    - currentCSV: rhacs-operator.v3.68.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:195a694dc96096166fe82e81ecb3ceb493f96cfb5029112e576e44ab39b56abd
          createdAt: "2022-06-15T22:13:22.427644+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.67.0 < 3.68.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:195a694dc96096166fe82e81ecb3ceb493f96cfb5029112e576e44ab39b56abd
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:f0d2f5f11c2a96d34d6b4baf249cbf821fe064eade73c3f95c22f3c6981ac4d1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:105d10bd50a6062c998ed107dea388dd78f0caddb0b57713c8692b1d07042774
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:081f2198ad70bfb2ec66153b50b48b950b9a5550ba996de413c5a2aff4052783
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c9d0a5a2871c24c0b1058e419f2b090429363a319856397855412fd2c0037395
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:9b9b9146d9683d6da0ff2abbc8ea82b96d945b7a8c3740d72d8c018fb2b04fba
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:0bb76363a3a25916dd4493e177e9b2d36eac9633473b07fe0fffb853fd072da3
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:711293438bab35b17d2e1f162e7f74f667a6e16dd32e45122f917b6817fefb4e
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:b5fc3dcaa72bbc8dea005189c940e1cf2d00c94a6cf7676cf0e1790e1191d45c
        version: 3.68.2
      entries:
      - name: rhacs-operator.v3.68.2
        version: 3.68.2
      - name: rhacs-operator.v3.68.1
        version: 3.68.1
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.68
    - currentCSV: rhacs-operator.v3.69.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9b2beceb255124855ba1b352d16009b2f2cd422901a19c311f0783dab1fcc3a0
          createdAt: "2022-06-21T09:07:40.705944+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.68.0 < 3.69.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:08a7d8c892fad69c589b81ad8fbd21d36c8fa4078530dba361cfa9a8f9f05950
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:9ae03669cfdc0f4fe56b021d911edb17159c07757720ae30a1f22d7f7a897c07
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:f167e928740d1c614e26948ea58f49e569f6a4fddc438cd0abe52f60538432d0
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:867613178d491f416e57975843817251490c5fbca2fc69d8b55b169bea675279
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:8db5206532e95a7ad7cb23f389155d41757aaf5f445ddce92b008a06627d3fd5
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:77f5db63e17b140a8d6b84e3822fd77463035be9c0c832b55b4977696da13ccc
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:da6dd486a20fdba61ea12d2c77ea299d9795a9f22d9057d65ff992b1f5896e07
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:b254746ed3faa1f1cbcafdf1343a1957bf0214311b36f4e892be3000cc2aecf3
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9b2beceb255124855ba1b352d16009b2f2cd422901a19c311f0783dab1fcc3a0
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:2afb2823edc6f3434e18db60ee37381b5b79f51234c1a261b6d0f8ed74cc8b63
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:cf5629823f2dd583495bcbbcf3b83b4f29f855ace3c388f5cb7b24a6d34ed267
        version: 3.69.2
      entries:
      - name: rhacs-operator.v3.69.2
        version: 3.69.2
      - name: rhacs-operator.v3.69.1
        version: 3.69.1
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.69
    - currentCSV: rhacs-operator.v3.70.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fb903bf9e110ba03e9c63977ebaf707e6884dd8fd413bcd6b793a0ed795cdd0e
          createdAt: "2022-06-20T12:30:26.999648+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.69.0 < 3.70.1'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:dfeb239931f7a33ade5c829a102cabc01f5c348e1aa35a7ddc24757cfd549629
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:df8cfdaa40b50ff2f2670d97c2115b47ae4bea443ae2831caaa468dd62f3c75f
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fb903bf9e110ba03e9c63977ebaf707e6884dd8fd413bcd6b793a0ed795cdd0e
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:6c32b86ed8bc18e384706eb6188406e66910bf34c8869bd155a60c3525d0fea7
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:d2e827a209840b7197960c8115720972237532eb9cf9d6fee7a36107053408ce
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:3d7c8b77671740e324ee6c309a84bbaa4d925e6750979fe4e46c561bc79f1b4d
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:b47c96af1c9ffe4d8728da510a93efe21adc221225fe2b822cfcace6c6672ba2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:08a7d8c892fad69c589b81ad8fbd21d36c8fa4078530dba361cfa9a8f9f05950
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:ee55a233d2a8fd77d89ec2c69f357721c7debead6470a9f494b3cd6a001829c4
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:6b14e850cda9a54618847858292ea669abff599a40692e27fed0e337c4a60526
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:e71f2f2347dd46dec790405b1bb98d54a348a4583949f142f5703be9a790f96e
        version: 3.70.1
      entries:
      - name: rhacs-operator.v3.70.1
        version: 3.70.1
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.70
    - currentCSV: rhacs-operator.v3.71.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:469b6c1592226eb9a521131ef758a33e6a6ac7c96bc67fcc903cea30469c125e
          createdAt: "2022-12-07T08:31:36.777654+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.70.0 < 3.71.3'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4a0dc8c337299d7f6440ff75c2d5219d3ab68183f4f117f008ef065757dd5d4b
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:469b6c1592226eb9a521131ef758a33e6a6ac7c96bc67fcc903cea30469c125e
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:f16f92e074f7125b88e9e2033c7c12b44421145b93d040034fa4688626869af1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:c3ec88cd99908894d0e67153779062f90ac1b007980d66040a79b2be13798a10
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:4d8bac9b3857af6ba351c7768f70ef30ef7cb99ae4df60cb226b5a793353a052
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:7f279bb6e0f34a5208d8711ba4d7b62c78f08b683815ce33f6aa28c23c4246da
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:9bc25e81c21cda6d8b535d8ac67c1286fe57a1596001850e730f8f73bf371111
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:e282a8e5e606003ea344d44ac5ee7699563848f7160eeae183aa58ad740e0e94
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:cb2d0f470968c2b725b0d176d423cc1866fd56893344d077d28f2c16581d3614
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:df7b41999b542a477848823a1d0e7e5c29ae1b00fbac035007a5a1f33fac139f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:5fdfd3690af60bd5ecf04741cd1f73259491da501700b066120976960567f0db
        version: 3.71.3
      entries:
      - name: rhacs-operator.v3.71.3
        version: 3.71.3
      - name: rhacs-operator.v3.71.2
        version: 3.71.2
      - name: rhacs-operator.v3.71.1
        version: 3.71.1
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.71
    - currentCSV: rhacs-operator.v3.72.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:3ff61e1cc77b66a22c8a8c0ab306f0e122cd74e762005285aede23ebb6c2f6a2
          createdAt: "2023-02-28T17:02:08.443613+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.71.0 < 3.72.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:575b6c3e2914751df1668ffff83cbda187ce6e56f32214414cdf036b7a62bbd8
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d56a00b3e20e60b7169f5c35d03fa6a1bd3b9980dd3ebc9d89d759e1d6d20ad2
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:cd475a6c2356d63cfe736a11114de7fedfbd8fdcfecc56224a00b6c819103be6
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:42e5f6c4ba789f124ac3ebe7b6d1a607ede8e763a35888c97ccbd0259bec4a1c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:648b14e4158da2a85eedb46b3456fb502d680d52d375ad34bf45b4b94f8b2ef6
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:3ff61e1cc77b66a22c8a8c0ab306f0e122cd74e762005285aede23ebb6c2f6a2
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:cd0c205ee44eddcdf99b84e80fcf4403851d956efc50532b9d12013bafe44988
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:0b69595405e4e598183bbf4197ac5b7593724fa6162413f225aa01a542af18c9
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:622f30b7f591194d86a5bf95deede0225e66f7b624db197f6289a5fbbbe887a9
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:c07fae3616cbc681e4d31839abea6623b4bf4dade93273201c8db7ba0d635aad
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:cb5df858e0c5cae8f77b285e23daeb83062bd72570c90ed5ea3ce4a02d2f5f08
        version: 3.72.4
      entries:
      - name: rhacs-operator.v3.72.4
        version: 3.72.4
      - name: rhacs-operator.v3.72.3
        version: 3.72.3
      - name: rhacs-operator.v3.72.2
        version: 3.72.2
      - name: rhacs-operator.v3.72.1
        version: 3.72.1
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.72
    - currentCSV: rhacs-operator.v3.73.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a8486b66f977a639d77ef110492e5ecdde47be8acaefbaa1cbb5417085c56839
          createdAt: "2023-05-26T21:27:04.033276+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.72.0 < 3.73.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:7bf9415aa4b0746d11ca5a59d4179bc382a0fac4ded969f0c4294171fde27d58
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:e6687c12d070069f655c17832d94c619c18274a259ccc59715036f9908e162f7
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:ee2a9e4ab74f5a554d1e4ff148833a53bdaab70ac1e2668cc9ee2fceca290ba9
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:082bd6f4b40209e0591be001dea358f328cd86a08691cdceb7f71ea59d861d9d
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:4d0ead8969f120a629104a92a454842bb7103aec137deb6c8209b7db9c48a1b2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:648b14e4158da2a85eedb46b3456fb502d680d52d375ad34bf45b4b94f8b2ef6
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a8486b66f977a639d77ef110492e5ecdde47be8acaefbaa1cbb5417085c56839
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:e2c89f735969d438f75a89ff63c404a8ec4361a03fde48e880fee73006b9c669
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:e4a27d3ffb173c87728fb637ca7fde92a9e78ea9ee3edcecccc2a5e3d3748b8c
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:dce5db4f162dd40c09cad114cb802cd6d7c72d0c6986581f2e74c977493f75b0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:c020a80ff82bf4d9b5cec3eeac5de9990de76041e22c444e664a20e15c46ce3b
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:5df359156b01505ce43e5f14c3b08722856bdb86f3d0d9b4fba785a2e580f346
        version: 3.73.5
      entries:
      - name: rhacs-operator.v3.73.5
        version: 3.73.5
      - name: rhacs-operator.v3.73.4
        version: 3.73.4
      - name: rhacs-operator.v3.73.3
        version: 3.73.3
      - name: rhacs-operator.v3.73.2
        version: 3.73.2
      - name: rhacs-operator.v3.73.1
        version: 3.73.1
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.73
    - currentCSV: rhacs-operator.v3.74.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
          createdAt: "2024-02-06T23:06:13.271846+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.73.0 < 3.74.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d546f2d3f668ce2f20e594394ade0a86d808f6c8c88cadbba227083d4abac940
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:264102d4bca7d9bd384dc04e5bbff1a7b43343ea8e59cc79047cdda50a3f299f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:0482d4464d6d39c01abab8df882604dd227c52b5420d36fe7414aaff8067cbec
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:2c0840f4573373a19eb8c9a26501138ef0e76efbd508fd69a6260e78194d86c8
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:74624389ac3261ed699b1f64e5717353da68895709d30a011bc8813b61a4a249
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:87eb62fe8aa7952d0d69dc13c4867ac2f51e821f56d149a9b588c826caec4c16
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:120b9891ee9db839ab0b2029a62e67ec4747c47c6ffe15af38e9035160f57b78
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:22bf00e1457b4a9e63901892aa4375a1d4a9d1b2c067c1f054d918a24b222b10
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:a108e9a7967001e0bdfed7124b01e9a3dbf1dec308444762973030aec57e1d5a
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:71b9fdcaf4cc761b1499be8e6ccff6c463cc6b9856e9deb1f93baad45ea59eb0
        version: 3.74.9
      entries:
      - name: rhacs-operator.v3.74.9
        version: 3.74.9
      - name: rhacs-operator.v3.74.8
        version: 3.74.8
      - name: rhacs-operator.v3.74.7
        version: 3.74.7
      - name: rhacs-operator.v3.74.6
        version: 3.74.6
      - name: rhacs-operator.v3.74.5
        version: 3.74.5
      - name: rhacs-operator.v3.74.4
        version: 3.74.4
      - name: rhacs-operator.v3.74.3
        version: 3.74.3
      - name: rhacs-operator.v3.74.2
        version: 3.74.2
      - name: rhacs-operator.v3.74.1
        version: 3.74.1
      - name: rhacs-operator.v3.74.0
        version: 3.74.0
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.74
    - currentCSV: rhacs-operator.v4.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9bf3a3e0fd573f91f03794eaf5291364a90392e5e7d4b5d4ed550ef4600ab826
          createdAt: "2023-10-20T07:04:50.531740+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.74.0 < 4.0.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:08f537cba262605a0eb0a510b6faf78e3ba26465f636d03cc1f7a6ed6b40d355
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:eee1aa054d4d15b23be293f31d3ea659e0ffc1d05174ef89a6c977dd3d4377cc
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:4c848eb7465e928251e05d32ffbc93cd14e7bb77972d22b8721bd6eaa4a93d2f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:3250d89e78de852a2627593f29469f0b3ff639ea3644b4b7d6d057d0c1ad3659
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:5a492b856e0b6d9fdb020b556d13186ef5e7095b0c736a2031500c9c00dad0c1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:139d80d398e7e5cfc85839e3039a3630b48c7ae6650d755754430fccd94d6a27
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0aedd8e865f72ffa8b79fd2bab48974caec3eea3e3a6f77c89fe6dacb8947057
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9bf3a3e0fd573f91f03794eaf5291364a90392e5e7d4b5d4ed550ef4600ab826
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:cfa10bb6b461f302674f185b26c8d6efaaeba7f3f65d2cf05ae1f069aa2710d3
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:c39621837cbc8419d187a43a9575ec8cb4021c58ea1b56f215003d7264eb5d03
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:09e79b5db8977520f9951cc43c339c8e40b4659d7da513347d5a88b2500f809d
        version: 4.0.5
      entries:
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.0
    - currentCSV: rhacs-operator.v4.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:8ffb6f0ff1bf016c3919c1eefc6c3db3d4a8fbe0e79919f72b4c281afe6d4b1c
          createdAt: "2024-01-13T18:55:22.011709+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 4.0.0 < 4.1.6'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:364c107c34694ca33ac724de78db3e41ee55f9098d9ade0775a51b11f1533065
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:8ffb6f0ff1bf016c3919c1eefc6c3db3d4a8fbe0e79919f72b4c281afe6d4b1c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a8ad0e76ce4cae9f9b36aecf125c4ed56fc2d776fe3df113d41bfc1c2f8b3602
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:8547b76cf0533dfe6da2aafbbd063021bad4ff78b32b2dfed4a7607496e0a904
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:8760d6dc68c86f0c5b53d8f0bfada0df012f80bc8e0023ba3522fe6829a05a41
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:762ad216e79be0837e98873202b0072c7fd399925ee94bde57cc61c68cbd5f4e
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:01b9f9798d3ecbefe3da1ac2ac782e8e8b930e7eeceb82fe3473b5bfc1ccd40a
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:7ac504b7824d17bd7ac55b1e59c26be308506f315a0ab0d8c45d3b4934c0a16c
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:2ec62f3da93a1346ad215744257820377ccf4ead84e7471c52e348a31a431d3c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:bdbe08ef2d01b114702780fab76905e6eb1851684692ffa0a4c84604bda3ddb3
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:34694b9f4fe7fffc0c547be4e311579b5bb0c1fa92d8257bf6d1581ffd655c46
        version: 4.1.6
      entries:
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.1
    - currentCSV: rhacs-operator.v4.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4253eca84cbb0c5fc033507ddc2deef235b054c0ba944fa271ebbba21545b872
          createdAt: "2024-03-13T14:51:55.084122+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.1.0 < 4.2.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:b5a15f2985b114bf9bc24a0d9942db5d0cb818e60fd15a2f9094aa00904cce77
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4253eca84cbb0c5fc033507ddc2deef235b054c0ba944fa271ebbba21545b872
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:12db245af3b8475c9b5680bb0035c566109819e05cfe62bd1dfdfb5d6ed25bea
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:b82f7f654235688fc07fb037422a73e1106e27024d39a68ef7fac1422b69496d
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:ff27f03064633d6d08f8383649069e39c868def4aeca8f2622750397e049d0ec
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:2f326cabf03f780a3e6120995d5442641764641f62430d82e05edbe21b30c7b2
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:3858ab936d4626a11bbf895d41bad181364ee5df3312f6dd29ebeb9b7d31ec77
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:e2b07cb7d690477d78a2b8e1875545998d4bdb946f80dd0533a14187da99dbbf
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:59f1862c8ce188906499f08201ae05deeb47bd4fc5eab4be0386ca44ec065da8
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:e79a5e5681ebfe2d6349e2763ee0c1fa521b20319846958e402d23229b1f257b
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:4b35f5f39865238a8e95d3cba7ebb5f15ccfbf6a2563c34f57d91a0b5c927066
        version: 4.2.5
      entries:
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.2
    - currentCSV: rhacs-operator.v4.3.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:266cbbc270efafa26d5d15f3e333c96eaedee3b47854ff337ae1c572d34901ce
          createdAt: "2024-06-06T22:31:19.365878+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.2.0 < 4.3.8'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c36a557c9dc2ae4479c432933a9a4dbef38f2be53360a68a5b6befb614d01dac
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:b0c9b3858fc0547dcb4eb3065dedd5bbcf1e1f6707a8ab812ed06882d4f6d6ad
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:b7eb17fbcf6c4a6e4205ad66f1ffd9b10a4634f6ef5045d407c266566b1f255d
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:bdf9785bff51194636c6e8a518448c6d12a297bd31c4bb2e79d07084409c18b4
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:266cbbc270efafa26d5d15f3e333c96eaedee3b47854ff337ae1c572d34901ce
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:9fba889021a876ab627ff8c496e4b6c46c221c61142cd0c76daf33c89762e2c1
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:10d0cc82b6eddd089f3a7af37b872c33248513d0b4d3f094b6f62c01f4460825
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:6b55a25eb3a611c648d9be883828834592229c16f4166fd28ef40a25fc667061
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:2c35cb6f518393d002edd2bf5f714c1d053cd0fab55f6152a2f35676e4979202
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:2bb643651c8d51525223df3b2b675947174bbe9f1eaea8688e1fa83044cfa41f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:5092c3fbb4a38c06a2f42d125dc718ef0e1005f304ad8046afdd2a5e7ccbc580
        version: 4.3.8
      entries:
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.3
    - currentCSV: rhacs-operator.v4.4.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:2b87569d7f4ef5e06c989b273c378e007c6d87b26eec41774ac1fc67dcd9e4e4
          createdAt: "2025-02-10T19:50:13.597164+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.3.0 < 4.4.8'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:65707fdad55767ff5598a8d3679fd0779a37f59501e58971b81bebcb23f1f9fc
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:d6b4f8caf342b105c8df14c5c9180cb36123e08d34756ee8fe4f2f76c34f8064
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:664a5dc40dc1fa06c47a11e13b0ab5e0612a981d7b69bcfb29422abf3d1a13b2
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:39cfd01236e905034f6645a73063427436c14e351e0e0bc735e73e62aa5a1c3c
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:467d069ee957aea07882b83b60ae9fcc62ccca159f5d643aac13e531ccf7df88
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:2b87569d7f4ef5e06c989b273c378e007c6d87b26eec41774ac1fc67dcd9e4e4
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:0df4e95c7406aee7812d145cdd4d8c61cec32857fd2079dbc4244954a9d8288a
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:183c75920ab8e7cb947cb966f1543e39b6fdb5e1bb0caa1b0cf5f89671f0328f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:76f9dc2e2ee5595c36530d97dec216e2dbefada2fe0747e3f6a3a725b75722ab
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:763e7568704971cb4656727dc223374de13e82b4304dd3f5387ddc0f4e42b8b6
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:69c12230a785619c765e3749a7d56f76dce9ce4f530c20e8a9387b99eb65627b
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:ba6e25a413c79131b966a097f72c36478933ea52deb55a62acda7011a96b3983
        version: 4.4.8
      entries:
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.4
    - currentCSV: rhacs-operator.v4.5.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a7dfbe3590668f91fe6694335b5a35b2fca07b4735348cc5202921755d2f654d
          createdAt: "2025-04-09T22:51:15.157672+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.4.0 < 4.5.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:5914548f6ce2160e92c5d96580ffba23cdf1ae940c1f73f3f6424d15c6e97c37
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:dc7f23341c8b209f1846f29f889d00ef232fb58da9d5b4bd561f3933093b2786
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:d928600f75461465598a10de8df2495a275fd5232a92b078f609d1f5774eb778
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:1817cf8d2c08b04391792321702ef53d16a4b1ca06459c00c9c053128f534358
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:c2656b794ec4236e36f48f71bf26bdedd9de27fc4acef39deba2bdcfcddab03b
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:cc4b329cc795752ecbb178d2e4efc152bde54b04181334eb5bd70304c3354cad
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:0ff948b5517df3773cae2df02b708c348d8707409899b1de4108588e31601c0c
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a7dfbe3590668f91fe6694335b5a35b2fca07b4735348cc5202921755d2f654d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:5109b304eb9458f97af85c2ba63e08f51672fa3f9406226b7f357b86b5828dc0
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:fde78d34ecd2944c7c16b4e823c76b3667bf6c5b468e9991128ecc4272ec0e2e
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:ed33e7396357f1db03a065573c020846a5872b12a21284439770686689222434
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:f07cba0738897e0e7ae74971be0470ce2a855657f133a6e1cdd8a7092f6151b0
        version: 4.5.9
      entries:
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.5
    - currentCSV: rhacs-operator.v4.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:c4d434e1ff052508e5594d2b95dafa8726fd8adf5317754774c013a46de224d8
          createdAt: "2025-09-26T09:24:28.046251+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.5.0 < 4.6.10'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:b5cd08349931669055f03d21331499ef2ea04597f8a90fb201615d43000271bc
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:ffe41363dc00b05b0daa0c6d16efb30e534086aa4e3d7ccc599c86d33813138c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:713e61d3c8b55de1edc9d7834d3e8576718e1327665f94a6b58178fa7496c0cc
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:65793c05aec8085a53fab8bcf33a85192c17b46c91a7d86e0a457d1e3a113878
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:899da9c23000bb821bc9d544ebb56bcda4f3245ceee5b1b9c12d09f580b4b9a1
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:8fad8ab2efdcbc99d9efa736a90cca6f236f0a9d45fdb355ae0b6b4dc42d77f0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:90c47c6f95571cfcca04aa4e05a18befad75c1b44a6f3a055fcdfdab749d29c5
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:16b61ca1b9ee4e0821e7fc82fe4def75ed22950e502730f5704ec4fe63bb2e59
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:c4d434e1ff052508e5594d2b95dafa8726fd8adf5317754774c013a46de224d8
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:b630d53ad0b862151be7e14f0483ca5adac570216deae66f60e75af87a4dfd46
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:19db8c5d2425dcf18ae63b025f946fc46adaec0578e9505415cb585797069e95
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:6da59c5b82625894f6efa3fa0bc5ea0378edd4f95cf42dee772e9c8a565dba81
        version: 4.6.10
      entries:
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.6
    - currentCSV: rhacs-operator.v4.7.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:5b60d8271d6fa70df3c8f7e0ecd0d64feecfdf674dc84036500fe5f145c9415a
          createdAt: "2025-09-26T11:05:05.900682+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.6.0 < 4.7.7'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:def71e1ddea5b8c1e54c8582f39dc9ecca855b38aa9423690150aa6ef0ff189f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8882c00c9be9b579654a9ee8cd0c38537cf92d2f4bd8ebded1678aa0b6ae827
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:34ea97c6e9fc2d23427d0c6f2628075c0eaa0e151ca1c381539b039c249217f7
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:98553e6f18d3c03a29212b4f7ea1d30a42bcf8b52ca12aecbcbc47769fc35b53
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:bac79889d4717233e89bb6c8fae6732ed7bec5b9a55fd3b3de1f1311d870d2b1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:6bb454d64a9e339b1aeb81ca17b26f33db4a84f19ced4ad81858e40fd01878dc
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:5b60d8271d6fa70df3c8f7e0ecd0d64feecfdf674dc84036500fe5f145c9415a
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:7596529cbdce7118226b79d69e0667ab35ee692569e4014b74abdf0fa7415222
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:5ed9902b52830be93436d7977c10e73d346b6290532f6d7245cfe7cca9c93b31
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:0c32318ee4cbd547e2c1fd64433cd23a629d1b06523d7c0f07ddb98c2baaa708
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:d8fcfe1267495801d23a0b94101c7b4d06b48346a949c97946ebcbffe148f4b6
        version: 4.7.7
      entries:
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.7
    - currentCSV: rhacs-operator.v4.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
          createdAt: "2025-09-12T11:06:11.361461+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.7.0 < 4.8.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:adf47cb6615d344533fcca0aee41774f5e8df9b952b9b89a57c2be2b7e0fc03c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:92ff7691963e620dc5bf6da0a7694e6bcd68c2545afc1642a59a443ba2ab5611
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:db691065e46f604b980b0b202f6c2e14c8b204094d784c4a15db1e4fecb2965e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:97df59280c12727a99fdec0c9bf56eb874e6a1cacd12d849709797d4be635ce9
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:e5c200f28187c1f6971562e41cae969183ee15f3d0feb076bb202b22ff67cbcb
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8964b1441251607b23f92918e9158724174013dc4af423988f56d6098dfb2d5
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:e949830f3034c41aa5434a3654b2b1ccba7dac0373bf43d7c50a9ea43bb686fb
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:16912df2c26e5fc2e6051c0501170722856db274415063608de716cf53851038
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:d1758b76c32f03b5a1a44951d18fb2bb0c4fd0a99d24f8c42fe9e6b059574889
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0c92d5b5508032d77559606927e45686179eaafc59e16d5496e088f2f212d258
        version: 4.8.4
      entries:
      - name: rhacs-operator.v4.8.4
        version: 4.8.4
      - name: rhacs-operator.v4.8.3
        version: 4.8.3
      - name: rhacs-operator.v4.8.2
        version: 4.8.2
      - name: rhacs-operator.v4.8.1
        version: 4.8.1
      - name: rhacs-operator.v4.8.0
        version: 4.8.0
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.8
    - currentCSV: rhacs-operator.v4.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
          createdAt: "2025-09-12T11:06:11.361461+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.7.0 < 4.8.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:e5c200f28187c1f6971562e41cae969183ee15f3d0feb076bb202b22ff67cbcb
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:e949830f3034c41aa5434a3654b2b1ccba7dac0373bf43d7c50a9ea43bb686fb
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:16912df2c26e5fc2e6051c0501170722856db274415063608de716cf53851038
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:adf47cb6615d344533fcca0aee41774f5e8df9b952b9b89a57c2be2b7e0fc03c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:97df59280c12727a99fdec0c9bf56eb874e6a1cacd12d849709797d4be635ce9
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:d1758b76c32f03b5a1a44951d18fb2bb0c4fd0a99d24f8c42fe9e6b059574889
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8964b1441251607b23f92918e9158724174013dc4af423988f56d6098dfb2d5
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0c92d5b5508032d77559606927e45686179eaafc59e16d5496e088f2f212d258
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:db691065e46f604b980b0b202f6c2e14c8b204094d784c4a15db1e4fecb2965e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:92ff7691963e620dc5bf6da0a7694e6bcd68c2545afc1642a59a443ba2ab5611
        version: 4.8.4
      entries:
      - name: rhacs-operator.v4.8.4
        version: 4.8.4
      - name: rhacs-operator.v4.8.3
        version: 4.8.3
      - name: rhacs-operator.v4.8.2
        version: 4.8.2
      - name: rhacs-operator.v4.8.1
        version: 4.8.1
      - name: rhacs-operator.v4.8.0
        version: 4.8.0
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: stable
    defaultChannel: stable
    packageName: rhacs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dns-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dns-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSRecord",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dns-operator",
                    "app.kubernetes.io/instance": "dnsrecord-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dnsrecord",
                    "app.kubernetes.io/part-of": "dns-operator"
                  },
                  "name": "dnsrecord-sample"
                },
                "spec": {
                  "endpoints": [
                    {
                      "dnsName": "dnsrecord-simple.example.com",
                      "recordTTL": 60,
                      "recordType": "A",
                      "targets": [
                        "52.215.108.61",
                        "52.30.101.221"
                      ]
                    }
                  ],
                  "providerRef": {
                    "name": "dns-provider-creds"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/dns-operator:v0.6.0
          createdAt: "2024-09-13T09:55:39Z"
          description: A Kubernetes Operator to manage the lifecycle of DNS resources
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kuadrant/dns-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSRecord is the Schema for the dnsrecords API
            displayName: DNSRecord
            kind: DNSRecord
            name: dnsrecords.kuadrant.io
            version: v1alpha1
        description: A Kubernetes Operator to manage the lifecycle of DNS resources
        displayName: DNS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dns
        - kuadrant
        links:
        - name: DNS Operator
          url: https://github.com/Kuadrant/dns-operator
        maintainers:
        - email: mnairn@redhat.com
          name: Michael Nairn
        - email: pbrookes@redhat.com
          name: Phil Brookes
        - email: cbrookes@redhat.com
          name: Craig Brookes
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/dns-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: dns-operator.v0.6.0
        version: 0.6.0
      name: stable
    defaultChannel: stable
    packageName: dns-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nexus-repository-ha-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-repository-ha-operator-certified.v3.84.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusRepo",
                "metadata": {
                  "name": "nexusrepo-sample"
                },
                "spec": {
                  "ingress": {
                    "additionalRules": null,
                    "annotations": null,
                    "defaultRule": false,
                    "dockerIngress": {
                      "annotations": null,
                      "enabled": false,
                      "host": "example.com",
                      "name": "nexus-docker-ingress"
                    },
                    "dockerSubdomain": false,
                    "enabled": false,
                    "host": "example.com",
                    "name": "nexus-ingress",
                    "tls": {
                      "enabled": false,
                      "secretName": "tlsSecretName"
                    }
                  },
                  "license": {
                    "fileContentsBase64": "your_license_file_contents_in_base_64",
                    "secretName": "nexus-repo-license"
                  },
                  "secret": {
                    "nexusSecret": {
                      "enabled" : false,
                      "name": "nexus-secret.json",
                      "nexusSecretsKeyId": "super_secret_key_id",
                      "secretKeyfileContentsBase64": "secretKeyfileContentsBase64",
                      "mountPath": "/var/nexus-repo-secrets"
                    }
                  },
                  "nexusData": {
                    "pvc": {
                      "accessMode": "ReadWriteOnce",
                      "size": "2Gi"
                    },
                    "storageClass": {
                      "enabled": false,
                      "name": "nexusrepo-storage",
                      "parameters": {},
                      "provisioner": "provisioner",
                      "reclaimPolicy": "Retain",
                      "volumeBindingMode": "WaitForFirstConsumer"
                    },
                    "volumeClaimTemplate": {
                      "enabled": false
                    }
                  },
                  "service": {
                    "docker": {
                      "enabled": false,
                      "name": "nexus-repo-docker-service",
                      "port": 9090,
                      "protocol": "TCP",
                      "targetPort": 9090,
                      "type": "NodePort"
                    },
                    "nexus": {
                      "enabled": false,
                      "name": "nexus-repo-service",
                      "port": 80,
                      "protocol": "TCP",
                      "targetPort": 8081,
                      "type": "NodePort"
                    }
                  },
                  "statefulset": {
                    "container": {
                      "containerPort": 8081,
                      "env": {
                        "clustered": false,
                        "install4jAddVmParams": "-Xms2703m -Xmx2703m",
                        "jdbcUrl": null,
                        "nexusInitialPassword": "nexusabc",
                        "password": "nexus",
                        "user": "nexus",
                        "zeroDowntimeEnabled": false
                      },
                      "imageName": "registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:fbad488ff2e368845473a11b4955992fe826a94811da3abf5cc3c6a33ff3974f",
                      "pullPolicy": "IfNotPresent",
                      "resources": {
                        "limits": {
                          "cpu": 4,
                          "memory": "8Gi"
                        },
                        "requests": {
                          "cpu": 4,
                          "memory": "8Gi"
                        }
                      },
                      "terminationGracePeriod": 120
                    },
                    "imagePullSecrets": {},
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 240,
                      "path": "/",
                      "periodSeconds": 60
                    },
                    "name": "nexusrepo-statefulset",
                    "readinessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 240,
                      "path": "/",
                      "periodSeconds": 60
                    },
                    "replicaCount": 1
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-certified@sha256:99f30aa5ba8469d96de3efe69d1b21e73ff022172fc79144bba496d1119ec8a2
          createdAt: "2023-08-14"
          description: |-
            Nexus Repository is the central source of control to efficiently manage all binaries
            and build artifacts across your DevOps pipeline.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.replaces: 3.80.0-2
          olm.skipRange: <3.84.0-1
          repository: https://github.com/sonatype/operator-nexus-repository
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus Repository
            displayName: NexusRepo
            kind: NexusRepo
            name: nexusrepos.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Repository is the central source of control to efficiently manage all binaries
          and build artifacts across your DevOps pipeline.
          The flow of open source and third-party components into and through an organization
          creates a complex software supply chain.
          Nexus Repository delivers speed, efficiency, and quality to the governance
          and management of all dependencies, libraries, and applications for your DevOps teams.

          ## Core Capabilities

          * **Dependency Management**:
            Improves reliability with repeatable, fast access to secure dependencies
          * **Developer Productivity**:
            Streamline developer workflows by enabling the sharing of components and applications across teams
          * **Supply Chain Performance**:
            Improve speed-to-market and reduced build times with release advanced staging and component tagging
          * **CI/CD Integrations**:
            Increase DevOps scalability with integrations to the most popular build and deployment tools

          Version control systems and package registries do not scale when managing proprietary,
          open source, and third-party components.
          Organizations need a central binary and build artifact repository to manage dependencies
          across the entire software supply chain.

          ## Limitations

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus Repository Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage
          By default, the example Custom Resource starts up a single instance of Nexus Repository in non-Pro mode using an H2 database.

          ### Connect to a PostgreSQL database
          We strongly recommend using Nexus Repository with a PostgreSQL database. To connect a Nexus Repository instance to a PostgreSQL database, do as follows:
          * Update the `license.fileContentsBase64` parameter with a base 64 representation of your Nexus Repository Pro license.
          * Set the `statefulset.container.env.jdbcUrl` parameter to your database url.
          * Set the `statefulset.container.env.user` parameter to your database user name.
          * Set the `statefulset.container.env.password` parameter to your database password.

          ### High Availability (HA) mode
          To start Nexus Repository in HA mode, you must edit the Custom Resource YAML as follows:

          * Update the `license.fileContentsBase64` parameter with a base 64 representation of your Nexus Repository Pro license.
          * Set the `statefulset.container.env.jdbcUrl` parameter to your database url.
          * Set the `statefulset.container.env.user` parameter to your database user name.
          * Set the `statefulset.container.env.password` parameter to your database password.
          * Set the `statefulset.container.env.clustered` parameter to true.
          * Set the `statefulset.replicaCount` parameter to the number of Nexus Repository pods you require in your HA cluster
          * You must use a persistent volume as described in the 'Storage' section below.

          ### Storage
          The default configuration uses an emptyDir volume for storing Nexus Repository logs. However, we strongly recommend that
          you configure dynamic provisioning of persistent storage based on your deployment environment as explained below.

          #### Cloud deployments (AWS/Azure)
          * Ensure the appropriate Container Storage Interface (CSI) driver(s) are installed for your chosen OpenShift cloud deployment.
            Please, refer to OpenShift documentation for details on configuring CSI drivers.
          Note: if you're using Red Hat OpenShift on AWS (ROSA), CSI drivers for dynamically provisioning EBS volumes are installed
          by default; you should see associated storage classes for them in your cluster web console.

          #### On premise deployments
          1. Attach separate disks (i.e. separate from the root disk) to your worker nodes.
          2. Install the Local Storage Operator. Please refer to Open Shift persistent storage documentation:
             https://docs.openshift.com/container-platform/4.13/storage/persistent_storage/persistent_storage_local/persistent-storage-local.html
          and the Local Storage Operator documentation: https://github.com/openshift/local-storage-operator.
          3. Use the Local Storage Operator to automatically create persistent volumes for your chosen storage class name as documented here:
             https://docs.openshift.com/container-platform/4.13/storage/persistent_storage/persistent_storage_local/persistent-storage-local.html
          4. Specify the storage class name from step 3 as the value of the `nexusData.storageClass.name` property and also set `nexusData.volumeClaimTemplate.enabled` to true.

          #### Configuring the operator for dynamic persistent volume provisioning
          * Set the `nexusData.storageClass.name` parameter to a storage class name. This could be one of the default storage classes
            automatically created in your OpenShift cluster (e.g., if you're using ROSA) or one that you would like the operator to create.
          * If you would like to create a dedicated storage class (i.e., you don't want to use the default), then in addition to
            specifying a value for `nexusData.storageClass.name`, you must also set `nexusData.storageClass.enabled` parameter to `true`.
          * Set the `nexusData.volumeClaimTemplate.enabled` parameter to `true`.

          ### Additional configuration
          * Once the operator has created the server instance and it is running, you can further configure Nexus Repository via the NexusRepo custom resource:

          | Parameter                                                   | Description                         | Default                                 |
          | ----------------------------------------------------------- | ----------------------------------  | ----------------------------------------|
          | `statefulset.name`                                          |  The suffix for the statefulset's service name    | nexusrepo-statefulset     |
          | `statefulset.replicaCount`                                  |  The desired number of Nexus Repository pods    | 3                           |
          | `statefulset.container.imageName`                           |  The Nexus repository image registry url    |
          | | | registry.connect.redhat.com/sonatype/nexus-repository-manager:<app version>-ubi-1|
          | `statefulset.container.resources.requests.cpu`              |  The minimum cpu the Nexus repository pod can request    | 4 |
          | `statefulset.container.resources.requests.memory`           |  The minimum memory the Nexus repository pod can request    | 8Gi|
          | `statefulset.container.resources.limits.cpu`                |  The maximum cpu the Nexus repository pod may get.    | 4|
          | `statefulset.container.resources.limits.memory`             |  The maximum memory the Nexus repository pod may get.  | 8Gi|
          | `statefulset.container.containerPort`                       |  The Nexus Repository container's HTTP port  | 8081|
          | `statefulset.container.pullPolicy`                          |  The Nexus Repository docker image pull policy  | IfNotPresent|
          | `statefulset.container.terminationGracePeriod`              |  The time given for the pod to gracefully shut down  | 120 seconds|
          | `statefulset.container.env.clustered`                       |  Controls whether Nexus Repository should start in clustered mode  | False|
          | `statefulset.container.env.install4jAddVmParams`            |  Xmx and Xms settings for JVM  | -Xms2703m -Xmx2703m|
          | `statefulset.container.env.user`                            |  The database user name | nexus (can be overridden via Kubernetes secret. See above)|
          | `statefulset.container.env.password`                        |  The database password  | nexus (can be overridden via Kubernetes secret. See above)|
          | `statefulset.container.env.jdbcUrl`                         |  The database url |
          | | | jdbc:postgresql://your_database_server:5432/nexus (must be overridden via Kubernetes |
          | | | secret object which feeds into env variable variables defined in the operator.yaml . See above)|
          | `statefulset.container.env.nexusInitialPassword`            |  The initial admin password for Nexus Repository UI. This is for first time |
          | | installs only. Subsequent installs/restarts on the same node ignores this setting | nexusabc|
          | `statefulset.livenessProbe.initialDelaySeconds`             |  The initial delay before the Nexus Repository application is pinged for a |
          | | response i.e. to determine whether the application is still alive | 240|
          | `statefulset.livenessProbe.path`                            |  The path to ping the application | |
          | `statefulset.livenessProbe.periodSeconds`                   |  The interval for checking that the application is still alive i.e. responsive, not deadlocked etc | 60|
          | `statefulset.livenessProbe.failureThreshold`                |  The maximum number of times the application ping can fail before the |
          | | application is considered unresponsive. Kubernetes will attempt to restart the application's pod in this case | 6|
          | `statefulset.readinessProbe.initialDelaySeconds`            |  The initial delay before the Nexus Repository application is checked |
          | | whether it is ready to receive traffic | 240|
          | `statefulset.readinessProbe.path`                           |  The path where the Nexus Repository application is checked for traffic readiness | |
          | `statefulset.readinessProbe.periodSeconds`                  |  The interval for checking the Nexus Repository application for traffic readiness | 60|
          | `statefulset.readinessProbe.failureThreshold`               |  The maximum number of times to check the Nexus Repository application for traffic readiness | 6|
          | `statefulset.imagePullSecrets`                              |  The pull secret for private image registries | {}|
          | `ingress.name`                                              |  The name of an Ingress object which listens for requests to Nexus Repository on port 80 | nexus-ingress|
          | `ingress.enabled`                                           |  Whether or not to create the Ingress | false|
          | `ingress.host`                                              |  Ingress host | "example.com"|
          | `ingress.defaultRule`                                       |  Whether or not to add a default rule for the Nexus Repository |
          | | Ingress which forwards traffic to a Service object  | false|
          | `ingress.additionalRules`                                   |  Additional rules to add to the ingress | null|
          | `ingress.tls.enabled`                                       |  Whether or not TLS should be enabled for the Ingress | false|
          | `ingress.tls.secretName`                                    | The name of a Secret object in which to store the TLS secret | tlsSecretName|
          | `ingress.dockerSubdomain`                                   |  Whether or not to add rules for docker subdomains | false|
          | `ingress.annotations`                                       |  Annotations for the Ingress object | null |
          | `license.fileContentsBase64`                                |  Your Base64 encoded Nexus Repository license | your_license_file_contents_in_base_64 |
          | `license.secretName`                                        |  The name of a secret object in which to store your |
          | | Base64-encoded Nexus Repository license | nexus-repo-license|
          | `ingress.dockerIngress.name`                                |  The name of an Ingress object which listens for requests to an |
          | | existing docker repository in Nexus Repository | nexus-docker-ingress|
          | `ingress.dockerIngress.enabled`                             |  Whether or not to create an Ingress for a docker repository | false|
          | `ingress.dockerIngress.host`                                |  The docker Ingress's host | "example.com"|
          | `ingress.dockerIngress.annotations`                         |  Annotations for the docker Ingress object | null |
          | `nexusData.pvc.size`                                        |  The volume size to request for storing Nexus logs | 2Gi |
          | `nexusData.pvc.accessMode`                                  |  The persistent volume claim access mode | ReadWriteOnce |
          | `nexusData.storageClass.name`                               |  The name of the storage class to use for dynamically requesting that |
          | | storage be provisioned when using dynamic provisioning or to randomly select a pre-created persisted volume when using the Local Storage Operator. | nexusrepo-storage |
          | `nexusData.storageClass.reclaimPolicy`                      |  The reclaim policy for persistent volume's created by the storage class | Retain |
          | `nexusData.storageClass.enabled`                            |  Whether or not a storage class object should be created. | false |
          | `nexusData.storageClass.volumeBindingMode`                  |  It controls when volume binding and dynamic provisioning should occur | WaitForFirstConsumer |
          | `nexusData.storageClass.provisioner`                        |  The name of the volume provisioner | provisioner |
          | `nexusData.storageClass.parameters`                         |  The additional parameters for the storage class | {} |
          | `nexusData.volumeClaimTemplate.enabled`                           |  Whether or not volume should be requested using a claim template. |
          | | You should set this property to true when using dynamic volume provisioning (cloud deployments) |
          | | or if you're using Local Storage Operator to automatically create persistent volumes for locally attached disks (i.e. not root disk) | false |
          | `license.secretName | A name for the Kubernetes secret object which stores the base64 representation of your |
          | | Nexus Repository license specified via `license.fileContentsBase64` is stored | nexus-repo-license |
          | `license.fileContentsBase64 | A base64 representation of your Nexus Repository license | your_license_file_contents_in_base_64 |
          | `service.nexus.name`                                              |  The name of a kubernetes Service Object which allows access to |
          | | Nexus Repository's main HTTP port | nexus-repo-service |
          | `service.nexus.enabled`                                           |  Whether or not to create the Service object | false|
          | `service.nexus.type`                                              |  The type of the Kubernetes Service | "NodePort"|
          | `service.nexus.protocol`                                       |  The protocol  | TCP |
          | `service.nexus.port`                                   |  The port to listen for incoming requests | 80|
          | `service.nexus.targetPort`                                   |  The port to forward requests to | 8081 |
          | `service.docker.name`                                              |  The name of a kubernetes Service Object which allows access to a |
          | | docker repository within Nexus Repository | nexus-repo-docker-service |
          | `service.docker.enabled`                                           |  Whether or not to create a Kubernetes Service object for a given |
          | | docker repository within Nexus Repository | false |
          | `service.docker.type`                                              |  The type of the Kubernetes Service | "NodePort"|
          | `service.docker.protocol`                                       |  The protocol  | TCP |
          | `service.docker.port`                                   |  The port to listen for incoming requests | 9090 |
          | `service.docker.targetPort`                                   |  The port to forward requests to | 9090 |
        displayName: Nexus Repository HA Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - repository
        - sonatype
        links:
        - name: Nexus Repository Manager
          url: https://www.sonatype.com/product-nexus-repository
        maintainers:
        - email: support@sonatype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-certified@sha256:99f30aa5ba8469d96de3efe69d1b21e73ff022172fc79144bba496d1119ec8a2
        - registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:fbad488ff2e368845473a11b4955992fe826a94811da3abf5cc3c6a33ff3974f
        version: 3.84.0-1
      entries:
      - name: nexus-repository-ha-operator-certified.v3.84.0-1
        version: 3.84.0-1
      name: stable
    defaultChannel: stable
    packageName: nexus-repository-ha-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-kubernetes.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "ClusterMongoDBRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-enterprise"
                  },
                  "name": "clustermongodbrole-sample"
                },
                "spec": {
                  "db": "admin",
                  "role": "rootMonitor",
                  "roles": [
                    {
                      "db": "admin",
                      "role": "root"
                    },
                    {
                      "db": "admin",
                      "role": "clusterMonitor"
                    }
                  ]
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "my-replica-set"
                },
                "spec": {
                  "credentials": "my-credentials",
                  "members": 3,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "sample-sharded-cluster"
                },
                "spec": {
                  "configServerCount": 3,
                  "credentials": "my-credentials",
                  "mongodsPerShardCount": 3,
                  "mongosCount": 2,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "shardCount": 1,
                  "type": "ShardedCluster",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBMultiCluster",
                "metadata": {
                  "name": "multi-replica-set"
                },
                "spec": {
                  "clusterSpecList": [
                    {
                      "clusterName": "e2e.cluster1.mongokubernetes.com",
                      "members": 2
                    },
                    {
                      "clusterName": "e2e.cluster2.mongokubernetes.com",
                      "members": 1
                    },
                    {
                      "clusterName": "e2e.cluster3.mongokubernetes.com",
                      "members": 2
                    }
                  ],
                  "credentials": "my-credentials",
                  "duplicateServiceObjects": false,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBOpsManager",
                "metadata": {
                  "name": "ops-manager"
                },
                "spec": {
                  "adminCredentials": "ops-manager-admin",
                  "applicationDatabase": {
                    "members": 3,
                    "podSpec": {
                      "cpu": 1
                    }
                  },
                  "configuration": {
                    "mms.fromEmailAddr": "admin@thecompany.com"
                  },
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "version": "6.0.3"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBUser",
                "metadata": {
                  "name": "my-replica-set-x509-user"
                },
                "spec": {
                  "db": "$external",
                  "mongodbResourceRef": {
                    "name": "my-replica-set"
                  },
                  "roles": [
                    {
                      "db": "admin",
                      "name": "dbOwner"
                    }
                  ],
                  "username": "CN=my-replica-set-x509-user,OU=cloud,O=MongoDB,L=New York,ST=New York,C=US"
                },
                "status": {
                  "phase": "Updated"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/mongodb/mongodb-kubernetes@sha256:745cc2c6bb893c1cb1e75d79b6f3bfc7b125a6a62f7431bdc43cf6cacd373998
          createdAt: "2025-09-16T14:56:17Z"
          description: The MongoDB Controllers for Kubernetes enable easy deploys
            of MongoDB into Kubernetes clusters, using our management, monitoring
            and backup platforms, Ops Manager and Cloud Manager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mongodb/mongodb-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterMongoDBRole
            name: clustermongodbroles.mongodb.com
            version: v1
          - description: MongoDB Deployment
            displayName: MongoDB Deployment
            kind: MongoDB
            name: mongodb.mongodb.com
            version: v1
          - kind: MongoDBCommunity
            name: mongodbcommunity.mongodbcommunity.mongodb.com
            version: v1
          - description: MongoDB Multi Deployment
            displayName: MongoDB Multi Deployment
            kind: MongoDBMultiCluster
            name: mongodbmulticluster.mongodb.com
            version: v1
          - kind: MongoDBSearch
            name: mongodbsearch.mongodb.com
            version: v1
          - description: MongoDB x509 User
            displayName: MongoDB User
            kind: MongoDBUser
            name: mongodbusers.mongodb.com
            version: v1
          - description: MongoDB Ops Manager
            displayName: MongoDB Ops Manager
            kind: MongoDBOpsManager
            name: opsmanagers.mongodb.com
            version: v1
        description: |
          MongoDB Controllers for Kubernetes (MCK) is an open-source project. It represents a merge of the previous
          MongoDB Community Operator (MCO) and MongoDB Enterprise Kubernetes Operator (MEKO).

          This release brings MongoDB Community and Enterprise editions together under a single, unified operator,
          making it easier to manage, scale, and upgrade your deployments. While the first version simply brings
          the capabilities of both into a single Operator, future changes will build on this to more closely align
          how Community and Enterprise are managed in Kubernetes, to offer an even more seamless and streamlined experience.
          As an open-source project, it now allows for community contributions, helping drive quicker bug fixes and ongoing innovation.
          The MongoDB Controllers for Kubernetes enable easy deploys of MongoDB
          into Kubernetes clusters, using our management, monitoring and backup
          platforms, Ops Manager and Cloud Manager.

          ## Before You Start

          To start using the operator you''ll need an account in MongoDB Cloud Manager or
          a MongoDB Ops Manager deployment.

          * [Create a Secret with your OpsManager API key](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-operator-credentials/#procedure)

          * [Create a ConfigMap with your OpsManager project ID and URL](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-project-using-configmap/)

          By installing this integration, you will be able to deploy MongoDB instances
          with a single simple command.

          ## Required Parameters

          * `opsManager` or `cloudManager` - Enter the name of the ConfigMap containing project information
          * `credentials` - Enter the name of the Secret containing your OpsManager credentials
          * `type` - Enter MongoDB Deployment Types ("Standalone", "ReplicaSet", "ShardedCluster"

          ## Supported MongoDB Deployment Types ##

          * Standalone: An instance of mongod that is running as a single server and
          not as part of a replica set, this is, it does not do any kind of
          replication.

          * Replica Set: A replica set in MongoDB is a group of mongod processes that
          maintain the same data set. Replica sets provide redundancy and high
          availability, and are the basis for all production deployments. This section
          introduces replication in MongoDB as well as the components and architecture
          of replica sets. The section also provides tutorials for common tasks
          related to replica sets.

          * Sharded Cluster: The set of nodes comprising a sharded MongoDB deployment.
          A sharded cluster consists of config servers, shards, and one or more mongos
          routing processes. Sharding is a A database architecture that partitions
          data by key ranges and distributes the data among two or more database
          instances. Sharding enables horizontal scaling.

          ## Requirements for deploying MongoDB OpsManager

          * In order to deploy resources of type MongoDB OpsManager, you will need to
          create a secret containing the [credentials](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/plan-om-resource/#om-rsrc-prereqs)
          for the Global Owner user

          ## Security ##

          The operator can enable TLS for all traffic between servers and also between
          clients and servers. Before enabling `security.tls.enabled` to `true` you
          should create your certificates.  or you can leave the operator to create all
          the certificates for you. The operator ability to create certs is been
          deprecated due to Kubernetes API changes.

          For more information, please read the official MongoDB
          Kubernetes Operator  [docs](https://docs.mongodb.com/kubernetes-operator/stable/).
        displayName: MongoDB Controllers for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mongodb
        - databases
        - nosql
        links:
        - name: Documentation
          url: https://docs.opsmanager.mongodb.com/current/tutorial/install-k8s-operator/index.html
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: stable
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-kubernetes-init-database@sha256:3ea9106c0e35c015bc8f15657c76d974bb1cadb9220dc58be2076c9699f26fb7
        - quay.io/mongodb/mongodb-agent@sha256:6e826bbe7dd812c101464e75295370c4b10e4e9c31477db89cfceb96759853e9
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:702f529cb9d8d9042e54263f8bc37b40c062ac2695efc3e7150819779886e323
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c48631742ad1fec8a9ee745fc3cc9ca1fe7d86ef7fa68ad8e8337c7a4566dc87
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7e5361c2a1604dec3fc671d629806fe6ddc6df524b69a441b86933a4b30093
        - quay.io/mongodb/mongodb-enterprise-server@sha256:85c1d669c073caf9cbaac815e4c8ea2ca18b71deefa03bf84cd6d8b8ebd53e12
        - quay.io/mongodb/mongodb-agent@sha256:fa30370a75e8e4ac648650ceb39a7a259a754aab8a48d7e2f2801cd1b777e91b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f1e72679b2468aed2f3c3106568327911e6594c9ee15f7348e953eaa0d8fd44a
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:85b4e954c9c4bbe1019774bad1e05e9c8dc175648b01d3badf0b64143f182019
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d323c1bcd14118cc537ac282996ea5a58756d20b6b73f1eec0efa8710d246077
        - quay.io/mongodb/mongodb-enterprise-server@sha256:67238129df09d33ee4746ff1acb5515d699a29d9f42350eae3f3f69ccb7fbf54
        - quay.io/mongodb/mongodb-kubernetes-database@sha256:f8c981a6223f70d7c6bbd30c503fc8b2497f351830b2b74c30e6f81863d78d2f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:a4590af027182305913f0de1a2e37ad8612bd73b8b31f4df11068b80089021ff
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3add8b68bfbaa42596fe5dbbee4623ba5ff1a826b18bb71301525adf574dcb94
        - quay.io/mongodb/mongodb-agent@sha256:ee9f3906d893004bb2aaaa6412646f58288b0684502b9e90f33cafadcfce4349
        - quay.io/mongodb/mongodb-enterprise-server@sha256:df6081325ea2fe01d1b27d0c70005dec8232aaf51a43d91661b722b8a1761263
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:146c4f5dc443f94f54d72808b4e4df93bb904e23589b222d17700166780776b4
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:09788976bdc1327acbf18cedf8efd5fca2159e43b506faa40e03ee42ad94928a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:655896bac9e82de8e9729d59a36c3a31f66235f22b93449260a2ccb7709de084
        - quay.io/mongodb/mongodb-enterprise-server@sha256:172e338a43df01ee5408d8d1e3cb038d2c55e77761c1a2ac6ed37959d1525140
        - quay.io/mongodb/mongodb-agent@sha256:3fa5bcfda3228e0ea880434a60ced829e3872e74763b347636c18475d866d0e1
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:33bbd29b74f95baad522b380d70bcc9f933a3fee57642c37b09376c86179c40f
        - quay.io/mongodb/mongodb-kubernetes@sha256:745cc2c6bb893c1cb1e75d79b6f3bfc7b125a6a62f7431bdc43cf6cacd373998
        - quay.io/mongodb/mongodb-enterprise-server@sha256:19925e1297ab652db59c35f53e7b7c8b2b277f8ecdc4215682961aaf2b3e924e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:f71b027fccf649f18f63f8df7a59b1bac421a2cdab6712e0c985a93c08ed970a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8445d59d491a6d3890ea9c8b421214bcd0fdcfbc420ed40497a790b79f0bc89e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:895dbdfb87ddc2e9693671c7108a31ab28e16fec2e5be6d046bdd2e30319eade
        - quay.io/mongodb/mongodb-enterprise-server@sha256:dfd4277f0c16ed476ac7de51894739e8c8d13a65ea5024b9a97b3dce0f6c6975
        - quay.io/mongodb/mongodb-enterprise-server@sha256:fb8397cbcdbf7c33f6ad142492792771127f0d9a8c46388f350b6927405ee590
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7885ab6e88214f9289d8babf5189136dde0dd4c84a15a5481ccd3ef512b2e66f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:5d299820bb903f0cdfdf2626a3c44099d2eaee16188cd8234d2ff3a57259b3a7
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:439d6fefb469613432e2d05149dfc96c32b16469d743835d5f402a2ada8be8cc
        - quay.io/mongodb/mongodb-agent@sha256:da77dcb60b927bf6059c2c9bf008f8c3c0eec45cf7d1b40c34a0c263e8cc18b8
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b8aa1edbc5fb02af312a863bc7db8535a282dfd692a07b1e321e5a1acbfeb6a1
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9a3a6496be0d2d446e6d654eb56386c7726e0cd2d7232f6dcec579923edf0014
        - quay.io/mongodb/mongodb-agent@sha256:89819704d70a3b6b4d2358745a22060ef740b33ad5c6debf8e6ac4d4a6fd63ba
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b7764c316eeacffd3f9ddae62d9fd2a74eb58484d2cfd5021e460095d604764f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:e7aeedb752dda77cf0e49df3430843b5a2bba64c8404628a983f46640d316808
        - quay.io/mongodb/mongodb-search@sha256:7149a7b6f1ee84646d3972837920f80fc5db49503c598406ce92e2f794bd5f0d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa8683500e8d3632c4588bbcdddb4e3186a01f48aefe1b6d95f8860dae8b6c76
        - quay.io/mongodb/mongodb-enterprise-server@sha256:15e31285c8949a4666f7702df53416b0905d76945a28575a6028a876cc8a6f0f
        - quay.io/mongodb/mongodb-kubernetes-init-ops-manager@sha256:93c4633cbc84e9bfa3563771ac2efbf5e6515f380dec2c085494cb8326546377
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5c8553e5ba115657fb878c0fe08704af9fd5decec572a87d89d2c8c156ffc9f9
        - quay.io/mongodb/mongodb-agent@sha256:5b064c470335df840b050b714b28b75b6061444843250f1eb9215cb2b8c1a149
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a86c4d8a0b9ca3c20cc524809676eb03a944ada43378ec98c107cd72cc64700f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:5e3982493e85ac1e40536758c3145b7c2fff9a7b4be9d053ee41201363a89920
        - quay.io/mongodb/mongodb-agent@sha256:5f5031e4c2a3729f618d026b0a7f34fd4509ea73d5a4e3068b551041dc291583
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:d81eac2129b6ab8346be1749c56ea67f6da64d4ee0fb9f3ff4171ba0f61918eb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3316015ac0699d66ac73f714e5dc0784738654e0349e80b1e1f4e953b3c88964
        - quay.io/mongodb/mongodb-agent@sha256:b387b7458f5640f907c9074fcb08ae4d4785be31d4bb65f2d7c459cd6762b2d6
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:adeab0dc766a27138db7df5f37cbb1da860c14179ef631b2e001e5bb875325e4
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ef732e3eea91cc86caf2eabedc0c9115ed235a8d204f33e74bbf090a0b0259fb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a48a4d5c275fd669a27469649fb39ea38c728dbcf9f0a9da4e7cf27a2ce912ed
        - quay.io/mongodb/mongodb-enterprise-server@sha256:720056b985c1a16d48b2aecb7a88a15cba3d8186e55b3c6c24729699a85c8232
        - quay.io/mongodb/mongodb-enterprise-server@sha256:523566f44cbb6bbf0cd12b1adae939a883f2254c78dc4b8c9d0098581337b1db
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f2d9a3264a8814f33143b73c56624572fb80afeab184133d3a901713a31c1a87
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5145c97d2281aabcd8538ff515c3833aa48acb66b0b739fb1881120c1985d88a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9bef77e905dadd1e61d0439dd53b21ed1f3606e49affd9ffc4ba8156578e8f0f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:cd869ff7d661059fad7593e30141d74e8cc751295b1579577f7a853396eb5d0b
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:fa4aa97767bd19cf3711a74bf21ecb9cbbc25fed6b6b09a2eb241ebd3fbc8f3a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d49417ea85ee1ddb84429552a5aac87a82c2843904d8c0300a43b657b7a49e0e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:33a30d11969eb21952681108092fe7f916e25c8e1de43e6a4e40e126025a2bec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:778f9e29e36d6687cd17dd95ec1f8127d2a732bde1a2dd5821212eb55c1c998a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1d418c65d82ccad205b7889be4ba161d3cbfa8b1fb75b21c8fcf02de38fad9ca
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b9ab7147b59b62e40a69c6400f87d0c7faecfa5462bed6e68ca653b00e50b887
        - quay.io/mongodb/mongodb-agent@sha256:6394391f78da5e3218bb2c694e7ec9f6682fd17a6dd7e7250b0c0b5e1729796d
        - quay.io/mongodb/mongodb-agent@sha256:a8358ce465adf463d8e73916d7944765d7aaf2b0ed10bd3b8119a87fc9080667
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:78f11c9cd5e63f2b0e10c67dc1491f4ce1279fa90178c108814725414defab9a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c752a7f0726d21d55f204a359b85fa29cd727a5fe571b37289ae3ea52d6206b0
        - quay.io/mongodb/mongodb-agent@sha256:b422829aac63f068aed4a6948fd14cbd3c3bfad6d47ab76523f6839bf20b0203
        - quay.io/mongodb/mongodb-agent@sha256:6c23febaa9dd52bc29ac1e52d087fc4b2b47c578f4d6ac3a1674bb750a38cb07
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:3bbd3d7fd45d107621e7770f2b4c2a758e148cd472e0ee21719442f71bf6ade0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0fde236fe26d5b20210d6f246aadacd7845874712def516cbe2eec11d79d5181
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d2d8b59c1f39c41951e7878963f8496eb494688720366f8edc7060887cdbe6a0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7a93a0276531ff9be4c90bb8fe8d104e0a9e930c29792aafe03cc6a76a9fa89c
        - quay.io/mongodb/mongodb-agent@sha256:6cc4e953df3f09d793f0dbc6e57e55a3c5277ef63177a6d8f691a41227facf64
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3fe62f3ad52707ccf0e9993c69891e44c569a4b4e1a3244cf60e3b5fdde57bd2
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1a77851485ad3b2566e765b4bb051ff0ddfa48703492cef6c6db021dde84d87f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:bfbcfcfc5f0de22e20520b2d96f826638e424adb028a909ccc192013d11f5833
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b00ac773623e1bf422b47e6d783263d2d08737865abe9abbe106d86d1ea25165
        - quay.io/mongodb/mongodb-enterprise-server@sha256:37af0913742b6de339c7e664b30f35dcef04506529e4a09a9bbbdd57643f7149
        - quay.io/mongodb/mongodb-agent@sha256:d8a2ef79d767f03162ec57fe028f9072294b292d376af8eb97d2967eb6d7047b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7266b405ebb2dd24a9f7c12d3921024f547c3d8c01831e6bcfc1d551d63be1
        - quay.io/mongodb/mongodb-agent@sha256:af3de4f631d0b00cae7f7e80c3900378ba7e449e01830bf9dc643b873d2e6d33
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa300abf739e26350a258f643fa8e57aedacd041055c7b923bd22b21ef554cdb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d21c65780cadc9d8668d2c28935d492625d6993f13ba562722200a819281cbcc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:376ad7e5fe42651d946cba5705c0eee44aa636c1b4ff258210e4ce2fe4205b4d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:454b1585c342375dea643b10ead688d0f62347841f5f30059fcd4f92e3297ee9
        - quay.io/mongodb/mongodb-enterprise-server@sha256:11ab324698f919f84a5dab5a964d93fe63f98a48146bb3b86974cba7606d30ec
        - quay.io/mongodb/mongodb-kubernetes-init-appdb@sha256:afb0d463304fc0bfb3d8dade91459414daac6bf6fcf742347a2012cbc9fea354
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c8aec83b76dfecc05b8f5d8f40fd61b24980c0287a07294f75f321c2d42c32b4
        - quay.io/mongodb/mongodb-agent@sha256:2d5652d0e263aa94554e7a068c93e4fd94b84ec223f077361d030d172bc93f67
        version: 1.4.0
      entries:
      - name: mongodb-kubernetes.v1.4.0
        version: 1.4.0
      - name: mongodb-kubernetes.v1.3.0
        version: 1.3.0
      - name: mongodb-kubernetes.v1.2.0
        version: 1.2.0
      - name: mongodb-kubernetes.v1.1.0
        version: 1.1.0
      - name: mongodb-kubernetes.v1.0.1
        version: 1.0.1
      - name: mongodb-kubernetes.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-apigateway-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-apigateway-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Resource",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "RestAPI",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Stage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "VPCLink",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/apigateway-controller:1.3.1
          createdAt: "2025-09-19T21:13:12Z"
          description: AWS API Gateway controller is a service controller for managing
            API Gateway resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIIntegrationResponse represents the state of an AWS apigateway
              APIIntegrationResponse resource.
            displayName: APIIntegrationResponse
            kind: APIIntegrationResponse
            name: apiintegrationresponses.apigateway.services.k8s.aws
            version: v1alpha1
          - description: APIKey represents the state of an AWS apigateway APIKey resource.
            displayName: APIKey
            kind: APIKey
            name: apikeys.apigateway.services.k8s.aws
            version: v1alpha1
          - description: APIMethodResponse represents the state of an AWS apigateway
              APIMethodResponse resource.
            displayName: APIMethodResponse
            kind: APIMethodResponse
            name: apimethodresponses.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Authorizer represents the state of an AWS apigateway Authorizer
              resource.
            displayName: Authorizer
            kind: Authorizer
            name: authorizers.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Deployment represents the state of an AWS apigateway Deployment
              resource.
            displayName: Deployment
            kind: Deployment
            name: deployments.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Integration represents the state of an AWS apigateway Integration
              resource.
            displayName: Integration
            kind: Integration
            name: integrations.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Method represents the state of an AWS apigateway Method resource.
            displayName: Method
            kind: Method
            name: methods.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Resource represents the state of an AWS apigateway Resource
              resource.
            displayName: Resource
            kind: Resource
            name: resources.apigateway.services.k8s.aws
            version: v1alpha1
          - description: RestAPI represents the state of an AWS apigateway RestAPI
              resource.
            displayName: RestAPI
            kind: RestAPI
            name: restapis.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Stage represents the state of an AWS apigateway Stage resource.
            displayName: Stage
            kind: Stage
            name: stages.apigateway.services.k8s.aws
            version: v1alpha1
          - description: VPCLink represents the state of an AWS apigateway VPCLink
              resource.
            displayName: VPCLink
            kind: VPCLink
            name: vpclinks.apigateway.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon API Gateway resources in AWS from within your Kubernetes cluster.

          **About Amazon API Gateway**

          Amazon API Gateway is an AWS service for creating, publishing, maintaining, monitoring, and securing REST, HTTP, and WebSocket APIs at any scale. API developers can create APIs that access AWS or other web services, as well as data stored in the AWS Cloud. As an API Gateway API developer, you can create APIs for use in your own client applications. Or you can make your APIs available to third-party app developers.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon API Gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apigateway
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon API Gateway Developer Resources
          url: https://aws.amazon.com/api-gateway/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: apigateway maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/apigateway-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-apigateway-controller.v1.3.1
        version: 1.3.1
      - name: ack-apigateway-controller.v1.3.0
        version: 1.3.0
      - name: ack-apigateway-controller.v1.2.5
        version: 1.2.5
      - name: ack-apigateway-controller.v1.2.4
        version: 1.2.4
      - name: ack-apigateway-controller.v1.2.3
        version: 1.2.3
      - name: ack-apigateway-controller.v1.2.2
        version: 1.2.2
      - name: ack-apigateway-controller.v1.2.1
        version: 1.2.1
      - name: ack-apigateway-controller.v1.1.0
        version: 1.1.0
      - name: ack-apigateway-controller.v1.0.6
        version: 1.0.6
      - name: ack-apigateway-controller.v1.0.5
        version: 1.0.5
      - name: ack-apigateway-controller.v1.0.4
        version: 1.0.4
      - name: ack-apigateway-controller.v1.0.2
        version: 1.0.2
      - name: ack-apigateway-controller.v1.0.1
        version: 1.0.1
      - name: ack-apigateway-controller.v1.0.0
        version: 1.0.0
      - name: ack-apigateway-controller.v0.0.3
        version: 0.0.3
      - name: ack-apigateway-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-apigateway-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: geunwookim
      provider-url: ""
    name: ncn-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ncn-operator.v0.3.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-1"
                },
                "spec": null
              },
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "label-1"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
          - description: Label is the Schema for the labels API
            displayName: Label
            kind: Label
            name: labels.cxl.antline.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: jwahn@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Antline
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/rasen708/ncn-operator:v0.4.0
        version: 0.3.18
      entries:
      - name: ncn-operator.v0.3.18
        version: 0.3.18
      - name: ncn-operator.v0.3.14
        version: 0.3.14
      name: alpha
    - currentCSV: ncn-operator.v0.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-1"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-19T01:21:08Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: jwahn@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        provider:
          name: Antline
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/rasen708/ncn-operator:v0.2.2
        version: 0.2.2
      entries:
      - name: ncn-operator.v0.2.2
        version: 0.2.2
      - name: ncn-operator.v0.2.1
        version: 0.2.1
      - name: ncn-operator.v0.2.0
        version: 0.2.0
      name: beta
    - currentCSV: ncn-operator.v0.9.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cmmd.gw.com/v1",
                "kind": "CMMD",
                "metadata": {
                  "name": "my-app",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "allocateMode": "auto",
                  "enable": true,
                  "payload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "metadata": {
                      "name": "my-app"
                    },
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "name": "my-app"
                        }
                      },
                      "template": {
                        "metadata": {
                          "labels": {
                            "name": "my-app"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "quay.io/my-account/my-app:latest",
                              "name": "my-app-01"
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "cmmd.gw.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "machineconfiguration.openshift.io/role": "cmmd"
                  },
                  "name": "cmmd-node-label",
                  "namespace": "socmmd"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          console.openshift.io/plugins: '["socmmd-console-plugin"]'
          containerImage: quay.io/rasen708/ncn-operator@sha256:a09963745bd9c1419532d42d91237b7496c7a78e6a1c460a8e714ea1a0ce4d45
          createdAt: "2024-08-01T05:45:38Z"
          description: Provides the ncn extend
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "cmmd.gw.com/v1",
              "kind": "Label",
              "metadata": {
                "name": "cmmd-node-label",
                "namespace": "socmmd",
                "labels": {}
              },
              "spec": {}
            }
          operatorframework.io/suggested-namespace: socmmd-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/geunwookim/ncn-operator
          support: geunwookim
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CMMD is the Schema for the cmmds API
            displayName: CMMD
            kind: CMMD
            name: cmmds.cmmd.gw.com
            version: v1
          - description: Label is the Schema for the labels API
            displayName: Node Label
            kind: Label
            name: labels.cmmd.gw.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: geunwookim@gmail.com
          name: geunwookim
        maturity: betav1
        provider:
          name: geunwookim
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/rasen708/ncn-operator@sha256:a09963745bd9c1419532d42d91237b7496c7a78e6a1c460a8e714ea1a0ce4d45
        version: 0.9.8
      entries:
      - name: ncn-operator.v0.9.8
        version: 0.9.8
      - name: ncn-operator.v0.9.2
        version: 0.9.2
      - name: ncn-operator.v0.8.1
        version: 0.8.1
      - name: ncn-operator.v0.5.6
        version: 0.5.6
      - name: ncn-operator.v0.4.3
        version: 0.4.3
      - name: ncn-operator.v0.4.2
        version: 0.4.2
      - name: ncn-operator.v0.3.18
        version: 0.3.18
      - name: ncn-operator.v0.3.14
        version: 0.3.14
      name: betav1
    defaultChannel: betav1
    packageName: ncn-operator
    provider:
      name: geunwookim
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CNCF
      provider-url: ""
    name: etcd
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: etcdoperator.v0.6.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          description: etcd is a distributed key value store providing a reliable
            way to store data across a cluster of machines.
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
        description: |
          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader. Your applications can read and write data into etcd.
          A simple use-case is to store database connection details or feature flags within etcd as key value pairs. These values can be watched, allowing your app to reconfigure itself when they change. Advanced uses take advantage of the consistency guarantees to implement database leader elections or do distributed locking across a cluster of workers.

          _The etcd Open Cloud Service is Public Alpha. The goal before Beta is to fully implement backup features._

          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` or with the API using the automatically generated Kubernetes Service.

          [Read the complete guide to using the etcd Open Cloud Service](https://coreos.com/tectonic/docs/latest/alm/etcd-ocs.html)

          ### Supported Features
          **High availability**
          Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.
          **Automated updates**
          Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.
          **Backups included**
          Coming soon, the ability to schedule backups to happen on or off cluster.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: support@coreos.com
          name: CoreOS, Inc
        maturity: alpha
        provider:
          name: CoreOS, Inc
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:bd944a211eaf8f31da5e6d69e8541e7cada8f16a9f7a5a570b22478997819943
        version: 0.6.1
      entries:
      - name: etcdoperator.v0.6.1
        version: 0.6.1
      name: alpha
    - currentCSV: etcdoperator.v0.9.4-clusterwide
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdCluster",
                "metadata": {
                  "name": "example",
                  "annotations": {
                    "etcd.database.coreos.com/scope": "clusterwide"
                  }
                },
                "spec": {
                  "size": 3,
                  "version": "3.2.13"
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdRestore",
                "metadata": {
                  "name": "example-etcd-cluster-restore"
                },
                "spec": {
                  "etcdCluster": {
                    "name": "example-etcd-cluster"
                  },
                  "backupStorageType": "S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdBackup",
                "metadata": {
                  "name": "example-etcd-cluster-backup"
                },
                "spec": {
                  "etcdEndpoints": ["<etcd-cluster-endpoints>"],
                  "storageType":"S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
          createdAt: "2019-02-28 01:03:00"
          description: Create and maintain highly-available etcd clusters on Kubernetes
          repository: https://github.com/coreos/etcd-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to backup an etcd cluster.
            displayName: etcd Backup
            kind: EtcdBackup
            name: etcdbackups.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to restore an etcd cluster from a backup.
            displayName: etcd Restore
            kind: EtcdRestore
            name: etcdrestores.etcd.database.coreos.com
            version: v1beta2
        description: |
          The etcd Operater creates and maintains highly-available etcd clusters on Kubernetes, allowing engineers to easily deploy and manage etcd clusters for their applications.

          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader.


          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` via port forwarding:

              $ kubectl --namespace default port-forward service/example-client 2379:2379
              $ etcdctl --endpoints http://127.0.0.1:2379 get /

          Or directly to the API using the automatically generated Kubernetes Service:

              $ etcdctl --endpoints http://example-client.default.svc:2379 get /

          Be sure to secure your etcd cluster (see Common Configurations) before exposing it outside of the namespace or cluster.


          ### Supported Features

          * **High availability** - Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.

          * **Automated updates** - Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.

          * **Backups included** - Create etcd backups and restore them through the etcd Operator.

          ### Common Configurations

          * **Configure TLS** - Specify [static TLS certs](https://github.com/coreos/etcd-operator/blob/master/doc/user/cluster_tls.md) as Kubernetes secrets.

          * **Set Node Selector and Affinity** - [Spread your etcd Pods](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-node-selector-and-anti-affinity-across-nodes) across Nodes and availability zones.

          * **Set Resource Limits** - [Set the Kubernetes limit and request](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-resource-requirement) values for your etcd Pods.

          * **Customize Storage** - [Set a custom StorageClass](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#custom-persistentvolumeclaim-definition) that you would like to use.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: etcd-dev@googlegroups.com
          name: etcd Community
        maturity: alpha
        provider:
          name: CNCF
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
        version: 0.9.4-clusterwide
      entries:
      - name: etcdoperator.v0.9.4-clusterwide
        version: 0.9.4-clusterwide
      - name: etcdoperator.v0.9.2-clusterwide
        version: 0.9.2-clusterwide
      - name: etcdoperator.v0.9.0
        version: 0.9.0
      name: clusterwide-alpha
    - currentCSV: etcdoperator.v0.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "size": 3,
                  "version": "3.2.13"
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdRestore",
                "metadata": {
                  "name": "example-etcd-cluster-restore"
                },
                "spec": {
                  "etcdCluster": {
                    "name": "example-etcd-cluster"
                  },
                  "backupStorageType": "S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdBackup",
                "metadata": {
                  "name": "example-etcd-cluster-backup"
                },
                "spec": {
                  "etcdEndpoints": ["<etcd-cluster-endpoints>"],
                  "storageType":"S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
          createdAt: "2019-02-28 01:03:00"
          description: Create and maintain highly-available etcd clusters on Kubernetes
          repository: https://github.com/coreos/etcd-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to backup an etcd cluster.
            displayName: etcd Backup
            kind: EtcdBackup
            name: etcdbackups.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to restore an etcd cluster from a backup.
            displayName: etcd Restore
            kind: EtcdRestore
            name: etcdrestores.etcd.database.coreos.com
            version: v1beta2
        description: |
          The etcd Operater creates and maintains highly-available etcd clusters on Kubernetes, allowing engineers to easily deploy and manage etcd clusters for their applications.

          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader.


          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` via port forwarding:

              $ kubectl --namespace default port-forward service/example-client 2379:2379
              $ etcdctl --endpoints http://127.0.0.1:2379 get /

          Or directly to the API using the automatically generated Kubernetes Service:

              $ etcdctl --endpoints http://example-client.default.svc:2379 get /

          Be sure to secure your etcd cluster (see Common Configurations) before exposing it outside of the namespace or cluster.


          ### Supported Features

          * **High availability** - Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.

          * **Automated updates** - Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.

          * **Backups included** - Create etcd backups and restore them through the etcd Operator.

          ### Common Configurations

          * **Configure TLS** - Specify [static TLS certs](https://github.com/coreos/etcd-operator/blob/master/doc/user/cluster_tls.md) as Kubernetes secrets.

          * **Set Node Selector and Affinity** - [Spread your etcd Pods](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-node-selector-and-anti-affinity-across-nodes) across Nodes and availability zones.

          * **Set Resource Limits** - [Set the Kubernetes limit and request](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-resource-requirement) values for your etcd Pods.

          * **Customize Storage** - [Set a custom StorageClass](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#custom-persistentvolumeclaim-definition) that you would like to use.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: etcd-dev@googlegroups.com
          name: etcd Community
        maturity: alpha
        provider:
          name: CNCF
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
        version: 0.9.4
      entries:
      - name: etcdoperator.v0.9.4
        version: 0.9.4
      - name: etcdoperator.v0.9.2
        version: 0.9.2
      - name: etcdoperator.v0.9.0
        version: 0.9.0
      name: singlenamespace-alpha
    defaultChannel: singlenamespace-alpha
    packageName: etcd
    provider:
      name: CNCF
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataStax
      provider-url: ""
    name: cass-operator-community
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cass-operator.v1.23.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cassandra.datastax.com/v1beta1",
                "kind": "CassandraDatacenter",
                "metadata": {
                  "name": "dc1"
                },
                "spec": {
                  "clusterName": "development",
                  "config": {
                    "cassandra-yaml": {
                      "authenticator": "PasswordAuthenticator",
                      "authorizer": "CassandraAuthorizer",
                      "num_tokens": 16,
                      "role_manager": "CassandraRoleManager"
                    },
                    "jvm-server-options": {
                      "initial_heap_size": "1G",
                      "max_heap_size": "1G"
                    }
                  },
                  "managementApiAuth": {
                    "insecure": {}
                  },
                  "podTemplateSpec": {
                    "spec": {
                      "containers": [
                        {
                          "name": "cassandra"
                        }
                      ],
                      "securityContext": {}
                    }
                  },
                  "racks": [
                    {
                      "name": "rack1"
                    },
                    {
                      "name": "rack2"
                    },
                    {
                      "name": "rack3"
                    }
                  ],
                  "resources": {
                    "requests": {
                      "cpu": "1000m",
                      "memory": "2Gi"
                    }
                  },
                  "serverType": "cassandra",
                  "serverVersion": "4.1.2",
                  "size": 3,
                  "storageConfig": {
                    "cassandraDataVolumeClaimSpec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "server-storage"
                    }
                  }
                }
              },
              {
                "apiVersion": "control.k8ssandra.io/v1alpha1",
                "kind": "CassandraTask",
                "metadata": {
                  "name": "example-cleanup"
                },
                "spec": {
                  "datacenter": {
                    "name": "dc2",
                    "namespace": "cass-operator"
                  },
                  "jobs": [
                    {
                      "command": "cleanup",
                      "name": "cleanup-run"
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database,Big Data,OpenShift Optional
          certified: "true"
          containerImage: cr.k8ssandra.io/k8ssandra/cass-operator:v1.23.2
          createdAt: "2025-03-20"
          description: |
            Simple provisioning, turn-key operations, and automated remediation of Apache Cassandra clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/k8ssandra/cass-operator
          support: DataStax
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CassandraDatacenter is the Schema for the cassandradatacenters
              API
            displayName: Cassandra Datacenter
            kind: CassandraDatacenter
            name: cassandradatacenters.cassandra.datastax.com
            version: v1beta1
          - description: CassandraTask is the Schema for the cassandrajobs API
            displayName: Cassandra Task
            kind: CassandraTask
            name: cassandratasks.control.k8ssandra.io
            version: v1alpha1
        description: "## Apache Cassandra\n\nThe Apache Cassandra database is the
          right choice when you need scalability and\nhigh availability without compromising
          performance. Linear scalability and\nproven fault-tolerance on commodity
          hardware or cloud infrastructure make it the\nperfect platform for mission-critical
          data. Cassandra's support for replicating\nacross multiple datacenters is
          best-in-class, providing lower latency for your\nusers and the peace of
          mind of knowing that you can survive regional outages.\n\n## DataStax Enterprise\n\nThe
          most advanced distribution of Apache Cassandra™ on the market, with the\nenterprise
          functionality needed for serious production systems and backed up and\nsupported
          by the best distributed-experts in the world. It's one platform for\nall
          types of applications anywhere, any cloud, any model: key-value, graph,\ntabular,
          JSON.\n\nDataStax Enterprise is a fully integrated and optimized database,
          with graph,\nanalytics, and search included, all with a unified security
          model. Simply put,\nit's the only database capable of meeting today's demanding
          requirements\n\n## Operator Details\n\n`cass-operator` is designed as a
          modular operator for Apache Cassandra and\nderived  distributions. Apache
          Cassandra is a distributed database consisting of\nmultiple nodes working
          in concert to store data and process queries along a\nnumber of fault domains.
          `cass-operator` has the deployment of a Cassandra\ncluster around the logical
          domain of a datacenter with the `CassandraDatacenter`\ncustom resource.\n
          \   \nUpon submission of one of these resources it handles provisioning
          the underlying\nstateful sets (analogous to C\\* logical racks), services,
          and configuration.\nAdditionally through monitoring pod state via Kubernetes
          callbacks it handles day to day\noperations such as restarting failed processes,
          scaling clusters up, and deploying\nconfiguration changes in a rolling,
          non-disruptive, fashion.\n    \nThis operator is designed to be `Namespace`
          scoped. A single Kubernetes cluster may\nbe running multiple instances of
          this operator, in separate namespaces, to support\na number of C\\* clusters
          and environments. Configuration is simple with the usage of\nYAML based
          overrides in the Custom Resource paired with an `init` container.\n    \nIn
          C\\* clusters ordering and timing of certain operations are important to
          keep the system\nevenly distributed. `cass-operator` takes advantage of
          a sidecar process within the\nmain container to handle the orchestration
          of starting our main C* process."
        displayName: DataStax Kubernetes Operator for Apache Cassandra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - k8ssandra
        - cass-operator
        - DataStax Enterprise
        - Apache Cassandra
        - DSE
        - C*
        - Cassandra
        links:
        - name: DataStax
          url: https://datastax.com
        - name: Operator GitHub Repo
          url: https://github.com/k8ssandra/cass-operator
        - name: Config Builder GitHub Repo
          url: https://github.com/datastax/cass-config-builder
        - name: Documentation [1]
          url: https://github.com/k8ssandra/cass-operator/tree/master/docs/user
        - name: Documentation [2]
          url: https://docs.datastax.com/en/cass-operator/doc/cass-operator/cassOperatorTOC.html
        maintainers:
        - email: cass-operator@datastax.com
          name: DataStax Operator Team
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: DataStax
        relatedImages:
        - cr.k8ssandra.io/k8ssandra/cass-operator:v1.23.2
        version: 1.23.2
      entries:
      - name: cass-operator.v1.23.2
        version: 1.23.2
      - name: cass-operator.v1.23.1
        version: 1.23.1
      - name: cass-operator.v1.23.0
        version: 1.23.0
      - name: cass-operator.v1.21.1
        version: 1.21.1
      - name: cass-operator.v1.19.1
        version: 1.19.1
      - name: cass-operator.v1.17.1
        version: 1.17.1
      - name: cass-operator.v1.16.0
        version: 1.16.0
      - name: cass-operator.v1.14.0
        version: 1.14.0
      - name: cass-operator.v1.13.0
        version: 1.13.0
      - name: cass-operator.v1.10.4
        version: 1.10.4
      - name: cass-operator.v1.10.1
        version: 1.10.1
      name: stable
    defaultChannel: stable
    packageName: cass-operator-community
    provider:
      name: DataStax
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: openstack-ansibleee-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openstack-ansibleee-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ansibleee.openstack.org/v1beta1",
                "kind": "OpenStackAnsibleEE",
                "metadata": {
                  "name": "openstackansibleee-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-09-23T20:57:34Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.3
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenStackAnsibleEE is the Schema for the openstackansibleees
              API
            displayName: OpenStack Ansible EE
            kind: OpenStackAnsibleEE
            name: openstackansibleees.ansibleee.openstack.org
            version: v1beta1
        description: The OpenStack Ansible Execution Environment Operator
        displayName: OpenStackAnsibleEE
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - execution
        links:
        - name: OpenStack Ansible EE Operator
          url: https://github.com/openstack-k8s-operators/openstack-ansibleee-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:24b468b09b43dd6cf77f5b08a68f2c6359d115722d28982fd8cdb1923e8122de
        - registry.redhat.io/rhoso-operators/openstack-ansibleee-rhel9-operator@sha256:bf096518547e92fb6fe0ecd526813bdc4a15f928a6d89eacadb71ffaad91db69
        - registry.redhat.io/rhoso-operators/ee-openstack-ansible-ee-rhel9@sha256:640efe595fbdb993ecc1c55c14457edc2b0e3e1c6dde8ad9ea1acf4b4503339f
        version: 1.0.3
      entries:
      - name: openstack-ansibleee-operator.v1.0.3
        version: 1.0.3
      - name: openstack-ansibleee-operator.v1.0.2
        version: 1.0.2
      - name: openstack-ansibleee-operator.v1.0.1
        version: 1.0.1
      - name: openstack-ansibleee-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: openstack-ansibleee-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: ""
    name: red-hat-hawtio-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: red-hat-hawtio-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              },
              {
                "apiVersion": "hawt.io/v2",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24,
                    "internalSSL": true
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
          createdAt: "2025-08-07T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v2
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
        version: 1.2.1
      entries:
      - name: red-hat-hawtio-operator.v1.2.1
        version: 1.2.1
      - name: red-hat-hawtio-operator.v1.2.0-0.1754874242.p
        version: 1.2.0+0.1754874242.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754338551.p
        version: 1.2.0+0.1754338551.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754317392.p
        version: 1.2.0+0.1754317392.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1753863723.p
        version: 1.2.0+0.1753863723.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752679472.p
        version: 1.2.0+0.1752679472.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752452299.p
        version: 1.2.0+0.1752452299.p
      - name: red-hat-hawtio-operator.v1.2.0
        version: 1.2.0
      - name: red-hat-hawtio-operator.v1.1.1
        version: 1.1.1
      - name: red-hat-hawtio-operator.v1.1.0
        version: 1.1.0
      - name: red-hat-hawtio-operator.v1.1.0-0.1742200932.p
        version: 1.1.0+0.1742200932.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1740358952.p
        version: 1.1.0+0.1740358952.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739994212.p
        version: 1.1.0+0.1739994212.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739992853.p
        version: 1.1.0+0.1739992853.p
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: latest
    - currentCSV: red-hat-hawtio-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              },
              {
                "apiVersion": "hawt.io/v2",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24,
                    "internalSSL": true
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
          createdAt: "2025-08-07T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v2
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
        version: 1.2.1
      entries:
      - name: red-hat-hawtio-operator.v1.2.1
        version: 1.2.1
      - name: red-hat-hawtio-operator.v1.2.0-0.1754874242.p
        version: 1.2.0+0.1754874242.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754338551.p
        version: 1.2.0+0.1754338551.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754317392.p
        version: 1.2.0+0.1754317392.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1753863723.p
        version: 1.2.0+0.1753863723.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752679472.p
        version: 1.2.0+0.1752679472.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752452299.p
        version: 1.2.0+0.1752452299.p
      - name: red-hat-hawtio-operator.v1.2.0
        version: 1.2.0
      - name: red-hat-hawtio-operator.v1.1.1
        version: 1.1.1
      - name: red-hat-hawtio-operator.v1.1.0
        version: 1.1.0
      - name: red-hat-hawtio-operator.v1.1.0-0.1742200932.p
        version: 1.1.0+0.1742200932.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1740358952.p
        version: 1.1.0+0.1740358952.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739994212.p
        version: 1.1.0+0.1739994212.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739992853.p
        version: 1.1.0+0.1739992853.p
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: stable-v1
    - currentCSV: red-hat-hawtio-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel8-operator@sha256:f04d81d24c757db7b678464382ad8c5182b1580e714123a9f4f5fc9fbd9b8c57
          createdAt: "2024-01-30T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-rhel8-operator@sha256:f04d81d24c757db7b678464382ad8c5182b1580e714123a9f4f5fc9fbd9b8c57
        version: 1.0.3
      entries:
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1
    packageName: red-hat-hawtio-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: namespace-configuration-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: namespace-configuration-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupConfig",
                "metadata": {
                  "name": "test-groupconfig"
                },
                "spec": {
                  "labelSelector": {
                    "matchLabels": {
                      "type": "team"
                    }
                  },
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-dev\n  labels:\n    group: {{ .Name }}  \n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-qa\n  labels:\n    group: {{ .Name }}          \n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-prod\n  labels:\n    group: {{ .Name }}                           \n"
                    },
                    {
                      "objectTemplate": "apiVersion: quota.openshift.io/v1\nkind: ClusterResourceQuota\nmetadata:\n  name: {{ .Name }}-quota\nspec:\n  quota:\n    hard:\n      pods: \"4\" \n      requests.cpu: \"1\" \n      requests.memory: 1Gi \n      requests.ephemeral-storage: 2Gi \n      limits.cpu: \"2\" \n      limits.memory: 2Gi \n      limits.ephemeral-storage: 4Gi\n  selector:\n    labels:\n      matchLabels:\n        group: {{ .Name }}\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NamespaceConfig",
                "metadata": {
                  "name": "test-namespaceconfig"
                },
                "spec": {
                  "labelSelector": {
                    "matchLabels": {
                      "type": "secure"
                    }
                  },
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: ResourceQuota\nmetadata:\n  name: standard-sandbox\n  namespace: {{ .Name }}\nspec:\n  hard:\n    pods: \"4\" \n    requests.cpu: \"1\" \n    requests.memory: 1Gi \n    requests.ephemeral-storage: 2Gi \n    limits.cpu: \"2\" \n    limits.memory: 2Gi \n    limits.ephemeral-storage: 4Gi\n"
                    },
                    {
                      "objectTemplate": "kind: EgressNetworkPolicy\napiVersion: network.openshift.io/v1\nmetadata:\n  name: air-gapped\n  namespace: {{ .Name }}  \nspec:\n  egress:\n  - type: Deny\n    to:\n      cidrSelector: 0.0.0.0/0\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "UserConfig",
                "metadata": {
                  "name": "test-user-config"
                },
                "spec": {
                  "providerName": "my-provider",
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-sandbox\n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: ResourceQuota\nmetadata:\n  name: standard-sandbox\n  namespace: {{ .Name }}-sandbox\nspec:\n  hard:\n    pods: \"4\" \n    requests.cpu: \"1\" \n    requests.memory: 1Gi \n    requests.ephemeral-storage: 2Gi \n    limits.cpu: \"2\" \n    limits.memory: 2Gi \n    limits.ephemeral-storage: 4Gi\n"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/namespace-configuration-operator@sha256:49ed7d6155342adaa2b12fd80c6761c3081d8e6149d187cb7ff91a247cdf2e7a
          createdAt: "2023-11-22T15:58:50Z"
          description: This operator provides a facility to define and enforce namespace
            configurations
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: namespace-configuration-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/namespace-configuration-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GroupConfig is the Schema for the groupconfigs API
            displayName: Group Config
            kind: GroupConfig
            name: groupconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: NamespaceConfig is the Schema for the namespaceconfigs API
            displayName: Namespace Config
            kind: NamespaceConfig
            name: namespaceconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: UserConfig is the Schema for the userconfigs API
            displayName: User Config
            kind: UserConfig
            name: userconfigs.redhatcop.redhat.io
            version: v1alpha1
        description: "The `namespace-configuration-operator` helps keeping configurations
          related to Users, Groups and Namespaces aligned with one of more policies
          specified as a CRs. The purpose is to provide the foundational building
          block to create an end-to-end onboarding process. \nBy onboarding process
          we mean all the provisioning steps needed to a developer team working on
          one or more applications to OpenShift.\n This usually involves configuring
          resources such as: Groups, RoleBindings, Namespaces, ResourceQuotas, NetworkPolicies,
          EgressNetworkPolicies, etc.... . Depending on the specific environment the
          list could continue.\nNaturally such a process should be as automatic and
          scalable as possible.\n\nWith the namespace-configuration-operator one can
          create rules that will react to the creation of Users, Groups and Namespace
          and will create and enforce a set of resources.\n\nHere are some examples
          of the type of onboarding processes that one could support:\n\n1. [developer
          sandbox](https://github.com/redhat-cop/namespace-configuration-operator/blob/master/examples/user-sandbox/readme.md)\n
          2. [team onboarding](https://github.com/redhat-cop/namespace-configuration-operator//blob/master/examples/team-onboarding/readme.md)
          with support of the entire SDLC in a multitentant environment.\n\nPolicies
          can be expressed with the following CRDs:\n\n| Watched Resource | CRD |\n|--|--|\n
          | Groups | [GroupConfig](#GroupConfig) |\n| Users | [UserConfig](#UserConfig)
          |\n| Namespace | [NamespaceConfig](#NamespaceConfig) |\n\nThese CRDs all
          share some commonalities:\n\n1. Templated Resources\n2. List of ignored
          jason path\n \n### Templated Resources\n\nEach has a parameter called `templatedResources`,
          which is an array. Each element of the array has two fields `objectTemplate`
          and `excludedPaths` (see below).\n\nThe `objectTemplate` field must contains
          a [go template](https://golang.org/pkg/text/template/) that resolves to
          a single API Resource expressed in `yaml`. The template is merged with the
          object selected by the CR. For example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: UserConfig\nmetadata:\n  name: test-user-config\nspec:\n  provider:
          corp-ldap\n  templates:\n  - objectTemplate: |\n      apiVersion: v1\n      kind:
          Namespace\n      metadata:\n        name: {{ .Name }}-sandbox\n```\n\nThis
          create a rule in which every time a user from the `corp-ldap` provider is
          created, a namespace called `<username>-sandbox` is also created.\n\n###
          Excluded Paths\n\nThe logic of the `namespace-configuration-operator` is
          to enforce that the resources resolved by processing the templates \"stays
          in place\". In other words if those resources are changed and/or deleted
          they will be reset by the operator.\nBut there are situations in which at
          least part of a resource is allowed to change. Common use cases are: annotations
          and in general the metadata section of a resource can be updated by the
          various operators watching that resources. The status field is often updated
          by the main operator managing that resources. Finally, when applicable the
          `spec.replicas` field should also be allowed to change.\n\nTo handle special
          use case, one can also specify additional *jsonpaths* that should be ignored
          when comparing the desired resource and the current resource and making
          a decision on whether that resource should be reset.\n\nThe following paths
          are always included:\n \n1. `.metadata`\n2. `.status`\n3. `.spec.replicas`\n\n##
          NamespaceConfig\n\n The `NamespaceConfig` CR allows specifying one or more
          objects that will be created in the selected namespaces.\n\nNamespaces can
          be selected by labels or annotations via a label selector for example:\n\n```yaml\napiVersion:
          redhatcop.redhat.io/v1alpha1\n kind: NamespaceConfig\nmetadata:\n  name:
          small-namespace\n  selector:\n    matchLabels:\n      size: small  \n  resources:\n
          \ - apiVersion: v1\n    kind: ResourceQuota\n    metadata:\n      name:
          small-size  \n    spec:\n      hard:\n        requests.cpu: \"4\"\n        requests.memory:
          \"2Gi\"\n```\n\nHere is a `NamespaceConfig` object using a `matchExpressions`
          selector:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n kind: NamespaceConfig\nmetadata:\n
          \ name: tier-config\nspec:\n  annotationSelector:\n    matchExpressions:\n
          \     - {key: tier, operator: In, values: [gold,silver]}\n ```\n\nAlthough
          not enforced by the operator the general expectation is that the NamespaceConfig
          CR will be used to create objects inside the selected namespace.\n \nExamples
          of NamespaceConfig usages can be found [here](https://github.com/redhat-cop/namespace-configuration-operator/blob/master/examples/namespace-config/readme.md)\n
          \n## GroupConfig\n\nThe `GroupConfig` CR allows specifying one or more objects
          that will be created in the selected Group.\nGroups can be selected by labels
          or annotations via a label selector, similarly to the `NamespaceConfig`.\n\nOften
          groups are created in OpenShift by a job that synchronizes an Identity Provider
          with OCP. So the idea is that when new groups are added or deleted the configuration
          in OpenShift will adapt automatically.\n\nAlthough not enforced by the operator,
          GroupConfig are expected to create cluster-scoped resources like Namespaces,
          ClusterResourceQuotas and potentially some namespaced resources like RoleBindings.\n\n##
          UserConfig\n \nIn OpenShift an external user is defined by two entities:
          Users and Identities. There is a relationship of on to many between Users
          and Identities. Given one user, there can be one Identity per authentication
          mechanism.\n\nThe `UserConfig` CR allows specifying one or more objects
          that will be created in the selected User.\nUsers can be selected by label
          or annotation like `NamespaceConfig` and `UserConfig`.\nUSers can also be
          selected by provider name (the name of the authentication mechanism) and
          identity extra field.\n\nHere is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          UserConfig\nmetadata:\n  name: test-user-config\nspec:\n  providerName:
          okta-provider\n  identityExtraFieldSelector:\n    matchLabels:\n      sandbox_enabled:
          \"true\"\n  templates:\n  - objectTemplate: |\n      apiVersion: v1\n      kind:
          Namespace\n      metadata:\n        name: {{ .Name }}-sandbox\n ```\n\nUser
          will be selected by this `UserConfig` only if they login via the *okta-provider*
          and if the extra field was populate with the label `sandbox_enabled: \"true\"
          `. Note that not all authentication provider allow populating the extra
          fields in the Identity object.\n"
        displayName: Namespace Configuration Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - namespace
        - user
        - group
        - configuration
        - policy
        - management
        links:
        - name: Namespace Configuration Operator
          url: https://github.com/redhat-cop/namespace-configuration-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/namespace-configuration-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/namespace-configuration-operator@sha256:49ed7d6155342adaa2b12fd80c6761c3081d8e6149d187cb7ff91a247cdf2e7a
        version: 1.2.6
      entries:
      - name: namespace-configuration-operator.v1.2.6
        version: 1.2.6
      - name: namespace-configuration-operator.v1.2.5
        version: 1.2.5
      - name: namespace-configuration-operator.v1.2.4
        version: 1.2.4
      - name: namespace-configuration-operator.v1.2.3
        version: 1.2.3
      - name: namespace-configuration-operator.v1.2.2
        version: 1.2.2
      - name: namespace-configuration-operator.v1.2.1
        version: 1.2.1
      - name: namespace-configuration-operator.v1.2.0
        version: 1.2.0
      - name: namespace-configuration-operator.v1.1.0
        version: 1.1.0
      - name: namespace-configuration-operator.v1.0.7
        version: 1.0.7
      - name: namespace-configuration-operator.v1.0.6
        version: 1.0.6
      - name: namespace-configuration-operator.v1.0.5
        version: 1.0.5
      - name: namespace-configuration-operator.v1.0.4
        version: 1.0.4
      - name: namespace-configuration-operator.v1.0.3
        version: 1.0.3
      - name: namespace-configuration-operator.v1.0.2
        version: 1.0.2
      - name: namespace-configuration-operator.v1.0.1
        version: 1.0.1
      - name: namespace-configuration-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: namespace-configuration-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator-certified.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a",
                  "imagePullPolicy": "Always",
                  "initImage": "registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:3c03fb18a5d73876576b80aca2b61e7c84ca052839ccb467d14e15f6d1f9d834
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a57926b53eef862b70a53818a7ccb0e9756f878f30d4128b608090b109bdc585
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator-certified.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator-certified.v1.20.0
        version: 1.20.0
      - name: percona-server-mongodb-operator-certified.v1.19.1
        version: 1.19.1
      - name: percona-server-mongodb-operator-certified.v1.19.0
        version: 1.19.0
      - name: percona-server-mongodb-operator-certified.v1.18.0
        version: 1.18.0
      - name: percona-server-mongodb-operator-certified.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator-certified
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Cloudera
      provider-url: https://cloudera.com
    name: cloudera-streams-messaging-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster"
                },
                "spec": {
                  "kafka": {
                    "version": "3.9.0.1.4",
                    "replicas": 3,
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2,
                      "inter.broker.protocol.version": "3.9"
                    },
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "zookeeper": {
                    "replicas": 3,
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "3.9.0.1.4",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
          createdAt: "2025-07-14 10:09:41"
          description: Run Apache Kafka with Cloudera Streams Messaging - Kubernetes
            Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: A valid license and current subscription
            for Cloudera Streaming - Kubernetes Operator is required. Contact Cloudera
            sales to obtain a license. https://www.cloudera.com/contact-sales.html
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |-
          Cloudera Streams Messaging - Kubernetes Operator brings enterprise-grade Kafka deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of Kafka for enterprise deployments. To learn more about Cloudera Streaming, [read this deep dive blog post](https://blog.cloudera.com/accelerating-deployments-of-streaming-pipelines-announcing-data-in-motion-on-kubernetes/).

          Cloudera Streams Messaging - Kubernetes Operator consists of and ships multiple components including Apache Kafka, Apache ZooKeeper, Cruise Control, Strimzi, and others.

          ## Installation prerequisites

          Installing Cloudera Streams Messaging - Kubernetes Operator requires a valid license as well as Cloudera credentials (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly. The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) where installation artifacts are pulled from.

          ### Obtaining a license

          Contact [Cloudera Sales](https://www.cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets

          Before installation, create  two `Secrets`. One containing your Cloudera credentials, and one containing your license. Cloudera credentials are provided to you together with your license.

          Create a `Secret` containing your Cloudera credentials.
          ```
          oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***] --docker-server=container.repository.cloudera.com
          ```
          Create a `Secret` containing your license.
          ```
          oc create secret generic csm-op-license --from-file=license.txt=[***PATH TO LICENSE FILE***]
          ```

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces where you plan to deploy Kafka or Kafka Connect clusters. Cloudera recommends that you create the `Secret` in all required namespaces before installation if you know what namespaces you will be using to deploy Kafka or Kafka Connect.

          ## Learn more
          To learn more, visit the [Cloudera Streams Messaging - Kubernetes Operator documentation library](https://docs.cloudera.com/csm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Streams Messaging - Kubernetes Operator, visit the following resources:

          * [Strimzi](https://strimzi.io/)
          * [Apache Kafka](https://kafka.apache.org/)
          * [Apache ZooKeeper](https://zookeeper.apache.org/)
          * [Cruise Control](https://github.com/linkedin/cruise-control)
        displayName: Cloudera Streams Messaging - Kubernetes Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Documentation
          url: https://docs.cloudera.com/csm-operator
        - name: Cloudera Community Forum
          url: https://community.cloudera.com/
        - name: Cloudera Stream Processing
          url: https://www.cloudera.com/products/stream-processing.html
        - name: Contact Cloudera
          url: https://www.cloudera.com/contact-sales.html
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
        - container.repository.cloudera.com/cloudera/kafka@sha256:aa64e161b9672153a6f6559f9ae9ef88d0e40d6040a80e21bd3dc8ee39df6523
        - container.repository.cloudera.com/cloudera/kafka@sha256:b675e7426a558c4681e53d8e03f278264d7ad555f7c02422017683a29708ac2b
        - container.repository.cloudera.com/cloudera/maven-builder@sha256:9f18203b175193e5079fdc0c2b71c14887641d997f92b4a1c918546f777166df
        version: 1.4.0-b199
      entries:
      - name: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
        version: 1.4.0-b199
      name: 1.2.x
    - currentCSV: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster"
                },
                "spec": {
                  "kafka": {
                    "version": "3.9.0.1.4",
                    "replicas": 3,
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2,
                      "inter.broker.protocol.version": "3.9"
                    },
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "zookeeper": {
                    "replicas": 3,
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "3.9.0.1.4",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
          createdAt: "2025-07-14 10:09:41"
          description: Run Apache Kafka with Cloudera Streams Messaging - Kubernetes
            Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: A valid license and current subscription
            for Cloudera Streaming - Kubernetes Operator is required. Contact Cloudera
            sales to obtain a license. https://www.cloudera.com/contact-sales.html
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |-
          Cloudera Streams Messaging - Kubernetes Operator brings enterprise-grade Kafka deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of Kafka for enterprise deployments. To learn more about Cloudera Streaming, [read this deep dive blog post](https://blog.cloudera.com/accelerating-deployments-of-streaming-pipelines-announcing-data-in-motion-on-kubernetes/).

          Cloudera Streams Messaging - Kubernetes Operator consists of and ships multiple components including Apache Kafka, Apache ZooKeeper, Cruise Control, Strimzi, and others.

          ## Installation prerequisites

          Installing Cloudera Streams Messaging - Kubernetes Operator requires a valid license as well as Cloudera credentials (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly. The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) where installation artifacts are pulled from.

          ### Obtaining a license

          Contact [Cloudera Sales](https://www.cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets

          Before installation, create  two `Secrets`. One containing your Cloudera credentials, and one containing your license. Cloudera credentials are provided to you together with your license.

          Create a `Secret` containing your Cloudera credentials.
          ```
          oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***] --docker-server=container.repository.cloudera.com
          ```
          Create a `Secret` containing your license.
          ```
          oc create secret generic csm-op-license --from-file=license.txt=[***PATH TO LICENSE FILE***]
          ```

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces where you plan to deploy Kafka or Kafka Connect clusters. Cloudera recommends that you create the `Secret` in all required namespaces before installation if you know what namespaces you will be using to deploy Kafka or Kafka Connect.

          ## Learn more
          To learn more, visit the [Cloudera Streams Messaging - Kubernetes Operator documentation library](https://docs.cloudera.com/csm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Streams Messaging - Kubernetes Operator, visit the following resources:

          * [Strimzi](https://strimzi.io/)
          * [Apache Kafka](https://kafka.apache.org/)
          * [Apache ZooKeeper](https://zookeeper.apache.org/)
          * [Cruise Control](https://github.com/linkedin/cruise-control)
        displayName: Cloudera Streams Messaging - Kubernetes Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Documentation
          url: https://docs.cloudera.com/csm-operator
        - name: Cloudera Community Forum
          url: https://community.cloudera.com/
        - name: Cloudera Stream Processing
          url: https://www.cloudera.com/products/stream-processing.html
        - name: Contact Cloudera
          url: https://www.cloudera.com/contact-sales.html
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/kafka@sha256:aa64e161b9672153a6f6559f9ae9ef88d0e40d6040a80e21bd3dc8ee39df6523
        - container.repository.cloudera.com/cloudera/kafka@sha256:b675e7426a558c4681e53d8e03f278264d7ad555f7c02422017683a29708ac2b
        - container.repository.cloudera.com/cloudera/maven-builder@sha256:9f18203b175193e5079fdc0c2b71c14887641d997f92b4a1c918546f777166df
        - container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
        version: 1.4.0-b199
      entries:
      - name: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
        version: 1.4.0-b199
      name: stable
    defaultChannel: stable
    packageName: cloudera-streams-messaging-kubernetes-operator
    provider:
      name: Cloudera
      url: https://cloudera.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: vertical-pod-autoscaler
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: verticalpodautoscaler.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "autoscaling.openshift.io/v1",
                "kind": "VerticalPodAutoscalerController",
                "metadata": {
                  "name": "default",
                  "namespace": "openshift-vertical-pod-autoscaler"
                },
                "spec": {
                  "podMinCPUMillicores": 25,
                  "podMinMemoryMb": 250,
                  "recommendationOnly": false,
                  "safetyMarginFraction": 0.15,
                  "minReplicas": 2
                }
              },
              {
                 "apiVersion": "autoscaling.k8s.io/v1",
                 "kind": "VerticalPodAutoscaler",
                 "metadata": {
                    "name": "myapp-vpa"
                 },
                 "spec": {
                    "targetRef": {
                       "apiVersion": "apps/v1",
                       "kind": "Deployment",
                       "name": "myapp-deployment"
                    },
                    "resourcePolicy": {
                       "containerPolicies": [
                          {
                             "containerName": "*",
                             "minAllowed": {
                                "cpu": "100m",
                                "memory": "50Mi"
                             },
                             "maxAllowed": {
                                "cpu": 1,
                                "memory": "500Mi"
                             },
                             "controlledResources": [
                                "cpu",
                                "memory"
                             ]
                          }
                       ]
                    }
                 }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          certifiedLevel: Primed
          containerImage: registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9-operator@sha256:edf76334695c9a4aee4a9c0297af40475a16dec7e35ec7113a35b8507724536e
          createdAt: 2019-10-14 12:17:13+00:00
          description: An operator to run the OpenShift Vertical Pod Autoscaler. Vertical
            Pod Autoscaler (VPA) can be configured to monitor a workload's resource
            utilization, and then adjust its CPU and memory limits by updating the
            pod (future) or restarting the pod with updated limits.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          healthIndex: B
          olm.skipRange: '>=4.5.0 <4.16.0-202509240123'
          operatorframework.io/suggested-namespace: openshift-vertical-pod-autoscaler
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/vertical-pod-autoscaler-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of the set of VPA controllers
            displayName: VPA Controller
            kind: VerticalPodAutoscalerController
            name: verticalpodautoscalercontrollers.autoscaling.openshift.io
            version: v1
          - description: An application whose resources are managed by the VPA
            displayName: VerticalPodAutoscaler
            kind: VerticalPodAutoscaler
            name: verticalpodautoscalers.autoscaling.k8s.io
            version: v1
          - description: Represents VPA checkpoints used by the VPA controllers to
              manage workloads
            displayName: VPA Checkpoint
            kind: VerticalPodAutoscalerCheckpoint
            name: verticalpodautoscalercheckpoints.autoscaling.k8s.io
            version: v1
        description: An operator to run the OpenShift Vertical Pod Autoscaler. Vertical
          Pod Autoscaler (VPA) can be configured to monitor a workload's resource
          utilization, and then adjust its CPU and memory limits by updating the pod
          (future) or restarting the pod with updated limits.
        displayName: VerticalPodAutoscaler
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Vertical Pod Autoscaler Documentation
          url: https://docs.openshift.com/container-platform/latest/nodes/pods/nodes-pods-vertical-autoscaler.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9-operator@sha256:edf76334695c9a4aee4a9c0297af40475a16dec7e35ec7113a35b8507724536e
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9@sha256:e6c8f34895184179bb87f1659c4ff30b96b131787fa7a64e7bc6b6adbb336c16
        version: 4.16.0-202509240123
      entries:
      - name: verticalpodautoscaler.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: verticalpodautoscaler.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: verticalpodautoscaler.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: verticalpodautoscaler.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: verticalpodautoscaler.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: verticalpodautoscaler.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: verticalpodautoscaler.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: verticalpodautoscaler.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: verticalpodautoscaler.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: verticalpodautoscaler.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: verticalpodautoscaler.v4.16.0-202501260334
        version: 4.16.0-202501260334
      - name: verticalpodautoscaler.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: verticalpodautoscaler.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: verticalpodautoscaler.v4.16.0-202412021735
        version: 4.16.0-202412021735
      - name: verticalpodautoscaler.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: verticalpodautoscaler.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: verticalpodautoscaler.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: verticalpodautoscaler.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: verticalpodautoscaler.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: verticalpodautoscaler.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: verticalpodautoscaler.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: verticalpodautoscaler.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: verticalpodautoscaler.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: verticalpodautoscaler.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: verticalpodautoscaler.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: verticalpodautoscaler.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: verticalpodautoscaler.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: verticalpodautoscaler.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: verticalpodautoscaler.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: verticalpodautoscaler.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: vertical-pod-autoscaler
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-businessautomation: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: businessautomation-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: businessautomation-operator.7.13.5-16
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.kiegroup.org/v2","kind":"KieApp","metadata":{"name":"rhpam-trial"},"spec":{"environment":"rhpam-trial"}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/rhpam-7/rhpam-rhel8-operator@sha256:3b4f420def9e5d2c768d5209264bd5f9af6d6a37b22ef9a988a167ff7d9959cf
          createdAt: "2025-07-30 19:54:27"
          description: Deploys and manages Red Hat Process Automation Manager and
            Red Hat Decision Manager environments.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Process Automation
            Manager", "Red Hat Decision Manager"]'
          repository: https://github.com/kiegroup/kie-cloud-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A project prescription running an RHPAM/RHDM environment.
            displayName: KieApp
            kind: KieApp
            name: kieapps.app.kiegroup.org
            version: v2
        description: |-
          Deploys and manages Red Hat Process Automation Manager and Red Hat Decision Manager environments.

          * **Red Hat Process Automation Manager** is a platform for developing containerized microservices and applications that automate business decisions and processes. It includes business process management (BPM), business rules management (BRM), and business resource optimization and complex event processing (CEP) technologies. It also includes a user experience platform to create engaging user interfaces for process and decision services with minimal coding.

          [See more](https://www.redhat.com/en/products/process-automation).
        displayName: Business Automation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kieapp
        - pam
        - decision
        - kie
        - cloud
        - bpm
        - process
        - automation
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-process-automation-manager
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_process_automation_manager/7.13/#category-deploying-red-hat-process-automation-manager-on-openshift
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-monitoring-rhel8@sha256:8d4481a11269565f1690ad007f152e24326bae4b2099cb26ddf7b21be67031e8
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:cb37360843e9b062c59755593d9d412e1bf2c1dc1d6902e8a65a2825812c0989
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhpam-7/rhpam-kieserver-rhel8@sha256:2526d12da63b67ae9ed9294c914e9406d833c090ffa777623303ea0be22755d8
        - registry.redhat.io/rhpam-7/rhpam-process-migration-rhel8@sha256:3d7bdbe38895e50549e97573bd40da47b11691932164db7dafff0155a051aab2
        - registry.redhat.io/rhpam-7/rhpam-controller-rhel8@sha256:3a901b54a9d35d38b5aca275df2160e3190a874e139fcb606b9969b38284b3dc
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        - registry.redhat.io/rhpam-7/rhpam-kieserver-rhel8@sha256:3fbef82d1c5da8da1606d1a2d5dcfd20000da574737eb32b772d4f0123b179cb
        - registry.redhat.io/rhpam-7/rhpam-smartrouter-rhel8@sha256:d1e5461db8983a13f98259152dc3cfa2aa677c29f799104190e5cefa83cf97b5
        - registry.redhat.io/rhpam-7/rhpam-dashbuilder-rhel8@sha256:46396dc8060f2f9cc6f87ad905ef55be5b8e5216e5162bdf54df98c9d70e29ef
        - registry.redhat.io/rhscl/mysql-80-rhel7@sha256:e4f535c78bc2fdb58914c52fb805185035ff31515452b4df714f9ab623ba383a
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-rhel8@sha256:228c6ab5b83e215a98f6eb76250feb2ae3096fd21398c44e5b2701257c9c3493
        - registry.redhat.io/rhpam-7/rhpam-smartrouter-rhel8@sha256:9d7efdd97f163770d8a84b1ce5236bd4f31508154dced1f11dc6efe75a7b5089
        - registry.redhat.io/rhscl/postgresql-13-rhel7@sha256:338dd09681dcf860e6a976a3bc48a1474c2376d698815e4eb611ae9c5aa68966
        - registry.redhat.io/rhpam-7/rhpam-rhel8-operator@sha256:3b4f420def9e5d2c768d5209264bd5f9af6d6a37b22ef9a988a167ff7d9959cf
        - registry.redhat.io/rhpam-7/rhpam-dashbuilder-rhel8@sha256:4edac282119898ab1395030eecaee584990064e88fc8e65fb999f19ea2ef3f8b
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-rhel8@sha256:1abbaba9d4b2ecaec38205bcccf5ef4c7b6388ed9725b4f4c6dc8dc664588eb1
        - registry.redhat.io/rhpam-7/rhpam-controller-rhel8@sha256:989dc6dc64a2d5b6723fb5c000b0411fbdc6f78c0056666d831ffeabeab4b612
        - registry.redhat.io/openshift4/ose-cli@sha256:c3eb8807caf73e30012ae53c6d4f5114471290a5596929c27acef5a699a5b813
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/rhpam-7/rhpam-process-migration-rhel8@sha256:bd5c6586b4859dec23ad9530ee98bab33497ab251d544f863fc0af01cbd45101
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:457fc73423fa94591bc7e35729a25c58d3465583afccf342fb338b6e11188575
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-monitoring-rhel8@sha256:32912270eba0f2464da88f315bde8bf1ef5bbded814db156811d7a415a458f86
        version: 7.13.5-16
      entries:
      - name: businessautomation-operator.7.13.5-16
        version: 7.13.5-16
      - name: businessautomation-operator.7.13.5-15
        version: 7.13.5-15
      - name: businessautomation-operator.7.13.5-14
        version: 7.13.5-14
      - name: businessautomation-operator.7.13.5-13
        version: 7.13.5-13
      - name: businessautomation-operator.7.13.5-12
        version: 7.13.5-12
      - name: businessautomation-operator.7.13.5-11
        version: 7.13.5-11
      - name: businessautomation-operator.7.13.5-10
        version: 7.13.5-10
      - name: businessautomation-operator.7.13.5-9
        version: 7.13.5-9
      - name: businessautomation-operator.7.13.5-8
        version: 7.13.5-8
      - name: businessautomation-operator.7.13.5-7
        version: 7.13.5-7
      - name: businessautomation-operator.7.13.5-6
        version: 7.13.5-6
      - name: businessautomation-operator.7.13.5-5
        version: 7.13.5-5
      - name: businessautomation-operator.7.13.5-4
        version: 7.13.5-4
      - name: businessautomation-operator.7.13.5-3
        version: 7.13.5-3
      - name: businessautomation-operator.7.13.5-2
        version: 7.13.5-2
      - name: businessautomation-operator.7.13.5-1
        version: 7.13.5-1
      - name: businessautomation-operator.7.13.4-4
        version: 7.13.4-4
      - name: businessautomation-operator.7.13.4-3
        version: 7.13.4-3
      - name: businessautomation-operator.7.13.4-2
        version: 7.13.4-2
      - name: businessautomation-operator.7.13.4-1
        version: 7.13.4-1
      - name: businessautomation-operator.7.13.3-5
        version: 7.13.3-5
      - name: businessautomation-operator.7.13.3-4
        version: 7.13.3-4
      - name: businessautomation-operator.7.13.3-3
        version: 7.13.3-3
      - name: businessautomation-operator.7.13.3-2
        version: 7.13.3-2
      - name: businessautomation-operator.7.13.3-1
        version: 7.13.3-1
      - name: businessautomation-operator.7.13.2-3
        version: 7.13.2-3
      - name: businessautomation-operator.7.13.2-2
        version: 7.13.2-2
      - name: businessautomation-operator.7.13.2-1
        version: 7.13.2-1
      - name: businessautomation-operator.7.13.1-2
        version: 7.13.1-2
      - name: businessautomation-operator.7.13.1-1
        version: 7.13.1-1
      - name: businessautomation-operator.7.13.0-1
        version: 7.13.0-1
      - name: businessautomation-operator.7.12.1-4
        version: 7.12.1-4
      - name: businessautomation-operator.7.12.1-3
        version: 7.12.1-3
      - name: businessautomation-operator.7.12.1-2
        version: 7.12.1-2
      - name: businessautomation-operator.7.12.1-1
        version: 7.12.1-1
      - name: businessautomation-operator.7.12.0-2
        version: 7.12.0-2
      - name: businessautomation-operator.7.12.0-1
        version: 7.12.0-1
      - name: businessautomation-operator.7.11.1-3
        version: 7.11.1-3
      - name: businessautomation-operator.7.11.1-2
        version: 7.11.1-2
      - name: businessautomation-operator.7.11.1-1
        version: 7.11.1-1
      - name: businessautomation-operator.7.11.0-2
        version: 7.11.0-2
      - name: businessautomation-operator.7.11.0-1
        version: 7.11.0-1
      - name: businessautomation-operator.7.10.1-2
        version: 7.10.1-2
      - name: businessautomation-operator.7.10.1-1
        version: 7.10.1-1
      - name: businessautomation-operator.7.10.0-2
        version: 7.10.0-2
      - name: businessautomation-operator.7.10.0-1
        version: 7.10.0-1
      - name: businessautomation-operator.7.9.1-1
        version: 7.9.1-1
      - name: businessautomation-operator.7.9.0
        version: 7.9.0
      - name: businessautomation-operator.7.9.0-2
        version: 7.9.0-2
      - name: businessautomation-operator.7.8.1
        version: 7.8.1
      - name: businessautomation-operator.7.8.0
        version: 7.8.0
      - name: businessautomation-operator.1.4.1
        version: 1.4.1
      - name: businessautomation-operator.1.4.0
        version: 1.4.0
      - name: businessautomation-operator.1.3.0
        version: 1.3.0
      - name: businessautomation-operator.1.2.1
        version: 1.2.1
      - name: businessautomation-operator.1.2.0
        version: 1.2.0
      name: stable
    defaultChannel: stable
    packageName: businessautomation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: quay-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:d73321dfa6dcced943405fae45c2ccf841663e0ab57e8be8cc1ddba251b7ce08
          createdAt: 2021-11-02 11:11 UTC
          description: Opinionated deployment of Red Hat Quay on Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:d73321dfa6dcced943405fae45c2ccf841663e0ab57e8be8cc1ddba251b7ce08
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:0e6bef51caf479074925115fa11621c41bcada375cc1925d0511a73084b93543
        - registry.redhat.io/rhel8/postgresql-10@sha256:b2695e9af379624621bef9ec785f0cb618bfdabae98c5f17eea82ff5f8d2d5cb
        - registry.redhat.io/rhel8/redis-5@sha256:8d7e66927ce4b3050ca15ab4df4dbaeb593cb855e16fb0a6f5dde950e9db7db9
        - registry.redhat.io/quay/quay-rhel8@sha256:ed8c4249d376b4226e35bf088ffe5bf688bf68b059718c5e218aa1184fe048d0
        - registry.redhat.io/quay/clair-rhel8@sha256:af5e2fd1543365b365b1a6cce00b86e44fe5e66cde6491e9b113324c21f629c8
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:2a471300f70d3995ab7e9fcc3df35c5513e73fcc1d140fb4d651a1c3eae949eb
        version: 3.4.7
      entries:
      - name: quay-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: quay-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:1f5e4cc9ec2d9b8b89d82f1a698acfa1ac2764ca4ebc5c85b3417da7730b5867
          createdAt: 2021-09-09 20:09 UTC
          description: Opinionated deployment of Red Hat Quay on Kubernetes.
          olm.skipRange: '>=3.4.x <3.5.7'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.5.7
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7fe408a1e428b2d0660cb578c1dcd4b961f838e37166fe7b21b2bed70d288594
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:5d72577ccfd63c6d9ad712b376b604c40d2d3ecb53e391e2ee0d18caa6e0225a
        - registry.redhat.io/rhel8/postgresql-10@sha256:a723bb3a7a5046527b9a5b628ddfb211c526e3814cc4242262c4f8a016f38edd
        - registry.redhat.io/rhel8/redis-5@sha256:7916760ff03af5ca230df5e96b2bf353fbf1829386de1c063c05f623837be7f7
        - registry.redhat.io/quay/quay-rhel8@sha256:c599892160ac20c744d833603d9375923af7aec1edfd86982b7513e02fb6d463
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:1f5e4cc9ec2d9b8b89d82f1a698acfa1ac2764ca4ebc5c85b3417da7730b5867
        - registry.redhat.io/quay/clair-rhel8@sha256:d242bd9d37f95b2a5a8267b3f1eba02c671d46121712baae43466c86d1da89f9
        version: 3.5.7
      entries:
      - name: quay-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: quay-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:3c0201e3cf88e711c921870324002a02cbd42c0d2eecafb4f242aa0601e24602
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.10.15
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:fc20b8bdac88c1bfce9aec993bdce6dd2cc132f8a9d4062d174b098819c74013
        - registry.redhat.io/quay/clair-rhel8@sha256:efa6e69500ea643042b544e6205f088214d79ab09a36864a612e73d07979d1c5
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:3c0201e3cf88e711c921870324002a02cbd42c0d2eecafb4f242aa0601e24602
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7e720c6a332ec4032e8576c31421af404f61d9baaf8e88f69f8333c52e5d1683
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:774cd1f181527787ba8631c7c14437eb2250c15ddb4a9c75fdfa35c5c1a19a2e
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        version: 3.10.15
      entries:
      - name: quay-operator.v3.10.15
        version: 3.10.15
      - name: quay-operator.v3.10.14
        version: 3.10.14
      - name: quay-operator.v3.10.13
        version: 3.10.13
      - name: quay-operator.v3.10.12
        version: 3.10.12
      - name: quay-operator.v3.10.11
        version: 3.10.11
      - name: quay-operator.v3.10.10
        version: 3.10.10
      - name: quay-operator.v3.10.9
        version: 3.10.9
      - name: quay-operator.v3.10.8
        version: 3.10.8
      - name: quay-operator.v3.10.7
        version: 3.10.7
      - name: quay-operator.v3.10.6
        version: 3.10.6
      - name: quay-operator.v3.10.5
        version: 3.10.5
      - name: quay-operator.v3.10.4
        version: 3.10.4
      - name: quay-operator.v3.10.3
        version: 3.10.3
      - name: quay-operator.v3.10.2
        version: 3.10.2
      - name: quay-operator.v3.10.1
        version: 3.10.1
      - name: quay-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:cb3a41caa3a572b7981d424be2a32b8e881127cc03b557ee0a36ab13f430b897
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.11.13
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:7230fa1d4970e877cb6de8716e99a9bd43302c0ff23b4501f2aa806a8d496d9b
        - registry.redhat.io/quay/clair-rhel8@sha256:3746677ef1cb0ddd4f237ab598b14cf0bc76dcce080ffde127b5bba87ffe6693
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:1a083cd72502609427428ed22fb60846613ddb41a4613496c5cddc797b382d3a
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:11b8c9d64c42f0f0c7b2df491bba859d0410666a8ef1d2fd41eb77e4ad9d2dc8
        - registry.redhat.io/rhel8/postgresql-13@sha256:35d86936c539007a78a89869a14b87b76788fe117df442a670f63936b978ed5b
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:cb3a41caa3a572b7981d424be2a32b8e881127cc03b557ee0a36ab13f430b897
        version: 3.11.13
      entries:
      - name: quay-operator.v3.11.13
        version: 3.11.13
      - name: quay-operator.v3.11.12
        version: 3.11.12
      - name: quay-operator.v3.11.11
        version: 3.11.11
      - name: quay-operator.v3.11.10
        version: 3.11.10
      - name: quay-operator.v3.11.9
        version: 3.11.9
      - name: quay-operator.v3.11.8
        version: 3.11.8
      - name: quay-operator.v3.11.7
        version: 3.11.7
      - name: quay-operator.v3.11.6
        version: 3.11.6
      - name: quay-operator.v3.11.5
        version: 3.11.5
      - name: quay-operator.v3.11.4
        version: 3.11.4
      - name: quay-operator.v3.11.3
        version: 3.11.3
      - name: quay-operator.v3.11.2
        version: 3.11.2
      - name: quay-operator.v3.11.1
        version: 3.11.1
      - name: quay-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:accb705b6327fb2dcc1ce657a99f61e343eca752c9873d63d89fea89c10c3187
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.12.12
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:accb705b6327fb2dcc1ce657a99f61e343eca752c9873d63d89fea89c10c3187
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:ccbd83e05aea0bb96bf099f63420175b36ed413bfc8a618f4c90b6a2abc53895
        - registry.redhat.io/quay/clair-rhel8@sha256:dc3d5fb35f2bd5d00480351dca705fe260a225d9a97958cf0276801af08d8307
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:8e5dd171e6e94de0137bb6a972533b1f93b56efc1731e19b80869d5e9b8a26a6
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:8cdd677449f3c723dad79bcd9c4191c8917d6eb0beeeeeda68fe19a0988bb2a8
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        version: 3.12.12
      entries:
      - name: quay-operator.v3.12.12
        version: 3.12.12
      - name: quay-operator.v3.12.11
        version: 3.12.11
      - name: quay-operator.v3.12.10
        version: 3.12.10
      - name: quay-operator.v3.12.9
        version: 3.12.9
      - name: quay-operator.v3.12.8
        version: 3.12.8
      - name: quay-operator.v3.12.7
        version: 3.12.7
      - name: quay-operator.v3.12.6
        version: 3.12.6
      - name: quay-operator.v3.12.5
        version: 3.12.5
      - name: quay-operator.v3.12.4
        version: 3.12.4
      - name: quay-operator.v3.12.3
        version: 3.12.3
      - name: quay-operator.v3.12.2
        version: 3.12.2
      - name: quay-operator.v3.12.1
        version: 3.12.1
      - name: quay-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:5481a4b914ce4aed6ba3ceee8432eddab6e264cad2caef528404795f904100eb
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.13.8
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/clair-rhel8@sha256:069ccf36346cb14f4e4ea54de1d59afb72548912b140cfef67a6bcb7534d25e9
        - registry.redhat.io/rhel8/postgresql-13@sha256:f17fd8dd4d56d8e916cb2ad3f403d05ad259f79fd6918d67a38a6e9e4a8896cf
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:966a3dafa2dd31ac1a39e826e08c1b988fe694189fe79449bbe14a48b2631213
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:5481a4b914ce4aed6ba3ceee8432eddab6e264cad2caef528404795f904100eb
        - registry.redhat.io/quay/quay-rhel8@sha256:8887c17547c7a4be61471523eade9820e5095cf966b1885a3758081da559871c
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:31a9095d8f291ce68e8e688877692e74a17143bdda8d998fe3a07bcee32575dd
        - registry.redhat.io/rhel8/postgresql-15@sha256:a5f36e7893a519c12752c6118233c92d4e83618670bc9ea7bd3a8a34a63bd50b
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:f62cbd6632454c227b7cbb01dadc8e501e8288b2d8e0abf8e7842b806fa3195c
        version: 3.13.8
      entries:
      - name: quay-operator.v3.13.8
        version: 3.13.8
      - name: quay-operator.v3.13.7
        version: 3.13.7
      - name: quay-operator.v3.13.6
        version: 3.13.6
      - name: quay-operator.v3.13.5
        version: 3.13.5
      - name: quay-operator.v3.13.4
        version: 3.13.4
      - name: quay-operator.v3.13.3
        version: 3.13.3
      - name: quay-operator.v3.13.2
        version: 3.13.2
      - name: quay-operator.v3.13.1
        version: 3.13.1
      - name: quay-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:8ebc1f0511f7d394ceaa39a4c993e955c560dc04588072cb3f8c8fab8c85e32b
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.14.5
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/clair-rhel8@sha256:e11e8a819d17bd14c67f093b5303b4958855e731f89ce67dcca70fcb7b77f45a
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:8ebc1f0511f7d394ceaa39a4c993e955c560dc04588072cb3f8c8fab8c85e32b
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-rhel8@sha256:33e984892650d315065313fcb5f8808e145aadec5a287219c7031d85b713235e
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:43ef8adddbf9b007b6e2396f53bd6a5475150d28c2f4254f031a0fe6c7de5826
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:a12a0918f6dea114b0aae231d7119a4acfd0cd297074d895f4ecf190d2aec43a
        version: 3.14.5
      entries:
      - name: quay-operator.v3.14.5
        version: 3.14.5
      - name: quay-operator.v3.14.4
        version: 3.14.4
      - name: quay-operator.v3.14.3
        version: 3.14.3
      - name: quay-operator.v3.14.2
        version: 3.14.2
      - name: quay-operator.v3.14.1
        version: 3.14.1
      - name: quay-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:05fb981d9848281b97b3683093f752b1c1dc5ced439a2b4ba18b26ff8e11231e
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.15.2
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:873a3daf1dadf2438eb96e151072326c31ba34ed48cac041614b604413d5f8e7
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:956f3110f944d880bc27bb6df9472d0804cd75f5dcf4ed11bcf721e9d1d4542b
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:aa8c9f7b6ce57441ca1fe9dbaaefaf2bb83fd44aa367647b663bb822bf890362
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:05fb981d9848281b97b3683093f752b1c1dc5ced439a2b4ba18b26ff8e11231e
        - registry.redhat.io/quay/clair-rhel8@sha256:861be31d30142bae125cd059bb716d79b28694e9c4336333c68adf4a3823b61d
        version: 3.15.2
      entries:
      - name: quay-operator.v3.15.2
        version: 3.15.2
      - name: quay-operator.v3.15.1
        version: 3.15.1
      - name: quay-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-operator.v3.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:7a6e592d7f5f760b4451f9c974dbb96f56b1d3de92f039adcbf284f78ba13a02
          createdAt: 2021-10-12 22:10 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.6.10
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:ec36241f0a9a2983b1894cd8501898e6b70379af51a81b1e09b0ba0e9b54b33f
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:75df6b8e6bef14844a08f7d994470fa6ec4ffa66d4b15faaa0a3ba373f612fbb
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:7a6e592d7f5f760b4451f9c974dbb96f56b1d3de92f039adcbf284f78ba13a02
        - registry.redhat.io/rhel8/postgresql-10@sha256:4b912c80085b88a03309aeb7907efcc29dd3342fa3952b6ea067afb1914bfe53
        - registry.redhat.io/rhel8/redis-6@sha256:d7c7852338717308cbb59e9303e1ea35cc8e5c01ceb2818569be20c15f3f943d
        - registry.redhat.io/quay/quay-rhel8@sha256:e75e4d084eba9bb65b9cc72e38294af314d49316b12a0beb75db808dc057830b
        - registry.redhat.io/quay/clair-rhel8@sha256:cbafddd1eb50c1c0f9159ffafd86f44a8c0b976194903c96c6f144bfe62e749d
        version: 3.6.10
      entries:
      - name: quay-operator.v3.6.10
        version: 3.6.10
      - name: quay-operator.v3.6.9
        version: 3.6.9
      - name: quay-operator.v3.6.8
        version: 3.6.8
      - name: quay-operator.v3.6.7
        version: 3.6.7
      - name: quay-operator.v3.6.6
        version: 3.6.6
      - name: quay-operator.v3.6.5
        version: 3.6.5
      - name: quay-operator.v3.6.4
        version: 3.6.4
      - name: quay-operator.v3.6.2
        version: 3.6.2
      - name: quay-operator.v3.6.1
        version: 3.6.1
      - name: quay-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: quay-operator.v3.7.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:a8283c2be661e07c5f2f2e5b683c520846bcb91d580d9a3a0aa27ce994ddfff6
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.7.14
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:b15335d88b240bd6bbedeb7fe26f0a1de2f6c330af9758b0fb2edab7ca69f6df
        - registry.redhat.io/quay/quay-rhel8@sha256:7bc9e115e9bace07505ffd225242cab83d5eb6d50358188d18fd9161bdf1a5a6
        - registry.redhat.io/quay/clair-rhel8@sha256:d676af83e96740220e6f18721c98817ba071fd023a62ebdc3815424e94838083
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:c54c77e09c95dc7d80d7b31e796857c8d06f2f733426629b53d3298206d23b20
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:04ac405c152bffb43b33fb2e7f45df8b3940eb21e055ebf6aa17bd81ebdeeab4
        - registry.redhat.io/rhel8/postgresql-10@sha256:a52ad402458ec8ef3f275972c6ebed05ad64398f884404b9bb8e3010c5c95291
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:a8283c2be661e07c5f2f2e5b683c520846bcb91d580d9a3a0aa27ce994ddfff6
        version: 3.7.14
      entries:
      - name: quay-operator.v3.7.14
        version: 3.7.14
      - name: quay-operator.v3.7.13
        version: 3.7.13
      - name: quay-operator.v3.7.12
        version: 3.7.12
      - name: quay-operator.v3.7.11
        version: 3.7.11
      - name: quay-operator.v3.7.10
        version: 3.7.10
      name: stable-3.7
    - currentCSV: quay-operator.v3.8.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:abec0fa7ee641c8e424e995b5fe07b98e522ad1eb92c3b90426f69941f33eb2e
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.8.15
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7e4e57071439d671971b0de46b597cf6cff8ef3879ab711125a6ae7ecb9c6d29
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:3c047c7b22fb42557bd3a9b9c9ac333f0df7d5e4a4d333a2b02a4313f57085be
        - registry.redhat.io/rhel8/postgresql-10@sha256:b7fa644277d5e406b642685ff2ace1b1de7aea2e2a1421b3cbcda7d60405f44d
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:abec0fa7ee641c8e424e995b5fe07b98e522ad1eb92c3b90426f69941f33eb2e
        - registry.redhat.io/rhel8/redis-6@sha256:444200241cb9186cd97d564d7acf077d83fc997a8ee5ee7fdae41b604895085c
        - registry.redhat.io/quay/quay-rhel8@sha256:de2b65506f99eebb5eaafca6bb42b4ebec08e9479dd0d50aa1286f197e42d6de
        - registry.redhat.io/quay/clair-rhel8@sha256:b327fcfd7472cdb9d83a564904976b26b285fa576dc0a76e5590990ec77fd871
        version: 3.8.15
      entries:
      - name: quay-operator.v3.8.15
        version: 3.8.15
      - name: quay-operator.v3.8.14
        version: 3.8.14
      - name: quay-operator.v3.8.13
        version: 3.8.13
      - name: quay-operator.v3.8.12
        version: 3.8.12
      - name: quay-operator.v3.8.11
        version: 3.8.11
      - name: quay-operator.v3.8.10
        version: 3.8.10
      - name: quay-operator.v3.8.9
        version: 3.8.9
      - name: quay-operator.v3.8.8
        version: 3.8.8
      - name: quay-operator.v3.8.7
        version: 3.8.7
      - name: quay-operator.v3.8.6
        version: 3.8.6
      - name: quay-operator.v3.8.5
        version: 3.8.5
      - name: quay-operator.v3.8.4
        version: 3.8.4
      - name: quay-operator.v3.8.3
        version: 3.8.3
      - name: quay-operator.v3.8.2
        version: 3.8.2
      - name: quay-operator.v3.8.1
        version: 3.8.1
      - name: quay-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:8a16803946272156b7aeb00eb2202ab6cb87d875fe4a7c327554f770f0f2ecb5
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.9.16
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/clair-rhel8@sha256:6c67163c783d7d9fa4d187963379885a8b76685347ecf1894c4f8ebd6656544b
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:af8ada0c88b6f19991e61ee86d04b49a148c19dacb57c282e4ab98630e7499b4
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:5a42689198701d845ff3f05502cbbebceb7e6ad7b12762ccc038c0131cf6479e
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:f87b7e7ee6c31f6ae80b08aef0e10729b21fe1b8656fe18f6dee2310bae28ec1
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:8a16803946272156b7aeb00eb2202ab6cb87d875fe4a7c327554f770f0f2ecb5
        version: 3.9.16
      entries:
      - name: quay-operator.v3.9.16
        version: 3.9.16
      - name: quay-operator.v3.9.15
        version: 3.9.15
      - name: quay-operator.v3.9.14
        version: 3.9.14
      - name: quay-operator.v3.9.13
        version: 3.9.13
      - name: quay-operator.v3.9.12
        version: 3.9.12
      - name: quay-operator.v3.9.11
        version: 3.9.11
      - name: quay-operator.v3.9.10
        version: 3.9.10
      - name: quay-operator.v3.9.9
        version: 3.9.9
      - name: quay-operator.v3.9.8
        version: 3.9.8
      - name: quay-operator.v3.9.7
        version: 3.9.7
      - name: quay-operator.v3.9.6
        version: 3.9.6
      - name: quay-operator.v3.9.5
        version: 3.9.5
      - name: quay-operator.v3.9.4
        version: 3.9.4
      - name: quay-operator.v3.9.3
        version: 3.9.3
      - name: quay-operator.v3.9.2
        version: 3.9.2
      - name: quay-operator.v3.9.1
        version: 3.9.1
      - name: quay-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: quay-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MinIO
      provider-url: https://min.io
    name: minio-key-manager-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-key-manager-operator.v2025.8.19175300
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aistor.min.io/v1alpha1",
                "kind": "KeyManager",
                "metadata": {
                  "name": "my-kms",
                  "namespace": "ns-1"
                },
                "spec": {
                  "apiKeySecret": {
                    "name": "my-api-key"
                  },
                  "configuration": {
                    "name": "my-kms-server-config"
                  },
                  "hsmSecret": {
                    "name": "my-kms-hsm"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "registry-creds"
                    }
                  ],
                  "replicas": 2,
                  "volumeClaimTemplate": {
                    "metadata": {
                      "name": "key-manager-volume"
                    },
                    "spec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "100Mi"
                        }
                      },
                      "storageClassName": "standard"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security, Storage
          createdAt: "2025-08-27T14:50:59Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KeyManager
            name: keymanagers.aistor.min.io
            version: v1alpha1
        description: |-
          Highly available, powerful and operationally simple, MinIO's Enterprise Key Management Server is optimized for large storage infrastructures where billions of cryptographic keys are required.

          MinIO's AIStor KMS establishes its foundational trust using the concept of an hardware security module (but given KES is software, this is only a concept). That module assumes a pivotal role in sealing and unsealing the KMS root encryption key. The module responsibility extends to safeguarding the integrity of KMS by allowing the unsealing of its encrypted on-disk state and facilitating communication among nodes within a KMS cluster.

          It solves the challenges associated with billions of cryptographic keys and hundreds of thousands of cryptographic operations per node per second - which are commonplace in larger deployments.

          <b>High Availability and Fault Tolerance</b>
          In the dynamic landscape of large-scale systems, network or node outages are inevitable. Taking down a cluster for maintenance is rarely feasible. MinIO's AIStor KMS ensures uninterrupted availability, even when faced with such disruptions, mitigating cascading effects that can take down the entire storage infrastructure. Specifically, you could lose all but one node of a cluster and still handle any encryption, decryption or data key generation requests.

          <b>Scalability</b>
          While the amount of data usually only increases, the load on a large-scale storage system may vary significantly from time to time. MinIO's AIStor KMS supports dynamic cluster resizing and nodes can be added or removed at any point without incurring any downtime.

          <b>Multi-Tenancy</b>
          Large-scale storage infrastructures are often used by many applications and teams across the entire organization. Isolating teams and groups into their own namespaces is a core requirement. MinIO's AIStor KMS supports namespacing in the form of enclaves. Each tenant can be assigned its own enclave which is completely independent and isolated from all other enclaves on the KMS cluster.

          <b>Predictable Behavior</b>
          MinIO's AIStor KMS is designed to be easily managed, providing operators with the ability to comprehend its state intuitively. Due to its simple design, MinIO's AIStor KMS is significantly easier to operate than similar solutions that rely on more complex consensus algorithms like Raft, or Paxos.

          <b>Consistent and Performant</b>
          The responsiveness of the KMS for GET/PUT operations directly influences the overall efficiency and speed of the storage system. MinIO's AIStor KMS nodes don’t have to coordinate when handling such requests from the storage system. Therefore, the performance of a MinIO's AIStor KMS cluster increases linearly with the number of nodes. Further, MinIO's AIStor KMS supports request pipelining to handle hundreds of thousands of cryptographic operations per node and second.

          <b>Simplicity</b>
          Operating a KMS cluster does not require expertise in cryptography or distributed systems. Everything can be done from the AIStor Console.

          <a href="https://resources.min.io/request-a-demo-marketplaces/">Request a demo today</a>
        displayName: Minio AIStor Key management System (KMS)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Storage
        - MinIO
        links:
        - name: Minio AIStor
          url: https://min.io
        - name: Support
          url: https://subnet.min.io
        - name: Github
          url: https://github.com/minio/aistor
        maintainers:
        - email: dev@min.io
          name: MinIO
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: MinIO
          url: https://min.io
        relatedImages:
        - quay.io/minio/aistor/operator@sha256:f80139efa062fdd83f78318b20f371a827b68e854b03dd9706b9eba4c493aeb2
        version: 2025.8.19175300
      entries:
      - name: minio-key-manager-operator.v2025.8.19175300
        version: 2025.8.19175300
      - name: minio-key-manager-operator.v2025.8.1184927
        version: 2025.8.1184927
      - name: minio-key-manager-operator.v2025.7.1011319
        version: 2025.7.1011319
      - name: minio-key-manager-operator.v2025.6.3065135
        version: 2025.6.3065135
      - name: minio-key-manager-operator.v2025.5.12190907
        version: 2025.5.12190907
      name: stable
    defaultChannel: stable
    packageName: minio-key-manager-operator
    provider:
      name: MinIO
      url: https://min.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-acmpca-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-acmpca-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acmpca.services.k8s.aws/v1alpha1",
                "kind": "CertificateAuthority",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/acmpca-controller:1.1.1
          createdAt: "2025-09-19T20:51:07Z"
          description: AWS ACM PCA controller is a service controller for managing
            ACM PCA resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CertificateAuthority represents the state of an AWS acmpca
              CertificateAuthority resource.
            displayName: CertificateAuthority
            kind: CertificateAuthority
            name: certificateauthorities.acmpca.services.k8s.aws
            version: v1alpha1
          - description: CertificateAuthorityActivation represents the state of an
              AWS acmpca CertificateAuthorityActivation resource.
            displayName: CertificateAuthorityActivation
            kind: CertificateAuthorityActivation
            name: certificateauthorityactivations.acmpca.services.k8s.aws
            version: v1alpha1
          - description: Certificate represents the state of an AWS acmpca Certificate
              resource.
            displayName: Certificate
            kind: Certificate
            name: certificates.acmpca.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ACM PCA resources in AWS from within your Kubernetes cluster.

          **About Amazon ACM PCA**

          AWS Private CA enables creation of private certificate authority (CA) hierarchies, including root and subordinate CAs, without the investment and maintenance costs of operating an on-premises CA. Your private CAs can issue end-entity X.509 certificates useful in scenarios including:
          - Creating encrypted TLS communication channels
          - Authenticating users, computers, API endpoints, and IoT devices
          - Cryptographically signing code
          - Implementing Online Certificate Status Protocol (OCSP) for obtaining certificate revocation status

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ACM PCA
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acmpca
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ACM PCA Developer Resources
          url: https://aws.amazon.com/private-ca/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: acmpca maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/acmpca-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-acmpca-controller.v1.1.1
        version: 1.1.1
      - name: ack-acmpca-controller.v1.1.0
        version: 1.1.0
      - name: ack-acmpca-controller.v1.0.10
        version: 1.0.10
      - name: ack-acmpca-controller.v1.0.9
        version: 1.0.9
      - name: ack-acmpca-controller.v1.0.8
        version: 1.0.8
      - name: ack-acmpca-controller.v1.0.7
        version: 1.0.7
      - name: ack-acmpca-controller.v1.0.6
        version: 1.0.6
      - name: ack-acmpca-controller.v1.0.5
        version: 1.0.5
      - name: ack-acmpca-controller.v1.0.4
        version: 1.0.4
      - name: ack-acmpca-controller.v1.0.3
        version: 1.0.3
      - name: ack-acmpca-controller.v1.0.2
        version: 1.0.2
      - name: ack-acmpca-controller.v1.0.1
        version: 1.0.1
      - name: ack-acmpca-controller.v1.0.0
        version: 1.0.0
      - name: ack-acmpca-controller.v0.0.20
        version: 0.0.20
      - name: ack-acmpca-controller.v0.0.16
        version: 0.0.16
      - name: ack-acmpca-controller.v0.0.15
        version: 0.0.15
      - name: ack-acmpca-controller.v0.0.14
        version: 0.0.14
      - name: ack-acmpca-controller.v0.0.13
        version: 0.0.13
      - name: ack-acmpca-controller.v0.0.12
        version: 0.0.12
      - name: ack-acmpca-controller.v0.0.11
        version: 0.0.11
      - name: ack-acmpca-controller.v0.0.10
        version: 0.0.10
      - name: ack-acmpca-controller.v0.0.9
        version: 0.0.9
      - name: ack-acmpca-controller.v0.0.8
        version: 0.0.8
      - name: ack-acmpca-controller.v0.0.5
        version: 0.0.5
      - name: ack-acmpca-controller.v0.0.4
        version: 0.0.4
      - name: ack-acmpca-controller.v0.0.3
        version: 0.0.3
      - name: ack-acmpca-controller.v0.0.2
        version: 0.0.2
      - name: ack-acmpca-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-acmpca-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MayaData
      provider-url: ""
    name: openebs
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openebsoperator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "openebs.io/v1alpha1",
                "kind": "OpenEBSInstallTemplate",
                "metadata": {
                  "name": "oebs",
                  "namespace": "openebs"
                },
                "spec": {
                  "rbac": {
                    "create": true,
                    "pspEnabled": false,
                    "kyvernoEnabled": false
                  },
                  "serviceAccount": {
                    "create": true,
                    "name": "openebs-maya-operator"
                  },
                  "imagePullSecrets": [],
                  "release": {
                    "version": "3.0.0"
                  },
                  "legacy": {
                    "enabled": false
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": ""
                  },
                  "apiserver": {
                    "enabled": true,
                    "image": "openebs/m-apiserver",
                    "imageTag": "2.12.2",
                    "replicas": 1,
                    "ports": {
                      "externalPort": 5656,
                      "internalPort": 5656
                    },
                    "sparse": {
                      "enabled": "false"
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "defaultStorageConfig": {
                    "enabled": "true"
                  },
                  "varDirectoryPath": {
                    "baseDir": "/var/openebs"
                  },
                  "provisioner": {
                    "enabled": true,
                    "image": "openebs/openebs-k8s-provisioner",
                    "imageTag": "2.12.2",
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "patchJivaNodeAffinity": "enabled",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "localprovisioner": {
                    "enabled": true,
                    "image": "openebs/provisioner-localpv",
                    "imageTag": "3.0.0",
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "enableDeviceClass": true,
                    "enableHostpathClass": true,
                    "basePath": "/var/openebs/local",
                    "waitForBDBindTimeoutRetryCount": "12",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "snapshotOperator": {
                    "enabled": true,
                    "controller": {
                      "image": "openebs/snapshot-controller",
                      "imageTag": "2.12.2",
                      "resources": {}
                    },
                    "provisioner": {
                      "image": "openebs/snapshot-provisioner",
                      "imageTag": "2.12.2",
                      "resources": {}
                    },
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "upgradeStrategy": "Recreate",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    }
                  },
                  "ndm": {
                    "enabled": true,
                    "image": "openebs/node-disk-manager",
                    "imageTag": "1.7.0",
                    "sparse": {
                      "path": "/var/openebs/sparse",
                      "size": "10737418240",
                      "count": "0"
                    },
                    "filters": {
                      "enableOsDiskExcludeFilter": true,
                      "osDiskExcludePaths": "/,/etc/hosts,/boot",
                      "enableVendorFilter": true,
                      "excludeVendors": "CLOUDBYT,OpenEBS",
                      "enablePathFilter": true,
                      "includePaths": "",
                      "excludePaths": "/dev/loop,/dev/fd0,/dev/sr0,/dev/ram,/dev/dm-,/dev/md,/dev/rbd,/dev/zd"
                    },
                    "probes": {
                      "enableSeachest": false
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "ndmOperator": {
                    "enabled": true,
                    "image": "openebs/node-disk-operator",
                    "imageTag": "1.7.0",
                    "replicas": 1,
                    "upgradeStrategy": "Recreate",
                    "nodeSelector": {},
                    "tolerations": [],
                    "healthCheck": {
                      "initialDelaySeconds": 15,
                      "periodSeconds": 20
                    },
                    "readinessCheck": {
                      "initialDelaySeconds": 5,
                      "periodSeconds": 10
                    },
                    "resources": {}
                  },
                  "ndmExporter": {
                    "enabled": false,
                    "image": {
                      "registry": null,
                      "repository": "openebs/node-disk-exporter",
                      "pullPolicy": "IfNotPresent",
                      "tag": "1.7.0"
                    },
                    "nodeExporter": {
                      "name": "ndm-node-exporter",
                      "podLabels": {
                        "name": "openebs-ndm-node-exporter"
                      },
                      "metricsPort": 9101
                    },
                    "clusterExporter": {
                      "name": "ndm-cluster-exporter",
                      "podLabels": {
                        "name": "openebs-ndm-cluster-exporter"
                      },
                      "metricsPort": 9100
                    }
                  },
                  "webhook": {
                    "enabled": true,
                    "image": "openebs/admission-server",
                    "imageTag": "2.12.2",
                    "failurePolicy": "Fail",
                    "replicas": 1,
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "hostNetwork": false,
                    "resources": {}
                  },
                  "helper": {
                    "image": "openebs/linux-utils",
                    "helperImageTag": "3.0.0"
                  },
                  "featureGates": {
                    "enabled": true,
                    "GPTBasedUUID": {
                      "enabled": true,
                      "featureGateFlag": "GPTBasedUUID"
                    },
                    "APIService": {
                      "enabled": false,
                      "featureGateFlag": "APIService",
                      "address": "0.0.0.0:9115"
                    },
                    "UseOSDisk": {
                      "enabled": false,
                      "featureGateFlag": "UseOSDisk"
                    },
                    "ChangeDetection": {
                      "enabled": false,
                      "featureGateFlag": "ChangeDetection"
                    }
                  },
                  "crd": {
                    "enableInstall": true
                  },
                  "policies": {
                    "monitoring": {
                      "enabled": true,
                      "image": "openebs/m-exporter",
                      "imageTag": "2.12.2"
                    }
                  },
                  "analytics": {
                    "enabled": true,
                    "pingInterval": "24h"
                  },
                  "jiva": {
                    "image": "openebs/jiva",
                    "imageTag": "2.12.2",
                    "replicas": 3,
                    "defaultStoragePath": "/var/openebs",
                    "enabled": false,
                    "openebsLocalpv": {
                      "enabled": false
                    },
                    "localpv-provisioner": {
                      "openebsNDM": {
                        "enabled": false
                      }
                    }
                  },
                  "cstor": {
                    "pool": {
                      "image": "openebs/cstor-pool",
                      "imageTag": "2.12.2"
                    },
                    "poolMgmt": {
                      "image": "openebs/cstor-pool-mgmt",
                      "imageTag": "2.12.2"
                    },
                    "target": {
                      "image": "openebs/cstor-istgt",
                      "imageTag": "2.12.2"
                    },
                    "volumeMgmt": {
                      "image": "openebs/cstor-volume-mgmt",
                      "imageTag": "2.12.2"
                    },
                    "enabled": false,
                    "openebsNDM": {
                      "enabled": false
                    }
                  },
                  "openebs-ndm": {
                    "enabled": false
                  },
                  "localpv-provisioner": {
                    "enabled": false,
                    "openebsNDM": {
                      "enabled": false
                    }
                  },
                  "zfs-localpv": {
                    "enabled": false
                  },
                  "lvm-localpv": {
                    "enabled": false
                  },
                  "nfs-provisioner": {
                    "enabled": false
                  },
                  "cleanup": {
                    "image": {
                      "registry": "",
                      "repository": "bitnami/kubectl",
                      "tag": "",
                      "imagePullSecrets": []
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: index.docker.io/openebs/helm-operator:v0.0.9
          createdAt: "2021-11-02T02:56:58Z"
          description: Creates and maintains OpenEBS Control Plane deployments
          repository: https://github.com/openebs/helm-operator
          support: https://slack.openebs.io/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a OpenEBS Install Operator
            displayName: OpenEBS Install Template
            kind: OpenEBSInstallTemplate
            name: openebsinstalltemplates.openebs.io
            version: v1alpha1
        description: |
          **OpenEBS** is a leading container attached storage solution that enables the use of containers for mission-critical, persistent workloads and for other stateful workloads such as logging or Prometheus for example.
          OpenEBS itself is deployed as just another container on your host and enables storage services that can be designated on a per pod, application, cluster or container level, including:
          * Data persistence across nodes
          * Synchronization of data across availability zones and cloud providers
          * A common layer whether you are running on the cloud, or your bare metal
          * Integration with Kubernetes, so developer and application intent flows into OpenEBS
          * Management of tiering to and from S3 and other targets.
          ## OpenEBS Operator
          OpenEBS primarily provides container attached block storage (iSCSI volumes) by leveraging/aggregating the storage on the nodes, with the storage controller itself running as a container. Different storage engines (Jiva & cStor) are supported, with tools available to dynamically provision Kubernetes Local PVs. The volumes are dynamically provisioned via PersistentVolumeClaims and are managed by a control plane component called "maya", which also runs as a deployment in the K8s cluster. In addition to maya, a typical OpenEBS installation comprises several other resources, which aid with various functionalities, ranging from snapshotting to disk management. All these components are described briefly below:
          * **Maya-API-Server** - A storage orchestrator which integrates into Kubernetes workflow to help provision
            and manage OpenEBS Jiva & cStor (storage engine) volumes

          * **Dynamic-OpenEBS-Provisioner** - A Kubernetes external storage provisioner that utilizes APIs exposed by maya-apiserver
            to perform provision & delete operations of Jiva & cStor volumes

          * **Dynamic-LocalPV-Provisioner** - A dynamic provisioner for Kubernetes Local PVs
          * **OpenEBS-Snapshot-Operator** - A Kubernetes snapshot controller that creates & restores OpenEBS volume snapshots
          * **Node-Disk-Manager** - A disk management controller which identifies available disks, maintains inventory, and
            dynamically attaches/detaches disks to backend storage pods

          The helm-based OpenEBS Operator eases the setup of all the above mentioned components, with a simple custom resource provided to define the install options, thereby enabling applications to start using the OpenEBS storageclasses in their PVCs. The OpenEBSInstallTemplate CR can be used to specify start-up parameters & also update/overwrite the definitions post install.
          ## Pre-Requisites
          Before installing OpenEBS control plane, perform the following steps to ensure successful deployment of the Node-Disk-Manager & the OpenEBS volume replicas, respectively.
          * Configure the OpenEBS service account on the openshift-operators namespace/project to use the privileged security context constraint.

            **Note**: The serviceaccount name is same as the one specified in the `spec.serviceAccount.name` field of the OpenEBSInstallTemplate CR.


            ```
            oc adm policy add-scc-to-user privileged system:serviceaccount:openshift-operators:openebs-maya-operator
            ```

          * Configure the default service account on the namespace/project in which the volume replicas are deployed to use privileged
            security context constraint.


            ```
            oc adm policy add-scc-to-user privileged system:serviceaccount:openshift-operators:default
            ```

          ## Getting Started
          * Try the quickstart [guide](https://github.com/openebs/helm-operator/blob/master/olm/README.md)
          * To learn how to contribute, please read the [contribution guide](https://github.com/openebs/helm-operator/blob/master/CONTRIBUTING.md)
          * OpenEBS welcomes your feedback and contributions in any form possible. [Join our Community](https://openebs.org/community)

          ## License
          OpenEBS is distributed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt)
        displayName: OpenEBS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenEBS
        - ContainerAttachedStorage
        links:
        - name: OpenEBS Website
          url: https://openebs.io
        - name: Operator Source Code
          url: https://github.com/openebs/helm-operator
        - name: Install Instructions
          url: https://github.com/openebs/helm-operator/blob/master/olm/README.md
        maintainers:
        - email: kiran.mova@openebs.io
          name: kmova
        maturity: alpha
        minKubeVersion: 1.12.0
        provider:
          name: MayaData
        relatedImages:
        - index.docker.io/openebs/helm-operator:v0.0.9
        version: 3.0.0
      entries:
      - name: openebsoperator.v3.0.0
        version: 3.0.0
      - name: openebsoperator.v2.12.2
        version: 2.12.2
      name: alpha
    defaultChannel: alpha
    packageName: openebs
    provider:
      name: MayaData
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-core
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Core",
                "metadata": {
                  "name": "core-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configService": {
                    "port": 8080
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "helper": {
                    "port": 8080
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "licenseService": {
                    "port": 8080
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "restStatsService": {
                    "port": 8080
                  },
                  "resultService": {
                    "port": 8080,
                    "storage": "64Gi"
                  },
                  "sessionManager": {
                    "port": 8080
                  },
                  "statsService": {
                    "port": 8080
                  },
                  "testResultsService": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:34:00Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Core
            name: cores.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Core Operator
        displayName: Keysight WAP Core Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Core
          url: https://keysight-wap-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-core@sha256:001b0d08cd84a9ebc13cd8fd35b68d431b96f84d71e933c1123ed6c0d244fd37
        version: 0.1.21
      entries:
      - name: keysight-wap-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-nats-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-nats-operator.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2023-09-19T12:44:10Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NatsCluster
            name: natsclusters.nats.io
            version: v1alpha2
          - kind: NatsServiceRole
            name: natsserviceroles.nats.io
            version: v1alpha2
        description: Keysight Nats Operator
        displayName: Keysight Nats Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Nats Operator
          url: https://keysight-nats-operator.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-nats-operator@sha256:2ee5dc8814274e83a5f82e89eabd0b643e8cde8e3d555a56535a61d05c72fd75
        version: 0.1.21
      entries:
      - name: keysight-nats-operator.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-nats-operator
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: github.com/stakater-ab/hibernation-operator
    name: hibernation-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hibernation-operator.v0.1.103
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Monitoring
          createdAt: "2025-10-02T05:44:42Z"
          description: Operator to engage and disengage hibernation on k8s resources
            using cron expressions
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.1.103
          operatorframework.io/suggested-namespace: hibernation-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Cluster Resource Supervisor
            kind: ClusterResourceSupervisor
            name: clusterresourcesupervisors.hibernation.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.hibernation.stakater.com
            version: v1beta1
        description: Schedule k8s resources using cron expressions via Hibernation
          Operator
        displayName: Hibernation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - redhat
        - k8s
        - stakater
        - hibernation
        - operator
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maintainers:
        - email: hello@stakater.com
          name: StakaterAB
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Stakater
          url: github.com/stakater-ab/hibernation-operator
        relatedImages:
        - registry.connect.redhat.com/stakater/hibernation-operator@sha256:3738f82834d53aec4f8fba5caad6258c7adb4bdc23a57bde47c7ec22030fae5c
        version: 0.1.103
      entries:
      - name: hibernation-operator.v0.1.103
        version: 0.1.103
      name: release-0.1
    defaultChannel: release-0.1
    packageName: hibernation-operator
    provider:
      name: Stakater
      url: github.com/stakater-ab/hibernation-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: runtime-component-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runtime-component.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/runtime-component-operator@sha256:d37b3f62575c9ad1ffcb22c89e2c3886275c326bf27aa49fb1808acba475853d
          createdAt: "2022-03-25T14:00:00Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          olm.skipRange: <0.8.1
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          This advanced Operator is capable of deploying any runtime component image with consistent, production-grade QoS. It enables enterprise architects to govern the way their applications get deployed & managed in the cluster, while dramatically reducing the learning curve for developers to deploy into Kubernetes - allowing them to focus on writing the code!
          Here are some key features:

          #### Application Lifecyle
          You can deploy your runtime component container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the Operator will watch for any updates and will re-deploy the modified image.

          #### Custom RBAC
          This Operator is capable of using a custom ServiceAccount from the caller, allowing it to follow RBAC restrictions. By default it creates a ServiceAccount if one is not specified, which can also be bound with specific roles.

          #### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as: labels, annotations, and environment variables from a ConfigMap, a Secret or a value.

          #### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configuration, such as TLS settings, are also easily enabled.

          #### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto scaling to create (and delete) instances based on resource consumption.

          #### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you.
          This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart.
          We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.

          #### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating kubernetes Secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to its required services without any intervention or interruption.

          #### Exposing metrics to Prometheus
          The Runtime Component Operator exposes the runtime container's metrics via the Prometheus Operator.
          You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `spec.monitoring` field to control configurations such as the poll interval and security credentials.

          #### Easily mount logs and transaction directories
          If you need to mount the logs and transaction data from your runtime component to an external volume such as NFS (or any storage supported in your cluster), simply add the following (to specify the volume size and the location to persist) to your RuntimeComponent CR:
          ``` storage: size: 2Gi mountPath: "/logs" ```

          #### Integration with OpenShift Serverless
          Deploy your serverless runtime component using a single toggle.  The Operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing your pod to automatically scale to 0 when it is idle.

          #### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and how they are connected.

          See our [**documentation**](https://github.com/application-stacks/runtime-component-operator/tree/main/doc/) for more information.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://github.com/application-stacks/runtime-component-operator/tree/main/doc
        - name: Repository
          url: https://github.com/application-stacks/runtime-component-operator
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: beta
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/runtime-component-operator@sha256:d37b3f62575c9ad1ffcb22c89e2c3886275c326bf27aa49fb1808acba475853d
        - registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4
        version: 0.8.1
      entries:
      - name: runtime-component.v0.8.1
        version: 0.8.1
      - name: runtime-component.v0.8.0
        version: 0.8.0
      name: beta2
    - currentCSV: runtime-component.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:1b818f1a9c27baab6a39ae42d86f78bac085d772cae5cd97b357d0ee5799cf85
          createdAt: "2023-08-28T16:46:50Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          olm.skipRange: '>=0.8.0 <1.2.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02
        - icr.io/appcafe/runtime-component-operator@sha256:1b818f1a9c27baab6a39ae42d86f78bac085d772cae5cd97b357d0ee5799cf85
        version: 1.2.2
      entries:
      - name: runtime-component.v1.2.2
        version: 1.2.2
      name: v1.2
    - currentCSV: runtime-component.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:682c5cb3e24566bf2933b1d4a4f6bf520d9b546edcedc1ee3214ec1ac4d48629
          createdAt: "2024-08-01T20:26:42Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.3.3'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/runtime-component-operator@sha256:682c5cb3e24566bf2933b1d4a4f6bf520d9b546edcedc1ee3214ec1ac4d48629
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91
        version: 1.3.3
      entries:
      - name: runtime-component.v1.3.3
        version: 1.3.3
      name: v1.3
    - currentCSV: runtime-component.v1.4.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:6f527c5173421c28d175b42d8c690eb63c4d1c5092c31d5350fd7a5e93661eaa
          createdAt: "2025-07-18T20:38:10Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.4.4'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/runtime-component-operator@sha256:6f527c5173421c28d175b42d8c690eb63c4d1c5092c31d5350fd7a5e93661eaa
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3
        version: 1.4.4
      entries:
      - name: runtime-component.v1.4.4
        version: 1.4.4
      name: v1.4
    - currentCSV: runtime-component.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:58bfceaed2fc828bce1c0e4e21a583d25f6d72f88abdc78df2510867a4f2020f
          createdAt: "2025-09-15T18:08:27Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.5.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8
        - icr.io/appcafe/runtime-component-operator@sha256:58bfceaed2fc828bce1c0e4e21a583d25f6d72f88abdc78df2510867a4f2020f
        version: 1.5.0
      entries:
      - name: runtime-component.v1.5.0
        version: 1.5.0
      name: v1.5
    defaultChannel: v1.5
    packageName: runtime-component-operator-certified
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: glance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: glance-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "glance.openstack.org/v1beta1",
                "kind": "Glance",
                "metadata": {
                  "name": "glance"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\nenabled_backends=foo:bar,foo1:bar1\n",
                  "databaseAccount": "glance",
                  "databaseInstance": "openstack",
                  "secret": "osp-secret",
                  "serviceUser": "glance",
                  "storage": {
                    "storageClass": "local-storage",
                    "storageRequest": "10G"
                  }
                }
              },
              {
                "apiVersion": "glance.openstack.org/v1beta1",
                "kind": "GlanceAPI",
                "metadata": {
                  "annotation": {
                    "keystoneEndpoint": "true"
                  },
                  "name": "default"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\nenabled_backends=foo:bar,foo1:bar1\n[foo]\nhere-foo-config\n[foo1]\nhere-foo1-config\n",
                  "databaseAccount": "glance",
                  "databaseHostname": "glance",
                  "preserveJobs": false,
                  "replicas": 1,
                  "secret": "glance-secret",
                  "serviceAccount": "glance",
                  "serviceUser": "glance",
                  "storage": {
                    "storageRequest": "10G"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:45Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GlanceAPI is the Schema for the glanceapis API
            displayName: Glance API
            kind: GlanceAPI
            name: glanceapis.glance.openstack.org
            version: v1beta1
          - description: Glance is the Schema for the glances API
            displayName: Glance
            kind: Glance
            name: glances.glance.openstack.org
            version: v1beta1
        description: Glance Operator
        displayName: Glance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Image
        - Glance
        links:
        - name: Glance Operator
          url: https://github.com/openstack-k8s-operators/glance-operator
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/glance-rhel9-operator@sha256:74f9907b59493239fbe9c5f534399ff513c693a2bda11a037aa2cc5dc8cc3e91
        - registry.redhat.io/rhoso/openstack-glance-api-rhel9@sha256:b8f84f9b432a70f7ca7aebda60455a4890347529f6bed9a440a3b197e377e0a1
        version: 1.0.6
      entries:
      - name: glance-operator.v1.0.6
        version: 1.0.6
      - name: glance-operator.v1.0.5
        version: 1.0.5
      - name: glance-operator.v1.0.4
        version: 1.0.4
      - name: glance-operator.v1.0.3
        version: 1.0.3
      - name: glance-operator.v1.0.2
        version: 1.0.2
      - name: glance-operator.v1.0.1
        version: 1.0.1
      - name: glance-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: glance-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kafka-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kafka-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kafka-controller:1.2.1
          createdAt: "2025-09-20T21:41:07Z"
          description: AWS Kafka controller is a service controller for managing Kafka
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster represents the state of an AWS kafka Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.kafka.services.k8s.aws
            version: v1alpha1
          - description: Configuration represents the state of an AWS kafka Configuration
              resource.
            displayName: Configuration
            kind: Configuration
            name: configurations.kafka.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCluster represents the state of an AWS kafka ServerlessCluster
              resource.
            displayName: ServerlessCluster
            kind: ServerlessCluster
            name: serverlessclusters.kafka.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Kafka resources in AWS from within your Kubernetes cluster.

          **About Amazon Kafka**

          Amazon Managed Streaming for Apache Kafka (Amazon MSK) is a fully managed AWS streaming data service makes it easy for you to build and run applications that use Apache Kafka to process streaming data without needing to become an expert in operating Apache Kafka clusters. Amazon MSK manages the provisioning, configuration, and maintenance of Apache Kafka clusters and Apache ZooKeeper nodes for you. Amazon MSK also shows key Apache Kafka performance metrics in the AWS Console.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Kafka Developer Resources
          url: https://aws.amazon.com/msk/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kafka maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/kafka-controller:1.2.1
        version: 1.2.1
      entries:
      - name: ack-kafka-controller.v1.2.1
        version: 1.2.1
      - name: ack-kafka-controller.v1.2.0
        version: 1.2.0
      - name: ack-kafka-controller.v1.1.3
        version: 1.1.3
      - name: ack-kafka-controller.v1.1.2
        version: 1.1.2
      - name: ack-kafka-controller.v1.1.1
        version: 1.1.1
      - name: ack-kafka-controller.v1.1.0
        version: 1.1.0
      - name: ack-kafka-controller.v1.0.14
        version: 1.0.14
      - name: ack-kafka-controller.v1.0.13
        version: 1.0.13
      - name: ack-kafka-controller.v1.0.12
        version: 1.0.12
      - name: ack-kafka-controller.v1.0.11
        version: 1.0.11
      - name: ack-kafka-controller.v1.0.10
        version: 1.0.10
      - name: ack-kafka-controller.v1.0.8
        version: 1.0.8
      - name: ack-kafka-controller.v1.0.7
        version: 1.0.7
      - name: ack-kafka-controller.v1.0.6
        version: 1.0.6
      - name: ack-kafka-controller.v1.0.5
        version: 1.0.5
      - name: ack-kafka-controller.v1.0.4
        version: 1.0.4
      - name: ack-kafka-controller.v1.0.3
        version: 1.0.3
      - name: ack-kafka-controller.v1.0.2
        version: 1.0.2
      - name: ack-kafka-controller.v1.0.1
        version: 1.0.1
      - name: ack-kafka-controller.v1.0.0
        version: 1.0.0
      - name: ack-kafka-controller.v0.0.14
        version: 0.0.14
      - name: ack-kafka-controller.v0.0.12
        version: 0.0.12
      - name: ack-kafka-controller.v0.0.11
        version: 0.0.11
      - name: ack-kafka-controller.v0.0.10
        version: 0.0.10
      - name: ack-kafka-controller.v0.0.9
        version: 0.0.9
      - name: ack-kafka-controller.v0.0.8
        version: 0.0.8
      - name: ack-kafka-controller.v0.0.7
        version: 0.0.7
      - name: ack-kafka-controller.v0.0.6
        version: 0.0.6
      - name: ack-kafka-controller.v0.0.5
        version: 0.0.5
      - name: ack-kafka-controller.v0.0.4
        version: 0.0.4
      - name: ack-kafka-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-kafka-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Apache Software Foundation
      provider-url: ""
    name: camel-k
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: camel-k-operator.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.0.1
          createdAt: "2023-09-04T12:32:38Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
          - kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.0.1
        version: 2.0.1
      entries:
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: candidate
    - currentCSV: camel-k-operator.v2.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "setBody": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationProfile",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "debug"
                },
                "spec": {
                  "traits": {
                    "logging": {
                      "level": "DEBUG"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example-source"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "setBody": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "timer-to-log-example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "log-sink"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world",
                      "period": 2000
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "timer-source"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.8.0
          createdAt: "2025-09-21T09:07:43Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: IntegrationProfile is the Schema for the integrationprofiles
              API
            displayName: Integration Profile
            kind: IntegrationProfile
            name: integrationprofiles.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Pipe is the Schema for the pipes API
            displayName: Pipe
            kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers:
          the easiest way to build and manage your Camel applications on Kubernetes.

          ## Installation

          You may need to specify the container registry where to store the container images created for the Camel applications.
          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          spec:
            build:
              registry:
                address: docker.io/my-org
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step when using Openshift cluster.

          ## Running an Integration

          Run a Camel application on the cluster by creating a sample `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            annotations:
              camel.apache.org/operator.id: camel-k
            name: test
          spec:
            flows:
            - from:
                parameters:
                  period: "1000"
                steps:
                - setBody:
                    simple: Hello Camel from ${routeId}
                - log: ${body}
                uri: timer:yaml
          ```
          Log the result of the running application:
          ```
          kubectl wait --for=condition=ready integration test --timeout=180s
          kubectl logs deployment/test -f
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.8.0
        version: 2.8.0
      entries:
      - name: camel-k-operator.v2.8.0
        version: 2.8.0
      - name: camel-k-operator.v2.7.0
        version: 2.7.0
      - name: camel-k-operator.v2.6.0
        version: 2.6.0
      - name: camel-k-operator.v2.5.1
        version: 2.5.1
      - name: camel-k-operator.v2.5.0
        version: 2.5.0
      - name: camel-k-operator.v2.4.0
        version: 2.4.0
      - name: camel-k-operator.v2.3.3
        version: 2.3.3
      - name: camel-k-operator.v2.3.2
        version: 2.3.2
      - name: camel-k-operator.v2.3.1
        version: 2.3.1
      - name: camel-k-operator.v2.3.0
        version: 2.3.0
      - name: camel-k-operator.v2.2.0
        version: 2.2.0
      - name: camel-k-operator.v2.1.0
        version: 2.1.0
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: latest
    - currentCSV: camel-k-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.12.0
          createdAt: "2023-02-27T09:01:55Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.12.0
        version: 1.12.0
      entries:
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable
    - currentCSV: camel-k-operator.v1.10.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.10.4
          createdAt: "2023-01-28T13:49:05Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.10.4
        version: 1.10.4
      entries:
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.10
    - currentCSV: camel-k-operator.v1.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.11.1
          createdAt: "2023-01-30T08:45:04Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.11.1
        version: 1.11.1
      entries:
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.11
    - currentCSV: camel-k-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.12.1
          createdAt: "2023-06-12T08:46:58Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.12.1
        version: 1.12.1
      entries:
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.12
    - currentCSV: camel-k-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.5.1
          createdAt: "2021-08-31T09:03:41Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.5.1
        version: 1.5.1
      entries:
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.5
    - currentCSV: camel-k-operator.v1.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.6.1
          createdAt: "2021-10-27T21:06:20Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.6.1
        version: 1.6.1
      entries:
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.6
    - currentCSV: camel-k-operator.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.7.0
          createdAt: "2021-11-16T10:54:12Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.8.1+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.7.0
        version: 1.7.0
      entries:
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.7
    - currentCSV: camel-k-operator.v1.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.8.2
          createdAt: "2022-03-07T07:13:19Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.8.2
        version: 1.8.2
      entries:
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.8
    - currentCSV: camel-k-operator.v1.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.9.2
          createdAt: "2022-05-23T09:05:22Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.9.2
        version: 1.9.2
      entries:
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.9
    - currentCSV: camel-k-operator.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.0.1
          createdAt: "2023-09-04T12:32:38Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
          - kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.0.1
        version: 2.0.1
      entries:
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-2.0
    - currentCSV: camel-k-operator.v2.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "setBody": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationProfile",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "debug"
                },
                "spec": {
                  "traits": {
                    "logging": {
                      "level": "DEBUG"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example-source"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "setBody": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "timer-to-log-example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "log-sink"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world",
                      "period": 2000
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "timer-source"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.8.0
          createdAt: "2025-09-21T09:07:43Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: IntegrationProfile is the Schema for the integrationprofiles
              API
            displayName: Integration Profile
            kind: IntegrationProfile
            name: integrationprofiles.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Pipe is the Schema for the pipes API
            displayName: Pipe
            kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers:
          the easiest way to build and manage your Camel applications on Kubernetes.

          ## Installation

          You may need to specify the container registry where to store the container images created for the Camel applications.
          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          spec:
            build:
              registry:
                address: docker.io/my-org
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step when using Openshift cluster.

          ## Running an Integration

          Run a Camel application on the cluster by creating a sample `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            annotations:
              camel.apache.org/operator.id: camel-k
            name: test
          spec:
            flows:
            - from:
                parameters:
                  period: "1000"
                steps:
                - setBody:
                    simple: Hello Camel from ${routeId}
                - log: ${body}
                uri: timer:yaml
          ```
          Log the result of the running application:
          ```
          kubectl wait --for=condition=ready integration test --timeout=180s
          kubectl logs deployment/test -f
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.8.0
        version: 2.8.0
      entries:
      - name: camel-k-operator.v2.8.0
        version: 2.8.0
      - name: camel-k-operator.v2.7.0
        version: 2.7.0
      - name: camel-k-operator.v2.6.0
        version: 2.6.0
      - name: camel-k-operator.v2.5.1
        version: 2.5.1
      - name: camel-k-operator.v2.5.0
        version: 2.5.0
      - name: camel-k-operator.v2.4.0
        version: 2.4.0
      - name: camel-k-operator.v2.3.3
        version: 2.3.3
      - name: camel-k-operator.v2.3.2
        version: 2.3.2
      - name: camel-k-operator.v2.3.1
        version: 2.3.1
      - name: camel-k-operator.v2.3.0
        version: 2.3.0
      - name: camel-k-operator.v2.2.0
        version: 2.2.0
      - name: camel-k-operator.v2.1.0
        version: 2.1.0
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-v2
    defaultChannel: stable-v2
    packageName: camel-k
    provider:
      name: The Apache Software Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-cpd-ccs
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-ccs.v10.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ccs.cpd.ibm.com/v1beta1",
                "kind": "CCS",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ccs-cr",
                    "app.kubernetes.io/managed-by": "ibm-cpd-ccs-operator",
                    "app.kubernetes.io/name": "ibm-cpd-ccs-operator"
                  },
                  "name": "ccs-cr"
                },
                "spec": {
                  "scaleConfig": "small",
                  "license": {
                    "accept": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
          createdAt: "2025-04-18"
          description: Common Core Services
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-ccs
          olm.skipRange: '>=7.0.0 <10.3.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM watsonx.ai"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: |-
            [
              "runtimeassemblies.runtimes.ibm.com"
            ]
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation Runtime Manager CRD. This product is not deployable
              standalone and it is part of the LICCR already accepted.  For additional
              details regarding install parameters check: http://ibm.biz/-readme.
              License By installing this product you accept the license terms https://ibm.biz/BdaaLC
              or https://ibm.biz/BdaaL9.'
            kind: RuntimeAssembly
            name: runtimeassemblies.runtimes.ibm.com
            version: v1alpha1
          - description: 'Documentation This product is not deployable standalone
              and it is part of the LICCR already accepted.  For additional details
              regarding install parameters check: http://ibm.biz/-readme. License
              By installing this product you accept the license terms https://ibm.biz/BdaaLC
              or https://ibm.biz/BdaaL9.'
            kind: CCS
            name: ccs.ccs.cpd.ibm.com
            version: v1beta1
        description: |-
          # Introduction Common Core Services Operator
          Put stuff here

          # Details
          Various Details
          ## Prerequisites
          Various Prerequisites
          ## Resources Required
          Various Resources Required
          ## Installing
          Various Installation details
          # Configuration
          Various Configuration information
          ### Limitations
          Various Limitations
          ### SecurityContextConstraints Requirements
          All pods will use the restricted SCC, I think.
        displayName: Common Core Services
        installModes:
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: OwnNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commonc-core-services
        - cpd40
        links:
        - name: Ccs
          url: https://ccs.domain
        maintainers:
        - email: test@ibm.com
          name: test
        maturity: v1
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/jobs-api@sha256:173b88f90078e1817686ac949f48387af70a706e23e6fdab4a7c6217db5ad86e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-bundle@sha256:94c7a1dd290e2abd55c5b8998cf08ad174d3fbfc659ae6fa6fb6cbab61d5fe14
        - registry.marketplace.redhat.com/cp/cpd/dap-base-job@sha256:f208686a3f56d2980e7b8640b6ebdea9057109e7520c6817a731e68d08506a36
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:bcdf9b7eea8753b3e463973a2690140ae7aa8c02c3c7726511f116bd751ec2d8
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:c268d2852580286e5c6770857fcfc3b06a698521f5acd17ed83b8e2cb9ab55cf
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:523dcc671f1c05e690d5e578a90300d28d76570f605d4de7f80645b590f4a462
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:81d8b25221d827160dd5e88efd383453f532057ad02d7a2ee3dcd9c6eb16ba65
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:0e0ae96aff8540323da8a2379af014d3af1770e4b0043b84b9b415dda44501f7
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:6f2e7cc917adb54b3112dc4909f645ac08c33e8c2c33266fbc8f5133cbcd1973
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:21c81ff7c4b8809accaf62b242fd9b3d52f235489ae427556b7f71cc21bf8926
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:351375de35d7fc9d940b905b867ce8a00b7ff1416e70d3a100918797fe3d8c56
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:4fb2127ba7c525a959bbbcf024de364ef266ebb3ca7e696f74d990d2552cda59
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:56315ce0a0a70f5899191a7eab856105489143f0e889c8cdd1b6b8c225f02eb7
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:a3a0d58f895fca4d832e602d0dee80018df4ad7bad8800ad4058eea07fda4817
        - registry.marketplace.redhat.com/cp/cpd/ngp-projects-api@sha256:df420e9f616ad8a123d9eedd117fb6cb007b30a772138db534f47ad8801fa96d
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:257bf301338f7b123099e2ffe31c3f3b793bf28d6a0570a25b889271239579ba
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:6f96b8b6afb00cb26fcf903b8e4c30b0d6344bb86d8bce2f2026fc8cb0b575ec
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:0621eb9b83c9b2b8bfe5cd09797fe08fa37a854a5506914935ff6f0df599b631
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:0ae22914dbaf09d6679bd06bc9b6a5728e744c9ce2309cbd5f01e28308c403ea
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:562281dbb2bec5191a7c6b595ab853092e6fe3ff1432e77b2471f71ccdffcbc8
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:54c86367bfe0ab658611856cc0a6c5de24c14a336bd129ac9f16293945a55238
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:f5527ddbbd0b9878555fce852996b13020fd91b1604d8c03d34a61ca2fdb2d1d
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:fbc6eab07fcfc2d0c1895f18e17f751e9481ad111875959e9ff08e14f81b9a35
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:740a1d87a43e61e32cfdea032d852af33100d959c8b57ff2f165f4baf77fc011
        - registry.marketplace.redhat.com/cp/cpd/task-credentials@sha256:e162828d5ad07ceae4db0589f930979842714e3b766552102bee373a3a570af3
        - registry.marketplace.redhat.com/cp/cpd/environments-ui@sha256:2923664475c7a8ea5e63055c90748db37825fe76c7975b845f1185d4b32ee4ea
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/cpd/runtime-manager-api@sha256:a37773df07fe757dd698e3f09f8687902e7b3592d97855e703dd92ecc832aab0
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:8a6ce0dada706056efda73ad0cc1547566003e0a7aab8ed379b50e4776861b94
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e12599ff99db058fdc8f9c1eebfba31160c636ff092138deb1f6f6705f44fd62
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:3f2f97e6b2b18526255e88ff02d1039b24312eca51ee22b7cbf76d5680f7aee1
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:d023f35b65816f244b6d699b41362e6031553f7d4c416efcee4a7f2395a61355
        - registry.marketplace.redhat.com/cp/cpd/spaces-ui-job@sha256:95dade51444e30ec034c12207ecbef675b70889b81970f49f10aa75160b2f947
        - registry.marketplace.redhat.com/cp/cpd/environment-specifications-sync@sha256:894b4624d8ead7286cf9068c6477300992c32d96339644fe764bcb29b1c8ef5c
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:ad19b65042fb26cc1244a92f81a96e5ceb4313f8c823181efe28b66e9fecf9c8
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:044b663e4b7508c62dbb28b7fb2aa267c9771fc62ab509b08d46ea1499d95431
        - registry.marketplace.redhat.com/cp/cpd/environments-api@sha256:6dd60b530f81d7a2d55a534211f7f128814abf87b8cd3575d528dba53442c44e
        - registry.marketplace.redhat.com/cp/cpd/jobs-ui@sha256:49daf607918b378593485da16ac417056ee295bca94f1a5cac37b334381ae096
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:6cffef7631ee54612fba71efc0eff4e17bc08defc3e2137aee9881d685444298
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:722944bfd0eba280255f4c198f65f562423a26e2c4ad3580591ee71cb6e9c2ef
        - registry.marketplace.redhat.com/cp/cpd/job-trigger@sha256:f2ea487072393a24e4c6864b2107c1fce072ebba07f9dbc2c139bd37fbd062da
        - registry.marketplace.redhat.com/cp/cpd/environments-init-job@sha256:6af8f900af7c11eee6894e2c3e52f740b3c700dac0ee753fadef330230bd42dd
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:6fe90f051d9261ad224dccf4d9f6bc9f19ce0e8d22fa010a46b2752374bce2b3
        version: 10.3.0
      entries:
      - name: ibm-cpd-ccs.v10.3.0
        version: 10.3.0
      name: v10.3
    defaultChannel: v10.3
    packageName: ibm-cpd-ccs
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: swift-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: swift-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "swift.openstack.org/v1beta1",
                "kind": "Swift",
                "metadata": {
                  "name": "swift"
                },
                "spec": {
                  "swiftProxy": {
                    "passwordSelectors": {
                      "service": "SwiftPassword"
                    },
                    "replicas": 1
                  },
                  "swiftRing": {
                    "ringReplicas": 1
                  },
                  "swiftStorage": {
                    "replicas": 1,
                    "storageClass": "local-storage"
                  }
                }
              },
              {
                "apiVersion": "swift.openstack.org/v1beta1",
                "kind": "SwiftProxy",
                "metadata": {
                  "name": "swift-proxy"
                },
                "spec": {
                  "passwordSelectors": {
                    "service": "SwiftPassword"
                  },
                  "replicas": 1,
                  "secret": "osp-secret",
                  "serviceUser": "swift"
                }
              },
              {
                "apiVersion": "swift.openstack.org/v1beta1",
                "kind": "SwiftRing",
                "metadata": {
                  "name": "swift-ring"
                },
                "spec": {
                  "ringReplicas": 1
                }
              },
              {
                "apiVersion": "swift.openstack.org/v1beta1",
                "kind": "SwiftStorage",
                "metadata": {
                  "name": "swift-storage"
                },
                "spec": {
                  "replicas": 1,
                  "storageClass": "local-storage",
                  "storageRequest": "10G"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:34Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SwiftProxy is the Schema for the swiftproxies API
            displayName: Swift Proxy
            kind: SwiftProxy
            name: swiftproxies.swift.openstack.org
            version: v1beta1
          - description: SwiftRing is the Schema for the swiftrings API
            displayName: Swift Ring
            kind: SwiftRing
            name: swiftrings.swift.openstack.org
            version: v1beta1
          - description: Swift is the Schema for the swifts API
            displayName: Swift
            kind: Swift
            name: swifts.swift.openstack.org
            version: v1beta1
          - description: SwiftStorage is the Schema for the swiftstorages API
            displayName: Swift Storage
            kind: SwiftStorage
            name: swiftstorages.swift.openstack.org
            version: v1beta1
        description: OpenStack Swift operator
        displayName: Swift operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Swift
        - Object Storage
        links:
        - name: Swift Operator
          url: https://github.com/openstack-k8s-operators/swift-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-swift-object-rhel9@sha256:656f23b25ca0bd80d7b77ba717a21455366dfaf2049df1bd6d6536925a18b94b
        - registry.redhat.io/rhoso/openstack-swift-proxy-server-rhel9@sha256:6e654db83d570a3f8feedb912d5d9dff35ecba4d53169481a3b4982d3a832798
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/swift-rhel9-operator@sha256:18207f02c049447836211b326f09ce660782c384cda92e592db256692e3c8e73
        - registry.redhat.io/rhoso/openstack-swift-account-rhel9@sha256:ecebf3eb430188392f7a7b8cda3d1c0c0f658bcbbaff531d99eba88a7e0c9889
        - registry.redhat.io/rhoso/openstack-swift-container-rhel9@sha256:782866f8e9661de50d6a74bc52262f4546d7e0055d0c4da83ba6b891a50c35ea
        version: 1.0.6
      entries:
      - name: swift-operator.v1.0.6
        version: 1.0.6
      - name: swift-operator.v1.0.5
        version: 1.0.5
      - name: swift-operator.v1.0.4
        version: 1.0.4
      - name: swift-operator.v1.0.3
        version: 1.0.3
      - name: swift-operator.v1.0.2
        version: 1.0.2
      - name: swift-operator.v1.0.1
        version: 1.0.1
      - name: swift-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: swift-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/component: kmm
      app.kubernetes.io/name: kmm
      app.kubernetes.io/part-of: kmm
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: kernel-module-management
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kernel-module-management.v2.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:1022205d712a2ed9d6e76968961c180d4e66a2e0fb991221b0b02ed94ed830d2
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.0.2'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel-module-management
        - out-of-tree
        - kernel
        - kmm
        - module
        - modules
        - oot
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:1022205d712a2ed9d6e76968961c180d4e66a2e0fb991221b0b02ed94ed830d2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62c44057a27abd41cf3daf60b8736da59a82dadb65fa6c32bef65d19bd49ea49
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:25f39f20f325f64676bf8ef14cd8a60cfbdc77116753a89d4bb667ea2a86172c
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:192e08b121a04ae69aea60c35199e8f5e3c42305e1ebd024a605a371de4ab472
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:94b4ddc45cb34d4ffe794a51b99c868cb709f90cc4eb8b28284d70c526a0a226
        version: 2.0.2
      entries:
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.0
    - currentCSV: kernel-module-management.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a85037778c53c3766f39efe4c4d2523d01ae7743e5c5693b05bd8b2cbb50b1d8
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.1.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a85037778c53c3766f39efe4c4d2523d01ae7743e5c5693b05bd8b2cbb50b1d8
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:611556d999ea97f9dbb78fec5963baa02de7a26626d7096ba63b177e3bcf9707
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:14e100e5853444114029116255ebfb7a54e5ed51503ba969740ad48b5ca80f70
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:1a9a19df17b464d1708b2905d03b8726e4480ca1f11aa97a15ff4c049a24ea4b
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:ea3e5fbf7eaf6bd29031e2876d7e24859779fb4d3d7ef27c26a1d310f137ad17
        version: 2.1.1
      entries:
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.1
    - currentCSV: kernel-module-management.v2.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a224708950f8defb259ef0da49011d361c5c0378404350640600c6e8fbdbd606
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.2.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:b1d98c0f14448ae8eb80821cbec236150c338674aae2a2b37f8d0ab0d5f1430c
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:53fddf0b226eb579c71c86112a8d9b6b1d90a1498d6f6985932b16b8be532f35
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:c3dc9ad4ebc68cd85a074b520903a9310e72a90a73bdf4aa7c42b7ed66b7bcba
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:ce1d68fd96f5003285f9fbc0c7abd8aa23830dbdd9b76629d38d4d2238e30032
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a224708950f8defb259ef0da49011d361c5c0378404350640600c6e8fbdbd606
        version: 2.2.1
      entries:
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.2
    - currentCSV: kernel-module-management.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:750dcc75f908cdda86932386fa1a28edaf65d7b5379ef6d0447586a064db906f
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.3.0'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:750dcc75f908cdda86932386fa1a28edaf65d7b5379ef6d0447586a064db906f
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:c5005b74402f680d11c9ea682a9a0d3f74b6ab0d01d25298d232f910f1a29c34
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:055c2d6e80bd1ec225f0cb2041572ebb28cfd1a4ed82ba47ff0d638d52ded2a8
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:f9fde334af68220943017a7491130e6a82f15e63522f5bb3931aa7d9b00ccc40
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:06ac089cc89b303e525fd71d6439b958f34ab944a72209f1cc8cb562ecfd1112
        version: 2.3.0
      entries:
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.3
    - currentCSV: kernel-module-management.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:41Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:8e0a3cfb1cb92d6d3c9ab5a3030593d8df1378c3ce755d598fb77b08ea5793d3
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        version: 2.4.1
      entries:
      - name: kernel-module-management.v2.4.1
        version: 2.4.1
      - name: kernel-module-management.v2.4.0
        version: 2.4.0
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.4
    - currentCSV: kernel-module-management.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:41Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:8e0a3cfb1cb92d6d3c9ab5a3030593d8df1378c3ce755d598fb77b08ea5793d3
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        version: 2.4.1
      entries:
      - name: kernel-module-management.v2.4.1
        version: 2.4.1
      - name: kernel-module-management.v2.4.0
        version: 2.4.0
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: kernel-module-management
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio Project
      provider-url: ""
    name: apicurito
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: 1.0.x
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: candidate
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: latest
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: latest
    packageName: apicurito
    provider:
      name: Apicurio Project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: mongodb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"opstreelabs.in/v1alpha1","kind":"MongoDB","metadata":{"name":"mongodb"},"spec":{"kubernetesConfig":{"image":"quay.io/opstree/mongo:v5.0.6","imagePullPolicy":"IfNotPresent","securityContext":{"fsGroup":1001}},"storage":{"accessModes":["ReadWriteOnce"],"storageSize":"1Gi","storageClass":"gp2"},"mongoDBSecurity":{"mongoDBAdminUser":"admin","secretRef":{"name":"mongodb-secret","key":"password"}}}},{"apiVersion":"opstreelabs.in/v1alpha1","kind":"MongoDBCluster","metadata":{"name":"mongodb"},"spec":{"clusterSize":3,"kubernetesConfig":{"image":"quay.io/opstree/mongo:v5.0.6","imagePullPolicy":"IfNotPresent","securityContext":{"fsGroup":1001}},"storage":{"accessModes":["ReadWriteOnce"],"storageSize":"1Gi","storageClass":"gp2"},"mongoDBSecurity":{"mongoDBAdminUser":"admin","secretRef":{"name":"mongodb-secret","key":"password"}}}}]'
          capabilities: Seamless Upgrades
          categories: Database,  OpenShift Optional, Storage
          certified: "false"
          containerImage: quay.io/opstree/mongodb-operator:v0.3.0
          createdAt: 05-04-2022
          description: A golang based operator to create and manage MongoDB standalone
            setup and cluster setup
          k8sMaxVersion: ""
          k8sMinVersion: 1.17.0
          repository: https://github.com/ot-container-kit/mongodb-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Mongo DB Cluster
            displayName: MongoDBCluster
            kind: MongoDBCluster
            name: mongodbclusters.opstreelabs.in
            version: v1alpha1
          - description: Mongo DB
            displayName: MongoDB
            kind: MongoDB
            name: mongodbs.opstreelabs.in
            version: v1alpha1
        description: |+
          MongoDB Operator is an operator created in Golang to create, update, and manage MongoDB standalone, replicated, and arbiter replicated setup on Kubernetes and Openshift clusters. This operator is capable of doing the setup for MongoDB with all the required best practices.

          ### Supported Features

          * MongoDB replicated cluster setup
          * MongoDB standalone setup
          * MongoDB replicated cluster failover and recovery
          * Monitoring support with MongoDB Exporter
          * Password based authentication for MongoDB
          * Kubernetes's resources for MongoDB standalone and cluster
          * Customizable configuration changes in MongoDB
          * Insightful Grafana dashboards

          ### Documentation

          **[Documentation](https://ot-mongodb-operator.netlify.app/)**
          ### Getting Help

          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.
          ### Contributing
          You can contribute to this project by:-

          * Raising Bugs and Feature Requests
          * Fixing issue by opening a Pull Request
          * Improving Documentation

          ### License

          MongoDB  Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/mongodb-operator/blob/master/LICENSE)


        displayName: MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - MongoDB
        - Opstree
        - Operatork
        - Cloud Native
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-mongodb-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: alpha
        minKubeVersion: 1.17.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/opstree/mongodb-operator:v0.3.0
        version: 0.3.0
      entries:
      - name: mongodb-operator.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: mongodb-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: https://konveyor.io
    name: move2kube-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: move2kube-operator.v0.3.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {
                  "deployment": {
                    "api": {
                      "cpu": "300m",
                      "imageTag": "",
                      "initContainer": false,
                      "memory": "1Gi",
                      "privilegedPods": false,
                      "replicas": 1
                    },
                    "authServer": {
                      "cpu": "300m",
                      "enable": false,
                      "memory": "1Gi",
                      "replicas": 1
                    },
                    "database": {
                      "cpu": "300m",
                      "enable": false,
                      "memory": "1Gi",
                      "replicas": 1,
                      "startEmpty": true
                    }
                  },
                  "ingress": {
                    "enable": true,
                    "host": "mydomain.com",
                    "preferRoute": false,
                    "tlsSecretName": ""
                  },
                  "persistentVolumeClaim": {
                    "createNew": true,
                    "enable": false,
                    "name": "",
                    "readWriteMany": true,
                    "storageClassName": "",
                    "storageSize": "1Gi"
                  },
                  "secret": {
                    "api": {
                      "configYAML": "",
                      "createNew": true,
                      "enable": false,
                      "name": ""
                    },
                    "authServer": {
                      "adminPassword": "password",
                      "adminUsername": "admin",
                      "createNew": true,
                      "databasePassword": "password",
                      "databaseUsername": "auth-server",
                      "name": "",
                      "realmJSON": "",
                      "standaloneHAXML": ""
                    }
                  },
                  "securityContextContraints": {
                    "enable": false
                  },
                  "serviceAccount": {
                    "createNew": true,
                    "enable": false,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.10
          createdAt: "2023-09-25T02:56:15Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/konveyor/move2kube-operator:v0.3.10
        version: 0.3.10
      entries:
      - name: move2kube-operator.v0.3.10
        version: 0.3.10
      - name: move2kube-operator.v0.3.10-rc.0
        version: 0.3.10-rc.0
      - name: move2kube-operator.v0.3.9
        version: 0.3.9
      - name: move2kube-operator.v0.3.8
        version: 0.3.8
      name: alpha
    - currentCSV: move2kube-operator.v0.3.15-rc.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.15-rc.0
          createdAt: "2024-10-18T16:10:56Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/konveyor/move2kube-operator:v0.3.15-rc.0
        version: 0.3.15-rc.0
      entries:
      - name: move2kube-operator.v0.3.15-rc.0
        version: 0.3.15-rc.0
      - name: move2kube-operator.v0.3.14-rc.0
        version: 0.3.14-rc.0
      - name: move2kube-operator.v0.3.13-rc.0
        version: 0.3.13-rc.0
      - name: move2kube-operator.v0.3.11-rc.0
        version: 0.3.11-rc.0
      name: prerelease
    - currentCSV: move2kube-operator.v0.3.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.15
          createdAt: "2025-03-07T06:09:48Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/konveyor/move2kube-operator:v0.3.15
        version: 0.3.15
      entries:
      - name: move2kube-operator.v0.3.15
        version: 0.3.15
      - name: move2kube-operator.v0.3.14
        version: 0.3.14
      - name: move2kube-operator.v0.3.13
        version: 0.3.13
      - name: move2kube-operator.v0.3.11
        version: 0.3.11
      name: stable
    defaultChannel: stable
    packageName: move2kube-operator
    provider:
      name: Konveyor
      url: https://konveyor.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Dell Technologies
      provider-url: https://github.com/dell/csm-operator
    name: dell-csm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dell-csm-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "authorization",
                  "namespace": "authorization"
                },
                "spec": {
                  "modules": [
                    {
                      "components": [
                        {
                          "enabled": true,
                          "name": "nginx"
                        },
                        {
                          "enabled": true,
                          "name": "cert-manager"
                        },
                        {
                          "authorizationController": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231",
                          "authorizationControllerReplicas": 1,
                          "certificate": "",
                          "controllerReconcileInterval": "5m",
                          "enabled": true,
                          "hostname": "csm-authorization.com",
                          "leaderElection": true,
                          "name": "proxy-server",
                          "opa": "docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095",
                          "opaKubeMgmt": "docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517",
                          "openTelemetryCollectorAddress": "",
                          "privateKey": "",
                          "proxyServerIngress": [
                            {
                              "annotations": {},
                              "hosts": [],
                              "ingressClassName": "nginx"
                            }
                          ],
                          "proxyService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682",
                          "proxyServiceReplicas": 1,
                          "roleService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c",
                          "roleServiceReplicas": 1,
                          "storageService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5",
                          "storageServiceReplicas": 1,
                          "tenantService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358",
                          "tenantServiceReplicas": 1
                        },
                        {
                          "commander": "docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a",
                          "name": "redis",
                          "redis": "docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f",
                          "redisCommander": "rediscommander",
                          "redisName": "redis-csm",
                          "redisReplicas": 5,
                          "redisSecretProviderClass": [
                            {
                              "redisPasswordKey": "",
                              "redisSecretName": "",
                              "redisUsernameKey": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "sentinel": "sentinel"
                        },
                        {
                          "configSecretProviderClass": [
                            {
                              "configSecretName": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "name": "config"
                        },
                        {
                          "name": "storage-system-credentials",
                          "secretProviderClasses": {
                            "vault": [
                              "secret-provider-class-1",
                              "secret-provider-class-2"
                            ]
                          }
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": true,
                      "forceRemoveModule": true,
                      "name": "authorization-proxy-server"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "isilon",
                  "namespace": "isilon"
                },
                "spec": {
                  "driver": {
                    "authSecret": "isilon-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VERBOSE",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_ISI_PORT",
                          "value": "8080"
                        },
                        {
                          "name": "X_CSI_ISI_PATH",
                          "value": "/ifs/data/csi"
                        },
                        {
                          "name": "X_CSI_ISI_NO_PROBE_ON_START",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_AUTOPROBE",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_AUTH_TYPE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_CUSTOM_TOPOLOGY_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOISILON_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "AZ_RECONCILE_INTERVAL",
                          "value": "1h"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_ISI_QUOTA_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_ACCESS_ZONE",
                          "value": "System"
                        },
                        {
                          "name": "X_CSI_ISI_VOLUME_PATH_PERMISSIONS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_IGNORE_UNRESOLVABLE_HOSTS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "isilon",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerscale"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSCALE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_PERFORMANCE_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSCALE_QUOTA_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "ISICLIENT_INSECURE",
                              "value": "true"
                            },
                            {
                              "name": "ISICLIENT_AUTH_TYPE",
                              "value": "1"
                            },
                            {
                              "name": "ISICLIENT_VERBOSE",
                              "value": "0"
                            },
                            {
                              "name": "POWERSCALE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSCALE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0",
                          "name": "metrics-powerscale"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-isilon/csi_sock",
                            "--mode=node",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powermax",
                  "namespace": "powermax"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powermax-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_MANAGED_ARRAYS",
                          "value": "000000000001,000000000002"
                        },
                        {
                          "name": "X_CSI_POWERMAX_PORTGROUPS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_TRANSPORT_PROTOCOL",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_K8S_CLUSTER_PREFIX",
                          "value": "CSM"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "X_CSI_VSPHERE_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VSPHERE_PORTGROUP",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VSPHERE_HOSTNAME",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VCENTER_HOST",
                          "value": ""
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "X_CSI_POWERMAX_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "X_CSI_POWERMAX_DEBUG",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_REVPROXY_USE_SECRET",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powermax",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERMAX_ISCSI_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_TOPOLOGY_CONTROL_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": [
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/memory-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/disk-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/network-unavailable",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REVPROXY_TLS_SECRET",
                              "value": "csirevproxy-tls-secret"
                            },
                            {
                              "name": "X_CSI_REVPROXY_PORT",
                              "value": "2222"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            },
                            {
                              "name": "DeployAsSidecar",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265",
                          "name": "csipowermax-reverseproxy"
                        }
                      ],
                      "configVersion": "v2.14.0",
                      "name": "csireverseproxy"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powermax"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERMAX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERMAX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414",
                          "name": "metrics-powermax"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/powermax.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powerstore",
                  "namespace": "powerstore"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powerstore-config",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_NODE_NAME_PREFIX",
                          "value": "csi-node"
                        },
                        {
                          "name": "X_CSI_FC_PORTS_FILTER_FILE_PATH",
                          "value": "/etc/fc-ports-filter"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOPOWERSTORE_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSM_AUTH_ENABLED",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_MULTI_NAS_FAILURE_THRESHOLD",
                          "value": "5"
                        },
                        {
                          "name": "X_CSI_MULTI_NAS_COOLDOWN_PERIOD",
                          "value": "5m"
                        },
                        {
                          "name": "X_CSI_NFS_ACLS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerstore",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-powerstore.dellemc.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerstore"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSTORE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_SPACE_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_ARRAY_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_FILE_SYSTEM_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSTORE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSTORE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "ZIPKIN_URI",
                              "value": ""
                            },
                            {
                              "name": "ZIPKIN_SERVICE_NAME",
                              "value": "metrics-powerstore"
                            },
                            {
                              "name": "ZIPKIN_PROBABILITY",
                              "value": "0.0"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60",
                          "name": "metrics-powerstore"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "unity",
                  "namespace": "unity"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_UNITY_ALLOW_MULTI_POD_ACCESS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_EPHEMERAL_STAGING_PATH",
                          "value": "/var/lib/kubelet/plugins/kubernetes.io/csi/pv/"
                        },
                        {
                          "name": "X_CSI_ISCSI_CHROOT",
                          "value": "/noderoot"
                        },
                        {
                          "name": "X_CSI_UNITY_SYNC_NODEINFO_INTERVAL",
                          "value": "15"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOUNITY_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOUNITY_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "TENANT_NAME",
                          "value": ""
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_UNITY_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "unity",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "vxflexos",
                  "namespace": "vxflexos"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLELISTVOLUMESNAPSHOT",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLESNAPSHOTCGDELETE",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "INFO"
                        },
                        {
                          "name": "GOSCALEIO_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOSCALEIO_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_QUOTA_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "INTERFACE_NAMES",
                          "value": null
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERFLEX_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "File",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerflex",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "initContainers": [
                      {
                        "envs": [
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          },
                          {
                            "name": "REPO_ADDRESS",
                            "value": "sftp://0.0.0.0"
                          },
                          {
                            "name": "REPO_USER",
                            "value": "sdcRepoUser"
                          },
                          {
                            "name": "MODULE_SIGCHECK",
                            "value": "0"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "imagePullPolicy": "IfNotPresent",
                        "name": "sdc"
                      }
                    ],
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_SDC_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_APPROVE_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_PREFIX",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_SDC_SFTP_REPO_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 1,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "enabled": false,
                        "envs": [
                          {
                            "name": "HOST_PID",
                            "value": "1"
                          },
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "name": "sdc-monitor"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "csi-external-health-monitor-controller"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERFLEX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_SDC_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_SDC_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERFLEX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d",
                          "name": "metrics-powerflex"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerflex"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--csisock=unix:/var/lib/kubelet/plugins/vxflexos.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
          createdAt: "2025-09-05T13:18:42Z"
          description: Easily install and manage Dell’s CSI Drivers and CSM
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/dell/csm-operator
          support: Dell Technologies
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ContainerStorageModule is the Schema for the containerstoragemodules
              API
            displayName: Container Storage Module
            kind: ContainerStorageModule
            name: containerstoragemodules.storage.dell.com
            version: v1
        description: "Dell Container Storage Modules (CSM) Operator is a Kubernetes
          Operator which can be \nused to install and manage Dell’s CSI drivers and
          CSM modules. \nBy using Dell CSM Operator, enterprises can quickly and easily
          deploy the \nCSM modules for CSI drivers making it easy for DevOps \nteams
          to build and optimize their Kubernetes workloads using enterprise grade
          storage \nwhile consistently aligning with the desired outcomes of the organization.\nThis
          operator is available as both community and certified operator for Openshift
          \nand as a community operator for upstream Kubernetes.\nOperator can be
          deployed via OperatorHub.io. using Operator Lifecycle Manager (OLM)\n\n###
          About Dell CSM\n\nDell CSM brings powerful enterprise storage features and
          functionality to Kubernetes for \neasier adoption of cloud native workloads,
          improved productivity and scalable operations. \n\nCSM delivers software
          modules for storage management that provide developers access to \nbuild
          automation for enhanced IT needs and other critical enterprise storage features.
          \n\nCurrently, the following modules are available as part of Dell CSM.
          \n*\tObservability - Delivers a high-level view of the storage capacity
          \n  and performance usage via Grafana dashboards to the Kubernetes users\n*\tReplication
          - Enables array replication capabilities for Kubernetes users\n*\tAuthorization
          - Provides both storage and Kubernetes administrators the ability \n  to
          apply RBAC for Dell CSI Drivers\n*\tResiliency - Designed to make Kubernetes
          applications more resilient to \n  node failures\n\nThese modules are available
          for various storage platforms in the Dell portfolio, \nsuch as Dell PowerStore,
          Dell PowerMax, Dell PowerFlex, Dell PowerScale, and Dell Unity XT.\n\nFor
          more information, visit [Dell’s Container Storage Module web page](https://www.dell.com/en-us/dt/solutions/kubernetes-containers/kubernetes-storage.htm)\n\nAdditional
          Materials\n*\t[Documentation](https://dell.github.io/csm-docs)\n*\t[CSM
          Solution Brief](https://www.delltechnologies.com/asset/en-us/products/storage/briefs-summaries/csm-solution-brief.pdf)\n"
        displayName: Dell Container Storage Modules
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dell Container Storage Modules
        - Dell CSI Driver
        - Dell CSM Modules
        - Powerflex
        - Powerscale
        - Powerstore
        - Unity
        - Authorization
        - Observability
        - Replication
        links:
        - name: Documentation
          url: https://dell.github.io/csm-docs
        maintainers:
        - email: container.storage.modules@dell.com
          name: Dell Container Storage Modules
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Dell Technologies
          url: https://github.com/dell/csm-operator
        relatedImages:
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5
        - registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334
        - docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517
        - registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
        - registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:59335acaca515fceed5f0350c46cfcfd6037475f0cf35ae8ab41d37333039e2e
        - registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d
        - registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60
        - ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0
        - registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265
        - registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414
        - docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f
        - quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540
        - docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231
        - registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16
        - registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        version: 1.10.0
      entries:
      - name: dell-csm-operator.v1.10.0
        version: 1.10.0
      - name: dell-csm-operator.v1.9.1
        version: 1.9.1
      - name: dell-csm-operator.v1.9.0
        version: 1.9.0
      - name: dell-csm-operator.v1.8.1
        version: 1.8.1
      - name: dell-csm-operator.v1.8.0
        version: 1.8.0
      - name: dell-csm-operator.v1.7.0
        version: 1.7.0
      - name: dell-csm-operator.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: dell-csm-operator
    provider:
      name: Dell Technologies
      url: https://github.com/dell/csm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: raftt.io
      provider-url: https:///www.raftt.io
    name: raftt-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: raftt-operator.v1.25.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "raftt.io/v1beta1",
                "kind": "Raftt",
                "metadata": {
                  "name": "raftt",
                  "namespace": "raftt"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/raftt/raftt-operator@sha256:73edc6d8416ee3815d7e59bb26c33e3d35d97cf1ceadd8fa457041be13e175ca
          createdAt: "2025-04-29T11:31:02Z"
          description: This operator installs and configures a Raftt instance
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/raftt-operator/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/raftt-operator/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: raftt
          operators.openshift.io/valid-subscription: Raftt commercial license
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Raftt systems
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Raftt represents an installation of Raftt cluster controller
              & supporting services
            displayName: Raftt
            kind: Raftt
            name: raftts.raftt.io
            version: v1beta1
        description: |
          Raftt allows fast code iterations and debugging over complex Kubernetes environments.
          Raftt provides a localhost development experience for applications running on Kubernetes.
          You can use Raftt to orchestrate the environments for you, so you get the optimal development experience -
          iterations that take under a second, OOTB debugging of your code as it runs in the cluster, easy introspection, and more.
        displayName: Raftt
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dx
        - development
        - debugging
        links:
        - name: Raftt Operator
          url: https://raftt-operator.domain
        - name: Documentation
          url: https://docs.raftt.io
        maintainers:
        - email: ops@raftt.io
          name: Raftt
        maturity: alpha
        provider:
          name: raftt.io
          url: https:///www.raftt.io
        relatedImages:
        - quay.io/raftt/raftt-operator@sha256:73edc6d8416ee3815d7e59bb26c33e3d35d97cf1ceadd8fa457041be13e175ca
        version: 1.25.2
      entries:
      - name: raftt-operator.v1.25.2
        version: 1.25.2
      name: alpha
    defaultChannel: alpha
    packageName: raftt-operator
    provider:
      name: raftt.io
      url: https:///www.raftt.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HashiCorp
      provider-url: https://www.hashicorp.com/
    name: vault-secrets-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vault-secrets-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultConnection",
                "metadata": {
                  "name": "vaultconnection-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "address": "http://vault.vault.svc.cluster.local:8200"
                }
              },
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultAuth",
                "metadata": {
                  "name": "vaultauth-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "vaultConnectionRef": "vaultconnection-sample",
                  "method": "kubernetes",
                  "mount": "kubernetes",
                  "kubernetes": {
                    "role": "sample",
                    "serviceAccount": "default"
                  }
                }
              },
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultStaticSecret",
                "metadata": {
                  "name": "vaultstaticsecret-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "vaultAuthRef": "vaultauth-sample",
                  "mount": "kvv2",
                  "type": "kv-v2",
                  "path": "secret",
                  "refreshAfter": "5s",
                  "destination": {
                    "create": true,
                    "name": "secret1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.connect.redhat.com/hashicorp/vault-secrets-operator@sha256:84ac7003c4cb90c35f1e296585bb58187a86d64f972133e2661839bdb532ddb8
          createdAt: "2025-09-26T21:50:51Z"
          description: The Vault Secrets Operator (VSO) allows Pods to consume Vault
            secrets natively from Kubernetes Secrets.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/hashicorp/vault-secrets-operator
          support: HashiCorp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CSISecrets
            name: csisecrets.secrets.hashicorp.com
            version: v1beta1
          - description: HCPAuth is the Schema for the hcpauths API
            displayName: HCPAuth
            kind: HCPAuth
            name: hcpauths.secrets.hashicorp.com
            version: v1beta1
          - description: HCPVaultSecretsApp is the Schema for the hcpvaultsecretsapps
              API
            displayName: HCPVault Secrets App
            kind: HCPVaultSecretsApp
            name: hcpvaultsecretsapps.secrets.hashicorp.com
            version: v1beta1
          - description: SecretTransformation is the Schema for the secrettransformations
              API
            displayName: Secret Transformation
            kind: SecretTransformation
            name: secrettransformations.secrets.hashicorp.com
            version: v1beta1
          - description: VaultAuthGlobal is the Schema for the vaultauthglobals API
            displayName: Vault Auth Global
            kind: VaultAuthGlobal
            name: vaultauthglobals.secrets.hashicorp.com
            version: v1beta1
          - description: VaultAuth is the Schema for the vaultauths API
            displayName: Vault Auth
            kind: VaultAuth
            name: vaultauths.secrets.hashicorp.com
            version: v1beta1
          - description: VaultConnection is the Schema for the vaultconnections API
            displayName: Vault Connection
            kind: VaultConnection
            name: vaultconnections.secrets.hashicorp.com
            version: v1beta1
          - description: VaultDynamicSecret is the Schema for the vaultdynamicsecrets
              API
            displayName: Vault Dynamic Secret
            kind: VaultDynamicSecret
            name: vaultdynamicsecrets.secrets.hashicorp.com
            version: v1beta1
          - description: VaultPKISecret is the Schema for the vaultpkisecrets API
            displayName: Vault PKISecret
            kind: VaultPKISecret
            name: vaultpkisecrets.secrets.hashicorp.com
            version: v1beta1
          - description: VaultStaticSecret is the Schema for the vaultstaticsecrets
              API
            displayName: Vault Static Secret
            kind: VaultStaticSecret
            name: vaultstaticsecrets.secrets.hashicorp.com
            version: v1beta1
        description: |-
          The Vault Secrets Operator (VSO) allows Pods to consume Vault secrets
          natively from Kubernetes Secrets.

          ## Overview

          The Vault Secrets Operator operates by watching for changes to its supported set of Custom Resource Definitions (CRD).
          Each CRD provides the specification required to allow the *Operator* to synchronize a Vault Secrets to a Kubernetes Secret.
          The *Operator* writes the *source* Vault secret data directly to the *destination* Kubernetes Secret, ensuring that any
          changes made to the *source* are replicated to the *destination* over its lifetime. In this way, an application only needs
          to have access to the *destination* secret in order to make use of the secret data contained within.

          See the developer docs for more info [here](https://developer.hashicorp.com/vault/docs/platform/k8s/vso), including
          [examples](https://developer.hashicorp.com/vault/docs/platform/k8s/vso/examples) and a
          [tutorial](https://developer.hashicorp.com/vault/tutorials/kubernetes/vault-secrets-operator)
          for getting started.
        displayName: Vault Secrets Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vault
        - security
        - secrets
        links:
        - name: Vault Secrets Operator
          url: https://github.com/hashicorp/vault-secrets-operator
        - name: Documentation
          url: https://developer.hashicorp.com/vault/docs/platform/k8s/vso
        maintainers:
        - email: support@hashicorp.com
          name: HashiCorp Support
        maturity: stable
        provider:
          name: HashiCorp
          url: https://www.hashicorp.com/
        relatedImages:
        - registry.connect.redhat.com/hashicorp/vault-secrets-operator@sha256:84ac7003c4cb90c35f1e296585bb58187a86d64f972133e2661839bdb532ddb8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:503d5ee0fbcd5bd3492785f4246e19cbdb92960be0d05f3c6f123a32fe0fe84d
        version: 1.0.1
      entries:
      - name: vault-secrets-operator.v1.0.1
        version: 1.0.1
      - name: vault-secrets-operator.v1.0.0
        version: 1.0.0
      - name: vault-secrets-operator.v0.10.0
        version: 0.10.0
      - name: vault-secrets-operator.v0.9.1
        version: 0.9.1
      - name: vault-secrets-operator.v0.9.0
        version: 0.9.0
      - name: vault-secrets-operator.v0.8.1
        version: 0.8.1
      - name: vault-secrets-operator.v0.8.0
        version: 0.8.0
      - name: vault-secrets-operator.v0.7.0
        version: 0.7.0
      - name: vault-secrets-operator.v0.6.0
        version: 0.6.0
      - name: vault-secrets-operator.v0.5.2
        version: 0.5.2
      - name: vault-secrets-operator.v0.5.1
        version: 0.5.1
      - name: vault-secrets-operator.v0.5.0
        version: 0.5.0
      - name: vault-secrets-operator.v0.4.3
        version: 0.4.3
      name: stable
    defaultChannel: stable
    packageName: vault-secrets-operator
    provider:
      name: HashiCorp
      url: https://www.hashicorp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ab Initio
      provider-url: ""
    name: abinitio-runtime-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: abinitio-runtime-operator.v4.3.3-1
      currentCSVDesc:
        annotations:
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: abinitio-system
          operators.operatorframework.io/builder: operator-sdk-v1.8.0-ocp
          operators.operatorframework.io/internal-objects: '["cooperatingsystemruntimepoolmetrics.cloud.abinitio.com",
            "cooperatingsystemruntimeprofiles.cloud.abinitio.com"]'
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Defines a *runtime claim*, an interface for reserving a runtime
              object (CoOperatingSystemRuntime) in a runtime pool (CoOperatingSystemRuntimePool)
              in order to launch a job.
            displayName: Co>Operating System Runtime Claim
            kind: CoOperatingSystemRuntimeClaim
            name: cooperatingsystemruntimeclaims.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Pool Metric
            kind: CoOperatingSystemRuntimePoolMetric
            name: cooperatingsystemruntimepoolmetrics.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime pool*, a set of templated runtime objects
              (CoOperatingSystemRuntime) that are available on demand for running
              jobs. A runtime pool's capacity is configurable and can be scheduled
              in response to usage patterns.
            displayName: Co>Operating System Runtime Pool
            kind: CoOperatingSystemRuntimePool
            name: cooperatingsystemruntimepools.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Profile
            kind: CoOperatingSystemRuntimeProfile
            name: cooperatingsystemruntimeprofiles.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime object*, a set of coordinated resources
              used to run a data processing job in parallel across a launcher pod
              and a dynamically scaled set of worker pods. Such a job leverages the
              full power of the Co>Operating System for processing complex enterprise
              data.
            displayName: Co>Operating System Runtime
            kind: CoOperatingSystemRuntime
            name: cooperatingsystemruntimes.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime template*, a reusable specification (CoOperatingSystemRuntimeSpec)
              that can be referenced and overridden by runtime objects as needed,
              making it possible to specify jobs concisely. The runtime objects in
              a runtime pool (CoOperatingSystemRuntimePool) are created from a runtime
              template.
            displayName: Co>Operating System Runtime Template
            kind: CoOperatingSystemRuntimeTemplate
            name: cooperatingsystemruntimetemplates.cloud.abinitio.com
            version: v1
        description: |-
          The Co>Operating System runtime operator brings the power of Ab Initio's Co>Operating
          System to Kubernetes by defining a set of custom API resources for provisioning and
          running data processing jobs.

          The operator is installed along with the following custom resource definitions (CRDs):

          - CoOperatingSystemRuntime
          - CoOperatingSystemRuntimeClaim
          - CoOperatingSystemRuntimePool
          - CoOperatingSystemRuntimeTemplate

          In combination, these resources give you significant flexibility in scheduling, running,
          and monitoring your jobs, which can run in parallel across a dynamically scaled set of
          containers. For reference documentation for these resources, see
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-FOUR-Reference/Co-Operating-System-runtime-operator-API-reference)
          in the Ab Initio Help.

          The following topics describe the basic requirements and procedures for creating custom
          resources for running jobs after you have installed the runtime operator.

          ## Requirements for using custom resources

          The runtime operator is fully configured when you install it in your OpenShift cluster
          through OperatorHub. Importantly, the operator is installed in unrestricted mode,
          which means it can manage the CoOperatingSystemRuntime family of CRDs in any namespace
          in your cluster.

          For creating custom resources in order to run jobs using the operator, you must meet
          the requirements and configure the settings described in the following sections.

          ### Container images for job agents

          To run jobs of kind CoOperatingSystemRuntime, you must have a way to build the following
          OCI container images:

          - A *processing* image with the Co>Operating System server software installed in it.
          - A *sandbox workspace* image with the graphs and other files needed to run your jobs.
          This image is based on the Co>Operating System image and functions as the job's launcher
          agent, which coordinates with a set of worker agents (processing containers) in order to
          run a job in parallel.

          For building and managing these container images, you need the following:

          - A Co>Operating System installation package.
          - Access to a technical repository containing the projects that make up your application.
          - A container-building tool.
          - Access to a private container registry for hosting the images you build. You need pull
          and push access to this registry from your container build context and pull access from
          your OpenShift cluster (for example, via a Kubernetes Secret).

          > WARNING! Never push container images with Ab Initio software to a public registry. Push
          them to a private container registry only.

          The Co>Operating System includes reference Dockerfiles and example scripts for building
          container images. Typically, you build these images as part of your CI/CD pipeline. For
          more information, contact your Ab Initio administrator or Ab Initio Support.

          ### Ab Initio key servers

          The key servers that you use to enable jobs must be reachable from your cluster. You
          specify these key servers via the `AB_HOSTNAME_KEYSERVER_URLS` configuration variable in
          the pod specifications for launcher and worker agents in the CoOperatingSystemRuntime
          resource.

          To discuss the software activation key requirements for jobs that run in Kubernetes,
          contact your key server administrator or Ab Initio Support.

          ### Namespace and role considerations

          The runtime operator is installed with the cluster-wide RBAC roles and role bindings that
          it needs in order to manage custom resources and run jobs.

          You typically use a separate namespace or namespaces for creating and managing custom
          resources to run jobs. Each of these namespaces must be equipped with a service
          account that can list pod limits and get and patch CoOperatingSystemRuntime resources.

          For each namespace in which you intend to create custom resources, create the needed RBAC
          role and role binding by applying the following Role and RoleBinding manifests:

          ```yaml
          apiVersion: rbac.authorization.k8s.io/v1
          kind: Role
          metadata:
            name: cooperatingsystemruntime-role
          rules:
          - apiGroups:
            - ""
            resources:
            - limitranges
            verbs:
            - list
          - apiGroups:
            - cloud.abinitio.com
            resources:
            - cooperatingsystemruntimes
            verbs:
            - get
            - patch
          ---
          apiVersion: rbac.authorization.k8s.io/v1
          kind: RoleBinding
          metadata:
            name: run-cooperatingsystemruntime-rolebinding
          subjects:
          - kind: ServiceAccount
            name: default
          roleRef:
            kind: Role
            name: cooperatingsystemruntime-role
            apiGroup: rbac.authorization.k8s.io
          ```

          One way to do this is to save these manifests to a file and specify a namespace when
          using `oc` to apply them to the cluster. In the following example, the manifests are
          saved to a file named `runtime-rbac.yaml` and the role and role binding are to be created
          for the `default` service account in the `hello-world` namespace:

          ```
          oc -n hello-world apply -f runtime-rbac.yaml
          ```

          To bind the role to a service account other than `default`, simply update the value of
          `subjects[0].name` in the provided RoleBinding manifest.

          ### Dynamic volume provisioning

          The launcher and worker template sections of the CoOperatingSystemRuntime resource
          require you to specify persistent volume claims (PVCs) for job recovery data. To ensure
          that these PVCs can be bound to approriate storage volumes, dynamic volume provisioning
          must be enabled in your cluster.

          Work with your cluster administrator to identify the storage needs of your jobs and to
          ensure that dynamic volume provisioning is enabled so that PVCs can be matched to
          persistent volumes of the right storage class.

          For more information about dynamic storage provisioning, see the
          [OpenShift documentation](https://docs.openshift.com/container-platform/latest/storage/dynamic-provisioning.html).

          ## Quick start and examples

          For an overview of the key concepts associated with running jobs in Kubernetes, see
          ["About running jobs in Kubernetes"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-TWO-Running-jobs-in-Kubernetes/About-running-jobs-in-Kubernetes)
          in the Ab Initio Help.

          As described in this topic, you might begin by creating a CoOperatingSystemRuntime
          manifest that defines the configuration and compute resources needed for a standalone test
          job. You could then use the specification for this job to create a CoOperatingSystemRuntimeTemplate
          resource, which would collect common settings and make it possible to concisely specify
          CoOperatingSystemRuntime jobs.

          You could also use a runtime pool (of kind CoOperatingSystemRuntime) to extend these basic
          resources: such a pool consists of an actively running group of templated runtime objects whose
          pods can be used to run jobs on demand. Finally, for reserving resources in a pool, you could
          create a CoOperatingSystemRuntimeClaim resource.

          For more information about these custom resources, refer to the
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-FOUR-Reference/Co-Operating-System-runtime-operator-API-reference).

          For a set of integrated examples (including complete YAML manifests) that show you how to
          create and use the custom resources, see ["Example manifests for running jobs"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-TWO-Running-jobs-in-Kubernetes/Configuring-and-running-Co-Operating-System-jobs-in-Kubernetes/Example-manifests-for-CoOperatingSystemRuntime-jobs).
        displayName: Co>Operating System Runtime Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - abinitio
        - runtime
        - operator
        links:
        - name: Ab Initio
          url: https://www.abinitio.com/
        maintainers:
        - email: support@abinitio.com
          name: Ab Initio Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Ab Initio
        relatedImages:
        - registry.connect.redhat.com/abinitio/cooperating-system-runtime-operator@sha256:665efa8cc2ba97ad31f219161cdd53e15c1aef16b6c05ac60506be9bbc9ffc9a
        version: 4.3.3-1
      entries:
      - name: abinitio-runtime-operator.v4.3.3-1
        version: 4.3.3-1
      - name: abinitio-runtime-operator.v4.3.1-0
        version: 4.3.1-0
      name: release-4.3
    - currentCSV: abinitio-runtime-operator.v4.2.3-0
      currentCSVDesc:
        annotations:
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: abinitio-system
          operators.operatorframework.io/builder: operator-sdk-v1.8.0-ocp
          operators.operatorframework.io/internal-objects: '["cooperatingsystemruntimepoolmetrics.cloud.abinitio.com",
            "cooperatingsystemruntimeprofiles.cloud.abinitio.com"]'
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Defines a *runtime claim*, an interface for reserving a runtime
              object (CoOperatingSystemRuntime) in a runtime pool (CoOperatingSystemRuntimePool)
              in order to launch a job.
            displayName: Co>Operating System Runtime Claim
            kind: CoOperatingSystemRuntimeClaim
            name: cooperatingsystemruntimeclaims.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Pool Metric
            kind: CoOperatingSystemRuntimePoolMetric
            name: cooperatingsystemruntimepoolmetrics.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime pool*, a set of templated runtime objects
              (CoOperatingSystemRuntime) that are available on demand for running
              jobs. A runtime pool's capacity is configurable and can be scheduled
              in response to usage patterns.
            displayName: Co>Operating System Runtime Pool
            kind: CoOperatingSystemRuntimePool
            name: cooperatingsystemruntimepools.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Profile
            kind: CoOperatingSystemRuntimeProfile
            name: cooperatingsystemruntimeprofiles.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime object*, a set of coordinated resources
              used to run a data processing job in parallel across a launcher pod
              and a dynamically scaled set of worker pods. Such a job leverages the
              full power of the Co>Operating System for processing complex enterprise
              data.
            displayName: Co>Operating System Runtime
            kind: CoOperatingSystemRuntime
            name: cooperatingsystemruntimes.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime template*, a reusable specification (CoOperatingSystemRuntimeSpec)
              that can be referenced and overridden by runtime objects as needed,
              making it possible to specify jobs concisely. The runtime objects in
              a runtime pool (CoOperatingSystemRuntimePool) are created from a runtime
              template.
            displayName: Co>Operating System Runtime Template
            kind: CoOperatingSystemRuntimeTemplate
            name: cooperatingsystemruntimetemplates.cloud.abinitio.com
            version: v1
        description: |-
          The Co>Operating System runtime operator brings the power of Ab Initio's Co>Operating
          System to Kubernetes by defining a set of custom API resources for provisioning and
          running data processing jobs.

          The operator is installed along with the following custom resource definitions (CRDs):

          - CoOperatingSystemRuntime
          - CoOperatingSystemRuntimeClaim
          - CoOperatingSystemRuntimePool
          - CoOperatingSystemRuntimeTemplate

          In combination, these resources give you significant flexibility in scheduling, running,
          and monitoring your jobs, which can run in parallel across a dynamically scaled set of
          containers. For reference documentation for these resources, see
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Co_Operating_System_runtime_operator_API_reference)
          in the Ab Initio Help.

          The following topics describe the basic requirements and procedures for creating custom
          resources for running jobs after you have installed the runtime operator.

          ## Requirements for using custom resources

          The runtime operator is fully configured when you install it in your OpenShift cluster
          through OperatorHub. Importantly, the operator is installed in unrestricted mode,
          which means it can manage the CoOperatingSystemRuntime family of CRDs in any namespace
          in your cluster.

          For creating custom resources in order to run jobs using the operator, you must meet
          the requirements and configure the settings described in the following sections.

          ### Container images for job agents

          To run jobs of kind CoOperatingSystemRuntime, you must have a way to build the following
          OCI container images:

          - A *processing* image with the Co>Operating System server software installed in it.
          - A *sandbox workspace* image with the graphs and other files needed to run your jobs.
          This image is based on the Co>Operating System image and functions as the job's launcher
          agent, which coordinates with a set of worker agents (processing containers) in order to
          run a job in parallel.

          For building and managing these container images, you need the following:

          - A Co>Operating System installation package.
          - Access to a technical repository containing the projects that make up your application.
          - A container-building tool.
          - Access to a private container registry for hosting the images you build. You need pull
          and push access to this registry from your container build context and pull access from
          your OpenShift cluster (for example, via a Kubernetes Secret).

          > WARNING! Never push container images with Ab Initio software to a public registry. Push
          them to a private container registry only.

          The Co>Operating System includes reference Dockerfiles and example scripts for building
          container images. Typically, you build these images as part of your CI/CD pipeline. For
          more information, contact your Ab Initio administrator or Ab Initio Support.

          ### Ab Initio key servers

          The key servers that you use to enable jobs must be reachable from your cluster. You
          specify these key servers via the `AB_HOSTNAME_KEYSERVER_URLS` configuration variable in
          the pod specifications for launcher and worker agents in the CoOperatingSystemRuntime
          resource.

          To discuss the software activation key requirements for jobs that run in Kubernetes,
          contact your key server administrator or Ab Initio Support.

          ### Namespace and role considerations

          The runtime operator is installed with the cluster-wide RBAC roles and role bindings that
          it needs in order to manage custom resources and run jobs.

          You typically use a separate namespace or namespaces for creating and managing custom
          resources to run jobs. Each of these namespaces must be equipped with a service
          account that can list pod limits and get and patch CoOperatingSystemRuntime resources.

          For each namespace in which you intend to create custom resources, create the needed RBAC
          role and role binding by applying the following Role and RoleBinding manifests:

          ```yaml
          apiVersion: rbac.authorization.k8s.io/v1
          kind: Role
          metadata:
            name: cooperatingsystemruntime-role
          rules:
          - apiGroups:
            - ""
            resources:
            - limitranges
            verbs:
            - list
          - apiGroups:
            - cloud.abinitio.com
            resources:
            - cooperatingsystemruntimes
            verbs:
            - get
            - patch
          ---
          apiVersion: rbac.authorization.k8s.io/v1
          kind: RoleBinding
          metadata:
            name: run-cooperatingsystemruntime-rolebinding
          subjects:
          - kind: ServiceAccount
            name: default
          roleRef:
            kind: Role
            name: cooperatingsystemruntime-role
            apiGroup: rbac.authorization.k8s.io
          ```

          One way to do this is to save these manifests to a file and specify a namespace when
          using `oc` to apply them to the cluster. In the following example, the manifests are
          saved to a file named `runtime-rbac.yaml` and the role and role binding are to be created
          for the `default` service account in the `hello-world` namespace:

          ```
          oc -n hello-world apply -f runtime-rbac.yaml
          ```

          To bind the role to a service account other than `default`, simply update the value of
          `subjects[0].name` in the provided RoleBinding manifest.

          ### Dynamic volume provisioning

          The launcher and worker template sections of the CoOperatingSystemRuntime resource
          require you to specify persistent volume claims (PVCs) for job recovery data. To ensure
          that these PVCs can be bound to approriate storage volumes, dynamic volume provisioning
          must be enabled in your cluster.

          Work with your cluster administrator to identify the storage needs of your jobs and to
          ensure that dynamic volume provisioning is enabled so that PVCs can be matched to
          persistent volumes of the right storage class.

          For more information about dynamic storage provisioning, see the
          [OpenShift documentation](https://docs.openshift.com/container-platform/latest/storage/dynamic-provisioning.html).

          ## Quick start and examples

          For an overview of the key concepts associated with running jobs in Kubernetes, see
          ["About running jobs in Kubernetes"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Configuring_and_running_Co_Operating_System_jobs_in_Kubernetes/About_running_jobs_in_Kubernetes)
          in the Ab Initio Help.

          As described in this topic, you might begin by creating a CoOperatingSystemRuntime
          manifest that defines the configuration and compute resources needed for a standalone test
          job. You could then use the specification for this job to create a CoOperatingSystemRuntimeTemplate
          resource, which would collect common settings and make it possible to concisely specify
          CoOperatingSystemRuntime jobs.

          You could also use a runtime pool (of kind CoOperatingSystemRuntime) to extend these basic
          resources: such a pool consists of an actively running group of templated runtime objects whose
          pods can be used to run jobs on demand. Finally, for reserving resources in a pool, you could
          create a CoOperatingSystemRuntimeClaim resource.

          For more information about these custom resources, refer to the
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Co_Operating_System_runtime_operator_API_reference).

          For a set of integrated examples (including complete YAML manifests) that show you how to
          create and use the custom resources, see ["Example manifests for running jobs"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Configuring_and_running_Co_Operating_System_jobs_in_Kubernetes/Creating_custom_resources_for_running_jobs/Example_manifests_for_running_jobs).
        displayName: Co>Operating System Runtime Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - abinitio
        - runtime
        - operator
        links:
        - name: Ab Initio
          url: https://www.abinitio.com/
        maintainers:
        - email: support@abinitio.com
          name: Ab Initio Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Ab Initio
        relatedImages:
        - registry.connect.redhat.com/abinitio/cooperating-system-runtime-operator@sha256:ee30210a153e8b3ba09080737216d881cc6a07e1a888e9684be0d17fd4ff791d
        version: 4.2.3-0
      entries:
      - name: abinitio-runtime-operator.v4.2.3-0
        version: 4.2.3-0
      - name: abinitio-runtime-operator.v4.2.2
        version: 4.2.2
      name: stable
    defaultChannel: release-4.3
    packageName: abinitio-runtime-operator
    provider:
      name: Ab Initio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: recipe
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: recipe.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "Recipe",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "recipe",
                    "app.kubernetes.io/instance": "recipe-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "recipe",
                    "app.kubernetes.io/part-of": "recipe"
                  },
                  "name": "recipe-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          createdAt: "2025-09-12T06:34:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Recipe is the Schema for the recipes API
            displayName: Recipe
            kind: Recipe
            name: recipes.ramendr.openshift.io
            version: v1alpha1
        description: |
          Recipe describes a workflow used for capturing or recovering Kubernetes resources. This can be referred to by a Ramen DRPlacementControl object when kubeObjectProtection is enabled.
        displayName: Recipe
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - recipe
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/recipe
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:af262b0890300183db1766bc392017b0bda723a9ba841dc25e3d928aa413db9a
        version: 4.16.16-rhodf
      entries:
      - name: recipe.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: recipe.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: recipe.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: recipe.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: recipe.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: recipe.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: recipe.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: recipe.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: recipe.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: recipe.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: recipe.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: recipe.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: recipe.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: recipe.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: recipe.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: recipe.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: recipe.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: recipe
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: New England Research Cloud
      provider-url: https://nerc.mghpcc.org/
    name: keycloak-permissions-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keycloak-permissions-operator.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.nerc.mghpcc.org/v1",
                "kind": "KeycloakAuthorization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keycloak-permissions-operator",
                    "app.kubernetes.io/instance": "keycloakauthorization-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "keycloakauthorization",
                    "app.kubernetes.io/part-of": "keycloak-permissions-operator"
                  },
                  "name": "example"
                },
                "spec": {
                  "cleanupUndefinedAuthorizations": false,
                  "description": "",
                  "keycloak": {
                    "adminSecret": {
                      "name": "keycloak-initial-admin",
                      "passwordKey": "password",
                      "usernameKey": "username"
                    }
                  },
                  "managementState": "Managed",
                  "realm": null,
                  "validateCerts": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          com.redhat.openshift.versions: v4.11
          containerImage: quay.io/nerc-images/keycloak-permissions-operator:1.4.1
          createdAt: "2024-05-07T21:09:02Z"
          description: Managing Keycloak resources, scopes, policies, and permissions
            for fine-grained resource permissions
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/nerc-images/keycloak-permissions-operator
          support: New England Research Cloud
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Keycloak resources, scopes, policies, and permissions
              for fine-grained resource permissions.
            displayName: Keycloak Authorization
            kind: KeycloakAuthorization
            name: keycloakauthorizations.keycloak.nerc.mghpcc.org
            version: v1
        description: "An OpenShift Operator for managing Keycloak resources, scopes,
          policies,\nand permissions for fine-grained resource permissions.\nAccording
          to the [Keycloak Authorization Services documentation](https://www.keycloak.org/docs/23.0.7/authorization_services/):\n\n-
          A resource can be a web page, a RESTFul resource, a file in your file system,
          or other entity you wish to authorize access to.\n- Scopes usually represent
          the actions that can be performed on a resource, but they are not limited
          to that. You can also use scopes to represent one or more attributes within
          a resource.\n- Policies define the conditions that must be satisfied to
          access or perform operations on something (resource or scope), but they
          are not tied to what they are protecting. They are generic and can be reused
          to build permissions or even more complex policies.\n- Permissions are coupled
          with the resource they are protecting. Here you specify what you want to
          protect (resource or scope) and the policies that must be satisfied to grant
          or deny permission.\n\n## Getting started with the Keycloak Permissions
          Operator\n\nStart by defining the name of your KeycloakAuthorization and
          the same namespace where Keycloak is deployed.\n\n```yaml\napiVersion: keycloak.nerc.mghpcc.org/v1\nkind:
          KeycloakAuthorization\nmetadata:\n  name: example\n  namespace: keycloak\n```\n\n##
          KeycloakAuthorization spec\n\nNext, define the spec. \n\n- `validateCerts:
          false` If your Keycloak does not have valid certificates, if you are deploying
          to OpenShift Local for example. You don't need this line if Keycloak has
          valid certificates.\n- `cleanupUndefinedAuthorizations: true` if you would
          like to remove any undefined authorizations that may have been added manually.
          This is useful for keeping in sync with GitOps. If you do not want the operator
          removing any resources, policies, or permissions, then remove this line.\n\n```yaml\nspec:\n
          \ validateCerts: false\n  cleanupUndefinedAuthorizations: true\n```\n\n##
          Keycloak admin credentials\n\nSpecify the details to connect to keycloak,
          including the `baseUrl`, and the Secret name containing the Keycloak admin
          usernameKey and passwordKey.\n\n```yaml\n  keycloak:\n    baseUrl: https://keycloak-service:8443\n
          \   adminSecret:\n      name: keycloak-initial-admin\n      usernameKey:
          username\n      passwordKey: password\n```\n\n## Realm\n\nSpecify the realm
          id you wish to add authorization configuration to. If you want to add authorization
          configuration to multiple realms, then create a KeycloakAuthorization resource
          for each realm.\n\n```yaml\n  realm:\n    id: NERC\n```\n\n## Clients\n\nFor
          each client in the realm you wish to configure, add authorizationSettings.\n\n```yaml\n
          \   clients:\n      - id: ai4cloudops\n        authorizationSettings:\n```\n\n##
          Authorization Scopes\n\nList the scopes you wish to define for the client.\n\n```yaml\n
          \         scopes:\n            - id: prod\n              name: prod\n              displayName:
          Production\n            - id: test\n              name: test\n              displayName:
          Test\n```\n\n## Authorization Resources\n\nList the resources and corresponding
          scopes you wish to define for the client.\n\n```yaml\n          resources:\n
          \           - name: cluster\n              displayName: cluster\n              scopes:\n
          \               - id: prod\n                  name: prod\n                  displayName:
          Production\n                - id: test\n                  name: test\n                  displayName:
          Test\n```\n\n## Authorization Policies\n\nDefine the policies you wish to
          use in your permissions. You can choose from all the different types of
          permissions, including: aggregated, client, client-scope, group, regex,
          role, time, and user.\n\n```yaml\n          policies:\n            - id:
          client-example\n              name: client-example\n              logic:
          POSITIVE\n              decisionStrategy: UNANIMOUS\n              type:\n
          \               client:\n                  clients:\n                    -
          example\n            - id: group-example\n              name: group-example\n
          \             logic: POSITIVE\n              decisionStrategy: UNANIMOUS\n
          \             type:\n                group:\n                  groups:\n
          \                   - id: example\n```\n\n## Authorization Permissions\n\nDefine
          the permissions, connecting the resources, scopes, and policies together.\n\n```yaml\n
          \         permissions:\n            - name: group-example-cluster-test\n
          \             client: test\n              policy: group-example\n              resource:
          cluster\n            - name: client-example-cluster-test\n              client:
          example\n              policy: client-test\n              resource: cluster\n
          \             scope: test\n```\n\n## Other details\n\nResources, scopes,
          policies, and permissions will all be applied to Keycloak if you have configured
          the admin credentials to Keycloak successfully.\nWhen the KeycloakAuthorization
          is deleted, then the configured resources, policies, and permissions will
          be deleted as well. "
        displayName: Keycloak Permissions Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - fine-grained resource permissions
        - keycloak
        - resources
        - scopes
        - policies
        - permissions
        links:
        - name: Keycloak Permissions Operator
          url: https://keycloak-permissions-operator.domain
        maintainers:
        - email: computate@computate.org
          name: Christopher Tate
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: New England Research Cloud
          url: https://nerc.mghpcc.org/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/nerc-images/keycloak-permissions-operator:1.4.1
        version: 1.4.1
      entries:
      - name: keycloak-permissions-operator.v1.4.1
        version: 1.4.1
      - name: keycloak-permissions-operator.v1.3.0
        version: 1.3.0
      - name: keycloak-permissions-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: keycloak-permissions-operator
    provider:
      name: New England Research Cloud
      url: https://nerc.mghpcc.org/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: service-registry-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-registry-operator.v2.0.7-0.1662549961.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-kafkasql"
                },
                "spec": {
                  "configuration": {
                    "kafkasql": {
                      "bootstrapServers": "\u003cservice name\u003e.\u003cnamespace\u003e.svc:9092"
                    },
                    "persistence": "kafkasql"
                  }
                }
              },
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-sql"
                },
                "spec": {
                  "configuration": {
                    "persistence": "sql",
                    "sql": {
                      "dataSource": {
                        "password": "\u003cpassword\u003e",
                        "url": "jdbc:postgresql://\u003cservice name\u003e.\u003cnamespace\u003e.svc:5432/\u003cdatabase name\u003e",
                        "userName": "postgres"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/integration/service-registry-rhel8-operator@sha256:6dfd539951a739ec06fb6620b79c42c6159d4ee148c0d9d336547e786cd4780c
          createdAt: "2022-06-24"
          description: Deploy and manage Red Hat Integration service registry on OpenShift.
          olm.substitutesFor: service-registry-operator.v2.0.7-0.1661784420.p
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents a Red Hat Integration Service
              Registry instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Service Registry
          Service Registry stores and retrieves API designs and event schemas, and gives you control of their evolution.
          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf - Provides a REST API and web UI to manage the artifacts and artifact meta-data - Includes Serializers and Deserializers for Kafka client integration - Configurable rules to control schema validity and evolution (compatibility) - Storage options: Kafka, PostgreSQL - Compatible with Confluent and IBM APIs - Runs on a lightweight Quarkus platform - Includes Maven plugin to integrate with Maven based builds
          ## Service Registry Operator
          Provides a quick and easy way to deploy and manage service registry on OpenShift.
          **Features**
          - Supports installation and configuration of the Registry for all storage options - Easily perform a rolling upgrade of the Registry
          ## Prerequisites
          This operator does not deploy storage for the Registry. Therefore, some storage options require that the chosen persistence service is already set up.
          ## License
          Red Hat Service Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - Service Registry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - service
        - service-registry
        links:
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/integration/service-registry-rhel8-operator@sha256:6dfd539951a739ec06fb6620b79c42c6159d4ee148c0d9d336547e786cd4780c
        - registry.redhat.io/integration/service-registry-kafkasql-rhel8@sha256:fc97a2a08b2f77d620ea86ec77bb4f750bfc0109ede3ad1080c2cb694019f8b8
        - registry.redhat.io/integration/service-registry-sql-rhel8@sha256:6c97fa8ead1f6f0323811f8f35a80ae33f0e160f6c58f2bdb31713dbeaa3f014
        version: 2.0.7+0.1662549961.p
      entries:
      - name: service-registry-operator.v2.0.7-0.1662549961.p
        version: 2.0.7+0.1662549961.p
      - name: service-registry-operator.v2.0.7-0.1661784420.p
        version: 2.0.7+0.1661784420.p
      - name: service-registry-operator.v2.0.7
        version: 2.0.7
      - name: service-registry-operator.v2.0.6
        version: 2.0.6
      - name: service-registry-operator.v2.0.5-0.1654503709.p
        version: 2.0.5+0.1654503709.p
      - name: service-registry-operator.v2.0.5-0.1652887585.p
        version: 2.0.5+0.1652887585.p
      - name: service-registry-operator.v2.0.5-0.1646309885.p
        version: 2.0.5+0.1646309885.p
      - name: service-registry-operator.v2.0.5-0.1647938417.p
        version: 2.0.5+0.1647938417.p
      - name: service-registry-operator.v2.0.5-0.1648813064.p
        version: 2.0.5+0.1648813064.p
      - name: service-registry-operator.v2.0.5-0.1651047781.p
        version: 2.0.5+0.1651047781.p
      - name: service-registry-operator.v2.0.5-0.1652429755.p
        version: 2.0.5+0.1652429755.p
      - name: service-registry-operator.v2.0.5
        version: 2.0.5
      - name: service-registry-operator.v2.0.5-0.1655396591.p
        version: 2.0.5+0.1655396591.p
      - name: service-registry-operator.v2.0.4
        version: 2.0.4
      - name: service-registry-operator.v2.0.3
        version: 2.0.3
      - name: service-registry-operator.v2.0.2
        version: 2.0.2
      - name: service-registry-operator.v2.0.1
        version: 2.0.1
      - name: service-registry-operator.v2.0.0
        version: 2.0.0
      name: 2.0.x
    - currentCSV: service-registry-operator.v2.6.13-r2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-kafkasql"
                },
                "spec": {
                  "configuration": {
                    "kafkasql": {
                      "bootstrapServers": "\u003cservice name\u003e.\u003cnamespace\u003e.svc:9092"
                    },
                    "persistence": "kafkasql"
                  }
                }
              },
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-sql"
                },
                "spec": {
                  "configuration": {
                    "persistence": "sql",
                    "sql": {
                      "dataSource": {
                        "password": "\u003cpassword\u003e",
                        "url": "jdbc:postgresql://\u003cservice name\u003e.\u003cnamespace\u003e.svc:5432/\u003cdatabase name\u003e",
                        "userName": "postgres"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/apicurio/apicurio-registry-rhel8-operator@sha256:61f00395b51819312dda7abcdaa47190a11a83b1e2b523101be0ba776b1c5b7c
          createdAt: "2025-09-15"
          description: Deploy and manage Red Hat build of Apicurio Registry on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations",
            "Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents a Red Hat build of Apicurio Registry
              instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Red Hat build of Apicurio Registry

          Red Hat build of Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Provides a REST API and web UI to manage the artifacts and artifact meta-data
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL
          - Compatible with Confluent and IBM APIs
          - Runs on a lightweight Quarkus platform
          - Includes Maven plugin to integrate with Maven based builds

          ## Red Hat build of Apicurio Registry Operator

          Provides a quick and easy way to deploy and manage Red Hat build of Apicurio Registry on OpenShift.

          **Features**
          - Supports installation and configuration of the Registry for all storage options
          - Easily perform a rolling upgrade of Red Hat build of Apicurio Registry

          ## Prerequisites

          This operator does not deploy storage for Red Hat build of Apicurio Registry. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Red Hat build of Apicurio Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/main/LICENSE)
        displayName: Red Hat build of Apicurio Registry
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - service
        - service-registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/apicurio/apicurio-registry-rhel8-operator@sha256:61f00395b51819312dda7abcdaa47190a11a83b1e2b523101be0ba776b1c5b7c
        - registry.redhat.io/apicurio/apicurio-registry-kafkasql-rhel8@sha256:f2da7cd92e126187d91043e2ae40521f8574586f8b3975be27d8370fe31b6732
        - registry.redhat.io/apicurio/apicurio-registry-sql-rhel8@sha256:a55425a9af1a16959ab5e528a5740d47ea06f6935593cc925e7727cded4e7d13
        version: 2.6.13-r2
      entries:
      - name: service-registry-operator.v2.6.13-r2
        version: 2.6.13-r2
      - name: service-registry-operator.v2.6.13-r1
        version: 2.6.13-r1
      - name: service-registry-operator.v2.6.12-r1
        version: 2.6.12-r1
      - name: service-registry-operator.v2.6.11-r1
        version: 2.6.11-r1
      - name: service-registry-operator.v2.6.10
        version: 2.6.10
      - name: service-registry-operator.v2.6.9
        version: 2.6.9
      - name: service-registry-operator.v2.6.8-0.1740477005.p
        version: 2.6.8+0.1740477005.p
      - name: service-registry-operator.v2.6.8-0.1739457262.p
        version: 2.6.8+0.1739457262.p
      - name: service-registry-operator.v2.6.8
        version: 2.6.8
      - name: service-registry-operator.v2.6.6-0.1736942038.p
        version: 2.6.6+0.1736942038.p
      - name: service-registry-operator.v2.6.6
        version: 2.6.6
      - name: service-registry-operator.v2.6.5-0.1731946870.p
        version: 2.6.5+0.1731946870.p
      - name: service-registry-operator.v2.6.5
        version: 2.6.5
      - name: service-registry-operator.v2.6.5-0.1729511936.p
        version: 2.6.5+0.1729511936.p
      - name: service-registry-operator.v2.6.5-0.1730982111.p
        version: 2.6.5+0.1730982111.p
      - name: service-registry-operator.v2.6.3
        version: 2.6.3
      - name: service-registry-operator.v2.6.1-0.1724927626.p
        version: 2.6.1+0.1724927626.p
      - name: service-registry-operator.v2.6.1
        version: 2.6.1
      - name: service-registry-operator.v2.3.5
        version: 2.3.5
      - name: service-registry-operator.v2.3.4-0.1716987426.p
        version: 2.3.4+0.1716987426.p
      - name: service-registry-operator.v2.3.4
        version: 2.3.4
      - name: service-registry-operator.v2.3.3
        version: 2.3.3
      - name: service-registry-operator.v2.3.2
        version: 2.3.2
      - name: service-registry-operator.v2.3.1-0.1706181988.p
        version: 2.3.1+0.1706181988.p
      - name: service-registry-operator.v2.3.1
        version: 2.3.1
      - name: service-registry-operator.v2.3.0
        version: 2.3.0
      - name: service-registry-operator.v2.2.3
        version: 2.2.3
      - name: service-registry-operator.v2.2.3-0.1698369839.p
        version: 2.2.3+0.1698369839.p
      - name: service-registry-operator.v2.2.2
        version: 2.2.2
      - name: service-registry-operator.v2.2.2-0.1697116886.p
        version: 2.2.2+0.1697116886.p
      - name: service-registry-operator.v2.2.1
        version: 2.2.1
      - name: service-registry-operator.v2.2.0
        version: 2.2.0
      - name: service-registry-operator.v2.1.5
        version: 2.1.5
      - name: service-registry-operator.v2.1.4
        version: 2.1.4
      - name: service-registry-operator.v2.1.3
        version: 2.1.3
      - name: service-registry-operator.v2.1.2-0.1676389596.p
        version: 2.1.2+0.1676389596.p
      - name: service-registry-operator.v2.1.2
        version: 2.1.2
      - name: service-registry-operator.v2.1.1
        version: 2.1.1
      - name: service-registry-operator.v2.1.0
        version: 2.1.0
      - name: service-registry-operator.v2.0.7
        version: 2.0.7
      - name: service-registry-operator.v2.0.7-0.1662549961.p
        version: 2.0.7+0.1662549961.p
      - name: service-registry-operator.v2.0.7-0.1661784420.p
        version: 2.0.7+0.1661784420.p
      - name: service-registry-operator.v2.0.6
        version: 2.0.6
      - name: service-registry-operator.v2.0.5-0.1648813064.p
        version: 2.0.5+0.1648813064.p
      - name: service-registry-operator.v2.0.5-0.1646309885.p
        version: 2.0.5+0.1646309885.p
      - name: service-registry-operator.v2.0.5-0.1654503709.p
        version: 2.0.5+0.1654503709.p
      - name: service-registry-operator.v2.0.5-0.1652887585.p
        version: 2.0.5+0.1652887585.p
      - name: service-registry-operator.v2.0.5-0.1652429755.p
        version: 2.0.5+0.1652429755.p
      - name: service-registry-operator.v2.0.5-0.1651047781.p
        version: 2.0.5+0.1651047781.p
      - name: service-registry-operator.v2.0.5-0.1655396591.p
        version: 2.0.5+0.1655396591.p
      - name: service-registry-operator.v2.0.5-0.1647938417.p
        version: 2.0.5+0.1647938417.p
      - name: service-registry-operator.v2.0.5
        version: 2.0.5
      - name: service-registry-operator.v2.0.4
        version: 2.0.4
      - name: service-registry-operator.v2.0.3
        version: 2.0.3
      - name: service-registry-operator.v2.0.2
        version: 2.0.2
      - name: service-registry-operator.v2.0.1
        version: 2.0.1
      - name: service-registry-operator.v2.0.0
        version: 2.0.0
      name: 2.x
    defaultChannel: 2.x
    packageName: service-registry-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microsoft
      provider-url: ""
    name: azure-service-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: azure-service-operator.v1.0.59040
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "APIMgmtAPI",
                "metadata": {
                  "name": "aso-apimgmtapi-sample"
                },
                "spec": {
                  "apiId": "apiId0",
                  "apiService": "apimservice-sample",
                  "location": "westus",
                  "properties": {
                    "apiVersionDescription": "API version description",
                    "description": "API description",
                    "displayName": "aso-apimgmtapi-sample",
                    "isCurrent": true,
                    "isOnline": true,
                    "path": "/api/test",
                    "protocols": [
                      "http"
                    ],
                    "subscriptionRequired": false
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ApimService",
                "metadata": {
                  "name": "apimservice-sample"
                },
                "spec": {
                  "appInsightsName": "appinsights-sample-wmm",
                  "appInsightsResourceGroup": "resourcegroup-azure-operators-wmm",
                  "location": "westus",
                  "publisherEmail": {
                    "Your email": ""
                  },
                  "publisherName": {
                    "Your name": ""
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "tier": "premium",
                  "vnetName": "virtualnetwork-sample",
                  "vnetResourceGroup": "resourcegroup-azure-operators",
                  "vnetSubnetName": "test1",
                  "vnetType": "none"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AppInsights",
                "metadata": {
                  "name": "appinsights-sample"
                },
                "spec": {
                  "applicationType": "other",
                  "kind": "other",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AppInsightsApiKey",
                "metadata": {
                  "name": "key-sample"
                },
                "spec": {
                  "appInsights": "appinsights-sample",
                  "authSDKControlChannel": true,
                  "readTelemetry": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "writeAnnotations": true
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureLoadBalancer",
                "metadata": {
                  "name": "lb-sample-01"
                },
                "spec": {
                  "backendAddressPoolName": "test",
                  "backendPort": 3000,
                  "frontendPortRangeEnd": 2000,
                  "frontendPortRangeStart": 1001,
                  "inboundNatPoolName": "test",
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "lb-pip-sample-01",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureNetworkInterface",
                "metadata": {
                  "name": "nic-sample-01"
                },
                "spec": {
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "nic-pip-sample-01",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnetName": "test1",
                  "vnetName": "vnet-sample-01"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureNetworkInterface",
                "metadata": {
                  "name": "nic-sample-02"
                },
                "spec": {
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "nic-pip-sample-02",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnetName": "test1",
                  "vnetName": "vnet-sample-02"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "lb-pip-sample-01"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "nic-pip-sample-01"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "nic-pip-sample-02"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLManagedUser",
                "metadata": {
                  "name": "azuresqlmanageduser-sample"
                },
                "spec": {
                  "dbName": "azuresqldatabase-sample",
                  "managedIdentityClientId": null,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "db_owner"
                  ],
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLUser",
                "metadata": {
                  "name": "sqluser-sample"
                },
                "spec": {
                  "dbName": "azuresqldatabase-sample1",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "db_owner"
                  ],
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLVNetRule",
                "metadata": {
                  "name": "azuresqlvnetrule-sample"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-vnet",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlAction",
                "metadata": {
                  "name": "azuresqlaction-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverName": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "azuresqldatabase-sample1"
                },
                "spec": {
                  "edition": 0,
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "valueA",
                    "tag2": "valueB"
                  },
                  "name": "azuresqldatabase-sample3"
                },
                "spec": {
                  "edition": 0,
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlFailoverGroup",
                "metadata": {
                  "name": "azuresqlfailovergroup-sample"
                },
                "spec": {
                  "databaselist": [
                    "azuresqldatabase-sample1",
                    "azuresqldatabase-sample2"
                  ],
                  "failovergraceperiod": 30,
                  "failoverpolicy": "Automatic",
                  "location": "eastus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "secondaryserver": "sqlserver-samplesec",
                  "secondaryserverresourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-samplepri"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlFirewallRule",
                "metadata": {
                  "name": "sqlf-allowazuresvcaccess"
                },
                "spec": {
                  "endipaddress": "0.0.0.0",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "startipaddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample-777"
                },
                "spec": {
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample777"
                },
                "spec": {
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVMScaleSet",
                "metadata": {
                  "name": "vmss-sample-01"
                },
                "spec": {
                  "adminUserName": "azureuser7",
                  "backendAddressPoolName": "test",
                  "capacity": 3,
                  "inboundNatPoolName": "test",
                  "loadBalancerName": "lb-sample-01",
                  "location": "SouthCentralUS",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": "{ssh public key}",
                  "subnetName": "test1",
                  "virtualNetworkName": "vnet-sample-01",
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachine",
                "metadata": {
                  "name": "hpfvm1"
                },
                "spec": {
                  "adminUserName": "azureuser",
                  "location": "SouthCentralUS",
                  "networkInterfaceName": "nic-sample-02",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": {
                    "ssh key": ""
                  },
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachine",
                "metadata": {
                  "name": "hpfvm20"
                },
                "spec": {
                  "adminUserName": "azureuser",
                  "location": "SouthCentralUS",
                  "networkInterfaceName": "nic-sample-01",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": "{ssh public key}",
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachineExtension",
                "metadata": {
                  "name": "vmext20a"
                },
                "spec": {
                  "autoUpgradeMinorVersion": true,
                  "forceUpdateTag": "Test2",
                  "location": "SouthCentralUS",
                  "protectedSettings": "{}",
                  "publisher": "Microsoft.Azure.Extensions",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "settings": "{\"commandToExecute\":\"echo 'hello, world'\"}",
                  "typeHandlerVersion": "2.1",
                  "typeName": "CustomScript",
                  "vmName": "vm20"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "BlobContainer",
                "metadata": {
                  "name": "blobcontainer-sample"
                },
                "spec": {
                  "accesslevel": "Container",
                  "accountname": "storageaccountsample777",
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ConsumerGroup",
                "metadata": {
                  "name": "consumergroup-sample-1"
                },
                "spec": {
                  "eventHub": "eventhub-sample-1",
                  "namespace": "eventhubnamespace-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "CosmosDB",
                "metadata": {
                  "name": "cosmosdb-sample-1"
                },
                "spec": {
                  "kind": "GlobalDocumentDB",
                  "location": "westus",
                  "properties": {
                    "databaseAccountOfferType": "Standard",
                    "isVirtualNetworkFilterEnabled": true
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "CosmosDB",
                "metadata": {
                  "labels": {
                    "CosmosAccountType": "Production"
                  },
                  "name": "cosmosdb-sample-2"
                },
                "spec": {
                  "ipRules": [
                    "104.42.195.92,40.76.54.131,52.176.6.30,52.169.50.45,52.187.184.26"
                  ],
                  "kind": "GlobalDocumentDB",
                  "location": "eastus",
                  "locations": [
                    {
                      "failoverPriority": 0,
                      "isZoneRedundant": true,
                      "locationName": "eastus"
                    },
                    {
                      "failoverPriority": 1,
                      "locationName": "westus"
                    }
                  ],
                  "properties": {
                    "databaseAccountOfferType": "Standard",
                    "enableMultipleWriteLocations": true
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "Eventhub",
                "metadata": {
                  "name": "eventhub-sample-1"
                },
                "spec": {
                  "authorizationRule": {
                    "name": "RootManageSharedAccessKey",
                    "rights": [
                      "Listen",
                      "Manage",
                      "Send"
                    ]
                  },
                  "location": "westus",
                  "namespace": "eventhubnamespace-sample-1",
                  "properties": {
                    "messageRetentionInDays": 7,
                    "partitionCount": 2
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "Eventhub",
                "metadata": {
                  "name": "eventhub-sample-2307-10-capture"
                },
                "spec": {
                  "authorizationRule": {
                    "name": "RootManageSharedAccessKey",
                    "rights": [
                      "Listen",
                      "Manage",
                      "Send"
                    ]
                  },
                  "location": "westus",
                  "namespace": "eventhubnamespace-sample-1",
                  "properties": {
                    "captureDescription": {
                      "destination": {
                        "archiveNameFormat": "{Namespace}/{EventHub}/{PartitionId}/{Year}/{Month}/{Day}/{Hour}/{Minute}/{Second}",
                        "blobContainer": "capturecontainer",
                        "name": "EventHubArchive.AzureBlockBlob",
                        "storageAccount": {
                          "accountName": "storageaccountauv1",
                          "resourceGroup": "my-resource-group"
                        }
                      },
                      "enabled": true,
                      "intervalInSeconds": 120,
                      "sizeLimitInBytes": 524288000
                    },
                    "messageRetentionInDays": 7,
                    "partitionCount": 2
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "EventhubNamespace",
                "metadata": {
                  "name": "eventhubnamespace-sample-1"
                },
                "spec": {
                  "location": "westus",
                  "properties": {
                    "isAutoInflateEnabled": false,
                    "kafkaEnabled": false,
                    "maximumThroughputUnits": 0
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "capacity": 1,
                    "name": "Standard",
                    "tier": "Standard"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVault",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "keyvaultsample123"
                },
                "spec": {
                  "accessPolicies": [
                    {
                      "objectID": "\u003cobjectID\u003e",
                      "permissions": {
                        "certificates": [
                          "list",
                          "get"
                        ],
                        "keys": [
                          "list",
                          "get"
                        ],
                        "secrets": [
                          "list",
                          "get"
                        ],
                        "storage": [
                          "list",
                          "get"
                        ]
                      },
                      "tenantID": "\u003ctenantID\u003e"
                    }
                  ],
                  "enableSoftDelete": false,
                  "location": "westus",
                  "networkPolicies": {
                    "bypass": "AzureServices",
                    "defaultAction": "Allow",
                    "ipRules": [
                      "172.16.0.0/24",
                      "172.16.1.0/24"
                    ],
                    "virtualNetworkRules": [
                      "/subscriptions/\u003csubid\u003e/resourceGroups/rg1/providers/Microsoft.Network/virtualNetworks/test-vnet/subnets/subnet1",
                      "/subscriptions/\u003csubid\u003e/resourceGroups/rg1/providers/Microsoft.Network/virtualNetworks/test-vnet/subnets/subnet2"
                    ]
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "name": "standard"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVault",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "keyvaultsample123-simple"
                },
                "spec": {
                  "enableSoftDelete": false,
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVaultKey",
                "metadata": {
                  "name": "keyvaultkey-sample"
                },
                "spec": {
                  "keySize": 4096,
                  "keyVault": "keyvaultsample123",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "type": "RSA"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLDatabase",
                "metadata": {
                  "name": "mysqldatabase-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLDatabase",
                "metadata": {
                  "name": "mysqldatabase-sample2"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLFirewallRule",
                "metadata": {
                  "name": "mysqlfirewallrule-sample"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLFirewallRule",
                "metadata": {
                  "name": "mysqlfirewallrule-sample2"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample2",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-replica"
                },
                "spec": {
                  "createMode": "Replica",
                  "location": "eastus2",
                  "replicaProperties": {
                    "sourceServerId": "/subscriptions/{SUBID}/resourceGroups/resourcegroup-azure-operators/providers/Microsoft.DBforMySQL/servers/mysqlserver-sample"
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-sample2"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLUser",
                "metadata": {
                  "name": "mysqluser-sample"
                },
                "spec": {
                  "dbName": "mysqldatabase-sample",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "SELECT"
                  ],
                  "server": "mysqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLVNetRule",
                "metadata": {
                  "name": "mysqlvnetrule-sample"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-vnet",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLDatabase",
                "metadata": {
                  "name": "postgresqldatabase-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLDatabase",
                "metadata": {
                  "name": "postgresqldatabase-sample2"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLFirewallRule",
                "metadata": {
                  "name": "postgresqlfirewallrule-sample"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLFirewallRule",
                "metadata": {
                  "name": "postgresqlfirewallrule-sample2"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample2",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-replica3"
                },
                "spec": {
                  "createMode": "Replica",
                  "location": "eastus",
                  "replicaProperties": {
                    "sourceServerId": "/subscriptions/{subscription ID}/resourceGroups/resourcegroup-azure-operators/providers/Microsoft.DBforPostgreSQL/servers/postgresqlserver-sample"
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-sample2"
                },
                "spec": {
                  "location": "southcentralus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 2,
                    "family": "Gen5",
                    "name": "B_Gen5_2",
                    "size": "51200",
                    "tier": "Basic"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLUser",
                "metadata": {
                  "name": "psqluser-sample"
                },
                "spec": {
                  "dbName": "postgresqldatabase-sample",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "azure_pg_admin"
                  ],
                  "server": "postgresqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLVNetRule",
                "metadata": {
                  "name": "postgresqlvnetrule-sample1"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-azure-operators",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCache",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "rediscache-sample-1"
                },
                "spec": {
                  "location": "westus",
                  "properties": {
                    "enableNonSslPort": true,
                    "sku": {
                      "capacity": 1,
                      "family": "C",
                      "name": "Basic"
                    }
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCacheAction",
                "metadata": {
                  "name": "rediscacheaction-sample-1"
                },
                "spec": {
                  "actionName": "rollallkeys",
                  "cacheName": "rediscache-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCacheFirewallRule",
                "metadata": {
                  "name": "rediscachefirewallrule"
                },
                "spec": {
                  "properties": {
                    "endIP": "0.0.0.0",
                    "startIP": "0.0.0.0"
                  },
                  "redisCache": "rediscache-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ResourceGroup",
                "metadata": {
                  "name": "resourcegroup-azure-operators"
                },
                "spec": {
                  "location": "westus"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "StorageAccount",
                "metadata": {
                  "name": "storageaccountsample777"
                },
                "spec": {
                  "accessTier": "Hot",
                  "kind": "StorageV2",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "name": "Standard_RAGRS"
                  },
                  "supportsHttpsTrafficOnly": true
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "VirtualNetwork",
                "metadata": {
                  "name": "vnet-sample-01"
                },
                "spec": {
                  "addressSpace": "10.0.0.0/8",
                  "location": "SouthCentralUS",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnets": [
                    {
                      "subnetAddressPrefix": "10.1.0.0/16",
                      "subnetName": "test1"
                    },
                    {
                      "subnetAddressPrefix": "10.2.0.0/16",
                      "subnetName": "test2"
                    }
                  ]
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "VirtualNetwork",
                "metadata": {
                  "name": "vnet-sample-02"
                },
                "spec": {
                  "addressSpace": "10.0.0.0/8",
                  "location": "SouthCentralUS",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnets": [
                    {
                      "subnetAddressPrefix": "10.1.0.0/16",
                      "subnetName": "test1"
                    },
                    {
                      "subnetAddressPrefix": "10.2.0.0/16",
                      "subnetName": "test2"
                    }
                  ]
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "BlobContainer",
                "metadata": {
                  "name": "blobcontainer-sample"
                },
                "spec": {
                  "accessLevel": "Container",
                  "accountName": "storageaccountsample777",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "MySQLServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "mysqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "postgresqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "azuresqldatabase-sample"
                },
                "spec": {
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "sku": {
                    "name": "Basic",
                    "tier": "Basic"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlFailoverGroup",
                "metadata": {
                  "name": "azuresqlfailovergroup-sample"
                },
                "spec": {
                  "databaseList": [
                    "azuresqldatabase-sample",
                    "azuresqldatabase-sample2"
                  ],
                  "failoverGracePeriod": 30,
                  "failoverPolicy": "automatic",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "secondaryServer": "sqlserver-samplesec",
                  "secondaryServerResourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-samplepri"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlFirewallRule",
                "metadata": {
                  "name": "sqlf-allowazuresvcaccess"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample-777"
                },
                "spec": {
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Cloud Provider
          certified: "false"
          containerImage: mcr.microsoft.com/k8s/azureserviceoperator:1.0.59040
          createdAt: Mon 12 Jun 2023 04:24:02 UTC
          description: The Azure Service Operator helps you provision Azure resources
            and connect your applications to them from within Kubernetes.
          displayName: Azure Service Operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Azure/azure-service-operator
          support: https://github.com/Azure/azure-service-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Creates an API with the specified properties in the specified
              API Management service.
            displayName: APIMgmtAPI
            kind: APIMgmtAPI
            name: apimgmtapis.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an API Management instance into a specified Resource
              Group at the specified location. It also provides the option to link
              to an Application Insights instance for logging, and to place the API
              Management instance in a specified Virtual Network.
            displayName: ApimService
            kind: ApimService
            name: apimservices.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Application Insights instance into a specified
              Resource Group at the specified location.
            displayName: AppInsights
            kind: AppInsights
            name: appinsights.azure.microsoft.com
            version: v1alpha1
          - description: Creates an Api Key to access a given Application Insights
              instance.
            displayName: AppInsightsApiKey
            kind: AppInsightsApiKey
            name: appinsightsapikeys.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Load Balancer (LB) into a specified Resource
              Group at the specified location.
            displayName: AzureLoadBalancer
            kind: AzureLoadBalancer
            name: azureloadbalancers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Network Interface (NIC) into a specified
              Resource Group at the specified location. Users can specify underlying
              public IP address and virtual network configurations in their NIC setup.
            displayName: AzureNetworkInterface
            kind: AzureNetworkInterface
            name: azurenetworkinterfaces.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Public IP Address (PIP) into a specified
              Resource Group at the specified location. Users can specify IP allocation
              method, idle timeout, IP address version, and SKU.
            displayName: AzurePublicIPAddress
            kind: AzurePublicIPAddress
            name: azurepublicipaddresses.azure.microsoft.com
            version: v1alpha1
          - description: Allows you to roll the password for the specified SQL server.
            displayName: AzureSqlAction
            kind: AzureSqlAction
            name: azuresqlactions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an SQL database given the SQL server.
            displayName: AzureSqlDatabase
            kind: AzureSqlDatabase
            name: azuresqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an SQL database given the SQL server
            displayName: AzureSqlDatabase
            kind: AzureSqlDatabase
            name: azuresqldatabases.azure.microsoft.com
            version: v1beta1
          - description: Deploys a failover group on a specified Azure SQL server
              given the secondary server and the databases to failover.
            displayName: AzureSqlFailoverGroup
            kind: AzureSqlFailoverGroup
            name: azuresqlfailovergroups.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a failover group on a specified Azure SQL server
              given the secondary server and the databases to failover.
            displayName: AzureSqlFailoverGroup
            kind: AzureSqlFailoverGroup
            name: azuresqlfailovergroups.azure.microsoft.com
            version: v1beta1
          - description: Deploys a firewall rule to allow access to the SQL server
              from specific IPs.
            displayName: AzureSqlFirewallRule
            kind: AzureSqlFirewallRule
            name: azuresqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the SQL server
              from specific IPs.
            displayName: AzureSqlFirewallRule
            kind: AzureSqlFirewallRule
            name: azuresqlfirewallrules.azure.microsoft.com
            version: v1beta1
          - description: Binds a specified Managed Identity as a user on the SQL database
            displayName: AzureSQLManagedUser
            kind: AzureSQLManagedUser
            name: azuresqlmanagedusers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure SQL server given the location and Resource
              Group.
            displayName: AzureSqlServer
            kind: AzureSqlServer
            name: azuresqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure SQL server given the location and Resource
              Group.
            displayName: AzureSqlServer
            kind: AzureSqlServer
            name: azuresqlservers.azure.microsoft.com
            version: v1beta1
          - description: Creates an user on the specified Azure SQL database and stores
              the username/password as secrets.
            displayName: AzureSQLUser
            kind: AzureSQLUser
            name: azuresqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a VirtualNetwork rule to allow access to the SQL
              server from specific VNets.
            displayName: AzureSQLVNetRule
            kind: AzureSQLVNetRule
            name: azuresqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys Azure Virtual Machine Extensions to provide post-deployment
              configuration and automation on Azure VMs.
            displayName: AzureVirtualMachineExtension
            kind: AzureVirtualMachineExtension
            name: azurevirtualmachineextensions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Machine (VM) into a specified Resource
              Group at the specified location. Users can specify platform image, size,
              user name and public SSH key, etc. for the VM.
            displayName: AzureVirtualMachine
            kind: AzureVirtualMachine
            name: azurevirtualmachines.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Machine Scale Set (VMSS) into a
              specified Resource Group at the specified location. Users can specify
              platform image, size, user name and public SSH key, etc. for the VMSS.
            displayName: AzureVMScaleSet
            kind: AzureVMScaleSet
            name: azurevmscalesets.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a Blob Container into a specified Resource Group
              at the specified location. A Storage Account must first be created prior
              to creating a Blob Container.
            displayName: BlobContainer
            kind: BlobContainer
            name: blobcontainers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a Blob Container into a specified Resource Group
              at the specified location. A Storage Account must first be created prior
              to creating a Blob Container.
            displayName: BlobContainer
            kind: BlobContainer
            name: blobcontainers.azure.microsoft.com
            version: v1alpha2
          - description: Deploys a consumer group given the Event Hub, Event Hubs
              namespace and Resource Group.
            displayName: ConsumerGroup
            kind: ConsumerGroup
            name: consumergroups.azure.microsoft.com
            version: v1alpha1
          - description: Provision a CosmosDB instance given the Cosmos DB instance
              type, location, and Resource Group.
            displayName: CosmosDB
            kind: CosmosDB
            name: cosmosdbs.azure.microsoft.com
            version: v1alpha1
          - kind: CosmosDBSQLDatabase
            name: cosmosdbsqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Event Hubs namespace given the Resource Group
              and location. Also has the ability to configure SKU, properties, and
              network rules.
            displayName: EventhubNamespace
            kind: EventhubNamespace
            name: eventhubnamespaces.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Event Hubs instance given the Event Hubs namespace,
              Resource Group and Location.
            displayName: Eventhub
            kind: Eventhub
            name: eventhubs.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Key Vault key given the location and Resource
              Group.
            displayName: KeyVaultKey
            kind: KeyVaultKey
            name: keyvaultkeys.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Key Vault given the location and Resource
              Group.
            displayName: KeyVault
            kind: KeyVault
            name: keyvaults.azure.microsoft.com
            version: v1alpha1
          - description: Add an AAD user to an existing MySQL database.
            kind: MySQLAADUser
            name: mysqlaadusers.azure.microsoft.com
            version: v1alpha1
          - description: Add an AAD user to an existing MySQL database.
            kind: MySQLAADUser
            name: mysqlaadusers.azure.microsoft.com
            version: v1alpha2
          - description: Deploys a database under the given Azure Database for MySQL
              server
            displayName: MySQLDatabase
            kind: MySQLDatabase
            name: mysqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the Azure Database
              for MySQL server from the specified IP range.
            displayName: MySQLFirewallRule
            kind: MySQLFirewallRule
            name: mysqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Create an AAD-based admininistrative user in a MySQL server.
            kind: MySQLServerAdministrator
            name: mysqlserveradministrators.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for MySQL server given the Location,
              Resource Group and other properties. Also helps creating read replicas
              for MySQL server.
            displayName: MySQLServer
            kind: MySQLServer
            name: mysqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for MySQL server given the Location,
              Resource Group and other properties. Also helps creating read replicas
              for MySQL server.
            displayName: MySQLServer
            kind: MySQLServer
            name: mysqlservers.azure.microsoft.com
            version: v1alpha2
          - description: Add a new user to an existing MySQL database.
            displayName: MySQLUser
            kind: MySQLUser
            name: mysqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Add a new user to an existing MySQL database.
            kind: MySQLUser
            name: mysqlusers.azure.microsoft.com
            version: v1alpha2
          - description: Add virtual network rules to the MySQL server.
            displayName: MySQLVNetRule
            kind: MySQLVNetRule
            name: mysqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a database under the given Azure Database for PostgreSQL
              server.
            displayName: PostgreSQLDatabase
            kind: PostgreSQLDatabase
            name: postgresqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the Azure Database
              for PostgreSQL server from the specified IP range.
            displayName: PostgreSQLFirewallRule
            kind: PostgreSQLFirewallRule
            name: postgresqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for PostgreSQL server given the
              Location and Resource Group.
            displayName: PostgreSQLServer
            kind: PostgreSQLServer
            name: postgresqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for PostgreSQL server given the
              Location and Resource Group.
            displayName: PostgreSQLServer
            kind: PostgreSQLServer
            name: postgresqlservers.azure.microsoft.com
            version: v1alpha2
          - description: Create and Manage Users for Azure Database for PostgreSQL
              server.
            displayName: PostgreSQLUser
            kind: PostgreSQLUser
            name: postgresqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Create and manage VNet service endpoints and VNet rules in
              Azure Database for PostgreSQL.
            displayName: PostgreSQLVNetRule
            kind: PostgreSQLVNetRule
            name: postgresqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Allows you to regenerate keys and reboot the RedisCache cluster.
            displayName: RedisCacheAction
            kind: RedisCacheAction
            name: rediscacheactions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the RedisCache
              from the specified IP range.
            displayName: RedisCacheFirewallRule
            kind: RedisCacheFirewallRule
            name: rediscachefirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Cache for Redis into a specified Resource
              Group at the specified location.
            displayName: RedisCache
            kind: RedisCache
            name: rediscaches.azure.microsoft.com
            version: v1alpha1
          - description: Provisions a Resource Group given the location and a name.
            displayName: ResourceGroup
            kind: ResourceGroup
            name: resourcegroups.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Storage Account into a specified Resource
              Group at the specified location.
            displayName: StorageAccount
            kind: StorageAccount
            name: storageaccounts.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Network into a specified Resource
              Group at the specified location, in the address space provided. Users
              are also able to add subnets to their virtual network through the operator.
            displayName: VirtualNetwork
            kind: VirtualNetwork
            name: virtualnetworks.azure.microsoft.com
            version: v1alpha1
        description: |
          ## Overview

          The Azure Service Operator comprises of:

          - The Custom Resource Definitions (CRDs) for each of the Azure services a Kubernetes user can provision.
          - The Kubernetes controller that watches for requests to create Custom Resources for each of these CRDs and creates them.

          The project was built using [Kubebuilder](https://book.kubebuilder.io/).

          Curious to see how it all works? Check out our [control flow diagrams](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/controlflow.md).

          **Note**: The current version of Azure Service Operator only supports watching all namespaces in the cluster.

          ## Supported Azure Services

          - [Resource Group](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/resourcegroup/resourcegroup.md)
          - [Event Hubs](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/eventhub/eventhub.md)
          - [Azure SQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/azuresql/azuresql.md)
          - [Azure Database for PostgreSQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/postgresql/postgresql.md)
          - [Azure Database for MySQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/mysql/mysql.md)
          - [Azure Key Vault](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/keyvault/keyvault.md)
          - [Azure Cache for Redis](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/rediscache/rediscache.md)
          - [Storage Account](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/storage/storageaccount.md)
          - [Blob Storage](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/storage/blobcontainer.md)
          - [Virtual Network](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/virtualnetwork/virtualnetwork.md)
          - [Application Insights](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/appinsights/appinsights.md)
          - [API Management](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/apimgmt/apimgmt.md)
          - [Cosmos DB](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/cosmosdb/cosmosdb.md)
          - [Virtual Machine](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/virtualmachine/virtualmachine.md)
          - [Virtual Machine Scale Set](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/vmscaleset/vmscaleset.md)

          ## Instructions - Please read before installing

          Before you begin, verify you're running the Azure CLI version 2.0.53 or later. Run `az --version` to find the version. If you need to install or upgrade, see [Install Azure CLI](https://docs.microsoft.com/en-us/cli/azure/install-azure-cli).

          1. Create an Azure Service Principal. You'll need this to grant Azure Service Operator permissions to create resources in your subscription.

              First, set the following environment variables to your Azure Tenant ID and Subscription ID with your values:
              ```yaml
              AZURE_TENANT_ID=<your-tenant-id-goes-here>
              AZURE_SUBSCRIPTION_ID=<your-subscription-id-goes-here>
              ```

              You can find these values by using the [Azure CLI](https://docs.microsoft.com/en-us/cli/azure/?view=azure-cli-latest):
              ```sh
              az account show
              ```

          2. Next, we'll create a Service Principal with Contributor permissions for your subscription, so ASO can create resources in your subscription on your behalf. Note that the [ServicePrincipal](https://docs.microsoft.com/en-us/cli/azure/create-an-azure-service-principal-azure-cli) you pass to the command below needs to have access to create resources in your subscription. If you'd like to use Managed Identity for authorization instead, check out instructions [here](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/managedidentity.md).

              ```sh
              az ad sp create-for-rbac -n "azure-service-operator" --role contributor \
                  --scopes /subscriptions/$AZURE_SUBSCRIPTION_ID
              ```

              This should give you output like the following:
              ```sh
              "appId": "xxxxxxxxxx",
              "displayName": "azure-service-operator",
              "name": "http://azure-service-operator",
              "password": "xxxxxxxxxxx",
              "tenant": "xxxxxxxxxxxxx"
              ```

          3.  Once you have created a Service Principal, gather the following values:

              `AZURE_TENANT_ID` is the Tenant ID from Step 1.

              `AZURE_SUBSCRIPTION_ID` is the Subscription ID from Step 1.

              `AZURE_CLIENT_ID` is the appID from the Service Principal created in Step 2.

              `AZURE_CLIENT_SECRET` is the password from the service Principal we created in Step 2.

              `AZURE_CLOUD_ENV` is the Azure Environment you'd like to use, i.e. AzurePublicCloud, AzureUSGovernmentCloud, AzureChinaCloud, AzureGermanCloud.

          4.  Set those values in a `Secret` called `azureoperatorsettings` within the `operators` Namespace:
              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name: azureoperatorsettings
                namespace: operators
              stringData:
                AZURE_TENANT_ID: <your-tenant-id-goes-here>
                AZURE_SUBSCRIPTION_ID: <your-subscription-id-goes-here>
                AZURE_CLIENT_ID: <your-client-id>
                AZURE_CLIENT_SECRET: <your-client-secret>
                AZURE_CLOUD_ENV: <your-azure-cloud-environment>
              ```

          5. Now you can proceed to install the Azure Service Operator to the `operators` Namespace via the Install button on the top right of this page. After the operator is installed, you will then see the Azure Service Operator pod running in your cluster:

              ```console
              $ kubectl get pods -n operators
              NAME                                                READY   STATUS    RESTARTS   AGE
              azureoperator-controller-manager-7dd75bbd97-mk4s9   2/2     Running   0          35s
              ```

          ## About the project

          This project maintains [releases of the Azure Service Operator](https://github.com/Azure/azure-service-operator/releases).

          Please see the [FAQ](https://github.com/Azure/azure-service-operator/blob/main/docs/faq.md) for answers to commonly asked questions about the Azure Service Operator.

          Have more questions? Feel free to consult our documentation [here](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/contents.md).

          ## Contributing

          The [contribution guide](https://github.com/Azure/azure-service-operator/blob/main/CONTRIBUTING.md) covers everything you need to know about how you can contribute to Azure Service Operators. The [developer guide](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/contents.md#developing-azure-service-operator) will help you onboard as a developer.

          ## Support

          Azure Service Operator is an open source project that is **not** covered by the [Microsoft Azure support policy](https://support.microsoft.com/en-us/help/2941892/support-for-linux-and-open-source-technology-in-azure). Please search open issues [here](https://github.com/Azure/azure-service-operator/issues). If your issue isn't already represented, please [open a new one](https://github.com/Azure/azure-service-operator/issues/new/choose). The Azure Service Operator project maintainers will respond to the best of their abilities.

          ## Code of conduct

          This project has adopted the [Microsoft Open Source Code of Conduct](https://opensource.microsoft.com/codeofconduct/). For more information, see the [Code of Conduct FAQ](https://opensource.microsoft.com/codeofconduct/faq) or contact [opencode@microsoft.com](mailto:opencode@microsoft.com) with any additional questions or comments.
        displayName: Azure Service Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - azure, microsoft, cloud service broker
        links:
        - name: Azure Service Operator
          url: https://github.com/Azure/azure-service-operator
        maintainers:
        - email: christian.muirhead@microsoft.com
          name: Christian Muirhead
        maturity: stable
        provider:
          name: Microsoft
        relatedImages:
        - mcr.microsoft.com/k8s/azureserviceoperator:1.0.59040
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 1.0.59040
      entries:
      - name: azure-service-operator.v1.0.59040
        version: 1.0.59040
      name: stable
    defaultChannel: stable
    packageName: azure-service-operator
    provider:
      name: Microsoft
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/kuadrant-operator
    name: kuadrant-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kuadrant-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "authScheme": {
                    "identity": [
                      {
                        "apiKey": {
                          "allNamespaces": true,
                          "selector": {
                            "matchLabels": {
                              "app": "toystore"
                            }
                          }
                        },
                        "credentials": {
                          "in": "authorization_header",
                          "keySelector": "APIKEY"
                        },
                        "name": "friends"
                      }
                    ],
                    "response": [
                      {
                        "json": {
                          "properties": [
                            {
                              "name": "userID",
                              "valueFrom": {
                                "authJSON": "auth.identity.metadata.annotations.secret\\.kuadrant\\.io/user-id"
                              }
                            }
                          ]
                        },
                        "name": "rate-limit-apikey",
                        "wrapper": "envoyDynamicMetadata",
                        "wrapperKey": "ext_auth_data"
                      }
                    ]
                  },
                  "rules": null,
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "rateLimits": [
                    {
                      "configurations": [
                        {
                          "actions": [
                            {
                              "metadata": {
                                "default_value": "no-user",
                                "descriptor_key": "userID",
                                "metadata_key": {
                                  "key": "envoy.filters.http.ext_authz",
                                  "path": [
                                    {
                                      "segment": {
                                        "key": "ext_auth_data"
                                      }
                                    },
                                    {
                                      "segment": {
                                        "key": "userID"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          ]
                        }
                      ],
                      "limits": [
                        {
                          "conditions": [],
                          "maxValue": 2,
                          "seconds": 10,
                          "variables": [
                            "userID"
                          ]
                        }
                      ]
                    }
                  ],
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/kuadrant-operator:v0.3.1
          createdAt: "2023-06-01T12:24:17Z"
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Enable AuthN and AuthZ based access control on workloads
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1beta1
          - description: Kuadrant is the Schema for the kuadrants API
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: Enable access control on workloads based on HTTP rate limiting
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1beta1
        description: A Kubernetes Operator to manage the lifecycle of the Kuadrant
          system
        displayName: Kuadrant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: Kuadrant Operator
          url: https://github.com/Kuadrant/kuadrant-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - quay.io/kuadrant/kuadrant-operator:v0.3.1
        - oci://quay.io/kuadrant/wasm-shim:v0.2.0
        version: 0.3.1
      entries:
      - name: kuadrant-operator.v0.3.1
        version: 0.3.1
      - name: kuadrant-operator.v0.3.0
        version: 0.3.0
      - name: kuadrant-operator.v0.2.1
        version: 0.2.1
      - name: kuadrant-operator.v0.2.0
        version: 0.2.0
      name: alpha
    - currentCSV: kuadrant-operator.v0.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSPolicy",
                "metadata": {
                  "name": "dnspolicy-sample"
                },
                "spec": {
                  "healthCheck": {
                    "endpoint": "/",
                    "protocol": "HTTP"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "TLSPolicy",
                "metadata": {
                  "name": "tlspolicy-sample"
                },
                "spec": {
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "self-signed-ca"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "kuadrant.io/v1beta2",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "rules": {
                    "authentication": {
                      "apikey": {
                        "apiKey": {
                          "selector": {}
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta2",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "limits": {
                    "toys": {
                      "rates": [
                        {
                          "duration": 1,
                          "limit": 50,
                          "unit": "minute"
                        }
                      ]
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/kuadrant-operator:v0.11.0
          createdAt: "2024-10-02T20:30:45Z"
          description: '[DEPRECATED] Follow https://docs.kuadrant.io instead'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthPolicy enables authentication and authorization for service
              workloads in a Gateway API network
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1beta2
          - description: DNSPolicy configures how North-South based traffic should
              be balanced and reach the gateways
            displayName: DNSPolicy
            kind: DNSPolicy
            name: dnspolicies.kuadrant.io
            version: v1alpha1
          - description: Kuadrant configures installations of Kuadrant Service Protection
              components
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: RateLimitPolicy enables rate limiting for service workloads
              in a Gateway API network
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1beta2
          - description: TLSPolicy provides tls for gateway listeners by managing
              the lifecycle of tls certificates
            displayName: TLSPolicy
            kind: TLSPolicy
            name: tlspolicies.kuadrant.io
            version: v1alpha1
        description: |-
          ## Deprecated
          This Kuadrant community operator is deprecated, Follow https://docs.kuadrant.io instead.
        displayName: '[DEPRECATED] Kuadrant Operator'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: Kuadrant Operator
          url: https://github.com/Kuadrant/kuadrant-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - quay.io/kuadrant/kuadrant-operator:v0.11.0
        - oci://quay.io/kuadrant/wasm-shim:v0.6.0
        version: 0.11.1
      entries:
      - name: kuadrant-operator.v0.11.1
        version: 0.11.1
      - name: kuadrant-operator.v0.11.0
        version: 0.11.0
      - name: kuadrant-operator.v0.10.0
        version: 0.10.0
      - name: kuadrant-operator.v0.7.2
        version: 0.7.2
      - name: kuadrant-operator.v0.4.1
        version: 0.4.1
      - name: kuadrant-operator.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: kuadrant-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/kuadrant-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Nutanix
      provider-url: ""
    name: nutanixcsioperator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nutanixcsioperator.v2.6.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crd.nutanix.com/v1alpha1",
                "kind": "NutanixCsiStorage",
                "metadata": {
                  "name": "nutanixcsistorage"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:859b6d1b2b36a48391e347c2a7431cafa08decafb56f801e21e544c3460a4b4b
          createdAt: "2024-10-30T17:10:05Z"
          description: An Operator for deploying and managing the Nutanix CSI Driver
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2.6.11
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi","disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/nutanix/csi-operator
          support: Nutanix Support Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a new instance of Nutanix CSI Driver
            displayName: Nutanix CSI Storage
            kind: NutanixCsiStorage
            name: nutanixcsistorages.crd.nutanix.com
            version: v1alpha1
        description: "# Overview \n\nThe Nutanix CSI Operator for Kubernetes packages,
          deploys, manages, upgrades the Nutanix CSI Driver on Kubernetes and OpenShift
          for dynamic provisioning of persistent volumes on Nutanix Enterprise Cloud
          platform.\n\nThe Nutanix Container Storage Interface (CSI) Driver for Kubernetes
          leverages Nutanix Volumes and Nutanix Files to provide scalable and persistent
          storage for stateful applications.\n\nWith the Nutanix CSI Provider you
          can:\n\n*  **Provide persistent storage to your containers**\n    * Leverage
          PVC resources to consume dynamically Nutanix storage\n    * With Files storage
          classes, applications on multiple pods can access the same storage, and
          also have the benefit of multi-pod read and write access. \n\n## Warning\n\nIf
          you upgrade from a previous version take care to validate your existing
          StorageClass parameters against the corresponding [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v2_6:CSI-Volume-Driver-v2_6).\n\n##
          Supported Access mode\n\n| Storage Class mode | ReadWriteOnce | ReadOnlyMany
          | ReadWriteMany |\n|--------------------|---------------|--------------|---------------|\n|
          Nutanix Volumes    | yes           | yes          | no            |\n| Nutanix
          Files      | yes           | yes          | yes           |\n\n\n## Supported
          orchestration platforms\nThe following table details orchestration platforms
          suitable for deployment of the Nutanix CSI driver.\n\n| Orchestration platform
          | Version | Architecture |\n|------------------------|---------|--------------|\n|
          Red Hat OpenShift      | 4.12    | x86          |\n| Red Hat OpenShift      |
          4.13    | x86          |\n| Red Hat OpenShift      | 4.14    | x86          |\n|
          Red Hat OpenShift      | 4.15    | x86          |\n| Red Hat OpenShift      |
          4.16    | x86          |\n| Red Hat OpenShift      | 4.17    | x86          |\n\n##
          Configuring k8s secret and storage class\nIn order to use this driver, deploy
          your NutanixCsiStorage instance and create the relevant storage classes
          and secrets, by following the [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v2_6:CSI-Volume-Driver-v2_6)
          or the example in below section:\n\n**1. Create a NutanixCsiStorage resource
          to deploy your driver**\n\nYou can do it directly inside the Operator UI
          with the `Create instance` button or with the following ressource\n```yaml\napiVersion:
          crd.nutanix.com/v1alpha1\nkind: NutanixCsiStorage\nmetadata:\n  name: nutanixcsistorage\n
          \ namespace: openshift-cluster-csi-drivers\nspec: {}\n```\n\n**2. Create
          a Nutanix secret**\n```yaml\napiVersion: v1\nkind: Secret\nmetadata:\n  name:
          ntnx-secret\n  namespace: openshift-cluster-csi-drivers\nstringData:\n  #
          prism-ip:prism-port:username:password.\n  key: 1.2.3.4:9440:admin:password\n```\n\n**3.
          Create storage classes**\n\n- volume mode example\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n  name: nutanix-volume\nprovisioner: csi.nutanix.com\nparameters:\n
          \ csi.storage.k8s.io/provisioner-secret-name: ntnx-secret\n  csi.storage.k8s.io/provisioner-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/node-publish-secret-name:
          ntnx-secret\n  csi.storage.k8s.io/node-publish-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/controller-expand-secret-name: ntnx-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/fstype: ext4\n  storageContainer:
          default-container\n  storageType: NutanixVolumes\n  #description: \"description
          added to each storage object created by the driver\"\n  #isSegmentedIscsiNetwork:
          \"false\"\n  #whitelistIPMode: ENABLED\n  #chapAuth: ENABLED\n  #isLVMVolume:
          \"false\"\n  #numLVMDisks: 4\nallowVolumeExpansion: true\nreclaimPolicy:
          Delete\n```\n\n- dynamic files mode example\n\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n    name: nutanix-dynfiles\nprovisioner: csi.nutanix.com\nparameters:\n
          \ dynamicProv: ENABLED\n  nfsServerName: fs\n  csi.storage.k8s.io/provisioner-secret-name:
          ntnx-secret\n  csi.storage.k8s.io/provisioner-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/node-publish-secret-name: ntnx-secret\n  csi.storage.k8s.io/node-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-expand-secret-name:
          ntnx-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace: openshift-cluster-csi-drivers\n
          \ storageType: NutanixFiles\n  squashType: \"none\"\n  #description: \"description
          added to each storage object created by the driver\"\nallowVolumeExpansion:
          true\n```"
        displayName: Nutanix CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Nutanix
        - CSI
        - Volumes
        - Files
        - Storage
        links:
        - name: Documentation
          url: https://portal.nutanix.com/page/documents/list?type=software&filterKey=product&filterVal=CSI
        maintainers:
        - email: cloudnative@nutanix.com
          name: Nutanix Cloud Native
        maturity: beta
        provider:
          name: Nutanix
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:9af9bf28430b00a0cedeb2ec29acadce45e6afcecd8bdf31c793c624cfa75fa7
        - registry.k8s.io/sig-storage/livenessprobe@sha256:82adbebdf5d5a1f40f246aef8ddbee7f89dea190652aefe83336008e69f9a89f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:e998f22243869416f9860fc6a1fb07d4202eac8846defc1b85ebd015c1207605
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:76a47d100de1f2afa23803bec503d9440605db03c7179b231f963e5663a3d12a
        - registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:859b6d1b2b36a48391e347c2a7431cafa08decafb56f801e21e544c3460a4b4b
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:2cddcc716c1930775228d56b0d2d339358647629701047edfdad5fcdfaf4ebcb
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:04c55b93a032ad8075cae540d60fc740a72c31006590dbdb61ad4ab9813959d0
        - registry.connect.redhat.com/nutanix/ntnx-csi-os@sha256:92962320ad682a7113aeadc8f9de1c7849996e06819acf00925aa8aa3bf51068
        version: 2.6.11
      entries:
      - name: nutanixcsioperator.v2.6.11
        version: 2.6.11
      name: stable
    - currentCSV: nutanixcsioperator.v3.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crd.nutanix.com/v1alpha1",
                "kind": "NutanixCsiStorage",
                "metadata": {
                  "name": "nutanixcsistorage"
                },
                "spec": {
                  "ntnxInitConfigMap": {
                    "usePC": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:9cc0ffc6e9c0a82320e207dad6b6b3ddc30ab8f83fa9e099a34450d5537ca435
          createdAt: "2025-06-18T11:04:03Z"
          description: An Operator for deploying and managing the Nutanix CSI Driver
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.3.4
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi","disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/nutanix/csi-operator
          support: Nutanix Support Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a new instance of Nutanix CSI Driver
            displayName: Nutanix CSI Storage
            kind: NutanixCsiStorage
            name: nutanixcsistorages.crd.nutanix.com
            version: v1alpha1
        description: "# Important\nPlease read the following instructions carefully
          before upgrading from 2.6.x to 3.3.4, for more information please refer
          to [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)
          \n- Please do not upgrade to the CSI 3.x operator if:\n    * You are using
          LVM volumes.\n    \n- To upgrade from the CSI 2.6.x to CSI 3.3 (interacting
          with Prism Central) operator\n    * Create a Nutanix Prism Central secret
          as explained below.\n    * Delete the csidriver object from the cluster:\n\n
          \       ```\n        oc delete csidriver csi.nutanix.com\n        ```\n\n
          \   * In the installed operators, go to Nutanix CSI Operator and change
          the subscription channel from stable to stable-3.x. \n    If you have installed
          the operator with automatic update approval, the operator will be automatically
          upgraded to CSI 3.3, and then the nutanixcsistorage resource will be upgraded.\n
          \   An update plan will be generated for manual updates. Upon approval,
          the operator will be successfully upgraded.\n    \n- Direct upgrades from
          CSI 2.6.x to CSI 3.3 interacting with Prism Element are not supported. \n
          \   The only solution is to recreate the nutanixcsistorage instance by following
          the below procedure:\n    - In the installed operators, go to Nutanix CSI
          Operator and delete the nutanixcsistorage instance.\n    - Next change the
          subscription channel from stable to stable-3.x.\n    - Verify the following
          points:\n        - Ensure a Nutanix Prism Element secret is present in the
          namespace.\n        - Ensure that all the storage classes with provisioner:
          csi.nutanix.com have a controller publish secret as explained below.\n\n
          \           ```\n            csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n            csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n            ```\n\n            If this secret
          is not present in the storage class please delete and recreate the storage
          classes with the required secrets.\n    - Create a new instance of nutanixcsistorage
          from this operator by specifying `usePC: false` in YAML spec section.\n
          \   - Caution: Moving from CSI driver interacting with Prism Central to
          CSI driver interacting with Prism Element is not supported.\n    \n- Troubleshooting:\n
          \   \n    If the upgrade was unsuccessful and you want to revert to version
          CSI 2.6.x, please delete the csidriver object as explained above, uninstall
          the operator (no need to delete the nutanixcsistorage custom resource),
          and install version CSI 2.6.x from the stable channel.\n# Overview \nThe
          Nutanix CSI Operator for Kubernetes packages, deploys, manages, and upgrades
          the Nutanix CSI Driver on Kubernetes and OpenShift for dynamic provisioning
          of persistent volumes on the Nutanix Enterprise Cloud platform.\nThe Nutanix
          Container Storage Interface (CSI) Driver for Kubernetes leverages Nutanix
          Volumes and Nutanix Files to provide scalable and persistent storage for
          stateful applications.\nWith the Nutanix CSI Provider, you can:\n*  **Provide
          persistent storage to your containers**\n    * Leverage PVC resources to
          consume dynamically Nutanix storage\n    * With Files storage classes, applications
          on multiple pods can access the same storage and have the benefit of multi-pod
          read and write access. \n## Warning\nIf you upgrade from a previous version,
          make sure to validate your existing StorageClass parameters against the
          corresponding [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)\n##
          Supported Access mode\n| Storage Class mode | ReadWriteOnce | ReadOnlyMany
          | ReadWriteMany |\n|--------------------|---------------|--------------|---------------|\n|
          Nutanix Volumes    | yes           | yes          | no            |\n| Nutanix
          Files      | yes           | yes          | yes           |\n## Supported
          orchestration platforms\nThe following table details orchestration platforms
          suitable for deployment of the Nutanix CSI driver.\n| Orchestration platform
          | Version | Architecture |\n|------------------------|---------|--------------|\n|
          Red Hat OpenShift      | 4.12    | x86          |\n| Red Hat OpenShift      |
          4.13    | x86          |\n| Red Hat OpenShift      | 4.14    | x86          |\n|
          Red Hat OpenShift      | 4.15    | x86          |\n| Red Hat OpenShift      |
          4.16    | x86          |\n| Red Hat OpenShift      | 4.17    | x86          |\n|
          Red Hat OpenShift      | 4.18    | x86          |\n| Red Hat OpenShift      |
          4.19    | x86          |\n## Configuring k8s secret and storage class\nTo
          use this driver, create the relevant storage classes and secrets and deploy
          your NutanixCsiStorage instance by following the [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)
          or the example in below section:\n**1. Create a Nutanix PE secret**\n```yaml\napiVersion:
          v1\nkind: Secret\nmetadata:\n  name: ntnx-pe-secret\n  namespace: openshift-cluster-csi-drivers\nstringData:\n
          \ # prism-ip:prism-port:username:password.\n  key: 1.2.3.4:9440:admin:password\n
          \ files-key: \"fileserver01.sample.com:csi:password1\"\n```\n**2. Create
          a Nutanix PC secret**\n```yaml\napiVersion: v1\nkind: Secret\nmetadata:\n
          \ name: ntnx-pc-secret\n  namespace: openshift-cluster-csi-drivers\nstringData:\n
          \ # prism-ip:prism-port:username:password.\n  key: 1.2.3.4:9440:admin:password\n```\n**3.
          Create a NutanixCsiStorage resource to deploy your driver**\nYou can do
          it directly inside the Operator UI with the `Create instance` button or
          with the following resource\n```yaml\napiVersion: crd.nutanix.com/v1alpha1\nkind:
          NutanixCsiStorage\nmetadata:\n  name: nutanixcsistorage\n  namespace: openshift-cluster-csi-drivers\nspec:
          \n  ntnxInitConfigMap:\n    usePC : true (false for PE based installations.
          Default is true.)\n```\n**4. Create storage classes**\n- volume mode example
          for 3.x installation\n```yaml\nkind: StorageClass\napiVersion: storage.k8s.io/v1\nmetadata:\n
          \ name: nutanix-volume\nprovisioner: csi.nutanix.com\nparameters:\n  csi.storage.k8s.io/provisioner-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/node-publish-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-expand-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/fstype: ext4\n  storageContainer:
          default-container\n  storageType: NutanixVolumes\n  #description: \"description
          added to each storage object created by the driver\"\n  #hypervisorAttached:
          ENABLED\nallowVolumeExpansion: true\nreclaimPolicy: Delete\n```\n- volume
          mode example for 2.x installation\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n  name: nutanix-volume\nprovisioner: csi.nutanix.com\nparameters:\n
          \ csi.storage.k8s.io/provisioner-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/node-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/controller-expand-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/fstype: ext4\n   storageContainer:
          default-container\n   storageType: NutanixVolumes\n   #description: \"description
          added to each storage object created by the driver\"\n   #isSegmentedIscsiNetwork:
          \"false\"\n   #whitelistIPMode: ENABLED\n   #chapAuth: ENABLED\n   #isLVMVolume:
          \"false\"\n   #numLVMDisks: 4\n allowVolumeExpansion: true\n reclaimPolicy:
          Delete\n```\n- dynamic files mode example\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n    name: nutanix-dynfiles\nprovisioner: csi.nutanix.com\nparameters:\n
          \ dynamicProv: ENABLED\n  nfsServerName: fs\n  csi.storage.k8s.io/provisioner-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/node-publish-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-expand-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  storageType: NutanixFiles\n  squashType:
          \"none\"\n  #description: \"description added to each storage object created
          by the driver\"\nallowVolumeExpansion: true\n```"
        displayName: Nutanix CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Nutanix
        - CSI
        - Volumes
        - Files
        - Storage
        links:
        - name: Documentation
          url: https://portal.nutanix.com/page/documents/list?type=software&filterKey=product&filterVal=CSI
        maintainers:
        - email: cloudnative@nutanix.com
          name: Nutanix Cloud Native
        maturity: beta
        provider:
          name: Nutanix
        relatedImages:
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:9af9bf28430b00a0cedeb2ec29acadce45e6afcecd8bdf31c793c624cfa75fa7
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:04c55b93a032ad8075cae540d60fc740a72c31006590dbdb61ad4ab9813959d0
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:9e581022a4e063b9d8b4f9466a36eabe4d0ced3badc8e353f13b700074ce4553
        - registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:9cc0ffc6e9c0a82320e207dad6b6b3ddc30ab8f83fa9e099a34450d5537ca435
        - registry.connect.redhat.com/nutanix/ntnx-csi-os@sha256:f0e191e7bf98829dbf718978fc971444436a1ecb9957ee22a5d8f9ba01a931b7
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0f64602ea791246712b51df334bbd701a0f31df9950a4cb9c28c059f367baa9e
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:f1bd6ee18c4021c1c94f29edfab89b49b6a4d1b800936c19dbef2d75f8202f2d
        - registry.connect.redhat.com/nutanix/ntnx-csi-os@sha256:9f121929227acf5581a813dc2f952105f517647c11a39cd7ea0863e768082586
        - registry.k8s.io/sig-storage/livenessprobe@sha256:82adbebdf5d5a1f40f246aef8ddbee7f89dea190652aefe83336008e69f9a89f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:3c116f543f0590aeff3299c8bb0683f250817d11a77d9e9071b15a0bffdabcd9
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:10624570c0aceb03f55f1eb07147b0c537e4676869cca2e9bd4bab113f810ac4
        - registry.k8s.io/sig-storage/csi-attacher@sha256:d7325367ab72b2d469a5091d87b4fc01142d2d13d1a28b2defbbe3e6fdbc4611
        version: 3.3.4
      entries:
      - name: nutanixcsioperator.v3.3.4
        version: 3.3.4
      name: stable-3.x
    defaultChannel: stable-3.x
    packageName: nutanixcsioperator
    provider:
      name: Nutanix
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/confidential-containers
    name: trustee-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trustee-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          containerImage: registry.redhat.io/build-of-trustee/trustee-rhel9-operator@sha256:fe8e23de14ad088c79b1bd5e7aab08a1b88a86679effcbf174be81e761bd9d6d
          createdAt: "2024-06-07T10:06:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trustee-operator-system
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/trustee-operator
          support: Confidential Containers Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KbsConfig is the Schema for the kbsconfigs API
            displayName: Kbs Config
            kind: KbsConfig
            name: kbsconfigs.confidentialcontainers.org
            version: v1alpha1
        description: Remote attestation operator for confidential containers
        displayName: Red Hat build of Trustee
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trustee
        - Confidential Computing
        - Attestation
        links:
        - name: Red Hat build of Trustee
          url: https://github.com/openshift/trustee-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
          url: https://github.com/confidential-containers
        relatedImages:
        - registry.redhat.io/build-of-trustee/trustee-rhel9-operator@sha256:fe8e23de14ad088c79b1bd5e7aab08a1b88a86679effcbf174be81e761bd9d6d
        version: 0.4.2
      entries:
      - name: trustee-operator.v0.4.2
        version: 0.4.2
      - name: trustee-operator.v0.4.1
        version: 0.4.1
      - name: trustee-operator.v0.4.0
        version: 0.4.0
      - name: trustee-operator.v0.3.0
        version: 0.3.0
      - name: trustee-operator.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: trustee-operator
    provider:
      name: Red Hat
      url: https://github.com/confidential-containers
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-keyspaces-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-keyspaces-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keyspaces.services.k8s.aws/v1alpha1",
                "kind": "Keyspace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "keyspaces.services.k8s.aws/v1alpha1",
                "kind": "Table",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/keyspaces-controller:1.1.1
          createdAt: "2025-09-20T01:23:26Z"
          description: AWS Keyspaces controller is a service controller for managing
            Keyspaces resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Keyspace represents the state of an AWS keyspaces Keyspace
              resource.
            displayName: Keyspace
            kind: Keyspace
            name: keyspaces.keyspaces.services.k8s.aws
            version: v1alpha1
          - description: Table represents the state of an AWS keyspaces Table resource.
            displayName: Table
            kind: Table
            name: tables.keyspaces.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Keyspaces resources in AWS from within your Kubernetes cluster.

          **About Amazon Keyspaces**

          Amazon Keyspaces (for Apache Cassandra) is a scalable, highly available, and managed Apache Cassandra–compatible database service. With Amazon Keyspaces, you don’t have to provision, patch, or manage servers, and you don’t have to install, maintain, or operate software.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Keyspaces
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keyspaces
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Keyspaces Developer Resources
          url: https://aws.amazon.com/keyspaces/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: keyspaces maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/keyspaces-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-keyspaces-controller.v1.1.1
        version: 1.1.1
      - name: ack-keyspaces-controller.v1.1.0
        version: 1.1.0
      - name: ack-keyspaces-controller.v1.0.14
        version: 1.0.14
      - name: ack-keyspaces-controller.v1.0.12
        version: 1.0.12
      - name: ack-keyspaces-controller.v1.0.11
        version: 1.0.11
      - name: ack-keyspaces-controller.v1.0.10
        version: 1.0.10
      - name: ack-keyspaces-controller.v1.0.9
        version: 1.0.9
      - name: ack-keyspaces-controller.v1.0.7
        version: 1.0.7
      - name: ack-keyspaces-controller.v1.0.6
        version: 1.0.6
      - name: ack-keyspaces-controller.v1.0.5
        version: 1.0.5
      - name: ack-keyspaces-controller.v1.0.3
        version: 1.0.3
      - name: ack-keyspaces-controller.v1.0.2
        version: 1.0.2
      - name: ack-keyspaces-controller.v1.0.1
        version: 1.0.1
      - name: ack-keyspaces-controller.v1.0.0
        version: 1.0.0
      - name: ack-keyspaces-controller.v0.0.16
        version: 0.0.16
      - name: ack-keyspaces-controller.v0.0.15
        version: 0.0.15
      - name: ack-keyspaces-controller.v0.0.14
        version: 0.0.14
      - name: ack-keyspaces-controller.v0.0.13
        version: 0.0.13
      - name: ack-keyspaces-controller.v0.0.12
        version: 0.0.12
      - name: ack-keyspaces-controller.v0.0.11
        version: 0.0.11
      - name: ack-keyspaces-controller.v0.0.10
        version: 0.0.10
      - name: ack-keyspaces-controller.v0.0.9
        version: 0.0.9
      - name: ack-keyspaces-controller.v0.0.8
        version: 0.0.8
      - name: ack-keyspaces-controller.v0.0.7
        version: 0.0.7
      - name: ack-keyspaces-controller.v0.0.6
        version: 0.0.6
      - name: ack-keyspaces-controller.v0.0.5
        version: 0.0.5
      - name: ack-keyspaces-controller.v0.0.4
        version: 0.0.4
      - name: ack-keyspaces-controller.v0.0.3
        version: 0.0.3
      - name: ack-keyspaces-controller.v0.0.2
        version: 0.0.2
      - name: ack-keyspaces-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-keyspaces-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Carlos Eduardo de Paula
      provider-url: ""
    name: lbconfig-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lbconfig-operator.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-infra-sample",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "http",
                    "path": "/healthz",
                    "port": 1936
                  },
                  "ports": [
                    80,
                    443
                  ],
                  "provider": {
                    "creds": "f5-creds",
                    "host": "https://192.168.1.35",
                    "partition": "Common",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "F5_BigIP"
                  },
                  "type": "infra",
                  "vip": "192.168.1.45"
                }
              },
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-infra-sample-shard",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "http",
                    "path": "/healthz",
                    "port": 1936
                  },
                  "nodelabels": {
                    "kubernetes.io/region": "DC1",
                    "node-role.kubernetes.io/infra": ""
                  },
                  "ports": [
                    80,
                    443
                  ],
                  "provider": {
                    "creds": "f5-creds",
                    "host": "https://192.168.1.35",
                    "partition": "Common",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "F5_BigIP"
                  },
                  "type": "infra",
                  "vip": "192.168.1.46"
                }
              },
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-master-sample",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "https",
                    "path": "/healthz",
                    "port": 6443
                  },
                  "ports": [
                    6443
                  ],
                  "provider": {
                    "creds": "netscaler-creds",
                    "host": "https://192.168.1.36",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "Citrix_ADC"
                  },
                  "type": "master",
                  "vip": "192.168.1.40"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Networking
          certified: "false"
          containerImage: quay.io/carlosedp/lbconfig-operator:v0.5.1
          createdAt: "2025-01-09T15:45:28Z"
          description: Manage External Load Balancers allowing creation/update for
            VIPs and Servers dynamically via API.
          k8sMaxVersion: ""
          k8sMinVersion: ""
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/carlosedp/lbconfig-operator
          support: https://github.com/carlosedp/lbconfig-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalLoadBalancer is the Schema for the externalloadbalancers
              API
            displayName: ExternalLoadBalancer Instance
            kind: ExternalLoadBalancer
            name: externalloadbalancers.lb.lbconfig.carlosedp.com
            version: v1
          required:
          - description: ExternalLoadBalancer represents a configured instance of
              an external Load-Balancer for a specific group of nodes of the cluster.
              The Instance has a VIP and ports to be balanced to the cluster nodes
              based on a set of node labels.
            displayName: External Load-Balancer Configuration Instance
            kind: ExternalLoadBalancer
            name: externalloadbalancers.lb.lbconfig.carlosedp.com
            version: v1
        description: |
          ## About the Operator

          The LBConfig Operator, manages the configuration of External Load Balancer instances (on third-party equipment via it's API) and creates VIPs and IP Pools with Monitors for a set of OpenShift or Kubernetes nodes like Master-nodes (Control-Plane), Infra nodes (where the Routers or Ingress controllers are located) or based on it's roles and/or labels.

          The operator dynamically handles creating, updating or deleting the IPs of the pools in the Load Balancer based on the Node IPs for each role or label. On every change of the operator configuration (CRDs) or addition/change/removal or cluster Nodes, the operator updates the Load Balancer properly.

          ## Main Use Cases

          The main users for this operator is enterprise deployments or clusters composed of multiple nodes having an external load-balancer providing the balancing and high-availability to access the cluster in both API and Application levels.

          ### Create ExternalLoadBalancer Instances

          Create the instances for each Load Balancer instance you need (for example one for Master Nodes and another for the Infra Nodes). **If installing on OpenShift or Kubernetes with OLM (or in a different namespace), adjust the sample YAMLs to match the created namespace**.

          **The provider `vendor` field can be (case-insensitive):**

          * **`F5_BigIP`** - Tested on F5 BigIP version 15
          * **`Citrix_ADC`** - Tested on Citrix ADC (Netscaler) version 13
          * **`HAProxy`** - HAProxy with Dataplane API. ([Docs](./docs/haproxy/))
          * **`Dummy`** - Dummy backend used for testing to only print log messages on operations

          Create the secret holding the Load Balancer API user and password:

          ```sh
          oc create secret generic f5-creds --from-literal=username=admin --from-literal=password=admin123 --namespace lbconfig-operator-system
          ```

          #### Sample CRDs and Available Fields

          Master Nodes using a Citrix ADC LB:

          ```yaml
          apiVersion: lb.lbconfig.carlosedp.com/v1
          kind: ExternalLoadBalancer
          metadata:
            name: externalloadbalancer-master-sample
            namespace: lbconfig-operator-system
          spec:
            vip: "192.168.1.40"
            type: "master"
            ports:
              - 6443
            monitor:
              path: "/healthz"
              port: 6443
              monitortype: "https"
            provider:
              vendor: Citrix_ADC
              host: "https://192.168.1.36"
              port: 443
              creds: netscaler-creds
              validatecerts: false
          ```

          Infra Nodes using a F5 BigIP LB:

          ```yaml
          apiVersion: lb.lbconfig.carlosedp.com/v1
          kind: ExternalLoadBalancer
          metadata:
            name: externalloadbalancer-infra-sample
            namespace: lbconfig-operator-system
          spec:
            vip: "192.168.1.45"
            type: "infra"
            ports:
              - 80
              - 443
            monitor:
              path: "/healthz"
              port: 1936
              monitortype: http
            provider:
              vendor: F5_BigIP
              host: "https://192.168.1.35"
              port: 443
              creds: f5-creds
              partition: "Common"
              validatecerts: false
          ```

          To choose the nodes which will be part of the server pool, you can set either `type` or `nodelabels` fields. The yaml field `type: "master"` or `type: "infra"` selects nodes with the role label `"node-role.kubernetes.io/master"` and `"node-role.kubernetes.io/infra"` respectively. If the field `nodelabels` array is used instead, the operator will use nodes which match all labels.

          Clusters with sharded routers or using arbitrary labels to determine where the Ingress Controllers run can be configured like:

          ```yaml
          spec:
            vip: "10.0.0.6"
            ports:
              - 80
            nodelabels:
              "node.kubernetes.io/ingress-controller": "production"
              "kubernetes.io/region": "DC1"
            ...
          ```

          ## Getting Help

          This project fully open-source and is hosted at [GitHub](https://github.com/carlosedp/lbconfig-operator). Always check the project page and issues for the latest updates and/or solutions to problems encountered. If a problem is detected, please fill an issue. Feature request and improvement issues are welcome.

          Some fields inside `providers` are optional and depend on the used backend. Check the [API docs](https://pkg.go.dev/github.com/carlosedp/lbconfig-operator/api/lb.lbconfig.carlosedp.com/v1?utm_source=gopls#Provider) which fields are backend-specific.

          ## Disclaimers

          * The operator does not check if the requested configuration (names, IPs) already exists and/or conflicts with existing configuration in the Load Balancer. The user is responsible for these checks before deployment;
          * I am not responsible if the operator changes/deletes existing configuration on the Load Balancer if existing names are already configured.
          * The operator creates the entries(Pools, VIPs, Monitors) in the provided Load Balancer with the `name` of the instance configured in the CustomResource prefixed with the type. Eg. For a CR with name `externalloadbalancer-master-sample`, the operator creates a server pool named `Pool-externalloadbalancer-master-sample-6443` (suffixed with the port), a monitor named `Monitor-externalloadbalancer-master-sample` and a VIP named `VIP-externalloadbalancer-master-sample-6443` (suffixed with the port).
        displayName: External Load-Balancer Configuration Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 'load-balance '
        - infrastructure
        links:
        - name: GitHub
          url: https://github.com/carlosedp/lbconfig-operator
        - name: Documentation
          url: https://github.com/carlosedp/lbconfig-operator/docs
        - name: Issues
          url: https://github.com/carlosedp/lbconfig-operator/issues
        maintainers:
        - email: carlosedp@gmail.com
          name: Carlos Eduardo de Paula
        maturity: beta
        minKubeVersion: 1.18.0
        provider:
          name: Carlos Eduardo de Paula
        relatedImages:
        - quay.io/carlosedp/kube-rbac-proxy:v0.15.0
        - quay.io/carlosedp/lbconfig-operator:v0.5.1
        version: 0.5.1
      entries:
      - name: lbconfig-operator.v0.5.1
        version: 0.5.1
      - name: lbconfig-operator.v0.5.0
        version: 0.5.0
      - name: lbconfig-operator.v0.4.1
        version: 0.4.1
      - name: lbconfig-operator.v0.4.0
        version: 0.4.0
      - name: lbconfig-operator.v0.2.0
        version: 0.2.0
      name: beta
    defaultChannel: beta
    packageName: lbconfig-operator
    provider:
      name: Carlos Eduardo de Paula
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: group-sync-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: group-sync-operator.v0.0.36
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "azure-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "azure": {
                        "credentialsSecret": {
                          "name": "azure-group-sync",
                          "namespace": "group-sync-operator"
                        }
                      },
                      "name": "azure"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "github-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "github": {
                        "credentialsSecret": {
                          "name": "github-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "organization": "redhat-cop"
                      },
                      "name": "github"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "gitlab-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "gitlab": {
                        "credentialsSecret": {
                          "name": "gitlab-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "url": "https://gitlab.apps.openshift.com"
                      },
                      "name": "gitlab"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "keycloak-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "keycloak": {
                        "credentialsSecret": {
                          "name": "keycloak-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "insecure": true,
                        "realm": "ocp",
                        "url": "https://keycloak.apps.openshift.com"
                      },
                      "name": "keycloak"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "ldap-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "ldap": {
                        "credentialsSecret": {
                          "name": "ldap-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "insecure": true,
                        "rfc2307": {
                          "groupMembershipAttributes": [
                            "member"
                          ],
                          "groupNameAttributes": [
                            "cn"
                          ],
                          "groupUIDAttribute": "dn",
                          "groupsQuery": {
                            "baseDN": "ou=Groups,dc=example,dc=com",
                            "derefAliases": "never",
                            "filter": "(objectClass=groupofnames)",
                            "scope": "sub"
                          },
                          "tolerateMemberNotFoundErrors": true,
                          "tolerateMemberOutOfScopeErrors": true,
                          "userNameAttributes": [
                            "cn"
                          ],
                          "userUIDAttribute": "dn",
                          "usersQuery": {
                            "baseDN": "ou=Users,dc=example,dc=com",
                            "derefAliases": "never",
                            "scope": "sub"
                          }
                        },
                        "url": "ldap://localhost:389"
                      },
                      "name": "ldap"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "okta-sync"
                },
                "spec": {
                  "providers": [
                    {
                      "name": "okta",
                      "okta": {
                        "appId": "okta-sync-app-id",
                        "credentialsSecret": {
                          "name": "okta-api-token",
                          "namespace": "group-sync-operator"
                        },
                        "extractLoginUsername": true,
                        "profileKey": "login",
                        "url": "https://example.okta.com/"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/group-sync-operator@sha256:43ef79efc2a77ee5edd4e3123f9dab156a30d3e00fd1fba9f89600b96f0f68d6
          createdAt: "2025-08-12T19:19:21Z"
          description: Synchronize groups and users from external providers
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: group-sync-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/group-sync-operator
          support: Red Hat Community of Practice
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GroupSync is the Schema for the groupsyncs API
            displayName: Group Sync
            kind: GroupSync
            name: groupsyncs.redhatcop.redhat.io
            version: v1alpha1
        description: |
          Synchronizes groups from external providers into OpenShift

          ## Overview

          The OpenShift Container Platform contains functionality to synchronize groups found in external identity providers into the platform. Currently, the functionality that is included in OpenShift to limited to synchronizing LDAP only. This operator is designed to integrate with external providers in order to provide new solutions.

          Group Synchronization is facilitated by creating a `GroupSync` resource. The following describes the high level schema for this resource:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: example-groupsync
          spec:
            providers:
              - <One or more providers to synchronize>
          ```

          ## Authentication

          In most cases, authentication details must be provided in order to communicate with providers. Authentication details are provider specific with regards to the required values. In supported providers, the secret can be referenced in the `credentialsSecret` by name and namespace where it has been created as shown below:

          ```
          credentialsSecret:
            name: <secret_name>
            namespace: <secret_namespace>
          ```

          ## Providers

          Integration with external systems is made possible through a set of pluggable external providers. The following providers are currently supported:

          * [Azure](https://azure.microsoft.com/)
          * [GitHub](https://github.com)
          * [GitLab](https://gitlab.com)
          * [LDAP](https://en.wikipedia.org/wiki/Lightweight_Directory_Access_Protocol)
          * [Keycloak](https://www.keycloak.org/)/[Red Hat Single Sign On](https://access.redhat.com/products/red-hat-single-sign-on)
          * [Okta](https://www.okta.com/)
          * [IBM Security Verify](https://docs.verify.ibm.com/verify)

          The following sections describe the configuration options available for each provider


          ### Azure

          Groups contained within Azure Active Directory can be synchronized into OpenShift. The following table describes the set of configuration options for the Azure provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `authorityHost` | Azure Active Directory Endpoint | `https://login.microsoftonline.com` | No |
          | `baseGroups` | List of groups to start searching from instead of listing all groups in the directory | | No |
          | `credentialsSecret` | Name of the secret containing authentication details (See below) | | Yes |
          | `filter` | Graph API filter | | No |
          | `groups` | List of groups to filter against | | No |
          | `userNameAttributes` | Fields on a user record to use as the User Name | `userPrincipalName` | No |
          | `prune` | Prune Whether to prune groups that are no longer in Azure | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Azure provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: azure-groupsync
          spec:
            providers:
            - name: azure
              azure:
                credentialsSecret:
                  name: azure-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to Azure

          Authentication to Azure can be performed using Application Registration with access to query group information in Azure Active Directory.

          The App Registration must be granted access to the following Microsoft Graph API's:

          * Group.Read.All
          * GroupMember.Read.All
          * User.Read.All

          A secret must be created in the same namespace that contains the `GroupSync` resource:

          The following keys must be defined in the secret

          * `AZURE_TENANT_ID` - Tenant ID
          * `AZURE_CLIENT_ID` - Client ID
          * `AZURE_CLIENT_SECRET` - Client Secret

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic azure-group-sync --from-literal=AZURE_SUBSCRIPTION_ID=<AZURE_SUBSCRIPTION_ID> --from-literal=AZURE_TENANT_ID=<AZURE_TENANT_ID> --from-literal=AZURE_CLIENT_ID=<AZURE_CLIENT_ID> --from-literal=AZURE_CLIENT_SECRET=<AZURE_CLIENT_SECRET>
          ```

          ### GitHub

          Teams stored within a GitHub organization can be synchronized into OpenShift. The following table describes the set of configuration options for the GitHub provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `caSecret` | Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `organization` | Organization to synchronize against | | Yes |
          | `teams` | List of teams to filter against | | No |
          | `url` | Base URL for the GitHub or GitHub Enterprise host (Must contain a trailing slash) | | No |
          | `prune` | Prune Whether to prune groups that are no longer in GitHub | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Github provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: github-groupsync
          spec:
            providers:
            - name: github
              github:
                organization: ocp
                credentialsSecret:
                  name: github-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to GitHub

          Authentication to GitLab can be performed using a [Token](https://docs.gitlab.com/ee/security/token_overview.html) or a Username and Password (Note: 2FA not supported). A secret must be created in the same namespace that contains the `GroupSync` resource:

          When using an OAuth token, the following token types are supported:

          * Personal Access Token
          * OAuth Token
          * Job Token

          the following key is required:

          * `token` - OAuth token

          Optionally, the `tokenType` key can be specified to indicate the type of token being provided from the following values:

          * OAuth - `oauth`
          * Personal Access Token - `personal`
          * Job Token - `job`

          If no `tokenType` is provided, `oauth` is used by default

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token>
          ```

          To specify a token type, such as a Personal Access Token, the following command can be executed:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token> --from-literal=tokenType=personal
          ```

          The following keys are required for username and password:

          * `username` - Username for authenticating with GitHub
          * `password` - Password for authenticating with GitHub

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic github-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### GitLab

          Groups stored within a GitLab can be synchronized into OpenShift. The following table describes the set of configuration options for the GitLab provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `groups` | List of groups to filter against | | No |
          | `prune` | Prune Whether to prune groups that are no longer in GitLab | `false` | No |
          | `scope` | Scope for group synchronization. Options are `one` for one level or `sub` to include subgroups | `sub` | No |
          | `url` | Base URL for the GitLab instance | `https://gitlab.com` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Github provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: gitlab-groupsync
          spec:
            providers:
            - name: gitlab
              gitlab:
                credentialsSecret:
                  name: gitlab-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to GitLab

          Authentication to GitLab can be performed using an OAuth Personal Access Token or a Username and Password (Note: 2FA not supported). A secret must be created in the same namespace that contains the `GroupSync` resource:

          When using an OAuth token, the following key is required:

          * `token` - OAuth token

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token>
          ```

          The following keys are required for username and password:

          * `username` - Username for authenticating with GitLab
          * `password` - Password for authenticating with GitLab

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### LDAP

          Groups stored within an [LDAP](https://en.wikipedia.org/wiki/Lightweight_Directory_Access_Protocol) server can be synchronized into OpenShift. The LDAP provider implements the included features of the [Syncing LDAP groups](https://docs.openshift.com/container-platform/latest/authentication/ldap-syncing.html) feature and makes use of the libraries from the [OpenShift Command Line](https://docs.openshift.com/container-platform/latest/cli_reference/openshift_cli/getting-started-cli.html) tool to streamline the migration to this operator based implementation.

          The configurations of the three primary schemas (`rfc2307`, `activeDirectory` and `augmentedActiveDirectory`) can be directly migrated as is without any modification.

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | No |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `groupUIDNameMapping` | User defined name mapping | | No |
          | `rfc2307` | Configuration using the [rfc2307](https://docs.openshift.com/container-platform/latest/authentication/ldap-syncing.html#ldap-syncing-rfc2307_ldap-syncing-groups) schema | | No |
          | `activeDirectory` | Configuration using the [activeDirectory](https://docs.openshift.com/container-platform/4.5/authentication/ldap-syncing.html#ldap-syncing-activedir_ldap-syncing-groups) schema | | No |
          | `augmentedActiveDirectory` | Configuration using the [activeDirectory](https://docs.openshift.com/container-platform/4.5/authentication/ldap-syncing.html#ldap-syncing-augmented-activedir_ldap-syncing-groups) schema | | No |
          | `url` | Connection URL for the LDAP server | `ldap://ldapserver:389` | No |
          | `whitelist` | Explicit list of groups to synchronize |  | No |
          | `blacklist` | Explicit list of groups to not synchronize |  | No |
          | `prune` | Prune Whether to prune groups that are no longer in LDAP | `false` | No |

          The following is an example using the `rfc2307` schema:

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
                credentialsSecret:
                  name: ldap-group-sync
                  namespace: group-sync-operator
                insecure: true
                rfc2307:
                  groupMembershipAttributes:
                  - member
                  groupNameAttributes:
                  - cn
                  groupUIDAttribute: dn
                  groupsQuery:
                    baseDN: ou=Groups,dc=example,dc=com
                    derefAliases: never
                    filter: (objectClass=groupofnames)
                    scope: sub
                  tolerateMemberNotFoundErrors: true
                  tolerateMemberOutOfScopeErrors: true
                  userNameAttributes:
                  - cn
                  userUIDAttribute: dn
                  usersQuery:
                    baseDN: ou=Users,dc=example,dc=com
                    derefAliases: never
                    scope: sub
                url: ldap://ldapserver:389
              name: ldap
          ```

          The examples provided in the OpenShift documented referenced previously can be used to construct the schemas for the other LDAP synchronization types.

          #### Authenticating to LDAP

          If authentication is required in order to communicate with the LDAP server, a secret should be created in the same namespace that contains the `GroupSync` resource. The following keys can be defined:

          * `username` - Username (Bind DN) for authenticating with the LDAP server
          * `password` - Password for authenticating with the LDAP server

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic ldap-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          #### Whitelists and Blacklists

          Groups can be explicitly whitelisted or blacklisted in order to control the groups that are eligible to be synchronized into OpenShift. When running LDAP group synchronization using the command line, this configuration is referenced via separate files, but these are instead specified in the `blacklist` and `whitelist` properties as shown below:

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
          ...
                whitelist:
                - cn=Online Corporate Banking,ou=Groups,dc=example,dc=com
          ...
              name: ldap
          ```

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
          ...
                blacklist:
                - cn=Finance,ou=Groups,dc=example,dc=com
          ...
              name: ldap
          ```

          ### Keycloak

          Groups stored within Keycloak can be synchronized into OpenShift. The following table describes the set of configuration options for the Keycloak provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `groups` | List of groups to filter against | | No |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `loginRealm` | Realm to authenticate against | `master` | No |
          | `realm` | Realm to synchronize | | Yes |
          | `scope` | Scope for group synchronization. Options are `one` for one level or `sub` to include subgroups | `sub` | No |
          | `url` | Base URL for the Keycloak server. Older versions (<17.0.0) including Red Hat SSO should include the context path `/auth` appended to the hostname  | | Yes |
          | `prune` | Prune Whether to prune groups that are no longer in Keycloak | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Keycloak provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                url: https://keycloak-keycloak-operator.apps.openshift.com/auth
          ```

          #### Authenticating to Keycloak

          A user with rights to query for Keycloak groups must be available. The following permissions must be associated to the user:

          * Password must be set (Temporary option unselected) on the _Credentials_ tab
          * On the _Role Mappings_ tab, select _master-realm_ or _realm-management_ next to the _Client Roles_ dropdown and then select **Query Groups** and **Query Users**.

          A secret must be created in the same namespace that contains the `GroupSync` resource. It must contain the following keys for the user previously created:

          * `username` - Username for authenticating with Keycloak
          * `password` - Password for authenticating with Keycloak

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic keycloak-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### Okta

          [Okta Groups](https://help.okta.com/en/prod/Content/Topics/users-groups-profiles/usgp-main.htm) assigned to [Okta Applications](https://help.okta.com/en/prod/Content/Topics/Apps/Apps_Apps.htm) can be synchronized into OpenShift. The developer docs for the Okta API that the Okta Syncer uses can be found [here](https://developer.okta.com/docs/reference/api/apps/#list-groups-assigned-to-application).
          The following table describes the set of configuration options for the Okta provider:

            | Name | Description | Defaults | Required |
            | ----- | ---------- | -------- | ----- |
            | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
            | `groups` | List of groups to filter against | | No |
            | `url` | Okta organization URL (Must contain a trailing slash) | | Yes |
            | `appId` | Application ID of App Groups are assigned to | | Yes |
            | `extractLoginUsername` | Bool to determine if you should extract username from okta login | | No |
            | `profileKey` | Attribute field on Okta User Profile you would like to use as identity | `login` | No |
            | `groupLimit` | Integer to set the maximum number of groups to retrieve from OKTA per request. | `1000` | No |
            | `prune` | Prune Whether to prune groups that are no longer in OKTA | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with an Okta provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: okta-sync
          spec:
            providers:
              - name: okta
                okta:
                  credentialsSecret:
                    name: okta-api-token
                    namespace: group-sync-operator
                  url: "https://example.okta.com/"
                  appId: okta-sync-app-id
          ```

          #### Authenticating to Okta

          A secret must be created in the same namespace as the group-sync-operator pod. It must contain the following key:

            * `okta-api-token` - Okta API Token for interacting with Okta

          The secret can be created by executing the following command:

            ```shell
            oc create secret generic okta-api-token --from-literal=okta-api-token=<OKTA_API_TOKEN> -n group-sync-operator
            ```

          ### IBM Security Verify

          Groups defined in [IBM Security Verify](https://help.okta.com/en/prod/Content/Topics/users-groups-profiles/usgp-main.htm) (ISV) can be synchronized into OpenShift. Currently only the `userName` field from ISV will be synchronized. The developer docs for the ISV API can be found [here](https://docs.verify.ibm.com/verify/page/api-documentation).
          The following table describes the set of configuration options for the provider:

            | Name | Description | Defaults | Required |
            | ----- | ---------- | -------- | ----- |
            | `credentialsSecret` | Reference to a secret containing authentication details (see below) | `''`  | Yes |
            | `groups` | List of groups to synchronize (see below) | `nil`  | Yes |
            | `tenantUrl` | The ISV tenant URL, for example `https://my-isv.verify.ibm.com`) | `''`  | Yes |

          The following is an example of a minimal configuration that can be applied to integrate with an Okta provider:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ibmsecurityverify-sync
          spec:
              providers:
              - name: ibmsecurityverify
                ibmsecurityverify:
                  credentialsSecret:
                    name: isv-group-sync
                    namespace: group-sync-operator
                  tenantUrl: https://my-isv.verify.ibm.com
                  groups:
                    - name: 'application owners'
                      id: 645001V3V9
                    - name: developer
                      id: 645001V3VA
          ```

          #### Group Objects
          Each group object in the `groups` array must contain an `id` field. The group ID can be retrieved by pulling the group information from the ISV API. Optionally, the object may also contain a `name` which corresponds to the group's display name. When defined, the operator will confirm that the name defined in the YAML matches that received from the API when synchronization occurs; as the group IDs are not human-friendly, using the name can confirm the correct groups are configured. If the names do not match an error will be logged.

          #### Group Names
          The name of each groups created in OpenShift will match the group name in ISV. Any whitespace in the ISV group name will be replaced with a hyphen.

          #### Authenticating to IBM Security Verify

          A secret must be created in the same namespace as the group-sync-operator pod. It must contain the following keys:

          * `clientId` - The API client ID.
          * `clientSecret`- The API client secret.

          See the IBM Security Verify [API documentation](https://docs.verify.ibm.com/verify/docs/api-access) for setting up authentication.

          ### Support for Additional Metadata (Beta)

          Additional metadata based on Keycloak group are also added to the OpenShift groups as Annotations including:

          * Parent/child relationship between groups and their subgroups
          * Group attributes

          ## CA Certificates

          Several providers allow for certificates to be provided in either a _ConfigMap_ or _Secret_ to communicate securely to the target host through the use of a property called `ca`.

          The certificate can be added to a Secret called _keycloak-certs_ using the key `ca.crt` representing the certificate using the following command.

          ```
          oc create secret generic keycloak-certs --from-file=ca.crt=<file>
          ```

          An example of how the CA certificate can be added to the Keycloak provider is shown below:


          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                caSecret:
                  kind: Secret
                  name: keycloak-certs
                  namespace: group-sync-operator
                  key: ca.crt
                url: https://keycloak-keycloak-operator.apps.openshift.com
          ```

          Alteratively, a _ConfigMap_ can be used instead instead of a _Secret_. This is useful when using the [Certificate injection using Operators](https://docs.openshift.com/container-platform/latest/networking/configuring-a-custom-pki.html#certificate-injection-using-operators_configuring-a-custom-pki) feature.

          The following command can be used to create a _ConfigMap_ containing the certificate:

          ```
          oc create configmap keycloak-certs --from-file=ca.crt=<file>
          ````

          An example of how the CA certificate can be added to the Keycloak provider is shown below:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                caSecret:
                  kind: ConfigMap
                  name: keycloak-certs
                  namespace: group-sync-operator
                  key: ca.crt
                url: https://keycloak-keycloak-operator.apps.openshift.com
          ```

          ## Scheduled Execution

          A cron style expression can be specified for which a synchronization event will occur. The following specifies that a synchronization should occur nightly at 3AM

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            schedule: "0 3 * * *"
            providers:
            - ...
          ```

          If a schedule is not provided, synchronization will occur only when the object is reconciled by the platform.

          ## Metrics

          Prometheus compatible metrics are exposed by the Operator and can be integrated into OpenShift's default cluster monitoring. To enable OpenShift cluster monitoring, label the namespace the operator is deployed in with the label `openshift.io/cluster-monitoring="true"`.

          ```
          oc label namespace <namespace> openshift.io/cluster-monitoring="true"
          ```
        displayName: Group Sync Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - group
        - user
        - security
        - authorization
        links:
        - name: Group Sync Operator
          url: https://github.com/redhat-cop/group-sync-operator
        maintainers:
        - email: ablock@redhat.com
          name: Andrew Block
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:b8a174be91f3df75dff9656d4c72de28af51a731008e9e96db5eef3c7af58aac
        - quay.io/redhat-cop/group-sync-operator@sha256:43ef79efc2a77ee5edd4e3123f9dab156a30d3e00fd1fba9f89600b96f0f68d6
        version: 0.0.36
      entries:
      - name: group-sync-operator.v0.0.36
        version: 0.0.36
      - name: group-sync-operator.v0.0.35
        version: 0.0.35
      - name: group-sync-operator.v0.0.34
        version: 0.0.34
      - name: group-sync-operator.v0.0.33
        version: 0.0.33
      - name: group-sync-operator.v0.0.32
        version: 0.0.32
      - name: group-sync-operator.v0.0.31
        version: 0.0.31
      - name: group-sync-operator.v0.0.30
        version: 0.0.30
      - name: group-sync-operator.v0.0.29
        version: 0.0.29
      - name: group-sync-operator.v0.0.28
        version: 0.0.28
      - name: group-sync-operator.v0.0.27
        version: 0.0.27
      - name: group-sync-operator.v0.0.26
        version: 0.0.26
      - name: group-sync-operator.v0.0.25
        version: 0.0.25
      - name: group-sync-operator.v0.0.24
        version: 0.0.24
      - name: group-sync-operator.v0.0.23
        version: 0.0.23
      - name: group-sync-operator.v0.0.22
        version: 0.0.22
      - name: group-sync-operator.v0.0.21
        version: 0.0.21
      - name: group-sync-operator.v0.0.20
        version: 0.0.20
      - name: group-sync-operator.v0.0.19
        version: 0.0.19
      - name: group-sync-operator.v0.0.18
        version: 0.0.18
      - name: group-sync-operator.v0.0.17
        version: 0.0.17
      - name: group-sync-operator.v0.0.16
        version: 0.0.16
      - name: group-sync-operator.v0.0.15
        version: 0.0.15
      - name: group-sync-operator.v0.0.14
        version: 0.0.14
      - name: group-sync-operator.v0.0.13
        version: 0.0.13
      name: alpha
    defaultChannel: alpha
    packageName: group-sync-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubecost
      provider-url: https://www.kubecost.com
    name: kubecost-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubecost-operator.v2.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kubecost.com/v1alpha1",
                "kind": "CostAnalyzer",
                "metadata": {
                  "name": "costanalyzer-sample"
                },
                "spec": {
                  "clusterController": {
                    "image": {
                      "repository": "registry.connect.redhat.com/kubecost/kubecost-cluster-controller@sha256",
                      "tag": "1a6f11dcd7341ab3f24eba23ec06c7b75646a534a7ed960153601806425469d1"
                    }
                  },
                  "forecasting": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-modeling@sha256:8f59f8fd8696964c1b5322ca79fae78befe97323fa56c2b509bdcc67ee4c0edd"
                  },
                  "global": {
                    "platforms": {
                      "openshift": {
                        "createMonitoringClusterRoleBinding": true,
                        "createMonitoringResourceReaderRoleBinding": true,
                        "enabled": true,
                        "scc": {
                          "clusterController": true,
                          "networkCosts": true,
                          "nodeExporter": true
                        }
                      }
                    },
                    "prometheus": {
                      "enabled": false,
                      "fqdn": "https://prometheus-k8s.openshift-monitoring.svc.cluster.local:9091",
                      "kubeRBACProxy": true
                    }
                  },
                  "grafana": {
                    "image": {
                      "repository": "registry.redhat.io/rhel9/grafana@sha256",
                      "tag": "9f71de028c18034e07a8d15dac13ef68afb91cb899b9e8e7d39703547dd0eac1"
                    },
                    "sidecar": {
                      "image": {
                        "repository": "quay.io/kiwigrid/k8s-sidecar@sha256",
                        "tag": "d3749777d5bedf62549b452a7e2ccb205da53588dc7f04927efc75d0da5b6d23"
                      }
                    }
                  },
                  "kubecostAggregator": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef"
                  },
                  "kubecostFrontend": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-frontend@sha256:f9b2196dc6ac017b8074f8fe40f1a58e7a2630282c5b7a019ca9f3dbdb670dc8"
                  },
                  "kubecostModel": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef"
                  },
                  "networkCosts": {
                    "image": {
                      "repository": "registry.connect.redhat.com/kubecost/kubecost-network-costs@sha256",
                      "tag": "c444d475be4f1aa36680bd84358e1e07303dcfe5d13a95544ad9e59f80d4a5f0"
                    }
                  },
                  "prometheusRule": {
                    "enabled": true
                  },
                  "serviceMonitor": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          createdAt: "2025-08-08T19:34:34Z"
          description: Deploys Kubecost 2.8.2 via the Helm chart.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Contact Kubecost for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Kubecost.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The CostAnalyzer API describes a deployment of Kubecost and
              its dependencies.
            displayName: CostAnalyzer
            kind: CostAnalyzer
            name: costanalyzers.charts.kubecost.com
            version: v1alpha1
        description: |2-

          IMPORTANT - Before installation, you must run the following command to allow monitoring of the Kubecost namespace by the local OpenShift Promtheus. If you install in a namespace other than "Kubecost", adjust accordingly.

          `oc label namespace kubecost openshift.io/cluster-monitoring=true`

          Kubecost provides real-time cost visibility and insights for teams using Kubernetes, helping you continuously reduce your cloud costs.

          This OpenShift Operator deploys Kubecost from the official Helm chart.

          Kubecost allows you to:

          * See allocated spend across all native Kubernetes concepts and provide your teams with transparent, accurate cost data reconciled with your actual cloud bill.
          * Join in-cluster costs like CPU and memory with out-of-cluster spend from other cloud services.
          * Get complete cost visibility across OpenShift, On-Prem, AWS, GCP, Azure, Oracle, AliCloud + more.
          * Gain insights you can use to save 30-50%+ on your infrastructure spend, without exposing your private information — your data never gets shared externally, even with us.
          * Have real-time alerts and recurring reports, empowering teams to take control of their Kubernetes-enabled infrastructure and stay within budgeted limits.

          Visit the Kubecost website for more details including features, pricing, documentation, and more at www.kubecost.com.
        displayName: Kubecost
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kubernetes
        - cost
        - kubernetes costs
        - kubernetes optimization
        - k8s
        - cost monitoring
        - cloud costs
        - k8s cost
        links:
        - name: Kubecost
          url: https://www.kubecost.com
        maintainers:
        - email: support@kubecost.com
          name: Kubecost Support
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Kubecost
          url: https://www.kubecost.com
        relatedImages:
        - registry.connect.redhat.com/kubecost/kubecost-network-costs@sha256:c444d475be4f1aa36680bd84358e1e07303dcfe5d13a95544ad9e59f80d4a5f0
        - registry.redhat.io/rhel9/grafana@sha256:9f71de028c18034e07a8d15dac13ef68afb91cb899b9e8e7d39703547dd0eac1
        - quay.io/kiwigrid/k8s-sidecar@sha256:d3749777d5bedf62549b452a7e2ccb205da53588dc7f04927efc75d0da5b6d23
        - registry.connect.redhat.com/kubecost/kubecost-operator-manager@sha256:582223a66a77f35d1f7ca33460032cfca1581a8ff30f75ead289d8cd4cf62bb9
        - registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef
        - registry.connect.redhat.com/kubecost/kubecost-frontend@sha256:f9b2196dc6ac017b8074f8fe40f1a58e7a2630282c5b7a019ca9f3dbdb670dc8
        - registry.connect.redhat.com/kubecost/kubecost-modeling@sha256:8f59f8fd8696964c1b5322ca79fae78befe97323fa56c2b509bdcc67ee4c0edd
        - registry.connect.redhat.com/kubecost/kubecost-cluster-controller@sha256:1a6f11dcd7341ab3f24eba23ec06c7b75646a534a7ed960153601806425469d1
        version: 2.8.2
      entries:
      - name: kubecost-operator.v2.8.2
        version: 2.8.2
      - name: kubecost-operator.v2.8.1
        version: 2.8.1
      - name: kubecost-operator.v2.8.0
        version: 2.8.0
      - name: kubecost-operator.v2.7.2
        version: 2.7.2
      - name: kubecost-operator.v2.7.1
        version: 2.7.1
      - name: kubecost-operator.v2.5.4
        version: 2.5.4
      name: alpha
    defaultChannel: alpha
    packageName: kubecost-operator
    provider:
      name: Kubecost
      url: https://www.kubecost.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-access-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-access-operator.v25.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyAccess",
                "metadata": {
                  "name": "ivia-sample"
                },
                "spec": {
                  "image": "icr.io/ivia/ivia-wrp@sha256:c2dd5ec2b362136896e47ccbd0f3ee033af945fc2794ddaf075da603e4d4d7b2",
                  "instance": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isva/verify-access-operator@sha256:ca1d7be14b4dd313d4675f18541b2cc670ea0a0adf3ec87e347e5e20a86bc39c
          createdAt: "2025-05-01T01:27:17Z"
          description: The IBM Verify Identity Access Operator manages the lifecycle
            of IBM Verify Identity Access worker containers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM-Security/verify-access-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyAccess is the Schema for the ibmsecurityverifyaccesses
              API.
            displayName: IBMSecurity Verify Access
            kind: IBMSecurityVerifyAccess
            name: ibmsecurityverifyaccesses.ibm.com
            version: v1
        description: |+
          In a world of highly fragmented access management environments, [IBM Verify Identity Access](https://www.ibm.com/au-en/products/verify-access) helps you simplify your users' access while more securely adopting web, mobile and cloud technologies. This solution helps you strike a balance between usability and security through the use of risk-based access, single sign-on, integrated access management control, identity federation and its mobile multi-factor authentication capability, IBM Verify. Take back control of your access management with IBM Verify Identity Access.

          The IBM Verify Identity Access operator provides lifecycle management of the lightweight containers which are used to protect an environment, namely:
          * [Web Reverse Proxy](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-web-reverse-proxy)
          * [Runtime](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-runtime)
          * [Distributed Session Cache](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-distributed-session-cache)

          The operator will manage the deployment of these lightweight IBM Verify Identity Access worker containers, and also control the rolling restart of these containers when a configuration snapshot is updated.

          See the project [Readme](https://github.com/IBM-Security/verify-access-operator/blob/master/README.md) for further information and details.

        displayName: IBM Verify Identity Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Identity Access Product Information
          url: https://www.ibm.com/au-en/products/verify-access
        - name: Verify Identity Access Documentation
          url: https://www.ibm.com/docs/en/sva
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Identity Access Development Team
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - icr.io/isva/verify-access-operator@sha256:ca1d7be14b4dd313d4675f18541b2cc670ea0a0adf3ec87e347e5e20a86bc39c
        - icr.io/ivia/ivia-wrp@sha256:c2dd5ec2b362136896e47ccbd0f3ee033af945fc2794ddaf075da603e4d4d7b2
        version: 25.4.0
      entries:
      - name: ibm-security-verify-access-operator.v25.4.0
        version: 25.4.0
      - name: ibm-security-verify-access-operator.v24.12.0
        version: 24.12.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-access-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Redis
      provider-url: ""
    name: redis-enterprise-operator-cert
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-enterprise-operator.v6.2.10-45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
        - registry.connect.redhat.com/redislabs/services-manager@sha256:8b46f2aacfb3e1b243c0b64975accc55fd80b4de9c9ac4769c3c67a5676addc5
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:ffdf10a1b08115fd4de06f1cd751d36644a9bc9a8c11034cd432e9b66a5d2f73
        version: 6.2.10-45
      entries:
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: 6.2.10
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.22.0
    - currentCSV: redis-enterprise-operator.v7.22.0-11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:088e4b57c6e1ca1680b359078e5eb1bba67343cdbd9a539af26a0ea23a830593
        - registry.connect.redhat.com/redislabs/services-manager@sha256:3b96cdc65c3e42fd22eb165950be96b907725903ac74fe4f82cfa0073e28e95a
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:d56ba97e95e799210e76615095409143a5dddd5c5a1e1be1d551d7e3f9e4d3ef
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:0cfb07f774db131b5b74e5aa91291579c916c4d4cd1a693d7aaed554f0177a68
        version: 7.22.0-11.3
      entries:
      - name: redis-enterprise-operator.v7.22.0-11.3
        version: 7.22.0-11.3
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.22.0-11.3
    - currentCSV: redis-enterprise-operator.v7.4.6-7.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:473b5661049a6c8830ba113504975726154c7b172376ba3b63d891674d59c78f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5f6f32a63f4dd8be85b9d469f0fedd26d6aa76e32b7735bfa7ab7203cb272cdb
        - registry.connect.redhat.com/redislabs/services-manager@sha256:ba20ab644a473bc742af671171d5313537292d88942b16336d03a6386c061d8a
        version: 7.4.6-7.2
      entries:
      - name: redis-enterprise-operator.v7.4.6-7.2
        version: 7.4.6-7.2
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      name: 7.4.6
    - currentCSV: redis-enterprise-operator.v7.8.2-6.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:36c7a36dabb23a4c003e112085cda2873296bfb22e5d11b470d7c1aba55b7f0f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:c6e1f2109a0dac76b51c25c1cbf6030b4894b21b98147e2ad1db2dd8e7693576
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:acf817c90b5f03bc2796fbc305cf43ce5c0644dcbe1b685752090b3f600a0f7b
        version: 7.8.2-6.8
      entries:
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.2
    - currentCSV: redis-enterprise-operator.v7.8.4-9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:803be04534e59cdc7b8898b3bb336676aad1cb77cf266051e226be82679fcf6a
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:4ba0c414a74268bed9c6757e447b340db7c912afa864cfb096ddf88f11feeaf9
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:66ec124da16cc9d0abd66687e088be0b358e20cfd48f5774d34bf5874e6a4e65
        version: 7.8.4-9.0
      entries:
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.4
    - currentCSV: redis-enterprise-operator.v7.8.6-5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:1fff01daece34aeb7befcb0760d5112c8c7d261f2b86388ece416ac94076a573
        - registry.connect.redhat.com/redislabs/services-manager@sha256:88d483f5e37b5a901d1ae3d965304edc000c172d2e5e307297d5b7bcd77ae73b
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5125e83d3c93b224d1983c599994e43899a0ba0c6f2fc4eed31476f755b5fc0b
        version: 7.8.6-5.1
      entries:
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.6
    - currentCSV: redis-enterprise-operator.v6.2.10-41
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/outbound",
                      "versionTag": "6.2.10-41"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager",
                      "versionTag": "6.2.10-41"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise",
                    "versionTag": "6.2.14-9.rhel8-openshift"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/outbound@sha256:3c251754ae8650157465a56ef3a7b41d5a0ad4e9db09ef042e0014076f25e8a2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/outbound@sha256:3c251754ae8650157465a56ef3a7b41d5a0ad4e9db09ef042e0014076f25e8a2
        version: 6.2.10-41
      entries:
      - name: redis-enterprise-operator.v6.2.10-41
        version: 6.2.10-41
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: preview
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: production
    defaultChannel: production
    packageName: redis-enterprise-operator-cert
    provider:
      name: Redis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: candidate
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: fast
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: stable
    defaultChannel: stable
    packageName: stackgres
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NooBaa
      provider-url: ""
    name: mcg-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mcg-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[{"kind":"NooBaa","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"noobaa","creationTimestamp":null},"spec":{"cleanupPolicy":{},"security":{"kms":{}},"loadBalancerSourceSubnets":{},"autoscaler":{}},"status":{"lastKeyRotateTime":null}},{"kind":"BackingStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"NamespaceStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"BucketClass","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{},"status":{}},{"kind":"NooBaaAccount","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"allow_bucket_creation":false},"status":{}}]'
          capabilities: Deep Insights
          categories: Storage,Big Data
          certified: "false"
          containerImage: registry.redhat.io/odf4/mcg-rhel9-operator@sha256:43ddea9d3c1cde5c8b38f74e8840a871dd0441ed9fd074528bacf78b439f024f
          createdAt: "2019-07-08T13:10:20.940Z"
          description: NooBaa is an object data service for hybrid and multi cloud
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operators.openshift.io/infrastructure-features: ֿ'["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/noobaa/noobaa-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A NooBaa system - Create this to start
            displayName: NooBaa
            kind: NooBaa
            name: noobaas.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos,
              PV's and more. Used in BucketClass to construct data placement policies.
            displayName: Backing Store
            kind: BackingStore
            name: backingstores.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos
              and more. Used in BucketClass to construct namespace policies.
            displayName: Namespace Store
            kind: NamespaceStore
            name: namespacestores.noobaa.io
            version: v1alpha1
          - description: Storage policy spec  tiering, mirroring, spreading, namespace
              policy. Combines BackingStores Or NamespaceStores. Referenced by ObjectBucketClaims.
            displayName: Bucket Class
            kind: BucketClass
            name: bucketclasses.noobaa.io
            version: v1alpha1
          - kind: NooBaaAccount
            name: noobaaaccounts.noobaa.io
            version: v1alpha1
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information.
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: "The noobaa operator creates and reconciles a NooBaa system in
          a Kubernetes/Openshift cluster.\n\nNooBaa provides an S3 object-store service
          abstraction and data placement policies to create hybrid and multi cloud
          data solutions.\n\nFor more information on using NooBaa refer to [Github](https://github.com/noobaa/noobaa-core)
          / [Website](https://www.noobaa.io) / [Articles](https://noobaa.desk.com).
          \n\n## How does it work?\n\n- The operator deploys the noobaa core pod and
          two services - Mgmt (UI/API) and S3 (object-store).\n- Both services require
          credentials which you will get from a secret that the operator creates -
          use describe noobaa to locate it.\n- The service addresses will also appear
          in the describe output - pick the one that is suitable for your client:\n
          \   - minikube - use the NodePort address.\n    - remote cluster - probably
          need one of the External addresses.\n    - connect an application on the
          same cluster - use Internal DNS (though any address should work)\n    \n-
          Feel free to email us or open github issues on any question.\n\n## Getting
          Started\n\n### Notes:\n- The following instructions are for **minikube**
          but it works on any Kubernetes/Openshift clusters.\n- This will setup noobaa
          in the **my-noobaa-operator** namespace.\n- You will need **jq**, **curl**,
          **kubectl** or **oc**, **aws-cli**.\n\n### 1. Install OLM (if you don't
          have it already):\n```\ncurl -sL https://github.com/operator-framework/operator-lifecycle-manager/releases/download/0.12.0/install.sh
          | bash -s 0.12.0\n```\n\n### 2. Install noobaa-operator:\n```\nkubectl create
          -f https://operatorhub.io/install/noobaa-operator.yaml\n```\nWait for it
          to be ready:\n```\nkubectl wait pod -n my-noobaa-operator -l noobaa-operator
          --for=condition=ready\n```\n\n### 3. Create noobaa system:\n```\ncurl -sL
          https://operatorhub.io/api/operator?packageName=noobaa-operator | \n    jq
          '.operator.customResourceDefinitions[0].yamlExample | .metadata.namespace=\"my-noobaa-operator\"'
          |\n    kubectl create -f -\n```\nWait for it to be ready:\n```\nkubectl
          wait pod -n my-noobaa-operator -l noobaa-core --for=condition=ready\nkubectl
          get noobaa -n my-noobaa-operator -w\n# NAME     PHASE   MGMT-ENDPOINTS                  S3-ENDPOINTS
          \                   IMAGE                    AGE\n# noobaa   **Ready**   [https://192.168.64.12:31121]
          \  [https://192.168.64.12:32557]   noobaa/noobaa-core:4.0   19m\n```\n\n###
          4. Get system information to your shell:\n```\nNOOBAA_SECRET=$(kubectl get
          noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.accounts.admin.secretRef.name'
          )\nNOOBAA_MGMT=$(kubectl get noobaa noobaa -n my-noobaa-operator -o json
          | jq -r '.status.services.serviceMgmt.nodePorts[0]' )\nNOOBAA_S3=$(kubectl
          get noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.services.serviceS3.nodePorts[0]'
          )\nNOOBAA_ACCESS_KEY=$(kubectl get secret $NOOBAA_SECRET -n my-noobaa-operator
          -o json | jq -r '.data.AWS_ACCESS_KEY_ID|@base64d')\nNOOBAA_SECRET_KEY=$(kubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq -r '.data.AWS_SECRET_ACCESS_KEY|@base64d')\n```\n\n###
          5. Connect to Mgmt UI:\n```\n# show email/password from the secret:\nkubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq '.data|map_values(@base64d)'\n\n#
          open mgmt UI login:\nopen $NOOBAA_MGMT\n```\n\n### 6. Connect to S3 with
          aws-cli:\n```\nalias s3='AWS_ACCESS_KEY_ID=$NOOBAA_ACCESS_KEY AWS_SECRET_ACCESS_KEY=$NOOBAA_SECRET_KEY
          aws --endpoint $NOOBAA_S3 --no-verify-ssl s3'\ns3 ls\ns3 sync /var/log/
          s3://first.bucket\ns3 ls s3://first.bucket\n```\n"
        displayName: NooBaa Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - noobaa
        - kubernetes
        - openshift
        - cloud
        - hybrid
        - multi
        - data
        - storage
        - s3
        - tiering
        - mirroring
        links:
        - name: Github
          url: https://github.com/noobaa/noobaa-core
        - name: Website
          url: https://www.noobaa.io
        - name: Articles
          url: https://noobaa.desk.com
        maintainers:
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: etamir@redhat.com
          name: Eran Tamir
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: NooBaa
        relatedImages:
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:f37e2c611f4cda32172d4e8938938e488076fd8a8f19c252c56a80c48044bc02
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhel8/postgresql-12@sha256:af1ab437b165fa4ee9dfd6ade637d7de5fa556d0b3851fb8edc78afe0e6fa95e
        - registry.redhat.io/odf4/mcg-rhel9-operator@sha256:43ddea9d3c1cde5c8b38f74e8840a871dd0441ed9fd074528bacf78b439f024f
        - registry.redhat.io/odf4/odf-cosi-sidecar-rhel9@sha256:bce36c7d33a85a11cd6b6b24f92887cffe6f38af70340e54d10e4f6f524aa782
        version: 4.15.20-rhodf
      entries:
      - name: mcg-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: mcg-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: mcg-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: mcg-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: mcg-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: mcg-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: mcg-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: mcg-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: mcg-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: mcg-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: mcg-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: mcg-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: mcg-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: mcg-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: mcg-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: mcg-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: mcg-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: mcg-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: mcg-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: mcg-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: mcg-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[{"kind":"NooBaa","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"noobaa","creationTimestamp":null},"spec":{"cleanupPolicy":{},"security":{"kms":{}},"loadBalancerSourceSubnets":{},"autoscaler":{}},"status":{"lastKeyRotateTime":null}},{"kind":"BackingStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"NamespaceStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"BucketClass","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{},"status":{}},{"kind":"NooBaaAccount","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"allow_bucket_creation":false},"status":{}}]'
          capabilities: Deep Insights
          categories: Storage,Big Data
          certified: "false"
          containerImage: registry.redhat.io/odf4/mcg-rhel9-operator@sha256:fdcdc8870b12d853973000877818463c3b4a5ed1d3316ff72a04a299577ed23d
          createdAt: "2019-07-08T13:10:20.940Z"
          description: NooBaa is an object data service for hybrid and multi cloud
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/infrastructure-features: ֿ'["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/noobaa/noobaa-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A NooBaa system - Create this to start
            displayName: NooBaa
            kind: NooBaa
            name: noobaas.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos,
              PV's and more. Used in BucketClass to construct data placement policies.
            displayName: Backing Store
            kind: BackingStore
            name: backingstores.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos
              and more. Used in BucketClass to construct namespace policies.
            displayName: Namespace Store
            kind: NamespaceStore
            name: namespacestores.noobaa.io
            version: v1alpha1
          - description: Storage policy spec  tiering, mirroring, spreading, namespace
              policy. Combines BackingStores Or NamespaceStores. Referenced by ObjectBucketClaims.
            displayName: Bucket Class
            kind: BucketClass
            name: bucketclasses.noobaa.io
            version: v1alpha1
          - kind: NooBaaAccount
            name: noobaaaccounts.noobaa.io
            version: v1alpha1
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information.
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: "The noobaa operator creates and reconciles a NooBaa system in
          a Kubernetes/Openshift cluster.\n\nNooBaa provides an S3 object-store service
          abstraction and data placement policies to create hybrid and multi cloud
          data solutions.\n\nFor more information on using NooBaa refer to [Github](https://github.com/noobaa/noobaa-core)
          / [Website](https://www.noobaa.io) / [Articles](https://noobaa.desk.com).
          \n\n## How does it work?\n\n- The operator deploys the noobaa core pod and
          two services - Mgmt (UI/API) and S3 (object-store).\n- Both services require
          credentials which you will get from a secret that the operator creates -
          use describe noobaa to locate it.\n- The service addresses will also appear
          in the describe output - pick the one that is suitable for your client:\n
          \   - minikube - use the NodePort address.\n    - remote cluster - probably
          need one of the External addresses.\n    - connect an application on the
          same cluster - use Internal DNS (though any address should work)\n    \n-
          Feel free to email us or open github issues on any question.\n\n## Getting
          Started\n\n### Notes:\n- The following instructions are for **minikube**
          but it works on any Kubernetes/Openshift clusters.\n- This will setup noobaa
          in the **my-noobaa-operator** namespace.\n- You will need **jq**, **curl**,
          **kubectl** or **oc**, **aws-cli**.\n\n### 1. Install OLM (if you don't
          have it already):\n```\ncurl -sL https://github.com/operator-framework/operator-lifecycle-manager/releases/download/0.12.0/install.sh
          | bash -s 0.12.0\n```\n\n### 2. Install noobaa-operator:\n```\nkubectl create
          -f https://operatorhub.io/install/noobaa-operator.yaml\n```\nWait for it
          to be ready:\n```\nkubectl wait pod -n my-noobaa-operator -l noobaa-operator
          --for=condition=ready\n```\n\n### 3. Create noobaa system:\n```\ncurl -sL
          https://operatorhub.io/api/operator?packageName=noobaa-operator | \n    jq
          '.operator.customResourceDefinitions[0].yamlExample | .metadata.namespace=\"my-noobaa-operator\"'
          |\n    kubectl create -f -\n```\nWait for it to be ready:\n```\nkubectl
          wait pod -n my-noobaa-operator -l noobaa-core --for=condition=ready\nkubectl
          get noobaa -n my-noobaa-operator -w\n# NAME     PHASE   MGMT-ENDPOINTS                  S3-ENDPOINTS
          \                   IMAGE                    AGE\n# noobaa   **Ready**   [https://192.168.64.12:31121]
          \  [https://192.168.64.12:32557]   noobaa/noobaa-core:4.0   19m\n```\n\n###
          4. Get system information to your shell:\n```\nNOOBAA_SECRET=$(kubectl get
          noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.accounts.admin.secretRef.name'
          )\nNOOBAA_MGMT=$(kubectl get noobaa noobaa -n my-noobaa-operator -o json
          | jq -r '.status.services.serviceMgmt.nodePorts[0]' )\nNOOBAA_S3=$(kubectl
          get noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.services.serviceS3.nodePorts[0]'
          )\nNOOBAA_ACCESS_KEY=$(kubectl get secret $NOOBAA_SECRET -n my-noobaa-operator
          -o json | jq -r '.data.AWS_ACCESS_KEY_ID|@base64d')\nNOOBAA_SECRET_KEY=$(kubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq -r '.data.AWS_SECRET_ACCESS_KEY|@base64d')\n```\n\n###
          5. Connect to Mgmt UI:\n```\n# show email/password from the secret:\nkubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq '.data|map_values(@base64d)'\n\n#
          open mgmt UI login:\nopen $NOOBAA_MGMT\n```\n\n### 6. Connect to S3 with
          aws-cli:\n```\nalias s3='AWS_ACCESS_KEY_ID=$NOOBAA_ACCESS_KEY AWS_SECRET_ACCESS_KEY=$NOOBAA_SECRET_KEY
          aws --endpoint $NOOBAA_S3 --no-verify-ssl s3'\ns3 ls\ns3 sync /var/log/
          s3://first.bucket\ns3 ls s3://first.bucket\n```\n"
        displayName: NooBaa Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - noobaa
        - kubernetes
        - openshift
        - cloud
        - hybrid
        - multi
        - data
        - storage
        - s3
        - tiering
        - mirroring
        links:
        - name: Github
          url: https://github.com/noobaa/noobaa-core
        - name: Website
          url: https://www.noobaa.io
        - name: Articles
          url: https://noobaa.desk.com
        maintainers:
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: etamir@redhat.com
          name: Eran Tamir
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: NooBaa
        relatedImages:
        - registry.redhat.io/odf4/mcg-rhel9-operator@sha256:fdcdc8870b12d853973000877818463c3b4a5ed1d3316ff72a04a299577ed23d
        - registry.redhat.io/odf4/odf-cosi-sidecar-rhel9@sha256:9fed589373d1586577d57f23b7a4f39ace5052fc55c99a03cb256da63a26c441
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:0084d9053ad2c74926ca02d2ede0461abd59227eee98d122c45e08a5ebaf629a
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhel8/postgresql-12@sha256:af1ab437b165fa4ee9dfd6ade637d7de5fa556d0b3851fb8edc78afe0e6fa95e
        version: 4.16.16-rhodf
      entries:
      - name: mcg-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: mcg-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: mcg-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: mcg-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: mcg-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: mcg-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: mcg-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: mcg-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: mcg-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: mcg-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: mcg-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: mcg-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: mcg-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: mcg-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: mcg-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: mcg-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: mcg-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: mcg-operator
    provider:
      name: NooBaa
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-eks-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-eks-controller.v1.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "FargateProfile",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Nodegroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/eks-controller:1.9.3
          createdAt: "2025-10-03T21:43:09Z"
          description: AWS EKS controller is a service controller for managing EKS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessEntry represents the state of an AWS eks AccessEntry
              resource.
            displayName: AccessEntry
            kind: AccessEntry
            name: accessentries.eks.services.k8s.aws
            version: v1alpha1
          - description: Addon represents the state of an AWS eks Addon resource.
            displayName: Addon
            kind: Addon
            name: addons.eks.services.k8s.aws
            version: v1alpha1
          - description: Cluster represents the state of an AWS eks Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.eks.services.k8s.aws
            version: v1alpha1
          - description: FargateProfile represents the state of an AWS eks FargateProfile
              resource.
            displayName: FargateProfile
            kind: FargateProfile
            name: fargateprofiles.eks.services.k8s.aws
            version: v1alpha1
          - description: IdentityProviderConfig represents the state of an AWS eks
              IdentityProviderConfig resource.
            displayName: IdentityProviderConfig
            kind: IdentityProviderConfig
            name: identityproviderconfigs.eks.services.k8s.aws
            version: v1alpha1
          - description: Nodegroup represents the state of an AWS eks Nodegroup resource.
            displayName: Nodegroup
            kind: Nodegroup
            name: nodegroups.eks.services.k8s.aws
            version: v1alpha1
          - description: PodIdentityAssociation represents the state of an AWS eks
              PodIdentityAssociation resource.
            displayName: PodIdentityAssociation
            kind: PodIdentityAssociation
            name: podidentityassociations.eks.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Kubernetes Service (EKS) resources in AWS from within your Kubernetes cluster.

          **About Amazon EKS**

          Amazon Elastic Kubernetes Service (Amazon EKS) is a managed service that you can use to run Kubernetes on AWS without needing to install, operate, and maintain your own Kubernetes control plane or nodes. Kubernetes is an open-source system for automating the deployment, scaling, and management of containerized applications.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EKS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eks
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EKS Developer Resources
          url: https://aws.amazon.com/eks/getting-started/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/eks-controller:1.9.3
        version: 1.9.3
      entries:
      - name: ack-eks-controller.v1.9.3
        version: 1.9.3
      - name: ack-eks-controller.v1.9.2
        version: 1.9.2
      - name: ack-eks-controller.v1.9.1
        version: 1.9.1
      - name: ack-eks-controller.v1.9.0
        version: 1.9.0
      - name: ack-eks-controller.v1.8.0
        version: 1.8.0
      - name: ack-eks-controller.v1.7.6
        version: 1.7.6
      - name: ack-eks-controller.v1.7.5
        version: 1.7.5
      - name: ack-eks-controller.v1.7.4
        version: 1.7.4
      - name: ack-eks-controller.v1.7.3
        version: 1.7.3
      - name: ack-eks-controller.v1.7.2
        version: 1.7.2
      - name: ack-eks-controller.v1.6.4
        version: 1.6.4
      - name: ack-eks-controller.v1.6.3
        version: 1.6.3
      - name: ack-eks-controller.v1.6.2
        version: 1.6.2
      - name: ack-eks-controller.v1.6.1
        version: 1.6.1
      - name: ack-eks-controller.v1.6.0
        version: 1.6.0
      - name: ack-eks-controller.v1.5.5
        version: 1.5.5
      - name: ack-eks-controller.v1.5.4
        version: 1.5.4
      - name: ack-eks-controller.v1.5.3
        version: 1.5.3
      - name: ack-eks-controller.v1.5.2
        version: 1.5.2
      - name: ack-eks-controller.v1.5.1
        version: 1.5.1
      - name: ack-eks-controller.v1.5.0
        version: 1.5.0
      - name: ack-eks-controller.v1.4.7
        version: 1.4.7
      - name: ack-eks-controller.v1.4.6
        version: 1.4.6
      - name: ack-eks-controller.v1.4.5
        version: 1.4.5
      - name: ack-eks-controller.v1.4.4
        version: 1.4.4
      - name: ack-eks-controller.v1.4.3
        version: 1.4.3
      - name: ack-eks-controller.v1.4.2
        version: 1.4.2
      - name: ack-eks-controller.v1.4.1
        version: 1.4.1
      - name: ack-eks-controller.v1.4.0
        version: 1.4.0
      - name: ack-eks-controller.v1.3.5
        version: 1.3.5
      - name: ack-eks-controller.v1.3.3
        version: 1.3.3
      - name: ack-eks-controller.v1.3.1
        version: 1.3.1
      - name: ack-eks-controller.v1.3.0
        version: 1.3.0
      - name: ack-eks-controller.v1.2.3
        version: 1.2.3
      - name: ack-eks-controller.v1.2.2
        version: 1.2.2
      - name: ack-eks-controller.v1.2.1
        version: 1.2.1
      - name: ack-eks-controller.v1.2.0
        version: 1.2.0
      - name: ack-eks-controller.v1.1.3
        version: 1.1.3
      - name: ack-eks-controller.v1.1.2
        version: 1.1.2
      - name: ack-eks-controller.v1.1.1
        version: 1.1.1
      - name: ack-eks-controller.v1.1.0
        version: 1.1.0
      - name: ack-eks-controller.v1.0.5
        version: 1.0.5
      - name: ack-eks-controller.v1.0.4
        version: 1.0.4
      - name: ack-eks-controller.v1.0.3
        version: 1.0.3
      - name: ack-eks-controller.v1.0.2
        version: 1.0.2
      - name: ack-eks-controller.v1.0.1
        version: 1.0.1
      - name: ack-eks-controller.v1.0.0
        version: 1.0.0
      - name: ack-eks-controller.v0.1.9
        version: 0.1.9
      - name: ack-eks-controller.v0.1.8
        version: 0.1.8
      - name: ack-eks-controller.v0.1.7
        version: 0.1.7
      - name: ack-eks-controller.v0.1.6
        version: 0.1.6
      - name: ack-eks-controller.v0.1.5
        version: 0.1.5
      - name: ack-eks-controller.v0.1.4
        version: 0.1.4
      - name: ack-eks-controller.v0.1.3
        version: 0.1.3
      - name: ack-eks-controller.v0.1.2
        version: 0.1.2
      - name: ack-eks-controller.v0.1.1
        version: 0.1.1
      - name: ack-eks-controller.v0.1.0
        version: 0.1.0
      - name: ack-eks-controller.v0.0.15
        version: 0.0.15
      - name: ack-eks-controller.v0.0.14
        version: 0.0.14
      - name: ack-eks-controller.v0.0.13
        version: 0.0.13
      - name: ack-eks-controller.v0.0.12
        version: 0.0.12
      - name: ack-eks-controller.v0.0.11
        version: 0.0.11
      - name: ack-eks-controller.v0.0.10
        version: 0.0.10
      - name: ack-eks-controller.v0.0.9
        version: 0.0.9
      - name: ack-eks-controller.v0.0.8
        version: 0.0.8
      - name: ack-eks-controller.v0.0.7
        version: 0.0.7
      - name: ack-eks-controller.v0.0.6
        version: 0.0.6
      - name: ack-eks-controller.v0.0.5
        version: 0.0.5
      - name: ack-eks-controller.v0.0.4
        version: 0.0.4
      - name: ack-eks-controller.v0.0.3
        version: 0.0.3
      - name: ack-eks-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-eks-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ODH
      provider-url: ""
    name: opendatahub-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opendatahub-operator.v2.35.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Managed",
                      "registriesNamespace": "odh-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "opendatahub",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "opendatahub"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v2.35.0
          createdAt: "2025-10-01T13:12:27Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",
            "workbenches.components.platform.opendatahub.io", "monitorings.services.platform.opendatahub.io",
            "servicemeshes.services.platform.opendatahub.io", "modelcontrollers.components.platform.opendatahub.io",
            "feastoperators.components.platform.opendatahub.io", "llamastackoperators.components.platform.opendatahub.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/opendatahub-io/opendatahub-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: ServiceMesh is the Schema for the servicemesh API
            displayName: Service Mesh
            kind: ServiceMesh
            name: servicemeshes.services.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack - Unified open-source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odh
        - notebooks
        - serving
        - training
        - pipelines
        - modelmesh
        - workbenches
        - dashboard
        - kserve
        - distributed workloads
        - trustyai
        - modelregistry
        - feast
        - featurestore
        - llamastack
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ODH
        relatedImages:
        - quay.io/opendatahub/opendatahub-operator:v2.35.0
        version: 2.35.0
      entries:
      - name: opendatahub-operator.v2.35.0
        version: 2.35.0
      - name: opendatahub-operator.v2.34.0
        version: 2.34.0
      - name: opendatahub-operator.v2.33.0
        version: 2.33.0
      - name: opendatahub-operator.v2.32.0
        version: 2.32.0
      - name: opendatahub-operator.v2.31.0
        version: 2.31.0
      - name: opendatahub-operator.v2.30.0
        version: 2.30.0
      - name: opendatahub-operator.v2.29.0
        version: 2.29.0
      - name: opendatahub-operator.v2.28.0
        version: 2.28.0
      - name: opendatahub-operator.v2.27.0
        version: 2.27.0
      - name: opendatahub-operator.v2.26.0
        version: 2.26.0
      - name: opendatahub-operator.v2.25.0
        version: 2.25.0
      - name: opendatahub-operator.v2.24.0
        version: 2.24.0
      - name: opendatahub-operator.v2.23.0
        version: 2.23.0
      - name: opendatahub-operator.v2.22.0
        version: 2.22.0
      - name: opendatahub-operator.v2.21.0
        version: 2.21.0
      - name: opendatahub-operator.v2.20.0
        version: 2.20.0
      - name: opendatahub-operator.v2.19.0
        version: 2.19.0
      - name: opendatahub-operator.v2.18.2
        version: 2.18.2
      - name: opendatahub-operator.v2.18.1
        version: 2.18.1
      - name: opendatahub-operator.v2.18.0
        version: 2.18.0
      - name: opendatahub-operator.v2.17.0
        version: 2.17.0
      - name: opendatahub-operator.v2.16.0
        version: 2.16.0
      - name: opendatahub-operator.v2.15.0
        version: 2.15.0
      - name: opendatahub-operator.v2.14.0
        version: 2.14.0
      - name: opendatahub-operator.v2.13.0
        version: 2.13.0
      - name: opendatahub-operator.v2.12.1
        version: 2.12.1
      - name: opendatahub-operator.v2.12.0
        version: 2.12.0
      - name: opendatahub-operator.v2.11.1
        version: 2.11.1
      - name: opendatahub-operator.v2.11.0
        version: 2.11.0
      - name: opendatahub-operator.v2.10.0
        version: 2.10.0
      name: fast
    - currentCSV: opendatahub-operator.v2.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "opendatahub",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "opendatahub"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v2.8.0
          createdAt: "2024-2-20T00:00:00Z"
          operatorframework.io/initialization-resource: |-
            {
            "apiVersion": "dscinitialization.opendatahub.io/v1",
            "kind": "DSCInitialization",
            "metadata": {
            "labels": {
             "app.kubernetes.io/created-by": "opendatahub-operator",
             "app.kubernetes.io/instance": "default",
             "app.kubernetes.io/managed-by": "kustomize",
             "app.kubernetes.io/name": "dscinitialization",
             "app.kubernetes.io/part-of": "opendatahub-operator"
            },
            "name": "default-dsci"
            },
            "spec": {
            "applicationsNamespace": "opendatahub",
            "monitoring": {
             "managementState": "Managed",
             "namespace": "opendatahub"
            },
            "serviceMesh": {
             "controlPlane": {
               "metricsCollection": "Istio",
               "name": "data-science-smcp",
               "namespace": "istio-system"
             },
             "managementState": "Managed"
            },
            "trustedCABundle": {
             "customCABundle": "",
             "managementState": "Managed"
            }
            }
            }
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: "The Open Data Hub is a machine-learning-as-a-service platform
          built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open
          Data Hub integrates multiple AI/ML open source components into one operator
          that can easily be downloaded and installed by OpenShift users.\n\nOpen
          Data Hub operator allows users to install and manage components of the Open
          Data Hub. Users can mix and match tools from each project to fulfill the
          needs of their use case. Each of the projects share some components, but
          can be mostly seen as an extension of each other to provide a complete solution
          for both novice and skilled enterprise users.\n    \n### Components\n\n*
          Open Data Hub Dashboard - A web dashboard that displays installed Open Data
          Hub components with easy access to component UIs and documentation\n* ODH
          Notebook Controller - Secure management of Jupyter Notebook in Kubernetes
          environments built on top of Kubeflow Notebook Controller with support for
          OAuth\n* Jupyter Notebooks - JupyterLab notebook that provide Python support
          for GPU workloads\n* Data Science Pipelines - Pipeline solution for end
          to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton\n*
          Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh,
          a general-purpose model serving management/routing layer\n* Distributed
          Workloads(Incubation) - Stack built to make managing distributed compute
          infrastructure in the cloud easy and intuitive for Data Scientists. This
          stack consists of three components \n                                      Codeflare
          , KubeRay and Kueue.\n* Kserve - Kserve is the Controller for for serving
          machine learning (ML) models on arbitrary frameworks"
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odh
        - notebooks
        - serving
        - training
        - pipelines
        - modelmesh
        - workbenches
        - dashboard
        - kserve
        - distributed-workloads
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: ODH
        relatedImages:
        - quay.io/opendatahub/opendatahub-operator:v2.8.0
        version: 2.8.1
      entries:
      - name: opendatahub-operator.v2.8.1
        version: 2.8.1
      name: odh-2.8.z
    - currentCSV: opendatahub-operator.v1.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kfdef.apps.kubeflow.org/v1",
                "kind": "KfDef",
                "metadata": {
                  "name": "opendatahub",
                  "namespace": "opendatahub"
                },
                "spec": {
                  "applications": [
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-common"
                        }
                      },
                      "name": "odh-common"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-dashboard"
                        }
                      },
                      "name": "odh-dashboard"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "data-science-pipelines-operator/"
                        }
                      },
                      "name": "data-science-pipelines-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-notebook-controller"
                        }
                      },
                      "name": "odh-notebook-controller"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "notebook-images"
                        }
                      },
                      "name": "notebook-images"
                    },
                    {
                      "kustomizeConfig": {
                        "parameters": [
                          {
                            "name": "monitoring-namespace",
                            "value": "opendatahub"
                          }
                        ],
                        "repoRef": {
                          "name": "manifests",
                          "path": "model-mesh"
                        }
                      },
                      "name": "model-mesh"
                    }
                  ],
                  "repos": [
                    {
                      "name": "manifests",
                      "uri": "file:///opt/manifests/odh-manifests.tar.gz"
                    }
                  ]
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhApplication",
                "metadata": {
                  "name": "example-odhapplication",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started,Model development,Model training,Notebook environments,Package management,Data Management"
                  }
                },
                "spec": {
                  "displayName": "ODHAPPLICATION DISPLAY NAME",
                  "provider": "PROVIDER_NAME",
                  "description": "Description of your ODHApplication",
                  "kfdefApplications": [
                    "<NAME OF THE ODH APPLICATION IN KFDEF>"
                  ],
                  "route": "<APPLICATION_OCP_ROUTE_OBJECT_NAME>",
                  "img": "<svg xmlns=\"http://www.w3.org/2000/svg\" role=\"img\" viewBox=\"-1.13 -1.13 58.50 48.94\"><title>Open Data Hub logo</title><g><g><g fill=\"#4d5258\" stroke-width=\".006\"><path fill=\"#bbb\" d=\"M112.30231 22.797083c-.2034 0-.35737.143517-.35737.337407 0 .192939.15397.339307.35737.339307.20339 0 .35736-.147318.35736-.339307 0-.19294-.15397-.337407-.35736-.337407zm.002.131161c.1112 0 .20149.08934.20149.206246 0 .116904-.0903.208146-.20149.208146-.1112 0-.2053-.09124-.2053-.208146 0-.116904.0931-.206246.2053-.206246zM113.04141 22.801836h-.28228v.665308h.15017v-.190088h.13211c.16918 0 .26612-.08934.26612-.244263 0-.147319-.0969-.230957-.26612-.230957zm-.008.349762h-.12451v-.225255h.12451c.0836 0 .13116.03707.13116.111202 0 .07603-.0475.114053-.13116.114053zM113.40328 22.801836v.665308h.51703v-.124508h-.36687v-.147318h.32125V23.07081h-.32125v-.144467h.35642v-.124507zM114.03814 22.801836v.665308h.14352v-.426748l.3222.426748h.13686v-.665308h-.14257v.427698L114.175 22.801836z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/><path d=\"M115.29685 22.801836h-.28228v.665308h.27658c.21004 0 .35546-.136863.35546-.332654 0-.196741-.14351-.332654-.34976-.332654zm.004.538899h-.13591v-.412491h.12926c.11405 0 .19959.08459.19959.207196 0 .121657-.0817.205295-.19294.205295zM116.21878 23.467144h.16062l-.28418-.665308h-.15492l-.28894.665308h.15493l.0523-.12926h.30794zm-.3108-.249966l.1055-.25852.10455.25852zM116.31525 22.801836v.127359h.20244v.537949h.15017v-.537949h.2034v-.127359zM117.37527 23.467144h.16063l-.28419-.665308h-.15492l-.28893.665308h.15492l.0523-.12926h.30795zm-.31079-.249966l.1055-.25852.10454.25852zM117.83433 22.801836v.665308h.15017v-.262322h.30129v.262322h.15017v-.665308h-.15017v.277528h-.30129v-.277528zM118.72025 23.188665v-.386829h-.15017v.386829c0 .175831.11785.285132.30224.285132.18343 0 .29844-.109301.29844-.285132v-.386829h-.15017v.386829c0 .09599-.0589.153971-.14827.153971-.0893 0-.15207-.05798-.15207-.153971zM119.61059 22.801836h-.30604v.665308h.31649c.15492 0 .25092-.07033.25092-.184385 0-.08744-.0532-.152071-.13876-.168228.0713-.01806.115-.07413.115-.146368 0-.102648-.0912-.166327-.23761-.166327zm-.008.268024h-.14826v-.145418h.14826c.0561 0 .0913.02566.0913.07128 0 .04657-.0352.07413-.0913.07413zm0 .274677h-.14826v-.158723h.14826c.0713 0 .11406.02946.11406.07794 0 .05132-.0428.08079-.11406.08079z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g><g fill=\"#58595b\" stroke-width=\".01\"><path d=\"M112.19192 24.15027h.0381l-.13717-.306759h-.0364l-.1376.306759h.0372l.0342-.07844h.16696zm-.18712-.110871l.0688-.157763.0697.157763zM112.27858 23.843511v.306759h.0351v-.306759zM112.64205 23.843511h-.11525v.306759h.0351v-.09729h.0802c.0758 0 .11964-.039.11964-.10649 0-.0653-.0438-.102983-.11964-.102983zm-.001.177044h-.0789V23.87594h.0789c.0552 0 .0872.02454.0872.07143 0 .0482-.032.07318-.0872.07318zM112.81775 23.825105v.325165h.0342v-.325165zM113.1137 24.15027l-.00044-.152503c-.00044-.04908-.032-.07932-.089-.07932-.0394 0-.0644.01139-.0942.03111l.0145.0241c.0241-.01709.0478-.02673.0732-.02673.0408 0 .0618.0206.0618.05434v.01446h-.0758c-.0543.000439-.0863.02586-.0863.06617 0 .03944.0307.06968.0784.06968.0368 0 .0662-.01227.0846-.03725v.03593zm-.11131-.02717c-.0324 0-.053-.01709-.053-.04251 0-.02585.0202-.03769.0587-.03769h.0719v.02805c-.009.03287-.0386.05215-.0776.05215zM113.29915 24.108201c-.0127.0074-.025.01139-.0368.01183-.0197 0-.0307-.01139-.0307-.03944v-.127086h.0732v-.02673h-.0732v-.06048h-.0337v.06048h-.0333v.02673h.0333v.132783c0 .04558.025.06573.0587.0653.0184 0 .0359-.0061.0521-.01709zM113.44228 23.854466c.0118 0 .0245.0044.0351.01183l.014-.02629c-.014-.01052-.032-.01622-.0495-.01622-.0386-.000438-.0741.02717-.0741.07713v.02586h-.0298v.02673h.0298v.196764h.0337v-.196764h.0662v-.02673h-.0662v-.03024c0-.02717.0184-.04251.0408-.04207zM113.59236 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM113.80323 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM114.22747 23.918448c-.0443.000438-.0767.01972-.0911.06004-.01-.03813-.0394-.06004-.0828-.06004-.0416.000438-.0727.01753-.0881.05346v-.05215h-.0337v.230508h.0337v-.113501c0-.04996.0285-.08458.0758-.08589.0403 0 .064.02454.064.06617v.133221h.0346v-.113501c0-.04996.0276-.08458.0754-.08589.0399 0 .064.02454.064.06617v.133221h.0342V24.00828c0-.05609-.032-.08984-.0859-.08984zM114.63474 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0487.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11088-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM114.89934 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM115.02844 23.919762l.0863.230508h.0355l.0675-.188876.0679.188876h.0355l.0855-.230508h-.0355l-.0675.195888-.0679-.195888h-.0359l-.067.195888-.0684-.195888zM115.53009 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM115.73075 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM115.95034 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11482.116569.0386 0 .071-.01402.0929-.03813l-.0188-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18843c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM116.29474 23.825105v.144615c-.0171-.03287-.0491-.05171-.0903-.05171-.0653 0-.11087.04777-.11087.116569 0 .0688.0456.117445.11131.117445.0412 0 .0727-.01884.0898-.05127v.04952h.0342v-.325165zm-.0833.297557c-.0491 0-.0837-.03637-.0837-.08721 0-.05083.0346-.08765.0837-.08765.0486 0 .0833.03681.0833.08765 0 .05083-.0346.08721-.0833.08721zM116.65984 23.918448c-.0403 0-.0719.01884-.089.05083v-.144177h-.0342v.325165h.0342v-.04952c.0171.03199.0486.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM116.89657 24.183138l.10605-.263376h-.0351l-.0732.191944-.0767-.191944h-.0359l.0955.228756-.01.02191c-.009.02191-.0241.03331-.0421.03331-.0114 0-.0219-.0035-.0324-.01095l-.0145.02717c.014.01139.0294.01665.0486.01665.0307 0 .0548-.01841.0692-.05346zM117.29947 23.842634c-.0898 0-.16083.06793-.16083.154257 0 .08633.071.154694.16083.154694.0894 0 .16039-.06836.16039-.154694 0-.08633-.071-.154257-.16039-.154257zm0 .03243c.0684 0 .12446.0539.12446.121828 0 .06792-.0561.122703-.12446.122703-.0684 0-.12533-.05478-.12533-.122703 0-.06793.057-.121828.12533-.121828zM117.64414 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0486.05083.0898.05083.0653 0 .10999-.04777.10999-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM117.9061 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM118.19616 23.918448c-.0421.000438-.0732.01709-.0894.05215v-.05084h-.0342v.230508h.0342v-.12358c.004-.0447.032-.07494.0776-.07537.0408 0 .0653.02454.0653.06573v.133221h.0342v-.141986c0-.05609-.0324-.08984-.0876-.08984zM118.65836 23.901795l.0149-.03111c-.0254-.01709-.0618-.02848-.0973-.02848-.0635 0-.10605.03068-.10605.07888 0 .109995.17266.06661.17222.148559 0 .0298-.0276.04777-.0719.04777-.0346 0-.0736-.01578-.1021-.04295l-.0158.03024c.0289.02761.0727.04689.11745.04689.0657 0 .11087-.03287.11087-.0837.00044-.11131-.17223-.06968-.17223-.149436 0-.02629.0259-.04207.0653-.04207.0245 0 .057.0075.0846.02542zM118.84168 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .11701-.04777.11701-.117007 0-.06836-.0487-.11613-.11701-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM119.18446 23.919762v.113063c0 .04996-.0281.08502-.0758.08589-.0399 0-.064-.02454-.064-.06573v-.133222h-.0342v.141986c0 .05565.0324.08984.0863.08984.0416-.000438.0723-.01753.0876-.05346v.05215h.0342v-.230508zM119.33586 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM119.62557 23.976294l.0188-.02366c-.021-.02147-.0517-.03418-.089-.03418-.0666 0-.11438.0482-.11438.116568 0 .0688.0473.116569.11438.116569.0399 0 .0719-.0149.0925-.03988l-.0197-.02016c-.0162.01928-.0412.0298-.0714.0298-.0478 0-.082-.03506-.082-.08633 0-.05083.0342-.08589.082-.08589.028 0 .0513.01008.0688.02717zM119.78997 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g></g><g><g><path fill=\"#a586f3\" d=\"m 7808.8125,2097.1875 c -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 v 329.875 c -0.8781,10.2195 7.1804,19 17.4375,19 10.2571,0 19.0227,-8.7805 18.1446,-19 v -77.7324 l 0.4179,-6.1426 c 0,-55.1597 44.2153,-99.5625 99.375,-99.5625 55.1596,0 99.5625,44.4028 99.5625,99.5625 0.01,0.1837 0.015,0.3674 0.027,0.5508 -0.026,0.4409 -0.035,0.8826 -0.027,1.3242 v 110 c 0,9.6822 7.849,17.5312 17.5312,17.5312 9.6823,0 17.5313,-7.849 17.5313,-17.5312 v -110 c 0,-0.3245 -0.012,-0.6488 -0.031,-0.9727 0.018,-0.3004 0.029,-0.6013 0.031,-0.9023 1e-4,-74.075 -60.55,-134.4374 -134.625,-134.4375 -39.9177,1e-4 -75.8465,17.5354 -100.5,45.293 V 2115 c 0.1749,-9.8356 -7.7881,-17.8833 -17.625,-17.8125 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><path fill=\"#5ed3ee\" d=\"m 7765.8027,2098.123 c -0.038,5e-4 -0.077,0 -0.1152,0 -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 V 2361 c 0,41.9333 -26.0162,79.2671 -65.4375,93.5625 -39.4213,14.2954 -83.5594,2.3741 -110.4375,-29.8125 -26.878,-32.1865 -30.6579,-77.5719 -9.5625,-113.8125 21.0955,-36.2406 62.4262,-55.2897 103.6875,-47.8125 9.5173,1.7397 18.6463,-4.5547 20.4043,-14.0686 1.758,-9.514 -4.5187,-18.6551 -14.0293,-20.4314 -55.6354,-10.082 -111.8054,15.8219 -140.25,64.6875 -28.4446,48.8656 -23.3039,110.5381 12.9375,153.9375 36.2415,43.3993 95.908,59.4005 149.0625,40.125 44.7667,-16.2337 77.0286,-54.658 86.0977,-100.1562 0,0 1.9812,-4.7686 2.5898,-9.1563 0.6086,-4.3877 0,-17.0625 0,-17.0625 v -245.0625 c 0.1769,-9.865 -7.8313,-17.9263 -17.6973,-17.8145 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><g fill=\"none\" stroke-width=\".365\" transform=\"translate(-449.265 -83.342) scale(5.91262)\"><circle cx=\"79.677\" cy=\"16.823\" r=\".645\" stroke=\"#c6ea70\"/><path stroke=\"#dff3b1\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 80.808656,18.234303 a 1.8088735,1.8088735 0 0 1 -2.309183,-0.03848 1.8088735,1.8088735 0 0 1 -0.389241,-2.276466 1.8088735,1.8088735 0 0 1 2.165211,-0.803532\"/><path stroke=\"#f2fadf\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 82.216272,17.248644 a 2.455857,2.4425189 0 0 1 -2.468226,2.016391 2.455857,2.4425189 0 0 1 -2.383743,-2.114547 2.455857,2.4425189 0 0 1 1.733108,-2.668494\"/></g></g></g></g></svg>",
                  "category": "Open Data Hub",
                  "support": "open data hub",
                  "docsLink": "http://REPLACE.THIS.URL",
                  "quickStart": "example-odhquickstart",
                  "getStartedLink": "http://REPLACE.THIS.URL",
                  "getStartedMarkDown": "# Header for Getting Started MarkDown section\nHere is where you add information about your ODHApplication in markdown format\n\n1. List item\n1. List item\n1. List item"
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhDocument",
                "metadata": {
                  "name": "example-odhdocument",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHDocument",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "type": "how-to",
                  "description": "This is an example of an ODHDocument custom resource",
                  "url": "http://REPLACE.THIS.URL",
                  "durationMinutes": 5
                }
              },
              {
                "apiVersion": "console.openshift.io/v1",
                "kind": "OdhQuickStart",
                "metadata": {
                  "name": "example-odhquickstart",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHQuickstart Name",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "durationMinutes": 5,
                  "description": "Description of an example quickstart",
                  "introduction": "### Introduction Header\nThis is the body of the introduction for an ODHQuickstart",
                  "tasks": [
                    {
                      "title": "Title of the first task",
                      "description": "### Description header for the first task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the first task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    },
                    {
                      "title": "Title of the second task",
                      "description": "### Description header for the second task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the second task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    }
                  ],
                  "conclusion": "Message for the conclusion of the ODHQuickstart",
                  "nextQuickStart": null
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v1.11.0
          createdAt: "2023-10-31T00:00:00Z"
          description: Open Data Hub Operator for deployment and management of Open
            Data Hub
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
          support: Open Data Hub
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AcceleratorProfile is the Schema for the acceleratorprofiles
            displayName: AcceleratorProfile
            kind: AcceleratorProfile
            name: acceleratorprofiles.dashboard.opendatahub.io
            version: v1alpha
          - description: KfDef is the Schema for the kfdefs API
            displayName: Kf Def
            kind: KfDef
            name: kfdefs.kfdef.apps.kubeflow.org
            version: v1
          - description: OdhApplication is the Schema for the odhapplications
            displayName: Open Data Hub Dashboard Application
            kind: OdhApplication
            name: odhapplications.dashboard.opendatahub.io
            version: v1
          - description: OdhDashboardConfig is the Schema for the odhdashboardconfigs
            displayName: Open Data Hub Dashboard Configuration
            kind: OdhDashboardConfig
            name: odhdashboardconfigs.opendatahub.io
            version: v1alpha
          - description: OdhDocument is the Schema for the odhdocuments
            displayName: Open Data Hub Dashboard Documentation
            kind: OdhDocument
            name: odhdocuments.dashboard.opendatahub.io
            version: v1
          - description: Extension for guiding user through various workflows in the
              Open Data Hub dashboard.
            displayName: Open Data Hub Dashboard QuickStart
            kind: OdhQuickStart
            name: odhquickstarts.console.openshift.io
            version: v1
        description: |
          The Open Data Hub is a machine-learning-as-a-service platform built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open Data Hub integrates multiple AI/ML open source components into one operator that can easily be downloaded and installed by OpenShift users.

          Open Data Hub operator allows users to install and manage components of the Open Data Hub. Users can mix and match tools from each project to fulfill the needs of their use case. Each of the
          projects share some components, but can be mostly seen as an extension of each other to provide a complete solution for both novice and skilled enterprise users.

          ### Components
          * Open Data Hub Dashboard - A web dashboard that displays installed Open Data Hub components with easy access to component UIs and documentation
          * ODH Notebook Controller - Secure management of Jupyter Notebook in Kubernetes environments built on top of Kubeflow Notebook Controller with support for OAuth
          * Jupyter Notebooks - JupyterLab notebook that provide Python support for GPU workloads
          * Data Science Pipelines - Pipeline solution for end to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton
          * Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh, a general-purpose model serving management/routing layer.
          * TrustyAI - TrustyAI is an open source AI Explainability (XAI) Toolkit

          To install one or multiple of these components use the default KfDef provided with the operator.
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Open Data Hub
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/opendatahub/opendatahub-operator:v1.11.0
        version: 1.11.0
      entries:
      - name: opendatahub-operator.v1.11.0
        version: 1.11.0
      - name: opendatahub-operator.v1.10.1
        version: 1.10.1
      - name: opendatahub-operator.v1.10.0
        version: 1.10.0
      - name: opendatahub-operator.v1.9.0
        version: 1.9.0
      - name: opendatahub-operator.v1.8.0
        version: 1.8.0
      - name: opendatahub-operator.v1.7.0
        version: 1.7.0
      - name: opendatahub-operator.v1.6.0
        version: 1.6.0
      - name: opendatahub-operator.v1.5.0
        version: 1.5.0
      - name: opendatahub-operator.v1.4.2
        version: 1.4.2
      - name: opendatahub-operator.v1.4.1
        version: 1.4.1
      - name: opendatahub-operator.v1.4.0
        version: 1.4.0
      - name: opendatahub-operator.v1.3.0
        version: 1.3.0
      - name: opendatahub-operator.v1.2.0
        version: 1.2.0
      - name: opendatahub-operator.v1.1.2
        version: 1.1.2
      - name: opendatahub-operator.v1.1.1
        version: 1.1.1
      name: rolling
    - currentCSV: opendatahub-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kfdef.apps.kubeflow.org/v1",
                "kind": "KfDef",
                "metadata": {
                  "name": "opendatahub",
                  "namespace": "opendatahub"
                },
                "spec": {
                  "applications": [
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-common"
                        }
                      },
                      "name": "odh-common"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-dashboard"
                        }
                      },
                      "name": "odh-dashboard"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "prometheus/cluster"
                        }
                      },
                      "name": "prometheus-cluster"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "prometheus/operator"
                        }
                      },
                      "name": "prometheus-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "grafana/cluster"
                        }
                      },
                      "name": "grafana-cluster"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "grafana/grafana"
                        }
                      },
                      "name": "grafana-instance"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-notebook-controller"
                        }
                      },
                      "name": "odh-notebook-controller"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "notebook-images"
                        }
                      },
                      "name": "notebook-images"
                    },
                    {
                      "kustomizeConfig": {
                        "overlays": [
                          "odh-model-controller"
                        ],
                        "repoRef": {
                          "name": "manifests",
                          "path": "model-mesh"
                        }
                      },
                      "name": "model-mesh"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "data-science-pipelines-operator"
                        }
                      },
                      "name": "data-science-pipelines-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "trustyai-service"
                        }
                      },
                      "name": "trustyai"
                    }
                  ],
                  "repos": [
                    {
                      "name": "manifests",
                      "uri": "https://github.com/opendatahub-io/odh-manifests/tarball/v1.5"
                    }
                  ]
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhApplication",
                "metadata": {
                  "name": "example-odhapplication",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started,Model development,Model training,Notebook environments,Package management,Data Management"
                  }
                },
                "spec": {
                  "displayName": "ODHAPPLICATION DISPLAY NAME",
                  "provider": "PROVIDER_NAME",
                  "description": "Description of your ODHApplication",
                  "kfdefApplications": [
                    "<NAME OF THE ODH APPLICATION IN KFDEF>"
                  ],
                  "route": "<APPLICATION_OCP_ROUTE_OBJECT_NAME>",
                  "img": "<svg xmlns=\"http://www.w3.org/2000/svg\" role=\"img\" viewBox=\"-1.13 -1.13 58.50 48.94\"><title>Open Data Hub logo</title><g><g><g fill=\"#4d5258\" stroke-width=\".006\"><path fill=\"#bbb\" d=\"M112.30231 22.797083c-.2034 0-.35737.143517-.35737.337407 0 .192939.15397.339307.35737.339307.20339 0 .35736-.147318.35736-.339307 0-.19294-.15397-.337407-.35736-.337407zm.002.131161c.1112 0 .20149.08934.20149.206246 0 .116904-.0903.208146-.20149.208146-.1112 0-.2053-.09124-.2053-.208146 0-.116904.0931-.206246.2053-.206246zM113.04141 22.801836h-.28228v.665308h.15017v-.190088h.13211c.16918 0 .26612-.08934.26612-.244263 0-.147319-.0969-.230957-.26612-.230957zm-.008.349762h-.12451v-.225255h.12451c.0836 0 .13116.03707.13116.111202 0 .07603-.0475.114053-.13116.114053zM113.40328 22.801836v.665308h.51703v-.124508h-.36687v-.147318h.32125V23.07081h-.32125v-.144467h.35642v-.124507zM114.03814 22.801836v.665308h.14352v-.426748l.3222.426748h.13686v-.665308h-.14257v.427698L114.175 22.801836z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/><path d=\"M115.29685 22.801836h-.28228v.665308h.27658c.21004 0 .35546-.136863.35546-.332654 0-.196741-.14351-.332654-.34976-.332654zm.004.538899h-.13591v-.412491h.12926c.11405 0 .19959.08459.19959.207196 0 .121657-.0817.205295-.19294.205295zM116.21878 23.467144h.16062l-.28418-.665308h-.15492l-.28894.665308h.15493l.0523-.12926h.30794zm-.3108-.249966l.1055-.25852.10455.25852zM116.31525 22.801836v.127359h.20244v.537949h.15017v-.537949h.2034v-.127359zM117.37527 23.467144h.16063l-.28419-.665308h-.15492l-.28893.665308h.15492l.0523-.12926h.30795zm-.31079-.249966l.1055-.25852.10454.25852zM117.83433 22.801836v.665308h.15017v-.262322h.30129v.262322h.15017v-.665308h-.15017v.277528h-.30129v-.277528zM118.72025 23.188665v-.386829h-.15017v.386829c0 .175831.11785.285132.30224.285132.18343 0 .29844-.109301.29844-.285132v-.386829h-.15017v.386829c0 .09599-.0589.153971-.14827.153971-.0893 0-.15207-.05798-.15207-.153971zM119.61059 22.801836h-.30604v.665308h.31649c.15492 0 .25092-.07033.25092-.184385 0-.08744-.0532-.152071-.13876-.168228.0713-.01806.115-.07413.115-.146368 0-.102648-.0912-.166327-.23761-.166327zm-.008.268024h-.14826v-.145418h.14826c.0561 0 .0913.02566.0913.07128 0 .04657-.0352.07413-.0913.07413zm0 .274677h-.14826v-.158723h.14826c.0713 0 .11406.02946.11406.07794 0 .05132-.0428.08079-.11406.08079z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g><g fill=\"#58595b\" stroke-width=\".01\"><path d=\"M112.19192 24.15027h.0381l-.13717-.306759h-.0364l-.1376.306759h.0372l.0342-.07844h.16696zm-.18712-.110871l.0688-.157763.0697.157763zM112.27858 23.843511v.306759h.0351v-.306759zM112.64205 23.843511h-.11525v.306759h.0351v-.09729h.0802c.0758 0 .11964-.039.11964-.10649 0-.0653-.0438-.102983-.11964-.102983zm-.001.177044h-.0789V23.87594h.0789c.0552 0 .0872.02454.0872.07143 0 .0482-.032.07318-.0872.07318zM112.81775 23.825105v.325165h.0342v-.325165zM113.1137 24.15027l-.00044-.152503c-.00044-.04908-.032-.07932-.089-.07932-.0394 0-.0644.01139-.0942.03111l.0145.0241c.0241-.01709.0478-.02673.0732-.02673.0408 0 .0618.0206.0618.05434v.01446h-.0758c-.0543.000439-.0863.02586-.0863.06617 0 .03944.0307.06968.0784.06968.0368 0 .0662-.01227.0846-.03725v.03593zm-.11131-.02717c-.0324 0-.053-.01709-.053-.04251 0-.02585.0202-.03769.0587-.03769h.0719v.02805c-.009.03287-.0386.05215-.0776.05215zM113.29915 24.108201c-.0127.0074-.025.01139-.0368.01183-.0197 0-.0307-.01139-.0307-.03944v-.127086h.0732v-.02673h-.0732v-.06048h-.0337v.06048h-.0333v.02673h.0333v.132783c0 .04558.025.06573.0587.0653.0184 0 .0359-.0061.0521-.01709zM113.44228 23.854466c.0118 0 .0245.0044.0351.01183l.014-.02629c-.014-.01052-.032-.01622-.0495-.01622-.0386-.000438-.0741.02717-.0741.07713v.02586h-.0298v.02673h.0298v.196764h.0337v-.196764h.0662v-.02673h-.0662v-.03024c0-.02717.0184-.04251.0408-.04207zM113.59236 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM113.80323 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM114.22747 23.918448c-.0443.000438-.0767.01972-.0911.06004-.01-.03813-.0394-.06004-.0828-.06004-.0416.000438-.0727.01753-.0881.05346v-.05215h-.0337v.230508h.0337v-.113501c0-.04996.0285-.08458.0758-.08589.0403 0 .064.02454.064.06617v.133221h.0346v-.113501c0-.04996.0276-.08458.0754-.08589.0399 0 .064.02454.064.06617v.133221h.0342V24.00828c0-.05609-.032-.08984-.0859-.08984zM114.63474 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0487.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11088-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM114.89934 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM115.02844 23.919762l.0863.230508h.0355l.0675-.188876.0679.188876h.0355l.0855-.230508h-.0355l-.0675.195888-.0679-.195888h-.0359l-.067.195888-.0684-.195888zM115.53009 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM115.73075 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM115.95034 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11482.116569.0386 0 .071-.01402.0929-.03813l-.0188-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18843c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM116.29474 23.825105v.144615c-.0171-.03287-.0491-.05171-.0903-.05171-.0653 0-.11087.04777-.11087.116569 0 .0688.0456.117445.11131.117445.0412 0 .0727-.01884.0898-.05127v.04952h.0342v-.325165zm-.0833.297557c-.0491 0-.0837-.03637-.0837-.08721 0-.05083.0346-.08765.0837-.08765.0486 0 .0833.03681.0833.08765 0 .05083-.0346.08721-.0833.08721zM116.65984 23.918448c-.0403 0-.0719.01884-.089.05083v-.144177h-.0342v.325165h.0342v-.04952c.0171.03199.0486.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM116.89657 24.183138l.10605-.263376h-.0351l-.0732.191944-.0767-.191944h-.0359l.0955.228756-.01.02191c-.009.02191-.0241.03331-.0421.03331-.0114 0-.0219-.0035-.0324-.01095l-.0145.02717c.014.01139.0294.01665.0486.01665.0307 0 .0548-.01841.0692-.05346zM117.29947 23.842634c-.0898 0-.16083.06793-.16083.154257 0 .08633.071.154694.16083.154694.0894 0 .16039-.06836.16039-.154694 0-.08633-.071-.154257-.16039-.154257zm0 .03243c.0684 0 .12446.0539.12446.121828 0 .06792-.0561.122703-.12446.122703-.0684 0-.12533-.05478-.12533-.122703 0-.06793.057-.121828.12533-.121828zM117.64414 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0486.05083.0898.05083.0653 0 .10999-.04777.10999-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM117.9061 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM118.19616 23.918448c-.0421.000438-.0732.01709-.0894.05215v-.05084h-.0342v.230508h.0342v-.12358c.004-.0447.032-.07494.0776-.07537.0408 0 .0653.02454.0653.06573v.133221h.0342v-.141986c0-.05609-.0324-.08984-.0876-.08984zM118.65836 23.901795l.0149-.03111c-.0254-.01709-.0618-.02848-.0973-.02848-.0635 0-.10605.03068-.10605.07888 0 .109995.17266.06661.17222.148559 0 .0298-.0276.04777-.0719.04777-.0346 0-.0736-.01578-.1021-.04295l-.0158.03024c.0289.02761.0727.04689.11745.04689.0657 0 .11087-.03287.11087-.0837.00044-.11131-.17223-.06968-.17223-.149436 0-.02629.0259-.04207.0653-.04207.0245 0 .057.0075.0846.02542zM118.84168 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .11701-.04777.11701-.117007 0-.06836-.0487-.11613-.11701-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM119.18446 23.919762v.113063c0 .04996-.0281.08502-.0758.08589-.0399 0-.064-.02454-.064-.06573v-.133222h-.0342v.141986c0 .05565.0324.08984.0863.08984.0416-.000438.0723-.01753.0876-.05346v.05215h.0342v-.230508zM119.33586 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM119.62557 23.976294l.0188-.02366c-.021-.02147-.0517-.03418-.089-.03418-.0666 0-.11438.0482-.11438.116568 0 .0688.0473.116569.11438.116569.0399 0 .0719-.0149.0925-.03988l-.0197-.02016c-.0162.01928-.0412.0298-.0714.0298-.0478 0-.082-.03506-.082-.08633 0-.05083.0342-.08589.082-.08589.028 0 .0513.01008.0688.02717zM119.78997 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g></g><g><g><path fill=\"#a586f3\" d=\"m 7808.8125,2097.1875 c -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 v 329.875 c -0.8781,10.2195 7.1804,19 17.4375,19 10.2571,0 19.0227,-8.7805 18.1446,-19 v -77.7324 l 0.4179,-6.1426 c 0,-55.1597 44.2153,-99.5625 99.375,-99.5625 55.1596,0 99.5625,44.4028 99.5625,99.5625 0.01,0.1837 0.015,0.3674 0.027,0.5508 -0.026,0.4409 -0.035,0.8826 -0.027,1.3242 v 110 c 0,9.6822 7.849,17.5312 17.5312,17.5312 9.6823,0 17.5313,-7.849 17.5313,-17.5312 v -110 c 0,-0.3245 -0.012,-0.6488 -0.031,-0.9727 0.018,-0.3004 0.029,-0.6013 0.031,-0.9023 1e-4,-74.075 -60.55,-134.4374 -134.625,-134.4375 -39.9177,1e-4 -75.8465,17.5354 -100.5,45.293 V 2115 c 0.1749,-9.8356 -7.7881,-17.8833 -17.625,-17.8125 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><path fill=\"#5ed3ee\" d=\"m 7765.8027,2098.123 c -0.038,5e-4 -0.077,0 -0.1152,0 -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 V 2361 c 0,41.9333 -26.0162,79.2671 -65.4375,93.5625 -39.4213,14.2954 -83.5594,2.3741 -110.4375,-29.8125 -26.878,-32.1865 -30.6579,-77.5719 -9.5625,-113.8125 21.0955,-36.2406 62.4262,-55.2897 103.6875,-47.8125 9.5173,1.7397 18.6463,-4.5547 20.4043,-14.0686 1.758,-9.514 -4.5187,-18.6551 -14.0293,-20.4314 -55.6354,-10.082 -111.8054,15.8219 -140.25,64.6875 -28.4446,48.8656 -23.3039,110.5381 12.9375,153.9375 36.2415,43.3993 95.908,59.4005 149.0625,40.125 44.7667,-16.2337 77.0286,-54.658 86.0977,-100.1562 0,0 1.9812,-4.7686 2.5898,-9.1563 0.6086,-4.3877 0,-17.0625 0,-17.0625 v -245.0625 c 0.1769,-9.865 -7.8313,-17.9263 -17.6973,-17.8145 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><g fill=\"none\" stroke-width=\".365\" transform=\"translate(-449.265 -83.342) scale(5.91262)\"><circle cx=\"79.677\" cy=\"16.823\" r=\".645\" stroke=\"#c6ea70\"/><path stroke=\"#dff3b1\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 80.808656,18.234303 a 1.8088735,1.8088735 0 0 1 -2.309183,-0.03848 1.8088735,1.8088735 0 0 1 -0.389241,-2.276466 1.8088735,1.8088735 0 0 1 2.165211,-0.803532\"/><path stroke=\"#f2fadf\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 82.216272,17.248644 a 2.455857,2.4425189 0 0 1 -2.468226,2.016391 2.455857,2.4425189 0 0 1 -2.383743,-2.114547 2.455857,2.4425189 0 0 1 1.733108,-2.668494\"/></g></g></g></g></svg>",
                  "category": "Open Data Hub",
                  "support": "open data hub",
                  "docsLink": "http://REPLACE.THIS.URL",
                  "quickStart": "example-odhquickstart",
                  "getStartedLink": "http://REPLACE.THIS.URL",
                  "getStartedMarkDown": "# Header for Getting Started MarkDown section\nHere is where you add information about your ODHApplication in markdown format\n\n1. List item\n1. List item\n1. List item"
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhDocument",
                "metadata": {
                  "name": "example-odhdocument",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHDocument",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "type": "how-to",
                  "description": "This is an example of an ODHDocument custom resource",
                  "url": "http://REPLACE.THIS.URL",
                  "durationMinutes": 5
                }
              },
              {
                "apiVersion": "console.openshift.io/v1",
                "kind": "OdhQuickStart",
                "metadata": {
                  "name": "example-odhquickstart",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHQuickstart Name",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "durationMinutes": 5,
                  "description": "Description of an example quickstart",
                  "introduction": "### Introduction Header\nThis is the body of the introduction for an ODHQuickstart",
                  "tasks": [
                    {
                      "title": "Title of the first task",
                      "description": "### Description header for the first task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the first task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    },
                    {
                      "title": "Title of the second task",
                      "description": "### Description header for the second task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the second task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    }
                  ],
                  "conclusion": "Message for the conclusion of the ODHQuickstart",
                  "nextQuickStart": null
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v1.5.0
          createdAt: "2022-10-18T00:00:00Z"
          description: Open Data Hub Operator for deployment and management of Open
            Data Hub
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
          support: Open Data Hub
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KfDef is the Schema for the kfdefs API
            displayName: Kf Def
            kind: KfDef
            name: kfdefs.kfdef.apps.kubeflow.org
            version: v1
          - description: OdhApplication is the Schema for the odhapplications
            displayName: Open Data Hub Dashboard Application
            kind: OdhApplication
            name: odhapplications.dashboard.opendatahub.io
            version: v1
          - description: OdhDashboardConfig is the Schema for the odhdashboardconfigs
            displayName: Open Data Hub Dashboard Configuration
            kind: OdhDashboardConfig
            name: odhdashboardconfigs.opendatahub.io
            version: v1alpha
          - description: OdhDocument is the Schema for the odhdocuments
            displayName: Open Data Hub Dashboard Documentation
            kind: OdhDocument
            name: odhdocuments.dashboard.opendatahub.io
            version: v1
          - description: Extension for guiding user through various workflows in the
              Open Data Hub dashboard.
            displayName: Open Data Hub Dashboard QuickStart
            kind: OdhQuickStart
            name: odhquickstarts.console.openshift.io
            version: v1
        description: |
          The Open Data Hub is a machine-learning-as-a-service platform built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open Data Hub integrates multiple AI/ML open source components into one operator that can easily be downloaded and installed by OpenShift users.

          Open Data Hub operator allows users to install and manage components of the Open Data Hub. Users can mix and match tools from each project to fulfill the needs of their use case. Each of the
          projects share some components, but can be mostly seen as an extension of each other to provide a complete solution for both novice and skilled enterprise users.

          ### Components
          * Open Data Hub Dashboard - A web dashboard that displays installed Open Data Hub components with easy access to component UIs and documentation
          * ODH Notebook Controller - Secure management of Jupyter Notebook in Kubernetes environments built on top of Kubeflow Notebook Controller with support for OAuth
          * Jupyter Notebooks - JupyterLab notebook that provide Python support for GPU workloads
          * Data Science Pipelines - Pipeline solution for end to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton
          * Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh, a general-purpose model serving management/routing layer.

          To install one or multiple of these components use the default KfDef provided with the operator.

          ### Available Channels

          #### Stable

          Channel `stable` offers the major releases of Open Data Hub operator and ODH manifests. It is based on Kubeflow Operator and Kustomize for deployment configuration. This version is compatible with [Open Data Hub manifests](https://github.com/opendatahub-io/odh-manifests/) as well as [Kubeflow manifests](https://github.com/opendatahub-io/manifests). This channel will offer new components and features that have been thoroughly tested by consumers of the `rolling` channel

          NOTE: The `stable` channel will be deprecated starting with the ODH 1.6 release and we will only be publishing updates on the `rolling` channel for all future releases

          #### Rolling

          Channel `rolling` will offer the latest release of Open Data Hub operator and manifests. Subscribing to this branch will provide faster updates and access to components that will be in developement and testing for the next stable release
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io
        maintainers:
        - email: maintainers@lists.opendatahub.io
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Open Data Hub
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/opendatahub/opendatahub-operator:v1.5.0
        version: 1.5.0
      entries:
      - name: opendatahub-operator.v1.5.0
        version: 1.5.0
      - name: opendatahub-operator.v1.4.2
        version: 1.4.2
      - name: opendatahub-operator.v1.4.1
        version: 1.4.1
      - name: opendatahub-operator.v1.4.0
        version: 1.4.0
      - name: opendatahub-operator.v1.3.0
        version: 1.3.0
      - name: opendatahub-operator.v1.2.0
        version: 1.2.0
      - name: opendatahub-operator.v1.1.2
        version: 1.1.2
      - name: opendatahub-operator.v1.1.1
        version: 1.1.1
      name: stable
    defaultChannel: fast
    packageName: opendatahub-operator
    provider:
      name: ODH
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Thales Group
      provider-url: https://cpl.thalesgroup.com/data-protection
    name: cte-k8s-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cte-k8s-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cte-k8s-operator.csi.cte.cpl.thalesgroup.com/v1",
                "kind": "CteK8sOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cte-k8s-operator"
                  },
                  "name": "ctek8soperator-sample"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-06-05T08:24:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CteK8sOperator is the Schema for the ctek8soperators API
            displayName: Cte K8s Operator
            kind: CteK8sOperator
            name: ctek8soperators.cte-k8s-operator.csi.cte.cpl.thalesgroup.com
            version: v1
        description: |-
          # Kubernetes Operator to deploy CTE for Kubernetes on an OCP/K8s Cluster.
          This Operator deploys CTE for Kubernetes on the OCP/K8s Cluster.

          ## About this operator

          It is a GO based operator that manages the deployment, upgrade and
          deletion of the CTE for Kubernetes on all the nodes of the Cluster on
          which it is installed.

          ## Prerequisites

          Applicable only on OCP cluster
            For an OCP cluster the operator requires a SecurityContextConstraints (SCC)
            to be created before installation. To create the SecurityContextConstraints
            you can use a definition provided by the deployment scripts at
            https://github.com/thalescpl-io/ciphertrust-transparent-encryption-kubernetes

            Look for cte-csi-scc.yaml file in deploy/kubernetes/X.Y.Z/operator-deploy/
            Where X.Y.Z is the last released version of the deploy scripts.

          For more information refer to the guide at:
            https://thalesdocs.com/ctp/cte-con/cte-k8s/latest/cte-4-k8s/index.html
        displayName: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Encryption
        - Data Protection
        - Security
        - Kubernetes
        - Openshift
        links:
        - name: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          url: https://cpl.thalesgroup.com/encryption/ciphertrust-transparent-data-encryption-kubernetes
        maintainers:
        - email: technical.support.DIS@thalesgroup.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Thales Group
          url: https://cpl.thalesgroup.com/data-protection
        relatedImages:
        - docker.io/thalesciphertrust/ciphertrust-transparent-encryption-kubernetes-operator@sha256:d58cbb65d758fd0b9d4cfba4cb48a35bdaebe2e239fb25edb9b491cd8c2c3f73
        version: 1.6.5
      entries:
      - name: cte-k8s-operator.v1.6.5
        version: 1.6.5
      name: candidate
    - currentCSV: cte-k8s-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cte-k8s-operator.csi.cte.cpl.thalesgroup.com/v1",
                "kind": "CteK8sOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cte-k8s-operator"
                  },
                  "name": "ctek8soperator-sample"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-06-05T08:24:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CteK8sOperator is the Schema for the ctek8soperators API
            displayName: Cte K8s Operator
            kind: CteK8sOperator
            name: ctek8soperators.cte-k8s-operator.csi.cte.cpl.thalesgroup.com
            version: v1
        description: |-
          # Kubernetes Operator to deploy CTE for Kubernetes on an OCP/K8s Cluster.
          This Operator deploys CTE for Kubernetes on the OCP/K8s Cluster.

          ## About this operator

          It is a GO based operator that manages the deployment, upgrade and
          deletion of the CTE for Kubernetes on all the nodes of the Cluster on
          which it is installed.

          ## Prerequisites

          Applicable only on OCP cluster
            For an OCP cluster the operator requires a SecurityContextConstraints (SCC)
            to be created before installation. To create the SecurityContextConstraints
            you can use a definition provided by the deployment scripts at
            https://github.com/thalescpl-io/ciphertrust-transparent-encryption-kubernetes

            Look for cte-csi-scc.yaml file in deploy/kubernetes/X.Y.Z/operator-deploy/
            Where X.Y.Z is the last released version of the deploy scripts.

          For more information refer to the guide at:
            https://thalesdocs.com/ctp/cte-con/cte-k8s/latest/cte-4-k8s/index.html
        displayName: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Encryption
        - Data Protection
        - Security
        - Kubernetes
        - Openshift
        links:
        - name: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          url: https://cpl.thalesgroup.com/encryption/ciphertrust-transparent-data-encryption-kubernetes
        maintainers:
        - email: technical.support.DIS@thalesgroup.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Thales Group
          url: https://cpl.thalesgroup.com/data-protection
        relatedImages:
        - docker.io/thalesciphertrust/ciphertrust-transparent-encryption-kubernetes-operator@sha256:d58cbb65d758fd0b9d4cfba4cb48a35bdaebe2e239fb25edb9b491cd8c2c3f73
        version: 1.6.5
      entries:
      - name: cte-k8s-operator.v1.6.5
        version: 1.6.5
      name: stable
    defaultChannel: stable
    packageName: cte-k8s-operator
    provider:
      name: Thales Group
      url: https://cpl.thalesgroup.com/data-protection
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-team-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-team-operator.v2024.10.2-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": null,
                        "host": null,
                        "process": null
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master",
                        "operator": "",
                        "value": ""
                      }
                    ]
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": null
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:12c1d2e9347ec447196a618b481a0970dacac7a8a72a3b38ec22efa69668d180
          createdAt: "2024-10-25T01:47:52Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2024.10.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2023.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                   operator: ""
                   value: ""

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:7c4868db021beaffa762eb0de43739bbc865a4bf5900f1d17f78caed8612c073
        - registry.connect.redhat.com/ca/uma-operator@sha256:12c1d2e9347ec447196a618b481a0970dacac7a8a72a3b38ec22efa69668d180
        version: 2024.10.2-12
      entries:
      - name: uma-team-operator.v2024.10.2-12
        version: 2024.10.2-12
      name: "2024.10"
    - currentCSV: uma-team-operator.v2025.1.3-33
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": null,
                        "host": null,
                        "process": null
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.1.1.33,990033",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master",
                        "operator": "",
                        "value": ""
                      }
                    ]
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": null
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:78b66a291f58aa94123cdd59f2aa07dce4928162c481fe9de5ce524782bc93de
          createdAt: "2025-02-18T05:35:06Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.1.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2023.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.1.1.26,990026"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                   operator: ""
                   value: ""

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:78b66a291f58aa94123cdd59f2aa07dce4928162c481fe9de5ce524782bc93de
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:ae91a28d57c8a359f81757e4bf79472f2ac189ac0ae3ee0d87e8ff9c433f1ccc
        version: 2025.1.3-33
      entries:
      - name: uma-team-operator.v2025.1.3-33
        version: 2025.1.3-33
      name: "2025.1"
    - currentCSV: uma-team-operator.v2025.2.1-25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.2.1.25,990025",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master",
                        "operator": "",
                        "value": ""
                      }
                    ]
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:d0a0090a871ff7e23d3c97b620a1b0d004a8317efe1581ba0437d073c5656825
          createdAt: "2025-03-05T05:51:44Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.2.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2023.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.1.1.26,990026"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                   operator: ""
                   value: ""

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:d0a0090a871ff7e23d3c97b620a1b0d004a8317efe1581ba0437d073c5656825
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:8bcf0dac889f38a0bf77eb9babe3f726e3b2564d5988c05bdc5c2d91ae189c61
        version: 2025.2.1-25
      entries:
      - name: uma-team-operator.v2025.2.1-25
        version: 2025.2.1-25
      name: "2025.2"
    - currentCSV: uma-team-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:60fb1f76b4d4aab569eec5f2c20f7a06d813759f5fb6ac952156b41c4d295a52
          createdAt: "2025-04-16T06:57:55Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.4.5
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.4.2.5,990005"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:60fb1f76b4d4aab569eec5f2c20f7a06d813759f5fb6ac952156b41c4d295a52
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        version: 2025.4.5-5
      entries:
      - name: uma-team-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-team-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:052055a6280f60a69caccf89f50fd8bf92da61cd59f1962f61d21467c4191970
          createdAt: "2025-07-24T14:54:58Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.5.2.12,990012"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:052055a6280f60a69caccf89f50fd8bf92da61cd59f1962f61d21467c4191970
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        version: 2025.5.4-12
      entries:
      - name: uma-team-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-team-operator.v2025.7.3-21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:e3ae2240aeb618f36f41073a5f65b03a7ad5dfdccc7f9feb830506fbd2bfd604
          createdAt: "2025-08-23T09:01:14Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.7.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.7.1.23,990023"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:e3ae2240aeb618f36f41073a5f65b03a7ad5dfdccc7f9feb830506fbd2bfd604
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.3-21
      entries:
      - name: uma-team-operator.v2025.7.3-21
        version: 2025.7.3-21
      name: "2025.7"
    - currentCSV: uma-team-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "hostmonitoring": {
                        "linux": {
                          "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:b8422d0f3a9e576948ced8244a91abc3283b26217e4c7efc7e2669ca5508930e
          createdAt: "2025-08-15T03:20:57Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 hostmonitoring:
                   linux:
                   # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                   groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.8.1.9,990009"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:b8422d0f3a9e576948ced8244a91abc3283b26217e4c7efc7e2669ca5508930e
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        version: 2025.8.2-9
      entries:
      - name: uma-team-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-team-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
          createdAt: "2025-09-24T12:24:28Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.12,990012"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-team-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-team-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "hostmonitoring": {
                        "linux": {
                          "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:8c12315745ad3564035a4db923698bf63fe32dddbc7d246f5a3867945305713e
          createdAt: "2025-07-30T08:27:50Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 hostmonitoring:
                   linux:
                   # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                   groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "99.99.0.serp_F152813,19"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:8c12315745ad3564035a4db923698bf63fe32dddbc7d246f5a3867945305713e
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        version: 2025.7.2-19
      entries:
      - name: uma-team-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-team-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:89e0505b5e526beffda9b13a1ca7b49e5ba144fd0166a7e171e9d9b084915a56
          createdAt: "2025-09-11T11:16:37Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.8,990008"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:89e0505b5e526beffda9b13a1ca7b49e5ba144fd0166a7e171e9d9b084915a56
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-team-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-team-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
          createdAt: "2025-09-24T12:24:28Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.12,990012"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-team-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-team-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ocs-client-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocs-client-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          console.openshift.io/plugins: '["odf-client-console"]'
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageClassClaim is the Schema for the storageclassclaims
              API
            displayName: Storage Class Claim
            kind: StorageClassClaim
            name: storageclassclaims.ocs.openshift.io
            version: v1alpha1
          - description: StorageClient is the Schema for the storageclients API
            displayName: Storage Client
            kind: StorageClient
            name: storageclients.ocs.openshift.io
            version: v1alpha1
        description: This operator is not intended to be directly installed
        displayName: OpenShift Data Foundation Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: OCS Client Operator
          url: https://github.com/red-hat-storage/ocs-client-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:975529d3fbdcf6f78cc1602333832f543da2918ac5c018d1ee6bea2b315106be
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:ef4276b80cacae825d4831de88d3fe6ba7cc324464087b7b28734bd2b8de3315
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel8@sha256:d83b44a1257e8015409cd261e8eed5bd8c4e581b788e8452f19114ba696a30c1
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:bc637bbbed3457c45fbb5cb1e8940b2a1dacfa64959e5bd0bfb58c5d669427fd
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:1dd7c0ef17d39ebf3311aee326202e09038cece1b7554cea23c1efe582620d0b
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:c9e5b48d84febf10ff4c9b78a8210df36c90905fbb92b6a6b5b32e0fd9a6a41d
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:9a0008f662b6afa9e2677a8a83bd662d806ad6f4d03fb49bccb78ed487ce4c7a
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:028120f2212738917f49e0330b60be5f79a208ce2974bea22b71f533aa6ce3be
        - registry.redhat.io/odf4/ocs-client-console-rhel9@sha256:ebf3f6098a9a0a06706d20fe8fc86513be34b98604bc70fd82910f0fcd17f191
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:ffcc0806ec2afb8b40d75c70ef2a6d98bb887879f5b3f573e900861316ce9e41
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel8@sha256:05cf61192ee75dc11f5777bebeacdd209232042445553329e14f542ac0084c71
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:e206dcc33863f4d8d0aa6df173a86cd50ee1d52e6bcf9efc43b39831acc20573
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:0dc0733df67bb49886ac9715b81bd4ec4a36330fe1afdde657587e1883f5ca7a
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel8@sha256:647074ada8cb88f7ed99ca14621d3b754db6a60346f12b7561cd035c05f283f7
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:390b527d10e9c7a8a9191251b0a1ba501128057cefaff35e950f0efe04ef3122
        - registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:cb93b68adb857c82f23f292cdee47d52656dc4fc5bb83fac8581110869cbe841
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:645d4fd09628f7ddcfc765e4ddacaa79fa6a7c35c03292ca2f6e317d48765d6b
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel8@sha256:1f6a17909f1fe6e09a01cad38b6a50b41d850b1c5cffd24ef9a87c8798ff15e7
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:1711a0aabd28937aac1424adebe96619a48b91f7f888f0588117f80891001da4
        version: 4.15.20-rhodf
      entries:
      - name: ocs-client-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: ocs-client-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: ocs-client-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: ocs-client-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: ocs-client-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: ocs-client-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: ocs-client-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: ocs-client-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: ocs-client-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: ocs-client-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: ocs-client-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: ocs-client-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: ocs-client-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: ocs-client-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: ocs-client-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: ocs-client-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: ocs-client-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: ocs-client-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: ocs-client-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: ocs-client-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: ocs-client-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-client-console"]'
          containerImage: registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:bd8454b5a0b235669732650ae8502831811793d93fbfbd0d0f53278fcdfed459
          createdAt: "2025-09-12T06:29:47Z"
          description: OpenShift Data Foundation client operator enables consumption
            of storage services from a remote centralized OpenShift Data Foundation
            provider cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["storageclaims.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/ocs-client-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageClaim is the Schema for the storageclaims API
            displayName: Storage Claim
            kind: StorageClaim
            name: storageclaims.ocs.openshift.io
            version: v1alpha1
          - description: StorageClient is the Schema for the storageclients API
            displayName: Storage Client
            kind: StorageClient
            name: storageclients.ocs.openshift.io
            version: v1alpha1
        description: OpenShift Data Foundation client operator enables consumption
          of storage services from a remote centralized OpenShift Data Foundation
          provider cluster.
        displayName: OpenShift Data Foundation Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: OCS Client Operator
          url: https://github.com/red-hat-storage/ocs-client-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:80c01633a6bfca44df37ac38f764035be6998fce965c8709e4f6d54d17eeaa7b
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:82a4b481028195cfd75471656fe40c1cbbade4c7b2d170d6bf9422d9d3b84fa7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel8@sha256:d83b44a1257e8015409cd261e8eed5bd8c4e581b788e8452f19114ba696a30c1
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel8@sha256:1f6a17909f1fe6e09a01cad38b6a50b41d850b1c5cffd24ef9a87c8798ff15e7
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:975529d3fbdcf6f78cc1602333832f543da2918ac5c018d1ee6bea2b315106be
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c11a599bfd09b9b22888e03eabdbe1054d4049718a591d2a45bf0ed89800111e
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6fee56851da3a87b9f7f477fc59e3385e118dc0b398861a8f13459c998a5af77
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:0afcb516eb528857887c6ee1a9cba081dc8f6b7d47d9f26a0dbb2dca0cefe07c
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:8d8326c74795c9c6bd8f33a9f9e81d2304a70b1f9be3c2f8d50e76be4b0292ff
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/odf4/ocs-client-console-rhel9@sha256:beb4f82d3923171b80aacb90dfe345a22626453bf2f3921b03a16839b204d377
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:1711a0aabd28937aac1424adebe96619a48b91f7f888f0588117f80891001da4
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:ef4276b80cacae825d4831de88d3fe6ba7cc324464087b7b28734bd2b8de3315
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:028120f2212738917f49e0330b60be5f79a208ce2974bea22b71f533aa6ce3be
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:645d4fd09628f7ddcfc765e4ddacaa79fa6a7c35c03292ca2f6e317d48765d6b
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:bc637bbbed3457c45fbb5cb1e8940b2a1dacfa64959e5bd0bfb58c5d669427fd
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:ffcc0806ec2afb8b40d75c70ef2a6d98bb887879f5b3f573e900861316ce9e41
        - registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:bd8454b5a0b235669732650ae8502831811793d93fbfbd0d0f53278fcdfed459
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:693a60d6a1162659213ae554b1c6ee200b0300a7da72c01415044cf507acab49
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:c9e5b48d84febf10ff4c9b78a8210df36c90905fbb92b6a6b5b32e0fd9a6a41d
        version: 4.16.16-rhodf
      entries:
      - name: ocs-client-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: ocs-client-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: ocs-client-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: ocs-client-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: ocs-client-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: ocs-client-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: ocs-client-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: ocs-client-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: ocs-client-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: ocs-client-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: ocs-client-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: ocs-client-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: ocs-client-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: ocs-client-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: ocs-client-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: ocs-client-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: ocs-client-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: ocs-client-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://your.domain
    name: amq-streams-proxy
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: alpha
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: amq-streams-3.0.x
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: amq-streams-3.x
    defaultChannel: alpha
    packageName: amq-streams-proxy
    provider:
      name: Red Hat
      url: https://your.domain
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-iam-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-iam-controller.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Role",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Group",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "OpenIDConnectProvider",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Policy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/iam-controller:1.5.2
          createdAt: "2025-10-03T22:22:57Z"
          description: AWS IAM controller is a service controller for managing IAM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Group represents the state of an AWS iam Group resource.
            displayName: Group
            kind: Group
            name: groups.iam.services.k8s.aws
            version: v1alpha1
          - description: InstanceProfile represents the state of an AWS iam InstanceProfile
              resource.
            displayName: InstanceProfile
            kind: InstanceProfile
            name: instanceprofiles.iam.services.k8s.aws
            version: v1alpha1
          - description: OpenIDConnectProvider represents the state of an AWS iam
              OpenIDConnectProvider resource.
            displayName: OpenIDConnectProvider
            kind: OpenIDConnectProvider
            name: openidconnectproviders.iam.services.k8s.aws
            version: v1alpha1
          - description: Policy represents the state of an AWS iam Policy resource.
            displayName: Policy
            kind: Policy
            name: policies.iam.services.k8s.aws
            version: v1alpha1
          - description: Role represents the state of an AWS iam Role resource.
            displayName: Role
            kind: Role
            name: roles.iam.services.k8s.aws
            version: v1alpha1
          - description: ServiceLinkedRole represents the state of an AWS iam ServiceLinkedRole
              resource.
            displayName: ServiceLinkedRole
            kind: ServiceLinkedRole
            name: servicelinkedroles.iam.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS iam User resource.
            displayName: User
            kind: User
            name: users.iam.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Identity and Access Management (IAM) resources in AWS from within your Kubernetes cluster.

          **About Amazon IAM**

          AWS Identity and Access Management (IAM) is a web service that helps you securely control access to AWS resources. You use IAM to control who is authenticated (signed in) and authorized (has permissions) to use resources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon IAM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iam
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon IAM Developer Resources
          url: https://aws.amazon.com/iam/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: iam maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/iam-controller:1.5.2
        version: 1.5.2
      entries:
      - name: ack-iam-controller.v1.5.2
        version: 1.5.2
      - name: ack-iam-controller.v1.5.1
        version: 1.5.1
      - name: ack-iam-controller.v1.5.0
        version: 1.5.0
      - name: ack-iam-controller.v1.4.4
        version: 1.4.4
      - name: ack-iam-controller.v1.4.3
        version: 1.4.3
      - name: ack-iam-controller.v1.4.2
        version: 1.4.2
      - name: ack-iam-controller.v1.4.1
        version: 1.4.1
      - name: ack-iam-controller.v1.4.0
        version: 1.4.0
      - name: ack-iam-controller.v1.3.22
        version: 1.3.22
      - name: ack-iam-controller.v1.3.21
        version: 1.3.21
      - name: ack-iam-controller.v1.3.20
        version: 1.3.20
      - name: ack-iam-controller.v1.3.19
        version: 1.3.19
      - name: ack-iam-controller.v1.3.18
        version: 1.3.18
      - name: ack-iam-controller.v1.3.17
        version: 1.3.17
      - name: ack-iam-controller.v1.3.15
        version: 1.3.15
      - name: ack-iam-controller.v1.3.14
        version: 1.3.14
      - name: ack-iam-controller.v1.3.13
        version: 1.3.13
      - name: ack-iam-controller.v1.3.12
        version: 1.3.12
      - name: ack-iam-controller.v1.3.11
        version: 1.3.11
      - name: ack-iam-controller.v1.3.10
        version: 1.3.10
      - name: ack-iam-controller.v1.3.9
        version: 1.3.9
      - name: ack-iam-controller.v1.3.8
        version: 1.3.8
      - name: ack-iam-controller.v1.3.6
        version: 1.3.6
      - name: ack-iam-controller.v1.3.2
        version: 1.3.2
      - name: ack-iam-controller.v1.3.1
        version: 1.3.1
      - name: ack-iam-controller.v1.3.0
        version: 1.3.0
      - name: ack-iam-controller.v1.2.6
        version: 1.2.6
      - name: ack-iam-controller.v1.2.5
        version: 1.2.5
      - name: ack-iam-controller.v1.2.3
        version: 1.2.3
      - name: ack-iam-controller.v1.2.2
        version: 1.2.2
      - name: ack-iam-controller.v1.2.1
        version: 1.2.1
      - name: ack-iam-controller.v1.2.0
        version: 1.2.0
      - name: ack-iam-controller.v1.1.3
        version: 1.1.3
      - name: ack-iam-controller.v1.1.2
        version: 1.1.2
      - name: ack-iam-controller.v1.1.1
        version: 1.1.1
      - name: ack-iam-controller.v1.1.0
        version: 1.1.0
      - name: ack-iam-controller.v1.0.0
        version: 1.0.0
      - name: ack-iam-controller.v0.1.2
        version: 0.1.2
      - name: ack-iam-controller.v0.1.1
        version: 0.1.1
      - name: ack-iam-controller.v0.1.0
        version: 0.1.0
      - name: ack-iam-controller.v0.0.22
        version: 0.0.22
      - name: ack-iam-controller.v0.0.21
        version: 0.0.21
      - name: ack-iam-controller.v0.0.20
        version: 0.0.20
      - name: ack-iam-controller.v0.0.19
        version: 0.0.19
      - name: ack-iam-controller.v0.0.18
        version: 0.0.18
      - name: ack-iam-controller.v0.0.17
        version: 0.0.17
      - name: ack-iam-controller.v0.0.16
        version: 0.0.16
      - name: ack-iam-controller.v0.0.15
        version: 0.0.15
      - name: ack-iam-controller.v0.0.14
        version: 0.0.14
      - name: ack-iam-controller.v0.0.13
        version: 0.0.13
      - name: ack-iam-controller.v0.0.12
        version: 0.0.12
      - name: ack-iam-controller.v0.0.11
        version: 0.0.11
      - name: ack-iam-controller.v0.0.10
        version: 0.0.10
      - name: ack-iam-controller.v0.0.9
        version: 0.0.9
      - name: ack-iam-controller.v0.0.8
        version: 0.0.8
      - name: ack-iam-controller.v0.0.7
        version: 0.0.7
      - name: ack-iam-controller.v0.0.6
        version: 0.0.6
      - name: ack-iam-controller.v0.0.5
        version: 0.0.5
      - name: ack-iam-controller.v0.0.4
        version: 0.0.4
      - name: ack-iam-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-iam-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: crunchy-postgres-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:c559c896e174674d151f93543ac92a1b5647eb4676313f7dc6cbb8faee786f18
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:4bdd0588ffbb3f628e7df1c72670e7a5087269438343b68779012bc3a13dea31
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-exporter@sha256:3b231902b2116cd38de05c7378202716c95813b9961fb8cee811b61d36cddf86
        - registry.connect.redhat.com/crunchydata/crunchy-upgrade@sha256:68bc93cce61d9b1e44822ae3e3ac8f7a9c2b712eddfaea090eabefacd60c67f6
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:0e283616c2d32e53258d62d02702bc4e179ef8c03933592154babed1a70bf666
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:56a5e3a5758982e086c9f78201a7137e0d934316f058a31cf30df44639a569a2
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:b2c3c74fdcc0e08f555e50c527f0d322292333c5d3dce061a0dee3680a1fbc56
        - registry.connect.redhat.com/crunchydata/crunchy-pgbackrest@sha256:b70da0035b74379f2507cfabdd9d1cc1e3c3eecb9066c22f292981c9a27916e1
        - registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
        - registry.connect.redhat.com/crunchydata/crunchy-pgadmin4@sha256:0d84acb1ce6a9b046e3f8b222a583ddcc91dda0a117c9f637ef173d0ac992f57
        - registry.connect.redhat.com/crunchydata/crunchy-pgbouncer@sha256:17cb58dec78447c2f70a414a2aeff6facbf921b42a0f366d0dbf7be95aa8987a
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:a7f62f367817fc08294a0b8b82ac9826a42c0ae668efdfacaa158a4287b8a688
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.4
        version: 5.4.4
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.4.0
        version: 5.4.0
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: crunchy-postgres-operator
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hbase-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hbase-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:e56051b66b0f419874d113c0320232dba7dd2547357256cf80eeaf18bbecb298
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hbase-operator@sha256:e56051b66b0f419874d113c0320232dba7dd2547357256cf80eeaf18bbecb298
        version: 24.11.1
      entries:
      - name: hbase-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: hbase-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:232ff30ffdf4bdc33f49d87950864e15436957510ba47f352dc0b133cb43279d
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hbase-operator@sha256:232ff30ffdf4bdc33f49d87950864e15436957510ba47f352dc0b133cb43279d
        version: 25.3.0
      entries:
      - name: hbase-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: hbase-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
        version: 25.7.0
      entries:
      - name: hbase-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hbase-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
        version: 25.7.0
      entries:
      - name: hbase-operator.v25.7.0
        version: 25.7.0
      - name: hbase-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hbase-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-secretsmanager-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-secretsmanager-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secretsmanager.services.k8s.aws/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/secretsmanager-controller:1.1.1
          createdAt: "2025-09-19T21:14:43Z"
          description: AWS Secrets Manager controller is a service controller for
            managing Secrets Manager resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret represents the state of an AWS secretsmanager Secret
              resource.
            displayName: Secret
            kind: Secret
            name: secrets.secretsmanager.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Secrets Manager resources in AWS from within your Kubernetes cluster.

          **About Amazon Secrets Manager**

          AWS Secrets Manager helps you manage, retrieve, and rotate database credentials, application credentials, OAuth tokens, API keys, and other secrets throughout their lifecycles. [Many AWS services that use secrets](https://docs.aws.amazon.com/secretsmanager/latest/userguide/integrating.html) store them in Secrets Manager.

          Secrets Manager helps you improve your security posture, because you no longer need hard-coded credentials in application source code. Storing the credentials in Secrets Manager helps avoid possible compromise by anyone who can inspect your application or the components. You replace hard-coded credentials with a runtime call to the Secrets Manager service to retrieve credentials dynamically when you need them.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Secrets Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secretsmanager
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Secrets Manager Developer Resources
          url: https://aws.amazon.com/secrets-manager/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: secretsmanager maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/secretsmanager-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-secretsmanager-controller.v1.1.1
        version: 1.1.1
      - name: ack-secretsmanager-controller.v1.1.0
        version: 1.1.0
      - name: ack-secretsmanager-controller.v1.0.14
        version: 1.0.14
      - name: ack-secretsmanager-controller.v1.0.13
        version: 1.0.13
      - name: ack-secretsmanager-controller.v1.0.12
        version: 1.0.12
      - name: ack-secretsmanager-controller.v1.0.11
        version: 1.0.11
      - name: ack-secretsmanager-controller.v1.0.10
        version: 1.0.10
      - name: ack-secretsmanager-controller.v1.0.9
        version: 1.0.9
      - name: ack-secretsmanager-controller.v1.0.8
        version: 1.0.8
      - name: ack-secretsmanager-controller.v1.0.7
        version: 1.0.7
      - name: ack-secretsmanager-controller.v1.0.6
        version: 1.0.6
      - name: ack-secretsmanager-controller.v1.0.5
        version: 1.0.5
      - name: ack-secretsmanager-controller.v1.0.4
        version: 1.0.4
      - name: ack-secretsmanager-controller.v1.0.3
        version: 1.0.3
      - name: ack-secretsmanager-controller.v1.0.1
        version: 1.0.1
      - name: ack-secretsmanager-controller.v1.0.0
        version: 1.0.0
      - name: ack-secretsmanager-controller.v0.0.11
        version: 0.0.11
      - name: ack-secretsmanager-controller.v0.0.10
        version: 0.0.10
      - name: ack-secretsmanager-controller.v0.0.9
        version: 0.0.9
      - name: ack-secretsmanager-controller.v0.0.8
        version: 0.0.8
      - name: ack-secretsmanager-controller.v0.0.7
        version: 0.0.7
      - name: ack-secretsmanager-controller.v0.0.2
        version: 0.0.2
      - name: ack-secretsmanager-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-secretsmanager-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Accuknox Inc.
      provider-url: https://accuknox.com
    name: accuknox-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: accuknox-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.accuknox.com/v1",
                "kind": "AccuknoxConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "accuknox-operator",
                    "app.kubernetes.io/instance": "accuknoxconfig",
                    "app.kubernetes.io/name": "accuknoxconfig-sample",
                    "app.kubernetes.io/part-of": "accuknox-operator"
                  },
                  "name": "accuknoxconfig-sample",
                  "namespace": "accuknox-agents"
                },
                "spec": {
                  "ppsHost": "PLEASE_FILL_IN",
                  "knoxGateway": "PLEASE_FILL_IN",
                  "spireHost": "PLEASE_FILL_IN"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Security
          containerImage: public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:b6fae25e229fb7fca757872fc0460642d52c4f9906f776afac50c54a69b05b68
          createdAt: "2024-02-09T17:06:30Z"
          operatorframework.io/suggested-namespace: accuknox-agents
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: public.ecr.aws/k9v9d5v2/accuknox-operator
          support: Accuknox Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccuknoxConfig is the Schema for the accuknoxconfigs API
            kind: AccuknoxConfig
            name: accuknoxconfigs.operator.accuknox.com
            version: v1
        description: |2-

          The AccuKnox Operator brings native K8s support for deploying and managing AccuKnox Components on a Kubernetes cluster.
          The AccuKnox Operator helps to easily onboard a k8s cluster to the AccuKnox SaaS.

          AccuKnox Operator deploys and manages the components:

          1. Discovery Engine

          2. Policy Enforcement Agent

          3. Shared Informer Agent

          4. Feeder Service

          5. Agents Operator

          ## Prerequisites for using this Operator

          1. Get AccuKnox licence - https://app.accuknox.com/sign-up

          2. After Successful Sign-up into AccuKnox Application.

             Please naviagate to Settings >> Manager Cluster >> Onboard Cluster >>

             After successful onboarding, Copy the join token created by AccuKnox

          3. Create a secret using jointoken by executing the below command:

          ```
          oc create secret generic accuknox-jointoken --from-literal=token=<joinToken> -n accuknox-agents
          ```

          ## Deploy AccuKnox Operator using AccuknoxConfig Custom Resource

          All the information to configure the AccuKnox Operator will be provided via Accuknox during Cluster Onboarding steps.

          ### AccuknoxConfig Spec

          | Spec | Description |
          | ---- | ----------- |
          | spireHost | "spire host url" |
          | ppsHost   |  "pps host url" |
          | knoxGateway | "knox gateway url" |
          | joinTokenSecretName | "join token secret name if it's different than accuknox-jointoken" |
          | joinTokenSecretKey | "key name used to store the jointoken with the secret" |

          *Example*

          ```
          apiVersion: operator.accuknox.com/v1
          kind: AccuknoxConfig
          metadata:
            name: accuknoxconfig-default
            namespace: accuknox-agents
          spec:
            ppsHost: <ppsHost url>
            knoxGateway: <knox gateway url>
            spireHost: <spire host url>
          ```
        displayName: Accuknox Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CNAPP
        - CSPM
        - CWPP
        - Cloud Native
        - Runtime Security
        - Zero-Trust
        maturity: alpha
        provider:
          name: Accuknox Inc.
          url: https://accuknox.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        - public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:b6fae25e229fb7fca757872fc0460642d52c4f9906f776afac50c54a69b05b68
        - public.ecr.aws/k9v9d5v2/feeder-service@sha256:f5bc340d9d3e34804dd54a03701d272e0ce5c04160d1aa41984cb3d877eb5b76
        - public.ecr.aws/k9v9d5v2/discovery-engine-discover@sha256:052efbe1064503853adab497415b9407f51578e1456931a8a0e60353aa53e0c2
        - public.ecr.aws/k9v9d5v2/discovery-engine-hardening@sha256:3eb28d44ef93e6ba331e9085e64475458e1c234017c77dc30b8bc802def434c3
        - public.ecr.aws/k9v9d5v2/agents-operator@sha256:656d6bac4371adef3e86658011ba68ae2f98514f6ddaea0667d4269eb88dd02e
        - public.ecr.aws/k9v9d5v2/shared-informer-agent@sha256:6d18d13312d4309b6e32b18c063d5b26f15fbf057fda2b9e6b7e8ebb1a5885e9
        - public.ecr.aws/k9v9d5v2/policy-enforcement-agent@sha256:f004fc3572a4d66da1bc0db9e05a279d761a42d6ed62f95de7a2bb42ce71e17b
        - public.ecr.aws/k9v9d5v2/discovery-engine-sumengine@sha256:4ca78f90b28b38badb0520f148ce702a3a6cd6b02bb62e8b1a6782a5b82e0868
        - public.ecr.aws/k9v9d5v2/discovery-engine-offloader@sha256:3c2291310e30e35a931d2af389f7ade656ce86f609d2e42ecdabf1000764695c
        version: 0.2.1
      entries:
      - name: accuknox-operator.v0.2.1
        version: 0.2.1
      name: alpha
    - currentCSV: accuknox-operator.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.accuknox.com/v1",
                "kind": "AccuknoxConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "accuknox-operator",
                    "app.kubernetes.io/instance": "accuknoxconfig",
                    "app.kubernetes.io/name": "accuknoxconfig-sample",
                    "app.kubernetes.io/part-of": "accuknox-operator"
                  },
                  "name": "accuknoxconfig-sample",
                  "namespace": "accuknox-agents"
                },
                "spec": {
                  "ppsHost": "PLEASE_FILL_IN",
                  "knoxGateway": "PLEASE_FILL_IN",
                  "spireHost": "PLEASE_FILL_IN"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Security
          containerImage: public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:a935ad806e82760bb838bada30d3ed1a54c8c61b95fe5c6109f48985c6faf4d7
          createdAt: "2024-08-23T04:25:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: accuknox-agents
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: docker.io/accuknox/accuknox-operator
          support: Accuknox Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccuknoxConfig is the Schema for the accuknoxconfigs API
            kind: AccuknoxConfig
            name: accuknoxconfigs.operator.accuknox.com
            version: v1
        description: |2-

          The AccuKnox Operator brings native K8s support for deploying and managing AccuKnox Components on a Kubernetes cluster.
          The AccuKnox Operator helps to easily onboard a k8s cluster to the AccuKnox SaaS.

          AccuKnox Operator deploys and manages the components:

          1. Discovery Engine

          2. Policy Enforcement Agent

          3. Shared Informer Agent

          4. Feeder Service

          5. Agents Operator

          ## Prerequisites for using this Operator

          1. Get AccuKnox licence - https://app.accuknox.com/sign-up

          2. After Successful Sign-up into AccuKnox Application.

             Please naviagate to Settings >> Manager Cluster >> Onboard Cluster >>

             After successful onboarding, Copy the join token created by AccuKnox

          3. Create a secret using jointoken by executing the below command:

          ```
          oc create secret generic accuknox-jointoken --from-literal=token=<joinToken> -n accuknox-agents
          ```

          ## Deploy AccuKnox Operator using AccuknoxConfig Custom Resource

          All the information to configure the AccuKnox Operator will be provided via Accuknox during Cluster Onboarding steps.

          ### AccuknoxConfig Spec

          | Spec | Description |
          | ---- | ----------- |
          | spireHost | "spire host url" |
          | ppsHost   |  "pps host url" |
          | knoxGateway | "knox gateway url" |
          | joinTokenSecretName | "join token secret name if it's different than accuknox-jointoken" |
          | joinTokenSecretKey | "key name used to store the jointoken with the secret" |

          *Example*

          ```
          apiVersion: operator.accuknox.com/v1
          kind: AccuknoxConfig
          metadata:
            name: accuknoxconfig-default
            namespace: accuknox-agents
          spec:
            ppsHost: <ppsHost url>
            knoxGateway: <knox gateway url>
            spireHost: <spire host url>
          ```
        displayName: Accuknox Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CNAPP
        - CSPM
        - CWPP
        - Cloud Native
        - Runtime Security
        - Zero-Trust
        maturity: stable
        provider:
          name: Accuknox Inc.
          url: https://accuknox.com
        relatedImages:
        - docker.io/accuknox/discovery-engine-discover@sha256:e2d37bfbd0e9d23c129c143eb7c9aab1c3ed55e9571b44888eea2b842f8a2361
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        - public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:a935ad806e82760bb838bada30d3ed1a54c8c61b95fe5c6109f48985c6faf4d7
        - docker.io/accuknox/discovery-engine-hardening@sha256:e596aaaabaec94e0abb4f92ad9266dc010c0e88a89d600e8750a360099cae3cc
        - docker.io/accuknox/feeder-service@sha256:295e897b6f27ce61c45fb0179b734ef6d4c9e91864bcc5d2884a2d1dd4db81b6
        - docker.io/accuknox/policy-enforcement-agent@sha256:a0364343afeac4537589bfc247f64fffd7aac16f8afe07ce050bc58b2af248fb
        - docker.io/accuknox/discovery-engine-offloader@sha256:59cc594009de89a574cd569e701b6765d59a6899a03ad07eee1c025331df4331
        - docker.io/accuknox/shared-informer-agent@sha256:9f53d7dba9569487c42c6cf9e88988a1adb4c0971742e8b521700dce2d22b8cc
        - docker.io/accuknox/discovery-engine-sumengine@sha256:f7927128b86fdfb635377bd599df06f0eb9b9f0d1b068a0047468e480855443e
        - docker.io/accuknox/agents-operator@sha256:b0ae2728f15ebae904efb778cd29215ccf07ffca945a59deb5484f6c19c1b21f
        version: 0.7.1
      entries:
      - name: accuknox-operator.v0.7.1
        version: 0.7.1
      name: stable
    defaultChannel: stable
    packageName: accuknox-operator-certified
    provider:
      name: Accuknox Inc.
      url: https://accuknox.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pachyderm, Inc.
      provider-url: https://pachyderm.com
    name: pachyderm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pachyderm-operator.v0.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "Pachyderm",
                "metadata": {
                  "name": "pachyderm-sample"
                },
                "spec": {
                  "console": {
                    "disable": true
                  },
                  "pachd": {
                    "metrics": {
                      "disable": false
                    },
                    "storage": {
                      "amazon": {
                        "credentialSecretName": "pachyderm-aws-secret"
                      },
                      "backend": "AMAZON"
                    }
                  },
                  "version": "v2.1.6"
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermExport",
                "metadata": {
                  "name": "pachydermexport-sample"
                },
                "spec": {
                  "backup": {
                    "target": "pachyderm-sample"
                  }
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermImport",
                "metadata": {
                  "name": "pachydermimport-sample"
                },
                "spec": {
                  "backup": "backups/pachyderm-backup-20220610.tar.gz",
                  "destination": {
                    "name": "pachyderm-restore",
                    "namespace": "pachyderm-workloads"
                  },
                  "storageSecret": "s3-credentials-secret"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
          createdAt: "2021-05-20T08:05:00Z"
          description: Pachyderm provides the data layer that allows data science
            teams to productionize and scale their machine learning lifecycle with
            data driven automation, petabyte scalability and end-to-end reproducibility.
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pachyderm/openshift-operator
          support: Pachyderm, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PachydermExport is the Schema for the pachydermexports API
            displayName: Pachyderm Export
            kind: PachydermExport
            name: pachydermexports.aiml.pachyderm.com
            version: v1beta1
          - description: PachydermImport is the Schema for the pachydermimports API
            displayName: Pachyderm Import
            kind: PachydermImport
            name: pachydermimports.aiml.pachyderm.com
            version: v1beta1
          - description: Pachyderm is the Schema for the pachyderms API
            displayName: Pachyderm
            kind: Pachyderm
            name: pachyderms.aiml.pachyderm.com
            version: v1beta1
        description: "The Pachyderm operator aims to make deploying and managing Pachyderm
          instances on Openshift container platform easier.\n\nPachyderm is the data
          foundation for machine learning.  Pachyderm provides industry leading data
          versioning, pipelines and lineage that allow data science teams to automate
          the machine learning lifecycle and optimize their machine learning operations
          (MLOps). \n\n## Getting started\n\n1. Pachyderm uses an object store to
          store your data. Set the pachd storage backend at `pachyderm.spec.pachd.storage.backend`
          to `AMAZON`, `MICROSOFT`, `GOOGLE` or `MINIO` depending on your storage
          provider. Additionally, create a user with read/write access to the object
          store.\n\n2. Setup the user credentials in the section, `pachyderm.spec.pachd.storage`.
          Required fields depend on the backend type set in step 1 above.\n\n3. Finally,
          deploy the pachyderm cluster in its own namespace.\n\nFor more information
          on getting started, reference the documentation.\n"
        displayName: Pachyderm
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pachyderm
        - versioning
        - machine learning
        - artificial intelligence
        links:
        - name: Documentation
          url: https://docs.pachyderm.com/latest/
        maintainers:
        - email: kyle@pachyderm.io
          name: Kyle White
        - email: eochieng@redhat.com
          name: Edmund Ochieng
        maturity: beta
        minKubeVersion: v1.19.0
        provider:
          name: Pachyderm, Inc.
          url: https://pachyderm.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - registry.connect.redhat.com/pachyderm/backup-handler@sha256:e2ef6bd8191e8ad58c765c0db8f5bb21ea7123caed0c0f58fe681a024201306d
        - registry.connect.redhat.com/pachyderm/etcd@sha256:7f330b51874379b1d4b71d94e37e0ecfa52b86d628296136970688cf204f53b5
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/pgbouncer@sha256:c3b6949227df24abb8934d111a34b44e87fb16194ac34a5d0fe284cff02e9ac0
        - registry.connect.redhat.com/pachyderm/init-utils@sha256:5c0258bfa0203079925648c0a7dda6d05a29b6f461854821a1eec541777ff359
        - registry.connect.redhat.com/pachyderm/worker@sha256:374f83d92899707bc87b0318fda6bffa78d2e9e523b827bb3d40693a2d1bdecd
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
        - registry.redhat.io/rhel8/postgresql-13@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/pachd@sha256:116477fefdf413ea02c6ac3d013f1f91d3f6e4b16aed2408c9890832cdb41430
        version: 0.1.4
      entries:
      - name: pachyderm-operator.v0.1.4
        version: 0.1.4
      name: stable
    defaultChannel: stable
    packageName: pachyderm-operator
    provider:
      name: Pachyderm, Inc.
      url: https://pachyderm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-listener-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: listener-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/tools@sha256:47204b2697dfbab5eaf9d6dcb4c160dd427a148d6b840727239b773810e74fc9
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/tools@sha256:47204b2697dfbab5eaf9d6dcb4c160dd427a148d6b840727239b773810e74fc9
        - quay.io/stackable/listener-operator@sha256:b9c7957e7e562889bddd08e679abff05bb677672b3f735932f68514b77eacfd5
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:a6d6631d35c32b874c995b6af889636ac8dd1f39dcb68810d9f9b10a4e2c1987
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:161e754437428c68490333337303fb30b695a9ac677c3a00a1dc27565b855ab8
        version: 24.11.1
      entries:
      - name: listener-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: listener-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/listener-operator@sha256:da53970954b889bf4d874ae3df7e411b9abf41950ef7d50d6289831826f045a0
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/listener-operator@sha256:da53970954b889bf4d874ae3df7e411b9abf41950ef7d50d6289831826f045a0
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:71f8465e35a67f4b96e4c4904e67ccac9bbb9ae347426afb8bbfaabd6709e983
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        version: 25.3.0
      entries:
      - name: listener-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: listener-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
        version: 25.7.0
      entries:
      - name: listener-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: listener-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        version: 25.7.0
      entries:
      - name: listener-operator.v25.7.0
        version: 25.7.0
      - name: listener-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: "25.7"
    packageName: stackable-listener-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      pod-security.kubernetes.io/audit: privileged
      pod-security.kubernetes.io/enforce: privileged
      pod-security.kubernetes.io/warn: privileged
      provider: NVIDIA Corporation
      provider-url: ""
    name: gpu-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gpu-operator-certified.v25.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "kernelModuleType": "auto",
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": []
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
          createdAt: Thu Sep 18 18:11:36 PDT 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <25.3.4'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:43aa99352e1dc47e9218f36340ae6328a6282ed1e7e9a61b4204475e980230ef
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:c22dc67465c33d2f17932931229e61a6e7d9e32c5ed2ecefe8d10e873ffedfeb
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:51c8f71d3b3c08ae4eb4853697e3f8e6f11e435e666e08210178e6a1faf8028f
        - nvcr.io/nvidia/driver@sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735
        - nvcr.io/nvidia/driver@sha256:44422f10e899d3419fbd611a02fc1c6127e071a2067bd36e3debb043bf251de0
        - nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:66c3d68ca2b35d47b5562300a10ff8711b953ba3a15df7dbc409d81613ebe5d0
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b689dfb7806ae51b0072620b47c26114f05477e8ad9725892ee10366e5d93d8d
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:119de9a331a47203858b99901f44d0c4a8052961b4e60327f4b100d0ab8c9df0
        - nvcr.io/nvidia/k8s-device-plugin@sha256:3c54348fe5a57e5700e7d8068e7531d2ef2d5f3ccb70c8f6bac0953432527abd
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:5c4e61f7ba83d7a64ff2523d447c209ce5bde1ddc79acaf1f32f19620b4912d6
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:6bb786c41fa7b1a6238c023566fdeaeeea3ba842a8daec015a7bd85a8037b249
        - nvcr.io/nvidia/cuda@sha256:d19fe621624c4eb6ac931b8558daa3ecc0c3f07f1e2a52e0267e083d22dceade
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3191cc0684e92d655e72fe9faf3da95fe6b8a6c2eb4ad396f75ba9b98c021aaf
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:619e7bf48168d76a7e087c3bf6190fda9b10449c9839309fa4f9ed5b8a9e8804
        version: 25.3.4
      entries:
      - name: gpu-operator-certified.v25.3.4
        version: 25.3.4
      - name: gpu-operator-certified.v25.3.3
        version: 25.3.3
      - name: gpu-operator-certified.v25.3.2
        version: 25.3.2
      - name: gpu-operator-certified.v25.3.1
        version: 25.3.1
      - name: gpu-operator-certified.v25.3.0
        version: 25.3.0
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: stable
    - currentCSV: gpu-operator-certified.v1.10.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "dcgmExporter": {
                    "config": {
                      "name": ""
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                  },
                  "devicePlugin": {
                  },
                  "driver": {
                    "enabled": true,
                    "use_ocp_driver_toolkit": true,
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "gfd": {
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "operator": {
                    "defaultRuntime": "crio",
                    "deployGFD": true,
                    "initContainer": {
                    }
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:c7f9074c1a7f58947c807f23f2eece3a8b04e11175127919156f8e864821d45a
          createdAt: Mon Apr 18 10:06:40 PDT 2022
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=v1.9.0 <v1.10.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cuda@sha256:7f0f7dc8c993a86865d4d9f9b0e36d113ea15603d5d5279d2879fa2bc21c8236
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:5b16056257acc51b517d9cdb1da3218693cefc214af93789e6e214fd2b4cacf1
        - nvcr.io/nvidia/k8s-device-plugin@sha256:aa95c16106280e36a5e32d2fe4c66e8b70f5a114860c6f4ed5b1a4085c63601b
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:e8f06ce53415e8572f633b2169289d973993cb1e49d4dcc686ac134194d88f33
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:24d804e8f005d7aeca8343aa13e5f92295d8642a4c47cb24e3ac86a22543bc37
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:67c48a8370c34f68b46eb57339e87d01c04198368a0423553fab7565c862c4df
        - nvcr.io/nvidia/driver@sha256:f99b66855e92451a59a722355dd52bf97eb985cc8a647172273de8178aeaa93c
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:c2e3d0fe41a0d227dbb70caec03c780cc76317515e5ab3875f31d50c63f41c66
        - nvcr.io/nvidia/gpu-operator@sha256:c7f9074c1a7f58947c807f23f2eece3a8b04e11175127919156f8e864821d45a
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:ac13256c295126b854a07421c45de0721e7a57974b38b3e66eb1fdc3a4e29f2e
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:9abb063e6b74085728fead5ed5bbe3bd7b3f57de60e87f6b964bba2d46f1bc67
        version: 1.10.1
      entries:
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v1.10
    - currentCSV: gpu-operator-certified.v1.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    },
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "dcgmExporter": {
                    "config": {
                      "name": ""
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                  },
                  "devicePlugin": {
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": true
                  },
                  "vgpuDeviceManager": {
                    "enabled": true,
                    "config": {
                      "name": "vgpu-devices-config",
                      "default": "default"
                      }
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:53180f81fb7e1e4bcd49a556bc6285c2262abbab8083b29d2688b02d4200dd2d
          createdAt: Thu Aug  4 11:57:56 PDT 2022
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=v1.9.0 <v1.11.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:6e2902d73c54cf57bec94edf166d206cfeb0b948b5aab70800ccbe025064b3a1
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:b90a519e9f30fbc013e1806bc081268724e9e80b4035c67c167ab3b43a37704a
        - nvcr.io/nvidia/gpu-operator@sha256:53180f81fb7e1e4bcd49a556bc6285c2262abbab8083b29d2688b02d4200dd2d
        - nvcr.io/nvidia/driver@sha256:cf8ce04907d0e8decf90a8dccfa6134af9c39509d2f8a21c355ed71223acacdb
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:39127d04c9faff3f22068dcf463ed48d02be684cd18e3f14db843bddbdd783a7
        - nvcr.io/nvidia/driver@sha256:b8799e7397fae5a27446737443091eed05b918816f53c400a48be71125ee22e8
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3804e2b03c44d25dd410e3882b7fb6d1fde41f55110d7702af47136b042a63ff
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:10d825ccbfcf87f8c74f841fa89ae345f2d381fb47a5b2b31f61b72ebbbb0915
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:f48e8b430248a9cadf716a9211bb0050d4a663d1f1386ea107e72917f82e5b5f
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:8490d555131a57a11311561667ea9ec8cf554a399d2492a6f9e6677c01ec4b9a
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7a0ccbcbe8d379bfc441e06504eb45d0d433a37e2a9e6dde83d13347296ebe81
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:e88e610d09f1a7528ab32cb54259aee340eba2b2bf66b92a28157626daadc82d
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:2d6e86db0e2b0db64c00566938e1851e25d4bb1019b1e7b696e2b199cba33397
        - nvcr.io/nvidia/driver@sha256:8c9fabc1cd773e9526eb7e8526084f065f6cca9d0fd8dc2a672a0bb717f1cc60
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:16ee5b1a8d9217f761c40a06c7ae598cbc893fbc35d99ff9b84fa886ea6b9b37
        - nvcr.io/nvidia/cuda@sha256:4400b56f81095b24fc9dde93ad7ad7d0ca6a7e9736c4c7f1d4e2bcbf0afbe45c
        version: 1.11.1
      entries:
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v1.11
    - currentCSV: gpu-operator-certified.v22.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:54d5846c0291db621931a8544c9cb6437512470a93278ca9130ba73f2e2f0055
          createdAt: Mon Jan 30 15:43:24 PST 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <22.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:c3cf59dd5d6160eba5d816ade2e81b35ebb10f4884df67971f6ace36f8e6efc1
        - nvcr.io/nvidia/driver@sha256:f250d1e4989fd4821e81bad9227eeeaa531ca8810f3a40e996f1ca2af7945f96
        - nvcr.io/nvidia/k8s-device-plugin@sha256:9c17d3a907eb77eb8f7b4f3faf52d8352e4252af92003f828083f80d629bd2c3
        - nvcr.io/nvidia/driver@sha256:b860ef04cc6cd8c481810fa19bef7221071f5a3862640b2ad6c2704728dc77c8
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:0d47dad29d2ef445b301c5c64717758eed43a606345b79f97bce2e64b40a91a8
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:64d757b4c80b910e64647a84a1d592fab2ea3313ff6dce30c25c3a08e180bd74
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:efb88937f73434994d1bbadc87b492a1df047aa9f8d6e9f5ec3b09536e6e7691
        - nvcr.io/nvidia/driver@sha256:f2ca4612f3418e0e3963be8c3771ffe463cdc1fc1b810437a3aad9d76bb6499d
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:bec9f026d9b3d9404c78d6091817a359015c6a7aa411735b34138c1518853b5d
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:00f1476548fbed9ee01961443a73bf65396c2e8bb2b84426f949dd56cb4d14cd
        - nvcr.io/nvidia/gpu-operator@sha256:54d5846c0291db621931a8544c9cb6437512470a93278ca9130ba73f2e2f0055
        - nvcr.io/nvidia/driver@sha256:c5a62bc2a6cce339798e4b7cf9cec47e289c2d30b40f451ba8c9e958a68e3e55
        - nvcr.io/nvidia/driver@sha256:268c26781c46b36fe691638e807eb5c3a2cd077029ddb095ac75264adac04925
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:9177a0ae30798b42d0387f6a20cd3ce3cd1799a91b7866bf812368764b05b1af
        - nvcr.io/nvidia/cuda@sha256:5f2a2d8977f2c59abe88394f05cc3c044194554f90994d8554b0e1117ab5900d
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:9a00cdfdddb73327ef8e8e0fa60e50926a388d380c551bba6a3d3012be40401d
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:50c2a2b998e467c82716efa79fe136aa6f7ea95fd23576cf384d251bb9628640
        version: 22.9.2
      entries:
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v22.9
    - currentCSV: gpu-operator-certified.v23.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:08ca33f491195b7ba867be1e6642b5ff8aff6a4b0b7cfc0ec87619252ccfddc7
          createdAt: Mon May 22 15:22:15 PDT 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <23.3.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:8ff0ee88fe6c4ff031ef642593f93625bfd9a48fe7be8a79a5c2736117e36941
        - nvcr.io/nvidia/k8s-device-plugin@sha256:80871275fbcf888d4683c58638c7a75b3c46091040cafc6210a8c93513adb146
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:03e43d6b1460cadb33509d5ee07a5e27603ab89d7ec35f67d8a51bb78745b463
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:cab21c93987a5c884075efe0fb4a8abaa1997e1696cbc773ba69889f42f8329b
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:84ce86490d0d313ed6517f2ac3a271e1179d7478d86c772da3846727d7feddc3
        - nvcr.io/nvidia/cuda@sha256:62d845b4bd6de4be65f0b6daaae6831c29d4529955f99db5d26f5f7179106285
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:5489125ceae5864280e4d6a9ab52ab0f650b3179349a7298c4a204feb60b661a
        - nvcr.io/nvidia/gpu-operator@sha256:08ca33f491195b7ba867be1e6642b5ff8aff6a4b0b7cfc0ec87619252ccfddc7
        - nvcr.io/nvidia/driver@sha256:73b8928a1b1fb650d2ca8d6451a7d7c551fd0509f7511c2fd5bfdcef83ae4f5d
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:2d7e32e3d30c2415b4eb0b48ff4ce5a4ccabaf69ede0486305ed51d26cab7713
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:21dfc9c56b5f8bce73e60361d6e83759c3fa14dc6afc2d5ebdf1b891a936daf6
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:0d47dad29d2ef445b301c5c64717758eed43a606345b79f97bce2e64b40a91a8
        - nvcr.io/nvidia/driver@sha256:917f92c589c3124dba54f66f89c44b93b0beaf418f01d0da6a546e1670fdf3a9
        - nvcr.io/nvidia/driver@sha256:3df118f2330e95db3c76173cca35e542b2da120ec0bb8f6aff83c4c96c0c2daf
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:8f03f74076b5aa6672a5d75a398fa7515fcde57fdb1cf8079e015b3366efd3b5
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:f10170b076badda8f669af8b0f7b47eb5470abaf35a67dda9c4e182d66138177
        version: 23.3.2
      entries:
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.3
    - currentCSV: gpu-operator-certified.v23.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:d2e0f8bd86e8e9f2ff3514e113657658584ca4a2d0b82dac2e47364cc0b30648
          createdAt: Thu Aug 31 12:22:44 PDT 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <23.6.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:878a6162a4fa0b61dae02c7cfc13ba4d28bb06f92fd62d0c521633ab507d675a
        - nvcr.io/nvidia/driver@sha256:bbdc4415fe2e71eeea1b8a6e50ee562e11704dabb46db1db4aea807102886a35
        - nvcr.io/nvidia/k8s-device-plugin@sha256:f1617c3fe54fd9f6d4a00358e5da1a9fa9e8963a4ef6b3ed06f47d4971ddff4b
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:b30f5c64791dc5406b57eaf51ac40195dc1051a18cd1f38fd0a43c65f1d6dba9
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:a229beba8540549c234a046e975bf4bf8c50d521dfe551af56f6e042b25ec017
        - nvcr.io/nvidia/gpu-operator@sha256:d2e0f8bd86e8e9f2ff3514e113657658584ca4a2d0b82dac2e47364cc0b30648
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:9484110986c80ab83bc404066ca4b7be115124ec04ca16bce775403e92bfd890
        - nvcr.io/nvidia/driver@sha256:b4517810f2884f345b4f2323e83e41d28f54cdc6e2caf0690bb7ed4f3947dbcc
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:bb845160b32fd12eb3fae3e830d2e6a7780bc7405e0d8c5b816242d48be9daa8
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:60086c33226bbe7f04b611c20d308fed5825226f8a0a5635bc58d8ed55cc08a5
        - nvcr.io/nvidia/cuda@sha256:64c0ff5fd4a27d4b1d46378634f7a7398c75f87c18c70c988dde18910d461560
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:00a91252be8e62da6f28cdf1ec3721f2e945ae296938b0b925ef9db2bf6dcc8b
        - nvcr.io/nvidia/driver@sha256:3382e254056f28831767bc6729bc2594353a5ff2a28fe9f2d94396c597bb23d8
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:0379c59b43e11861837787a0da22e70b528718373bb10de9e8b7752849bd4e7a
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:bf06c8ad380a36e5014607bd3ab05b80e8e5e7e77d23d0dd89fe53a5564152fa
        version: 23.6.1
      entries:
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.6
    - currentCSV: gpu-operator-certified.v23.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:7481a3697783dcdca9ae78e7b548a6900e86ea33ab49ec14f0ba55db2fdb1a2e",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:3988b3b570cf65b0d338cb29bd24b0b9bf1885e1563fbb398ef23cba0b59e6bb
          createdAt: Wed Mar 6 16:43:36 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <23.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:6a05d6a1923fda756aed0dddf7ed23a83c30cf1e6c519fc39dd70c0309ec8257
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:a67d8e92861a2dce5649105c07561e4422e9fe4ba81a6525dc0d70a7ef85f9c0
        - nvcr.io/nvidia/driver@sha256:56c79482582cdfbc58d3134e8672637c5bf05f328880f76898f526143d04c6af
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:8d1ec78f2b1ddb7f0c47453d0427231190747bda411733a7dd0c8f5196f09e9c
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:27c44f4720a4abf780217bd5e7903e4a008ebdbcf71238c4f106a0c22654776c
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:925f4a47710e4318ed457930f5406174c1f6d28b1bf6b1bc310687fec0fde712
        - nvcr.io/nvidia/driver@sha256:abda1ac56371d55917b96ff330109980f468e133c9d5705da0ef87429f14ccd7
        - nvcr.io/nvidia/k8s-device-plugin@sha256:bb81ae8d034d1e1c6a342091a5f1dda5884dd43ab47e96624487a133d1edd4b3
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:387021553dfb16aab633228d42f63f04fa932b4f46add07527f296dfe97e5148
        - nvcr.io/nvidia/driver@sha256:7481a3697783dcdca9ae78e7b548a6900e86ea33ab49ec14f0ba55db2fdb1a2e
        - nvcr.io/nvidia/gpu-operator@sha256:3988b3b570cf65b0d338cb29bd24b0b9bf1885e1563fbb398ef23cba0b59e6bb
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:011fb450af3fa2e8fe5d28d590e4c653631447bc23d149591ced3d89089c4f2c
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:9aefef081c3ab1123556374d2b15d0429f3990af2fbaccc3c9827801e1042703
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:59a3875e7a37eb370385e654184efa3a1b193c9ea352165818496b19cbe14aa4
        - nvcr.io/nvidia/cuda@sha256:714547d54e5fe4191019a1e5f1daffc7fab7481b619b79c378541dafc76c9e5d
        version: 23.9.2
      entries:
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.9
    - currentCSV: gpu-operator-certified.v24.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:96f25c67e5b1072d5981080e12d65ec37eb9ef2fc0494499416aa801b0a34da3",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:041e75a3df84039c2dbbd4b9d67763bd212138822dbb6dbc0008858c1c6eff8d
          createdAt: Tue Apr 30 14:18:26 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.3.0'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:ecb867c5787e44f1889b257e6c03b07748689a7b293c8f1affb1d4cb449b235c
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:f0e4e14a3081417be8e6a5d855fb0cb69e1b63bc54a74f17e6d7084abe275588
        - nvcr.io/nvidia/driver@sha256:56c79482582cdfbc58d3134e8672637c5bf05f328880f76898f526143d04c6af
        - nvcr.io/nvidia/driver@sha256:a836ccbe21da735aee1c39b81060ed5e2fdb4ffa339874baaf4634f1e9259f74
        - nvcr.io/nvidia/driver@sha256:96f25c67e5b1072d5981080e12d65ec37eb9ef2fc0494499416aa801b0a34da3
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:2edc1d4ed555830e70010c82558936198f5faa86fc29ecf5698219145102cfcc
        - nvcr.io/nvidia/k8s-device-plugin@sha256:80674b19898ecf9ed6949e39674da769d6feeeb01bea54b914ef9ff502834f49
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:d8774d6afd37110199512636172823ae8749e5ea2e13760f57c255a74f47018c
        - nvcr.io/nvidia/cuda@sha256:ae0623ec8634b6c88f815b88037763def160cbbac15013b77ddef257fc276c9a
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:1ae0b923bc57f47bab046b50c50110f6914bbaffbfef704df34b3fe332db2e31
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:0c77725de8c42d248ed825453efd2e005f4900f4be384fd23084f6b721ddd0e0
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:7798eb9b9424e09959fc808596f87bdb68ac5740174123c6646dee2166d3e7d2
        - nvcr.io/nvidia/gpu-operator@sha256:041e75a3df84039c2dbbd4b9d67763bd212138822dbb6dbc0008858c1c6eff8d
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:2cc2119d20515f8ca7a61a0f3932578d69a45bcdea49e6c320a89c56f105e7d9
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:2442fd2ec0ee29746abbced727c53741f017e8f3f615321c1ec42c1c3ffa0b12
        version: 24.3.0
      entries:
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.3
    - currentCSV: gpu-operator-certified.v24.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:858de27c152669f5a3cf4287406405b16dd5bb70c0373324eb735511997bb415",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:8e0969cffc030a89c4acd68e64d41dd54e3bce8a794106b178d4dbd636a07f1c
          createdAt: Wed Sep 25 08:25:16 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.6.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:21f4c8b88716e8e6f732f9fb4c2efaef937c227491a8631c5e55036f80f39a4d
        - nvcr.io/nvidia/driver@sha256:858de27c152669f5a3cf4287406405b16dd5bb70c0373324eb735511997bb415
        - nvcr.io/nvidia/gpu-operator@sha256:8e0969cffc030a89c4acd68e64d41dd54e3bce8a794106b178d4dbd636a07f1c
        - nvcr.io/nvidia/cuda@sha256:de5b598bca89850567c4c104411d66bb52f47c9179199e6a3be6829b7ac586a2
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:740abc3ff657545c10effd5354f09af525200ed9a1b7623f0c2e8c7bd9e4a4e2
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:96380b95396b7f29d2ed2ec8ef1ad5a8bcf9f55051db723295dc0a20db845331
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:ae63fac9a4057a7646f0cf0ee0566e8928529adde05c4c0a017cda0599e381b2
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:33de74eb590f071403c17b6c210c02963245851971168bc0c07c06c100a9f376
        - nvcr.io/nvidia/driver@sha256:07e11f85d54d49ec9648fb06e148b8d832ee1f9c3549a915eee853c9ef2949c2
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:781fb47e264d9e0fbc8da5bd046e5e678316c866bc36ddd4b56d4eb0de682d5b
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:969147c01d63be5d1fe458f32f1cc0c7408cf3062531db91408e2fc57b4d8a67
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:15dab1273345df4a5844c4c761d064dbc4b592101251dc39174e597137123027
        - nvcr.io/nvidia/driver@sha256:a6d12fb5753f267dda25dfd38910f972bc632c006a24107fa50e20bba3642d7c
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7ad2c9f71fe06f9f7745ac8635f46740fbdff4f11edd468addfab81afcdfa534
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:7bcc188703f2fac630f7a4ff8960e6733ac3a29adf6a946533b796d9a27b8acf
        version: 24.6.2
      entries:
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.6
    - currentCSV: gpu-operator-certified.v24.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:6fe74322562c726c8fade184d8c45ebae3da7b1ea0a21f0ff9dc42c66c65e692",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:92f5262bb4934e0ea29eb10ec0ad4fc293164f2f2f94d48a935922791abbfdac
          createdAt: Mon Jan 27 16:10:52 PST 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cuda@sha256:b3b340847db9abd0c39703278f5638d3e65e1369e9a526a19b3138d5a6563527
        - nvcr.io/nvidia/gpu-operator@sha256:92f5262bb4934e0ea29eb10ec0ad4fc293164f2f2f94d48a935922791abbfdac
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:34eab99d7992a57c35803ec5e0afd538d16e997341e951fef4c4019afb08793e
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:e8398b2451996e5c64fc3855e7a3c93b667a85f69c0e259049d3b0012c21545d
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:d934033f9ff89087d555850eb1da150eaaa27b11e5ad21db1b874b8bc0b8d35b
        - nvcr.io/nvidia/driver@sha256:6fe74322562c726c8fade184d8c45ebae3da7b1ea0a21f0ff9dc42c66c65e692
        - nvcr.io/nvidia/driver@sha256:38b66a8d44cab9e2c62da9e101f32cd9dbcb5e02d8e57b47671284d374ca3695
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:4ffa1cd2a6497eb647a89ed259dcfb007554737b9d80f69bc173a2c3cd72a1da
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:40830d3561c14743f484d45b498141f9e86b1308e16fae3978110783927264ab
        - nvcr.io/nvidia/driver@sha256:026f8f2d29b7058ecaaa5b98666e28e1b3646b73e7d7cda3dc8026d47a929152
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:cf39d78ced7fb5727a9668ee2cd44b14bb7a23a95b83d5464b7d755740e02121
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7089559ce6153018806857f5049085bae15b3bf6f1c8bd19d8b12f707d087dea
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b072c5793be65eee556eaff1b9cbbd115a1ef29982be95b2959adfcb4bc72382
        - nvcr.io/nvidia/driver@sha256:d55b57938866e538acc3a71ca32f8cf87e71c591abd4a34695ee428e7ec2fa73
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:7edd7a0413dcb39b6e3bcefaf06812f3293c8e480ca10783e821a561ed686200
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:857f7669ae5ce2a43a3b5691fa970085098321ef8e90a896fa8dda3dee5b1d2b
        version: 24.9.2
      entries:
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.9
    - currentCSV: gpu-operator-certified.v25.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "kernelModuleType": "auto",
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": []
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
          createdAt: Thu Sep 18 18:11:36 PDT 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <25.3.4'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cuda@sha256:d19fe621624c4eb6ac931b8558daa3ecc0c3f07f1e2a52e0267e083d22dceade
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:619e7bf48168d76a7e087c3bf6190fda9b10449c9839309fa4f9ed5b8a9e8804
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:66c3d68ca2b35d47b5562300a10ff8711b953ba3a15df7dbc409d81613ebe5d0
        - nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:5c4e61f7ba83d7a64ff2523d447c209ce5bde1ddc79acaf1f32f19620b4912d6
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:51c8f71d3b3c08ae4eb4853697e3f8e6f11e435e666e08210178e6a1faf8028f
        - nvcr.io/nvidia/driver@sha256:44422f10e899d3419fbd611a02fc1c6127e071a2067bd36e3debb043bf251de0
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3191cc0684e92d655e72fe9faf3da95fe6b8a6c2eb4ad396f75ba9b98c021aaf
        - nvcr.io/nvidia/driver@sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735
        - nvcr.io/nvidia/k8s-device-plugin@sha256:3c54348fe5a57e5700e7d8068e7531d2ef2d5f3ccb70c8f6bac0953432527abd
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:6bb786c41fa7b1a6238c023566fdeaeeea3ba842a8daec015a7bd85a8037b249
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:c22dc67465c33d2f17932931229e61a6e7d9e32c5ed2ecefe8d10e873ffedfeb
        - nvcr.io/nvidia/driver@sha256:43aa99352e1dc47e9218f36340ae6328a6282ed1e7e9a61b4204475e980230ef
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:119de9a331a47203858b99901f44d0c4a8052961b4e60327f4b100d0ab8c9df0
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b689dfb7806ae51b0072620b47c26114f05477e8ad9725892ee10366e5d93d8d
        version: 25.3.4
      entries:
      - name: gpu-operator-certified.v25.3.4
        version: 25.3.4
      - name: gpu-operator-certified.v25.3.3
        version: 25.3.3
      - name: gpu-operator-certified.v25.3.2
        version: 25.3.2
      - name: gpu-operator-certified.v25.3.1
        version: 25.3.1
      - name: gpu-operator-certified.v25.3.0
        version: 25.3.0
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v25.3
    defaultChannel: v25.3
    packageName: gpu-operator-certified
    provider:
      name: NVIDIA Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Yugabyte Inc
      provider-url: ""
    name: yugabyte-platform-operator-bundle-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: yugabyte-platform-operator.v0.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "yugabyte.com/v1alpha1",
                "kind": "YBPlatform",
                "metadata": {
                  "name": "ybplatform-sample"
                },
                "spec": {
                  "domainName": "cluster.local",
                  "helm": {
                    "package": "/opt/yugabyte/helm/yugabyte-latest.tgz",
                    "timeout": 900
                  },
                  "helm2Legacy": false,
                  "image": {
                    "commonRegistry": "",
                    "nginx": {
                      "name": "nginxinc/nginx-unprivileged",
                      "registry": "",
                      "tag": "1.17.4-amd64"
                    },
                    "postgres": {
                      "name": "postgres",
                      "registry": "",
                      "tag": 11.5
                    },
                    "prometheus": {
                      "name": "prom/prometheus",
                      "registry": "",
                      "tag": "v2.2.1"
                    },
                    "pullPolicy": "IfNotPresent",
                    "pullSecret": "yugabyte-k8s-pull-secret",
                    "repository": "registry.connect.redhat.com/yugabytedb/yugabyte-platform",
                    "tag": "latest",
                    "thirdparty-deps": {
                      "name": "yugabyte/thirdparty-deps",
                      "registry": "quay.io",
                      "tag": "latest"
                    }
                  },
                  "ip_version_support": "v4_only",
                  "ocpCompatibility": {
                    "enabled": true
                  },
                  "rbac": {
                    "create": true
                  },
                  "securityContext": {
                    "enabled": false,
                    "fsGroup": 1001,
                    "runAsUser": 1001
                  },
                  "tls": {
                    "certificate": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tDQpNSUlHTHpDQ0JSZWdBd0lCQWdJUUQ2R1ZmL0ViUDhSMFQvcWxjRGZyM3pBTkJna3Foa2lHOXcwQkFRc0ZBREJlDQpNUXN3Q1FZRFZRUUdFd0pWVXpFVk1CTUdBMVVFQ2hNTVJHbG5hVU5sY25RZ1NXNWpNUmt3RndZRFZRUUxFeEIzDQpkM2N1WkdsbmFXTmxjblF1WTI5dE1SMHdHd1lEVlFRREV4UlNZWEJwWkZOVFRDQlNVMEVnUTBFZ01qQXhPREFlDQpGdzB4T1RBNE1ERXdNREF3TURCYUZ3MHlNVEEzTXpFeE1qQXdNREJhTUI4eEhUQWJCZ05WQkFNTUZDb3VZMnh2DQpkV1F1ZVhWbllXSjVkR1V1WTI5dE1JSUJJakFOQmdrcWhraUc5dzBCQVFFRkFBT0NBUThBTUlJQkNnS0NBUUVBDQp3SmswU25ub1pRYW1hWm0rMVVJdjdRYjBLalhZeVEwckpJSGpsbE03elhFYUFoL0JqVjZ1M0h1TWxrTlZSekVFDQpGeEtGUm5nSlpXTVlPcy9sWDU1QmtBS0RocXNmaDl0ZVA4ZFVQU1RTVTN6bWRVc0dTcDJIMGpKbVhuRWNvMFpzDQpwYlVyYWhlcWlxSUUyZ1gwQlpvYldvWGRHc1EydXJiUGxyMGlVSnUxVnp4SGRoTS9nTnFQWEN5bE5YdENrbVI2DQo0eXpIa0Rob21pRTAySTMwTXBlOERxczl4d3lKd3F6T3J2ZnA0eUE0Ukhna2g1clovK1Erc0VQNkIyVzBHQXc2DQpzZDNTYytOTytYVUZoVnlFRDhqZkNsTndPVHZodi9keE5IWHZ4WHd0RWNjZTd1TVBBTFJzSFh1RW1kTkN5N2pvDQpibzJuUWFObzZBQ1c0U2s5RWJVZXB3SURBUUFCbzRJREpqQ0NBeUl3SHdZRFZSMGpCQmd3Rm9BVVU4b1hXZnhyDQp3QU1oTHhxdTVLcW9ISUpXMm5Vd0hRWURWUjBPQkJZRUZHemw3bjdOMy9GbFdDSDFvdWFCWldKMjI1UStNQjhHDQpBMVVkRVFRWU1CYUNGQ291WTJ4dmRXUXVlWFZuWVdKNWRHVXVZMjl0TUE0R0ExVWREd0VCL3dRRUF3SUZvREFkDQpCZ05WSFNVRUZqQVVCZ2dyQmdFRkJRY0RBUVlJS3dZQkJRVUhBd0l3UGdZRFZSMGZCRGN3TlRBem9ER2dMNFl0DQphSFIwY0RvdkwyTmtjQzV5WVhCcFpITnpiQzVqYjIwdlVtRndhV1JUVTB4U1UwRkRRVEl3TVRndVkzSnNNRXdHDQpBMVVkSUFSRk1FTXdOd1lKWUlaSUFZYjliQUVDTUNvd0tBWUlLd1lCQlFVSEFnRVdIR2gwZEhCek9pOHZkM2QzDQpMbVJwWjJsalpYSjBMbU52YlM5RFVGTXdDQVlHWjRFTUFRSUJNSFVHQ0NzR0FRVUZCd0VCQkdrd1p6QW1CZ2dyDQpCZ0VGQlFjd0FZWWFhSFIwY0RvdkwzTjBZWFIxY3k1eVlYQnBaSE56YkM1amIyMHdQUVlJS3dZQkJRVUhNQUtHDQpNV2gwZEhBNkx5OWpZV05sY25SekxuSmhjR2xrYzNOc0xtTnZiUzlTWVhCcFpGTlRURkpUUVVOQk1qQXhPQzVqDQpjblF3Q1FZRFZSMFRCQUl3QURDQ0FYNEdDaXNHQVFRQjFua0NCQUlFZ2dGdUJJSUJhZ0ZvQUhVQTdrdTl0M1hPDQpZTHJoUW1rZnErR2VacU1QZmwrd2N0aURBTVI3aVhxby9jc0FBQUZzVHEwL013QUFCQU1BUmpCRUFpQlg3NXJ0DQpwNkdQbUQxbVl1S1hvcXVRVytIcUtRNE8rV1RXbTNCbVBteWF1d0lnQUoya01WdmduaUx1dEh5Y3AyYzBwdDRsDQptL3M5SWN2Yi9aN09jMU82b2xJQWRnQ0hkYi9uV1h6NGpFT1pYNzN6YnY5V2pVZFdOdjlLdFdEQnRPci9YcUNEDQpEd0FBQVd4T3JUOXNBQUFFQXdCSE1FVUNJSHphY3NCcmpEN1FER1ZRSHA5YUZPMnJ4VEx4NU9GOVE2MTFYaUU2DQpZUEk3QWlFQW9xd3pTd1BVaGt4bVhYNzN0SDVEcVNwaUxuOHdrMXFWeWhZOTVmSmxZa0VBZHdCRWxHVXVzTzdPDQpyOFJBQjlpby9pakEydWFDdnRqTE1iVS8wek9XdGJhQnFBQUFBV3hPclQ1M0FBQUVBd0JJTUVZQ0lRQ1ZLaHQ5DQpWTWhlNTUvZEwrSGN6Qzk5a2xheG9DcW5zOStRNGVOTDZ2Smd4QUloQUsyZmp4a1RadGw4WFZ4WHEwc2NSTWNJDQpiSktHcGJYdEV0N3dFVzRkZDNGS01BMEdDU3FHU0liM0RRRUJDd1VBQTRJQkFRQm1GcDNCK1g2Y21Cb0FLWkhrDQpIWWxzanhIZk01bDQyZWJtSFdGdlQyTi9aVkZ3MEZ3YndBNExyNzlvd01QdU5FbVNNa2pmVVhxR011Q2VlVTdODQo0MlRwUFdPaGNzVHF6ZkxVa0hhMU9rK011L3RUZHBNcmhROEVqbkhMY3FKVE9MbmRSVy8zcE5IM2xrK2FJejNoDQpDVGFueElRVEE2cUxRNThRakszWmFMVzdqWVVNcVJFcnBGVDlBVXkxMDAxMjlOam5HVkhtYmU1bW1ZbEwvQlhMDQpmdHBJUUJJZnhCNXpHbHpTcjRHT3p5RDBES05YU040RU0yM2pIM1ljeUg4eFQrODU4ZkxSaXQ4SFlOQ3RQbkloDQoyRzR6UjV3emZFVDlnZUh6aWtuY0ZMWUpMMEpKMVIrZGVNRE91bmJWOUJqcmZJb3NNSk53TUV2ejFZczlBWXhHDQpxSFNYDQotLS0tLUVORCBDRVJUSUZJQ0FURS0tLS0tDQotLS0tLUJFR0lOIENFUlRJRklDQVRFLS0tLS0KTUlJRXNUQ0NBNW1nQXdJQkFnSVFDS1dpUnMxTFhJeUQxd0swdTZ0VFNUQU5CZ2txaGtpRzl3MEJBUXNGQURCaApNUXN3Q1FZRFZRUUdFd0pWVXpFVk1CTUdBMVVFQ2hNTVJHbG5hVU5sY25RZ1NXNWpNUmt3RndZRFZRUUxFeEIzCmQzY3VaR2xuYVdObGNuUXVZMjl0TVNBd0hnWURWUVFERXhkRWFXZHBRMlZ5ZENCSGJHOWlZV3dnVW05dmRDQkQKUVRBZUZ3MHhOekV4TURZeE1qSXpNek5hRncweU56RXhNRFl4TWpJek16TmFNRjR4Q3pBSkJnTlZCQVlUQWxWVApNUlV3RXdZRFZRUUtFd3hFYVdkcFEyVnlkQ0JKYm1NeEdUQVhCZ05WQkFzVEVIZDNkeTVrYVdkcFkyVnlkQzVqCmIyMHhIVEFiQmdOVkJBTVRGRkpoY0dsa1UxTk1JRkpUUVNCRFFTQXlNREU0TUlJQklqQU5CZ2txaGtpRzl3MEIKQVFFRkFBT0NBUThBTUlJQkNnS0NBUUVBNVMyb2loRW85bm5wZXpvemlEdHg0V1dMTENsbC9lMHQxRVllbUU1bgorTWdQNXZpYUhMeStWcEhQK25kWDVEMThJTkl1dUFWOHdGcTI2S0Y1VTBXTklaaVFwNm1MdElXalVlV0RQQTI4Ck9leWhUbGo5VExrMmJleXRidEZVNnlwYnBXVWx0bXZZNVY4bmdzcEM3bkZSTkNqcGZuREVEMmtSeUp6Tzh5b0sKTUZ6NEo0SkU4TjdOQTF1SndVRUZNVXZITHMwc2NMb1Baa0tjZXdJUm0xUlYyQXhtRlF4SmtkZjdZTjlQY2traQpmMlhnbTNiNDhCWm4wemYwcVhzU2VHdTg0dWE5Z3d6anpJN3RiVEJqYXlUcFQrL1hwV3VCVnY2ZnZhckk2YmlrCktCODU5T1NHUXV3NzNYWGdldUZ3RVBIVElSb1V0a3p1My9FUStMdHd6bmtrZFFJREFRQUJvNElCWmpDQ0FXSXcKSFFZRFZSME9CQllFRkZQS0YxbjhhOEFESVM4YXJ1U3FxQnlDVnRwMU1COEdBMVVkSXdRWU1CYUFGQVBlVURWVwowVXk3WnZDajRoc2J3NWV5UGRGVk1BNEdBMVVkRHdFQi93UUVBd0lCaGpBZEJnTlZIU1VFRmpBVUJnZ3JCZ0VGCkJRY0RBUVlJS3dZQkJRVUhBd0l3RWdZRFZSMFRBUUgvQkFnd0JnRUIvd0lCQURBMEJnZ3JCZ0VGQlFjQkFRUW8KTUNZd0pBWUlLd1lCQlFVSE1BR0dHR2gwZEhBNkx5OXZZM053TG1ScFoybGpaWEowTG1OdmJUQkNCZ05WSFI4RQpPekE1TURlZ05hQXpoakZvZEhSd09pOHZZM0pzTXk1a2FXZHBZMlZ5ZEM1amIyMHZSR2xuYVVObGNuUkhiRzlpCllXeFNiMjkwUTBFdVkzSnNNR01HQTFVZElBUmNNRm93TndZSllJWklBWWI5YkFFQ01Db3dLQVlJS3dZQkJRVUgKQWdFV0hHaDBkSEJ6T2k4dmQzZDNMbVJwWjJsalpYSjBMbU52YlM5RFVGTXdDd1lKWUlaSUFZYjliQUVCTUFnRwpCbWVCREFFQ0FUQUlCZ1puZ1F3QkFnSXdEUVlKS29aSWh2Y05BUUVMQlFBRGdnRUJBSDRqeC9MS05XNVprbEZjCllXczhFamJtMG55ektlWkMyS09WWVI3UDhnZXZLeXNsV200WG80QlN6S3IyMzVGc0o0YUZ0NnlBaXYxZVkwdFoKL1pOMThiT0dTR1N0b0VjL0pFNG9jSXpyOFA1TWcxMWtSWUhibWdZbnIxUnhla2k1bVNlYjM5REd4VHBKRDRrRwpoczVsWE5vbzRjb25VaWlKd0thcUg3dmgyYmFyeWQ4cE1JU2FnODNKVXF5VkdjMnRXUHBPMDMyOS9DV3Eya3J5CnF2NjZPU01qd3VsVXowZFhmNE9IUWFzUjdDTmZJcis0S1NjYzZBQmxRNVJERjg2UEdlRTZrZHdTUWtGaUIvY1EKeXNOeXEwakVEUVRrZmEycGptdVd0TUNOYkJuaEZYQlllamZ1YkloYVViRXYyRk9RQjNkQ2F2K0ZQZzVlRXZlWApUVnlNbkdvPQotLS0tLUVORCBDRVJUSUZJQ0FURS0tLS0tCg==",
                    "enabled": false,
                    "hostname": "localhost",
                    "key": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQpNSUlFb3dJQkFBS0NBUUVBd0prMFNubm9aUWFtYVptKzFVSXY3UWIwS2pYWXlRMHJKSUhqbGxNN3pYRWFBaC9CCmpWNnUzSHVNbGtOVlJ6RUVGeEtGUm5nSlpXTVlPcy9sWDU1QmtBS0RocXNmaDl0ZVA4ZFVQU1RTVTN6bWRVc0cKU3AySDBqSm1YbkVjbzBac3BiVXJhaGVxaXFJRTJnWDBCWm9iV29YZEdzUTJ1cmJQbHIwaVVKdTFWenhIZGhNLwpnTnFQWEN5bE5YdENrbVI2NHl6SGtEaG9taUUwMkkzME1wZThEcXM5eHd5Sndxek9ydmZwNHlBNFJIZ2toNXJaCi8rUStzRVA2QjJXMEdBdzZzZDNTYytOTytYVUZoVnlFRDhqZkNsTndPVHZodi9keE5IWHZ4WHd0RWNjZTd1TVAKQUxSc0hYdUVtZE5DeTdqb2JvMm5RYU5vNkFDVzRTazlFYlVlcHdJREFRQUJBb0lCQUJDbDdaZzJlM2tvSHdCbgowdEFhWFl3dW1ySi9YeExaK1kyNVVKUHc4KzFwUm1kQ21YZC81THVZVG1tNnZrZGVIRTFtUTNENDlVNzNSdVBtCmpwdzZVTTZDK1IyTjh6YktKcFo2dzNSWmwzaXlCalhEVzNMZHRSemRVRzMvZHd0Q0RpZ3d4ZVR3ME1RVkJ6aFYKWldlS2o2eStmNE5OekxGTkZocWdsTFBGRWVMTlk3WUgvYUdHdTZwaVpERkFMT0liMlFVUUVvVVRkUEJabDMxSwpKL0Y0cHE1RkV6ekQ0aHMwYXJzSlIxeUd4L0NCRWxqQWRSNWE4Y3FHdGhxYjFMdXJTMVcxT0dNUGkzNFdDblJKCnYwTC9zSU9ORCt5cjdjd3ZQaFFzcmI3Tk9NbGxRSDYwUGdyQ0JyOUVwalFiRm1ZN3NCTS9jYUZTWHVDTmZnTzMKT3hTQjF0RUNnWUVBNHpSMTh2aXl3Ykt5aE5xeEhkNzJ1UkxHS2k4SWZhVmhsNGxFeGdsR0pLNHQ0KzYreXhObAp2OFBkdGg3elJjL1pyQUlBUEtBRnZxa1RnVzkrc2llL3JPcE9PUzA2UnVIWVpJMjdpS3ZiL054N29YTm9TM0Y2CitXRlNJSVJqRFo0eHc2SU1iNUJFS0hOVGtKU1pBMXBkaXhYeFgzeDVTM2hCWTlmekl2NGdVVDhDZ1lFQTJRSDAKbnU5REZ0MDNmWmpvcjRMUkZCQS84NEV1dmdBL2RxMkZ4TzVvMjIrdk43Qk9LRW9IWTRuTzhYdFN2L2MzaVFZdApmTUdGZUR2eXdqRnFPU05SS2JxenMza3ZmY2gxNEhHdFltT29DVTF6Smd2VkF4SytWOFF4YWxMbmhlbjZsaTgwCnJKYmtWWUJXYnFHZHU1d2FPM1RyeGsvbHBMejFvSnliK0FQbmNKa0NnWUVBeHd3anRTeEtOc1hkL2VRMXZReWoKcTdvVjM4ajZzRkRkOGNzeGRDbVZ5UGJMZjErZXlQeUNiSUVrMExQR3ZzZDhWR1FnWHZ4MmEvVm1BRVIwZzVHdApLOGpOcE92Sm5XV29sVmRTRDBjNDRBc3RTR0Y1NG5aVm5QcTAzNFg3M1dkVkE2WUpMUWYvbmt5Q3ppeFE3UTRTCjI1SHF0NU91RS84VGttd1ROdm5aeURNQ2dZQlUxV0xmendFWEt4dktTY0NNL0VMbjZRSlhHeVUrdWJJVm52Z3cKYUZNdGxaZVYyYUowTlp4MTJTWEJXVWljS3ZUVzZFZFlBZjgvWlpBZFFsd1lRaW8wZWZ4T0tRM2J5a1hxbkFYLwo5SjE0SllMaVdMWEpJY2F2WU05QWxYQlU5WWMwdHZlTWtPQnFYR1Z0a2xpclF2Y3VUeUxzdjhjeWczd3ljWWEzCmlpOHI0UUtCZ0NaODZtcEZuUlN2aTdpTWw1MUdybDFQZER2YzZVMURRVkI5empTQlVzak1lamNTQXk4WmVJZlQKUDVoSGc3REFaUEhmUGpYLzZxUnpNQW8zWlFlZVBMdTNEYVgwbFBoekNhalF1bkpIbWZvYlA2QjFOUG45T056bQo0dE0wdEU1YTJVVmJuZW5haVZyaDRhSXlJTTFxVTdHT1ArcDNMNi9hNzVnSVQ2SVJRVlloCi0tLS0tRU5EIFJTQSBQUklWQVRFIEtFWS0tLS0tCg==",
                    "sslProtocols": ""
                  },
                  "yugaware": {
                    "health": {
                      "email": "",
                      "password": "",
                      "username": ""
                    },
                    "multiTenant": false,
                    "replicas": 1,
                    "service": {
                      "enabled": true,
                      "ip": "",
                      "type": "LoadBalancer"
                    },
                    "serviceAccount": "yugaware",
                    "storage": "100Gi",
                    "storageClass": "standard"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/yugabytedb/yugabyte-platform-operator@sha256:f227c9e2c625b04160d952349ddba552ac66dae8e7da35f56746f3447c270ce3
          description: Operator for Yugabyte Platform makes it easy to deploy and
            upgrade Yugabyte Platform on OpenShift environments.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/yugabyte-platform-operator-bundle-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/yugabyte-platform-operator-bundle-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/yugabyte/yugabyte-platform-operator
          support: YugabyteDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy and manage Yugabyte Platform as a Kubernetes native
              custom resource.
            displayName: Yugabyte Platform
            kind: YBPlatform
            name: ybplatforms.yugabyte.com
            version: v1alpha1
        description: |2

          Operator for Yugabyte Platform makes it easy to deploy and upgrade Yugabyte Platform on OpenShift environments.
          Use Yugabyte Platform's orchestration and monitoring to manage YugabyteDB universes. Yugabyte Platform can create a YugabyteDB universe with many instances (VMs, pods, machines, etc., provided by IaaS), logically grouped together to form one logical distributed database. Each universe includes one or more clusters. A universe is comprised of one primary cluster and, optionally, one or more read replica clusters. All instances belonging to a cluster run on the same type of cloud provider instance type.

          ### Yugabyte Platform Features
          * **YugabyteDB-as-a-Service on Kubernetes** - Deploy your own YugabyteDB-as-a-Service running on any Kubernetes infrastructure, Yugabyte Platform provides central management for multiple database clusters and Includes 24/7 production support.
          * **Monitoring & Alerting** - Comprehensive monitoring & alerting without any configuration.
          * **Scheduled Backups** - Use Yugabyte Platform to perform a regular scheduled backup of a full YugabyteDB universe or selected tables.

          ### Getting Started
          Here are a few resources for getting started with Yugabyte platform
          * [Quick start guide for Yugabyte Platform](https://docs.yugabyte.com/latest/yugabyte-platform/deploy/) - Install Yugabyte platform on OpenShift.
          * [Quick Start for YugabyteDB](http://docs.yugabyte.com/latest/quick-start/) - install, create a local cluster and read/write from YugabyteDB.
          * [Explore core features](https://docs.yugabyte.com/latest/explore/) - automatic sharding & re-balancing, linear scalability, fault tolerance, tunable reads etc.
          * [Ecosystem integrations](https://docs.yugabyte.com/latest/develop/ecosystem-integrations/) - integrations with Apache Kafka/KSQL, Apache Spark, JanusGraph, KairosDB, Presto and more.
          * [Real world apps](https://docs.yugabyte.com/latest/develop/realworld-apps/) - sample real-world, end-to-end applications built using YugabyteDB.

          ### Help
          Cannot find what you are looking for? Have a question? We love to hear from you - [please file a GitHub issue](https://github.com/YugaByte/yugabyte-db/issues) or talk to us live at [Slack](https://www.yugabyte.com/slack).
        displayName: Yugabyte Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - distributed sql
        links:
        - name: Yugabyte Platform Operator
          url: https://www.yugabyte.com/
        maintainers:
        - email: oss-maintainers@yugabyte.com
          name: Yugabyte
        maturity: alpha
        provider:
          name: Yugabyte Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9211b70edbb93d718428e1c9a992e5fc74a20cb213f8a0592ffe340c362890fe
        - registry.marketplace.redhat.com/rhm/yugabytedb/yugabyte-platform-operator@sha256:f227c9e2c625b04160d952349ddba552ac66dae8e7da35f56746f3447c270ce3
        version: 0.0.6
      entries:
      - name: yugabyte-platform-operator.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: yugabyte-platform-operator-bundle-rhmp
    provider:
      name: Yugabyte Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: cinder-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cinder-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cinder.openstack.org/v1beta1",
                "kind": "Cinder",
                "metadata": {
                  "name": "cinder",
                  "namespace": "openstack"
                },
                "spec": {
                  "cinderAPI": {},
                  "cinderBackup": {},
                  "cinderScheduler": {},
                  "cinderVolumes": {
                    "volume1": {}
                  },
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "cinder",
                  "databaseInstance": "openstack",
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "cinder-secret",
                  "serviceUser": "cinder"
                }
              },
              {
                "apiVersion": "cinder.openstack.org/v1beta1",
                "kind": "CinderAPI",
                "metadata": {
                  "name": "cinderapi-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cinder.openstack.org/v1beta1",
                "kind": "CinderBackup",
                "metadata": {
                  "name": "cinderbackup-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cinder.openstack.org/v1beta1",
                "kind": "CinderScheduler",
                "metadata": {
                  "name": "cinderscheduler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cinder.openstack.org/v1beta1",
                "kind": "CinderVolume",
                "metadata": {
                  "name": "cindervolume-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:43Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CinderAPI is the Schema for the cinderapis API
            displayName: Cinder API
            kind: CinderAPI
            name: cinderapis.cinder.openstack.org
            version: v1beta1
          - description: CinderBackup is the Schema for the cinderbackups API
            displayName: Cinder Backup
            kind: CinderBackup
            name: cinderbackups.cinder.openstack.org
            version: v1beta1
          - description: Cinder is the Schema for the cinders API
            displayName: Cinder
            kind: Cinder
            name: cinders.cinder.openstack.org
            version: v1beta1
          - description: CinderScheduler is the Schema for the cinderschedulers API
            displayName: Cinder Scheduler
            kind: CinderScheduler
            name: cinderschedulers.cinder.openstack.org
            version: v1beta1
          - description: CinderVolume is the Schema for the cindervolumes API
            displayName: Cinder Volume
            kind: CinderVolume
            name: cindervolumes.cinder.openstack.org
            version: v1beta1
        description: Cinder Operator
        displayName: Cinder Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Block Storage
        - Cinder
        links:
        - name: Cinder Operator
          url: https://github.com/openstack-k8s-operators/cinder-operator
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/cinder-rhel9-operator@sha256:e0e457500af9db0831c7080b10e010d11213cd800d89e40683974aaf7966a68d
        - registry.redhat.io/rhoso/openstack-cinder-api-rhel9@sha256:ec6835efb7ffc4fa7eed10bd029440263237e8f792e7e43d4cda337d00fecdb6
        - registry.redhat.io/rhoso/openstack-cinder-backup-rhel9@sha256:936e10d80300c325d0f58b4d6ca290fcf3c5cad774eb5cb696330a20c0151561
        - registry.redhat.io/rhoso/openstack-cinder-scheduler-rhel9@sha256:753775cc575321c727b9a6371daedc8774ba3a2287c96f3d44ad8f7025f63841
        - registry.redhat.io/rhoso/openstack-cinder-volume-rhel9@sha256:47b19d993bdfc790c508861dbcc87f0ec51302db8975698af8e1476a0f5209c7
        version: 1.0.6
      entries:
      - name: cinder-operator.v1.0.6
        version: 1.0.6
      - name: cinder-operator.v1.0.5
        version: 1.0.5
      - name: cinder-operator.v1.0.4
        version: 1.0.4
      - name: cinder-operator.v1.0.3
        version: 1.0.3
      - name: cinder-operator.v1.0.2
        version: 1.0.2
      - name: cinder-operator.v1.0.1
        version: 1.0.1
      - name: cinder-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: cinder-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: node-healthcheck-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-healthcheck-operator.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:60d95ea8240ecf701ba974b6df3ee296a7fb3ef5393e0dcf43af15e782f2d153
          createdAt: "2023-11-02 11:44:50"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.3.0 <0.4.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://docs.openshift.com/container-platform/4.12/nodes/nodes/eco-node-health-check-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:60d95ea8240ecf701ba974b6df3ee296a7fb3ef5393e0dcf43af15e782f2d153
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:8d32dee5c72ec1c94e4e92f8d1da935a384e40e379f237c432086a2781f4ecb9
        version: 0.4.1
      entries:
      - name: node-healthcheck-operator.v0.4.1
        version: 0.4.1
      name: 4.12-eus
    - currentCSV: node-healthcheck-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:90cceb176fa903cc8e9aa4c0f7316560a8a354a3138dc26abbfc9884c1c576b5
          createdAt: "2023-10-30 14:21:41"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.4.0 <0.6.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/node-health-check-operator
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:90cceb176fa903cc8e9aa4c0f7316560a8a354a3138dc26abbfc9884c1c576b5
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:b8d5db068bfc81031c2eef86e8649e4fc008b7480e08c9cfab2ab4497098eecf
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel8@sha256:85054b34e119c16cc95cda9dc3f5d63f2d7bfc70634f70973410b2a1b1fa58d4
        version: 0.6.1
      entries:
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.14-eus
    - currentCSV: node-healthcheck-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:4811046e26846b1f12a3b20ddc6209980897c11b171ddd61d26b43630cc1d4b4
          createdAt: "2024-08-09T14:54:37Z"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.8.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          SNR is installed automatically when installing NHC.

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/node-health-check-operator
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:4811046e26846b1f12a3b20ddc6209980897c11b171ddd61d26b43630cc1d4b4
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:942f69efc7ae05d6c4d9c33335e4be71bfae2e6f9ef062dcd7887eebfe02120e
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel8@sha256:062cd9def39a087ae3802cd5208e2f0d5bfca29ba76ebd5139d88348e8c9e923
        version: 0.8.2
      entries:
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.16-eus
    - currentCSV: node-healthcheck-operator.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:f174ada653bb5fb7e58219510d0b10d868273a70db5c486e0b3ad05a1be50640
          createdAt: "2025-08-04 12:38:05"
          description: Detect failed Nodes and trigger remediation with a remediation
            operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <0.9.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.7/html/remediation_fencing_and_maintenance
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel9@sha256:8e9926558c3d8082f298e9e5f58bc5b473020acd44c2ce853c2916cdad71d649
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:f174ada653bb5fb7e58219510d0b10d868273a70db5c486e0b3ad05a1be50640
        - registry.redhat.io/workload-availability/node-remediation-console-rhel9@sha256:37f62407284fe574b973a1a685371d191b710ad70a6bfe8172f19a4568c11cf6
        version: 0.9.1
      entries:
      - name: node-healthcheck-operator.v0.9.1
        version: 0.9.1
      - name: node-healthcheck-operator.v0.9.0
        version: 0.9.0
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.18-eus
    - currentCSV: node-healthcheck-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:8e1dbc35126101e1025bd4c0eac4f4c614eb779ccafbb55050541a27ba7e9706
          createdAt: "2022-08-15 12:34:34"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.2.0 <0.3.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-node-health-check-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b5786bbbef725badf3dfcc2c2c7a86ead5ebb584c978c47aae8b9a62e241b80d
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:8e1dbc35126101e1025bd4c0eac4f4c614eb779ccafbb55050541a27ba7e9706
        version: 0.3.1
      entries:
      - name: node-healthcheck-operator.v0.3.1
        version: 0.3.1
      name: candidate
    - currentCSV: node-healthcheck-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:2160f2688e5b098c0bae7597cebcd380019b82ae02f4506d07c2ed85f0b3664d
          createdAt: "2025-09-01 18:07:56"
          description: Detect failed Nodes and trigger remediation with a remediation
            operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.8/html/remediation_fencing_and_maintenance
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:2160f2688e5b098c0bae7597cebcd380019b82ae02f4506d07c2ed85f0b3664d
        - registry.redhat.io/workload-availability/node-remediation-console-rhel9@sha256:cc0f671b126cde10476f8cc2061c6903932e7bf5e3dbf4e8bdda9add70d5deea
        version: 0.10.0
      entries:
      - name: node-healthcheck-operator.v0.10.0
        version: 0.10.0
      - name: node-healthcheck-operator.v0.9.1
        version: 0.9.1
      - name: node-healthcheck-operator.v0.9.0
        version: 0.9.0
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: stable
    defaultChannel: stable
    packageName: node-healthcheck-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-mq-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-mq-controller.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mq.services.k8s.aws/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/mq-controller:1.1.2
          createdAt: "2025-10-03T18:49:51Z"
          description: AWS MQ controller is a service controller for managing MQ resources
            in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker represents the state of an AWS mq Broker resource.
            displayName: Broker
            kind: Broker
            name: brokers.mq.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon MQ resources in AWS from within your Kubernetes cluster.

          **About Amazon MQ**

          Amazon MQ is a managed message broker service for Apache ActiveMQ and RabbitMQ that makes it easy to set up and operate message brokers on AWS. Amazon MQ reduces your operational responsibilities by managing the provisioning, setup, and maintenance of message brokers for you. Because Amazon MQ connects to your current applications with industry-standard APIs and protocols, you can easily migrate to AWS without having to rewrite code.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon MQ
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mq
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon MQ Developer Resources
          url: https://docs.aws.amazon.com/amazon-mq/latest/developer-guide/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: mq maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/mq-controller:1.1.2
        version: 1.1.2
      entries:
      - name: ack-mq-controller.v1.1.2
        version: 1.1.2
      - name: ack-mq-controller.v1.1.1
        version: 1.1.1
      - name: ack-mq-controller.v1.1.0
        version: 1.1.0
      - name: ack-mq-controller.v1.0.13
        version: 1.0.13
      - name: ack-mq-controller.v1.0.11
        version: 1.0.11
      - name: ack-mq-controller.v1.0.10
        version: 1.0.10
      - name: ack-mq-controller.v1.0.9
        version: 1.0.9
      - name: ack-mq-controller.v1.0.8
        version: 1.0.8
      - name: ack-mq-controller.v1.0.7
        version: 1.0.7
      - name: ack-mq-controller.v1.0.6
        version: 1.0.6
      - name: ack-mq-controller.v1.0.5
        version: 1.0.5
      - name: ack-mq-controller.v1.0.2
        version: 1.0.2
      - name: ack-mq-controller.v1.0.1
        version: 1.0.1
      - name: ack-mq-controller.v1.0.0
        version: 1.0.0
      - name: ack-mq-controller.v0.0.36
        version: 0.0.36
      - name: ack-mq-controller.v0.0.35
        version: 0.0.35
      - name: ack-mq-controller.v0.0.34
        version: 0.0.34
      - name: ack-mq-controller.v0.0.31
        version: 0.0.31
      - name: ack-mq-controller.v0.0.29
        version: 0.0.29
      - name: ack-mq-controller.v0.0.27
        version: 0.0.27
      - name: ack-mq-controller.v0.0.26
        version: 0.0.26
      - name: ack-mq-controller.v0.0.25
        version: 0.0.25
      - name: ack-mq-controller.v0.0.24
        version: 0.0.24
      - name: ack-mq-controller.v0.0.23
        version: 0.0.23
      - name: ack-mq-controller.v0.0.22
        version: 0.0.22
      - name: ack-mq-controller.v0.0.21
        version: 0.0.21
      - name: ack-mq-controller.v0.0.20
        version: 0.0.20
      - name: ack-mq-controller.v0.0.19
        version: 0.0.19
      - name: ack-mq-controller.v0.0.18
        version: 0.0.18
      - name: ack-mq-controller.v0.0.17
        version: 0.0.17
      - name: ack-mq-controller.v0.0.16
        version: 0.0.16
      - name: ack-mq-controller.v0.0.15
        version: 0.0.15
      - name: ack-mq-controller.v0.0.14
        version: 0.0.14
      - name: ack-mq-controller.v0.0.13
        version: 0.0.13
      - name: ack-mq-controller.v0.0.12
        version: 0.0.12
      - name: ack-mq-controller.v0.0.11
        version: 0.0.11
      - name: ack-mq-controller.v0.0.10
        version: 0.0.10
      - name: ack-mq-controller.v0.0.9
        version: 0.0.9
      - name: ack-mq-controller.v0.0.8
        version: 0.0.8
      - name: ack-mq-controller.v0.0.7
        version: 0.0.7
      name: alpha
    defaultChannel: alpha
    packageName: ack-mq-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: volsync-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: volsync-product.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
          createdAt: "2025-08-27T13:12:05Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.13.1'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
        version: 0.13.1
      entries:
      - name: volsync-product.v0.13.1
        version: 0.13.1
      - name: volsync-product.v0.13.0
        version: 0.13.0
      - name: volsync-product.v0.12.2
        version: 0.12.2
      - name: volsync-product.v0.12.1
        version: 0.12.1
      - name: volsync-product.v0.12.1-0.1749677898.p
        version: 0.12.1+0.1749677898.p
      - name: volsync-product.v0.12.0
        version: 0.12.0
      name: stable
    - currentCSV: volsync-product.v0.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:d345825e953c4c170db45a5d92639182b7ee18933287aaeea035f2a2bad13c37
          createdAt: 09 Jan 2025, 00:47
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.10.2'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:d345825e953c4c170db45a5d92639182b7ee18933287aaeea035f2a2bad13c37
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 0.10.2
      entries:
      - name: volsync-product.v0.10.2
        version: 0.10.2
      - name: volsync-product.v0.10.1
        version: 0.10.1
      - name: volsync-product.v0.10.1-0.1730236570.p
        version: 0.10.1+0.1730236570.p
      - name: volsync-product.v0.10.0
        version: 0.10.0
      name: stable-0.10
    - currentCSV: volsync-product.v0.11.2-0.1749674925.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:9da8572c79b75767952cf699e5b1699b9f46d83586aa2d6a17fb23c092713b62
          createdAt: 16 Apr 2025, 06:42
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.11.2'
          olm.substitutesFor: volsync-product.v0.11.2
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:9da8572c79b75767952cf699e5b1699b9f46d83586aa2d6a17fb23c092713b62
        version: 0.11.2+0.1749674925.p
      entries:
      - name: volsync-product.v0.11.2
        version: 0.11.2
      - name: volsync-product.v0.11.2-0.1749674925.p
        version: 0.11.2+0.1749674925.p
      - name: volsync-product.v0.11.1
        version: 0.11.1
      - name: volsync-product.v0.11.0
        version: 0.11.0
      name: stable-0.11
    - currentCSV: volsync-product.v0.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:7d1ac69fe3d0fddc66acdfce9361a4fc8dd3f549f66c98450f7a0e4c5cc893a0
          createdAt: "2025-07-07T13:46:18Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.12.2'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:7d1ac69fe3d0fddc66acdfce9361a4fc8dd3f549f66c98450f7a0e4c5cc893a0
        version: 0.12.2
      entries:
      - name: volsync-product.v0.12.2
        version: 0.12.2
      - name: volsync-product.v0.12.1
        version: 0.12.1
      - name: volsync-product.v0.12.1-0.1749677898.p
        version: 0.12.1+0.1749677898.p
      - name: volsync-product.v0.12.0
        version: 0.12.0
      name: stable-0.12
    - currentCSV: volsync-product.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
          createdAt: "2025-08-27T13:12:05Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.13.1'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
        version: 0.13.1
      entries:
      - name: volsync-product.v0.13.1
        version: 0.13.1
      - name: volsync-product.v0.13.0
        version: 0.13.0
      name: stable-0.13
    - currentCSV: volsync-product.v0.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:ef44675fe70ee9ec8109fb91c6559f0c5518b45219e19175a18a5004e40b65f7
          createdAt: 22 Aug 2024, 22:06
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.9.3'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ReplicationDestination defines the destination for a replicated
              volume
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: ReplicationSource defines the source for a replicated volume
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:ef44675fe70ee9ec8109fb91c6559f0c5518b45219e19175a18a5004e40b65f7
        version: 0.9.3
      entries:
      - name: volsync-product.v0.9.3
        version: 0.9.3
      - name: volsync-product.v0.9.2
        version: 0.9.2
      - name: volsync-product.v0.9.1
        version: 0.9.1
      - name: volsync-product.v0.9.0
        version: 0.9.0
      name: stable-0.9
    defaultChannel: stable
    packageName: volsync-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: ""
    name: gitops-primer
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitops-primer.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "primer.gitops.io/v1alpha1",
                "kind": "Export",
                "metadata": {
                  "name": "primer"
                },
                "spec": {
                  "method": "download"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/migtools/gitops-primer:v0.0.11
          operatorframework.io/suggested-namespace: gitops-primer-system
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Export is the Schema for the exports API
            displayName: Export
            kind: Export
            name: exports.primer.gitops.io
            version: v1alpha1
        description: Export and normalize Kubernetes YAML files for GitOps
        displayName: gitops-primer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - operator
        links:
        - name: Gitops Primer
          url: https://github.com/migtools/gitops-primer
        maturity: alpha
        provider:
          name: Konveyor
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6d57bfd91fac9b68eb72d27226bc297472ceb136c996628b845ecc54a48b31cb
        - quay.io/migtools/gitops-primer:v0.0.11
        - quay.io/migtools/gitops-primer-webhook:v0.0.11
        - quay.io/migtools/gitops-primer-downloader:v0.0.11
        - quay.io/migtools/gitops-primer-export:v0.0.11
        - quay.io/openshift/origin-oauth-proxy:4.9
        version: 0.0.11
      entries:
      - name: gitops-primer.v0.0.11
        version: 0.0.11
      - name: gitops-primer.v0.0.10
        version: 0.0.10
      - name: gitops-primer.v0.0.9
        version: 0.0.9
      - name: gitops-primer.v0.0.8
        version: 0.0.8
      - name: gitops-primer.v0.0.7
        version: 0.0.7
      - name: gitops-primer.v0.0.6
        version: 0.0.6
      - name: gitops-primer.v0.0.5
        version: 0.0.5
      - name: gitops-primer.v0.0.4
        version: 0.0.4
      - name: gitops-primer.v0.0.3
        version: 0.0.3
      - name: gitops-primer.v0.0.2
        version: 0.0.2
      - name: gitops-primer.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: gitops-primer
    provider:
      name: Konveyor
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Craig Trought
      provider-url: ""
    name: prometheus
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometheusoperator.0.56.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "Prometheus",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "replicas": 2,
                  "serviceAccountName": "prometheus-k8s",
                  "serviceMonitorSelector": {},
                  "ruleSelector": {},
                  "podMonitorSelector": {},
                  "probeSelector": {},
                  "alerting": {
                    "alertmanagers": [
                      {
                        "namespace": "monitoring",
                        "name": "alertmanager-main",
                        "port": "web"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "ServiceMonitor",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "operated-prometheus": "true"
                    }
                  },
                  "endpoints": [
                    {
                      "port": "web",
                      "interval": "30s"
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "PodMonitor",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "app": "prometheus"
                    }
                  },
                  "podMetricsEndpoints": [
                    {
                      "port": "web",
                      "interval": "30s"
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "Alertmanager",
                "metadata": {
                  "name": "alertmanager-main"
                },
                "spec": {
                  "replicas": 3,
                  "alertmanagerConfigSelector": {}
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "PrometheusRule",
                "metadata": {
                  "name": "prometheus-example-rules"
                },
                "spec": {
                  "groups": [
                    {
                      "name": "./example.rules",
                      "rules": [
                        {
                          "alert": "ExampleAlert",
                          "expr": "vector(1)"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1alpha1",
                "kind": "AlertmanagerConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "receivers": [
                    {
                      "name": "example"
                    }
                  ],
                  "route": {
                    "receiver": "example"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/prometheus-operator/prometheus-operator:v0.56.3
          createdAt: "2022-05-24T9:40:00Z"
          description: Manage the full lifecycle of configuring and managing Prometheus
            and Alertmanager servers.
          repository: https://github.com/prometheus-operator/prometheus-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A running Prometheus instance
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.coreos.com
            version: v1
          - description: A Prometheus Rule configures groups of sequentially evaluated
              recording and alerting rules.
            displayName: Prometheus Rule
            kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - description: Configures prometheus to monitor a particular k8s service
            displayName: Service Monitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - description: Configures prometheus to monitor a particular pod
            displayName: Pod Monitor
            kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
          - description: Configures an Alertmanager for the namespace
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.coreos.com
            version: v1
          - description: Configures subsections of Alertmanager configuration specific
              to each namespace
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.coreos.com
            version: v1alpha1
          - description: Defines how groups of ingresses and static targets should
              be monitored.
            displayName: Probe
            kind: Probe
            name: probes.monitoring.coreos.com
            version: v1
          - description: Allows recording and alerting rules to be processed across
              multiple Prometheus instances.
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.coreos.com
            version: v1
        description: |
          The Prometheus Operator for Kubernetes provides easy monitoring definitions for Kubernetes services and deployment and management of Prometheus instances.

          Once installed, the Prometheus Operator provides the following features:

          * **Create/Destroy**: Easily launch a Prometheus instance for your Kubernetes namespace, a specific application or team easily using the Operator.
          * **Simple Configuration**: Configure the fundamentals of Prometheus like versions, persistence, retention policies, and replicas from a native Kubernetes resource.
          * **Target Services via Labels**: Automatically generate monitoring target configurations based on familiar Kubernetes label queries; no need to learn a Prometheus specific configuration language.

          ### Other Supported Features

          **High availability**

          Multiple instances are run across failure zones and data is replicated. This keeps your monitoring available during an outage, when you need it most.

          **Updates via automated operations**

          New Prometheus versions are deployed using a rolling update with no downtime, making it easy to stay up to date.

          **Handles the dynamic nature of containers**

          Alerting rules are attached to groups of containers instead of individual instances, which is ideal for the highly dynamic nature of container deployment.
        displayName: Prometheus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - prometheus
        - monitoring
        - tsdb
        - alerting
        - observability
        links:
        - name: Prometheus
          url: https://www.prometheus.io/
        - name: Documentation
          url: https://github.com/prometheus-operator/prometheus-operator/tree/master/Documentation
        - name: Prometheus Operator
          url: https://github.com/prometheus-operator/prometheus-operator
        maintainers:
        - email: k8s@trought.ca
          name: Craig Trought
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Craig Trought
        relatedImages:
        - quay.io/prometheus-operator/prometheus-operator:v0.56.3
        version: 0.56.3
      entries:
      - name: prometheusoperator.0.56.3
        version: 0.56.3
      - name: prometheusoperator.0.47.0
        version: 0.47.0
      name: beta
    defaultChannel: beta
    packageName: prometheus
    provider:
      name: Craig Trought
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC
      provider-url: ""
    name: nspc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nspc-operator.v1.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.nec.com/v1",
                "kind": "NSPC",
                "metadata": {
                  "name": "nspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.connect.redhat.com/nec/nspc-operator@sha256:50fe10f05d8aa7db1b97280e5842bb0b67c6f9bfb01c0d17c5be3cd11b78e853
          createdAt: "2025-07-24T00:00:00Z"
          description: An operator for managing NEC Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: NEC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NSPC is the Schema for the nspcs API
            displayName: NSPC
            kind: NSPC
            name: nspcs.csi.nec.com
            version: v1
        description: |-
          ## About
          NEC Storage Plug-in for Containers (NSPC) is a plugin that integrates NEC Storage V Series into Kubernetes based clusters.
          NSPC provides persistent volume provisioning capabilities from NEC Storage V Series arrays.

          For full documentation, refer to the reference guide.

          ## Requirements

          ### Supported Driver Version

          * NSPC v3.16.0
        displayName: NEC Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - NEC
        - Storage
        - iStorage
        - V Series
        - Container Storage Interface
        - CSI
        maturity: stable
        provider:
          name: NEC
        relatedImages:
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:81d32545fb1b2c319bfeb36a70db4b4d3aea9c51c0574b209b8f64750b32d2f0
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.connect.redhat.com/nec/nspc-operator@sha256:50fe10f05d8aa7db1b97280e5842bb0b67c6f9bfb01c0d17c5be3cd11b78e853
        - registry.connect.redhat.com/nec/nspc-csi-driver@sha256:e50d60e025b3c8ae522a8c5ae78a07f8a624a1210ed7a078c8084efd76ccc646
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        version: 1.16.0
      entries:
      - name: nspc-operator.v1.16.0
        version: 1.16.0
      name: stable
    defaultChannel: stable
    packageName: nspc-operator
    provider:
      name: NEC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: tempo-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tempo-operator.v0.18.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoMonolithic",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "jaegerui": {
                    "enabled": true,
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    }
                  },
                  "storage": {
                    "traces": {
                      "backend": "memory"
                    }
                  }
                }
              },
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoStack",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "resources": {
                    "total": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2Gi"
                      }
                    }
                  },
                  "storage": {
                    "secret": {
                      "name": "my-storage-secret",
                      "type": "s3"
                    }
                  },
                  "storageSize": "1Gi",
                  "template": {
                    "queryFrontend": {
                      "jaegerQuery": {
                        "enabled": true,
                        "ingress": {
                          "type": "route"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/rhosdt/tempo-rhel8-operator@sha256:9bfa5a2ffb09b501f001a4275379210621f06620adcf6454a9327c65538872f5
          createdAt: 22 Sep 2025, 11:15
          description: Create and manage deployments of Tempo, a high-scale distributed
            tracing backend.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.18.0-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-tempo-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/tempo-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TempoMonolithic manages a Tempo deployment in monolithic
              mode.
            displayName: TempoMonolithic
            kind: TempoMonolithic
            name: tempomonolithics.tempo.grafana.com
            version: v1alpha1
          - description: TempoStack manages a Tempo deployment in microservices mode.
            displayName: TempoStack
            kind: TempoStack
            name: tempostacks.tempo.grafana.com
            version: v1alpha1
        description: |-
          Red Hat OpenShift distributed tracing platform based on Tempo. Tempo is an open-source, easy-to-use, and highly scalable distributed tracing backend. It provides observability for microservices architectures by allowing developers to track requests as they flow through distributed systems. Tempo is optimized to handle large volumes of trace data and is designed to be highly performant even under heavy loads.
          It can ingest common open source tracing protocols including Jaeger, Zipkin, and OpenTelemetry and requires only object storage to operate.

          The Tempo Operator supports Tempo deployments in Microservices mode (`TempoStack` CR) and Monolithic mode (`TempoMonolithic` CR).

          ### Core capabilities
          Tempo is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed transaction monitoring
          * Root cause analysis
          * Performance / latency optimization

          ### Operator features
          * **Resource Limits** - Specify overall resource requests and limits in the `TempoStack` CR; the operator assigns fractions of it to each component
          * **AuthN and AuthZ** - Supports OpenID Control (OIDC) and role-based access control (RBAC)
          * **Managed upgrades** - Updating the operator will automatically update all managed Tempo clusters
          * **Multitenancy** - Multiple tenants can send traces to the same Tempo cluster
          * **mTLS** - Communication between the Tempo components can be secured via mTLS
          * **Jaeger UI** - Traces can be visualized in Jaeger UI and exposed via Ingress or OpenShift Route
          * **Observability** - The operator and `TempoStack` operands expose telemetry (metrics, traces) and integrate with Prometheus `ServiceMonitor` and `PrometheusRule`

          ### Before you start
          `TempoStack` requires object storage to store its traces. `TempoMonolithic` can store traces in-memory, in a Persistent Volume and in object storage.
          Please ensure that your system has a compatible object storage solution that is supported, such as OpenShift Data Foundation, Minio, AWS S3, Azure Storage, or Google Cloud Storage.

          ### Examples
          There is a list of examples to help you deploy different configurations in [this GitHub repository](https://github.com/os-observability/redhat-rhosdt-samples).

          ### Support & Troubleshooting
          Tempo Operator is available and supported as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/distributed_tracing) is recommended when installing, configuring, and managing the Operator and instances.
        displayName: Tempo Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tempo
        - tracing
        - observability
        - monitoring
        - database
        links:
        - name: Tempo Operator
          url: https://github.com/grafana/tempo-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6bcfa981a7a75c242a16afa6dab36bf6e3b8a1442fa863875919ad551f48eceb
        - registry.redhat.io/rhosdt/tempo-rhel8@sha256:530a0ef3aaea3f07d6210a683a771c416cea8ce5accac8f5f751dcd7232cb341
        - registry.redhat.io/rhosdt/tempo-query-rhel8@sha256:5b89188bc587a81af3ecdca3caabd6e48356e9ba4f9070860f5160544ce30eb9
        - registry.redhat.io/rhosdt/tempo-rhel8-operator@sha256:9bfa5a2ffb09b501f001a4275379210621f06620adcf6454a9327c65538872f5
        - registry.redhat.io/rhosdt/tempo-jaeger-query-rhel8@sha256:53b62fea39ce211842026e8365d61f579b8e01e8587a1ce639423d72f3b44f45
        - registry.redhat.io/rhosdt/tempo-gateway-rhel8@sha256:095ec83219660bb1955758b94a9431b3add572da4dc2cba61b42b59fa6b4bba0
        - registry.redhat.io/rhosdt/tempo-gateway-opa-rhel8@sha256:22e876837d00371d24feac3f989ead3d4f034400b292f043b18f3839bfde564a
        version: 0.18.0-1
      entries:
      - name: tempo-operator.v0.18.0-1
        version: 0.18.0-1
      - name: tempo-operator.v0.16.0-2
        version: 0.16.0-2
      - name: tempo-operator.v0.16.0-1
        version: 0.16.0-1
      - name: tempo-operator.v0.15.4-1
        version: 0.15.4-1
      - name: tempo-operator.v0.15.3-1
        version: 0.15.3-1
      - name: tempo-operator.v0.14.1-2
        version: 0.14.1-2
      - name: tempo-operator.v0.14.1-1
        version: 0.14.1-1
      - name: tempo-operator.v0.13.0-1-0.1730817419.p
        version: 0.13.0-1+0.1730817419.p
      - name: tempo-operator.v0.13.0-1
        version: 0.13.0-1
      - name: tempo-operator.v0.10.0-8
        version: 0.10.0-8
      - name: tempo-operator.v0.10.0-7
        version: 0.10.0-7
      - name: tempo-operator.v0.10.0-6
        version: 0.10.0-6
      - name: tempo-operator.v0.8.0-3
        version: 0.8.0-3
      - name: tempo-operator.v0.8.0-2
        version: 0.8.0-2
      - name: tempo-operator.v0.6.0-2
        version: 0.6.0-2
      - name: tempo-operator.v0.3.1-3
        version: 0.3.1-3
      - name: tempo-operator.v0.1.0-6
        version: 0.1.0-6
      - name: tempo-operator.v0.1.0-6-0.1687199947.p
        version: 0.1.0-6+0.1687199947.p
      name: stable
    defaultChannel: stable
    packageName: tempo-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: deployment-validation-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: deployment-validation-operator.v0.7.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Application Runtime, Monitoring, Security
          certified: "false"
          containerImage: quay.io/deployment-validation-operator/dv-operator:0.7.12
          createdAt: "2025-09-11T00:00:00Z"
          description: The deployment validation operator
          ignore-check.kube-linter.io/minimum-three-replicas: This deployment uses
            1 pod as currently replicating does not replicate metric data causing
            installation issues
          repository: https://github.com/app-sre/deployment-validation-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Deployment Validation Operator (DVO) checks deployments and other resources against a curated collection of best practices.
          These best practices focus mainly on ensuring that the applications are fault-tolerant. DVO reports failed validations via Prometheus metrics.
          If the best-practice check has failed, the metrics will report `1`.
        displayName: Deployment Validation Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: false
          type: MultiNamespace
        keywords:
        - dvo
        links:
        - name: repository
          url: https://github.com/app-sre/deployment-validation-operator
        - name: containerImage
          url: https://quay.io/deployment-validation-operator/dv-operator:0.7.12
        maintainers:
        - email: dvo-owners@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/deployment-validation-operator/dv-operator:0.7.12
        version: 0.7.12
      entries:
      - name: deployment-validation-operator.v0.7.12
        version: 0.7.12
      - name: deployment-validation-operator.v0.7.9
        version: 0.7.9
      - name: deployment-validation-operator.v0.7.8
        version: 0.7.8
      - name: deployment-validation-operator.v0.7.7
        version: 0.7.7
      - name: deployment-validation-operator.v0.7.6
        version: 0.7.6
      - name: deployment-validation-operator.v0.7.5
        version: 0.7.5
      - name: deployment-validation-operator.v0.7.4
        version: 0.7.4
      - name: deployment-validation-operator.v0.7.3
        version: 0.7.3
      - name: deployment-validation-operator.v0.7.2
        version: 0.7.2
      - name: deployment-validation-operator.v0.7.1
        version: 0.7.1
      - name: deployment-validation-operator.v0.7.0
        version: 0.7.0
      - name: deployment-validation-operator.v0.6.0
        version: 0.6.0
      - name: deployment-validation-operator.v0.5.0
        version: 0.5.0
      - name: deployment-validation-operator.v0.4.0
        version: 0.4.0
      - name: deployment-validation-operator.v0.3.0
        version: 0.3.0
      - name: deployment-validation-operator.v0.2.2
        version: 0.2.2
      name: alpha
    defaultChannel: alpha
    packageName: deployment-validation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Google
      provider-url: https://cloud.google.com/alloydb/omni
    name: alloydb-omni-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: alloydb-omni-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "backupPlanRef": "backupplan1",
                  "dbclusterRef": "dbcluster-sample",
                  "manual": true,
                  "physicalBackupSpec": {
                    "backupType": "full"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "BackupPlan",
                "metadata": {
                  "name": "backupplan1"
                },
                "spec": {
                  "backupRetainDays": 14,
                  "backupSchedules": {
                    "full": "0 0 * * 0",
                    "incremental": "0 21 * * *"
                  },
                  "dbclusterRef": "dbcluster-sample",
                  "paused": false
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "clone1"
                },
                "spec": {
                  "clonedDBClusterConfig": {
                    "dbclusterName": "new-dbcluster-sample"
                  },
                  "pointInTime": "2024-02-23T19:59:43Z",
                  "sourceDBCluster": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "DBCluster",
                "metadata": {
                  "name": "dbcluster-sample"
                },
                "spec": {
                  "databaseVersion": "16.3.0",
                  "primarySpec": {
                    "adminUser": {
                      "passwordRef": {
                        "name": "db-pw-dbcluster-sample"
                      }
                    },
                    "resources": {
                      "cpu": 1,
                      "disks": [
                        {
                          "name": "DataDisk",
                          "size": "10Gi"
                        }
                      ],
                      "memory": "5Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "DBInstance",
                "metadata": {
                  "name": "dbcluster-sample-rp-1"
                },
                "spec": {
                  "dbcParent": {
                    "name": "dbcluster-sample"
                  },
                  "instanceType": "ReadPool",
                  "nodeCount": 2,
                  "resources": {
                    "cpu": 2,
                    "disks": [
                      {
                        "name": "DataDisk",
                        "size": "15Gi"
                      }
                    ],
                    "memory": "6Gi"
                  },
                  "schedulingconfig": {
                    "nodeaffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "preference": {
                            "matchExpressions": [
                              {
                                "key": "another-node-label-key",
                                "operator": "In",
                                "values": [
                                  "another-node-label-value"
                                ]
                              }
                            ]
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "podAffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "podAffinityTerm": {
                            "labelSelector": {
                              "matchExpressions": [
                                {
                                  "key": "app",
                                  "operator": "In",
                                  "values": [
                                    "store"
                                  ]
                                }
                              ]
                            },
                            "topologyKey": "kubernetes.io/hostname"
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "podAntiAffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "podAffinityTerm": {
                            "labelSelector": {
                              "matchExpressions": [
                                {
                                  "key": "security",
                                  "operator": "In",
                                  "values": [
                                    "S1"
                                  ]
                                }
                              ]
                            },
                            "topologyKey": "topology.kubernetes.io/zone"
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane",
                        "operator": "Exists"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Failover",
                "metadata": {
                  "name": "failover-sample"
                },
                "spec": {
                  "dbclusterRef": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "PgBouncer",
                "metadata": {
                  "name": "mypgbouncer"
                },
                "spec": {
                  "allowSuperUserAccess": true,
                  "dbclusterRef": "dbcluster-sample",
                  "parameters": {
                    "default_pool_size": "15",
                    "ignore_startup_parameters": "extra_float_digits",
                    "max_client_conn": "800",
                    "max_db_connections": "160",
                    "pool_mode": "transaction"
                  },
                  "podSpec": {
                    "image": "gcr.io/alloydb-omni/operator/g-pgbouncer:1.5.0",
                    "resources": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "replicaCount": 1,
                  "serviceOptions": {
                    "type": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Replication",
                "metadata": {
                  "name": "replication-downstream-sample"
                },
                "spec": {
                  "dbcluster": {
                    "name": "dbcluster-sample"
                  },
                  "downstream": {
                    "control": "setup",
                    "host": "10.10.10.10",
                    "password": {
                      "name": "ha-rep-pw-dbcluster-sample"
                    },
                    "port": 5432,
                    "replicationSlotName": "dbcluster_sample_replication_upstream_sample",
                    "username": "alloydbreplica"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Replication",
                "metadata": {
                  "name": "replication-upstream-sample"
                },
                "spec": {
                  "dbcluster": {
                    "name": "dbcluster-sample"
                  },
                  "upstream": {}
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backup": "backup1",
                  "sourceDBCluster": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Sidecar",
                "metadata": {
                  "name": "sidecar-sample"
                },
                "spec": {
                  "sidecars": [
                    {
                      "args": [
                        "-c",
                        "while [ true ]\ndo\ndate\nset -x\nls -lh /logs/diagnostic\nset +x\ndone\n"
                      ],
                      "command": [
                        "/bin/sh"
                      ],
                      "image": "busybox",
                      "name": "sidecar-sample",
                      "volumeMounts": [
                        {
                          "mountPath": "/logs",
                          "name": "obsdisk"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Switchover",
                "metadata": {
                  "name": "switchover-sample"
                },
                "spec": {
                  "dbclusterRef": "dbcluster-sample",
                  "newPrimary": "aaaa-dbcluster-sample"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: gcr.io/alloydb-omni/operator/fleet-operator:1.5.0
          createdAt: "2025-07-25T04:12:04Z"
          operatorframework.io/suggested-namespace: alloydb-omni-system
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["backuprepositories.alloydbomni.internal.dbadmin.goog",
            "failovers.alloydbomni.internal.dbadmin.goog", "instancebackupplans.alloydbomni.internal.dbadmin.goog","instancebackups.alloydbomni.internal.dbadmin.goog","instancerestores.alloydbomni.internal.dbadmin.goog","instances.alloydbomni.internal.dbadmin.goog","instanceswitchovers.alloydbomni.internal.dbadmin.goog","instanceuserdefinedauthentications.alloydbomni.internal.dbadmin.goog","lrojobs.alloydbomni.internal.dbadmin.goog","replicationconfigs.alloydbomni.internal.dbadmin.goog","sidecars.alloydbomni.internal.dbadmin.goog","createstandbyjobs.alloydbomni.internal.dbadmin.goog","deletestandbyjobs.alloydbomni.internal.dbadmin.goog"]'
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BackupPlan is the Schema for the backupplans API.
            displayName: Backup Plan
            kind: BackupPlan
            name: backupplans.alloydbomni.dbadmin.goog
            version: v1
          - description: Backup is the Schema for the backups API.
            displayName: Backup
            kind: Backup
            name: backups.alloydbomni.dbadmin.goog
            version: v1
          - description: CreateStandbyJob is an internal workflow tracking object.
              Users should not directly interact with this.
            displayName: (Internal) Create Standby Job
            kind: CreateStandbyJob
            name: createstandbyjobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: DBCluster is the Schema for the DBClusters API.
            displayName: DB Cluster
            kind: DBCluster
            name: dbclusters.alloydbomni.dbadmin.goog
            version: v1
          - description: DBInstance is the Schema for the DBInstances API.
            displayName: DB Instance
            kind: DBInstance
            name: dbinstances.alloydbomni.dbadmin.goog
            version: v1
          - description: DeleteStandbyJob is an internal workflow tracking object.
              Users should not directly interact with this.
            displayName: (Internal) Delete Standby Job
            kind: DeleteStandbyJob
            name: deletestandbyjobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Failover represents the parameters and status of a single
              failover operation.
            displayName: Failover
            kind: Failover
            name: failovers.alloydbomni.dbadmin.goog
            version: v1
          - description: PgBouncer is the Schema for the PgBouncer API.
            displayName: PgBouncer
            kind: PgBouncer
            name: pgbouncers.alloydbomni.dbadmin.goog
            version: v1
          - description: Replication is the Schema for the Replications API.
            displayName: Replication
            kind: Replication
            name: replications.alloydbomni.dbadmin.goog
            version: v1
          - description: Restore is the Schema for the Restores API.
            displayName: Restore
            kind: Restore
            name: restores.alloydbomni.dbadmin.goog
            version: v1
          - description: Sidecar is the Schema for the Sidecar API.
            displayName: Sidecar
            kind: Sidecar
            name: sidecars.alloydbomni.dbadmin.goog
            version: v1
          - description: Switchover represents the parameters and status of a single
              switchover.
            displayName: Switchover
            kind: Switchover
            name: switchovers.alloydbomni.dbadmin.goog
            version: v1
          - description: UserDefinedAuthentication is the Schema for the UserDefinedAuthentications
              API.
            displayName: User Defined Authentication
            kind: UserDefinedAuthentication
            name: userdefinedauthentications.alloydbomni.dbadmin.goog
            version: v1
          - description: BackupRepository is the Schema for the backuprepositories
              API.
            displayName: (Internal) Backup Repository
            kind: BackupRepository
            name: backuprepositories.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Failover is the Schema for the failover API.
            displayName: (Internal) Failover
            kind: Failover
            name: failovers.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceBackupPlan is the Schema for the InstanceBackupPlan
              API.
            displayName: (Internal) Instance Backup Plan
            kind: InstanceBackupPlan
            name: instancebackupplans.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceBackup is the Schema for the InstanceBackup API.
            displayName: (Internal) Instance Backup
            kind: InstanceBackup
            name: instancebackups.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceRestore is the Schema for the InstanceRestore API.
            displayName: (Internal) Instance Restore
            kind: InstanceRestore
            name: instancerestores.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Instance is the Schema for the Instance API.
            displayName: (Internal) Instance
            kind: Instance
            name: instances.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceSwitchover is the Schema for the InstanceSwitchover
              API.
            displayName: (Internal) Instance Switchover
            kind: InstanceSwitchover
            name: instanceswitchovers.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceUserDefinedAuthentication is the Schema for the InstanceUserDefinedAuthentication
              API.
            displayName: (Internal) Instance User Defined Authentication
            kind: InstanceUserDefinedAuthentication
            name: instanceuserdefinedauthentications.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: LROJob is the Schema for the LROJob API.
            displayName: (Internal) LRO Job
            kind: LROJob
            name: lrojobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: ReplicationConfig is the Schema for the ReplicationConfig
              API.
            displayName: (Internal) Replication Config
            kind: ReplicationConfig
            name: replicationconfigs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Sidecar is the Schema for the Sidecar API.
            displayName: (Internal) Sidecar
            kind: Sidecar
            name: sidecars.alloydbomni.internal.dbadmin.goog
            version: v1
        description: |-
          AlloyDB Omni is a database that lets you deploy a streamlined edition of AlloyDB for PostgreSQL in your own computing environment. While being 100% compatible with PostgreSQL applications and tools, AlloyDB Omni delivers higher performance for both OLTP and OLAP workloads and makes administration easy with a number of autopilots. AlloyDB AI provides powerful tools to build generative AI applications with the ability to create, store, and index vector embeddings right in your database.

          ## Post-installation

          After installing the operator, you need to manually create the cert-manager resources in your cluster as this is a strict dependency for our operator. For this, follow these steps:

          ```
          kubectl create ns ${NAMESPACE:?}
          ```

          ```
          kubectl apply -f - <<EOF
          apiVersion: cert-manager.io/v1
          kind: ClusterIssuer
          metadata:
            name: alloydbomni-selfsigned-cluster-issuer
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Issuer
          metadata:
            name: fleet-selfsigned-issuer
            namespace: ${NAMESPACE:?}
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Certificate
          metadata:
            name: fleet-serving-cert
            namespace: ${NAMESPACE:?}
          spec:
            dnsNames:
            - fleet-webhook-service.alloydb-omni-system.svc
            - fleet-webhook-service.alloydb-omni-system.svc.cluster.local
            issuerRef:
              kind: Issuer
              name: fleet-selfsigned-issuer
            secretName: fleet-webhook-server-cert
          ---
          apiVersion: cert-manager.io/v1
          kind: Issuer
          metadata:
            name: local-selfsigned-issuer
            namespace: ${NAMESPACE:?}
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Certificate
          metadata:
            name: local-serving-cert
            namespace: ${NAMESPACE:?}
          spec:
            dnsNames:
            - local-webhook-service.alloydb-omni-system.svc
            - local-webhook-service.alloydb-omni-system.svc.cluster.local
            issuerRef:
              kind: Issuer
              name: local-selfsigned-issuer
            secretName: local-webhook-server-cert
          EOF
          ```

          Replace the following:
          * `NAMESPACE` : the namespace where you have installed your operator. Suggested namespace: `alloydb-omni-system`.

          ## Creating a database cluster

          After you install the AlloyDB Omni operator on your Kubernetes cluster, you can create an AlloyDB Omni database cluster. For each database cluster you create, you must first create a password secret using the following format:

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: db-pw-${DB_CLUSTER_NAME:s?}
            namespace: ${DB_CLUSTER_NAMESPACE:?}
          type: Opaque
          data:
            ${DB_CLUSTER_NAME:?}: "ENCODED_PASSWORD"
          ```

          Replace the following:
          * `DB_CLUSTER_NAME`: the name of this database cluster. For example, `my-db-cluster`.
          * `DB_CLUSTER_NAMESPACE` (Optional): the namespace where you want to create the database cluster. For example, `my-db-cluster-namespace`.
          * `ENCODED_PASSWORD`: the database login password for the default postgres user role, encoded as a base64 string. For example, `Q2hhbmdlTWUxMjM=` for password `ChangeMe123`.

          Follow [these][database-cluster] steps for creating a database cluster.

          ## Connect to AlloyDB Omni running on Kubernetes
          Follow [these][connect] steps to connect to AlloyDB Omni.

          [database-cluster]: https://cloud.google.com/alloydb/omni/docs/deploy-kubernetes#create
          [connect]: https://cloud.google.com/alloydb/omni/docs/run-connect#connect-kubernetes
        displayName: AlloyDB Omni Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - sql
        - postgres
        - vector
        - ai
        - analytics
        - postgresql
        - alloydb
        links:
        - name: Documentation
          url: https://cloud.google.com/alloydb/omni/docs
        - name: Release Notes
          url: https://cloud.google.com/alloydb/docs/release-notes
        maintainers:
        - email: alloydb-omni-ext@google.com
          name: Google Cloud
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Google
          url: https://cloud.google.com/alloydb/omni
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.1
        - gcr.io/alloydb-omni/operator/local-operator:1.5.0
        - gcr.io/alloydb-omni/operator/fleet-operator:1.5.0
        version: 1.5.0
      entries:
      - name: alloydb-omni-operator.v1.5.0
        version: 1.5.0
      - name: alloydb-omni-operator.v1.4.1
        version: 1.4.1
      - name: alloydb-omni-operator.v1.4.0
        version: 1.4.0
      - name: alloydb-omni-operator.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: alloydb-omni-operator
    provider:
      name: Google
      url: https://cloud.google.com/alloydb/omni
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Trilio
      provider-url: https://trilio.io/
    name: tvo-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tvo-operator.6.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOControlPlane",
                "metadata": {
                  "name": "tvocontrolplane-sample"
                },
                "spec": {
                  "crdType": "TVOControlPlane",
                  "common": {
                    "dmapi_workers": "16",
                    "memcached_servers": "",
                    "my_ip": {
                      "host_interface": null
                    },
                    "nova_group_id": "42436",
                    "nova_user_id": "42436",
                    "oslo_enable_proxy_headers_parsing": false,
                    "trustee_role": "creator"
                  },
                  "database": {
                    "common": {
                      "host": "EDIT_DB_HOSTNAME",
                      "root_password": "EDIT_PASSWORD",
                      "root_user_name": "root"
                    },
                    "datamover_api": {
                      "database": "dmapi",
                      "password": "EDIT_PASSWORD",
                      "user": "dmapi"
                    },
                    "wlm_api": {
                      "database": "workloadmgr",
                      "password": "EDIT_PASSWORD",
                      "user": "workloadmgr"
                    }
                  },
                  "helm3_hook": true,
                  "images": {
                    "triliovault_datamover_api": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_datamover_db_sync": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_cloud_trust": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_cron": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_db_sync": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_scheduler": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_workloads": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1"
                  },
                  "keystone": {
                    "common": {
                      "admin_role_name": "admin",
                      "auth_uri": "",
                      "auth_url": "",
                      "ca_cert": "CERT MULTIPLE LINES\n",
                      "cloud_admin_domain_name": "Default",
                      "cloud_admin_project_name": "admin",
                      "cloud_admin_user_name": "admin",
                      "cloud_admin_user_password": "",
                      "is_self_signed_ssl_cert": true,
                      "keystone_auth_host": "",
                      "keystone_auth_port": "",
                      "keystone_auth_protocol": "",
                      "region_name": "regionOne",
                      "service_project_domain_name": "Default",
                      "service_project_name": "service"
                    },
                    "datamover_api": {
                      "internal_endpoint": "ENDPOINT_PROTOCOL_INTERNAL://ENDPOINT_HOST_INTERNAL:8784/v2",
                      "password": "EDIT_PASSWORD",
                      "public_auth_host": "",
                      "public_endpoint": "ENDPOINT_PROTOCOL_PUBLIC://ENDPOINT_HOST_PUBLIC:8784/v2",
                      "service_desc": "TrilioVault Datamover Api Service",
                      "service_name": "dmapi",
                      "service_type": "datamover",
                      "user": "dmapi"
                    },
                    "identity": null,
                    "keystone_interface": "internal",
                    "wlm_api": {
                      "internal_endpoint": "ENDPOINT_PROTOCOL_INTERNAL://ENDPOINT_HOST_INTERNAL:8781/v1/$(tenant_id)s",
                      "password": "EDIT_PASSWORD",
                      "public_auth_host": "",
                      "public_endpoint": "ENDPOINT_PROTOCOL_PUBLIC://ENDPOINT_HOST_PUBLIC:8781/v1/$(tenant_id)s",
                      "service_desc": "TrilioVault Workloadmgr Api Service",
                      "service_name": "TrilioVaultWLM",
                      "service_type": "workloads",
                      "user": "triliovault"
                    }
                  },
                  "labels": {
                    "datamover_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_cron": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_scheduler": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_workloads": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    }
                  },
                  "pod": {
                    "replicas": {
                      "triliovault_datamover_api": 3,
                      "triliovault_object_store": 3,
                      "triliovault_wlm_api": 3,
                      "triliovault_wlm_cron": 1,
                      "triliovault_wlm_scheduler": 3,
                      "triliovault_wlm_workloads": 3
                    },
                    "security_context": {
                      "triliovault": {
                        "container": {
                          "triliovault_datamover_api": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 0,
                            "runAsUser": 63630
                          },
                          "triliovault_datamover_api_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 63630,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_api": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_cron": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_scheduler": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_workloads": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "wlm_cloud_trust_creation": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 42436
                          }
                        }
                      }
                    }
                  },
                  "rabbitmq": {
                    "common": {
                      "admin_password": "EDIT_PASSWORD",
                      "admin_user": "admin",
                      "driver": "messagingv2",
                      "host": "EDIT_HOST",
                      "port": "5672"
                    },
                    "datamover_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "",
                      "user": "dmapi",
                      "vhost": "dmapi"
                    },
                    "wlm_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "",
                      "user": "workloadmgr",
                      "vhost": "workloadmgr"
                    }
                  },
                  "triliovault_backup_targets": [
                    {
                      "backup_target_name": "S3_BT1",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY1",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET1",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL1",
                      "s3_region_name": "REGION1",
                      "s3_secret_key": "SECRETKEY1",
                      "s3_self_signed_cert": false,
                      "s3_signature_version": "default",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "S3_BT2",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY2",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET2",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL2",
                      "s3_region_name": "REGION2",
                      "s3_secret_key": "SECRETKEY2",
                      "s3_self_signed_cert": true,
                      "s3_signature_version": "default",
                      "s3_ssl_ca_cert": "SSL_CERT_LINE_1\n",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "NFS_BT3",
                      "backup_target_type": "nfs",
                      "is_default": false,
                      "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10",
                      "nfs_shares": "11.30.1.10:/mnt/share"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:3d80cad27a1043fb61735e7a6fb156c98b269cfcc68f8b9712713a7d4f6ad6bd
          createdAt: "2025-01-14T11:20:30Z"
          description: Cloud-Native Data Protection for OpenStack
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trilio-openstack
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trilioData/triliovault-cfg-scripts
          support: Trilio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trilio for OpenStack Control Plane. Includes all control
              plane components like Datamover Api, Trilio object store and Workloadmgr
              services.
            displayName: Trilio for OpenStack Control Plane
            kind: TVOControlPlane
            name: tvocontrolplanes.tvo.trilio.io
            version: v1
          - description: Trilio for OpenStack Backup Target. This CR is responsible
              to add a new backup target.
            displayName: Trilio for OpenStack Backup Target
            kind: TVOBackupTarget
            name: tvobackuptargets.tvo.trilio.io
            version: v1
        description: |-
          Trilio for OpenStack provides scalable, native, agentless, multi-tenant backup and recovery solutions, ensuring seamless protection of workloads with flexible storage options for dynamic cloud environments.

            * OpenStack Backup and Recovery: Recover any OpenStack application quickly in the event of data loss or misconfiguration. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Such as One-Click-Restore to entire project recovery on different clusters.
            * Disaster Recovery: Recover any OpenStack application quickly in the event of a disaster. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Trilio makes OpenStack disaster recovery easy.

            * Migration: A tenant or administrator can capture an application and its data and non-disruptively migrate that exact point-in-time to another OpenStack cloud. The need to migrate OpenStack workloads can be driven by economics, security or test/dev.

            * Network Functions Virtualization backup and recovery: Business critical services like NFV can quickly be configured to any state, assuring the optimal quality of service. Trilio provides recovery from data damage and allows restoration of entire VNFs, selected VNFs, or individual data items using a point-in-time working copy. The platform also allows recovery from a disaster by restoring the OpenStack network topology and populating it with a working copy of the entire NFV.

          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Backup
        - Recovery
        - RHOSO Backup
        - Disaster Recovery
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/openstack-backup-and-recovery/
        - name: Operator Documentation
          url: https://docs.trilio.io/openstack
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        provider:
          name: Trilio
          url: https://trilio.io/
        relatedImages:
        - registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82
        - registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1
        - registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:3d80cad27a1043fb61735e7a6fb156c98b269cfcc68f8b9712713a7d4f6ad6bd
        version: 6.1.1
      entries:
      - name: tvo-operator.6.1.1
        version: 6.1.1
      name: 6.0.x
    - currentCSV: tvo-operator.6.1.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOControlPlane",
                "metadata": {
                  "name": "tvocontrolplane-sample",
                  "namespace": "trilio-openstack"
                },
                "spec": {
                  "common": {
                    "dmapi_workers": "16",
                    "memcached_servers": "memcached-0.memcached.openstack.svc:11212",
                    "nova_group_id": "42436",
                    "nova_user_id": "42436",
                    "oslo_enable_proxy_headers_parsing": false,
                    "trustee_role": "member"
                  },
                  "crdType": "TVOControlPlane",
                  "database": {
                    "common": {
                      "host": "openstack.openstack.svc",
                      "port": 3306,
                      "root_password": "openstack",
                      "root_user_name": "root"
                    },
                    "datamover_api": {
                      "database": "dmapi",
                      "password": "dsw21aSyH5gHH6XE72k11Y3t3C308jY5rP02ndwJW",
                      "user": "dmapi"
                    },
                    "wlm_api": {
                      "database": "workloadmgr",
                      "password": "dfeeYb72OwUzX5YuRKcVjxXJaELV8YT0jZczIQwys0v",
                      "user": "workloadmgr"
                    }
                  },
                  "helm3_hook": true,
                  "images": {
                    "triliovault_datamover_api": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_datamover_db_sync": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_cloud_trust": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_cron": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_db_sync": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_scheduler": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_workloads": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd"
                  },
                  "keystone": {
                    "common": {
                      "admin_role_name": "admin",
                      "auth_uri": "https://keystone-internal.openstack.svc:5000/v3",
                      "auth_url": "https://keystone-internal.openstack.svc:5000",
                      "cloud_admin_domain_name": "Default",
                      "cloud_admin_project_name": "admin",
                      "cloud_admin_user_name": "admin",
                      "cloud_admin_user_password": "openstack",
                      "is_self_signed_ssl_cert": true,
                      "keystone_auth_host": "keystone-internal.openstack.svc",
                      "keystone_auth_port": "5000",
                      "keystone_auth_protocol": "https",
                      "region_name": "regionOne",
                      "service_project_domain_name": "Default",
                      "service_project_name": "service"
                    },
                    "datamover_api": {
                      "internal_endpoint": "https://triliovault-datamover-internal.trilio-openstack.svc:8784/v2",
                      "password": "defajgh48zgt88wFzzhwEt3xHGeCOdExYbLGQN2XRm",
                      "public_auth_host": "triliovault-datamover-public-trilio-openstack.apps.trilio.trilio.bos2",
                      "public_endpoint": "https://triliovault-datamover-public-trilio-openstack.apps.trilio.trilio.bos2/v2",
                      "service_desc": "TrilioVault Datamover Api Service",
                      "service_name": "dmapi",
                      "service_type": "datamover",
                      "user": "dmapi"
                    },
                    "keystone_interface": "internal",
                    "wlm_api": {
                      "internal_endpoint": "https://triliovault-wlm-internal.trilio-openstack.svc:8781/v1/$(tenant_id)s",
                      "password": "dsdeJl8axy0gv14fURwUUxy0qUOQbugUUOfXikz6j",
                      "public_auth_host": "triliovault-wlm-public-trilio-openstack.apps.trilio.trilio.bos2",
                      "public_endpoint": "https://triliovault-wlm-public-trilio-openstack.apps.trilio.trilio.bos2/v1/$(tenant_id)s",
                      "service_desc": "TrilioVault Workloadmgr Api Service",
                      "service_name": "TrilioVaultWLM",
                      "service_type": "workloads",
                      "user": "triliovault"
                    }
                  },
                  "labels": {
                    "datamover_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_cron": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_scheduler": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_workloads": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    }
                  },
                  "pod": {
                    "replicas": {
                      "triliovault_datamover_api": 3,
                      "triliovault_object_store": 3,
                      "triliovault_wlm_api": 3,
                      "triliovault_wlm_cron": 1,
                      "triliovault_wlm_scheduler": 3,
                      "triliovault_wlm_workloads": 3
                    },
                    "security_context": {
                      "triliovault": {
                        "container": {
                          "triliovault_datamover_api": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 0,
                            "runAsUser": 63630
                          },
                          "triliovault_datamover_api_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 63630,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_api": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_cron": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_scheduler": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_workloads": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "wlm_cloud_trust_creation": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 42436
                          }
                        }
                      }
                    }
                  },
                  "rabbitmq": {
                    "common": {
                      "admin_password": "EDIT_PASSWORD",
                      "admin_user": "admin",
                      "driver": "messagingv2",
                      "host": "EDIT_HOST",
                      "port": "15671",
                      "ssl": true
                    },
                    "datamover_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "rabbit://dmapi:${PASSWORD}@${RABBITMQ_HOST}:5671/dmapi?ssl=1",
                      "user": "dmapi",
                      "vhost": "dmapi"
                    },
                    "wlm_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "rabbit://workloadmgr:${PASSWORD}@${RABBITMQ_HOST}:5671/workloadmgr?ssl=1",
                      "user": "workloadmgr",
                      "vhost": "workloadmgr"
                    }
                  },
                  "triliovault_backup_targets": [
                    {
                      "backup_target_name": "BT1_S3",
                      "backup_target_type": "s3",
                      "is_default": false,
                      "s3_access_key": "",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "",
                      "s3_region_name": "",
                      "s3_secret_key": "",
                      "s3_self_signed_cert": false,
                      "s3_signature_version": "default",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "BT2_Other_S3",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY2",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET2",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL2",
                      "s3_region_name": "REGION2",
                      "s3_secret_key": "SECRETKEY2",
                      "s3_self_signed_cert": true,
                      "s3_signature_version": "default",
                      "s3_ssl_ca_cert": "SSL_CERT_LINE_1\nSSL_CERT_LINE_2\nSSL_CERT_LINE_N\n",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "other_s3"
                    },
                    {
                      "backup_target_name": "BT3_NFS",
                      "backup_target_type": "nfs",
                      "is_default": false,
                      "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10",
                      "nfs_shares": "192.168.2.3:/nfs/share"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOBackupTarget",
                "metadata": {
                  "name": "tvobackuptarget-sample",
                  "namespace": "trilio-openstack"
                },
                "spec": {
                  "crdType": "TVOBackupTarget",
                  "images": {
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd"
                  },
                  "triliovault_backup_target": {
                    "backup_target_name": "S3_BT4",
                    "backup_target_type": "s3",
                    "is_default": false,
                    "s3_type": "other_s3",
                    "s3_access_key": "",
                    "s3_secret_key": "",
                    "s3_region_name": "",
                    "s3_bucket": "",
                    "s3_endpoint_url": "",
                    "s3_signature_version": "default",
                    "s3_auth_version": "DEFAULT",
                    "s3_bucket_object_lock_enabled": false,
                    "s3_ssl_enabled": true,
                    "s3_ssl_verify": true,
                    "s3_self_signed_cert": true,
                    "s3_ssl_ca_cert": "SSL_CERT_LINE_1\nSSL_CERT_LINE_2\nSSL_CERT_LINE_N",
                    "nfs_shares": "192.168.2.30:/nfs/share",
                    "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:e282bb056a1780e5170db3e63efdf5099a5ac5240ad44162bc7af83cac893ac7
          createdAt: "2025-01-14T11:20:30Z"
          description: Cloud-Native Data Protection for OpenStack
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: tvo-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trilioData/triliovault-cfg-scripts
          support: Trilio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trilio for OpenStack Control Plane. Includes all control
              plane components like Datamover Api, Trilio object store and Workloadmgr
              services.
            displayName: Trilio for OpenStack Control Plane
            kind: TVOControlPlane
            name: tvocontrolplanes.tvo.trilio.io
            version: v1
          - description: Trilio for OpenStack Backup Target. This CR is responsible
              to add a new backup target.
            displayName: Trilio for OpenStack Backup Target
            kind: TVOBackupTarget
            name: tvobackuptargets.tvo.trilio.io
            version: v1
        description: |-
          Trilio for OpenStack provides scalable, native, agentless, multi-tenant backup and recovery solutions, ensuring seamless protection of workloads with flexible storage options for dynamic cloud environments.

            * OpenStack Backup and Recovery: Recover any OpenStack application quickly in the event of data loss or misconfiguration. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Such as One-Click-Restore to entire project recovery on different clusters.
            * Disaster Recovery: Recover any OpenStack application quickly in the event of a disaster. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Trilio makes OpenStack disaster recovery easy.

            * Migration: A tenant or administrator can capture an application and its data and non-disruptively migrate that exact point-in-time to another OpenStack cloud. The need to migrate OpenStack workloads can be driven by economics, security or test/dev.

            * Network Functions Virtualization backup and recovery: Business critical services like NFV can quickly be configured to any state, assuring the optimal quality of service. Trilio provides recovery from data damage and allows restoration of entire VNFs, selected VNFs, or individual data items using a point-in-time working copy. The platform also allows recovery from a disaster by restoring the OpenStack network topology and populating it with a working copy of the entire NFV.

          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Backup
        - Recovery
        - RHOSO Backup
        - Disaster Recovery
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/openstack-backup-and-recovery/
        - name: Operator Documentation
          url: https://docs.trilio.io/openstack
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        provider:
          name: Trilio
          url: https://trilio.io/
        relatedImages:
        - registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd
        - registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:e282bb056a1780e5170db3e63efdf5099a5ac5240ad44162bc7af83cac893ac7
        - registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82
        version: 6.1.9
      entries:
      - name: tvo-operator.6.1.9
        version: 6.1.9
      name: 6.1.x
    defaultChannel: 6.0.x
    packageName: tvo-operator
    provider:
      name: Trilio
      url: https://trilio.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: elasticsearch-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: elasticsearch-operator.v5.8.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Elasticsearch",
                "metadata": {
                  "name": "elasticsearch"
                },
                "spec": {
                  "indexManagement": {
                    "mappings": [
                      {
                        "aliases": [
                          "infra",
                          "logs.infra"
                        ],
                        "name": "infra",
                        "policyRef": "infra-policy"
                      }
                    ],
                    "policies": [
                      {
                        "name": "infra-policy",
                        "phases": {
                          "delete": {
                            "minAge": "2d",
                            "namespaceSpec": [
                              {
                                "minAge": "5h",
                                "namespace": "openshift-monitoring"
                              }
                            ],
                            "pruneNamespacesInterval": "24h"
                          },
                          "hot": {
                            "actions": {
                              "rollover": {
                                "maxAge": "8h"
                              }
                            }
                          }
                        },
                        "pollInterval": "30m"
                      }
                    ]
                  },
                  "managementState": "Managed",
                  "nodeSpec": {
                    "resources": {
                      "limits": {
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "512Mi"
                      }
                    }
                  },
                  "nodes": [
                    {
                      "nodeCount": 1,
                      "roles": [
                        "client",
                        "data",
                        "master"
                      ],
                      "storage": {
                        "size": "20G"
                      }
                    }
                  ],
                  "redundancyPolicy": "ZeroRedundancy"
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Kibana",
                "metadata": {
                  "name": "kibana"
                },
                "spec": {
                  "managementState": "Managed",
                  "nodeSelector": {},
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "memory": "512Mi"
                    },
                    "requests": {
                      "memory": "512Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
          createdAt: "2024-09-19T15:14:33Z"
          description: |
            The Elasticsearch Operator for OCP provides a means for configuring and managing an Elasticsearch cluster for tracing and cluster logging.
            ## Prerequisites and Requirements
            ### Elasticsearch Operator Namespace
            The Elasticsearch Operator must be deployed to the global operator group namespace
            ### Memory Considerations
            Elasticsearch is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Elasticsearch cluster.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory. Each ES node can operate with a
            lower memory setting though this is not recommended for production deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.21'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Cluster Logging, Jaeger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Elasticsearch cluster instance
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.openshift.io
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.openshift.io
            version: v1
        description: "### IN OCP 4.16 AND FORWARD \n\nELASTICSEARCH IS ONLY SUPPORTED
          FOR **SERVICEMESH, TRACING, AND KIALI**, AND THIS OPERATOR **WILL BE REMOVED
          FROM THE OPENSHIFT OPERATOR CATALOG IN NOVEMBER 2025**.  ELASTICSEARCH IS
          **NO LONGER SUPPORTED FOR LOG STORAGE**. **KIBANA IS NO LONGER SUPPORTED**
          IN OCP 4.16 AND FORWARD. \n\n### About \n\nThe Elasticsearch Operator for
          OCP provides a means for configuring and managing an Elasticsearch cluster
          for use in tracing \nand cluster logging as well as a Kibana instance to
          connect to it.\nThis operator only supports OCP Cluster Logging and Jaeger.
          \ It is tightly coupled to each and is not currently capable of\nbeing used
          as a general purpose manager of Elasticsearch clusters running on OCP.\n\nPlease
          note: For a general purpose Elasticsearch operator, please use Elastic's
          Elasticsearch (ECK) Operator [here](https://catalog.redhat.com/software/containers/elastic/eck-operator/5fabf6d1ecb52450895164be?container-tabs=gti)\n\nIt
          is recommended that this operator be installed in the `openshift-operators-redhat`
          namespace to \nproperly support the Cluster Logging and Jaeger use cases.\n\nOnce
          installed, the operator provides the following features for **Elasticsearch**:\n*
          **Create/Destroy**: Deploy an Elasticsearch cluster to the same namespace
          in which the elasticsearch CR is created.\n* **Update**: Changes to the
          elasticsearch CR will be scheduled and applied to the cluster in a controlled
          manner (most often as a rolling upgrade).\n* **Cluster health**: The operator
          will periodically poll the cluster to evaluate its current health (such
          as the number of active shards and if any cluster nodes have reached their
          storage watermark usage).\n* **Redeploys**: In the case where the provided
          secrets are updated, the Elasticsearch Operator will schedule and perform
          a full cluster restart.\n* **Index management**: The Elasticsearch Operator
          will create cronjobs to perform index management such as roll over and deletion.\n\nOnce
          installed, the operator provides the following features for **Kibana**:\n*
          **Create/Destroy**: Deploy a Kibana instance to the same namespace in which
          the kibana CR is created (this should be the same namespace as the elasticsearch
          CR).\n* **Update**: Changes to the kibana CR will be scheduled and applied
          to the cluster in a controlled manner.\n* **Redeploys**: In the case where
          the provided secrets are updated, the Elasticsearch Operator will perform
          a restart.\n\n### Additionally provided features\n* Out of the box multitenancy
          that is integrated with OCP user access control.\n* Document Level Security\n*
          mTLS communication between Elasticsearch, Kibana, Index Management cronjobs,
          and CLO's Fluentd\n* OCP prometheus dashboard for Elasticsearch clusters\n*
          Prometheus Alerting rules  \n"
        displayName: OpenShift Elasticsearch Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - jaeger
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Elasticsearch Operator
          url: https://github.com/openshift/elasticsearch-operator
        maintainers:
        - email: aos-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8d603100978c655020a8f4d81c527162348f468d93989edbc00463b8297fb8a5
        - registry.redhat.io/openshift-logging/kibana6-rhel8@sha256:a2c200667738fdb079ba11ed3b0df3c5901373428d5e560bc1efea9d5e1ddd67
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
        - registry.redhat.io/openshift-logging/logging-curator5-rhel9@sha256:394e0f46fb8be44c720b29aed716443747a636d2758226c9386f930f95758470
        - registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:b26d2218aa6b1efea83a1717d62f81009c97e8082434b918937c609f5cdab912
        version: 5.8.21
      entries:
      - name: elasticsearch-operator.v5.8.21
        version: 5.8.21
      - name: elasticsearch-operator.v5.8.20
        version: 5.8.20
      - name: elasticsearch-operator.v5.8.19
        version: 5.8.19
      - name: elasticsearch-operator.v5.8.18
        version: 5.8.18
      - name: elasticsearch-operator.v5.8.17
        version: 5.8.17
      - name: elasticsearch-operator.v5.8.16
        version: 5.8.16
      - name: elasticsearch-operator.v5.8.15
        version: 5.8.15
      - name: elasticsearch-operator.v5.8.14
        version: 5.8.14
      - name: elasticsearch-operator.v5.8.13
        version: 5.8.13
      - name: elasticsearch-operator.v5.8.12
        version: 5.8.12
      - name: elasticsearch-operator.v5.8.11
        version: 5.8.11
      - name: elasticsearch-operator.v5.8.10
        version: 5.8.10
      - name: elasticsearch-operator.v5.8.9
        version: 5.8.9
      - name: elasticsearch-operator.v5.8.8
        version: 5.8.8
      - name: elasticsearch-operator.v5.8.7
        version: 5.8.7
      - name: elasticsearch-operator.v5.8.6
        version: 5.8.6
      - name: elasticsearch-operator.v5.8.5
        version: 5.8.5
      - name: elasticsearch-operator.v5.8.4
        version: 5.8.4
      - name: elasticsearch-operator.v5.8.3
        version: 5.8.3
      - name: elasticsearch-operator.v5.8.2
        version: 5.8.2
      - name: elasticsearch-operator.v5.8.1
        version: 5.8.1
      - name: elasticsearch-operator.v5.8.0
        version: 5.8.0
      name: stable
    - currentCSV: elasticsearch-operator.v5.8.22
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Elasticsearch",
                "metadata": {
                  "name": "elasticsearch"
                },
                "spec": {
                  "indexManagement": {
                    "mappings": [
                      {
                        "aliases": [
                          "infra",
                          "logs.infra"
                        ],
                        "name": "infra",
                        "policyRef": "infra-policy"
                      }
                    ],
                    "policies": [
                      {
                        "name": "infra-policy",
                        "phases": {
                          "delete": {
                            "minAge": "2d",
                            "namespaceSpec": [
                              {
                                "minAge": "5h",
                                "namespace": "openshift-monitoring"
                              }
                            ],
                            "pruneNamespacesInterval": "24h"
                          },
                          "hot": {
                            "actions": {
                              "rollover": {
                                "maxAge": "8h"
                              }
                            }
                          }
                        },
                        "pollInterval": "30m"
                      }
                    ]
                  },
                  "managementState": "Managed",
                  "nodeSpec": {
                    "resources": {
                      "limits": {
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "512Mi"
                      }
                    }
                  },
                  "nodes": [
                    {
                      "nodeCount": 1,
                      "roles": [
                        "client",
                        "data",
                        "master"
                      ],
                      "storage": {
                        "size": "20G"
                      }
                    }
                  ],
                  "redundancyPolicy": "ZeroRedundancy"
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Kibana",
                "metadata": {
                  "name": "kibana"
                },
                "spec": {
                  "managementState": "Managed",
                  "nodeSelector": {},
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "memory": "512Mi"
                    },
                    "requests": {
                      "memory": "512Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.12-v4.17
          containerImage: registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:ce1f01f7a7fd8fa3d52f6bca62f63f65aebcb4c9a8ff8f6afaca7e5d121f0406
          createdAt: "2025-09-23T09:54:59Z"
          description: |
            The Elasticsearch Operator for OCP provides a means for configuring and managing an Elasticsearch cluster for tracing and cluster logging.
            ## Prerequisites and Requirements
            ### Elasticsearch Operator Namespace
            The Elasticsearch Operator must be deployed to the global operator group namespace
            ### Memory Considerations
            Elasticsearch is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Elasticsearch cluster.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory. Each ES node can operate with a
            lower memory setting though this is not recommended for production deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.22'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-5.8
          operators.operatorframework.io.bundle.channels.v1: stable-5.8
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: elasticsearch-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Cluster Logging, Jaeger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Elasticsearch cluster instance
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.openshift.io
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.openshift.io
            version: v1
        description: |
          ### IN OCP 4.16 AND FORWARD

          IN OCP 4.16 AND FORWARD ELASTICSEARCH and KIBANA ARE **NO LONGER SUPPORTED
          FOR LOG STORAGE** AND ELASTICSEARCH IS ONLY SUPPORTED FOR **SERVICEMESH,
          TRACING, AND KIALI**

          THIS OPERATOR **WILL REACH END OF LIFE ON NOVEMBER 2025**

          ### About

          The Elasticsearch Operator for OCP provides a means for configuring and
          managing an Elasticsearch cluster for use in tracing and cluster logging as
          well as a Kibana instance to connect to it. This operator only supports OCP
          Cluster Logging and Jaeger.  It is tightly coupled to each and is not
          currently capable of being used as a general purpose manager of
          Elasticsearch clusters running on OCP.

          Please note: For a general purpose Elasticsearch operator, please use
          Elastic's Elasticsearch (ECK) Operator
          [here](https://catalog.redhat.com/software/containers/elastic/eck-operator/5fabf6d1ecb52450895164be?container-tabs=gti)

          It is recommended that this operator be installed in the
          `openshift-operators-redhat` namespace to properly support the Cluster
          Logging and Jaeger use cases.

          Once installed, the operator provides the following features for
          **Elasticsearch**: * **Create/Destroy**: Deploy an Elasticsearch cluster to
          the same namespace in which the elasticsearch CR is created.
          * **Update**: Changes to the elasticsearch CR will be scheduled and applied
          to the cluster in a controlled manner (most often as a rolling upgrade).
          * **Cluster health**: The operator will periodically poll the cluster to
          evaluate its current health (such as the number of active shards and if any
          cluster nodes have reached their storage watermark usage).
          * **Redeploys**: In the case where the provided secrets are updated, the
          Elasticsearch Operator will schedule and perform a full cluster restart.
          * **Index management**: The Elasticsearch Operator will create cronjobs to
          perform index management such as roll over and deletion.

          Once installed, the operator provides the following features for **Kibana**:

          * **Create/Destroy**: Deploy a Kibana instance to the same namespace in
          which the kibana CR is created (this should be the same namespace as the
          elasticsearch CR).
          * **Update**: Changes to the kibana CR will be scheduled and applied to the
          cluster in a controlled manner.
          * **Redeploys**: In the case where the provided secrets are updated, the
          Elasticsearch Operator will perform a restart.

          ### Additionally provided features

          * Out of the box multitenancy that is integrated with OCP user access
          control.
          * Document Level Security
          * mTLS communication between Elasticsearch, Kibana, Index Management
          cronjobs, and CLO's Fluentd
          * OCP prometheus dashboard for Elasticsearch clusters
          * Prometheus Alerting rules
        displayName: OpenShift Elasticsearch Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - jaeger
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Elasticsearch Operator
          url: https://github.com/openshift/elasticsearch-operator
        maintainers:
        - email: aos-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:6388562dc496e524a297c733720dea23a2889468fa25394812b3247f5f08c907
        - registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:69878c9e40178530baf837a28d3f78e63eb6fba54865dbf05f30dda1a661d8d0
        - registry.redhat.io/openshift-logging/kibana6-rhel8@sha256:7cfb21861cbc3c6d8c6b6f0192ac3c687e4ac9b16fbccf56841e179a08553c84
        - registry.redhat.io/openshift-logging/logging-curator5-rhel9@sha256:dd1cef6a91ca25231af46e0417c7929663098a3b4ad1230dddf4a8171871ce0e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:ce1f01f7a7fd8fa3d52f6bca62f63f65aebcb4c9a8ff8f6afaca7e5d121f0406
        version: 5.8.22
      entries:
      - name: elasticsearch-operator.v5.8.22
        version: 5.8.22
      - name: elasticsearch-operator.v5.8.21
        version: 5.8.21
      - name: elasticsearch-operator.v5.8.20
        version: 5.8.20
      - name: elasticsearch-operator.v5.8.19
        version: 5.8.19
      - name: elasticsearch-operator.v5.8.18
        version: 5.8.18
      - name: elasticsearch-operator.v5.8.17
        version: 5.8.17
      - name: elasticsearch-operator.v5.8.16
        version: 5.8.16
      - name: elasticsearch-operator.v5.8.15
        version: 5.8.15
      - name: elasticsearch-operator.v5.8.14
        version: 5.8.14
      - name: elasticsearch-operator.v5.8.13
        version: 5.8.13
      - name: elasticsearch-operator.v5.8.12
        version: 5.8.12
      - name: elasticsearch-operator.v5.8.11
        version: 5.8.11
      - name: elasticsearch-operator.v5.8.10
        version: 5.8.10
      - name: elasticsearch-operator.v5.8.9
        version: 5.8.9
      - name: elasticsearch-operator.v5.8.8
        version: 5.8.8
      - name: elasticsearch-operator.v5.8.7
        version: 5.8.7
      - name: elasticsearch-operator.v5.8.6
        version: 5.8.6
      - name: elasticsearch-operator.v5.8.5
        version: 5.8.5
      - name: elasticsearch-operator.v5.8.4
        version: 5.8.4
      - name: elasticsearch-operator.v5.8.3
        version: 5.8.3
      - name: elasticsearch-operator.v5.8.2
        version: 5.8.2
      - name: elasticsearch-operator.v5.8.1
        version: 5.8.1
      - name: elasticsearch-operator.v5.8.0
        version: 5.8.0
      name: stable-5.8
    defaultChannel: stable-5.8
    packageName: elasticsearch-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: logging-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: logging-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Elasticsearch","metadata":{"name":"elasticsearch"},"spec":{"esClusterName":"prod","esVersion":"7.17.0","esSecurity":{"autoGeneratePassword":true,"tlsEnabled":true}}},{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Fluentd","metadata":{"name":"fluentd"},"spec":{"esCluster":{"host":"elasticsearch-master"},"esSecurity":{"tlsEnabled":true,"existingSecret":"elasticsearch-password"},"indexNameStrategy":"namespace_name"}},{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Kibana","metadata":{"name":"kibana"},"spec":{"replicas":1,"esCluster":{"host":"https://elasticsearch-master:9200","esVersion":"7.16.0","clusterName":"elasticsearch"},"esSecurity":{"tlsEnabled":true,"existingSecret":"elasticsearch-password"}}}]'
          capabilities: Seamless Upgrades
          categories: Database,  Monitoring, Big Data
          certified: "false"
          containerImage: quay.io/opstree/logging-operator:v0.4.0
          createdAt: 04-07-2022
          description: A golang-based operator to create and manage EFK (Elasticsearch,
            Fluentd, and Kibana) stack on Kubernetes.
          k8sMaxVersion: ""
          k8sMinVersion: ""
          repository: https://github.com/ot-container-kit/logging-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Elasticsearch
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Index Life Cycle
            displayName: IndexLifeCycle
            kind: IndexLifeCycle
            name: indexlifecycles.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Index Template
            displayName: IndexTemplate
            kind: IndexTemplate
            name: indextemplates.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Fluentd
            displayName: Fluentd
            kind: Fluentd
            name: fluentds.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Kibana
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.logging.opstreelabs.in
            version: v1beta1
        description: |2+

          Logging Operator is an operator created in Golang to set up and manage EFK(Elasticsearch, Fluentd, and Kibana) cluster inside Kubernetes and Openshift environment. This operator is capable of setting up each individual component of EFK cluster separately.
          ### Supported Features
          - Setup of elasticsearch cluster with different nodes type:- master, data, ingestion, and client.
          - Customizable configuration for elasticsearch, fluentd stack.
          - Setup of fluentd as light-weight log-shipper as Daemonset.
          - Kibana will be set up as visualization tool for elastic stack.
          - Seamless upgrades of elasticsearch, fluentd, and kibana.
          - Security best practices support for complete stack such as TLS, elastic security.
          - Kubernetes resources objects support like:- resources, securityContext, affinity, tolerations, etc.
          - Elasticsearch plugins and keystore seamless support with operator
          ### Documentation
          **[Documentation](https://ot-logging-operator.netlify.app/)**
          ### Getting Help
          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.
          ### Contributing
          You can contribute to this project by:-
          * Raising Bugs and Feature Requests
          * Fixing issue by opening a Pull Request
          * Improving Documentation
          ### License
          MongoDB Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/logging-operator/blob/master/LICENSE)

        displayName: Logging Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Operator
        - Elasticsearch
        - Fluentd
        - Kibana
        - Opstree
        - Logging
        - Monitoring
        - Cloud Native
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-logging-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep
        maturity: beta
        minKubeVersion: 1.17.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/opstree/logging-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: logging-operator.v0.4.0
        version: 0.4.0
      - name: logging-operator.v0.3.0
        version: 0.3.0
      name: beta
    defaultChannel: beta
    packageName: logging-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-licensing-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-licensing-operator.v8.0.220
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Developer"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:feeabfe4ecd36205306e3f3bf6c11517cfee735d7ae7eb4350abd01e9d76610b
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          olm.skipRange: <8.0.220
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:2f3de2bf46920cfe8c251011fc99841b43999561c1d3fe44850b205e94585a68
        version: 8.0.220
      entries:
      - name: infoscale-licensing-operator.v8.0.220
        version: 8.0.220
      - name: infoscale-licensing-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-licensing-operator.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-licensing-operator.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:b357b2203466516684c8fbc209757a98a45ec2376981f8699faf7dd5e37486a3
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:b357b2203466516684c8fbc209757a98a45ec2376981f8699faf7dd5e37486a3
        version: 8.0.330
      entries:
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-licensing-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Arctera
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Arctera
        - InfoScale
        - Licensing
        - Arctera Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
        version: 8.0.410
      entries:
      - name: infoscale-licensing-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-licensing-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-licensing-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-licensing-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Arctera
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Arctera
        - InfoScale
        - Licensing
        - Arctera Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
        version: 8.0.410
      entries:
      - name: infoscale-licensing-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-licensing-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-licensing-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-licensing-operator.v8.0.110
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Developer"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:e7b8d3c44df6151e61ded91336273e35eeaec04bd19124cf6313dc16d1224660
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:e7b8d3c44df6151e61ded91336273e35eeaec04bd19124cf6313dc16d1224660
        version: 8.0.110
      entries:
      - name: infoscale-licensing-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-licensing-operator.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-licensing-operator
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kedify
      provider-url: kedify.io
    name: kedify-agent-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kedify-agent-helm-operator.v0.2.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kedify.io/v1alpha1",
                "kind": "KedifyAgent",
                "metadata": {
                  "name": "kedifyagent-sample"
                },
                "spec": {
                  "agent": {
                    "agentId": "",
                    "apiKey": "kfy_xxxxxxxxxxxxxxxxxxxxx",
                    "createCrds": true,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v0.2.16@sha256:bfcc0b24ff75de17464ec213406f48bc211a932308e888c450ed23ee1786d37a"
                    },
                    "kedifyProxy": {
                      "globalValues": {
                        "openshift": true
                      },
                      "image": {
                        "tag": "v1.34.1@sha256:683e2f32809657c8d4a1508648545481130ac5409cd1ce63ccc795ad8c71202e"
                      }
                    },
                    "kubectlImage": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3"
                    },
                    "orgId": "00000000-0000-0000-0000-000000000000",
                    "securityContext": {
                      "readOnlyRootFilesystem": true
                    }
                  },
                  "keda": {
                    "enabled": true,
                    "image": {
                      "keda": {
                        "tag": "v2.17.2-0@sha256:802592b691a7b81a249055fa117994927c2fd376b5023a3888bcc7973c65ba52"
                      },
                      "metricsApiServer": {
                        "tag": "v2.17.2-0@sha256:df44bc1de75324158017d623a9d0adf3e3e9f1d64db6afa7133f777adc7b6389"
                      },
                      "pullPolicy": "IfNotPresent",
                      "webhooks": {
                        "tag": "v2.17.2-0@sha256:73e180bd5793dec3785fcb9ecd7e7184e645f7c37d561a5728970e5303c19a41"
                      }
                    }
                  },
                  "keda-add-ons-http": {
                    "enabled": true,
                    "interceptor": {
                      "pullPolicy": "IfNotPresent",
                      "scaledObject": {
                        "waitForCrd": true
                      }
                    },
                    "kubectlImage": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3"
                    },
                    "pinnedImages": {
                      "interceptor": {
                        "tag": "v0.10.0-17@sha256:30ccf20e79c6a62c14781872a73ad0a811063add7136b77df986bf403d6ca6bb"
                      },
                      "scaler": {
                        "tag": "v0.10.0-17@sha256:5863ee0ca09efeca225966c2eb8781f90f2dfb4aed872d84a6075b711ab042b2"
                      }
                    },
                    "scaler": {
                      "pullPolicy": "IfNotPresent"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: '{"kedify-openshift-default":{"description":"Default
            Kedify Agent CR for OpenShift. Register at https://dashboard.kedify.io
            to get started."}}'
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:d61332d50587ce2c75aff29033fdd5cb1bf0ec49d199ab72323a27d9d19ff402
          createdAt: "2025-08-07T14:04:11Z"
          description: Kedify Autoscaler, powered by proven open source technologies
            and KEDA, is an enterprise-grade service that allows teams to autoscale
            any workload using various event sources including HTTP events
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: keda
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: http://github.com/kedify/kedify-agent-helm-operator
          support: support@kedify.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KedifyAgent
            name: kedifyagents.charts.kedify.io
            version: v1alpha1
        description: Kedify Autoscaler, powered by proven open source technologies
          and KEDA, is an enterprise-grade service that allows teams to autoscale
          any workload using various event sources including HTTP events
        displayName: kedify-agent-helm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - autoscaling
        - keda
        - kedify
        links:
        - name: Kedify Homepage
          url: https://kedify.io
        maintainers:
        - email: support@kedify.io
          name: Kedify
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Kedify
          url: kedify.io
        relatedImages:
        - ghcr.io/kedify/agent:v0.2.16@sha256:9a57d5eafe368aab1d94f0eec3101e02f0ff2e356d8499984030ab89161e42ae
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:d61332d50587ce2c75aff29033fdd5cb1bf0ec49d199ab72323a27d9d19ff402
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:8e3a458912ec6540cb414446425ad7b1546528f04a5d84fcd64d8ee5424392e6
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:fc91d1f7f8b632614f31dea14c0c8d4a5f8311a83fdef82303857aa8b43b5782
        - envoyproxy/envoy:v1.34.1@sha256:007da57c2c328a90bd4e6d99b70bd899132f1b4a9426ccafe25437cf84a60c14
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:4b7277ad53dfb01bf5f2201dd31d94f24519da00d008638c3c36026b96ea78b4
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:9300a5382373b179f7f541a663cd30e50284e54f352993e8cca34460acbd890d
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:c1c093a6fda7efc9bc30f110c55f8bb075250db929dbdb2534b9071efbfc3169
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:30ccf20e79c6a62c14781872a73ad0a811063add7136b77df986bf403d6ca6bb
        - envoyproxy/envoy:v1.34.1@sha256:683e2f32809657c8d4a1508648545481130ac5409cd1ce63ccc795ad8c71202e
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:a38fdaf3a49eb4c71c4f3ab1211befa22fa5de56391f88efc11335f0bb3b1d29
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:82b041027dc900fdab717e8db6a522139067801c04cd1ebfe950572e714938ab
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:720b90a4f112a91a94b1bdafe445df1681657ddaa4fb20aa6d2904a0eb664a2a
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:5863ee0ca09efeca225966c2eb8781f90f2dfb4aed872d84a6075b711ab042b2
        - envoyproxy/envoy:v1.34.1@sha256:75545b91d050aa51b52cb0673be172f2f263c971dae3d2a82a70d128d92e87c2
        - ghcr.io/kedify/agent:v0.2.16@sha256:bfcc0b24ff75de17464ec213406f48bc211a932308e888c450ed23ee1786d37a
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:df44bc1de75324158017d623a9d0adf3e3e9f1d64db6afa7133f777adc7b6389
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:73e180bd5793dec3785fcb9ecd7e7184e645f7c37d561a5728970e5303c19a41
        - ghcr.io/kedify/agent:v0.2.16@sha256:c59d765091a31659236b8d80f4033912f5cf47d86c2e2da96bcfa133b9f6fe25
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:33cd5a8c6dd5e90614341dc1863e1f041f2c2475bd61edf5176a2a0d98ea54f9
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:850922994153104a8e382416561c251b46a95ff2b2a216aacbd7fcb4be7f6382
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:802592b691a7b81a249055fa117994927c2fd376b5023a3888bcc7973c65ba52
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:4b63409c022ba137760e7a534b6308fc1e3639f06fdb53ee1f6df32787b68400
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:ea6b22bb12c8ff7cbab4633c2f4a0e49eaf3883b07ab1f47dbebff2a3e07f029
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:9fcc76f983035eb0ceff7beb941bf32fe7fc28b341ec1aac5191ffc1471f401b
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:cb40290842976fd09f2008d8e2e64ebf90a215f0584656d4886122d322006873
        version: 0.2.16
      entries:
      - name: kedify-agent-helm-operator.v0.2.16
        version: 0.2.16
      name: stable
    defaultChannel: stable
    packageName: kedify-agent-helm-operator
    provider:
      name: Kedify
      url: kedify.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-operator.v2024.10.2-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "24.10.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "false",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": null
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": null
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:f4b9eb4f209f5428b0940056c760efcd9939553f3baac29924d1554701bc2dce
          createdAt: "2024-10-25T01:46:32Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2024.10.2
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2024.3.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "99.99.0.nextgen_uma_beta,3"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:f4b9eb4f209f5428b0940056c760efcd9939553f3baac29924d1554701bc2dce
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:7c4868db021beaffa762eb0de43739bbc865a4bf5900f1d17f78caed8612c073
        version: 2024.10.2-12
      entries:
      - name: uma-operator.v2024.10.2-12
        version: 2024.10.2-12
      name: "2024.10"
    - currentCSV: uma-operator.v2025.1.3-33
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.1.1.33,990033",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "false",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": null
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": null
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperDataQueueSize": 300,
                            "scrapperProcessingRoutines": 50,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:4a1bc124f22cf05dc6b42edbcc971d6309862e7be9c39b2ab6b1e3d50de61f3f
          createdAt: "2025-02-18T05:33:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.1.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2024.3.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       scrapperProcessingRoutines: 50
                       scrapperDataQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.1.1.26,990026"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:4a1bc124f22cf05dc6b42edbcc971d6309862e7be9c39b2ab6b1e3d50de61f3f
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:ae91a28d57c8a359f81757e4bf79472f2ac189ac0ae3ee0d87e8ff9c433f1ccc
        version: 2025.1.3-33
      entries:
      - name: uma-operator.v2025.1.3-33
        version: 2025.1.3-33
      name: "2025.1"
    - currentCSV: uma-operator.v2025.2.1-25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.2.1.25,990025",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mixedMode": "false",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperDataQueueSize": 300,
                            "scrapperProcessingRoutines": 50,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:d8f4ca750252f1570a2dd89cabf018cc4f0b2c0cd6cadfedbf48507a34d4aee9
          createdAt: "2025-03-05T05:49:42Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.2.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2024.3.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       scrapperProcessingRoutines: 50
                       scrapperDataQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.1.1.26,990026"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:d8f4ca750252f1570a2dd89cabf018cc4f0b2c0cd6cadfedbf48507a34d4aee9
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:8bcf0dac889f38a0bf77eb9babe3f726e3b2564d5988c05bdc5c2d91ae189c61
        version: 2025.2.1-25
      entries:
      - name: uma-operator.v2025.2.1-25
        version: 2025.2.1-25
      name: "2025.2"
    - currentCSV: uma-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:8ff4aab6d0043ccd2542a86b7e9c24d60670a3c376021ef2d02fc1a289130275
          createdAt: "2025-04-16T06:54:48Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.4.5
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.4.2.5,990005"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:8ff4aab6d0043ccd2542a86b7e9c24d60670a3c376021ef2d02fc1a289130275
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        version: 2025.4.5-5
      entries:
      - name: uma-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:90b4b21d93bce41f2530699cec01f03e5daa73bfbf4f30cbcc8fbd6e75e97da6
          createdAt: "2025-07-24T15:07:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.5.2.12,990012""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:90b4b21d93bce41f2530699cec01f03e5daa73bfbf4f30cbcc8fbd6e75e97da6
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        version: 2025.5.4-12
      entries:
      - name: uma-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-operator.v2025.7.4-23
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:35f74109aa1916524c1fb483972caa31cb84381c3cff7c3b26fd9b676f9d748c
          createdAt: "2025-08-25T09:00:19Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.3
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.7.1.23,990023"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:35f74109aa1916524c1fb483972caa31cb84381c3cff7c3b26fd9b676f9d748c
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.4-23
      entries:
      - name: uma-operator.v2025.7.4-23
        version: 2025.7.4-23
      name: "2025.7"
    - currentCSV: uma-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:adea186632253865406e95a0062abbb312dee23c92ff108f5914e325dc0b8833
          createdAt: "2025-08-15T03:19:46Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.8.1.9,990009"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:adea186632253865406e95a0062abbb312dee23c92ff108f5914e325dc0b8833
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        version: 2025.8.2-9
      entries:
      - name: uma-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
          createdAt: "2025-09-24T12:22:41Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.12,990012"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:c892787785c8c5e0de1692709a9e3f54eef36f1f00c53f1b9feabb613769a48c
          createdAt: "2025-07-30T08:06:11Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "99.99.0.serp_F152813,19"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:c892787785c8c5e0de1692709a9e3f54eef36f1f00c53f1b9feabb613769a48c
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        version: 2025.7.2-19
      entries:
      - name: uma-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:dc15dfdc8b82ee159b52036cc9d057e6525b08c478ed759939cf5c3fc7daeeec
          createdAt: "2025-09-11T11:15:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.8,990008"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:dc15dfdc8b82ee159b52036cc9d057e6525b08c478ed759939cf5c3fc7daeeec
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
          createdAt: "2025-09-24T12:22:41Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.12,990012"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: EDB
      provider-url: ""
    name: cloud-native-postgresql
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloud-native-postgresql.v1.27.0-rc1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:5f5bebbf4d8021dafd0fb7485d629b30aaa362e899e69eb52fa5e74757af6e36
          createdAt: "2025-07-30T09:28:05Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:5f5bebbf4d8021dafd0fb7485d629b30aaa362e899e69eb52fa5e74757af6e36
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        version: 1.27.0-rc1
      entries:
      - name: cloud-native-postgresql.v1.27.0-rc1
        version: 1.27.0-rc1
      - name: cloud-native-postgresql.v1.26.0-rc3
        version: 1.26.0-rc3
      - name: cloud-native-postgresql.v1.26.0-rc2
        version: 1.26.0-rc2
      - name: cloud-native-postgresql.v1.26.0-rc1
        version: 1.26.0-rc1
      name: candidate
    - currentCSV: cloud-native-postgresql.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
          createdAt: "2025-08-19T12:00:36Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
        version: 1.27.0
      entries:
      - name: cloud-native-postgresql.v1.27.0
        version: 1.27.0
      - name: cloud-native-postgresql.v1.26.1
        version: 1.26.1
      - name: cloud-native-postgresql.v1.26.0
        version: 1.26.0
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      - name: cloud-native-postgresql.v1.24.1
        version: 1.24.1
      - name: cloud-native-postgresql.v1.24.0
        version: 1.24.0
      - name: cloud-native-postgresql.v1.23.3
        version: 1.23.3
      - name: cloud-native-postgresql.v1.23.2
        version: 1.23.2
      - name: cloud-native-postgresql.v1.23.1
        version: 1.23.1
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      - name: cloud-native-postgresql.v1.21.1
        version: 1.21.1
      - name: cloud-native-postgresql.v1.21.0
        version: 1.21.0
      - name: cloud-native-postgresql.v1.20.2
        version: 1.20.2
      - name: cloud-native-postgresql.v1.20.1
        version: 1.20.1
      - name: cloud-native-postgresql.v1.20.0
        version: 1.20.0
      - name: cloud-native-postgresql.v1.19.1
        version: 1.19.1
      - name: cloud-native-postgresql.v1.19.0
        version: 1.19.0
      name: fast
    - currentCSV: cloud-native-postgresql.v1.25.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
          createdAt: "2025-07-26T06:26:53Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.25.3
      entries:
      - name: cloud-native-postgresql.v1.25.3
        version: 1.25.3
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      - name: cloud-native-postgresql.v1.22.9
        version: 1.22.9
      - name: cloud-native-postgresql.v1.22.8
        version: 1.22.8
      - name: cloud-native-postgresql.v1.22.7
        version: 1.22.7
      - name: cloud-native-postgresql.v1.22.6
        version: 1.22.6
      - name: cloud-native-postgresql.v1.22.5
        version: 1.22.5
      - name: cloud-native-postgresql.v1.22.4
        version: 1.22.4
      - name: cloud-native-postgresql.v1.22.3
        version: 1.22.3
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      - name: cloud-native-postgresql.v1.18.13
        version: 1.18.13
      - name: cloud-native-postgresql.v1.18.12
        version: 1.18.12
      - name: cloud-native-postgresql.v1.18.11
        version: 1.18.11
      - name: cloud-native-postgresql.v1.18.10
        version: 1.18.10
      - name: cloud-native-postgresql.v1.18.9
        version: 1.18.9
      - name: cloud-native-postgresql.v1.18.8
        version: 1.18.8
      - name: cloud-native-postgresql.v1.18.7
        version: 1.18.7
      name: stable
    - currentCSV: cloud-native-postgresql.v1.18.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:69e439a2a961881309cc122174eaca273cf6d7873e9e871efbdb2d60804f40e4
          createdAt: "2024-06-13T16:08:55Z"
          description: EDB Postgres for Kubernetes is an open source operator designed
            to manage highly available PostgreSQL databases with a primary/standby
            architecture on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 1.18.0 < 1.18.13'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloud
        - native
        - edb
        - enterprisedb
        - cnp
        - pg4k
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: john.long@enterprisedb.com
          name: John Long
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: The EDB Postgres for Kubernetes Contributors
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:69e439a2a961881309cc122174eaca273cf6d7873e9e871efbdb2d60804f40e4
        version: 1.18.13
      entries:
      - name: cloud-native-postgresql.v1.18.13
        version: 1.18.13
      - name: cloud-native-postgresql.v1.18.12
        version: 1.18.12
      - name: cloud-native-postgresql.v1.18.11
        version: 1.18.11
      - name: cloud-native-postgresql.v1.18.10
        version: 1.18.10
      - name: cloud-native-postgresql.v1.18.9
        version: 1.18.9
      - name: cloud-native-postgresql.v1.18.8
        version: 1.18.8
      - name: cloud-native-postgresql.v1.18.7
        version: 1.18.7
      name: stable-v1.18
    - currentCSV: cloud-native-postgresql.v1.19.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:2a717edd0523798bd4d0b14c557b8d6b19ae3ad1f019c0cc4f61db49f10c3dad
          createdAt: "2023-11-08T19:02:47Z"
          description: Operator to manage Postgres high availability clusters with
            a primary/standby architecture.
          olm.skipRange: '>=1.15.3 < 1.19.6'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup manager
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster manager
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PgBouncer Pooler Manager
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL backups schedule
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: "EDB Postgres for Kubernetes is an operator designed, developed,
          and supported by EDB that covers the full \nlifecycle of a highly available
          Postgres database clusters with a primary/standby architecture, using native\nstreaming
          replication. It is based on the open source CloudNativePG operator, and
          provides additional value\nsuch as compatibility with Oracle using EDB Postgres
          Advanced Server and additional supported platforms such\nas IBM Power and
          OpenShift.\n\nKey features available include:\n\n* Kubernetes API integration
          for high availability\n* Self-healing through failover and automated recreation
          of replicas\n* Capacity management with scale up/down capabilities\n* Planned
          switchovers for scheduled maintenance\n* Read-only and read-write Kubernetes
          services definitions\n* Rolling updates for Postgres minor versions and
          operator upgrades\n* Continuous backup and point-in-time recovery\n* Connection
          Pooling with PgBouncer\n* Integrated metrics exporter out of the box\n*
          PostgreSQL replication across multiple Kubernetes clusters\n* Red Hat certified
          operator for OpenShift\n\nThe operator has been renamed from Cloud Native
          PostgreSQL. Existing users of Cloud Native PostgreSQL will not\nexperience
          any change, as the underlying components and resources have not changed.\n"
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloud
        - native
        - edb
        - enterprisedb
        - cnp
        - pg4k
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: sales@enterprisedb.com
          name: Sales
        - email: support@enterprisedb.com
          name: Support
        maturity: stable
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:2a717edd0523798bd4d0b14c557b8d6b19ae3ad1f019c0cc4f61db49f10c3dad
        version: 1.19.6
      entries:
      - name: cloud-native-postgresql.v1.19.6
        version: 1.19.6
      - name: cloud-native-postgresql.v1.19.5
        version: 1.19.5
      - name: cloud-native-postgresql.v1.19.4
        version: 1.19.4
      - name: cloud-native-postgresql.v1.19.3
        version: 1.19.3
      - name: cloud-native-postgresql.v1.19.2
        version: 1.19.2
      - name: cloud-native-postgresql.v1.19.1
        version: 1.19.1
      - name: cloud-native-postgresql.v1.19.0
        version: 1.19.0
      name: stable-v1.19
    - currentCSV: cloud-native-postgresql.v1.20.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:919c6669f8bfa333b19fdd1ca03aaeceebe0c5ca74d91cb66b2369680422e916
          createdAt: "2024-02-03T17:26:23Z"
          description: Operator to manage Postgres high availability clusters with
            a primary/standby architecture.
          olm.skipRange: '>=1.15.3 < 1.20.6'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup manager
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster manager
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PgBouncer Pooler Manager
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL backups schedule
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: "EDB Postgres for Kubernetes is an operator designed, developed,
          and supported by EDB that covers the full \nlifecycle of a highly available
          Postgres database clusters with a primary/standby architecture, using native\nstreaming
          replication. It is based on the open source CloudNativePG operator, and
          provides additional value\nsuch as compatibility with Oracle using EDB Postgres
          Advanced Server and additional supported platforms such\nas IBM Power and
          OpenShift.\n\nKey features available include:\n\n* Kubernetes API integration
          for high availability\n* Self-healing through failover and automated recreation
          of replicas\n* Capacity management with scale up/down capabilities\n* Planned
          switchovers for scheduled maintenance\n* Read-only and read-write Kubernetes
          services definitions\n* Rolling updates for Postgres minor versions and
          operator upgrades\n* Continuous backup and point-in-time recovery\n* Connection
          Pooling with PgBouncer\n* Integrated metrics exporter out of the box\n*
          PostgreSQL replication across multiple Kubernetes clusters\n* Red Hat certified
          operator for OpenShift\n\nThe operator has been renamed from Cloud Native
          PostgreSQL. Existing users of Cloud Native PostgreSQL will not\nexperience
          any change, as the underlying components and resources have not changed.\n"
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloud
        - native
        - edb
        - enterprisedb
        - cnp
        - pg4k
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: sales@enterprisedb.com
          name: Sales
        - email: support@enterprisedb.com
          name: Support
        maturity: stable
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:919c6669f8bfa333b19fdd1ca03aaeceebe0c5ca74d91cb66b2369680422e916
        version: 1.20.6
      entries:
      - name: cloud-native-postgresql.v1.20.6
        version: 1.20.6
      - name: cloud-native-postgresql.v1.20.5
        version: 1.20.5
      - name: cloud-native-postgresql.v1.20.4
        version: 1.20.4
      - name: cloud-native-postgresql.v1.20.3
        version: 1.20.3
      - name: cloud-native-postgresql.v1.20.2
        version: 1.20.2
      - name: cloud-native-postgresql.v1.20.1
        version: 1.20.1
      - name: cloud-native-postgresql.v1.20.0
        version: 1.20.0
      name: stable-v1.20
    - currentCSV: cloud-native-postgresql.v1.21.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:8cc328b1fc8efa3911d3dba689bc824ff789b305f7eb3284bf21ffcdc52a9683
          createdAt: "2024-06-13T16:41:26Z"
          description: EDB Postgres for Kubernetes is an open source operator designed
            to manage highly available PostgreSQL databases with a primary/standby
            architecture on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 1.18.0 < 1.21.6'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: john.long@enterprisedb.com
          name: John Long
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: The EDB Postgres for Kubernetes Contributors
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:8cc328b1fc8efa3911d3dba689bc824ff789b305f7eb3284bf21ffcdc52a9683
        version: 1.21.6
      entries:
      - name: cloud-native-postgresql.v1.21.6
        version: 1.21.6
      - name: cloud-native-postgresql.v1.21.5
        version: 1.21.5
      - name: cloud-native-postgresql.v1.21.4
        version: 1.21.4
      - name: cloud-native-postgresql.v1.21.3
        version: 1.21.3
      - name: cloud-native-postgresql.v1.21.2
        version: 1.21.2
      - name: cloud-native-postgresql.v1.21.1
        version: 1.21.1
      - name: cloud-native-postgresql.v1.21.0
        version: 1.21.0
      name: stable-v1.21
    - currentCSV: cloud-native-postgresql.v1.22.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:4d55ad54cc7e7807bc99f5a6f96bc92e88b4060f424ffd2cb6552be9505f38f4
          createdAt: "2025-07-25T19:02:56Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images. Only
              available in 1.25 or later
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster.
              Only available in 1.25 or later
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster. Only available in 1.25 or later
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication.
              Only available in 1.25 or later
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:4d55ad54cc7e7807bc99f5a6f96bc92e88b4060f424ffd2cb6552be9505f38f4
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.22.11
      entries:
      - name: cloud-native-postgresql.v1.22.11
        version: 1.22.11
      - name: cloud-native-postgresql.v1.22.10
        version: 1.22.10
      - name: cloud-native-postgresql.v1.22.9
        version: 1.22.9
      - name: cloud-native-postgresql.v1.22.8
        version: 1.22.8
      - name: cloud-native-postgresql.v1.22.7
        version: 1.22.7
      - name: cloud-native-postgresql.v1.22.6
        version: 1.22.6
      - name: cloud-native-postgresql.v1.22.5
        version: 1.22.5
      - name: cloud-native-postgresql.v1.22.4
        version: 1.22.4
      - name: cloud-native-postgresql.v1.22.3
        version: 1.22.3
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      name: stable-v1.22
    - currentCSV: cloud-native-postgresql.v1.23.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:934e01ed5cd35cf681414b9136a337514a1b73c0110871f1c78ba98447badb0f
          createdAt: "2024-12-23T16:37:54Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 1.18.0 < 1.23.6'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:934e01ed5cd35cf681414b9136a337514a1b73c0110871f1c78ba98447badb0f
        version: 1.23.6
      entries:
      - name: cloud-native-postgresql.v1.23.6
        version: 1.23.6
      - name: cloud-native-postgresql.v1.23.5
        version: 1.23.5
      - name: cloud-native-postgresql.v1.23.4
        version: 1.23.4
      - name: cloud-native-postgresql.v1.23.3
        version: 1.23.3
      - name: cloud-native-postgresql.v1.23.2
        version: 1.23.2
      - name: cloud-native-postgresql.v1.23.1
        version: 1.23.1
      name: stable-v1.23
    - currentCSV: cloud-native-postgresql.v1.24.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:a1acaaf5a413db0f33f3824a15aefe598974d1593172bcbccd0fc3d39d7b46d0
          createdAt: "2025-05-23T15:20:37Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:a1acaaf5a413db0f33f3824a15aefe598974d1593172bcbccd0fc3d39d7b46d0
        version: 1.24.4
      entries:
      - name: cloud-native-postgresql.v1.24.4
        version: 1.24.4
      - name: cloud-native-postgresql.v1.24.3
        version: 1.24.3
      - name: cloud-native-postgresql.v1.24.2
        version: 1.24.2
      - name: cloud-native-postgresql.v1.24.1
        version: 1.24.1
      - name: cloud-native-postgresql.v1.24.0
        version: 1.24.0
      name: stable-v1.24
    - currentCSV: cloud-native-postgresql.v1.25.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
          createdAt: "2025-07-26T06:26:53Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
        version: 1.25.3
      entries:
      - name: cloud-native-postgresql.v1.25.3
        version: 1.25.3
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      name: stable-v1.25
    - currentCSV: cloud-native-postgresql.v1.26.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:47d70233908ea855ddab3ff79c7a226719c3f2be5b8a5018c7934cd44bf0f546
          createdAt: "2025-07-26T07:59:14Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:47d70233908ea855ddab3ff79c7a226719c3f2be5b8a5018c7934cd44bf0f546
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.26.1
      entries:
      - name: cloud-native-postgresql.v1.26.1
        version: 1.26.1
      - name: cloud-native-postgresql.v1.26.0
        version: 1.26.0
      name: stable-v1.26
    - currentCSV: cloud-native-postgresql.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
          createdAt: "2025-08-19T12:00:36Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.27.0
      entries:
      - name: cloud-native-postgresql.v1.27.0
        version: 1.27.0
      name: stable-v1.27
    defaultChannel: fast
    packageName: cloud-native-postgresql
    provider:
      name: EDB
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: http://marketplace.redhat.com
    name: marketplace-games-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marketplace-games-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "game.marketplace.redhat.com/v1",
                "kind": "Arcade",
                "metadata": {
                  "name": "arcade-sample"
                },
                "spec": {
                  "size": 1
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/marketplace-games-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/marketplace-games-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Arcade is the Schema for the arcades API
            displayName: Arcade Instance
            kind: Arcade
            name: arcades.game.marketplace.redhat.com
            version: v1
        description: RHM Arcade Operator - a simple operator used to showcase how
          to build an operator around an existing web application that can can run
          on Openshift, Vanilla k8s, or on-boarded to Red Hat Marketplace.
        displayName: Arcade Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - rhm
        - red hat
        - marketplace
        - games
        - arcade
        links:
        - name: Marketplace Games Operator
          url: https://github.com/redhat-marketplace/marketplace-games-operator
        maturity: alpha
        provider:
          name: Red Hat Marketplace
          url: http://marketplace.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/rhm-labs/arcade-operator@sha256:d6959be6693728d2dc4cd38b69251551d23c5002cb52db2684df31a0a8f7b0ab
        version: 1.0.0
      entries:
      - name: marketplace-games-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: marketplace-games-operator-rhmp
    provider:
      name: Red Hat Marketplace
      url: http://marketplace.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: application-services-metering-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: application-services-metering-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "applicationservices.redhat.com/v1",
                "kind": "Meter",
                "metadata": {
                  "name": "app-services-metrics-example"
                },
                "spec": {
                  "includeInfrastructure": false,
                  "meterCollectionEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Application Runtime
          certified: "false"
          containerImage: quay.io/redhat-developer/application-services-metering-operator:0.6.1
          createdAt: "2021-09-28T13:00:00Z"
          description: Collect the core usage of products from the Application Services
            portfolio into a single metric labeled by Runtimes, Integration, and Automation.
          repository: https://github.com/redhat-developer/application-services-metering-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Meter
            displayName: Meter
            kind: Meter
            name: meters.applicationservices.redhat.com
            version: v1
        description: |-
          ## About the managed application

          There is no managed application created with this operator.

          ## About this Operator

          The metering operator watches pods for predetermined labels, across a cluster or namespace(s).
          The label schema for Red Hat Application Services is in the format:

          * `com.company=Red_Hat`
          * `rht.comp=PAM`
          * `rht.comp_ver=7.10`
          * `rht.prod_name=Red_Hat_Process_Automation`
          * `rht.prod_ver=7.11`
          * `rht.subcomp=rhpam-kogito-runtime`
          * `rht.subcomp_t=application`

          It's important not to remove labels as the operator will cease monitoring those pods.

          The operator exposes a metric named: `appsvcs_cpu_usage_cores`
          The metric has a label for product category, these are Runtimes, Integration and Automation.
          There is a prometheus rule using the metric to group by product name, `appsvcs_cores_by_product:sum`.

          The operator utilises the OpenShift Monitoring Stack to provide usage observability via cloud.redhat.com

          ## Prerequisites for enabling this Operator

          None.
        displayName: Application Services Metering Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - metering
        - operator
        links:
        - name: Red Hat Application Services portfolio
          url: https://www.redhat.com/en/products/middleware
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/redhat-developer/application-services-metering-operator:0.6.1
        version: 0.6.1
      entries:
      - name: application-services-metering-operator.v0.6.1
        version: 0.6.1
      - name: application-services-metering-operator.v0.6.0
        version: 0.6.0
      name: alpha
    defaultChannel: alpha
    packageName: application-services-metering-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CloudBees
      provider-url: https://www.cloudbees.com/products/continuous-integration
    name: cloudbees-ci
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudbees-ci.v3.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.cloudbees.com/v1alpha1",
                "kind": "CloudBeesCI",
                "metadata": {
                  "name": "cloudbeesci-sample"
                },
                "spec": {
                  "Agents": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-agent:2.289.3.2"
                    },
                    "ImagePullSecrets": null,
                    "SeparateNamespace": {
                      "Create": false,
                      "Enabled": false,
                      "Name": null
                    }
                  },
                  "Hibernation": {
                    "Enabled": false,
                    "Image": {
                      "dockerImage": "cloudbees/managed-master-hibernation-monitor:247.c5dfce00a179",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "NodeSelector": {},
                    "Tolerations": []
                  },
                  "Master": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-mm:2.289.3.2"
                    },
                    "JavaOpts": null,
                    "OperationsCenterNamespace": null
                  },
                  "NetworkPolicy": {
                    "Enabled": false,
                    "JMXSelectors": [],
                    "ingressControllerSelector": []
                  },
                  "OperationsCenter": {
                    "AgentListenerPort": 50000,
                    "Annotations": {},
                    "CSRF": {
                      "ProxyCompatibility": false
                    },
                    "CasC": {
                      "ConfigMapName": "oc-casc-bundle",
                      "Enabled": false
                    },
                    "ContainerPort": 8080,
                    "ContextPath": null,
                    "Enabled": true,
                    "ExtraConfigMaps": [],
                    "ExtraContainers": [],
                    "ExtraGroovyConfiguration": {},
                    "ExtraVolumeMounts": [],
                    "ExtraVolumes": [],
                    "HealthProbeLivenessFailureThreshold": 12,
                    "HealthProbes": true,
                    "HostName": null,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-cloud-core-oc:2.289.3.2",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "Ingress": {
                      "Annotations": {
                        "kubernetes.io/tls-acme": "false"
                      },
                      "Class": "nginx",
                      "tls": {
                        "Enable": false,
                        "SecretName": null
                      }
                    },
                    "JavaOpts": null,
                    "JenkinsOpts": null,
                    "LoadBalancerIP": null,
                    "LoadBalancerSourceRanges": [
                      "0.0.0.0/0"
                    ],
                    "Name": "cjoc",
                    "NodeSelector": {},
                    "Platform": "standard",
                    "Protocol": "http",
                    "Resources": {
                      "Limits": {
                        "Cpu": 1,
                        "Memory": "2G"
                      },
                      "Requests": {
                        "Cpu": 1,
                        "Memory": "2G"
                      }
                    },
                    "Route": {
                      "tls": {
                        "CACertificate": null,
                        "Certificate": null,
                        "DestinationCACertificate": null,
                        "Enable": false,
                        "InsecureEdgeTerminationPolicy": "Redirect",
                        "Key": null,
                        "Termination": "edge"
                      }
                    },
                    "ServiceAgentListenerPort": 50000,
                    "ServiceAnnotations": {},
                    "ServicePort": 80,
                    "ServiceType": "ClusterIP",
                    "Tolerations": []
                  },
                  "Persistence": {
                    "AccessMode": "ReadWriteOnce",
                    "Annotations": {},
                    "Size": "20Gi",
                    "StorageClass": null
                  },
                  "PodSecurityPolicy": {
                    "Annotations": {},
                    "Enabled": false
                  },
                  "Subdomain": false,
                  "ingress-nginx": {
                    "Enabled": false,
                    "controller": {
                      "admissionWebhooks": {
                        "patch": {
                          "nodeSelector": {
                            "kubernetes.io/os": "linux"
                          }
                        }
                      },
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "nginx-ingress": {
                    "Enabled": false,
                    "controller": {
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "rbac": {
                    "agentsServiceAccountName": "jenkins-agents",
                    "hibernationMonitorServiceAccountName": "managed-master-hibernation-monitor",
                    "install": true,
                    "masterServiceAccountName": "jenkins",
                    "serviceAccountName": "cjoc"
                  },
                  "sidecarinjector": {
                    "Enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudBeesCI
            name: cloudbeescis.charts.cloudbees.com
            version: v1alpha1
        description: CloudBees CI
        displayName: CloudBees CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudbees
        - ci
        - enterprise
        - continuous
        - integration
        - jenkins
        - automation
        - pipeline
        - devops
        links:
        - name: CloudBees CI
          url: https://www.cloudbees.com/products/continuous-integration
        maintainers:
        - email: support@cloudbees.com
          name: Support
        - email: sales@cloudbees.com
          name: Sales
        maturity: alpha
        provider:
          name: CloudBees
          url: https://www.cloudbees.com/products/continuous-integration
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.connect.redhat.com/cloudbees/cloudbees-ci-operator@sha256:bd0bd72e9e7807b0fa307de8a93a56967221fe2dd181bac02383f0f9be9120e4
        version: 3.34.1
      entries:
      - name: cloudbees-ci.v3.34.1
        version: 3.34.1
      name: alpha
    defaultChannel: alpha
    packageName: cloudbees-ci
    provider:
      name: CloudBees
      url: https://www.cloudbees.com/products/continuous-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: https://github.com/Mellanox/network-operator/
    name: nvidia-network-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nvidia-network-operator.v25.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "doca3.1.0-25.07-0.9.7.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "nvcr.io/nvidia/mellanox",
                    "version": "sha256:a87096761d155eeb6f470e042d2d167bb466d57e63b4aba957f57d745e15a9b2"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:51270965284cceeda659fb1098e70a888b39664ceade7e9d455fb95ac85914b7
          createdAt: "2025-08-21T19:12:20Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/mellanox/nvidia-k8s-ipam@sha256:3f692273c1598525b9c698c9d9cd463e5ddc55b638c6c031639f38786fc67475
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:9a831bfdf85f313b1f5749b7c9b2673bb8fff18b4ff768c9242dabaa4468e449
        - nvcr.io/nvidia/mellanox/k8s-rdma-shared-dev-plugin@sha256:a87096761d155eeb6f470e042d2d167bb466d57e63b4aba957f57d745e15a9b2
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:5ad37cd5ae96153e6f87670f437ff5f6e44febe7bb6eaeb1a0f219e28284e53d
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:d41eea65b102dc5bb5792871ae212b59a65dae492ca109c52138d5626eba1ffd
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:13c2a2d055e501c03c6451413b85476ceafbc2fbffc7443952294c196f3f89f3
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:30354799264480e1aee057397eef2c52b277bbc6872184e2b7b811f4152b3a67
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:423aea08b4a0f01e2e19b6a709e14347bd08f8274f40002e0224090dfa4904fc
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:f9d4b8358c68ba28f849ca710bf769c4327a639733c0ff30282ca33a744078f6
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:51270965284cceeda659fb1098e70a888b39664ceade7e9d455fb95ac85914b7
        - nvcr.io/nvidia/mellanox/nic-feature-discovery@sha256:1330e3a7ea2491bc310bfc4766769389ceb56d050831bf9c145b7bfb6aba36c4
        - nvcr.io/nvidia/mellanox/network-operator-init-container@sha256:02ee60ac6ee163acc97706bd4229d268f608a95bcbcb6fccee90ceb1d8a6c62b
        - nvcr.io/nvidia/mellanox/ipoib-cni@sha256:7564ad1142c863e61f1dd84f5714deba9a7c6f6d85efb383ac3a2566eb4f4b71
        - nvcr.io/nvidia/mellanox/sriov-network-device-plugin@sha256:7508a160f3f124e0b78d2b4d5a52deca6b9f56e226fee957c4d156e425eb101a
        - nvcr.io/nvidia/mellanox/nic-configuration-operator-daemon@sha256:b78c15904cecd343fa48374757b3619b5cbdc22059d5ee9731da87e35b94a2e2
        - nvcr.io/nvidia/mellanox/nic-configuration-operator@sha256:03f34f7fde3f65ca7b214b66c6fdbc1250d4f3863f9c02b0fba78e8aa84efca3
        - nvcr.io/nvidia/mellanox/ib-kubernetes@sha256:1fa251979b5f5201af50bc7aa6936829103d742f42bfbf752f20852f5fe44dfb
        version: 25.7.0
      entries:
      - name: nvidia-network-operator.v25.7.0
        version: 25.7.0
      - name: nvidia-network-operator.v25.4.0
        version: 25.4.0
      name: stable
    - currentCSV: nvidia-network-operator.v24.10.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "24.10-0.7.0.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "v1.5.2"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:23aeb481d662496210d1c13083c65cfa4ffdcf67f973e5f884b1b936f693c0f5
          createdAt: "2025-01-27T08:16:45Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - ghcr.io/mellanox/ib-kubernetes@sha256:100e6addf99d5b44e81cb3aa8d2fc892fced1de3183190931961b92ee924207b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:e36c46cf2c813c081bf7cc22b47fb4cf391a3d91804e47a6ce87f31d62f42535
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:4bb13a8b850c91011013509334d5ae4f4b7aa1d1b0c12958625e161cd50309d6
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:9f468fdc4449e65e4772575f83aa85840a00f97165f9a00ba34695c91d610fbd
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:5a87d04872d5d4530cab3cbb3db528722fa24435f3375debd8ed102b04e631c2
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:23aeb481d662496210d1c13083c65cfa4ffdcf67f973e5f884b1b936f693c0f5
        - ghcr.io/mellanox/ipoib-cni@sha256:81646c07e27ce1df2603050a75dd267f264685fdc64c285eca3925ae98ed5392
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:77e3b946c20f45728dfa6dd1d115356de5ea317f91cddfd1540b7121550f66dc
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:47e1bb84ac97f9af49f97b49c2de44b4ba82b890b487880edc5c9adf86a176f7
        - ghcr.io/mellanox/network-operator-init-container@sha256:1699d23027ea30c9fa59575a914114bdfd5a87a359caf8c0a9b16d409ec0d068
        - ghcr.io/mellanox/nic-feature-discovery@sha256:92dda9434519de39be0f94c80787766e3399e8b238d14839360aff4d6232a09b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:db5797121735eeb935d8f11e305b50c80ba07aa83c291d76d05dea7e303bc740
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:ecb2cfdc510b7ed09a59daed4817ee353e83a94d2d8581c3c052f4fa4b990574
        version: 24.10.1
      entries:
      - name: nvidia-network-operator.v24.10.1
        version: 24.10.1
      name: v24.10
    - currentCSV: nvidia-network-operator.v24.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "24.10-0.7.0.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "v1.5.2"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:f258e8665d094c09159654759bc9ae41fbe0f7fffe95139ec02964759c3d8a87
          createdAt: "2024-11-26T09:58:46Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:5a87d04872d5d4530cab3cbb3db528722fa24435f3375debd8ed102b04e631c2
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:ecb2cfdc510b7ed09a59daed4817ee353e83a94d2d8581c3c052f4fa4b990574
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:9f468fdc4449e65e4772575f83aa85840a00f97165f9a00ba34695c91d610fbd
        - ghcr.io/mellanox/nic-feature-discovery@sha256:92dda9434519de39be0f94c80787766e3399e8b238d14839360aff4d6232a09b
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:f258e8665d094c09159654759bc9ae41fbe0f7fffe95139ec02964759c3d8a87
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:db5797121735eeb935d8f11e305b50c80ba07aa83c291d76d05dea7e303bc740
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:e36c46cf2c813c081bf7cc22b47fb4cf391a3d91804e47a6ce87f31d62f42535
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:4bb13a8b850c91011013509334d5ae4f4b7aa1d1b0c12958625e161cd50309d6
        - ghcr.io/mellanox/ib-kubernetes@sha256:100e6addf99d5b44e81cb3aa8d2fc892fced1de3183190931961b92ee924207b
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:77e3b946c20f45728dfa6dd1d115356de5ea317f91cddfd1540b7121550f66dc
        - ghcr.io/mellanox/ipoib-cni@sha256:81646c07e27ce1df2603050a75dd267f264685fdc64c285eca3925ae98ed5392
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:47e1bb84ac97f9af49f97b49c2de44b4ba82b890b487880edc5c9adf86a176f7
        - ghcr.io/mellanox/network-operator-init-container@sha256:1699d23027ea30c9fa59575a914114bdfd5a87a359caf8c0a9b16d409ec0d068
        version: 24.10.0
      entries:
      - name: nvidia-network-operator.v24.10.0
        version: 24.10.0
      name: v24.10.0
    - currentCSV: nvidia-network-operator.v24.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "docaTelemetryService": {
                    "image": "doca_telemetry",
                    "repository": "nvcr.io/nvidia/doca",
                    "version": "1.16.5-doca2.6.0-host"
                  },
                  "nicFeatureDiscovery": {
                    "image": "nic-feature-discovery",
                    "repository": "ghcr.io/mellanox",
                    "version": "v0.0.1"
                  },
                  "nvIpam": {
                    "enableWebhook": false,
                    "image": "nvidia-k8s-ipam",
                    "repository": "ghcr.io/mellanox",
                    "version": "v0.2.0"
                  },
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "24.07-0.6.1.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "v1.5.1"
                  },
                  "secondaryNetwork": {
                    "ipoib": {
                      "image": "ipoib-cni",
                      "repository": "ghcr.io/mellanox",
                      "version": "v1.2.0"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:fcee44995d64058cb7bf14c02da604879443e191a382402b8a6ebfe10a3a2601
          createdAt: "2024-08-29T13:23:34Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - ghcr.io/mellanox/network-operator-init-container@sha256:1699d23027ea30c9fa59575a914114bdfd5a87a359caf8c0a9b16d409ec0d068
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:93e856dde88abb0aeb8e0b991e5f6d9f41884d8beea792f4d70cda06f18e783f
        - ghcr.io/mellanox/ib-kubernetes@sha256:b56bb88963dadaa2290a17cf0b57b86c55c09e7c075b7271c184959532555c91
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:c41eaf76c06186e1dabb0d870225d946ad3ecda2a072fa4be197dd131b7695f2
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:8324c009c3d694fe7702af36305ede78b903d839190b1e4506e82362f9aa5c02
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:fcee44995d64058cb7bf14c02da604879443e191a382402b8a6ebfe10a3a2601
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:47e1bb84ac97f9af49f97b49c2de44b4ba82b890b487880edc5c9adf86a176f7
        - ghcr.io/mellanox/nic-feature-discovery@sha256:92dda9434519de39be0f94c80787766e3399e8b238d14839360aff4d6232a09b
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - ghcr.io/mellanox/ipoib-cni@sha256:81646c07e27ce1df2603050a75dd267f264685fdc64c285eca3925ae98ed5392
        version: 24.7.0
      entries:
      - name: nvidia-network-operator.v24.7.0
        version: 24.7.0
      name: v24.7.0
    - currentCSV: nvidia-network-operator.v24.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "docaTelemetryService": {
                    "image": "doca_telemetry",
                    "repository": "nvcr.io/nvidia/doca",
                    "version": "1.16.5-doca2.6.0-host"
                  },
                  "nicFeatureDiscovery": {
                    "image": "nic-feature-discovery",
                    "repository": "ghcr.io/mellanox",
                    "version": "v0.0.1"
                  },
                  "nvIpam": {
                    "enableWebhook": false,
                    "image": "nvidia-k8s-ipam",
                    "repository": "ghcr.io/mellanox",
                    "version": "v0.2.0"
                  },
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "24.07-0.6.1.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "v1.5.1"
                  },
                  "secondaryNetwork": {
                    "ipoib": {
                      "image": "ipoib-cni",
                      "repository": "ghcr.io/mellanox",
                      "version": "v1.2.0"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:b8a07696d05fb8e5292991ba731758579313d6983ed3d75dd95e7356024d987c
          createdAt: "2024-09-04T06:38:33Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:d5a163d36ce60630e14ae8620ab437510ff0821924e043b88f0d7b67cee3bd77
        - ghcr.io/mellanox/network-operator-init-container@sha256:1699d23027ea30c9fa59575a914114bdfd5a87a359caf8c0a9b16d409ec0d068
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:403be1bcf708984f90b3ccfdf24f58594554336d69f878d3843535f4fa61ceca
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:7facd05f60e2be29925ffce9d0a6b2c712c0670c8889b815425582c46c82efdc
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:8324c009c3d694fe7702af36305ede78b903d839190b1e4506e82362f9aa5c02
        - ghcr.io/mellanox/ib-kubernetes@sha256:b56bb88963dadaa2290a17cf0b57b86c55c09e7c075b7271c184959532555c91
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:10684c0c36364b1d897a03a346abb2f691a18541ae1242b596d002f34f0d7390
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:7f87324adb351513d87fa27dfd0c9d9371ae42529d76cce8ecd68ce41c4d263f
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:676c74fcbe6d2c2341a092913b1afca65ba1a2f568ee6796b371542657e7a174
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:22e6ff1e9245f0cdf67e7904235d1a6181ee2bd3f39cedf76b39d4aeaa4a6c1d
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:b3f503b1ac39ebc55840a436973f142eb7cf67e48a070284129fd1e30390c032
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:93e856dde88abb0aeb8e0b991e5f6d9f41884d8beea792f4d70cda06f18e783f
        - ghcr.io/mellanox/ipoib-cni@sha256:81646c07e27ce1df2603050a75dd267f264685fdc64c285eca3925ae98ed5392
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:47e1bb84ac97f9af49f97b49c2de44b4ba82b890b487880edc5c9adf86a176f7
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:b8a07696d05fb8e5292991ba731758579313d6983ed3d75dd95e7356024d987c
        - ghcr.io/mellanox/nic-feature-discovery@sha256:92dda9434519de39be0f94c80787766e3399e8b238d14839360aff4d6232a09b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:c41eaf76c06186e1dabb0d870225d946ad3ecda2a072fa4be197dd131b7695f2
        version: 24.7.1
      entries:
      - name: nvidia-network-operator.v24.7.1
        version: 24.7.1
      name: v24.7.1
    - currentCSV: nvidia-network-operator.v25.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "25.01-0.6.0.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "sha256:9f468fdc4449e65e4772575f83aa85840a00f97165f9a00ba34695c91d610fbd"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:12f0576f372f5adc17b60795952ae47388e50d8fad563b9595f3091eadd8eb29
          createdAt: "2025-02-26T16:55:35Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:f5c1741cc6368fcec5b250285552d9c80ff7254414f68102cc89162e12ac0400
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:cabce074d10a0f1d62135e2cc5442d65b49094b95b8297fdd024a1a5f461319f
        - ghcr.io/mellanox/nic-feature-discovery@sha256:92dda9434519de39be0f94c80787766e3399e8b238d14839360aff4d6232a09b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:5b91974c74f08fc5b1ad2edd4ca97d81b2dc6cc897045c3eac777826fe5d40d8
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:3edf0cec91eea10df17c9f182b2db430cb383887ac144e27ac44043b308284d7
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:68b3735c4edca2d889c456beba80ac2077b0cc0edba11a32769dacd3ece64892
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:a257aae9c4e02b62f9f4ddeea66c4349c517519a544dd02fe4c8d3fee5229521
        - ghcr.io/mellanox/ib-kubernetes@sha256:100e6addf99d5b44e81cb3aa8d2fc892fced1de3183190931961b92ee924207b
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:47e1bb84ac97f9af49f97b49c2de44b4ba82b890b487880edc5c9adf86a176f7
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:7604a30cbd0927eba76ddc74ea1cfbe113563ccf068e6ab3fe78123e4deec28d
        - ghcr.io/mellanox/ipoib-cni@sha256:56f19687b0b956ba45bf079cf3a655b0d1bad0314da9d5ecfe0f4d98ff31fc86
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:12f0576f372f5adc17b60795952ae47388e50d8fad563b9595f3091eadd8eb29
        - ghcr.io/mellanox/network-operator-init-container@sha256:67e93ccf3ecb61f17597567faf0f72e1b8ddcf73c5d7440baeadcc1cb6bb811b
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:9f468fdc4449e65e4772575f83aa85840a00f97165f9a00ba34695c91d610fbd
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        version: 25.1.0
      entries:
      - name: nvidia-network-operator.v25.1.0
        version: 25.1.0
      name: v25.1
    - currentCSV: nvidia-network-operator.v25.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "25.04-0.6.1.0-2"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "sha256:1eb5ab7f4b6f0cf89c954cf7ac8eb2e6fb2290a4f5f57417136a5302f20b12c8"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:93c2063f4da6a93810784be95df0083f07618486699d93245c97a29a63861dee
          createdAt: "2025-05-28T19:40:27Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:1b20b78f889339834ed74e0da621fc5da582719b2537b36d8967ddc6a04679b8
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:56a41010b4dbee0dc31e8f9c1467a739241a2d248ed8b5054d4b418f7f79f04d
        - ghcr.io/mellanox/nic-configuration-operator-daemon@sha256:0e8478898e0b30aec13030b11081b205308666c2a9c5c5bf2c00121079901b41
        - ghcr.io/mellanox/ipoib-cni@sha256:513d88d639a01ef98b65365f711cee7fbaa166137a56245b7ab32818f745418e
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:9bc09051aca763f83cde99ee8909cf989b1d6c5d6ab683fbb1ca441d47f61fdf
        - ghcr.io/mellanox/network-operator-init-container@sha256:67e93ccf3ecb61f17597567faf0f72e1b8ddcf73c5d7440baeadcc1cb6bb811b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:4a620bd39bc70928cd6e249855176ced0f4feb585c8c7624c0b9f5a3e57664c9
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:69e6ab6a0249be918f60c1e04fc53a07f280b33563f8edc953806538a9cc3c30
        - ghcr.io/mellanox/nic-configuration-operator@sha256:353ab27f5f5fba08197cbe8968222e1c115431cc1b9b8d8445f1facb3c3c4d7b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:35ff3b173b2694198e57b61720545be67e944c1da18ac004007a5b4e68562adc
        - ghcr.io/mellanox/ib-kubernetes@sha256:9774f774df6ede9ddf780b9193ca9d919618056e682d985e6b41e477b13c3432
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:cabce074d10a0f1d62135e2cc5442d65b49094b95b8297fdd024a1a5f461319f
        - ghcr.io/mellanox/nic-feature-discovery@sha256:35b1820fe94a833a43613ef5c9db1f58b293ea0dbcf96ec0742b0987f23f133c
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:76529c24ea17e68b89a2c0d4823d66504fa6f53759760a449308fda396574097
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:93c2063f4da6a93810784be95df0083f07618486699d93245c97a29a63861dee
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:1eb5ab7f4b6f0cf89c954cf7ac8eb2e6fb2290a4f5f57417136a5302f20b12c8
        version: 25.4.0
      entries:
      - name: nvidia-network-operator.v25.4.0
        version: 25.4.0
      name: v25.4
    defaultChannel: stable
    packageName: nvidia-network-operator
    provider:
      name: NVIDIA
      url: https://github.com/Mellanox/network-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: node-observability-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-observability-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservability",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "labels": {
                    "kubernetes.io/hostname": "worker-a"
                  },
                  "type": "crio-kubelet"
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservabilityMachineConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "debug": {
                    "enableCrioProfiling": true
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservabilityRun",
                "metadata": {
                  "name": "nodeobservabilityrun-sample",
                  "namespace": "node-observability-operator"
                },
                "spec": {
                  "nodeObservabilityRef": {
                    "name": "cluster"
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservability",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "nodeSelector": {
                    "kubernetes.io/hostname": "worker-a"
                  },
                  "type": "crio-kubelet"
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservabilityMachineConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "debug": {
                    "enableCrioProfiling": true
                  },
                  "nodeSelector": {
                    "kubernetes.io/hostname": "worker-a"
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservabilityRun",
                "metadata": {
                  "name": "nodeobservabilityrun-sample",
                  "namespace": "node-observability-operator"
                },
                "spec": {
                  "nodeObservabilityRef": {
                    "name": "cluster"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/noo/node-observability-rhel8-operator@sha256:53ba77f648154ac8aca1ae6c83815078dfde211dba1e42f597e69844e4954c01
          createdAt: 2022-12-05T09:25:09
          olm.skipRange: <0.2.0
          operatorframework.io/suggested-namespace: node-observability-operator
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeObservability prepares a subset of worker nodes (identified
              by a label) for running node observability queries, such as profiling
            displayName: Node Observability
            kind: NodeObservability
            name: nodeobservabilities.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservability prepares a subset of worker nodes (identified
              by a label) for running node observability queries, such as profiling
            displayName: Node Observability
            kind: NodeObservability
            name: nodeobservabilities.nodeobservability.olm.openshift.io
            version: v1alpha2
          - description: NodeObservabilityMachineConfig is the Schema for the nodeobservabilitymachineconfigs
              API
            displayName: Node Observability Machine Config
            kind: NodeObservabilityMachineConfig
            name: nodeobservabilitymachineconfigs.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservabilityMachineConfig is the Schema for the nodeobservabilitymachineconfigs
              API
            displayName: Node Observability Machine Config
            kind: NodeObservabilityMachineConfig
            name: nodeobservabilitymachineconfigs.nodeobservability.olm.openshift.io
            version: v1alpha2
          - description: NodeObservabilityRun is a request to run observability actions
              on the nodes previously selected in NodeObservability resource
            displayName: Node Observability Run
            kind: NodeObservabilityRun
            name: nodeobservabilityruns.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservabilityRun is a request to run observability actions
              on the nodes previously selected in NodeObservability resource
            displayName: Node Observability Run
            kind: NodeObservabilityRun
            name: nodeobservabilityruns.nodeobservability.olm.openshift.io
            version: v1alpha2
        description: |-
          An Operator that will be able to gather debugging/profiling data over
          a custom period of time which would be helpful to troubleshoot and resolve issues
          for OpenShift customers.
        displayName: Node Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node-observability-operator
        links:
        - name: Source Code
          url: https://github.com/openshift/node-observability-operator
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/noo/node-observability-rhel8-operator@sha256:53ba77f648154ac8aca1ae6c83815078dfde211dba1e42f597e69844e4954c01
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/noo/node-observability-agent-rhel8@sha256:a8e60d9f63e2ed0cf1e27309aac54014e3f1efc947b51ca0a99376338d220875
        version: 0.2.0
      entries:
      - name: node-observability-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: node-observability-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/
    name: varnish-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: varnish-operator.v0.27.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "caching.ibm.com/v1alpha1",
                "kind": "VarnishCluster",
                "metadata": {
                  "name": "varnishcluster-sample"
                },
                "spec": {
                  "backend": {
                    "port": "web",
                    "selector": {
                      "app": "nginx"
                    }
                  },
                  "replicas": 1,
                  "service": {
                    "port": 80
                  },
                  "varnish": {
                    "args": [
                      "-p",
                      "default_ttl=3600",
                      "-p",
                      "default_grace=3600"
                    ]
                  },
                  "vcl": {
                    "configMapName": "vcl-files",
                    "entrypointFileName": "entrypoint.vcl"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: ibmcom/varnish-operator:0.27.2
          createdAt: 2021-02-15T11:15:31+0000
          description: Run and manage Varnish clusters on Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/varnish-operator
          support: https://ibm.github.io/varnish-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VarnishCluster is the Schema for the varnishclusters API
            displayName: Varnish Cluster
            kind: VarnishCluster
            name: varnishclusters.caching.ibm.com
            version: v1alpha1
        description: |
          Run and manage Varnish clusters on Kubernetes

          The operator manages the whole lifecycle of the cluster: creating, deleting and keeping the cluster configuration up to date.

          The operator is responsible for building the VCL configuration using templates defined by the users and keeping the configuration up to date when relevant events occur (backend pod failure, scaling of the deployment, VCL configuration change).

          Example of a simple `VarnishCluster`:

          ```yaml
          apiVersion: caching.ibm.com/v1alpha1
          kind: VarnishCluster
          metadata:
            name: varnishcluster-sample
            namespace: varnish-ns
          spec:
            vcl:
              configMapName: vcl-files
              entrypointFileName: entrypoint.vcl
            replicas: 3
            backend:
              selector:
                app: nginx
            service:
              port: 80
          ```

          Full documentation can be found [here](https://ibm.github.io/varnish-operator/)
        displayName: Varnish Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - varnish
        - cache
        links:
        - name: Varnish Operator Docs
          url: https://ibm.github.io/varnish-operator
        - name: Varnish Operator Source Code
          url: https://github.com/IBM/varnish-operator
        maintainers:
        - email: tomash.sidei@ibm.com
          name: Tomash Sidei
        - email: cingram@us.ibm.com
          name: Craig Ingram
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/
        relatedImages:
        - ibmcom/varnish-operator:0.27.2
        version: 0.27.2
      entries:
      - name: varnish-operator.v0.27.2
        version: 0.27.2
      name: alpha
    defaultChannel: alpha
    packageName: varnish-operator
    provider:
      name: IBM
      url: https://www.ibm.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Odigos
      provider-url: https://odigos.io
    name: odigos-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odigos-operator.v1.0.212
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.odigos.io/v1alpha1",
                "kind": "Odigos",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "odigos-operator"
                  },
                  "name": "odigos-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing
          containerImage: registry.connect.redhat.com/odigos/odigos-certified-operator-ubi9@sha256:c221eb2be22c2fcabcee728429bbca1255e0e31642c572e4dc40096d1d35e0d1
          createdAt: "2025-07-18T13:23:10Z"
          description: Odigos enables automatic distributed tracing with OpenTelemetry
            and eBPF.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Odigos Enterprise subscription
            (for enterprise features)
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Odigos
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Odigos is the Schema for the odigos API
            displayName: Odigos
            kind: Odigos
            name: odigos.operator.odigos.io
            version: v1alpha1
        description: |-
          The Odigos Operator provides options for installing and configuring Odigos.

          Upon creation of an Odigos resource, the Odigos Operator installs the following Odigos components:

          - Instrumentor
          - Odiglet
          - Autoscaler
          - Scheduler
          - Frontend UI

          With Odigos installed, follow the Odigos docs at docs.odigos.io for instructions on instrumenting applications and configuring destinations.
        displayName: Odigos Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - OpenTelemetry
        - eBPF
        - tracing
        - observability
        - distributed tracing
        - otel
        - monitoring
        links:
        - name: Odigos
          url: https://odigos.io
        - name: Odigos Documentation
          url: https://docs.odigos.io
        - name: Odigos on Github
          url: https://github.com/odigos-io/odigos
        maintainers:
        - email: mike@odigos.io
          name: Mike Dame
        maturity: alpha
        minKubeVersion: 1.20.15
        provider:
          name: Odigos
          url: https://odigos.io
        relatedImages:
        - registry.connect.redhat.com/odigos/odigos-certified-operator-ubi9@sha256:c221eb2be22c2fcabcee728429bbca1255e0e31642c572e4dc40096d1d35e0d1
        - registry.connect.redhat.com/odigos/odigos-ui-ubi9@sha256:b2a27cdfc94f1968a06ae18de1977127bb7d4e7d553e875d09a212693302f9c6
        - registry.connect.redhat.com/odigos/odigos-enterprise-instrumentor-ubi9@sha256:647e526cdbe38023bb221a802b180f0b95cb25734a0b39fa11cbbe22a07482a3
        - registry.connect.redhat.com/odigos/odigos-scheduler-ubi9@sha256:9e613c9b22d2fbfa3d21cc24f3b1e2dbc2e7287790e1f4286f1817c92565c865
        - registry.connect.redhat.com/odigos/odigos-instrumentor-ubi9@sha256:16712373c8e927b7208e43e9fe6e21180289975dbf3a81df60578e8550cf7150
        - registry.connect.redhat.com/odigos/odigos-odiglet-ubi9@sha256:770193901b4e5abcec59833f56f1530b43d4ec9514f3b5a5cb0bbb4356e06c67
        - registry.connect.redhat.com/odigos/odigos-enterprise-odiglet-ubi9@sha256:aec9b66461375994dc9aa92e5b2cc7d4151cf0332150726840cec049a7cbe7d9
        - registry.connect.redhat.com/odigos/odigos-autoscaler-ubi9@sha256:692193851b40f72ca79f7c82903136686d721c7fed77ef599037269ac6e55fdb
        - registry.connect.redhat.com/odigos/odigos-collector-ubi9@sha256:b67138d3ffd9e721602c5bf83f0c03ae1dcee230eca5e15335dcb9abced281ba
        version: 1.0.212
      entries:
      - name: odigos-operator.v1.0.212
        version: 1.0.212
      name: alpha
    defaultChannel: alpha
    packageName: odigos-operator
    provider:
      name: Odigos
      url: https://odigos.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-superset-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: superset-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:e0c9be4ac01a3d02f908f71f47f96bb1423fee14ba454c1bc627b35654d9cae3
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:e0c9be4ac01a3d02f908f71f47f96bb1423fee14ba454c1bc627b35654d9cae3
        version: 24.11.1
      entries:
      - name: superset-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: superset-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:f12cc3ed35135c14a82d49fcafed55465f0e66d46c2a36d532e5e09b9dece31e
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:f12cc3ed35135c14a82d49fcafed55465f0e66d46c2a36d532e5e09b9dece31e
        version: 25.3.0
      entries:
      - name: superset-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: superset-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
        version: 25.7.0
      entries:
      - name: superset-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: superset-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
        version: 25.7.0
      entries:
      - name: superset-operator.v25.7.0
        version: 25.7.0
      - name: superset-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-superset-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: LINBIT
      provider-url: ""
    name: linstor-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: linstor-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorController",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerImage": "",
                  "dbConnectionURL": "etcd://linstor-etcd:2379",
                  "drbdRepoCred": "",
                  "priorityClassName": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorCSIDriver",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerReplicas": 1,
                  "controllerEndpoint": "http://linstor:3370",
                  "csiControllerServiceAccountName": "csi-controller",
                  "csiNodeServiceAccountName": "csi-node",
                  "linstorPluginImage": "",
                  "imagePullSecret": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorSatelliteSet",
                "metadata": {
                  "name": "linstor-satellites"
                },
                "spec": {
                  "automaticStorageType": "None",
                  "controllerEndpoint": "http://linstor:3370",
                  "drbdRepoCred": "",
                  "kernelModuleInjectionMode": "ShippedModules",
                  "priorityClassName": "",
                  "satelliteImage": ""
                },
                "status": {
                  "errors": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          createdAt: "2021-06-22T07:57:53+00:00"
          description: LINSTOR Kubernetes Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/piraeusdatastore/piraeus-operator
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LinstorController is the Schema for the linstorcontrollersets
              API
            kind: LinstorController
            name: linstorcontrollers.linstor.linbit.com
            version: v1
          - description: LinstorCSIDriver is the Schema for the linstorcsidrivers
              API
            kind: LinstorCSIDriver
            name: linstorcsidrivers.linstor.linbit.com
            version: v1
          - description: LinstorSatelliteSet is the Schema for the linstornodesets
              API
            kind: LinstorSatelliteSet
            name: linstorsatellitesets.linstor.linbit.com
            version: v1
        description: |-
          LINSTOR is a configuration management system for storage on Linux systems.
          It manages LVM logical volumes and/or ZFS ZVOLs on a cluster of nodes.
          It leverages DRBD for replication between different nodes and to provide block
          storage devices to users and applications. It manages snapshots, encryption and
          caching of HDD backed data in SSDs via bcache.

          LINBIT provides a certified LINSTOR operator to ease deployment of LINSTOR
          on Openshift by installing DRBD, managing Satellite and Controller pods,
          and integrating with Openshift to provision persistent storage for your workloads.

          For detailed instructions and more configuration options, see our [user guide].

          [user guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift

          ## Install

          Unlike deployment via the helm chart, the certified Openshift
          operator does not deploy the needed etcd cluster. You must deploy this
          yourself ahead of time. We do this via the etcd operator available in the
          OperatorHub.

          IMPORTANT: It is advised that the etcd deployment uses persistent
          storage of some type. Either use an existing storage provisioner with
          a default `StorageClass` or simply use `hostPath` volumes.

          ### Installing the operator

          Hit "Install", select the stable update channel and a namespace for the
          operator. Use of a new namespace is recommended.

          Hit "Install" again. At this point you should have just one pod, the
          operator pod, running. Next we needs to configure the remaining provided APIs.

          #### A note on operator namespaces
          The LINSTOR operator can only watch for events and manage
          custom resources that are within the same namespace it is deployed
          within (OwnNamsespace). This means the LINSTOR Controller, LINSTOR
          Satellites, and LINSTOR CSI Driver pods all need to be deployed in the
          same namsepace as the LINSTOR Operator pod.

          ### Deploying the LINSTOR Controller

          Navigate to the left-hand control pane of the Openshift Web
          Console. Expand the "Operators" section, selecting "Installed Operators".
          Find the entry for the "Linstor Operator", then select the "LinstorController"
          from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorController". Click the "Create LinstorController" button.

          Here you will be presented with options to configure the LINSTOR
          Controller. Either via the web-form view or the YAML View. Regardless
          of which view you select, make sure that the `dbConnectionURL` matches
          the endpoint provided from your etcd deployment. Otherwise, the
          defaults are usually fine for most purposes.

          Lastly hit "Create", you should now see a linstor-controller pod
          running.

          ### Deploying the LINSTOR Satellites

          Next we need to deploy the Satellites Set. Just as before navigate
          to the left-hand control pane of the Openshift Web Console. Expand the
          "Operators" section, but this time select "Installed Operators". Find
          the entry for the "Linstor Operator", then select the
          "LinstorSatelliteSet" from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorSatelliteSet". Click the "Create LinstorSatelliteSet" button.

          Here you will be presented with the options to configure the LINSTOR
          Satellites. The defaults should be enough to get you started.
          Make sure the `controllerEndpoint` matches what is available in the
          openshift endpoints. The default is usually correct here.

          You can edit the `storagePools` section to configure LINSTOR storage pools,
          including preparing the backing devices. See our [storage guide].

          [storage guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-storage

          Below is an example manifest:

          ```yaml
          apiVersion: linstor.linbit.com/v1
          kind: LinstorSatelliteSet
          metadata:
            name: linstor-satellites
          spec:
            satelliteImage: ''
            automaticStorageType: None
            storagePools:
              lvmThinPools:
              - name: openshift-pool
                volumeGroup: ""
                thinVolume: openshift
                devicePaths:
                - /dev/vdb
            drbdRepoCred: ''
            kernelModuleInjectionMode: ShippedModules
            controllerEndpoint: 'http://linstor:3370'
            priorityClassName: ''
          ```

          Lastly hit "Create", you should now see a linstor-node pod
          running on every worker node.

          ### Deploying the LINSTOR CSI driver

          Last bit left is the CSI pods to bridge the layer between the CSI and
          LINSTOR. Just as before navigate to the left-hand control pane of the
          Openshift Web Console. Expand the "Operators" section, but this time
          select "Installed Operators". Find the entry for the "Linstor Operator",
          then select the "LinstorCSIDriver" from the "Provided APIs" column on the
          right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorCSIDriver". Click the "Create LinstorCSIDriver" button.

          Again, you will be presented with the options. Make sure that the
          `controllerEndpoint` is correct. Otherwise the defaults are fine for
          most use cases.

          Lastly hit "Create". You will now see a single "linstor-csi-controller" pod,
          as well as a "linstor-csi-node" pod on all worker nodes.

          ## Interacting with LINSTOR in Openshift.

          The Controller pod includes a LINSTOR Client, making it easy to interact directly with LINSTOR.
          For instance:

          ```
          oc exec deployment/linstor-cs-controller -- linstor storage-pool list
          ```

          This should only be necessary for investigating problems and accessing advanced functionality.
          Regular operation such as creating volumes should be achieved via the [Openshift/Kubernetes integration].

          [Openshift/Kubernetes integration]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-basic-configuration-and-deployment
        displayName: Linstor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - DRBD
        links:
        - name: LINBIT
          url: https://www.linbit.com/
        - name: User Guide
          url: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift
        maintainers:
        - email: sales@linbit.com
          name: LINBIT
        maturity: beta
        provider:
          name: LINBIT
        relatedImages:
        - registry.marketplace.redhat.com/rhm/linbit/linstor-operator@sha256:75e3ec9a9331072dc88c13979f0cacdf6232f8b3a6bfaf55c95b2ee3a03a5d5b
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:395a5a4aa4cfe3a0093d2225ce2e67acdcec0fd894e4b61e30a750f22931448d
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:bba872052e42fe0d6cbb2eea6541bf57f035575cbc7d838a3260b8582b8fe572
        - registry.marketplace.redhat.com/rhm/linbit/linstor-controller@sha256:1084ba14eaec7511682dd87c562398ce5e6690ded700c964373188b73bfe3b17
        - registry.marketplace.redhat.com/rhm/linbit/drbd-9@sha256:6337b3adfa3de87e74adcc974edafa09b618cb8e0315658c1cbdf8dd1c6c4c32
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:ba217e7e51f0748b54f477486263ff4983b0029635fd1f7ccc67160a09f81608
        - registry.marketplace.redhat.com/rhm/linbit/linstor-csi@sha256:034e3641376aa7ad47ed309fb389083c3ff82974f246f4d71b8789be8f8f618d
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:53b0f48eeecaa3ff1e211a2a413cd49266b7be85a5aa3a0431ab72d32e3a4e5c
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:8e52a6b31f9f3834d709dbca3614c2110b8ebb4af382c76b7056921221c87600
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:91cfe63acc12c91c3053ba7cc12d72cce404d33aab4afcfac285311f5ddfc7fe
        - registry.marketplace.redhat.com/rhm/linbit/linstor-satellite@sha256:4a3503092c8cb9435ccfb4837ee6a9183b680eb18a35c7d40227fa5c8f4da996
        version: 1.5.1
      entries:
      - name: linstor-operator.v1.5.1
        version: 1.5.1
      name: alpha
    - currentCSV: linstor-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorController",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerImage": "",
                  "dbConnectionURL": "etcd://linstor-etcd:2379",
                  "drbdRepoCred": "",
                  "priorityClassName": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorCSIDriver",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerReplicas": 1,
                  "controllerEndpoint": "http://linstor:3370",
                  "csiControllerServiceAccountName": "csi-controller",
                  "csiNodeServiceAccountName": "csi-node",
                  "linstorPluginImage": "",
                  "imagePullSecret": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorSatelliteSet",
                "metadata": {
                  "name": "linstor-satellites"
                },
                "spec": {
                  "automaticStorageType": "None",
                  "controllerEndpoint": "http://linstor:3370",
                  "drbdRepoCred": "",
                  "kernelModuleInjectionMode": "ShippedModules",
                  "priorityClassName": "",
                  "satelliteImage": ""
                },
                "status": {
                  "errors": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          createdAt: "2021-06-22T07:57:53+00:00"
          description: LINSTOR Kubernetes Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/piraeusdatastore/piraeus-operator
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LinstorController is the Schema for the linstorcontrollersets
              API
            kind: LinstorController
            name: linstorcontrollers.linstor.linbit.com
            version: v1
          - description: LinstorCSIDriver is the Schema for the linstorcsidrivers
              API
            kind: LinstorCSIDriver
            name: linstorcsidrivers.linstor.linbit.com
            version: v1
          - description: LinstorSatelliteSet is the Schema for the linstornodesets
              API
            kind: LinstorSatelliteSet
            name: linstorsatellitesets.linstor.linbit.com
            version: v1
        description: |-
          LINSTOR is a configuration management system for storage on Linux systems.
          It manages LVM logical volumes and/or ZFS ZVOLs on a cluster of nodes.
          It leverages DRBD for replication between different nodes and to provide block
          storage devices to users and applications. It manages snapshots, encryption and
          caching of HDD backed data in SSDs via bcache.

          LINBIT provides a certified LINSTOR operator to ease deployment of LINSTOR
          on Openshift by installing DRBD, managing Satellite and Controller pods,
          and integrating with Openshift to provision persistent storage for your workloads.

          For detailed instructions and more configuration options, see our [user guide].

          [user guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift

          ## Install

          Unlike deployment via the helm chart, the certified Openshift
          operator does not deploy the needed etcd cluster. You must deploy this
          yourself ahead of time. We do this via the etcd operator available in the
          OperatorHub.

          IMPORTANT: It is advised that the etcd deployment uses persistent
          storage of some type. Either use an existing storage provisioner with
          a default `StorageClass` or simply use `hostPath` volumes.

          ### Installing the operator

          Hit "Install", select the stable update channel and a namespace for the
          operator. Use of a new namespace is recommended.

          Hit "Install" again. At this point you should have just one pod, the
          operator pod, running. Next we needs to configure the remaining provided APIs.

          #### A note on operator namespaces
          The LINSTOR operator can only watch for events and manage
          custom resources that are within the same namespace it is deployed
          within (OwnNamsespace). This means the LINSTOR Controller, LINSTOR
          Satellites, and LINSTOR CSI Driver pods all need to be deployed in the
          same namsepace as the LINSTOR Operator pod.

          ### Deploying the LINSTOR Controller

          Navigate to the left-hand control pane of the Openshift Web
          Console. Expand the "Operators" section, selecting "Installed Operators".
          Find the entry for the "Linstor Operator", then select the "LinstorController"
          from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorController". Click the "Create LinstorController" button.

          Here you will be presented with options to configure the LINSTOR
          Controller. Either via the web-form view or the YAML View. Regardless
          of which view you select, make sure that the `dbConnectionURL` matches
          the endpoint provided from your etcd deployment. Otherwise, the
          defaults are usually fine for most purposes.

          Lastly hit "Create", you should now see a linstor-controller pod
          running.

          ### Deploying the LINSTOR Satellites

          Next we need to deploy the Satellites Set. Just as before navigate
          to the left-hand control pane of the Openshift Web Console. Expand the
          "Operators" section, but this time select "Installed Operators". Find
          the entry for the "Linstor Operator", then select the
          "LinstorSatelliteSet" from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorSatelliteSet". Click the "Create LinstorSatelliteSet" button.

          Here you will be presented with the options to configure the LINSTOR
          Satellites. The defaults should be enough to get you started.
          Make sure the `controllerEndpoint` matches what is available in the
          openshift endpoints. The default is usually correct here.

          You can edit the `storagePools` section to configure LINSTOR storage pools,
          including preparing the backing devices. See our [storage guide].

          [storage guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-storage

          Below is an example manifest:

          ```yaml
          apiVersion: linstor.linbit.com/v1
          kind: LinstorSatelliteSet
          metadata:
            name: linstor-satellites
          spec:
            satelliteImage: ''
            automaticStorageType: None
            storagePools:
              lvmThinPools:
              - name: openshift-pool
                volumeGroup: ""
                thinVolume: openshift
                devicePaths:
                - /dev/vdb
            drbdRepoCred: ''
            kernelModuleInjectionMode: ShippedModules
            controllerEndpoint: 'http://linstor:3370'
            priorityClassName: ''
          ```

          Lastly hit "Create", you should now see a linstor-node pod
          running on every worker node.

          ### Deploying the LINSTOR CSI driver

          Last bit left is the CSI pods to bridge the layer between the CSI and
          LINSTOR. Just as before navigate to the left-hand control pane of the
          Openshift Web Console. Expand the "Operators" section, but this time
          select "Installed Operators". Find the entry for the "Linstor Operator",
          then select the "LinstorCSIDriver" from the "Provided APIs" column on the
          right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorCSIDriver". Click the "Create LinstorCSIDriver" button.

          Again, you will be presented with the options. Make sure that the
          `controllerEndpoint` is correct. Otherwise the defaults are fine for
          most use cases.

          Lastly hit "Create". You will now see a single "linstor-csi-controller" pod,
          as well as a "linstor-csi-node" pod on all worker nodes.

          ## Interacting with LINSTOR in Openshift.

          The Controller pod includes a LINSTOR Client, making it easy to interact directly with LINSTOR.
          For instance:

          ```
          oc exec deployment/linstor-cs-controller -- linstor storage-pool list
          ```

          This should only be necessary for investigating problems and accessing advanced functionality.
          Regular operation such as creating volumes should be achieved via the [Openshift/Kubernetes integration].

          [Openshift/Kubernetes integration]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-basic-configuration-and-deployment
        displayName: Linstor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - DRBD
        links:
        - name: LINBIT
          url: https://www.linbit.com/
        - name: User Guide
          url: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift
        maintainers:
        - email: sales@linbit.com
          name: LINBIT
        maturity: beta
        provider:
          name: LINBIT
        relatedImages:
        - registry.marketplace.redhat.com/rhm/linbit/linstor-csi@sha256:034e3641376aa7ad47ed309fb389083c3ff82974f246f4d71b8789be8f8f618d
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:8e52a6b31f9f3834d709dbca3614c2110b8ebb4af382c76b7056921221c87600
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:395a5a4aa4cfe3a0093d2225ce2e67acdcec0fd894e4b61e30a750f22931448d
        - registry.marketplace.redhat.com/rhm/linbit/linstor-controller@sha256:1084ba14eaec7511682dd87c562398ce5e6690ded700c964373188b73bfe3b17
        - registry.marketplace.redhat.com/rhm/linbit/linstor-satellite@sha256:4a3503092c8cb9435ccfb4837ee6a9183b680eb18a35c7d40227fa5c8f4da996
        - registry.marketplace.redhat.com/rhm/linbit/linstor-operator@sha256:75e3ec9a9331072dc88c13979f0cacdf6232f8b3a6bfaf55c95b2ee3a03a5d5b
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:53b0f48eeecaa3ff1e211a2a413cd49266b7be85a5aa3a0431ab72d32e3a4e5c
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:bba872052e42fe0d6cbb2eea6541bf57f035575cbc7d838a3260b8582b8fe572
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:ba217e7e51f0748b54f477486263ff4983b0029635fd1f7ccc67160a09f81608
        - registry.marketplace.redhat.com/rhm/linbit/drbd-9@sha256:6337b3adfa3de87e74adcc974edafa09b618cb8e0315658c1cbdf8dd1c6c4c32
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:91cfe63acc12c91c3053ba7cc12d72cce404d33aab4afcfac285311f5ddfc7fe
        version: 1.5.1
      entries:
      - name: linstor-operator.v1.5.1
        version: 1.5.1
      name: stable
    defaultChannel: stable
    packageName: linstor-operator-rhmp
    provider:
      name: LINBIT
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lib-bucket-provisioner
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lib-bucket-provisioner.v2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage,Big Data
          certified: "false"
          containerImage: quay.io/noobaa/pause
          createdAt: "2020-07-01T06:28:48Z"
          description: Library for the dynamic provisioning of object store buckets
            to be used by object store providers.
          repository: https://github.com/kube-object-storage/lib-bucket-provisioner
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: ObjectBucketClaim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information
            displayName: ObjectBucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: |
          ### CRD-only Operator

          This operator package is **CRD-only** and the operator is a no-op operator.

          Instead, bucket provisioners using this library are using these CRD's and using CSV [required-crds](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/Documentation/design/building-your-csv.md#required-crds) them so that OLM can install it as a dependency.

          ### Generic Bucket Provisioning

          Kubernetes natively supports dynamic provisioning for many types of file and block storage, but lacks support for object bucket provisioning.

          This repo is a placeholder for an object store bucket provisioning library, very similar to the Kubernetes [sig-storage-lib-external-provisioner](https://github.com/kubernetes-sigs/sig-storage-lib-external-provisioner/blob/master/controller/controller.go) library.

          ### Known Provisioners
          - https://github.com/noobaa/noobaa-operator (NooBaa)
          - https://github.com/rook/rook (Rook-Ceph)
          - https://github.com/yard-turkey/aws-s3-provisioner (AWS-S3)
        displayName: lib-bucket-provisioner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - object
        - bucket
        - storage
        - cloud
        - s3
        links:
        - name: Github
          url: https://github.com/kube-object-storage/lib-bucket-provisioner
        maintainers:
        - email: jcope@redhat.com
          name: Jon Cope
        - email: jvance@redhat.com
          name: Jeff Vance
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: dzaken@redhat.com
          name: Danny Zaken
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/noobaa/pause
        version: 2.0.0
      entries:
      - name: lib-bucket-provisioner.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: lib-bucket-provisioner
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: dapr.io
      provider-url: https://dapr.io
    name: dapr-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dapr-kubernetes-operator.v0.0.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprControlPlane",
                "metadata": {
                  "name": "dapr-control-plane",
                  "namespace": "dapr-system"
                },
                "spec": {
                  "values": {}
                }
              },
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprCruiseControl",
                "metadata": {
                  "name": "dapr-cruise-control",
                  "namespace": "dapr-system"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprInstance",
                "metadata": {
                  "name": "dapr-instance",
                  "namespace": "dapr-system"
                },
                "spec": {
                  "values": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          containerImage: docker.io/daprio/kubernetes-operator@sha256:da599a273943bf28812314a11df1052411237dd1887edf0fe6104c5857f439b0
          createdAt: "2025-10-06T18:47:59Z"
          description: Dapr Control Plane Operator
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/dapr/kubernetes-operator
          support: contact@dapr.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DaprControlPlane is the Schema for the Dapr ControlPlane
              API.
            displayName: DaprControlPlane
            kind: DaprControlPlane
            name: daprcontrolplanes.operator.dapr.io
            version: v1alpha1
          - kind: DaprCruiseControl
            name: daprcruiscontrols.operator.dapr.io
            version: v1alpha1
          - description: DaprInstance is the Schema for the Dapr Instance API.
            displayName: DaprInstance
            kind: DaprInstance
            name: daprinstances.operator.dapr.io
            version: v1alpha1
        description: Dapr Control Plane Operator
        displayName: Dapr Control Plane Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dapr
        links:
        - name: Dapr
          url: https://dapr.io
        - name: Dapr Kubernetes Operator
          url: https://github.com/dapr/kubernetes-operator
        maintainers:
        - email: lburgazzoli@gmail.com
          name: Luca Burgazzoli
        - email: salaboy@gmail.com
          name: Mauricio Salatino
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: dapr.io
          url: https://dapr.io
        relatedImages:
        - docker.io/daprio/kubernetes-operator@sha256:da599a273943bf28812314a11df1052411237dd1887edf0fe6104c5857f439b0
        version: 0.0.12
      entries:
      - name: dapr-kubernetes-operator.v0.0.12
        version: 0.0.12
      - name: dapr-kubernetes-operator.v0.0.11
        version: 0.0.11
      - name: dapr-kubernetes-operator.v0.0.10
        version: 0.0.10
      - name: dapr-kubernetes-operator.v0.0.9
        version: 0.0.9
      - name: dapr-kubernetes-operator.v0.0.8
        version: 0.0.8
      - name: dapr-kubernetes-operator.v0.0.6
        version: 0.0.6
      - name: dapr-kubernetes-operator.v0.0.5
        version: 0.0.5
      - name: dapr-kubernetes-operator.v0.0.4
        version: 0.0.4
      - name: dapr-kubernetes-operator.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: dapr-kubernetes-operator
    provider:
      name: dapr.io
      url: https://dapr.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: gitwebhook-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitwebhook-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitWebhook",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "gitwebhook-operator",
                    "app.kubernetes.io/instance": "gitwebhook-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gitwebhook",
                    "app.kubernetes.io/part-of": "gitwebhook-operator"
                  },
                  "name": "gitwebhook-sample"
                },
                "spec": null
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/redhat-cop/gitwebhook-operator@sha256:2810d75eebb24069dbbedf6a82285136c6f16b4cc58b59f1b0fbdc38560cf8a8
          createdAt: "2024-05-10T18:17:13Z"
          description: An operator to create git webhooks declaratively
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: gitwebhook-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/gitwebhook-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitWebhook is the Schema for the gitwebhooks API
            displayName: Git Webhook
            kind: GitWebhook
            name: gitwebhooks.redhatcop.redhat.io
            version: v1alpha1
        description: "This operator automates the setting up of git webhooks. Its
          main purpose is to stream line the configuration of operators such as Tekton
          or ArgoCD which rely on git webhooks for some of their functionalities.\n\n##
          The GitWebhook CRD\n \nA CRD is provided by this operator to create a git
          web hook, here is an example for github\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GitWebhook\nmetadata:\n  name: gitwebhook-github\nspec:\n  gitHub:\n    gitServerCredentials:\n
          \     name: github-pat\n  repositoryOwner: ${repo_owner}\n  ownerType: user\n
          \ repositoryName: ${repo_name}\n  webhookURL: https://hellowebhook.com\n
          \ insecureSSL: false\n  webhookSecret:\n    name: webhook-secret\n  events:\n
          \   - push  \n  contentType: json\n  active: true  \n```    \n\nhere is
          an explanation of each field:\n\n- `gihub` specifies how to connect to the
          git api server. It also requires a local reference to a secret (in the same
          namespace) containing a key `token` with a valid github token to be used
          to authenticate. A similar `gitLab` section exists when connecting to gitlab.
          Only one of `gitLab` or `gitHub` can be defined. \n- `repositoryOwner` and
          `repositoryName` identify the repository for which we want to receive events.\n-
          `ownerType` can have two values: `user` and `organization` and identifies
          the kind of owner.\n- `webhookURL` is the URL for to be called.\n - `insecureSSL`
          defines whether the target URL certificate should be validated (default
          `false`). \n- `webhookSecret` defines a local reference to a secret containing
          the `secret` key. The value is a shared secret between the webhook caller
          and the received for farther validation or identification of the caller.\n-
          `events` is the list of the repo-level events that the webhook should generate.
          The list of valid events for github can be found [here](https://docs.github.com/en/developers/webhooks-and-events/webhooks/webhook-events-and-payloads).
          The list of valid events for gitlab can be found [here](https://docs.gitlab.com/ee/user/project/integrations/webhook_events.html).\n
          - `contentType` defines the format of the webhook payload (default `json`)
          (github).\n - `active` whether the webhook should be turned on (default
          `true`) (github).\n - `pushEventBranchFilter` a regular expression to filter
          from which branches push events should be generated (gitlab only).\n\n##
          Security Considerations\n\nThis operator does not own credentials for the
          git server, but instead always allocate a new connection based on the credentials
          referenced in the CR and every reconcile cycle. As a result there is no
          risk of security escalation or credential leaking between tenants of a cluster
          using this operator. On the other hand it is the responsibility of the namespace
          owners or the platform owner to ensure that valid git credentials are always
          available in the namespace where the GitWebhook CRs need to defined.\n\n##
          Current support\n\nCurrently this operator support creating repo-level webhooks
          for github and gitlab. Potentially this operator could be extended to support
          org-level webhook or other git systems. Contributions are welcome.  "
        displayName: GitWebhook operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - webhook
        - github
        - gitlab
        links:
        - name: GitWbehook Operator
          url: https://github.com/redhat-cop/gitwebhook-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/gitwebhook-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/gitwebhook-operator@sha256:2810d75eebb24069dbbedf6a82285136c6f16b4cc58b59f1b0fbdc38560cf8a8
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        version: 0.1.2
      entries:
      - name: gitwebhook-operator.v0.1.2
        version: 0.1.2
      - name: gitwebhook-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: gitwebhook-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ec2-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ec2-controller.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "DHCPOptions",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "ElasticIPAddress",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "Instance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "InternetGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "NATGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "RouteTable",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "SecurityGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "Subnet",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "TransitGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "VPC",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "VPCEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ec2-controller:1.7.0
          createdAt: "2025-09-27T00:04:55Z"
          description: AWS EC2 controller is a service controller for managing EC2
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CapacityReservation represents the state of an AWS ec2 CapacityReservation
              resource.
            displayName: CapacityReservation
            kind: CapacityReservation
            name: capacityreservations.ec2.services.k8s.aws
            version: v1alpha1
          - description: DHCPOptions represents the state of an AWS ec2 DHCPOptions
              resource.
            displayName: DHCPOptions
            kind: DHCPOptions
            name: dhcpoptions.ec2.services.k8s.aws
            version: v1alpha1
          - description: ElasticIPAddress represents the state of an AWS ec2 ElasticIPAddress
              resource.
            displayName: ElasticIPAddress
            kind: ElasticIPAddress
            name: elasticipaddresses.ec2.services.k8s.aws
            version: v1alpha1
          - description: FlowLog represents the state of an AWS ec2 FlowLog resource.
            displayName: FlowLog
            kind: FlowLog
            name: flowlogs.ec2.services.k8s.aws
            version: v1alpha1
          - description: Instance represents the state of an AWS ec2 Instance resource.
            displayName: Instance
            kind: Instance
            name: instances.ec2.services.k8s.aws
            version: v1alpha1
          - description: InternetGateway represents the state of an AWS ec2 InternetGateway
              resource.
            displayName: InternetGateway
            kind: InternetGateway
            name: internetgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: LaunchTemplate represents the state of an AWS ec2 LaunchTemplate
              resource.
            displayName: LaunchTemplate
            kind: LaunchTemplate
            name: launchtemplates.ec2.services.k8s.aws
            version: v1alpha1
          - description: NATGateway represents the state of an AWS ec2 NATGateway
              resource.
            displayName: NATGateway
            kind: NATGateway
            name: natgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: NetworkACL represents the state of an AWS ec2 NetworkACL
              resource.
            displayName: NetworkACL
            kind: NetworkACL
            name: networkacls.ec2.services.k8s.aws
            version: v1alpha1
          - description: RouteTable represents the state of an AWS ec2 RouteTable
              resource.
            displayName: RouteTable
            kind: RouteTable
            name: routetables.ec2.services.k8s.aws
            version: v1alpha1
          - description: SecurityGroup represents the state of an AWS ec2 SecurityGroup
              resource.
            displayName: SecurityGroup
            kind: SecurityGroup
            name: securitygroups.ec2.services.k8s.aws
            version: v1alpha1
          - description: Subnet represents the state of an AWS ec2 Subnet resource.
            displayName: Subnet
            kind: Subnet
            name: subnets.ec2.services.k8s.aws
            version: v1alpha1
          - description: TransitGateway represents the state of an AWS ec2 TransitGateway
              resource.
            displayName: TransitGateway
            kind: TransitGateway
            name: transitgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: TransitGatewayVPCAttachment represents the state of an AWS
              ec2 TransitGatewayVPCAttachment resource.
            displayName: TransitGatewayVPCAttachment
            kind: TransitGatewayVPCAttachment
            name: transitgatewayvpcattachments.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCEndpoint represents the state of an AWS ec2 VPCEndpoint
              resource.
            displayName: VPCEndpoint
            kind: VPCEndpoint
            name: vpcendpoints.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCEndpointServiceConfiguration represents the state of an
              AWS ec2 VPCEndpointServiceConfiguration resource.
            displayName: VPCEndpointServiceConfiguration
            kind: VPCEndpointServiceConfiguration
            name: vpcendpointserviceconfigurations.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCPeeringConnection represents the state of an AWS ec2 VPCPeeringConnection
              resource.
            displayName: VPCPeeringConnection
            kind: VPCPeeringConnection
            name: vpcpeeringconnections.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPC represents the state of an AWS ec2 VPC resource.
            displayName: VPC
            kind: VPC
            name: vpcs.ec2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Compute Cloud (EC2) resources in AWS from within your Kubernetes cluster.

          **About Amazon EC2**

          Amazon Elastic Compute Cloud (Amazon EC2) provides scalable computing capacity in the Amazon Web Services (AWS) Cloud. Using Amazon EC2 eliminates your need to invest in hardware up front, so you can develop and deploy applications faster. You can use Amazon EC2 to launch as many or as few virtual servers as you need, configure security and networking, and manage storage. Amazon EC2 enables you to scale up or down to handle changes in requirements or spikes in popularity, reducing your need to forecast traffic.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EC2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ec2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EC2 Developer Resources
          url: https://aws.amazon.com/ec2/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ec2 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ec2-controller:1.7.0
        version: 1.7.0
      entries:
      - name: ack-ec2-controller.v1.7.0
        version: 1.7.0
      - name: ack-ec2-controller.v1.6.2
        version: 1.6.2
      - name: ack-ec2-controller.v1.6.1
        version: 1.6.1
      - name: ack-ec2-controller.v1.6.0
        version: 1.6.0
      - name: ack-ec2-controller.v1.4.11
        version: 1.4.11
      - name: ack-ec2-controller.v1.4.10
        version: 1.4.10
      - name: ack-ec2-controller.v1.4.9
        version: 1.4.9
      - name: ack-ec2-controller.v1.4.8
        version: 1.4.8
      - name: ack-ec2-controller.v1.4.7
        version: 1.4.7
      - name: ack-ec2-controller.v1.4.6
        version: 1.4.6
      - name: ack-ec2-controller.v1.4.5
        version: 1.4.5
      - name: ack-ec2-controller.v1.4.4
        version: 1.4.4
      - name: ack-ec2-controller.v1.4.2
        version: 1.4.2
      - name: ack-ec2-controller.v1.4.1
        version: 1.4.1
      - name: ack-ec2-controller.v1.4.0
        version: 1.4.0
      - name: ack-ec2-controller.v1.3.7
        version: 1.3.7
      - name: ack-ec2-controller.v1.3.6
        version: 1.3.6
      - name: ack-ec2-controller.v1.3.5
        version: 1.3.5
      - name: ack-ec2-controller.v1.3.4
        version: 1.3.4
      - name: ack-ec2-controller.v1.3.3
        version: 1.3.3
      - name: ack-ec2-controller.v1.3.1
        version: 1.3.1
      - name: ack-ec2-controller.v1.3.0
        version: 1.3.0
      - name: ack-ec2-controller.v1.2.28
        version: 1.2.28
      - name: ack-ec2-controller.v1.2.27
        version: 1.2.27
      - name: ack-ec2-controller.v1.2.26
        version: 1.2.26
      - name: ack-ec2-controller.v1.2.25
        version: 1.2.25
      - name: ack-ec2-controller.v1.2.24
        version: 1.2.24
      - name: ack-ec2-controller.v1.2.23
        version: 1.2.23
      - name: ack-ec2-controller.v1.2.22
        version: 1.2.22
      - name: ack-ec2-controller.v1.2.21
        version: 1.2.21
      - name: ack-ec2-controller.v1.2.20
        version: 1.2.20
      - name: ack-ec2-controller.v1.2.17
        version: 1.2.17
      - name: ack-ec2-controller.v1.2.16
        version: 1.2.16
      - name: ack-ec2-controller.v1.2.15
        version: 1.2.15
      - name: ack-ec2-controller.v1.2.14
        version: 1.2.14
      - name: ack-ec2-controller.v1.2.13
        version: 1.2.13
      - name: ack-ec2-controller.v1.2.12
        version: 1.2.12
      - name: ack-ec2-controller.v1.2.11
        version: 1.2.11
      - name: ack-ec2-controller.v1.2.10
        version: 1.2.10
      - name: ack-ec2-controller.v1.2.9
        version: 1.2.9
      - name: ack-ec2-controller.v1.2.8
        version: 1.2.8
      - name: ack-ec2-controller.v1.2.7
        version: 1.2.7
      - name: ack-ec2-controller.v1.2.6
        version: 1.2.6
      - name: ack-ec2-controller.v1.2.4
        version: 1.2.4
      - name: ack-ec2-controller.v1.2.1
        version: 1.2.1
      - name: ack-ec2-controller.v1.2.0
        version: 1.2.0
      - name: ack-ec2-controller.v1.1.2
        version: 1.1.2
      - name: ack-ec2-controller.v1.1.1
        version: 1.1.1
      - name: ack-ec2-controller.v1.1.0
        version: 1.1.0
      - name: ack-ec2-controller.v1.0.7
        version: 1.0.7
      - name: ack-ec2-controller.v1.0.6
        version: 1.0.6
      - name: ack-ec2-controller.v1.0.4
        version: 1.0.4
      - name: ack-ec2-controller.v1.0.3
        version: 1.0.3
      - name: ack-ec2-controller.v1.0.2
        version: 1.0.2
      - name: ack-ec2-controller.v1.0.1
        version: 1.0.1
      - name: ack-ec2-controller.v1.0.0
        version: 1.0.0
      - name: ack-ec2-controller.v0.1.1
        version: 0.1.1
      - name: ack-ec2-controller.v0.1.0
        version: 0.1.0
      - name: ack-ec2-controller.v0.0.21
        version: 0.0.21
      - name: ack-ec2-controller.v0.0.20
        version: 0.0.20
      - name: ack-ec2-controller.v0.0.19
        version: 0.0.19
      - name: ack-ec2-controller.v0.0.18
        version: 0.0.18
      - name: ack-ec2-controller.v0.0.17
        version: 0.0.17
      - name: ack-ec2-controller.v0.0.16
        version: 0.0.16
      - name: ack-ec2-controller.v0.0.15
        version: 0.0.15
      - name: ack-ec2-controller.v0.0.14
        version: 0.0.14
      - name: ack-ec2-controller.v0.0.13
        version: 0.0.13
      - name: ack-ec2-controller.v0.0.12
        version: 0.0.12
      - name: ack-ec2-controller.v0.0.11
        version: 0.0.11
      - name: ack-ec2-controller.v0.0.10
        version: 0.0.10
      - name: ack-ec2-controller.v0.0.9
        version: 0.0.9
      - name: ack-ec2-controller.v0.0.8
        version: 0.0.8
      - name: ack-ec2-controller.v0.0.7
        version: 0.0.7
      - name: ack-ec2-controller.v0.0.6
        version: 0.0.6
      - name: ack-ec2-controller.v0.0.5
        version: 0.0.5
      - name: ack-ec2-controller.v0.0.4
        version: 0.0.4
      - name: ack-ec2-controller.v0.0.3
        version: 0.0.3
      - name: ack-ec2-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-ec2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: global-load-balancer-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: global-load-balancer-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSRecord",
                "metadata": {
                  "name": "hello-global-record"
                },
                "spec": {
                  "endpoints": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    }
                  ],
                  "globalZoneRef": {
                    "name": "external-dns-zone"
                  },
                  "loadBalancingPolicy": "Multivalue",
                  "name": "hello.my.base.domain",
                  "ttl": 60
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSRecord",
                "metadata": {
                  "name": "route53-geoproximity-global-record-with-healthcheck"
                },
                "spec": {
                  "endpoints": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    }
                  ],
                  "globalZoneRef": {
                    "name": "route53-global-dns-zone"
                  },
                  "healthCheck": {
                    "failureThreshold": 3,
                    "httpGet": {
                      "host": "geoproximity-hc.${global_base_domain}",
                      "path": "/healthz",
                      "port": 80,
                      "scheme": "HTTP"
                    },
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 2
                  },
                  "loadBalancingPolicy": "Geoproximity",
                  "name": "geoproximity-hc.my.base.domain",
                  "ttl": 60
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSZone",
                "metadata": {
                  "name": "external-dns-zone"
                },
                "spec": {
                  "domain": "myglobalzone.mybasedomain.io",
                  "provider": {
                    "externalDNS": {
                      "annotations": {
                        "type": "global"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSZone",
                "metadata": {
                  "name": "route53-global-dns-zone"
                },
                "spec": {
                  "domain": "myglobalzone.mybasedomain.io",
                  "provider": {
                    "route53": {
                      "credentialsSecretRef": {
                        "name": "route53-global-zone-credentials",
                        "namespace": "mynamespace"
                      },
                      "zoneID": "Z083892239ROKNEP70BDY"
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalRouteDiscovery",
                "metadata": {
                  "name": "external-dns-globalroutediscovery"
                },
                "spec": {
                  "clusters": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3"
                    }
                  ],
                  "defaultLoadBalancingPolicy": "Multivalue",
                  "defaultTTL": 30,
                  "globalZoneRef": {
                    "name": "external-dns-zone"
                  },
                  "routeSelector": {
                    "matchLabels": {
                      "route-type": "global"
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalRouteDiscovery",
                "metadata": {
                  "name": "route53-globalroutediscovery"
                },
                "spec": {
                  "clusters": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3"
                    }
                  ],
                  "defaultLoadBalancingPolicy": "Multivalue",
                  "defaultTTL": 30,
                  "globalZoneRef": {
                    "name": "route53-global-dns-zone"
                  },
                  "routeSelector": {
                    "matchLabels": {
                      "route-type": "global"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Cloud Provider
          certified: "false"
          containerImage: quay.io/redhat-cop/global-load-balancer-operator@sha256:c0b6ac9b03571a74a56b28302d720e8aa4442233f7867d05ed365e022f7cb8f5
          createdAt: "2021-12-02T17:33:23Z"
          description: This operator creates automation around a DNS to operate as
            a global load balancer for a set of OpenShift clusters.
          operatorframework.io/suggested-namespace: global-load-balancer-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/global-load-balancer-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DNSEndpoint
            name: dnsendpoints.externaldns.k8s.io
            version: v1alpha1
          - description: GlobalDNSRecord is the Schema for the globaldnsrecords API
            displayName: Global DNSRecord
            kind: GlobalDNSRecord
            name: globaldnsrecords.redhatcop.redhat.io
            version: v1alpha1
          - description: GlobalDNSZone is the Schema for the globaldnszones API
            displayName: Global DNSZone
            kind: GlobalDNSZone
            name: globaldnszones.redhatcop.redhat.io
            version: v1alpha1
          - description: GlobalRouteDiscovery is the Schema for the globalroutediscoveries
              API
            displayName: Global Route Discovery
            kind: GlobalRouteDiscovery
            name: globalroutediscoveries.redhatcop.redhat.io
            version: v1alpha1
        description: "The global-load-balancer-operator implements automation to program
          a DNS service to act as global load balancer for applications deployed to
          multiple OpenShift clusters.\nThis operator is designed to be deployed to
          a control cluster which will watch the load balanced clusters (controlled
          clusters).\nThere are two main concepts (APIs) provided by this operator:\n\n1.
          GlobalDNSZone\n2. GlobalDNSRecord\n\n## GlobalDNSZone\n\nThe `GlobalDNSZone`
          CR allows you to configure a zone which will contain global load balanced
          records and the provider used to populate it.\nHere is an example of GlobalDNSZone:\n\n```yaml\napiVersion:
          redhatcop.redhat.io/v1alpha1\nkind: GlobalDNSZone\nmetadata:\n  name: external-dns-zone\nspec:\n
          \ # Add fields here\n  domain: global.myzone.io\n  provider:\n    externalDNS:\n
          \     annotations:\n        type: global\n```\n\nHere is a table summarizing
          the supported providers and their capabilities:\n\n| Provider  | Supports
          Health Checks  | Supports Multivalue LB | Supports Latency LB  | Supports
          GeoProximity LB  |\n|:--:|:--:|:---:|:---:|:---:|\n| External-dns  | no
          \ | yes | no  | no  |\n| Route53  | yes | yes(*)  | yes(*)  | yes(*)  |\n\n(*)
          only if all controlled clusters run on AWS.\n\n## GlobalDNSRecord\n\nThe
          `GlobalDNSRecord` CR allows you to specify the intention to create a global
          dns record. Here is an example\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GlobalDNSRecord\nmetadata:\n  name: hello-global-record\nspec:\n  name:
          hello.global.myzone.io\n  endpoints:\n  - clusterName: cluster1\n    clusterCredentialRef:\n
          \     name: kubeconfig\n      namespace: cluster1\n    loadBalancerServiceRef:\n
          \     name: router-default\n      namespace: openshift-ingress\n  - clusterName:
          cluster2\n    clusterCredentialRef:\n      name: kubeconfig\n      namespace:
          cluster2\n    loadBalancerServiceRef:\n      name: router-default\n      namespace:
          openshift-ingress\n  - clusterName: cluster3\n    clusterCredentialRef:\n
          \     name: kubeconfig\n      namespace: cluster3\n    loadBalancerServiceRef:\n
          \     name: router-default\n      namespace: openshift-ingress\n  ttl: 60\n
          \ loadBalancingPolicy: Multivalue\n  globalZoneRef:\n    name: external-dns-zone\n
          \ healthCheck:\n    failureThreshold: 3\n    httpGet:\n      host: hello.global.myzone.io\n
          \     httpHeaders:\n        - name: ciao\n          value: hello\n      path:
          /readyz\n      port: 80\n      scheme: HTTP\n    periodSeconds: 10\n    successThreshold:
          1\n    timeoutSeconds: 1     \n```\n\nThe list of endpoints identifies the
          set of `LoadBalancer` type of services to watch on the remote clusters (at
          the moment only LoadBalancer services are supported). These are the `LoadBalancer`
          services created by the [ingress controller](https://docs.openshift.com/container-platform/4.5/networking/ingress-operator.html)
          on which the routers rely. Each entry of this list requires a reference
          to the loadbalancer service on the remote cluster and a reference to a local
          secret containing the credential to connect to the remote cluster.\n\nThis
          secret must contain one entry called `kubeconfig` with a kubeconfig file
          with a default context pointing to the remote cluster. The account used
          by the kubeconfig file (presumably a service account) will require at a
          minimum cluster-level permissions as described in this [cluster role](./docs/scripts/remote-account-cluster-role.yaml).\n\nThe
          `globalZoneRef` field refers to a local (same namespace) `GlobalZone` CR.
          The DNS record represented by this `GlobalDNSRecord`, will be created in
          the referenced zone.\n\n`ttl` is the TTL of the crated DNS record.\n\n`loadBalancingPolicy`
          is the load balancing policy for this global record. It must match one of
          the policy supported by the provider of the referenced `GlobalZone`.\n\nFinally,
          `healthcheck` represent a [probe](https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.18/#probe-v1-core)
          to be used to test the health of a record. This field will be ignored if
          the provider does not support health checks.\n\n## External DNS Provider\n\nThe
          [external-dns](https://github.com/kubernetes-sigs/external-dns) provider
          delegates to external-dns the creation of the actual DNS records by creating
          a `DNSEndpoint` object.\nThe `DNSEndpoint` object will be created in the
          same namespace as the `GlobalDNSRecord` and will be owned by it.\nThe `DNSEdnpoint`
          object will have the same labels as the `GlobalDNSRecord` and the annotations
          specified in the GlobalDNSZone configuration.\nExternal-dns should be configured
          to watch for DNSEnpoints at the cluster level and to point to the desired
          provider.\nDetails on configuration can be found at the external-dns git
          repository.\nThe External-dns should be used as a fall back option when
          other options are not available as it does not support health checks and
          advanced load balancing policies.\n\n## AWS Route53 provider\n\nAWS Route53
          provider uses the Route53 service as a global loadbalancer and offers advanced
          routing capabilities via [route53 traffic policies](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/traffic-flow.html)
          (note that traffic policies will trigger an expense).\nThe following routing
          polices are currently supported:\n\n1. [Multivalue](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-multivalue)\n2.
          [Geoproximity](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-geoproximity)\n3.
          [Latency](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-latency)\n\nAWS
          Route53 provider at the moment requires that all the controlled clusters
          run in AWS.\n\nIf health checks are defined, a route53 health check originating
          from any reason (you have to ensure connectivity) will be created for each
          of the endpoint. Because the endpoint represent s a shared ELB (shared with
          other apps, that is) and the health check is app specific, we cannot use
          the ELB health check, so the route53 endpoint is created with one of the
          two IP exposed by the ELB. This is suboptimal, but it works in most situations.\n\n##
          Global Route Auto Discovery\n\nThe aboev examples showed how to create global
          DNS records. This can be good in some situations, but most of the times
          in an openshift deployment global DNS records will point to routes that
          are intended to be global.\nThe global-load-balancer operator can auto-discover
          these routes and automatically create the corresponding `GloablDNSRecord`.
          The `GlobalRouteDiscovery` CRD is used to configure the discovery process,
          here is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GlobalRouteDiscovery\nmetadata:\n  name: route53-globalroutediscovery\nspec:\n
          \ clusters:\n  - clusterName: cluster1\n    clusterCredentialRef:\n      name:
          ${cluster1_secret_name}\n      namespace: cluster1\n...\n  routeSelector:\n
          \   matchLabels:\n      route-type: global\n  defaultLoadBalancingPolicy:
          Multivalue\n  defaultTTL: 30 \n  globalZoneRef:\n    name: route53-global-dns-zone\n```\n\nThis
          global discovery route will discover routes in the provided list of cluster.
          Only the routes that match the route selector will be considered global.
          \nThe default load balancing policy and default TTL can be expressed in
          the `GlobalRouteDiscovery` CR. However with the following annotations, it's
          possible to configure route-specific values:\n\n* `global-load-balancer-operator.redhat-cop.io/load-balancing-policy`
          to set the load balancing policy\n* `global-load-balancer-operator.redhat-cop.io/ttl`
          to set the TTL\n\nThe globalZoneRef refers to the global zone to be used
          for the created `GlobalDNSRecords`.\n\nHealth checks will also be automatically
          discovered. If the pods behind the route expose a readiness check of httpGet
          kind, that configuration will be used to create the GlobalDNSRecord health
          check.\nWhen more than one container is present in the pod, by default the
          first container will be examined for health check. This behavior can be
          overridden with the this annotation on the route: `global-load-balancer-operator.redhat-cop.io/container-probe`
          where the value will container the name of the container with teh correct
          readiness probe.\n\nIf routes with the same namespace and name exists in
          multiple cluster, the following conditions must be met:\n\n* all host names
          must be the same\n* all load balancing policy must be the same\n* all TTLs
          must be the same\n* all discovered readiness checks must be the same\n\n##
          Examples\n\nThese examples are intended to help you setting up working configuration
          with each of the providers\n\n### Cluster Set up\n\nTwo approaches for cluster
          setup are provided\n\n1. [One cluster, three ingress-gateways.](https://github.com/redhat-cop/global-load-balancer-operator/docs/one-cluster-three-ingress-gateways.md)
          This approach is intended for development purposes and has the objective
          to keep resource consumption at the minimum.\n2. [Control cluster and three
          controlled clusters in different regions.](https://github.com/redhat-cop/global-load-balancer-operator/docs/three-clusters.md).
          This approach represents a more realistic set-up albeit it consumes more
          resources.\n\nYou can also set up the cluster on your own, at the end the
          following conditions must be met:\n\nThree namespace `cluster1` `cluster2`
          `cluster3` are created.\nthe following environment variables are initialized
          for each cluster:\n\n1. <cluster>_secret_name. Pointing to a secret in each
          of the cluster namespaces containing a valid kubeconfig fot that cluster\n2.
          <cluster>_service_name.  Pointing to the name of the loadbalancer service
          to be used for that cluster.\n3. <cluster>_service_namespace. Pointing to
          the namespace of the loadbalancer service to be used for that cluster.\n\nHere
          are examples for the supported provider:\n\n1. [Setting up external-dns
          as provider](https://github.com/redhat-cop/global-load-balancer-operator/docs/external-dns-provider.md)\n2.
          [Setting up route53 as a provider](https://github.com/redhat-cop/global-load-balancer-operator/docs/aws-route53-provider.md)\n"
        displayName: Global Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Networking
        - Cloud Provider
        - Global Load Balancing
        - GLB
        - GTM
        - route53
        - DNS
        links:
        - name: GigHub Repository
          url: https://github.com/redhat-cop/global-load-balancer-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/global-load-balancer-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/global-load-balancer-operator@sha256:c0b6ac9b03571a74a56b28302d720e8aa4442233f7867d05ed365e022f7cb8f5
        version: 1.4.2
      entries:
      - name: global-load-balancer-operator.v1.4.2
        version: 1.4.2
      - name: global-load-balancer-operator.v1.4.1
        version: 1.4.1
      - name: global-load-balancer-operator.v1.4.0
        version: 1.4.0
      - name: global-load-balancer-operator.v1.3.0
        version: 1.3.0
      - name: global-load-balancer-operator.v1.2.0
        version: 1.2.0
      - name: global-load-balancer-operator.v1.1.0
        version: 1.1.0
      - name: global-load-balancer-operator.v1.0.7
        version: 1.0.7
      - name: global-load-balancer-operator.v1.0.6
        version: 1.0.6
      - name: global-load-balancer-operator.v1.0.5
        version: 1.0.5
      - name: global-load-balancer-operator.v1.0.4
        version: 1.0.4
      - name: global-load-balancer-operator.v1.0.3
        version: 1.0.3
      - name: global-load-balancer-operator.v1.0.2
        version: 1.0.2
      - name: global-load-balancer-operator.v1.0.1
        version: 1.0.1
      - name: global-load-balancer-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: global-load-balancer-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Netscaler
      provider-url: https://www.netscaler.com
    name: netscaler-adm-agent-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netscaler-adm-agent-operator.v141.37.44
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "AdmAgent",
                "metadata": {
                  "name": "admagent"
                },
                "spec": {
                  "accessSecret": "",
                  "admURL": "adm.cloud.com",
                  "affinity": {},
                  "clusterPrefix": "",
                  "core": {
                    "livenessProbe": {
                      "failureThreshold": 5,
                      "initialDelaySeconds": 600,
                      "periodSeconds": 60,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "readinessProbe": {
                      "failureThreshold": 30,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 30,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "100m",
                        "memory": "2000Mi"
                      }
                    }
                  },
                  "fullnameOverride": "",
                  "helper": {
                    "image": "{{ .Values.helper.imageRegistry }}/{{ .Values.helper.imageRepository }}:{{ .Values.helper.imageTag }}",
                    "imageRegistry": "quay.io",
                    "imageRepository": "citrix/adm-agent-helper",
                    "imageTag": "2.0.1",
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "200Mi"
                      }
                    }
                  },
                  "hpa": {
                    "maxReplicas": 10,
                    "minReplicas": 1,
                    "required": false,
                    "targetCPUUtilization": "1000m",
                    "targetMemoryUtilization": "4000Mi"
                  },
                  "image": "{{ .Values.imageRegistry }}/{{ .Values.imageRepository }}:{{ .Values.imageTag }}",
                  "imagePullSecrets": [],
                  "imageRegistry": "quay.io",
                  "imageRepository": "citrix/adm-agent",
                  "imageTag": "14.1-37.44",
                  "kad": {
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "50m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "lic": {
                    "livenessProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 600,
                      "periodSeconds": 20,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "100m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "loginSecret": "",
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "redis": {
                    "livenessProbe": {
                      "failureThreshold": 5,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "readinessProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 20,
                      "periodSeconds": 10,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true
                  },
                  "sharding": {
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-adm-agent-operator@sha256:1c462c2d6161877f87d5eed4888f8d94c733501e67a94757ab0873d983fb3a4a
          createdAt: "2024-11-18T06:40:58Z"
          description: This is an operator to install ADM Agent
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler ADM Agent Operator
            displayName: NetScaler ADM Agent Operator
            kind: AdmAgent
            name: admagents.netscaler.com
            version: v1
        description: Netscaler ADM Agent Operator
        displayName: Netscaler ADM Agent Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        - ADC
        - Kubernetes
        - OpenShift
        - Citrix
        - Controller
        - Observability
        - Metrices
        - adm
        - agent
        - monitoring
        - proxy
        links:
        - name: Netscaler Adm Agent Operator
          url: https://github.com/netscaler/netscaler-helm-charts/blob/master/adm-agent/README.md
        maintainers:
        - email: pavan.belani@cloud.com
          name: Pavan Belani
        - email: swapnil.khanderaokakade@cloud.com
          name: Swapnil Khanderao
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Netscaler
          url: https://www.netscaler.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6b8c101124579dd46086ec83715618dd9a6aebe66176d676242e6c0b0900a53c
        - registry.connect.redhat.com/citrix/netscaler-adm-agent-operator@sha256:1c462c2d6161877f87d5eed4888f8d94c733501e67a94757ab0873d983fb3a4a
        - quay.io/citrix/adm-agent@sha256:45cab90524687c76a6d53f014b55063fa25c8aaeec0261f0f4f0efa6bff42968
        - quay.io/citrix/adm-agent-helper@sha256:f567a0637e8024f9d1674977f6c013741cc0fd12cf78703f8dbe01126561c3c4
        version: 141.37.44
      entries:
      - name: netscaler-adm-agent-operator.v141.37.44
        version: 141.37.44
      - name: netscaler-adm-agent-operator.v141.31.38
        version: 141.31.38
      - name: netscaler-adm-agent-operator.v141.25.33
        version: 141.25.33
      - name: netscaler-adm-agent-operator.v141.23.31
        version: 141.23.31
      - name: netscaler-adm-agent-operator.v141.21.41
        version: 141.21.41
      - name: netscaler-adm-agent-operator.v141.16.39
        version: 141.16.39
      - name: netscaler-adm-agent-operator.v141.13.23
        version: 141.13.23
      name: stable
    defaultChannel: stable
    packageName: netscaler-adm-agent-operator
    provider:
      name: Netscaler
      url: https://www.netscaler.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: pg4k-pgd
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pg4k-pgd.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgd.k8s.enterprisedb.io/v1beta1",
                "kind": "PGDGroup",
                "metadata": {
                  "name": "region-a"
                },
                "spec": {
                  "backup": null,
                  "cnp": {
                    "storage": {
                      "size": "1Gi"
                    },
                    "walStorage": {
                      "size": "1Gi"
                    }
                  },
                  "connectivity": {
                    "dns": {
                      "additional": [
                        {
                          "domain": "alternate.domain"
                        },
                        {
                          "domain": "my.domain",
                          "hostSuffix": "-dc1"
                        }
                      ]
                    },
                    "tls": {
                      "clientCert": {
                        "caCertSecret": "client-ca-key-pair",
                        "certManager": {
                          "spec": {
                            "issuerRef": {
                              "group": "cert-manager.io",
                              "kind": "Issuer",
                              "name": "client-ca-issuer"
                            }
                          }
                        }
                      },
                      "mode": "verify-ca",
                      "serverCert": {
                        "caCertSecret": "server-ca-key-pair",
                        "certManager": {
                          "spec": {
                            "issuerRef": {
                              "group": "cert-manager.io",
                              "kind": "Issuer",
                              "name": "server-ca-issuer"
                            }
                          }
                        }
                      }
                    }
                  },
                  "instances": 2,
                  "pgd": {
                    "parentGroup": {
                      "create": true,
                      "name": "world"
                    }
                  },
                  "proxyInstances": 2,
                  "witness": {
                    "storage": {
                      "size": "1Gi"
                    },
                    "walStorage": {
                      "size": "1Gi"
                    }
                  },
                  "witnessInstances": 1
                }
              },
              {
                "apiVersion": "pgd.k8s.enterprisedb.io/v1beta1",
                "kind": "PGDGroupCleanup",
                "metadata": {
                  "name": "region-b-cleanup"
                },
                "spec": {
                  "executor": "region-a",
                  "target": "region-b"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: docker.enterprisedb.com/k8s_enterprise_pgd/pg4k-pgd-openshift@sha256:5127d913011619d61fd06b27be3c709df2236a4b5c0254fa58632698d58b50fb
          createdAt: "2025-09-26T01:49:29Z"
          description: EDB Postgres Distributed for Kubernetes is an operator designed
            to manage EDB Postgres Distributed (PGD) v5 workloads on Kubernetes both
            in a single Kubernetes cluster or across multiple ones.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterImageCatalog
            name: clusterimagecatalogs.pgd.k8s.enterprisedb.io
            version: v1beta1
          - kind: ImageCatalog
            name: imagecatalogs.pgd.k8s.enterprisedb.io
            version: v1beta1
          - description: Object to clean up a PGD Group
            displayName: PGD Group Cleanup
            kind: PGDGroupCleanup
            name: pgdgroupcleanups.pgd.k8s.enterprisedb.io
            version: v1beta1
          - description: A group of Postgres Clusters
            displayName: PGD Group
            kind: PGDGroup
            name: pgdgroups.pgd.k8s.enterprisedb.io
            version: v1beta1
        description: |
          EDB Postgres Distributed for Kubernetes is an operator designed to manage EDB Postgres Distributed (PGD) v5 workloads on Kubernetes both in a single Kubernetes cluster or across multiple ones.

          EDB Postgres Distributed is an extension for PostgreSQL that provides multi-master replication and data distribution with advanced conflict management, data-loss protection, and throughput up to 5X faster than native logical replication. It enables distributed PostgreSQL clusters with high availability up to five 9s.

          This operator requires access to the EDB private repositories.
        displayName: EDB Postgres Distributed for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - distributed
        - postgres
        - database
        - sql
        - cloud
        - native
        - edb
        - enterprisedb
        - cnp
        - pg4k
        - pgd
        links:
        - name: 'Documentation: EDB Postgres Distributed v5'
          url: https://www.enterprisedb.com/docs/pgd/latest/
        - name: 'Documentation: EDB Postgres Distributed for Kubernetes'
          url: https://www.enterprisedb.com/docs/postgres_distributed_for_kubernetes/latest/
        maintainers:
        - email: sales@enterprisedb.com
          name: Sales
        - email: support@enterprisedb.com
          name: Support
        maturity: alpha
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - docker.enterprisedb.com/k8s_enterprise_pgd/pg4k-pgd-openshift@sha256:5127d913011619d61fd06b27be3c709df2236a4b5c0254fa58632698d58b50fb
        version: 1.1.3
      entries:
      - name: pg4k-pgd.v1.1.3
        version: 1.1.3
      - name: pg4k-pgd.v1.1.2
        version: 1.1.2
      name: stable
    defaultChannel: stable
    packageName: pg4k-pgd
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Mantisnet
      provider-url: https://www.mantisnet.com
    name: tawon-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tawon-operator.v2.29.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Directive",
                "metadata": {
                  "name": "amf-h2"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "amf"
                    }
                  },
                  "container": {
                    "resources": {
                      "limits": {
                        "cpu": "500m"
                      }
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  },
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 8080"
                      },
                      "task": "payload"
                    },
                    {
                      "task": "h2"
                    },
                    {
                      "config": {
                        "addr": "nats://nats:4222",
                        "topic": "hello"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "TopologyAggregator",
                "metadata": {
                  "name": "topology-aggregator"
                },
                "spec": {
                  "messaging": {
                    "nats": {
                      "url": "nats://nats:4222"
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-11-13T01:16:29Z"
          olm.skipRange: '>=2.23.0 <2.29.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Directive allows to define a Tawon task pipeline that will
              be applied to a Kubernetes cluster.
            displayName: Tawon Directive
            kind: Directive
            name: directives.tawon.mantisnet.com
            version: v1alpha1
          - description: TopologyAggregator is the Schema for the topologyaggregators
              API
            displayName: Tawon Topology Aggregator
            kind: TopologyAggregator
            name: topologyaggregators.tawon.mantisnet.com
            version: v1alpha1
        description: Tawon is a cloud-native observability fabric for deeply instrumenting
          microservice-based applications.
        displayName: Tawon
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - networking
        - metrics
        - ebpf
        links:
        - name: Tawon Docs
          url: https://docs.tawon.xmantis.net
        maintainers:
        - email: elliott@mantisnet.com
          name: Elliott Starin
        - email: nicolas@mantisnet.com
          name: Nicolas Paton
        - email: thiago@mantisnet.com
          name: Thiago Navarro
        maturity: alpha
        provider:
          name: Mantisnet
          url: https://www.mantisnet.com
        relatedImages:
        - quay.io/mantisnet/kernel-headers@sha256:242e837af831b6eed98d522b6bae7416928b9fbe451fb6e02ffa678686771068
        - quay.io/mantisnet/kernel-headers@sha256:2aa3bc7871fe0fc0a79ac812d941e55cd0d25a5f0d1d9b3e2435aab6f3d70c67
        - quay.io/mantisnet/tawonctl@sha256:d436ef902bb08d830f64f34fdb86b20d2bd18d86233c8240014f0722c60e5e94
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - quay.io/mantisnet/tawon-operator@sha256:d5669dde46bb30411964baa86844d97721655886769b96809626246a4123caad
        - quay.io/mantisnet/tawon@sha256:83575079fe00ae6e53bb13005b94ef0047cc1974d95015c19e64dd53e13bcb93
        - quay.io/mantisnet/nats@sha256:6afc7ced92181a7c12d709e9693740847297dc5d60ab78cbac7b9b6c859ecdd1
        version: 2.29.1
      entries:
      - name: tawon-operator.v2.29.1
        version: 2.29.1
      name: alpha
    - currentCSV: tawon-operator.v2.39.27-rc6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "ClusterDirective",
                "metadata": {
                  "name": "capture-coredns"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "coredns"
                    }
                  },
                  "debug": {
                    "logLevel": "debug"
                  },
                  "streams": [
                    {
                      "maxage": "6h0m0s",
                      "maxmsgs": 1000,
                      "name": "coredns",
                      "retentionPolicy": "Delete"
                    }
                  ],
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 53"
                      },
                      "task": "capture"
                    },
                    {
                      "config": {
                        "name": "coredns",
                        "type": "stream"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Dashboard",
                "metadata": {
                  "name": "tawon-dashboard"
                },
                "spec": null
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Directive",
                "metadata": {
                  "name": "capture-coredns"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "coredns"
                    }
                  },
                  "debug": {
                    "logLevel": "debug"
                  },
                  "streams": [
                    {
                      "maxage": "6h0m0s",
                      "maxmsgs": 1000,
                      "name": "coredns",
                      "retentionPolicy": "Delete"
                    }
                  ],
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 53"
                      },
                      "task": "capture"
                    },
                    {
                      "config": {
                        "name": "coredns",
                        "type": "stream"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Stream",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "tawon-operator",
                    "app.kubernetes.io/instance": "stream-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "stream",
                    "app.kubernetes.io/part-of": "tawon-operator"
                  },
                  "name": "stream-sample"
                },
                "spec": {
                  "maxage": "6h",
                  "maxbytes": "1G",
                  "maxmsgs": 100000,
                  "store": {
                    "name": "mystreamstore"
                  }
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "StreamStore",
                "metadata": {
                  "name": "mystreamstore"
                },
                "spec": {
                  "defaults": {
                    "maxage": "240h",
                    "maxbytes": "1Gi",
                    "maxsize": 1000000
                  },
                  "limits": {
                    "max": 10000000,
                    "maxage": "2400h",
                    "maxbytes": "2Gi"
                  },
                  "persistentVolumeClaimRetentionPolicy": {
                    "whenDeleted": "Delete"
                  },
                  "volumeClaimTemplate": {
                    "metadata": {
                      "name": "mypvclaim"
                    },
                    "spec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "ssd-csi"
                    }
                  }
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "TopologyAggregator",
                "metadata": {
                  "name": "topology-aggregator"
                },
                "spec": {
                  "messaging": {
                    "nats": {
                      "url": "nats://nats:4222"
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing, Monitoring, Networking
          containerImage: quay.io/mantisnet/tawon-operator@sha256:d8ce5c8966d3590ce5f286d8e778463b85582b970ad7730ddf7828bc68843a90
          createdAt: "2025-03-08T18:58:04Z"
          description: Tawon is a cloud-native observability fabric for deeply instrumenting
            microservice-based applications.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Mantisnet Tawon Operator Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: MantisNet
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterDirective is the Schema for the clusterdirectives
              API.
            displayName: Cluster Tawon Directive
            kind: ClusterDirective
            name: clusterdirectives.tawon.mantisnet.com
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.tawon.mantisnet.com
            version: v1alpha1
          - description: A Directive allows to define a Tawon task pipeline that will
              be applied to a Kubernetes cluster.
            displayName: Tawon Directive
            kind: Directive
            name: directives.tawon.mantisnet.com
            version: v1alpha1
          - description: Stream is a persistent Tawon data stream.
            displayName: Stream
            kind: Stream
            name: streams.tawon.mantisnet.com
            version: v1alpha1
          - description: StreamStore is the Schema for the streamstores API
            displayName: Stream Store
            kind: StreamStore
            name: streamstores.tawon.mantisnet.com
            version: v1alpha1
          - description: TopologyAggregator is the Schema for the topologyaggregators
              API
            displayName: Tawon Topology Aggregator
            kind: TopologyAggregator
            name: topologyaggregators.tawon.mantisnet.com
            version: v1alpha1
        description: Tawon is a cloud-native observability fabric for deeply instrumenting
          microservice-based applications.
        displayName: Tawon
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - networking
        - metrics
        - ebpf
        links:
        - name: Tawon Docs
          url: https://docs.tawon.xmantis.net
        maintainers:
        - email: elliott@mantisnet.com
          name: Elliott Starin
        - email: nicolas@mantisnet.com
          name: Nicolas Paton
        - email: thiago@mantisnet.com
          name: Thiago Navarro
        maturity: alpha
        provider:
          name: Mantisnet
          url: https://www.mantisnet.com
        relatedImages:
        - quay.io/mantisnet/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/mantisnet/kernel-headers@sha256:ca3baeba7ab8ce968c1ee129112532fc8b8d6fb6741e42831b6100f0a2451629
        - quay.io/mantisnet/tawon-operator@sha256:d8ce5c8966d3590ce5f286d8e778463b85582b970ad7730ddf7828bc68843a90
        - quay.io/mantisnet/diagnose@sha256:ab15f4eaeda32f510d81a3cc865b0fd77405b55f6a0c6f285131331dcc38f2ca
        - quay.io/mantisnet/tawonctl@sha256:8bf40ca0cc9984a252363e87d6135e43ac52839113b762399cc54a085af171d8
        - quay.io/mantisnet/sengat@sha256:00c3e1194c2c3e343d16c1543b7ce2a1ee9829f6ddd1c2870785624596b46ebf
        - quay.io/mantisnet/kernel-headers@sha256:9290eb56b186ee0edf65da615c1a8aabc66f7d290168e49ad8764ab8fc2d209a
        - quay.io/mantisnet/tawon@sha256:46e654c6f275be2be604254102aaee9d927859cfd05dfa16b7cb571d36b0f1a2
        - quay.io/mantisnet/tawon-tests@sha256:cd49c24deb7cc9f4b439cb9baa885208dfe91de385cd8c5d312fd08280615c2c
        - quay.io/openshift/origin-oauth-proxy@sha256:28856aae9e481fe763776df713b4b335632645687e6fd2e31619f76d11e0f47c
        - quay.io/mantisnet/tawon@sha256:5a8bf475ff18abacb96f631e4d3ea01c94bef8a38f84c00d1dc4047073b6cc14
        - quay.io/mantisnet/nats@sha256:5f9e7f8f685fd0c1279a3465b61c7215f4600c99d81785dda86a71c966cf829a
        - quay.io/mantisnet/kernel-headers@sha256:5d00bfa9f60c95afcfa363ede59a9625fef9925cd1cb4235327a1de94a0fc632
        version: 2.39.27-rc6
      entries:
      - name: tawon-operator.v2.39.27-rc6
        version: 2.39.27-rc6
      name: candidate-v2
    - currentCSV: tawon-operator.v2.39.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "ClusterDirective",
                "metadata": {
                  "name": "capture-coredns"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "coredns"
                    }
                  },
                  "debug": {
                    "logLevel": "debug"
                  },
                  "streams": [
                    {
                      "maxage": "6h0m0s",
                      "maxmsgs": 1000,
                      "name": "coredns",
                      "retentionPolicy": "Delete"
                    }
                  ],
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 53"
                      },
                      "task": "capture"
                    },
                    {
                      "config": {
                        "name": "coredns",
                        "type": "stream"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Dashboard",
                "metadata": {
                  "name": "tawon-dashboard"
                },
                "spec": null
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Directive",
                "metadata": {
                  "name": "capture-coredns"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "coredns"
                    }
                  },
                  "debug": {
                    "logLevel": "debug"
                  },
                  "streams": [
                    {
                      "maxage": "6h0m0s",
                      "maxmsgs": 1000,
                      "name": "coredns",
                      "retentionPolicy": "Delete"
                    }
                  ],
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 53"
                      },
                      "task": "capture"
                    },
                    {
                      "config": {
                        "name": "coredns",
                        "type": "stream"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Stream",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "tawon-operator",
                    "app.kubernetes.io/instance": "stream-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "stream",
                    "app.kubernetes.io/part-of": "tawon-operator"
                  },
                  "name": "stream-sample"
                },
                "spec": {
                  "maxage": "6h",
                  "maxbytes": "1G",
                  "maxmsgs": 100000,
                  "store": {
                    "name": "mystreamstore"
                  }
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "StreamStore",
                "metadata": {
                  "name": "mystreamstore"
                },
                "spec": {
                  "defaults": {
                    "maxage": "240h",
                    "maxbytes": "1Gi",
                    "maxsize": 1000000
                  },
                  "limits": {
                    "max": 10000000,
                    "maxage": "2400h",
                    "maxbytes": "2Gi"
                  },
                  "persistentVolumeClaimRetentionPolicy": {
                    "whenDeleted": "Delete"
                  },
                  "volumeClaimTemplate": {
                    "metadata": {
                      "name": "mypvclaim"
                    },
                    "spec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "ssd-csi"
                    }
                  }
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "TopologyAggregator",
                "metadata": {
                  "name": "topology-aggregator"
                },
                "spec": {
                  "messaging": {
                    "nats": {
                      "url": "nats://nats:4222"
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing, Monitoring, Networking
          containerImage: quay.io/mantisnet/tawon-operator@sha256:f831abf0f9fb3a4e096743d4ff7356984738da25790e305325e740ca97f0be5b
          createdAt: "2025-07-18T16:24:07Z"
          description: Tawon is a cloud-native observability fabric for deeply instrumenting
            microservice-based applications.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.23.0 <2.39.32'
          operators.openshift.io/valid-subscription: Mantisnet Tawon Operator Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: MantisNet
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterDirective is the Schema for the clusterdirectives
              API.
            displayName: Cluster Tawon Directive
            kind: ClusterDirective
            name: clusterdirectives.tawon.mantisnet.com
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.tawon.mantisnet.com
            version: v1alpha1
          - description: A Directive allows to define a Tawon task pipeline that will
              be applied to a Kubernetes cluster.
            displayName: Tawon Directive
            kind: Directive
            name: directives.tawon.mantisnet.com
            version: v1alpha1
          - description: Stream is a persistent Tawon data stream.
            displayName: Stream
            kind: Stream
            name: streams.tawon.mantisnet.com
            version: v1alpha1
          - description: StreamStore is the Schema for the streamstores API
            displayName: Stream Store
            kind: StreamStore
            name: streamstores.tawon.mantisnet.com
            version: v1alpha1
          - description: TopologyAggregator is the Schema for the topologyaggregators
              API
            displayName: Tawon Topology Aggregator
            kind: TopologyAggregator
            name: topologyaggregators.tawon.mantisnet.com
            version: v1alpha1
        description: Tawon is a cloud-native observability fabric for deeply instrumenting
          microservice-based applications.
        displayName: Tawon
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - networking
        - metrics
        - ebpf
        links:
        - name: Tawon Docs
          url: https://docs.tawon.xmantis.net
        maintainers:
        - email: elliott@mantisnet.com
          name: Elliott Starin
        - email: nicolas@mantisnet.com
          name: Nicolas Paton
        - email: thiago@mantisnet.com
          name: Thiago Navarro
        maturity: alpha
        provider:
          name: Mantisnet
          url: https://www.mantisnet.com
        relatedImages:
        - quay.io/mantisnet/nats@sha256:5f9e7f8f685fd0c1279a3465b61c7215f4600c99d81785dda86a71c966cf829a
        - quay.io/mantisnet/tawon-tests@sha256:294df1580816919502dc592c1c28ffff32b13ee892686f783eff181863fa9d18
        - quay.io/mantisnet/tawon@sha256:1640758505d9b4e0f684ab502ef3344135a270192054e256e7e711209fa51adf
        - quay.io/mantisnet/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/mantisnet/nginx@sha256:c9f91949187fa1c2b4615b88d3acf7902c7e2d4a2557f33ca0cf90164269a7ae
        - quay.io/mantisnet/tawonctl@sha256:88ac81b4fe00d360c5ff127f77456f9e36990e435442a9194d3dbf50b20e7206
        - quay.io/mantisnet/diagnose@sha256:ab15f4eaeda32f510d81a3cc865b0fd77405b55f6a0c6f285131331dcc38f2ca
        - quay.io/mantisnet/tawon@sha256:846d03ada7e6f6648fc74fd0ec9e68c3c6b17e4c155df22829a3db3ca556c395
        - quay.io/mantisnet/kernel-headers@sha256:9290eb56b186ee0edf65da615c1a8aabc66f7d290168e49ad8764ab8fc2d209a
        - quay.io/mantisnet/tawon-operator@sha256:f831abf0f9fb3a4e096743d4ff7356984738da25790e305325e740ca97f0be5b
        - quay.io/mantisnet/alpine@sha256:ec1b05d1eac264d9204a57f4ad9d4dc35e9e756e9fedaea0674aefc7edb1d6a4
        - quay.io/mantisnet/kernel-headers@sha256:ca3baeba7ab8ce968c1ee129112532fc8b8d6fb6741e42831b6100f0a2451629
        - quay.io/mantisnet/kernel-headers@sha256:5d00bfa9f60c95afcfa363ede59a9625fef9925cd1cb4235327a1de94a0fc632
        - quay.io/mantisnet/kernel-headers@sha256:b0975a44cc94bc1d75679f288a9e23a65c2c0b350ce1fb2f5eb421a76a9567c6
        - quay.io/mantisnet/sengat@sha256:483fbce2bacbd95b7458773e2278b936f06a19e65acb8987adffa8403d76e40a
        - quay.io/openshift/origin-oauth-proxy@sha256:28856aae9e481fe763776df713b4b335632645687e6fd2e31619f76d11e0f47c
        version: 2.39.32
      entries:
      - name: tawon-operator.v2.39.32
        version: 2.39.32
      - name: tawon-operator.v2.39.31
        version: 2.39.31
      - name: tawon-operator.v2.39.29
        version: 2.39.29
      - name: tawon-operator.v2.39.28
        version: 2.39.28
      - name: tawon-operator.v2.39.27
        version: 2.39.27
      name: fast-v2
    defaultChannel: fast-v2
    packageName: tawon-operator
    provider:
      name: Mantisnet
      url: https://www.mantisnet.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Perficient Inc
      provider-url: http://www.perficient.com
    name: devopsinabox
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devopsinabox.v0.0.45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "DevSecOpsManager",
                "metadata": {
                  "name": "devsecopsmanager-sample"
                },
                "spec": {
                  "config": {
                    "repoPath": "",
                    "repoUrl": ""
                  },
                  "forwarding": {
                    "fluentd": {
                      "affinity": {},
                      "caFile": "files/default-openshift-logging-fluentd.crt",
                      "image": "registry.redhat.io/openshift4/ose-logging-fluentd:v4.4",
                      "keyFile": "files/default-openshift-logging-fluentd.key",
                      "loglevel": "warn",
                      "nodeSelector": {},
                      "passphrase": "",
                      "persistence": {
                        "accessMode": "ReadWriteOnce",
                        "enabled": false,
                        "size": "10Gi",
                        "storageClass": ""
                      },
                      "port": 24224,
                      "replicas": 2,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "512Mi"
                        }
                      },
                      "sharedkey": "splunkforwarding",
                      "ssl": true,
                      "tolerations": [],
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      }
                    },
                    "preserveDefault": true
                  },
                  "fullnameOverride": "",
                  "git": {
                    "password": "",
                    "server": "https://gitlab.com",
                    "token": "",
                    "username": ""
                  },
                  "nameOverride": "",
                  "prometheus": {
                    "token": ""
                  }
                }
              },
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "ManagedNamespace",
                "metadata": {
                  "name": "managednamespace-sample"
                },
                "spec": {
                  "config": {
                    "repoBranch": "master",
                    "repoPath": "",
                    "repoUrl": ""
                  },
                  "container": {
                    "password": "",
                    "registry": "https://quay.io",
                    "username": ""
                  },
                  "fullnameOverride": "",
                  "git": {
                    "password": "",
                    "server": "https://gitlab.com",
                    "token": "",
                    "username": ""
                  },
                  "manager": {
                    "namespace": ""
                  },
                  "mirrors": {
                    "maven": ""
                  },
                  "nameOverride": ""
                }
              },
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "SpringBootApp",
                "metadata": {
                  "name": "springbootapp-sample"
                },
                "spec": {
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "config": {
                    "repo": ""
                  },
                  "container": {
                    "image": "placeholder",
                    "registry": ""
                  },
                  "fullnameOverride": "",
                  "git": {
                    "branch": "master",
                    "module": "",
                    "url": ""
                  },
                  "management": {
                    "port": 8081
                  },
                  "maven": {
                    "mirror": ""
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "replicaCount": 1,
                  "resources": {},
                  "service": {
                    "port": 8080
                  },
                  "serviceAccount": {
                    "annotations": {}
                  },
                  "version": {
                    "branch": "",
                    "image": "placeholder",
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration,Developer Tools,Application Runtime
          containerImage: quay.io/mrethers/devopsinabox-operator:v0.0.45
          createdAt: 2021-03-03 10:51:09Z
          description: Perficient's DevSecOps Manager For Kubernetes Applications
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/prft-rh/devopsinabox-operator
          support: Matthieu Rethers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DevSecOpsManager
            name: devsecopsmanagers.apimc.com.perficient
            version: v1alpha1
          - kind: ManagedNamespace
            name: managednamespaces.apimc.com.perficient
            version: v1alpha1
          - kind: SpringBootApp
            name: springbootapps.apimc.com.perficient
            version: v1alpha1
        description: "The devops-in-a-box Operator manages the entire lifecycle of
          your applications in Kubernetes. It encapsulates all the best devsecops
          practices gathered by Perficient over the years working on many large projects.
          Use it to migrate existing applications with minimum changes or deploy new
          ones with little to no knowledge of Kubernetes.\n\nOnce installed, the Devops-in-a-box
          operator provides the following features:\n\n* **Create/Destroy**: Easily
          launch a new DevSecOpsManager instance to bootstrap your cluster and start
          deploying new applications with total confidence.\n\n* **Gitops Simplified**:
          Configure your cluster and individual applications through Git and watch
          your changes propagate instantly. Just point to where your code is and let
          the operator do the rest with preconfigured cloud-native CI/CD pipelines.\n\n*
          **Production-Ready**: Automatically register your applications with Openshift
          logging and monitoring systems and enjoy special additions for Spring Boot
          like custom dashboards and searchable logs.\n\n*Contact us on Github for
          installation and usage instruction*\n\n### Other Supported Features\n\n**Blue/Green,
          Canary, A/B Deployments**\n\nRun multiple versions of your applications
          simultaneously and configure traffic rules to target specific groups of
          users using Service Mesh.\n\n**Updates via automated operations**\n\nGet
          the latest enhancements automatically through the Operator Lifecycle Manager
          and upgrade all your running applications at once. "
        displayName: Devops-in-a-box
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devops
        - devsecops
        - spring
        - boot
        - java
        - gitops
        - cicd
        - argo
        links:
        - name: Devops-in-a-box Operator
          url: https://github.com/prft-rh/devopsinabox-operator
        maintainers:
        - email: matthieu.rethers@perficient.com
          name: Matthieu Rethers
        maturity: alpha
        provider:
          name: Perficient Inc
          url: http://www.perficient.com
        relatedImages:
        - quay.io/mrethers/devopsinabox-operator:v0.0.45
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        version: 0.0.45
      entries:
      - name: devopsinabox.v0.0.45
        version: 0.0.45
      name: alpha
    defaultChannel: alpha
    packageName: devopsinabox
    provider:
      name: Perficient Inc
      url: http://www.perficient.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc
      provider-url: ""
    name: bare-metal-event-relay
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bare-metal-event-relay.v4.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "event.redhat-cne.org/v1alpha1",
                "kind": "HardwareEvent",
                "metadata": {
                  "name": "hardwareevent-sample"
                },
                "spec": {
                  "nodeSelector": {
                    "node-role.kubernetes.io/worker": ""
                  },
                  "storageType": "emptyDir",
                  "transportHost": "http://hw-event-publisher-service.openshift-bare-metal-events.svc.cluster.local:9043"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/openshift4/baremetal-hardware-event-proxy-rhel9@sha256:aaaea040164e1fd801e13c74024b2307997cdde954c1a5e35efd90023d2d7376
          createdAt: "2023-02-23"
          description: This software manages the lifecycle of the hw-event-proxy container.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.22.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/hw-event-proxy
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HardwareEvent is the Schema for the hardwareevents API
            displayName: Hardware Event
            kind: HardwareEvent
            name: hardwareevents.event.redhat-cne.org
            version: v1alpha1
        description: |
          ## Funtionality
          Install and manage a hardware event delivery framework.
          ## Installation
          To start using Hardware event proxy, install the operator and then create the following HardwareEvent:
          ```
          apiVersion: "event.redhat-cne.org/v1alpha1"
          kind: "HardwareEvent"
          metadata:
            name: "hardware-event"
          spec:
            nodeSelector:
              node-role.kubernetes.io/worker: ""
            transportHost: "http://hw-event-publisher-service.openshift-bare-metal-events.svc.cluster.local:9043"

          ```
          This operator needs a secret to be created to access Redfish
          Message Registry.
          The secret name and spec the operand expects are under the same namespace
          the operand is deployed on. For example:
          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: redfish-basic-auth
          type: Opaque
          stringData:
            username: <bmc_username>
            password: <bmc_password>
            # BMC host DNS or IP address
            hostaddr: <bmc_host_ip_address>
          ```
        displayName: Bare Metal Event Relay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hardware
        - event
        - cloudevents
        links:
        - name: Hardware Event Proxy Operator
          url: https://github.com/redhat-cne/hw-event-proxy-operator
        maturity: alpha
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:6b7beed3cba9093b3aeb3109f23ec38c20f5568915dff774ad5e9e6bb5db3a58
        - registry.redhat.io/openshift4/bare-metal-event-relay-rhel9-operator@sha256:2219f2f623ee27ef83518d28857995287de0a70cefa193792f28acb352ed8fc0
        - registry.redhat.io/openshift4/baremetal-hardware-event-proxy-rhel9@sha256:aaaea040164e1fd801e13c74024b2307997cdde954c1a5e35efd90023d2d7376
        - registry.redhat.io/openshift4/ose-cloud-event-proxy-rhel9@sha256:c99b448820cf03a33e9237030ac74c1d1ad6a4ce17f1b2840c69b039097c5733
        version: 4.16.1
      entries:
      - name: bare-metal-event-relay.v4.16.1
        version: 4.16.1
      name: "4.16"
    - currentCSV: bare-metal-event-relay.v4.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "event.redhat-cne.org/v1alpha1",
                "kind": "HardwareEvent",
                "metadata": {
                  "name": "hardwareevent-sample"
                },
                "spec": {
                  "nodeSelector": {
                    "node-role.kubernetes.io/worker": ""
                  },
                  "storageType": "emptyDir",
                  "transportHost": "http://hw-event-publisher-service.openshift-bare-metal-events.svc.cluster.local:9043"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/openshift4/baremetal-hardware-event-proxy-rhel9@sha256:aaaea040164e1fd801e13c74024b2307997cdde954c1a5e35efd90023d2d7376
          createdAt: "2023-02-23"
          description: This software manages the lifecycle of the hw-event-proxy container.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.22.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/hw-event-proxy
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HardwareEvent is the Schema for the hardwareevents API
            displayName: Hardware Event
            kind: HardwareEvent
            name: hardwareevents.event.redhat-cne.org
            version: v1alpha1
        description: |
          ## Funtionality
          Install and manage a hardware event delivery framework.
          ## Installation
          To start using Hardware event proxy, install the operator and then create the following HardwareEvent:
          ```
          apiVersion: "event.redhat-cne.org/v1alpha1"
          kind: "HardwareEvent"
          metadata:
            name: "hardware-event"
          spec:
            nodeSelector:
              node-role.kubernetes.io/worker: ""
            transportHost: "http://hw-event-publisher-service.openshift-bare-metal-events.svc.cluster.local:9043"

          ```
          This operator needs a secret to be created to access Redfish
          Message Registry.
          The secret name and spec the operand expects are under the same namespace
          the operand is deployed on. For example:
          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: redfish-basic-auth
          type: Opaque
          stringData:
            username: <bmc_username>
            password: <bmc_password>
            # BMC host DNS or IP address
            hostaddr: <bmc_host_ip_address>
          ```
        displayName: Bare Metal Event Relay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hardware
        - event
        - cloudevents
        links:
        - name: Hardware Event Proxy Operator
          url: https://github.com/redhat-cne/hw-event-proxy-operator
        maturity: alpha
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:6b7beed3cba9093b3aeb3109f23ec38c20f5568915dff774ad5e9e6bb5db3a58
        - registry.redhat.io/openshift4/bare-metal-event-relay-rhel9-operator@sha256:2219f2f623ee27ef83518d28857995287de0a70cefa193792f28acb352ed8fc0
        - registry.redhat.io/openshift4/ose-cloud-event-proxy-rhel9@sha256:c99b448820cf03a33e9237030ac74c1d1ad6a4ce17f1b2840c69b039097c5733
        - registry.redhat.io/openshift4/baremetal-hardware-event-proxy-rhel9@sha256:aaaea040164e1fd801e13c74024b2307997cdde954c1a5e35efd90023d2d7376
        version: 4.16.1
      entries:
      - name: bare-metal-event-relay.v4.16.1
        version: 4.16.1
      name: stable
    defaultChannel: stable
    packageName: bare-metal-event-relay
    provider:
      name: Red Hat Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ocs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocs-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |2-

            [
                {
                    "apiVersion": "ocs.openshift.io/v1",
                    "kind": "StorageCluster",
                    "metadata": {
                        "name": "example-storagecluster",
                        "namespace": "openshift-storage"
                    },
                    "spec": {
                        "manageNodes": false,
                        "monPVCTemplate": {
                            "spec": {
                                "accessModes": [
                                    "ReadWriteOnce"
                                ],
                                "resources": {
                                    "requests": {
                                        "storage": "10Gi"
                                    }
                                },
                                "storageClassName": "gp2-csi"
                            }
                        },
                        "storageDeviceSets": [
                            {
                                "count": 3,
                                "dataPVCTemplate": {
                                    "spec": {
                                        "accessModes": [
                                            "ReadWriteOnce"
                                        ],
                                        "resources": {
                                            "requests": {
                                                "storage": "1Ti"
                                            }
                                        },
                                        "storageClassName": "gp2-csi",
                                        "volumeMode": "Block"
                                    }
                                },
                                "name": "example-deviceset",
                                "placement": {},
                                "portable": true,
                                "resources": {}
                            }
                        ]
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Storage
          containerImage: registry.redhat.io/odf4/ocs-rhel9-operator@sha256:b1a9492bd9b314f86b4e192cf2c83144885ff863d9932e386a41691f50be6193
          description: Red Hat OpenShift Container Storage provides hyperconverged
            storage for applications within an OpenShift cluster.
          external.features.ocs.openshift.io/export-script: |-
            IiIiCkNvcHlyaWdodCAyMDIwIFRoZSBSb29rIEF1dGhvcnMuIEFsbCByaWdodHMgcmVzZXJ2ZWQu
            CgpMaWNlbnNlZCB1bmRlciB0aGUgQXBhY2hlIExpY2Vuc2UsIFZlcnNpb24gMi4wICh0aGUgIkxp
            Y2Vuc2UiKTsKeW91IG1heSBub3QgdXNlIHRoaXMgZmlsZSBleGNlcHQgaW4gY29tcGxpYW5jZSB3
            aXRoIHRoZSBMaWNlbnNlLgpZb3UgbWF5IG9idGFpbiBhIGNvcHkgb2YgdGhlIExpY2Vuc2UgYXQK
            CglodHRwOi8vd3d3LmFwYWNoZS5vcmcvbGljZW5zZXMvTElDRU5TRS0yLjAKClVubGVzcyByZXF1
            aXJlZCBieSBhcHBsaWNhYmxlIGxhdyBvciBhZ3JlZWQgdG8gaW4gd3JpdGluZywgc29mdHdhcmUK
            ZGlzdHJpYnV0ZWQgdW5kZXIgdGhlIExpY2Vuc2UgaXMgZGlzdHJpYnV0ZWQgb24gYW4gIkFTIElT
            IiBCQVNJUywKV0lUSE9VVCBXQVJSQU5USUVTIE9SIENPTkRJVElPTlMgT0YgQU5ZIEtJTkQsIGVp
            dGhlciBleHByZXNzIG9yIGltcGxpZWQuClNlZSB0aGUgTGljZW5zZSBmb3IgdGhlIHNwZWNpZmlj
            IGxhbmd1YWdlIGdvdmVybmluZyBwZXJtaXNzaW9ucyBhbmQKbGltaXRhdGlvbnMgdW5kZXIgdGhl
            IExpY2Vuc2UuCiIiIgoKaW1wb3J0IGVycm5vCmltcG9ydCBzeXMKaW1wb3J0IGpzb24KaW1wb3J0
            IGFyZ3BhcnNlCmltcG9ydCByZQppbXBvcnQgc3VicHJvY2VzcwppbXBvcnQgaG1hYwpmcm9tIGhh
            c2hsaWIgaW1wb3J0IHNoYTEgYXMgc2hhCmZyb20gb3MgaW1wb3J0IGxpbmVzZXAgYXMgTElORVNF
            UApmcm9tIG9zIGltcG9ydCBwYXRoCmZyb20gZW1haWwudXRpbHMgaW1wb3J0IGZvcm1hdGRhdGUK
            aW1wb3J0IHJlcXVlc3RzCmZyb20gcmVxdWVzdHMuYXV0aCBpbXBvcnQgQXV0aEJhc2UKCnB5M2sg
            PSBGYWxzZQppZiBzeXMudmVyc2lvbl9pbmZvLm1ham9yID49IDM6CiAgICBweTNrID0gVHJ1ZQog
            ICAgaW1wb3J0IHVybGxpYi5wYXJzZQogICAgZnJvbSBpcGFkZHJlc3MgaW1wb3J0IGlwX2FkZHJl
            c3MsIElQdjRBZGRyZXNzCgpNb2R1bGVOb3RGb3VuZEVycm9yID0gSW1wb3J0RXJyb3IKCnRyeToK
            ICAgIGltcG9ydCByYWRvcwpleGNlcHQgTW9kdWxlTm90Rm91bmRFcnJvciBhcyBub01vZEVycjoK
            ICAgIHByaW50KGYiRXJyb3I6IHtub01vZEVycn1cbkV4aXRpbmcgdGhlIHNjcmlwdC4uLiIpCiAg
            ICBzeXMuZXhpdCgxKQoKdHJ5OgogICAgaW1wb3J0IHJiZApleGNlcHQgTW9kdWxlTm90Rm91bmRF
            cnJvciBhcyBub01vZEVycjoKICAgIHByaW50KGYiRXJyb3I6IHtub01vZEVycn1cbkV4aXRpbmcg
            dGhlIHNjcmlwdC4uLiIpCiAgICBzeXMuZXhpdCgxKQoKdHJ5OgogICAgIyBmb3IgMi43LngKICAg
            IGZyb20gU3RyaW5nSU8gaW1wb3J0IFN0cmluZ0lPCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9y
            OgogICAgIyBmb3IgMy54CiAgICBmcm9tIGlvIGltcG9ydCBTdHJpbmdJTwoKdHJ5OgogICAgIyBm
            b3IgMi43LngKICAgIGZyb20gdXJscGFyc2UgaW1wb3J0IHVybHBhcnNlCiAgICBmcm9tIHVybGxp
            YiBpbXBvcnQgdXJsZW5jb2RlIGFzIHVybGVuY29kZQpleGNlcHQgTW9kdWxlTm90Rm91bmRFcnJv
            cjoKICAgICMgZm9yIDMueAogICAgZnJvbSB1cmxsaWIucGFyc2UgaW1wb3J0IHVybHBhcnNlCiAg
            ICBmcm9tIHVybGxpYi5wYXJzZSBpbXBvcnQgdXJsZW5jb2RlIGFzIHVybGVuY29kZQoKdHJ5Ogog
            ICAgZnJvbSBiYXNlNjQgaW1wb3J0IGVuY29kZXN0cmluZwpleGNlcHQ6CiAgICBmcm9tIGJhc2U2
            NCBpbXBvcnQgZW5jb2RlYnl0ZXMgYXMgZW5jb2Rlc3RyaW5nCgoKY2xhc3MgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbihFeGNlcHRpb24pOgogICAgcGFzcwoKCiMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMgRHVtbXlSYWRv
            cyAjIyMjIyMjIyMjIyMjIyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjCiMgdGhpcyBpcyBtYWlubHkgZm9yIHRlc3RpbmcgYW5kIGNvdWxkIGJlIHVz
            ZWQgd2hlcmUgJ3JhZG9zJyBpcyBub3QgYXZhaWxhYmxlCgoKY2xhc3MgRHVtbXlSYWRvcyhvYmpl
            Y3QpOgogICAgZGVmIF9faW5pdF9fKHNlbGYpOgogICAgICAgIHNlbGYucmV0dXJuX3ZhbCA9IDAK
            ICAgICAgICBzZWxmLmVycl9tZXNzYWdlID0gIiIKICAgICAgICBzZWxmLnN0YXRlID0gImNvbm5l
            Y3RlZCIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwID0ge30KICAgICAgICBzZWxmLmNtZF9u
            YW1lcyA9IHt9CiAgICAgICAgc2VsZi5faW5pdF9jbWRfb3V0cHV0X21hcCgpCiAgICAgICAgc2Vs
            Zi5kdW1teV9ob3N0X2lwX21hcCA9IHt9CgogICAgZGVmIF9pbml0X2NtZF9vdXRwdXRfbWFwKHNl
            bGYpOgogICAgICAgIGpzb25fZmlsZV9uYW1lID0gInRlc3QtZGF0YS9jZXBoLXN0YXR1cy1vdXQi
            CiAgICAgICAgc2NyaXB0X2RpciA9IHBhdGguYWJzcGF0aChwYXRoLmRpcm5hbWUoX19maWxlX18p
            KQogICAgICAgIGNlcGhfc3RhdHVzX3N0ciA9ICIiCiAgICAgICAgd2l0aCBvcGVuKAogICAgICAg
            ICAgICBwYXRoLmpvaW4oc2NyaXB0X2RpciwganNvbl9maWxlX25hbWUpLCBtb2RlPSJyIiwgZW5j
            b2Rpbmc9IlVURi04IgogICAgICAgICkgYXMganNvbl9maWxlOgogICAgICAgICAgICBjZXBoX3N0
            YXR1c19zdHIgPSBqc29uX2ZpbGUucmVhZCgpCiAgICAgICAgc2VsZi5jbWRfbmFtZXNbImZzIGxz
            Il0gPSAiIiJ7ImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJmcyBscyJ9IiIiCiAgICAgICAg
            c2VsZi5jbWRfbmFtZXNbInF1b3J1bV9zdGF0dXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3Jt
            YXQiOiAianNvbiIsICJwcmVmaXgiOiAicXVvcnVtX3N0YXR1cyJ9IiIiCiAgICAgICAgKQogICAg
            ICAgIHNlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXSA9ICgKICAgICAgICAgICAgIiIieyJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAibWdyIHNlcnZpY2VzIn0iIiIKICAgICAgICApCiAg
            ICAgICAgIyBhbGwgdGhlIGNvbW1hbmRzIGFuZCB0aGVpciBvdXRwdXQKICAgICAgICBzZWxmLmNt
            ZF9vdXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJmcyBscyJdXSA9ICgKICAgICAgICAgICAgIiIi
            W3sibmFtZSI6Im15ZnMiLCJtZXRhZGF0YV9wb29sIjoibXlmcy1tZXRhZGF0YSIsIm1ldGFkYXRh
            X3Bvb2xfaWQiOjIsImRhdGFfcG9vbF9pZHMiOlszXSwiZGF0YV9wb29scyI6WyJteWZzLXJlcGxp
            Y2F0ZWQiXX1dIiIiCiAgICAgICAgKQogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbc2VsZi5j
            bWRfbmFtZXNbInF1b3J1bV9zdGF0dXMiXV0gPSAoCiAgICAgICAgICAgICIiInsiZWxlY3Rpb25f
            ZXBvY2giOjMsInF1b3J1bSI6WzBdLCJxdW9ydW1fbmFtZXMiOlsiYSJdLCJxdW9ydW1fbGVhZGVy
            X25hbWUiOiJhIiwicXVvcnVtX2FnZSI6MTQzODUsImZlYXR1cmVzIjp7InF1b3J1bV9jb24iOiI0
            NTQwMTM4MjkyODM2Njk2MDYzIiwicXVvcnVtX21vbiI6WyJrcmFrZW4iLCJsdW1pbm91cyIsIm1p
            bWljIiwib3NkbWFwLXBydW5lIiwibmF1dGlsdXMiLCJvY3RvcHVzIl19LCJtb25tYXAiOnsiZXBv
            Y2giOjEsImZzaWQiOiJhZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5ZDBmMWMiLCJtb2Rp
            ZmllZCI6IjIwMjAtMDUtMDdUMDM6MzY6MzkuOTE4MDM1WiIsImNyZWF0ZWQiOiIyMDIwLTA1LTA3
            VDAzOjM2OjM5LjkxODAzNVoiLCJtaW5fbW9uX3JlbGVhc2UiOjE1LCJtaW5fbW9uX3JlbGVhc2Vf
            bmFtZSI6Im9jdG9wdXMiLCJmZWF0dXJlcyI6eyJwZXJzaXN0ZW50IjpbImtyYWtlbiIsImx1bWlu
            b3VzIiwibWltaWMiLCJvc2RtYXAtcHJ1bmUiLCJuYXV0aWx1cyIsIm9jdG9wdXMiXSwib3B0aW9u
            YWwiOltdfSwibW9ucyI6W3sicmFuayI6MCwibmFtZSI6ImEiLCJwdWJsaWNfYWRkcnMiOnsiYWRk
            cnZlYyI6W3sidHlwZSI6InYyIiwiYWRkciI6IjEwLjExMC4yMDUuMTc0OjMzMDAiLCJub25jZSI6
            MH0seyJ0eXBlIjoidjEiLCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OSIsIm5vbmNlIjowfV19
            LCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OS8wIiwicHVibGljX2FkZHIiOiIxMC4xMTAuMjA1
            LjE3NDo2Nzg5LzAiLCJwcmlvcml0eSI6MCwid2VpZ2h0IjowfV19fSIiIgogICAgICAgICkKICAg
            ICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXV0g
            PSAoCiAgICAgICAgICAgICIiInsiZGFzaGJvYXJkIjoiaHR0cHM6Ly9jZXBoLWRhc2hib2FyZDo4
            NDQzLyIsInByb21ldGhldXMiOiJodHRwOi8vY2VwaC1kYXNoYm9hcmQtZGI6OTI4My8ifSIiIgog
            ICAgICAgICkKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNh
            cHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMiLCAib3Nk
            IiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0
            YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250
            cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJdLCAiZW50aXR5Ijog
            ImNsaWVudC5oZWFsdGhjaGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRo
            IGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuaGVh
            bHRoY2hlY2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRB
            PT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cyIs
            Im9zZCI6InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5j
            b250cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJ9fV0iIiIKICAg
            ICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIs
            ICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAib3NkIiwgInBy
            b2ZpbGUgcmJkIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtbm9kZSIsICJmb3JtYXQiOiAi
            anNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIi
            W3siZW50aXR5IjoiY2xpZW50LmNzaS1yYmQtbm9kZSIsImtleSI6IkFRQk9nck5lSGJLMUF4QUF1
            YllCZVY4UzFVL0dQenE1U1ZlcTZnPT0iLCJjYXBzIjp7Im1vbiI6InByb2ZpbGUgcmJkLCBhbGxv
            dyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsIm9zZCI6InByb2ZpbGUgcmJkIn19XSIiIgogICAg
            ICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsibW9uIiwg
            InByb2ZpbGUgcmJkLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxs
            b3cgcnciLCAib3NkIiwgInByb2ZpbGUgcmJkIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQt
            cHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNy
            ZWF0ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktcmJkLXByb3Zp
            c2lvbmVyIiwia2V5IjoiQVFCTmdyTmUxZ2V5S3hBQThla1ZpUmRFK2hzczVPd2VZQmt3Tmc9PSIs
            ImNhcHMiOnsibWdyIjoiYWxsb3cgcnciLCJtb24iOiJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFu
            ZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJwcm9maWxlIHJiZCJ9fV0iIiIKICAgICAgICBzZWxm
            LmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyBy
            LCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3Nk
            IiwgImFsbG93IHJ3IHRhZyBjZXBoZnMgKj0qIiwgIm1kcyIsICJhbGxvdyBydyJdLCAiZW50aXR5
            IjogImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4Ijog
            ImF1dGggZ2V0LW9yLWNyZWF0ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVu
            dC5jc2ktY2VwaGZzLW5vZGUiLCJrZXkiOiJBUUJPZ3JOZUVOdW5LeEFBUENtZ0U3UjZHOERjWG5h
            SjFGMzJxZz09IiwiY2FwcyI6eyJtZHMiOiJhbGxvdyBydyIsIm1nciI6ImFsbG93IHJ3IiwibW9u
            IjoiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBy
            dyB0YWcgY2VwaGZzICo9KiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAg
            ICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kICdvc2Qg
            YmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3NkIiwgImFsbG93IHJ3IHRhZyBjZXBo
            ZnMgbWV0YWRhdGE9KiJdLCAiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVy
            IiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgog
            ICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIs
            ImtleSI6IkFRQk9nck5lQUZnY0dCQUF2R3FLT0FEMEQzeHhtVlkwUjkxMmRnPT0iLCJjYXBzIjp7
            Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2Nr
            bGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoifX1dIiIiCiAgICAg
            ICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAi
            YWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdyIiwgImFsbG93IHJ3
            IiwgIm9zZCIsICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiXSwgImVudGl0eSI6ICJj
            bGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVuc2hpZnQtc3RvcmFnZSIsICJmb3JtYXQi
            OiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0g
            IiIiW3siZW50aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0
            b3JhZ2UiLCJrZXkiOiJCUUJPZ3JOZUFGZ2NHQkFBdkdxS09BRDBEM3h4bVZZMFI5MTJkZz09Iiwi
            Y2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29z
            ZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIn19XSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsi
            bW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciIsICJh
            bGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT1teWZzIl0sICJl
            bnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0b3JhZ2Ut
            bXlmcyIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0i
            IiIKICAgICAgICBdID0gIiIiW3siZW50aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9u
            ZXItb3BlbnNoaWZ0LXN0b3JhZ2UtbXlmcyIsImtleSI6IkNRQk9nck5lQUZnY0dCQUF2R3FLT0FE
            MEQzeHhtVlkwUjkxMmRnPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cg
            ciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2Vw
            aGZzIG1ldGFkYXRhPW15ZnMifX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAg
            ICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCBxdW9y
            dW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIiwgImFsbG93IGNvbW1hbmQg
            Y29uZmlnIiwgIm9zZCIsICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFsbG93IHJ3eCBwb29sPWRl
            ZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9ZGVm
            YXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJndy5sb2csIGFsbG93IHgg
            cG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4Il0sICJlbnRpdHkiOiAiY2xpZW50LmhlYWx0
            aGNoZWNrZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0
            ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5oZWFsdGhjaGVja2VyIiwi
            a2V5IjoiQVFERmtiTmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsi
            bW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFu
            ZCB2ZXJzaW9uIiwgIm1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsICJvc2QiOiAicHJvZmls
            ZSByYmQtcmVhZC1vbmx5LCBhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhLCBhbGxvdyBy
            IHBvb2w9LnJndy5yb290LCBhbGxvdyBydyBwb29sPWRlZmF1bHQucmd3LmNvbnRyb2wsIGFsbG93
            IHJ4IHBvb2w9ZGVmYXVsdC5yZ3cubG9nLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0
            cy5pbmRleCJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAi
            IiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMs
            IGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cgY29tbWFuZCBjb25maWciLCAi
            b3NkIiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5j
            b250cm9sLCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1
            bHQucmd3LmJ1Y2tldHMuaW5kZXgiXSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIs
            ICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBjYXBzIn0iIiIKICAgICAgICBdID0g
            IiIiW3siZW50aXR5IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0NWJG
            UkFBVG5kTE5VU1JLcnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIjog
            ImFsbG93IGNvbW1hbmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFs
            bG93IHJ3eCBwb29sPWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFs
            bG93IHJ3IHBvb2w9ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJn
            dy5sb2csIGFsbG93IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4In19XSIiIgogICAg
            ICAgIHNlbGYuY21kX291dHB1dF9tYXBbIiIieyJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            bWdyIHNlcnZpY2VzIn0iIiJdID0gKAogICAgICAgICAgICAiIiJ7ImRhc2hib2FyZCI6ICJodHRw
            Oi8vcm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6NzAwMC8iLCAicHJvbWV0aGV1cyI6
            ICJodHRwOi8vcm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6OTI4My8ifSIiIgogICAg
            ICAgICkKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0
            eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            YXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJ7ImRhc2hib2FyZCI6ICJodHRwOi8vcm9vay1j
            ZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6NzAwMC8iLCAicHJvbWV0aGV1cyI6ICJodHRwOi8v
            cm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6OTI4My8ifSIiIgogICAgICAgIHNlbGYu
            Y21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNsaWVudC5oZWFsdGhj
            aGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldCJ9IiIiCiAgICAg
            ICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5oZWFsdGhjaGVja2VyIiwia2V5IjoiQVFERmti
            TmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjogImFsbG93
            IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFuZCB2ZXJzaW9uIiwg
            Im1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsICJvc2QiOiAicHJvZmlsZSByYmQtcmVhZC1v
            bmx5LCBhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhLCBhbGxvdyByIHBvb2w9LnJndy5y
            b290LCBhbGxvdyBydyBwb29sPWRlZmF1bHQucmd3LmNvbnRyb2wsIGFsbG93IHJ4IHBvb2w9ZGVm
            YXVsdC5yZ3cubG9nLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJ9fV0i
            IiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0eSI6
            ICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJh
            dXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0
            X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtbm9kZSIsICJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJb
            XSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5
            IjogImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRf
            b3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMt
            cHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIK
            ICAgICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAg
            ICAgICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVuc2hp
            ZnQtc3RvcmFnZSIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgog
            ICAgICAgIF0gPSAiIiJbXSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAg
            ICAgICIiInsiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlm
            dC1zdG9yYWdlLW15ZnMiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0i
            IiIKICAgICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAg
            ICAgICAgICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJb
            eyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsImtleSI6IkFRREZrYk5l
            ZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRBPT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIi
            LCAibWdyIjoiYWxsb3cgcnciLCAib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0q
            In19XSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2Fw
            cyI6IFsibW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1n
            ciIsICJhbGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIl0s
            ICJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpz
            b24iLCAicHJlZml4IjogImF1dGggY2FwcyJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmtiTmVmdDViRlJBQVRu
            ZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cgciwgIGFsbG93IGNv
            bW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciI6ImFsbG93IHJ3IiwgIm9zZCI6ImFsbG93IHJ3
            IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFw
            Wyd7ImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJzdGF0dXMifSddID0gY2VwaF9zdGF0dXNf
            c3RyCgogICAgZGVmIHNodXRkb3duKHNlbGYpOgogICAgICAgIHBhc3MKCiAgICBkZWYgZ2V0X2Zz
            aWQoc2VsZik6CiAgICAgICAgcmV0dXJuICJhZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5
            ZDBmMWMiCgogICAgZGVmIGNvbmZfcmVhZF9maWxlKHNlbGYpOgogICAgICAgIHBhc3MKCiAgICBk
            ZWYgY29ubmVjdChzZWxmKToKICAgICAgICBwYXNzCgogICAgZGVmIHBvb2xfZXhpc3RzKHNlbGYs
            IHBvb2xfbmFtZSk6CiAgICAgICAgcmV0dXJuIFRydWUKCiAgICBkZWYgbW9uX2NvbW1hbmQoc2Vs
            ZiwgY21kLCBvdXQpOgogICAgICAgIGpzb25fY21kID0ganNvbi5sb2FkcyhjbWQpCiAgICAgICAg
            anNvbl9jbWRfc3RyID0ganNvbi5kdW1wcyhqc29uX2NtZCwgc29ydF9rZXlzPVRydWUpCiAgICAg
            ICAgY21kX291dHB1dCA9IHNlbGYuY21kX291dHB1dF9tYXBbanNvbl9jbWRfc3RyXQogICAgICAg
            IHJldHVybiBzZWxmLnJldHVybl92YWwsIGNtZF9vdXRwdXQsIHN0cihzZWxmLmVycl9tZXNzYWdl
            LmVuY29kZSgidXRmLTgiKSkKCiAgICBkZWYgX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoc2VsZiwg
            aG9zdF9uYW1lKToKICAgICAgICBpcF9yZWdfeCA9IHJlLmNvbXBpbGUociJcZHsxLDN9LlxkezEs
            M30uXGR7MSwzfS5cZHsxLDN9IikKICAgICAgICAjIGlmIHByb3ZpZGVkIGhvc3QgaXMgZGlyZWN0
            bHkgYW4gSVAgYWRkcmVzcywgcmV0dXJuIHRoZSBzYW1lCiAgICAgICAgaWYgaXBfcmVnX3gubWF0
            Y2goaG9zdF9uYW1lKToKICAgICAgICAgICAgcmV0dXJuIGhvc3RfbmFtZQogICAgICAgIGltcG9y
            dCByYW5kb20KCiAgICAgICAgaG9zdF9pcCA9IHNlbGYuZHVtbXlfaG9zdF9pcF9tYXAuZ2V0KGhv
            c3RfbmFtZSwgIiIpCiAgICAgICAgaWYgbm90IGhvc3RfaXA6CiAgICAgICAgICAgIGhvc3RfaXAg
            PSBmIjE3Mi45LntyYW5kb20ucmFuZGludCgwLCAyNTQpfS57cmFuZG9tLnJhbmRpbnQoMCwgMjU0
            KX0iCiAgICAgICAgICAgIHNlbGYuZHVtbXlfaG9zdF9pcF9tYXBbaG9zdF9uYW1lXSA9IGhvc3Rf
            aXAKICAgICAgICBkZWwgcmFuZG9tCiAgICAgICAgcmV0dXJuIGhvc3RfaXAKCiAgICBAY2xhc3Nt
            ZXRob2QKICAgIGRlZiBSYWRvcyhjb25mZmlsZT1Ob25lKToKICAgICAgICByZXR1cm4gRHVtbXlS
            YWRvcygpCgoKY2xhc3MgUzNBdXRoKEF1dGhCYXNlKToKICAgICIiIkF0dGFjaGVzIEFXUyBBdXRo
            ZW50aWNhdGlvbiB0byB0aGUgZ2l2ZW4gUmVxdWVzdCBvYmplY3QuIiIiCgogICAgc2VydmljZV9i
            YXNlX3VybCA9ICJzMy5hbWF6b25hd3MuY29tIgoKICAgIGRlZiBfX2luaXRfXyhzZWxmLCBhY2Nl
            c3Nfa2V5LCBzZWNyZXRfa2V5LCBzZXJ2aWNlX3VybD1Ob25lKToKICAgICAgICBpZiBzZXJ2aWNl
            X3VybDoKICAgICAgICAgICAgc2VsZi5zZXJ2aWNlX2Jhc2VfdXJsID0gc2VydmljZV91cmwKICAg
            ICAgICBzZWxmLmFjY2Vzc19rZXkgPSBzdHIoYWNjZXNzX2tleSkKICAgICAgICBzZWxmLnNlY3Jl
            dF9rZXkgPSBzdHIoc2VjcmV0X2tleSkKCiAgICBkZWYgX19jYWxsX18oc2VsZiwgcik6CiAgICAg
            ICAgIyBDcmVhdGUgZGF0ZSBoZWFkZXIgaWYgaXQgaXMgbm90IGNyZWF0ZWQgeWV0LgogICAgICAg
            IGlmICJkYXRlIiBub3QgaW4gci5oZWFkZXJzIGFuZCAieC1hbXotZGF0ZSIgbm90IGluIHIuaGVh
            ZGVyczoKICAgICAgICAgICAgci5oZWFkZXJzWyJkYXRlIl0gPSBmb3JtYXRkYXRlKHRpbWV2YWw9
            Tm9uZSwgbG9jYWx0aW1lPUZhbHNlLCB1c2VnbXQ9VHJ1ZSkKICAgICAgICBzaWduYXR1cmUgPSBz
            ZWxmLmdldF9zaWduYXR1cmUocikKICAgICAgICBpZiBweTNrOgogICAgICAgICAgICBzaWduYXR1
            cmUgPSBzaWduYXR1cmUuZGVjb2RlKCJ1dGYtOCIpCiAgICAgICAgci5oZWFkZXJzWyJBdXRob3Jp
            emF0aW9uIl0gPSBmIkFXUyB7c2VsZi5hY2Nlc3Nfa2V5fTp7c2lnbmF0dXJlfSIKICAgICAgICBy
            ZXR1cm4gcgoKICAgIGRlZiBnZXRfc2lnbmF0dXJlKHNlbGYsIHIpOgogICAgICAgIGNhbm9uaWNh
            bF9zdHJpbmcgPSBzZWxmLmdldF9jYW5vbmljYWxfc3RyaW5nKHIudXJsLCByLmhlYWRlcnMsIHIu
            bWV0aG9kKQogICAgICAgIGlmIHB5M2s6CiAgICAgICAgICAgIGtleSA9IHNlbGYuc2VjcmV0X2tl
            eS5lbmNvZGUoInV0Zi04IikKICAgICAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZy5lbmNv
            ZGUoInV0Zi04IikKICAgICAgICBlbHNlOgogICAgICAgICAgICBrZXkgPSBzZWxmLnNlY3JldF9r
            ZXkKICAgICAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZwogICAgICAgIGggPSBobWFjLm5l
            dyhrZXksIG1zZywgZGlnZXN0bW9kPXNoYSkKICAgICAgICByZXR1cm4gZW5jb2Rlc3RyaW5nKGgu
            ZGlnZXN0KCkpLnN0cmlwKCkKCiAgICBkZWYgZ2V0X2Nhbm9uaWNhbF9zdHJpbmcoc2VsZiwgdXJs
            LCBoZWFkZXJzLCBtZXRob2QpOgogICAgICAgIHBhcnNlZHVybCA9IHVybHBhcnNlKHVybCkKICAg
            ICAgICBvYmplY3RrZXkgPSBwYXJzZWR1cmwucGF0aFsxOl0KCiAgICAgICAgYnVja2V0ID0gcGFy
            c2VkdXJsLm5ldGxvY1s6IC1sZW4oc2VsZi5zZXJ2aWNlX2Jhc2VfdXJsKV0KICAgICAgICBpZiBs
            ZW4oYnVja2V0KSA+IDE6CiAgICAgICAgICAgICMgcmVtb3ZlIGxhc3QgZG90CiAgICAgICAgICAg
            IGJ1Y2tldCA9IGJ1Y2tldFs6LTFdCgogICAgICAgIGludGVyZXN0aW5nX2hlYWRlcnMgPSB7ImNv
            bnRlbnQtbWQ1IjogIiIsICJjb250ZW50LXR5cGUiOiAiIiwgImRhdGUiOiAiIn0KICAgICAgICBm
            b3Iga2V5IGluIGhlYWRlcnM6CiAgICAgICAgICAgIGxrID0ga2V5Lmxvd2VyKCkKICAgICAgICAg
            ICAgdHJ5OgogICAgICAgICAgICAgICAgbGsgPSBsay5kZWNvZGUoInV0Zi04IikKICAgICAgICAg
            ICAgZXhjZXB0OgogICAgICAgICAgICAgICAgcGFzcwogICAgICAgICAgICBpZiBoZWFkZXJzW2tl
            eV0gYW5kICgKICAgICAgICAgICAgICAgIGxrIGluIGludGVyZXN0aW5nX2hlYWRlcnMua2V5cygp
            IG9yIGxrLnN0YXJ0c3dpdGgoIngtYW16LSIpCiAgICAgICAgICAgICk6CiAgICAgICAgICAgICAg
            ICBpbnRlcmVzdGluZ19oZWFkZXJzW2xrXSA9IGhlYWRlcnNba2V5XS5zdHJpcCgpCgogICAgICAg
            ICMgSWYgeC1hbXotZGF0ZSBpcyB1c2VkIGl0IHN1cGVyc2VkZXMgdGhlIGRhdGUgaGVhZGVyLgog
            ICAgICAgIGlmIG5vdCBweTNrOgogICAgICAgICAgICBpZiAieC1hbXotZGF0ZSIgaW4gaW50ZXJl
            c3RpbmdfaGVhZGVyczoKICAgICAgICAgICAgICAgIGludGVyZXN0aW5nX2hlYWRlcnNbImRhdGUi
            XSA9ICIiCiAgICAgICAgZWxzZToKICAgICAgICAgICAgaWYgIngtYW16LWRhdGUiIGluIGludGVy
            ZXN0aW5nX2hlYWRlcnM6CiAgICAgICAgICAgICAgICBpbnRlcmVzdGluZ19oZWFkZXJzWyJkYXRl
            Il0gPSAiIgoKICAgICAgICBidWYgPSBmInttZXRob2R9XG4iCiAgICAgICAgZm9yIGtleSBpbiBz
            b3J0ZWQoaW50ZXJlc3RpbmdfaGVhZGVycy5rZXlzKCkpOgogICAgICAgICAgICB2YWwgPSBpbnRl
            cmVzdGluZ19oZWFkZXJzW2tleV0KICAgICAgICAgICAgaWYga2V5LnN0YXJ0c3dpdGgoIngtYW16
            LSIpOgogICAgICAgICAgICAgICAgYnVmICs9IGYie2tleX06e3ZhbH1cbiIKICAgICAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgICAgIGJ1ZiArPSBmInt2YWx9XG4iCgogICAgICAgICMgYXBwZW5k
            IHRoZSBidWNrZXQgaWYgaXQgZXhpc3RzCiAgICAgICAgaWYgYnVja2V0ICE9ICIiOgogICAgICAg
            ICAgICBidWYgKz0gZiIve2J1Y2tldH0iCgogICAgICAgICMgYWRkIHRoZSBvYmplY3RrZXkuIGV2
            ZW4gaWYgaXQgZG9lc24ndCBleGlzdCwgYWRkIHRoZSBzbGFzaAogICAgICAgIGJ1ZiArPSBmIi97
            b2JqZWN0a2V5fSIKCiAgICAgICAgcmV0dXJuIGJ1ZgoKCmNsYXNzIFJhZG9zSlNPTjoKICAgIEVY
            VEVSTkFMX1VTRVJfTkFNRSA9ICJjbGllbnQuaGVhbHRoY2hlY2tlciIKICAgIEVYVEVSTkFMX1JH
            V19BRE1JTl9PUFNfVVNFUl9OQU1FID0gInJndy1hZG1pbi1vcHMtdXNlciIKICAgIEVNUFRZX09V
            VFBVVF9MSVNUID0gIkVtcHR5IG91dHB1dCBsaXN0IgogICAgREVGQVVMVF9SR1dfUE9PTF9QUkVG
            SVggPSAiZGVmYXVsdCIKICAgIERFRkFVTFRfTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUID0gIjky
            ODMiCgogICAgQGNsYXNzbWV0aG9kCiAgICBkZWYgZ2VuX2FyZ19wYXJzZXIoY2xzLCBhcmdzX3Rv
            X3BhcnNlPU5vbmUpOgogICAgICAgIGFyZ1AgPSBhcmdwYXJzZS5Bcmd1bWVudFBhcnNlcigpCgog
            ICAgICAgIGNvbW1vbl9ncm91cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJjb21tb24iKQog
            ICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoIi0tdmVyYm9zZSIsICItdiIsIGFjdGlv
            bj0ic3RvcmVfdHJ1ZSIsIGRlZmF1bHQ9RmFsc2UpCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9h
            cmd1bWVudCgKICAgICAgICAgICAgIi0tY2VwaC1jb25mIiwgIi1jIiwgaGVscD0iUHJvdmlkZSBh
            IGNlcGggY29uZiBmaWxlLiIsIHR5cGU9c3RyCiAgICAgICAgKQogICAgICAgIGNvbW1vbl9ncm91
            cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLWtleXJpbmciLCAiLWsiLCBoZWxwPSJQYXRo
            IHRvIGNlcGgga2V5cmluZyBmaWxlLiIsIHR5cGU9c3RyCiAgICAgICAgKQogICAgICAgIGNvbW1v
            bl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJ1bi1hcy11c2VyIiwKICAgICAg
            ICAgICAgIi11IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgdHlwZT1zdHIs
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIGEgdXNlciBuYW1lIHRvIGNoZWNrIHRoZSBjbHVz
            dGVyJ3MgaGVhbHRoIHN0YXR1cywgbXVzdCBiZSBwcmVmaXhlZCBieSAnY2xpZW50LiciLAogICAg
            ICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1j
            bHVzdGVyLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJL
            dWJlcm5ldGVzIGNsdXN0ZXIgbmFtZShsZWdhY3kgZmxhZyksIE5vdGU6IEVpdGhlciB1c2UgdGhp
            cyBvciAtLWs4cy1jbHVzdGVyLW5hbWUiLAogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAu
            YWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1rOHMtY2x1c3Rlci1uYW1lIiwgZGVmYXVsdD0i
            IiwgaGVscD0iS3ViZXJuZXRlcyBjbHVzdGVyIG5hbWUiCiAgICAgICAgKQogICAgICAgIGNvbW1v
            bl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLW5hbWVzcGFjZSIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIGhlbHA9Ik5hbWVzcGFjZSB3aGVyZSBDZXBoQ2x1
            c3RlciBpcyBydW5uaW5nIiwKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tcmd3LXBvb2wtcHJlZml4IiwgZGVmYXVsdD0iIiwgaGVscD0i
            UkdXIFBvb2wgcHJlZml4IgogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3Vt
            ZW50KAogICAgICAgICAgICAiLS1yZXN0cmljdGVkLWF1dGgtcGVybWlzc2lvbiIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlJlc3RyaWN0IGNlcGhDU0lLZXly
            aW5ncyBhdXRoIHBlcm1pc3Npb25zIHRvIHNwZWNpZmljIHBvb2xzLCBjbHVzdGVyLiIKICAgICAg
            ICAgICAgKyAiTWFuZGF0b3J5IGZsYWdzIHRoYXQgbmVlZCB0byBiZSBzZXQgYXJlIC0tcmJkLWRh
            dGEtcG9vbC1uYW1lLCBhbmQgLS1rOHMtY2x1c3Rlci1uYW1lLiIKICAgICAgICAgICAgKyAiLS1j
            ZXBoZnMtZmlsZXN5c3RlbS1uYW1lIGZsYWcgY2FuIGFsc28gYmUgcGFzc2VkIGluIGNhc2Ugb2Yg
            Y2VwaGZzIHVzZXIgcmVzdHJpY3Rpb24sIHNvIGl0IGNhbiByZXN0cmljdCB1c2VyIHRvIHBhcnRp
            Y3VsYXIgY2VwaGZzIGZpbGVzeXN0ZW0iCiAgICAgICAgICAgICsgInNhbXBsZSBydW46IGBweXRo
            b24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwtY2x1c3Rlci1yZXNvdXJjZXMucHkgLS1jZXBo
            ZnMtZmlsZXN5c3RlbS1uYW1lIG15ZnMgLS1yYmQtZGF0YS1wb29sLW5hbWUgcmVwbGljYXBvb2wg
            LS1rOHMtY2x1c3Rlci1uYW1lIHJvb2tzdG9yYWdlIC0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Np
            b24gdHJ1ZWAiCiAgICAgICAgICAgICsgIk5vdGU6IFJlc3RyaWN0aW5nIHRoZSBjc2ktdXNlcnMg
            cGVyIHBvb2wsIGFuZCBwZXIgY2x1c3RlciB3aWxsIHJlcXVpcmUgY3JlYXRpbmcgbmV3IGNzaS11
            c2VycyBhbmQgbmV3IHNlY3JldHMgZm9yIHRoYXQgY3NpLXVzZXJzLiIKICAgICAgICAgICAgKyAi
            U28gYXBwbHkgdGhlc2Ugc2VjcmV0cyBvbmx5IHRvIG5ldyBgQ29uc3VtZXIgY2x1c3RlcmAgZGVw
            bG95bWVudCB3aGlsZSB1c2luZyB0aGUgc2FtZSBgU291cmNlIGNsdXN0ZXJgLiIsCiAgICAgICAg
            KQogICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXYyLXBv
            cnQtZW5hYmxlIiwKICAgICAgICAgICAgYWN0aW9uPSJzdG9yZV90cnVlIiwKICAgICAgICAgICAg
            ZGVmYXVsdD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iRW5hYmxlIHYyIG1vbiBwb3J0KDMzMDAp
            IGZvciBtb25zIiwKICAgICAgICApCgogICAgICAgIG91dHB1dF9ncm91cCA9IGFyZ1AuYWRkX2Fy
            Z3VtZW50X2dyb3VwKCJvdXRwdXQiKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQo
            CiAgICAgICAgICAgICItLWZvcm1hdCIsCiAgICAgICAgICAgICItdCIsCiAgICAgICAgICAgIGNo
            b2ljZXM9WyJqc29uIiwgImJhc2giXSwKICAgICAgICAgICAgZGVmYXVsdD0ianNvbiIsCiAgICAg
            ICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBvdXRwdXQgZm9ybWF0IChqc29uIHwgYmFzaCkiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1vdXRwdXQiLAogICAgICAgICAgICAiLW8iLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAg
            ICAgICAgICBoZWxwPSJPdXRwdXQgd2lsbCBiZSBzdG9yZWQgaW50byB0aGUgcHJvdmlkZWQgZmls
            ZSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAog
            ICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgQ2VwaCBmaWxlc3lzdGVt
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tY2VwaGZzLW1ldGFkYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIs
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBjZXBoZnMgbWV0YWRh
            dGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAg
            ICAgICAgICAgICItLWNlcGhmcy1kYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9
            IiIsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBjZXBoZnMgZGF0
            YSBwb29sIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAg
            ICAgICAgICAgIi0tcmJkLWRhdGEtcG9vbC1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwK
            ICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRo
            ZSBuYW1lIG9mIHRoZSBSQkQgZGF0YXBvb2wiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3Jv
            dXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1hbGlhcy1yYmQtZGF0YS1wb29sLW5hbWUi
            LAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAg
            ICAgICAgICAgaGVscD0iUHJvdmlkZXMgYW4gYWxpYXMgZm9yIHRoZSAgUkJEIGRhdGEgcG9vbCBu
            YW1lLCBuZWNlc3NhcnkgaWYgYSBzcGVjaWFsIGNoYXJhY3RlciBpcyBwcmVzZW50IGluIHRoZSBw
            b29sIG5hbWUgc3VjaCBhcyBhIHBlcmlvZCBvciB1bmRlcnNjb3JlIiwKICAgICAgICApCiAgICAg
            ICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmd3LWVuZHBvaW50
            IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAg
            ICAgICAgICAgIGhlbHA9IlJBRE9TIEdhdGV3YXkgZW5kcG9pbnQgKGluIGA8SVB2ND46PFBPUlQ+
            YCBvciBgPFtJUHY2XT46PFBPUlQ+YCBvciBgPEZRRE4+OjxQT1JUPmAgZm9ybWF0KSIsCiAgICAg
            ICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJn
            dy10bHMtY2VydC1wYXRoIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVx
            dWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlJBRE9TIEdhdGV3YXkgZW5kcG9pbnQgVExT
            IGNlcnRpZmljYXRlIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVu
            dCgKICAgICAgICAgICAgIi0tcmd3LXNraXAtdGxzIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFs
            c2UsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9Iklnbm9yZSBU
            TFMgY2VydGlmaWNhdGlvbiB2YWxpZGF0aW9uIHdoZW4gYSBzZWxmLXNpZ25lZCBjZXJ0aWZpY2F0
            ZSBpcyBwcm92aWRlZCAoTk9UIFJFQ09NTUVOREVEIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tbW9uaXRvcmluZy1lbmRwb2ludCIs
            CiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAg
            ICAgICAgICBoZWxwPSJDZXBoIE1hbmFnZXIgcHJvbWV0aGV1cyBleHBvcnRlciBlbmRwb2ludHMg
            KGNvbW1hIHNlcGFyYXRlZCBsaXN0IG9mIChmb3JtYXQgYDxJUHY0PmAgb3IgYDxbSVB2Nl0+YCBv
            ciBgPEZRRE4+YCkgZW50cmllcyBvZiBhY3RpdmUgYW5kIHN0YW5kYnkgbWdycykiLAogICAgICAg
            ICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1tb25p
            dG9yaW5nLWVuZHBvaW50LXBvcnQiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAg
            ICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iQ2VwaCBNYW5hZ2VyIHByb21ldGhl
            dXMgZXhwb3J0ZXIgcG9ydCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJn
            dW1lbnQoCiAgICAgICAgICAgICItLXNraXAtbW9uaXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAg
            ICAgICAgIGhlbHA9IkRvIG5vdCBjaGVjayBmb3IgYSBtb25pdG9yaW5nIGVuZHBvaW50IGZvciB0
            aGUgQ2VwaCBjbHVzdGVyIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tcmJkLW1ldGFkYXRhLWVjLXBvb2wtbmFtZSIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBo
            ZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiBlcmFzdXJlIGNvZGVkIFJCRCBtZXRhZGF0YSBwb29s
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tZHJ5LXJ1biIsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFj
            dGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGhlbHA9IkRyeSBydW4gcHJpbnRzIHRoZSBl
            eGVjdXRlZCBjb21tYW5kcyB3aXRob3V0IHJ1bm5pbmcgdGhlbSIsCiAgICAgICAgKQogICAgICAg
            IG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJhZG9zLW5hbWVzcGFj
            ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAog
            ICAgICAgICAgICBoZWxwPSJEaXZpZGVzIGEgcG9vbCBpbnRvIHNlcGFyYXRlIGxvZ2ljYWwgbmFt
            ZXNwYWNlcywgdXNlZCBmb3IgY3JlYXRpbmcgUkJEIFBWQyBpbiBhIENlcGhCbG9ja1Bvb2xSYWRv
            c05hbWVzcGFjZSAoc2hvdWxkIGJlIGxvd2VyIGNhc2UpIiwKICAgICAgICApCiAgICAgICAgb3V0
            cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tc3Vidm9sdW1lLWdyb3VwIiwK
            ICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAg
            ICAgICAgIGhlbHA9InByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBzdWJ2b2x1bWUgZ3JvdXAiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1yZ3ctcmVhbG0tbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJl
            cXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUg
            cmd3LXJlYWxtIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgK
            ICAgICAgICAgICAgIi0tcmd3LXpvbmUtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAg
            ICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUg
            bmFtZSBvZiB0aGUgcmd3LXpvbmUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS1yZ3ctem9uZWdyb3VwLW5hbWUiLAogICAgICAgICAg
            ICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVs
            cD0icHJvdmlkZXMgdGhlIG5hbWUgb2YgdGhlIHJndy16b25lZ3JvdXAiLAogICAgICAgICkKCiAg
            ICAgICAgdXBncmFkZV9ncm91cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJ1cGdyYWRlIikK
            ICAgICAgICB1cGdyYWRlX2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tdXBncmFk
            ZSIsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9
            RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlVwZ3JhZGVzIHRoZSBjZXBoQ1NJS2V5cmluZ3MoRm9y
            IGV4YW1wbGU6IGNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyKSBhbmQgY2xpZW50LmhlYWx0
            aGNoZWNrZXIgY2VwaCB1c2VycyB3aXRoIG5ldyBwZXJtaXNzaW9ucyBuZWVkZWQgZm9yIHRoZSBu
            ZXcgY2x1c3RlciB2ZXJzaW9uIGFuZCBvbGRlciBwZXJtaXNzaW9uIHdpbGwgc3RpbGwgYmUgYXBw
            bGllZC4iCiAgICAgICAgICAgICsgIlNhbXBsZSBydW46IGBweXRob24zIC9ldGMvY2VwaC9jcmVh
            dGUtZXh0ZXJuYWwtY2x1c3Rlci1yZXNvdXJjZXMucHkgLS11cGdyYWRlYCwgdGhpcyB3aWxsIHVw
            Z3JhZGUgYWxsIHRoZSBkZWZhdWx0IGNzaSB1c2Vycyhub24tcmVzdHJpY3RlZCkiCiAgICAgICAg
            ICAgICsgIkZvciByZXN0cmljdGVkIHVzZXJzKEZvciBleGFtcGxlOiBjbGllbnQuY3NpLWNlcGhm
            cy1wcm92aXNpb25lci1vcGVuc2hpZnQtc3RvcmFnZS1teWZzKSwgdXNlcnMgY3JlYXRlZCB1c2lu
            ZyAtLXJlc3RyaWN0ZWQtYXV0aC1wZXJtaXNzaW9uIGZsYWcgbmVlZCB0byBwYXNzIG1hbmRhdG9y
            eSBmbGFncyIKICAgICAgICAgICAgKyAibWFuZGF0b3J5IGZsYWdzOiAnLS1yYmQtZGF0YS1wb29s
            LW5hbWUsIC0tazhzLWNsdXN0ZXItbmFtZSBhbmQgLS1ydW4tYXMtdXNlcicgZmxhZ3Mgd2hpbGUg
            dXBncmFkaW5nIgogICAgICAgICAgICArICJpbiBjYXNlIG9mIGNlcGhmcyB1c2VycyBpZiB5b3Ug
            aGF2ZSBwYXNzZWQgLS1jZXBoZnMtZmlsZXN5c3RlbS1uYW1lIGZsYWcgd2hpbGUgY3JlYXRpbmcg
            dXNlciB0aGVuIHdoaWxlIHVwZ3JhZGluZyBpdCB3aWxsIGJlIG1hbmRhdG9yeSB0b28iCiAgICAg
            ICAgICAgICsgIlNhbXBsZSBydW46IGBweXRob24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwt
            Y2x1c3Rlci1yZXNvdXJjZXMucHkgLS11cGdyYWRlIC0tcmJkLWRhdGEtcG9vbC1uYW1lIHJlcGxp
            Y2Fwb29sIC0tazhzLWNsdXN0ZXItbmFtZSByb29rc3RvcmFnZSAgLS1ydW4tYXMtdXNlciBjbGll
            bnQuY3NpLXJiZC1ub2RlLXJvb2tzdG9yYWdlLXJlcGxpY2Fwb29sYCIKICAgICAgICAgICAgKyAi
            UFM6IEFuIGV4aXN0aW5nIG5vbi1yZXN0cmljdGVkIHVzZXIgY2Fubm90IGJlIGNvbnZlcnRlZCB0
            byBhIHJlc3RyaWN0ZWQgdXNlciBieSB1cGdyYWRpbmcuIgogICAgICAgICAgICArICJVcGdyYWRl
            IGZsYWcgc2hvdWxkIG9ubHkgYmUgdXNlZCB0byBhcHBlbmQgbmV3IHBlcm1pc3Npb25zIHRvIHVz
            ZXJzLCBpdCBzaG91bGRuJ3QgYmUgdXNlZCBmb3IgY2hhbmdpbmcgdXNlciBhbHJlYWR5IGFwcGxp
            ZWQgcGVybWlzc2lvbiwgZm9yIGV4YW1wbGUgeW91IHNob3VsZG4ndCBjaGFuZ2UgaW4gd2hpY2gg
            cG9vbCB1c2VyIGhhcyBhY2Nlc3MiLAogICAgICAgICkKCiAgICAgICAgaWYgYXJnc190b19wYXJz
            ZToKICAgICAgICAgICAgYXNzZXJ0ICgKICAgICAgICAgICAgICAgIHR5cGUoYXJnc190b19wYXJz
            ZSkgPT0gbGlzdAogICAgICAgICAgICApLCAiQXJndW1lbnQgdG8gJ2dlbl9hcmdfcGFyc2VyJyBz
            aG91bGQgYmUgYSBsaXN0IgogICAgICAgIGVsc2U6CiAgICAgICAgICAgIGFyZ3NfdG9fcGFyc2Ug
            PSBzeXMuYXJndlsxOl0KICAgICAgICByZXR1cm4gYXJnUC5wYXJzZV9hcmdzKGFyZ3NfdG9fcGFy
            c2UpCgogICAgZGVmIHZhbGlkYXRlX3JiZF9tZXRhZGF0YV9lY19wb29sX25hbWUoc2VsZik6CiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lOgogICAg
            ICAgICAgICByYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5yYmRf
            bWV0YWRhdGFfZWNfcG9vbF9uYW1lCiAgICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9h
            cmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQoKICAgICAgICAgICAgaWYgcmJkX3Bvb2xfbmFt
            ZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24o
            CiAgICAgICAgICAgICAgICAgICAgIkZsYWcgJy0tcmJkLWRhdGEtcG9vbC1uYW1lJyBzaG91bGQg
            bm90IGJlIGVtcHR5IgogICAgICAgICAgICAgICAgKQoKICAgICAgICAgICAgaWYgcmJkX21ldGFk
            YXRhX2VjX3Bvb2xfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIkZsYWcgJy0tcmJkLW1ldGFkYXRh
            LWVjLXBvb2wtbmFtZScgc2hvdWxkIG5vdCBiZSBlbXB0eSIKICAgICAgICAgICAgICAgICkKCiAg
            ICAgICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAib3NkIGR1bXAiLCAiZm9ybWF0IjogImpz
            b24ifQogICAgICAgICAgICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1v
            bl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBs
            ZW4oanNvbl9vdXQpID09IDA6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJl
            RXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYie2NtZF9qc29uWydwcmVmaXgnXX0gY29t
            bWFuZCBmYWlsZWQuXG4iCiAgICAgICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYg
            cmV0X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAgICAgICAg
            ICApCiAgICAgICAgICAgIG1ldGFkYXRhX3Bvb2xfZXhpc3QsIHBvb2xfZXhpc3QgPSBGYWxzZSwg
            RmFsc2UKCiAgICAgICAgICAgIGZvciBrZXkgaW4ganNvbl9vdXRbInBvb2xzIl06CiAgICAgICAg
            ICAgICAgICAjIGlmIGVyYXN1cmVfY29kZV9wcm9maWxlIGlzIGVtcHR5IGFuZCBwb29sIG5hbWUg
            ZXhpc3RzIHRoZW4gaXQgcmVwbGljYSBwb29sCiAgICAgICAgICAgICAgICBpZiAoCiAgICAgICAg
            ICAgICAgICAgICAga2V5WyJlcmFzdXJlX2NvZGVfcHJvZmlsZSJdID09ICIiCiAgICAgICAgICAg
            ICAgICAgICAgYW5kIGtleVsicG9vbF9uYW1lIl0gPT0gcmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFt
            ZQogICAgICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgICAgICBtZXRhZGF0YV9wb29sX2V4
            aXN0ID0gVHJ1ZQogICAgICAgICAgICAgICAgIyBpZiBlcmFzdXJlX2NvZGVfcHJvZmlsZSBpcyBu
            b3QgZW1wdHkgYW5kIHBvb2wgbmFtZSBleGlzdHMgdGhlbiBpdCBpcyBlYyBwb29sCiAgICAgICAg
            ICAgICAgICBpZiBrZXlbImVyYXN1cmVfY29kZV9wcm9maWxlIl0gYW5kIGtleVsicG9vbF9uYW1l
            Il0gPT0gcmJkX3Bvb2xfbmFtZToKICAgICAgICAgICAgICAgICAgICBwb29sX2V4aXN0ID0gVHJ1
            ZQoKICAgICAgICAgICAgaWYgbm90IG1ldGFkYXRhX3Bvb2xfZXhpc3Q6CiAgICAgICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJQ
            cm92aWRlZCByYmRfZWNfbWV0YWRhdGFfcG9vbCBuYW1lLCIKICAgICAgICAgICAgICAgICAgICBm
            IiB7cmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZX0sIGRvZXMgbm90IGV4aXN0IgogICAgICAgICAg
            ICAgICAgKQogICAgICAgICAgICBpZiBub3QgcG9vbF9leGlzdDoKICAgICAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgZiJQcm92
            aWRlZCByYmRfZGF0YV9wb29sIG5hbWUsIHtyYmRfcG9vbF9uYW1lfSwgZG9lcyBub3QgZXhpc3Qi
            CiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIHJldHVybiByYmRfbWV0YWRhdGFfZWNfcG9v
            bF9uYW1lCgogICAgZGVmIGRyeV9ydW4oc2VsZiwgbXNnKToKICAgICAgICBpZiBzZWxmLl9hcmdf
            cGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHByaW50KCJFeGVjdXRlOiAiICsgIiciICsgbXNn
            ICsgIiciKQoKICAgIGRlZiB2YWxpZGF0ZV9yZ3dfZW5kcG9pbnRfdGxzX2NlcnQoc2VsZik6CiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZ3dfdGxzX2NlcnRfcGF0aDoKICAgICAgICAgICAg
            d2l0aCBvcGVuKHNlbGYuX2FyZ19wYXJzZXIucmd3X3Rsc19jZXJ0X3BhdGgsIGVuY29kaW5nPSJ1
            dGY4IikgYXMgZjoKICAgICAgICAgICAgICAgIGNvbnRlbnRzID0gZi5yZWFkKCkKICAgICAgICAg
            ICAgICAgIHJldHVybiBjb250ZW50cy5yc3RyaXAoKQoKICAgIGRlZiBfY2hlY2tfY29uZmxpY3Rp
            bmdfb3B0aW9ucyhzZWxmKToKICAgICAgICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNlci51cGdyYWRl
            IGFuZCBub3Qgc2VsZi5fYXJnX3BhcnNlci5yYmRfZGF0YV9wb29sX25hbWU6CiAgICAgICAgICAg
            IHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAiRWl0aGVy
            ICctLXVwZ3JhZGUnIG9yICctLXJiZC1kYXRhLXBvb2wtbmFtZSA8cG9vbF9uYW1lPicgc2hvdWxk
            IGJlIHNwZWNpZmllZCIKICAgICAgICAgICAgKQoKICAgIGRlZiBfaW52YWxpZF9lbmRwb2ludChz
            ZWxmLCBlbmRwb2ludF9zdHIpOgogICAgICAgICMgZXh0cmFjdCB0aGUgcG9ydCBieSBnZXR0aW5n
            IHRoZSBsYXN0IHNwbGl0IG9uIGA6YCBkZWxpbWl0ZXIKICAgICAgICB0cnk6CiAgICAgICAgICAg
            IGVuZHBvaW50X3N0cl9pcCwgcG9ydCA9IGVuZHBvaW50X3N0ci5yc3BsaXQoIjoiLCAxKQogICAg
            ICAgIGV4Y2VwdCBWYWx1ZUVycm9yOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJl
            RXhjZXB0aW9uKGYiTm90IGEgcHJvcGVyIGVuZHBvaW50OiB7ZW5kcG9pbnRfc3RyfSIpCgogICAg
            ICAgIHRyeToKICAgICAgICAgICAgaWYgZW5kcG9pbnRfc3RyX2lwWzBdID09ICJbIjoKICAgICAg
            ICAgICAgICAgIGVuZHBvaW50X3N0cl9pcCA9IGVuZHBvaW50X3N0cl9pcFsxIDogbGVuKGVuZHBv
            aW50X3N0cl9pcCkgLSAxXQogICAgICAgICAgICBpcF90eXBlID0gKAogICAgICAgICAgICAgICAg
            IklQdjQiIGlmIHR5cGUoaXBfYWRkcmVzcyhlbmRwb2ludF9zdHJfaXApKSBpcyBJUHY0QWRkcmVz
            cyBlbHNlICJJUHY2IgogICAgICAgICAgICApCiAgICAgICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAg
            ICAgICAgICAgIGlwX3R5cGUgPSAiRlFETiIKICAgICAgICBpZiBub3QgcG9ydC5pc2RpZ2l0KCk6
            CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oZiJQb3J0IG5vdCB2
            YWxpZDoge3BvcnR9IikKICAgICAgICBpbnRQb3J0ID0gaW50KHBvcnQpCiAgICAgICAgaWYgaW50
            UG9ydCA8IDEgb3IgaW50UG9ydCA+IDIqKjE2IC0gMToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0
            aW9uRmFpbHVyZUV4Y2VwdGlvbihmIk91dCBvZiByYW5nZSBwb3J0IG51bWJlcjoge3BvcnR9IikK
            CiAgICAgICAgcmV0dXJuIGlwX3R5cGUKCiAgICBkZWYgZW5kcG9pbnRfZGlhbChzZWxmLCBlbmRw
            b2ludF9zdHIsIGlwX3R5cGUsIHRpbWVvdXQ9MywgY2VydD1Ob25lKToKICAgICAgICAjIGlmIHRo
            ZSAnY2x1c3RlcicgaW5zdGFuY2UgaXMgYSBkdW1teSBvbmUsCiAgICAgICAgIyBkb24ndCB0cnkg
            dG8gcmVhY2ggb3V0IHRvIHRoZSBlbmRwb2ludAogICAgICAgIGlmIGlzaW5zdGFuY2Uoc2VsZi5j
            bHVzdGVyLCBEdW1teVJhZG9zKToKICAgICAgICAgICAgcmV0dXJuICIiLCAiIiwgIiIKICAgICAg
            ICBpZiBpcF90eXBlID09ICJJUHY2IjoKICAgICAgICAgICAgdHJ5OgogICAgICAgICAgICAgICAg
            ZW5kcG9pbnRfc3RyX2lwLCBlbmRwb2ludF9zdHJfcG9ydCA9IGVuZHBvaW50X3N0ci5yc3BsaXQo
            IjoiLCAxKQogICAgICAgICAgICBleGNlcHQgVmFsdWVFcnJvcjoKICAgICAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgZiJOb3Qg
            YSBwcm9wZXIgZW5kcG9pbnQ6IHtlbmRwb2ludF9zdHJ9IgogICAgICAgICAgICAgICAgKQogICAg
            ICAgICAgICBpZiBlbmRwb2ludF9zdHJfaXBbMF0gIT0gIlsiOgogICAgICAgICAgICAgICAgZW5k
            cG9pbnRfc3RyX2lwID0gIlsiICsgZW5kcG9pbnRfc3RyX2lwICsgIl0iCiAgICAgICAgICAgIGVu
            ZHBvaW50X3N0ciA9ICI6Ii5qb2luKFtlbmRwb2ludF9zdHJfaXAsIGVuZHBvaW50X3N0cl9wb3J0
            XSkKCiAgICAgICAgcHJvdG9jb2xzID0gWyJodHRwIiwgImh0dHBzIl0KICAgICAgICByZXNwb25z
            ZV9lcnJvciA9IE5vbmUKICAgICAgICBmb3IgcHJlZml4IGluIHByb3RvY29sczoKICAgICAgICAg
            ICAgdHJ5OgogICAgICAgICAgICAgICAgZXAgPSBmIntwcmVmaXh9Oi8ve2VuZHBvaW50X3N0cn0i
            CiAgICAgICAgICAgICAgICB2ZXJpZnkgPSBOb25lCiAgICAgICAgICAgICAgICAjIElmIHZlcmlm
            eSBpcyBzZXQgdG8gYSBwYXRoIHRvIGEgZGlyZWN0b3J5LAogICAgICAgICAgICAgICAgIyB0aGUg
            ZGlyZWN0b3J5IG11c3QgaGF2ZSBiZWVuIHByb2Nlc3NlZCB1c2luZyB0aGUgY19yZWhhc2ggdXRp
            bGl0eSBzdXBwbGllZCB3aXRoIE9wZW5TU0wuCiAgICAgICAgICAgICAgICBpZiBwcmVmaXggPT0g
            Imh0dHBzIiBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfc2tpcF90bHM6CiAgICAgICAgICAgICAg
            ICAgICAgdmVyaWZ5ID0gRmFsc2UKICAgICAgICAgICAgICAgICAgICByID0gcmVxdWVzdHMuaGVh
            ZChlcCwgdGltZW91dD10aW1lb3V0LCB2ZXJpZnk9RmFsc2UpCiAgICAgICAgICAgICAgICBlbGlm
            IHByZWZpeCA9PSAiaHR0cHMiIGFuZCBjZXJ0OgogICAgICAgICAgICAgICAgICAgIHZlcmlmeSA9
            IGNlcnQKICAgICAgICAgICAgICAgICAgICByID0gcmVxdWVzdHMuaGVhZChlcCwgdGltZW91dD10
            aW1lb3V0LCB2ZXJpZnk9Y2VydCkKICAgICAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAg
            ICAgICAgciA9IHJlcXVlc3RzLmhlYWQoZXAsIHRpbWVvdXQ9dGltZW91dCkKICAgICAgICAgICAg
            ICAgIGlmIHIuc3RhdHVzX2NvZGUgPT0gMjAwOgogICAgICAgICAgICAgICAgICAgIHJldHVybiBw
            cmVmaXgsIHZlcmlmeSwgIiIKICAgICAgICAgICAgZXhjZXB0IEV4Y2VwdGlvbiBhcyBlcnI6CiAg
            ICAgICAgICAgICAgICByZXNwb25zZV9lcnJvciA9IGVycgogICAgICAgICAgICAgICAgY29udGlu
            dWUKICAgICAgICBzeXMuc3RkZXJyLndyaXRlKAogICAgICAgICAgICBmInVuYWJsZSB0byBjb25u
            ZWN0IHRvIGVuZHBvaW50OiB7ZW5kcG9pbnRfc3RyfSwgZmFpbGVkIGVycm9yOiB7cmVzcG9uc2Vf
            ZXJyb3J9IgogICAgICAgICkKICAgICAgICByZXR1cm4gKAogICAgICAgICAgICAiIiwKICAgICAg
            ICAgICAgIiIsCiAgICAgICAgICAgICgiLTEiKSwKICAgICAgICApCgogICAgZGVmIF9faW5pdF9f
            KHNlbGYsIGFyZ19saXN0PU5vbmUpOgogICAgICAgIHNlbGYub3V0X21hcCA9IHt9CiAgICAgICAg
            c2VsZi5fZXhjbHVkZWRfa2V5cyA9IHNldCgpCiAgICAgICAgc2VsZi5fYXJnX3BhcnNlciA9IHNl
            bGYuZ2VuX2FyZ19wYXJzZXIoYXJnc190b19wYXJzZT1hcmdfbGlzdCkKICAgICAgICBzZWxmLl9j
            aGVja19jb25mbGljdGluZ19vcHRpb25zKCkKICAgICAgICBzZWxmLnJ1bl9hc191c2VyID0gc2Vs
            Zi5fYXJnX3BhcnNlci5ydW5fYXNfdXNlcgogICAgICAgIHNlbGYub3V0cHV0X2ZpbGUgPSBzZWxm
            Ll9hcmdfcGFyc2VyLm91dHB1dAogICAgICAgIHNlbGYuY2VwaF9jb25mID0gc2VsZi5fYXJnX3Bh
            cnNlci5jZXBoX2NvbmYKICAgICAgICBzZWxmLmNlcGhfa2V5cmluZyA9IHNlbGYuX2FyZ19wYXJz
            ZXIua2V5cmluZwogICAgICAgICMgaWYgdXNlciBub3QgcHJvdmlkZWQsIGdpdmUgYSBkZWZhdWx0
            IHVzZXIKICAgICAgICBpZiBub3Qgc2VsZi5ydW5fYXNfdXNlciBhbmQgbm90IHNlbGYuX2FyZ19w
            YXJzZXIudXBncmFkZToKICAgICAgICAgICAgc2VsZi5ydW5fYXNfdXNlciA9IHNlbGYuRVhURVJO
            QUxfVVNFUl9OQU1FCiAgICAgICAgaWYgbm90IHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJl
            Zml4IGFuZCBub3Qgc2VsZi5fYXJnX3BhcnNlci51cGdyYWRlOgogICAgICAgICAgICBzZWxmLl9h
            cmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeCA9IHNlbGYuREVGQVVMVF9SR1dfUE9PTF9QUkVGSVgK
            ICAgICAgICBpZiBzZWxmLmNlcGhfY29uZjoKICAgICAgICAgICAga3dhcmdzID0ge30KICAgICAg
            ICAgICAgaWYgc2VsZi5jZXBoX2tleXJpbmc6CiAgICAgICAgICAgICAgICBrd2FyZ3NbImNvbmYi
            XSA9IHsia2V5cmluZyI6IHNlbGYuY2VwaF9rZXlyaW5nfQogICAgICAgICAgICBzZWxmLmNsdXN0
            ZXIgPSByYWRvcy5SYWRvcyhjb25mZmlsZT1zZWxmLmNlcGhfY29uZiwgKiprd2FyZ3MpCiAgICAg
            ICAgZWxzZToKICAgICAgICAgICAgc2VsZi5jbHVzdGVyID0gcmFkb3MuUmFkb3MoKQogICAgICAg
            ICAgICBzZWxmLmNsdXN0ZXIuY29uZl9yZWFkX2ZpbGUoKQogICAgICAgIHNlbGYuY2x1c3Rlci5j
            b25uZWN0KCkKCiAgICBkZWYgc2h1dGRvd24oc2VsZik6CiAgICAgICAgaWYgc2VsZi5jbHVzdGVy
            LnN0YXRlID09ICJjb25uZWN0ZWQiOgogICAgICAgICAgICBzZWxmLmNsdXN0ZXIuc2h1dGRvd24o
            KQoKICAgIGRlZiBnZXRfZnNpZChzZWxmKToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRy
            eV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGggZnNpZCIpCiAgICAg
            ICAgcmV0dXJuIHN0cihzZWxmLmNsdXN0ZXIuZ2V0X2ZzaWQoKSkKCiAgICBkZWYgX2NvbW1vbl9j
            bWRfanNvbl9nZW4oc2VsZiwgY21kX2pzb24pOgogICAgICAgIGNtZCA9IGpzb24uZHVtcHMoY21k
            X2pzb24sIHNvcnRfa2V5cz1UcnVlKQogICAgICAgIHJldF92YWwsIGNtZF9vdXQsIGVycl9tc2cg
            PSBzZWxmLmNsdXN0ZXIubW9uX2NvbW1hbmQoY21kLCBiIiIpCiAgICAgICAgaWYgc2VsZi5fYXJn
            X3BhcnNlci52ZXJib3NlOgogICAgICAgICAgICBwcmludChmIkNvbW1hbmQgSW5wdXQ6IHtjbWR9
            IikKICAgICAgICAgICAgcHJpbnQoCiAgICAgICAgICAgICAgICBmIlJldHVybiBWYWw6IHtyZXRf
            dmFsfVxuQ29tbWFuZCBPdXRwdXQ6IHtjbWRfb3V0fVxuIgogICAgICAgICAgICAgICAgZiJFcnJv
            ciBNZXNzYWdlOiB7ZXJyX21zZ31cbi0tLS0tLS0tLS1cbiIKICAgICAgICAgICAgKQogICAgICAg
            IGpzb25fb3V0ID0ge30KICAgICAgICAjIGlmIHRoZXJlIGlzIG5vIGVycm9yIChpLmU7IHJldF92
            YWwgaXMgWkVSTykgYW5kICdjbWRfb3V0JyBpcyBub3QgZW1wdHkKICAgICAgICAjIHRoZW4gY29u
            dmVydCAnY21kX291dCcgdG8gYSBqc29uIG91dHB1dAogICAgICAgIGlmIHJldF92YWwgPT0gMCBh
            bmQgY21kX291dDoKICAgICAgICAgICAganNvbl9vdXQgPSBqc29uLmxvYWRzKGNtZF9vdXQpCiAg
            ICAgICAgcmV0dXJuIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnCgogICAgZGVmIGdldF9jZXBo
            X2V4dGVybmFsX21vbl9kYXRhKHNlbGYpOgogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAi
            cXVvcnVtX3N0YXR1cyIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgaWYgc2VsZi5fYXJnX3Bh
            cnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKCJjZXBoICIgKyBj
            bWRfanNvblsicHJlZml4Il0pCiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBz
            ZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICMgaWYgdGhlcmUgaXMg
            YW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihq
            c29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgICAgICIncXVvcnVtX3N0YXR1cycgY29tbWFuZCBmYWlsZWQuXG4iCiAg
            ICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxm
            LkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgKQogICAgICAgIHFfbGVhZGVyX25hbWUg
            PSBqc29uX291dFsicXVvcnVtX2xlYWRlcl9uYW1lIl0KICAgICAgICBxX2xlYWRlcl9kZXRhaWxz
            ID0ge30KICAgICAgICBxX2xlYWRlcl9tYXRjaGluZ19saXN0ID0gWwogICAgICAgICAgICBsIGZv
            ciBsIGluIGpzb25fb3V0WyJtb25tYXAiXVsibW9ucyJdIGlmIGxbIm5hbWUiXSA9PSBxX2xlYWRl
            cl9uYW1lCiAgICAgICAgXQogICAgICAgIGlmIGxlbihxX2xlYWRlcl9tYXRjaGluZ19saXN0KSA9
            PSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKCJObyBtYXRj
            aGluZyAnbW9uJyBkZXRhaWxzIGZvdW5kIikKICAgICAgICBxX2xlYWRlcl9kZXRhaWxzID0gcV9s
            ZWFkZXJfbWF0Y2hpbmdfbGlzdFswXQogICAgICAgICMgZ2V0IHRoZSBhZGRyZXNzIHZlY3RvciBv
            ZiB0aGUgcXVvcnVtLWxlYWRlcgogICAgICAgIHFfbGVhZGVyX2FkZHJ2ZWMgPSBxX2xlYWRlcl9k
            ZXRhaWxzLmdldCgicHVibGljX2FkZHJzIiwge30pLmdldCgiYWRkcnZlYyIsIFtdKQogICAgICAg
            ICMgaWYgdGhlIHF1b3J1bS1sZWFkZXIgaGFzIG9ubHkgb25lIGFkZHJlc3MgaW4gdGhlIGFkZHJl
            c3MtdmVjdG9yCiAgICAgICAgIyBhbmQgaXQgaXMgb2YgdHlwZSAndjInIChpZTsgd2l0aCA8SVA+
            OjMzMDApLAogICAgICAgICMgcmFpc2UgYW4gZXhjZXB0aW9uIHRvIG1ha2UgdXNlciBhd2FyZSB0
            aGF0CiAgICAgICAgIyB0aGV5IGhhdmUgdG8gZW5hYmxlICd2MScgKGllOyB3aXRoIDxJUD46Njc4
            OSkgdHlwZSBhcyB3ZWxsCiAgICAgICAgaWYgbGVuKHFfbGVhZGVyX2FkZHJ2ZWMpID09IDEgYW5k
            IHFfbGVhZGVyX2FkZHJ2ZWNbMF1bInR5cGUiXSA9PSAidjIiOgogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgIk9ubHkgJ3YyJyBhZGRy
            ZXNzIHR5cGUgaXMgZW5hYmxlZCwgdXNlciBzaG91bGQgYWxzbyBlbmFibGUgJ3YxJyB0eXBlIGFz
            IHdlbGwiCiAgICAgICAgICAgICkKICAgICAgICBpcF9hZGRyID0gc3RyKHFfbGVhZGVyX2RldGFp
            bHNbInB1YmxpY19hZGRyIl0uc3BsaXQoIi8iKVswXSkKCiAgICAgICAgaWYgc2VsZi5fYXJnX3Bh
            cnNlci52Ml9wb3J0X2VuYWJsZToKICAgICAgICAgICAgaWYgbGVuKHFfbGVhZGVyX2FkZHJ2ZWMp
            ID4gMToKICAgICAgICAgICAgICAgIGlmIHFfbGVhZGVyX2FkZHJ2ZWNbMF1bInR5cGUiXSA9PSAi
            djIiOgogICAgICAgICAgICAgICAgICAgIGlwX2FkZHIgPSBxX2xlYWRlcl9hZGRydmVjWzBdWyJh
            ZGRyIl0KICAgICAgICAgICAgICAgIGVsaWYgcV9sZWFkZXJfYWRkcnZlY1sxXVsidHlwZSJdID09
            ICJ2MiI6CiAgICAgICAgICAgICAgICAgICAgaXBfYWRkciA9IHFfbGVhZGVyX2FkZHJ2ZWNbMV1b
            ImFkZHIiXQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0
            ZSgKICAgICAgICAgICAgICAgICAgICAiJ3YyJyBhZGRyZXNzIHR5cGUgbm90IHByZXNlbnQsIGFu
            ZCAndjItcG9ydC1lbmFibGUnIGZsYWcgaXMgcHJvdmlkZWQiCiAgICAgICAgICAgICAgICApCgog
            ICAgICAgIHJldHVybiBmIntzdHIocV9sZWFkZXJfbmFtZSl9PXtpcF9hZGRyfSIKCiAgICBkZWYg
            X2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoc2VsZiwgaG9zdF9uYW1lLCBwb3J0LCBpcF90eXBlKToK
            ICAgICAgICAjIGlmICdjbHVzdGVyJyBpbnN0YW5jZSBpcyBhIGR1bW15IHR5cGUsCiAgICAgICAg
            IyBjYWxsIHRoZSBkdW1teSBpbnN0YW5jZSdzICJjb252ZXJ0IiBtZXRob2QKICAgICAgICBpZiBu
            b3QgaG9zdF9uYW1lOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KCJFbXB0eSBob3N0bmFtZSBwcm92aWRlZCIpCiAgICAgICAgaWYgaXNpbnN0YW5jZShzZWxmLmNs
            dXN0ZXIsIER1bW15UmFkb3MpOgogICAgICAgICAgICByZXR1cm4gc2VsZi5jbHVzdGVyLl9jb252
            ZXJ0X2hvc3RuYW1lX3RvX2lwKGhvc3RfbmFtZSkKCiAgICAgICAgaWYgaXBfdHlwZSA9PSAiRlFE
            TiI6CiAgICAgICAgICAgICMgY2hlY2sgd2hpY2ggaXAgRlFETiBzaG91bGQgYmUgY29udmVydGVk
            IHRvLCBJUHY0IG9yIElQdjYKICAgICAgICAgICAgIyBjaGVjayB0aGUgaG9zdCBpcCwgdGhlIGVu
            ZHBvaW50IGlwIHR5cGUgd291bGQgYmUgc2ltaWxhciB0byBob3N0IGlwCiAgICAgICAgICAgIGNt
            ZF9qc29uID0geyJwcmVmaXgiOiAib3JjaCBob3N0IGxzIiwgImZvcm1hdCI6ICJqc29uIn0KICAg
            ICAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pz
            b25fZ2VuKGNtZF9qc29uKQogICAgICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1
            bCBhdHRlbXB0LAogICAgICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9
            PSAwOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAg
            ICAgICAgICAgICAgICAgICAiJ29yY2ggaG9zdCBscycgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAg
            ICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYgcmV0X3ZhbCAhPSAwIGVsc2Ugc2Vs
            Zi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGhvc3Rf
            YWRkciA9IGpzb25fb3V0WzBdWyJhZGRyIl0KICAgICAgICAgICAgIyBhZGQgOjgwIHNhbXBsZSBw
            b3J0IGluIGlwX3R5cGUsIGFzIF9pbnZhbGlkX2VuZHBvaW50IGFsc28gdmVyaWZ5IHBvcnQKICAg
            ICAgICAgICAgaG9zdF9pcF90eXBlID0gc2VsZi5faW52YWxpZF9lbmRwb2ludChob3N0X2FkZHIg
            KyAiOjgwIikKICAgICAgICAgICAgaW1wb3J0IHNvY2tldAoKICAgICAgICAgICAgIyBleGFtcGxl
            IG91dHB1dCBbKDxBZGRyZXNzRmFtaWx5LkFGX0lORVQ6IDI+LCA8U29ja2V0S2luZC5TT0NLX1NU
            UkVBTTogMT4sIDYsICcnLCAoJzkzLjE4NC4yMTYuMzQnLCA4MCkpLCAuLi5dCiAgICAgICAgICAg
            ICMgd2UgbmVlZCB0byBnZXQgOTMuMTg0LjIxNi4zNCBzbyBpdCB3b3VsZCBiZSBpcFswXVs0XVsw
            XQogICAgICAgICAgICBpZiBob3N0X2lwX3R5cGUgPT0gIklQdjYiOgogICAgICAgICAgICAgICAg
            aXAgPSBzb2NrZXQuZ2V0YWRkcmluZm8oCiAgICAgICAgICAgICAgICAgICAgaG9zdF9uYW1lLCBw
            b3J0LCBmYW1pbHk9c29ja2V0LkFGX0lORVQ2LCBwcm90bz1zb2NrZXQuSVBQUk9UT19UQ1AKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxpZiBob3N0X2lwX3R5cGUgPT0gIklQdjQiOgog
            ICAgICAgICAgICAgICAgaXAgPSBzb2NrZXQuZ2V0YWRkcmluZm8oCiAgICAgICAgICAgICAgICAg
            ICAgaG9zdF9uYW1lLCBwb3J0LCBmYW1pbHk9c29ja2V0LkFGX0lORVQsIHByb3RvPXNvY2tldC5J
            UFBST1RPX1RDUAogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBkZWwgc29ja2V0CiAgICAg
            ICAgICAgIHJldHVybiBpcFswXVs0XVswXQogICAgICAgIHJldHVybiBob3N0X25hbWUKCiAgICBk
            ZWYgZ2V0X2FjdGl2ZV9hbmRfc3RhbmRieV9tZ3JzKHNlbGYpOgogICAgICAgIGlmIHNlbGYuX2Fy
            Z19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuICIiLCBzZWxmLmRyeV9ydW4oImNl
            cGggc3RhdHVzIikKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQgPSBzZWxmLl9hcmdf
            cGFyc2VyLm1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydAogICAgICAgIG1vbml0b3JpbmdfZW5kcG9p
            bnRfaXBfbGlzdCA9IHNlbGYuX2FyZ19wYXJzZXIubW9uaXRvcmluZ19lbmRwb2ludAogICAgICAg
            IHN0YW5kYnlfbWdycyA9IFtdCiAgICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBf
            bGlzdDoKICAgICAgICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJzdGF0dXMiLCAiZm9ybWF0
            IjogImpzb24ifQogICAgICAgICAgICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYu
            X2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgICAgICMgaWYgdGhlcmUgaXMg
            YW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBs
            ZW4oanNvbl9vdXQpID09IDA6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJl
            RXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICInbWdyIHNlcnZpY2VzJyBjb21tYW5kIGZh
            aWxlZC5cbiIKICAgICAgICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFs
            ICE9IDAgZWxzZSBzZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgICAgICkKICAg
            ICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludCA9ICgKICAgICAgICAgICAgICAgIGpzb25fb3V0
            LmdldCgibWdybWFwIiwge30pLmdldCgic2VydmljZXMiLCB7fSkuZ2V0KCJwcm9tZXRoZXVzIiwg
            IiIpCiAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnQ6
            CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAg
            ICAgICAgICAgICAgICJjYW4ndCBmaW5kIG1vbml0b3JpbmdfZW5kcG9pbnQsIHByb21ldGhldXMg
            bW9kdWxlIG1pZ2h0IG5vdCBiZSBlbmFibGVkLCAiCiAgICAgICAgICAgICAgICAgICAgImVuYWJs
            ZSB0aGUgbW9kdWxlIGJ5IHJ1bm5pbmcgJ2NlcGggbWdyIG1vZHVsZSBlbmFibGUgcHJvbWV0aGV1
            cyciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICMgbm93IGNoZWNrIHRoZSBzdGFuZC1i
            eSBtZ3ItcwogICAgICAgICAgICBzdGFuZGJ5X2FyciA9IGpzb25fb3V0LmdldCgibWdybWFwIiwg
            e30pLmdldCgic3RhbmRieXMiLCBbXSkKICAgICAgICAgICAgZm9yIGVhY2hfc3RhbmRieSBpbiBz
            dGFuZGJ5X2FycjoKICAgICAgICAgICAgICAgIGlmICJuYW1lIiBpbiBlYWNoX3N0YW5kYnkua2V5
            cygpOgogICAgICAgICAgICAgICAgICAgIHN0YW5kYnlfbWdycy5hcHBlbmQoZWFjaF9zdGFuZGJ5
            WyJuYW1lIl0pCiAgICAgICAgICAgIHRyeToKICAgICAgICAgICAgICAgIHBhcnNlZF9lbmRwb2lu
            dCA9IHVybHBhcnNlKG1vbml0b3JpbmdfZW5kcG9pbnQpCiAgICAgICAgICAgIGV4Y2VwdCBWYWx1
            ZUVycm9yOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigK
            ICAgICAgICAgICAgICAgICAgICBmImludmFsaWQgZW5kcG9pbnQ6IHttb25pdG9yaW5nX2VuZHBv
            aW50fSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9p
            cF9saXN0ID0gcGFyc2VkX2VuZHBvaW50Lmhvc3RuYW1lCiAgICAgICAgICAgIGlmIG5vdCBtb25p
            dG9yaW5nX2VuZHBvaW50X3BvcnQ6CiAgICAgICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50
            X3BvcnQgPSBzdHIocGFyc2VkX2VuZHBvaW50LnBvcnQpCgogICAgICAgICMgaWYgbW9uaXRvcmlu
            ZyBlbmRwb2ludCBwb3J0IGlzIG5vdCBzZXQsIHB1dCBhIGRlZmF1bHQgbW9uIHBvcnQKICAgICAg
            ICBpZiBub3QgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0OgogICAgICAgICAgICBtb25pdG9yaW5n
            X2VuZHBvaW50X3BvcnQgPSBzZWxmLkRFRkFVTFRfTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUCgog
            ICAgICAgICMgdXNlciBjb3VsZCBnaXZlIGNvbW1hIGFuZCBzcGFjZSBzZXBhcmF0ZWQgaW5wdXRz
            IChsaWtlIC0tbW9uaXRvcmluZy1lbmRwb2ludD0iPGlwMT4sIDxpcDI+IikKICAgICAgICBtb25p
            dG9yaW5nX2VuZHBvaW50X2lwX2xpc3QgPSBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3QucmVw
            bGFjZSgiLCIsICIgIikKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXQg
            PSBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Quc3BsaXQoKQogICAgICAgICMgaWYgbW9uaXRv
            cmluZy1lbmRwb2ludCBjb3VsZCBub3QgYmUgZm91bmQsIHJhaXNlIGFuIGVycm9yCiAgICAgICAg
            aWYgbGVuKG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdF9zcGxpdCkgPT0gMDoKICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigiTm8gJ21vbml0b3JpbmctZW5kcG9p
            bnQnIGZvdW5kIikKICAgICAgICAjIGZpcnN0IGlwIGlzIHRyZWF0ZWQgYXMgdGhlIG1haW4gbW9u
            aXRvcmluZy1lbmRwb2ludAogICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfaXAgPSBtb25pdG9y
            aW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXRbMF0KICAgICAgICAjIHJlc3Qgb2YgdGhlIGlwLXMg
            YXJlIGFkZGVkIHRvIHRoZSAnc3RhbmRieV9tZ3JzJyBsaXN0CiAgICAgICAgc3RhbmRieV9tZ3Jz
            LmV4dGVuZChtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXRbMTpdKQogICAgICAgIGZh
            aWxlZF9pcCA9IG1vbml0b3JpbmdfZW5kcG9pbnRfaXAKCiAgICAgICAgbW9uaXRvcmluZ19lbmRw
            b2ludCA9ICI6Ii5qb2luKAogICAgICAgICAgICBbbW9uaXRvcmluZ19lbmRwb2ludF9pcCwgbW9u
            aXRvcmluZ19lbmRwb2ludF9wb3J0XQogICAgICAgICkKICAgICAgICBpcF90eXBlID0gc2VsZi5f
            aW52YWxpZF9lbmRwb2ludChtb25pdG9yaW5nX2VuZHBvaW50KQogICAgICAgIHRyeToKICAgICAg
            ICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcCA9IHNlbGYuX2NvbnZlcnRfaG9zdG5hbWVfdG9f
            aXAoCiAgICAgICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X2lwLCBtb25pdG9yaW5nX2Vu
            ZHBvaW50X3BvcnQsIGlwX3R5cGUKICAgICAgICAgICAgKQogICAgICAgICAgICAjIGNvbGxlY3Qg
            YWxsIHRoZSAnc3RhbmQtYnknIG1nciBpcHMKICAgICAgICAgICAgbWdyX2lwcyA9IFtdCiAgICAg
            ICAgICAgIGZvciBlYWNoX3N0YW5kYnlfbWdyIGluIHN0YW5kYnlfbWdyczoKICAgICAgICAgICAg
            ICAgIGZhaWxlZF9pcCA9IGVhY2hfc3RhbmRieV9tZ3IKICAgICAgICAgICAgICAgIG1ncl9pcHMu
            YXBwZW5kKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAo
            CiAgICAgICAgICAgICAgICAgICAgICAgIGVhY2hfc3RhbmRieV9tZ3IsIG1vbml0b3JpbmdfZW5k
            cG9pbnRfcG9ydCwgaXBfdHlwZQogICAgICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAg
            ICkKICAgICAgICBleGNlcHQ6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNl
            cHRpb24oCiAgICAgICAgICAgICAgICBmIkNvbnZlcnNpb24gb2YgaG9zdDoge2ZhaWxlZF9pcH0g
            dG8gSVAgZmFpbGVkLiAiCiAgICAgICAgICAgICAgICAiUGxlYXNlIGVudGVyIHRoZSBJUCBhZGRy
            ZXNzZXMgb2YgYWxsIHRoZSBjZXBoLW1ncnMgd2l0aCB0aGUgJy0tbW9uaXRvcmluZy1lbmRwb2lu
            dCcgZmxhZyIKICAgICAgICAgICAgKQoKICAgICAgICBfLCBfLCBlcnIgPSBzZWxmLmVuZHBvaW50
            X2RpYWwobW9uaXRvcmluZ19lbmRwb2ludCwgaXBfdHlwZSkKICAgICAgICBpZiBlcnIgPT0gIi0x
            IjoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihlcnIpCiAgICAg
            ICAgIyBhZGQgdGhlIHZhbGlkYXRlZCBhY3RpdmUgbWdyIElQIGludG8gdGhlIGZpcnN0IGluZGV4
            CiAgICAgICAgbWdyX2lwcy5pbnNlcnQoMCwgbW9uaXRvcmluZ19lbmRwb2ludF9pcCkKICAgICAg
            ICBhbGxfbWdyX2lwc19zdHIgPSAiLCIuam9pbihtZ3JfaXBzKQogICAgICAgIHJldHVybiBhbGxf
            bWdyX2lwc19zdHIsIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydAoKICAgIGRlZiBjaGVja191c2Vy
            X2V4aXN0KHNlbGYsIHVzZXIpOgogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAiYXV0aCBn
            ZXQiLCAiZW50aXR5IjogZiJ7dXNlcn0iLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIHJldF92
            YWwsIGpzb25fb3V0LCBfID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAg
            ICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICBy
            ZXR1cm4gIiIKICAgICAgICByZXR1cm4gc3RyKGpzb25fb3V0WzBdWyJrZXkiXSkKCiAgICBkZWYg
            Z2V0X2NlcGhmc19wcm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50
            aXR5ID0gImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIgogICAgICAgIGNhcHMgPSB7CiAg
            ICAgICAgICAgICJtb24iOiAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCci
            LAogICAgICAgICAgICAibWdyIjogImFsbG93IHJ3IiwKICAgICAgICAgICAgIm9zZCI6ICJhbGxv
            dyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiLAogICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9h
            cmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uOgogICAgICAgICAgICBrOHNfY2x1
            c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1lCiAgICAgICAgICAg
            IGlmIGs4c19jbHVzdGVyX25hbWUgPT0gIiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRp
            b25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJrOHNfY2x1c3Rlcl9uYW1l
            IG5vdCBmb3VuZCwgcGxlYXNlIHNldCB0aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW0gPSBzZWxmLl9hcmdf
            cGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICAgICAgaWYgY2VwaGZzX2ZpbGVz
            eXN0ZW0gPT0gIiI6CiAgICAgICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1
            c3Rlcl9uYW1lfSIKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYi
            e2VudGl0eX0te2s4c19jbHVzdGVyX25hbWV9LXtjZXBoZnNfZmlsZXN5c3RlbX0iCiAgICAgICAg
            ICAgICAgICBjYXBzWyJvc2QiXSA9IGYiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT17Y2Vw
            aGZzX2ZpbGVzeXN0ZW19IgoKICAgICAgICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVmIGdl
            dF9jZXBoZnNfbm9kZV9jYXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gImNs
            aWVudC5jc2ktY2VwaGZzLW5vZGUiCiAgICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6
            ICJhbGxvdyByLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJt
            Z3IiOiAiYWxsb3cgcnciLAogICAgICAgICAgICAib3NkIjogImFsbG93IHJ3IHRhZyBjZXBoZnMg
            Kj0qIiwKICAgICAgICAgICAgIm1kcyI6ICJhbGxvdyBydyIsCiAgICAgICAgfQogICAgICAgIGlm
            IHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAg
            IGs4c19jbHVzdGVyX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAg
            ICAgICAgICAgaWYgazhzX2NsdXN0ZXJfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIms4c19jbHVz
            dGVyX25hbWUgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1rOHMtY2x1c3Rlci1uYW1lJyBm
            bGFnIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBjZXBoZnNfZmlsZXN5c3RlbSA9IHNl
            bGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgICAgICBpZiBjZXBo
            ZnNfZmlsZXN5c3RlbSA9PSAiIjoKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0t
            e2s4c19jbHVzdGVyX25hbWV9IgogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgZW50
            aXR5ID0gZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0te2NlcGhmc19maWxlc3lzdGVtfSIK
            ICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJdID0gZiJhbGxvdyBydyB0YWcgY2VwaGZzICo9e2Nl
            cGhmc19maWxlc3lzdGVtfSIKCiAgICAgICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBn
            ZXRfZW50aXR5KAogICAgICAgIHNlbGYsCiAgICAgICAgZW50aXR5LAogICAgICAgIHJiZF9wb29s
            X25hbWUsCiAgICAgICAgYWxpYXNfcmJkX3Bvb2xfbmFtZSwKICAgICAgICBrOHNfY2x1c3Rlcl9u
            YW1lLAogICAgICAgIHJhZG9zX25hbWVzcGFjZSwKICAgICk6CiAgICAgICAgaWYgKAogICAgICAg
            ICAgICByYmRfcG9vbF9uYW1lLmNvdW50KCIuIikgIT0gMAogICAgICAgICAgICBvciByYmRfcG9v
            bF9uYW1lLmNvdW50KCJfIikgIT0gMAogICAgICAgICAgICBvciBhbGlhc19yYmRfcG9vbF9uYW1l
            ICE9ICIiCiAgICAgICAgICAgICMgY2hlY2tpbmcgYWxpYXNfcmJkX3Bvb2xfbmFtZSBpcyBub3Qg
            ZW1wdHkgYXMgdGhlcmUgbWF5YmUgYSBzcGVjaWFsIGNoYXJhY3RlciB1c2VkIG90aGVyIHRoYW4g
            LiBvciBfCiAgICAgICAgKToKICAgICAgICAgICAgaWYgYWxpYXNfcmJkX3Bvb2xfbmFtZSA9PSAi
            IjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAg
            ICAgICAgICAgICAgICAgInBsZWFzZSBzZXQgdGhlICctLWFsaWFzLXJiZC1kYXRhLXBvb2wtbmFt
            ZScgZmxhZyBhcyB0aGUgcmJkIGRhdGEgcG9vbCBuYW1lIGNvbnRhaW5zICcuJyBvciAnXyciCiAg
            ICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmICgKICAgICAgICAgICAgICAgIGFsaWFzX3Ji
            ZF9wb29sX25hbWUuY291bnQoIi4iKSAhPSAwCiAgICAgICAgICAgICAgICBvciBhbGlhc19yYmRf
            cG9vbF9uYW1lLmNvdW50KCJfIikgIT0gMAogICAgICAgICAgICApOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJy0t
            YWxpYXMtcmJkLWRhdGEtcG9vbC1uYW1lJyBmbGFnIHZhbHVlIHNob3VsZCBub3QgY29udGFpbiAn
            Licgb3IgJ18nIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRp
            dHl9LXtrOHNfY2x1c3Rlcl9uYW1lfS17YWxpYXNfcmJkX3Bvb2xfbmFtZX0iCiAgICAgICAgZWxz
            ZToKICAgICAgICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0te3Ji
            ZF9wb29sX25hbWV9IgoKICAgICAgICBpZiByYWRvc19uYW1lc3BhY2U6CiAgICAgICAgICAgIGVu
            dGl0eSA9IGYie2VudGl0eX0te3JhZG9zX25hbWVzcGFjZX0iCiAgICAgICAgcmV0dXJuIGVudGl0
            eQoKICAgIGRlZiBnZXRfcmJkX3Byb3Zpc2lvbmVyX2NhcHNfYW5kX2VudGl0eShzZWxmKToKICAg
            ICAgICBlbnRpdHkgPSAiY2xpZW50LmNzaS1yYmQtcHJvdmlzaW9uZXIiCiAgICAgICAgY2FwcyA9
            IHsKICAgICAgICAgICAgIm1vbiI6ICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJs
            b2NrbGlzdCciLAogICAgICAgICAgICAibWdyIjogImFsbG93IHJ3IiwKICAgICAgICAgICAgIm9z
            ZCI6ICJwcm9maWxlIHJiZCIsCiAgICAgICAgfQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIu
            cmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBz
            ZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgICAgICBhbGlhc19yYmRf
            cG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5hbGlhc19yYmRfZGF0YV9wb29sX25hbWUKICAg
            ICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJf
            bmFtZQogICAgICAgICAgICByYWRvc19uYW1lc3BhY2UgPSBzZWxmLl9hcmdfcGFyc2VyLnJhZG9z
            X25hbWVzcGFjZQogICAgICAgICAgICBpZiByYmRfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAg
            ICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAg
            ICAibWFuZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1yYmQtZGF0YS1w
            b29sLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmIGs4c19jbHVz
            dGVyX25hbWUgPT0gIiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhj
            ZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBs
            ZWFzZSBzZXQgdGhlICctLWs4cy1jbHVzdGVyLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgIGVudGl0eSA9IHNlbGYuZ2V0X2VudGl0eSgKICAgICAgICAgICAgICAgIGVu
            dGl0eSwKICAgICAgICAgICAgICAgIHJiZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBhbGlh
            c19yYmRfcG9vbF9uYW1lLAogICAgICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAg
            ICAgICAgICAgIHJhZG9zX25hbWVzcGFjZSwKICAgICAgICAgICAgKQogICAgICAgICAgICBpZiBy
            YWRvc19uYW1lc3BhY2UgIT0gIiI6CiAgICAgICAgICAgICAgICBjYXBzWyJvc2QiXSA9ICgKICAg
            ICAgICAgICAgICAgICAgICBmInByb2ZpbGUgcmJkIHBvb2w9e3JiZF9wb29sX25hbWV9IG5hbWVz
            cGFjZT17cmFkb3NfbmFtZXNwYWNlfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxz
            ZToKICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJdID0gZiJwcm9maWxlIHJiZCBwb29sPXtyYmRf
            cG9vbF9uYW1lfSIKCiAgICAgICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfcmJk
            X25vZGVfY2Fwc19hbmRfZW50aXR5KHNlbGYpOgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3Np
            LXJiZC1ub2RlIgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAgICJtb24iOiAicHJvZmlsZSBy
            YmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwKICAgICAgICAgICAgIm9zZCI6ICJw
            cm9maWxlIHJiZCIsCiAgICAgICAgfQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJp
            Y3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9h
            cmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9u
            YW1lID0gc2VsZi5fYXJnX3BhcnNlci5hbGlhc19yYmRfZGF0YV9wb29sX25hbWUKICAgICAgICAg
            ICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQog
            ICAgICAgICAgICByYWRvc19uYW1lc3BhY2UgPSBzZWxmLl9hcmdfcGFyc2VyLnJhZG9zX25hbWVz
            cGFjZQogICAgICAgICAgICBpZiByYmRfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAibWFu
            ZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1yYmQtZGF0YS1wb29sLW5h
            bWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmIGs4c19jbHVzdGVyX25h
            bWUgPT0gIiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KAogICAgICAgICAgICAgICAgICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBz
            ZXQgdGhlICctLWs4cy1jbHVzdGVyLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAg
            ICAgICAgIGVudGl0eSA9IHNlbGYuZ2V0X2VudGl0eSgKICAgICAgICAgICAgICAgIGVudGl0eSwK
            ICAgICAgICAgICAgICAgIHJiZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBhbGlhc19yYmRf
            cG9vbF9uYW1lLAogICAgICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAgICAgICAg
            ICAgIHJhZG9zX25hbWVzcGFjZSwKICAgICAgICAgICAgKQogICAgICAgICAgICBpZiByYWRvc19u
            YW1lc3BhY2UgIT0gIiI6CiAgICAgICAgICAgICAgICBjYXBzWyJvc2QiXSA9ICgKICAgICAgICAg
            ICAgICAgICAgICBmInByb2ZpbGUgcmJkIHBvb2w9e3JiZF9wb29sX25hbWV9IG5hbWVzcGFjZT17
            cmFkb3NfbmFtZXNwYWNlfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxzZToKICAg
            ICAgICAgICAgICAgIGNhcHNbIm9zZCJdID0gZiJwcm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9u
            YW1lfSIKCiAgICAgICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfZGVmYXVsdFVz
            ZXJfY2Fwc19hbmRfZW50aXR5KHNlbGYpOgogICAgICAgIGVudGl0eSA9IHNlbGYucnVuX2FzX3Vz
            ZXIKICAgICAgICBjYXBzID0gewogICAgICAgICAgICAibW9uIjogImFsbG93IHIsIGFsbG93IGNv
            bW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFuZCB2ZXJzaW9uIiwKICAgICAgICAgICAg
            Im1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsCiAgICAgICAgICAgICJvc2QiOiBmInByb2Zp
            bGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9e3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bv
            b2xfcHJlZml4fS5yZ3cubWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9v
            bD17c2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5jb250cm9sLCBhbGxvdyBy
            eCBwb29sPXtzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmxvZywgYWxsb3cg
            eCBwb29sPXtzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmJ1Y2tldHMuaW5k
            ZXgiLAogICAgICAgIH0KCiAgICAgICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRf
            Y2Fwc19hbmRfZW50aXR5KHNlbGYsIHVzZXJfbmFtZSk6CiAgICAgICAgaWYgImNsaWVudC5jc2kt
            Y2VwaGZzLXByb3Zpc2lvbmVyIiBpbiB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQu
            Y3NpLWNlcGhmcy1wcm92aXNpb25lciIgIT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAg
            ICAgcmV0dXJuIHNlbGYuZ2V0X2NlcGhmc19wcm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoKQog
            ICAgICAgIGlmICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiBpbiB1c2VyX25hbWU6CiAgICAgICAg
            ICAgIGlmICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiAhPSB1c2VyX25hbWU6CiAgICAgICAgICAg
            ICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQog
            ICAgICAgICAgICByZXR1cm4gc2VsZi5nZXRfY2VwaGZzX25vZGVfY2Fwc19hbmRfZW50aXR5KCkK
            ICAgICAgICBpZiAiY2xpZW50LmNzaS1yYmQtcHJvdmlzaW9uZXIiIGluIHVzZXJfbmFtZToKICAg
            ICAgICAgICAgaWYgImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiAhPSB1c2VyX25hbWU6CiAg
            ICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9u
            ID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5nZXRfcmJkX3Byb3Zpc2lvbmVyX2NhcHNf
            YW5kX2VudGl0eSgpCiAgICAgICAgaWYgImNsaWVudC5jc2ktcmJkLW5vZGUiIGluIHVzZXJfbmFt
            ZToKICAgICAgICAgICAgaWYgImNsaWVudC5jc2ktcmJkLW5vZGUiICE9IHVzZXJfbmFtZToKICAg
            ICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24g
            PSBUcnVlCiAgICAgICAgICAgIHJldHVybiBzZWxmLmdldF9yYmRfbm9kZV9jYXBzX2FuZF9lbnRp
            dHkoKQogICAgICAgIGlmICJjbGllbnQuaGVhbHRoY2hlY2tlciIgaW4gdXNlcl9uYW1lOgogICAg
            ICAgICAgICBpZiAiY2xpZW50LmhlYWx0aGNoZWNrZXIiICE9IHVzZXJfbmFtZToKICAgICAgICAg
            ICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVl
            CiAgICAgICAgICAgIHJldHVybiBzZWxmLmdldF9kZWZhdWx0VXNlcl9jYXBzX2FuZF9lbnRpdHko
            KQoKICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICBm
            Im5vIHVzZXIgZm91bmQgd2l0aCB1c2VyX25hbWU6IHt1c2VyX25hbWV9LCAiCiAgICAgICAgICAg
            ICJnZXRfY2Fwc19hbmRfZW50aXR5IGNvbW1hbmQgZmFpbGVkLlxuIgogICAgICAgICkKCiAgICBk
            ZWYgY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3VzZXIoc2VsZiwgdXNlcik6CiAgICAgICAgIiIiCiAg
            ICAgICAgY29tbWFuZDogY2VwaCBhdXRoIGdldC1vci1jcmVhdGUgY2xpZW50LmNzaS1jZXBoZnMt
            cHJvdmlzaW9uZXIgbW9uICdhbGxvdyByJyBtZ3IgJ2FsbG93IHJ3JyBvc2QgJ2FsbG93IHJ3IHRh
            ZyBjZXBoZnMgbWV0YWRhdGE9KicKICAgICAgICAiIiIKICAgICAgICBjYXBzLCBlbnRpdHkgPSBz
            ZWxmLmdldF9jYXBzX2FuZF9lbnRpdHkodXNlcikKICAgICAgICBjbWRfanNvbiA9IHsKICAgICAg
            ICAgICAgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUiLAogICAgICAgICAgICAiZW50aXR5
            IjogZW50aXR5LAogICAgICAgICAgICAiY2FwcyI6IFtjYXAgZm9yIGNhcF9saXN0IGluIGxpc3Qo
            Y2Fwcy5pdGVtcygpKSBmb3IgY2FwIGluIGNhcF9saXN0XSwKICAgICAgICAgICAgImZvcm1hdCI6
            ICJqc29uIiwKICAgICAgICB9CgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoK
            ICAgICAgICAgICAgcmV0dXJuICgKICAgICAgICAgICAgICAgIHNlbGYuZHJ5X3J1bigKICAgICAg
            ICAgICAgICAgICAgICAiY2VwaCAiCiAgICAgICAgICAgICAgICAgICAgKyBjbWRfanNvblsicHJl
            Zml4Il0KICAgICAgICAgICAgICAgICAgICArICIgIgogICAgICAgICAgICAgICAgICAgICsgY21k
            X2pzb25bImVudGl0eSJdCiAgICAgICAgICAgICAgICAgICAgKyAiICIKICAgICAgICAgICAgICAg
            ICAgICArICIgIi5qb2luKGNtZF9qc29uWyJjYXBzIl0pCiAgICAgICAgICAgICAgICApLAogICAg
            ICAgICAgICAgICAgIiIsCiAgICAgICAgICAgICkKICAgICAgICAjIGNoZWNrIGlmIHVzZXIgYWxy
            ZWFkeSBleGlzdAogICAgICAgIHVzZXJfa2V5ID0gc2VsZi5jaGVja191c2VyX2V4aXN0KGVudGl0
            eSkKICAgICAgICBpZiB1c2VyX2tleSAhPSAiIjoKICAgICAgICAgICAgcmV0dXJuIHVzZXJfa2V5
            LCBmIntlbnRpdHkuc3BsaXQoJy4nLCAxKVsxXX0iCiAgICAgICAgICAgICMgZW50aXR5LnNwbGl0
            KCcuJywxKVsxXSB0byByZW5hbWUgZW50aXR5KGNsaWVudC5jc2ktcmJkLW5vZGUpIGFzIGNzaS1y
            YmQtbm9kZQoKICAgICAgICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1v
            bl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nl
            c3NmdWwgYXR0ZW1wdCwKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9
            PSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAg
            ICAgICAgICAgZiInYXV0aCBnZXQtb3ItY3JlYXRlIHt1c2VyfScgY29tbWFuZCBmYWlsZWQuXG4i
            CiAgICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBz
            ZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgKQogICAgICAgIHJldHVybiBzdHIo
            anNvbl9vdXRbMF1bImtleSJdKSwgZiJ7ZW50aXR5LnNwbGl0KCcuJywgMSlbMV19IgogICAgICAg
            ICMgZW50aXR5LnNwbGl0KCcuJywxKVsxXSB0byByZW5hbWUgZW50aXR5KGNsaWVudC5jc2ktcmJk
            LW5vZGUpIGFzIGNzaS1yYmQtbm9kZQoKICAgIGRlZiBnZXRfY2VwaGZzX2RhdGFfcG9vbF9kZXRh
            aWxzKHNlbGYpOgogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAiZnMgbHMiLCAiZm9ybWF0
            IjogImpzb24ifQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAg
            ICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigiY2VwaCAiICsgY21kX2pzb25bInByZWZpeCJdKQogICAg
            ICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dl
            bihjbWRfanNvbikKICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0
            LCByZXBvcnQgYW4gZXJyb3IKICAgICAgICBpZiByZXRfdmFsICE9IDA6CiAgICAgICAgICAgICMg
            aWYgZnMgYW5kIGRhdGFfcG9vbCBhcmd1bWVudHMgYXJlIG5vdCBzZXQsIHNpbGVudGx5IHJldHVy
            bgogICAgICAgICAgICBpZiAoCiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhm
            c19maWxlc3lzdGVtX25hbWUgPT0gIiIKICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFy
            c2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZSA9PSAiIgogICAgICAgICAgICApOgogICAgICAgICAg
            ICAgICAgcmV0dXJuCiAgICAgICAgICAgICMgaWYgdXNlciBoYXMgcHJvdmlkZWQgYW55IG9mIHRo
            ZQogICAgICAgICAgICAjICctLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUnIG9yICctLWNlcGhmcy1k
            YXRhLXBvb2wtbmFtZScgYXJndW1lbnRzLAogICAgICAgICAgICAjIHJhaXNlIGFuIGV4Y2VwdGlv
            biBhcyB3ZSBhcmUgdW5hYmxlIHRvIHZlcmlmeSB0aGUgYXJncwogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiInZnMgbHMnIGNlcGgg
            Y2FsbCBmYWlsZWQgd2l0aCBlcnJvcjoge2Vycl9tc2d9IgogICAgICAgICAgICApCgogICAgICAg
            IG1hdGNoaW5nX2pzb25fb3V0ID0ge30KICAgICAgICAjIGlmICctLWNlcGhmcy1maWxlc3lzdGVt
            LW5hbWUnIGFyZ3VtZW50IGlzIHByb3ZpZGVkLAogICAgICAgICMgY2hlY2sgd2hldGhlciB0aGUg
            cHJvdmlkZWQgZmlsZXN5c3RlbS1uYW1lIGV4aXN0cyBvciBub3QKICAgICAgICBpZiBzZWxmLl9h
            cmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWU6CiAgICAgICAgICAgICMgZ2V0IHRoZSBt
            YXRjaGluZyBsaXN0CiAgICAgICAgICAgIG1hdGNoaW5nX2pzb25fb3V0X2xpc3QgPSBbCiAgICAg
            ICAgICAgICAgICBtYXRjaGVkCiAgICAgICAgICAgICAgICBmb3IgbWF0Y2hlZCBpbiBqc29uX291
            dAogICAgICAgICAgICAgICAgaWYgc3RyKG1hdGNoZWRbIm5hbWUiXSkgPT0gc2VsZi5fYXJnX3Bh
            cnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lCiAgICAgICAgICAgIF0KICAgICAgICAgICAgIyB1
            bmFibGUgdG8gZmluZCBhIG1hdGNoaW5nIGZzLW5hbWUsIHJhaXNlIGFuIGVycm9yCiAgICAgICAg
            ICAgIGlmIGxlbihtYXRjaGluZ19qc29uX291dF9saXN0KSA9PSAwOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBmIkZp
            bGVzeXN0ZW0gcHJvdmlkZWQsICd7c2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9u
            YW1lfScsICIKICAgICAgICAgICAgICAgICAgICBmImlzIG5vdCBmb3VuZCBpbiB0aGUgZnMtbGlz
            dDoge1tzdHIoeFsnbmFtZSddKSBmb3IgeCBpbiBqc29uX291dF19IgogICAgICAgICAgICAgICAg
            KQogICAgICAgICAgICBtYXRjaGluZ19qc29uX291dCA9IG1hdGNoaW5nX2pzb25fb3V0X2xpc3Rb
            MF0KICAgICAgICAjIGlmIGNlcGhmcyBmaWxlc3lzdGVtIG5hbWUgaXMgbm90IHByb3ZpZGVkLAog
            ICAgICAgICMgdHJ5IHRvIGdldCBhIGRlZmF1bHQgZnMgbmFtZSBieSBkb2luZyB0aGUgZm9sbG93
            aW5nCiAgICAgICAgZWxzZToKICAgICAgICAgICAgIyBhLiBjaGVjayBpZiB0aGVyZSBpcyBvbmx5
            IG9uZSBmaWxlc3lzdGVtIGlzIHByZXNlbnQKICAgICAgICAgICAgaWYgbGVuKGpzb25fb3V0KSA9
            PSAxOgogICAgICAgICAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXQgPSBqc29uX291dFswXQogICAg
            ICAgICAgICAjIGIuIG9yIGVsc2UsIGNoZWNrIGlmIGRhdGFfcG9vbCBuYW1lIGlzIHByb3ZpZGVk
            CiAgICAgICAgICAgIGVsaWYgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWU6
            CiAgICAgICAgICAgICAgICAjIGFuZCBpZiBwcmVzZW50LCBjaGVjayB3aGV0aGVyIHRoZXJlIGV4
            aXN0cyBhIGZzIHdoaWNoIGhhcyB0aGUgZGF0YV9wb29sCiAgICAgICAgICAgICAgICBmb3IgZWFj
            aEogaW4ganNvbl9vdXQ6CiAgICAgICAgICAgICAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5j
            ZXBoZnNfZGF0YV9wb29sX25hbWUgaW4gZWFjaEpbImRhdGFfcG9vbHMiXToKICAgICAgICAgICAg
            ICAgICAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXQgPSBlYWNoSgogICAgICAgICAgICAgICAgICAg
            ICAgICBicmVhawogICAgICAgICAgICAgICAgIyBpZiB0aGVyZSBpcyBubyBtYXRjaGluZyBmcyBl
            eGlzdHMsIHRoYXQgbWVhbnMgcHJvdmlkZWQgZGF0YV9wb29sIG5hbWUgaXMgaW52YWxpZAogICAg
            ICAgICAgICAgICAgaWYgbm90IG1hdGNoaW5nX2pzb25fb3V0OgogICAgICAgICAgICAgICAgICAg
            IHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgICAg
            IGYiUHJvdmlkZWQgZGF0YV9wb29sIG5hbWUsIHtzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRh
            X3Bvb2xfbmFtZX0sIgogICAgICAgICAgICAgICAgICAgICAgICAiIGRvZXMgbm90IGV4aXN0cyIK
            ICAgICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICMgYy4gaWYgbm90aGluZyBpcyBzZXQg
            YW5kIGNvdWxkbid0IGZpbmQgYSBkZWZhdWx0LAogICAgICAgICAgICBlbHNlOgogICAgICAgICAg
            ICAgICAgIyBqdXN0IHJldHVybiBzaWxlbnRseQogICAgICAgICAgICAgICAgcmV0dXJuCgogICAg
            ICAgIGlmIG1hdGNoaW5nX2pzb25fb3V0OgogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNl
            cGhmc19maWxlc3lzdGVtX25hbWUgPSBzdHIobWF0Y2hpbmdfanNvbl9vdXRbIm5hbWUiXSkKICAg
            ICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfbWV0YWRhdGFfcG9vbF9uYW1lID0gc3Ry
            KAogICAgICAgICAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXRbIm1ldGFkYXRhX3Bvb2wiXQogICAg
            ICAgICAgICApCgogICAgICAgIGlmIGlzaW5zdGFuY2UobWF0Y2hpbmdfanNvbl9vdXRbImRhdGFf
            cG9vbHMiXSwgbGlzdCk6CiAgICAgICAgICAgICMgaWYgdGhlIHVzZXIgaGFzIGFscmVhZHkgcHJv
            dmlkZWQgZGF0YS1wb29sLW5hbWUsCiAgICAgICAgICAgICMgdGhyb3VnaCAtLWNlcGhmcy1kYXRh
            LXBvb2wtbmFtZQogICAgICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bv
            b2xfbmFtZToKICAgICAgICAgICAgICAgICMgaWYgdGhlIHByb3ZpZGVkIG5hbWUgaXMgbm90IG1h
            dGNoaW5nIHdpdGggdGhlIG9uZSBpbiB0aGUgbGlzdAogICAgICAgICAgICAgICAgaWYgKAogICAg
            ICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lCiAg
            ICAgICAgICAgICAgICAgICAgbm90IGluIG1hdGNoaW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0K
            ICAgICAgICAgICAgICAgICk6CiAgICAgICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAgICAgZiJQcm92aWRlZCBkYXRhLXBv
            b2wtbmFtZTogJ3tzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZX0nLCAiCiAg
            ICAgICAgICAgICAgICAgICAgICAgICJkb2Vzbid0IG1hdGNoIGZyb20gdGhlIGRhdGEtcG9vbHMg
            bGlzdDogIgogICAgICAgICAgICAgICAgICAgICAgICBmIntbc3RyKHgpIGZvciB4IGluIG1hdGNo
            aW5nX2pzb25fb3V0WydkYXRhX3Bvb2xzJ11dfSIKICAgICAgICAgICAgICAgICAgICApCiAgICAg
            ICAgICAgICMgaWYgZGF0YV9wb29sIG5hbWUgaXMgbm90IHByb3ZpZGVkLAogICAgICAgICAgICAj
            IHRoZW4gdHJ5IHRvIGZpbmQgYSBkZWZhdWx0IGRhdGEgcG9vbCBuYW1lCiAgICAgICAgICAgIGVs
            c2U6CiAgICAgICAgICAgICAgICAjIGlmIG5vIGRhdGFfcG9vbHMgZXhpc3QsIHNpbGVudGx5IHJl
            dHVybgogICAgICAgICAgICAgICAgaWYgbGVuKG1hdGNoaW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xz
            Il0pID09IDA6CiAgICAgICAgICAgICAgICAgICAgcmV0dXJuCiAgICAgICAgICAgICAgICBzZWxm
            Ll9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZSA9IHN0cigKICAgICAgICAgICAgICAg
            ICAgICBtYXRjaGluZ19qc29uX291dFsiZGF0YV9wb29scyJdWzBdCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgICMgaWYgdGhlcmUgYXJlIG1vcmUgdGhhbiBvbmUgJ2RhdGFfcG9vbHMnIGV4
            aXN0LAogICAgICAgICAgICAjIHRoZW4gd2FybiB0aGUgdXNlciB0aGF0IHdlIGFyZSB1c2luZyB0
            aGUgc2VsZWN0ZWQgbmFtZQogICAgICAgICAgICBpZiBsZW4obWF0Y2hpbmdfanNvbl9vdXRbImRh
            dGFfcG9vbHMiXSkgPiAxOgogICAgICAgICAgICAgICAgcHJpbnQoCiAgICAgICAgICAgICAgICAg
            ICAgIldBUk5JTkc6IE11bHRpcGxlIGRhdGEgcG9vbHMgZGV0ZWN0ZWQ6ICIKICAgICAgICAgICAg
            ICAgICAgICBmIntbc3RyKHgpIGZvciB4IGluIG1hdGNoaW5nX2pzb25fb3V0WydkYXRhX3Bvb2xz
            J11dfVxuIgogICAgICAgICAgICAgICAgICAgIGYiVXNpbmcgdGhlIGRhdGEtcG9vbDogJ3tzZWxm
            Ll9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZX0nXG4iCiAgICAgICAgICAgICAgICAp
            CgogICAgZGVmIGNyZWF0ZV9jaGVja2VyS2V5KHNlbGYsIHVzZXIpOgogICAgICAgIGNhcHMsIGVu
            dGl0eSA9IHNlbGYuZ2V0X2NhcHNfYW5kX2VudGl0eSh1c2VyKQogICAgICAgIGNtZF9qc29uID0g
            ewogICAgICAgICAgICAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0ZSIsCiAgICAgICAgICAg
            ICJlbnRpdHkiOiBlbnRpdHksCiAgICAgICAgICAgICJjYXBzIjogW2NhcCBmb3IgY2FwX2xpc3Qg
            aW4gbGlzdChjYXBzLml0ZW1zKCkpIGZvciBjYXAgaW4gY2FwX2xpc3RdLAogICAgICAgICAgICAi
            Zm9ybWF0IjogImpzb24iLAogICAgICAgIH0KCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5k
            cnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKAogICAgICAgICAgICAgICAg
            ImNlcGggIgogICAgICAgICAgICAgICAgKyBjbWRfanNvblsicHJlZml4Il0KICAgICAgICAgICAg
            ICAgICsgIiAiCiAgICAgICAgICAgICAgICArIGNtZF9qc29uWyJlbnRpdHkiXQogICAgICAgICAg
            ICAgICAgKyAiICIKICAgICAgICAgICAgICAgICsgIiAiLmpvaW4oY21kX2pzb25bImNhcHMiXSkK
            ICAgICAgICAgICAgKQogICAgICAgICMgY2hlY2sgaWYgdXNlciBhbHJlYWR5IGV4aXN0CiAgICAg
            ICAgdXNlcl9rZXkgPSBzZWxmLmNoZWNrX3VzZXJfZXhpc3QoZW50aXR5KQogICAgICAgIGlmIHVz
            ZXJfa2V5ICE9ICIiOgogICAgICAgICAgICByZXR1cm4gdXNlcl9rZXkKCiAgICAgICAgcmV0X3Zh
            bCwganNvbl9vdXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29u
            KQogICAgICAgICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAg
            aWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmFpc2Ug
            RXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgIGYiJ2F1dGggZ2V0LW9y
            LWNyZWF0ZSB7c2VsZi5ydW5fYXNfdXNlcn0nIGNvbW1hbmQgZmFpbGVkXG4iCiAgICAgICAgICAg
            ICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09V
            VFBVVF9MSVNUfSIKICAgICAgICAgICAgKQogICAgICAgIHJldHVybiBzdHIoanNvbl9vdXRbMF1b
            ImtleSJdKQoKICAgIGRlZiBnZXRfY2VwaF9kYXNoYm9hcmRfbGluayhzZWxmKToKICAgICAgICBj
            bWRfanNvbiA9IHsicHJlZml4IjogIm1nciBzZXJ2aWNlcyIsICJmb3JtYXQiOiAianNvbiJ9CiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2Vs
            Zi5kcnlfcnVuKCJjZXBoICIgKyBjbWRfanNvblsicHJlZml4Il0pCiAgICAgICAgcmV0X3ZhbCwg
            anNvbl9vdXQsIF8gPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAg
            ICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3Zh
            bCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmV0dXJuIE5vbmUKICAg
            ICAgICBpZiAiZGFzaGJvYXJkIiBub3QgaW4ganNvbl9vdXQ6CiAgICAgICAgICAgIHJldHVybiBO
            b25lCiAgICAgICAgcmV0dXJuIGpzb25fb3V0WyJkYXNoYm9hcmQiXQoKICAgIGRlZiBjcmVhdGVf
            cmd3X2FkbWluX29wc191c2VyKHNlbGYpOgogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgInJh
            ZG9zZ3ctYWRtaW4iLAogICAgICAgICAgICAidXNlciIsCiAgICAgICAgICAgICJjcmVhdGUiLAog
            ICAgICAgICAgICAiLS11aWQiLAogICAgICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9P
            UFNfVVNFUl9OQU1FLAogICAgICAgICAgICAiLS1kaXNwbGF5LW5hbWUiLAogICAgICAgICAgICAi
            Um9vayBSR1cgQWRtaW4gT3BzIHVzZXIiLAogICAgICAgICAgICAiLS1jYXBzIiwKICAgICAgICAg
            ICAgImJ1Y2tldHM9Kjt1c2Vycz0qO3VzYWdlPXJlYWQ7bWV0YWRhdGE9cmVhZDt6b25lPXJlYWQi
            LAogICAgICAgICAgICAiLS1yZ3ctcmVhbG0iLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd19yZWFsbV9uYW1lLAogICAgICAgICAgICAiLS1yZ3ctem9uZWdyb3VwIiwKICAgICAgICAg
            ICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZWdyb3VwX25hbWUsCiAgICAgICAgICAgICItLXJn
            dy16b25lIiwKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1lLAogICAg
            ICAgIF0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJl
            dHVybiBzZWxmLmRyeV9ydW4oImNlcGggIiArICIgIi5qb2luKGNtZCkpCiAgICAgICAgdHJ5Ogog
            ICAgICAgICAgICBvdXRwdXQgPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1z
            dWJwcm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vy
            cm9yIGFzIGV4ZWNFcnI6CiAgICAgICAgICAgICMgaWYgdGhlIHVzZXIgYWxyZWFkeSBleGlzdHMs
            IHdlIGp1c3QgcXVlcnkgaXQKICAgICAgICAgICAgaWYgZXhlY0Vyci5yZXR1cm5jb2RlID09IGVy
            cm5vLkVFWElTVDoKICAgICAgICAgICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgICAgICAgICAi
            cmFkb3Nndy1hZG1pbiIsCiAgICAgICAgICAgICAgICAgICAgInVzZXIiLAogICAgICAgICAgICAg
            ICAgICAgICJpbmZvIiwKICAgICAgICAgICAgICAgICAgICAiLS11aWQiLAogICAgICAgICAgICAg
            ICAgICAgIHNlbGYuRVhURVJOQUxfUkdXX0FETUlOX09QU19VU0VSX05BTUUsCiAgICAgICAgICAg
            ICAgICAgICAgIi0tcmd3LXJlYWxtIiwKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFy
            c2VyLnJnd19yZWFsbV9uYW1lLAogICAgICAgICAgICAgICAgICAgICItLXJndy16b25lZ3JvdXAi
            LAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1l
            LAogICAgICAgICAgICAgICAgICAgICItLXJndy16b25lIiwKICAgICAgICAgICAgICAgICAgICBz
            ZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUsCiAgICAgICAgICAgICAgICBdCiAgICAgICAg
            ICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVj
            a19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAgICAgICAgZXhj
            ZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yIGFzIGV4ZWNFcnI6CiAgICAgICAgICAg
            ICAgICAgICAgZXJyX21zZyA9ICgKICAgICAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8g
            ZXhlY3V0ZSBjb21tYW5kIHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAg
            ICAgICAgICAgICAgICAgICAgZiJDb2RlOiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtl
            eGVjRXJyLnN0ZGVycn0iCiAgICAgICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAg
            IHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykKICAgICAgICAgICAgICAgICAgICByZXR1cm4gTm9u
            ZSwgTm9uZSwgRmFsc2UsICItMSIKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVy
            cl9tc2cgPSAoCiAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5k
            IHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICBm
            IkNvZGU6IHtleGVjRXJyLnJldHVybmNvZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykK
            ICAgICAgICAgICAgICAgIHJldHVybiBOb25lLCBOb25lLCBGYWxzZSwgIi0xIgoKICAgICAgICAj
            IGlmIGl0IGlzIHB5dGhvbjIsIGRvbid0IGNoZWNrIGZvciBjZXBoIHZlcnNpb24gZm9yIGFkZGlu
            ZyBgaW5mbz1yZWFkYCBjYXAocmd3X3ZhbGlkYXRpb24pCiAgICAgICAgaWYgc3lzLnZlcnNpb25f
            aW5mby5tYWpvciA8IDM6CiAgICAgICAgICAgIGpzb25vdXRwdXQgPSBqc29uLmxvYWRzKG91dHB1
            dCkKICAgICAgICAgICAgcmV0dXJuICgKICAgICAgICAgICAgICAgIGpzb25vdXRwdXRbImtleXMi
            XVswXVsiYWNjZXNzX2tleSJdLAogICAgICAgICAgICAgICAganNvbm91dHB1dFsia2V5cyJdWzBd
            WyJzZWNyZXRfa2V5Il0sCiAgICAgICAgICAgICAgICBGYWxzZSwKICAgICAgICAgICAgICAgICIi
            LAogICAgICAgICAgICApCgogICAgICAgICMgc2VwYXJhdGVseSBhZGQgaW5mbz1yZWFkIGNhcHMg
            Zm9yIHJndy1lbmRwb2ludCBpcCB2YWxpZGF0aW9uCiAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVk
            ID0gVHJ1ZQogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAg
            ICAgICAgICAiY2FwcyIsCiAgICAgICAgICAgICJhZGQiLAogICAgICAgICAgICAiLS11aWQiLAog
            ICAgICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9PUFNfVVNFUl9OQU1FLAogICAgICAg
            ICAgICAiLS1jYXBzIiwKICAgICAgICAgICAgImluZm89cmVhZCIsCiAgICAgICAgICAgICItLXJn
            dy1yZWFsbSIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUsCiAg
            ICAgICAgICAgICItLXJndy16b25lZ3JvdXAiLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd196b25lZ3JvdXBfbmFtZSwKICAgICAgICAgICAgIi0tcmd3LXpvbmUiLAogICAgICAgICAg
            ICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUsCiAgICAgICAgXQogICAgICAgIHRyeToK
            ICAgICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9
            c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2VwdCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NF
            cnJvciBhcyBleGVjRXJyOgogICAgICAgICAgICAjIGlmIHRoZSBjZXBoIHZlcnNpb24gbm90IHN1
            cHBvcnRlZCBmb3IgYWRkaW5nIGBpbmZvPXJlYWRgIGNhcChyZ3dfdmFsaWRhdGlvbikKICAgICAg
            ICAgICAgaWYgKAogICAgICAgICAgICAgICAgImNvdWxkIG5vdCBhZGQgY2FwczogdW5hYmxlIHRv
            IGFkZCBjYXBzOiBpbmZvPXJlYWRcbiIKICAgICAgICAgICAgICAgIGluIGV4ZWNFcnIuc3RkZXJy
            LmRlY29kZSgidXRmLTgiKQogICAgICAgICAgICAgICAgYW5kIGV4ZWNFcnIucmV0dXJuY29kZSA9
            PSAyNDQKICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgIGluZm9fY2FwX3N1cHBvcnRlZCA9
            IEZhbHNlCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICBlcnJfbXNnID0gKAogICAg
            ICAgICAgICAgICAgICAgIGYiZmFpbGVkIHRvIGV4ZWN1dGUgY29tbWFuZCB7Y21kfS4gT3V0cHV0
            OiB7ZXhlY0Vyci5vdXRwdXR9LiAiCiAgICAgICAgICAgICAgICAgICAgZiJDb2RlOiB7ZXhlY0Vy
            ci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtleGVjRXJyLnN0ZGVycn0iCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKGVycl9tc2cpCiAgICAgICAgICAgICAg
            ICByZXR1cm4gTm9uZSwgTm9uZSwgRmFsc2UsICItMSIKCiAgICAgICAganNvbm91dHB1dCA9IGpz
            b24ubG9hZHMob3V0cHV0KQogICAgICAgIHJldHVybiAoCiAgICAgICAgICAgIGpzb25vdXRwdXRb
            ImtleXMiXVswXVsiYWNjZXNzX2tleSJdLAogICAgICAgICAgICBqc29ub3V0cHV0WyJrZXlzIl1b
            MF1bInNlY3JldF9rZXkiXSwKICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkLAogICAgICAg
            ICAgICAiIiwKICAgICAgICApCgogICAgZGVmIHZhbGlkYXRlX3JiZF9wb29sKHNlbGYpOgogICAg
            ICAgIGlmIG5vdCBzZWxmLmNsdXN0ZXIucG9vbF9leGlzdHMoc2VsZi5fYXJnX3BhcnNlci5yYmRf
            ZGF0YV9wb29sX25hbWUpOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0
            aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcG9vbCwgJ3tzZWxmLl9hcmdfcGFy
            c2VyLnJiZF9kYXRhX3Bvb2xfbmFtZX0nLCBkb2VzIG5vdCBleGlzdCIKICAgICAgICAgICAgKQoK
            ICAgIGRlZiBpbml0X3JiZF9wb29sKHNlbGYpOgogICAgICAgIGlmIGlzaW5zdGFuY2Uoc2VsZi5j
            bHVzdGVyLCBEdW1teVJhZG9zKToKICAgICAgICAgICAgcmV0dXJuCiAgICAgICAgcmJkX3Bvb2xf
            bmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgaW9jdHgg
            PSBzZWxmLmNsdXN0ZXIub3Blbl9pb2N0eChyYmRfcG9vbF9uYW1lKQogICAgICAgIHJiZF9pbnN0
            ID0gcmJkLlJCRCgpCiAgICAgICAgcmJkX2luc3QucG9vbF9pbml0KGlvY3R4LCBUcnVlKQoKICAg
            IGRlZiB2YWxpZGF0ZV9yYWRvc19uYW1lc3BhY2Uoc2VsZik6CiAgICAgICAgcmJkX3Bvb2xfbmFt
            ZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgcmFkb3NfbmFt
            ZXNwYWNlID0gc2VsZi5fYXJnX3BhcnNlci5yYWRvc19uYW1lc3BhY2UKICAgICAgICBpZiByYWRv
            c19uYW1lc3BhY2UgPT0gIiI6CiAgICAgICAgICAgIHJldHVybgogICAgICAgIGlmIHJhZG9zX25h
            bWVzcGFjZS5pc2xvd2VyKCkgPT0gRmFsc2U6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCByYWRvcyBOYW1l
            c3BhY2UsICd7cmFkb3NfbmFtZXNwYWNlfScsICIKICAgICAgICAgICAgICAgIGYiY29udGFpbnMg
            dXBwZXIgY2FzZSIKICAgICAgICAgICAgKQogICAgICAgIHJiZF9pbnN0ID0gcmJkLlJCRCgpCiAg
            ICAgICAgaW9jdHggPSBzZWxmLmNsdXN0ZXIub3Blbl9pb2N0eChyYmRfcG9vbF9uYW1lKQogICAg
            ICAgIGlmIHJiZF9pbnN0Lm5hbWVzcGFjZV9leGlzdHMoaW9jdHgsIHJhZG9zX25hbWVzcGFjZSkg
            aXMgRmFsc2U6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAg
            ICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCByYWRvcyBOYW1lc3BhY2UsICd7cmFkb3NfbmFt
            ZXNwYWNlfScsICIKICAgICAgICAgICAgICAgIGYiaXMgbm90IGZvdW5kIGluIHRoZSBwb29sICd7
            cmJkX3Bvb2xfbmFtZX0nIgogICAgICAgICAgICApCgogICAgZGVmIGdldF9vcl9jcmVhdGVfc3Vi
            dm9sdW1lX2dyb3VwKHNlbGYsIHN1YnZvbHVtZV9ncm91cCwgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFt
            ZSk6CiAgICAgICAgY21kID0gWwogICAgICAgICAgICAiY2VwaCIsCiAgICAgICAgICAgICJmcyIs
            CiAgICAgICAgICAgICJzdWJ2b2x1bWVncm91cCIsCiAgICAgICAgICAgICJnZXRwYXRoIiwKICAg
            ICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAgICAgICAgICAgc3Vidm9sdW1lX2dy
            b3VwLAogICAgICAgIF0KICAgICAgICB0cnk6CiAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNo
            ZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1
            YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICBjbWQgPSBbCiAgICAgICAg
            ICAgICAgICAiY2VwaCIsCiAgICAgICAgICAgICAgICAiZnMiLAogICAgICAgICAgICAgICAgInN1
            YnZvbHVtZWdyb3VwIiwKICAgICAgICAgICAgICAgICJjcmVhdGUiLAogICAgICAgICAgICAgICAg
            Y2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAgICAgICAgICAgICAgIHN1YnZvbHVtZV9ncm91cCwK
            ICAgICAgICAgICAgXQogICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBfID0gc3VicHJv
            Y2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAg
            ICBleGNlcHQgc3VicHJvY2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3I6CiAgICAgICAgICAgICAgICBy
            YWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYic3Vi
            dm9sdW1lIGdyb3VwIHtzdWJ2b2x1bWVfZ3JvdXB9IGlzIG5vdCBhYmxlIHRvIGdldCBjcmVhdGVk
            IgogICAgICAgICAgICAgICAgKQoKICAgIGRlZiBwaW5fc3Vidm9sdW1lKAogICAgICAgIHNlbGYs
            IHN1YnZvbHVtZV9ncm91cCwgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwgcGluX3R5cGUsIHBpbl9z
            ZXR0aW5nCiAgICApOgogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgImNlcGgiLAogICAgICAg
            ICAgICAiZnMiLAogICAgICAgICAgICAic3Vidm9sdW1lZ3JvdXAiLAogICAgICAgICAgICAicGlu
            IiwKICAgICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAgICAgICAgICAgc3Vidm9s
            dW1lX2dyb3VwLAogICAgICAgICAgICBwaW5fdHlwZSwKICAgICAgICAgICAgcGluX3NldHRpbmcs
            CiAgICAgICAgXQogICAgICAgIHRyeToKICAgICAgICAgICAgXyA9IHN1YnByb2Nlc3MuY2hlY2tf
            b3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAgICAgICBleGNlcHQgc3VicHJv
            Y2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3I6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1
            cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmInN1YnZvbHVtZSBncm91cCB7c3Vidm9sdW1l
            X2dyb3VwfSBpcyBub3QgYWJsZSB0byBnZXQgcGlubmVkIgogICAgICAgICAgICApCgogICAgZGVm
            IGdldF9yZ3dfZnNpZChzZWxmLCBiYXNlX3VybCwgdmVyaWZ5KToKICAgICAgICBhY2Nlc3Nfa2V5
            ID0gc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJfQUNDRVNTX0tFWSJdCiAgICAgICAg
            c2VjcmV0X2tleSA9IHNlbGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX1NFQ1JFVF9LRVki
            XQogICAgICAgIHJnd19lbmRwb2ludCA9IHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50CiAg
            ICAgICAgYmFzZV91cmwgPSBiYXNlX3VybCArICI6Ly8iICsgcmd3X2VuZHBvaW50ICsgIi9hZG1p
            bi9pbmZvPyIKICAgICAgICBwYXJhbXMgPSB7ImZvcm1hdCI6ICJqc29uIn0KICAgICAgICByZXF1
            ZXN0X3VybCA9IGJhc2VfdXJsICsgdXJsZW5jb2RlKHBhcmFtcykKCiAgICAgICAgdHJ5OgogICAg
            ICAgICAgICByID0gcmVxdWVzdHMuZ2V0KAogICAgICAgICAgICAgICAgcmVxdWVzdF91cmwsCiAg
            ICAgICAgICAgICAgICBhdXRoPVMzQXV0aChhY2Nlc3Nfa2V5LCBzZWNyZXRfa2V5LCByZ3dfZW5k
            cG9pbnQpLAogICAgICAgICAgICAgICAgdmVyaWZ5PXZlcmlmeSwKICAgICAgICAgICAgKQogICAg
            ICAgIGV4Y2VwdCByZXF1ZXN0cy5leGNlcHRpb25zLlRpbWVvdXQ6CiAgICAgICAgICAgIHN5cy5z
            dGRlcnIud3JpdGUoCiAgICAgICAgICAgICAgICBmImludmFsaWQgZW5kcG9pbnQ6LCBub3QgYWJs
            ZSB0byBjYWxsIGFkbWluLW9wcyBhcGl7cmd3X2VuZHBvaW50fSIKICAgICAgICAgICAgKQogICAg
            ICAgICAgICByZXR1cm4gIiIsICItMSIKICAgICAgICByMSA9IHIuanNvbigpCiAgICAgICAgaWYg
            cjEgaXMgTm9uZSBvciByMS5nZXQoImluZm8iKSBpcyBOb25lOgogICAgICAgICAgICBzeXMuc3Rk
            ZXJyLndyaXRlKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmd3IEVuZHBvaW50LCAn
            e3NlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50fScsIGlzIGludmFsaWQuIgogICAgICAgICAg
            ICApCiAgICAgICAgICAgIHJldHVybiAoCiAgICAgICAgICAgICAgICAiIiwKICAgICAgICAgICAg
            ICAgICItMSIsCiAgICAgICAgICAgICkKCiAgICAgICAgcmV0dXJuIHIxWyJpbmZvIl1bInN0b3Jh
            Z2VfYmFja2VuZHMiXVswXVsiY2x1c3Rlcl9pZCJdLCAiIgoKICAgIGRlZiB2YWxpZGF0ZV9yZ3df
            ZW5kcG9pbnQoc2VsZiwgaW5mb19jYXBfc3VwcG9ydGVkKToKICAgICAgICAjIGlmIHRoZSAnY2x1
            c3RlcicgaW5zdGFuY2UgaXMgYSBkdW1teSBvbmUsCiAgICAgICAgIyBkb24ndCB0cnkgdG8gcmVh
            Y2ggb3V0IHRvIHRoZSBlbmRwb2ludAogICAgICAgIGlmIGlzaW5zdGFuY2Uoc2VsZi5jbHVzdGVy
            LCBEdW1teVJhZG9zKToKICAgICAgICAgICAgcmV0dXJuCgogICAgICAgIHJnd19lbmRwb2ludCA9
            IHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50CgogICAgICAgICMgdmFsaWRhdGUgcmd3IGVu
            ZHBvaW50IG9ubHkgaWYgaXAgYWRkcmVzcyBpcyBwYXNzZWQKICAgICAgICBpcF90eXBlID0gc2Vs
            Zi5faW52YWxpZF9lbmRwb2ludChyZ3dfZW5kcG9pbnQpCgogICAgICAgICMgY2hlY2sgaWYgdGhl
            IHJndyBlbmRwb2ludCBpcyByZWFjaGFibGUKICAgICAgICBjZXJ0ID0gTm9uZQogICAgICAgIGlm
            IG5vdCBzZWxmLl9hcmdfcGFyc2VyLnJnd19za2lwX3RscyBhbmQgc2VsZi52YWxpZGF0ZV9yZ3df
            ZW5kcG9pbnRfdGxzX2NlcnQoKToKICAgICAgICAgICAgY2VydCA9IHNlbGYuX2FyZ19wYXJzZXIu
            cmd3X3Rsc19jZXJ0X3BhdGgKICAgICAgICBiYXNlX3VybCwgdmVyaWZ5LCBlcnIgPSBzZWxmLmVu
            ZHBvaW50X2RpYWwocmd3X2VuZHBvaW50LCBpcF90eXBlLCBjZXJ0PWNlcnQpCiAgICAgICAgaWYg
            ZXJyICE9ICIiOgogICAgICAgICAgICByZXR1cm4gIi0xIgoKICAgICAgICAjIGNoZWNrIGlmIHRo
            ZSByZ3cgZW5kcG9pbnQgYmVsb25ncyB0byB0aGUgc2FtZSBjbHVzdGVyCiAgICAgICAgIyBvbmx5
            IGNoZWNrIGlmIGBpbmZvYCBjYXAgaXMgc3VwcG9ydGVkCiAgICAgICAgaWYgaW5mb19jYXBfc3Vw
            cG9ydGVkOgogICAgICAgICAgICBmc2lkID0gc2VsZi5nZXRfZnNpZCgpCiAgICAgICAgICAgIHJn
            d19mc2lkLCBlcnIgPSBzZWxmLmdldF9yZ3dfZnNpZChiYXNlX3VybCwgdmVyaWZ5KQogICAgICAg
            ICAgICBpZiBlcnIgPT0gIi0xIjoKICAgICAgICAgICAgICAgIHJldHVybiAiLTEiCiAgICAgICAg
            ICAgIGlmIGZzaWQgIT0gcmd3X2ZzaWQ6CiAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRl
            KAogICAgICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHJndyBFbmRwb2ludCwgJ3tzZWxm
            Ll9hcmdfcGFyc2VyLnJnd19lbmRwb2ludH0nLCBpcyBpbnZhbGlkLiBXZSBhcmUgdmFsaWRhdGlu
            ZyBieSBjYWxsaW5nIHRoZSBhZG1pbm9wcyBhcGkgdGhyb3VnaCByZ3ctZW5kcG9pbnQgYW5kIHZh
            bGlkYXRpbmcgdGhlIGNsdXN0ZXJfaWQgJ3tyZ3dfZnNpZH0nIGlzIGVxdWFsIHRvIHRoZSBjZXBo
            IGNsdXN0ZXIgZnNpZCAne2ZzaWR9JyIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAg
            IHJldHVybiAiLTEiCgogICAgICAgICMgY2hlY2sgaWYgdGhlIHJndyBlbmRwb2ludCBwb29sIGV4
            aXN0CiAgICAgICAgIyBvbmx5IHZhbGlkYXRlIGlmIHJnd19wb29sX3ByZWZpeCBpcyBwYXNzZWQg
            ZWxzZSBpdCB3aWxsIHRha2UgZGVmYXVsdCB2YWx1ZSBhbmQgd2UgZG9uJ3QgY3JlYXRlIHRoZXNl
            IGRlZmF1bHQgcG9vbHMKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZp
            eCAhPSAiZGVmYXVsdCI6CiAgICAgICAgICAgIHJnd19wb29sc190b192YWxpZGF0ZSA9IFsKICAg
            ICAgICAgICAgICAgIGYie3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cubWV0
            YSIsCiAgICAgICAgICAgICAgICAiLnJndy5yb290IiwKICAgICAgICAgICAgICAgIGYie3NlbGYu
            X2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cuY29udHJvbCIsCiAgICAgICAgICAgICAg
            ICBmIntzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmxvZyIsCiAgICAgICAg
            ICAgIF0KICAgICAgICAgICAgZm9yIF9yZ3dfcG9vbF90b192YWxpZGF0ZSBpbiByZ3dfcG9vbHNf
            dG9fdmFsaWRhdGU6CiAgICAgICAgICAgICAgICBpZiBub3Qgc2VsZi5jbHVzdGVyLnBvb2xfZXhp
            c3RzKF9yZ3dfcG9vbF90b192YWxpZGF0ZSk6CiAgICAgICAgICAgICAgICAgICAgc3lzLnN0ZGVy
            ci53cml0ZSgKICAgICAgICAgICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcG9vbCwgJ3tf
            cmd3X3Bvb2xfdG9fdmFsaWRhdGV9JywgZG9lcyBub3QgZXhpc3QiCiAgICAgICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICAgICAgICAgIHJldHVybiAiLTEiCgogICAgICAgIHJldHVybiAiIgoK
            ICAgIGRlZiB2YWxpZGF0ZV9yZ3dfbXVsdGlzaXRlKHNlbGYsIHJnd19tdWx0aXNpdGVfY29uZmln
            X25hbWUsIHJnd19tdWx0aXNpdGVfY29uZmlnKToKICAgICAgICBpZiByZ3dfbXVsdGlzaXRlX2Nv
            bmZpZyAhPSAiIjoKICAgICAgICAgICAgY21kID0gWwogICAgICAgICAgICAgICAgInJhZG9zZ3ct
            YWRtaW4iLAogICAgICAgICAgICAgICAgcmd3X211bHRpc2l0ZV9jb25maWcsCiAgICAgICAgICAg
            ICAgICAiZ2V0IiwKICAgICAgICAgICAgICAgICItLXJndy0iICsgcmd3X211bHRpc2l0ZV9jb25m
            aWcsCiAgICAgICAgICAgICAgICByZ3dfbXVsdGlzaXRlX2NvbmZpZ19uYW1lLAogICAgICAgICAg
            ICBdCiAgICAgICAgICAgIHRyeToKICAgICAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNoZWNr
            X291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgICAgIGV4Y2VwdCBz
            dWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJvciBhcyBleGVjRXJyOgogICAgICAgICAgICAgICAg
            ZXJyX21zZyA9ICgKICAgICAgICAgICAgICAgICAgICBmImZhaWxlZCB0byBleGVjdXRlIGNvbW1h
            bmQge2NtZH0uIE91dHB1dDoge2V4ZWNFcnIub3V0cHV0fS4gIgogICAgICAgICAgICAgICAgICAg
            IGYiQ29kZToge2V4ZWNFcnIucmV0dXJuY29kZX0uIEVycm9yOiB7ZXhlY0Vyci5zdGRlcnJ9Igog
            ICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZShlcnJfbXNn
            KQogICAgICAgICAgICAgICAgcmV0dXJuICItMSIKICAgICAgICByZXR1cm4gIiIKCiAgICBkZWYg
            X2dlbl9vdXRwdXRfbWFwKHNlbGYpOgogICAgICAgIGlmIHNlbGYub3V0X21hcDoKICAgICAgICAg
            ICAgcmV0dXJuCiAgICAgICAgIyBzdXBwb3J0IGxlZ2FjeSBmbGFnIHdpdGggdXBncmFkZXMKICAg
            ICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNsdXN0ZXJfbmFtZToKICAgICAgICAgICAgc2VsZi5f
            YXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5jbHVzdGVyX25h
            bWUKICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUgPSAoCiAgICAgICAg
            ICAgIHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZS5sb3dlcigpCiAgICAgICAgKSAg
            IyBhbHdheXMgY29udmVydCBjbHVzdGVyIG5hbWUgdG8gbG93ZXJjYXNlIGNoYXJhY3RlcnMKICAg
            ICAgICBzZWxmLnZhbGlkYXRlX3JiZF9wb29sKCkKICAgICAgICBzZWxmLmluaXRfcmJkX3Bvb2wo
            KQogICAgICAgIHNlbGYudmFsaWRhdGVfcmFkb3NfbmFtZXNwYWNlKCkKICAgICAgICBzZWxmLl9l
            eGNsdWRlZF9rZXlzLmFkZCgiSzhTX0NMVVNURVJfTkFNRSIpCiAgICAgICAgc2VsZi5nZXRfY2Vw
            aGZzX2RhdGFfcG9vbF9kZXRhaWxzKCkKICAgICAgICBzZWxmLm91dF9tYXBbIk5BTUVTUEFDRSJd
            ID0gc2VsZi5fYXJnX3BhcnNlci5uYW1lc3BhY2UKICAgICAgICBzZWxmLm91dF9tYXBbIks4U19D
            TFVTVEVSX05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQogICAgICAg
            IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9GU0lEIl0gPSBzZWxmLmdldF9mc2lkKCkKICAg
            ICAgICBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUk5BTUUiXSA9IHNlbGYucnVuX2Fz
            X3VzZXIKICAgICAgICBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfQ0VQSF9NT05fREFUQSJd
            ID0gc2VsZi5nZXRfY2VwaF9leHRlcm5hbF9tb25fZGF0YSgpCiAgICAgICAgc2VsZi5vdXRfbWFw
            WyJST09LX0VYVEVSTkFMX1VTRVJfU0VDUkVUIl0gPSBzZWxmLmNyZWF0ZV9jaGVja2VyS2V5KAog
            ICAgICAgICAgICAiY2xpZW50LmhlYWx0aGNoZWNrZXIiCiAgICAgICAgKQogICAgICAgIHNlbGYu
            b3V0X21hcFsiUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSyJdID0gc2VsZi5nZXRfY2VwaF9k
            YXNoYm9hcmRfbGluaygpCiAgICAgICAgKAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9S
            QkRfTk9ERV9TRUNSRVQiXSwKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVf
            U0VDUkVUX05BTUUiXSwKICAgICAgICApID0gc2VsZi5jcmVhdGVfY2VwaENTSUtleXJpbmdfdXNl
            cigiY2xpZW50LmNzaS1yYmQtbm9kZSIpCiAgICAgICAgKAogICAgICAgICAgICBzZWxmLm91dF9t
            YXBbIkNTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUIl0sCiAgICAgICAgICAgIHNlbGYub3V0X21h
            cFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICkgPSBzZWxmLmNy
            ZWF0ZV9jZXBoQ1NJS2V5cmluZ191c2VyKCJjbGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIpCiAg
            ICAgICAgc2VsZi5vdXRfbWFwWyJDRVBIRlNfUE9PTF9OQU1FIl0gPSBzZWxmLl9hcmdfcGFyc2Vy
            LmNlcGhmc19kYXRhX3Bvb2xfbmFtZQogICAgICAgIHNlbGYub3V0X21hcFsiQ0VQSEZTX01FVEFE
            QVRBX1BPT0xfTkFNRSJdID0gKAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19t
            ZXRhZGF0YV9wb29sX25hbWUKICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJDRVBIRlNf
            RlNfTkFNRSJdID0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lCiAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJSRVNUUklDVEVEX0FVVEhfUEVSTUlTU0lPTiJdID0gKAogICAgICAg
            ICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uCiAgICAgICAg
            KQogICAgICAgIHNlbGYub3V0X21hcFsiUkFET1NfTkFNRVNQQUNFIl0gPSBzZWxmLl9hcmdfcGFy
            c2VyLnJhZG9zX25hbWVzcGFjZQogICAgICAgIHNlbGYub3V0X21hcFsiU1VCVk9MVU1FX0dST1VQ
            Il0gPSBzZWxmLl9hcmdfcGFyc2VyLnN1YnZvbHVtZV9ncm91cAogICAgICAgIHNlbGYub3V0X21h
            cFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVCJdID0gIiIKICAgICAgICBzZWxmLm91dF9tYXBbIkNT
            SV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUIl0gPSAiIgogICAgICAgICMgY3JlYXRlIENlcGhG
            UyBub2RlIGFuZCBwcm92aXNpb25lciBrZXlyaW5nIG9ubHkgd2hlbiBNRFMgZXhpc3RzCiAgICAg
            ICAgaWYgc2VsZi5vdXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdIGFuZCBzZWxmLm91dF9tYXBbIkNF
            UEhGU19QT09MX05BTUUiXToKICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgc2VsZi5vdXRf
            bWFwWyJDU0lfQ0VQSEZTX05PREVfU0VDUkVUIl0sCiAgICAgICAgICAgICAgICBzZWxmLm91dF9t
            YXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICApID0gc2VsZi5j
            cmVhdGVfY2VwaENTSUtleXJpbmdfdXNlcigiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIpCiAgICAg
            ICAgICAgICgKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9WSVNJ
            T05FUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19Q
            Uk9WSVNJT05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICApID0gc2VsZi5jcmVhdGVfY2Vw
            aENTSUtleXJpbmdfdXNlcigiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiKQogICAgICAg
            ICAgICAjIGNyZWF0ZSB0aGUgZGVmYXVsdCAiY3NpIiBzdWJ2b2x1bWVncm91cAogICAgICAgICAg
            ICBzZWxmLmdldF9vcl9jcmVhdGVfc3Vidm9sdW1lX2dyb3VwKAogICAgICAgICAgICAgICAgImNz
            aSIsIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgICAgICAp
            CiAgICAgICAgICAgICMgcGluIHRoZSBkZWZhdWx0ICJjc2kiIHN1YnZvbHVtZWdyb3VwCiAgICAg
            ICAgICAgIHNlbGYucGluX3N1YnZvbHVtZSgKICAgICAgICAgICAgICAgICJjc2kiLCBzZWxmLl9h
            cmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUsICJkaXN0cmlidXRlZCIsICIxIgogICAg
            ICAgICAgICApCiAgICAgICAgICAgIGlmIHNlbGYub3V0X21hcFsiU1VCVk9MVU1FX0dST1VQIl06
            CiAgICAgICAgICAgICAgICBzZWxmLmdldF9vcl9jcmVhdGVfc3Vidm9sdW1lX2dyb3VwKAogICAg
            ICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuc3Vidm9sdW1lX2dyb3VwLAogICAgICAg
            ICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgIHNlbGYucGluX3N1YnZvbHVtZSgKICAgICAg
            ICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnN1YnZvbHVtZV9ncm91cCwKICAgICAgICAg
            ICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAg
            ICAgICAgICAgICAgICAgImRpc3RyaWJ1dGVkIiwKICAgICAgICAgICAgICAgICAgICAiMSIsCiAg
            ICAgICAgICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJSR1dfVExTX0NFUlQiXSA9ICIi
            CiAgICAgICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UIl0gPSAiIgogICAgICAg
            IHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUIl0gPSAiIgogICAgICAgIGlm
            IG5vdCBzZWxmLl9hcmdfcGFyc2VyLnNraXBfbW9uaXRvcmluZ19lbmRwb2ludDoKICAgICAgICAg
            ICAgKAogICAgICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UIl0s
            CiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJd
            LAogICAgICAgICAgICApID0gc2VsZi5nZXRfYWN0aXZlX2FuZF9zdGFuZGJ5X21ncnMoKQogICAg
            ICAgIHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdID0gc2VsZi5fYXJnX3BhcnNlci5yYmRf
            ZGF0YV9wb29sX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBbIlJCRF9NRVRBREFUQV9FQ19QT09M
            X05BTUUiXSA9ICgKICAgICAgICAgICAgc2VsZi52YWxpZGF0ZV9yYmRfbWV0YWRhdGFfZWNfcG9v
            bF9uYW1lKCkKICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJSR1dfUE9PTF9QUkVGSVgi
            XSA9IHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4CiAgICAgICAgc2VsZi5vdXRfbWFw
            WyJSR1dfRU5EUE9JTlQiXSA9ICIiCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZ3dfZW5k
            cG9pbnQ6CiAgICAgICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAg
            ICAgICAgIHNlbGYuY3JlYXRlX3Jnd19hZG1pbl9vcHNfdXNlcigpCiAgICAgICAgICAgIGVsc2U6
            CiAgICAgICAgICAgICAgICBpZiAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfcmVhbG1fbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdf
            cGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBz
            ZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICk6CiAg
            ICAgICAgICAgICAgICAgICAgZXJyID0gc2VsZi52YWxpZGF0ZV9yZ3dfbXVsdGlzaXRlKAogICAg
            ICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lLCAicmVh
            bG0iCiAgICAgICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAgIGVyciA9IHNlbGYu
            dmFsaWRhdGVfcmd3X211bHRpc2l0ZSgKICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJn
            X3BhcnNlci5yZ3dfem9uZWdyb3VwX25hbWUsICJ6b25lZ3JvdXAiCiAgICAgICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICAgICAgICAgIGVyciA9IHNlbGYudmFsaWRhdGVfcmd3X211bHRpc2l0
            ZSgKICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1l
            LCAiem9uZSIKICAgICAgICAgICAgICAgICAgICApCgogICAgICAgICAgICAgICAgaWYgKAogICAg
            ICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUgPT0gIiIKICAg
            ICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZWdyb3VwX25hbWUg
            PT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZV9u
            YW1lID09ICIiCiAgICAgICAgICAgICAgICApIG9yICgKICAgICAgICAgICAgICAgICAgICBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lICE9ICIiCiAgICAgICAgICAgICAgICAgICAgYW5k
            IHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lICE9ICIiCiAgICAgICAgICAgICAg
            ICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVfbmFtZSAhPSAiIgogICAgICAgICAg
            ICAgICAgKToKICAgICAgICAgICAgICAgICAgICAoCiAgICAgICAgICAgICAgICAgICAgICAgIHNl
            bGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX0FDQ0VTU19LRVkiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJfU0VDUkVUX0tFWSJd
            LAogICAgICAgICAgICAgICAgICAgICAgICBpbmZvX2NhcF9zdXBwb3J0ZWQsCiAgICAgICAgICAg
            ICAgICAgICAgICAgIGVyciwKICAgICAgICAgICAgICAgICAgICApID0gc2VsZi5jcmVhdGVfcmd3
            X2FkbWluX29wc191c2VyKCkKICAgICAgICAgICAgICAgICAgICBlcnIgPSBzZWxmLnZhbGlkYXRl
            X3Jnd19lbmRwb2ludChpbmZvX2NhcF9zdXBwb3J0ZWQpCiAgICAgICAgICAgICAgICAgICAgaWYg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfdGxzX2NlcnRfcGF0aDoKICAgICAgICAgICAgICAgICAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJSR1dfVExTX0NFUlQiXSA9ICgKICAgICAgICAgICAgICAgICAgICAg
            ICAgICAgIHNlbGYudmFsaWRhdGVfcmd3X2VuZHBvaW50X3Rsc19jZXJ0KCkKICAgICAgICAgICAg
            ICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAgICMgaWYgdGhlcmUgaXMgbm8gZXJyb3Is
            IHNldCB0aGUgUkdXX0VORFBPSU5UCiAgICAgICAgICAgICAgICAgICAgaWYgZXJyICE9ICItMSI6
            CiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiUkdXX0VORFBPSU5UIl0gPSBz
            ZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludAogICAgICAgICAgICAgICAgZWxzZToKICAgICAg
            ICAgICAgICAgICAgICBlcnIgPSAiUGxlYXNlIHByb3ZpZGUgYWxsIHRoZSBSR1cgbXVsdGlzaXRl
            IHBhcmFtZXRlcnMgb3Igbm9uZSBvZiB0aGVtIgogICAgICAgICAgICAgICAgICAgIHN5cy5zdGRl
            cnIud3JpdGUoZXJyKQoKICAgIGRlZiBnZW5fc2hlbGxfb3V0KHNlbGYpOgogICAgICAgIHNlbGYu
            X2dlbl9vdXRwdXRfbWFwKCkKICAgICAgICBzaE91dElPID0gU3RyaW5nSU8oKQogICAgICAgIGZv
            ciBrLCB2IGluIHNlbGYub3V0X21hcC5pdGVtcygpOgogICAgICAgICAgICBpZiB2IGFuZCBrIG5v
            dCBpbiBzZWxmLl9leGNsdWRlZF9rZXlzOgogICAgICAgICAgICAgICAgc2hPdXRJTy53cml0ZShm
            ImV4cG9ydCB7a309e3Z9e0xJTkVTRVB9IikKICAgICAgICBzaE91dCA9IHNoT3V0SU8uZ2V0dmFs
            dWUoKQogICAgICAgIHNoT3V0SU8uY2xvc2UoKQogICAgICAgIHJldHVybiBzaE91dAoKICAgIGRl
            ZiBnZW5fanNvbl9vdXQoc2VsZik6CiAgICAgICAgc2VsZi5fZ2VuX291dHB1dF9tYXAoKQogICAg
            ICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuICIiCiAg
            ICAgICAganNvbl9vdXQgPSBbCiAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICJuYW1lIjog
            InJvb2stY2VwaC1tb24tZW5kcG9pbnRzIiwKICAgICAgICAgICAgICAgICJraW5kIjogIkNvbmZp
            Z01hcCIsCiAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAiZGF0
            YSI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9DRVBIX01PTl9EQVRBIl0sCiAgICAgICAg
            ICAgICAgICAgICAgIm1heE1vbklkIjogIjAiLAogICAgICAgICAgICAgICAgICAgICJtYXBwaW5n
            IjogInt9IiwKICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgIH0sCiAgICAgICAgICAgIHsK
            ICAgICAgICAgICAgICAgICJuYW1lIjogInJvb2stY2VwaC1tb24iLAogICAgICAgICAgICAgICAg
            ImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAg
            ICAgICAgICJhZG1pbi1zZWNyZXQiOiAiYWRtaW4tc2VjcmV0IiwKICAgICAgICAgICAgICAgICAg
            ICAiZnNpZCI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9GU0lEIl0sCiAgICAgICAgICAg
            ICAgICAgICAgIm1vbi1zZWNyZXQiOiAibW9uLXNlY3JldCIsCiAgICAgICAgICAgICAgICB9LAog
            ICAgICAgICAgICB9LAogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFtZSI6ICJyb29r
            LWNlcGgtb3BlcmF0b3ItY3JlZHMiLAogICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwK
            ICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICJ1c2VySUQiOiBz
            ZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUk5BTUUiXSwKICAgICAgICAgICAgICAgICAg
            ICAidXNlcktleSI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9VU0VSX1NFQ1JFVCJdLAog
            ICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgfSwKICAgICAgICBdCgogICAgICAgICMgaWYg
            J01PTklUT1JJTkdfRU5EUE9JTlQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAnbW9uaXRvcmluZy1l
            bmRwb2ludCcgdG8gQ2x1c3RlcgogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFw
            WyJNT05JVE9SSU5HX0VORFBPSU5UIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsiTU9O
            SVRPUklOR19FTkRQT0lOVF9QT1JUIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291dC5h
            cHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAibW9u
            aXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiQ2VwaENsdXN0
            ZXIiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAg
            ICAiTW9uaXRvcmluZ0VuZHBvaW50Ijogc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5U
            Il0sCiAgICAgICAgICAgICAgICAgICAgICAgICJNb25pdG9yaW5nUG9ydCI6IHNlbGYub3V0X21h
            cFsiTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUIl0sCiAgICAgICAgICAgICAgICAgICAgfSwKICAg
            ICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQoKICAgICAgICAjIGlmICdDU0lfUkJEX05PREVf
            U0VDUkVUJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ3Jvb2stY3NpLXJiZC1wcm92aXNpb25lcicg
            U2VjcmV0CiAgICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9E
            RV9TRUNSRVQiXQogICAgICAgICAgICBhbmQgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VD
            UkVUX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAg
            ICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9vay17c2VsZi5vdXRf
            bWFwWydDU0lfUkJEX05PREVfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICJr
            aW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAg
            ICAgICAgICAgICAgICJ1c2VySUQiOiBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9TRUNSRVRf
            TkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNlbGYub3V0X21hcFsi
            Q1NJX1JCRF9OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAg
            ICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdDU0lfUkJEX1BST1ZJU0lPTkVSX1NF
            Q1JFVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdyb29rLWNzaS1yYmQtcHJvdmlzaW9uZXInIFNl
            Y3JldAogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJ
            U0lPTkVSX1NFQ1JFVCJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9SQkRfUFJP
            VklTSU9ORVJfU0VDUkVUX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFw
            cGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9v
            ay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAg
            ICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJk
            YXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJD
            U0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAg
            ICJ1c2VyS2V5Ijogc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAog
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAg
            ICAgICAjIGlmICdDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCcgZXhpc3RzLCB0aGVuIG9u
            bHkgYWRkICdyb29rLWNzaS1jZXBoZnMtcHJvdmlzaW9uZXInIFNlY3JldAogICAgICAgIGlmICgK
            ICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJd
            CiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VD
            UkVUX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAg
            ICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9vay17c2VsZi5vdXRf
            bWFwWydDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAg
            ICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewog
            ICAgICAgICAgICAgICAgICAgICAgICAiYWRtaW5JRCI6IHNlbGYub3V0X21hcFsiQ1NJX0NFUEhG
            U19QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAiYWRt
            aW5LZXkiOiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUIl0sCiAg
            ICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAg
            ICAgICMgaWYgJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAn
            cm9vay1jc2ktY2VwaGZzLW5vZGUnIFNlY3JldAogICAgICAgIGlmICgKICAgICAgICAgICAgc2Vs
            Zi5vdXRfbWFwWyJDU0lfQ0VQSEZTX05PREVfU0VDUkVUIl0KICAgICAgICAgICAgYW5kIHNlbGYu
            b3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FIl0KICAgICAgICApOgogICAgICAg
            ICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAg
            ICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9O
            QU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAg
            ICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAiYWRtaW5JRCI6IHNl
            bGYub3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FIl0sCiAgICAgICAgICAgICAg
            ICAgICAgICAgICJhZG1pbktleSI6IHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JF
            VCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAg
            ICkKICAgICAgICAjIGlmICdST09LX0VYVEVSTkFMX0RBU0hCT0FSRF9MSU5LJyBleGlzdHMsIHRo
            ZW4gb25seSBhZGQgJ3Jvb2stY2VwaC1kYXNoYm9hcmQtbGluaycgU2VjcmV0CiAgICAgICAgaWYg
            c2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0RBU0hCT0FSRF9MSU5LIl06CiAgICAgICAgICAg
            IGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAi
            bmFtZSI6ICJyb29rLWNlcGgtZGFzaGJvYXJkLWxpbmsiLAogICAgICAgICAgICAgICAgICAgICJr
            aW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAg
            ICAgICAgICAgICAgICJ1c2VySUQiOiAiY2VwaC1kYXNoYm9hcmQtbGluayIsCiAgICAgICAgICAg
            ICAgICAgICAgICAgICJ1c2VyS2V5Ijogc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0RBU0hC
            T0FSRF9MSU5LIl0sCiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAg
            ICAgICAgICAgKQogICAgICAgICMgaWYgJ1JBRE9TX05BTUVTUEFDRScgZXhpc3RzLCB0aGVuIG9u
            bHkgYWRkIHRoZSAiUkFET1NfTkFNRVNQQUNFIiBuYW1lc3BhY2UKICAgICAgICBpZiAoCiAgICAg
            ICAgICAgIHNlbGYub3V0X21hcFsiUkFET1NfTkFNRVNQQUNFIl0KICAgICAgICAgICAgYW5kIHNl
            bGYub3V0X21hcFsiUkVTVFJJQ1RFRF9BVVRIX1BFUk1JU1NJT04iXQogICAgICAgICAgICBhbmQg
            bm90IHNlbGYub3V0X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdCiAgICAgICAgKToK
            ICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAg
            ICAgICAgICAgICJuYW1lIjogInJhZG9zLW5hbWVzcGFjZSIsCiAgICAgICAgICAgICAgICAgICAg
            ImtpbmQiOiAiQ2VwaEJsb2NrUG9vbFJhZG9zTmFtZXNwYWNlIiwKICAgICAgICAgICAgICAgICAg
            ICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgInJhZG9zTmFtZXNwYWNlTmFtZSI6
            IHNlbGYub3V0X21hcFsiUkFET1NfTkFNRVNQQUNFIl0sCiAgICAgICAgICAgICAgICAgICAgICAg
            ICJwb29sIjogc2VsZi5vdXRfbWFwWyJSQkRfUE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAg
            ICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICAgICBqc29uX291
            dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAi
            Y2VwaC1yYmQtcmFkb3MtbmFtZXNwYWNlIiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJT
            dG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAg
            ICAgICAgICAgICAicG9vbCI6IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAogICAgICAg
            ICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1u
            YW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05B
            TUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRy
            b2xsZXItZXhwYW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRf
            UFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3Np
            LnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0
            X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICB9
            LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgZWxzZToKICAgICAgICAg
            ICAgaWYgc2VsZi5vdXRfbWFwWyJSQkRfTUVUQURBVEFfRUNfUE9PTF9OQU1FIl06CiAgICAgICAg
            ICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICAgICAgewogICAgICAgICAg
            ICAgICAgICAgICAgICAibmFtZSI6ICJjZXBoLXJiZCIsCiAgICAgICAgICAgICAgICAgICAgICAg
            ICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAgICAgICAgICAgICJkYXRhIjog
            ewogICAgICAgICAgICAgICAgICAgICAgICAgICAgImRhdGFQb29sIjogc2VsZi5vdXRfbWFwWyJS
            QkRfUE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYu
            b3V0X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdLAogICAgICAgICAgICAgICAgICAg
            ICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNyZXQtbmFtZSI6IGYi
            cm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwK
            ICAgICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vY29udHJvbGxl
            ci1leHBhbmQtc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9QUk9W
            SVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiY3Np
            LnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0
            X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAg
            ICAgfSwKICAgICAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICAgICApCiAgICAgICAgICAg
            IGVsc2U6CiAgICAgICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICAg
            ICAgewogICAgICAgICAgICAgICAgICAgICAgICAibmFtZSI6ICJjZXBoLXJiZCIsCiAgICAgICAg
            ICAgICAgICAgICAgICAgICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAgICAg
            ICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAgICAgInBvb2wiOiBzZWxm
            Lm91dF9tYXBbIlJCRF9QT09MX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJj
            c2kuc3RvcmFnZS5rOHMuaW8vcHJvdmlzaW9uZXItc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYu
            b3V0X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAg
            ICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNl
            Y3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VD
            UkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4
            cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9S
            QkRfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgICAgIH0sCiAgICAg
            ICAgICAgICAgICAgICAgfQogICAgICAgICAgICAgICAgKQoKICAgICAgICAjIGlmICdDRVBIRlNf
            RlNfTkFNRScgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdjZXBoZnMnIFN0b3JhZ2VDbGFzcwogICAg
            ICAgIGlmIHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXToKICAgICAgICAgICAganNvbl9v
            dXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjog
            ImNlcGhmcyIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwKICAg
            ICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImZzTmFt
            ZSI6IHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAg
            ICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIkNFUEhGU19QT09MX05BTUUiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNyZXQtbmFtZSI6
            IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1F
            J119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250cm9s
            bGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfQ0VQSEZT
            X1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNz
            aS5zdG9yYWdlLms4cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91
            dF9tYXBbJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAg
            ICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICMgaWYgJ1JHV19F
            TkRQT0lOVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdjZXBoLXJndycgU3RvcmFnZUNsYXNzCiAg
            ICAgICAgaWYgc2VsZi5vdXRfbWFwWyJSR1dfRU5EUE9JTlQiXToKICAgICAgICAgICAganNvbl9v
            dXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjog
            ImNlcGgtcmd3IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAog
            ICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAiZW5k
            cG9pbnQiOiBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdLAogICAgICAgICAgICAgICAgICAg
            ICAgICAicG9vbFByZWZpeCI6IHNlbGYub3V0X21hcFsiUkdXX1BPT0xfUFJFRklYIl0sCiAgICAg
            ICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAg
            ICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAg
            ICAgIm5hbWUiOiAicmd3LWFkbWluLW9wcy11c2VyIiwKICAgICAgICAgICAgICAgICAgICAia2lu
            ZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAg
            ICAgICAgICAgICAiYWNjZXNzS2V5Ijogc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJf
            QUNDRVNTX0tFWSJdLAogICAgICAgICAgICAgICAgICAgICAgICAic2VjcmV0S2V5Ijogc2VsZi5v
            dXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJfU0VDUkVUX0tFWSJdLAogICAgICAgICAgICAgICAg
            ICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdSR1df
            VExTX0NFUlQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCB0aGUgImNlcGgtcmd3LXRscy1jZXJ0IiBz
            ZWNyZXQKICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VSVCJdOgogICAgICAgICAg
            ICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAg
            Im5hbWUiOiAiY2VwaC1yZ3ctdGxzLWNlcnQiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjog
            IlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAg
            ICAgICAgICJjZXJ0Ijogc2VsZi5vdXRfbWFwWyJSR1dfVExTX0NFUlQiXSwKICAgICAgICAgICAg
            ICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCgogICAgICAgIHJldHVy
            biBqc29uLmR1bXBzKGpzb25fb3V0KSArIExJTkVTRVAKCiAgICBkZWYgdXBncmFkZV91c2Vyc19w
            ZXJtaXNzaW9ucyhzZWxmKToKICAgICAgICB1c2VycyA9IFsKICAgICAgICAgICAgImNsaWVudC5j
            c2ktY2VwaGZzLW5vZGUiLAogICAgICAgICAgICAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9u
            ZXIiLAogICAgICAgICAgICAiY2xpZW50LmNzaS1yYmQtbm9kZSIsCiAgICAgICAgICAgICJjbGll
            bnQuY3NpLXJiZC1wcm92aXNpb25lciIsCiAgICAgICAgICAgICJjbGllbnQuaGVhbHRoY2hlY2tl
            ciIsCiAgICAgICAgXQogICAgICAgIGlmIHNlbGYucnVuX2FzX3VzZXIgIT0gIiIgYW5kIHNlbGYu
            cnVuX2FzX3VzZXIgbm90IGluIHVzZXJzOgogICAgICAgICAgICB1c2Vycy5hcHBlbmQoc2VsZi5y
            dW5fYXNfdXNlcikKICAgICAgICBmb3IgdXNlciBpbiB1c2VyczoKICAgICAgICAgICAgc2VsZi51
            cGdyYWRlX3VzZXJfcGVybWlzc2lvbnModXNlcikKCiAgICBkZWYgZ2V0X3Jnd19wb29sX25hbWVf
            ZHVyaW5nX3VwZ3JhZGUoc2VsZiwgdXNlciwgY2Fwcyk6CiAgICAgICAgaWYgdXNlciA9PSAiY2xp
            ZW50LmhlYWx0aGNoZWNrZXIiOgogICAgICAgICAgICAjIHdoZW4gYWRtaW4gaGFzIG5vdCBwcm92
            aWRlZCByZ3cgcG9vbCBuYW1lIGR1cmluZyB1cGdyYWRlLAogICAgICAgICAgICAjIGdldCB0aGUg
            cmd3IHBvb2wgbmFtZSBmcm9tIGNsaWVudC5oZWFsdGhjaGVja2VyIHVzZXIgd2hpY2ggd2FzIHVz
            ZWQgZHVyaW5nIGNvbm5lY3Rpb24KICAgICAgICAgICAgaWYgbm90IHNlbGYuX2FyZ19wYXJzZXIu
            cmd3X3Bvb2xfcHJlZml4OgogICAgICAgICAgICAgICAgIyBUbyBnZXQgdmFsdWUgJ2RlZmF1bHQn
            IHdoaWNoIGlzIHJndyBwb29sIG5hbWUgZnJvbSAnYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YScKICAgICAgICAgICAgICAgIHBhdHRlcm4gPSByInBvb2w9KC4qPylcLnJnd1wubWV0YSIK
            ICAgICAgICAgICAgICAgIG1hdGNoID0gcmUuc2VhcmNoKHBhdHRlcm4sIGNhcHMpCiAgICAgICAg
            ICAgICAgICBpZiBtYXRjaDoKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJn
            d19wb29sX3ByZWZpeCA9IG1hdGNoLmdyb3VwKDEpCiAgICAgICAgICAgICAgICBlbHNlOgogICAg
            ICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAg
            ICAgICAgICAgICAgICAgICJmYWlsZWQgdG8gZ2V0IHJndyBwb29sIG5hbWUgZm9yIHVwZ3JhZGUi
            CiAgICAgICAgICAgICAgICAgICAgKQoKICAgIGRlZiB1cGdyYWRlX3VzZXJfcGVybWlzc2lvbnMo
            c2VsZiwgdXNlcik6CiAgICAgICAgIyBjaGVjayB3aGV0aGVyIHRoZSBnaXZlbiB1c2VyIGV4aXN0
            cyBvciBub3QKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4IjogImF1dGggZ2V0IiwgImVudGl0
            eSI6IGYie3VzZXJ9IiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICByZXRfdmFsLCBqc29uX291
            dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAg
            aWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgcHJpbnQo
            ZiJ1c2VyIHt1c2VyfSBub3QgZm91bmQgZm9yIHVwZ3JhZGluZy4iKQogICAgICAgICAgICByZXR1
            cm4KICAgICAgICBleGlzdGluZ19jYXBzID0ganNvbl9vdXRbMF1bImNhcHMiXQogICAgICAgIHNl
            bGYuZ2V0X3Jnd19wb29sX25hbWVfZHVyaW5nX3VwZ3JhZGUodXNlciwgc3RyKGV4aXN0aW5nX2Nh
            cHMpKQogICAgICAgIG5ld19jYXAsIF8gPSBzZWxmLmdldF9jYXBzX2FuZF9lbnRpdHkodXNlcikK
            ICAgICAgICBjYXBfa2V5cyA9IFsibW9uIiwgIm1nciIsICJvc2QiLCAibWRzIl0KICAgICAgICBj
            YXBzID0gW10KICAgICAgICBmb3IgZWFjaENhcCBpbiBjYXBfa2V5czoKICAgICAgICAgICAgY3Vy
            X2NhcF92YWx1ZXMgPSBleGlzdGluZ19jYXBzLmdldChlYWNoQ2FwLCAiIikKICAgICAgICAgICAg
            bmV3X2NhcF92YWx1ZXMgPSBuZXdfY2FwLmdldChlYWNoQ2FwLCAiIikKICAgICAgICAgICAgY3Vy
            X2NhcF9wZXJtX2xpc3QgPSBbCiAgICAgICAgICAgICAgICB4LnN0cmlwKCkgZm9yIHggaW4gY3Vy
            X2NhcF92YWx1ZXMuc3BsaXQoIiwiKSBpZiB4LnN0cmlwKCkKICAgICAgICAgICAgXQogICAgICAg
            ICAgICBuZXdfY2FwX3Blcm1fbGlzdCA9IFsKICAgICAgICAgICAgICAgIHguc3RyaXAoKSBmb3Ig
            eCBpbiBuZXdfY2FwX3ZhbHVlcy5zcGxpdCgiLCIpIGlmIHguc3RyaXAoKQogICAgICAgICAgICBd
            CiAgICAgICAgICAgICMgYXBwZW5kIG5ld19jYXBfbGlzdCB0byBjdXJfY2FwX2xpc3QgdG8gbWFp
            bnRhaW4gdGhlIG9yZGVyIG9mIGNhcHMKICAgICAgICAgICAgY3VyX2NhcF9wZXJtX2xpc3QuZXh0
            ZW5kKG5ld19jYXBfcGVybV9saXN0KQogICAgICAgICAgICAjIGVsaW1pbmF0ZSBkdXBsaWNhdGVz
            IHdpdGhvdXQgdXNpbmcgJ3NldCcKICAgICAgICAgICAgIyBzZXQgcmUtb3JkZXJzIGl0ZW1zIGlu
            IHRoZSBsaXN0IGFuZCB3ZSBoYXZlIHRvIGtlZXAgdGhlIG9yZGVyCiAgICAgICAgICAgIG5ld19j
            YXBfbGlzdCA9IFtdCiAgICAgICAgICAgIFtuZXdfY2FwX2xpc3QuYXBwZW5kKHgpIGZvciB4IGlu
            IGN1cl9jYXBfcGVybV9saXN0IGlmIHggbm90IGluIG5ld19jYXBfbGlzdF0KICAgICAgICAgICAg
            ZXhpc3RpbmdfY2Fwc1tlYWNoQ2FwXSA9ICIsICIuam9pbihuZXdfY2FwX2xpc3QpCiAgICAgICAg
            ICAgIGlmIGV4aXN0aW5nX2NhcHNbZWFjaENhcF06CiAgICAgICAgICAgICAgICBjYXBzLmFwcGVu
            ZChlYWNoQ2FwKQogICAgICAgICAgICAgICAgY2Fwcy5hcHBlbmQoZXhpc3RpbmdfY2Fwc1tlYWNo
            Q2FwXSkKICAgICAgICBjbWRfanNvbiA9IHsKICAgICAgICAgICAgInByZWZpeCI6ICJhdXRoIGNh
            cHMiLAogICAgICAgICAgICAiZW50aXR5IjogdXNlciwKICAgICAgICAgICAgImNhcHMiOiBjYXBz
            LAogICAgICAgICAgICAiZm9ybWF0IjogImpzb24iLAogICAgICAgIH0KICAgICAgICByZXRfdmFs
            LCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24p
            CiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiInYXV0aCBjYXBzIHt1c2VyfScgY29tbWFu
            ZCBmYWlsZWQuXG4gRXJyb3I6IHtlcnJfbXNnfSIKICAgICAgICAgICAgKQogICAgICAgIHByaW50
            KGYiVXBkYXRlZCB1c2VyIHt1c2VyfSBzdWNjZXNzZnVsbHkuIikKCiAgICBkZWYgbWFpbihzZWxm
            KToKICAgICAgICBnZW5lcmF0ZWRfb3V0cHV0ID0gIiIKICAgICAgICBpZiBzZWxmLl9hcmdfcGFy
            c2VyLnVwZ3JhZGU6CiAgICAgICAgICAgIHNlbGYudXBncmFkZV91c2Vyc19wZXJtaXNzaW9ucygp
            CiAgICAgICAgZWxpZiBzZWxmLl9hcmdfcGFyc2VyLmZvcm1hdCA9PSAianNvbiI6CiAgICAgICAg
            ICAgIGdlbmVyYXRlZF9vdXRwdXQgPSBzZWxmLmdlbl9qc29uX291dCgpCiAgICAgICAgZWxpZiBz
            ZWxmLl9hcmdfcGFyc2VyLmZvcm1hdCA9PSAiYmFzaCI6CiAgICAgICAgICAgIGdlbmVyYXRlZF9v
            dXRwdXQgPSBzZWxmLmdlbl9zaGVsbF9vdXQoKQogICAgICAgIGVsc2U6CiAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlVuc3VwcG9y
            dGVkIGZvcm1hdDoge3NlbGYuX2FyZ19wYXJzZXIuZm9ybWF0fSIKICAgICAgICAgICAgKQogICAg
            ICAgIHByaW50KGdlbmVyYXRlZF9vdXRwdXQpCiAgICAgICAgaWYgc2VsZi5vdXRwdXRfZmlsZSBh
            bmQgZ2VuZXJhdGVkX291dHB1dDoKICAgICAgICAgICAgZk91dCA9IG9wZW4oc2VsZi5vdXRwdXRf
            ZmlsZSwgbW9kZT0idyIsIGVuY29kaW5nPSJVVEYtOCIpCiAgICAgICAgICAgIGZPdXQud3JpdGUo
            Z2VuZXJhdGVkX291dHB1dCkKICAgICAgICAgICAgZk91dC5jbG9zZSgpCgoKIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyBNQUlOICMjIyMjIyMjIyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMKaWYgX19uYW1lX18gPT0gIl9fbWFpbl9fIjoKICAgIHJqT2Jq
            ID0gUmFkb3NKU09OKCkKICAgIHRyeToKICAgICAgICByak9iai5tYWluKCkKICAgIGV4Y2VwdCBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uIGFzIGVycjoKICAgICAgICBwcmludChmIkV4ZWN1dGlv
            biBGYWlsZWQ6IHtlcnJ9IikKICAgICAgICByYWlzZSBlcnIKICAgIGV4Y2VwdCBLZXlFcnJvciBh
            cyBrRXJyOgogICAgICAgIHByaW50KGYiS2V5RXJyb3I6IHtrRXJyfSIpCiAgICBleGNlcHQgT1NF
            cnJvciBhcyBvc0VycjoKICAgICAgICBwcmludChmIkVycm9yIHdoaWxlIHRyeWluZyB0byBvdXRw
            dXQgdGhlIGRhdGE6IHtvc0Vycn0iKQogICAgZmluYWxseToKICAgICAgICByak9iai5zaHV0ZG93
            bigpCg==
          external.features.ocs.openshift.io/supported-platforms: '["BareMetal", "None",
            "VSphere", "OpenStack", "oVirt"]'
          external.features.ocs.openshift.io/validation: '{"secrets":["rook-ceph-operator-creds",
            "rook-csi-rbd-node", "rook-csi-rbd-provisioner"], "configMaps": ["rook-ceph-mon-endpoints",
            "rook-ceph-mon"], "storageClasses": ["ceph-rbd"], "cephClusters": ["monitoring-endpoint"]}'
          features.ocs.openshift.io/disabled: '["ss-list", "install-wizard", "block-pool",
            "mcg-resource", "odf-dashboard", "common", "storage-provider", "storage-provisioner",
            "dashboard-resources", "csv-actions", "inventory-item", "alert-actions"]'
          features.ocs.openshift.io/enabled: '["kms", "arbiter", "flexible-scaling",
            "multus", "pool-management", "namespace-store", "mcg-standalone", "taint-nodes",
            "vault-sa-kms", "hpcs-kms"]'
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operatorframework.io/initialization-resource: "\n    {\n        \"apiVersion\":
            \"ocs.openshift.io/v1\",\n        \"kind\": \"StorageCluster\",\n        \"metadata\":
            {\n            \"name\": \"example-storagecluster\",\n            \"namespace\":
            \"openshift-storage\"\n        },\n        \"spec\": {\n            \"manageNodes\":
            false,\n            \"monPVCTemplate\": {\n                \"spec\": {\n
            \                   \"accessModes\": [\n                        \"ReadWriteOnce\"\n
            \                   ],\n                    \"resources\": {\n                        \"requests\":
            {\n                            \"storage\": \"10Gi\"\n                        }\n
            \                   },\n                    \"storageClassName\": \"gp2-csi\"\n
            \               }\n            },\n            \"storageDeviceSets\":
            [\n                {\n                    \"count\": 3,\n                    \"dataPVCTemplate\":
            {\n                        \"spec\": {\n                            \"accessModes\":
            [\n                                \"ReadWriteOnce\"\n                            ],\n
            \                           \"resources\": {\n                                \"requests\":
            {\n                                    \"storage\": \"1Ti\"\n                                }\n
            \                           },\n                            \"storageClassName\":
            \"gp2-csi\",\n                            \"volumeMode\": \"Block\"\n
            \                       }\n                    },\n                    \"name\":
            \"example-deviceset\",\n                    \"placement\": {},\n                    \"portable\":
            true,\n                    \"resources\": {}\n                }\n            ]\n
            \       }\n    }\n\t"
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:03e50284f7fe090c57eddf90feea8ff48d78aff8ce95d1b297aa91e71c3ee577
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.4
          operators.operatorframework.io/internal-objects: '["ocsinitializations.ocs.openshift.io","storageclassrequests.ocs.openshift.io","storageconsumers.ocs.openshift.io","storageprofiles.ocs.openshift.io","cephclusters.ceph.rook.io","cephobjectstores.ceph.rook.io","cephobjectstoreusers.ceph.rook.io","cephnfses.ceph.rook.io","cephclients.ceph.rook.io","cephfilesystems.ceph.rook.io","cephfilesystemmirrors.ceph.rook.io","cephrbdmirrors.ceph.rook.io","cephobjectrealms.ceph.rook.io","cephobjectzonegroups.ceph.rook.io","cephobjectzones.ceph.rook.io","cephbucketnotifications.ceph.rook.io","cephbuckettopics.ceph.rook.io","cephfilesystemsubvolumegroups.ceph.rook.io","cephblockpoolradosnamespaces.ceph.rook.io","cephcosidrivers.ceph.rook.io"]'
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/ocs-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OCSInitialization represents the initial data to be created
              when the operator is installed.
            displayName: OCS Initialization
            kind: OCSInitialization
            name: ocsinitializations.ocs.openshift.io
            version: v1
          - description: StorageClassRequest is the Schema for the StorageClassRequests
              API
            displayName: Storage Class Request
            kind: StorageClassRequest
            name: storageclassrequests.ocs.openshift.io
            version: v1alpha1
          - description: StorageCluster represents a cluster including Ceph Cluster,
              NooBaa and all the storage and compute resources required.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.ocs.openshift.io
            version: v1
          - description: StorageConsumer is the Schema for the storageconsumers API
            displayName: Storage Consumer
            kind: StorageConsumer
            name: storageconsumers.ocs.openshift.io
            version: v1alpha1
          - kind: StorageProfile
            name: storageprofiles.ocs.openshift.io
            version: v1
          - description: Represents a Ceph cluster.
            displayName: Ceph Cluster
            kind: CephCluster
            name: cephclusters.ceph.rook.io
            version: v1
          - description: Represents a Ceph Block Pool.
            displayName: Block Pools
            kind: CephBlockPool
            name: cephblockpools.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store.
            displayName: Ceph Object Store
            kind: CephObjectStore
            name: cephobjectstores.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store User.
            displayName: Ceph Object Store User
            kind: CephObjectStoreUser
            name: cephobjectstoreusers.ceph.rook.io
            version: v1
          - description: Represents a cluster of Ceph NFS ganesha gateways.
            displayName: Ceph NFS
            kind: CephNFS
            name: cephnfses.ceph.rook.io
            version: v1
          - description: Represents a Ceph User.
            displayName: Ceph Client
            kind: CephClient
            name: cephclients.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem.
            displayName: Ceph Filesystem
            kind: CephFilesystem
            name: cephfilesystems.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem Mirror.
            displayName: Ceph Filesystem Mirror
            kind: CephFilesystemMirror
            name: cephfilesystemmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph RBD Mirror.
            displayName: Ceph RBD Mirror
            kind: CephRBDMirror
            name: cephrbdmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Realm.
            displayName: Ceph Object Store Realm
            kind: CephObjectRealm
            name: cephobjectrealms.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone Group.
            displayName: Ceph Object Store Zone Group
            kind: CephObjectZoneGroup
            name: cephobjectzonegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone.
            displayName: Ceph Object Store Zone
            kind: CephObjectZone
            name: cephobjectzones.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Notification.
            displayName: Ceph Bucket Notification
            kind: CephBucketNotification
            name: cephbucketnotifications.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Topic.
            displayName: Ceph Bucket Topic
            kind: CephBucketTopic
            name: cephbuckettopics.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem SubVolumeGroup.
            displayName: Ceph Filesystem SubVolumeGroup
            kind: CephFilesystemSubVolumeGroup
            name: cephfilesystemsubvolumegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph BlockPool Rados Namespace.
            displayName: Ceph BlockPool Rados Namespace
            kind: CephBlockPoolRadosNamespace
            name: cephblockpoolradosnamespaces.ceph.rook.io
            version: v1
          - description: Represents a Ceph COSI Driver.
            displayName: Ceph COSI Driver
            kind: CephCOSIDriver
            name: cephcosidrivers.ceph.rook.io
            version: v1
        description: |
          **Red Hat OpenShift Container Storage** deploys three operators.

          ### OpenShift Container Storage operator

          The OpenShift Container Storage operator is the primary operator for OpenShift Container Storage. It serves to facilitate the other operators in OpenShift Container Storage by performing administrative tasks outside their scope as well as watching and configuring their CustomResources.

          ### Rook

          [Rook][1] deploys and manages Ceph on OpenShift, which provides block and file storage.

          # Core Capabilities

          * **Self-managing service:** No matter which supported storage technologies you choose, OpenShift Container Storage ensures that resources can be deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Container Storage you can either build dedicated storage clusters or hyper-converged clusters where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Container Storage:** OpenShift Container Storage integrates Ceph with multiple storage presentations including object storage (compatible with S3), block storage, and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Container Storage efficiently distributes and replicates your data across your cluster to minimize the risk of data loss. With snapshots, cloning, and versioning, no more losing sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data storage for your workloads, using a single namespace.

          [1]: https://rook.io
        displayName: OpenShift Container Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - rook
        - ceph
        - block storage
        - shared filesystem
        - object storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ocs-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:1711a0aabd28937aac1424adebe96619a48b91f7f888f0588117f80891001da4
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:f37e2c611f4cda32172d4e8938938e488076fd8a8f19c252c56a80c48044bc02
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:028120f2212738917f49e0330b60be5f79a208ce2974bea22b71f533aa6ce3be
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:f75a12fc0c449008028d7fc8e18ca7c1afc37ba585885ac754fed860047794a0
        - registry.redhat.io/odf4/ocs-metrics-exporter-rhel9@sha256:c58c67263c08d84a68c8e45cfd77d59a35aee2151bf79b44377016ca0fb429ee
        - registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:9a0008f662b6afa9e2677a8a83bd662d806ad6f4d03fb49bccb78ed487ce4c7a
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:645d4fd09628f7ddcfc765e4ddacaa79fa6a7c35c03292ca2f6e317d48765d6b
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:ef4276b80cacae825d4831de88d3fe6ba7cc324464087b7b28734bd2b8de3315
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:0dc0733df67bb49886ac9715b81bd4ec4a36330fe1afdde657587e1883f5ca7a
        - registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:03e50284f7fe090c57eddf90feea8ff48d78aff8ce95d1b297aa91e71c3ee577
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:ffcc0806ec2afb8b40d75c70ef2a6d98bb887879f5b3f573e900861316ce9e41
        - registry.redhat.io/odf4/ocs-rhel9-operator@sha256:b1a9492bd9b314f86b4e192cf2c83144885ff863d9932e386a41691f50be6193
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:f8bfbf64b4bd030474e239aed5e03af9dfe920eeae444c5cc8ac3dcfa640fcc1
        version: 4.15.20-rhodf
      entries:
      - name: ocs-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: ocs-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: ocs-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: ocs-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: ocs-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: ocs-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: ocs-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: ocs-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: ocs-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: ocs-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: ocs-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: ocs-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: ocs-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: ocs-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: ocs-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: ocs-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: ocs-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: ocs-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: ocs-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: ocs-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: ocs-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |2-

            [
                {
                    "apiVersion": "ocs.openshift.io/v1",
                    "kind": "StorageCluster",
                    "metadata": {
                        "name": "example-storagecluster",
                        "namespace": "openshift-storage"
                    },
                    "spec": {
                        "manageNodes": false,
                        "monPVCTemplate": {
                            "spec": {
                                "accessModes": [
                                    "ReadWriteOnce"
                                ],
                                "resources": {
                                    "requests": {
                                        "storage": "10Gi"
                                    }
                                },
                                "storageClassName": "gp2-csi"
                            }
                        },
                        "storageDeviceSets": [
                            {
                                "count": 3,
                                "dataPVCTemplate": {
                                    "spec": {
                                        "accessModes": [
                                            "ReadWriteOnce"
                                        ],
                                        "resources": {
                                            "requests": {
                                                "storage": "1Ti"
                                            }
                                        },
                                        "storageClassName": "gp2-csi",
                                        "volumeMode": "Block"
                                    }
                                },
                                "name": "example-deviceset",
                                "placement": {},
                                "portable": true,
                                "resources": {}
                            }
                        ]
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Storage
          containerImage: registry.redhat.io/odf4/ocs-rhel9-operator@sha256:3ef1c2679e9229d51ebc04bd2f2b22ce9b11e8f7fdd7bddee69630fe5b1de47c
          description: Red Hat OpenShift Container Storage provides hyperconverged
            storage for applications within an OpenShift cluster.
          external.features.ocs.openshift.io/supported-platforms: '["BareMetal", "None",
            "VSphere", "OpenStack", "oVirt"]'
          external.features.ocs.openshift.io/validation: '{"secrets":["rook-ceph-operator-creds",
            "rook-csi-rbd-node", "rook-csi-rbd-provisioner"], "configMaps": ["rook-ceph-mon-endpoints",
            "rook-ceph-mon"], "storageClasses": ["ceph-rbd"], "cephClusters": ["monitoring-endpoint"]}'
          features.ocs.openshift.io/disabled: '["ss-list", "install-wizard", "block-pool",
            "mcg-resource", "odf-dashboard", "common", "storage-provider", "storage-provisioner",
            "dashboard-resources", "csv-actions", "inventory-item", "alert-actions"]'
          features.ocs.openshift.io/enabled: '["kms", "arbiter", "flexible-scaling",
            "multus", "pool-management", "namespace-store", "mcg-standalone", "taint-nodes",
            "vault-sa-kms", "hpcs-kms"]'
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operatorframework.io/initialization-resource: "\n    {\n        \"apiVersion\":
            \"ocs.openshift.io/v1\",\n        \"kind\": \"StorageCluster\",\n        \"metadata\":
            {\n            \"name\": \"example-storagecluster\",\n            \"namespace\":
            \"openshift-storage\"\n        },\n        \"spec\": {\n            \"manageNodes\":
            false,\n            \"monPVCTemplate\": {\n                \"spec\": {\n
            \                   \"accessModes\": [\n                        \"ReadWriteOnce\"\n
            \                   ],\n                    \"resources\": {\n                        \"requests\":
            {\n                            \"storage\": \"10Gi\"\n                        }\n
            \                   },\n                    \"storageClassName\": \"gp2-csi\"\n
            \               }\n            },\n            \"storageDeviceSets\":
            [\n                {\n                    \"count\": 3,\n                    \"dataPVCTemplate\":
            {\n                        \"spec\": {\n                            \"accessModes\":
            [\n                                \"ReadWriteOnce\"\n                            ],\n
            \                           \"resources\": {\n                                \"requests\":
            {\n                                    \"storage\": \"1Ti\"\n                                }\n
            \                           },\n                            \"storageClassName\":
            \"gp2-csi\",\n                            \"volumeMode\": \"Block\"\n
            \                       }\n                    },\n                    \"name\":
            \"example-deviceset\",\n                    \"placement\": {},\n                    \"portable\":
            true,\n                    \"resources\": {}\n                }\n            ]\n
            \       }\n    }\n\t"
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:b7758b197cf423573de96bdca0cb3045852b20f34fc9c17ad1632b9c3a3cf3d0
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.4
          operators.operatorframework.io/internal-objects: '["ocsinitializations.ocs.openshift.io","storageconsumers.ocs.openshift.io","storageprofiles.ocs.openshift.io","storagerequests.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/ocs-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OCSInitialization represents the initial data to be created
              when the operator is installed.
            displayName: OCS Initialization
            kind: OCSInitialization
            name: ocsinitializations.ocs.openshift.io
            version: v1
          - description: StorageCluster represents a cluster including Ceph Cluster,
              NooBaa and all the storage and compute resources required.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.ocs.openshift.io
            version: v1
          - description: StorageConsumer is the Schema for the storageconsumers API
            displayName: Storage Consumer
            kind: StorageConsumer
            name: storageconsumers.ocs.openshift.io
            version: v1alpha1
          - description: StorageProfile is the Schema for the storageprofiles API
            displayName: Storage Profile
            kind: StorageProfile
            name: storageprofiles.ocs.openshift.io
            version: v1
          - description: StorageRequest is the Schema for the StorageRequests API
            displayName: Storage Request
            kind: StorageRequest
            name: storagerequests.ocs.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Container Storage** deploys three operators.

          ### OpenShift Container Storage operator

          The OpenShift Container Storage operator is the primary operator for OpenShift Container Storage. It serves to facilitate the other operators in OpenShift Container Storage by performing administrative tasks outside their scope as well as watching and configuring their CustomResources.

          ### Rook

          [Rook][1] deploys and manages Ceph on OpenShift, which provides block and file storage.

          # Core Capabilities

          * **Self-managing service:** No matter which supported storage technologies you choose, OpenShift Container Storage ensures that resources can be deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Container Storage you can either build dedicated storage clusters or hyper-converged clusters where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Container Storage:** OpenShift Container Storage integrates Ceph with multiple storage presentations including object storage (compatible with S3), block storage, and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Container Storage efficiently distributes and replicates your data across your cluster to minimize the risk of data loss. With snapshots, cloning, and versioning, no more losing sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data storage for your workloads, using a single namespace.

          [1]: https://rook.io
        displayName: OpenShift Container Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - rook
        - ceph
        - block storage
        - shared filesystem
        - object storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ocs-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d
        - registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:b7758b197cf423573de96bdca0cb3045852b20f34fc9c17ad1632b9c3a3cf3d0
        - registry.redhat.io/odf4/ocs-rhel9-operator@sha256:3ef1c2679e9229d51ebc04bd2f2b22ce9b11e8f7fdd7bddee69630fe5b1de47c
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:70a3c6bf4bde84c7e43fd58c60abc5612d25c792042b902f6abee94bfaa63f80
        - registry.redhat.io/odf4/ocs-metrics-exporter-rhel9@sha256:3332a93e3ea2e83af15a7edab02a9e00d70e81e44338b63a0eacdea54142ad00
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:0084d9053ad2c74926ca02d2ede0461abd59227eee98d122c45e08a5ebaf629a
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:cf7f72469d25fb37301a8159406613673f8968c82af1035fd45a470e0ae8acfa
        version: 4.16.16-rhodf
      entries:
      - name: ocs-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: ocs-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: ocs-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: ocs-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: ocs-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: ocs-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: ocs-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: ocs-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: ocs-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: ocs-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: ocs-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: ocs-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: ocs-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: ocs-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: ocs-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: ocs-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: ocs-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: ocs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Redis
      provider-url: ""
    name: redis-enterprise-operator-cert-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-enterprise-operator.v6.2.10-45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:ffdf10a1b08115fd4de06f1cd751d36644a9bc9a8c11034cd432e9b66a5d2f73
        - registry.connect.redhat.com/redislabs/services-manager@sha256:8b46f2aacfb3e1b243c0b64975accc55fd80b4de9c9ac4769c3c67a5676addc5
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
        version: 6.2.10-45
      entries:
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: 6.2.10
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.22.0
    - currentCSV: redis-enterprise-operator.v7.22.0-11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:088e4b57c6e1ca1680b359078e5eb1bba67343cdbd9a539af26a0ea23a830593
        - registry.connect.redhat.com/redislabs/services-manager@sha256:3b96cdc65c3e42fd22eb165950be96b907725903ac74fe4f82cfa0073e28e95a
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:d56ba97e95e799210e76615095409143a5dddd5c5a1e1be1d551d7e3f9e4d3ef
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:0cfb07f774db131b5b74e5aa91291579c916c4d4cd1a693d7aaed554f0177a68
        version: 7.22.0-11.3
      entries:
      - name: redis-enterprise-operator.v7.22.0-11.3
        version: 7.22.0-11.3
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.22.0-11.3
    - currentCSV: redis-enterprise-operator.v7.4.6-7.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:473b5661049a6c8830ba113504975726154c7b172376ba3b63d891674d59c78f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5f6f32a63f4dd8be85b9d469f0fedd26d6aa76e32b7735bfa7ab7203cb272cdb
        - registry.connect.redhat.com/redislabs/services-manager@sha256:ba20ab644a473bc742af671171d5313537292d88942b16336d03a6386c061d8a
        version: 7.4.6-7.2
      entries:
      - name: redis-enterprise-operator.v7.4.6-7.2
        version: 7.4.6-7.2
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      name: 7.4.6
    - currentCSV: redis-enterprise-operator.v7.8.2-6.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:36c7a36dabb23a4c003e112085cda2873296bfb22e5d11b470d7c1aba55b7f0f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:c6e1f2109a0dac76b51c25c1cbf6030b4894b21b98147e2ad1db2dd8e7693576
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:acf817c90b5f03bc2796fbc305cf43ce5c0644dcbe1b685752090b3f600a0f7b
        version: 7.8.2-6.8
      entries:
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.2
    - currentCSV: redis-enterprise-operator.v7.8.4-9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:803be04534e59cdc7b8898b3bb336676aad1cb77cf266051e226be82679fcf6a
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:4ba0c414a74268bed9c6757e447b340db7c912afa864cfb096ddf88f11feeaf9
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:66ec124da16cc9d0abd66687e088be0b358e20cfd48f5774d34bf5874e6a4e65
        version: 7.8.4-9.0
      entries:
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.4
    - currentCSV: redis-enterprise-operator.v7.8.6-5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:1fff01daece34aeb7befcb0760d5112c8c7d261f2b86388ece416ac94076a573
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5125e83d3c93b224d1983c599994e43899a0ba0c6f2fc4eed31476f755b5fc0b
        - registry.connect.redhat.com/redislabs/services-manager@sha256:88d483f5e37b5a901d1ae3d965304edc000c172d2e5e307297d5b7bcd77ae73b
        version: 7.8.6-5.1
      entries:
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.6
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: production
    defaultChannel: production
    packageName: redis-enterprise-operator-cert-rhmp
    provider:
      name: Redis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Sigstore Authors
      provider-url: https://sigstore.dev
    name: sigstore-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sigstore-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.sigstore.dev/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "name": "rekor"
                },
                "spec": {
                  "server": {
                    "ingress": {
                      "hostname": "rekor.example.com"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          containerImage: ghcr.io/sigstore/sigstore-helm-operator@sha256:d9c6eb162b97a1aa1ad3bf515db65d01a96e55b6f8bcd646864d3e64a3ee2b9e
          createdAt: "2021-09-01T21:20:28Z"
          description: Helm Charts for Sigstore Project.
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/sigstore/sigstore-helm-operator
          support: The Sigstore Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Rekor
            name: rekors.helm.sigstore.dev
            version: v1alpha1
        description: |-
          Helm Charts for the sigstore project

          ## Charts

          The following charts are available:

          * [Rekor](https://github.com/sigstore/helm-charts/tree/main/charts/rekor)
        displayName: sigstore
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        links:
        - name: Sigstore Helm Operator
          url: https://github.com/sabre1041/sigstore-helm-operator
        maintainers:
        - email: ablock@redhat.com
          name: Andrew Block
        maturity: alpha
        provider:
          name: The Sigstore Authors
          url: https://sigstore.dev
        relatedImages:
        - quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - ghcr.io/sigstore/sigstore-helm-operator@sha256:d9c6eb162b97a1aa1ad3bf515db65d01a96e55b6f8bcd646864d3e64a3ee2b9e
        version: 0.0.1
      entries:
      - name: sigstore-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: sigstore-helm-operator
    provider:
      name: The Sigstore Authors
      url: https://sigstore.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Loki SIG Operator
      provider-url: ""
    name: loki-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: docker.io/grafana/loki-operator:0.8.0
          createdAt: "2025-03-17T11:00:08Z"
          description: The Community Loki Operator provides Kubernetes native deployment
            and management of Loki and related logging components.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/loki/tree/main/operator
          support: Grafana Loki SIG Operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |-
          The Community Loki Operator provides Kubernetes native deployment and management of Loki and related logging components.
          The purpose of this project is to simplify and automate the configuration of a Loki based logging stack for Kubernetes clusters.

          ### Operator features

          The Loki Operator includes, but is not limited to, the following features:

          * Kubernetes Custom Resources: Use Kubernetes custom resources to deploy and manage Loki, Alerting rules, Recording rules, and related components.
          * Simplified Deployment Configuration: Configure the fundamentals of Loki like tenants, limits, replication factor and storage from a native Kubernetes resource.

          ### Feature Gates

          The Loki Operator Bundle provides a set of feature gates that enable/disable specific feature depending on the target Kubernetes distribution. The following feature gates are enabled by default:
          * `serviceMonitors`: Enables creating a Prometheus-Operator managed ServiceMonitor resource per LokiStack component.
          * `serviceMonitorTlsEndpoints`: Enables TLS for the ServiceMonitor endpoints.
          * `lokiStackAlerts`: Enables creating PrometheusRules for common Loki alerts.
          * `httpEncryption`: Enables TLS encryption for all HTTP LokiStack services.
          * `grpcEncryption`: Enables TLS encryption for all GRPC LokiStack services.
          * `builtInCertManagement`: Enables the built-in facility for generating and rotating TLS client and serving certificates for all LokiStack services and internal clients
          * `lokiStackGateway`: Enables reconciling the reverse-proxy lokistack-gateway component for multi-tenant authentication/authorization traffic control to Loki.
          * `runtimeSeccompProfile`: Enables the restricted seccomp profile on all Lokistack components.
          * `defaultNodeAffinity`: Enable the operator will set a default node affinity on all pods. This will limit scheduling of the pods to Nodes with Linux.
          * `lokiStackWebhook`: Enables the LokiStack CR validation and conversion webhooks.
          * `alertingRuleWebhook`: Enables the AlertingRule CR validation webhook.
          * `recordingRuleWebhook`: Enables the RecordingRule CR validation webhook.
          * `rulerConfigWebhook`: Enables the RulerConfig CR validation webhook.

          In addition it enables the following OpenShift-only related feature gates:
          * `servingCertsService`: Enables OpenShift ServiceCA annotations on the lokistack-gateway service only.
          * `ruleExtendedValidation`: Enables extended validation of AlertingRule and RecordingRule to enforce tenancy in an OpenShift context.
          * `clusterTLSPolicy`: Enables usage of TLS policies set in the API Server.
          * `clusterProxy`: Enables usage of the proxy variables set in the proxy resource.

          ### Before you start

          1. Ensure that the appropriate object storage solution, that will be used by Loki, is avaliable and configured.
        displayName: Community Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Documentation
          url: https://loki-operator.dev/
        maintainers:
        - email: loki-operator-team@googlegroups.com
          name: Grafana Loki SIG Operator
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Grafana Loki SIG Operator
        relatedImages:
        - docker.io/grafana/loki-operator:0.8.0
        - quay.io/openshift/origin-kube-rbac-proxy:latest
        - quay.io/observatorium/api:latest
        - quay.io/observatorium/opa-openshift:latest
        - docker.io/grafana/loki:3.4.2
        version: 0.8.0
      entries:
      - name: loki-operator.v0.8.0
        version: 0.8.0
      - name: loki-operator.v0.7.1
        version: 0.7.1
      - name: loki-operator.v0.7.0
        version: 0.7.0
      - name: loki-operator.v0.6.2
        version: 0.6.2
      - name: loki-operator.v0.6.1
        version: 0.6.1
      - name: loki-operator.v0.6.0
        version: 0.6.0
      - name: loki-operator.v0.5.0
        version: 0.5.0
      - name: loki-operator.v0.4.0
        version: 0.4.0
      - name: loki-operator.v0.3.0
        version: 0.3.0
      - name: loki-operator.v0.2.0
        version: 0.2.0
      - name: loki-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: loki-operator
    provider:
      name: Grafana Loki SIG Operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-license-service-reporter-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-license-service-reporter-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "IBMLicenseServiceReporter",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-license-service-reporter-operator",
                    "app.kubernetes.io/instance": "ibmlicenseservicereporter-instance",
                    "app.kubernetes.io/managed-by": "operator",
                    "app.kubernetes.io/name": "ibmlicenseservicereporter",
                    "app.kubernetes.io/part-of": "ibm-license-service-reporter-operator"
                  },
                  "name": "instance"
                },
                "spec": {
                  "authentication": {
                    "useradmin": {
                      "enabled": true
                    }
                  },
                  "license": {
                    "accept": false
                  },
                  "version": "4.2.14"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:277e2ab849907553e4992c2dccc11f03ae1e0302f20aae48c620968aa97d6afc
          createdAt: "2025-02-26T12:10:52Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <4.2.14'
          operatorframework.io/suggested-namespace: ibm-ls-reporter
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "IBM License Service Reporter is a singleton which may be
              used in a single or multi-cluster environment. It aggregates data pushed
              from IBM License Services, deployed on clusters and from ILMT. \n Documentation:
              https://ibm.biz/lsvc-rprtr. \n License: Please refer to the IBM Terms
              website (ibm.biz/lsvc-lic) to find the license terms for the particular
              IBM product for which you are deploying this component. \n IBM License
              Service Reporter is a free, optionally installed add-on – one of services
              of Cloud Pak Foundational Services. Thanks to the IBM License Service
              Reporter customer can: \n - see IBM software deployments on a single
              dashboard, verify and maintain license compliance and avoid audit infractions,
              \n - see historical data on IBM software deployments to support making
              informed decisions for future purchases, \n - see details of software
              deployments, including source (ILMT for VMs, License Service for containers),
              cluster and quantities of licenses deployed, \n - identify which software
              deployments are VMs vs. containers and can use this information to evaluate
              workloads for modernization."
            displayName: IBMLicense Service Reporter
            kind: IBMLicenseServiceReporter
            name: ibmlicenseservicereporters.operator.ibm.com
            version: v1alpha1
        description: |-
          IBM License Service Reporter is a free, optionally installed add-on – one of services of Cloud Pak Foundational Services. Thanks to the IBM License Service Reporter customer can:
           * see IBM software deployments on a single dashboard, verify and maintain license compliance and avoid audit infractions,
           * see historical data on IBM software deployments to support making informed decisions for future purchases,
           * see details of software deployments, including source (ILMT for VMs, License Service for containers), cluster and quantities of licenses deployed,
           * identify which software deployments are VMs vs. containers and can use this information to evaluate workloads for modernization.

          IBM License Service Reporter is a singleton which may be used in a single or multi-cluster environment. It aggregates data pushed from IBM License Services, deployed on clusters and from ILMT.

          **Supported platforms**

          Red Hat OpenShift Container Platform 4.10 or newer installed on Linux x86_64.

          **Prerequisites**

          Prerequisites depend on the integration of the License Service Reporter with an IBM Cloud Pak. For more information, see the applicable IBM Cloud Pak documentation.
        displayName: IBM License Service Reporter
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - licensing
        - usage
        - IBMLicenseServiceReporter
        - IBM
        - Cloud
        - License
        - License Service
        - Service
        - Reporter
        - License Service Reporter
        - License Reporter
        - Service Reporter
        maintainers:
        - email: talk2sam@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-ui@sha256:213889e3c4c1d7bb10c88926ac8276d7e78c558397993d098e36f6c5da617de1
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-oauth2-proxy@sha256:c403c24857e29ddc306ba5d9bc28c3d556b181c5265fee16879fd2b4d946f899
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator-bundle@sha256:648f92bacd52fb51655925e34da1d0bdd0da720da8489b79064e73a47dcdd54a
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:277e2ab849907553e4992c2dccc11f03ae1e0302f20aae48c620968aa97d6afc
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-postgresql@sha256:b6e11080774863f56dd16d5748d03c3f89f6f4872f662ee218d6bf600491a35f
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter@sha256:87e62695d733a60bfc2084cd751810df6801a5fbe1ab87b0a421f6814f1b887a
        version: 4.2.14
      entries:
      - name: ibm-license-service-reporter-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-license-service-reporter-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: orchestrator-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: orchestrator-operator.v1.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Orchestrator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "orchestrator-sample"
                  },
                  "name": "orchestrator-sample"
                },
                "spec": {
                  "argocd": {
                    "enabled": false,
                    "namespace": "orchestrator-gitops"
                  },
                  "platform": {
                    "eventing": {
                      "broker": {}
                    },
                    "monitoring": {
                      "enabled": false
                    },
                    "namespace": "sonataflow-infra",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "64Mi"
                      }
                    }
                  },
                  "postgres": {
                    "authSecret": {
                      "name": "sonataflow-psql-postgresql",
                      "passwordKey": "postgres-password",
                      "userKey": "postgres-username"
                    },
                    "database": "sonataflow",
                    "name": "sonataflow-psql-postgresql",
                    "namespace": "sonataflow-infra"
                  },
                  "rhdh": {
                    "devMode": true,
                    "installOperator": true,
                    "name": "my-rhdh",
                    "namespace": "rhdh",
                    "plugins": {
                      "notificationsEmail": {
                        "enabled": false,
                        "port": 587,
                        "replyTo": "",
                        "sender": ""
                      }
                    }
                  },
                  "serverless": {
                    "installOperator": true
                  },
                  "serverlessLogic": {
                    "installOperator": true
                  },
                  "tekton": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-08-15T00:51:04Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/rhdhorchestrator/orchestrator-go-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Orchestrator
            name: orchestrators.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub Orchestrator is a plugin that enables serverless asynchronous workflows to Backstage.

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://rhdhorchestrator.io)  displayName: Orchestrator Operator
        displayName: Orchestrator Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Orchestrator
        - RHDH
        links:
        - name: Orchestrator Operator
          url: https://github.com/rhdhorchestrator/orchestrator-go-operator
        maintainers:
        - email: jubah@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/rhdh-orchestrator-dev-preview-beta/controller-rhel9-operator@sha256:9f5423829ac55faf2395e2d0d3683b238339dc03285696e1b350c5c5a1c4df3a
        version: 1.6.1
      entries:
      - name: orchestrator-operator.v1.6.1
        version: 1.6.1
      - name: orchestrator-operator.v1.6.0
        version: 1.6.0
      - name: orchestrator-operator.v1.5.2
        version: 1.5.2
      - name: orchestrator-operator.v1.5.1
        version: 1.5.1
      - name: orchestrator-operator.v1.5.0
        version: 1.5.0
      - name: orchestrator-operator.v1.4.0
        version: 1.4.0
      - name: orchestrator-operator.v1.3.1
        version: 1.3.1
      - name: orchestrator-operator.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: orchestrator-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtv-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:d0158c37686bf378347dea5368c0c7838eec9674fbb0ff11f369e6fd99753652
          createdAt: "2024-09-23T16:42:11Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.7.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.5 is supported on OpenShift 4.12 and 4.13

          Migration Toolkit for Virtualization 2.6 is supported on OpenShift 4.14 and 4.15

          Migration Toolkit for Virtualization 2.7 is supported on OpenShift 4.15 and 4.16

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:d0158c37686bf378347dea5368c0c7838eec9674fbb0ff11f369e6fd99753652
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:f8b8d2ce23c1be9dd74d18c7a90324288213e91630da529ded561fbfe5d59a5d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:fffa2fd9f8c804d021264748a910ceb17cee388b4d55cc763a8598234db87f15
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:31e20712c82ae77807eb5dfba5e8616c045280ff164027443214ab8f259d964a
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:ac13419a8c9993bbe656563f478095526dedd34a93d6463979d26852d08ed217
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:19909db45872b2d665eac31f20dac2f0fa3bd9b9a343f2be839747106de6bc5b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:90742a924a1782dced8ae639e68aef93df3c14e4a7b65e2e6b1b73e22860e7a7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:3a07a1e7c74783b0de3c190717f17091d4c9d573776419b46c2ef0c55328d769
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:045cfaec032648cb1a734398fd625de7ac1cc25bfd621b7558e641da82316b58
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:8ce4bba6a69de762f221792c03f6a759afcd16dc576de4b9b1f341d660cb82b0
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:1b06a93c3458b92d40d8eb437043c9abbb9ec5290fe85b2f6ef12297bea0c838
        version: 2.7.0
      entries:
      - name: mtv-operator.v2.7.0
        version: 2.7.0
      name: candidate
    - currentCSV: mtv-operator.2.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui": "true",
                  "feature_validation": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "plan": {
                      "name": "example-plan",
                      "namespace": "openshift-mtv"
                   }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {}
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "type": "vmware",
                   "secret": {
                      "name": "",
                      "namespace": ""
                   },
                   "url": ""
                }
              },
              {
                "kind": "Host",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                     "namespace": "",
                     "name": ""
                   },
                   "id": "",
                   "ipAddress": ""
                }
              },
              {
                "kind": "NetworkMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-networkmap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "type": "pod",
                            "namespace": "",
                            "name": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "StorageMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-storagemap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "storageClass": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "Hook",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "kind": "Provisioner",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "cinder",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "name": "kubernetes.io/cinder",
                  "volumeModes": [
                    {
                      "name": "Block",
                      "priority": 0,
                      "accessModes": [
                        {
                          "name": "ReadWriteOnce",
                          "priority": 0
                        },
                        {
                          "name": "ReadWriteMany",
                          "priority": 1
                        }
                      ]
                    },
                    {
                      "name": "Filesystem",
                      "priority": 1,
                      "accessModes": [
                        {
                          "name": "ReadWriteOnce",
                          "priority": 0
                        },
                        {
                          "name": "ReadWriteMany",
                          "priority": 1
                        }
                      ]
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:74a740878707470ea79c7bf03a7f045c2ea77d924d0890042ba3554dc5613a33
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.1.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui": "true",
                "feature_validation": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          repository: https://github.com/konveyor/forklift-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: Storage Provisioner Capabilities
            displayName: Provisioner
            kind: Provisioner
            name: provisioners.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator fully manages the deployment and life cycle of Migration Toolkit for Virtualization on OpenShift.

          If you experience any issues or have feature requests, please [file an issue in our Github repository](https://github.com/konveyor/forklift-operator/issues)
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        links:
        - name: Konveyor Forklift Operator
          url: https://github.com/konveyor/forklift-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:584a696270554f4fb64755cdc5b525bc6852925e6780f776fb309b1bd8d81d0f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ui-rhel8@sha256:8089b0889f8956f2abf5a69f4ab8f4382919bc15a48663b98551fb0b865d8af7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel8@sha256:aead9b494562985e20fa466888f28119ab302d64e93d7f23bc7d8da1c60267e7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel8@sha256:8c6758c2d81f807c3875f41cec14d9e39f1fe7710584794de57a1449510110db
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:74a740878707470ea79c7bf03a7f045c2ea77d924d0890042ba3554dc5613a33
        version: 2.1.0
      entries:
      - name: mtv-operator.2.1.0
        version: 2.1.0
      name: release-v2.1.0
    - currentCSV: mtv-operator.v2.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui": "true",
                  "feature_validation": "true",
                  "feature_must_gather_api": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "plan": {
                      "name": "example-plan",
                      "namespace": "openshift-mtv"
                   }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {}
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "type": "vmware",
                   "secret": {
                      "name": "",
                      "namespace": ""
                   },
                   "url": ""
                }
              },
              {
                "kind": "Host",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                     "namespace": "",
                     "name": ""
                   },
                   "id": "",
                   "ipAddress": ""
                }
              },
              {
                "kind": "NetworkMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-networkmap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "type": "pod",
                            "namespace": "",
                            "name": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "StorageMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-storagemap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "storageClass": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "Hook",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "kind": "Provisioner",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "cinder",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "name": "kubernetes.io/cinder",
                  "volumeModes": [
                    {
                      "name": "Block",
                      "priority": 0,
                      "accessModes": [
                        {
                          "name": "ReadWriteOnce",
                          "priority": 0
                        },
                        {
                          "name": "ReadWriteMany",
                          "priority": 1
                        }
                      ]
                    },
                    {
                      "name": "Filesystem",
                      "priority": 1,
                      "accessModes": [
                        {
                          "name": "ReadWriteOnce",
                          "priority": 0
                        },
                        {
                          "name": "ReadWriteMany",
                          "priority": 1
                        }
                      ]
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:3d28138500d6b715e7d11983684db12ced28dfef9c97c619044f0b830534f4ed
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.2.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui": "true",
                "feature_validation": "true",
                "feature_must_gather_api": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          repository: https://github.com/konveyor/forklift-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: Storage Provisioner Capabilities
            displayName: Provisioner
            kind: Provisioner
            name: provisioners.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation
          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.
          * Must-gather-api, a service to generate targeted must-gather archives.

          ### Compatibility
          Migration Toolkit for Virtualization 2.0 is supported in OpenShift 4.7
          Migration Toolkit for Virtualization 2.1 is supported in OpenShift 4.8
          Migration Toolkit for Virtualization 2.2 is supported in OpenShift 4.9

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [Konveyor Community](https://github.com/konveyor/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/konveyor/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Konveyor Forklift Operator
          url: https://github.com/konveyor/forklift-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:3d28138500d6b715e7d11983684db12ced28dfef9c97c619044f0b830534f4ed
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel8@sha256:097a3c667e594060844cbe5d16adcc278d35402d1965ff3fca249c8f84a688dc
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:02ec1b0dbc12ba31fce85ee164c97560ba3a3184316324ab7c871989a319bf8d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-api-rhel8@sha256:d6bf2bb018c471f8f86f85c62ef52f4a99510b748ac8b6b7b177fc44fa53fb44
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ui-rhel8@sha256:697b20d6423ac25c0d2f41b2598f4909b6863e85265a4f52bc23b8ce77afc732
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel8@sha256:220d8f97173292dc8389bf4b1a350843757f01ea7861bbcebcf75749366b450b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel8@sha256:6f332be2991b026d5ce2b22e5bdef8d17da06c8b437fb3e5b26e8d35b75e3f95
        version: 2.2.0
      entries:
      - name: mtv-operator.v2.2.0
        version: 2.2.0
      name: release-v2.2.0
    - currentCSV: mtv-operator.v2.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui": "true",
                  "feature_validation": "true",
                  "feature_must_gather_api": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "plan": {
                      "name": "example-plan",
                      "namespace": "openshift-mtv"
                   }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {}
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "type": "vmware",
                   "secret": {
                      "name": "",
                      "namespace": ""
                   },
                   "url": ""
                }
              },
              {
                "kind": "Host",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                     "namespace": "",
                     "name": ""
                   },
                   "id": "",
                   "ipAddress": ""
                }
              },
              {
                "kind": "NetworkMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-networkmap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "type": "pod",
                            "namespace": "",
                            "name": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "StorageMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-storagemap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "storageClass": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "Hook",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:fc6a4503d00719a0d22736a78689b5828a8ebaf8c7895689a56df5cf7b43b97f
          createdAt: "2022-11-08T18:56:59Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.3.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui": "true",
                "feature_validation": "true",
                "feature_must_gather_api": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/forklift-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.
          * Must-gather-api, a service to generate targeted must-gather archives.

          ### Compatibility

          Migration Toolkit for Virtualization 2.2 is supported in OpenShift 4.9

          Migration Toolkit for Virtualization 2.3 is supported in OpenShift 4.10

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [Konveyor Community](https://github.com/konveyor/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/konveyor/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/konveyor/forklift-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ui-rhel8@sha256:e6c2337e610e0f634e57a200a92f492ebe1f111d9fccaa019f8756ba907b067b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel8@sha256:4423034aac33409721ce1b5344f6d5330c174c9b47fc6837b8de7b47018036d7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel8@sha256:872e2a803742071bf81b70d9a58aafd0ceb6ce1d80e651d876690820835854c0
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:fc6a4503d00719a0d22736a78689b5828a8ebaf8c7895689a56df5cf7b43b97f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel8@sha256:59ad5149ce8691eb7944fa291be75e5701c8e74160a0b9a74d098872be704a67
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:a0c201d3474996a08ce662b074ebb069e5082da36f3e7977cf54e3c4e1e54b1d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-api-rhel8@sha256:a7a2d6216ee87284a4f9e07d4c1973b9e001292c504c3b933515826e90073167
        version: 2.3.4
      entries:
      - name: mtv-operator.v2.3.4
        version: 2.3.4
      name: release-v2.3
    - currentCSV: mtv-operator.v2.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui": "true",
                  "feature_validation": "true",
                  "feature_must_gather_api": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "plan": {
                      "name": "example-plan",
                      "namespace": "openshift-mtv"
                   }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {}
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "type": "vmware",
                   "secret": {
                      "name": "",
                      "namespace": ""
                   },
                   "url": ""
                }
              },
              {
                "kind": "Host",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                     "namespace": "",
                     "name": ""
                   },
                   "id": "",
                   "ipAddress": ""
                }
              },
              {
                "kind": "NetworkMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-networkmap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "type": "pod",
                            "namespace": "",
                            "name": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "StorageMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-storagemap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "storageClass": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "Hook",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:fc6a4503d00719a0d22736a78689b5828a8ebaf8c7895689a56df5cf7b43b97f
          createdAt: "2022-11-08T18:56:59Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.3.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui": "true",
                "feature_validation": "true",
                "feature_must_gather_api": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/forklift-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.
          * Must-gather-api, a service to generate targeted must-gather archives.

          ### Compatibility

          Migration Toolkit for Virtualization 2.2 is supported in OpenShift 4.9

          Migration Toolkit for Virtualization 2.3 is supported in OpenShift 4.10

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [Konveyor Community](https://github.com/konveyor/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/konveyor/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/konveyor/forklift-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel8@sha256:4423034aac33409721ce1b5344f6d5330c174c9b47fc6837b8de7b47018036d7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel8@sha256:872e2a803742071bf81b70d9a58aafd0ceb6ce1d80e651d876690820835854c0
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel8@sha256:59ad5149ce8691eb7944fa291be75e5701c8e74160a0b9a74d098872be704a67
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:a0c201d3474996a08ce662b074ebb069e5082da36f3e7977cf54e3c4e1e54b1d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-api-rhel8@sha256:a7a2d6216ee87284a4f9e07d4c1973b9e001292c504c3b933515826e90073167
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ui-rhel8@sha256:e6c2337e610e0f634e57a200a92f492ebe1f111d9fccaa019f8756ba907b067b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:fc6a4503d00719a0d22736a78689b5828a8ebaf8c7895689a56df5cf7b43b97f
        version: 2.3.4
      entries:
      - name: mtv-operator.v2.3.4
        version: 2.3.4
      name: release-v2.3.0
    - currentCSV: mtv-operator.v2.4.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_must_gather_api": "true",
                  "feature_ui": "false",
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:f17fb7a2a840bda51551683e18d01ad1fae3ea73c293de1711f382ed309b8668
          createdAt: "2023-10-23T12:16:09Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.4.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui": "false",
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_must_gather_api": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.
          * Must-gather-api, a service to generate targeted must-gather archives.

          ### Compatibility

          Migration Toolkit for Virtualization 2.2 is supported in OpenShift 4.9

          Migration Toolkit for Virtualization 2.3 is supported in OpenShift 4.10 and 4.11

          Migration Toolkit for Virtualization 2.4 is supported in OpenShift 4.11 and 4.12

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-api-rhel8@sha256:cf22690c6bb420b9acc09ee5f90329ece6a35c18e5d3d6f6e40a320e2385ad85
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:7b4eb4de5dbc7e1af9c74d47fb6bba39f4973dc2d41531989c9fcf454b715b14
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:dc7ed113fae6ea3d5baaf165b08b5404a15f3a70f08089caca65b837623af4fb
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:8d3aeb8f6cfb17f5694f92f56a2361e08d220e2608183fb39ac94e92187b3fc7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-warm-rhel8@sha256:a6f91bc0ffc5e2f545a18b8888715281db7a751e1eff5f9eb89e8d319268a3ec
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:f17fb7a2a840bda51551683e18d01ad1fae3ea73c293de1711f382ed309b8668
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:8e7a7b3c2dc31b90e2102a82633f1eda95355496ff657964fedb26157f9ec229
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:ad9bacdfa4f10e020902d8e0381bdd84e48ce89ffab0f4144bd3c03fc91f7259
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:86716db780d3e43fa1e3029ab095ae6a557cf023797f4d79565a1aa5ebcb9b93
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:a7caa55c9d211a869b42e6f0101cd5099d64f5f3fd423f8a06091bad98d84aff
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ui-rhel8@sha256:54638e44269175e112a510fe2558f9a1dda4758bf3b3fb5e5b54c6617a5917bd
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:20a7b16f9f5d3dd58dbc7c061567590618a24bd595911ed089034095cc919d96
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:9aec00a487545eef76dcc4701680c5ec5bdc5e74302edbdfbc12c821b8ab4997
        version: 2.4.3
      entries:
      - name: mtv-operator.v2.4.3
        version: 2.4.3
      name: release-v2.4
    - currentCSV: mtv-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_must_gather_api": "true",
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:7dac245977fa0beeb7539b8ea5b3be68c457da132c979fd36405509fc879c7e1
          createdAt: "2023-08-31T15:03:42Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.5.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_must_gather_api": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.
          * Must-gather-api, a service to generate targeted must-gather archives.

          ### Compatibility

          Migration Toolkit for Virtualization 2.4 is supported on OpenShift 4.11 and 4.12

          Migration Toolkit for Virtualization 2.5 is supported on OpenShift 4.12 and 4.13

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:00fac23d8a514208c906ab2c911f188947368cb5e3076aacf3dc049e27b6c366
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:7dac245977fa0beeb7539b8ea5b3be68c457da132c979fd36405509fc879c7e1
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:8cc711f92258cad8f0747d3b98fef432d20661b90c4b67df8aa91650f2be3388
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-warm-rhel8@sha256:b5e0a88008c5c355844ed16b0889b0d73e63748fffafe57d0d1c52bc39c50f87
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:349cb43a620b2db9e724d874b231ef2ad896260009e3b0a93539121ac9cbe84b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:c157d6ebfb12f37a646f5f3bfbfdc90b7c8c669afe08ad342d34e474c39aa094
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:68cde15ed754189d74a528fc61cc2191861a27a4a9be112fb3ce1de3e836a77c
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-api-rhel8@sha256:c702dc948e42b9ba1e87b988f603be3df249a44dfdd6a11b0b9b49cb25f6de0c
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:55d773cad8fcf2919f5b0a5387dea62feb4ce676fb94eba310fc4d459f729d73
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:18591334cd99eb9e9f7e3d2609ff8666a9c0625e996d2f8751a14ebb9386457a
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:6097cc8e88905793bc61e48a62d77decffbbc59c1d4de83bb2eb5340ea1e8697
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:e2fca2e132bfc9a90a7c1f7e8e6b6d6c4fd8a63ec4177f5dea3f06c0c529e722
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:e665d8c5da798dc0dcd66b691db087de49564333e3c6e238d16a356c8c62d080
        version: 2.5.0
      entries:
      - name: mtv-operator.v2.5.0
        version: 2.5.0
      name: release-v2.5
    - currentCSV: mtv-operator.v2.6.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:1dfbb0993d3009608a6b272f178197f8a88edb047db4f4ff84aee8c023191d65
          createdAt: "2024-09-13T15:42:40Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.6.7'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.5 is supported on OpenShift 4.12 and 4.13

          Migration Toolkit for Virtualization 2.6 is supported on OpenShift 4.14 and 4.15

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:d34826b2a9370a1b6502d8e74a1254b2bd786684b25d90c2f0a960be0e99777d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-warm-rhel8@sha256:a46eb455027d248bcf591b612f34eb2ff4cd887cbbc3d116f78187b579ac31ab
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:3fef1c267f1f89d51ba7cd54bf39e17d7fe8ba5a4b6ecf9cdedcbe6d49d28bcc
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:1dfbb0993d3009608a6b272f178197f8a88edb047db4f4ff84aee8c023191d65
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:537f5fed16dcc46ee9cea05285a68c0e037c9b439289ff537590352865eced92
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:7dbf0f8d3d11cdb73033377552f999951331ed61d28c98cfd836903faf5601f9
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:a30fca5c5f260b94181c03a8c4c1cfe340048af87515e86571e9676e836d4b9f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:c98635ac256743873c4d74d63f4278cd26794dbaa1d56a82b15089dab510e905
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:a24acbfeca235270e75bd78046aa1dded9ff2a325a725cb46fc52ed3ff6880b8
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:403a5f8fd64926a9638342dbe568b557a7e3dc4b700cde79e64ecd9d948002a7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:6e0b0a7aca1a4afef62cdf511195e6a567feb644d3e5128c9899afc351807212
        version: 2.6.7
      entries:
      - name: mtv-operator.v2.6.7
        version: 2.6.7
      - name: mtv-operator.v2.6.6
        version: 2.6.6
      - name: mtv-operator.v2.6.5
        version: 2.6.5
      - name: mtv-operator.v2.6.4
        version: 2.6.4
      - name: mtv-operator.v2.6.3
        version: 2.6.3
      - name: mtv-operator.v2.6.2
        version: 2.6.2
      - name: mtv-operator.v2.6.1
        version: 2.6.1
      - name: mtv-operator.v2.6.0
        version: 2.6.0
      name: release-v2.6
    - currentCSV: mtv-operator.v2.7.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:db4cbc7214cdc7bf176ff16e562120111219ad8e663b7d4385dfd4c164253b7f
          createdAt: "2025-04-22T18:23:33Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.7.12'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.7 is supported on OpenShift 4.15, 4.16 and 4.17

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:a0360971c4cf448f99eb9ca79c9a153aec479e59e5985b91e019ee478382510b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:ed8cefd01f3b03df7faaf1ba0774cc18b7fc08b25f359d1e52c794b4ad60441e
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:33bae3e4c5b96df3319287d409d8d25db7db33704c7e9f8a58aec797bf20d12a
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:db4cbc7214cdc7bf176ff16e562120111219ad8e663b7d4385dfd4c164253b7f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:97ba4a4b479957e57eb296598d0c8797233f3fd0dbac20fb4686db02d3a8ed9f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:1622759bca392e2f7f66611df7988e7c63138807f917f8b34aefcc60467051ec
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:a32fba94e62f4c979e3b3596ab3a70750bd3bbf841e6004b5fd5305ac796027f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:51fdb9274911e93ae9dff9ab616a04a8a2c8fce4b6b21c05cf5f0145b1f93de0
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:a0b713409ad1d5018832cd50a2ffa3ce70ebf6cf1a4e532b3676b6f039316b52
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:ec2b67b3aff9b74f2246c310f1871198bbea8428b4e563d0e8c254a785b46a27
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:112a465da1695340286a7ff6e4dd66144209ce4389134f03813f786d30ba7508
        version: 2.7.12
      entries:
      - name: mtv-operator.v2.7.12
        version: 2.7.12
      - name: mtv-operator.v2.7.11
        version: 2.7.11
      - name: mtv-operator.v2.7.10
        version: 2.7.10
      - name: mtv-operator.v2.7.9
        version: 2.7.9
      - name: mtv-operator.v2.7.8
        version: 2.7.8
      - name: mtv-operator.v2.7.7
        version: 2.7.7
      - name: mtv-operator.v2.7.6
        version: 2.7.6
      - name: mtv-operator.v2.7.5
        version: 2.7.5
      - name: mtv-operator.v2.7.4
        version: 2.7.4
      - name: mtv-operator.v2.7.3
        version: 2.7.3
      - name: mtv-operator.v2.7.2
        version: 2.7.2
      - name: mtv-operator.v2.7.1
        version: 2.7.1
      - name: mtv-operator.v2.7.0
        version: 2.7.0
      name: release-v2.7
    - currentCSV: mtv-operator.v2.8.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "\u003cbase64-encoded-playbook\u003e"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "konveyor-forklift"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:e4145efb22dfc336d8b81f24f5b488db602d6d547479d615bf82b4fdce96c784
          createdAt: "2025-08-26T14:06:48Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: VSphere Xcopy Volume Populator
            displayName: VSphereXcopyVolumePopulator
            kind: VSphereXcopyVolumePopulator
            name: vspherexcopyvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.8 is supported on OpenShift 4.16, 4.17 and 4.18

          Migration Toolkit for Virtualization 2.9 is supported on OpenShift 4.17, 4.18 and 4.19

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/support/policy/updates/migration-toolkit-for-virtualization).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:e4145efb22dfc336d8b81f24f5b488db602d6d547479d615bf82b4fdce96c784
        version: 2.8.7
      entries:
      - name: mtv-operator.v2.8.7
        version: 2.8.7
      - name: mtv-operator.v2.8.6
        version: 2.8.6
      - name: mtv-operator.v2.8.5
        version: 2.8.5
      - name: mtv-operator.v2.8.4
        version: 2.8.4
      - name: mtv-operator.v2.8.3
        version: 2.8.3
      - name: mtv-operator.v2.8.2
        version: 2.8.2
      - name: mtv-operator.v2.8.1
        version: 2.8.1
      - name: mtv-operator.v2.8.0
        version: 2.8.0
      name: release-v2.8
    defaultChannel: release-v2.8
    packageName: mtv-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: serverless-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: serverless-operator.v1.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.36.0 <1.36.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:52ba6b1252b6f3d41e2a0b3b880cb768010ab9dd925ac34596f6adaedcbbe07c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-source-rhel8@sha256:88323da16360981f27b8e4686fbbb64053e4e0e3c79c65d695edb317b334e1e0
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:3fe326ad9a00b112fc888ffd345c8d5042f0a22cf32ea575931192c39ce1fece
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:259e5dab50d36cdbfc117a42325128c4e16f34c048ba64b36d4a89132ddb4c1a
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:4344716eb2940c6a3fa9a278250f936602bb1549fece6037912837719c178b24
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-transform-jsonata-rhel8@sha256:3af4c48030263eb5769e5f36a6dd13fd390d8566efaab37532f39e1df1c95c67
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sns-sink-rhel8@sha256:51b5e18e8d4c1ca81d7c5ddfea6352a3887f757acfcc53bd2f8bd3b01337ed42
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-ddb-streams-source-rhel8@sha256:a903480159ebf1d8652eeaa702d9885fb76bb106c02d75cec5c82c31d77d6410
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:e3ae1cc6012f19631c012a07b113fb015977d2ee4f9faab3207fbc2464c25aa7
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:596a7272d6de714d7dbc6e00e9e50a2872846c269547465252d953172ef0b009
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:f4d86b196457cdeebdc380b6f1b707c10f8dee789c7779c90ee1332e511af8c4
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:f2d112ba9fdfccea988589b99b463c8a7157475f7a3f03678e8999e65b678cc3
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-sink-rhel8@sha256:afa634e029bcf79a972f49372498d44b45bcfbdb6eb28ecaf4f0b0b2a0a1824b
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:dcfe5fbc57b9f8520adbff368c8bf29eb491c9c26d1a1737a1722cffea83eb1b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-source-rhel8@sha256:27e74287774cda208f667bc2a386debe6b6255e8991abfb37c959c8525f2af53
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:9b328fff0cc8545fb9bb1f24f11d78cfac3ee76dedc11fa4def4262d79a60e8c
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:82e69a1890042f5636ac7bb645b1aaeaba0a4894851eccc1fa1c63296918e31a
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:3d780ad1ab280b5092b90e277485d589da2a18e8b08090af7d95aecd17af545d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:631eb3a8dd9d83ceafa7d8a4da4b1e2d4f2740e6017c766828dcc7575cf1322f
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-log-sink-rhel8@sha256:30b8045ef5f925a36c58941386f1899bec125d005a08aeeeba5edd49b0859149
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:c7c2e738e0d9da1d2d1993f68938e72c4709b1f28a54fa92224ef6a0e456b0aa
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:a1834b620f875681388f0ae224dd0771773a8b61b1d8e5214a942f232854b87b
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:074ebfc4494581611647ba90049655a470ca43743610a27be77db66a8633bac6
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:429d5d9568cca31fbe83c663276a300286af4cdbf3d35685afdbe5f6f2f8c6e0
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:5352e4eb4e426a41521276decadf204b6d23b4762a298e4d3dc73969e3eddc68
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:bf82b37ab8fd29c7a7cd846008a16d14b4cfdc1c09df4f731d71be3957e90123
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:6b7316cce3432d391adc6ac38c92ff0ee86dd7ae0101957cc65cd338cc69e352
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:3b04e9d8cd4b5e43a07ece838d10478723a6c6529350f010550b4a0d1d0950d1
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:6210a946f373b9c36c9b31293debe6fed090008b8234c0ba17d37fa81646172e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:c39cb88afaddc295d93a3774b8a20d001b82392df540fb502ee02316d82460a9
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:0c2cdf917bbad1f5bca8a2a2ef12a96cfcd0c77c190cc35687eb57169b54ab6f
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:187fb9d64b53c55a029ffce3c0f9e7c3304683fa8a0e832cfb7134e8a3d0c237
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:d6f338e08fa937c8a4ccfec692424de4acded565f7ca3971ea507dffeb351503
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:f3d87a7fa5b5f2818effcac3304bd56eddcdb0db97a18ac0f35a06fc7e223e7d
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:4aeb68a22f63300e54391ec6b4cda8c4e665d904b5dcdd3c40155040525742a0
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:ed60e889a677fb139f2eb4bd9ab4fda6ecc39fd050ca2b6cc1c19717e428535a
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:77e4689b2ce7a713e905553e0e42b8e128045f40cb6c6d4ee082f126cbc46b00
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:4274004b6289fa9bea3fa6e56d64077a328a6b5e63a9f0e9c41f6d561f682bd7
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:a6f565d73375e827d830dca85838f4c4cb7f08b0c078acb0c0d567b25f0d256a
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-timer-source-rhel8@sha256:8fed5ef9a1f39a55077bb5026c8854d0ee7fc8d8ff5da277ce9c1bfc2ab2474b
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:78fee4c596f2848487ca704da3fa1074e3c9c99b3f96f66833aeba63836d7b93
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:a47b6b5237a87ccd90e5a49a53d06485fd55c87eebf12d01c335b87fc3a8c494
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-sink-rhel8@sha256:d9b944e64391a996104afbe8a21f66316a948e9fb7b91734e861847285be59c7
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:ee6533b71cf280877c9d7489dfd821c9209c6802bed7fa3fdb4f8f00fbeacb09
        version: 1.36.1
      entries:
      - name: serverless-operator.v1.36.1
        version: 1.36.1
      - name: serverless-operator.v1.36.0
        version: 1.36.0
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable
    - currentCSV: serverless-operator.v1.29.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2023-07-28T05:42:31Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.29.0 <1.29.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of a particular version of Knative
              Serving
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Eventing
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Kafka components
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:a3c5d5fd8014c432830be912aedebbc58f9c60a8cabd87ae6c42c7b8d27db8f9
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:a59ba5aeb8e66db3d8ba9074b8e553603c59e6b0e35b57b82c4ce3de589b94cc
        - registry.redhat.io/ocp-tools-4-tech-preview/source-to-image-rhel8@sha256:98d8cb3a255641ca6a1bce854e5e2460c20de9fb9b28e3cc67eb459f122873dd
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:56e48afb1dcbf4e266ace944105f3c45042550d8d4a0528c982225a793d26d56
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:31d45f10a8519addec9aac115c5edc8a2e14f9621c875d95f123d1291776ef4b
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:b780634fffe2257eb845f8c1d25498d9c654d0327c1f8841b40d90c176a2bfe0
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:04866e16881f64ff2b5edd379eab0731c9d66a47c25ebfe79c8d8ad33d8116e6
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:61fb98802490323bb03ff4a7060334d8fbda9e9f95cd6366b29b1b6494e4ef8b
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:3b7449287adf98f0cf0a1dc74e03b69218be3bcf5538327bc34ef776cc1c69c4
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:10788ca67809e5a2d76d6d72df7dd64ea4407773999707a7681e0cf1b5c4695f
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:216f58dc8c8166c0c69fc0312b667daad14cf39b7aeb7bc0fe1e313376f1a5fd
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:aea5f200a845e1d4dc718d76465fdbb1eeac2b20ecae224d5d20a98214ad0357
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:e5db11a0adb889f907cad797dad9cf2de09c433e5522fa333193cb35575802d8
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:c4cff5ad5b0ac79e71a16b88640a312ab320817041317f50caf575f7e12a864e
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:c2cb80bce34a74afa76151d2b44dc25ece73fb274cdfc6fdf243b1b0cb7d989f
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:3aab995872336e6ccb590ef6488c097cff38a91348055031145427a00aa6b732
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:b520305189c7567b2f8ac867b20468259fb22cd3edc5b4576af2bfcdb0c11855
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:8273842d8e1714fa03c7e0809924b1ffc5ef282f79b9606bd8ac88c03e7a16bd
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a10e76aa15707e708f82c8d9e8b5179bf67780d99f3c3e2e99b84ae2a0aa4ecb
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:6342358e6aed2339333ebf37fc223a41a0a3e199a0e4a840f3f66e5accf8f233
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:594c3ac24bf59802e2d3fa127e4051277bc07b6444f934f535fddeb668ccc144
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:4408d1fcf34c0ffe9327d4c9c0bc772afa969d831bc5de1097164d10b4e8d900
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:03a3d6f0314acf4da4f3d3d85f1e273a3ef25828d28f5b22f394441d55a645f3
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:7dd11d385b3d5a178566d901650ecd45e4af395faec2141d57b98af8fc1e5151
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:a4943c0563c45ceccfda0a01e26be3a01c1437c89d927714c577bf6060495958
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:fd86ea2a34de43014dea3ba2b8d705a3d94347d3a24a370a8bfd032ed239dc56
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:f37f494956adf0892ffca34d2a4c9b839ce28dc9f2779ca32e8801fb4f0e9105
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:557e73127326be051aa6257e7daeb9721e2af04d2f06cc26cfc7a1073666b989
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:fe916630003194dbcd3894dd880cad8b15a041a5e534cd8f76fde6296db00797
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:397eb425f383c1273c7976bea69a603530676a00a61f461ceb5f48d67901c4cd
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:bd134e64c156dc5c81d14522c9c9efcc1f56b2e972ac8ecb1bb1b130c76adc60
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:ab90ec9b3cd9203fdcdd7c43abf95f48c9d7479a9c7af479eb1e4b6599ba9642
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:c12c471902b8e2596588d9e553ad59b501d979dd04f301d45f7d45e794737a8c
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:de0634d14962bdb8701a8c29e80b7b70684824462bc5203e5fa7f139839cb73b
        - registry.redhat.io/rhel8/buildah@sha256:a1e5cc0fb334e333e5eab69689223e8bd1f0c060810d260603b26cf8c0da2023
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:0229a0a4f026f04f8fa0a1cd8e7afea10117fe3f12e21986ec6089769c8afcaf
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ad399867cbb712111db6affaf5c9e3573699ed0eb0def8ce6057de02054ced4f
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:3d7801f7ffe6420595779673a170235921a9090333f8ca51d5b7f664e19c82b2
        version: 1.29.1
      entries:
      - name: serverless-operator.v1.29.1
        version: 1.29.1
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.29
    - currentCSV: serverless-operator.v1.30.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2023-10-30T14:03:25Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.30.0 <1.30.2'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of a particular version of Knative
              Serving
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Eventing
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Kafka components
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:9a28cb1d4c4f4dd7d494829eeeb7122db2174a27747f915244672c95dc8f9d12
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:4800c85e8de031bf77f6fae0e4c8e6c1a9912ccdfb77799cc2ea574c37d870ac
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:66149df8549380673c24e1487d0c2c8eb638b7e502455671ca64ed6f32ff82c3
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:a0c5a91348ec71697747113990652feb024143fbbd0cdd67e8ae672730383a79
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:2d3e314a6584fc6974517c1a992fb59f5133f54c2f6cfaa42f579c0d332e0300
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:ce4b36c90955a356a5774410356aafdac2b60795526a3488b0b76ac2ec1a018c
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:1f7e65009168383cdcb6b2caa22fe4db49e03c71574973a8086f6c332dc5e26e
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:32445e4a73bc7c1e8007775adeb147175d934b17e7d3d4dcf422104dbe80a52b
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:c9353a022e6f883b7c0b2d87b1bc4653172feb13806ebf0aa075d6f9f43c264e
        - registry.redhat.io/rhel8/buildah@sha256:a1e5cc0fb334e333e5eab69689223e8bd1f0c060810d260603b26cf8c0da2023
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:92a39ca90984cebb1938447a31c9170b52994ae98f83caa860acdb2c3ebb75c7
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:06e9e073f092dce67d4227e282fd8eea6dc9f38b90a8d444cf448d80d027b616
        - registry.redhat.io/openshift-serverless-1-tech-preview/eventing-istio-controller-rhel8@sha256:b6ef281fca2b80d379b1e28fe3f23119df8ff765fb2cca291d514053fb512da6
        - registry.redhat.io/ubi8/python-39@sha256:0ff58549041d32b685cc850aaa85da32d1ef73f99cd4b6e0f6882dad5f0399bf
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:21a638b481639c3c58e8b026d875917c30a5874b08fa4ceac40d447076b2bfa4
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:124087e6402caa375b24b976c3b9fd7d53890e20ac3b0b57ba3ac646e352c7ce
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:16939b802f4c16a3d9d941917d0363325f40cdef28f6e1a982bc95401117674c
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:e9505aba7c44cffc3f261a5dc7617a0e54ed98e78f73ea3c9c7a508560b792ed
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:e8c756b659f6187790bc1a5115e0f71791bff96630859a905bc5b75d14596e0e
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:a9c95fc208cc03625cf49bcebdaac73da823d1fa3bfae4239d7db20d3a0b8723
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:3fb6e7626d2b0c92c2a2fef35e9b8171382b4bdf54393f2ab1281ae7ab441761
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/ocp-tools-4-tech-preview/source-to-image-rhel8@sha256:98d8cb3a255641ca6a1bce854e5e2460c20de9fb9b28e3cc67eb459f122873dd
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:12bc62246142a605f4efed0f6bb8843aa2df7699cc1c5ab53debb94b0682e9d1
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:3c8bf480dbb05e09e8d07cab12e4736643f61bf9b5d98efa6bc0557eb31887bc
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:1455606c8edf611d996a48e29ead53e0bb4078f55384b7689058bb25fc59fb2c
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:6057a2d18406e00b0b7b1731870f707844e39193d4cce6de50137b77f17d220b
        - registry.redhat.io/ubi8/nodejs-16@sha256:8abeddc6ccfffbe313f87e5e069e18cb7047f2c7880ad63cf15bcd0a79ba71ff
        - registry.redhat.io/ubi8/openjdk-17@sha256:af305e087061cbaa4107ce0b304a1ef987f6b54bae3f75bfd529cfd084cba9b5
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:2264d2cd4ee2a720667d625ff42b6f8be4f2cd65b57a91eb6044da086a110740
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:69ea22c646c09208ca68345d4b3508453d8c49f7f7fd9a4ff4c4cb78be7bff43
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:a5154f9d13a1c373b8ba47e2bcfb9d4474fd03e08db4134c92aa4f049536f1a8
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:96d6f52711ba433ab2f273be8080840a88e626e989027480424d19c2fe1e9454
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ef116868c9c60e5ee246060688da49832571ff6a7901cfe143a42d048743736e
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:01c8fee92889a84a19136b0843e4fed5db31c1f2a6d997669f103373fd995e36
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:5c933641efc6bf919b1945be5a01dfb7774ae865975b1bb1f44653b3e4ca9f1f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d03f2bf8f5e7cba47102ba26ea0149ae6719076c645d5e451058392d6044b738
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:3508434d5ab6b38c8a7466256f1158ebcd75e8b751afcc5255da8c21566804f0
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:cd1abeb2067afa44a0f1eca301cce37307582d56c1147820ff86b9b64a40588b
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:85a369d5696771d0932b61760336e52ae3f7c66f9a3def5312aefcac03ba516e
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:96c135ced2e3bff434addb012368d96d0778ef1d294ba507ae7e624811e20d46
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:e97eb2bb7eb529cbfe67c14e823ffa1269d614e2d35723896e89efa12a499c6a
        version: 1.30.2
      entries:
      - name: serverless-operator.v1.30.2
        version: 1.30.2
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.30
    - currentCSV: serverless-operator.v1.31.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-02-01T19:15:42Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.31.0 <1.31.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:9221a833ecc66fe2354aee7aaec2a2470b21e2d89cb6ed287ece32edab2a2edb
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:22ffe08d6a0d5461158613f2b2a2ac6936025357e282a09197a750c42b5e6922
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:3dab790fa845ed7eda088e5abdc87799da1354884a53ff1ff728739b8e278418
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:1ee675125f7141e91f4ee57ebce082967834bbdabe1a870c221f31aeb8750be4
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:486bf60d17a5d265ebd65d590a6ee1a7e13bf3953ea797b2f6c2b5910789858f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:bf693eaa845373e1a06041626a9314ec00519a6fd2437b12f7f4f3f104864039
        - registry.redhat.io/ubi8/openjdk-17@sha256:a8165bc2cd5051a96d6937e25ed03155bbd4b731da6e58cebfe2ea83209c16d8
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:565becc77dbe3bf03d89b1399d6e1853253276a71e80b6b14069285d03cb14b2
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:433821e6c7f43da5ca2e38c0550722feb179a0bab021decd9a97ad020cb78b76
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:e49220eaf3143a8958f0cb1f22115b2e834fb7177755433bdc592b753a54a8ec
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:1fcecc11b55007b9c6969a24940a391968378e4102eebf7ceabbe7efda138575
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:b37dd2cf4b459fecb6d681cac36918e486c3bd22501df50b90e20b15fe164d56
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:b2480154951109a159acfb88c5472a312a515d6a6b060d832b3f7bfc16f90ac5
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:887480662b1bfe09626fbca8ec7c5da879d097c46dca11764ea98474ec78ab5d
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:cda04e3b7c2c1f37164733ad162ea0c4e52f14fdc8eb8bdef8507f0ed5454640
        - registry.redhat.io/ubi8/nodejs-16@sha256:f90b28179d1d5de68d42ef1a9166d70a9dfea85a1c6a4bf12dce2c59eeb2fbb1
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:677be039353b2cee51d5faa5d5a2abd6a7bd46992ffe9521eb033eea9925c785
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:3181ad775ead55f317f96466c2719b2b73c5b35d31933733da21273955d160c4
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:9e120ecdbea54eecbc5e0412c1774f8a2ee70d7fca682ecd0757b8627f70680b
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:eacdad8db61b23634c3f89bffa3f1c30ac5c7bbb5e938c6208003ff00f9595f1
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:df1aee67a161fa19d7f49ba56d53e0682cc1996436e985a12a2068fa419b54ea
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:3b01ebdb0704dc6a1a993ac076819489495ab3fe32e3c14609e42d1358856be2
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:d8bdd96c6a9b31f48c961ca438c254511db785fd34bd10b98084b702baf0ebab
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:63e08279d2852d494d4f8836c54e2ce942a342f452a23ef437df9640bb2f5bc9
        - registry.redhat.io/openshift-serverless-1-tech-preview/eventing-istio-controller-rhel8@sha256:5b38e5c71e422161e40f9c37dd2f46212cb80e8e37b00247767276b957080397
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:80361e052d366ffffe9a99893cd83965bd69a9ff81d8c9262976de145bbc6184
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:d2a34100f6309c3cbac7c9ccd452b7bdab2a533afc99463bfbf91e3ba2706d54
        - registry.redhat.io/rhel8/buildah@sha256:b48f410efa0ff8ab0db6ead420a5d8d866d64af846fece5efb185230d7ecf591
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:0a2989f688b966277652957b312df736ada43fff78fd89b17f60b2137e38f115
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c078071cd51c7fd5584039b54a2004331c05a897ee8b0689bfcb9204a8ebae80
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:8353ccb4f0d70597d61e4fd1f799d4bb74e5bc038f6b9916059e712aa70b97c1
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:3be4fc173b489d0beb768739f60274e07d95ffee86d70a8423e1f59678c805ea
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:3a4e7412322755b4c278540c965a403f9e178f138bd4ad7929c1c24a8036dd14
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:de62762efa0df95ed23df56f5520490d8a231e47ce13c433c512254ceff81844
        - registry.redhat.io/ubi8/python-39@sha256:4f35cbcc6f4e108c69f7d2b70cce5afb138741b0c961506376d524e001b922f4
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:99f5dcb92a33c047f0b68236a3a42ffaec78037dd0a5eb79d1144d62d6bfcf43
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:2572d98363cca2dd94984d6cefa573c5054fc206ee70048f901ad4b846d8dd52
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:dbd9c0cac86f79b8d91ab65ae7543854144781d4cb17ef16183931ec9bc234a7
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:7941b421b3e18bda8c565305241621fcdf1212cbbe0dc02f38f4ab0565d2bfcb
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:af5c30fc9f631d9ed110641b815315d8b4aef539f687277a3821d9939ca5dd46
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:ef7d28580e75d746449a90cc0e5343ef3f4ddd75bd9fb866d97836bfbc3b314d
        version: 1.31.1
      entries:
      - name: serverless-operator.v1.31.1
        version: 1.31.1
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.31
    - currentCSV: serverless-operator.v1.32.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-08-01T10:56:58Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.32.0 <1.32.2'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:16c026f09b377c44253cdb902b9a5fa0ae766fd0185cca1863798a212e0daee1
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:2f0f84767f7d76dbbf4387ca2e272766267a2c46b5bc29f5d01f05dfa39cb944
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:20566173a8e907980e603d82fae4d96fcdfd5c43d9004c2e506cf0e09a47b91e
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:04e5b3882223d45a5f5383ce6481efb7b2cbedea57ebd894caafb1454ad8c09d
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:e1e2273fdf6384d95e1128e54c0794d7d620291a272bfb1f91e0c2723be7f1b7
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:22b4d47725ed8cdd663ee26a03f31198fff9e71dc703532e70c42bdfb4840811
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:e7e6cbb46720eeca3f0f4b9829dfeeb3db373a83c89cd808b667ded4d5e139a4
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:a5a7b2be50f211d4d9d555b2a2e2302096c139ff5011d984cc386f1530007f16
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:0729de0f3f03ed220216c6cb718fc34beece0ceacd926b5e2480bdfcf44df5cd
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:ab9a1b952e9603b4ecc5f197715a1d871abc25ebb0c5e6470d4f0b365ae94276
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:2a30ae808d5c32d812406ff4ac9b472e82b6bb0555e62eee4c2c6a972f8750c2
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:b008da825ee2429f04cf39258389740c68056f82cf4bde0cbd3db850b6256885
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:5930caa25ea875d29dff0ec5e235eaa13cb6dc98980d8045335eba1a09ae761d
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:da5f813dea324ffe9679524ea7f3801536fa48f7dbbd8da56a8492b916b15fc3
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:990962ad245a6d62cd11551bd4e1f8b14dc08abbebe47beddb2f4874f78d4f47
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:b0db9f0fc85fcdcd5da980633a3a44be2bbabae068f89d0965263de38789e812
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:9d36cde68bd822016e967148eebd2ef3e7ffab9dcdd37d17f33017e9b703d9b2
        - registry.redhat.io/ubi8/openjdk-17@sha256:a166bd2fedb99cb62ebf4695ab46e19c7e451f887940cdf0b8245c81e403aea2
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:9b6524f9c0e7f78e6ffb372a8587eaf5f4f1281eaf64f93dd00b63be84a4c21e
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:42f31eb8f12dd36541c8c0b6572d942c62e2064d5418428482d676d117a6c426
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:7c8cefd5823ad953f112b905c651f12a97491aafc008b2ba0111f0c9e83e50ae
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:3be20d63d7e82f35ac88274532d6ab7839f9a7306ca272e7d532ac03215faea9
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:46b8dc1c8ea6c40401d4d1a9868ab3edbe8b9e92364a43add6722489e2eefa04
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:ab09a1710aa2331e24df9e7283f3909fdb104c1b314dea88e919d0a396f334d3
        - registry.redhat.io/ubi8/openjdk-21@sha256:bb95ea866269ac8652f666ee1587d71f4f43da33493c2be7f6a9548bcfd2407c
        - registry.redhat.io/ubi8/nodejs-16@sha256:69439fe83fa957089cb6fd3c1042ab71d719f81dd457a5d1fbc56c1e3ea600da
        - registry.redhat.io/ubi8/python-39@sha256:8c96e8648cc6651d0cefbd20496a89f208621601f9664554d3fdeef8ced25453
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:e55c70591c0498b6b85f0d68e304e88b830ea0b35219ba8881143555757be4b7
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:25cbdf78b67e7f5a360313ccfbe3b6e2c98873b4e9ddee46c115a623e77c7abb
        - registry.redhat.io/rhel8/buildah@sha256:55f5c7d88c77e1ab5a2ab8e877172b2b66cb196e21f10dc45148470d0ee8bbae
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:1a1de677297525c0c5d317fe8a7a42684fe3f1b422a7d6bea1f70f0924653b08
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:0507ea8745267fdb53329a5274d59d23c387225d2594af51dd2603b4ad99db0f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:4c9faea6fd162db3340aee8c11af11a08c9d79bbed104967c25df26c4933c8ae
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:14ca32e999767776c3e585ff5eb1aa568f7d76fd3e5ae24cae05ce7564678080
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:8033b7c1e903ba313134645c626eb3bf7ab1a0d68dc2575bc8e26884ad4f805f
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:e70dd3a9e067075eb56e5f2772642545e308348cd49b89357c02de3979b1a5be
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:3ffcf746df914f0949477b46f90cb2992bb63f8198d73ffd54208e7a777a9cb1
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:1befc9969edd2062b3df69f45aee9771b33a2659c3a14db0aaaf5c3df985af58
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:0551f1657dc158bd41eb2338dd6514d55c39fc77300cdf1744d8f5ab14e9ed1b
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:4933b76d95d411d597d4dc460fc46c92875d28e107cda661946206a9d934b8b5
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:70c26c7ac48d1eaaf1f95ca5c5b8ba8ee6b047a3c9c4a2c2db9c49a506bb2205
        version: 1.32.2
      entries:
      - name: serverless-operator.v1.32.2
        version: 1.32.2
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.32
    - currentCSV: serverless-operator.v1.33.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-11-20T08:38:34Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.33.0 <1.33.3'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:e1cb1c24333605e52364456f9bd8d86d787d1bae17ef75dd04be61257fdb2256
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:6dbfa2573a2c78e450d3c1c04c6e7cbc50d9272233e828fe9b42e575ecf386c0
        - registry.redhat.io/ubi8/nodejs-20-minimal@sha256:acee3b00ccf8ff49bcebcf10e028e34a84d45f28f855363bafb3b93c75e84f11
        - registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:e1cb1c24333605e52364456f9bd8d86d787d1bae17ef75dd04be61257fdb2256
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:eb95deb2105f14c6a9f3bb2cdd7306b446109a9b142a7bfc726b6429f270d9e5
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:ea9e4de6228d5c3ed494c2ce5317ec290543ad40bcf7c43c2a3d6565e5d94e5d
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:359f8b9fbb4c3c2f79e8c70f3b50c994693b930f6f078c34c87112ebb93037e3
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:c3dee2598088bda36ef474f4748943a29169ed9589aeb0f77bc79a7f6bef9907
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:672022b6f2a080483d9d6f6f4a5f42e52e522a5797105f3465731809c6cf4947
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:28c9d05a3ba73b76f293e8f4902b574deb44f77a2bae8cf92473b8a8b10cee92
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:a0399a993155e755409355740b260d9b0019559a62be3b6253d8693a4fb59286
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:42d16e4d9f14a427e7a5b7e3a4ea07c258e0f2c74984baeee3b4926a8c5e2640
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:a8821b59d495e24598b27c6510b32ef19de2bb04ea1e568fc5afb9135d772036
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:9b850afa750a67120f68a49f4834ba7f105563df22cbcbc563317fe1545640d1
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:09060de091d617bd4606f7a289b4528265ae0b7a04737f383e79efd5b0d3eb79
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:dd1f7c850f7cf2a4d70403deedd8a57ecc6e0ee5455d68fdfc49ff175fe42dde
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:93b2f5211e79abc38bf92f0b8fb8adf046303efa2fe62c47a2fef9dfb5662f55
        - registry.redhat.io/ubi8/openjdk-21@sha256:095175ef8573de8055e67bd5c8a21098448ac73db4e0d6ccd5269dd36b4d263c
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:b871bd08b782c9168e9a54340efffc0b0effadde8079915b848e4eff72bb3e95
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:879fca768b46676f1df89de1cdc4ded8dac310209c146c2f381ed1deb19b06be
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:993597ff83a4d62a23f642fadc92f182944b953770716b73b7d19492b5a462c3
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:eab4aa93b9d175045b1f01bff09681107d839285e0d633cd5146d2a2efab373c
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:af66cc7dbed88907d96a390fe5c8807a2a60bbe3276b8cd13b4b061cbaf7b162
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:b12272e1a9cb4f4bb62c4badc62b0bab110abb99484612dc781c46bfeac13cc0
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:b2be693c71cfc4ce934d5b01efb863dc6e391f93a6b15fe7eeb15914063a8294
        - registry.redhat.io/openshift-serverless-1-tech-preview/backstage-plugins-eventmesh-rhel8@sha256:28ae309e4cf7feabaf78ac6067e1173f9f5b33a71c06d6c27cf25e01a0b1ed64
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:05e3dd6255294b691783523f7722cfaf13aca1a1e1dab344a25eeea72aaf9f9e
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:e9b38f1df225238836aa829f3a2016372859fb8709cdd2b30496498669a90be1
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:510e98837a549f489b2d74939f3b9c7ed90a8260ba914910f681d5c0dec7655c
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:d8ae7bc888f11121fbbba65074389b879bc22453bbabc551d79d0cee68fc33a3
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:5c7f29554d04eb9348932953344f2ea0806d14e7704355619381bf395d993f58
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:76a47d100de1f2afa23803bec503d9440605db03c7179b231f963e5663a3d12a
        - registry.redhat.io/ubi8/python-39@sha256:ac12e4706e0fdc464e791a3f274474ac1bb14d5af8f217c440fbb0c76d6a5d00
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:ace72435e3e63ab21dff2345f958cf6bd171537ef2846c5d5d91cf821877140d
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:9b94875e0e5d7abdb3b9ba7b04e217a2a9b7b4410d38d1608c86d1c21f929cc1
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:56c6ac59a68976741d2ba676a53e59fc494d04dc44b513d72fdf8a9ad894a54d
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:62bf4bee0ea17ecc60366ecc71431adc79d6f0a5b9f59489913b474128c4adb0
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ccbda564735555785815f27f8f25ad9c6c27dabe4e262586fc51609331692527
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:a9c84ac7e7c3fb1e5dedaf949d54a3aee8a62397d2bd6022094671de00b651da
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:2bd241898bd54dcb248ae86470430042dc545d45e7519ed470a1a9c8c63bdc37
        version: 1.33.3
      entries:
      - name: serverless-operator.v1.33.3
        version: 1.33.3
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.33
    - currentCSV: serverless-operator.v1.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-11-20T08:53:53Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.34.0 <1.34.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:a4b3d75903de1929326041e9885213074473f07fda91b299548d45680c295c00
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi8/python-39@sha256:ac12e4706e0fdc464e791a3f274474ac1bb14d5af8f217c440fbb0c76d6a5d00
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:633ea30ea3b6d5152bbdb61acb4271e5905d50fbe3c731fb4c8d8640a57f1970
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:586fc8ed88f91298dbff31bc2fc4c63e0d6d9356e3358b2f853f49511c3ff9bf
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:d8b94bbce98325fe8079c11c0b7a6c59bad8a7edd0532a7d27ef1632ff0de708
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:ef7f34a025e40f93fbd51f3a4596998a6dc508db824f137ffa1fbaf90e88f4a8
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:2adb0124f04d6c4a552ade79dd1499c8ca997edb45855bf438ca8043b2709e19
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:f94f0e8d21a6b6341b1db00513a1a2ed5cd84fcd898a376ecde942e940a518e7
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:01ea7a06332d60ac16ae0a27adcaaede3f680a98a22ff995259a6c42ac0e358d
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:b07010595201d7cfe0ec4f18e612d323eb38464c74eb187f275617a744d9df12
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:144883b09d8a999fd4d2a109d074e44d7d8b959987512379aa71c71e4060ca28
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:a9e8583c19ca3e01593f8c0e54731cb73df9b46b2ed47a2b1c711828c2295f82
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:76a47d100de1f2afa23803bec503d9440605db03c7179b231f963e5663a3d12a
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:6d042c614431aa9a9e77cf8ba0064423e903f7c418e367ecc23bc02a24a796c0
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:86ffb8be1762782d643c147cefcfbc2b4723b0f97d017314a99bd2275d05a1cf
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:30cb41f09067959cfe3792e070fca62820fcca4035683e5c43dd46d1cce97c01
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:45bc50a993d04b775a153e985a43d587d5455883fc4c6f31b772408297cf9040
        - registry.redhat.io/openshift-serverless-1-tech-preview/backstage-plugins-eventmesh-rhel8@sha256:09dd94090eea9cac7a1935a6d671cb97fd9fd5714c715c9d2aed461e468ca54a
        - registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:a4b3d75903de1929326041e9885213074473f07fda91b299548d45680c295c00
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:946daeff59b41a6f8c4842f8c2668f97ff1b49e2a8afae5d64ba3baa916b1d95
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:75228e70fb1383bbdbcc64d0cf24bf5d977c09112e3f32316824188264c3a2bc
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:33d62d13e16f47db9951ac826671f39fdc048fe7c3fb019eeb4bf9bd0f814222
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:d5ad3c9db7ce8e4fc950261f7c388476214a03d92e0043c7bbc4f0add2b1e69e
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:b9d85323e110d8eb6600223c38f9d4cbe89b50d59e03bf1e6b8808f7cef9a3a5
        - registry.redhat.io/ubi8/nodejs-20-minimal@sha256:acee3b00ccf8ff49bcebcf10e028e34a84d45f28f855363bafb3b93c75e84f11
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:512c4d93dc13cc53d923a6990770a7527d6fb657278e2c5d8f0ece3cd5020155
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:0f126893074cd0398654c4018df6e882b7610e148c82a41ac814f4bf3830bf9d
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:a5950f81cbc64d1edb64ae7f7f120c3435d36afcd14ffc34ae522af297d87e1e
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:3d17f12cc9da578549254ba1b8934b5afc66e453b1d5d667bb90281e45b8795c
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:c0376504edae12e7af6b6cce02ac53b6b374f5058ebefcbdbd32628b551bb0c4
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:1c82d20ccf5eb5f8e7725839f1060f4efdfd1c1b4773a99bd905e690a6e36f38
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:1252960c0c0a4dc7b39090cde0fd5742a19453b1a65a08b737bc41d044a0882a
        - registry.redhat.io/ubi8/openjdk-21@sha256:095175ef8573de8055e67bd5c8a21098448ac73db4e0d6ccd5269dd36b4d263c
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:d40daf67b36387faa9f55e70e50d10d02accfd19470c0ac11b78bf969947fed1
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:84994eabeb7342a0771c7cb1bdd70e7ee3928dd45874eafa64cde5bb385fe964
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:0f8d006d0e7468783739942a1c82c7aec376c9948f23841132c8ddece8ecba06
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:b801ed7b1cb32d4a1522a959e97f280f2eff6fc92e47fea2ea022f180d37831a
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:e520d1b6b87f85dfffab5e53cfb70562ab7edf450061c07609da39e3b2498fdd
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:4b6f161c14c68536ec2c7679d30681afa9a1fa2d512548f0f591791a10a8b894
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:76480201188dd142206040af845bd55b44155af92e0c5ee9bcdb8660bea0cdb8
        version: 1.34.1
      entries:
      - name: serverless-operator.v1.34.1
        version: 1.34.1
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.34
    - currentCSV: serverless-operator.v1.35.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.35.0 <1.35.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:f23d3ab058667c6fe9ee44dc249aa7e8449c1bb9581380ddbc66624511b286d2
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:b2de6b658a1595084a202203c90704c3f9de018a7a904ca276112ad606e374f1
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:0751b9ea7ca197480c4cbca17654d0653d1c0c760ca3c818c2a1018dc952cdc8
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:da8d975527ba100313b92e87d6d2690fddae9988aa8d82d728221295955b7224
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:ab8f85bb18311b0bfee00561ef6b5d9faf810358c954ac2c6f39cf2a7739e07c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:6b842b4eddfc71a2c8f6c1fcebb8f23a05f50c95be1ba2d20a1be660d3014c8e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:e743dd93b7ae855825716403979db3f34cd300caad130e7d3011172b1aa4eaf7
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:0e58edf5ee340fdfb282faeb6c9bee8cd6323e7f29a9b282dbb91c2622131a9c
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:477b1ac0db0abf1e16ae31bb22fdb73dd98f4dbd243fa0a7e6e94852b16a813c
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:b66bb9531f05cd46360f787955c849208496e3e45541cb50b23bfb6b68570aa8
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:9b176477be4c782ce60d81a72bacb566069a499579ce6df5a987a0841f05ad5c
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:86386b54cc10743023884193b7d6c7159b90052bc5e4a0608a6cdfd47865c966
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:cb2ebe98481808e788ca70e7f0a9b3402818b622158f327b99a1e87c23226d2d
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:d0650e4d7e72241e07a3796f20931adba231fbf5e401a45d794761f2c5afae48
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:2cbe84afebac23b5397d144b4d5660af33290df451141fa3625607b6033983e0
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:020264063c50f2a4590f3d7524b271f0f77350a1f8fa3c35ca6e28f0162aa9a5
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:d00a0059f6134b2519f9b1e034b81a943cae7ce391109703f24db21e56ee923d
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:d68fae6fc07efa9da9af0532aadb49f897a15afcdf9a98dc7517ff5e17735454
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:3130a5159adb9b4eaab75a0209d2ac639156d5e9f2346a8b2e6e0e22fad4b45f
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:80a155d2637f47e249fd69760b250c9b7a1eeb6781de9308b54162bb3e1bce71
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:ffaad2214fa897e89e92e574eb5cd1c671edab7f6eab87cf2ffd4ebf162b3cf4
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:b5db9a51bbb46640b23712c61e9c5a1c7bbdbcd9a53b0b92fa36b80dcaffffc0
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:f23d3ab058667c6fe9ee44dc249aa7e8449c1bb9581380ddbc66624511b286d2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:56e7a0158a713c70765ead847732db4e5f8439aacf31e9a4a22672b583028064
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:e7aa54986c3d388c1bb00828c9ccb294ca0b6c26376c7593ed6bfddda0ef4e5d
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:ab0342b325ee67cc2a7d321aa1eda9c6797dc91fe2b6f26e3cbab53fadf564c7
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:732827652b7a7a49181e86d561831001c28b72b7e12d5e3c1d71b2e496003229
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:018459cbba2e3169bc4be6564cf3747c9cd0677445912abf237f8fe6d58e46ac
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:6f4fa254d923f3260ab0f7d6a0b459554f83b1abfcb575a5ce2c0135fba32427
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:6722aecbbbefa80dfaf93143e033c562cea0982514e2eea27654dd4638e1d072
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:2980eca437e4c1cd22a2b6c2ccae9c4d6b55921b3d575fbd4b9241f1f0a8a3a4
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:dfcd11686f161d44786a205d12f9d628d9e847d40ffdb628ec9aa4a0da0b86f0
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:acb28d84cf03b8551efa32646864005dc3bc9bb341a696544892dfd0c87fbda6
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:818dc7bd69b453ba5882b4ccb82841f2d48c7552368417bac20bef1df64f52ff
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:4ca1bfaf2d7e9477583cddfa40f0b2ce59c20ff9ab20022038774dc3cc49662b
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:f56d21089566330ecbd9aa8acf14da332d3db7360aaf6da00c9a19df83f7334c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:8a4e5e0934197c32e0dfa1efc922d750940587fb61e0df49d476a2d035c53413
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:86074674b7926ad9bfc95556e50820792fcbc5e11ab02879378625ffabdd1889
        version: 1.35.1
      entries:
      - name: serverless-operator.v1.35.1
        version: 1.35.1
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.35
    - currentCSV: serverless-operator.v1.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.36.0 <1.36.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-ddb-streams-source-rhel8@sha256:a903480159ebf1d8652eeaa702d9885fb76bb106c02d75cec5c82c31d77d6410
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-timer-source-rhel8@sha256:8fed5ef9a1f39a55077bb5026c8854d0ee7fc8d8ff5da277ce9c1bfc2ab2474b
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:77e4689b2ce7a713e905553e0e42b8e128045f40cb6c6d4ee082f126cbc46b00
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:3d780ad1ab280b5092b90e277485d589da2a18e8b08090af7d95aecd17af545d
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:429d5d9568cca31fbe83c663276a300286af4cdbf3d35685afdbe5f6f2f8c6e0
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:4344716eb2940c6a3fa9a278250f936602bb1549fece6037912837719c178b24
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:f2d112ba9fdfccea988589b99b463c8a7157475f7a3f03678e8999e65b678cc3
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:5352e4eb4e426a41521276decadf204b6d23b4762a298e4d3dc73969e3eddc68
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:187fb9d64b53c55a029ffce3c0f9e7c3304683fa8a0e832cfb7134e8a3d0c237
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:6210a946f373b9c36c9b31293debe6fed090008b8234c0ba17d37fa81646172e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-transform-jsonata-rhel8@sha256:3af4c48030263eb5769e5f36a6dd13fd390d8566efaab37532f39e1df1c95c67
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-sink-rhel8@sha256:d9b944e64391a996104afbe8a21f66316a948e9fb7b91734e861847285be59c7
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:631eb3a8dd9d83ceafa7d8a4da4b1e2d4f2740e6017c766828dcc7575cf1322f
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:a6f565d73375e827d830dca85838f4c4cb7f08b0c078acb0c0d567b25f0d256a
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:a47b6b5237a87ccd90e5a49a53d06485fd55c87eebf12d01c335b87fc3a8c494
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:ed60e889a677fb139f2eb4bd9ab4fda6ecc39fd050ca2b6cc1c19717e428535a
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:074ebfc4494581611647ba90049655a470ca43743610a27be77db66a8633bac6
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:dcfe5fbc57b9f8520adbff368c8bf29eb491c9c26d1a1737a1722cffea83eb1b
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:f3d87a7fa5b5f2818effcac3304bd56eddcdb0db97a18ac0f35a06fc7e223e7d
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:4274004b6289fa9bea3fa6e56d64077a328a6b5e63a9f0e9c41f6d561f682bd7
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-source-rhel8@sha256:27e74287774cda208f667bc2a386debe6b6255e8991abfb37c959c8525f2af53
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:3b04e9d8cd4b5e43a07ece838d10478723a6c6529350f010550b4a0d1d0950d1
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:ee6533b71cf280877c9d7489dfd821c9209c6802bed7fa3fdb4f8f00fbeacb09
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:6b7316cce3432d391adc6ac38c92ff0ee86dd7ae0101957cc65cd338cc69e352
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:c39cb88afaddc295d93a3774b8a20d001b82392df540fb502ee02316d82460a9
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:82e69a1890042f5636ac7bb645b1aaeaba0a4894851eccc1fa1c63296918e31a
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:52ba6b1252b6f3d41e2a0b3b880cb768010ab9dd925ac34596f6adaedcbbe07c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-sink-rhel8@sha256:afa634e029bcf79a972f49372498d44b45bcfbdb6eb28ecaf4f0b0b2a0a1824b
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:f4d86b196457cdeebdc380b6f1b707c10f8dee789c7779c90ee1332e511af8c4
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:259e5dab50d36cdbfc117a42325128c4e16f34c048ba64b36d4a89132ddb4c1a
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-log-sink-rhel8@sha256:30b8045ef5f925a36c58941386f1899bec125d005a08aeeeba5edd49b0859149
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:3fe326ad9a00b112fc888ffd345c8d5042f0a22cf32ea575931192c39ce1fece
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:e3ae1cc6012f19631c012a07b113fb015977d2ee4f9faab3207fbc2464c25aa7
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:bf82b37ab8fd29c7a7cd846008a16d14b4cfdc1c09df4f731d71be3957e90123
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:d6f338e08fa937c8a4ccfec692424de4acded565f7ca3971ea507dffeb351503
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:4aeb68a22f63300e54391ec6b4cda8c4e665d904b5dcdd3c40155040525742a0
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:596a7272d6de714d7dbc6e00e9e50a2872846c269547465252d953172ef0b009
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:0c2cdf917bbad1f5bca8a2a2ef12a96cfcd0c77c190cc35687eb57169b54ab6f
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:c7c2e738e0d9da1d2d1993f68938e72c4709b1f28a54fa92224ef6a0e456b0aa
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:a1834b620f875681388f0ae224dd0771773a8b61b1d8e5214a942f232854b87b
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:9b328fff0cc8545fb9bb1f24f11d78cfac3ee76dedc11fa4def4262d79a60e8c
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:78fee4c596f2848487ca704da3fa1074e3c9c99b3f96f66833aeba63836d7b93
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sns-sink-rhel8@sha256:51b5e18e8d4c1ca81d7c5ddfea6352a3887f757acfcc53bd2f8bd3b01337ed42
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-source-rhel8@sha256:88323da16360981f27b8e4686fbbb64053e4e0e3c79c65d695edb317b334e1e0
        version: 1.36.1
      entries:
      - name: serverless-operator.v1.36.1
        version: 1.36.1
      - name: serverless-operator.v1.36.0
        version: 1.36.0
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.36
    defaultChannel: stable
    packageName: serverless-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-pipes-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-pipes-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pipes.services.k8s.aws/v1alpha1",
                "kind": "Pipe",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/pipes-controller:1.1.1
          createdAt: "2025-09-19T21:15:33Z"
          description: AWS EventBridge Pipes controller is a service controller for
            managing Pipes resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pipe represents the state of an AWS pipes Pipe resource.
            displayName: Pipe
            kind: Pipe
            name: pipes.pipes.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EventBridge Pipes resources in AWS from within your Kubernetes cluster.

          **About Amazon EventBridge Pipes**

          Amazon EventBridge Pipes connects sources to targets. Pipes are intended for point-to-point integrations between supported [sources](https://docs.aws.amazon.com/eventbridge/latest/userguide/eb-pipes-event-source.html) and [targets](https://docs.aws.amazon.com/eventbridge/latest/userguide/eb-pipes-event-target.html), with support for advanced transformations and [enrichment](https://docs.aws.amazon.com/eventbridge/latest/userguide/pipes-enrichment.html). It reduces the need for specialized knowledge and integration code when developing event-driven architectures, fostering consistency across your company’s applications. To set up a pipe, you choose the source, add optional filtering, define optional enrichment, and choose the target for the event data.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EventBridge Pipes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pipes
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EventBridge Pipes Developer Resources
          url: https://aws.amazon.com/eventbridge/pipes/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: pipes maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/pipes-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-pipes-controller.v1.1.1
        version: 1.1.1
      - name: ack-pipes-controller.v1.1.0
        version: 1.1.0
      - name: ack-pipes-controller.v1.0.25
        version: 1.0.25
      - name: ack-pipes-controller.v1.0.24
        version: 1.0.24
      - name: ack-pipes-controller.v1.0.23
        version: 1.0.23
      - name: ack-pipes-controller.v1.0.22
        version: 1.0.22
      - name: ack-pipes-controller.v1.0.21
        version: 1.0.21
      - name: ack-pipes-controller.v1.0.20
        version: 1.0.20
      - name: ack-pipes-controller.v1.0.18
        version: 1.0.18
      - name: ack-pipes-controller.v1.0.17
        version: 1.0.17
      - name: ack-pipes-controller.v1.0.16
        version: 1.0.16
      - name: ack-pipes-controller.v1.0.15
        version: 1.0.15
      - name: ack-pipes-controller.v1.0.14
        version: 1.0.14
      - name: ack-pipes-controller.v1.0.13
        version: 1.0.13
      - name: ack-pipes-controller.v1.0.12
        version: 1.0.12
      - name: ack-pipes-controller.v1.0.11
        version: 1.0.11
      - name: ack-pipes-controller.v1.0.10
        version: 1.0.10
      - name: ack-pipes-controller.v1.0.9
        version: 1.0.9
      - name: ack-pipes-controller.v1.0.8
        version: 1.0.8
      - name: ack-pipes-controller.v1.0.6
        version: 1.0.6
      - name: ack-pipes-controller.v1.0.3
        version: 1.0.3
      - name: ack-pipes-controller.v1.0.2
        version: 1.0.2
      - name: ack-pipes-controller.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-pipes-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: PingCAP
      provider-url: ""
    name: tidb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tidb-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pingcap.com/v1alpha1",
                "kind": "TidbCluster",
                "metadata": {
                  "name": "basic"
                },
                "spec": {
                  "version": "v5.4.0",
                  "timezone": "UTC",
                  "pvReclaimPolicy": "Delete",
                  "discovery": {},
                  "pd": {
                    "baseImage": "pingcap/pd",
                    "replicas": 1,
                    "requests": {
                      "storage": "1Gi"
                    },
                    "config": {}
                  },
                  "tikv": {
                    "baseImage": "pingcap/tikv",
                    "replicas": 1,
                    "requests": {
                      "storage": "1Gi"
                    },
                    "config": {
                      "storage": {
                        "reserve-space": "0MB"
                      }
                    }
                  },
                  "tidb": {
                    "baseImage": "pingcap/tidb",
                    "replicas": 1,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "config": {}
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: pingcap/tidb-operator:v1.3.1
          createdAt: 2021-2-15T07-18-39Z
          description: TiDB Operator manages TiDB clusters on Kubernetes and automates
            tasks related to operating a TiDB cluster. It makes TiDB a truly cloud-native
            database.
          repository: https://github.com/pingcap/tidb-operator
          support: KanShiori
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup
            displayName: Backup
            kind: Backup
            name: backups.pingcap.com
            version: v1alpha1
          - description: Backup Schedule
            displayName: BackupSchedule
            kind: BackupSchedule
            name: backupschedules.pingcap.com
            version: v1alpha1
          - description: DM Cluster
            displayName: DMCluster
            kind: DMCluster
            name: dmclusters.pingcap.com
            version: v1alpha1
          - description: Restore
            displayName: Restore
            kind: Restore
            name: restores.pingcap.com
            version: v1alpha1
          - description: Tidb Cluster Auto Scaler
            displayName: TidbClusterAutoScaler
            kind: TidbClusterAutoScaler
            name: tidbclusterautoscalers.pingcap.com
            version: v1alpha1
          - description: Tidb Cluster
            displayName: TidbCluster
            kind: TidbCluster
            name: tidbclusters.pingcap.com
            version: v1alpha1
          - description: Tidb Initializer
            displayName: TidbInitializer
            kind: TidbInitializer
            name: tidbinitializers.pingcap.com
            version: v1alpha1
          - description: Tidb Monitor
            displayName: TidbMonitor
            kind: TidbMonitor
            name: tidbmonitors.pingcap.com
            version: v1alpha1
          - description: Tidb NG Monitoring
            displayName: TidbNGMonitoring
            kind: TidbNGMonitoring
            name: tidbngmonitorings.pingcap.com
            version: v1alpha1
        description: |
          ## About the managed application
          TiDB ("Ti" stands for Titanium) is an open-source, distributed, NewSQL database that supports Hybrid Transactional and Analytical Processing (HTAP) workloads. It is MySQL compatible and features horizontal scalability, strong consistency, and high availability. TiDB can be deployed on-premise or in-cloud.
          ## About this Operator

          TiDB Operator manages TiDB clusters on Kubernetes and automates tasks related to operating a TiDB cluster. It makes TiDB a truly cloud-native database.

          ## Features

          * Safely scaling the TiDB cluster

          TiDB Operator empowers TiDB with horizontal scalability on the cloud.

          * Rolling update of the TiDB cluster

          Gracefully perform rolling updates for the TiDB cluster in order, achieving zero-downtime of the TiDB cluster.

          * Multi-tenant support

          Users can deploy and manage multiple TiDB clusters on a single Kubernetes cluster easily.

          * Automatic failover

          TiDB Operator automatically performs failover for your TiDB cluster when node failures occur.

          * Kubernetes package manager support

          By embracing Kubernetes package manager Helm, users can easily deploy TiDB clusters with only one command.

          * Automatically monitoring TiDB cluster at creating

          Automatically deploy Prometheus, Grafana for TiDB cluster monitoring.

          ## Quick Start
          You can follow our [Get Started](https://docs.pingcap.com/tidb-in-kubernetes/stable/get-started) guide to quickly start a testing Kubernetes cluster and play with TiDB Operator on your own machine.

          ## Documentation

          You can see our documentation at PingCAP website for more in-depth installation and instructions for production:

          - [English](https://docs.pingcap.com/tidb-in-kubernetes/stable)
          - [简体中文](https://docs.pingcap.com/zh/tidb-in-kubernetes/stable)

          All the TiDB Operator documentation is maintained in the [docs-tidb-operator repository](https://github.com/pingcap/docs-tidb-operator).

          ## Community

          Feel free to reach out if you have any questions. The maintainers of this project are reachable via:

          - [TiDB Community Slack](https://pingcap.com/tidbslack/) in the [#sig-k8s](https://tidbcommunity.slack.com/archives/CHD0HA3LZ) channel
          - [Filing an issue](https://github.com/pingcap/tidb-operator/issue) against this repo

          Pull Requests are welcome! Check the [issue tracker](https://github.com/pingcap/tidb-operator/issue) for `status/help-wanted` issues if you're unsure where to start.

          If you're planning a new feature, please file an issue or join [#sig-k8s](https://tidbcommunity.slack.com/archives/CHD0HA3LZ) channel to discuss first.
          ## Prerequisites for enabling this Operator
        displayName: TiDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TiDB
        - SQL
        - Database
        links:
        - name: GitHub
          url: https://github.com/pingcap/tidb-operator
        - name: Documentation
          url: https://docs.pingcap.com/tidb-in-kubernetes/stable
        maintainers:
        - email: zhanghailong@pingcap.com
          name: Daniel Zhang
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: PingCAP
        relatedImages:
        - pingcap/tidb-operator:v1.3.1
        version: 1.3.1
      entries:
      - name: tidb-operator.v1.3.1
        version: 1.3.1
      name: stable
    defaultChannel: stable
    packageName: tidb-operator
    provider:
      name: PingCAP
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v0.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: gcr.io/datadoghq/operator:0.8.4
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          olm.skipRange: <0.8.4
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - gcr.io/datadoghq/operator:0.8.4
        version: 0.8.4
      entries:
      - name: datadog-operator.v0.8.4
        version: 0.8.4
      - name: datadog-operator.v0.8.2
        version: 0.8.2
      - name: datadog-operator.v0.8.1
        version: 0.8.1
      - name: datadog-operator.v0.8.0
        version: 0.8.0
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: gcr.io/datadoghq/operator:1.18.0
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - gcr.io/datadoghq/operator:1.18.0
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.2
        version: 1.4.2
      - name: datadog-operator.v1.4.1
        version: 1.4.1
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      - name: datadog-operator.v1.2.0
        version: 1.2.0
      - name: datadog-operator.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: datadog-operator
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: servicemeshoperator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: servicemeshoperator.v2.6.10
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"maistra.io/v2\",\n    \"kind\":
            \"ServiceMeshControlPlane\",\n    \"metadata\": {\n      \"name\": \"basic\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"version\": \"v2.6\",\n      \"policy\": {\n         \"type\":
            \"Istiod\"\n      },\n      \"telemetry\": {\n         \"type\": \"Istiod\"\n
            \     },\n      \"addons\": {\n        \"prometheus\": {\n          \"enabled\":
            true\n        },\n        \"kiali\": {\n          \"enabled\": true\n
            \       },\n        \"grafana\": {\n          \"enabled\": true\n        }\n
            \     }\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n    \"kind\":
            \"ServiceMeshMemberRoll\",\n    \"metadata\": {\n      \"name\": \"default\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"members\": [\n        \"your-project\",\n        \"another-of-your-projects\"
            \n      ]\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n
            \   \"kind\": \"ServiceMeshMember\",\n    \"metadata\": {\n      \"name\":
            \"default\",\n      \"namespace\": \"application-namespace\"\n    },\n
            \   \"spec\": {\n      \"controlPlaneRef\": {\n        \"name\": \"basic\",\n
            \       \"namespace\": \"control-plane-namespace\"\n      }\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
          createdAt: 2025-08-11T09:32:44CEST
          description: The OpenShift Service Mesh 2 Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh 2.6
            and earlier releases. OpenShift Service Mesh 2 is based on the open source
            Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2.6.10-0'
          operators.openshift.io/infrastructure-features: '["Disconnected","fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/maistra/istio-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Istio control plane installation
            displayName: Istio Service Mesh Control Plane
            kind: ServiceMeshControlPlane
            name: servicemeshcontrolplanes.maistra.io
            version: v2
          - description: Marks the containing namespace as a member of the referenced
              Service Mesh
            displayName: Istio Service Mesh Member
            kind: ServiceMeshMember
            name: servicemeshmembers.maistra.io
            version: v1
          - description: A list of namespaces in Service Mesh
            displayName: Istio Service Mesh Member Roll
            kind: ServiceMeshMemberRoll
            name: servicemeshmemberrolls.maistra.io
            version: v1
        description: |-
          Red Hat OpenShift Service Mesh 2 is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh 2, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh 2
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh 2 provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh 2 supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### Joining Projects Into a Mesh

          Once an instance of Red Hat OpenShift Service Mesh 2 has been installed, it will only exercise control over services within its own
          project.  Other projects may be added into the mesh using one of two methods:

          1. A **ServiceMeshMember** resource may be created in other projects to add those projects into the mesh.  The
            **ServiceMeshMember** specifies a reference to the **ServiceMeshControlPlane** object that was used to install
            the control plane.  The user creating the **ServiceMeshMember** resource must have permission to *use* the
            **ServiceMeshControlPlane** object.  The adminstrator for the project containing the control plane can grant
            individual users or groups the *use* permissions.

          2. A **ServiceMeshMemberRoll** resource may be created in the project containing the control plane.  This resource
            contains a single *members* list of all the projects that should belong in the mesh.  The resource must be named
            *default*.  The user creating the resource must have *edit* or *admin* permissions for all projects in the
            *members* list.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 2
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - maistra
        - servicemesh
        links:
        - name: Red Hat OpenShift Service Mesh 2
          url: https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html
        - name: Istio
          url: https://istio.io/
        - name: Operator Source Code
          url: https://github.com/Maistra/istio-operator
        - name: Bugs
          url: https://issues.redhat.com/projects/OSSM
        maintainers:
        - email: istio-feedback@redhat.com
          name: Red Hat, OpenShift Service Mesh
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:4ed63a5440753b1b22d44191246c82e7671e3f545c1dbdae4cb4fad163ca3a46
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6b4623c416fc85bd8e38904c16494fa3c4aeae25c90dc0da9763a0a87141bc04
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:515c5d9e6700fae981b0ea3807436b250acb1534e9ecf50043e5df8474e66dce
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:b134578e0677963197e353d8acd0ac3330df1b6d2752aa41cb072e86b5b4b91b
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6fb7d50b188dc27ffb9a396edabbfb21767fe438344fd70d4754d1ec7cf27957
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:e734ca310e09f737b4f06f3ecfeb9ac12971c63884fdcc27e3bebef9d5b18722
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:a36d5db6097749c55e946bba3b8ddd13a84b1f6f68b62d0ccbad5a803e9b46fb
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:fc500a282c6668ff987a632efbbd19ecc3aebdc021f831a7b386e360a5c8324d
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:57eb7ba03f24781501a7ceed2e160ff95317869af6c2caf8f9a83990e6a6c7b9
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:1ca80927334c6af8e386de8d44704f568b4b35d3f7afbe785a999d0fa4efd0bf
        - registry.redhat.io/openshift4/ose-prometheus@sha256:423ae2e81cfe660b7217f7662754f366a91d457d6dd1589b4fc40733896eb9d6
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:d16a60204d266bf4cc266c005fabf175b533b827500f171013884a713c88857a
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel8@sha256:3d773d16fed72278949b9e3b54dc59e346e95341dc94b715bcfc328d3c9e83b2
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:6a420228e684596ce2b900614a874f3e1284a6f2bbbba4c5a71ade158867affe
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:c77183e50a34d33b5e2fede60350f54b18ab3536c3189e331569d5df862d9979
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:a3d65983d46dcf47b94f26e7265ed32082208ce08abcf3587a9a4e15c19b274a
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:b241452cce5c9af5aa0f12d1db3a7094cc004ce355988b8a5207091810a6f84b
        - registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:3dc1bddb070fd530ce6251666989b9a849de90591718a578da4a9900f98d93ec
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:c34b4cd9a82fef40708a25f0568f6dc632878affba35d5d631a48ec31da08cd3
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:c1cadf40bdb9511d612a0257c12eb3a9738823592ebb6e32a4f51b3a8dfe08db
        - registry.redhat.io/openshift-service-mesh/prometheus-rhel8@sha256:71e193a2d403bb06a7a76686b376985315aa0c32f495b8b9dacd73c7ccce2af3
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d95762c62087072cadee5948c706619f09202a349e1310f0ed81bfb42f11203f
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:4669724b430683bf2e895c5cf941c2d99536117815999d38131e5b6f162bbef0
        version: 2.6.10-0
      entries:
      - name: servicemeshoperator.v2.6.10
        version: 2.6.10-0
      - name: servicemeshoperator.v2.6.9
        version: 2.6.9-0
      - name: servicemeshoperator.v2.6.8
        version: 2.6.8-0
      - name: servicemeshoperator.v2.6.7
        version: 2.6.7-0
      - name: servicemeshoperator.v2.6.6
        version: 2.6.6-0
      - name: servicemeshoperator.v2.6.5
        version: 2.6.5-0
      - name: servicemeshoperator.v2.6.4
        version: 2.6.4-0
      - name: servicemeshoperator.v2.6.3
        version: 2.6.3-0
      - name: servicemeshoperator.v2.6.2
        version: 2.6.2-0
      - name: servicemeshoperator.v2.6.1
        version: 2.6.1-0
      - name: servicemeshoperator.v2.6.0
        version: 2.6.0-0
      - name: servicemeshoperator.v2.5.2
        version: 2.5.2-0
      - name: servicemeshoperator.v2.5.1
        version: 2.5.1-0
      - name: servicemeshoperator.v2.5.0
        version: 2.5.0-0
      - name: servicemeshoperator.v2.4.5
        version: 2.4.5-0
      - name: servicemeshoperator.v2.4.4
        version: 2.4.4-0
      - name: servicemeshoperator.v2.4.3
        version: 2.4.3-0
      - name: servicemeshoperator.v2.4.2
        version: 2.4.2-0
      - name: servicemeshoperator.v2.4.1
        version: 2.4.1-0
      - name: servicemeshoperator.v2.4.0
        version: 2.4.0-0
      - name: servicemeshoperator.v2.3.3
        version: 2.3.3-0
      - name: servicemeshoperator.v2.3.2
        version: 2.3.2-0
      - name: servicemeshoperator.v2.3.1
        version: 2.3.1-0
      - name: servicemeshoperator.v2.3.0
        version: 2.3.0-0
      - name: servicemeshoperator.v2.2.3
        version: 2.2.3-0
      - name: servicemeshoperator.v2.2.2
        version: 2.2.2-0
      - name: servicemeshoperator.v2.2.1
        version: 2.2.1-0
      - name: servicemeshoperator.v2.2.0
        version: 2.2.0-0
      - name: servicemeshoperator.v2.1.3
        version: 2.1.3-0
      - name: servicemeshoperator.v2.1.2
        version: 2.1.2-0
      - name: servicemeshoperator.v2.1.1
        version: 2.1.1-0
      - name: servicemeshoperator.v2.1.0
        version: 2.1.0-0
      - name: servicemeshoperator.v2.0.8
        version: 2.0.8-0
      name: "1.0"
    - currentCSV: servicemeshoperator.v2.6.10
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"maistra.io/v2\",\n    \"kind\":
            \"ServiceMeshControlPlane\",\n    \"metadata\": {\n      \"name\": \"basic\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"version\": \"v2.6\",\n      \"policy\": {\n         \"type\":
            \"Istiod\"\n      },\n      \"telemetry\": {\n         \"type\": \"Istiod\"\n
            \     },\n      \"addons\": {\n        \"prometheus\": {\n          \"enabled\":
            true\n        },\n        \"kiali\": {\n          \"enabled\": true\n
            \       },\n        \"grafana\": {\n          \"enabled\": true\n        }\n
            \     }\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n    \"kind\":
            \"ServiceMeshMemberRoll\",\n    \"metadata\": {\n      \"name\": \"default\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"members\": [\n        \"your-project\",\n        \"another-of-your-projects\"
            \n      ]\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n
            \   \"kind\": \"ServiceMeshMember\",\n    \"metadata\": {\n      \"name\":
            \"default\",\n      \"namespace\": \"application-namespace\"\n    },\n
            \   \"spec\": {\n      \"controlPlaneRef\": {\n        \"name\": \"basic\",\n
            \       \"namespace\": \"control-plane-namespace\"\n      }\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
          createdAt: 2025-08-11T09:32:44CEST
          description: The OpenShift Service Mesh 2 Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh 2.6
            and earlier releases. OpenShift Service Mesh 2 is based on the open source
            Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2.6.10-0'
          operators.openshift.io/infrastructure-features: '["Disconnected","fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/maistra/istio-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Istio control plane installation
            displayName: Istio Service Mesh Control Plane
            kind: ServiceMeshControlPlane
            name: servicemeshcontrolplanes.maistra.io
            version: v2
          - description: Marks the containing namespace as a member of the referenced
              Service Mesh
            displayName: Istio Service Mesh Member
            kind: ServiceMeshMember
            name: servicemeshmembers.maistra.io
            version: v1
          - description: A list of namespaces in Service Mesh
            displayName: Istio Service Mesh Member Roll
            kind: ServiceMeshMemberRoll
            name: servicemeshmemberrolls.maistra.io
            version: v1
        description: |-
          Red Hat OpenShift Service Mesh 2 is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh 2, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh 2
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh 2 provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh 2 supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### Joining Projects Into a Mesh

          Once an instance of Red Hat OpenShift Service Mesh 2 has been installed, it will only exercise control over services within its own
          project.  Other projects may be added into the mesh using one of two methods:

          1. A **ServiceMeshMember** resource may be created in other projects to add those projects into the mesh.  The
            **ServiceMeshMember** specifies a reference to the **ServiceMeshControlPlane** object that was used to install
            the control plane.  The user creating the **ServiceMeshMember** resource must have permission to *use* the
            **ServiceMeshControlPlane** object.  The adminstrator for the project containing the control plane can grant
            individual users or groups the *use* permissions.

          2. A **ServiceMeshMemberRoll** resource may be created in the project containing the control plane.  This resource
            contains a single *members* list of all the projects that should belong in the mesh.  The resource must be named
            *default*.  The user creating the resource must have *edit* or *admin* permissions for all projects in the
            *members* list.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 2
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - maistra
        - servicemesh
        links:
        - name: Red Hat OpenShift Service Mesh 2
          url: https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html
        - name: Istio
          url: https://istio.io/
        - name: Operator Source Code
          url: https://github.com/Maistra/istio-operator
        - name: Bugs
          url: https://issues.redhat.com/projects/OSSM
        maintainers:
        - email: istio-feedback@redhat.com
          name: Red Hat, OpenShift Service Mesh
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:c1cadf40bdb9511d612a0257c12eb3a9738823592ebb6e32a4f51b3a8dfe08db
        - registry.redhat.io/openshift4/ose-prometheus@sha256:423ae2e81cfe660b7217f7662754f366a91d457d6dd1589b4fc40733896eb9d6
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:4ed63a5440753b1b22d44191246c82e7671e3f545c1dbdae4cb4fad163ca3a46
        - registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:57eb7ba03f24781501a7ceed2e160ff95317869af6c2caf8f9a83990e6a6c7b9
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:3dc1bddb070fd530ce6251666989b9a849de90591718a578da4a9900f98d93ec
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel8@sha256:3d773d16fed72278949b9e3b54dc59e346e95341dc94b715bcfc328d3c9e83b2
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:a3d65983d46dcf47b94f26e7265ed32082208ce08abcf3587a9a4e15c19b274a
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:6a420228e684596ce2b900614a874f3e1284a6f2bbbba4c5a71ade158867affe
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:c77183e50a34d33b5e2fede60350f54b18ab3536c3189e331569d5df862d9979
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:d16a60204d266bf4cc266c005fabf175b533b827500f171013884a713c88857a
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:1ca80927334c6af8e386de8d44704f568b4b35d3f7afbe785a999d0fa4efd0bf
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6b4623c416fc85bd8e38904c16494fa3c4aeae25c90dc0da9763a0a87141bc04
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:515c5d9e6700fae981b0ea3807436b250acb1534e9ecf50043e5df8474e66dce
        - registry.redhat.io/openshift-service-mesh/prometheus-rhel8@sha256:71e193a2d403bb06a7a76686b376985315aa0c32f495b8b9dacd73c7ccce2af3
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d95762c62087072cadee5948c706619f09202a349e1310f0ed81bfb42f11203f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:e734ca310e09f737b4f06f3ecfeb9ac12971c63884fdcc27e3bebef9d5b18722
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:a36d5db6097749c55e946bba3b8ddd13a84b1f6f68b62d0ccbad5a803e9b46fb
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:b134578e0677963197e353d8acd0ac3330df1b6d2752aa41cb072e86b5b4b91b
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6fb7d50b188dc27ffb9a396edabbfb21767fe438344fd70d4754d1ec7cf27957
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:4669724b430683bf2e895c5cf941c2d99536117815999d38131e5b6f162bbef0
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:fc500a282c6668ff987a632efbbd19ecc3aebdc021f831a7b386e360a5c8324d
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:c34b4cd9a82fef40708a25f0568f6dc632878affba35d5d631a48ec31da08cd3
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:b241452cce5c9af5aa0f12d1db3a7094cc004ce355988b8a5207091810a6f84b
        version: 2.6.10-0
      entries:
      - name: servicemeshoperator.v2.6.10
        version: 2.6.10-0
      - name: servicemeshoperator.v2.6.9
        version: 2.6.9-0
      - name: servicemeshoperator.v2.6.8
        version: 2.6.8-0
      - name: servicemeshoperator.v2.6.7
        version: 2.6.7-0
      - name: servicemeshoperator.v2.6.6
        version: 2.6.6-0
      - name: servicemeshoperator.v2.6.5
        version: 2.6.5-0
      - name: servicemeshoperator.v2.6.4
        version: 2.6.4-0
      - name: servicemeshoperator.v2.6.3
        version: 2.6.3-0
      - name: servicemeshoperator.v2.6.2
        version: 2.6.2-0
      - name: servicemeshoperator.v2.6.1
        version: 2.6.1-0
      - name: servicemeshoperator.v2.6.0
        version: 2.6.0-0
      - name: servicemeshoperator.v2.5.2
        version: 2.5.2-0
      - name: servicemeshoperator.v2.5.1
        version: 2.5.1-0
      - name: servicemeshoperator.v2.5.0
        version: 2.5.0-0
      - name: servicemeshoperator.v2.4.5
        version: 2.4.5-0
      - name: servicemeshoperator.v2.4.4
        version: 2.4.4-0
      - name: servicemeshoperator.v2.4.3
        version: 2.4.3-0
      - name: servicemeshoperator.v2.4.2
        version: 2.4.2-0
      - name: servicemeshoperator.v2.4.1
        version: 2.4.1-0
      - name: servicemeshoperator.v2.4.0
        version: 2.4.0-0
      - name: servicemeshoperator.v2.3.3
        version: 2.3.3-0
      - name: servicemeshoperator.v2.3.2
        version: 2.3.2-0
      - name: servicemeshoperator.v2.3.1
        version: 2.3.1-0
      - name: servicemeshoperator.v2.3.0
        version: 2.3.0-0
      - name: servicemeshoperator.v2.2.3
        version: 2.2.3-0
      - name: servicemeshoperator.v2.2.2
        version: 2.2.2-0
      - name: servicemeshoperator.v2.2.1
        version: 2.2.1-0
      - name: servicemeshoperator.v2.2.0
        version: 2.2.0-0
      - name: servicemeshoperator.v2.1.3
        version: 2.1.3-0
      - name: servicemeshoperator.v2.1.2
        version: 2.1.2-0
      - name: servicemeshoperator.v2.1.1
        version: 2.1.1-0
      - name: servicemeshoperator.v2.1.0
        version: 2.1.0-0
      - name: servicemeshoperator.v2.0.8
        version: 2.0.8-0
      name: stable
    defaultChannel: stable
    packageName: servicemeshoperator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Labs
      provider-url: https://grafana.com
    name: grafana-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: grafana-operator.v5.19.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "Grafana",
                "metadata": {
                  "labels": {
                    "dashboards": "grafana-a",
                    "folders": "grafana-a"
                  },
                  "name": "grafana-a"
                },
                "spec": {
                  "config": {
                    "auth": {
                      "disable_login_form": "false"
                    },
                    "log": {
                      "mode": "console"
                    },
                    "security": {
                      "admin_password": "start",
                      "admin_user": "root"
                    }
                  }
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaAlertRuleGroup",
                "metadata": {
                  "name": "grafanaalertrulegroup-sample"
                },
                "spec": {
                  "folderRef": "test-folder-from-operator",
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "interval": "5m",
                  "rules": [
                    {
                      "condition": "B",
                      "data": [
                        {
                          "datasourceUid": "grafanacloud-demoinfra-prom",
                          "model": {
                            "datasource": {
                              "type": "prometheus",
                              "uid": "grafanacloud-demoinfra-prom"
                            },
                            "editorMode": "code",
                            "expr": "weather_temp_c{}",
                            "instant": true,
                            "intervalMs": 1000,
                            "legendFormat": "__auto",
                            "maxDataPoints": 43200,
                            "range": false,
                            "refId": "A"
                          },
                          "refId": "A",
                          "relativeTimeRange": {
                            "from": 600
                          }
                        },
                        {
                          "datasourceUid": "__expr__",
                          "model": {
                            "conditions": [
                              {
                                "evaluator": {
                                  "params": [
                                    0
                                  ],
                                  "type": "gt"
                                },
                                "operator": {
                                  "type": "and"
                                },
                                "query": {
                                  "params": [
                                    "C"
                                  ]
                                },
                                "reducer": {
                                  "params": [],
                                  "type": "last"
                                },
                                "type": "query"
                              }
                            ],
                            "datasource": {
                              "type": "__expr__",
                              "uid": "__expr__"
                            },
                            "expression": "A",
                            "intervalMs": 1000,
                            "maxDataPoints": 43200,
                            "refId": "B",
                            "type": "threshold"
                          },
                          "refId": "B",
                          "relativeTimeRange": {
                            "from": 600
                          }
                        }
                      ],
                      "execErrState": "Error",
                      "for": "5m0s",
                      "noDataState": "NoData",
                      "title": "Temperature below freezing",
                      "uid": "4843de5c-4f8a-4af0-9509-23526a04faf8"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaContactPoint",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "grafana-operator",
                    "app.kubernetes.io/instance": "grafanacontactpoint-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "grafanacontactpoint",
                    "app.kubernetes.io/part-of": "grafana-operator"
                  },
                  "name": "grafanacontactpoint-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "name": "grafanacontactpoint-sample",
                  "settings": {
                    "email": null
                  },
                  "type": "email"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaDashboard",
                "metadata": {
                  "name": "grafanadashboard-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "json": "{\n\n  \"id\": null,\n  \"title\": \"Simple Dashboard\",\n  \"tags\": [],\n  \"style\": \"dark\",\n  \"timezone\": \"browser\",\n  \"editable\": true,\n  \"hideControls\": false,\n  \"graphTooltip\": 1,\n  \"panels\": [],\n  \"time\": {\n    \"from\": \"now-6h\",\n    \"to\": \"now\"\n  },\n  \"timepicker\": {\n    \"time_options\": [],\n    \"refresh_intervals\": []\n  },\n  \"templating\": {\n    \"list\": []\n  },\n  \"annotations\": {\n    \"list\": []\n  },\n  \"refresh\": \"5s\",\n  \"schemaVersion\": 17,\n  \"version\": 0,\n  \"links\": []\n}\n"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaDatasource",
                "metadata": {
                  "name": "grafanadatasource-sample"
                },
                "spec": {
                  "datasource": {
                    "access": "proxy",
                    "isDefault": true,
                    "jsonData": {
                      "timeInterval": "5s",
                      "tlsSkipVerify": true
                    },
                    "name": "prometheus",
                    "type": "prometheus",
                    "url": "http://prometheus-service:9090"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "plugins": [
                    {
                      "name": "grafana-clock-panel",
                      "version": "1.3.0"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaFolder",
                "metadata": {
                  "name": "grafanafolder-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "title": "Example Folder"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaLibraryPanel",
                "metadata": {
                  "name": "grafana-library-panel-inline-envs"
                },
                "spec": {
                  "envs": [
                    {
                      "name": "CUSTOM_RANGE_ENV",
                      "value": "now - 12h"
                    }
                  ],
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "jsonnet": "local myRange = std.extVar('CUSTOM_RANGE_ENV'); {\n\n  model: {}\n}\n",
                  "plugins": [
                    {
                      "name": "grafana-piechart-panel",
                      "version": "1.3.9"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaMuteTiming",
                "metadata": {
                  "name": "mutetiming-sample"
                },
                "spec": {
                  "editable": false,
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "name": "mutetiming-sample",
                  "time_intervals": [
                    {
                      "days_of_month": [
                        "1",
                        "15"
                      ],
                      "location": "Asia/Shanghai",
                      "times": [
                        {
                          "end_time": "06:00",
                          "start_time": "00:00"
                        }
                      ],
                      "weekdays": [
                        "saturday"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationPolicy",
                "metadata": {
                  "name": "grafananotificationpolicy-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "route": {
                    "group_by": [
                      "grafana_folder",
                      "alertname"
                    ],
                    "receiver": "Grafana Cloud OnCall",
                    "routes": [
                      {
                        "object_matchers": [
                          [
                            "foo",
                            "=",
                            "bar"
                          ]
                        ],
                        "receiver": "grafana-default-email",
                        "routes": [
                          {
                            "object_matchers": [
                              [
                                "severity",
                                "=",
                                "critical"
                              ]
                            ],
                            "receiver": "Grafana Cloud OnCall"
                          }
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationPolicyRoute",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "grafana-operator",
                    "app.kubernetes.io/instance": "grafananotificationpolicyroute-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "grafananotificationpolicyroute",
                    "app.kubernetes.io/part-of": "grafana-operator"
                  },
                  "name": "grafananotificationpolicyroute-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationTemplate",
                "metadata": {
                  "name": "test"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "name": "test",
                  "template": "{{ define \"SlackAlert\" }}\n  [{{.Status}}] {{ .Labels.alertname }}\n  {{ .Annotations.AlertValues }}\n{{ end }}\n\n{{ define \"SlackAlertMessage\" }}\n  {{ if gt (len .Alerts.Firing) 0 }}\n    {{ len .Alerts.Firing }} firing:\n    {{ range .Alerts.Firing }} {{ template \"SlackAlert\" . }} {{ end }}\n  {{ end }}\n  {{ if gt (len .Alerts.Resolved) 0 }}\n    {{ len .Alerts.Resolved }} resolved:\n    {{ range .Alerts.Resolved }} {{ template \"SlackAlert\" . }} {{ end }}\n  {{ end }}\n{{ end }}\n\n{{ template \"SlackAlertMessage\" . }}\n"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: ghcr.io/grafana/grafana-operator@sha256:b678bc3faad11f4ef53c240e2a650ce939b9e8920dd72a47681b2efcadc7b837
          createdAt: "2025-08-21T08:38:00Z"
          description: Deploys and manages Grafana instances, dashboards and data
            sources
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/grafana/grafana-operator
          support: Grafana Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GrafanaAlertRuleGroup is the Schema for the grafanaalertrulegroups
              API
            displayName: Grafana Alert Rule Group
            kind: GrafanaAlertRuleGroup
            name: grafanaalertrulegroups.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaContactPoint is the Schema for the grafanacontactpoints
              API
            displayName: Grafana Contact Point
            kind: GrafanaContactPoint
            name: grafanacontactpoints.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaDashboard is the Schema for the grafanadashboards
              API
            displayName: Grafana Dashboard
            kind: GrafanaDashboard
            name: grafanadashboards.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaDatasource is the Schema for the grafanadatasources
              API
            displayName: Grafana Datasource
            kind: GrafanaDatasource
            name: grafanadatasources.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaFolder is the Schema for the grafanafolders API
            displayName: Grafana Folder
            kind: GrafanaFolder
            name: grafanafolders.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaLibraryPanel
            name: grafanalibrarypanels.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaMuteTiming
            name: grafanamutetimings.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaNotificationPolicy is the Schema for the GrafanaNotificationPolicy
              API
            displayName: Grafana Notification policy
            kind: GrafanaNotificationPolicy
            name: grafananotificationpolicies.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaNotificationPolicyRoute is the Schema for the grafananotificationpolicyroutes
              API
            displayName: Grafana Notification Policy Route
            kind: GrafanaNotificationPolicyRoute
            name: grafananotificationpolicyroutes.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaNotificationTemplate
            name: grafananotificationtemplates.grafana.integreatly.org
            version: v1beta1
          - description: Grafana is the Schema for the grafanas API
            displayName: Grafana
            kind: Grafana
            name: grafanas.grafana.integreatly.org
            version: v1beta1
        description: Deploys and manages Grafana instances, dashboards and data sources
        displayName: Grafana Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Grafana
        - Metrics
        - Observability
        links:
        - name: Grafana Operator
          url: https://grafana.github.io/grafana-operator
        maintainers:
        - email: grafana-operator@grafana.com
          name: Grafana-operator maintainers
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Grafana Labs
          url: https://grafana.com
        relatedImages:
        - ghcr.io/grafana/grafana-operator@sha256:b678bc3faad11f4ef53c240e2a650ce939b9e8920dd72a47681b2efcadc7b837
        - docker.io/grafana/grafana@sha256:6ac590e7cabc2fbe8d7b8fc1ce9c9f0582177b334e0df9c927ebd9670469440f
        version: 5.19.4
      entries:
      - name: grafana-operator.v5.19.4
        version: 5.19.4
      - name: grafana-operator.v5.18.0
        version: 5.18.0
      - name: grafana-operator.v5.17.1
        version: 5.17.1
      - name: grafana-operator.v5.17.0
        version: 5.17.0
      - name: grafana-operator.v5.16.0
        version: 5.16.0
      - name: grafana-operator.v5.15.1
        version: 5.15.1
      - name: grafana-operator.v5.14.0
        version: 5.14.0
      name: v5
    defaultChannel: v5
    packageName: grafana-operator
    provider:
      name: Grafana Labs
      url: https://grafana.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: netobserv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netobserv-operator.v1.6.3-community
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowMetric",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netobserv-operator",
                    "app.kubernetes.io/instance": "flowmetric-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "flowmetric",
                    "app.kubernetes.io/part-of": "netobserv-operator"
                  },
                  "name": "flowmetric-sample"
                },
                "spec": {
                  "charts": [
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "",
                          "promQL": "sum(rate($METRIC[2m]))"
                        }
                      ],
                      "title": "External ingress traffic",
                      "type": "SingleStat",
                      "unit": "Bps"
                    },
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "{{DstK8S_Namespace}} / {{DstK8S_OwnerName}}",
                          "promQL": "sum(rate($METRIC{DstK8S_Namespace!=\"\"}[2m])) by (DstK8S_Namespace, DstK8S_OwnerName)"
                        }
                      ],
                      "sectionName": "External",
                      "title": "Top external ingress traffic per workload",
                      "type": "StackArea",
                      "unit": "Bps"
                    }
                  ],
                  "direction": "Ingress",
                  "filters": [
                    {
                      "field": "SrcSubnetLabel",
                      "matchType": "Absence"
                    }
                  ],
                  "labels": [
                    "DstK8S_HostName",
                    "DstK8S_Namespace",
                    "DstK8S_OwnerName",
                    "DstK8S_OwnerType"
                  ],
                  "metricName": "cluster_external_ingress_bytes_total",
                  "type": "Counter",
                  "valueField": "Bytes"
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "app"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "infra"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "enable": true,
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "includeList": [
                        "node_ingress_bytes_total",
                        "workload_ingress_bytes_total",
                        "namespace_flows_total"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta2",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "metrics": {
                        "server": {
                          "port": 9400
                        }
                      },
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "eBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "Disabled"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "\"app\""
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "\"infra\""
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "\"Pod\"",
                          "src_kind": "\"Pod\""
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "\"Service\""
                        },
                        "name": "Services network"
                      }
                    ],
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    }
                  },
                  "deploymentModel": "Direct",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "enable": true,
                    "lokiStack": {
                      "name": "loki"
                    },
                    "mode": "Monolithic",
                    "monolithic": {
                      "tenantID": "netobserv",
                      "tls": {
                        "caCert": {
                          "certFile": "service-ca.crt",
                          "name": "loki-gateway-ca-bundle",
                          "type": "configmap"
                        },
                        "enable": false
                      },
                      "url": "http://loki.netobserv.svc:3100/"
                    },
                    "readTimeout": "30s",
                    "writeBatchSize": 10485760,
                    "writeBatchWait": "1s",
                    "writeTimeout": "10s"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "Flows",
                    "metrics": {
                      "disableAlerts": [],
                      "server": {
                        "port": 9401
                      }
                    },
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "prometheus": {
                    "querier": {
                      "enable": true,
                      "mode": "Auto",
                      "timeout": "30s"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:1.6.2-community
          createdAt: "2024-08-22T18:26:56Z"
          description: Network flows collector and monitoring solution
          operatorframework.io/initialization-resource: '{"apiVersion":"flows.netobserv.io/v1beta2",
            "kind":"FlowCollector","metadata":{"name":"cluster"},"spec": {}}'
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
          support: NetObserv team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta2
          - description: '`FlowMetric` is the schema for the custom metrics API, which
              allows to generate more metrics out of flow logs. You can find examples
              here: https://github.com/netobserv/network-observability-operator/tree/1.6.2-community/config/samples/flowmetrics'
            displayName: Flow Metric
            kind: FlowMetric
            name: flowmetrics.flows.netobserv.io
            version: v1alpha1
        description: |-
          DEPRECATED: this distribution channel is deprecated and doesn't receive new updates. Please read https://github.com/orgs/netobserv/discussions/996 to learn more.

          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline that consists in:
          - an eBPF agent, that generates network flows from captured packets
          - flowlogs-pipeline, a component that collects, enriches and exports these flows
          - when used in OpenShift, a Console plugin for flows visualization with powerful filtering options, a topology representation and more

          Flow data is then available in multiple ways, each optional:

          - As Prometheus metrics
          - As raw flow logs stored in Grafana Loki
          - As raw flow logs exported to a collector

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, can optionally be used as the backend to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 10GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          If you prefer to not use Loki, you must set `spec.loki.enable` to `false` in `FlowCollector`.
          In that case, you can still get the Prometheus metrics or export raw flows to a custom collector. But be aware that some of the Console plugin features will be disabled. For instance, you will not be able to view raw flows there, and the metrics / topology will have a more limited level of details, missing information such as pods or IPs.

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for example [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/config/samples/flows_v1beta2_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide on every node, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you might have to configure differently if you used another installation method. Make sure to disable it (`spec.loki.enable`) if you don't want to use Loki.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`) an optional list of exporters to which to send enriched flows. KAFKA and IPFIX exporters are supported. This allows you to define any custom storage or processing that can read from Kafka or use the IPFIX standard.

          - To enable availability zones awareness, set `spec.processor.addZone` to `true`.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - More information on [configuring metrics](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/Metrics.md).
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: '[DEPRECATED] NetObserv Operator'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/network-observability-operator:1.6.2-community
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/netobserv/netobserv-ebpf-agent:v1.6.2-community
        - quay.io/netobserv/flowlogs-pipeline:v1.6.2-community
        - quay.io/netobserv/network-observability-console-plugin:v1.6.2-community
        version: 1.6.3-community
      entries:
      - name: netobserv-operator.v1.6.3-community
        version: 1.6.3-community
      - name: netobserv-operator.v1.6.2-community
        version: 1.6.2-community
      - name: netobserv-operator.v1.6.1-community
        version: 1.6.1-community
      - name: netobserv-operator.v1.6.0-community
        version: 1.6.0-community
      - name: netobserv-operator.v1.5.0-community
        version: 1.5.0-community
      name: community
    - currentCSV: netobserv-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "ipfix": {
                      "cacheActiveTimeout": "20s",
                      "cacheMaxFlows": 400,
                      "clusterNetworkOperator": {
                        "namespace": "openshift-network-operator"
                      },
                      "ovnKubernetes": {
                        "containerName": "ovnkube-node",
                        "daemonSetName": "ovnkube-node",
                        "namespace": "ovn-kubernetes"
                      },
                      "sampling": 400
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "staticLabels": {
                      "app": "netobserv-flowcollector"
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "enableKubeProbes": true,
                    "healthPort": 8080,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:0.2.1
          createdAt: "2022-12-07T09:00:48Z"
          description: Network flows collector and monitoring solution
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
        description: |-
          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline to collect and enrich network flows. These flows can be produced by the NetObserv eBPF agent, or by any device or CNI able to export flows in IPFIX format, such as OVN-Kubernetes.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When used in OpenShift, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For a normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionnaly be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/0.2.1/config/samples/flows_v1alpha1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Agent (`spec.agent.type`) can be `EBPF` (default) or `IPFIX`. eBPF is recommended, as it should work in more situations and offers better performances. If you can't, or don't want to use eBPF, note that the IPFIX option is fully functional only when using [OVN-Kubernetes](https://github.com/ovn-org/ovn-kubernetes/) CNI. Other CNIs are not officially supported, but you may still be able to configure them manually if they allow IPFIX exports.

          - Sampling (`spec.agent.ebpf.sampling` and `spec.agent.ipfix.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50) for eBPF and 400 (1:400) for IPFIX. Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub more more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - A [F.A.Q](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: NetObserv Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: mmaciasl@redhat.com
          name: Mario Macias
        - email: ocazade@redhat.com
          name: Olivier Cazade
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/netobserv-ebpf-agent:v0.2.2
        - quay.io/netobserv/flowlogs-pipeline:v0.1.5
        - quay.io/netobserv/network-observability-console-plugin:v0.1.6
        - quay.io/netobserv/network-observability-operator:0.2.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        version: 0.2.1
      entries:
      - name: netobserv-operator.v0.2.1
        version: 0.2.1
      - name: netobserv-operator.v0.2.0
        version: 0.2.0
      - name: netobserv-operator.v0.1.4
        version: 0.1.4
      - name: netobserv-operator.v0.1.3
        version: 0.1.3
      - name: netobserv-operator.v0.1.2
        version: 0.1.2
      - name: netobserv-operator.v0.1.1
        version: 0.1.1
      - name: netobserv-operator.v0.1.0
        version: 0.1.0
      name: latest
    - currentCSV: netobserv-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "ipfix": {
                      "cacheActiveTimeout": "20s",
                      "cacheMaxFlows": 400,
                      "clusterNetworkOperator": {
                        "namespace": "openshift-network-operator"
                      },
                      "ovnKubernetes": {
                        "containerName": "ovnkube-node",
                        "daemonSetName": "ovnkube-node",
                        "namespace": "ovn-kubernetes"
                      },
                      "sampling": 400
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "staticLabels": {
                      "app": "netobserv-flowcollector"
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "enableKubeProbes": true,
                    "healthPort": 8080,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:0.2.1
          createdAt: "2022-12-07T09:00:48Z"
          description: Network flows collector and monitoring solution
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
        description: |-
          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline to collect and enrich network flows. These flows can be produced by the NetObserv eBPF agent, or by any device or CNI able to export flows in IPFIX format, such as OVN-Kubernetes.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When used in OpenShift, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For a normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionnaly be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/0.2.1/config/samples/flows_v1alpha1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Agent (`spec.agent.type`) can be `EBPF` (default) or `IPFIX`. eBPF is recommended, as it should work in more situations and offers better performances. If you can't, or don't want to use eBPF, note that the IPFIX option is fully functional only when using [OVN-Kubernetes](https://github.com/ovn-org/ovn-kubernetes/) CNI. Other CNIs are not officially supported, but you may still be able to configure them manually if they allow IPFIX exports.

          - Sampling (`spec.agent.ebpf.sampling` and `spec.agent.ipfix.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50) for eBPF and 400 (1:400) for IPFIX. Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub more more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - A [F.A.Q](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: NetObserv Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: mmaciasl@redhat.com
          name: Mario Macias
        - email: ocazade@redhat.com
          name: Olivier Cazade
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/network-observability-operator:0.2.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/netobserv/network-observability-console-plugin:v0.1.6
        - quay.io/netobserv/netobserv-ebpf-agent:v0.2.2
        - quay.io/netobserv/flowlogs-pipeline:v0.1.5
        version: 0.2.1
      entries:
      - name: netobserv-operator.v0.2.1
        version: 0.2.1
      - name: netobserv-operator.v0.2.0
        version: 0.2.0
      - name: netobserv-operator.v0.1.4
        version: 0.1.4
      - name: netobserv-operator.v0.1.3
        version: 0.1.3
      - name: netobserv-operator.v0.1.2
        version: 0.1.2
      - name: netobserv-operator.v0.1.1
        version: 0.1.1
      - name: netobserv-operator.v0.1.0
        version: 0.1.0
      name: v0.2.x
    defaultChannel: community
    packageName: netobserv-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-license-service-scanner-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-license-service-scanner-operator.v4.2.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1",
                "kind": "IBMLicenseServiceScanner",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-licensing-scanner-operator",
                    "app.kubernetes.io/instance": "ibm-licensing-scanner-instance",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ibm-licensing-scanner-instance",
                    "app.kubernetes.io/part-of": "ibm-licensing-scanner-operator"
                  },
                  "name": "ibm-licensing-scanner-instance"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "license-service-upload-secret": "ibm-licensing-scanner-upload-secret",
                  "log-level": "INFO",
                  "registry-pull-secret": "ibm-licensing-scanner-registry-pull-secret",
                  "scan": {
                    "namespaces": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:77c9db26a13dba25438a6fc5a41290606418481e08ea2ecc0a9a397062af74fb
          createdAt: "2025-05-14T12:59:00Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <4.2.16'
          operatorframework.io/suggested-namespace: ibm-licensing-scanner
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMLicenseServiceScanner is the Schema for the ibmlicenseservicescanners
              API
            displayName: IBMLicense Service Scanner
            kind: IBMLicenseServiceScanner
            name: ibmlicenseservicescanners.operator.ibm.com
            version: v1
        description: IBM License Service Scanner
        displayName: IBM License Service Scanner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - licensing
        - usage
        - IBMLicenseServiceScanner
        - IBM
        - Cloud
        - License
        - License Service
        - Service
        - Scanner
        - License Service Scanner
        - License Scanner
        - Service Scanner
        links:
        - name: Ibm License Service Scanner Operator
          url: https://ibm-license-service-scanner-operator.domain
        maintainers:
        - email: talk2sam@us.ibm.com
          name: IBM
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-bundle@sha256:b6cb9e5976b749dd7818ac19d1cfd5e24673f3c052a41d6e2fe1d2d7094694b4
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-catalog@sha256:c016904e8c5cfcf8270241ad5e28b52f528859984ed894cdd5e1283fa90cd6e8
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:77c9db26a13dba25438a6fc5a41290606418481e08ea2ecc0a9a397062af74fb
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:f3503f72a90ff8abbadaef3e78a16db6e1fbff4c90b475c33ead17bf1e69990c
        version: 4.2.16
      entries:
      - name: ibm-license-service-scanner-operator.v4.2.16
        version: 4.2.16
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-license-service-scanner-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/component: operator
      app.kubernetes.io/managed-by: kustomize
      app.kubernetes.io/name: zabbix-operator-certified
      app.kubernetes.io/part-of: zabbix-operator-certified
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zabbix LLC
      provider-url: ""
    name: zabbix-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zabbix-operator-certified.v6.0.40
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent",
                "metadata": {
                  "labels": {
                    "app": "agent",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "log_remote_commands": true,
                  "max_lines_per_second": 20,
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent2",
                "metadata": {
                  "labels": {
                    "app": "agent2",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent2"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "enable_persistent_buffer": false,
                  "enable_status_port": false,
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "persistent_buffer_claim": "",
                  "persistent_buffer_period": "1h",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixFull",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-full"
                },
                "spec": {
                  "builtin_report_service": true,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysql_volumeclaim": "mysql-full-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxyMysql",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-mysql"
                },
                "spec": {
                  "db_server_port": 3306,
                  "enable_snmp_traps": false,
                  "internal_db": true,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix_proxy",
                  "proxy": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "zabbix_mysql_volumeclaim": "mysql-proxy-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxySqlite",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-sqlite"
                },
                "spec": {
                  "db_data_volumeclaim": "zabbix-proxy-sqlite",
                  "enable_snmp_traps": false,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "proxy": {
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "node_name_as_db_name": true,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "proxy_size": 1,
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  }
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixServer",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-server"
                },
                "spec": {
                  "builtin_report_service": true,
                  "db_server_host": "mysql-server",
                  "db_server_port": 3306,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix",
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              }
            ]
          ansible.operator-sdk/reconcile-period: 5s
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          certified: "true"
          containerImage: registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
          createdAt: "2025-05-26T07:41:22Z"
          description: Zabbix operator with multiple deployment variants and different
            components
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <6.0.40
          operatorframework.io/suggested-namespace: zabbix
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: ""
          support: Zabbix
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Zabbix agent 2 is a new generation of Zabbix agent and may
              be used in place of Zabbix agent
            displayName: Zabbix agent 2
            kind: ZabbixAgent2
            name: zabbixagent2s.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix agent is deployed on a monitoring nodes to actively
              monitor local resources and applications
            displayName: Zabbix agent
            kind: ZabbixAgent
            name: zabbixagents.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database server, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Full
            kind: ZabbixFull
            name: zabbixfulls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with MySQL database server
            displayName: Zabbix proxy (MySQL)
            kind: ZabbixProxyMysql
            name: zabbixproxymysqls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with SQLite database
            displayName: Zabbix proxy (SQLite3)
            kind: ZabbixProxySqlite
            name: zabbixproxysqlites.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database support, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Server
            kind: ZabbixServer
            name: zabbixservers.kubernetes.zabbix.com
            version: v1alpha1
        description: "## About this Operator\n\nZabbix helps you to real-time monitoring
          of millions of metrics collected from tens of thousands of servers, virtual
          machines and network devices.\nThe Zabbix Operator allows users to easily
          deploy, manage, and maintain Zabbix deployments on OpenShift. By installing
          this integration you will be able to deploy Zabbix server / proxies and
          other components with a single command.\n## Supported Features\n* **Zabbix
          Server** - Simple Zabbix installation with included Zabbix server, Zabbix
          web-interface and Zabbix Java Gateway with MySQL database support. The feature
          does not provide MySQL service and requires an external MySQL database.
          \n* **Zabbix Server (Full)** - Zabbix installation with included Zabbix
          server, Zabbix web-interface, Zabbix Java Gateway and MySQL server instance.\n*
          **Zabbix proxy (SQLite3)** - Very simple way to gain power of Zabbix proxy.
          The feature has  SQLite3 support for Zabbix proxies and allow to specify
          amount of proxies. \n* **Zabbix proxy (MySQL)** - Another option of Zabbix
          proxy. The option support and deliver MySQL database.\n* **Zabbix agent**
          - Zabbix agent can be deployed on each available node for stability and
          performance monitoring on remote nodes. It allows to gather metrics with
          full automation!* \n## Prerequisites\nAll deployment options are require
          additional information during deployment. Please, check the following instructions
          and provide required configuration:\n* **Zabbix Server** - MySQL database
          host information and MySQL database credentials in specially formatted *Secret*.
          Additionally it is possible to specify SSL certificates for HTTPS support
          in *Secret*.\n* **Zabbix Server (Full)** - MySQL database credentials in
          specially formatted *Secret*. MySQL database volume name information. Additionally
          it is possible to specify SSL certificates for HTTPS support in *Secret*.\n*
          **Zabbix proxy (SQLite3)** - Zabbix server host information only.\n* **Zabbix
          proxy (MySQL)** - MySQL database credentials in specially formatted *Secrets*
          and Zabbix server host information.\n* **Zabbix agent** - Zabbix server
          host information only for outgoing and incoming connections. Zabbix agent
          uses \"privileged\" mode to monitor node resources! For example, running
          processes.* \n### MySQL credentials\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-full-secrets\ndata:\n  mysql_root_password: emFiYml4X3N1cGVyX3Jvb3Q=
          [1]\n  mysql_zabbix_password: emFiYml4X3VzZXJfcGFzc3dvcmQ= [2]\n  mysql_zabbix_username:
          emFiYml4 [3]\ntype: Opaque\n```\nUsing MySQL root password (*mysqlrootpassword*
          [1]) Zabbix server / proxy will try to create MySQL database schema with
          grant permissions to *mysqlzabbixusername* [2] and *mysqlzabbixpassword*
          [3].\n### SSL certificates for HTTPS\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-web-sslsecret\ndata:\n  ssl.crt: >-\n   < ssl.crt data>\n
          \ ssl.key: >-\n\t < ssl.key data >\n  dhparam.pem: >-\n   <  dhparam.pem
          data >\n```\nFiles *ssl.crt*, *ssl.key* and *dhparam.perm* are required
          for Zabbix web-interface for SSL support.\n"
        displayName: Zabbix Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - zabbix
        - appliance
        - agent
        - metrics
        links:
        - name: Zabbix
          url: https://www.zabbix.com
        - name: Zabbix Official Documentation
          url: https://www.zabbix.com/documentation/6.0/en/manual/quickstart
        - name: Zabbix Operator Installation
          url: https://www.zabbix.com/documentation/6.0/en/manual/installation/containers/openshift
        - name: Downloads
          url: https://www.zabbix.com/download
        maintainers:
        - email: alexey.pustovalov@zabbix.com
          name: Alexey Pustovalov
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Zabbix LLC
        relatedImages:
        - registry.connect.redhat.com/zabbix/zabbix-agent2-60@sha256:424003d52f3cb4360adf2b3619d8b9f8de9196c630ee7e4ceb5d177c1c5a36e4
        - registry.connect.redhat.com/zabbix/zabbix-snmptraps-60@sha256:e61e6b1c551dec8488cce0998f94d83d6d10b5d89a465d7119b088313032cb81
        - registry.connect.redhat.com/zabbix/zabbix-server-mysql-60@sha256:637484faec6e9a55049c13dbd1f100c03d757c3a21749388c0605040f02d05ce
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
        - quay.io/zabbix/zabbix-operator@sha256:43d26488a19967bdb2a8aafde5f43820be95011811da9cc5ec623c9029801675
        - registry.connect.redhat.com/zabbix/zabbix-web-mysql-60@sha256:611d2670e1bf95b81226976c0a1f64434b2a7766f196dadbd646a7d8319f4cb1
        - registry.connect.redhat.com/zabbix/zabbix-java-gateway-60@sha256:a4feb725c12c4ac90eb9244c6fafdb6708679f2e0e7f6aa683c765127306412e
        - registry.connect.redhat.com/zabbix/zabbix-proxy-mysql-60@sha256:c2ae74d97c083ff348dc9c55932a33da4779cd9efdd9d7c5a79acfffae5b9ed0
        - registry.connect.redhat.com/zabbix/zabbix-proxy-sqlite-60@sha256:62e62ebc8c1bc13fb610f7a83412e65d852f74b3c2c212baa75ea27fc7fa03eb
        - registry.connect.redhat.com/zabbix/zabbix-web-service-60@sha256:4a333094e110713470a897b6406c5e340e962caf32a429dddf19865041fcc448
        - registry.redhat.io/rhel9/mysql-80@sha256:c4b5246d0c0d089a766ea07768c22d9d6be8049ac1379bd345db99390ffe2a36
        - registry.connect.redhat.com/zabbix/zabbix-agent-60@sha256:2895b5e45785aaf528d4a6f78c8af0a0d494dbe9362403fdd470673d911d5004
        version: 6.0.40
      entries:
      - name: zabbix-operator-certified.v6.0.40
        version: 6.0.40
      - name: zabbix-operator-certified.v6.0.38
        version: 6.0.38
      - name: zabbix-operator-certified.v6.0.36
        version: 6.0.36
      - name: zabbix-operator-certified.v6.0.33
        version: 6.0.33
      - name: zabbix-operator-certified.v6.0.29
        version: 6.0.29
      - name: zabbix-operator-certified.v6.0.27
        version: 6.0.27
      - name: zabbix-operator-certified.v6.0.26
        version: 6.0.26
      - name: zabbix-operator-certified.v6.0.17
        version: 6.0.17
      - name: zabbix-operator-certified.v6.0.14
        version: 6.0.14
      - name: zabbix-operator-certified.v6.0.13
        version: 6.0.13
      name: lts
    defaultChannel: lts
    packageName: zabbix-operator-certified
    provider:
      name: Zabbix LLC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/multiarch-tuning-operator
    name: multiarch-tuning-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multiarch-tuning-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multiarch.openshift.io/v1beta1",
                "kind": "ClusterPodPlacementConfig",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logVerbosity": "Normal",
                  "namespaceSelector": {
                    "matchExpressions": [
                      {
                        "key": "multiarch.openshift.io/exclude-pod-placement",
                        "operator": "DoesNotExist"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Other
          console.openshift.io/disable-operand-delete: "false"
          containerImage: registry.redhat.io/multiarch-tuning/multiarch-tuning-rhel9-operator@sha256:1a1d3f5258d4b91f1a0380b0fe2c809823e323d9ff74ae4171468168d255e7e0
          createdAt: "2025-03-05T12:27:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operator.openshift.io/uninstall-message: You must remove all Operands before
            uninstalling the operator.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-multiarch-tuning-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/multiarch-tuning-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPodPlacementConfig defines the configuration for the
              architecture aware pod placement operand. Users can only deploy a single
              object named "cluster". Creating the object enables the operand.
            displayName: Cluster Pod Placement Config
            kind: ClusterPodPlacementConfig
            name: clusterpodplacementconfigs.multiarch.openshift.io
            version: v1alpha1
          - description: ClusterPodPlacementConfig defines the configuration for the
              architecture aware pod placement operand. Users can only deploy a single
              object named "cluster". Creating the object enables the operand.
            displayName: Cluster Pod Placement Config
            kind: ClusterPodPlacementConfig
            name: clusterpodplacementconfigs.multiarch.openshift.io
            version: v1beta1
        description: |
          The Multiarch Tuning Operator optimizes workload management within multi-architecture clusters and in
          single-architecture clusters transitioning to multi-architecture environments.

          By default, the scheduler does not consider the architecture of a pod's container images
          when determining the placement of new pods onto nodes.

          The Multiarch Tuning Operator introduces the ClusterPodPlacementConfig custom resource to
          enable architecture-aware scheduling, ensuring that pods are assigned to nodes with
          a corresponding architecture for their container images.

          To enable this functionality, you must create a ClusterPodPlacementConfig resource.

          When the ClusterPodPlacementConfig object is created, the Multiarch Tuning Operator deploys the necessary operands
          to support architecture-aware workload scheduling.
        displayName: Multiarch Tuning Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multiarch-tuning-operator
        - pod-placement
        links:
        - name: Multiarch Tuning Operator Documentation
          url: https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/postinstallation_configuration/configuring-multi-architecture-compute-machines-on-an-openshift-cluster#multiarch-tuning-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/multiarch-tuning-operator
        relatedImages:
        - registry.redhat.io/multiarch-tuning/multiarch-tuning-rhel9-operator@sha256:1a1d3f5258d4b91f1a0380b0fe2c809823e323d9ff74ae4171468168d255e7e0
        version: 1.1.0
      entries:
      - name: multiarch-tuning-operator.v1.1.0
        version: 1.1.0
      - name: multiarch-tuning-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: multiarch-tuning-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/multiarch-tuning-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: infra-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infra-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instanceha.openstack.org/v1beta1",
                "kind": "InstanceHa",
                "metadata": {
                  "name": "instanceha"
                },
                "spec": {
                  "caBundleSecretName": "combined-ca-bundle"
                }
              },
              {
                "apiVersion": "memcached.openstack.org/v1beta1",
                "kind": "Memcached",
                "metadata": {
                  "name": "memcached"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "network.openstack.org/v1beta1",
                "kind": "DNSData",
                "metadata": {
                  "name": "allhosts"
                },
                "spec": {
                  "hosts": [
                    {
                      "hostnames": [
                        "edpm-compute-0.ctlplane"
                      ],
                      "ip": "192.168.122.100"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-0.internalapi"
                      ],
                      "ip": "172.17.0.100"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-0.storage"
                      ],
                      "ip": "172.18.0.100"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-0.tenant"
                      ],
                      "ip": "172.19.0.100"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-1.ctlplane"
                      ],
                      "ip": "192.168.122.101"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-1.internalapi"
                      ],
                      "ip": "172.17.0.101"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-1.storage"
                      ],
                      "ip": "172.18.0.101"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-1.tenant"
                      ],
                      "ip": "172.19.0.101"
                    },
                    {
                      "hostnames": [
                        "keystone-internal.openstack.svc",
                        "placement-internal.openstack.svc",
                        "cinder-internal.openstack.svc",
                        "glance-internal.openstack.svc",
                        "nova-internal.openstack.svc",
                        "nova-metadata-internal.openstack.svc",
                        "neutron-internal.openstack.svc"
                      ],
                      "ip": "172.17.0.80"
                    },
                    {
                      "hostnames": [
                        "rabbitmq.openstack.svc"
                      ],
                      "ip": "172.17.0.85"
                    },
                    {
                      "hostnames": [
                        "rabbitmq-cell1.openstack.svc"
                      ],
                      "ip": "172.17.0.86"
                    }
                  ]
                }
              },
              {
                "apiVersion": "network.openstack.org/v1beta1",
                "kind": "DNSMasq",
                "metadata": {
                  "name": "dnsmasq"
                },
                "spec": {
                  "debug": {
                    "service": false
                  },
                  "options": [
                    {
                      "key": "server",
                      "values": [
                        "192.168.122.1"
                      ]
                    },
                    {
                      "key": "local",
                      "values": [
                        "/example.com/"
                      ]
                    }
                  ],
                  "override": {
                    "service": {
                      "metadata": {
                        "annotations": {
                          "metallb.universe.tf/address-pool": "ctlplane",
                          "metallb.universe.tf/allow-shared-ip": "ctlplane",
                          "metallb.universe.tf/loadBalancerIPs": "192.168.122.80"
                        }
                      },
                      "spec": {
                        "type": "LoadBalancer"
                      }
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "network.openstack.org/v1beta1",
                "kind": "IPSet",
                "metadata": {
                  "name": "edpm-compute-0"
                },
                "spec": {
                  "networks": [
                    {
                      "defaultRoute": true,
                      "name": "CtlPlane",
                      "subnetName": "subnet1"
                    },
                    {
                      "name": "InternalApi",
                      "subnetName": "subnet1"
                    },
                    {
                      "name": "Storage",
                      "subnetName": "subnet1"
                    },
                    {
                      "name": "Tenant",
                      "subnetName": "subnet1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "network.openstack.org/v1beta1",
                "kind": "NetConfig",
                "metadata": {
                  "name": "netconfig"
                },
                "spec": {
                  "networks": [
                    {
                      "dnsDomain": "ctlplane.example.com",
                      "name": "ctlplane",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "192.168.122.120",
                              "start": "192.168.122.100"
                            },
                            {
                              "end": "192.168.122.200",
                              "start": "192.168.122.150"
                            }
                          ],
                          "cidr": "192.168.122.0/24",
                          "gateway": "192.168.122.1",
                          "name": "subnet1"
                        }
                      ]
                    },
                    {
                      "dnsDomain": "internalapi.example.com",
                      "name": "internalapi",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "172.17.0.250",
                              "start": "172.17.0.100"
                            }
                          ],
                          "cidr": "172.17.0.0/24",
                          "name": "subnet1",
                          "vlan": 20
                        }
                      ]
                    },
                    {
                      "dnsDomain": "external.example.com",
                      "name": "external",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "10.0.0.250",
                              "start": "10.0.0.100"
                            }
                          ],
                          "cidr": "10.0.0.0/24",
                          "gateway": "10.0.0.1",
                          "name": "subnet1"
                        }
                      ]
                    },
                    {
                      "dnsDomain": "storage.example.com",
                      "name": "storage",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "172.18.0.250",
                              "start": "172.18.0.100"
                            }
                          ],
                          "cidr": "172.18.0.0/24",
                          "name": "subnet1",
                          "vlan": 21
                        }
                      ]
                    },
                    {
                      "dnsDomain": "storagemgmt.example.com",
                      "name": "storagemgmt",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "172.20.0.250",
                              "start": "172.20.0.100"
                            }
                          ],
                          "cidr": "172.20.0.0/24",
                          "name": "subnet1",
                          "vlan": 23
                        }
                      ]
                    },
                    {
                      "dnsDomain": "tenant.example.com",
                      "name": "tenant",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "172.19.0.250",
                              "start": "172.19.0.100"
                            }
                          ],
                          "cidr": "172.19.0.0/24",
                          "name": "subnet1",
                          "vlan": 22
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "network.openstack.org/v1beta1",
                "kind": "Reservation",
                "metadata": {
                  "name": "edpm-compute-0"
                },
                "spec": {
                  "ipSetRef": {
                    "name": "edpm-compute-0",
                    "namespace": "openstack",
                    "uid": "56753ffb-e486-43ab-a79a-f85f27b2ac14"
                  },
                  "reservation": {
                    "CtlPlane": {
                      "address": "192.168.122.100",
                      "network": "CtlPlane",
                      "subnet": "subnet1"
                    },
                    "InternalApi": {
                      "address": "172.17.0.100",
                      "network": "InternalApi",
                      "subnet": "subnet1"
                    },
                    "Storage": {
                      "address": "172.18.0.100",
                      "network": "Storage",
                      "subnet": "subnet1"
                    },
                    "StorageMgmt": {
                      "address": "172.20.0.100",
                      "network": "StorageMgmt",
                      "subnet": "subnet1"
                    },
                    "Tenant": {
                      "address": "172.19.0.100",
                      "network": "Tenant",
                      "subnet": "subnet1"
                    }
                  }
                }
              },
              {
                "apiVersion": "rabbitmq.openstack.org/v1beta1",
                "kind": "TransportURL",
                "metadata": {
                  "name": "transporturl"
                },
                "spec": {
                  "rabbitmqClusterName": "rabbitmq"
                }
              },
              {
                "apiVersion": "redis.openstack.org/v1beta1",
                "kind": "Redis",
                "metadata": {
                  "name": "redis"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:45Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSData is the Schema for the dnsdata API
            displayName: DNSData
            kind: DNSData
            name: dnsdata.network.openstack.org
            version: v1beta1
          - description: DNSMasq is the Schema for the dnsmasqs API
            displayName: DNSMasq
            kind: DNSMasq
            name: dnsmasqs.network.openstack.org
            version: v1beta1
          - description: InstanceHa is the Schema for the instancehas API
            displayName: Instance Ha
            kind: InstanceHa
            name: instancehas.instanceha.openstack.org
            version: v1beta1
          - description: IPSet is the Schema for the ipsets API
            displayName: IPSet
            kind: IPSet
            name: ipsets.network.openstack.org
            version: v1beta1
          - description: Memcached is the Schema for the memcacheds API
            displayName: Memcached
            kind: Memcached
            name: memcacheds.memcached.openstack.org
            version: v1beta1
          - description: NetConfig is the Schema for the netconfigs API
            displayName: Net Config
            kind: NetConfig
            name: netconfigs.network.openstack.org
            version: v1beta1
          - description: Redis is the Schema for the redises API
            displayName: Redis
            kind: Redis
            name: redises.redis.openstack.org
            version: v1beta1
          - description: Reservation is the Schema for the reservations API
            displayName: Reservation
            kind: Reservation
            name: reservations.network.openstack.org
            version: v1beta1
          - description: TransportURL is the Schema for the transporturls API
            displayName: Transport URL
            kind: TransportURL
            name: transporturls.rabbitmq.openstack.org
            version: v1beta1
        description: OpenStack Infrastructure
        displayName: OpenStack Infra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Infrastructure
        links:
        - name: Infra Operator
          url: https://github.com/openstack-k8s-operators/infra-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/infra-rhel9-operator@sha256:177077029ad8baa959e41a785cd29ed63745e8f39be618aa97509e6867fbff7c
        - registry.redhat.io/rhoso/openstack-memcached-rhel9@sha256:962a314226009a8ca5c95fb66a75b15176506e02110847d0f42aeed6dd26a751
        - registry.redhat.io/rhoso/openstack-redis-rhel9@sha256:b83d23df9f2b367e604f404d97574fd6db9e0ed530eda23e9e0bbe01292c8db8
        - registry.redhat.io/rhoso/openstack-neutron-server-rhel9@sha256:806e5e009a77a29c82699487e539669e1ec2136514a0332d222b5cb3fed6f452
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: infra-operator.v1.0.6
        version: 1.0.6
      - name: infra-operator.v1.0.5
        version: 1.0.5
      - name: infra-operator.v1.0.4
        version: 1.0.4
      - name: infra-operator.v1.0.3
        version: 1.0.3
      - name: infra-operator.v1.0.2
        version: 1.0.2
      - name: infra-operator.v1.0.1
        version: 1.0.1
      - name: infra-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: infra-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: service-binding-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-binding-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:b355d2370926063a3de8c7a09c0ca8c2348a8d158d6a7095d0e5e8381cc95be5
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/redhat-developer/servicebinding-operator@sha256:b355d2370926063a3de8c7a09c0ca8c2348a8d158d6a7095d0e5e8381cc95be5
        version: 0.10.0
      entries:
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: beta
    - currentCSV: service-binding-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindablekinds-sample"
                }
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:c9c9e45321db2d4d3cdc2d364337af844f69c9893d0cc74d2fa56c846ffcf3da
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/redhat-developer/servicebinding-operator@sha256:c9c9e45321db2d4d3cdc2d364337af844f69c9893d0cc74d2fa56c846ffcf3da
        version: 0.11.0
      entries:
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: candidate
    - currentCSV: service-binding-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindable-kinds"
                },
                "spec": {}
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  },
                  "type": "postgres",
                  "workload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:ac47f496fb7ecdcbc371f8c809fad2687ec0c35bbc8c522a7ab63b3e5ffd90ea
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'.
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1alpha3
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1beta1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
          - description: Service Binding implementing community specification (https://github.com/k8s-service-bindings/spec)
              describing the connection between a backing service and an application
              workload.
            displayName: Service Binding (Spec API)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1beta1
        description: |
          The Service Binding Operator manages the data plane for applications and backing services. It enables developers to connect their
          applications to backing services (REST API, databases, event buses and others) with a consistent and predictable
          experience. The operator reads data made available by the control plane of backing services and projects the bindings
          data to applications either as accessible files mounted into the application’s container or environment variables.

          In Kubernetes, each service suggests a different way to access their secrets and developers will consume them in their
          application in a custom way. While this provides a good deal of flexibility, it also creates discrepancies and
          complexity dealing with each specific services, which leads large development teams to lose overall velocity.

          Service Binding Operator removes the error-prone manual configuration of binding information, provides service operators
          a low-touch administrative experience to provision and manage access to services and enriches development lifecycle.

          For additional details refer to the project [documentation](https://redhat-developer.github.io/service-binding-operator).

          Get Started with Service Binding Operator
          * Find out about SBO’s [prerequisites](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/installing-service-binding.html).
          * Refer to the [Quick Start Guide](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/quick-start.html) to see SBO in action in a simple scenario.

          Need Help
          * Raise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)

          Licence
          * Service Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE)

          Service Binding Operator is providing and supporting two different APIs.
          * `binding.operators.coreos.com/v1alpha1`: This API is compliant with the Service Binding specification. It is fully supported and considered as the mature API. It’s the API we recommend in production environments.
          * `servicebinding.io/v1beta1`: This API is the one defined in the Service Binding specification. It is available in Service Binding Operator 1.x.

          We’ll consider supporting only the API from the specification, once the Service Binding specification will be officially GA. Of course, that will be a breaking change and handle in a safe way for you.
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Documentation
          url: https://redhat-developer.github.io/service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/redhat-developer/servicebinding-operator@sha256:ac47f496fb7ecdcbc371f8c809fad2687ec0c35bbc8c522a7ab63b3e5ffd90ea
        version: 1.4.0
      entries:
      - name: service-binding-operator.v1.4.0
        version: 1.4.0
      - name: service-binding-operator.v1.3.4
        version: 1.3.4
      - name: service-binding-operator.v1.3.3
        version: 1.3.3
      - name: service-binding-operator.v1.3.2
        version: 1.3.2
      - name: service-binding-operator.v1.3.1
        version: 1.3.1
      - name: service-binding-operator.v1.3.0
        version: 1.3.0
      - name: service-binding-operator.v1.2.0
        version: 1.2.0
      - name: service-binding-operator.v1.1.1
        version: 1.1.1
      - name: service-binding-operator.v1.1.0
        version: 1.1.0
      - name: service-binding-operator.v1.0.1
        version: 1.0.1
      - name: service-binding-operator.v1.0.0
        version: 1.0.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: service-binding-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Skupper Project
      provider-url: https://skupper.io
    name: skupper-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: alpha
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1
    - currentCSV: skupper-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:e5f8b2ddf378cf3e40f4c7b7ec8482eb3321947cd46481c868d727932c449d78
          createdAt: "2024-02-26T18:22:07Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/config-sync@sha256:659fe28d1df7c99bba3bb854b1e9991c715ae4de99d91c88d5db5e4f9ef4c852
        - quay.io/skupper/flow-collector@sha256:3870c5bc89e623b7cebe21856bfefc22de76313705efa0ec551a5963789fff3b
        - quay.io/skupper/site-controller@sha256:e5f8b2ddf378cf3e40f4c7b7ec8482eb3321947cd46481c868d727932c449d78
        - quay.io/prometheus/prometheus@sha256:d2ab0a27783fd4ad96a8853e2847b99a0be0043687b8a5d1ebfb2dd3fa4fd1b8
        - quay.io/openshift/origin-oauth-proxy@sha256:a7dff785d821acc4b972d94964947713fade45f007dedb5deee17a2869bf5d24
        - quay.io/skupper/skupper-router@sha256:bf8d5955ce4e406275a4934278efe51c1af9a1a269a70e8fe642b7d94d80cb81
        - quay.io/skupper/service-controller@sha256:345edf2d9d72b5a97bcce0e8cea8cda82d120f7bf83756d73f9ac114c2fcced8
        version: 1.6.0
      entries:
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.6
    - currentCSV: skupper-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:4809db7263d77a95797061798e910f1bdbb5d5fe2db2bb3c743b97f5d0c7b3af
          createdAt: "2024-06-12T14:36:59Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/prometheus/prometheus@sha256:d2ab0a27783fd4ad96a8853e2847b99a0be0043687b8a5d1ebfb2dd3fa4fd1b8
        - quay.io/openshift/origin-oauth-proxy@sha256:a7dff785d821acc4b972d94964947713fade45f007dedb5deee17a2869bf5d24
        - quay.io/skupper/skupper-router@sha256:5eb258a0927e7706e7c035f1def2eb5211c8c33dd371fafa7210517fd894f68b
        - quay.io/skupper/service-controller@sha256:624e88e695297df28e485db7e802945ad1978ac6b0c9e9edca8b5e6013af8991
        - quay.io/skupper/site-controller@sha256:4809db7263d77a95797061798e910f1bdbb5d5fe2db2bb3c743b97f5d0c7b3af
        - quay.io/skupper/config-sync@sha256:a884b26a32179af5a2bfa85118bf873852eec6e2930e92da778b57506c0a0880
        - quay.io/skupper/flow-collector@sha256:bf646b0cb75e1c29bbe0261aa9943e06d33118f5e6a5ccb7503c60c974167692
        version: 1.7.1
      entries:
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.7
    - currentCSV: skupper-operator.v1.8.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:ffe3ea5067d350b64b69120bdf6d9e3ee9a30e8a2559b77170a10d66635623b8
          createdAt: "2025-03-27T20:11:53Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/service-controller@sha256:a45de211b8f8a8294a3588932a0a72766cf0d5ce64b01f4f744832c8d09aaa46
        - quay.io/skupper/config-sync@sha256:9e96d6a585f037f60ca3d89b5efee27ec92444d353e8b289ca35b94aeed1e309
        - quay.io/skupper/flow-collector@sha256:252244ab6aaf6b80118bfc06b690d4f5e8d7829b999fca60d194a84452fee7f6
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/openshift/origin-oauth-proxy@sha256:c15baf57dbd17dcc517cd2fa2299462d62124f689d6f5a48a3c9fd154fc77ccf
        - quay.io/skupper/skupper-router@sha256:7633b143f1844ee559322e4642ca80e8248a08711c8e814d489c26c39ecf9bd0
        - quay.io/skupper/site-controller@sha256:ffe3ea5067d350b64b69120bdf6d9e3ee9a30e8a2559b77170a10d66635623b8
        version: 1.8.4
      entries:
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.8
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.9
    defaultChannel: stable
    packageName: skupper-operator
    provider:
      name: Skupper Project
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Harness Inc.
      provider-url: https://harness.io
    name: harness-gitops-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: harness-gitops-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gitops.harness.io/v1alpha1",
                "kind": "GitopsAgent",
                "metadata": {
                  "annotations": {
                    "containerImage": "harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde",
                    "createdAt": "2025-07-31T12:11:18Z"
                  },
                  "name": "gitops-agent-sample"
                },
                "spec": {
                  "identity": {
                    "accountIdentifier": "accountIdentifier",
                    "agentIdentifier": "agentIdentifier",
                    "agentName": "agentName"
                  },
                  "replicas": 1,
                  "secrets": {
                    "agentSecret": "agentSecret"
                  },
                  "target": {
                    "http": "https://app.harness.io/gitops/",
                    "protocol": "HTTP1"
                  },
                  "upgrade": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde
          createdAt: "2025-07-31T06:41:18Z"
          description: Harness operator to manage and provision GitOps Agent
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Harness Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Schema for the Gitops Agents API
            displayName: Gitops Agent
            kind: GitopsAgent
            name: gitopsagents.gitops.harness.io
            version: v1alpha1
        description: |-
          ## Overview
          Harness Gitops Operator manages the full lifecycle for [Gitops Agent](https://developer.harness.io/docs/continuous-delivery/gitops/connect-and-manage/install-a-harness-git-ops-agent/)

          The operator aims to provide the following:

            * Easy configuration and installation of Harness GitOps Agent.
            * Provide seamless upgrades for Harness GitOps Agent.
            * Monitoring and autorecovery for GitOps Agent.

          ## Usage
          Deploy gitops-agent in namespace where operator is installed with basic setup.

          ### Prerequisites:
            * Go to your Harness account and create a new gitops agent
            * Copy account, org and project identifiers
            * Retrieve the agent secret key

          ```
          apiVersion: gitops.harness.io/v1alpha1
          kind: GitopsAgent
          metadata:
            name: gitops-agent-sample
          spec:
            identity:
              accountIdentifier: accountIdentifier
              agentIdentifier: agentIdentifier
              agentName: agentName
            secrets:
              agentSecret: agentSecret
            target:
              http: https://app.harness.io/gitops/
              protocol: HTTP1
            replicas: 1
            upgrade:
              enabled: true
          ```
          **Note:** Gitops Agent require an active ArgoCD instance running in the same namespace where agent is installed.

          See [Harness Documentation](https://developer.harness.io/) for more information.
        displayName: Harness GitOps Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - harness
        - agent
        - cd
        - argocd
        - kubernetes
        links:
        - name: Harness Gitops Operator
          url: https://harness-gitops-operator.domain
        maintainers:
        - email: alexandru.bosii@harness.io
          name: Alexandru Bosii
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Harness Inc.
          url: https://harness.io
        relatedImages:
        - harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde
        - docker.io/harness/gitops-agent@sha256:c9f57338be294d32a73a4cbecf90dfff50f5d36a699f3ec32d962dcb64d8d085
        version: 0.3.0
      entries:
      - name: harness-gitops-operator.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: harness-gitops-operator
    provider:
      name: Harness Inc.
      url: https://harness.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: maximilianopizarro
      provider-url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
    name: jhipster-online-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jhipster-online-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maximilianopizarro.github.io/v1alpha1",
                "kind": "JhipsterOnline",
                "metadata": {
                  "name": "jhipsteronline-sample"
                },
                "spec": {
                  "affinity": {},
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "env": {
                    "APPLICATION_GITHUB_CLIENT-ID": "CLIENT-ID",
                    "APPLICATION_GITHUB_CLIENT-SECRET": "CLIENT-SECRET",
                    "APPLICATION_GITHUB_HOST": "https://github.com",
                    "APPLICATION_JHIPSTER-CMD_CMD": "jhipster-quarkus",
                    "JAVA_APP_JAR": "jhonline-2.33.0.war",
                    "SPRING_DATASOURCE_PASSWORD": "jhipster",
                    "SPRING_DATASOURCE_URL": "jdbc:mariadb://mariadb:3306/jhipsteronline",
                    "SPRING_DATASOURCE_USERNAME": "jhipster"
                  },
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "quay.io/maximilianopizarro/jhipster-online",
                    "tag": "quarkus"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "livenessProbe": {
                    "httpGet": {
                      "path": "/jdl-studio/",
                      "port": 8080
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "readinessProbe": {
                    "httpGet": {
                      "path": "/jdl-studio/",
                      "port": 8080
                    }
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "securityContext": {},
                  "service": {
                    "port": 8080,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automount": false,
                    "create": false,
                    "name": ""
                  },
                  "tolerations": [],
                  "volumeMounts": [],
                  "volumes": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/maximilianopizarro/jhipster-online-operator:v0.1.0
          createdAt: "2025-05-30T22:42:19Z"
          description: JHipster Online provides an intuitive web interface for generating
            JHipster projects, allowing you to define your application architecture
            using a JDL (JHipster Domain Language) model and generate the code directly
            into a GitHub repository.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/maximilianoPizarro/jhipster-online-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: JhipsterOnline
            name: jhipsteronlines.maximilianopizarro.github.io
            version: v1alpha1
        description: |
          JHipster Online Operator: This repository hosts the JHipster Online Operator, designed to seamlessly deploy JHipster Online on Red Hat OpenShift and Kubernetes environments.

          JHipster Online provides an intuitive web interface for generating JHipster projects, allowing you to define your application architecture using a JDL (JHipster Domain Language) model and generate the code directly into a GitHub repository.

          Overview: The JHipster Online Operator manages the deployment and lifecycle of the JHipster Online application. It leverages the power of Kubernetes Operators to provide a native Kubernetes experience for JHipster Online users.

          This artifact includes:
          - JHipster 8.8.0. for generate Spring Boot 3.4.1 projects.
          - generator-jhipster-quarkus 3.4.0 for generate Quarkus 3.11.1 projects.
          - JDL Studio for add JDL models files by PR on your repo.

          Usage: Once the operator is running, you can create instances of JHipster Online by defining `JhipsterOnline` custom resources. An example `JhipsterOnline` custom resource is provided in the `alm-examples`. You would apply this (or a similar) YAML file to your cluster to create an instance of JHipster Online. Feel free to customize the `JhipsterOnline` Custom Resource Definition (CRD) instance to suit your needs, such as configuring GitHub integration, database connections, and resource allocations.

          Configuration: The `JhipsterOnline` custom resource supports various configuration options including `replicaCount`, `image`, `imagePullSecrets`, `serviceAccount`, `podAnnotations`, `podLabels`, `podSecurityContext`, `securityContext`, `service` type and port, `ingress` settings (enabled, `className`, `annotations`, `hosts`, `tls`), OpenShift `Route` (enabled), `autoscaling` (enabled, `minReplicas`, `maxReplicas`, `targetCPUUtilizationPercentage`), `resource` requests and limits, `nodeSelector`, `tolerations`, `affinity`, `liveness` and `readiness` probes, `env` variables, `volumes`, and `volumeMounts`. Refer to the detailed `README.md` for a full table of parameters.
        displayName: JHipster Online Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - java
        - spring-boot
        - jhipster
        - quarkus
        links:
        - name: Jhipster Online Operator
          url: https://maximilianopizarro.github.io/jhipster-online-operator/
        maintainers:
        - email: maximiliano.pizarro.5@gmail.com
          name: Maximiliano Pizarro
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: maximilianopizarro
          url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/maximilianopizarro/jhipster-online-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: jhipster-online-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: jhipster-online-operator
    provider:
      name: maximilianopizarro
      url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: postgresql
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: registry.developers.crunchydata.com/crunchydata/postgres-operator:ubi9-5.8.3-0
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.developers.crunchydata.com/crunchydata/postgres-operator:ubi9-5.8.3-0
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: postgresql
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Vertica
      provider-url: http://www.vertica.com
    name: verticadb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: verticadb-operator.v1.11.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "EventTrigger",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "verticadb-operator",
                    "app.kubernetes.io/instance": "eventtrigger-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eventtrigger",
                    "app.kubernetes.io/part-of": "verticadb-operator"
                  },
                  "name": "eventtrigger-sample"
                },
                "spec": {
                  "matches": [
                    {
                      "condition": {
                        "status": "True",
                        "type": "DBInitialized"
                      }
                    }
                  ],
                  "references": [
                    {
                      "object": {
                        "apiVersion": "vertica.com/v1beta1",
                        "kind": "VerticaDB",
                        "name": "verticadb-sample"
                      }
                    }
                  ],
                  "template": {
                    "metadata": {
                      "generateName": "create-user-table-"
                    },
                    "spec": {
                      "template": {
                        "spec": {
                          "containers": [
                            {
                              "command": [
                                "/opt/vertica/bin/vsql",
                                "-h",
                                "verticadb-sample-defaultsubcluster",
                                "-c",
                                "CREATE TABLE T1 (C1 INT);"
                              ],
                              "image": "vertica/vertica-k8s:latest",
                              "name": "main"
                            }
                          ],
                          "restartPolicy": "OnFailure"
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "serviceName": "defaultsubcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaDB",
                "metadata": {
                  "name": "verticadb-sample"
                },
                "spec": {
                  "communal": {
                    "endpoint": "https://s3.amazonaws.com"
                  },
                  "image": "vertica/vertica-k8s:23.3.0-0-minimal",
                  "subclusters": [
                    {
                      "name": "default_subcluster",
                      "size": 3
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: docker.io/vertica/verticadb-operator:1.11.2
          createdAt: "2023-08-02T11:29:39Z"
          description: Operator that manages a Vertica Eon Mode database.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/vertica/vertica-kubernetes
          support: Vertica
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventTrigger is the Schema for the eventtriggers API
            displayName: Event Trigger
            kind: EventTrigger
            name: eventtriggers.vertica.com
            version: v1beta1
          - description: VerticaAutoscaler is a CR that allows you to autoscale one
              or more subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1beta1
          - description: VerticaDB is the CR that defines a Vertica Eon mode cluster
              that is managed by the verticadb-operator.
            displayName: Vertica DB
            kind: VerticaDB
            name: verticadbs.vertica.com
            version: v1beta1
        description: |-
          ### What is Vertica?
          Vertica is a unified analytics platform, based on a massively scalable architecture with the broadest set of analytical functions spanning event and time series, pattern matching, geospatial and end-to-end in-database machine learning. Vertica enables you to easily apply these powerful functions to the largest and most demanding analytical workloads, arming you and your customers with predictive business insights faster than any analytics data warehouse in the market. Vertica provides a unified analytics platform across major public clouds and on-premises data centers and integrates data in cloud object storage and HDFS without forcing you to move any of your data.

          ### What does this operator do?
          This operator manages a Vertica Eon Mode database, and automates the following administrative tasks:

          * Installing Vertica
          * Creating and reviving a Vertica database
          * Restarting and rescheduling DOWN pods to maintain quorum
          * Upgrading Vertica to a new version while keeping the database online
          * Subcluster scaling, both manually and automatically with the Horizontal Pod Autoscaler
          * Service management and health monitoring for pods
          * Load balancing for internal and external traffic
          * Expose a metrics endpoint to integrate with Prometheus to monitor the operator's health

          For a brief overview on how to create a cluster, see the [Vertica GitHub repository](https://github.com/vertica/vertica-kubernetes). For an in-depth look at Vertica on Kubernetes, see the [Vertica documentation](https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm).
        displayName: VerticaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - database
        - vertica
        links:
        - name: Documentation
          url: https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm
        - name: Vertica Container Images
          url: https://hub.docker.com/u/vertica
        maintainers:
        - email: matt.spilchen@vertica.com
          name: Matt Spilchen
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Vertica
          url: http://www.vertica.com
        relatedImages:
        - docker.io/vertica/verticadb-operator:1.11.2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 1.11.2
      entries:
      - name: verticadb-operator.v1.11.2
        version: 1.11.2
      - name: verticadb-operator.v1.11.1
        version: 1.11.1
      - name: verticadb-operator.v1.11.0
        version: 1.11.0
      - name: verticadb-operator.v1.10.2
        version: 1.10.2
      - name: verticadb-operator.v1.10.1
        version: 1.10.1
      - name: verticadb-operator.v1.10.0
        version: 1.10.0
      - name: verticadb-operator.v1.9.0
        version: 1.9.0
      - name: verticadb-operator.v1.8.0
        version: 1.8.0
      - name: verticadb-operator.v1.7.0
        version: 1.7.0
      - name: verticadb-operator.v1.6.0
        version: 1.6.0
      - name: verticadb-operator.v1.5.0
        version: 1.5.0
      - name: verticadb-operator.v1.4.0
        version: 1.4.0
      - name: verticadb-operator.v1.3.1
        version: 1.3.1
      - name: verticadb-operator.v1.3.0
        version: 1.3.0
      - name: verticadb-operator.v1.2.0
        version: 1.2.0
      name: stable
    - currentCSV: verticadb-operator.v25.3.1-0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vertica.com/v1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "customAutoscaler": {},
                  "scalingGranularity": "Subcluster",
                  "serviceName": "default-subcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1",
                "kind": "VerticaDB",
                "metadata": {
                  "name": "verticadb-sample"
                },
                "spec": {
                  "communal": {
                    "endpoint": "https://s3.amazonaws.com"
                  },
                  "image": "opentext/vertica-k8s:24.1.0-0-minimal",
                  "subclusters": [
                    {
                      "name": "default_subcluster",
                      "size": 3
                    }
                  ]
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "EventTrigger",
                "metadata": {
                  "name": "eventtrigger-sample"
                },
                "spec": {
                  "matches": [
                    {
                      "condition": {
                        "status": "True",
                        "type": "DBInitialized"
                      }
                    }
                  ],
                  "references": [
                    {
                      "object": {
                        "apiVersion": "vertica.com/v1beta1",
                        "kind": "VerticaDB",
                        "name": "verticadb-sample"
                      }
                    }
                  ],
                  "template": {
                    "metadata": {
                      "generateName": "create-user-table-"
                    },
                    "spec": {
                      "template": {
                        "spec": {
                          "containers": [
                            {
                              "command": [
                                "/opt/vertica/bin/vsql",
                                "-h",
                                "verticadb-sample-defaultsubcluster",
                                "-c",
                                "CREATE TABLE T1 (C1 INT);"
                              ],
                              "image": "opentext/vertica-k8s:23.4.0-0-minimal",
                              "name": "main"
                            }
                          ],
                          "restartPolicy": "OnFailure"
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "customAutoscaler": {},
                  "scalingGranularity": "Subcluster",
                  "serviceName": "default-subcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaReplicator",
                "metadata": {
                  "name": "verticareplicator-sample"
                },
                "spec": {
                  "source": {
                    "verticaDB": "sourceDBName"
                  },
                  "target": {
                    "verticaDB": "targetDBName"
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaRestorePointsQuery",
                "metadata": {
                  "name": "verticarestorepointsquery-sample"
                },
                "spec": {
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaScrutinize",
                "metadata": {
                  "name": "verticascrutinize-sample"
                },
                "spec": {
                  "verticaDBName": "verticadb-sample"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: docker.io/opentext/verticadb-operator:25.3.1-0
          createdAt: "2025-09-24T00:58:06Z"
          description: Operator that manages a Vertica Eon Mode database.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/vertica/vertica-kubernetes
          support: Vertica
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventTrigger is the Schema for the eventtriggers API
            displayName: Event Trigger
            kind: EventTrigger
            name: eventtriggers.vertica.com
            version: v1beta1
          - description: |-
              VerticaAutoscaler is a CR that allows you to autoscale one or more
              subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1
          - description: |-
              VerticaAutoscaler is a CR that allows you to autoscale one or more
              subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1beta1
          - description: VerticaDB is the CR that defines a Vertica Eon mode cluster
              that is managed by the verticadb-operator.
            displayName: Vertica DB
            kind: VerticaDB
            name: verticadbs.vertica.com
            version: v1
          - description: VerticaReplicator is the Schema for the verticareplicators
              API
            displayName: Vertica Replicator
            kind: VerticaReplicator
            name: verticareplicators.vertica.com
            version: v1beta1
          - description: VerticaRestorePointsQuery is the Schema for the verticarestorepointsqueries
              API
            displayName: Vertica Restore Points Query
            kind: VerticaRestorePointsQuery
            name: verticarestorepointsqueries.vertica.com
            version: v1beta1
          - description: VerticaScrutinize is the schema for verticascrutinize API
            displayName: Vertica Scrutinize
            kind: VerticaScrutinize
            name: verticascrutinizers.vertica.com
            version: v1beta1
        description: |-
          ### What is Vertica?
          Vertica is a unified analytics platform, based on a massively scalable architecture with the broadest set of analytical functions spanning event and time series, pattern matching, geospatial and end-to-end in-database machine learning. Vertica enables you to easily apply these powerful functions to the largest and most demanding analytical workloads, arming you and your customers with predictive business insights faster than any analytics data warehouse in the market. Vertica provides a unified analytics platform across major public clouds and on-premises data centers and integrates data in cloud object storage and HDFS without forcing you to move any of your data.

          ### What does this operator do?
          This operator manages a Vertica Eon Mode database, and automates the following administrative tasks:

          * Installing Vertica
          * Creating and reviving a Vertica database
          * Restarting and rescheduling DOWN pods to maintain quorum
          * Upgrading Vertica to a new version while keeping the database online
          * Subcluster scaling, both manually and automatically with the Horizontal Pod Autoscaler
          * Service management and health monitoring for pods
          * Load balancing for internal and external traffic
          * Expose a metrics endpoint to integrate with Prometheus to monitor the operator's health

          For a brief overview on how to create a cluster, see the [Vertica GitHub repository](https://github.com/vertica/vertica-kubernetes). For an in-depth look at Vertica on Kubernetes, see the [Vertica documentation](https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm).
        displayName: VerticaDB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - vertica
        links:
        - name: Documentation
          url: https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm
        - name: Vertica Container Images
          url: https://hub.docker.com/u/vertica
        maintainers:
        - email: cchen8@opentext.com
          name: Cai Chen
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Vertica
          url: http://www.vertica.com
        relatedImages:
        - docker.io/opentext/verticadb-operator:25.3.1-0
        version: 25.3.1-0
      entries:
      - name: verticadb-operator.v25.3.1-0
        version: 25.3.1-0
      - name: verticadb-operator.v25.3.0-0
        version: 25.3.0-0
      - name: verticadb-operator.v25.2.1-0
        version: 25.2.1-0
      - name: verticadb-operator.v25.2.0-0
        version: 25.2.0-0
      - name: verticadb-operator.v25.1.0-0
        version: 25.1.0-0
      - name: verticadb-operator.v24.4.0-0
        version: 24.4.0-0
      - name: verticadb-operator.v2.2.0
        version: 2.2.0
      - name: verticadb-operator.v2.1.3
        version: 2.1.3
      - name: verticadb-operator.v2.1.2
        version: 2.1.2
      - name: verticadb-operator.v2.1.1
        version: 2.1.1
      - name: verticadb-operator.v2.1.0
        version: 2.1.0
      - name: verticadb-operator.v2.0.2
        version: 2.0.2
      - name: verticadb-operator.v2.0.1
        version: 2.0.1
      - name: verticadb-operator.v2.0.0
        version: 2.0.0
      name: v2-stable
    defaultChannel: v2-stable
    packageName: verticadb-operator
    provider:
      name: Vertica
      url: http://www.vertica.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: F5 Networks Inc.
      provider-url: ""
    name: f5-ipam-controller-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: f5-ipam-controller-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fic.f5.com/v1",
                "kind": "F5IpamCtlr",
                "metadata": {
                  "name": "f5ipamctlr-sample"
                },
                "spec": {
                  "args": {
                    "infoblox_grid_host": "IP_ADDRESS_OF_SERVER",
                    "infoblox_labels": "CIDR_Labels",
                    "infoblox_netview": "NetView",
                    "infoblox_password": "PASSWORD",
                    "infoblox_username": "USERNAME",
                    "infoblox_wapi_version": "VERSION",
                    "insecure": true,
                    "ip_range": "IP_RANGE_LABELS",
                    "log_level": "DEBUG_OR_INFO",
                    "orchestration": "kubernetes",
                    "provider": "INFOBLOX_OR_F5-IP-PROVIDER"
                  },
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "f5-ipam-controller",
                    "user": "f5networks",
                    "version": "latest"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "serviceAccount": {
                    "create": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
          createdAt: "2025-08-08T03:48:02Z"
          description: Operator to install F5 IPAM Controller.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/f5-ipam-controller
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This Operator installs F5 IPAM controller(FIC)
            kind: F5IpamCtlr
            name: f5ipamctlrs.fic.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 IPAM Controller in your OpenShift Cluster. This enables to configure and deploy FIC using Helm Charts.
          ## F5 IPAM Controller
          The F5 IPAM Controller is a Docker container that runs in an orchestration environment and interfaces with an IPAM system. It allocates IP addresses from an IPAM system’s address pool for hostnames in an orchestration environment. The F5 IPAM Controller watches orchestration-specific resources and consumes the hostnames within each resource.
          ## Documentation
          Refer to F5 IPAM Controller documentation
          - https://github.com/F5Networks/f5-ipam-controller/blob/main/README.md
          ## Prerequisites
          ### Configure Infoblox with network and netview [Provider: Infoblox]
          Refer Infoblox documentation
          ### Create persistent volume and persistent volume claim [Provider: f5-ip-provider]
          Create persistent volume and persistent volume claim
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/f5-ipam-controller/main/docs/config_examples/f5-ip-provider/localstorage-pv-pvc-example.yaml
          ```
        displayName: F5 Ipam controller
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipam
        - ip
        - ipaddress
        - controller
        - ipv4
        - ipv6
        - F5
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/userguide/ipam/
        - name: Github Repo
          url: https://github.com/F5Networks/f5-ipam-controller
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
        - registry.connect.redhat.com/f5networks/f5-ipam-controller@sha256:09a358db300d82e00fd3b4e255917df82c02e1bc58425f8cff21cc2c77eba1b6
        version: 0.0.8
      entries:
      - name: f5-ipam-controller-operator.v0.0.8
        version: 0.0.8
      name: alpha
    - currentCSV: f5-ipam-controller-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fic.f5.com/v1",
                "kind": "F5IpamCtlr",
                "metadata": {
                  "name": "f5ipamctlr-sample"
                },
                "spec": {
                  "args": {
                    "infoblox_grid_host": "IP_ADDRESS_OF_SERVER",
                    "infoblox_labels": "CIDR_Labels",
                    "infoblox_netview": "NetView",
                    "infoblox_password": "PASSWORD",
                    "infoblox_username": "USERNAME",
                    "infoblox_wapi_version": "VERSION",
                    "insecure": true,
                    "ip_range": "IP_RANGE_LABELS",
                    "log_level": "DEBUG_OR_INFO",
                    "orchestration": "kubernetes",
                    "provider": "INFOBLOX_OR_F5-IP-PROVIDER"
                  },
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "f5-ipam-controller",
                    "user": "f5networks",
                    "version": "latest"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "serviceAccount": {
                    "create": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
          createdAt: "2025-08-08T03:48:02Z"
          description: Operator to install F5 IPAM Controller.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/f5-ipam-controller
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This Operator installs F5 IPAM controller(FIC)
            kind: F5IpamCtlr
            name: f5ipamctlrs.fic.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 IPAM Controller in your OpenShift Cluster. This enables to configure and deploy FIC using Helm Charts.
          ## F5 IPAM Controller
          The F5 IPAM Controller is a Docker container that runs in an orchestration environment and interfaces with an IPAM system. It allocates IP addresses from an IPAM system’s address pool for hostnames in an orchestration environment. The F5 IPAM Controller watches orchestration-specific resources and consumes the hostnames within each resource.
          ## Documentation
          Refer to F5 IPAM Controller documentation
          - https://github.com/F5Networks/f5-ipam-controller/blob/main/README.md
          ## Prerequisites
          ### Configure Infoblox with network and netview [Provider: Infoblox]
          Refer Infoblox documentation
          ### Create persistent volume and persistent volume claim [Provider: f5-ip-provider]
          Create persistent volume and persistent volume claim
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/f5-ipam-controller/main/docs/config_examples/f5-ip-provider/localstorage-pv-pvc-example.yaml
          ```
        displayName: F5 Ipam controller
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipam
        - ip
        - ipaddress
        - controller
        - ipv4
        - ipv6
        - F5
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/userguide/ipam/
        - name: Github Repo
          url: https://github.com/F5Networks/f5-ipam-controller
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
        - registry.connect.redhat.com/f5networks/f5-ipam-controller@sha256:09a358db300d82e00fd3b4e255917df82c02e1bc58425f8cff21cc2c77eba1b6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        version: 0.0.8
      entries:
      - name: f5-ipam-controller-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: f5-ipam-controller-operator
    provider:
      name: F5 Networks Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: smart-gateway-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smart-gateway-operator.v5.0.1757650606
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "smartgateway.infra.watch/v2",
                "kind": "SmartGateway",
                "metadata": {
                  "name": "demo"
                },
                "spec": {
                  "applications": [
                    {
                      "config": "host: localhost\nport: 8081\nwithtimestamp: false\n",
                      "name": "prometheus"
                    }
                  ],
                  "blockEventBus": false,
                  "bridge": {
                    "amqpBlock": false,
                    "amqpUrl": "amqp://amq-interconnect:5672/collectd",
                    "enabled": true,
                    "ringBufferCount": 15000,
                    "ringBufferSize": 2048,
                    "socketBlock": true,
                    "statsPeriod": 60,
                    "stopCount": 0,
                    "unixSocketPath": "/tmp",
                    "verbose": true
                  },
                  "handleErrors": false,
                  "logLevel": "info",
                  "pluginDir": "/usr/lib64/sg-core",
                  "services": [
                    {
                      "name": "prometheus",
                      "ports": [
                        {
                          "name": "prom-http",
                          "port": 8081,
                          "protocol": "TCP",
                          "targetPort": 8081
                        }
                      ]
                    }
                  ],
                  "size": 1,
                  "transports": [
                    {
                      "config": "path: /tmp/smartgateway\n",
                      "handlers": [
                        {
                          "config": "",
                          "name": "collectd-metrics"
                        }
                      ],
                      "name": "socket"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/stf/smart-gateway-rhel9-operator@sha256:3363abcb91319f8642212b9b80ca49bb9f90cae69255b1f75f03929af3d3a8c3
          createdAt: "2025-09-12T04:17:59Z"
          description: Operator for managing the Smart Gateway Custom Resources, resulting
            in deployments of the Smart Gateway.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenStack Platform", "Cloud
            Infrastructure", "Cloud Suite"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/infrawatch/smart-gateway-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of the Smart Gateway
            displayName: SmartGateway
            kind: SmartGateway
            name: smartgateways.smartgateway.infra.watch
            version: v2
        description: |-
          # Smart Gateway for Service Telemetry Framework

          The Smart Gateway is middleware for OpenShift that takes metrics and events
          data streams from an AMQP 1.x message bus, and provides a method to having
          that data stream stored within supported data stores such as Prometheus and
          ElasticSearch.

          ## Core capabilities

          The Smart Gateway provides two modes:

          * metrics: provides an HTTP scrape endpoint for Prometheus
          * events: writes events directly to an ElasticSearch endpoint

          ## Operator features

          * **Two Service Types** - Supports the `metrics` and `events` service types which defines the type of Smart Gateway to be deployed

          * **Configuration** - Configuration of the Smart Gateway to allow for it to be connected to an AMQP 1.x bus and pointed at the appropriate storage mechanism based on service type.

          ## Before getting started

          1. Ensure an AMQP 1.x bus has been setup (such as AMQ Interconnect)
          2. If the Smart Gateway is running in events mode, be sure the ElasticSearch Operator and ElasticSearch instance have been pre-deployed.
        displayName: Smart Gateway Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - servicetelemetry
        - monitoring
        - telemetry
        - notifications
        - smartgateway
        links:
        - name: Source Code
          url: https://github.com/infrawatch/smart-gateway-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/stf/smart-gateway-rhel9-operator@sha256:3363abcb91319f8642212b9b80ca49bb9f90cae69255b1f75f03929af3d3a8c3
        - registry.redhat.io/stf/sg-core-rhel9@sha256:0e12903da0b15c64421102b5b914ca4c82d107295485c9b76ed19e91a45a4420
        - registry.redhat.io/stf/sg-bridge-rhel9@sha256:7fae342cf6cecc676eaa0a8fb64ee380ad101d667d6e8f0d89c943938bb6baf4
        - registry.redhat.io/openshift4/ose-oauth-proxy:latest@sha256:dc24aa382e02c4ac058d4b824a21f0e90a6259d6d817d0341b2a41809b5daa78
        version: 5.0.1757650606
      entries:
      - name: smart-gateway-operator.v5.0.1757650606
        version: 5.0.1757650606
      - name: smart-gateway-operator.v5.0.1743671602
        version: 5.0.1743671602
      - name: smart-gateway-operator.v5.0.1740051852
        version: 5.0.1740051852
      - name: smart-gateway-operator.v5.0.1731336910
        version: 5.0.1731336910
      - name: smart-gateway-operator.v5.0.1725953602
        version: 5.0.1725953602
      name: stable-1.5
    defaultChannel: stable-1.5
    packageName: smart-gateway-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-pipelines-operator-rh
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-pipelines-operator-rh.v1.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    },
                    {
                      "name": "legacyPipelineRbac",
                      "value": "true"
                    },
                    {
                      "name": "createCABundleConfigMaps",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.19.0 <1.20.0'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.3.2
          - Tekton Triggers: v0.33.0
          - Pipelines as Code: v0.37.0
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.22.3
          - Tekton Results (tech-preview): v0.16.0
          - Manual Approval Gate (tech-preview): v0.7.0
          - Tekton Pruner (tech-preview): v0.2.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.20/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:5021158a4752a26f8b33adea7da66695b22d2ad155d346d35dde5f971bc1a55b
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:5683321a5376e63c3ff3932270a1850d5e261f76212020ce867334166f90d2f7
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:a712bb411d2433b01793dff25785c1e265fa865a52f7c8732c9fb4418ab61757
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e7998e113f75e8f29602bae210c8b66ed06634b699f1d99c5875ef879220d733
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:96db721939b22572d7d8d8b95dcb5c92fdedb98e9cd0aabb7d17b703a915b07f
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:115eea1a5ca87f3b6b72db1df05b7386bf73dc70c9c2e4dc5e1a66700d960c37
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:980ed7e81ea55522e1681dc1e9b170b68d16f07061dd1f1ffd142367d59552f1
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:1b6aecfce45ed3163227b845b212e83385cdbe999edfc2e9387a0c76f3f8d8e9
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:161df929e2bd4c5acff594be31d5458365de5ae0d2189bb2f647c3fc6657fa28
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:bc10e7d6122ede5cf4f2882c200447683255465a1080e826761e90bfb983506e
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:df15eb1f1f5f237bfe481c16ab3d40bf7fb4123f3c930bcfb121e2cc8dc42882
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:d56d6b46331b4569077ab1e7a7b609ba01f0a1dd62b695d62f8aa0f32ea1b621
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:c0f438e6095aa81f3f7d820614e0414cc43374014c1505d7641bc1dba69a3903
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:818bfa4a3a874974b5f7f8bcf2faeb6c05e1795381ee1ceacf89f17b1ec0ef7f
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:9415e889ef98df6d2b2dca1e5d798fb765eeea0d9828e48424374d5267a6b0d1
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:441e1c7124c1674e3d4c38c3e82a4fe5568e76d13c89c04bf009d6b97435de09
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:7eafffa6e4e04f10c1d4e84e02d51c00eb8649c60f8f9a13da14492f24634891
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:bc403f93ef6c5526fa30152eab384c4a25bb0bc6ab4158a1e24f637b87a57259
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:2d496752b19f765f1ea36ec94b873ebd8fa604764a56aa67d3ed94b33e5fe75c
        - registry.redhat.io/rhel9/buildah@sha256:18453ab0a62154283aaf4c45efd7543d1c8ecca1aebae46dedc7ddc0f410b232
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:036250bcdc6fb95273bf07948dfdd4dce765ab84c803ceca0d0b4abf6ca19dec
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:8f0494341244a75546a4b80b065702be04e6d200421aa4e7ec90385c9dabc6ab
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:c4d05e546c466c57379fff502816012f615be073eb8cdce54677b4ab206a3f5c
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:bbd1918660be7b3c78777dcccc87d3b26bf5adfc077844ceffbbdd53696dc0a3
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:e3e64abef53151f6b15f15bcbff85668167a571e98badbd8a2ca252ff3a5b8dc
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:dd941acd357ae2c0fd7e15fb8d32d7ee6b2847435101dd1e1f107db48af18bf3
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:823e018e0329d764ac694f177ce88af9074f364c29a1c17929f9dd0b596d03dd
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:a7163a9e1fe891de021b1003558e6e64fe17ed17ab0820ba16bba09aa112401e
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:e63298d07ccec5804dfab4e42bcc8163c29cb3ac371c91248777423a4edd31ce
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:3cf990fb2fc422f3527a7c8b43d06ccd4d32bb5487ec8cb2d4a97f0f5d13f170
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:c3013c347bca8a520443a44ac2817fbba5744a9e91069ac5f09e91447ebf6b28
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:901b5108ceb669c985e75f5fb033c496e9f028863df99ab32651d1b08433a352
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:14dab8d87323fc7d112213f7f4db3ee7cf3841747de669ce710ae6a794053457
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:458f865f1b246c5f0682f5956795771a0515b89edfcffbbc0da1a3d3f532645e
        - registry.redhat.io/ubi9/ubi-minimal@sha256:2f06ae0e6d3d9c4f610d32c480338eef474867f435d8d28625f2985e8acde6e8
        - registry.redhat.io/ubi9/openjdk-17@sha256:555c8bcd2596637befd1a6da4aa949c5a2640cf66e09c2dfb932afe37e39095e
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:c31af237afb84214bf2a4c06a7636ca071fe97203b41da90d7878109ccd64784
        - registry.redhat.io/openshift-pipelines/pipelines-opc-rhel9@sha256:c1bee4a2bb495de6d4da962b6e7b2447caa51cc36c3005e4083bd31420ef5102
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:ec74a562c18fefe36ec7ae3ba083e4eb76c56b562f916578900f8767ae98e6c2
        - registry.redhat.io/rhel9/skopeo@sha256:a65a413f8a2864389a09dc750690d97afbcdc5c70821e0f85e99e8adba7954e0
        version: 1.20.0
      entries:
      - name: openshift-pipelines-operator-rh.v1.20.0
        version: 1.20.0
      - name: openshift-pipelines-operator-rh.v1.19.3
        version: 1.19.3
      - name: openshift-pipelines-operator-rh.v1.19.2
        version: 1.19.2
      - name: openshift-pipelines-operator-rh.v1.19.1
        version: 1.19.1
      - name: openshift-pipelines-operator-rh.v1.19.0
        version: 1.19.0
      - name: openshift-pipelines-operator-rh.v1.18.1
        version: 1.18.1
      - name: openshift-pipelines-operator-rh.v1.18.0
        version: 1.18.0
      - name: openshift-pipelines-operator-rh.v1.17.1
        version: 1.17.1
      - name: openshift-pipelines-operator-rh.v1.17.0
        version: 1.17.0
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      - name: openshift-pipelines-operator-rh.v1.14.1
        version: 1.14.1
      - name: openshift-pipelines-operator-rh.v1.13.0
        version: 1.13.0
      - name: openshift-pipelines-operator-rh.v1.12.2
        version: 1.12.2
      - name: openshift-pipelines-operator-rh.v1.12.1
        version: 1.12.1
      - name: openshift-pipelines-operator-rh.v1.12.0
        version: 1.12.0
      - name: openshift-pipelines-operator-rh.v1.11.1
        version: 1.11.1
      name: latest
    - currentCSV: openshift-pipelines-operator-rh.v1.14.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "db_user": "result",
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.14.6'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.56.4
          - Tekton Triggers: v0.26.2
          - Pipelines as Code: v0.24.6
          - Tekton Chains: v0.20.1
          - Tekton Hub (tech-preview): v1.16.0
          - Tekton Results (tech-preview): v0.9.2

          ## Getting Started
          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/container-platform/OPENSHIFT_DOCS_VERSION/cicd/pipelines/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)
          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi8/openjdk-17@sha256:a8165bc2cd5051a96d6937e25ed03155bbd4b731da6e58cebfe2ea83209c16d8
        - registry.redhat.io/ubi8/ubi-minimal@sha256:6825946f10e8c81586fa7815c4f06016af05cfc231d921ec9699402a6655d201
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:059c08316152c209de13017e322d6fe5eacfde81c3b0039e55b5de8f2f27183a
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:d01ba7ead670d5ab9c1355b90010657988e46b4d1aa0bb68b194eaf4001b2d7a
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:6327385e3d3c9df1ea50e803955d980b4cd35fb15593fcfec0d477d91a45021f
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel8@sha256:6e7c7a1ca458e08ad85e1146815ec00a30016f709b45e28b464043e875c2a3e8
        - registry.redhat.io/rhel8/skopeo@sha256:e297b84ba6f1d5eb2c37e2c3b31c9a9a4c39f9590e7d7c32388165245bde4835
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:ae0260abd2725c2f4bb25c0c07d86290978892a49050e0fd23f466401aba668c
        - registry.redhat.io/rhel8/buildah@sha256:b48f410efa0ff8ab0db6ead420a5d8d866d64af846fece5efb185230d7ecf591
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:2261f4f657f70759f4d0c898e44ffafceb8dbe2144063d50227140463bdbf2c5
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:7a63e7a762b9c7b15057ed9d5bcbea08acb4c8933a300c6ec336824d49a6e937
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:8d4450d68ac03d8275e337d2c55692ef35e954400d0c9f041a943d163a8b7680
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:bf6cf2e87fb19f7aa9a490b83c16af69834c0721220a643710a1b077959e91ca
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:aa6eb4c7ade9d6add1e4ab94197c2ce72a253fea0d7461698ca0e37854f1dcbe
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:71d91c7b80b6d196025f68749bd8518b8b0affd8d5064a7c5c70d125656fbc36
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:db3ff043410d1dd0621fef4483761727f7241d788b136e446fb861953350a2cb
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:5e7f5a2813a591e7b13fea9fc4e2f599b5ec855a015cd5a4c287c4e178358756
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:9c29f6eff1f7019993446d85d226f04c5614a18351dfb4042a8bcb62276810de
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:d9972bc64dfa7e260bd9b949bbb85ef5ae532e7c6df8d874d3d3963af1097726
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:bf0c38beefa24ca3d04ebfa6140a1cc5a9a56fba81ae05799279a559093b9033
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel8@sha256:b3bedee6084e9d9da549d3e637d2bf176aa365b5818e6b6377525577edd710dc
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:82c8a91d521d7e6d81fb462fa2b804f420f52ee062df5cda39545165a2f19e0c
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:ce0ea093c522af329acdc85148c2473ea45cd44270c15f8dd32965b0d56f304f
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:abf4442b6acc714c2392b86af046d5f32d247ee0c3763bb18c57d0473edfcf02
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:e2bd929246d62038a24a0685cb3e4873c131c9b7713d6a5268ca09a8614538be
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:83913744a1bb7ff23091cfa541b807cfb82b38dde0c37e0f1dd9c17cade99775
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:63fc1351191a54d12d4e463a1bb4cc732620cba3000af86ba7f7a579514f5fe4
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:e23c901868229e4f6a4016aa7798f1aba879ff35ffe594f5b0e4d39f1bff9e52
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:f68dc18a4dc6a1886507097551f4a6fce39062d03c6eb86c067d317812b0d34a
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:f9fed20cdfebbb226e035cc62f037174f5ecdb3139a5edc7e39f019f6d2b1e98
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:ddf0cb64ea57de872515e19de63b2dec5f56738c9d7eaa283a4a489545ef4902
        - registry.redhat.io/rhel8/postgresql-13@sha256:3903d562425dd564986123280dc076548fb3d24c74138dcd857652f53052e0a7
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:2e75096ffe20d1261e8a645fefd32e37295d727aa6a280f10041a44443917469
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:57fb86ba112d79a228cbacc96cd0525c123464bcac956354722e3f2aeff42508
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:ffa8c6d8fb5bacb59da27f4ebde87385359b3f97a6a023b4a4488ced2d46c574
        version: 1.14.6
      entries:
      - name: openshift-pipelines-operator-rh.v1.14.6
        version: 1.14.6
      - name: openshift-pipelines-operator-rh.v1.14.5
        version: 1.14.5
      - name: openshift-pipelines-operator-rh.v1.14.4
        version: 1.14.4
      - name: openshift-pipelines-operator-rh.v1.14.3
        version: 1.14.3
      - name: openshift-pipelines-operator-rh.v1.14.2
        version: 1.14.2
      - name: openshift-pipelines-operator-rh.v1.14.1
        version: 1.14.1
      - name: openshift-pipelines-operator-rh.v1.13.0
        version: 1.13.0
      - name: openshift-pipelines-operator-rh.v1.12.2
        version: 1.12.2
      - name: openshift-pipelines-operator-rh.v1.12.1
        version: 1.12.1
      - name: openshift-pipelines-operator-rh.v1.12.0
        version: 1.12.0
      - name: openshift-pipelines-operator-rh.v1.11.1
        version: 1.11.1
      name: pipelines-1.14
    - currentCSV: openshift-pipelines-operator-rh.v1.15.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "db_user": "result",
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.15.3'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.59.6
          - Tekton Triggers: v0.27.0
          - Pipelines as Code: v0.27.2
          - Tekton Chains: v0.20.1
          - Tekton Hub (tech-preview): v1.17.1
          - Tekton Results (tech-preview): v0.10.0
          - Manual Approval Gate (tech-preview): v0.2.2

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.15/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:322ef2385620661a7914f47c73a6d145ec0c9ce1ccde30645bb66148545ddc96
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:49c346bf7fc69a9796690236dc8597e35a874343b02fb74c5d2f1c05d356e41c
        - registry.redhat.io/rhel8/skopeo@sha256:29465586e92ed04d2878dcdeda2d508c7e8e7539762e10296b52d631929960aa
        - registry.redhat.io/rhel8/buildah@sha256:aac6629389db17e99894c5bee0da01d4c8065d11d8c6f6e1602f9484290baa70
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:bf94c350923da706c6e0a427e2d91af3e13e751a12388a4ed75d1c34dd7b3177
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:ac1a6e057a5599fda7e048ff77030e98c92b5d51cac2472c1ef3910964fc17e6
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:61c22119f37e620a3807056cde35bc07c90867765b149294056326b1d11494f1
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:abac279a0a5ecde57952b6bef634f8ba7527cd8c8956a2ea30e9449174ad52bd
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:166789a990df9f6ab04fd3df0a8cd301bc19b9bedde94215cb27306060eb1e2c
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:83eb08bb9579cc19132af282c455bca32a0fd4f5436a35e7ad23b5efa2faa3cf
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:479e3eed93e0aa13e3489514fe00375a519eb45969d099c490b4acd056ea5adf
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:d1f86458079449c86407ac716c3cbf608cad06de29edeecbde6586f76628daa9
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:e749db81c9d7e11e2802833a2a02e93afc357c83d1a2a827e537262df281a4f7
        - registry.redhat.io/ubi8/ubi-minimal@sha256:f729a7f5685ea823e87ffd68aff988f2b8ff8d52126ade4e6de7c68088f28ebd
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:7c81ae208a16e17d0102074867add37dfb1207ee95cbbc6c732d88e61bbc59fb
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:7152c3c174d25f2a5239456ba677ac5315bfb82bc3b3ac5ee0eb8f9f74fb6632
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:aca0024cac7a8c41b1ed151eb1e1188f09bc636866f4e03f1b973f90890b37cf
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:bf6cf2e87fb19f7aa9a490b83c16af69834c0721220a643710a1b077959e91ca
        - registry.redhat.io/rhel8/postgresql-13@sha256:a92a579f1aef66ac188d24fd489c456a1a3e311d95dcce652da6b81e28fbf725
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:7f00ff1b407d9ea107d3d969988a6279b28bde86943f490a007fa922dd40a1c3
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:eed464a3b4e23f0b2623aa1b923c10232d235f4272f562955926c1748e295d2b
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:7958925cecce1b4b02ac4649e3bfd7fa3b5361c16ab3593700b8da93fdf1c79a
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:312d1f90765d05a86845856c677e1ece4c57e81a9bfe3fd1e76868d624e3d72d
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:1300d171dbf09c63071ba5586e223491a13acfc6dbab27eff04648aeb788d455
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:538c63c8e7af3cba91526f2f9ac99fb8f6b0f4781db31771c959612c209dd090
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:601fcac1abc6c8efa1248c5ca229213d6660708470a1c416d41c6aff0ba4691a
        - registry.redhat.io/ubi8/openjdk-17@sha256:e8cc2e476282b75d89c73057bfa713db22d72bdb2808d62d981a84c33beb2575
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:e7b000a2735bbf98ee79bd093f7b703ed09d7c1e11ed2a5b22c4b8ac8d440386
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:827313c2691fb8833e42418b65b4a3360a11cabc2c92a7d8e7b20183f06cbf68
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:f02af8475c33f6f38a2f90ff1be1fb1b2094860c4d5e300aefdd7eb84b4a529a
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:acf8017029f4b2e7a1c3da9bc62c65239d313d24e3012712356a191bea9a2e6f
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:75d6228eb656b8778fb3bb8dc68c686ac1c9b787e5343ae3c2467f8e724af83e
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:1db14f23077785108660dde38c98e2e30c8301f8e2e1c594dd4636e0ba29f84f
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:70bb0ebcfe5da45e5c239aa011dc5f1ef48844116d7346bb05433b503107268a
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:f04ea12675e0eb2354635cd2f9e99d6e59619d1c611e0ea95c3ce8222d615322
        version: 1.15.3
      entries:
      - name: openshift-pipelines-operator-rh.v1.15.3
        version: 1.15.3
      - name: openshift-pipelines-operator-rh.v1.15.2
        version: 1.15.2
      - name: openshift-pipelines-operator-rh.v1.15.1
        version: 1.15.1
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      - name: openshift-pipelines-operator-rh.v1.14.1
        version: 1.14.1
      - name: openshift-pipelines-operator-rh.v1.13.0
        version: 1.13.0
      - name: openshift-pipelines-operator-rh.v1.12.2
        version: 1.12.2
      - name: openshift-pipelines-operator-rh.v1.12.1
        version: 1.12.1
      - name: openshift-pipelines-operator-rh.v1.12.0
        version: 1.12.0
      - name: openshift-pipelines-operator-rh.v1.11.1
        version: 1.11.1
      name: pipelines-1.15
    - currentCSV: openshift-pipelines-operator-rh.v1.16.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.16.4'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.62.9
          - Tekton Triggers: v0.29.2
          - Pipelines as Code: v0.28.2
          - Tekton Chains: v0.22.3
          - Tekton Hub (tech-preview): v1.18.2
          - Tekton Results (tech-preview): v0.12.3
          - Manual Approval Gate (tech-preview): v0.3.1

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.16/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:81a8b903d0fc4b81e6a23fac862d18da4cc18b14cca8213665df37dfaabbc2e9
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:5672f3a375c6a274200b423eb2643c818aba75488dec0b210fec2532814ba51e
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:47f094ccd12db7a43c8a91c1a3d2a190bc2b83f1745e7eb8ce097d0c8e1e2c88
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:80d46bc5ce8c931af22a97ccbce05b59740e313167c88ad910febcc2f2dc0320
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:90e2a7ea498da756e09c7e4d19c7fb91d429cc28c57f97c6958958e3466cb58c
        - registry.redhat.io/rhel8/skopeo@sha256:1dcb4189c92da21ee58b3ca5433615e0208194c8412f7f1fb45fa491179aaf9b
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:d2bb3f9f79ef719d8fdafef3a639e3eb8bc199ee56fcfb33e85975d536d2a6ac
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:7edbd11e5a449408d673cae1fc2b9b5bab492ecff2ac6e640c8cd69a9b90ff08
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:a5a7b2be50f211d4d9d555b2a2e2302096c139ff5011d984cc386f1530007f16
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:f9122d71c3862cd2afafc8eb7f1a39c83eee306e5baf38514e6e39215c70a72b
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel8@sha256:8cb87920e6fc8185ead49aafb7799fead6a4a510eafdae0ee29e85248dae3945
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:aacdabe243dc573f3161a3d04fbf014b77d9962207305d70fcafce475f7eea93
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:92042dcacd4ab0f2a78aab03fd91a9a39d7d7366b5398fd797ecb145acf5fce4
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:2fea7d8e29d70c5893ddb6caf2bf1d4144787f5f0de8739351d2097ed09ee751
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/rhel8/buildah@sha256:5c7cd7c9a3d49e8905fc98693f6da605aeafae36bde5622dc78e12f31db3cd59
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:474658c262b539dd1c83ac22bb090b3fdac5de9cffc06587d36dbc6fa4ce208c
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:e1fe4a86211cf22e5ad2b23c989d24b69f597db4c3604e84777dbd0c585ec186
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:6acc31f179b08658976cddcb9ef42f377edb29933d1d83a2ecebfd455bb68571
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:f3e6df07fd86acb91037d55a1970e5f8c6ab12a7517752692c224c7687022812
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:8129e62906c68f04075a44c939375a7bafdeaeac673a20a4fa74ffd6285187ac
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:7d92dca9da27c066cb530172e1811493d8c916e67d8068f5ddc471fa3f3f7856
        - registry.redhat.io/ubi8/ubi-minimal@sha256:a47c89f02b39a98290f88204ed3d162845db0a0c464b319c2596cfd1e94b444e
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:b94bd566ae3d9ce5f47273d867164486826b19fca0ce11d1a7d18312413e87ba
        - registry.redhat.io/ubi8/openjdk-17@sha256:632e78989471859ca4ed1148b951a911f7e3b6e6516482be20967c4171612c55
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel8@sha256:46635d636ea2e9a6ba0760bf8c528028700ee40b0ba8466797034371185b16d0
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:856229e9e37683c926e3449658d76f93d910384eff62c2ca192768a653b71533
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:c6d654c71aba35c3c7f5c1a2ca5b8401e95adf6f649c599a9aff25177c2b3323
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:549d74a18ec52db91fa949cac9b8e3d943d478d5dec9850d2ba0cb1bb526f578
        - registry.redhat.io/rhel8/postgresql-13@sha256:a308263f5d1071be8ac5c13fd84b40609988ca30ed56c22a8bc6769e107ce09b
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:b41b888fc1a3b7215e6b71ee97bf64a1816914a8f33e96ad3930a4e2d5b5ffa7
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:4e0e4c358faa9cafc3964d6d40ee4aa73f8dfe8b6338dfb8206613d7c5241842
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:dcab4a3d519975340ade59a0ca0222104e68acd052a61f2aeb8c736be4b672cb
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:e84c2d80e92815c6bf384e1504a194dec9b796a3aceac2eed555fff241249571
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:d36dbb60c7b5caca7f3be08705f8ab91d7f01abaf8bfee4aee73bdbbb3e7e71a
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:7da7f929d2711fc55dda01fe51a474bfa7adf5aaf78be6d65a6d65850958d6e7
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:19b3dd69c1d7e17890b7163aef322c1efbc76fed7f2b1faf26aaa05fd8161184
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:aeeed543412cde0bd3a9baec965cb99d80f2b6acb77ae6356ed51b4b1262d006
        version: 1.16.4
      entries:
      - name: openshift-pipelines-operator-rh.v1.16.4
        version: 1.16.4
      - name: openshift-pipelines-operator-rh.v1.16.3
        version: 1.16.3
      - name: openshift-pipelines-operator-rh.v1.16.2
        version: 1.16.2
      - name: openshift-pipelines-operator-rh.v1.16.1
        version: 1.16.1
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      - name: openshift-pipelines-operator-rh.v1.14.1
        version: 1.14.1
      - name: openshift-pipelines-operator-rh.v1.13.0
        version: 1.13.0
      - name: openshift-pipelines-operator-rh.v1.12.2
        version: 1.12.2
      - name: openshift-pipelines-operator-rh.v1.12.1
        version: 1.12.1
      - name: openshift-pipelines-operator-rh.v1.12.0
        version: 1.12.0
      - name: openshift-pipelines-operator-rh.v1.11.1
        version: 1.11.1
      name: pipelines-1.16
    - currentCSV: openshift-pipelines-operator-rh.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.16.0 <1.17.2'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.65.7
          - Tekton Triggers: v0.30.1
          - Pipelines as Code: v0.29.1
          - Tekton Chains: v0.23.1
          - Tekton Hub (tech-preview): v1.19.2
          - Tekton Results (tech-preview): v0.13.4
          - Manual Approval Gate (tech-preview): v0.4.1

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.17/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:a14270967460883f283ec1ce17ff8b82db1e0777dda4f6d690f15ac38dcbad85
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:650391ffd3e4c7b173bf9d718405b2375c6ff2027abc0b7ce4f67e79e7ef4b81
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:37109f6de75dda750496f6668164d0dceaf9db5278e8c034708436f39de3276e
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:a392d2c85baedd7428c579e91d9f8acf8232399b90523b1093dc3447aaf56a16
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:8e52b4b31b0d9ec772abe6e6ef9c6698e89893395f364968216bebc3367da4b2
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:4b87782b35e7284a365c49f1c094aae98531ac630db1958294009c041a2d8440
        - registry.redhat.io/ubi8/openjdk-17@sha256:e5fc9e34382da41b10d4ea566639d31707493abff9e8de81f5accfd66876409e
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:4224040b00cfd76e85f80aecc5095098564cc62e9c477fe2f7a428b30df7d4d9
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:41ec6db921812fd1e80b05d51c224ddbce63bf0c57f02ccf07461d2202e97b90
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:4fabae1312c1aaf8a57bd2de63bd040956faa0c728453f2a4b4002705fba0f0c
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:7abc88bea01dd6513f6e48334f0b98501f909a1dc1ab9e83efeb3c1a834f8f92
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:eab798472b76bd51235198fd3b95189740f8d650ebc6bb01fe653c025c1defaa
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:61c9e8b1660d8935ee316ecc113e8141daf05149d172685cabd861ae8aeed3eb
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:4240501c076605a01d969328bbb8ad885c598b44ec1dc1c066312491a0eb6715
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:7c710882752f12be3c22bedc8d32096c111e42c5b9b392163f61cf8d72be3ec0
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:66309ae437c4647f48efe3d6f2b696df83859323700427d0025b94795751a859
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:41e2f2c7397a8b9e093e18977666e93943734f51780c769c99958ae1d20af53e
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:4a96f1bb7e3a5c7ca62dff2199fa15ef964b129e4e36e9619837c1221538f496
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:3696cf724ce7865626ae408d82fd87727c741e4d3eb3475c271ec6d56f0882a4
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:44da76d75f567c2ba623bae646e1d552d2c0fded1accdde3aab4f4bcdf3adc28
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:26f7b695fc1430359ab2bfd2915cf0b7e88b06d8695e26761bdc2ce8935f0fb3
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:0b6ee6cd98cb7e835cfc9544996c6f85288d7252071c100bed853eb593d6befb
        - registry.redhat.io/rhel8/buildah@sha256:6d2dcb651ba680cf4ec74331f8349dec43d071d420625a1703370acc8d984e9e
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:d851a2dd8ec95bfd7ac6ec5515c8cdadfcbfc6e6ec84b40630b4fe595f978fa5
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:429bc3d6044ca7dab780ff4b5eb91d42d38a29eb7ef428a25c53b7dda49871bf
        - registry.redhat.io/rhel8/postgresql-13@sha256:757d8dd6f243717d820846da7144a49913f28f223b74ec305b2215f54720f8a0
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel8@sha256:b5cceead6b8e9c99655c9ba810679b9d6fb978e5ff18d020a46d475c4c26fa29
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:c92ebe828b6ced6c93a3df5f4ccdfba022471729ad08b184de2f7410ecf592cc
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:df11cdbab2a1c6b9b800ef124c73f2fb5e3981402b21bc822ccc13e95836197a
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:db27cb4f0c350d99546b0f1bdf02d1ba271a8053a4552d88ff302adca60a3e29
        - registry.redhat.io/rhel8/skopeo@sha256:7fd65200b4847fc84fba0466519b87f915e28c0e9bc90d3022beecfebffd4af8
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:99f5383d8c6bc046a0bf48fd6c2064a5d055645085248327305f3ee4ae71979c
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:e673c20fff2fc684f5098282dd6ac6cb265045945249e995b81289532ce8c856
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:9668d852123b160e49511242371edc6e941a8e8b8c2ffa60080d91cc6b97d3d7
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:84eed8ba1189448e02ffd284c7d913e7b3cb7efe7a8e5b0edf890f7eb17fc439
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:922b9e86a674a6f3eb82f510569ab9cde97a1ad833c263335cd4164bf4164b66
        - registry.redhat.io/ubi8/ubi-minimal@sha256:c12e67af6a7e15113d76bc72f10bef2045c026c71ec8b7124c8a075458188a83
        version: 1.17.2
      entries:
      - name: openshift-pipelines-operator-rh.v1.17.2
        version: 1.17.2
      - name: openshift-pipelines-operator-rh.v1.17.1
        version: 1.17.1
      - name: openshift-pipelines-operator-rh.v1.17.0
        version: 1.17.0
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      - name: openshift-pipelines-operator-rh.v1.14.1
        version: 1.14.1
      - name: openshift-pipelines-operator-rh.v1.13.0
        version: 1.13.0
      - name: openshift-pipelines-operator-rh.v1.12.2
        version: 1.12.2
      - name: openshift-pipelines-operator-rh.v1.12.1
        version: 1.12.1
      - name: openshift-pipelines-operator-rh.v1.12.0
        version: 1.12.0
      - name: openshift-pipelines-operator-rh.v1.11.1
        version: 1.11.1
      name: pipelines-1.17
    - currentCSV: openshift-pipelines-operator-rh.v1.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.17.0 <1.18.1'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.68.1
          - Tekton Triggers: v0.31.0
          - Pipelines as Code: v0.33.2
          - Tekton Chains: v0.24.0
          - Tekton Hub (tech-preview): v1.20.1
          - Tekton Results: v0.14.0
          - Manual Approval Gate (tech-preview): v0.5.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.18/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi9/openjdk-17@sha256:5d619e202695acf76a5f32c07ace6709f5ee5d3999bf977e1bf95ad34197f44a
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:86ea94b12e8923035e3e4f39b6a53b3ab32c5f77d7eea170aa9b34aecc6f8577
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:0a60ce46c05256c74f770901e0bb79ad89e88ba882b643c1ff7b2a1738463acc
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:3964c3248dfb1006ea9a46295baf7f8cfa71952d6b9863b52f87e8f3d7863b1e
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:36a4c6a3273f2371ab5e679f1955221c286ea753e29e6ab8da7c983681a68f02
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:b550a8bc2a87d07a5b22a3ec3fb83efd8eb26ca6e80d404f2e453cc9b8434da7
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:cd28c2c90dd340324b0d8709640cd7636f28286114b0023649b6e21be13e03c2
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:fa5c456fc91319c55fac9a011dc7bb32ef85db293c73d74436e903b76ceb4349
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:409c14649d8e7007d4b72b6feffb71dabd26322bf2a7c77f5307f2546fdfe81b
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:96f5e1922fb9d6dc17e271576dda9bd032bd6dcc04e7496fb15a2f6b5871cd67
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:16d799503674641d4690822edc9c399c61429ccd31621d295eee4a4d1ef1c076
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:cbd89c531953a43b8055bda72f7696be80f57fd1d782567a91d82e0fbcb3d26d
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:8fec8e7a8b3e5182c67104de4e3abe32593b682c4a8986d1e9ab835dafb95ff4
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:3696cf724ce7865626ae408d82fd87727c741e4d3eb3475c271ec6d56f0882a4
        - registry.redhat.io/rhel9/skopeo@sha256:d077e35100d6d7b3795f8e3594293dbf8b65c6bc18b909aa555fe56f048b356a
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:1b57b64657eca0002145100bae262a19f02eb40068e07994df467e28a1618f90
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:20e19c02efdecf94a7b16483047e5db81c50dc2a0aa59d4895067e6c7769efcc
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e2b4c3b505c4b5aa36f8a5e661389b0de4e799c74c2c7f16a065e5522b438e89
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:80d441ec5d07c5059277a3a9173a35dfc0b8372d6e4d7823514131a94bae3fd7
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:d72842a7c6a1e50a68576c93f60047afec801cf83db64d7d3e34b25bf3c753ec
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:e6a7738cdb0e8aa9e1af381273b1c18ec1414b08a66fc851f6e11e5aa7a6706f
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:fcd5fd10de7334bd59a8da4997f91008ced95c6cde40d56d50f3b3e32c0f19e1
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:c20ab6fa719506bca942669ea84b113335b05df3ed275257b88be234cade7c8d
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:2ba6782af0d7e5c4d623de31814f43f4f7c3c38d09b57705021957d334067f87
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:d055d2b35a663aef3e1aafdbed0b12957867c0670c946ebae66e9c44a768bda2
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:72901cf4e091fe5f6a577144fdffaea8570712e9887a8bd715d5ae5bc346ff1d
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:b9675b22bc8759903472c1eb0e03f90c75460f02799d894294c161641cad9414
        - registry.redhat.io/rhel9/postgresql-13@sha256:ecd5a5583db922c517c735797db3fca587c0ef0b410e7dd494ace2267d89cf05
        - registry.redhat.io/rhel9/buildah@sha256:6886997da8c92214e99e3e8da47a0f08239172d2544b4edae058d5cb018effab
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:c90a9c3a20f088241980b76986ba4f43b0243fd6d94a826fa7b5888e01c13093
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:776d34b0ea2e35c11ad4050e6b40cf4b40c84dbe0d5f3e6ad669c494a210ea8e
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:8b4bf043e0fb237553503d63b129c432600bb4e65b2096b246b257b404b42be9
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:eaad37453a5ebef51c28d4a068f56e9f394e710c72bb86342cd2189001420836
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:8ae0ca6411e4fcd5c0605344bc6930cf122ffd80e71199509d58423a8b41de1d
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:dbc462cb50385dec3a220b5d0caed0e0211c166182026f86e3c51a1351ba0de4
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:6abc903c784bb455e1f2fde45a48afe775239668e1d0dbc1b3cc6afca3c0b477
        - registry.redhat.io/ubi9/ubi-minimal@sha256:ac61c96b93894b9169221e87718733354dd3765dd4a62b275893c7ff0d876869
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:e27a30d471dda4983a5b8192709d283a9d082558f3caf31b30db2214e418b017
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:e63d956a0534f2e726301e00e877c4fe2f34ca4559e59d2df7567c04ded311a3
        version: 1.18.1
      entries:
      - name: openshift-pipelines-operator-rh.v1.18.1
        version: 1.18.1
      - name: openshift-pipelines-operator-rh.v1.18.0
        version: 1.18.0
      name: pipelines-1.18
    - currentCSV: openshift-pipelines-operator-rh.v1.19.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.18.0 <1.19.3'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.0.0
          - Tekton Triggers: v0.32.0
          - Pipelines as Code: v0.35.3
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.21.0
          - Tekton Results (tech-preview): v0.15.3
          - Manual Approval Gate (tech-preview): v0.6.0
          - Tekton Pruner (tech-preview): v0.1.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.19/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:22cfadc42e793553182acf05a34fb6a7ae0d7055431c2e393699ecfeb171b173
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:7a0c6794e9284209dc131f3087ffc01fd5946e20a9261422808c6dd3938407de
        - registry.redhat.io/rhel9/skopeo@sha256:993f799d58c48151ecae7a84e63860f6c0039e8e18efed5ad67e7752f4b7e408
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:fd6f4cd7801d7c20cf5a6e546022136404212da5d9997750e643f6e640b00e3d
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:2c3074ecc0c74d508020bf1e11667d868150915b55d436a43c2b4fc22815bf48
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:80af5fa2d87e7c405f16575519293ca7fccc49d635b4f3d0d334dea4fbd28721
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:25bcdb30f0fe34300c4f61f62513f4853f65fe11bafde58bcf19b7b3c3fb9af7
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:251c87e2c48e7f377bcb399ed64d50dff11de3459629a73a63bac17c353eb16b
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:363e28a82ed022e507c3976d3a42ee63c053eb7871712c213b12235d8234b7ef
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:b247e9f3a2db32b575999387ef4f91d685a67c1f3950a6e265e1fe14c1fac196
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:6aaabb9af37e46a856371e29f141de6b0fac623128099bb927fc30f8f61e9db4
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:d000daec36a6cdbe41a85ee4eac24a712f7628441a412f751baf8d7778626f62
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:8bcaeaa751c61eb927d6353984f81d6a92b4bd4f9a5be9dc620a11b06ff9e6ba
        - registry.redhat.io/rhel9/buildah@sha256:a6e04c061b8d261dba85d01dcb64d3f7dba8c0b9765e03f05ec79850a248d75a
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:41d669fbe8bb48481890fb9fb52de9d4cac98cb74e6a6b284d046a8d21b06e58
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:28258903330a0715141e4685b3760ff18b0d75da0694985bba7c3345d7808b5f
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:90a5eb2c0decb72234cbea5354bb7737cebc53c467229b2c16a1425b4bf408dc
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:3f7e1d83eb1e21860cc2ae0bfe3abe8c9058e92a69698dba9d5871069016e90d
        - registry.redhat.io/ubi9/openjdk-17@sha256:cdbd64ba1b7826d09bc180fa19aa6fbc09ddb5bff5e02b7e1351353bbabebb78
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:a90a6c269b3fc4c77652d4862878211eee5eca1d8047e561d07e3ef1292e52f5
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:31a40f8095a3f8cd04750810d964572339965ae0b2d357c57c0b671a95e1cdbe
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:1f870e70669ad584db1ee31b1dcd822d8b8d0cc03028c7cb09238fb7abf19a42
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:b8cdf2a1873a3df1fb8b8f6d58d06235cef28d228abd70d0bb83cdcfbe3bc230
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:3b1c6f8acdf17a5185413147e0248baa559378ed79e08cb098c724036020576e
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:c9fbb018a70df9a542236ace556fa3343ae350f43c0629f13e525854508f73f9
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:92b8c92eb0bec102215b97f1d60d76c7a57747aeec9594ce1e0e46f7c27391d6
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:5f39f9e61ddbce92629e6502bbef737365c8a1bb2dc2ecaf9f9d8092dea844c9
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:aa1b6f374643ed4e89ba194afde8a5138e401c8cdeb45a6bb5df5b608f58358b
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:22095c436562ff92771d9349f57d007d82415dae519e3da1de7d51e3f1042a93
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:198cf9e5c9641757a0453dec2747a24b018a9b85e10eb0dfdc7a2e75ed09cf6d
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:ba596c629fe8b5b70db97923ff11392dc12115c8e7afefee0d48c12440b8dd86
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:10c620aa3a9c4bfbca185fb460e2521473a5fe6d518342b89a3dada5cf14e65a
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:13608cfe2b87535776c65728aa40cfaa84ccf63afe64d1f2e799ea0e26819239
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:68d9007c5abb2ad3b7666249a61d9428c2cf07165bd8090a6e6def0b7cf70fa3
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:37e2f1fb8a2b4c7ac3d5390ba68a9bf98a0c83a38567b6a9778764061b3d2bab
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:c3ba6b9652205407cc4d4d47b4a928fe8fa86ce4427e6b44ee1d53faf3455f21
        - registry.redhat.io/ubi9/ubi-minimal@sha256:14f14e03d68f7fd5f2b18a13478b6b127c341b346c86b6e0b886ed2b7573b8e0
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:d21cc7e094aa46ba7f6ea717a3d7927da489024a46a6c1224c0b3c5834dcb7a6
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:2d2499b7b12aace07c4b5ab4121caad445aab5bc8dd73377171683b36d417d73
        version: 1.19.3
      entries:
      - name: openshift-pipelines-operator-rh.v1.19.3
        version: 1.19.3
      - name: openshift-pipelines-operator-rh.v1.19.2
        version: 1.19.2
      - name: openshift-pipelines-operator-rh.v1.19.1
        version: 1.19.1
      - name: openshift-pipelines-operator-rh.v1.19.0
        version: 1.19.0
      name: pipelines-1.19
    - currentCSV: openshift-pipelines-operator-rh.v1.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    },
                    {
                      "name": "legacyPipelineRbac",
                      "value": "true"
                    },
                    {
                      "name": "createCABundleConfigMaps",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.19.0 <1.20.0'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.3.2
          - Tekton Triggers: v0.33.0
          - Pipelines as Code: v0.37.0
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.22.3
          - Tekton Results (tech-preview): v0.16.0
          - Manual Approval Gate (tech-preview): v0.7.0
          - Tekton Pruner (tech-preview): v0.2.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.20/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:bc403f93ef6c5526fa30152eab384c4a25bb0bc6ab4158a1e24f637b87a57259
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:980ed7e81ea55522e1681dc1e9b170b68d16f07061dd1f1ffd142367d59552f1
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:bc10e7d6122ede5cf4f2882c200447683255465a1080e826761e90bfb983506e
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:c3013c347bca8a520443a44ac2817fbba5744a9e91069ac5f09e91447ebf6b28
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:1b6aecfce45ed3163227b845b212e83385cdbe999edfc2e9387a0c76f3f8d8e9
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:5683321a5376e63c3ff3932270a1850d5e261f76212020ce867334166f90d2f7
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/ubi9/ubi-minimal@sha256:2f06ae0e6d3d9c4f610d32c480338eef474867f435d8d28625f2985e8acde6e8
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:7eafffa6e4e04f10c1d4e84e02d51c00eb8649c60f8f9a13da14492f24634891
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:d56d6b46331b4569077ab1e7a7b609ba01f0a1dd62b695d62f8aa0f32ea1b621
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:e63298d07ccec5804dfab4e42bcc8163c29cb3ac371c91248777423a4edd31ce
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:8f0494341244a75546a4b80b065702be04e6d200421aa4e7ec90385c9dabc6ab
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:c31af237afb84214bf2a4c06a7636ca071fe97203b41da90d7878109ccd64784
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:115eea1a5ca87f3b6b72db1df05b7386bf73dc70c9c2e4dc5e1a66700d960c37
        - registry.redhat.io/rhel9/skopeo@sha256:a65a413f8a2864389a09dc750690d97afbcdc5c70821e0f85e99e8adba7954e0
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:dd941acd357ae2c0fd7e15fb8d32d7ee6b2847435101dd1e1f107db48af18bf3
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:5021158a4752a26f8b33adea7da66695b22d2ad155d346d35dde5f971bc1a55b
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:823e018e0329d764ac694f177ce88af9074f364c29a1c17929f9dd0b596d03dd
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:14dab8d87323fc7d112213f7f4db3ee7cf3841747de669ce710ae6a794053457
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:bbd1918660be7b3c78777dcccc87d3b26bf5adfc077844ceffbbdd53696dc0a3
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:a7163a9e1fe891de021b1003558e6e64fe17ed17ab0820ba16bba09aa112401e
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e7998e113f75e8f29602bae210c8b66ed06634b699f1d99c5875ef879220d733
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:ec74a562c18fefe36ec7ae3ba083e4eb76c56b562f916578900f8767ae98e6c2
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:c4d05e546c466c57379fff502816012f615be073eb8cdce54677b4ab206a3f5c
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:c0f438e6095aa81f3f7d820614e0414cc43374014c1505d7641bc1dba69a3903
        - registry.redhat.io/rhel9/buildah@sha256:18453ab0a62154283aaf4c45efd7543d1c8ecca1aebae46dedc7ddc0f410b232
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:818bfa4a3a874974b5f7f8bcf2faeb6c05e1795381ee1ceacf89f17b1ec0ef7f
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:9415e889ef98df6d2b2dca1e5d798fb765eeea0d9828e48424374d5267a6b0d1
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:a712bb411d2433b01793dff25785c1e265fa865a52f7c8732c9fb4418ab61757
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:3cf990fb2fc422f3527a7c8b43d06ccd4d32bb5487ec8cb2d4a97f0f5d13f170
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:458f865f1b246c5f0682f5956795771a0515b89edfcffbbc0da1a3d3f532645e
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:96db721939b22572d7d8d8b95dcb5c92fdedb98e9cd0aabb7d17b703a915b07f
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:036250bcdc6fb95273bf07948dfdd4dce765ab84c803ceca0d0b4abf6ca19dec
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:161df929e2bd4c5acff594be31d5458365de5ae0d2189bb2f647c3fc6657fa28
        - registry.redhat.io/ubi9/openjdk-17@sha256:555c8bcd2596637befd1a6da4aa949c5a2640cf66e09c2dfb932afe37e39095e
        - registry.redhat.io/openshift-pipelines/pipelines-opc-rhel9@sha256:c1bee4a2bb495de6d4da962b6e7b2447caa51cc36c3005e4083bd31420ef5102
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:901b5108ceb669c985e75f5fb033c496e9f028863df99ab32651d1b08433a352
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:df15eb1f1f5f237bfe481c16ab3d40bf7fb4123f3c930bcfb121e2cc8dc42882
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:441e1c7124c1674e3d4c38c3e82a4fe5568e76d13c89c04bf009d6b97435de09
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:e3e64abef53151f6b15f15bcbff85668167a571e98badbd8a2ca252ff3a5b8dc
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:2d496752b19f765f1ea36ec94b873ebd8fa604764a56aa67d3ed94b33e5fe75c
        version: 1.20.0
      entries:
      - name: openshift-pipelines-operator-rh.v1.20.0
        version: 1.20.0
      name: pipelines-1.20
    defaultChannel: latest
    packageName: openshift-pipelines-operator-rh
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataTrucker.IO
      provider-url: https://www.datatrucker.io/
    name: datatrucker-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datatrucker-operator.v2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datatrucker.datatrucker.io/v1",
                "kind": "DatatruckerConfig",
                "metadata": {
                  "name": "datatruckerconfig-sample"
                },
                "spec": {
                  "API": {
                    "crypto": {
                      "CredentailEncryption": {
                        "Algorithm": "aes-256-cbc",
                        "IV": "eThWmZq4t7w!z$C\u0026",
                        "Key": "JaNdRgUkXp2s5u8x/A?D(G+KbPeShVmY"
                      },
                      "JWT": {
                        "authorization_keys": {
                          "private_key": "privateJWT.key",
                          "public_key": "publicJWT.key"
                        },
                        "signOptions": {
                          "algorithm": "RS256",
                          "expiresIn": "60m",
                          "issuer": "datatrucker"
                        },
                        "verify": {
                          "issuer": "datatrucker"
                        }
                      },
                      "keycloak": {
                        "client_id": "nodejs",
                        "client_secret": "7c5542ae-7264-420d-bf5f-766f4ab39ad3",
                        "jwt_publickey_verifier": "keycloak.key",
                        "realm": "TestRealm",
                        "url": "http://localhost:30015"
                      },
                      "passwordHash": {
                        "Algorithm": "sha256",
                        "SaltLength": 16
                      }
                    },
                    "cryptokeys": {
                      "keycloakkey": "LS0tLS1CRUdJTiBQVUJMSUMgS0VZLS0tLS0KTUlJQklqQU5CZ2txaGtpRzl3MEJBUUVGQUFPQ0FROEFNSUlCQ2dLQ0FRRUF4ZnhGY1gxb3RmTUN6SWlkZ09tMndZakZWZXZHN1BQT1h5dTNrTEFuTWtTVkUxUXl4TzJ4SFFvSGcydDFVcUU4SVNXeTN3ZkZTbERzdU1VK0J1VVkvaWI5L04vSlpBV3k5MGRET0YxREdIeGVwM21HMVBTUllMQWFGQ0I3NTVsbWZjL1dna3VzZFdLdmxOTmZuNkd6S1J6OXBmV3RNUnhWWkFrejhmNlVUeG0zWE02Q0tOMUludThubEcxU3ZzaFc5cmt3Lyt6UmVvOFVCV2pPcTh5RjdPV0tnS0ZZZlhENFREa090OWFxN3RaaSsxaExYOE5iV2xxRnJZYmhJcFZINXlLSko3ZTA5cUVIRkVEelg5QXJzVVNHamNjbFFFYTl1dmJxU2d4Rzc3RUZVbk82aDZxUkNnU3lNNnVYdk9qcWlIMHp4RTNmRHB4R2NhY2pvTW1EMlFJREFRQUIKLS0tLS1FTkQgUFVCTElDIEtFWS0tLS0t",
                      "privateJWTkey": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQ0KTUlJRW9RSUJBQUtDQVFCVWd1ZDN1OFlXcUJCWWtYZ3M2K25pMS84bnM5YjRGNkZudzA4RDRKdGRlalg5L1ZaMg0KWFdOamJ1ekMxUkRHMnVseFM4QjZIdTZYeEhYWU9XU2ZzOEY0UHZ2eC9TUWdmRWtWUlNDL1pQN3pzM1ZETllaMA0Kc1JScDNSZzUza29qUG80T3o2azFaeGprbFRFZEpMWUtqYlZud3U1M3A3MTFsVTJkV3B4R3VEeGYxVTJuUWgrWA0KaUhKZkhyUmxTZVA1aVhYQkpxd21VTDFkK25lTU5BOXlGakR0MnNZMG5iSDRySDVxbk1VRk4xMEpsZ3V4NFdiUw0KaGFUNklNN3h1dUVHbDRDckVENmdTejBNNGF1UVk4ekZxQWxJNlFWUTlML28zQnBCRmxYUUMwWmowb3lwQVFKLw0KQ1VLaU9FcEhLRHgram02RU9KOE1aejlMQ01yTmdLME5KSWlaQWdNQkFBRUNnZ0VBRTBMMnNLSjZob1Z0bkZJag0KZXRHV01CNmN5QmJFSXkxM215ZTNwYUdXcDVxSUhjaVU0VXR0c0tjQ2pYRjBibTVNa0liOGFwa01Oa2tQdHVWMQ0KMnJSa254eHozZmFkT2s0NWRnQ1ljeWI0dU9DcEZGVnB2TVR6YXlGNk9RNUNCWUJuYmxiRU1TaTdVclh5U0xzcQ0KUm9OLzJoTml1S0RobktBWDNZcGVYZnYxdnIyaWIrMGtiRDVvMVZLcnB6SlJ5emlsWHFwQWJwNDQ4RExFZHVGeA0KTnJKc0tWb1dXMVhmS1FvR2VDMXZqNFJudzdhWkNyU1BjMEhvMUdKb29xSXlrNjRXZnhGWjhpVElsWDJMU1gzbQ0KQm1pN1ZXZnVJK0QvUXRtWEVpN05oTE1jUmExVXB4QkdiK1ZPZXlucEVyMEpxM3VZSmlva1VIZzloOGIzSDM1Yg0KK0lJaEFRS0JnUUNtbzUrRUNpOEJUcFRHdlpqQXBtdmxGZ055NkpWd0RvVHJLSGdCcHJHVHpsVGt6Q0tLYzJhTw0KTTMvQmVIVUQ1NVFqcTdUM094eGgzQUk3cG9kT0tCWS84eXgzenlYUExQUmVCMlR0eTNVQjE5YU5CclhNUGQ5RQ0KTG5hbUQ0SmZQeTV0WHpWWEloWVB0UzBiNjdGbjZnZFl4MlNBbGFYYWFjejU4eURLZnlZL3VRS0JnUUNCMUxNYg0KTFZNTU1Pd1U4R2xsVVB1V0Q5VnBsMnlURytCN2Nnd0ZiYTlrWVlVSFQ4TmlvaUdxQzlLZWxqeklKNnlUdyt3aA0KQisvNGNJU1o5Y0JDNHB0VnRlTWJwUkg3UTQwK20xU3M1SmQzU1VFcDBMOW5VbmdVNDhrMzlpSTV5aEgxU3JTRw0KREtBamVka2Rjb3hNVWFwUHB6cVprOVNKdGVHckVsTHY5WmcvNFFLQmdHcExQTmt3MWNMenFSc01uWDlscUxVOA0KVVg3ZXRWUUJMdkFCcEdzSk5ZU2V2TEFsTm5NeUlWemJwTVo0Vzh2ZDBEVGd0bEp0dGpFRjhiYXpCdVBuMWlsOA0KcUFUVWNiMnhZejhIWUczK3RlZ3pFWHhycDFOY3RDbjlxVmxMMDFHR0w4czhTLzRFbENYNDM0bEwwcUlDT2lWNw0KS3JTS3dIeU51RjVIbzhjSFRCS2hBb0dBVjhVRVpENEs0N1g3RTlqSFY3QWN0UUQxUkpJUzI0ekFBSlJFSGpOZA0KbkIzMVFkRWhoSCtaQkkxRGtPUitLTTN4ZEYvQ05UNVpjM29WbXNPeGUwWVl2eVBjYWViNU5nVWY2cEExK1QvOA0KU2E4ak4zS2d3dWZVdUVLWjJKUXo1UjkvYm5WWGZ6dnYwcXZnSkhFNCthdEgyTnR6MlRsQU9PelZwOElXZ1pSbA0KbmVFQ2dZQXpVS0RLS1ZKaXBYdC9TRC8xd2ZzWElCazFxYWpmdENHOFVTcFlGWGp5VzVYMlZ5bTlVTldmVEdJcA0KQlJkVldXamdFbWhyOU5Da1UrVEU4WEVjTXJNRnhWbXBEeTQ4YmxsQWltWFJtOEhNMlhKODQ4TlArTG5GdEtHdA0KTWswT3JuUHdPYitualZGcXdWUllNQ0QzeFN0Z2RnZUppUHFYcWd3SDhjR0M0OGVmZ0E9PQ0KLS0tLS1FTkQgUlNBIFBSSVZBVEUgS0VZLS0tLS0=",
                      "publicJWTkey": "LS0tLS1CRUdJTiBQVUJMSUMgS0VZLS0tLS0NCk1JSUJJVEFOQmdrcWhraUc5dzBCQVFFRkFBT0NBUTRBTUlJQkNRS0NBUUJVZ3VkM3U4WVdxQkJZa1hnczYrbmkNCjEvOG5zOWI0RjZGbncwOEQ0SnRkZWpYOS9WWjJYV05qYnV6QzFSREcydWx4UzhCNkh1Nlh4SFhZT1dTZnM4RjQNClB2dngvU1FnZkVrVlJTQy9aUDd6czNWRE5ZWjBzUlJwM1JnNTNrb2pQbzRPejZrMVp4amtsVEVkSkxZS2piVm4NCnd1NTNwNzExbFUyZFdweEd1RHhmMVUyblFoK1hpSEpmSHJSbFNlUDVpWFhCSnF3bVVMMWQrbmVNTkE5eUZqRHQNCjJzWTBuYkg0ckg1cW5NVUZOMTBKbGd1eDRXYlNoYVQ2SU03eHV1RUdsNENyRUQ2Z1N6ME00YXVRWTh6RnFBbEkNCjZRVlE5TC9vM0JwQkZsWFFDMFpqMG95cEFRSi9DVUtpT0VwSEtEeCtqbTZFT0o4TVp6OUxDTXJOZ0swTkpJaVoNCkFnTUJBQUU9DQotLS0tLUVORCBQVUJMSUMgS0VZLS0tLS0=",
                      "truckerapicrt": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSUQxVENDQXIyZ0F3SUJBZ0lVUFVzZ2NYVXZRYWtuQldjeUFDeEw2dmt3b204d0RRWUpLb1pJaHZjTkFRRUwKQlFBd2RqRUxNQWtHQTFVRUJoTUNRMEV4Q3pBSkJnTlZCQWdNQWs5T01ROHdEUVlEVlFRSERBWk5hV3gwYjI0eApGekFWQmdOVkJBb01Ea1JoZEdGVWNuVmphMlZ5TGtsUE1SUXdFZ1lEVlFRTERBdEZibWRwYm1WbGNtbHVaekVhCk1CZ0dBMVVFQXd3UlJHRjBZVlJ5ZFdOclpYSlNiMjkwUTBFd0hoY05NakF4TURFeU1UWXlOalV5V2hjTk1qTXcKTVRFMU1UWXlOalV5V2pCM01Rc3dDUVlEVlFRR0V3SkRRVEVMTUFrR0ExVUVDQXdDVDA0eER6QU5CZ05WQkFjTQpCazFwYkhSdmJqRVhNQlVHQTFVRUNnd09SR0YwWVZSeWRXTnJaWEl1U1U4eEZEQVNCZ05WQkFzTUMwVnVaMmx1ClpXVnlhVzVuTVJzd0dRWURWUVFEREJKa1lYUmhkSEoxWTJ0bGNpMWhjR2t1YVc4d2dnRWlNQTBHQ1NxR1NJYjMKRFFFQkFRVUFBNElCRHdBd2dnRUtBb0lCQVFEakM4ZmVCVEg4L3NnejFQcTJQYU9QQjc5YWtyeUZvWVpGV292bQpWOXNvMmJNRHp4NVBZSFJydVFGYmJQRzlVVlc2QWZpSHdEb1p2Z3NTVXpyS3ZyR1RlYjluNUhSVlVOMTl0NExTCjdtUFppNzFRbGE4bVR3QmdvTnE4TFppcnZUdzV0S2I4K3NmcFJxQUlUQkVsRmxVUEZISzllcDl2eFdvaHhwdGgKS1dPVkwvV2ZtWldhWXlRdjlLNDBkWUJpRWpGcTFKZVB4MDlnUFdHM2pOT1VYTTZQUVlMaHJDUXFFcy9VME9GWQpwVFRHUmNXZEx4TFVFZVNkcHJKK1ZEc3lEK1lVeUFRR2phYk14VmVTVlRUN1FGbFBjLzc3OVZsRkdXMC8xWVVaCmpqakNsSkFLUWZOTHg5bG85cWJERVBZL3grZmhnRnYyWUVaSTNTdGVyZWUvVDJoWEFnTUJBQUdqV2pCWU1COEcKQTFVZEl3UVlNQmFBRkNJVy9JYkZIUUNBeEdxZC9FVFpXbVhGeGMvQ01Ba0dBMVVkRXdRQ01BQXdDd1lEVlIwUApCQVFEQWdUd01CMEdBMVVkRVFRV01CU0NFbVJoZEdGMGNuVmphMlZ5TFdGd2FTNXBiekFOQmdrcWhraUc5dzBCCkFRc0ZBQU9DQVFFQUxrMlJDbnVweDA5Wmx0SmN3M1pUaFIveUZGNS9KVlVsWkdHTnc3MlAwQmR6SEloQncrVk4KRkNucGFPMjJ0WGF1NlJVeTVNUWhwc3JrbFNKYVdJT25jOUg0Uk0vbGh3YWsyRXN0MkFhdUFwNUdYSVFJMDVBeAo3NmtDZXEyMnUyalNXRWpFS1hGRnRiZXVDZnFpWlpqNWFLcGF1YlZWRG4ydEZTWjJRMFRib2RRN2dpTFdNUTBLCk5ycVgwQWNSU2J0ZnRSNjBhdGQ1YW41aFNXY0dyZUg5ZHdFUVRvVVNFblg2RFRZSURoeElsZStjQ2V5VjVSWk8KVS9hck5RZ0lFY3ZrZUM1d1lZaFFFL1FUN3lNb1lNajFnckxmOTZSZkRhbEVFNE1seUtDWm9wYWNJWUdzSmhuKwpLMWFISEQvbGZxTC9zbnNtY3pXdHlLUncxTDF1WkxPcjJnPT0KLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=",
                      "truckerapikey": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQpNSUlFb3dJQkFBS0NBUUVBNHd2SDNnVXgvUDdJTTlUNnRqMmpqd2UvV3BLOGhhR0dSVnFMNWxmYktObXpBODhlClQyQjBhN2tCVzJ6eHZWRlZ1Z0g0aDhBNkdiNExFbE02eXI2eGszbS9aK1IwVlZEZGZiZUMwdTVqMll1OVVKV3YKSms4QVlLRGF2QzJZcTcwOE9iU20vUHJINlVhZ0NFd1JKUlpWRHhSeXZYcWZiOFZxSWNhYllTbGpsUy8xbjVtVgptbU1rTC9TdU5IV0FZaEl4YXRTWGo4ZFBZRDFodDR6VGxGek9qMEdDNGF3a0toTFAxTkRoV0tVMHhrWEZuUzhTCjFCSGtuYWF5ZmxRN01nL21GTWdFQm8ybXpNVlhrbFUwKzBCWlQzUCsrL1ZaUlJsdFA5V0ZHWTQ0d3BTUUNrSHoKUzhmWmFQYW13eEQyUDhmbjRZQmI5bUJHU04wclhxM252MDlvVndJREFRQUJBb0lCQUZSbFo2aiswMk1RcmdpZQpUWWQ2aUJhOXNZS1FjdVZWY2x6TFdKK0ozdHhqZCtpZGo4Mi9zK2JyQVRjajJ3UWRmTGxFekVaaHpQY09ybnprClZVRlIreDR2cmxlSEV6czNjTzJ2VGtEci9rKzZZdWZXakNHNFVoS1pHRlhFdVRCMWlqS0ZGV0dLTXM5aXdrL3oKUHBoRG9lQ1dJRVNRbFQxdDY1S2NpZHRuRUM0K3oxaURxR21ZZEVEa2xEZFFIVHorRGpseS80VzV2MFpVQWN2Vwo1Unkvd0JldmNURlBXQy9uZGFrN3hGUnVrU2xNZUlhTWhvTlllOU5uVCt2Nmh5VVh1MXp6cG45UEFJRTVpQWpOCnBsZkFEZitwakwrZFJDbW9sWW1LR3NQVjEzSmtCY1p2cEVoUWgrUmJnaWFwczJLc2plQ0NmTmtST0hndHc0MU0KeGxOTU1Ha0NnWUVBOHM1akpuZ2FMblIvTzZtakZEb0wvd1p2UWdnT1h3Z1RBN2tsQ0trbjBYM0xJS0hhaTZESgpyOVRGQktMdkhUazVFK1Y4NkhIZ29VMUFQQnYyMWJpaHhuQXhUYm8wSHZMdG5QOXU3YnZJejFNOWMyd2ZlWUgzCmliMG1FNkV0Y2wvNnVIUHdMVlZtRGh5bllneEZpeTRRRXp1ZHZwaFIvSnlPazQvUytQUzh2ZXNDZ1lFQTcySW8KWDlkY0xvUmYzSk0vRTF1SXhrRGRLZm1VakN6Mk1NN3pyRkpNUERobTFSUlJhQjdidHNxUzl4Mm5KeTI5amtoSQpndTR0Y2s5STdncm9WRXY0cEtkU0lOVzVNVUg0MXJPckxTbFBobCtMaTM5QXNrTERFVXVQbkFmYko3YWZibFFHCjNFWE1Ka1d2WFQxRG90aE11MUpUMTFPMjhmYlBmb3R3YVRucXFFVUNnWUE1RnArblh0dkJDbkVUNVJuTDJxNFIKWTBEcmhKdGthckNWMG9WTG1RSWN0bmY1RHpVa1dVU3lpOU5hK0xoQjViOXUxeWxjT2RJMytzTGNuektWSTVrZAo0Q2tuZmwxYzNYOVN4NWdjVzJlbW40cTJSYmpNeCtzY3ZwVmlUaVl5V3pESEI3Tno1VDFlbWd3YytJZXRac0l1CjAzaU9aSi82UmhFYnp0Z3E0dGd4eXdLQmdCanhwSDFnSm9YZVcxTFRzUlhtU3RsRnR3WUJzTDN1alhUR2k5T2YKKzd1bXNPaHhKbzV2OExzYVhyZ0xTZmltSm9ObCt0QlIyMEIzRElWcGM0NGRJSHFmeU1vTkd6QUx6TThxaDZLOApGRUpQUUc0M3cxa1A1ZDJ1ZnZkbVJpcmNONzY2V01DcE1UMm5zejlHKy81REJiUGJQMHRIdU45N3RXRmRURWhhCnRaMzFBb0dCQU5xZnVycEV4ZHozdXRNUm91RnJjTmVkV2s4VlVpVVZIV1hZMjhNNWQvRjcxMExIOTZuOW12V08Kc1cwcDFjcXBJMEsyQko2WlYwNVpybnlyRHBsVVpVdVpSQWxGR3hYR1dLU2trQzlmWm90SkFOWDBkR29KSitqTwpEYUNLNGxlcXdOMHNPRE1YZFErMlFPaUtlSUxkNU5KcDNUSW1YV1Z0RFlXUkRwOW9iSUdlCi0tLS0tRU5EIFJTQSBQUklWQVRFIEtFWS0tLS0tCg=="
                    },
                    "jobsServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "debug"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": true,
                        "FileSFTP": true,
                        "IOTKafka": true,
                        "IOTProxy": false,
                        "IOTRedis": true,
                        "Login": false,
                        "ScriptJS": true,
                        "ScriptSSH": true,
                        "ScriptShell": true,
                        "UtilEcho": true,
                        "UtilFuzzy": true,
                        "UtilSentiment": true,
                        "credentials": false,
                        "jobs": true,
                        "resources": false,
                        "statuschecks": true,
                        "ui": false,
                        "usermanagement": false
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "loginServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "info"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": false,
                        "FileSFTP": false,
                        "IOTKafka": false,
                        "IOTProxy": false,
                        "IOTRedis": false,
                        "Login": true,
                        "ScriptJS": false,
                        "ScriptSSH": false,
                        "ScriptShell": false,
                        "UtilEcho": false,
                        "UtilFuzzy": false,
                        "UtilSentiment": false,
                        "credentials": false,
                        "jobs": false,
                        "resources": false,
                        "statuschecks": true,
                        "ui": false,
                        "usermanagement": false
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "managementServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "debug"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": false,
                        "FileSFTP": false,
                        "IOTKafka": false,
                        "IOTProxy": false,
                        "IOTRedis": false,
                        "Login": true,
                        "ScriptJS": false,
                        "ScriptSSH": false,
                        "ScriptShell": false,
                        "UtilEcho": false,
                        "UtilFuzzy": false,
                        "UtilSentiment": false,
                        "credentials": true,
                        "jobs": false,
                        "resources": false,
                        "statuschecks": true,
                        "ui": true,
                        "usermanagement": true
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "name": "API",
                    "resource": {
                      "Cache": {
                        "Creds": false,
                        "CredsFlush": 10,
                        "Resource": false,
                        "ResourceFlush": 10
                      },
                      "Templates": {
                        "Scripts": "scripts",
                        "Shell": "shell",
                        "keys": "keys",
                        "resourcedefinitions": "resourcedefinitions"
                      }
                    }
                  },
                  "DB": {
                    "databasename": "userdb",
                    "hostname": "db",
                    "password": "password",
                    "port": 5432,
                    "type": "pg",
                    "user": "testuser"
                  },
                  "TempDB": {
                    "enabled": false,
                    "pvc": "samplepvc"
                  }
                }
              },
              {
                "apiVersion": "datatrucker.datatrucker.io/v1",
                "kind": "DatatruckerFlow",
                "metadata": {
                  "name": "datatruckerflow-sample"
                },
                "spec": {
                  "API": {
                    "Image": {
                      "imageName": "datatrucker-server/datatrucker-api",
                      "imageNameUI": "datatruckerio/datatrucker-ui",
                      "repository": "registry.gitlab.com/datatruckerio",
                      "tagName": "latest"
                    },
                    "name": "api"
                  },
                  "DatatruckerConfig": "datatruckerconfig-sample",
                  "JobDefinitions": [
                    {
                      "name": "echoapi",
                      "restmethod": "GET",
                      "tenant": "Admin",
                      "type": "Util-Echo",
                      "validations": {
                        "properties": {
                          "voila": {
                            "enum": [
                              "true",
                              "false"
                            ],
                            "type": "string"
                          }
                        },
                        "type": "object"
                      }
                    }
                  ],
                  "Keys": {
                    "configmap": "placeholder"
                  },
                  "Replicas": 1,
                  "Resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "256Mi"
                    },
                    "requests": {
                      "cpu": "250m",
                      "memory": "256Mi"
                    }
                  },
                  "Scripts": {
                    "configmap": "placeholder"
                  },
                  "Type": "Job"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/datatruckerio/datatrucker-operator:2.0.0
          createdAt: "2021-09-08 01:01:01"
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: hhttps://gitlab.com/datatruckerio/datatrucker-server.git
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              - creates secrets to hold server configs
              - create tempDB for protoypting
              - create configs for seperating Login, management and Jobs endpoints
              - Create crypto configs
            displayName: Datatrucker Config
            kind: DatatruckerConfig
            name: datatruckerconfigs.datatrucker.datatrucker.io
            version: v1
          - description: '- creates an API as per Job Definitions'
            displayName: Datatrucker Flow
            kind: DatatruckerFlow
            name: datatruckerflows.datatrucker.datatrucker.io
            version: v1
        description: "## DataTrucker.IO\n\nDataTrucker.IO is a simple no-code / less-code
          API Backend\n\n### What does it do ?\n\nDatatrucker.IO is a product capable
          of reading simple json/yaml configs and building th code necessary for converting
          it into an API.\nAlong with building the code, it will also host the code
          base on a nodejs server , i.e. it will immediately make it available for
          consumption.\n\n### What does it offload from a developer ?\n\nDataTrucker
          is capable of removing the most common activities which a developer needs
          to do on every new project A few of such common activities are\n\n- Creating
          an API endpoint with a specified business logic (using simple plugins)\n-
          Applying standard RBAC\n- Applying Authorization logic\n- Applying hardening
          on endpoints\n- Log management\n- Connecting to variety of systems\n- Modularizing
          business logic\n- The best of doing it with little to no code\n\n### watch
          the quick elevator pitch to understand datatrucker.io\n\n https://youtu.be/DKLFDqhjs5M\n "
        displayName: DataTrucker.IO
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - datatrucker
        - operator
        - nocode
        - apibuilder
        links:
        - name: Website
          url: https://www.datatrucker.io/
        - name: Git Repo
          url: https://gitlab.com/datatruckerio/datatrucker-server.git
        maintainers:
        - email: gauravshankar.can@gmail.com
          name: Gaurav Shankar
        maturity: alpha
        provider:
          name: DataTrucker.IO
          url: https://www.datatrucker.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - docker.io/datatruckerio/datatrucker-operator:2.0.0
        version: 2.0.0
      entries:
      - name: datatrucker-operator.v2.0.0
        version: 2.0.0
      - name: datatrucker-operator.v1.3.2
        version: 1.3.2
      - name: datatrucker-operator.v1.3.1
        version: 1.3.1
      - name: datatrucker-operator.v1.3.0
        version: 1.3.0
      name: alpha
    defaultChannel: alpha
    packageName: datatrucker-operator
    provider:
      name: DataTrucker.IO
      url: https://www.datatrucker.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Syndesis Team
      provider-url: https://github.com/syndesisio/syndesis
    name: syndesis
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: syndesis.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta2",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.12
          createdAt: "2021-08-11T12:00:00Z"
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta2
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.12
        - centos/postgresql-10-centos7
        version: 1.12.0
      entries:
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: 1.12.x
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.13
        - centos/postgresql-10-centos7
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: 1.13.x
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - centos/postgresql-10-centos7
        - quay.io/syndesis/syndesis-operator:1.13
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: candidate
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.13
        - centos/postgresql-10-centos7
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: latest
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.13
        - centos/postgresql-10-centos7
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: stable
    defaultChannel: latest
    packageName: syndesis
    provider:
      name: Syndesis Team
      url: https://github.com/syndesisio/syndesis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.22.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.5"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.22
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.22
        version: 42.22.0
      entries:
      - name: t8c-operator.v42.22.0
        version: 42.22.0
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: stable
    - currentCSV: t8c-operator.v42.10.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.10
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.10
        version: 42.10.0
      entries:
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.10.0
    - currentCSV: t8c-operator.v42.10.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.10
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.10
        version: 42.10.0
      entries:
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.12.0
    - currentCSV: t8c-operator.v42.13.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.7"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.13
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.13
        version: 42.13.0
      entries:
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.13.0
    - currentCSV: t8c-operator.v42.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.7"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.14
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.14
        version: 42.14.0
      entries:
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.14.0
    - currentCSV: t8c-operator.v42.15.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.6.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.15
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.15
        version: 42.15.0
      entries:
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.15.0
    - currentCSV: t8c-operator.v42.16.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.16
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.16
        version: 42.16.0
      entries:
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.16.0
    - currentCSV: t8c-operator.v42.17.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.6.6"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.17
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.17
        version: 42.17.0
      entries:
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.17.0
    - currentCSV: t8c-operator.v42.18.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.1"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.18
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.18
        version: 42.18.0
      entries:
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.18.0
    - currentCSV: t8c-operator.v42.19.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.2"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.19
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.19
        version: 42.19.0
      entries:
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.19.0
    - currentCSV: t8c-operator.v42.20.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.20
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.20
        version: 42.20.0
      entries:
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.20.0
    - currentCSV: t8c-operator.v42.21.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.21
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.21
        version: 42.21.0
      entries:
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.21.0
    - currentCSV: t8c-operator.v42.22.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.5"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.22
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.22
        version: 42.22.0
      entries:
      - name: t8c-operator.v42.22.0
        version: 42.22.0
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.22.0
    - currentCSV: t8c-operator.v42.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.4.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.6
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.6
        version: 42.6.0
      entries:
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.6.0
    - currentCSV: t8c-operator.v42.7.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.4.6"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.7
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.7
        version: 42.7.0
      entries:
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.7.0
    - currentCSV: t8c-operator.v42.8.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.0"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.8
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.8
        version: 42.8.0
      entries:
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.8.0
    - currentCSV: t8c-operator.v42.9.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.1"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.9
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.9
        version: 42.9.0
      entries:
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.9.0
    defaultChannel: stable
    packageName: t8c
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-db
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-db.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Db",
                "metadata": {
                  "name": "db-sample"
                },
                "spec": {
                  "clusterDomain": "cluster.local",
                  "databases": [
                    {
                      "name": "wapdb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "grafanadb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "brokerdb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "nimbusmosaic",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    }
                  ],
                  "image": {
                    "migrationJobImage": "wap-ose-cli",
                    "migrationJobRepository": "prod.harbor.keysight.digital/loadcore",
                    "migrationJobTag": "v4.13.0-3747-35-ocp",
                    "pgImage": "postgresql-ubi-image",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "pgTag": "14.2.0-3747-32"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "postgresql": {
                    "image": {
                      "pullSecrets": [
                        "load-core-secret"
                      ],
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "postgresql-ubi-image",
                      "tag": "14.2.0-3747-32"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "pgHbaConfiguration": "local all all trust\nhost all all 0.0.0.0/0 trust",
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "storageClass": "managed-loadcore-storage"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "volumePermissions": {
                        "containerSecurityContext": {
                          "runAsUser": "auto"
                        }
                      }
                    },
                    "serviceAccount": {
                      "automountServiceAccountToken": false,
                      "name": "load-core-service-account"
                    }
                  },
                  "storage": "8Gi"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:30:46Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Db
            name: dbs.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP DB Operator
        displayName: Keysight WAP DB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Db
          url: https://keysight-wap-db.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-db@sha256:5689d6a4da2fdd97223bc64aa61a8574bba810f93d09ab804cb45b4475f80ff5
        version: 0.1.21
      entries:
      - name: keysight-wap-db.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-db
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-certified-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "tag": "5.4.6",
                  "oem": "",
                  "imagePullSecrets": "",
                  "psp": false,
                  "rbac": true,
                  "serviceAccount": "default",
                  "leastPrivilege": false,
                  "global": {
                    "cattle": {
                      "url": ""
                    }
                  },
                  "bootstrapPassword": "",
                  "autoGenerateCert": true,
                  "defaultValidityPeriod": 365,
                  "internal": {
                    "certmanager": {
                      "enabled": false,
                      "secretname": "neuvector-internal"
                    },
                    "autoGenerateCert": true,
                    "autoRotateCert": true
                  },
                  "controller": {
                    "enabled": true,
                    "annotations": {},
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": "registry.connect.redhat.com/neuvector/controller@sha256:65ee3723c7c007db5ce3cd0493a785d4163d972cf957de6dec956b842807a551",
                    "imagePullPolicy": "IfNotPresent",
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "searchRegistries": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "nodePort": "",
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      }
                    },
                    "ranchersso": {
                      "enabled": false
                    },
                    "pvc": {
                      "enabled": false,
                      "existingClaim": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {
                        "userinitcfg.yaml": {
                          "users": [
                            {
                              "Fullname": "admin",
                              "Password": "",
                              "Role": "admin"
                            }
                          ]
                        }
                      }
                    },
                    "certupgrader": {
                      "env": [],
                      "schedule": "",
                      "imagePullPolicy": "IfNotPresent",
                      "timeout": 3600,
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "prime": {
                      "enabled": false,
                      "image": {
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "1.0.7"
                      }
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/enforcer@sha256:72a628a5c811c95bb267bbd0a5ffc9d74dbaca03a92aae615d43c7601b4d3b43",
                    "imagePullPolicy": "IfNotPresent",
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "env": [],
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/etcd"
                      }
                    ],
                    "resources": {},
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "manager": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/manager@sha256:89a42c9ee86f98ef56b5e695e23bdcc6334361ebd29cfb7729d7c86f04da98b0",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "",
                    "env": {
                      "ssl": true,
                      "envs": []
                    },
                    "svc": {
                      "type": "ClusterIP",
                      "nodePort": "",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": "",
                      "tls": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "topologySpreadConstraints": [],
                    "affinity": {},
                    "podLabels": {},
                    "podAnnotations": {},
                    "tolerations": [],
                    "nodeSelector": {},
                    "runAsUser": "",
                    "probes": {
                      "enabled": false,
                      "timeout": 1,
                      "periodSeconds": 10,
                      "startupFailureThreshold": 30
                    }
                  },
                  "cve": {
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "cacert": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                      "image": "registry.access.redhat.com/ubi8",
                      "imagePullPolicy": "IfNotPresent",
                      "schedule": "0 0 * * *",
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": "registry.connect.redhat.com/neuvector/scanner@sha256:b632f1d31998a4c8bad9f97216e8aa56cd7e6446bd70d2089d6c66845b964e8d",
                      "imagePullPolicy": "Always",
                      "priorityClassName": "",
                      "resources": {},
                      "topologySpreadConstraints": [],
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      },
                      "volumes": "",
                      "volumeMounts": ""
                    }
                  },
                  "resources": {},
                  "runtimePath": "",
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhooksvc": {
                    "enabled": true
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  },
                  "lease": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "true"
          containerImage: registry.connect.redhat.com/neuvector/neuvector-operator@sha256:e2f19c50ebc4127a83c943dbcfa1312ec46dec71bd36f6d99178b320d8da970b
          createdAt: "2020-12-02 14:19:21"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: Neuvector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm/tree/master/charts/core) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart.  NeuVector Operator versions are tied to NeuVector product versions. Version 1.5.2 of the NeuVector Certified Operator deploys version 5.4.6 of NeuVector.

          **Complete below steps to Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Privileged SCC is added to Service Account specified in the deploment yaml by Operator version 1.3.6 and above in new deployment. In case of upgrading NeuVector Operator from previous version to 1.3.6, please delete Privileged SCC before upgrading.

                   oc delete rolebinding -nneuvector system:openshift:scc:privileged

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s                    neuvector/default


          #Deploying the NeuVector Operator#


          Please refer to the instructions [here](https://github.com/neuvector/neuvector-operator/blob/master/README.md)


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Networking
        - Security
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - registry.connect.redhat.com/neuvector/manager@sha256:89a42c9ee86f98ef56b5e695e23bdcc6334361ebd29cfb7729d7c86f04da98b0
        - registry.connect.redhat.com/neuvector/scanner@sha256:b632f1d31998a4c8bad9f97216e8aa56cd7e6446bd70d2089d6c66845b964e8d
        - registry.access.redhat.com/ubi8@sha256:534c2c0efa4150ede18e3f9d7480d3b9ec2a52e62bc91cd54e08ee7336819619
        - registry.connect.redhat.com/neuvector/neuvector-operator@sha256:e2f19c50ebc4127a83c943dbcfa1312ec46dec71bd36f6d99178b320d8da970b
        - registry.connect.redhat.com/neuvector/controller@sha256:65ee3723c7c007db5ce3cd0493a785d4163d972cf957de6dec956b842807a551
        - registry.connect.redhat.com/neuvector/enforcer@sha256:72a628a5c811c95bb267bbd0a5ffc9d74dbaca03a92aae615d43c7601b4d3b43
        version: 1.5.2
      entries:
      - name: neuvector-operator.v1.5.2
        version: 1.5.2
      - name: neuvector-operator.v1.5.1
        version: 1.5.1
      - name: neuvector-operator.v1.5.0
        version: 1.5.0
      - name: neuvector-operator.v1.4.9
        version: 1.4.9
      - name: neuvector-operator.v1.4.8
        version: 1.4.8
      - name: neuvector-operator.v1.4.7
        version: 1.4.7
      - name: neuvector-operator.v1.4.6
        version: 1.4.6
      - name: neuvector-operator.v1.4.5
        version: 1.4.5
      - name: neuvector-operator.v1.4.4
        version: 1.4.4
      - name: neuvector-operator.v1.4.3
        version: 1.4.3
      - name: neuvector-operator.v1.4.2
        version: 1.4.2
      - name: neuvector-operator.v1.4.1
        version: 1.4.1
      - name: neuvector-operator.v1.4.0
        version: 1.4.0
      - name: neuvector-operator.v1.3.9
        version: 1.3.9
      - name: neuvector-operator.v1.3.8
        version: 1.3.8
      - name: neuvector-operator.v1.3.7
        version: 1.3.7
      - name: neuvector-operator.v1.3.6
        version: 1.3.6
      - name: neuvector-operator.v1.3.5
        version: 1.3.5
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-certified-operator
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: model-builder-for-vision-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-modelbuilder-for-vision.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "Modelbuilder",
                "metadata": {
                  "name": "modelbuilder-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "backup_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "metadata_storage": {
                    "storage_class": "ibmc-block-bronze",
                    "storage_size": "20G"
                  },
                  "photo_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "500G"
                  },
                  "in_memory_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              },
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "ModelbuilderAWS",
                "metadata": {
                  "name": "modelbuilder-aws-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "postgres": {
                    "storage_size": "20G",
                    "storage_class": "gp2",
                    "backup_storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
          createdAt: "2021-08-28T12:00:00Z"
          olm.skipRange: <1.0.3
          operators.operatorframework.io/builder: operator-sdk-v1.11.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/ibm-modelbuilder-for-vision/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Model Builder for Vision on AWS
            displayName: Deploy-Modelbuilder on AWS
            kind: ModelbuilderAWS
            name: modelbuildersaws.modelbuilder.com
            version: v1alpha1
          - description: Deploy Model Builder for Vision on IBM
            displayName: Deploy-Modelbuilder on IBM
            kind: Modelbuilder
            name: modelbuilders.modelbuilder.com
            version: v1alpha1
        description: IBM Services Software Model Builder for Vision is an AI training
          platform that can use the capabilities of IBM Cloud and Amazon Web Services
          and their GPUs to quickly train computer vision models for compatible mobile
          apps. IBM Services Software Inspection Workbench iOS and iPadOS app will
          be used as the exclusive UI for labeling and training. Specifically, IBM
          Model Builder for Vision will train computer vision models that can then
          be deployed for use on the IBM Services Software Inspector Portable , IBM
          Services Software Inspector Wearable and the IBM Maximo Visual Inspection
          Mobile apps.
        displayName: IBM Model Builder for Vision
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Model Builder
        - AI
        - Computer Vision
        - Inspection
        - Mobile
        - Core ML
        - TensorFlow Lite
        links:
        - name: IBM Model Builder for Vision
          url: http://ibm.biz/model_builder
        maintainers:
        - email: isshelp@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:2f65f46b25d84a3cb41a6a10f93786bcab47b7354ca9ac13a8e602ea833eba60
        - registry.connect.redhat.com/ibm/modelbuilder-image-server@sha256:ecdb665af651417c50335e7259b89ef8096521abec431976bd7abe4c1edc49d6
        - registry.connect.redhat.com/ibm/modelbuilder-cv-training-service@sha256:c3c9080b08b87022c784fd98ac3e2774ec89ce29a2dd19d91dee039e3fdb2b36
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d4c60531f59953f399c6a2b0f8239128cdc2640f72f2d1e9ac4d6add618f9f01
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.connect.redhat.com/ibm/modelbuilder-augmentor@sha256:ef5d0dde3eb334d919c2773af63007f1ef5e735ae040ef77fffe3440f854f068
        - registry.connect.redhat.com/ibm/modelbuilder-auth-service@sha256:f96b444fdcf7781775c9eba942bcc5e5ebdcf5f50820447f5eb26061c4c674cd
        - registry.connect.redhat.com/ibm/modelbuilder-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
        version: 1.0.3
      entries:
      - name: ibm-modelbuilder-for-vision.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: model-builder-for-vision-certified
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Coroot Inc.
      provider-url: https://coroot.com
    name: coroot-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: coroot-operator.1.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "coroot.com/v1",
                "kind": "Coroot",
                "metadata": {
                  "name": "coroot",
                  "namespace": "coroot"
                },
                "spec": {
                  "enterpriseEdition": {
                    "licenseKey": "<COROOT_LICENSE_KEY> from https://coroot.com/account"
                  },
                  "ingress": {
                    "host": "coroot.example.com"
                  },
                  "storage": {
                    "size": "50Gi"
                  },
                  "clickhouse": {
                    "replicas": 2,
                    "shards": 2,
                    "storage": {
                      "size": "100Gi"
                    }
                  },
                  "prometheus": {
                    "storage": {
                      "size": "100Gi"
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring, Logging & Tracing
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/coroot/coroot-operator
          support: support@coroot.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Coroot is the Schema for the coroots API
            displayName: Coroot
            kind: Coroot
            name: coroots.coroot.com
            version: v1
        description: An open-source observability platform using eBPF to gather telemetry
          data, providing actionable insights for quick issue detection and resolution
          in Kubernetes and VM environments.
        displayName: Coroot
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - coroot
        - observability
        links:
        - name: Documentation
          url: https://docs.coroot.com
        maintainers:
        - email: support@coroot.com
          name: Coroot Inc.
        maturity: stable
        provider:
          name: Coroot Inc.
          url: https://coroot.com
        relatedImages:
        - ghcr.io/coroot/kube-state-metrics@sha256:080a85e77b1e530df56f84c8e0ee2523c1f740adea152015367b87edbfac078d
        - ghcr.io/coroot/coroot-ee@sha256:09b9177ac07c2d26f4bd16050d07bb1e2132431055c32e047563083ee09194a8
        - ghcr.io/coroot/coroot-node-agent@sha256:3d552e98ca2ff8c6e125496857306172ec6f8e134ff13886c7f799abef29d1b8
        - ghcr.io/coroot/coroot-cluster-agent@sha256:674bcb147354873eaa0ee9c1328ddf758d43adbbe532c9fbbe41ddae7addc114
        - ghcr.io/coroot/clickhouse@sha256:afed1054626598371975c948e5b9e4dcaac54e8059b7ad9f0b905ef1c99f2260
        - ghcr.io/coroot/coroot-operator@sha256:13f1a4a07877c8fe8230a0bb7282ff47451007c9cf689a995a37d16697b9268a
        - ghcr.io/coroot/prometheus@sha256:e7117bd86e1088fc02ac12758bc94dba1029de54a5bc1ee04954569d8b79625d
        version: 1.1.5
      entries:
      - name: coroot-operator.1.1.5
        version: 1.1.5
      name: stable
    defaultChannel: stable
    packageName: coroot-operator
    provider:
      name: Coroot Inc.
      url: https://coroot.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: cpd-platform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cpd-platform-operator.v6.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cpd.ibm.com/v1",
                "kind": "Ibmcpd",
                "metadata": {
                  "name": "ibmcpd-sample"
                },
                "spec": {
                  "license": {
                    "accept": false,
                    "license": "EE"
                  },
                  "storageClass": "managed-nfs-storage"
                }
              }
            ]
          capabilities: Seamless Upgrades
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.0 <6.1.3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ibmcpd
            name: ibmcpds.cpd.ibm.com
            version: v1
          required:
          - kind: CommonService
            name: commonservices.operator.ibm.com
            version: v3
        description: "IBM Cloud Pak for Data is a cloud-native solution that enables
          you to put your data to work quickly and efficiently. \nCloud Pak for Data
          includes a suite of services that support you on your journey to AI.\n#
          Introduction\nWith Cloud Pak for Data, your users can collaborate from a
          single, unified interface that supports a large number of \nservices that
          are designed to work together. You can use the Cloud Pak for Data Operator
          to quickly provision services for \nimmediate use. These are just a few
          of the services that are available in Cloud Pak for Data:     \n- Watson
          Studio\n- Watson Machine Learning\n- Watson Knowledge Catalog\n- Watson
          OpenScale\n- Watson Machine Learning Accelerator\n- Data Virtualization\n-
          Db2 Data Management Console\n- Db2 Warehouse\n- Cognos® Dashboards\n- Analytics
          Engine Powered by Apache Spark\n- IBM Match 360 with Watson\n- Data Refinery\n-
          Db2® Big SQL\n- Decision Optimization\n- SPSS® Modeler\n- Db2 Data Gate\n\n#
          Details\n\n## Prerequisites\n- Red Hat OpenShift Container Platform 4.6
          or newer installed on x86_64\n- A user with cluster administrator role\n-
          No storage is required to install the operator, but you need storage and
          resources to install Cloud Pak for Data services. For more information,
          see [Storage considerations](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=planning-storage-considerations).\n-
          See [Pre-installation](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-pre-installation-tasks)\n-
          To install the Cloud Pak for Data control plane, review the [system\nrequirements
          for IBM Cloud Pak for Data](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=planning-system-requirements).\n-
          Before you install Cloud Pak for Data services, review the [system requirements\nfor
          services](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=requirements-system-services).\n-
          For air-gapped installations, see [Installing the Cloud Pak for Data Operator
          on air-gapped clusters](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=tasks-mirroring-images-your-container-registry).\n##
          Resources Required\n- The operator requires 100m CPU and 256Mi memory.\n-
          For the minimum requirements for services, see the [system requirements
          for services](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=requirements-system-services).\n##
          Installing\n- [Installing Cloud Pak for Data Control Plane] (https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-cloud-pak-data)\n-
          [Installing Cloud Pak for Data Service Subscriptions](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=tasks-creating-operator-subscriptions)\n-
          [Installing Cloud Pak for Data Services](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=integrations-services)\n##
          Configuration\n See [Cloud Pak for Data Post Installation](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-post-installation-tasks)\n##
          Limitations\n- `x86-64`, `PPCLE`, and `x390` platforms are supported by
          this operator.\n- See [limitations](https://www.ibm.com/docs/en/search/limitations?scope=SSQNUZ_4.0)
          in Cloud Pak for Data services\n## SecurityContextConstraints Requirements\nThe
          `restricted` security context constraint (SCC) is used for the operator.\n##
          Documentation\nSee the [Cloud Pak for Data documentation](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest)
          and [license information](https://ibm.biz/BdfVHT).\n"
        displayName: Cloud Pak for Data Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cloud Pak
        - Data
        - Analytics
        - IBM
        - AI
        - Machine Learning
        - Cloud Pak for Data
        - Auto AI
        - Db2
        - Watson
        - Data Governance
        links:
        - name: Cloud Pak for Data Operator
          url: https://www.ibm.com/products/cloud-pak-for-data/deployment-model-options
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:bbb457d4f6b7bab065e909756c2dae31af7162dce10bfb92f274aedbc98d0762
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:716b5823661ac8f1eacda11ffbe5f0ff08a0e9d2296d04acffb646b7526128a9
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:1f82599d91ab20d2024abe134d2f56caca1f9bca1f02d64352c81b5f1e0f1730
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:d6575de3156c7a9b0553887abaf8713a758d9c9916f9cdcef611923c50632867
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:113fdfb9395c96f7eb65f9a20e18ede8a2eae4aa6c91660be8fd3e517656e156
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:d0e0144eaf27e03aec1499e53cae5366e11a3da284536cd3b27ba5691b4f70d0
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:6a996123e69c7f2c3254f334e278ee0daf75fc53980183e1a72358323e41349c
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:1d6dc7a3039e6fa044053d4e89ef54ba2e83ceadd57bcc4613a6d447d027ae2f
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:7861185d55a5fd18d896dbc57501ae5cbc1de0bc02f1ae3fe558a36f0f80b4fc
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:46bbfdc1f5e418db95ba806228fa5696be22c4cf9eefc185ce27bfa68ec1d247
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:a95e42a4f9138f1f97937a74982c98d466c171d0e3e79c522ededba6c25b39a3
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:23ba1b277cf76bc28f952ce6f5bb02d6177f76eca8a8b4ff7925ef119a994a85
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:670a36f3400164cc9eec5d23422d184340e0c1edfb13a634e58f0345c325c21b
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:0dc719e58694fb501b8b36f3a375b3c0462624d27fe49afe6b835821e31393f3
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:257bf301338f7b123099e2ffe31c3f3b793bf28d6a0570a25b889271239579ba
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:d61b7b8e6a9785b59630d9c5776f72a29ffbf2e81eb753d7b11dcf2d9d77f1f0
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:30a5b2c4e3c231373fa18d8aaab16d3505cc11a8a5d4be36eaf99d7c0dee35f2
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:fbc6eab07fcfc2d0c1895f18e17f751e9481ad111875959e9ff08e14f81b9a35
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:aadebf0598c492639e7c1109ce681256731215b686fec808578bb09558a40315
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:6cd1825ba40071bca7f46c1af3e813ae7ea44a393dd3d10e3c83e62834563ffd
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:b3387bae415fe3a66db2264ee766cf43aa8f31e2b615b200baa9228216596a0a
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:508412a22fbd65074964150d6b270f38d6df898e17eb7ec5e28f279c4c2e1062
        - registry.marketplace.redhat.com/cpopen/cpd/cpd-toolbox@sha256:dcd74c128502dc301482eb302b34c21dc02173ae1cf522356d5b500ba42b4e25
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:f00158405adec00c7e992efa7d7a19a520df485cbbcf39972e579ed3c6e4d289
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-hook-utils@sha256:ec44be69b4c585e9a3d507e29bd287f5b464360d82c8e0969c5fb4ed3e95abc0
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:31470f873a0931d130484d5255692d445e3127398aae0e777a3c3cb702917743
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:f561f089c4e5037f7c04971d96fa024ba4df5e6d17834f0c44545d0e53c772d7
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:20863de342e07dddc9aa8eeb214fba5cb17478bb88613e4e1f167d77cd0af214
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:9460c98c7202ab368e732951c13bb6f5064cf18f69d36b3d26eedcfed233ec81
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:5aecbb4de6fca34f026de72b7480f35dbb2ee8d8ef1f22e524a85a71bea99a1d
        - registry.marketplace.redhat.com/cp/cpd/jobs-ui@sha256:49daf607918b378593485da16ac417056ee295bca94f1a5cac37b334381ae096
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e924abf025141f3b3437ed89830957c5915152d2610213090e33fde988bd879e
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:e6fc9d7a70877100b42e290cd091011f2e1c674a27985fce180ee245373391dc
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:8bd9fe1d35785fa0125b8fb3a60333219a50ee8be7a7ec07386e8e9d65316061
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-bundle@sha256:94c7a1dd290e2abd55c5b8998cf08ad174d3fbfc659ae6fa6fb6cbab61d5fe14
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:ad027c91c719fd1d931cbac8c76735474f676b228f9ab23d1f243201246b3238
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:2f40d8188dc8b41fb072256ce8d109167095469af416b322ac023071814cda8f
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:e4eb6a75e0d207232e479693d9c99ba863f8e62830f3191ca4d1e7ceaf3586d1
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:847751338f3b78cc06e29a4a26a18dc0f5d0bc24cbda23b765198a6a5c4503c1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:ca4da5a8a3c1e7add6ac630c3fe4713a08739f6b891b2290bb17026b94c31317
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:c268d2852580286e5c6770857fcfc3b06a698521f5acd17ed83b8e2cb9ab55cf
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:fb37fe521693bdb8b96400fd491a39c573545f35b4131fdfab501b7ba29713f6
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:4a5ed208b68613d5d88936687f6912b3aad0e93c75f0d311120a8e55200f5795
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:52ce08ac2bce7c02dc30e2c0aa74df4bde8f156debccc59960f02b4b0040ffea
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:d6cdb71edd7b7451a206d1620b35c321f88b3ba8028468329595cd538a45abf6
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:0f01a67524fa46829e455f654d3b9b00ceac44f3d870b21c022051dfcc8a00eb
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:b646ea3330f3fd9bb3174224de69f20ef58589a464d50d1c9399211141f67bd5
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:650e68c41252300836740ad495ad6776a7da94331e8070520e60a1c5e2c5f3ca
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:6f2e7cc917adb54b3112dc4909f645ac08c33e8c2c33266fbc8f5133cbcd1973
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:5e161ff695c0552d1333a1f993048bd055168a9be9e0a75adb7c4464c9c2b603
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:ff5a5be492347e75b309194661d7742d1c909c3765c9947c8b41f4c3af62b08c
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:7ac32270044a9ec08ddaa8dee62b61618d9363c8773fe44ba1dd86e1b53acbd5
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:c70d3c86d4c61bd87a427f7a7571f37fc4d0966ff4f420b73ac4a2af73ed1379
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:978cf4bf299cd3e8cde9d7ee7aa75b73fd9d9e6755567651d6c1c49036b8978b
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-usage@sha256:8438c2229b500fc9c40629280701856345e228fc645a5b468bee2135c61b2387
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:498ad7832e989e9356aeffbee96f310e3edaa6a7466320fda9b4f75034e73eec
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:7661b04bc59d64a4fafb2f660cffb99e3d326772749759ed0d2a1940ba147078
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:1f68dadaf2ab779dece300037f3abda1b0bb33603436b84336f8fb7d53d33644
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:0f46b288d060a15a48ecad8f328a8046bf232b8629c3d79da6a82ea712b363f4
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:cab91a231068a6bbef03ee8db32fadf11654317c4573c6714721a15d0e2a98dd
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:6c226a0ccde697b56c81427a970062872929eab73d99791c9dc670f4aa176ab8
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:67bfdf9115fbee782e35ac61dd93801bcb9e622d441806f1144208b4447f1430
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:2668a562eb8df1117c40acbc3600d6735469d210e5aae4aa03bf9f54d0ebfe46
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:e4888fadfadb3462e3a9ad0395a526a2631872f70caf291dc18b60269ded199f
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:f0c68715f78bd20c352c97dd48ac5a912cad58e19a00159b5e6878f6beeb00fb
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:b0735e396c09fbd08f328ce0686cd55fe9eafd5b759f36cf5677326b24cbb859
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:abd96661823553d623eb0ea1fac8271d64391243dd0f13eb4a4fcc858dfe10a5
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:48d237cd3d363da8a206ca5e092d0f91e8a5bc931ea81763498dcf6f28c0434b
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:c078bd65fb18b31ecfe03c832643bd6fd85f7330b94a1ca26f557d32d89903c1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:631a18f4b437e5092b9671fb4c86af05e99effe5540c4d6f964dc626466764a3
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:b63e2c10478735b51b6828bb11709569a2fce62a19b70157a8a5b1964820164d
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:47a3745c78280f98f2ee352bb9192cdcd2fcd6d4c1a3627e7aa42178a0f107fd
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:a28b71ef64c1967cdd36a16e01b611fff6a387b6a741dbcaf7556ed6be9b1f69
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:095aae63ced410e033d8f7f3d27bd14dc5b93a10068ef5a389f181115c5d07c2
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:00f8bcee4ba7e0c5cf5f0c52e9e7707df367041c9f4dd5508dbefe0dbdba8877
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:ec137115210dae9b69cb2899ed8a5437d94c8a658ae1ae5eb295232b34b0925e
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-velero-plugin@sha256:fe562192e86558a17d1c88a8fe21148bc000fa93c8928f7a3a897acce62e6798
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:e857323acd10f7b99c5b7fb4b0a5b89a808f68036a1d9a224eff55a7ae7744b6
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:1cf983f9b9cda18257066fc82e199bd3e22e2fcba13daf0f60c92b50a4a0e078
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:96d382b1b26a1e1e4074b7a1aa48b973cd4865bb5c3cf02795306e92c752a8b4
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:dadfa9ef372b8d165276fc5f1ee9c63c3f01367d156700810d63e1ba19f3da79
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:45194b56f9fcb3af83158eb0297248364db40e8164fb4fe4f5656c58aeac3fa5
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:0ceaaab514f65f9a1c40717e97ca652506f8b3595186222d470a99d653afcead
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:04f9ed67696fc193877f9da3d73ee3f6f4cadc7b8d09fd07fa12d0ee15f1152d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:9f7e6a1ca3faa5af1db2afe8693f6bf9182296be2a23e305ba6c990aee3fc301
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:562281dbb2bec5191a7c6b595ab853092e6fe3ff1432e77b2471f71ccdffcbc8
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-bundle@sha256:552fc53664722903eccaa7d3782a3e84413509fd357725fe0b0f4a077c48ab2c
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:cb9a33db3707c0c6045e6c823ecb21b0f229c17ff868da7d36602d0026eff351
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:d4738f64e36a8f0c88d40043eebaae1d561193895d438fb6d8bf1a042d0a1007
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:cb3620eb9aec24f7ee54c3ec6709fc223b685177500c503af8ffeb23642dcdb5
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:c330a321438943fdd264b292b68fdfb8be89c3bde93a7bd987bd92c9dd2512b0
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:9a970756cd76ec748b555d5bed97889b8c9f410db6765eedee84fad4976cd523
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:4b3ffae1ddb7a2c28f98199be6343ed826f8fd21b64fc9c584566341f44e44d9
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:48b7cfa31b9a185176c20afc119e37346a6f9380d74000edf6c729b4677ca44e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:a7c050ae29cb93754f59d8f83676275ee9a91ae9a9956395b97abf8c5e20e96e
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:65108559a5edc9a7c4f82962ab2293017e0873872730d60d8487963970761941
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:d34d561bbf8afbfba9362d2a9a58f2919faed1502abf72e11882dd9cd4fc9711
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:4f0cd4920e7cef64e9f49ae1abc5d87303102fc96819e0c3b699d86b8f8a3980
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-operator-bundle@sha256:086eaeaaaf440a1d4a254984afacd8d04583aafb9ec11ac072feb42a87b986f7
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:a2949c3868c2c91bf35e50c64b51fcc24372986744b55520e21e9ce31fde59f0
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:e21a346d15b286956f62bdd1fbb6d7b77a554ee10a475cb698a8796c42f3a267
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:6298c1561c066eae3d420f20212a3c7a8f9069a6996408685437588bf8335dec
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:47cd9776a46ae5fdac3a101d79a779baefc802ddf8e1f8fb02216f47feb55fa9
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:8444703b2a7822f82141afcf4f36bf255650d755321b9e9fe2cbd1bdea41d5c7
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:ea57c7cb1ccd5bce17f6cb35e326ab974718be8cd19b0cd38ea8112e91c51939
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:5d889be3dff96d1e888eff94bba94228aa46f4e94e5f30ba33077ef00692509b
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:ecbc65705de76804a3c7642339ecfb4cc0f5a61891058e07c20bf2b63f1b82b8
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:690e9b5218a200de93a7c8477fd2980d8d341f22be75002967201db09b5a5524
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:9c2093089ca0651bcabc1d01ee2b69d2fb0cf2a277a8b966cec9533950fdcdae
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:a1754e003811afb578dace78f9eaef91e7917e5468e590fcb41920fa71bcc22d
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:010bd5bb9943ffdffa8fd00cc8226e59d9583bb3baaeacfe70e5b00e805878f1
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:3d5c6573faf4ae00cbb21cbfe9b2ad5cd8ffb658df0c5ba09bd98b3bf54f9979
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:8cba5f256747002b62fcedb9066f7ea6dfdb1071eb359e9137fd413d20b29ff8
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:6f96b8b6afb00cb26fcf903b8e4c30b0d6344bb86d8bce2f2026fc8cb0b575ec
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:56315ce0a0a70f5899191a7eab856105489143f0e889c8cdd1b6b8c225f02eb7
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:d733c5a10a21cac2df2fff6378035151e5d0df57e3ed1f96bb3ac17c0c95b8eb
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:d06ef794046dac2c58163c201f55489b69081dd15dd91cab16387510a50e24b2
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:2c39ff1850b7a9d32a11e99393039ef74ab363f3fffec1e4fd542f2ae148a6b4
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:61ba2b05a6826f618853ba048496e2c9558e509a01648c74b50cce6d28d0d764
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:e0feb4d305c5ec9501b30ccdf8f5826cd8eb1ba5ba1d62d4d685101c785a708f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:c8786cffb4a27decb5dff3d2a9682d07d11b7096a689bda0d7cf105aa3156299
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:51e1275f6e8297418997b0ca9615642153371c29b3466f155e15d7fb9e80cda9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:f1aa6add76f59afeccdff84c9ea89a46b49c6696f62acd75ef8ff2f31a3080be
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:132670412a0ce6dd010869e569504d2e89aee0983ff221da371c9e30d5f8d8a9
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:b5a8fb14c3bb7c89ff970ca85f06dfcdca832e2b45ad86e8bb1b6c0a0e7e0764
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:ecac5e0327d939ca84d304b2fe95739ae43a06745ef775ee59e4e5378ced65c2
        - registry.marketplace.redhat.com/cp/cpd/environments-init-job@sha256:6af8f900af7c11eee6894e2c3e52f740b3c700dac0ee753fadef330230bd42dd
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:369f66883e040accb73c3deae984ed479e90eed242051b7035723f0ba21b51ac
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:7ae4dac24ea054ce527a507dea6c77a3233aca0047672ba5a367fd5742448517
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:8d7625a0ba656425f601b6ec9a398a6d30511fe1de119cf0e29f680c704ea225
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:7b2e9c373f9bfd91e16eff68ef2313171c69b94fe5a21d3d3158470c74f54a05
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:ad2215adfb8dd2dc600bf5097b7f88c257d5fbaf541409c01b36cdb67c130236
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:c41e4d3314fac0d2cf5f08e7db85801cb2098d37d0aa8104a347e9e82526e914
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:2e70036d7299168331c94af70a22e3b2063c56160f616362b1a2187c12d003a4
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter@sha256:be5fd8a8b43ae3792c58de1342cca91836287fb584bab003f7f9aeeeaca3f65f
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:54c86367bfe0ab658611856cc0a6c5de24c14a336bd129ac9f16293945a55238
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:603a6bf0e616f16b609bf157b9e21415802478aefc9c606e124008d0224a1188
        - registry.marketplace.redhat.com/cp/cpd/task-credentials@sha256:e162828d5ad07ceae4db0589f930979842714e3b766552102bee373a3a570af3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:4aa81bd0043d0bac60c9f1b0ab56e45a3035c88314614e60eaea48b2a29ffa16
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:587af68ec8b4974b29da46ede525c70054703fce8d5934f0cbe68739c6058fe4
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:fa01c7d11c23a47bb84375389e69619455c2879b5ed6bb5b9b753b945dcb02c3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:08dcf99453b4dc6dfb5620410b8f7e4eeb42fa824aedb2e480e2fb40ec8ce347
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:0164d8fbea591bbd9d67d84603162c3781b08551eddcc2a8a842aafc828af18c
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:c957787ae587a31e6a497b46649da9cd46102a2a3dff29243fd5cbac96e7e2d5
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:4e65e96603d5043e78a517a4c6e5debc5df7e3747fad4d8025ef1fdf1cd6a9da
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:e3d8d4d5285d058130345d6108ed4056ea9c871ed8c0575601a94629e86ec691
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr@sha256:43cf735b51e2ef45a911ac7032f03c74cf0125e2d2801258d4aa04831e1c6ce0
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:4a87dbcb7d2ca2d121512fbf57b645eaa9ab745439b79722f474b84efa9eb391
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:6fe90f051d9261ad224dccf4d9f6bc9f19ce0e8d22fa010a46b2752374bce2b3
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:9a537eae7e581322e2f0764ed38a797ba7cd260724a285f24d5095058b7b47a0
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:3be006596f8e1d36865a9000c77df3223c264e807bf0aae6049a23307e8079e8
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:d28e6a3c65b580051ffbe287492dfad28dbb1c669cb76aa3df23bd9428f4bbe9
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:8cb37bb0f0461efc603cab89b5420c56990aa104fe39c9b77f0393bc6b641172
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-bundle@sha256:98cf2a6c7b7af5d7cbfbc61ce140201272b30b12edabaec2b2d4287b0690d89e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:20ceaa70a3ab5ed8a2e451a1e2960d04c201f0087fba1679fb80c57c6be6117a
        - registry.marketplace.redhat.com/cp/cpd/job-trigger@sha256:f2ea487072393a24e4c6864b2107c1fce072ebba07f9dbc2c139bd37fbd062da
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:83038bb7e3f4b4637f0d5095b42dc906e7fc27bbffcbd66f6ad93dc9b1cd6bbb
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:75ca37701dea1bd079df8f8cf302cf95e8dd4ed817fe126b192ca368e92fb78c
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:b10ca594f31751faad4050b4f7717e0e98c7977fb7e26d4866e1c14125422225
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:b2034fae4f2287c8997104154064f2712b98d3c4382d1faea3df777338711f23
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:849ecfeaa36fe365c7dd8f1a76fa0d263a1196fd16af805dc66035643e8e9d9e
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-api@sha256:a76d3114418a3df1d4d3cfe03da587c37028ca1697795e2535cac9f6f190457d
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:481391c273cffdeca36f91543732ba1f082dc409ee851ae95d3c1d205a7a6a08
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:c730fd4a7baa3f46c03f6c73e2c6c7c0ff24a8900b79dfc0f2fd319909edce74
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:9d4ae40d34e38a48ad2ad6556e997a42f6308006d5f6b8638cb7bc4ce5b1e40a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:6e0a23c3434046195428792c45c95869d81fb4ac4ba75444e4343963fda6014b
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:deb66a64288a370c5da5c7cc0ba4e2ba4ab816d7c5a080a0c505b05f55fa6072
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:443b51f8b10acc85bdefde7193e0f45b1c423fa7fbdcaa28342805815c43db3d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:53d1fba68aa0227c2f4c8de26837f6c114c56ca4a397826e1f2e93bb857e5b7a
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:42ffa27ae47b577cdda214e0fbc46c10bd108211e7dc58777d8236c59cb78766
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:99589ee11b50af8ed6355a5f9c2272ce1fc7f9a13e80fda8eda5883a332782a3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:a0273c013c5943c9b9a223cf1aafbc96b3cdd2dbb147c73d2a9ed49befdbdfc2
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:9a0c76fd9ef9b2dd9af5b0f29e1813567ab7f7d35c1efdf71da1a315f2942b45
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:e3138406cfe81021370924992dca891cba22b3be2eef84546faa0e8f8aa57ad6
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:62e2b059f565f87beded59b36962c457c9662d1e401a78766fda328ea1aeb78f
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:166a3918b942a4ea7b6c371a1e30fd4c928586d7ffd5b6919c2fc728a086ccbb
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:d30397b74780cdc14930a3a02dfcf6b99264847702afe0513ce536983c32e2b1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:f662be48eeb6dc93a8a4b0c3374f59bc342b5b8f5360437978cbac8f74c3a084
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:12eece1fc0a2e14244a2f57dcf7c198a792ced9ad0d02c9606645bce1f295877
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:130872119947fe91cbfb8dd82b3baaf78f7adcaca5933ca394eca3f9ee84c8cc
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:f3f8c6c395ae860fed5761f67ce092a122e7073edbc205c9167c27a4ba85094f
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:aed24dcc7ec24707c978a17b86501b65c6f0a77bd3ea0d173c28684b9df9cd91
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:8b7941a3f518e952766a2e6f88acfa7c9914b6d2bcd4ad9ddd54df313fa30347
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:722944bfd0eba280255f4c198f65f562423a26e2c4ad3580591ee71cb6e9c2ef
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:0341c499ae144ee81dc2b92e8b99ee313ff6759374b7c1a6469156c20f522921
        - registry.marketplace.redhat.com/cp/cpd/runtime-manager-api@sha256:a37773df07fe757dd698e3f09f8687902e7b3592d97855e703dd92ecc832aab0
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bb074bdf89cbc7c703fadb17639608599ab4f66db9be5617e5f79a0550676790
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:e673f1d0c5a0e32e9abc7ac40eff905ce411f9f3d8ca0b03160e23ba5661e041
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:f2f256a2fb787419b25e99c765003d25d1edbbe60042151d91b46a5d029fc0e7
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:a16b822b41f5a745b475ceb353a459e998f9f43bb9bb867ca1640866ba51780e
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:8a6ce0dada706056efda73ad0cc1547566003e0a7aab8ed379b50e4776861b94
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:ec5e4a002cfc63c1ff1bec340111e8a1755c8624066445db36c23fdd1c3061f0
        - registry.marketplace.redhat.com/cp/cpd/environments-api@sha256:6dd60b530f81d7a2d55a534211f7f128814abf87b8cd3575d528dba53442c44e
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:b0c8bb70a01bb19c685d019e89f83d8ed327bfa670a13b78d3e803fed77499f3
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:2f2590734343c99183dba7aa9fa4debc73377ee707f9bb090680750e62acdd68
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:25420a21686a32c62e83b99fcb215df015dd13fec10f73f397f927c61226f964
        - registry.marketplace.redhat.com/cp/cpd/spaces-ui-job@sha256:95dade51444e30ec034c12207ecbef675b70889b81970f49f10aa75160b2f947
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:686700e561181ebe1431f71eb9aa11b550bc4361dab8a3668ce0c7ead217f79a
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:3d83ebdfddcbe6eca38705fc3d575f2e70580cf9f7e8735941afee0a33d3b5c5
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:99c863002ae1e93e97792fa606128d05063abbedad9e9404acf28c745ae1a5c4
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7d9b2d61a4eaac645490d1c9e6db97ba193d65015b6efa8d5a1f3ec6cccb1ad1
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:06403c4ad99655cfc10fde46b21e84f90c05ae143bb6d08464dbc8954486897a
        - registry.marketplace.redhat.com/cpopen/cpd/cpdtool@sha256:8ed09b7d65178aab93b77b9d39371bc86b1022f32c5e63a9192baef02c71016d
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:ded16bf3cc993bfa5af8907a1517cc34ecb511948e4d9ac73bab68888af5543b
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:107911ed6a60a024d151ec3724ddd6d2d1402aa70a519b9dbec1a519d88231fe
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:0157fe80f2c67786cacaaf1302c334a6e0ec0b941afd870e0c4ba163ce394216
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a24f4cb635e8b5759a5e2bda20ddefc6d5ea9b16fb1081592885c54aefb5480c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:0711e1c6001586c950e74fbdeef8f04d20ce7b93de2ad033e82d2d5421338936
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:ad892ce5e6a3f69ef484eed68e92d9c972597a86420e22ff80c247689ba02905
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:39700bf47b93c7f0db1894148c16c60eaf406196034351f54573f0a94c4c8aac
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:3112fa03dc9a530642d329666a0cc64c06a04308beb1ddca0f7472915719ef5d
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:66c1d2493e6721094e85a0c65011523004f9589f58b7f68cc6f33c5e5adb6984
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:68ae72930a346d0fa0871cff573c64285266311208f2bf64101697a540f60272
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:2712755961a1c1e23f30d49341c037858c0965dbfc065aebda88c84526291f21
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:4b6bc5640d9e68519a7f543ddbe3f16ca200f32a74ce9c07e996cf122da3b59f
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:39e6aee0bb64fe6dc18ce81496936b113de6d6746221655cea102e04c11dfa33
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:07142ce0d259412d0a9908b63dd78d845ef6b7c5f8a24442b93964b4c38372cd
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:525924934f74b28783dacf630a6401ccd4c88d253aafa2d312b7993d8b9a14c1
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:73c40c6c72b7f2399fb0ed7e462baa72c3e8b5aa246fd2fe1082904b4afc0796
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:fcb79d76f965e74f7e3459abb92e3561ba2f016b34e2ebb6f88298bb19f7bc11
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:46893c166cabdd8b3422a190e220ee2e34861e797aa068100ccf53da349e3fe7
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:a477d47eaf8f8e569601d3c23c1def66f31c098ac3f5a1ae350618c460917dd8
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d51033417b786dd2dfbc94e648db00ba68f316c23b06a0bc9e36a5c22f579124
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:86cdb5b6543e7184361cbf0b0fa556277232460737ad254e920756c6d365e47c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:e6ce3c998642b235080daa6c0b954733bd3fe973fc8d85a10ef0cdae2a73aeb0
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
        - registry.marketplace.redhat.com/cpopen/cpfs/cpdbr-hook-utils@sha256:63a5606df81013b8a7b5cb07bd1f29bf032c8998db843885184fe86a76fb93bb
        - registry.marketplace.redhat.com/cp/cpd/environment-specifications-sync@sha256:894b4624d8ead7286cf9068c6477300992c32d96339644fe764bcb29b1c8ef5c
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:3a872814d01ded278afc8dfa6b1569d9e933e8fbda28c55a991da010b4304c69
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:351375de35d7fc9d940b905b867ce8a00b7ff1416e70d3a100918797fe3d8c56
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:00c30fe372524fa82b5ed4a637509dd9d4373dc727a9abc706221f832858f9bd
        - registry.marketplace.redhat.com/cpopen/ds-operator-bundle@sha256:be2ddb821274938e5ab926627bd998279b6d19b7789d9683513e85a4b05bc2d8
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:edfb732843a53254ebf7d4a2f109af109623be8fb9bbae94366d30febd2d0494
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:ee15594f82d374da81dab59c559567af74153524a6d9bfeb28ee085bd4109b37
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:3ddc6bbed1ddae1be71043d177cbcb58830d0dad542b06567fcfa254dd11e626
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:75595bf2753fb43391e2a496fb6d59faa8785babebb4aeb1da34cd5e72c93699
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:bcfb9b7d3b379ee2bb0b449cdf5af3fbc88f99c77641441ca392ebed04403e00
        - registry.marketplace.redhat.com/cp/cpd/environments-ui@sha256:2923664475c7a8ea5e63055c90748db37825fe76c7975b845f1185d4b32ee4ea
        - registry.marketplace.redhat.com/cpopen/cpd/ibm-network-performance@sha256:30e0b58531f876a47174c14bc6f94a3c38a40f757403775a06745d4ae992060c
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:119be0bf4c60230c43bd64c0ea252108a0173adfffc9f05e5257ec260b335599
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:3c1e08c56430e9bfc0c831be2394fa28e395a3341ce5cb5201b08828298deb60
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:2698150edf65a99a1d24bf3e8336f02b57fd964fcad283e182794634b033af89
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:077107dd9cede2c49beec429decb66760372a229c476ca1089e42e4ed5ef40d9
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:9782a9de863ad10ed36341d07ea5087e40f6f213cd4453c3ac2ec100d8733ce2
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:894d67753dd0c5b8625369e53635a1ff53066d3910b73d91e5625dd8907bff43
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:6bec687b2bd5c4d04403a25f56d2245f822e914907771f26c25166462976258e
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:993f16d744a81c95e73339d3c8d5880b5cdbcf29beb914fcb2c6403af6d628a1
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:21ce62d80df2575c3aa7c8dfde706cf8af01b2fc86f72ea68f767bc1770fe733
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:2565a8ba9aa103956d6902f2bcb53ac13328b647a2d7a150bf80ea2749d5b972
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:17bac3307ad3843afea77ed213279603b5f775f24dcf4857095baf5029409c18
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:749dd365ce85a4c9ac8e6ea7f90a34b791a7bc5300319daa85d427b04f40af65
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:cff7466aa6063a2bca677815df01208822930f8c12784c1c406fa352d6bd6cb4
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:d5feb4aecf8f36531db4c52e672558c04f7e314a1a791b27b5638202f9477bf5
        - registry.marketplace.redhat.com/cpopen/cpd/ibm-service-functionality@sha256:ea4c32a19b1549698c8a68fa030d7036d3ea3cf2d000f3523b6e9fc7c6c708f1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:0e563d51d601adbd710833cbde26ab8f857092f85686213adfd6fb3fe5e2828f
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:f7e68137e1ca58a63924967a8e4c872912557fa68af8a7392e3e0d4aaaef1e78
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:009d98ff8e3cf353cf8ba71a973228e1e409e4de21373498534a6cce01c07ba5
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:fada47a5f4824d3cd5bdc262b92018ca0e697e18dd373f92fd93186fb8719a1e
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:f88a7294aa200d3c633c0a25c4e4ccda52c75de174ba508d847ffa3092d5b8d1
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:2d320dee224cc4818a84493723f2262d259891bef89dedef09e1f1e6ae5686cb
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:489d7f80d83c9890600e3c9c4cb9442a6f42b9c6c4f13dafdf0f614b72e67143
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:934bc42fb29cbe616edcd93d2a060dbffab5e5a8a2ef4caa7b2e538f26167ee3
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:96e6db93b2d16ba590252d6568b68f254548aaed73ec17e4bdd8d530ea668be6
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:5f0974565de8b045d3faf6acac7c8ebce6cd35327dd76290b39fe48b27ad5b62
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d4dff1a8968c18d984e16c4c9f1517c3a10e23a11bf166c64d734773324d4c9d
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:c411456aa41fe705b47dec60eb49c6d65d4b7bd1a64d1e0545817efc95afc67f
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:f5527ddbbd0b9878555fce852996b13020fd91b1604d8c03d34a61ca2fdb2d1d
        - registry.marketplace.redhat.com/cp/cpd/jobs-api@sha256:173b88f90078e1817686ac949f48387af70a706e23e6fdab4a7c6217db5ad86e
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:85669f2a35ecc4ebf6b46fd793ba6a6bdf996164c00662e65cd7ceece8d60cda
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ed81ba4ab2caa7ffd1cb90e11aa8915df1c5bc1b76aeb4c87532c84097bb102
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:23337786ca20ab624eb428b0d53490975e2b08a2db642ccc76d974dc0e455f55
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:b8c83fe5215b7412765181aaba84162b6cb188b5765e45972a6f29b2a9f19826
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:811d1cbda1d0287dab48cff90c7091dd1bc8aea385f562c896f2b09f075cd1f6
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:7a0e784b0722ab0a4c62754a4e817a1f25e5a2d5e960de76387971fa1ff7da41
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:36e3b8e88cbb895ca89288ed385fb800b888743844f50fb2233ebe5e6c7498c3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:de7094acd65a80daa67669d8eab809220dc715b9cd6aaa96c58e5fcbe34c460a
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90046eb6170e12439ccc5aa343f19a4d4fc99a10d98b383cf1dc0e020ebeed55
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:164f6063f3a50a36e28a240ae7c060c4f90266b7ed799bfdc075f90628bea7ff
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:78fffd8a74eb6495f104ef346bdd03a5636134e2d5e8da9ef5c2647a9ebc7624
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:19bb32e0389361e8314e627bcad2e9bd93984289485cf632466834353d9c3449
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-oadp@sha256:35594a16332ba3ac00226414be9b85243bb25dd0648e256d7ef9d42137dcf914
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:787e01e29abcbb3d1fe58e0b6051067b309085459983f38d6bd18d3811d608d9
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:2f45aabc99eb9646db4c74686569a7f2af3f27cc44362a30c92d171540bb274c
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:97b9ecef8caef156a0eb073b6eb2a34b63209b44511f5694a243f11285d88662
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:bcc579eac4be59609102836252ded82a17ee0088e5ce28c684fff41c80232a0e
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:da38999d66199df0eeabd7c9e934205fb858cf2c52be8bbaf6b34adb894dd38e
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:e29885c1ee083a73b952b0de8d1cd675b2b2830abf44d69a8577f56972a601fb
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:17f016c11039c34f90f67279b45ee488740cc3d033112a3bd589bb1a60ce6812
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:bcadf3d367f76d3454e119a9d39f6fdeeb6cb806c282f21a1bc3c38bed41b427
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:ce57bc323b95db548dcd2f90ec938e708988cf6103579e793ee39d36359d60cf
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:5f63e062489b8e1bf49b5943e7a836174fb47af8dbbc8092ace5c8b72baabb99
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:740a1d87a43e61e32cfdea032d852af33100d959c8b57ff2f165f4baf77fc011
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:d72cef9b19ea48f03f7658dd55cabe6e118be0209852025845d426b1fbf0e0ca
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:1a1bc54977b160b70cc00cc4eb1380250dcd639fa3d741363212b8081b35f72b
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:ac128758328325ae56273800ea010f9559902281cc60226c8f99de51b636b180
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:26500e941ccdddc08dea07d3a4b9aa9eae7840fd5b683191320060f6dfc88c97
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:19ee4cb8e3e60d9c1e8fa63b0bf3e218bd683b6c05b9e9553d9477c68c493a9d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:7dbff355db7739152961cb6a97887d2e43bc960ac58837c126e03ed1a4480a3a
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:075b6bc4866b812c4b5394c8902b012978449a710ce71a30f55231fead941105
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:697baab21258166a1897ce4e056e39bc4667ff837bb97cadf92dad5863242ceb
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:ef03bb6ed647d92b9c84d34810376da7049aad88589eb2e35798d82a22f5b994
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:6df671937bc55c410a9df13aba15687b2832c7c19c0eff4686c3f83648528068
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4dd54195a3283a820d260eaeaf83463722e49bee05d875333c1b99201a7e2b85
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0a773245bec514cb893f730086f39a3789d0204b82cd11f5d11a5c270b92f70f
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:fb15cbade10fb37b14afe063a93decefd9a40583895b7b6b436f36f08bd73aee
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:a6e98064effd12ffda0613db98e795e056b83ef77b5deee932a762b178ceec3f
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:be5679938413dfb7766f765fba7923d3079d2d9b6f464485a8309e8e306f19e6
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:0cc0283b768c2968849f24cbf7178213926e084da5d0c90c18d38e2260e19138
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:82f60c881048618473c07970d6463b3186518bc18e4b263e16ca61d15448ab30
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:e76b61f9f0efe16facbddd80808d80bc32d0d41b4c16aed566116f1366152987
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:b368c22956707b8e594441c0e9bac20981339a5713d4fdfbe3c64bbfdf677b35
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:9fa6eea66a3daff39b7c0c7839ef5349b483b29dc219e1db384aef6c4c31ecb4
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:dc257df716b270858ca5d07909bda9ab635a388c9887d30b015391f60cc7b5c1
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:bb45d0217b4bc4352b209e3771756cb82923b8d260846c75b33ec270b4a8b8dd
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:b797e00eb32ff9a6a56e71d355bea653ecdecfecc121ddb197a6e347b89dbf8c
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:0e0ae96aff8540323da8a2379af014d3af1770e4b0043b84b9b415dda44501f7
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:37853d2979d0ca5f88fed76fed5194b628507ef0480e668dfd6eb034fa1d2023
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:c0459e5d3acb3b2a6d9863efe6ca1a24e529b38a6ddd3cc1e13d143924e872e7
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:582ced3f60212f736f720558524225b0d4a1c79ad4890656c47dd58009ffe0de
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:351358f551dfcf817c5936c0b43c2e28db61adcde65d7d2c81cae666630426b8
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:07c13b18197ed0bdd8b97d65312a342f95d7a7fc43d50af746d788032a67f83a
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:67515618d631b0ec00ad95ecb9bcb48d7f199855af9a64e86c4b46e7ae3be69e
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:932825fc7af8b96c381d504cd0976c751661ce078257430e5c79d565f416bacb
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:5ec262d1ada13a7d5d519ad98cf82dfbbd915664bb7560408558001ea6bb8451
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:3be3ee728cbae6bb8cf29d47c37ed1a4cd5791b54e1262135089424e5bda7ea1
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-usage@sha256:fe5bc4a40472926ec0df3023777b2aa8105507128bd9ac02eac2c68ad08723c1
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:f6e8de9f11696ca22302b8576ffd5bc2c024df0c8dc3734c404cc799565f351a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:36f4c57c6594f997383c4731e1e62c4765b8642c0b5c7373ba22118220928d09
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:2d8f924a7e4ef373df9760b4f1f3ad8050facec7de0c0efccd8c27d46b756e91
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:719b0871f0d32da894ec80dc2a23483623b7802a5c6f935cf3672468f99faa40
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:b372540b630507f4cb74f5eab38d0335d7b5bd64197e3f83852b2c7c5ba9067a
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:f9c9202e106d5a159c8f6b8fd883da4d6ac0d81e75b558fa52e0f7776269404c
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:095c969026e191bc4fceb020b5760e7595e439cca91b88f0d8c3e422591d56aa
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:8d9e35c3f29f8f56d55d81a2e6cc4d1b0cc352ddc8d33d83dced83c4a111a0cb
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:88d3c1ee02acbffedbf6ddcc27b80be6727f85425c361d45cc5e5c4056cf6733
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:1279b116fbcebb1c052eb27a72c67011b44a12fe673a86c1ece28202afd911d4
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:13a9ea6ee24d9b0ac1d5fb09c3e42eff731779867e7e71ce40429b43ba2779f9
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:5e5176376a664755fa776efe623a1e11bc5d3c31a3d0af980b165e54cf06620d
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:6fc976bc622c2165dd8bf93c9244e1c29b7c973a4da7a5ce4d0a097a1c1ea4fb
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:18f44ab15289625fa2c6b72ece6e06afb51058dcba30b8f38b4fdeedf8c9e51b
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:4fb2127ba7c525a959bbbcf024de364ef266ebb3ca7e696f74d990d2552cda59
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:a99b1f5a4eb0306b83562de0184fc826c34835ffbf46d929ceb9e70e0d325361
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:2011f32eb1e68ed5b75a4e8a99c992eac546995e219701d17a511596a5a09361
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:99d0b3192e6e46167423505f48655e7d49af1a461ae00b95c9c7c9d4209e5b65
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:65d327900e3b7a874dfe4bfbd0f2c7a1ee402346aba0e8cbc628b406e5b5329b
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:bb85f2567aaf7b612343fb9819fe05e2e771ccca8383316d34282ec37fcd1383
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:b2ce782efc5349a49d9c7326e2f72b2c1d99d30eccfb6975f7cba10008f9b2d3
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:267b7da5936493b0daf8ce824e3e6b50ef3b2bcb6e3126ed18fa9e09cdb0da43
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:27195e9758afdf8521971866fd60ea3dafed91a146c3c4165876e5ea476a4622
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-bundle@sha256:cd92fd1c85a72560241fffcb3bb62cf965f96bf2a4e1f5cca84c0b52624ae962
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:a34169228887ff93161584753c990bb82a5520ad7945dfa1ff4078e70aede592
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:7dc67256c002e86ee416c60ebb2942b35d8dfe0a5f96b4db495eff7361d21466
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:7de2c2f7cd478baba2011461c8d28c1a4c70aa485404d4523d1166dcb94c0a79
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:427e431d5b8b56011a91bc4f1e1586cd5490944615011fa6ac20abc8f45f92a7
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:96451d4a9d1c4b759d82dff2eaa9aba2a6fd58518f6ac6306cc082ebb7766a64
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:eb157bfe50005f9ffd3560ce1ef7ee0fc283ddfe265c42b35a71f18409e3f561
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e1b10435f94bbb7165703720f58aa8c68acbdaa18006a2c36d1e24ffdd72cfc3
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:411a4c7164b59ab3eed86e67aa457e45bd028067c937d35d0a227de63870cf87
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:3743af183b7b801f72b279a658b0d32da0bdbfa526a00d191a9a44de6df2284f
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:2472053d912f693a5fcdd00b7d9f0a7ce98697c47eae955e31c0b602fd8a4a37
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:b2ca6bc1119571cce732f1ce2ccb9ce406b298550b55144f950bb46cf985df3a
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:7cec55c00827b747f23c2f1703d99e1ca5b30a0c445f8ad0471b25b0c9cca57c
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-ui@sha256:b9a255ea269a11c85f52fe556291b0e869efbff9918b488b65e4297b958c6059
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a4daf28771178b0af391965684337805f8de0019a677753112961710f6011991
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:b7468b824d519e46e3ea23f5a803a8ace030b74d71d4cd282819027665ff0e12
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:3f2f97e6b2b18526255e88ff02d1039b24312eca51ee22b7cbf76d5680f7aee1
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:183ee860bf48076ea155db08032eed7004c1204a363625f54218734b00175441
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:62102f79bec5a8f66c600d184478f2b42c8acfde059fc71bc262b16dd46975cb
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:3129d9c7b096f9fce9cc12002eafb5c3d731a162ddef61578455dcf8acd31fac
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:a72c2b70a9f0a93047d47247cddc1616138cbc87337289ddb208565af6e4369d
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:27210840cbff8a7c61acac2ccc328a415a405b4738422b4f47eb9a0110127d22
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:381170c9c497b1bda33eef3ddeb1bafefa07be1e6ca75e75c9eaef9dc87b8d4a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:6bbcb64cf9adb4d3b6a4c259eb910e7da5c3c58732643e60d416399b2083cc0d
        - registry.marketplace.redhat.com/cpopen/ds-operator-bundle@sha256:e30b09797e341a5fdcc3ec377657e0e65be5855fb850dfc31f35b200f43130d8
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:ef24cc79265473a8b6fea231b1930c5763293dca79eedd46c89ec63dd95812c5
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:a31e0fb7a05e969e31c10a74cc858c483d5d4da14ce3abc36f5318100e60e792
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:aba8f1d08d143a8901b4d2cd8790e067fc7b9eca9987696a1ec6f9f505a686ba
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e12599ff99db058fdc8f9c1eebfba31160c636ff092138deb1f6f6705f44fd62
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bcf193d6f30ce1c7f9337ba8efa8d9b853eb7961e0008aa82635aa383fbe9d76
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:ab31869bdab5180b816a89c45008328cb7013c26b09c531bdbe5b00382f0dd7d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8ccecc00f3a9656a3dd1b07af76ef065e301e4101bda4ae8dc5bb91097cad7ad
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:c2b9d370b6c3e3c8e30079465502ce376055261cbf0f95c48d8a2ae10ca36c76
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:a6d5bd253959be1f91b50f5f2c56ed77e1c82b06483e62094aa86ff57cd46eec
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:78206c6634f2be2bd36806eb717bb8e4db2c755263cefa6d026a425a19329324
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:673320deb0ae96c6132545032f6b3e569d40204d3a5565a568dd97c5c1a4a552
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:3f6e9eeb659857799e30011f222924cef35301c59eaa5327ab5e8d212af75ce4
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:be55d9ad6ae76bd061046af448d032e65926abaeef95c831c978e01e2b0355c6
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:f6f3758287dd31889933ed0461fcf02dfa85d872c3d193669608573ca5269524
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:14da175e95c91864142ae00e9087ac2fe0e2083fc1a3bcb20a6b29dbab99d8a3
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:0cdba17b95a959b6d04889179d105515ed3176ea3449bc104f18b86374876048
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:b19bbc36033c3e60f387a9924711aab1ed81f150ca9d8f435e5433fed7f6a5c8
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:a3a0d58f895fca4d832e602d0dee80018df4ad7bad8800ad4058eea07fda4817
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:8d744af7dfeb9892378b7cc67d0e85e29d12b6f79c8b6b2a53f7a53aeffe8231
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:fa3d8a54b1ba33465171240d66f62271286cb120c51c54267d2b4f8b731c6177
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:7a90c8e0180da12172b23f3b54536bd1afe35affe37d908ddd78996a81435751
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:d2680c5f05c886b44df644aef6cb2fdc9534bb6a238066f6ebbb97bbeaac431e
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:6ccbb20c9fc7ce00ddfbac19c8808417957481473eeb1ee62ecda165e0e275cc
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:ca175109562cd9140d98bea3be88d6b34490ccb8d79f0cbbde08c821db9c5fe9
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:4f0165926685f49cde22a93387fea401bc1799ee846041fe54bca148dfc2f62f
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:70d23c45cc636af2148fadc1c7a08802692fed14e1a31893214b8d2d82f06e1e
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:9ad526bd56829ef24834b8a790b29c50809866c1dfb60cf8e0600fb9a903d388
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:0ab7d6651cdeda78fd0416137eff72eb58a621c8cd55bbbcb16f269808dec7cd
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:545b41cc6c044d6807bfeb7453a045e074a5670cfff9f5c311f4918946901fcb
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:17f7039e49f28abb7239e0c4f3689de58c75f34c9bde9e7b33a6ffb93ba92677
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:0fb78efb9e92dd0c2e52ca2488e6f1e7431e197f87e6994e15f7d621b90319f2
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:4deb399f28d3385c50a256c6e927cf27115c311be5943021733e4b281c12a419
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:5f086cdc7a272384c8a9b680f8f90fd471799e9e96175b40c7f16126b73169e7
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:786b21db2a9ce6cb0eabd5cb7c93dc41a8b1db6c1f2d07725a5bdd75b3a1b7d5
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:cdfb097f2438b5739bb88905b8318d15d15c4c64d37cfa6fdb06f44dfe976e6d
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-velero-plugin@sha256:2dd02d8f15dca1dce7cecad098cd8791cf1cbeddd2abd8e8c6d8f066d85f9e2e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:9c635194a9c63fed440e7c0ab0be1fe6cf59b1c3f954216dfb9150fe9e0fc207
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:044b663e4b7508c62dbb28b7fb2aa267c9771fc62ab509b08d46ea1499d95431
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:9c24d02b1cbb3857d017453744bbb2e68fbe133210c35e7da5ce0341921d5b08
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:cf2b88be76a4aad436f45a79694f32d2e2175587c3630543a6d24d7725b60c5a
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:9b91aaeacb196806646a1f196d4adc83b5159de767fad1a2eab699f735938911
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:d023f35b65816f244b6d699b41362e6031553f7d4c416efcee4a7f2395a61355
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:1019e13e6f072c07da6d31277c8ac639e8ff2200599ece1666a3e34745e763d4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ee161954766211651d7cb13f3f5a3c025ac04094ba742be453ce789c6f0e24a
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:4d24fb274b5429ba61fc4bc6833e93b14b92b8d525556db64e2172b45433de31
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:772f98763527c5452c662b2ae516de59c1f6edcf339da46231f4be8aab6c580d
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:bb6492c452fa6b481551628da26bb85bad151d12afd34a89c1a272a9587ecd1c
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:00cc7d072ea6b6f5e9f3e5b95b9cb4ab8ad7bd942d8c259e70f8ba0e9c09a464
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:9c160303552a9b8a97b951311018de2ada410883468680dd61fb41519b9a9e7a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:80c0913e2c84530d9b7c6a2152de76ded5754c134f6768754d00944f46d0ed4f
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:777e090e0394ffdf39abdde285b5b2749d6835e61458f117dbdac108186c34c8
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:d7ca81372914a28c8298de5f9d3c35cdbd6d8d8582ac9b8fda5a85d22e3c0def
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:07c5e4b3dc18b921339a7a5c78b8fe9179205678eb662377e8124e46479d7f25
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:2a50cd6ee3cb352ba186a57ba73fb9df321cf98f0b79b3d81024cbe57e0c7f35
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:866ebafb730e660062cf7401023d82c5242a2b80cca0126d47a68193c84670e0
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:c2aa3566f8fce992f00f43867173e5f47593e8618357369b23044fdf3c67ba53
        - registry.marketplace.redhat.com/cpopen/cpd/cpd-toolbox@sha256:1b3469f6fcd037aacabde3e05605b908e22f012d331d7365d0450f6d933d3efd
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-api@sha256:dd9e4909317512a6fada5ecfc7cdb47c6807f266fa456a01b60d7b3486faeeb7
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:fb850a3d660729fe8281967c2cb466fbad3f01fdc302db17988c27682217ad29
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:6caf5203489492755e5cf6bffcba53be0bd32662e8cd331dada68a85b2c083ca
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:7693312b2bcee7b8502de0da8786be985607fbfe6616bc9b0da4af8a6fdedea8
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:4ca30e2017a0ff2d64100a03a84bee8d601c765f0e624e1f485cbb97748bb8a2
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:c35eaf4921bfd046985ea7a03fc325f79766bf73e3ddba69408f4467fbd7994f
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:4184f25155643f01850f286f15cc92294a92d8fdc8e5020b1b522977ba429c2c
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:61155acdb8982648ae5748be9f5944ba913db1590ec3e773206d6fbb1d8952cb
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:474187fe2837da4b86ed9482a96fbca0a7f83db149859f8ef23db84de9a6d7fb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:55d4c6a13452f76bbab485363288efcace23975623b8a4b6b83f3c549f844c58
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator-bundle@sha256:c743dd7de2edbd08c6c73e20cf15f3b147491ad77a81677339e88caffb93e84b
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:b8886b9b54e6ee4c6b2981ef092e54348a0286577d9924cf136cee881f5bf1f4
        - registry.marketplace.redhat.com/cpopen/operand-deployment-lifecycle-manager-bundle@sha256:68f71f2767a40ca9333f0a670e07493e8a8c13a5c7ca877ea5bd8a544b352d9a
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:2754384c6936b36ffd64f89df751cbe47993f73dd700a184af35f412b3408d72
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:887348b5035f5a70f1c31aede992a5c76397f5eddd2aebeb27074c3ea6248c00
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:1b29e04308bb8e3c990a8c41a5d3d1e854e197bbdf0cffb2833fea4866772ba9
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90d45b7a3c5ffcd30a1ca1447693ba4aedfb400983cf49779a8421a42dc0eecd
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:60dbe69693314c937a69bd213de438c3af0a8a290aa9b1cfaad8a57e9d04c448
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:ab686e9e4c38b7909994f9fb496943b745e1a07492fe61a61b9e8a78f756eb1c
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:abec4867d6b9294d674cc02b9f7f254e65e7de8eea0231fa94f50513f2b7c45d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:b6500764d4f6d281f1b42458eb223d21f6a413b5db55f2a2ba1a7973aad8465c
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:3c18707afb90a7dc1bd2ed1fff05c53cce89b7c85062e72261f538a12834ab37
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:245f81b08980bfd4790cc3661e609e2cf53f3dd4d4c1f988e5db4dcf316d90fb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:d76476d2ecd88ea7b4fd857e591cc9b71be9d31fe0fb0057baa1b21e6113f730
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:186e849f3cf782f0d78660eb4433853e1e61296234d230a9dc5921fa4bd41060
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:6e055b6e07cb1adaa652b8e7515c4ce91438593a56ab0ae051435f55373d96ab
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:24dff64a80b910f0b4c5442d4f611f9df7602a16c5231eff2c9d08f07b2a9f9a
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:0b640bda3afa672011d5da390af798bbd3e6dd9126029a9eef03846d41582743
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:62dda55cc870279d491679f6b43be177986991b6eb85d9dfeba3ed2e72b3f10e
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:25c98c67d09cfaa3a4d332c461fb6766700a65876a48a1aabbd73b58bc2653f3
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:bd237f42078822b038c13271f35e6710273c2a4f3fe2f3eb74429c93cd2a0e41
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:bc39ff644108a458d1b7e29fcd88bf53982ab4070ce2ea0e75b0908678b9205d
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:21dffe4500172f74345fdafb9beb476e0ccffa55f7bca59f372f312f8d28d247
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:8f9d3438f58604cf29d643691b125a4138275b2c09cca4afc8e4b29289febd77
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:e8e0ea5cb1a070443011794988f2f6c9d67758605dc2847bf0183e19bfde52eb
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:9f835bb2984329db9db7eb2ac5521371ce0ced216a7875e45ef3ec31d255b648
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:234a8dcd8a9c5a68533bce6e0f793324e3ad7a660fdedf95895c5ed73a4cee3d
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:48828f128c98e73f8e345d87bc228f0920a3ff2cdc325d4a4ffe73ad95c5c5e0
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:ceee171c5ce0a8e580dfeccfe0dbfb73474c109becd45a562411678f3ab08595
        - registry.marketplace.redhat.com/cp/cpd/dap-base-job@sha256:f208686a3f56d2980e7b8640b6ebdea9057109e7520c6817a731e68d08506a36
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:31edab3e5662489dc2fe9528d078966663c8d49836425e648b7e112f871808dc
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:bcdf9b7eea8753b3e463973a2690140ae7aa8c02c3c7726511f116bd751ec2d8
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:63d4a9fa16231cc0c01f1b8bb9363221ac1e93f1571ea132866f170b156731e8
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:8fe85b93b9dcab9c33306ae358e6910685e109f32662d02e2b6bcd8dae243968
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:da0190fee910cb8245cf7a61d6e7aa2bf599d41c679eeedc183950918ea2611f
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:ad19b65042fb26cc1244a92f81a96e5ceb4313f8c823181efe28b66e9fecf9c8
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:78e33fff4bfc2e33ea179eaed9902df9bbd11463b9152d0281d1e0924fe09563
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-bundle@sha256:842f2ac568db4db79cf0450b413b4b09b5f2180ef3541385b46fb650ac7066c0
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:c43770715b76e0aa3fd9b3c2b21998ee369f2142e3c5d04320c677c9e9b29add
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:4eab846d6e80d68ce488e2cda36329cec1a6bf2f81feadf666b1005fed294e48
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:aa8d647f007d921db565a505d52583511e25de4a6a1187fc6b4df15427944364
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:276d5d06661ed401f0b0672dbcfa3d9f07fa2cde662fbb532fcdffbe3e332227
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:c6862158a2a5ae0a4e4206bfdc7f73e1c2359544966c72cf4fc3a617acb13697
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:4794a0fdae9d060def1e0d75a98560e713ca18de2625e4d9e5b5fa4c4877fac2
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9669ed5ba26ef32ad36a6784b2658553c00ac2a89cf4e9caddadf7bf46c1592e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:8305f26c2ebc493b54abfea0757d65b679f6c96bc3a0840c89f08a034327ed9c
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:0ae22914dbaf09d6679bd06bc9b6a5728e744c9ce2309cbd5f01e28308c403ea
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:209ea4b77b3b0863c591bcdec6f53c4c39dafcb10ff13e27ae4e1cb986a59727
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:483bc8f74d02efcbfb40ceb6abb0ab0b6581523cdea576b34512709b98879233
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:f40a9375fe5722972f995a82b50c00f254e3fe37f202b0aa1bebf66b0504099f
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:e8eee5083cbe317b368c002f47ce09f98e0c5e3407e5a4712e1c5473c4090ba3
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:523dcc671f1c05e690d5e578a90300d28d76570f605d4de7f80645b590f4a462
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:7ca8d085dd9aeb568eec2c4da974c03aa72bbfe4648f5d66a62d0444a2d1c839
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:9cd2ba7ab3ecab7094c892086a5276bbde22ff96d26ce0d087df2c1791235b02
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:ab99b62d40ba05cef8e699ebc7cc65271ea6fcb5900a76fe17526e7d55b50b08
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:a64a93104dee64e0e7e3e381144340511f903155f72b7f619c82c1d440712227
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:aa0068aec14e37986620c982295c083df6fc6bbe95facba582ae3c6f769a8e3a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:a10a15f07b856a7803ce0dcd30a542741dcb591804dc74c2a5fd50c55de0a983
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:10a935e41cc2dc09f8f3939cd512d9f6bc71d68457f77793a6957c5420cf4bf5
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:b87fa8374976ccd3dad90ddbf6b65cd6ae5a028ba896562f284f089380ebe44a
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:81d8b25221d827160dd5e88efd383453f532057ad02d7a2ee3dcd9c6eb16ba65
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:973eda4612c905661ff5ebc28c7cac6d9fe16fe5edaa544548d2311496ed938e
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:b82a10ce8f75d4441e869bf62fbfb5da349a2f228b52178bb29723af62f13938
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:5436a25b6c5b419b77643a76eb073bbbf8f09517299dbf4a992ae14c1965972c
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:40064d2893ce54b358c02023f4cc3603d3d28202ba56a177b20636abd74cb559
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:ae5e6858ed698d67b4b3917c0d41d0b4f7e2564585694f834099584c21a3c012
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:268c763fec7defe810624bbd31c96407fa37fef47b904758bce4c58d95aac043
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-bundle@sha256:94662b9dba5d4bafeec53b8ad334e7a1235385b21ac0952e12975f4f0a7623a3
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:c59837981b9a2a645c9ffd84c40d4f9ddc3797911f7a785ce4e3a2bfdd0bbe5c
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:bcd879a7cc3442ef62bb9677cc5534a2bfc9e7ea248e566aa4f1f6a0242431a8
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:b87b6c32a2050f1d5ff311d163a7b45efa7d5de1d352984b2f70fa578301a952
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:b9c8a65c1742c72c6351b1c8f4a858940c2bfa00ac5ec271c57b2ee7e5e7a66b
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:78682d9b9f19eae363cec0b6d31e696b2a5fd1bd1f8429df6f65a197310b6778
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:96ebe954a21f697ebaecfe5aa2c497c60b2ddec17e9ec4b376f435e6c53b3c29
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:21c81ff7c4b8809accaf62b242fd9b3d52f235489ae427556b7f71cc21bf8926
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-oauth2-proxy@sha256:bf4052b099393b563332acd5dcba6ba37ac424aa431d2d4f21532838b5ed6d6f
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:f5248b19c4a9b605a9160ecdb5dbbe0906841b667ede6fd6cf8e1d6c4369fc8e
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator-bundle@sha256:e4340bd9481da998a6b89e94c884c46e17cfb50d9d0c1099f39a3c54e500d2b2
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-oadp@sha256:9aec2bab8c7a6941e97c18352cb9e838a99aec08b06269e8053ac764638a90b9
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:0a32e127dd305fc445808d0469147db6c9704cd15d924d57d791e035475ecb60
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:4e3ec61a34fc7c71d213999b68b1f5c3f866f97e19c12fb001d506d4166a50f1
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:99d3b5a88d20380bcb4161e3b182f305af2e4cb405460d0f2a91965dd6f99f80
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:173ef1b032594c4e070e2e013db469115ec58d970fa8b62f3b748163b46d6d3e
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:a65ed5ac55cd5be99cc43c5a8d33bf127e8694917fce5f83fb8401c04e31b1f1
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:83ca1235dc054bf9ead64932c85333ea0dbade076b4a0cd1c3fe2b6bc2e05896
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:4ccd786884e39641a74096d783742c219fb277caafe80b667a80396c3a9cf6c3
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:6cffef7631ee54612fba71efc0eff4e17bc08defc3e2137aee9881d685444298
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:a3dc5fcff6a2aeee7c7e562fe9f6e0cf0652fd7ad4ec61496021d366206e2544
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:e296fa15dc14ee827d5046e8b1434d7f6838059c41d355471e1275e09f087411
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:68a98f52b7eba7e41c78aa91602c4e88d21d4a2a2a11b9b7592f181ce30e41ea
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:dd76a8484628970b62238334e122768aeea820ab18e24d62d23a6328c83b868c
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:e9e408a13bd103fb46536f896c51e7a27cba38d9b224231b20328a28efe1e616
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:9ced60f18740063ab281b836d4caf21035242ac3f659c6bb022ddb546770bff9
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:ebcfc8941330da6991480dc18ac34d6df4fe879e562ee537e5c36df73f4bc536
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:bdbbd76c93f0fb8f838edd9f124b02cc646a6f735a68a430a00f025a06c5f407
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:c0b6ad3b88f216379fc8f14aa5f075f5adc06ad06cc46495d3a40cc765f3deba
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:bb2f19e1797a900d60fb9f05d9f115faf20d514000aa3acac55fdba701581d81
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:93f73afa23f9fe8b73fd389081281fe1b0fd74bba6aa9bf7a6df663053afd3b8
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:54d8ceaaa06196e1ff524b7be595cbe90811f3be1a57c767b1ba74f53edb2f0e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:20f05c662659be9b2b54407c02f1805b93f09a3c204011981b7f17a76caad8ee
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:758f9e49dccaec5ebd125fc0c383fb5a656ba9f2f301984f993d50b103208d12
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:7655f531c354cfcce438c61cc5cde2c7690bfee4e2eadf9f99e4b157f163cd73
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:66c11636fdf68631e37bda7467536093b9402e883999e7989dbdcc4b4d2ba7d7
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:0621eb9b83c9b2b8bfe5cd09797fe08fa37a854a5506914935ff6f0df599b631
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:57e3433f860ec2b65f5c5f86574614fa9273c09b6b37db96a9e1888f4f6932f4
        - registry.marketplace.redhat.com/cp/cpd/ngp-projects-api@sha256:df420e9f616ad8a123d9eedd117fb6cb007b30a772138db534f47ad8801fa96d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:6d903fd4bd0ef3ef361d22b209ba7b686601b49d6f8d4e29b5d030da8d949dd5
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:d904761e70d3969a4ccacd71a60ad4d77dda939c556615df8d65782ccc3273aa
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:74aadd4fe99e66b89471222f853e442fa8338a23d9c24b4e908d38536c66456a
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:c91fc0992cff4b674f5e3690398e7dfb70662c807b930b9b0e65ba507ff925d2
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:24ca69c2f80a4369e29aa382a94d64fd3546b110787df1142c45a20b8246e844
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:9021a04805bb7f21d68fc6211f844a21c1d06f700c332ec3238073065b7c224b
        - registry.marketplace.redhat.com/cpopen/cpd/cpdtool@sha256:714caed1042ee03461efc9256a3da1046bd8f43819caf3a91ec38b118d459869
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:2346f67ffdc1a802720b325e010018cdb908aa458f372fe50ac6d8a92d8fb733
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:ef38f12712c406f8206742dffae9378e344d66c774e5b37b4fcb0b657db456ca
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:7505eb74c6bf165d40f6a8f3511f8e891cc34c0201fee38240c693a08e002ab1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8da8b240176c00f8f6eadf32e5614c663cc1d56d546c73996cca614b715acf4a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:fd35c33c7361bf04a66e69112b05a078ad65fdbe8f8ca8b7f4ecc2143b1643e2
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:67a02a9264d9c6e82edc15ae36d4f99b3fa9099697c550088d27c76511189f37
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:25a4bb9cea1c3bdbeeeb205622512bcba98163f9629eba35625756346ce9d652
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:860c17d2b2c6b0e013d0f4c035a388e45d8ef1483f53aa7ad022913b025f73b0
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:7d8fb7d8e2551bf949edd33589393fd12f634871b486e26464cd089e3e4b429d
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9186c350fa54e8b792d72e47be6b54a67b610f726175c9641795010d45ad51f9
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:f289ac0d44803ff12d41c3659e6fdf04c404e0b3b233b6bc44eec671abb1b47e
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:c8d2825596b7e675c247158bdd06911e1cc0ad9af3364c9c175a5b4cf16d5d8e
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:f42395ea365edc2c1bbef914a55b091e6442d17bd87bf3730a8f9f6d401bf4f5
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:8160c688f6b66206f8428f78dcf5e21b65c3a1cadc38b765f22ba8403a87f4d6
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:aea8fcb72664a41c3881839086a47fef1b54dd672300d4aaf583fd0b575a64b4
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:b341d9af40ad6aceb4ecc792ec75295313128a1eea3a0974aac879aed24a89ba
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:838f7649df5a7c13d6155394e211ac565aa86233bd30a3a2aab9706ec4a74eec
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:8542b9e449b054e455abf6e7884dcfda708a6213015e932844568d1853b65eb2
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:3af0967b0fb0a28a83a3e007b2c78fe27ad56a5e9c153806c6a7b17d7d9b1df5
        version: 6.1.3
      entries:
      - name: cpd-platform-operator.v6.1.3
        version: 6.1.3
      name: v6.1
    defaultChannel: v6.1
    packageName: cpd-platform-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Arkmq-org
      provider-url: https://arkmq-org.io
    name: activemq-artemis-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: activemq-artemis-operator.v2.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/arkmq-org/activemq-artemis-operator:2.0.5
          createdAt: "2025-04-04 14:35:00"
          description: An operator for managing the Apache ActiveMQ Artemis message
            broker
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/arkmq-org/activemq-artemis-operator
          support: arkmq-org
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: ArkMQ Operator manages ActiveMQ Artemis messaging broker in the
          cloud
        displayName: ArkMQ Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - messaging
        - cloud
        - kubernetes
        - activemq
        - artemis
        - integration
        - operator
        links:
        - name: Documentation
          url: https://arkmq-org.io/docs
        - name: Repository
          url: https://github.com/arkmq-org/activemq-artemis-operator
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Arkmq-org
          url: https://arkmq-org.io
        relatedImages:
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:4bba8828e72fd7c9d38c22ba1c2a4f7130817a4e7eac8fb63c13e86465a8e1e7
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:b71d1a57c8b3820a8133cf7a25a6edde7b29e1c4d32d7c30325183415a8bacd1
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:49f34122ab46ab5d7ead094b6e40817b62c75b6a809f67db07d68bce750da51e
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:a4a0d0453f0761b9e114d0d4cc7c14091a2a9c30790f18ffa6bf1e365d5083fe
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:93168fd8cd4fb32f041795afb8fb0e21cf81081d6cc463a44356e153dc1ad032
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:e2438670831bad39bb08147bbebe67ef9465899fb7750f1a5a185569f99d5fb5
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:8fbdd1ba728058d5f437186b9055ca28596140068ebc69ecff4fa52717b19a82
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:81590ec33b2f00775d3c6596f610887b670de7c0b89148b09eab7584f5b0b0ec
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:ddaf00914e3a412d21ec3a814e24f3d4cc382b5c4d6c21e9b1b94dde16ddf194
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:2895b2cf6292c951a0cdde78e755514f5d7cb43ce9f7b6e7dc381d8dfeec31ab
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.7
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.0
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:3bbfb6177e4839087f615a01ee54f9e06a95acaddbb083e11851b6af9315dbf0
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.0
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.0
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:42230ce9d11a1949a5bb73b6ee218b1581f585bcd553485ee438e1c5b50fcf8b
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.10
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.7
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:8f592b2454d092c82b5f61b4e04c3aff173db246539b3c06e15a71960d3bd235
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:f10983c68aff358e285d89d827c27b3487a3ee3b8e121cdaa295bf5fa8501145
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:b293c1269166cc6181f68b8d803fdfcf20fdbd5df2787ed0453724b22e298fb2
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:f25db980c3229e66cf48690aa175f6f93e3dce3974c0e70ab23af6462e7d1893
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:9bbb915855f29ea16950a3fd50d0b9bac25f82cba7d197579e04d9c33a5b02c4
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:de82344daa0caeabc6559656da0af285436a687e8f371744c0c51bdbc309167d
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:e438bcacf099a92b143793e88c00c6ac0afb31effe4b0eed49a935c9265ad943
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:2d155bffd246f26f516a6075ec170babc1373c72bb664cba4e3c58521ffd3558
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:264b33a9aad90d32a41258904f35092a141d76dcef0c115d414475444100aace
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:56bae938a6b4cd87f9f5284931f440f9b9d116cda3b63b13f83a4b1c6c641959
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:da95bff69deb63545d9782ed20cbdd17c1a79c6b16165580bf96a1d5191e6a59
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:f2c853ff91809282c04d5b1381ceae366518bf07b16649a61aa6c60c47f10714
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.2
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:b6f61188379def8c796c738549c099bd40bcd9bb44f5c30682c5edd136347aca
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:fc9b9a4c7550d84fd90335dd4f9e0d8d39c09136c6034b28490a368368cb11f1
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.2
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:9e1fefc9a94d72a7b5a3f4e2d53ebd23e5ab221df17a5afde42ce973277838f6
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.1
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.7
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:634f3d8fd620607b56e0360a191d23a9cb8a32ba0ba1f24fe20791ec52343586
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:4be60cfa2a300b2eef9b5f10ea756bd3f55d75fbd8890232d907f8d64080449b
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:37ab7354273f4dcd19bad6936f10e92220d98b8c0266002221477a0e2d146c81
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:719178251893def955e444e706ff07042ea98d294671315d95da504c1f47de24
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.4
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:70d4c799c7d905fac27457d63e662ea72bca9e899a692e425609e4eaa45129e6
        - quay.io/arkmq-org/activemq-artemis-operator:2.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:5e768aefade6d18cd65cb1eb5ee233844f9f7e7b52631f93a0982b67d382239d
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:ffa0d7cb4068df84319776dbfb66210297da1b8466ccd5451318fc3de79736c3
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:6f2da9ecd8f5021964d0dfa85144830efcfbd0462d18f18f036971e5f17b3bb6
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:decb3932ebb5b848b90a4a23b083f0cb5237eec2844095d8c58a6798da8bb999
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:88a0098a628fe43a9b422f8ebb774086fb65ed389a195e2779482a4a63447e28
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:a4628c2591ec06caf21a3467f6e3faeaede7fa402b4f137d943b25b4dd4ad343
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.2
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:52f337cd5d7d5d199eebf4fc722f8c5279f652506f009efbc0f8f6de88e18718
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:1fd99a46758cd29fce846d0d590021758b2a7a3efedb817bb205359caa7d4f67
        version: 2.0.5
      entries:
      - name: activemq-artemis-operator.v2.0.5
        version: 2.0.5
      name: stable
    defaultChannel: stable
    packageName: activemq-artemis-operator
    provider:
      name: Arkmq-org
      url: https://arkmq-org.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: ScaleOps
      provider-url: https://scaleops.com
    name: scaleops-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: scaleops-operator.v1.7.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1alpha1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops-sample"
                },
                "spec": {
                  "scaleopsToken": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: ghcr.io/scaleops-sh/operator-bundle@sha256:ee627042deefa4faceb63efd56aa6f6338c90f6c13877f29d7161efe603e02f4
          createdAt: "2024-09-19T07:04:10Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: ScaleOps
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1alpha1
        description: ScaleOps Cluster Optimization
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: vlad@scaleops.com
          name: Vladislav Shub
        - email: guy@scaleops.com
          name: Guy Baron
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - ghcr.io/scaleops-sh/operator@sha256:7de0651a76d6eb63fee81e9177df93f126906d50063d7891f12c4a72c817dc86
        - ghcr.io/scaleops-sh/kube-state-metrics/ubi@sha256:d8cb65795861ce257f08726b7a3aa60a869cc44ed152f559431191bc79e98908
        - ghcr.io/scaleops-sh/operator-bundle@sha256:ee627042deefa4faceb63efd56aa6f6338c90f6c13877f29d7161efe603e02f4
        - ghcr.io/scaleops-sh/scaleops-operator-bundle@sha256:128771d92c852f7ace78103493db079a474e06d1352f06cbde7a7d4666684610
        - ghcr.io/scaleops-sh/configmap-reload/ubi@sha256:fed19f1aad43228560912abc36372a6c124d9ddf5e3e39476269dc51604d26de
        - ghcr.io/scaleops-sh/prometheus/ubi@sha256:8aeeae2a33251493d4d6e2f02a2fff1d4c2f32ab54554e6a84619ecd8b4aa69a
        version: 0.0.1
      entries:
      - name: scaleops-operator.v1.7.5
        version: 0.0.1
      name: alpha
    - currentCSV: scaleops-operator.v1.21.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops",
                  "namespace": "scaleops-system"
                },
                "spec": {
                  "scaleopsToken": "${SCALEOPS_TOKEN}"
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: registry.connect.redhat.com/scaleops-sh/operator@sha256:528f2e09d2dad4f48e01a272b51baa1753c12069515831223e3f97c7dfe03e3c
          createdAt: "2025-09-29T13:35:24Z"
          description: |-
            ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

            ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

            Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

            Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

            Contact [us](mailto:team@scaleops.com) to start your free trial today
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.21.8
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://scaleops.com
          support: ScaleOps Support support@scaleops.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This will install an instance of the ScaleOps Platform on
              your cluster
            displayName: ScaleOps Platform
            kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1
        description: |-
          ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

          ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

          Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

          Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

          Contact [us](mailto:team@scaleops.com) to start your free trial today
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        - application runtime
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: support@scaleops.com
          name: ScaleOps
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - registry.connect.redhat.com/scaleops-sh/operator@sha256:528f2e09d2dad4f48e01a272b51baa1753c12069515831223e3f97c7dfe03e3c
        - registry.connect.redhat.com/scaleops-sh/scaleops-sh@sha256:57d5d06f4f4b5470562c4f24ea5d37203633e1f4b65cd497e71d9066fd2b9e41
        - registry.connect.redhat.com/scaleops-sh/scaleops-gpu@sha256:724aef7683ea4cb245ab0a464517a6d542efa4634e5469ff35b7a7b7f00b563e
        version: 1.21.8
      entries:
      - name: scaleops-operator.v1.21.8
        version: 1.21.8
      - name: scaleops-operator.v1.21.7
        version: 1.21.7
      - name: scaleops-operator.v1.21.6
        version: 1.21.6
      - name: scaleops-operator.v1.21.5
        version: 1.21.5
      - name: scaleops-operator.v1.21.3
        version: 1.21.3
      - name: scaleops-operator.v1.21.2
        version: 1.21.2
      - name: scaleops-operator.v1.21.1
        version: 1.21.1
      - name: scaleops-operator.v1.21.0
        version: 1.21.0
      - name: scaleops-operator.v1.20.9
        version: 1.20.9
      - name: scaleops-operator.v1.20.8
        version: 1.20.8
      - name: scaleops-operator.v1.20.7
        version: 1.20.7
      - name: scaleops-operator.v1.20.6
        version: 1.20.6
      - name: scaleops-operator.v1.20.5
        version: 1.20.5
      - name: scaleops-operator.v1.20.4
        version: 1.20.4
      - name: scaleops-operator.v1.20.2
        version: 1.20.2
      - name: scaleops-operator.v1.20.1
        version: 1.20.1
      - name: scaleops-operator.v1.19.8
        version: 1.19.8
      - name: scaleops-operator.v1.19.7
        version: 1.19.7
      - name: scaleops-operator.v1.19.6
        version: 1.19.6
      - name: scaleops-operator.v1.19.5
        version: 1.19.5
      - name: scaleops-operator.v1.19.4
        version: 1.19.4
      - name: scaleops-operator.v1.19.1
        version: 1.19.1
      - name: scaleops-operator.v1.19.0
        version: 1.19.0
      - name: scaleops-operator.v1.18.8
        version: 1.18.8
      - name: scaleops-operator.v1.18.7
        version: 1.18.7
      - name: scaleops-operator.v1.18.6
        version: 1.18.6
      - name: scaleops-operator.v1.18.5
        version: 1.18.5
      - name: scaleops-operator.v1.18.4
        version: 1.18.4
      - name: scaleops-operator.v1.18.3
        version: 1.18.3
      - name: scaleops-operator.v1.18.2
        version: 1.18.2
      name: stable
    - currentCSV: scaleops-operator.v1.10.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops",
                  "namespace": "scaleops-system"
                },
                "spec": {
                  "scaleopsToken": "${SCALEOPS_TOKEN}"
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: registry.connect.redhat.com/scaleops-sh/operator@sha256:8def662a546babd6740e5417401836d647d92d52b15d559b1b3e2cebd7f77f33
          createdAt: "2025-02-12T19:06:04Z"
          description: |-
            ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

            ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

            Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

            Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

            Contact [us](mailto:team@scaleops.com) to start your free trial today
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.10.32
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://scaleops.com
          support: ScaleOps Support support@scaleops.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This will install an instance of the ScaleOps Platform on
              your cluster
            displayName: ScaleOps Platform
            kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1
        description: |-
          ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

          ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

          Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

          Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

          Contact [us](mailto:team@scaleops.com) to start your free trial today
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        - application runtime
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: support@scaleops.com
          name: ScaleOps
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - registry.connect.redhat.com/scaleops-sh/operator@sha256:8def662a546babd6740e5417401836d647d92d52b15d559b1b3e2cebd7f77f33
        - registry.connect.redhat.com/scaleops-sh/scaleops-sh@sha256:5d77e81d7959483121c208072eb61b063eb3ad9abc49e7df3c7efb8e2b77a3ae
        version: 1.10.32
      entries:
      - name: scaleops-operator.v1.10.32
        version: 1.10.32
      name: stable-1.10
    - currentCSV: scaleops-operator.v1.12.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops",
                  "namespace": "scaleops-system"
                },
                "spec": {
                  "scaleopsToken": "${SCALEOPS_TOKEN}"
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: registry.connect.redhat.com/scaleops-sh/operator@sha256:4bd74a56a8680f47aa1da7a377f74529fe7ca1568995a4e74018af0cf67763ed
          createdAt: "2025-03-11T18:10:07Z"
          description: |-
            ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

            ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

            Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

            Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

            Contact [us](mailto:team@scaleops.com) to start your free trial today
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.12.6
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://scaleops.com
          support: ScaleOps Support support@scaleops.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This will install an instance of the ScaleOps Platform on
              your cluster
            displayName: ScaleOps Platform
            kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1
        description: |-
          ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

          ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

          Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

          Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

          Contact [us](mailto:team@scaleops.com) to start your free trial today
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        - application runtime
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: support@scaleops.com
          name: ScaleOps
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - registry.connect.redhat.com/scaleops-sh/operator@sha256:4bd74a56a8680f47aa1da7a377f74529fe7ca1568995a4e74018af0cf67763ed
        - registry.connect.redhat.com/scaleops-sh/scaleops-sh@sha256:5ea4e26f2fecee3570d63adc3e2588d34f9610345d703975d4febac9c94b4bd0
        version: 1.12.6
      entries:
      - name: scaleops-operator.v1.12.6
        version: 1.12.6
      name: stable-1.12
    - currentCSV: scaleops-operator.v1.14.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops",
                  "namespace": "scaleops-system"
                },
                "spec": {
                  "scaleopsToken": "${SCALEOPS_TOKEN}"
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: registry.connect.redhat.com/scaleops-sh/operator@sha256:9267f49ce774ee4d401458531599bbd6729c6313cecc4f2bcd3bf0b5eecd92e1
          createdAt: "2025-06-08T16:15:46Z"
          description: |-
            ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

            ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

            Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

            Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

            Contact [us](mailto:team@scaleops.com) to start your free trial today
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.14.14
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://scaleops.com
          support: ScaleOps Support support@scaleops.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This will install an instance of the ScaleOps Platform on
              your cluster
            displayName: ScaleOps Platform
            kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1
        description: |-
          ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

          ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

          Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

          Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

          Contact [us](mailto:team@scaleops.com) to start your free trial today
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        - application runtime
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: support@scaleops.com
          name: ScaleOps
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - registry.connect.redhat.com/scaleops-sh/operator@sha256:9267f49ce774ee4d401458531599bbd6729c6313cecc4f2bcd3bf0b5eecd92e1
        - registry.connect.redhat.com/scaleops-sh/scaleops-sh@sha256:9c6395202ccb70e8edae3a1b540de31991ec9232b7b235dd92755588b8491720
        version: 1.14.14
      entries:
      - name: scaleops-operator.v1.14.14
        version: 1.14.14
      name: stable-1.14
    defaultChannel: stable
    packageName: scaleops-operator
    provider:
      name: ScaleOps
      url: https://scaleops.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataStax
      provider-url: ""
    name: cass-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cass-operator.v1.23.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cassandra.datastax.com/v1beta1",
                "kind": "CassandraDatacenter",
                "metadata": {
                  "name": "dc1"
                },
                "spec": {
                  "clusterName": "development",
                  "config": {
                    "cassandra-yaml": {
                      "authenticator": "PasswordAuthenticator",
                      "authorizer": "CassandraAuthorizer",
                      "num_tokens": 16,
                      "role_manager": "CassandraRoleManager"
                    },
                    "jvm-server-options": {
                      "initial_heap_size": "1G",
                      "max_heap_size": "1G"
                    }
                  },
                  "managementApiAuth": {
                    "insecure": {}
                  },
                  "podTemplateSpec": {
                    "spec": {
                      "containers": [
                        {
                          "name": "cassandra"
                        }
                      ],
                      "securityContext": {}
                    }
                  },
                  "racks": [
                    {
                      "name": "rack1"
                    },
                    {
                      "name": "rack2"
                    },
                    {
                      "name": "rack3"
                    }
                  ],
                  "resources": {
                    "requests": {
                      "cpu": "1000m",
                      "memory": "2Gi"
                    }
                  },
                  "serverType": "cassandra",
                  "serverVersion": "4.1.2",
                  "size": 3,
                  "storageConfig": {
                    "cassandraDataVolumeClaimSpec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "server-storage"
                    }
                  }
                }
              },
              {
                "apiVersion": "control.k8ssandra.io/v1alpha1",
                "kind": "CassandraTask",
                "metadata": {
                  "name": "example-cleanup"
                },
                "spec": {
                  "datacenter": {
                    "name": "dc2",
                    "namespace": "cass-operator"
                  },
                  "jobs": [
                    {
                      "command": "cleanup",
                      "name": "cleanup-run"
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database,Big Data,OpenShift Optional
          certified: "true"
          containerImage: registry.connect.redhat.com/datastax/cass-operator@sha256:6abc6253b50a0d7f6bdd4b394d2bf96320a8845952aeb8f0bbed99863192ec52
          createdAt: "2025-03-20"
          description: |
            Simple provisioning, turn-key operations, and automated remediation of Apache Cassandra clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/k8ssandra/cass-operator
          support: DataStax
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CassandraDatacenter is the Schema for the cassandradatacenters
              API
            displayName: Cassandra Datacenter
            kind: CassandraDatacenter
            name: cassandradatacenters.cassandra.datastax.com
            version: v1beta1
          - description: CassandraTask is the Schema for the cassandrajobs API
            displayName: Cassandra Task
            kind: CassandraTask
            name: cassandratasks.control.k8ssandra.io
            version: v1alpha1
        description: "## Apache Cassandra\n\nThe Apache Cassandra database is the
          right choice when you need scalability and\nhigh availability without compromising
          performance. Linear scalability and\nproven fault-tolerance on commodity
          hardware or cloud infrastructure make it the\nperfect platform for mission-critical
          data. Cassandra's support for replicating\nacross multiple datacenters is
          best-in-class, providing lower latency for your\nusers and the peace of
          mind of knowing that you can survive regional outages.\n\n## DataStax Enterprise\n\nThe
          most advanced distribution of Apache Cassandra™ on the market, with the\nenterprise
          functionality needed for serious production systems and backed up and\nsupported
          by the best distributed-experts in the world. It's one platform for\nall
          types of applications anywhere, any cloud, any model: key-value, graph,\ntabular,
          JSON.\n\nDataStax Enterprise is a fully integrated and optimized database,
          with graph,\nanalytics, and search included, all with a unified security
          model. Simply put,\nit's the only database capable of meeting today's demanding
          requirements\n\n## Operator Details\n\n`cass-operator` is designed as a
          modular operator for Apache Cassandra and\nderived  distributions. Apache
          Cassandra is a distributed database consisting of\nmultiple nodes working
          in concert to store data and process queries along a\nnumber of fault domains.
          `cass-operator` has the deployment of a Cassandra\ncluster around the logical
          domain of a datacenter with the `CassandraDatacenter`\ncustom resource.\n
          \   \nUpon submission of one of these resources it handles provisioning
          the underlying\nstateful sets (analogous to C\\* logical racks), services,
          and configuration.\nAdditionally through monitoring pod state via Kubernetes
          callbacks it handles day to day\noperations such as restarting failed processes,
          scaling clusters up, and deploying\nconfiguration changes in a rolling,
          non-disruptive, fashion.\n    \nThis operator is designed to be `Namespace`
          scoped. A single Kubernetes cluster may\nbe running multiple instances of
          this operator, in separate namespaces, to support\na number of C\\* clusters
          and environments. Configuration is simple with the usage of\nYAML based
          overrides in the Custom Resource paired with an `init` container.\n    \nIn
          C\\* clusters ordering and timing of certain operations are important to
          keep the system\nevenly distributed. `cass-operator` takes advantage of
          a sidecar process within the\nmain container to handle the orchestration
          of starting our main C* process."
        displayName: DataStax Kubernetes Operator for Apache Cassandra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - k8ssandra
        - cass-operator
        - DataStax Enterprise
        - Apache Cassandra
        - DSE
        - C*
        - Cassandra
        links:
        - name: DataStax
          url: https://datastax.com
        - name: Operator GitHub Repo
          url: https://github.com/k8ssandra/cass-operator
        - name: Config Builder GitHub Repo
          url: https://github.com/datastax/cass-config-builder
        - name: Documentation [1]
          url: https://github.com/k8ssandra/cass-operator/tree/master/docs/user
        - name: Documentation [2]
          url: https://docs.datastax.com/en/cass-operator/doc/cass-operator/cassOperatorTOC.html
        maintainers:
        - email: cass-operator@datastax.com
          name: DataStax Operator Team
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: DataStax
        relatedImages:
        - registry.connect.redhat.com/datastax/cass-operator@sha256:6abc6253b50a0d7f6bdd4b394d2bf96320a8845952aeb8f0bbed99863192ec52
        - registry.connect.redhat.com/datastax/system-logger@sha256:3408620786e6265aa835be56f85aadc48b20060b2038dc21126f614685f77c6b
        version: 1.23.2
      entries:
      - name: cass-operator.v1.23.2
        version: 1.23.2
      name: stable
    defaultChannel: stable
    packageName: cass-operator
    provider:
      name: DataStax
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
    name: aikit-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aikit-operator.v2021.2.102120
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aikit.intel/v1alpha1",
                "kind": "AIKitContainer",
                "metadata": {
                  "name": "intel-aikit-container",
                  "namespace": "redhat-ods-applications"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagestream": {
                    "name": "oneapi-aikit",
                    "namespace": "redhat-ods-applications",
                    "registry": {
                      "root": "registry.connect.redhat.com",
                      "repo": "intel",
                      "name": "oneapi-aikit",
                      "version": "latest"
                    }
                  },
                  "nameOverride": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aikit-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aikit-operator-rhmp/support?utm_source=openshift_console
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AIKitContainer
            name: aikitcontainers.aikit.intel
            version: v1alpha1
        description: |
          The Intel® oneAPI AI Analytics Toolkit gives data scientists, AI developers, and researchers familiar Python* tools and frameworks to accelerate end-to-end data science and machine learning pipelines on Intel® architectures. The components are built using oneAPI libraries for low-level compute optimizations. This toolkit maximizes performance from preprocessing through machine and deep learning phases and provides interoperability for efficient model development and deployment across single and multinodes.
        displayName: Intel® oneAPI AI Analytics Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aikit-operator
        links:
        - name: Aikit Operator
          url: https://aikit-operator.domain
        maintainers:
        - email: kam.d.kasravi@intel.com
          name: Kam Kasravi
        - email: abolfazl.shahbazi@intel.com
          name: Abolfazl Shahbazi
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Intel
          url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
        relatedImages:
        - registry.marketplace.redhat.com/rhm/intel/aikit-operator@sha256:7abb5e6164ffb3e4b87c39061f45d09848218fcfb74b48bf5e1035d56b41ee0c
        version: 2021.2.102120
      entries:
      - name: aikit-operator.v2021.2.102120
        version: 2021.2.102120
      - name: aikit-operator.v2021.2.100410
        version: 2021.2.100410
      - name: aikit-operator.v2021.2.92911
        version: 2021.2.92911
      - name: aikit-operator.v2021.2.92814
        version: 2021.2.92814
      - name: aikit-operator.v2021.2.92808-dev-a60e5ba
        version: 2021.2.92808-dev-a60e5ba
      name: alpha
    defaultChannel: alpha
    packageName: aikit-operator-rhmp
    provider:
      name: Intel
      url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cluster-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-manager.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "cluster-manager"
                },
                "spec": {
                  "addOnManagerImagePullSpec": "quay.io/open-cluster-management/addon-manager:v0.13.1",
                  "deployOption": {
                    "mode": "Default"
                  },
                  "placementImagePullSpec": "quay.io/open-cluster-management/placement:v0.13.1",
                  "registrationConfiguration": {
                    "featureGates": [
                      {
                        "feature": "DefaultClusterSet",
                        "mode": "Enable"
                      }
                    ]
                  },
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "cluster-manager"
                },
                "spec": {
                  "addOnManagerImagePullSpec": "quay.io/open-cluster-management/addon-manager:v0.13.1",
                  "deployOption": {
                    "hosted": {
                      "registrationWebhookConfiguration": {
                        "address": "management-control-plane",
                        "port": 30443
                      },
                      "workWebhookConfiguration": {
                        "address": "management-control-plane",
                        "port": 31443
                      }
                    },
                    "mode": "Hosted"
                  },
                  "placementImagePullSpec": "quay.io/open-cluster-management/placement:v0.13.1",
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/registration-operator:v0.13.1
          createdAt: "2024-04-10T15:27:49Z"
          description: Manages the installation and upgrade of the ClusterManager.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management-io/ocm
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterManager defines the configurations of controllers
              running hub that govern registration and work distribution for attached
              Klusterlets
            displayName: ClusterManager
            kind: ClusterManager
            name: clustermanagers.operator.open-cluster-management.io
            version: v1
        description: "Note: This operator will stop updating, please use the latest
          operator in OperatorHub.io https://operatorhub.io/operator/cluster-manager.
          \n\nThe Cluster Manager provides the multicluster hub, which can manage
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. This operator supports the installation and upgrade of ClusterManager.\n\n##
          How to Install\nInstall ClusterManager operator by following instructions
          in top right button `Install`. After installing the operator, create an
          instance of the ClusterManager resource to install the ClusterManager.\n\nThe
          pods of ClusterManager are created in the `open-cluster-management-hub`
          namespace:\n\n```\n$ kubectl get pods -n open-cluster-management-hub\nNAME
          \                                                      READY   STATUS    RESTARTS
          \  AGE\ncluster-manager-addon-manager-controller-7fcc67bfdf-x7hbz  1/1     Running
          \  0          61m\ncluster-manager-placement-controller-dbcfd4975-26jpw
          \      1/1     Running   0          61m\ncluster-manager-registration-controller-7949d97899-wwtg4
          \  1/1     Running   0          61m\ncluster-manager-registration-webhook-7f675bc4bb-v4rhj
          \     1/1     Running   0          61m\ncluster-manager-work-webhook-6c9d65fb95-jh97b
          \             1/1     Running   0          61m\n```\n\nThere is a ManagedCluster
          created in the Hub cluster when the managed cluster is registered to the
          Hub cluster.\n\n```\n$ kubectl get managedclusters.cluster.open-cluster-management.io\nNAME
          \                  HUB ACCEPTED   MANAGED CLUSTER URLS     JOINED   AVAILABLE
          \  AGE\ncluster-dev            true                                    True
          \    True        7h20m\ncluster-test           false          https://127.0.0.1:8443
          \                       7h25m\n```\n"
        displayName: Cluster Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open-cluster-management
        - ClusterManager
        links:
        - name: Registration Operator
          url: https://github.com/open-cluster-management-io/ocm
        maintainers:
        - email: zyin@redhat.com
          name: Zhiwei Yin
        - email: deads@redhat.com
          name: David Eads
        - email: jqiu@redhat.com
          name: Jian Qiu
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/open-cluster-management/registration-operator:v0.13.1
        version: 0.13.1
      entries:
      - name: cluster-manager.v0.13.1
        version: 0.13.1
      - name: cluster-manager.v0.13.0
        version: 0.13.0
      - name: cluster-manager.v0.12.0
        version: 0.12.0
      - name: cluster-manager.v0.11.0
        version: 0.11.0
      - name: cluster-manager.v0.10.0
        version: 0.10.0
      - name: cluster-manager.v0.9.1
        version: 0.9.1
      - name: cluster-manager.v0.8.0
        version: 0.8.0
      - name: cluster-manager.v0.7.0
        version: 0.7.0
      - name: cluster-manager.v0.6.0
        version: 0.6.0
      - name: cluster-manager.v0.5.0
        version: 0.5.0
      - name: cluster-manager.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: cluster-manager
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: cloudnative-pg
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudnative-pg.v1.27.0-rc1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: ghcr.io/cloudnative-pg/cloudnative-pg@sha256:c2212e806e950ed5b667455f476481aef179c8e467e435352b624dc670000443
          createdAt: "2025-07-31T12:14:29Z"
          description: CloudNativePG is an open source operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/cloudnative-pg/cloudnative-pg
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.cnpg.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.cnpg.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.cnpg.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.cnpg.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.cnpg.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.cnpg.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.cnpg.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnpg` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: CloudNativePG
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnpg
        links:
        - name: CloudNativePG
          url: https://cloudnative-pg.io/
        - name: Documentation
          url: https://cloudnative-pg.io/docs/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: The CloudNativePG Contributors
        relatedImages:
        - ghcr.io/cloudnative-pg/cloudnative-pg@sha256:c2212e806e950ed5b667455f476481aef179c8e467e435352b624dc670000443
        - ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b
        - ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891
        version: 1.27.0-rc1
      entries:
      - name: cloudnative-pg.v1.27.0-rc1
        version: 1.27.0-rc1
      - name: cloudnative-pg.v1.26.0-rc3
        version: 1.26.0-rc3
      - name: cloudnative-pg.v1.26.0-rc2
        version: 1.26.0-rc2
      - name: cloudnative-pg.v1.26.0-rc1
        version: 1.26.0-rc1
      name: candidate
    - currentCSV: cloudnative-pg.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: ghcr.io/cloudnative-pg/cloudnative-pg@sha256:53995e9d6a7760c77d77e5298f78b49029427494f4dd933d5fa7a74f7af3bb44
          createdAt: "2025-08-12T14:08:35Z"
          description: CloudNativePG is an open source operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/cloudnative-pg/cloudnative-pg
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.cnpg.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.cnpg.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.cnpg.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.cnpg.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.cnpg.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.cnpg.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.cnpg.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnpg` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: CloudNativePG
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnpg
        links:
        - name: CloudNativePG
          url: https://cloudnative-pg.io/
        - name: Documentation
          url: https://cloudnative-pg.io/docs/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - ghcr.io/cloudnative-pg/cloudnative-pg@sha256:53995e9d6a7760c77d77e5298f78b49029427494f4dd933d5fa7a74f7af3bb44
        - ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b
        - ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891
        version: 1.27.0
      entries:
      - name: cloudnative-pg.v1.27.0
        version: 1.27.0
      - name: cloudnative-pg.v1.26.1
        version: 1.26.1
      - name: cloudnative-pg.v1.26.0
        version: 1.26.0
      - name: cloudnative-pg.v1.25.1
        version: 1.25.1
      - name: cloudnative-pg.v1.25.0
        version: 1.25.0
      - name: cloudnative-pg.v1.24.1
        version: 1.24.1
      - name: cloudnative-pg.v1.24.0
        version: 1.24.0
      - name: cloudnative-pg.v1.23.3
        version: 1.23.3
      - name: cloudnative-pg.v1.23.2
        version: 1.23.2
      - name: cloudnative-pg.v1.23.1
        version: 1.23.1
      name: stable-v1
    defaultChannel: stable-v1
    packageName: cloudnative-pg
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cincinnati-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: update-service-operator.v5.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "updateservice.operator.openshift.io/v1",
                "kind": "UpdateService",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          description: Creates and maintains an OpenShift Update Service instance
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-update-service
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: UpdateService is the Schema for the updateservices API.
            displayName: Update Service
            kind: UpdateService
            name: updateservices.updateservice.operator.openshift.io
            version: v1
        description: |-
          # Use Case
          Running an Update Service instance in a cluster is appealing for offline OpenShift
          clusters or for admins that want to provide their own graph data instead of
          using [https://github.com/openshift/cincinnati-graph-data](https://github.com/openshift/cincinnati-graph-data/).

          # About Update Service
          Update Service uses **Cincinnati** which is an update protocol designed to facilitate
          automatic updates. It describes a particular method for representing transitions between
          releases of a project and allowing a client to perform automatic updates between these
          releases.

          ## Components
          A **Graph Builder** iterates over the release payloads hosted by the
          storage component and builds a DAG of the releases. It is responsible for
          verifying that the graph described by the releases is acyclic and connected.

          A **Policy Engine** is in charge of altering a client's view of the graph
          by applying a set of filters which are defined within the particular Policy
          Engine instance. Both the input to and the output from Policy Engines is a
          graph, allowing multiple Policy Engines to be chained together. The first
          Policy Engine in a chain will fetch its graph from the Graph Builder and the
          last Policy Engine in a chain will serve the modified graph to the client.

          An **Update Service client** is the end consumer of the release payloads. The
          client periodically queries the Policy Engine for updates and applys them if
          available.

          # Query OpenShift's Update Service Endpoint
          $ curl --silent --header 'Accept:application/json' 'https://api.openshift.com/api/upgrades_info/v1/graph?arch=amd64&channel=stable-4.2' | jq '. as $graph | $graph.nodes | map(.version == "4.2.13") | index(true) as $orig | $graph.edges | map(select(.[0] == $orig)[1]) | map($graph.nodes[.])'
        displayName: OpenShift Update Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Update Service
        - UpdateService
        - update-service
        - OSUS
        - Cincinnati
        - version
        - upgrade
        links:
        - name: Update Service Operator
          url: https://github.com/openshift/cincinnati-operator
        - name: OpenShift Update Service
          url: https://github.com/openshift/cincinnati
        - name: OpenShift Graph Data
          url: https://github.com/openshift/cincinnati-graph-data
        maintainers:
        - email: aos-team-ota@redhat.com
          name: OpenShift Update Service maintainers
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-update-service/openshift-update-service-rhel8-operator@sha256:759f4de107a5b429933c5e44046ffb4ed75c674aea75440513a6df8fa93a83a3
        - registry.redhat.io/openshift-update-service/openshift-update-service-rhel8@sha256:a9468a662babb272917df92bc9c827535b4b37883edb440b6d2d7c32ffb2145e
        version: 5.0.3
      entries:
      - name: update-service-operator.v5.0.3
        version: 5.0.3
      - name: update-service-operator.v5.0.2
        version: 5.0.2
      - name: update-service-operator.v5.0.1
        version: 5.0.1
      - name: update-service-operator.v5.0.0
        version: 5.0.0
      - name: update-service-operator.v4.9.1
        version: 4.9.1
      - name: update-service-operator.v4.9.0
        version: 4.9.0
      name: v1
    defaultChannel: v1
    packageName: cincinnati-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: www.redhat.com
    name: dpu-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dpu-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.openshift.io/v1",
                "kind": "DpuOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dpu-operator",
                    "app.kubernetes.io/instance": "dpuoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dpuoperatorconfig",
                    "app.kubernetes.io/part-of": "dpu-operator"
                  },
                  "name": "dpuoperatorconfig-sample"
                },
                "spec": {
                  "mode": "host"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-13T13:42:44Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: dpu-operator-system
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DpuOperatorConfig is the Schema for the dpuoperatorconfigs
              API
            displayName: Dpu Operator Config
            kind: DpuOperatorConfig
            name: dpuoperatorconfigs.config.openshift.io
            version: v1
        description: This operator will manage and configure data processing unit
          (DPUs) to be used in accelerating/offloading k8s networking functions
        displayName: DPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dpu
        - dpu-operator
        - ipu
        links:
        - name: Dpu Operator
          url: https://github.com/openshift/dpu-operator
        maintainers:
        - email: wizhao@redhat.com
          name: William Zhao
        - email: bnemeth@redhat.com
          name: Balazs Nemeth
        maturity: alpha
        provider:
          name: Red Hat
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-dpu-rhel9-operator@sha256:9d8283238af545c332781b2edef6b3a2857bcf4a264c9bae26720cc48e338a4b
        - registry.redhat.io/openshift4/ose-dpu-daemon-rhel9@sha256:a9a375f70dfb4a730aca0f7e58619fb05ef226261de6980d04eeffb8a043e6a4
        version: 4.16.0-202509240123
      entries:
      - name: dpu-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: dpu-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: dpu-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: dpu-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: dpu-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: dpu-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: dpu-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: dpu-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: dpu-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: dpu-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: dpu-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: dpu-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: dpu-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: dpu-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: dpu-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: dpu-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: dpu-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: dpu-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: dpu-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: dpu-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: dpu-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: dpu-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: dpu-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: dpu-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: dpu-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: dpu-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: dpu-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: dpu-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: dpu-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      name: stable
    defaultChannel: stable
    packageName: dpu-operator
    provider:
      name: Red Hat
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: odr-cluster-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odr-cluster-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "VolumeReplicationGroup",
                "metadata": {
                  "name": "volumereplicationgroup-sample"
                },
                "spec": {
                  "async": {
                    "schedulingInterval": "10m"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ],
                        "name": "cpd"
                      },
                      {
                        "includedResources": [
                          "Deployment"
                        ],
                        "name": "deployments"
                      },
                      {
                        "excludedResources": [
                          ""
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includeClusterResources": true,
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "cpd",
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      },
                      {
                        "backupName": "deployments",
                        "includedResources": [
                          "Deployment"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "excludedResources": [
                          "ConfigMap",
                          "Secret",
                          "Deployment",
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      }
                    ]
                  },
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  },
                  "replicationState": "primary",
                  "s3Profiles": [
                    "s3-profile-of-east",
                    "s3-profile-of-west"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: MaintenanceMode
            name: maintenancemodes.ramendr.openshift.io
            version: v1alpha1
          - kind: ProtectedVolumeReplicationGroupList
            name: protectedvolumereplicationgrouplists.ramendr.openshift.io
            version: v1alpha1
          - description: VolumeReplicationGroup is the Schema for the volumereplicationgroups
              API
            displayName: Volume Replication Group
            kind: VolumeReplicationGroup
            name: volumereplicationgroups.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Cluster is a disaster-recovery orchestrator for
          stateful applications, that operates from an Advanced Cluster Management
          (ACM) managed cluster and is controlled by Openshift DR Hub operator to
          orchestrate the life-cycle of an application, and its state on the managed
          cluster.
        displayName: Openshift DR Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:769acf15d016d4992793c9c527d7c268d2a67502dd7e0fadc240e4896398195b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 4.15.20-rhodf
      entries:
      - name: odr-cluster-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odr-cluster-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odr-cluster-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odr-cluster-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odr-cluster-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odr-cluster-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odr-cluster-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odr-cluster-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odr-cluster-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odr-cluster-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odr-cluster-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odr-cluster-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odr-cluster-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odr-cluster-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odr-cluster-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odr-cluster-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odr-cluster-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odr-cluster-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odr-cluster-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odr-cluster-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odr-cluster-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "VolumeReplicationGroup",
                "metadata": {
                  "name": "volumereplicationgroup-sample"
                },
                "spec": {
                  "async": {
                    "schedulingInterval": "10m"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ],
                        "name": "cpd"
                      },
                      {
                        "includedResources": [
                          "Deployment"
                        ],
                        "name": "deployments"
                      },
                      {
                        "excludedResources": [
                          ""
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includeClusterResources": true,
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "cpd",
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      },
                      {
                        "backupName": "deployments",
                        "includedResources": [
                          "Deployment"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "excludedResources": [
                          "ConfigMap",
                          "Secret",
                          "Deployment",
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      }
                    ]
                  },
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  },
                  "replicationState": "primary",
                  "s3Profiles": [
                    "s3-profile-of-east",
                    "s3-profile-of-west"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: MaintenanceMode
            name: maintenancemodes.ramendr.openshift.io
            version: v1alpha1
          - kind: ProtectedVolumeReplicationGroupList
            name: protectedvolumereplicationgrouplists.ramendr.openshift.io
            version: v1alpha1
          - description: VolumeReplicationGroup is the Schema for the volumereplicationgroups
              API
            displayName: Volume Replication Group
            kind: VolumeReplicationGroup
            name: volumereplicationgroups.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Cluster is a disaster-recovery orchestrator for
          stateful applications, that operates from an Advanced Cluster Management
          (ACM) managed cluster and is controlled by Openshift DR Hub operator to
          orchestrate the life-cycle of an application, and its state on the managed
          cluster.
        displayName: Openshift DR Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:af262b0890300183db1766bc392017b0bda723a9ba841dc25e3d928aa413db9a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        version: 4.16.16-rhodf
      entries:
      - name: odr-cluster-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odr-cluster-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odr-cluster-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odr-cluster-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odr-cluster-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odr-cluster-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odr-cluster-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odr-cluster-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odr-cluster-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odr-cluster-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odr-cluster-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odr-cluster-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odr-cluster-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odr-cluster-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odr-cluster-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odr-cluster-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odr-cluster-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odr-cluster-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Apache Software Foundation
      provider-url: ""
    name: camel-karavan-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: camel-karavan-operator.v3.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Karavan",
                "metadata": {
                  "labels": {
                    "app": "karavan"
                  },
                  "name": "karavan"
                },
                "spec": {
                  "instances": 1,
                  "auth": "public",
                  "environment": "dev",
                  "runtimes": "quarkus,spring-boot",
                  "gitPullInterval": "10s"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/apache/camel-karavan-operator:3.20.1
          repository: https://github.com/apache/camel-karavan
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Karavan
            name: karavans.camel.apache.org
            version: v1
        description: |
          Apache Camel Karavan
          ====================
          Karavan is an Integration Toolkit for Apache Camel, which makes integration easy and fun through the visualization of pipelines,
          integration with runtimes and package, image build and deploy to kubernetes out-of-the-box.

          ## Installation

          1. Install Apache Camel Karavan Operator
          2. Install Tekton operator
          3. Create namespace, ex: `karavan`
          4. Create `Secret` in the namespace, ex:
          ```
          kind: Secret
          apiVersion: v1
          metadata:
            name: karavan
            namespace: karavan
          type: Opaque
          stringData:
            master-password: karavan
            oidc-secret: secret
            oidc-server-url: https://hostname/auth/realms/karavan
            oidc-frontend-url: https://hostname/auth
            git-repository: https://github.com/repository/projects.git
            git-password: password
            git-username: username
            git-branch: main
            image-registry: image-registry.openshift-image-registry.svc:5000
          ```
          5. Create `Karavan` instance in the namespace, ex:
          ```
          apiVersion: camel.apache.org/v1
          kind: Karavan
          metadata:
            name: karavan
            namespace: karavan
          spec:
            instances: 1
            auth: public
            environment: dev
            runtimes: quarkus,spring-boot
            gitPullInterval: 10s
          ```
        displayName: Camel Karavan Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - camel
        - karavan
        - integration
        - microservices
        - low-code
        links:
        - name: Apache Camel Karavan source code repository
          url: https://github.com/apache/camel-karavan
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - ghcr.io/apache/camel-karavan-operator:3.20.1
        version: 3.20.1
      entries:
      - name: camel-karavan-operator.v3.20.1
        version: 3.20.1
      - name: camel-karavan-operator.v3.20.0
        version: 3.20.0
      - name: camel-karavan-operator.v3.18.5
        version: 3.18.5
      name: alpha
    defaultChannel: alpha
    packageName: camel-karavan-operator
    provider:
      name: The Apache Software Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Upbound Inc.
      provider-url: https://upbound.io
    name: universal-crossplane
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: universal-crossplane.1.5.1-up.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"pkg.crossplane.io/v1",
                "kind":"Configuration",
                "metadata":{
                  "name":"xp-getting-started-with-aws"
                },
                "spec":{
                  "package":"registry.upbound.io/xp/getting-started-with-aws:v1.2.2"
                }
              },
              {
                "apiVersion":"apiextensions.crossplane.io/v1",
                "kind":"CompositeResourceDefinition",
                "metadata":{
                  "name":"compositepostgresqlinstances.database.example.org"
                },
                "spec":{
                  "group":"database.example.org",
                  "names":{
                    "kind":"CompositePostgreSQLInstance",
                    "plural":"compositepostgresqlinstances"
                  },
                  "claimNames":{
                    "kind":"PostgreSQLInstance",
                    "plural":"postgresqlinstances"
                  },
                  "connectionSecretKeys":[
                      "username",
                      "password",
                      "endpoint",
                      "port"
                  ],
                  "versions":[
                    {
                      "name":"v1alpha1",
                      "served":true,
                      "referenceable":true,
                      "schema":{
                        "openAPIV3Schema":{
                          "type":"object",
                          "properties":{
                            "spec":{
                              "type":"object",
                              "properties":{
                                "parameters":{
                                  "type":"object",
                                  "properties":{
                                    "storageGB":{
                                      "type":"integer"
                                    }
                                  },
                                  "required":[
                                      "storageGB"
                                  ]
                                }
                              },
                              "required":[
                                  "parameters"
                              ]
                            }
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion":"apiextensions.crossplane.io/v1",
                "kind":"Composition",
                "metadata":{
                  "name":"compositepostgresqlinstances.aws.database.example.org",
                  "labels":{
                    "provider":"aws",
                    "guide":"quickstart",
                    "vpc":"default"
                  }
                },
                "spec":{
                  "writeConnectionSecretsToNamespace":"crossplane-system",
                  "compositeTypeRef":{
                    "apiVersion":"database.example.org/v1alpha1",
                    "kind":"CompositePostgreSQLInstance"
                  },
                  "resources":[
                    {
                      "name":"rdsinstance",
                      "base":{
                        "apiVersion":"database.aws.crossplane.io/v1beta1",
                        "kind":"RDSInstance",
                        "spec":{
                          "forProvider":{
                            "region":"us-east-1",
                            "dbInstanceClass":"db.t2.small",
                            "masterUsername":"masteruser",
                            "engine":"postgres",
                            "engineVersion":"9.6",
                            "skipFinalSnapshotBeforeDeletion":true,
                            "publiclyAccessible":true
                          },
                          "writeConnectionSecretToRef":{
                            "namespace":"crossplane-system"
                          }
                        }
                      },
                      "patches":[
                        {
                          "fromFieldPath":"metadata.uid",
                          "toFieldPath":"spec.writeConnectionSecretToRef.name",
                          "transforms":[
                            {
                              "type":"string",
                              "string":{
                                "fmt":"%s-postgresql"
                              }
                            }
                          ]
                        },
                        {
                          "fromFieldPath":"spec.parameters.storageGB",
                          "toFieldPath":"spec.forProvider.allocatedStorage"
                        }
                      ],
                      "connectionDetails":[
                        {
                          "fromConnectionSecretKey":"username"
                        },
                        {
                          "fromConnectionSecretKey":"password"
                        },
                        {
                          "fromConnectionSecretKey":"endpoint"
                        },
                        {
                          "fromConnectionSecretKey":"port"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion":"pkg.crossplane.io/v1",
                "kind":"Provider",
                "metadata":{
                  "name":"provider-aws"
                },
                "spec":{
                  "package":"crossplane/provider-aws:master"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime, Cloud Provider, Database, Integration &
            Delivery, Networking, Storage
          com.redhat.component: universal-crossplane
          com.redhat.delivery.backport: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.6
          containerImage: upbound/uxp-bootstrapper
          createdAt: 2021-05-18T00:00:00+0000
          description: Upbound Universal Crossplane (UXP) is Upbound's official enterprise-grade
            distribution of Crossplane.
          io.k8s.display-name: universal-crossplane
          io.openshift.tags: uxp,crossplane,upbound
          maintainer: Upbound Inc. <info@upbound.io>
          name: universal-crossplane
          ocs.tags: v4.6
          operatorframework.io/initialization-resource: |
            apiVersion: v1
            kind: Secret
            metadata:
              name: upbound-control-plane-token
              namespace: upbound-system
            type: Opaque
            data:
              token: "Output of `up cloud controlplane attach | base64` command"
          operatorframework.io/suggested-namespace: upbound-system
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io.bundle.channels.v1: stable
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: universal-crossplane
          operators.operatorframework.io/internal-objects: '["locks.pkg.crossplane.io"]'
          repository: https://github.com/upbound/universal-crossplane
          summary: Upbound Universal Crossplane (UXP) is Upbound's official enterprise-grade
            distribution of Crossplane.
          support: Upbound Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: CompositeResourceDefinition
            kind: CompositeResourceDefinition
            name: compositeresourcedefinitions.apiextensions.crossplane.io
            version: v1
          - displayName: CompositionRevision
            kind: CompositionRevision
            name: compositionrevisions.apiextensions.crossplane.io
            version: v1alpha1
          - displayName: Composition
            kind: Composition
            name: compositions.apiextensions.crossplane.io
            version: v1
          - displayName: ConfigurationRevision
            kind: ConfigurationRevision
            name: configurationrevisions.pkg.crossplane.io
            version: v1
          - displayName: Configuration
            kind: Configuration
            name: configurations.pkg.crossplane.io
            version: v1
          - displayName: ControllerConfig
            kind: ControllerConfig
            name: controllerconfigs.pkg.crossplane.io
            version: v1alpha1
          - displayName: Lock
            kind: Lock
            name: locks.pkg.crossplane.io
            version: v1alpha1
          - displayName: ProviderRevision
            kind: ProviderRevision
            name: providerrevisions.pkg.crossplane.io
            version: v1
          - displayName: Provider
            kind: Provider
            name: providers.pkg.crossplane.io
            version: v1
        description: |
          ## Overview

          Upbound Universal Crossplane (UXP) is [Upbound's][upbound] official
          enterprise-grade distribution of [Crossplane][crossplane]. It's fully compatible
          with upstream Crossplane, open source, capable of connecting to Upbound Cloud
          for real-time dashboard visibility, and maintained by Upbound. It's the easiest
          way for both individual community members and enterprises to build their
          production control planes.

          ## Quick Start

          1. Install the [Upbound CLI][upbound-cli].
              ```console
              curl -sL https://cli.upbound.io | sh
              # Install with Homebrew
              brew install upbound/tap/up
              ```

          2. [Create an Upbound account][create-account] for a free dashboard for UXP.

          3. Connect UXP to Upbound Cloud.
              ```console
              # The name of your new UXP control plane.
              UXP_NAME=mycrossplane

              up cloud login
              up cloud controlplane attach ${UXP_NAME} | up uxp connect -
              ```

          4. Manage your UXP control plane by [signing in][login] to your Upbound account.

          ## Additional Resources

          - The [UXP Documentation][uxp-documentation] provides additional information
            about UXP and resources for developers, like examples.
          - The [developer guide][developer-guide] describes how to build and run UXP
            locally from source.
          - [UXP Slack][uxp-slack] is where you can go to get all of your UXP questions
            answered.

          [upbound]: https://upbound.io
          [crossplane]: https://crossplane.io/
          [upbound-cli]: https://github.com/upbound/up
          [create-account]: https://cloud.upbound.io/register
          [login]: https://cloud.upbound.io/login
          [uxp-documentation]: https://cloud.upbound.io/docs/uxp
          [developer-guide]: https://github.com/upbound/universal-crossplane/blob/main/docs/developer-guide.md
          [uxp-slack]: https://crossplane.slack.com/archives/upbound
        displayName: Upbound Universal Crossplane (UXP)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - infrastructure
        - services
        - application
        - database
        - cache
        - bucket
        - infra
        - app
        - ops
        - oam
        - gcp
        - azure
        - aws
        - alibaba
        - cloudsql
        - rds
        - s3
        - azuredatabase
        - asparadb
        - gke
        - aks
        - eks
        links:
        - name: GitHub
          url: https://github.com/upbound/universal-crossplane
        - name: Website
          url: https://upbound.io
        - name: Twitter
          url: https://twitter.com/upbound_io
        - name: Slack
          url: https://slack.crossplane.io/
        maintainers:
        - email: info@upbound.io
          name: Upbound Inc.
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Upbound Inc.
          url: https://upbound.io
        relatedImages:
        - upbound/uxp-bootstrapper:v1.5.1-up.1
        - upbound/crossplane:v1.5.1-up.1
        - upbound/xgql:v0.1.5
        version: 1.5.1-up.1
      entries:
      - name: universal-crossplane.1.5.1-up.1
        version: 1.5.1-up.1
      - name: universal-crossplane.1.4.3-up.1
        version: 1.4.3-up.1
      - name: universal-crossplane.1.3.3-up.1
        version: 1.3.3-up.1
      - name: universal-crossplane.1.3.1-up.1
        version: 1.3.1-up.1
      - name: universal-crossplane.1.2.3-up.1
        version: 1.2.3-up.1
      - name: universal-crossplane.1.2.2-up.1
        version: 1.2.2-up.1
      - name: universal-crossplane.1.2.1-up.4
        version: 1.2.1-up.4
      - name: universal-crossplane.1.2.1-up.3
        version: 1.2.1-up.3
      name: stable
    defaultChannel: stable
    packageName: universal-crossplane
    provider:
      name: Upbound Inc.
      url: https://upbound.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Palo Alto Networks
      provider-url: ""
    name: pcc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pcc-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "name": "pcc-console",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "ConsoleDefender",
                "metadata": {
                  "name": "pcc-consoledefender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "defenderConfig": {
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Defender",
                "metadata": {
                  "name": "pcc-defender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "defenderConfig": {
                    "clusterAddress": "twistlock-console.example.com",
                    "consoleAddress": "https://twistlock-console.example.com:8083",
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: quay.io/prismacloud/pcc-operator@sha256:51a343d5ea0115e3c43ccaa546492e85ab74cf2cf2658d725c1ea7099fc89191
          description: Deploy Prisma Cloud Compute for cloud-native security in your
            clusters
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator
          support: pcc-integrations@paloaltonetworks.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Install the Prisma Cloud Compute Console and Defenders
            displayName: Console and Defenders
            kind: ConsoleDefender
            name: consoledefenders.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install the Prisma Cloud Compute Console
            displayName: Console
            kind: Console
            name: consoles.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install Prisma Cloud Compute Defenders
            displayName: Defenders
            kind: Defender
            name: defenders.pcc.paloaltonetworks.com
            version: v1alpha1
        description: |
          ## Features
          - Deploy Console
              - Create initial local administrator user
              - Add license
          - Deploy Defenders
          - Support for offline deployments
          - Upgrade Console
          - Upgrade Defenders

          ## Installation
          See the [Prisma Cloud Compute Operator documentation](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator) for installation guides.

          ## Support
          This operator is available "as is," and should be seen as community-supported; however, Palo Alto Networks will continue contributing whenever possible.
          Palo Alto Networks does not provide technical support or help with using or troubleshooting the operator through normal support channels.
          We encourage you to open GitHub [issues](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/issues) and [pull requests](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/pulls) to track bugs and feature requests.

          ## Other links
          [Prisma Cloud Compute product documentation](https://docs.paloaltonetworks.com/prisma/prisma-cloud/prisma-cloud-admin-compute.html)

          [Prisma Cloud Compute API documentation](https://prisma.pan.dev/api/cloud/cwpp)
        displayName: Prisma Cloud Compute Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - palo
        - palo alto
        - palo alto networks
        - paloalto
        - twistlock
        - compute
        - prisma
        - prisma cloud
        - prisma cloud compute
        - security
        - image
        - images
        - container
        - containers
        - docker
        - podman
        - vulnerability
        - vulnerabilities
        - compliance
        - runtime
        - scan
        - scanning
        - monitor
        - monitoring
        - alert
        - alerting
        links:
        - name: Prisma Cloud
          url: https://www.paloaltonetworks.com/prisma/cloud
        maintainers:
        - email: pcc-integrations@paloaltonetworks.com
          name: Wyatt Gill
        maturity: alpha
        provider:
          name: Palo Alto Networks
        relatedImages:
        - quay.io/prismacloud/pcc-operator@sha256:51a343d5ea0115e3c43ccaa546492e85ab74cf2cf2658d725c1ea7099fc89191
        version: 0.2.3
      entries:
      - name: pcc-operator.v0.2.3
        version: 0.2.3
      name: stable
    defaultChannel: stable
    packageName: pcc-operator
    provider:
      name: Palo Alto Networks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-iam-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-iam-operator.v4.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "Authentication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-iam-operator",
                    "app.kubernetes.io/managed-by": "ibm-iam-operator",
                    "app.kubernetes.io/name": "ibm-iam-operator"
                  },
                  "name": "example-authentication"
                },
                "spec": {
                  "auditService": {
                    "imageName": "auditService no longer used - ignore",
                    "imageRegistry": "auditService no longer used - ignore",
                    "imageTag": "auditService no longer used - ignore"
                  },
                  "authService": {
                    "imageName": "icp-platform-auth",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "ldapsCACert": "platform-auth-ldaps-ca-cert",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "650Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "400Mi",
                        "memory": "350Mi"
                      }
                    },
                    "routerCertSecret": "saml-auth-secret"
                  },
                  "clientRegistration": {
                    "imageName": "im-initcontainer",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "158Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "config": {
                    "authUniqueHosts": "internal-ip1 internal-ip2 mycluster.icp",
                    "bootstrapUserId": "kubeadmin",
                    "claimsMap": "name=\"givenName\" family_name=\"givenName\" given_name=\"givenName\" preferred_username=\"displayName\" display_name=\"displayName\"",
                    "claimsSupported": "name,family_name,display_name,given_name,preferred_username",
                    "clusterCADomain": "mycluster.icp",
                    "clusterExternalAddress": "10.0.0.1",
                    "clusterInternalAddress": "10.0.0.1",
                    "clusterName": "mycluster",
                    "defaultAdminPassword": "password",
                    "defaultAdminUser": "cpadmin",
                    "defaultLogin": "",
                    "enableImpersonation": false,
                    "fipsEnabled": true,
                    "ibmCloudSaas": false,
                    "icpPort": 8443,
                    "installType": "fresh",
                    "isOpenshiftEnv": true,
                    "nonceEnabled": true,
                    "oidcIssuerURL": "https://127.0.0.1:443/idauth/oidc/endpoint/OP",
                    "openshiftPort": 443,
                    "preferredLogin": "",
                    "providerIssuerURL": "",
                    "roksEnabled": true,
                    "roksURL": "https://roks.domain.name:443",
                    "roksUserPrefix": "changeme",
                    "saasClientRedirectUrl": "",
                    "scimAdminPassword": "scimpassword",
                    "scimAdminUser": "scimadmin",
                    "scopeClaim": "profile=\"name,family_name,display_name,given_name,preferred_username\"",
                    "wlpClientID": "4444be3a738841016ab76d71b650e836",
                    "wlpClientRegistrationSecret": "f1362ca4d20b8389af2d1ea68042c9af",
                    "wlpClientSecret": "aa73bf39752053bf723d1143fb4cf8a2",
                    "xframeDomain": "",
                    "zenFrontDoor": false
                  },
                  "identityManager": {
                    "imageName": "icp-identity-manager",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "masterNodesList": "10.0.0.1",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "550Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "ephemeral-storage": "300Mi",
                        "memory": "150Mi"
                      }
                    }
                  },
                  "identityProvider": {
                    "imageName": "icp-identity-provider",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "550Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "ephemeral-storage": "300Mi",
                        "memory": "150Mi"
                      }
                    }
                  },
                  "initMongodb": {
                    "imageName": "im-initcontainer",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "178Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "operatorVersion": "0.14.1",
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
          createdAt: "2025-02-20T00:03:25Z"
          description: The IAM operator provides a simple Kubernetes CRD-Based API
            to manage the lifecycle of IAM services. With this operator, you can simply
            deploy and upgrade the IAM services
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.11.0
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-iam-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Authentication is the Schema for the authentications API
            displayName: Authentication
            kind: Authentication
            name: authentications.operator.ibm.com
            version: v1alpha1
          - description: Client is the Schema for the clients API
            displayName: Client
            kind: Client
            name: clients.oidc.security.ibm.com
            version: v1
        description: |-
          The `ibm-iam-operator` installs the IBM Cloud Platform Common Services Identity and access management (IAM) service.

          **Important:** Do not install this operator directly. Install this operator only by using the IBM Common Service Operator. For more information about installing this operator and other Common Services operators, see [Installer documentation](http://ibm.biz/cpcs_opinstall). Additionally, you can exit this panel and navigate to the IBM Common Services tile in OperatorHub to learn more about the operator. If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For more information about IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).

          You can use the `ibm-iam-operator` to install the authentication and authorization services for the IBM Cloud Platform Common Services.

          With these services, you can configure security for IBM Cloud Platform Common Services, IBM Certified Containers (IBM products), or IBM Cloud Paks that are installed.

          For more information about the available IBM Cloud Platform Common Services, see the [IBM Knowledge Center](http://ibm.biz/cpcsdocs).

          ## Supported platforms

           - Red Hat OpenShift Container Platform 4.3 or newer installed on one of the following platforms:

              - Linux x86_64
              - Linux on Power (ppc64le)
              - Linux on IBM Z and LinuxONE

          ## Prerequisites

          Before you install this operator, you need to first install the operator dependencies and prerequisites:

          - For the list of operator dependencies, see the IBM Knowledge Center [Common Services dependencies documentation](http://ibm.biz/cpcs_opdependencies).

          - For the list of prerequisites for installing the operator, see the IBM Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).

          ## Documentation

          To install the operator by using the IBM Common Services Operator, follow the installation and configuration instructions that are in the IBM Knowledge Center.

          - If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).
          - If you are using the operator with an IBM Containerized Software, see the IBM Cloud Platform Common Services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall).
        displayName: IBM IM Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:b8886b9b54e6ee4c6b2981ef092e54348a0286577d9924cf136cee881f5bf1f4
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-bundle@sha256:98cf2a6c7b7af5d7cbfbc61ce140201272b30b12edabaec2b2d4287b0690d89e
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:f289ac0d44803ff12d41c3659e6fdf04c404e0b3b233b6bc44eec671abb1b47e
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:52ce08ac2bce7c02dc30e2c0aa74df4bde8f156debccc59960f02b4b0040ffea
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:97b9ecef8caef156a0eb073b6eb2a34b63209b44511f5694a243f11285d88662
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:0f46b288d060a15a48ecad8f328a8046bf232b8629c3d79da6a82ea712b363f4
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:7cec55c00827b747f23c2f1703d99e1ca5b30a0c445f8ad0471b25b0c9cca57c
        version: 4.11.0
      entries:
      - name: ibm-iam-operator.v4.11.0
        version: 4.11.0
      name: v4.10
    defaultChannel: v4.10
    packageName: ibm-iam-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: vfunction-operator
      app.kubernetes.io/managed-by: operator
      app.kubernetes.io/name: vfunction
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      name: vfunction-server-operator
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: vFunction
      provider-url: ""
    name: vfunction-server-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vfunction-server-operator.v4.3.1762
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vfunction.com/v1",
                "kind": "VfunctionServer",
                "metadata": {
                  "name": "vfunction-server",
                  "namespace": "vfunction",
                  "labels": {
                      "name": "vfunction",
                      "app.kubernetes.io/name": "vfunction",
                      "app.kubernetes.io/instance": "vfunction-server",
                      "app.kubernetes.io/managed-by": "operator"
                  }
                },
                "spec": {
                  "admin": {
                    "email": "admin@mycompany.com",
                    "name": "Admin",
                    "password": "Password1!"
                  },
                  "host": "http://my.domain.com",
                  "org_name": "MyCompany",
                  "upgrade": "Never",
                  "measurements": {
                    "min_number_of_services": "2",
                    "max_number_of_services": "10",
                    "max_pod_memory_capacity": "16G",
                    "max_pod_cpu_capacity": "1",
                    "pod_ephemeral_storage": "1G",
                    "S3": {
                      "bucket": "",
                      "key": "",
                      "secret": "",
                      "region": ""
                    }
                  },
                  "smtp": {
                    "identity": "",
                    "password": "",
                    "url": "",
                    "user": ""
                  },
                  "authentication": {
                    "authority": "",
                    "client_id": "",
                    "client_secret": "",
                    "jwks_url": "",
                    "issuer": "",
                    "ca_root_crt": ""
                  },
                  "tls": {
                    "use_letsencrypt": "No",
                    "crt": "",
                    "key": ""
                  },
                  "nginx": {
                    "force_http": "No",
                    "ipv6_disabled": "No",
                    "service_type": ""
                  },
                  "route": {
                    "use_route": "Yes",
                    "tls_termination": "edge",
                    "use_tls_certificate": "No"
                  },
                  "mysql": {
                    "external_mysql_ip": "",
                    "external_mysql_user": "",
                    "external_mysql_password": "",
                    "max_pod_memory_capacity": "4G",
                    "max_pod_cpu_capacity": "1",
                    "pod_ephemeral_storage": "1G"
                  },
                  "storage": {
                    "storage_class": "",
                    "mysql_persistent_volume": "",
                    "mysql_persistent_volume_claim": "",
                    "mysql_persistent_volume_claim_access_mode": "ReadWriteOnce",
                    "storage_persistent_volume": "",
                    "storage_persistent_volume_claim": "",
                    "storage_persistent_volume_claim_access_mode": "ReadWriteOnce",
                    "lets_encrypt_persistent_volume": "",
                    "lets_encrypt_persistent_volume_claim": "",
                    "lets_encrypt_persistent_volume_claim_access_mode": "ReadWriteOnce"
                  },
                  "distributed": {
                    "otlp": {
                      "min_number_of_services": "1",
                      "max_number_of_services": "10",
                      "max_pod_memory_capacity": "8G",
                      "max_pod_cpu_capacity": "1",
                      "pod_ephemeral_storage": "250M",
                      "max_folder_size_in_mb": "200"
                    },
                    "dm": {
                      "min_number_of_services": "1",
                      "max_number_of_services": "10",
                      "max_pod_memory_capacity": "8G",
                      "max_pod_cpu_capacity": "1",
                      "pod_ephemeral_storage": "1G"
                    }
                  },
                  "backup": {
                    "mysql": {
                      "restore_during_installation": "No",
                      "S3": {
                        "bucket": "",
                        "key": "",
                        "secret": "",
                        "region": ""
                      }
                    }
                  },
                  "security": {
                    "image_pull_policy": "",
                    "force_hashed_images": "No",
                    "set_pod_resources": "Yes",
                    "proxy": {
                      "http_proxy": "",
                      "https_proxy": "",
                      "additional_no_proxy": ""
                    },
                    "use_rate_limiting": "No",
                    "requests_per_second": "10"
                  },
                  "affinity": {
                    "node_affinity": [
                      {
                        "key": "",
                        "operator": "",
                        "values": "",
                        "services": ""
                      }
                    ],
                    "node_selectors": [
                      {
                        "key": "",
                        "value": "",
                        "services": ""
                      }
                    ],
                    "additional_labels": [
                      {
                        "key": "",
                        "value": "",
                        "services": ""
                      }
                    ],
                    "additional_annotations": [
                      {
                        "key": "",
                        "value": "",
                        "services": ""
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, AI/Machine Learning
          certified: Red Hat certified
          containerImage: registry.connect.redhat.com/vfunction/vfunction-server-operator@sha256:a5a0f47f7fb9219139f8d749da14950a8c39698b340644011485dfbec54ebe8b
          createdAt: "2025-08-04T16:10:32Z"
          description: vFunction is a cutting-edge code analysis, machine learning,
            and automation to boost your Java modernization projects.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.3.1762
          support: vFunction
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of vFunction running on your cluster.
            displayName: vFunction
            kind: VfunctionServer
            name: vfunctionservers.vfunction.com
            version: v1
        description: |2-

          ## Introduction

          [vFunction](https://www.vfunction.com/) is a cloud-native modernization platform that combines dynamic and static code analysis, machine learning, and automation to automatically identify and extract services from existing applications. vFunction is the only platform purpose-built for modernization of Java applications.

          The vFunction modernization process starts by learning the running monolithic application, and surfacing the interdependencies within it. Using AI, the platform analyzes and identifies services that can be separated from the application. This decomposition can present a range of micro, mini, or even macro services, depending on your application environment, each being an independently deployable and scalable application component.

          vFunction automates the extraction of these services, enabling you to modernize your monolith, quickly and easily.

          ## The vFunction platform

          The platform consists of 3 basic components; the server, the controller package, and a tools package. The server runs as an operand on an OpenShift environment. The controller package is installed on the machine that runs the monolithic application which can be either a Linux or a Windows machine, and the tools are run on a development machine, with access to the code of the monolithic application.

          The controller package consists of three elements: the vFunction agent, that collects data during the dynamic analysis phase; the vFunction Viper application, that performs static analysis on the binaries of the application; and the vFunction controller that handles all the communication between the agent, Viper, and the vFunction server.

          The vFunction agent is a mix of a Java and native agent, and needs to run on the JVM that is currently running your application. [Refer to the vFunction OOB Support Matrix document](https://drive.google.com/file/d/1ccq8LFab1FrYAimDUxwgjiCOdk4QuVzs/view) for a list of supported application servers and JVMs.

          This operator installs a vFunction server instance to be connected later with one or more vFunction controllers installed on your application machine(s).

          ## Before You Start

          The vFunction operator requires a Red Hat OpenShift Kubernetes Application Platform 4.x.

          The cluster recommended configuration (which is the same as the minimal one) should consist of:

          - *1 master node*

          - *1 worker node*

          - *Storage capacity enough for dynamically provision 2 PVs of 50G each*

          - *A default StorageClass configured*

          vFunction doesn’t provide any encryption mechanism for data stored on any attached storage. If required, encryption can be achieved by the client by using its own external means on the storage itself.

          vFunction supports backup of all critical data as part of the application. Nevertheless, for complete backup of entire data it is recommended that the client apply its own storage based backup mechanism.

          ## Install Prerequisites

          ###Cluster:
          The operator was certified and tested on OCP 4.6 - 4.17.
          If you encounter any issue with other OCP versions or cloud providers please contact info@vfunction.com.

          ###Storage:
          The operator creates two new PersistentVolumeClaims (PVCs) during installation time. For their creation, it relies on the default StorageClass to dynamically provision two PersistentVolumes (PVs). The system must have a default StorageClass in place.
          Both PersistentVolumes are defaulted to access with ReadWriteOnce mode, but you can change this by specifying a different access mode in the YAML file.
          Notes:
          - There is an option to specify a specific StorageClass for the PersistentVolumeClaims.
          - There is an option to specify the use of a specific pre-created PersistentVolume for each PersistentVolumeClaim
          - There is an option to specify a pre-created PersistentVolumeClaim for each Deployment.
          - There is an option to change the access mode for each PersistentVolumeClaim.

          ###Project:
          The operator should be installed in a new and dedicated project (namespace). If you intend to install more than one vFunction operands in the same cluster, please use different projects for each one. Installation of more than one operand on the same cluster will allow scaling by load balancing multiple applications to different vFunction servers.

          ## Installation instructions

          1. Prepare the YAML file to use in the installation:

              a. Copy the following YAML template into a text editor:

              ```yaml
                  apiVersion: vfunction.com/v1
                  kind: VfunctionServer
                  metadata:
                    name: vfunction
                    namespace: vfunction
                    labels:
                      name: vfunction
                      app.kubernetes.io/name: vfunction
                      app.kubernetes.io/instance: vfunction-server
                      app.kubernetes.io/managed-by: operator
                  spec:
                    host: "http://my.domain.com"
                    org_name: "MyCompany"
                    upgrade: "Never"
                    admin:
                      email: "admin@mycompany.com"
                      name: "Admin"
                      password: "Password1!"
                    measurements:
                      min_number_of_services: "2"
                      max_number_of_services: "10"
                      max_pod_memory_capacity: "16G"
                      max_pod_cpu_capacity: "1"
                      S3:
                        bucket: ""
                        key: ""
                        secret: ""
                    smtp:
                      password: ""
                      url: ""
                      identity: ""
                      user: ""
                    authentication:
                      authority: ""
                      client_id: ""
                      client_secret: ""
                      jwks_url: ""
                      issuer: ""
                      ca_root_crt: |
                        -----BEGIN CERTIFICATE-----
                        ...
                        -----END CERTIFICATE-----
                    tls:
                      use_letsencrypt: "No"
                      crt: |
                        -----BEGIN CERTIFICATE-----
                        ...
                        -----END CERTIFICATE-----
                      key: |
                        -----BEGIN PRIVATE KEY-----
                        ...
                        -----END PRIVATE KEY-----
                    nginx:
                      force_http: "No"
                      ipv6_disabled: "No"
                      service_type: ""
                    route:
                      use_route: "Yes"
                      tls_termination: "edge"
                      use_tls_certificate: "No"
                    mysql:
                      external_mysql_ip: ""
                      external_mysql_user: ""
                      external_mysql_password: ""
                      max_pod_memory_capacity: "4G"
                      max_pod_cpu_capacity: "1"
                    storage:
                      storage_class: ""
                      mysql_persistent_volume: ""
                      mysql_persistent_volume_claim: ""
                      mysql_persistent_volume_claim_access_mode: "ReadWriteOnce"
                      storage_persistent_volume: ""
                      storage_persistent_volume_claim: ""
                      storage_persistent_volume_claim_access_mode: "ReadWriteOnce"
                      lets_encrypt_persistent_volume: ""
                      lets_encrypt_persistent_volume_claim: ""
                      lets_encrypt_persistent_volume_claim_access_mode: "ReadWriteOnce"
                    distributed:
                        otlp:
                          min_number_of_services: "1"
                          max_number_of_services: "10"
                          max_pod_memory_capacity: "8G"
                          max_pod_cpu_capacity: "1"
                          max_folder_size_in_mb: "200"
                        dm:
                          min_number_of_services: "1"
                          max_number_of_services: "10"
                          max_pod_memory_capacity: "8G"
                          max_pod_cpu_capacity: "1"
                    backup:
                      mysql:
                        restore_during_installation: "No"
                        S3:
                          bucket: ""
                          key: ""
                          secret: ""
                    security:
                      image_pull_policy: ""
                      force_hashed_images: "No"
                      set_pod_resources: "Yes"
                      proxy:
                        http_proxy: ""
                        https_proxy: ""
                        additional_no_proxy: ""
                      use_rate_limiting: "No"
                      requests_per_second: "10"
                    affinity:
                      node_affinity:
                        - key: ""
                          operator: ""
                          values: ""
                          services: ""
                      node_selectors:
                        - key: ""
                          value: ""
                          services: ""
                      additional_labels:
                        - key: ""
                          value: ""
                          services: ""
                      additional_annotations:
                        - key: ""
                          value: ""
                          services: ""
              ```
              b. Customize the template fields by replacing with your information:

              - **host**: Enter the FQDN for accessing the vFunction dashboard.
                          Make sure that you use a domain name and not an IP address. Writing "https://" will mandate TLS while writing "http://" will expose the server through HTTP.
              - **org_name**: Enter your organization name.
              - **upgrade**: Choose auto-upgrade mode.
                  - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically.
                  - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
                  - **Never**: The automatic upgrading mechanism is off (default).
                  - **Maintenance**: Put the operator in a maintenance mode.
              - **admin.email**: Enter the email address of your administrator.
              - **admin.name**: Enter the name of your administrator.
              - **admin.password**: Enter the password you want to use for the vFunction administrator.
                                    The password should be at least 8 characters long, and consist of at least one lowercase letter, at least one uppercase letter, at least one number, and at least one special character.
              - **measurements.min\_num\_of\_services**: Minimum Number of measurement service pods.
              - **measurements.max\_num\_of\_services**: Maximum Number of measurement service pods.
              - **measurements.max\_pod\_memory\_capacity**: Maximum memory capacity limit for measurement service pods. (Default is "8G").
              - **measurements.max\_pod\_cpu\_capacity**: Maximum CPU capacity limit for measurement service pods. (Default is "1").
              - **measurements.pod\_ephemeral\_storage**: Ephemeral Storage Capacity for measurement service pods. (Default is "1G").
              - **measurements.S3**: Save all measurements data to S3, instead of local PV (optional).
                  - **measurements.S3.bucket**: Measurements S3 bucket name.
                  - **measurements.S3.key**: Measurements S3 key.
                  - **measurements.S3.secret**: Measurements S3 secret.
                  - **measurements.S3.region**: Measurements S3 region.
              - **smtp.user**: Enter the email address for a designated user for the SMTP server (optional).
              - **smtp.password**: Enter this user’s password (optional).
              - **smtp.identity**: Enter this user’s password (optional).
              - **smtp.url**: Enter the SMTP server URL (optional).
              - **authentication.authority**: Authority for authentication (Optional).
              - **authentication.client_id**: Client ID for authentication (Optional).
              - **authentication.client_secret**: Client Secret for authentication (Optional).
              - **authentication.jwks_url**: JWKS URL for authentication (Optional).
              - **authentication.issuer**: Issuer for authentication (Optional).
              - **authentication.ca\_root\_crt**: CA root certificate for authentication (Optional).
              - **tls.use_letsencrypt**: If your host FQDN starts with “https”, choose whether to use Let’s Encrypt to obtain certificates.
              - **tls.crt**: If your host FQDN starts with “https”, paste in the certifications you have for using the TLS connection (Applicable only if you don’t use Let’s Encrypt).
              - **tls.key**: If your host FQDN starts with “https”, enter the key you have for using the TLS connection (Applicable only if you don’t use Let’s Encrypt).
              - **nginx.force_http**: Force internal NGINX to listen on http even if the host FQDN starts with “https”.
                                      This is useful if you want to use a TLS termination proxy in front of vFunction, other then the route provided by the operator.
              - **nginx.ipv6_disabled**: Disable IPv6 on internal NGINX.
              - **nginx.service_type**: Set the service type for the internal NGINX service - ClusterIP, NodePort or LoadBalancer.
              - **route.use_route**: Use OpenShift route to expose vFunction. (Default is "Yes").
              - **route.tls_termination**: Set the TLS termination type for the OpenShift route - edge, passthrough or reencrypt. (Default is "edge").
              - **route.use\_tls\_certificate**: Use TLS certificate for OpenShift route (as provided in tls.crt and tls.key) - instead of the default one defined for the route controller in your cluster. (Default is "No").
              - **mysql.external\_mysql\_ip**: If you want to use an external MySQL database, enter its IP address.
              - **mysql.external\_mysql\_user**: If you want to use an external MySQL database, enter its user name.
              - **mysql.external\_mysql\_password**: If you want to use an external MySQL database, enter its password.
              - **mysql.max\_pod\_memory\_capacity**: Maximum memory capacity limit for MySQL pod. (Default is "2G").
              - **mysql.max\_pod\_cpu\_capacity**: Maximum CPU capacity limit for MySQL pod. (Default is "1").
              - **mysql.pod\_ephemeral\_storage**: Ephemeral Storage Capacity for MySQL pod. (Default is "1G").
              - **storage.storage\_class**: If you want to use a specific StorageClass for all PVCs, other than the default one, enter its name.
              - **storage.mysql\_persistent\_volume**: If you want to use a specific PV for mysql PVC, other than the dynamically provisioned one, enter its name.
              - **storage.mysql\_persistent\_volume\_claim**: If you want to use a specific pre-created PVC for mysql deployment, other than the dynamically provisioned one, enter its name.
              - **storage.mysql\_persistent\_volume\_claim\access\_mode**: If you want to use a specific access mode for mysql PVC, other than the default one, enter its name. (Default is "ReadWriteOnce").
              - **storage.storage\_persistent\_volume**: If you want to use a specific PV for storage PVC, other than the dynamically provisioned one, enter its name.
              - **storage.storage\_persistent\_volume\_claim**: If you want to use a specific pre-created PVC for storage deployment, other than the dynamically provisioned one, enter its name.
              - **storage.storage\_persistent\_volume\_claim\access\_mode**: If you want to use a specific access mode for storage PVC, other than the default one, enter its name. (Default is "ReadWriteOnce").
              - **storage.lets\_encrypt\_persistent\_volume**: If you want to use a specific PV for lets encrypt PVC, other than the dynamically provisioned one, enter its name.
              - **storage.lets\_encrypt\_persistent\_volume\_claim**: If you want to use a specific pre-created PVC for lets encrypt deployment, other than the dynamically provisioned one, enter its name.
              - **storage.lets\_encrypt\_persistent\_volume\_claim\access\_mode**: If you want to use a specific access mode for lets encrypt PVC, other than the default one, enter its name. (Default is "ReadWriteOnce").
              - **distributed.otlp.min\_number\_of\_services**: Minimum Number of OTLP pods.
              - **distributed.otlp.max\_number\_of\_services**: Maximum Number of OTLP pods.
              - **distributed.otlp.max\_pod\_memory\_capacity**: Maximum memory capacity for OTLP pods. (Default is "1G").
              - **distributed.otlp.max\_pod\_cpu\_capacity**: Maximum CPU capacity for OTLP pods. (Default is "1").
              - **distributed.otlp.pod\_ephemeral\_storage**: Ephemeral Storage Capacity for OTLP service pods. (Default is "250M").
              - **distributed.otlp.max\_folder\_size\_in\_mb**: Maximum folder size in MB for OTLP pods. (Default is "200").
              - **distributed.dm.min\_number\_of\_services**: Minimum Number of dm pods.
              - **distributed.dm.max\_number\_of\_services**: Maximum Number of dm pods.
              - **distributed.dm.max\_pod\_memory\_capacity**: Maximum memory capacity for dm pods. (Default is "8G").
              - **distributed.dm.max\_pod\_cpu\_capacity**: Maximum CPU capacity for dm pods. (Default is "1").
              - **distributed.dm.pod\_ephemeral\_storage**: Ephemeral Storage Capacity for dm service pods. (Default is "1G").
              - **backup.mysql.restore\_during\_installation**: If you want, during installation, to restore the backup from the latest kept backup with the same major version (if available on attached DB storage or S3 bucket), choose "Yes".
              - **backup.mysql.S3**: Backup MySQL to S3 (optional).
                  - **backup.mysql.S3.bucket**: MySQL backup S3 bucket name.
                  - **backup.mysql.S3.key**: MySQL backup S3 key.
                  - **backup.mysql.S3.secret**: MySQL backup S3 secret.
                  - **backup.mysql.S3.region**: MySQL backup S3 region.
              - **security.image\_pull\_policy**: If you want to use a custom image pull policy for all deployments, enter it here (optional).
              - **security.force\_hashed\_images**: If you want to use hashed images for all deployments, choose "Yes".
              - **security.set\_pod\_resources**: If you want to not set resource requests and limits for all deployments, choose "No".
              - **security.proxy.http\_proxy**: Enter the address of an HTTP proxy (optional).
              - **security.proxy.https\_proxy**: Enter the address of an HTTPS proxy (optional).
              - **security.proxy.additional\_no\_proxy**: Enter additional addresses to exclude from proxying (optional).
              - **security.use\_rate\_limiting**: If you want to use rate limiting in nginx, choose "Yes".
              - **security.requests\_per\_second**: Enter the number of requests per second for rate limiting.
              - **affinity.node\_affinity**: Enter the node affinity for the deployments (optional). There is only one nodeSelectorTerm as requiredDuringSchedulingIgnoredDuringExecution, but it can contain multiple matchExpressions (which are ANDed together).
                  - **key**: Enter the key for the node affinity.
                  - **operator**: Enter the operator for the node affinity (In, NotIn, Exists, DoesNotExist, Gt, Lt).
                  - **values**: Enter the values for the node affinity. A comma-separated list of values.
                  - **services**: Enter the services for the node affinity. A comma-separated list of services to which the node affinity applies. If not specified, the node affinity will apply to all services. List of possible services: mysql, nginx, storage, vfapi-parser, vfapi-users, vfapi-organizations, vfapi-idp, vfapi-mcp, vfapi-measurements, vfapi-otlp, vfapi-dm. (optional).
              - **affinity.node\_selectors**: Enter the node selectors for the deployments (optional).
                  - **key**: Enter the key for the node selector.
                  - **value**: Enter the value for the node selector.
                  - **services**: Enter the services for the node selector. A comma-separated list of services to which the node selector applies. If not specified, the node selector will apply to all services. List of possible services: mysql, nginx, storage, vfapi-parser, vfapi-users, vfapi-organizations, vfapi-idp, vfapi-mcp, vfapi-measurements, vfapi-otlp, vfapi-dm. (optional).
              - **affinity.additional\_labels**: Enter additional labels for the deployments (optional).
                  - **key**: Enter the key for the additional label.
                  - **value**: Enter the value for the additional label.
                  - **services**: Enter the services for the additional label. See above at the node selectors section for more details. (optional).
              - **affinity.additional\_annotations**: Enter additional annotations for the deployments (optional).
                  - **key**: Enter the key for the additional annotation.
                  - **value**: Enter the value for the additional annotation.
                  - **services**: Enter the services for the additional annotation. See above at the node selectors section for more details. (optional).
          1. In the Installed Operators window choose and click on vFunction Operator and then Create Instance.
          1. In the Create vFunctionServer window, paste the YAML file that you created in step 1.
          1. Click Create.
          1. You can now click on your new created vFunction operand in order to see its details and installation progress.

          ## Verifying the installation

          Check that the operand installed successfully:

          1. Check that the "Operand State" property shows “Working”.
          1. Check that the "Successfully Installed" property shows "Yes".

          If "Operand State" shows "Failed", it indicates the operand wasn’t installed correctly. Check the events and logs for all pods (operator and image containers) for any issues. In the event that you cannot troubleshoot, contact vFunction support.

          ## After installation

          The vFunction site is now accessible via the newly created vFunction application custom address, for example, my.domain.com. There are two ways you can access the vFunction dashboard.

          ### Access using the router canonical hostname

          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address, and to the ”vfunction” subdomain of the OpenShift canonical router hostname as the alias.
          For example, ***my.domain.com.   CNAME   vfunction.apps.ocp4.my-openshift.com.***
          2. Find your cluster Router Canonical Hostname address in the newly created vfunction-route-xxx under your vFunction OpenShift project -> **Networking** -> **Routes** -> **vfunction-route-xxx** route -> **Router Canonical Hostname** field.
          You can now access the dashboard using your defined "host" spec property (as above).

          ### Access using the nginx service location

          You may use this access method if your OpenShift is installed on a provider that supports exposing LoadBalancer-type services.
          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address and to the vfunction-nginx-xxx service location.
          For example, ***my.domain.com.   CNAME   a05951ed7cdf-1394239323.us-east-1.elb.amazonaws.com.***
          1. Point your custom domain to the vfunction-nginx-xxx service's external IP location, which you can find in the OpenShift project > **Networking** -> **Services** -> **vfunction-nginx-xxx** service > **Service Address** > **Location** field.

          ## Upgrade and Rollback

          The vFunction operator includes a built-in auto-upgrade mechanism.

          You can choose one of 4 upgrading modes:
          - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically.
          - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
          - **Never**: The automatic upgrading mechanism is off (default).
          - **Maintenance**: Put the operator in a maintenance mode.

          For an on-demand upgrade,  you can change anytime the upgrading mode from “Never” to “Always”, wait for the operand to be upgraded, and change again to “Never”.

          During the upgrade, the "Operand State" property will change to “Upgrading...” and back again to “Working” after a successful upgrade.

          In case of a failure upgrading the operand, an automatic rollback to the last working version will occur. The failed version will be marked as defective and the operator will not try to upgrade to it again.

          Choosing the "Maintenance" option will tell the operator to take down all pods, so you can safely fix storage issues, return volumes from snapshots, etc.
          After finishing the maintenance, return the upgrade mode to your original desired policy.
        displayName: vFunction Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - vfunction
        - analyzing
        - services
        links:
        - name: vFunction
          url: https://vfunction.com
        - name: Terms
          url: https://www.vfunction.com/eula/
        maintainers:
        - email: raviv@vfunction.com
          name: Raviv Halivni
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: vFunction
        relatedImages:
        - registry.connect.redhat.com/vfunction/vfunction-server-operator@sha256:a5a0f47f7fb9219139f8d749da14950a8c39698b340644011485dfbec54ebe8b
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-users@sha256:c471d450d016aa0801ffb3de4648735245f2cf6901576106aed06cacc948d6f6
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-parser@sha256:b70fe4340d0de33c03a3b3458ea29e4cbff30cdd43f0e8d94265568e9a8d1ac7
        - registry.connect.redhat.com/vfunction/vfunction-storage@sha256:8c36770f1fa5034ba089ae81c36c587cb73a5176f63969023c8ab6f90d8d7abd
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-idp@sha256:17fe2e3ef706a4bbd8fdf0e0e08bd021c49104b331674aae7b8978b31730a03d
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-organizations@sha256:8dd4d043806da1e85a4565b2f14bb9e24671fc158f8c1c2adc3804fda5213571
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-dm@sha256:2f82294b15af873c012094a8984cb0c157bf322c28cbe58b38253adc917b6662
        - registry.connect.redhat.com/vfunction/vfunction-mysql@sha256:b4b6543eb02e396e1339877f61b74cc783977dd607cd3679edce8ac6b627e9e9
        - registry.connect.redhat.com/vfunction/vfunction-nginx@sha256:8370b871d05a24da9841a93fb255bf47945428b0a192b96b4eb95b9c0efb86a0
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-mcp@sha256:d50142c1b7f8a52944f9051a13bb821c0a790f137ebef09b0a84256bf7e5d794
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-measurements@sha256:81edae1c7344c13916530f20c8be09512ac8e20542770949783b0025c37079f8
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-otlp@sha256:80ca6d64b0d74f2d4f1b267002d5666f7c602956c656fd36a5c430f77eef0672
        version: 4.3.1762
      entries:
      - name: vfunction-server-operator.v4.3.1762
        version: 4.3.1762
      - name: vfunction-server-operator.v4.0.1576
        version: 4.0.1576
      name: stable
    defaultChannel: stable
    packageName: vfunction-server-operator
    provider:
      name: vFunction
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent
      provider-url: ""
    name: cilium
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cilium.v1.10.15-x084926b
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/hubble-ui@sha256:dec6e7e541cac91ab859245dd2a7fa02de5363b75caa3509a79cf76d4f622550
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/hubble-ui-backend@sha256:a07e13df3e9d7ec0dcc3b158f243cda5a08f6eca38a25a5aa3171f6b3ec491ed
        - quay.io/cilium/operator-generic@sha256:7f2f7b59b541e4f660d65ed0a21d25ab69e49022172529771dd1134315e95c82
        - quay.io/cilium/clustermesh-apiserver@sha256:e73151c3295589de00dccae75819aabdff8fd907260817020f9e8fe3738e78b2
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - registry.connect.redhat.com/isovalent/cilium-olm:970b1d74dbe2b87c28e8dc1c68936de33123a4df-v1.10.15@sha256:afa3ac140fae24b151ab6ff5a8cdef11566a58f7234d702a6a45a398822e6d34
        - quay.io/cilium/cilium@sha256:adbb60f9d19d641a5496fcfc4769ebb55c1d4cd7699fef752f7c94754194cf83
        - quay.io/cilium/hubble-relay@sha256:67a9183fd1febd9e51f4d02bd35092b81c8612a1716015988f69fffba79d999a
        version: 1.10.15+x084926b
      entries:
      - name: cilium.v1.10.15-x084926b
        version: 1.10.15+x084926b
      - name: cilium.v1.10.13-xa8bb1e5
        version: 1.10.13+xa8bb1e5
      - name: cilium.v1.10.12-x362ea11
        version: 1.10.12+x362ea11
      - name: cilium.v1.10.10-x6adc6ff
        version: 1.10.10+x6adc6ff
      name: "1.10"
    - currentCSV: cilium.v1.11.20-xdb5ed32
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/cilium/cilium@sha256:60df3cb7155886e0b62060c7a4a31e457933c6e35af79febad5fd6e43bab2a99
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:3677840d57e1d74e3ac14db0dfceb83088e6ebd70775d775855cdd3bfed68059
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/operator-generic@sha256:1439954acf620f048ef663524ae70b4a25693c58527a2f2cee51124496e29f90
        - quay.io/cilium/hubble-ui@sha256:bbf8b041152041d3ef000d6074a8dd5d40ed910b5c0067f365b00e52e449d467
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/clustermesh-apiserver@sha256:46760182f8c98227cfac27627275616987b71509227775350573d834133a6d49
        - quay.io/cilium/hubble-relay@sha256:e2f38b901fd8bd5adc9a765a5e68836364ebd1e7dfb85c2bcd8a5488b23c3470
        - quay.io/cilium/hubble-ui-backend@sha256:46070270b8ffe7ac3f1794721be0b5a09f906dff3492f8a199a5f995db362d2a
        version: 1.11.20+xdb5ed32
      entries:
      - name: cilium.v1.11.20-xdb5ed32
        version: 1.11.20+xdb5ed32
      name: "1.11"
    - currentCSV: cilium.v1.12.19-xef55de8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.12.0 <1.12.19+xef55de8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/hubble-ui-backend@sha256:f88a73a120521eeafbcdbd908f517117f6557bf61e115847853fac371f0d774c
        - quay.io/cilium/hubble-relay@sha256:19f2dee371523888985e6e34b30d07ef3f84f2a9206ad79bdfc8dede96e25d5c
        - quay.io/cilium/clustermesh-apiserver@sha256:61f8e871760dceaa0c0459bbf44a3049dd0f4ab15e57e063d9b3284434cb3dd8
        - quay.io/cilium/hubble-ui@sha256:e15af59a2ded739e420be82de6cbdd0ce22f8d3f00e3a10b3d2d2734e074a394
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/cilium/cilium@sha256:b997994ab759443d9fc9807eb39b4784bf03e13730a8132ffe5df3edef22a312
        - quay.io/cilium/operator-generic@sha256:da4d72facecde862c88d08bf99c3f476845beb407052a0cc4c06b68cc93b7570
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:d35994cf5cf319b88ff5a7985cc972edd23df5a282360f9ba5516d610390ebf8
        version: 1.12.19+xef55de8
      entries:
      - name: cilium.v1.12.19-xef55de8
        version: 1.12.19+xef55de8
      - name: cilium.v1.12.18-x63f4f4c
        version: 1.12.18+x63f4f4c
      - name: cilium.v1.12.17-x6cc0c18
        version: 1.12.17+x6cc0c18
      - name: cilium.v1.12.16-x14b1d1d
        version: 1.12.16+x14b1d1d
      - name: cilium.v1.12.15-x7d84617
        version: 1.12.15+x7d84617
      - name: cilium.v1.12.14-x17ed60a
        version: 1.12.14+x17ed60a
      - name: cilium.v1.12.13-xb414403
        version: 1.12.13+xb414403
      - name: cilium.v1.12.12-xd280cc3
        version: 1.12.12+xd280cc3
      - name: cilium.v1.12.11-xc8fb7eb
        version: 1.12.11+xc8fb7eb
      - name: cilium.v1.12.10-xe372776
        version: 1.12.10+xe372776
      - name: cilium.v1.12.9-x2ccb2b6
        version: 1.12.9+x2ccb2b6
      - name: cilium.v1.12.8-xa53bac1
        version: 1.12.8+xa53bac1
      - name: cilium.v1.12.7-xba75145
        version: 1.12.7+xba75145
      - name: cilium.v1.12.0-xb42b669
        version: 1.12.0+xb42b669
      name: "1.12"
    - currentCSV: cilium.v1.13.12-x6e56d47
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.13.12+x6e56d47'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:b0d22d4376b4c2b2b3b988955227f1ed2ad49a363c5367e53f64edae2cec0ff5
        - quay.io/cilium/cilium@sha256:d99204aa7b3b7bd2c9ab47fd398cc9f40290799bc0c7a4386c8dc5c1780cd3d3
        - quay.io/cilium/clustermesh-apiserver@sha256:f1b1d0a85bab65e7d6adc90d000513a56ac58bdb071aa391a8580d73d20e6b6a
        - quay.io/cilium/operator-generic@sha256:f83734bbe270f961d545c7929152785507ce04a05d818ebc9776941723736d02
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/hubble-relay@sha256:01b23ea40bcd81145dde6bfcbfc4d542749d08c2a1c6348954c85123a8d2b1fc
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        version: 1.13.12+x6e56d47
      entries:
      - name: cilium.v1.13.12-x6e56d47
        version: 1.13.12+x6e56d47
      - name: cilium.v1.13.11-xba1ab0b
        version: 1.13.11+xba1ab0b
      - name: cilium.v1.13.10-x412ae64
        version: 1.13.10+x412ae64
      - name: cilium.v1.13.9-x3ae7099
        version: 1.13.9+x3ae7099
      - name: cilium.v1.13.8-xe0355c7
        version: 1.13.8+xe0355c7
      - name: cilium.v1.13.7-x37f9269
        version: 1.13.7+x37f9269
      - name: cilium.v1.13.6-xdb5afcd
        version: 1.13.6+xdb5afcd
      - name: cilium.v1.13.5-xaf5caa6
        version: 1.13.5+xaf5caa6
      - name: cilium.v1.13.4-x129f2a7
        version: 1.13.4+x129f2a7
      - name: cilium.v1.13.3-x5854d64
        version: 1.13.3+x5854d64
      - name: cilium.v1.13.2-x82c79c4
        version: 1.13.2+x82c79c4
      - name: cilium.v1.13.1-x10a5de9
        version: 1.13.1+x10a5de9
      - name: cilium.v1.13.0-x32540df
        version: 1.13.0+x32540df
      name: "1.13"
    - currentCSV: cilium.v1.14.7-x3522df7
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.14.7+x3522df7'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:291a05c58de8d7daefeb0689c23cadd92c9368aea09751de70eaa817f7fcc358
        - quay.io/cilium/operator-generic@sha256:37ef0bd85c27c765c637cd58c3ff4a559f8734ae39f9d1839a3ac7803de7b952
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/cilium@sha256:45ce2b87696082ecf7d53ba1c64ceeb4217578033e5ef28ac479ec049a48bc32
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/hubble-relay@sha256:46762393daf4a0aaef76b106614c2615942f98f10aeacd435ea3fb1a0bdf69e4
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/clustermesh-apiserver@sha256:28f3ffe53365ca79831af600f09a95c0b3e9959f5f891b416dab8cedd90c263d
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        version: 1.14.7+x3522df7
      entries:
      - name: cilium.v1.14.7-x3522df7
        version: 1.14.7+x3522df7
      - name: cilium.v1.14.6-x11464e1
        version: 1.14.6+x11464e1
      - name: cilium.v1.14.5-x3583390
        version: 1.14.5+x3583390
      - name: cilium.v1.14.4-xb89471d
        version: 1.14.4+xb89471d
      - name: cilium.v1.14.3-x041dc10
        version: 1.14.3+x041dc10
      - name: cilium.v1.14.2-xedd5ce2
        version: 1.14.2+xedd5ce2
      - name: cilium.v1.14.1-xdd859b7
        version: 1.14.1+xdd859b7
      name: "1.14"
    - currentCSV: cilium.v1.15.1-x7095b76
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.15.1+x7095b76'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:9ab6be29447125e886300e9258b9a06bedf0a9d87405832aa8b6565ed1ba4215
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        - quay.io/cilium/cilium@sha256:351d6685dc6f6ffbcd5451043167cfa8842c6decf80d8c8e426a417c73fb56d4
        - quay.io/cilium/clustermesh-apiserver@sha256:b353badd255c2ce47eaa8f394ee4cbf70666773d7294bd887693e0c33503dc37
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/operator-generic@sha256:819c7281f5a4f25ee1ce2ec4c76b6fbc69a660c68b7825e9580b1813833fa743
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/hubble-relay@sha256:3254aaf85064bc1567e8ce01ad634b6dd269e91858c83be99e47e685d4bb8012
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        version: 1.15.1+x7095b76
      entries:
      - name: cilium.v1.15.1-x7095b76
        version: 1.15.1+x7095b76
      name: "1.15"
    - currentCSV: cilium.v1.9.14-x70dc065
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Basic Install
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/hubble-relay@sha256:fd6ab1aea260abc5f64eca26c1b1e7009983e4aaa8e5d098e8d442f7659603fb
        - quay.io/cilium/clustermesh-apiserver@sha256:a0da5edf0372899647da51de1b277f0bab8e676d694aee7f939cddfdd3172010
        - quay.io/cilium/hubble-ui-backend@sha256:2cf1c7b1fde6161f575f3147c2ef885739e60b8c59a1ed444fe74edbdf22c7c1
        - quay.io/cilium/cilium@sha256:2c6ce93fa7e625979043a387eb998c17ad57df8768d89facb9b715da42a4c51c
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:3a9ca9a4c2f78109f265c4758f5f693d01865f2d0e31bc5f1ecbc065e878628b
        - docker.io/envoyproxy/envoy@sha256:55a141c0b8e39497b32074cb758b4e5dc2384bd6d9926144969a4452eaa611fd
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/hubble-ui@sha256:c267219946b84f79a5fd2849f5cc71f39453e29d93039e57b9cef2b5802e9c50
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/cilium/operator-generic@sha256:bdcfd2eade99933f2fda55ef79ea697ddfad3512f65b15bcd0ba7702518c1ba3
        version: 1.9.14+x70dc065
      entries:
      - name: cilium.v1.9.14-x70dc065
        version: 1.9.14+x70dc065
      name: "1.9"
    - currentCSV: cilium.v1.10.4-x5bfd7b3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Basic Install
          categories: Networking,Security
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:b45ff6a6552d71cbc26e358e0b25759d33818fd92301f1656453c5bcf826ff13
        version: 1.10.4+x5bfd7b3
      entries:
      - name: cilium.v1.10.4-x5bfd7b3
        version: 1.10.4+x5bfd7b3
      name: stable
    defaultChannel: "1.15"
    packageName: cilium
    provider:
      name: Isovalent
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Couchbase
      provider-url: ""
    name: couchbase-enterprise-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: couchbase-operator.v2.6.4-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator@sha256:1f0a6865c360a14ba98706fd6fe8dff3bdbdede32e3622f93f19c82e377ca33f
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        version: 2.6.4-7
      entries:
      - name: couchbase-operator.v2.6.4-7
        version: 2.6.4-7
      - name: couchbase-operator.v2.6.4-6
        version: 2.6.4-6
      name: 2.6.4
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator@sha256:707571a7881bc12d5f1b0b1a94eb3c675d5c0ea0114c994a47ceb5aa79df4013
        version: 2.7.0-5
      entries:
      - name: couchbase-operator.v2.7.0-5
        version: 2.7.0-5
      name: 2.7.0
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:a79e1b6781d5f8badf664e7efea6e2889e43fae6e0d37169d766685dcd6ff394
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        version: 2.7.1-1
      entries:
      - name: couchbase-operator.v2.7.1-1
        version: 2.7.1-1
      name: 2.7.1
    - currentCSV: couchbase-operator.v2.8.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:df61928b064014d851c6e7dff6dcd63a433a27fa5623254b6e389c3eea170e36
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        version: 2.8.0-2
      entries:
      - name: couchbase-operator.v2.8.0-2
        version: 2.8.0-2
      - name: couchbase-operator.v2.8.0-1
        version: 2.8.0-1
      name: 2.8.0
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        version: 2.8.1-1
      entries:
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: 2.8.1
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        version: 2.8.1-1
      entries:
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: stable
    defaultChannel: stable
    packageName: couchbase-enterprise-certified-rhmp
    provider:
      name: Couchbase
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: topology-aware-lifecycle-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: topology-aware-lifecycle-manager.v4.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ran.openshift.io/v1alpha1",
                "kind": "ClusterGroupUpgrade",
                "metadata": {
                  "name": "ClusterGroupUpgrade-sample"
                },
                "spec": {
                  "preCaching": true
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:bf24cad2940a227f5bd2f188be5d298ff19176c634fcf609c7225d7e0a6f2fd1
          description: The topology aware lifecycle manager is an operator that facilitates
            platform upgrades of group of clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-operators",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/cluster-group-upgrades-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterGroupUpgrade is the Schema for the ClusterGroupUpgrades
              API
            displayName: Cluster Group Upgrade
            kind: ClusterGroupUpgrade
            name: clustergroupupgrades.ran.openshift.io
            version: v1alpha1
          - description: ImageBasedGroupUpgrade is the schema for upgrading a group
              of clusters using IBU
            displayName: Image-Based Group Upgrade
            kind: ImageBasedGroupUpgrade
            name: imagebasedgroupupgrades.lcm.openshift.io
            version: v1alpha1
          - description: PreCachingConfig is the Schema for the precachingconfigs
              API
            displayName: Pre-caching Config
            kind: PreCachingConfig
            name: precachingconfigs.ran.openshift.io
            version: v1alpha1
        description: |-
          Topology Aware Lifecycle Manager is an operator that facilitates
          platform and operator upgrades of group of clusters
        displayName: Topology Aware Lifecycle Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm lifecyclemanagement upgrades cluster
        links:
        - name: Cluster Group Upgrades Operator
          url: https://cluster-group-upgrades-operator.domain
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:eed8234b0fe47083ab941f513709f778bfcb80e32747b71a65fd5f62328a5ca0
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:bf24cad2940a227f5bd2f188be5d298ff19176c634fcf609c7225d7e0a6f2fd1
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-aztp-rhel9@sha256:f6a37a0de10c847a53e22406990858dbaa4cdd3be075bc047e8dcf5bc6a2f2f7
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-precache-rhel9@sha256:7673a0d7f8944b245124376f2153bf987299f5bd7350b33f0d4bd1ac18543929
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-recovery-rhel9@sha256:37e234f247c841cd12c4617f92bf8bd5a5dbcbf46fa1be042d56b966994efaad
        version: 4.16.5
      entries:
      - name: topology-aware-lifecycle-manager.v4.16.5
        version: 4.16.5
      - name: topology-aware-lifecycle-manager.v4.16.4
        version: 4.16.4
      - name: topology-aware-lifecycle-manager.v4.16.3
        version: 4.16.3
      - name: topology-aware-lifecycle-manager.v4.16.2
        version: 4.16.2
      - name: topology-aware-lifecycle-manager.v4.16.1
        version: 4.16.1
      - name: topology-aware-lifecycle-manager.v4.16.0
        version: 4.16.0
      name: "4.16"
    - currentCSV: topology-aware-lifecycle-manager.v4.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ran.openshift.io/v1alpha1",
                "kind": "ClusterGroupUpgrade",
                "metadata": {
                  "name": "ClusterGroupUpgrade-sample"
                },
                "spec": {
                  "preCaching": true
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:bf24cad2940a227f5bd2f188be5d298ff19176c634fcf609c7225d7e0a6f2fd1
          description: The topology aware lifecycle manager is an operator that facilitates
            platform upgrades of group of clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-operators",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/cluster-group-upgrades-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterGroupUpgrade is the Schema for the ClusterGroupUpgrades
              API
            displayName: Cluster Group Upgrade
            kind: ClusterGroupUpgrade
            name: clustergroupupgrades.ran.openshift.io
            version: v1alpha1
          - description: ImageBasedGroupUpgrade is the schema for upgrading a group
              of clusters using IBU
            displayName: Image-Based Group Upgrade
            kind: ImageBasedGroupUpgrade
            name: imagebasedgroupupgrades.lcm.openshift.io
            version: v1alpha1
          - description: PreCachingConfig is the Schema for the precachingconfigs
              API
            displayName: Pre-caching Config
            kind: PreCachingConfig
            name: precachingconfigs.ran.openshift.io
            version: v1alpha1
        description: |-
          Topology Aware Lifecycle Manager is an operator that facilitates
          platform and operator upgrades of group of clusters
        displayName: Topology Aware Lifecycle Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm lifecyclemanagement upgrades cluster
        links:
        - name: Cluster Group Upgrades Operator
          url: https://cluster-group-upgrades-operator.domain
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:eed8234b0fe47083ab941f513709f778bfcb80e32747b71a65fd5f62328a5ca0
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:bf24cad2940a227f5bd2f188be5d298ff19176c634fcf609c7225d7e0a6f2fd1
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-aztp-rhel9@sha256:f6a37a0de10c847a53e22406990858dbaa4cdd3be075bc047e8dcf5bc6a2f2f7
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-precache-rhel9@sha256:7673a0d7f8944b245124376f2153bf987299f5bd7350b33f0d4bd1ac18543929
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-recovery-rhel9@sha256:37e234f247c841cd12c4617f92bf8bd5a5dbcbf46fa1be042d56b966994efaad
        version: 4.16.5
      entries:
      - name: topology-aware-lifecycle-manager.v4.16.5
        version: 4.16.5
      - name: topology-aware-lifecycle-manager.v4.16.4
        version: 4.16.4
      - name: topology-aware-lifecycle-manager.v4.16.3
        version: 4.16.3
      - name: topology-aware-lifecycle-manager.v4.16.2
        version: 4.16.2
      - name: topology-aware-lifecycle-manager.v4.16.1
        version: 4.16.1
      - name: topology-aware-lifecycle-manager.v4.16.0
        version: 4.16.0
      name: stable
    defaultChannel: stable
    packageName: topology-aware-lifecycle-manager
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Korrel8r
      provider-url: https://github.com/korrel8r/korrel8r
    name: korrel8r
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: korrel8r.v0.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "korrel8r.openshift.io/v1alpha1",
                "kind": "Korrel8r",
                "metadata": {
                  "name": "korrel8r",
                  "namespace": "korrel8r"
                },
                "spec": {
                  "debug": {
                    "verbose": 1
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional,Monitoring,Logging & Tracing
          containerImage: quay.io/korrel8r/operator:0.1.8
          createdAt: "2024-09-27T17:45:00Z"
          description: Correlate cluster resources and observability data.
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "korrel8r.openshift.io/v1alpha1",
              "kind": "Korrel8r",
              "metadata": {
                "name": "korrel8r",
                "namespace": "korrel8r"
              }
            }
          operatorframework.io/suggested-namespace: korrel8r
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/korrel8r/korrel8r
          support: Korrel8r Project
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Korrel8r is a service that correlates observabililty signals
              in the cluster.
            displayName: Korrel8r
            kind: Korrel8r
            name: korrel8rs.korrel8r.openshift.io
            version: v1alpha1
        description: |
          Korrel8r helps navigate from problem symptoms to related resources and signal data that can reveal the cause.
          It can follow relationships between disjointed observability "silos" (logs, metrics, alerts and more) to bring together all the data available to solve a problem.
          This operator allows Korrel8r to be deployed as a cluster service with a REST API.
          It can be used as a back-end for data correlation by the OpenShift console and other tools.
        displayName: Korrel8r
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - signal
        - correlation
        links:
        - name: Documentation
          url: https://korrel8r.github.io/korrel8r
        - name: Source Code and Issues
          url: https://github.com/korrel8r/korrel8r
        maintainers:
        - email: korrel8r@gmail.com
          name: Korrel8r Project
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Korrel8r
          url: https://github.com/korrel8r/korrel8r
        relatedImages:
        - quay.io/korrel8r/operator:0.1.8
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 0.1.8
      entries:
      - name: korrel8r.v0.1.8
        version: 0.1.8
      - name: korrel8r.v0.1.7
        version: 0.1.7
      - name: korrel8r.v0.1.5
        version: 0.1.5
      - name: korrel8r.v0.1.3
        version: 0.1.3
      - name: korrel8r.v0.1.2
        version: 0.1.2
      - name: korrel8r.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: korrel8r
    provider:
      name: Korrel8r
      url: https://github.com/korrel8r/korrel8r
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app: enmasse
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq-online
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-online.1.7.4-0.1657725339.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "StandardInfraConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "broker": {
                    "resources": {
                      "memory": "1Gi",
                      "storage": "5Gi"
                    },
                    "addressFullPolicy": "FAIL"
                  },
                  "router": {
                    "linkCapacity": 50,
                    "resources": {
                      "memory": "512Mi"
                    }
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "BrokeredInfraConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "broker": {
                    "resources": {
                      "memory": "4Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta2",
                "kind": "AddressPlan",
                "metadata": {
                  "name": "standard-small-queue"
                },
                "spec": {
                  "addressType": "queue",
                  "shortDescription": "Small Queue",
                  "resources": {
                    "router": 0.01,
                    "broker": 0.1
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta2",
                "kind": "AddressSpacePlan",
                "metadata": {
                  "name": "standard-small"
                },
                "spec": {
                  "addressSpaceType": "standard",
                  "infraConfigRef": "default",
                  "shortDescription": "Small Address Space Plan",
                  "resourceLimits": {
                    "router": 1.0,
                    "broker": 2.0,
                    "aggregate": 2.0
                  },
                  "addressPlans": [
                    "standard-small-queue"
                  ]
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "AuthenticationService",
                "metadata": {
                  "name": "standard-authservice"
                },
                "spec": {
                  "type": "standard",
                  "standard": {
                    "storage": {
                      "claimName": "standard-authservice",
                      "deleteClaim": true,
                      "size": "1Gi",
                      "type": "persistent-claim"
                    }
                  }
                }
              },
              {
                "apiVersion": "enmasse.io/v1beta1",
                "kind": "AddressSpace",
                "metadata": {
                  "name": "myspace"
                },
                "spec": {
                  "plan": "standard-small",
                  "type": "standard"
                }
              },
              {
                "apiVersion": "enmasse.io/v1beta1",
                "kind": "Address",
                "metadata": {
                  "name": "myspace.myqueue"
                },
                "spec": {
                  "address": "myqueue",
                  "plan": "standard-small-queue",
                  "type": "queue"
                }
              },
              {
                "apiVersion": "user.enmasse.io/v1beta1",
                "kind": "MessagingUser",
                "metadata": {
                  "name": "myspace.user"
                },
                "spec": {
                  "authentication": {
                    "password": "ZW5tYXNzZQ==",
                    "type": "password"
                  },
                  "authorization": [
                    {
                      "addresses": [
                        "myqueue"
                      ],
                      "operations": [
                        "send",
                        "recv"
                      ]
                    }
                  ],
                  "username": "user"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator@sha256:2aaf87224a96aa56dc595e96d8e7ec8d138aaf20240afa25a1e805b405941355
          createdAt: "2019-02-19T00:00:00Z"
          description: Red Hat Integration - AMQ Online provides messaging as a managed
            service on OpenShift
          olm.substitutesFor: amq-online.1.7.4
          operators.openshift.io/capability: '["license-required"]'
          operators.openshift.io/valid-licenses: '["Red Hat AMQ", "Red Hat Integration"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: ""
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A messaging user that can connect to an Address Space
            displayName: Messaging User
            kind: MessagingUser
            name: messagingusers.user.enmasse.io
            version: v1beta1
          - description: A resource representing the available schema of plans and
              authentication services
            displayName: AddressSpaceSchema
            kind: AddressSpaceSchema
            name: addressspaceschemas.enmasse.io
            version: v1beta1
          - description: A group of messaging addresses that can be accessed via the
              same endpoint
            displayName: Address Space
            kind: AddressSpace
            name: addressspaces.enmasse.io
            version: v1beta1
          - description: A messaging address that can be used to send/receive messages
              to/from
            displayName: Address
            kind: Address
            name: addresses.enmasse.io
            version: v1beta1
          - description: Infrastructure configuration template for the standard address
              space type
            displayName: Standard Infra Config
            kind: StandardInfraConfig
            name: standardinfraconfigs.admin.enmasse.io
            version: v1beta1
          - description: Infrastructure configuration template for the brokered address
              space type
            displayName: Brokered Infra Config
            kind: BrokeredInfraConfig
            name: brokeredinfraconfigs.admin.enmasse.io
            version: v1beta1
          - description: Plan describing the resource usage of a given address type
            displayName: Address Plan
            kind: AddressPlan
            name: addressplans.admin.enmasse.io
            version: v1beta2
          - description: Plan describing the capabilities and resource limits of a
              given address space type
            displayName: Address Space Plan
            kind: AddressSpacePlan
            name: addressspaceplans.admin.enmasse.io
            version: v1beta2
          - description: Authentication service configuration available to address
              spaces.
            displayName: Authentication Service
            kind: AuthenticationService
            name: authenticationservices.admin.enmasse.io
            version: v1beta1
          - description: Console Service Singleton for deploying global console.
            displayName: Console Service
            kind: ConsoleService
            name: consoleservices.admin.enmasse.io
            version: v1beta1
        description: "**Red Hat Integration - AMQ Online** provides messaging as a
          managed service on OpenShift.\nWith Red Hat Integration - AMQ Online, administrators
          can configure a cloud-native,\nmulti-tenant messaging service either in
          the cloud or on premise.\nDevelopers can provision messaging using the Red
          Hat Integration - AMQ Online Console.\nMultiple development teams can provision
          the brokers and queues from the\nconsole, without requiring each team to
          install, configure, deploy,\nmaintain, or patch any software. \n\n**The
          core capabilities include:**\n\n  * **Built-in authentication and authorization**
          - Use the built-in authentication service or\n    plug in your own authentication
          service for authentication and\n    authorization of messaging clients.\n\n
          \ * **Self-service messaging for applications** - The service administrator
          deploys\n    and manages the messaging infrastructure, while applications
          can request\n    messaging resources regardless of the messaging infrastructure.\n\n
          \ * **Support for a wide variety of messaging patterns** - Choose between\n
          \   JMS-style messaging with strict guarantees, or messaging that supports\n
          \   a larger number of connections and higher throughput.\n\n## Post-installation
          tasks\n\nTo fully use Red Hat Integration - AMQ Online, you need to create
          a few\ninfrastructure components after the installation, including:\n\n
          \ * An authentication service\n  * Infrastructure configuration for the
          standard and brokered address space\n  * Address and address space plans\n
          \ * (Optional) Create RBAC roles to allow users to discover available plans\n
          \ * (Optional) Create RBAC roles to allow users to self-manage addresses
          and\n    address spaces.\n\nFor a complete overview of how to configure
          Red Hat Integration - AMQ Online, see\n[Configuring Red Hat Integration
          - AMQ Online](https://access.redhat.com/documentation/en-us/red_hat_amq/7.7/html/installing_and_managing_amq_online_on_openshift/configuring-messaging).\n\n###
          Quickstart infrastructure configuration\n\nIf you simply want to get started
          quickly, you can import the following\nYAML by saving the content to a file
          and apply it by running the \n`oc apply -f <file>` command. You can also
          split the content (at the `---` marker)\nand import the single YAML document
          using the Web UI: \n\n~~~yaml\n---\napiVersion: admin.enmasse.io/v1beta1\nkind:
          StandardInfraConfig\nmetadata:\n  name: default\n---\napiVersion: admin.enmasse.io/v1beta2\nkind:
          AddressPlan\nmetadata:\n  name: standard-small-queue\nspec:\n  addressType:
          queue\n  resources:\n    router: 0.01\n    broker: 0.1\n---\napiVersion:
          admin.enmasse.io/v1beta2\nkind: AddressSpacePlan\nmetadata:\n  name: standard-small\nspec:\n
          \ addressSpaceType: standard\n  infraConfigRef: default\n  addressPlans:\n
          \ - standard-small-queue\n  resourceLimits:\n    router: 2.0\n    broker:
          3.0\n    aggregate: 4.0\n---\napiVersion: admin.enmasse.io/v1beta1\nkind:
          AuthenticationService\nmetadata:\n  name: standard-authservice\nspec:\n
          \ type: standard\n  standard:\n    storage:\n      claimName: standard-authservice\n
          \     deleteClaim: true\n      size: 1Gi\n      type: persistent-claim\n~~~\n\n###
          Create RBAC roles to allow users to discover available plans\n\nFor users
          to discover the available plans, cluster-wide roles to read the available\nschema
          can be created.  Import the following YAML by saving the content to a file
          and apply it by running the \n`oc apply -f <file>` command. You can also
          split the content (at the `---` marker)\nand import the single YAML document
          using the Web UI: \n\n~~~yaml\n---\napiVersion: rbac.authorization.k8s.io/v1\nkind:
          ClusterRole\nmetadata:\n  name: enmasse.io:schema\n  labels:\n    app: enmasse\nrules:\n
          \ - apiGroups: [ \"enmasse.io\" ]\n    resources: [ \"addressspaceschemas\"
          ]\n    verbs: [ \"get\", \"list\", \"watch\" ]\n---\napiVersion: rbac.authorization.k8s.io/v1\nkind:
          ClusterRoleBinding\nmetadata:\n  name: \"enmasse.io:schema\"\n  labels:\n
          \   app: enmasse\nroleRef:\n  apiGroup: rbac.authorization.k8s.io\n  kind:
          ClusterRole\n  name: enmasse.io:schema\nsubjects:\n- kind: Group\n  name:
          system:authenticated\n~~~\n\n### Creating infrastructure using the Web UI\n\nYou
          must create a new instance of each of the following custom resources. You
          can\nuse the example data directly, which is provided when using the\nWeb
          UI:\n\n  * *Authentication Service* – Create an authentication service.\n
          \ * *Brokered Infra Config* – Create the broker infrastructure configuration.\n
          \ * *Standard Infra Config* – Create the standard infrastructure\n    configuration.\n\nYou
          must also create at least one address space plan and one address plan.\n\n*Note*:
          The name of the address space plan and address plan is required\nlater when
          creating new addresses. Some examples use specific plan\nnames, which might
          not be available in your environment. You can\ncreate those plans, or edit
          the examples to use different plan names.\n\n## Configuration for messaging
          tenants\n\nWhile service administrators perform the infrastructure configuration,
          the following\nresources are for the actual users of the system, the messaging
          tenants.\n\nYou need to create those resources to satisfy your particular
          use case.\n\n  * *Address space* – A container for addresses\n  * *Address*
          – A messaging address (address, topic, queue, and so on)\n  * *Messaging
          user* – Manages access to an address\n"
        displayName: Red Hat Integration - AMQ Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amq
        - messaging
        - amqp
        - iot
        - mqtt
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/7.7/html/using_amq_online_on_openshift/index
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-online-1-topic-forwarder@sha256:de33a701339d91a4e9275f7cf65f917ff75368a59b8ab36dd66c90bc306875aa
        - registry.redhat.io/amq7/amq-online-1-broker-plugin@sha256:2466a336c0d78b196635de3fad636aeff2d9265839f7271b8ef94f3ba97d9bc0
        - registry.redhat.io/amq7/amq-interconnect@sha256:5ebd4c4a0ed0db2725ae466550c915531b32933c420344cc2083c616da3f326f
        - registry.redhat.io/redhat-sso-7/sso73-openshift@sha256:18d15980f28f0d0c6b8582131a1f062d84463577512875b5391a612c3d7d1096
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6d489a7897bd831ec273034f1ff0e223427a40afd01b9835ebb020a8bf9ddc66
        - registry.redhat.io/amq7/amq-broker@sha256:1ec0437bbd82722148fdbb02dc55682b114c311cd6f521a8bc621a8c1cb4b32b
        - registry.redhat.io/amq7/amq-online-1-address-space-controller@sha256:9acc555c8e869aec7d72f53b60320e1a96cc82d748638e0d50970837229fdbee
        - registry.redhat.io/amq7/amq-online-1-none-auth-service@sha256:a013fb1806c002b408a7753ff943cee24fbcdf8a657eff1e9582c42b5a7b0844
        - registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator@sha256:2aaf87224a96aa56dc595e96d8e7ec8d138aaf20240afa25a1e805b405941355
        - registry.redhat.io/amq7/amq-online-1-standard-controller@sha256:070349af30cec52fa8efc185fc940077a64bb40e3e36310d7cc6e39ce0971436
        - registry.redhat.io/amq7/amq-online-1-console-init@sha256:0970804909b82fdc3d4e7d5d88280e41157d743107de8bfa2d1baa1367543176
        - registry.redhat.io/openshift3/oauth-proxy@sha256:59b20f23c03d0012ec5c95c4725892ee64277c22dedc9d0a90a7b6b0bc80fc72
        - registry.redhat.io/amq7/amq-online-1-console-server-rhel7@sha256:1dcb3d0cc7bc6e84f4f96962ff88849511b7a7686a1f5dbd9300c4c2c8c05e2f
        - registry.redhat.io/amq7/amq-online-1-agent@sha256:f98ed8ee97ac59d8fdd40ec8b2c063a5afd360a3e183057c186e7a77a32801ee
        - registry.redhat.io/amq7/amq-online-1-auth-plugin@sha256:68926f434b2f6ce4d9698938bc3c53d514c21c3f76c73b6b4cdf7fcf51f7cb0d
        version: 1.7.4+0.1657725339.p
      entries:
      - name: amq-online.1.7.4
        version: 1.7.4
      - name: amq-online.1.7.4-0.1657725339.p
        version: 1.7.4+0.1657725339.p
      - name: amq-online.1.7.3
        version: 1.7.3
      - name: amq-online.1.7.2
        version: 1.7.2
      - name: amq-online.1.7.1
        version: 1.7.1
      - name: amq-online.1.7.1-0.1628610187.p
        version: 1.7.1+0.1628610187.p
      - name: amq-online.1.7.0
        version: 1.7.0
      name: stable
    defaultChannel: stable
    packageName: amq-online
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: sandboxed-containers-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sandboxed-containers-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:051bd7f1dad8cc3251430fee32184be8d64077aba78580184cef0255d267bdcf
          createdAt: 2021-06-07-13:38:06
          olm.skipRange: <1.0.2
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          OpenShift sandboxed containers based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift’s Defense-in-Depth strategy. For more information
          [see](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5).

          # Features & benefits
          - **Isolated Developer Environments & Priviliges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as CAP_ADMIN or CAP_BPF. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select preview-1.0 from the list of available Update Channel options.
            This ensures that you install the version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              openshift-sandboxed-containers-operator namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              openshift-sandboxed-containers-operator causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a KataConfig CRD instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.8/sandboxed_containers/understanding-sandboxed-containers.html).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kernel-isolated containers
        links:
        - name: OpenShift sandboxed containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        provider:
          name: Red Hat
          url: https://www.github.com/openshift/sandboxed-containers-operator
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:051bd7f1dad8cc3251430fee32184be8d64077aba78580184cef0255d267bdcf
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:c6c589d5e47ba9564c66c84fc2bc7e5e046dae1d56a3dc99d7343f01e42e4d31
        version: 1.0.2
      entries:
      - name: sandboxed-containers-operator.v1.0.2
        version: 1.0.2
      name: preview-1.0
    - currentCSV: sandboxed-containers-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:b1e824e126c579db0f56d04c3d1796d82ed033110c6bc923de66d95b67099611
          createdAt: "2021-10-08T13:20:20.250789Z"
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift’s Defense-in-Depth strategy. For more information
          [see](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5).

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `preview-1.1` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CRD instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.8/sandboxed_containers/understanding-sandboxed-containers.html).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kernel-isolated containers
        links:
        - name: OpenShift sandboxed containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        provider:
          name: Red Hat
          url: https://www.github.com/openshift/sandboxed-containers-operator
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:b1e824e126c579db0f56d04c3d1796d82ed033110c6bc923de66d95b67099611
        version: 1.1.0
      entries:
      - name: sandboxed-containers-operator.v1.1.0
        version: 1.1.0
      - name: sandboxed-containers-operator.v1.0.2
        version: 1.0.2
      name: preview-1.1
    - currentCSV: sandboxed-containers-operator.v1.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-12T14:30:28Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.10.3'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '["peerpods.confidentialcontainers.org","peerpodconfigs.confidentialcontainers.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/sandboxed-containers-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
          - kind: PeerPod
            name: peerpods.confidentialcontainers.org
            version: v1alpha1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `candidate` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the [OpenShift sandboxed containers documentation](https://docs.redhat.com/en/documentation/openshift_sandboxed_containers/).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-cloud-api-adaptor-rhel9@sha256:7d1fa1e85af0d34c23e75868b1388eff17abe633c04a8dc55fbefcc470195c05
        - registry.redhat.io/openshift-sandboxed-containers/osc-cloud-api-adaptor-webhook-rhel9@sha256:6802f00a939f24ecd6dad150433db5ea3fd72d171f0c410cb894d3af910fc7fd
        - registry.redhat.io/openshift-sandboxed-containers/osc-podvm-builder-rhel9@sha256:c15f36a639fa07fee11f74c6563696ec8ef4da3fb1a99b7bd3f308dd04881339
        - registry.redhat.io/openshift-sandboxed-containers/osc-podvm-payload-rhel9@sha256:8d6b1262c6cedd04e33441264539842d82ed057f64eda5c4bce59566bcb72558
        - registry.redhat.io/openshift-sandboxed-containers/osc-dm-verity-image@sha256:e633bc27853f811d919e64461d0c917e659f5430458abe2304a4c85b0d871c80
        - registry.redhat.io/openshift-sandboxed-containers/osc-monitor-rhel9@sha256:037cbc221ae8fda83cdf1093127795ea806ecda53893deb49cdf981e7863774c
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel9-operator@sha256:1ffb432d9a8833854fa5e5262dcbf42530dfbaadfddf6978f7bc7804b84d2e3f
        version: 1.10.3
      entries:
      - name: sandboxed-containers-operator.v1.10.3
        version: 1.10.3
      - name: sandboxed-containers-operator.v1.10.2
        version: 1.10.2
      - name: sandboxed-containers-operator.v1.10.1
        version: 1.10.1
      - name: sandboxed-containers-operator.v1.10.0
        version: 1.10.0
      - name: sandboxed-containers-operator.v1.9.0
        version: 1.9.0
      - name: sandboxed-containers-operator.v1.8.1
        version: 1.8.1
      - name: sandboxed-containers-operator.v1.7.0
        version: 1.7.0
      - name: sandboxed-containers-operator.v1.6.0
        version: 1.6.0
      - name: sandboxed-containers-operator.v1.5.3
        version: 1.5.3
      - name: sandboxed-containers-operator.v1.5.2
        version: 1.5.2
      - name: sandboxed-containers-operator.v1.5.1
        version: 1.5.1
      - name: sandboxed-containers-operator.v1.5.0
        version: 1.5.0
      name: stable
    - currentCSV: sandboxed-containers-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:8f30a9129d817c3f4e404d2c43fb47e196d8c8da3badba4c48f65d440a4d7584
          createdAt: "2021-11-22T14:44:50.817336Z"
          olm.skipRange: '>=1.1.0 <1.2.2'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `stable-1.2` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.9/sandboxed_containers/understanding-sandboxed-containers.html)
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:a660f0b54b9139bed9a3aeef3408001c0d50ba60648364a98a09059b466fbcc1
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:8f30a9129d817c3f4e404d2c43fb47e196d8c8da3badba4c48f65d440a4d7584
        version: 1.2.2
      entries:
      - name: sandboxed-containers-operator.v1.2.2
        version: 1.2.2
      name: stable-1.2
    - currentCSV: sandboxed-containers-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          olm.skipRange: '>=1.1.0 <1.3.3'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/sandboxed-containers-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `stable-1.3` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/latest/sandboxed_containers/understanding-sandboxed-containers.html)
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:5d2b03721043e5221dfb0cf164cf59eba396ba3aae40a56c53aa3496c625eea0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:cb836456974e510eb4bccbffadbc6d99d5f57c36caec54c767a158ffd8a025d5
        version: 1.3.3
      entries:
      - name: sandboxed-containers-operator.v1.3.3
        version: 1.3.3
      name: stable-1.3
    defaultChannel: stable
    packageName: sandboxed-containers-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/stolostron/multicluster-global-hub
    name: multicluster-global-hub-operator-rh
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-global-hub-operator-rh.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:01e8dc62d5dc48af84b92f5315fab22574f648f2f705ac94f0516f4a7dab9237
          createdAt: 27 Aug 2024, 22:21
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.1.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```


          ### Documentation

          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: release-1.1
        provider:
          name: Red Hat
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:01e8dc62d5dc48af84b92f5315fab22574f648f2f705ac94f0516f4a7dab9237
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:78628c4f7a0286417023410261cc3b36797b7cdf72b38d0f4a2533a0f3db89dc
        - registry.redhat.io/rhel9/postgresql-13@sha256:c67d00643295697d196dd70dea457948ebb9e9d9d61a42019490b350c8286b50
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:5c8cea421be39712503378e32c54c332a7a0bf77a1705781d10fc123224bd60e
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:4ea8d3a5317d3b9968fa93dd68416e07e12e00e492860ff20ed90f88ca6ba225
        version: 1.1.2
      entries:
      - name: multicluster-global-hub-operator-rh.v1.1.2
        version: 1.1.2
      - name: multicluster-global-hub-operator-rh.v1.1.1
        version: 1.1.1
      - name: multicluster-global-hub-operator-rh.v1.1.0
        version: 1.1.0
      name: release-1.1
    - currentCSV: multicluster-global-hub-operator-rh.v1.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:98d470c8683c346107ab943776991554e3173c9a6431202bdacc48a75c4cf18f
          createdAt: 15 May 2025, 21:22
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          kafka-zookeeper-0                                       1/1     Running   0          21m
          kafka-zookeeper-1                                       1/1     Running   0          21m
          kafka-zookeeper-2                                       1/1     Running   0          21m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgres-0                      2/2     Running   0          22m
          ```


          ### Documentation

          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: release-1.2
        provider:
          name: Red Hat
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:c83a7b01bfe5a39d6cdf585180a011448e935d109768cb51560308b29d9ffe14
        - registry.redhat.io/rhel9/postgresql-13@sha256:b1af637cbe0e4041cd2fe6fdbcf665c5183e2991d3cc3e5b7384235f801504d9
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:16bb76ad0c68c0d6b3149b2741126a96e154150c1059621b14c2e6c30a214f16
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:98d470c8683c346107ab943776991554e3173c9a6431202bdacc48a75c4cf18f
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:92f44cff160874efe57a985609e599b13c49c0c4cef846dccc80833af2ae74f3
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:64e5ade4fbcd4cb70d62a6a62d775e051ff49838b891b682574646d4e61d9751
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:9b98fc01b08c3262e5ab99096e74fb55982f2a9c0de61e8fe84e419522e725b0
        version: 1.2.3
      entries:
      - name: multicluster-global-hub-operator-rh.v1.2.3
        version: 1.2.3
      - name: multicluster-global-hub-operator-rh.v1.2.2
        version: 1.2.2
      - name: multicluster-global-hub-operator-rh.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator-rh.v1.2.0
        version: 1.2.0
      name: release-1.2
    - currentCSV: multicluster-global-hub-operator-rh.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:0ef24ded54588aa49e366ba1cafc0971162e9d32024045bbcd7ef16e9f38a2d5
          createdAt: 01 Apr 2025, 20:25
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          kafka-zookeeper-0                                       1/1     Running   0          21m
          kafka-zookeeper-1                                       1/1     Running   0          21m
          kafka-zookeeper-2                                       1/1     Running   0          21m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgres-0                      2/2     Running   0          22m
          ```


          ### Documentation

          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: release-1.3
        provider:
          name: Red Hat
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:a139da58c681de987d847b233dd5c1e6ced5d6e86f3aab0f1b4181a77fc583d4
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:a17e90abc5e23bbc3908613c56c4208ad554eb27ed9679c4ccd06c4a1e9fcbcc
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:0ef24ded54588aa49e366ba1cafc0971162e9d32024045bbcd7ef16e9f38a2d5
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:619d8cb6b3811169af183bd8bc26ed98cdb25486ec428d88b25fee0ef9e6c552
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:e848a90d17e4f00d8021cbde47e488b82c447d8a5f6588f85f3d7cfbd6877f0f
        - registry.redhat.io/rhel9/postgresql-13@sha256:76c49b7dbd7c4ed8b430c4931b32f546db6d07ddc2146232f5d17eb17d3243b4
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-kessel-inventory-api-rhel9@sha256:dbd3924eb24db7ef28301b17851b5b5cd20e044cf259f0c556c67a0721561803
        version: 1.3.3
      entries:
      - name: multicluster-global-hub-operator-rh.v1.3.3
        version: 1.3.3
      - name: multicluster-global-hub-operator-rh.v1.3.2
        version: 1.3.2
      - name: multicluster-global-hub-operator-rh.v1.3.1
        version: 1.3.1
      - name: multicluster-global-hub-operator-rh.v1.3.0
        version: 1.3.0
      name: release-1.3
    - currentCSV: multicluster-global-hub-operator-rh.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:c5b3b3907e8135d3eedba40e27ba18a232ea4f30854d652059fcb14f82a8bf15
          createdAt: "2025-07-23T09:04:08Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: release-1.4
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:047b6c378e2b37339dd498c9025060f20812014bde4d0c98183f1a1201c21e66
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:ea906dcf8d45956251df22e269fe0b8fa8d49734e96a61b10db7b25d626b2e6a
        - registry.redhat.io/rhel9/postgresql-16@sha256:2161182c88022f38d6d9b8b467c0d59156edcededc908d383b80caa4eedf57f9
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:2be8bc79667018386235823e322ae10788e4e7e889c10e63566ae27a65dcabf1
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:0f4683d18404b4a9412e42ed417f6d3ac0024b3c57aa72c9162ff72d6ea8d394
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:c5b3b3907e8135d3eedba40e27ba18a232ea4f30854d652059fcb14f82a8bf15
        version: 1.4.2
      entries:
      - name: multicluster-global-hub-operator-rh.v1.4.2
        version: 1.4.2
      - name: multicluster-global-hub-operator-rh.v1.4.1
        version: 1.4.1
      - name: multicluster-global-hub-operator-rh.v1.4.0
        version: 1.4.0
      name: release-1.4
    - currentCSV: multicluster-global-hub-operator-rh.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:8af7e57b8c6285ad232084e7776fef32d026e080eb0bf14b91986ac28d00bb6a
          createdAt: "2025-08-18T08:30:32Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: release-1.5
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:8af7e57b8c6285ad232084e7776fef32d026e080eb0bf14b91986ac28d00bb6a
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:c95052569312d805b6750477c8bc57f7c72e1efd26143f3899ea0eddc2aed37b
        - registry.redhat.io/rhel9/postgresql-16@sha256:2161182c88022f38d6d9b8b467c0d59156edcededc908d383b80caa4eedf57f9
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:921163d139bca9e47bf95df43d7b408ff6fe2bf9b0902fb6af9f4cfc9d4598cc
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:ab3ed25a4448a7d545648ae4cea52b493918cc4d3b0c5adc56bb2510c0838c80
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:eae390ec42c3582f7896a8346df8782b287d64440794983ae93755490639e21e
        version: 1.5.1
      entries:
      - name: multicluster-global-hub-operator-rh.v1.5.1
        version: 1.5.1
      - name: multicluster-global-hub-operator-rh.v1.5.0
        version: 1.5.0
      name: release-1.5
    defaultChannel: release-1.5
    packageName: multicluster-global-hub-operator-rh
    provider:
      name: Red Hat, Inc
      url: https://github.com/stolostron/multicluster-global-hub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: metallb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: metallb-operator.v4.16.0-202509230119
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "basic",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64512,
                        "neighbors": [
                          {
                            "address": "172.30.0.3",
                            "asn": 64512,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            },
                            "toReceive": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "two-neighs",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64512,
                        "neighbors": [
                          {
                            "address": "172.30.0.3",
                            "asn": 4200000000,
                            "ebgpMultiHop": true,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          },
                          {
                            "address": "172.18.0.6",
                            "asn": 4200000000,
                            "port": 179,
                            "toAdvertise": {
                              "allowed": {
                                "prefixes": [
                                  "192.168.2.0/24"
                                ]
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "two-routers-vrf",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64513,
                        "neighbors": [
                          {
                            "address": "172.32.0.3",
                            "asn": 4200000000,
                            "ebgpMultiHop": true,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ]
                      },
                      {
                        "asn": 64515,
                        "neighbors": [
                          {
                            "address": "172.31.0.3",
                            "asn": 4200000000,
                            "port": 179,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ],
                        "vrf": "red"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRNodeState",
                "metadata": {
                  "name": "frr-k8s-worker"
                },
                "status": {
                  "lastConversionResult": "success",
                  "lastReloadResult": "success",
                  "runningConfig": "Building configuration...\n\nCurrent configuration:\n!\nfrr version 9.0.2_git\nfrr defaults traditional\nhostname frr-k8s-worker\nlog file /etc/frr/frr.log informational\nlog timestamp precision 3\nno ipv6 forwarding\nservice integrated-vtysh-config\n!\nip nht resolve-via-default\n!\nipv6 nht resolve-via-default\n!\nend\n"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BFDProfile",
                "metadata": {
                  "name": "bfd-profile-sample",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "receiveInterval": 380,
                  "transmitInterval": 270
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "peers": [
                    "ebgp-single-hop0"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPoolSelectors": [
                    {
                      "matchLabels": {
                        "test": "ipv6"
                      }
                    },
                    {
                      "matchLabels": {
                        "test": "ipv4"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "communities": [
                    "65535:65282"
                  ],
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "localPref": 50
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample4",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "communities": [
                    "NO_ADVERTISE"
                  ],
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "localPref": 50
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample5",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "nodeSelectors": [
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-control-plane"
                      }
                    },
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-worker"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-beta1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "Community",
                "metadata": {
                  "name": "community1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "communities": [
                    {
                      "name": "NO_ADVERTISE",
                      "value": "65535:65282"
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "name": "ip-addresspool-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "192.168.10.0/24",
                    "192.168.9.1-192.168.9.5",
                    "fc00:f853:0ccd:e799::/124"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "labels": {
                    "test": "ipv4"
                  },
                  "name": "ip-addresspool-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "172.20.0.100/24"
                  ],
                  "autoAssign": false
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "labels": {
                    "test": "ipv6"
                  },
                  "name": "ip-addresspool-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "2002:2:2::1-2002:2:2::100"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "L2Advertisement",
                "metadata": {
                  "name": "l2-adv-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "L2Advertisement",
                "metadata": {
                  "name": "l2-adv-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "nodeSelectors": [
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-control-plane"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "MetalLB",
                "metadata": {
                  "name": "metallb",
                  "namespace": "metallb-system"
                },
                "spec": {}
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "bfdProfile": "bfd-profile-sample",
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "holdTime": "9m0s",
                  "keepaliveTime": "0s",
                  "myASN": 64512,
                  "passwordSecret": {
                    "name": "secretname",
                    "namespace": "metallb-system"
                  },
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3",
                  "peerPort": 180
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:b07c046ba2ae71b93f27d5cdca7cb93f51f3198cb84bb6e65dd6320bcda268c8
          createdAt: "2023-06-06T15:25:00Z"
          description: An operator for deploying MetalLB on a kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.8.0 <4.16.0'
          operatorframework.io/suggested-namespace: metallb-system
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/metallb-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BFDProfile is the Schema for the bfdprofiles API
            kind: BFDProfile
            name: bfdprofiles.metallb.io
            version: v1beta1
          - description: BGPAdvertisement is the Schema for the bgpadvertisements
              API
            kind: BGPAdvertisement
            name: bgpadvertisements.metallb.io
            version: v1beta1
          - description: BGPPeer is the Schema for the bgppeers API
            kind: BGPPeer
            name: bgppeers.metallb.io
            version: v1beta1
          - description: BGPPeer is the Schema for the bgppeers API
            kind: BGPPeer
            name: bgppeers.metallb.io
            version: v1beta2
          - description: Community is the Schema for the communities API
            kind: Community
            name: communities.metallb.io
            version: v1beta1
          - description: FRRConfiguration is the Schema for the frrconfigurations
              API
            kind: FRRConfiguration
            name: frrconfigurations.frrk8s.metallb.io
            version: v1beta1
          - description: FRRNodeState is the Schema for the frrnodestates API
            kind: FRRNodeState
            name: frrnodestates.frrk8s.metallb.io
            version: v1beta1
          - description: IPAddressPool is the Schema for the ipaddresspools API
            kind: IPAddressPool
            name: ipaddresspools.metallb.io
            version: v1beta1
          - description: L2Advertisement is the Schema for the l2advertisements API
            kind: L2Advertisement
            name: l2advertisements.metallb.io
            version: v1beta1
          - description: MetalLB is the Schema for the metallbs API
            displayName: MetalLB
            kind: MetalLB
            name: metallbs.metallb.io
            version: v1beta1
        description: An operator for deploying MetalLB on a kubernetes cluster.
        displayName: MetalLB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - metallb-operator
        links:
        - name: MetalLB Operator
          url: https://github.com/openshift/metallb-operator
        maintainers:
        - email: fpaoline@redhat.com
          name: Federico Paolinelli
        - email: obraunsh@redhat.com
          name: Ori Braunshtein
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/metallb-rhel9@sha256:95abda97d689078bd124ef59c92cc12e661222cac8e8b0e526579470ab54d05e
        - registry.redhat.io/openshift4/frr-rhel9@sha256:48a361644b2fa8fc632f49117172db2e9c12022774e7e0fd409561d884a6c237
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:b07c046ba2ae71b93f27d5cdca7cb93f51f3198cb84bb6e65dd6320bcda268c8
        version: 4.16.0-202509230119
      entries:
      - name: metallb-operator.v4.16.0-202509230119
        version: 4.16.0-202509230119
      - name: metallb-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: metallb-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: metallb-operator.v4.16.0-202507221936
        version: 4.16.0-202507221936
      - name: metallb-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: metallb-operator.v4.16.0-202506090335
        version: 4.16.0-202506090335
      - name: metallb-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: metallb-operator.v4.16.0-202505051339
        version: 4.16.0-202505051339
      - name: metallb-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: metallb-operator.v4.16.0-202503252033
        version: 4.16.0-202503252033
      - name: metallb-operator.v4.16.0-202502260004
        version: 4.16.0-202502260004
      - name: metallb-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: metallb-operator.v4.16.0-202501232204
        version: 4.16.0-202501232204
      - name: metallb-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: metallb-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: metallb-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: metallb-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: metallb-operator.v4.16.0-202411110236
        version: 4.16.0-202411110236
      - name: metallb-operator.v4.16.0-202410292005
        version: 4.16.0-202410292005
      - name: metallb-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: metallb-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: metallb-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: metallb-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: metallb-operator.v4.16.0-202409161407
        version: 4.16.0-202409161407
      - name: metallb-operator.v4.16.0-202409062036
        version: 4.16.0-202409062036
      - name: metallb-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: metallb-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: metallb-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: metallb-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: metallb-operator.v4.16.0-202408051837
        version: 4.16.0-202408051837
      - name: metallb-operator.v4.16.0-202407221107
        version: 4.16.0-202407221107
      - name: metallb-operator.v4.16.0-202407180936
        version: 4.16.0-202407180936
      - name: metallb-operator.v4.16.0-202407091337
        version: 4.16.0-202407091337
      - name: metallb-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: metallb-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: metallb-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: metallb-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: metallb-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ian Boden
      provider-url: ""
    name: eclipse-amlen-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eclipse-amlen-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eclipse.org/v1",
                "kind": "Amlen",
                "metadata": {
                  "name": "amlen-sample"
                },
                "spec": {
                  "_amlen_image_tag": "latest",
                  "_amlen_monitor_image_tag": "latest",
                  "_amlen_name": "amlen-sample",
                  "_amlen_namespace": "osdk-test",
                  "_amlen_persistence": "false",
                  "config": "amlen-config",
                  "device_cert_issuer": {
                    "mode": "automatic",
                    "name": "amlen-selfsigned-issuer"
                  },
                  "ha": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "size": 2
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/amlen/operator:v1.0.1a
          createdAt: "2022-11-24T15:31:00Z"
          description: An operator to run the Eclipse Amlen MQTT Broker
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/eclipse/amlen
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Amlen
            name: amlens.eclipse.org
            version: v1
        description: Operator for the eclipse Amlen MQTT Broker\n\n### Supported Feature\n\nCurrently
          supports:\n* Install - simple install process\n* Upgrade - non-disruptive
          upgrade\n*High Availability - can be enabled/disabled via the CRD\n\n###
          Documentation\nThe main Amlen website has links to product documentation,
          forum, blog and the slack channel\nhttp://www.eclipse.org/amlen\n\n### Getting
          help\nThe quickest way to get help is to use the slack channel but the forum
          is monitored as well\n\nIssues can also be added to the github repository
          https://github.com/eclipse/amlen\n\n### License\nEclipse Amlen Operator
          is licensed under the [Eclipse Public License - v 2.0](https://www.eclipse.org/legal/epl-2.0/)
        displayName: eclipse-amlen-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amlen
        - mqtt
        - broker
        links:
        - name: Eclipse Amlen Operator
          url: https://eclipse.org/amlen
        maintainers:
        - email: ianboden@uk.ibm.com
          name: Ian Boden
        maturity: stable
        provider:
          name: Ian Boden
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/amlen/operator:v1.0.1a
        version: 1.0.0
      entries:
      - name: eclipse-amlen-operator.v1.0.0
        version: 1.0.0
      - name: eclipse-amlen-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: eclipse-amlen-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eclipse.org/v1",
                "kind": "Amlen",
                "metadata": {
                  "name": "amlen-sample"
                },
                "spec": {
                  "config": "amlen-config",
                  "device_cert_issuer": {
                    "mode": "automatic",
                    "name": "amlen-selfsigned-issuer"
                  },
                  "ha": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "ldap": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "size": 2
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/amlen/operator:v1.1.3
          createdAt: "2025-05-22T10:00:00Z"
          description: An operator to run the Eclipse Amlen MQTT Broker
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/eclipse-amlen/amlen
          support: Eclipse
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Amlen
            name: amlens.eclipse.org
            version: v1
        description: |-
          Operator for the eclipse Amlen MQTT Broker

          ### Supported Feature

          Currently supports:
          * Install - simple install process
          * Upgrade - non-disruptive upgrade
          * High Availability - can be enabled/disabled via the CRD

          ### Documentation
          The main Amlen website has links to product documentation, forum, blog and the slack channel
          http://www.eclipse.org/amlen

          ### Getting help
          The quickest way to get help is to use the slack channel but the forum is monitored as well
          Issues can also be added to the github repository https://github.com/eclipse-amlen/amlen

          ### License
          Eclipse Amlen Operator is licensed under [Eclipse Public License - v 2.0](https://www.eclipse.org/legal/epl-2.0/)
        displayName: eclipse-amlen-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amlen
        - mqtt
        - broker
        links:
        - name: Eclipse Amlen Operator
          url: https://eclipse.org/amlen
        maintainers:
        - email: ianboden@uk.ibm.com
          name: Ian Boden
        maturity: stable
        provider:
          name: Eclipse Amlen
          url: https://eclipse.org/amlen
        relatedImages:
        - controller:latest
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        version: 1.1.3
      entries:
      - name: eclipse-amlen-operator.v1.1.3
        version: 1.1.3
      - name: eclipse-amlen-operator.v1.1.0
        version: 1.1.0
      - name: eclipse-amlen-operator.v1.0.2
        version: 1.0.2
      - name: eclipse-amlen-operator.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: alpha
    packageName: eclipse-amlen-operator
    provider:
      name: Ian Boden
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: vault-config-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vault-config-operator.v0.8.34
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AuthEngineMount",
                "metadata": {
                  "name": "authenginemount-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kube-authengine-mount-sample",
                  "type": "kubernetes"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azureauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azureauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azureauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureCredentials": {
                    "passwordKey": "clientsecret",
                    "secret": {
                      "name": "aad-credentials"
                    },
                    "usernameKey": "clientid"
                  },
                  "clientID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "environment": "AzurePublicCloud",
                  "maxRetries": 3,
                  "maxRetryDelay": 60,
                  "path": "azure/test",
                  "resource": "https://management.azure.com/",
                  "retryDelay": 4,
                  "tenantID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azureauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azureauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azureauthenginerole-sample"
                },
                "spec": {
                  "BoundResourceGroups": [
                    "resourcegroup1",
                    "resourcegroup2"
                  ],
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "boundGroupIDs": [
                    "group1",
                    "group2"
                  ],
                  "boundLocations": [
                    "location1",
                    "location2"
                  ],
                  "boundScaleSets": [
                    "scaleset1",
                    "scaleset1"
                  ],
                  "boundServicePrincipalIDs": [
                    "sp1",
                    "sp2"
                  ],
                  "boundSubscriptionIDs": [
                    "subscription1",
                    "subscription2"
                  ],
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "dev-role",
                  "path": "azure",
                  "policies": [
                    "policy1",
                    "policy2"
                  ],
                  "tokenBoundCIDRs": [
                    "10.20.102.13",
                    "10.20.102.15"
                  ],
                  "tokenExplicitMaxTTL": "",
                  "tokenMaxTTL": "",
                  "tokenNoDefaultPolicy": false,
                  "tokenNumUses": 0,
                  "tokenPeriod": 0,
                  "tokenPolicies": [
                    "policy1",
                    "policy2"
                  ],
                  "tokenTTL": "",
                  "tokenType": ""
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureSecretEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azuresecretengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azuresecretengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azuresecretengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureCredentials": {
                    "passwordKey": "clientsecret",
                    "secret": {
                      "name": "aad-credentials"
                    },
                    "usernameKey": "clientid"
                  },
                  "clientID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "environment": "AzurePublicCloud",
                  "passwordPolicy": "",
                  "path": "azure",
                  "rootPasswordTTL": "128d",
                  "subscriptionID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "tenantID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureSecretEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azuresecretenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azuresecretenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azuresecretenginerole-sample"
                },
                "spec": {
                  "TTL": "",
                  "applicationObjectID": "",
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureGroups": "",
                  "azureRoles": "",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "maxTTL": "",
                  "name": "azure-role",
                  "path": "azure",
                  "permanentlyDelete": "",
                  "persistApp": "",
                  "signInAudience": "",
                  "tags": ""
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "CertAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "certauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "certauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "certauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "test",
                  "path": "cert"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "CertAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "certauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "certauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "certauthenginerole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "certificate": "-----BEGIN CERTIFICATE-----\nMIIDJTCCAg2gAwIBAgIUF119kmCNlnMwK6yOOpfRs4JGhAAwDQYJKoZIhvcNAQEL\nBQAwFjEUMBIGA1UEAwwLZXhhbXBsZS5jb20wHhcNMjUwMTIxMTM0MjExWhcNMjYw\nMTIxMTM0MjExWjAWMRQwEgYDVQQDDAtleGFtcGxlLmNvbTCCASIwDQYJKoZIhvcN\nAQEBBQADggEPADCCAQoCggEBALad4Sx97VyFSfOepJoXUGYljg7hL+z20pVL2/M8\n8HXAQl4qf/i45b//Qz8VogMxdg3Y9J7O/M0n9hxbL18i1gm270VjEFyzjJCyUq0R\n72WMM4Sr3OjQPnVtJoQkWnH7wR3zsWK2N4NuIPdE3Ofb48Olyachffv7xCI+MLIa\n/adcyRggM4Z5f2nBCMcXL8D/eY03qCgd/U1aaPJ0WyJTZHLL4euL0ewKlKtB1JnU\n4MrRa5mbPQ3B+OhxcaD/7BNV6hPXAe2zftEv8LcimhbVoj3XmLJ+om58YIaEMadk\nJj61xnXbf+F/iV41onzwvTPNWCSWiLowDl8P3zkabuS583sCAwEAAaNrMGkwHQYD\nVR0OBBYEFH971CwrqkTvUz2qrUncWMS5ULwKMB8GA1UdIwQYMBaAFH971CwrqkTv\nUz2qrUncWMS5ULwKMA8GA1UdEwEB/wQFMAMBAf8wFgYDVR0RBA8wDYILZXhhbXBs\nZS5jb20wDQYJKoZIhvcNAQELBQADggEBAIQzT113TQC9BqitOCOoDFCJqcM+BHmf\nITEKlEhlqubPvA40lD4wkz67RP2SdQbF1glzZW5sDTxMjpHL+8nhdmo65CJpsQGP\nyMPwnOvx7W5JWQJa1Yt+48JNhgIymDMun+xwDsjNnlBqp0Kr7HkOxPimMPWN4R6B\nydG96puDGBYqxB1ujx5seOGfGvxPBBr3h8kNg054rtIIDrUG2iUTfpNIILJrd8Q2\nshop9//3VcIpUJO7Z1brvK/Hp4DrcKBZVo7vd8n+5uCx3vBA0tjqNAXYtFnSRvDZ\n6XV+rCZecDBFkaLuJ5YgYEK15gGe6lQAez6DTVLXMAKhZ52kceefnwM=\n-----END CERTIFICATE-----\n",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "test-role",
                  "path": "cert/test"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineConfig",
                "metadata": {
                  "name": "my-postgresql-database"
                },
                "spec": {
                  "allowedRoles": [
                    "read-only"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "connectionURL": "postgresql://{{username}}:{{password}}@my-postgresql-database.test-vault-config-operator.svc:5432",
                  "path": "test-vault-config-operator/database",
                  "pluginName": "postgresql-database-plugin",
                  "rootCredentials": {
                    "passwordKey": "postgresql-password",
                    "secret": {
                      "name": "postgresql-admin-password"
                    }
                  },
                  "username": "postgres"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineRole",
                "metadata": {
                  "name": "read-only"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "creationStatements": [
                    "CREATE ROLE \"{{name}}\" WITH LOGIN PASSWORD '{{password}}' VALID UNTIL '{{expiration}}'; GRANT SELECT ON ALL TABLES IN SCHEMA public TO \"{{name}}\";"
                  ],
                  "dBName": "my-postgresql-database",
                  "defaultTTL": "1h",
                  "maxTTL": "24h",
                  "path": "test-vault-config-operator/database"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineStaticRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "databasesecretenginestaticrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "databasesecretenginestaticrole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "databasesecretenginestaticrole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "dBName": "db-name",
                  "path": "test-vault-config-operator/database",
                  "rotationPeriod": "24h",
                  "rotationStatements": [
                    "ALTER ROLE \"{{name}}\" WITH PASSWORD '{{password}}';git add"
                  ],
                  "username": "db-user"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GCPAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "gcpauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gcpauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "gcpauthengineconfig-sample"
                },
                "spec": {
                  "GCEalias": "role_id",
                  "GCEmetadata": "default",
                  "GCPCredentials": {
                    "passwordKey": "credentials",
                    "secret": {
                      "name": "gcp-serviceaccount-credentials"
                    },
                    "usernameKey": "serviceaccount"
                  },
                  "IAMalias": "default",
                  "IAMmetadata": "default",
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "customEndpoint": {},
                  "path": "azure",
                  "serviceAccount": "vault-sa"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GCPAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "gcpauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gcpauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "gcpauthenginerole-sample"
                },
                "spec": {
                  "addGroupAliases": false,
                  "allowGCEInference": false,
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "boundInstanceGroups": [],
                  "boundLabels": [],
                  "boundProjects": {},
                  "boundRegions": [],
                  "boundServiceAccounts": {},
                  "boundZones": [],
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "maxJWTExp": "",
                  "name": "gcp-iam-role",
                  "path": "gcp",
                  "policies": [],
                  "tokenBoundCIDRs": [],
                  "tokenExplicitMaxTTL": "",
                  "tokenMaxTTL": "",
                  "tokenNoDefaultPolicy": false,
                  "tokenNumUses": 0,
                  "tokenPeriod": 0,
                  "tokenPolicies": [],
                  "tokenTTL": "",
                  "tokenType": "",
                  "type": "iam"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitHubSecretEngineConfig",
                "metadata": {
                  "name": "githubsecretengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitHubSecretEngineRole",
                "metadata": {
                  "name": "githubsecretenginerole-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Group",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "group-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "group",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "group-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "metadata": {
                    "team": "team-abc"
                  },
                  "policies": [
                    "team-abc-access"
                  ],
                  "type": "external"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupAlias",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "groupalias-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "groupalias",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "groupalias-sample"
                },
                "spec": {
                  "authEngineMountPath": "kubernetes",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "groupName": "group-sample"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "JWTOIDCAuthEngineConfig",
                "metadata": {
                  "name": "azure-oidc-config"
                },
                "spec": {
                  "OIDCClientID": "000000000-0000-0000-0000-000000000",
                  "OIDCCredentials": {
                    "passwordKey": "oidc_client_secret",
                    "secret": {
                      "name": "oidccredentials"
                    },
                    "usernameKey": "oidc_client_id"
                  },
                  "OIDCDiscoveryURL": "https://login.microsoftonline.com/000000000-0000-0000-0000-000000000/v2.0",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "oidc/azuread-oidc/"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "JWTOIDCAuthEngineRole",
                "metadata": {
                  "name": "azure-oidc-dev-role"
                },
                "spec": {
                  "OIDCScopes": [
                    "https://graph.microsoft.com/.default"
                  ],
                  "allowedRedirectURIs": [
                    "http://localhost:8250/oidc/callback",
                    "http://localhost:8200/ui/vault/auth/oidc/azuread-oidc/oidc/callback"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "groupsClaim": "groups",
                  "name": "dev-role",
                  "path": "oidc-aad/azuread-oidc",
                  "roleType": "oidc",
                  "tokenPolicies": [
                    "dev",
                    "prod"
                  ],
                  "userClaim": "email"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesAuthEngineConfig",
                "metadata": {
                  "name": "kubernetesauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kube-authengine-mount-sample"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesAuthEngineRole",
                "metadata": {
                  "name": "database-engine-admin"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kubernetes",
                  "policies": [
                    "database-engine-admin"
                  ],
                  "targetNamespaces": {
                    "targetNamespaceSelector": {
                      "matchLabels": {
                        "database-engine-admin": "true"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesSecretEngineConfig",
                "metadata": {
                  "name": "kubernetessecretengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesSecretEngineRole",
                "metadata": {
                  "name": "kubernetessecretenginerole-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "LDAPAuthEngineConfig",
                "metadata": {
                  "name": "ldapauthengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "LDAPAuthEngineGroup",
                "metadata": {
                  "name": "ldapauthenginegroup-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin",
                    "serviceAccount": {
                      "name": "default"
                    }
                  },
                  "name": "test-group",
                  "path": "ldap/test"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PKISecretEngineConfig",
                "metadata": {
                  "name": "pkisecretengineconfig-sample"
                },
                "spec": {
                  "TTL": "8760h",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "commonName": "pki-vault-demo.internal.io",
                  "crlDistributionPoints": [
                    "https://vault-internal.vault.svc:8200/v1/test-vault-config-operator/pki/crl\""
                  ],
                  "issuingCertificates": [
                    "https://vault-internal.vault.svc:8200/v1/test-vault-config-operator/pki/ca"
                  ],
                  "path": "test-vault-config-operator/pki",
                  "privateKeyType": "internal",
                  "type": "root"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PKISecretEngineRole",
                "metadata": {
                  "name": "pkisecretenginerole-sample"
                },
                "spec": {
                  "allowGlobDomains": true,
                  "allowSubdomains": true,
                  "allowedDomains": [
                    "internal.io",
                    "pki-vault-demo.svc",
                    "example.com"
                  ],
                  "allowedOtherSans": "*",
                  "allowedURISans": [
                    "*-pki-vault-demo.apps.example.com"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "maxTTL": "8760h",
                  "path": "test-vault-config-operator/pki"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PasswordPolicy",
                "metadata": {
                  "name": "simple-password-policy"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "passwordPolicy": "length = 20\nrule \"charset\" {\n  charset = \"abcdefghijklmnopqrstuvwxyz\"\n}"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Policy",
                "metadata": {
                  "name": "database-engine-admin"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "policy": "# mount database secrets engines\npath \"/sys/mounts/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n  allowed_parameters = {\n    \"type\" = [\"database\"]\n    \"*\"   = []\n  }\n}\n\n# tune database secrets engines\npath \"/sys/mounts/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/tune\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n\n# Configure database secrets engines\npath \"/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/config/+\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n\n# Configure database roles\npath \"/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/roles/+\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineConfig",
                "metadata": {
                  "name": "quaysecretengineconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineRole",
                "metadata": {
                  "name": "quaysecretenginerole-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineStaticRole",
                "metadata": {
                  "name": "quaysecretenginestaticrole-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RabbitMQSecretEngineConfig",
                "metadata": {
                  "name": "rabbitmqsecretengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "rabbitMQ-engine-admin"
                  },
                  "connectionURI": "https://test.com",
                  "leaseMaxTTL": 86400,
                  "leaseTTL": 86400,
                  "path": "vault-config-operator/rabbitmq"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RabbitMQSecretEngineRole",
                "metadata": {
                  "name": "rabbitmqsecretenginerole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "rabbitmq-engine-admin"
                  },
                  "path": "vault-config-operator/rabbitmq",
                  "tags": "administrator",
                  "vhostTopics": [
                    {
                      "topics": [
                        {
                          "permissions": {
                            "configure": ".*",
                            "read": ".*",
                            "write": ".*"
                          },
                          "topicName": "my-topic"
                        },
                        {
                          "permissions": {
                            "read": ".*"
                          },
                          "topicName": "my-read-topic"
                        }
                      ],
                      "vhostName": "/"
                    }
                  ],
                  "vhosts": [
                    {
                      "permissions": {
                        "configure": ".*",
                        "read": ".*",
                        "write": ".*"
                      },
                      "vhostName": "/"
                    },
                    {
                      "permissions": {
                        "read": "my-queue",
                        "write": "my-exchange"
                      },
                      "vhostName": "my-vhost"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RandomSecret",
                "metadata": {
                  "name": "randomsecret-password"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "secret-writer"
                  },
                  "path": "test-vault-config-operator/kv",
                  "secretFormat": {
                    "passwordPolicyName": "simple-password-policy"
                  },
                  "secretKey": "password"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "SecretEngineMount",
                "metadata": {
                  "name": "database"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "path": "test-vault-config-operator",
                  "type": "database"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "VaultSecret",
                "metadata": {
                  "name": "vaultsecret-sample"
                },
                "spec": {
                  "output": {
                    "annotations": {
                      "refresh": "every-minute"
                    },
                    "labels": {
                      "app": "test-vault-config-operator"
                    },
                    "name": "randomsecret",
                    "stringData": {
                      "anotherpassword": "{{ .anotherrandomsecret.password }}",
                      "password": "{{ .randomsecret.password }}"
                    },
                    "type": "Opaque"
                  },
                  "refreshPeriod": "3m0s",
                  "vaultSecretDefinitions": [
                    {
                      "authentication": {
                        "path": "kubernetes",
                        "role": "secret-reader",
                        "serviceAccount": {
                          "name": "default"
                        }
                      },
                      "name": "randomsecret",
                      "path": "test-vault-config-operator/kv/randomsecret-password"
                    },
                    {
                      "authentication": {
                        "path": "kubernetes",
                        "role": "secret-reader",
                        "serviceAccount": {
                          "name": "default"
                        }
                      },
                      "name": "anotherrandomsecret",
                      "path": "test-vault-config-operator/kv/another-password"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/vault-config-operator@sha256:490852417fac37f3003ab7c1ee797f9035fd2905c354cc46eca5e33d2ad780ce
          createdAt: "2025-09-18T18:56:36Z"
          description: This operator provides primitives to declaratively configure
            Hashicorp Vault.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: vault-config-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/vault-config-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthEngineMount is the Schema for the authenginemounts API
            displayName: Auth Engine Mount
            kind: AuthEngineMount
            name: authenginemounts.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureAuthEngineConfig is the Schema for the azureauthengineconfigs
              API
            displayName: Azure Auth Engine Config
            kind: AzureAuthEngineConfig
            name: azureauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureAuthEngineRole is the Schema for the azureauthengineroles
              API
            displayName: Azure Auth Engine Role
            kind: AzureAuthEngineRole
            name: azureauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - kind: AzureSecretEngineConfig
            name: azuresecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureSecretEngineRole is the Schema for the azuresecretengineroles
              API
            displayName: Azure Secret Engine Role
            kind: AzureSecretEngineRole
            name: azuresecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: CertAuthEngineConfig is the Schema for the certauthengineconfigs
              API
            displayName: Cert Auth Engine Config
            kind: CertAuthEngineConfig
            name: certauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: CertAuthEngineRole is the Schema for the certauthengineroles
              API
            displayName: Cert Auth Engine Role
            kind: CertAuthEngineRole
            name: certauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineConfig is the Schema for the databasesecretengineconfigs
              API
            displayName: Database Secret Engine Config
            kind: DatabaseSecretEngineConfig
            name: databasesecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineRole is the Schema for the databasesecretengineroles
              API
            displayName: Database Secret Engine Role
            kind: DatabaseSecretEngineRole
            name: databasesecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineStaticRole is the Schema for the databasesecretenginestaticroles
              API
            displayName: Database Secret Engine Static Role
            kind: DatabaseSecretEngineStaticRole
            name: databasesecretenginestaticroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GCPAuthEngineConfig is the Schema for the gcpauthengineconfigs
              API
            displayName: GCPAuth Engine Config
            kind: GCPAuthEngineConfig
            name: gcpauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: GCPAuthEngineRole is the Schema for the gcpauthengineroles
              API
            displayName: GCPAuth Engine Role
            kind: GCPAuthEngineRole
            name: gcpauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GitHubSecretEngineConfig is the Schema for the githubsecretengineconfigs
              API
            displayName: Git Hub Secret Engine Config
            kind: GitHubSecretEngineConfig
            name: githubsecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: GitHubSecretEngineRole is the Schema for the githubsecretengineroles
              API
            displayName: Git Hub Secret Engine Role
            kind: GitHubSecretEngineRole
            name: githubsecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GroupAlias is the Schema for the groupaliases API
            displayName: Group Alias
            kind: GroupAlias
            name: groupaliases.redhatcop.redhat.io
            version: v1alpha1
          - description: Group is the Schema for the groups API
            displayName: Group
            kind: Group
            name: groups.redhatcop.redhat.io
            version: v1alpha1
          - description: JWTOIDCAuthEngineConfig is the Schema for the jwtoidcauthengineconfigs
              API
            displayName: JWTOIDCAuth Engine Config
            kind: JWTOIDCAuthEngineConfig
            name: jwtoidcauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: JWTOIDCAuthEngineRole is the Schema for the jwtoidcauthengineroles
              API
            displayName: JWTOIDCAuth Engine Role
            kind: JWTOIDCAuthEngineRole
            name: jwtoidcauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesAuthEngineConfig is the Schema for the kubernetesauthengineconfigs
              API
            displayName: Kubernetes Auth Engine Config
            kind: KubernetesAuthEngineConfig
            name: kubernetesauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesAuthEngineRole can be used to define a KubernetesAuthEngineRole
              for the kube-auth authentication method
            displayName: Kubernetes Auth Engine Role
            kind: KubernetesAuthEngineRole
            name: kubernetesauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesSecretEngineConfig is the Schema for the kubernetessecretengineconfigs
              API
            displayName: Kubernetes Secret Engine Config
            kind: KubernetesSecretEngineConfig
            name: kubernetessecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesSecretEngineRole is the Schema for the kubernetessecretengineroles
              API
            displayName: Kubernetes Secret Engine Role
            kind: KubernetesSecretEngineRole
            name: kubernetessecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: LDAPAuthEngineConfig is the Schema for the ldapauthengineconfigs
              API
            displayName: LDAPAuth Engine Config
            kind: LDAPAuthEngineConfig
            name: ldapauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: LDAPAuthEngineGroup is the Schema for the ldapauthenginegroups
              API
            displayName: LDAPAuth Engine Group
            kind: LDAPAuthEngineGroup
            name: ldapauthenginegroups.redhatcop.redhat.io
            version: v1alpha1
          - description: PasswordPolicy is the Schema for the passwordpolicies API
            displayName: Password Policy
            kind: PasswordPolicy
            name: passwordpolicies.redhatcop.redhat.io
            version: v1alpha1
          - description: PKISecretEngineConfig is the Schema for the pkisecretengineconfigs
              API
            displayName: PKISecret Engine Config
            kind: PKISecretEngineConfig
            name: pkisecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: PKISecretEngineRole is the Schema for the pkisecretengineroles
              API
            displayName: PKISecret Engine Role
            kind: PKISecretEngineRole
            name: pkisecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: Policy is the Schema for the policies API
            displayName: Policy
            kind: Policy
            name: policies.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineConfig is the Schema for the quaysecretengineconfigs
              API
            displayName: Quay Secret Engine Config
            kind: QuaySecretEngineConfig
            name: quaysecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineRole is the Schema for the quaysecretengineroles
              API
            displayName: Quay Secret Engine Role
            kind: QuaySecretEngineRole
            name: quaysecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineStaticRole is the Schema for the quaysecretenginestaticroles
              API
            displayName: Quay Secret Engine Static Role
            kind: QuaySecretEngineStaticRole
            name: quaysecretenginestaticroles.redhatcop.redhat.io
            version: v1alpha1
          - description: RabbitMQSecretEngineConfig is the Schema for the rabbitmqsecretengineconfigs
              API
            displayName: Rabbit MQSecret Engine Config
            kind: RabbitMQSecretEngineConfig
            name: rabbitmqsecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: RabbitMQSecretEngineRole is the Schema for the rabbitmqsecretengineroles
              API
            displayName: Rabbit MQSecret Engine Role
            kind: RabbitMQSecretEngineRole
            name: rabbitmqsecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: RandomSecret is the Schema for the randomsecrets API
            displayName: Random Secret
            kind: RandomSecret
            name: randomsecrets.redhatcop.redhat.io
            version: v1alpha1
          - description: SecretEngineMount is the Schema for the secretenginemounts
              API
            displayName: Secret Engine Mount
            kind: SecretEngineMount
            name: secretenginemounts.redhatcop.redhat.io
            version: v1alpha1
          - description: VaultSecret is the Schema for the vaultsecrets API
            displayName: Vault Secret
            kind: VaultSecret
            name: vaultsecrets.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator helps set up Vault Configurations. The main intent is to do so such that subsequently pods can consume the secrets made available.
          There are two main principles through all of the capabilities of this operator:

          1. high-fidelity API. The CRD exposed by this operator reflect field by field the Vault APIs. This is because we don't want to make any assumption on the kinds of configuration workflow that user will set up. That being said the Vault API is very extensive and we are starting with enough API coverage to support, we think, some simple and very common configuration workflows.
          2. attention to security (after all we are integrating with a security tool). To prevent credential leaks we give no permissions to the operator itself against Vault. All APIs exposed by this operator contains enough information to authenticate to Vault using a local service account (local to the namespace where the API exist). In other word for a namespace user to be abel to successfully configure Vault, a service account in that namespace must have been previously given the needed Vault permissions.

          Currently this operator supports the following CRDs:

          1. [Policy](https://github.com/redhat-cop/vault-config-operator#policy) Configures Vault [Policies](https://www.vaultproject.io/docs/concepts/policies)
          2. [KubernetesAuthEngineRole](https://github.com/redhat-cop/vault-config-operator#KubernetesAuthEngineRole) Configures a Vault [Kubernetes Authentication](https://www.vaultproject.io/docs/auth/kubernetes) Role
          3. [SecretEngineMount](https://github.com/redhat-cop/vault-config-operator#SecretEngineMount) Configures a Mount point for a [SecretEngine](https://www.vaultproject.io/docs/secrets)
          4. [DatabaseSecretEngineConfig](https://github.com/redhat-cop/vault-config-operator#DatabaseSecretEngineConfig) Configures a [Database Secret Engine](https://www.vaultproject.io/docs/secrets/databases) Connection
          5. [DatabaseSecretEngineRole](https://github.com/redhat-cop/vault-config-operator#DatabaseSecretEngineRole) Configures a [Database Secret Engine](https://www.vaultproject.io/docs/secrets/databases) Role
          6. [RandomSecret](https://github.com/redhat-cop/vault-config-operator#RandomSecret) Creates a random secret in a vault [kv Secret Engine](https://www.vaultproject.io/docs/secrets/kv) with one password field generated using a [PasswordPolicy](https://www.vaultproject.io/docs/concepts/password-policies)vault-config-operator
        displayName: Vault Config Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - secrets
        - gitops
        - policy
        - credentials
        links:
        - name: Vault Config Operator
          url: https://github.com/redhat-cop/vault-config-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/vault-config-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/vault-config-operator@sha256:490852417fac37f3003ab7c1ee797f9035fd2905c354cc46eca5e33d2ad780ce
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 0.8.34
      entries:
      - name: vault-config-operator.v0.8.34
        version: 0.8.34
      - name: vault-config-operator.v0.8.33
        version: 0.8.33
      - name: vault-config-operator.v0.8.32
        version: 0.8.32
      - name: vault-config-operator.v0.8.31
        version: 0.8.31
      - name: vault-config-operator.v0.8.30
        version: 0.8.30
      - name: vault-config-operator.v0.8.29
        version: 0.8.29
      - name: vault-config-operator.v0.8.28
        version: 0.8.28
      - name: vault-config-operator.v0.8.27
        version: 0.8.27
      - name: vault-config-operator.v0.8.25
        version: 0.8.25
      - name: vault-config-operator.v0.8.24
        version: 0.8.24
      - name: vault-config-operator.v0.8.23
        version: 0.8.23
      - name: vault-config-operator.v0.8.22
        version: 0.8.22
      - name: vault-config-operator.v0.8.21
        version: 0.8.21
      - name: vault-config-operator.v0.8.20
        version: 0.8.20
      - name: vault-config-operator.v0.8.19
        version: 0.8.19
      - name: vault-config-operator.v0.8.18
        version: 0.8.18
      - name: vault-config-operator.v0.8.17
        version: 0.8.17
      - name: vault-config-operator.v0.8.16
        version: 0.8.16
      - name: vault-config-operator.v0.8.15
        version: 0.8.15
      - name: vault-config-operator.v0.8.14
        version: 0.8.14
      - name: vault-config-operator.v0.8.13
        version: 0.8.13
      - name: vault-config-operator.v0.8.12
        version: 0.8.12
      - name: vault-config-operator.v0.8.11
        version: 0.8.11
      - name: vault-config-operator.v0.8.10
        version: 0.8.10
      - name: vault-config-operator.v0.8.9
        version: 0.8.9
      - name: vault-config-operator.v0.8.7
        version: 0.8.7
      - name: vault-config-operator.v0.8.5
        version: 0.8.5
      - name: vault-config-operator.v0.8.4
        version: 0.8.4
      - name: vault-config-operator.v0.8.3
        version: 0.8.3
      - name: vault-config-operator.v0.8.2
        version: 0.8.2
      - name: vault-config-operator.v0.8.0
        version: 0.8.0
      - name: vault-config-operator.v0.7.1
        version: 0.7.1
      - name: vault-config-operator.v0.6.5
        version: 0.6.5
      - name: vault-config-operator.v0.6.4
        version: 0.6.4
      - name: vault-config-operator.v0.6.3
        version: 0.6.3
      - name: vault-config-operator.v0.6.2
        version: 0.6.2
      - name: vault-config-operator.v0.6.1
        version: 0.6.1
      - name: vault-config-operator.v0.6.0
        version: 0.6.0
      - name: vault-config-operator.v0.5.1
        version: 0.5.1
      - name: vault-config-operator.v0.5.0
        version: 0.5.0
      - name: vault-config-operator.v0.4.1
        version: 0.4.1
      - name: vault-config-operator.v0.4.0
        version: 0.4.0
      - name: vault-config-operator.v0.3.1
        version: 0.3.1
      - name: vault-config-operator.v0.3.0
        version: 0.3.0
      - name: vault-config-operator.v0.2.0
        version: 0.2.0
      - name: vault-config-operator.v0.1.2
        version: 0.1.2
      - name: vault-config-operator.v0.1.1
        version: 0.1.1
      - name: vault-config-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: vault-config-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-rest-api-browser-v1
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-rest-api-browser-v1.v0.1.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rest.loadcore.keysight.com/v1alpha1",
                "kind": "Apibrowser",
                "metadata": {
                  "name": "apibrowser-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-06-29T14:08:05Z"
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Apibrowser
            name: apibrowsers.rest.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Rest API Browser Operator v1
        displayName: Keysight Rest API Browser Operator v1
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Rest Api Browser V1
          url: https://keysight-rest-api-browser-v1.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5d5061dbc2ec5082cf14b6c600fb5400b83cf91d7ccebfa80680a238d275db
        - prod.harbor.keysight.digital/loadcore/keysight-rest-api-browser-v1@sha256:3ada6a0fa630acb0fa962a11d2c272ff679dded1106c4693d2c8f5ec9ab414a4
        version: 0.1.11
      entries:
      - name: keysight-rest-api-browser-v1.v0.1.11
        version: 0.1.11
      name: alpha
    defaultChannel: alpha
    packageName: keysight-rest-api-browser-v1
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Qualcomm Technologies, Inc.
      provider-url: https://github.com/quic/aic-operator
    name: aic-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aic-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aic.quicinc.com/v1",
                "kind": "AIC",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "aic-operator",
                    "app.kubernetes.io/instance": "aic-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "aic",
                    "app.kubernetes.io/part-of": "aic-operator"
                  },
                  "name": "aic-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "false"
          containerImage: ghcr.io/quic/cloud_ai_openshift_operator@sha256:9cf581e9b5c9cd08819fab8f8b6313836a0a20098ad4c99b80d93dbcf5ab32bf
          description: Automate the management of OpenShift Operator for Qualcomm
            AI 100 accelerators.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: Qualcomm Technologies, Inc.
          repository: https://github.com/quic/aic-operator
          support: Qualcomm Technologies, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AIC is the Schema for the aics API
            displayName: AIC
            kind: AIC
            name: aics.aic.quicinc.com
            version: v1
        description: The Openshift Operator enables the Qualcomm® Cloud AI series
          of hardware on OpenShift clusters by automating the configuration and installation
          of their Linux device drivers and its Kubernetes Device Plugin.
        displayName: AIC Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AIC
        - AIC-Operator
        links:
        - name: AIC Operator
          url: https://github.com/quic/aic-operator
        maintainers:
        - email: quic_vmellemp@quicinc.com
          name: quic_vmellemp
        - email: quic_carls@quicinc.com
          name: quic_carls
        maturity: alpha
        provider:
          name: Qualcomm Technologies, Inc.
          url: https://github.com/quic/aic-operator
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:7a943db9bcb22f817d75341621472f4bb30abc5625c53b9a781d97d431bc4d09
        - ghcr.io/quic/cloud_ai_openshift_operator@sha256:9cf581e9b5c9cd08819fab8f8b6313836a0a20098ad4c99b80d93dbcf5ab32bf
        version: 0.3.1
      entries:
      - name: aic-operator.v0.3.1
        version: 0.3.1
      name: alpha
    defaultChannel: alpha
    packageName: aic-operator
    provider:
      name: Qualcomm Technologies, Inc.
      url: https://github.com/quic/aic-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kong Inc.
      provider-url: https://konghq.com
    name: kong-gateway-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kong-gateway-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "ControlPlane",
                "metadata": {
                  "name": "controlplane-example"
                },
                "spec": {
                  "dataplane": "dataplane-example",
                  "gatewayClass": "kong"
                }
              },
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "DataPlane",
                "metadata": {
                  "name": "dataplane-example"
                },
                "spec": {
                  "env": [
                    {
                      "name": "KONG_LOG_LEVEL",
                      "value": "debug"
                    }
                  ]
                }
              },
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "GatewayConfiguration",
                "metadata": {
                  "name": "kong",
                  "namespace": "default"
                },
                "spec": {
                  "controlPlaneDeploymentOptions": {
                    "env": [
                      {
                        "name": "CONTROLLER_LOG_LEVEL",
                        "value": "debug"
                      }
                    ]
                  },
                  "dataPlaneDeploymentOptions": {
                    "containerImage": "kong",
                    "version": "2.8"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking,Integration & Delivery
          operatorframework.io/suggested-namespace: kong-system
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ControlPlane is the Schema for the controlplanes API
            displayName: Control Plane
            kind: ControlPlane
            name: controlplanes.gateway-operator.konghq.com
            version: v1alpha1
          - description: DataPlane is the Schema for the dataplanes API
            displayName: Data Plane
            kind: DataPlane
            name: dataplanes.gateway-operator.konghq.com
            version: v1alpha1
          - description: GatewayConfiguration is the Schema for the gatewayconfigurations
              API
            displayName: Gateway Configuration
            kind: GatewayConfiguration
            name: gatewayconfigurations.gateway-operator.konghq.com
            version: v1alpha1
        description: Kong Operator for Kubernetes. Technical preview.
        displayName: Kong
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - gateway
        - ingress
        - kong
        - kubernetes
        - go
        links:
        - name: Github repository for the operator
          url: https://github.com/kong/gateway-operator
        maintainers:
        - email: support@konghq.com
          name: Kong Inc.
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Kong Inc.
          url: https://konghq.com
        relatedImages:
        - ghcr.io/kong/gateway-operator@sha256:76b227715a3e1915a28f8aaa1e2ff8495349d29d0d61e7b85a72d74b03fbcc7f
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.connect.redhat.com/kong/kong@sha256:54e7c7e57bf93810a7cc36327023b8699f754c4c40c5c289548d54170caa5db2
        - registry.connect.redhat.com/kong/kong-ingress-controller@sha256:f691db070e7c604da9e163800903d0942ec6fda879eb9dcd5d8e90618ae787da
        version: 0.0.2
      entries:
      - name: kong-gateway-operator.v0.0.2
        version: 0.0.2
      - name: kong-gateway-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: kong-gateway-operator
    provider:
      name: Kong Inc.
      url: https://konghq.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: https://www.turbonomic.com/
    name: t8c-tsc-client-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-tsc-client-certified.v1.4.2-candidate-20250806145822000
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "TurbonomicClient",
                "metadata": {
                  "name": "turbonomicclient-release"
                },
                "spec": {
                  "global": {
                    "version": "8.16.1"
                  },
                  "probes": {
                    "actionScript": {
                      "enabled": false
                    },
                    "actionStreamKafka": {
                      "enabled": false
                    },
                    "appDynamics": {
                      "enabled": false
                    },
                    "compellent": {
                      "enabled": false
                    },
                    "datadog": {
                      "enabled": false
                    },
                    "dynatrace": {
                      "enabled": false
                    },
                    "flexera": {
                      "enabled": false
                    },
                    "hds": {
                      "enabled": false
                    },
                    "hpe3par": {
                      "enabled": false
                    },
                    "hyperFlex": {
                      "enabled": false
                    },
                    "hyperV": {
                      "enabled": false
                    },
                    "ibmStorageFlashSystem": {
                      "enabled": false
                    },
                    "instana": {
                      "enabled": false
                    },
                    "jvm": {
                      "enabled": false
                    },
                    "mssql": {
                      "enabled": false
                    },
                    "mysql": {
                      "enabled": false
                    },
                    "netApp": {
                      "enabled": false
                    },
                    "newRelic": {
                      "enabled": false
                    },
                    "nutanix": {
                      "enabled": false
                    },
                    "oneView": {
                      "enabled": false
                    },
                    "oracle": {
                      "enabled": false
                    },
                    "powerVM": {
                      "enabled": false
                    },
                    "pure": {
                      "enabled": false
                    },
                    "scaleIO": {
                      "enabled": false
                    },
                    "serviceNow": {
                      "enabled": false
                    },
                    "snmp": {
                      "enabled": false
                    },
                    "terraform": {
                      "enabled": false
                    },
                    "tomcat": {
                      "enabled": false
                    },
                    "ucs": {
                      "enabled": false
                    },
                    "vCenter": {
                      "enabled": false
                    },
                    "vmax": {
                      "enabled": false
                    },
                    "vmm": {
                      "enabled": false
                    },
                    "vplex": {
                      "enabled": false
                    },
                    "webLogic": {
                      "enabled": false
                    },
                    "webSphere": {
                      "enabled": false
                    },
                    "wmi": {
                      "enabled": false
                    },
                    "xtremIO": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "VersionManager",
                "metadata": {
                  "name": "versionmanager-release"
                },
                "spec": {
                  "url": "http://remote-nginx-tunnel:9080/cluster-manager/clusterConfiguration"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-client-operator@sha256:8db7396c956e7a04aee92e1e90491cd3eb533cebe5f5616ca43ff16d9f3ed6e2
          createdAt: "2025-08-07T07:15:28Z"
          description: Turbonomic Secure Connect for Hybrid environment management
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.4.2-candidate-20250806145822000'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TurbonomicClient is the Schema for the turbonomicclients
              API
            displayName: Turbonomic Client
            kind: TurbonomicClient
            name: turbonomicclients.clients.turbonomic.ibm.com
            version: v1alpha1
          - description: VersionManager is the Schema for the versionmanagers API
            displayName: Version Manager
            kind: VersionManager
            name: versionmanagers.clients.turbonomic.ibm.com
            version: v1alpha1
        description: "### Turbonomic Secure Connect for Hybrid environment management\nTurbonomic
          Workload Automation for Multicloud simultaneously optimizes performance,
          compliance, and cost in real-time. Workloads are precisely resourced, automatically,
          to perform while satisfying business constraints:\n* Continuous placement
          of workload across multiple clouds both on-prem and public clouds providers.\n*
          Continuous scaling for applications and the underlying infrastructure.\n*
          Capability to connect both public cloud and on-prem targets on single saas
          instance\nIt assures application performance by giving workloads the resources
          they need when they need them.\n\n### How do I start?\nTurbonomic documentation
          can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm.\nTo
          setup Turbonomic Secure Connect client, follow the instructions below:\n1.
          Click Install on the landing page\n2. Upon moving to Install Operator page,
          do not modify the default selection for any of the configuration\n3. Select
          the Installed namespace from the drop down or add a new namespace and click
          on create project to create the entered namespace.\n4. Click Install on
          the Install Operator page.\nAt this point operator is installed.\n\nOn the
          Details tab, under Provided APIs: \n1. Click Create instance for the Turbonomic
          Client API. Enable/Disable the probes based on the requirement, and click
          Create.\n2. Click Create instance for the Version Manager API. Do not modify
          the JSON/YAML file, click on Create.\n\n### How does it work?\nTurbonomic
          uses a public APIs already exposed by application and infrastructure instrumentation
          to discover and monitor your environment.\nTurbonomic determines the right
          actions that drive continuous health, including continuous placement and
          continuous scaling for applications and the underlying cluster.\nTurbonomic
          leverages the built-on orchestration provided by the application and infrastructure
          deployment tools and automates the execution of these actions to continiously
          meet the respective service level objective of each application service."
        displayName: Turbonomic Secure Connect Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - turbonomic
        - turbo secure connect
        links:
        - name: T8c Tsc Client Certified
          url: https://t8c-tsc-certified.domain
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        relatedImages:
        - icr.io/cpopen/t8c-client-operator@sha256:8db7396c956e7a04aee92e1e90491cd3eb533cebe5f5616ca43ff16d9f3ed6e2
        version: 1.4.2-candidate-20250806145822000
      entries:
      - name: t8c-tsc-client-certified.v1.4.2-candidate-20250806145822000
        version: 1.4.2-candidate-20250806145822000
      name: candidate
    - currentCSV: t8c-tsc-client-certified.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "TurbonomicClient",
                "metadata": {
                  "name": "turbonomicclient-release"
                },
                "spec": {
                  "global": {
                    "version": "8.16.1"
                  },
                  "probes": {
                    "actionScript": {
                      "enabled": false
                    },
                    "actionStreamKafka": {
                      "enabled": false
                    },
                    "appDynamics": {
                      "enabled": false
                    },
                    "compellent": {
                      "enabled": false
                    },
                    "datadog": {
                      "enabled": false
                    },
                    "dynatrace": {
                      "enabled": false
                    },
                    "flexera": {
                      "enabled": false
                    },
                    "hds": {
                      "enabled": false
                    },
                    "hpe3par": {
                      "enabled": false
                    },
                    "hyperFlex": {
                      "enabled": false
                    },
                    "hyperV": {
                      "enabled": false
                    },
                    "ibmStorageFlashSystem": {
                      "enabled": false
                    },
                    "instana": {
                      "enabled": false
                    },
                    "jvm": {
                      "enabled": false
                    },
                    "mssql": {
                      "enabled": false
                    },
                    "mysql": {
                      "enabled": false
                    },
                    "netApp": {
                      "enabled": false
                    },
                    "newRelic": {
                      "enabled": false
                    },
                    "nutanix": {
                      "enabled": false
                    },
                    "oneView": {
                      "enabled": false
                    },
                    "oracle": {
                      "enabled": false
                    },
                    "powerVM": {
                      "enabled": false
                    },
                    "pure": {
                      "enabled": false
                    },
                    "scaleIO": {
                      "enabled": false
                    },
                    "serviceNow": {
                      "enabled": false
                    },
                    "snmp": {
                      "enabled": false
                    },
                    "terraform": {
                      "enabled": false
                    },
                    "tomcat": {
                      "enabled": false
                    },
                    "ucs": {
                      "enabled": false
                    },
                    "vCenter": {
                      "enabled": false
                    },
                    "vmax": {
                      "enabled": false
                    },
                    "vmm": {
                      "enabled": false
                    },
                    "vplex": {
                      "enabled": false
                    },
                    "webLogic": {
                      "enabled": false
                    },
                    "webSphere": {
                      "enabled": false
                    },
                    "wmi": {
                      "enabled": false
                    },
                    "xtremIO": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "VersionManager",
                "metadata": {
                  "name": "versionmanager-release"
                },
                "spec": {
                  "url": "http://remote-nginx-tunnel:9080/cluster-manager/clusterConfiguration"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-client-operator@sha256:626e024327388410f7a77bd9b8e5a5df29993fcafdb0176cd0bea08b05713c61
          createdAt: "2025-08-12T23:15:08Z"
          description: Turbonomic Secure Connect for Hybrid environment management
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.4.2'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TurbonomicClient is the Schema for the turbonomicclients
              API
            displayName: Turbonomic Client
            kind: TurbonomicClient
            name: turbonomicclients.clients.turbonomic.ibm.com
            version: v1alpha1
          - description: VersionManager is the Schema for the versionmanagers API
            displayName: Version Manager
            kind: VersionManager
            name: versionmanagers.clients.turbonomic.ibm.com
            version: v1alpha1
        description: "### Turbonomic Secure Connect for Hybrid environment management\nTurbonomic
          Workload Automation for Multicloud simultaneously optimizes performance,
          compliance, and cost in real-time. Workloads are precisely resourced, automatically,
          to perform while satisfying business constraints:\n* Continuous placement
          of workload across multiple clouds both on-prem and public clouds providers.\n*
          Continuous scaling for applications and the underlying infrastructure.\n*
          Capability to connect both public cloud and on-prem targets on single saas
          instance\nIt assures application performance by giving workloads the resources
          they need when they need them.\n\n### How do I start?\nTurbonomic documentation
          can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm.\nTo
          setup Turbonomic Secure Connect client, follow the instructions below:\n1.
          Click Install on the landing page\n2. Upon moving to Install Operator page,
          do not modify the default selection for any of the configuration\n3. Select
          the Installed namespace from the drop down or add a new namespace and click
          on create project to create the entered namespace.\n4. Click Install on
          the Install Operator page.\nAt this point operator is installed.\n\nOn the
          Details tab, under Provided APIs: \n1. Click Create instance for the Turbonomic
          Client API. Enable/Disable the probes based on the requirement, and click
          Create.\n2. Click Create instance for the Version Manager API. Do not modify
          the JSON/YAML file, click on Create.\n\n### How does it work?\nTurbonomic
          uses a public APIs already exposed by application and infrastructure instrumentation
          to discover and monitor your environment.\nTurbonomic determines the right
          actions that drive continuous health, including continuous placement and
          continuous scaling for applications and the underlying cluster.\nTurbonomic
          leverages the built-on orchestration provided by the application and infrastructure
          deployment tools and automates the execution of these actions to continiously
          meet the respective service level objective of each application service."
        displayName: Turbonomic Secure Connect Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - turbonomic
        - turbo secure connect
        links:
        - name: T8c Tsc Client Certified
          url: https://t8c-tsc-certified.domain
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        relatedImages:
        - icr.io/cpopen/t8c-client-operator@sha256:626e024327388410f7a77bd9b8e5a5df29993fcafdb0176cd0bea08b05713c61
        version: 1.4.2
      entries:
      - name: t8c-tsc-client-certified.v1.4.2
        version: 1.4.2
      name: stable
    defaultChannel: stable
    packageName: t8c-tsc-client-certified
    provider:
      name: Turbonomic, Inc.
      url: https://www.turbonomic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-recyclebin-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-recyclebin-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "recyclebin.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/recyclebin-controller:1.1.1
          createdAt: "2025-09-20T01:20:03Z"
          description: AWS Recycle Bin controller is a service controller for managing
            Recycle Bin resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Rule represents the state of an AWS recyclebin Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.recyclebin.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Recycle Bin resources in AWS from within your Kubernetes cluster.

          **About Amazon Recycle Bin**

          Recycle Bin is a data recovery feature that enables you to restore accidentally deleted Amazon EBS snapshots and EBS-backed AMIs. When using Recycle Bin, if your resources are deleted, they are retained in the Recycle Bin for a time period that you specify before being permanently deleted.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Recycle Bin
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - recyclebin
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Recycle Bin Developer Resources
          url: https://docs.aws.amazon.com/ebs/latest/userguide/recycle-bin.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: recyclebin maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/recyclebin-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-recyclebin-controller.v1.1.1
        version: 1.1.1
      - name: ack-recyclebin-controller.v1.1.0
        version: 1.1.0
      - name: ack-recyclebin-controller.v1.0.13
        version: 1.0.13
      - name: ack-recyclebin-controller.v1.0.12
        version: 1.0.12
      - name: ack-recyclebin-controller.v1.0.11
        version: 1.0.11
      - name: ack-recyclebin-controller.v1.0.10
        version: 1.0.10
      - name: ack-recyclebin-controller.v1.0.9
        version: 1.0.9
      - name: ack-recyclebin-controller.v1.0.8
        version: 1.0.8
      - name: ack-recyclebin-controller.v1.0.7
        version: 1.0.7
      - name: ack-recyclebin-controller.v1.0.6
        version: 1.0.6
      - name: ack-recyclebin-controller.v1.0.5
        version: 1.0.5
      - name: ack-recyclebin-controller.v1.0.2
        version: 1.0.2
      - name: ack-recyclebin-controller.v1.0.1
        version: 1.0.1
      - name: ack-recyclebin-controller.v1.0.0
        version: 1.0.0
      - name: ack-recyclebin-controller.v0.0.3
        version: 0.0.3
      - name: ack-recyclebin-controller.v0.0.2
        version: 0.0.2
      - name: ack-recyclebin-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-recyclebin-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicast-community-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicast-community-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/apicast-operator:v0.8.0
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway built on top of NGINX. It is part
            of the Red Hat 3scale API Management Platform
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |
          The APIcast Operator creates and maintains the Red Hat 3scale API Gateway in several deployment configurations.

          APIcast is an API gateway built on top of [NGINX](https://www.nginx.com/). It is part of the [Red Hat 3scale API Management Platform](https://www.redhat.com/en/technologies/jboss-middleware/3scale).

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator/blob/v0.8.0/doc/operator-user-guide.md#upgrading-APIcast) for more information.

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/apicast-operator/tree/v0.8.0).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: APIcast
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://github.com/3scale/apicast-operator/tree/v0.8.0
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/apicast-operator:v0.8.0
        version: 0.8.0
      entries:
      - name: apicast-community-operator.v0.8.0
        version: 0.8.0
      - name: apicast-community-operator.v0.7.1
        version: 0.7.1
      - name: apicast-community-operator.v0.6.0
        version: 0.6.0
      - name: apicast-community-operator.v0.5.1
        version: 0.5.1
      name: alpha
    - currentCSV: apicast-community-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/apicast-operator:v0.8.0
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway built on top of NGINX. It is part
            of the Red Hat 3scale API Management Platform
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |
          The APIcast Operator creates and maintains the Red Hat 3scale API Gateway in several deployment configurations.

          APIcast is an API gateway built on top of [NGINX](https://www.nginx.com/). It is part of the [Red Hat 3scale API Management Platform](https://www.redhat.com/en/technologies/jboss-middleware/3scale).

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator/blob/v0.8.0/doc/operator-user-guide.md#upgrading-APIcast) for more information.

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/apicast-operator/tree/v0.8.0).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: APIcast
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://github.com/3scale/apicast-operator/tree/v0.8.0
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/apicast-operator:v0.8.0
        version: 0.8.0
      entries:
      - name: apicast-community-operator.v0.8.0
        version: 0.8.0
      - name: apicast-community-operator.v0.7.1
        version: 0.7.1
      - name: apicast-community-operator.v0.6.0
        version: 0.6.0
      - name: apicast-community-operator.v0.5.1
        version: 0.5.1
      name: stable
    defaultChannel: stable
    packageName: apicast-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-appsec
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-appsec.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Appsec",
                "metadata": {
                  "name": "appsec-sample"
                },
                "spec": {
                  "appsecDataModel": {
                    "port": 8080
                  },
                  "appsecResourceService": {
                    "port": 8080
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "remoteRepository": "docker.io",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "reportTemplates": {
                    "enabled": false
                  },
                  "systemMonitor": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:39:34Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Appsec
            name: appsecs.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP APPSEC Operator
        displayName: Keysight WAP APPSEC Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Appsec
          url: https://keysight-wap-appsec.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-appsec@sha256:e39ff225c26b750f86553693dda47d0dcc9834a4ab0975aea154361ee3527314
        version: 0.1.21
      entries:
      - name: keysight-wap-appsec.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-appsec
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-operator.v23.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerify",
                "metadata": {
                  "name": "ibmsecurityverify-sample"
                },
                "spec": {
                  "clientSecret": "--secret--",
                  "sessionLifetime": 3600,
                  "ssoPath": "/verify-sso"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isv-saas/verify-operator@sha256:c3bed2e5db48e54884d2278d24ca4864f389a56a667db481c627d01ab5ffaba7
          createdAt: "2023-03-22T02:49:11Z"
          description: The IBM Security Verify operator can consistently enforce policy-driven
            security, including authentication and authorization, using the Ingress
            networking capability of OpenShift.
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.9"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM-Security/verify-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerify is the Schema for the ibmsecurityverifies
              API.
            displayName: IBMSecurity Verify
            kind: IBMSecurityVerify
            name: ibmsecurityverifies.ibm.com
            version: v1
        description: "[IBM Security Verify](https://www.ibm.com/products/verify-saas)
          is an Identity-as-a-Service platform that allows IT, security and business
          leaders to protect their digital users, assets and data in a hybrid, multi-cloud
          world by enabling technical agility and operational efficiency. IBM Security
          Verify SaaS provides single sign-on (SSO), multi-factor authentication (MFA),
          AI-powered context for risk-based authentication for adaptive access decisions,
          user management, access recertification campaigns and identity analytics.\n\nAs
          part of using Red Hat OpenShift, you are entitled to Verify SaaS SSO for
          unlimited apps and users.  If you do not already have a Verify SaaS SSO
          tenant a new one can be [created](https://www.ibm.com/account/reg/signup?formid=urx-51255).\n\nFor
          a detailed description of IBM Security Verify refer to the [Official documentation](https://www.ibm.com/docs/en/security-verify).\n\nThe
          IBM Security Verify operator can consistently enforce policy-driven security
          by using the Ingress networking capability of OpenShift in conjunction with
          the [Nginx Ingress operator](https://www.nginx.com/blog/getting-started-nginx-ingress-operator-red-hat-openshift/).
          With this approach, you can enforce authentication and authorization policies
          for all of the applications in your cluster at the same time, without ever
          changing your application code. You can also dynamically register your application
          to start protecting them centrally from the cloud via Verify SaaS. \n\nSee
          the project [Readme](https://github.com/IBM-Security/verify-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Security Verify Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Product Information
          url: https://www.ibm.com/products/verify-saas
        - name: Verify Documentation
          url: https://www.ibm.com/docs/en/security-verify
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Development Team
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - icr.io/isv-saas/verify-operator@sha256:c3bed2e5db48e54884d2278d24ca4864f389a56a667db481c627d01ab5ffaba7
        version: 23.3.0
      entries:
      - name: ibm-security-verify-operator.v23.3.0
        version: 23.3.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hitachi
      provider-url: ""
    name: hspc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hspc-operator.v1.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.hitachi.com/v1",
                "kind": "HSPC",
                "metadata": {
                  "name": "hspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.hitachivantara.com/hitachicsi-oci-oss/hspc-operator@sha256:4dda4a43cefc19c8b1992756b499b1f9300920c5250dd77379f523997cfd5d54
          createdAt: "2025-06-13T10:30:39Z"
          description: An operator for managing Hitachi Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: Hitachi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HSPC is the Schema for the hspcs API
            displayName: HSPC
            kind: HSPC
            name: hspcs.csi.hitachi.com
            version: v1
        description: |-
          ## About
          Hitachi Storage Plug-in for Containers is a software plugin component for Kubernetes and OpenShift environments that is used to create and manage persistent block storage from Hitachi Virtual Storage Platform One storages.


          For full documentation, go to our [Product Documentation site](https://docs.hitachivantara.com/home) and refer to the reference guide for HSPC v3.16.0.

          ## Requirements

          ### Supported Driver Version

          * HSPC v3.16.0
        displayName: Hitachi Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hitachi
        - Storage
        - Container Storage Interface
        - CSI
        links:
        - name: Product Documentation
          url: https://docs.hitachivantara.com/home
        maintainers:
        - email: Container-info@hitachivantara.com
          name: Hitachi
        maturity: stable
        provider:
          name: Hitachi
        relatedImages:
        - registry.hitachivantara.com/hitachicsi-oci-oss/hspc-operator@sha256:4dda4a43cefc19c8b1992756b499b1f9300920c5250dd77379f523997cfd5d54
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - registry.hitachivantara.com/hitachicsi-oci-oss/hspc-csi-driver@sha256:2c580d98be55fffd8f6083379b6f23cbfe2aeb42fc4aec9e890072bc831c87c4
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:81d32545fb1b2c319bfeb36a70db4b4d3aea9c51c0574b209b8f64750b32d2f0
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        version: 1.16.0
      entries:
      - name: hspc-operator.v1.16.0
        version: 1.16.0
      name: stable
    defaultChannel: stable
    packageName: hspc-operator
    provider:
      name: Hitachi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CyberArk
      provider-url: https://www.cyberark.com/
    name: conjur-follower-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: conjur-follower-operator.v2.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:75fb5de2c111084ffcc3d973e6c63f1adfae78e29f6fd02152419fbca553ce7b",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:bee2cd31b3620a31e98f9d21326a2cdcecb36479099b06cc9660e4e83354d940",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:383776c23556a0a327a0bacc843d22c4b1516082ddd0cfb8a8275500f2eb841a",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:a9b109c7f3bfba5553aa4366338fa211896564f8e33856d75b4918acf8429f1e",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:b3096c17e7f3b9d723c906619dde293ea084e78c0da48c534335a8eee07b23c7",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:680a2044bc0e94fbc9376cd1eed9d067de91657f566c9752654480c939447841",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:ed53aea1866b1e2262386ac38cb9e1e47cc3f9047ae47f49070e2f506af6ca5a"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0 <2.5.3'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:cf46bdb4238cdb1dc51329822da55b7836958be2344d26f9e1819c0aa97fece6
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:383776c23556a0a327a0bacc843d22c4b1516082ddd0cfb8a8275500f2eb841a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:75fb5de2c111084ffcc3d973e6c63f1adfae78e29f6fd02152419fbca553ce7b
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:bee2cd31b3620a31e98f9d21326a2cdcecb36479099b06cc9660e4e83354d940
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:a9b109c7f3bfba5553aa4366338fa211896564f8e33856d75b4918acf8429f1e
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:b3096c17e7f3b9d723c906619dde293ea084e78c0da48c534335a8eee07b23c7
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:680a2044bc0e94fbc9376cd1eed9d067de91657f566c9752654480c939447841
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:ed53aea1866b1e2262386ac38cb9e1e47cc3f9047ae47f49070e2f506af6ca5a
        version: 2.5.3
      entries:
      - name: conjur-follower-operator.v2.5.3
        version: 2.5.3
      name: "2.5"
    - currentCSV: conjur-follower-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:c5337633eb3cc272f989e3e70c2e9bace9c819bba9e817537c7a691a8b6ee596
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813
        version: 2.6.3
      entries:
      - name: conjur-follower-operator.v2.6.3
        version: 2.6.3
      - name: conjur-follower-operator.v2.6.2
        version: 2.6.2
      - name: conjur-follower-operator.v2.6.1
        version: 2.6.1
      - name: conjur-follower-operator.v2.6.0
        version: 2.6.0
      - name: conjur-follower-operator.v2.5.0
        version: 2.5.0
      - name: conjur-follower-operator.v2.4.4
        version: 2.4.4
      - name: conjur-follower-operator.v2.4.3
        version: 2.4.3
      - name: conjur-follower-operator.v2.4.2
        version: 2.4.2
      - name: conjur-follower-operator.v2.3.5
        version: 2.3.5
      - name: conjur-follower-operator.v2.3.4
        version: 2.3.4
      - name: conjur-follower-operator.v2.3.1-290
        version: 2.3.1
      - name: conjur-follower-operator.v2.3.0
        version: 2.3.0
      - name: conjur-follower-operator.v2.2.4
        version: 2.2.4
      - name: conjur-follower-operator.v2.2.2
        version: 2.2.2
      - name: conjur-follower-operator.v2.2.1
        version: 2.2.1
      name: preview
    - currentCSV: conjur-follower-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:c5337633eb3cc272f989e3e70c2e9bace9c819bba9e817537c7a691a8b6ee596
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f
        version: 2.6.3
      entries:
      - name: conjur-follower-operator.v2.6.3
        version: 2.6.3
      - name: conjur-follower-operator.v2.6.2
        version: 2.6.2
      - name: conjur-follower-operator.v2.6.1
        version: 2.6.1
      - name: conjur-follower-operator.v2.6.0
        version: 2.6.0
      - name: conjur-follower-operator.v2.5.0
        version: 2.5.0
      - name: conjur-follower-operator.v2.4.4
        version: 2.4.4
      - name: conjur-follower-operator.v2.4.3
        version: 2.4.3
      - name: conjur-follower-operator.v2.4.2
        version: 2.4.2
      - name: conjur-follower-operator.v2.3.5
        version: 2.3.5
      - name: conjur-follower-operator.v2.3.4
        version: 2.3.4
      - name: conjur-follower-operator.v2.3.1-290
        version: 2.3.1
      - name: conjur-follower-operator.v2.3.0
        version: 2.3.0
      - name: conjur-follower-operator.v2.2.4
        version: 2.2.4
      - name: conjur-follower-operator.v2.2.2
        version: 2.2.2
      - name: conjur-follower-operator.v2.2.1
        version: 2.2.1
      name: stable
    defaultChannel: stable
    packageName: conjur-follower-operator
    provider:
      name: CyberArk
      url: https://www.cyberark.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: ovms-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openvino-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "ModelServer",
                "metadata": {
                  "name": "model-server-sample"
                },
                "spec": {
                  "image_name": "registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0",
                  "deployment_parameters": {
                    "replicas": 1
                  },
                  "service_parameters": {
                    "grpc_port": 8080,
                    "rest_port": 8081
                  },
                  "models_settings": {
                    "single_model_mode": true,
                    "config_configmap_name": "",
                    "config_path": "",
                    "model_name": "resnet",
                    "model_path": "gs://<bucket_name>/<model_dir>",
                    "nireq": 0,
                    "plugin_config": "{\"PERFORMANCE_HINT\":\"LATENCY\"}",
                    "batch_size": "",
                    "shape": "",
                    "model_version_policy": "{\"latest\": { \"num_versions\":1 }}",
                    "layout": "",
                    "target_device": "CPU",
                    "is_stateful": false,
                    "idle_sequence_cleanup": false,
                    "low_latency_transformation": true,
                    "max_sequence_number": 0
                  },
                  "server_settings": {
                    "file_system_poll_wait_seconds": 0,
                    "sequence_cleaner_poll_wait_minutes": 0,
                    "log_level": "INFO",
                    "grpc_workers": 1,
                    "rest_workers": 0
                  },
                  "models_repository": {
                    "https_proxy": "",
                    "http_proxy": "",
                    "storage_type": "google",
                    "models_host_path": "",
                    "models_volume_claim": "",
                    "aws_secret_access_key": "",
                    "aws_access_key_id": "",
                    "aws_region": "",
                    "s3_compat_api_endpoint": "",
                    "gcp_creds_secret_name": "",
                    "azure_storage_connection_string": ""
                  }
                }
              },
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "Notebook",
                "metadata": {
                  "name": "v2024.1"
                },
                "spec": {
                  "git_ref": "2024.1",
                  "auto_update_image": false
                }
              }
            ]
          alm-examples-metadata: '{ "model-server-sample":{"description":"Example
            of the model server configuration"}, "v2024":{"description":"Default notebook
            configuration"} }'
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
          createdAt: "2024-05-14T09:29:35Z"
          description: OpenVINO Toolkit Operator manages OpenVINO components in OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/ovms-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/ovms-operator-rhmp/support?utm_source=openshift_console
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.34.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Provides inference serving through gRPC & REST
            displayName: Model Server
            kind: ModelServer
            name: modelservers.intel.com
            version: v1alpha1
          - description: Adds a collection of ready-to-run Jupyter notebooks for learning
              and experimenting with the OpenVINO™ Toolkit to Red Hat® OpenShift®
              Data Science
            displayName: Notebook
            kind: Notebook
            name: notebooks.intel.com
            version: v1alpha1
        description: |
          OpenVINO Toolkit Operator manages OpenVINO components in OpenShift.

          Currently there available components are ModelServer and Notebook.

          # Model Server
          [OpenVINO™ Model Server](https://github.com/openvinotoolkit/model_server) (OVMS) is a scalable, high-performance solution for serving machine learning models optimized for Intel® architectures. The server provides an inference service via gRPC or REST API for any models trained in a framework that is supported by [OpenVINO](https://software.intel.com/en-us/openvino-toolki://docs.openvino.ai/latest/index.html).
          Model Server configuration parameters is explained [here](https://github.com/openvinotoolkit/operator/blob/main/docs/modelserver_params.md).
          ## Using the cluster
          OpenVINO Model Server can be consumed as a `Service`. It is called like with `ModelServer` resource with `-ovms` suffix.
          The suffix is omitted when `ovms` phrase is included in the name.
          The service exposes gRPC and REST API interfaces to run inference requests.
          ```
          oc get pods
          NAME                                        READY   STATUS    RESTARTS   AGE
          model-server-sample-ovms-586f6f76df-dpps4   1/1     Running   0          8h

          oc get services
          NAME                       TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)             AGE
          model-server-sample-ovms   ClusterIP   172.25.199.210   <none>        8080/TCP,8081/TCP   8h
          ```

          # Notebook
          The Notebook resource integrates JupyterHub from OpenShift Data Science or Open Data Hub with a container image that includes [developer
          tools](https://github.com/openvinotoolkit/openvino/blob/master/docs/install_guides/pypi-openvino-dev.md)
          from the OpenVINO toolkit and a set of Jupyter notebook tutorials. It enables selecting a defined image `OpenVINO™ Toolkit` from
          the Jupyter Spawner choice list.

          Create the `Notebook` resource in the same project with JupyterHub and RedHat OpenShift Data Science operator as seen in [documentation](https://github.com/openvinotoolkit/operator/blob/main/docs/notebook_in_rhods.md).
          It builds the image in the cluster based on Dockerfile from [openvino_notebooks](https://github.com/openvinotoolkit/openvino_notebooks).

          ## References
          OpenVINO Model Server on [Github](https://github.com/openvinotoolkit/model_server)

          OpenVINO Model Server Operator on [Github](https://github.com/openvinotoolkit/operator)
        displayName: OpenVINO Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Inference OpenVINO
        links:
        - name: OpenVINO Toolkit Operator
          url: https://github.com/openvinotoolkit/operator
        - name: OpenVINO Model Server
          url: https://github.com/openvinotoolkit/model_server
        - name: OpenVINO Notebooks
          url: https://github.com/openvinotoolkit/openvino_notebooks
        maintainers:
        - email: openvino-operator-support@intel.com
          name: Operator Support
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5714b93ac39277affdc10c3ba142d8c032ed0ddb2ba61c04cbddac710464c4
        - registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
        - registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0
        version: 1.2.0
      entries:
      - name: openvino-operator.v1.2.0
        version: 1.2.0
      - name: openvino-operator.v1.1.0
        version: 1.1.0
      - name: openvino-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: ovms-operator-rhmp
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Aerospike
      provider-url: ""
    name: aerospike-kubernetes-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aerospike-kubernetes-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "memory-size": 3000000000,
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:6.3.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: quay.io/aerospike/aerospike-kubernetes-operator@sha256:a3033b50ae434b97de4de1ca03013dabcfac580d810a6ee8c0fe05b6a9301952
          createdAt: 2023-05-04T09:32:47UTC
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: <=2.4.0
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Aerospike
        relatedImages:
        - quay.io/aerospike/aerospike-kubernetes-operator@sha256:a3033b50ae434b97de4de1ca03013dabcfac580d810a6ee8c0fe05b6a9301952
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        version: 2.5.0
      entries:
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      name: alpha
    - currentCSV: aerospike-kubernetes-operator-rhmp.v4.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "data-size": 1073741824,
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:8.1.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackup",
                "metadata": {
                  "name": "aerospikebackup",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "aerospike-cluster": {
                      "aerospike-aerospikebackup-test-cluster": {
                        "credentials": {
                          "password": "admin123",
                          "user": "admin"
                        },
                        "seed-nodes": [
                          {
                            "host-name": "aerocluster.aerospike.svc.cluster.local",
                            "port": 3000
                          }
                        ]
                      }
                    },
                    "backup-routines": {
                      "aerospike-aerospikebackup-test-routine": {
                        "backup-policy": "test-policy",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "local"
                      },
                      "aerospike-aerospikebackup-test-routine1": {
                        "backup-policy": "test-policy1",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "s3Storage"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackupService",
                "metadata": {
                  "name": "aerospikebackupservice-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "config": {
                    "backup-policies": {
                      "test-policy": {
                        "parallel": 3
                      },
                      "test-policy1": {
                        "parallel": 3
                      }
                    },
                    "service": {
                      "http": {
                        "port": 8081
                      }
                    },
                    "storage": {
                      "local": {
                        "local-storage": {
                          "path": "/tmp/localStorage"
                        }
                      },
                      "s3Storage": {
                        "s3-storage": {
                          "bucket": "aerospike-kubernetes-operator-test",
                          "s3-profile": "default",
                          "s3-region": "us-east-1"
                        }
                      }
                    }
                  },
                  "image": "aerospike/aerospike-backup-service:3.2.0",
                  "secrets": [
                    {
                      "secretName": "aws-secret",
                      "volumeMount": {
                        "mountPath": "/root/.aws/credentials",
                        "name": "aws-secret",
                        "subPath": "credentials"
                      }
                    }
                  ],
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeRestore",
                "metadata": {
                  "name": "aerospikerestore-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "backup-data-path": "aerospike-aerospikebackup-test-routine/backup/1733506015827/data/test",
                    "destination": {
                      "credentials": {
                        "password": "admin123",
                        "user": "admin"
                      },
                      "label": "destinationCluster",
                      "seed-nodes": [
                        {
                          "host-name": "aerocluster.test.svc.cluster.local",
                          "port": 3000
                        }
                      ]
                    },
                    "policy": {
                      "no-generation": true,
                      "no-indexes": true,
                      "parallel": 3
                    },
                    "source": {
                      "local-storage": {
                        "path": "/tmp/localStorage"
                      }
                    }
                  },
                  "type": "Full"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/aerospike/aerospike-kubernetes-operator@sha256:e6d6bd7c58e036b3728e52475a9b764ca8c46b83cbc4751cac66294495bf2b31
          createdAt: "2025-09-15T10:14:05Z"
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
          support: Aerospike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeBackup is the Schema for the aerospikebackup API
            displayName: Aerospike Backup
            kind: AerospikeBackup
            name: aerospikebackups.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeBackupService is the Schema for the aerospikebackupservices
              API
            displayName: Aerospike Backup Service
            kind: AerospikeBackupService
            name: aerospikebackupservices.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1
          - description: AerospikeRestore is the Schema for the aerospikerestores
              API
            displayName: Aerospike Restore
            kind: AerospikeRestore
            name: aerospikerestores.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
           -   Aerospike cluster monitoring
           -   Backup and restore Aerospike clusters
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Aerospike
        relatedImages:
        - quay.io/aerospike/aerospike-kubernetes-operator@sha256:e6d6bd7c58e036b3728e52475a9b764ca8c46b83cbc4751cac66294495bf2b31
        version: 4.1.1
      entries:
      - name: aerospike-kubernetes-operator-rhmp.v4.1.1
        version: 4.1.1
      - name: aerospike-kubernetes-operator-rhmp.v4.1.0
        version: 4.1.0
      - name: aerospike-kubernetes-operator-rhmp.v4.0.2
        version: 4.0.2
      - name: aerospike-kubernetes-operator-rhmp.v4.0.1
        version: 4.0.1
      - name: aerospike-kubernetes-operator-rhmp.v4.0.0
        version: 4.0.0
      - name: aerospike-kubernetes-operator-rhmp.v3.4.1
        version: 3.4.1
      - name: aerospike-kubernetes-operator-rhmp.v3.4.0
        version: 3.4.0
      - name: aerospike-kubernetes-operator-rhmp.v3.3.1
        version: 3.3.1
      - name: aerospike-kubernetes-operator-rhmp.v3.3.0
        version: 3.3.0
      - name: aerospike-kubernetes-operator-rhmp.v3.2.2
        version: 3.2.2
      - name: aerospike-kubernetes-operator-rhmp.v3.2.1
        version: 3.2.1
      - name: aerospike-kubernetes-operator.v3.2.0
        version: 3.2.0
      - name: aerospike-kubernetes-operator.v3.1.0
        version: 3.1.0
      - name: aerospike-kubernetes-operator.v3.0.0
        version: 3.0.0
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      name: stable
    defaultChannel: stable
    packageName: aerospike-kubernetes-operator-rhmp
    provider:
      name: Aerospike
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: samsung
      provider-url: ""
    name: socmmd
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: socmmd.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cmmd.samsung.com/v1",
                "kind": "CMMD",
                "metadata": {
                  "name": "my-app",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "allocateMode": "auto",
                  "enable": true,
                  "payload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "metadata": {
                      "name": "my-app"
                    },
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "name": "my-app"
                        }
                      },
                      "template": {
                        "metadata": {
                          "labels": {
                            "name": "my-app"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "quay.io/my-account/my-app:latest",
                              "name": "my-app-01"
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "cmmd.samsung.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "machineconfiguration.openshift.io/role": "cmmd"
                  },
                  "name": "cmmd-node-label",
                  "namespace": "socmmd"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          console.openshift.io/plugins: '["socmmd-console-plugin"]'
          containerImage: quay.io/samsung_sag/socmmd@sha256:2ef361fb15f0fec99e0a8f1ef83ad4dd14f701d579c900cce084c4a62cf805f8
          createdAt: "2024-08-01T05:45:38Z"
          description: Samsung Operator For CMM-D
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "cmmd.samsung.com/v1",
              "kind": "Label",
              "metadata": {
                "name": "cmmd-node-label",
                "namespace": "socmmd",
                "labels": {}
              },
              "spec": {}
            }
          operatorframework.io/suggested-namespace: socmmd-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenMPDK/SMDK/tree/main/lib/redhat/socmmd
          support: samsung
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CMMD is the Schema for the cmmds API.
            displayName: CMMD
            kind: CMMD
            name: cmmds.cmmd.samsung.com
            version: v1
          - description: Label is the Schema for the labels API.
            displayName: Node Label
            kind: Label
            name: labels.cmmd.samsung.com
            version: v1
        description: Samsung Operator For CMM-D
        displayName: socmmd
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - samsung
        - memory
        - extend
        - cmm-d
        links:
        - name: Samsung Operator For CMM-D
          url: https://github.com/OpenMPDK/SMDK/tree/main/lib/redhat/socmmd
        maintainers:
        - email: junhyeok.im@samsung.com
          name: junhyeok-im
        maturity: stable
        provider:
          name: samsung
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/samsung_sag/socmmd@sha256:2ef361fb15f0fec99e0a8f1ef83ad4dd14f701d579c900cce084c4a62cf805f8
        version: 0.0.2
      entries:
      - name: socmmd.v0.0.2
        version: 0.0.2
      - name: socmmd.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: socmmd
    provider:
      name: samsung
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/file-integrity-operator
    name: file-integrity-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: file-integrity-operator.v0.1.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <0.1.13'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator
          url: https://file-integrity-operator.domain
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/openshift4/file-integrity-rhel8-operator@sha256:a65f1244636649734763164a5d6155a6f6a7b79b3950ecc03fb3c58188583599
        version: 0.1.13
      entries:
      - name: file-integrity-operator.v0.1.13
        version: 0.1.13
      name: "4.7"
    - currentCSV: file-integrity-operator.v0.1.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <0.1.32'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:c3e34e6198eadb48c007cd949d8be38e26ef7d15d2f9047a8c3c89c03201e4ea
        version: 0.1.32
      entries:
      - name: file-integrity-operator.v0.1.32
        version: 0.1.32
      name: release-0.1
    - currentCSV: file-integrity-operator.v1.3.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          createdAt: "2024-11-21T18:02:08Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.3.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:f7039c6aaaa3e60dc4a3965cc1341c5217837e4b4030eb943771967e0b206578
        version: 1.3.6
      entries:
      - name: file-integrity-operator.v1.3.6
        version: 1.3.6
      - name: file-integrity-operator.v1.3.5
        version: 1.3.5
      - name: file-integrity-operator.v1.3.4
        version: 1.3.4
      - name: file-integrity-operator.v1.3.3
        version: 1.3.3
      name: stable
    - currentCSV: file-integrity-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <1.0.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:dd1dfeccf59214a0de653c3a4c4dbbc8c41640954318d13386cda73891660e4f
        version: 1.0.0
      entries:
      - name: file-integrity-operator.v1.0.0
        version: 1.0.0
      name: v1
    defaultChannel: stable
    packageName: file-integrity-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/file-integrity-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Jason Skrzypek
      provider-url: https://github.com/shpwrck
    name: openshift-node-upgrade-mutex-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-node-upgrade-mutex-operator.v0.2.66
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexRule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutexrule-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutexrule",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutexrule-sample"
                },
                "spec": {
                  "jobName": "example",
                  "jobNamespace": "node-upgrade-mutex-operator",
                  "jobSpec": {
                    "selector": {},
                    "template": {
                      "metadata": {
                        "name": "mutex"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              "20"
                            ],
                            "image": "ubi9/toolbox",
                            "name": "mutex"
                          }
                        ],
                        "restartPolicy": "Never"
                      }
                    }
                  },
                  "type": "kubernetes"
                }
              },
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexRunConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutexrunconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutexrunconfig",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutexrunconfig-sample"
                },
                "spec": {
                  "canary": {
                    "name": "canary-example"
                  },
                  "mutexRules": [
                    {
                      "name": "mutexrule-sample",
                      "namespace": "node-upgrade-mutex-operator"
                    }
                  ],
                  "mutexTargets": [
                    {
                      "name": "mutextarget-sample",
                      "namespace": "node-upgrade-mutex-operator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexTarget",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutextarget-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutextarget",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutextarget-sample"
                },
                "spec": {
                  "machineConfigPools": [
                    "worker"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/jskrzypek/openshift-node-upgrade-mutex-operator
          createdAt: "2025-05-30T17:49:00Z"
          description: |
            This software injects required operations into the worker node upgrade process.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: node-upgrade-mutex-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          provider: Jason Skrzypek
          repository: https://github.com/shpwrck/openshift-upgrade-rollout-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A node mutex rule definition in the form of a kubernetes
              job spec
            displayName: Mutex Rule
            kind: MutexRule
            name: mutexrules.openshift-optional.skrzypek.io
            version: v1alpha1
          - description: Lists of Targets and Rules to apply to them, as well as an
              upgrade facilitating Machine Config Pool
            displayName: Mutex Run Config
            kind: MutexRunConfig
            name: mutexrunconfigs.openshift-optional.skrzypek.io
            version: v1alpha1
          - description: A list of Machine Config Pools to enforce upgrade procedure.
            displayName: Mutex Target
            kind: MutexTarget
            name: mutextargets.openshift-optional.skrzypek.io
            version: v1alpha1
        description: |
          ## Introduction

          This operator injects required operations in the form of a Kubernetes job into the worker node upgrade process using MachineConfigPools. In short, when provided a list of operations (`MutexRules`) and a list of MachineConfigPools (`MutexTargets`) , the OpenShift Node Upgrade Mutex Operator will execute each operation for a given target, enable updates for the given target by modifying it's MachineConfigPool membership, and wait until the node has been upgraded before proceeding onto subsequent operations and targets.

          ## Dependencies

          ### Release Version Label

          In order to confirm the successful upgrade of each target node, the NFD Operator is used to generate a version feature on every worker node. In order to proceed through given targets, each node must have this label
          `'machineconfiguration.openshift.io/release-image-version'` with a key providing the semantic version of this node (e.g. `'4.16.4'`)

          There are several ways to accomplish this, but an example MachineConfig has been provided [here](https://github.com/shpwrck/openshift-node-upgrade-mutex-operator/blob/main/machineconfig.yaml).

          ### Canary Machine Config Pool

          In addition to node labels, a MachineConfigPool that will be used to accomodate targets subject to upgrades must be provided matching the following criteria:
          * selects only worker nodes
          * is not paused

          Applying the example below will create a compatible canary pool:

          ```yaml
          ---
          apiVersion: machineconfiguration.openshift.io/v1
          kind: MachineConfigPool
          metadata:
            name: canary
          spec:
            machineConfigSelector:
              matchLabels:
                machineconfiguration.openshift.io/role: worker
            nodeSelector:
              matchLabels:
                node-role.kubernetes.io/worker: ''
              matchExpressions:
                - key: 'isCanary'
                  operator: Exists
            paused: false
          ```

          ## Usage / Custom Resources Definitions

          #### MutexRule:

          A resource storing a single rule referring to a template for a Kubernetes job to be created.
          This resource **does not** modify existing resources.

          ```yaml
          ---
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexRule
          metadata:
            name: k8s-example
          spec:
            jobName: "example" # Jobs will be run with this prefix name
            jobNamespace: "node-upgrade-mutex-operator" # Jobs will be run in this namespace
            jobSpec: # Accepts full 'batch/v1' job spec
              selector: {}
              template:
                metadata:
                  name: mutex
                spec:
                  containers:
                  - command:
                    - sleep
                    args:
                    - "30"
                    image: ubi9/toolbox
                    name: mutex
                  restartPolicy: Never
            cleanUpJobSpec: # Accepts full 'batch/v1' job spec (Optional)
              selector: {}
              template:
                metadata:
                  name: mutex
                spec:
                  containers:
                  - command:
                    - sleep
                    args:
                    - "30"
                    image: ubi9/toolbox
                    name: mutex
                  restartPolicy: Never
            type: kubernetes
          ```

          #### MutexTarget:

          A resource storing a set of MachineConfigPools to pause and incrementally upgrade.
          This resource **does not** modify existing resources.

          ```yaml
          ---
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexTarget
          metadata:
            name: worker-example
          spec:
            machineConfigPools:
              - worker # Name of existing machineconfigpool
          ```

          #### MutexRunConfig:

          A resource that references existing MutexRules and MutexTargets as well as an existing MachineConfigPool that will allow members to be upgraded. Every rule defined in the `mutexRules` list will be added as a mandatory task to every node that is a member of the each MachineConfigPool defined in the `mutexTargets` list.
          This resource **does** modify existing resources.

          ```yaml
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexRunConfig
          metadata:
            name: example
          spec:
            skipTLSVerify: false # Defaults to false if omitted
            canary:
              name: canary # Name of unpaused machineconfigpool to modify
            mutexRules: # List of existing mutexRules
              - name: k8s-example
                namespace: node-upgrade-mutex-operator
            mutexTargets: # List of existing mutexTargets
              - name: worker-example
                namespace: node-upgrade-mutex-operator
          ```

          ## Procedure

          #### MutexRule

          1. Upon creation/modification, the schema of each K8s rule is validated.
          2. Status is updated to indicate valid or invalid configuration.

          #### MutexTargets

          1. Upon creation/modification, the existence of each target is validated.
          2. Status is updated to indicate valid or invalid configuration.
          3. The nodes assigned to each target are added to the status field `nodes`.

          #### MutexRunConfigs

          1. Upon creation/modification, the existence as well as validation status of each rule, target, and canary pool is confirmed.
          2. If valid, the existence of an upgrade is queried. An upgrade is available if the following is true:
              * The `status.history` of the `cluster` ClusterVersion equals `Completed`
              * The `release-image-version` of any worker node is less than the `spec.desiredUpdate.version` of the `cluster` ClusterVersion.
          3. If an update exists, each target MachineConfigPool has `spec.paused` set to `true`.
          4. If each pool is paused, each node is updated to include a label for each rule with the following format:
              * `mutexrule.openshift-optional.skrzypek.io/{{MutexRule_Name}}.{{MutexRule_Namespace}}`
          5. If labels have been applied, the provided canary pool is modified to exclude nodes with labels provided by step (4).
          6. Assuming steps (2-5) have completed successfully, then the process for running jobs and updating nodes begins. The process per node is as follows:
              * Launch each K8S job sequentially
              * Remove the current node's rule labels
              * Wait until the `release-image-version` of the node matches the `spec.desiredUpdate.version` of the `cluster` ClusterVersion
              * Launch each K8S cleanup job sequentially
          7. If the update succeeds, all machine config pools are unpaused and all match expressions are removed.

          ## Permissions Granted

          In addition to the permissions required to operate the included CRDs, the following access is granted to the operator's service account.

          | Resource | Permissions |
          | :---: | :---: |
          | Jobs | Create & Read |
          | Nodes | Read & Update |
          | ClusterVersion | Read |
          | MachineConfigPools | Read & Update |

          ## Notes

          * This operator was built using the operator-sdk's ansible plugin and therefore has all of the inherited capabilities and restrictions.
          * This operator has undergone minimal testing and should not be run in production environments without acknowledging the risks.
          * This operator is currently alpha and subject to change based on requirements.

          ## Additional Information
          For additional information refer to the [GitHub Repository](https://github.com/shpwrck/openshift-node-upgrade-mutex-operator)
        displayName: OpenShift Node Upgrade Mutex Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - OpenShift Optional
        - Upgrade
        - Mutex
        links:
        - name: Openshift Node Upgrade Mutex Operator
          url: https://github.com/shpwrck/openshift-node-upgrade-mutex-operator
        maintainers:
        - email: jskrzype@redhat.com
          name: Jason Skrzypek
        maturity: alpha
        minKubeVersion: 1.27.0
        provider:
          name: Jason Skrzypek
          url: https://github.com/shpwrck
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - quay.io/jskrzypek/openshift-node-upgrade-mutex-operator@sha256:d4c28e8b256c00429d9999595c2386f1a25de597ee8e6c33871bb7b22b84c4af
        version: 0.2.66
      entries:
      - name: openshift-node-upgrade-mutex-operator.v0.2.66
        version: 0.2.66
      - name: openshift-node-upgrade-mutex-operator.v0.2.65
        version: 0.2.65
      - name: openshift-node-upgrade-mutex-operator.v0.2.60
        version: 0.2.60
      - name: openshift-node-upgrade-mutex-operator.v0.2.59
        version: 0.2.59
      - name: openshift-node-upgrade-mutex-operator.v0.2.53
        version: 0.2.53
      - name: openshift-node-upgrade-mutex-operator.v0.2.38
        version: 0.2.38
      - name: openshift-node-upgrade-mutex-operator.v0.2.34
        version: 0.2.34
      - name: openshift-node-upgrade-mutex-operator.v0.2.15
        version: 0.2.15
      - name: openshift-node-upgrade-mutex-operator.v0.2.14
        version: 0.2.14
      - name: openshift-node-upgrade-mutex-operator.v0.1.22
        version: 0.1.22
      name: alpha
    defaultChannel: alpha
    packageName: openshift-node-upgrade-mutex-operator
    provider:
      name: Jason Skrzypek
      url: https://github.com/shpwrck
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Digital.ai
      provider-url: https://digital.ai
    name: digitalai-deploy-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: digitalai-deploy-operator.v25.3.0-beta.813
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "quorum",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.0-debian-12-r4"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.0-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.5.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.2-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabsunsupported/deploy-operator@sha256:9ccca8ab999d4b6058c979bbb70568a63cf43a257ae65a787b2ff4d9bf470c39 '
          createdAt: "2025-09-02T05:43:37Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabsunsupported/deploy-operator@sha256:9ccca8ab999d4b6058c979bbb70568a63cf43a257ae65a787b2ff4d9bf470c39
        - docker.io/xebialabsunsupported/xl-deploy@sha256:41feb1d24d70a0c78a560b338d05c96c5383d570e3ef1ff37daff0eb8c92b386
        - docker.io/xebialabsunsupported/deploy-task-engine@sha256:15322f2263c2b9b70336177aec170e59497fa320281c2d73605918b30272b8f1
        - docker.io/xebialabsunsupported/central-configuration@sha256:3a35c2aada71a7f5ab8fe7d5c68441fe58b0b2c778b6ea85ac6f3f7b351f80ad
        version: 25.3.0-beta.813
      entries:
      - name: digitalai-deploy-operator.v25.3.0-beta.813
        version: 25.3.0-beta.813
      - name: digitalai-deploy-operator.v25.3.0-beta.711
        version: 25.3.0-beta.711
      - name: digitalai-deploy-operator.v25.3.0-beta.609
        version: 25.3.0-beta.609
      - name: digitalai-deploy-operator.v25.3.0-beta.512
        version: 25.3.0-beta.512
      name: beta
    - currentCSV: digitalai-deploy-operator.v24.1.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "ingressClass": "haproxy-daid",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "tag": "1.26.0-debian-12-r1"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "tag": "1.10.1-debian-12-r5"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "tag": "15.6.0-debian-12-r7"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "tag": "3.12.13-debian-12-r2"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator without SCC creation and with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:2b3413cd41f2b437e6027c8ae2cff12d1113e0ac1be69fc24840d68a05d3a665 '
          createdAt: "2024-10-16T05:39:21Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind:
          DigitalaiDeploy\nmetadata:\n  name: daid-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  securityContextConstraints:\n
          \   enabled: false\n  centralConfiguration:\n    replicaCount: 1\n    podSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  master:\n    replicaCount: 1\n    persistence:\n
          \     storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n      runAsUser:
          null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  worker:\n    replicaCount: 1\n    persistence:\n
          \     storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n      runAsUser:
          null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  route:\n    enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n        runAsUser: null\n        runAsGroup: null\n
          \       fsGroup: null\n      containerSecurityContext:\n        enabled:
          false\n        runAsUser: null\n        runAsGroup: null\n      securityContextConstraints:\n
          \       enabled: false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n
          \   install: true\n    persistence:\n      size: 1Gi\n      storageClass:
          ''\n    replicaCount: 1\n    podSecurityContext:\n      enabled: false\n
          \     runAsUser: null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     enabled: false\n      runAsUser: null\n      runAsGroup: null\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"daid-doc-digitalai-deploy-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xld-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Embedded PostgreSQL for
          DB management\n- Embedded RabbitMQ for message queue management\n\n### Security
          Configuration Details\n\n- `securityContextConstraints.enabled: false` -
          disables creation of SCCs;\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/deploy-operator@sha256:2b3413cd41f2b437e6027c8ae2cff12d1113e0ac1be69fc24840d68a05d3a665
        - docker.io/xebialabs/deploy-task-engine@sha256:8a381fd84483ea919fa7f8bee5cf19d9dec0d1d61cce73ec6b7c56fa504c6c93
        - docker.io/xebialabs/xl-deploy@sha256:77fdf6a3909367b060eba2ef698baddb77e4086d4d977f9c617d1e26a4427472
        - docker.io/xebialabs/central-configuration@sha256:83359e49c66e4fab502cc75663b456eb37d81d7cb512f2c48667f4d6a324e0df
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:771a9a173e033a3ad8b46f5c00a7036eaa88c8d8d1fbd89217325168998113ea
        version: 24.1.10
      entries:
      - name: digitalai-deploy-operator.v24.1.10
        version: 24.1.10
      name: stable
    - currentCSV: digitalai-deploy-operator.v24.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "image": {
                        "tag": "v0.14.7"
                      },
                      "ingressClass": "haproxy-daid",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.26.2-debian-12-r10"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "15.6.0-debian-12-r20"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.12.14-debian-12-r7"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator without SCC creation and with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:7d505a855728c761433ef7d46f0a64ffe2b4192f4819bad86c23582375099ac9 '
          createdAt: "2025-08-25T04:53:23Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind:
          DigitalaiDeploy\nmetadata:\n  name: daid-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  securityContextConstraints:\n
          \   enabled: false\n  centralConfiguration:\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  master:\n    replicaCount:
          1\n    persistence:\n      storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  worker:\n    replicaCount:
          1\n    persistence:\n      storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  route:\n    enabled: false\n
          \   annotations:\n      haproxy.router.openshift.io/cookie_name: SESSION_XLD\n
          \     haproxy.router.openshift.io/disable_cookies: \"false\"\n      haproxy.router.openshift.io/rewrite-target:
          /\n    hostname: '<mandatory-deploy-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n      securityContextConstraints:\n
          \       enabled: false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n
          \   install: true\n    persistence:\n      size: 1Gi\n      storageClass:
          ''\n    replicaCount: 1\n    podSecurityContext:\n      enabled: false\n
          \   containerSecurityContext:\n      enabled: false\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"daid-doc-digitalai-deploy-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xld-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Embedded PostgreSQL for
          DB management\n- Embedded RabbitMQ for message queue management\n\n### Security
          Configuration Details\n\n- `securityContextConstraints.enabled: false` -
          disables creation of SCCs;\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/xl-deploy@sha256:c2fe209f198fb2f42dc78a2e1544c026a6891d60682b4e47634234bbf7346540
        - docker.io/xebialabs/central-configuration@sha256:581a0bc71ae3652ceb643d81927fcf8cabea92f1d301f5d354076baff32fbcde
        - docker.io/xebialabs/deploy-operator@sha256:7d505a855728c761433ef7d46f0a64ffe2b4192f4819bad86c23582375099ac9
        - docker.io/xebialabs/deploy-task-engine@sha256:a97d67ea53e1b954ffb4d9514fd61426dcc0cab08922009eef02f32162026d6a
        version: 24.1.21
      entries:
      - name: digitalai-deploy-operator.v24.1.21
        version: 24.1.21
      - name: digitalai-deploy-operator.v24.1.20
        version: 24.1.20
      - name: digitalai-deploy-operator.v24.1.19
        version: 24.1.19
      - name: digitalai-deploy-operator.v24.1.18
        version: 24.1.18
      - name: digitalai-deploy-operator.v24.1.17
        version: 24.1.17
      - name: digitalai-deploy-operator.v24.1.14
        version: 24.1.14
      - name: digitalai-deploy-operator.v24.1.10
        version: 24.1.10
      name: stable-v24.1
    - currentCSV: digitalai-deploy-operator.v24.3.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "16.3.0-debian-12-r23"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n{{ if eq .Values.queueType \"quorum\" -}}\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n{{- end }}\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.13.7-debian-12-r5"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "queueType": "classic",
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:41ffc26c1d99a808e537f2366d92473ecf76dbffbac699dfa04785c3016f7470 '
          createdAt: "2025-09-26T18:18:41Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/central-configuration@sha256:4b83d4e9a523ae3615579907c65790478a3b22315ca5afce594793424d1d33b1
        - docker.io/xebialabs/deploy-operator@sha256:41ffc26c1d99a808e537f2366d92473ecf76dbffbac699dfa04785c3016f7470
        - docker.io/xebialabs/xl-deploy@sha256:8536efdf530dde45b84b14191ebfdc7ec5ebb1844f84871032fe2d335b13908d
        - docker.io/xebialabs/deploy-task-engine@sha256:75fe942305bf3bb403eb243b157b88844363eb702f03227a61e76ad4321bf0ee
        version: 24.3.11
      entries:
      - name: digitalai-deploy-operator.v24.3.11
        version: 24.3.11
      - name: digitalai-deploy-operator.v24.3.10
        version: 24.3.10
      - name: digitalai-deploy-operator.v24.3.9
        version: 24.3.9
      - name: digitalai-deploy-operator.v24.3.8
        version: 24.3.8
      - name: digitalai-deploy-operator.v24.3.7
        version: 24.3.7
      - name: digitalai-deploy-operator.v24.3.6
        version: 24.3.6
      - name: digitalai-deploy-operator.v24.3.5
        version: 24.3.5
      name: stable-v24.3
    - currentCSV: digitalai-deploy-operator.v25.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "quorum",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.1-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.4.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.3-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:d192ef9008d72c9792e1b7b988bdae8f1bbce45ed0e44afc48e3e268b657b38b '
          createdAt: "2025-09-26T16:30:53Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/deploy-task-engine@sha256:13bb779d6041836acb95c95f1eed75f4b5b2940d75687306183a9f8a2177cab3
        - docker.io/xebialabs/central-configuration@sha256:0c46275ffa3b251ccb76d2623e0e5a4cd1441bf7f94901845bda743f087bc6e4
        - docker.io/xebialabs/xl-deploy@sha256:e4212d20fad19f09cbe53796371bec7fc368cfe63316f1b7bccecf5faf6ea733
        - docker.io/xebialabs/deploy-operator@sha256:d192ef9008d72c9792e1b7b988bdae8f1bbce45ed0e44afc48e3e268b657b38b
        version: 25.1.5
      entries:
      - name: digitalai-deploy-operator.v25.1.5
        version: 25.1.5
      - name: digitalai-deploy-operator.v25.1.4
        version: 25.1.4
      - name: digitalai-deploy-operator.v25.1.3
        version: 25.1.3
      - name: digitalai-deploy-operator.v25.1.2
        version: 25.1.2
      - name: digitalai-deploy-operator.v25.1.1
        version: 25.1.1
      - name: digitalai-deploy-operator.v25.1.0
        version: 25.1.0
      name: stable-v25.1
    defaultChannel: stable-v24.3
    packageName: digitalai-deploy-operator
    provider:
      name: Digital.ai
      url: https://digital.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://access.redhat.com/products/red-hat-fuse
    name: fuse-online
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-online.v7.10.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:ed4faedcb97bb6a45f568586ff19da19da31ecbc304f12e0ae7978381247744a
          createdAt: 2022-01-31 15:55:46 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:ed4faedcb97bb6a45f568586ff19da19da31ecbc304f12e0ae7978381247744a
        - registry.redhat.io/rhel8/postgresql-12@sha256:44215c5b244b190c82d9d371a8ffd39b0c83576bd85c980cc72782a5cf9e6e1b
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:5048cd4fab2f98e6000bc60835c539cc40b70dbe265e927cfe185595347d8054
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:5dd1bc8f6b7e724736f4aeac86409bda3e0f64f62b736688e9572d14411232f1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:fcdd0292ebfc7e9edc36241f63b54a9d924a7605329c934ff2b10373af26dbbf
        - registry.redhat.io/openshift4/ose-prometheus@sha256:7f679d6c03535e300dac366f4bc6dd8d781d96dc47592836713f8bb76be36fa9
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:4f8849362e9ed6ff95d318a04e287b7ba33254a2cf7abdaf9b90cab7cbba6646
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:df2d55dd8092e9e94d034166890117cf5d1cfd5cf52ef0f90cfa2374cc9e2d06
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:3f0277e2c278740307c9b2d966f539ba12317e306198cb6c4c1c52a2ad9f4f40
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:96810c770f75057b8bba6b78a5ba414c8289315d8af1a387f681ae5e4fb7d6dd
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:ef229a4c8851ec495f28d0eeab83b137cd02698e283f76ea1b32271b7222621b
        version: 7.10.4
      entries:
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.10.x
    - currentCSV: fuse-online.v7.11.1-0.1687357787.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:658af3fc7819efb7cc39cd3a0504dcddc2ce7788c9b1e5178df869f7c9a742a8
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.11.1-0.1684718286.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:a2b1cd0256f30aff8fa37c332419b4483c0bfe34fb4f574ad8f18b0858969342
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:4de91e4527022e1b25df8d3835ddf778f773c20334314be9ba01a885c82287f5
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:8b9bd777bcbd20b1a1bbbc18274a5076a67bea1f928d772cb243c57fb1da7408
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:ac514973ec428d815d7a248d09d27f2a84f762aedfea63b3b48de9e633267ed5
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:658af3fc7819efb7cc39cd3a0504dcddc2ce7788c9b1e5178df869f7c9a742a8
        - registry.redhat.io/rhel8/postgresql-12@sha256:9248c4eaa8aeedacc1c06d7e3141ca1457147eef59e329273eb78e32fcd27e79
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:251c29e81f8e338f980b475b841647a977807908c388f4c86e18da7bff1b198a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:400a6ee03c3200c19b9b15c5d037a849633156a62e13922fb2e62838024ae69c
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:3d98d775bec85704c9a10c2f0d8e4f8116b7a41494f73c0b763e010262e5cd4d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        version: 7.11.1+0.1687357787.p
      entries:
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.11.x
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.12.x
    - currentCSV: fuse-online.v7.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta2",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:c70c5e76dc3db559c722bea2a767eb7dcba1eda786e1ef3a45c4cf32e1aca9ee
          createdAt: 2021-07-21 12:25:15 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta2
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to the product documentation at https://access.redhat.com/documentation/en-us/red_hat_fuse
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:d6df09b15599e776b49e1d46aa01728fd02d5629cee65e77dc277c0166554e0a
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:c668fe1f5d25baddf627acd79df4c8d2ab8b05309d3d41784ddc0a10dd5f34ad
        - registry.redhat.io/openshift4/ose-prometheus@sha256:13e9ea359cbcd3e0e94a6589b86b708dc84197f152701fd4092fffa234d11b58
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:60d5cf00411b03af3f1149a2bb3aeeb196c063d19625f876a094ae6baf46b54b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:f65600506db985689e5b7cd8b838e646d4268a0c4e272e19d7db3577a8dd7ac4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:75d5772c792a7c795281acec3d59e90021c109e9d7d1110868be0b882768aefb
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:c70c5e76dc3db559c722bea2a767eb7dcba1eda786e1ef3a45c4cf32e1aca9ee
        - registry.redhat.io/rhscl/postgresql-12-rhel7@sha256:c3e95e808eecba8ac57632452d3553509131ca783f62c68754a582cf5c55efab
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:eff9f8edf28ccbadcacab2022fff9e6bc93926af97521b49e3968eaf7b6fda65
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:d6f0c804c179fa559d3baed2e2d106c916a0439dc2215c4d34b878bc2e9bc281
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bafd85b0fd1256b8a60a2bb368a7e68f5dbad72718e5059a65660e01a425d8e2
        version: 7.9.3
      entries:
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.9.x
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: candidate
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: latest
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: stable
    defaultChannel: latest
    packageName: fuse-online
    provider:
      name: Red Hat
      url: https://access.redhat.com/products/red-hat-fuse
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: resource-locker-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: resource-locker-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "ResourceLocker",
                "metadata": {
                  "name": "combined"
                },
                "spec": {
                  "patches": [
                    {
                      "id": "patch1",
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 0).metadata.name }}: {{ (index . 1).metadata.name }}\n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "kind": "Namespace",
                          "name": "resource-locker-test"
                        },
                        {
                          "apiVersion": "v1",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "resource-locker-test"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "resource-locker-test"
                      }
                    },
                    {
                      "id": "patch2",
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 0) }}: {{ (index . 1) }}\n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "fieldPath": "$.metadata.uid",
                          "kind": "Namespace",
                          "name": "resource-locker-test"
                        },
                        {
                          "apiVersion": "v1",
                          "fieldPath": "$.metadata.uid",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "resource-locker-test"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "resource-locker-test"
                      }
                    }
                  ],
                  "resources": [
                    {
                      "object": {
                        "apiVersion": "networking.k8s.io/v1",
                        "kind": "NetworkPolicy",
                        "metadata": {
                          "name": "allow-from-same-namespace",
                          "namespace": "resource-locker-test"
                        },
                        "spec": {
                          "ingress": [
                            {
                              "from": [
                                {
                                  "podSelector": {}
                                }
                              ]
                            }
                          ],
                          "podSelector": {}
                        }
                      }
                    },
                    {
                      "object": {
                        "apiVersion": "networking.k8s.io/v1",
                        "kind": "NetworkPolicy",
                        "metadata": {
                          "name": "allow-from-default-namespace",
                          "namespace": "resource-locker-test"
                        },
                        "spec": {
                          "ingress": [
                            {
                              "from": [
                                {
                                  "namespaceSelector": {
                                    "matchLabels": {
                                      "name": "default"
                                    }
                                  }
                                }
                              ]
                            }
                          ],
                          "podSelector": {}
                        }
                      }
                    }
                  ],
                  "serviceAccountRef": {
                    "name": "default"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/resource-locker-operator@sha256:2176db02f561c24012c5179495c89c3819500dff21e89a62052d465eb3314b44
          createdAt: "2021-11-30T19:38:40Z"
          description: This operator provides a facility to lock resources and / or
            patches into the state described by its custom resource preventing any
            drifts.
          operatorframework.io/suggested-namespace: resource-locker-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/resource-locker-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ResourceLocker Represents the intetion to create and enforce
              a set of resources and/or patches.
            displayName: Resource Locker
            kind: ResourceLocker
            name: resourcelockers.redhatcop.redhat.io
            version: v1alpha1
        description: |
          The resource locker operator allows you to specify a set of configurations that the operator will "keep in place" (lock) preventing any drifts.
          Two types of configurations may be specified:
          * Resources. This will instruct the operator to create and enforce the specified resource. In this case the operator "owns" the created resources.
          * Patches to resources. This will instruct the operator to patch- and enforce the change on- a pre-existing resource. In this case the operator does not "own" the resource.
          Locked resources are defined with the `ResourceLocker` CRD. Here is the high-level structure of this CRD:
          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: ResourceLocker
          metadata:
            name: test-simple-resource
          spec:
            resources:
            - object:
                apiVersion: v1
            ...
            patches:
            - targetObjectRef:
              ...
              patchTemplate: |
                metadata:
                  annotations:
                    ciao: hello
            ...
            serviceAccountRef:
              name: default
          ```
          It contains:
          * `resources`: representing an array of resources
          * `patches`: representing an array of patches
          * `serviceAccountRef`: a reference to a service account defined in the same namespace as the ResourceLocker CR, that will be used to create the resources and apply the patches. If not specified the service account will be defaulted to: `default`
          For each ResourceLocker a manager is dynamically allocated. For each resource and patch a controller with the needed watches is created and associated with the previously created manager.
          **Note**: creating a `ResourceLocker` CR will not work without due preparation (of the service account in particular). Follow these [instructions](https://github.com/redhat-cop/resource-locker-operator/blob/master/EXAMPLES.md) for a step by step walkthrough of the needed prerequisites to make this operator work.
        displayName: Resource Locker Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - resource_management
        - enforcing
        - patches
        - gitops
        links:
        - name: Resource Locker Operator
          url: https://github.com/redhat-cop/resource-locker-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/resource-locker-operator:latest
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/resource-locker-operator@sha256:2176db02f561c24012c5179495c89c3819500dff21e89a62052d465eb3314b44
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.1.3
      entries:
      - name: resource-locker-operator.v1.1.3
        version: 1.1.3
      - name: resource-locker-operator.v1.1.2
        version: 1.1.2
      - name: resource-locker-operator.v1.1.1
        version: 1.1.1
      - name: resource-locker-operator.v1.1.0
        version: 1.1.0
      - name: resource-locker-operator.v1.0.9
        version: 1.0.9
      name: alpha
    defaultChannel: alpha
    packageName: resource-locker-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: APIMatic.io
      provider-url: https://www.apimatic.io/
    name: apimatic-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apimatic-kubernetes-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicodegen.apimatic.io/v1beta1",
                "kind": "APIMatic",
                "metadata": {
                  "name": "apimatic-sample"
                },
                "spec": {
                  "licensespec": {
                    "licenseSourceName": "apimaticlicense",
                    "licenseSourceType": "ConfigMap"
                  },
                  "podspec": {
                    "apimaticContainerSpec": {
                      "image": "apimaticio/apimatic-codegen",
                      "imagePullSecret": "apimaticimagesecret"
                    }
                  },
                  "replicas": 3,
                  "servicespec": {
                    "apimaticserviceport": {
                      "nodePort": 32000,
                      "port": 8070
                    },
                    "servicetype": "NodePort"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: docker.io/apimaticio/apimatic-kubernetes-operator:v0.0.1
          createdAt: "2021-08-24 12:00:00"
          description: Generate client SDKs and interactive Documentation for your
            APIs in minutes
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/apimatic/apimatic-kubernetes-operator
          support: APIMatic.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIMatic is the Schema for the apimatics API
            displayName: APIMatic App
            kind: APIMatic
            name: apimatics.apicodegen.apimatic.io
            version: v1beta1
        description: "##About APIMatic \n\n APIMatic is a Developer Experience   platform
          for Web APIs that empowers users to create multi-language SDKs with API
          reference docs, dynamic code samples, assisted authentication and more -
          all auto-generated to provide a comprehensive developer experience beyond
          any API documentation tool.\n\n## Table of contents\n\n* [Introduction](#introduction)\n*
          [Features](#features)\n* [Running the Sample](#running-the-sample)\n  *
          [Prerequisites](#prerequisites)\n  * [Steps for Direct Deployment](#steps-for-direct-deployment)\n
          \ * [Steps for OLM Deployment](#steps-for-olm-deployment)\n* [Technical
          Support](#technical-support)\n* [Copyrights](#copyrights)\n\n## Introduction\n\nAPIMatic
          Operator simplifies the configuration and lifecycle management of the APIMatic
          code and docs generation solution on different Kubernetes distributions
          and OpenShift. The Operator encapsulates key operational knowledge on how
          to configure and upgrade the APIMatic CodeGen application, making it easy
          to get up and running.\n\n\nMore information about the underlying APIMatic
          CodeGen API that is exposed\nby this operator can be found [here](https://apimatic-core-v3-docs.netlify.app/#/http/getting-started/overview-apimatic-core).\n\n##
          Features\n\nAPIMatic Operator provides the following features:\n- Deploys
          the APIMatic CodeGen Web API service within the Kubernetes or OpenShift
          cluster.\n- Exposes the APIMatic CodeGen API external to the cluster, using
          Service type as [NodePort](https://kubernetes.io/docs/concepts/services-networking/service/#nodeport),
          [LoadBalancer](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          or [ExternalName](https://kubernetes.io/docs/concepts/services-networking/service/#externalname).\n-
          For exposing the service through an ingress resource, create an Ingress
          resource in the namespace of your APIMatic CR and set owned APIMatic service
          created by the operator as a backed service. More information can be found
          [here](https://kubernetes.io/docs/concepts/services-networking/ingress/).\n-
          Manual horizontal scaling of pods.\n  ```sh\n  kubectl scale apm apimatic-sample--replicas=2\n
          \ ```\n\n## Running the Sample \n\n### Prerequisites\n\nPlease contact APIMatic
          at [support@apimatic.io](mailto:support@apimatic.io) to register with the
          APIMatic CodeGen Docker Hub registry and acquire a valid license to run
          the APIMatic CodeGen API.\n\nFurther prerequisites for running the sample
          include:\n\n- [go](https://golang.org/) v1.16.*\n- [git](https://git-scm.com/)\n-
          [make](https://www.gnu.org/software/make/)\n- [Operator SDK](https://sdk.operatorframework.io/docs/overview/)\n-
          A running Kubernetes cluster with [kubectl](https://kubernetes.io/docs/reference/kubectl/overview/)
          on client. For testing purposes, you can use [Minikube](https://minikube.sigs.k8s.io/docs/)
          or [kind](https://kind.sigs.k8s.io/)\n- For checking the service created
          by the APIMatic operator on-prem, you can use [MetalLB](https://metallb.org/)\n\n###
          Steps for Direct Deployment\n\nTo run the sample for checking the APIMatic
          operator:\n\n- Clone the APIMatic repository into your working directory
          using the following command:\n  ```sh\n  git clone https://github.com/apimatic/apimatic-kubernetes-operator.git\n
          \ ```\n- Run `make deploy` to set up the APIMatic operator resources. This
          will deploy the '`apimatic-system`' namespace as well as the CRD and the
          RBAC manifests.\n\n- Create a secret named '`apimaticimagesecret`' to allow
          pulling the APIMatic CodeGen image using the Docker Hub username registered
          with APIMatic. If you have not done so, please contact support@apimatic.io
          for the steps required.\n  ```sh\n  kubectl create secret docker-registry
          apimaticimagesecret --docker-server=https://index.docker.io/v1/ --docker-username=<your-name>
          --docker-password=<your-pword> --docker-email=<your-email>\n  ```\n- Create
          a [configmap](https://kubernetes.io/docs/concepts/configuration/configmap/)
          resource named '`apimaticlicense`' that will serve as the volume storing
          the APIMatic license. An example of this is given below which will create
          the ConfigMap using the *License.lic* file located at */usr/local/apimatic/license/*\n
          \ ```sh\n  kubectl create configmap apimaticlicense --from-file /usr/local/apimatic/license/License.lic\n
          \ ```\n- This will deploy a ConfigMap resource with the following definition:\n
          \ ```sh\n  apiVersion: v1\n  data:\n    License.lic: \\\"<License file contents
          here>\\\"\n  kind: ConfigMap\n  metadata:\n    name: apimaticlicense\n    namespace:
          default\n  ```\n- Now run the sample using the following command:\n  ```sh\n
          \ kubectl apply -f config/samples/apicodegen_v1beta1_apimatic.yaml\n  ```\n-
          You will now see a new [Deployment](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/)
          with replica count of 3 and [Service](https://kubernetes.io/docs/concepts/services-networking/service/)
          of type `NodePort` created, both named ***apimatic-sample***. Accessing
          http://localhost:32000 from your browser (or using curl from within the
          Minikube or Kind cluster, if using that) should now show the APIMatic Titan
          page.\n- You can now use the exposed APIMatic CodeGen API to generate API
          SDKs and Docs using [curl](https://curl.se/), [Postman](https://www.postman.com/)
          or your own custom Web application that consumes the APIMatic CodeGen API
          service.\n\n- Once done, you can remove the APIMatic resources using the
          following command:\n  ```sh\n  make undeploy\n  ```\n\n### Steps for OLM
          Deployment\n\nThe following steps can be used to utilize [Operator LifeCycle
          Manager (OLM)](https://olm.operatorframework.io/docs/) to deploy the operator
          and run the sample. The steps are as follows:\n\n- If not already done so,
          clone the APIMatic repository into your working directory:\n  ```sh\n  git
          clone https://github.com/apimatic/apimatic-kubernetes-operator.git\n  ```
          \n- [Install OLM in your Kubernetes cluster](https://olm.operatorframework.io/docs/getting-started/#installing-olm-in-your-cluster).\n\n-
          Run the following script to install the resources required by OLM to deploy
          the APIMatic operator in the Kubernetes cluster within the `apimatic-system`
          namespace. Information about the different resources required can be found
          using the steps given [here](https://olm.operatorframework.io/docs/tasks/).\n
          \ ```sh\n  kubectl apply -f olm/manifests.yaml\n  ```\n\n- This should spin
          up the `ClusterServiceVersion` of the operator in the `apimatic-system`
          namespace, following which the operator pod will spin up. To ensure the
          that the operator installed successfully, check for the `ClusterServiceVersion`
          and the operator deployment in the namespace it was installed in.\n  ```sh\n
          \ kubectl get csv -n apimatic-system\n  \n  kubectl get deployment -n apimatic-system\n
          \ ```\n\n  - Create a secret named `apimaticimagesecret` to allow pulling
          the APIMatic CodeGen image using the Docker Hub username registered with
          APIMatic. If you have not done so, please contact support@apimatic.io for
          the steps required.\n  ```sh\n  kubectl create secret docker-registry apimaticimagesecret
          --docker-server=https://index.docker.io/v1/ --docker-username=<your-name>
          --docker-password=<your-pword> --docker-email=<your-email>\n  ```\n- Create
          a [configmap](https://kubernetes.io/docs/concepts/configuration/configmap/)
          resource named `apimaticlicense` that will serve as the volume storing the
          APIMatic license. An example of this is given below which will create the
          ConfigMap using the *License.lic* file located at */usr/local/apimatic/license/*\n
          \ ```sh\n  kubectl create configmap apimaticlicense --from-file /usr/local/apimatic/license/License.lic\n
          \ ```\n- This will deploy a ConfigMap resource with the following definition:\n
          \ ```sh\n  apiVersion: v1\n  data:\n    License.lic: \\\"<License file contents
          here>\\\"\n  kind: ConfigMap\n  metadata:\n    name: apimaticlicense\n    namespace:
          default\n  ```\n- Now run the sample using the following command:\n  ```sh\n
          \ kubectl apply -f config/samples/apicodegen_v1beta1_apimatic.yaml\n  ```\n\n-
          You will now see a new [Deployment](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/)
          with replica count of 3 and [Service](https://kubernetes.io/docs/concepts/services-networking/service/)
          of type `NodePort` created, both named ***apimatic-sample***. Accessing
          http://localhost:32000 from your browser (or using curl from within the
          Minikube or Kind cluster, if using that) should now show the APIMatic Titan
          page.\n- You can now use the exposed APIMatic CodeGen API to generate API
          SDKs and Docs using [curl](https://curl.se/), [Postman](https://www.postman.com/)
          or your own custom Web application that consumes the APIMatic CodeGen API
          service.\n- Once done, you can remove the APIMatic operator resources using
          the follow script:\n  ```sh\n  kubectl delete -f olm/manifests.yaml\n  ```\n\n##
          Technical Support\n\n- To request additional features in the future, or
          if you notice any discrepancy regarding this document, please drop an email
          to [support@apimatic.io](mailto:support@apimatic.io).\n\n### Copyrights\n\n&copy;
          2021 APIMatic.io\n"
        displayName: APIMatic Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Rest APIs
        - API SDK Generator
        - API Docs Generator
        - API Developer Portal
        - APIMatic
        - Developer Tools
        links:
        - name: Overview
          url: https://www.apimatic.io/
        - name: Developer Experience Portal
          url: https://www.apimatic.io/developer-experience-portal
        - name: Code-Gen as a Service
          url: https://www.apimatic.io/continuous-code-generation
        - name: API Transformer
          url: https://www.apimatic.io/transformer
        - name: Contact
          url: https://www.apimatic.io/contact
        - name: Documentation
          url: https://docs.apimatic.io/
        - name: Blogs
          url: https://www.apimatic.io/blog/
        maintainers:
        - email: support@apimatic.io
          name: Support
        maturity: alpha
        provider:
          name: APIMatic.io
          url: https://www.apimatic.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - docker.io/apimaticio/apimatic-kubernetes-operator:v0.0.1
        version: 0.0.1
      entries:
      - name: apimatic-kubernetes-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: apimatic-kubernetes-operator
    provider:
      name: APIMatic.io
      url: https://www.apimatic.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-openshift-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.14
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-06-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:154f85cc6721a70909dee755ddd88897ad51038afe2538b6dee4d26e7d8b9afb
        version: 0.0.14
      entries:
      - name: joget-dx8-openshift-operator.v0.0.14
        version: 0.0.14
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-openshift-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sumo Logic
      provider-url: https://www.sumologic.com/
    name: sumologic-kubernetes-collection-helm-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "helm-operator.sumologic.com/v1alpha1",
                    "kind": "SumologicCollection",
                    "metadata": {
                        "name": "collection"
                    },
                    "spec": {
                        "sumologic": {
                            "accessId": "",
                            "accessKey": "",
                            "clusterName": "kubernetes",
                            "scc": {
                                "create": true
                            }
                        },
                        "kube-prometheus-stack": {
                            "prometheus-node-exporter": {
                                "service": {
                                    "port": 9200,
                                    "targetPort": 9200
                                }
                            }
                        },
                        "otellogs": {
                            "daemonset": {
                                "containers": {
                                    "otelcol": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                },
                                "initContainers": {
                                    "changeowner": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                }
                            }
                        }
                    }
                }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing,Monitoring
          containerImage: registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
          createdAt: "2024-07-19T10:48:00Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/sumologic-kubernetes-collection-helm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/sumologic-kubernetes-collection-helm-operator-rhmp/support?utm_source=openshift_console
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator.git
          support: Sumo Logic
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configuration for Sumo Logic Kubernetes Collection
            displayName: SumologicCollection
            kind: SumologicCollection
            name: sumologiccollections.helm-operator.sumologic.com
            version: v1alpha1
        description: Helm Operator for the Sumo Logic Kubernetes Collection Helm Chart
          v4.13.0-2
        displayName: Sumo Logic Kubernetes Collection Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - collection
        - logging
        - metrics
        - monitoring
        links:
        - name: Sumo Logic Kubernetes Collection Helm Operator
          url: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator
        maintainers:
        - email: collection@sumologic.com
          name: Sumo Logic
        maturity: alpha
        provider:
          name: Sumo Logic
          url: https://www.sumologic.com/
        relatedImages:
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
        - registry.connect.redhat.com/sumologic/kube-state-metrics@sha256:3551ededa327a35db798ec8286cd843580e3ae395d2231e041e8c4b7621127ee
        - registry.connect.redhat.com/sumologic/metrics-server@sha256:65d2a3ddc4b4183f438fa18b655079c93c905cb9e3b8c54977336205dc1fff07
        - registry.connect.redhat.com/sumologic/prometheus@sha256:0ef6eaf570a40499dd5de92a79593e950bc771b014e0ee6ff7419b4b82a36fcc
        - registry.connect.redhat.com/sumologic/autoinstrumentation-python@sha256:f0fa45fa45e64c7df88f21181ae18ee96fe83cd3988258663d0806ea6f39e6bb
        - registry.connect.redhat.com/sumologic/kubernetes-setup@sha256:f07eaf68fa11ffc708e4485a83bd3763fc5712f7d73d29edfe5b9c9e08f05df6
        - registry.connect.redhat.com/sumologic/sumologic-otel-collector@sha256:4554d43dba204e3ec8e91b36442a438f10234fa0cb4db7fb7880b257a6485112
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:e1f49fcf5b7e86d9eb49c16dbbf9484cb486f55f999fd51d7c738c3b1f2666cf
        - registry.connect.redhat.com/sumologic/autoinstrumentation-dotnet@sha256:f4f385c4489c74b5c43f2fe6b6620a9eb8d9ecab470567236b3c45f636106640
        - registry.connect.redhat.com/sumologic/tailing-sidecar@sha256:738d653a78dc415243ab3627db9135768175fb7f51d0803566f5d9e824c8744b
        - registry.connect.redhat.com/sumologic/telegraf-operator@sha256:ff958222be01c0c3966474463e43bbd5bcf348790dd34bb761718cb1cea2d793
        - registry.connect.redhat.com/sumologic/autoinstrumentation-nodejs@sha256:c206eef4fdecaa58b4f01e5e7f0dbd74e9606eab2b4debf47461d36191e248d1
        - registry.connect.redhat.com/sumologic/node-exporter@sha256:11e9e604432a66d9b388cb1bf70eac077c7f0372d1264342ef927d76b8ef0272
        - registry.connect.redhat.com/sumologic/kubernetes-tools-kubectl@sha256:d3155b864cc74f327adedf7d013a8f0fcb11c2e73a5655b7cafa848824e80172
        - registry.connect.redhat.com/sumologic/thanos@sha256:16faf7191668e86d8b576bba055dd0cbdeb009635976b9e07bd00f926032f11c
        - registry.connect.redhat.com/sumologic/opentelemetry-operator@sha256:5587fe5b7ebfa57ad320ad2521103f908317388d1741c40e5f7be5ed9e5f9ad7
        - registry.connect.redhat.com/sumologic/nginx-unprivileged@sha256:7260555d16c94c4fe1e016626b4cb8a5c0de6b700e2eea28744331fb298f71ae
        - registry.connect.redhat.com/sumologic/prometheus-operator@sha256:58a329e6582456b32abebb470086191f8366b927af4703741613993465f116b8
        - registry.connect.redhat.com/sumologic/telegraf@sha256:75efd2e509c6017ea0fc5c9103ce068cbc89087b16506a4ef48cb5ff00eab84e
        - registry.connect.redhat.com/sumologic/tailing-sidecar-operator@sha256:c16b8b18e9b263fc3eb24059bba512bcfc3201b32105af6e5518392826ddd66c
        - registry.connect.redhat.com/sumologic/prometheus-config-reloader@sha256:0b605d36782afd8f50fd631d0e9c7b19d39ce962af89b7a737b2f09a8ceff291
        - registry.connect.redhat.com/sumologic/autoinstrumentation-java@sha256:242958b7d8905f4ac02f6ad38897ba58b9badf2a6213a389e67dcc71fee3630c
        - registry.connect.redhat.com/sumologic/busybox@sha256:209bc8e0b1d958fc699f3bb75b1248544e372ad35cdc46c991bd1698f641e1f9
        version: 4.13.0-2
      entries:
      - name: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
        version: 4.13.0-2
      name: alpha
    defaultChannel: alpha
    packageName: sumologic-kubernetes-collection-helm-operator-rhmp
    provider:
      name: Sumo Logic
      url: https://www.sumologic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-bamoe-businessautomation: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: bamoe-businessautomation-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bamoe-businessautomation-operator.8.0.8-1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.kiegroup.org/v2","kind":"KieApp","metadata":{"name":"rhpam-trial"},"spec":{"environment":"rhpam-trial"}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/ibm-bamoe/bamoe-rhel8-operator@sha256:0646da8707d10b3cc2e2344269abd8797db34127a3326bf76bbc1547aa4bbd73
          createdAt: "2025-07-22 17:40:17"
          description: Deploys and manages IBM Business Automation Manager Open Editions
            environment.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM Business Automation Manager
            Open Edition"]'
          repository: https://github.com/kiegroup/kie-cloud-operator
          support: https://www.ibm.com/mysupport
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A project prescription running an IBM BAMOE environment.
            displayName: KieApp
            kind: KieApp
            name: kieapps.app.kiegroup.org
            version: v2
        description: |-
          Deploys and manages IBM Business Automation Manager Open Editions environment.

          * **IBM Process Automation Manager** is a platform for developing containerized microservices and applications that automate business decisions and processes. It includes business process management (BPM), business rules management (BRM), and business resource optimization and complex event processing (CEP) technologies. It also includes a user experience platform to create engaging user interfaces for process and decision services with minimal coding.
        displayName: IBM Business Automation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kieapp
        - pam
        - decision
        - kie
        - cloud
        - bpm
        - process
        - automation
        - operator
        links:
        - name: Product Page
          url: https://ibm.com
        - name: Documentation
          url: https://www.ibm.com/docs/en/ibamoe
        maintainers:
        - email: mysphelp@us.ibm.com
          name: IBM
        maturity: 8.x-stable
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-monitoring-rhel8@sha256:ef1d77bb085172f927ee53d1674c85e2acaf1f580d7cad55bae8bf2c33c49c01
        - registry.redhat.io/openshift4/ose-cli@sha256:b2b93def4eb5edeccfa2d8ba22e493646325842674dff0446b45677bf641a342
        - registry.redhat.io/rhscl/postgresql-13-rhel7@sha256:338dd09681dcf860e6a976a3bc48a1474c2376d698815e4eb611ae9c5aa68966
        - registry.redhat.io/ibm-bamoe/bamoe-process-migration-rhel8@sha256:3fd45f2dabca7220772d823275a5697b42dd35f53c27671b093117933ef2d4f6
        - registry.redhat.io/rhscl/mysql-80-rhel7@sha256:e4f535c78bc2fdb58914c52fb805185035ff31515452b4df714f9ab623ba383a
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-rhel8@sha256:c3553707936cb9eee8a6a622643f72c91e1b0c11add5ec9ad5865242a9e133b4
        - registry.redhat.io/ibm-bamoe/bamoe-kieserver-rhel8@sha256:01bf97beffb3647968aae052cdd401d6d562c35f3e78ac483dbf55c44a71dc2f
        - registry.redhat.io/jboss-datagrid-7/datagrid73-openshift@sha256:4056b1cf8ff5995f56eecfdc087d9ece56a216dba6bd73acc5693f0778e8d573
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-rhel8@sha256:82d8ebde7eb3ecfbf72e5b71e7e392d3a3597fc084b6cf35e8fec23bfe41210d
        - registry.redhat.io/amq7/amq-broker@sha256:792c1cb19e62519cc47d834650fe5a6a6ca589a37d0b53d4f4549e80cabae3f9
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:98e9a6618558bc8a1b9cb91d1c9b3380968d335092f6e953d1abe6af96abf65c
        - registry.redhat.io/ibm-bamoe/bamoe-smartrouter-rhel8@sha256:2f1a7519617b5c0d082f943de0db295370649d968d699d76cfd829f4c9826a0f
        - registry.redhat.io/ibm-bamoe/bamoe-rhel8-operator@sha256:0646da8707d10b3cc2e2344269abd8797db34127a3326bf76bbc1547aa4bbd73
        - registry.redhat.io/ibm-bamoe/bamoe-process-migration-rhel8@sha256:a287a2fbaded371b8fcae5e4124b2d905bb228f94da0fc899ca2600b378dca0d
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-monitoring-rhel8@sha256:25025e489c15108edffe5270d257e1ed9be0a67aa067bcfc184647793aac2617
        - registry.redhat.io/ibm-bamoe/bamoe-dashbuilder-rhel8@sha256:efa55e17190b0ebe1618452f786876ef354b8fbec25c316aed7ed5bb1ce88116
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:457fc73423fa94591bc7e35729a25c58d3465583afccf342fb338b6e11188575
        - registry.redhat.io/ibm-bamoe/bamoe-smartrouter-rhel8@sha256:5ed8821328413ec11859ae317679f0692afff9945f134fc8909e07a3ac387794
        - registry.redhat.io/ibm-bamoe/bamoe-kieserver-rhel8@sha256:c9f08ddc6111cdd74faa0db5721d5730fc7716fccf4dc2a658595f7e0c184c10
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:f714b9c1555e7f27a3c9cf2cf687a74188f36b1283d4e09cec4db3c084d9fe64
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhscl/postgresql-12-rhel7@sha256:191237ccffc476e764714a69e35d9421dcee5ddd7184442d7a8306a06b97e6bb
        - registry.redhat.io/ibm-bamoe/bamoe-dashbuilder-rhel8@sha256:b3b7388b3b116ef37aabbb04b5d2c3e36e58d2fc95f1e59795b423010e071fc3
        - registry.redhat.io/ibm-bamoe/bamoe-controller-rhel8@sha256:3fcd9ae0bc194c198cbb5ee94272e80f302658733e467766bd1789357e9fd015
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:fb326be806b337b0ae1711f974d15e073aebcbad77aaa1123a63cf4f331da5f2
        - registry.redhat.io/ibm-bamoe/bamoe-controller-rhel8@sha256:6cd0f07856d612dca46b0a01a93b796f7a0f112b98cedfd6754f9fd770ee843d
        version: 8.0.8-1
      entries:
      - name: bamoe-businessautomation-operator.8.0.8-1
        version: 8.0.8-1
      - name: bamoe-businessautomation-operator.8.0.7-2
        version: 8.0.7-2
      - name: bamoe-businessautomation-operator.8.0.7-1
        version: 8.0.7-1
      - name: bamoe-businessautomation-operator.8.0.6-3
        version: 8.0.6-3
      - name: bamoe-businessautomation-operator.8.0.6-2
        version: 8.0.6-2
      - name: bamoe-businessautomation-operator.8.0.6-1
        version: 8.0.6-1
      - name: bamoe-businessautomation-operator.8.0.5-3
        version: 8.0.5-3
      - name: bamoe-businessautomation-operator.8.0.5-2
        version: 8.0.5-2
      - name: bamoe-businessautomation-operator.8.0.5-1
        version: 8.0.5-1
      - name: bamoe-businessautomation-operator.8.0.4-3
        version: 8.0.4-3
      - name: bamoe-businessautomation-operator.8.0.4-2
        version: 8.0.4-2
      - name: bamoe-businessautomation-operator.8.0.4-1
        version: 8.0.4-1
      - name: bamoe-businessautomation-operator.8.0.3-4
        version: 8.0.3-4
      - name: bamoe-businessautomation-operator.8.0.3-3
        version: 8.0.3-3
      - name: bamoe-businessautomation-operator.8.0.3-2
        version: 8.0.3-2
      - name: bamoe-businessautomation-operator.8.0.3-1
        version: 8.0.3-1
      - name: bamoe-businessautomation-operator.8.0.2-3
        version: 8.0.2-3
      - name: bamoe-businessautomation-operator.8.0.2-2
        version: 8.0.2-2
      - name: bamoe-businessautomation-operator.8.0.2-1
        version: 8.0.2-1
      - name: bamoe-businessautomation-operator.8.0.1-2
        version: 8.0.1-2
      - name: bamoe-businessautomation-operator.8.0.1-1
        version: 8.0.1-1
      - name: bamoe-businessautomation-operator.8.0.0-1
        version: 8.0.0-1
      name: 8.x-stable
    defaultChannel: 8.x-stable
    packageName: bamoe-businessautomation-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: eG Innovations
      provider-url: https://www.eginnovations.com
    name: eginnovations-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eginnovations-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eginnovations.com/v1beta1",
                "kind": "EgUniversalAgent",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eginnovations-operator"
                  },
                  "name": "eguniversalagent",
                  "namespace": "egagent"
                },
                "spec": {
                  "agent": {
                    "env": [
                      {
                        "name": "EG_MANAGER",
                        "value": "Replace with eG Manager IP"
                      },
                      {
                        "name": "EG_MANAGER_PORT",
                        "value": "Replace with eG Manager Port"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "true"
          containerImage: docker.io/eginnovations/universal-agent-operator:0.0.11
          createdAt: "2025-09-11T20:08:00Z"
          description: eG Innovations provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with eG Universal Agent Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eginnovations/universal-agent-operator
          support: eG Innovations
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgUniversalAgent is the Schema for the eguniversalagents
              API
            displayName: Eg Universal Agent
            kind: EgUniversalAgent
            name: eguniversalagents.eginnovations.com
            version: v1beta1
        description: |
          eG Innovations’ converged application and infrastructure performance monitoring capabilities provide end-to-end visibility and correlated analytics for every layer, every tier of your IT landscape.

          The eG Enterprise Universal Agent Operator automatically configures a host agent on every Kubernetes worker node. The host agent auto-discovers the worker nodes and application containers running as Pods on each node and tracks their performance and utilization levels. In-depth monitoring of applications running on containers is also provided using the same host agent. No additional agents are required for the containers.

          ### Capabilities
            * Rolls out the egagent pod per node to monitor its pods and the node itself
            * Automatic BTM profiler injection using webhook based injection for deep application monitoring and transection tracing.

          ### Installation
            1. Create the namespace "egagent" for installing eG Agent on OpenShift Cluster.

               <br />``` $ oc create ns egagent ```
            2. Add "egagent" serviceaccount to scc( Security Context Constraints ) of OpenShift Cluster as privileged using below command.

               <br />``` $ oc adm policy add-scc-to-user privileged -z egagent -n egagent ```
            2. Install eG Innovations Universal Agent operator from [OpenShift Container Platform](https://www.openshift.com/) OperatorHub.
            3. Create the eG Universal Agent CRD (Custom Resource Definition)
               a. Click "View YAML Example" and copy the yaml contents and save it to (filename).yaml on your host.
               b. To apply the CRD run the following command

               <br />``` $ oc apply -f (filename).yaml ```

          ### Required Parameters
            * EG\_MANAGER, EG\_MANAGER\_PORT - provide the ip and port of your eG Manager Installation.

          ### Advanced Options
            * Disable Certificate Checking - disable any certificate validation that may interact poorly with proxies within your cluster
            * Image Override - use a copy of the eG Iniversal Agent container image from a registry other that Docker's or Red Hat's
            * Environment variables - define environment variables for the eG Universal Agent container of for APM profiler.
            * Namespace Selectors- select a subset of namesapces to enable instrumentation for applications.

          For complete list of supported parameters and the detailed documentation available in [eG Innovations Docs](https://docs.eginnovations.com)
        displayName: eG Innovations Universal Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Tracing
        links:
        - name: eG Innovations Kubernetes Monitoring Overview
          url: https://www.eginnovations.com/documentation/Kubernetes/Introduction-to-Monitoring-Kubernetes.htm
        maintainers:
        - email: devops@eginnovations.com
          name: eG Innovations
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: eG Innovations
          url: https://www.eginnovations.com
        relatedImages:
        - docker.io/eginnovations/universal-agent-operator:0.0.11
        version: 0.0.11
      entries:
      - name: eginnovations-operator.v0.0.11
        version: 0.0.11
      - name: eginnovations-operator.v0.0.10
        version: 0.0.10
      - name: eginnovations-operator.v0.0.9
        version: 0.0.9
      - name: eginnovations-operator.v0.0.8
        version: 0.0.8
      name: beta
    defaultChannel: beta
    packageName: eginnovations-operator
    provider:
      name: eG Innovations
      url: https://www.eginnovations.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fortanix Inc.
      provider-url: https://fortanix.com
    name: ccm-node-agent-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ccm-node-agent-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1",
                "kind": "CcmNodeAgent",
                "metadata": {
                  "name": "ccmnodeagent-sample"
                },
                "spec": {
                  "ccmJoinToken": {
                    "secretName": "ccm-join-token",
                    "token": ""
                  },
                  "fullnameOverride": "",
                  "image": {
                    "ccmNodeAgent": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/ccm-node-agent-ubi7",
                      "tag": "latest"
                    },
                    "ccmNodeAgentInit": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "busybox",
                      "tag": "latest"
                    },
                    "sgxSoftwareInstaller": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/sgx-software-installer-ubi7",
                      "tag": "latest"
                    }
                  },
                  "log": {
                    "debug": false
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "ports": {
                    "containerPort": 9092,
                    "hostPort": 9092
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CCM Node Agent software enables registration of the compute
              nodes to Fortanix CCM when installed in the Openshift Cluster Nodes<U+200B>.
            displayName: CCM Node Agent
            kind: CcmNodeAgent
            name: ccmnodeagents.charts.operatorhub.io
            version: v1
        description: "CCM Node Agent software enables registration of the compute
          nodes to Fortanix CCM when installed in the Openshift Cluster Nodes​. \nIt
          also assists with application attestation and visibility for Fortanix CCM.\n\n##
          About this Operator\nThis Operator is based on a Helm chart for CCM Node
          Agent. It enrolls all the nodes in the openshift cluster to a particular
          CCM account. You have to provide the base64 encode of join token of CCM
          account while creating CcmNodeAgent resource.\n\n## Prerequisites\nAll the
          worker nodes in the cluster must have SGX capabilities.\n\n## Installation\n*
          Choose a namespace to install the Operator. Operator will be available in
          all namespaces.\n* Select Approval Strategy: Available options are Automatic
          or Manual. Automatic approval will automatically upgrade the running instance
          of your Operator without human intervention when a new version is available.
          In case of Manual approval, when a new version is available, OLM will create
          an update request. You must manually approve the request to update the Operator
          to a new version.\n* Generate your CCM Account Join Token: Please log in
          to https://ccm.fortanix.com, and in the Management Console tab, click the
          + ENROLL NODE button. In the ENROLL NODE screen, a Join Token will be generated
          in the text box for \"Get a join token to register an SGX compute node\".
          This Join Token is used by the compute node to authenticate itself.\n* Base64
          encode the Join Token:\n```shell\necho -n <join-token> | openssl base64
          -A\n```\n* Create an instance of CcmNodeAgent by setting the `spec.ccmJoinToken.Token`
          to the base64 encoded CCM account join token.\n* After CcmNodeAgent resource
          is installed successfully you should be able to see all the worker nodes
          of the cluster in your CCM account under Compute Node Tab.\n\n## Getting
          Started\nNow to run your application inside SGX enclaves on your cluster
          follow the [documentation](https://support.fortanix.com/hc/en-us/articles/360043527431-User-s-Guide-Add-and-Edit-an-Application).\n"
        displayName: CCM Node Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dockerfile
        - Helm
        - CCM
        - Confidential Compute
        links:
        - name: CCM Node Agent Operator
          url: https://support.fortanix.com/hc/en-us/categories/360003107511-Confidential-Computing-Manager
        maintainers:
        - email: info@fortanix.com
          name: Fortanix Inc.
        maturity: alpha
        provider:
          name: Fortanix Inc.
          url: https://fortanix.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62575404000cb77fa8d2284c87b31e39e667a55b0fb47821263b23b3d1135a7e
        - docker.io/fortanix/ccm-node-agent-operator@sha256:e2a3d52551198102ddbe7f4ebd9621edc357c57ddeff47b7ed9f2cf56a1bed39
        version: 0.0.2
      entries:
      - name: ccm-node-agent-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ccm-node-agent-operator
    provider:
      name: Fortanix Inc.
      url: https://fortanix.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: https://intel.com
    name: intel-device-plugins-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: intel-device-plugins-operator.v0.29.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "GpuDevicePlugin",
                "metadata": {
                  "name": "gpudeviceplugin-sample"
                },
                "spec": {
                  "enableMonitoring": true,
                  "image": "registry.connect.redhat.com/intel/intel-gpu-plugin@sha256:963caf1e6e16b5e297671c536eec04dc57cf45f5cc5cf458bdcafa56bdb97972",
                  "preferredAllocationPolicy": "none",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/gpu": "true"
                  },
                  "sharedDevNum": 1
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "QatDevicePlugin",
                "metadata": {
                  "name": "qatdeviceplugin-sample"
                },
                "spec": {
                  "dpdkDriver": "vfio-pci",
                  "image": "registry.connect.redhat.com/intel/intel-qat-plugin@sha256:53617c3a72af169031cbaba1380cd6a25c9d2b50c9a2da822cf37b473c6e7a8f",
                  "initImage": "registry.connect.redhat.com/intel/intel-qat-initcontainer@sha256:14f282a0adae43262657f24ce99a620b671c9a2ce12a8216163e6b22ac6e252b",
                  "kernelVfDrivers": [
                    "c6xxvf",
                    "4xxxvf"
                  ],
                  "logLevel": 4,
                  "maxNumDevices": 128,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/qat": "true"
                  }
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "DsaDevicePlugin",
                "metadata": {
                  "name": "dsadeviceplugin-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/intel/intel-dsa-plugin@sha256:092575515739bf9a0f5982c3605b63382bc4d1250128269434c2bfe4a27ec99d",
                  "initImage": "registry.connect.redhat.com/intel/intel-idxd-config-initcontainer@sha256:649e4f4e7f362ba054fb85a67e3baa23ec9649c68df356d1e33721ca3eef2f6b",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/dsa": "true"
                  },
                  "sharedDevNum": 10
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "SgxDevicePlugin",
                "metadata": {
                  "name": "sgxdeviceplugin-sample"
                },
                "spec": {
                  "enclaveLimit": 110,
                  "image": "registry.connect.redhat.com/intel/intel-sgx-plugin@sha256:e54cdf61feb0b67e330ad24d92e7fa4000953c2ec4d4300d052e032d419d5e0a",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/sgx": "true"
                  },
                  "provisionLimit": 110
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Security, AI/Machine Learning
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/intel/intel-technology-enabling-for-openshift
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GpuDevicePlugin is the Schema for the gpudeviceplugins API.
              It represents the GPU device plugin responsible for advertising Intel
              GPU hardware resources to the kubelet.
            displayName: Intel GPU Device Plugin
            kind: GpuDevicePlugin
            name: gpudeviceplugins.deviceplugin.intel.com
            version: v1
          - description: DsaDevicePlugin is the Schema for the dsadeviceplugins API.
              It represents the DSA device plugin responsible for advertising Intel
              DSA hardware resources to the kubelet.
            displayName: Intel DSA Device Plugin
            kind: DsaDevicePlugin
            name: dsadeviceplugins.deviceplugin.intel.com
            version: v1
          - description: QatDevicePlugin is the Schema for the qatdeviceplugins API.
              It represents the QAT device plugin responsible for advertising Intel
              QuickAssist Technology hardware resources to the kubelet.
            displayName: Intel QuickAssist Technology Device Plugin
            kind: QatDevicePlugin
            name: qatdeviceplugins.deviceplugin.intel.com
            version: v1
          - description: SgxDevicePlugin is the Schema for the sgxdeviceplugins API.
              It represents the SGX device plugin responsible for advertising SGX
              device nodes to the kubelet.
            displayName: Intel Software Guard Extensions Device Plugin
            kind: SgxDevicePlugin
            name: sgxdeviceplugins.deviceplugin.intel.com
            version: v1
        description: |
          [Intel Device Plugins for Kubernetes](https://github.com/intel/intel-technology-enabling-for-openshift) is a collection of
          [device plugins](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/) advertising Intel specific hardware resources
          to the kubelet. Currently the operator has basic support for Intel SGX, Intel GPU, Intel QAT and Intel DSA. Intel IAA device plugin will be supported in a future release.
        displayName: Intel Device Plugins Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - qat
        - sgx
        - gpu
        - dsa
        - dlb
        links:
        - name: Intel Device Plugins Operator
          url: https://github.com/intel/intel-technology-enabling-for-openshift
        maintainers:
        - email: martin.xu@intel.com
          name: Martin Xu
        - email: chaitanya.kulkarni@intel.com
          name: Chaitanya Kulkarni
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Intel® Corporation
          url: https://intel.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - registry.connect.redhat.com/intel/intel-sgx-plugin@sha256:e54cdf61feb0b67e330ad24d92e7fa4000953c2ec4d4300d052e032d419d5e0a
        - registry.connect.redhat.com/intel/intel-gpu-plugin@sha256:963caf1e6e16b5e297671c536eec04dc57cf45f5cc5cf458bdcafa56bdb97972
        - registry.connect.redhat.com/intel/intel-qat-plugin@sha256:53617c3a72af169031cbaba1380cd6a25c9d2b50c9a2da822cf37b473c6e7a8f
        - registry.connect.redhat.com/intel/intel-qat-initcontainer@sha256:14f282a0adae43262657f24ce99a620b671c9a2ce12a8216163e6b22ac6e252b
        - registry.connect.redhat.com/intel/intel-dsa-plugin@sha256:092575515739bf9a0f5982c3605b63382bc4d1250128269434c2bfe4a27ec99d
        - registry.connect.redhat.com/intel/intel-idxd-initcontainer@sha256:649e4f4e7f362ba054fb85a67e3baa23ec9649c68df356d1e33721ca3eef2f6b
        - registry.connect.redhat.com/intel/intel-deviceplugin-operator@sha256:283e8b60d7ab8560f8ec8875b905e6446b59a91dd5e0d97da72e3a9aab235cc4
        version: 0.29.1
      entries:
      - name: intel-device-plugins-operator.v0.29.1
        version: 0.29.1
      - name: intel-device-plugins-operator.v0.29.0
        version: 0.29.0
      - name: intel-device-plugins-operator.v0.28.0
        version: 0.28.0
      name: alpha
    defaultChannel: alpha
    packageName: intel-device-plugins-operator
    provider:
      name: Intel® Corporation
      url: https://intel.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: designate-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: designate-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "designate.openstack.org/v1beta1",
                "kind": "Designate",
                "metadata": {
                  "name": "designate"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "designate",
                  "databaseInstance": "openstack",
                  "designateAPI": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate"
                  },
                  "designateBackendbind9": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "networkAttachments": [
                      "designate"
                    ],
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate",
                    "storageClass": "local-storage",
                    "storageRequest": "1G"
                  },
                  "designateCentral": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate"
                  },
                  "designateMdns": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "networkAttachments": [
                      "designate"
                    ],
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate"
                  },
                  "designateProducer": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate"
                  },
                  "designateUnbound": {
                    "replicas": 1
                  },
                  "designateWorker": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "networkAttachments": [
                      "designate"
                    ],
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate"
                  },
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "osp-secret",
                  "serviceUser": "designate"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:48Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DesignateAPI is the Schema for the designateapis API
            displayName: Designate API
            kind: DesignateAPI
            name: designateapis.designate.openstack.org
            version: v1beta1
          - kind: DesignateBackendbind9
            name: designatebackendbind9s.designate.openstack.org
            version: v1beta1
          - description: DesignateCentral is the Schema for the designatecentral API
            displayName: Designate Central
            kind: DesignateCentral
            name: designatecentrals.designate.openstack.org
            version: v1beta1
          - description: DesignateMdns is the Schema for the designatemdnses API
            displayName: Designate Mdns
            kind: DesignateMdns
            name: designatemdnses.designate.openstack.org
            version: v1beta1
          - description: DesignateProducer is the Schema for the designateproducer
              API
            displayName: Designate Producer
            kind: DesignateProducer
            name: designateproducers.designate.openstack.org
            version: v1beta1
          - description: Designate is the Schema for the designates API
            displayName: Designate
            kind: Designate
            name: designates.designate.openstack.org
            version: v1beta1
          - description: DesignateUnbound is the Schema for the designateunbound API
            displayName: Designate Unbound
            kind: DesignateUnbound
            name: designateunbounds.designate.openstack.org
            version: v1beta1
          - description: DesignateWorker is the Schema for the designateworker API
            displayName: Designate Worker
            kind: DesignateWorker
            name: designateworkers.designate.openstack.org
            version: v1beta1
        description: Designate Operator
        displayName: Designate Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Network
        - Designate
        links:
        - name: Designate Operator
          url: https://github.com/openstack-k8s-operators/designate-operator
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso/openstack-designate-api-rhel9@sha256:1a9baa4a0f99029eb8a367ddcce264581aa37fdf1834ba277c30a3f0ae33f400
        - registry.redhat.io/rhoso/openstack-designate-central-rhel9@sha256:ad610cdc29b31cd82a81cea22793176351574f43c0f7705273c8ffc03ad2fdac
        - registry.redhat.io/rhoso/openstack-designate-producer-rhel9@sha256:4beb652a073a65965f96f125d0fcb454255efa5e53a84bd0dd9120b57e59fcf3
        - registry.redhat.io/rhoso-operators/designate-rhel9-operator@sha256:e6c1ba511cfd0278fc8cd2836c6324d97ae3f6e9affc992600978062fa2d4549
        - registry.redhat.io/rhoso/openstack-designate-mdns-rhel9@sha256:a4109755360f340aadf6bc15d24599f882d82bd59739065f87275f9274d9eeb9
        - registry.redhat.io/rhoso/openstack-designate-worker-rhel9@sha256:94ef291ca85b06dffc93d149b09b412d9190c99f7a1b0b300d50bd70346401ad
        - registry.redhat.io/rhoso/openstack-designate-backend-bind9-rhel9@sha256:89ce30fc1eff339867ce422203e6e0c63296667e102ce85d823d44c30f7556ad
        - registry.redhat.io/rhoso/openstack-unbound-rhel9@sha256:ad95d40227a1e52f739f09a2ce8db647d6a2fa7ef63e58a2444688728fa5bdfd
        version: 1.0.6
      entries:
      - name: designate-operator.v1.0.6
        version: 1.0.6
      - name: designate-operator.v1.0.5
        version: 1.0.5
      - name: designate-operator.v1.0.4
        version: 1.0.4
      - name: designate-operator.v1.0.3
        version: 1.0.3
      - name: designate-operator.v1.0.2
        version: 1.0.2
      - name: designate-operator.v1.0.1
        version: 1.0.1
      - name: designate-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: designate-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-cert-manager-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-manager-operator.v1.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
          createdAt: 2025-08-06T08:21:40
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.16.1 <1.17.0'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.17.4](https://github.com/cert-manager/cert-manager/tree/v1.17.4), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:96d51e3a64bf30cbd92836c7cbd82f06edca16eef78ab1432757d34c16628659
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b
        version: 1.17.0
      entries:
      - name: cert-manager-operator.v1.17.0
        version: 1.17.0
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      - name: cert-manager-operator.v1.14.1
        version: 1.14.1
      - name: cert-manager-operator.v1.14.0
        version: 1.14.0
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1
    - currentCSV: cert-manager-operator.v1.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "\u003cACCESS_KEY_ID\u003e",
                            "hostedZoneID": "\u003cHOSTED_ZONE_ID\u003e",
                            "region": "\u003cAWS_REGION\u003e",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4e854c14fecbd27107b91b76164ce2e1575ce60202d835abba5962fb5d386560
          createdAt: 2024-04-30T13:29:33
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.13.0 <1.13.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/infrastructure-features: '["proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.13.6](https://github.com/cert-manager/cert-manager/tree/v1.13.6), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:afc82895c09d75e3d6a932aea2d3e3ef998f765306da92070de1dab26baf0593
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4e854c14fecbd27107b91b76164ce2e1575ce60202d835abba5962fb5d386560
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:578f74c3e96db7fc2f4e94dd9fbc7d6499d87b0bc55576930ee17de08a95022c
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:7915f8049d2451b425e3674eb60e596b9b11946ffda1497f5d5607799e2bf431
        version: 1.13.1
      entries:
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1.13
    - currentCSV: cert-manager-operator.v1.14.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:5c593bfd70309f87cd05716e233e65aa55394eafcdaf85f178698912cb29d79d
          createdAt: 2025-04-09T05:19:10
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.14.1 <1.14.2'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/infrastructure-features: '["proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.14.7](https://github.com/cert-manager/cert-manager/tree/v1.14.7), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:5c593bfd70309f87cd05716e233e65aa55394eafcdaf85f178698912cb29d79d
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:5df082f8e3787e6ac2edf395113806999ae192425ce18b4ddba9ab20c135fa5b
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:49fb8f710319d6f3479b98d2c5dd7d0c9df8c1678b2c00727314dea1a8e933a1
        version: 1.14.2
      entries:
      - name: cert-manager-operator.v1.14.2
        version: 1.14.2
      - name: cert-manager-operator.v1.14.1
        version: 1.14.1
      - name: cert-manager-operator.v1.14.0
        version: 1.14.0
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1.14
    - currentCSV: cert-manager-operator.v1.15.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:f328263e2d29e34ede65e4501f0447b2d9f84e9445a365c2fa2fbb253939e274
          createdAt: 2025-03-13T18:26:37
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.15.0 <1.15.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/infrastructure-features: '["proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.15.5](https://github.com/cert-manager/cert-manager/tree/v1.15.2), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:f328263e2d29e34ede65e4501f0447b2d9f84e9445a365c2fa2fbb253939e274
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9573d74bd2b926ec94af76f813e6358f14c5b2f4e0eedab7c1ff1070b7279a5c
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:6af3ee8b2a5a87042fb7158bda8d6cf2e6324d1e265974acf77214d4cd0ea0d3
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:20efff60a0caf5eafb38986fd21611697b5bc534c2e789da233983a9739938ed
        version: 1.15.1
      entries:
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      - name: cert-manager-operator.v1.14.1
        version: 1.14.1
      - name: cert-manager-operator.v1.14.0
        version: 1.14.0
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1.15
    - currentCSV: cert-manager-operator.v1.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:d30e755a266bb035d50e2b257e7a26c1fe6a5ccb66125576db99d537aa03ec27
          createdAt: 2025-07-10T08:52:43
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.15.1 <1.16.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.16.5](https://github.com/cert-manager/cert-manager/tree/v1.16.5), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:831e4b1e5aa6eed6fbb22f779ddbcb9c73e4bf6cc079d0c6652737a6a44e6320
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:d30e755a266bb035d50e2b257e7a26c1fe6a5ccb66125576db99d537aa03ec27
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:33308217d2226c2555eb18020710d16175295a8e349263a5435e4b158bbc735f
        version: 1.16.1
      entries:
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      - name: cert-manager-operator.v1.14.1
        version: 1.14.1
      - name: cert-manager-operator.v1.14.0
        version: 1.14.0
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1.16
    - currentCSV: cert-manager-operator.v1.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
          createdAt: 2025-08-06T08:21:40
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.16.1 <1.17.0'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.17.4](https://github.com/cert-manager/cert-manager/tree/v1.17.4), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:96d51e3a64bf30cbd92836c7cbd82f06edca16eef78ab1432757d34c16628659
        version: 1.17.0
      entries:
      - name: cert-manager-operator.v1.17.0
        version: 1.17.0
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      - name: cert-manager-operator.v1.14.1
        version: 1.14.1
      - name: cert-manager-operator.v1.14.0
        version: 1.14.0
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1.17
    defaultChannel: stable-v1
    packageName: openshift-cert-manager-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Odoo Community Association
      provider-url: ""
    name: odoo-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odoo-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"odoo-community.org/v1alpha1","kind":"Odoo","metadata":{"name":"example"},"spec":{"postgresqlVolumeSize":"4Gi","odooVolumeSize":"4Gi"}}]'
          capabilities: Full Lifecycle
          categories: Application Runtime
          certified: "false"
          containerImage: ursa/odoo-operator:0.0.2
          createdAt: "2021-04-26T12:00:00Z"
          description: Operator to manage Odoo instances within an OpenShift cluster
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          repository: https://github.com/ursais/odoo-operator
          support: Open Source Integrators
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Odoo
            displayName: Odoo
            kind: Odoo
            name: odoos.odoo-community.org
            version: v1alpha1
        description: |
          ## About Odoo

          [Odoo](https://odoo-community.org), is a suite of business management software tools including CRM, e-commerce, billing, accounting, manufacturing, warehouse, project management, and inventory management to name a few.
          ## About the Odoo Operator

          This is an Ansible-based operator created with operator-sdk v1.5.0.

          The upgrade is handled by the Odoo container itself using [Marabunta](https://pypi.org/project/marabunta/) and [Anthem](https://pypi.org/project/anthem/).

          ## Roadmap
          * Add horizontal pod autoscaler
          * Add cronjob for backup
          ## Prerequisites for enabling this Operator

          Please refer to the README file on the [GitHub project](https://github.com/ursais/odoo-operator).
        displayName: Odoo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ERP
        - Business Software
        links:
        - name: Odoo Community Association
          url: https://odoo-community.org
        - name: GitHub Repository
          url: https://github.com/ursais/odoo-operator
        - name: Open Source Integrators
          url: https://opensourceintegrators.com
        maintainers:
        - email: mchambreuil@opensourceintegrators.com
          name: Maxime Chambreuil
        maturity: alpha
        provider:
          name: Odoo Community Association
        relatedImages:
        - ursa/odoo-operator:0.0.2
        version: 0.0.2
      entries:
      - name: odoo-operator.v0.0.2
        version: 0.0.2
      - name: odoo-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: odoo-operator
    provider:
      name: Odoo Community Association
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operator-kogitocloud: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kogito-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kogito-operator.v1.44.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "process-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogito-infinispan-infra"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "infinispan.org/v1",
                    "kind": "Infinispan"
                  }
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/kiegroup/kogito-operator:1.44.1
          createdAt: "2019-08-22T13:12:22Z"
          description: Kogito Operator for deployment and management of Kogito services.
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.app.kiegroup.org
            version: v1beta1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.app.kiegroup.org
            version: v1beta1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.app.kiegroup.org
            version: v1beta1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.app.kiegroup.org
            version: v1beta1
        description: |-
          Kogito Operator is designed to deploy Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Creates two kinds of build configuration: Native or JVM. For more information refer to [Native build and deployment](https://docs.jboss.org/kogito/release/latest/html_single/#_native_build_and_deployment) section on our docs
          * Builds a Kogito service from a remote git repository and creates a custom image for the given runtime: Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Installs the [Kogito Data Index Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-data-index-service_kogito-deploying-on-openshift) to enable data index feature for Kogito services.
          * Installs the [Kogito Jobs Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-jobs-service_kogito-deploying-on-openshift) to enable jobs scheduling.

          For Kubernetes deployment of Kogito services, please refer to the section [Deploying Kogito services on Kubernetes](https://docs.jboss.org/kogito/release/latest/html_single/#proc-kogito-deploying-on-kubernetes_kogito-deploying-on-openshift) on our docs.

          **IMPORTANT!** Kogito Operator might need Infinispan Operator and Strimzi to implement persistence and messaging use cases if you do not provide these services beforehand.
          If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the Kogito Operator.
        displayName: Kogito
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://kogito.kie.org/
        - name: Documentation
          url: https://github.com/kiegroup/kogito-operator/blob/main/README.md
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/kiegroup/kogito-operator:1.44.1
        version: 1.44.1
      entries:
      - name: kogito-operator.v1.44.1
        version: 1.44.1
      - name: kogito-operator.v1.43.0
        version: 1.43.0
      - name: kogito-operator.v1.42.0
        version: 1.42.0
      - name: kogito-operator.v1.41.0
        version: 1.41.0
      - name: kogito-operator.v1.40.0
        version: 1.40.0
      - name: kogito-operator.v1.39.0
        version: 1.39.0
      - name: kogito-operator.v1.38.0
        version: 1.38.0
      - name: kogito-operator.v1.37.0
        version: 1.37.0
      - name: kogito-operator.v1.36.0
        version: 1.36.0
      - name: kogito-operator.v1.35.0
        version: 1.35.0
      - name: kogito-operator.v1.34.0
        version: 1.34.0
      - name: kogito-operator.v1.33.0
        version: 1.33.0
      - name: kogito-operator.v1.32.0
        version: 1.32.0
      - name: kogito-operator.v1.31.0
        version: 1.31.0
      - name: kogito-operator.v1.30.0
        version: 1.30.0
      - name: kogito-operator.v1.29.0
        version: 1.29.0
      - name: kogito-operator.v1.28.0
        version: 1.28.0
      - name: kogito-operator.v1.27.0
        version: 1.27.0
      - name: kogito-operator.v1.26.0
        version: 1.26.0
      - name: kogito-operator.v1.25.0
        version: 1.25.0
      - name: kogito-operator.v1.24.0
        version: 1.24.0
      - name: kogito-operator.v1.23.0
        version: 1.23.0
      - name: kogito-operator.v1.22.1
        version: 1.22.1
      - name: kogito-operator.v1.22.0
        version: 1.22.0
      - name: kogito-operator.v1.21.0
        version: 1.21.0
      - name: kogito-operator.v1.20.0
        version: 1.20.0
      - name: kogito-operator.v1.19.0
        version: 1.19.0
      - name: kogito-operator.v1.18.0
        version: 1.18.0
      - name: kogito-operator.v1.17.0
        version: 1.17.0
      - name: kogito-operator.v1.16.1
        version: 1.16.1
      - name: kogito-operator.v1.16.0
        version: 1.16.0
      - name: kogito-operator.v1.15.0
        version: 1.15.0
      - name: kogito-operator.v1.14.0
        version: 1.14.0
      - name: kogito-operator.v1.13.0
        version: 1.13.0
      - name: kogito-operator.v1.12.0
        version: 1.12.0
      - name: kogito-operator.v1.11.1
        version: 1.11.1
      - name: kogito-operator.v1.11.0
        version: 1.11.0
      - name: kogito-operator.v1.10.0
        version: 1.10.0
      - name: kogito-operator.v1.9.1
        version: 1.9.1
      - name: kogito-operator.v1.9.0
        version: 1.9.0
      - name: kogito-operator.v1.8.0
        version: 1.8.0
      - name: kogito-operator.v1.7.0
        version: 1.7.0
      - name: kogito-operator.v1.6.0
        version: 1.6.0
      - name: kogito-operator.v1.5.0
        version: 1.5.0
      - name: kogito-operator.v1.4.0
        version: 1.4.0
      name: 1.x
    - currentCSV: kogito-operator.v1.44.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "process-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogito-infinispan-infra"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "infinispan.org/v1",
                    "kind": "Infinispan"
                  }
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/kiegroup/kogito-operator:1.44.1
          createdAt: "2019-08-22T13:12:22Z"
          description: Kogito Operator for deployment and management of Kogito services.
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.app.kiegroup.org
            version: v1beta1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.app.kiegroup.org
            version: v1beta1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.app.kiegroup.org
            version: v1beta1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.app.kiegroup.org
            version: v1beta1
        description: |-
          Kogito Operator is designed to deploy Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Creates two kinds of build configuration: Native or JVM. For more information refer to [Native build and deployment](https://docs.jboss.org/kogito/release/latest/html_single/#_native_build_and_deployment) section on our docs
          * Builds a Kogito service from a remote git repository and creates a custom image for the given runtime: Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Installs the [Kogito Data Index Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-data-index-service_kogito-deploying-on-openshift) to enable data index feature for Kogito services.
          * Installs the [Kogito Jobs Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-jobs-service_kogito-deploying-on-openshift) to enable jobs scheduling.

          For Kubernetes deployment of Kogito services, please refer to the section [Deploying Kogito services on Kubernetes](https://docs.jboss.org/kogito/release/latest/html_single/#proc-kogito-deploying-on-kubernetes_kogito-deploying-on-openshift) on our docs.

          **IMPORTANT!** Kogito Operator might need Infinispan Operator and Strimzi to implement persistence and messaging use cases if you do not provide these services beforehand.
          If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the Kogito Operator.
        displayName: Kogito
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://kogito.kie.org/
        - name: Documentation
          url: https://github.com/kiegroup/kogito-operator/blob/main/README.md
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kiegroup/kogito-operator:1.44.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        version: 1.44.1
      entries:
      - name: kogito-operator.v1.44.1
        version: 1.44.1
      - name: kogito-operator.v1.43.0
        version: 1.43.0
      - name: kogito-operator.v1.42.0
        version: 1.42.0
      - name: kogito-operator.v1.41.0
        version: 1.41.0
      - name: kogito-operator.v1.40.0
        version: 1.40.0
      - name: kogito-operator.v1.39.0
        version: 1.39.0
      - name: kogito-operator.v1.38.0
        version: 1.38.0
      - name: kogito-operator.v1.37.0
        version: 1.37.0
      - name: kogito-operator.v1.36.0
        version: 1.36.0
      - name: kogito-operator.v1.35.0
        version: 1.35.0
      - name: kogito-operator.v1.34.0
        version: 1.34.0
      - name: kogito-operator.v1.33.0
        version: 1.33.0
      - name: kogito-operator.v1.32.0
        version: 1.32.0
      - name: kogito-operator.v1.31.0
        version: 1.31.0
      - name: kogito-operator.v1.30.0
        version: 1.30.0
      - name: kogito-operator.v1.29.0
        version: 1.29.0
      - name: kogito-operator.v1.28.0
        version: 1.28.0
      - name: kogito-operator.v1.27.0
        version: 1.27.0
      - name: kogito-operator.v1.26.0
        version: 1.26.0
      - name: kogito-operator.v1.25.0
        version: 1.25.0
      - name: kogito-operator.v1.24.0
        version: 1.24.0
      - name: kogito-operator.v1.23.0
        version: 1.23.0
      - name: kogito-operator.v1.22.1
        version: 1.22.1
      - name: kogito-operator.v1.22.0
        version: 1.22.0
      - name: kogito-operator.v1.21.0
        version: 1.21.0
      - name: kogito-operator.v1.20.0
        version: 1.20.0
      - name: kogito-operator.v1.19.0
        version: 1.19.0
      - name: kogito-operator.v1.18.0
        version: 1.18.0
      - name: kogito-operator.v1.17.0
        version: 1.17.0
      - name: kogito-operator.v1.16.1
        version: 1.16.1
      - name: kogito-operator.v1.16.0
        version: 1.16.0
      - name: kogito-operator.v1.15.0
        version: 1.15.0
      - name: kogito-operator.v1.14.0
        version: 1.14.0
      - name: kogito-operator.v1.13.0
        version: 1.13.0
      - name: kogito-operator.v1.12.0
        version: 1.12.0
      - name: kogito-operator.v1.11.1
        version: 1.11.1
      - name: kogito-operator.v1.11.0
        version: 1.11.0
      - name: kogito-operator.v1.10.0
        version: 1.10.0
      - name: kogito-operator.v1.9.1
        version: 1.9.1
      - name: kogito-operator.v1.9.0
        version: 1.9.0
      - name: kogito-operator.v1.8.0
        version: 1.8.0
      - name: kogito-operator.v1.7.0
        version: 1.7.0
      - name: kogito-operator.v1.6.0
        version: 1.6.0
      - name: kogito-operator.v1.5.0
        version: 1.5.0
      - name: kogito-operator.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: 1.x
    packageName: kogito-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: neutron-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neutron-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "neutron.openstack.org/v1beta1",
                "kind": "NeutronAPI",
                "metadata": {
                  "name": "neutron",
                  "namespace": "openstack"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "neutron",
                  "databaseInstance": "openstack",
                  "memcachedInstance": "memcached",
                  "preserveJobs": false,
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "neutron-secret",
                  "serviceUser": "neutron"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:57Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NeutronAPI is the Schema for the neutronapis API
            displayName: Neutron API
            kind: NeutronAPI
            name: neutronapis.neutron.openstack.org
            version: v1beta1
        description: Neutron Operator
        displayName: Neutron Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Network
        - Neutron
        links:
        - name: Neutron Operator
          url: https://github.com/openstack-k8s-operators/neutron-operator
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/neutron-rhel9-operator@sha256:16b1a5f83692f3d193754df664992e5b88cb357f158b69261eb802430f3a69f1
        - registry.redhat.io/rhoso/openstack-neutron-server-rhel9@sha256:806e5e009a77a29c82699487e539669e1ec2136514a0332d222b5cb3fed6f452
        version: 1.0.6
      entries:
      - name: neutron-operator.v1.0.6
        version: 1.0.6
      - name: neutron-operator.v1.0.5
        version: 1.0.5
      - name: neutron-operator.v1.0.4
        version: 1.0.4
      - name: neutron-operator.v1.0.3
        version: 1.0.3
      - name: neutron-operator.v1.0.2
        version: 1.0.2
      - name: neutron-operator.v1.0.1
        version: 1.0.1
      - name: neutron-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: neutron-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ram-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ram-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ram-controller:1.1.1
          createdAt: "2025-09-19T22:18:09Z"
          description: AWS RAM controller is a service controller for managing RAM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Permission represents the state of an AWS ram Permission
              resource.
            displayName: Permission
            kind: Permission
            name: permissions.ram.services.k8s.aws
            version: v1alpha1
          - description: ResourceShare represents the state of an AWS ram ResourceShare
              resource.
            displayName: ResourceShare
            kind: ResourceShare
            name: resourceshares.ram.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon RAM resources in AWS from within your Kubernetes cluster.

          **About Amazon RAM**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon RAM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ram
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon RAM Developer Resources
          url: https://aws.amazon.com/RAM/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ram maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ram-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ram-controller.v1.1.1
        version: 1.1.1
      - name: ack-ram-controller.v1.1.0
        version: 1.1.0
      - name: ack-ram-controller.v1.0.11
        version: 1.0.11
      - name: ack-ram-controller.v1.0.10
        version: 1.0.10
      name: alpha
    defaultChannel: alpha
    packageName: ack-ram-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: project-quay
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-operator.v3.10.0-nightly.20231014
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.10.0-nightly.20231014
          createdAt: 2023-10-14 06:26 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.10.0-nightly.20231014'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.10.0-nightly.20231014
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.10.0-nightly.20231014
        version: 3.10.0-nightly.20231014
      entries:
      - name: quay-operator.v3.10.0-nightly.20231014
        version: 3.10.0-nightly.20231014
      - name: quay-operator.v3.10.0-nightly.20231013
        version: 3.10.0-nightly.20231013
      - name: quay-operator.v3.10.0-nightly.20231012
        version: 3.10.0-nightly.20231012
      - name: quay-operator.v3.10.0-nightly.20231011
        version: 3.10.0-nightly.20231011
      - name: quay-operator.v3.10.0-nightly.20231010
        version: 3.10.0-nightly.20231010
      - name: quay-operator.v3.10.0-nightly.20231009
        version: 3.10.0-nightly.20231009
      - name: quay-operator.v3.10.0-nightly.20231008
        version: 3.10.0-nightly.20231008
      - name: quay-operator.v3.10.0-nightly.20231007
        version: 3.10.0-nightly.20231007
      - name: quay-operator.v3.10.0-nightly.20231006
        version: 3.10.0-nightly.20231006
      - name: quay-operator.v3.10.0-nightly.20231005
        version: 3.10.0-nightly.20231005
      - name: quay-operator.v3.10.0-nightly.20231002
        version: 3.10.0-nightly.20231002
      - name: quay-operator.v3.10.0-nightly.20230930
        version: 3.10.0-nightly.20230930
      - name: quay-operator.v3.10.0-nightly.20230929
        version: 3.10.0-nightly.20230929
      - name: quay-operator.v3.10.0-nightly.20230928
        version: 3.10.0-nightly.20230928
      - name: quay-operator.v3.10.0-nightly.20230926
        version: 3.10.0-nightly.20230926
      - name: quay-operator.v3.10.0-nightly.20230925
        version: 3.10.0-nightly.20230925
      - name: quay-operator.v3.10.0-nightly.20230923
        version: 3.10.0-nightly.20230923
      - name: quay-operator.v3.10.0-nightly.20230922
        version: 3.10.0-nightly.20230922
      - name: quay-operator.v3.10.0-nightly.20230921
        version: 3.10.0-nightly.20230921
      - name: quay-operator.v3.10.0-nightly.20230920
        version: 3.10.0-nightly.20230920
      - name: quay-operator.v3.10.0-nightly.20230919
        version: 3.10.0-nightly.20230919
      - name: quay-operator.v3.10.0-nightly.20230918
        version: 3.10.0-nightly.20230918
      - name: quay-operator.v3.10.0-nightly.20230917
        version: 3.10.0-nightly.20230917
      - name: quay-operator.v3.10.0-nightly.20230916
        version: 3.10.0-nightly.20230916
      - name: quay-operator.v3.10.0-nightly.20230915
        version: 3.10.0-nightly.20230915
      - name: quay-operator.v3.10.0-nightly.20230914
        version: 3.10.0-nightly.20230914
      - name: quay-operator.v3.10.0-nightly.20230913
        version: 3.10.0-nightly.20230913
      name: candidate-3.10
    - currentCSV: quay-operator.v3.9.0-nightly.20230812
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.9.0-nightly.20230812
          createdAt: 2023-08-12 06:06 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.9.0-nightly.20230812'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.9.0-nightly.20230812
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.9.0-nightly.20230812
        version: 3.9.0-nightly.20230812
      entries:
      - name: quay-operator.v3.9.0-nightly.20230812
        version: 3.9.0-nightly.20230812
      - name: quay-operator.v3.9.0-nightly.20230724
        version: 3.9.0-nightly.20230724
      - name: quay-operator.v3.9.0-nightly.20230723
        version: 3.9.0-nightly.20230723
      - name: quay-operator.v3.9.0-nightly.20230722
        version: 3.9.0-nightly.20230722
      - name: quay-operator.v3.9.0-nightly.20230721
        version: 3.9.0-nightly.20230721
      - name: quay-operator.v3.9.0-nightly.20230720
        version: 3.9.0-nightly.20230720
      - name: quay-operator.v3.9.0-nightly.20230717
        version: 3.9.0-nightly.20230717
      - name: quay-operator.v3.9.0-nightly.20230716
        version: 3.9.0-nightly.20230716
      - name: quay-operator.v3.9.0-nightly.20230715
        version: 3.9.0-nightly.20230715
      - name: quay-operator.v3.9.0-nightly.20230714
        version: 3.9.0-nightly.20230714
      - name: quay-operator.v3.9.0-nightly.20230712
        version: 3.9.0-nightly.20230712
      - name: quay-operator.v3.9.0-nightly.20230711
        version: 3.9.0-nightly.20230711
      name: candidate-3.9
    - currentCSV: quay-operator.v3.7.0-rc.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.7.0-rc.4
          createdAt: 2022-05-17 14:05 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.3.x <3.7.0-rc.4'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.7.0-rc.4
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.7.0-rc.4
        version: 3.7.0-rc.4
      entries:
      - name: quay-operator.v3.7.0-rc.4
        version: 3.7.0-rc.4
      - name: quay-operator.v3.7.0-rc.3
        version: 3.7.0-rc.3
      - name: quay-operator.v3.7.0-rc.2
        version: 3.7.0-rc.2
      - name: quay-operator.v3.7.0-rc.1
        version: 3.7.0-rc.1
      name: preview-3.7
    - currentCSV: quay-operator.v3.8.0-rc.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.8.0-rc.7
          createdAt: 2022-11-29 15:11 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.8.0-rc.7'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.8.0-rc.7
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.8.0-rc.7
        version: 3.8.0-rc.7
      entries:
      - name: quay-operator.v3.8.0-rc.7
        version: 3.8.0-rc.7
      - name: quay-operator.v3.8.0-rc.6
        version: 3.8.0-rc.6
      - name: quay-operator.v3.8.0-rc.5
        version: 3.8.0-rc.5
      - name: quay-operator.v3.8.0-rc.4
        version: 3.8.0-rc.4
      - name: quay-operator.v3.8.0-rc.3
        version: 3.8.0-rc.3
      - name: quay-operator.v3.8.0-rc.2
        version: 3.8.0-rc.2
      - name: quay-operator.v3.8.0-rc.1
        version: 3.8.0-rc.1
      name: preview-3.8
    - currentCSV: quay-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.10.15
          createdAt: "2025-09-25T14:06:05Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.10.15
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.10.15
        version: 3.10.15
      entries:
      - name: quay-operator.v3.10.15
        version: 3.10.15
      - name: quay-operator.v3.10.14
        version: 3.10.14
      - name: quay-operator.v3.10.13
        version: 3.10.13
      - name: quay-operator.v3.10.12
        version: 3.10.12
      - name: quay-operator.v3.10.11
        version: 3.10.11
      - name: quay-operator.v3.10.10
        version: 3.10.10
      - name: quay-operator.v3.10.9
        version: 3.10.9
      - name: quay-operator.v3.10.8
        version: 3.10.8
      - name: quay-operator.v3.10.7
        version: 3.10.7
      - name: quay-operator.v3.10.6
        version: 3.10.6
      - name: quay-operator.v3.10.5
        version: 3.10.5
      - name: quay-operator.v3.10.4
        version: 3.10.4
      - name: quay-operator.v3.10.3
        version: 3.10.3
      - name: quay-operator.v3.10.2
        version: 3.10.2
      - name: quay-operator.v3.10.1
        version: 3.10.1
      - name: quay-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.11.13
          createdAt: "2025-09-02T06:02:15Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.11.13
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.11.13
        version: 3.11.13
      entries:
      - name: quay-operator.v3.11.13
        version: 3.11.13
      - name: quay-operator.v3.11.12
        version: 3.11.12
      - name: quay-operator.v3.11.11
        version: 3.11.11
      - name: quay-operator.v3.11.10
        version: 3.11.10
      - name: quay-operator.v3.11.9
        version: 3.11.9
      - name: quay-operator.v3.11.8
        version: 3.11.8
      - name: quay-operator.v3.11.7
        version: 3.11.7
      - name: quay-operator.v3.11.6
        version: 3.11.6
      - name: quay-operator.v3.11.5
        version: 3.11.5
      - name: quay-operator.v3.11.4
        version: 3.11.4
      - name: quay-operator.v3.11.3
        version: 3.11.3
      - name: quay-operator.v3.11.2
        version: 3.11.2
      - name: quay-operator.v3.11.1
        version: 3.11.1
      - name: quay-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.12.12
          createdAt: "2025-09-23T09:59:30Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.12.12
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.12.12
        version: 3.12.12
      entries:
      - name: quay-operator.v3.12.12
        version: 3.12.12
      - name: quay-operator.v3.12.11
        version: 3.12.11
      - name: quay-operator.v3.12.10
        version: 3.12.10
      - name: quay-operator.v3.12.9
        version: 3.12.9
      - name: quay-operator.v3.12.8
        version: 3.12.8
      - name: quay-operator.v3.12.7
        version: 3.12.7
      - name: quay-operator.v3.12.6
        version: 3.12.6
      - name: quay-operator.v3.12.5
        version: 3.12.5
      - name: quay-operator.v3.12.4
        version: 3.12.4
      - name: quay-operator.v3.12.3
        version: 3.12.3
      - name: quay-operator.v3.12.2
        version: 3.12.2
      - name: quay-operator.v3.12.1
        version: 3.12.1
      - name: quay-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.13.8
          createdAt: "2025-08-28T03:28:40Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.13.8
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.13.8
        version: 3.13.8
      entries:
      - name: quay-operator.v3.13.8
        version: 3.13.8
      - name: quay-operator.v3.13.7
        version: 3.13.7
      - name: quay-operator.v3.13.6
        version: 3.13.6
      - name: quay-operator.v3.13.5
        version: 3.13.5
      - name: quay-operator.v3.13.4
        version: 3.13.4
      - name: quay-operator.v3.13.3
        version: 3.13.3
      - name: quay-operator.v3.13.2
        version: 3.13.2
      - name: quay-operator.v3.13.1
        version: 3.13.1
      - name: quay-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.14.5
          createdAt: "2025-09-22T14:19:49Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.14.5
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.14.5
        version: 3.14.5
      entries:
      - name: quay-operator.v3.14.5
        version: 3.14.5
      - name: quay-operator.v3.14.4
        version: 3.14.4
      - name: quay-operator.v3.14.3
        version: 3.14.3
      - name: quay-operator.v3.14.2
        version: 3.14.2
      - name: quay-operator.v3.14.1
        version: 3.14.1
      - name: quay-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.15.2
          createdAt: "2025-09-22T13:44:32Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.15.2
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.15.2
        version: 3.15.2
      entries:
      - name: quay-operator.v3.15.2
        version: 3.15.2
      - name: quay-operator.v3.15.1
        version: 3.15.1
      - name: quay-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-operator.v3.6.2
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"quay.redhat.com/v1\",\n    \"kind\":
            \"QuayRegistry\",\n    \"metadata\": {\n      \"name\": \"example-registry\"\n
            \   },\n    \"spec\": {\n      \"components\": [\n        {\"kind\": \"clair\",
            \"managed\": true},\n        {\"kind\": \"postgres\", \"managed\": true},\n
            \       {\"kind\": \"objectstorage\", \"managed\": true},\n        {\"kind\":
            \"redis\", \"managed\": true},\n        {\"kind\": \"horizontalpodautoscaler\",
            \"managed\": true},\n        {\"kind\": \"route\", \"managed\": true},\n
            \       {\"kind\": \"mirror\", \"managed\": true},\n        {\"kind\":
            \"monitoring\", \"managed\": true},\n        {\"kind\": \"tls\", \"managed\":
            true}\n      ]\n    }\n  }, \n  {\n    \"apiVersion\": \"redhatcop.redhat.io/v1alpha1\",\n
            \   \"kind\": \"QuayEcosystem\",\n    \"metadata\": {\n      \"name\":
            \"example-quayecosystem\"\n    },\n    \"spec\": {}\n  }\n]"
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:v3.6.2
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.5.x <3.6.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.6.2
          repository: https://github.com/quay/quay-operator
          support: Quay Team
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Project Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        - quay
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:v3.6.2
        version: 3.6.2
      entries:
      - name: quay-operator.v3.6.2
        version: 3.6.2
      name: stable-3.6
    - currentCSV: quay-operator.v3.7.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.7.11
          createdAt: 2023-02-22 10:02 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.3.x <3.7.11'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.7.11
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.7.11
        version: 3.7.11
      entries:
      - name: quay-operator.v3.7.11
        version: 3.7.11
      - name: quay-operator.v3.7.10
        version: 3.7.10
      - name: quay-operator.v3.7.9
        version: 3.7.9
      - name: quay-operator.v3.7.8
        version: 3.7.8
      - name: quay-operator.v3.7.7
        version: 3.7.7
      - name: quay-operator.v3.7.6
        version: 3.7.6
      - name: quay-operator.v3.7.5
        version: 3.7.5
      - name: quay-operator.v3.7.4
        version: 3.7.4
      - name: quay-operator.v3.7.3
        version: 3.7.3
      - name: quay-operator.v3.7.2
        version: 3.7.2
      - name: quay-operator.v3.7.1
        version: 3.7.1
      - name: quay-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: quay-operator.v3.8.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.8.13
          createdAt: 2023-10-17 16:13 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.8.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.8.13
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.8.13
        version: 3.8.13
      entries:
      - name: quay-operator.v3.8.13
        version: 3.8.13
      - name: quay-operator.v3.8.12
        version: 3.8.12
      - name: quay-operator.v3.8.11
        version: 3.8.11
      - name: quay-operator.v3.8.10
        version: 3.8.10
      - name: quay-operator.v3.8.8
        version: 3.8.8
      - name: quay-operator.v3.8.7
        version: 3.8.7
      - name: quay-operator.v3.8.6
        version: 3.8.6
      - name: quay-operator.v3.8.5
        version: 3.8.5
      - name: quay-operator.v3.8.4
        version: 3.8.4
      - name: quay-operator.v3.8.3
        version: 3.8.3
      - name: quay-operator.v3.8.2
        version: 3.8.2
      - name: quay-operator.v3.8.1
        version: 3.8.1
      - name: quay-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.9.16
          createdAt: "2025-09-25T14:31:58Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.9.16
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.9.16
        version: 3.9.16
      entries:
      - name: quay-operator.v3.9.16
        version: 3.9.16
      - name: quay-operator.v3.9.15
        version: 3.9.15
      - name: quay-operator.v3.9.14
        version: 3.9.14
      - name: quay-operator.v3.9.13
        version: 3.9.13
      - name: quay-operator.v3.9.12
        version: 3.9.12
      - name: quay-operator.v3.9.11
        version: 3.9.11
      - name: quay-operator.v3.9.10
        version: 3.9.10
      - name: quay-operator.v3.9.9
        version: 3.9.9
      - name: quay-operator.v3.9.6
        version: 3.9.6
      - name: quay-operator.v3.9.5
        version: 3.9.5
      - name: quay-operator.v3.9.4
        version: 3.9.4
      - name: quay-operator.v3.9.3
        version: 3.9.3
      - name: quay-operator.v3.9.2
        version: 3.9.2
      - name: quay-operator.v3.9.1
        version: 3.9.1
      - name: quay-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: project-quay
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-ui
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-ui.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Ui",
                "metadata": {
                  "name": "ui-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "volumeMounts": [
                    {
                      "mountPath": "/opt/app-root/src/assets/config.json",
                      "name": "config",
                      "subPath": "config.json"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:37:58Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ui
            name: uis.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP UI Operator
        displayName: Keysight WAP UI Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Ui
          url: https://keysight-wap-ui.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-ui@sha256:876a394f5cc2f745091ad2ee9ca0d08404cfb95284222e1f3a8607e193bd986a
        version: 0.1.21
      entries:
      - name: keysight-wap-ui.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-ui
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator-certified-rhmp.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator-certified-rhmp.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator-certified-rhmp.v2.6.1
        version: 2.6.1
      - name: percona-postgresql-operator-certified-rhmp.v2.6.0
        version: 2.6.0
      - name: percona-postgresql-operator-certified-rhmp.v2.5.0
        version: 2.5.0
      name: stable
    - currentCSV: percona-postgresql-operator-certified-rhmp.v2.7.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        version: 2.7.0-cw
      entries:
      - name: percona-postgresql-operator-certified-rhmp.v2.7.0-cw
        version: 2.7.0-cw
      - name: percona-postgresql-operator-certified-rhmp.v2.6.1-cw
        version: 2.6.1-cw
      - name: percona-postgresql-operator-certified-rhmp.v2.6.0-cw
        version: 2.6.0-cw
      - name: percona-postgresql-operator-certified-rhmp.v2.5.0-cw
        version: 2.5.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-postgresql-operator-certified-rhmp
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Tempo Operator SIG
      provider-url: ""
    name: tempo-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tempo-operator.v0.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoMonolithic",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "jaegerui": {
                    "enabled": true,
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    }
                  },
                  "storage": {
                    "traces": {
                      "backend": "memory"
                    }
                  }
                }
              },
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoStack",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "resources": {
                    "total": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2Gi"
                      }
                    }
                  },
                  "storage": {
                    "secret": {
                      "name": "my-storage-secret",
                      "type": "s3"
                    }
                  },
                  "storageSize": "1Gi",
                  "template": {
                    "queryFrontend": {
                      "jaegerQuery": {
                        "enabled": true,
                        "ingress": {
                          "type": "route"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          containerImage: ghcr.io/grafana/tempo-operator/tempo-operator:v0.18.0
          createdAt: "2025-09-17T14:23:28Z"
          description: Create and manage deployments of Tempo, a high-scale distributed
            tracing backend.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: tempo-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/tempo-operator
          support: Grafana Tempo Operator SIG
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TempoMonolithic manages a Tempo deployment in monolithic
              mode.
            displayName: TempoMonolithic
            kind: TempoMonolithic
            name: tempomonolithics.tempo.grafana.com
            version: v1alpha1
          - description: TempoStack manages a Tempo deployment in microservices mode.
            displayName: TempoStack
            kind: TempoStack
            name: tempostacks.tempo.grafana.com
            version: v1alpha1
        description: |-
          Tempo is an open source, easy-to-use, and high-scale distributed tracing backend.
          It can ingest common open source tracing protocols including Jaeger, Zipkin, and OpenTelemetry and requires only object storage to operate.
          Please refer to the [Tempo documentation](https://grafana.com/docs/tempo/latest/) for more information about Tempo.

          The Community Tempo Operator can manage Tempo deployments in Microservices mode (`TempoStack` CR) and Monolithic mode (`TempoMonolithic` CR).

          ### Operator features
          * **Resource Limits** - Specify overall resource requests and limits in the `TempoStack` CR; the operator assigns fractions of it to each component
          * **AuthN and AuthZ** - Supports OpenID Control (OIDC) and role-based access control (RBAC)
          * **Managed upgrades** - Updating the operator will automatically update all managed Tempo clusters
          * **Multitenancy** - Multiple tenants can send traces to the same Tempo cluster
          * **mTLS** - Communication between the Tempo components can be secured via mTLS
          * **Jaeger UI** - Traces can be visualized in Jaeger UI and exposed via Ingress or OpenShift Route
          * **Observability** - The operator and `TempoStack` operands expose telemetry (metrics, traces) and integrate with Prometheus `ServiceMonitor` and `PrometheusRule`

          ### Prerequisites
          `TempoStack` requires object storage to store its traces. `TempoMonolithic` can store traces in-memory, in a Persistent Volume and in object storage.
          Please ensure that an object storage solution is available and configured.
        displayName: Community Tempo Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tempo
        - tracing
        - observability
        - monitoring
        - database
        links:
        - name: Tempo Operator
          url: https://github.com/grafana/tempo-operator
        maintainers:
        - email: ruben.vp8510@gmail.com
          name: Ruben Vargas
        - email: p.loffay@gmail.com
          name: Pavol Loffay
        - email: bongartz@klimlive.de
          name: Benedikt Bongartz
        - email: andreas@gerstmayr.me
          name: Andreas Gerstmayr
        - email: iblancas@redhat.com
          name: Israel Blancas Alvarez
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Grafana Tempo Operator SIG
        relatedImages:
        - docker.io/jaegertracing/jaeger-query:1.68.0
        - docker.io/grafana/tempo-query:2.8.2
        - quay.io/observatorium/api:main-2025-08-25-7489eb0
        - quay.io/observatorium/opa-openshift:main-2025-06-16-ecdeca0
        - quay.io/openshift/origin-oauth-proxy:4.14
        - ghcr.io/grafana/tempo-operator/tempo-operator:v0.18.0
        - docker.io/grafana/tempo:2.8.2
        version: 0.18.0
      entries:
      - name: tempo-operator.v0.18.0
        version: 0.18.0
      - name: tempo-operator.v0.17.1
        version: 0.17.1
      - name: tempo-operator.v0.15.3
        version: 0.15.3
      - name: tempo-operator.v0.15.2
        version: 0.15.2
      - name: tempo-operator.v0.15.1
        version: 0.15.1
      - name: tempo-operator.v0.15.0
        version: 0.15.0
      - name: tempo-operator.v0.14.2
        version: 0.14.2
      - name: tempo-operator.v0.14.1
        version: 0.14.1
      - name: tempo-operator.v0.14.0
        version: 0.14.0
      - name: tempo-operator.v0.13.0
        version: 0.13.0
      - name: tempo-operator.v0.12.0
        version: 0.12.0
      - name: tempo-operator.v0.11.1
        version: 0.11.1
      - name: tempo-operator.v0.11.0
        version: 0.11.0
      - name: tempo-operator.v0.10.0
        version: 0.10.0
      - name: tempo-operator.v0.9.0
        version: 0.9.0
      - name: tempo-operator.v0.8.0
        version: 0.8.0
      - name: tempo-operator.v0.7.0
        version: 0.7.0
      - name: tempo-operator.v0.6.0
        version: 0.6.0
      - name: tempo-operator.v0.5.0
        version: 0.5.0
      - name: tempo-operator.v0.4.0
        version: 0.4.0
      - name: tempo-operator.v0.3.0
        version: 0.3.0
      - name: tempo-operator.v0.2.0
        version: 0.2.0
      - name: tempo-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: tempo-operator
    provider:
      name: Grafana Tempo Operator SIG
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-application-gateway-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-application-gateway-operator.v25.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMApplicationGateway",
                "metadata": {
                  "name": "iag-instance"
                },
                "spec": {
                  "configuration": [
                    {
                      "dataKey": "config",
                      "name": "test-config",
                      "type": "configmap"
                    },
                    {
                      "discoveryEndpoint": "https://isam.mmfa.ibm.com/mga/sps/oauth/oauth20/metadata/test",
                      "postData": [
                        {
                          "name": "redirect_uris",
                          "values": [
                            "https://isam.mmfa.ibm.com/pkmsoidc"
                          ]
                        },
                        {
                          "name": "client_name",
                          "value": "OperatorTest"
                        },
                        {
                          "name": "enforce_pkce",
                          "value": "false"
                        },
                        {
                          "name": "all_users_entitled",
                          "value": "true"
                        },
                        {
                          "name": "consent_action",
                          "value": "never_prompt"
                        }
                      ],
                      "secret": "oidc-client",
                      "type": "oidc_registration"
                    },
                    {
                      "headers": [
                        {
                          "name": "Authorization",
                          "secretKey": "value",
                          "type": "secret",
                          "value": "githubsecret"
                        }
                      ],
                      "type": "web",
                      "url": "https://raw.github.com/ibm-security/iag-config/master/test/sample1.yaml"
                    },
                    {
                      "type": "literal",
                      "value": "version: \"22.07\"\n\nserver:\n  local_applications:\n    cred_viewer:\n      path_segment: creds\n      enable_html: true\n\nidentity:\n  oidc:\n    discovery_endpoint: \"https://iag-dev.ibmcloudsecurity.com/oidc/endpoint/default/.well-known/openid-configuration\"\n    client_id: 1cbfe647-9e5f-4d99-8e05-8ed1c862eb47\n    client_secret: uPP8rM7N0e\n"
                    }
                  ],
                  "deployment": {
                    "image": "icr.io/ibmappgateway/ibm-application-gateway@sha256:1144b79a2930fe942ce627b0abcc3bed2f0b0770ae8313cb77ffe6772511c1f9",
                    "imagePullPolicy": "Always",
                    "imagePullSecrets": [
                      {
                        "name": "regcred"
                      }
                    ],
                    "lang": "C",
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 8,
                      "periodSeconds": 9,
                      "successThreshold": 7,
                      "timeoutSeconds": 1
                    },
                    "readinessProbe": {
                      "failureThreshold": 2,
                      "initialDelaySeconds": 7,
                      "periodSeconds": 8,
                      "successThreshold": 4,
                      "timeoutSeconds": 5
                    },
                    "serviceAccountName": "iag"
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/ibmappgateway/ibm-application-gateway-operator@sha256:6a16d4ef802a866942f504b11a948565245d303f322944200c1a1269fe7c371c
          createdAt: "2025-02-05T21:15:37Z"
          description: The IBM Application Gateway operator manages the lifecycle
            of IBM Application Gateway containers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM-Security/ibm-application-gateway-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMApplicationGateway is the Schema for the ibmapplicationgateways
              API
            displayName: IBMApplication Gateway
            kind: IBMApplicationGateway
            name: ibmapplicationgateways.ibm.com
            version: v1
        description: "The [IBM Application Gateway (IAG)](https://ibm.biz/ibm-app-gateway)
          image provides a containerized secure Web Reverse proxy which is designed
          to sit in front of your application, seamlessly adding authentication and
          authorization protection to your application.\n\nAn IAG instance deployed
          on Kubernetes can be a complex deployment.  In particular the configuration
          can be defined externally in one or more locations, and changes to this
          configuration may require all instances to be reloaded for the changes to
          take effect. \n\nThe internal Kubernetes deployment controller does not
          have any knowledge of how an IBM Application Gateway instance should behave
          when the configuration changes. As such this IBM Application Gateway specific
          Kubernetes operator is available to be deployed. Once deployed the operator
          can manage the lifecycle of the IBM Application Gateway instances.\n\nSee
          the project [Readme](https://github.com/IBM-Security/ibm-application-gateway-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Application Gateway Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Security
        - Identity and Access
        links:
        - name: IBM Application Gateway product documentation
          url: https://ibm.biz/ibm-app-gateway
        maintainers:
        - email: isamdev@au1.ibm.com
          name: IBM Application Gateway Development Team
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - icr.io/ibmappgateway/ibm-application-gateway-operator@sha256:6a16d4ef802a866942f504b11a948565245d303f322944200c1a1269fe7c371c
        - icr.io/ibmappgateway/ibm-application-gateway@sha256:1144b79a2930fe942ce627b0abcc3bed2f0b0770ae8313cb77ffe6772511c1f9
        version: 25.2.0
      entries:
      - name: ibm-application-gateway-operator.v25.2.0
        version: 25.2.0
      - name: ibm-application-gateway-operator.v24.4.0
        version: 24.4.0
      name: stable
    defaultChannel: stable
    packageName: ibm-application-gateway-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cockroach Labs
      provider-url: ""
    name: cockroachdb-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroach-operator.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v21.1.10",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cockroachdb/cockroachdb-operator@sha256:9fba7cd5fdf589317911925cb69338508bf9de928b1817c6f117c612635bbab3
          createdAt: "2021-10-19T15:47:03Z"
          description: CockroachDB Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:e67299fc0e79cff5b6caf9c0df154180dbaedc57b84f072acf68b7b2c958668d
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ec1672ce8091b677689a31cf09dbde5d1697f4caddae0151040b0fb156722e9d
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:e39bce1d9b08a7d62eb05d53e29fabc668bf6a4e4ae3eb9994eebf5b642ff6cf
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ab47ddd22328e0a1564814f049616c3e3569d24c0a290abee9308392c6a0de23
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:49421968126f9c98499bc0cf0c65729370ab566da3fbd1588c100b20052b972e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:c56f980b87a066d1e65ac8cff3d75040af9bd13093494cf4b4d30b5234d56edc
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:de94036e9cc8618e0493313512d880ae145e94d3bff54ed51b0de66a88da1319
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:3edda46eccce8350440a0ea78dad2e41a3eaa0838fe57945ad5318739e968516
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:0bd22d69db5ac570f30d3379afc67f6c2e46fc5b027c8aab42c3f562085d4672
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:bcf3f976437c4ee52e922aa5d7e382a86a031d780b711c08c6c158dbf2d26cfa
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:0ef0234b6fd42977b8a9eda2c59ca8194cc1c8cf6425b99d7bf8fde8feb826c5
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:87dcdbdc07904b217880e70484582c9d87dde7e4071e01cf9e2e6da43111190e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:2664d78fbe88c368927ddbf80704f430e6ca2ce1f72cb4b7e1906dfe72be7cd0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:fe0c615b4d1ef4ec6e089e85deae2b6bd85410be8ed525e2317c759669417f47
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1b56420753eac1126039b8ad4be82ddb59c89a4ec72fb7b6198cea88a10f3a91
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1ed18af1d3ffc65c24946c817d7d04d5f01025d3fb6c9eb78940e343b989bed0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:162a356cc8b423926634f83a1372b6321104d8de7a2c5b9b7b2681c769a88cc7
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:288ae92ebdfc848540ff80ef682b74e50809e9742cafce22b028112326d66b65
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:867c46bb4d8ca3f2fb8dc94063facc99758344c7b9d04f10e50932023ef5d262
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ba9da9aa9d662c0fa9fd16a40551eedf100869b008cf9c0c09cbb87a09fda697
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:25791b76b0d4b7220dadab3a405b3fdc330ba264ef6c9fbfdfb924c154c4bb5e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:cdab9fc62e07dd349703ae394adb48ab3d2281610df35e081f4b76e208b38c8e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:a5414e54c4c513c7877f925b09eadc4c28eb592d3e120bcf9c65ab10e07cd607
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:98b76cf63aabc39d026317905d6ba72eec3f0d35c0f770ebb8b789c3fa0709ee
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:9f5dfcde260804cd2d6976a6b3544ddabdd899365be2104210dc78ba13755ec3
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:27ef61d9f320a52ee9617f2ab8a1d1a1eacc576a0add2d85d99f9413912b651f
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:38e940144e34b69dfb7d1cf31456e7a0016c5dd40d5626baa346446c7678ceb0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1cea12847a0d236437f7fc28fba476fe512c343cd020d21bb50316e50da6ffd6
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:c1a07a8686bbadcb8c68ffba4e40e1285e581e24d43e5811f38a22d57fb9cc56
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:d290099496d070f135b5cc1e6bd59607512524d399df81698f11cfa09b8dca4a
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:b02549e5c3673b8154441990afa05a478d1ecc2bd0a70af64fd693cb6b469cb0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:5eb59093ee81f8115d82ec6a8a1c35eefad055cbb3a175978cef7afe9196e6a0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:fec6b0f56f2b94f141c8553a63b2e08b9494d4e6c5dc109d73268621ca445c56
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:d9a756df2ed2536b938d7c24fbf3de149cd1a0f44a3a1a0cc5d042fe8980362e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroachdb-operator@sha256:9fba7cd5fdf589317911925cb69338508bf9de928b1817c6f117c612635bbab3
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:6667919a41d304d5d4ade3ded4f11b42d722a995a4283e11e15320529f7f9abf
        version: 2.3.0
      entries:
      - name: cockroach-operator.v2.3.0
        version: 2.3.0
      name: beta
    - currentCSV: cockroach-operator.v2.18.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v25.2.2",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
          createdAt: "2025-07-22T11:30:56Z"
          description: CockroachDB Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9ab968e4ca62f23bf9848e5475ef4076df7318a8560f63d5b2090a32af6fe4e2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:90cf21c36d19d87b97d06edd1c0912a3bb288c769e6159651877db3206aa7355
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8f1f2fd1aca914a1ef007117c85ac944055bbfb873a468a7374d239ef20b065e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9b5a2e0c006eb4be8e1444734a98743551f278a24e81c20bebc96f0f3b631eb0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6371377718325235a2720dba0c97eb4321ee29b31a28cb893f76abd62baa339
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:68fe64fe3e112288ccfc2a7dffd69b9b2786a5da3d3bdedd6addf41659d460c9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5240f9f591263569fad7ee934e81fa8a1617de3c25da94ddd59ee5b2b3407f90
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f29c31bb267451b389608af3e8a4a28d6ca0691e67b77e8f57bb5e461b064893
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8737b67b91983817e3e4ff3f47f0b070bd36162d309b60993529426c6979cb24
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a3ac544d87a9d8e7c92897c4c23f803753c7ab31467db91b5c51244afc2c6be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf001173bbf00cdf443b8d64a3ad724389f1cf83e3628d0b609019c8cddad35a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce3d94a3fbdcde5858beb3c265987759971d991c397c6486e19904bd4706f44d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7d6320ac7fb0925f10cf6f524a13de629cbf13ccc371912e4fe8c777bc15d1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bd75aa918355e77cd671c3e11ee807861378e96245245f4fdf243cba5749f9a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d590946b2f7c7ec0c7ec057a338e81f864d6a9e3d3ac4e0769aa5756a8e13fc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:32fdca575c334822e4356aab36a7ed97b685c065925fe85f1b8ba8425c57159b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc4b75ed71845d5b18e1d9163d480fd9d0d3cb5ebbcfed00a2e4e1b174c0a5de
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f74c8a5dea4560ad59c6f6f6ab1e699f063334ee8db704460f870508dbeaeff0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4b3aa590623a38f435b075de4c09f8a343d8ece8b673e6bf7c6df5d5eb4bdf4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:46d9287ef11e3b21a36abfa31a26da3b93040bb6163b9223972a5dd1930fecff
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e0f21abea34825e55622c432a306f2e8aafd64d913eb78f2f601c51dc878a1da
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2b5c028f293c6f394e58a5495cb3dbfbedc215d0d03778d0217e045586d55a46
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08cec123813304ab03bd66d72a09949667a874a42320bedafda391ebcdb2e56c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6a2ee8d62c586eb74ebbef6bef256e06e1ba5cb51c1ae55b1654adafc2ca8859
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c2a62179e8ca293d7e30b9da599e48795345cf07a72abd3a83063107b53b03cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:72ea3d0ef72972dba93362dc31366c5946a9898f5c16780b877316a97068ef8f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1304d7f871a95c8b087bad1b548217562792edc66c8fe07c621e05242b8cd53e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9492894eaa5f43d00ac80b60029b32a0b0a85ae88ba609b46f98702c9349a5bb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bf36dab9b3257a4bba24d986f1451fc51a3a37023fbb988281f1941fd3faedd
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d467383af41aa80c26172964c34152abeba45121599804e502984655b72179f0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:022c6e1a0de875474bc953718a481a0af526c49dc910094dcdce0b5677aced18
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2d89398f03736f8ffaa82b10fed0f19fba166d3d2145b26f25cb71f3c39acf07
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e23af8f9d32c3f80a52d79ceb4b359dda1d75ba0dd4cb8ed4b66ac2f4e40c69b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:61e602b0e3582bf00460eb9328808ccf66e3ff1b1f6df07ee6c5c76a91846876
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8443c15eb1653dd9e5ec7aad54d4b8b092d41a0977e42b76e372e5e21280746e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0ec1dc184a2199736b4bd7de3241a357be7427caea5da1c931b08bc06f5c3dc0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c9a31bdf19175880d0981e21dd19a1458b6e8cde8d4acbcb554f3bea3578a64e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3f587e7b9f5a611013af2608e2151ed782859a52501b00adbc95739c56a8ff1e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7ad12df8e258b03c4099d375d02d272bd774bf188d108901d5cf6cb865e57ef3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4f753151af4c3d3f5a1df9014461316249a6d0621b8fba5c4430307b3355251
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:58de9d0b2ce005c7ad3e99152100e59b7677398557028a7a59e56e669fabb81c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1f6b1ec92251397dbcf7ed4a82f1f05619d64676aae2717f4eff1c33e4c2c27a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:266165ab87b19a05d6796e8b15ff0eba632d6ac96e6393400464c6136dd500ec
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:086d7b435993fdf21fd59b0093f52105a9028d6b769398b0033f98a65dfb7e79
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc1ab77a9537ccf236ce6dc370940dfedcfaa91b6e3db37bf2d5eade7ff84708
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4d8e804b35b73d7978c112748aea2d2bb68f65081b90bca62c39f2b6e2989378
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:744193037d6b64cb73fecab15ebe022abc2b79ab0335b76ec1415ac6379b7dfa
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5181196fff3bf8a5ec8fcfe4f1958cc69f285f0beb4f0b476b4b5e0912f3db42
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:40f00476b63cb9e370fe85f01407173693213a9242fc117d28b06378ca0d98e0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fd6c9b8ca82ff4f2d0fe29166d1fc9bca6439cb40907be60f4d8f1b9735de2f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:85a0e085901c2a73192556d5d70795d4b66cbcfe01d24c73b6fe6db8d5f2e5ca
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6bb15b36d64eebb6e4c1db5a5466e108b271d53383c58e0b6c78cec214756a9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b83eaff2e217c06098509709cead525b55b00ccd48019d8357e6231892008055
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:858dc82977b3c9e989a3a20626671b0c67c834c97132094a9212bb227ab55929
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da367cf0ac52045002e1c186f8e6964267ad87d5c25f9e72fd2c9b9a98a32702
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b77b80eed7ce149d2a3f228e5f188e76d73f32bc50ecde20bba9cfa7bdc4714c
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:53c7b960a2ed70a0998a4f8e28a4f78f7f676e01cffebe665e91e3a7629d88ed
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1558a1f5eab79f1a8c6c85d81740cc539dc23025ea457c0019d93a72dc05d57a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db4e5c2f27ba23e95a363c8174c01f0f9c6b6a27e0e581c86a38f3a1f68dc43c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:37bf9e1ec7e3eff47755fe7521223f31ec6aefc6e63901db3881b30898cf78f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:35ede65457d845e8c063ad7d9515d3a69f98e0751424ecbac484963678a0b1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:50393ad005fed19618d04f0716c6d55e2f1afce651e785c396638e852cac83b3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cad60044ad6573bd80b65e892a9ec0510dbe4fcbbfc6b51010ecf419f56f1024
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f42fa22aac3cc41ce72e144a72530ffbee765a392b24593dafd06cc098f8410b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:31b161f9bf9028a1a62cd2eef6abda01257d8889bdc3159822642549a75e3429
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9c85c25d5392817d17b68714dca6195ad53255d43e9c65b021d75b12d97e0579
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a6e2fe24ad52913e2160f85b9fe665a4e4c4d79b164b4b899f69dd9d92f999e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cc20270212f01fc231ff788e20b0476b8228c440bfc40430c0836d5eb702a042
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1374cbdc4ecf38e4b4226328f532e830b8932b34d9449bf2fbb36e60d0310521
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bd336715332f776a72aaf5c3ca52e62dd53a9e8521c1feb6e135cc868d694a39
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7481c0cf3b579c8d98d94d4aa0334ca94bf2ccfcd25ad70b57c42b22bbf4bcee
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4635550222775fc75737be5e5c0359303e3a23c4d1e1c296d7408cd15ad94aad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7daaf1368ab2c25694cdea0a598e1de59247b04a4b4b4b8348ed6edbaa194f9d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fe4a40006d1e8439423de1417cc9bc55716d790b882485074bb2a5ec68d7d1ad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5b3de46526d169fcd1318354545b3a18861843fa2022a23ed04d795493825287
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acc083f7dde1fbc557d858d03fc26b68b2056997fc1a457e4695650214d5b303
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0b0e0588c1aadf0c3826cd584bd2b3a7d6781882dcb5c5e037b47ff5cb84509b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3bf017b2b2506724500f5e29017d1b956825dd0f61a7c0049b0ff084c3554b5b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce1209ca6cc8724baee981c97c66e2852e512423bc0c0a16cdd80a6838b54031
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e82684b828256b3c2814930210031b3b70bc818514d4359a25f87ffdbf22be6e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:19f36f53f7da67755eb86da77098c5eeb84b381f9c18aa29ba573793dc498564
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f5a0ccc02dc9e938e484d5b5282ff650d1890d5f754c30a0c0d717989ed5d600
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9d0bfb642090ff55a3cc3da0f72663cae18063622a4ec785ba99bc71ef4db116
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f045fe5037d44aab1a0b2209241b3931ec7cb9897cb3c6a96a9f8d5280e55dd2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:44a0f40a6c46ba08297b482f3d96defe87b181c944803e5381c9acc42cabf97f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db07461824e73410456ef540b0bce3c7fe550aa67a73d6174300ca74b8f75276
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0754f7901feea9017bc5041cb91b7ac1cdbca10c2064f37c3648a2c395f24f83
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:28ae9661b38a72decf413c6a6c3758e61098debf5162197c27a12a070c7a096d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:743cda1eda6e273c798e3405a9a97e07dd6aa5ad4d5a52789be3b0f39a32a039
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:91f7e59f9e35a0395756eafed906963c535e886125294d671d930ce959a8f511
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d78c2e0709214f96794564f13386861f4b541e55b5bd316ad495cf104f7d40b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da3bd2e9ff2b08d5bb86b1f5faa95a865f85c5e8912df00b6b79190c1e518a2f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f3ffebd812881a830ff46c860984814dd61edcd51b478ebd3667759fb7710682
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5d6901b6b6fd6594d42bfc1842d050c7cffcb0867a5591a67469f0f860b1f21b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e8f26fd132affbaeb8ef6b1807a646c83b301e5dc334a592dfb2cf8dc3dbaf10
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c74cf927a2b2c5098ad74ad0880996f8b042186a9911d888a231ef5b9e2a0715
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5a54d564e08baefc2786f4b742b2507d31984664cd00333406aa7a3aa5b9c913
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:310310515625f099a928545865f7096997871ee4a16650a01c76c3799a18b684
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a0faa52a999289b308dc4caf050c6d9fe2f42e112aa30e4af282ed1f21c5c346
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a59cafff8d4ff2c62204ece15880117011c30978823c232902c7182d747ec827
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1e5d85f5db10fa79d09263db90daa44549bcc3b025e9f3f98cf0d4e51766394b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6ca927f137be28481e3a226e6c6bcb7c5ac27664984f17ffc6c1419cd7d08eb7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e50aab39722af22ccbf3d1db19e37972d52b0fc9a40998a64618dc966b2bac57
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d978678b3c254abd52f69f226b0aacc4b02c2aaca328c54ef10f6f9bb453582d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6408f89b0a88ec154f7ee98c314be574971c4c9aaf906b473d10124d822889be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:468b0a40a33be1dd5a2a89ce745d7e9012bc7371f2f01cc0f8f12c0ac887db43
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b06bd7f3802e8fa2af14bb0b07804cad4cc0fbbc97da18d09a9106bf074ed6d8
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f438fbbf9c87ca0d54c170e2c5105b1e265182e1c2ede48adea370e2be9dc895
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d8cfb221f67516675fdfde061f6ec5dfee1224ca7eb3d8be375c27d483f306af
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e151bded670f4cbd165a8ef864324805667d3e84ce3ace0cdf8f9aad087fccf1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cbc74889e9ca5edbd5aaa243d7649bca0c4da2719239f8d0ec9827868112a771
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a6f1c43486d044fed9f59406118f429c8037b389bd8c8d7a1c7aeb89ed2661f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2af60025ecb3bb933b61328be4c2b2bfd0e7d26f53b72430629208505e45c6d2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:79d688bebe0643ef0dd6ba7108cdb34c0442470c0c20ae0df88b5d02a3476cd0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9097e135159f46396ef4649f724d61bb5b9113c488f989d404586fbb7486fd85
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d14f77f70aea0422a906510c1ac2652ff76e131ed705cb002f80e13e887b0cd9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0fea6ed8a78ea624240ee6ffb3573d5cf1115186aa9180fcd7d1273351deaaa3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e18d73af3257c8cdf00245f2f718b56667524c1cc69d3db28e9b868a539ae16f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:dd521d7bad62da054eec54179be75c643a5492d981e11c0e66db41902cd2182c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:363d8feb560324a848c5771edd224dc19020013af7641979ee81ae8ee536d7e3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7b907d03b001b758d8a4b5e898725c73dda2ec2aa4b7425b685320b947154d11
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4b56c0311af14220a466aa11f25fd10fa38280825539bd02215c945bbafb861c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:42c9930d6861f6a2147ed6e01827eddd730279966763ec68868567a19cc164ab
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:740ba10f1ef0a6c5fa6ad31ab18855fc75f8e024ce4c23b0676c5a9c5b62cd75
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8e8c081c231b5b15eca483ceec16ea9190ba73aae6e421865162446eb584519a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:56109e57ee0379cf48644bcf8226a9238e01139cacc5499002c99f973f121911
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:742a6f69a3a209a7080a1e911fd96467d2a91aacb3f91d6be80a6a6e2e54b2ae
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0e2c7023f1f97acc20c46a4944635a9c5f57bb9bb3dd534cf41fea1a9cfdfdbe
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf7d3bea47638b763b312c7c11ca034a7ac62d993d75a13edceec74799b97e52
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:226d7f1c3f0d70ac6532ad9490bebcdef44a40959b6aa7c11b0f2d1383a46217
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6cd9f9b9fcb6e8f9a931a6ad01cb66329a74d40b22f9029fbc50e57248bad7cb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1ce71e80c0cd0e0063eba275ab91f77982f2e92a8aef8e89cab99b93feab0a88
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1b6d934a7b6c42a61e5ad51380e69af341a575b42205299fd58bc0c6ff306f6a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b046aa0a42dae992a03685ed82a20835532d6d4bf048e2366ff7382be4b3138
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acbfee4492327a720fe7da76929a86192d8af2442bbb76f0a089f6a81a890be3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:818f592c50d4ed5599faa053344f7ecd963003352f1faa3b72dc0f5e00a0e78b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4e5f7df1dc1e1db398c36d590431e7e5782897b209972d8e9e4671971c10d1b6
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5458a8fd89aa664b60e59c35bd933975cd37d18e19b11589e4ca0c2844ea0ed4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2a128b570352a5a7f82b27353f9b774caccce83e0e7bb2673e4802d96ad0bd51
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2171e1c75d0c665fe33e1ea334e605604c51de6dbc22d88a3a0ac604259c448f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ec6eb7c28c213cc83b2d7919cd87988f9a07f12276eb7351d0915f1567a5b095
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e5814748c5ab229a1dea5662a007063c0df06fbbfcfe041fd9a6df9ef67e07cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:82d8104a37349c513447e68ea3790efb256370632353176e5fc13c2e8df1bd44
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b78a4d3864e16d270979d9d7756012abf09e6cdb7f1eb4832f6497c26281099
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9f42e929f2cc7e9b6d24dd61e19c5d93dcfeabd34b9c8b935afbe78577efb18b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8fc8de4f3671a73f9aa747e5b23ec973d192fff6f2869c38985cd656063c1e3d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:02e1ef01a023da0eee653f63b314786fd66a78d184a97a314422f6f9c3ef8e7e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:65923c37fecda2d3174212fa9d9a6f34241d065389336ea69a7be99ba16cedc2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bed27f3f2c2d7627d1af4500d7ea0302da5c3fd3d875a6480fb6d98da481312e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:10b9df672260c023dfb37166e156c0d7e0e13144dc3e485af922e8151fd3ab05
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f05824bb8eaa5a93e54059449eaa236dd83d729049ad00065ef3242f5f74c91e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:87db6e079b96fb392d98ef5fc7063836bff10717efd75061016ce688be0ba38e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:29abb169ffefe36f6d69edd438a7135ee23059cedab8f87ad3b2d52361b96b35
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a4f932c3e9ff3aeb70ba1f7a92aa4642bd14cbb7d37d04ff750ed517b916cdb9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08496b84b3cc691909227be17e67bcd6f4d5185a93712f44d8a0985bd2c5fcd7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f53b07212320c6939d3bb4ddd4e6a0bce311ab5b63858eb1a8728531f0dbd510
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b0057de903654ea0315a2c96affcc8f50a92ede8ad121ed30edb3aaf0fb7b554
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d0014b1c19890b9e8ef1a4f3efe7eb5b269be5610ba6ababd4258b52d2f85e8a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:aeac9cd8e8acdbd2b0c339c48b85079c159b04f0ade24df8e3f9cae3f48b4c70
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6366ba29f301a82215c05e60d78e1de66bcee85440ce2fc3c692712d4da32685
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8a70ba12f610ac0e62066e417a0833bc1d01e91ac99d0dd794a2c0cd2a350164
        version: 2.18.2
      entries:
      - name: cockroach-operator.v2.18.2
        version: 2.18.2
      name: stable
    defaultChannel: stable
    packageName: cockroachdb-certified-rhmp
    provider:
      name: Cockroach Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      control-plane: odfmo-controller-manager
      full_version: 4.15.4-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-compose-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-compose-operator.v4.15.4-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "labels": {
                    "control-plane": "odfmo-controller-manager"
                  },
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              }
            ]
          capabilities: Basic Install
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.4-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          This is a CSV used for mirroring the entire product properly. Do not install it please.
        displayName: ODF Compose (DO NOT INSTALL)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: branto@redhat.com
          name: Boris Ranto
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-operator-bundle@sha256:3773158dbd141530cd851f36c3c8a5f54dc0892d003633118e123f5862923666
        - registry.redhat.io/odf4/odr-cluster-operator-bundle@sha256:4c7049b6137b3d49fa9b5e51e0cf7cd430d70c8790e0ed017e6bd811fcc506c8
        - registry.redhat.io/odf4/odf-csi-addons-operator-bundle@sha256:b2b8144c34df667b13987773b11df2fc8ac96d418999292303ef2c03cd3c79d6
        - registry.redhat.io/odf4/mcg-operator-bundle@sha256:9ce0ab75f56ebeeac63aefd0f48cba671da3377ac628830b3c03be9569f8c87a
        - registry.redhat.io/odf4/ocs-operator-bundle@sha256:1e91c5c6c122fdccab52c37ff2cef76317af80970dbdc06aec752115117b2fa0
        - registry.redhat.io/odf4/ocs-client-operator-bundle@sha256:8904c171ea317249e9a19c3a7638f688357a0b238582330b1bddc23a11a9925e
        - registry.redhat.io/odf4/odf-multicluster-operator-bundle@sha256:f7e1c75ee6b1a62fddbcc1e6a718a9cbd5ecf364e4ef10f29da7a15bc82d5b17
        - registry.redhat.io/odf4/odf-multicluster-rhel8-operator@sha256:861e56167698558109197b515aff93691f8d0fdba19311be18a32512f6896139
        - registry.redhat.io/odf4/odr-hub-operator-bundle@sha256:862a9d983013c16b7dfd5194ab5752c26a473ca5a4575bbcd80be59f97f8ac27
        version: 4.15.4-rhodf
      entries:
      - name: odf-compose-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      name: stable-4.15
    defaultChannel: stable-4.15
    packageName: odf-compose-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: unsupported
      operatorframework.io/arch.ppc64le: unsupported
      operatorframework.io/arch.s390x: unsupported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: mta-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mta-operator.v6.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:f5d10cd2528bebf5094c7744e47d9de6fe827fd09024c5ffc0e1b74f1a6cc3b4
          createdAt: "2023-04-26T14:04:20Z"
          description: MTA contains tools that support the modernization and migration
            of applications to Kubernetes and OpenShift
          olm.skipRange: '>=0.0.0 <6.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/tackle2-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is a collection of tools that supports large-scale application modernization and migration projects to Kubernetes and OpenShift.


          Migration Toolkit for Applications allows users to maintain their portfolio of applications with a full set of metadata and to assess their suitability for modernization leveraging a questionnaire based approach.


          ### Install

          By default, MTA requires a RWX capable storage class for caching purposes, this can be disabled via the rwx_supported parameter in Tackle CR.

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-windup-addon-rhel8@sha256:a4200fc0a39df67eae31c53ee9ae0c05769fb410d0312ebd0bbf8bf1567b2a48
        - registry.redhat.io/mta/mta-hub-rhel8@sha256:c75e1d9d3307d95be392a53440d9123601a6ada3693e3f2573e03022565e4339
        - registry.redhat.io/rhel8/postgresql-12@sha256:3a69efa85af6a6f95665eca69085b007fa1fcb645a37f356895aeb4c9527a621
        - registry.redhat.io/mta/mta-pathfinder-rhel8@sha256:a66dd030eb99207b6aaacc6b7c7cb5182901da836ebf50373c14f8e5ecb7b642
        - registry.redhat.io/mta/mta-ui-rhel8@sha256:de6cb95ce3663922ecc725b186876b043a9c7794e95510459ffbcba7d668fc26
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:f5d10cd2528bebf5094c7744e47d9de6fe827fd09024c5ffc0e1b74f1a6cc3b4
        - registry.redhat.io/mta/mta-admin-addon-rhel8@sha256:812b8dc621a06bb99cf2c59abb8257da3ceeb165c563e14d258e69803a557cc8
        version: 6.0.3
      entries:
      - name: mta-operator.v6.0.3
        version: 6.0.3
      - name: mta-operator.v6.0.2
        version: 6.0.2
      - name: mta-operator.v6.0.1
        version: 6.0.1
      - name: mta-operator.v6.0.0
        version: 6.0.0
      name: stable-v6.0
    - currentCSV: mta-operator.v6.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:6da71d4caecfbe049e63576d4325d4919002196da2e875f79703d8ff0d5a1705
          createdAt: "2023-10-26T15:29:26Z"
          description: MTA contains tools that support the modernization and migration
            of applications to Kubernetes and OpenShift
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0 <6.1.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/tackle2-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is a collection of tools that supports large-scale application modernization and migration projects to Kubernetes and OpenShift.


          Migration Toolkit for Applications allows users to maintain their portfolio of applications with a full set of metadata and to assess their suitability for modernization leveraging a questionnaire based approach.


          ### Install

          By default, MTA requires a RWX capable storage class for caching purposes, this can be disabled via the rwx_supported parameter in Tackle CR.

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:6da71d4caecfbe049e63576d4325d4919002196da2e875f79703d8ff0d5a1705
        - registry.redhat.io/mta/mta-hub-rhel8@sha256:32624f2bddfe8b6b9c6c1c3f2be541c8c61be29680d13b6d446e8874bdf8e219
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/mta/mta-pathfinder-rhel8@sha256:1df881da1e20c2501caf53e1cf59b6b366281d842f0948ca5f91694ec1db829d
        - registry.redhat.io/mta/mta-ui-rhel8@sha256:3845d6d246003ebeb49bcf1d95d6be2eb9cd05a280a5bf8320b201e878a41427
        - registry.redhat.io/mta/mta-windup-addon-rhel8@sha256:6bcaf8e8b63e169652d6568d77e91a289084e621a586105a2d429b5726569058
        version: 6.1.4
      entries:
      - name: mta-operator.v6.1.4
        version: 6.1.4
      - name: mta-operator.v6.1.3
        version: 6.1.3
      - name: mta-operator.v6.1.2
        version: 6.1.2
      - name: mta-operator.v6.1.1
        version: 6.1.1
      - name: mta-operator.v6.1.0
        version: 6.1.0
      name: stable-v6.1
    - currentCSV: mta-operator.v6.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:6938c3fc6dd8fb032e1765fc11d9586dbfd1aaba4b423b0d87da4120ee0bb60d
          createdAt: "2024-06-10T22:12:58Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0 <6.2.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-windup-addon-rhel9@sha256:d1f4939b92dd13c15cb7344ec47a571669f6d108dc2e7d795ce29350f2cb3838
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:6938c3fc6dd8fb032e1765fc11d9586dbfd1aaba4b423b0d87da4120ee0bb60d
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:e59725832547e80cab58077a1b6e42c313a4057621defa841f35c0a7627f149d
        - registry.redhat.io/rhel8/postgresql-12@sha256:c4716bee5e2cf2e43698bc99e5aa42f16a0f2d68c1fbdcd3602696bf88e408dc
        - registry.redhat.io/mta/mta-pathfinder-rhel9@sha256:a2a8b6d204d99494a6a005e8cb1fbd92afdc84dc01e4c1de4ad7d8798bd83741
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:c35f3ed429a7d351cb8c2d07600b4ad35615a5ba08e5cc9f8b9bd504aea3f7ee
        version: 6.2.3
      entries:
      - name: mta-operator.v6.2.3
        version: 6.2.3
      - name: mta-operator.v6.2.2
        version: 6.2.2
      - name: mta-operator.v6.2.1
        version: 6.2.1
      - name: mta-operator.v6.2.0
        version: 6.2.0
      name: stable-v6.2
    - currentCSV: mta-operator.v7.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:268ecf5efc51af71f9b2bc2d01b410f8cfcf81a948b82d2d476fef6c0cd3e033
          createdAt: "2024-05-10T19:37:30Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:268ecf5efc51af71f9b2bc2d01b410f8cfcf81a948b82d2d476fef6c0cd3e033
        - registry.redhat.io/mta/mta-windup-shim-rhel9@sha256:e2ea89055936d7ad229b5e890c6ba234c6cb22d8b688515ed4d6a34848efea83
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:d554cc36748fc05e30eb10e6fe816ff6aab2fe07b61cc60e9d93de5c3cf17748
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:0e1ec99af5666a060176adc52612f20209e3398a9a5a2aa1dcff11ef260b2006
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:1b5e016046d93b6a2674d63f08e0a46f8050e785bd58fd94ba7a70487efe7e2a
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:18c611415589a093e16e9c2f2fba9f3c83d9514f79c8e4fde9704ef0679ea4d5
        - registry.redhat.io/mta/mta-analyzer-lsp-rhel9@sha256:0c8f59f35a950ea8b45f16cf71362128ccc335b58a996a099384fd737dd6ff90
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:af879d8b9b5e396a5308fb675a1d6dba6d70151b9bdbe7829362e081a6adab1c
        - registry.redhat.io/rhel8/postgresql-12@sha256:da619f6617d99790b4fe00fdb23ab4aea0cdb0592590162d51b3ef94aca01a7b
        version: 7.0.3
      entries:
      - name: mta-operator.v7.0.3
        version: 7.0.3
      - name: mta-operator.v7.0.2
        version: 7.0.2
      - name: mta-operator.v7.0.1
        version: 7.0.1
      - name: mta-operator.v7.0.0
        version: 7.0.0
      name: stable-v7.0
    - currentCSV: mta-operator.v7.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:08b33b44140397785f8aa1b736796e09cf6e49ac30cf7c386a2fbd8b678d47b1
          createdAt: "2024-09-16T23:45:42Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.1.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:16cb28d683159ba71a5d7c233404ac1d6b7a07a374b46d24effe551a13a596c3
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:23c97ac122eb7cd1e953df8da968da420f6db8708baf772759517aca4929748b
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:e4e70201df23c9e1168646d3c6013a8ffa22481a85b07df491f240310c13ebe1
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:08b33b44140397785f8aa1b736796e09cf6e49ac30cf7c386a2fbd8b678d47b1
        - registry.redhat.io/rhel8/postgresql-15@sha256:16faf1ee82382263408f713ee4002f47ff13c82d4545e8a69d5b0ebd1d289330
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:30a3b1d5a70dec74a9253ca34f18bd8f2cdfaa68b85b0b25dc395158c2596330
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel8@sha256:582f28e609c1a0466d30c5586d6b6023f1b1de3c28f76cabb498322914ab24d4
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:ac38f1c3d34bc718928eb20743835c99783d846f63a3197d13839ea22f1634c8
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:4ad19b50cdd6be21e010155f2026abf0dc865e72aac1353e35bbb5eeb8320a4e
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:ee2c8ce2af0be097353758d9ee7eaa7e6b3fcd5f2bf5e04280e450e2b619d2d8
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        version: 7.1.1
      entries:
      - name: mta-operator.v7.1.1
        version: 7.1.1
      - name: mta-operator.v7.1.0
        version: 7.1.0
      name: stable-v7.1
    - currentCSV: mta-operator.v7.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:917d6cd534c158f6c4a7e9c83f0c66fa85ba81037c352d8a039e198581126a7f
          createdAt: "2025-06-17T16:04:15Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.2.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:917d6cd534c158f6c4a7e9c83f0c66fa85ba81037c352d8a039e198581126a7f
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel8@sha256:c0643c05833dd1b70ed361bf56c46f56ed6536aac3106b92f0d85b83da1b9d15
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:10ef4b218d603432508b5e1f3685fb32a80cbce6c327e35952cdcccb0465468b
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:9d6c02f6463771e747a30126de95fc0728fd95fff4ddfa6c70d828c55e0b6790
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:2dfc51d5398a90e73273b3e332f70d7ed94bac05e12b142fba22104ecc929436
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:b10fdc78c6f453a34a3d306313dc29fcb3152688a6a1d54cd236d00d9fab88af
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:4cde2c2df482293bd4d6b86a344336e13d2f9f81b29323698f603e656b14b190
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:70ff3b5e429a8fe0c68752ee05c47ead885b59bf9c8d7235337801270e7fc1de
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:c9447659195b9eebb8f8a265fd5052ca9b2611d57080017448297a33409863c6
        - registry.redhat.io/rhel8/postgresql-15@sha256:dd4aec2775da1cf187c0839384cb16d69268e9781d4330874bbc865f98153a7a
        version: 7.2.3
      entries:
      - name: mta-operator.v7.2.3
        version: 7.2.3
      - name: mta-operator.v7.2.2
        version: 7.2.2
      - name: mta-operator.v7.2.1
        version: 7.2.1
      - name: mta-operator.v7.2.0
        version: 7.2.0
      name: stable-v7.2
    - currentCSV: mta-operator.v7.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel9-operator@sha256:6abd6541a955543207c1ff4b903928debcc89dd6ea3d7aeaa3e47465849f7b37
          createdAt: "2025-08-15T16:32:16Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.3.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-15@sha256:854710b428c4c1132c448cbe9ad30087a8a4b3c33ae9c0c3a4de21a15c0ae642
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel9@sha256:9748d954ad8654ac203257286497ccf19cfdfa211fee40449cbe732ff48c1176
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:32328fc0459099ef1de0f92e064c15d29a46d88dfe2f54c88f822d44ab82e3f4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        - registry.redhat.io/mta/mta-rhel9-operator@sha256:6abd6541a955543207c1ff4b903928debcc89dd6ea3d7aeaa3e47465849f7b37
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:b2ff154a4b4ead22311005b7115c02346928b5ebf77d776ef909ad4af1f350bc
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:6118c1db15667851f26eb2acf6060e045ed9065459e3e649bbdd0e2625d825bd
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:072f676323e9abcb2ff22eb5e7a70f6681e9d885d8666ddd81060fbad54c443e
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:8ca00460b2fb64583a247f5b906c19743b7732b60ff8215f1baec12a766c61d1
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:80aa13813fe121da49e166d81d09d63749393f3375c3b96d762d0b09c6685ca9
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:ef57a58e609e244333d35681a164a1d3bf382c0eee37d33c993b018439814198
        version: 7.3.2
      entries:
      - name: mta-operator.v7.3.2
        version: 7.3.2
      - name: mta-operator.v7.3.1
        version: 7.3.1
      - name: mta-operator.v7.3.0
        version: 7.3.0
      name: stable-v7.3
    - currentCSV: mta-operator.v8.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel9-operator@sha256:83c2b804667d28f782d64132a0c393194083d10c820b88df49c2503d3351fd8b
          createdAt: "2025-09-26T03:03:31Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.3.0 <8.0.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Schema
            displayName: Schema
            kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-platform-addon-rhel9@sha256:dcb35178810b7eddc2f2713434d377003f8f4fb7022a40f6c1926b74b13fe23a
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:a77ee9b76225b6cca667fa3402492b0845587fe5e400fa8d141b5ab636c56811
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel9@sha256:2a6f0b392b86fe16244d5bf9fcb84ed9770a22329a96e9861f03872b7209fb04
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ef10b2ecbfa0c4ac8568af9ee64ee4a554e3fdc9089ee5683b6c2c3a74caafbe
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:616d04cfbab92341157c42eee0683f31867d81c5a289723da8da8208d1054774
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:aaf7bf10b1af107ad55289caab2cf9dd052e69211968d2e8d05c1d25f21ed2b3
        - registry.redhat.io/mta/mta-rhel9-operator@sha256:83c2b804667d28f782d64132a0c393194083d10c820b88df49c2503d3351fd8b
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:b1b38e218ea37b529f344eaa1f0f3b55b716ad7953e852c19f34f4a25ab6c5a7
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:af2c1790f2b98ebcfadb5d6bdd43cb9e7e96cdabd5b653d48d0fb67901781c49
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:460d1d45ed44ae9328ed93c5fed240e154549bbec662dc1fc55569ff7256aca9
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:8ae0887fa27d6f4281e5bb9b629bf6b64a9c460948a002839ba42da9b25b5521
        - registry.redhat.io/mta/mta-solution-server-rhel9@sha256:e9d48d11ae203108229d67bc96721daa1c3bb10160b9d6d957cd2e0ba68a2ef7
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        version: 8.0.0
      entries:
      - name: mta-operator.v8.0.0
        version: 8.0.0
      name: stable-v8.0
    defaultChannel: stable-v8.0
    packageName: mta-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: aws-efs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-efs-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"aws-efs.managed.openshift.io/v1alpha1","kind":"SharedVolume","metadata":{"name":"sv1"},"spec":{"accessPointID":
            "fsap-0123456789abcdef", "fileSystemID": "fs-0123cdef"}}]'
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/app-sre/aws-efs-operator:9c87267
          createdAt: "2021-12-20T19:51:50Z"
          description: Manage read-write-many access to AWS EFS volumes in an OpenShift
            cluster.
          repository: https://github.com/openshift/aws-efs-operator
          support: https://github.com/openshift/aws-efs-operator/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Shared Volume
            displayName: SharedVolume
            kind: SharedVolume
            name: sharedvolumes.aws-efs.managed.openshift.io
            version: v1alpha1
        description: |-
          ## About the managed application
          AWS provides Elastic File System, an NFS-based shared storage solution. A couple of mechanisms exist to enable mounting such a file system in containers. Curating these options and making them work in a cluster is nontrivial.
          ## About this Operator
          The aws-efs-operator minimizes the barrier to entry for container use of EFS by:
          - Installing the [AWS EFS CSI Driver](https://github.com/kubernetes-sigs/aws-efs-csi-driver) and supporting infrastructure.
          - Reducing the identification of a shared storage resource to a simple Custom Resource Definition with only two fields.
          - Creating the necessary PersistentVolume and PersistentVolumeClaim resources that can simply be referenced from a Pod specification.

          More information can be found in the operator's [source repository](https://github.com/openshift/aws-efs-operator).
          ## Prerequisites for enabling this Operator
          An OpenShift cluster backed by AWS.
        displayName: AWS EFS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AWS
        - EFS
        - NFS
        - Shared storage
        links:
        - name: README
          url: https://github.com/openshift/aws-efs-operator/blob/master/README.md
        - name: Design
          url: https://github.com/openshift/aws-efs-operator/blob/master/DESIGN.md
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/app-sre/aws-efs-operator:9c87267
        version: 0.0.8
      entries:
      - name: aws-efs-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: aws-efs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: opdev.github.io
    name: spark-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spark-helm-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apachespark.apache.org/v1alpha1",
                "kind": "Spark",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/install-disable-crds": "true"
                  },
                  "name": "spark-operator"
                },
                "spec": {
                  "affinity": {},
                  "batchScheduler": {
                    "enable": false
                  },
                  "controllerThreads": 10,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/googlecloudplatform/spark-operator",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "ingressUrlFormat": "",
                  "istio": {
                    "enabled": false
                  },
                  "labelSelectorFilter": "",
                  "leaderElection": {
                    "lockName": "spark-operator-lock",
                    "lockNamespace": ""
                  },
                  "logLevel": 2,
                  "metrics": {
                    "enable": true,
                    "endpoint": "/metrics",
                    "port": 10254,
                    "portName": "metrics",
                    "prefix": ""
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podMonitor": {
                    "enable": false,
                    "jobLabel": "spark-operator-podmonitor",
                    "labels": {},
                    "podMetricsEndpoint": {
                      "interval": "5s",
                      "scheme": "http"
                    }
                  },
                  "podSecurityContext": {},
                  "rbac": {
                    "create": false,
                    "createClusterRole": true,
                    "createRole": true
                  },
                  "replicaCount": 1,
                  "resourceQuotaEnforcement": {
                    "enable": false
                  },
                  "resources": {},
                  "resyncInterval": 30,
                  "securityContext": {},
                  "serviceAccounts": {
                    "spark": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "sparkoperator": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    }
                  },
                  "sparkJobNamespace": "",
                  "tolerations": [],
                  "uiService": {
                    "enable": true
                  },
                  "volumeMounts": [],
                  "volumes": [],
                  "webhook": {
                    "cleanupAnnotations": {
                      "helm.sh/hook": "pre-delete, pre-upgrade",
                      "helm.sh/hook-delete-policy": "hook-succeeded"
                    },
                    "enable": false,
                    "initAnnotations": {
                      "helm.sh/hook": "pre-install, pre-upgrade",
                      "helm.sh/hook-weight": "50"
                    },
                    "namespaceSelector": "",
                    "port": 8080,
                    "timeout": 30
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "ScheduledSparkApplication",
                "metadata": {
                  "name": "spark-pi-scheduled",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "concurrencyPolicy": "Allow",
                  "schedule": "@every 5m",
                  "template": {
                    "driver": {
                      "coreLimit": "1200m",
                      "cores": 1,
                      "labels": {
                        "version": "3.1.1"
                      },
                      "memory": "512m",
                      "serviceAccount": "spark-operator-spark",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "executor": {
                      "cores": 1,
                      "instances": 1,
                      "labels": {
                        "version": "3.1.1"
                      },
                      "memory": "512m",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "image": "gcr.io/spark-operator/spark:v3.1.1",
                    "imagePullPolicy": "Always",
                    "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.12-3.1.1.jar",
                    "mainClass": "org.apache.spark.examples.SparkPi",
                    "mode": "cluster",
                    "restartPolicy": {
                      "type": "Never"
                    },
                    "sparkVersion": "3.1.1",
                    "type": "Scala",
                    "volumes": [
                      {
                        "hostPath": {
                          "path": "/tmp",
                          "type": "Directory"
                        },
                        "name": "test-volume"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "pyspark-pi",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "driver": {
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "serviceAccount": "spark-operator-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "gcr.io/spark-operator/spark-py:v3.1.1",
                  "imagePullPolicy": "Always",
                  "mainApplicationFile": "local:///opt/spark/examples/src/main/python/pi.py",
                  "mode": "cluster",
                  "pythonVersion": "3",
                  "restartPolicy": {
                    "onFailureRetries": 3,
                    "onFailureRetryInterval": 10,
                    "onSubmissionFailureRetries": 5,
                    "onSubmissionFailureRetryInterval": 20,
                    "type": "OnFailure"
                  },
                  "sparkVersion": "3.1.1",
                  "type": "Python",
                  "volumes": [
                    {
                      "hostPath": {
                        "path": "/tmp",
                        "type": "Directory"
                      },
                      "name": "test-volume"
                    }
                  ]
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "spark-pi",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "driver": {
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "serviceAccount": "spark-operator-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "gcr.io/spark-operator/spark:v3.1.1",
                  "imagePullPolicy": "Always",
                  "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.12-3.1.1.jar",
                  "mainClass": "org.apache.spark.examples.SparkPi",
                  "mode": "cluster",
                  "restartPolicy": {
                    "type": "Never"
                  },
                  "sparkVersion": "3.1.1",
                  "type": "Scala",
                  "volumes": [
                    {
                      "hostPath": {
                        "path": "/tmp",
                        "type": "Directory"
                      },
                      "name": "test-volume"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/spark-helm-operator-controller:0.0.2
          createdAt: "2023-07-14T17:57:44Z"
          description: An experimental operator that installs the spark on k8s operator
            from its helm chart.
          operatorframework.io/initialization-resource: "{\n  \"apiVersion\": \"apachespark.apache.org/v1alpha1\",\n
            \ \"kind\": \"Spark\",\n  \"metadata\": {\n    \"name\": \"spark-operator\"\n
            \ },\n  \"spec\": {\n    \"affinity\": {},\n    \"batchScheduler\": {\n
            \     \"enable\": false\n    },\n    \"controllerThreads\": 10,\n    \"fullnameOverride\":
            \"\",\n    \"image\": {\n      \"pullPolicy\": \"IfNotPresent\",\n      \"repository\":
            \"ghcr.io/googlecloudplatform/spark-operator\",\n      \"tag\": \"\"\n
            \   },\n    \"imagePullSecrets\": [],\n    \"ingressUrlFormat\": \"\",\n
            \   \"istio\": {\n      \"enabled\": false\n    },\n    \"labelSelectorFilter\":
            \"\",\n    \"leaderElection\": {\n      \"lockName\": \"spark-operator-lock\",\n
            \     \"lockNamespace\": \"\"\n    },\n    \"logLevel\": 2,\n    \"metrics\":
            {\n      \"enable\": true,\n      \"endpoint\": \"/metrics\",\n      \"port\":
            10254,\n      \"portName\": \"metrics\",\n      \"prefix\": \"\"\n    },\n
            \   \"nameOverride\": \"\",\n    \"nodeSelector\": {},\n    \"podAnnotations\":
            {},\n    \"podLabels\": {},\n    \"podMonitor\": {\n      \"enable\":
            false,\n      \"jobLabel\": \"spark-operator-podmonitor\",\n      \"labels\":
            {},\n      \"podMetricsEndpoint\": {\n        \"interval\": \"5s\",\n
            \       \"scheme\": \"http\"\n      }\n    },\n    \"podSecurityContext\":
            {},\n    \"rbac\": {\n      \"create\": false,\n      \"createClusterRole\":
            true,\n      \"createRole\": true\n    },\n    \"replicaCount\": 1,\n
            \   \"resourceQuotaEnforcement\": {\n      \"enable\": false\n    },\n
            \   \"resources\": {},\n    \"resyncInterval\": 30,\n    \"securityContext\":
            {},\n    \"serviceAccounts\": {\n      \"spark\": {\n        \"annotations\":
            {},\n        \"create\": true,\n        \"name\": \"\"\n      },\n      \"sparkoperator\":
            {\n        \"annotations\": {},\n        \"create\": true,\n        \"name\":
            \"\"\n      }\n    },\n    \"sparkJobNamespace\": \"\",\n    \"tolerations\":
            [],\n    \"uiService\": {\n      \"enable\": true\n    },\n    \"volumeMounts\":
            [],\n    \"volumes\": [],\n    \"webhook\": {\n      \"cleanupAnnotations\":
            {\n        \"helm.sh/hook\": \"pre-delete, pre-upgrade\",\n        \"helm.sh/hook-delete-policy\":
            \"hook-succeeded\"\n      },\n      \"enable\": false,\n      \"initAnnotations\":
            {\n        \"helm.sh/hook\": \"pre-install, pre-upgrade\",\n        \"helm.sh/hook-weight\":
            \"50\"\n      },\n      \"namespaceSelector\": \"\",\n      \"port\":
            8080,\n      \"timeout\": 30\n    }\n  }\n}    "
          operatorframework.io/suggested-namespace: spark-operator
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys a scheduled spark application based on the spec provided
            displayName: Scheduled Spark Application
            kind: ScheduledSparkApplication
            name: scheduledsparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - description: Deploys a spark application based on the spec provided
            displayName: Spark Application
            kind: SparkApplication
            name: sparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - description: Deploys the spark operator based on the google cloud platform
              helm chart
            displayName: Spark Operator
            kind: Spark
            name: sparks.apachespark.apache.org
            version: v1alpha1
        description: An experimental helm based operator that wraps the spark on k8s
          helm chart from google cloud platform enabling installation and workload
          management of Apache Spark
        displayName: Spark Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - big data
        - map reduce
        - data science
        links:
        - name: Spark Helm Operator
          url: https://github.com/skattoju/spark-helm-operator
        maintainers:
        - email: skattoju@redhat.com
          name: Sid Kattoju
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: opdev.github.io
        relatedImages:
        - quay.io/opdev/spark-helm-operator-controller:0.0.2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 0.0.2
      entries:
      - name: spark-helm-operator.v0.0.2
        version: 0.0.2
      - name: spark-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: spark-helm-operator
    provider:
      name: opdev
      url: opdev.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-test-core
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-test-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Testcore",
                "metadata": {
                  "name": "testcore-sample"
                },
                "spec": {
                  "agentController": {
                    "port": 8080
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "dataseqService": {
                    "ingressEnabled": true,
                    "port": 8080,
                    "providers": {
                      "loadCore": {
                        "enabled": true
                      }
                    }
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "ingress": {
                    "backwardsCompatibilityEnabled": true
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "systemMonitor": {
                    "enabled": false
                  },
                  "trafficController": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:35:34Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Testcore
            name: testcores.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Test Core Operator
        displayName: Keysight WAP Test Core Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Test Core
          url: https://keysight-wap-test-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-test-core@sha256:9e2d81061abff49112f908e71dace03a748e283f32597aa22ad58471fff30caf
        version: 0.1.21
      entries:
      - name: keysight-wap-test-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-test-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp
      provider-url: https://www.netapp.com/
    name: netapp-spark-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netapp-spark-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.netapp.com/v1alpha1",
                "kind": "SparkOperator",
                "metadata": {
                  "name": "sparkoperator-sample"
                },
                "spec": {
                  "spark-operator": {
                    "affinity": {},
                    "controllerThreads": 20,
                    "disableExecutorReporting": false,
                    "hostNetwork": false,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "public.ecr.aws/f4k1p1n4/spark-operator",
                      "tag": "108810d-ubi9-rel20240916"
                    },
                    "imagePullSecrets": [],
                    "istio": {
                      "enabled": true
                    },
                    "jobAffinity": {},
                    "podLabels": {},
                    "pprof": {
                      "enable": true,
                      "port": 6060,
                      "portName": "pprof"
                    },
                    "rbac": {
                      "createClusterRole": true,
                      "createRole": true
                    },
                    "resources": {
                      "requests": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      }
                    },
                    "securityContext": {},
                    "serviceAccounts": {
                      "spark": {
                        "create": true
                      }
                    },
                    "sparkJobNamespace": "",
                    "tolerations": [],
                    "uiService": {
                      "enable": false
                    },
                    "webhook": {
                      "enable": true,
                      "jobTTLSecondsAfterFinished": 60,
                      "port": 9443
                    },
                    "workqueue": {
                      "maxDelay": {
                        "enabled": false,
                        "value": "6h"
                      },
                      "tokenBucketSize": 500,
                      "tokenRefillRate": 50
                    }
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "ScheduledSparkApplication",
                "metadata": {
                  "name": "spark-pi-scheduled"
                },
                "spec": {
                  "concurrencyPolicy": "Allow",
                  "schedule": "@every 5m",
                  "template": {
                    "arguments": [
                      "1000"
                    ],
                    "driver": {
                      "annotations": {
                        "openshift.io/required-scc": "sparkoperator-sample-spark"
                      },
                      "coreLimit": "1200m",
                      "cores": 1,
                      "labels": {
                        "version": "3.5.0"
                      },
                      "memory": "512m",
                      "serviceAccount": "sparkoperator-sample-spark",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "executor": {
                      "annotations": {
                        "openshift.io/required-scc": "sparkoperator-sample-spark"
                      },
                      "cores": 1,
                      "instances": 1,
                      "labels": {
                        "version": "3.5.0"
                      },
                      "memory": "512m",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "image": "public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce",
                    "imagePullPolicy": "IfNotPresent",
                    "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.13-3.5.0.jar",
                    "mainClass": "org.apache.spark.examples.SparkPi",
                    "mode": "cluster",
                    "restartPolicy": {
                      "type": "Never"
                    },
                    "sparkVersion": "3.5.0",
                    "type": "Scala",
                    "volumes": [
                      {
                        "emptyDir": {},
                        "name": "test-volume"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "spark-pi"
                },
                "spec": {
                  "arguments": [
                    "1000"
                  ],
                  "driver": {
                    "annotations": {
                      "openshift.io/required-scc": "sparkoperator-sample-spark"
                    },
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.5.0"
                    },
                    "memory": "512m",
                    "serviceAccount": "sparkoperator-sample-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "annotations": {
                      "openshift.io/required-scc": "sparkoperator-sample-spark"
                    },
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.5.0"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce",
                  "imagePullPolicy": "IfNotPresent",
                  "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.13-3.5.0.jar",
                  "mainClass": "org.apache.spark.examples.SparkPi",
                  "mode": "cluster",
                  "restartPolicy": {
                    "type": "Never"
                  },
                  "sparkVersion": "3.5.0",
                  "type": "Scala",
                  "volumes": [
                    {
                      "emptyDir": {},
                      "name": "test-volume"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Big Data, Integration & Delivery, AI/Machine Learning, Streaming
            & Messaging
          com.redhat.openshift.versions: v4.14
          containerImage: public.ecr.aws/f4k1p1n4/netapp-spark-operator-controller@sha256:9fe1d0793f11b1c525a9a92cbcde6f4bc1bf3c887617a8496bed3a5a0008a67e
          createdAt: "2024-10-21T13:47:47Z"
          description: Red Hat and NetApp have partnered to offer enterprise-grade
            supported Apache Spark on OpenShift.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ScheduledSparkApplication
            name: scheduledsparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - kind: SparkApplication
            name: sparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - kind: SparkOperator
            name: sparkoperators.charts.netapp.com
            version: v1alpha1
        description: |-
          # NetApp Supported Apache Spark on OpenShift

          Red Hat and NetApp have partnered to offer enterprise-grade supported Apache Spark on OpenShift.
          Use this certified operator to launch Apache Spark applications on OpenShift.

          ## Documentation

          Please find the official documentation at https://www.instaclustr.com/support/documentation/apache-spark/apache-spark-on-red-hat-openshift/getting-started/

          ## Support

          With questions about Apache Spark, contact NetApp's Instaclustr support team on the [web](https://support.instaclustr.com/) or by email at support@instaclustr.com

          With questions about Operators in the OpenShift OperatorHub, contact Red Hat's support team on the [web](https://www.redhat.com/en/services/support) or by email at customerservice@redhat.com

          With questions about pricing, email ng-netapp-apache-spark-support-redhat-openshift@netapp.com
        displayName: NetApp Supported Apache Spark on OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Spark
        - Big Data
        - Apache Spark
        - AI
        - ML
        - Data Engineering
        links:
        - name: NetApp Supported Apache Spark on OpenShift
          url: https://www.instaclustr.com/support-solutions
        maintainers:
        - email: thorsteinn.sigurdsson@netapp.com
          name: Thorsteinn Sigurdsson
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: NetApp
          url: https://www.netapp.com/
        relatedImages:
        - public.ecr.aws/f4k1p1n4/netapp-spark-operator-controller@sha256:9fe1d0793f11b1c525a9a92cbcde6f4bc1bf3c887617a8496bed3a5a0008a67e
        - public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce
        version: 1.1.3
      entries:
      - name: netapp-spark-operator.v1.1.3
        version: 1.1.3
      name: alpha
    defaultChannel: alpha
    packageName: netapp-spark-operator
    provider:
      name: NetApp
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Trilio
      provider-url: ""
    name: k8s-triliovault
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k8s-triliovault-stable.3.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/trilio/k8s-triliovault-operator@sha256:0e3094945427231ecbb884b903dc4d347dfaa73e5a04eb4c5cbaed6fce1cbbb8
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          olm.skipRange: '>=3.0.0 <3.1.3'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          TrilioVault for Kubernetes is an enterprise-grade, cloud-native platform purpose built for data protection and management of Kubernetes applications for IT managers, administrators and developers. TrilioVault supports upstream Kubernetes and OpenShift environments and offers the following features:
            * Application-Centric - protects both data and metadata for Helm, Operator or custom Label based applications.
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any Storage (CSI, NFS, S3), or any Cloud (Private or Public).
            * Application Deployment & Tooling: Helm, Operators, Labels, Prometheus, Fluentd.
          Tutorials
          ------------
          Please click the link below to access the TrilioVault for Kubernetes "How-To" series for deployment, best practice and use-case videos
          <a href="https://www.trilio.io/triliovault-for-kubernetes">TrilioVault for Kubernetes - Tutorials and How-To </a>”
          Licensing
          --------
          Customers can download a 30-day Free Trial or a 10-node Basic Edition at no cost.  You can also connect with the Trilio team for an Enterprise Edition license with no limitations and Premium Support.
          For more information on license plans please vist:
          <a href="www.trilio.io/plans"> Trilio Vault for Kubernetes licensing and plans </a>
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our TrilioVault technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. www.trilio.io and @triliodata on Twitter.
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://www.trilio.io/
        - name: Product Licensing
          url: https://www.trilio.io/plans
        - name: Tutorials
          url: https://www.trilio.io/triliovault-for-kubernetes/
        - name: Support for Free/Basic Users
          url: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Trilio
        relatedImages:
        - registry.connect.redhat.com/trilio/k8s-triliovault-operator@sha256:0e3094945427231ecbb884b903dc4d347dfaa73e5a04eb4c5cbaed6fce1cbbb8
        - registry.connect.redhat.com/trilio/trilio-webhook-init@sha256:c45e9b6b75862451d873d3eba29d6bd11cafc64823855f82ccb8c0b3d779ebb0
        - registry.connect.redhat.com/trilio/web@sha256:76865d8afa8059d40810cb94df29420950a7593a40e87d142d771c57b3319e1d
        - registry.connect.redhat.com/trilio/minio@sha256:cee0f976edc79ff9e7a3f3574a7115dd960042241f3ed5b4636240ad609f9bca
        - registry.connect.redhat.com/trilio/control-plane@sha256:01e36ed5b07104865f922d2996d2211b2b6000adc90266197d9bac74d908db97
        - registry.connect.redhat.com/trilio/datamover@sha256:c7bc15f8217899ed73a812d278eb118633edbee1bf284132054b6f9ca2afd801
        - registry.connect.redhat.com/trilio/dex@sha256:b6331cf3cb3a6a66d9b9ea639cda3c2559fbf1f92f7174032352c1c7bb5176d9
        - registry.connect.redhat.com/trilio/nats@sha256:47396bfcff3e65efb3410fa8369f11f20d794ab543f9ada87b451448b1813178
        - registry.connect.redhat.com/trilio/event-stack@sha256:37c33ae1adddfe739c98fd2b2183c49ec7cfdafc88dec2970ad7c670016ebaac
        version: 3.1.3
      entries:
      - name: k8s-triliovault-stable.3.1.3
        version: 3.1.3
      - name: k8s-triliovault-stable.3.1.2
        version: 3.1.2
      - name: k8s-triliovault-stable.3.1.1
        version: 3.1.1
      - name: k8s-triliovault-stable.3.1.0
        version: 3.1.0
      name: 3.1.x
    - currentCSV: k8s-triliovault-stable.4.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:69871613ceecefff7f691f21bb0b9147236af5a5dc2c7b36ad5e0cd8551f0e6a
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <4.0.5'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/event-stack@sha256:c793418df5a1fc1a49f629d335f4684d9cde1a5f2b462f7c9483ffba904456d2
        - quay.io/triliodata/control-plane@sha256:53368a72a60a38d3ce4293236a841346da6da2380ffd2525e5ffdb8dbcf11c7f
        - quay.io/triliodata/web@sha256:20de1a2f9477ad238bf24f48c8c16a38521c3a86164f2eb6b10d20f09839016e
        - quay.io/triliodata/datamover@sha256:e6848d94ba9f81843bde882a030b0f2556aae4bf16d79e2c5b199417cf626c85
        - quay.io/triliodata/k8s-triliovault-operator@sha256:69871613ceecefff7f691f21bb0b9147236af5a5dc2c7b36ad5e0cd8551f0e6a
        - quay.io/triliodata/operator-webhook-init@sha256:aa85b787f7f37a71e909f049d2abdda746cdd0627e2f74aaf7476d5e5a7f4401
        - quay.io/triliodata/dex@sha256:b4b2372fd03be586b9ff8cd269b9001797d813a34155ab37de31a44fe3cab1ea
        - quay.io/triliodata/nats@sha256:e6723595bcc5674b6201091d45d711d29a5cd1421c8b97c76cf00e468379fddf
        version: 4.0.5
      entries:
      - name: k8s-triliovault-stable.4.0.5
        version: 4.0.5
      - name: k8s-triliovault-stable.4.0.4
        version: 4.0.4
      - name: k8s-triliovault-stable.4.0.3
        version: 4.0.3
      - name: k8s-triliovault-stable.4.0.2
        version: 4.0.2
      - name: k8s-triliovault-stable.4.0.1
        version: 4.0.1
      - name: k8s-triliovault-stable.4.0.0
        version: 4.0.0
      name: 4.0.x
    - currentCSV: k8s-triliovault-stable.5.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:8d582842dbef1c64bbe3b0ce5eeb64790910bdc0416c5e5561557642dede7022
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <5.0.4'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/web@sha256:04c4baadae9c05f4add098ff62774d00102d56581e261626840d17f74effe438
        - quay.io/triliodata/datamover@sha256:104bdca9513c78f81c4340e9a3219df0a3bc8a5b026e4a252363a7e2c10977ba
        - quay.io/triliodata/dex@sha256:b4b2372fd03be586b9ff8cd269b9001797d813a34155ab37de31a44fe3cab1ea
        - quay.io/triliodata/nats@sha256:e6723595bcc5674b6201091d45d711d29a5cd1421c8b97c76cf00e468379fddf
        - quay.io/triliodata/event-stack@sha256:859535e19e0ebebc8261c649c2446f8ca5fe4b4e5b6236e69b780dcc0d0bb5b9
        - quay.io/triliodata/k8s-triliovault-operator@sha256:8d582842dbef1c64bbe3b0ce5eeb64790910bdc0416c5e5561557642dede7022
        - quay.io/triliodata/operator-webhook-init@sha256:ffcc55e85f884b884947b73fa7d2fb257704ff70ccd7f2dc37d07e02292abed6
        - quay.io/triliodata/control-plane@sha256:e60ef0a2c7a5b7879c9718e9a14c223f5290b4710b598fae19466a7558484824
        version: 5.0.4
      entries:
      - name: k8s-triliovault-stable.5.0.4
        version: 5.0.4
      - name: k8s-triliovault-stable.5.0.3
        version: 5.0.3
      - name: k8s-triliovault-stable.5.0.2
        version: 5.0.2
      - name: k8s-triliovault-stable.5.0.1
        version: 5.0.1
      - name: k8s-triliovault-stable.5.0.0
        version: 5.0.0
      name: 5.0.x
    - currentCSV: k8s-triliovault-stable.5.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "2Gi",
                      "cpu": "1"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:f0cf3f586561414671668210c59ac44071fde6abc954917e6ca04df5777a9a2f
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <5.1.1'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/k8s-triliovault-operator@sha256:f0cf3f586561414671668210c59ac44071fde6abc954917e6ca04df5777a9a2f
        - quay.io/triliodata/event-stack@sha256:548ae1a748919f518e811774ec87f128aa8857be234837e09308c295de165726
        - quay.io/triliodata/control-plane@sha256:2e7e44cf6bc70dd9b978dce2cb2d9a36b8886c0f55286d5a6f5eb0a5182c4350
        - quay.io/triliodata/datamover@sha256:8a1350c8975610f23dd0b40d604d218f59dca79db6beb81ffe78a2682e40864f
        - quay.io/triliodata/dex@sha256:6b9d1667916968fb1f187ad5945fcc08db702dcab4c947e309dd9d2881a7b9cf
        - quay.io/triliodata/nats@sha256:72283871b6b9343aeabc3ff090305979422f192d30a6cbc4b0f20b2de7533996
        - quay.io/triliodata/filerecovery-vm-ssh-server@sha256:09186be228bd0b59e83119b28f3239c6173e410a3cf36957fe2d9b56d25200f8
        - quay.io/triliodata/web@sha256:3908a39eed44d702c6b6804d392d08a570b3f5562590d050f17ddc7a5c87b37f
        - quay.io/triliodata/filerecovery@sha256:42559ac5d949d5ecf819cd4768a71fa688c72230502ff0b56539a8bb5e2f5838
        version: 5.1.1
      entries:
      - name: k8s-triliovault-stable.5.1.1
        version: 5.1.1
      - name: k8s-triliovault-stable.5.1.0
        version: 5.1.0
      name: 5.1.x
    defaultChannel: 5.1.x
    packageName: k8s-triliovault
    provider:
      name: Trilio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Couchbase
      provider-url: ""
    name: couchbase-enterprise-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: couchbase-operator.v2.6.4-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator@sha256:1f0a6865c360a14ba98706fd6fe8dff3bdbdede32e3622f93f19c82e377ca33f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        version: 2.6.4-7
      entries:
      - name: couchbase-operator.v2.6.4-7
        version: 2.6.4-7
      - name: couchbase-operator.v2.6.4-6
        version: 2.6.4-6
      name: 2.6.4
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:707571a7881bc12d5f1b0b1a94eb3c675d5c0ea0114c994a47ceb5aa79df4013
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        version: 2.7.0-5
      entries:
      - name: couchbase-operator.v2.7.0-5
        version: 2.7.0-5
      name: 2.7.0
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:a79e1b6781d5f8badf664e7efea6e2889e43fae6e0d37169d766685dcd6ff394
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        version: 2.7.1-1
      entries:
      - name: couchbase-operator.v2.7.1-1
        version: 2.7.1-1
      name: 2.7.1
    - currentCSV: couchbase-operator.v2.8.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator@sha256:df61928b064014d851c6e7dff6dcd63a433a27fa5623254b6e389c3eea170e36
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        version: 2.8.0-2
      entries:
      - name: couchbase-operator.v2.8.0-2
        version: 2.8.0-2
      - name: couchbase-operator.v2.8.0-1
        version: 2.8.0-1
      name: 2.8.0
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        version: 2.8.1-2
      entries:
      - name: couchbase-operator.v2.8.1-2
        version: 2.8.1-2
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: 2.8.1
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        version: 2.8.1-2
      entries:
      - name: couchbase-operator.v2.8.1-2
        version: 2.8.1-2
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: stable
    defaultChannel: stable
    packageName: couchbase-enterprise-certified
    provider:
      name: Couchbase
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubemq.io
      provider-url: ""
    name: kubemq-operator-marketplace-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubemq-operator.v1.7.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
              "apiVersion": "core.k8s.kubemq.io/v1beta1",
              "kind": "KubemqCluster",
              "metadata": {
                "name": "kubemq-cluster"
              },
              "spec": {
                "replicas": 3
              }
            },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqDashboard",
                "metadata": {
                  "name": "kubemq-dashboard"
                }
              },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqConnector",
                "metadata": {
                  "name": "kubemq-bridges"
                },
                "spec": {
                  "type": "bridges",
                  "config": "bindings: null"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/kubemq/kubemq-operator:v1.7.2
          createdAt: 2020-03-11 12:00:00+00:00
          description: Kubernetes Messaging Platform
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/kubemq-operator-marketplace-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/kubemq-operator-marketplace-rhmp/support?utm_source=openshift_console
          support: support@kubemq.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kubemq Cluster
            displayName: Kubemq Cluster
            kind: KubemqCluster
            name: kubemqclusters.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Cluster Dashboard
            displayName: Kubemq Dashboard
            kind: KubemqDashboard
            name: kubemqdashboards.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Connector
            displayName: Kubemq Connector
            kind: KubemqConnector
            name: kubemqconnectors.core.k8s.kubemq.io
            version: v1beta1
        description: |2-

          ## Kubernetes Messaging Platform
          A message-based architecture for microservices providing flexible connectivity for hybrid environments: multi-cloud, on-premises, and at the edge. Powered by enterprise-grade message broker and message queue with ready to use connectors, bridges, and control center.

          Messaging platform components:
            - **KubeMQ Cluster** - Enterprise-grade message broker and message queue, scalable, high available, and secured.
            - **KubeMQ Targets Connector** - connects a KubeMQ cluster with external systems and cloud services.
            - **KubeMQ Bridges Connector** - bridge, replicate, aggregate, and transform messages between KubeMQ clusters.
            - **KubeMQ Sources Connector** - transform external systems and cloud services data to KubeMQ cluster messages.
            - **KubeMQ Dashboard** - monitor KubeMQ cluster metrics.

          ### Kubernetes Native
          Innovative and modern message queue and message broker in a lightweight container developed to run in Kubernetes, certified in the CNCF landscape, and connect natively to the cloud-native ecosystem.

          ### Ease of use

          Simple deployment in Kubernetes in less than 1 minute. Developer friendly by simple to use SDKs and elimination of the many developers and DevOps-centered challenges to define exchanges, brokers, channels, routes, and predefined topics

          ### All messaging patterns

          KubeMQ is available with all messaging patterns such as Queue, Stream, Pub/Sub, and RPC, Saving the need to maintain multiple messaging platforms for different use cases.

          ### Hybrid Environments

          KubeMQ enables the microservices from multiple environments to communicate and build one hybrid infrastructure solution across clouds, on-prem, and at the edge.

          ### KubeMQ Cluster key features

          - All-batteries included Messaging Queue Broker for the Kubernetes environment
          - Deployed with Operator for full life cycle operation
          - Blazing fast (written in Go), small and lightweight Docker container
          - Asynchronous and Synchronous messaging with support for  `At Most Once Delivery` and `At Least Once Delivery` models
          - Supports durable FIFO based Queue, Publish-Subscribe Events, Publish-Subscribe with Persistence (Events Store), RPC Command and Query messaging patterns
          - Supports gRPC, Rest and WebSocket Transport protocols with TLS support (both RPC and Stream modes)
          - Supports Access control Authorization and Authentication
          - Supports message masticating and smart routing
          - No Message broker configuration needed (i.e., queues, exchanges)
          - .Net, Java, Python, Go, and NodeJS SDKs

          ### KubeMQ Targets Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major services types (databases, cache, messaging, serverless, HTTP, etc.)
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-targets)

          ### KubeMQ Bridges Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Build Any Topology - connects KubeMQ clusters in 1:1, 1:n , n:1, n:n
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-bridges)

          ### KubeMQ Sources Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major messaging services such as Kafka, ActiveMQ, SQS, and more.
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-sources)

          ## Getting Started

          [Start with OpenShift](https://docs.kubemq.io/getting-started/create-cluster/openshift)

          ## Get Enterprise Trial License

          [Register](https://account.kubemq.io/login/register)

          ## Request a Demo

          [See KubeMQ in Action](https://kubemq.io/request-a-demo/)
        displayName: Kubemq Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - messaging
        - queue
        - broker
        - pubsub
        - rpc
        links:
        - name: Learn more about Kubemq
          url: https://kubemq.io/
        - name: Register for free license
          url: https://account.kubemq.io/login/register
        - name: Contact Sales
          url: https://kubemq.io/contact-sales/
        - name: Documentation
          url: https://docs.kubemq.io/
        - name: Get Help
          url: https://kubemq.io/contact-support/
        maintainers:
        - email: support@kubemq.io
          name: kubemq
        maturity: alpha
        provider:
          name: Kubemq.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-grafana:v4.3
        - registry.marketplace.redhat.com/rhm/kubemq/kubemq-operator:v1.7.2
        - registry.marketplace.redhat.com/rhm/kubemq/kubemq-uni:v2.2.12
        - registry.redhat.io/openshift4/ose-prometheus:v4.3
        version: 1.7.8
      entries:
      - name: kubemq-operator.v1.7.8
        version: 1.7.8
      name: alpha
    defaultChannel: alpha
    packageName: kubemq-operator-marketplace-rhmp
    provider:
      name: Kubemq.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: crunchy-postgres-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/crunchy-postgres-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/crunchy-postgres-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-exporter@sha256:3b231902b2116cd38de05c7378202716c95813b9961fb8cee811b61d36cddf86
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:0e283616c2d32e53258d62d02702bc4e179ef8c03933592154babed1a70bf666
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:a7f62f367817fc08294a0b8b82ac9826a42c0ae668efdfacaa158a4287b8a688
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:c559c896e174674d151f93543ac92a1b5647eb4676313f7dc6cbb8faee786f18
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:56a5e3a5758982e086c9f78201a7137e0d934316f058a31cf30df44639a569a2
        - registry.connect.redhat.com/crunchydata/crunchy-pgadmin4@sha256:0d84acb1ce6a9b046e3f8b222a583ddcc91dda0a117c9f637ef173d0ac992f57
        - registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:b2c3c74fdcc0e08f555e50c527f0d322292333c5d3dce061a0dee3680a1fbc56
        - registry.connect.redhat.com/crunchydata/crunchy-pgbackrest@sha256:b70da0035b74379f2507cfabdd9d1cc1e3c3eecb9066c22f292981c9a27916e1
        - registry.connect.redhat.com/crunchydata/crunchy-pgbouncer@sha256:17cb58dec78447c2f70a414a2aeff6facbf921b42a0f366d0dbf7be95aa8987a
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:4bdd0588ffbb3f628e7df1c72670e7a5087269438343b68779012bc3a13dea31
        - registry.connect.redhat.com/crunchydata/crunchy-upgrade@sha256:68bc93cce61d9b1e44822ae3e3ac8f7a9c2b712eddfaea090eabefacd60c67f6
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: crunchy-postgres-operator-rhmp
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-networkfirewall-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-networkfirewall-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "Firewall",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "FirewallPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "RuleGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/networkfirewall-controller:1.1.1
          createdAt: "2025-09-19T21:03:12Z"
          description: AWS Network Firewall controller is a service controller for
            managing Network Firewall resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FirewallPolicy represents the state of an AWS networkfirewall
              FirewallPolicy resource.
            displayName: FirewallPolicy
            kind: FirewallPolicy
            name: firewallpolicies.networkfirewall.services.k8s.aws
            version: v1alpha1
          - description: Firewall represents the state of an AWS networkfirewall Firewall
              resource.
            displayName: Firewall
            kind: Firewall
            name: firewalls.networkfirewall.services.k8s.aws
            version: v1alpha1
          - description: RuleGroup represents the state of an AWS networkfirewall
              RuleGroup resource.
            displayName: RuleGroup
            kind: RuleGroup
            name: rulegroups.networkfirewall.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Network Firewall resources in AWS from within your Kubernetes cluster.

          **About Amazon Network Firewall**

          AWS Network Firewall is a stateful, managed, network firewall and intrusion detection and prevention service for your virtual private cloud (VPC) that you create in Amazon Virtual Private Cloud (Amazon VPC).
          With Network Firewall, you can filter traffic at the perimeter of your VPC. This includes filtering traffic going to and coming from an internet gateway, NAT gateway, or over VPN or AWS Direct Connect. Network Firewall uses the open source intrusion prevention system (IPS), Suricata, for stateful inspection. Network Firewall supports Suricata compatible rules. For more information, see [Working with stateful rule groups in AWS Network Firewall](https://docs.aws.amazon.com/network-firewall/latest/developerguide/stateful-rule-groups-ips.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Network Firewall
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networkfirewall
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Network Firewall Developer Resources
          url: https://aws.amazon.com/network-firewall/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: network firewall maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/networkfirewall-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-networkfirewall-controller.v1.1.1
        version: 1.1.1
      - name: ack-networkfirewall-controller.v1.1.0
        version: 1.1.0
      - name: ack-networkfirewall-controller.v1.0.13
        version: 1.0.13
      - name: ack-networkfirewall-controller.v1.0.12
        version: 1.0.12
      - name: ack-networkfirewall-controller.v1.0.11
        version: 1.0.11
      - name: ack-networkfirewall-controller.v1.0.10
        version: 1.0.10
      - name: ack-networkfirewall-controller.v1.0.9
        version: 1.0.9
      - name: ack-networkfirewall-controller.v1.0.8
        version: 1.0.8
      - name: ack-networkfirewall-controller.v1.0.7
        version: 1.0.7
      - name: ack-networkfirewall-controller.v1.0.6
        version: 1.0.6
      - name: ack-networkfirewall-controller.v1.0.5
        version: 1.0.5
      - name: ack-networkfirewall-controller.v1.0.4
        version: 1.0.4
      - name: ack-networkfirewall-controller.v1.0.1
        version: 1.0.1
      - name: ack-networkfirewall-controller.v1.0.0
        version: 1.0.0
      - name: ack-networkfirewall-controller.v0.0.10
        version: 0.0.10
      - name: ack-networkfirewall-controller.v0.0.9
        version: 0.0.9
      - name: ack-networkfirewall-controller.v0.0.8
        version: 0.0.8
      - name: ack-networkfirewall-controller.v0.0.7
        version: 0.0.7
      - name: ack-networkfirewall-controller.v0.0.6
        version: 0.0.6
      - name: ack-networkfirewall-controller.v0.0.5
        version: 0.0.5
      - name: ack-networkfirewall-controller.v0.0.2
        version: 0.0.2
      - name: ack-networkfirewall-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-networkfirewall-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: self-node-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: self-node-remediation.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:cb03286b9d9cda4a47e5b1c0058ca315bf082bcc8697506ea88d97ba77452bea
          createdAt: "2023-10-20 16:38:06"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          olm.skipRange: '>=0.4.0 <0.5.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the heath detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-self-node-remediation-operator.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:cb03286b9d9cda4a47e5b1c0058ca315bf082bcc8697506ea88d97ba77452bea
        version: 0.5.1
      entries:
      - name: self-node-remediation.v0.5.1
        version: 0.5.1
      name: 4.12-eus
    - currentCSV: self-node-remediation.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:e2048a8a2ec37e10b57dafeba2e131311e61982242cda4beb989e5e30c1a5537
          createdAt: "2023-10-20 15:32:20"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          olm.skipRange: '>=0.5.0 <0.7.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:e2048a8a2ec37e10b57dafeba2e131311e61982242cda4beb989e5e30c1a5537
        version: 0.7.1
      entries:
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.14-eus
    - currentCSV: self-node-remediation.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:6bd86a47d69f7662e24a93ce205a8179018152a97696a9223d41ed1887c35257
          createdAt: "2025-07-02 16:40:20"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.9.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:6bd86a47d69f7662e24a93ce205a8179018152a97696a9223d41ed1887c35257
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ef91935704bb9f0af4cdad07333e802587e3d458b2cd911e04d557b1defe3412
        version: 0.9.1
      entries:
      - name: self-node-remediation.v0.9.1
        version: 0.9.1
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.16-eus
    - currentCSV: self-node-remediation.v0.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
          createdAt: "2025-07-02 18:34:25"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        version: 0.10.2
      entries:
      - name: self-node-remediation.v0.10.2
        version: 0.10.2
      - name: self-node-remediation.v0.10.1
        version: 0.10.1
      - name: self-node-remediation.v0.10.0
        version: 0.10.0
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.18-eus
    - currentCSV: self-node-remediation.v0.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
          createdAt: "2025-07-02 18:34:25"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        version: 0.10.2
      entries:
      - name: self-node-remediation.v0.10.2
        version: 0.10.2
      - name: self-node-remediation.v0.10.1
        version: 0.10.1
      - name: self-node-remediation.v0.10.0
        version: 0.10.0
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: stable
    defaultChannel: stable
    packageName: self-node-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: aws-efs-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-efs-csi-driver-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-aws-efs-csi-driver-rhel9-operator@sha256:35ef24108df5f24d24fefe915f501965c98edea934fa6c2895e375ed006e5e7b
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure AWS EFS CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0-0 <4.16.0-202509240123'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/aws-efs-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          AWS EFS CSI Driver Operator provides AWS Elastic File Service (EFS) CSI Driver that enables you to create and mount AWS EFS PersistentVolumes.

          Please follow [AWS EFS CSI Driver Operator documentation](https://docs.openshift.com/container-platform/4.16/storage/container_storage_interface/persistent-storage-csi-aws-efs.html) to:

          1. [Install the operator](https://docs.openshift.com/container-platform/4.16/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#persistent-storage-csi-olm-operator-install_persistent-storage-csi-aws-efs). This section also covers details about configuring an AWS role in OCP cluster that uses Secure Token Services (STS).

          2. [Install the AWS EFS CSI driver using the operator](https://docs.openshift.com/container-platform/4.16/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#persistent-storage-csi-efs-driver-install_persistent-storage-csi-aws-efs)

          3. Finally, [create a StorageClass](https://docs.openshift.com/container-platform/4.16/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#storage-create-storage-class_persistent-storage-csi-aws-efs) that enables dynamic provisioning of PersistentVolumes.
        displayName: AWS EFS CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - efs
        links:
        - name: Documentation
          url: https://github.com/openshift/aws-efs-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/aws-efs-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-aws-efs-csi-driver-rhel9-operator@sha256:35ef24108df5f24d24fefe915f501965c98edea934fa6c2895e375ed006e5e7b
        - registry.redhat.io/openshift4/ose-aws-efs-csi-driver-container-rhel9@sha256:c857dda8f399fa104bd4f392a354b74365657acea6b8fc5b023fdb4bf55057c2
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:9c3c3d2b857ce289324aee56df1ba4ab5e5ea1e76cd86ac988233463915834b4
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0627c09fe3f5a015f117fb30af1f4d17340aefa48b23994737d3901a804b63d3
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:a5704d26d0b562dba9633d1c107f34e05c750c0e3a64dbcace4dab7b964fff8d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        version: 4.16.0-202509240123
      entries:
      - name: aws-efs-csi-driver-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: aws-efs-csi-driver-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: aws-efs-csi-driver-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: aws-efs-csi-driver-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: aws-efs-csi-driver-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: aws-efs-csi-driver-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: aws-efs-csi-driver-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: aws-efs-csi-driver-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: aws-efs-csi-driver-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: aws-efs-csi-driver-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: aws-efs-csi-driver-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: aws-efs-csi-driver-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: aws-efs-csi-driver-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: aws-efs-csi-driver-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: aws-efs-csi-driver-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: aws-efs-csi-driver-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: aws-efs-csi-driver-operator.v4.16.0-202410280205
        version: 4.16.0-202410280205
      - name: aws-efs-csi-driver-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: aws-efs-csi-driver-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: aws-efs-csi-driver-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: aws-efs-csi-driver-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: aws-efs-csi-driver-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: aws-efs-csi-driver-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: aws-efs-csi-driver-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: aws-efs-csi-driver-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: aws-efs-csi-driver-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: aws-efs-csi-driver-operator.v4.16.0-202408061508
        version: 4.16.0-202408061508
      - name: aws-efs-csi-driver-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: aws-efs-csi-driver-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: aws-efs-csi-driver-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: aws-efs-csi-driver-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: aws-efs-csi-driver-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: aws-efs-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: proactive-node-scaling-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: proactive-node-scaling-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NodeScalingWatermark",
                "metadata": {
                  "name": "ai-ml"
                },
                "spec": {
                  "nodeSelector": {
                    "machine.openshift.io/cluster-api-machine-type": "ai-ml"
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "workload",
                      "operator": "Equal",
                      "value": "ai-ml"
                    }
                  ],
                  "watermarkPercentage": 20
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NodeScalingWatermark",
                "metadata": {
                  "name": "us-west-2a"
                },
                "spec": {
                  "nodeSelector": {
                    "topology.kubernetes.io/zone": "us-west-2a"
                  },
                  "watermarkPercentage": 20
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "false"
          containerImage: quay.io/redhat-cop/proactive-node-scaling-operator@sha256:892303babb45090a6458cd4b1c61a01364012c43ea0bd6ed4d10bdcb8041e2ab
          createdAt: "2022-01-22T14:14:23Z"
          description: This operator provides automation proactively scaling nodes
            before pods get stuck waiting for the cluster autoscaler.
          operatorframework.io/suggested-namespace: proactive-node-scaling-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/proactive-node-scaling-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeScalingWatermark is the Schema for the nodescalingwatermarks
              API
            displayName: Node Scaling Watermark
            kind: NodeScalingWatermark
            name: nodescalingwatermarks.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator makes the [cluster autoscaler](https://github.com/kubernetes/autoscaler/tree/master/cluster-autoscaler) more proactive. As of now the cluster auto scaler will create new nodes only when a pod is pending because it cannot be allocated due to lack of capacity. This is not a goos user experience as the pending workload has to wait for several minutes as the new node is create and joins the cluster.

          The Proactive Node Scaling Operator improves the user experience by allocating low priority pods that don't do anything. When the cluster is full and a new user pod is created the following happens:

          1. some of the low priority pods are de-scheduled to make room for the user pod, which can then be scheduled. The user workload does not have to wait in this case.

          2. the de-scheduled low priority pods are rescheduled and in doing so the trigger the cluster autoscaler to add new nodes.

          Essentially this operator allows you to trade wasted resources for faster response time.

          In order for this operator to work correctly [pod priorities](https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/) must be defined. The default name for the priority class used by this operator is "proactive-node-autoscaling-pods" and it should have the lowest possible priority, 0. To ensure your regular workloads get a normal priority you should also define a PriorityClass for those with a higher priority than 0 and set globalDefault to true.

          For example:

          ```yaml
          apiVersion: scheduling.k8s.io/v1
          kind: PriorityClass
          metadata:
            name: proactive-node-autoscaling-pods
          value: 0
          globalDefault: false
          description: "This priority class is the priority class used for Proactive Node Scaling Pods."
          ---
          apiVersion: scheduling.k8s.io/v1
          kind: PriorityClass
          metadata:
            name: normal-workload
          value: 1000
          globalDefault: true
          description: "This priority classis the cluster default and should be used for normal workloads."
          ```

          Also for this operator to work the [cluster autoscaler](https://github.com/kubernetes/autoscaler/tree/master/cluster-autoscaler) must be active, see OpenShift instructions [here](https://docs.openshift.com/container-platform/4.6/machine_management/applying-autoscaling.html) on how to turn it on.

          To activate the proactive autoscaling, a CR must be defined, here is an example:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: NodeScalingWatermark
          metadata:
            name: us-west-2a
          spec:
            priorityClassName: proactive-node-autoscaling-pods
            watermarkPercentage: 20
            nodeSelector:
              topology.kubernetes.io/zone: us-west-2a
          ```

          The `nodeSelector` selects the nodes observed by this operator, which are also the nodes on which the low priority pods will be scheduled. The nodes observed by the cluster autoscaler should coincide with the nodes selected by this operator CR.

          The `watermarkPercentage` define the percentage of capacity of user workload that will be allocated to low priority pods. So in this example 20% of the user allocated capacity will be allocated via low priority pods. This also means that when the user workload reaches 80% capacity of the nodes selected by this CR (and the autoscaler), the cluster will start to scale.
        displayName: Proactive Node Scaling Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Capacity Management
        - Node Autoscaling
        - Proactive Autoscaling
        - Elastic Capacity
        links:
        - name: Proactive Node Scaling Operator
          url: https://github.com/redhat-cop/proactive-node-scaling-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/proactive-node-scaling-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/proactive-node-scaling-operator@sha256:892303babb45090a6458cd4b1c61a01364012c43ea0bd6ed4d10bdcb8041e2ab
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 0.4.2
      entries:
      - name: proactive-node-scaling-operator.v0.4.2
        version: 0.4.2
      - name: proactive-node-scaling-operator.v0.4.0
        version: 0.4.0
      - name: proactive-node-scaling-operator.v0.3.10
        version: 0.3.10
      - name: proactive-node-scaling-operator.v0.3.9
        version: 0.3.9
      - name: proactive-node-scaling-operator.v0.3.8
        version: 0.3.8
      - name: proactive-node-scaling-operator.v0.3.7
        version: 0.3.7
      - name: proactive-node-scaling-operator.v0.3.6
        version: 0.3.6
      - name: proactive-node-scaling-operator.v0.3.5
        version: 0.3.5
      - name: proactive-node-scaling-operator.v0.3.4
        version: 0.3.4
      - name: proactive-node-scaling-operator.v0.3.3
        version: 0.3.3
      - name: proactive-node-scaling-operator.v0.3.2
        version: 0.3.2
      - name: proactive-node-scaling-operator.v0.3.1
        version: 0.3.1
      - name: proactive-node-scaling-operator.v0.3.0
        version: 0.3.0
      - name: proactive-node-scaling-operator.v0.2.0
        version: 0.2.0
      - name: proactive-node-scaling-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: proactive-node-scaling-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Debezium Authors
      provider-url: https://debezium.io/
    name: debezium-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: debezium-operator.v2.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.4.2.Final
          createdAt: "2023-12-13T12:20:56Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.4.2.Final
        version: 2.4.2
      entries:
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.4.x
    - currentCSV: debezium-operator.v2.5.3-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.5.3.Final
          createdAt: "2024-03-19T08:17:17Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.5.3.Final
        version: 2.5.3-final
      entries:
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.5.x
    - currentCSV: debezium-operator.v2.6.1-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.6.1.Final
          createdAt: "2024-04-12T17:18:16Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.5.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.6.1.Final
        version: 2.6.1-final
      entries:
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.6.x
    - currentCSV: debezium-operator.v2.7.3-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.7.3.Final
          createdAt: "2024-09-20T10:31:16Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.7.3.Final
        version: 2.7.3-final
      entries:
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.7.x
    - currentCSV: debezium-operator.v3.0.0-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:3.0.0.Final
          createdAt: "2024-10-02T09:55:38Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.7.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:3.0.0.Final
        version: 3.0.0-final
      entries:
      - name: debezium-operator.v3.0.0-final
        version: 3.0.0-final
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-3.0.x
    - currentCSV: debezium-operator.v3.0.0-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:3.0.0.Final
          createdAt: "2024-10-02T09:55:38Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.7.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:3.0.0.Final
        version: 3.0.0-final
      entries:
      - name: debezium-operator.v3.0.0-final
        version: 3.0.0-final
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-latest
    defaultChannel: debezium-latest
    packageName: debezium-operator
    provider:
      name: Debezium Authors
      url: https://debezium.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-notifications
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-notifications.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Notifications",
                "metadata": {
                  "name": "notifications-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configuration": {
                    "port": 8080
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp",
                    "test": "wap-rhel8-toolbox:8.7.0-3747-36-ocp",
                    "testRepository": "prod.harbor.keysight.digital/loadcore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:34:47Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Notifications
            name: notifications.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Notifications Operator
        displayName: Keysight WAP Notifications Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Notifications
          url: https://keysight-wap-notifications.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-notifications@sha256:50c24180b3c16cf3eb82f6f4eeb61b91a62a198c122f7e59aa680a25bc488c2d
        version: 0.1.21
      entries:
      - name: keysight-wap-notifications.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-notifications
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-stats-dashboards
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-stats-dashboards.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Statsdashboard",
                "metadata": {
                  "name": "statsdashboard-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "statsDashboardService": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:41:08Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Statsdashboard
            name: statsdashboards.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Stats Dashboard Operator
        displayName: Keysight WAP Stats Dashboard Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Stats Dashboards
          url: https://keysight-wap-stats-dashboards.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-stats-dashboards@sha256:bcdc68bf0b9e9c22d775d704320829d258a9ebc77dfe41e2414061740f87f236
        version: 0.1.21
      entries:
      - name: keysight-wap-stats-dashboards.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-stats-dashboards
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Saison Technology Co., Ltd.
      provider-url: https://www.hulft.com/
    name: hulft
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hulft.v10.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hulft.saison-technology.com/v1",
                "kind": "Hulft",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "hulft",
                    "app.kubernetes.io/instance": "sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hulft",
                    "app.kubernetes.io/part-of": "hulft"
                  },
                  "name": "sample",
                  "namespace": "user-apps"
                },
                "spec": {
                  "applicationName": "hulft-apps",
                  "control": {
                    "allowExternalAccess": true,
                    "replicas": {
                      "max": 1,
                      "min": 1
                    }
                  },
                  "db": {
                    "authorizationSecret": "secret-hulft-db-authorization",
                    "connectionRetryCount": 30,
                    "hostname": "database.example.com",
                    "name": "hulft",
                    "port": 3306,
                    "tablePrefix": "hulft"
                  },
                  "licenseSecret": "secret-hulft-license",
                  "resourceIdentifier": "sample",
                  "transfer": {
                    "control": {
                      "retryCount": 10,
                      "retryInterval": 10000,
                      "timeout": 60000
                    },
                    "gracefulShutdownWait": 3600,
                    "observe": {
                      "websocket": {
                        "allowExternalAccess": false
                      }
                    },
                    "replicas": {
                      "max": 5,
                      "min": 1
                    },
                    "s3": {
                      "defaultRegion": "ap-northeast-1"
                    },
                    "websocket": {
                      "allowExternalAccess": false
                    }
                  }
                },
                "status": {
                  "applicationName": "hulft-apps",
                  "control": {
                    "allowExternalAccess": true,
                    "apiClientTlsSecret": "secret-hulft-sample-transfer-control-server-tls",
                    "consoleLoginSecret": "secret-hulft-sample-console-login",
                    "consoleTlsSecret": "secret-hulft-sample-control-console-tls",
                    "hostname": "",
                    "httpPort": 30080,
                    "httpsPort": 30443,
                    "replicas": {
                      "max": 1,
                      "min": 1
                    },
                    "samlSecret": ""
                  },
                  "db": {
                    "authorizationSecret": "secret-hulft-db-authorization",
                    "connectionRetryCount": 30,
                    "encryptionKeySecret": "secret-hulft-sample-db-encryption-key",
                    "hostname": "database.example.com",
                    "name": "hulft",
                    "port": 3306,
                    "tablePrefix": "hulft"
                  },
                  "licenseSecret": "secret-hulft-license",
                  "resourceIdentifier": "sample",
                  "transfer": {
                    "control": {
                      "clientTlsSecret": "secret-hulft-sample-control-console-tls",
                      "port": 30443,
                      "retryCount": 10,
                      "retryInterval": 10000,
                      "serverTlsSecret": "secret-hulft-sample-transfer-control-server-tls",
                      "timeout": 60000
                    },
                    "gracefulShutdownWait": 3600,
                    "hostname": "hulft-sample-transfer-wss-ci.apps.openshift.example.com",
                    "hulft": {
                      "port": 30000
                    },
                    "observe": {
                      "port": 31000,
                      "websocket": {
                        "allowExternalAccess": true,
                        "port": 33443,
                        "serverTlsSecret": "secret-hulft-sample-transfer-observe-wss-server-tls"
                      }
                    },
                    "replicas": {
                      "max": 5,
                      "min": 1
                    },
                    "s3": {
                      "authenticationSecret": "",
                      "defaultRegion": "ap-northeast-1"
                    },
                    "websocket": {
                      "allowExternalAccess": true,
                      "clientTlsSecret": "",
                      "port": 31443,
                      "serverTlsSecret": "secret-hulft-sample-transfer-wss-server-tls"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/saison-technology/hulft-operator@sha256:8ee0a29d0ce96f4ec230f387467b241dccbb2bd350cd642b275ef5f5d3b64972
          createdAt: "2025-09-30T07:54:26Z"
          description: HULFT10 for Container Platform is file integration middleware
            that provides transfer functions for the data accumulated in your IT system
            during the daily operations of your business, such as sales data, customer
            information, and various logs. File transfer from/to on-premises HULFT
            is also available.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: To use HULFT10 for Container
            Platform, you must purchase the product and have a key issued. Please
            refer to [Start Up Guide](https://www.hulft.com/help/ja-jp/HULFT-V10/CONP-SUGOS)
            for the deployment procedure.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Saison Technology Co., Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hulft is the Schema for the hulfts API
            displayName: Hulft
            kind: Hulft
            name: hulfts.hulft.saison-technology.com
            version: v1
        description: |
          HULFT10 for Container Platform is file integration middleware that provides transfer functions for the data accumulated in your IT system during the daily operations of your business, such as sales data, customer information, and various logs. File transfer from/to on-premises HULFT is also available.
          To use HULFT10 for Container Platform, you must purchase the product and have a key issued. Please refer to [Start Up Guide](https://www.hulft.com/help/ja-jp/HULFT-V10/CONP-SUGOS) for the deployment procedure.
        displayName: HULFT10 for Container Platform
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backup & Recovery
        - Storage
        - Migration
        links:
        - name: Product page
          url: https://www.hulft.com/software/hulft-10
        - name: Product manual
          url: https://www.hulft.com/help/ja-jp/HULFT-V10/index.htm
        maintainers:
        - email: salessupport_mp@hulft.com
          name: Saison Technology Co., Ltd.
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Saison Technology Co., Ltd.
          url: https://www.hulft.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fa0acefda43e309e7f9f31f5247c933d0b868ff1573ece2f38e0549d7e2ec942
        - registry.connect.redhat.com/saison-technology/hulft-operator@sha256:8ee0a29d0ce96f4ec230f387467b241dccbb2bd350cd642b275ef5f5d3b64972
        version: 10.5.0
      entries:
      - name: hulft.v10.5.0
        version: 10.5.0
      - name: hulft.v10.4.0
        version: 10.4.0
      - name: hulft.v10.3.0
        version: 10.3.0
      name: stable
    defaultChannel: stable
    packageName: hulft
    provider:
      name: Saison Technology Co., Ltd.
      url: https://www.hulft.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC Corporation
      provider-url: https://jpn.nec.com/webotx/
    name: webotx-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: webotx-operator.v11.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "webotx.nec.com/v3",
                "kind": "ApplicationServer",
                "metadata": {
                  "name": "webotx-express"
                },
                "spec": {
                  "deployment": {
                    "image": "registry.connect.redhat.com/nec/webotx-express:11.11.00.00@sha256:73a5dc48b4e3340951dbe207d744f5ac68474701dd134fa7aa05e65e3510c8ad",
                    "replicas": 1
                  },
                  "service": {
                    "externalTrafficPolicy": "Cluster",
                    "ports": [
                      {
                        "name": "http",
                        "nodePort": 30080,
                        "port": 8080,
                        "protocol": "TCP",
                        "targetPort": 8080
                      },
                      {
                        "name": "https",
                        "nodePort": 30443,
                        "port": 8443,
                        "protocol": "TCP",
                        "targetPort": 8443
                      }
                    ],
                    "sessionAffinity": "ClientIP",
                    "sessionAffinityConfig": {
                      "clientIP": {
                        "timeoutSeconds": 10800
                      }
                    },
                    "type": "NodePort"
                  },
                  "webotx": {
                    "metering": {
                      "enabled": false,
                      "port": 8443,
                      "scheme": "https"
                    },
                    "sessionReplication": {
                      "enabled": false,
                      "useAllProcessGroups": false,
                      "useJmxAgent": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApplicationServer is the Schema for the applicationservers
              API
            displayName: Application Server
            kind: ApplicationServer
            name: applicationservers.webotx.nec.com
            version: v3
        description: WebOTX Operator is the Operator for WebOTX Application Server.
          WebOTX Operator enables to deploy and administrate WebOTX Application Server
          automatically.
        displayName: WebOTX Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - webotx
        - WebOTX
        - nec
        - NEC
        links:
        - name: WebOTX Operator for Kubernetes
          url: https://github.com/webotx/operator
        maintainers:
        - email: info-webotx@isd.jp.nec.com
          name: NEC Corporation
        maturity: stable
        provider:
          name: NEC Corporation
          url: https://jpn.nec.com/webotx/
        relatedImages:
        - registry.connect.redhat.com/nec/webotx-operator:11.11.00.00@sha256:df0d0315ed88be875ca7a6474500f43f17855f763c5c7d0f4d3ec58a2e46f6fb
        version: 11.11.0
      entries:
      - name: webotx-operator.v11.11.0
        version: 11.11.0
      name: stable
    defaultChannel: stable
    packageName: webotx-operator
    provider:
      name: NEC Corporation
      url: https://jpn.nec.com/webotx/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator-certified-rhmp.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a",
                  "imagePullPolicy": "Always",
                  "initImage": "registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-server-mongodb-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-server-mongodb-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:3c03fb18a5d73876576b80aca2b61e7c84ca052839ccb467d14e15f6d1f9d834
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a57926b53eef862b70a53818a7ccb0e9756f878f30d4128b608090b109bdc585
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator-certified-rhmp.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator-certified-rhmp.v1.20.0
        version: 1.20.0
      - name: percona-server-mongodb-operator-certified-rhmp.v1.19.1
        version: 1.19.1
      - name: percona-server-mongodb-operator-certified-rhmp.v1.19.0
        version: 1.19.0
      - name: percona-server-mongodb-operator-certified-rhmp.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator-certified-rhmp
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kms-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kms-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Alias",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Key",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kms-controller:1.1.1
          createdAt: "2025-09-19T22:02:02Z"
          description: AWS KMS controller is a service controller for managing KMS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alias represents the state of an AWS kms Alias resource.
            displayName: Alias
            kind: Alias
            name: aliases.kms.services.k8s.aws
            version: v1alpha1
          - description: Grant represents the state of an AWS kms Grant resource.
            displayName: Grant
            kind: Grant
            name: grants.kms.services.k8s.aws
            version: v1alpha1
          - description: Key represents the state of an AWS kms Key resource.
            displayName: Key
            kind: Key
            name: keys.kms.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Key Management Service (KMS) resources in AWS from within your Kubernetes cluster.

          **About Amazon KMS**

          AWS Key Management Service (AWS KMS) is a managed service that makes it easy for you to create and control the cryptographic keys that are used to protect your data. AWS KMS uses hardware security modules (HSM) to protect and validate your AWS KMS keys under the [FIPS 140-2 Cryptographic Module Validation Program](https://csrc.nist.gov/projects/cryptographic-module-validation-program/Certificate/3139), except in the China (Beijing) and China (Ningxia) Regions.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon KMS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kms
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon KMS Developer Resources
          url: https://aws.amazon.com/kms/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kms maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/kms-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-kms-controller.v1.1.1
        version: 1.1.1
      - name: ack-kms-controller.v1.1.0
        version: 1.1.0
      - name: ack-kms-controller.v1.0.31
        version: 1.0.31
      - name: ack-kms-controller.v1.0.30
        version: 1.0.30
      - name: ack-kms-controller.v1.0.29
        version: 1.0.29
      - name: ack-kms-controller.v1.0.28
        version: 1.0.28
      - name: ack-kms-controller.v1.0.27
        version: 1.0.27
      - name: ack-kms-controller.v1.0.26
        version: 1.0.26
      - name: ack-kms-controller.v1.0.25
        version: 1.0.25
      - name: ack-kms-controller.v1.0.24
        version: 1.0.24
      - name: ack-kms-controller.v1.0.23
        version: 1.0.23
      - name: ack-kms-controller.v1.0.22
        version: 1.0.22
      - name: ack-kms-controller.v1.0.19
        version: 1.0.19
      - name: ack-kms-controller.v1.0.18
        version: 1.0.18
      - name: ack-kms-controller.v1.0.17
        version: 1.0.17
      - name: ack-kms-controller.v1.0.16
        version: 1.0.16
      - name: ack-kms-controller.v1.0.15
        version: 1.0.15
      - name: ack-kms-controller.v1.0.14
        version: 1.0.14
      - name: ack-kms-controller.v1.0.13
        version: 1.0.13
      - name: ack-kms-controller.v1.0.12
        version: 1.0.12
      - name: ack-kms-controller.v1.0.8
        version: 1.0.8
      - name: ack-kms-controller.v1.0.7
        version: 1.0.7
      - name: ack-kms-controller.v1.0.5
        version: 1.0.5
      - name: ack-kms-controller.v1.0.4
        version: 1.0.4
      - name: ack-kms-controller.v1.0.3
        version: 1.0.3
      - name: ack-kms-controller.v1.0.2
        version: 1.0.2
      - name: ack-kms-controller.v1.0.0
        version: 1.0.0
      - name: ack-kms-controller.v0.1.5
        version: 0.1.5
      - name: ack-kms-controller.v0.1.4
        version: 0.1.4
      - name: ack-kms-controller.v0.1.3
        version: 0.1.3
      - name: ack-kms-controller.v0.1.2
        version: 0.1.2
      - name: ack-kms-controller.v0.1.1
        version: 0.1.1
      - name: ack-kms-controller.v0.1.0
        version: 0.1.0
      - name: ack-kms-controller.v0.0.20
        version: 0.0.20
      - name: ack-kms-controller.v0.0.19
        version: 0.0.19
      - name: ack-kms-controller.v0.0.18
        version: 0.0.18
      - name: ack-kms-controller.v0.0.17
        version: 0.0.17
      - name: ack-kms-controller.v0.0.16
        version: 0.0.16
      - name: ack-kms-controller.v0.0.15
        version: 0.0.15
      - name: ack-kms-controller.v0.0.14
        version: 0.0.14
      - name: ack-kms-controller.v0.0.13
        version: 0.0.13
      - name: ack-kms-controller.v0.0.12
        version: 0.0.12
      - name: ack-kms-controller.v0.0.11
        version: 0.0.11
      - name: ack-kms-controller.v0.0.10
        version: 0.0.10
      - name: ack-kms-controller.v0.0.9
        version: 0.0.9
      - name: ack-kms-controller.v0.0.8
        version: 0.0.8
      - name: ack-kms-controller.v0.0.7
        version: 0.0.7
      - name: ack-kms-controller.v0.0.6
        version: 0.0.6
      - name: ack-kms-controller.v0.0.5
        version: 0.0.5
      - name: ack-kms-controller.v0.0.4
        version: 0.0.4
      - name: ack-kms-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-kms-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: boxboat-github-practice
      provider-url: https://github.com/boxboat-github-practice
    name: github-arc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: github-arc-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "github-practice.boxboat.com/v1alpha1",
                "kind": "ActionsRunnerController",
                "metadata": {
                  "name": "arc-sample"
                },
                "spec": {
                  "authSecret": {
                    "name": "ghauth"
                  },
                  "createRunnerNamespaces": true,
                  "openshift": true,
                  "runnerNamespaces": [
                    "test1",
                    "test2"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery, OpenShift Optional
          containerImage: ghcr.io/boxboat-github-practice/github-arc-operator:1.1.0
          createdAt: "2023-03-03T20:49:16Z"
          description: Operator for deploying the github actions runner controller
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/boxboat-github-practice/github-arc-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActionsRunnerController is the schema for the ActionsRunnerController
              API
            displayName: Actions Runner Controller
            kind: ActionsRunnerController
            name: actionsrunnercontrollers.github-practice.boxboat.com
            version: v1alpha1
        description: helm based operator for github actions runner controller
        displayName: github-arc-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - github
        - actions
        - boxboat
        - ibm
        - self-hosted
        - runners
        links:
        - name: Github Arc Operator
          url: https://github.com/boxboat-github-practice/github-arc-operator
        - name: Actions Runner Controller
          url: https://github.com/actions/actions-runner-controller
        - name: Boxboat
          url: https://boxboat.com/
        maintainers:
        - email: Michael.McMullen@ibm.com
          name: mbmcmullen27
        maturity: alpha
        provider:
          name: boxboat-github-practice
          url: https://github.com/boxboat-github-practice
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - ghcr.io/boxboat-github-practice/github-arc-operator:1.1.0
        version: 1.1.0
      entries:
      - name: github-arc-operator.v1.1.0
        version: 1.1.0
      - name: github-arc-operator.v1.0.4
        version: 1.0.4
      name: alpha
    defaultChannel: alpha
    packageName: github-arc-operator
    provider:
      name: boxboat-github-practice
      url: https://github.com/boxboat-github-practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: special-resource-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: special-resource-operator.v4.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sro.openshift.io/v1beta1",
                "kind": "SpecialResource",
                "metadata": {
                  "name": "simple-kmod"
                },
                "spec": {
                  "chart": {
                    "name": "simple-kmod",
                    "repository": {
                      "name": "example",
                      "url": "file:///charts/example"
                    },
                    "version": "0.0.1"
                  },
                  "driverContainer": {
                    "source": {
                      "git": {
                        "ref": "master",
                        "uri": "https://github.com/openshift-psap/kvc-simple-kmod.git"
                      }
                    }
                  },
                  "namespace": "simple-kmod",
                  "set": {
                    "apiVersion": "sro.openshift.io/v1beta1",
                    "buildArgs": [
                      {
                        "name": "KMODVER",
                        "value": "SRO"
                      }
                    ],
                    "kind": "Values",
                    "kmodNames": [
                      "simple-kmod",
                      "simple-procfs-kmod"
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional, Storage
          containerImage: quay.io/openshift-psap/special-resource-operator:release-4.9
          description: The special resource operator is an orchestrator for resources
            in a cluster required to deploy special hardware and software. Examples
            include hardware accelerators or software defined storage, where extra
            management or kernel modules are needed.
          olm.skipRange: '>=4.6.0 <4.9.0'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpecialResource is the Schema for the specialresources API
            displayName: Special Resource
            kind: SpecialResource
            name: specialresources.sro.openshift.io
            version: v1beta1
          required:
          - description: Discovers special resources and labels the nodes with the
              PCI vendor ID.
            displayName: Node Feature Discovery
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
        description: "## About the managed application \nThe SRO will deploy the complete
          stack for enabling the accelerator in a cluster. \nIt manages the drivers,
          device-plugin, node-exporter, grafana and feature discovery.\n## About this
          Operator\n README: [https://github.com/openshift-psap/special-resource-operator](https://github.com/openshift-psap/special-resource-operator)\n\n##
          Prerequisites for enabling this Operator\nThe SRO has a dependency on NFD
          (Node Feature Discovery)\n"
        displayName: Special Resource Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GPI
        links:
        - name: README
          url: https://github.com/openshift-psap/special-resource-operator
        maintainers:
        - email: zkaiser@redhat.com
          name: Zvonko Kaiser
        - email: dagray@redhat.com
          name: David Gray
        maturity: alpha
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy
        - quay.io/openshift-psap/special-resource-operator:release-4.9
        version: 4.9.0
      entries:
      - name: special-resource-operator.v4.9.0
        version: 4.9.0
      - name: special-resource-operator.v4.8.2
        version: 4.8.2
      - name: special-resource-operator.v4.8.1
        version: 4.8.1
      - name: special-resource-operator.v4.8.0
        version: 4.8.0
      name: alpha
    defaultChannel: alpha
    packageName: special-resource-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Digital.ai
      provider-url: https://digital.ai
    name: digitalai-release-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: digitalai-release-operator.v25.3.0-beta.813
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 5,
                    "metrics": [],
                    "minReplicas": 3
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "cluster": {
                    "management": {
                      "ssl": {
                        "enabled": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    },
                    "manager": "xlr-legacy",
                    "mode": "full",
                    "name": "xl-release_cluster",
                    "node": {
                      "datacenter": "default"
                    },
                    "remote": {
                      "sslEngine": {
                        "enabled": false,
                        "enabledAlgorithms": [
                          "TLS_DHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_DHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_RSA_WITH_AES_128_CBC_SHA"
                        ],
                        "hostnameVerification": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "requireMutualAuthentication": false,
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"${XLR_CLUSTER_NAME}\"\n    manager = \"${XLR_CLUSTER_MANAGER}\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n      {{- if .Values.storage.reportingEngine.downingProviderClass }}\n      cluster {\n        downing-provider-class = {{.Values.storage.reportingEngine.downingProviderClass | quote}}\n      }\n      {{- end }}\n\n      {{- if .Values.cluster.remote.sslEngine.enabled }}\n      remote {\n        artery {  \n          ssl {  \n            config-ssl-engine {  \n              enabled-algorithms = [\n                {{- $last := sub (len .Values.cluster.remote.sslEngine.enabledAlgorithms) 1 }}\n                {{ range $index, $v := .Values.cluster.remote.sslEngine.enabledAlgorithms }}\n                {{ $v | quote }}{{ if ne $index $last }},{{ end }}\n                {{- end }}\n              ]  \n              key-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEY_PASSWORD}\"  \n              key-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.keystore \"default\" (printf \"cluster_remote_keystore.%s\" .Values.cluster.remote.sslEngine.keystoreType) \"context\" $) }}\"  \n              key-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEYSTORE_PASSWORD}\"  \n              trust-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.truststore \"default\" (printf \"cluster_remote_truststore.%s\" .Values.cluster.remote.sslEngine.truststoreType) \"context\" $) }}\"  \n              trust-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_TRUSTSTORE_PASSWORD}\"  \n              hostname-verification = {{ if .Values.cluster.remote.sslEngine.hostnameVerification }}on{{ else }}off{{ end }}\n              require-mutual-authentication = {{ if .Values.cluster.remote.sslEngine.requireMutualAuthentication }}on{{ else }}off{{ end }}  \n            }  \n          }  \n          transport = tls-tcp  \n        }  \n      }\n      {{- end }}\n    }\n    {{- if .Values.cluster.management.ssl.enabled }}\n    xlr-pekko-native {\n      pekko {\n        management {\n          http {  \n            route-providers-read-only = true  \n          }  \n          ssl {  \n            enabled = true\n            keyPassword = \"${?CLUSTER_MANAGEMENT_SSL_KEY_PASSWORD}\"  \n            keyStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.keystore \"default\" (printf \"cluster_management_keystore.%s\" .Values.cluster.management.ssl.keystoreType) \"context\" $) }}\"  \n            keyStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_KEYSTORE_PASSWORD}\"  \n            keystoreType = {{ .Values.cluster.management.ssl.keystoreType }}  \n            trustStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.truststore \"default\" (printf \"cluster_management_truststore.%s\" .Values.cluster.management.ssl.truststoreType) \"context\" $) }}\"  \n            trustStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_TRUSTSTORE_PASSWORD}\"  \n            trustStoreType = {{ .Values.cluster.management.ssl.truststoreType }}  \n          }  \n        }  \n      }\n    }\n    {{- end }}\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\nmanagement.endpoint.prometheus.enabled = {{ .Values.metrics.prometheus.enabled | default false }}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      datacenter = ${XLR_CLUSTER_DATACENTER}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "cd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "livenessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/liveness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "jmx": {
                      "enabled": false
                    },
                    "prometheus": {
                      "enabled": false,
                      "podAnnotations": {
                        "prometheus.io/port": "{{ .Values.service.ports.metrics }}",
                        "prometheus.io/scrape": "true"
                      },
                      "rule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "serviceMonitor": {
                        "annotations": {},
                        "basicAuth": {
                          "password": {
                            "key": "releasePassword",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          },
                          "username": {
                            "key": "releaseUsername",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          }
                        },
                        "enabled": false,
                        "honorLabels": false,
                        "interval": "30s",
                        "jobLabel": "",
                        "labels": {},
                        "metricRelabelings": [],
                        "namespace": "",
                        "params": {},
                        "path": "/s/actuator/prometheus",
                        "podTargetLabels": {},
                        "relabelings": [],
                        "scrapeTimeout": "",
                        "selector": {
                          "service-type": "normal"
                        },
                        "targetLabels": {}
                      }
                    }
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.0-debian-12-r4"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.0-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/extra/lib",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.5.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rbac": {
                    "create": true
                  },
                  "readinessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {
                      "service-type": "normal"
                    },
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "startupProbe": {
                    "enabled": true,
                    "failureThreshold": 12,
                    "initialDelaySeconds": 60,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 450,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnamilegacy/os-shell",
                      "tag": "12-debian-12-r49"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabsunsupported/release-operator@sha256:52914f17dc55dfaef55aed382e7269408fb676a213d12117eafa906d81506824 '
          createdAt: "2025-09-02T05:48:19Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql.\n\n```yaml\napiVersion:
          xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n
          \ global:\n    security:\n      allowInsecureImages: true\n  k8sSetup:\n
          \   platform: Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula:
          true\n  keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n```\n\n#### Configuration Details\n\nThe sample configuration
          uses:\n\n- Autogenerated limited license\n- Simple administrator initial
          password\n- Default keystore setup\n- Only one replica is set for all pods\n-
          All Pods are missing resource limit setup\n- Default storage class and minimal
          size for persistent storage\n- Release pods are using persistence access
          mode `ReadWriteOnce`, for Release pods the requirement is to have `ReadWriteMany`\n-
          Embedded PostgreSQL for DB management\n\n#### Security Configuration Details\n\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabsunsupported/release-operator@sha256:52914f17dc55dfaef55aed382e7269408fb676a213d12117eafa906d81506824
        - docker.io/xebialabsunsupported/xl-release@sha256:0548a1ba63da5ae09c59b67a30240bf48877ab8950e09b0f0f6fd7c88ba34c13
        version: 25.3.0-beta.813
      entries:
      - name: digitalai-release-operator.v25.3.0-beta.813
        version: 25.3.0-beta.813
      - name: digitalai-release-operator.v25.3.0-beta.711
        version: 25.3.0-beta.711
      - name: digitalai-release-operator.v25.3.0-beta.609
        version: 25.3.0-beta.609
      - name: digitalai-release-operator.v25.3.0-beta.512
        version: 25.3.0-beta.512
      name: beta
    - currentCSV: digitalai-release-operator.v24.1.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "fsGroup": null,
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "fsGroup": null,
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  jvm {\n      override-jvm-options = {{ .Values.jvm.overrideOptions }}\n      networkaddress.cache.ttl = {{ index .Values.jvm.options \"networkaddress.cache.ttl\" }}\n  }\n}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "ingressClass": "haproxy-dair",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-client/xl",
                        "kube",
                        "upgrade",
                        "--skip-context-check",
                        "--local-repo",
                        "/opt/xebialabs/xl-op-blueprints",
                        "--answers",
                        "/opt/xebialabs/xl-client/generated_answers.yaml"
                      ],
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    },
                    "overrideOptions": true
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "tag": "1.26.0-debian-12-r1"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "tag": "1.10.1-debian-12-r5"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "tag": "15.6.0-debian-12-r7"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "tag": "3.12.13-debian-12-r2"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnami/os-shell",
                      "tag": "12-debian-12-r16"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator without SCC creation and with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: docker.io/xebialabs/release-operator@sha256:47d6e31f66fce52c6566240b85fb028890409ad3addeea134120d1afa4814883
          createdAt: "2024-10-16T04:25:19Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind:
          DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    runAsUser: null\n    runAsGroup: null\n
          \   fsGroup: null\n  containerSecurityContext:\n    runAsUser: null\n    runAsGroup:
          null\n  securityContextConstraints:\n    enabled: false\n  volumePermissions:\n
          \   enabled: false\n  route:\n    enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n        runAsUser: null\n        runAsGroup: null\n        fsGroup:
          null\n      containerSecurityContext:\n        enabled: false\n        runAsUser:
          null\n        runAsGroup: null\n      securityContextConstraints:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n      runAsUser: null\n
          \     runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     enabled: false\n      runAsUser: null\n      runAsGroup: null\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"dair-doc-digitalai-release-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xlr-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Release pods are using
          persistence access mode `ReadWriteOnce`, for Release pods the requirement
          is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB management\n- Embedded
          RabbitMQ for message queue management\n\n#### Security Configuration Details\n\n-
          `securityContextConstraints.enabled: false` - disables creation of SCCs;\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:771a9a173e033a3ad8b46f5c00a7036eaa88c8d8d1fbd89217325168998113ea
        - docker.io/xebialabs/release-operator@sha256:47d6e31f66fce52c6566240b85fb028890409ad3addeea134120d1afa4814883
        - docker.io/xebialabs/xl-release@sha256:d1b65493ed524d37c60483f44b8bcc877ff398c81d36c7e767f2254dadec98cc
        version: 24.1.10
      entries:
      - name: digitalai-release-operator.v24.1.10
        version: 24.1.10
      name: stable
    - currentCSV: digitalai-release-operator.v24.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  jvm {\n      override-jvm-options = {{ .Values.jvm.overrideOptions }}\n      networkaddress.cache.ttl = {{ index .Values.jvm.options \"networkaddress.cache.ttl\" }}\n  }\n}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "image": {
                        "tag": "v0.14.7"
                      },
                      "ingressClass": "haproxy-dair",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-client/xl",
                        "kube",
                        "upgrade",
                        "--skip-context-check",
                        "--local-repo",
                        "/opt/xebialabs/xl-op-blueprints",
                        "--answers",
                        "/opt/xebialabs/xl-client/generated_answers.yaml"
                      ],
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    },
                    "overrideOptions": true
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.26.2-debian-12-r10"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "15.6.0-debian-12-r20"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.12.14-debian-12-r7"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnamilegacy/os-shell",
                      "tag": "12-debian-12-r49"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator without SCC creation and with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:aca76f0f5db170ddfc6ad692ea2874b9efb83fe732540946d88a0c078907ec8c '
          createdAt: "2025-08-22T08:54:25Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind:
          DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  securityContextConstraints:\n    enabled: false\n
          \ volumePermissions:\n    enabled: false\n  route:\n    enabled: false\n
          \   annotations:\n      haproxy.router.openshift.io/cookie_name: JSESSIONID\n
          \     haproxy.router.openshift.io/disable_cookies: 'false'\n      haproxy.router.openshift.io/rewrite-target:
          /\n      haproxy.router.openshift.io/timeout: 120s\n    hostname: '<mandatory-release-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n      securityContextConstraints:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   securityContextConstraints:\n      enabled: false\n    volumePermissions:\n
          \     enabled: false\n```\n\nIf SCC is required, it must be managed manually.
          The necessary permissions for accessing SecurityContextConstraints must
          be granted when SCC is enabled; otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"dair-doc-digitalai-release-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xlr-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Release pods are using
          persistence access mode `ReadWriteOnce`, for Release pods the requirement
          is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB management\n- Embedded
          RabbitMQ for message queue management\n\n#### Security Configuration Details\n\n-
          `securityContextConstraints.enabled: false` - disables creation of SCCs;\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/release-operator@sha256:aca76f0f5db170ddfc6ad692ea2874b9efb83fe732540946d88a0c078907ec8c
        - docker.io/xebialabs/xl-release@sha256:1cadd1d6ad7389ac9fe398ffba246b78238e54c180045b9dd44fdeefa3cf2af5
        version: 24.1.21
      entries:
      - name: digitalai-release-operator.v24.1.21
        version: 24.1.21
      - name: digitalai-release-operator.v24.1.20
        version: 24.1.20
      - name: digitalai-release-operator.v24.1.19
        version: 24.1.19
      - name: digitalai-release-operator.v24.1.18
        version: 24.1.18
      - name: digitalai-release-operator.v24.1.17
        version: 24.1.17
      - name: digitalai-release-operator.v24.1.14
        version: 24.1.14
      - name: digitalai-release-operator.v24.1.10
        version: 24.1.10
      name: stable-v24.1
    - currentCSV: digitalai-release-operator.v24.3.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "#!/bin/sh\n\ncd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "16.3.0-debian-12-r23"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n{{ if eq .Values.queueType \"quorum\" -}}\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n{{- end }}\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.13.7-debian-12-r5"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "queueType": "classic",
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-release",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:725fb5f7cfa4098c4110a40f54dc646f7663267bcba53603a7c1303c44051803 '
          createdAt: "2025-09-26T18:24:33Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n
          \ name: dair-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n```\n\n#### Configuration
          Details\n\nThe sample configuration uses:\n\n- Autogenerated limited license\n-
          Simple administrator initial password\n- Default keystore setup\n- Only
          one replica is set for all pods\n- All Pods are missing resource limit setup\n-
          Default storage class and minimal size for persistent storage\n- Release
          pods are using persistence access mode `ReadWriteOnce`, for Release pods
          the requirement is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB
          management\n- Embedded RabbitMQ for message queue management\n\n#### Security
          Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined ranges (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/release-operator@sha256:725fb5f7cfa4098c4110a40f54dc646f7663267bcba53603a7c1303c44051803
        - docker.io/xebialabs/xl-release@sha256:e9e87975888fcd4c8c164db90cb0a5c7abde88bf272c59d97e465af7b49bc6be
        version: 24.3.11
      entries:
      - name: digitalai-release-operator.v24.3.11
        version: 24.3.11
      - name: digitalai-release-operator.v24.3.10
        version: 24.3.10
      - name: digitalai-release-operator.v24.3.9
        version: 24.3.9
      - name: digitalai-release-operator.v24.3.8
        version: 24.3.8
      - name: digitalai-release-operator.v24.3.7
        version: 24.3.7
      - name: digitalai-release-operator.v24.3.6
        version: 24.3.6
      - name: digitalai-release-operator.v24.3.5
        version: 24.3.5
      name: stable-v24.3
    - currentCSV: digitalai-release-operator.v25.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "cluster": {
                    "management": {
                      "ssl": {
                        "enabled": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    },
                    "manager": "xlr-legacy",
                    "mode": "full",
                    "name": "xl-release_cluster",
                    "node": {
                      "datacenter": "default"
                    },
                    "remote": {
                      "sslEngine": {
                        "enabled": false,
                        "enabledAlgorithms": [
                          "TLS_DHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_DHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_RSA_WITH_AES_128_CBC_SHA"
                        ],
                        "hostnameVerification": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "requireMutualAuthentication": false,
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"${XLR_CLUSTER_NAME}\"\n    manager = \"${XLR_CLUSTER_MANAGER}\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n      {{- if .Values.storage.reportingEngine.downingProviderClass }}\n      cluster {\n        downing-provider-class = {{.Values.storage.reportingEngine.downingProviderClass | quote}}\n      }\n      {{- end }}\n\n      {{- if .Values.cluster.remote.sslEngine.enabled }}\n      remote {\n        artery {  \n          ssl {  \n            config-ssl-engine {  \n              enabled-algorithms = [\n                {{- $last := sub (len .Values.cluster.remote.sslEngine.enabledAlgorithms) 1 }}\n                {{ range $index, $v := .Values.cluster.remote.sslEngine.enabledAlgorithms }}\n                {{ $v | quote }}{{ if ne $index $last }},{{ end }}\n                {{- end }}\n              ]  \n              key-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEY_PASSWORD}\"  \n              key-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.keystore \"default\" (printf \"cluster_remote_keystore.%s\" .Values.cluster.remote.sslEngine.keystoreType) \"context\" $) }}\"  \n              key-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEYSTORE_PASSWORD}\"  \n              trust-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.truststore \"default\" (printf \"cluster_remote_truststore.%s\" .Values.cluster.remote.sslEngine.truststoreType) \"context\" $) }}\"  \n              trust-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_TRUSTSTORE_PASSWORD}\"  \n              hostname-verification = {{ if .Values.cluster.remote.sslEngine.hostnameVerification }}on{{ else }}off{{ end }}\n              require-mutual-authentication = {{ if .Values.cluster.remote.sslEngine.requireMutualAuthentication }}on{{ else }}off{{ end }}  \n            }  \n          }  \n          transport = tls-tcp  \n        }  \n      }\n      {{- end }}\n    }\n    {{- if .Values.cluster.management.ssl.enabled }}\n    xlr-pekko-native {\n      pekko {\n        management {\n          http {  \n            route-providers-read-only = true  \n          }  \n          ssl {  \n            enabled = true\n            keyPassword = \"${?CLUSTER_MANAGEMENT_SSL_KEY_PASSWORD}\"  \n            keyStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.keystore \"default\" (printf \"cluster_management_keystore.%s\" .Values.cluster.management.ssl.keystoreType) \"context\" $) }}\"  \n            keyStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_KEYSTORE_PASSWORD}\"  \n            keystoreType = {{ .Values.cluster.management.ssl.keystoreType }}  \n            trustStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.truststore \"default\" (printf \"cluster_management_truststore.%s\" .Values.cluster.management.ssl.truststoreType) \"context\" $) }}\"  \n            trustStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_TRUSTSTORE_PASSWORD}\"  \n            trustStoreType = {{ .Values.cluster.management.ssl.truststoreType }}  \n          }  \n        }  \n      }\n    }\n    {{- end }}\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\nmanagement.endpoint.prometheus.enabled = {{ .Values.metrics.prometheus.enabled | default false }}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      datacenter = ${XLR_CLUSTER_DATACENTER}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "#!/bin/sh\n\ncd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "livenessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/liveness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "jmx": {
                      "enabled": false
                    },
                    "prometheus": {
                      "enabled": false,
                      "podAnnotations": {
                        "prometheus.io/port": "{{ .Values.service.ports.metrics }}",
                        "prometheus.io/scrape": "true"
                      },
                      "rule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "serviceMonitor": {
                        "annotations": {},
                        "basicAuth": {
                          "password": {
                            "key": "releasePassword",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          },
                          "username": {
                            "key": "releaseUsername",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          }
                        },
                        "enabled": false,
                        "honorLabels": false,
                        "interval": "30s",
                        "jobLabel": "",
                        "labels": {},
                        "metricRelabelings": [],
                        "namespace": "",
                        "params": {},
                        "path": "/s/actuator/prometheus",
                        "podTargetLabels": {},
                        "relabelings": [],
                        "scrapeTimeout": "",
                        "selector": {
                          "service-type": "normal"
                        },
                        "targetLabels": {}
                      }
                    }
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.1-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.4.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.3-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "readinessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {
                      "service-type": "normal"
                    },
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "startupProbe": {
                    "enabled": true,
                    "failureThreshold": 12,
                    "initialDelaySeconds": 60,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 450,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-release",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:838443baf39514c5e0ac0a09cab411942a3cc4a13e261c8c60138788fe8d96e4 '
          createdAt: "2025-09-26T17:28:37Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n
          \ name: dair-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n```\n\n#### Configuration
          Details\n\nThe sample configuration uses:\n\n- Autogenerated limited license\n-
          Simple administrator initial password\n- Default keystore setup\n- Only
          one replica is set for all pods\n- All Pods are missing resource limit setup\n-
          Default storage class and minimal size for persistent storage\n- Release
          pods are using persistence access mode `ReadWriteOnce`, for Release pods
          the requirement is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB
          management\n- Embedded RabbitMQ for message queue management\n\n#### Security
          Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined ranges (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/release-operator@sha256:838443baf39514c5e0ac0a09cab411942a3cc4a13e261c8c60138788fe8d96e4
        - docker.io/xebialabs/xl-release@sha256:9b63789cdb87745832f1283d3f0e28ac5a4514d767895e0914a6be6672f2d14f
        version: 25.1.5
      entries:
      - name: digitalai-release-operator.v25.1.5
        version: 25.1.5
      - name: digitalai-release-operator.v25.1.4
        version: 25.1.4
      - name: digitalai-release-operator.v25.1.3
        version: 25.1.3
      - name: digitalai-release-operator.v25.1.2
        version: 25.1.2
      - name: digitalai-release-operator.v25.1.1
        version: 25.1.1
      - name: digitalai-release-operator.v25.1.0
        version: 25.1.0
      name: stable-v25.1
    defaultChannel: stable-v24.3
    packageName: digitalai-release-operator
    provider:
      name: Digital.ai
      url: https://digital.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Confluent, Inc.
      provider-url: https://confluent.io
    name: confluent-for-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: confluent-for-kubernetes.v2.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.8.4",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:0ca8769b04d6134ee231af84dfe98fb65161148d0d32e85f18dad6a3a345ed33
          createdAt: "2025-10-01T13:12:32Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0 <2.10.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - confluentinc/confluent-operator@sha256:0ca8769b04d6134ee231af84dfe98fb65161148d0d32e85f18dad6a3a345ed33
        - docker.io/confluentinc/confluent-operator@sha256:0ca8769b04d6134ee231af84dfe98fb65161148d0d32e85f18dad6a3a345ed33
        - docker.io/confluentinc/cp-server@sha256:45eb6e4f9c1201202d90216fbbdfcd8655eb378c937f09ac0129cd259956337d
        - docker.io/confluentinc/cp-server-connect@sha256:24aaa1cc8eb0df643f1efa37999b7ac33d844bfd1e4502da0d678996e4ac5ddd
        - docker.io/confluentinc/confluent-init-container@sha256:175ba68c876eca5804597b1591fdf551b2091d2f1c50e4daed2c32958af5d649
        - docker.io/confluentinc/cp-zookeeper@sha256:3e0af6ac6d34dcb8b9ed8f7a08109a48b2b2acd331c1dc4ea875222685c9afbd
        - docker.io/confluentinc/cp-schema-registry@sha256:bba639a3f9e2bf5f99b5d056ed520d2c268a125bee4f546d6cbca7a2c58f7dbb
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:2024d117dd413e207e49f37558d79dc882fd29f6b39e2d722708f18076813361
        - docker.io/confluentinc/cp-kafka-rest@sha256:8218e8db4731219ceaf9dbc378ead9c49aa86f98bc5106f62fbf70958dfd1ee4
        - docker.io/confluentinc/cp-ksqldb-server@sha256:093e87359c176d13d81317e972888a835cd01092d061bfc03fd60f0d1b4c8428
        version: 2.10.3
      entries:
      - name: confluent-for-kubernetes.v2.10.3
        version: 2.10.3
      name: "2.10"
    - currentCSV: confluent-for-kubernetes.v2.11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.9.3",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:85b512d3f20dc6e4a4da1759ee5ed306a4da0604592563a9af35be2cdf9ba1b6
          createdAt: "2025-10-01T13:10:39Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.10.0 <2.11.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/confluent-operator@sha256:85b512d3f20dc6e4a4da1759ee5ed306a4da0604592563a9af35be2cdf9ba1b6
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:351b68b062824c15db7ef999ba04062defdd4ea1ebe9665932147f99cf74879d
        - docker.io/confluentinc/cp-ksqldb-server@sha256:14a44cbbc13094974e7a5e14b33a829480ec3da75868130fdc7abaeda707dc5e
        - confluentinc/confluent-operator@sha256:85b512d3f20dc6e4a4da1759ee5ed306a4da0604592563a9af35be2cdf9ba1b6
        - docker.io/confluentinc/cp-server@sha256:e61e6c122b50ed345bc3e4e39cda6c93430e669535f8a70957e221563a7af1bd
        - docker.io/confluentinc/confluent-init-container@sha256:bccc7b723afdbe4e04c668720cdb1570835f8fd58ab7dd9a06ecf807f1176019
        - docker.io/confluentinc/cp-kafka-rest@sha256:cc72b3585ef9e7557910ab3efe3082444abe1fee191cdfb8747907fd2643971f
        - docker.io/confluentinc/cp-schema-registry@sha256:bd974dda5abf35fcad66c25fad92ee684e82305435b291b953d221f74d2bc119
        - docker.io/confluentinc/cp-zookeeper@sha256:03eadde2afa390f7339707e6baef2ecf7affb898aaa432378f61da54f5636ba3
        - docker.io/confluentinc/cp-server-connect@sha256:0d3b81e511cdf820daac4a2458b3cada40b4a7c3d142062a48880d38186bbcc4
        version: 2.11.3
      entries:
      - name: confluent-for-kubernetes.v2.11.3
        version: 2.11.3
      name: "2.11"
    - currentCSV: confluent-for-kubernetes.v2.8.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.6.4",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:551b14777b121b1e083524a0d5cce03cbe2fd2833537751bcd6be6cf9801f95e
          createdAt: "2024-12-02T06:51:43Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.0 <2.8.5'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-server@sha256:1b2ae325f87047b83d37d506066f073c27b0c42a0fd3c3409aa079ba736d1975
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:01bcde2aa35c0ad062ece49c63d41bf487618ea3c33e42ea4c23fde7194b26c9
        - docker.io/confluentinc/cp-schema-registry@sha256:dec88d30c8f21bc7fb38d28603c97c81c666ccaee63ff0136ecff48b3b3ec775
        - docker.io/confluentinc/confluent-init-container@sha256:b1ad1aa4daa81967a93b02323172f96db6a50c7135866f016344b2eb1a736a55
        - docker.io/confluentinc/cp-server-connect@sha256:255875bc57d81c322d47d5f6f11bc217e99ae709d1fdc5aec8e11829dd34279a
        - confluentinc/confluent-operator@sha256:551b14777b121b1e083524a0d5cce03cbe2fd2833537751bcd6be6cf9801f95e
        - docker.io/confluentinc/cp-zookeeper@sha256:7e728feebfc90a693d028216da68a5d79e5e8212855c63bac08e5b10b141d143
        - docker.io/confluentinc/cp-kafka-rest@sha256:2a5bc0d39693682e25d79f5036e18b71376df0941fedc2cb962a4b4d5a5eff82
        - docker.io/confluentinc/cp-ksqldb-server@sha256:50e4834c44759ab4a2a0bbab0f20545e124325b7ee12ba1a275af73b57084104
        - docker.io/confluentinc/confluent-operator@sha256:551b14777b121b1e083524a0d5cce03cbe2fd2833537751bcd6be6cf9801f95e
        version: 2.8.5
      entries:
      - name: confluent-for-kubernetes.v2.8.5
        version: 2.8.5
      name: "2.8"
    - currentCSV: confluent-for-kubernetes.v2.9.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.7.5",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:20648758fdb98d0f8913e75fb919ad0b622de8ecda9ab394912103e5e17d4007
          createdAt: "2025-10-03T06:52:55Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0 <2.9.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-server-connect@sha256:560833472dbdf23cf85b780e8306ae1bfe16ce6e2c0a9b326c64a513117525f1
        - docker.io/confluentinc/cp-zookeeper@sha256:2bd46d1f5d1eec0675bb5d1c08634d6458bbac7cacfb233aa61c68137fd4a7ce
        - docker.io/confluentinc/cp-kafka-rest@sha256:b6ac629a880474d81182eaf8459b4d83fe8126db796f1f285798beecda55a50a
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:257c38a75b3f4241780e20f16a7adc026dcca590f2e9c8da7c4e86b961e7611f
        - confluentinc/confluent-operator@sha256:20648758fdb98d0f8913e75fb919ad0b622de8ecda9ab394912103e5e17d4007
        - docker.io/confluentinc/confluent-operator@sha256:20648758fdb98d0f8913e75fb919ad0b622de8ecda9ab394912103e5e17d4007
        - docker.io/confluentinc/cp-ksqldb-server@sha256:6817a555ae4c041869f9b422a115c22894b2cd3988e82dc601694ce713b3ccd4
        - docker.io/confluentinc/confluent-init-container@sha256:b138d954318cc98767c10b4564a1412a22766317a83b9a452285bb0b3d893c63
        - docker.io/confluentinc/cp-server@sha256:79243e7e8b721f9fdf26fc57330ebb27d69c3b4686dbdf5319f6f39c2c7be836
        - docker.io/confluentinc/cp-schema-registry@sha256:6506a8c659d4154ae3e2d702f08b4ccac63513392ba454cbc9cf384d070c46eb
        version: 2.9.7
      entries:
      - name: confluent-for-kubernetes.v2.9.7
        version: 2.9.7
      name: "2.9"
    - currentCSV: confluent-for-kubernetes.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:8.0.1",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-ng-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "alertManagerClient": {
                      "url": "http://controlcenter-next-gen.confluent.svc.cluster.local:9093"
                    },
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    },
                    "prometheusClient": {
                      "url": "http://controlcenter-next-gen.confluent.svc.cluster.local:9090"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center-next-gen:2.2.0",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "replicas": 1,
                  "services": {
                    "alertmanager": {
                      "image": "confluentinc/cp-enterprise-alertmanager:2.2.0"
                    },
                    "prometheus": {
                      "image": "confluentinc/cp-enterprise-prometheus:2.2.0",
                      "pvc": {
                        "dataVolumeCapacity": "10Gi"
                      }
                    }
                  },
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:8.0.1",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:8.0.1",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:8.0.1",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:8.0.1",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:8.0.1",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:55b61717690fe334b5eebf5e7a55590537abddbd404f2eede81d4dd8cd29eda4
          createdAt: "2025-10-03T07:59:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <3.0.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-kafka-rest@sha256:f45f23e891ae073037ed7670659ecb6c3f72f9eea33e109f16243eec24a287a3
        - docker.io/confluentinc/cp-server-connect@sha256:4739f3a07d9f8eadd58f448a8e05b1392ef9b1dc9e9085d25a71f32b5b806a6e
        - docker.io/confluentinc/cp-ksqldb-server@sha256:df94bb805cff4f92315135a042f650c42f887bca70a40df8ecef2603532f11a7
        - docker.io/confluentinc/cp-enterprise-prometheus@sha256:db0989ef9859dbc88bc59f9b8ade86f2ba13e9dc36533e654566218328e5ff57
        - docker.io/confluentinc/cp-enterprise-control-center-next-gen@sha256:6378a0174433685ee52a7fe2e48424d3f598827ed2a5570ec74b916102485e0d
        - confluentinc/confluent-operator@sha256:55b61717690fe334b5eebf5e7a55590537abddbd404f2eede81d4dd8cd29eda4
        - docker.io/confluentinc/confluent-operator@sha256:55b61717690fe334b5eebf5e7a55590537abddbd404f2eede81d4dd8cd29eda4
        - docker.io/confluentinc/cp-schema-registry@sha256:2152975dfb62afbdf1d438dd0eb1f7fc3c929adf5804fe4790eccda0c96b9208
        - docker.io/confluentinc/cp-server@sha256:207cabc12e0d122199da875b2b0e39d1cf6f7fe435d5f184a9c8d7dc57ea81d9
        - docker.io/confluentinc/cp-enterprise-alertmanager@sha256:fcce6f5311c26cd58427a039530a9f8dfd009249ff5a519694aa945d34502252
        - docker.io/confluentinc/confluent-init-container@sha256:bf38b3bd9851bfd3ef0635e93b9e8b9ada570d393f650f60d8ae8943537730dc
        version: 3.0.1
      entries:
      - name: confluent-for-kubernetes.v3.0.1
        version: 3.0.1
      name: "3.0"
    defaultChannel: "3.0"
    packageName: confluent-for-kubernetes
    provider:
      name: Confluent, Inc.
      url: https://confluent.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pachyderm, Inc.
      provider-url: https://pachyderm.com
    name: pachyderm-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pachyderm-operator.v0.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "Pachyderm",
                "metadata": {
                  "name": "pachyderm-sample"
                },
                "spec": {
                  "console": {
                    "disable": true
                  },
                  "pachd": {
                    "metrics": {
                      "disable": false
                    },
                    "storage": {
                      "amazon": {
                        "credentialSecretName": "pachyderm-aws-secret"
                      },
                      "backend": "AMAZON"
                    }
                  },
                  "version": "v2.1.6"
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermExport",
                "metadata": {
                  "name": "pachydermexport-sample"
                },
                "spec": {
                  "backup": {
                    "target": "pachyderm-sample"
                  }
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermImport",
                "metadata": {
                  "name": "pachydermimport-sample"
                },
                "spec": {
                  "backup": "backups/pachyderm-backup-20220610.tar.gz",
                  "destination": {
                    "name": "pachyderm-restore",
                    "namespace": "pachyderm-workloads"
                  },
                  "storageSecret": "s3-credentials-secret"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
          createdAt: "2021-05-20T08:05:00Z"
          description: Pachyderm provides the data layer that allows data science
            teams to productionize and scale their machine learning lifecycle with
            data driven automation, petabyte scalability and end-to-end reproducibility.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/pachyderm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/pachyderm-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pachyderm/openshift-operator
          support: Pachyderm, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PachydermExport is the Schema for the pachydermexports API
            displayName: Pachyderm Export
            kind: PachydermExport
            name: pachydermexports.aiml.pachyderm.com
            version: v1beta1
          - description: PachydermImport is the Schema for the pachydermimports API
            displayName: Pachyderm Import
            kind: PachydermImport
            name: pachydermimports.aiml.pachyderm.com
            version: v1beta1
          - description: Pachyderm is the Schema for the pachyderms API
            displayName: Pachyderm
            kind: Pachyderm
            name: pachyderms.aiml.pachyderm.com
            version: v1beta1
        description: "The Pachyderm operator aims to make deploying and managing Pachyderm
          instances on Openshift container platform easier.\n\nPachyderm is the data
          foundation for machine learning.  Pachyderm provides industry leading data
          versioning, pipelines and lineage that allow data science teams to automate
          the machine learning lifecycle and optimize their machine learning operations
          (MLOps). \n\n## Getting started\n\n1. Pachyderm uses an object store to
          store your data. Set the pachd storage backend at `pachyderm.spec.pachd.storage.backend`
          to `AMAZON`, `MICROSOFT`, `GOOGLE` or `MINIO` depending on your storage
          provider. Additionally, create a user with read/write access to the object
          store.\n\n2. Setup the user credentials in the section, `pachyderm.spec.pachd.storage`.
          Required fields depend on the backend type set in step 1 above.\n\n3. Finally,
          deploy the pachyderm cluster in its own namespace.\n\nFor more information
          on getting started, reference the documentation.\n"
        displayName: Pachyderm
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pachyderm
        - versioning
        - machine learning
        - artificial intelligence
        links:
        - name: Documentation
          url: https://docs.pachyderm.com/latest/
        maintainers:
        - email: kyle@pachyderm.io
          name: Kyle White
        - email: eochieng@redhat.com
          name: Edmund Ochieng
        maturity: beta
        minKubeVersion: v1.19.0
        provider:
          name: Pachyderm, Inc.
          url: https://pachyderm.com
        relatedImages:
        - registry.connect.redhat.com/pachyderm/worker@sha256:374f83d92899707bc87b0318fda6bffa78d2e9e523b827bb3d40693a2d1bdecd
        - registry.connect.redhat.com/pachyderm/etcd@sha256:7f330b51874379b1d4b71d94e37e0ecfa52b86d628296136970688cf204f53b5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
        - registry.connect.redhat.com/pachyderm/backup-handler@sha256:e2ef6bd8191e8ad58c765c0db8f5bb21ea7123caed0c0f58fe681a024201306d
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/pachd@sha256:116477fefdf413ea02c6ac3d013f1f91d3f6e4b16aed2408c9890832cdb41430
        - registry.redhat.io/rhel8/postgresql-13@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/init-utils@sha256:5c0258bfa0203079925648c0a7dda6d05a29b6f461854821a1eec541777ff359
        - registry.connect.redhat.com/pachyderm/pgbouncer@sha256:c3b6949227df24abb8934d111a34b44e87fb16194ac34a5d0fe284cff02e9ac0
        version: 0.1.4
      entries:
      - name: pachyderm-operator.v0.1.4
        version: 0.1.4
      name: stable
    defaultChannel: stable
    packageName: pachyderm-operator-rhmp
    provider:
      name: Pachyderm, Inc.
      url: https://pachyderm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: local-storage-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: local-storage-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "local.storage.openshift.io/v1",
                "kind": "LocalVolume",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "storageClassDevices": [
                    {
                      "devicePaths": [
                        "/dev/disk/by-id/ata-crucial"
                      ],
                      "fsType": "ext4",
                      "storageClassName": "foobar",
                      "volumeMode": "Filesystem"
                    }
                  ]
                }
              },
              {
                "apiVersion": "local.storage.openshift.io/v1alpha1",
                "kind": "LocalVolumeSet",
                "metadata": {
                  "name": "example-localvolumeset"
                },
                "spec": {
                  "deviceInclusionSpec": {
                    "deviceMechanicalProperties": [
                      "Rotational",
                      "NonRotational"
                    ],
                    "deviceTypes": [
                      "disk"
                    ],
                    "maxSize": "100G",
                    "minSize": "10G"
                  },
                  "maxDeviceCount": 10,
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "storageClassName": "example-storageclass",
                  "volumeMode": "Block"
                }
              },
              {
                "apiVersion": "local.storage.openshift.io/v1alpha1",
                "kind": "LocalVolumeDiscovery",
                "metadata": {
                  "name": "auto-discover-devices"
                },
                "spec": {
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-local-storage-rhel9-operator@sha256:6f6491f023ffa21eff6a9b938cb1869f362d6b4a5b04c27dc5f13f49622ca178
          createdAt: "2019-08-14T00:00:00Z"
          description: Configure and use local storage volumes.
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.17"}]'
          olm.skipRange: '>=4.3.0 <4.16.0-202509240123'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-local-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["localvolumediscoveryresults.local.storage.openshift.io"]'
          repository: https://github.com/openshift/local-storage-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manage local storage volumes for OpenShift
            displayName: Local Volume
            kind: LocalVolume
            name: localvolumes.local.storage.openshift.io
            version: v1
          - description: A Local Volume set allows you to filter a set of storage
              volumes, group them and create a dedicated storage class to consume
              storage from the set of volumes.
            displayName: Local Volume Set
            kind: LocalVolumeSet
            name: localvolumesets.local.storage.openshift.io
            version: v1alpha1
          - description: Discover list of potentially usable disks on the chosen set
              of nodes
            displayName: Local Volume Discovery
            kind: LocalVolumeDiscovery
            name: localvolumediscoveries.local.storage.openshift.io
            version: v1alpha1
          - description: Disc inventory of available disks from selected nodes
            displayName: Local Volume Discovery Result
            kind: LocalVolumeDiscoveryResult
            name: localvolumediscoveryresults.local.storage.openshift.io
            version: v1alpha1
        description: |
          Operator that configures local storage volumes for use in Kubernetes and OpenShift. OpenShift 4.2 and above are the only supported OpenShift versions.
        displayName: Local Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - local storage
        links:
        - name: Documentation
          url: https://github.com/openshift/local-storage-operator/tree/master/docs
        - name: Source Repository
          url: https://github.com/openshift/local-storage-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-local-storage-rhel9-operator@sha256:6f6491f023ffa21eff6a9b938cb1869f362d6b4a5b04c27dc5f13f49622ca178
        - registry.redhat.io/openshift4/ose-local-storage-diskmaker-rhel9@sha256:d8d8b21fce5930a36c9755d82037fce1ab81f5ae2d5fb1bfd643d500c4ec7fba
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel9@sha256:ec5847d5006d66a014315decd579e10104b086a06bfefa33430545b4ffcece71
        version: 4.16.0-202509240123
      entries:
      - name: local-storage-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: local-storage-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: local-storage-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: local-storage-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: local-storage-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: local-storage-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: local-storage-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: local-storage-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: local-storage-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: local-storage-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: local-storage-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: local-storage-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: local-storage-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: local-storage-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: local-storage-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: local-storage-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: local-storage-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: local-storage-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: local-storage-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: local-storage-operator.v4.16.0-202410031007
        version: 4.16.0-202410031007
      - name: local-storage-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: local-storage-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: local-storage-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: local-storage-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: local-storage-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: local-storage-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: local-storage-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: local-storage-operator.v4.16.0-202408021139
        version: 4.16.0-202408021139
      - name: local-storage-operator.v4.16.0-202407190036
        version: 4.16.0-202407190036
      - name: local-storage-operator.v4.16.0-202407111006
        version: 4.16.0-202407111006
      - name: local-storage-operator.v4.16.0-202407081937
        version: 4.16.0-202407081937
      - name: local-storage-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: local-storage-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: local-storage-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: local-storage-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: local-storage-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: security-profiles-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: security-profiles-operator.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "AppArmorProfile",
                "metadata": {
                  "annotations": {
                    "description": "Block writing to any files in the disk."
                  },
                  "name": "test-profile"
                },
                "spec": {
                  "abstract": {
                    "filesystem": {
                      "readOnlyPaths": [],
                      "readWritePaths": [],
                      "writeOnlyPaths": []
                    }
                  }
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "SecurityProfilesOperatorDaemon",
                "metadata": {
                  "name": "spod",
                  "namespace": "security-profiles-operator"
                },
                "spec": {
                  "enableAppArmor": false,
                  "enableLogEnricher": false,
                  "enableSelinux": false
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "ProfileRecording",
                "metadata": {
                  "name": "test-recording"
                },
                "spec": {
                  "kind": "SeccompProfile",
                  "podSelector": {
                    "matchLabels": {
                      "app": "alpine"
                    }
                  },
                  "recorder": "bpf"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "ProfileBinding",
                "metadata": {
                  "name": "profile-binding"
                },
                "spec": {
                  "image": "nginx:1.19.1",
                  "profileRef": {
                    "kind": "SeccompProfile",
                    "name": "profile-allow-unsafe"
                  }
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha2",
                "kind": "RawSelinuxProfile",
                "metadata": {
                  "name": "errorlogger"
                },
                "spec": {
                  "policy": "(blockinherit container)\n(allow process var_log_t ( dir ( open read getattr lock search ioctl add_name remove_name write ))) \n(allow process var_log_t ( file ( getattr read write append ioctl lock map open create  ))) \n(allow process var_log_t ( sock_file ( getattr read write append open  ))) \n"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "Blocks all syscalls."
                  },
                  "name": "profile-block-all"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_ERRNO"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "UNSAFE: Allows all syscalls whilst logging their use. Similar to running as unconfined in terms of enforcement."
                  },
                  "name": "profile-complain-unsafe"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_LOG"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "UNSAFE: Allows all syscalls. Similar to running as unconfined as it provides no enforcement."
                  },
                  "name": "profile-allow-unsafe"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_ALLOW"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "Enables complain mode whilst blocking high-risk syscalls. Some essential syscalls are allowed to decrease log noise."
                  },
                  "name": "profile-complain-block-high-risk"
                },
                "spec": {
                  "architectures": [
                    "SCMP_ARCH_X86_64",
                    "SCMP_ARCH_PPC64LE"
                  ],
                  "defaultAction": "SCMP_ACT_LOG",
                  "syscalls": [
                    {
                      "action": "SCMP_ACT_ALLOW",
                      "names": [
                        "exit",
                        "exit_group",
                        "futex",
                        "nanosleep"
                      ]
                    },
                    {
                      "action": "SCMP_ACT_ERRNO",
                      "names": [
                        "acct",
                        "add_key",
                        "bpf",
                        "clock_adjtime",
                        "clock_settime",
                        "create_module",
                        "delete_module",
                        "finit_module",
                        "get_kernel_syms",
                        "get_mempolicy",
                        "init_module",
                        "ioperm",
                        "iopl",
                        "kcmp",
                        "kexec_file_load",
                        "kexec_load",
                        "keyctl",
                        "lookup_dcookie",
                        "mbind",
                        "mount",
                        "move_pages",
                        "name_to_handle_at",
                        "nfsservctl",
                        "open_by_handle_at",
                        "perf_event_open",
                        "personality",
                        "pivot_root",
                        "process_vm_readv",
                        "process_vm_writev",
                        "ptrace",
                        "query_module",
                        "quotactl",
                        "reboot",
                        "request_key",
                        "set_mempolicy",
                        "setns",
                        "settimeofday",
                        "stime",
                        "swapoff",
                        "swapon",
                        "_sysctl",
                        "sysfs",
                        "umount2",
                        "umount",
                        "unshare",
                        "uselib",
                        "userfaultfd",
                        "ustat",
                        "vm86old",
                        "vm86"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha2",
                "kind": "SelinuxProfile",
                "metadata": {
                  "name": "errorlogger"
                },
                "spec": {
                  "allow": {
                    "var_log_t": {
                      "dir": [
                        "open",
                        "read",
                        "getattr",
                        "lock",
                        "search",
                        "ioctl",
                        "add_name",
                        "remove_name",
                        "write"
                      ],
                      "file": [
                        "getattr",
                        "read",
                        "write",
                        "append",
                        "ioctl",
                        "lock",
                        "map",
                        "open",
                        "create"
                      ],
                      "sock_file": [
                        "getattr",
                        "read",
                        "write",
                        "append",
                        "open"
                      ]
                    }
                  },
                  "inherit": [
                    {
                      "name": "container"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          createdAt: "2025-02-25T13:34:54Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.1 <0.9.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-security-profiles
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ProfileBinding is the Schema for the profilebindings API.
            displayName: Profile Binding
            kind: ProfileBinding
            name: profilebindings.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: ProfileRecording is the Schema for the profilerecordings
              API.
            displayName: Profile Recording
            kind: ProfileRecording
            name: profilerecordings.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: SeccompProfile is a cluster level specification for a seccomp
              profile. See https://github.com/opencontainers/runtime-spec/blob/master/config-linux.md#seccomp
            displayName: Seccomp Profile
            kind: SeccompProfile
            name: seccompprofiles.security-profiles-operator.x-k8s.io
            version: v1beta1
          - description: SecurityProfileNodeStatus is a per-node status of a security
              profile
            displayName: Security Profile Node Status
            kind: SecurityProfileNodeStatus
            name: securityprofilenodestatuses.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: SecurityProfilesOperatorDaemon is the Schema to configure
              the spod deployment.
            displayName: Security Profiles Operator Daemon
            kind: SecurityProfilesOperatorDaemon
            name: securityprofilesoperatordaemons.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - kind: RawSelinuxProfile
            name: rawselinuxprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha2
          - description: SelinuxProfile is the Schema for the selinuxprofiles API.
            displayName: Selinux Profile
            kind: SelinuxProfile
            name: selinuxprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha2
          - description: AppArmorProfile is a cluster level specification for an AppArmor
              profile.
            displayName: App Armor Profile
            kind: AppArmorProfile
            name: apparmorprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha1
        description: SPO is an operator which aims to make it easier for users to
          use SELinux, seccomp and AppArmor in Kubernetes clusters
        displayName: Security Profiles Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - selinux
        - seccomp
        - apparmor
        - ebpf
        links:
        - name: Security Profiles Operator
          url: https://github.com/openshift/security-profiles-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-security-profiles-rhel8-operator@sha256:828274d71c65ef820e6a8b652e3f102624d09a0609fb00bae7b32dc62ff37fd1
        - registry.redhat.io/compliance/openshift-selinuxd-rhel9@sha256:3278d69814a8d0d1e31dcdef454c24828aadbb468a3d5d3e15910c76481765bd
        - registry.redhat.io/compliance/openshift-selinuxd-rhel8@sha256:81273ff149482109417064b40d333edb131481c05547e4ffa6485475460a48ba
        version: 0.9.0
      entries:
      - name: security-profiles-operator.v0.9.0
        version: 0.9.0
      - name: security-profiles-operator.v0.8.6
        version: 0.8.6
      - name: security-profiles-operator.v0.8.5
        version: 0.8.5
      - name: security-profiles-operator.v0.8.4
        version: 0.8.4
      - name: security-profiles-operator.v0.8.2
        version: 0.8.2
      name: release-alpha-rhel-8
    defaultChannel: release-alpha-rhel-8
    packageName: security-profiles-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: prometheus-exporter-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometheus-exporter-operator.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.3scale.net/v1alpha1",
                "kind": "PrometheusExporter",
                "metadata": {
                  "name": "example-memcached"
                },
                "spec": {
                  "dbHost": "your-memcached-host",
                  "dbPort": 11211,
                  "grafanaDashboard": {
                    "label": {
                      "key": "autodiscovery",
                      "value": "enabled"
                    }
                  },
                  "type": "memcached"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/3scale/prometheus-exporter-operator:v0.7.0
          createdAt: "2023-09-07 00:00:00"
          description: Operator to setup 3rd party prometheus exporters, with a collection
            of grafana dashboards
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/3scale-ops/prometheus-exporter-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configures a prometheus exporter to monitor a memcached instance
            displayName: PrometheusExporter
            kind: PrometheusExporter
            name: prometheusexporters.monitoring.3scale.net
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK to centralize the setup of 3rd party prometheus exporters on **Kubernetes/OpenShift**, with a collection of grafana dashboards.

          By just providing a few parameters like *dbHost* or *dbPort* (operator manage the container image, port, argument, command, volumes... and also prometheus `ServiceMonitor` and `GrafanaDashboard` k8s objects), you can setup different prometheus exporters to monitor:
          * The **internals from different databases**
          * **HTTP/TCP endpoints** (availability, latency, SSL/TLS certificate expiration...)
          * Any available **cloudwatch metric from any AWS Service**
          * Sendgrid email statistics (delivered, bounces, errors, spam...)

          Current prometheus exporters types supported, managed by same prometheus-exporter-operator:
          * memcached
          * redis
          * mysql
          * postgresql
          * sphinx
          * es (elasticsearch)
          * cloudwatch
          * probe (blackbox)
          * sendgrid

          The operator manages the lifecycle of the following objects:
          * Deployment (one per CR)
          * Service (one per CR)
          * ServiceMonitor (optional, one per CR)
          * GrafanaDashboard (optional, one per Namespace)

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale-ops/prometheus-exporter-operator#documentation).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale-ops/prometheus-exporter-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:
          * Raising any issues you find using Prometheus Exporter Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale-ops/prometheus-exporter-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving [documentation](https://github.com/3scale-ops/prometheus-exporter-operator)
          * Talking about Prometheus Exporter Operator
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale-ops/prometheus-exporter-operator/issues).

          ### License
          Prometheus Exporter Operator is licensed under the [Apache 2.0 license](https://github.com/3scale-ops/prometheus-exporter-operator/blob/main/LICENSE)
        displayName: Prometheus Exporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - metrics
        - observability
        - prometheus
        - exporter
        - grafana
        - mysql
        - postgresql
        - memcached
        - redis
        - sphinx
        - elasticsearch
        - cloudwatch
        - blackbox
        - sendgrid
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/prometheus-exporter-operator
        maintainers:
        - email: 3scale-operations@redhat.com
          name: 3scale Ops
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/3scale/prometheus-exporter-operator:v0.7.0
        version: 0.7.0
      entries:
      - name: prometheus-exporter-operator.v0.7.0
        version: 0.7.0
      - name: prometheus-exporter-operator.v0.3.4
        version: 0.3.4
      name: alpha
    defaultChannel: alpha
    packageName: prometheus-exporter-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: mariadb-operator
      provider-url: https://github.com/mariadb-operator/mariadb-operator
    name: mariadb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-operator.v25.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.mariadb.mmontes.io/v1alpha1",
                "kind": "MariadbOperator",
                "metadata": {
                  "name": "mariadb-operator"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "caValidity": "35064h",
                    "certValidity": "8766h",
                    "enabled": true,
                    "extrArgs": [],
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "ha": {
                      "enabled": false,
                      "replicas": 3
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "lookaheadValidity": "2160h",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podSecurityContext": {},
                    "requeueDuration": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "automount": true,
                      "enabled": true,
                      "extraLabels": {},
                      "name": ""
                    },
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    },
                    "tolerations": []
                  },
                  "clusterName": "cluster.local",
                  "extrArgs": [],
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "fullnameOverride": "",
                  "ha": {
                    "enabled": false,
                    "replicas": 3
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "logLevel": "INFO",
                  "metrics": {
                    "enabled": false,
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "rbac": {
                    "enabled": true
                  },
                  "resources": {},
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "automount": true,
                    "enabled": true,
                    "extraLabels": {},
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "annotations": {},
                    "cert": {
                      "caPath": "/tmp/k8s-webhook-server/certificate-authority",
                      "certManager": {
                        "duration": "",
                        "enabled": false,
                        "issuerRef": {},
                        "renewBefore": ""
                      },
                      "path": "/tmp/k8s-webhook-server/serving-certs",
                      "secretAnnotations": {}
                    },
                    "extrArgs": [],
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "ha": {
                      "enabled": false,
                      "replicas": 3
                    },
                    "hostNetwork": false,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podSecurityContext": {},
                    "port": 10250,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "automount": true,
                      "enabled": true,
                      "extraLabels": {},
                      "name": ""
                    },
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    },
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup"
                },
                "spec": {
                  "args": [
                    "--single-transaction",
                    "--all-databases"
                  ],
                  "logLevel": "info",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "backups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Connection",
                "metadata": {
                  "name": "connection"
                },
                "spec": {
                  "database": "mariadb",
                  "healthCheck": {
                    "interval": "10s",
                    "retryInterval": "3s"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "params": {
                    "parseTime": "true"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "secretName": "connection",
                  "secretTemplate": {
                    "annotations": {
                      "k8s.mariadb.com/connection": "sample"
                    },
                    "databaseKey": "database",
                    "hostKey": "host",
                    "key": "dsn",
                    "labels": {
                      "k8s.mariadb.com/connection": "sample"
                    },
                    "passwordKey": "password",
                    "portKey": "port",
                    "usernameKey": "username"
                  },
                  "serviceName": "mariadb",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Database",
                "metadata": {
                  "name": "data-test"
                },
                "spec": {
                  "characterSet": "utf8",
                  "collate": "utf8_general_ci",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "retryInterval": "5s"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "ExternalMariaDB",
                "metadata": {
                  "name": "external-mariadb"
                },
                "spec": {
                  "connection": {
                    "healthCheck": {
                      "interval": "5s"
                    },
                    "secretName": "external-mariadb"
                  },
                  "host": "mariadb.default.svc.cluster.local",
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "port": 3306,
                  "username": "root"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "grant"
                },
                "spec": {
                  "database": "*",
                  "grantOption": true,
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "privileges": [
                    "SELECT",
                    "INSERT",
                    "UPDATE"
                  ],
                  "requeueInterval": "30s",
                  "retryInterval": "5s",
                  "table": "*",
                  "username": "user"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "MariaDB",
                "metadata": {
                  "name": "mariadb"
                },
                "spec": {
                  "connection": {
                    "secretName": "mariadb-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "database": "mariadb",
                  "galera": {
                    "enabled": true
                  },
                  "metrics": {
                    "enabled": true,
                    "passwordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "mariadb-metrics"
                    }
                  },
                  "myCnf": "[mariadb]\nbind-address=*\ndefault_storage_engine=InnoDB\nbinlog_format=row\ninnodb_autoinc_lock_mode=2\ninnodb_buffer_pool_size=1024M\nmax_allowed_packet=256M\n",
                  "passwordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-password"
                  },
                  "primaryConnection": {
                    "secretName": "mariadb-conn-primary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "primaryService": {
                    "type": "ClusterIP"
                  },
                  "replicas": 3,
                  "rootPasswordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-root"
                  },
                  "secondaryConnection": {
                    "secretName": "mariadb-conn-secondary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "secondaryService": {
                    "type": "ClusterIP"
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "storage": {
                    "size": "1Gi"
                  },
                  "updateStrategy": {
                    "type": "ReplicasFirstPrimaryLast"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "MaxScale",
                "metadata": {
                  "name": "maxscale-galera"
                },
                "spec": {
                  "admin": {
                    "guiEnabled": true,
                    "port": 8989
                  },
                  "auth": {
                    "generate": true
                  },
                  "config": {
                    "sync": {
                      "database": "mysql",
                      "interval": "5s",
                      "timeout": "10s"
                    }
                  },
                  "connection": {
                    "port": 3306,
                    "secretName": "mxs-galera-conn"
                  },
                  "guiKubernetesService": {
                    "metadata": {
                      "annotations": {
                        "metallb.universe.tf/loadBalancerIPs": "172.18.0.231"
                      }
                    },
                    "type": "LoadBalancer"
                  },
                  "kubernetesService": {
                    "annotations": {
                      "metallb.universe.tf/loadBalancerIPs": "172.18.0.224"
                    },
                    "type": "LoadBalancer"
                  },
                  "mariaDbRef": {
                    "name": "mariadb-galera"
                  },
                  "monitor": {
                    "cooperativeMonitoring": "majority_of_all",
                    "interval": "2s",
                    "params": {
                      "available_when_donor": "false",
                      "disable_master_failback": "false",
                      "disable_master_role_setting": "false"
                    }
                  },
                  "replicas": 3,
                  "requeueInterval": "10s",
                  "services": [
                    {
                      "listener": {
                        "params": {
                          "connection_metadata": "tx_isolation=auto"
                        },
                        "port": 3306,
                        "protocol": "MariaDBProtocol"
                      },
                      "name": "rw-router",
                      "params": {
                        "master_accept_reads": "true",
                        "max_replication_lag": "3s",
                        "max_slave_connections": "255",
                        "transaction_replay": "true",
                        "transaction_replay_attempts": "10",
                        "transaction_replay_timeout": "5s"
                      },
                      "router": "readwritesplit"
                    },
                    {
                      "listener": {
                        "port": 3307
                      },
                      "name": "rconn-master-router",
                      "params": {
                        "master_accept_reads": "true",
                        "max_replication_lag": "3s",
                        "router_options": "master"
                      },
                      "router": "readconnroute"
                    },
                    {
                      "listener": {
                        "port": 3308
                      },
                      "name": "rconn-slave-router",
                      "params": {
                        "max_replication_lag": "3s",
                        "router_options": "slave"
                      },
                      "router": "readconnroute"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "PhysicalBackup",
                "metadata": {
                  "name": "physicalbackup"
                },
                "spec": {
                  "compression": "bzip2",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "podAffinity": true,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "immediate": true,
                    "suspend": false
                  },
                  "serviceAccountName": "backup",
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "physicalbackups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "region": "us-east-1",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  },
                  "timeout": "1h"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore"
                },
                "spec": {
                  "backupRef": {
                    "name": "backup"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "targetRecoveryTime": "2023-12-19T09:00:00Z"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "SqlJob",
                "metadata": {
                  "name": "03-stars"
                },
                "spec": {
                  "database": "mariadb",
                  "dependsOn": [
                    {
                      "name": "01-users"
                    },
                    {
                      "name": "02-repos"
                    }
                  ],
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "sql": "CREATE TABLE IF NOT EXISTS stars (\n  id bigint PRIMARY KEY AUTO_INCREMENT,\n  user_id bigint NOT NULL,\n  repo_id bigint NOT NULL,\n  FOREIGN KEY (user_id) REFERENCES users(id) ON DELETE CASCADE,\n  FOREIGN KEY (repo_id) REFERENCES repos(id) ON DELETE CASCADE,\n  UNIQUE KEY (user_id, repo_id)\n);\nINSERT INTO stars(user_id, repo_id) \n  VALUES((SELECT id FROM users ORDER BY RAND() LIMIT 1), (SELECT id FROM repos ORDER BY RAND() LIMIT 1))\n  ON DUPLICATE KEY UPDATE id=id;\nDELETE FROM stars WHERE id = (SELECT id FROM stars ORDER BY RAND() LIMIT 1);\nSELECT r.name AS repo, COUNT(*) AS stars\nFROM stars s\nJOIN repos r\nON s.repo_id = r.id\nGROUP BY r.id\nORDER BY stars DESC;",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "user"
                },
                "spec": {
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxUserConnections": 20,
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "retryInterval": "5s"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: ghcr.io/mariadb-operator/mariadb-operator-helm:25.8.4
          createdAt: "2025-09-19T13:24:35Z"
          description: Run and operate MariaDB in a cloud native way
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mariadb-operator/mariadb-operator
          support: mariadb-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configures a backup
            displayName: Backup
            kind: Backup
            name: backups.k8s.mariadb.com
            version: v1alpha1
          - description: Configures a connection
            displayName: Connection
            kind: Connection
            name: connections.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a logical database
            displayName: Database
            kind: Database
            name: databases.k8s.mariadb.com
            version: v1alpha1
          - description: Configures an external MariaDB resources
            displayName: ExternalMariaDB
            kind: ExternalMariaDB
            name: externalmariadbs.k8s.mariadb.com
            version: v1alpha1
          - description: Grants permissions to an user in a database
            displayName: Grant
            kind: Grant
            name: grants.k8s.mariadb.com
            version: v1alpha1
          - description: Configures MariaDB helm chart based operator
            displayName: MariadbOperator
            kind: MariadbOperator
            name: mariadboperators.helm.mariadb.mmontes.io
            version: v1alpha1
          - description: Provisions a MariaDB instance
            displayName: MariaDB
            kind: MariaDB
            name: mariadbs.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a MaxScale database proxy
            displayName: MaxScale
            kind: MaxScale
            name: maxscales.k8s.mariadb.com
            version: v1alpha1
          - description: Configures a physical backup
            displayName: PhysicalBackup
            kind: PhysicalBackup
            name: physicalbackups.k8s.mariadb.com
            version: v1alpha1
          - description: Restores a backup
            displayName: Restore
            kind: Restore
            name: restores.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a SQL job
            displayName: SqlJob
            kind: SqlJob
            name: sqljobs.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a user
            displayName: User
            kind: User
            name: users.k8s.mariadb.com
            version: v1alpha1
        description: |
          Install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) via [OLM](https://olm.operatorframework.io/) using the [helm chart](https://artifacthub.io/packages/helm/mariadb-operator/mariadb-operator).

          This helm operator provides provides a 1:1 mapping between the official helm chart and the [`MariadbOperator`](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/config/samples/helm_v1alpha1_mariadboperator.yaml) CRD, allowing to install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) via OLM without having to do any change in the helm chart.

          Normally, you would install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) providing this `values.yaml` to the helm chart:
          ```yaml
          image:
            repository: ghcr.io/mariadb-operator/mariadb-operator
            pullPolicy: IfNotPresent
          logLevel: INFO
          ha:
            enabled: true
          metrics:
            enabled: true
            serviceMonitor:
              enabled: true
          webhook:
            cert:
              certManager:
                enabled: true
          ```

          This helm chart installation is abstracted in the [`MariadbOperator`](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/config/samples/helm_v1alpha1_mariadboperator.yaml) CRD, which will be reconciled by the helm operator:
          ```yaml
          apiVersion: helm.k8s.mariadb.com/v1alpha1
          kind: MariadbOperator
          metadata:
            name: mariadb-operator
          spec:
            image:
              repository: ghcr.io/mariadb-operator/mariadb-operator
              pullPolicy: IfNotPresent
            logLevel: INFO
            ha:
              enabled: true
            metrics:
              enabled: true
              serviceMonitor:
                enabled: true
            webhook:
              cert:
                certManager:
                  enabled: true
          ```

          Once you have installed the operator, you will able to install a [`MariaDB`](https://github.com/mariadb-operator/mariadb-operator/blob/main/examples/manifests/mariadb_v1alpha1_mariadb.yaml) instance. Refer to the main [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) documentation for getting started with the rest of CRDs.

          ## Documentation
          * [mariadb-operator](https://github.com/mariadb-operator/mariadb-operator/blob/main/README.md)
          * [mariadb-operator-helm](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/README.md)

          ## Releases
          This operator is automatically published in the following repositories whenever a new version of the [helm chart](https://artifacthub.io/packages/helm/mariadb-operator/mariadb-operator) is released:
          - [k8s-operatorhub/community-operators](https://github.com/k8s-operatorhub/community-operators)
          - [redhat-openshift-ecosystem/community-operators-prod](https://github.com/redhat-openshift-ecosystem/community-operators-prod)

          ## Roadmap
          Take a look at our [roadmap](https://github.com/mariadb-operator/mariadb-operator/blob/main/ROADMAP.md) and feel free to open an issue to suggest new features.

          ## Contributing
          We welcome and encourage contributions to this project! Please check our [contributing](https://github.com/mariadb-operator/mariadb-operator/blob/main/CONTRIBUTING.md) and [development](https://github.com/mariadb-operator/mariadb-operator/blob/main/docs/DEVELOPMENT.md) guides. PRs welcome!

          ## Get in touch
          Join us on Slack: **[MariaDB Community Slack](https://r.mariadb.com/join-community-slack)**.
        displayName: MariaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mariadb
        - mysql
        - operator
        - mariadb-operator
        - database
        - maxscale
        links:
        - name: GitHub
          url: https://github.com/mariadb-operator/mariadb-operator
        maintainers:
        - email: mariadb-operator@proton.me
          name: mmontes11
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: mariadb-operator
          url: https://github.com/mariadb-operator/mariadb-operator
        relatedImages:
        - ghcr.io/mariadb-operator/mariadb-operator-helm:25.8.4
        version: 25.8.4
      entries:
      - name: mariadb-operator.v25.8.4
        version: 25.8.4
      - name: mariadb-operator.v25.8.3
        version: 25.8.3
      - name: mariadb-operator.v25.8.2
        version: 25.8.2
      - name: mariadb-operator.v25.8.1
        version: 25.8.1
      - name: mariadb-operator.v25.8.0
        version: 25.8.0
      - name: mariadb-operator.v0.38.1
        version: 0.38.1
      - name: mariadb-operator.v0.38.0
        version: 0.38.0
      - name: mariadb-operator.v0.37.1
        version: 0.37.1
      - name: mariadb-operator.v0.37.0
        version: 0.37.0
      - name: mariadb-operator.v0.36.0
        version: 0.36.0
      - name: mariadb-operator.v0.35.1
        version: 0.35.1
      - name: mariadb-operator.v0.35.0
        version: 0.35.0
      - name: mariadb-operator.v0.34.0
        version: 0.34.0
      - name: mariadb-operator.v0.33.0
        version: 0.33.0
      - name: mariadb-operator.v0.32.0
        version: 0.32.0
      - name: mariadb-operator.v0.31.0
        version: 0.31.0
      - name: mariadb-operator.v0.30.0
        version: 0.30.0
      - name: mariadb-operator.v0.29.0
        version: 0.29.0
      - name: mariadb-operator.v0.28.1
        version: 0.28.1
      - name: mariadb-operator.v0.27.0
        version: 0.27.0
      - name: mariadb-operator.v0.26.0
        version: 0.26.0
      - name: mariadb-operator.v0.24.0
        version: 0.24.0
      - name: mariadb-operator.v0.23.1
        version: 0.23.1
      - name: mariadb-operator.v0.23.0
        version: 0.23.0
      - name: mariadb-operator.v0.22.0
        version: 0.22.0
      - name: mariadb-operator.v0.21.0
        version: 0.21.0
      - name: mariadb-operator.v0.20.0
        version: 0.20.0
      - name: mariadb-operator.v0.19.0
        version: 0.19.0
      - name: mariadb-operator.v0.18.0
        version: 0.18.0
      - name: mariadb-operator.v0.17.0
        version: 0.17.0
      - name: mariadb-operator.v0.16.0
        version: 0.16.0
      - name: mariadb-operator.v0.15.1
        version: 0.15.1
      - name: mariadb-operator.v0.15.0
        version: 0.15.0
      - name: mariadb-operator.v0.14.0
        version: 0.14.0
      - name: mariadb-operator.v0.13.0
        version: 0.13.0
      - name: mariadb-operator.v0.11.0
        version: 0.11.0
      - name: mariadb-operator.v0.10.0
        version: 0.10.0
      - name: mariadb-operator.v0.9.1
        version: 0.9.1
      - name: mariadb-operator.v0.9.0
        version: 0.9.0
      - name: mariadb-operator.v0.8.0
        version: 0.8.0
      - name: mariadb-operator.v0.7.1
        version: 0.7.1
      - name: mariadb-operator.v0.7.0
        version: 0.7.0
      - name: mariadb-operator.v0.6.7
        version: 0.6.7
      - name: mariadb-operator.v0.6.6
        version: 0.6.6
      - name: mariadb-operator.v0.6.5
        version: 0.6.5
      - name: mariadb-operator.v0.6.4
        version: 0.6.4
      name: alpha
    defaultChannel: alpha
    packageName: mariadb-operator
    provider:
      name: mariadb-operator
      url: https://github.com/mariadb-operator/mariadb-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Anchore Inc.
      provider-url: ""
    name: anchore-engine
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anchore-engine.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.anchore.io/v1alpha1",
                "kind": "AnchoreEngine",
                "metadata": {
                  "name": "anchoreengine-sample"
                },
                "spec": {
                  "anchoreGlobal": {
                    "defaultAdminEmail": "example@email.com"
                  },
                  "postgresql": {
                    "extraEnv": [
                      {
                        "name": "POSTGRESQL_USER",
                        "value": "anchoreengine"
                      },
                      {
                        "name": "POSTGRESQL_PASSWORD",
                        "value": "anchore-postgres,123"
                      },
                      {
                        "name": "POSTGRESQL_DATABASE",
                        "value": "anchore"
                      },
                      {
                        "name": "PGUSER",
                        "value": "postgres"
                      }
                    ],
                    "postgresDatabase": "anchore",
                    "postgresPassword": "anchore-postgres,123",
                    "postgresUser": "anchoreengine"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/anchore/engine-operator@sha256:baf0f5acd319bc4d0132c88922ee3f357fdcdc278d931f243d0cbec3f6d416c6
          createdAt: "2021-10-06T23:53:51Z"
          description: Anchore Engine - container image scanning service for policy-based
            security, best-practice and compliance enforcement.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/anchore/engine-operator
          support: anchore.slack.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys Anchore Engine
            displayName: Anchore Engine Operator
            kind: AnchoreEngine
            name: anchoreengines.charts.anchore.io
            version: v1alpha1
        description: "Anchore Engine is an open source software system that provides
          a centralized service for analyzing container images, \n scanning for security
          vulnerabilities, and enforcing deployment policies. Anchore Engine is provided
          as a Docker \n container image that can be run standalone or with an orchestration
          platform such as Kubernetes. Anchore Engine allows \n users to perform detailed
          analysis on their container images, run queries, produce reports, and define
          policies that \n can be used in CI/CD pipelines. Users can extend Anchore
          Engine with plugins that add new queries, image analysis, and \n policies.
          Anchore Engine can be accessed directly through a RESTful API or via the
          Anchore CLI.\n\n The Anchore Engine Operator is based on the official stable
          [Helm Chart](https://github.com/anchore/anchore-charts/tree/master/stable/anchore-engine)."
        displayName: Anchore Engine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - analysis
        - docker
        - anchore
        - anchore-engine
        - image
        - security
        - vulnerability
        - scanner
        links:
        - name: Anchore Inc.
          url: https://www.anchore.com
        - name: Anchore Engine
          url: https://github.com/anchore/anchore-engine
        maintainers:
        - email: bradyt@anchore.com
          name: Brady Todhunter
        - email: zach@anchore.com
          name: Zach Hill
        maturity: alpha
        provider:
          name: Anchore Inc.
        relatedImages:
        - registry.connect.redhat.com/anchore/engine-operator@sha256:baf0f5acd319bc4d0132c88922ee3f357fdcdc278d931f243d0cbec3f6d416c6
        - registry.connect.redhat.com/anchore/engine0@sha256:68e09d5ff695b46f24de3d7610f88dfc14b77b38758bbd68d03350dbae3deb0f
        - registry.redhat.io/rhel8/postgresql-96@sha256:e8177c5de05ccdd6d12e360b65c63c889b2de725d573ffdedd8914230b118639
        version: 1.0.1
      entries:
      - name: anchore-engine.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: anchore-engine
    provider:
      name: Anchore Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: https://github.com/Mellanox/maintenance-operator/
    name: nvidia-maintenance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nvidia-maintenance-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "MaintenanceOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "maintenanceoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "maintenanceoperatorconfig",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "maintenanceoperatorconfig-sample"
                },
                "spec": {
                  "logLevel": "info",
                  "maxNodeMaintenanceTimeSeconds": 1600,
                  "maxParallelOperations": 1,
                  "maxUnavailable": "30%"
                }
              },
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "nodemaintenance-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodemaintenance",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "cordon": true,
                  "drainSpec": {
                    "deleteEmptyDir": true,
                    "force": true,
                    "podEvictionFilters": [
                      {
                        "byResourceNameRegex": "nvidia.com/gpu-*"
                      },
                      {
                        "byResourceNameRegex": "nvidia.com/rdma*"
                      }
                    ],
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  },
                  "nodeName": "worker-node-01",
                  "requestorID": "some-cluster-component.vendor.com",
                  "waitForPodCompletion": {
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
          createdAt: "2025-07-13T12:33:06Z"
          description: Node maintenance in K8s cluster in a coordinated manner
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-maintenance-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: NVIDIA
          repository: https://github.com/Mellanox/maintenance-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MaintenanceOperatorConfig is the Schema for the maintenanceoperatorconfigs
              API
            displayName: Maintenance Operator Config
            kind: MaintenanceOperatorConfig
            name: maintenanceoperatorconfigs.maintenance.nvidia.com
            version: v1alpha1
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.maintenance.nvidia.com
            version: v1alpha1
        description: Coordinates And Performs Common Node Preparations For Maintenance
        displayName: NVIDIA Maintenance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node
        - maintenance
        - drain
        links:
        - name: NVIDIA Maintenance Operator
          url: https://github.com/Mellanox/maintenance-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/maintenance-operator/
        relatedImages:
        - ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
        version: 0.2.3
      entries:
      - name: nvidia-maintenance-operator.v0.2.3
        version: 0.2.3
      name: stable
    - currentCSV: nvidia-maintenance-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "MaintenanceOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "maintenanceoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "maintenanceoperatorconfig",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "maintenanceoperatorconfig-sample"
                },
                "spec": {
                  "logLevel": "info",
                  "maxNodeMaintenanceTimeSeconds": 1600,
                  "maxParallelOperations": 1,
                  "maxUnavailable": "30%"
                }
              },
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "nodemaintenance-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodemaintenance",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "cordon": true,
                  "drainSpec": {
                    "deleteEmptyDir": true,
                    "force": true,
                    "podEvictionFilters": [
                      {
                        "byResourceNameRegex": "nvidia.com/gpu-*"
                      },
                      {
                        "byResourceNameRegex": "nvidia.com/rdma*"
                      }
                    ],
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  },
                  "nodeName": "worker-node-01",
                  "requestorID": "some-cluster-component.vendor.com",
                  "waitForPodCompletion": {
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
          createdAt: "2025-07-13T12:33:06Z"
          description: Node maintenance in K8s cluster in a coordinated manner
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-maintenance-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: NVIDIA
          repository: https://github.com/Mellanox/maintenance-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MaintenanceOperatorConfig is the Schema for the maintenanceoperatorconfigs
              API
            displayName: Maintenance Operator Config
            kind: MaintenanceOperatorConfig
            name: maintenanceoperatorconfigs.maintenance.nvidia.com
            version: v1alpha1
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.maintenance.nvidia.com
            version: v1alpha1
        description: Coordinates And Performs Common Node Preparations For Maintenance
        displayName: NVIDIA Maintenance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node
        - maintenance
        - drain
        links:
        - name: NVIDIA Maintenance Operator
          url: https://github.com/Mellanox/maintenance-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/maintenance-operator/
        relatedImages:
        - ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
        version: 0.2.3
      entries:
      - name: nvidia-maintenance-operator.v0.2.3
        version: 0.2.3
      name: v0.2
    defaultChannel: v0.2
    packageName: nvidia-maintenance-operator
    provider:
      name: NVIDIA
      url: https://github.com/Mellanox/maintenance-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: mariadb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mariadb.openstack.org/v1beta1",
                "kind": "Galera",
                "metadata": {
                  "name": "openstack"
                },
                "spec": {
                  "replicas": 3,
                  "secret": "osp-secret",
                  "storageClass": "local-storage",
                  "storageRequest": "500M"
                }
              },
              {
                "apiVersion": "mariadb.openstack.org/v1beta1",
                "kind": "MariaDBAccount",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "mariadb-operator",
                    "app.kubernetes.io/instance": "mariadbaccount-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mariadbaccount",
                    "app.kubernetes.io/part-of": "mariadb-operator",
                    "mariaDBDatabaseName": "neutron"
                  },
                  "name": "neutron1"
                },
                "spec": {
                  "secret": "neutrondb-secret",
                  "userName": "neutron1"
                }
              },
              {
                "apiVersion": "mariadb.openstack.org/v1beta1",
                "kind": "MariaDBDatabase",
                "metadata": {
                  "labels": {
                    "dbName": "mariadb"
                  },
                  "name": "scott"
                },
                "spec": {
                  "name": "scott",
                  "secret": "tiger"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:39Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Galera is the Schema for the galeras API
            displayName: Galera
            kind: Galera
            name: galeras.mariadb.openstack.org
            version: v1beta1
          - description: MariaDBAccount is the Schema for the mariadbaccounts API
            displayName: Maria DBAccount
            kind: MariaDBAccount
            name: mariadbaccounts.mariadb.openstack.org
            version: v1beta1
          - description: MariaDBDatabase is the Schema for the mariadbdatabases API
            displayName: Maria DBDatabase
            kind: MariaDBDatabase
            name: mariadbdatabases.mariadb.openstack.org
            version: v1beta1
        description: MariaDB Operator
        displayName: MariaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Identity
        - MariaDB
        links:
        - name: MariaDB Operator
          url: https://github.com/openstack-k8s-operators/mariadb-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/mariadb-rhel9-operator@sha256:5ae64a3cf9bf5eaf244548c3bf2e8ef89c9436cb676721c1f1d1c2ff8285e7e5
        - registry.redhat.io/rhoso/openstack-mariadb-rhel9@sha256:13119fe0ec56263a2bff3fc9c3892ea6386b837955280d8488f05d8ea6c4e44d
        version: 1.0.6
      entries:
      - name: mariadb-operator.v1.0.6
        version: 1.0.6
      - name: mariadb-operator.v1.0.5
        version: 1.0.5
      - name: mariadb-operator.v1.0.4
        version: 1.0.4
      - name: mariadb-operator.v1.0.3
        version: 1.0.3
      - name: mariadb-operator.v1.0.2
        version: 1.0.2
      - name: mariadb-operator.v1.0.1
        version: 1.0.1
      - name: mariadb-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: mariadb-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-community-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.5.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "registry": "docker.io",
                  "tag": "5.4.6",
                  "oem": "",
                  "imagePullSecrets": "",
                  "psp": false,
                  "rbac": true,
                  "serviceAccount": "default",
                  "leastPrivilege": false,
                  "global": {
                    "cattle": {
                      "url": ""
                    },
                    "azure": {
                      "enabled": false,
                      "identity": {
                        "clientId": "DONOTMODIFY"
                      },
                      "marketplace": {
                        "planId": "DONOTMODIFY"
                      },
                      "extension": {
                        "resourceId": "DONOTMODIFY"
                      },
                      "serviceAccount": "csp",
                      "imagePullSecrets": "",
                      "images": {
                        "neuvector_csp_pod": {
                          "tag": "latest",
                          "image": "neuvector-billing-azure-by-suse-llc",
                          "registry": "registry.suse.de/suse/sle-15-sp5/update/pubclouds/images",
                          "imagePullPolicy": "IfNotPresent"
                        },
                        "controller": {
                          "tag": "5.2.4",
                          "image": "controller",
                          "registry": "docker.io/neuvector"
                        },
                        "manager": {
                          "tag": "5.2.4",
                          "image": "manager",
                          "registry": "docker.io/neuvector"
                        },
                        "enforcer": {
                          "tag": "5.2.4",
                          "image": "enforcer",
                          "registry": "docker.io/neuvector"
                        }
                      }
                    },
                    "aws": {
                      "enabled": false,
                      "accountNumber": "",
                      "roleName": "",
                      "serviceAccount": "csp",
                      "annotations": {},
                      "imagePullSecrets": "",
                      "image": {
                        "digest": "",
                        "repository": "neuvector/neuvector-csp-adapter",
                        "tag": "latest",
                        "imagePullPolicy": "IfNotPresent"
                      }
                    }
                  },
                  "bootstrapPassword": "",
                  "autoGenerateCert": true,
                  "defaultValidityPeriod": 365,
                  "internal": {
                    "certmanager": {
                      "enabled": false,
                      "secretname": "neuvector-internal"
                    },
                    "autoGenerateCert": true,
                    "autoRotateCert": true
                  },
                  "controller": {
                    "enabled": true,
                    "annotations": {},
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": {
                      "repository": "neuvector/controller",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "searchRegistries": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "nodePort": "",
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      }
                    },
                    "ranchersso": {
                      "enabled": false
                    },
                    "pvc": {
                      "enabled": false,
                      "existingClaim": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {
                        "userinitcfg.yaml": {
                          "users": [
                            {
                              "Fullname": "admin",
                              "Password": "",
                              "Role": "admin"
                            }
                          ]
                        }
                      }
                    },
                    "certupgrader": {
                      "env": [],
                      "schedule": "",
                      "imagePullPolicy": "IfNotPresent",
                      "timeout": 3600,
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "prime": {
                      "enabled": false,
                      "image": {
                        "repository": "neuvector/compliance-config",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "1.0.7",
                        "hash": ""
                      }
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": {
                      "repository": "neuvector/enforcer",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "env": [],
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/etcd"
                      }
                    ],
                    "resources": {},
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "manager": {
                    "enabled": true,
                    "image": {
                      "repository": "neuvector/manager",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "priorityClassName": "",
                    "env": {
                      "ssl": true,
                      "envs": []
                    },
                    "svc": {
                      "type": "ClusterIP",
                      "nodePort": "",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": "",
                      "tls": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "topologySpreadConstraints": [],
                    "affinity": {},
                    "podLabels": {},
                    "podAnnotations": {},
                    "tolerations": [],
                    "nodeSelector": {},
                    "runAsUser": "",
                    "probes": {
                      "enabled": false,
                      "timeout": 1,
                      "periodSeconds": 10,
                      "startupFailureThreshold": 30
                    }
                  },
                  "cve": {
                    "adapter": {
                      "enabled": false,
                      "image": {
                        "repository": "neuvector/registry-adapter",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "0.1.9",
                        "hash": ""
                      },
                      "priorityClassName": "",
                      "resources": {},
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt"
                      },
                      "harbor": {
                        "protocol": "https",
                        "secretName": ""
                      },
                      "svc": {
                        "type": "ClusterIP",
                        "loadBalancerIP": "",
                        "annotations": {}
                      },
                      "route": {
                        "enabled": true,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      },
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ingressClassName": "",
                        "path": "/",
                        "annotations": {
                          "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                        },
                        "tls": false,
                        "secretName": ""
                      },
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      }
                    },
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "cacert": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                      "image": {
                        "registry": "",
                        "repository": "neuvector/updater",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "0.0.5",
                        "hash": ""
                      },
                      "schedule": "0 0 * * *",
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": {
                        "registry": "",
                        "repository": "neuvector/scanner",
                        "imagePullPolicy": "Always",
                        "tag": "6",
                        "hash": ""
                      },
                      "priorityClassName": "",
                      "resources": {},
                      "topologySpreadConstraints": [],
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      },
                      "volumes": "",
                      "volumeMounts": ""
                    }
                  },
                  "resources": {},
                  "runtimePath": "",
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhooksvc": {
                    "enabled": true
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  },
                  "lease": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "false"
          containerImage: docker.io/neuvector/neuvector-operator@sha256:6815df6435eafaf7b2bc9d5e3f0c9b65f0a50ad4e05d2862e56b227e19aeca27
          createdAt: "2019-11-18T02:09:59Z"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: NeuVector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart. NeuVector Operator version 1.5.5 deploys NeuVector app version 5.4.6.

          **Complete below steps to create namespace and Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Privileged SCC is added to Service Account specified in the deploment yaml by Operator version 1.3.4 and above in new deployment. In case of upgrading NeuVector Operator from previous version to 1.3.4, please delete Privileged SCC before upgrading.

                   oc delete rolebinding -nneuvector system:openshift:scc:privileged

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s  neuvector/default


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - wavefront
        - wavefront collector
        - wavefront proxy
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - docker.io/neuvector/neuvector-operator@sha256:6815df6435eafaf7b2bc9d5e3f0c9b65f0a50ad4e05d2862e56b227e19aeca27
        version: 1.5.5
      entries:
      - name: neuvector-operator.v1.5.5
        version: 1.5.5
      - name: neuvector-operator.v1.5.4
        version: 1.5.4
      - name: neuvector-operator.v1.5.3
        version: 1.5.3
      - name: neuvector-operator.v1.5.2
        version: 1.5.2
      - name: neuvector-operator.v1.5.1
        version: 1.5.1
      - name: neuvector-operator.v1.5.0
        version: 1.5.0
      - name: neuvector-operator.v1.4.9
        version: 1.4.9
      - name: neuvector-operator.v1.4.8
        version: 1.4.8
      - name: neuvector-operator.v1.4.7
        version: 1.4.7
      - name: neuvector-operator.v1.4.6
        version: 1.4.6
      - name: neuvector-operator.v1.4.5
        version: 1.4.5
      - name: neuvector-operator.v1.4.4
        version: 1.4.4
      - name: neuvector-operator.v1.4.3
        version: 1.4.3
      - name: neuvector-operator.v1.4.2
        version: 1.4.2
      - name: neuvector-operator.v1.4.1
        version: 1.4.1
      - name: neuvector-operator.v1.4.0
        version: 1.4.0
      - name: neuvector-operator.v1.3.9
        version: 1.3.9
      - name: neuvector-operator.v1.3.8
        version: 1.3.8
      - name: neuvector-operator.v1.3.7
        version: 1.3.7
      - name: neuvector-operator.v1.3.6
        version: 1.3.6
      - name: neuvector-operator.v1.3.5
        version: 1.3.5
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-community-operator
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "docker.io/percona/percona-postgresql-operator:2.7.0-ppg17.5.2-postgres",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "docker.io/percona/percona-pgbouncer:1.24.1",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "docker.io/percona/percona-pgbackrest:2.55.0",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:3.2.0",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: docker.io/percona/percona-postgresql-operator:2.7.0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          olm.skipRange: <v2.7.0
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native

          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - docker.io/percona/percona-postgresql-operator:2.7.0
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator.v2.6.1
        version: 2.6.1
      - name: percona-postgresql-operator.v2.6.0
        version: 2.6.0
      - name: percona-postgresql-operator.v2.5.0
        version: 2.5.0
      name: stable
    defaultChannel: stable
    packageName: percona-postgresql-operator
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: GitLab, Inc.
      provider-url: https://gitlab.com
    name: gitlab-runner-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-runner-operator.v1.41.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta2",
                "kind": "Runner",
                "metadata": {
                  "name": "gitlab-runner-example",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "gitlabUrl": "https://gitlab.example.com",
                  "token": "runner-token-secret",
                  "tags": "kubernetes,example",
                  "concurrent": 10,
                  "imagePullPolicy": "IfNotPresent"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery, Developer Tools
          containerImage: registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
          createdAt: "2025-09-25T15:08:40Z"
          description: GitLab Runner operator manages lifecycle of GitLab Runner instances
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.11.0 <1.41.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: none
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
          support: GitLab, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Runner is the open source project used to run your jobs and
              send the results back to GitLab
            displayName: GitLab Runner
            kind: Runner
            name: runners.apps.gitlab.com
            version: v1beta2
        description: |-
          GitLab Runner is the lightweight, highly-scalable agent that runs your build jobs and sends the results back to a GitLab instance. GitLab Runner works in conjunction with GitLab CI/CD, the open-source continuous integration service included with GitLab.

              The GitLab Runner operator manages the lifecycle of GitLab Runner in Kubernetes or Openshift clusters. The operator aims to automate the tasks needed to run your CI/CD jobs in your container orchestration platform.

              ## Prerequisites

                For Kubernetes cluster, install cert-manager:

                ```shell
                kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.7.1/cert-manager.yaml
                ```

              ## GitLab Runner version

              This version of **GitLab Runner Operator** ships with **GitLab Runner v18.4.0**.

              To use a different version of **GitLab Runner** change the [`runnerImage` and `helperImage` properties](https://docs.gitlab.com/runner/configuration/configuring_runner_operator.html#operator-properties).

              ## Usage

               To link a GitLab Runner instance to a self-hosted GitLab instance or the hosted [GitLab](https://gitlab.com), you first need to:

               1. Create a secret containing the `runner-token` from your GitLab project.

                ```
                cat > gitlab-runner-secret.yml << EOF
                apiVersion: v1
                kind: Secret
                metadata:
                  name: gitlab-runner-secret
                type: Opaque
                # Only one of the following fields must be set. The Operator will fail to register Runner if both are provided.
                # # NOTE: runner-registration-token is deprecated and will be removed in GitLab 18.0. You should use runner-token instead.
                stringData:
                  runner-token: REPLACE_ME # your project runner token
                  # runner-registration-token: "" # your project runner secret
                EOF
                ```

                ```
                oc apply -f gitlab-runner-secret.yml
                ```

               2. Create the Custom Resource Definition (CRD) file and include the following information. The tags value must be openshift for the job to run.

                 ```
                 cat > gitlab-runner.yml << EOF
                 apiVersion: apps.gitlab.com/v1beta2
                 kind: Runner
                 metadata:
                   name: gitlab-runner
                 spec:
                   gitlabUrl: https://gitlab.example.com
                   buildImage: alpine
                   token: gitlab-runner-secret
                   tags: openshift
                 EOF
                 ```

                ```
                oc apply -f gitlab-runner.yml
                ```

              ## Full documentation

              Visit [Install GitLab Runner Operator](https://docs.gitlab.com/runner/install/operator.html)
        displayName: GitLab Runner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GitLab
        - CI/CD
        - DevOps
        - SAST
        - DAST
        links:
        - name: Gitlab Runner Operator
          url: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
        - name: GitLab Docs
          url: https://docs.gitlab.com
        - name: GitLab Runner
          url: https://docs.gitlab.com/runner/
        - name: GitLab CI/CD
          url: https://docs.gitlab.com/ee/ci/quick_start/
        maintainers:
        - email: support@gitlab.com
          name: Georgi N. Georgiev
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: GitLab, Inc.
          url: https://gitlab.com
        relatedImages:
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-ocp@sha256:0f2a5a3b6779d719eecbf079b99cf06b652603d74b563f57d753e32966d253dd
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-helper-ocp@sha256:29727fae83a226bcf57607240d4046a0b8c208b5ff42ee1f8ca0f115bb188809
        version: 1.41.1
      entries:
      - name: gitlab-runner-operator.v1.41.1
        version: 1.41.1
      - name: gitlab-runner-operator.v1.41.0
        version: 1.41.0
      - name: gitlab-runner-operator.v1.40.1
        version: 1.40.1
      - name: gitlab-runner-operator.v1.40.0
        version: 1.40.0
      - name: gitlab-runner-operator.v1.39.0
        version: 1.39.0
      name: stable
    defaultChannel: stable
    packageName: gitlab-runner-operator
    provider:
      name: GitLab, Inc.
      url: https://gitlab.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Omar Gaye IBM
      provider-url: omarg.net
    name: memcached-operator-ogaye
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: memcached-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.omarg.net/v1alpha1",
                "kind": "Memcached",
                "metadata": {
                  "name": "memcached-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.6.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Memcached is the Schema for the memcacheds API
            displayName: Memcached
            kind: Memcached
            name: memcacheds.cache.omarg.net
            version: v1alpha1
        description: Omar memcached operator
        displayName: memcached-operator-ogaye
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Omar
        - Memcached
        links:
        - name: Memcached Operator OGaye
          url: https://memcached-operator.domain
        maturity: alpha
        provider:
          name: Omar Gaye IBM
          url: omarg.net
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9211b70edbb93d718428e1c9a992e5fc74a20cb213f8a0592ffe340c362890fe
        - quay.io/ogaye/memcached-operator@sha256:5340ac11ce676dd3204bf3807a4c2333a8dddcbcec918db0aeafbc977163c1c6
        version: 0.0.1
      entries:
      - name: memcached-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: memcached-operator-ogaye
    provider:
      name: Omar Gaye IBM
      url: omarg.net
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Aqua Security, Inc.
      provider-url: ""
    name: aqua-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aqua-operator.2022.4.836
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaCsp",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "namespace": "aqua",
                    "version": "2022.4",
                    "requirements": true
                  },
                  "common": {
                    "dbDiskSize": 10,
                    "databaseSecret": {
                      "key": "db-password",
                      "name": "aqua-database-password"
                    }
                  },
                  "database": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "gateway": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "server": {
                    "replicas": 1,
                    "service": "LoadBalancer"
                  },
                  "route": true,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaDatabase",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "splitDB": false
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "diskSize": 10,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "runAsNonRoot": false,
                  "gateway": {
                    "host": "aqua-gateway",
                    "port": 8443
                  },
                  "token": "<<your-token>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaKubeEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "config": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "aqua-secure",
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "service": "ClusterIP"
                  },
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaGateway",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  }
                },
                "runAsNonRoot": false
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaScanner",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "deploy": {
                    "replicas": 1
                  },
                  "runAsNonRoot": false,
                  "login": {
                    "username": "<<YOUR AQUA USER NAME>>",
                    "password": "<<YOUR AQUA USER PASSWORD>>",
                    "host": "http://aqua-server:8080"
                  }
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaServer",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "LoadBalancer"
                  }
                },
                "runAsNonRoot": false
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaLightning",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "global": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "Default-cluster-name"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "kubeEnforcer": {
                    "infra": {
                      "serviceAccount": "aqua-kube-enforcer-sa"
                    },
                    "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                    "deploy": {
                      "service": "ClusterIP"
                    },
                    "starboard": {
                      "infra": {
                        "serviceAccount": "aqua-kube-enforcer-sa"
                      },
                      "config": {
                        "imagePullSecret": "aqua-registry"
                      },
                      "deploy": {
                        "replicas": 1
                      }
                    },
                    "env": []
                  },
                  "enforcer": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "deploy": null,
                    "token": "<<ENFORCER_GROUP_TOKEN>>",
                    "env": [
                    ],
                    "secret": {
                      "name": null,
                      "key": null
                    },
                    "runAsNonRoot": null,
                    "rhcosVersion": null
                  }
                }
              },
            {
                  "apiVersion": "operator.aquasec.com/v1alpha1",
                  "kind": "AquaCloudConnector",
                  "metadata": {
                    "name": "aqua"
                  },
                  "spec": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "common": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    },
                    "runAsNonRoot": false,
                    "gateway": {
                      "host": "aqua-gateway",
                      "port": "8443"
                    },
                    "login": {
                      "host": "aqua-gateway:8443",
                      "password": "<<YOUR AQUA USER PASSWORD>>",
                      "token": "<<YOUR AQUA CLOUD CONNECTOR TOKEN>>",
                      "username": "<<YOUR AQUA USER NAME>>"
                    },
                    "tunnels": [
                      {
                        "host": "<account>.azurecr.io",
                        "port": "443"
                      },
                      {
                        "host": "gke.gcr.io",
                        "port": "443"
                      },
                      {
                        "host": "<account>.jfrog.io",
                        "port": "443"
                      },
                      {
                        "host": "1.1.1.1",
                        "port": "443"
                      },
                      {
                        "region": "us-east1",
                        "type": "ecr"
                      }
                    ]
                  }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/aquasec/aquasec@sha256:c9f2c770a51e9664006c12d124ca18aab76bea8c32a6a253fe31d44c02633642
          createdAt: ""
          description: The Aqua Security Operator runs within a Openshift cluster
            and provides a means to deploy and manage Aqua Security cluster and components.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2022.4.836'
          operators.operatorframework.io/internal-objects: '["aquastarboards.aquasecurity.github.io","configauditreports.aquasecurity.github.io",
            "clusterconfigauditreports.aquasecurity.github.io"]'
          repository: https://github.com/aquasecurity/aqua-operator
          support: Aqua Security, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aqua Security CSP Deployment with Aqua Operator
            displayName: AquaCsp
            kind: AquaCsp
            name: aquacsps.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Lightning Enforcers Deployment with Aqua Operator
            displayName: AquaLightningEnforcers
            kind: AquaLightning
            name: aqualightnings.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Database Deployment with Aqua Operator
            displayName: AquaDatabase
            kind: AquaDatabase
            name: aquadatabases.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Enforcer Deployment with Aqua Operator
            displayName: AquaEnforcer
            kind: AquaEnforcer
            name: aquaenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security KubeEnforcer Deployment with Aqua Operator
            displayName: AquaKubeEnforcer
            kind: AquaKubeEnforcer
            name: aquakubeenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Gateway Deployment with Aqua Operator
            displayName: AquaGateway
            kind: AquaGateway
            name: aquagateways.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Scanner Deployment with Aqua Operator
            displayName: AquaScanner
            kind: AquaScanner
            name: aquascanners.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Server Deployment with Aqua Operator
            displayName: AquaServer
            kind: AquaServer
            name: aquaservers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Starboard for kube-enforcer.
            displayName: AquaStarboard
            kind: AquaStarboard
            name: aquastarboards.aquasecurity.github.io
            version: v1alpha1
          - description: Aqua Cloud Connector Deployment with Aqua Operator
            displayName: AquaCloudConnector
            kind: AquaCloudConnector
            name: aquacloudconnectors.operator.aquasec.com
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ClusterConfigAuditReport
            kind: ClusterConfigAuditReport
            name: clusterconfigauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ConfigAuditReport
            kind: ConfigAuditReport
            name: configauditreports.aquasecurity.github.io
            version: v1alpha1
        description: |-
          The Aqua Security Operator runs within an OpenShift cluster and provides a means to deploy and manage the Aqua Security cluster and components :
          * Server (aka “console”)
          * Database (for production environments we recommend to use an external database and not the Aqua default database)
          * Gateway
          * Enforcer (aka “agent”)
          * KubeEnforcer
          * Scanner
          * CSP (package that contains the Server, Database, and Gateway)
          * LightningEnforcer (package that contains Enforcer and Kube Enforcer) - Do not deploy using Certified Operator.
          * AquaCloudConnector.

          Use the Aqua-Operator to
          * Deploy Aqua Security components on OpenShift
          * Scale up Aqua Security components with extra replicas
          * Assign metadata tags to Aqua Security components

          The Aqua Operator provides a few [Custom Resources](https://github.com/aquasecurity/aqua-operator/tree/master/config/crd) for managing the Aqua CSP platform.

          ## Prerequisites
          There are only 4 prerequisites:
          1. Verify that the OpenShift version is officially supported [Verify Here](https://docs.aquasec.com/saas/workload-protection/product-information/system-requirements/)
          2. Make sure you have a license and to obtain one, please contact Aqua Security at [Contact Us](mailto:cloudsales@aquasec.com).
          3. Create a new project for aqua

              ```oc new-project aqua```

          4. Create the secret for Aqua Database password. If you are using an External Database, make sure you use those credentials.
          NOTE: This step is optional and you can specify the Database password when creating the CRs.

              ```oc create secret generic aqua-database-password --from-literal=db-password=<password> -n aqua```

          Please note that for the certified operator, the Docker registry secret is NOT needed. For the Red Hat certified operator, the Aqua application images are stored in the Red Hat Connect registry and will be automatically pulled by the Aqua operator.

          ## Choosing the right channel

          Channels are a flexible publishing mechanism that offers you different Aqua versions to deploy, at any point. These channels map to their respective Aqua versions, describing the maturity based on where they are in their lifecycle:
          * 4.6: This maps to the older version of Aqua 4.6.0
          * 5.0: This maps to the Aqua version 5.0.0
          * 5.3: This maps to the Aqua version 5.3.0
          * 6.0: This maps to the Aqua version 6.0.0
          * 6.2: This maps to the Aqua version 6.2.0
          * 6.5: This maps to the Aqua version 6.5.0
          * 2022.4: This is the latest and greatest version of Aqua (default)

          ## Deploying the Aqua Operator

          Aqua Operator follows an easy push button deployment that installs the operator in its own aqua namespace.
          Please refer to the instructions [here](https://github.com/aquasecurity/aqua-operator/blob/2022.4.0/docs/DeployOpenShiftOperator.md).

          ## Configuring the Aqua Operator

          Operators leverage CRDs as a configuration mechanism for the application. Aqua offers a number of CRDs that can be tweaked and configured as per your needs. Please refer to this [link](https://github.com/aquasecurity/aqua-operator/blob/6.2.0/docs/DeployOpenShiftOperator.md#aquacsp-crds) for Custom Resource examples for popular scenarios.

          ## Support

          For support please contact support@aquasec.com.
        displayName: Aqua Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aqua-security
        - scanning
        - security
        - runtime-security
        links:
        - name: Aqua Security
          url: https://www.aquasec.com/
        - name: Aqua Operator Github
          url: https://github.com/aquasecurity/aqua-operator
        maintainers:
        - email: support@aquasec.com
          name: Aqua Support, Aqua Security
        maturity: alpha
        provider:
          name: Aqua Security, Inc.
        relatedImages:
        - registry.connect.redhat.com/aquasec/console@sha256:ed371053e46e89ad68ed48282ddf28abc668a613a9c32e3aa45e5111512b8466
        - registry.connect.redhat.com/aquasec/kube-bench@sha256:bbb4e8a61033b303525af6764f685e62a2f441453c3f0eaaf4cbab6997bf93ee
        - registry.connect.redhat.com/aquasec/gateway@sha256:11aa01e681986f6b5bd53b41a07db4456023080a753970d8c872e000f68f3b9b
        - registry.connect.redhat.com/aquasec/kube-enforcer@sha256:c456b65a3f1dc2de7209f3d9cbcf1bb523130021433b8fd740c41d9bd6f7db9f
        - registry.connect.redhat.com/aquasec/aquasec@sha256:c9f2c770a51e9664006c12d124ca18aab76bea8c32a6a253fe31d44c02633642
        - registry.connect.redhat.com/aquasec/enforcer@sha256:50128b750e0a094de266c9a2cb42ce06a757f441a68d206a8a9c17f7b7119066
        - registry.connect.redhat.com/aquasec/scanner@sha256:c086d735c7e6ea842ad1aaeb1eb19d3a951e6e88d397a20f2cd559c6bc042d26
        - registry.connect.redhat.com/aquasec/cloud-connector@sha256:f9da1936ac5e4b8e4537dca69f920c682fb9fac98d091680e01d2122f2b0f85f
        - registry.connect.redhat.com/aquasec/database@sha256:c91ae82b0179c3ec4e156835a9f23b66767ba15514ff33fb268b35d8f04d6fde
        - registry.connect.redhat.com/aquasec/starboard@sha256:61673f7ba1d983ff4bb4bd76036ac55e5b81696c9a3e3aa4ab1a49a27a7a2106
        version: 2022.4.836
      entries:
      - name: aqua-operator.2022.4.836
        version: 2022.4.836
      - name: aqua-operator.2022.4.829
        version: 2022.4.829
      - name: aqua-operator.2022.4.818
        version: 2022.4.818
      - name: aqua-operator.2022.4.805
        version: 2022.4.805
      - name: aqua-operator.2022.4.760
        version: 2022.4.760
      - name: aqua-operator.2022.4.759
        version: 2022.4.759
      - name: aqua-operator.2022.4.720
        version: 2022.4.720
      - name: aqua-operator.2022.4.683
        version: 2022.4.683
      - name: aqua-operator.2022.4.662
        version: 2022.4.662
      - name: aqua-operator.2022.4.588
        version: 2022.4.588
      - name: aqua-operator.2022.4.587
        version: 2022.4.587
      - name: aqua-operator.2022.4.519
        version: 2022.4.519
      - name: aqua-operator.2022.4.518
        version: 2022.4.518
      name: "2022.4"
    defaultChannel: "2022.4"
    packageName: aqua-operator-certified
    provider:
      name: Aqua Security, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: netobserv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: network-observability-operator.v1.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowMetric",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netobserv-operator",
                    "app.kubernetes.io/instance": "flowmetric-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "flowmetric",
                    "app.kubernetes.io/part-of": "netobserv-operator"
                  },
                  "name": "flowmetric-sample"
                },
                "spec": {
                  "charts": [
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "",
                          "promQL": "sum(rate($METRIC[2m]))"
                        }
                      ],
                      "title": "External ingress traffic",
                      "type": "SingleStat",
                      "unit": "Bps"
                    },
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "{{DstK8S_Namespace}} / {{DstK8S_OwnerName}}",
                          "promQL": "sum(rate($METRIC{DstK8S_Namespace!=\"\"}[2m])) by (DstK8S_Namespace, DstK8S_OwnerName)"
                        }
                      ],
                      "sectionName": "External",
                      "title": "Top external ingress traffic per workload",
                      "type": "StackArea",
                      "unit": "Bps"
                    }
                  ],
                  "direction": "Ingress",
                  "filters": [
                    {
                      "field": "SrcSubnetLabel",
                      "matchType": "Absence"
                    }
                  ],
                  "labels": [
                    "DstK8S_HostName",
                    "DstK8S_Namespace",
                    "DstK8S_OwnerName",
                    "DstK8S_OwnerType"
                  ],
                  "metricName": "cluster_external_ingress_bytes_total",
                  "type": "Counter",
                  "valueField": "Bytes"
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "app"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "infra"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "enable": true,
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "includeList": [
                        "node_ingress_bytes_total",
                        "workload_ingress_bytes_total",
                        "namespace_flows_total"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta2",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "metrics": {
                        "server": {
                          "port": 9400
                        }
                      },
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "eBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "Disabled"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "\"app\""
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "\"infra\""
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "\"Pod\"",
                          "src_kind": "\"Pod\""
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "\"Service\""
                        },
                        "name": "Services network"
                      }
                    ],
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    }
                  },
                  "deploymentModel": "Direct",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "enable": true,
                    "lokiStack": {
                      "name": "loki"
                    },
                    "mode": "Monolithic",
                    "monolithic": {
                      "tenantID": "netobserv",
                      "tls": {
                        "caCert": {
                          "certFile": "service-ca.crt",
                          "name": "loki-gateway-ca-bundle",
                          "type": "configmap"
                        },
                        "enable": false
                      },
                      "url": "http://loki.netobserv.svc:3100/"
                    },
                    "readTimeout": "30s",
                    "writeBatchSize": 10485760,
                    "writeBatchWait": "1s",
                    "writeTimeout": "10s"
                  },
                  "namespace": "netobserv",
                  "networkPolicy": {
                    "additionalNamespaces": [],
                    "enable": false
                  },
                  "processor": {
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "Flows",
                    "metrics": {
                      "disableAlerts": [],
                      "server": {
                        "port": 9401
                      }
                    },
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "prometheus": {
                    "querier": {
                      "enable": true,
                      "mode": "Auto",
                      "timeout": "30s"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking, Observability, OpenShift Optional
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:13b6dd0ddf4bdd2836b0d3ec9c79119a902456dcee100cfac6898768e1f22aab
          createdAt: 2025-09-08T21:48:09
          description: Network flows collector and monitoring solution
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: '{"apiVersion":"flows.netobserv.io/v1beta2",
            "kind":"FlowCollector","metadata":{"name":"cluster"},"spec": {}}'
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
          support: NetObserv team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta2
          - description: '`FlowMetric` is the schema for the custom metrics API, which
              allows to generate more metrics out of flow logs. You can find examples
              here: https://github.com/netobserv/network-observability-operator/tree/main/config/samples/flowmetrics'
            displayName: Flow Metric
            kind: FlowMetric
            name: flowmetrics.flows.netobserv.io
            version: v1alpha1
        description: |
          Network Observability is an OpenShift operator that deploys a monitoring pipeline consisting in:
          - an eBPF agent, that generates network flows from captured packets
          - flowlogs-pipeline, a component that collects, enriches and exports these flows
          - a Console plugin for flows visualization with powerful filtering options, a topology representation and more

          Flow data is then available in multiple ways, each optional:

          - As Cluster Monitoring metrics
          - As raw flow logs stored in Grafana Loki
          - As raw flow logs exported to a collector

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, can optionally be used as the backend to store all collected flows. The Network Observability operator does not install Loki directly, however we provide some guidance to help you there.

          - For a production or production-like environment usage, refer to [the operator documentation](https://docs.openshift.com/container-platform/latest/network_observability/installing-operators.html).

          - For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 10GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          oc create namespace netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/1-storage.yaml) -n netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          If you prefer to not use Loki, you must set `spec.loki.enable` to `false` in `FlowCollector`.
          In that case, you can still get the Prometheus metrics or export raw flows to a custom collector. But be aware that some of the Console plugin features will be disabled. For instance, you will not be able to view raw flows there, and the metrics / topology will have a more limited level of details, missing information such as pods or IPs.

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for example [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/config/samples/flows_v1beta2_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          oc edit flowcollector cluster
          ```

          As it operates cluster-wide on every node, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you might have to configure differently if you used another installation method. Make sure to disable it (`spec.loki.enable`) if you don't want to use Loki.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`) an optional list of exporters to which to send enriched flows. KAFKA and IPFIX exporters are supported. This allows you to define any custom storage or processing that can read from Kafka or use the IPFIX standard.

          - To enable availability zones awareness, set `spec.processor.addZone` to `true`.

          ## Resource considerations

          The following table outlines examples of resource considerations for clusters with certain workload sizes.
          The examples outlined in the table demonstrate scenarios that are tailored to specific workloads. Consider each example only as a baseline from which adjustments can be made to accommodate your workload needs.


          | Resource recommendations                        | Extra small (10 nodes) | Small (25 nodes)       | Medium (65 nodes) **    | Large (120 nodes) **          |
          | ----------------------------------------------- | ---------------------- | ---------------------- | ----------------------- | ----------------------------- |
          | *Worker Node vCPU and memory*                   | 4 vCPUs\| 16GiB mem *  | 16 vCPUs\| 64GiB mem * | 16 vCPUs\| 64GiB mem  * |16 vCPUs\| 64GiB Mem *         |
          | *LokiStack size*                                | `1x.extra-small`       | `1x.small`             | `1x.small`              | `1x.medium`                   |
          | *Network Observability controller memory limit* | 400Mi (default)        | 400Mi (default)        | 400Mi (default)         | 800Mi                         |
          | *eBPF sampling rate*                            | 50 (default)           | 50 (default)           | 50 (default)            | 50 (default)                  |
          | *eBPF memory limit*                             | 800Mi (default)        | 800Mi (default)        | 2000Mi                  | 800Mi (default)               |
          | *FLP memory limit*                              | 800Mi (default)        | 800Mi (default)        | 800Mi (default)         | 800Mi (default)               |
          | *FLP Kafka partitions*                          | N/A                    | 48                     | 48                      | 48                            |
          | *Kafka consumer replicas*                       | N/A                    | 24                     | 24                      | 24                            |
          | *Kafka brokers*                                 | N/A                    | 3 (default)            | 3 (default)             | 3 (default)                   |

          *. Tested with AWS M6i instances.
          **. In addition to this worker and its controller, 3 infra nodes (size `M6i.12xlarge`) and 1 workload node (size `M6i.8xlarge`) were tested.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - More information on [configuring metrics](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/Metrics.md).
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: Network Observability
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/network-observability/network-observability-console-plugin-compat-rhel9@sha256:d130f94cdd3cc863a31c59f395a6c4f663e8e0da23747304353c6674b20d81be
        - registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:13b6dd0ddf4bdd2836b0d3ec9c79119a902456dcee100cfac6898768e1f22aab
        - registry.redhat.io/network-observability/network-observability-ebpf-agent-rhel9@sha256:0b4337f4db1e74a2bea1021a617cb42690251036f583bb771cc5ec888f38c97f
        - registry.redhat.io/network-observability/network-observability-flowlogs-pipeline-rhel9@sha256:09454618709a5c83beb4252f1c14c149716d1c09b6567c0a571d5390e8a639a2
        - registry.redhat.io/network-observability/network-observability-console-plugin-rhel9@sha256:7d92eadb3cd08a08672a65a196a5d3181e764c68e8c20cab2554933cdb10fd0c
        version: 1.9.3
      entries:
      - name: network-observability-operator.v1.9.3
        version: 1.9.3
      - name: network-observability-operator.v1.9.2
        version: 1.9.2
      - name: network-observability-operator.v1.9.1
        version: 1.9.1
      - name: network-observability-operator.v1.9.0
        version: 1.9.0
      - name: network-observability-operator.v1.8.1
        version: 1.8.1
      - name: network-observability-operator.v1.8.0
        version: 1.8.0
      - name: network-observability-operator.v1.7.0
        version: 1.7.0
      - name: network-observability-operator.v1.6.2
        version: 1.6.2
      - name: network-observability-operator.v1.6.1
        version: 1.6.1
      - name: network-observability-operator.v1.6.0
        version: 1.6.0
      - name: network-observability-operator.v1.5.0
        version: 1.5.0
      - name: network-observability-operator.v1.4.2
        version: 1.4.2
      - name: network-observability-operator.v1.4.1
        version: 1.4.1
      - name: network-observability-operator.v1.4.0
        version: 1.4.0
      - name: network-observability-operator.v1.3.0
        version: 1.3.0
      - name: network-observability-operator.v1.2.0
        version: 1.2.0
      - name: network-observability-operator.v1.1.0-202302110050
        version: 1.1.0
      - name: network-observability-operator.v1.0.0-202301070345
        version: 1.0.0
      name: stable
    - currentCSV: network-observability-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:fd943746840cba26996dabb47b1beb3445571fa43cb1b859fb073e4aa114b376
          createdAt: 2023-06-22T09:31:50
          description: Network flows collector and monitoring solution
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
          - description: FlowCollector is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
        description: |
          Network Observability is an OpenShift operator that deploys a monitoring pipeline to collect and enrich network flows that are produced by the Network Observability eBPF agent.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When a `FlowCollector` instance is created, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For normal usage, we recommend two options:

          - Installing the [Loki Operator](https://docs.openshift.com/container-platform/4.11//logging/cluster-logging-loki.html). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          oc create namespace netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.0.2/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.0.2/config/samples/flows_v1beta1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          oc edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.0.2/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: Network Observability
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:fd943746840cba26996dabb47b1beb3445571fa43cb1b859fb073e4aa114b376
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:37b3c9b2eed351d9d1a39c99045bbc7993ccb819f67f7b390deaa34930f45946
        - registry.redhat.io/network-observability/network-observability-console-plugin-rhel9@sha256:880c0812a2cd4b7194f417e31fbdcee951ce76776fa06e33c5c9cd479a854daa
        - registry.redhat.io/network-observability/network-observability-ebpf-agent-rhel9@sha256:e31acc12d8690c339e13b5dd22d33e707701362bdd31086c9b4c0f48cad62cc3
        - registry.redhat.io/network-observability/network-observability-flowlogs-pipeline-rhel9@sha256:7c7c42fff63e0599bb943a9e9617b43181360d5a5254b27f368010af275cf0c3
        version: 1.3.0
      entries:
      - name: network-observability-operator.v1.3.0
        version: 1.3.0
      - name: network-observability-operator.v1.2.0
        version: 1.2.0
      - name: network-observability-operator.v1.1.0-202302110050
        version: 1.1.0
      - name: network-observability-operator.v1.0.0-202301070345
        version: 1.0.0
      name: v1.0.x
    defaultChannel: stable
    packageName: netobserv-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: openshift-builds-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-builds-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "openshift-builds"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-rhel8-operator@sha256:7100a83aa053a976ceb77ace79f090ec8601bc9b2152ff6f7d00f0b1deea7f0b
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n\n## Usage\n\nTo deploy and manage [builds
          for Red Hat OpenShift](https://github.com/shipwright-io/build) in your cluster,\nfirst
          make sure this operator is installed and running on your cluster.\n\nNext,
          create the following:\n\n```yaml\n---\napiVersion: operator.shipwright.io/v1alpha1\nkind:
          ShipwrightBuild\nmetadata:\n  name: openshift-builds-operator\nspec:\n  targetNamespace:
          openshift-builds\n```\n\nThe operator will deploy Shipwright Build in the
          provided `targetNamespace`.\nWhen `.spec.targetNamespace` is not set, the
          namespace will default to ` openshift-builds`.\nRefer to the [ShipwrightBuild
          operator documentation](https://github.com/shipwright-io/operator/blob/main/docs/shipwrightbuild.md)
          for more information about this custom resource.\n"
        displayName: builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://shipwright.io
        - name: Shipwright Builds
          url: https://github.com/shipwright-io/build
        maintainers:
        - email: pipeline-integrations-team@redhat.com
          name: Red Hat Pipelines Integration Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-rhel8-operator@sha256:7100a83aa053a976ceb77ace79f090ec8601bc9b2152ff6f7d00f0b1deea7f0b
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel8@sha256:73efcb6ad82f698d7463718b06c86f2b2b3b5326689a62dc39c9f5d68f8943f2
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel8@sha256:0cf079a24b8b3a4b3e1f30710fea85039d9cf323cfedd06e7d1e33e6336fb848
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel8@sha256:22930726a25b11de2571da0395b53352939d142a16f4db6fd5460abd4cc37ba5
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel8@sha256:fd903804672ec08acf9a2b49aa7ecc5f9b543c12f7462637ef756733171b3817
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel8@sha256:f95348e5fc48676d93ebc919af2c27a041dfd632ec285296246d886daab11576
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel8@sha256:d01c1749ce76dd4bf5fefaa6c553725d33533d6e70e884e916f12217b674c3cf
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:97cade2c1ee468261aec5400728c8d44de387b459134aec7a4c3b5ec5a335d2c
        version: 1.0.2
      entries:
      - name: openshift-builds-operator.v1.0.2
        version: 1.0.2
      - name: openshift-builds-operator.v1.0.1
        version: 1.0.1
      - name: openshift-builds-operator.v1.0.0
        version: 1.0.0
      name: builds-1.0
    - currentCSV: openshift-builds-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          createdAt: "2025-01-23T14:48:43Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.openshift.com/builds/latest/about/overview-openshift-builds.html
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:778ad1946bdad68e1430a2fb366cee25cd98d87e44cf5ad34b87603c2f4171d0
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:e8582857566020c71653228355cbadc2c1cda41e36fc767c204a4edb05e7f42a
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:fa039b62ebd1fc6e86db8ec1c7ae5fe0dd3a6f6f642b63b4dc89af32a78cea6b
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:1ec9a1dcfc08b707df48175bba76f8a6cbfcc8396d9e9940cc3d7100e6731701
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:0e139db4370c37e91ef7eb3e0cf9a2587cb81f37709009a37f2724d4461ed2f3
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:949a1a5150642167b6c7d2b8f6d92a2885afe453319c02495f351b2e79eff3dd
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:7cfffdd43fdd33a20d115f35c2ba492da3c1d0e3ab970587bc888c68c47dfacf
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:97cade2c1ee468261aec5400728c8d44de387b459134aec7a4c3b5ec5a335d2c
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:e9f7b8741f265a332008424a2c5b3dae9ea154e623c469085c2abafb53fc62d4
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:5babdf590730b217deb59acbccf142c0500bc692542131810714cc00724ee02c
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:5ec474ab9e62c92e44c1a6635913032aead6246faec0895dcfbdb28ab0dc8a1a
        version: 1.1.1
      entries:
      - name: openshift-builds-operator.v1.1.1
        version: 1.1.1
      - name: openshift-builds-operator.v1.1.0
        version: 1.1.0
      name: builds-1.1
    - currentCSV: openshift-builds-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          createdAt: "2025-01-23T13:47:38Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.2
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:ee2792972dc52d81814603e26103e0d1117194db5ab3b5577736b0fc3285a8ab
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3fa22124916523b958c67af8ad652e73a2c3d68bb5579da1cba1ade537f3b7ae
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:8ace52981058ed51cec5b893e6d773de963b12ab715258ac1f2cbc6430ad52e7
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:775c15b93b19ba855791d34c4a95ef203dd9f8e51110c67ebf131e38eda49422
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:2b288e36e3606cf8a0d7723772edb8d4ac11772ec450539ef3448ad517155ecb
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:4bee8594d17172e5b9893a865a9922a1c7869027633ad64b3f684ec44755bac6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3fcd8e2bf0bcb8ff8c93a87af2c59a3bcae7be8792f9d3236c9b5bbd9b6db3b2
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:e8582857566020c71653228355cbadc2c1cda41e36fc767c204a4edb05e7f42a
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:78ee35ccc70b78638c667152b7b6247417c0aa1bb6d3f817d59b22eb179899ca
        - registry.redhat.io/ubi8/buildah@sha256:56c71ecd450f284325ba8434cbf5b3006078b11f0625c6258689699fa811c95d
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:ebfe4f8a97f6f0f9afe38953ae8cc2fdf2ec0b76401c819bb12c1df7cd671e9e
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:016c93578908f1dbdc3052333a1d5a9ab0c03104070ac36a6ddf99350d7510f4
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:87c4c0e9fe4d9baf8ae4f191c51d90c46729273e182fb5b3dfd951632e7bfad3
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:a91c2bb1239ef7fbce8c1ab9308da03b533df9cea75c7d978fd76f38087bc912
        version: 1.2.2
      entries:
      - name: openshift-builds-operator.v1.2.2
        version: 1.2.2
      - name: openshift-builds-operator.v1.2.1
        version: 1.2.1
      - name: openshift-builds-operator.v1.2.0
        version: 1.2.0
      name: builds-1.2
    - currentCSV: openshift-builds-operator.v1.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-01-29T08:58:54Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.2
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:6e44758fd4a7b34814ac7c71e52d93633781d39f7897ecab98a090859b487b4f
        - registry.redhat.io/ubi9/buildah@sha256:27d837f9bc69ad3c3651cf3315e2501b137c11baa553d9d46140e5cf7fa7873a
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:0783132b14ed8050664810f37e7ad6388fefbc586c45ecc7ba69dd5041a55c9f
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:e8582857566020c71653228355cbadc2c1cda41e36fc767c204a4edb05e7f42a
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:04b8bfc0dc6f260ef367ca1cce2bcc84032bc8be17b4a595be12a344813a2632
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:1e9a54212690f7930d2b681db87a6f79f5a2acc5c982734d4b283a2472b59db5
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:3523e7c2215e156081e5f05443884d65efad9b7b9ca81a35416695020231916a
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:d4b3f451becb7a42341c98dd5b9fc5d2b582c93cac2728070ef4bb83b64abb4c
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:a4ba591c57bbc815c93023f4a080213be32581ba6a1c2d94b0ecbc5c90b07a1e
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:0b1dd35f55181f34775188bc71b3ff8dff66c70acfcfabad19353a02e19351f7
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:6214708f3b517cb901af9f9bd0502ba54f8e5183d0d909a3b559deee6394026e
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:ed3fdd1e85f5c5434de6ec0feaec8c7fe4e680ca6c64258287f18ec2886bb71f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3fcd8e2bf0bcb8ff8c93a87af2c59a3bcae7be8792f9d3236c9b5bbd9b6db3b2
        version: 1.3.2
      entries:
      - name: openshift-builds-operator.v1.3.2
        version: 1.3.2
      - name: openshift-builds-operator.v1.3.1
        version: 1.3.1
      - name: openshift-builds-operator.v1.3.0
        version: 1.3.0
      name: builds-1.3
    - currentCSV: openshift-builds-operator.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2024-11-15T22:28:01Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:9d12e6cbf26e31a1e4458c99a972ed36125d759c5ec57773f203cd8dd0985de7
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:ed3fdd1e85f5c5434de6ec0feaec8c7fe4e680ca6c64258287f18ec2886bb71f
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:538416bc1384eb987e8a7e80b57eda3feddaf4260b196709a6e9a9125e3af39e
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:1a6ca51c2c68950af605ff367356ab6c0ef7d1b2fd6fb8f00aa665196ff7af84
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:f4b935c16dcbbc0393733437747497c1d24f2cc782c40c49738e5f37c037b3e7
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:c60ca44275297d2ee8c499d9d7fb4056d104f4bd5dec8a02e7388241ac466c44
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:1c7c24111f1709d7648156873ced09462638aa4b3cf7e7aee13297e51bcdfb3e
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:6251cb303774e29d137a37d6527875c371fe1d3f6a5fa0383e80f0e9ef5c02be
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:f87c2069105121927fb0c0a2e37f8f72bc6cb5e035fa8d22f7b41ce0e86c0631
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:4e9c152023b2db04ee5d55f102490294b393bf78baa3aa8917f85f8547fbe3c8
        - registry.redhat.io/ubi9/buildah@sha256:6886997da8c92214e99e3e8da47a0f08239172d2544b4edae058d5cb018effab
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:adc4a31d63b7044db7094fddcb1a4b235d848e842257f5baf22b6ead03fd6c91
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:5e87ba403f9e9fff3002a8275aba0d6c36aed54fb4754d90fa0516ad989f187a
        version: 1.4.1
      entries:
      - name: openshift-builds-operator.v1.4.1
        version: 1.4.1
      - name: openshift-builds-operator.v1.4.0
        version: 1.4.0
      name: builds-1.4
    - currentCSV: openshift-builds-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-08-14T12:12:15Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              OpenShiftBuild describes the desired state of Builds for OpenShift, and the status of
              all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I), Buildah, and Buildpacks. You can create and apply
          build resources, view logs of build runs, \nand manage builds in your OpenShift
          Container Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:38560dea7627632d8b24fa670c1d7ac1218cf86975329908786a460f76de558d
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ce7050192d9850ed4c35cb3295ccdeec1aa593e27e3a3da6d6aca97cde782abe
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c1a84feb88d53e93bdcf2cd5f76e2cbb18541c8b0e2979c132a888d6c280b664
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:ddcfcebd5d788d11f756939f07215b2e3507a1d8fe704a1052e15bb4ebe592e8
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:753e24a1cd57ca22552447363aed48f064b525b49e52f774f3e25f41d8dcb329
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:e2a7a42cd5916c05d95d3fb0decc34d0840420532270422aa7f6b7bdd94de1f2
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:bcae84267f582278591d93aecf94c8d7fb1376521d9b5208201c959a0fe56774
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:f4a75c115b5dc8853845f666f3bd063b2736e01e428a605fb5cd0f02736ecce8
        - registry.redhat.io/ubi9/buildah@sha256:8ea2b81b32f70b9502d88722a5f69de8c0cf5efa6f30df041873546e9f9438cb
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:4dc123fa703006662efcc5569a843a043bb31cf69b6868d6a9b6321c66bfca18
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:1d5a0bfe0013ab7602efe95be8a814be1e5586728fdf0e198c97739cc12f1d87
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:504179e8ee9f88a23f1d4c28c7683ab7395f4c1c4868765bf3c1406669fd52fc
        version: 1.5.1
      entries:
      - name: openshift-builds-operator.v1.5.1
        version: 1.5.1
      - name: openshift-builds-operator.v1.5.0
        version: 1.5.0
      name: builds-1.5
    - currentCSV: openshift-builds-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-09-23T13:38:59Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I), Buildah, and Buildpacks. You can create and apply
          build resources, view logs of build runs, \nand manage builds in your OpenShift
          Container Platform namespaces.\n\n## Prerequisites\n\n* OpenShift Pipelines
          operator must be installed before installing this operator\n\nRead more:
          [https://shipwright.io](https://shipwright.io)\n\n## Features\n\n* Standard
          Kubernetes-native API for building container images from source code and
          Dockerfile\n\n* Support for Source-to-Image (S2I) and Buildah build strategies\n\n*
          Extensibility with your own custom build strategies\n\n* Execution of builds
          from source code in a local directory\n\n* Shipwright CLI for creating and
          viewing logs, and managing builds on the cluster\n\n* Integrated user experience
          with the Developer perspective of the OpenShift Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:c6805e85ad4dc29f3a0582a1fbf99647a90c1af0b3924647839ad881cbb199fa
        - registry.redhat.io/ubi9/buildah@sha256:8ea2b81b32f70b9502d88722a5f69de8c0cf5efa6f30df041873546e9f9438cb
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:f9f43795c7ed1e416b92497cf2fe647beb5e45d755f6a6d89c3915b7a9c76570
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:1a61d4a8f69fa845ee678dd96f934685821828e7575852f97b6e7b1dd79a8297
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c1a84feb88d53e93bdcf2cd5f76e2cbb18541c8b0e2979c132a888d6c280b664
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:3d2e501abe1aa757e4e6ba03fe8e7a45bbfdb4303360ed7130c6344051887ed8
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:9f38cd42591b6a6b65e78a3cc5c165f8498ad3c79460a6c47acfb8744d2dd883
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:0ea2b0153ae07ee15972abbbe034ef7feb0dcaa403ac993afcb171906d9cc796
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:23e3f29ae63546db018701abb29b444b9b825fb3a2bda5084542649d9827d4ba
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:dd96e8c857bc52097d4d194db8fa2cb368bbdcbb6594722660d4e906d0bdc420
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ce7050192d9850ed4c35cb3295ccdeec1aa593e27e3a3da6d6aca97cde782abe
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:fd7c805c6d0638f2814503c6779719279119c7268b1eb0eda9b7b58ac775efa6
        version: 1.6.0
      entries:
      - name: openshift-builds-operator.v1.6.0
        version: 1.6.0
      name: builds-1.6
    - currentCSV: openshift-builds-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-09-23T13:38:59Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I), Buildah, and Buildpacks. You can create and apply
          build resources, view logs of build runs, \nand manage builds in your OpenShift
          Container Platform namespaces.\n\n## Prerequisites\n\n* OpenShift Pipelines
          operator must be installed before installing this operator\n\nRead more:
          [https://shipwright.io](https://shipwright.io)\n\n## Features\n\n* Standard
          Kubernetes-native API for building container images from source code and
          Dockerfile\n\n* Support for Source-to-Image (S2I) and Buildah build strategies\n\n*
          Extensibility with your own custom build strategies\n\n* Execution of builds
          from source code in a local directory\n\n* Shipwright CLI for creating and
          viewing logs, and managing builds on the cluster\n\n* Integrated user experience
          with the Developer perspective of the OpenShift Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:0ea2b0153ae07ee15972abbbe034ef7feb0dcaa403ac993afcb171906d9cc796
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:c6805e85ad4dc29f3a0582a1fbf99647a90c1af0b3924647839ad881cbb199fa
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:3d2e501abe1aa757e4e6ba03fe8e7a45bbfdb4303360ed7130c6344051887ed8
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:fd7c805c6d0638f2814503c6779719279119c7268b1eb0eda9b7b58ac775efa6
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:23e3f29ae63546db018701abb29b444b9b825fb3a2bda5084542649d9827d4ba
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ce7050192d9850ed4c35cb3295ccdeec1aa593e27e3a3da6d6aca97cde782abe
        - registry.redhat.io/ubi9/buildah@sha256:8ea2b81b32f70b9502d88722a5f69de8c0cf5efa6f30df041873546e9f9438cb
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:f9f43795c7ed1e416b92497cf2fe647beb5e45d755f6a6d89c3915b7a9c76570
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:1a61d4a8f69fa845ee678dd96f934685821828e7575852f97b6e7b1dd79a8297
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:9f38cd42591b6a6b65e78a3cc5c165f8498ad3c79460a6c47acfb8744d2dd883
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c1a84feb88d53e93bdcf2cd5f76e2cbb18541c8b0e2979c132a888d6c280b664
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:dd96e8c857bc52097d4d194db8fa2cb368bbdcbb6594722660d4e906d0bdc420
        version: 1.6.0
      entries:
      - name: openshift-builds-operator.v1.6.0
        version: 1.6.0
      - name: openshift-builds-operator.v1.5.1
        version: 1.5.1
      - name: openshift-builds-operator.v1.5.0
        version: 1.5.0
      - name: openshift-builds-operator.v1.4.1
        version: 1.4.1
      - name: openshift-builds-operator.v1.4.0
        version: 1.4.0
      - name: openshift-builds-operator.v1.3.2
        version: 1.3.2
      - name: openshift-builds-operator.v1.3.1
        version: 1.3.1
      - name: openshift-builds-operator.v1.3.0
        version: 1.3.0
      - name: openshift-builds-operator.v1.2.2
        version: 1.2.2
      - name: openshift-builds-operator.v1.2.1
        version: 1.2.1
      - name: openshift-builds-operator.v1.2.0
        version: 1.2.0
      - name: openshift-builds-operator.v1.1.1
        version: 1.1.1
      - name: openshift-builds-operator.v1.1.0
        version: 1.1.0
      - name: openshift-builds-operator.v1.0.2
        version: 1.0.2
      - name: openshift-builds-operator.v1.0.1
        version: 1.0.1
      - name: openshift-builds-operator.v1.0.0
        version: 1.0.0
      name: latest
    defaultChannel: latest
    packageName: openshift-builds-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: VMware Tanzu
      provider-url: ""
    name: rabbitmq-messaging-topology-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-messaging-topology-operator.v1.17.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [{"apiVersion":"rabbitmq.com/v1beta1","kind":"Binding","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Exchange","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Federation","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Permission","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Policy","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"OperatorPolicy","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Queue","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"SchemaReplication","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Shovel","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1alpha1","kind":"SuperStream","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"User","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Vhost","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"TopicPermission","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}}]
          capabilities: Full Lifecycle
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/rabbitmqoperator/messaging-topology-operator:1.17.4
          createdAt: 2025-08-21T11:15:42
          description: Kubernetes operator to allow developers to create and manage
            RabbitMQ messaging topologies within a RabbitMQ cluster using a declarative
            Kubernetes API.
          repository: https://github.com/rabbitmq/messaging-topology-operator
          support: VMware Tanzu
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Binding
            displayName: Binding
            kind: Binding
            name: bindings.rabbitmq.com
            version: v1beta1
          - description: Exchange
            displayName: Exchange
            kind: Exchange
            name: exchanges.rabbitmq.com
            version: v1beta1
          - description: Federation
            displayName: Federation
            kind: Federation
            name: federations.rabbitmq.com
            version: v1beta1
          - description: Permission
            displayName: Permission
            kind: Permission
            name: permissions.rabbitmq.com
            version: v1beta1
          - description: TopicPermission
            displayName: TopicPermission
            kind: TopicPermission
            name: topicpermissions.rabbitmq.com
            version: v1beta1
          - description: Policy
            displayName: Policy
            kind: Policy
            name: policies.rabbitmq.com
            version: v1beta1
          - description: OperatorPolicy
            displayName: OperatorPolicy
            kind: OperatorPolicy
            name: operatorpolicies.rabbitmq.com
            version: v1beta1
          - description: Queue
            displayName: Queue
            kind: Queue
            name: queues.rabbitmq.com
            version: v1beta1
          - description: Schema Replication
            displayName: SchemaReplication
            kind: SchemaReplication
            name: schemareplications.rabbitmq.com
            version: v1beta1
          - description: Shovel
            displayName: Shovel
            kind: Shovel
            name: shovels.rabbitmq.com
            version: v1beta1
          - description: Super Stream
            displayName: SuperStream
            kind: SuperStream
            name: superstreams.rabbitmq.com
            version: v1alpha1
          - description: User
            displayName: User
            kind: User
            name: users.rabbitmq.com
            version: v1beta1
          - description: Vhost
            displayName: Vhost
            kind: Vhost
            name: vhosts.rabbitmq.com
            version: v1beta1
          required:
          - description: Rabbitmq Cluster
            displayName: RabbitmqCluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: |
          ## About the managed application
          RabbitMQ is an open-source message-broker software that originally implemented the Advanced Message Queuing Protocol and has since been extended with a plug-in architecture to support Streaming Text Oriented Messaging Protocol, MQ Telemetry Transport, and other protocols.
          ## About this Operator
          Kubernetes operator to allow developers to create and manage RabbitMQ messaging topologies within a RabbitMQ cluster using a declarative Kubernetes API. A Messaging topology is the collection of objects such as exchanges, queues, bindings and policies that provides specific messaging or streaming scenario. This operator is used with RabbitMQ clusters deployed via the RabbitMQ Cluster Kubernetes Operator. This repository contains custom controllers and custom resource definitions (CRDs) enabling a declarative API for RabbitMQ messaging topologies.
          ## Prerequisites for enabling this Operator
          The operator works on top of the RabbitMQ cluster operator.
          Messaging Topology Operator is tested with the latest release of RabbitMQ Cluster Operator. It uses the generated default user secret from RabbitmqCluster (set in rabbitmqcluster.status.binding) to authenticate with RabbitMQ server.
        displayName: rabbitmq-messaging-topology-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        - messaging
        - amqp
        - amqp10
        - stomp
        - mqtt
        - mqtt5
        - queues
        - streaming
        links:
        - name: GitHub
          url: https://github.com/rabbitmq/messaging-topology-operator
        - name: Documentation
          url: https://www.rabbitmq.com/kubernetes/operator/install-topology-operator.html
        maintainers:
        - email: rabbitmq-users@googlegroups.com
          name: VMware Tanzu
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        provider:
          name: VMware Tanzu
        relatedImages:
        - quay.io/rabbitmqoperator/messaging-topology-operator:1.17.4
        version: 1.17.4
      entries:
      - name: rabbitmq-messaging-topology-operator.v1.17.4
        version: 1.17.4
      - name: rabbitmq-messaging-topology-operator.v1.16.0
        version: 1.16.0
      - name: rabbitmq-messaging-topology-operator.v1.15.0
        version: 1.15.0
      - name: rabbitmq-messaging-topology-operator.v1.14.2
        version: 1.14.2
      - name: rabbitmq-messaging-topology-operator.v1.14.1
        version: 1.14.1
      - name: rabbitmq-messaging-topology-operator.v1.13.0
        version: 1.13.0
      - name: rabbitmq-messaging-topology-operator.v1.12.2
        version: 1.12.2
      - name: rabbitmq-messaging-topology-operator.v1.12.1
        version: 1.12.1
      name: stable
    defaultChannel: stable
    packageName: rabbitmq-messaging-topology-operator
    provider:
      name: VMware Tanzu
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: kaoto.io
      provider-url: https://kaoto.io
    name: kaoto-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kaoto-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "designer.kaoto.io/v1alpha1",
                "kind": "Kaoto",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "kaoto-operator",
                    "app.kubernetes.io/instance": "kaoto-demo",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kaoto",
                    "app.kubernetes.io/part-of": "kaoto-operator"
                  },
                  "name": "kaoto-demo"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          containerImage: quay.io/kaotoio/kaoto-operator@sha256:590d4958f148954ea1edb9825fb7fb4935d041ddcdb4a99302da14a605380071
          createdAt: "2025-09-03T12:48:58Z"
          description: Integration Editor for Apache Camel
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/KaotoIO/kaoto-operator
          support: contact@kaoto.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kaoto is the Schema for the kaotoes API.
            displayName: Kaoto
            kind: Kaoto
            name: kaotoes.designer.kaoto.io
            version: v1alpha1
        description: Kaoto is a low code and no code integration designer to edit
          and deploy based on Apache Camel . It is designed to be extendable and flexible
          and adaptable to different usecases.
        displayName: Kaoto Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - kaoto
        - camel-k
        - designer
        links:
        - name: Kaoto Operator
          url: https://kaoto-operator.domain
        maintainers:
        - email: info@kaoto.io
          name: kaoto
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: kaoto.io
          url: https://kaoto.io
        relatedImages:
        - quay.io/kaotoio/kaoto-operator@sha256:590d4958f148954ea1edb9825fb7fb4935d041ddcdb4a99302da14a605380071
        version: 2.7.0
      entries:
      - name: kaoto-operator.v2.7.0
        version: 2.7.0
      - name: kaoto-operator.v2.6.0
        version: 2.6.0
      - name: kaoto-operator.v2.5.0
        version: 2.5.0
      - name: kaoto-operator.v2.4.0
        version: 2.4.0
      - name: kaoto-operator.v2.3.0
        version: 2.3.0
      - name: kaoto-operator.v2.1.0
        version: 2.1.0
      - name: kaoto-operator.v2.0.1
        version: 2.0.1
      - name: kaoto-operator.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: kaoto-operator
    provider:
      name: kaoto.io
      url: https://kaoto.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: fence-agents-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fence-agents-remediation.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/fence-agents-remediation-operator:v0.4.0
          createdAt: "2024-05-23 19:57:04"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://www.medik8s.io/remediation/fence-agents-remediation/fence-agents-remediation/
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/medik8s/fence-agents-remediation-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      - name: fence-agents-remediation.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: fence-agents-remediation
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: hpe-filex-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-filex-csi-operator.v1.0.0-beta3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.hpe.com/v1",
                "kind": "HPEGreenLakeFileCSIDriver",
                "metadata": {
                  "name": "hpegreenlakefilecsidriver-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "disableNodeConformance": false,
                  "imagePullPolicy": "IfNotPresent",
                  "images": {
                    "csiAttacher": "registry.k8s.io/sig-storage/csi-attacher@sha256:6e54dae32284f60a2de1645c527715e4c934e5ce7899560c0d845bac957118dc",
                    "csiControllerDriver": "quay.io/hpestorage/filex-csi-driver@sha256:8e96df8f21db969dcd507c6fef8ecc82b7f4d9b33a9159f16345ebe7e02d8161",
                    "csiNodeDriver": "quay.io/hpestorage/filex-csi-driver@sha256:8e96df8f21db969dcd507c6fef8ecc82b7f4d9b33a9159f16345ebe7e02d8161",
                    "csiNodeDriverRegistrar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f",
                    "csiNodeInit": "quay.io/hpestorage/filex-csi-init@sha256:65a014b344d262f628100dcd96fecd228c42162bbfa67037952a5a67f16a1437",
                    "csiProvisioner": "registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f",
                    "csiResizer": "registry.k8s.io/sig-storage/csi-resizer@sha256:ab774734705a906561e15b67f6a96538f3ad520335d636f793aaafe87a3b5200",
                    "csiSnapshotter": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:dd788d79cf4c1b8edee6d9b80b8a1ebfc51a38a365c5be656986b129be9ac784"
                  },
                  "kubeletRootDir": "/var/lib/kubelet",
                  "node": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/hpestorage/filex-csi-driver-operator@sha256:e00066fccf2c673547ccafffb0a280130b6825ab6a82a52bd43d752535c63775
          createdAt: "2025-01-10T18:20:35Z"
          description: A Container Storage Interface (CSI) driver for HPE GreenLake
            File Storage.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.0.0-beta3'
          operatorhub.io/ui-metadata-max-k8s-version: "1.32"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/hpe-storage/co-deployments
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE GreenLake File Storage CSI Driver
            displayName: HPEGreenLakeFileCSIDriver
            kind: HPEGreenLakeFileCSIDriver
            name: hpegreenlakefilecsidrivers.storage.hpe.com
            version: v1
        description: |
          The HPE GreenLake File Storage CSI Operator deploys the  HPE GreenLake File Storage CSI Driver on OpenShift for dynamic  provisioning of persistent volumes.
          Please refer to our CSI driver [documentation](https://scod.hpedev.io/filex_csi_driver/index.html) for supported HPE platforms and full list of supported CSI features.
          ## Installation
          Refer to the HPE GreenLake File Storage CSI Operator [documentation](https://scod.hpedev.io/filex_csi_driver/deployment.html#operator).
        displayName: HPE GreenLake File Storage CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE CSI
        - HPE Storage
        - CSI
        links:
        - name: Documentation
          url: https://scod.hpedev.io/filex_csi_driver/index.html
        maintainers:
        - email: dev-hi-containers@hpe.com
          name: HPE Storage Containers Team
        maturity: beta
        minKubeVersion: 1.25.0
        provider:
          name: HPE
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - quay.io/hpestorage/filex-csi-init@sha256:65a014b344d262f628100dcd96fecd228c42162bbfa67037952a5a67f16a1437
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:dd788d79cf4c1b8edee6d9b80b8a1ebfc51a38a365c5be656986b129be9ac784
        - registry.k8s.io/sig-storage/csi-resizer@sha256:ab774734705a906561e15b67f6a96538f3ad520335d636f793aaafe87a3b5200
        - quay.io/hpestorage/filex-csi-driver@sha256:8e96df8f21db969dcd507c6fef8ecc82b7f4d9b33a9159f16345ebe7e02d8161
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3fcd8e2bf0bcb8ff8c93a87af2c59a3bcae7be8792f9d3236c9b5bbd9b6db3b2
        - registry.connect.redhat.com/hpestorage/filex-csi-driver-operator@sha256:e00066fccf2c673547ccafffb0a280130b6825ab6a82a52bd43d752535c63775
        - registry.k8s.io/sig-storage/csi-attacher@sha256:6e54dae32284f60a2de1645c527715e4c934e5ce7899560c0d845bac957118dc
        version: 1.0.0-beta3
      entries:
      - name: hpe-filex-csi-operator.v1.0.0-beta3
        version: 1.0.0-beta3
      name: beta
    defaultChannel: beta
    packageName: hpe-filex-csi-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SmartX
      provider-url: https://www.smartx.com
    name: iomesh-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: iomesh-operator.0.10.1-rc4
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
          createdAt: 2021-08-17 00:00:00+00:00
          repository: https://github.com/iomesh/docs
          support: SmartX, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: iomesh
            displayName: iomesh
            kind: IOMeshCluster
            name: iomeshclusters.iomesh.com
            version: v1alpha1
        description: "IOMesh (https://www.iomesh.com/) is a cloud-native storage product
          specifically designed and developed for Kubernetes. It takes ZBS as its
          core, a “production-ready” distributed storage system developed independently
          by SmartX, to provide production-level high performance and reliable persistent
          storage for business-critical applications such as MySQL, Cassandra, MongoDB,
          etc., contributing to the containerized transformation of stateful applications.
          \     \n## Pre-Requisites\n* According to https://docs.iomesh.com/deploy/prerequisites,
          make sure that your openshift cluster meets the \"Installation Requirements\"
          and set up the worker nodes correctly according to the \"Setup Worker Node\"
          chapter.\n* Run IOMesh Operator pre-installation script in an environment
          where oc or kubectl can be used to access the openshift cluster, the script
          will install the dependencies of IOMesh Operator and specific IOMesh settings
          for the openshift cluster:\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-pre-install-openshift.sh
          | sh -\n```\n## Install IOMesh Operator and IOMesh Cluster\n* Click install
          button on the current web page to install IOMesh Operator\n* On the **Installed
          Operators** -> **IOMesh Operator** -> **Create instance** -> **YAML view**,
          fill in an IOMesh Custom Resources according https://iomesh.run/iomesh.yaml,
          change the `spec.*.dataCIDR` to your own data network CIDR:\n## Post Install\n*
          Run IOMesh Operator post-installation script in an environment where oc
          or kubectl can be used to access the openshift cluster, the script will
          install IOMesh CSI Driver\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-post-install-openshift.sh
          | sh -\n```\n## Getting Started\n* Try to quickstart setup and use IOMesh:
          https://docs.iomesh.com/deploy/setup-iomesh\n"
        displayName: IOMesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iomesh-operator
        links:
        - name: IOMesh Operator
          url: https://iomesh.com
        maintainers:
        - email: na.wang@smartx.com
          name: Na Wang
        maturity: alpha
        provider:
          name: SmartX
          url: https://www.smartx.com
        relatedImages:
        - iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
        version: 0.10.1-rc4
      entries:
      - name: iomesh-operator.0.10.1-rc4
        version: 0.10.1-rc4
      name: stable
    defaultChannel: stable
    packageName: iomesh-operator
    provider:
      name: SmartX
      url: https://www.smartx.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-activemqartemis: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/
    name: amq-broker-rhel9
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-broker-operator.v7.13.1-opr-1-0.1758027574.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 9 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-broker-operator.v7.13.1-opr-1-0.1756381365.p
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 9 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel9-operator@sha256:f34672152017a94ed51056435e218341e22787788ceed52439c1ff066480c715
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9cc48eecf1442ae04b8543fa5d4381a13bc2831390850828834d387006d1342b
        - registry.redhat.io/amq7/amq-broker-rhel9@sha256:755d0f7421efba76f52818bb3a8667cd7d2dcff2c7054dac1a4637b74ecac51d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:47fd4ce2533496828aba37bd1f9715e2164d5c90bd0fc6b25e7e0786d723bf01
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b2402179b1d0ff9c950defe5d8808195ef9a4c3e13e3c469cae31ab51d5f5014
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:bb6fbd68475a7852b4d99eea6c4ab313f9267da7963162f0d75375d7063409e7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:40920687402fd747ca02d384e63692a5e1db625723a2bf3056a607c92ee8eee1
        - registry.redhat.io/amq7/amq-broker-rhel9@sha256:4304a5414911fe55ce4c0133723be234ff288463f4916bd18e743d6359489138
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:7f9b921d89b6d4d91b365659afff127d99f1455b94b48a10a6a7f4c38dd8b700
        - registry.redhat.io/amq7/amq-broker-init-rhel9@sha256:ff80e0a7accda6825bf834d9aec27f48205dd6329991890dd017b4d21d0da5c8
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:16253895bc113f395dc6ab5cd511901e43b6ee676369ec80d598a8741c3c9bf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d42d713da0ce6806fdc6492b6342586783e6865a82a8647d3c4288439b1751ee
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:dc13871dfed52402c40fd192900fc3cb5a49da4fa08e4ec4d0e23a879e83db25
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ffffa9875f0379e9373f89f05eb06e5a193273bb04bc3aa5f85b044357b79098
        - registry.redhat.io/amq7/amq-broker-init-rhel9@sha256:e2cb996e0bf131e71fb8e80c128af18ab2b9330c9fca7a5437842796158e2870
        version: 7.13.1-opr-1+0.1758027574.p
      entries:
      - name: amq-broker-operator.v7.13.1-opr-1
        version: 7.13.1-opr-1
      - name: amq-broker-operator.v7.13.1-opr-1-0.1754564924.p
        version: 7.13.1-opr-1+0.1754564924.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1756374447.p
        version: 7.13.1-opr-1+0.1756374447.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1756381365.p
        version: 7.13.1-opr-1+0.1756381365.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1758024989.p
        version: 7.13.1-opr-1+0.1758024989.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1758027574.p
        version: 7.13.1-opr-1+0.1758027574.p
      - name: amq-broker-operator.v7.13.0-opr-1
        version: 7.13.0-opr-1
      - name: amq-broker-operator.v7.13.0-opr-1-0.1747315188.p
        version: 7.13.0-opr-1+0.1747315188.p
      - name: amq-broker-operator.v7.13.0-opr-1-0.1747658456.p
        version: 7.13.0-opr-1+0.1747658456.p
      name: 7.13.x
    defaultChannel: 7.13.x
    packageName: amq-broker-rhel9
    provider:
      name: Red Hat
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: quay-bridge-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-bridge-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "my-quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecretName": "<NAMESPACE>/<SECRET>",
                  "quayHostname": "https://<QUAY_URL>",
                  "insecureRegistry": false,
                  "whitelistNamespaces": ["default"]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:501fe9b003b5066b3936def0ef4de05abd1a12e795e718d6c9a62d2c6c804c2e
          createdAt: 2021-11-02 11:11 UTC
          description: Enhance OCP using Red Hat Quay container registry
          repository: https://github.com/redhat-cop/quay-openshift-registry-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a connected Quay instance.
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.redhatcop.redhat.io
            version: v1alpha1
        description: "## Overview\n\nThe OpenShift Container Platform contains an
          in cluster image registry for pushing and pulling images that either produced
          within the cluster or are references from external sources. Quay is a highly
          available, enterprise grade image registry with enhanced role based access
          control (RBAC) features. The goal of this operator is to support replicating
          the features found within OpenShift intgrated image registry with Quay.\n\n##
          Functionality\n\nThe following capabilities are enabled through this operator:\n\n*
          Synchronization of OpenShift namespaces as Quay organizations\n    * Creation
          of Robot accounts for each default namespace service account\n    * Creation
          of Secrets for each created Robot Account\n        * Associate Robot Secret
          to Service Account as Mountable and Image Pull Secret\n    * Synchronization
          of OpenShift ImageStreams as Quay Repositories\n* Automatically rewrite
          new Builds making use of ImageStreams to output to Quay\n* Automatically
          import ImageStream tag once build completes\n\n## Prerequisites\n\nThe following
          requirements must be satisfied prior to setting and configuring the integration
          between OpenShift and Quay:\n\n* A Red Hat Quay environment with a user
          with superuser permissions\n* A Red Hat OpenShift Container Platform environment
          (at least version 4.0) with cluster-admin permissions on an OpenShift cluster
          \n*\tOpenShift Command Line tool (CLI)\n\n## Setup and Configuration\n\nThe
          setup process involves steps within both Quay and OpenShift as described
          below:\n\n### Quay Setup\n\nThe integration between OpenShift and Quay makes
          extensive use of the Quay RESTFul API. To allow for other methods aside
          from user credentials to be used for authenticating against the API, an
          application is required to be created. When an application is created, an
          OAuth token is generated which can be used by the operator in OpenShift.
          Applications are created within a Quay organization. While an existing organization
          can be utilized, it is recommended that a dedicated organization be used.\n\n1.\tLogin
          to Quay as a user with superuser access and select the organization for
          which the external application will be configured within.\n2.\tOn the left
          hand side of the page, select **Applications**\n3.\tCreate a new application
          by selecting the **Create New Application** button and entering a name for
          the application\n4.\tAfter the new application has been created, select
          the application that was created\n5.\tOn the lefthand navigation bar, select
          the **Generate Token** button to create a new OAuth2 token\n6.\tSelect all
          of the checkboxes to grant the necessary access that is needed for the integration\n7.\tReview
          the permissions that will be assigned and then select the **Authorize Application**
          button\n8.\tTake note of the generated Access Token as it will be needed
          in the following section\n\n### OpenShift Setup\n\nTo complete the process
          to begin to integrate OpenShift and Quay, several steps are required. Before
          beginning, ensure that you have the OpenShift Command Line tool installed
          and that you are logged into OpenShift as a cluster administrator. \n\n####
          Create A Secret for the Quay OAuth Token\n\nThe Operator will use the previously
          obtained Access Token to communicate with Quay. Store this token within
          OpenShift as a secret.\n\nExecute the following command to create a secret
          called `quay-integration` with a key called `token` containing the access
          token:\n\n```\n$ oc create secret generic quay-integration --from-literal=token=<access_token>\n```\n\nThis
          token will be referenced in the final section.\n\n#### MutatingWebhookConfiguration
          Support\n\nSupport for dynamic interception of API requests that are performed
          as part of OpenShift’s typical build process is facilitated through a [MutatingWebhookConfiguration](https://docs.openshift.com/container-platform/3.11/architecture/additional_concepts/dynamic_admission_controllers.html).
          A MutatingWebhookConfiguration allows for invoking an API running within
          a project on OpenShift when certain API requests are received. In particular,
          to support Quay’s integration with OpenShift, any new Build requests should
          be intercepted so that the output can be modified to target Quay instead
          of OpenShift’s integrated registry. \n\nKubernetes requires that the webhook
          endpoint is secured via SSL using a certificate that makes use of the certificate
          authority for the cluster. Fortunately, OpenShift provides support for generating
          a certificate signed by the cluster. A script is available in the documentation
          called `webhook-create-signed-cert.sh` that automates the process for requesting
          a new certificate.\n\nFirst, create a `Service` for the Operator's webhook
          endpoint:\n```yaml\napiVersion: v1\nkind: Service\nmetadata:\n  labels:\n
          \   name: quay-bridge-operator\n  name: quay-bridge-operator\n  namespace:
          openshift-operators\nspec:\n  ports:\n    - name: https\n      port: 443\n
          \     protocol: TCP\n      targetPort: 8443\n  selector:\n    name: quay-bridge-operator\n
          \ sessionAffinity: None\n  type: ClusterIP\n```\n\nExecute the following
          command to request the certificate:\n\n```\n./webhook-create-signed-cert.sh
          --namespace openshift-operators --secret quay-bridge-operator-webhook-certs
          --service quay-bridge-operator\n```\n\nThe result will place the newly created
          private key and certificate within a secret within the secret specified.
          The secret will be mounted into the appropriate located within the operator
          as declared in the Deployment of the Operator. \n\nOnce the secret has been
          created, focus can shift to the `MutatingWebhookConfiguration` (see below).
          The majority of the content should be in place to apply to the cluster.
          The one section that requires configuration is the caBundle property. This
          refers to the Certificate Authority (CA) for the OpenShift environment.
          The OpenShift CA is available as a _ConfigMap_ within the `kube-system`
          namespace.\n\nExecute the following command to retrieve the CA and format
          the result as a single line so that it can be entered into the MutatingWebhookConfiguration
          resource:\n\n```\noc get configmap -n kube-system extension-apiserver-authentication
          -o=jsonpath='{.data.client-ca-file}' | base64 | tr -d '\\n'\n```\n\nReplace
          the `${CA_BUNDLE}` variable in the following `MutatingWebhookConfiguration`
          YAML and create it:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\nkind:
          MutatingWebhookConfiguration \nmetadata:\n  name: quay-bridge-operator\nwebhooks:\n
          \ - name: quayintegration.redhatcop.redhat.io\n    clientConfig: \n      service:\n
          \       namespace: openshift-operators\n        name: quay-bridge-operator\n
          \       path: \"/admissionwebhook\"\n      caBundle: \"${CA_BUNDLE}\"\n
          \   rules: \n    - operations:  [ \"CREATE\" ]\n      apiGroups: [ \"build.openshift.io\"
          ]\n      apiVersions: [\"v1\" ]\n      resources: [ \"builds\" ]\n    failurePolicy:
          Fail\n```\n\n_Note: Until the operator is running, new requests for builds
          will fail since the webserver the MutatingWebhookConfiguration invokes is
          not available and a proper is response is required in order for the object
          to be persisted in etcd._\n\n#### Create the QuayIntegration Custom Resource\n\nFinally,
          to complete the integration between OpenShift and Quay, a `QuayIntegration`
          custom resource needs to be created.\n\nThe following is an example of a
          basic definition of a `QuayIntegration` resource associated from the associated
          CRD.\n\n```\napiVersion: redhatcop.redhat.io/v1alpha1\nkind: QuayIntegration\nmetadata:\n
          \ name: example-quayintegration\nspec:\n  clusterID: openshift\n  credentialsSecretName:
          <NAMESPACE>/<SECRET>\n  quayHostname: https://<QUAY_URL>\n```\n\nThe _clusterID_
          is a value which should be unique across the entire ecosystem. This value
          is optional and defaults to `openshift`.\n\nThe _credentialsSecretName_
          is a NamespacedName value of the secret containing the token that was previously
          created.\n\nNote: If Quay is using self signed certificates, the property
          `insecureRegistry: true`\n\nOrganizations within Quay should be created
          for the related namespaces in OpenShift."
        displayName: Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Source Code
          url: https://github.com/redhat-cop/quay-openshift-registry-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:501fe9b003b5066b3936def0ef4de05abd1a12e795e718d6c9a62d2c6c804c2e
        version: 3.4.7
      entries:
      - name: quay-bridge-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: quay-bridge-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "my-quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecretName": "<NAMESPACE>/<SECRET>",
                  "quayHostname": "https://<QUAY_URL>",
                  "insecureRegistry": false,
                  "whitelistNamespaces": ["default"]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:9149df5ffc83dce6593072b7f72805e07e700b84527e45b4b7486b9c38f70731
          createdAt: 2021-09-15 23:09 UTC
          description: Enhance OCP using Red Hat Quay container registry
          olm.skipRange: '>=3.4.x <3.5.7'
          repository: https://github.com/redhat-cop/quay-openshift-registry-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a connected Quay instance.
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.redhatcop.redhat.io
            version: v1alpha1
        description: "## Overview\n\nThe OpenShift Container Platform contains an
          in cluster image registry for pushing and pulling images that either produced
          within the cluster or are references from external sources. Quay is a highly
          available, enterprise grade image registry with enhanced role based access
          control (RBAC) features. The goal of this operator is to support replicating
          the features found within OpenShift intgrated image registry with Quay.\n\n##
          Functionality\n\nThe following capabilities are enabled through this operator:\n\n*
          Synchronization of OpenShift namespaces as Quay organizations\n    * Creation
          of Robot accounts for each default namespace service account\n    * Creation
          of Secrets for each created Robot Account\n        * Associate Robot Secret
          to Service Account as Mountable and Image Pull Secret\n    * Synchronization
          of OpenShift ImageStreams as Quay Repositories\n* Automatically rewrite
          new Builds making use of ImageStreams to output to Quay\n* Automatically
          import ImageStream tag once build completes\n\n## Prerequisites\n\nThe following
          requirements must be satisfied prior to setting and configuring the integration
          between OpenShift and Quay:\n\n* A Red Hat Quay environment with a user
          with superuser permissions\n* A Red Hat OpenShift Container Platform environment
          (at least version 4.0) with cluster-admin permissions on an OpenShift cluster
          \n*\tOpenShift Command Line tool (CLI)\n\n## Setup and Configuration\n\nThe
          setup process involves steps within both Quay and OpenShift as described
          below:\n\n### Quay Setup\n\nThe integration between OpenShift and Quay makes
          extensive use of the Quay RESTFul API. To allow for other methods aside
          from user credentials to be used for authenticating against the API, an
          application is required to be created. When an application is created, an
          OAuth token is generated which can be used by the operator in OpenShift.
          Applications are created within a Quay organization. While an existing organization
          can be utilized, it is recommended that a dedicated organization be used.\n\n1.\tLogin
          to Quay as a user with superuser access and select the organization for
          which the external application will be configured within.\n2.\tOn the left
          hand side of the page, select **Applications**\n3.\tCreate a new application
          by selecting the **Create New Application** button and entering a name for
          the application\n4.\tAfter the new application has been created, select
          the application that was created\n5.\tOn the lefthand navigation bar, select
          the **Generate Token** button to create a new OAuth2 token\n6.\tSelect all
          of the checkboxes to grant the necessary access that is needed for the integration\n7.\tReview
          the permissions that will be assigned and then select the **Authorize Application**
          button\n8.\tTake note of the generated Access Token as it will be needed
          in the following section\n\n### OpenShift Setup\n\nTo complete the process
          to begin to integrate OpenShift and Quay, several steps are required. Before
          beginning, ensure that you have the OpenShift Command Line tool installed
          and that you are logged into OpenShift as a cluster administrator. \n\n####
          Create A Secret for the Quay OAuth Token\n\nThe Operator will use the previously
          obtained Access Token to communicate with Quay. Store this token within
          OpenShift as a secret.\n\nExecute the following command to create a secret
          called `quay-integration` with a key called `token` containing the access
          token:\n\n```\n$ oc create secret generic quay-integration --from-literal=token=<access_token>\n```\n\nThis
          token will be referenced in the final section.\n\n#### MutatingWebhookConfiguration
          Support\n\nSupport for dynamic interception of API requests that are performed
          as part of OpenShift’s typical build process is facilitated through a [MutatingWebhookConfiguration](https://docs.openshift.com/container-platform/3.11/architecture/additional_concepts/dynamic_admission_controllers.html).
          A MutatingWebhookConfiguration allows for invoking an API running within
          a project on OpenShift when certain API requests are received. In particular,
          to support Quay’s integration with OpenShift, any new Build requests should
          be intercepted so that the output can be modified to target Quay instead
          of OpenShift’s integrated registry. \n\nKubernetes requires that the webhook
          endpoint is secured via SSL using a certificate that makes use of the certificate
          authority for the cluster. Fortunately, OpenShift provides support for generating
          a certificate signed by the cluster. A script is available in the documentation
          called `webhook-create-signed-cert.sh` that automates the process for requesting
          a new certificate.\n\nFirst, create a `Service` for the Operator's webhook
          endpoint:\n```yaml\napiVersion: v1\nkind: Service\nmetadata:\n  labels:\n
          \   name: quay-bridge-operator\n  name: quay-bridge-operator\n  namespace:
          openshift-operators\nspec:\n  ports:\n    - name: https\n      port: 443\n
          \     protocol: TCP\n      targetPort: 8443\n  selector:\n    name: quay-bridge-operator\n
          \ sessionAffinity: None\n  type: ClusterIP\n```\n\nExecute the following
          command to request the certificate:\n\n```\n./webhook-create-signed-cert.sh
          --namespace openshift-operators --secret quay-bridge-operator-webhook-certs
          --service quay-bridge-operator\n```\n\nThe result will place the newly created
          private key and certificate within a secret within the secret specified.
          The secret will be mounted into the appropriate located within the operator
          as declared in the Deployment of the Operator. \n\nOnce the secret has been
          created, focus can shift to the `MutatingWebhookConfiguration` (see below).
          The majority of the content should be in place to apply to the cluster.
          The one section that requires configuration is the caBundle property. This
          refers to the Certificate Authority (CA) for the OpenShift environment.
          The OpenShift CA is available as a _ConfigMap_ within the `kube-system`
          namespace.\n\nExecute the following command to retrieve the CA and format
          the result as a single line so that it can be entered into the MutatingWebhookConfiguration
          resource:\n\n```\noc get configmap -n kube-system extension-apiserver-authentication
          -o=jsonpath='{.data.client-ca-file}' | base64 | tr -d '\\n'\n```\n\nReplace
          the `${CA_BUNDLE}` variable in the following `MutatingWebhookConfiguration`
          YAML and create it:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\nkind:
          MutatingWebhookConfiguration \nmetadata:\n  name: quay-bridge-operator\nwebhooks:\n
          \ - name: quayintegration.redhatcop.redhat.io\n    clientConfig: \n      service:\n
          \       namespace: openshift-operators\n        name: quay-bridge-operator\n
          \       path: \"/admissionwebhook\"\n      caBundle: \"${CA_BUNDLE}\"\n
          \   rules: \n    - operations:  [ \"CREATE\" ]\n      apiGroups: [ \"build.openshift.io\"
          ]\n      apiVersions: [\"v1\" ]\n      resources: [ \"builds\" ]\n    failurePolicy:
          Fail\n    sideEffects: None\n    admissionReviewVersions: [v1beta1]\n    matchPolicy:
          Exact\n    timeoutSeconds: 30  \n```\n\n_Note: Until the operator is running,
          new requests for builds will fail since the webserver the MutatingWebhookConfiguration
          invokes is not available and a proper is response is required in order for
          the object to be persisted in etcd._\n\n#### Create the QuayIntegration
          Custom Resource\n\nFinally, to complete the integration between OpenShift
          and Quay, a `QuayIntegration` custom resource needs to be created.\n\nThe
          following is an example of a basic definition of a `QuayIntegration` resource
          associated from the associated CRD.\n\n```\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          QuayIntegration\nmetadata:\n  name: example-quayintegration\nspec:\n  clusterID:
          openshift\n  credentialsSecretName: <NAMESPACE>/<SECRET>\n  quayHostname:
          https://<QUAY_URL>\n```\n\nThe _clusterID_ is a value which should be unique
          across the entire ecosystem. This value is optional and defaults to `openshift`.\n\nThe
          _credentialsSecretName_ is a NamespacedName value of the secret containing
          the token that was previously created.\n\nNote: If Quay is using self signed
          certificates, the property `insecureRegistry: true`\n\nOrganizations within
          Quay should be created for the related namespaces in OpenShift."
        displayName: Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Source Code
          url: https://github.com/redhat-cop/quay-openshift-registry-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:9149df5ffc83dce6593072b7f72805e07e700b84527e45b4b7486b9c38f70731
        version: 3.5.7
      entries:
      - name: quay-bridge-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: quay-bridge-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:33dc7673155f01728225221768f6b90d87cc7609dcda46b394dba2421c136a23
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.10.15
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:33dc7673155f01728225221768f6b90d87cc7609dcda46b394dba2421c136a23
        version: 3.10.15
      entries:
      - name: quay-bridge-operator.v3.10.15
        version: 3.10.15
      - name: quay-bridge-operator.v3.10.14
        version: 3.10.14
      - name: quay-bridge-operator.v3.10.13
        version: 3.10.13
      - name: quay-bridge-operator.v3.10.12
        version: 3.10.12
      - name: quay-bridge-operator.v3.10.11
        version: 3.10.11
      - name: quay-bridge-operator.v3.10.10
        version: 3.10.10
      - name: quay-bridge-operator.v3.10.9
        version: 3.10.9
      - name: quay-bridge-operator.v3.10.8
        version: 3.10.8
      - name: quay-bridge-operator.v3.10.7
        version: 3.10.7
      - name: quay-bridge-operator.v3.10.6
        version: 3.10.6
      - name: quay-bridge-operator.v3.10.5
        version: 3.10.5
      - name: quay-bridge-operator.v3.10.4
        version: 3.10.4
      - name: quay-bridge-operator.v3.10.3
        version: 3.10.3
      - name: quay-bridge-operator.v3.10.2
        version: 3.10.2
      - name: quay-bridge-operator.v3.10.1
        version: 3.10.1
      - name: quay-bridge-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-bridge-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:511f8471db70e8d5fd9d0b547c73f1416bbac6f0fe6cdeb6c59a9ef8bd845ac2
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.11.13
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:511f8471db70e8d5fd9d0b547c73f1416bbac6f0fe6cdeb6c59a9ef8bd845ac2
        version: 3.11.13
      entries:
      - name: quay-bridge-operator.v3.11.13
        version: 3.11.13
      - name: quay-bridge-operator.v3.11.12
        version: 3.11.12
      - name: quay-bridge-operator.v3.11.11
        version: 3.11.11
      - name: quay-bridge-operator.v3.11.10
        version: 3.11.10
      - name: quay-bridge-operator.v3.11.9
        version: 3.11.9
      - name: quay-bridge-operator.v3.11.8
        version: 3.11.8
      - name: quay-bridge-operator.v3.11.7
        version: 3.11.7
      - name: quay-bridge-operator.v3.11.6
        version: 3.11.6
      - name: quay-bridge-operator.v3.11.5
        version: 3.11.5
      - name: quay-bridge-operator.v3.11.4
        version: 3.11.4
      - name: quay-bridge-operator.v3.11.3
        version: 3.11.3
      - name: quay-bridge-operator.v3.11.2
        version: 3.11.2
      - name: quay-bridge-operator.v3.11.1
        version: 3.11.1
      - name: quay-bridge-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-bridge-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:c6e04de3ad3b8c0ded95146d090e120bffed7f9dc5e74332a183fe1c2e45d16d
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.12.12
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:c6e04de3ad3b8c0ded95146d090e120bffed7f9dc5e74332a183fe1c2e45d16d
        version: 3.12.12
      entries:
      - name: quay-bridge-operator.v3.12.12
        version: 3.12.12
      - name: quay-bridge-operator.v3.12.11
        version: 3.12.11
      - name: quay-bridge-operator.v3.12.10
        version: 3.12.10
      - name: quay-bridge-operator.v3.12.9
        version: 3.12.9
      - name: quay-bridge-operator.v3.12.8
        version: 3.12.8
      - name: quay-bridge-operator.v3.12.7
        version: 3.12.7
      - name: quay-bridge-operator.v3.12.6
        version: 3.12.6
      - name: quay-bridge-operator.v3.12.5
        version: 3.12.5
      - name: quay-bridge-operator.v3.12.4
        version: 3.12.4
      - name: quay-bridge-operator.v3.12.3
        version: 3.12.3
      - name: quay-bridge-operator.v3.12.2
        version: 3.12.2
      - name: quay-bridge-operator.v3.12.1
        version: 3.12.1
      - name: quay-bridge-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-bridge-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:e8ee1944d8e0e9d515a6e63a3e39901c304abd421fcad2abbf9a3c43c7ca6243
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.13.8
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:e8ee1944d8e0e9d515a6e63a3e39901c304abd421fcad2abbf9a3c43c7ca6243
        version: 3.13.8
      entries:
      - name: quay-bridge-operator.v3.13.8
        version: 3.13.8
      - name: quay-bridge-operator.v3.13.7
        version: 3.13.7
      - name: quay-bridge-operator.v3.13.6
        version: 3.13.6
      - name: quay-bridge-operator.v3.13.5
        version: 3.13.5
      - name: quay-bridge-operator.v3.13.4
        version: 3.13.4
      - name: quay-bridge-operator.v3.13.3
        version: 3.13.3
      - name: quay-bridge-operator.v3.13.2
        version: 3.13.2
      - name: quay-bridge-operator.v3.13.1
        version: 3.13.1
      - name: quay-bridge-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-bridge-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b12f0dfb8fdeae3f521e41fcf60ef118bdfd08ad7f68bd87c5494b067b7c7c58
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.14.5
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b12f0dfb8fdeae3f521e41fcf60ef118bdfd08ad7f68bd87c5494b067b7c7c58
        version: 3.14.5
      entries:
      - name: quay-bridge-operator.v3.14.5
        version: 3.14.5
      - name: quay-bridge-operator.v3.14.4
        version: 3.14.4
      - name: quay-bridge-operator.v3.14.3
        version: 3.14.3
      - name: quay-bridge-operator.v3.14.2
        version: 3.14.2
      - name: quay-bridge-operator.v3.14.1
        version: 3.14.1
      - name: quay-bridge-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-bridge-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b2ac4a48c80209d7bebe8b1b2d066571f99a9f47b78604e47a8300d027941b66
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.15.2
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b2ac4a48c80209d7bebe8b1b2d066571f99a9f47b78604e47a8300d027941b66
        version: 3.15.2
      entries:
      - name: quay-bridge-operator.v3.15.2
        version: 3.15.2
      - name: quay-bridge-operator.v3.15.1
        version: 3.15.1
      - name: quay-bridge-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-bridge-operator.v3.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:0ad7af440c6e00fb578ef40e6d600964324fd94ce7835a28d0ec355b7415093f
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.6.10
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:99dc6b5af811bb33892c17334dd4e1a7eddc6982d49a793b1789f505e00f49b1
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:0ad7af440c6e00fb578ef40e6d600964324fd94ce7835a28d0ec355b7415093f
        version: 3.6.10
      entries:
      - name: quay-bridge-operator.v3.6.10
        version: 3.6.10
      - name: quay-bridge-operator.v3.6.9
        version: 3.6.9
      - name: quay-bridge-operator.v3.6.8
        version: 3.6.8
      - name: quay-bridge-operator.v3.6.7
        version: 3.6.7
      - name: quay-bridge-operator.v3.6.6
        version: 3.6.6
      - name: quay-bridge-operator.v3.6.5
        version: 3.6.5
      - name: quay-bridge-operator.v3.6.4
        version: 3.6.4
      - name: quay-bridge-operator.v3.6.2
        version: 3.6.2
      - name: quay-bridge-operator.v3.6.1
        version: 3.6.1
      - name: quay-bridge-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: quay-bridge-operator.v3.7.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:6a0dd80dc48ad0aaa576b3de78b2b265e29a3c10e4996aa6b8f129592fe20f09
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.7.14
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:975f287513039b2244894b9031e590f665d48f9af480dfdc4b85d4f96dd4bc37
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:6a0dd80dc48ad0aaa576b3de78b2b265e29a3c10e4996aa6b8f129592fe20f09
        version: 3.7.14
      entries:
      - name: quay-bridge-operator.v3.7.14
        version: 3.7.14
      - name: quay-bridge-operator.v3.7.13
        version: 3.7.13
      - name: quay-bridge-operator.v3.7.12
        version: 3.7.12
      - name: quay-bridge-operator.v3.7.11
        version: 3.7.11
      - name: quay-bridge-operator.v3.7.10
        version: 3.7.10
      - name: quay-bridge-operator.v3.7.9
        version: 3.7.9
      - name: quay-bridge-operator.v3.7.8
        version: 3.7.8
      - name: quay-bridge-operator.v3.7.7
        version: 3.7.7
      - name: quay-bridge-operator.v3.7.6
        version: 3.7.6
      - name: quay-bridge-operator.v3.7.5
        version: 3.7.5
      - name: quay-bridge-operator.v3.7.4
        version: 3.7.4
      - name: quay-bridge-operator.v3.7.3
        version: 3.7.3
      - name: quay-bridge-operator.v3.7.2
        version: 3.7.2
      - name: quay-bridge-operator.v3.7.1
        version: 3.7.1
      - name: quay-bridge-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: quay-bridge-operator.v3.8.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:f8544a546dd6d209355c3a1cbe35138f6910491fc4584b401634431da4414ec0
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.8.15
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:975f287513039b2244894b9031e590f665d48f9af480dfdc4b85d4f96dd4bc37
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:f8544a546dd6d209355c3a1cbe35138f6910491fc4584b401634431da4414ec0
        version: 3.8.15
      entries:
      - name: quay-bridge-operator.v3.8.15
        version: 3.8.15
      - name: quay-bridge-operator.v3.8.14
        version: 3.8.14
      - name: quay-bridge-operator.v3.8.13
        version: 3.8.13
      - name: quay-bridge-operator.v3.8.12
        version: 3.8.12
      - name: quay-bridge-operator.v3.8.11
        version: 3.8.11
      - name: quay-bridge-operator.v3.8.10
        version: 3.8.10
      - name: quay-bridge-operator.v3.8.9
        version: 3.8.9
      - name: quay-bridge-operator.v3.8.8
        version: 3.8.8
      - name: quay-bridge-operator.v3.8.7
        version: 3.8.7
      - name: quay-bridge-operator.v3.8.6
        version: 3.8.6
      - name: quay-bridge-operator.v3.8.5
        version: 3.8.5
      - name: quay-bridge-operator.v3.8.4
        version: 3.8.4
      - name: quay-bridge-operator.v3.8.3
        version: 3.8.3
      - name: quay-bridge-operator.v3.8.2
        version: 3.8.2
      - name: quay-bridge-operator.v3.8.1
        version: 3.8.1
      - name: quay-bridge-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-bridge-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:875e0b63fdb7d44879650eb2f788ea705d753520841a5d875428f0b3ca70a36c
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.9.16
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:875e0b63fdb7d44879650eb2f788ea705d753520841a5d875428f0b3ca70a36c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 3.9.16
      entries:
      - name: quay-bridge-operator.v3.9.16
        version: 3.9.16
      - name: quay-bridge-operator.v3.9.15
        version: 3.9.15
      - name: quay-bridge-operator.v3.9.14
        version: 3.9.14
      - name: quay-bridge-operator.v3.9.13
        version: 3.9.13
      - name: quay-bridge-operator.v3.9.12
        version: 3.9.12
      - name: quay-bridge-operator.v3.9.11
        version: 3.9.11
      - name: quay-bridge-operator.v3.9.10
        version: 3.9.10
      - name: quay-bridge-operator.v3.9.9
        version: 3.9.9
      - name: quay-bridge-operator.v3.9.8
        version: 3.9.8
      - name: quay-bridge-operator.v3.9.7
        version: 3.9.7
      - name: quay-bridge-operator.v3.9.6
        version: 3.9.6
      - name: quay-bridge-operator.v3.9.5
        version: 3.9.5
      - name: quay-bridge-operator.v3.9.4
        version: 3.9.4
      - name: quay-bridge-operator.v3.9.3
        version: 3.9.3
      - name: quay-bridge-operator.v3.9.2
        version: 3.9.2
      - name: quay-bridge-operator.v3.9.1
        version: 3.9.1
      - name: quay-bridge-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: quay-bridge-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/securesign/secure-sign-operator
    name: rhtas-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhtas-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": {
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname-intermediate",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname-leaf",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname-root",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname-intermediate",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname-leaf",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname-root",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": {
                  "database": {
                    "create": true
                  },
                  "server": {
                    "replicas": 1
                  },
                  "signer": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:0ebc6b3bd992309a914cc0da8ad8fc628224f583b66f1ca2cd10d2281b781cde
          createdAt: "2025-09-30T11:33:48Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:e7f5a9a2ea76fdfde34f6aac264b51c596cf4ab53f6922b58514c77a1977677e
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:f2798c21dbcb3334b256bf1370a42c013f8f3eefb0641e1c2f250428a82ddb82
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:2dd696771c0188a119d746711551b1643a14a579d0fc66d70918cca9de0c0107
        - registry.redhat.io/rhtas/createtree-rhel9@sha256:244758721aa1c63adfffbae31df58059ff8a06311f6cb60515eeae14b20e78ce
        - registry.redhat.io/rhtas/rekor-monitor-rhel9@sha256:da3aa5c441653f00c4b558ef6ebf21fef518731e70f152a1fd3a750b1145d901
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:b311fb891203fef071d53c9dab67a45961638af98bdfd8e02026dc2dbc25fd00
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:8f8a5d4abf7878ba34151f8632aa7db25218300dbb9f17b121c5f61e0e5c5e03
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:eaae082559c81c5e1253f1b5cf9e298b14235470c67e9111c18ece46078228d5
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:2d91dceb216d8cb8dacefcfe86f357f4ff4eb5e5e998f264d1ad231821a92fad
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:043103f7c888d150053b873276b92b2ee1907c8a7238652ad0770e36d44c10e2
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:0ebc6b3bd992309a914cc0da8ad8fc628224f583b66f1ca2cd10d2281b781cde
        - registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:08302a0af012c1a1f04ef15de86558c9eef55620ab7b179fbda170025f0584be
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:497203991dcbca71275d691ec970ba480cb8e80c9ada6af876f5559abe0d6155
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:110509f5f24240d79afff7cb0d8a4ead0ed038c6100df8e435b01f270b7d83b2
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:abaa247777eb05bf8c97ba1ffc30c00d0d84590fac2fdfbffa06ba374b924888
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:fcb225019f865dd4b7af1f5134a9aed9c5c8a425f3ee43b74e0517ce602e209c
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:e6a5bf7804f1c4371e4de06805d6e1ddfc98e356e355ac3b55a3be25c5636b4d
        version: 1.3.0
      entries:
      - name: rhtas-operator.v1.3.0
        version: 1.3.0
      - name: rhtas-operator.v1.2.1
        version: 1.2.1
      - name: rhtas-operator.v1.2.0
        version: 1.2.0
      - name: rhtas-operator.v1.1.2
        version: 1.1.2
      - name: rhtas-operator.v1.1.1
        version: 1.1.1
      - name: rhtas-operator.v1.1.0
        version: 1.1.0
      - name: rhtas-operator.v1.0.2
        version: 1.0.2
      - name: rhtas-operator.v1.0.1
        version: 1.0.1
      - name: rhtas-operator.v1.0.0
        version: 1.0.0
      name: stable
    - currentCSV: rhtas-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e1a78636cd7cc8e7227b6b0a5ddfd93d002db0464b15e966e88337d38989ea5a
          createdAt: "2024-06-27T09:45:39Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e1a78636cd7cc8e7227b6b0a5ddfd93d002db0464b15e966e88337d38989ea5a
        version: 1.0.2
      entries:
      - name: rhtas-operator.v1.0.2
        version: 1.0.2
      - name: rhtas-operator.v1.0.1
        version: 1.0.1
      - name: rhtas-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: rhtas-operator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e52d07e2cd8cbdc2ccd0336a4fd29cbb2e5b8a1be628350294a54579cc2636ed
          createdAt: "2025-04-08T09:29:59Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e52d07e2cd8cbdc2ccd0336a4fd29cbb2e5b8a1be628350294a54579cc2636ed
        version: 1.1.2
      entries:
      - name: rhtas-operator.v1.1.2
        version: 1.1.2
      - name: rhtas-operator.v1.1.1
        version: 1.1.1
      - name: rhtas-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: rhtas-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
          createdAt: "2025-08-17T21:52:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:02fbd4772c185edf38fcee4e5c5e1c9dad82b77c9da824addaebc445b67d35ac
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
        - registry.redhat.io/ubi9/httpd-24@sha256:badeee0aa3477e61de64f6a2b93e8a3673edd38fc860125f7536506f3d592132
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:c22cdb9a17f85e13c00412cba83aa644e90a0faea0236029d0f1c53fe212e227
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:12b438ea70f83ee6edb12c96b507f92510f4afba824eac34c8f50793c4199db6
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:fc018a45b7eb48690c46c83f52cac55fa40089e6b02686bc21265dc7c2205b8f
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:1d782a127d2f8bdab633f287cb2df2e400ae91bf5d4b913473539d8512dfaab6
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:09b4aeeb607c88c72f69e6f87cb840c82ef5752c64a58fa13551db749fac2530
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:eb38e98dbb9828fe033a5388609132f7246fe77d2f4d258d015a94ea30752b24
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:6270ba29adf779de5d3916d5e82f1cd9491ba89d4ae2042347b6ad6d1819dd26
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:1c2201d50469d70ec6e21546c7c74bd52251d420dc1dcfa5375c1cf61dd3a9fd
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:030986ea26f33db3a192c67a93cd2bde352f23c68a17b062fc062955675c3c51
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:531603563c69aac6905cd6b8157d30781040527404ac2313acba5a9310e0b713
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:fe31830df2d99359749cb86cd1ac2f32fcc3de07aae427d2f2cd3998d31d2833
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:a93df3206409de75613d398e7dcd8b30ff7cc36eb00349c435784f73f00a2335
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:640633823a54c11fe2a8b3d05e571399fda415b6f0b3adf1ba9806882a94bdd4
        version: 1.2.1
      entries:
      - name: rhtas-operator.v1.2.1
        version: 1.2.1
      - name: rhtas-operator.v1.2.0
        version: 1.2.0
      name: stable-v1.2
    - currentCSV: rhtas-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": {
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname-intermediate",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname-leaf",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname-root",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname-intermediate",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname-leaf",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname-root",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": {
                  "database": {
                    "create": true
                  },
                  "server": {
                    "replicas": 1
                  },
                  "signer": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:0ebc6b3bd992309a914cc0da8ad8fc628224f583b66f1ca2cd10d2281b781cde
          createdAt: "2025-09-30T11:33:48Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:abaa247777eb05bf8c97ba1ffc30c00d0d84590fac2fdfbffa06ba374b924888
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:eaae082559c81c5e1253f1b5cf9e298b14235470c67e9111c18ece46078228d5
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:2d91dceb216d8cb8dacefcfe86f357f4ff4eb5e5e998f264d1ad231821a92fad
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:e6a5bf7804f1c4371e4de06805d6e1ddfc98e356e355ac3b55a3be25c5636b4d
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:8f8a5d4abf7878ba34151f8632aa7db25218300dbb9f17b121c5f61e0e5c5e03
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:2dd696771c0188a119d746711551b1643a14a579d0fc66d70918cca9de0c0107
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:110509f5f24240d79afff7cb0d8a4ead0ed038c6100df8e435b01f270b7d83b2
        - registry.redhat.io/rhtas/createtree-rhel9@sha256:244758721aa1c63adfffbae31df58059ff8a06311f6cb60515eeae14b20e78ce
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:b311fb891203fef071d53c9dab67a45961638af98bdfd8e02026dc2dbc25fd00
        - registry.redhat.io/rhtas/rekor-monitor-rhel9@sha256:da3aa5c441653f00c4b558ef6ebf21fef518731e70f152a1fd3a750b1145d901
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:08302a0af012c1a1f04ef15de86558c9eef55620ab7b179fbda170025f0584be
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:043103f7c888d150053b873276b92b2ee1907c8a7238652ad0770e36d44c10e2
        - registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:e7f5a9a2ea76fdfde34f6aac264b51c596cf4ab53f6922b58514c77a1977677e
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:fcb225019f865dd4b7af1f5134a9aed9c5c8a425f3ee43b74e0517ce602e209c
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:0ebc6b3bd992309a914cc0da8ad8fc628224f583b66f1ca2cd10d2281b781cde
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:497203991dcbca71275d691ec970ba480cb8e80c9ada6af876f5559abe0d6155
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:f2798c21dbcb3334b256bf1370a42c013f8f3eefb0641e1c2f250428a82ddb82
        version: 1.3.0
      entries:
      - name: rhtas-operator.v1.3.0
        version: 1.3.0
      name: stable-v1.3
    defaultChannel: stable
    packageName: rhtas-operator
    provider:
      name: Red Hat
      url: https://github.com/securesign/secure-sign-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-lambda-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-lambda-controller.v1.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "Alias",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "CodeSigningConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "EventSourceMapping",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/lambda-controller:1.9.1
          createdAt: "2025-09-19T19:23:07Z"
          description: AWS Lambda controller is a service controller for managing
            Lambda resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alias represents the state of an AWS lambda Alias resource.
            displayName: Alias
            kind: Alias
            name: aliases.lambda.services.k8s.aws
            version: v1alpha1
          - description: CodeSigningConfig represents the state of an AWS lambda CodeSigningConfig
              resource.
            displayName: CodeSigningConfig
            kind: CodeSigningConfig
            name: codesigningconfigs.lambda.services.k8s.aws
            version: v1alpha1
          - description: EventSourceMapping represents the state of an AWS lambda
              EventSourceMapping resource.
            displayName: EventSourceMapping
            kind: EventSourceMapping
            name: eventsourcemappings.lambda.services.k8s.aws
            version: v1alpha1
          - description: Function represents the state of an AWS lambda Function resource.
            displayName: Function
            kind: Function
            name: functions.lambda.services.k8s.aws
            version: v1alpha1
          - description: FunctionURLConfig represents the state of an AWS lambda FunctionURLConfig
              resource.
            displayName: FunctionURLConfig
            kind: FunctionURLConfig
            name: functionurlconfigs.lambda.services.k8s.aws
            version: v1alpha1
          - description: LayerVersion represents the state of an AWS lambda LayerVersion
              resource.
            displayName: LayerVersion
            kind: LayerVersion
            name: layerversions.lambda.services.k8s.aws
            version: v1alpha1
          - description: Version represents the state of an AWS lambda Version resource.
            displayName: Version
            kind: Version
            name: versions.lambda.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Lambda resources in AWS from within your Kubernetes cluster.

          **About Amazon Lambda**

          Lambda is a compute service that lets you run code without provisioning or managing servers. Lambda runs your code on a high-availability compute infrastructure and performs all of the administration of the compute resources, including server and operating system maintenance, capacity provisioning and automatic scaling, code monitoring and logging. With Lambda, you can run code for virtually any type of application or backend service. All you need to do is supply your code in one of the [languages that Lambda supports](https://docs.aws.amazon.com/lambda/latest/dg/lambda-runtimes.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Lambda
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lambda
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Lambda Developer Resources
          url: https://aws.amazon.com/lambda/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: lambda maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/lambda-controller:1.9.1
        version: 1.9.1
      entries:
      - name: ack-lambda-controller.v1.9.1
        version: 1.9.1
      - name: ack-lambda-controller.v1.9.0
        version: 1.9.0
      - name: ack-lambda-controller.v1.8.0
        version: 1.8.0
      - name: ack-lambda-controller.v1.7.7
        version: 1.7.7
      - name: ack-lambda-controller.v1.7.6
        version: 1.7.6
      - name: ack-lambda-controller.v1.7.5
        version: 1.7.5
      - name: ack-lambda-controller.v1.7.4
        version: 1.7.4
      - name: ack-lambda-controller.v1.7.2
        version: 1.7.2
      - name: ack-lambda-controller.v1.7.1
        version: 1.7.1
      - name: ack-lambda-controller.v1.7.0
        version: 1.7.0
      - name: ack-lambda-controller.v1.6.3
        version: 1.6.3
      - name: ack-lambda-controller.v1.6.2
        version: 1.6.2
      - name: ack-lambda-controller.v1.6.1
        version: 1.6.1
      - name: ack-lambda-controller.v1.6.0
        version: 1.6.0
      - name: ack-lambda-controller.v1.5.8
        version: 1.5.8
      - name: ack-lambda-controller.v1.5.7
        version: 1.5.7
      - name: ack-lambda-controller.v1.5.6
        version: 1.5.6
      - name: ack-lambda-controller.v1.5.5
        version: 1.5.5
      - name: ack-lambda-controller.v1.5.4
        version: 1.5.4
      - name: ack-lambda-controller.v1.5.3
        version: 1.5.3
      - name: ack-lambda-controller.v1.5.2
        version: 1.5.2
      - name: ack-lambda-controller.v1.5.1
        version: 1.5.1
      - name: ack-lambda-controller.v1.5.0
        version: 1.5.0
      - name: ack-lambda-controller.v1.4.2
        version: 1.4.2
      - name: ack-lambda-controller.v1.4.0
        version: 1.4.0
      - name: ack-lambda-controller.v1.3.4
        version: 1.3.4
      - name: ack-lambda-controller.v1.3.2
        version: 1.3.2
      - name: ack-lambda-controller.v1.3.1
        version: 1.3.1
      - name: ack-lambda-controller.v1.3.0
        version: 1.3.0
      - name: ack-lambda-controller.v1.2.1
        version: 1.2.1
      - name: ack-lambda-controller.v1.2.0
        version: 1.2.0
      - name: ack-lambda-controller.v1.1.1
        version: 1.1.1
      - name: ack-lambda-controller.v1.1.0
        version: 1.1.0
      - name: ack-lambda-controller.v1.0.1
        version: 1.0.1
      - name: ack-lambda-controller.v1.0.0
        version: 1.0.0
      - name: ack-lambda-controller.v0.1.7
        version: 0.1.7
      - name: ack-lambda-controller.v0.1.6
        version: 0.1.6
      - name: ack-lambda-controller.v0.1.5
        version: 0.1.5
      - name: ack-lambda-controller.v0.1.3
        version: 0.1.3
      - name: ack-lambda-controller.v0.1.2
        version: 0.1.2
      - name: ack-lambda-controller.v0.1.1
        version: 0.1.1
      - name: ack-lambda-controller.v0.1.0
        version: 0.1.0
      - name: ack-lambda-controller.v0.0.17
        version: 0.0.17
      - name: ack-lambda-controller.v0.0.16
        version: 0.0.16
      - name: ack-lambda-controller.v0.0.15
        version: 0.0.15
      - name: ack-lambda-controller.v0.0.14
        version: 0.0.14
      - name: ack-lambda-controller.v0.0.13
        version: 0.0.13
      - name: ack-lambda-controller.v0.0.12
        version: 0.0.12
      - name: ack-lambda-controller.v0.0.11
        version: 0.0.11
      - name: ack-lambda-controller.v0.0.10
        version: 0.0.10
      - name: ack-lambda-controller.v0.0.9
        version: 0.0.9
      - name: ack-lambda-controller.v0.0.8
        version: 0.0.8
      - name: ack-lambda-controller.v0.0.7
        version: 0.0.7
      - name: ack-lambda-controller.v0.0.6
        version: 0.0.6
      - name: ack-lambda-controller.v0.0.5
        version: 0.0.5
      - name: ack-lambda-controller.v0.0.4
        version: 0.0.4
      - name: ack-lambda-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-lambda-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-websocket-service
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-websocket-service.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "websocket.loadcore.keysight.com/v1alpha1",
                "kind": "Websocket",
                "metadata": {
                  "name": "websocket-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "websocket": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:26:39Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Websocket
            name: websockets.websocket.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Websocket Operator
        displayName: Keysight Websocket Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Websocket Service
          url: https://keysight-websocket-service.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-websocket-service@sha256:c91da3557ddcb8f052c9caa514d2dbf6c22b4963c080117a9d1859f1458fc404
        version: 0.1.21
      entries:
      - name: keysight-websocket-service.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-websocket-service
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/in-en/products/cloud-pak-for-data
    name: ibm-cpd-datastage-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-datastage-operator.v7.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ds.cpd.ibm.com/v1",
                "kind": "DataStage",
                "metadata": {
                  "name": "datastage"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "fileStorageClass": "managed-nfs-storage",
                  "scaleConfig": "small"
                }
              }
            ]
          capabilities: Seamless Upgrades
          cloudpakId: 49a42b864bb94569bef0188ead948f11
          cloudpakName: IBM DataStage Enterprise Plus Cartridge for IBM Cloud Pak
            for Data
          containerImage: registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
          description: The IBM DataStage Operator is used to deploy IBM DataStage
            services
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-datastage-operator
          olm.skipRange: '>=1.0.0 <7.3.0'
          operatorframework.io/arch.amd64: supported
          operatorframework.io/arch.s390x: supported
          operatorframework.io/os.linux: supported
          operatorframework.io/suggested-namespace: ibm-common-services
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation: [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)
              License: [IBM DataStage Enterprise Cartridge for Cloud Pak for Data](https://ibm.biz/Bdahic),
              [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)'
            kind: DataStage
            name: datastages.ds.cpd.ibm.com
            version: v1
          - description: Documentation [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)
              License [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)
            kind: PXRuntime
            name: pxruntimes.ds.cpd.ibm.com
            version: v1
        description: |-
          # Introduction
          Industry-leading batch and real-time data integration service for data engineers to build trusted pipelines (ETL/ELT) for AI and operational use cases across on-premises and hybrid cloud environments.
          With DataStage, you can design and run data flows that move and transform data anywhere, at any scale.
          No matter how complex your data landscape, DataStage can streamline your data movement costs and increase productivity. DataStage offers:
          <ul><li>A best-in-breed parallel processing engine that enables you to process your data where it resides</li><li>Automated job design</li><li>Simple integration with cloud data lakes, real-time data sources, relational databases, big data, and NoSQL data stores</li></ul>
          The DataStage service uses Cloud Pak for Data platform connections and integration points, with services like Data Virtualization, to simplify the process of connecting to and accessing your data.

          With DataStage, Data Engineers can use the simple user interface to build no-code/low-code data pipelines. The interface offers hundreds of functions and connectors that reduce development time and inconsistencies across pipelines. The interface also makes it easy to collaborate with your peers and control access to specific analytics projects.
          The service also provides automatic workload balancing to provide high performance pipelines that make efficient use of available compute resources.

          ## Details
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=SSQNUZ_4.0/svc-welcome/dsent.html)

          ## Documentation

          [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)

          ## Installing
          To install the operator follow the the installation and configuration instructions within the [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=datastage-installing).

          ## Prerequisites

          - OpenShift Version >= 4.6
          - IBM Cloud Pak for Data >= 4.0.2
          - IBM Cloud Pak for Data Common Core Services (CCS) >= 2.0.2 (The DataStage service automatically installs the CCS if they are not already installed.)

          ### Resources Required

          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=requirements-hardware)

          ## Configuration

          To scale DataStage services, you can specify spec.scaleConfig in the DataStage custom resource to small, medium, or large.

          ## License
          [IBM DataStage Enterprise Cartridge for Cloud Pak for Data](https://ibm.biz/Bdahic)
          [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)

          ## Limitations
          - Only the `x86-64` platform is supported by this operator.
          - See [Limitations and known issues in DataStage](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=issues-datastage)

          ## SecurityContextConstraints Requirements
          Custom SecurityContextConstraints definition:
          ```
          apiVersion: security.openshift.io/v1
          kind: SecurityContextConstraints
          allowHostDirVolumePlugin: false
          allowHostIPC: false
          allowHostNetwork: false
          allowHostPID: false
          allowHostPorts: false
          allowPrivilegeEscalation: true
          allowPrivilegedContainer: false
          allowedCapabilities: null
          defaultAddCapabilities: null
          fsGroup:
            type: MustRunAs
          metadata:
            annotations:
              include.release.openshift.io/self-managed-high-availability: "true"
              kubernetes.io/description: restrict denies access to all host features and requires pods to be run with a UID, and SELinux context that are allocated to the namespace.  This is the most restrictive SCC and it is used by default for authenticated users.
              release.openshift.io/create-only: "true"
            name: restrict
          priority: null
          readOnlyRootFilesystem: false
          requiredDropCapabilities:
          - KILL
          - MKNOD
          - SETUID
          - SETGID
          runAsUser:
            type: MustRunAsRange
          seLinuxContext:
            type: MustRunAs
          supplementalGroups:
            type: RunAsAny
          users: []
          volumes:
          - configMap
          - downwardAPI
          - emptyDir
          - persistentVolumeClaim
          - projected
          - secret
          ```
        displayName: IBM DataStage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DataStage
        - Analytics
        - Cloud Pak
        - Data
        - Cloud Pak for Data
        - IBM
        maintainers:
        - email: cp4d@ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/in-en/products/cloud-pak-for-data
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:046eb99c8f1f37dd4cffa1c6c9a93aea50bb5d10a549aa12c3c90134d03c6d1c
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:4604a076c837600cd4b3e29b31e22470915fe5b6b1191ae9401407f8620f2472
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:5f5faf7d15f46d79eea873c12ff631ed59bc55bc57f1f7aee9fd0a8f35fdd3cf
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:6c569c4687546a03a5f4e127acc13b219ce3efa057fbdfd55e96868e549c70cb
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:8b72cafe9de355634bc3218cc1dc701ffb376a29aecb5cd12a5719d43a1304bd
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:3f8133fcd904befa579b99d3a828fe737faf13ae14e842ed0408eeeee72de5fb
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:7015b410b947338dcafde98b68b95cb4199c08511ae46e733691e8fa835ad988
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:3eecc54f26cd87870c201b5ade1d037d2733ccfda40fc1aedd9a2164e582275d
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:001bf35df08cd397397c3fdd1ddbd7953cdf4bde06965427ae640e5608e60ddf
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:1b39e1d5d7c13451252250ceeb29b91c1e7ef3739fc17baf9958e1bf07c4c2fd
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:ef03bb6ed647d92b9c84d34810376da7049aad88589eb2e35798d82a22f5b994
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:10492010b28f57f8badbc7cadbe15d1447269052b387d878c6d9ecbd2f62923b
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:99fc418fb1a09ee6c45e90dfba47f73ea8e1da1b19f9f570d2a749b68118b7a3
        version: 7.3.0
      entries:
      - name: ibm-cpd-datastage-operator.v7.3.0
        version: 7.3.0
      name: v7.3
    defaultChannel: v7.3
    packageName: ibm-cpd-datastage-operator
    provider:
      name: IBM
      url: https://www.ibm.com/in-en/products/cloud-pak-for-data
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Advanced Micro Devices, Inc.
      provider-url: ""
    name: amd-gpu-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amd-gpu-operator.v1.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "amd.com/v1alpha1",
                "kind": "DeviceConfig",
                "metadata": {
                  "name": "test-deviceconfig",
                  "namespace": "openshift-amd-gpu"
                },
                "spec": {
                  "devicePlugin": {
                    "devicePluginImage": "rocm/k8s-device-plugin:rhubi-latest",
                    "nodeLabellerImage": "rocm/k8s-node-labeller:rhubi-latest"
                  },
                  "driver": {
                    "enable": "true",
                    "version": "6.4.2"
                  },
                  "selector": {
                    "feature.node.kubernetes.io/amd-gpu": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning,Monitoring
          containerImage: docker.io/rocm/gpu-operator@sha256:1583d4ec172d50d4416f4030eb85910ec54633b7e8e061d4bd79b14dbf98d28c
          createdAt: "2025-08-01T01:55:48Z"
          description: |-
            Operator responsible for deploying AMD GPU kernel drivers, device plugin, device test runner and device metrics exporter
            For more information, visit [documentation](https://instinct.docs.amd.com/projects/gpu-operator/en/latest/)
          deviceConfigManagerImage: docker.io/rocm/device-config-manager@sha256:d2896271eacfc664ac7651425f46a85af4f4ef98d713988c2a4fee4a846548ee
          devicePluginImage: docker.io/rocm/k8s-device-plugin@sha256:840c4ea1bf58df60ef3e80e3bf27bbf620a393854e42f58b2224b288767e3dd4
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          metricsExporterImage: docker.io/rocm/device-metrics-exporter@sha256:6c946c6ca10d5c0f40e67b5759759f12d51c45e3b934cb1af912821d5fd76e6e
          nodelabellerImage: docker.io/rocm/k8s-node-labeller@sha256:75f916c5a407e40e9ffdd26cdfc65b7da855710568937e740a969094e2991065
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-amd-gpu
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ROCm/gpu-operator
          support: Advanced Micro Devices, Inc.
          testRunnerImage: docker.io/rocm/test-runner@sha256:98e698263b4df26959341b14b3ddded9df51e504e7c6819d5c7c56c434855e9d
          utilsContainerImage: docker.io/rocm/gpu-operator-utils@sha256:9cbe60dce5b3a2ffd93f7148a363ed5f072a320377d5e5bcbbfb1e05d4959287
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DeviceConfig describes how to enable AMD GPU device
            displayName: DeviceConfig
            kind: DeviceConfig
            name: deviceconfigs.amd.com
            version: v1alpha1
        description: |-
          Operator responsible for deploying AMD GPU kernel drivers, device plugin, device test runner and device metrics exporter
          For more information, visit [documentation](https://instinct.docs.amd.com/projects/gpu-operator/en/latest/)
        displayName: amd-gpu-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AMD
        - GPU
        - AI
        - Deep Learning
        - Hardware
        - Driver
        - Monitoring
        links:
        - name: AMD GPU Operator
          url: https://github.com/ROCm/gpu-operator
        maintainers:
        - email: Yan.Sun3@amd.com
          name: Yan Sun
        - email: farshad.ghodsian@amd.com
          name: Farshad Ghodsian
        - email: shrey.ajmera@amd.com
          name: Shrey Ajmera
        maturity: stable
        provider:
          name: Advanced Micro Devices, Inc.
        relatedImages:
        - docker.io/rocm/gpu-operator@sha256:1583d4ec172d50d4416f4030eb85910ec54633b7e8e061d4bd79b14dbf98d28c
        - docker.io/rocm/device-metrics-exporter@sha256:6c946c6ca10d5c0f40e67b5759759f12d51c45e3b934cb1af912821d5fd76e6e
        - docker.io/rocm/k8s-device-plugin@sha256:840c4ea1bf58df60ef3e80e3bf27bbf620a393854e42f58b2224b288767e3dd4
        - docker.io/rocm/k8s-node-labeller@sha256:75f916c5a407e40e9ffdd26cdfc65b7da855710568937e740a969094e2991065
        - docker.io/rocm/gpu-operator-utils@sha256:9cbe60dce5b3a2ffd93f7148a363ed5f072a320377d5e5bcbbfb1e05d4959287
        - docker.io/rocm/test-runner@sha256:98e698263b4df26959341b14b3ddded9df51e504e7c6819d5c7c56c434855e9d
        - docker.io/rocm/device-config-manager@sha256:d2896271eacfc664ac7651425f46a85af4f4ef98d713988c2a4fee4a846548ee
        version: 1.3.2
      entries:
      - name: amd-gpu-operator.v1.3.2
        version: 1.3.2
      - name: amd-gpu-operator.v1.3.1
        version: 1.3.1
      - name: amd-gpu-operator.v1.2.1
        version: 1.2.1
      - name: amd-gpu-operator.v1.1.1
        version: 1.1.1
      name: alpha
    defaultChannel: alpha
    packageName: amd-gpu-operator
    provider:
      name: Advanced Micro Devices, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-zookeeper-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zookeeper-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:1d36865fe1cac88c9edc63c0e35e42a80456486668de5a575f10cc4b445d47db
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:1d36865fe1cac88c9edc63c0e35e42a80456486668de5a575f10cc4b445d47db
        version: 24.11.1
      entries:
      - name: zookeeper-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: zookeeper-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:6a4729289275e473907a3792c001dbe7dd2dc0498aabc719cf4175d956e3c87f
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:6a4729289275e473907a3792c001dbe7dd2dc0498aabc719cf4175d956e3c87f
        version: 25.3.0
      entries:
      - name: zookeeper-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: zookeeper-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
        version: 25.7.0
      entries:
      - name: zookeeper-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: zookeeper-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
        version: 25.7.0
      entries:
      - name: zookeeper-operator.v25.7.0
        version: 25.7.0
      - name: zookeeper-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-zookeeper-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cluster-kube-descheduler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clusterkubedescheduleroperator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "KubeDescheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-kube-descheduler-operator"
                },
                "spec": {
                  "deschedulingIntervalSeconds": 3600,
                  "managementState": "Managed",
                  "mode": "Predictive",
                  "profiles": ["AffinityAndTaints"]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-cluster-kube-descheduler-operator@sha256:ba0b71ff2a30a069b4a8a8f3c1e0898aaadc6db112e4cc12aff7c77ced7a0405
          createdAt: 2021/10/13
          description: An operator to run descheduler in Openshift cluster.
          olm.skipRange: '>=4.10.0 <4.13.0-202309181427'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-kube-descheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeDescheduler is the Schema for the deschedulers API
            displayName: Kube Descheduler
            kind: KubeDescheduler
            name: kubedeschedulers.operator.openshift.io
            version: v1
        description: |
          The Kube Descheduler Operator provides the ability to evict a running pod so that the pod can be rescheduled onto a more suitable node.

          There are several situations where descheduling can benefit your cluster:

          * Nodes are underutilized or overutilized.
          * Pod and node affinity requirements, such as taints or labels, have changed and the original scheduling decisions are no longer appropriate for certain nodes.
          * Node failure requires pods to be moved.
          * New nodes are added to clusters.

          ## Descheduler Profiles

          Once the operator is installed, you can configure one or more profiles to identify pods to evict. The scheduler will schedule the replacement of the evicted pods.

          The following profiles are available:

          * AffinityAndTaints
          * TopologyAndDuplicates
          * SoftTopologyAndDuplicates
          * LifecycleAndUtilization
          * EvictPodsWithPVC
          * EvictPodsWithLocalStorage

          These profiles are documented in detail in the [descheduler operator README](https://github.com/openshift/cluster-kube-descheduler-operator#profiles).

          ## Additional Parameters

          In addition to the profiles, the following parameters can be configured:

          * `deschedulingIntervalSeconds` - Set the number of seconds between descheduler runs. A value of `0` in this field runs the descheduler once and exits.
          * `profileCustomizations` - Allows certain profile parameters to be tweaked, such as `podLifetime` (see [README](https://github.com/openshift/cluster-kube-descheduler-operator#profile-customizations) for more info).
        displayName: Kube Descheduler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - scheduling
        - descheduler
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/cluster-kube-descheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-cluster-kube-descheduler-operator@sha256:ba0b71ff2a30a069b4a8a8f3c1e0898aaadc6db112e4cc12aff7c77ced7a0405
        - registry.redhat.io/openshift4/ose-descheduler@sha256:45dc69ad93ab50bdf9ce1bb79f6d98f849e320db68af30475b10b7f5497a1b13
        version: 4.13.0-202309181427
      entries:
      - name: clusterkubedescheduleroperator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: "4.13"
    - currentCSV: clusterkubedescheduleroperator.v5.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "KubeDescheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-kube-descheduler-operator"
                },
                "spec": {
                  "deschedulingIntervalSeconds": 3600,
                  "managementState": "Managed",
                  "mode": "Predictive",
                  "profiles": ["AffinityAndTaints"]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/kube-descheduler-operator/kube-descheduler-rhel9-operator@sha256:03404d74da227fdf5805f61657272af2c915ce0014d6a902ef384b7a70bd10e2
          createdAt: 2024/10/30
          description: An operator to run descheduler in Openshift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.12.0 <5.0.2'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-kube-descheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeDescheduler is the Schema for the deschedulers API
            displayName: Kube Descheduler
            kind: KubeDescheduler
            name: kubedeschedulers.operator.openshift.io
            version: v1
        description: |
          The Kube Descheduler Operator provides the ability to evict a running pod so that the pod can be rescheduled onto a more suitable node.

          There are several situations where descheduling can benefit your cluster:

          * Nodes are underutilized or overutilized.
          * Pod and node affinity requirements, such as taints or labels, have changed and the original scheduling decisions are no longer appropriate for certain nodes.
          * Node failure requires pods to be moved.
          * New nodes are added to clusters.

          ## Descheduler Profiles

          Once the operator is installed, you can configure one or more profiles to identify pods to evict. The scheduler will schedule the replacement of the evicted pods.

          The following profiles are available:

          * AffinityAndTaints
          * TopologyAndDuplicates
          * SoftTopologyAndDuplicates
          * LifecycleAndUtilization
          * EvictPodsWithPVC
          * EvictPodsWithLocalStorage

          These profiles are documented in detail in the [descheduler operator README](https://github.com/openshift/cluster-kube-descheduler-operator#profiles).

          ## Additional Parameters

          In addition to the profiles, the following parameters can be configured:

          * `deschedulingIntervalSeconds` - Set the number of seconds between descheduler runs. A value of `0` in this field runs the descheduler once and exits.
          * `profileCustomizations` - Allows certain profile parameters to be tweaked, such as `podLifetime` (see [README](https://github.com/openshift/cluster-kube-descheduler-operator#profile-customizations) for more info).
        displayName: Kube Descheduler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - scheduling
        - descheduler
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/cluster-kube-descheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/kube-descheduler-operator/kube-descheduler-rhel9-operator@sha256:03404d74da227fdf5805f61657272af2c915ce0014d6a902ef384b7a70bd10e2
        - registry.redhat.io/kube-descheduler-operator/descheduler-rhel9@sha256:7ba3f46cbbec2e47edcc255fae405bdbf2b9df3a232fc6809ce65fb1827a3b45
        version: 5.0.2
      entries:
      - name: clusterkubedescheduleroperator.v5.0.2
        version: 5.0.2
      - name: clusterkubedescheduleroperator.4.14.0-202311021650
        version: 4.14.0-202311021650
      - name: clusterkubedescheduleroperator.4.14.0-202310201027
        version: 4.14.0-202310201027
      name: stable
    defaultChannel: stable
    packageName: cluster-kube-descheduler-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: machine-deletion-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: machine-deletion-remediation.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:420a4d4a8610a3694fd3ef9809c55a4f140ec8cea7f09219647f82a84d1f411b
          createdAt: "2023-10-30 13:59:28"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          olm.skipRange: '>=0.1.0 <0.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:420a4d4a8610a3694fd3ef9809c55a4f140ec8cea7f09219647f82a84d1f411b
        version: 0.2.1
      entries:
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.14-eus
    - currentCSV: machine-deletion-remediation.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:1445e6df8f49672aa729806f0e257043ecc22e19e666bb5a799286c6b878b1dd
          createdAt: "2024-08-11 06:47:20"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.2.0 <0.3.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f4381aaf002b761a2ba614332a2af973f4de5f6a4c84dd96b8808725e561fa89
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:1445e6df8f49672aa729806f0e257043ecc22e19e666bb5a799286c6b878b1dd
        version: 0.3.1
      entries:
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.16-eus
    - currentCSV: machine-deletion-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-29 09:35:10"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.3.1 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.6/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel9-operator@sha256:dbe9095d494ebc7b56982cb1f74329d7a7fd0cbb7544e9f90207eb35de7997dc
        version: 0.4.1
      entries:
      - name: machine-deletion-remediation.v0.4.1
        version: 0.4.1
      - name: machine-deletion-remediation.v0.4.0
        version: 0.4.0
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.18-eus
    - currentCSV: machine-deletion-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:7e1cad02f1e74a1355c168fc53925bd0fbcc55f2535a46f734c8de082eb49400
          createdAt: "2023-04-24 08:22:17"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3af07bfe0fb73b1a7f00487eba302d6da442e742ab4007fc50e52efd4d45dae5
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:7e1cad02f1e74a1355c168fc53925bd0fbcc55f2535a46f734c8de082eb49400
        version: 0.1.0
      entries:
      - name: machine-deletion-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: machine-deletion-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-29 09:35:10"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.3.1 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.6/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel9-operator@sha256:dbe9095d494ebc7b56982cb1f74329d7a7fd0cbb7544e9f90207eb35de7997dc
        version: 0.4.1
      entries:
      - name: machine-deletion-remediation.v0.4.1
        version: 0.4.1
      - name: machine-deletion-remediation.v0.4.0
        version: 0.4.0
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: stable
    defaultChannel: stable
    packageName: machine-deletion-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: sailoperator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sailoperator.v1.28.0-nightly-2025-10-06
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:07ea72de6ff9184f346ee2948228e73cfe4180c8d05fe7ee972c22378ad450ef
          createdAt: "2025-10-06T03:29:39Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0
          - master
          - v1.28-alpha.8de61b9a

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-testing/ztunnel@sha256:38b5902bf36f5163274ce0963dfb1e106000ac639d429f05846448de8454492d
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - quay.io/sail-dev/sail-operator@sha256:07ea72de6ff9184f346ee2948228e73cfe4180c8d05fe7ee972c22378ad450ef
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - gcr.io/istio-testing/install-cni@sha256:9525a3711201bcca04ce2ca9b1576d4094a03779d60db0a36b15a956020bbec1
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-testing/proxyv2@sha256:11cf6a3083a3c1ffa539a21cee30ddc0abe47fb44bb16520cdc2f4141dcfee03
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-testing/pilot@sha256:13e2125870f59f752dc5834b702c2dab28f36edc00dccdeeaf872023018e3df0
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        version: 1.28.0-nightly-2025-10-06
      entries:
      - name: sailoperator.v1.28.0-nightly-2025-10-06
        version: 1.28.0-nightly-2025-10-06
      - name: sailoperator.v1.28.0-nightly-2025-10-05
        version: 1.28.0-nightly-2025-10-05
      - name: sailoperator.v1.28.0-nightly-2025-10-04
        version: 1.28.0-nightly-2025-10-04
      - name: sailoperator.v1.28.0-nightly-2025-10-03
        version: 1.28.0-nightly-2025-10-03
      - name: sailoperator.v1.28.0-nightly-2025-10-02
        version: 1.28.0-nightly-2025-10-02
      - name: sailoperator.v1.28.0-nightly-2025-10-01
        version: 1.28.0-nightly-2025-10-01
      - name: sailoperator.v1.28.0-nightly-2025-09-30
        version: 1.28.0-nightly-2025-09-30
      - name: sailoperator.v1.28.0-nightly-2025-09-29
        version: 1.28.0-nightly-2025-09-29
      - name: sailoperator.v1.28.0-nightly-2025-09-28
        version: 1.28.0-nightly-2025-09-28
      - name: sailoperator.v1.28.0-nightly-2025-09-27
        version: 1.28.0-nightly-2025-09-27
      - name: sailoperator.v1.28.0-nightly-2025-09-26
        version: 1.28.0-nightly-2025-09-26
      - name: sailoperator.v1.28.0-nightly-2025-09-25
        version: 1.28.0-nightly-2025-09-25
      - name: sailoperator.v1.28.0-nightly-2025-09-23
        version: 1.28.0-nightly-2025-09-23
      - name: sailoperator.v1.28.0-nightly-2025-09-22
        version: 1.28.0-nightly-2025-09-22
      name: 1.28-nightly
    - currentCSV: sailoperator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.23.0"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.23.0"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/maistra-dev/sail-operator:0.1-latest
          createdAt: "2024-08-20T11:54:30Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1alpha1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1alpha1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1alpha1
          - description: |-
              RemoteIstio represents a remote Istio Service Mesh deployment consisting of one or more
              remote control plane instances (represented by one or more IstioRevision objects).
            displayName: Remote Istio
            kind: RemoteIstio
            name: remoteistios.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.23.0
          - v1.22.4

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - docker.io/istio/proxyv2:1.22.4
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.16.0
        - quay.io/maistra-dev/sail-operator:0.1-latest
        - docker.io/istio/pilot:1.23.0
        - docker.io/istio/proxyv2:1.23.0
        - docker.io/istio/install-cni:1.22.4
        - docker.io/istio/pilot:1.22.4
        - docker.io/istio/ztunnel:1.23.0
        - docker.io/istio/ztunnel:1.22.4
        - docker.io/istio/install-cni:1.23.0
        version: 0.1.0
      entries:
      - name: sailoperator.v0.1.0
        version: 0.1.0
      - name: sailoperator.v0.1.0-rc.1
        version: 0.1.0-rc.1
      name: candidates
    - currentCSV: sailoperator.v1.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
          createdAt: "2025-09-05T10:12:19Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        version: 1.27.1
      entries:
      - name: sailoperator.v1.27.1
        version: 1.27.1
      - name: sailoperator.v1.27.0
        version: 1.27.0
      - name: sailoperator.v1.26.3
        version: 1.26.3
      - name: sailoperator.v1.26.2
        version: 1.26.2
      - name: sailoperator.v1.26.1
        version: 1.26.1
      - name: sailoperator.v1.26.0
        version: 1.26.0
      - name: sailoperator.v1.25.2
        version: 1.25.2
      - name: sailoperator.v1.25.1
        version: 1.25.1
      - name: sailoperator.v1.25.0
        version: 1.25.0
      name: stable
    - currentCSV: sailoperator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.3"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.3"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator:1.0-latest
          createdAt: "2025-03-07T10:17:30Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.24-latest
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.23-latest
          - v1.23.5
          - v1.23.4
          - v1.23.3
          - v1.23.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/sail-dev/sail-operator:1.0-latest
        - gcr.io/istio-release/ztunnel:1.24.1
        - gcr.io/istio-release/proxyv2:1.24.2
        - gcr.io/istio-release/install-cni:1.24.3
        - gcr.io/istio-release/install-cni:1.23.0
        - gcr.io/istio-release/proxyv2:1.23.5
        - gcr.io/istio-release/proxyv2:1.23.4
        - gcr.io/istio-release/pilot:1.24.2
        - gcr.io/istio-release/pilot:1.23.0
        - gcr.io/istio-release/install-cni:1.24.2
        - gcr.io/istio-release/pilot:1.23.4
        - gcr.io/istio-release/proxyv2:1.23.3
        - gcr.io/istio-release/proxyv2:1.23.0
        - gcr.io/istio-release/pilot:1.23.3
        - gcr.io/istio-release/pilot:1.24.1
        - gcr.io/istio-release/pilot:1.23.5
        - gcr.io/istio-release/proxyv2:1.24.3
        - gcr.io/istio-release/ztunnel:1.24.3
        - gcr.io/istio-release/ztunnel:1.23.4
        - gcr.io/istio-release/install-cni:1.23.5
        - gcr.io/istio-release/ztunnel:1.23.3
        - gcr.io/istio-release/install-cni:1.24.1
        - gcr.io/istio-release/proxyv2:1.24.1
        - gcr.io/istio-release/ztunnel:1.23.0
        - gcr.io/istio-release/install-cni:1.23.4
        - gcr.io/istio-release/ztunnel:1.23.5
        - gcr.io/istio-release/ztunnel:1.24.2
        - gcr.io/istio-release/install-cni:1.23.3
        - gcr.io/istio-release/pilot:1.24.3
        version: 1.0.0
      entries:
      - name: sailoperator.v1.0.0
        version: 1.0.0
      name: stable-1.0
    - currentCSV: sailoperator.v1.25.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.25.2"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.25.2"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator:1.25.2
          createdAt: "2025-05-07T19:44:51Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.25.2'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.25-latest
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0
          - v1.23-latest
          - v1.23.5
          - v1.23.4
          - v1.23.3
          - v1.23.2
          - v1.22-latest
          - v1.22.8
          - v1.22.7
          - v1.22.6
          - v1.22.5
          - v1.21.6

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/ztunnel:1.22.6
        - gcr.io/istio-release/pilot:1.22.8
        - gcr.io/istio-release/install-cni:1.23.5
        - quay.io/sail-dev/sail-operator:1.25.2
        - gcr.io/istio-release/pilot:1.22.6
        - gcr.io/istio-release/install-cni:1.24.2
        - gcr.io/istio-release/install-cni:1.24.3
        - gcr.io/istio-release/install-cni:1.23.3
        - gcr.io/istio-release/ztunnel:1.22.5
        - gcr.io/istio-release/proxyv2:1.23.3
        - gcr.io/istio-release/pilot:1.23.2
        - gcr.io/istio-release/install-cni:1.25.1
        - gcr.io/istio-release/pilot:1.24.3
        - gcr.io/istio-release/ztunnel:1.25.1
        - gcr.io/istio-release/install-cni:1.22.8
        - gcr.io/istio-release/proxyv2:1.22.8
        - gcr.io/istio-release/ztunnel:1.21.6
        - gcr.io/istio-release/pilot:1.23.5
        - gcr.io/istio-release/pilot:1.24.1
        - gcr.io/istio-release/proxyv2:1.21.6
        - gcr.io/istio-release/proxyv2:1.25.1
        - gcr.io/istio-release/ztunnel:1.23.4
        - gcr.io/istio-release/proxyv2:1.23.5
        - gcr.io/istio-release/pilot:1.25.2
        - gcr.io/istio-release/install-cni:1.21.6
        - gcr.io/istio-release/ztunnel:1.23.3
        - gcr.io/istio-release/pilot:1.22.5
        - gcr.io/istio-release/proxyv2:1.25.2
        - gcr.io/istio-release/pilot:1.24.4
        - gcr.io/istio-release/pilot:1.23.3
        - gcr.io/istio-release/pilot:1.25.1
        - gcr.io/istio-release/proxyv2:1.23.4
        - gcr.io/istio-release/install-cni:1.22.7
        - gcr.io/istio-release/ztunnel:1.24.3
        - gcr.io/istio-release/ztunnel:1.24.4
        - gcr.io/istio-release/proxyv2:1.22.6
        - gcr.io/istio-release/proxyv2:1.24.2
        - gcr.io/istio-release/ztunnel:1.23.2
        - gcr.io/istio-release/ztunnel:1.24.2
        - gcr.io/istio-release/ztunnel:1.25.2
        - gcr.io/istio-release/install-cni:1.24.4
        - gcr.io/istio-release/pilot:1.22.7
        - gcr.io/istio-release/install-cni:1.25.2
        - gcr.io/istio-release/ztunnel:1.22.7
        - gcr.io/istio-release/proxyv2:1.22.7
        - gcr.io/istio-release/ztunnel:1.22.8
        - gcr.io/istio-release/pilot:1.24.0
        - gcr.io/istio-release/pilot:1.21.6
        - gcr.io/istio-release/install-cni:1.24.1
        - gcr.io/istio-release/install-cni:1.22.5
        - gcr.io/istio-release/install-cni:1.23.2
        - gcr.io/istio-release/install-cni:1.24.0
        - gcr.io/istio-release/ztunnel:1.23.5
        - gcr.io/istio-release/proxyv2:1.24.0
        - gcr.io/istio-release/install-cni:1.22.6
        - gcr.io/istio-release/proxyv2:1.23.2
        - gcr.io/istio-release/proxyv2:1.24.1
        - gcr.io/istio-release/pilot:1.24.2
        - gcr.io/istio-release/proxyv2:1.22.5
        - gcr.io/istio-release/pilot:1.23.4
        - gcr.io/istio-release/ztunnel:1.24.0
        - gcr.io/istio-release/install-cni:1.23.4
        - gcr.io/istio-release/ztunnel:1.24.1
        - gcr.io/istio-release/proxyv2:1.24.4
        - gcr.io/istio-release/proxyv2:1.24.3
        version: 1.25.2
      entries:
      - name: sailoperator.v1.25.2
        version: 1.25.2
      - name: sailoperator.v1.25.1
        version: 1.25.1
      - name: sailoperator.v1.25.0
        version: 1.25.0
      name: stable-1.25
    - currentCSV: sailoperator.v1.26.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.3"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.3"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:d37258646d76a9f708a69fbaa081a0a95908a2b0c8837a14d7ecbfcf505f9bd4
          createdAt: "2025-08-04T06:23:45Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.26-latest
          - v1.26.3
          - v1.26.2
          - v1.26.1
          - v1.26.0
          - v1.25-latest
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/proxyv2@sha256:fd734e6031566b4fb92be38f0f6bb02fdba6c199c45c2db5dc988bbc4fdee026
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/pilot@sha256:c4e33921e030bb22d3b8fabc693851d52c07a7af79a9534addc2f50005fa462a
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - quay.io/sail-dev/sail-operator@sha256:d37258646d76a9f708a69fbaa081a0a95908a2b0c8837a14d7ecbfcf505f9bd4
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/ztunnel@sha256:2d567893e5b452d5c0b9c43a43efd4f8e28e639a344e28d90498f13858f13d46
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/install-cni@sha256:8e782f5077f63c0a15fd4b9a4fa75f8543033c28ccf2457e25afbf1aefb6a72a
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        version: 1.26.3
      entries:
      - name: sailoperator.v1.26.3
        version: 1.26.3
      - name: sailoperator.v1.26.2
        version: 1.26.2
      - name: sailoperator.v1.26.1
        version: 1.26.1
      - name: sailoperator.v1.26.0
        version: 1.26.0
      name: stable-1.26
    - currentCSV: sailoperator.v1.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
          createdAt: "2025-09-05T10:12:19Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        version: 1.27.1
      entries:
      - name: sailoperator.v1.27.1
        version: 1.27.1
      - name: sailoperator.v1.27.0
        version: 1.27.0
      name: stable-1.27
    defaultChannel: stable
    packageName: sailoperator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: keepalived-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keepalived-operator.v1.5.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KeepalivedGroup",
                "metadata": {
                  "name": "keepalivedgroup-workers"
                },
                "spec": {
                  "blacklistRouterIDs": [
                    1,
                    2
                  ],
                  "image": "registry.redhat.io/openshift4/ose-keepalived-ipfailover",
                  "interface": "ens3",
                  "nodeSelector": {
                    "node-role.kubernetes.io/loadbalancer": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Networking, Cloud Provider
          certified: "false"
          containerImage: quay.io/redhat-cop/keepalived-operator@sha256:cd0c2f7df82df1178795e4a18a0942da33adb9670b192a7b7a2b1483a5b4ccf8
          createdAt: "2023-03-11T18:48:22Z"
          description: This operator provides a simple way to create self-hosted load
            balancers using keepalived.
          operatorframework.io/suggested-namespace: keepalived-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/keepalived-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KeepalivedGroup
            name: keepalivedgroups.redhatcop.redhat.io
            version: v1alpha1
        description: "The objective of the keepalived operator is to allow for a way
          to create self-hosted load balancers in an automated way. From a user experience
          point of view the behavior is the same as of when creating [`LoadBalancer`](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          services with a cloud provider able to manage them.\n\nThe keepalived operator
          can be used in all environments that allows nodes to advertise additional
          IPs on their NICs (and at least for now, in networks that allow multicast),
          however it's mainly aimed at supporting LoadBalancer services and ExternalIPs
          on bare metal installations (or other installation environments where a
          cloud provider is not available).\n\nOne possible use of the keepalived
          operator is also to support [OpenShift Ingresses](https://docs.openshift.com/container-platform/4.5/networking/configuring_ingress_cluster_traffic/overview-traffic.html)
          in environments where an external load balancer cannot be provisioned.\n\n##
          How it works\n\nThe keepalived operator will create one or more [VIPs](https://en.wikipedia.org/wiki/Virtual_IP_address)
          (an HA IP that floats between multiple nodes), based on the [`LoadBalancer`](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          services and/or services requesting [`ExternalIPs`](https://kubernetes.io/docs/concepts/services-networking/service/#external-ips).\n\nFor
          `LoadBalancer` services the IPs found at `.Status.LoadBalancer.Ingress[].IP`
          will become VIPs.\n\nFor services requesting a `ExternalIPs`, the IPs found
          at `.Spec.ExternalIPs[]` will become VIPs.\n\nNote that a service can be
          of `LoadBalancer` type and also request `ExternalIPs`, it this case both
          sets of IPs will become VIPs.\n\nDue to a [keepalived](https://www.keepalived.org/manpage.html)
          limitation a single keepalived cluster can manage up to 256 VIP configurations.
          Multiple keepalived clusters can coexists in the same network as long as
          they use different multicast ports [TODO verify this statement].\n\nTo address
          this limitation the `KeepalivedGroup` [CRD](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/)
          has been introduced. This CRD is supposed to be configured by an administrator
          and allows you to specify a node selector to pick on which nodes the keepalived
          pods should be deployed. Here is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          KeepalivedGroup\nmetadata:\n  name: keepalivedgroup-router\nspec:\n  image:
          registry.redhat.io/openshift4/ose-keepalived-ipfailover\n  interface: ens3\n
          \ nodeSelector:\n    node-role.kubernetes.io/loadbalancer: \"\"\n  blacklistRouterIDs:\n
          \ - 1\n  - 2  \n```\n\nThis KeepalivedGroup will be deployed on all the
          nodes with role `loadbalancer`. One must also specify the network device
          on which the VIPs will be exposed, it is assumed that all the nodes have
          the same network device configuration.\n\nServices must be annotated to
          opt-in to being observed by the keepalived operator and to specify which
          KeepalivedGroup they refer to. The annotation looks like this:\n\n`keepalived-operator.redhat-cop.io/keepalivedgroup:
          <keepalivedgroup namespace>/<keepalivedgroup-name>`\n\nThe image used for
          the keepalived containers can be specified with `.Spec.Image` it will default
          to `registry.redhat.io/openshift4/ose-keepalived-ipfailover` if undefined.
          \n\n## Requirements\n\n### Security Context Constraints\n\nEach KeepalivedGroup
          deploys a [daemonset](https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/)
          that requires the [privileged scc](https://docs.openshift.com/container-platform/4.5/authentication/managing-security-context-constraints.html),
          this permission must be given to the `default` service account in the namespace
          where the keepalived group is created by and administrator.\n\n```shell\noc
          adm policy add-scc-to-user privileged -z default -n keepalived-operator\n```\n\n###
          Cluster Network Operator\n\nIn Openshift, use of an external IP address
          is governed by the following fields in the `Network.config.openshift.io`
          CR named `cluster`\n\n* `spec.externalIP.autoAssignCIDRs` defines an IP
          address block used by the load balancer when choosing an external IP address
          for the service. OpenShift supports only a single IP address block for automatic
          assignment.\n\n* `spec.externalIP.policy` defines the permissible IP address
          blocks when manually specifying an IP address. OpenShift does not apply
          policy rules to IP address blocks defined by `spec.externalIP.autoAssignCIDRs`\n\nThe
          following patch can be used to configure the Cluster Network Operator:\n\n```yaml\nspec:\n
          \ externalIP:\n    policy:\n      allowedCIDRs:\n      - ${ALLOWED_CIDR}\n
          \   autoAssignCIDRs:\n      - \"${AUTOASSIGNED_CIDR}\"\n```\n\nHere is an
          example of how to apply the patch:\n\n```shell\nexport ALLOWED_CIDR=\"192.168.131.128/26\"\nexport
          AUTOASSIGNED_CIDR=\"192.168.131.192/26\"\noc patch network cluster -p \"$(envsubst
          < ./network-patch.yaml | yq r -j -)\" --type=merge\n```\nAdditionally, the
          fields can be edited manually via `oc edit Network.config.openshift.io cluster`\n\n##
          Blacklisting router IDs\n\nIf the Keepalived pods are deployed on nodes
          which are in the same network (same broadcast domain to be precise) with
          other keepalived the process, it's necessary to ensure that there is no
          collision between the used routers it.\nFor this purpose it is possible
          to provide a `blacklistRouterIDs` field with a list of black-listed IDs
          that will not be used.\n\n## OpenShift RHV, vSphere, OSP and bare metal
          IPI instructions\n\nWhen IPI is used for RHV, vSphere, OSP or bare metal
          platforms, three keepalived VIPs are deployed. To make sure that keepalived-operator
          can work in these environment we need to discover and blacklist the corresponding
          VRRP router IDs.\n\nTo discover the VRRP router IDs being used, run the
          following command, you can run this command from you laptop:\n\n```shell\npodman
          run quay.io/openshift/origin-baremetal-runtimecfg:4.5 vr-ids <cluster_name>\n```\n\nIf
          you don't know your cluster name, run this command:\n\n```shell\npodman
          run quay.io/openshift/origin-baremetal-runtimecfg:4.5 vr-ids $(oc get cm
          cluster-config-v1 -n kube-system -o jsonpath='{.data.install-config}'| yq
          -r .metadata.name)\n```\n\nThen use these [instructions](#Blacklisting-router-IDs)
          to blacklist those VRRP router IDs.\n\n## Verbatim Configurations\n\nKeepalived
          has dozens of [configurations](https://www.keepalived.org/manpage.html).
          At the early stage of this project it's difficult to tell which one should
          be modeled in the API. Yet, users of this project may still need to use
          them. To account for that there is a way to pass verbatim options both at
          the keepalived group level (which maps to the keepalived config `global_defs`
          section) and at the service level (which maps to the keepalived config `vrrp_instance`
          section).\n\nKeepalivedGroup-level verbatim configurations can be passed
          as in the following example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          KeepalivedGroup\nmetadata:\n  name: keepalivedgroup-router\nspec:\n  interface:
          ens3\n  nodeSelector:\n    node-role.kubernetes.io/loadbalancer: \"\"\n
          \ verbatimConfig:  \n    vrrp_iptables: my-keepalived\n```\n\nthis will
          map to the following `global_defs`:\n\n```\n    global_defs {\n        router_id
          keepalivedgroup-router\n        vrrp_iptables my-keepalived\n    }\n```\n\nService-level
          verbatim configurations can be passed as in the following example:\n\n```yaml\napiVersion:
          v1\nkind: Service\nmetadata:\n  annotations:\n    keepalived-operator.redhat-cop.io/keepalivedgroup:
          keepalived-operator/keepalivedgroup-router\n    keepalived-operator.redhat-cop.io/verbatimconfig:
          '{ \"track_src_ip\": \"\" }'\n```\n\nthis will map to the following `vrrp_instance`
          section\n\n```\n    vrrp_instance openshift-ingress/router-default {\n        interface
          ens3\n        virtual_router_id 1  \n        virtual_ipaddress {\n          192.168.131.129\n
          \       }\n        track_src_ip\n    }\n```\n\n## Metrics collection\n\nEach
          keepalived pod exposes a [Prometheus](https://prometheus.io/) metrics port
          at `9650`. Metrics are collected with [keepalived_exporter](github.com/gen2brain/keepalived_exporter),
          the available metrics are described in the project documentation.\n\nWhen
          a keepalived group is created a [`PodMonitor`](https://github.com/coreos/prometheus-operator/blob/master/Documentation/api.md#podmonitor)
          rule to collect those metrics. All PodMonitor resources created that way
          have the label: `metrics: keepalived`. It is up to you to make sure your
          Prometheus instance watches for those `PodMonitor` rules. Here is an example
          of a fragment of a `Prometheus` CR configured to collect the keepalived
          pod metrics:\n\n```yaml\n  podMonitorSelector:\n    matchLabels:\n      metrics:
          keepalived\n```\n"
        displayName: Keepalived Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadbalancer
        - keepalived
        - ipfailover
        - externalIP
        - self-hosted
        links:
        - name: Keepalived Operator
          url: https://github.com/redhat-cop/keepalived-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/keepalived-operator
        maintainers:
        - email: rspazzol@redhat.com'
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/keepalived-operator@sha256:cd0c2f7df82df1178795e4a18a0942da33adb9670b192a7b7a2b1483a5b4ccf8
        version: 1.5.4
      entries:
      - name: keepalived-operator.v1.5.4
        version: 1.5.4
      - name: keepalived-operator.v1.5.3
        version: 1.5.3
      - name: keepalived-operator.v1.5.2
        version: 1.5.2
      - name: keepalived-operator.v1.5.1
        version: 1.5.1
      - name: keepalived-operator.v1.5.0
        version: 1.5.0
      - name: keepalived-operator.v1.4.2
        version: 1.4.2
      - name: keepalived-operator.v1.4.1
        version: 1.4.1
      - name: keepalived-operator.v1.4.0
        version: 1.4.0
      - name: keepalived-operator.v1.3.4
        version: 1.3.4
      - name: keepalived-operator.v1.3.3
        version: 1.3.3
      - name: keepalived-operator.v1.3.2
        version: 1.3.2
      - name: keepalived-operator.v1.3.1
        version: 1.3.1
      - name: keepalived-operator.v1.3.0
        version: 1.3.0
      - name: keepalived-operator.v1.2.2
        version: 1.2.2
      - name: keepalived-operator.v1.2.1
        version: 1.2.1
      - name: keepalived-operator.v1.2.0
        version: 1.2.0
      - name: keepalived-operator.v1.1.0
        version: 1.1.0
      - name: keepalived-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: keepalived-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel Corporation
      provider-url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
    name: sriov-fec
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sriov-fec.v2.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecClusterConfig",
                "metadata": {
                  "name": "acc100SampleConfig",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "acceleratorSelector": {
                    "deviceID": "someDevice",
                    "driver": "someDriver",
                    "maxVirtualFunctions": 2,
                    "pciAddress": "somePciAddress",
                    "vendorID": "someVendor"
                  },
                  "drainSkip": false,
                  "nodeSelector": {
                    "expectedLabel1": "valueOfExpectedLabel1",
                    "expectedLabelN": "valueOfExpectedLabelN"
                  },
                  "physicalFunction": {
                    "bbDevConfig": {
                      "acc100": {
                        "downlink4G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 0
                        },
                        "downlink5G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 4
                        },
                        "maxQueueSize": 1024,
                        "numVfBundles": 16,
                        "pfMode": true,
                        "uplink4G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 0
                        },
                        "uplink5G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 4
                        }
                      }
                    },
                    "pfDriver": "pci-pf-stub",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  },
                  "priority": 100
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecClusterConfig",
                "metadata": {
                  "name": "n3000SampleConfig",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "acceleratorSelector": {
                    "deviceID": "someDevice",
                    "driver": "someDriver",
                    "maxVirtualFunctions": 2,
                    "pciAddress": "somePciAddress",
                    "vendorID": "someVendor"
                  },
                  "drainSkip": false,
                  "nodeSelector": {
                    "expectedLabel1": "valueOfExpectedLabel1",
                    "expectedLabelN": "valueOfExpectedLabelN"
                  },
                  "physicalFunction": {
                    "bbDevConfig": {
                      "n3000": {
                        "downlink": {
                          "bandwidth": 3,
                          "loadBalance": 128,
                          "queues": {
                            "vf0": 16,
                            "vf1": 16,
                            "vf2": 0,
                            "vf3": 0,
                            "vf4": 0,
                            "vf5": 0,
                            "vf6": 0,
                            "vf7": 0
                          }
                        },
                        "flrTimeout": 610,
                        "networkType": "FPGA_5GNR",
                        "pfMode": true,
                        "uplink": {
                          "bandwidth": 3,
                          "loadBalance": 128,
                          "queues": {
                            "vf0": 16,
                            "vf1": 16,
                            "vf2": 0,
                            "vf3": 0,
                            "vf4": 0,
                            "vf5": 0,
                            "vf6": 0,
                            "vf7": 0
                          }
                        }
                      }
                    },
                    "pfDriver": "pci-pf-stub",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  },
                  "priority": 100
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecNodeConfig",
                "metadata": {
                  "name": "acc100-worker",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "drainSkip": false,
                  "physicalFunctions": [
                    {
                      "bbDevConfig": {
                        "acc100": {
                          "downlink4G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 0
                          },
                          "downlink5G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 4
                          },
                          "maxQueueSize": 1024,
                          "numVfBundles": 16,
                          "pfMode": true,
                          "uplink4G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 0
                          },
                          "uplink5G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 4
                          }
                        }
                      },
                      "pci_addr": "somePciAddress",
                      "pf_driver": "pci-pf-stub",
                      "vf_amount": 2,
                      "vf_driver": "vfio-pci"
                    }
                  ]
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecNodeConfig",
                "metadata": {
                  "name": "n3000-worker",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "drainSkip": false,
                  "physicalFunctions": [
                    {
                      "bbDevConfig": {
                        "n3000": {
                          "downlink": {
                            "bandwidth": 3,
                            "loadBalance": 128,
                            "queues": {
                              "vf0": 16,
                              "vf1": 16,
                              "vf2": 0,
                              "vf3": 0,
                              "vf4": 0,
                              "vf5": 0,
                              "vf6": 0,
                              "vf7": 0
                            }
                          },
                          "flrTimeout": 610,
                          "networkType": "FPGA_5GNR",
                          "pfMode": true,
                          "uplink": {
                            "bandwidth": 3,
                            "loadBalance": 128,
                            "queues": {
                              "vf0": 16,
                              "vf1": 16,
                              "vf2": 0,
                              "vf3": 0,
                              "vf4": 0,
                              "vf5": 0,
                              "vf6": 0,
                              "vf7": 0
                            }
                          }
                        }
                      },
                      "pci_addr": "somePciAddress",
                      "pf_driver": "pci-pf-stub",
                      "vf_amount": 2,
                      "vf_driver": "vfio-pci"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/intel/sriov-fec-operator@sha256:adb6a4c2da6399c8ed40c906f790bce13bfcefae3326a047f6a50592b3f5e485
          createdAt: "2025-01-15"
          description: An operator for Intel Wireless FEC Accelerator to orchestrate
            and manage the resources/devices exposed by a range of Intel's vRAN FEC
            acceleration devices/hardware within the OpenShift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/intel/sriov-fec-operator
          support: Intel Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovFecClusterConfig is the Schema for the sriovfecclusterconfigs
              API
            displayName: SriovFecClusterConfig
            kind: SriovFecClusterConfig
            name: sriovfecclusterconfigs.sriovfec.intel.com
            version: v2
          - description: SriovFecNodeConfig is the Schema for the sriovfecnodeconfigs
              API
            displayName: SriovFecNodeConfig
            kind: SriovFecNodeConfig
            name: sriovfecnodeconfigs.sriovfec.intel.com
            version: v2
          - description: SriovVrbClusterConfig is the Schema for the sriovvrbclusterconfigs
              API
            displayName: SriovVrbClusterConfig
            kind: SriovVrbClusterConfig
            name: sriovvrbclusterconfigs.sriovvrb.intel.com
            version: v1
          - description: SriovVrbNodeConfig is the Schema for the sriovvrbnodeconfigs
              API
            displayName: SriovVrbNodeConfig
            kind: SriovVrbNodeConfig
            name: sriovvrbnodeconfigs.sriovvrb.intel.com
            version: v1
        description: The Intel vRAN Boost family of accelerators ACC100/VRB1/VRB2,
          to offload and accelerate the computing-intensive process of forward error
          correction (FEC) and more for 4G/LTE and 5G technology, freeing up processing
          power. It allows the optimization of data plane performance to reduce total
          cost of ownership while maintaining a high degree of flexibility.  The Intel
          vRAN Boost accelerators ACC100/VRB1/VRB2 plays a key role in accelerating
          5G and network functions virtualization (NFV) adoption for ecosystem partners
          such as telecommunications equipment manufacturers (TEMs) virtual network
          functions (VNF) vendors, system integrators and telcos, to bring scalable
          and high-performance solutions to market. The Intel ACC100 includes a variant
          that is design to be Network Equipment Building System (NEBS)-friendly,
          and features a Root-of-Trust device that helps protect systems from FPGA
          host security exploits.
        displayName: SRIOV-FEC Operator for Intel vRAN Boost Accelerators
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ACC100
        - vRAN
        - ORAN
        links:
        - name: SRIOV-FEC source code
          url: https://github.com/intel/sriov-fec-operator
        maturity: alpha
        provider:
          name: Intel Corporation
          url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
        relatedImages:
        - registry.connect.redhat.com/intel/sriov-fec-daemon@sha256:5b0c60b67bcf5d7b2797deab0d5e3fa079577179f0acc6b0062533584ea0bc23
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1ed3b514bf9af6ac0d386e3739874d1d964e667d54e041c2d67edd769b0009fb
        - registry.connect.redhat.com/intel/sriov-fec-operator@sha256:adb6a4c2da6399c8ed40c906f790bce13bfcefae3326a047f6a50592b3f5e485
        - registry.connect.redhat.com/intel/n3000-labeler@sha256:b417cfcedc998a35491784005cc7eee3d9c647b84eac5ca41cb459076870b8f6
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin@sha256:d0fcb6e9135d9a19ea26851bcdb4ac72c04bd747706fc5123020460e9dbfc21b
        version: 2.11.1
      entries:
      - name: sriov-fec.v2.11.1
        version: 2.11.1
      - name: sriov-fec.v2.11.0
        version: 2.11.0
      - name: sriov-fec.v2.10.0
        version: 2.10.0
      - name: sriov-fec.v2.9.0
        version: 2.9.0
      - name: sriov-fec.v2.8.0
        version: 2.8.0
      - name: sriov-fec.v2.7.2
        version: 2.7.2
      - name: sriov-fec.v2.7.1
        version: 2.7.1
      name: stable
    defaultChannel: stable
    packageName: sriov-fec
    provider:
      name: Intel Corporation
      url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: spyre-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spyre-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreClusterPolicy",
                "metadata": {
                  "name": "spyreclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2",
                      "spyreFilter": ".",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "latest"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "spyre-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "spyre-device-plugin-init",
                      "repository": "icr.io/ibmaiu_internal",
                      "version": "0.1.1"
                    },
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "spyre-exporter",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.2"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "spyre-webhook-validator",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  },
                  "scheduler": {
                    "image": "spyre-scheduler",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  }
                }
              },
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu_internal/spyre-operator@sha256:0f657c98668a8358c37b3fc2ed3f31ec78b7ef7f9a9678e7cd35d383d22ce2da
          createdAt: "2025-07-31T10:32:48Z"
          description: Operator for IBM Spyre devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM Spyre
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["spyrenodestates.spyre.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpyreClusterPolicy is the Schema for the Spyre API
            displayName: Spyre Cluster Policy
            kind: SpyreClusterPolicy
            name: spyreclusterpolicies.spyre.ibm.com
            version: v1alpha1
          - description: SpyreNodeState is the Schema for the SpyreNodeState API
            displayName: Spyre Node State
            kind: SpyreNodeState
            name: spyrenodestates.spyre.ibm.com
            version: v1alpha1
        description: Operator for IBM Spyre devices
        displayName: IBM Spyre Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: IBM Spyre Operator
          url: https://spyre-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: Satyam.Gupta.3@ibm.com
          name: Satyam Gupta
        - email: Rehan.Khan7@ibm.com
          name: Rehan Khan
        - email: Nishan.Acharya@ibm.com
          name: Nishan Acharya
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu_internal/spyre-operator@sha256:0f657c98668a8358c37b3fc2ed3f31ec78b7ef7f9a9678e7cd35d383d22ce2da
        - icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2
        version: 0.1.2
      entries:
      - name: spyre-operator.v0.1.2
        version: 0.1.2
      name: stable-v0.1
    - currentCSV: spyre-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreClusterPolicy",
                "metadata": {
                  "name": "spyreclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb",
                      "spyreFilter": ".",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "latest"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "spyre-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "spyre-device-plugin-init",
                      "repository": "quay.io/ibm-aiu",
                      "version": "0.2.0"
                    },
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "spyre-exporter",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "spyre-webhook-validator",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "scheduler": {
                    "image": "spyre-scheduler",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.1"
                  }
                }
              },
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/ibm-aiu/spyre-operator@sha256:d0cdd449f1d8f0dd9281c4d9a809333db6c4ed201bb14187b6c70ae686a30e99
          createdAt: "2025-09-02T13:25:34Z"
          description: Operator for IBM Spyre devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM Spyre
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["spyrenodestates.spyre.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpyreClusterPolicy is the Schema for the Spyre API
            displayName: Spyre Cluster Policy
            kind: SpyreClusterPolicy
            name: spyreclusterpolicies.spyre.ibm.com
            version: v1alpha1
          - description: SpyreNodeState is the Schema for the SpyreNodeState API
            displayName: Spyre Node State
            kind: SpyreNodeState
            name: spyrenodestates.spyre.ibm.com
            version: v1alpha1
        description: Operator for IBM Spyre devices
        displayName: IBM Spyre Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: IBM Spyre Operator
          url: https://spyre-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: Satyam.Gupta.3@ibm.com
          name: Satyam Gupta
        - email: Rehan.Khan7@ibm.com
          name: Rehan Khan
        - email: Nishan.Acharya@ibm.com
          name: Nishan Acharya
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - quay.io/ibm-aiu/spyre-operator@sha256:d0cdd449f1d8f0dd9281c4d9a809333db6c4ed201bb14187b6c70ae686a30e99
        version: 0.2.1
      entries:
      - name: spyre-operator.v0.2.1
        version: 0.2.1
      name: stable-v0.2
    - currentCSV: spyre-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreClusterPolicy",
                "metadata": {
                  "name": "spyreclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "config": {
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb",
                      "spyreFilter": ".",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "latest"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "spyre-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "spyre-device-plugin-init",
                      "repository": "quay.io/ibm-aiu",
                      "version": "1.0.0"
                    },
                    "repository": "quay.io/ibm-aiu",
                    "version": "1.0.0"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "spyre-exporter",
                    "repository": "quay.io/ibm-aiu",
                    "version": "1.0.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "spyre-webhook-validator",
                    "repository": "quay.io/ibm-aiu",
                    "version": "1.0.0"
                  },
                  "scheduler": {
                    "image": "spyre-scheduler",
                    "repository": "quay.io/ibm-aiu",
                    "version": "1.0.0"
                  }
                }
              },
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/ibm-aiu/spyre-operator@sha256:dd262e8c65f35b12c5d79571e8b5ddd54dd0ee2693f28bea05a6fc02ac1fe6f0
          createdAt: "2025-10-01T00:40:23Z"
          description: Operator for IBM Spyre devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: spyre-operator
          operators.openshift.io/valid-subscription: use with IBM Spyre
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["spyrenodestates.spyre.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpyreClusterPolicy is the Schema for the Spyre API
            displayName: Spyre Cluster Policy
            kind: SpyreClusterPolicy
            name: spyreclusterpolicies.spyre.ibm.com
            version: v1alpha1
          - description: SpyreNodeState is the Schema for the SpyreNodeState API
            displayName: Spyre Node State
            kind: SpyreNodeState
            name: spyrenodestates.spyre.ibm.com
            version: v1alpha1
        description: Operator for IBM Spyre devices
        displayName: IBM Spyre Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: IBM Spyre Operator
          url: https://spyre-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: Satyam.Gupta.3@ibm.com
          name: Satyam Gupta
        - email: Rehan.Khan7@ibm.com
          name: Rehan Khan
        - email: Nishan.Acharya@ibm.com
          name: Nishan Acharya
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - quay.io/ibm-aiu/spyre-operator@sha256:dd262e8c65f35b12c5d79571e8b5ddd54dd0ee2693f28bea05a6fc02ac1fe6f0
        version: 1.0.0
      entries:
      - name: spyre-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: spyre-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hive-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hive-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:e71fc728632622f7ea97357ec361dd2bb73e63c0707e8014ca8ecf8ec4ab93f1
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:e71fc728632622f7ea97357ec361dd2bb73e63c0707e8014ca8ecf8ec4ab93f1
        version: 24.11.1
      entries:
      - name: hive-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: hive-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:5530f5034cd83c2e8b860fce737ac491a2adc6d22246356bee0c79768ed64706
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:5530f5034cd83c2e8b860fce737ac491a2adc6d22246356bee0c79768ed64706
        version: 25.3.0
      entries:
      - name: hive-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: hive-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
        version: 25.7.0
      entries:
      - name: hive-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hive-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
        version: 25.7.0
      entries:
      - name: hive-operator.v25.7.0
        version: 25.7.0
      - name: hive-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hive-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CSI Addons Community
      provider-url: ""
    name: odf-csi-addons-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-csi-addons-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          createdAt: "2025-09-12T07:33:01Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIAddonsNode is the Schema for the csiaddonsnode API
            displayName: CSIAddons Node
            kind: CSIAddonsNode
            name: csiaddonsnodes.csiaddons.openshift.io
            version: v1alpha1
          - description: NetworkFence is the Schema for the networkfences API
            displayName: Network Fence
            kind: NetworkFence
            name: networkfences.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceCronJob is the Schema for the reclaimspacecronjobs
              API
            displayName: Reclaim Space CronJob
            kind: ReclaimSpaceCronJob
            name: reclaimspacecronjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceJob is the Schema for the reclaimspacejobs API
            displayName: Reclaim Space Job
            kind: ReclaimSpaceJob
            name: reclaimspacejobs.csiaddons.openshift.io
            version: v1alpha1
          - description: VolumeReplicationClass is the Schema for the volumereplicationclasses
              API
            displayName: Volume Replication Class
            kind: VolumeReplicationClass
            name: volumereplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplication is the Schema for the volumereplications
              API
            displayName: Volume Replication
            kind: VolumeReplication
            name: volumereplications.replication.storage.openshift.io
            version: v1alpha1
        description: CSI Addons provides the CSI Addons Controller that enables advanced
          storage operations for CSI-drivers.
        displayName: CSI Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: CSI Addons
          url: https://csi-addons.github.io
        maintainers:
        - email: csi-addons@redhat.com
          name: CSI Addons Community
        maturity: alpha
        provider:
          name: CSI Addons Community
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/odf4/odf-csi-addons-rhel9-operator@sha256:8b0c16c21fccca9b674c4d980a793b6cde482b86363b793e173126b030bdcc55
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:9a0008f662b6afa9e2677a8a83bd662d806ad6f4d03fb49bccb78ed487ce4c7a
        version: 4.15.20-rhodf
      entries:
      - name: odf-csi-addons-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odf-csi-addons-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odf-csi-addons-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odf-csi-addons-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odf-csi-addons-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odf-csi-addons-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odf-csi-addons-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odf-csi-addons-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odf-csi-addons-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odf-csi-addons-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odf-csi-addons-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odf-csi-addons-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odf-csi-addons-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odf-csi-addons-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odf-csi-addons-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odf-csi-addons-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odf-csi-addons-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odf-csi-addons-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odf-csi-addons-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odf-csi-addons-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odf-csi-addons-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          createdAt: "2025-09-12T06:32:07Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIAddonsNode is the Schema for the csiaddonsnode API
            displayName: CSIAddons Node
            kind: CSIAddonsNode
            name: csiaddonsnodes.csiaddons.openshift.io
            version: v1alpha1
          - description: NetworkFence is the Schema for the networkfences API
            displayName: Network Fence
            kind: NetworkFence
            name: networkfences.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceCronJob is the Schema for the reclaimspacecronjobs
              API
            displayName: Reclaim Space CronJob
            kind: ReclaimSpaceCronJob
            name: reclaimspacecronjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceJob is the Schema for the reclaimspacejobs API
            displayName: Reclaim Space Job
            kind: ReclaimSpaceJob
            name: reclaimspacejobs.csiaddons.openshift.io
            version: v1alpha1
          - description: VolumeReplicationClass is the Schema for the volumereplicationclasses
              API
            displayName: Volume Replication Class
            kind: VolumeReplicationClass
            name: volumereplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplication is the Schema for the volumereplications
              API
            displayName: Volume Replication
            kind: VolumeReplication
            name: volumereplications.replication.storage.openshift.io
            version: v1alpha1
        description: CSI Addons provides the CSI Addons Controller that enables advanced
          storage operations for CSI-drivers.
        displayName: CSI Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: CSI Addons
          url: https://csi-addons.github.io
        maintainers:
        - email: csi-addons@redhat.com
          name: CSI Addons Community
        maturity: alpha
        provider:
          name: CSI Addons Community
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/odf4/odf-csi-addons-rhel9-operator@sha256:d370f19578d1c466d39d00da82f1f3fe3d9f13d7021ed0c39f759b24b0ed7573
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:0afcb516eb528857887c6ee1a9cba081dc8f6b7d47d9f26a0dbb2dca0cefe07c
        version: 4.16.16-rhodf
      entries:
      - name: odf-csi-addons-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odf-csi-addons-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odf-csi-addons-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odf-csi-addons-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odf-csi-addons-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odf-csi-addons-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odf-csi-addons-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odf-csi-addons-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odf-csi-addons-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odf-csi-addons-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odf-csi-addons-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odf-csi-addons-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odf-csi-addons-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odf-csi-addons-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odf-csi-addons-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odf-csi-addons-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odf-csi-addons-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odf-csi-addons-operator
    provider:
      name: CSI Addons Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Weights & Biases
      provider-url: https://wandb.ai
    name: wandb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: wandb-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.wandb.com/v1",
                "kind": "WeightsAndBiases",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "wandb-operator",
                    "app.kubernetes.io/instance": "weightsandbiases-sample",
                    "app.kubernetes.io/managed-by": "olm",
                    "app.kubernetes.io/name": "weightsandbiases",
                    "app.kubernetes.io/part-of": "wandb-operator"
                  },
                  "name": "weightsandbiases-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning,Developer Tools
          certified: "false"
          containerImage: quay.io/wandb_tools/wandb-k8s-operator:1.0.0
          createdAt: "2024-10-10T14:46:54Z"
          description: Operator for Weights and Biases, enabling seamless integration
            and management of machine learning experiments in Kubernetes environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: wandb-system
          repository: https://github.com/wandb/operator
          support: WeightsAndBiases
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: WeightsAndBiases is a custom resource for managing W&B configurations.
            displayName: Weights And Biases
            kind: WeightsAndBiases
            name: weightsandbiases.apps.wandb.com
            version: v1
        description: |-
          The Weights and Biases Operator simplifies the management of machine learning experiments, tracking, and infrastructure. It integrates with Kubernetes to deploy and manage Weights and Biases instances.
          Features:
          - Experiment Tracking
          - Hyperparameter Tuning
          - Model Management
          - Integration with popular ML frameworks
        displayName: Weights & Biases Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - machine learning
        - experiment tracking
        - ML infrastructure
        - weights and biases
        links:
        - name: Weights & Biases Documentation
          url: https://docs.wandb.ai/
        maintainers:
        - email: abhinav.garg@domain.com
          name: Abhinav Garg
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Weights & Biases
          url: https://wandb.ai
        relatedImages:
        - quay.io/wandb_tools/wandb-k8s-operator:1.0.0
        version: 1.0.0
      entries:
      - name: wandb-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: wandb-operator
    provider:
      name: Weights & Biases
      url: https://wandb.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: multicluster-operators-subscription
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-operators-subscription.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.3.0
          createdAt: "2021-07-19T18:00:00Z"
          description: An operator to subscribe resources from object Bucket, helm
            and github types of channels
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-application:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-channel:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-placementrule:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-deployable:2.3.0
        version: 0.3.0
      entries:
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.3
    - currentCSV: multicluster-operators-subscription.v0.3.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.4.0
          createdAt: "2021-09-29T10:23:00Z"
          description: An operator to subscribe resources from object Bucket, helm
            and github types of channels
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-placementrule:2.4.0
        - quay.io/open-cluster-management/multicloud-integrations:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-deployable:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-application:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-channel:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.4.0
        version: 0.3.6
      entries:
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.4
    - currentCSV: multicluster-operators-subscription.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.5.0
          createdAt: "2022-03-23T18:00:00Z"
          description: An operator for deploying resources from object Bucket, helm
            repo or github repo to multiple managed clusters
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management-io/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: subscription status per cluster per app
            displayName: App Subscription Report
            kind: SubscriptionReport
            name: subscriptionreports.apps.open-cluster-management.io
            version: v1alpha1
          - description: subscription status per package
            displayName: App Subscription Status
            kind: SubscriptionStatus
            name: subscriptionstatuses.apps.open-cluster-management.io
            version: v1alpha1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.5.0
        - quay.io/open-cluster-management/multicloud-integrations:2.5.0
        - quay.io/open-cluster-management/multicluster-operators-application:2.5.0
        - quay.io/open-cluster-management/multicluster-operators-channel:2.5.0
        version: 0.4.0
      entries:
      - name: multicluster-operators-subscription.v0.4.0
        version: 0.4.0
      - name: multicluster-operators-subscription.v0.3.9
        version: 0.3.9
      - name: multicluster-operators-subscription.v0.3.8
        version: 0.3.8
      - name: multicluster-operators-subscription.v0.3.7
        version: 0.3.7
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.5
    - currentCSV: multicluster-operators-subscription.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-operators-subscription:2.6.0
          createdAt: "2022-06-06T18:00:00Z"
          description: An operator for deploying resources from object Bucket, helm
            repo or github repo to multiple managed clusters
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management-io/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: subscription status per cluster per app
            displayName: App Subscription Report
            kind: SubscriptionReport
            name: subscriptionreports.apps.open-cluster-management.io
            version: v1alpha1
          - description: subscription status per package
            displayName: App Subscription Status
            kind: SubscriptionStatus
            name: subscriptionstatuses.apps.open-cluster-management.io
            version: v1alpha1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/multicluster-operators-subscription:2.6.0
        - quay.io/stolostron/multicloud-integrations:2.6.0
        - quay.io/stolostron/multicluster-operators-application:2.6.0
        - quay.io/stolostron/multicluster-operators-channel:2.6.0
        version: 0.4.1
      entries:
      - name: multicluster-operators-subscription.v0.4.1
        version: 0.4.1
      - name: multicluster-operators-subscription.v0.4.0
        version: 0.4.0
      - name: multicluster-operators-subscription.v0.3.9
        version: 0.3.9
      - name: multicluster-operators-subscription.v0.3.8
        version: 0.3.8
      - name: multicluster-operators-subscription.v0.3.7
        version: 0.3.7
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.6
    defaultChannel: release-2.5
    packageName: multicluster-operators-subscription
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.90.0-beta.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"externalDbIP":
            "10.0.2.15","customImageNames": false,"repository": "icr.io/cpopen/turbonomic","tag":"8.17.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
          createdAt: "2023-03-24T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=42.53.0-beta.1 <42.90.0-beta.2'
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints. Turbonomic platform install and configuration
              documentation can be found online here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How do I start?

          Turbonomic platform installation and configuration documentation can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
          Carefully read through the prerequisites before beginning the installation.  Once ready, on the `Details` tab for the Turbonomic Platform Operator, under `Provided APIs`, click `Create instance`.
          For samples of common configurations, refer to sample custom resource YAMLs provided here: https://github.com/turbonomic/t8c-install/tree/master/samples
          Looking to upgrade your current deployment? Refer to documentation here: https://www.ibm.com/docs/en/tarm/latest?topic=version-updating-using-red-hat-openshift-operatorhub

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
        version: 42.90.0-beta.2
      entries:
      - name: t8c-operator.v42.90.0-beta.2
        version: 42.90.0-beta.2
      name: beta
    - currentCSV: t8c-operator.v42.90.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"externalDbIP":
            "10.0.2.15","customImageNames": false,"repository": "icr.io/cpopen/turbonomic","tag":"8.17.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
          createdAt: "2023-03-24T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=42.53.0 <42.90.0'
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints. Turbonomic platform install and configuration
              documentation can be found online here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How do I start?

          Turbonomic platform installation and configuration documentation can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
          Carefully read through the prerequisites before beginning the installation.  Once ready, on the `Details` tab for the Turbonomic Platform Operator, under `Provided APIs`, click `Create instance`.
          For samples of common configurations, refer to sample custom resource YAMLs provided here: https://github.com/turbonomic/t8c-install/tree/master/samples
          Looking to upgrade your current deployment? Refer to documentation here: https://www.ibm.com/docs/en/tarm/latest?topic=version-updating-using-red-hat-openshift-operatorhub

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
        version: 42.90.0
      entries:
      - name: t8c-operator.v42.90.0
        version: 42.90.0
      name: stable
    defaultChannel: stable
    packageName: t8c-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: https://stakater.com
    name: tenant-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tenant-operator.v0.12.66
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: =0.12.65
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:9f806a065b25229741dac8437a32ec45a99acecf4e6bcbe6e7d1fa4192bc560d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a35b96eacb4e46114b39354f66429f8f05df09c55c98c9416b4ffa2768310ef7
        version: 0.12.66
      entries:
      - name: tenant-operator.v0.12.66
        version: 0.12.66
      name: release-0.12
    - currentCSV: tenant-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2024-10-16T08:15:54Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.0
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:feb71a5045e621f8a8cbfe7e41831100b6b85af9812138071b6217cc4396860a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a35b96eacb4e46114b39354f66429f8f05df09c55c98c9416b4ffa2768310ef7
        version: 1.0.0
      entries:
      - name: tenant-operator.v1.0.0
        version: 1.0.0
      name: release-1.0
    - currentCSV: tenant-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-01-27T13:23:27Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.0
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:f412b6d05038a185f5fe35f9c917388194a5ea66afb538a513423c24390983ce
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a35b96eacb4e46114b39354f66429f8f05df09c55c98c9416b4ffa2768310ef7
        version: 1.1.0
      entries:
      - name: tenant-operator.v1.1.0
        version: 1.1.0
      name: release-1.1
    - currentCSV: tenant-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-03-19T13:30:50Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.2.1
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a35b96eacb4e46114b39354f66429f8f05df09c55c98c9416b4ffa2768310ef7
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:a899d60b31efd5cf2358efcd48737b1c4f4a0356ab6c0abe7a5f62ea825a52cd
        version: 1.2.1
      entries:
      - name: tenant-operator.v1.2.1
        version: 1.2.1
      name: release-1.2
    - currentCSV: tenant-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-06-11T15:25:04Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.3.0
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:e88aa5b008ece406bca35ec22fdc87c842544049d613d4d81c8bc5dd2fb50812
        version: 1.3.0
      entries:
      - name: tenant-operator.v1.3.0
        version: 1.3.0
      name: release-1.3
    - currentCSV: tenant-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-09-08T08:14:28Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.4.2
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:7ee14c0974d1eda849d2c199a56786b0ae777ef6569085be917449f1998b1158
        version: 1.4.2
      entries:
      - name: tenant-operator.v1.4.2
        version: 1.4.2
      name: release-1.4
    defaultChannel: release-1.4
    packageName: tenant-operator
    provider:
      name: Stakater
      url: https://stakater.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.0.4
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
        version: 1.0.4
      entries:
      - name: datadog-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      name: stable
    defaultChannel: stable
    packageName: datadog-operator-certified-rhmp
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp, Inc.
      provider-url: https://www.netapp.com/
    name: trident-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trident-operator.v25.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "trident.netapp.io/v1",
                "kind": "TridentOrchestrator",
                "metadata": {
                  "name": "trident"
                },
                "spec": {
                  "IPv6": false,
                  "debug": true,
                  "nodePrep": [],
                  "imagePullSecrets": [],
                  "imageRegistry": "",
                  "namespace": "trident",
                  "silenceAutosupport": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: docker.io/netapp/trident-operator:25.06.2
          createdAt: "2025-08-03"
          description: Trident Operator, to manage NetApp Trident installations
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          repository: https://github.com/netapp/trident
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Used to deploy Astra Trident.
            displayName: Trident Orchestrator
            kind: TridentOrchestrator
            name: tridentorchestrators.trident.netapp.io
            version: v1
          - description: Automates AWS FSxN backend configuration
            displayName: Trident Configurator
            kind: TridentConfigurator
            name: tridentconfigurators.trident.netapp.io
            version: v1
        description: |
          NetApp Trident is an open source storage provisioner and orchestrator maintained by NetApp. It enables you to create storage volumes for containerized applications managed by Docker and Kubernetes. For full release information, including patch release changes, see https://docs.netapp.com/us-en/trident/trident-rn.html.
        displayName: NetApp Trident
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - NetApp
        - ONTAP
        - Trident
        - storage
        - performance
        - open source
        - Astra
        - Cloud Native
        - AWS
        - Azure
        - GCP
        - Cloud Volumes
        links:
        - name: GitHub Repository
          url: https://github.com/NetApp/trident
        - name: Trident documentation
          url: https://docs.netapp.com/us-en/trident/index.html
        - name: Support policy
          url: https://mysupport.netapp.com/site/info/version-support
        - name: Release Notes
          url: https://docs.netapp.com/us-en/trident/trident-rn.html
        maintainers:
        - email: ng-astra-product@netapp.com
          name: NetApp
        provider:
          name: NetApp, Inc.
          url: https://www.netapp.com/
        relatedImages:
        - docker.io/netapp/trident-operator:25.06.2
        version: 25.6.2
      entries:
      - name: trident-operator.v25.6.2
        version: 25.6.2
      - name: trident-operator.v25.6.1
        version: 25.6.1
      - name: trident-operator.v25.6.0
        version: 25.6.0
      - name: trident-operator.v25.2.1
        version: 25.2.1
      - name: trident-operator.v25.2.0
        version: 25.2.0
      - name: trident-operator.v24.10.1
        version: 24.10.1
      - name: trident-operator.v24.10.0
        version: 24.10.0
      - name: trident-operator.v24.6.0
        version: 24.6.0
      - name: trident-operator.v24.2.0
        version: 24.2.0
      name: stable
    defaultChannel: stable
    packageName: trident-operator
    provider:
      name: NetApp, Inc.
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeVirt project
      provider-url: ""
    name: community-kubevirt-hyperconverged
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubevirt-hyperconverged-operator.v1.10.7
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployTektonTaskResources":true,"deployVmConsoleProxy":true},"tektonPipelines":{"namespace":"kubevirt"},"tektonTasks":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:0ee30557d5d5a42dec19121e70ed77fbdacd8932f55b11e2c7c1a3d5ad5df4aa
          createdAt: "2024-02-09 14:34:30"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:38293a1050bce9c7cb786cbe8719be1edb91a922a6c583eedb61e56729ed4c72
        - quay.io/kubevirt/virt-controller@sha256:0789fafed2913b35a771e3db882748502b3250be04ece86d97f30201779b4e54
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:e79027973b09aac75860f267c7d7f830599978c38a081001f404b9a1c3d2990f
        - quay.io/kubevirt/mtq-lock-server@sha256:9614df0d5af7c998bcceaf73b548f2e3b29ff8ee4ca64c32613bfbdd6e452f6c
        - quay.io/kubevirt/cdi-uploadproxy@sha256:551221d79902a5053d1c734b81163d69f087217e2ac13c49bdf6900336ef0786
        - quay.io/kubevirt/virt-operator@sha256:cbca7f93afbb18a840f7e763e4c03aeeb1fd97967f381c12c443cc179b51da29
        - quay.io/kubevirt/hostpath-provisioner@sha256:eae30f37b3e837478a4500144b7bc8eb4f5ddf15f49fa7354bc9d5e240ea8aea
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5ff5392cb762a41ee9bc4767e3a693249ae08b46a4322f623f24ad6b63dd06b5
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:57573a24923e5588bca6bc337a8b2b08406c5b77583974365d2cf063c0dd5d06
        - quay.io/kubevirt/cni-default-plugins@sha256:825e3f9fec1996c54a52cec806154945b38f76476b160d554c36e38dfffe5e61
        - quay.io/kubevirt/kubevirt-template-validator@sha256:a148b6e812b70f326b48d0439b1e44a967a6e1f8cf2f98b1a08e9294a93674eb
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:f64ab5981d9ddef1662dd0a05888783f5dbadda8ede596585ba80bfc97a0f51e
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:f7567bea7f24184da06c3ed87b396c804342fcfd89f54989622780ce0b7b0724
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:3fbcc32bd4e4d15bd93c96def784a229cd84cca27942bf4858b581f31c97ee02
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4dc0b87ccd69f9865b89234d8555d3a614ab0a16ed94a3016ffd27f8106132ce
        - quay.io/kubevirt/virtio-container-disk@sha256:f8e71a3a930ce95d646d80862d5ff6f9d3ed7ef2541dd2e2acec0bce26274734
        - quay.io/kubevirt/virt-api@sha256:707003b221496b4432da2f507d1e36e528b45888b5d321e06d460f0678da44ae
        - quay.io/kubevirt/virt-artifacts-server:1.10.7-unstable
        - quay.io/kubevirt/virt-handler@sha256:138dfda5fea8622f3da0d6413fe214fef80c2fd6a6f9533592a0dbfa7e1865b5
        - quay.io/kubevirt/hostpath-csi-driver@sha256:0ca8b50b0adb20afd4e20da9974200f3290b6d51aa7c1129de6e1b953bf26594
        - quay.io/kubevirt/mtq-controller@sha256:96dd73ea0d7847d9f6322192559747bd85655e25ee25482d112e6e3c6fef21b2
        - quay.io/kubevirt/libguestfs-tools@sha256:b7e248be783691daad1326cf41641d5042be80c5c057ebc5315ca7a1619b0581
        - quay.io/kubevirt/cdi-apiserver@sha256:e9e39408413b1478d2e98eba68913f9e20c93000558b190b47de73bdfd1d9ac4
        - quay.io/kubevirt/ovs-cni-plugin@sha256:08f72edf2bef876bba0b0f5513d30225304ad5e7ad6912a61c083664acdb99ff
        - quay.io/kubevirt/macvtap-cni@sha256:434420511e09b2b5ede785a2c9062b6658ffbc26fbdd4629ce06110f9039c600
        - quay.io/kubevirt/cdi-importer@sha256:3143bbc67cdc6267eb48b7eaac664b8551ac4c11401dfbf4921efd3f233e6ce9
        - ghcr.io/kubevirt/kubesecondarydns@sha256:e87e829380a1e576384145f78ccaa885ba1d5690d5de7d0b73d40cfb804ea24d
        - quay.io/kubevirt/virt-exportproxy@sha256:f14444b0200a85efb4b3c176caefe70aabba7d295e33d2af14027a59ce297e24
        - quay.io/kubevirt/cdi-operator@sha256:6c63521d835578b0fbb77d3145b648f95ab0932d5d1b36ff2068ed8fcd91bc5a
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:16ecabb1d733aa17084cd16d88963c5255d04f389e0f4d41941d67e8be973594
        - quay.io/kubevirt/virt-launcher@sha256:4c5fce3de2e2589197de72fb0c9436490ea318aca952c05a622c43e067023f35
        - quay.io/kubevirt/bridge-marker@sha256:bba066e3b5ff3fb8c5e20861fe8abe51e3c9b50ad6ce3b2616af9cb5479a06d0
        - quay.io/kubevirt/cdi-uploadserver@sha256:30f1827d3696cf996b081c22c3267ca78e7219c872fdb54950198fa54359f6ee
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - quay.io/kubevirt/cdi-controller@sha256:27c47883a08226f83757971d3adafb0cd9bcb26e58fbcf7208236070e0adf37e
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:0ee30557d5d5a42dec19121e70ed77fbdacd8932f55b11e2c7c1a3d5ad5df4aa
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:86c10c122767661e31ee26a9e9de7ea606c2c1168b32793b4ea567ede2ae3b21
        - quay.io/kubevirt/virt-exportserver@sha256:73311f79a9c71007f8572b3cc40cd6f6da404c7ef0a9c6509fb717d979546582
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f6717ce72a2615c7fbc746b4068f788e78579c54c43b8716e5ce650d97af2df1
        - quay.io/kubevirt/kubemacpool@sha256:3214020c21204c189cdf2d767beb8efb61ee3ef65dd8de7f0f1edc61e13e8adf
        - quay.io/kubevirt/ssp-operator@sha256:49cac00844f091d6e80301573e00235d685415185ec29c89b4bd7361f938711e
        - quay.io/kubevirt/mtq-operator@sha256:ef5c4bb1fa72f9e6e75d5585ca352f7563720389e6c9876aa8b07d12d1be6194
        - quay.io/kubevirt/cdi-cloner@sha256:9d31b14f23259398c5bac636f5ead13ad0afd6fe8eeab4499e8e047b4d85074f
        version: 1.10.7
      entries:
      - name: kubevirt-hyperconverged-operator.v1.10.7
        version: 1.10.7
      name: 1.10.7
    - currentCSV: kubevirt-hyperconverged-operator.v1.11.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployTektonTaskResources":true,"deployVmConsoleProxy":true},"tektonPipelines":{"namespace":"kubevirt"},"tektonTasks":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:dd6538e720f18b723b0807037ca7cc1014cba09e57b20706e10e6eb641ac6acc
          createdAt: "2024-03-06 15:11:00"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: quay.io/kubevirt/must-gather
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - quay.io/kubevirt/aaq-server@sha256:0c6b01c48030e03e99b3ff2b319658c2bf2b2001f4e3e030ba70d1143b25717a
        - quay.io/kubevirt/virt-exportserver@sha256:f8473fc824f8bac5f21c64b71cdb4227de28f9bb25c049932f1259216b874a86
        - quay.io/kubevirt/cdi-importer@sha256:7320dce01f4628363992997b592200b4804daf15a9f237cd50a4a6acab7798e8
        - quay.io/kubevirt/virt-artifacts-server@sha256:22fa3218deca38dc98fa0e905df16173a47ccec85707eb91404a367da6480817
        - quay.io/kubevirt/virt-operator@sha256:0e6d56ea0ebdf9bb832f29ab0dc4603f0f619474bb8bc02a01e0a06e9d8d6f4a
        - quay.io/kubevirt/mtq-operator@sha256:90c7608b0ba085582a65ccdc11d089049c5e6b0438d018dd9dc76eb3490383c7
        - quay.io/kubevirt/cni-default-plugins@sha256:c884d6d08f8c0db98964f1eb3877b44ade41fa106083802a9914775df17d5291
        - quay.io/kubevirt/bridge-marker@sha256:bba066e3b5ff3fb8c5e20861fe8abe51e3c9b50ad6ce3b2616af9cb5479a06d0
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:dd6538e720f18b723b0807037ca7cc1014cba09e57b20706e10e6eb641ac6acc
        - quay.io/kubevirt/virt-handler@sha256:238550bb263608ba334b44a15647cbf875482ccfb46049713c14af5d5f0cc788
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:f1bd6ee18c4021c1c94f29edfab89b49b6a4d1b800936c19dbef2d75f8202f2d
        - quay.io/kubevirt/kubemacpool@sha256:cf8daa57ae6603b776d3af512331b143fa03bc2f4b72f28420fddcf5e4156d0a
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:4ed1f37116cbaf883edd8783201d091fbbe121fcb84e7184ee531d3b1abd77ab
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - registry.k8s.io/sig-storage/livenessprobe@sha256:5baeb4a6d7d517434292758928bb33efc6397368cbb48c8a4cf29496abf4e987
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:bf17da674ed6dac95aa4fd7637173fd4e026ebb4523b66f529c0eff394a25bcc
        - quay.io/kubevirt/cdi-uploadserver@sha256:16d257e575b2bd504f148f63b57dc181fa6058d33a11ada64888183c3e4d4863
        - quay.io/kubevirt/virt-exportproxy@sha256:942c548d4848a87945eba51683002797eb4e1f5c463a1873115b2ec0e1e6e169
        - quay.io/kubevirt/ovs-cni-plugin@sha256:e16ac74343da21abb8fb668ce71e728053d00503a992dae2164b9e94a280113e
        - quay.io/kubevirt/hostpath-csi-driver@sha256:8d84b9a33e0ec3749aece7c5f9879d1e0d248e8028ab3f776528a044d387033c
        - ghcr.io/kubevirt/kubesecondarydns@sha256:e87e829380a1e576384145f78ccaa885ba1d5690d5de7d0b73d40cfb804ea24d
        - quay.io/kubevirt/mtq-controller@sha256:ae7dd07e4018c63414b021eba34aa4776a6c7d6ff4fcdf9fd2d7ca30aae98a29
        - quay.io/kubevirt/cdi-apiserver@sha256:95f6e71b36883abcb9ec463cfea2332f5490ec9eb3e4868bdf0a4da769279094
        - quay.io/kubevirt/cdi-controller@sha256:606de89e68be2ac911466e6713345932bda496c1e4f20e34ef5efa270456c9fd
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:dd2948ca8e62d17eafba40bdcaad8367a97f092fd3b8d9849b2d62aec3152a72
        - quay.io/kubevirt/aaq-operator@sha256:590d24742401eeff3fb6a9c8ed977c63442f8f0a0505509fae69aa1bf8aa84ae
        - quay.io/kubevirt/macvtap-cni@sha256:850b89343ace7c7ea6b18dd8e11964613974e9d1f7377af03854d407fb15230a
        - quay.io/kubevirt/virtio-container-disk@sha256:f8e71a3a930ce95d646d80862d5ff6f9d3ed7ef2541dd2e2acec0bce26274734
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:3fbcc32bd4e4d15bd93c96def784a229cd84cca27942bf4858b581f31c97ee02
        - quay.io/kubevirt/virt-controller@sha256:4f576e30b83db7b9f96dde77896898bde87db3517963c4227634d0c3e96bb86d
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:70efc16ab2bfec04f73ccc292064d080870659cfd47c2e0c9017ff7b6fd78863
        - quay.io/kubevirt/cdi-operator@sha256:8a27e1450c3bf8c1d8924e74470a62419d02ed2202924b082a74e2414b1fc21a
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:c53535af8a7f7e3164609838c4b191b42b2d81238d75c1b2a2b582ada62a9780
        - quay.io/kubevirt/libguestfs-tools@sha256:4971af1ee42959d366a3db6129ab684d11064e00bbd3c2b36324ad14e90cb903
        - quay.io/kubevirt/cdi-cloner@sha256:b22985af0ee7b1ce4e1dedded08c853d90d9e7209f98c5d470897bbefce2a063
        - quay.io/kubevirt/ssp-operator@sha256:d57ef05fea09c4690dbe728ebe6a39d9df1a3574dfa7d18346bf3fe66afed702
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:57573a24923e5588bca6bc337a8b2b08406c5b77583974365d2cf063c0dd5d06
        - quay.io/kubevirt/virt-launcher@sha256:4fbfd76561667e98fafc75c0b804f77df251f96471d1121d69f149cf8fb9a4d8
        - quay.io/kubevirt/kubevirt-template-validator@sha256:f6db539fe4d4b752644c6bc8a308f5a80ae09a71da5c5d535114dc4c3cf18e0a
        - quay.io/kubevirt/aaq-controller@sha256:cb2fa52ae3f6a1a99b2de77b58c5db54c43ba8ee86c6540b4844c901180ce611
        - quay.io/kubevirt/virt-api@sha256:735662c824da3d6069624d40674163d758c2256f8043cf6f02f505c3a724afa3
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:e2def4213ec0657e72eb790ae8a115511d5b8f164a62d3568d2f1bff189917e8
        - quay.io/kubevirt/mtq-lock-server@sha256:8e686107dd9990dab1a3d90ee1068541938f673bf54ea9e1308b2ed7218546b0
        - quay.io/kubevirt/hostpath-provisioner@sha256:adb1d33e95e8991dc47107926172469b5ecbeace509391ce021ca42d305a41d3
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:38293a1050bce9c7cb786cbe8719be1edb91a922a6c583eedb61e56729ed4c72
        - quay.io/kubevirt/cdi-uploadproxy@sha256:b89104ef5513d63cab43b34ac52f752c3d935cb986d5527d920c2c31d3b6f239
        version: 1.11.0
      entries:
      - name: kubevirt-hyperconverged-operator.v1.11.0
        version: 1.11.0
      name: 1.11.0
    - currentCSV: kubevirt-hyperconverged-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:7afeef450521fb7bbc918e874d9297dc3762cdeb28db0bef96d66d48b8752a9d
          createdAt: "2024-08-07 15:33:27"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: quay.io/kubevirt/must-gather
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:7afeef450521fb7bbc918e874d9297dc3762cdeb28db0bef96d66d48b8752a9d
        - quay.io/kubevirt/virt-artifacts-server@sha256:cc3e1e17bc3fa2a895f934210c9106bee01c3f9bf001f988d920eb58628e347b
        - quay.io/kubevirt/virt-operator@sha256:1ecc62c25f869e17027c930150d5c9b52b72d7e1fb434376bc0be0d886952b2e
        - quay.io/kubevirt/aaq-operator@sha256:002513b78dfd795f5195534b651eace136ead3d73c8ac8cae145607e79c8a8ba
        - quay.io/kubevirt/hostpath-csi-driver@sha256:b7ccc469ad7e640fed2e98b50eae57a64e81c7b9862a799b7a9892d14ade8937
        - quay.io/kubevirt/virt-controller@sha256:3471b00eb9f4774ae33867c7d0b83e1d26a065a2bd19d305b5d3946ab665b5e9
        - quay.io/kubevirt/aaq-controller@sha256:170abd7a75397ddce07ba0d64886ead25183c49cacbccf0d191a4dd37ca013fd
        - quay.io/kubevirt/ssp-operator@sha256:a15a53d0f3ef530fe469acf3b7673486011605d0021d4dbe29af94096b3b65ee
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:413a6c229ebb608868e8b2c7c06b251c6b1d43c07af12a76b893292cbed8bef9
        - quay.io/kubevirt/ovs-cni-plugin@sha256:e16ac74343da21abb8fb668ce71e728053d00503a992dae2164b9e94a280113e
        - quay.io/kubevirt/kubevirt-template-validator@sha256:8f7c389a7cd7ef34f2edc84245bc1d0ebcf13e86da5e0b8893f56733c482b10d
        - quay.io/kubevirt/hostpath-provisioner@sha256:2b93b6c4ce0307a8f0dbdae2226da4acb1bae781788fbde7f0c32d6ad6421aa0
        - quay.io/kubevirt/mtq-lock-server@sha256:8b5189518ac6f68161d6a259e2a6f70e7902855ea1510aa260e79093520a7803
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac
        - quay.io/kubevirt/libguestfs-tools@sha256:0de3de89720d4e5433a33944f00dac590daed6ab12e4d26a657ef082d0573747
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:e2def4213ec0657e72eb790ae8a115511d5b8f164a62d3568d2f1bff189917e8
        - quay.io/kubevirt/mtq-operator@sha256:859524721ae1a3b1e5dd6e29036b4c80acff13f74a4ee6769d867283e9ccbba5
        - registry.k8s.io/sig-storage/livenessprobe@sha256:5baeb4a6d7d517434292758928bb33efc6397368cbb48c8a4cf29496abf4e987
        - quay.io/kubevirt/virtio-container-disk@sha256:f8e71a3a930ce95d646d80862d5ff6f9d3ed7ef2541dd2e2acec0bce26274734
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:935475c2850466aa5ac57e4de627fb177515cb2c402a95842ead095d82b6df5f
        - quay.io/kubevirt/mtq-controller@sha256:04b38720b65479046270c8132cff5ab0f2dabc9cae91de1cc95fecf167efc356
        - quay.io/kubevirt/virt-exportserver@sha256:acac33a19528592cc3ef0cd3a25fa4c72bcf992e4c2c247437007fcb9de0aa6d
        - quay.io/kubevirt/cdi-uploadproxy@sha256:86285df5cf52533732ddf4fdb74674cbce877ba69833cb8ed058c2a157cb5573
        - quay.io/kubevirt/cdi-cloner@sha256:05da8df0b64024f9e2a48a1a54f06c821d79a3efe7828fad2c4fda6a01d0cd86
        - quay.io/kubevirt/cni-default-plugins@sha256:0c354fa9d695b8cab97b459e8afea2f7662407a987e83f6f6f1a8af4b45726be
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:3c70cb833819d8346e5531b10ac516283bab7098e782708768a6fe00ae0ddfac
        - quay.io/kubevirt/cdi-operator@sha256:f68e4f2a082fc72d92bc1df0340285af03cd12d8c7721b38c619d1de0f4c175a
        - ghcr.io/kubevirt/kubesecondarydns@sha256:6268d84154e2483fbce8c1adacbdaf6f0839117b2d48d9fa4687cc8f76bd5130
        - quay.io/kubevirt/virt-api@sha256:aeeae94c9dd3b352ff2fbcdba77a6b084fb4ebe0fed35fe7af9481e9e2b1f24c
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:83b460502671fb4f34116363a1a39b2ddfc9d14a920ee0a6413bfc3bd0580404
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - quay.io/kubevirt/cdi-apiserver@sha256:4ba004f57e7d589ddb5372a9729360460a97effecc6e5ecaccd7fbf16c6fcf13
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:d51821f001c959a79943cb22f5ea47589d7cf5c032e11ec1d571e8db7bf29b78
        - quay.io/kubevirt/virt-exportproxy@sha256:dd3b5c12314fa76b88d55d978210de81562fb888351748ea3997551c8c96a877
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:c4b6b02737bc24906fcce57fe6626d1a36cb2b91baa971af2a5e5a919093c34e
        - quay.io/kubevirt/cdi-importer@sha256:de5b89c879a531165482d540deffdc2dec0c188fd0f7b5fbd49c0298cf9a5b1d
        - quay.io/kubevirt/kubemacpool@sha256:20e156be33e6d3692c456081acbb91a4349f94de448f7d1f1cddd0228931b31f
        - quay.io/kubevirt/virt-launcher@sha256:cd37543ed2bc6baadbcbffa6951d740e4c65dc0ea366dec0a8593193c384e6de
        - quay.io/kubevirt/virt-handler@sha256:2a7d96da4de3ea51339b3a047abb4b05e088e3d1080110b42a85ea5baff4a0cc
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:246b3819cfa3833f9a8cb01e8cb859aa61a1924a85ea4a112e43e8414f32c429
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:3fbcc32bd4e4d15bd93c96def784a229cd84cca27942bf4858b581f31c97ee02
        - quay.io/kubevirt/cdi-controller@sha256:392ead2fdc5050c584765ebf99706cc5c0ab9c2b424b692ad35005e77d110b58
        - quay.io/kubevirt/cdi-uploadserver@sha256:d7fb653bb33dade84583fff6048092346639532d6675aaf9f88070171cb0fe7c
        - quay.io/kubevirt/bridge-marker@sha256:bba066e3b5ff3fb8c5e20861fe8abe51e3c9b50ad6ce3b2616af9cb5479a06d0
        - quay.io/kubevirt/aaq-server@sha256:9db6ca9147b1d75ba4d6bb1326c4ec289327eb9f8aadbcd951fcc689e584cd0d
        - quay.io/kubevirt/macvtap-cni@sha256:850b89343ace7c7ea6b18dd8e11964613974e9d1f7377af03854d407fb15230a
        version: 1.12.0
      entries:
      - name: kubevirt-hyperconverged-operator.v1.12.0
        version: 1.12.0
      - name: kubevirt-hyperconverged-operator.v1.11.1
        version: 1.11.1
      - name: kubevirt-hyperconverged-operator.v1.10.0
        version: 1.10.0
      - name: kubevirt-hyperconverged-operator.v1.9.0
        version: 1.9.0
      - name: kubevirt-hyperconverged-operator.v1.8.2
        version: 1.8.2
      - name: kubevirt-hyperconverged-operator.v1.8.1
        version: 1.8.1
      - name: kubevirt-hyperconverged-operator.v1.8.0
        version: 1.8.0
      - name: kubevirt-hyperconverged-operator.v1.7.0
        version: 1.7.0
      - name: kubevirt-hyperconverged-operator.v1.6.0
        version: 1.6.0
      - name: kubevirt-hyperconverged-operator.v1.5.2
        version: 1.5.2
      - name: kubevirt-hyperconverged-operator.v1.5.1
        version: 1.5.1
      - name: kubevirt-hyperconverged-operator.v1.5.0
        version: 1.5.0
      - name: kubevirt-hyperconverged-operator.v1.4.4
        version: 1.4.4
      - name: kubevirt-hyperconverged-operator.v1.4.3
        version: 1.4.3
      - name: kubevirt-hyperconverged-operator.v1.4.2
        version: 1.4.2
      - name: kubevirt-hyperconverged-operator.v1.4.1
        version: 1.4.1
      - name: kubevirt-hyperconverged-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: community-kubevirt-hyperconverged
    provider:
      name: KubeVirt project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: StreamNative
      provider-url: https://streamnative.io
    name: sn-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
        - docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21
        - docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      name: alpha
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
        - docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21
        - docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      name: beta
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21
        - quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
        - docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      name: stable
    defaultChannel: alpha
    packageName: sn-operator
    provider:
      name: StreamNative
      url: https://streamnative.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzo-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzo-operator.v2.0.101
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzo.cambridgesemantics.com/v1",
                "kind": "Anzo",
                "metadata": {
                  "name": "agent01"
                },
                "spec": {
                  "nodeConfig": {
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "app": "anzo"
                        }
                      },
                      "serviceName": "anzo-agent01",
                      "template": {
                        "metadata": {
                          "labels": {
                            "app": "anzo"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo@sha256:d1c2d54ca205d669c5ea1bb6aae105afe2385cc5695cecd78423ffa1d98929f8",
                              "name": "anzo",
                              "resources": {
                                "limits": {
                                  "cpu": "4000m",
                                  "memory": "12Gi"
                                },
                                "requests": {
                                  "cpu": "4000m",
                                  "memory": "12Gi"
                                }
                              }
                            }
                          ],
                          "serviceAccountName": "anzo-operator"
                        }
                      }
                    }
                  },
                  "role": "AnzoAgent"
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-operator@sha256:563e964f5771b9bfd860ea993b6e0936a54504908f5b2d1dff6f99ade74bb7a0
          createdAt: "2021-08-02"
          description: kubernetes operator for Anzo
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzo-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzo-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Anzo is the Schema for the anzos API
            displayName: Anzo Operator
            kind: Anzo
            name: anzos.anzo.cambridgesemantics.com
            version: v1
        description: |-
          The Anzo Operator provides the way to install and configure an anzo agent setup on Red Hat K8S environment.
          Currently, this is supported only through when deployed as an Anzo Agent as part of an Anzo Unstructured deployment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v2.0.0/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers through [ Cambridge Semantics Customer Center ]( https://customercenter.cambridgesemantics.com/ ).
        displayName: Anzo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        links:
        - name: Anzo Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v2.0.0/README_openshift.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-operator@sha256:563e964f5771b9bfd860ea993b6e0936a54504908f5b2d1dff6f99ade74bb7a0
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo@sha256:d1c2d54ca205d669c5ea1bb6aae105afe2385cc5695cecd78423ffa1d98929f8
        version: 2.0.0
      entries:
      - name: anzo-operator.v2.0.101
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anzo-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: odr-hub-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odr-hub-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east"
                },
                "spec": {
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east1"
                },
                "spec": {
                  "cidrs": [
                    "192.168.39.45/32"
                  ],
                  "clusterFence": "ManuallyFenced",
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east1"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPlacementControl",
                "metadata": {
                  "name": "drplacementcontrol-sample",
                  "namespace": "application-namespace"
                },
                "spec": {
                  "drPolicyRef": {
                    "name": "drpolicy-sample"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "excludedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "includedResources": [
                          "*"
                        ]
                      }
                    ]
                  },
                  "placementRef": {
                    "kind": "PlacementRule",
                    "name": "application-placement-rule"
                  },
                  "preferredCluster": "east",
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east",
                    "west"
                  ],
                  "replicationClassSelector": {
                    "matchLabels": {
                      "class": "ramen"
                    }
                  },
                  "schedulingInterval": "1h"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "metrodr-drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east1",
                    "east2"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DRCluster is the Schema for the drclusters API
            displayName: DRCluster
            kind: DRCluster
            name: drclusters.ramendr.openshift.io
            version: v1alpha1
          - description: DRPlacementControl is the Schema for the drplacementcontrols
              API
            displayName: DRPlacement Control
            kind: DRPlacementControl
            name: drplacementcontrols.ramendr.openshift.io
            version: v1alpha1
          - description: DRPolicy is the Schema for the drpolicies API
            displayName: DRPolicy
            kind: DRPolicy
            name: drpolicies.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Hub is a disaster-recovery orchestrator for stateful
          applications. It operates from an Advanced Cluster Management (ACM) hub
          cluster to orchestrate the recovery of application state, and scheduling
          of ACM PlacementRule for disaster recovery operations.
        displayName: Openshift DR Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:769acf15d016d4992793c9c527d7c268d2a67502dd7e0fadc240e4896398195b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 4.15.20-rhodf
      entries:
      - name: odr-hub-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odr-hub-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odr-hub-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odr-hub-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odr-hub-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odr-hub-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odr-hub-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odr-hub-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odr-hub-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odr-hub-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odr-hub-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odr-hub-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odr-hub-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odr-hub-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odr-hub-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odr-hub-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odr-hub-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odr-hub-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odr-hub-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odr-hub-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odr-hub-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east"
                },
                "spec": {
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east1"
                },
                "spec": {
                  "cidrs": [
                    "192.168.39.45/32"
                  ],
                  "clusterFence": "ManuallyFenced",
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east1"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPlacementControl",
                "metadata": {
                  "name": "drplacementcontrol-sample",
                  "namespace": "application-namespace"
                },
                "spec": {
                  "drPolicyRef": {
                    "name": "drpolicy-sample"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "excludedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "includedResources": [
                          "*"
                        ]
                      }
                    ]
                  },
                  "placementRef": {
                    "kind": "PlacementRule",
                    "name": "application-placement-rule"
                  },
                  "preferredCluster": "east",
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east",
                    "west"
                  ],
                  "replicationClassSelector": {
                    "matchLabels": {
                      "class": "ramen-replication-class"
                    }
                  },
                  "schedulingInterval": "1h",
                  "volumeSnapshotClassSelector": {
                    "matchLabels": {
                      "class": "ramen-snapshot-class"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "metrodr-drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east1",
                    "east2"
                  ],
                  "replicationClassSelector": {},
                  "schedulingInterval": {},
                  "volumeSnapshotClassSelector": {}
                }
              }
            ]
          capabilities: Deep Insights
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DRCluster is the Schema for the drclusters API
            displayName: DRCluster
            kind: DRCluster
            name: drclusters.ramendr.openshift.io
            version: v1alpha1
          - description: DRPlacementControl is the Schema for the drplacementcontrols
              API
            displayName: DRPlacement Control
            kind: DRPlacementControl
            name: drplacementcontrols.ramendr.openshift.io
            version: v1alpha1
          - description: DRPolicy is the Schema for the drpolicies API
            displayName: DRPolicy
            kind: DRPolicy
            name: drpolicies.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Hub is a disaster-recovery orchestrator for stateful
          applications. It operates from an Advanced Cluster Management (ACM) hub
          cluster to orchestrate the recovery of application state, and scheduling
          of ACM PlacementRule for disaster recovery operations.
        displayName: Openshift DR Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:af262b0890300183db1766bc392017b0bda723a9ba841dc25e3d928aa413db9a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        version: 4.16.16-rhodf
      entries:
      - name: odr-hub-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odr-hub-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odr-hub-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odr-hub-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odr-hub-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odr-hub-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odr-hub-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odr-hub-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odr-hub-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odr-hub-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odr-hub-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odr-hub-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odr-hub-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odr-hub-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odr-hub-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odr-hub-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odr-hub-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odr-hub-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: container-security-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: container-security-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secscan.quay.redhat.com/v1alpha1",
                "kind": "ImageManifestVuln",
                "metadata": {
                  "selfLink": "/apis/secscan.quay.redhat.com/v1alpha1/namespaces/openshift-cluster-version/imagemanifestvulns/sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18",
                  "resourceVersion": "14565395",
                  "name": "sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18",
                  "uid": "3f16a188-e69a-406c-aa06-df340195409c",
                  "creationTimestamp": "2021-03-17T12:07:16Z",
                  "generation": 2,
                  "managedFields": [
                    {
                      "apiVersion": "secscan.quay.redhat.com/v1alpha1",
                      "fieldsType": "FieldsV1",
                      "fieldsV1": {
                        "f:metadata": {
                          "f:labels": {
                            ".": {},
                            "f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": {}
                          }
                        },
                        "f:spec": {
                          ".": {},
                          "f:features": {},
                          "f:image": {},
                          "f:manifest": {}
                        },
                        "f:status": {
                          ".": {},
                          "f:affectedPods": {
                            ".": {},
                            "f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": {}
                          },
                          "f:fixableCount": {},
                          "f:highCount": {},
                          "f:highestSeverity": {},
                          "f:lastUpdate": {},
                          "f:lowCount": {},
                          "f:mediumCount": {}
                        }
                      },
                      "manager": "security-labeller",
                      "operation": "Update",
                      "time": "2021-03-17T13:07:20Z"
                    }
                  ],
                  "namespace": "openshift-cluster-version",
                  "labels": {
                    "openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": "true"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:447e05859134a0455ca55f8aae16c8a27ce06af4c18c191b41f6339df553f1bf
          createdAt: 2021-11-02 11:11 UTC
          description: Identify image vulnerabilities in Kubernetes pods
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Quay Container Security Operator (CSO) brings Quay and Clair
          metadata to Kubernetes / OpenShift. Starting with vulnerability information
          the scope will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Quay Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Quay Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:447e05859134a0455ca55f8aae16c8a27ce06af4c18c191b41f6339df553f1bf
        version: 3.4.7
      entries:
      - name: container-security-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: container-security-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"secscan.quay.redhat.com/v1alpha1\",\n
            \   \"kind\": \"ImageManifestVuln\",\n    \"metadata\": {\n      \"selfLink\":
            \"/apis/secscan.quay.redhat.com/v1alpha1/namespaces/openshift-cluster-version/imagemanifestvulns/sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18\",\n
            \     \"resourceVersion\": \"14565395\",\n      \"name\": \"sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18\",\n
            \     \"uid\": \"3f16a188-e69a-406c-aa06-df340195409c\",\n      \"creationTimestamp\":
            \"2021-03-17T12:07:16Z\",\n      \"generation\": 2,\n      \"managedFields\":
            [\n        {\n          \"apiVersion\": \"secscan.quay.redhat.com/v1alpha1\",\n
            \         \"fieldsType\": \"FieldsV1\",\n          \"fieldsV1\": {\n            \"f:metadata\":
            {\n              \"f:labels\": {\n                \".\": {},\n                \"f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            {}\n              }\n            },\n            \"f:spec\": {\n              \".\":
            {},\n              \"f:features\": {},\n              \"f:image\": {},\n
            \             \"f:manifest\": {}\n            },\n            \"f:status\":
            {\n              \".\": {},\n              \"f:affectedPods\": {\n                \".\":
            {},\n                \"f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            {}\n              },\n              \"f:fixableCount\": {},\n              \"f:highCount\":
            {},\n              \"f:highestSeverity\": {},\n              \"f:lastUpdate\":
            {},\n              \"f:lowCount\": {},\n              \"f:mediumCount\":
            {}\n            }\n          },\n          \"manager\": \"security-labeller\",\n
            \         \"operation\": \"Update\",\n          \"time\": \"2021-03-17T13:07:20Z\"\n
            \       }\n      ],\n      \"namespace\": \"openshift-cluster-version\",\n
            \     \"labels\": {\n        \"openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            \"true\"\n      }\n    },\n    \"spec\": {\n      \"features\": [\n        {\n
            \         \"name\": \"musl\",\n          \"namespaceName\": \"alpine:v3.14\",\n
            \         \"version\": \"1.2.2-r3\",\n          \"versionformat\": \"dpkg\",\n
            \         \"vulnerabilities\": [\n            {\n              \"fixedby\":
            \"1.2.2_pre2-r0\",\n              \"link\": \"https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28928\",\n
            \             \"metadata\": \"{\\\"NVD\\\": {\\\"CVSSv3\\\": {\\\"Vectors\\\":
            \\\"CVSS:3.0/AV:L/AC:L/PR:L/UI:N/S:U/C:N/I:N/A:H\\\", \\\"ImpactScore\\\":
            3.6, \\\"Score\\\": 5.5, \\\"ExploitabilityScore\\\": 1.8}, \\\"CVSSv2\\\":
            {\\\"Score\\\": 2.1, \\\"PublishedDateTime\\\": \\\"2020-11-24T18:15Z\\\",
            \\\"Vectors\\\": \\\"AV:L/AC:L/Au:N/C:N/I:N/A:P\\\"}}}\",\n              \"name\":
            \"CVE-2020-28928\",\n              \"namespaceName\": \"alpine:v3.14\",\n
            \             \"severity\": \"Low\"\n            }\n          ]\n        }\n
            \     ],\n      \"image\": \"quay.io/harishg/pruned-redhat-operators-index\",\n
            \     \"manifest\": \"sha256:19631c86642609254777ff59fc9e128aa5d0a94a7dbb504ef1fcfb2bbc6e486e\"\n
            \   }\n  }\n]    "
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:6b3581e62550d6a615d36eee6eaf4c89b9c35ba53cab8433bd93c68393a9d115
          createdAt: 2021-09-10 23:09 UTC
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.4.x <3.5.7'
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Quay Container Security Operator (CSO) brings Quay and Clair
          metadata to Kubernetes / OpenShift. Starting with vulnerability information
          the scope will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Quay Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Quay Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:6b3581e62550d6a615d36eee6eaf4c89b9c35ba53cab8433bd93c68393a9d115
        version: 3.5.7
      entries:
      - name: container-security-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: container-security-operator.v3.10.15
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:de7c66a0004f3c0643c09d4c3bbb4a0f1f8f2cbf264250802cdc58028fc72b50
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.10.15
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:de7c66a0004f3c0643c09d4c3bbb4a0f1f8f2cbf264250802cdc58028fc72b50
        version: 3.10.15
      entries:
      - name: container-security-operator.v3.10.15
        version: 3.10.15
      - name: container-security-operator.v3.10.14
        version: 3.10.14
      - name: container-security-operator.v3.10.13
        version: 3.10.13
      - name: container-security-operator.v3.10.12
        version: 3.10.12
      - name: container-security-operator.v3.10.11
        version: 3.10.11
      - name: container-security-operator.v3.10.10
        version: 3.10.10
      - name: container-security-operator.v3.10.9
        version: 3.10.9
      - name: container-security-operator.v3.10.8
        version: 3.10.8
      - name: container-security-operator.v3.10.7
        version: 3.10.7
      - name: container-security-operator.v3.10.6
        version: 3.10.6
      - name: container-security-operator.v3.10.5
        version: 3.10.5
      - name: container-security-operator.v3.10.4
        version: 3.10.4
      - name: container-security-operator.v3.10.3
        version: 3.10.3
      - name: container-security-operator.v3.10.2
        version: 3.10.2
      - name: container-security-operator.v3.10.1
        version: 3.10.1
      - name: container-security-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: container-security-operator.v3.11.13
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:3914389fd53c371627921298c5dc3564999d92bc720f27a240f74c034ebbf5ec
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.11.13
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:3914389fd53c371627921298c5dc3564999d92bc720f27a240f74c034ebbf5ec
        version: 3.11.13
      entries:
      - name: container-security-operator.v3.11.13
        version: 3.11.13
      - name: container-security-operator.v3.11.12
        version: 3.11.12
      - name: container-security-operator.v3.11.11
        version: 3.11.11
      - name: container-security-operator.v3.11.10
        version: 3.11.10
      - name: container-security-operator.v3.11.9
        version: 3.11.9
      - name: container-security-operator.v3.11.8
        version: 3.11.8
      - name: container-security-operator.v3.11.7
        version: 3.11.7
      - name: container-security-operator.v3.11.6
        version: 3.11.6
      - name: container-security-operator.v3.11.5
        version: 3.11.5
      - name: container-security-operator.v3.11.4
        version: 3.11.4
      - name: container-security-operator.v3.11.3
        version: 3.11.3
      - name: container-security-operator.v3.11.2
        version: 3.11.2
      - name: container-security-operator.v3.11.1
        version: 3.11.1
      - name: container-security-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: container-security-operator.v3.12.12
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:1a15328f0b2c64d3a857f2f155ad4def095027e809610de98545e303eec63225
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.12.12
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:1a15328f0b2c64d3a857f2f155ad4def095027e809610de98545e303eec63225
        version: 3.12.12
      entries:
      - name: container-security-operator.v3.12.12
        version: 3.12.12
      - name: container-security-operator.v3.12.11
        version: 3.12.11
      - name: container-security-operator.v3.12.10
        version: 3.12.10
      - name: container-security-operator.v3.12.9
        version: 3.12.9
      - name: container-security-operator.v3.12.8
        version: 3.12.8
      - name: container-security-operator.v3.12.7
        version: 3.12.7
      - name: container-security-operator.v3.12.6
        version: 3.12.6
      - name: container-security-operator.v3.12.5
        version: 3.12.5
      - name: container-security-operator.v3.12.4
        version: 3.12.4
      - name: container-security-operator.v3.12.3
        version: 3.12.3
      - name: container-security-operator.v3.12.2
        version: 3.12.2
      - name: container-security-operator.v3.12.1
        version: 3.12.1
      - name: container-security-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: container-security-operator.v3.13.8
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:ba50621739581da9a661c2c2e355849c328cfa33b7311b06abb148e1183a6321
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.13.8
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:ba50621739581da9a661c2c2e355849c328cfa33b7311b06abb148e1183a6321
        version: 3.13.8
      entries:
      - name: container-security-operator.v3.13.8
        version: 3.13.8
      - name: container-security-operator.v3.13.7
        version: 3.13.7
      - name: container-security-operator.v3.13.6
        version: 3.13.6
      - name: container-security-operator.v3.13.5
        version: 3.13.5
      - name: container-security-operator.v3.13.4
        version: 3.13.4
      - name: container-security-operator.v3.13.3
        version: 3.13.3
      - name: container-security-operator.v3.13.2
        version: 3.13.2
      - name: container-security-operator.v3.13.1
        version: 3.13.1
      - name: container-security-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: container-security-operator.v3.14.5
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:c42487ccf098df3ec6dbeb5d409346bdd30e92781cf3ffea5fdb6dd0b196d0cf
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.14.5
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:c42487ccf098df3ec6dbeb5d409346bdd30e92781cf3ffea5fdb6dd0b196d0cf
        version: 3.14.5
      entries:
      - name: container-security-operator.v3.14.5
        version: 3.14.5
      - name: container-security-operator.v3.14.4
        version: 3.14.4
      - name: container-security-operator.v3.14.3
        version: 3.14.3
      - name: container-security-operator.v3.14.2
        version: 3.14.2
      - name: container-security-operator.v3.14.1
        version: 3.14.1
      - name: container-security-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: container-security-operator.v3.15.2
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:81e929ec095f791a83310ee7fcda239d5c348ba69a3cf5adcf1b3a77c349ce77
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.15.2
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:81e929ec095f791a83310ee7fcda239d5c348ba69a3cf5adcf1b3a77c349ce77
        version: 3.15.2
      entries:
      - name: container-security-operator.v3.15.2
        version: 3.15.2
      - name: container-security-operator.v3.15.1
        version: 3.15.1
      - name: container-security-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: container-security-operator.v3.6.10
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:641b80df8d7102d49a5dfd5b99245669f60393880a6f1bc8ba8450b88a640d96
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.6.10
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:641b80df8d7102d49a5dfd5b99245669f60393880a6f1bc8ba8450b88a640d96
        version: 3.6.10
      entries:
      - name: container-security-operator.v3.6.10
        version: 3.6.10
      - name: container-security-operator.v3.6.9
        version: 3.6.9
      - name: container-security-operator.v3.6.8
        version: 3.6.8
      - name: container-security-operator.v3.6.7
        version: 3.6.7
      - name: container-security-operator.v3.6.6
        version: 3.6.6
      - name: container-security-operator.v3.6.5
        version: 3.6.5
      - name: container-security-operator.v3.6.4
        version: 3.6.4
      - name: container-security-operator.v3.6.2
        version: 3.6.2
      - name: container-security-operator.v3.6.1
        version: 3.6.1
      - name: container-security-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: container-security-operator.v3.7.14
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:10736a06beb503b4eb73260a87a2cde9bd83353374ee0ac9e8dd0a6260cf2665
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.7.14
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:10736a06beb503b4eb73260a87a2cde9bd83353374ee0ac9e8dd0a6260cf2665
        version: 3.7.14
      entries:
      - name: container-security-operator.v3.7.14
        version: 3.7.14
      - name: container-security-operator.v3.7.13
        version: 3.7.13
      - name: container-security-operator.v3.7.12
        version: 3.7.12
      - name: container-security-operator.v3.7.11
        version: 3.7.11
      - name: container-security-operator.v3.7.10
        version: 3.7.10
      - name: container-security-operator.v3.7.9
        version: 3.7.9
      - name: container-security-operator.v3.7.8
        version: 3.7.8
      - name: container-security-operator.v3.7.7
        version: 3.7.7
      - name: container-security-operator.v3.7.6
        version: 3.7.6
      - name: container-security-operator.v3.7.5
        version: 3.7.5
      - name: container-security-operator.v3.7.4
        version: 3.7.4
      - name: container-security-operator.v3.7.3
        version: 3.7.3
      - name: container-security-operator.v3.7.2
        version: 3.7.2
      - name: container-security-operator.v3.7.1
        version: 3.7.1
      - name: container-security-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: container-security-operator.v3.8.15
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:108bd033543beb7cce077130c96f992c202f7a258229c239795f8f19e5fae5e5
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.8.15
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:108bd033543beb7cce077130c96f992c202f7a258229c239795f8f19e5fae5e5
        version: 3.8.15
      entries:
      - name: container-security-operator.v3.8.15
        version: 3.8.15
      - name: container-security-operator.v3.8.14
        version: 3.8.14
      - name: container-security-operator.v3.8.13
        version: 3.8.13
      - name: container-security-operator.v3.8.12
        version: 3.8.12
      - name: container-security-operator.v3.8.11
        version: 3.8.11
      - name: container-security-operator.v3.8.10
        version: 3.8.10
      - name: container-security-operator.v3.8.9
        version: 3.8.9
      - name: container-security-operator.v3.8.8
        version: 3.8.8
      - name: container-security-operator.v3.8.7
        version: 3.8.7
      - name: container-security-operator.v3.8.6
        version: 3.8.6
      - name: container-security-operator.v3.8.5
        version: 3.8.5
      - name: container-security-operator.v3.8.4
        version: 3.8.4
      - name: container-security-operator.v3.8.3
        version: 3.8.3
      - name: container-security-operator.v3.8.2
        version: 3.8.2
      - name: container-security-operator.v3.8.1
        version: 3.8.1
      - name: container-security-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: container-security-operator.v3.9.16
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:df8e1d246273d21049c6bb8169a2e5f3fc8d69c85c564359abea7364969c6acc
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.9.16
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:df8e1d246273d21049c6bb8169a2e5f3fc8d69c85c564359abea7364969c6acc
        version: 3.9.16
      entries:
      - name: container-security-operator.v3.9.16
        version: 3.9.16
      - name: container-security-operator.v3.9.15
        version: 3.9.15
      - name: container-security-operator.v3.9.14
        version: 3.9.14
      - name: container-security-operator.v3.9.13
        version: 3.9.13
      - name: container-security-operator.v3.9.12
        version: 3.9.12
      - name: container-security-operator.v3.9.11
        version: 3.9.11
      - name: container-security-operator.v3.9.10
        version: 3.9.10
      - name: container-security-operator.v3.9.9
        version: 3.9.9
      - name: container-security-operator.v3.9.8
        version: 3.9.8
      - name: container-security-operator.v3.9.7
        version: 3.9.7
      - name: container-security-operator.v3.9.6
        version: 3.9.6
      - name: container-security-operator.v3.9.5
        version: 3.9.5
      - name: container-security-operator.v3.9.4
        version: 3.9.4
      - name: container-security-operator.v3.9.3
        version: 3.9.3
      - name: container-security-operator.v3.9.2
        version: 3.9.2
      - name: container-security-operator.v3.9.1
        version: 3.9.1
      - name: container-security-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: container-security-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sfn-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sfn-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sfn.services.k8s.aws/v1alpha1",
                "kind": "Activity",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sfn.services.k8s.aws/v1alpha1",
                "kind": "StateMachine",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sfn-controller:1.1.1
          createdAt: "2025-09-19T18:41:33Z"
          description: AWS Step Functions controller is a service controller for managing
            Step Functions resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Activity represents the state of an AWS sfn Activity resource.
            displayName: Activity
            kind: Activity
            name: activities.sfn.services.k8s.aws
            version: v1alpha1
          - description: StateMachine represents the state of an AWS sfn StateMachine
              resource.
            displayName: StateMachine
            kind: StateMachine
            name: statemachines.sfn.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Step Functions resources in AWS from within your Kubernetes cluster.

          **About Amazon Step FUnctions**

          Step Functions is a serverless orchestration service that lets you combine AWS Lambda functions and other AWS services to build business-critical applications. Through Step Functions' graphical console, you see your application’s workflow as a series of event-driven steps.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Step Functions
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sfn
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Step Functions Developer Resources
          url: https://aws.amazon.com/step-functions/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: step functions maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sfn-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-sfn-controller.v1.1.1
        version: 1.1.1
      - name: ack-sfn-controller.v1.1.0
        version: 1.1.0
      - name: ack-sfn-controller.v1.0.29
        version: 1.0.29
      - name: ack-sfn-controller.v1.0.28
        version: 1.0.28
      - name: ack-sfn-controller.v1.0.27
        version: 1.0.27
      - name: ack-sfn-controller.v1.0.26
        version: 1.0.26
      - name: ack-sfn-controller.v1.0.25
        version: 1.0.25
      - name: ack-sfn-controller.v1.0.24
        version: 1.0.24
      - name: ack-sfn-controller.v1.0.23
        version: 1.0.23
      - name: ack-sfn-controller.v1.0.22
        version: 1.0.22
      - name: ack-sfn-controller.v1.0.21
        version: 1.0.21
      - name: ack-sfn-controller.v1.0.20
        version: 1.0.20
      - name: ack-sfn-controller.v1.0.19
        version: 1.0.19
      - name: ack-sfn-controller.v1.0.18
        version: 1.0.18
      - name: ack-sfn-controller.v1.0.17
        version: 1.0.17
      - name: ack-sfn-controller.v1.0.16
        version: 1.0.16
      - name: ack-sfn-controller.v1.0.15
        version: 1.0.15
      - name: ack-sfn-controller.v1.0.14
        version: 1.0.14
      - name: ack-sfn-controller.v1.0.13
        version: 1.0.13
      - name: ack-sfn-controller.v1.0.11
        version: 1.0.11
      - name: ack-sfn-controller.v1.0.10
        version: 1.0.10
      - name: ack-sfn-controller.v1.0.9
        version: 1.0.9
      - name: ack-sfn-controller.v1.0.8
        version: 1.0.8
      - name: ack-sfn-controller.v1.0.6
        version: 1.0.6
      - name: ack-sfn-controller.v1.0.5
        version: 1.0.5
      - name: ack-sfn-controller.v1.0.2
        version: 1.0.2
      - name: ack-sfn-controller.v1.0.1
        version: 1.0.1
      - name: ack-sfn-controller.v1.0.0
        version: 1.0.0
      - name: ack-sfn-controller.v0.1.3
        version: 0.1.3
      - name: ack-sfn-controller.v0.1.2
        version: 0.1.2
      - name: ack-sfn-controller.v0.1.1
        version: 0.1.1
      - name: ack-sfn-controller.v0.1.0
        version: 0.1.0
      - name: ack-sfn-controller.v0.0.19
        version: 0.0.19
      - name: ack-sfn-controller.v0.0.18
        version: 0.0.18
      - name: ack-sfn-controller.v0.0.17
        version: 0.0.17
      - name: ack-sfn-controller.v0.0.16
        version: 0.0.16
      - name: ack-sfn-controller.v0.0.15
        version: 0.0.15
      - name: ack-sfn-controller.v0.0.14
        version: 0.0.14
      - name: ack-sfn-controller.v0.0.13
        version: 0.0.13
      - name: ack-sfn-controller.v0.0.12
        version: 0.0.12
      - name: ack-sfn-controller.v0.0.11
        version: 0.0.11
      - name: ack-sfn-controller.v0.0.10
        version: 0.0.10
      - name: ack-sfn-controller.v0.0.9
        version: 0.0.9
      - name: ack-sfn-controller.v0.0.8
        version: 0.0.8
      - name: ack-sfn-controller.v0.0.7
        version: 0.0.7
      - name: ack-sfn-controller.v0.0.6
        version: 0.0.6
      - name: ack-sfn-controller.v0.0.5
        version: 0.0.5
      name: alpha
    defaultChannel: alpha
    packageName: ack-sfn-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: authorino-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: authorino-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/authorino-operator:v0.8.0
          createdAt: 2021-12-08T10-00-00Z
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta1
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: The operator to manage instances of Authorino
        displayName: Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: Authorino Operator
          url: https://github.com/Kuadrant/authorino-operator
        - name: Authorino
          url: https://github.com/Kuadrant/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/authorino-operator:v0.8.0
        version: 0.8.0
      entries:
      - name: authorino-operator.v0.8.0
        version: 0.8.0
      - name: authorino-operator.v0.7.0
        version: 0.7.0
      - name: authorino-operator.v0.6.0
        version: 0.6.0
      - name: authorino-operator.v0.5.0
        version: 0.5.0
      - name: authorino-operator.v0.4.1
        version: 0.4.1
      - name: authorino-operator.v0.4.0
        version: 0.4.0
      - name: authorino-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: authorino-operator.v0.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta3",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/authorino-operator:v0.16.0
          createdAt: "2024-11-25T13:22:24Z"
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta3
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: The operator to manage instances of Authorino
        displayName: Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: Authorino Operator
          url: https://github.com/Kuadrant/authorino-operator
        - name: Authorino
          url: https://github.com/Kuadrant/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/authorino-operator:v0.16.0
        - quay.io/kuadrant/authorino:v0.20.0
        version: 0.16.0
      entries:
      - name: authorino-operator.v0.16.0
        version: 0.16.0
      - name: authorino-operator.v0.15.1
        version: 0.15.1
      - name: authorino-operator.v0.15.0
        version: 0.15.0
      - name: authorino-operator.v0.14.0
        version: 0.14.0
      - name: authorino-operator.v0.13.1
        version: 0.13.1
      - name: authorino-operator.v0.13.0
        version: 0.13.0
      - name: authorino-operator.v0.12.0
        version: 0.12.0
      - name: authorino-operator.v0.11.1
        version: 0.11.1
      - name: authorino-operator.v0.10.0
        version: 0.10.0
      - name: authorino-operator.v0.9.0
        version: 0.9.0
      - name: authorino-operator.v0.8.0
        version: 0.8.0
      - name: authorino-operator.v0.7.0
        version: 0.7.0
      - name: authorino-operator.v0.6.0
        version: 0.6.0
      - name: authorino-operator.v0.5.0
        version: 0.5.0
      - name: authorino-operator.v0.4.1
        version: 0.4.1
      - name: authorino-operator.v0.4.0
        version: 0.4.0
      - name: authorino-operator.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: authorino-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Eclipse Foundation
      provider-url: ""
    name: eclipse-che
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eclipse-che.v7.110.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "org.eclipse.che/v1",
                "kind": "CheCluster",
                "metadata": {
                  "name": "eclipse-che",
                  "namespace": "eclipse-che"
                },
                "spec": {
                  "auth": {
                    "identityProviderURL": "",
                    "oAuthClientName": "",
                    "oAuthSecret": ""
                  },
                  "database": {
                    "externalDb": false
                  },
                  "k8s": {
                    "ingressDomain": null,
                    "tlsSecretName": null
                  },
                  "metrics": {
                    "enable": true
                  },
                  "server": {
                    "externalDevfileRegistries": [
                      {
                        "url": "https://registry.devfile.io"
                      }
                    ],
                    "externalDevfileRegistry": true,
                    "externalPluginRegistry": true,
                    "workspaceNamespaceDefault": "<username>-che"
                  },
                  "storage": {
                    "pvcStrategy": "common"
                  }
                }
              },
              {
                "apiVersion": "org.eclipse.che/v2",
                "kind": "CheCluster",
                "metadata": {
                  "name": "eclipse-che",
                  "namespace": "eclipse-che"
                },
                "spec": {
                  "components": {
                    "devfileRegistry": {
                      "disableInternalRegistry": true,
                      "externalDevfileRegistries": [
                        {
                          "url": "https://registry.devfile.io"
                        }
                      ]
                    },
                    "pluginRegistry": {
                      "disableInternalRegistry": true
                    }
                  },
                  "containerRegistry": {},
                  "devEnvironments": {},
                  "gitServices": {},
                  "networking": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/eclipse/che-operator@sha256:10c4d05c2cb848d04b97caf752c7f4d181e6973683263e45adec417342164e51
          createdAt: "2025-10-02T14:44:20Z"
          description: A Kube-native development solution that delivers portable and
            collaborative developer workspaces.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eclipse-che/che-operator
          support: Eclipse Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              The `CheCluster` custom resource allows defining and managing Eclipse Che server installation.
              Based on these settings, the  Operator automatically creates and maintains several ConfigMaps:
              `che`, `plugin-registry` that will contain the appropriate environment variables
              of the various components of the installation. These generated ConfigMaps must NOT be updated manually.
            displayName: Eclipse Che instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v2
          - description: The `CheCluster` custom resource allows defining and managing
              a Che server installation
            displayName: Eclipse Che instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v1
        description: |
          A collaborative Kubernetes-native development solution that delivers OpenShift workspaces and in-browser IDE for rapid cloud application development.
          This operator installs the Plugin and Devfile registries, Dashboard, Gateway and the Eclipse Che server, and configures these services.
          OpenShift OAuth is used directly for authentication. TLS mode is on.

          ## How to Install
          Press the **Install** button, choose the channel and the upgrade strategy, and wait for the **Installed** Operator status.
          When the operator is installed, create a new Custom Resource (CR) of Kind CheCluster (click the **Create New** button).
          The CR spec contains all defaults. You can start using Eclipse Che when the CR status is set to **Available**, and you see a URL to Eclipse Che.

          ## Defaults
          By default, the operator deploys Eclipse Che with:
          * 10Gi storage
          * Auto-generated passwords
          * Bundled Plugin and Devfile registries

          Use `oc edit checluster/eclipse-che -n eclipse-che` to update Eclipse Che default installation options.
          See more in the [Installation guide](https://www.eclipse.org/che/docs/stable/administration-guide/configuring-che/).

          ### Certificates
          Operator uses a default router certificate to secure Eclipse Che routes.
          Follow the [guide](https://www.eclipse.org/che/docs/stable/administration-guide/importing-untrusted-tls-certificates/)
          to import certificates into Eclipse Che.
        displayName: Eclipse Che
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eclipse che
        - workspaces
        - devtools
        - developer
        - ide
        - java
        links:
        - name: Product Page
          url: https://www.eclipse.org/che
        - name: Documentation
          url: https://www.eclipse.org/che/docs
        - name: Operator GitHub Repo
          url: https://github.com/eclipse-che/che-operator
        maintainers:
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Eclipse Foundation
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy@sha256:738c854322f56d63ebab75de5210abcdd5e0782ce2d30c0ecd4620f63b24694d
        - quay.io/devfile/universal-developer-image@sha256:3cf53fb9eab60ee3156e9edd4ec5af72b70e55d5ffb28f41ecb5b26e01dc9838
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:354fc75eb7a21a934381e93d03ef9d42bc2c8ae8989fdcacecfb39b863b96ced
        - quay.io/eclipse/che-operator@sha256:10c4d05c2cb848d04b97caf752c7f4d181e6973683263e45adec417342164e51
        - quay.io/eclipse/che--traefik@sha256:4cf907247939b5d20bf4eff73abd21cb413c339600dde76dbc94a874b2578a27
        - quay.io/che-incubator/che-code@sha256:cfcfbe4d665b18b11401c3c35b23b04d292c31e7f4be6d78a5e6d60fce776ed0
        - quay.io/che-incubator/che-code-server@sha256:357c95297ce2182ef0a792038ea122075fdeb6c8e835099be09e8d4f105eac2a
        - quay.io/eclipse/che-tls-secret-creator@sha256:cbb2fd8ce903d822377f2093af1aa2543672ed8c516d607a7a9dd857379dac5d
        - quay.io/che-incubator/configbump@sha256:396714521307ae9f79453c492fb8064cf940cd6b4e6183a825f1056045623874
        - quay.io/che-incubator/che-idea-dev-server@sha256:87c8b53cc295e3d7eb35b326e221a949553be3345135a4b6c1999d66f345a8b2
        - quay.io/che-incubator/che-code-sshd@sha256:7eb5381fba8878b8f252523caf70c9957b9f9e699db4564061017fb0a023cc1a
        - quay.io/che-incubator/che-idea@sha256:8aae69dc4b0c122491a75400639af0fe92b5e214c6e68ac97cda29fb58b44151
        - quay.io/eclipse/che-plugin-registry@sha256:a70a0d98505ba5aeed691fc70c3c59aed3304155fba0ab3ad70a6d0453b22093
        - quay.io/che-incubator/che-code-server@sha256:2093baab1a73e9682f5892f80582481abfbef99d744472ee6098e2d82fb2899f
        - quay.io/che-incubator/header-rewrite-proxy@sha256:bd7873b8feef35f218f54c6251ea224bea2c8bf202a328230019a0ba2941245d
        - quay.io/oauth2-proxy/oauth2-proxy@sha256:dcb6ff8dd21bf3058f6a22c6fa385fa5b897a9cd3914c88a2cc2bb0a85f8065d
        - quay.io/che-incubator/che-code@sha256:ac62a2f42ae6d6f6628620a00133afece954fcbed2df309fa3de79e4c28cc029
        - quay.io/openshift/origin-oauth-proxy@sha256:870bfe92a4663720775c0dfe5728ecbb10a17f0644eef5f57276ec135034c6a1
        - quay.io/eclipse/che-dashboard@sha256:2db3255752aebd75dceea578cd27b7e8ecb41c2290827ff7104ed77e1b61feb9
        - quay.io/eclipse/che-server@sha256:b83d536cdd3cb2b86def5599495c40870dde20c194a003736a2b47a5e5e17a06
        version: 7.110.0
      entries:
      - name: eclipse-che.v7.110.0
        version: 7.110.0
      - name: eclipse-che.v7.109.0
        version: 7.109.0
      - name: eclipse-che.v7.108.0
        version: 7.108.0
      - name: eclipse-che.v7.107.0
        version: 7.107.0
      - name: eclipse-che.v7.106.0
        version: 7.106.0
      - name: eclipse-che.v7.105.0
        version: 7.105.0
      - name: eclipse-che.v7.104.0
        version: 7.104.0
      - name: eclipse-che.v7.103.0
        version: 7.103.0
      - name: eclipse-che.v7.102.0
        version: 7.102.0
      - name: eclipse-che.v7.101.0
        version: 7.101.0
      - name: eclipse-che.v7.100.0
        version: 7.100.0
      - name: eclipse-che.v7.99.0
        version: 7.99.0
      - name: eclipse-che.v7.98.0
        version: 7.98.0
      - name: eclipse-che.v7.97.0
        version: 7.97.0
      - name: eclipse-che.v7.96.0
        version: 7.96.0
      - name: eclipse-che.v7.95.0
        version: 7.95.0
      - name: eclipse-che.v7.94.0
        version: 7.94.0
      - name: eclipse-che.v7.93.0
        version: 7.93.0
      - name: eclipse-che.v7.92.0
        version: 7.92.0
      - name: eclipse-che.v7.91.0
        version: 7.91.0
      - name: eclipse-che.v7.90.0
        version: 7.90.0
      - name: eclipse-che.v7.89.0
        version: 7.89.0
      - name: eclipse-che.v7.88.0
        version: 7.88.0
      - name: eclipse-che.v7.87.0
        version: 7.87.0
      - name: eclipse-che.v7.86.0
        version: 7.86.0
      - name: eclipse-che.v7.85.0
        version: 7.85.0
      - name: eclipse-che.v7.84.0
        version: 7.84.0
      - name: eclipse-che.v7.83.0
        version: 7.83.0
      - name: eclipse-che.v7.82.0
        version: 7.82.0
      - name: eclipse-che.v7.81.0
        version: 7.81.0
      - name: eclipse-che.v7.80.0
        version: 7.80.0
      - name: eclipse-che.v7.79.0
        version: 7.79.0
      - name: eclipse-che.v7.78.0
        version: 7.78.0
      - name: eclipse-che.v7.77.0
        version: 7.77.0
      - name: eclipse-che.v7.76.0
        version: 7.76.0
      - name: eclipse-che.v7.75.0
        version: 7.75.0
      - name: eclipse-che.v7.74.0
        version: 7.74.0
      - name: eclipse-che.v7.73.0
        version: 7.73.0
      - name: eclipse-che.v7.72.0
        version: 7.72.0
      - name: eclipse-che.v7.71.0
        version: 7.71.0
      - name: eclipse-che.v7.70.0
        version: 7.70.0
      - name: eclipse-che.v7.69.0
        version: 7.69.0
      - name: eclipse-che.v7.68.0
        version: 7.68.0
      - name: eclipse-che.v7.67.0
        version: 7.67.0
      - name: eclipse-che.v7.66.0
        version: 7.66.0
      - name: eclipse-che.v7.65.0
        version: 7.65.0
      - name: eclipse-che.v7.64.0
        version: 7.64.0
      - name: eclipse-che.v7.63.0
        version: 7.63.0
      - name: eclipse-che.v7.62.0
        version: 7.62.0
      - name: eclipse-che.v7.61.0
        version: 7.61.0
      - name: eclipse-che.v7.60.1
        version: 7.60.1
      - name: eclipse-che.v7.60.0
        version: 7.60.0
      - name: eclipse-che.v7.59.0
        version: 7.59.0
      - name: eclipse-che.v7.58.0
        version: 7.58.0
      - name: eclipse-che.v7.57.0
        version: 7.57.0
      - name: eclipse-che.v7.56.0
        version: 7.56.0
      - name: eclipse-che.v7.55.0
        version: 7.55.0
      - name: eclipse-che.v7.54.0
        version: 7.54.0
      - name: eclipse-che.v7.53.0
        version: 7.53.0
      - name: eclipse-che.v7.52.0
        version: 7.52.0
      - name: eclipse-che.v7.51.0
        version: 7.51.0
      - name: eclipse-che.v7.50.0
        version: 7.50.0
      - name: eclipse-che.v7.49.0
        version: 7.49.0
      - name: eclipse-che.v7.48.1
        version: 7.48.1
      - name: eclipse-che.v7.47.0
        version: 7.47.0
      - name: eclipse-che.v7.46.0
        version: 7.46.0
      - name: eclipse-che.v7.45.0
        version: 7.45.0
      - name: eclipse-che.v7.44.0
        version: 7.44.0
      - name: eclipse-che.v7.42.0
        version: 7.42.0
      - name: eclipse-che.v7.41.2
        version: 7.41.2
      - name: eclipse-che.v7.41.1
        version: 7.41.1
      - name: eclipse-che.v7.41.0
        version: 7.41.0
      - name: eclipse-che.v7.40.2
        version: 7.40.2
      - name: eclipse-che.v7.40.1
        version: 7.40.1
      name: stable
    defaultChannel: stable
    packageName: eclipse-che
    provider:
      name: Eclipse Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solution-Soft
      provider-url: www.solution-soft.com
    name: timemachine-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: timemachine-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tm.solution-soft.com/v1alpha1",
                "kind": "TimeMachine",
                "metadata": {
                  "name": "timemachine-sample"
                },
                "spec": {
                  "cluster_provider": "ibm",
                  "aws": {
                    "shared_volume_name": ""
                  },
                  "ibm": {
                    "persistentvolume": {
                      "storageclass": "ibmc-file-silver-gid"
                    }
                  },
                  "tm_lichost": "52.152.148.40",
                  "tm_licpass": "qwaszx",
                  "tm_licport": "57777",
                  "imagePullSecrets": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
          createdAt: "2024-01-30T10:20:30Z"
          description: Time Machine Operator enables time travel of OpenShift pods,
            without the need to change the system clock on any of the cluster nodes
          olm.skipRange: '>=0.0.4 <1.0.17'
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: SolutionSoft Systems, Inc. - support@solution-soft.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Time Machine
            displayName: Time Machine
            kind: TimeMachine
            name: timemachines.tm.solution-soft.com
            version: v1alpha1
        description: |-
          Time Machine operator enables time travel of OpenShift pods/deployments.

          To time travel a pod where your application is, you just need to install the Time Machine operator and create a Time Machine custom resource in the same namespace, without the need of changing the container images of the application you’re using, or adding containers to the pod with your application.

          Time travelling is done on the namespace level where all the (configured) target pods/deployments see the same virtual time.
        displayName: Time Machine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - time shift
        - time travel
        - virtual clocks
        - time machine
        - virtual time
        - solution-soft
        links:
        - name: Timemachine Operator
          url: https://catalog.redhat.com/software/operators/detail/5e98727a6c5dcb34dfbb19e3
        maintainers:
        - email: dtrajkovic@solution-soft.com
          name: Dragan Trajkovic
        maturity: alpha
        provider:
          name: Solution-Soft
          url: www.solution-soft.com
        relatedImages:
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
        - registry.connect.redhat.com/rhel-timemachine-container/virtual-clock-testing-tool@sha256:a81a823328b0ee795c40668c89a08bcf545c5192d9eed0146071e3721a0a8372
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0c304c11986790f446604319dfc29db494b15f5f9fa13f2c436688ec29c92dbe
        version: 1.0.17
      entries:
      - name: timemachine-operator.v1.0.17
        version: 1.0.17
      - name: timemachine-operator.v1.0.16
        version: 1.0.16
      name: alpha
    defaultChannel: alpha
    packageName: timemachine-operator
    provider:
      name: Solution-Soft
      url: www.solution-soft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: antrea.io
      provider-url: ""
    name: antrea-by-broadcom-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: antrea-by-broadcom-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.antrea.vmware.com/v1",
                "kind": "AntreaInstall",
                "metadata": {
                  "name": "antrea-install",
                  "namespace": "antrea-operator"
                },
                "spec": {
                  "antreaAgentConfig": "# FeatureGates is a map of feature names to bools that enable or disable experimental features.\nfeatureGates:\n# AllAlpha is a global toggle for alpha features. Per-feature key values override the default set by AllAlpha.\n#  AllAlpha: false\n\n# AllBeta is a global toggle for beta features. Per-feature key values override the default set by AllBeta.\n#  AllBeta: false\n\n# Enable AntreaProxy which provides ServiceLB for in-cluster Services in antrea-agent.\n# It should be enabled on Windows, otherwise NetworkPolicy will not take effect on\n# Service traffic.\n#  AntreaProxy: true\n\n# Enable EndpointSlice support in AntreaProxy. Don't enable this feature unless that EndpointSlice\n# API version v1 is supported and set as enabled in Kubernetes. If AntreaProxy is not enabled,\n# this flag will not take effect.\n#  EndpointSlice: false\n\n# Enable TopologyAwareHints in AntreaProxy. This requires AntreaProxy and EndpointSlice to be\n# enabled, otherwise this flag will not take effect.\n#  TopologyAwareHints: false\n\n# Enable traceflow which provides packet tracing feature to diagnose network issue.\n#  Traceflow: true\n\n# Enable NodePortLocal feature to make the Pods reachable externally through NodePort\n#  NodePortLocal: true\n\n# Enable Antrea ClusterNetworkPolicy feature to complement K8s NetworkPolicy for cluster admins\n# to define security policies which apply to the entire cluster, and Antrea NetworkPolicy\n# feature that supports priorities, rule actions and externalEntities in the future.\n#  AntreaPolicy: true\n\n# Enable flowexporter which exports polled conntrack connections as IPFIX flow records from each\n# agent to a configured collector.\n#  FlowExporter: false\n\n# Enable collecting and exposing NetworkPolicy statistics.\n#  NetworkPolicyStats: true\n\n# Enable controlling SNAT IPs of Pod egress traffic.\n#  Egress: true\n\n# Enable AntreaIPAM, which can allocate IP addresses from IPPools. AntreaIPAM is required by the\n# bridging mode and allocates IPs to Pods in bridging mode. It is also required to use Antrea for\n# IPAM when configuring secondary network interfaces with Multus.\n#  AntreaIPAM: false\n\n# Enable multicast traffic.\n#  Multicast: false\n\n# Enable Antrea Multi-cluster Gateway to support cross-cluster traffic.\n# This feature is supported only with encap mode.\n#  Multicluster: false\n\n# Enable support for provisioning secondary network interfaces for Pods (using\n# Pod annotations). At the moment, Antrea can only create secondary network\n# interfaces using SR-IOV VFs on baremetal Nodes.\n#  SecondaryNetwork: false\n\n# Enable managing external IPs of Services of LoadBalancer type.\n#  ServiceExternalIP: false\n\n# Enable mirroring or redirecting the traffic Pods send or receive.\n#  TrafficControl: false\n\n# Enable certificated-based authentication for IPsec.\n#  IPsecCertAuth: false\n\n# Enable collecting support bundle files with SupportBundleCollection CRD.\n#  SupportBundleCollection: false\n\n# Enable users to protect their applications by specifying how they are allowed to communicate with others, taking\n# into account application context.\n#  L7NetworkPolicy: false\n\n# Name of the OpenVSwitch bridge antrea-agent will create and use.\n# Make sure it doesn't conflict with your existing OpenVSwitch bridges.\novsBridge: \"br-int\"\n\n# Datapath type to use for the OpenVSwitch bridge created by Antrea. At the moment, the only\n# supported value is 'system', which corresponds to the kernel datapath.\n#ovsDatapathType: system\n\n# Name of the interface antrea-agent will create and use for host \u003c--\u003e pod communication.\n# Make sure it doesn't conflict with your existing interfaces.\nhostGateway: \"antrea-gw0\"\n\n# Determines how traffic is encapsulated. It has the following options:\n# encap(default):    Inter-node Pod traffic is always encapsulated and Pod to external network\n#                    traffic is SNAT'd.\n# noEncap:           Inter-node Pod traffic is not encapsulated; Pod to external network traffic is\n#                    SNAT'd if noSNAT is not set to true. Underlying network must be capable of\n#                    supporting Pod traffic across IP subnets.\n# hybrid:            noEncap if source and destination Nodes are on the same subnet, otherwise encap.\n# networkPolicyOnly: Antrea enforces NetworkPolicy only, and utilizes CNI chaining and delegates Pod\n#                    IPAM and connectivity to the primary CNI.\n#\ntrafficEncapMode: \"encap\"\n\n# Whether or not to SNAT (using the Node IP) the egress traffic from a Pod to the external network.\n# This option is for the noEncap traffic mode only, and the default value is false. In the noEncap\n# mode, if the cluster's Pod CIDR is reachable from the external network, then the Pod traffic to\n# the external network needs not be SNAT'd. In the networkPolicyOnly mode, antrea-agent never\n# performs SNAT and this option will be ignored; for other modes it must be set to false.\nnoSNAT: false\n\n# Tunnel protocols used for encapsulating traffic across Nodes. If WireGuard is enabled in trafficEncryptionMode,\n# this option will not take effect. Supported values:\n# - geneve (default)\n# - vxlan\n# - gre\n# - stt\n# Note that \"gre\" is not supported for IPv6 clusters (IPv6-only or dual-stack clusters).\ntunnelType: \"geneve\"\n\n# TunnelPort is the destination port for UDP and TCP based tunnel protocols (Geneve, VXLAN, and STT).\n# If zero, it will use the assigned IANA port for the protocol, i.e. 6081 for Geneve, 4789 for VXLAN,\n# and 7471 for STT.\ntunnelPort: 0\n\n# TunnelCsum determines whether to compute UDP encapsulation header (Geneve or VXLAN) checksums on outgoing\n# packets. For Linux kernel before Mar 2021, UDP checksum must be present to trigger GRO on the receiver for better\n# performance of Geneve and VXLAN tunnels. The issue has been fixed by\n# https://github.com/torvalds/linux/commit/89e5c58fc1e2857ccdaae506fb8bc5fed57ee063, thus computing UDP checksum is\n# no longer necessary.\n# It should only be set to true when you are using an unpatched Linux kernel and observing poor transfer performance.\ntunnelCsum: false\n\n# Determines how tunnel traffic is encrypted. Currently encryption only works with encap mode.\n# It has the following options:\n# - none (default):  Inter-node Pod traffic will not be encrypted.\n# - ipsec:           Enable IPsec (ESP) encryption for Pod traffic across Nodes. Antrea uses\n#                    Preshared Key (PSK) for IKE authentication. When IPsec tunnel is enabled,\n#                    the PSK value must be passed to Antrea Agent through an environment\n#                    variable: ANTREA_IPSEC_PSK.\n# - wireGuard:       Enable WireGuard for tunnel traffic encryption.\ntrafficEncryptionMode: \"none\"\n\n# Enable bridging mode of Pod network on Nodes, in which the Node's transport interface is connected\n# to the OVS bridge, and cross-Node/VLAN traffic of AntreaIPAM Pods (Pods whose IP addresses are\n# allocated by AntreaIPAM from IPPools) is sent to the underlay network, and forwarded/routed by the\n# underlay network.\n# This option requires the `AntreaIPAM` feature gate to be enabled. At this moment, it supports only\n# IPv4 and Linux Nodes, and can be enabled only when `ovsDatapathType` is `system`,\n# `trafficEncapMode` is `noEncap`, and `noSNAT` is true.\nenableBridgingMode: false\n\n# Disable TX checksum offloading for container network interfaces. It's supposed to be set to true when the\n# datapath doesn't support TX checksum offloading, which causes packets to be dropped due to bad checksum.\n# It affects Pods running on Linux Nodes only.\ndisableTXChecksumOffload: false\n\n# Default MTU to use for the host gateway interface and the network interface of each Pod.\n# If omitted, antrea-agent will discover the MTU of the Node's primary interface and\n# also adjust MTU to accommodate for tunnel encapsulation overhead (if applicable).\ndefaultMTU: 0\n\n# wireGuard specifies WireGuard related configurations.\nwireGuard:\n  # The port for WireGuard to receive traffic.\n  port: 51820\n\negress:\n  # exceptCIDRs is the CIDR ranges to which outbound Pod traffic will not be SNAT'd by Egresses.\n  exceptCIDRs:\n\n# ClusterIP CIDR range for Services. It's required when AntreaProxy is not enabled, and should be\n# set to the same value as the one specified by --service-cluster-ip-range for kube-apiserver. When\n# AntreaProxy is enabled, this parameter is not needed and will be ignored if provided.\nserviceCIDR: \"\"\n\n# ClusterIP CIDR range for IPv6 Services. It's required when using kube-proxy to provide IPv6 Service in a Dual-Stack\n# cluster or an IPv6 only cluster. The value should be the same as the configuration for kube-apiserver specified by\n# --service-cluster-ip-range. When AntreaProxy is enabled, this parameter is not needed.\n# No default value for this field.\nserviceCIDRv6: \"\"\n\n# The port for the antrea-agent APIServer to serve on.\n# Note that if it's set to another value, the `containerPort` of the `api` port of the\n# `antrea-agent` container must be set to the same value.\napiPort: 10350\n\n# Enable metrics exposure via Prometheus. Initializes Prometheus metrics listener.\nenablePrometheusMetrics: true\n\n# Provide the IPFIX collector address as a string with format \u003cHOST\u003e:[\u003cPORT\u003e][:\u003cPROTO\u003e].\n# HOST can either be the DNS name or the IP of the Flow Collector. For example,\n# \"flow-aggregator.flow-aggregator.svc\" can be provided as DNS name to connect\n# to the Antrea Flow Aggregator service. If IP, it can be either IPv4 or IPv6.\n# However, IPv6 address should be wrapped with [].\n# If PORT is empty, we default to 4739, the standard IPFIX port.\n# If no PROTO is given, we consider \"tls\" as default. We support \"tls\", \"tcp\" and\n# \"udp\" protocols. \"tls\" is used for securing communication between flow exporter and\n# flow aggregator.\nflowCollectorAddr: \"flow-aggregator.flow-aggregator.svc:4739:tls\"\n\n# Provide flow poll interval as a duration string. This determines how often the\n# flow exporter dumps connections from the conntrack module. Flow poll interval\n# should be greater than or equal to 1s (one second).\n# Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\", \"m\", \"h\".\nflowPollInterval: \"5s\"\n\n# Provide the active flow export timeout, which is the timeout after which a flow\n# record is sent to the collector for active flows. Thus, for flows with a continuous\n# stream of packets, a flow record will be exported to the collector once the elapsed\n# time since the last export event is equal to the value of this timeout.\n# Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\", \"m\", \"h\".\nactiveFlowExportTimeout: \"5s\"\n\n# Provide the idle flow export timeout, which is the timeout after which a flow\n# record is sent to the collector for idle flows. A flow is considered idle if no\n# packet matching this flow has been observed since the last export event.\n# Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\", \"m\", \"h\".\nidleFlowExportTimeout: \"15s\"\n\nnodePortLocal:\n# Enable NodePortLocal, a feature used to make Pods reachable using port forwarding on the host. To\n# enable this feature, you need to set \"enable\" to true, and ensure that the NodePortLocal feature\n# gate is also enabled (which is the default).\n  enable: false\n# Provide the port range used by NodePortLocal. When the NodePortLocal feature is enabled, a port\n# from that range will be assigned whenever a Pod's container defines a specific port to be exposed\n# (each container can define a list of ports as pod.spec.containers[].ports), and all Node traffic\n# directed to that port will be forwarded to the Pod.\n  portRange: \"61000-62000\"\n\n# Provide the address of Kubernetes apiserver, to override any value provided in kubeconfig or InClusterConfig.\n# Defaults to \"\". It must be a host string, a host:port pair, or a URL to the base of the apiserver.\nkubeAPIServerOverride: \"\"\n\n# Provide the address of DNS server, to override the kube-dns service. It's used to resolve hostname in FQDN policy.\n# Defaults to \"\". It must be a host string or a host:port pair of the DNS server (e.g. 10.96.0.10, 10.96.0.10:53,\n# [fd00:10:96::a]:53).\ndnsServerOverride: \"\"\n\n# Comma-separated list of Cipher Suites. If omitted, the default Go Cipher Suites will be used.\n# https://golang.org/pkg/crypto/tls/#pkg-constants\n# Note that TLS1.3 Cipher Suites cannot be added to the list. But the apiserver will always\n# prefer TLS1.3 Cipher Suites whenever possible.\ntlsCipherSuites: \"\"\n\n# TLS min version from: VersionTLS10, VersionTLS11, VersionTLS12, VersionTLS13.\ntlsMinVersion: \"\"\n\n# The name of the interface on Node which is used for tunneling or routing the traffic across Nodes.\n# If there are multiple IP addresses configured on the interface, the first one is used. The IP\n# address used for tunneling or routing traffic to remote Nodes is decided in the following order of\n# preference (from highest to lowest):\n# 1. transportInterface\n# 2. transportInterfaceCIDRs\n# 3. The Node IP\ntransportInterface: \"\"\n\nmulticast:\n# The names of the interfaces on Nodes that are used to forward multicast traffic.\n# Defaults to transport interface if not set.\n  multicastInterfaces:\n\n# The interval at which the antrea-agent sends IGMP queries to Pods.\n# Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\", \"m\", \"h\".\n  igmpQueryInterval: \"125s\"\n\n# The network CIDRs of the interface on Node which is used for tunneling or routing the traffic across\n# Nodes. If there are multiple interfaces configured the same network CIDR, the first one is used. The\n# IP address used for tunneling or routing traffic to remote Nodes is decided in the following order of\n# preference (from highest to lowest):\n# 1. transportInterface\n# 2. transportInterfaceCIDRs\n# 3. The Node IP\ntransportInterfaceCIDRs:\n\n# Option antreaProxy contains AntreaProxy related configuration options.\nantreaProxy:\n  # ProxyAll tells antrea-agent to proxy all Service traffic, including NodePort, LoadBalancer, and ClusterIP traffic,\n  # regardless of where they come from. Therefore, running kube-proxy is no longer required. This requires the AntreaProxy\n  # feature to be enabled.\n  # Note that this option is experimental. If kube-proxy is removed, option kubeAPIServerOverride must be used to access\n  # apiserver directly.\n  proxyAll: false\n  # A string array of values which specifies the host IPv4/IPv6 addresses for NodePort. Values can be valid IP blocks.\n  # (e.g. 1.2.3.0/24, 1.2.3.4/32). An empty string slice is meant to select all host IPv4/IPv6 addresses.\n  # Note that the option is only valid when proxyAll is true.\n  nodePortAddresses:\n  # An array of string values to specify a list of Services which should be ignored by AntreaProxy (traffic to these\n  # Services will not be load-balanced). Values can be a valid ClusterIP (e.g. 10.11.1.2) or a Service name\n  # with Namespace (e.g. kube-system/kube-dns)\n  skipServices:\n  # When ProxyLoadBalancerIPs is set to false, AntreaProxy no longer load-balances traffic destined to the\n  # External IPs of LoadBalancer Services. This is useful when the external LoadBalancer provides additional\n  # capabilities (e.g. TLS termination) and it is desirable for Pod-to-ExternalIP traffic to be sent to the\n  # external LoadBalancer instead of being load-balanced to an Endpoint directly by AntreaProxy.\n  # Note that setting ProxyLoadBalancerIPs to false usually only makes sense when ProxyAll is set to true and\n  # kube-proxy is removed from the cluser, otherwise kube-proxy will still load-balance this traffic.\n  proxyLoadBalancerIPs: true\n\n# IPsec tunnel related configurations.\nipsec:\n  # The authentication mode of IPsec tunnel. It has the following options:\n  # - psk (default): Use pre-shared key (PSK) for IKE authentication.\n  # - cert:          Use CA-signed certificates for IKE authentication. This option requires the `IPsecCertAuth`\n  #                  feature gate to be enabled.\n  authenticationMode: \"psk\"\n\nmulticluster:\n# Enable Antrea Multi-cluster Gateway to support cross-cluster traffic.\n# This feature is supported only with encap mode.\n  enable: false\n# The Namespace where Antrea Multi-cluster Controller is running.\n# The default is antrea-agent's Namespace.\n  namespace: \"\"\n# Enable StretchedNetworkPolicy which could be enforced on cross-cluster traffic.\n# This feature is supported only with encap mode.\n  enableStretchedNetworkPolicy: false\n",
                  "antreaCNIConfig": "{\n    \"cniVersion\":\"0.3.0\",\n    \"name\": \"antrea\",\n    \"plugins\": [\n        {\n            \"type\": \"antrea\",\n            \"ipam\": {\n                \"type\": \"host-local\"\n            }\n        }\n        ,\n        {\n            \"type\": \"portmap\",\n            \"capabilities\": {\"portMappings\": true}\n        }\n        ,\n        {\n            \"type\": \"bandwidth\",\n            \"capabilities\": {\"bandwidth\": true}\n        }\n    ]\n}\n",
                  "antreaControllerConfig": "# FeatureGates is a map of feature names to bools that enable or disable experimental features.\nfeatureGates:\n# AllAlpha is a global toggle for alpha features. Per-feature key values override the default set by AllAlpha.\n#  AllAlpha: false\n\n# AllBeta is a global toggle for beta features. Per-feature key values override the default set by AllBeta.\n#  AllBeta: false\n\n# Enable traceflow which provides packet tracing feature to diagnose network issue.\n#  Traceflow: true\n\n# Enable Antrea ClusterNetworkPolicy feature to complement K8s NetworkPolicy for cluster admins\n# to define security policies which apply to the entire cluster, and Antrea NetworkPolicy\n# feature that supports priorities, rule actions and externalEntities in the future.\n#  AntreaPolicy: true\n\n# Enable collecting and exposing NetworkPolicy statistics.\n#  NetworkPolicyStats: true\n\n# Enable multicast traffic.\n#  Multicast: false\n\n# Enable controlling SNAT IPs of Pod egress traffic.\n#  Egress: true\n\n# Run Kubernetes NodeIPAMController with Antrea.\n#  NodeIPAM: false\n\n# Enable AntreaIPAM, which can allocate IP addresses from IPPools. AntreaIPAM is required by the\n# bridging mode and allocates IPs to Pods in bridging mode. It is also required to use Antrea for\n# IPAM when configuring secondary network interfaces with Multus.\n#  AntreaIPAM: false\n\n# Enable managing external IPs of Services of LoadBalancer type.\n#  ServiceExternalIP: false\n\n# Enable certificated-based authentication for IPsec.\n#  IPsecCertAuth: false\n\n# Enable managing ExternalNode for unmanaged VM/BM.\n#  ExternalNode: false\n\n# Enable collecting support bundle files with SupportBundleCollection CRD.\n#  SupportBundleCollection: false\n\n# Enable multi-cluster features.\n#  Multicluster: false\n\n# Enable users to protect their applications by specifying how they are allowed to communicate with others, taking\n# into account application context.\n#  L7NetworkPolicy: false\n\n# The port for the antrea-controller APIServer to serve on.\n# Note that if it's set to another value, the `containerPort` of the `api` port of the\n# `antrea-controller` container must be set to the same value.\napiPort: 10349\n\n# Enable metrics exposure via Prometheus. Initializes Prometheus metrics listener.\nenablePrometheusMetrics: true\n\n# Indicates whether to use auto-generated self-signed TLS certificate.\n# If false, a Secret named \"antrea-controller-tls\" must be provided with the following keys:\n#   ca.crt: \u003cCA certificate\u003e\n#   tls.crt: \u003cTLS certificate\u003e\n#   tls.key: \u003cTLS private key\u003e\nselfSignedCert: true\n\n# Comma-separated list of Cipher Suites. If omitted, the default Go Cipher Suites will be used.\n# https://golang.org/pkg/crypto/tls/#pkg-constants\n# Note that TLS1.3 Cipher Suites cannot be added to the list. But the apiserver will always\n# prefer TLS1.3 Cipher Suites whenever possible.\ntlsCipherSuites: \"\"\n\n# TLS min version from: VersionTLS10, VersionTLS11, VersionTLS12, VersionTLS13.\ntlsMinVersion: \"\"\n\nnodeIPAM:\n  # Enable the integrated Node IPAM controller within the Antrea controller.\n  enableNodeIPAM: false\n  # CIDR ranges for Pods in cluster. String array containing single CIDR range, or multiple ranges.\n  # The CIDRs could be either IPv4 or IPv6. At most one CIDR may be specified for each IP family.\n  # Value ignored when enableNodeIPAM is false.\n  clusterCIDRs:\n  # CIDR ranges for Services in cluster. It is not necessary to specify it when there is no overlap with clusterCIDRs.\n  # Value ignored when enableNodeIPAM is false.\n  serviceCIDR: \"\"\n  serviceCIDRv6: \"\"\n  # Mask size for IPv4 Node CIDR in IPv4 or dual-stack cluster. Value ignored when enableNodeIPAM is false\n  # or when IPv4 Pod CIDR is not configured. Valid range is 16 to 30.\n  nodeCIDRMaskSizeIPv4: 24\n  # Mask size for IPv6 Node CIDR in IPv6 or dual-stack cluster. Value ignored when enableNodeIPAM is false\n  # or when IPv6 Pod CIDR is not configured. Valid range is 64 to 126.\n  nodeCIDRMaskSizeIPv6: 64\n\nipsecCSRSigner:\n  # Determines the auto-approve policy of Antrea CSR signer for IPsec certificates management.\n  # If enabled, Antrea will auto-approve the CertificateSingingRequest (CSR) if its subject and x509 extensions\n  # are permitted, and the requestor can be validated. If K8s `BoundServiceAccountTokenVolume` feature is enabled,\n  # the Pod identity will also be validated to provide maximum security.\n  # If set to false, Antrea will not auto-approve CertificateSingingRequests and they need to be approved\n  # manually by `kubectl certificate approve`.\n  autoApprove: true\n  # Indicates whether to use auto-generated self-signed CA certificate.\n  # If false, a Secret named \"antrea-ipsec-ca\" must be provided with the following keys:\n  #   tls.crt: \u003cCA certificate\u003e\n  #   tls.key: \u003cCA private key\u003e\n  selfSignedCA: true\n\nmulticluster:\n  # Enable StretchedNetworkPolicy which allow Antrea-native policies to select peers\n  # from other clusters in a ClusterSet.\n  enableStretchedNetworkPolicy: false\n",
                  "antreaImage": "antrea/antrea-ubi:v2.1.0",
                  "antreaPlatform": "openshift"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: projects.packages.broadcom.com/antreainterworking/antrea-operator:v2.1.0_vmware.3@sha256:70f129499c801f164adf8772e7ccd5d1fc659df5f298d7851e6bf02ee760f2a9
          createdAt: "2025-04-16T12:00:00Z"
          description: An operator which installs Antrea network CNI plugin on the
            Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          support: projectantrea-maintainers@googlegroups.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AntreaInstall is the Schema for the antreainstalls API
            displayName: Antrea Install
            kind: AntreaInstall
            name: antreainstalls.operator.antrea.vmware.com
            version: v1
        description: An operator which installs Antrea network CNI plugin on the Kubernetes
          cluster.
        displayName: Antrea-by-Broadcom Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        links:
        - name: Antrea Operator For Kubernetes
          url: https://github.com/vmware/antrea-operator-for-kubernetes
        maintainers:
        - email: projectantrea-maintainers@googlegroups.com
          name: Project Antrea Maintainers
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: antrea.io
        relatedImages:
        - projects.packages.broadcom.com/antreainterworking/antrea-operator:v2.1.0_vmware.3@sha256:70f129499c801f164adf8772e7ccd5d1fc659df5f298d7851e6bf02ee760f2a9
        version: 1.10.0
      entries:
      - name: antrea-by-broadcom-operator.v1.10.0
        version: 1.10.0
      name: alpha
    defaultChannel: alpha
    packageName: antrea-by-broadcom-operator
    provider:
      name: antrea.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Nokia
      provider-url: https://docs.eda.dev
    name: eda-k8s-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eda-openshift-operator.v4.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.eda.nokia.com/v1",
                "kind": "ConnectControllerConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "operator"
                  },
                  "name": "connectcontrollerconfig-sample",
                  "namespace": "eda-connect-k8s"
                },
                "spec": {
                  "connectPluginName": "",
                  "controllerImagePullSecret": "",
                  "edaNamespace": "eda",
                  "heartBeat": "",
                  "logLevel": "",
                  "skipTLSVerify": false,
                  "tlsCertificateData": "",
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:c43a7fa7475db83011cc08b728982511fc0279a259e1aacc8ddde9e15f947725
          createdAt: "2025-06-10T15:47:27Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: eda-connect-k8s
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["connectnetworkattachmentstatuses.connect.eda.nokia.com",
            connectnetworkdefinitions.connect.eda.nokia.com]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConnectNetworkAttachmentStatus is the Schema for the connectnetworkattachmentstatuses
              API
            displayName: Connect NetworkAttachment Status
            kind: ConnectNetworkAttachmentStatus
            name: connectnetworkattachmentstatuses.connect.eda.nokia.com
            version: v1
          - description: ConnectNetworkDefinition is the Schema for the connectnetworkdefinitions
              API
            displayName: Connect Network Definition
            kind: ConnectNetworkDefinition
            name: connectnetworkdefinitions.connect.eda.nokia.com
            version: v1
          - description: ConnectControllerConfig is the Schema for the connectcontrollerconfigs
              API
            displayName: Connect Controller Config
            kind: ConnectControllerConfig
            name: connectcontrollerconfigs.config.eda.nokia.com
            version: v1
        description: "K8s operator to deploy eda connect controller to configure EDA
          Connect.\eb"
        displayName: eda-openshift-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - EDA Connect
        - K8s Controller
        - EDA K8s Controller Operator
        links:
        - name: EDA Connect K8s Controller Operator
          url: https://docs.eda.dev/connect/openshift-plugin/
        maintainers:
        - email: support@nokia.com
          name: nokia
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Nokia
          url: https://docs.eda.dev
        relatedImages:
        - ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:c43a7fa7475db83011cc08b728982511fc0279a259e1aacc8ddde9e15f947725
        version: 4.0.0
      entries:
      - name: eda-openshift-operator.v4.0.0
        version: 4.0.0
      name: alpha
    - currentCSV: eda-openshift-operator.v4.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.eda.nokia.com/v1",
                "kind": "ConnectControllerConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "operator"
                  },
                  "name": "connectcontrollerconfig-sample",
                  "namespace": "eda-connect-k8s"
                },
                "spec": {
                  "connectPluginName": "",
                  "controllerImagePullSecret": "",
                  "edaNamespace": "eda",
                  "heartBeat": "",
                  "logLevel": "",
                  "skipTLSVerify": false,
                  "tlsCertificateData": "",
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:ec264b52d581b8fd76e737b9811ab894957a4edf2f627f90bd9fe707890e6bf5
          createdAt: "2025-06-10T15:47:27Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: eda-connect-k8s
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["connectnetworkattachmentstatuses.connect.eda.nokia.com",
            connectnetworkdefinitions.connect.eda.nokia.com]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConnectNetworkAttachmentStatus is the Schema for the connectnetworkattachmentstatuses
              API
            displayName: Connect NetworkAttachment Status
            kind: ConnectNetworkAttachmentStatus
            name: connectnetworkattachmentstatuses.connect.eda.nokia.com
            version: v1
          - description: ConnectNetworkDefinition is the Schema for the connectnetworkdefinitions
              API
            displayName: Connect Network Definition
            kind: ConnectNetworkDefinition
            name: connectnetworkdefinitions.connect.eda.nokia.com
            version: v1
          - description: ConnectControllerConfig is the Schema for the connectcontrollerconfigs
              API
            displayName: Connect Controller Config
            kind: ConnectControllerConfig
            name: connectcontrollerconfigs.config.eda.nokia.com
            version: v1
        description: "K8s operator to deploy eda connect controller to configure EDA
          Connect.\eb"
        displayName: eda-openshift-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - EDA Connect
        - K8s Controller
        - EDA K8s Controller Operator
        links:
        - name: EDA Connect K8s Controller Operator
          url: https://docs.eda.dev/connect/openshift-plugin/
        maintainers:
        - email: support@nokia.com
          name: nokia
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Nokia
          url: https://docs.eda.dev
        relatedImages:
        - ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:ec264b52d581b8fd76e737b9811ab894957a4edf2f627f90bd9fe707890e6bf5
        version: 4.0.1
      entries:
      - name: eda-openshift-operator.v4.0.1
        version: 4.0.1
      name: stable
    defaultChannel: alpha
    packageName: eda-k8s-operator
    provider:
      name: Nokia
      url: https://docs.eda.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: prometurbo-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometurbo-operator.v8.17.40-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5-beta.1 <8.17.40-beta.1'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
          - kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.

          ### Get Started
          Please follow [this](https://www.ibm.com/docs/en/tarm/8.15.x?topic=configuration-prometheus) for an overview of Prometurbo and deployment instructions.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/prometurbo
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/prometurbo-operator@sha256:a1605506884f6bce7471d79c6243728bce991ceb381d7fe0c51f1cdcbeb96e22
        version: 8.17.40-beta.1
      entries:
      - name: prometurbo-operator.v8.17.40-beta.1
        version: 8.17.40-beta.1
      name: beta
    - currentCSV: prometurbo-operator.v8.17.40
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5 <8.17.40'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
          - kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.

          ### Get Started
          Please follow [this](https://www.ibm.com/docs/en/tarm/8.15.x?topic=configuration-prometheus) for an overview of Prometurbo and deployment instructions.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/prometurbo
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/prometurbo-operator@sha256:cbcf90e44ffcedec8e3f161eff4c95b581924a1db195bf55d13da3714346b26f
        version: 8.17.40
      entries:
      - name: prometurbo-operator.v8.17.40
        version: 8.17.40
      name: stable
    defaultChannel: stable
    packageName: prometurbo-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: AppDynamics LLC
      provider-url: https://www.appdynamics.com/
    name: appdynamics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appdynamics-operator.v24.7.0-861
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clusteragent",
                "metadata": {
                  "name": "k8s-cluster-agent",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "appName": "<app-name>",
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/cluster-agent:tag"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "InfraViz",
                "metadata": {
                  "name": "appdynamics-infraviz",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/machine-agent:latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3c420cce8b626d8f5cf424410527c6a7a17ca3348e903dd8cdfcd74819b8de41
          createdAt: "2023-02-06T14:26:07Z"
          description: AppDynamics Operator to install AppDynamics Cluster Agent &
            AppDynamics Machine Agent for K8s Monitoring for various Cloud Providers
            like EKS, AKS etc.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: support@appdynamics.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clusteragent is the Schema for the clusteragents API
            displayName: Clusteragent
            kind: Clusteragent
            name: clusteragents.cluster.appdynamics.com
            version: v1alpha1
          - description: InfraViz is the Schema for the infravizs API
            displayName: Infra Viz
            kind: InfraViz
            name: infravizs.cluster.appdynamics.com
            version: v1alpha1
        description: |-
          AppDynamics Operator to install AppDynamics Cluster Agent & Machine Agent for K8s Monitoring on various Cloud Providers like EKS, AKS, GCP, Openshift etc.
          **About AppDynamics Operator**
          AppDynamics Operator simplifies the installation and lifecycle management of the AppDynamics ClusterAgent and the AppDynamics Machine Agent for K8s Monitoring on different Cloud Providers like EKS, AKS, GCP, OpenShift etc.
          **Installation Steps**
          Please refer below links for installation guide:
          [Install AppDynamics Cluster Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli)
          [Install AppDynamics Machine Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli)
          [Install AppDynamics Cluster Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts)
          [Install Appdynamics Machine Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts)
        displayName: AppDynamics Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AppDynamics
        - Operator
        - AppDynamics Cluster Agent
        - AppDynamics Machine Agent
        - AppDynamics K8s Monitoring
        links:
        - name: AppDynamics Kubernetes Monitoring Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent
        - name: AppDynamics Cluster Agent Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/overview-of-cluster-monitoring
        - name: AppDynamics Cluster Agent Requirements
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/cluster-agent-requirements-and-supported-environments
        - name: Install AppDynamics Cluster Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli
        - name: Install AppDynamics Cluster Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts
        - name: Install AppDynamics Machine Agent & NetViz Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli
        - name: Install AppDynamics Machine Agent & NetViz Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: AppDynamics LLC
          url: https://www.appdynamics.com/
        relatedImages:
        - registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3c420cce8b626d8f5cf424410527c6a7a17ca3348e903dd8cdfcd74819b8de41
        version: 24.7.0-861
      entries:
      - name: appdynamics-operator.v24.7.0-861
        version: 24.7.0-861
      name: alpha
    - currentCSV: appdynamics-operator.v25.9.0-1218
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clusteragent",
                "metadata": {
                  "name": "k8s-cluster-agent",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "appName": "<app-name>",
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/cluster-agent:tag"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "InfraViz",
                "metadata": {
                  "name": "appdynamics-infraviz",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/machine-agent:latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3320c2c8d5f60adc541befd099dac1ef96257b16e0df28b727b0e589f78d981a
          createdAt: "2023-02-06T14:26:07Z"
          description: AppDynamics Operator to install AppDynamics Cluster Agent &
            AppDynamics Machine Agent for K8s Monitoring for various Cloud Providers
            like EKS, AKS etc.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: support@appdynamics.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clusteragent is the Schema for the clusteragents API
            displayName: Clusteragent
            kind: Clusteragent
            name: clusteragents.cluster.appdynamics.com
            version: v1alpha1
          - description: InfraViz is the Schema for the infravizs API
            displayName: Infra Viz
            kind: InfraViz
            name: infravizs.cluster.appdynamics.com
            version: v1alpha1
        description: |-
          AppDynamics Operator to install AppDynamics Cluster Agent & Machine Agent for K8s Monitoring on various Cloud Providers like EKS, AKS, GCP, Openshift etc.
          **About AppDynamics Operator**
          AppDynamics Operator simplifies the installation and lifecycle management of the AppDynamics ClusterAgent and the AppDynamics Machine Agent for K8s Monitoring on different Cloud Providers like EKS, AKS, GCP, OpenShift etc.
          **Installation Steps**
          Please refer below links for installation guide:
          [Install AppDynamics Cluster Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli)
          [Install AppDynamics Machine Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli)
          [Install AppDynamics Cluster Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts)
          [Install Appdynamics Machine Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts)
        displayName: AppDynamics Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - AppDynamics
        - Operator
        - AppDynamics Cluster Agent
        - AppDynamics Machine Agent
        - AppDynamics K8s Monitoring
        links:
        - name: AppDynamics Kubernetes Monitoring Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent
        - name: AppDynamics Cluster Agent Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/overview-of-cluster-monitoring
        - name: AppDynamics Cluster Agent Requirements
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/cluster-agent-requirements-and-supported-environments
        - name: Install AppDynamics Cluster Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli
        - name: Install AppDynamics Cluster Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts
        - name: Install AppDynamics Machine Agent & NetViz Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli
        - name: Install AppDynamics Machine Agent & NetViz Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: AppDynamics LLC
          url: https://www.appdynamics.com/
        relatedImages:
        - registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3320c2c8d5f60adc541befd099dac1ef96257b16e0df28b727b0e589f78d981a
        version: 25.9.0-1218
      entries:
      - name: appdynamics-operator.v25.9.0-1218
        version: 25.9.0-1218
      name: stable
    defaultChannel: alpha
    packageName: appdynamics-operator
    provider:
      name: AppDynamics LLC
      url: https://www.appdynamics.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: www.redhat.com
    name: odf-node-recovery-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-node-recovery-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "NodeRecovery",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "odf-node-recovery-operator",
                    "app.kubernetes.io/instance": "noderecovery-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "noderecovery",
                    "app.kubernetes.io/part-of": "odf-node-recovery-operator"
                  },
                  "name": "noderecovery-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Storage,OpenShift Optional
          description: ODF Node Recovery is an operator that assist in the recovery
            of and ODF cluster that has a device or a node replaced
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/jordigilh/odf-node-recovery-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeRecovery is the Schema for the noderecoveries API
            displayName: Node Recovery
            kind: NodeRecovery
            name: noderecoveries.odf.openshift.io
            version: v1alpha1
        description: |
          ODF Node Recovery is an operator that assist in the recovery of an
          ODF cluster in these 2 scenarios:
          * <a href="https://docs.redhat.com/en/documentation/red_hat_openshift_data_foundation/4.14/html-single/replacing_devices/index"> Device replacement<a/>
          * <a href="https://docs.redhat.com/en/documentation/red_hat_openshift_data_foundation/4.14/html-single/replacing_nodes/index"> Node replacement</a>
        displayName: ODF Node Recovery Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - node recovery
        links:
        - name: GitHub Repository
          url: https://github.com/jordigilh/odf-node-recovery-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
          url: www.redhat.com
        relatedImages:
        - quay.io/jordigilh/odf-node-recovery-controller@sha256:4164fafa9bb1c4d6121f4c5fc700fa2827c3994ca98a65d641471a2a46d4f59c
        version: 1.1.0
      entries:
      - name: odf-node-recovery-operator.v1.1.0
        version: 1.1.0
      - name: odf-node-recovery-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: odf-node-recovery-operator
    provider:
      name: Red Hat, Inc.
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-prometheus-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-prometheus-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:b6071d6a1ec7cdd5f65b6d7296c1e9cd5660972a8ed614da4277e3a4ab5d1df1
          createdAt: "2024-03-01T23:43:00Z"
          description: Manage the full lifecycle of configuring and managing Prometheus
            and Alertmanager servers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/prometheus-operator/prometheus-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Prometheus Operator for Kubernetes provides easy monitoring definitions for Kubernetes services and deployment and management of Prometheus instances.

          Once installed, the Prometheus Operator provides the following features:

          * **Create/Destroy**: Easily launch a Prometheus instance for your Kubernetes namespace, a specific application or team easily using the Operator.
          * **Simple Configuration**: Configure the fundamentals of Prometheus like versions, persistence, retention policies, and replicas from a native Kubernetes resource.
          * **Target Services via Labels**: Automatically generate monitoring target configurations based on familiar Kubernetes label queries; no need to learn a Prometheus specific configuration language.

          ### Other Supported Features

          **High availability**

          Multiple instances are run across failure zones and data is replicated. This keeps your monitoring available during an outage, when you need it most.

          **Updates via automated operations**

          New Prometheus versions are deployed using a rolling update with no downtime, making it easy to stay up to date.

          **Handles the dynamic nature of containers**

          Alerting rules are attached to groups of containers instead of individual instances, which is ideal for the highly dynamic nature of container deployment.
        displayName: Prometheus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - prometheus
        - monitoring
        - tsdb
        - alerting
        - observability
        links:
        - name: Prometheus
          url: https://www.prometheus.io/
        - name: Documentation
          url: https://github.com/prometheus-operator/prometheus-operator/tree/master/Documentation
        - name: Prometheus Operator
          url: https://github.com/prometheus-operator/prometheus-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:b6071d6a1ec7cdd5f65b6d7296c1e9cd5660972a8ed614da4277e3a4ab5d1df1
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:413f7da5cb2e78abf33eb4e43f5fd483c19e791416eec7f5d4c3dbeb54db33a0
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader-rhel9@sha256:eb7614528884f2eefe1ae5e33a09624590c23e0e75510b4d88a274880cf090bc
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager-rhel9@sha256:7d1a3392872fd3990ae3acce79da659ba2238b30421824b7be533dcf8b93e89c
        version: 4.16.16-rhodf
      entries:
      - name: odf-prometheus-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odf-prometheus-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odf-prometheus-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odf-prometheus-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odf-prometheus-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odf-prometheus-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odf-prometheus-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odf-prometheus-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odf-prometheus-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odf-prometheus-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odf-prometheus-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odf-prometheus-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odf-prometheus-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odf-prometheus-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odf-prometheus-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odf-prometheus-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odf-prometheus-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odf-prometheus-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solace Corporation
      provider-url: www.solace.com
    name: pubsubplus-eventbroker-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pubsubplus-eventbroker-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pubsubplus.solace.com/v1beta1",
                "kind": "PubSubPlusEventBroker",
                "metadata": {
                  "name": "non-ha-standalone-example"
                },
                "spec": {
                  "redundancy": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "true"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.10
          containerImage: registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator@sha256:d68db92bcb217d94dc0ad0a1cea625a0a9c9299d886d325a903163f9f887dcd5
          createdAt: "2025-07-31T19:48:36Z"
          description: The Solace PubSub+ Event Broker Operator deploys and manages
            the lifecycle of PubSub+ Event Brokers
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
          support: Solace Products
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PubSub+ Event Broker
            displayName: PubSub+ Event Broker
            kind: PubSubPlusEventBroker
            name: pubsubpluseventbrokers.pubsubplus.solace.com
            version: v1beta1
        description: |
          ## Solace PubSub+ Platform
          Solace [PubSub+ Platform](https://solace.com/products/platform/) is a complete event streaming and management platform
          for real-time enterprises. The [PubSub+ Software Event Broker](https://solace.com/products/event-broker/software/) efficiently streams
          event-driven information between applications, IoT devices, and user interfaces running in the cloud, on-premises,
          and in hybrid environments using open APIs and protocols like AMQP, JMS, MQTT, REST and WebSocket.
          It can be installed into a variety of public and private clouds, PaaS, and on-premises environments.
          Event brokers in multiple locations can be linked together in an [Event Mesh](https://solace.com/what-is-an-event-mesh/)
          to dynamically share events across the distributed enterprise.

          ## Solace PubSub+ Event Broker Operator
          Solace Pubsub+ Event Broker Operator automatically deploys and manages PubSub+ Software Event Brokers on Kubernetes and OpenShift environments.

          Features include:
          * Initial deployment configuration: Production-ready HA, non-HA or for developers
          * Rolling update for configuration changes or broker upgrades
          * Prometheus monitoring support

          ## Getting started
          Refer to the [Solace PubSub+ Event Broker Operator Quick Start](https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/README.md)
        displayName: Solace PubSub+ Event Broker Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - solace
        - pubsubplus
        - pubsub+
        - pubsub
        - messaging
        - advanced event broker
        - event broker
        - event mesh
        - event streaming
        - data streaming
        - event integration
        - middleware
        links:
        - name: Operator QuickStart
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
        - name: Operator Documentation
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/docs/EventBrokerOperatorUserGuide.md
        - name: Solace.com
          url: https://solace.com
        - name: Understanding Solace Technology
          url: https://solace.com/products/platform/
        - name: Solace PubSub+ Technical Documentation
          url: https://docs.solace.com/
        - name: Solace Developer Portal
          url: https://www.solace.dev/
        - name: Solace Community
          url: https://solace.community/
        maintainers:
        - email: support@solace.com
          name: Solace R&D
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Solace Corporation
          url: www.solace.com
        relatedImages:
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator@sha256:d68db92bcb217d94dc0ad0a1cea625a0a9c9299d886d325a903163f9f887dcd5
        version: 1.4.0
      entries:
      - name: pubsubplus-eventbroker-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: pubsubplus-eventbroker-operator
    provider:
      name: Solace Corporation
      url: www.solace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: StreamNative
      provider-url: https://streamnative.io
    name: function-mesh
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: function-mesh.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "java-function-sample",
                  "namespace": "default"
                },
                "spec": {
                  "autoAck": true,
                  "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                  "clusterName": "test-pulsar",
                  "forwardSourceMessageProperty": true,
                  "input": {
                    "topics": [
                      "persistent://public/default/java-function-input-topic"
                    ],
                    "typeClassName": "java.lang.String"
                  },
                  "java": {
                    "extraDependenciesDir": "random-dir/",
                    "jar": "pulsar-functions-api-examples.jar",
                    "jarLocation": "public/default/nlu-test-java-function"
                  },
                  "logTopic": "persistent://public/default/logging-function-logs",
                  "maxPendingAsyncRequests": 1000,
                  "maxReplicas": 5,
                  "output": {
                    "topic": "persistent://public/default/java-function-output-topic",
                    "typeClassName": "java.lang.String"
                  },
                  "pod": {
                    "annotations": {
                      "managed-function": "true"
                    },
                    "env": [
                      {
                        "name": "EXAMPLE_VARIABLE",
                        "value": "exampleValue"
                      }
                    ],
                    "imagePullSecrets": [
                      {
                        "name": "regcred"
                      }
                    ],
                    "initContainers": [
                      {
                        "command": [
                          "sh",
                          "-c",
                          "echo The app is running! \u0026\u0026 sleep 30"
                        ],
                        "image": "busybox:1.28",
                        "name": "init-function"
                      }
                    ],
                    "labels": {
                      "locaction": "mtv"
                    },
                    "sidecars": [
                      {
                        "command": [
                          "sh",
                          "-c",
                          "echo The app is running! \u0026\u0026 sleep 30000"
                        ],
                        "image": "busybox:1.28",
                        "name": "sidecar-function"
                      }
                    ],
                    "volumes": [
                      {
                        "emptyDir": {},
                        "name": "cache-volume"
                      }
                    ]
                  },
                  "pulsar": {
                    "authSecret": "test-auth",
                    "pulsarConfig": "test-pulsar",
                    "tlsSecret": "test-tls"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "secretsMap": {
                    "name": {
                      "key": "username",
                      "path": "test-secret"
                    },
                    "pwd": {
                      "key": "password",
                      "path": "test-secret"
                    }
                  },
                  "volumeMounts": [
                    {
                      "mountPath": "/cache",
                      "name": "cache-volume"
                    }
                  ]
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "FunctionMesh",
                "metadata": {
                  "name": "functionmesh-sample"
                },
                "spec": {
                  "functions": [
                    {
                      "autoAck": true,
                      "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                      "clusterName": "test-pulsar",
                      "forwardSourceMessageProperty": true,
                      "input": {
                        "topics": [
                          "persistent://public/default/functionmesh-input-topic"
                        ],
                        "typeClassName": "java.lang.String"
                      },
                      "java": {
                        "jar": "pulsar-functions-api-examples.jar",
                        "jarLocation": "public/default/nlu-test-functionmesh-ex1"
                      },
                      "logTopic": "persistent://public/default/logging-function-log",
                      "maxReplicas": 1,
                      "name": "ex1",
                      "output": {
                        "topic": "persistent://public/default/mid-topic",
                        "typeClassName": "java.lang.String"
                      },
                      "pulsar": {
                        "pulsarConfig": "mesh-test-pulsar"
                      },
                      "replicas": 1,
                      "resources": {
                        "limits": {
                          "cpu": "0.2",
                          "memory": "1.1G"
                        },
                        "requests": {
                          "cpu": "0.1",
                          "memory": "1G"
                        }
                      }
                    },
                    {
                      "autoAck": true,
                      "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                      "clusterName": "test-pulsar",
                      "forwardSourceMessageProperty": true,
                      "input": {
                        "topics": [
                          "persistent://public/default/mid-topic"
                        ],
                        "typeClassName": "java.lang.String"
                      },
                      "java": {
                        "jar": "pulsar-functions-api-examples.jar",
                        "jarLocation": "public/default/nlu-test-functionmesh-ex2"
                      },
                      "logTopic": "persistent://public/default/logging-function-logs",
                      "maxReplicas": 1,
                      "name": "ex2",
                      "output": {
                        "topic": "persistent://public/default/functionmesh-output-topic",
                        "typeClassName": "java.lang.String"
                      },
                      "pulsar": {
                        "pulsarConfig": "mesh-test-pulsar"
                      },
                      "replicas": 1,
                      "resources": {
                        "limits": {
                          "cpu": "0.2",
                          "memory": "1.1G"
                        },
                        "requests": {
                          "cpu": "0.1",
                          "memory": "1G"
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Sink",
                "metadata": {
                  "name": "sink-sample"
                },
                "spec": {
                  "autoAck": true,
                  "className": "org.apache.pulsar.io.elasticsearch.ElasticSearchSink",
                  "clusterName": "test-pulsar",
                  "image": "streamnative/pulsar-io-elastic-search:2.10.0.0-rc10",
                  "input": {
                    "topics": [
                      "persistent://public/default/input"
                    ],
                    "typeClassName": "[B"
                  },
                  "java": {
                    "jar": "connectors/pulsar-io-elastic-search-2.10.0.0-rc10.nar",
                    "jarLocation": ""
                  },
                  "maxReplicas": 1,
                  "pulsar": {
                    "pulsarConfig": "test-sink"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "sinkConfig": {
                    "elasticSearchUrl": "http://quickstart-es-http.default.svc.cluster.local:9200",
                    "indexName": "my_index",
                    "password": "wJ757TmoXEd941kXm07Z2GW3",
                    "typeName": "doc",
                    "username": "elastic"
                  }
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Source",
                "metadata": {
                  "name": "source-sample"
                },
                "spec": {
                  "className": "org.apache.pulsar.io.debezium.mongodb.DebeziumMongoDbSource",
                  "clusterName": "test-pulsar",
                  "forwardSourceMessageProperty": true,
                  "image": "streamnative/pulsar-io-debezium-mongodb:2.10.0.0-rc10",
                  "java": {
                    "jar": "connectors/pulsar-io-debezium-mongodb-2.10.0.0-rc10.nar",
                    "jarLocation": ""
                  },
                  "maxReplicas": 1,
                  "output": {
                    "producerConf": {
                      "maxPendingMessages": 1000,
                      "maxPendingMessagesAcrossPartitions": 50000,
                      "useThreadLocalProducers": true
                    },
                    "topic": "persistent://public/default/destination",
                    "typeClassName": "org.apache.pulsar.common.schema.KeyValue"
                  },
                  "pulsar": {
                    "pulsarConfig": "test-source"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "sourceConfig": {
                    "database.whitelist": "inventory",
                    "mongodb.hosts": "rs0/mongo-dbz-0.mongo.default.svc.cluster.local:27017,rs0/mongo-dbz-1.mongo.default.svc.cluster.local:27017,rs0/mongo-dbz-2.mongo.default.svc.cluster.local:27017",
                    "mongodb.name": "dbserver1",
                    "mongodb.password": "dbz",
                    "mongodb.task.id": "1",
                    "mongodb.user": "debezium",
                    "pulsar.service.url": "pulsar://test-pulsar-broker.default.svc.cluster.local:6650"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/function-mesh@sha256:3e13e5cc3ac4d7387ee0adc02cb5183aad6cf2301f76ab230482609c4da1c0b7
          createdAt: "2022-10-27T07:49:28Z"
          description: The Function Mesh Operator manages the Pulsar Functions and
            Connectors deployed on a Kubernetes cluster.
          olm.properties: '[{"type":"olm.maxOpenShiftVersion","value":"4.8"}]'
          olm.skipRange: <0.7.0
          operatorhub.io/ui-metadata-max-k8s-version: "1.21"
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/streamnative/function-mesh
          support: StreamNative
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FunctionMesh is the Schema for the functionmeshes API
            displayName: Mesh
            kind: FunctionMesh
            name: functionmeshes.compute.functionmesh.io
            version: v1alpha1
          - description: Function is the Schema for the functions API
            displayName: Function
            kind: Function
            name: functions.compute.functionmesh.io
            version: v1alpha1
          - description: Sink is the Schema for the sinks API
            displayName: Sink
            kind: Sink
            name: sinks.compute.functionmesh.io
            version: v1alpha1
          - description: Source is the Schema for the sources API
            displayName: Source
            kind: Source
            name: sources.compute.functionmesh.io
            version: v1alpha1
        description: |
          [Function Mesh](https://functionmesh.io/) is a serverless framework purpose-built for stream processing applications. It brings powerful event-streaming capabilities to your applications by orchestrating multiple [Pulsar Functions](http://pulsar.apache.org/docs/en/next/functions-overview/) and [Pulsar IO connectors](http://pulsar.apache.org/docs/en/next/io-overview/) for complex stream processing jobs on Kubernetes.
          Prerequisites:
          - Install cert-manager operator first.
        displayName: Function Mesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pulsar
        - function-mesh
        - serverless
        - pulsar-functions
        - pulsar-io-connectors
        - StreamNative
        links:
        - name: Function Mesh
          url: https://function-mesh.io
        maintainers:
        - email: function-mesh@streamnative.io
          name: Function Mesh
        maturity: alpha
        minKubeVersion: v1.17.0
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/function-mesh@sha256:3e13e5cc3ac4d7387ee0adc02cb5183aad6cf2301f76ab230482609c4da1c0b7
        - docker.cloudsmith.io/streamnative/mirrors/gcr.io/kubebuilder/kube-rbac-proxy@sha256:67ecb332573384515406ebd71816781366b70adb0eb66345e5980e92603373e1
        version: 0.7.0
      entries:
      - name: function-mesh.v0.7.0
        version: 0.7.0
      name: alpha
    defaultChannel: alpha
    packageName: function-mesh
    provider:
      name: StreamNative
      url: https://streamnative.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.3.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo-operator@sha256:32178a16dd2115e0a11a2e3dfbfc130ae1d0afd8e293131e76b6fc1a79b36c12
          createdAt: "2019-05-01T00:00:00Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/kubeturbo-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/kubeturbo-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo-operator@sha256:32178a16dd2115e0a11a2e3dfbfc130ae1d0afd8e293131e76b6fc1a79b36c12
        - registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo@sha256:671be694176d9cc31b8a22c11bbfefc2504ae72d27fd010b315b6839f9ab59d5
        version: 8.3.3
      entries:
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: stable
    defaultChannel: stable
    packageName: kubeturbo-certified-rhmp
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: devspaces
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devspacesoperator.v3.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "org.eclipse.che/v1",
                "kind": "CheCluster",
                "metadata": {
                  "name": "devspaces",
                  "namespace": "openshift-devspaces"
                },
                "spec": {
                  "auth": {
                    "identityProviderURL": "",
                    "oAuthClientName": "",
                    "oAuthSecret": ""
                  },
                  "database": {
                    "externalDb": false
                  },
                  "k8s": {
                    "ingressDomain": null,
                    "tlsSecretName": null
                  },
                  "metrics": {
                    "enable": true
                  },
                  "server": {
                    "workspaceNamespaceDefault": "<username>-devspaces"
                  },
                  "storage": {
                    "pvcStrategy": "common"
                  }
                }
              },
              {
                "apiVersion": "org.eclipse.che/v2",
                "kind": "CheCluster",
                "metadata": {
                  "name": "devspaces",
                  "namespace": "openshift-devspaces"
                },
                "spec": {
                  "components": {},
                  "containerRegistry": {},
                  "devEnvironments": {},
                  "gitServices": {},
                  "networking": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/devspaces/devspaces-rhel9-operator@sha256:39c95279bced68928d028ef026f82eafe53eddb3dbe57703c8007cb5a170a70a
          createdAt: "2025-09-11T22:30:49+00:00"
          description: Devfile v2 and v1 development solution, 1 instance per cluster,
            for portable, collaborative k8s workspaces.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/devspaces-images/
          skipRange: '>=3.0.0 <3.23.0'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              The `CheCluster` custom resource allows defining and managing Red Hat OpenShift Dev Spaces server installation.
              Based on these settings, the  Operator automatically creates and maintains several ConfigMaps:
              `che`, `plugin-registry` that will contain the appropriate environment variables
              of the various components of the installation. These generated ConfigMaps must NOT be updated manually.
            displayName: Red Hat OpenShift Dev Spaces instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v2
          - description: The `CheCluster` custom resource allows defining and managing
              a Che server installation
            displayName: Red Hat OpenShift Dev Spaces instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v1
        description: |
          A collaborative Kubernetes-native development solution that delivers OpenShift workspaces and in-browser IDE for rapid cloud application development.
          This operator installs the Plugin and Devfile registries, Dashboard, Gateway and the Red Hat OpenShift Dev Spaces server, and configures these services.
          OpenShift OAuth is used directly for authentication. TLS mode is on.

          ## How to Install
          Press the **Install** button, choose the channel and the upgrade strategy, and wait for the **Installed** Operator status.
          When the operator is installed, create a new Custom Resource (CR) of Kind CheCluster (click the **Create New** button).
          The CR spec contains all defaults. You can start using Red Hat OpenShift Dev Spaces when the CR status is set to **Available**, and you see a URL to Red Hat OpenShift Dev Spaces.

          ## Defaults
          By default, the operator deploys Red Hat OpenShift Dev Spaces with:
          * 10Gi storage
          * Auto-generated passwords
          * Bundled Plugin and Devfile registries

          Use `oc edit checluster/devspaces -n openshift-devspaces` to update Red Hat OpenShift Dev Spaces default installation options.
          See more in the [Installation guide](https://docs.redhat.com/en/documentation/red_hat_openshift_dev_spaces/3.23/html/administration_guide/configuring-devspaces).

          ### Certificates
          Operator uses a default router certificate to secure Red Hat OpenShift Dev Spaces routes.
          Follow the [guide](https://docs.redhat.com/en/documentation/red_hat_openshift_dev_spaces/3.23/html/administration_guide/configuring-devspaces#importing-untrusted-tls-certificates)
          to import certificates into Red Hat OpenShift Dev Spaces.
        displayName: Red Hat OpenShift Dev Spaces
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eclipse che
        - workspaces
        - devtools
        - developer
        - ide
        - java
        - devspaces
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/openshift-dev-spaces/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_openshift_dev_spaces
        - name: Operator GitHub Repo
          url: https://github.com/redhat-developer/devspaces-images/
        maintainers:
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/devspaces/pluginregistry-rhel9@sha256:3ee131041761f5250b29d497043ecf7820e02e03e73ed0f929c17e2247c0d7a3
        - registry.redhat.io/devspaces/traefik-rhel9@sha256:95cf6a28028ec7a546a3d8d78377f204bb559c3d1022a73f08c646e76b85ce85
        - registry.redhat.io/devspaces-tech-preview/jetbrains-ide-rhel9@sha256:46bf7bea716df74f14d2629fb7e0e2ab692564cec348ffb8a6f4fea1739b23c2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5a6edfecf1636805e6f0bcad02df10bceec764f0580f3ac8fe62aba978d50006
        - registry.redhat.io/devspaces/server-rhel9@sha256:c0ec8c792f88aacf6ee31a4b6d2d847d864707a084148b48e4e402388989f868
        - registry.redhat.io/devspaces/devspaces-rhel9-operator@sha256:39c95279bced68928d028ef026f82eafe53eddb3dbe57703c8007cb5a170a70a
        - registry.redhat.io/devspaces/code-rhel9@sha256:abc9b12131c8cdca0719e0404d354310f53769990b089f1a3adea31193798a8e
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6bcfa981a7a75c242a16afa6dab36bf6e3b8a1442fa863875919ad551f48eceb
        - registry.redhat.io/devspaces/configbump-rhel9@sha256:cecbc3796058618da9ad0473a6f937aae5aca25776bb434894286f22baa4677c
        - registry.redhat.io/devspaces/dashboard-rhel9@sha256:c34cf5ac9bf95c1c5082afd5a2941b5e412be8ca43aef6a72e8fe5a279a43271
        - registry.redhat.io/devspaces/udi-rhel9@sha256:f7325ee7259a0585e8017d4442d778a37f4c8a67d81fe70c241a014001e37539
        - registry.redhat.io/devspaces-tech-preview/idea-rhel9@sha256:df24ebb239a2cc0d4543309f51a7268c7d524d428a536e0b3443c51089c3f87e
        version: 3.23.0
      entries:
      - name: devspacesoperator.v3.23.0
        version: 3.23.0
      - name: devspacesoperator.v3.22.0
        version: 3.22.0
      - name: devspacesoperator.v3.21.0
        version: 3.21.0
      - name: devspacesoperator.v3.20.0
        version: 3.20.0
      - name: devspacesoperator.v3.19.0
        version: 3.19.0
      - name: devspacesoperator.v3.18.1-0.1739475915.p
        version: 3.18.1+0.1739475915.p
      - name: devspacesoperator.v3.18.1
        version: 3.18.1
      - name: devspacesoperator.v3.18.0
        version: 3.18.0
      - name: devspacesoperator.v3.17.0
        version: 3.17.0
      - name: devspacesoperator.v3.16.1
        version: 3.16.1
      - name: devspacesoperator.v3.16.0
        version: 3.16.0
      - name: devspacesoperator.v3.15.0
        version: 3.15.0
      - name: devspacesoperator.v3.14.0
        version: 3.14.0
      - name: devspacesoperator.v3.13.0
        version: 3.13.0
      - name: devspacesoperator.v3.12.0-0.1712003563.p
        version: 3.12.0+0.1712003563.p
      - name: devspacesoperator.v3.12.0
        version: 3.12.0
      - name: devspacesoperator.v3.11.0
        version: 3.11.0
      - name: devspacesoperator.v3.10.1
        version: 3.10.1
      - name: devspacesoperator.v3.10.0
        version: 3.10.0
      - name: devspacesoperator.v3.9.1
        version: 3.9.1
      - name: devspacesoperator.v3.9.0
        version: 3.9.0
      - name: devspacesoperator.v3.8.0
        version: 3.8.0
      - name: devspacesoperator.v3.7.2
        version: 3.7.2
      - name: devspacesoperator.v3.7.1-0.1691407091.p
        version: 3.7.1+0.1691407091.p
      - name: devspacesoperator.v3.7.1
        version: 3.7.1
      - name: devspacesoperator.v3.7.1-0.1691548593.p
        version: 3.7.1+0.1691548593.p
      - name: devspacesoperator.v3.7.0
        version: 3.7.0
      - name: devspacesoperator.v3.6.0
        version: 3.6.0
      - name: devspacesoperator.v3.6.0-0.1685320856.p
        version: 3.6.0+0.1685320856.p
      - name: devspacesoperator.v3.6.0-0.1685322634.p
        version: 3.6.0+0.1685322634.p
      - name: devspacesoperator.v3.5.0
        version: 3.5.0
      - name: devspacesoperator.v3.4.0-0.1675914643.p
        version: 3.4.0+0.1675914643.p
      - name: devspacesoperator.v3.4.0
        version: 3.4.0
      - name: devspacesoperator.v3.3.0-0.1673528130.p
        version: 3.3.0+0.1673528130.p
      - name: devspacesoperator.v3.3.0
        version: 3.3.0
      name: stable
    defaultChannel: stable
    packageName: devspaces
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Alvearie
      provider-url: https://alvearie.github.io
    name: alvearie-imaging-ingestion
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: imaging-ingestion-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomEventBridge",
                "metadata": {
                  "name": "events"
                },
                "spec": {
                  "dicomEventDrivenIngestionName": "core",
                  "eventBridge": {},
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "events",
                  "natsTokenSecret": "nats-events-secure-bound-token",
                  "natsUrl": "jetstream.imaging-ingestion.svc.cluster.local:4222",
                  "role": "hub"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomEventDrivenIngestion",
                "metadata": {
                  "name": "core"
                },
                "spec": {
                  "databaseConfigName": "db-config",
                  "databaseSecretName": "db-secret",
                  "eventProcessor": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  },
                  "imagePullPolicy": "Always"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomInstanceBinding",
                "metadata": {
                  "name": "stow"
                },
                "spec": {
                  "bindingConfigName": "instance-binding-config",
                  "bindingSecretName": "instance-binding-secret",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "instanceBinding": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomStudyBinding",
                "metadata": {
                  "name": "fhir"
                },
                "spec": {
                  "bindingConfigName": "study-binding-config",
                  "bindingSecretName": "study-binding-secret",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "studyBinding": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomwebIngestionService",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "bucketConfigName": "imaging-ingestion",
                  "bucketSecretName": "imaging-ingestion",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "providerName": "provider",
                  "stowService": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  },
                  "wadoService": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DimseIngestionService",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "applicationEntityTitle": "DICOM-INGEST",
                  "bucketConfigName": "imaging-ingestion",
                  "bucketSecretName": "imaging-ingestion",
                  "dicomEventDrivenIngestionName": "core",
                  "dimseService": {},
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "DIMSE",
                  "natsTokenSecret": "ingestion-nats-secure-bound-token",
                  "natsUrl": "nats-secure.imaging-ingestion.svc.cluster.local:4222",
                  "providerName": "provider"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DimseProxy",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "applicationEntityTitle": "DCM4CHEE",
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "DIMSE",
                  "natsTokenSecret": "ingestion-nats-secure-bound-token",
                  "natsUrl": "nats-secure.imaging-ingestion.svc.cluster.local:4222",
                  "proxy": {},
                  "targetDimseHost": "arc.dcm4chee.svc.cluster.local",
                  "targetDimsePort": 11112
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: alvearie/imaging-ingestion-operator:0.0.3
          createdAt: "2021-04-08T15:29:55Z"
          description: The Alvearie Imaging Ingestion provides a collection of components
            for extending enterprise medical imaging informatics systems to Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Alvearie/imaging-ingestion
          support: alvearie.slack.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DicomEventBridge is the Schema for the dicomeventbridges
              API
            displayName: Dicom Event Bridge
            kind: DicomEventBridge
            name: dicomeventbridges.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Event driven manifest of all DICOM data across all associated
              S3 buckets
            displayName: Dicom Event Driven Ingestion
            kind: DicomEventDrivenIngestion
            name: dicomeventdriveningestions.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Fan-out notification of DICOM instance data (STOW-RS)
            displayName: Dicom Instance Binding
            kind: DicomInstanceBinding
            name: dicominstancebindings.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Fan-out notification of DICOM studies (FHIRv4 ImagingStudy)
            displayName: Dicom Study Binding
            kind: DicomStudyBinding
            name: dicomstudybindings.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides DICOMweb WADO-RS and STOW-RS access to a S3 bucket
            displayName: Dicomweb Ingestion Service
            kind: DicomwebIngestionService
            name: dicomwebingestionservices.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides a proxied DIMSE Application Entity (AE) in the cluster
              for C-STORE operations to a storage space
            displayName: Dimse Ingestion Service
            kind: DimseIngestionService
            name: dimseingestionservices.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides a bidirectional proxied DIMSE Application Entity
              (AE) in the cluster
            displayName: Dimse Proxy
            kind: DimseProxy
            name: dimseproxies.imaging-ingestion.alvearie.org
            version: v1alpha1
        description: |
          The Alvearie Imaging Ingestion provides a collection of components for extending enterprise medical imaging informatics systems to Kubernetes.

          ## About this Operator
          This operator enables standards-based medical imaging application interfaces such as DICOM Message Service Element (DIMSE) and DICOMweb.
          The API's provided allow for extending enterprise imaging systems to the cloud with custom, fit for purpose deployments.
          This is possible at speed, low cost, and little or zero integration effort.or extending enterprise medical imaging informatics systems to a modern cloud.


          Some potential usages:

          - Perform lightweight secure bridging of DICOM from the enterprise imaging zone to a cloud availability zone.
          - Utilize S3 buckets for segregating DICOM data to be used for clinical or scientific purposes.
          - Fan-out DICOM data to a number of different DICOM subscribers
          - Generate FHIRv4 *ImagingStudy* resources from ingested content and publish to a FHIRv4 based patient logitudinal record.
          - Distribute imaging study arrival and update notifications to imaging study subscribers


          ## Prerequisites

          - [Knative](https://operatorhub.io/operator/knative-operator)
          - ProgresSQL such as [Crunchy](https://operatorhub.io/operator/postgresql)
          - S3 storage such as [MinIO](https://operatorhub.io/operator/minio-operator)
          - [NATS operator](https://github.com/nats-io/nats-operator/) (Optional:  only required for DIMSE Proxy component)



          ## Documentation

          Visit the [official github repository](https://github.com/Alvearie/imaging-ingestion) for documentation and examples.
        displayName: Alvearie Imaging Ingestion Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - dicom
        - medical imaging
        - fhir
        - dimse-c
        - dimse
        - imaging study
        links:
        - name: Alvearie Imaging Ingestion Operator
          url: https://github.com/Alvearie/imaging-ingestion
        maintainers:
        - email: rduggan@ca.ibm.com
          name: Richard Duggan
        - email: jjacob@us.ibm.com
          name: Jeesmon Jacob
        maturity: alpha
        provider:
          name: Alvearie
          url: https://alvearie.github.io
        relatedImages:
        - alvearie/imaging-ingestion-operator:0.0.3
        version: 0.0.3
      entries:
      - name: imaging-ingestion-operator.v0.0.3
        version: 0.0.3
      - name: imaging-ingestion-operator.v0.0.2
        version: 0.0.2
      - name: imaging-ingestion-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: alvearie-imaging-ingestion
    provider:
      name: Alvearie
      url: https://alvearie.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-licensing
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-licensing.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Licensing",
                "metadata": {
                  "name": "licensing-sample"
                },
                "spec": {
                  "environment": {
                    "deploymentType": "openshift"
                  },
                  "hostid": {
                    "enabled": true,
                    "environment": {},
                    "hostMounts": {},
                    "image": "prod.harbor.keysight.digital/loadcore/kcos-licensing-hostid-vmuuid@sha256:076a2c6aeec875470414af7905613a25b56bcb2aca5edcb1c4c53037cd01b332",
                    "imagePullPolicy": "IfNotPresent"
                  },
                  "image": {
                    "dbimage": "prod.harbor.keysight.digital/loadcore/postgresql-ubi-image@sha256:17f6a8f775d06650939d2c73e82b45d73d50190cc735ca70695d082f7790834b",
                    "initOwnership": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf",
                    "repository": "prod.harbor.keysight.digital/loadcore"
                  },
                  "ingress": {
                    "annotations": {
                      "nginx.ingress.kubernetes.io/auth-method": "POST",
                      "nginx.ingress.kubernetes.io/auth-response-headers": "X-AUTH-USER",
                      "nginx.ingress.kubernetes.io/auth-url": "https://127.0.0.1/token/validate"
                    },
                    "enabled": true
                  },
                  "nodeSelector": {
                    "node-role.kubernetes.io/kcos-licensing": "kcos-licensing"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:29:04Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Licensing
            name: licensings.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Licensing Operator
        displayName: Keysight KCOS Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Licensing
          url: https://keysight-kcos-licensing.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-licensing@sha256:a09291923ad8d7533b33e40e1a3c4e735c2a4ae3b2b87be7c9d1a1a9244b3dad
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - prod.harbor.keysight.digital/loadcore/postgresql-ubi-image@sha256:17f6a8f775d06650939d2c73e82b45d73d50190cc735ca70695d082f7790834b
        - prod.harbor.keysight.digital/loadcore/kcos-licensing-hostid-vmuuid@sha256:076a2c6aeec875470414af7905613a25b56bcb2aca5edcb1c4c53037cd01b332
        version: 0.1.21
      entries:
      - name: keysight-kcos-licensing.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-licensing
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: multicluster-engine
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-engine.v2.5.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-08-06T12:21:10Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0 <2.5.10'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '[]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine is the Schema for the multiclusterengines
              API
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:80f33a762535ff69f57ecb04de043b8fe249a63767712f4942ac6685450e8f28
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:1af88c3703b668020137f90657e6262b73e407aa858bbc6d9c12ba3a1cab8043
        - registry.redhat.io/multicluster-engine/agent-service-rhel8@sha256:92464c57700dcb13ccf5431c104a3fba6845f651776aa5beb1ce7171087dc1ea
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:cb2c57c51edc8a2320a10ac217cca6a401db09de55c0c9d0c9a7b4545e2636e9
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:7f32ed997e03dba1746b3627d5a33fc69282b2e831edd21ec74020fe4fa1ca5d
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:90e792c92ef8c75ffb97545394206c9197997d474493d93f09fca273f65133e4
        - registry.redhat.io/multicluster-engine/hive-rhel8@sha256:27eefe6ad7d93e752f4d76b776884129615c2e87ff24116c401a886e98abfcc5
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:1e9eff48ea837a1d19c595986d765d74810d6d1397602ad79e2e820b1cbb5c6e
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:297d8d0d5855db90ad40af702d6ad5f3f32a4aab486686919d86bbd21c367ca9
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:b0a70eee404ba1267c0ae1a8eb140251d054e722305d680685223cfda360aa37
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:53b8d48e540c3d9bb5d948af0375c26377271e595fffdb3d5f921e3a358fa298
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:36abf06619cd3971da8f42f4193621a1384970365fce0f48e86a65e2840bca38
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel8@sha256:360d3dfb823d8e199574cf1c3fbf890fa74c71fc48644e22b47339363e941b3e
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:c8d48e7ccdce43ce72154e464e74c32477efb0094aab327ed0eaab6c1244774a
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:515eda1e4e1ced222709103789af14f7b4e99d4da037deddd8499aad3b45451b
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:3e3f7ee898e0f6b0f74026399753d7fde41013ade76f7963d5c25e578eb1906e
        - registry.redhat.io/rhel8/postgresql-12@sha256:da619f6617d99790b4fe00fdb23ab4aea0cdb0592590162d51b3ef94aca01a7b
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel8@sha256:6bf850d8fb250740113bfb004b19edbd5cf560c8175b81e2f3766280b3482e27
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:43cc171404571bb3c87275a86e9ab4409861c1c01b5f1441c455565d56e1921d
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:d71c1441577b1cd360d074581ad19c2b0cfa99cdfe1869af1777ff7de6821735
        - registry.redhat.io/multicluster-engine/cluster-api-provider-aws-rhel9@sha256:7f65119dc5d81b8ba485b651efe1d9013b62ec57e4725dbf8bdc2e3b7748f677
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:1a5ff852d96bd15c631a6f89e2b0047cd8dde6264646f51106f2b523decc64ba
        - registry.redhat.io/multicluster-engine/aws-encryption-provider-rhel9@sha256:22cefc43e205e4bdd7874c7e1fccba60fe556df4b2bab055d01c834d62f194bd
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:53ad9402a0a13c894b485f418dd1353c8870792893a5ce33f9d6c3cc9839ebc4
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:2497e077a00dfa94b913359be0c696126e932db1a45e066ead65dee76a1db8fd
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel8@sha256:1e755b64358c10309804a3bf44ca640a7d5bb6e0474d33774a8d76095f2b8c3d
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:25b30d4f4ccc23fda707db88969c0ae036ecb4a2baf15b82221e3351c23292f9
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:a4f58b94d45c1ae1c4ba69b12030b9945b37f3ce6e96c27919f256ae887ec28b
        - registry.redhat.io/multicluster-engine/cluster-api-rhel9@sha256:0e329ca39b1f18520da093a5194ef6e1d5527270c74bf260517620577a21b6c7
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:19dba8be2295f54a93853dbe566132011f0120b2c4b48de2816509f5d201e420
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:bd6d0189ea2bbc87b5745f9d7f9c82118f8dcc9ee63fe0a36bb5e7dfc0e56021
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:424ca238d9c8b20dc434a07aa6a27fe6e7c63b3e2480eead284570864d682175
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:3d420ba8bc5e1dbe629c0d4aa65c0a79a7914ac349828046994e245f4f2adb2b
        - registry.redhat.io/multicluster-engine/assisted-installer-reporter-rhel8@sha256:5d41ca13f4ae102760b65375578c2f632aa6ce41092732300d6727f195da3d55
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:460d37a0032813aa85a3b18bd6f152e4ed6b89f802d17d787c3365082420d379
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:d7ccc8ddbeb11ac8a3a0d4946920a0ab87d9e9595d7c61740bd57f3f4a19a711
        version: 2.5.10
      entries:
      - name: multicluster-engine.v2.5.10
        version: 2.5.10
      - name: multicluster-engine.v2.5.9
        version: 2.5.9
      - name: multicluster-engine.v2.5.8
        version: 2.5.8
      - name: multicluster-engine.v2.5.7
        version: 2.5.7
      - name: multicluster-engine.v2.5.6
        version: 2.5.6
      - name: multicluster-engine.v2.5.5
        version: 2.5.5
      - name: multicluster-engine.v2.5.4
        version: 2.5.4
      - name: multicluster-engine.v2.5.3
        version: 2.5.3
      - name: multicluster-engine.v2.5.2
        version: 2.5.2
      - name: multicluster-engine.v2.5.1
        version: 2.5.1
      - name: multicluster-engine.v2.5.0-103
        version: 2.5.0-103
      name: stable-2.5
    - currentCSV: multicluster-engine.v2.6.8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-08-04T11:16:16Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0 <2.6.8'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:69a40339bf3b3dd3c3b6359721895313147a5f682b62faf5151288ba63b65812
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:f6b15f827aeabc33074de1ff9ed20edaf04271b4e4d2eb16abbe786c78e033ca
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:9e1ea96b998b0c7854ca3c7c8a505fc96cbc1b35e6ce9f49b1dbec703fdc8070
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:7dde47f7fe16a26676a6e5b1739954d76cad1cddbefce2d6869ffb2af399e403
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:f563ff0acb4ea9d1bb112765b67bee535063304a98a2aa2cb7f21b0155d49262
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:54b8e17cacae2d355d82677ed334b4981b50e825579ff30406039933833e828b
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:a8117c99b33eedc702baeab32f0347892b587dfc1860f467cb1900565c3fdb3b
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:00ea8fe4945e93d2cbeceda4440acd6a53a48b4dd420dfe2cbd5287e20f0362f
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel8@sha256:42fee1066ff00066a172f82754ad9f08e8c29029e75f76712cb283d9fe5d1d9a
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:1d098eab26421f99ef71ffce82e4fa59ef0f9043de04ee74ca43486cbdb82006
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:b39819aa916974f4a408d23177e99cfb71e21af8ff9b0eacf6c37159c354ba70
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:0737e57ae8af6a62ac580710276d2c1620ac1b1283d3b84369db43e76e518610
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:5f9d5bdfd45f931989d069dcb1dea91a69020d83011fe0846ddfe22bba14b001
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel8@sha256:a630a20a633d057e475b4c39b342512119a71021e199aa0becf80899d8840ad7
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:c3c654be0b20cac501af1e31631464c32aecdbd16cbfb88f865900c70ccf77c2
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:7c5f339e762db1714b3f6cf1fc1bc8f08892cac433ab5a46e73b530fa08486fb
        - registry.redhat.io/multicluster-engine/cluster-api-rhel9@sha256:bd4eb05427ba2ea35f01227600cebd0fe8598d4de456348c24bb5e88d9ed8720
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:aff06db527e4ddc713a2fb8885824f0482647cbdcd844b311ccc3880a0dd2c3f
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:5ef8acc612817d478cede3b1bda55abafac00d0debebd68d948d157ee5b09286
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:294928c58b33ae8e776f86df211dac3b63a2cd9fc00aa995a0de037a29cb2c7d
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:2d807348774b8c6c0255c4bc931a956fc987c961fae6be689835410e8e301fd0
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel8@sha256:4d96379ad99d299d2a89eb1b0eb8a22db8bdada9a16248503cd8a9a5900eab35
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:d2c1d90a4e829b181fdf3d2838764c8f2b61c22b761279f53cb22316ad46ca03
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:93d65bfa0102c7ad7ec6ebfaad9ab062a31098b7ad7026313bcd8c6a495d5ca7
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:cd23f237c91dd8585946b31e17e38193f2e5ac8993d65459bb703a68f6fa8ca0
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:80a6615500949cc8c7a5b2a1e8c8c5f4564e9cdb9f995787a44f462e5bbd62cb
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel8@sha256:a703f2b1f4eb538aa02dbc0f4ff793506076480e94905c367b0d4ae352b87b65
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:5d488c8a644f2bfce97444d770516de10ceecc988cb3e9fbf0072d0b0a2eeb83
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:be1aa4cf5e09914774dcf69ab7ccdaf67504309d40c7e8cf4d91e9c1080e706f
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:82c0af2d285ba943dfa1da9f0856b26ec7e4c2c431577fcd4ab031ae040cfaea
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:87f5db53ad18f6d84501a5c2294cff73a5425276cdbde92225909826091b5ffa
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:46c4ac0cacef587b6e678b6cb0f68c29f5ac63bd59209d42a3ae0b7807e99972
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:d7c8bef4c3dec710ccb299482a10bfdf6efacd4e2ed2e6e323d448ed505000b3
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:d71ef80913d4e34389bce56c4a282cf7d2be705f003575dab87c07ca71f8c2d6
        version: 2.6.8
      entries:
      - name: multicluster-engine.v2.6.8
        version: 2.6.8
      - name: multicluster-engine.v2.6.7
        version: 2.6.7
      - name: multicluster-engine.v2.6.6
        version: 2.6.6
      - name: multicluster-engine.v2.6.5
        version: 2.6.5
      - name: multicluster-engine.v2.6.4
        version: 2.6.4
      - name: multicluster-engine.v2.6.3
        version: 2.6.3
      - name: multicluster-engine.v2.6.2
        version: 2.6.2
      - name: multicluster-engine.v2.6.1
        version: 2.6.1
      - name: multicluster-engine.v2.6.0
        version: 2.6.0
      name: stable-2.6
    - currentCSV: multicluster-engine.v2.7.5
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-21T22:32:21Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0 <2.7.5'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:44b07a18116ad827c23ca5a98b9fe8ef74a99abe4787cd01d848de42df44178d
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:15c2873703afcc74e41ae27dbb4761ad19c4bd2594ca6de82d3b6c02752e0ce3
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:816712bb3ccb58b705a85e5f4a55e519316fd8b9d7d9295d1042420ce44a4bad
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:e6a5e71f0aee8a682179cfd8f85b45a2ee04a1de680fc03dabdb659a2d2c0396
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:2a78ab192a554e78f3fc367c47e571f96b650ee9fb5ba0d06674f340f4320d94
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:9d1be5eb4fa357fe6bd97a52a53b3eeb537c431daf7af8a30efc3e0b6c5d5d66
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:2e3fe5e3b39721cc77b7d32b5682300eb60b1bc6940b8e83662b458d88489140
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:a24e9855a11d879d5308e29af6e3b2947365b7e466e7fa12dc57997140753604
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:b124242a55b1b2a2258cb12094617a0e009cf1cf58a60ab1b42d48d4ea2514ac
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:b66f2f4b5e0d77d501f9cc9da483f976f3c0eaf0ca72a5e9c9435c92d58e1388
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:00d415071f4c10575f9e5068868a7f80b0c0fb9c8614da86bc515923bed136a3
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:e2e16ff6213c0274126c9d64d495c0abbe17379e841ab730de66520908f1a190
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:4a0f1d3e696310f7a5bccbb358c538023a28f5bbc23b5743d10fd61b452d45d0
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:041044c08a1d526b349b5ad5fc4c4457f3632b1353530ed925c65ef2cce8e9c5
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:2025e424c11c6fe4a4395945c2e7d124d07cde43bb8d8cbe70286fe6ae345754
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:8c420720a48ea7c3aa096ae1929dcaa979c39e7865c5d43cba899e294745fc94
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:081151ca906b0f1afbcd3ee9c6d9e5bfe234494223c660ef376aa4004b0508f4
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:3defab1abc7395c9f9c22912ec6825be11b120a1416404eb0599e91f7f9c679a
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:c0fb7985620d5bd0dfadd57cb75664ee6777b8dac5a3463967e9b649e9eb836c
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:9e5c66ffdf39ae923ecccf79be538b883f2bdc8019169d55057565f077997d67
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:4fa34ab0aa8854eb5e942f2ed315041bf2026ce0bed86f57125878af0c73dc8b
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:7369bc79a283ff8b34c3189ea83e55f88a00d619cc78f8dad03d302dcb34a550
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:fedc699c0f786fb303d3956769f9c5fca1473ec4bf014ec87c5a4d40e34dcdb3
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:b82b1837dfc24a367f68741dece6d19c89c576706f006ee8d5123a0c6782c492
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:90c19d2aaa50e7cc9184f2f220b568c34c9fdcfdcba314da555885c999078057
        - registry.redhat.io/multicluster-engine/cluster-api-rhel9@sha256:cd8d6e0056fb0436275c12cc6b42f11c0ccd06729c7108e964b03d6c48d063bd
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:0684bcde2c018ae893bd71be8df3448afa09d199415e83285d2c740cff66192d
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:2216b015943eb8d9f25bf4ac37fbad2b7215a5c0003bd7c3b3d97b3b0a5ba708
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:555fcfce226e3b059f8f911d44a48578b898e085b6ad2575479feee71ec8baaf
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:a6736c7cac00f78c0e8877549ae2b74b7207cc348f24f175133db49fd38f240a
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:8e41d489e1def9b540ddd6387a1cc151b24e91a5fa44a74fbbaa0d1978588d57
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:45571dbb55bd213400eec6d2065b40615185bf3e52efdf495ab73201ec94faff
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:1c176688f13052783662115d5cabe8d614b8d9f79507dc2bdd4be1604e2eecc0
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:2cbd8c470dc4b263fd11952aa792094a7550f4b99e10f22186d51621bc8d3b3f
        version: 2.7.5
      entries:
      - name: multicluster-engine.v2.7.5
        version: 2.7.5
      - name: multicluster-engine.v2.7.4
        version: 2.7.4
      - name: multicluster-engine.v2.7.3
        version: 2.7.3
      - name: multicluster-engine.v2.7.2
        version: 2.7.2
      - name: multicluster-engine.v2.7.1
        version: 2.7.1
      - name: multicluster-engine.v2.7.0
        version: 2.7.0
      name: stable-2.7
    - currentCSV: multicluster-engine.v2.8.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-09-08T13:08:30Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0 <2.8.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:f869555beec0875c74f30a88428ac9641029dfed0365d45b23d5cae469cf7d0f
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:7ca7e87615e12b9d7fa75befd05f4f45bdc4bf8ef1b576c69f36a23c66c90cd6
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:988eb4aa01ba0d7e6728b23d043e4c0babe872ac23b6520f4daac714619a360b
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:f809a8774b0e3f891af46931dd4d2c11b95da4e7f5a8aee1814152d9a468d357
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:c5ac848e3b7eb88cdbdbf0b5f4aa96698d852b3a8118cf7146eb7e1948afaeba
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:a7f63077799a62fe6ece76a12346a87b5cbae966ab69df6ec6d951259dbf0e4a
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:c3b62ab1234257d30f41fa1ab33cb319321777a18de8ed751be1fcd48b204f2e
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:318ca0c7e9cd9a7336178519cadbf21eaeb474cd4a8eb5cf411a640d1c77372e
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:5ea86c20383a6932c593e020fd3b7b1a2a8e666bc374bff8b4bca0ee2d4ecc33
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:aa2754d35c8460a2879d4d523ebf32351d36e8fd7886335b5c0d2ca4befb8e8e
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:bee8f0c047cf8dbfa3f28712ce91bea55e926b2edb26c394f18a2d780e142083
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:7c86c89ff57091d4112783d82d1add47f2d56dba584418d5664c198ba048e0c2
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:f708f3e1007ed2b0e420e51d8b7ff9dd9bccaea9234755f4375f5514018211b4
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:3788b2674dda0b2cb4e9ca21bdce0b9e44702d72458b2396ef0acdad37fa866b
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:18b0e86305afe2eae3c26e637755fa9e8ab2abf30bc15e5428467a9a81ec99b0
        - registry.redhat.io/openshift4/ose-aws-cluster-api-controllers-rhel9@sha256:ada8b480a86d0a9e3ecc58f8b1cb53dbd7452b2b9183b3ba2b15a6a58c4436d1
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:f694b6e9e76941b11cc8a15153d9005cffe9bbbf58d0962f8ff546e90a63d93c
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:806b6c956f7bad824f0fe9acc3b534964e8f5f3399a1813f75511bc0f3904414
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:cf457b5a13cdd54c45a7b88f0f150a04e756bfac727c4365cb12f113d904937d
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:5461643a61850d95409d989e81c304053372f1dc5dbba291f3f3386de9ecab1c
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:261b99b93d98825ade0a0d3b9c4e465a785f35d2a310dedd4bbd6e4aa1f33ef9
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:0392c81a404be343d658dbc64153cbc46a64d0adf69aa8024c3f727351102d75
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:8379827635dfc4b51e04145d8048f7dcf9c6956ea1a22371bd22f948ae01b183
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:ee014c576263e19798df98460ebb5f1a21cb6249f09fe8a267b2dfa98b07d1eb
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:d4d11d0aeb3e1c9ff2a80e769d92dce509a0202a2ef1c9d2430030acdbf85bce
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:83a82f9558c406899e573308947c15b1f36de454006bb0e2ac44709f0c2c28d7
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:ea2ba066abd170e1087558c7b701faeb1be2e1667943dc14440f19c2ee5b0c30
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:fb2b8cef9916f387d26e65d9b53c7567ae1433d43d16e879ddb582ddc248eaa3
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:5ad96dad74c6f4967bd5fd397d6697e01b3b25bb3b8bae123a38e96964ddfa17
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:b9afa4d5328f17c7c3ce1dc105d177ecacf3447092d51a50f250a4b0ea692e2c
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:c6cd578123c1c86fcb3305ecca117543106f943cbcc4baf9065683988b2e10ff
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:e5e459e665afc14a1540e300015222b7c1ac896e2bee889283d67c3e19ca01ac
        - registry.redhat.io/openshift4/ose-cluster-api-rhel9@sha256:ec58463239eaa9bd15b31cbcfea0a34fd650201bac6120aa7ddaf72734451e06
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:ab0898d513b9b5ca2f5ca5879adc4e73da3232d515c0647a2a0ec28792264db1
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:0d1dfcb06eaa26de74994891371a5657aded7a59972f006f8c4a3859369877f3
        version: 2.8.3
      entries:
      - name: multicluster-engine.v2.8.3
        version: 2.8.3
      - name: multicluster-engine.v2.8.2
        version: 2.8.2
      - name: multicluster-engine.v2.8.1
        version: 2.8.1
      - name: multicluster-engine.v2.8.0
        version: 2.8.0
      name: stable-2.8
    - currentCSV: multicluster-engine.v2.9.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-30T19:22:43Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0 <2.9.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/open-cluster-management/backplane-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational components for managing multiple Kubernetes-based clusters. The deployment of the multicluster engine components is
              determined based on the configuration that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:8901e984b3b47f0650cddc1c36bdd89af22390f3d4a562d562fde7774d8ba6db
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:f3d65f68538e94260fad54f145433a36366f8c88f9c9f80b7fc5545ba62e1773
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:0a9c0ccb6a75273474fb04b785aa8fc58aec2f9ffbff8c232d359e14892bbae2
        - registry.redhat.io/multicluster-engine/capoa-control-plane-rhel9@sha256:b510a25a6370ca6b8d2f74165814ebfdae0ce9898cc343e73fab6344eebf2bed
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:8db1386d12d40dfbc80054c023897e5bfc0a987f6d07792b0f9f50f59f1a68d8
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:57699ecb529cb46843de847deef3e3757ca252910249d8deac16fcb00c01b636
        - registry.redhat.io/multicluster-engine/capoa-bootstrap-rhel9@sha256:a8b1278a2df9b17cee66813715995385bb3539af05e4df9b5bf7d9ccd508c313
        - registry.redhat.io/multicluster-engine/mce-capi-webhook-config-rhel9@sha256:7f1f70813e6d753b4acb04c7e9d7acb0047ad91f8a9a10e0fe2fff6c4411c8b4
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:a91276fcf5a589ef11f01b39df7d33ceb58ebcb47f95dfab772114bce7c1b4a8
        - registry.redhat.io/openshift4/ose-cluster-api-rhel9@sha256:9a72b8081a57846d942981ced07da86ad8e197309a1bc3c9d53876528b58f8ec
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:6a75df76fb0d9516b13fa6b079bc6c4d49dc6774f0d4f31343abfb8d70edcc23
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:ce0fd7bdc73cd2c84d1a9c5af528916488ca8116dd05d83dacf72f2af4248e65
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:ac43feaf079fa0d250d53cd9685fb68677d888dd091730faa32757dcdf2c95e4
        - registry.redhat.io/rhel8/postgresql-12@sha256:4773a29755ea826c0d3b58a8b8297c1d7388019516195cae9bd815f28ec4bca0
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:5e6594464bf7ddc2a66f127f645d93ad35d8cd95dd70544fa7fcec791f67f45c
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:36a6000b6482966f8bbf86e45c25335c066736ae1f8695a8b68e6b76b7c491ac
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:1b218f0aa0ea1f494094e3bc783ef40968cbff780e22de186d578a1d64c33bf8
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:3c450cb6a3ba6f2e3028856f14967803c1a691d95b8f54bd9d29d93691d6e795
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:8076db8ccd5578cb5aba2d66f53790563d3f42b93a6cfd14c7094d3a25707982
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:ad382d5eb347c448c46495a0a17503d789db4eb1096ad71a588b316b235e437f
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:6b95560a6a6dc1ddd41c76f275f83baec52ec9d79de04bfe54feaed02db123a3
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:ee75e805f0f9d73f073ce870f5d46962c0033d63158017841d17410febfc2c57
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:25ec9a0cc4ac92d4bed75c38fa37ef8d5ff9478cc475049b303d74eb58687a81
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:f8c3898e29f8c0a20c3be92bc4ee5a9443b9fc8218db95ba541fe3e57a89c40d
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:bf343592bf2f4399b5ac61709f56528435990e903812e264cbef845ff4433231
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:9555fd5f577f32a1d8ebe25f0d03e453c8ea925e4a12dbbe1015c58d3c8528b0
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:1cc6fddaf9e55637e90d5c3e598467df8967b8a362f8b8d46286b256bad6dd3f
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:a41cafc65e30deb073e051542702e258901b552ad00e42eacab05d149d6f0565
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:ea57247cf11f6ab966893ce4e6d2948232a4314588af2ca33c1d9689da7482dd
        - registry.redhat.io/openshift4/ose-baremetal-cluster-api-controllers-rhel9@sha256:f855d5ce17d3ac4560afc73c35d03f045f84551fdccc00312c890288a889b9ce
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:72a4ac73bb5e7797329e1856f785751d48815a64e82350a5d694bb5a8e121956
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:cd9dc89b933b394b08fe05ab5f0bbecea727bfb2c29770f11ec1db4aa5740b3c
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:37087ad8678f3e8ec8e2b2d035d3a12087ce9e0c993194a04a9874fd07e80d91
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:cecec1f7d7f0d37cf17fb25ccf34602728f2a09c61af7e212967b4ee28d2dec3
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:d068d73f41d8f8ecbec36752d0a432913bfbe2580f489abefe513c06233e80c9
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:edbb174cb69dd31339a53542d24c4fd00552e41fa09d7e2553baa7a1ad8dd810
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:1bbd134d44839258172e013fe84bfd174d36d6b7232e2cc8a3d8556998181d9d
        - registry.redhat.io/openshift4/ose-aws-cluster-api-controllers-rhel9@sha256:cf13b0658271cd0dbd1ce6dcc350e64c5459ef02ba4dc88a0031de9d6f26ffd1
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:8d189b565e4c3c434988aa4ff0ecad4c5f86b2a94120325c142ff111b25facde
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:902c5b3b4a4e8cb0a14fa5aba8451ec02830d5167ff1b91323ded62255983cd9
        version: 2.9.0
      entries:
      - name: multicluster-engine.v2.9.0
        version: 2.9.0
      name: stable-2.9
    defaultChannel: stable-2.9
    packageName: multicluster-engine
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Jooho Lee
      provider-url: https://github.com/jooho/nfs-provisioner-operator
    name: nfs-provisioner-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfs-provisioner-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.jhouse.com/v1alpha1",
                "kind": "NFSProvisioner",
                "metadata": {
                  "name": "nfsprovisioner-sample"
                },
                "spec": {
                  "nfsImageConfiguration": {
                    "image": "k8s.gcr.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439",
                    "imagePullPolicy": "IfNotPresent"
                  },
                  "scForNFS": "nfs",
                  "scForNFSPvc": "local-sc",
                  "storageSize": "1G"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: quay.io/jooholee/nfs-provisioner-operator@sha256:d9c013967421ec72644a588a155975bf856a856f1ef38ba71dfea306fdb47acd
          createdAt: "2025-08-08T20:23:50Z"
          description: Create and manage NFS Server and Provisioner
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/jooho/nfs-provisioner-operator
          support: NFS
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NFSProvisioner is the Schema for the nfsprovisioners API
            displayName: NFS Provisioner App
            kind: NFSProvisioner
            name: nfsprovisioners.cache.jhouse.com
            version: v1alpha1
        description: "This operator deploy NFS server with local storage and also
          provide provisioner for storageClass.\n### Core Capabilities\n* **NFS Server:**
          Deployed\n* **NFS Provisioner:** Help customers to create PV using StorageClass\n*
          **StorageClass:** Dynamically create PV for requested PVC\n### NFS Provisioner
          Operator Features\n* NFS Server can use localStorage PVC or HostPath on
          the node\n### Must-Gather \n> isv-cli must-gather --image=quay.io/jooholee/nfs-provisioner-operator-must-gather:0.0.1\n"
        displayName: NFS Provisioner Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nfs
        - storage
        links:
        - name: Nfs Provisioner Operator
          url: https://github.com/jooho/nfs-provisioner-operator
        maintainers:
        - email: ljhiyh@gmail.com
          name: Jooho
        maturity: alpha
        provider:
          name: Jooho Lee
          url: https://github.com/jooho/nfs-provisioner-operator
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - quay.io/jooholee/nfs-provisioner-operator@sha256:d9c013967421ec72644a588a155975bf856a856f1ef38ba71dfea306fdb47acd
        version: 0.0.8
      entries:
      - name: nfs-provisioner-operator.v0.0.8
        version: 0.0.8
      - name: nfs-provisioner-operator.v0.0.7
        version: 0.0.7
      - name: nfs-provisioner-operator.v0.0.6
        version: 0.0.6
      - name: nfs-provisioner-operator.v0.0.5
        version: 0.0.5
      - name: nfs-provisioner-operator.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: nfs-provisioner-operator
    provider:
      name: Jooho Lee
      url: https://github.com/jooho/nfs-provisioner-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator-certified.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator-certified.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator-certified.v2.6.1
        version: 2.6.1
      - name: percona-postgresql-operator-certified.v2.6.0
        version: 2.6.0
      - name: percona-postgresql-operator-certified.v2.5.0
        version: 2.5.0
      name: stable
    - currentCSV: percona-postgresql-operator-certified.v2.7.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        version: 2.7.0-cw
      entries:
      - name: percona-postgresql-operator-certified.v2.7.0-cw
        version: 2.7.0-cw
      - name: percona-postgresql-operator-certified.v2.6.1-cw
        version: 2.6.1-cw
      - name: percona-postgresql-operator-certified.v2.6.0-cw
        version: 2.6.0-cw
      - name: percona-postgresql-operator-certified.v2.5.0-cw
        version: 2.5.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-postgresql-operator-certified
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Starburst Data, Inc.
      provider-url: https://www.starburstdata.com
    name: starburst-enterprise-helm-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: starburst-enterprise-helm-operator.v402.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstEnterprise",
                "metadata": {
                  "name": "starburstenterprise-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "cache": {
                    "diskUsagePercentage": 80,
                    "enabled": false,
                    "ttl": "7d",
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "catalogs": {
                    "tpch": "connector.name=tpch\ntpch.splits-per-node=4"
                  },
                  "commonLabels": {},
                  "coordinator": {
                    "additionalProperties": "",
                    "affinity": {},
                    "deploymentAnnotations": {},
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=true\nnode-scheduler.include-coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://localhost:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "resources": {
                      "cpu": 16,
                      "memory": "60Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "tolerations": []
                  },
                  "environment": null,
                  "expose": {
                    "clusterIp": {
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      }
                    },
                    "ingress": {
                      "annotations": {},
                      "host": null,
                      "ingressClassName": null,
                      "ingressName": "coordinator-ingress",
                      "path": "/",
                      "pathType": "ImplementationSpecific",
                      "serviceName": "starburst",
                      "servicePort": 8080,
                      "tls": {
                        "enabled": true,
                        "secretName": null
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "nodePort": 30080,
                          "port": 8080
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "externalSecrets": {
                    "enabled": false,
                    "goDaddy": {
                      "backendType": "secretsManager"
                    },
                    "secretPrefix": "external0",
                    "type": "goDaddy"
                  },
                  "extraArguments": [],
                  "extraSecret": {
                    "file": null,
                    "name": null
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise",
                    "tag": "402-e.1"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "initImage": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise-init",
                    "tag": "1.3.2"
                  },
                  "internal": {
                    "ports": {
                      "http": {
                        "port": 8080
                      },
                      "https": {
                        "port": 8443
                      }
                    }
                  },
                  "internalTls": false,
                  "licenseManager": {
                    "cronjob": {
                      "affinity": {},
                      "nodeSelector": {},
                      "priorityClassName": null,
                      "securityContext": {},
                      "tolerations": []
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "registry.connect.redhat.com/starburst/license-manager",
                      "tag": "0.0.2"
                    },
                    "url": "https://license-api.dogfood.eng.starburstdata.net/test/license/checkout"
                  },
                  "livenessProbe": null,
                  "prometheus": {
                    "agent": {
                      "config": "/etc/starburst/telemetry/prometheus.yaml",
                      "port": 8081,
                      "version": "0.16.1"
                    },
                    "enabled": true,
                    "rules": [
                      {
                        "attrNameSnakeCase": true,
                        "name": "$1",
                        "pattern": "trino.execution<name=QueryManager><>(running_queries|queued_queries)",
                        "type": "GAUGE"
                      },
                      {
                        "name": "failed_queries",
                        "pattern": "trino.execution<name=QueryManager><>FailedQueries\\.TotalCount",
                        "type": "COUNTER"
                      }
                    ]
                  },
                  "query": {
                    "maxConcurrentQueries": 3
                  },
                  "readinessProbe": null,
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "securityContext": {},
                  "sharedSecret": null,
                  "siac": {
                    "additionalHeapSizePercentage": 15,
                    "enabled": false
                  },
                  "spilling": {
                    "enabled": false,
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "startupProbe": null,
                  "usageMetrics": {
                    "enabled": false,
                    "usageClient": {
                      "initialDelay": "1m",
                      "interval": "1m"
                    }
                  },
                  "userDatabase": {
                    "enabled": false,
                    "users": [
                      {
                        "password": "46991b33f7a75ff79213c0dc0e610610",
                        "username": "admin"
                      }
                    ]
                  },
                  "worker": {
                    "additionalProperties": "",
                    "affinity": {},
                    "autoscaling": {
                      "enabled": false,
                      "maxReplicas": 100,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "deploymentAnnotations": {},
                    "deploymentTerminationGracePeriodSeconds": 300,
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://coordinator:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "kedaScaler": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.connect.redhat.com/starburst/keda-trino-scaler",
                        "tag": "0.1.7"
                      },
                      "port": 8021,
                      "scaledObjectSpec": {
                        "advanced": {},
                        "cooldownPeriod": 300,
                        "idleReplicaCount": 0,
                        "maxReplicaCount": 100,
                        "minReplicaCount": 1,
                        "pollingInterval": 30,
                        "scaleTargetRef": {
                          "name": "worker"
                        },
                        "triggers": [
                          {
                            "metadata": {
                              "numberOfQueriesPerWorker": "10",
                              "scaleInToIdleReplicaCountIfNoQueuedQueriesLeft": "true",
                              "scaleMethod": "query_queue"
                            },
                            "type": "external"
                          }
                        ]
                      }
                    },
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "replicas": 2,
                    "resources": {
                      "cpu": 16,
                      "memory": "100Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "starburstWorkerShutdownGracePeriodSeconds": 120,
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstHive",
                "metadata": {
                  "name": "starbursthive-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "affinity": {},
                  "commonLabels": {},
                  "database": {
                    "external": {
                      "driver": null,
                      "jdbcUrl": null,
                      "password": null,
                      "user": null
                    },
                    "internal": {
                      "databaseName": "hive",
                      "databasePassword": "HivePass1234",
                      "databaseUser": "hive",
                      "driver": "org.postgresql.Driver",
                      "env": [],
                      "envFrom": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.redhat.io/rhel8/postgresql-10",
                        "tag": "1"
                      },
                      "port": 5432,
                      "resources": {
                        "limits": {
                          "cpu": 2,
                          "memory": "1Gi"
                        },
                        "requests": {
                          "cpu": 2,
                          "memory": "1Gi"
                        }
                      },
                      "volume": {
                        "emptyDir": {}
                      }
                    },
                    "type": "internal"
                  },
                  "deploymentAnnotations": {},
                  "env": [],
                  "envFrom": [],
                  "expose": {
                    "clusterIp": {
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "nodePort": 30083,
                          "port": 9083
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "extraArguments": [],
                  "hdfs": {
                    "hadoopUserName": null
                  },
                  "heapSizePercentage": 85,
                  "hiveMetastoreStorageAuthorization": true,
                  "hiveMetastoreWarehouseDir": null,
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/hive",
                    "tag": "3.1.2-e.22"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "nodeSelector": {},
                  "objectStorage": {
                    "adl": {
                      "oauth2": {
                        "clientId": null,
                        "credential": null,
                        "refreshUrl": null
                      }
                    },
                    "awsS3": {
                      "accessKey": null,
                      "endpoint": null,
                      "pathStyleAccess": false,
                      "region": null,
                      "secretKey": null
                    },
                    "azure": {
                      "abfs": {
                        "accessKey": {
                          "accessKey": null,
                          "storageAccount": null
                        },
                        "authType": "accessKey",
                        "oauth": {
                          "clientId": null,
                          "endpoint": null,
                          "secret": null
                        }
                      },
                      "wasb": {
                        "accessKey": null,
                        "storageAccount": null
                      }
                    },
                    "gs": {
                      "cloudKeyFileSecret": null
                    }
                  },
                  "podAnnotations": {},
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "resources": {
                    "limits": {
                      "cpu": 1,
                      "memory": "1Gi"
                    },
                    "requests": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "securityContext": {},
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
          description: Deploys StarburstEnterprise Helm Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Starburst Data
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: StarburstEnterprise
            name: starburstenterprises.charts.starburstdata.com
            version: v1alpha1
          - kind: StarburstHive
            name: starbursthives.charts.starburstdata.com
            version: v1alpha1
        description: CRs for deploying starburst enterprise and hive
        displayName: Starburst Enterprise Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - starburst
        - trino
        - sep
        - bigdata
        - data
        links:
        - name: Starburst Enterprise Helm Operator
          url: https://www.starburstdata.com
        maintainers:
        - email: hello@starburstdata.com
          name: Starburst Data, Inc.
        maturity: alpha
        provider:
          name: Starburst Data, Inc.
          url: https://www.starburstdata.com
        relatedImages:
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
        version: 402.1.0
      entries:
      - name: starburst-enterprise-helm-operator.v402.1.0
        version: 402.1.0
      name: alpha
    defaultChannel: alpha
    packageName: starburst-enterprise-helm-operator-rhmp
    provider:
      name: Starburst Data, Inc.
      url: https://www.starburstdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: keystone-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keystone-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keystone.openstack.org/v1beta1",
                "kind": "KeystoneAPI",
                "metadata": {
                  "name": "keystone"
                },
                "spec": {
                  "adminProject": "admin",
                  "adminUser": "admin",
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "keystone",
                  "databaseInstance": "openstack",
                  "preserveJobs": false,
                  "region": "regionOne",
                  "resources": {
                    "requests": {
                      "cpu": "1.0",
                      "memory": "500Mi"
                    }
                  },
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "keystone.openstack.org/v1beta1",
                "kind": "KeystoneEndpoint",
                "metadata": {
                  "name": "placement"
                },
                "spec": {
                  "endpoints": {
                    "admin": "http://placement-admin-openstack.apps-crc.testing",
                    "internal": "http://placement-internal-openstack.apps-crc.testing",
                    "public": "http://placement-public-openstack.apps-crc.testing"
                  },
                  "serviceName": "placement"
                }
              },
              {
                "apiVersion": "keystone.openstack.org/v1beta1",
                "kind": "KeystoneService",
                "metadata": {
                  "name": "placement"
                },
                "spec": {
                  "enabled": true,
                  "passwordSelector": "PlacementPassword",
                  "secret": "osp-secret",
                  "serviceDescription": "Placement service",
                  "serviceName": "placement",
                  "serviceType": "placement",
                  "serviceUser": "placement"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:01Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KeystoneAPI is the Schema for the keystoneapis API
            displayName: Keystone API
            kind: KeystoneAPI
            name: keystoneapis.keystone.openstack.org
            version: v1beta1
          - description: KeystoneEndpoint is the Schema for the keystoneendpoints
              API
            displayName: Keystone Endpoint
            kind: KeystoneEndpoint
            name: keystoneendpoints.keystone.openstack.org
            version: v1beta1
          - description: KeystoneService is the Schema for the keystoneservices API
            displayName: Keystone Service
            kind: KeystoneService
            name: keystoneservices.keystone.openstack.org
            version: v1beta1
        description: Keystone Operator
        displayName: Keystone Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Identity
        - Keystone
        links:
        - name: Keystone Operator
          url: https://github.com/openstack-k8s-operators/keystone-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-keystone-rhel9@sha256:740e92a78d1e3874d2a40e37c970c6163d9e295dbc236914f1204baed070c595
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/keystone-rhel9-operator@sha256:8b055945a3808f7397265a10e6a6463af0ee3ef80fdff9d66402b488e5b70523
        version: 1.0.6
      entries:
      - name: keystone-operator.v1.0.6
        version: 1.0.6
      - name: keystone-operator.v1.0.5
        version: 1.0.5
      - name: keystone-operator.v1.0.4
        version: 1.0.4
      - name: keystone-operator.v1.0.3
        version: 1.0.3
      - name: keystone-operator.v1.0.2
        version: 1.0.2
      - name: keystone-operator.v1.0.1
        version: 1.0.1
      - name: keystone-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: keystone-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: ""
    name: ingress-node-firewall
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ingress-node-firewall.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ingressnodefirewall.openshift.io/v1alpha1",
                "kind": "IngressNodeFirewall",
                "metadata": {
                  "name": "ingressnodefirewall-demo2"
                },
                "spec": {
                  "ingress": [
                    {
                      "rules": [
                        {
                          "action": "Deny",
                          "order": 10,
                          "protocolConfig": {
                            "icmp": {
                              "icmpType": 8
                            },
                            "protocol": "ICMP"
                          }
                        },
                        {
                          "action": "Deny",
                          "order": 20,
                          "protocolConfig": {
                            "protocol": "TCP",
                            "tcp": {
                              "ports": "8000-9000"
                            }
                          }
                        }
                      ],
                      "sourceCIDRs": [
                        "172.16.0.0/12"
                      ]
                    },
                    {
                      "rules": [
                        {
                          "action": "Deny",
                          "order": 10,
                          "protocolConfig": {
                            "icmpv6": {
                              "icmpType": 128
                            },
                            "protocol": "ICMPv6"
                          }
                        }
                      ],
                      "sourceCIDRs": [
                        "fc00:f853:ccd:e793::0/64"
                      ]
                    }
                  ],
                  "interfaces": [
                    "eth0"
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "do-node-ingress-firewall": "true"
                    }
                  }
                }
              },
              {
                "apiVersion": "ingressnodefirewall.openshift.io/v1alpha1",
                "kind": "IngressNodeFirewallConfig",
                "metadata": {
                  "name": "ingressnodefirewallconfig",
                  "namespace": "ingress-node-firewall-system"
                },
                "spec": {
                  "nodeSelector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ingress-node-firewall-rhel9-operator@sha256:144403b009d1b5f2b2ee25ed3cebc525f3588f21bfa671b25c662cf800798839
          createdAt: "2024-03-13T12:36:01Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.11.0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-ingress-node-firewall
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/ingress-node-firewall
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: IngressNodeFirewallConfig
            name: ingressnodefirewallconfigs.ingressnodefirewall.openshift.io
            version: v1alpha1
          - kind: IngressNodeFirewallNodeState
            name: ingressnodefirewallnodestates.ingressnodefirewall.openshift.io
            version: v1alpha1
          - kind: IngressNodeFirewall
            name: ingressnodefirewalls.ingressnodefirewall.openshift.io
            version: v1alpha1
        description: The Ingress node firewall operator, manages the configuration
          of cluster scoped Ingress Node firewall rules for Kubernetes.
        displayName: Ingress Node Firewall Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ingressnodefirewall
        links:
        - name: Ingress Node Firewall
          url: https://github.com/openshift/ingress-node-firewall
        maintainers:
        - email: mmahmoud@redhat.com
          name: msherif1234
        maturity: alpha
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ingress-node-firewall-rhel9@sha256:4e3d69793f9280f6910a6749e80e8f6fe5d20c7a2c160d5682f310309c497e64
        - registry.redhat.io/openshift4/ingress-node-firewall-rhel9-operator@sha256:144403b009d1b5f2b2ee25ed3cebc525f3588f21bfa671b25c662cf800798839
        version: 4.16.0-202509240123
      entries:
      - name: ingress-node-firewall.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: ingress-node-firewall.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: ingress-node-firewall.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: ingress-node-firewall.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: ingress-node-firewall.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: ingress-node-firewall.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: ingress-node-firewall.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: ingress-node-firewall.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: ingress-node-firewall.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: ingress-node-firewall.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: ingress-node-firewall.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: ingress-node-firewall.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: ingress-node-firewall.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: ingress-node-firewall.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: ingress-node-firewall.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: ingress-node-firewall.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: ingress-node-firewall.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: ingress-node-firewall.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: ingress-node-firewall.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: ingress-node-firewall.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: ingress-node-firewall.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: ingress-node-firewall.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: ingress-node-firewall.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: ingress-node-firewall.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: ingress-node-firewall.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: ingress-node-firewall.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: ingress-node-firewall.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: ingress-node-firewall.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: ingress-node-firewall.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: ingress-node-firewall.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: ingress-node-firewall.v4.16.0-202405281506
        version: 4.16.0-202405281506
      name: stable
    defaultChannel: stable
    packageName: ingress-node-firewall
    provider:
      name: Red Hat, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cluster-observability-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-observability-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e
          createdAt: "2025-06-24T11:30:26Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=..0 <1.2.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-cluster-observability-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs",
              "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs",
              "prometheusagents.monitoring.rhobs",
              "perses.perses.dev"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
          support: Cluster Observability (https://issues.redhat.com/projects/COO/)
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: Perses is the Schema for the perses API
            displayName: Perses
            kind: Perses
            name: perses.perses.dev
            version: v1alpha1
          - description: A Perses Dashboard
            displayName: Perses Dashboard
            kind: PersesDashboard
            name: persesdashboards.perses.dev
            version: v1alpha1
          - description: A Perses Datasource
            displayName: Perses Datasource
            kind: PersesDatasource
            name: persesdatasources.perses.dev
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
          - description: UIPlugin defines a console plugin for observability
            displayName: UIPlugin
            kind: UIPlugin
            name: uiplugins.observability.openshift.io
            version: v1alpha1
        description: |-
          Cluster Observability Operator is a Go based Kubernetes operator to easily setup and manage various observability tools.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          - Setup console plugins
          - Setup korrel8r
          - Setup Perses
          - Setup Cluster Health Analyzer
          ### Documentation
          - **[Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/cluster_observability_operator/index)**
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)
        displayName: Cluster Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad
        - registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf
        - registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd
        - registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a
        - registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552
        - registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:e2681bce57dc9c15701f5591532c2dfe8f19778606661339553a28dc003dbca5
        - registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e
        - registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:e54c1e1301be66933f3ecb01d5a0ca27f58aabfd905b18b7d057bbf23bdb7b0d
        - registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:8597c48fc71fc6ec8e87dbe40dace4dbb7b817c1039db608af76a0d90f7ac2d0
        - registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404
        - registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223
        - registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00
        - registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1
        - registry.redhat.io/cluster-observability-operator/perses-0-1-rhel9-operator@sha256:bfed9f442aea6e8165644f1dc615beea06ec7fd84ea3f8ca393a63d3627c6a7c
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed
        version: 1.2.2
      entries:
      - name: cluster-observability-operator.v1.2.2
        version: 1.2.2
      - name: cluster-observability-operator.v1.2.1
        version: 1.2.1
      - name: cluster-observability-operator.v1.2.0
        version: 1.2.0
      - name: cluster-observability-operator.v1.1.1
        version: 1.1.1
      - name: cluster-observability-operator.v1.1.0
        version: 1.1.0
      - name: cluster-observability-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: cluster-observability-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JohnSnowLabs
      provider-url: ""
    name: nlp-server
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nlp-server.v0.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.johnsnowlabs.com/v1",
                "kind": "Nlpserver",
                "metadata": {
                  "name": "nlpserver-sample"
                },
                "spec": {
                  "ExtraEnv": {
                    "johnsnowlabs_client_id": "sI4MKSmLHOX2Pg7XhM3McJS2oyKG5PHcp0BlANEW",
                    "nb_uid": "1000",
                    "nb_user": "johnsnowlabs",
                    "pyspark_version": "3.0.2",
                    "spark_nlp_version": "3.4.4"
                  },
                  "affinity": {},
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "johnsnowlabs/nlp-server",
                    "tag": "0.8.0rc1"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {
                      "kubernetes.io/ingress.class": "nginx",
                      "nginx.ingress.kubernetes.io/ssl-redirect": "true"
                    },
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "nlp-server.jsl.aiplatform.eks.test.johnsnowlabs.com",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": [
                      {
                        "hosts": [
                          "nlp-server.jsl.aiplatform.eks.test.johnsnowlabs.com"
                        ],
                        "secretName": "annotationlab-tls-secret"
                      }
                    ]
                  },
                  "livenessProbe": "httpGet:\n  path: /\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 3\n",
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "readinessProbe": "httpGet:\n  path: /\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 3\n",
                  "replicaCount": 1,
                  "resources": {},
                  "securityContext": {},
                  "service": {
                    "port": 5000,
                    "portName": "http",
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": "nlp-server-sa"
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.johnsnowlabs.com/nlp-server/nlp-server-operator:0.8.6
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nlpserver
            name: nlpservers.apps.johnsnowlabs.com
            version: v1
        description: NLPServer from JohnSnow labs
        displayName: nlp-server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nlp-server
        links:
        - name: Nlp Server
          url: https://johnsnowlabs.com
        maturity: alpha
        provider:
          name: JohnSnowLabs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - registry.johnsnowlabs.com/nlp-server/nlp-server-operator:0.8.6
        version: 0.8.6
      entries:
      - name: nlp-server.v0.8.6
        version: 0.8.6
      name: alpha
    defaultChannel: alpha
    packageName: nlp-server
    provider:
      name: JohnSnowLabs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Seldon Technologies
      provider-url: ""
    name: seldon-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: seldon-operator.v1.14.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1",
                "kind": "SeldonDeployment",
                "metadata": {
                  "labels": {
                    "app": "seldon",
                    "app.kubernetes.io/instance": "seldon1",
                    "app.kubernetes.io/name": "seldon",
                    "app.kubernetes.io/version": "v0.5"
                  },
                  "name": "seldon-model"
                },
                "spec": {
                  "name": "test-deployment",
                  "predictors": [
                    {
                      "componentSpecs": [
                        {
                          "spec": {
                            "containers": [
                              {
                                "image": "seldonio/mock_classifier:1.6.0",
                                "name": "classifier"
                              }
                            ]
                          }
                        }
                      ],
                      "graph": {
                        "children": [],
                        "name": "classifier",
                        "type": "MODEL"
                      },
                      "name": "example",
                      "replicas": 1
                    }
                  ]
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: docker.io/seldonio/seldon-core-operator:1.14.1
          createdAt: "2019-05-21 15:00:00"
          description: The Seldon operator for management, monitoring and operations
            of machine learning systems through the Seldon Engine. Once installed,
            the Seldon Operator provides multiple functions which facilitate the productisation,
            monitoring and maintenance of machine learning systems at scale.
          olm.skipRange: <1.13.1
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/SeldonIO/seldon-core
          support: Clive Cox
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A seldon engine deployment
            displayName: Seldon Deployment
            kind: SeldonDeployment
            name: seldondeployments.machinelearning.seldon.io
            version: v1
        description: "The Seldon operator enables for native operation of production
          machine learning workloads, including monitoring and operations of language-agnostic
          models with the benefits of real-time metrics and log analysis.\n   \n##
          Overview\nSeldon Core is an open source platform for deploying machine learning
          models on a Kubernetes cluster.\n\n* Deploy machine learning models in the
          cloud or on-premise.\n* Get metrics and ensure proper governance and compliance
          for your running machine learning models.\n* Create powerful inference graphs
          made up of multiple components.\n* Provide a consistent serving layer for
          models built using heterogeneous ML toolkits.\n\nYou can get started by
          following the guides in our documentation at https://docs.seldon.io/projects/seldon-core/en/latest/workflow/README.html\n"
        displayName: Seldon Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - aiops
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://docs.seldon.io/projects/seldon-core/en/latest/
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Seldon Technologies
        relatedImages:
        - docker.io/seldonio/seldon-core-operator:1.14.1
        - ""
        version: 1.14.1
      entries:
      - name: seldon-operator.v1.14.1
        version: 1.14.1
      name: stable
    defaultChannel: stable
    packageName: seldon-operator
    provider:
      name: Seldon Technologies
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: keycloak-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keycloak-operator.v19.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "False"
          containerImage: quay.io/keycloak/keycloak-operator:19.0.3-legacy
          createdAt: "2022-10-07 17:00:00"
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
          * Install Extensions
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/docs/latest/server_installation/index.html#_operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@lists.jboss.org
          name: Keycloak DEV mailing list
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/keycloak/keycloak-operator:19.0.3-legacy
        version: 19.0.3
      entries:
      - name: keycloak-operator.v19.0.3
        version: 19.0.3
      - name: keycloak-operator.v19.0.2
        version: 19.0.2
      - name: keycloak-operator.v19.0.1
        version: 19.0.1
      - name: keycloak-operator.v19.0.0
        version: 19.0.0
      - name: keycloak-operator.v18.0.2
        version: 18.0.2
      - name: keycloak-operator.v18.0.1
        version: 18.0.1
      - name: keycloak-operator.v18.0.0
        version: 18.0.0
      - name: keycloak-operator.v17.0.1
        version: 17.0.1
      - name: keycloak-operator.v17.0.0
        version: 17.0.0
      - name: keycloak-operator.v16.1.0
        version: 16.1.0
      - name: keycloak-operator.v16.0.0
        version: 16.0.0
      - name: keycloak-operator.v15.1.1
        version: 15.1.1
      - name: keycloak-operator.v15.1.0
        version: 15.1.0
      - name: keycloak-operator.v15.0.2
        version: 15.0.2
      - name: keycloak-operator.v15.0.1
        version: 15.0.1
      - name: keycloak-operator.v15.0.0
        version: 15.0.0
      - name: keycloak-operator.v14.0.0
        version: 14.0.0
      name: alpha
    - currentCSV: keycloak-operator.v20.0.0-alpha.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "False"
          containerImage: quay.io/keycloak/keycloak-operator:19.0.3
          createdAt: 10/07/22 17:00:00
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/docs/latest/server_installation/index.html#_operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@lists.jboss.org
          name: Keycloak DEV mailing list
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/keycloak/keycloak-operator:19.0.3
        version: 20.0.0-alpha.7
      entries:
      - name: keycloak-operator.v20.0.0-alpha.7
        version: 20.0.0-alpha.7
      - name: keycloak-operator.v20.0.0-alpha.6
        version: 20.0.0-alpha.6
      - name: keycloak-operator.v20.0.0-alpha.5
        version: 20.0.0-alpha.5
      - name: keycloak-operator.v20.0.0-alpha.4
        version: 20.0.0-alpha.4
      - name: keycloak-operator.v20.0.0-alpha.3
        version: 20.0.0-alpha.3
      - name: keycloak-operator.v20.0.0-alpha.2
        version: 20.0.0-alpha.2
      - name: keycloak-operator.v20.0.0-alpha.1
        version: 20.0.0-alpha.1
      name: candidate
    - currentCSV: keycloak-operator.v26.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/keycloak/keycloak-operator:26.4.0
          createdAt: "2025-09-30T11:58:06Z"
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/keycloak/keycloak-operator:26.4.0
        version: 26.4.0
      entries:
      - name: keycloak-operator.v26.4.0
        version: 26.4.0
      - name: keycloak-operator.v26.3.5
        version: 26.3.5
      - name: keycloak-operator.v26.3.4
        version: 26.3.4
      - name: keycloak-operator.v26.3.3
        version: 26.3.3
      - name: keycloak-operator.v26.3.2
        version: 26.3.2
      - name: keycloak-operator.v26.3.1
        version: 26.3.1
      - name: keycloak-operator.v26.3.0
        version: 26.3.0
      - name: keycloak-operator.v26.2.5
        version: 26.2.5
      - name: keycloak-operator.v26.2.4
        version: 26.2.4
      - name: keycloak-operator.v26.2.3
        version: 26.2.3
      - name: keycloak-operator.v26.2.2
        version: 26.2.2
      - name: keycloak-operator.v26.2.1
        version: 26.2.1
      - name: keycloak-operator.v26.2.0
        version: 26.2.0
      - name: keycloak-operator.v26.1.5
        version: 26.1.5
      - name: keycloak-operator.v26.1.4
        version: 26.1.4
      - name: keycloak-operator.v26.1.3
        version: 26.1.3
      - name: keycloak-operator.v26.1.2
        version: 26.1.2
      - name: keycloak-operator.v26.1.1
        version: 26.1.1
      - name: keycloak-operator.v26.1.0
        version: 26.1.0
      - name: keycloak-operator.v26.0.8
        version: 26.0.8
      - name: keycloak-operator.v26.0.7
        version: 26.0.7
      - name: keycloak-operator.v26.0.6
        version: 26.0.6
      - name: keycloak-operator.v26.0.5
        version: 26.0.5
      - name: keycloak-operator.v26.0.4
        version: 26.0.4
      - name: keycloak-operator.v26.0.2
        version: 26.0.2
      - name: keycloak-operator.v26.0.1
        version: 26.0.1
      - name: keycloak-operator.v26.0.0
        version: 26.0.0
      - name: keycloak-operator.v25.0.6
        version: 25.0.6
      - name: keycloak-operator.v25.0.5
        version: 25.0.5
      - name: keycloak-operator.v25.0.4
        version: 25.0.4
      - name: keycloak-operator.v25.0.2
        version: 25.0.2
      - name: keycloak-operator.v25.0.1
        version: 25.0.1
      - name: keycloak-operator.v25.0.0
        version: 25.0.0
      - name: keycloak-operator.v24.0.5
        version: 24.0.5
      - name: keycloak-operator.v24.0.4
        version: 24.0.4
      - name: keycloak-operator.v24.0.3
        version: 24.0.3
      - name: keycloak-operator.v24.0.2
        version: 24.0.2
      - name: keycloak-operator.v24.0.1
        version: 24.0.1
      - name: keycloak-operator.v24.0.0
        version: 24.0.0
      - name: keycloak-operator.v23.0.7
        version: 23.0.7
      - name: keycloak-operator.v23.0.6
        version: 23.0.6
      - name: keycloak-operator.v23.0.5
        version: 23.0.5
      - name: keycloak-operator.v23.0.4
        version: 23.0.4
      - name: keycloak-operator.v23.0.3
        version: 23.0.3
      - name: keycloak-operator.v23.0.2
        version: 23.0.2
      - name: keycloak-operator.v23.0.1
        version: 23.0.1
      - name: keycloak-operator.v23.0.0
        version: 23.0.0
      - name: keycloak-operator.v22.0.5
        version: 22.0.5
      - name: keycloak-operator.v22.0.4
        version: 22.0.4
      - name: keycloak-operator.v22.0.3
        version: 22.0.3
      - name: keycloak-operator.v22.0.2
        version: 22.0.2
      - name: keycloak-operator.v22.0.1
        version: 22.0.1
      - name: keycloak-operator.v22.0.0
        version: 22.0.0
      - name: keycloak-operator.v21.1.2
        version: 21.1.2
      - name: keycloak-operator.v21.1.1
        version: 21.1.1
      - name: keycloak-operator.v21.1.0
        version: 21.1.0
      - name: keycloak-operator.v21.0.2
        version: 21.0.2
      - name: keycloak-operator.v21.0.1
        version: 21.0.1
      - name: keycloak-operator.v21.0.0
        version: 21.0.0
      - name: keycloak-operator.v20.0.3
        version: 20.0.3
      - name: keycloak-operator.v20.0.2
        version: 20.0.2
      - name: keycloak-operator.v20.0.1
        version: 20.0.1
      - name: keycloak-operator.v20.0.0
        version: 20.0.0
      name: fast
    defaultChannel: fast
    packageName: keycloak-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: redis-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"Redis","metadata":{"name":"redis"},"spec":{"mode":"cluster","size":3,"global":{"image":"quay.io/opstree/redis:v6.2","imagePullPolicy":"IfNotPresent","password":"Opstree@1234","resources":{"requests":{"cpu":"100m","memory":"128Mi"},"limits":{"cpu":"100m","memory":"128Mi"}}},"master":{"service":{"type":"ClusterIP"},"redisConfig":{}},"slave":{"service":{"type":"ClusterIP"},"redisConfig":{}},"service":{"type":"ClusterIP"},"redisConfig":{},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:1.0","imagePullPolicy":"Always","resources":{"requests":{"cpu":"100m","memory":"128Mi"},"limits":{"cpu":"100m","memory":"128Mi"}}},"storage":{"volumeClaimTemplate":{"spec":{"storageClassName":"buildpiper-storage","accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}}}}]'
          capabilities: Seamless Upgrades
          categories: Database, Storage
          certified: "false"
          containerImage: quay.io/opstree/redis-operator:v0.4.0
          createdAt: 06-02-2021
          description: A Golang based redis operator that will make/oversee Redis
            standalone/cluster mode setup on top of the Kubernetes.
          repository: https://github.com/ot-container-kit/redis-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Redis
            displayName: Redis
            kind: Redis
            name: redis.redis.redis.opstreelabs.in
            version: v1beta1
        description: |+
          A Golang based redis operator that will make/oversee Redis standalone/cluster mode setup on top of the Kubernetes. It can create a redis cluster setup with best practices on Cloud as well as the Bare metal environment. Also, it provides an in-built monitoring capability using redis-exporter.

          ### Supported Features

          Here the features which are supported by this operator:-
          - Redis cluster/standalone mode setup
          - Inbuilt monitoring with prometheus exporter
          - Dynamic storage provisioning with pvc template
          - Resources restrictions with k8s requests and limits
          - Password/Password-less setup
          - Node selector and affinity
          - Priority class to manage setup priority
          - SecurityContext to manipulate kernel parameters
          ### Documentation

          [Documentation](https://docs.opstreelabs.in/redis-operator/)

          ### Getting Help

          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.

          ### Contributing

          You can contribute to this project by:-
          - Raising Bugs and Feature Requests
          - Fixing issue by opening a Pull Request
          - Improving Documentation

          ### License

          Logging Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/redis-operator/blob/master/LICENSE)


        displayName: Redis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - Redis
        - Opstree
        - CRD
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://docs.opstreelabs.in/redis-operator/
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: beta
        minKubeVersion: 1.11.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/opstree/redis-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: redis-operator.v0.4.0
        version: 0.4.0
      - name: redis-operator.v0.3.0
        version: 0.3.0
      name: preview
    - currentCSV: redis-operator.v0.15.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisCluster","metadata":{"name":"redis-cluster"},"spec":{"clusterSize":3,"clusterVersion":"v7","securityContext":{"runAsUser":1000,"fsGroup":1000},"persistenceEnabled":true,"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"Redis","metadata":{"name":"redis-standalone"},"spec":{"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"IfNotPresent"},"securityContext":{"runAsUser":1000,"fsGroup":1000}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisSentinel","metadata":{"name":"redis-sentinel"},"spec":{"clusterSize":3,"podSecurityContext":{"runAsUser":1000,"fsGroup":1000},"redisSentinelConfig":{"redisReplicationName":"redis-replication"},"kubernetesConfig":{"image":"quay.io/opstree/redis-sentinel:v7.0.7","imagePullPolicy":"IfNotPresent","resources":{"requests":{"cpu":"101m","memory":"128Mi"},"limits":{"cpu":"101m","memory":"128Mi"}}}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisReplication","metadata":{"name":"redis-replication"},"spec":{"clusterSize":3,"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"Always"},"podSecurityContext":{"runAsUser":1000,"fsGroup":1000}}}]'
          capabilities: Seamless Upgrades
          categories: Database, Storage
          certified: "false"
          containerImage: quay.io/opstree/redis-operator:v0.15.1
          createdAt: 24-09-2023
          description: A Golang based redis operator that will make/oversee Redis
            standalone/cluster mode setup on top of the Kubernetes.
          repository: https://github.com/ot-container-kit/redis-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Redis Cluster
            displayName: RedisCluster
            kind: RedisCluster
            name: redisclusters.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis Replication
            displayName: RedisReplication
            kind: RedisReplication
            name: redisreplications.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis Sentinel
            displayName: RedisSentinel
            kind: RedisSentinel
            name: redissentinels.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis
            displayName: Redis
            kind: Redis
            name: redis.redis.redis.opstreelabs.in
            version: v1beta1
        description: "A Golang based redis operator that will make/oversee Redis\nstandalone/cluster
          mode setup on top of the Kubernetes. It can create a\nredis cluster setup
          with best practices on Cloud as well as the Bare metal\nenvironment. Also,
          it provides an in-built monitoring capability using\nredis-exporter.\n\nOrganizations
          that are using Redis Operator to manage their redis workload \ncan be found
          here. If your organization is also using Redis Operator, please \nfree to
          add by creating a PR https://github.com/OT-CONTAINER-KIT/redis-operator/pulls.\n\nThis
          operator only supports versions of redis =>6.\n\n### Supported Features\n\nHere
          the features which are supported by this operator:-\n- Redis cluster and
          standalone mode setup\n- Redis cluster failover and recovery\n- Redis leader
          and follower setup with Sentinel for High Availability\n- Inbuilt monitoring
          with redis exporter\n- Password and password-less setup of redis\n- TLS
          support for additional security layer\n- Ipv4 and Ipv6 support for redis
          setup\n- External configuration support for leader and follower\n- Detailed
          monitoring grafana dashboard\n\n### Documentation\n\n[Documentation](https://ot-redis-operator.netlify.app/)\n\n###
          Getting Help\nThis project is managed by OpsTree Solutions. If you have
          any queries or\nsuggestions, mail us at opensource@opstree.com.\n\n### Contributing\n\nYou
          can contribute to this project by:-\n- Raising Bugs and Feature Requests\n-
          Fixing issue by opening a Pull Request\n- Improving Documentation\n\n###
          License\n\nLogging Operator is licensed under [Apache License, Version\n2.0](https://github.com/OT-CONTAINER-KIT/redis-operator/blob/master/LICENSE)\n"
        displayName: Redis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - Redis
        - Opstree
        - CRD
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-redis-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/opstree/redis-operator:v0.15.1
        version: 0.15.1
      entries:
      - name: redis-operator.v0.15.1
        version: 0.15.1
      - name: redis-operator.v0.15.0
        version: 0.15.0
      - name: redis-operator.v0.14.0
        version: 0.14.0
      - name: redis-operator.v0.13.0
        version: 0.13.0
      - name: redis-operator.v0.8.0
        version: 0.8.0
      - name: redis-operator.v0.6.0
        version: 0.6.0
      - name: redis-operator.v0.5.0
        version: 0.5.0
      - name: redis-operator.v0.4.0
        version: 0.4.0
      - name: redis-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: redis-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Veeam Kasten
      provider-url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
    name: k10-kasten-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k10-kasten-operator-rhmp.v8.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apik10.kasten.io/v1alpha1",
                "kind": "K10",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/reconcile-period": "2m",
                    "helm.sdk.operatorframework.io/rollback-force": "false"
                  },
                  "name": "k10"
                },
                "spec": {
                  "auth": {
                    "basicAuth": {
                      "enabled": false,
                      "htpasswd": "",
                      "secretName": ""
                    },
                    "tokenAuth": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "persistence": {
                      "catalog": {
                        "size": ""
                      },
                      "storageClass": ""
                    }
                  },
                  "metering": {
                    "mode": ""
                  },
                  "route": {
                    "enabled": false,
                    "host": "",
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "apik10.kasten.io/v1alpha1",
                "kind": "K10restore",
                "metadata": {
                  "name": "k10restore"
                },
                "spec": {
                  "pointInTime": null,
                  "profileName": "",
                  "skipResource": null,
                  "sourceClusterID": ""
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          console.openshift.io/plugins: '["veeam-kasten-console-plugin"]'
          containerImage: registry.connect.redhat.com/kasten/k10-operator@sha256:4e8ccf38e47f92d2a308587591f85fb4f15aea1e7fd5c248803764c94eaecd4a
          createdAt: Wed Sep 24 02:11:59 PM UTC 2025
          description: Kasten data management platform for OpenShift, The Market Leader
            for Kubernetes Backup and DR
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/k10-kasten-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/k10-kasten-operator-rhmp/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: kasten-io
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: contact@kasten.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              k10restore can recover the K10 platform in case of a variety of disasters such as
              the accidental deletion of K10, failure of underlying storage that K10 uses for its catalog,
              or even the accidental destruction of the Kubernetes cluster on which K10 is deployed.

              **Before creating k10restore**: [K10 Disaster Recovery documentation](https://docs.kasten.io/latest/operating/dr.html#recovering-with-the-operator)
            kind: K10restore
            name: k10restores.apik10.kasten.io
            version: v1alpha1
          - description: |-
              [Kasten's k10](https://docs.kasten.io/) is a data lifecycle management system for all your persistence enabled container-based applications.

              **Note: The complete list of values supported by the K10 spec field in the "YAML View" are documented [here](https://docs.kasten.io/latest/install/advanced.html#complete-list-of-k10-helm-options).**
            kind: K10
            name: k10s.apik10.kasten.io
            version: v1alpha1
        description: |-
          Kasten data management platform for OpenShift, The Market Leader for Kubernetes Backup and DR.

          The supported Kubernetes versions are listed here - https://docs.kasten.io/latest/operating/support.html
          Please review release notes before an upgrade - https://docs.kasten.io/latest/releasenotes.html
        displayName: Veeam Kasten (Free)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kubernetes
        - storage
        - data-management
        links:
        - name: Kasten Documentation
          url: https://docs.kasten.io
        - name: Veeam Kasten
          url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
        maintainers:
        - email: contact@kasten.io
          name: Veeam Kasten
        maturity: stable
        provider:
          name: Veeam Kasten
          url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:503d5ee0fbcd5bd3492785f4246e19cbdb92960be0d05f3c6f123a32fe0fe84d
        - registry.connect.redhat.com/kasten/bloblifecyclemanager@sha256:7b209314b96bc8ebaa77e97426bb34176d05844cc3e8d4b462cfba5c4a8d3798
        - registry.connect.redhat.com/kasten/kanister-tools@sha256:7af2e37c9027f46c565901d35a3a7da4d9c7ddd2812628b67275d4acf006b80c
        - registry.connect.redhat.com/kasten/kanister@sha256:1d69ff91cec8e7c131dfbe1ad3e282aab854d3412f5f6e3b808d25eb7b8c38bc
        - registry.connect.redhat.com/kasten/metric-sidecar@sha256:20c142258de56cf8c60d4b0afa5e72b9317023ebf4ab109fb84e0d57fae854a4
        - registry.connect.redhat.com/kasten/controllermanager@sha256:ab3b815bb12cd3c263cbfdba3939a95097450f1864431cca7380827986d704ed
        - registry.connect.redhat.com/kasten/configmap-reload@sha256:01010f2bb580eb131d63d3271efe40eadd8269b84cdb02bdea36750d1359a43f
        - registry.connect.redhat.com/kasten/aggregatedapis@sha256:db868afd5c6d4a8d2e7c8376ed6943963c4be299fca6b26931f907430763d244
        - registry.connect.redhat.com/kasten/jobs@sha256:c6358878e981baee4ac72abcc1269c206ee014af2c16584d1b41fe18e0f0e57f
        - registry.connect.redhat.com/kasten/upgrade@sha256:bf2200697d38f3ee95ba2863dfbae4d982273d2fd8450c9ffbc3a85be7d6d87d
        - registry.connect.redhat.com/kasten/prometheus@sha256:29d913bb06c3245cca5798772dba7d7d5dc9b76ed2e0700f5eb4d20416e24fa2
        - registry.connect.redhat.com/kasten/k10tools@sha256:59caa1920c76a1bc86e7733731602e3f0fc08026d9a16000bb748503a4d9cd9b
        - registry.connect.redhat.com/kasten/garbagecollector@sha256:6f2972bb255f5a2de33ace5ce805b15d5e04d3c259c9c8d018c1257c74562878
        - registry.connect.redhat.com/kasten/events@sha256:7cae1c0909ea7a468e96484d99300677bdad49f34cc57c412ebaf342cb176264
        - registry.connect.redhat.com/kasten/logging@sha256:014bfdfc3620cc2fe6540b9993b9b9e7d70c9b969714477db4de91a90bdb4a38
        - registry.connect.redhat.com/kasten/k10-operator@sha256:4e8ccf38e47f92d2a308587591f85fb4f15aea1e7fd5c248803764c94eaecd4a
        - registry.connect.redhat.com/kasten/auth@sha256:f8543e67de3d5276c597c26f16053bf9bbcbd9cbdad4652214935a1e40a26817
        - registry.connect.redhat.com/kasten/frontend@sha256:a021964ed85133897e4956014b5c5013da3332a36a7e96a0de2e2e1b67888491
        - registry.connect.redhat.com/kasten/init@sha256:cc7fbbda1b6946e93dd083fb7092f38789f626c26ac68c226b61fe57064a8f11
        - registry.connect.redhat.com/kasten/dex@sha256:4c8522a27c0d4f0977baa074800bba96a9bd9aefc57473de283b1d95586f7517
        - registry.connect.redhat.com/kasten/catalog@sha256:11a77ac968634850c72cd10f2bde55afc632c68cdc0881aeae83c100288c6aec
        - registry.connect.redhat.com/kasten/dashboardbff@sha256:e1208c60146faa8c3c7ef73b2290663840276ed8d3b3c990f495cee2967104ee
        - registry.connect.redhat.com/kasten/state@sha256:191383c76c4ff22805c06217e9f32a2f41a1206afab5f16acec4a5dd56e4a6b4
        - registry.connect.redhat.com/kasten/repositories@sha256:924a607874e1807a877cb454e00539d64f4487e7435d4bfb524f3e569ad1f942
        - registry.connect.redhat.com/kasten/metering@sha256:f809696ac1309fc54fde87e5d2030267d266386fb73a56287868bbac433756d8
        - registry.connect.redhat.com/kasten/restorectl@sha256:9dc3c834bd3d4543115ae1ad586ee1248fc2b637dbc2d1e2c66ecc2a459db67e
        - registry.connect.redhat.com/kasten/datamover@sha256:c9e2dc5b646aa0c465a7b2daf011462273a6f5b53fdff772d879293208e2d57c
        - registry.connect.redhat.com/kasten/vbrintegrationapi@sha256:1a20b0da4efb7e802516224675b37751bd56d406e57b16216388d2b5167c648d
        - registry.connect.redhat.com/kasten/executor@sha256:32c4ac1aa2ca802ff77e913991949e6841149963863ad527a67f6bab6c94a5af
        - registry.connect.redhat.com/kasten/gateway@sha256:dc2d7bdeb8abaf9454470afd53af2cd5430505351a97b6501ff65a60dd2a6db1
        - registry.connect.redhat.com/kasten/ocpconsoleplugin@sha256:69f6bfb9421f4264dd7d066d809311022c0db57c41319a6242532f6acfbd2145
        - registry.connect.redhat.com/kasten/crypto@sha256:86bfbbf4c4f06bab1e4b091f9f312d536813e2c11aa899690b3277ff73110a09
        version: 8.0.9
      entries:
      - name: k10-kasten-operator-rhmp.v8.0.9
        version: 8.0.9
      - name: k10-kasten-operator-rhmp.v8.0.8
        version: 8.0.8
      - name: k10-kasten-operator-rhmp.v8.0.7
        version: 8.0.7
      - name: k10-kasten-operator-rhmp.v8.0.6
        version: 8.0.6
      - name: k10-kasten-operator-rhmp.v8.0.5
        version: 8.0.5
      - name: k10-kasten-operator-rhmp.v8.0.4
        version: 8.0.4
      - name: k10-kasten-operator-rhmp.v8.0.3
        version: 8.0.3
      - name: k10-kasten-operator-rhmp.v8.0.2
        version: 8.0.2
      - name: k10-kasten-operator-rhmp.v8.0.1
        version: 8.0.1
      - name: k10-kasten-operator-rhmp.v8.0.0
        version: 8.0.0
      - name: k10-kasten-operator-rhmp.v7.5.10
        version: 7.5.10
      - name: k10-kasten-operator-rhmp.v7.5.9
        version: 7.5.9
      - name: k10-kasten-operator-rhmp.v7.5.8
        version: 7.5.8
      - name: k10-kasten-operator-rhmp.v7.5.7
        version: 7.5.7
      - name: k10-kasten-operator-rhmp.v7.5.4
        version: 7.5.4
      - name: k10-kasten-operator-rhmp.v7.5.3
        version: 7.5.3
      - name: k10-kasten-operator-rhmp.v7.5.2
        version: 7.5.2
      - name: k10-kasten-operator-rhmp.v7.5.1
        version: 7.5.1
      - name: k10-kasten-operator-rhmp.v7.5.0
        version: 7.5.0
      - name: k10-kasten-operator-rhmp.v7.0.14
        version: 7.0.14
      - name: k10-kasten-operator-rhmp.v7.0.13
        version: 7.0.13
      - name: k10-kasten-operator-rhmp.v7.0.12
        version: 7.0.12
      - name: k10-kasten-operator-rhmp.v7.0.11
        version: 7.0.11
      - name: k10-kasten-operator-rhmp.v7.0.9
        version: 7.0.9
      - name: k10-kasten-operator-rhmp.v7.0.8
        version: 7.0.8
      - name: k10-kasten-operator-rhmp.v7.0.7
        version: 7.0.7
      - name: k10-kasten-operator-rhmp.v7.0.6
        version: 7.0.6
      - name: k10-kasten-operator-rhmp.v7.0.5
        version: 7.0.5
      - name: k10-kasten-operator-rhmp.v7.0.4
        version: 7.0.4
      - name: k10-kasten-operator-rhmp.v7.0.3
        version: 7.0.3
      - name: k10-kasten-operator-rhmp.v7.0.1
        version: 7.0.1
      - name: k10-kasten-operator-rhmp.v6.5.14
        version: 6.5.14
      - name: k10-kasten-operator-rhmp.v6.5.13
        version: 6.5.13
      - name: k10-kasten-operator-rhmp.v6.5.12
        version: 6.5.12
      - name: k10-kasten-operator-rhmp.v6.5.11
        version: 6.5.11
      - name: k10-kasten-operator-rhmp.v6.5.10
        version: 6.5.10
      - name: k10-kasten-operator-rhmp.v6.5.8
        version: 6.5.8
      - name: k10-kasten-operator-rhmp.v6.5.6
        version: 6.5.6
      - name: k10-kasten-operator-rhmp.v6.5.5
        version: 6.5.5
      - name: k10-kasten-operator-rhmp.v6.5.4
        version: 6.5.4
      - name: k10-kasten-operator-rhmp.v6.5.3
        version: 6.5.3
      - name: k10-kasten-operator-rhmp.v6.5.1
        version: 6.5.1
      - name: k10-kasten-operator-rhmp.v6.5.0
        version: 6.5.0
      - name: k10-kasten-operator-rhmp.v6.0.12
        version: 6.0.12
      - name: k10-kasten-operator-rhmp.v6.0.11
        version: 6.0.11
      - name: k10-kasten-operator-rhmp.v6.0.9
        version: 6.0.9
      - name: k10-kasten-operator-rhmp.v6.0.8
        version: 6.0.8
      - name: k10-kasten-operator-rhmp.v6.0.7
        version: 6.0.7
      - name: k10-kasten-operator-rhmp.v6.0.6
        version: 6.0.6
      - name: k10-kasten-operator-rhmp.v6.0.5
        version: 6.0.5
      - name: k10-kasten-operator-rhmp.v6.0.4
        version: 6.0.4
      - name: k10-kasten-operator-rhmp.v6.0.3
        version: 6.0.3
      - name: k10-kasten-operator-rhmp.v6.0.2
        version: 6.0.2
      - name: k10-kasten-operator-rhmp.v6.0.1
        version: 6.0.1
      - name: k10-kasten-operator-rhmp.v6.0.0
        version: 6.0.0
      name: stable
    defaultChannel: stable
    packageName: k10-kasten-operator-rhmp
    provider:
      name: Veeam Kasten
      url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kinesis-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kinesis-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kinesis.services.k8s.aws/v1alpha1",
                "kind": "Stream",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kinesis-controller:1.1.1
          createdAt: "2025-09-19T20:56:43Z"
          description: AWS Kinesis controller is a service controller for managing
            Kinesis resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Stream represents the state of an AWS kinesis Stream resource.
            displayName: Stream
            kind: Stream
            name: streams.kinesis.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Kinesis resources in AWS from within your Kubernetes cluster.

          **About Amazon Kinesis**

          You can use Amazon Kinesis Data Streams to collect and process large streams of data records in real time. You can create data-processing applications, known as Kinesis Data Streams applications. A typical Kinesis Data Streams application reads data from a data stream as data records. These applications can use the Kinesis Client Library, and they can run on Amazon EC2 instances.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Kinesis
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kinesis
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Kinesis Developer Resources
          url: https://aws.amazon.com/kinesis/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kinesis maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/kinesis-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-kinesis-controller.v1.1.1
        version: 1.1.1
      - name: ack-kinesis-controller.v1.1.0
        version: 1.1.0
      - name: ack-kinesis-controller.v1.0.16
        version: 1.0.16
      - name: ack-kinesis-controller.v1.0.15
        version: 1.0.15
      - name: ack-kinesis-controller.v1.0.14
        version: 1.0.14
      - name: ack-kinesis-controller.v1.0.13
        version: 1.0.13
      - name: ack-kinesis-controller.v1.0.12
        version: 1.0.12
      - name: ack-kinesis-controller.v1.0.11
        version: 1.0.11
      - name: ack-kinesis-controller.v1.0.10
        version: 1.0.10
      - name: ack-kinesis-controller.v1.0.9
        version: 1.0.9
      - name: ack-kinesis-controller.v1.0.8
        version: 1.0.8
      - name: ack-kinesis-controller.v1.0.7
        version: 1.0.7
      - name: ack-kinesis-controller.v1.0.6
        version: 1.0.6
      - name: ack-kinesis-controller.v1.0.5
        version: 1.0.5
      - name: ack-kinesis-controller.v1.0.4
        version: 1.0.4
      - name: ack-kinesis-controller.v1.0.3
        version: 1.0.3
      - name: ack-kinesis-controller.v1.0.2
        version: 1.0.2
      - name: ack-kinesis-controller.v1.0.1
        version: 1.0.1
      - name: ack-kinesis-controller.v1.0.0
        version: 1.0.0
      - name: ack-kinesis-controller.v0.0.19
        version: 0.0.19
      - name: ack-kinesis-controller.v0.0.18
        version: 0.0.18
      - name: ack-kinesis-controller.v0.0.17
        version: 0.0.17
      - name: ack-kinesis-controller.v0.0.16
        version: 0.0.16
      - name: ack-kinesis-controller.v0.0.15
        version: 0.0.15
      - name: ack-kinesis-controller.v0.0.10
        version: 0.0.10
      - name: ack-kinesis-controller.v0.0.9
        version: 0.0.9
      - name: ack-kinesis-controller.v0.0.8
        version: 0.0.8
      - name: ack-kinesis-controller.v0.0.6
        version: 0.0.6
      - name: ack-kinesis-controller.v0.0.5
        version: 0.0.5
      - name: ack-kinesis-controller.v0.0.4
        version: 0.0.4
      - name: ack-kinesis-controller.v0.0.3
        version: 0.0.3
      - name: ack-kinesis-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-kinesis-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cockroach Labs
      provider-url: ""
    name: cockroachdb-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroach-operator.v2.18.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v25.2.2",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
          createdAt: "2025-07-22T11:30:55Z"
          description: CockroachDB Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1304d7f871a95c8b087bad1b548217562792edc66c8fe07c621e05242b8cd53e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6408f89b0a88ec154f7ee98c314be574971c4c9aaf906b473d10124d822889be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:65923c37fecda2d3174212fa9d9a6f34241d065389336ea69a7be99ba16cedc2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3bf017b2b2506724500f5e29017d1b956825dd0f61a7c0049b0ff084c3554b5b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:85a0e085901c2a73192556d5d70795d4b66cbcfe01d24c73b6fe6db8d5f2e5ca
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bd75aa918355e77cd671c3e11ee807861378e96245245f4fdf243cba5749f9a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e8f26fd132affbaeb8ef6b1807a646c83b301e5dc334a592dfb2cf8dc3dbaf10
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6ca927f137be28481e3a226e6c6bcb7c5ac27664984f17ffc6c1419cd7d08eb7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da367cf0ac52045002e1c186f8e6964267ad87d5c25f9e72fd2c9b9a98a32702
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc1ab77a9537ccf236ce6dc370940dfedcfaa91b6e3db37bf2d5eade7ff84708
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e0f21abea34825e55622c432a306f2e8aafd64d913eb78f2f601c51dc878a1da
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1b6d934a7b6c42a61e5ad51380e69af341a575b42205299fd58bc0c6ff306f6a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a6f1c43486d044fed9f59406118f429c8037b389bd8c8d7a1c7aeb89ed2661f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b06bd7f3802e8fa2af14bb0b07804cad4cc0fbbc97da18d09a9106bf074ed6d8
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:740ba10f1ef0a6c5fa6ad31ab18855fc75f8e024ce4c23b0676c5a9c5b62cd75
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c9a31bdf19175880d0981e21dd19a1458b6e8cde8d4acbcb554f3bea3578a64e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf7d3bea47638b763b312c7c11ca034a7ac62d993d75a13edceec74799b97e52
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f438fbbf9c87ca0d54c170e2c5105b1e265182e1c2ede48adea370e2be9dc895
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9ab968e4ca62f23bf9848e5475ef4076df7318a8560f63d5b2090a32af6fe4e2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d978678b3c254abd52f69f226b0aacc4b02c2aaca328c54ef10f6f9bb453582d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:dd521d7bad62da054eec54179be75c643a5492d981e11c0e66db41902cd2182c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d0014b1c19890b9e8ef1a4f3efe7eb5b269be5610ba6ababd4258b52d2f85e8a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5a54d564e08baefc2786f4b742b2507d31984664cd00333406aa7a3aa5b9c913
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5181196fff3bf8a5ec8fcfe4f1958cc69f285f0beb4f0b476b4b5e0912f3db42
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9f42e929f2cc7e9b6d24dd61e19c5d93dcfeabd34b9c8b935afbe78577efb18b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:266165ab87b19a05d6796e8b15ff0eba632d6ac96e6393400464c6136dd500ec
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:46d9287ef11e3b21a36abfa31a26da3b93040bb6163b9223972a5dd1930fecff
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acbfee4492327a720fe7da76929a86192d8af2442bbb76f0a089f6a81a890be3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6371377718325235a2720dba0c97eb4321ee29b31a28cb893f76abd62baa339
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:79d688bebe0643ef0dd6ba7108cdb34c0442470c0c20ae0df88b5d02a3476cd0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e23af8f9d32c3f80a52d79ceb4b359dda1d75ba0dd4cb8ed4b66ac2f4e40c69b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:58de9d0b2ce005c7ad3e99152100e59b7677398557028a7a59e56e669fabb81c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7daaf1368ab2c25694cdea0a598e1de59247b04a4b4b4b8348ed6edbaa194f9d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:35ede65457d845e8c063ad7d9515d3a69f98e0751424ecbac484963678a0b1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f42fa22aac3cc41ce72e144a72530ffbee765a392b24593dafd06cc098f8410b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a4f932c3e9ff3aeb70ba1f7a92aa4642bd14cbb7d37d04ff750ed517b916cdb9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:19f36f53f7da67755eb86da77098c5eeb84b381f9c18aa29ba573793dc498564
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4635550222775fc75737be5e5c0359303e3a23c4d1e1c296d7408cd15ad94aad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f3ffebd812881a830ff46c860984814dd61edcd51b478ebd3667759fb7710682
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2a128b570352a5a7f82b27353f9b774caccce83e0e7bb2673e4802d96ad0bd51
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f74c8a5dea4560ad59c6f6f6ab1e699f063334ee8db704460f870508dbeaeff0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bd336715332f776a72aaf5c3ca52e62dd53a9e8521c1feb6e135cc868d694a39
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5240f9f591263569fad7ee934e81fa8a1617de3c25da94ddd59ee5b2b3407f90
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d14f77f70aea0422a906510c1ac2652ff76e131ed705cb002f80e13e887b0cd9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:10b9df672260c023dfb37166e156c0d7e0e13144dc3e485af922e8151fd3ab05
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2af60025ecb3bb933b61328be4c2b2bfd0e7d26f53b72430629208505e45c6d2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:32fdca575c334822e4356aab36a7ed97b685c065925fe85f1b8ba8425c57159b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:42c9930d6861f6a2147ed6e01827eddd730279966763ec68868567a19cc164ab
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4b3aa590623a38f435b075de4c09f8a343d8ece8b673e6bf7c6df5d5eb4bdf4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:02e1ef01a023da0eee653f63b314786fd66a78d184a97a314422f6f9c3ef8e7e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:310310515625f099a928545865f7096997871ee4a16650a01c76c3799a18b684
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6a2ee8d62c586eb74ebbef6bef256e06e1ba5cb51c1ae55b1654adafc2ca8859
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acc083f7dde1fbc557d858d03fc26b68b2056997fc1a457e4695650214d5b303
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b78a4d3864e16d270979d9d7756012abf09e6cdb7f1eb4832f6497c26281099
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:61e602b0e3582bf00460eb9328808ccf66e3ff1b1f6df07ee6c5c76a91846876
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2d89398f03736f8ffaa82b10fed0f19fba166d3d2145b26f25cb71f3c39acf07
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5458a8fd89aa664b60e59c35bd933975cd37d18e19b11589e4ca0c2844ea0ed4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:50393ad005fed19618d04f0716c6d55e2f1afce651e785c396638e852cac83b3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ec6eb7c28c213cc83b2d7919cd87988f9a07f12276eb7351d0915f1567a5b095
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6366ba29f301a82215c05e60d78e1de66bcee85440ce2fc3c692712d4da32685
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5d6901b6b6fd6594d42bfc1842d050c7cffcb0867a5591a67469f0f860b1f21b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4d8e804b35b73d7978c112748aea2d2bb68f65081b90bca62c39f2b6e2989378
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:743cda1eda6e273c798e3405a9a97e07dd6aa5ad4d5a52789be3b0f39a32a039
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4f753151af4c3d3f5a1df9014461316249a6d0621b8fba5c4430307b3355251
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce1209ca6cc8724baee981c97c66e2852e512423bc0c0a16cdd80a6838b54031
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f5a0ccc02dc9e938e484d5b5282ff650d1890d5f754c30a0c0d717989ed5d600
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f045fe5037d44aab1a0b2209241b3931ec7cb9897cb3c6a96a9f8d5280e55dd2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08cec123813304ab03bd66d72a09949667a874a42320bedafda391ebcdb2e56c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5b3de46526d169fcd1318354545b3a18861843fa2022a23ed04d795493825287
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d590946b2f7c7ec0c7ec057a338e81f864d6a9e3d3ac4e0769aa5756a8e13fc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0754f7901feea9017bc5041cb91b7ac1cdbca10c2064f37c3648a2c395f24f83
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db07461824e73410456ef540b0bce3c7fe550aa67a73d6174300ca74b8f75276
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c74cf927a2b2c5098ad74ad0880996f8b042186a9911d888a231ef5b9e2a0715
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7d6320ac7fb0925f10cf6f524a13de629cbf13ccc371912e4fe8c777bc15d1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:28ae9661b38a72decf413c6a6c3758e61098debf5162197c27a12a070c7a096d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:226d7f1c3f0d70ac6532ad9490bebcdef44a40959b6aa7c11b0f2d1383a46217
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:44a0f40a6c46ba08297b482f3d96defe87b181c944803e5381c9acc42cabf97f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b046aa0a42dae992a03685ed82a20835532d6d4bf048e2366ff7382be4b3138
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1ce71e80c0cd0e0063eba275ab91f77982f2e92a8aef8e89cab99b93feab0a88
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:29abb169ffefe36f6d69edd438a7135ee23059cedab8f87ad3b2d52361b96b35
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:68fe64fe3e112288ccfc2a7dffd69b9b2786a5da3d3bdedd6addf41659d460c9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8443c15eb1653dd9e5ec7aad54d4b8b092d41a0977e42b76e372e5e21280746e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b83eaff2e217c06098509709cead525b55b00ccd48019d8357e6231892008055
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bed27f3f2c2d7627d1af4500d7ea0302da5c3fd3d875a6480fb6d98da481312e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b0057de903654ea0315a2c96affcc8f50a92ede8ad121ed30edb3aaf0fb7b554
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cbc74889e9ca5edbd5aaa243d7649bca0c4da2719239f8d0ec9827868112a771
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1558a1f5eab79f1a8c6c85d81740cc539dc23025ea457c0019d93a72dc05d57a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf001173bbf00cdf443b8d64a3ad724389f1cf83e3628d0b609019c8cddad35a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:87db6e079b96fb392d98ef5fc7063836bff10717efd75061016ce688be0ba38e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:72ea3d0ef72972dba93362dc31366c5946a9898f5c16780b877316a97068ef8f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1374cbdc4ecf38e4b4226328f532e830b8932b34d9449bf2fbb36e60d0310521
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2171e1c75d0c665fe33e1ea334e605604c51de6dbc22d88a3a0ac604259c448f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db4e5c2f27ba23e95a363c8174c01f0f9c6b6a27e0e581c86a38f3a1f68dc43c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d8cfb221f67516675fdfde061f6ec5dfee1224ca7eb3d8be375c27d483f306af
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:aeac9cd8e8acdbd2b0c339c48b85079c159b04f0ade24df8e3f9cae3f48b4c70
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:086d7b435993fdf21fd59b0093f52105a9028d6b769398b0033f98a65dfb7e79
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9d0bfb642090ff55a3cc3da0f72663cae18063622a4ec785ba99bc71ef4db116
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3f587e7b9f5a611013af2608e2151ed782859a52501b00adbc95739c56a8ff1e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:82d8104a37349c513447e68ea3790efb256370632353176e5fc13c2e8df1bd44
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c2a62179e8ca293d7e30b9da599e48795345cf07a72abd3a83063107b53b03cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d467383af41aa80c26172964c34152abeba45121599804e502984655b72179f0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e5814748c5ab229a1dea5662a007063c0df06fbbfcfe041fd9a6df9ef67e07cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a6e2fe24ad52913e2160f85b9fe665a4e4c4d79b164b4b899f69dd9d92f999e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6cd9f9b9fcb6e8f9a931a6ad01cb66329a74d40b22f9029fbc50e57248bad7cb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:56109e57ee0379cf48644bcf8226a9238e01139cacc5499002c99f973f121911
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f53b07212320c6939d3bb4ddd4e6a0bce311ab5b63858eb1a8728531f0dbd510
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce3d94a3fbdcde5858beb3c265987759971d991c397c6486e19904bd4706f44d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e151bded670f4cbd165a8ef864324805667d3e84ce3ace0cdf8f9aad087fccf1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6bb15b36d64eebb6e4c1db5a5466e108b271d53383c58e0b6c78cec214756a9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0e2c7023f1f97acc20c46a4944635a9c5f57bb9bb3dd534cf41fea1a9cfdfdbe
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:91f7e59f9e35a0395756eafed906963c535e886125294d671d930ce959a8f511
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f05824bb8eaa5a93e54059449eaa236dd83d729049ad00065ef3242f5f74c91e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b77b80eed7ce149d2a3f228e5f188e76d73f32bc50ecde20bba9cfa7bdc4714c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cad60044ad6573bd80b65e892a9ec0510dbe4fcbbfc6b51010ecf419f56f1024
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8fc8de4f3671a73f9aa747e5b23ec973d192fff6f2869c38985cd656063c1e3d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:022c6e1a0de875474bc953718a481a0af526c49dc910094dcdce0b5677aced18
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:468b0a40a33be1dd5a2a89ce745d7e9012bc7371f2f01cc0f8f12c0ac887db43
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4e5f7df1dc1e1db398c36d590431e7e5782897b209972d8e9e4671971c10d1b6
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7481c0cf3b579c8d98d94d4aa0334ca94bf2ccfcd25ad70b57c42b22bbf4bcee
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08496b84b3cc691909227be17e67bcd6f4d5185a93712f44d8a0985bd2c5fcd7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fd6c9b8ca82ff4f2d0fe29166d1fc9bca6439cb40907be60f4d8f1b9735de2f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:742a6f69a3a209a7080a1e911fd96467d2a91aacb3f91d6be80a6a6e2e54b2ae
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e82684b828256b3c2814930210031b3b70bc818514d4359a25f87ffdbf22be6e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8737b67b91983817e3e4ff3f47f0b070bd36162d309b60993529426c6979cb24
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4b56c0311af14220a466aa11f25fd10fa38280825539bd02215c945bbafb861c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9097e135159f46396ef4649f724d61bb5b9113c488f989d404586fbb7486fd85
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1e5d85f5db10fa79d09263db90daa44549bcc3b025e9f3f98cf0d4e51766394b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a0faa52a999289b308dc4caf050c6d9fe2f42e112aa30e4af282ed1f21c5c346
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1f6b1ec92251397dbcf7ed4a82f1f05619d64676aae2717f4eff1c33e4c2c27a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a59cafff8d4ff2c62204ece15880117011c30978823c232902c7182d747ec827
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:858dc82977b3c9e989a3a20626671b0c67c834c97132094a9212bb227ab55929
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:744193037d6b64cb73fecab15ebe022abc2b79ab0335b76ec1415ac6379b7dfa
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d78c2e0709214f96794564f13386861f4b541e55b5bd316ad495cf104f7d40b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9b5a2e0c006eb4be8e1444734a98743551f278a24e81c20bebc96f0f3b631eb0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e50aab39722af22ccbf3d1db19e37972d52b0fc9a40998a64618dc966b2bac57
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:818f592c50d4ed5599faa053344f7ecd963003352f1faa3b72dc0f5e00a0e78b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0b0e0588c1aadf0c3826cd584bd2b3a7d6781882dcb5c5e037b47ff5cb84509b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:53c7b960a2ed70a0998a4f8e28a4f78f7f676e01cffebe665e91e3a7629d88ed
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f29c31bb267451b389608af3e8a4a28d6ca0691e67b77e8f57bb5e461b064893
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bf36dab9b3257a4bba24d986f1451fc51a3a37023fbb988281f1941fd3faedd
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7b907d03b001b758d8a4b5e898725c73dda2ec2aa4b7425b685320b947154d11
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8f1f2fd1aca914a1ef007117c85ac944055bbfb873a468a7374d239ef20b065e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:37bf9e1ec7e3eff47755fe7521223f31ec6aefc6e63901db3881b30898cf78f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:40f00476b63cb9e370fe85f01407173693213a9242fc117d28b06378ca0d98e0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8a70ba12f610ac0e62066e417a0833bc1d01e91ac99d0dd794a2c0cd2a350164
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2b5c028f293c6f394e58a5495cb3dbfbedc215d0d03778d0217e045586d55a46
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da3bd2e9ff2b08d5bb86b1f5faa95a865f85c5e8912df00b6b79190c1e518a2f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7ad12df8e258b03c4099d375d02d272bd774bf188d108901d5cf6cb865e57ef3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cc20270212f01fc231ff788e20b0476b8228c440bfc40430c0836d5eb702a042
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:31b161f9bf9028a1a62cd2eef6abda01257d8889bdc3159822642549a75e3429
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0fea6ed8a78ea624240ee6ffb3573d5cf1115186aa9180fcd7d1273351deaaa3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9c85c25d5392817d17b68714dca6195ad53255d43e9c65b021d75b12d97e0579
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a3ac544d87a9d8e7c92897c4c23f803753c7ab31467db91b5c51244afc2c6be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e18d73af3257c8cdf00245f2f718b56667524c1cc69d3db28e9b868a539ae16f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:363d8feb560324a848c5771edd224dc19020013af7641979ee81ae8ee536d7e3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fe4a40006d1e8439423de1417cc9bc55716d790b882485074bb2a5ec68d7d1ad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:90cf21c36d19d87b97d06edd1c0912a3bb288c769e6159651877db3206aa7355
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0ec1dc184a2199736b4bd7de3241a357be7427caea5da1c931b08bc06f5c3dc0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8e8c081c231b5b15eca483ceec16ea9190ba73aae6e421865162446eb584519a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc4b75ed71845d5b18e1d9163d480fd9d0d3cb5ebbcfed00a2e4e1b174c0a5de
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9492894eaa5f43d00ac80b60029b32a0b0a85ae88ba609b46f98702c9349a5bb
        version: 2.18.2
      entries:
      - name: cockroach-operator.v2.18.2
        version: 2.18.2
      name: stable
    defaultChannel: stable
    packageName: cockroachdb-certified
    provider:
      name: Cockroach Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Elastic
      provider-url: ""
    name: elasticsearch-eck-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: elasticsearch-eck-operator-certified.v3.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "elasticsearch.k8s.elastic.co/v1",
                    "kind": "Elasticsearch",
                    "metadata": {
                        "name": "elasticsearch-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "nodeSets": [
                            {
                                "name": "default",
                                "config": {
                                    "node.roles": ["master", "data"],
                                    "node.attr.attr_name": "attr_value",
                                    "node.store.allow_mmap": false
                                },
                                "podTemplate": {
                                    "metadata": {
                                        "labels": {
                                            "foo": "bar"
                                        }
                                    },
                                    "spec": {
                                        "containers": [
                                            {
                                                "name": "elasticsearch",
                                                "resources": {
                                                    "requests": {
                                                        "memory": "4Gi",
                                                        "cpu": 1
                                                    },
                                                    "limits": {
                                                        "memory": "4Gi",
                                                        "cpu": 2
                                                    }
                                                }
                                            }
                                        ]
                                    }
                                },
                                "count": 3
                            }
                        ]
                    }
                },
                {
                    "apiVersion": "kibana.k8s.elastic.co/v1",
                    "kind": "Kibana",
                    "metadata": {
                        "name": "kibana-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        },
                        "podTemplate": {
                            "metadata": {
                                "labels": {
                                    "foo": "bar"
                                }
                            },
                            "spec": {
                                "containers": [
                                    {
                                        "name": "kibana",
                                        "resources": {
                                            "requests": {
                                                "memory": "1Gi",
                                                "cpu": 0.5
                                            },
                                            "limits": {
                                                "memory": "2Gi",
                                                "cpu": 2
                                            }
                                        }
                                    }
                                ]
                            }
                        }
                    }
                },
                {
                    "apiVersion": "apm.k8s.elastic.co/v1",
                    "kind": "ApmServer",
                    "metadata": {
                        "name": "apmserver-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                    "apiVersion": "enterprisesearch.k8s.elastic.co/v1",
                    "kind": "EnterpriseSearch",
                    "metadata": {
                        "name": "ent-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "config": {
                            "ent_search.external_url": "https://localhost:3002"
                        },
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                  "apiVersion": "beat.k8s.elastic.co/v1beta1",
                  "kind": "Beat",
                  "metadata": {
                    "name": "heartbeat-sample"
                  },
                  "spec": {
                    "type": "heartbeat",
                    "version": "9.1.0",
                    "elasticsearchRef": {
                      "name": "elasticsearch-sample"
                    },
                    "config": {
                      "heartbeat.monitors": [
                        {
                          "type": "tcp",
                          "schedule": "@every 5s",
                          "hosts": [
                            "elasticsearch-sample-es-http.default.svc:9200"
                          ]
                        }
                      ]
                    },
                    "deployment": {
                      "replicas": 1,
                      "podTemplate": {
                        "spec": {
                          "securityContext": {
                            "runAsUser": 0
                          }
                        }
                      }
                    }
                  }
                },
                {
                  "apiVersion": "agent.k8s.elastic.co/v1alpha1",
                  "kind": "Agent",
                  "metadata": {
                    "name": "agent-sample"
                  },
                  "spec": {
                    "version": "9.1.0",
                    "elasticsearchRefs": [
                      {
                        "name": "elasticsearch-sample"
                      }
                    ],
                    "daemonSet": {},
                    "config": {
                      "inputs": [
                        {
                          "name": "system-1",
                          "revision": 1,
                          "type": "system/metrics",
                          "use_output": "default",
                          "meta": {
                            "package": {
                              "name": "system",
                              "version": "0.9.1"
                            }
                          },
                          "data_stream": {
                            "namespace": "default"
                          },
                          "streams": [
                            {
                              "id": "system/metrics-system.cpu",
                              "data_stream": {
                                "dataset": "system.cpu",
                                "type": "metrics"
                              },
                              "metricsets": [
                                "cpu"
                              ],
                              "cpu.metrics": [
                                "percentages",
                                "normalized_percentages"
                              ],
                              "period": "10s"
                            }
                          ]
                        }
                      ]
                    }
                  }
                },
                {
                    "apiVersion": "maps.k8s.elastic.co/v1alpha1",
                    "kind": "ElasticMapsServer",
                    "metadata": {
                        "name": "ems-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                    "apiVersion": "logstash.k8s.elastic.co/v1alpha1",
                    "kind": "Logstash",
                    "metadata" : {
                        "name": "logstash-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.connect.redhat.com/elastic/eck-operator@sha256:1750178c5dd716c031155f2e7058e19d4b7f1d3fdbb3e8a806b177a71be88500
          createdAt: "2025-07-30 11:10:34"
          description: Run Elasticsearch, Kibana, APM Server, Beats, Enterprise Search,
            Elastic Agent, Elastic Maps Server and Logstash on Kubernetes and OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Elastic Basic license
          repository: https://github.com/elastic/cloud-on-k8s
          support: elastic.co
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Elastic Agent instance
            displayName: Elastic Agent
            kind: Agent
            name: agents.agent.k8s.elastic.co
            version: v1alpha1
          - description: APM Server instance
            displayName: APM Server
            kind: ApmServer
            name: apmservers.apm.k8s.elastic.co
            version: v1
          - description: Beats instance
            displayName: Beats
            kind: Beat
            name: beats.beat.k8s.elastic.co
            version: v1beta1
          - description: Elastic Maps Server instance
            displayName: Elastic Maps Server
            kind: ElasticMapsServer
            name: elasticmapsservers.maps.k8s.elastic.co
            version: v1alpha1
          - description: Instance of an Elasticsearch autoscaler
            displayName: Elasticsearch Autoscaler
            kind: ElasticsearchAutoscaler
            name: elasticsearchautoscalers.autoscaling.k8s.elastic.co
            version: v1alpha1
          - description: Instance of an Elasticsearch cluster
            displayName: Elasticsearch Cluster
            kind: Elasticsearch
            name: elasticsearches.elasticsearch.k8s.elastic.co
            version: v1
          - description: Enterprise Search instance
            displayName: Enterprise Search
            kind: EnterpriseSearch
            name: enterprisesearches.enterprisesearch.k8s.elastic.co
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.kibana.k8s.elastic.co
            version: v1
          - description: Logstash instance
            displayName: Logstash
            kind: Logstash
            name: logstashes.logstash.k8s.elastic.co
            version: v1alpha1
          - description: Elastic Stack Config Policy
            displayName: Elastic Stack Config Policy
            kind: StackConfigPolicy
            name: stackconfigpolicies.stackconfigpolicy.k8s.elastic.co
            version: v1alpha1
        description: |-
          Elastic Cloud on Kubernetes (ECK) is the official operator by Elastic for automating the deployment, provisioning, management, and orchestration of Elasticsearch, Kibana, APM Server, Beats, Enterprise Search, Elastic Agent, Elastic Maps Server, and Logstash on Kubernetes.

          Current features:

          *  Elasticsearch, Kibana, APM Server, Enterprise Search, Beats, Elastic Agent, Elastic Maps Server, and Logstash deployments
          *  TLS Certificates management
          *  Safe Elasticsearch cluster configuration and topology changes
          *  Persistent volumes usage
          *  Custom node configuration and attributes
          *  Secure settings keystore updates

          Supported versions:

          * Kubernetes 1.29-1.33
          * OpenShift 4.15-4.19
          * Google Kubernetes Engine (GKE), Azure Kubernetes Service (AKS), and Amazon Elastic Kubernetes Service (EKS)
          * Elasticsearch, Kibana, APM Server: 7.17+, 8+, 9+
          * Enterprise Search: 7.17+, 8+
          * Beats: 7.17+, 8+, 9+
          * Elastic Agent: 7.17+, 8+, 9+
          * Elastic Maps Server: 7.17+, 8+, 9+
          * Logstash 8.12+, 9+

          ECK should work with all conformant installers as listed in these [FAQs](https://github.com/cncf/k8s-conformance/blob/master/faq.md#what-is-a-distribution-hosted-platform-and-an-installer). Distributions include source patches and so may not work as-is with ECK.
          Alpha, beta, and stable API versions follow the same [conventions used by Kubernetes](https://kubernetes.io/docs/concepts/overview/kubernetes-api/#api-versioning).
          See the full [Elastic support matrix](https://www.elastic.co/support/matrix#matrix_kubernetes) for more information.
          See the [Quickstart](https://www.elastic.co/guide/en/cloud-on-k8s/3.1/k8s-quickstart.html) to get started with ECK.
        displayName: Elasticsearch (ECK) Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - analytics
        - search
        - database
        - apm
        - logstash
        links:
        - name: Documentation
          url: https://www.elastic.co/guide/en/cloud-on-k8s/3.1/index.html
        maintainers:
        - email: eck@elastic.co
          name: Elastic
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Elastic
        relatedImages:
        - registry.connect.redhat.com/elastic/eck-operator@sha256:1750178c5dd716c031155f2e7058e19d4b7f1d3fdbb3e8a806b177a71be88500
        version: 3.1.0
      entries:
      - name: elasticsearch-eck-operator-certified.v3.1.0
        version: 3.1.0
      - name: elasticsearch-eck-operator-certified.v3.0.0
        version: 3.0.0
      - name: elasticsearch-eck-operator-certified.v2.16.1
        version: 2.16.1
      - name: elasticsearch-eck-operator-certified.v2.16.0
        version: 2.16.0
      - name: elasticsearch-eck-operator-certified.v2.15.0
        version: 2.15.0
      - name: elasticsearch-eck-operator-certified.v2.14.0
        version: 2.14.0
      - name: elasticsearch-eck-operator-certified.v2.13.0
        version: 2.13.0
      - name: elasticsearch-eck-operator-certified.v2.12.1
        version: 2.12.1
      - name: elasticsearch-eck-operator-certified.v2.12.0
        version: 2.12.0
      - name: elasticsearch-eck-operator-certified.v2.11.1
        version: 2.11.1
      - name: elasticsearch-eck-operator-certified.v2.11.0
        version: 2.11.0
      - name: elasticsearch-eck-operator-certified.v2.10.0
        version: 2.10.0
      - name: elasticsearch-eck-operator-certified.v2.9.0
        version: 2.9.0
      - name: elasticsearch-eck-operator-certified.v2.8.0
        version: 2.8.0
      - name: elasticsearch-eck-operator-certified.v2.7.0
        version: 2.7.0
      - name: elasticsearch-eck-operator-certified.v2.6.2
        version: 2.6.2
      - name: elasticsearch-eck-operator-certified.v2.6.1
        version: 2.6.1
      - name: elasticsearch-eck-operator-certified.v2.5.0
        version: 2.5.0
      - name: elasticsearch-eck-operator-certified.v2.4.0
        version: 2.4.0
      - name: elasticsearch-eck-operator-certified.v2.3.0
        version: 2.3.0
      - name: elasticsearch-eck-operator-certified.v2.2.0
        version: 2.2.0
      - name: elasticsearch-eck-operator-certified.v2.1.0
        version: 2.1.0
      - name: elasticsearch-eck-operator-certified.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: elasticsearch-eck-operator-certified
    provider:
      name: Elastic
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Portworx by PureStorage
      provider-url: ""
    name: portworx-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: portworx-operator.v25.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageCluster",
                "metadata": {
                  "name": "portworx",
                  "namespace": "test-operator",
                  "annotations": {
                    "portworx.io/is-openshift": "true"
                  }
                },
                "spec": {}
              },
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageNode",
                "metadata": {
                  "name": "example",
                  "namespace": "test-operator"
                },
                "spec": {}
              }
            ]
          capabilities: Auto Pilot
          categories: Storage
          certified: "true"
          console.openshift.io/plugins: '["portworx"]'
          containerImage: registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
          createdAt: "2025-09-06T19:38:43Z"
          description: Cloud native storage solution for production workloads
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "core.libopenstorage.org/v1",
              "kind": "StorageCluster",
              "metadata": {
                "name": "portworx",
                "annotations": {
                  "portworx.io/is-openshift": "true"
                }
              }
            }
          support: Portworx, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Storage Cluster installs Portworx in the cluster. It has
              all the necessary configurations to setup and update a Portworx cluster.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.core.libopenstorage.org
            version: v1
          - description: Do not create Storage Node as it is internally created by
              the operator. It represents the status of a Portworx node.
            displayName: Storage Node
            kind: StorageNode
            name: storagenodes.core.libopenstorage.org
            version: v1
        description: |
          Portworx-Enterprise is the most widely-used and reliable cloud-native
          storage solution for production workloads and provides high-availability,
          data protection and security for containerized applications.

          Portworx Enterprise enables you to migrate entire applications, including
          data, between clusters in a single data center or cloud, or between clouds,
          with a single kubectl command.

          The cloud native storage and data management platform that enterprises trust
          to manage data in containers now has an operator which simplifies the install,
          configuration, upgrades and manages the Portworx Enterprise cluster lifecycle.

          Learn more about the Portworx Enterprise
          [the data platform for Kubernetes](https://portworx.com/products/introduction)

          To learn more about the platform features, please visit our
          [product features page](https://portworx.com/products/features)

          ### About Portworx

          Portworx is the solution for running stateful containers in production,
          designed with DevOps in mind. With Portworx, users can manage any database
          or stateful service on any infrastructure using any container scheduler,
          including Kubernetes, Mesosphere DC/OS, and Docker Swarm. Portworx solves
          the five most common problems DevOps teams encounter when running stateful
          services in production: persistence, high availability, data automation,
          security, and support for multiple data stores and infrastructure.

          ### How to install StorageCluster

          To customize your cluster's configuration (specification), use the
          [Spec Generator](https://central.portworx.com/) from PX-Central.

          ### Prerequisite

          Ensure ports 17001-17020 on worker nodes are reachable from master and other worker nodes.

          ### Tutorials

          * [Portworx Enterprise on Openshift](https://portworx.com/openshift)

          * [Stateful applications on Kubernetes](https://docs.portworx.com/portworx-enterprise/concepts/kubernetes-storage-101/applications)

          * [Portworx Enterprise on Kubernetes](https://docs.portworx.com/portworx-install-with-kubernetes)

          * [Kafka on Kubernetes](https://portworx.com/kafka-kubernetes)

          * [Elastisearch on Kubernetes](https://portworx.com/elasticsearch-kubernetes)

          * [PostgreSQL on Kubernetes](https://portworx.com/postgres-kubernetes/)

          * [MongoDB on Kubernetes](https://portworx.com/mongodb-kubernetes/)

          * [Cassandra on Kubernetes](https://portworx.com/cassandra-kubernetes/)

          * [Kubernetes backup and recovery](https://portworx.com/kubernetes-backup/)

          * [Disaster Recovery for Kubernetes](https://portworx.com/kubernetes-disaster-recovery/)

          ### Uninstall

          Deleting the StorageCluster object for Portworx cluster does not stop Portworx
          service running on the nodes, to avoid application downtime.

          To uninstall Portworx completely without wiping the data, you should add the
          following delete strategy to the StorageCluster spec:
          ```
          spec:
            deleteStrategy:
              type: Uninstall
          ```
          **Caution:** To uninstall Portworx and **wipe all the data**, you should use the following
          delete strategy:
          ```
          spec:
            deleteStrategy:
              type: UninstallAndWipe
          ```
        displayName: Portworx Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - portworx
        - persistent storage
        - storage
        - cloud native
        - open source
        links:
        - name: Product Features
          url: https://portworx.com/products/features
        - name: Documentation
          url: https://docs.portworx.com/portworx-enterprise/platform/openshift
        - name: Support
          url: https://docs.portworx.com/portworx-enterprise/support/contact-support
        - name: Source Code
          url: https://github.com/libopenstorage/operator
        maintainers:
        - email: support@portworx.com
          name: Portworx
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Portworx by PureStorage
        relatedImages:
        - registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
        version: 25.3.1
      entries:
      - name: portworx-operator.v25.3.1
        version: 25.3.1
      - name: portworx-operator.v25.3.0
        version: 25.3.0
      - name: portworx-operator.v25.2.2
        version: 25.2.2
      - name: portworx-operator.v25.2.1
        version: 25.2.1
      - name: portworx-operator.v25.2.0
        version: 25.2.0
      - name: portworx-operator.v24.2.4
        version: 24.2.4
      - name: portworx-operator.v24.2.3
        version: 24.2.3
      - name: portworx-operator.v24.2.2
        version: 24.2.2
      - name: portworx-operator.v24.2.1
        version: 24.2.1
      - name: portworx-operator.v24.2.0
        version: 24.2.0
      - name: portworx-operator.v24.1.3
        version: 24.1.3
      - name: portworx-operator.v24.1.2
        version: 24.1.2
      - name: portworx-operator.v24.1.1
        version: 24.1.1
      - name: portworx-operator.v24.1.0
        version: 24.1.0
      - name: portworx-operator.v23.10.5
        version: 23.10.5
      - name: portworx-operator.v23.10.4
        version: 23.10.4
      - name: portworx-operator.v23.10.3
        version: 23.10.3
      - name: portworx-operator.v23.10.2
        version: 23.10.2
      - name: portworx-operator.v23.10.1
        version: 23.10.1
      - name: portworx-operator.v23.7.0
        version: 23.7.0
      - name: portworx-operator.v23.5.1
        version: 23.5.1
      - name: portworx-operator.v23.5.0
        version: 23.5.0
      - name: portworx-operator.v23.4.0
        version: 23.4.0
      - name: portworx-operator.v23.3.1
        version: 23.3.1
      - name: portworx-operator.v23.3.0
        version: 23.3.0
      - name: portworx-operator.v1.10.5
        version: 1.10.5
      - name: portworx-operator.v1.10.4
        version: 1.10.4
      - name: portworx-operator.v1.10.3
        version: 1.10.3
      - name: portworx-operator.v1.10.2
        version: 1.10.2
      - name: portworx-operator.v1.10.1
        version: 1.10.1
      - name: portworx-operator.v1.10.0
        version: 1.10.0
      - name: portworx-operator.v1.9.1
        version: 1.9.1
      - name: portworx-operator.v1.9.0
        version: 1.9.0
      - name: portworx-operator.v1.8.1
        version: 1.8.1
      - name: portworx-operator.v1.8.0
        version: 1.8.0
      - name: portworx-operator.v1.7.1
        version: 1.7.1
      - name: portworx-operator.v1.6.1
        version: 1.6.1
      name: alpha
    - currentCSV: portworx-operator.v25.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageCluster",
                "metadata": {
                  "name": "portworx",
                  "namespace": "test-operator",
                  "annotations": {
                    "portworx.io/is-openshift": "true"
                  }
                },
                "spec": {}
              },
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageNode",
                "metadata": {
                  "name": "example",
                  "namespace": "test-operator"
                },
                "spec": {}
              }
            ]
          capabilities: Auto Pilot
          categories: Storage
          certified: "true"
          console.openshift.io/plugins: '["portworx"]'
          containerImage: registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
          createdAt: "2025-09-06T19:38:43Z"
          description: Cloud native storage solution for production workloads
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "core.libopenstorage.org/v1",
              "kind": "StorageCluster",
              "metadata": {
                "name": "portworx",
                "annotations": {
                  "portworx.io/is-openshift": "true"
                }
              }
            }
          support: Portworx, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Storage Cluster installs Portworx in the cluster. It has
              all the necessary configurations to setup and update a Portworx cluster.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.core.libopenstorage.org
            version: v1
          - description: Do not create Storage Node as it is internally created by
              the operator. It represents the status of a Portworx node.
            displayName: Storage Node
            kind: StorageNode
            name: storagenodes.core.libopenstorage.org
            version: v1
        description: |
          Portworx-Enterprise is the most widely-used and reliable cloud-native
          storage solution for production workloads and provides high-availability,
          data protection and security for containerized applications.

          Portworx Enterprise enables you to migrate entire applications, including
          data, between clusters in a single data center or cloud, or between clouds,
          with a single kubectl command.

          The cloud native storage and data management platform that enterprises trust
          to manage data in containers now has an operator which simplifies the install,
          configuration, upgrades and manages the Portworx Enterprise cluster lifecycle.

          Learn more about the Portworx Enterprise
          [the data platform for Kubernetes](https://portworx.com/products/introduction)

          To learn more about the platform features, please visit our
          [product features page](https://portworx.com/products/features)

          ### About Portworx

          Portworx is the solution for running stateful containers in production,
          designed with DevOps in mind. With Portworx, users can manage any database
          or stateful service on any infrastructure using any container scheduler,
          including Kubernetes, Mesosphere DC/OS, and Docker Swarm. Portworx solves
          the five most common problems DevOps teams encounter when running stateful
          services in production: persistence, high availability, data automation,
          security, and support for multiple data stores and infrastructure.

          ### How to install StorageCluster

          To customize your cluster's configuration (specification), use the
          [Spec Generator](https://central.portworx.com/) from PX-Central.

          ### Prerequisite

          Ensure ports 17001-17020 on worker nodes are reachable from master and other worker nodes.

          ### Tutorials

          * [Portworx Enterprise on Openshift](https://portworx.com/openshift)

          * [Stateful applications on Kubernetes](https://docs.portworx.com/portworx-enterprise/concepts/kubernetes-storage-101/applications)

          * [Portworx Enterprise on Kubernetes](https://docs.portworx.com/portworx-install-with-kubernetes)

          * [Kafka on Kubernetes](https://portworx.com/kafka-kubernetes)

          * [Elastisearch on Kubernetes](https://portworx.com/elasticsearch-kubernetes)

          * [PostgreSQL on Kubernetes](https://portworx.com/postgres-kubernetes/)

          * [MongoDB on Kubernetes](https://portworx.com/mongodb-kubernetes/)

          * [Cassandra on Kubernetes](https://portworx.com/cassandra-kubernetes/)

          * [Kubernetes backup and recovery](https://portworx.com/kubernetes-backup/)

          * [Disaster Recovery for Kubernetes](https://portworx.com/kubernetes-disaster-recovery/)

          ### Uninstall

          Deleting the StorageCluster object for Portworx cluster does not stop Portworx
          service running on the nodes, to avoid application downtime.

          To uninstall Portworx completely without wiping the data, you should add the
          following delete strategy to the StorageCluster spec:
          ```
          spec:
            deleteStrategy:
              type: Uninstall
          ```
          **Caution:** To uninstall Portworx and **wipe all the data**, you should use the following
          delete strategy:
          ```
          spec:
            deleteStrategy:
              type: UninstallAndWipe
          ```
        displayName: Portworx Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - portworx
        - persistent storage
        - storage
        - cloud native
        - open source
        links:
        - name: Product Features
          url: https://portworx.com/products/features
        - name: Documentation
          url: https://docs.portworx.com/portworx-enterprise/platform/openshift
        - name: Support
          url: https://docs.portworx.com/portworx-enterprise/support/contact-support
        - name: Source Code
          url: https://github.com/libopenstorage/operator
        maintainers:
        - email: support@portworx.com
          name: Portworx
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Portworx by PureStorage
        relatedImages:
        - registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
        version: 25.3.1
      entries:
      - name: portworx-operator.v25.3.1
        version: 25.3.1
      - name: portworx-operator.v25.3.0
        version: 25.3.0
      - name: portworx-operator.v25.2.2
        version: 25.2.2
      - name: portworx-operator.v25.2.1
        version: 25.2.1
      - name: portworx-operator.v25.2.0
        version: 25.2.0
      - name: portworx-operator.v24.2.4
        version: 24.2.4
      - name: portworx-operator.v24.2.3
        version: 24.2.3
      - name: portworx-operator.v24.2.2
        version: 24.2.2
      - name: portworx-operator.v24.2.1
        version: 24.2.1
      - name: portworx-operator.v24.2.0
        version: 24.2.0
      - name: portworx-operator.v24.1.3
        version: 24.1.3
      - name: portworx-operator.v24.1.2
        version: 24.1.2
      - name: portworx-operator.v24.1.1
        version: 24.1.1
      - name: portworx-operator.v24.1.0
        version: 24.1.0
      - name: portworx-operator.v23.10.5
        version: 23.10.5
      - name: portworx-operator.v23.10.4
        version: 23.10.4
      - name: portworx-operator.v23.10.3
        version: 23.10.3
      - name: portworx-operator.v23.10.2
        version: 23.10.2
      - name: portworx-operator.v23.10.1
        version: 23.10.1
      - name: portworx-operator.v23.7.0
        version: 23.7.0
      - name: portworx-operator.v23.5.1
        version: 23.5.1
      - name: portworx-operator.v23.5.0
        version: 23.5.0
      - name: portworx-operator.v23.4.0
        version: 23.4.0
      - name: portworx-operator.v23.3.1
        version: 23.3.1
      - name: portworx-operator.v23.3.0
        version: 23.3.0
      - name: portworx-operator.v1.10.5
        version: 1.10.5
      - name: portworx-operator.v1.10.4
        version: 1.10.4
      - name: portworx-operator.v1.10.3
        version: 1.10.3
      - name: portworx-operator.v1.10.2
        version: 1.10.2
      - name: portworx-operator.v1.10.1
        version: 1.10.1
      - name: portworx-operator.v1.10.0
        version: 1.10.0
      - name: portworx-operator.v1.9.1
        version: 1.9.1
      - name: portworx-operator.v1.9.0
        version: 1.9.0
      - name: portworx-operator.v1.8.1
        version: 1.8.1
      - name: portworx-operator.v1.8.0
        version: 1.8.0
      - name: portworx-operator.v1.7.1
        version: 1.7.1
      - name: portworx-operator.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: portworx-certified
    provider:
      name: Portworx by PureStorage
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: clusterresourceoverride
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clusterresourceoverride-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.autoscaling.openshift.io/v1",
                "kind": "ClusterResourceOverride",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "podResourceOverride": {
                    "spec": {
                      "memoryRequestToLimitPercent": 50,
                      "cpuRequestToLimitPercent": 25,
                      "limitCPUToMemoryPercent": 200
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certifiedLevel: Primed
          containerImage: registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9-operator@sha256:40b19c94f688c3a58189fcb404df35a40a8acbbdbf73fadb8818a86fd14a9ab9
          createdAt: 2019/11/15
          description: An operator to manage the OpenShift ClusterResourceOverride
            Mutating Admission Webhook Server
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          healthIndex: B
          olm.skipRange: '>=4.3.0 <4.16.0-202509240123'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-resource-override-admission-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of ClusterResourceOverride Admission
              Webhook
            displayName: ClusterResourceOverride
            kind: ClusterResourceOverride
            name: clusterresourceoverrides.operator.autoscaling.openshift.io
            version: v1
        description: "ClusterResourceOverride\n==============\n\nContainers can specify
          compute resource requests and limits. Requests are used for scheduling your
          container and provide a minimum service guarantee. Limits constrain the
          amount of compute resource that may be consumed on your node.\n\nThe scheduler
          attempts to optimize the compute resource use across all nodes in your cluster.
          It places pods onto specific nodes, taking the pods' compute resource requests
          and nodes' available capacity into consideration.\n\nRequests and limits
          enable administrators to allow and manage the overcommitment of resources
          on a node, which may be desirable in development environments where a trade
          off of guaranteed performance for capacity is acceptable.\n\n### Requests
          and Limits\n\nFor each compute resource, a container may specify a resource
          request and limit. Scheduling decisions are made based on the request to
          ensure that a node has enough capacity available to meet the requested value.
          If a container specifies limits, but omits requests, the requests are defaulted
          to the limits. A container is not able to exceed the specified limit on
          the node.\n\nThe enforcement of limits is dependent upon the compute resource
          type. If a container makes no request or limit, the container is scheduled
          to a node with no resource guarantees. In practice, the container is able
          to consume as much of the specified resource as is available with the lowest
          local priority. In low resource situations, containers that specify no resource
          requests are given the lowest quality of service.\n\n### Compute Resources\n\nThe
          node-enforced behavior for compute resources is specific to the resource
          type.\n\n#### CPU\n\nA container is guaranteed the amount of CPU it requests
          and is additionally able to consume excess CPU available on the node, up
          to any limit specified by the container. If multiple containers are attempting
          to use excess CPU, CPU time is distributed based on the amount of CPU requested
          by each container.\n\nFor example, if one container requested 500m of CPU
          time and another container requested 250m of CPU time, then any extra CPU
          time available on the node is distributed among the containers in a 2:1
          ratio. If a container specified a limit, it will be throttled not to use
          more CPU than the specified limit.\n\nCPU requests are enforced using the
          CFS shares support in the Linux kernel. By default, CPU limits are enforced
          using the CFS quota support in the Linux kernel over a 100ms measuring interval,
          though this can be disabled.\n\n#### Memory\n\nA container is guaranteed
          the amount of memory it requests. A container may use more memory than requested,
          but once it exceeds its requested amount, it could be killed in a low memory
          situation on the node.\n\nIf a container uses less memory than requested,
          it will not be killed unless system tasks or daemons need more memory than
          was accounted for in the node’s resource reservation. If a container specifies
          a limit on memory, it is immediately killed if it exceeds the limit amount.\n\n###
          Configuring the Cluster for Overcommitment\n\nScheduling is based on resources
          requested, while quota and hard limits refer to resource limits, which can
          be set higher than requested resources. The difference between request and
          limit determines the level of overcommit; for instance, if a container is
          given a memory request of 1Gi and a memory limit of 2Gi, it is scheduled
          based on the 1Gi request being available on the node, but could use up to
          2Gi; so it is 200% overcommitted. If OpenShift Container Platform administrators
          would like to control the level of overcommit and manage container density
          on nodes, ClusterResourceOverride Admission Webhook can be configured to
          override the ratio between request and limit set on developer containers.
          In conjunction with a per-project LimitRange specifying limits and defaults,
          this adjusts the container limit and request to achieve the desired level
          of overcommit.\n\nThis requires creating a custom resource of `ClusterResourceOverride`
          type as in the following example:\n\n    \n    - apiVersion: operator.autoscaling.openshift.io/v1\n
          \   - kind: ClusterResourceOverride\n    - metadata:\n    -   name: cluster\n
          \   - spec:\n    -   podResourceOverride:\n    -     spec:\n    -       memoryRequestToLimitPercent:
          25\n    -       cpuRequestToLimitPercent: 25\n    -       limitCPUToMemoryPercent:
          200\n    \n **memoryRequestToLimitPercent**: (optional, 1-100) If a container
          memory limit has been specified or defaulted, the memory request is overridden
          to this percentage of the limit.\n\n **cpuRequestToLimitPercent**: (optional,
          1-100) If a container CPU limit has been specified or defaulted, the CPU
          request is overridden to this percentage of the limit.\n\n **limitCPUToMemoryPercent**:
          (optional, positive integer) If a container memory limit has been specified
          or defaulted, the CPU limit is overridden to a percentage of the memory
          limit, with a 100 percentage scaling 1Gi of RAM to equal 1 CPU core. This
          is processed prior to overriding CPU request (if configured).\n\n Note that
          these overrides have no effect if no limits have been set on containers.
          [Create a LimitRange object] (https://docs.openshift.com/container-platform/3.3/admin_guide/limits.html#admin-guide-limits)
          with default limits (per individual project, or in the [project template](https://docs.openshift.com/container-platform/3.3/admin_guide/managing_projects.html#modifying-the-template-for-new-projects))
          in order to ensure that the overrides apply.\n\nWhen configured, overrides
          can be enabled per-project by applying the following label.\n    ```\n    clusterresourceoverrides.admission.autoscaling.openshift.io/enabled:
          \"true\"\n    ```\n\n"
        displayName: ClusterResourceOverride Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - deschedule
        - scale
        - binpack
        - efficiency
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9-operator@sha256:40b19c94f688c3a58189fcb404df35a40a8acbbdbf73fadb8818a86fd14a9ab9
        - registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9@sha256:e4992c556fe6be50fa30a45b05cffd6029ebd8914eca7c1c683564ff869e0e7d
        version: 4.16.0-202509240123
      entries:
      - name: clusterresourceoverride-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: clusterresourceoverride-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: clusterresourceoverride-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: clusterresourceoverride-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: clusterresourceoverride-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: clusterresourceoverride-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: clusterresourceoverride-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: clusterresourceoverride-operator.v4.16.0-202505092110
        version: 4.16.0-202505092110
      - name: clusterresourceoverride-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: clusterresourceoverride-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: clusterresourceoverride-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: clusterresourceoverride-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: clusterresourceoverride-operator.v4.16.0-202501241104
        version: 4.16.0-202501241104
      - name: clusterresourceoverride-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: clusterresourceoverride-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: clusterresourceoverride-operator.v4.16.0-202412021735
        version: 4.16.0-202412021735
      - name: clusterresourceoverride-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: clusterresourceoverride-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: clusterresourceoverride-operator.v4.16.0-202411081636
        version: 4.16.0-202411081636
      - name: clusterresourceoverride-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: clusterresourceoverride-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: clusterresourceoverride-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: clusterresourceoverride-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: clusterresourceoverride-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: clusterresourceoverride-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: clusterresourceoverride-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: clusterresourceoverride-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: clusterresourceoverride-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: clusterresourceoverride-operator.v4.16.0-202407040206
        version: 4.16.0-202407040206
      - name: clusterresourceoverride-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: clusterresourceoverride-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: clusterresourceoverride-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: clusterresourceoverride-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: clusterresourceoverride
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicurio-registry-3
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-3.v3.0.15-r1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "simple"
                },
                "spec": {
                  "app": {
                    "ingress": {
                      "host": "simple-app.apps.cluster.example"
                    }
                  },
                  "ui": {
                    "ingress": {
                      "host": "simple-ui.apps.cluster.example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/apicurio/apicurio-registry-rhel9-operator@sha256:e6d2ceab8accdb6e1b9766733039e24df620d7128eb1e13e60a61339130d52a7
          createdAt: "2025-09-29"
          description: Deploy and manage Red Hat build of Apicurio Registry 3 on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations",
            "Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Red Hat build
              of Apicurio Registry version 3.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Red Hat build of Apicurio Registry 3

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Allows you manage event schemas and API designs using the Apicurio Registry web console, REST API, Maven plug-in, or Java client
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, MySQL\*, SQL Server\*, in-memory\*
          - Compatible with Confluent APIs
          - Runs on a lightweight Quarkus platform

          _Note: Storage options marked with * have not yet been tested by Red Hat for production deployments._

          ## Red Hat build of Apicurio Registry 3 Operator

          Provides a quick and easy way to deploy and manage Red Hat build of Apicurio Registry 3 on OpenShift.

          **Features**
          - Supports installation and configuration of Red Hat build of Apicurio Registry 3
          - Perform a rolling upgrade of Red Hat build of Apicurio Registry 3

          ## Prerequisites

          This operator does not deploy storage for Red Hat build of Apicurio Registry 3. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Red Hat build of Apicurio Registry 3 and Apicurio Registry 3 Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Red Hat build of Apicurio Registry 3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/apicurio/apicurio-registry-ui-rhel9@sha256:e4791546fd371b34c7580c6aca2faed8527c985c8029fa8087fb46fd94ab8feb
        - registry.redhat.io/apicurio/apicurio-registry-rhel9-operator@sha256:e6d2ceab8accdb6e1b9766733039e24df620d7128eb1e13e60a61339130d52a7
        - registry.redhat.io/apicurio/apicurio-registry-rhel9@sha256:c86d73d5e2b4146f016ce7ec67265fe80bd596f472247eb9f5590cc0c89dba75
        version: 3.0.15-r1
      entries:
      - name: apicurio-registry-3.v3.0.15-r1
        version: 3.0.15-r1
      - name: apicurio-registry-3.v3.0.9-r3
        version: 3.0.9-r3
      - name: apicurio-registry-3.v3.0.9-r2
        version: 3.0.9-r2
      - name: apicurio-registry-3.v3.0.9-r1
        version: 3.0.9-r1
      - name: apicurio-registry-3.v3.0.7-r2
        version: 3.0.7-r2
      - name: apicurio-registry-3.v3.0.7-r1
        version: 3.0.7-r1
      - name: apicurio-registry-3.v3.0.6-beta.r1
        version: 3.0.6-beta.r1
      - name: apicurio-registry-3.v3.0.6-beta.r1-0.1740495889.p
        version: 3.0.6-beta.r1+0.1740495889.p
      - name: apicurio-registry-3.v3.0.6-beta
        version: 3.0.6-beta
      name: 3.x
    defaultChannel: 3.x
    packageName: apicurio-registry-3
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SusQL Operator Contributors
      provider-url: https://github.com/sustainable-computing-io/susql-operator
    name: susql-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: susql-operator.v0.0.35
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "susql.ibm.com/v1",
                "kind": "LabelGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "susql-operator",
                    "app.kubernetes.io/instance": "labelgroup-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "susql-operator",
                    "app.kubernetes.io/part-of": "susql-operator"
                  },
                  "name": "labelgroup-sample"
                },
                "spec": {
                  "labels": [
                    "labelgroup-sample-1",
                    "labelgroup-sample-2"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: quay.io/sustainable_computing_io/susql_operator:0.0.34
          createdAt: "2025-07-10T01:20:37Z"
          description: 'Aggregates energy and CO2 emission data for pods tagged with
            SusQL labels '
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/sustainable-computing-io/susql-operator
          support: https://github.com/sustainable-computing-io/susql-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LabelGroup is the Schema for the LabelGroups API
            displayName: Label Group
            kind: LabelGroup
            name: labelgroups.susql.ibm.com
            version: v1
        description: |-
          ### About this Operator

          SusQL is a Kubernetes operator that aggregates energy and estimated
          carbon dioxide emission data for pods tagged with SusQL specific
          labels. The energy measurements are taken from Kepler which should
          be deployed in the cluster before using SusQL.

          ### Prerequisites

          1. Deployment of [Kepler](https://sustainable-computing.io/) on the cluster
          2. Ensure that [User Project Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          is enabled to monitor energy consumed in user projects.
          3. Creation of a SusQL Service Monitor:
                `oc apply -n <NAMESPACE> -f https://raw.githubusercontent.com/sustainable-computing-io/susql-operator/main/hack/susql-servicemonitor.yaml`

          ### API Backward Compatibility

          Until Operator supports **seamless upgrades**, a new version of the operator may
          introduce a change that is **NOT** backwards compatible with the previous version.
          Thus, to upgrade the operator, uninstall the already installed version completely
          (including CRDs) and install the new version. For example to upgrade from 0.1.0
          to 0.2.0, you must first uninstall 0.1.0.

          ### LabelGroup spec Example

          ```
          apiVersion: susql.ibm.com/v1
          kind: LabelGroup
          metadata:
            name: labelgroup-name
            namespace: susql-operator
          spec:
            labels:
              - my-label-1
              - my-label-2
          ```

          ### Documentation

          Documentation can be found at [SusQL Operator](https://github.com/sustainable-computing-io/susql-operator#readme).

          ### Getting help

          If you encounter any issues while using the SusQL Operator, you can create an
          [issue](https://github.com/sustainable-computing-io/susql-operator/issues)
          on our GitHub repo for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by raising [issues](https://github.com/sustainable-computing-io/susql-operator/issues)
          related to susql-operator.

          ### License
          susql-operator is licensed under the Apache 2.0 license
        displayName: SusQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - aggregation
        - sustainability
        - energy
        - kepler
        - susql
        - carbon dioxide (CO2) emissions
        - green computing
        links:
        - name: SusQL Operator
          url: https://github.com/sustainable-computing-io/susql-operator
        - name: Sustainable Computing
          url: https://github.com/sustainable-computing-io
        maintainers:
        - email: trent@jp.ibm.com
          name: Scott Trent
        - email: mohammed.munir.abdi@ibm.com
          name: Mohammed Abdi
        maturity: alpha
        provider:
          name: SusQL Operator Contributors
          url: https://github.com/sustainable-computing-io/susql-operator
        relatedImages:
        - quay.io/sustainable_computing_io/susql_operator:0.0.34
        version: 0.0.35
      entries:
      - name: susql-operator.v0.0.35
        version: 0.0.35
      - name: susql-operator.v0.0.34
        version: 0.0.34
      - name: susql-operator.v0.0.33
        version: 0.0.33
      - name: susql-operator.v0.0.32
        version: 0.0.32
      - name: susql-operator.v0.0.30
        version: 0.0.30
      - name: susql-operator.v0.0.29
        version: 0.0.29
      - name: susql-operator.v0.0.28
        version: 0.0.28
      - name: susql-operator.v0.0.27
        version: 0.0.27
      name: alpha
    defaultChannel: alpha
    packageName: susql-operator
    provider:
      name: SusQL Operator Contributors
      url: https://github.com/sustainable-computing-io/susql-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CodeFlare
      provider-url: https://github.com/project-codeflare
    name: codeflare-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: codeflare-operator.v1.16.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: quay.io/project-codeflare/codeflare-operator:v1.16.0
          createdAt: "2025-06-16T10:53:51Z"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/project-codeflare/codeflare-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AppWrapper
            name: appwrappers.workload.codeflare.dev
            version: v1beta2
        description: CodeFlare allows you to scale complex pipelines anywhere
        displayName: CodeFlare Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Pipelines
        - Scaling
        - App
        - Jobs
        links:
        - name: Codeflare Operator
          url: https://codeflare.dev/
        maintainers:
        - email: aasthana@redhat.com
          name: Anish Asthana
        - email: kpostlet@redhat.com
          name: Kevin Postlethwait
        - email: meyceoz@redhat.com
          name: Mustafa Eyceoz
        maturity: alpha
        provider:
          name: CodeFlare
          url: https://github.com/project-codeflare
        relatedImages:
        - quay.io/project-codeflare/codeflare-operator:v1.16.0
        version: 1.16.0
      entries:
      - name: codeflare-operator.v1.16.0
        version: 1.16.0
      - name: codeflare-operator.v1.15.0
        version: 1.15.0
      - name: codeflare-operator.v1.14.0
        version: 1.14.0
      - name: codeflare-operator.v1.13.0
        version: 1.13.0
      - name: codeflare-operator.v1.12.0
        version: 1.12.0
      - name: codeflare-operator.v1.11.0
        version: 1.11.0
      - name: codeflare-operator.v1.10.0
        version: 1.10.0
      - name: codeflare-operator.v1.9.0
        version: 1.9.0
      - name: codeflare-operator.v1.8.0
        version: 1.8.0
      - name: codeflare-operator.v1.7.0
        version: 1.7.0
      - name: codeflare-operator.v1.6.0
        version: 1.6.0
      - name: codeflare-operator.v1.5.0
        version: 1.5.0
      - name: codeflare-operator.v1.4.4
        version: 1.4.4
      - name: codeflare-operator.v1.4.3
        version: 1.4.3
      - name: codeflare-operator.v1.4.2
        version: 1.4.2
      - name: codeflare-operator.v1.4.1
        version: 1.4.1
      - name: codeflare-operator.v1.4.0
        version: 1.4.0
      - name: codeflare-operator.v1.3.1
        version: 1.3.1
      - name: codeflare-operator.v1.3.0
        version: 1.3.0
      - name: codeflare-operator.v1.2.0
        version: 1.2.0
      - name: codeflare-operator.v1.1.0
        version: 1.1.0
      - name: codeflare-operator.v1.0.1
        version: 1.0.1
      - name: codeflare-operator.v1.0.0
        version: 1.0.0
      - name: codeflare-operator.v1.0.0-rc.4
        version: 1.0.0-rc.4
      - name: codeflare-operator.v1.0.0-rc.3
        version: 1.0.0-rc.3
      - name: codeflare-operator.v1.0.0-rc.2
        version: 1.0.0-rc.2
      - name: codeflare-operator.v1.0.0-rc.1
        version: 1.0.0-rc.1
      - name: codeflare-operator.v0.2.3
        version: 0.2.3
      - name: codeflare-operator.v0.2.2
        version: 0.2.2
      - name: codeflare-operator.v0.2.1
        version: 0.2.1
      - name: codeflare-operator.v0.2.0
        version: 0.2.0
      - name: codeflare-operator.v0.1.0
        version: 0.1.0
      - name: codeflare-operator.v0.0.6
        version: 0.0.6
      - name: codeflare-operator.v0.0.5
        version: 0.0.5
      - name: codeflare-operator.v0.0.4
        version: 0.0.4
      - name: codeflare-operator.v0.0.3
        version: 0.0.3
      - name: codeflare-operator.v0.0.2
        version: 0.0.2
      - name: codeflare-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: codeflare-operator
    provider:
      name: CodeFlare
      url: https://github.com/project-codeflare
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: opdev.github.io
    name: airflow-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airflow-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "workflow.apache.org/v1alpha1",
                "kind": "AirFlow",
                "metadata": {
                  "name": "airflow-helm"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/airflow-helm-operator:0.0.1
          createdAt: "2023-06-28T16:57:21Z"
          description: An experimental operator that installs Apache Airflow.
          operatorframework.io/suggested-namespace: airflow-helm
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys Apache Airflow based on the official helm chart in
              openshift
            kind: AirFlow
            name: airflows.workflow.apache.org
            version: v1alpha1
        description: Deploys Apache Airflow based on the official helm chart with
          openshift specific modifications applied by default
        displayName: Airflow Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        - data
        - batch
        - scheduler
        - data workflow
        links:
        - name: Airflow Helm Operator
          url: https://airflow-helm-operator.domain
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: opdev.github.io
        relatedImages:
        - quay.io/opdev/airflow-helm-operator:0.0.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 0.0.1
      entries:
      - name: airflow-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: airflow-helm-operator
    provider:
      name: opdev
      url: opdev.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-sso
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-sso.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Sso",
                "metadata": {
                  "name": "sso-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "kcos-keycloak": {
                    "auth": {
                      "baseURL": "https://keycloak.kcos-sso.svc.cluster.local:8443",
                      "enabled": true,
                      "sessionCookie": "AT"
                    },
                    "authproxy": {
                      "image": "prod.harbor.keysight.digital/loadcore/kcos-auth-proxy@sha256:a777ee5116280fbeb1e1342445fef5135f9ca61e983a3e2821f9ca4b2038b044",
                      "initContainer": {
                        "image": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf"
                      }
                    },
                    "crsCleanupJob": {
                      "clientExec": "oc",
                      "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "realm": {
                      "accessTokenLifespan": 266400,
                      "accessTokenLifespanForImplicitFlow": 266400,
                      "admin": {
                        "password": "admin"
                      },
                      "offlineSessionIdleTimeout": 2592000,
                      "offlineSessionMaxLifespan": 5184000,
                      "registrationAllowed": true,
                      "ssoSessionIdleTimeout": 266400,
                      "ssoSessionMaxLifespan": 266400
                    }
                  },
                  "subcharts": {
                    "enable": {
                      "kcos-keycloak": true,
                      "keycloak-operator": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:29:57Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Sso
            name: ssoes.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS SSO Operator
        displayName: Keysight KCOS SSO Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Sso
          url: https://keysight-kcos-sso.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-sso@sha256:298417fac2fc053bdc352f77b59b14e39eadd66c871cfb547304a517489f294e
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - prod.harbor.keysight.digital/loadcore/kcos-auth-proxy@sha256:a777ee5116280fbeb1e1342445fef5135f9ca61e983a3e2821f9ca4b2038b044
        version: 0.1.21
      entries:
      - name: keysight-kcos-sso.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-sso
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-cpd-wspipelines
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-wspipelines.v10.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wspipelines.cpd.ibm.com/v1beta1",
                "kind": "WSPipelines",
                "metadata": {
                  "name": "wspipelines-cr",
                  "labels": {
                    "app.kubernetes.io/name": "wspipelines",
                    "app.kubernetes.io/instance": "wspipelines.wspipelines.cpd.ibm.com",
                    "app.kubernetes.io/managed-by": "wspipelines-operator",
                    "wspipelinesBuildNumber": "2017"
                }
                },
                "spec": {
                  "version": "5.1.3",
                  "scaleConfig": "small",
                  "license": {
                    "accept": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:f9f616d17f816468957b69bbb09e6c92d678acb4ba7a40f1cb493d08fb98deb8
          createdAt: "2025-02-11"
          description: WS Pipelines CSV file
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-wspipelines
          olm.skipRange: '>=1.0.0 <10.3.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM watsonx.ai"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation For additional details regarding install parameters
              check: https://ibm.biz/BdaaL9 or https://ibm.biz/BdaaLC. License By
              installing this product you accept the license terms https://ibm.biz/BdaaL9
              or https://ibm.biz/BdaaLC.'
            kind: WSPipelines
            name: wspipelines.wspipelines.cpd.ibm.com
            version: v1beta1
        description: |-
          # Name
          IBM&reg; Watson Pipelines
          # Introduction
          ## Summary
          The Watson Pipelines editor provides a graphical interface for orchestrating an end-to-end flow of assets from creation through deployment. It enables to automate an end-to-end flow to prepare data, then create, train, deploy, monitor and update machine learning models and Python scripts. To design a pipeline you drag nodes onto the canvas, specify objects and parameters, then run and monitor the pipeline.
          ## Features
          Watson Pipelines are built off of Kubeflow pipelines on Tekton runtime and are fully integrated into the Watson Studio platform, allowing users to combine tools including:
          - Notebooks
          - Data refinery flows
          - AutoAI experiments
          - Web service / online deployments
          - Batch deployments
          - Import and export of Project / Space assets
          into repeatable and automated pipelines.
          # Details
          ## Prerequisites
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-pre-installation-tasks)
          ### Resources Required
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=planning-system-requirements)
          # Installing
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing)
          ## Configuration
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing)
          ## Storage
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=planning-storage-considerations)
          ## Limitations
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=issues-watson-studio)
          ## Documentation
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=overview)
          ## PodSecurityPolicy Requirements
          Custom PodSecurityPolicy definition:
          ```
          Not Applicable
          ```
          ## SecurityContextConstraints Requirements
          Custom SecurityContextConstraints definition:
          ```
          apiVersion: security.openshift.io/v1
          kind: SecurityContextConstraints
          allowHostDirVolumePlugin: false
          allowHostIPC: false
          allowHostNetwork: false
          allowHostPID: false
          allowHostPorts: false
          allowPrivilegeEscalation: false
          allowPrivilegedContainer: false
          allowedCapabilities: null
          defaultAddCapabilities: null
          fsGroup:
            type: MustRunAs
          metadata:
            annotations:
              include.release.openshift.io/self-managed-high-availability: "true"
              kubernetes.io/description: restrict denies access to all host features and requires pods to be run with a UID, and SELinux context that are allocated to the namespace.  This is the most restrictive SCC and it is used by default for authenticated users.
              release.openshift.io/create-only: "true"
            name: restrict
          priority: null
          readOnlyRootFilesystem: false
          requiredDropCapabilities:
          - ALL
          runAsUser:
            type: MustRunAsRange
          seLinuxContext:
            type: MustRunAs
          supplementalGroups:
            type: RunAsAny
          users: []
          volumes:
          - configMap
          - downwardAPI
          - emptyDir
          - persistentVolumeClaim
          - projected
          - secret
          ```
        displayName: Watson Pipelines
        installModes:
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: OwnNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - wspipelines
        - WSPipelines
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: v1
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:a477d47eaf8f8e569601d3c23c1def66f31c098ac3f5a1ae350618c460917dd8
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:0164d8fbea591bbd9d67d84603162c3781b08551eddcc2a8a842aafc828af18c
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:164f6063f3a50a36e28a240ae7c060c4f90266b7ed799bfdc075f90628bea7ff
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:ebcfc8941330da6991480dc18ac34d6df4fe879e562ee537e5c36df73f4bc536
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a24f4cb635e8b5759a5e2bda20ddefc6d5ea9b16fb1081592885c54aefb5480c
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:29cd5c9cd60bbcfe211337786616622b911e2319db9870eb9d38d4f630c03e75
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:f0c68715f78bd20c352c97dd48ac5a912cad58e19a00159b5e6878f6beeb00fb
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:a2949c3868c2c91bf35e50c64b51fcc24372986744b55520e21e9ce31fde59f0
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-bundle@sha256:842f2ac568db4db79cf0450b413b4b09b5f2180ef3541385b46fb650ac7066c0
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:f9f616d17f816468957b69bbb09e6c92d678acb4ba7a40f1cb493d08fb98deb8
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:61155acdb8982648ae5748be9f5944ba913db1590ec3e773206d6fbb1d8952cb
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:8d7625a0ba656425f601b6ec9a398a6d30511fe1de119cf0e29f680c704ea225
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:e4888fadfadb3462e3a9ad0395a526a2631872f70caf291dc18b60269ded199f
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:48b7cfa31b9a185176c20afc119e37346a6f9380d74000edf6c729b4677ca44e
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:19bb32e0389361e8314e627bcad2e9bd93984289485cf632466834353d9c3449
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:e29885c1ee083a73b952b0de8d1cd675b2b2830abf44d69a8577f56972a601fb
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:00f8bcee4ba7e0c5cf5f0c52e9e7707df367041c9f4dd5508dbefe0dbdba8877
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:99d0b3192e6e46167423505f48655e7d49af1a461ae00b95c9c7c9d4209e5b65
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:25a4bb9cea1c3bdbeeeb205622512bcba98163f9629eba35625756346ce9d652
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:6fc976bc622c2165dd8bf93c9244e1c29b7c973a4da7a5ce4d0a097a1c1ea4fb
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:ad2215adfb8dd2dc600bf5097b7f88c257d5fbaf541409c01b36cdb67c130236
        version: 10.3.0
      entries:
      - name: ibm-cpd-wspipelines.v10.3.0
        version: 10.3.0
      name: v10.3
    defaultChannel: v10.3
    packageName: ibm-cpd-wspipelines
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-enterprise
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-enterprise.v1.33.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "my-replica-set"
                },
                "spec": {
                  "credentials": "my-credentials",
                  "members": 3,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "sample-sharded-cluster"
                },
                "spec": {
                  "configServerCount": 3,
                  "credentials": "my-credentials",
                  "mongodsPerShardCount": 3,
                  "mongosCount": 2,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "shardCount": 1,
                  "type": "ShardedCluster",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBMultiCluster",
                "metadata": {
                  "name": "multi-replica-set"
                },
                "spec": {
                  "clusterSpecList": [
                    {
                      "clusterName": "e2e.cluster1.mongokubernetes.com",
                      "members": 2
                    },
                    {
                      "clusterName": "e2e.cluster2.mongokubernetes.com",
                      "members": 1
                    },
                    {
                      "clusterName": "e2e.cluster3.mongokubernetes.com",
                      "members": 2
                    }
                  ],
                  "credentials": "my-credentials",
                  "duplicateServiceObjects": false,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBOpsManager",
                "metadata": {
                  "name": "ops-manager"
                },
                "spec": {
                  "adminCredentials": "ops-manager-admin",
                  "applicationDatabase": {
                    "members": 3,
                    "podSpec": {
                      "cpu": 1
                    }
                  },
                  "configuration": {
                    "mms.fromEmailAddr": "admin@thecompany.com"
                  },
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "version": "6.0.3"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBUser",
                "metadata": {
                  "name": "my-replica-set-x509-user"
                },
                "spec": {
                  "db": "$external",
                  "mongodbResourceRef": {
                    "name": "my-replica-set"
                  },
                  "roles": [
                    {
                      "db": "admin",
                      "name": "dbOwner"
                    }
                  ],
                  "username": "CN=my-replica-set-x509-user,OU=cloud,O=MongoDB,L=New York,ST=New York,C=US"
                },
                "status": {
                  "phase": "Updated"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/mongodb/mongodb-enterprise-operator-ubi@sha256:6d394a5247cc346c4421bd9b37948c0378970dd3f4fc5d6eaac681f4c9655957
          createdAt: "2025-04-15T11:32:47Z"
          description: The MongoDB Enterprise Kubernetes Operator enables easy deploys
            of MongoDB into Kubernetes clusters, using our management, monitoring
            and backup platforms, Ops Manager and Cloud Manager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mongodb/mongodb-enterprise-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MongoDB Deployment
            displayName: MongoDB Deployment
            kind: MongoDB
            name: mongodb.mongodb.com
            version: v1
          - description: MongoDB Multi Deployment
            displayName: MongoDB Multi Deployment
            kind: MongoDBMultiCluster
            name: mongodbmulticluster.mongodb.com
            version: v1
          - description: MongoDB x509 User
            displayName: MongoDB User
            kind: MongoDBUser
            name: mongodbusers.mongodb.com
            version: v1
          - description: MongoDB Ops Manager
            displayName: MongoDB Ops Manager
            kind: MongoDBOpsManager
            name: opsmanagers.mongodb.com
            version: v1
        description: |
          The MongoDB Enterprise Kubernetes Operator enables easy deploys of MongoDB
          into Kubernetes clusters, using our management, monitoring and backup
          platforms, Ops Manager and Cloud Manager.

          ## Before You Start

          To start using the operator you''ll need an account in MongoDB Cloud Manager or
          a MongoDB Ops Manager deployment.

          * [Create a Secret with your OpsManager API key](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-operator-credentials/#procedure)

          * [Create a ConfigMap with your OpsManager project ID and URL](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-project-using-configmap/)

          By installing this integration, you will be able to deploy MongoDB instances
          with a single simple command.

          ## Required Parameters

          * `opsManager` or `cloudManager` - Enter the name of the ConfigMap containing project information
          * `credentials` - Enter the name of the Secret containing your OpsManager credentials
          * `type` - Enter MongoDB Deployment Types ("Standalone", "ReplicaSet", "ShardedCluster"

          ## Supported MongoDB Deployment Types ##

          * Standalone: An instance of mongod that is running as a single server and
          not as part of a replica set, this is, it does not do any kind of
          replication.

          * Replica Set: A replica set in MongoDB is a group of mongod processes that
          maintain the same data set. Replica sets provide redundancy and high
          availability, and are the basis for all production deployments. This section
          introduces replication in MongoDB as well as the components and architecture
          of replica sets. The section also provides tutorials for common tasks
          related to replica sets.

          * Sharded Cluster: The set of nodes comprising a sharded MongoDB deployment.
          A sharded cluster consists of config servers, shards, and one or more mongos
          routing processes. Sharding is a A database architecture that partitions
          data by key ranges and distributes the data among two or more database
          instances. Sharding enables horizontal scaling.

          ## Requirements for deploying MongoDB OpsManager

          * In order to deploy resources of type MongoDB OpsManager, you will need to
          create a secret containing the [credentials](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/plan-om-resource/#om-rsrc-prereqs)
          for the Global Owner user

          ## Security ##

          The operator can enable TLS for all traffic between servers and also between
          clients and servers. Before enabling `security.tls.enabled` to `true` you
          should create your certificates.  or you can leave the operator to create all
          the certificates for you. The operator ability to create certs is been
          deprecated due to Kubernetes API changes.

          For more information, please read the official MongoDB
          Kubernetes Operator  [docs](https://docs.mongodb.com/kubernetes-operator/stable/).
        displayName: MongoDB Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mongodb
        - databases
        - nosql
        links:
        - name: Documentation
          url: https://docs.opsmanager.mongodb.com/current/tutorial/install-k8s-operator/index.html
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: stable
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-enterprise-server@sha256:fb8397cbcdbf7c33f6ad142492792771127f0d9a8c46388f350b6927405ee590
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa8683500e8d3632c4588bbcdddb4e3186a01f48aefe1b6d95f8860dae8b6c76
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:7d93aee462c60cacc4e5e66729f5131c5c0fa434e6e6400c8e2ccac4bd5e20ba
        - quay.io/mongodb/mongodb-agent-ubi@sha256:85f883076bc8edc516800704e2290c54b1bee14a9a533df84928c7ce1ae807d2
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0b2d01d543d38feddcc9d351e711df33a0a5187b22d95cdb926f6015ee1093bd
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ca2aaf57d70a0b9d743a6533c4e1b88324056197453de540d3352e679a3c9c92
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7e5361c2a1604dec3fc671d629806fe6ddc6df524b69a441b86933a4b30093
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b2bd8eb7f8b96adb874468ff1b0ebae2db4ebe490b6e1da4c9a6c5108f813ff5
        - quay.io/mongodb/mongodb-agent-ubi@sha256:766e6cbeec27f995211aa509ff7f118b6945f3e728141f28f9c4ea0dd2297af5
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3add8b68bfbaa42596fe5dbbee4623ba5ff1a826b18bb71301525adf574dcb94
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0fde236fe26d5b20210d6f246aadacd7845874712def516cbe2eec11d79d5181
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:dfec259209995bb040a1eff1fcafb1d6c3fdd145129f12c04a56ec74834b80a6
        - quay.io/mongodb/mongodb-enterprise-server@sha256:110a560f5a0ebe1fde1b2c20ae5ff5638682229d1c8656b993bdaf7b71eb94bc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:19925e1297ab652db59c35f53e7b7c8b2b277f8ecdc4215682961aaf2b3e924e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa300abf739e26350a258f643fa8e57aedacd041055c7b923bd22b21ef554cdb
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:83af86f8098e3c81b6e0879eabd582592ef6cbb1d6592f5e175c0794b9df2a0e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5421e00c0935b7a9ced51a11e935773ef5df4d259f3e1b1930fbf2f1b9d46586
        - quay.io/mongodb/mongodb-agent-ubi@sha256:38175af108178d89e4d3359ee2754534e5e9b24e843f12711fc14fe1dead516a
        - quay.io/mongodb/mongodb-agent-ubi@sha256:6ebe232ed0ea87045792e578ecb11dd46411333ad0ea61383c247704b4a0efe9
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:d3248506b6bd6b01c5973ab7bdac51ee41efa0ec8dd6211be4cc0b76a53a40c1
        - quay.io/mongodb/mongodb-enterprise-init-database-ubi@sha256:ab12a330e6b40735ee516168073887a4f37b58694bd10026a451eb9972e55500
        - quay.io/mongodb/mongodb-agent-ubi@sha256:5119a1ceeae6440dbe88b89550a7ae18de8e85c9a2059f467ad66adf7ec5d75f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:11ab324698f919f84a5dab5a964d93fe63f98a48146bb3b86974cba7606d30ec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c752a7f0726d21d55f204a359b85fa29cd727a5fe571b37289ae3ea52d6206b0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8f8735d377d214f64961381b5b1ed7cd3c306e6354dbc6b45c51134a25444b93
        - quay.io/mongodb/mongodb-enterprise-server@sha256:db63f6d93e03a3e644ef6ef4846ed7d24576c688ffb584226c78e1d1fba38878
        - quay.io/mongodb/mongodb-agent-ubi@sha256:575a43be4c667f3249f46f1fdcc06dd2e68cacc9eed989da0f6edd7e214d51b7
        - quay.io/mongodb/mongodb-agent-ubi@sha256:23d962d038888c6aba5540c8f5d81d7e86c4ccfaa07b9e159825d98676609f8e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0ee80032abe873b330537d4b64511ee76c5efb512af95287f28ddeffab13bb6d
        - quay.io/mongodb/mongodb-agent-ubi@sha256:8c11348eac180f5faa113db2804817302ec78847472e1891b063744adba4aa60
        - quay.io/mongodb/mongodb-agent-ubi@sha256:6d007271faed24bdf06eb64f2b54e5cdc3e11ce71840a7f586dabe2ee02c640e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:83db493ac6158d654d3ce55619f312d53cfbac258e0faf22e9110ca004e174bd
        - quay.io/mongodb/mongodb-agent-ubi@sha256:2395fe9d38a1a9160f7cb2f400b776ea3fc2e72187d23b6dee7fd1e4fb9ca721
        - quay.io/mongodb/mongodb-agent-ubi@sha256:66b49d9dce414dcd657c3550b26f71964b82edfb67a41a3ce8d2386b04ef07e8
        - quay.io/mongodb/mongodb-agent-ubi@sha256:d927afa9d3d569d943518f0a684e19afd4adb0875f55df7c904b9dd019589e5e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:c96ad084edbdc6a6d907f57795fa38f39199b730328dbc64997d56ea50d55dfc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:67238129df09d33ee4746ff1acb5515d699a29d9f42350eae3f3f69ccb7fbf54
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:f64d99a77d6a6b9ed10de9a28b1b8b8155414eca2e561f3fdb04bd66914fb3c5
        - quay.io/mongodb/mongodb-enterprise-server@sha256:15e31285c8949a4666f7702df53416b0905d76945a28575a6028a876cc8a6f0f
        - quay.io/mongodb/mongodb-agent-ubi@sha256:31cfefd2660da3efe1b6d63e39090464c28ae006d6ed1dffb2189a09183b9c61
        - quay.io/mongodb/mongodb-agent-ubi@sha256:97bb6bb5ecd5a802e8a91e69aa10d8bde0f648b106dc5591d49ee94209822daa
        - quay.io/mongodb/mongodb-agent-ubi@sha256:f5f3b14ee6533b21d89d9ed6d4af47dbdee9e071f909dc686dfcdf6742179e46
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:985f770667e6c4c083e0c4bb79a948b83efa0cdc4ed3012b1dbf72fe48b38986
        - quay.io/mongodb/mongodb-agent-ubi@sha256:9d1f21df008b14ca721c9ed0ff905af23e1a56b64cd4eddab4df1e8e9b63df8c
        - quay.io/mongodb/mongodb-enterprise-server@sha256:523566f44cbb6bbf0cd12b1adae939a883f2254c78dc4b8c9d0098581337b1db
        - quay.io/mongodb/mongodb-agent-ubi@sha256:2dbaaece4dd48087370e43d9bcf6b0a0312f86cb5dfe916e8725323bd5401992
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9a3a6496be0d2d446e6d654eb56386c7726e0cd2d7232f6dcec579923edf0014
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1d418c65d82ccad205b7889be4ba161d3cbfa8b1fb75b21c8fcf02de38fad9ca
        - quay.io/mongodb/mongodb-enterprise-server@sha256:172e338a43df01ee5408d8d1e3cb038d2c55e77761c1a2ac6ed37959d1525140
        - quay.io/mongodb/mongodb-enterprise-server@sha256:655896bac9e82de8e9729d59a36c3a31f66235f22b93449260a2ccb7709de084
        - quay.io/mongodb/mongodb-agent-ubi@sha256:695074fbbfcadf454230b80fbf1094260e631f6108cefb2050473edaee2a217a
        - quay.io/mongodb/mongodb-agent-ubi@sha256:b175972792aa1be268d5fb3cf3ad0156f4aecffa323e1b473360fc2a3a8d8891
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d876f57ea45582e84748b02459748e440af443a7eb0882585638973409c08f6a
        - quay.io/mongodb/mongodb-agent-ubi@sha256:411f85772e7a4b2c76e21d26d320b7743612080bf0cb4c7900d9a626cfeea33b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a48a4d5c275fd669a27469649fb39ea38c728dbcf9f0a9da4e7cf27a2ce912ed
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:2be0c6c89898b18618932b6a6303aeefdde60f48ec37d5c1fca5b8d04a00fd78
        - quay.io/mongodb/mongodb-agent-ubi@sha256:bd7edb8d0931dd952524f1f0ef99f61b0f69385d9e61efd034a4f45fd315a40b
        - quay.io/mongodb/mongodb-agent-ubi@sha256:ee150e969c0bb786969c7e0a45fb1812766b752aee539072e28648817826433b
        - quay.io/mongodb/mongodb-agent-ubi@sha256:7144ac5e5038e92a6d142bf39eba46554c1b5fafeb8ea59f0390896a0d35ed17
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:2127d7f2684ae36b123422f55237055aad02fdb71c4c21a5b8f0b0f6d8baedfc
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:c9a211a2c31b97eed3a25a814198547f2aa3470df93d661ca3700b828100d8de
        - quay.io/mongodb/mongodb-agent-ubi@sha256:c03e4a7663206f54c5ef39974502816f66d8d64936c19849f0337ea4547e614e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:35b263861d6ea66b541abc190ce1aaf4d4e36f128e5d8ba2a5e9fa8e25d84446
        - quay.io/mongodb/mongodb-enterprise-server@sha256:85c1d669c073caf9cbaac815e4c8ea2ca18b71deefa03bf84cd6d8b8ebd53e12
        - quay.io/mongodb/mongodb-enterprise-server@sha256:37af0913742b6de339c7e664b30f35dcef04506529e4a09a9bbbdd57643f7149
        - quay.io/mongodb/mongodb-agent-ubi@sha256:442f5d99d02674099525abaa69daa42cec758ce5554198f978927aa5c49006c8
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f1bae6375f846af6fca7faeeff4aedf9fab41a9d45b515ba74e9fd7336408617
        - quay.io/mongodb/mongodb-enterprise-server@sha256:605dcca06a45e5ea127edb9ff262a1c3ddafdab2f30b6d3a18f73bfb8a6acdb0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ef732e3eea91cc86caf2eabedc0c9115ed235a8d204f33e74bbf090a0b0259fb
        - quay.io/mongodb/mongodb-agent-ubi@sha256:0f8d102780a9d2ce54b7453f9575d0e24b15f014cd9b74816b1e6a4090b979b2
        - quay.io/mongodb/mongodb-enterprise-init-appdb-ubi@sha256:be4295843d92ce0f325b8daaeb3a5490c11772d0348cbd9ea35a6126d038b7fb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:33a30d11969eb21952681108092fe7f916e25c8e1de43e6a4e40e126025a2bec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7aa00dca44559318356627e88440765576dd6706187044edd1f21d0d0dcb0bf6
        - quay.io/mongodb/mongodb-enterprise-server@sha256:df6081325ea2fe01d1b27d0c70005dec8232aaf51a43d91661b722b8a1761263
        - quay.io/mongodb/mongodb-enterprise-server@sha256:dfd4277f0c16ed476ac7de51894739e8c8d13a65ea5024b9a97b3dce0f6c6975
        - quay.io/mongodb/mongodb-enterprise-server@sha256:584255582db31d0bb1f9593feb3b675d934d8750766fc98d3a1b3340feae4995
        - quay.io/mongodb/mongodb-enterprise-database-ubi@sha256:e2ca129f04ad6bf4e0dbb988badb083187725f28bac01bb7080cc80cad9ac316
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7885ab6e88214f9289d8babf5189136dde0dd4c84a15a5481ccd3ef512b2e66f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f2d9a3264a8814f33143b73c56624572fb80afeab184133d3a901713a31c1a87
        - quay.io/mongodb/mongodb-agent-ubi@sha256:03763de5018fd5fa14843444d24adfc5e387605970b531801ef2d0916233a010
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d21c65780cadc9d8668d2c28935d492625d6993f13ba562722200a819281cbcc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b00ac773623e1bf422b47e6d783263d2d08737865abe9abbe106d86d1ea25165
        - quay.io/mongodb/mongodb-enterprise-operator-ubi@sha256:6d394a5247cc346c4421bd9b37948c0378970dd3f4fc5d6eaac681f4c9655957
        - quay.io/mongodb/mongodb-agent-ubi@sha256:b23cfa639c5a4d20381245772dfe396ec915af3a3046df5d9b10016b900431f5
        - quay.io/mongodb/mongodb-enterprise-server@sha256:cd0cdf1b2a58f4c3b514e602cc1d96543410f631825045bbf65bd3fb84113f1f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8445d59d491a6d3890ea9c8b421214bcd0fdcfbc420ed40497a790b79f0bc89e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c54eeb910855594ed4bf5dcbbbcc14aa6f733e97313896a8381fb4be6aeac583
        - quay.io/mongodb/mongodb-agent-ubi@sha256:36c84115c6c645fd7d65d1004d2507ef3592aa1f07176645b1ee8e8b2a4f5f73
        - quay.io/mongodb/mongodb-enterprise-init-ops-manager-ubi@sha256:6ccc8148c266a9037ea744e8d19875146b1c231909f29a8f01c7ce415ca15936
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d49417ea85ee1ddb84429552a5aac87a82c2843904d8c0300a43b657b7a49e0e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:bd868284d9dfcb7c5a20e1160218d0172358384229b556c37e0789af0030eb24
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c07f20f8e7a5479e43cac7c74fc9f18cb4d4a1bd8b6b46f7dcc6f7ec7525fe28
        - quay.io/mongodb/mongodb-enterprise-server@sha256:13186ad45a2ba3ab55ed5c2a008eca494ee3937c8519475379476cc8e6f1d055
        - quay.io/mongodb/mongodb-agent-ubi@sha256:3b57a03dc17a301fda86c8afed1917c10d9b98947680116399c975ece944f5de
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d323c1bcd14118cc537ac282996ea5a58756d20b6b73f1eec0efa8710d246077
        - quay.io/mongodb/mongodb-agent-ubi@sha256:8467b1c2924a19929fc101ef2ff8d7e53cb0115a0cdc2e371dc0733bd0262395
        version: 1.33.0
      entries:
      - name: mongodb-enterprise.v1.33.0
        version: 1.33.0
      - name: mongodb-enterprise.v1.32.0
        version: 1.32.0
      - name: mongodb-enterprise.v1.31.0
        version: 1.31.0
      - name: mongodb-enterprise.v1.30.0
        version: 1.30.0
      - name: mongodb-enterprise.v1.29.0
        version: 1.29.0
      - name: mongodb-enterprise.v1.28.0
        version: 1.28.0
      - name: mongodb-enterprise.v1.27.0
        version: 1.27.0
      - name: mongodb-enterprise.v1.26.0
        version: 1.26.0
      - name: mongodb-enterprise.v1.25.0
        version: 1.25.0
      - name: mongodb-enterprise.v1.24.0
        version: 1.24.0
      - name: mongodb-enterprise.v1.23.0
        version: 1.23.0
      - name: mongodb-enterprise.v1.22.0
        version: 1.22.0
      - name: mongodb-enterprise.v1.21.0
        version: 1.21.0
      - name: mongodb-enterprise.v1.20.1
        version: 1.20.1
      - name: mongodb-enterprise.v1.19.1
        version: 1.19.1
      - name: mongodb-enterprise.v1.18.0
        version: 1.18.0
      - name: mongodb-enterprise.v1.17.2
        version: 1.17.2
      - name: mongodb-enterprise.v1.17.1
        version: 1.17.1
      - name: mongodb-enterprise.v1.17.0
        version: 1.17.0
      - name: mongodb-enterprise.v1.16.4
        version: 1.16.4
      - name: mongodb-enterprise.v1.16.3
        version: 1.16.3
      - name: mongodb-enterprise.v1.16.2
        version: 1.16.2
      - name: mongodb-enterprise.v1.16.1
        version: 1.16.1
      name: stable
    defaultChannel: stable
    packageName: mongodb-enterprise
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: MariaDB Corporation
      provider-url: ""
    name: mariadb-enterprise-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-enterprise-operator.v25.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup"
                },
                "spec": {
                  "args": [
                    "--single-transaction",
                    "--all-databases"
                  ],
                  "logLevel": "info",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "0 0 * * *",
                    "suspend": false
                  },
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "backups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Connection",
                "metadata": {
                  "name": "connection"
                },
                "spec": {
                  "database": "mariadb",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "secretName": "connection",
                  "secretTemplate": {
                    "key": "dsn"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Database",
                "metadata": {
                  "name": "database"
                },
                "spec": {
                  "characterSet": "utf8",
                  "collate": "utf8_general_ci",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "name": "my-logical-database"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "grant"
                },
                "spec": {
                  "database": "*",
                  "grantOption": true,
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "privileges": [
                    "SELECT",
                    "INSERT",
                    "UPDATE"
                  ],
                  "table": "*",
                  "username": "my-user"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "MariaDB",
                "metadata": {
                  "name": "mariadb"
                },
                "spec": {
                  "connection": {
                    "secretName": "mariadb-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "database": "mariadb",
                  "galera": {
                    "enabled": true
                  },
                  "metrics": {
                    "enabled": true,
                    "passwordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "mariadb-metrics"
                    }
                  },
                  "myCnf": "[mariadb]\nbind-address=*\ndefault_storage_engine=InnoDB\nbinlog_format=row\ninnodb_autoinc_lock_mode=2\ninnodb_buffer_pool_size=1024M\nmax_allowed_packet=256M\n",
                  "passwordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-password"
                  },
                  "primaryConnection": {
                    "secretName": "mariadb-conn-primary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "primaryService": {
                    "type": "ClusterIP"
                  },
                  "replicas": 3,
                  "rootPasswordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-root"
                  },
                  "secondaryConnection": {
                    "secretName": "mariadb-conn-secondary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "secondaryService": {
                    "type": "ClusterIP"
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "storage": {
                    "size": "1Gi"
                  },
                  "updateStrategy": {
                    "type": "ReplicasFirstPrimaryLast"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "MaxScale",
                "metadata": {
                  "name": "maxscale"
                },
                "spec": {
                  "admin": {
                    "guiEnabled": true,
                    "port": 8989
                  },
                  "auth": {
                    "adminPasswordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "maxscale-admin"
                    },
                    "adminUsername": "mariadb-enterprise-operator",
                    "generate": true
                  },
                  "config": {
                    "sync": {
                      "database": "mysql",
                      "interval": "5s",
                      "timeout": "10s"
                    }
                  },
                  "connection": {
                    "port": 3306,
                    "secretName": "maxscale-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "guiKubernetesService": {
                    "type": "ClusterIP"
                  },
                  "kubernetesService": {
                    "type": "ClusterIP"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "metrics": {
                    "enabled": true
                  },
                  "monitor": {
                    "cooperativeMonitoring": "majority_of_all",
                    "interval": "2s",
                    "params": {
                      "available_when_donor": "false",
                      "disable_master_failback": "false",
                      "disable_master_role_setting": "false"
                    }
                  },
                  "replicas": 2,
                  "requeueInterval": "10s",
                  "services": [
                    {
                      "listener": {
                        "port": 3306,
                        "protocol": "MariaDBProtocol"
                      },
                      "name": "rw-router",
                      "params": {
                        "transaction_replay": "true",
                        "transaction_replay_attempts": "10",
                        "transaction_replay_timeout": "5s"
                      },
                      "router": "readwritesplit"
                    }
                  ]
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "PhysicalBackup",
                "metadata": {
                  "name": "physicalbackup"
                },
                "spec": {
                  "compression": "bzip2",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "podAffinity": true,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "0 0 * * *",
                    "immediate": true,
                    "suspend": false
                  },
                  "serviceAccountName": "backup",
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "physicalbackups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "region": "us-east-1",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  },
                  "timeout": "1h"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore"
                },
                "spec": {
                  "backupRef": {
                    "name": "backup"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "targetRecoveryTime": "2023-12-19T09:00:00Z"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "SqlJob",
                "metadata": {
                  "name": "sqljob"
                },
                "spec": {
                  "database": "mariadb",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "sql": "CREATE TABLE IF NOT EXISTS orders (\n  id bigint PRIMARY KEY AUTO_INCREMENT,\n  customer_id INT,\n  order_date DATE,\n  total_amount_cents INT\n);\nINSERT INTO orders (customer_id, order_date, total_amount_cents) VALUES\n  (101, '2024-05-15', 5000),\n  (102, '2024-05-14', 7525),\n  (103, '2024-05-13', 10050),\n  (104, '2024-05-12', 2575);",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "user"
                },
                "spec": {
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxUserConnections": 20,
                  "name": "my-user",
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.mariadb.com/mariadb-enterprise-operator@sha256:1effbbfa7632be92c17ef61c18a25f010b187840beec42c31d2410d0209e6985
          createdAt: "2025-08-28T11:35:03Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["MariaDB Enterprise"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: MariaDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API. It is used to define
              backup jobs and its storage.
            displayName: Backup
            kind: Backup
            name: backups.enterprise.mariadb.com
            version: v1alpha1
          - description: Connection is the Schema for the connections API. It is used
              to configure connection strings for the applications connecting to MariaDB.
            displayName: Connection
            kind: Connection
            name: connections.enterprise.mariadb.com
            version: v1alpha1
          - description: Database is the Schema for the databases API. It is used
              to define a logical database as if you were running a 'CREATE DATABASE'
              statement.
            displayName: Database
            kind: Database
            name: databases.enterprise.mariadb.com
            version: v1alpha1
          - description: Grant is the Schema for the grants API. It is used to define
              grants as if you were running a 'GRANT' statement.
            displayName: Grant
            kind: Grant
            name: grants.enterprise.mariadb.com
            version: v1alpha1
          - description: MariaDB is the Schema for the mariadbs API. It is used to
              define MariaDB clusters.
            displayName: Maria DB
            kind: MariaDB
            name: mariadbs.enterprise.mariadb.com
            version: v1alpha1
          - description: MaxScale is the Schema for the maxscales API. It is used
              to define MaxScale clusters.
            displayName: Max Scale
            kind: MaxScale
            name: maxscales.enterprise.mariadb.com
            version: v1alpha1
          - description: PhysicalBackup is the Schema for the physicalbackups API.
              It is used to define physical backup jobs and its storage.
            displayName: Physical Backup
            kind: PhysicalBackup
            name: physicalbackups.enterprise.mariadb.com
            version: v1alpha1
          - description: Restore is the Schema for the restores API. It is used to
              define restore jobs and its restoration source.
            displayName: Restore
            kind: Restore
            name: restores.enterprise.mariadb.com
            version: v1alpha1
          - description: SqlJob is the Schema for the sqljobs API. It is used to run
              sql scripts as jobs.
            displayName: Sql Job
            kind: SqlJob
            name: sqljobs.enterprise.mariadb.com
            version: v1alpha1
          - description: User is the Schema for the users API.  It is used to define
              grants as if you were running a 'CREATE USER' statement.
            displayName: User
            kind: User
            name: users.enterprise.mariadb.com
            version: v1alpha1
        description: |
          MariaDB Enterprise Operator for Openshift is an operator that manages MariaDB Enterprise operands declaratively via Custom Resources, with support for the following features:

          * Provision and Configure MariaDB and MaxScale Declaratively: Define MariaDB Enterprise Server and MaxScale clusters in YAML manifests and deploy them with ease in Kubernetes.
          * High Availability with Galera: Ensure availability with MariaDB Enterprise Cluster, providing synchronous multi-master replication.
          * Query and Connection-Based Routing with MaxScale: MaxScale provides query routing and connection load balancing for improved application performance.
          * Cluster-Aware Rolling Updates: Perform rolling updates on MariaDB and MaxScale clusters, ensuring zero-downtime upgrades with no disruptions to your applications.
          * Flexible Storage Configuration and Volume Expansion: Easily configure storage for MariaDB instances, including the ability to expand volumes as needed.
          * Physical Backups based on [mariadb-backup](https://mariadb.com/docs/server/server-usage/backup-and-restore/mariadb-backup/full-backup-and-restore-with-mariadb-backup) and [Kubernetes VolumeSnapshots](https://kubernetes.io/docs/concepts/storage/volume-snapshots/). By leveraging the [BACKUP STAGE](https://mariadb.com/docs/server/reference/sql-statements/administrative-sql-statements/backup-commands/backup-stage) feature, backups are taken without long read locks or service interruptions.
          * Logical Backups based on [mariadb-dump](https://mariadb.com/docs/server/clients-and-utilities/backup-restore-and-import-clients/mariadb-dump).
          * Backup Management: Take, restore, and schedule backups with multiple storage types supported: S3, PVCs, Kubernetes volumes and VolumeSnapshots..
          * Policy-Driven Backup Retention: Implement backup retention policies with bzip2 and gzip compression.
          * Target Recovery Time: Restore your database to the closest available backup based on a specified recovery time.
          * Bootstrap New Instances: Initialize new MariaDB instances from backups, S3, PVCs or VolumeSnapshots to quickly spin up new clusters.
          * TLS Certificate Management: Issue, configure, and rotate TLS certificates and Certificate Authorities (CAs) for secure connections.
          * Advanced TLS Support: customize certificate lifetime, private key algorithm and TLS version.
          * Native Integration with cert-manager: Leverage [cert-manager](https://cert-manager.io/docs/), the de-facto standard for managing certificates in Kubernetes, to enable issuance with private CAs, public CAs and HashiCorp Vault.
          * Prometheus Metrics: Expose metrics using the MariaDB and MaxScale Prometheus exporters.
          * Native Integration with prometheus-operator: Leverage [prometheus-operator](https://github.com/prometheus-operator/prometheus-operator) to scrape metrics from MariaDB and MaxScale instances.
          * Declarative User and Database Management: Manage users, grants, and logical databases in a declarative manner using Kubernetes resources.
          * Secure, immutable and lightweight images based on Red Hat UBI, available for multiple architectires (amd64, arm64 and ppc64le).
          * [Operator certified ](https://catalog.redhat.com/en/software/container-stacks/detail/65789bcbe17f1b31944acb1d#overview) by Red Hat.

          ## Installation

          In order to use this product, you will need a [MariaDB Enterprise subscription](https://mariadb.com/products/enterprise/).

          To be able to pull MariaDB Enterprise images, you will need to configure the MariaDB registry credentials in Openshift using the [MariaDB customer download token](https://customers.mariadb.com/downloads/token/) as described below:

          - Extract your [Openshift global pull secret](https://docs.redhat.com/en/documentation/openshift_container_platform/4.18/html/images/managing-images#images-update-global-pull-secret_using-image-pull-secrets):
            ```bash
            oc extract secret/pull-secret -n openshift-config --confirm
            ```

          - Login in the MariaDB registry providing the customer download token as password:
            ```bash
            oc registry login \
              --registry="docker.mariadb.com" \
              --auth-basic="<email>:<customer-download-token>" \
              --to=.dockerconfigjson
            ```

          - Update the global pull secret:
            ```bash
            oc set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson
            ```

          Alternatively, you can also create a `Secret` named `mariadb-enterprise` in the namespace where are you installing the operator by using the following command:
            ```bash
            oc create secret docker-registry mariadb-enterprise \
              --docker-server=docker.mariadb.com \
              --docker-username=<email> \
              --docker-password=<customer-download-token> \
              -n <operator-namespace>
            ```

          At this point, you can proceed to install the MariaDB Enterprise Operator, either via the Openshift console or by creating a `Subscription` object as described in the [Openshift installation docs](https://mariadb.com/kb/en/openshift/).

          Refer to the [quickstart guide](https://mariadb.com/kb/en/quickstart/), the [examples catalog](https://mariadb.com/kb/en/examples-catalog/) and the [API reference](https://mariadb.com/kb/en/api-reference/) for getting started with the Custom Resources offered by this operator.
        displayName: MariaDB Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mariadb
        - mysql
        - database
        - operator
        - mariadb-enterprise-operator
        - database
        links:
        - name: MariaDB Enterprise
          url: https://mariadb.com/products/enterprise/
        - name: Documentation
          url: https://mariadb.com/kb/en/mariadb-enterprise-operator/
        maintainers:
        - email: operator@mariadb.com
          name: MariaDB Corporation
        - email: support@mariadb.com
          name: Support
        - email: sales-EMEA@mariadb.com
          name: Sales EMEA
        - email: sales-APAC@mariadb.com
          name: Sales APAC
        - email: sales-nam@mariadb.com
          name: Sales Americas
        maturity: beta
        minKubeVersion: 1.29.0
        provider:
          name: MariaDB Corporation
        relatedImages:
        - docker.mariadb.com/enterprise-server@sha256:9872721d35ebe8d820b1aa9c6dd83b363a7b2df9af11a65bff6b80ff4144460a
        - docker.mariadb.com/maxscale@sha256:d9e04a7a406353dab2051e190be0c32ed20b780ed71e56e55cedd4ccb61af416
        - mariadb/mariadb-prometheus-exporter-ubi@sha256:f072175198cad050e80617d13e63f415bacdfe47e2dd177fde7283a96ee56c07
        - mariadb/maxscale-prometheus-exporter-ubi@sha256:cc17c1554795ad01b4d8db799f6aa90e00e41ede431f09d30eec64d9dd4479d9
        - docker.mariadb.com/mariadb-enterprise-operator@sha256:1effbbfa7632be92c17ef61c18a25f010b187840beec42c31d2410d0209e6985
        version: 25.8.1-1
      entries:
      - name: mariadb-enterprise-operator.v25.8.1-1
        version: 25.8.1-1
      - name: mariadb-enterprise-operator.v25.8.0
        version: 25.8.0
      - name: mariadb-enterprise-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: mariadb-enterprise-operator
    provider:
      name: MariaDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Instana
      provider-url: ""
    name: instana-agent-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: instana-agent-operator.v2.1.36
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instana.io/v1",
                "kind": "InstanaAgent",
                "metadata": {
                  "name": "instana-agent",
                  "namespace": "instana-agent"
                },
                "spec": {
                  "agent": {
                    "configuration_yaml": "# You can leave this empty, or use this to configure your instana agent.\n# See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.\n",
                    "endpointHost": "ingress-red-saas.instana.io",
                    "endpointPort": "443",
                    "env": {},
                    "key": "replace-key"
                  },
                  "cluster": {
                    "name": "my-cluster"
                  },
                  "opentelemetry": {
                    "grpc": {
                      "enabled": true
                    },
                    "http": {
                      "enabled": true
                    }
                  },
                  "zone": {
                    "name": "edited-zone"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring,Logging & Tracing,OpenShift Optional
          certified: "false"
          containerImage: icr.io/instana/instana-agent-operator@sha256:ab8b35f0b45c68f69210d1b17c5dd609799f6bb0bc9f12d0d4e5302c910607a1
          description: Fully automated Application Performance Monitoring (APM) for
            microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/instana-agent-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/instana-agent-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/instana/instana-agent-operator
          support: Instana
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InstanaAgent is the Schema for the agents API
            displayName: Instana Agent
            kind: InstanaAgent
            name: agents.instana.io
            version: v1
          - description: InstanaAgentRemote is the Schema for the agents API
            displayName: Instana Agent Remote
            kind: InstanaAgentRemote
            name: agentsremote.instana.io
            version: v1
        description: |
          # Instana

          Instana is an [APM solution](https://www.instana.com/product-overview/) built for microservices that enables IT Ops to build applications faster and deliver higher quality services by automating monitoring, tracing and root cause analysis. The solution is optimized for [Kubernetes](https://www.instana.com/automatic-kubernetes-monitoring/) and [OpenShift](https://www.instana.com/blog/automatic-root-cause-analysis-for-openshift-applications/).

          ## Instana Agent Operator

          This is the Kubernetes Operator for installing the Instana Agent on Kubernetes or OpenShift.

          ## Prerequisites for OpenShift

          Before the agent will be able to run in OpenShift, you need to perform a couple of extra configuration steps.

          You need to set up a project for the Instana Agent and configure it's permissions.

          The project you create here needs to be the namespace where you create the Instana Agent custom resource that the operator will use to deploy the agent.

          For example, create the `instana-agent` project:

              oc new-project instana-agent

          Then, ensure the `instana-agent` service account is in the privileged security context:

              oc adm policy add-scc-to-user privileged -z instana-agent

          The `instana-agent-remote` service account must be allowed to use any userid:

              oc adm policy add-scc-to-user anyuid -z instana-agent-remote -n instana-agent

          This service account will be created by the operator.

          Now you can proceed with installing the operator for the Instana agent.

          ## Installation and Configuration

          First, install this operator from [OperatorHub.io](https://operatorhub.io/), [OpenShift Container Platform](https://www.openshift.com/), or [OKD](https://www.okd.io/).

          Second, create a custom resource with the agent configuration in the target namespace (for now, this MUST always be the `instana-agent` namespace). The operator will pick up the custom resource and install the Instana agent accordingly.

          The following is a minimal template of the custom resource:

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgent
          metadata:
            name: instana-agent
            namespace: instana-agent
          spec:
            zone:
              name: my-zone # (optional) name of the zone of the host
            cluster:
              name: replace-me # replace with the name of your Kubernetes cluster
            agent:
              key: replace-me # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io # the monitoring ingress endpoint
              endpointPort: "443" # the monitoring ingress endpoint port, wrapped in quotes
              env:
                INSTANA_AGENT_TAGS: example
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://ibm.biz/monitoring-k8s
          ```

          Save the template in a file `instana-agent.yaml` and edit the following values:

          * If your target namespace is not `instana-agent`, replace the `namespace:` accordingly.
          * `agent.key` must be set with your Instana agent key.
          * `agent.endpointHost` must be set with the monitoring ingress endpoint, generally either `saas-us-west-2.instana.io` or `saas-eu-west-1.instana.io`.
          * `agent.endpointPort` must be set with the monitoring ingress port, generally "443" (wrapped in quotes).
          * `zone.name` should be set with the name of the Kubernetes cluster that is be displayed in Instana.

          For advanced configuration, you can edit the contents of the `configuration.yaml` file. View documentation [here](https://ibm.biz/monitoring-k8s).

          Apply the custom resource with `kubectl apply -f instana-agent.yaml`. After some time, you should see `instana-agent` Pods being created on each node of your cluster, and your cluster should show on the infrastructure map on your Instana Web interface.

          In addition to the InstanaAgent custom resource, a remote agent can be deployed. This agent will not monitor workload running inside the cluster, but can deploy a single replica of an agent which can get an independent configuration for running remote monitoring for certain sensors.

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgentRemote
          metadata:
            name: remote-agent
            namespace: "instana-agent"
          spec:
            zone:
              name: edited-zone # name of the zone of the host
            hostname:
              name: use-custom-hostname-for-remote-agent # (optional) give your remote agent pod a custom hostname
            agent:
              key: replace-key # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io
              endpointPort: "443"
              env: {}
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.
          ```

          ## Uninstalling

          In order to uninstall the Instana agent, simply remove the custom resource with `kubectl delete -f instana-agent.yaml`.

          ## Source Code

          The Instana agent operator is an open source project hosted on [https://github.com/instana/instana-agent-operator](https://github.com/instana/instana-agent-operator/).
        displayName: Instana Agent Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - APM
        - Instana
        links:
        - name: GitHub Repository
          url: https://github.com/instana/instana-agent-operator
        maintainers:
        - email: support@instana.com
          name: Instana
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: Instana
        relatedImages:
        - icr.io/instana/instana-agent-operator@sha256:ab8b35f0b45c68f69210d1b17c5dd609799f6bb0bc9f12d0d4e5302c910607a1
        - icr.io/instana/agent@sha256:ed71e6c3e5bdb155853a2db8c916460e6a867fb02627e3a332186188cbbd98dd
        version: 2.1.36
      entries:
      - name: instana-agent-operator.v2.1.36
        version: 2.1.36
      - name: instana-agent-operator.v2.1.35
        version: 2.1.35
      - name: instana-agent-operator.v2.1.34
        version: 2.1.34
      - name: instana-agent-operator.v2.1.33
        version: 2.1.33
      - name: instana-agent-operator.v2.1.32
        version: 2.1.32
      - name: instana-agent-operator.v2.1.31
        version: 2.1.31
      - name: instana-agent-operator.v2.1.30
        version: 2.1.30
      - name: instana-agent-operator.v2.1.28
        version: 2.1.28
      - name: instana-agent-operator.v2.1.27
        version: 2.1.27
      - name: instana-agent-operator.v2.1.24
        version: 2.1.24
      - name: instana-agent-operator.v2.1.23
        version: 2.1.23
      - name: instana-agent-operator.v2.1.22
        version: 2.1.22
      - name: instana-agent-operator.v2.1.21
        version: 2.1.21
      - name: instana-agent-operator.v2.1.20
        version: 2.1.20
      - name: instana-agent-operator.v2.1.19
        version: 2.1.19
      - name: instana-agent-operator.v2.1.18
        version: 2.1.18
      - name: instana-agent-operator.v2.1.17
        version: 2.1.17
      - name: instana-agent-operator.v2.1.16
        version: 2.1.16
      - name: instana-agent-operator.v2.1.15
        version: 2.1.15
      - name: instana-agent-operator.v2.1.14
        version: 2.1.14
      - name: instana-agent-operator.v2.1.13
        version: 2.1.13
      - name: instana-agent-operator.v2.1.12
        version: 2.1.12
      - name: instana-agent-operator.v2.1.11
        version: 2.1.11
      - name: instana-agent-operator.v2.1.10
        version: 2.1.10
      - name: instana-agent-operator.v2.1.9
        version: 2.1.9
      - name: instana-agent-operator.v2.1.8
        version: 2.1.8
      - name: instana-agent-operator.v2.1.7
        version: 2.1.7
      - name: instana-agent-operator.v2.1.6
        version: 2.1.6
      - name: instana-agent-operator.v2.1.5
        version: 2.1.5
      - name: instana-agent-operator.v2.1.4
        version: 2.1.4
      - name: instana-agent-operator.v2.1.3
        version: 2.1.3
      - name: instana-agent-operator.v2.1.2
        version: 2.1.2
      - name: instana-agent-operator.v2.1.1
        version: 2.1.1
      - name: instana-agent-operator.v2.1.0
        version: 2.1.0
      - name: instana-agent-operator.v2.0.32
        version: 2.0.32
      - name: instana-agent-operator.v2.0.31
        version: 2.0.31
      - name: instana-agent-operator-rhmp.v2.0.30
        version: 2.0.30
      - name: instana-agent-operator.v2.0.27
        version: 2.0.27
      - name: instana-agent-operator.v2.0.19
        version: 2.0.19
      - name: instana-agent-operator.v2.0.18
        version: 2.0.18
      - name: instana-agent-operator.v2.0.17
        version: 2.0.17
      - name: instana-agent-operator.v2.0.13
        version: 2.0.13
      - name: instana-agent-operator.v2.0.12
        version: 2.0.12
      - name: instana-agent-operator.v2.0.11
        version: 2.0.11
      - name: instana-agent-operator.v2.0.10
        version: 2.0.10
      - name: instana-agent-operator.v2.0.9
        version: 2.0.9
      - name: instana-agent-operator.v2.0.8
        version: 2.0.8
      - name: instana-agent-operator.v2.0.7
        version: 2.0.7
      - name: instana-agent-operator.v2.0.5
        version: 2.0.5
      - name: instana-agent-operator.v2.0.4
        version: 2.0.4
      name: stable
    defaultChannel: stable
    packageName: instana-agent-operator-rhmp
    provider:
      name: Instana
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: pelorus-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pelorus-operator.v0.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.pelorus.dora-metrics.io/v1alpha1",
                "kind": "Pelorus",
                "metadata": {
                  "name": "pelorus-sample"
                },
                "spec": {
                  "exporters": {
                    "global": {},
                    "instances": [
                      {
                        "app_name": "deploytime-exporter",
                        "exporter_type": "deploytime"
                      },
                      {
                        "app_name": "failuretime-exporter",
                        "enabled": false,
                        "exporter_type": "failure"
                      },
                      {
                        "app_name": "committime-exporter",
                        "exporter_type": "committime"
                      },
                      {
                        "app_name": "webhook-exporter",
                        "exporter_type": "webhook"
                      }
                    ]
                  },
                  "federate_openshift_monitoring": {
                    "enabled": false
                  },
                  "openshift_prometheus_basic_auth_pass": "changeme",
                  "openshift_prometheus_htpasswd_auth": "internal:{SHA}+pvrmeQCmtWmYVOZ57uuITVghrM=",
                  "prometheus_retention": "1y",
                  "prometheus_retention_size": "1GB",
                  "prometheus_storage": false,
                  "prometheus_storage_pvc_capacity": "2Gi",
                  "prometheus_storage_pvc_storageclass": "gp2"
                }
              }
            ]
          capabilities: Basic Install
          categories: |
            Modernization & Migration,Developer Tools,Monitoring,Integration & Delivery
          containerImage: quay.io/pelorus/pelorus-operator:0.0.9
          createdAt: "2024-05-27T20:04:01Z"
          description: |
            Tool that helps IT organizations measure their impact on the overall performance of their organization
          operatorframework.io/suggested-namespace: pelorus
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/dora-metrics/pelorus/
          support: Pelorus Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pelorus is the Schema for pelorus API to adopt instance to
              the requested workflow
            displayName: Pelorus
            kind: Pelorus
            name: peloruses.charts.pelorus.dora-metrics.io
            version: v1alpha1
        description: |
          Pelorus is a tool that helps IT organizations measure their impact on the overall performance of their organization. It does this by gathering metrics about team and organizational behaviors over time in some key areas of IT that have been shown to impact the value they deliver to the organization as a whole. Some of the key outcomes Pelorus can focus on are:

          - Software Delivery Performance
          - Product Quality and Sustainability
          - Customer experience

          For more background on the project you can read [@trevorquinn](https://github.com/trevorquinn)'s blog post on [Metrics Driven Transformation](https://www.openshift.com/blog/exploring-a-metrics-driven-approach-to-transformation).

          ## Software Delivery Performance as an outcome

          Currently, Pelorus functionality can capture proven metrics that measure Software Delivery Performance -- a significant outcome that IT organizations aim to deliver.

          Pelorus is a Grafana dashboard that can easily be deployed to an OpenShift cluster, and provides an organizational-level view of the [four critical measures of software delivery performance](https://blog.openshift.com/exploring-a-metrics-driven-approach-to-transformation/).

          ## Software Delivery Metrics Dashboard

          A short video describing each of metrics provided by Pelorus is available [here](https://www.youtube.com/watch?v=7-iB_KhUaQg).

          ## Demo

          [YouTube Video](https://www.youtube.com/watch?v=VPCOIfDcgso) with the Pelorus in action recorded during online Konveyor Community event.

          ## Documentation

          Pelorus documentation is available at [pelorus.readthedocs.io](https://pelorus.readthedocs.io/en/latest/).

          ## Contributing to Pelorus

          If you are interested in contributing to the Pelorus project, please review our Contribution guide which can be found in the [contribution guide](https://github.com/dora-metrics/pelorus/blob/master/CONTRIBUTING.md)

          ## Statement of Support

          Our support policy can be found in the [Upstream Support statement](https://github.com/dora-metrics/pelorus/blob/master/docs/UpstreamSupport.md)

          ## Code of Conduct
          Refer to dora-metrics's Code of Conduct [here](https://github.com/dora-metrics/pelorus/blob/master/CODE_OF_CONDUCT.md).

          ## License

          This repository is licensed under the terms of [Apache-2.0 License](https://raw.githubusercontent.com/dora-metrics/pelorus/master/LICENSE).
        displayName: Pelorus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pelorus-operator
        - dora
        - dora metrics
        - pelorus
        - metrics
        - transformation
        - devops
        links:
        - name: Pelorus Docs
          url: https://pelorus.readthedocs.io/en/latest/
        - name: Pelorus Getting Started
          url: https://pelorus.readthedocs.io/en/latest/GettingStarted/Overview/
        - name: Pelorus GIT repository
          url: https://github.com/dora-metrics/pelorus/
        maintainers:
        - email: kgranger@redhat.com
          name: Kevin Granger
        - email: whayutin@redhat.com
          name: Wesley Hayutin
        - email: mpryc@redhat.com
          name: Michal Pryc
        - email: msouzaol@redhat.com
          name: Mateus Souza Oliveira
        - email: esauer@redhat.com
          name: Eric Sauer
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - quay.io/pelorus/pelorus-operator:0.0.9
        version: 0.0.9
      entries:
      - name: pelorus-operator.v0.0.9
        version: 0.0.9
      - name: pelorus-operator.v0.0.8
        version: 0.0.8
      - name: pelorus-operator.v0.0.7
        version: 0.0.7
      - name: pelorus-operator.v0.0.6
        version: 0.0.6
      - name: pelorus-operator.v0.0.5
        version: 0.0.5
      - name: pelorus-operator.v0.0.4
        version: 0.0.4
      - name: pelorus-operator.v0.0.3
        version: 0.0.3
      - name: pelorus-operator.v0.0.2
        version: 0.0.2
      - name: pelorus-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: pelorus-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: horizon-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: horizon-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "horizon.openstack.org/v1beta1",
                "kind": "Horizon",
                "metadata": {
                  "name": "horizon"
                },
                "spec": {
                  "customServiceConfig": "DEBUG = True\nLOGGING['handlers']['console']['level'] = 'DEBUG'\n",
                  "replicas": 1,
                  "secret": "osp-secret"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:00Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Horizon is the Schema for the horizons API
            displayName: Horizon
            kind: Horizon
            name: horizons.horizon.openstack.org
            version: v1beta1
        description: Horizon Operator
        displayName: Horizon Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - UI
        - Horizon
        links:
        - name: Horizon Operator
          url: https://github.com/openstack-k8s-operators/horizon-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-horizon-rhel9@sha256:b02c115d40aacc30e6763e295da90d69681b24fba20ebbc53133be852393b399
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/horizon-rhel9-operator@sha256:b76038b96a064c275721c1d45b95d7b0dfdb9fc626a2b3250d0adb1f97f89319
        version: 1.0.6
      entries:
      - name: horizon-operator.v1.0.6
        version: 1.0.6
      - name: horizon-operator.v1.0.5
        version: 1.0.5
      - name: horizon-operator.v1.0.4
        version: 1.0.4
      - name: horizon-operator.v1.0.3
        version: 1.0.3
      - name: horizon-operator.v1.0.2
        version: 1.0.2
      - name: horizon-operator.v1.0.1
        version: 1.0.1
      - name: horizon-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: horizon-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq-streams
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amqstreams.v2.2.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.3",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.3",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:f908d848b5e67de4b162597b190ef18d449ffd4b6540457b24b491445d5c5ea3
          createdAt: "2022-07-01 10:33:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.1.0-0 <2.2.2-2'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ", "Red Hat Application Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from AMQ Streams 1.7, migration to `v1beta2` needs to be completed for all AMQ Streams CRDs and CRs before the upgrade to 2.2 is done! If upgrading from AMQ Streams version earlier than 1.7, you need to first install the CRDs from AMQ Streams 1.7 and complete the migration to `v1beta2` for all AMQ Streams CRDs and CRs before the upgrade to 2.2 is done! For more details about the CRD upgrades, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.2/html/deploying_and_upgrading_amq_streams_on_openshift/assembly-upgrade-str#assembly-upgrade-resources-str).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.2/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: Red Hat Integration - AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        - registry.redhat.io/amq7/amq-streams-kafka-31-rhel8@sha256:d0607285fae95365b9d34b9474466824fb5ad012871204e8b13e52e438b66bf5
        - registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:f908d848b5e67de4b162597b190ef18d449ffd4b6540457b24b491445d5c5ea3
        - registry.redhat.io/amq7/amq-streams-kafka-32-rhel8@sha256:e3bbfff505f12c0360e175f4640c2dd9a7adcf3dce1ae2f586f2bde805ed8b1f
        - registry.redhat.io/amq7/amq-streams-bridge-rhel8@sha256:5d49bd23da40590e74422424b9457398bad89da8276718ff6a8a088e901c8345
        version: 2.2.2-2
      entries:
      - name: amqstreams.v2.2.2-2
        version: 2.2.2-2
      - name: amqstreams.v2.2.2-1
        version: 2.2.2-1
      - name: amqstreams.v2.2.2-0
        version: 2.2.2-0
      - name: amqstreams.v2.2.1-7
        version: 2.2.1-7
      - name: amqstreams.v2.2.1-6
        version: 2.2.1-6
      - name: amqstreams.v2.2.1-5
        version: 2.2.1-5
      - name: amqstreams.v2.2.1-4
        version: 2.2.1-4
      - name: amqstreams.v2.2.1-3
        version: 2.2.1-3
      - name: amqstreams.v2.2.1-2
        version: 2.2.1-2
      - name: amqstreams.v2.2.1-1
        version: 2.2.1-1
      - name: amqstreams.v2.2.1-0
        version: 2.2.1-0
      - name: amqstreams.v2.2.0-5
        version: 2.2.0-5
      name: amq-streams-2.2.x
    - currentCSV: amqstreams.v2.3.0-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.3.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.3.1",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:2e46a25144d697761f0119deaf528c6f1c2ccd04e2acd95c661b0df4f32e1a9f
          createdAt: "2022-10-31 17:05:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.2.0-0 <2.3.0-4'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.3.0 release, AMQ Streams supports only OCP 4.8 to 4.12!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 2.2 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more details, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.3/html/deploying_and_upgrading_amq_streams_on_openshift/assembly-upgrade-str#assembly-upgrade-resources-str

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.3/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-streams-kafka-32-rhel8@sha256:0152a85f2d0bc85881dacaee62c6e8b0b8b04f57a542b9b82dc93ad5303d6bf9
        - registry.redhat.io/amq7/amq-streams-kafka-33-rhel8@sha256:046c2539fef32bd928fc1f4f4c41e63f9a143b371fcc201d78c66f48fc34ea32
        - registry.redhat.io/amq7/amq-streams-bridge-rhel8@sha256:c822b3d87f7dea233fc523238d8d0516e42f9a3d252f50f4d427c934995f9b22
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        - registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:2e46a25144d697761f0119deaf528c6f1c2ccd04e2acd95c661b0df4f32e1a9f
        version: 2.3.0-4
      entries:
      - name: amqstreams.v2.3.0-4
        version: 2.3.0-4
      name: amq-streams-2.3.x
    - currentCSV: amqstreams.v2.4.0-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:cd1fefcd199400a504b4b9942c24ba6becafcecc952a9e39fa2729228be79e77
          createdAt: "2022-10-31 17:05:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.3.0-0 <2.4.0-3'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.4.0 release, AMQ Streams supports only OCP 4.8 to 4.12!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more details, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.4/html/deploying_and_managing_amq_streams_on_openshift/assembly-upgrade-str)

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.4/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:8d467fc746b78440df06f6d88d266f6ac9b5a4574a2b17fd870c173a107f8d65
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:cd1fefcd199400a504b4b9942c24ba6becafcecc952a9e39fa2729228be79e77
        - registry.redhat.io/amq-streams/kafka-33-rhel8@sha256:aa780a250444c2426be79e95366050a33f4a4c9d68c08b58743413ea67a2d6d7
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:50b3be789993cfeedc4923a8e760c1cf52498b2178be3abfa3719fc2d15ce723
        version: 2.4.0-3
      entries:
      - name: amqstreams.v2.4.0-3
        version: 2.4.0-3
      - name: amqstreams.v2.4.0-2
        version: 2.4.0-2
      - name: amqstreams.v2.4.0-1
        version: 2.4.0-1
      name: amq-streams-2.4.x
    - currentCSV: amqstreams.v2.5.2-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
          createdAt: "2023-07-10 21:47:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.2.0-0 <2.5.2-8'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.5.2 release, AMQ Streams supports only OCP 4.12 to 4.16!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more information, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:8c7e8ccb6c2d5da32b70d9d0c3919c303ad73a25447f6993d964f7b19cca1072
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:4ede5eeed159bddacfe830a768bab4996003d4c57a496fc8ec9b41df916d4acc
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:95bc35951eb545b918114af1e92a7139e2e3323b7d054dcb2a060843c7326d28
        version: 2.5.2-8
      entries:
      - name: amqstreams.v2.5.2-8
        version: 2.5.2-8
      - name: amqstreams.v2.5.2-7
        version: 2.5.2-7
      - name: amqstreams.v2.5.2-6
        version: 2.5.2-6
      - name: amqstreams.v2.5.2-5
        version: 2.5.2-5
      - name: amqstreams.v2.5.2-4
        version: 2.5.2-4
      - name: amqstreams.v2.5.2-3
        version: 2.5.2-3
      - name: amqstreams.v2.5.2-2
        version: 2.5.2-2
      - name: amqstreams.v2.5.2-1
        version: 2.5.2-1
      - name: amqstreams.v2.5.2-0
        version: 2.5.2-0
      - name: amqstreams.v2.5.1-5
        version: 2.5.1-5
      - name: amqstreams.v2.5.1-4
        version: 2.5.1-4
      - name: amqstreams.v2.5.1-3
        version: 2.5.1-3
      - name: amqstreams.v2.5.1-2
        version: 2.5.1-2
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.5.x
    - currentCSV: amqstreams.v2.6.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.6.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:dacf7ddffe7cbe2dc35e7d3290637b3334117e09dde986e6eaeaceded7069fbf
          createdAt: "2023-10-14 18:18:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.6.0-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: "**Red Hat AMQ Streams** is a massively scalable, distributed,
          and high performance data streaming platform based on the Apache Kafka®
          project. AMQ Streams provides an event streaming backbone that allows microservices
          and other application components to exchange data with extremely high throughput
          and low latency.\n**From the 2.6.0 release, AMQ Streams supports only OCP
          4.11 to 4.14! \n**The core capabilities include:** * A pub/sub messaging
          model, similar to a traditional enterprise messaging system, in which application
          components publish and consume events to/from an ordered stream\n* The long
          term, fault-tolerant storage of events\n* The ability for a consumer to
          replay streams of events\n* The ability to partition topics for horizontal
          scalability\n### Upgrades\n**!!! IMPORTANT !!!** **Direct upgrade from AMQ
          Stream 1.7 or earlier is not supported!** You have to upgrade first to one
          of the previous versions of AMQ Streams. You will also need to convert the
          CRD resources. For more information, see the [documentation for AMQ Streams
          on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).\n###
          Supported Features\n* **Manages the Kafka Cluster** - Deploys and manages
          all of the components of this complex application, including dependencies
          like Apache ZooKeeper® that are traditionally hard to administer.\n* **Includes
          Kafka Connect** - Allows for configuration of common data sources and sinks
          to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for mirroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashboards\n### Upgrading your Clusters\nThe
          AMQ Streams Operator understands how to run and upgrade between a set of
          Kafka versions. When specifying a new version in your config, check to make
          sure you aren't using any features that may have been removed. For more
          information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).\n###
          Storage\nAn efficient data storage infrastructure is essential to the optimal
          performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires
          block storage. The use of file storage (for example, NFS) is not recommended.\nThe
          AMQ Streams Operator supports three types of data storage:\n* Ephemeral
          (Recommended for development only!)\n* Persistent\n* JBOD (Just a Bunch
          of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nAMQ Streams
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation for the current release
          can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).\n"
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:dacf7ddffe7cbe2dc35e7d3290637b3334117e09dde986e6eaeaceded7069fbf
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:be7466ffd7f624f769eedca2e4106c597d2bf1d95cc9679f235e6dcdec66cfa9
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:fa8361e5089ab4449261f5a0ad57bd0bd059dc254b94bd9ab68340faae386947
        - registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:8b10bfb697b48ba3ed246fb846384f4cc67f05f670e9521edf0f47f829869404
        version: 2.6.0-2
      entries:
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.6.x
    - currentCSV: amqstreams.v2.7.0-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:671c1a2fcb0856cb208c0ead539fff9571fa2281a1e3d45b3df9ebd3df8c787e
          createdAt: "2024-03-09 21:27:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0-0 <2.7.0-7'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.6.0 release, Streams supports only OCP 4.12 to 4.14!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:671c1a2fcb0856cb208c0ead539fff9571fa2281a1e3d45b3df9ebd3df8c787e
        - registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:66362c800b78003d26df9471b7f21f5c9ad7fca8e8fd2511428dfc549cdf46c2
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:be9b82af13cbc120bb415abf4fcc97420e543172d6450b836c47344e4dcacd00
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:a7620559f3e5a0a4d5ecee44538619f698158847d9d0a06c7168404ed6f7cd36
        - registry.redhat.io/amq-streams/kafka-36-rhel9@sha256:8890c3217343d9aa3dd2ab9c7fa1742f960b5a26e09906581ce6075e61dc2888
        version: 2.7.0-7
      entries:
      - name: amqstreams.v2.7.0-7
        version: 2.7.0-7
      - name: amqstreams.v2.7.0-6
        version: 2.7.0-6
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.7.x
    - currentCSV: amqstreams.v2.8.0-0-0.1738265624.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:675d4acc375fb20e1a3fd5755184a318341c15f9337a11ea41e902fcda6d0bec
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.0-0 <2.8.0-0'
          olm.substitutesFor: amqstreams.v2.8.0-0
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.8.0 release, Streams supports only OCP 4.12 to 4.14!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:675d4acc375fb20e1a3fd5755184a318341c15f9337a11ea41e902fcda6d0bec
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:b46457a0a7f63518bc12b2789f75afb0e44496e62de918b9b584129c7078de86
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:4df906cf1490922a4587d96250f3be6ff605fe2caf87dfba7b812c144c5c5774
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:e771a84a458388c07ecf970ec625e4203c18e5d2bd91e7571d9c44ebb09313de
        - registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:17b2d255439f38c6ef54c0a56929ae636743b12c688d42223ac3fb58729e51ad
        version: 2.8.0-0+0.1738265624.p
      entries:
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.8.x
    - currentCSV: amqstreams.v2.9.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.9.2-1'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.9.0 release, Streams supports only OCP 4.14 to 4.18!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:f5985592b7f05c9b3c64caa0ac90f34f26b9b4277b342b6bd4a6415d2c563f9a
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:86ea2e823cd83c476521f2a1e9da5619c4810661c60d28295d0f5482c8aa519d
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:ddfb50ec2f1cf45ccccc632182bf0ff549c14969da35b7248802eab67d9c2e9d
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:67647bff1a76adb5b51f43c99d5f88a8aa38055c1e87190add8972bb621ddc15
        version: 2.9.2-1
      entries:
      - name: amqstreams.v2.9.2-1
        version: 2.9.2-1
      - name: amqstreams.v2.9.2-0
        version: 2.9.2-0
      - name: amqstreams.v2.9.1-2
        version: 2.9.1-2
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.9.x
    - currentCSV: amqstreams.v2.9.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.9.2-1'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.9.0 release, Streams supports only OCP 4.14 to 4.18!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:86ea2e823cd83c476521f2a1e9da5619c4810661c60d28295d0f5482c8aa519d
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:ddfb50ec2f1cf45ccccc632182bf0ff549c14969da35b7248802eab67d9c2e9d
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:67647bff1a76adb5b51f43c99d5f88a8aa38055c1e87190add8972bb621ddc15
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:f5985592b7f05c9b3c64caa0ac90f34f26b9b4277b342b6bd4a6415d2c563f9a
        version: 2.9.2-1
      entries:
      - name: amqstreams.v2.9.2-1
        version: 2.9.2-1
      - name: amqstreams.v2.9.2-0
        version: 2.9.2-0
      - name: amqstreams.v2.9.1-2
        version: 2.9.1-2
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.x
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-3.0.x
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-3.x
    - currentCSV: amqstreams.v2.5.2-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
          createdAt: "2023-07-10 21:47:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.2.0-0 <2.5.2-8'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.5.2 release, AMQ Streams supports only OCP 4.12 to 4.16!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more information, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:4ede5eeed159bddacfe830a768bab4996003d4c57a496fc8ec9b41df916d4acc
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:95bc35951eb545b918114af1e92a7139e2e3323b7d054dcb2a060843c7326d28
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:8c7e8ccb6c2d5da32b70d9d0c3919c303ad73a25447f6993d964f7b19cca1072
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
        version: 2.5.2-8
      entries:
      - name: amqstreams.v2.5.2-8
        version: 2.5.2-8
      - name: amqstreams.v2.5.2-7
        version: 2.5.2-7
      - name: amqstreams.v2.5.2-6
        version: 2.5.2-6
      - name: amqstreams.v2.5.2-5
        version: 2.5.2-5
      - name: amqstreams.v2.5.2-4
        version: 2.5.2-4
      - name: amqstreams.v2.5.2-3
        version: 2.5.2-3
      - name: amqstreams.v2.5.2-2
        version: 2.5.2-2
      - name: amqstreams.v2.5.2-1
        version: 2.5.2-1
      - name: amqstreams.v2.5.2-0
        version: 2.5.2-0
      - name: amqstreams.v2.5.1-5
        version: 2.5.1-5
      - name: amqstreams.v2.5.1-4
        version: 2.5.1-4
      - name: amqstreams.v2.5.1-3
        version: 2.5.1-3
      - name: amqstreams.v2.5.1-2
        version: 2.5.1-2
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-lts
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: stable
    defaultChannel: stable
    packageName: amq-streams
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: manila-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: manila-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "manila.openstack.org/v1beta1",
                "kind": "Manila",
                "metadata": {
                  "name": "manila",
                  "namespace": "openstack"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "manila",
                  "databaseInstance": "openstack",
                  "manilaAPI": {},
                  "manilaScheduler": {},
                  "manilaShares": {
                    "share1": {}
                  },
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "osp-secret",
                  "serviceUser": "manila"
                }
              },
              {
                "apiVersion": "manila.openstack.org/v1beta1",
                "kind": "ManilaAPI",
                "metadata": {
                  "name": "manilaapi-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "manila.openstack.org/v1beta1",
                "kind": "ManilaScheduler",
                "metadata": {
                  "name": "manilascheduler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "manila.openstack.org/v1beta1",
                "kind": "ManilaShare",
                "metadata": {
                  "name": "manilashare-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:57Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManilaAPI is the Schema for the manilaapis API
            displayName: Manila API
            kind: ManilaAPI
            name: manilaapis.manila.openstack.org
            version: v1beta1
          - description: Manila is the Schema for the manilas API
            displayName: Manila
            kind: Manila
            name: manilas.manila.openstack.org
            version: v1beta1
          - description: ManilaScheduler is the Schema for the manilaschedulers API
            displayName: Manila Scheduler
            kind: ManilaScheduler
            name: manilaschedulers.manila.openstack.org
            version: v1beta1
          - description: ManilaShare is the Schema for the manilashares API
            displayName: Manila Share
            kind: ManilaShare
            name: manilashares.manila.openstack.org
            version: v1beta1
        description: Manila Operator
        displayName: Manila Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Share
        - Manila
        links:
        - name: Manila Operator
          url: https://github.com/openstack-k8s-operators/manila-operator
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/manila-rhel9-operator@sha256:1208676b94d8a726c277088591d898e08a3a8331a0125ba7bff101152afba934
        - registry.redhat.io/rhoso/openstack-manila-scheduler-rhel9@sha256:7a8f645126c073213ce066cdcac2164df51276f8aa4a3f89954664d3010e6fae
        - registry.redhat.io/rhoso/openstack-manila-share-rhel9@sha256:83a6d292785805a0ac2895810ea466cfb6e1cb7aa1a53ae8f4e22671381490df
        - registry.redhat.io/rhoso/openstack-manila-api-rhel9@sha256:4cbf725169614c2974d6d1a683cfc0179bfe06ca29cf30f7dca88caeb368c51a
        version: 1.0.6
      entries:
      - name: manila-operator.v1.0.6
        version: 1.0.6
      - name: manila-operator.v1.0.5
        version: 1.0.5
      - name: manila-operator.v1.0.4
        version: 1.0.4
      - name: manila-operator.v1.0.3
        version: 1.0.3
      - name: manila-operator.v1.0.2
        version: 1.0.2
      - name: manila-operator.v1.0.1
        version: 1.0.1
      - name: manila-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: manila-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JFrog
      provider-url: ""
    name: openshiftxray-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: xray-operator.v2.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftXray",
                "metadata": {
                  "name": "openshiftxray"
                },
                "spec": {
                  "xray": {
                    "common": {
                      "xrayUserId": "1000721035",
                      "xrayGroupId": "1000721035"
                    },
                    "analysis": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-analysis"
                      },
                      "name": "xray-analysis",
                      "podManagementPolicy": "Parallel",
                      "preStartCommand": null,
                      "updateStrategy": "RollingUpdate"
                    },
                    "database": {
                      "password": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "global": {
                      "postgresqlTlsSecret": null
                    },
                    "indexer": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-indexer"
                      },
                      "name": "xray-indexer",
                      "podManagementPolicy": "Parallel",
                      "updateStrategy": "RollingUpdate"
                    },
                    "persist": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-persist"
                      },
                      "name": "xray-persist",
                      "persistence": {
                        "size": "10Gi"
                      },
                      "podManagementPolicy": "Parallel",
                      "preStartCommand": null,
                      "updateStrategy": "RollingUpdate"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "rabbitmq": {
                      "enabled": true,
                      "replicaCount": 1,
                      "rbac": {
                        "create": true
                      },
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-rabbitmq"
                      },
                      "podSecurityContext": {
                        "fsGroup": 1000721001,
                        "runAsUser": 1000721001
                      },
                      "auth": {
                        "username": "xray",
                        "password": "xray",
                        "erlangCookie": "XRAYRABBITMQCLUSTER",
                        "tls": {
                          "enabled": false,
                          "failIfNoPeerCert": true,
                          "sslOptionsVerify": "verify_peer",
                          "caCertificate": "",
                          "serverCertificate": "",
                          "serverKey": ""
                        }
                      }
                    },
                    "rabbitmq-ha": {
                      "enabled": false,
                      "image": {
                        "repository": "registry.connect.redhat.com/jfrog/xray-rabbitmq"
                      },
                      "rabbitmqUsername": "guest",
                      "rabbitmqPassword": "guest",
                      "managementUsername": "guest",
                      "managementPassword": "guest",
                      "initContainer": {
                        "enabled": false
                      },
                      "securityContext": {
                        "fsGroup": 1000721035,
                        "runAsUser": 1000721035,
                        "runAsGroup": 1000721035
                      },
                      "replicaCount": 1
                    },
                    "replicaCount": 1,
                    "unifiedUpgradeAllowed":  "true",
                    "router": {
                      "image": {
                        "imagePullPolicy": "IfNotPresent",
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-router"
                      },
                      "name": "router"
                    },
                    "server": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-server"
                      },
                      "name": "xray-server",
                      "podManagementPolicy": "Parallel",
                      "replicaCount": 1,
                      "updateStrategy": "RollingUpdate"
                    },
                    "xray": {
                      "consoleLog": false,
                      "jfrogUrl": "OVERRIDE",
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/xray-operator@sha256:da7f6977170876b335875f4074ddfc0266b665351249295ce3c53f0e95891cc2
          createdAt: 2020-05-22 00:00:00+00:00
          description: JFrog Xray Enterprise deploys Xray continuous security scanner
            into Openshift (Requires Jfrog Artifactory)
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Xray Instances
            displayName: Xray
            kind: OpenshiftXray
            name: openshiftxrays.charts.helm.k8s.io
            version: v1
        description: "## Overview Openshift Operator to deploy JFrog Xray Continuous
          Security scanner into your Openshift cluster. NOTE: Artifactory is required
          for JFrog Xray to connect with and scan artifacts against.\n### Major version
          v2.0.4 Breaking change notification!!\nThis release has major bug fix for
          Rabbitmq Pod. Please update to version 2.0.4 of the operator.\n### Major
          version v2.0.0 Breaking change notification!! Xray is dependent upon Rabbitmq.
          The version of Rabbitmq that this chart uses has significantly changed from
          version 1.1.8 of the operator. For full details on how to migrate please
          visit: `https://github.com/jfrog/charts/blob/master/stable/xray/RABBITMQ_MIGRATION_NOTES.md`\nIf
          you are using operator version < 2.1.0 there is a issue with upgrade where
          xray version is not updated if you update operator version.As workaround
          you can update the Custom Resource OpenshiftXray which has hardcored image
          tag value.After upgrading to version 2.1.0 you can remove the image tags
          parameters from your CR so that future upgrade works without workaround.\n###
          Security Context Constraints To deploy this helm chart you will need to
          be a cluster admin w/ access to the anyuid scc and add the operator service
          account to the anyuid scc. ``` oc adm policy add-scc-to-user anyuid -z xray-operator
          ``` ``` oc adm policy add-scc-to-user anyuid -z openshiftxray ``` ``` oc
          adm policy add-scc-to-user anyuid -z openshiftxray-rabbitmq-ha ``` ## Usage
          An external DB is required. The operator will not deploy a DB but will require
          you to specify the configuration values to connect to it. Search for JFrog
          and click JFrog Xray Operator to install. Go to the Installed Operators.
          Wait for the JFrog Xray Operator to complete the installation. Open the
          Operator and click on the provided API: Xray Click Create New Instance and
          provide the following parameters for your DB configuration: ``` JFROG_URL
          DATABASE_URL DATABASE_USER DATABASE_PASSWORD ``` JFROG_URL is the external
          ip or DNS of your Artifactory to connect Xray to. Artifactory is required
          to use this operator. You can set your JFROG_URL to the service name of
          your Artifactory Nginx: ``` oc get svc -n my_namespace | grep nginx ```
          DATABASE_URL must be a Postgresql URL in the format: `` postgres://postgres-postgresql:5432/xraydb?sslmode=disable
          `` DATABASE_USER and DATABASE_PASSWORD must supply a valid user on Postgresql.
          Click Create for Xray to deploy into OpenShift. Open Artifactory in a web
          browser to complete the onboarding wizard for Xray! ### Air gap environments
          To use Xray Operator in an air gap environment you will need to download
          the images as image streams into your Openshift air gap cluster manually.
          Use the image overrides to then specify the image stream locations that
          are local to your cluster. Next you will need to setup the database of index
          data for xray to use when scanning artifacts. Follow the link below for
          instructions on setup: https://www.jfrog.com/confluence/display/JFROG/Configuring+Xray#ConfiguringXray-SynchronizingtheDatabase "
        displayName: JFrog Xray Continuous Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Xray Continuous Security
          url: https://jfrog.com/xray/
        - name: JFrog Xray Continuous Security Video
          url: https://youtu.be/wGzyKlQklKQ
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: alpha
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/xray-indexer@sha256:7f40d2dfe4a35b25d16dfcbfa04b4a58f227c132d4c3c86345e7b2bd19a97311
        - registry.connect.redhat.com/jfrog/xray-router@sha256:019235289e17d0e22dcd23a1505f71a78710827dfab338d8fa980f64c14f162f
        - registry.connect.redhat.com/jfrog/xray-rabbitmq@sha256:459d7d5e78192a6f45a5527e8fbaf1d8b30299b10ad4c42515d3d2bb5fcc75f4
        - registry.connect.redhat.com/jfrog/init@sha256:3f27e44b44c3556c50318110dbd542e34ef1061e28d7f0e0e7b8baa73c5f3e87
        - registry.connect.redhat.com/jfrog/xray-server@sha256:c5584c4e9528f0199e840b2b1ad1144220850ca9595ebf3e270bc6f86de4a71a
        - registry.connect.redhat.com/jfrog/xray-analysis@sha256:bcd8211f1bc753bfd7735165ce39b476e4efff198bffa877883620b41fcb33b5
        - registry.connect.redhat.com/jfrog/xray-persist@sha256:93fae0c124919a57d5162e32e88d3456ef6254889e454f89c21c89e69e6ae323
        - registry.connect.redhat.com/jfrog/xray-operator@sha256:da7f6977170876b335875f4074ddfc0266b665351249295ce3c53f0e95891cc2
        version: 2.1.6
      entries:
      - name: xray-operator.v2.1.6
        version: 2.1.6
      name: alpha
    defaultChannel: alpha
    packageName: openshiftxray-operator
    provider:
      name: JFrog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Aerospike
      provider-url: ""
    name: aerospike-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aerospike-kubernetes-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "memory-size": 3000000000,
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:6.3.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: docker.io/aerospike/aerospike-kubernetes-operator:2.5.0
          createdAt: 2023-05-04T09:32:47UTC
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Aerospike
        relatedImages:
        - docker.io/aerospike/aerospike-kubernetes-operator:2.5.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 2.5.0
      entries:
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      - name: aerospike-kubernetes-operator.v2.4.0
        version: 2.4.0
      - name: aerospike-kubernetes-operator.v2.3.0
        version: 2.3.0
      - name: aerospike-kubernetes-operator.v2.2.1
        version: 2.2.1
      - name: aerospike-kubernetes-operator.v2.2.0
        version: 2.2.0
      - name: aerospike-kubernetes-operator.v2.1.0
        version: 2.1.0
      name: alpha
    - currentCSV: aerospike-kubernetes-operator.v4.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "data-size": 1073741824,
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:8.1.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackup",
                "metadata": {
                  "name": "aerospikebackup",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "aerospike-cluster": {
                      "aerospike-aerospikebackup-test-cluster": {
                        "credentials": {
                          "password": "admin123",
                          "user": "admin"
                        },
                        "seed-nodes": [
                          {
                            "host-name": "aerocluster.aerospike.svc.cluster.local",
                            "port": 3000
                          }
                        ]
                      }
                    },
                    "backup-routines": {
                      "aerospike-aerospikebackup-test-routine": {
                        "backup-policy": "test-policy",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "local"
                      },
                      "aerospike-aerospikebackup-test-routine1": {
                        "backup-policy": "test-policy1",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "s3Storage"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackupService",
                "metadata": {
                  "name": "aerospikebackupservice-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "config": {
                    "backup-policies": {
                      "test-policy": {
                        "parallel": 3
                      },
                      "test-policy1": {
                        "parallel": 3
                      }
                    },
                    "service": {
                      "http": {
                        "port": 8081
                      }
                    },
                    "storage": {
                      "local": {
                        "local-storage": {
                          "path": "/tmp/localStorage"
                        }
                      },
                      "s3Storage": {
                        "s3-storage": {
                          "bucket": "aerospike-kubernetes-operator-test",
                          "s3-profile": "default",
                          "s3-region": "us-east-1"
                        }
                      }
                    }
                  },
                  "image": "aerospike/aerospike-backup-service:3.2.0",
                  "secrets": [
                    {
                      "secretName": "aws-secret",
                      "volumeMount": {
                        "mountPath": "/root/.aws/credentials",
                        "name": "aws-secret",
                        "subPath": "credentials"
                      }
                    }
                  ],
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeRestore",
                "metadata": {
                  "name": "aerospikerestore-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "backup-data-path": "aerospike-aerospikebackup-test-routine/backup/1733506015827/data/test",
                    "destination": {
                      "credentials": {
                        "password": "admin123",
                        "user": "admin"
                      },
                      "label": "destinationCluster",
                      "seed-nodes": [
                        {
                          "host-name": "aerocluster.test.svc.cluster.local",
                          "port": 3000
                        }
                      ]
                    },
                    "policy": {
                      "no-generation": true,
                      "no-indexes": true,
                      "parallel": 3
                    },
                    "source": {
                      "local-storage": {
                        "path": "/tmp/localStorage"
                      }
                    }
                  },
                  "type": "Full"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: docker.io/aerospike/aerospike-kubernetes-operator:4.1.1
          createdAt: "2025-09-15T10:14:05Z"
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          olm.skipRange: '>=4.0.0 <4.1.1'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
          support: Aerospike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeBackup is the Schema for the aerospikebackup API
            displayName: Aerospike Backup
            kind: AerospikeBackup
            name: aerospikebackups.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeBackupService is the Schema for the aerospikebackupservices
              API
            displayName: Aerospike Backup Service
            kind: AerospikeBackupService
            name: aerospikebackupservices.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1
          - description: AerospikeRestore is the Schema for the aerospikerestores
              API
            displayName: Aerospike Restore
            kind: AerospikeRestore
            name: aerospikerestores.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
           -   Aerospike cluster monitoring
           -   Backup and restore Aerospike clusters
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Aerospike
        relatedImages:
        - docker.io/aerospike/aerospike-kubernetes-operator:4.1.1
        version: 4.1.1
      entries:
      - name: aerospike-kubernetes-operator.v4.1.1
        version: 4.1.1
      - name: aerospike-kubernetes-operator.v4.1.0
        version: 4.1.0
      - name: aerospike-kubernetes-operator.v4.0.2
        version: 4.0.2
      - name: aerospike-kubernetes-operator.v4.0.1
        version: 4.0.1
      - name: aerospike-kubernetes-operator.v4.0.0
        version: 4.0.0
      - name: aerospike-kubernetes-operator.v3.4.1
        version: 3.4.1
      - name: aerospike-kubernetes-operator.v3.4.0
        version: 3.4.0
      - name: aerospike-kubernetes-operator.v3.3.1
        version: 3.3.1
      - name: aerospike-kubernetes-operator.v3.3.0
        version: 3.3.0
      - name: aerospike-kubernetes-operator.v3.2.2
        version: 3.2.2
      - name: aerospike-kubernetes-operator.v3.2.1
        version: 3.2.1
      - name: aerospike-kubernetes-operator.v3.2.0
        version: 3.2.0
      - name: aerospike-kubernetes-operator.v3.1.0
        version: 3.1.0
      - name: aerospike-kubernetes-operator.v3.0.0
        version: 3.0.0
      name: stable
    defaultChannel: stable
    packageName: aerospike-kubernetes-operator
    provider:
      name: Aerospike
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: 3scale-community-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: 3scale-community-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.8.2
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Improving [documentation](https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.8.2
        version: 0.8.2
      entries:
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.11
    - currentCSV: 3scale-community-operator.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.9.0
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.9.0/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.9.0/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.9.0
        version: 0.9.0
      entries:
      - name: 3scale-community-operator.v0.9.0
        version: 0.9.0
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.12
    - currentCSV: 3scale-community-operator.v0.10.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "ProductCRName": "product1-sample",
                  "Production": true
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.10.1
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.10.1/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Improving [documentation](https://github.com/3scale/3scale-operator/blob/master/doc/operator-user-guide.md)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.10.1/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.10.1
        version: 0.10.1
      entries:
      - name: 3scale-community-operator.v0.10.1
        version: 0.10.1
      - name: 3scale-community-operator.v0.9.0
        version: 0.9.0
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.13
    defaultChannel: threescale-2.13
    packageName: 3scale-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: https://www.intel.com
    name: ipm-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ipm-controller.v25.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ipm.intel.com/v1",
                "kind": "CPUPowerProv",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ipm-controller",
                    "node-label": "ipm-provisioning"
                  },
                  "name": "cpupowerprov-sample"
                },
                "spec": {
                  "cstates": {
                    "cores": {
                      "5": {
                        "C1": true,
                        "C1E": true,
                        "C6": false
                      },
                      "9,12": {
                        "C1": true,
                        "C6": true
                      }
                    },
                    "idleGovernor": "ladder"
                  },
                  "pstates": {
                    "cores": {
                      "5": {
                        "energyPerformanceBias": "normal",
                        "energyPerformancePreference": "performance",
                        "freqGovernor": "performance",
                        "maxFreqMHz": "3500",
                        "minFreqMHz": "800"
                      },
                      "9,12": {
                        "freqGovernor": "powersave",
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    },
                    "global": {
                      "conservative": {
                        "downThreshold": 20,
                        "freqStep": 5,
                        "ignoreNiceLoad": 0,
                        "samplingDownFactor": 1,
                        "samplingRate": 8888,
                        "upThreshold": 88
                      },
                      "maxPerfPct": 95,
                      "minPerfPct": 45,
                      "ondemand": {
                        "ignoreNiceLoad": 0,
                        "ioIsBusy": 1,
                        "powersaveBias": 0,
                        "samplingDownFactor": 1,
                        "samplingRate": 8888,
                        "upThreshold": 88
                      },
                      "schedutil": {
                        "rateLimitUs": 5000
                      },
                      "turbo": true
                    }
                  },
                  "uncoreFreqCPU": {
                    "cpu0": [
                      {
                        "maxFreqMHz": "1600",
                        "minFreqMHz": "800"
                      }
                    ],
                    "cpu1": [
                      {
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    ]
                  },
                  "uncoreFreqIO": {
                    "cpu0": [
                      {
                        "maxFreqMHz": "1600",
                        "minFreqMHz": "800"
                      }
                    ],
                    "cpu1": [
                      {
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/ipm-controller@sha256:32135c888f86a2c725d093ada9a59d222229bee5640d37f6e19a9e9812b4b67a
          createdAt: "2025-08-07T13:13:25Z"
          description: IPM Controller manages resources via REST API interactions.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.operatorframework.io/required-sccs: '["privileged"]'
          olm.securityContextConstraints.required: privileged
          operators.openshift.io/infrastructure-features: RestrictedNetwork
          operators.openshift.io/valid-subscription: Intel OBL Limited Distribution
            (Commercial Use) License Agreement [v2022.12.20]
          operators.operatorframework.io.bundle.package.v1: ipm-controller
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Intel® Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CPUPowerProv is the Schema for the cpupowerprovs API
            displayName: CPUPower Prov
            kind: CPUPowerProv
            name: cpupowerprovs.ipm.intel.com
            version: v1
        description: Infrastructure Power Manager (IPM) governs platform power technologies
          based on input from application workloads, the platform itself and customer
          input
        displayName: Intel® Infrastructure Power Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - intel
        - infrastructure
        - power
        - manager
        - management
        links:
        - name: IPM
          url: https://builders.intel.com/intel-technologies/power-management
        maintainers:
        - email: infrastructure.power.manager.operator@intel.com
          name: Infrastructure Power Manager Operator
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Intel® Corporation
          url: https://www.intel.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.13.0@sha256:471549d9cc07fcec0f285a05745b3d39a48e9a503b01b4047156a86ec1a30c00
        - registry.connect.redhat.com/intel/ipm-controller@sha256:32135c888f86a2c725d093ada9a59d222229bee5640d37f6e19a9e9812b4b67a
        - registry.connect.redhat.com/intel/ipm-agent@sha256:54de45cdc2e34768f6cdbf1cdee219406684143dddc4d236877709dc2536fe9c
        - registry.connect.redhat.com/intel/ipm-watcher@sha256:51eb773d5a4d1f401d4cbf2b42b8df4d0359beb85d5e274b8392dd7c5165e6ff
        version: 25.7.0
      entries:
      - name: ipm-controller.v25.7.0
        version: 25.7.0
      name: stable
    defaultChannel: stable
    packageName: ipm-controller
    provider:
      name: Intel® Corporation
      url: https://www.intel.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: ""
    name: nim-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nim-operator-certified.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMCache",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "source": {
                    "ngc": {
                      "modelPuller": "nvcr.io/nim/meta/llama-3.2-1b-instruct@sha256:1ff21ad32124d2382f8c733879394e7849cc3df5172651a24fcb671ac4fe77c6",
                      "pullSecret": "ngc-secret",
                      "authSecret": "ngc-api-secret",
                      "model": {
                        "lora": false,
                        "precision": "fp16",
                        "engine": "tensorrt_llm",
                        "qosProfile": "throughput",
                        "gpus": [
                          {
                            "product": "a100",
                            "ids": []
                          }
                        ]
                      }
                    }
                  },
                  "storage": {
                    "pvc": {
                      "create": true,
                      "storageClass": "local-path",
                      "size": "50Gi",
                      "volumeAccessMode": "ReadWriteOnce"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMService",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "image": {
                    "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                    "tag": "1.12.0",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [
                      "ngc-secret"
                    ]
                  },
                  "authSecret": "ngc-api-secret",
                  "storage": {
                    "nimCache": {
                      "name": "meta-llama3-1b-instruct",
                      "profile": ""
                    }
                  },
                  "scale": {
                    "enabled": false,
                    "hpa": {
                      "minReplicas": 1,
                      "maxReplicas": 10
                    }
                  },
                  "resources": {
                    "limits": {
                      "nvidia.com/gpu": 1
                    }
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMPipeline",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "k8s-nim-operator",
                    "app.kubernetes.io/managed-by": "k8s-nim-operator"
                  },
                  "name": "nimpipeline-sample"
                },
                "spec": {
                  "services": [
                    {
                      "name": "meta-llama3-1b-instruct",
                      "enabled": true,
                      "spec": {
                        "image": {
                          "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                          "tag": "1.12.0",
                          "pullPolicy": "IfNotPresent",
                          "pullSecrets": [
                            "ngc-secret"
                          ]
                        },
                        "authSecret": "ngc-api-secret",
                        "storage": {
                          "nimCache": {
                            "name": "meta-llama3-1b-instruct",
                            "profile": ""
                          }
                        },
                        "replicas": 1,
                        "resources": {
                          "limits": {
                            "nvidia.com/gpu": 1
                          }
                        },
                        "expose": {
                          "service": {
                            "type": "ClusterIP",
                            "port": 8000
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMBuild",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "nimCache": {
                    "name": "meta-llama3-1b-instruct",
                    "profile": "7cc8597690a35aba19a3636f35e7f1c7e7dbc005fe88ce9394cad4a4adeed414"
                  },
                  "image": {
                    "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                    "tag": "1.12.0",
                    "pullSecrets": [
                     "ngc-secret"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoCustomizer",
                "metadata": {
                  "name": "nemocustomizer-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "scheduler": {
                    "type": "volcano"
                  },
                  "wandb": {
                    "secretName": "wandb-secret",
                    "apiKeyKey": "apiKey",
                    "encryptionKey": "encryptionKey"
                  },
                  "otel": {
                    "enabled": true,
                    "exporterOtlpEndpoint": "http://customizer-otel-opentelemetry-collector.nemo.svc.cluster.local:4317"
                  },
                  "databaseConfig": {
                    "credentials": {
                      "user": "ncsuser",
                      "secretName": "customizer-pg-existing-secret",
                      "passwordKey": "password"
                    },
                    "host": "customizer-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "ncsdb"
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/customizer-api",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "entitystore": {
                    "endpoint": "http://nemoentitystore-sample.nemo.svc.cluster.local:8000"
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000"
                  },
                  "mlflow": {
                    "endpoint": "http://mlflow-tracking.nemo.svc.cluster.local:80"
                  },
                  "nemoDatastoreTools": {
                    "image": "nvcr.io/nvidia/nemo-microservices/nds-v2-huggingface-cli@sha256:7958e5244217de8af0ed723dfe7ea8e291b3afbb2a0aa73f5c27b218a1149b11"
                  },
                  "modelDownloadJobs": {
                    "image": "nvcr.io/nvidia/nemo-microservices/customizer-api@sha256:c590fa19121af2aac0379933deace3991f27c3ef88a8043560d6e3a172273b0b",
                    "ngcAPISecret": {
                      "name": "ngc-api-secret",
                      "key": "NGC_API_KEY"
                    },
                    "securityContext": {
                      "fsGroup": 1000,
                      "runAsNonRoot": true,
                      "runAsUser": 1000,
                      "runAsGroup": 1000
                    },
                    "ttlSecondsAfterFinished": 600,
                    "pollIntervalSeconds": 15
                  },
                  "modelConfig": {
                    "name": "nemo-model-config"
                  },
                  "trainingConfig": {
                    "configMap": {
                      "name": "nemo-training-config"
                    },
                    "modelPVC": {
                      "create": true,
                      "name": "finetuning-ms-models-pvc",
                      "storageClass": "",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": 50
                    },
                    "workspacePVC": {
                      "storageClass": "local-path",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": 10,
                      "mountPath": "/pvc/workspace"
                    },
                    "image": {
                      "repository": "nvcr.io/nvidia/nemo-microservices/customizer",
                      "tag": "25.08"
                    },
                    "env": [
                      { "name": "LOG_LEVEL", "value": "INFO" }
                    ],
                    "networkConfig": [
                      { "name": "NCCL_IB_SL", "value": "0" },
                      { "name": "NCCL_IB_TC", "value": "41" },
                      { "name": "NCCL_IB_QPS_PER_CONNECTION", "value": "4" },
                      { "name": "UCX_TLS", "value": "TCP" },
                      { "name": "UCX_NET_DEVICES", "value": "eth0" },
                      { "name": "HCOLL_ENABLE_MCAST_ALL", "value": "0" },
                      { "name": "NCCL_IB_GID_INDEX", "value": "3" }
                    ],
                    "ttlSecondsAfterFinished": 3600,
                    "timeout": 3600,
                    "tolerations": [
                      {
                        "key": "nvidia.com/gpu",
                        "operator": "Exists",
                        "effect": "NoSchedule"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoGuardrail",
                "metadata": {
                  "name": "nemoguardrails-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "configStore": {
                    "pvc": {
                      "name": "pvc-guardrail-config",
                      "create": true,
                      "storageClass": "",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": "1Gi"
                    }
                  },
                  "nimEndpoint": {
                    "baseURL": "http://meta-llama3-1b-instruct.nemo.svc.cluster.local:8000/v1"
                  },
                  "databaseConfig": {
                    "host": "guardrail-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "guardraildb",
                    "credentials": {
                      "user": "guardrailuser",
                      "secretName": "guardrail-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/guardrails",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [
                      "ngc-secret"
                    ]
                  },
                  "metrics": {
                    "serviceMonitor": {}
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "ephemeral-storage": "10Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoEvaluator",
                "metadata": {
                  "name": "nemoevaluator-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "evaluationImages": {
                    "bigcodeEvalHarness": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-bigcode@sha256:62f74c32d1ed07925ddb906de8975025bf63d73e82e7faefb840339aaf21a9dc",
                    "lmEvalHarness": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-lm-eval-harness@sha256:58e7c3da7f1edfed2ff508fa8abab7660366c2c17bb55d17aa68159d6e16be6d",
                    "similarityMetrics": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-custom-eval@sha256:eb51edf668481d1f5d2b100c850958dadf184fd8779433e45da40fa2a35ec4b9",
                    "llmAsJudge": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055",
                    "mtBench": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055",
                    "retriever": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-retriever@sha256:cb99fb3ea5b4e9989832679d4c6028573c28b99e43744e008ac5c9748b7dd250",
                    "rag": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-rag@sha256:b5e35cf277185e0901a4bbba35780b3673be71ae2c5e8efcdbab7710a8a6f783",
                    "bfcl": "nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-bfcl@sha256:95f45f66140d481d544b36b16bc3de7ad5e090d76225fd73ed70461b9bd9907c",
                    "agenticEval": "nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-agentic-eval@sha256:b38bbc5e61df0930204efa096cf4f40db4655b92d03741190513b019d67fc7dc"
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/evaluator",
                    "tag": "25.06",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "argoWorkflows": {
                    "endpoint": "https://argo-workflows-server.nemo.svc.cluster.local:2746",
                    "serviceAccount": "argo-workflows-executor"
                  },
                  "vectorDB": {
                    "endpoint": "http://milvus.nemo.svc.cluster.local:19530"
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000/v1/hf"
                  },
                  "entitystore": {
                    "endpoint": "http://nemoentitystore-sample.nemo.svc.cluster.local:8000"
                  },
                  "databaseConfig": {
                    "host": "evaluator-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "evaldb",
                    "credentials": {
                      "user": "evaluser",
                      "secretName": "evaluator-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "otel": {
                    "enabled": true,
                    "exporterOtlpEndpoint": "http://evaluator-otel-opentelemetry-collector.nemo.svc.cluster.local:4317"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoEntitystore",
                "metadata": {
                  "name": "nemoentitystore-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/entity-store",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "databaseConfig": {
                    "databaseName": "nesdb",
                    "host": "entity-store-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "credentials": {
                      "user": "nesuser",
                      "secretName": "entity-store-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000"
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoDatastore",
                "metadata": {
                  "name": "nemodatastore-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "secrets": {
                    "datastoreConfigSecret": "nemo-ms-nemo-datastore",
                    "datastoreInitSecret": "nemo-ms-nemo-datastore-init",
                    "datastoreInlineConfigSecret": "nemo-ms-nemo-datastore-inline-config",
                    "giteaAdminSecret": "gitea-admin-credentials",
                    "lfsJwtSecret": "nemo-ms-nemo-datastore--lfs-jwt"
                  },
                  "databaseConfig": {
                    "credentials": {
                      "user": "ndsuser",
                      "secretName": "datastore-pg-existing-secret",
                      "passwordKey": "password"
                    },
                    "host": "datastore-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "ndsdb"
                  },
                  "pvc": {
                    "name": "pvc-shared-data",
                    "create": true,
                    "storageClass": "",
                    "volumeAccessMode": "ReadWriteOnce",
                    "size": "10Gi"
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/datastore",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "replicas": 1,
                  "resources": {
                    "requests": {
                      "memory": "256Mi",
                      "cpu": "500m"
                    },
                    "limits": {
                      "memory": "512Mi",
                      "cpu": "1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: nvcr.io/nvidia/cloud-native/k8s-nim-operator@sha256:f42dd5c3dc269553ee46725ed4b1ab2d2f253ef06117f1d9d52faaa22d8fe60d
          createdAt: Wed Sep 10 05:49:48 PDT 2025
          description: An Operator for deployment and maintenance of NVIDIA NIMs and
            NeMo microservices in a Kubernetes environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          k8sMinVersion: 1.26.0
          repository: github.com/NVIDIA/k8s-nim-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NIM Cache
            displayName: NIMCache
            kind: NIMCache
            name: nimcaches.apps.nvidia.com
            version: v1alpha1
          - description: NIM Service
            displayName: NIMService
            kind: NIMService
            name: nimservices.apps.nvidia.com
            version: v1alpha1
          - description: NIM Pipeline
            displayName: NIMPipeline
            kind: NIMPipeline
            name: nimpipelines.apps.nvidia.com
            version: v1alpha1
          - description: NIM Build
            displayName: NIMBuild
            kind: NIMBuild
            name: nimbuilds.apps.nvidia.com
            version: v1alpha1
          - description: NEMO guardrail Service
            displayName: NemoGuardrail
            kind: NemoGuardrail
            name: nemoguardrails.apps.nvidia.com
            version: v1alpha1
          - description: NEMO datastore Service
            displayName: NemoDatastore
            kind: NemoDatastore
            name: nemodatastores.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Eval Service
            displayName: NemoEvaluator
            kind: NemoEvaluator
            name: nemoevaluators.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Customizer Service
            displayName: NemoCustomizer
            kind: NemoCustomizer
            name: nemocustomizers.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Entitystore Service
            displayName: NemoEntitystore
            kind: NemoEntitystore
            name: nemoentitystores.apps.nvidia.com
            version: v1alpha1
        description: |-
          ## NVIDIA NIM Operator
          The NVIDIA NIM Operator enables Kubernetes cluster administrators to operate the software components and services necessary to run NVIDIA NIMs in various domains such as reasoning, retrieval, speech, and biology. Additionally, it allows the use of NeMo Microservices to fine-tune, evaluate, or apply guardrails to your models.
          The Operator manages the life cycle of the following microservices and the models they use:
          ### NVIDIA NIM models, such as:
          * Reasoning LLMs
          * Retrieval - Embedding, Reranking, etc.
          * Speech
          * Biology
          ### NeMo core microservices:
          * NeMo Customizer
          * NeMo Evaluator
          * NeMo Guardrails
          ### NeMo platform component microservices:
          * NeMo Data Store
          * NeMo Entity Store
        displayName: The NVIDIA NIM Operator for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        links:
        - name: NVIDIA NIM
          url: https://docs.nvidia.com/nim/index.html
        - name: Documentation
          url: https://docs.nvidia.com/nim-operator/latest
        maintainers:
        - email: operator_feedback@nvidia.com
          name: shivamerla
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: NVIDIA
        relatedImages:
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-retriever@sha256:cb99fb3ea5b4e9989832679d4c6028573c28b99e43744e008ac5c9748b7dd250
        - nvcr.io/nvidia/nemo-microservices/customizer-api@sha256:c590fa19121af2aac0379933deace3991f27c3ef88a8043560d6e3a172273b0b
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-lm-eval-harness@sha256:58e7c3da7f1edfed2ff508fa8abab7660366c2c17bb55d17aa68159d6e16be6d
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-custom-eval@sha256:eb51edf668481d1f5d2b100c850958dadf184fd8779433e45da40fa2a35ec4b9
        - nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-bfcl@sha256:95f45f66140d481d544b36b16bc3de7ad5e090d76225fd73ed70461b9bd9907c
        - nvcr.io/nvidia/cloud-native/k8s-nim-operator@sha256:f42dd5c3dc269553ee46725ed4b1ab2d2f253ef06117f1d9d52faaa22d8fe60d
        - nvcr.io/nvidia/nemo-microservices/nds-v2-huggingface-cli@sha256:7958e5244217de8af0ed723dfe7ea8e291b3afbb2a0aa73f5c27b218a1149b11
        - nvcr.io/nim/meta/llama-3.2-1b-instruct@sha256:1ff21ad32124d2382f8c733879394e7849cc3df5172651a24fcb671ac4fe77c6
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-bigcode@sha256:62f74c32d1ed07925ddb906de8975025bf63d73e82e7faefb840339aaf21a9dc
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-rag@sha256:b5e35cf277185e0901a4bbba35780b3673be71ae2c5e8efcdbab7710a8a6f783
        - nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-agentic-eval@sha256:b38bbc5e61df0930204efa096cf4f40db4655b92d03741190513b019d67fc7dc
        version: 3.0.0
      entries:
      - name: nim-operator-certified.v3.0.0
        version: 3.0.0
      - name: nim-operator-certified.v2.0.2
        version: 2.0.2
      - name: nim-operator-certified.v2.0.1
        version: 2.0.1
      - name: nim-operator-certified.v2.0.0
        version: 2.0.0
      - name: nim-operator-certified.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: stable
    packageName: nim-operator-certified
    provider:
      name: NVIDIA
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-druid-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: druid-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:909a03d95f8173c418f6547092421a5531ff8957aa4319ce5fc890f32fe1f7be
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:909a03d95f8173c418f6547092421a5531ff8957aa4319ce5fc890f32fe1f7be
        version: 24.11.1
      entries:
      - name: druid-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: druid-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:253239268b8d32dbea10233fe6ab202bbe8a074c369a929630681dea37935383
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:253239268b8d32dbea10233fe6ab202bbe8a074c369a929630681dea37935383
        version: 25.3.0
      entries:
      - name: druid-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: druid-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
        version: 25.7.0
      entries:
      - name: druid-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: druid-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
        version: 25.7.0
      entries:
      - name: druid-operator.v25.7.0
        version: 25.7.0
      - name: druid-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-druid-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: eap
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eap-operator.v3.2.2
      currentCSVDesc:
        annotations:
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "false"
          containerImage: registry.redhat.io/jboss-eap/eap-rhel9-operator@sha256:2361e73b8c517177e80f60d204996ebf65b166e4e8b60b329323a2b409c54f6c
          createdAt: "2025-07-10T14:38:00Z"
          description: Operator that creates and manages Java applications running
            on JBoss EAP.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat JBoss Enterprise Application
            Platform", "Red Hat Runtimes", "Red Hat Integration", "Red Hat Process
            Automation"]'
          repository: https://github.com/wildfly/wildfly-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An application running on EAP application runtime.
            displayName: WildFlyServer
            kind: WildFlyServer
            name: wildflyservers.wildfly.org
            version: v1alpha1
        description: "# Red Hat JBoss Enterprise Application Platform (JBoss EAP)\n\nRed
          Hat® JBoss® Enterprise Application Platform (JBoss EAP) delivers enterprise-grade
          security, performance, and scalability in any environment.  Some features
          of JBoss EAP are:\n\n**Optimized for cloud and containers**: with simplified
          deployment and full Jakarta EE performance for applications in any environment.\n\n**Lightweight,
          flexible architecture**: which is well-suited for microservices and traditional
          applications.\n\n**Support and standardize microservices development**:
          With the JBoss Enterprise Application Platform expansion pack that supports
          Eclipse MicroProfile application programming interfaces (APIs).\n\n# JBoss
          EAP Operator\nOperators are software components that extend the Kubernetes
          API to customize how to manage applications and their components.\n\nThe
          JBoss EAP Operator facilitates the deployment and management of JBoss EAP
          on container orchestration platforms such as Red Hat OpenShift. The operator
          acts as an automated management tool, simplifying the process of provisioning,
          scaling, and updating JBoss EAP applications.\n\n# JBoss EAP Operator capabilities\nThe
          JBoss EAP Operator deploys and manages multiple instances of the JBoss EAP
          Java application in a Red Hat OpenShift cluster. \n\n**Simplifies management**:
          Creates, configures, manages, and seamlessly upgrades instances of complex
          stateful applications. No more manual configuration, making management a
          breeze.\n\n**Efficient scaling**: Effortlessly scales the instances of your
          JBoss EAP Java application based on demand.\n\n**Faster updates**: Seamlessly
          updates, enabling you to roll out new operator features and bug fixes without
          downtime\n\n**Safe transaction recovery**: Provides safe transaction recovery
          in your application cluster by verifying all transactions are completed
          before scaling down the replicas and marking a pod as clean for termination.\n\n#
          Operator installation and configuration\nLet's walk through the steps to
          install the JBoss EAP Operator on your OpenShift Cluster and deploy your
          first example application:\n\n**Step 1: Install JBoss EAP from OperatorHub**\n1.
          Click Install in the JBoss EAP operator window.\n2. In the Install Operator
          window, for Installation mode, select All namespaces on the cluster The
          operator is available in openshift-operators namespace.\n\n**Step 2: Build
          an application image**\n\nThe JBoss EAP Operator deploys an application
          image. If you do not have your JBoss EAP application image built yet, you
          can use the following example to build the image using Helm Charts.\n\n1.
          In the main navigation, click the dropdown menu (perspective switcher) and
          select Developer.\n2. Click Add > Helm Chart.\n3. Search for \"JBoss EAP
          8\" (Provided by Red Hat) and create a new Helm Chart release.\n4. On the
          Release name text box, type “eap-application”. \n5. Switch to the YAML view
          and replace the existing code with the following:\n\n    ```\n    build:\n
          \     uri: https://github.com/jboss-eap-up-and-running/eap8-getting-started\n
          \   deploy:\n      enabled: false\n    ```\n\n6. Click Create to build the
          application image.\n\n    The application image is created with the name
          eap-application:latest.\n\n**Step 3: Deploy your application with JBoss
          EAP Operator**\n\n1. In the main navigation, click the dropdown menu (perspective
          switcher) and select Developer.\n2. Click Add > Operator Backed > WildFlyServer
          > Create.\n3. Switch to the YAML view and replace the existing code with
          the following:\n\n    ```\n    apiVersion: wildfly.org/v1alpha1\n    kind:
          WildFlyServer\n    metadata:\n      name: eap-application\n    spec:\n      applicationImage:
          eap-application:latest\n      replicas: 1\n    ```\n\n 4. Click Create to
          deploy the application with JBoss EAP Operator.\n"
        displayName: JBoss EAP
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eap
        - java
        - open source
        - application runtime
        links:
        - name: JBoss EAP Homepage
          url: https://www.redhat.com/en/technologies/jboss-middleware/application-platform
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_jboss_enterprise_application_platform/
        - name: Blog
          url: https://developers.redhat.com/blog/
        - name: Quickstart
          url: https://github.com/jboss-developer/jboss-eap-quickstarts
        - name: Red Hat Customer Portal
          url: https://access.redhat.com/support
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat, Inc.
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/jboss-eap/eap-rhel9-operator@sha256:2361e73b8c517177e80f60d204996ebf65b166e4e8b60b329323a2b409c54f6c
        version: 3.2.2
      entries:
      - name: eap-operator.v3.2.2
        version: 3.2.2
      - name: eap-operator.v3.2.1
        version: 3.2.1
      - name: eap-operator.v3.2.0
        version: 3.2.0
      - name: eap-operator.v3.1.9
        version: 3.1.9
      - name: eap-operator.v3.1.8
        version: 3.1.8
      - name: eap-operator.v3.1.7
        version: 3.1.7
      - name: eap-operator.v3.1.6
        version: 3.1.6
      - name: eap-operator.v3.1.5
        version: 3.1.5
      - name: eap-operator.v3.1.4
        version: 3.1.4
      - name: eap-operator.v3.1.3
        version: 3.1.3
      - name: eap-operator.v3.1.2
        version: 3.1.2
      - name: eap-operator.v3.1.1
        version: 3.1.1
      - name: eap-operator.v3.1.0
        version: 3.1.0
      - name: eap-operator.v3.0.2
        version: 3.0.2
      - name: eap-operator.v3.0.1
        version: 3.0.1
      - name: eap-operator.v3.0.0
        version: 3.0.0
      - name: eap-operator.v2.4.3
        version: 2.4.3
      - name: eap-operator.v2.4.2
        version: 2.4.2
      - name: eap-operator.v2.4.1
        version: 2.4.1
      - name: eap-operator.v2.4.0
        version: 2.4.0
      - name: eap-operator.v2.3.10
        version: 2.3.10
      - name: eap-operator.v2.3.9
        version: 2.3.9
      - name: eap-operator.v2.3.8
        version: 2.3.8
      - name: eap-operator.v2.3.7
        version: 2.3.7
      - name: eap-operator.v2.3.6
        version: 2.3.6
      - name: eap-operator.v2.3.5
        version: 2.3.5
      - name: eap-operator.v2.3.4
        version: 2.3.4
      - name: eap-operator.v2.3.3
        version: 2.3.3
      - name: eap-operator.v2.3.2
        version: 2.3.2
      - name: eap-operator.v2.3.1
        version: 2.3.1
      - name: eap-operator.v2.3.0
        version: 2.3.0
      - name: eap-operator.v2.2.2
        version: 2.2.2
      - name: eap-operator.v2.2.1
        version: 2.2.1
      - name: eap-operator.v2.2.0
        version: 2.2.0
      name: stable
    defaultChannel: stable
    packageName: eap
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: m88i Labs
      provider-url: ""
    name: nexus-operator-m88i
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.m88i.io/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "name": "nexus3"
                },
                "spec": {
                  "networking": {
                    "expose": false
                  },
                  "persistence": {
                    "persistent": false
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    },
                    "requests": {
                      "cpu": "1",
                      "memory": "2Gi"
                    }
                  },
                  "useRedHatImage": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/m88i/nexus-operator:0.6.0
          createdAt: "2019-11-16T13:12:22Z"
          description: Nexus Operator to deploy and manage Nexus 3.x servers
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/m88i/nexus-operator
          support: m88i Labs
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: Nexus
            kind: Nexus
            name: nexus.apps.m88i.io
            version: v1alpha1
        description: |-
          Creates a new Nexus 3.x deployment in a Kubernetes cluster. Will help DevOps to have a quick Nexus application exposed to the world that can be used in a CI/CD process:

          * Deploys a new Nexus 3.x server based on either Community or Red Hat images
          * Creates an [Ingress](https://kubernetes.io/docs/concepts/services-networking/ingress/) in Kubernetes (1.14+) environments to expose the application to the world
          * On OpenShift, creates a Route to expose the service outside the cluster
          * Automatically creates Apache, Red Hat and JBoss Maven repositories
          * Automatically updates Nexus within the same minor version

          [See our documentation](https://github.com/m88i/nexus-operator/blob/main/README.md) for more installation and usage scenarios.

          If you experience any issues or have any ideas for new features, please [file an issue in our Github repository](https://github.com/m88i/nexus-operator/issues) or send an email to our maillist: [nexus-operator@googlegroups.com](mailto:nexus-operator@googlegroups.com)

          *Please note that the operator is an individual work and it's not provided nor supported by Sonatype.*
        displayName: Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nexus
        - sonatype
        - maven
        - docker
        - ci
        - continuous integration
        - continuous delivery
        - repository
        - repository manager
        - dev tools
        links:
        - name: Documentation
          url: https://github.com/m88i/nexus-operator/blob/main/README.md
        - name: Nexus Operator source code repository
          url: https://github.com/m88i/nexus-operator
        maintainers:
        - email: nexus-operator@googlegroups.com
          name: m88i Labs
        maturity: alpha
        provider:
          name: m88i Labs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/m88i/nexus-operator:0.6.0
        version: 0.6.0
      entries:
      - name: nexus-operator.v0.6.0
        version: 0.6.0
      name: alpha
    defaultChannel: alpha
    packageName: nexus-operator-m88i
    provider:
      name: m88i Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
    name: flink-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flink-kubernetes-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"flink.apache.org/v1beta1",
                "kind":"FlinkDeployment",
                "metadata":{
                    "name":"basic-example"
                },
                "spec":{
                    "image":"flink:1.16",
                    "flinkVersion":"v1_16",
                    "flinkConfiguration":{
                        "taskmanager.numberOfTaskSlots":"2"
                    },
                    "serviceAccount":"flink",
                    "jobManager":{
                        "resource":{
                            "memory":"2048m",
                            "cpu":1
                        }
                    },
                    "taskManager":{
                        "resource":{
                            "memory":"2048m",
                            "cpu":1
                        }
                    },
                    "job":{
                        "jarURI":"local:///opt/flink/examples/streaming/StateMachineExample.jar",
                        "parallelism":2,
                        "upgradeMode":"stateless"
                    }
                }
              },
              {
                "apiVersion":"flink.apache.org/v1beta1",
                "kind":"FlinkSessionJob",
                "metadata":{
                  "name":"basic-session-job-example2"
                },
                "spec":{
                  "deploymentName":"basic-session-cluster",
                  "job":{
                      "jarURI":"https://repo1.maven.org/maven2/org/apache/flink/flink-examples-streaming_2.12/1.16.1/flink-examples-streaming_2.12-1.16.1.jar",
                      "parallelism":2,
                      "upgradeMode":"stateless",
                      "entryClass":"org.apache.flink.streaming.examples.statemachine.StateMachineExample"
                  }
                }
              },
              {
                "apiVersion": "flink.apache.org/v1beta1",
                "kind": "FlinkStateSnapshot",
                "metadata": {
                  "name": "example-savepoint"
                },
                "spec": {
                  "savepoint": {
                    "alreadyExists": true,
                    "disposeOnDelete": false,
                    "path": "/flink-data/savepoints/savepoint-98682a-21931bf10c39"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "basic-example": {
                "description": "Example flink deployment"
              },
              "basic-session-job-example2": {
                "description": "Example flink session job"
              }
            }
          capabilities: Deep Insights
          categories: Big Data, Streaming & Messaging
          containerImage: ghcr.io/apache/flink-kubernetes-operator:1.12.0
          createdAt: "2025-06-23 18:42:54"
          description: Flink Kubernetes Operator manages the complete deployment lifecycle
            of Apache Flink applications.
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/apache/flink-kubernetes-operator.git
          support: Apache Software Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The FlinkDeployment CR defines Flink Application and Session
              cluster deployments
            displayName: Flink Deployment
            kind: FlinkDeployment
            name: flinkdeployments.flink.apache.org
            version: v1beta1
          - description: The FlinkSessionJob CR defines the session job on the Session
              cluster and each Session cluster can run multiple FlinkSessionJob.
            displayName: Flink Session Job
            kind: FlinkSessionJob
            name: flinksessionjobs.flink.apache.org
            version: v1beta1
          - description: The FlinkStateSnapshots CR is used to describe savepoint
              or checkpoint for a Flink job.
            displayName: Flink State Snapshot
            kind: FlinkStateSnapshot
            name: flinkstatesnapshots.flink.apache.org
            version: v1beta1
        description: |-
          Flink Kubernetes Operator acts as a control plane to manage the complete deployment lifecycle of Apache Flink applications.
          ## Installation
          This operator can be installed from [OperatorHub.io](https://operatorhub.io/operator/flink-kubernetes-operator).

          By default, the Flink operator monitors and run Flink applications in the same namespace as the operator.
          To run Flink jobs in another namespace, users are responsible for setting up the RBAC in that namespace.

          For example, users need to do the following for each additional namespace that runs the Flink applications:

          1. Switch to the namespace by running:

              ```sh
              kubectl config set-context --current --namespace=CHANGEIT
              ```

          2. Create the service account, role, and role binding in the namespace using the commands below:

              ```sh
              kubectl apply -f - <<EOF
              apiVersion: v1
              kind: ServiceAccount
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink
              ---
              apiVersion: rbac.authorization.k8s.io/v1
              kind: Role
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink
              rules:
              - apiGroups:
                - ""
                resources:
                - pods
                - configmaps
                verbs:
                - '*'
              - apiGroups:
                - apps
                resources:
                - deployments
                - deployments/finalizers
                verbs:
                - '*'
              ---
              apiVersion: rbac.authorization.k8s.io/v1
              kind: RoleBinding
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink-role-binding
              roleRef:
                apiGroup: rbac.authorization.k8s.io
                kind: Role
                name: flink
              subjects:
              - kind: ServiceAccount
                name: flink
              EOF
              ```

          3. Proceed deploying Flink applications using the the Custom Resource Definition below.

              ```sh
              kubectl create -f https://raw.githubusercontent.com/apache/flink-kubernetes-operator/release-1.12.0/examples/basic.yaml
              ```

          See [Flink custom resources](https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/custom-resource/overview/#overview) for more detail.
        displayName: Flink Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flink
        - streaming
        links:
        - name: Website
          url: https://flink.apache.org/
        - name: Documentation
          url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/
        - name: Mailing list
          url: https://lists.apache.org/list.html?dev@flink.apache.org
        - name: Slack
          url: https://apache-flink.slack.com/join/shared_invite/zt-1llkzbgyt-K2nNGGg88rfsDGLkT09Qzg#/shared-invite/email
        - name: GitHub
          url: https://github.com/apache/flink-kubernetes-operator
        maintainers:
        - email: mbalassi@apache.org
          name: Marton Balassi
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: jbusche@us.ibm.com
          name: James Busche
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Community
          url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
        relatedImages:
        - ghcr.io/apache/flink-kubernetes-operator:1.12.0
        version: 1.12.0
      entries:
      - name: flink-kubernetes-operator.v1.12.0
        version: 1.12.0
      - name: flink-kubernetes-operator.v1.11.0
        version: 1.11.0
      - name: flink-kubernetes-operator.v1.10.0
        version: 1.10.0
      - name: flink-kubernetes-operator.v1.6.0
        version: 1.6.0
      - name: flink-kubernetes-operator.v1.5.0
        version: 1.5.0
      - name: flink-kubernetes-operator.v1.4.0
        version: 1.4.0
      - name: flink-kubernetes-operator.v1.3.1
        version: 1.3.1
      - name: flink-kubernetes-operator.v1.3.0
        version: 1.3.0
      name: alpha
    defaultChannel: alpha
    packageName: flink-kubernetes-operator
    provider:
      name: Community
      url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeRocketCI
      provider-url: https://docs.kuberocketci.io/
    name: edp-keycloak-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: edp-keycloak-operator.v1.19.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "keycloak-sample"
                },
                "spec": {
                  "secret": "my-keycloak-secret",
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakAuthFlow",
                "metadata": {
                  "name": "keycloakauthflow-sample"
                },
                "spec": {
                  "alias": "MyBrowser",
                  "authenticationExecutions": [
                    {
                      "authenticator": "auth-cookie",
                      "priority": 0,
                      "requirement": "ALTERNATIVE"
                    },
                    {
                      "authenticator": "identity-provider-redirector",
                      "authenticatorConfig": {
                        "alias": "my-alias",
                        "config": {
                          "defaultProvider": "my-alias"
                        }
                      },
                      "priority": 1,
                      "requirement": "REQUIRED"
                    }
                  ],
                  "builtIn": false,
                  "description": "browser with idp",
                  "providerId": "basic-flow",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "topLevel": true
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-sample"
                },
                "spec": {
                  "advancedProtocolMappers": true,
                  "clientId": "agocd",
                  "defaultClientScopes": [
                    "argocd_groups"
                  ],
                  "directAccess": true,
                  "public": false,
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-name:client-secret-key",
                  "webUrl": "https://argocd.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClientScope",
                "metadata": {
                  "name": "keycloakclientscope-sample"
                },
                "spec": {
                  "description": "Group Membership",
                  "name": "groups",
                  "protocol": "openid-connect",
                  "protocolMappers": [
                    {
                      "config": {
                        "access.token.claim": "true",
                        "claim.name": "groups",
                        "full.path": "false",
                        "id.token.claim": "true",
                        "userinfo.token.claim": "true"
                      },
                      "name": "groups",
                      "protocol": "openid-connect",
                      "protocolMapper": "oidc-group-membership-mapper"
                    }
                  ],
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "keycloakrealm-sample"
                },
                "spec": {
                  "id": "d1-id-kc-realm-name",
                  "keycloakRef": {
                    "kind": "Keycloak",
                    "name": "keycloak-sample"
                  },
                  "passwordPolicy": [
                    {
                      "type": "forceExpiredPasswordChange",
                      "value": "365"
                    },
                    {
                      "type": "length",
                      "value": "8"
                    }
                  ],
                  "realmEventConfig": {
                    "adminEventsDetailsEnabled": false,
                    "adminEventsEnabled": true,
                    "enabledEventTypes": [
                      "UPDATE_CONSENT_ERROR",
                      "CLIENT_LOGIN"
                    ],
                    "eventsEnabled": true,
                    "eventsExpiration": 15000,
                    "eventsListeners": [
                      "jboss-logging"
                    ]
                  },
                  "realmName": "d2-id-kc-realm-name",
                  "ssoAutoRedirectEnabled": false,
                  "ssoRealmEnabled": false,
                  "ssoRealmName": "openshift"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmComponent",
                "metadata": {
                  "name": "keycloakrealmcomponent-sample"
                },
                "spec": {
                  "config": {
                    "allowPasswordAuthentication": [
                      "true"
                    ],
                    "cachePolicy": [
                      "EVICT_WEEKLY"
                    ],
                    "debug": [
                      "true"
                    ],
                    "editMode": [
                      "READ_ONLY"
                    ],
                    "enabled": [
                      "true"
                    ],
                    "evictionDay": [
                      "3"
                    ],
                    "evictionHour": [
                      "5"
                    ],
                    "evictionMinute": [
                      "7"
                    ],
                    "kerberosRealm": [
                      "test-realm"
                    ],
                    "keyTab": [
                      "test-key-tab"
                    ],
                    "priority": [
                      "0"
                    ],
                    "serverPrincipal": [
                      "srv-principal-test"
                    ],
                    "updateProfileFirstLogin": [
                      "true"
                    ]
                  },
                  "name": "cr-kerb-test",
                  "providerId": "kerberos",
                  "providerType": "org.keycloak.storage.UserStorageProvider",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmGroup",
                "metadata": {
                  "name": "keycloakrealmgroup-sample"
                },
                "spec": {
                  "name": "ArgoCDAdmins",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "keycloakrealmidentityprovider-sample"
                },
                "spec": {
                  "alias": "instagram",
                  "authenticateByDefault": false,
                  "config": {
                    "clientId": "foo",
                    "clientSecret": "$secretName:secretKey",
                    "hideOnLoginPage": "true",
                    "syncMode": "IMPORT",
                    "useJwksUrl": "true"
                  },
                  "enabled": true,
                  "firstBrokerLoginFlowAlias": "first broker login",
                  "mappers": [
                    {
                      "config": {
                        "role": "role-tr",
                        "syncMode": "INHERIT"
                      },
                      "identityProviderAlias": "instagram",
                      "identityProviderMapper": "oidc-hardcoded-role-idp-mapper",
                      "name": "test3212"
                    },
                    {
                      "config": {
                        "attribute": "foo",
                        "attribute.value": "bar",
                        "syncMode": "IMPORT"
                      },
                      "identityProviderAlias": "instagram",
                      "identityProviderMapper": "hardcoded-attribute-idp-mapper",
                      "name": "test-33221"
                    }
                  ],
                  "providerId": "instagram",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRole",
                "metadata": {
                  "name": "keycloakrealmrole-sample"
                },
                "spec": {
                  "composite": true,
                  "description": "default developer role",
                  "name": "developer",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRoleBatch",
                "metadata": {
                  "name": "keycloakrealmrolebatch-sample"
                },
                "spec": {
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "roles": [
                    {
                      "composite": true,
                      "description": "default developer role",
                      "isDefault": false,
                      "name": "developer"
                    },
                    {
                      "composite": true,
                      "description": "default administrator role",
                      "isDefault": false,
                      "name": "administrator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmUser",
                "metadata": {
                  "name": "keycloakrealmuser-sample"
                },
                "spec": {
                  "attributes": {
                    "baz": "jazz",
                    "foo": "bar"
                  },
                  "email": "john.snow13@example.com",
                  "emailVerified": true,
                  "enabled": true,
                  "firstName": "John",
                  "keepResource": true,
                  "lastName": "Snow",
                  "password": "12345678",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "requiredUserActions": [
                    "UPDATE_PASSWORD"
                  ],
                  "username": "john.snow13"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloak",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloak-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloak",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloak-sample"
                },
                "spec": {
                  "secret": "keycloak-access",
                  "url": "https://keycloak.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloakRealm",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloakrealm-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloakrealm",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloakrealm-sample"
                },
                "spec": {
                  "clusterKeycloakRef": "clusterkeycloak-sample",
                  "realmName": "realm-sample"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          containerImage: docker.io/epamedp/keycloak-operator:1.19.0
          createdAt: "2023-11-15T16:04:45Z"
          description: An Operator for managing Keycloak
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-keycloak-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterKeycloakRealm is the Schema for the clusterkeycloakrealms
              API.
            displayName: Cluster Keycloak Realm
            kind: ClusterKeycloakRealm
            name: clusterkeycloakrealms.v1.edp.epam.com
            version: v1alpha1
          - description: ClusterKeycloak is the Schema for the clusterkeycloaks API.
            displayName: Cluster Keycloak
            kind: ClusterKeycloak
            name: clusterkeycloaks.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakAuthFlow is the Schema for the keycloak authentication
              flow API.
            displayName: Keycloak Auth Flow
            kind: KeycloakAuthFlow
            name: keycloakauthflows.v1.edp.epam.com
            version: v1
          - description: KeycloakClient is the Schema for the keycloak clients API.
            displayName: Keycloak Client
            kind: KeycloakClient
            name: keycloakclients.v1.edp.epam.com
            version: v1
          - description: KeycloakClientScope is the Schema for the keycloakclientscopes
              API.
            displayName: Keycloak Client Scope
            kind: KeycloakClientScope
            name: keycloakclientscopes.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmComponent is the Schema for the keycloak component
              API.
            displayName: Keycloak Realm Component
            kind: KeycloakRealmComponent
            name: keycloakrealmcomponents.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmGroup is the Schema for the keycloak group API.
            displayName: Keycloak Realm Group
            kind: KeycloakRealmGroup
            name: keycloakrealmgroups.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmIdentityProvider is the Schema for the keycloak
              realm identity provider API.
            displayName: Keycloak Realm Identity Provider
            kind: KeycloakRealmIdentityProvider
            name: keycloakrealmidentityproviders.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRoleBatch is the Schema for the keycloak roles
              API.
            displayName: Keycloak Realm Role Batch
            kind: KeycloakRealmRoleBatch
            name: keycloakrealmrolebatches.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRole is the Schema for the keycloak group API.
            displayName: Keycloak Realm Role
            kind: KeycloakRealmRole
            name: keycloakrealmroles.v1.edp.epam.com
            version: v1
          - description: KeycloakRealm is the Schema for the keycloak realms API.
            displayName: Keycloak Realm
            kind: KeycloakRealm
            name: keycloakrealms.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmUser is the Schema for the keycloak user API.
            displayName: Keycloak Realm User
            kind: KeycloakRealmUser
            name: keycloakrealmusers.v1.edp.epam.com
            version: v1
          - description: Keycloak is the Schema for the keycloaks API.
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.v1.edp.epam.com
            version: v1
        description: |
          Keycloak Operator is an operator that is responsible for establishing
          a connection to provided Keycloak Server, reconciling Keycloak entities (realms,
          roles, groups, users, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Keycloak `Master` realm, and assign a `create-realm` role.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  keycloak-access
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Keycloak` with Keycloak instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: Keycloak
              metadata:
                name: keycloak-sample
              spec:
                secret: keycloak-access             # Secret name
                url: https://keycloak.example.com   # Keycloak URL
              ```

              Wait for the `.status` field with  `status.connected: true`

          4. Create Keycloak realm and group using Custom Resources:

            ```yaml
            apiVersion: v1.edp.epam.com/v1
            kind: KeycloakRealm
            metadata:
              name: keycloakrealm-sample
            spec:
              realmName: realm-sample
              keycloakOwner: keycloak-sample   # the name of `kind: Keycloak`
              ```

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: KeycloakRealmGroup
              metadata:
                name: argocd-admins
              spec:
                name: ArgoCDAdmins
                realm: keycloakrealm-sample   # the name of `kind: KeycloakRealm`
              ```
        displayName: EDP Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - authentication
        - authorization
        - edp
        - idp
        - keycloak
        - oauth
        - oidc
        - operator
        - saml
        - sso
        links:
        - name: Edp Keycloak Operator
          url: https://github.com/epam/edp-keycloak-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/keycloak-operator:1.19.0
        version: 1.19.0
      entries:
      - name: edp-keycloak-operator.v1.19.0
        version: 1.19.0
      - name: edp-keycloak-operator.v1.18.2
        version: 1.18.2
      - name: edp-keycloak-operator.v1.17.1
        version: 1.17.1
      - name: edp-keycloak-operator.v1.13.0
        version: 1.13.0
      name: alpha
    - currentCSV: edp-keycloak-operator.v1.29.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "keycloak-sample"
                },
                "spec": {
                  "secret": "my-keycloak-secret",
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakAuthFlow",
                "metadata": {
                  "name": "keycloakauthflow-sample"
                },
                "spec": {
                  "alias": "MyBrowser",
                  "authenticationExecutions": [
                    {
                      "authenticator": "auth-cookie",
                      "priority": 0,
                      "requirement": "ALTERNATIVE"
                    },
                    {
                      "authenticator": "identity-provider-redirector",
                      "authenticatorConfig": {
                        "alias": "my-alias",
                        "config": {
                          "defaultProvider": "my-alias"
                        }
                      },
                      "priority": 1,
                      "requirement": "REQUIRED"
                    }
                  ],
                  "builtIn": false,
                  "description": "browser with idp",
                  "providerId": "basic-flow",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "topLevel": true
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-authorization-sample"
                },
                "spec": {
                  "adminFineGrainedPermissionsEnabled": true,
                  "authorization": {
                    "permissions": [
                      {
                        "decisionStrategy": "AFFIRMATIVE",
                        "description": "Resource permission",
                        "logic": "POSITIVE",
                        "name": "resource-permission",
                        "policies": [
                          "role-policy"
                        ],
                        "resources": [
                          "resource1"
                        ],
                        "type": "resource"
                      },
                      {
                        "decisionStrategy": "CONSENSUS",
                        "description": "Scope permission",
                        "logic": "POSITIVE",
                        "name": "scope-permission",
                        "policies": [
                          "role-policy"
                        ],
                        "scopes": [
                          "scope1"
                        ],
                        "type": "scope"
                      }
                    ],
                    "policies": [
                      {
                        "decisionStrategy": "AFFIRMATIVE",
                        "description": "Role policy",
                        "logic": "POSITIVE",
                        "name": "role-policy",
                        "rolePolicy": {
                          "roles": [
                            {
                              "name": "developer",
                              "required": true
                            }
                          ]
                        },
                        "type": "role"
                      },
                      {
                        "aggregatedPolicy": {
                          "policies": [
                            "policy1",
                            "policy2"
                          ]
                        },
                        "description": "Aggregate policy",
                        "name": "aggregate-policy",
                        "type": "aggregate"
                      },
                      {
                        "clientPolicy": {
                          "clients": [
                            "client1",
                            "client2"
                          ]
                        },
                        "description": "Client policy",
                        "name": "client-policy",
                        "type": "client"
                      },
                      {
                        "description": "Group policy",
                        "groupPolicy": {
                          "groups": [
                            {
                              "extendChildren": true,
                              "name": "group1"
                            }
                          ]
                        },
                        "name": "group-policy",
                        "type": "group"
                      },
                      {
                        "description": "Role policy",
                        "name": "role-policy",
                        "rolePolicy": {
                          "roles": [
                            {
                              "name": "developer",
                              "required": true
                            }
                          ]
                        },
                        "type": "role"
                      },
                      {
                        "description": "Time policy",
                        "name": "time-policy",
                        "timePolicy": {
                          "notBefore": "2021-01-01T00:00:00Z",
                          "notOnOrAfter": "2021-12-31T23:59:59Z"
                        },
                        "type": "time"
                      },
                      {
                        "description": "User policy",
                        "name": "user-policy",
                        "type": "user",
                        "userPolicy": {
                          "users": [
                            "user1",
                            "user2"
                          ]
                        }
                      }
                    ],
                    "resources": [
                      {
                        "displayName": "Resource 1",
                        "iconUri": "https://example.com/icon.png",
                        "name": "resource1",
                        "scopes": [
                          "scope1"
                        ],
                        "type": "test"
                      }
                    ],
                    "scopes": [
                      "scope1"
                    ]
                  },
                  "authorizationServicesEnabled": true,
                  "clientId": "authorization-sample",
                  "directAccess": true,
                  "permission": {
                    "scopePermissions": [
                      {
                        "name": "token-exchange",
                        "policies": [
                          "policy1"
                        ]
                      }
                    ]
                  },
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-authorization-sample:client-secret-key",
                  "serviceAccount": {
                    "enabled": true
                  },
                  "webUrl": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-sample"
                },
                "spec": {
                  "advancedProtocolMappers": true,
                  "authenticationFlowBindingOverrides": {
                    "browser": "browser",
                    "directGrant": "direct grant"
                  },
                  "clientId": "agocd",
                  "defaultClientScopes": [
                    "argocd_groups"
                  ],
                  "directAccess": true,
                  "public": false,
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-name:client-secret-key",
                  "webUrl": "https://argocd.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClientScope",
                "metadata": {
                  "name": "keycloakclientscope-sample"
                },
                "spec": {
                  "description": "Group Membership",
                  "name": "groups",
                  "protocol": "openid-connect",
                  "protocolMappers": [
                    {
                      "config": {
                        "access.token.claim": "true",
                        "claim.name": "groups",
                        "full.path": "false",
                        "id.token.claim": "true",
                        "userinfo.token.claim": "true"
                      },
                      "name": "groups",
                      "protocol": "openid-connect",
                      "protocolMapper": "oidc-group-membership-mapper"
                    }
                  ],
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "keycloakrealm-sample"
                },
                "spec": {
                  "id": "d1-id-kc-realm-name",
                  "keycloakRef": {
                    "kind": "Keycloak",
                    "name": "keycloak-sample"
                  },
                  "organizationsEnabled": true,
                  "passwordPolicy": [
                    {
                      "type": "forceExpiredPasswordChange",
                      "value": "365"
                    },
                    {
                      "type": "length",
                      "value": "8"
                    }
                  ],
                  "realmEventConfig": {
                    "adminEventsDetailsEnabled": false,
                    "adminEventsEnabled": true,
                    "enabledEventTypes": [
                      "UPDATE_CONSENT_ERROR",
                      "CLIENT_LOGIN"
                    ],
                    "eventsEnabled": true,
                    "eventsExpiration": 15000,
                    "eventsListeners": [
                      "jboss-logging"
                    ]
                  },
                  "realmName": "d2-id-kc-realm-name",
                  "userProfileConfig": {
                    "attributes": [
                      {
                        "annotations": {
                          "inputType": "text"
                        },
                        "displayName": "Test Attribute",
                        "group": "test-group",
                        "multivalued": true,
                        "name": "test-attribute",
                        "permissions": {
                          "edit": [
                            "admin"
                          ],
                          "view": [
                            "admin",
                            "user"
                          ]
                        },
                        "required": {
                          "roles": [
                            "admin"
                          ],
                          "scopes": [
                            "profile"
                          ]
                        },
                        "selector": {
                          "scopes": [
                            "profile"
                          ]
                        },
                        "validations": {
                          "email": {
                            "max-local-length": {
                              "intVal": 64
                            }
                          },
                          "local-date": {},
                          "multivalued": {
                            "max": {
                              "stringVal": "10"
                            },
                            "min": {
                              "stringVal": "1"
                            }
                          },
                          "options": {
                            "options": {
                              "sliceVal": [
                                "option1",
                                "option2"
                              ]
                            }
                          }
                        }
                      }
                    ],
                    "groups": [
                      {
                        "annotations": {
                          "groupAnnotation": "groupAnnotation"
                        },
                        "displayDescription": "Test Group",
                        "displayHeader": "Test Group",
                        "name": "test-group"
                      }
                    ],
                    "unmanagedAttributePolicy": "ENABLED"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmComponent",
                "metadata": {
                  "name": "keycloakrealmcomponent-sample"
                },
                "spec": {
                  "config": {
                    "allowPasswordAuthentication": [
                      "true"
                    ],
                    "cachePolicy": [
                      "EVICT_WEEKLY"
                    ],
                    "debug": [
                      "true"
                    ],
                    "editMode": [
                      "READ_ONLY"
                    ],
                    "enabled": [
                      "true"
                    ],
                    "evictionDay": [
                      "3"
                    ],
                    "evictionHour": [
                      "5"
                    ],
                    "evictionMinute": [
                      "7"
                    ],
                    "kerberosRealm": [
                      "test-realm"
                    ],
                    "keyTab": [
                      "test-key-tab"
                    ],
                    "priority": [
                      "0"
                    ],
                    "serverPrincipal": [
                      "srv-principal-test"
                    ],
                    "updateProfileFirstLogin": [
                      "true"
                    ]
                  },
                  "name": "cr-kerb-test",
                  "providerId": "kerberos",
                  "providerType": "org.keycloak.storage.UserStorageProvider",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmGroup",
                "metadata": {
                  "name": "keycloakrealmgroup-sample"
                },
                "spec": {
                  "name": "ArgoCDAdmins",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "test-org-idp",
                  "namespace": "default"
                },
                "spec": {
                  "alias": "test-org-idp",
                  "config": {
                    "clientId": "test-org-client-id",
                    "clientSecret": "test-org-client-secret",
                    "kc.org.broker.redirect.mode.email-matches": "true",
                    "kc.org.domain": "example.com"
                  },
                  "enabled": true,
                  "providerId": "github",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "test-org-realm"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "keycloakrealmidentityprovider-sample"
                },
                "spec": {
                  "alias": "github",
                  "authenticateByDefault": false,
                  "config": {
                    "clientId": "foo",
                    "clientSecret": "$secretName:secretKey",
                    "syncMode": "IMPORT",
                    "useJwksUrl": "true"
                  },
                  "enabled": true,
                  "firstBrokerLoginFlowAlias": "first broker login",
                  "mappers": [
                    {
                      "config": {
                        "role": "role-tr",
                        "syncMode": "INHERIT"
                      },
                      "identityProviderAlias": "github",
                      "identityProviderMapper": "oidc-hardcoded-role-idp-mapper",
                      "name": "test3212"
                    },
                    {
                      "config": {
                        "attribute": "foo",
                        "attribute.value": "bar",
                        "syncMode": "IMPORT"
                      },
                      "identityProviderAlias": "github",
                      "identityProviderMapper": "hardcoded-attribute-idp-mapper",
                      "name": "test-33221"
                    }
                  ],
                  "providerId": "github",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRole",
                "metadata": {
                  "name": "keycloakrealmrole-sample"
                },
                "spec": {
                  "composite": true,
                  "description": "default developer role",
                  "name": "developer",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRoleBatch",
                "metadata": {
                  "name": "keycloakrealmrolebatch-sample"
                },
                "spec": {
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "roles": [
                    {
                      "composite": true,
                      "description": "default developer role",
                      "isDefault": false,
                      "name": "developer"
                    },
                    {
                      "composite": true,
                      "description": "default administrator role",
                      "isDefault": false,
                      "name": "administrator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmUser",
                "metadata": {
                  "name": "keycloakrealmuser-sample"
                },
                "spec": {
                  "attributes": {
                    "baz": "jazz",
                    "foo": "bar"
                  },
                  "email": "john.snow13@example.com",
                  "emailVerified": true,
                  "enabled": true,
                  "firstName": "John",
                  "keepResource": true,
                  "lastName": "Snow",
                  "password": "12345678",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "requiredUserActions": [
                    "UPDATE_PASSWORD"
                  ],
                  "username": "john.snow13"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloak",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloak-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloak",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloak-sample"
                },
                "spec": {
                  "secret": "keycloak-access",
                  "url": "https://keycloak.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloakRealm",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloakrealm-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloakrealm",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloakrealm-sample"
                },
                "spec": {
                  "authenticationFlows": {
                    "browserFlow": "browserFlow-sample"
                  },
                  "clusterKeycloakRef": "clusterkeycloak-sample",
                  "organizationsEnabled": false,
                  "realmName": "realm-sample"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "KeycloakOrganization",
                "metadata": {
                  "name": "test-keycloak-organization",
                  "namespace": "default"
                },
                "spec": {
                  "alias": "test-org",
                  "attributes": {
                    "department": [
                      "engineering",
                      "qa"
                    ],
                    "location": [
                      "us-east"
                    ]
                  },
                  "description": "Test organization",
                  "domains": [
                    "example.com",
                    "test.com"
                  ],
                  "identityProviders": [
                    {
                      "alias": "test-org-idp"
                    }
                  ],
                  "name": "Test Organization",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "test-org-realm"
                  },
                  "redirectUrl": "https://example.com/redirect"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          containerImage: docker.io/epamedp/keycloak-operator:1.29.0
          createdAt: "2025-09-26T12:16:55Z"
          description: An Operator for managing Keycloak
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/epam/edp-keycloak-operator
          support: KubeRocketCI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterKeycloakRealm is the Schema for the clusterkeycloakrealms
              API.
            displayName: Cluster Keycloak Realm
            kind: ClusterKeycloakRealm
            name: clusterkeycloakrealms.v1.edp.epam.com
            version: v1alpha1
          - description: ClusterKeycloak is the Schema for the clusterkeycloaks API.
            displayName: Cluster Keycloak
            kind: ClusterKeycloak
            name: clusterkeycloaks.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakAuthFlow is the Schema for the keycloak authentication
              flow API.
            displayName: Keycloak Auth Flow
            kind: KeycloakAuthFlow
            name: keycloakauthflows.v1.edp.epam.com
            version: v1
          - description: KeycloakClient is the Schema for the keycloak clients API.
            displayName: Keycloak Client
            kind: KeycloakClient
            name: keycloakclients.v1.edp.epam.com
            version: v1
          - description: KeycloakClientScope is the Schema for the keycloakclientscopes
              API.
            displayName: Keycloak Client Scope
            kind: KeycloakClientScope
            name: keycloakclientscopes.v1.edp.epam.com
            version: v1
          - description: KeycloakOrganization is the Schema for the organizations
              API.
            displayName: Keycloak Organization
            kind: KeycloakOrganization
            name: keycloakorganizations.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakRealmComponent is the Schema for the keycloak component
              API.
            displayName: Keycloak Realm Component
            kind: KeycloakRealmComponent
            name: keycloakrealmcomponents.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmGroup is the Schema for the keycloak group API.
            displayName: Keycloak Realm Group
            kind: KeycloakRealmGroup
            name: keycloakrealmgroups.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmIdentityProvider is the Schema for the keycloak
              realm identity provider API.
            displayName: Keycloak Realm Identity Provider
            kind: KeycloakRealmIdentityProvider
            name: keycloakrealmidentityproviders.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRoleBatch is the Schema for the keycloak roles
              API.
            displayName: Keycloak Realm Role Batch
            kind: KeycloakRealmRoleBatch
            name: keycloakrealmrolebatches.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRole is the Schema for the keycloak group API.
            displayName: Keycloak Realm Role
            kind: KeycloakRealmRole
            name: keycloakrealmroles.v1.edp.epam.com
            version: v1
          - description: KeycloakRealm is the Schema for the keycloak realms API.
            displayName: Keycloak Realm
            kind: KeycloakRealm
            name: keycloakrealms.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmUser is the Schema for the keycloak user API.
            displayName: Keycloak Realm User
            kind: KeycloakRealmUser
            name: keycloakrealmusers.v1.edp.epam.com
            version: v1
          - description: Keycloak is the Schema for the keycloaks API.
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.v1.edp.epam.com
            version: v1
        description: |
          Keycloak Operator is an operator that is responsible for establishing
          a connection to provided Keycloak Server, reconciling Keycloak entities (realms,
          roles, groups, users, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Keycloak `Master` realm, and assign a `create-realm` role.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  keycloak-access
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Keycloak` with Keycloak instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: Keycloak
              metadata:
                name: keycloak-sample
              spec:
                secret: keycloak-access             # Secret name
                url: https://keycloak.example.com   # Keycloak URL
              ```

              Wait for the `.status` field with  `status.connected: true`

          4. Create Keycloak realm and group using Custom Resources:

            ```yaml
            apiVersion: v1.edp.epam.com/v1
            kind: KeycloakRealm
            metadata:
              name: keycloakrealm-sample
            spec:
              realmName: realm-sample
              keycloakOwner: keycloak-sample   # the name of `kind: Keycloak`
              ```

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: KeycloakRealmGroup
              metadata:
                name: argocd-admins
              spec:
                name: ArgoCDAdmins
                realm: keycloakrealm-sample   # the name of `kind: KeycloakRealm`
              ```
        displayName: EDP Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - authentication
        - authorization
        - edp
        - idp
        - keycloak
        - kuberocketci
        - oauth
        - oidc
        - operator
        - saml
        - sso
        links:
        - name: Edp Keycloak Operator
          url: https://github.com/epam/edp-keycloak-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: KubeRocketCI
          url: https://docs.kuberocketci.io/
        relatedImages:
        - docker.io/epamedp/keycloak-operator:1.29.0
        version: 1.29.0
      entries:
      - name: edp-keycloak-operator.v1.29.0
        version: 1.29.0
      - name: edp-keycloak-operator.v1.28.0
        version: 1.28.0
      - name: edp-keycloak-operator.v1.27.1
        version: 1.27.1
      - name: edp-keycloak-operator.v1.26.0
        version: 1.26.0
      - name: edp-keycloak-operator.v1.25.0
        version: 1.25.0
      - name: edp-keycloak-operator.v1.24.0
        version: 1.24.0
      - name: edp-keycloak-operator.v1.23.0
        version: 1.23.0
      name: stable
    defaultChannel: stable
    packageName: edp-keycloak-operator
    provider:
      name: KubeRocketCI
      url: https://docs.kuberocketci.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: minhokim
      provider-url: ""
    name: node-discovery-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-discovery-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.wunicorns.com/v1alpha1",
                "kind": "NodeDiscovery",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "node-discovery-operator",
                    "app.kubernetes.io/instance": "nodediscovery-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodediscovery",
                    "app.kubernetes.io/part-of": "node-discovery-operator"
                  },
                  "name": "nodediscovery-sample"
                },
                "spec": {
                  "containers": [
                    {
                      "args": [
                        "tail",
                        "-f",
                        "/dev/null"
                      ],
                      "image": "alicek106/stress",
                      "imagePullPolicy": "IfNotPresent",
                      "name": "stress",
                      "resources": {
                        "limits": {
                          "cpu": "100m",
                          "memory": "1000Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "1000Mi"
                        }
                      },
                      "terminationGracePeriodSeconds": 10
                    }
                  ],
                  "cpus": "0",
                  "enabled": "true",
                  "mems": "0",
                  "nodeName": "worker3"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-18T01:32:51Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeDiscovery is the Schema for the nodediscoveries API
            displayName: Node Discovery
            kind: NodeDiscovery
            name: nodediscoveries.cache.wunicorns.com
            version: v1alpha1
        description: find cgroup infor and set value.
        displayName: node-discovery-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cgroup
        - cpulist
        - mems
        - cpus
        - numa
        links:
        - name: Node Discovery Operator
          url: https://node-discovery-operator.domain
        maintainers:
        - email: wunicorns1436@gmail.com
          name: minhokim
        maturity: alpha
        provider:
          name: minhokim
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/caspell/cgroup-discovery@sha256:b437dcde0107124e43a4e2fb8026ec191f44b5e761bedc2c295f47ff489cf343
        version: 0.0.1
      entries:
      - name: node-discovery-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: node-discovery-operator
    provider:
      name: minhokim
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HashiCorp
      provider-url: https://www.hashicorp.com/
    name: hcp-terraform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hcp-terraform-operator.v2.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "AgentPool",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "agentTokens": [
                    {
                      "name": "token"
                    }
                  ],
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Module",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "module": {
                    "source": "redeux/terraform-cloud-agent/kubernetes",
                    "version": "1.0.1"
                  },
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  },
                  "workspace": {
                    "id": "WORKSPACE_ID"
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Project",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Workspace",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.connect.redhat.com/hashicorp/hcp-terraform-operator@sha256:8885a6a5eae092117a670ec5541a3e4a24c6b23807fd17b0446d91b01eb6e421
          createdAt: "2025-05-28T13:58:52Z"
          description: HCP Terraform Operator for Kubernetes allows managing HCP Terraform
            / Terraform Enterprise resources via Kubernetes Custom Resources.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Valid HCP Terraform / Terraform
            Enterprise subscription is required
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hashicorp/hcp-terraform-operator
          support: HashiCorp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              AgentPool manages HCP Terraform Agent Pools, HCP Terraform Agent Tokens and can perform HCP Terraform Agent scaling.
              More infromation:
                - https://developer.hashicorp.com/terraform/cloud-docs/agents/agent-pools
                - https://developer.hashicorp.com/terraform/cloud-docs/users-teams-organizations/api-tokens#agent-api-tokens
                - https://developer.hashicorp.com/terraform/cloud-docs/agents
            displayName: Agent Pool
            kind: AgentPool
            name: agentpools.app.terraform.io
            version: v1alpha2
          - description: |-
              Module implements API-driven Run Workflows.
              More information:
                - https://developer.hashicorp.com/terraform/cloud-docs/run/api
            displayName: Module
            kind: Module
            name: modules.app.terraform.io
            version: v1alpha2
          - description: |-
              Project manages HCP Terraform Projects.
              More information:
              - https://developer.hashicorp.com/terraform/cloud-docs/projects/manage
            displayName: Project
            kind: Project
            name: projects.app.terraform.io
            version: v1alpha2
          - description: |-
              Workspace manages HCP Terraform Workspaces.
              More information:
                - https://developer.hashicorp.com/terraform/cloud-docs/workspaces
            displayName: Workspace
            kind: Workspace
            name: workspaces.app.terraform.io
            version: v1alpha2
        description: HCP Terraform Operator for Kubernetes allows managing HCP Terraform
          / Terraform Enterprise resources via Kubernetes Custom Resources.
        displayName: HCP Terraform Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - terraform
        links:
        - name: HCP Terraform Operator
          url: https://github.com/hashicorp/hcp-terraform-operator
        maintainers:
        - email: hcp-tf-operator@hashicorp.com
          name: HashiCorp
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: HashiCorp
          url: https://www.hashicorp.com/
        relatedImages:
        - registry.connect.redhat.com/hashicorp/hcp-terraform-operator@sha256:8885a6a5eae092117a670ec5541a3e4a24c6b23807fd17b0446d91b01eb6e421
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        version: 2.9.2
      entries:
      - name: hcp-terraform-operator.v2.9.2
        version: 2.9.2
      - name: hcp-terraform-operator.v2.9.1
        version: 2.9.1
      - name: hcp-terraform-operator.v2.9.0
        version: 2.9.0
      - name: hcp-terraform-operator.v2.8.1
        version: 2.8.1
      - name: hcp-terraform-operator.v2.8.0
        version: 2.8.0
      name: alpha
    defaultChannel: alpha
    packageName: hcp-terraform-operator
    provider:
      name: HashiCorp
      url: https://www.hashicorp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Appian Corporation
      provider-url: https://appian.com/
    name: appian-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appian-operator.v0.184.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: AI/Machine Learning, Application Runtime, Big Data, Integration
            & Delivery, Modernization & Migration
          createdAt: "2025-08-07T15:50:26Z"
          description: The Appian operator facilitates managing instances of Appian
            on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appian-operator
          operators.openshift.io/valid-subscription: '["Appian License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Appian Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Appian
            name: appians.crd.k8s.appian.com
            version: v1beta1
        description: The Appian operator facilitates managing instances of Appian
          on Kubernetes.
        displayName: Appian operator
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - Appian
        links:
        - name: Appian
          url: https://appian.com/
        - name: Documentation
          url: https://docs.appian.com/suite/help/latest/k8s
        maintainers:
        - email: support@appian.com
          name: Appian Corporation
        maturity: planning
        provider:
          name: Appian Corporation
          url: https://appian.com/
        relatedImages:
        - 937620572175.dkr.ecr.us-east-1.amazonaws.com/appian/appian-operator@sha256:26c2ba27f9360226f34f289109cb0d10e9bf44777146ccfd5f367adf2367a400
        version: 0.184.0
      entries:
      - name: appian-operator.v0.184.0
        version: 0.184.0
      - name: appian-operator.v0.171.0
        version: 0.171.0
      name: alpha
    defaultChannel: alpha
    packageName: appian-operator
    provider:
      name: Appian Corporation
      url: https://appian.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: aiu-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aiu-operator.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "labels": {
                    "operator-version": "2.1.0-dev"
                  },
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "aiu": {
                    "strategy": "single"
                  },
                  "devicePlugin": {
                    "aiuSelector": {
                      "deviceID": "06a7",
                      "vendor": "1014"
                    },
                    "image": "aiu-device-plugin",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "deviceType": "vfio-pci",
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation"
                  ],
                  "featureDiscovery": {
                    "image": "aiu-feature-discovery",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "isRdma": false,
                  "loglevel": "info",
                  "metricsExporter": {
                    "enabled": true,
                    "image": "aiu-exporter",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "numVfs": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
          createdAt: "2024-10-07T13:07:51Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 2.1.1
      entries:
      - name: aiu-operator.v2.1.1
        version: 2.1.1
      name: candidate-2.1
    - currentCSV: aiu-operator.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "labels": {
                    "operator-version": "2.1.0-dev"
                  },
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "aiu": {
                    "strategy": "single"
                  },
                  "devicePlugin": {
                    "aiuSelector": {
                      "deviceID": "06a7",
                      "vendor": "1014"
                    },
                    "image": "aiu-device-plugin",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "deviceType": "vfio-pci",
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation"
                  ],
                  "featureDiscovery": {
                    "image": "aiu-feature-discovery",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "isRdma": false,
                  "loglevel": "info",
                  "metricsExporter": {
                    "enabled": true,
                    "image": "aiu-exporter",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "numVfs": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
          createdAt: "2024-10-07T13:07:51Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
        version: 2.1.1
      entries:
      - name: aiu-operator.v2.1.1
        version: 2.1.1
      name: fast-2.1
    - currentCSV: aiu-operator.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "labels": {
                    "operator-version": "2.1.0-dev"
                  },
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "aiu": {
                    "strategy": "single"
                  },
                  "devicePlugin": {
                    "aiuSelector": {
                      "deviceID": "06a7",
                      "vendor": "1014"
                    },
                    "image": "aiu-device-plugin",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "deviceType": "vfio-pci",
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation"
                  ],
                  "featureDiscovery": {
                    "image": "aiu-feature-discovery",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "isRdma": false,
                  "loglevel": "info",
                  "metricsExporter": {
                    "enabled": true,
                    "image": "aiu-exporter",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "numVfs": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
          createdAt: "2024-10-07T13:07:51Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 2.1.1
      entries:
      - name: aiu-operator.v2.1.1
        version: 2.1.1
      name: stable-2.1
    - currentCSV: aiu-operator.v2.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "labels": {
                    "operator-version": "2.3.3"
                  },
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "aiu-device-plugin",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.1"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "featureDiscovery": {
                    "image": "aiu-feature-discovery",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "aiu-exporter",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "aiu-webhook-validator",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "scheduler": {
                    "image": "aiu-kube-scheduler-ocp-4.16",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.2"
                  }
                }
              },
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuNodeState",
                "metadata": {
                  "labels": {
                    "operator-version": "2.3.3"
                  },
                  "name": "aiunodestate"
                },
                "spec": {
                  "nodeName": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2025_03/aiu-operator@sha256:2ea8268e94d44581b3490cd4bbbd5525359073129bae42c7f007c05727afe1c3
          createdAt: "2025-04-03T01:19:38Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["aiunodestates.aiu.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu/release_2025_03/aiu-operator@sha256:2ea8268e94d44581b3490cd4bbbd5525359073129bae42c7f007c05727afe1c3
        version: 2.3.3
      entries:
      - name: aiu-operator.v2.3.3
        version: 2.3.3
      name: stable-v2.3
    - currentCSV: aiu-operator.v2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "aiuFilter": ".",
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "release_2025_05"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "aiu-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "e2e_stable",
                      "repository": "icr.io/ibmaiu_internal/x86_64/dd2",
                      "version": "release_2025_05"
                    },
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "aiu-exporter",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "aiu-webhook-validator",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "scheduler": {
                    "image": "aiu-kube-scheduler-ocp-4.16",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  }
                }
              },
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2025_05/aiu-operator@sha256:9796ab1c39d1dd55f8aa79c9f585be0f988b5026745184bdc890dfd1846909df
          createdAt: "2025-05-28T05:53:45Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["aiunodestates.aiu.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu/release_2025_05/aiu-operator@sha256:9796ab1c39d1dd55f8aa79c9f585be0f988b5026745184bdc890dfd1846909df
        - icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453
        version: 2.4.0
      entries:
      - name: aiu-operator.v2.4.0
        version: 2.4.0
      name: stable-v2.4
    defaultChannel: stable-v2.4
    packageName: aiu-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lifecycle-agent
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lifecycle-agent.v4.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "ImageBasedUpgrade",
                "metadata": {
                  "name": "upgrade"
                },
                "spec": {
                  "autoRollbackOnFailure": {},
                  "extraManifests": [
                    {
                      "name": "sno-extramanifests",
                      "namespace": "openshift-lifecycle-agent"
                    }
                  ],
                  "oadpContent": [
                    {
                      "name": "oadp-cm-sno-backup",
                      "namespace": "openshift-adp"
                    }
                  ],
                  "seedImageRef": {
                    "image": "quay.io/xyz",
                    "version": "4.16.0"
                  },
                  "stage": "Idle"
                }
              },
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "SeedGenerator",
                "metadata": {
                  "name": "seedimage"
                },
                "spec": {
                  "seedImage": "quay.io/xyz"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:6c603c89401823086d7a02d7ad221da44d8c84000ba63dd30675ae0cbcacc95f
          description: The Lifecycle Agent for OpenShift provides local lifecycle
            management services for Single Node Openshift (SNO) clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.3'
          operatorframework.io/suggested-namespace: openshift-lifecycle-agent
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-lifecycle-agent",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/lifecycle-agent
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ImageBasedUpgrade is the Schema for the ImageBasedUpgrades
              API
            displayName: Image-based Cluster Upgrade
            kind: ImageBasedUpgrade
            name: imagebasedupgrades.lca.openshift.io
            version: v1
          - description: SeedGenerator is the Schema for the seedgenerators API
            displayName: Seed Generator
            kind: SeedGenerator
            name: seedgenerators.lca.openshift.io
            version: v1
        description: "# Lifecycle Agent for OpenShift\nThe Lifecycle Agent for OpenShift
          provides local lifecycle management services \nfor Single Node Openshift
          (SNO) clusters.\n\n## Where to find more information\nYou can find additional
          guidance in the [agent repository](https://github.com/openshift-kni/lifecycle-agent).\n"
        displayName: Lifecycle Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - upgrade
        links:
        - name: Lifecycle Agent
          url: https://lifecycle-agent.domain
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift4/recert-rhel9@sha256:c91def5d3da4e777d2555c354e39bf021d9f499993f86614969dee5ecc00fc49
        - registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:6c603c89401823086d7a02d7ad221da44d8c84000ba63dd30675ae0cbcacc95f
        version: 4.16.3
      entries:
      - name: lifecycle-agent.v4.16.3
        version: 4.16.3
      - name: lifecycle-agent.v4.16.2
        version: 4.16.2
      - name: lifecycle-agent.v4.16.1
        version: 4.16.1
      - name: lifecycle-agent.v4.16.0
        version: 4.16.0
      name: "4.16"
    - currentCSV: lifecycle-agent.v4.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "ImageBasedUpgrade",
                "metadata": {
                  "name": "upgrade"
                },
                "spec": {
                  "autoRollbackOnFailure": {},
                  "extraManifests": [
                    {
                      "name": "sno-extramanifests",
                      "namespace": "openshift-lifecycle-agent"
                    }
                  ],
                  "oadpContent": [
                    {
                      "name": "oadp-cm-sno-backup",
                      "namespace": "openshift-adp"
                    }
                  ],
                  "seedImageRef": {
                    "image": "quay.io/xyz",
                    "version": "4.16.0"
                  },
                  "stage": "Idle"
                }
              },
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "SeedGenerator",
                "metadata": {
                  "name": "seedimage"
                },
                "spec": {
                  "seedImage": "quay.io/xyz"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:6c603c89401823086d7a02d7ad221da44d8c84000ba63dd30675ae0cbcacc95f
          description: The Lifecycle Agent for OpenShift provides local lifecycle
            management services for Single Node Openshift (SNO) clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.3'
          operatorframework.io/suggested-namespace: openshift-lifecycle-agent
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-lifecycle-agent",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/lifecycle-agent
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ImageBasedUpgrade is the Schema for the ImageBasedUpgrades
              API
            displayName: Image-based Cluster Upgrade
            kind: ImageBasedUpgrade
            name: imagebasedupgrades.lca.openshift.io
            version: v1
          - description: SeedGenerator is the Schema for the seedgenerators API
            displayName: Seed Generator
            kind: SeedGenerator
            name: seedgenerators.lca.openshift.io
            version: v1
        description: "# Lifecycle Agent for OpenShift\nThe Lifecycle Agent for OpenShift
          provides local lifecycle management services \nfor Single Node Openshift
          (SNO) clusters.\n\n## Where to find more information\nYou can find additional
          guidance in the [agent repository](https://github.com/openshift-kni/lifecycle-agent).\n"
        displayName: Lifecycle Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - upgrade
        links:
        - name: Lifecycle Agent
          url: https://lifecycle-agent.domain
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:6c603c89401823086d7a02d7ad221da44d8c84000ba63dd30675ae0cbcacc95f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift4/recert-rhel9@sha256:c91def5d3da4e777d2555c354e39bf021d9f499993f86614969dee5ecc00fc49
        version: 4.16.3
      entries:
      - name: lifecycle-agent.v4.16.3
        version: 4.16.3
      - name: lifecycle-agent.v4.16.2
        version: 4.16.2
      - name: lifecycle-agent.v4.16.1
        version: 4.16.1
      - name: lifecycle-agent.v4.16.0
        version: 4.16.0
      name: stable
    defaultChannel: stable
    packageName: lifecycle-agent
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-opensearchservice-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-opensearchservice-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opensearchservice.services.k8s.aws/v1alpha1",
                "kind": "Domain",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/opensearchservice-controller:1.1.1
          createdAt: "2025-09-19T18:05:50Z"
          description: AWS OpenSearch Service controller is a service controller for
            managing OpenSearch Service resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Domain represents the state of an AWS opensearchservice Domain
              resource.
            displayName: Domain
            kind: Domain
            name: domains.opensearchservice.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage OpenSearch Service resources in AWS from within your Kubernetes cluster.

          **About Amazon OpenSearch Service**

          Amazon OpenSearch Service (successor to Amazon Elasticsearch Service) is a managed service that makes it easy to deploy, operate, and scale OpenSearch clusters in the AWS Cloud. Amazon OpenSearch Service supports OpenSearch and legacy Elasticsearch OSS. When you create a cluster, you have the option of which search engine to use. OpenSearch Service offers broad compatibility with Elasticsearch OSS 7.10, the final open source version of the software. For information about what changed with the recent service renaming, and the actions you might need to take, see [Amazon OpenSearch Service - Summary of changes](https://docs.aws.amazon.com/opensearch-service/latest/developerguide/rename.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon OpenSearch Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opensearchservice
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon OpenSearch Service Developer Resources
          url: https://aws.amazon.com/opensearch-service/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: opensearch service maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/opensearchservice-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-opensearchservice-controller.v1.1.1
        version: 1.1.1
      - name: ack-opensearchservice-controller.v1.1.0
        version: 1.1.0
      - name: ack-opensearchservice-controller.v1.0.15
        version: 1.0.15
      - name: ack-opensearchservice-controller.v1.0.12
        version: 1.0.12
      - name: ack-opensearchservice-controller.v1.0.10
        version: 1.0.10
      - name: ack-opensearchservice-controller.v1.0.9
        version: 1.0.9
      - name: ack-opensearchservice-controller.v1.0.8
        version: 1.0.8
      - name: ack-opensearchservice-controller.v1.0.7
        version: 1.0.7
      - name: ack-opensearchservice-controller.v1.0.6
        version: 1.0.6
      - name: ack-opensearchservice-controller.v1.0.5
        version: 1.0.5
      - name: ack-opensearchservice-controller.v1.0.4
        version: 1.0.4
      - name: ack-opensearchservice-controller.v1.0.2
        version: 1.0.2
      - name: ack-opensearchservice-controller.v1.0.1
        version: 1.0.1
      - name: ack-opensearchservice-controller.v1.0.0
        version: 1.0.0
      - name: ack-opensearchservice-controller.v0.0.28
        version: 0.0.28
      - name: ack-opensearchservice-controller.v0.0.27
        version: 0.0.27
      - name: ack-opensearchservice-controller.v0.0.26
        version: 0.0.26
      - name: ack-opensearchservice-controller.v0.0.24
        version: 0.0.24
      - name: ack-opensearchservice-controller.v0.0.23
        version: 0.0.23
      - name: ack-opensearchservice-controller.v0.0.19
        version: 0.0.19
      - name: ack-opensearchservice-controller.v0.0.17
        version: 0.0.17
      - name: ack-opensearchservice-controller.v0.0.15
        version: 0.0.15
      - name: ack-opensearchservice-controller.v0.0.14
        version: 0.0.14
      - name: ack-opensearchservice-controller.v0.0.13
        version: 0.0.13
      - name: ack-opensearchservice-controller.v0.0.12
        version: 0.0.12
      - name: ack-opensearchservice-controller.v0.0.11
        version: 0.0.11
      - name: ack-opensearchservice-controller.v0.0.10
        version: 0.0.10
      - name: ack-opensearchservice-controller.v0.0.9
        version: 0.0.9
      - name: ack-opensearchservice-controller.v0.0.8
        version: 0.0.8
      - name: ack-opensearchservice-controller.v0.0.7
        version: 0.0.7
      - name: ack-opensearchservice-controller.v0.0.6
        version: 0.0.6
      - name: ack-opensearchservice-controller.v0.0.5
        version: 0.0.5
      - name: ack-opensearchservice-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-opensearchservice-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-migration
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-migration.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Migration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {
                  "auth": {
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configuration": {
                    "port": 8080
                  },
                  "database": {
                    "image": {
                      "name": "postgresql-ubi-image",
                      "tag": "14.2.0-3747-32"
                    }
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "esinfo": {
                    "batchSize": 10000,
                    "inputIndex": "wireless_stats"
                  },
                  "image": {
                    "remoteRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "systemMonitor": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:36:21Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Migration
            name: migrations.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Migration Operator
        displayName: Keysight WAP Migration Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Migration
          url: https://keysight-wap-migration.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-migration@sha256:118b466f1c59832a3620fa9b153507cc66ba38ee08b337b3a1a4225b158bebc5
        version: 0.1.21
      entries:
      - name: keysight-wap-migration.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-migration
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-xtradb-cluster-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator-certified.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator-certified.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator-certified.v1.17.0
        version: 1.17.0
      - name: percona-xtradb-cluster-operator-certified.v1.16.1
        version: 1.16.1
      - name: percona-xtradb-cluster-operator-certified.v1.16.0
        version: 1.16.0
      - name: percona-xtradb-cluster-operator-certified.v1.15.1
        version: 1.15.1
      - name: percona-xtradb-cluster-operator-certified.v1.15.0
        version: 1.15.0
      name: stable
    - currentCSV: percona-xtradb-cluster-operator-certified.v1.18.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        version: 1.18.0-cw
      entries:
      - name: percona-xtradb-cluster-operator-certified.v1.18.0-cw
        version: 1.18.0-cw
      - name: percona-xtradb-cluster-operator-certified.v1.17.0-cw
        version: 1.17.0-cw
      - name: percona-xtradb-cluster-operator-certified.v1.16.1-cw
        version: 1.16.1-cw
      - name: percona-xtradb-cluster-operator-certified.v1.16.0-cw
        version: 1.16.0-cw
      - name: percona-xtradb-cluster-operator-certified.v1.15.0-cw
        version: 1.15.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator-certified
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-storage-odf-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-storage-odf-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:962ca15393dd2c48f8a7189c1801b43e9fd1cd073a1971286d485f4b96060426
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.6.0'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:962ca15393dd2c48f8a7189c1801b43e9fd1cd073a1971286d485f4b96060426
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:2f785096aaa63b042e415aa0b64ace2a085cb7c219853f089d803bc37d8295ce
        version: 1.6.0
      entries:
      - name: ibm-storage-odf-operator.v1.6.0
        version: 1.6.0
      name: stable-v1.6
    - currentCSV: ibm-storage-odf-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:2dd8608645b13f81330466e25664b132d46af480a3ab2b0dd48fb1feba38870a
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.7.1'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:2dd8608645b13f81330466e25664b132d46af480a3ab2b0dd48fb1feba38870a
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:24eae9d1502ce98bb024d3eba154ddec07fdac0d78ecad0c3a64b55f5b3600f7
        version: 1.7.1
      entries:
      - name: ibm-storage-odf-operator.v1.7.1
        version: 1.7.1
      name: stable-v1.7
    - currentCSV: ibm-storage-odf-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:7432340ac156e620240eea73a19321a871fa89cc513c340a91d97d3f3fd4ded2
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.8.0'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:7432340ac156e620240eea73a19321a871fa89cc513c340a91d97d3f3fd4ded2
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:3344c7e987674fb98173bbc6ae1b189613e3ccefb468f1af9866734ddbf0408f
        version: 1.8.0
      entries:
      - name: ibm-storage-odf-operator.v1.8.0
        version: 1.8.0
      name: stable-v1.8
    defaultChannel: stable-v1.8
    packageName: ibm-storage-odf-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-ingress
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-ingress.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Ingress",
                "metadata": {
                  "name": "ingress-sample"
                },
                "spec": {
                  "admissionWebhooks": {
                    "patch": {
                      "image": {
                        "repository": "docker.io/jettech/kube-webhook-certgen"
                      }
                    }
                  },
                  "configMapMerger": {
                    "registry": "prod.harbor.keysight.digital/loadcore",
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "environment": "openshift",
                  "ingress-nginx": {
                    "controller": {
                      "config": {
                        "global-auth-signin": "https://$host/eula/static/",
                        "global-auth-url": "http://kcos-framework-v1-kcos-eula.kcos-framework.svc.cluster.local/v1/check"
                      },
                      "daemonset": {
                        "useHostPort": false
                      },
                      "hostNetwork": false,
                      "image": {
                        "registry": "prod.harbor.keysight.digital/loadcore"
                      },
                      "ingressClassResource": {
                        "default": true
                      },
                      "nodeSelector": "",
                      "service": {
                        "nodePorts": {
                          "http": 80,
                          "https": 443
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "defaultBackend": {
                      "image": {
                        "repository": "k8s.gcr.io/defaultbackend-amd64"
                      }
                    }
                  },
                  "jobDelWebhook": {
                    "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:28:18Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ingress
            name: ingresses.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Ingress Operator
        displayName: Keysight KCOS Ingress Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Ingress
          url: https://keysight-kcos-ingress.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-ingress@sha256:c6f6f8b44276efd8f5ffaa099b393e7e97d9cfa627952c0d049b98d92dc3ab59
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        version: 0.1.21
      entries:
      - name: keysight-kcos-ingress.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-ingress
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sns-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sns-controller.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "PlatformApplication",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "PlatformEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sns-controller:1.2.2
          createdAt: "2025-09-30T21:19:34Z"
          description: AWS SNS controller is a service controller for managing SNS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PlatformApplication represents the state of an AWS sns PlatformApplication
              resource.
            displayName: PlatformApplication
            kind: PlatformApplication
            name: platformapplications.sns.services.k8s.aws
            version: v1alpha1
          - description: PlatformEndpoint represents the state of an AWS sns PlatformEndpoint
              resource.
            displayName: PlatformEndpoint
            kind: PlatformEndpoint
            name: platformendpoints.sns.services.k8s.aws
            version: v1alpha1
          - description: Subscription represents the state of an AWS sns Subscription
              resource.
            displayName: Subscription
            kind: Subscription
            name: subscriptions.sns.services.k8s.aws
            version: v1alpha1
          - description: Topic represents the state of an AWS sns Topic resource.
            displayName: Topic
            kind: Topic
            name: topics.sns.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Simple Notification Service (SNS) resources in AWS from within your Kubernetes cluster.

          **About Amazon SNS**

          Amazon Simple Notification Service (Amazon SNS) is a managed service that provides message delivery from publishers to subscribers (also known as producers and consumers). Publishers communicate asynchronously with subscribers by sending messages to a topic, which is a logical access point and communication channel. Clients can subscribe to the SNS topic and receive published messages using a supported endpoint type, such as Amazon Kinesis Data Firehose, Amazon SQS, AWS Lambda, HTTP, email, mobile push notifications, and mobile text messages (SMS).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SNS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sns
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SNS Developer Resources
          url: https://aws.amazon.com/sns/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sns maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sns-controller:1.2.2
        version: 1.2.2
      entries:
      - name: ack-sns-controller.v1.2.2
        version: 1.2.2
      - name: ack-sns-controller.v1.2.1
        version: 1.2.1
      - name: ack-sns-controller.v1.2.0
        version: 1.2.0
      - name: ack-sns-controller.v1.1.16
        version: 1.1.16
      - name: ack-sns-controller.v1.1.15
        version: 1.1.15
      - name: ack-sns-controller.v1.1.14
        version: 1.1.14
      - name: ack-sns-controller.v1.1.13
        version: 1.1.13
      - name: ack-sns-controller.v1.1.12
        version: 1.1.12
      - name: ack-sns-controller.v1.1.11
        version: 1.1.11
      - name: ack-sns-controller.v1.1.10
        version: 1.1.10
      - name: ack-sns-controller.v1.1.9
        version: 1.1.9
      - name: ack-sns-controller.v1.1.8
        version: 1.1.8
      - name: ack-sns-controller.v1.1.7
        version: 1.1.7
      - name: ack-sns-controller.v1.1.6
        version: 1.1.6
      - name: ack-sns-controller.v1.1.5
        version: 1.1.5
      - name: ack-sns-controller.v1.1.4
        version: 1.1.4
      - name: ack-sns-controller.v1.1.3
        version: 1.1.3
      - name: ack-sns-controller.v1.1.2
        version: 1.1.2
      - name: ack-sns-controller.v1.1.1
        version: 1.1.1
      - name: ack-sns-controller.v1.1.0
        version: 1.1.0
      - name: ack-sns-controller.v1.0.14
        version: 1.0.14
      - name: ack-sns-controller.v1.0.13
        version: 1.0.13
      - name: ack-sns-controller.v1.0.12
        version: 1.0.12
      - name: ack-sns-controller.v1.0.11
        version: 1.0.11
      - name: ack-sns-controller.v1.0.9
        version: 1.0.9
      - name: ack-sns-controller.v1.0.6
        version: 1.0.6
      - name: ack-sns-controller.v1.0.5
        version: 1.0.5
      - name: ack-sns-controller.v1.0.4
        version: 1.0.4
      - name: ack-sns-controller.v1.0.3
        version: 1.0.3
      - name: ack-sns-controller.v1.0.2
        version: 1.0.2
      - name: ack-sns-controller.v1.0.1
        version: 1.0.1
      - name: ack-sns-controller.v1.0.0
        version: 1.0.0
      - name: ack-sns-controller.v0.0.12
        version: 0.0.12
      - name: ack-sns-controller.v0.0.11
        version: 0.0.11
      - name: ack-sns-controller.v0.0.10
        version: 0.0.10
      - name: ack-sns-controller.v0.0.9
        version: 0.0.9
      - name: ack-sns-controller.v0.0.8
        version: 0.0.8
      - name: ack-sns-controller.v0.0.7
        version: 0.0.7
      - name: ack-sns-controller.v0.0.5
        version: 0.0.5
      - name: ack-sns-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-sns-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Helm Community
      provider-url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
    name: cockroachdb
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroachdb.v5.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1alpha1",
                "kind": "Cockroachdb",
                "metadata": {
                  "name": "cockroachdb-sample"
                },
                "spec": {
                  "clusterDomain": "cluster.local",
                  "conf": {
                    "attrs": [],
                    "cache": "25%",
                    "cluster-name": "",
                    "disable-cluster-name-verification": false,
                    "http-port": 8080,
                    "join": [],
                    "locality": "",
                    "logtostderr": "INFO",
                    "max-sql-memory": "25%",
                    "port": 26257,
                    "single-node": false,
                    "sql-audit-dir": ""
                  },
                  "image": {
                    "credentials": {},
                    "pullPolicy": "IfNotPresent",
                    "repository": "cockroachdb/cockroach",
                    "tag": "v20.2.4"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [],
                    "labels": {},
                    "paths": [
                      "/"
                    ],
                    "tls": []
                  },
                  "init": {
                    "affinity": {},
                    "annotations": {},
                    "labels": {
                      "app.kubernetes.io/component": "init"
                    },
                    "nodeSelector": {},
                    "resources": {},
                    "tolerations": []
                  },
                  "labels": {},
                  "networkPolicy": {
                    "enabled": false,
                    "ingress": {
                      "grpc": [],
                      "http": []
                    }
                  },
                  "service": {
                    "discovery": {
                      "annotations": {},
                      "labels": {
                        "app.kubernetes.io/component": "cockroachdb"
                      }
                    },
                    "ports": {
                      "grpc": {
                        "external": {
                          "name": "grpc",
                          "port": 26257
                        },
                        "internal": {
                          "name": "grpc-internal",
                          "port": 26257
                        }
                      },
                      "http": {
                        "name": "http",
                        "port": 8080
                      }
                    },
                    "public": {
                      "annotations": {},
                      "labels": {
                        "app.kubernetes.io/component": "cockroachdb"
                      },
                      "type": "ClusterIP"
                    }
                  },
                  "serviceMonitor": {
                    "annotations": {},
                    "enabled": false,
                    "interval": "10s",
                    "labels": {}
                  },
                  "statefulset": {
                    "annotations": {},
                    "args": [],
                    "budget": {
                      "maxUnavailable": 1
                    },
                    "env": [],
                    "labels": {
                      "app.kubernetes.io/component": "cockroachdb"
                    },
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "podAffinity": {},
                    "podAntiAffinity": {
                      "topologyKey": "kubernetes.io/hostname",
                      "type": "soft",
                      "weight": 100
                    },
                    "podManagementPolicy": "Parallel",
                    "priorityClassName": "",
                    "replicas": 3,
                    "resources": {},
                    "secretMounts": [],
                    "tolerations": [],
                    "topologySpreadConstraints": {
                      "maxSkew": 1,
                      "topologyKey": "topology.kubernetes.io/zone",
                      "whenUnsatisfiable": "ScheduleAnyway"
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    }
                  },
                  "storage": {
                    "hostPath": "",
                    "persistentVolume": {
                      "annotations": {},
                      "enabled": true,
                      "labels": {},
                      "size": "100Gi",
                      "storageClass": ""
                    }
                  },
                  "tls": {
                    "certs": {
                      "clientRootSecret": "cockroachdb-root",
                      "nodeSecret": "cockroachdb-node",
                      "provided": false,
                      "tlsSecret": false
                    },
                    "enabled": false,
                    "init": {
                      "image": {
                        "credentials": {},
                        "pullPolicy": "IfNotPresent",
                        "repository": "cockroachdb/cockroach-k8s-request-cert",
                        "tag": "0.4"
                      }
                    },
                    "serviceAccount": {
                      "create": true,
                      "name": ""
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          certified: "false"
          containerImage: quay.io/helmoperators/cockroachdb:v5.0.4
          createdAt: 2021-01-23T10-00-00Z
          description: CockroachDB Operator based on the CockroachDB helm chart
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/dmesser/cockroachdb-operator
          support: dmesser
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a CockroachDB cluster
            displayName: CockroachDB
            kind: Cockroachdb
            name: cockroachdbs.charts.operatorhub.io
            version: v1alpha1
        description: |
          CockroachDB is a scalable, survivable, strongly-consistent SQL database.
          ## About this Operator
          This Operator is based on a Helm chart for CockroachDB. It supports the same parameters like the official [CockroachDB Helm Chart](https://github.com/cockroachdb/helm-charts/tree/master/cockroachdb).

          ## Core capabilities
          * **StatefulSet** - Sets up a dynamically scalable CockroachDB cluster using a Kubernetes StatefulSet
          * **Expand Replicas** - Supports expanding the set of replicas by simply editing your object
          * **Dashboard** - Installs the CockroachDB user interface to administer your cluster. Easily expose it via an Ingress rule.
          Review all of the [confiuguration options](https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration) to best run your database instance. The example configuration is derived from the chart's [`values.yaml`](https://github.com/helm/charts/blob/master/stable/cockroachdb/values.yaml).
          ## Using the cluster
          The resulting cluster endpoint can be consumed from a `Service` that follows the pattern: `<StatefulSet-name>-public`. For example to connect using the command line client, use something like the following to obtain the name of the service:
          ```
          kubectl get service -l chart=cockroachdb-5.0.4
          NAME                                           TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)              AGE
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb          ClusterIP   None             <none>        26257/TCP,8080/TCP   24m
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public   ClusterIP   10.106.249.134   <none>        26257/TCP,8080/TCP   24m
          ```
          Then you can use the CockroachDB command line client to connect to the database cluster:
          ```
          kubectl run -it --rm cockroach-client --image=cockroachdb/cockroach --restart=Never --command -- ./cockroach sql --insecure --host example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public
          ```
          ## Before you start
          This Operator requires a cluster with PV support in order to run correctly.

          Note that for a production cluster, you will likely want to override the following parameters in [`values.yaml`](values.yaml) with your own values.

          - `statefulset.resources.requests.memory` and `statefulset.resources.limits.memory` allocate memory resources to CockroachDB pods in your cluster.
          - `conf.cache` and `conf.max-sql-memory` are memory limits that we recommend setting to 1/4 of the above resource allocation. When running CockroachDB, you must set these limits explicitly to avoid running out of memory.
          - `storage.persistentVolume.size` defaults to `100Gi` of disk space per pod, which you may increase or decrease for your use case.
          - `storage.persistentVolume.storageClass` uses the default storage class for your environment. We strongly recommend that you specify a storage class which uses an SSD.
          - `tls.enabled` must be set to `yes`/`true` to deploy in secure mode.

          For more information on overriding the `values.yaml` parameters, please see:
          > <https://www.cockroachlabs.com/docs/stable/orchestrate-cockroachdb-with-kubernetes.html#step-2-start-cockroachdb>

          If you are running in secure mode (with configuration parameter `tls.enabled` set to `yes`/`true`) and `tls.certs.provided` set to `no`/`false`), then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          When `tls.certs.provided` is set to `yes`/`true`, this chart will use certificates created outside of Kubernetes. You may want to use this if you want to use a different certificate authority from the one being used by Kubernetes or if your Kubernetes cluster doesn't fully support certificate-signing requests. To use this, first set up your certificates and load them into your Kubernetes cluster as Secrets using the commands below:

          ```
          mkdir certs
          mkdir my-safe-directory
          cockroach cert create-ca --certs-dir=certs --ca-key=my-safe-directory/ca.key
          cockroach cert create-client root --certs-dir=certs --ca-key=my-safe-directory/ca.key
          kubectl create secret generic cockroachdb-root --from-file=certs
          cockroach cert create-node --certs-dir=certs --ca-key=my-safe-directory/ca.key localhost 127.0.0.1 eerie-horse-cockroachdb-public eerie-horse-cockroachdb-public.default eerie-horse-cockroachdb-public.default.svc.cluster.local *.eerie-horse-cockroachdb *.eerie-horse-cockroachdb.default *.eerie-horse-cockroachdb.default.svc.cluster.local
          kubectl create secret generic cockroachdb-node --from-file=certs
          ```

          Set `tls.certs.tlsSecret` to `yes/true` if you make use of [cert-manager][3] in your cluster.

          [cert-manager][3] stores generated certificates in dedicated TLS secrets. Thus, they are always named:

          * `ca.crt`
          * `tls.crt`
          * `tls.key`

          On the other hand, CockroachDB also demands dedicated certificate filenames:

          * `ca.crt`
          * `node.crt`
          * `node.key`
          * `client.root.crt`
          * `client.root.key`

          By activating `tls.certs.tlsSecret` we benefit from projected secrets and convert the TLS secret filenames to their according CockroachDB filenames.

          If you are running in secure mode, then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          Confirm that all pods are `Running` successfully and init has been completed:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS      RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running     0          1m
          my-release-cockroachdb-1            1/1       Running     0          1m
          my-release-cockroachdb-2            1/1       Running     0          1m
          my-release-cockroachdb-init-k6jcr   0/1       Completed   0          1m
          ```

          Confirm that persistent volumes are created and claimed for each pod:

          ```shell
          kubectl get pv
          ```

          ```
          NAME                                       CAPACITY   ACCESS MODES   RECLAIM POLICY   STATUS    CLAIM                                      STORAGECLASS   REASON    AGE
          pvc-64878ebf-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-0   standard                 51s
          pvc-64945b4f-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-1   standard                 51s
          pvc-649d920d-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-2   standard                 51s
          ```

          ## Upgrading the cluster

          ### Chart version 3.0.0 and after

          Launch a temporary interactive pod and start the built-in SQL client:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          Set `cluster.preserve_downgrade_option`, where `$current_version` is the CockroachDB version currently running (e.g., `19.2`):

          ```sql
          > SET CLUSTER SETTING cluster.preserve_downgrade_option = '$current_version';
          ```

          Exit the shell and delete the temporary pod:

          ```sql
          > \q
          ```

          Kick off the upgrade process by changing the new Docker image, where `$new_version` is the CockroachDB version to which you are upgrading:

          ```shell
          kubectl delete job my-release-cockroachdb-init
          ```

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"image":{"tag":"<new version tag here>"}}}' --type merge
          ```

          Kubernetes will carry out a safe [rolling upgrade](https://kubernetes.io/docs/tutorials/stateful-application/basic-stateful-set/#updating-statefulsets) of your CockroachDB nodes one-by-one. Monitor the cluster's pods until all have been successfully restarted:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS              RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running             0          2m
          my-release-cockroachdb-1            1/1       Running             0          3m
          my-release-cockroachdb-2            1/1       Running             0          3m
          my-release-cockroachdb-3            0/1       ContainerCreating   0          25s
          my-release-cockroachdb-init-nwjkh   0/1       ContainerCreating   0          6s
          ```

          ```shell
          kubectl get pods \
          -o jsonpath='{range .items[*]}{.metadata.name}{"\t"}{.spec.containers[0].image}{"\n"}'
          ```

          ```
          my-release-cockroachdb-0    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-1    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-2    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-3    cockroachdb/cockroach:v20.2.4
          ```

          Resume normal operations. Once you are comfortable that the stability and performance of the cluster is what you'd expect post-upgrade, finalize the upgrade:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          ```sql
          > RESET CLUSTER SETTING cluster.preserve_downgrade_option;
          > \q
          ```

          ## Configuration

          The following table lists the configurable parameters of the CockroachDB chart and their default values.
          For details see the [`values.yaml`](values.yaml) file.

          | Parameter                                                 | Description                                                     | Default                                          |
          | ---------                                                 | -----------                                                     | -------                                          |
          | `clusterDomain`                                           | Cluster's default DNS domain                                    | `cluster.local`                                  |
          | `conf.attrs`                                              | CockroachDB node attributes                                     | `[]`                                             |
          | `conf.cache`                                              | Size of CockroachDB's in-memory cache                           | `25%`                                            |
          | `conf.cluster-name`                                       | Name of CockroachDB cluster                                     | `""`                                             |
          | `conf.disable-cluster-name-verification`                  | Disable CockroachDB cluster name verification                   | `no`                                             |
          | `conf.join`                                               | List of already-existing CockroachDB instances                  | `[]`                                             |
          | `conf.max-disk-temp-storage`                              | Max storage capacity for temp data                              | `0`                                              |
          | `conf.max-offset`                                         | Max allowed clock offset for CockroachDB cluster                | `500ms`                                          |
          | `conf.max-sql-memory`                                     | Max memory to use processing SQL querie                         | `25%`                                            |
          | `conf.locality`                                           | Locality attribute for this deployment                          | `""`                                             |
          | `conf.single-node`                                        | Disable CockroachDB clustering (standalone mode)                | `no`                                             |
          | `conf.sql-audit-dir`                                      | Directory for SQL audit log                                     | `""`                                             |
          | `conf.port`                                               | CockroachDB primary serving port in Pods                        | `26257`                                          |
          | `conf.http-port`                                          | CockroachDB HTTP port in Pods                                   | `8080`                                           |
          | `image.repository`                                        | Container image name                                            | `cockroachdb/cockroach`                          |
          | `image.tag`                                               | Container image tag                                             | `v20.2.4`                                        |
          | `image.pullPolicy`                                        | Container pull policy                                           | `IfNotPresent`                                   |
          | `image.credentials`                                       | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `statefulset.replicas`                                    | StatefulSet replicas number                                     | `3`                                              |
          | `statefulset.updateStrategy`                              | Update strategy for StatefulSet Pods                            | `{"type": "RollingUpdate"}`                      |
          | `statefulset.podManagementPolicy`                         | `OrderedReady`/`Parallel` Pods creation/deletion order          | `Parallel`                                       |
          | `statefulset.budget.maxUnavailable`                       | k8s PodDisruptionBudget parameter                               | `1`                                              |
          | `statefulset.args`                                        | Extra command-line arguments                                    | `[]`                                             |
          | `statefulset.env`                                         | Extra env vars                                                  | `[]`                                             |
          | `statefulset.secretMounts`                                | Additional Secrets to mount at cluster members                  | `[]`                                             |
          | `statefulset.labels`                                      | Additional labels of StatefulSet and its Pods                   | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `statefulset.annotations`                                 | Additional annotations of StatefulSet Pods                      | `{}`                                             |
          | `statefulset.nodeAffinity`                                | [Node affinity rules][2] of StatefulSet Pods                    | `{}`                                             |
          | `statefulset.podAffinity`                                 | [Inter-Pod affinity rules][1] of StatefulSet Pods               | `{}`                                             |
          | `statefulset.podAntiAffinity`                             | [Anti-affinity rules][1] of StatefulSet Pods                    | auto                                             |
          | `statefulset.podAntiAffinity.topologyKey`                 | The topologyKey for auto [anti-affinity rules][1]               | `kubernetes.io/hostname`                         |
          | `statefulset.podAntiAffinity.type`                        | Type of auto [anti-affinity rules][1]                           | `soft`                                           |
          | `statefulset.podAntiAffinity.weight`                      | Weight for `soft` auto [anti-affinity rules][1]                 | `100`                                            |
          | `statefulset.nodeSelector`                                | Node labels for StatefulSet Pods assignment                     | `{}`                                             |
          | `statefulset.priorityClassName`                           | [PriorityClassName][4] for StatefulSet Pods                     | `""`                                             |
          | `statefulset.tolerations`                                 | Node taints to tolerate by StatefulSet Pods                     | `[]`                                             |
          | `statefulset.topologySpreadConstraints`                   | [Topology Spread Constraints rules][5] of StatefulSet Pods      | auto                                             |
          | `statefulset.topologySpreadConstraints.maxSkew`           | Degree to which Pods may be unevenly distributed                | `1`                                              |
          | `statefulset.topologySpreadConstraints.topologyKey`       | The key of node labels                                          | `topology.kubernetes.io/zone`                    |
          | `statefulset.topologySpreadConstraints.whenUnsatisfiable` | `ScheduleAnyway`/`DoNotSchedule` for unsatisfiable constraints  | `ScheduleAnyway`                                 |
          | `statefulset.resources`                                   | Resource requests and limits for StatefulSet Pods               | `{}`                                             |
          | `service.ports.grpc.external.port`                        | CockroachDB primary serving port in Services                    | `26257`                                          |
          | `service.ports.grpc.external.name`                        | CockroachDB primary serving port name in Services               | `grpc`                                           |
          | `service.ports.grpc.internal.port`                        | CockroachDB inter-communication port in Services                | `26257`                                          |
          | `service.ports.grpc.internal.name`                        | CockroachDB inter-communication port name in Services           | `grpc-internal`                                  |
          | `service.ports.http.port`                                 | CockroachDB HTTP port in Services                               | `8080`                                           |
          | `service.ports.http.name`                                 | CockroachDB HTTP port name in Services                          | `http`                                           |
          | `service.public.type`                                     | Public Service type                                             | `ClusterIP`                                      |
          | `service.public.labels`                                   | Additional labels of public Service                             | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.public.annotations`                              | Additional annotations of public Service                        | `{}`                                             |
          | `service.discovery.labels`                                | Additional labels of discovery Service                          | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.discovery.annotations`                           | Additional annotations of discovery Service                     | `{}`                                             |
          | `ingress.enabled`                                         | Enable ingress resource for CockroachDB                         | `false`                                          |
          | `ingress.labels`                                          | Additional labels of Ingress                                    | `{}`                                             |
          | `ingress.annotations`                                     | Additional annotations of Ingress                               | `{}`                                             |
          | `ingress.paths`                                           | Paths for the default host                                      | `[/]`                                            |
          | `ingress.hosts`                                           | CockroachDB Ingress hostnames                                   | `[]`                                             |
          | `ingress.tls[0].hosts`                                    | CockroachDB Ingress tls hostnames                               | `nil`                                            |
          | `ingress.tls[0].secretName`                               | CockroachDB Ingress tls secret name                             | `nil`                                            |
          | `serviceMonitor.enabled`                                  | Create [ServiceMonitor](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/design.md#servicemonitor) Resource for scraping metrics using [PrometheusOperator](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/user-guides/getting-started.md#prometheus-operator)                     | `false`                                             |
          | `serviceMonitor.labels`                                   | Additional labels of ServiceMonitor                             | `{}`                                             |
          | `serviceMonitor.annotations`                              | Additional annotations of ServiceMonitor                        | `{}`                                             |
          | `serviceMonitor.interval`                                 | ServiceMonitor scrape metrics interval                          | `10s`                                            |
          | `serviceMonitor.scrapeTimeout`                            | ServiceMonitor scrape timeout                                   | `nil`                                            |
          | `storage.hostPath`                                        | Absolute path on host to store data                             | `""`                                             |
          | `storage.persistentVolume.enabled`                        | Whether to use PersistentVolume to store data                   | `yes`                                            |
          | `storage.persistentVolume.size`                           | PersistentVolume size                                           | `100Gi`                                          |
          | `storage.persistentVolume.storageClass`                   | PersistentVolume class                                          | `""`                                             |
          | `storage.persistentVolume.labels`                         | Additional labels of PersistentVolumeClaim                      | `{}`                                             |
          | `storage.persistentVolume.annotations`                    | Additional annotations of PersistentVolumeClaim                 | `{}`                                             |
          | `init.labels`                                             | Additional labels of init Job and its Pod                       | `{"app.kubernetes.io/component": "init"}`        |
          | `init.annotations`                                        | Additional labels of the Pod of init Job                        | `{}`                                             |
          | `init.affinity`                                           | [Affinity rules][2] of init Job Pod                             | `{}`                                             |
          | `init.nodeSelector`                                       | Node labels for init Job Pod assignment                         | `{}`                                             |
          | `init.tolerations`                                        | Node taints to tolerate by init Job Pod                         | `[]`                                             |
          | `init.resources`                                          | Resource requests and limits for the Pod of init Job            | `{}`                                             |
          | `tls.enabled`                                             | Whether to run securely using TLS certificates                  | `no`                                             |
          | `tls.serviceAccount.create`                               | Whether to create a new RBAC service account                    | `yes`                                            |
          | `tls.serviceAccount.name`                                 | Name of RBAC service account to use                             | `""`                                             |
          | `tls.certs.provided`                                      | Bring your own certs scenario, i.e certificates are provided    | `no`                                             |
          | `tls.certs.clientRootSecret`                              | If certs are provided, secret name for client root cert         | `cockroachdb-root`                               |
          | `tls.certs.nodeSecret`                                    | If certs are provided, secret name for node cert                | `cockroachdb-node`                               |
          | `tls.certs.tlsSecret`                                     | Own certs are stored in TLS secret                              | `no`                                             |
          | `tls.init.image.repository`                               | Image to use for requesting TLS certificates                    | `cockroachdb/cockroach-k8s-request-cert`         |
          | `tls.init.image.tag`                                      | Image tag to use for requesting TLS certificates                | `0.4`                                            |
          | `tls.init.image.pullPolicy`                               | Requesting TLS certificates container pull policy               | `IfNotPresent`                                   |
          | `tls.init.image.credentials`                              | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `networkPolicy.enabled`                                   | Enable NetworkPolicy for CockroachDB's Pods                     | `no`                                             |
          | `networkPolicy.ingress.grpc`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |
          | `networkPolicy.ingress.http`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |


          Override the default parameters using the `CockroachDB` custom resource.

          ## Deep dive

          ### Connecting to the CockroachDB cluster

          Once you've created the cluster, you can start talking to it by connecting to its `-public` Service. CockroachDB is PostgreSQL wire protocol compatible, so there's a [wide variety of supported clients](https://www.cockroachlabs.com/docs/install-client-drivers.html). As an example, we'll open up a SQL shell using CockroachDB's built-in shell and play around with it a bit, like this (likely needing to replace `my-release-cockroachdb-public` with the name of the `-public` Service that was created with your installed chart):

          ```shell
          kubectl run cockroach-client --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host my-release-cockroachdb-public
          ```

          ```
          Waiting for pod default/cockroach-client to be running, status is Pending,
          pod ready: false
          If you don't see a command prompt, try pressing enter.
          root@my-release-cockroachdb-public:26257> SHOW DATABASES;
          +--------------------+
          |      Database      |
          +--------------------+
          | information_schema |
          | pg_catalog         |
          | system             |
          +--------------------+
          (3 rows)
          root@my-release-cockroachdb-public:26257> CREATE DATABASE bank;
          CREATE DATABASE
          root@my-release-cockroachdb-public:26257> CREATE TABLE bank.accounts (id INT
          PRIMARY KEY, balance DECIMAL);
          CREATE TABLE
          root@my-release-cockroachdb-public:26257> INSERT INTO bank.accounts VALUES
          (1234, 10000.50);
          INSERT 1
          root@my-release-cockroachdb-public:26257> SELECT * FROM bank.accounts;
          +------+---------+
          |  id  | balance |
          +------+---------+
          | 1234 | 10000.5 |
          +------+---------+
          (1 row)
          root@my-release-cockroachdb-public:26257> \q
          Waiting for pod default/cockroach-client to terminate, status is Running
          pod "cockroach-client" deleted
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          ### Cluster health

          Because our pod spec includes regular health checks of the CockroachDB processes, simply running `kubectl get pods` and looking at the `STATUS` column is sufficient to determine the health of each instance in the cluster.

          If you want more detailed information about the cluster, the best place to look is the Admin UI.

          ### Accessing the Admin UI

          If you want to see information about how the cluster is doing, you can try pulling up the CockroachDB Admin UI by port-forwarding from your local machine to one of the pods (replacing `my-release-cockroachdb-0` with the name of one of your pods:

          ```shell
          kubectl port-forward my-release-cockroachdb-0 8080
          ```

          You should then be able to access the Admin UI by visiting <http://localhost:8080/> in your web browser.

          ### Failover

          If any CockroachDB member fails, it is restarted or recreated automatically by the Kubernetes infrastructure, and will re-join the cluster automatically when it comes back up. You can test this scenario by killing any of the CockroachDB pods:

          ```shell
          kubectl delete pod my-release-cockroachdb-1
          ```

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          After a while:

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-1  1/1       Running       0          20s
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          You can check the state of re-joining from the new pod's logs:

          ```shell
          kubectl logs my-release-cockroachdb-1
          ```

          ```
          [...]
          I161028 19:32:09.754026 1 server/node.go:586  [n1] node connected via gossip and
          verified as part of cluster {"35ecbc27-3f67-4e7d-9b8f-27c31aae17d6"}
          [...]
          cockroachdb-0.my-release-cockroachdb.default.svc.cluster.local:26257
          build:      beta-20161027-55-gd2d3c7f @ 2016/10/28 19:27:25 (go1.7.3)
          admin:      http://0.0.0.0:8080
          sql:
          postgresql://root@my-release-cockroachdb-1.my-release-cockroachdb.default.svc.cluster.local:26257?sslmode=disable
          logs:       cockroach-data/logs
          store[0]:   path=cockroach-data
          status:     restarted pre-existing node
          clusterID:  {35ecbc27-3f67-4e7d-9b8f-27c31aae17d6}
          nodeID:     2
          [...]
          ```

          ### NetworkPolicy

          To enable NetworkPolicy for CockroachDB, install [a networking plugin that implements the Kubernetes NetworkPolicy spec](https://kubernetes.io/docs/tasks/administer-cluster/declare-network-policy#before-you-begin), and set `networkPolicy.enabled` to `yes`/`true`.

          For Kubernetes v1.5 & v1.6, you must also turn on NetworkPolicy by setting the `DefaultDeny` Namespace annotation. Note: this will enforce policy for _all_ pods in the Namespace:

          ```shell
          kubectl annotate namespace default "net.beta.kubernetes.io/network-policy={\"ingress\":{\"isolation\":\"DefaultDeny\"}}"
          ```

          For more precise policy, set `networkPolicy.ingress.grpc` and `networkPolicy.ingress.http` rules. This will only allow pods that match the provided rules to connect to CockroachDB.

          ### Scaling

          Scaling should be managed via the `helm upgrade` command. After resizing your cluster on your cloud environment (e.g., GKE or EKS), run the following command to add a pod. This assumes you scaled from 3 to 4 nodes:

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"statefulset":{"replicas":"4"}}}' --type merge
          ```

          Note, that if you are running in secure mode (`tls.enabled` is `yes`/`true`) and increase the size of your cluster, you will also have to approve the CSR (certificate-signing request) of each new node (using `kubectl get csr` and `kubectl certificate approve`).

          [1]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [2]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#node-affinity
          [3]: https://cert-manager.io/
          [4]: https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass
          [5]: https://kubernetes.io/docs/concepts/workloads/pods/pod-topology-spread-constraints/
        displayName: CockroachDB Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cockroach
        - cockroachdb
        links:
        - name: Helm Chart Source
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb
        - name: Configuration Options
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration
        - name: CockroachDB Source
          url: https://github.com/cockroachdb/cockroach
        maintainers:
        - email: dmesser@redhat.com
          name: Daniel Messer
        maturity: stable
        minKubeVersion: 1.8.0
        provider:
          name: Helm Community
          url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/helmoperators/cockroachdb:v5.0.4
        version: 5.0.4
      entries:
      - name: cockroachdb.v5.0.4
        version: 5.0.4
      - name: cockroachdb.v5.0.3
        version: 5.0.3
      name: stable-5.x
    - currentCSV: cockroachdb.v6.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1alpha1",
                "kind": "Cockroachdb",
                "metadata": {
                  "name": "cockroachdb-sample"
                },
                "spec": {"image":{"repository":"cockroachdb/cockroach","tag":"v22.2.4","pullPolicy":"IfNotPresent","credentials":{}},"labels":{},"clusterDomain":"cluster.local","conf":{"attrs":[],"cache":"25%","cluster-name":"","disable-cluster-name-verification":false,"join":[],"log":{"enabled":false,"config":{}},"logtostderr":"INFO","max-sql-memory":"25%","locality":"","single-node":false,"sql-audit-dir":"","port":26257,"http-port":8080,"path":"cockroach-data","store":{"enabled":false,"type":null,"size":null,"attrs":null}},"statefulset":{"replicas":3,"updateStrategy":{"type":"RollingUpdate"},"podManagementPolicy":"Parallel","budget":{"maxUnavailable":1},"args":[],"env":[],"secretMounts":[],"labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{},"nodeAffinity":{},"podAffinity":{},"podAntiAffinity":{"topologyKey":"kubernetes.io/hostname","type":"soft","weight":100},"nodeSelector":{},"priorityClassName":"","tolerations":[],"topologySpreadConstraints":{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"},"resources":{},"customLivenessProbe":{},"customReadinessProbe":{}},"service":{"ports":{"grpc":{"external":{"port":26257,"name":"grpc"},"internal":{"port":26257,"name":"grpc-internal"}},"http":{"port":8080,"name":"http"}},"public":{"type":"ClusterIP","labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{}},"discovery":{"labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{}}},"ingress":{"enabled":false,"labels":{},"annotations":{},"paths":["/"],"hosts":[],"tls":[]},"prometheus":{"enabled":true},"serviceMonitor":{"enabled":false,"labels":{},"annotations":{},"interval":"10s","namespaced":false},"storage":{"hostPath":"","persistentVolume":{"enabled":true,"size":"100Gi","storageClass":"","labels":{},"annotations":{}}},"init":{"labels":{"app.kubernetes.io/component":"init"},"jobAnnotations":{},"annotations":{},"affinity":{},"nodeSelector":{},"tolerations":[],"resources":{},"provisioning":{"enabled":false,"clusterSettings":null,"users":[],"databases":[]}},"tls":{"enabled":true,"serviceAccount":{"create":true,"name":""},"copyCerts":{"image":"busybox"},"certs":{"provided":false,"clientRootSecret":"cockroachdb-root","nodeSecret":"cockroachdb-node","tlsSecret":false,"selfSigner":{"enabled":true,"caProvided":false,"caSecret":"","minimumCertDuration":"624h","caCertDuration":"43800h","caCertExpiryWindow":"648h","clientCertDuration":"672h","clientCertExpiryWindow":"48h","nodeCertDuration":"8760h","nodeCertExpiryWindow":"168h","rotateCerts":true,"readinessWait":"30s","podUpdateTimeout":"2m"},"certManager":false,"certManagerIssuer":{"group":"cert-manager.io","kind":"Issuer","name":"cockroachdb","clientCertDuration":"672h","clientCertExpiryWindow":"48h","nodeCertDuration":"8760h","nodeCertExpiryWindow":"168h"},"useCertManagerV1CRDs":false},"selfSigner":{"image":{"repository":"cockroachlabs-helm-charts/cockroach-self-signer-cert","tag":"1.3","pullPolicy":"IfNotPresent","credentials":{},"registry":"gcr.io"}}},"networkPolicy":{"enabled":false,"ingress":{"grpc":[],"http":[]}},"iap":{"enabled":false}}
              }
            ]
          capabilities: Basic Install
          categories: Database
          certified: "false"
          containerImage: quay.io/cockroachdb/cockroach-helm-operator:6.0.0
          createdAt: 2021-01-23T10-00-00Z
          description: CockroachDB Operator based on the CockroachDB helm chart
          olm.skipRange: <6.0.0
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/cockroachdb/helm-charts
          support: CockroachDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a CockroachDB cluster
            displayName: CockroachDB
            kind: Cockroachdb
            name: cockroachdbs.charts.operatorhub.io
            version: v1alpha1
        description: |
          CockroachDB is a scalable, survivable, strongly-consistent SQL database.
          ## About this Operator
          This Operator is based on a Helm chart for CockroachDB. It supports reconfiguration for some parameters, but notably does not handle scale down of the replica count in a seamless manner. Scale up works great.

          ## Core capabilities
          * **StatefulSet** - Sets up a dynamically scalable CockroachDB cluster using a Kubernetes StatefulSet
          * **Expand Replicas** - Supports expanding the set of replicas by simply editing your object
          * **Dashboard** - Installs the CockroachDB user interface to administer your cluster. Easily expose it via an Ingress rule.
          Review all of the [confiuguration options](https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration) to best run your database instance. The example configuration is derived from the chart's [`values.yaml`](https://github.com/helm/charts/blob/master/stable/cockroachdb/values.yaml).
          ## Using the cluster
          The resulting cluster endpoint can be consumed from a `Service` that follows the pattern: `<StatefulSet-name>-public`. For example to connect using the command line client, use something like the following to obtain the name of the service:
          ```
          kubectl get service -l chart=cockroachdb-6.0.0
          NAME                                           TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)              AGE
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb          ClusterIP   None             <none>        26257/TCP,8080/TCP   24m
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public   ClusterIP   10.106.249.134   <none>        26257/TCP,8080/TCP   24m
          ```
          Then you can use the CockroachDB command line client to connect to the database cluster:
          ```
          kubectl run -it --rm cockroach-client --image=cockroachdb/cockroach --restart=Never --command -- ./cockroach sql --insecure --host example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public
          ```
          ## Before you start
          This Operator requires a cluster with PV support in order to run correctly.

          Note that for a production cluster, you will likely want to override the following parameters in [`values.yaml`](values.yaml) with your own values.

          - `statefulset.resources.requests.memory` and `statefulset.resources.limits.memory` allocate memory resources to CockroachDB pods in your cluster.
          - `conf.cache` and `conf.max-sql-memory` are memory limits that we recommend setting to 1/4 of the above resource allocation. When running CockroachDB, you must set these limits explicitly to avoid running out of memory.
          - `storage.persistentVolume.size` defaults to `100Gi` of disk space per pod, which you may increase or decrease for your use case.
          - `storage.persistentVolume.storageClass` uses the default storage class for your environment. We strongly recommend that you specify a storage class which uses an SSD.
          - `tls.enabled` must be set to `yes`/`true` to deploy in secure mode.

          For more information on overriding the `values.yaml` parameters, please see:
          > <https://www.cockroachlabs.com/docs/stable/orchestrate-cockroachdb-with-kubernetes.html#step-2-start-cockroachdb>

          If you are running in secure mode (with configuration parameter `tls.enabled` set to `yes`/`true`) and `tls.certs.provided` set to `no`/`false`), then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          When `tls.certs.provided` is set to `yes`/`true`, this chart will use certificates created outside of Kubernetes. You may want to use this if you want to use a different certificate authority from the one being used by Kubernetes or if your Kubernetes cluster doesn't fully support certificate-signing requests. To use this, first set up your certificates and load them into your Kubernetes cluster as Secrets using the commands below:

          ```
          mkdir certs
          mkdir my-safe-directory
          cockroach cert create-ca --certs-dir=certs --ca-key=my-safe-directory/ca.key
          cockroach cert create-client root --certs-dir=certs --ca-key=my-safe-directory/ca.key
          kubectl create secret generic cockroachdb-root --from-file=certs
          cockroach cert create-node --certs-dir=certs --ca-key=my-safe-directory/ca.key localhost 127.0.0.1 eerie-horse-cockroachdb-public eerie-horse-cockroachdb-public.default eerie-horse-cockroachdb-public.default.svc.cluster.local *.eerie-horse-cockroachdb *.eerie-horse-cockroachdb.default *.eerie-horse-cockroachdb.default.svc.cluster.local
          kubectl create secret generic cockroachdb-node --from-file=certs
          ```

          Set `tls.certs.tlsSecret` to `yes/true` if you make use of [cert-manager][3] in your cluster.

          [cert-manager][3] stores generated certificates in dedicated TLS secrets. Thus, they are always named:

          * `ca.crt`
          * `tls.crt`
          * `tls.key`

          On the other hand, CockroachDB also demands dedicated certificate filenames:

          * `ca.crt`
          * `node.crt`
          * `node.key`
          * `client.root.crt`
          * `client.root.key`

          By activating `tls.certs.tlsSecret` we benefit from projected secrets and convert the TLS secret filenames to their according CockroachDB filenames.

          If you are running in secure mode, then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          Confirm that all pods are `Running` successfully and init has been completed:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS      RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running     0          1m
          my-release-cockroachdb-1            1/1       Running     0          1m
          my-release-cockroachdb-2            1/1       Running     0          1m
          my-release-cockroachdb-init-k6jcr   0/1       Completed   0          1m
          ```

          Confirm that persistent volumes are created and claimed for each pod:

          ```shell
          kubectl get pv
          ```

          ```
          NAME                                       CAPACITY   ACCESS MODES   RECLAIM POLICY   STATUS    CLAIM                                      STORAGECLASS   REASON    AGE
          pvc-64878ebf-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-0   standard                 51s
          pvc-64945b4f-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-1   standard                 51s
          pvc-649d920d-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-2   standard                 51s
          ```

          ## Upgrading the cluster

          ### Chart version 3.0.0 and after

          Launch a temporary interactive pod and start the built-in SQL client:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          Set `cluster.preserve_downgrade_option`, where `$current_version` is the CockroachDB version currently running (e.g., `19.2`):

          ```sql
          > SET CLUSTER SETTING cluster.preserve_downgrade_option = '$current_version';
          ```

          Exit the shell and delete the temporary pod:

          ```sql
          > \q
          ```

          Kick off the upgrade process by changing the new Docker image, where `$new_version` is the CockroachDB version to which you are upgrading:

          ```shell
          kubectl delete job my-release-cockroachdb-init
          ```

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"image":{"tag":"<new version tag here>"}}}' --type merge
          ```

          Kubernetes will carry out a safe [rolling upgrade](https://kubernetes.io/docs/tutorials/stateful-application/basic-stateful-set/#updating-statefulsets) of your CockroachDB nodes one-by-one. Monitor the cluster's pods until all have been successfully restarted:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS              RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running             0          2m
          my-release-cockroachdb-1            1/1       Running             0          3m
          my-release-cockroachdb-2            1/1       Running             0          3m
          my-release-cockroachdb-3            0/1       ContainerCreating   0          25s
          my-release-cockroachdb-init-nwjkh   0/1       ContainerCreating   0          6s
          ```

          ```shell
          kubectl get pods \
          -o jsonpath='{range .items[*]}{.metadata.name}{"\t"}{.spec.containers[0].image}{"\n"}'
          ```

          ```
          my-release-cockroachdb-0    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-1    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-2    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-3    cockroachdb/cockroach:v20.2.4
          ```

          Resume normal operations. Once you are comfortable that the stability and performance of the cluster is what you'd expect post-upgrade, finalize the upgrade:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          ```sql
          > RESET CLUSTER SETTING cluster.preserve_downgrade_option;
          > \q
          ```

          ## Configuration

          The following table lists the configurable parameters of the CockroachDB chart and their default values.
          For details see the [`values.yaml`](values.yaml) file.

          | Parameter                                                 | Description                                                     | Default                                          |
          | ---------                                                 | -----------                                                     | -------                                          |
          | `clusterDomain`                                           | Cluster's default DNS domain                                    | `cluster.local`                                  |
          | `conf.attrs`                                              | CockroachDB node attributes                                     | `[]`                                             |
          | `conf.cache`                                              | Size of CockroachDB's in-memory cache                           | `25%`                                            |
          | `conf.cluster-name`                                       | Name of CockroachDB cluster                                     | `""`                                             |
          | `conf.disable-cluster-name-verification`                  | Disable CockroachDB cluster name verification                   | `no`                                             |
          | `conf.join`                                               | List of already-existing CockroachDB instances                  | `[]`                                             |
          | `conf.max-disk-temp-storage`                              | Max storage capacity for temp data                              | `0`                                              |
          | `conf.max-offset`                                         | Max allowed clock offset for CockroachDB cluster                | `500ms`                                          |
          | `conf.max-sql-memory`                                     | Max memory to use processing SQL querie                         | `25%`                                            |
          | `conf.locality`                                           | Locality attribute for this deployment                          | `""`                                             |
          | `conf.single-node`                                        | Disable CockroachDB clustering (standalone mode)                | `no`                                             |
          | `conf.sql-audit-dir`                                      | Directory for SQL audit log                                     | `""`                                             |
          | `conf.port`                                               | CockroachDB primary serving port in Pods                        | `26257`                                          |
          | `conf.http-port`                                          | CockroachDB HTTP port in Pods                                   | `8080`                                           |
          | `image.repository`                                        | Container image name                                            | `cockroachdb/cockroach`                          |
          | `image.tag`                                               | Container image tag                                             | `v20.2.4`                                        |
          | `image.pullPolicy`                                        | Container pull policy                                           | `IfNotPresent`                                   |
          | `image.credentials`                                       | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `statefulset.replicas`                                    | StatefulSet replicas number                                     | `3`                                              |
          | `statefulset.updateStrategy`                              | Update strategy for StatefulSet Pods                            | `{"type": "RollingUpdate"}`                      |
          | `statefulset.podManagementPolicy`                         | `OrderedReady`/`Parallel` Pods creation/deletion order          | `Parallel`                                       |
          | `statefulset.budget.maxUnavailable`                       | k8s PodDisruptionBudget parameter                               | `1`                                              |
          | `statefulset.args`                                        | Extra command-line arguments                                    | `[]`                                             |
          | `statefulset.env`                                         | Extra env vars                                                  | `[]`                                             |
          | `statefulset.secretMounts`                                | Additional Secrets to mount at cluster members                  | `[]`                                             |
          | `statefulset.labels`                                      | Additional labels of StatefulSet and its Pods                   | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `statefulset.annotations`                                 | Additional annotations of StatefulSet Pods                      | `{}`                                             |
          | `statefulset.nodeAffinity`                                | [Node affinity rules][2] of StatefulSet Pods                    | `{}`                                             |
          | `statefulset.podAffinity`                                 | [Inter-Pod affinity rules][1] of StatefulSet Pods               | `{}`                                             |
          | `statefulset.podAntiAffinity`                             | [Anti-affinity rules][1] of StatefulSet Pods                    | auto                                             |
          | `statefulset.podAntiAffinity.topologyKey`                 | The topologyKey for auto [anti-affinity rules][1]               | `kubernetes.io/hostname`                         |
          | `statefulset.podAntiAffinity.type`                        | Type of auto [anti-affinity rules][1]                           | `soft`                                           |
          | `statefulset.podAntiAffinity.weight`                      | Weight for `soft` auto [anti-affinity rules][1]                 | `100`                                            |
          | `statefulset.nodeSelector`                                | Node labels for StatefulSet Pods assignment                     | `{}`                                             |
          | `statefulset.priorityClassName`                           | [PriorityClassName][4] for StatefulSet Pods                     | `""`                                             |
          | `statefulset.tolerations`                                 | Node taints to tolerate by StatefulSet Pods                     | `[]`                                             |
          | `statefulset.topologySpreadConstraints`                   | [Topology Spread Constraints rules][5] of StatefulSet Pods      | auto                                             |
          | `statefulset.topologySpreadConstraints.maxSkew`           | Degree to which Pods may be unevenly distributed                | `1`                                              |
          | `statefulset.topologySpreadConstraints.topologyKey`       | The key of node labels                                          | `topology.kubernetes.io/zone`                    |
          | `statefulset.topologySpreadConstraints.whenUnsatisfiable` | `ScheduleAnyway`/`DoNotSchedule` for unsatisfiable constraints  | `ScheduleAnyway`                                 |
          | `statefulset.resources`                                   | Resource requests and limits for StatefulSet Pods               | `{}`                                             |
          | `service.ports.grpc.external.port`                        | CockroachDB primary serving port in Services                    | `26257`                                          |
          | `service.ports.grpc.external.name`                        | CockroachDB primary serving port name in Services               | `grpc`                                           |
          | `service.ports.grpc.internal.port`                        | CockroachDB inter-communication port in Services                | `26257`                                          |
          | `service.ports.grpc.internal.name`                        | CockroachDB inter-communication port name in Services           | `grpc-internal`                                  |
          | `service.ports.http.port`                                 | CockroachDB HTTP port in Services                               | `8080`                                           |
          | `service.ports.http.name`                                 | CockroachDB HTTP port name in Services                          | `http`                                           |
          | `service.public.type`                                     | Public Service type                                             | `ClusterIP`                                      |
          | `service.public.labels`                                   | Additional labels of public Service                             | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.public.annotations`                              | Additional annotations of public Service                        | `{}`                                             |
          | `service.discovery.labels`                                | Additional labels of discovery Service                          | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.discovery.annotations`                           | Additional annotations of discovery Service                     | `{}`                                             |
          | `ingress.enabled`                                         | Enable ingress resource for CockroachDB                         | `false`                                          |
          | `ingress.labels`                                          | Additional labels of Ingress                                    | `{}`                                             |
          | `ingress.annotations`                                     | Additional annotations of Ingress                               | `{}`                                             |
          | `ingress.paths`                                           | Paths for the default host                                      | `[/]`                                            |
          | `ingress.hosts`                                           | CockroachDB Ingress hostnames                                   | `[]`                                             |
          | `ingress.tls[0].hosts`                                    | CockroachDB Ingress tls hostnames                               | `nil`                                            |
          | `ingress.tls[0].secretName`                               | CockroachDB Ingress tls secret name                             | `nil`                                            |
          | `serviceMonitor.enabled`                                  | Create [ServiceMonitor](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/design.md#servicemonitor) Resource for scraping metrics using [PrometheusOperator](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/user-guides/getting-started.md#prometheus-operator)                     | `false`                                             |
          | `serviceMonitor.labels`                                   | Additional labels of ServiceMonitor                             | `{}`                                             |
          | `serviceMonitor.annotations`                              | Additional annotations of ServiceMonitor                        | `{}`                                             |
          | `serviceMonitor.interval`                                 | ServiceMonitor scrape metrics interval                          | `10s`                                            |
          | `serviceMonitor.scrapeTimeout`                            | ServiceMonitor scrape timeout                                   | `nil`                                            |
          | `storage.hostPath`                                        | Absolute path on host to store data                             | `""`                                             |
          | `storage.persistentVolume.enabled`                        | Whether to use PersistentVolume to store data                   | `yes`                                            |
          | `storage.persistentVolume.size`                           | PersistentVolume size                                           | `100Gi`                                          |
          | `storage.persistentVolume.storageClass`                   | PersistentVolume class                                          | `""`                                             |
          | `storage.persistentVolume.labels`                         | Additional labels of PersistentVolumeClaim                      | `{}`                                             |
          | `storage.persistentVolume.annotations`                    | Additional annotations of PersistentVolumeClaim                 | `{}`                                             |
          | `init.labels`                                             | Additional labels of init Job and its Pod                       | `{"app.kubernetes.io/component": "init"}`        |
          | `init.annotations`                                        | Additional labels of the Pod of init Job                        | `{}`                                             |
          | `init.affinity`                                           | [Affinity rules][2] of init Job Pod                             | `{}`                                             |
          | `init.nodeSelector`                                       | Node labels for init Job Pod assignment                         | `{}`                                             |
          | `init.tolerations`                                        | Node taints to tolerate by init Job Pod                         | `[]`                                             |
          | `init.resources`                                          | Resource requests and limits for the Pod of init Job            | `{}`                                             |
          | `tls.enabled`                                             | Whether to run securely using TLS certificates                  | `no`                                             |
          | `tls.serviceAccount.create`                               | Whether to create a new RBAC service account                    | `yes`                                            |
          | `tls.serviceAccount.name`                                 | Name of RBAC service account to use                             | `""`                                             |
          | `tls.certs.provided`                                      | Bring your own certs scenario, i.e certificates are provided    | `no`                                             |
          | `tls.certs.clientRootSecret`                              | If certs are provided, secret name for client root cert         | `cockroachdb-root`                               |
          | `tls.certs.nodeSecret`                                    | If certs are provided, secret name for node cert                | `cockroachdb-node`                               |
          | `tls.certs.tlsSecret`                                     | Own certs are stored in TLS secret                              | `no`                                             |
          | `tls.init.image.repository`                               | Image to use for requesting TLS certificates                    | `cockroachdb/cockroach-k8s-request-cert`         |
          | `tls.init.image.tag`                                      | Image tag to use for requesting TLS certificates                | `0.4`                                            |
          | `tls.init.image.pullPolicy`                               | Requesting TLS certificates container pull policy               | `IfNotPresent`                                   |
          | `tls.init.image.credentials`                              | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `networkPolicy.enabled`                                   | Enable NetworkPolicy for CockroachDB's Pods                     | `no`                                             |
          | `networkPolicy.ingress.grpc`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |
          | `networkPolicy.ingress.http`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |


          Override the default parameters using the `CockroachDB` custom resource.

          ## Deep dive

          ### Connecting to the CockroachDB cluster

          Once you've created the cluster, you can start talking to it by connecting to its `-public` Service. CockroachDB is PostgreSQL wire protocol compatible, so there's a [wide variety of supported clients](https://www.cockroachlabs.com/docs/install-client-drivers.html). As an example, we'll open up a SQL shell using CockroachDB's built-in shell and play around with it a bit, like this (likely needing to replace `my-release-cockroachdb-public` with the name of the `-public` Service that was created with your installed chart):

          ```shell
          kubectl run cockroach-client --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host my-release-cockroachdb-public
          ```

          ```
          Waiting for pod default/cockroach-client to be running, status is Pending,
          pod ready: false
          If you don't see a command prompt, try pressing enter.
          root@my-release-cockroachdb-public:26257> SHOW DATABASES;
          +--------------------+
          |      Database      |
          +--------------------+
          | information_schema |
          | pg_catalog         |
          | system             |
          +--------------------+
          (3 rows)
          root@my-release-cockroachdb-public:26257> CREATE DATABASE bank;
          CREATE DATABASE
          root@my-release-cockroachdb-public:26257> CREATE TABLE bank.accounts (id INT
          PRIMARY KEY, balance DECIMAL);
          CREATE TABLE
          root@my-release-cockroachdb-public:26257> INSERT INTO bank.accounts VALUES
          (1234, 10000.50);
          INSERT 1
          root@my-release-cockroachdb-public:26257> SELECT * FROM bank.accounts;
          +------+---------+
          |  id  | balance |
          +------+---------+
          | 1234 | 10000.5 |
          +------+---------+
          (1 row)
          root@my-release-cockroachdb-public:26257> \q
          Waiting for pod default/cockroach-client to terminate, status is Running
          pod "cockroach-client" deleted
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          ### Cluster health

          Because our pod spec includes regular health checks of the CockroachDB processes, simply running `kubectl get pods` and looking at the `STATUS` column is sufficient to determine the health of each instance in the cluster.

          If you want more detailed information about the cluster, the best place to look is the Admin UI.

          ### Accessing the Admin UI

          If you want to see information about how the cluster is doing, you can try pulling up the CockroachDB Admin UI by port-forwarding from your local machine to one of the pods (replacing `my-release-cockroachdb-0` with the name of one of your pods:

          ```shell
          kubectl port-forward my-release-cockroachdb-0 8080
          ```

          You should then be able to access the Admin UI by visiting <http://localhost:8080/> in your web browser.

          ### Failover

          If any CockroachDB member fails, it is restarted or recreated automatically by the Kubernetes infrastructure, and will re-join the cluster automatically when it comes back up. You can test this scenario by killing any of the CockroachDB pods:

          ```shell
          kubectl delete pod my-release-cockroachdb-1
          ```

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          After a while:

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-1  1/1       Running       0          20s
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          You can check the state of re-joining from the new pod's logs:

          ```shell
          kubectl logs my-release-cockroachdb-1
          ```

          ```
          [...]
          I161028 19:32:09.754026 1 server/node.go:586  [n1] node connected via gossip and
          verified as part of cluster {"35ecbc27-3f67-4e7d-9b8f-27c31aae17d6"}
          [...]
          cockroachdb-0.my-release-cockroachdb.default.svc.cluster.local:26257
          build:      beta-20161027-55-gd2d3c7f @ 2016/10/28 19:27:25 (go1.7.3)
          admin:      http://0.0.0.0:8080
          sql:
          postgresql://root@my-release-cockroachdb-1.my-release-cockroachdb.default.svc.cluster.local:26257?sslmode=disable
          logs:       cockroach-data/logs
          store[0]:   path=cockroach-data
          status:     restarted pre-existing node
          clusterID:  {35ecbc27-3f67-4e7d-9b8f-27c31aae17d6}
          nodeID:     2
          [...]
          ```

          ### NetworkPolicy

          To enable NetworkPolicy for CockroachDB, install [a networking plugin that implements the Kubernetes NetworkPolicy spec](https://kubernetes.io/docs/tasks/administer-cluster/declare-network-policy#before-you-begin), and set `networkPolicy.enabled` to `yes`/`true`.

          For Kubernetes v1.5 & v1.6, you must also turn on NetworkPolicy by setting the `DefaultDeny` Namespace annotation. Note: this will enforce policy for _all_ pods in the Namespace:

          ```shell
          kubectl annotate namespace default "net.beta.kubernetes.io/network-policy={\"ingress\":{\"isolation\":\"DefaultDeny\"}}"
          ```

          For more precise policy, set `networkPolicy.ingress.grpc` and `networkPolicy.ingress.http` rules. This will only allow pods that match the provided rules to connect to CockroachDB.

          ### Scaling

          Scaling should be managed via the `helm upgrade` command. After resizing your cluster on your cloud environment (e.g., GKE or EKS), run the following command to add a pod. This assumes you scaled from 3 to 4 nodes:

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"statefulset":{"replicas":"4"}}}' --type merge
          ```

          Note, that if you are running in secure mode (`tls.enabled` is `yes`/`true`) and increase the size of your cluster, you will also have to approve the CSR (certificate-signing request) of each new node (using `kubectl get csr` and `kubectl certificate approve`).

          [1]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [2]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#node-affinity
          [3]: https://cert-manager.io/
          [4]: https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass
          [5]: https://kubernetes.io/docs/concepts/workloads/pods/pod-topology-spread-constraints/
        displayName: CockroachDB Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cockroach
        - cockroachdb
        links:
        - name: Helm Chart Source
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb
        - name: Configuration Options
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration
        - name: CockroachDB Source
          url: https://github.com/cockroachdb/cockroach
        maintainers:
        - email: yandu@cockroachlabs.com
          name: Yandu Oppacher
        maturity: stable
        minKubeVersion: 1.8.0
        provider:
          name: Helm Community
          url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/cockroachdb/cockroach-helm-operator:6.0.0
        version: 6.0.0
      entries:
      - name: cockroachdb.v6.0.0
        version: 6.0.0
      name: stable-v6.x
    defaultChannel: stable-v6.x
    packageName: cockroachdb
    provider:
      name: Helm Community
      url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CrowdStrike
      provider-url: https://crowdStrike.com
    name: falcon-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: falcon-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconContainer",
                "metadata": {
                  "name": "falcon-sidecar-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "sidecar"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconNodeSensor",
                "metadata": {
                  "name": "falcon-node-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "daemonset"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security,Monitoring
          containerImage: quay.io/crowdstrike/falcon-operator:0.8.1
          createdAt: "2024-06-06T10:00:00Z"
          description: '[DEPRECATED] Use the CrowdStrike Falcon Operator from the
            certified channel instead'
          operatorframework.io/suggested-namespace: falcon-operator
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/CrowdStrike/falcon-operator
          support: Community Only
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FalconContainer is the Schema for the falconcontainers API
            displayName: Falcon Container
            kind: FalconContainer
            name: falconcontainers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconNodeSensor is the Schema for the falconnodesensors
              API
            displayName: Falcon Node Sensor
            kind: FalconNodeSensor
            name: falconnodesensors.falcon.crowdstrike.com
            version: v1alpha1
        description: |-
          ## Deprecated
          This CrowdStrike community operator is deprecated, use the CrowdStrike Falcon Operator from the certified channel instead.

          ## About the CrowdStrike Falcon Operator
          The CrowdStrike Falcon Operator deploys CrowdStrike Falcon Workload Protection to the cluster. The operator exposes 2 custom resources that allows you to deploy either the Falcon Container Sensor or Falcon Node Sensor.

          ## About Custom Resources

          | Custom Resource                                       | Description                                                      |
          | :--------                                             | :------------                                                    |
          | [FalconContainer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/container/README.md) | Manages installation of Falcon Container Sensor on the cluster   |
          | [FalconNodeSensor](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/node/README.md)     | Manages installation of Falcon Linux Sensor on the cluster nodes |

          ## Installation and Deployment

          For installation and deployment of the CrowdStrike Falcon Operator and its Custom Resources, please read the [Installation and Deployment Guide](./docs/install_guide.md) and choose the deployment method that is right for your target environment.

          ## Getting Help
          If you encounter any issues while using the Falcon Operator, you can create an issue on our [Github repo](https://github.com/CrowdStrike/falcon-operator) for bugs, enhancements, or other requests.

          ## Contributing
          You can contribute by:

          * Raising any issues you find using Falcon Operator
          * Fixing issues by opening [Pull Requests](https://github.com/CrowdStrike/falcon-operator/pulls)
          * Improving documentation
          * Talking about the CrowdStrike Falcon Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/CrowdStrike/falcon-operator/issues).

          ## Additional Resources
           - CrowdStrike Container Security: [Product Page](https://www.crowdstrike.com/products/cloud-security/falcon-cloud-workload-protection/container-security/)
           - So You Think Your Containers Are Secure? Four Steps to Ensure a Secure Container Deployment: [Blog Post](https://www.crowdstrike.com/blog/four-steps-to-ensure-a-secure-containter-deployment/)
           - Container Security With CrowdStrike: [Blog Post](https://www.crowdstrike.com/blog/tech-center/container-security/)
           - To learn more about Falcon Container Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/support/documentation/146/falcon-container-sensor-for-linux), [Release Notes](https://falcon.crowdstrike.com/support/news/release-notes-falcon-container-sensor-for-linux)
           - To learn more about Falcon Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/documentation/20/falcon-sensor-for-linux#kubernetes-support)
           - [Developer Documentation](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/developer_guide.md)
        displayName: '[DEPRECATED] CrowdStrike Operator'
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - runtime-protection
        links:
        - name: Falcon Operator
          url: https://github.com/CrowdStrike/falcon-operator
        maintainers:
        - email: integrations@crowdstrike.com
          name: CrowdStrike Solutions Architecture
        maturity: alpha
        minKubeVersion: 1.17.0
        provider:
          name: CrowdStrike
          url: https://crowdStrike.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/crowdstrike/falcon-operator:0.8.1
        version: 0.8.2
      entries:
      - name: falcon-operator.v0.8.2
        version: 0.8.2
      - name: falcon-operator.v0.8.1
        version: 0.8.1
      name: alpha
    defaultChannel: alpha
    packageName: falcon-operator
    provider:
      name: CrowdStrike
      url: https://crowdStrike.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Foundation Model Stack
      provider-url: ""
    name: multi-nic-cni-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multi-nic-cni-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.3.0",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.3.0",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.3.0
          createdAt: "2024-04-17T04:51:40Z"
          description: Automate container multi-network configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: http://github.com/foundation-model-stack/multi-nic-cni
          support: Foundation Model Stack Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Foundation Model Stack
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.3.0
        version: 1.3.0
      entries:
      - name: multi-nic-cni-operator.v1.3.0
        version: 1.3.0
      - name: multi-nic-cni-operator.v1.2.9
        version: 1.2.9
      - name: multi-nic-cni-operator.v1.2.8
        version: 1.2.8
      - name: multi-nic-cni-operator.v1.2.2
        version: 1.2.2
      - name: multi-nic-cni-operator.v1.2.1
        version: 1.2.1
      - name: multi-nic-cni-operator.v1.0.2
        version: 1.0.2
      name: alpha
    - currentCSV: multi-nic-cni-operator.v1.2.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.7",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.7",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.7
          createdAt: "2025-05-09T12:45:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: minzhang@redhat.com
          name: Min Zhang
        - email: ebendror@redhat.com
          name: Eran Ben Dror
        maturity: alpha
        provider:
          name: Foundation Model Stack
        relatedImages:
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.7
        version: 1.2.7
      entries:
      - name: multi-nic-cni-operator.v1.2.7
        version: 1.2.7
      - name: multi-nic-cni-operator.v1.2.3
        version: 1.2.3
      - name: multi-nic-cni-operator.v1.2.0
        version: 1.2.0
      - name: multi-nic-cni-operator.v1.1.0
        version: 1.1.0
      name: beta
    - currentCSV: multi-nic-cni-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.6",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.6",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.6
          createdAt: "2025-03-24T14:49:54Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: minzhang@redhat.com
          name: Min Zhang
        - email: ebendror@redhat.com
          name: Eran Ben Dror
        maturity: alpha
        provider:
          name: Foundation Model Stack
        relatedImages:
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.6
        version: 1.2.6
      entries:
      - name: multi-nic-cni-operator.v1.2.6
        version: 1.2.6
      - name: multi-nic-cni-operator.v1.2.5
        version: 1.2.5
      - name: multi-nic-cni-operator.v1.2.4
        version: 1.2.4
      - name: multi-nic-cni-operator.v1.0.5
        version: 1.0.5
      - name: multi-nic-cni-operator.v1.0.4
        version: 1.0.4
      - name: multi-nic-cni-operator.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: multi-nic-cni-operator
    provider:
      name: Foundation Model Stack
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: assisted-service-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: assisted-service-operator.v0.7.74
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-x86_64-live.x86_64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-aarch64-live.aarch64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-s390x-live.s390x.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-x86_64-live.x86_64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-aarch64-live.aarch64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-s390x-live.s390x.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-x86_64-live.x86_64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-aarch64-live.aarch64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-ppc64le-live.ppc64le.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-s390x-live.s390x.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-x86_64-live.x86_64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-aarch64-live.aarch64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-ppc64le-live.ppc64le.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-s390x-live.s390x.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-s390x-live.s390x.iso",
                      "version": "418.94.202409162337-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service:latest
          createdAt: "2024-10-16T17:43:34Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.74'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AgentClassification is the Schema for the AgentClassifications
              API
            displayName: Agent Classification
            kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - description: Agent is the Schema for the hosts API
            displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: Infra Env
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service:latest
        version: 0.7.74
      entries:
      - name: assisted-service-operator.v0.7.74
        version: 0.7.74
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: assisted-service-operator.v0.7.71
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-x86_64-live.x86_64.iso",
                      "version": "414.92.202310170514-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-aarch64-live.aarch64.iso",
                      "version": "414.92.202310170514-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202309201615-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-s390x-live.s390x.iso",
                      "version": "414.92.202309201615-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-s390x-live.s390x.iso",
                      "version": "415.92.202310310037-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:cffbdc5fd51c08f16ec2aa484cf226e3b29510919a8aa9668d766007de8f3d79
          createdAt: "2023-12-15T15:26:08Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.71'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service@sha256:cffbdc5fd51c08f16ec2aa484cf226e3b29510919a8aa9668d766007de8f3d79
        - quay.io/edge-infrastructure/assisted-image-service@sha256:027fa9c5a881848786745aafeac5e48575eefe22417abe29d9e8eb17940b8f80
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/edge-infrastructure/assisted-installer@sha256:0f4ecf9a75b6755d02a45f9b70357ac45b6b4db66f24f0e83f888269d64f4293
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:40bafed94547a5fe96deb5492f10225501b965198227c9634869931692bda492
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:98740646fbfc0520f9d85ac5b483d1a898f22e164c5234901be51970bef7fcf3
        version: 0.7.71
      entries:
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.10
    - currentCSV: assisted-service-operator.v0.7.73
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-x86_64-live.x86_64.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-aarch64-live.aarch64.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-s390x-live.s390x.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-x86_64-live.x86_64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-aarch64-live.aarch64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-s390x-live.s390x.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.16.0-ec.4/rhcos-4.16.0-ec.4-x86_64-live.x86_64.iso",
                      "version": "416.94.202403071059-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-aarch64-live.aarch64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-s390x-live.s390x.iso",
                      "version": "415.92.202402130021-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:641fb65d8f0117cc6007567337ae38d5e912add454a76905617b60db99b60a64
          createdAt: "2024-03-20T10:08:49Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.73'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:2cc14304ba27c51d61b86db05445f3f0acc68593a4d76081da8eb9d6bae41757
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:d0da07ec81877302fc376e8049dca69c1acdbf56a37aa3e320cd75f2b85a2ed0
        - quay.io/edge-infrastructure/assisted-service@sha256:641fb65d8f0117cc6007567337ae38d5e912add454a76905617b60db99b60a64
        - quay.io/edge-infrastructure/assisted-image-service@sha256:b33068bbe88e38c0444fb158e233e5644fb127416198d52fdc809a3f2e684fde
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/edge-infrastructure/assisted-installer@sha256:234e4680b0ff4376f2c11d6f94a75a5dd6ae16c3218d6b63d2003cf43c65e185
        version: 0.7.73
      entries:
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.11
    - currentCSV: assisted-service-operator.v0.7.74
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-x86_64-live.x86_64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-aarch64-live.aarch64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-s390x-live.s390x.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-x86_64-live.x86_64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-aarch64-live.aarch64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-s390x-live.s390x.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-x86_64-live.x86_64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-aarch64-live.aarch64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-ppc64le-live.ppc64le.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-s390x-live.s390x.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-x86_64-live.x86_64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-aarch64-live.aarch64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-ppc64le-live.ppc64le.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-s390x-live.s390x.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-s390x-live.s390x.iso",
                      "version": "418.94.202409162337-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service:latest
          createdAt: "2024-10-16T17:43:34Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.74'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AgentClassification is the Schema for the AgentClassifications
              API
            displayName: Agent Classification
            kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - description: Agent is the Schema for the hosts API
            displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: Infra Env
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service:latest
        version: 0.7.74
      entries:
      - name: assisted-service-operator.v0.7.74
        version: 0.7.74
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.13
    - currentCSV: assisted-service-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "osImages": [
                    {
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202105062123-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/ocpmetal/assisted-service@sha256:64e88c5aa41c606bc2db2b0991acdbb1166559306474e13f5e2810f325616d64
          createdAt: "2021-09-23T11:11:20Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.0 <0.0.8'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled, for
              all other names a "Warning" event will be emitted + message logged and
              it will be ignored forever.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent".
          No other name will be accepted.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/ocpmetal/assisted-service@sha256:64e88c5aa41c606bc2db2b0991acdbb1166559306474e13f5e2810f325616d64
        - quay.io/ocpmetal/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/ocpmetal/assisted-installer@sha256:db97f7a565a8d5af61f3876d6465996e90c7f79dad6ec71c5bb5cce837444974
        - quay.io/ocpmetal/assisted-installer-agent@sha256:a391892766d1ce81300bf82943318e7c3be0d2945fc19065527ab9f64fc24cf8
        - quay.io/ocpmetal/assisted-installer-controller@sha256:43b65c310eee985ee0fc20ad8ee684f7557ffe704fd1a2fa53831e3dbc41e849
        version: 0.0.8
      entries:
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.3
    - currentCSV: assisted-service-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.2/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.2/rhcos-4.8.2-x86_64-live.x86_64.iso",
                      "version": "48.84.202107202156-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-x86_64-live.x86_64.iso",
                      "version": "49.84.202110081407-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:c3a852842b88f47c51cf1c62d3a19b102729fb355df19ced0e03e8265739c39c
          createdAt: "2022-04-07T11:23:06Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.1.3'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled, for
              all other names a "Warning" event will be emitted + message logged and
              it will be ignored forever.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent".
          No other name will be accepted.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-image-service@sha256:add19e251cafdfc4f34264d0aec5b660f0ea2dcc13326f4a2c8d50a2e465c0a2
        - quay.io/edge-infrastructure/assisted-service@sha256:c3a852842b88f47c51cf1c62d3a19b102729fb355df19ced0e03e8265739c39c
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/edge-infrastructure/assisted-installer@sha256:461d60b21d1d3ccb15a170f937f4871088cd0f83108b2271e84bdf2883a66870
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:979650f2e03e457d71958e8813285fcb7185fe0035f21a6ba956b4277cb29b70
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:658787d1e6c1f733c5f03863655aff315258c0b38e1972573c3878df46e4e463
        version: 0.1.3
      entries:
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.4
    - currentCSV: assisted-service-operator.v0.3.30
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-x86_64-live.x86_64.iso",
                      "version": "49.84.202110081407-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-aarch64-live-rootfs.aarch64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-aarch64-live.aarch64.iso",
                      "version": "49.84.202110080947-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.3/rhcos-4.10.3-x86_64-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.3/rhcos-4.10.3-x86_64-live.x86_64.iso",
                      "version": "410.84.202201251210-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:7c13ce52d82f761aec47bc5a6639d7baabe99d0f4f2e2106c4341018c606b458
          createdAt: "2022-06-13T23:16:28Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.3.30'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service@sha256:7c13ce52d82f761aec47bc5a6639d7baabe99d0f4f2e2106c4341018c606b458
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/edge-infrastructure/assisted-installer@sha256:57cf8d5fe49067a8c1345bedaf90f2b63aededdd7d615878496182c35a13892e
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:6271d639e4c0647bd28fbc85cea54f31143a876c3237376b9c670ec254a9c2fc
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:2267d2d7be38f81d530fed977b0c46f6c976369cd89c123c8e036c3b1dd988d5
        - quay.io/edge-infrastructure/assisted-image-service@sha256:d6396f9360b877a4f57139ba5b4969f243278db709be4de3573f2fba75e50018
        version: 0.3.30
      entries:
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.5
    - currentCSV: assisted-service-operator.v0.4.41
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.40/rhcos-4.9.40-x86_64-live.x86_64.iso",
                      "version": "49.84.202206171736-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.16/rhcos-4.10.16-x86_64-live.x86_64.iso",
                      "version": "410.84.202205191234-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.16/rhcos-4.10.16-aarch64-live.aarch64.iso",
                      "version": "410.84.202205191023-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.11.0-rc.2/rhcos-4.11.0-rc.2-x86_64-live.x86_64.iso",
                      "version": "411.86.202207090519-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.11.0-rc.2/rhcos-4.11.0-rc.2-aarch64-live.aarch64.iso",
                      "version": "411.86.202207090159-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:0fa869095185759c26c391f0e353f293c903ea739576f160d6668f5f20404bdb
          createdAt: "2022-09-11T14:56:07Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.4.41'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:56484f99799d22f345215bb65c6b81b6ee14e07d26b625c3f283a2ddfdc37238
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:951b9d625bc300b5b72a3916cb382015527ea786f10b7acbeebf83ae854ca51a
        - quay.io/edge-infrastructure/assisted-service@sha256:0fa869095185759c26c391f0e353f293c903ea739576f160d6668f5f20404bdb
        - quay.io/edge-infrastructure/assisted-image-service@sha256:104d2aae313c724fb10d3ef3bba78b2370fc0597de31335453894b679a450ab8
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:4ca00219efbc0d9dcce415dadc1039e9495f3a3b00ab3bd575b41554d7934899
        - quay.io/edge-infrastructure/assisted-installer@sha256:803168aeb7e01db77e3b2ff7418ceea3d93daccaceeb95f8fc02027144ebac75
        version: 0.4.41
      entries:
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.6
    - currentCSV: assisted-service-operator.v0.4.50
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-x86_64-live.x86_64.iso",
                      "version": "411.86.202210041459-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-aarch64-live.aarch64.iso",
                      "version": "411.86.202210032347-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.12.0-ec.2/rhcos-4.12.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "412.86.202208101039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.12.0-ec.2/rhcos-4.12.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "412.86.202208101040-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:0ca1cfe04f088cd4c7318cb91978754d6472ada29e3436d05e22335740bfff49
          createdAt: "2022-11-23T19:17:47Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.4.50'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service@sha256:0ca1cfe04f088cd4c7318cb91978754d6472ada29e3436d05e22335740bfff49
        - quay.io/edge-infrastructure/assisted-image-service@sha256:71e8c9c76caa8fe01255fdb4a1172cda372969f7fe1399a3a54fb93e967e68cf
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:54b39c6cafefb70d258ec1d491b6a0f1dd888cd85bd442c8f40a4805cf4721ca
        - quay.io/edge-infrastructure/assisted-installer@sha256:f225c86242b0bfbf58753123ef33b7c07f367a5099641d19cfda5a7121382586
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:34e0341c0471b1772db11535107db06e511c8d5d9816edc3971e745928164ecd
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:c1c7281b9f06e56330cc55ac8d74185d42ee82dff105ae747db677b8eb059aea
        version: 0.4.50
      entries:
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.7
    - currentCSV: assisted-service-operator.v0.6.65
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-x86_64-live.x86_64.iso",
                      "version": "411.86.202210041459-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-aarch64-live.aarch64.iso",
                      "version": "411.86.202210032347-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-s390x-live.s390x.iso",
                      "version": "411.86.202210171842-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202210171844-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-x86_64-live.x86_64.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-aarch64-live.aarch64.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-s390x-live.s390x.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-x86_64-live.x86_64.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-aarch64-live.aarch64.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-s390x-live.s390x.iso",
                      "version": "413.92.202303281804-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:2dfaf2c3ac58b2987b0088df302c7cf717d0b6640cdb8b2327370aa110120784
          createdAt: "2023-05-24T11:15:57Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.6.65'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-image-service@sha256:b1ba1076fe16b2f868ed31a660b16583a957a83634c7ec3cf03574b56e02144c
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:1287d947464fbf5275feb31d106d2feb856c6117f0eaec8d947f2e03bd8618c5
        - quay.io/edge-infrastructure/assisted-installer@sha256:f8a7e7cb6ff1a4a7daa0cb26f0fab8da17f109653af5979c9d53c35b7ca2c1bb
        - quay.io/edge-infrastructure/assisted-service@sha256:2dfaf2c3ac58b2987b0088df302c7cf717d0b6640cdb8b2327370aa110120784
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:1ad91b13c763437863480df68f5b65576e95c558d4537886432a81c0b8d3cedc
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:a6f99547564cc2e047df74003019817b2dbd3dad992a0c8e53aaaba4f5b02ab6
        version: 0.6.65
      entries:
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.8
    - currentCSV: assisted-service-operator.v0.7.67
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-s390x-live.s390x.iso",
                      "version": "414.92.202305090606-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:fb0397a1402d42a64e3128e66a0817aa9466e67eb407cb792dda0d380bd5a189
          createdAt: "2023-10-08T07:14:20Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.67'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service@sha256:fb0397a1402d42a64e3128e66a0817aa9466e67eb407cb792dda0d380bd5a189
        - quay.io/edge-infrastructure/assisted-image-service@sha256:69ff087be8e5d8458a97f49d88b8b2025ee341d362dcf5618988c011b66688e0
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/edge-infrastructure/assisted-installer@sha256:d85be59bc379927f46d0a88ef04172c5c6733e31e1e4a8188de2d434d5bedb66
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:f2aeb4f7d193ea2c5702cae0f6039c02991b9f417d091468d46974408c97efe4
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:35c118d8c07dd72fe3307424cc051d9e0426e916412171449a05e41445b9f599
        version: 0.7.67
      entries:
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.9
    defaultChannel: alpha
    packageName: assisted-service-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Strimzi
      provider-url: ""
    name: strimzi-kafka-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: strimzi-cluster-operator.v0.48.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.1.0",
                      "metadataVersion": "4.1-IV1",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.1.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
          createdAt: "2025-09-25 10:32:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.48.0 supports only Kubernetes 1.27 and newer.**
            Kubernetes 1.25 and 1.26 are not supported anymore

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.48.0
          * Support for Kafka 4.1.0
          * Add monitoring of custom resources using kubernetes-state-metrics (KSM)
          * Support for type: custom client authentication
          * Added support for Strimzi Metrics Reporter to Kafka Connect, Mirror Maker 2 and Kafka Bridge.
          * Added support for KIP-745 in Kafka Connector and Mirror Maker 2, allowing the usage of includeTasks and onlyFailed arguments in Kafka connectors restart.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
        - quay.io/strimzi/kafka@sha256:34450afc1a3399a9026fc3c6ec90e9f6adfec9faf6520545b4d84e90cda34964
        - quay.io/strimzi/kafka-bridge@sha256:53034f64f0b672f10b5bacea1c7a25132f118df7fd5c9032c4dbf702ed93796a
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:64380efcb8da87fdc0e0de238a20456be9abb830174c2f2331882b7a84657e57
        - quay.io/strimzi/kafka@sha256:48db05ad2e9fa39eb388089e7e25654bcbf4da1ddc3a5635d675056c28bbcb96
        version: 0.48.0
      entries:
      - name: strimzi-cluster-operator.v0.48.0
        version: 0.48.0
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: stable
    - currentCSV: strimzi-cluster-operator.v0.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnectS2I",
                "metadata":{
                  "name":"my-connect-s2i-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "whitelist":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 2,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": 1,
                        "offset.storage.replication.factor": 1,
                        "status.storage.replication.factor": 1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "NetworkOutboundCapacityGoal",
                    "CpuCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:59d939f342f96d0e8a54c96fd187832c4b19cd4137bf55ace53d486728ae6c5f
          createdAt: "2021-05-14 11:37:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster with Source 2 Image support
            displayName: Kafka Connect Source to Image
            kind: KafkaConnectS2I
            name: kafkaconnects2is.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  Starting with Strimzi 0.23, we support only the `v1beta2` API version of all Strimzi custom resources.  This is a required for migration to `apiextensions/v1` which is needed because Kubernetes 1.22 will remove support for `apiextensions/v1beta1`.  Migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs **before upgrading to Strimzi 0.23**. For more details about the CRD upgrades, see the [documentation](https://strimzi.io/docs/operators/0.23.0/deploying.html#assembly-upgrade-resources-str).
          ### New in 0.23
          * Migrate to CRD v1 (required by Kubernetes 1.22+)
          * Strimzi API versions `v1alpha1` and `v1beta1` were removed from all Strimzi custom resources apart from `KafkaTopic` and `KafkaUser` (use `v1beta2` versions instead)
          * Add support for Kafka 2.8.0 and 2.6.2, remove support for Kafka 2.5.x
          * Make it possible to configure maximum number of connections and maximum connection creation rate in listener configuration
          * Add support for configuring finalizers for `loadbalancer` type listeners
          * Use dedicated Service Account for Kafka Connect Build on Kubernetes
          * Remove direct ZooKeeper access for handling user quotas in the User Operator. Add usage of Admin Client API instead.
          * Support for configuring custom Authorizer implementation
          * Changed Reconciliation interval for Topic Operator from 90 to 120 seconds (to keep it the same as for other operators)
          * Changed Zookeeper session timeout default value to 18 seconds for Topic and User Operators (for improved resiliency)
          * Removed requirement for replicas and partitions KafkaTopic spec making these parameters optional
          * Support to configure a custom filter for parent CR's labels propagation into subresources
          * Allow disabling service links (environment variables describing Kubernetes services) in Pod template
          * Add support for separate control plane listener (disabled by default, available via the `ControlPlaneListener` feature gate)
          * Support for Dual Stack networking
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _master_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/master/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/master/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:59d939f342f96d0e8a54c96fd187832c4b19cd4137bf55ace53d486728ae6c5f
        - quay.io/strimzi/kafka@sha256:699bff48ee445d8ba381a791cd0340015e3ecede8d169f41bc151095ab7fb5cc
        - quay.io/strimzi/kafka@sha256:d2745f48ed262290ce66c3362ffb9b9a8cfeb9c76beac40c995f72f3f06ad10e
        - quay.io/strimzi/kafka@sha256:dbc53bb374d2cb9513a5bb8f80fea751c91b249cabd36d8b3c04af5e5c961738
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/kafka@sha256:e90695b35d6dd5db20c9900dd0dab95d1ef908f3145b4f9f436b903303c88549
        - quay.io/strimzi/kafka-bridge@sha256:bed07186355a4c04ee50e90048e3bd3a023264093a9ef28f0278cce44640e93f
        - quay.io/strimzi/kafka@sha256:641f46114c59d61577bcb3f4ef5af16aea3ed42c2a575f8d3641c82848df85e8
        - quay.io/strimzi/jmxtrans@sha256:b4ec1844f4cc5eb74cd0aac83d74b81b977e178bf670d1136d1af0b68663d0a3
        version: 0.23.0
      entries:
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.23.x
    - currentCSV: strimzi-cluster-operator.v0.24.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnectS2I",
                "metadata":{
                  "name":"my-connect-s2i-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 2,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:786ea69c77e0651f009a73a8b3ce0b3d41b3cd71b917dc0d9bdc61c725f494a6
          createdAt: "2021-06-24 13:13:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster with Source 2 Image support
            displayName: Kafka Connect Source to Image
            kind: KafkaConnectS2I
            name: kafkaconnects2is.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: "Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org)
          cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/)
          in various deployment configurations. See our [website](https://strimzi.io)
          for more details about the project.\n### CRD Upgrades\n**!!! IMPORTANT !!!**
          \ This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.
          \ If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed
          for all Strimzi CRDs and CRs before the upgrade to 0.24 is done!  If upgrading
          from Strimzi version earlier than 0.22, you need to first install the CRDs
          from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi
          CRDs and CRs before the upgrade to 0.24 is done!  For more details about
          the CRD upgrades, see the documentation.\n### New in 0.24\n* Add support
          for Kafka 2.7.1 and remove support for 2.6.0, 2.6.1, and 2.6.2\n* Add support
          for [Kubernetes Configuration Provider for Apache Kafka](https://github.com/strimzi/kafka-kubernetes-config-provider)\n*
          Use Red Hat UBI8 base image\n* Support for patching of service accounts
          and configuring their labels and annotations. The feature is disabled by
          default and enabled using the new `ServiceAccountPatching` feature gate.\n*
          Added support for configuring cluster-operator's worker thread pool size
          that is used for various sync and async tasks\n* Add Kafka Quotas plugin
          with produce, consume, and storage quotas\n* Support pausing reconciliation
          of KafkaTopic CR with annotation `strimzi.io/pause-reconciliation`\n* Update
          cruise control to 2.5.57\n* Update to Strimzi Kafka Bridge to 0.20.1\n*
          Support for broker load information added to the rebalance optimization
          proposal. Information on the load difference, before and after a rebalance
          is stored in a ConfigMap\n* Add support for selectively changing the verbosity
          of logging for individual CRs, using markers.\n* Added support for `controller_mutation_rate'
          quota. Creation/Deletion of topics and creation of partitions can be configured
          through this.\n* Use newer version of Kafka Exporter with different bugfixes
          \n### Supported Features\n* **Manages the Kafka Cluster** - Deploys and
          manages all of the components of this complex application, including dependencies
          like Apache ZooKeeper® that are traditionally hard to administer.\n* **Includes
          Kafka Connect** - Allows for configuration of common data sources and sinks
          to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for morroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashabords\n### Upgrading your Clusters\nThe
          Strimzi Operator understands how to run and upgrade between a set of Kafka
          versions. When specifying a new version in your config, check to make sure
          you aren't using any features that may have been removed. See [the upgrade
          guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str)
          for more information.\n### Storage\nAn efficient data storage infrastructure
          is essential to the optimal performance of Apache Kafka®. Apache Kafka®
          deployed via Strimzi requires block storage. The use of file storage (for
          example, NFS) is not recommended.\nThe Strimzi Operator supports three types
          of data storage:\n* Ephemeral (Recommended for development only!)\n* Persistent\n*
          JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nStrimzi
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation to the current _main_
          branch as well as all releases can be found on our [website](https://strimzi.io/documentation).\n###
          Getting help\nIf you encounter any issues while using Strimzi, you can get
          help using:\n* [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          Contributing\nYou can contribute by:\n* Raising any issues you find using
          Strimzi\n* Fixing issues by opening Pull Requests\n* Improving documentation\n*
          Talking about Strimzi\n\nAll bugs, tasks or enhancements are tracked as
          [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues).
          Issues which  might be a good start for new contributors are marked with
          [\"good-start\"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start)
          label.\n\nThe [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md)
          describes how to build Strimzi and how to  test your changes before submitting
          a patch or opening a PR.\n\nThe [Documentation Contributor Guide](https://strimzi.io/contributing/guide/)
          describes how to contribute to Strimzi documentation.\n\nIf you want to
          get in touch with us first before contributing, you can use:\n* [Strimzi
          mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          License\nStrimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).\n"
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/kafka@sha256:95cfe9000afda2f7def269ca46472d3803ee85146c521884884d8505a7187daf
        - quay.io/strimzi/kafka@sha256:8959b7968ab8b3306906cdbff2ebb8d63329af37e58124a601843795c4ef5bd6
        - quay.io/strimzi/kafka@sha256:fbb08410d9595029bc4a02ed859971264e6ce2dc85dd6a9855eaa7bb58b52a25
        - quay.io/strimzi/kafka-bridge@sha256:537ea10e87619251204747ebc1005c632a6c6ed270f5ef06f86c0c3ae4a7553b
        - quay.io/strimzi/operator@sha256:786ea69c77e0651f009a73a8b3ce0b3d41b3cd71b917dc0d9bdc61c725f494a6
        - quay.io/strimzi/jmxtrans@sha256:2c54adabf6f31c087dbb898057f9c5e200aaf4c5b39fb3a47bbaf7f6587df72d
        version: 0.24.0
      entries:
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.24.x
    - currentCSV: strimzi-cluster-operator.v0.25.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:afaf1dd1f2f696ae80674d74159661ea239ae609f78a73a1423c7a17a39e5044
          createdAt: "2021-06-24 13:13:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.25 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.25 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.25
          * Move from Scala 2.12 to Scala 2.13
          * Open Policy Agent authorizer updated to a new version supporting Scala 2.13.
            The new version of the authroizer uses different format of the input data sent to the OPA server and might require you to update your policies.
            See [release notes](https://github.com/strimzi/strimzi-kafka-operator/releases/tag/0.25.0) for more details.

          * Allow a custom password to be set for SCRAM-SHA-512 users by referencing a secret in the `KafkaUser` resource
          * Add support for [EnvVar Configuration Provider for Apache Kafka](https://github.com/strimzi/kafka-env-var-config-provider)
          * Add support for `tls-external` authentication to User Operator to allow management of ACLs and Quotas for TLS users with user certificates generated externally
          * Support for disabling the automatic generation of network policies by the Cluster Operator
          * Configure fixed size limit for `emptyDir` volumes used for temporary files
          * Update Strimzi Kafka Bridge to 0.20.2
          * The `KafkaConnectS2I` resource has been removed and is no longer supported by the operator.
            Please use the [migration guide](https://strimzi.io/docs/operators/0.24.0/full/using.html#proc-migrating-kafka-connect-s2i-str) to migrate your `KafkaConnectS2I` deployments to [`KafkaConnect` Build](https://strimzi.io/docs/operators/latest/full/deploying.html#creating-new-image-using-kafka-connect-build-str) instead.

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:d8991608ec71f17e850c33714ee02ca87ba94bc0f077be62ad1ed127ca244823
        - quay.io/strimzi/kafka@sha256:ee17eb3611019b0cfa5e9e196610edc1dbd7084471fb20990f055c89da8eef45
        - quay.io/strimzi/kafka@sha256:56a33815f7d50a97ea9bb4cce1e3a6d04d918b841b5aa285f16b792d8930059e
        - quay.io/strimzi/kafka-bridge@sha256:a005e20d7fbe04d6f587f2093e21a00556ec62b850df757b9f7d9d57e3141c40
        - quay.io/strimzi/jmxtrans@sha256:af5eb8459ac9a0acbca51e67c6e0217f62ba404cba9f2251968c927dcef75252
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/operator@sha256:afaf1dd1f2f696ae80674d74159661ea239ae609f78a73a1423c7a17a39e5044
        version: 0.25.0
      entries:
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.25.x
    - currentCSV: strimzi-cluster-operator.v0.26.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.0.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"3.0",
                        "inter.broker.protocol.version":"3.0"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.0.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:5d90caa66f57939ba62a9a50efa00fce319ba596e719a3f8a42170d7635ee3f4
          createdAt: "2021-12-14 00:16:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.26 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.26 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.26.1
          * **Updated Log4j2 to fix CVE-2021-44228**
          ### New in 0.26.0
          * Add support for Kafka 2.8.1 and 3.0.0; remove Kafka 2.7.0 and 2.7.1
          * Add support for exposing JMX in Zookeeper
          * Allow configuring labels and annotations for JMX authentication secrets
          * Enable Cruise Control anomaly.detection configurations
          * Add support for building connector images from the Maven coordinates
          * Allow Kafka Connect Build artifacts to be downloaded from insecure servers
          * Add option to specify pull secret in Kafka Connect Build on OpenShift
          * Configurable authentication, authorization, and SSL for Cruise Control API
          * Allow to configure `/tmp` volume size via Pod template. By default `1Mi` is used.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:d3899c79747e4cb3a6df5a9fcd80b81f038ca0634a97329c1d160499469876dd
        - quay.io/strimzi/jmxtrans@sha256:bff5d3bab575d0b9ed95dd79d2d22518cc00c4d1c9dbb5614ed16149d4920a1b
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/maven-builder@sha256:663eb81388ae8f824e7920c272f6d2e2274cf6c140d61416607261cdce9d50e2
        - quay.io/strimzi/kafka@sha256:8c022a8535742b2eea691737f863af5ddef1e8df08b415b8139ecd02a79052a5
        - quay.io/strimzi/kafka@sha256:bd22a134b0c351ddfc78a3263850c2f16733c52f57ac92984cbe0ffff477c724
        - quay.io/strimzi/operator@sha256:5d90caa66f57939ba62a9a50efa00fce319ba596e719a3f8a42170d7635ee3f4
        - quay.io/strimzi/kafka@sha256:6367b0fbba9593e3ac736040ea2d7ef71296b1ec144d3b54d9370352141b90e4
        version: 0.26.1
      entries:
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.26.x
    - currentCSV: strimzi-cluster-operator.v0.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.0.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"3.0",
                        "inter.broker.protocol.version":"3.0"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.0.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:9ea49407a04006d0bddfaaee822adfb95f5857af32ffcee81ba60ce891b07ba8
          createdAt: "2022-01-14 20:00:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.27 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.27 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.27.1
          * Updated Log4j2 to version 2.17.1 and Fabric8 Kubernetes Client to 5.10.2
          ### New in 0.27.0
          * Updated Log4j2 to version 2.17.0 (adresses CVE-2021-44228, CVE-2021-45046, CVE-2021-45105)
          * (Experimental) support for AArch64 / ARM64 architecture
          * `ControlPlaneListener` and `ServiceAccountPatching` feature gates are now in the beta phase and are enabled by default
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a07ebe6600d6baba09980abdde2931b3b6abd2f75aeb2cac5c6a2d22a37aa807
        - quay.io/strimzi/maven-builder@sha256:9acd178cc504be8acaa70a39f3c17830aae1f0e93a656110790f9476b09260d5
        - quay.io/strimzi/kafka@sha256:f4d68bb94447c6612f70de7f6587e9e7ef712c83769ea1f11a8dbda0d241a059
        - quay.io/strimzi/kafka@sha256:a3ec567ca529df2fa41f8986d21e86dde66be94aa4baa56faa66db4dcb3b258e
        - quay.io/strimzi/kafka@sha256:5682a6bfcc4ee13d53a30d587c8f96bc40abe3618880989336c5af274ea3cefa
        - quay.io/strimzi/kafka-bridge@sha256:55a52316aead02efa2004174e11ad4728bb8b6f1c1415f725f597f1d2f2da188
        - quay.io/strimzi/operator@sha256:9ea49407a04006d0bddfaaee822adfb95f5857af32ffcee81ba60ce891b07ba8
        - quay.io/strimzi/jmxtrans@sha256:d2d2477517eb1ac98b51ad6d64d8c1c578141637afa8d46037076f0013183c0b
        version: 0.27.1
      entries:
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.27.x
    - currentCSV: strimzi-cluster-operator.v0.28.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.1.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.1"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.1.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.1.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:8b010c5fb8244b742a37f264feefc116d75bb2abd2d6895375cbeea250727b39
          createdAt: "2022-02-12 23:02:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.28 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.28 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.28.0
          * Support for Kafka 3.1.0
          * Support for `StrimziPodSet` resources (disabled by default through the `UseStrimziPodSets` feature gate)
          * Support custom authentication mechanisms in Kafka listeners
          * Intra-broker disk balancing using Cruise Control
          * Add support for disabling the FIPS mode in OpenJDK
          * Fix renewing your own CA certificates
          * Added the option `createBootstrapService` to disable the creation of the bootstrap service for the Load Balancer Type Listener
          * Dependency updates (Strimzi Kafka Bridge, Strimzi OAuth, Open Policy Agent Authroizer plugin and others)
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a07ebe6600d6baba09980abdde2931b3b6abd2f75aeb2cac5c6a2d22a37aa807
        - quay.io/strimzi/maven-builder@sha256:3d5d664e089651944712361a6d8ba42b578dbd9bb3dfd9972f614520c9cfab75
        - quay.io/strimzi/kafka@sha256:7c04f65e216bb941458aeadffb85e9eeee43d5ca67da566017c2c93a0664c98e
        - quay.io/strimzi/kafka@sha256:06e9bda15495feb2b92d9950e834cdeed6ef20064f021a879610f1b59dd46950
        - quay.io/strimzi/kafka-bridge@sha256:c345c783d8fbbee4edeca2ebc47e7c454fcc2e26686820b8037de97369280ba0
        - quay.io/strimzi/jmxtrans@sha256:3a55a81790dc88b90d1bf858e2cfd7e591ad0e79f763c6d1667b2d3eb0d0c79d
        - quay.io/strimzi/operator@sha256:8b010c5fb8244b742a37f264feefc116d75bb2abd2d6895375cbeea250727b39
        version: 0.28.0
      entries:
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.28.x
    - currentCSV: strimzi-cluster-operator.v0.29.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:afde94c6d2544207cb7f3c94b54455643558a9d5daaa13b9f736cc8da20e319c
          createdAt: "2022-05-15 00:01:07"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.29 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.29 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.29.0
          * Add support for Apache Kafka 3.0.1, 3.1.1 and 3.2.0
          * Renew user certificates in User Operator only during maintenance windows
          * Remove TLS sidecar from Cruise Control pod. Cruise Control is now configured to not using ZooKeeper, so the TLS sidecar is not needed anymore.
          * Allow Cruise Control topic names to be configured
          * Add support for `spec.rack.topologyKey` property in Mirror Maker 2 to enable "fetch from the closest replica" feature.
          * Experiemntal support for the s390x platform
          * Added new rebalancing modes on the `KafkaRebalance` custom resource to add or remove brokers
          * Experimental KRaft mode (ZooKeeper-less Kafka). Use it for development and testing only!
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:544c4987f7f3ca558b04e9ef4357be204f088212b7b5b9ed0f909401a8bc5238
        - quay.io/strimzi/maven-builder@sha256:0c096a59faf8fd224298c01415a895043db4647534c060cb7e53207e9448bf65
        - quay.io/strimzi/operator@sha256:afde94c6d2544207cb7f3c94b54455643558a9d5daaa13b9f736cc8da20e319c
        - quay.io/strimzi/kafka@sha256:95678e843bc87cbe571f3f299045dd075725816b59e1b87821cd6025876ec559
        - quay.io/strimzi/kafka@sha256:935caab03f1d0365b47ddefec0c1d8fa449fc3de715ccdd298aedef5ce968c73
        - quay.io/strimzi/kafka@sha256:ac190141120c1800cc7131d3a1f701e64efdcd4a5d87f0b2e0e2cadfdef8b438
        - quay.io/strimzi/kafka@sha256:9dfc09273ad8fc4bdc75cbe93b4a061a05a5defb5884a221ae1e178ba602787f
        - quay.io/strimzi/kafka@sha256:284cde40784b6ff7bfb81b6dd28928e1e8c598fe6a72504f025592e5b43fa94c
        - quay.io/strimzi/kafka-bridge@sha256:b7bc3dd90e87223766f608a0087b521f0bba79d1c21eaa833bc2708c71148b7f
        - quay.io/strimzi/jmxtrans@sha256:f89e3c9618430908a8b390f2490f5aa43687b12889b434472a962ce0170c6463
        version: 0.29.0
      entries:
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.29.x
    - currentCSV: strimzi-cluster-operator.v0.30.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:587db6f4389f56209f48853beedbfe249f6d590a4cb0000be2c42b57e43c2adb
          createdAt: "2022-07-16 00:00:10"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: "Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org)
          cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/)
          in various deployment configurations. See our [website](https://strimzi.io)
          for more details about the project.\n### CRD Upgrades\n**!!! IMPORTANT !!!**
          \ This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.
          \ If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed
          for all Strimzi CRDs and CRs before the upgrade to 0.30 is done!  If upgrading
          from Strimzi version earlier than 0.22, you need to first install the CRDs
          from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi
          CRDs and CRs before the upgrade to 0.30 is done!  For more details about
          the CRD upgrades, see the documentation.\n### New in 0.30.0\n* The `UseStrimziPodSets`
          feature gate moves to beta stage. By default, StrimziPodSets are used instead
          of StatefulSets. If needed, `UseStrimziPodSets` can be disabled in the feature
          gates configuration in the Cluster Operator.\n* The `ServiceAccountPatching`
          feature gate moves to GA. It cannot be disabled anymore and will be permanently
          enabled.\n* Remove support for Kafka 3.0.0 and 3.0.1\n* Add support for
          `simple` authorization and for the User Operator to the experimental `UseKRaft`
          feature gate. _(Note: Due to [KAFKA-13909](https://issues.apache.org/jira/browse/KAFKA-13909),
          broker restarts currently don't work when authorization is enabled.)_\n*
          Add CPU and network capacity overrides for Cruise Control capacity config\n*
          Use better encryption and digest algorithms when creating the PKCS12 stores.
          For existing clusters, the certificates will not be updated during upgrade
          but only next time the PKCS12 store is created. \n* Operator emits Kafka
          events to explain why it restarted a Kafka broker\n* Better configurability
          of the Kafka Admin client in the User Operator\n* Update Strimzi Kafka Bridge
          to 0.21.6\n### Supported Features\n* **Manages the Kafka Cluster** - Deploys
          and manages all of the components of this complex application, including
          dependencies like Apache ZooKeeper® that are traditionally hard to administer.\n*
          **Includes Kafka Connect** - Allows for configuration of common data sources
          and sinks to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for mirroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashboards\n### Upgrading your Clusters\nThe
          Strimzi Operator understands how to run and upgrade between a set of Kafka
          versions. When specifying a new version in your config, check to make sure
          you aren't using any features that may have been removed. See [the upgrade
          guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str)
          for more information.\n### Storage\nAn efficient data storage infrastructure
          is essential to the optimal performance of Apache Kafka®. Apache Kafka®
          deployed via Strimzi requires block storage. The use of file storage (for
          example, NFS) is not recommended.\nThe Strimzi Operator supports three types
          of data storage:\n* Ephemeral (Recommended for development only!)\n* Persistent\n*
          JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nStrimzi
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation to the current _main_
          branch as well as all releases can be found on our [website](https://strimzi.io/documentation).\n###
          Getting help\nIf you encounter any issues while using Strimzi, you can get
          help using:\n* [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          Contributing\nYou can contribute by:\n* Raising any issues you find using
          Strimzi\n* Fixing issues by opening Pull Requests\n* Improving documentation\n*
          Talking about Strimzi\nAll bugs, tasks or enhancements are tracked as [GitHub
          issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues
          which  might be a good start for new contributors are marked with [\"good-start\"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start)
          label.\nThe [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md)
          describes how to build Strimzi and how to  test your changes before submitting
          a patch or opening a PR.\nThe [Documentation Contributor Guide](https://strimzi.io/contributing/guide/)
          describes how to contribute to Strimzi documentation.\nIf you want to get
          in touch with us first before contributing, you can use:\n* [Strimzi mailing
          list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n* [Strimzi
          Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          License\nStrimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).\n"
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:544c4987f7f3ca558b04e9ef4357be204f088212b7b5b9ed0f909401a8bc5238
        - quay.io/strimzi/maven-builder@sha256:f2cb45234f3a7a010824a370d03032b706271cd0d1d260da929f046f9259d7e0
        - quay.io/strimzi/kafka@sha256:68d6f4cb5cc5ccad44de799bd9cf417c469a09e1024dbe42298942f25a09e004
        - quay.io/strimzi/kafka@sha256:d9258ddd453c3a53ca8e737b989d27725818da95f7a4fc879c9f07fe743c0cef
        - quay.io/strimzi/kafka@sha256:5d881605425ab0b37d0b9cac6957d00ba904df0db185ca6181c810151bdab8e7
        - quay.io/strimzi/operator@sha256:587db6f4389f56209f48853beedbfe249f6d590a4cb0000be2c42b57e43c2adb
        - quay.io/strimzi/kafka-bridge@sha256:77fc68da171c74d3a4b75b1030eaaa5eae12e200002f7aa2ee1969385eba1d66
        - quay.io/strimzi/jmxtrans@sha256:de1ca88b713b926b72df535aca2cfad24fb1b2bf40481065d6fa0c421b872320
        version: 0.30.0
      entries:
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.30.x
    - currentCSV: strimzi-cluster-operator.v0.31.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.3",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.3",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:6e5e207c70c25c595d3ce60268c0214c7518e95836c9b1212ef1eb2f008c7a76
          createdAt: "2022-09-20 21:07:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          **This is the last version of Strimzi with support for Kubernetes 1.16, 1.17 and 1.18!**
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.31 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.31 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.31.1
          * Kafka 3.1.2 and 3.2.3 (fixes CVE-2022-34917)
          * Make `sasl.server.max.receive.size` broker option user configurable
          * Update Strimzi Kafka Bridge to 0.22.1
          ### New in 0.31.0
          * Support for Apache Kafka 3.2.1
          * Pluggable Pod Security Profiles with built-in support for _restricted_ Kubernetes Security Profile
          * Support for leader election and running multiple operator replicas (1 active leader replicas and one or more stand-by replicas)
          * Support for IPv6 addresses being used in Strimzi issued certificates
          * `StrimziPodSet` reconciliation metrics
          * Updated Strimzi Kafka Bridge to 0.22.0 and Kafka Exporter to 1.6.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:03cc48bc5cf367f898338665b253855bed7d6096fe577adfe344da2f2cddcfe4
        - quay.io/strimzi/operator@sha256:6e5e207c70c25c595d3ce60268c0214c7518e95836c9b1212ef1eb2f008c7a76
        - quay.io/strimzi/kaniko-executor@sha256:06581fb615f0b13ddfe27fae26ba5836b12c6e28afd55a81cba7685c9ca4bf79
        - quay.io/strimzi/maven-builder@sha256:670ca70900f2229ef140a592b77f2107934601f8b050b79250eab17ff62099b6
        - quay.io/strimzi/kafka-bridge@sha256:3ddac62a4f8c8a3ebb72d5841d3826f7fd3fbb0e00eb75fb40bced0a576b86a9
        - quay.io/strimzi/kafka@sha256:b7c059bd72d8b063a96dde788d91d1ad87498bad3e73acba600929d595aeb7d6
        - quay.io/strimzi/kafka@sha256:430b3eb70e47a91d7b5b6f230d5be024c2de853099cc0437e219b443e0ead9f5
        - quay.io/strimzi/kafka@sha256:9553aae22d35448681cbbf7a831a28149fb159f3df58ccf8c8ec4ca7eb9c1fdf
        - quay.io/strimzi/kafka@sha256:83acc1e2cc1e64a88bdbb54b9818a6ad3658c3da60e72c4be365f824c5fcf57c
        - quay.io/strimzi/jmxtrans@sha256:52dd107062d78c29bb58c905ac654c39df72bbe3c25613023aa33e834930606d
        - quay.io/strimzi/kafka@sha256:652924a1dba46aa716caba8381f1e791c09035e31ce62fdcc81f9b69af36cd02
        version: 0.31.1
      entries:
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.31.x
    - currentCSV: strimzi-cluster-operator.v0.32.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.3.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.3.1",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:c5e0e5dca7503a8371b8d366db0f49e3fc68f62ea480262e39b6cc277991c97c
          createdAt: "2022-10-31 17:05:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **From the 0.32.0 release, Strimzi supports only Kubernetes 1.19 and newer! Kubernetes 1.16, 1.17 and 1.18 are not supported anymore.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.32.0/deploying.html#assembly-upgrade-str).
          ### New in 0.32.0
          * Add support for Kafka 3.3.1 and remove support for Kafka 3.1.0, 3.1.1, and 3.1.2
          * Update KafkaConnector CR status so the 'NotReady' condition is added if the connector or any tasks are reporting a 'FAILED' state.
          * Add auto-approval mechanism on KafkaRebalance resource when an optimization proposal is ready
          * The ControlPlaneListener feature gate moves to GA
          * Add client rack-awareness support to Strimzi Bridge pods
          * Add support for OpenTelemetry for distributed tracing
          * ZookeeperRoller considers unready pods
          * Support multiple operations per ACLRule
          * Add new listener type "cluster-ip" which is using per-broker services instead of the pod DNS names
          * Update Strimzi OAuth library to 0.11.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:8c9f38b45dd814e617c9c51cc90fd51e23d7dd23e8efe7e9eff013522bd6a5a0
        - quay.io/strimzi/kafka@sha256:bb97008b7f928ed98939b2008a7baf8909c88e726a59aaf607f2eeba55375cc0
        - quay.io/strimzi/kafka@sha256:680ae1958dbcb9da8ee4128a67c1163a6ee2744221f7d29f73d8bcc237fd0173
        - quay.io/strimzi/maven-builder@sha256:593e9980b786ca8ad3223151c3e17f16c7652ee0e86d722af4c34b023038ecc2
        - quay.io/strimzi/operator@sha256:c5e0e5dca7503a8371b8d366db0f49e3fc68f62ea480262e39b6cc277991c97c
        - quay.io/strimzi/jmxtrans@sha256:208928133b2854f60bb34912e07561a3511de895b815960e73d565be2d553918
        - quay.io/strimzi/kafka@sha256:4bacf5b54041bd0ab9199c8c5d51e8058758139db4b7abc380cf0400737716ef
        - quay.io/strimzi/kafka-bridge@sha256:aad0cee4381cc1c8c6638ab5732b135acd271b5889486ca9a33312cce67e4a02
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        version: 0.32.0
      entries:
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.32.x
    - currentCSV: strimzi-cluster-operator.v0.33.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:cf0d41fdb99b20e91063e23eb1d664817c18866c9a2f5c4430eac897e78d5d53
          createdAt: "2023-02-18 21:21:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.33 supports only Kubernetes 1.19 and newer! Kubernetes versions 1.16, 1.17 and 1.18 are not supported anymore since Strimzi 0.32.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.33.2/deploying.html#assembly-upgrade-str).
          ### New in 0.33
          _Note: Strimzi 0.33.2 add support for Kafka 3.4.0 which fixes CVE-2023-25194._
          * Add support for Kafka 3.3.2
          * Automatic restarts of failed Connect or Mirror Maker 2 connectors
          * Redesign of Strimzi User Operator to improve its scalability
          * Move to Java 17
          * Improved FIPS support
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/maven-builder@sha256:de9e51365e7acc63feae745618804e6286fd730f41c3b31582cbb7319a93b483
        - quay.io/strimzi/kafka@sha256:116cf543fbadadd0b45007ce24d44a0c79a62280e8faa1fc80df40291fbdf05d
        - quay.io/strimzi/kafka@sha256:9eca149cfc436580e3cdd6cee0ca318366b8d0ec53c2cf007e4e2c614b988921
        - quay.io/strimzi/kafka@sha256:301dc52ed5891ed4fa33e79ca3c344b1ac0eec7469dd46a5a7ac22064e053fc2
        - quay.io/strimzi/kafka-bridge@sha256:3a380d5f3ea509f90cdf95280d83b42f5de4f86ee3a512e02234e7639f50c9eb
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/kafka@sha256:30945aecb9d63cf8863c672222f130469c89aefa07078c95aaed52c40fd16146
        - quay.io/strimzi/jmxtrans@sha256:5ebdafa7a5e3718ed3274d6254c9e655613c272340094b997b1c97fde2598b4a
        - quay.io/strimzi/operator@sha256:cf0d41fdb99b20e91063e23eb1d664817c18866c9a2f5c4430eac897e78d5d53
        - quay.io/strimzi/kafka@sha256:fe9d3f8b029ff32d5d725e57384592914957314d63e92fda3abf4028fe4e66f6
        - quay.io/strimzi/kafka@sha256:40a847f8455ec6ea9ff9fe1fff20b1ed01f848326dc13977d13ac911f0cb40b7
        version: 0.33.2
      entries:
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.33.x
    - currentCSV: strimzi-cluster-operator.v0.34.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:f4ee3ca482a2dd19785f59634578a0850273444c1173a5699a99cf02f3a018fe
          createdAt: "2023-03-17 13:02:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.34 supports only Kubernetes 1.19 and newer! Kubernetes versions 1.16, 1.17 and 1.18 are not supported anymore since Strimzi 0.32.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.34.0/deploying.html#assembly-upgrade-str).
          ### New in 0.34
          * Add support for Apache Kafka 3.4.0 (fixes CVE-2023-25194) and remove support for Apache Kafka 3.2.x
          * Stable Pod identities for Kafka Connect and MirrorMaker 2 (disabled by default)
          * Use JDK HTTP client in the Kubernetes client instead of the OkHttp client
          * Add truststore configuration for HTTPS connections to Open Policy Agent server
          * Update Strimzi HTTP Bridge to 0.25.0 and strimzi Oauth to 0.12.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:37cf8d0fea1d2078511920190a72b49977ff4047b6ecbfed6b394c6f9e1e2ff0
        - quay.io/strimzi/kafka-bridge@sha256:51babca0f7a328f4db005c6a16daa8ac4a4306c46fe3ca89382919134aa09179
        - quay.io/strimzi/jmxtrans@sha256:62272e8a03da2144340bd7267f7a824d1ed70f4f3e30dc7aadfba9ee1b9baf76
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/maven-builder@sha256:f781b6b857d437c2de62585462e97b9fae357912c9f6711c1df39f431278878d
        - quay.io/strimzi/operator@sha256:f4ee3ca482a2dd19785f59634578a0850273444c1173a5699a99cf02f3a018fe
        - quay.io/strimzi/kafka@sha256:1da35fd7b36b113d1343e254fdc6efd97ca8660cc7977d8985aa3a3547d5ebb1
        - quay.io/strimzi/kafka@sha256:1ace2cc0db9653ac1b388477fa56c3c1cd97c7b3846eba3c7824dc58ea179476
        version: 0.34.0
      entries:
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.34.x
    - currentCSV: strimzi-cluster-operator.v0.35.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:06a94a3021cf028ccc1a49271f35f79216029e344536e664f196c1725ff2c663
          createdAt: "2023-06-07 16:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.35 is the last version with support for Kubernetes 1.19 and 1.20. From Strimzi 0.36, only Kubernetes 1.21 and newer will be supported.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.35.1/deploying.html#assembly-upgrade-str).
          ### New in 0.35
          * Move feature gate `UseStrimziPodSets` to GA and remove support for StatefulSets
          * Remove support for JMX Trans
          * Redesigned the Cluster and User Operator configuration to make it more efficient and flexible
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:39778b90c2b2afc30261e4ad5135805e1a10a2b60e2e53108fb9f80487f1208a
        - quay.io/strimzi/maven-builder@sha256:88a79eff3b3a386880a630658964b7754caed9e99dd6e645a4c0d23d0fdb47ee
        - quay.io/strimzi/operator@sha256:06a94a3021cf028ccc1a49271f35f79216029e344536e664f196c1725ff2c663
        - quay.io/strimzi/kafka@sha256:4de4874a7b722ad813f4dcc58acf509527bca0609999b81e70d81e3b38534d9d
        - quay.io/strimzi/kafka@sha256:0d910e7138cb49e1cd8cd84cef88bce35698b93ddd683a3398f1d485a3162693
        - quay.io/strimzi/kafka@sha256:54c6b25b31f51ef401c1b6e2a1b27432911e819bf3e502e05186f01be3f798e5
        - quay.io/strimzi/kafka-bridge@sha256:d6be183e492f8f88157ab9fe0af53950df8b6711a8a8c33da465de6064f6f86e
        version: 0.35.1
      entries:
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.35.x
    - currentCSV: strimzi-cluster-operator.v0.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:e9e03b31007c0f9b67a6a33e6b644e8c1ba35bd9976804ee3d22a04e4bf511b1
          createdAt: "2023-07-10 21:47:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!! Strimzi 0.36.1 supports only Kubernetes 1.21 and newer! Kubernetes versions 1.19 and 1.20 are not supported anymore since Strimzi 0.36.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.35.1/deploying.html#assembly-upgrade-str).
          ### New in 0.36
          * Add support for Apache Kafka 3.4.1, 3.5.0, and 3.5.1
          * Remove support for 3.3.1 and 3.3.2
          * Add support for _Kafka node pools_
          * Add support for _Unidirectional Topic Operator_
          * Enable SCRAM-SHA authentication in KRaft mode (supported in Apache Kafka 3.5.0 and newer)
          * Add support for insecure flag in Maven artifacts in Kafka Connect Build
          * Improve Kafka rolling update to avoid rolling broker in log recovery
          * Added support for Kafka Exporter topic exclude and consumer group exclude parameters
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:9f35f40be3d47b0cb1f6628f7ffd9c6a687ab9f564829f341dd66fd0759c064f
        - quay.io/strimzi/kafka@sha256:67c2b563a65c85c9393608f9af7eaf12ea48dcc0ef5f726d51f27273a3845186
        - quay.io/strimzi/operator@sha256:e9e03b31007c0f9b67a6a33e6b644e8c1ba35bd9976804ee3d22a04e4bf511b1
        - quay.io/strimzi/kafka@sha256:ea35817ff1ef295c682d70562f209ef658fcd87c61f467a4549d7a2f39e71700
        - quay.io/strimzi/kafka-bridge@sha256:fefee252895e9f94756ed5c9ea71749e3fa89281df01bc55878a244ca7d17697
        - quay.io/strimzi/kaniko-executor@sha256:5ff9fc7d0b7b694875ad0150fc787c582d42da3b4864d807ee8a6524e47b0e4e
        - quay.io/strimzi/maven-builder@sha256:2a469b40122027b981914db5b5c102be48280e039c3089963422c181174af226
        - quay.io/strimzi/kafka@sha256:a0da6a48b5bcbff355365f07df2d99ef6443d22c5690495f10a6417e761b3799
        version: 0.36.1
      entries:
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.36.x
    - currentCSV: strimzi-cluster-operator.v0.37.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:52f376e64b9b31d74162e26248c278e9bd301d2d8b4240ee5bcf780dfe2f8ef8
          createdAt: "2023-09-03 17:45:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.37.0/deploying.html#assembly-upgrade-str).
          ### New in 0.37
          * The StableConnectIdentites feature gate moves to a beta stage.
          * Support for the ppc64le platform
          * Version fields to the Kafka custom resource status to track installation and upgrade state
          * Infinite auto-restarts of Kafka Connect and Kafka Mirror Maker 2 connectors
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:1de6e073e8625a193fcfd11dd1a6717cc17596911c10bcebd792c00b08a7ebaf
        - quay.io/strimzi/kafka@sha256:2997ba1c169eb792a6004d7a41b674560515f3ea6e62ab8b9a1fdb747eab40c1
        - quay.io/strimzi/kafka@sha256:dfeb1eb7b94107c443b6bc82fcfee7fcf782516c83c6fc64451fb6a7e2de74a6
        - quay.io/strimzi/kafka-bridge@sha256:5b821f6f3396002ffa7c5804391b627ff9da3b494b2777dde518c55c2b73cefc
        - quay.io/strimzi/kaniko-executor@sha256:5729f1f0511da95e64e358cfabb3980ad9a660af2968118a87d6755f5d652a67
        - quay.io/strimzi/maven-builder@sha256:995be06a50b86c992fcee9f00e44f6a21c3406975e4e35802f10e88b72ffb9df
        - quay.io/strimzi/operator@sha256:52f376e64b9b31d74162e26248c278e9bd301d2d8b4240ee5bcf780dfe2f8ef8
        - quay.io/strimzi/kafka@sha256:38bbadcdb196a373e1698be890caa1dc01134ce91763d8f95c71768a117e1932
        version: 0.37.0
      entries:
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.37.x
    - currentCSV: strimzi-cluster-operator.v0.38.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:aee45bbe439335e45a0423e6565bae0d0727d85c7b807812949bfa127d467fe6
          createdAt: "2023-10-14 18:18:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.38.0/deploying.html#assembly-upgrade-str).
          ### New in 0.38
          * Add support for Apache Kafka 3.6.0 and remove support for 3.4.0 and 3.4.1
          * Container signing and SBOMs
          * Stopping Connect and MirrorMaker 2 connectors
          * Manual rolling updates of Kafka Connect and Kafka Mirror Maker 2 pods using annotation
          * Prevent scale-down of brokers that still contain any replicas
          * Run ZooKeeper- and KRaft-based clusters in parallel
          * Unidirectional Topic Operator improvements
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:0cc01438e8596643407b6b8ff05fafcea7ef54fa9410d4ae2492a9dbf41118b1
        - quay.io/strimzi/kafka@sha256:20d451fb32ec82eab3c16891198ace46bb9b9193f1f5511bb60ce249080dd7d4
        - quay.io/strimzi/kafka-bridge@sha256:13838889cc20a3a382ea64e749cad94f997c0b928840588f2d1befcf19247bea
        - quay.io/strimzi/kaniko-executor@sha256:4fe7a3f3f78e93369284a5ab456873c911230926ca071d1d88167e63d8ce376c
        - quay.io/strimzi/maven-builder@sha256:922d91dab9e2b8e14b6db25d8b48d7f4e3521bedd435468cc22039599ac4f888
        - quay.io/strimzi/operator@sha256:aee45bbe439335e45a0423e6565bae0d0727d85c7b807812949bfa127d467fe6
        - quay.io/strimzi/kafka@sha256:50e24c797455f9cdbb3cdde8e4f26bb8e1f0503469456da5d179017fd3e03ffe
        version: 0.38.0
      entries:
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.38.x
    - currentCSV: strimzi-cluster-operator.v0.39.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.6.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:50b49f9b5e554cca517fc2794e11319d5aaba1676494b285baa3a74fc79eb501
          createdAt: "2023-12-14 20:28:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.39.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.39 is the last minor release with support for Kubernetes 1.21 and 1.22.** From Strimzi 0.40 on, only Kubernetes 1.23 and newer will be supported.
          ### New in 0.39
          * Add support for Apache Kafka 3.5.2 and 3.6.1
          * The `StableConnectIdentities` feature gate moves to GA stage and is now permanently enabled without the possibility to disable it.
            All Connect and Mirror Maker 2 operands will now use StrimziPodSets.

          * The `KafkaNodePools` feature gate moves to the beta stage and is enabled by default.
            If needed, `KafkaNodePools` can be disabled in the feature gates configuration in the Cluster Operator.

          * The `UnidirectionalTopicOperator` feature gate moves to the beta stage and is enabled by default.
            If needed, `UnidirectionalTopicOperator` can be disabled in the feature gates configuration in the Cluster Operator.

          * Improved Kafka Connect metrics and dashboard example files
          * Allow specifying and managing KRaft metadata version
          * Add support for KRaft to KRaft upgrades (Apache Kafka upgrades for the KRaft-based clusters)
          * Improved Kafka Mirror Maker 2 dashboard example file
          * Support for rolling updates of KRaft controller nodes
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:dd1706f36031c96dfcd0da7b3f804168e5f076220caa302973455fdcd76bd03e
        - quay.io/strimzi/kaniko-executor@sha256:723b7add2473e73c45c81361e1708af4971c321b0391c1298a0557fb8ca4f7f6
        - quay.io/strimzi/maven-builder@sha256:6ce3c16284379d5db9e42a6f3e2b765c981091164a28fdd6bf927b6c5d4ab557
        - quay.io/strimzi/kafka@sha256:49b7530ada1175750ad5a706e545ba3594d85e1e3a90f98b0200a73fa9875057
        - quay.io/strimzi/kafka@sha256:84691b3afb67b312400e01572ff08b64c1e1b1d32a4276e0c21bba10c71c36fe
        - quay.io/strimzi/kafka-bridge@sha256:13838889cc20a3a382ea64e749cad94f997c0b928840588f2d1befcf19247bea
        - quay.io/strimzi/kafka@sha256:53b960411e852f5cb182ecc5d395a3562865f83e15e6251f3a9af0eb413feeed
        - quay.io/strimzi/kafka@sha256:2badc65532b483bba6c1f6a8b0c9895e993ce24e9ebd3cd953d0d17fbe36d27a
        - quay.io/strimzi/operator@sha256:50b49f9b5e554cca517fc2794e11319d5aaba1676494b285baa3a74fc79eb501
        version: 0.39.0
      entries:
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.39.x
    - currentCSV: strimzi-cluster-operator.v0.40.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:2228d89b53e3524e69d1bcf5c2cd373884ce696d9005389a6c78dad23310e837
          createdAt: "2024-03-09 21:27:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.40.0/deploying.html#assembly-upgrade-str).

          **From Strimzi 0.40 on, we support only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.40
          * Add support for Apache Kafka 3.7.0. Remove support for Apache Kafka 3.5.0, 3.5.1, and 3.5.2.
          * The `UseKRaft` feature gate moves to beta stage and is enabled by default. If needed, `UseKRaft` can be disabled in the feature gates configuration in the Cluster Operator.
          * Add support for ZooKeeper to KRaft migration by enabling the users to move from using ZooKeeper to store metadata to KRaft.
          * Add support for moving from dedicated controller-only KRaft nodes to mixed KRaft nodes
          * Added support for Kafka Exporter `offset.show-all` parameter
          * Prevent removal of the `broker` process role from KRaft mixed-nodes that have assigned partition-replicas
          * Improve broker scale-down prevention to continue in reconciliation when scale-down cannot be executed
          * Added support for Tiered Storage by enabling the configuration of custom storage plugins through the Kafka custom resource.
          * Update HTTP bridge to the latest 0.28.0 release
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:73deb1af0d93e9fa8955a5a412d0e7138384bdc80e7574e815d21796e446b636
        - quay.io/strimzi/kaniko-executor@sha256:3fb4ace22589fcf8d3f01b74c37b10b32d87e15d103ea59e0933b61c4e46ff4a
        - quay.io/strimzi/maven-builder@sha256:d02a9b48c77a44eeab6426c9af4adf75405217994bcb01ce0196e616041350d8
        - quay.io/strimzi/operator@sha256:2228d89b53e3524e69d1bcf5c2cd373884ce696d9005389a6c78dad23310e837
        - quay.io/strimzi/kafka@sha256:2c143eb10b66037ef877311eb767457e4b877d5ff2c47d393c76babec155caff
        - quay.io/strimzi/kafka@sha256:e2b9f2835498831bbf30ac2d651de143d57088bcb23943101b066e941db92aca
        - quay.io/strimzi/kafka@sha256:c491a95f9b9b58f406461faf222ef612ad8d2f4d7b53f7f3645b20c99d7d1583
        version: 0.40.0
      entries:
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.40.x
    - currentCSV: strimzi-cluster-operator.v0.41.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:4c687a4fac96ffc22433ca6d00fb1bd31722767b44c9d87c465f95b57fb70089
          createdAt: "2024-05-09 22:23:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.41.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.41 supports only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.41
          * Support for Apache Kafka 3.6.2
          * Metrics to monitor CA expiration
          * Support for JBOD storage in KRaft mode. (Note: JBOD support in KRaft mode is considered early access in Apache Kafka 3.7.x)
          * Support for changing topic replication factor by changing the `KafkaTopic` resources.
          * The `KafkaNodePools` feature gate moves to GA stage and is permanently enabled without the possibility to disable it. To use the Kafka Node Pool resources, you still need to use the `strimzi.io/node-pools: enabled` annotation on the `Kafka` custom resources.
          * Support for configuring the `externalIPs` field in node port type services.
          * The `UnidirectionalTopicOperator` feature gate moves to GA stage and is permanently enabled without the possibility to disable it. If the topics whose names start with `strimzi-store-topic` and `strimzi-topic-operator` still exist, you can delete them.
          * Improve validation of `KafkaMirrorMaker2` resource to make sure the Connect cluster is set to the target cluster
          * Continue reconciliation after failed manual rolling update using the `strimzi.io/manual-rolling-update` annotation (when the `ContinueReconciliationOnManualRollingUpdateFailure` feature gate is enabled)
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/maven-builder@sha256:17f09b07fc41a059e6a048cb766046d43a072cfd3a8266c97388a37543dadee8
        - quay.io/strimzi/kafka@sha256:c3a346235ce4dea42986648ca07f2482d8fc0bf654c1d441b1301a23999304db
        - quay.io/strimzi/kafka@sha256:bf49499980bbe5d4dd399de0dd396089b0fd9dd29bc46e14d39b20f1e826aec1
        - quay.io/strimzi/kafka@sha256:41811bb106d6182978a889eaedffeadffaa5ab1599c4cd8e4f5015b131c7ac30
        - quay.io/strimzi/kafka@sha256:6243d98890c665ef65b5f5fa2ffd046412281ffef31bc7a07cac37bd77b2579d
        - quay.io/strimzi/kafka-bridge@sha256:c833f930ead2d1e9054912a20689ee979d335f5e5ca8769e3c6b8a5514a0df96
        - quay.io/strimzi/operator@sha256:4c687a4fac96ffc22433ca6d00fb1bd31722767b44c9d87c465f95b57fb70089
        - quay.io/strimzi/kaniko-executor@sha256:9404037bd5431b012c8c633ec3345907667453c5532e5bfaaedc78286de22c62
        version: 0.41.0
      entries:
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.41.x
    - currentCSV: strimzi-cluster-operator.v0.42.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:ca5c1b70e1196deec4bf519921bfa55a967aa1ea15bd4c182bc56bea08d2e114
          createdAt: "2024-07-03 12:34:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.42.0/deploying.html#assembly-upgrade-str).

          **From version 0.40.0, Strimzi supports only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.42
          * Support for Apache Kafka 3.7.1
          * The UseKRaft feature gate moves to GA stage and is permanently enabled without the possibility of disabling it. To use KRaft (ZooKeeper-less Apache Kafka), you still need to use the strimzi.io/kraft: enabled annotation on the Kafka custom resources or migrate from an existing ZooKeeper-based cluster.
          * Update the base image used by Strimzi containers from UBI8 to UBI9
          * Add support for filename patterns when configuring trusted certificates
          * Enhance KafkaBridge resource with consumer inactivity timeout and HTTP consumer/producer enablement.
          * Add support for setting publishNotReadyAddresses on services for listener types other than internal.
          * Added support for configuring the quotas plugin with type strimzi or kafka in the Kafka custom resource. The Strimzi Quotas plugin version was updated to 0.3.1.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:6527c426f753521c07f3a3df2ee432273a41b109594096b4c6cae2e8e609d669
        - quay.io/strimzi/kafka@sha256:6c519531cdd5f40f326b295d8827344140efe2607a1826f7f1cab8c3427f43f7
        - quay.io/strimzi/kafka-bridge@sha256:fdddffb3de59ba48c05964bb0e760593088d24d1ee35b090112fe6cc6b553cd8
        - quay.io/strimzi/kafka@sha256:bc49c10b3e839e161e515c3a1e411d1f8df9de5e2c8d71aa7b63ccdbbae9e8d2
        - quay.io/strimzi/kaniko-executor@sha256:11821b0078a270612f1bcc37f5dc716ab273c974648c40f95fa4804352e647a7
        - quay.io/strimzi/operator@sha256:ca5c1b70e1196deec4bf519921bfa55a967aa1ea15bd4c182bc56bea08d2e114
        - quay.io/strimzi/maven-builder@sha256:342e8d0086c53cf7334127798e5bd3764b28f5e4909ba67d65f6d937dc3dba05
        - quay.io/strimzi/kafka@sha256:841a46e9c10481df3ef702412caf362b55ae07c30a3a5927073356cd36dc049f
        - quay.io/strimzi/kafka@sha256:c56c5abf94897f20f976603cba0678a33cdfbbcd6f9030a193f7684cb9894a6c
        version: 0.42.0
      entries:
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.42.x
    - currentCSV: strimzi-cluster-operator.v0.43.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:d3732245f34afaa6b571f1f912e1cd017891b22003d6367ae8ca059b2a5cb13c
          createdAt: "2024-08-22 15:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.43.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.43 is the last version with support for Kubernetes 1.23 and 1.24.** From Strimzi 0.44, only Kubernetes versions 1.25 and newer will be supported.
          ### New in 0.43
          * Support for Apache Kafka 3.8.0
          * Removed support for Apache Kafka 3.6.0, 3.6.1, and 3.6.2
          * Specifying additional volumes and volume mounts in Strimzi custom resources
          * Support for custom Cruise Control API users
          * Unregistration of KRaft nodes after scale-down
          * Kafka Exporter updated to 1.8.0 (and the Grafana dashboard updated to work with it)
          * Service account based authentication
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:80f329ecd3ba96fac12e75be874dce0debc8395337737581de34238ce11c8fb1
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:6414f8f3985f221b0eda53867c00494140108d17579c4dcbbc93934bb00e5d49
        - quay.io/strimzi/kafka@sha256:d27cf5bf3624e18ba705246b14f92d12b8a15eb8551b13abf027cf71b549baf0
        - quay.io/strimzi/kafka@sha256:ad06ead5125bce87498b1582a76e78143a6d21408ac4b12bb7e6007aa4cabd84
        - quay.io/strimzi/operator@sha256:d3732245f34afaa6b571f1f912e1cd017891b22003d6367ae8ca059b2a5cb13c
        - quay.io/strimzi/kafka@sha256:b2f4e8895e9ed66228512ce14aec685101fa75dc329128d47cb1c15e3aef3c9c
        version: 0.43.0
      entries:
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.43.x
    - currentCSV: strimzi-cluster-operator.v0.44.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:b07b81f7e282dea2e4e29a7c93cfdd911d4715a5c32fe4c4e3d7c71cba3091e8
          createdAt: "2024-08-22 15:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.44.0/deploying.html#assembly-upgrade-str).

          **From version 0.44, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**

          **ZooKeeper support will be soon removed from Apache Kafka and Strimzi.** Currently, the last Strimzi version with ZooKeeper support is expected to be Strimzi 0.45. Please plan your migration to KRaft (ZooKeeper-less Apache Kafka) accordingly. Follow the [documentation](https://strimzi.io/docs/operators/latest/full/deploying.html#assembly-kraft-mode-str) for more details.

          **Kafka Mirror Maker 1 support will be soon removed from Apache Kafka and Strimzi.** Currently, the last Strimzi version with Mirror Maker 1 support is expected to be Strimzi 0.45. Please plan your migration to Mirror Maker 2 or another mirroring tool.

          ### New in 0.44
          * Managing offsets in Connect and Mirror Maker 2 connectors through custom resources
          * Templating `advertisedHost` and `host` fields in listener configuration
          * Configuring environment variables in operands based on Secrets and ConfigMaps
          * Disabling Pod disruption Budget management
          * Auto-rebalancing at scale-up/scale-down
          * `ContinueReconciliationOnManualRollingUpdateFailure` feature gate moves to the beta stage and is enabled by default
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:9a4224635e09ece2c1d3ffe9d715813cf63037461c9f06c1b5f53554aed5ecdc
        - quay.io/strimzi/kafka-bridge@sha256:cc95ce9f496057f3b18d5e8297b40a410cc66f9fd4a2227615971458a26bac25
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:7b55f971a3b62923a9a61a61c82f542e07deeb6ffdbaede68244ba358a0853f2
        - quay.io/strimzi/operator@sha256:b07b81f7e282dea2e4e29a7c93cfdd911d4715a5c32fe4c4e3d7c71cba3091e8
        - quay.io/strimzi/kafka@sha256:238bb080825b7ecebaa819be09788bc3d056ae8162ebcd41444076dc22f85740
        - quay.io/strimzi/kafka@sha256:98d332b7fa5c9e48f60b516e00ed5812bd9268e014fd2337284e7cc3e664f1eb
        version: 0.44.0
      entries:
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.44.x
    - currentCSV: strimzi-cluster-operator.v0.45.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:58e8722216b0b0e8b1dd79109725238fa4016a16cc6a9fca0e327bccb1506810
          createdAt: "2025-06-27 01:38:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.45.1/deploying.html#assembly-upgrade-str).

          **From version 0.45, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**

          **ZooKeeper and Kafka Mirror Maker 1 support will be removed from Apache Kafka and Strimzi.** Strimzi 0.45 is the last minor Strimzi version with support for ZooKeeper-based Apache Kafka clusters and MirrorMaker 1 deployments. Please make sure to [migrate to KRaft](https://strimzi.io/docs/operators/latest/full/deploying.html#assembly-kraft-mode-str) and MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          **Strimzi EnvVar Configuration Provider and Strimzi MirrorMaker 2 Extensions will be removed from Strimzi.** Strimzi 0.45 is the last Strimzi version to include the Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0). Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.
          ### New in 0.45.1
          * Add support for Kafka 3.9.1
          * Update HTTP bridge to 0.31.2
          * Update Strimzi Kafka OAuth library to 0.15.1
          * Fixed bug which may lead to metadata loss within the cluster when restarting a KRaft migration after a previous rollback due to missing `/migration` znode deletion.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:e98fbe49cadbc5d3afbd8f17abea218f216c95d52f4fa1fe514a7a988624da24
        - quay.io/strimzi/kafka@sha256:ba52ed046b1dccdbd96f4e68057ce014d862a7c9c1fc670760c023b9aa09f23f
        - quay.io/strimzi/kafka-bridge@sha256:c462441a966f0732855e684b4840563bfad45391345c9511fa49e1630ba9c3e3
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:770e410f1e7e0e92c31bae6c19831ba36d59dd11b44e90c159aef16a965aa998
        - quay.io/strimzi/kafka@sha256:173cb5793756f75739c51a83bbe06d7f0094fd003a2982160a5cc6e216cd6d28
        - quay.io/strimzi/kafka@sha256:6a05924839a2c7da0431408e6d6223a055e75d88b3f057ac5caa7f9097227f56
        - quay.io/strimzi/operator@sha256:58e8722216b0b0e8b1dd79109725238fa4016a16cc6a9fca0e327bccb1506810
        version: 0.45.1
      entries:
      - name: strimzi-cluster-operator.v0.45.1
        version: 0.45.1
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.45.x
    - currentCSV: strimzi-cluster-operator.v0.46.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.0.0",
                      "metadataVersion": "4.0-IV3",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:6a22e4968e9d409405423d12c04a0e3fecaea2ee2794bd27de7cf21f59453265
          createdAt: "2025-06-26 11:58:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.46.0/deploying.html#assembly-upgrade-str).

          **From version 0.46, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**
          **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed. Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!
          **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images. Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.
          **Support for MirrorMaker 1 has been removed. Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.
          ### New in 0.46.1
          * Update CSV file for OLM installation to add missing permission for broker Pods to "get" nodes Kubernetes resource.
          * Allow the injection of annotations on the deployment object in Helm charts.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:3e1815aed3f74b8573cebfced66b2c78ca0cf7c746c26db2002380cd39d09ae6
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:b532234c234fc30fecec992ad1a6154c5ef08ebf9a20ec8bc62e0c88d791ff36
        - quay.io/strimzi/kafka@sha256:f3837f8cb2a60e906e84486128c41c398424f4085df2348e9e0486bfc12089fd
        - quay.io/strimzi/kafka@sha256:82273d873ce21a73f8f3118bf61627cb15734b43ea7ca92d43b290340b44e485
        - quay.io/strimzi/operator@sha256:6a22e4968e9d409405423d12c04a0e3fecaea2ee2794bd27de7cf21f59453265
        version: 0.46.1
      entries:
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.46.x
    - currentCSV: strimzi-cluster-operator.v0.47.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.0.0",
                      "metadataVersion": "4.0-IV3",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:589c4d63641d9a944462dd682f6e5febe8b38ff55b9253949b061aca16feb154
          createdAt: "2025-07-12 17:21:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.47 is the last Strimzi version with support for Kubernetes 1.25 and 1.26.**
            From Strimzi 0.48.0 on, we will support only Kubernetes 1.27 and newer.

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.47.0
          * Support for Kafka 3.9.1
          * Progress tracking for Cruise Control rebalances
          * Support for Kubernetes Image Volumes to mount custom plugins
          * Support for [JsonTemplateLayout](https://logging.apache.org/log4j/2.x/manual/json-template-layout.html) in Operators, Cruise Control, and Kafka 4.0.0
          * Support for Strimzi Metrics Reporter for the Kafka brokers / controllers
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/operator@sha256:589c4d63641d9a944462dd682f6e5febe8b38ff55b9253949b061aca16feb154
        - quay.io/strimzi/maven-builder@sha256:c67a3ad3d92420ac8b7f8435eeec53ebffaea4e894e65414a7c47c20e3faba27
        - quay.io/strimzi/kafka@sha256:978e2ed5f0d315e8bf0ab0503ee839acbd9da01ca427e6581b59ac25d36bc9c9
        - quay.io/strimzi/kafka@sha256:8a793cdf5084cdb72c6663419530985158b16330a518720e9e9d1d96babbaf7e
        - quay.io/strimzi/kafka@sha256:551882f47a6d2dc8300bcfca251fffffc3379acb2724a7efe16f17c0c4f529fb
        - quay.io/strimzi/kafka-bridge@sha256:428c5587f547dabfc995a73745569d113380c7e6b8bf0b2504c6a9a00d6c4608
        version: 0.47.0
      entries:
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.47.x
    - currentCSV: strimzi-cluster-operator.v0.48.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.1.0",
                      "metadataVersion": "4.1-IV1",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.1.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
          createdAt: "2025-09-25 10:32:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.48.0 supports only Kubernetes 1.27 and newer.**
            Kubernetes 1.25 and 1.26 are not supported anymore

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.48.0
          * Support for Kafka 4.1.0
          * Add monitoring of custom resources using kubernetes-state-metrics (KSM)
          * Support for type: custom client authentication
          * Added support for Strimzi Metrics Reporter to Kafka Connect, Mirror Maker 2 and Kafka Bridge.
          * Added support for KIP-745 in Kafka Connector and Mirror Maker 2, allowing the usage of includeTasks and onlyFailed arguments in Kafka connectors restart.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:64380efcb8da87fdc0e0de238a20456be9abb830174c2f2331882b7a84657e57
        - quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
        - quay.io/strimzi/kafka@sha256:48db05ad2e9fa39eb388089e7e25654bcbf4da1ddc3a5635d675056c28bbcb96
        - quay.io/strimzi/kafka@sha256:34450afc1a3399a9026fc3c6ec90e9f6adfec9faf6520545b4d84e90cda34964
        - quay.io/strimzi/kafka-bridge@sha256:53034f64f0b672f10b5bacea1c7a25132f118df7fd5c9032c4dbf702ed93796a
        version: 0.48.0
      entries:
      - name: strimzi-cluster-operator.v0.48.0
        version: 0.48.0
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.48.x
    defaultChannel: stable
    packageName: strimzi-kafka-operator
    provider:
      name: Strimzi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-nifi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nifi-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:05946a5e8f040e8ee46d56f2d4cca3d97e02d086e6d8224200204de4d76851fa
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:05946a5e8f040e8ee46d56f2d4cca3d97e02d086e6d8224200204de4d76851fa
        version: 24.11.1
      entries:
      - name: nifi-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: nifi-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:e5c3c4e2ad1c455fbb04a60658936bfb44c2bc9d2b616737ce1a0eba201b7a3f
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiClusterSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:e5c3c4e2ad1c455fbb04a60658936bfb44c2bc9d2b616737ce1a0eba201b7a3f
        version: 25.3.0
      entries:
      - name: nifi-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: nifi-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiClusterSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
        version: 25.7.0
      entries:
      - name: nifi-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: nifi-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiClusterSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
        version: 25.7.0
      entries:
      - name: nifi-operator.v25.7.0
        version: 25.7.0
      - name: nifi-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-nifi-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/en/products/edge/validated-patterns
    name: patterns-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: patterns-operator.v0.0.62
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gitops.hybrid-cloud-patterns.io/v1alpha1",
                "kind": "Pattern",
                "metadata": {
                  "name": "pattern-sample"
                },
                "spec": {
                  "clusterGroupName": "hub",
                  "gitSpec": {
                    "targetRepo": "https://github.com/validatedpatterns/multicloud-gitops",
                    "targetRevision": "main"
                  },
                  "multiSourceConfig": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/hybridcloudpatterns/patterns-operator:0.0.62
          createdAt: "2025-09-02 16:14:40"
          description: "An operator to deploy and manage architecture patterns from
            https://validatedpatterns.io. \nThis operator collects some Analytics
            like cluster version, cloud type, etc. To disable this\nadd \"ANALYTICS:
            false\" to the environment variables in the subscription."
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/validatedpatterns/patterns-operator
          support: validatedpatterns@googlegroups.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pattern is the Schema for the patterns API
            displayName: Pattern
            kind: Pattern
            name: patterns.gitops.hybrid-cloud-patterns.io
            version: v1alpha1
        description: Deploys and manages architecture patterns from https://validatedpatterns.io
        displayName: Validated Patterns Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Validated
        - Pattern
        - Blueprint
        - Architecture
        links:
        - name: Hybrid Cloud Patterns
          url: https://validatedpatterns.io
        maintainers:
        - email: validatedpatterns@googlegroups.com
          name: patterns-team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/en/products/edge/validated-patterns
        relatedImages:
        - quay.io/hybridcloudpatterns/kube-rbac-proxy@sha256:409a464692c7db3cf68893376479dca1b53802412237637725927f7e78aa0844
        - quay.io/hybridcloudpatterns/patterns-operator@sha256:15ed44938b112bcf9b3a9a566749b39ba2b9053a6a8b9b499ef247977c18496c
        version: 0.0.62
      entries:
      - name: patterns-operator.v0.0.62
        version: 0.0.62
      - name: patterns-operator.v0.0.61
        version: 0.0.61
      - name: patterns-operator.v0.0.60
        version: 0.0.60
      - name: patterns-operator.v0.0.59
        version: 0.0.59
      - name: patterns-operator.v0.0.58
        version: 0.0.58
      - name: patterns-operator.v0.0.57
        version: 0.0.57
      - name: patterns-operator.v0.0.56
        version: 0.0.56
      - name: patterns-operator.v0.0.55
        version: 0.0.55
      - name: patterns-operator.v0.0.54
        version: 0.0.54
      - name: patterns-operator.v0.0.53
        version: 0.0.53
      - name: patterns-operator.v0.0.52
        version: 0.0.52
      - name: patterns-operator.v0.0.51
        version: 0.0.51
      - name: patterns-operator.v0.0.50
        version: 0.0.50
      - name: patterns-operator.v0.0.49
        version: 0.0.49
      - name: patterns-operator.v0.0.48
        version: 0.0.48
      - name: patterns-operator.v0.0.47
        version: 0.0.47
      - name: patterns-operator.v0.0.46
        version: 0.0.46
      - name: patterns-operator.v0.0.45
        version: 0.0.45
      - name: patterns-operator.v0.0.44
        version: 0.0.44
      - name: patterns-operator.v0.0.43
        version: 0.0.43
      - name: patterns-operator.v0.0.42
        version: 0.0.42
      - name: patterns-operator.v0.0.41
        version: 0.0.41
      - name: patterns-operator.v0.0.40
        version: 0.0.40
      - name: patterns-operator.v0.0.39
        version: 0.0.39
      - name: patterns-operator.v0.0.38
        version: 0.0.38
      - name: patterns-operator.v0.0.37
        version: 0.0.37
      - name: patterns-operator.v0.0.36
        version: 0.0.36
      - name: patterns-operator.v0.0.35
        version: 0.0.35
      - name: patterns-operator.v0.0.34
        version: 0.0.34
      - name: patterns-operator.v0.0.33
        version: 0.0.33
      - name: patterns-operator.v0.0.32
        version: 0.0.32
      - name: patterns-operator.v0.0.31
        version: 0.0.31
      - name: patterns-operator.v0.0.30
        version: 0.0.30
      - name: patterns-operator.v0.0.29
        version: 0.0.29
      - name: patterns-operator.v0.0.28
        version: 0.0.28
      - name: patterns-operator.v0.0.27
        version: 0.0.27
      - name: patterns-operator.v0.0.26
        version: 0.0.26
      - name: patterns-operator.v0.0.25
        version: 0.0.25
      - name: patterns-operator.v0.0.24
        version: 0.0.24
      - name: patterns-operator.v0.0.23
        version: 0.0.23
      - name: patterns-operator.v0.0.22
        version: 0.0.22
      - name: patterns-operator.v0.0.21
        version: 0.0.21
      - name: patterns-operator.v0.0.20
        version: 0.0.20
      - name: patterns-operator.v0.0.19
        version: 0.0.19
      - name: patterns-operator.v0.0.18
        version: 0.0.18
      - name: patterns-operator.v0.0.17
        version: 0.0.17
      - name: patterns-operator.v0.0.16
        version: 0.0.16
      - name: patterns-operator.v0.0.15
        version: 0.0.15
      - name: patterns-operator.v0.0.14
        version: 0.0.14
      - name: patterns-operator.v0.0.13
        version: 0.0.13
      - name: patterns-operator.v0.0.12
        version: 0.0.12
      - name: patterns-operator.v0.0.11
        version: 0.0.11
      - name: patterns-operator.v0.0.10
        version: 0.0.10
      - name: patterns-operator.v0.0.9
        version: 0.0.9
      - name: patterns-operator.v0.0.8
        version: 0.0.8
      - name: patterns-operator.v0.0.7
        version: 0.0.7
      - name: patterns-operator.v0.0.6
        version: 0.0.6
      - name: patterns-operator.v0.0.5
        version: 0.0.5
      - name: patterns-operator.v0.0.4
        version: 0.0.4
      - name: patterns-operator.v0.0.3
        version: 0.0.3
      - name: patterns-operator.v0.0.2
        version: 0.0.2
      - name: patterns-operator.v0.0.1
        version: 0.0.1
      name: fast
    defaultChannel: fast
    packageName: patterns-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com/en/products/edge/validated-patterns
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: ""
    name: hawtio-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hawtio-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.14.0",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/hawtio/operator:0.5.0
          createdAt: "2022-04-08T00:00:00Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Hawtio Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - docker.io/hawtio/operator:0.5.0
        version: 0.5.0
      entries:
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: alpha
    - currentCSV: hawtio-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/hawtio/operator:1.2.0
          createdAt: "2024-12-02T17:22:48Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          olm.skipRange: '>=1.0.0 <1.0.2'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Hawtio console that eases the discovery and management
              of hawtio-enabled applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#jvm-agent).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/hawtio/operator:1.2.0
        version: 1.2.0
      entries:
      - name: hawtio-operator.v1.2.0
        version: 1.2.0
      - name: hawtio-operator.v1.1.1
        version: 1.1.1
      - name: hawtio-operator.v1.1.0
        version: 1.1.0
      - name: hawtio-operator.v1.0.1
        version: 1.0.1
      - name: hawtio-operator.v1.0.0
        version: 1.0.0
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: latest
    - currentCSV: hawtio-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/hawtio/operator:1.2.0
          createdAt: "2024-12-02T17:22:48Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          olm.skipRange: '>=1.0.0 <1.0.2'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Hawtio console that eases the discovery and management
              of hawtio-enabled applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#jvm-agent).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/hawtio/operator:1.2.0
        version: 1.2.0
      entries:
      - name: hawtio-operator.v1.2.0
        version: 1.2.0
      - name: hawtio-operator.v1.1.1
        version: 1.1.1
      - name: hawtio-operator.v1.1.0
        version: 1.1.0
      - name: hawtio-operator.v1.0.1
        version: 1.0.1
      - name: hawtio-operator.v1.0.0
        version: 1.0.0
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: stable-v1
    defaultChannel: stable-v1
    packageName: hawtio-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "docker.io/percona/percona-server-mongodb:7.0.18-11",
                  "imagePullPolicy": "Always",
                  "initImage": "docker.io/percona/percona-server-mongodb-operator:1.20.1",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:2.44.1",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "docker.io/percona/percona-backup-mongodb:2.9.1",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.io/percona/percona-server-mongodb-operator:1.20.1
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.20.1
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - docker.io/percona/percona-server-mongodb-operator:1.20.1
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator.v1.20.0
        version: 1.20.0
      - name: percona-server-mongodb-operator.v1.19.1
        version: 1.19.1
      - name: percona-server-mongodb-operator.v1.19.0
        version: 1.19.0
      - name: percona-server-mongodb-operator.v1.18.0
        version: 1.18.0
      - name: percona-server-mongodb-operator.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: directpv
      provider-url: ""
    name: minio-volume-manager-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-volume-manager-operator.v5.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.quay.io/v1alpha1",
                "kind": "AistorVolumemanager",
                "metadata": {
                  "name": "aistorvolumemanager-sample"
                },
                "spec": {
                  "annotations": {},
                  "controller": {
                    "podAnnotations": {},
                    "podLabels": {},
                    "replicas": 3
                  },
                  "images": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": []
                  },
                  "labels": {},
                  "license": "",
                  "nodeserver": {
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-10-05T23:18:24Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AistorVolumemanager
            name: aistorvolumemanagers.charts.quay.io
            version: v1alpha1
        description: directpv
        displayName: directpv
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - directpv
        links:
        - name: Private Dpv Operator
          url: https://private-dpv-operator.domain
        maintainers:
        - email: cesarcelis@minio.io
          name: celis
        maturity: alpha
        provider:
          name: directpv
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - quay.io/cniackz4/directpv-operator@sha256:af06e57661a7e05809e44c3fac571ed949cc04fd656fdb0c8324b533778eb604
        version: 5.0.3
      entries:
      - name: minio-volume-manager-operator.v5.0.3
        version: 5.0.3
      - name: private-dpv-operator.v5.0.0
        version: 5.0.0
      name: stable
    defaultChannel: stable
    packageName: minio-volume-manager-operator
    provider:
      name: directpv
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-commonui-operator-app
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-commonui-operator.v4.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operators.ibm.com/v1alpha1",
                "kind": "CommonWebUI",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "example-commonwebui",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator"
                  },
                  "name": "example-commonwebui"
                },
                "spec": {
                  "commonWebUIConfig": {
                    "cpuLimits": "300",
                    "cpuMemory": "256",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.8.0",
                    "ingressPath": "/common-nav",
                    "landingPage": "",
                    "requestLimits": "300",
                    "requestMemory": "256",
                    "serviceName": "common-web-ui"
                  },
                  "globalUIConfig": {
                    "cloudPakVersion": "3.6.0",
                    "defaultAdminUser": "admin",
                    "defaultAuth": "",
                    "enterpriseLDAP": "",
                    "enterpriseSAML": "",
                    "osAuth": "",
                    "sessionPollingInterval": 5000
                  },
                  "operatorVersion": "4.8.0",
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "256Mi"
                    },
                    "requests": {
                      "cpu": "300m",
                      "memory": "256Mi",
                      "ephemeral-storage": "251Mi"
                    }
                  },
                  "version": "4.8.0"
                }
              },
              {
                "apiVersion": "foundation.ibm.com/v1",
                "kind": "NavConfiguration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "common-web-ui-config",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator",
                    "default": "true",
                    "name": "common-web-ui-config"
                  },
                  "name": "common-web-ui-config"
                },
                "spec": {
                  "about": {
                    "copyright": "© 2018, 2020 IBM. All rights reserved.",
                    "licenses": [
                      "yq, version 3.3.0, MIT+GPL",
                      "MongoDB, version 4.0.16 Community Edition, SSPL",
                      "Ansible: © 2017 Red Hat, Inc., http://www.redhat.com; © Henry Graham (hzgraham) \u003cHenry.Graham@mail.wvu.edu\u003e",
                      "calico-bird: © 1998–2008, Martin Mares \u003cmj@ucw.cz\u003e; © 1998–2000 Pavel Machek \u003cpavel@ucw.cz\u003e;",
                      "© 1998–2008, Ondrej Filip \u003cfeela@network.cz\u003e; © 2009–2013,  CZ.NIC z.s.p.o.",
                      "chrony: © Richard P. Curnow  1997-2003, GPL v2",
                      "collectd, © 2017-2018, version 5.7.2, GPL v2, \u003chttps://github.com/collectd/collectd/tree/collectd-5.7.2\u003e",
                      "crudini: © Pádraig Brady \u003cP@draigBrady.com\u003e",
                      "Galera-3: © 2007–2014 Codership Oy \u003cinfo@codership.com\u003e",
                      "glusterfs: © 2010–2013+ James Shubin \u003chttps://ttboj.wordpress.com/\u003e",
                      "haproxy: © 2000–2013  Willy Tarreau \u003cw@1wt.eu\u003e",
                      "heketi v6.0.0: © 2015 The heketi Authors, GPL v2",
                      "heketi v8.0.0: © 2015 The heketi Authors, GPL v2",
                      "heketi-master/apps.app.go: © 2015 The heketi Authors",
                      "heketi-master/client/api/go-client/backup.go: © 2016 The heketi Authors",
                      "heketi-master/doc/man/heketi-cli.8: © 2016 The heketi Authors",
                      "heketi-master/extras/docker/gluster/gluster-setup.sh: © 2016 Red Hat, Inc. \u003chttp://www.redhat.com\u003e",
                      "ieee-data: © 2013 Luciano Bello \u003cluciano@debian.org\u003e",
                      "javax.mail: © 2017 Oracle and/or its affiliates. All rights reserved.",
                      "keepalived: © 2001-2017 Alexandre Cassen \u003cacassen@gmail.com\u003e",
                      "libonig2: © 2006–2008 Max Kellermann \u003cmax@duempel.org\u003e; © 2014–2015 Jörg Frings-Fürst \u003cdebian@jff-webhosting.net\u003e",
                      "libtomcrypt: © 2004 Sam Hocevar \u003csam@hocevar.net\u003e, GPL v2",
                      "mariadb-common: © 2018 MariaDB. All rights reserved.",
                      "mariaDB: © 2018 MariaDB. All rights reserved. \u003chttps://mariadb.com/\u003e",
                      "mariadb-server: © 2018 MariaDB. All rights reserved.",
                      "minitar: © 2004 Mauricio Julio Fernandez Pradier and Austin Ziegler",
                      "MongoDB: © 2007 Free Software Foundation, Inc. \u003chttp://fsf.org/\u003e",
                      "nvmi-cli: © 1989, 1991 Free Software Foundation, Inc., GPL v2",
                      "OpenJDK: © 2018 Oracle Corporation and/or its affiliates",
                      "openshift-mariadb-galera: © 2007 Free Software Foundation, Inc. \u003chttp://fsf.org/\u003e",
                      "percona-xtrabackup: © 2006–2018 Percona LLC.",
                      "pwgen: © Christian Thöing \u003cc.thoeing@web.de\u003e",
                      "rdoc: © 2001–2003 Dave Thomas, The Pragmatic Programmers",
                      "readline: © Chet Ramey \u003cchet.ramey@case.edu\u003e",
                      "John the Ripper password cracker: © 1996–2013 by Solar Designer \u003csolar@openwall.com\u003e",
                      "spdx-exceptions: © 2018 SPDX Workgroup a Linux Foundation Project. All rights reserved.",
                      "socat: © 2001–2010 Gerhard Rieger",
                      "sshpass: © 2006, 2008 Lingnu Open Source Consulting Ltd. \u003chttp://www.lingnu.com\u003e",
                      "timelimit: © 2001, 2007 - 2010  Peter Pentchev, GPL v2",
                      "ua-parser-js: © 2012-2018 Faisal Salman \u003cf@faisalman.com\u003e, GPL v2",
                      "ubuntu-cloud-keyring: © 2010 Michael Vogt \u003cmichael.vogt@canonical.com\u003e",
                      "unboundid-ldapsdk: © 2015 UnboundID. The LDAP SDK for Java is developed by UnboundID. \u003cinfo@unboundid.com\u003e",
                      "xmpp4r: © Lucas Nussbaum \u003clucas@lucas-nussbaum.net\u003e, Stephan Maka \u003cstephan@spaceboyz.net\u003e, and others.",
                      "module-assistant: © 2003-2008 Eduard Bloch \u003cblade@debian.org\u003e, version 0.11.8, GPL v2;",
                      "© 2009 Cyril Brulebois \u003ckibi@debian.org\u003e, version 0.11.8, GPL v2; © 2013-2018 Andreas Beckmann \u003canbe@debian.org\u003e, version 0.11.8, GPL v2",
                      "module-init-tools: © 2011 ProFUSION embedded systems, version 22, GPL v2",
                      "thin: © 2017 Marc-Andre Cournoyer \u003cmacournoyer@gmail.com\u003e, version 1.7.2, GPL v2",
                      "gosu, © 1999-2014, version 1.1, GPL v3",
                      "mercurial (Python), © 2006-2018 ,version v4.5.3, GPL v2",
                      "garden-runc, © 2015-Present CloudFoundry.org Foundation, Inc. All Rights Reserved, version 1.17.0, GPLv2",
                      "libtomcrypt0, © 2003-2007 Tom St Denis \u003ctomstdenis@gmail.com\u003e, version 1.17-7, GPLv2",
                      "console-setup-min, © 1999,2000,2001,2002,2003,2006,2007,2008,2009,2010,2011 Anton Zinoviev, \u003canton@lml.bas.bg\u003e,version 1.108, GPLv2",
                      "dracut, © 2009 Harald Hoyer \u003charald@redhat.com\u003e, version 044+3-3, GPLv2",
                      "dracut-core, © 2009 Harald Hoyer \u003charald@redhat.com\u003e, version 044+3-3, GPLv2",
                      "g++, version 5.4.0-6ubuntu, GPL v2",
                      "libstdc++6, version 5.4.0-6ubuntu, GPL v3",
                      "libstdc++-5-dev, version 5.4.0-6ubuntu, GPL v3",
                      "docker-engine-selinux, version 3b5fac4, GPLv2",
                      "unorm, version 1.5.0, GPL v2",
                      "psmisc, version 22.20, GPL v2",
                      "lvm2-devel, version 2.0.2, GPL v2",
                      "nfs-utils, version 1.3, GPL v2",
                      "popt-static, version 1.13, GPL v2",
                      "sysvinit-tools, version 2.88, GPL v2",
                      "stunnel, version 5.53, GPL v2",
                      "stunnel, version 5.39, GPL v2",
                      "LVM2, version 2.02.180-10.el7_6.2, GPL v2",
                      "sysdig, version 2c43237, GPL",
                      "chisels, version 9722dbc, GPL",
                      "MongoDB, version 4.0.12, SSPL",
                      "ffi (Ruby Gem), 1.11.1, GPL",
                      "inotify-tools, v3.14, GPL v2",
                      "logrotate, v3.8.6, GPL v2",
                      "checker-qual, version 2.0.0, GPLv2",
                      "ocrad-bower, v1.0.0, GPL v3",
                      "Grafana, v7.5.12, AGPL",
                      "IBM Semeru Runtime Open Edition binaries, GPL v2",
                      "checker-compat-qual 2.5.5, GPL-2.0",
                      "javax.annotation-api 1.3.1, GPL-2.0"
                    ],
                    "logoUrl": "IBM Cloud Pak | Administration Hub",
                    "logoAltText": "IBM Cloud Pak | Administration Hub"
                  },
                  "header": {
                    "disabledItems": [
                      "createResource",
                      "catalog"
                    ],
                    "docUrlMapping": "http://ibm.biz/cpcs_adminui",
                    "logoAltText": "IBM Cloud Pak Administration Hub",
                    "logoHeight": "47px",
                    "logoUrl": "/common-nav/graphics/ibm-cloudpack-logo.svg",
                    "logoWidth": "190px"
                  },
                  "login": {
                    "loginDialog": {
                      "acceptText": "Your acceptance text here",
                      "dialogText": "You must set your dialog for this environment",
                      "enable": false,
                      "headerText": "Header text here"
                    },
                    "logoAltText": "Cloud Pak",
                    "logoHeight": "47px",
                    "logoUrl": "/common-nav/api/graphics/logincloudpak.svg",
                    "logoWidth": "190px"
                  },
                  "navItems": [
                    {
                      "id": "providers",
                      "label": "Identity providers",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/realms"
                    }
                  ]
                }
              },
              {
                "apiVersion": "foundation.ibm.com/v1",
                "kind": "NavConfiguration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "icp4i",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator",
                    "name": "icp4i"
                  },
                  "name": "icp4i"
                },
                "spec": {
                  "header": {
                    "disabledItems": [
                      "createResource",
                      "catalog",
                      "bookmark"
                    ],
                    "logoAltText": "Cloud Pak for Integration",
                    "logoUrl": "/common-nav/graphics/ibm-cloudpak-integration.svg"
                  },
                  "navItems": [
                    {
                      "detectionServiceName": true,
                      "id": "metering",
                      "label": "Metering",
                      "serviceId": "metering-ui",
                      "serviceName": "metering-ui",
                      "url": "/metering/dashboard?ace_config={ 'showClusterData': false }\u0026dashboard=cpi.icp.main"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "monitoring",
                      "isAuthorized": [
                        "Administrator",
                        "ClusterAdministrator",
                        "Operator"
                      ],
                      "label": "Monitoring",
                      "serviceId": "monitoring-ui",
                      "serviceName": "ibm-monitoring-grafana",
                      "target": "_blank",
                      "url": "/grafana"
                    },
                    {
                      "id": "id-access",
                      "label": "Identity and access",
                      "serviceId": "webui-nav"
                    },
                    {
                      "id": "providers",
                      "parentId": "id-access",
                      "label": "Identity providers",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/realms?useNav=icp4i"
                    },
                    {
                      "id": "teams-ids",
                      "parentId": "id-access",
                      "label": "Teams and service IDs",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/teams?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "logging",
                      "label": "Logging",
                      "serviceId": "kibana",
                      "serviceName": "kibana",
                      "target": "_blank",
                      "url": "/kibana"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "releases",
                      "label": "Helm Releases",
                      "serviceId": "catalog-ui",
                      "serviceName": "catalog-ui",
                      "url": "/catalog/instances?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "repos",
                      "label": "Helm Repositories",
                      "serviceId": "catalog-ui",
                      "serviceName": "catalog-ui",
                      "url": "/catalog/repositories?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "licensing",
                      "label": "Licensing",
                      "serviceId": "ibm-license-service-reporter",
                      "url": "/license-service-reporter",
                      "isAuthorized": [
                        "ClusterAdministrator",
                        "CloudPakAdministrator"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operators.ibm.com/v1alpha1",
                "kind": "SwitcherItem",
                "metadata": {
                  "name": "example-switcheritem"
                },
                "spec": {
                  "cloudPakInfo": {
                    "display": "Administration",
                    "label": "administration-hub",
                    "landingPage": "/common-nav/dashboard",
                    "logoURL": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
          description: The IBM Common Web UI delivers the common header API and the
            identity and access pages for IBM Cloud Platform Common Services.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.8.0
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-commonui-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            displayName: CommonWebUI service
            kind: CommonWebUI
            name: commonwebuis.operators.ibm.com
            version: v1alpha1
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            kind: NavConfiguration
            name: navconfigurations.foundation.ibm.com
            version: v1
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            kind: SwitcherItem
            name: switcheritems.operators.ibm.com
            version: v1alpha1
        description: "**Important:** Do not install this operator directly. Only install
          this operator using the IBM Common Services Operator. For more information
          about installing this operator and other Common Services operators, see
          [Installer documentation](http://ibm.biz/cpcs_opinstall). Additionally,
          you can exit this panel and navigate to the IBM Common Services tile in
          OperatorHub to learn more about the operator.\n\n If you are using this
          operator as part of an IBM Cloud Pak, see the documentation for that IBM
          Cloud Pak to learn more about how to install and use the operator service.
          For more information about IBM Cloud Paks, see [IBM Cloud Paks that use
          Common Services](http://ibm.biz/cpcs_cloudpaks).\n\n You can use the ibm-commonui-operator
          to install the Common Web UI service for the IBM Cloud Platform Common Services
          and access the Common Web UI console. You can use the Common Web UI console
          to access information and features from other IBM Cloud Platform Common
          Services or IBM Cloud Paks that you install. \n\nFor more information about
          the available IBM Cloud Platform Common Services, see the [IBM Knowledge
          Center](http://ibm.biz/cpcsdocs). \n## Supported platforms \n\n Red Hat
          OpenShift Container Platform 4.2 or newer installed on one of the following
          platforms: \n\n- Linux x86_64 \n- Linux on Power (ppc64le) \n- Linux on
          IBM Z and LinuxONE \n## Prerequisites\n\n The Common Web UI service has
          dependencies on other IBM Cloud Platform Common Services. Before you install
          this operator, you need to first install the operator dependencies and prerequisites:
          \n For the list of operator dependencies, see the IBM Knowledge Center [Common
          Services dependencies documentation](http://ibm.biz/cpcs_opdependencies).
          \n For the list of prerequisites for installing the operator, see the IBM
          Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).
          \n## Documentation \n\n To install the operator with the IBM Common Services
          Operator follow the the installation and configuration instructions within
          the IBM Knowledge Center. \n- If you are using the operator as part of an
          IBM Cloud Pak, see the documentation for that IBM Cloud Pak, for a list
          of IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).
          \n- If you are using the operator with an IBM Containerized Software, see
          the IBM Cloud Platform Common Services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall)."
        displayName: Ibm Common UI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        - Web Console
        - Common Services
        links:
        - name: GitHub
          url: https://github.com/IBM/ibm-commonui-operator
        maintainers:
        - email: siddarth.patel@ibm.com
          name: Siddarth Patel
        - email: sgrube@us.ibm.com
          name: Steve Grube
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:e4eb6a75e0d207232e479693d9c99ba863f8e62830f3191ca4d1e7ceaf3586d1
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator-bundle@sha256:e4340bd9481da998a6b89e94c884c46e17cfb50d9d0c1099f39a3c54e500d2b2
        version: 4.8.0
      entries:
      - name: ibm-commonui-operator.v4.8.0
        version: 4.8.0
      name: v4.7
    defaultChannel: v4.7
    packageName: ibm-commonui-operator-app
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent
      provider-url: ""
    name: cilium-enterprise
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cilium.v1.10.13-xed86eae
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"prometheus":{"serviceMonitor":{"enabled":false}},"mode":"cluster-pool"},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium Enterprise - eBPF-based Networking, Security, and Observability
        displayName: Cilium Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Enterprise Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Cilium Enterprise
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-ui-enterprise@sha256:c9921c65892a5293a4f1bef787636d6ae58cec1af65da06862a47e3450afdc6e
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/isovalent/hubble-relay@sha256:9145b03acc954efde54e1ab9da878462526c5e820b83a92e32cc24291f784d33
        - quay.io/isovalent/clustermesh-apiserver@sha256:eb3ee3d2cb07e6c5966e4bc2e50615949303be5d5128e156d5c0dbe8404dc093
        - quay.io/isovalent/hubble-enterprise-metadata@sha256:72c7aa3961f14df413569cd156820a45553bb24d393b11246a18bf8b6553e022
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        - quay.io/isovalent/cilium@sha256:6887537789603946543aaa370a3ec592287a235231bf16aa386e091823285515
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:cc1f0e6b27ccda8a843dfb7fa4a8fef052ba29716646557815677c5d1bea2a1a
        - quay.io/isovalent/hubble-export-fluentd@sha256:05cb13f533f2c89b98a47cd58badda07add54077cd22283ca5028cb39f2e4486
        - quay.io/isovalent/hubble-export-s3@sha256:ce4f71f7f13301133b7d471b87fb81108b2c2da7fc29865ef96bc8431b2be72b
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/oauth2-proxy/oauth2-proxy@sha256:ecd26b74a01f2b547ddaed4d32d35f8f5e09c378d5c1fc6cfa63f0adf659ac2b
        - quay.io/isovalent/hubble-enterprise-operator@sha256:0d5e6a5d9c07780c18e5aa81d8f1c95c399bff2d36a44fa5f61ac89a788ca83b
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/hubble-ui-enterprise-backend@sha256:e892cba6dafcdf6704df3135a0b38d11894a5f276ace40683e74cb62b9c311d8
        - quay.io/isovalent/operator-generic@sha256:1cc0b0d2e71564ad8759be326c66ea13b4955b9441fb67e906da534e1b38aefa
        - quay.io/cilium/hubble-export-stdout@sha256:f1b1475a3c89950481264f5f8ca63eea3b911a45dc83fdf15122d695a0fba9b3
        - quay.io/isovalent/hubble-enterprise@sha256:1e1d3e6c199546b2398f4166ff64210c04075bb407c8f07c5be8d3db08560d4a
        - quay.io/isovalent/hubble-rbac@sha256:03f3304c157024356a6de13c3d91b4506b8e54d587920d80ebcc76cbcfeaeda6
        version: 1.10.13+xed86eae
      entries:
      - name: cilium.v1.10.13-xed86eae
        version: 1.10.13+xed86eae
      - name: cilium.v1.10.12-xb203fae
        version: 1.10.12+xb203fae
      - name: cilium.v1.10.11-x10ae342
        version: 1.10.11+x10ae342
      name: "1.10"
    - currentCSV: cilium.v1.11.21-xd0d07b1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium Enterprise - eBPF-based Networking, Security, and Observability
        displayName: Cilium Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Enterprise Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Cilium Enterprise
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - cgr.dev/chainguard/oauth2-proxy:7.4.0@sha256:54857c70321004d3e0b8b8d7208ce5422c1f71c18db5be02d2543f187ea8e8e2
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:372511f6fa3523dadfe9128caf77520cf74ce24df4a19e1583ee197375e06014
        - quay.io/isovalent/hubble-ui-enterprise:v0.23.0@sha256:02635d1392cca7e09ab7797f69147880068ba7e3b0d865be97eeab33edd646a0
        - quay.io/cilium/certgen:v0.1.5@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/startup-script:62093c5c233ea914bfa26a10ba41f8780d9b737f@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/isovalent/clustermesh-apiserver:v1.11.21-cee.1@sha256:e16d8375ed8a1c3a872047b3aacf3469bde43552f1a51ee3e159ef2e9fea3e97
        - quay.io/coreos/etcd:v3.4.13@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/isovalent/hubble-relay:v1.11.21-cee.1@sha256:5933dd1f8111d6bca0c68db9247385b5d215d5b613095bafab8244098f2c5a26
        - quay.io/isovalent/hubble-ui-enterprise-backend:v0.23.0@sha256:2fbdad86bb5244075f31487262d790f6fa75cdd993f868b952252905c118621c
        - quay.io/isovalent/hubble-rbac:v1.3.0@sha256:4c106dd133861a678eb24cf98b35d5023cccca0834c954c49b68514c366aedf0
        - quay.io/isovalent/hubble-enterprise-operator:v1.10.5@sha256:31b18a1ea7002936f04dd070549e0c9a02f8a33394666fc5b0523ab1350c6120
        - quay.io/isovalent/hubble-export-stdout:v1.0.3@sha256:bef600089c1671fce11bb4bcb98b81f25e65152a1a2f13b64ea637cfc5779c11
        - quay.io/isovalent/hubble-export-s3:v1.0.0@sha256:89c178c0a010166c68819ab5de32ab59b64a88aa23a05d8e4f78adb1a4b6c3d3
        - quay.io/isovalent/hubble-export-fluentd:v1.5.1@sha256:eaf42a84821509b69e2ef2b658796062e79a90e63045a98689da3da99a12e492
        - quay.io/isovalent/hubble-enterprise:v1.10.5@sha256:340700d4e28ded2aa9c50ff9f938996bb43faa84e0b2ba38ec79452648d7e5d7
        - quay.io/isovalent/operator-generic:v1.11.21-cee.1@sha256:25b668c1febe1d1a70b63dfe926be4e55fbf77fd186e9df7581e41d7e9813469
        - quay.io/isovalent/cilium:v1.11.21-cee.1@sha256:f41241a15cd96d26727aa1ec413ace643728ddbe42f7fb7b234dbda22bf3950b
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:d3148bf3b9765467936ee067b25b5ef3683d3e55260f517c7fe271ac0e19cb47
        version: 1.11.21+xd0d07b1
      entries:
      - name: cilium.v1.11.21-xd0d07b1
        version: 1.11.21+xd0d07b1
      - name: cilium.v1.11.20-x0f6d966
        version: 1.11.20+x0f6d966
      name: "1.11"
    - currentCSV: cilium.v1.12.21-x3861686
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.12.14 <1.12.21'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/cilium:v1.12.21-cee.1@sha256:3a462308cf12b572db987c03d190ee33691ccb04775d01beb192703a52f8d47b
        - quay.io/isovalent/hubble-relay:v1.12.21-cee.1@sha256:9d8dadc3f330acf0fa169c428d5fdf1d138d0294065eec5c628f6647106fe8b6
        - quay.io/coreos/etcd:v3.5.4@sha256:795d8660c48c439a7c3764c2330ed9222ab5db5bb524d8d0607cac76f7ba82a3
        - quay.io/isovalent/hubble-export-s3:v1.0.0@sha256:89c178c0a010166c68819ab5de32ab59b64a88aa23a05d8e4f78adb1a4b6c3d3
        - quay.io/isovalent/hubble-enterprise-operator:v1.10.5@sha256:31b18a1ea7002936f04dd070549e0c9a02f8a33394666fc5b0523ab1350c6120
        - quay.io/isovalent/clustermesh-apiserver:v1.12.21-cee.1@sha256:3f6549409966086a49e338c7829b9141a73ef2ebeebaac91db430aeb3ccd2e22
        - quay.io/isovalent/hubble-ui-enterprise:v0.28.1@sha256:11dfba0e4fcf43415aa6e8b3b6e98a5c4f468ec67b377328ad48b92037bbfbb0
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:f95c9e8002d8e381d604752b1d59b75595badbdb0b11237ac1cac4ecd04f4800
        - quay.io/isovalent/hubble-ui-enterprise-backend:v0.28.1@sha256:515aeb8166dd6995b04d4cd3333c94a740182943ba4e90a38871ee89c0b64e92
        - quay.io/isovalent/hubble-export-fluentd:v1.5.1@sha256:eaf42a84821509b69e2ef2b658796062e79a90e63045a98689da3da99a12e492
        - quay.io/cilium/certgen:v0.1.8@sha256:4a456552a5f192992a6edcec2febb1c54870d665173a33dc7d876129b199ddbd
        - quay.io/isovalent/hubble-enterprise:v1.10.5@sha256:340700d4e28ded2aa9c50ff9f938996bb43faa84e0b2ba38ec79452648d7e5d7
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/isovalent/hubble-rbac:v1.3.0@sha256:4c106dd133861a678eb24cf98b35d5023cccca0834c954c49b68514c366aedf0
        - quay.io/isovalent/operator-generic:v1.12.21-cee.1@sha256:4e19cda799a747b69f53b5e5a860ba256399d18221d5a4bbf243142ba1c69bb8
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/hubble-export-stdout:v1.0.3@sha256:bef600089c1671fce11bb4bcb98b81f25e65152a1a2f13b64ea637cfc5779c11
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:b6e40d76f0925e76f0a5b181be3f6c62728fa3e4aaef979caf9ea245f4ed03da
        - quay.io/cilium/startup-script:62093c5c233ea914bfa26a10ba41f8780d9b737f@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        version: 1.12.21+x3861686
      entries:
      - name: cilium.v1.12.21-x3861686
        version: 1.12.21+x3861686
      - name: cilium.v1.12.19-xca684e0
        version: 1.12.19+xca684e0
      - name: cilium.v1.12.18-xcb1adcf
        version: 1.12.18+xcb1adcf
      - name: cilium.v1.12.17-xae21c14
        version: 1.12.17+xae21c14
      - name: cilium.v1.12.16-x38e0b62
        version: 1.12.16+x38e0b62
      - name: cilium.v1.12.15-x8ca64a3
        version: 1.12.15+x8ca64a3
      - name: cilium.v1.12.14-xc86b260
        version: 1.12.14+xc86b260
      - name: cilium.v1.12.13-x2b5cec9
        version: 1.12.13+x2b5cec9
      - name: cilium.v1.12.12-xbd7b835
        version: 1.12.12+xbd7b835
      - name: cilium.v1.12.11-xffdea76
        version: 1.12.11+xffdea76
      - name: cilium.v1.12.10-x4a0e9e7
        version: 1.12.10+x4a0e9e7
      - name: cilium.v1.12.9-x4be66e4
        version: 1.12.9+x4be66e4
      - name: cilium.v1.12.8-x87b90a9
        version: 1.12.8+x87b90a9
      - name: cilium.v1.12.7-x53a2aab
        version: 1.12.7+x53a2aab
      name: "1.12"
    - currentCSV: cilium-enterprise.v1.13.21-x96449fb
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.7 <1.13.21+x96449fb'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/startup-script:c54c7edeab7fde4da68e59acd319ab24af242c3f@sha256:a9d0f451353f1b2f29ed3868e230ecfae10150eab9fd521de67adb3c8ca9c33a
        - quay.io/cilium/certgen:v0.1.13@sha256:01802e6a153a9473b06ebade7ee5730f8f2c6cc8db8768508161da3cdd778641
        - quay.io/isovalent/hubble-export-fluentd:v1.11.0@sha256:50cd2e4c2d04e8befcc2ae1a1463f07cba01edb9b546fb455fe7976311165385
        - quay.io/isovalent/hubble-enterprise:v1.11.6@sha256:eb60d59d1fb1b136855829274b31805c4206f8d050c3bf068e63a552576c1b38
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:b04e75456359b50b18e75464da5d7b467fdfeb058fc08ce3a10f057ba5f60107
        - quay.io/isovalent/cilium:v1.13.21-cee.2@sha256:f760cd85fa64eb8a197ec282644a8db630745a8a2854ee08aa2662cf67188db5
        - quay.io/isovalent/hubble-export-s3:v1.0.2@sha256:518daea1074d260d824c3e03a93ab6b2a7a5f4c22b8e07938770add8257f4a52
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:fc7046dee1a204163e68a6d9a42fec200b02129493a7d166f899dd12e2763364
        - quay.io/isovalent/hubble-rbac:v1.5.6@sha256:739ca45ea09829b0fa8a6198901b34eb5254df55be071bc854bbd37cae12a557
        - quay.io/isovalent/hubble-relay:v1.13.21-cee.2@sha256:aaf5f959a76d6b517e627d2d0b09dab86ae942616819a9a95f057ed4f45ea480
        - quay.io/isovalent/operator-generic:v1.13.21-cee.2@sha256:6f2d0d575e3f74395c959833992cd5fbb6971f4e292df1b7cf45be8b07be5e7a
        - quay.io/isovalent/clustermesh-apiserver:v1.13.21-cee.2@sha256:980b0235fd16817ecc4457ad4317a4601993b76df2e81b07a53f6f437c2705af
        - quay.io/isovalent/hubble-export-stdout:v1.0.4@sha256:a0ff77efa890478f18f5deeba2cd1e2dfd6c5c77981adac57b14d7734de71ff1
        - quay.io/isovalent/hubble-ui-enterprise:v1.0.3@sha256:55bf787c96561f4a460e2bfe5966421e2d85f3f9a38ac2487c6f6486c97cd5af
        - quay.io/isovalent/hubble-ui-enterprise-backend:v1.0.3@sha256:6756d2b808c38efce4765f90641d26aa2e0e79329ecc378a123cd3dc2c439640
        - quay.io/isovalent/hubble-enterprise-operator:v1.11.6@sha256:38253c2e17a8b2e88a226f0976c9cf4a00445212e2886e6f07cbb3a52757324f
        - quay.io/coreos/etcd:v3.5.4@sha256:795d8660c48c439a7c3764c2330ed9222ab5db5bb524d8d0607cac76f7ba82a3
        version: 1.13.21+x96449fb
      entries:
      - name: cilium-enterprise.v1.13.21-x96449fb
        version: 1.13.21+x96449fb
      - name: cilium-enterprise.v1.13.20-x5b70650
        version: 1.13.20+x5b70650
      - name: cilium-enterprise.v1.13.19-xdb27142
        version: 1.13.19+xdb27142
      - name: cilium-enterprise.v1.13.18-x0c21a94
        version: 1.13.18+x0c21a94
      - name: cilium-enterprise.v1.13.17-xc8f905f
        version: 1.13.17+xc8f905f
      - name: cilium-enterprise.v1.13.16-xcfe62ed
        version: 1.13.16+xcfe62ed
      - name: cilium-enterprise.v1.13.15-x1dfd26f
        version: 1.13.15+x1dfd26f
      - name: cilium.v1.13.14-xb70933f
        version: 1.13.14+xb70933f
      - name: cilium.v1.13.12-x453f6ad
        version: 1.13.12+x453f6ad
      - name: cilium.v1.13.11-xf6d6935
        version: 1.13.11+xf6d6935
      - name: cilium.v1.13.10-x9d57fc7
        version: 1.13.10+x9d57fc7
      - name: cilium.v1.13.9-xe8a359b
        version: 1.13.9+xe8a359b
      - name: cilium.v1.13.8-x96e1ca9
        version: 1.13.8+x96e1ca9
      - name: cilium.v1.13.7-x7320e7d
        version: 1.13.7+x7320e7d
      - name: cilium.v1.13.6-xd3714ef
        version: 1.13.6+xd3714ef
      - name: cilium.v1.13.5-x8386cae
        version: 1.13.5+x8386cae
      - name: cilium.v1.13.4-xc8b22ea
        version: 1.13.4+xc8b22ea
      - name: cilium.v1.13.3-x69e9561
        version: 1.13.3+x69e9561
      - name: cilium.v1.13.2-x4b1c6c8
        version: 1.13.2+x4b1c6c8
      name: "1.13"
    - currentCSV: cilium-enterprise.v1.14.22-x6d86c88
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.14.22+x6d86c88'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-enterprise:v1.11.6@sha256:eb60d59d1fb1b136855829274b31805c4206f8d050c3bf068e63a552576c1b38
        - quay.io/isovalent/hubble-rbac:v1.5.2@sha256:29309837e2246a1b6849d7a09d01ceebfdb43fa0ade973d14584f0e52b856b95
        - quay.io/coreos/etcd:v3.5.4@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:148b1029eac14049e2ba84d4d75be1ca43ef2d89951adbdeddec392ad7c9fe1e
        - quay.io/isovalent/hubble-ui-enterprise:v1.2.2@sha256:3e03a58f6eea5d2df195f42ee2817f7b28fa20bbe140f5086937eee8da1b8d56
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/isovalent/hubble-enterprise-operator:v1.11.6@sha256:38253c2e17a8b2e88a226f0976c9cf4a00445212e2886e6f07cbb3a52757324f
        - quay.io/cilium/startup-script:c54c7edeab7fde4da68e59acd319ab24af242c3f@sha256:a9d0f451353f1b2f29ed3868e230ecfae10150eab9fd521de67adb3c8ca9c33a
        - quay.io/isovalent/operator-generic:v1.14.22-cee.1@sha256:a997fe24ff0b7f7df3ee12748413ae53d6e07c803162b0db4d812b916b0cfc74
        - quay.io/isovalent/hubble-export-fluentd:v1.8.1@sha256:ce0cce63a8473d5ae73872fc65d82c54f2875511bb03f0bc8ee8480148c43ee1
        - quay.io/cilium/certgen:v0.1.17@sha256:07959f3125dae659ab9036c14df046e2536f7a4f6c92dabda837388b0344dc74
        - quay.io/isovalent/hubble-relay:v1.14.22-cee.1@sha256:d7ffc880d2e42e0876417ec26115b0e9155ad9a2e7f47c02c434746afa42b29a
        - quay.io/isovalent/clustermesh-apiserver:v1.14.22-cee.1@sha256:588e2b173ab409288eda2d404571184562d3d9fd12e2d9afe7282a5a7d79113a
        - quay.io/cilium/cilium-envoy:v1.32.5-1743506100-0821ef0acdf9f824d47d34e02932be522b3e7233@sha256:305c8b7da56ef1b64e4deb0642acf93a3944f2712e40c99a40bb005e6dc9761b
        - quay.io/isovalent/hubble-ui-enterprise-backend:v1.2.2@sha256:4258f7e0bff24ff1e0b214f62b37097032b6426ba28edd3dfc4f451a34de0110
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:fc7046dee1a204163e68a6d9a42fec200b02129493a7d166f899dd12e2763364
        - quay.io/isovalent/hubble-export-stdout:v1.0.4@sha256:a0ff77efa890478f18f5deeba2cd1e2dfd6c5c77981adac57b14d7734de71ff1
        - quay.io/isovalent/cilium:v1.14.22-cee.1@sha256:0e664a5a9e0edba38ab1a25eb7fa50861bb15f69ed0bbe3f3c792bd909240da3
        - quay.io/isovalent/hubble-export-s3:v1.0.2@sha256:518daea1074d260d824c3e03a93ab6b2a7a5f4c22b8e07938770add8257f4a52
        version: 1.14.22+x6d86c88
      entries:
      - name: cilium-enterprise.v1.14.22-x6d86c88
        version: 1.14.22+x6d86c88
      - name: cilium-enterprise.v1.14.21-xe0c9553
        version: 1.14.21+xe0c9553
      - name: cilium-enterprise.v1.14.20-xe01e743
        version: 1.14.20+xe01e743
      - name: cilium-enterprise.v1.14.19-xc1b9aff
        version: 1.14.19+xc1b9aff
      name: "1.14"
    - currentCSV: cilium-enterprise.v1.15.21-x2661a01
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.15.21+x2661a01'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:bd1a4d47ef48a6eb84c49d979e8f5ce867af5932828dfd3233fa9f78c2eac1f4
        - quay.io/isovalent/cilium-envoy-ubi:v1.33.4-1752151664-7c2edb0b44cf95f326d628b837fcdd845102ba68@sha256:e81fdde48c3d5cd93111e1d51d9f8554adb32ee937601a9124354e0894933b99
        - quay.io/isovalent/operator-generic-ubi:v1.15.21-cee.1@sha256:97daac94735e497e1e897beaa01eb50866abace1aa9a3919fc57971e45f538fe
        - quay.io/isovalent/hubble-ui-frontend-ubi:v1.3.7@sha256:f2b4e935abe3eb94f82dab41c9d3d929fca71d82fa39a6c8002683375b5303a4
        - quay.io/isovalent/oauth2-proxy:7.12.0-20250819@sha256:fe8f76728c36536d498b2a5c11172cb18dc30bbb555df2a6d14c036b29cb9f10
        - quay.io/isovalent/cilium:v1.15.21-cee.1@sha256:39ef22d467783481eccb376fcb5d6bc254ed8f4b3b25032121498db40278f0b8
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.15.21-cee.1@sha256:36d00eb9f8ed45a3d8892942cfae097782dd6509889ef7d1a451f9b0334d17e1
        - quay.io/isovalent/certgen-ubi:v0.1.19@sha256:c32e96b0972c9bcd3896945c46ea23e330c9dea407afee3a637daae637296f4c
        - quay.io/isovalent/hubble-ui-backend-ubi:v1.3.7@sha256:a4b5e7a032c1bbcd969a6873bc12935f8b6aa4c2d3a2ed87f5494fcfcdfd02d4
        - quay.io/isovalent/hubble-relay-ubi:v1.15.21-cee.1@sha256:8461f720ca95a9c543a298bee1c1e56241f56d5426030c13afed8262f2f5c5ff
        - quay.io/isovalent/startup-script-ubi:1755531540-60ee83e@sha256:2e1d5048cbfcd3c77f732da8203a6c792cbf19897b5cf137b9ad23a6a267f6b2
        version: 1.15.21+x2661a01
      entries:
      - name: cilium-enterprise.v1.15.21-x2661a01
        version: 1.15.21+x2661a01
      - name: cilium-enterprise.v1.15.20-x6df37bb
        version: 1.15.20+x6df37bb
      - name: cilium-enterprise.v1.15.19-x0be49ab
        version: 1.15.19+x0be49ab
      - name: cilium-enterprise.v1.15.18-xeeaa70b
        version: 1.15.18+xeeaa70b
      - name: cilium-enterprise.v1.15.17-xdfd9482
        version: 1.15.17+xdfd9482
      - name: cilium-enterprise.v1.15.16-xc7e31ca
        version: 1.15.16+xc7e31ca
      - name: cilium-enterprise.v1.15.15-xf37f22c
        version: 1.15.15+xf37f22c
      - name: cilium-enterprise.v1.15.14-x0e512e3
        version: 1.15.14+x0e512e3
      - name: cilium-enterprise.v1.15.13-x134fa29
        version: 1.15.13+x134fa29
      name: "1.15"
    - currentCSV: cilium-enterprise.v1.16.15-x6419e26
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.16.0 <1.16.15+x6419e26'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/startup-script-ubi:v1.16.15-cee.1@sha256:5d59a46be44c4d899576e69e0f20e148245e17d905a0769a73fd01f2432d1bab
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:ec8a17956e75e91b60e2999d10e67ef76f9b4cfd754db415aaf387d21690da87
        - quay.io/isovalent/cilium-envoy-ubi:v1.16.15-cee.1@sha256:843e6fdf6845f3d5f841dbd2dce75224b16c9d34c2b0ce7f329d9666466bfc6b
        - quay.io/isovalent/cilium-ubi:v1.16.15-cee.1@sha256:1d7bbbcb613ef20cd1719bd77473e71c198960ee5d8e563aec9c65cfd3c2efe2
        - quay.io/isovalent/hubble-relay-ubi:v1.16.15-cee.1@sha256:c992d549627e8f9180c272e7d96ff3783de184796633ed5d0b6fe2cf7e74055d
        - quay.io/isovalent/operator-generic-ubi:v1.16.15-cee.1@sha256:970aa156b6f5c390b6835a626190e18656f02db6a5cf27aaa1ac9c7872f4b15a
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.16.15-cee.1@sha256:f7573ab4145d306051b5d6fa7ff1af395fc71935f5204839ab1c9f574aee17ad
        - quay.io/isovalent/certgen-ubi:v1.16.15-cee.1@sha256:e3f1fa5ee59887d3a15c43947a853c8a3f4fc650ed72bcd891b277037ea16a9d
        version: 1.16.15+x6419e26
      entries:
      - name: cilium-enterprise.v1.16.15-x6419e26
        version: 1.16.15+x6419e26
      - name: cilium-enterprise.v1.16.13-xb2de075
        version: 1.16.13+xb2de075
      - name: cilium-enterprise.v1.16.12-xd373871
        version: 1.16.12+xd373871
      - name: cilium-enterprise.v1.16.11-xbd1cd23
        version: 1.16.11+xbd1cd23
      - name: cilium-enterprise.v1.16.10-x89ab4b0
        version: 1.16.10+x89ab4b0
      - name: cilium-enterprise.v1.16.9-x8ec4137
        version: 1.16.9+x8ec4137
      - name: cilium-enterprise.v1.16.8-yd2b50b4
        version: 1.16.8+yd2b50b4
      - name: cilium-enterprise.v1.16.7-x4867421
        version: 1.16.7+x4867421
      - name: cilium-enterprise.v1.16.6-x7e6eb1c
        version: 1.16.6+x7e6eb1c
      name: "1.16"
    defaultChannel: "1.16"
    packageName: cilium-enterprise
    provider:
      name: Isovalent
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-commons-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: commons-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:6768e4189fa8deebf337e4732e5f12bb2ad4acdc112d9d05f5dc1a4ed81f08de
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for S3ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for S3BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/commons-operator@sha256:6768e4189fa8deebf337e4732e5f12bb2ad4acdc112d9d05f5dc1a4ed81f08de
        version: 24.11.1
      entries:
      - name: commons-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: commons-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:adc2b25a676f7357ca3971c13ecf5c9ed3cb1af24ffc902b03733cc814e8946f
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for S3ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for S3BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/commons-operator@sha256:adc2b25a676f7357ca3971c13ecf5c9ed3cb1af24ffc902b03733cc814e8946f
        version: 25.3.0
      entries:
      - name: commons-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: commons-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
        version: 25.7.0
      entries:
      - name: commons-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: commons-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
        version: 25.7.0
      entries:
      - name: commons-operator.v25.7.0
        version: 25.7.0
      - name: commons-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-commons-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: web-terminal
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: web-terminal.v1.11.1-0.1747215995.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: registry.redhat.io/web-terminal/web-terminal-rhel9-operator@sha256:0478d14e92df84fdfbf4584384d34dc9a71427a6487c2564d2eb7815ba1ac12b
          createdAt: "2021-10-26T07:24:32Z"
          description: Start a web terminal in your browser with common CLI tools
            for interacting with the cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.substitutesFor: web-terminal.v1.11.1
          operatorframework.io/suggested-namespace: openshift-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/redhat-developer/web-terminal-operator/
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - kind: DevWorkspaceRouting
            name: devworkspaceroutings.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha1
        description: |
          Start a web terminal in your browser with common CLI tools for interacting with
          the cluster.

          **Note:** The Web Terminal Operator integrates with the OpenShift Console in
          OpenShift 4.5.3 and higher to simplify web terminal instance creation and
          automate OpenShift login. In earlier versions of OpenShift, the operator can
          be installed but web terminals will have to be created and accessed manually.

          ## Description
          The Web Terminal Operator leverages the
          [DevWorkspace Operator](https://github.com/devfile/devworkspace-operator)
          to provision enviroments which support common cloud CLI tools. When this
          operator is installed, the DevWorkspace Operator will be installed as a
          dependency.

          ## How to Install
          Press the **Install** button, choose the upgrade strategy, and wait for the
          **Installed** Operator status.

          When the operator is installed, you will see a terminal button appear on the
          top right of the console after refreshing the OpenShift console window.

          ## How to Uninstall
          The Web Terminal Operator requires manual steps to fully uninstall the operator.
          As the Web Terminal Operator is designed as a way to access the OpenShift
          cluster, web terminal instances store user credentials. To avoid exposing these
          credentials to unwanted parties, the operator deploys webhooks and finalizers
          that aren't removed when the operator is uninstalled. See the
          [uninstall guide](https://docs.openshift.com/container-platform/latest/web_console/web_terminal/uninstalling-web-terminal.html)
          for more details.

          ## Documentation
          Documentation for this Operator is available at https://docs.openshift.com/container-platform/latest/web\_console/web\_terminal/installing-web-terminal.html
        displayName: Web Terminal
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - workspace
        - devtools
        - developer
        - ide
        - terminal
        links:
        - name: Web Terminal Repo
          url: https://github.com/redhat-developer/web-terminal-operator/
        maintainers:
        - email: aobuchow@redhat.com
          name: Andrew Obuchowicz
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/web-terminal/web-terminal-tooling-rhel9@sha256:5c24220f884dcdf1b1e5ac1e20dc6b7c8c4300bb89e8f118c0b11c331a56ab3f
        - registry.redhat.io/web-terminal/web-terminal-exec-rhel9@sha256:cfc8200340655a045f45d02fa327538f87e98d6369bdef2b46cf447053c44426
        - registry.redhat.io/web-terminal/web-terminal-rhel9-operator@sha256:0478d14e92df84fdfbf4584384d34dc9a71427a6487c2564d2eb7815ba1ac12b
        version: 1.11.1+0.1747215995.p
      entries:
      - name: web-terminal.v1.11.1-0.1747215995.p
        version: 1.11.1+0.1747215995.p
      - name: web-terminal.v1.11.1
        version: 1.11.1
      - name: web-terminal.v1.11.0
        version: 1.11.0
      - name: web-terminal.v1.10.1
        version: 1.10.1
      - name: web-terminal.v1.10.1-0.1740684238.p
        version: 1.10.1+0.1740684238.p
      - name: web-terminal.v1.10.0-0.1731481377.p
        version: 1.10.0+0.1731481377.p
      - name: web-terminal.v1.10.0-0.1732652667.p
        version: 1.10.0+0.1732652667.p
      - name: web-terminal.v1.10.0
        version: 1.10.0
      - name: web-terminal.v1.10.0-0.1727169028.p
        version: 1.10.0+0.1727169028.p
      - name: web-terminal.v1.10.0-0.1720435222.p
        version: 1.10.0+0.1720435222.p
      - name: web-terminal.v1.10.0-0.1720402943.p
        version: 1.10.0+0.1720402943.p
      - name: web-terminal.v1.9.0-0.1708477317.p
        version: 1.9.0+0.1708477317.p
      - name: web-terminal.v1.9.0
        version: 1.9.0
      - name: web-terminal.v1.8.0-0.1708477299.p
        version: 1.8.0+0.1708477299.p
      - name: web-terminal.v1.8.0-0.1701199376.p
        version: 1.8.0+0.1701199376.p
      - name: web-terminal.v1.8.0-0.1692219820.p
        version: 1.8.0+0.1692219820.p
      - name: web-terminal.v1.8.0
        version: 1.8.0
      - name: web-terminal.v1.7.0-0.1682321121.p
        version: 1.7.0+0.1682321121.p
      - name: web-terminal.v1.7.0-0.1692219820.p
        version: 1.7.0+0.1692219820.p
      - name: web-terminal.v1.7.0
        version: 1.7.0
      - name: web-terminal.v1.7.0-0.1681197295.p
        version: 1.7.0+0.1681197295.p
      - name: web-terminal.v1.7.0-0.1708477265.p
        version: 1.7.0+0.1708477265.p
      - name: web-terminal.v1.7.0-0.1684429884.p
        version: 1.7.0+0.1684429884.p
      - name: web-terminal.v1.6.0
        version: 1.6.0
      - name: web-terminal.v1.6.0-0.1692219820.p
        version: 1.6.0+0.1692219820.p
      - name: web-terminal.v1.5.1-0.1661829403.p
        version: 1.5.1+0.1661829403.p
      - name: web-terminal.v1.5.1
        version: 1.5.1
      - name: web-terminal.v1.5.0-0.1657220207.p
        version: 1.5.0+0.1657220207.p
      - name: web-terminal.v1.5.0
        version: 1.5.0
      - name: web-terminal.v1.4.0
        version: 1.4.0
      - name: web-terminal.v1.3.0
        version: 1.3.0
      name: fast
    defaultChannel: fast
    packageName: web-terminal
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM Quantum
      provider-url: ""
    name: openshift-qiskit-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-qiskit-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "qiskit.ibm.com/v1alpha1",
                "kind": "QiskitPlayground",
                "metadata": {
                  "name": "qiskitplayground-sample"
                },
                "spec": {
                  "image": "qiskit/jupyter:0.1",
                  "imagePullPolicy": "Always",
                  "resources": {
                    "requests": {
                      "memory": "2G",
                      "cpu": "1"
                    },
                    "limits": {
                      "memory": "3G",
                      "cpu": "1.5"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: docker.io/qiskit/operator:0.1
          createdAt: 2021-004-23 22:25:53
          description: |-
            Launches a developement environment Jupyter notebok development environment for
            developing quantum applications using Qiskit.
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator
          support: https://github.com/qiskit-community/openshift-quantum-operators/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QiskitPlayground is the Schema for the qiskitplaygrounds
              API
            displayName: Qiskit Playground
            kind: QiskitPlayground
            name: qiskitplaygrounds.qiskit.ibm.com
            version: v1alpha1
        description: |-
          ### Introduction
          Launches a developement Jupiter lab environment with Qiskit.

          QiskitPlayground is a Jupyter notebook with the pre installed qiskit libraries for implementing quantum application development.
          It comes with pre installed python packages for visualising results. QiskitPlayground can run quantum
          applications on simulator or on real quantum machines using IBM Q account.

          ### Documentation
          For detailed installation guide [follow](https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator)

          ### Contributing
          You can contribute by
            * Raising any [issues](https://github.com/qiskit-community/openshift-quantum-operators/issues) you find using ibm-quantum-operator community operator
            * Fixing issues by opening [Pull Requests](https://github.com/qiskit-community/openshift-quantum-operators/pulls)
            * Talking about openshift-qiskit-operator

          ### License
          kubernetes-qiskit-operator is licensed under the [Apache 2.0 license](https://github.com/qiskit-community/openshift-quantum-operators/blob/master/LICENSE)

          #### Note
          Any api could be changed any time without any notice. That said, your feedback is very important and appreciated to make this project more stable and useful.
        displayName: QiskitPlayground
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Qiskit
        - Open Source
        - Quantum
        links:
        - name: GitHub
          url: https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator
        maintainers:
        - email: blublinsky@ibm.com
          name: boris
        maturity: alpha
        provider:
          name: IBM Quantum
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - docker.io/qiskit/operator:0.1
        version: 0.2.1
      entries:
      - name: openshift-qiskit-operator.v0.2.1
        version: 0.2.1
      - name: openshift-qiskit-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: openshift-qiskit-operator
    provider:
      name: IBM Quantum
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: servicemeshoperator3
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: servicemeshoperator3.v3.0.0-tp.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh-tech-preview/istio-rhel9-operator@sha256:f5dfe495ba92ecb880485b66cfafbf6451f8142b8e0d5fd7f29db94171c51336
          createdAt: "2024-12-12T10:28:38Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1alpha1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1alpha1
          - description: IstioRevisionTag references a Istio or IstioRevision object
              and serves as an alias for sidecar injection.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1alpha1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-rhel9-operator@sha256:f5dfe495ba92ecb880485b66cfafbf6451f8142b8e0d5fd7f29db94171c51336
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-pilot-rhel9@sha256:376f64eab0ebeb06035983a60d1e7647c5d41354f31a49cb34f4deb7b6708e3e
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-pilot-rhel9@sha256:ea9372fabe06cd8e65ea7b4567ce3c8fbddd9d673dc34a12743fd95c548059d1
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-proxyv2-rhel9@sha256:c21481d6dae58b82dac93b1b47b9d9f53a30617fa9d930a0b5d075620cffa74b
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-must-gather-rhel9@sha256:2bf3704ba8d5945973ccf6bd13418457e929836352acf0ea6a47976ffa246616
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-proxyv2-rhel9@sha256:cbd6bc47341ab8618609f89d063552e17232541c727429855c01deab73e4d06c
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-must-gather-rhel9@sha256:e3978aecaafe2678355e21a7303f8620f53d8b774f6592f10527c15bae957224
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-cni-rhel9@sha256:d0ab36246c556249b6d67346eb0c72ebd22483a6502caa0fc75112a8d327b053
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-cni-rhel9@sha256:f16cd8a29ddcbb3de77cbfcf7e1da78e42cc0c1c502132d477a01c52689cdf23
        version: 3.0.0-tp.2
      entries:
      - name: servicemeshoperator3.v3.0.0-tp.2
        version: 3.0.0-tp.2
      - name: servicemeshoperator3.v3.0.0-tp.1
        version: 3.0.0-tp.1
      name: candidates
    - currentCSV: servicemeshoperator3.v3.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.4"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.4"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
          createdAt: "2025-09-10T06:22:42Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:366266f10e658c4cea86bddf6ee847e1908ea4dd780cb5f7fb0e466bac8995f1
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:0d6930075b8906bee0ab1aeac82f6e2431c16163cf87f7f0c68c10b2c40552a1
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:26747627ae22bbdffcf9de58077454fc0c890cda83659d7834b15dea2b5aaaf2
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2ea112ab90b8540f11e9949d77c3e7e3b3ef57ac3bf23f6cf1e883a88430e1f9
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:028e10651db0d1ddb769a27c9483c6d41be6ac597f253afd9d599f395d9c82d8
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:ecc6a22a471f4f4b836648f3a72ea7f57a4ea960ebcbc7dbe3fe729500902d0b
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d518f3d1539f45e1253c5c9fa22062802804601d4998cd50344e476a3cc388fe
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:33feac0c9c888b375e207e86b3f9a1a289126ca46b64632a7f693e2191cfbbfd
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:2ca96716812d339775d3dfd396c863834dd3810720e98bfcf87bc67c5fbd29b5
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:6bb44c84363a4113c5bce2ae0d6528b345fddd12bc559bbfaacf6222060ddcb0
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:6881fb9f5a4b447be4c468b8585fff86171bd3d82671a48559f3ceabcb9dad24
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:14c5a52faf20267baa43d9a72ee6416f56fbc41dd640adc2aba3c4043802a0e9
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:b757b81d34711cb8abfcb0e8896c023d39d4778f237dd80b637ba4bf613df893
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:f7ab868455b5e887e2d89b7078678531290365f9166cae99ce56119f31fa2ba4
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d757e700a124d7583b9456ec9e08e5e7e249c1de202bc4e973102922b9f69191
        version: 3.1.2
      entries:
      - name: servicemeshoperator3.v3.1.2
        version: 3.1.2
      - name: servicemeshoperator3.v3.1.1
        version: 3.1.1
      - name: servicemeshoperator3.v3.1.0
        version: 3.1.0
      - name: servicemeshoperator3.v3.0.5
        version: 3.0.5
      - name: servicemeshoperator3.v3.0.4
        version: 3.0.4
      - name: servicemeshoperator3.v3.0.3
        version: 3.0.3
      - name: servicemeshoperator3.v3.0.2
        version: 3.0.2
      - name: servicemeshoperator3.v3.0.1
        version: 3.0.1
      - name: servicemeshoperator3.v3.0.0
        version: 3.0.0
      name: stable
    - currentCSV: servicemeshoperator3.v3.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.6"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.6"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:d716f3558fbdf8514f225de708135ba87fdcac4c718ef101f394daf546514afb
          createdAt: "2025-09-03T07:53:34Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:d716f3558fbdf8514f225de708135ba87fdcac4c718ef101f394daf546514afb
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        version: 3.0.5
      entries:
      - name: servicemeshoperator3.v3.0.5
        version: 3.0.5
      - name: servicemeshoperator3.v3.0.4
        version: 3.0.4
      - name: servicemeshoperator3.v3.0.3
        version: 3.0.3
      - name: servicemeshoperator3.v3.0.2
        version: 3.0.2
      - name: servicemeshoperator3.v3.0.1
        version: 3.0.1
      - name: servicemeshoperator3.v3.0.0
        version: 3.0.0
      name: stable-3.0
    - currentCSV: servicemeshoperator3.v3.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.4"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.4"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
          createdAt: "2025-09-10T06:22:42Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:028e10651db0d1ddb769a27c9483c6d41be6ac597f253afd9d599f395d9c82d8
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d518f3d1539f45e1253c5c9fa22062802804601d4998cd50344e476a3cc388fe
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:366266f10e658c4cea86bddf6ee847e1908ea4dd780cb5f7fb0e466bac8995f1
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:33feac0c9c888b375e207e86b3f9a1a289126ca46b64632a7f693e2191cfbbfd
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:ecc6a22a471f4f4b836648f3a72ea7f57a4ea960ebcbc7dbe3fe729500902d0b
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:b757b81d34711cb8abfcb0e8896c023d39d4778f237dd80b637ba4bf613df893
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:14c5a52faf20267baa43d9a72ee6416f56fbc41dd640adc2aba3c4043802a0e9
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:6bb44c84363a4113c5bce2ae0d6528b345fddd12bc559bbfaacf6222060ddcb0
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:0d6930075b8906bee0ab1aeac82f6e2431c16163cf87f7f0c68c10b2c40552a1
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:6881fb9f5a4b447be4c468b8585fff86171bd3d82671a48559f3ceabcb9dad24
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d757e700a124d7583b9456ec9e08e5e7e249c1de202bc4e973102922b9f69191
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2ea112ab90b8540f11e9949d77c3e7e3b3ef57ac3bf23f6cf1e883a88430e1f9
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:26747627ae22bbdffcf9de58077454fc0c890cda83659d7834b15dea2b5aaaf2
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:f7ab868455b5e887e2d89b7078678531290365f9166cae99ce56119f31fa2ba4
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:2ca96716812d339775d3dfd396c863834dd3810720e98bfcf87bc67c5fbd29b5
        version: 3.1.2
      entries:
      - name: servicemeshoperator3.v3.1.2
        version: 3.1.2
      - name: servicemeshoperator3.v3.1.1
        version: 3.1.1
      - name: servicemeshoperator3.v3.1.0
        version: 3.1.0
      name: stable-3.1
    defaultChannel: stable
    packageName: servicemeshoperator3
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: heat-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: heat-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "heat.openstack.org/v1beta1",
                "kind": "Heat",
                "metadata": {
                  "name": "heat"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = True\n",
                  "databaseInstance": "openstack",
                  "heatAPI": {},
                  "heatCfnAPI": {},
                  "heatEngine": {},
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "heat.openstack.org/v1beta1",
                "kind": "HeatAPI",
                "metadata": {
                  "name": "heat-api"
                },
                "spec": {
                  "databaseAccount": "heat",
                  "databaseHostname": "openstack",
                  "passwordSelectors": {
                    "authEncryptionKey": "HeatAuthEncryptionKey",
                    "service": "HeatPassword"
                  },
                  "replicas": 1,
                  "resources": {},
                  "secret": "osp-secret",
                  "serviceAccount": "heat-heat",
                  "serviceUser": "heat",
                  "transportURLSecret": "rabbitmq-transport-url-heat-heat-transport"
                }
              },
              {
                "apiVersion": "heat.openstack.org/v1beta1",
                "kind": "HeatCfnAPI",
                "metadata": {
                  "name": "heat-cfnapi"
                },
                "spec": {
                  "databaseAccount": "heat",
                  "databaseHostname": "openstack",
                  "passwordSelectors": {
                    "authEncryptionKey": "HeatAuthEncryptionKey",
                    "service": "HeatPassword"
                  },
                  "replicas": 1,
                  "resources": {},
                  "secret": "osp-secret",
                  "serviceAccount": "heat-heat",
                  "serviceUser": "heat",
                  "transportURLSecret": "rabbitmq-transport-url-heat-heat-transport"
                }
              },
              {
                "apiVersion": "heat.openstack.org/v1beta1",
                "kind": "HeatEngine",
                "metadata": {
                  "name": "heat-engine"
                },
                "spec": {
                  "databaseAccount": "heat",
                  "databaseHostname": "openstack",
                  "passwordSelectors": {
                    "authEncryptionKey": "HeatAuthEncryptionKey",
                    "service": "HeatPassword"
                  },
                  "replicas": 1,
                  "resources": {},
                  "secret": "osp-secret",
                  "serviceAccount": "heat-heat",
                  "serviceUser": "heat",
                  "transportURLSecret": "rabbitmq-transport-url-heat-heat-transport"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:16Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HeatAPI is the Schema for the heatapis API
            displayName: Heat API
            kind: HeatAPI
            name: heatapis.heat.openstack.org
            version: v1beta1
          - description: HeatCfnAPI is the Schema for the heatcfnapis API
            displayName: Heat Cfn API
            kind: HeatCfnAPI
            name: heatcfnapis.heat.openstack.org
            version: v1beta1
          - description: HeatEngine is the Schema for the heatengines API
            displayName: Heat Engine
            kind: HeatEngine
            name: heatengines.heat.openstack.org
            version: v1beta1
          - description: Heat is the Schema for the heats API
            displayName: Heat
            kind: Heat
            name: heats.heat.openstack.org
            version: v1beta1
        description: Heat Operator
        displayName: Heat Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Orchestration
        - Heat
        links:
        - name: Heat Operator
          url: https://github.com/openstack-k8s-operators/heat-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/heat-rhel9-operator@sha256:fa61aeee8d63087d150a7641d1e8d4b67dd067bcfed6a140401fe396f1e32403
        - registry.redhat.io/rhoso/openstack-heat-api-rhel9@sha256:dba236e5fed89fbe46397b6ab01de5795e66e9194f5cd99fdf137c37a7b96135
        - registry.redhat.io/rhoso/openstack-heat-api-cfn-rhel9@sha256:95548184e68325a5fe1604d5c4d35d1e9cca7f39b84303fce50479e0e35f8280
        - registry.redhat.io/rhoso/openstack-heat-engine-rhel9@sha256:0b6fc782986bf11b26c0839fd81c52257f4a31aa5e5a6728ce251f3670537aba
        version: 1.0.6
      entries:
      - name: heat-operator.v1.0.6
        version: 1.0.6
      - name: heat-operator.v1.0.5
        version: 1.0.5
      - name: heat-operator.v1.0.4
        version: 1.0.4
      - name: heat-operator.v1.0.3
        version: 1.0.3
      - name: heat-operator.v1.0.2
        version: 1.0.2
      - name: heat-operator.v1.0.1
        version: 1.0.1
      - name: heat-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: heat-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CrowdStrike
      provider-url: https://crowdStrike.com
    name: falcon-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: falcon-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconAdmission",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-admission",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconadmission",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-kac"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "admission_controller"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconContainer",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-sidecar-sensor",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconcontainer",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-container-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "sidecar"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconDeployment",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcondeployment-sample",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falcon-deployment",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-deployment"
                },
                "spec": {
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconImageAnalyzer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "falcon-operator",
                    "app.kubernetes.io/instance": "falconimageanalyzer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "falconimageanalyzer",
                    "app.kubernetes.io/part-of": "falcon-operator"
                  },
                  "name": "falcon-image-analyzer"
                },
                "spec": {
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconNodeSensor",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-node-sensor",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconnodesensor",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-node-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "daemonset"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security,Monitoring
          containerImage: registry.connect.redhat.com/crowdstrike/falcon-operator@sha256:ffe00e9e22e209d1e08f29c3d257bbd51937fcbde68a1ed93b542a26dd436dee
          createdAt: "2025-06-03T22:34:20Z"
          description: Falcon Operator installs CrowdStrike Falcon Sensors on the
            cluster
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: falcon-operator
          operators.openshift.io/valid-subscription: Not required to use this operator
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: registry.connect.redhat.com/crowdstrike/falcon-operator
          support: CrowdStrike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FalconAdmission is the Schema for the falconadmissions API
            displayName: Falcon Admission
            kind: FalconAdmission
            name: falconadmissions.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconContainer is the Schema for the falconcontainers API
            displayName: Falcon Container
            kind: FalconContainer
            name: falconcontainers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconDeployment is the Schema for the falcondeployments
              API
            displayName: Falcon Deployment
            kind: FalconDeployment
            name: falcondeployments.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconImageAnalyzer is the Schema for the falconImageAnalyzers
              API
            displayName: Falcon Image Analyzer
            kind: FalconImageAnalyzer
            name: falconimageanalyzers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconNodeSensor is the Schema for the falconnodesensors
              API
            displayName: Falcon Node Sensor
            kind: FalconNodeSensor
            name: falconnodesensors.falcon.crowdstrike.com
            version: v1alpha1
        description: |-
          The CrowdStrike Falcon Operator installs CrowdStrike Falcon custom resources on a Kubernetes cluster.

          ## About the CrowdStrike Falcon Operator
          The CrowdStrike Falcon Operator deploys CrowdStrike Falcon to the cluster. The operator exposes custom resources that allow you to protect your Kubernetes clusters when deployed.

          ## About Custom Resources

          | Custom Resource                                       | Description                                                      |
          | :--                                             | :--                                                    |
          | [FalconAdmission](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/admission/README.md) | Manages installation of Falcon Admission Controller on the cluster |
          | [FalconImageAnalyzer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/imageanalyzer/README.md) | Manages installation of Falcon Image Assessment at Runtime on the cluster |
          | [FalconContainer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/container/README.md) | Manages installation of Falcon Container Sensor on the cluster   |
          | [FalconNodeSensor](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/node/README.md)     | Manages installation of Falcon Linux Sensor on the cluster nodes |
          | [FalconDeployment](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/falcondeployment/README.md)       | Deploys FalconAdmission, FalconImageAnalyzer, FalconContainer, and FalconNodeSensor CRs from a single manifest |

          ## Installation and Deployment

          For installation and deployment of the CrowdStrike Falcon Operator and its Custom Resources, please read the [Installation and Deployment Guide](./docs/install_guide.md) and choose the deployment method that is right for your target environment.

          ## Getting Help
          If you encounter any issues while using the Falcon Operator, you can create an issue on our [Github repo](https://github.com/CrowdStrike/falcon-operator) for bugs, enhancements, or other requests.

          ## Contributing
          You can contribute by:

          * Raising any issues you find using Falcon Operator
          * Fixing issues by opening [Pull Requests](https://github.com/CrowdStrike/falcon-operator/pulls)
          * Improving documentation
          * Talking about the CrowdStrike Falcon Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/CrowdStrike/falcon-operator/issues).

          ## Additional Resources
           - CrowdStrike Container Security: [Product Page](https://www.crowdstrike.com/products/cloud-security/falcon-cloud-workload-protection/container-security/)
           - So You Think Your Containers Are Secure? Four Steps to Ensure a Secure Container Deployment: [Blog Post](https://www.crowdstrike.com/blog/four-steps-to-ensure-a-secure-containter-deployment/)
           - Container Security With CrowdStrike: [Blog Post](https://www.crowdstrike.com/blog/tech-center/container-security/)
           - To learn more about Falcon Container Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/support/documentation/146/falcon-container-sensor-for-linux), [Release Notes](https://falcon.crowdstrike.com/support/news/release-notes-falcon-container-sensor-for-linux)
           - To learn more about Falcon Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/documentation/20/falcon-sensor-for-linux#kubernetes-support)
           - [Developer Documentation](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/developer_guide.md)
        displayName: CrowdStrike Falcon Platform - Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - runtime-protection
        links:
        - name: Falcon Operator
          url: https://github.com/CrowdStrike/falcon-operator
        maintainers:
        - email: integrations@crowdstrike.com
          name: CrowdStrike Solutions Architecture
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: CrowdStrike
          url: https://crowdStrike.com
        relatedImages:
        - registry.connect.redhat.com/crowdstrike/falcon-operator@sha256:ffe00e9e22e209d1e08f29c3d257bbd51937fcbde68a1ed93b542a26dd436dee
        - registry.crowdstrike.com/falcon-container/us-1/release/falcon-sensor@sha256:92d222cd604c9f2c9e9f2c9b402b9fb0858fcb04dac38c15e25125978f43c99b
        - registry.crowdstrike.com/falcon-kac/us-1/release/falcon-kac@sha256:5049128dcfe9565f39fc6e28fb94f5bf05fb7c51cf787fc6fe5c63a2b04f800d
        - registry.crowdstrike.com/falcon-imageanalyzer/us-1/release/falcon-imageanalyzer@sha256:f08603648658041ad52ccf0cfd058da40b385904830030aa82dc50609aed7ceb
        - registry.crowdstrike.com/falcon-sensor/us-1/release/falcon-sensor@sha256:ed4f3fe8d9f5f491600f18d4440dc1d8e3a07d4b9c08a1f17382a23a3a5cf523
        version: 1.8.0
      entries:
      - name: falcon-operator.v1.8.0
        version: 1.8.0
      - name: falcon-operator.v1.7.0
        version: 1.7.0
      - name: falcon-operator.v1.6.0
        version: 1.6.0
      - name: falcon-operator.v1.5.0
        version: 1.5.0
      - name: falcon-operator.v1.4.0
        version: 1.4.0
      - name: falcon-operator.v1.3.1
        version: 1.3.1
      - name: falcon-operator.v1.3.0
        version: 1.3.0
      - name: falcon-operator.v1.2.0
        version: 1.2.0
      - name: falcon-operator.v1.1.0
        version: 1.1.0
      - name: falcon-operator.v1.0.1
        version: 1.0.1
      - name: falcon-operator.v1.0.0
        version: 1.0.0
      name: certified-1.0
    defaultChannel: certified-1.0
    packageName: falcon-operator
    provider:
      name: CrowdStrike
      url: https://crowdStrike.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Airlock
      provider-url: ""
    name: airlock-microgateway
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airlock-microgateway.v4.5.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControl","metadata":{"name":"access-control-example"},"spec":{"policies":[{"requestConditions":{"path":{"matcher":{"exact":"/secure"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}},{"oidc":{"claim":{"name":"roles","value":{"matcher":{"contains":"\"admin\""}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"oidc-example"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"identity-propagation-example"}}],"onFailure":"Pass"}},{"authorization":{}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurity","metadata":{"name":"content-security-example"},"spec":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1"},"spec":{"value":{"name":"ext-authz","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}}}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"lifetime":"16h"},"prefix":"session_prefix_example"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SidecarGateway","metadata":{"name":"sidecar-gateway-example"},"spec":{"podSelector":{"matchLabels":{"app":"example-app"}},"sessionHandlingRef":{"name":"session-handling-example"},"applications":[{"containerPort":8443,"routes":[{"pathPrefix":"/metrics","unsecured":{}},{"pathPrefix":"/","secured":{"accessControlRef":{"name":"access-control-example"},"contentSecurityRef":{"name":"content-security-example"}}}],"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}},"restrictions":{"http":{"headersLength":"80Ki"}},"protocol":{"auto":{}},"tls":{"enable":true,"secretRef":{"name":"example-tls-secret"},"protocol":{"minimum":"TLSv1_3"},"clientCertificate":{"ignored":{}},"xfcc":"AlwaysForwardOnly"}},"upstream":{"tls":{"enable":true,"protocol":{"minimum":"TLSv1_2"}}},"envoyHTTPFilterRefs":{"prepend":[{"name":"envoy-http-filter-example-1"}]},"telemetryRef":{"name":"telemetry-example"}}],"envoyClusterRefs":[{"name":"envoy-cluster-example-1"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-06-13T17:33:31Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: AccessControl
            name: accesscontrols.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurity
            name: contentsecurities.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: SidecarGateway
            name: sidecargateways.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056)
          * Gateway API
          * (Optional) Airlock Microgateway CNI (Required for [data plane mode sidecar](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000137)).

          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ### (Optional) Install Airlock Microgateway CNI
          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          Refer to [our documentation](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000140) for installation instructions

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD_MONITOR\_CREATE | "false" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.5/
        - name: Airlock Microgateway Labs
          url: https://play.instruqt.com/airlock/invite/hyi9fy4b4jzc?icp_referrer=openshift
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:806366a0b3f78844fe14775fc290fbfbde8d2a701d34a6478b019a0e7140b699
        version: 4.5.4
      entries:
      - name: airlock-microgateway.v4.5.4
        version: 4.5.4
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.5"
    - currentCSV: airlock-microgateway.v4.6.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControl","metadata":{"name":"access-control-example"},"spec":{"policies":[{"requestConditions":{"path":{"matcher":{"exact":"/secure"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}},{"oidc":{"claim":{"name":"roles","value":{"matcher":{"contains":"\"admin\""}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"oidc-example"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"token-exchange-example"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"identity-propagation-example"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"jwt-example"}}}}},{"authorization":{}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurity","metadata":{"name":"content-security-example"},"spec":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1"},"spec":{"value":{"name":"ext-authz","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}}}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"lifetime":"16h"},"prefix":"session_prefix_example"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SidecarGateway","metadata":{"name":"sidecar-gateway-example"},"spec":{"podSelector":{"matchLabels":{"app":"example-app"}},"sessionHandlingRef":{"name":"session-handling-example"},"applications":[{"containerPort":8443,"routes":[{"pathPrefix":"/metrics","unsecured":{}},{"pathPrefix":"/","secured":{"accessControlRef":{"name":"access-control-example"},"contentSecurityRef":{"name":"content-security-example"}}}],"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}},"restrictions":{"http":{"headersLength":"80Ki"}},"protocol":{"auto":{}},"tls":{"enable":true,"secretRef":{"name":"example-tls-secret"},"protocol":{"minimum":"TLSv1_3"},"clientCertificate":{"ignored":{}},"xfcc":"AlwaysForwardOnly"}},"upstream":{"tls":{"enable":true,"protocol":{"minimum":"TLSv1_2"}}},"envoyHTTPFilterRefs":{"prepend":[{"name":"envoy-http-filter-example-1"}]},"telemetryRef":{"name":"telemetry-example"}}],"envoyClusterRefs":[{"name":"envoy-cluster-example-1"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-23T06:27:21Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: AccessControl
            name: accesscontrols.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurity
            name: contentsecurities.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: SidecarGateway
            name: sidecargateways.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056)
          * Gateway API
          * (Optional) Airlock Microgateway CNI (Required for [data plane mode sidecar](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000137)).

          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.3.0/standard-install.yaml
          ```

          ### (Optional) Install Airlock Microgateway CNI
          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          Refer to [our documentation](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000140) for installation instructions

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD_MONITOR\_CREATE | "false" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.6/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:8fa300ced5b69acbad824c037d2b344fd1926bcfb39e264e9f0f055ceb9bd68d
        version: 4.6.4
      entries:
      - name: airlock-microgateway.v4.6.4
        version: 4.6.4
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.6"
    - currentCSV: airlock-microgateway.v4.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1","namespace":"default"},"spec":{"value":{"name":"envoycluster/default/envoy-cluster-example-1","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyExtensionPolicy","metadata":{"name":"envoy-extension-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"envoyHTTPFilterRefs":{"prepend":[{"name":"geoip-blocker"}],"append":[{"name":"lua-transform-request"}]},"envoyClusterRefs":[{"name":"geoip-server"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"volumes":[{"name":"vol-1","emptyDir":{}}],"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}},"volumeMounts":[{"name":"vol-1","mountPath":"/tmp/data"}]}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"telemetryRef":{"name":"telemetry-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"mode":"Enforce","persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"idle":"30s","lifetime":"16h"},"prefix":"session_prefix_example","cookie":{"name":"SESS","attributes":{"domain":"example.org","path":"/","sameSite":"Strict","secure":true}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"},"session":{"id":"%SESSION_ID%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-03T13:02:23Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyExtensionPolicy
            name: envoyextensionpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ***Note:*** Sidecar-based data plane mode is *not* supported by this OLM Bundle.
          If you want to use the Sidecar-based data plane mode, use Helm instead. Please note that the Airlock Microgateway CNI and the Helm charts are not certified for RedHat OpenShift.
          Refer to https://github.com/airlock/microgateway for installation instruction with Helm.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056)
          * [Gateway API](https://gateway-api.sigs.k8s.io/)

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ***Note:*** *This step is only required for OpenShift versions below 4.19, as Gateway API is included by default in OpenShift 4.19+.*

          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD\_MONITOR\_CREATE | "true" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gateway API
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.7/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:d4ce35184922953c209a9ebb234ca3370dbfc5054de411b5503c26ac2730b7ae
        version: 4.7.1
      entries:
      - name: airlock-microgateway.v4.7.1
        version: 4.7.1
      - name: airlock-microgateway.v4.7.0
        version: 4.7.0
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.7"
    - currentCSV: airlock-microgateway.v4.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1","namespace":"default"},"spec":{"value":{"name":"envoycluster/default/envoy-cluster-example-1","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyExtensionPolicy","metadata":{"name":"envoy-extension-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"envoyHTTPFilterRefs":{"prepend":[{"name":"geoip-blocker"}],"append":[{"name":"lua-transform-request"}]},"envoyClusterRefs":[{"name":"geoip-server"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"volumes":[{"name":"vol-1","emptyDir":{}}],"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}},"volumeMounts":[{"name":"vol-1","mountPath":"/tmp/data"}]}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"telemetryRef":{"name":"telemetry-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"mode":"Enforce","persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"idle":"30s","lifetime":"16h"},"prefix":"session_prefix_example","cookie":{"name":"SESS","attributes":{"domain":"example.org","path":"/","sameSite":"Strict","secure":true}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"},"session":{"id":"%SESSION_ID%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-03T13:02:23Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyExtensionPolicy
            name: envoyextensionpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ***Note:*** Sidecar-based data plane mode is *not* supported by this OLM Bundle.
          If you want to use the Sidecar-based data plane mode, use Helm instead. Please note that the Airlock Microgateway CNI and the Helm charts are not certified for RedHat OpenShift.
          Refer to https://github.com/airlock/microgateway for installation instruction with Helm.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056)
          * [Gateway API](https://gateway-api.sigs.k8s.io/)

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ***Note:*** *This step is only required for OpenShift versions below 4.19, as Gateway API is included by default in OpenShift 4.19+.*

          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD\_MONITOR\_CREATE | "true" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gateway API
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.7/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:d4ce35184922953c209a9ebb234ca3370dbfc5054de411b5503c26ac2730b7ae
        version: 4.7.1
      entries:
      - name: airlock-microgateway.v4.7.1
        version: 4.7.1
      - name: airlock-microgateway.v4.7.0
        version: 4.7.0
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: stable
    defaultChannel: "4.7"
    packageName: airlock-microgateway
    provider:
      name: Airlock
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: must-gather-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: must-gather-operator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "MustGather",
                "metadata": {
                  "name": "example-mustgather"
                },
                "spec": {
                  "caseID": "02527285",
                  "caseManagementAccountSecretRef": {
                    "name": "case-management-creds"
                  },
                  "serviceAccountRef": {
                    "name": "must-gather-admin"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security, Monitoring
          certified: "false"
          containerImage: quay.io/redhat-cop/must-gather-operator@sha256:aebef122cdb8409f5a5d70d3cdc585e02272704d57e103553a84bc6fc238e646
          createdAt: "2021-12-02T17:13:53Z"
          description: This operator provides a facility to easily upload must-gather
            reports to a Red Hat case.
          operatorframework.io/suggested-namespace: global-load-balancer-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/must-gather-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: MustGather
            name: mustgathers.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The must gather operator helps collecting must gather information on a cluster and uploading it to a case.
          To use the operator a cluster administrator can create the following must gather CR:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: MustGather
          metadata:
            name: example-mustgather
          spec:
            caseID: '02527285'
            caseManagementAccountSecretRef:
              name: case-management-creds
          ```

          this request will collect the standard must gather info and upload it to case `#02527285` using the credentials found in the `caseManagementCreds` secret.

          A more complex example:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: MustGather
          metadata:
            name: full-mustgather
          spec:
            caseID: '02527285'
            caseManagementAccountSecretRef:
              name: case-management-creds
            serviceAccountRef:
              name: must-gather-admin
            mustGatherImages:
            - quay.io/kubevirt/must-gather:latest
            - quay.io/ocs-dev/ocs-must-gather
          ```

          in this example we are using a specific service account (which must have cluster admin permissions as per must-gather requirements) and we are specifying a couple of additional must gather images to be run for the `kubevirt` and `ocs` subsystem. If not specified serviceAccountRef.Name will default to `default`. Also the standard must gather image: `quay.io/openshift/origin-must-gather:latest` is always added by default.
        displayName: Must Gather Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - case_management
        - support
        - must_gather
        links:
        - name: Must Gather Operator
          url: https://github.com/redhat-cop/must-gather-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/must-gather-operator
        maintainers:
        - email: rspazzol@redhat.com'
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/must-gather-operator@sha256:aebef122cdb8409f5a5d70d3cdc585e02272704d57e103553a84bc6fc238e646
        version: 1.1.2
      entries:
      - name: must-gather-operator.v1.1.2
        version: 1.1.2
      - name: must-gather-operator.v1.1.1
        version: 1.1.1
      - name: must-gather-operator.v1.1.0
        version: 1.1.0
      - name: must-gather-operator.v1.0.6
        version: 1.0.6
      - name: must-gather-operator.v1.0.5
        version: 1.0.5
      - name: must-gather-operator.v1.0.4
        version: 1.0.4
      - name: must-gather-operator.v1.0.3
        version: 1.0.3
      - name: must-gather-operator.v1.0.2
        version: 1.0.2
      - name: must-gather-operator.v1.0.1
        version: 1.0.1
      - name: must-gather-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: must-gather-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: machine-deletion-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: machine-deletion-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/machine-deletion-remediation-operator:v0.1.0
          createdAt: "2023-05-11 10:46:12"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/machine-deletion-remediation-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: machine-deletion-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: machine-deletion-remediation.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/machine-deletion-remediation-operator:v0.3.0
          createdAt: "2024-03-05 18:45:16"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/medik8s/machine-deletion-remediation-operator:v0.3.0
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        version: 0.3.0
      entries:
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      - name: machine-deletion-remediation.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: machine-deletion-remediation
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-nimbusmosaic-dusim
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-nimbusmosaic-dusim.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Nimbusmosaicdusim",
                "metadata": {
                  "name": "nimbusmosaicdusim-sample"
                },
                "spec": {
                  "development": false,
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "microservices": {
                    "jvm": "-Xmx300M -Xms64M -Xss256k"
                  },
                  "service": {
                    "type": "ClusterIP"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:41:56Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nimbusmosaicdusim
            name: nimbusmosaicdusims.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Nimbus Mosaic DUSim Operator
        displayName: Keysight Nimbus Mosaic DUSim Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Nimbusmosaic Dusim
          url: https://keysight-nimbusmosaic-dusim.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-nimbusmosaic-dusim@sha256:655dd6595a9c8e82947bed24991ac52f8fa55be99ca410e7d9f0026c51ef4520
        version: 0.1.21
      entries:
      - name: keysight-nimbusmosaic-dusim.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-nimbusmosaic-dusim
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-namespace-scope-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-namespace-scope-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1",
                "kind": "NamespaceScope",
                "metadata": {
                  "name": "namespacescope"
                },
                "spec": {
                  "namespaceMembers": [
                    "ibm-common-services",
                    "default"
                  ],
                  "restartLabels": {
                    "intent": "projected"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
          createdAt: "2025-02-19T00:54:24Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.2.14
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/IBM/ibm-namespace-scope-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NamespaceScope is the Schema for the namespacescopes API
            displayName: Namespace Scope
            kind: NamespaceScope
            name: namespacescopes.operator.ibm.com
            version: v1
        description: This operator automates the extension of operator watch and service
          account permission scope to other namespaces in an openshift cluster.
        displayName: IBM NamespaceScope Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Namespace Scope Operator
          url: https://github.com/IBM/ibm-namespace-scope-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:0774b1d916ebc354050612474800dda3a501c66a7a5aab6d18698c8c4e81ac4e
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-bundle@sha256:b6cb9e5976b749dd7818ac19d1cfd5e24673f3c052a41d6e2fe1d2d7094694b4
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:7a2cd72736bf9d30341763fd1652ff4b9ce3c785da5123645ee7664ee38c7283
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-catalog@sha256:c016904e8c5cfcf8270241ad5e28b52f528859984ed894cdd5e1283fa90cd6e8
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator-bundle@sha256:c743dd7de2edbd08c6c73e20cf15f3b147491ad77a81677339e88caffb93e84b
        version: 4.2.14
      entries:
      - name: ibm-namespace-scope-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-namespace-scope-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hitachi
      provider-url: ""
    name: hnas-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hnas-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.csi.hitachi.com/v1",
                "kind": "HNAS",
                "metadata": {
                  "name": "hnas",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-operator@sha256:9d3f9f10fc7172ee9cebbd9ee318906756d7e96909f2e66e0534e98bb53e0a7b
          createdAt: "2025-07-30T15:12:23Z"
          description: An operator for managing Hitachi NAS CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HNAS is the Schema for the hnas API
            displayName: HNAS
            kind: HNAS
            name: hnas.charts.csi.hitachi.com
            version: v1
        description: |-
          The Hitachi NAS CSI driver operator deploys the Hitachi NAS CSI driver on Kubernetes and OpenShift environments.  The driver can be used to create and manage persistent storage hosted on Hitachi NAS products, including those in the Hitachi VSP One File 3x range and the HNAS range of products.

          For the complete documentation, go to our [Product Documentation site](https://docs.hitachivantara.com/home) and refer to the Hitachi NAS CSI Driver User Guide - MK-92HNAS091.  The guide covers prerequisites for the NAS system and covers the driver functionality and all configuration options available for the driver.
        displayName: Hitachi NAS CSI Driver Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hitachi
        - Storage
        - Container Storage Interface
        - CSI
        - NAS
        links:
        - name: Product Documentation
          url: https://docs.hitachivantara.com/home
        maturity: stable
        minKubeVersion: 1.28.0
        provider:
          name: Hitachi
        relatedImages:
        - registry.k8s.io/sig-storage/csi-resizer@sha256:ab774734705a906561e15b67f6a96538f3ad520335d636f793aaafe87a3b5200
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:dd788d79cf4c1b8edee6d9b80b8a1ebfc51a38a365c5be656986b129be9ac784
        - registry.k8s.io/sig-storage/livenessprobe@sha256:2c5f9dc4ea5ac5509d93c664ae7982d4ecdec40ca7b0638c24e5b16243b8360f
        - registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-operator@sha256:9d3f9f10fc7172ee9cebbd9ee318906756d7e96909f2e66e0534e98bb53e0a7b
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:d7138bcc3aa5f267403d45ad4292c95397e421ea17a0035888850f424c7de25d
        - registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-driver@sha256:6b6f6816c9f4e608e93e7fbc0347255415426a0ff0c9a53b7750e5ba54d91262
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d5e46da8aff7d73d6f00c761dae94472bcda6e78f4f17b3802dc89d44de0111b
        - registry.k8s.io/sig-storage/csi-attacher@sha256:a399393ff5bd156277c56bae0c08389b1a1b95b7fd6ea44a316ce55e0dd559d7
        version: 1.0.0
      entries:
      - name: hnas-csi-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: hnas-csi-operator
    provider:
      name: Hitachi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: marin3r
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marin3r.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyConfig",
                "metadata": {
                  "name": "envoyconfig-example",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "nodeID": "example",
                  "resources": [
                    {
                      "type": "cluster",
                      "value": {
                        "connect_timeout": "0.01s",
                        "dns_lookup_family": "V4_ONLY",
                        "lb_policy": "ROUND_ROBIN",
                        "load_assignment": {
                          "cluster_name": "example",
                          "endpoints": [
                            {
                              "lb_endpoints": [
                                {
                                  "endpoint": {
                                    "address": {
                                      "socket_address": {
                                        "address": "example",
                                        "port_value": 8080
                                      }
                                    }
                                  }
                                }
                              ]
                            }
                          ]
                        },
                        "name": "example",
                        "type": "STRICT_DNS"
                      }
                    },
                    {
                      "type": "route",
                      "value": {
                        "name": "local",
                        "virtual_hosts": [
                          {
                            "domains": [
                              "*"
                            ],
                            "name": "all",
                            "routes": [
                              {
                                "match": {
                                  "prefix": "/"
                                },
                                "route": {
                                  "cluster": "example"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    {
                      "type": "listener",
                      "value": {
                        "address": {
                          "socket_address": {
                            "address": "0.0.0.0",
                            "port_value": 8443
                          }
                        },
                        "filter_chains": [
                          {
                            "filters": [
                              {
                                "name": "envoy.filters.network.http_connection_manager",
                                "typed_config": {
                                  "@type": "type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager",
                                  "access_log": [
                                    {
                                      "name": "envoy.access_loggers.file",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.access_loggers.file.v3.FileAccessLog",
                                        "path": "/dev/stdout"
                                      }
                                    }
                                  ],
                                  "http_filters": [
                                    {
                                      "name": "envoy.filters.http.router",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.filters.http.router.v3.Router"
                                      }
                                    }
                                  ],
                                  "rds": {
                                    "config_source": {
                                      "ads": {},
                                      "resource_api_version": "V3"
                                    },
                                    "route_config_name": "local"
                                  },
                                  "stat_prefix": "https"
                                }
                              }
                            ],
                            "transport_socket": {
                              "name": "envoy.transport_sockets.tls",
                              "typed_config": {
                                "@type": "type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext",
                                "common_tls_context": {
                                  "tls_certificate_sds_secret_configs": [
                                    {
                                      "name": "example.default.svc",
                                      "sds_config": {
                                        "ads": {},
                                        "resource_api_version": "V3"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          }
                        ],
                        "name": "https"
                      }
                    },
                    {
                      "blueprint": "tlsCertificate",
                      "generateFromTlsSecret": "example.default.svc",
                      "type": "secret"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "DiscoveryService",
                "metadata": {
                  "name": "discoveryservice-example",
                  "namespace": "my-namespace"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyDeployment",
                "metadata": {
                  "name": "envoydeployment-sample",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "discoveryServiceRef": "discoveryservice-example",
                  "envoyConfigRef": "envoyconfig-example",
                  "ports": [
                    {
                      "name": "https",
                      "port": 8443
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: quay.io/3scale/marin3r
          createdAt: "2024-08-05T16:08:25Z"
          description: Lighweight, CRD based Envoy control plane for Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["envoyconfigrevisions.marin3r.3scale.net","discoveryservicecertificates.operator.marin3r.3scale.net"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/3scale-ops/marin3r
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DiscoveryServiceCertificate is an internal resource used
              to create certificates. This resource is used by the DiscoveryService
              controller to create the required certificates for the different components.
              Direct use of DiscoveryServiceCertificate objects is discouraged.
            displayName: DiscoveryServiceCertificate
            kind: DiscoveryServiceCertificate
            name: discoveryservicecertificates.operator.marin3r.3scale.net
            version: v1alpha1
          - description: DiscoveryService represents an envoy discovery service server.
              Only one instance per namespace is currently supported.
            displayName: DiscoveryService
            kind: DiscoveryService
            name: discoveryservices.operator.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfigRevision is an internal resource that stores a
              specific version of an EnvoyConfig resource. EnvoyConfigRevisions are
              automatically created and deleted by the EnvoyConfig controller and
              are not intended to be directly used. Use EnvoyConfig objects instead.
            displayName: EnvoyConfigRevision
            kind: EnvoyConfigRevision
            name: envoyconfigrevisions.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfig holds the configuration for a given envoy nodeID.
              The spec of an EnvoyConfig object holds the Envoy resources that conform
              the desired configuration for the given nodeID and that the discovery
              service will send to any envoy client that identifies itself with that
              nodeID.
            displayName: EnvoyConfig
            kind: EnvoyConfig
            name: envoyconfigs.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyDeployment is a resource to deploy and manage a Kubernetes
              Deployment of Envoy Pods.
            displayName: EnvoyDeployment
            kind: EnvoyDeployment
            name: envoydeployments.operator.marin3r.3scale.net
            version: v1alpha1
        description: |
          MARIN3R implements a control plane to deploy, configure and operate a fleet of envoy instances within a Kubernetes cluster.
          It operates a group of gateways and Pod sidecar containers to provide general purpose functionality required to operate production systems.

          Features:

          * Deploy and manage an Envoy xDS server using the DiscoveryService custom resource.
          * Inject Envoy sidecar containers based on Pod annotations.
          * Deploy Envoy as a Kubernetes Deployment using the EnvoyDeployment custom resource.
          * Dynamic Envoy configuration and re-configuration using the EnvoyConfig custom resource.
          * Use Kubernetes Secrets as certificate sources.
          * Syntactic validation of Envoy configurations.
          * Lifecycle management of Envoy containers (graceful shutdown and connection draining).
          * Self-healing.

          Check the [docs](https://github.com/3scale-ops/marin3r#readme) for more information.

          ## License
          MARIN3R is licensed under the [Apache 2.0 license](https://github.com/3scale/prometheus-exporter-operator/blob/master/LICENSE)
        displayName: MARIN3R
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - envoy
        - discovery
        - controlplane
        - proxy
        - sidecar
        - xds
        - network
        - gateway
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/marin3r
        maintainers:
        - email: 3scale-operations+marin3r@redhat.com
          name: 3scale Operations Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/marin3r:v0.13.1
        version: 0.13.1
      entries:
      - name: marin3r.v0.13.1
        version: 0.13.1
      - name: marin3r.v0.12.3
        version: 0.12.3
      - name: marin3r.v0.11.1
        version: 0.11.1
      - name: marin3r.v0.10.0
        version: 0.10.0
      name: alpha
    - currentCSV: marin3r.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyConfig",
                "metadata": {
                  "name": "envoyconfig-example",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "nodeID": "example",
                  "resources": [
                    {
                      "type": "cluster",
                      "value": {
                        "connect_timeout": "0.01s",
                        "dns_lookup_family": "V4_ONLY",
                        "lb_policy": "ROUND_ROBIN",
                        "load_assignment": {
                          "cluster_name": "example",
                          "endpoints": [
                            {
                              "lb_endpoints": [
                                {
                                  "endpoint": {
                                    "address": {
                                      "socket_address": {
                                        "address": "example",
                                        "port_value": 8080
                                      }
                                    }
                                  }
                                }
                              ]
                            }
                          ]
                        },
                        "name": "example",
                        "type": "STRICT_DNS"
                      }
                    },
                    {
                      "type": "route",
                      "value": {
                        "name": "local",
                        "virtual_hosts": [
                          {
                            "domains": [
                              "*"
                            ],
                            "name": "all",
                            "routes": [
                              {
                                "match": {
                                  "prefix": "/"
                                },
                                "route": {
                                  "cluster": "example"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    {
                      "type": "listener",
                      "value": {
                        "address": {
                          "socket_address": {
                            "address": "0.0.0.0",
                            "port_value": 8443
                          }
                        },
                        "filter_chains": [
                          {
                            "filters": [
                              {
                                "name": "envoy.filters.network.http_connection_manager",
                                "typed_config": {
                                  "@type": "type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager",
                                  "access_log": [
                                    {
                                      "name": "envoy.access_loggers.file",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.access_loggers.file.v3.FileAccessLog",
                                        "path": "/dev/stdout"
                                      }
                                    }
                                  ],
                                  "http_filters": [
                                    {
                                      "name": "envoy.filters.http.router",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.filters.http.router.v3.Router"
                                      }
                                    }
                                  ],
                                  "rds": {
                                    "config_source": {
                                      "ads": {},
                                      "resource_api_version": "V3"
                                    },
                                    "route_config_name": "local"
                                  },
                                  "stat_prefix": "https"
                                }
                              }
                            ],
                            "transport_socket": {
                              "name": "envoy.transport_sockets.tls",
                              "typed_config": {
                                "@type": "type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext",
                                "common_tls_context": {
                                  "tls_certificate_sds_secret_configs": [
                                    {
                                      "name": "example.default.svc",
                                      "sds_config": {
                                        "ads": {},
                                        "resource_api_version": "V3"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          }
                        ],
                        "name": "https"
                      }
                    },
                    {
                      "blueprint": "tlsCertificate",
                      "generateFromTlsSecret": "example.default.svc",
                      "type": "secret"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "DiscoveryService",
                "metadata": {
                  "name": "discoveryservice-example",
                  "namespace": "my-namespace"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyDeployment",
                "metadata": {
                  "name": "envoydeployment-sample",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "discoveryServiceRef": "discoveryservice-example",
                  "envoyConfigRef": "envoyconfig-example",
                  "ports": [
                    {
                      "name": "https",
                      "port": 8443
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: quay.io/3scale/marin3r
          createdAt: "2024-08-05T16:08:25Z"
          description: Lighweight, CRD based Envoy control plane for Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["envoyconfigrevisions.marin3r.3scale.net","discoveryservicecertificates.operator.marin3r.3scale.net"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/3scale-ops/marin3r
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DiscoveryServiceCertificate is an internal resource used
              to create certificates. This resource is used by the DiscoveryService
              controller to create the required certificates for the different components.
              Direct use of DiscoveryServiceCertificate objects is discouraged.
            displayName: DiscoveryServiceCertificate
            kind: DiscoveryServiceCertificate
            name: discoveryservicecertificates.operator.marin3r.3scale.net
            version: v1alpha1
          - description: DiscoveryService represents an envoy discovery service server.
              Only one instance per namespace is currently supported.
            displayName: DiscoveryService
            kind: DiscoveryService
            name: discoveryservices.operator.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfigRevision is an internal resource that stores a
              specific version of an EnvoyConfig resource. EnvoyConfigRevisions are
              automatically created and deleted by the EnvoyConfig controller and
              are not intended to be directly used. Use EnvoyConfig objects instead.
            displayName: EnvoyConfigRevision
            kind: EnvoyConfigRevision
            name: envoyconfigrevisions.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfig holds the configuration for a given envoy nodeID.
              The spec of an EnvoyConfig object holds the Envoy resources that conform
              the desired configuration for the given nodeID and that the discovery
              service will send to any envoy client that identifies itself with that
              nodeID.
            displayName: EnvoyConfig
            kind: EnvoyConfig
            name: envoyconfigs.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyDeployment is a resource to deploy and manage a Kubernetes
              Deployment of Envoy Pods.
            displayName: EnvoyDeployment
            kind: EnvoyDeployment
            name: envoydeployments.operator.marin3r.3scale.net
            version: v1alpha1
        description: |
          MARIN3R implements a control plane to deploy, configure and operate a fleet of envoy instances within a Kubernetes cluster.
          It operates a group of gateways and Pod sidecar containers to provide general purpose functionality required to operate production systems.

          Features:

          * Deploy and manage an Envoy xDS server using the DiscoveryService custom resource.
          * Inject Envoy sidecar containers based on Pod annotations.
          * Deploy Envoy as a Kubernetes Deployment using the EnvoyDeployment custom resource.
          * Dynamic Envoy configuration and re-configuration using the EnvoyConfig custom resource.
          * Use Kubernetes Secrets as certificate sources.
          * Syntactic validation of Envoy configurations.
          * Lifecycle management of Envoy containers (graceful shutdown and connection draining).
          * Self-healing.

          Check the [docs](https://github.com/3scale-ops/marin3r#readme) for more information.

          ## License
          MARIN3R is licensed under the [Apache 2.0 license](https://github.com/3scale/prometheus-exporter-operator/blob/master/LICENSE)
        displayName: MARIN3R
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - envoy
        - discovery
        - controlplane
        - proxy
        - sidecar
        - xds
        - network
        - gateway
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/marin3r
        maintainers:
        - email: 3scale-operations+marin3r@redhat.com
          name: 3scale Operations Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/marin3r:v0.13.1
        version: 0.13.1
      entries:
      - name: marin3r.v0.13.1
        version: 0.13.1
      - name: marin3r.v0.12.3
        version: 0.12.3
      - name: marin3r.v0.11.1
        version: 0.11.1
      - name: marin3r.v0.10.0
        version: 0.10.0
      name: stable
    defaultChannel: stable
    packageName: marin3r
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: GitLab, Inc.
      provider-url: https://gitlab.com
    name: gitlab-runner-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-runner-operator.v1.41.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta2",
                "kind": "Runner",
                "metadata": {
                  "name": "gitlab-runner-example",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "gitlabUrl": "https://gitlab.example.com",
                  "token": "runner-token-secret",
                  "tags": "kubernetes,example",
                  "concurrent": 10,
                  "imagePullPolicy": "IfNotPresent"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery, Developer Tools
          containerImage: registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
          createdAt: "2025-09-25T15:08:40Z"
          description: GitLab Runner operator manages lifecycle of GitLab Runner instances
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.11.0 <1.41.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: none
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
          support: GitLab, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Runner is the open source project used to run your jobs and
              send the results back to GitLab
            displayName: GitLab Runner
            kind: Runner
            name: runners.apps.gitlab.com
            version: v1beta2
        description: |-
          GitLab Runner is the lightweight, highly-scalable agent that runs your build jobs and sends the results back to a GitLab instance. GitLab Runner works in conjunction with GitLab CI/CD, the open-source continuous integration service included with GitLab.

              The GitLab Runner operator manages the lifecycle of GitLab Runner in Kubernetes or Openshift clusters. The operator aims to automate the tasks needed to run your CI/CD jobs in your container orchestration platform.

              ## Prerequisites

                For Kubernetes cluster, install cert-manager:

                ```shell
                kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.7.1/cert-manager.yaml
                ```

              ## GitLab Runner version

              This version of **GitLab Runner Operator** ships with **GitLab Runner v18.4.0**.

              To use a different version of **GitLab Runner** change the [`runnerImage` and `helperImage` properties](https://docs.gitlab.com/runner/configuration/configuring_runner_operator.html#operator-properties).

              ## Usage

               To link a GitLab Runner instance to a self-hosted GitLab instance or the hosted [GitLab](https://gitlab.com), you first need to:

               1. Create a secret containing the `runner-token` from your GitLab project.

                ```
                cat > gitlab-runner-secret.yml << EOF
                apiVersion: v1
                kind: Secret
                metadata:
                  name: gitlab-runner-secret
                type: Opaque
                # Only one of the following fields must be set. The Operator will fail to register Runner if both are provided.
                # # NOTE: runner-registration-token is deprecated and will be removed in GitLab 18.0. You should use runner-token instead.
                stringData:
                  runner-token: REPLACE_ME # your project runner token
                  # runner-registration-token: "" # your project runner secret
                EOF
                ```

                ```
                oc apply -f gitlab-runner-secret.yml
                ```

               2. Create the Custom Resource Definition (CRD) file and include the following information. The tags value must be openshift for the job to run.

                 ```
                 cat > gitlab-runner.yml << EOF
                 apiVersion: apps.gitlab.com/v1beta2
                 kind: Runner
                 metadata:
                   name: gitlab-runner
                 spec:
                   gitlabUrl: https://gitlab.example.com
                   buildImage: alpine
                   token: gitlab-runner-secret
                   tags: openshift
                 EOF
                 ```

                ```
                oc apply -f gitlab-runner.yml
                ```

              ## Full documentation

              Visit [Install GitLab Runner Operator](https://docs.gitlab.com/runner/install/operator.html)
        displayName: GitLab Runner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GitLab
        - CI/CD
        - DevOps
        - SAST
        - DAST
        links:
        - name: Gitlab Runner Operator
          url: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
        - name: GitLab Docs
          url: https://docs.gitlab.com
        - name: GitLab Runner
          url: https://docs.gitlab.com/runner/
        - name: GitLab CI/CD
          url: https://docs.gitlab.com/ee/ci/quick_start/
        maintainers:
        - email: support@gitlab.com
          name: Georgi N. Georgiev
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: GitLab, Inc.
          url: https://gitlab.com
        relatedImages:
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-ocp@sha256:0f2a5a3b6779d719eecbf079b99cf06b652603d74b563f57d753e32966d253dd
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-helper-ocp@sha256:29727fae83a226bcf57607240d4046a0b8c208b5ff42ee1f8ca0f115bb188809
        version: 1.41.1
      entries:
      - name: gitlab-runner-operator.v1.41.1
        version: 1.41.1
      - name: gitlab-runner-operator.v1.41.0
        version: 1.41.0
      - name: gitlab-runner-operator.v1.40.1
        version: 1.40.1
      - name: gitlab-runner-operator.v1.40.0
        version: 1.40.0
      - name: gitlab-runner-operator.v1.39.0
        version: 1.39.0
      name: stable
    defaultChannel: stable
    packageName: gitlab-runner-operator
    provider:
      name: GitLab, Inc.
      url: https://gitlab.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Anjuna Security
      provider-url: https://www.anjuna.io
    name: anjuna-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anjuna-operator.v2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anjuna.io/v1",
                "kind": "AnjunaConfig",
                "metadata": {
                  "name": "anjuna-config-remote"
                },
                "spec": {
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "runtimeImage": "registry.connect.redhat.com/anjuna/anjuna-k8s-sev-runtime@sha256:1531b1ba955b2d3ea5d9238d2e20617a534bb1bab73f5fadadf8c9d3aee410b4"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-02-19T01:11:18Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          operatorframework.io/suggested-namespace: anjuna-operator
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnjunaConfig operand instructs the Anjuna Seaglass Operator
              to know (1) what needs to be installed into the cluster and (2) what
              nodes will be affected.
            displayName: Anjuna Config
            kind: AnjunaConfig
            name: anjunaconfigs.anjuna.io
            version: v1
        description: "The Anjuna Seaglass Operator for OpenShift allows running any
          Pod in\nOpenShift in a Confidential Pod leveraging the full power of the
          Confidential\nComputing technology.\n\n# Anjuna Seaglass\nAnjuna Seaglass™
          is the Universal Confidential Computing Platform to run applications in
          any cloud, and on-premises, with complete data security and privacy. \nIt
          isolates workloads in a hardware-backed environment that creates a zero-trust
          model for code and data.\n\nThe Anjuna Seaglass platform allows users to
          leverage the power of Confidential Computing:\n\nFor any Confidential Computing
          technology\nIn any Cloud Service Provider, as well as on-premises\nWith
          no changes to the application\n\nRunning a Confidential Pod with Anjuna
          Seaglass protects the Pod's memory with hardware-enforced encryption and
          integrity protection, and enables you to authenticate and trust the workload
          using Remote Attestation.\n\n# Data-in-use protection\nWhen protecting sensitive
          data managed by a Pod in OpenShift, one can protect the data at rest by
          encrypting the data written to the persistent storage, and protecting the
          data in transit when sending data over the network.\n\nHowever, when the
          data is used in the Pod's memory, it cannot be both encrypted and used at
          the same time.\nThis means that if a malicious insider (either human or
          software) had access to the OpenShift Node, they could dump the Pod's memory
          and access all of its sensitive data. This includes accessing the secrets
          used to encrypt the data at rest and in transit.\n\nRunning the Pod in a
          Confidential Pod with the Anjuna Seaglass platform solves this problem.
          Anjuna Seaglass runs the Pod inside a Trusted Execution Environment, where
          the Pod memory is encrypted by the hardware so that no privileged human
          or software on the Pod can access this memory in the clear.\n\n# Enabling
          zero trust for the Pod\nWhen running a Pod in OpenShift, the Pod may need
          to create trust with other Pods or with software running outside of the
          OpenShift cluster.\nThe standard way to create this trust is to identify
          with a token or credentials that could be easily read and stolen from the
          Pod disk or by providing a machine identity, a process name, or an IP that
          could be spoofed or reused.\n\nConfidential Computing technology provides
          a mechanism called Remote Attestation, which provides an identity to the
          software via the Hardware—the Software Measurement.\n\nWhen running a Pod
          as a Confidential Pod with the Anjuna Seaglass platform, a user can leverage
          the Remote Attestation to perform trusted authentication with the Confidential
          Pod, in a way that cannot be spoofed.\nThe Anjuna Policy Manager allows
          you to protect the first secret of a Confidential Pod and remove any secret
          from the Pod image or Kubernetes control plane that could be read by malicious
          insiders.\n\n# Quick start and configuration instructions\nFollow the directions
          in Installing the [Anjuna Kubernetes Toolset to OpenShift on GCP](https://docs.anjuna.io/k8s-sev/latest/installation/openshift_gcp.html)
          \nto learn how to deploy the Operator into your OpenShift cluster.\n\nFollow
          the directions in [Quickstart guides for OpenShift on GCP](https://docs.anjuna.io/k8s-sev/latest/applications/gcp.html)
          \nto learn how to deploy your first Confidential Pod in your OpenShift cluster.\n\nFollow
          the directions in [Attestation with Anjuna Policy Manager](https://docs.anjuna.io/k8s-sev/latest/attestation.html)
          \nto learn how to leverage Remote Attestation to protect your Pod's first
          secret."
        displayName: Anjuna Seaglass™ Operator for OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Security
        links:
        - name: Anjuna Operator Documentation
          url: https://docs.anjuna.io/k8s-sev/latest/index.html
        - name: Anjuna Seaglass
          url: https://www.anjuna.io/essence-of-anjuna-seaglass
        - name: Anjuna Security
          url: https://www.anjuna.io
        maintainers:
        - email: support@anjuna.io
          name: Anjuna Security
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Anjuna Security
          url: https://www.anjuna.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - registry.connect.redhat.com/anjuna/anjuna-k8s-sev-tools@sha256:e5ba31c240aaff8ca8f6addf12577fd6045bf6619a1d2cb2987b2ff5d75f4028
        - registry.connect.redhat.com/anjuna/anjuna-k8s-sev-runtime@sha256:1531b1ba955b2d3ea5d9238d2e20617a534bb1bab73f5fadadf8c9d3aee410b4
        version: 2.0.0
      entries:
      - name: anjuna-operator.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anjuna-operator
    provider:
      name: Anjuna Security
      url: https://www.anjuna.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzograph-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzograph-operator.v3.2.500
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzograph.clusters.cambridgesemantics.com/v2",
                "kind": "AnzoGraph",
                "metadata": {
                  "name": "azg01"
                },
                "spec": {
                  "db": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_data": "anzograph-data-grpc",
                            "app_mgmt": "anzograph-mgmt-grpc"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_data": "anzograph-data-grpc",
                              "app_mgmt": "anzograph-mgmt-grpc"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzograph-db@sha256:30ed2ec337135701ab8b54e75bb48ad8928a2c48818643c23b9dba4484255a2e",
                                "name": "db",
                                "resources": {
                                  "limits": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  },
                                  "requests": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  }
                                }
                              }
                            ],
                            "securityContext": {
                              "fsGroup": 1000,
                              "runAsNonRoot": true,
                              "runAsUser": 1000,
                              "seccompProfile": {
                                "type": "RuntimeDefault"
                              }
                            },
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  },
                  "deployFrontend": false,
                  "frontend": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_mgmt": "anzograph-frontend"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_mgmt": "anzograph-frontend"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzograph-frontend@sha256:e5446b3c04889f2189e955702c4165eb243eb89f47ac797eb417b60b1459647f",
                                "name": "frontend",
                                "resources": {
                                  "limits": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "securityContext": {
                              "fsGroup": 1000,
                              "runAsNonRoot": true,
                              "runAsUser": 1000,
                              "seccompProfile": {
                                "type": "RuntimeDefault"
                              }
                            },
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/anzograph-operator@sha256:2333154a14ab7f3a03a601f00c017feeafb9d9814aff5e59c95cef615ab2abd6
          createdAt: "2025-09-22T12:03:18Z"
          description: kubernetes operator for Altair Graph Lakehouse DB
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Altair Engineering Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoGraph is the Schema for the anzographs API
            displayName: Altair Graph Lakehouse Operator
            kind: AnzoGraph
            name: anzographs.anzograph.clusters.cambridgesemantics.com
            version: v2
        description: |-
          The Altair Graph Lakehouse Operator provides the way to install and configure an Altair Graph Lakehouse
          cluster on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v3.2.5/README_openshift_marketplace.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzograph/userdoc/ )

          ### Support

          We offer Support to our customers with the Altair Graph Lakehouse db Enterprise Edition License [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Lakehouse Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        - Database
        links:
        - name: Anzograph Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v3.2.5/README_openshift_marketplace.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzograph-operator@sha256:2333154a14ab7f3a03a601f00c017feeafb9d9814aff5e59c95cef615ab2abd6
        - registry.connect.redhat.com/cambridgesemantics/anzograph-db@sha256:30ed2ec337135701ab8b54e75bb48ad8928a2c48818643c23b9dba4484255a2e
        - registry.connect.redhat.com/cambridgesemantics/anzograph-frontend@sha256:e5446b3c04889f2189e955702c4165eb243eb89f47ac797eb417b60b1459647f
        version: 3.2.500
      entries:
      - name: anzograph-operator.v3.2.500
        version: 3.2.500
      - name: anzograph-operator.v3.2.300
        version: 3.2.300
      - name: anzograph-operator.v3.2.200
        version: 3.2.200
      - name: anzograph-operator.v3.2.104
        version: 3.2.104
      name: stable
    defaultChannel: stable
    packageName: anzograph-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-memorydb-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-memorydb-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "ACL",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "ParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "Snapshot",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "SubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/memorydb-controller:1.1.1
          createdAt: "2025-09-19T19:36:49Z"
          description: Amazon MemoryDB for Redis controller is a service controller
            for managing MemoryDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ACL represents the state of an AWS memorydb ACL resource.
            displayName: ACL
            kind: ACL
            name: acls.memorydb.services.k8s.aws
            version: v1alpha1
          - description: Cluster represents the state of an AWS memorydb Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.memorydb.services.k8s.aws
            version: v1alpha1
          - description: ParameterGroup represents the state of an AWS memorydb ParameterGroup
              resource.
            displayName: ParameterGroup
            kind: ParameterGroup
            name: parametergroups.memorydb.services.k8s.aws
            version: v1alpha1
          - description: Snapshot represents the state of an AWS memorydb Snapshot
              resource.
            displayName: Snapshot
            kind: Snapshot
            name: snapshots.memorydb.services.k8s.aws
            version: v1alpha1
          - description: SubnetGroup represents the state of an AWS memorydb SubnetGroup
              resource.
            displayName: SubnetGroup
            kind: SubnetGroup
            name: subnetgroups.memorydb.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS memorydb User resource.
            displayName: User
            kind: User
            name: users.memorydb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon MemoryDB for Redis resources in AWS from within your Kubernetes cluster.

          **About Amazon MemoryDB for Redis**

          MemoryDB for Redis is a durable, in-memory database service that delivers ultra-fast performance. It is purpose-built for modern applications with microservices architectures.
          MemoryDB is compatible with Redis, a popular open source data store, enabling you to quickly build applications using the same flexible and friendly Redis data structures, APIs, and commands that they already use today. With MemoryDB, all of your data is stored in memory, which enables you to achieve microsecond read and single-digit millisecond write latency and high throughput. MemoryDB also stores data durably across multiple Availability Zones (AZs) using a Multi-AZ transactional log to enable fast failover, database recovery, and node restarts.
          Delivering both in-memory performance and Multi-AZ durability, MemoryDB can be used as a high-performance primary database for your microservices applications, eliminating the need to separately manage both a cache and durable database.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - MemoryDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - memorydb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon MemoryDB for Redis Developer Resources
          url: https://docs.aws.amazon.com/memorydb/?id=docs_gateway
        maintainers:
        - email: ack-maintainers@amazon.com
          name: memorydb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/memorydb-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-memorydb-controller.v1.1.1
        version: 1.1.1
      - name: ack-memorydb-controller.v1.1.0
        version: 1.1.0
      - name: ack-memorydb-controller.v1.0.15
        version: 1.0.15
      - name: ack-memorydb-controller.v1.0.13
        version: 1.0.13
      - name: ack-memorydb-controller.v1.0.12
        version: 1.0.12
      - name: ack-memorydb-controller.v1.0.11
        version: 1.0.11
      - name: ack-memorydb-controller.v1.0.10
        version: 1.0.10
      - name: ack-memorydb-controller.v1.0.9
        version: 1.0.9
      - name: ack-memorydb-controller.v1.0.8
        version: 1.0.8
      - name: ack-memorydb-controller.v1.0.7
        version: 1.0.7
      - name: ack-memorydb-controller.v1.0.6
        version: 1.0.6
      - name: ack-memorydb-controller.v1.0.5
        version: 1.0.5
      - name: ack-memorydb-controller.v1.0.4
        version: 1.0.4
      - name: ack-memorydb-controller.v1.0.3
        version: 1.0.3
      - name: ack-memorydb-controller.v1.0.2
        version: 1.0.2
      - name: ack-memorydb-controller.v1.0.1
        version: 1.0.1
      - name: ack-memorydb-controller.v1.0.0
        version: 1.0.0
      - name: ack-memorydb-controller.v0.0.12
        version: 0.0.12
      - name: ack-memorydb-controller.v0.0.11
        version: 0.0.11
      - name: ack-memorydb-controller.v0.0.10
        version: 0.0.10
      - name: ack-memorydb-controller.v0.0.9
        version: 0.0.9
      - name: ack-memorydb-controller.v0.0.8
        version: 0.0.8
      - name: ack-memorydb-controller.v0.0.7
        version: 0.0.7
      - name: ack-memorydb-controller.v0.0.6
        version: 0.0.6
      - name: ack-memorydb-controller.v0.0.5
        version: 0.0.5
      - name: ack-memorydb-controller.v0.0.4
        version: 0.0.4
      - name: ack-memorydb-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-memorydb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibmcloud-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibmcloud-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"ibmcloud.ibm.com/v1","kind":"Service","metadata":{"annotations":{"ibmcloud.ibm.com/self-healing":"enabled"},"name":"mytranslator"},"spec":{"plan":"lite","serviceClass":"language-translator"}},{"apiVersion":"ibmcloud.ibm.com/v1","kind":"Binding","metadata":{"name":"binding-translator"},"spec":{"parameters":[{"name":"serviceid_crn","value":"crn:v1:bluemix:public:iam-identity::a/ACCOUNT_ID::serviceid:SERVICE_ID"}],"secretName":"translator-secret","serviceName":"mytranslator"}}]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: cloudoperators/ibmcloud-operator:1.1.0
          createdAt: "2022-03-17T04:17:40Z"
          description: The IBM Cloud Operator provides a Kubernetes CRD-Based API
            to manage the lifecycle of IBM public cloud services.
          repository: https://github.com/IBM/cloud-operators
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Binding is an instance of a service binding resource on IBM
              Cloud. A Binding creates a secret with the service instance credentials.
            displayName: Binding
            kind: Binding
            name: bindings.ibmcloud.ibm.com
            version: v1
          - description: Service is an instance of a Service resource on IBM Cloud.
            displayName: Service
            kind: Service
            name: services.ibmcloud.ibm.com
            version: v1
        description: |-
          With the IBM Cloud Operator, you can provision and bind [IBM public cloud services](https://cloud.ibm.com/catalog#services) to your Kubernetes cluster in a Kubernetes-native way. The IBM Cloud Operator is based on the [Kubernetes custom resource definition (CRD) API](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/) so that your applications can create, update, and delete IBM Cloud services from within the cluster by calling Kubnernetes APIs, instead of needing to use several IBM Cloud APIs in addition to configuring your app for Kubernetes.


          ## Features

          * **Service provisioning**: Create any service with any plan that is available in the [IBM Cloud catalog](https://cloud.ibm.com/catalog#services).

          * **Bindings management**: Automatically create Kubernetes secrets in your Kubernetes cluster with the credentials to bind to any provisioned service to the cluster.


          ## Upgrading the operator


          To upgrade an OperatorHub installation, use the OpenShift web console to perform the upgrade. If upgrading from the `alpha` channel subscription to `stable`, follow the additional upgrade instructions below.

          ### Upgrading to version v0.3 or v1.0
          **IMPORTANT NOTICE:** v0.1 and v0.2 used a different naming scheme for secrets and configmaps. Before you update the IBM Cloud Operator, create secret and configmap resources with these names and copy the contents of your previous resources to the new resources. Then, the upgraded operator recognizes and continues to update the resources.

          Existing `Service` and `Binding` resources **do not** need modification for the upgrade.

          | Previous names (v0.1 and v0.2)         | **Current names (v0.3 and v1.0)**             | Description                                                                           |
          |:---------------------------------------|:--------------------------------------------|:--------------------------------------------------------------------------------------|
          | secret-ibm-cloud-operator              | **ibmcloud-operator-secret**                | Secret with the API key, scoped to the namespace.                                     |
          | config-ibm-cloud-operator              | **ibmcloud-operator-defaults**              | ConfigMap with the default values for new resources.                                  |
          | ibm-cloud-operator                     | **ibmcloud-operator-config**                | ConfigMap with the management namespace configuration.                                |
          | ${namespace}-secret-ibm-cloud-operator | **${namespace}-ibmcloud-operator-secret**   | Management namespace Secret with the API key for ${namespace}.                        |
          | ${namespace}-config-ibm-cloud-operator | **${namespace}-ibmcloud-operator-defaults** | Management namespace ConfigMap with default values for new resources in ${namespace}. |

          **Tip:** Forgot to update the secret and configmap names before upgrading? The operator will not take action on Service and Binding resources until the new secrets and configmaps have been created. Creating these after the upgrade will also work.


          ### OperatorHub stable channel
          Continuous updates and bug fixes are provided in the latest stable channel. Subscribing to the `stable` channel in OperatorHub means you automatically get the latest updates without breaking backward compatibility.

          ### Changelog

          For a list of recent changes, see the [releases page](https://github.com/IBM/cloud-operators/releases).


          ## Prerequisites

          1.  Have an [IBM Cloud account](https://cloud.ibm.com/registration).
          2.  Have a cluster that runs Kubernetes version 1.11 or later (OpenShift 3.11 or later).
          3.  Install the required command line tools.
              *   [IBM Cloud CLI](https://cloud.ibm.com/docs/cli?topic=cloud-cli-getting-started) (`ibmcloud`)
              *   [Kubernetes CLI](https://kubernetes.io/docs/tasks/tools/install-kubectl/) (`kubectl`)
          4.  Log in to your IBM Cloud account from the CLI.

              ```bash
              ibmcloud login
              ```

          5.  Target the appropriate resource group (`-g`) and default region (`-r`) for provisioning services:

              ```bash
              ibmcloud target -g default -r us-south
              ```

          6.  If you also want to provision Cloud Foundry services, target the appropriate `org` and `space` (`--cf`).

              ```bash
              ibmcloud target --cf
              ```

          7.  Set the Kubernetes context of your CLI to your cluster so that you can run `kubectl` commands. For example, if your cluster runs OpenShift, use the `oc login` command.


          ## Setting up the operator

          You can use an installation script to set up the IBM Cloud Operator.
          By default, the installation script stores an API key in a Kubernetes secret in your cluster that can be accessed by the IBM Cloud Operator.
          Next, the script sets default values that are used to provision IBM Cloud services, like the resource group and region to provision the services in. Later, you can override any default value in the `Service` custom resource. Finally, the script deploys the operator in your cluster.

          To use your own API key, set the `IBMCLOUD_API_KEY` environment variable to the key before running the installation script:

              export IBMCLOUD_API_KEY="CUSTOM_API_KEY"

          If installed with Operator Hub, the operator will run in the `openshift-operators` namespace. Otherwise, it will run in the `ibmcloud-operator-system` namespace.

          Prefer to create the secrets and defaults yourself? See the [IBM Cloud Operator installation guide](https://github.com/IBM/cloud-operators/blob/master/docs/install.md).

          ### Using a service ID

          By default, the installation script creates an IBM Cloud API key that impersonates your user credentials, to use to set up the IBM Cloud Operator. However, you might want to create a service ID in IBM Cloud Identity and Access Managment (IAM). By using a service ID, you can control access for the IBM Cloud Operator without having the permissions tied to a particular user, such as if that user leaves the company. For more information, see the [IBM Cloud docs](https://cloud.ibm.com/docs/account?topic=account-serviceids).

          1.  Create a service ID in IBM Cloud IAM. If possible, do not use spaces in the names for your IAM credentials. When you use the operator binding feature, any spaces are replaced with underscores.

              ```bash
              ibmcloud iam service-id-create serviceid-ico -d service-ID-for-ibm-cloud-operator
              ```

          2.  Assign the service ID access to the required permissions to work with the IBM Cloud services that you want the operator to manage. The required permissions vary with each IBM Cloud service. You can also scope an access policy to different regions or resource groups. For example, the following command grants the service ID the **Administrator** platform role in the default resource group in the US South (Dallas) region. For more information, see the [IBM Cloud docs](https://cloud.ibm.com/docs/account?topic=account-userroles).

              ```bash
              ibmcloud iam service-policy-create serviceid-ico --roles Administrator --resource-group-name default --region us-south
              ```

          3.  Create an API key for the service ID.

              ```bash
              ibmcloud iam service-api-key-create apikey-ico serviceid-ico -d api-key-for-ibm-cloud-operator
              ```

          4.  Set the API key of the service ID as your CLI environment variable. Now, when you run the installation script, the script uses the service ID's API key. The following command is an example for macOS.

              ```bash
              export IBMCLOUD_API_KEY=<apikey-ico-value>
              ```

          5.  Confirm that the API key environment variable is set in your CLI.

              ```bash
              echo $IBMCLOUD_API_KEY
              ```

          6.  Follow the [prerequisite steps](https://github.com/IBM/cloud-operators/blob/master/README.md#prerequisites) to log in to the IBM Cloud account that owns the service ID.


          ### Installing the operator for OpenShift clusters

          Before you begin, complete the [prerequisite steps](https://github.com/IBM/cloud-operators/blob/master/README.md#prerequisites) to log in to IBM Cloud and your cluster, and optionally set up a [service ID API key](https://github.com/IBM/cloud-operators/blob/master/README.md#using-a-service-id).

          To configure the latest release for OpenShift before installing via the OperatorHub, run the following script:

          *   **Latest release**:

              ```bash
              curl -sL https://raw.githubusercontent.com/IBM/cloud-operators/master/hack/configure-operator.sh | bash
              ```

          *   **Specific release**: Replace `-v 0.0.0` with the specific version that you want to install.

              ```bash
              curl -sL https://raw.githubusercontent.com/IBM/cloud-operators/master/hack/configure-operator.sh | bash -s -- -v 0.0.0 store-creds
              ```


          ## Using the IBM Cloud Operator

          To use the IBM Cloud Operator, create a service instance and then bind the service to your cluster. For more information, see the [sample configuration files](https://github.com/IBM/cloud-operators/blob/master/config/samples), [user guide](https://github.com/IBM/cloud-operators/blob/master/docs/user-guide.md), and [reference documentation](https://github.com/IBM/cloud-operators/blob/master/README.md#reference).

          #### Step 1: Creating a service instance

          1.  To create an instance of an IBM public cloud service, first create a `Service` custom resource file. For more options, see the [Service properties](https://github.com/IBM/cloud-operators/blob/master/README.md#service-properties) reference doc.
              *   `<SERVICE_CLASS>` is the IBM Cloud service that you want to create. To list IBM Cloud services, run `ibmcloud catalog service-marketplace` and use the **Name** value from the output.
              *   `<PLAN>` is the plan for the IBM Cloud service that you want to create, such as `free` or `standard`. To list available plans, run `ibmcloud catalog service <SERVICE_CLASS> | grep plan`.

              ```yaml
              apiVersion: ibmcloud.ibm.com/v1
              kind: Service
              metadata:
                  name: myservice
              spec:
                  plan: <PLAN>
                  serviceClass: <SERVICE_CLASS>
              ```

          2.  Create the service instance in your cluster.

              ```bash
              kubectl apply -f filepath/myservice.yaml
              ```

          3.  Check that your service status is **Online** in your cluster.

              ```bash
              kubectl get services.ibmcloud
              NAME           STATUS   AGE
              myservice      Online   12s
              ```

          4.  Verify that your service instance is created in IBM Cloud.

              ```bash
              ibmcloud resource service-instances | grep myservice
              ```

          #### Step 2: Binding the service instance

          1.  To bind your service to the cluster so that your apps can use the service, create a `Binding` custom resource, where the `serviceName` field is the name of the `Service` custom resource that you previously created. For more options, see [Binding properties](https://github.com/IBM/cloud-operators/blob/master/README.md#binding-properties).

              ```yaml
              apiVersion: ibmcloud.ibm.com/v1
              kind: Binding
              metadata:
                  name: mybinding
              spec:
                  serviceName: myservice
              ```

          2.  Create the binding in your cluster.

              ```bash
              kubectl apply -f filepath/mybinding.yaml
              ```

          3.  Check that your service status is **Online**.

              ```bash
              kubectl get bindings.ibmcloud
              NAME         STATUS   AGE
              mybinding    Online   25s
              ```

          4.  Check that a secret of the same name as your binding is created. The secret contains the service credentials that apps in your cluster can use to access the service.

              ```bash
              kubectl get secrets
              NAME         TYPE      DATA   AGE
              mybinding    Opaque    6      102s
              ```
        displayName: IBM Cloud Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: Source on GitHub
          url: https://github.com/ibm/cloud-operators
        maintainers:
        - email: johnstarich@johnstarich.com
          name: John Starich
        - email: mvaziri@us.ibm.com
          name: vazirim
        - email: greg.lanthier@ca.ibm.com
          name: Greg Lanthier
        - email: halacs@users.noreply.github.com
          name: Halacs
        - email: artberger@users.noreply.github.com
          name: Art
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: IBM
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - cloudoperators/ibmcloud-operator:1.1.0
        version: 1.1.0
      entries:
      - name: ibmcloud-operator.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: ibmcloud-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Infinispan
      provider-url: ""
    name: infinispan
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infinispan-operator.v2.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.2.5.Final
          description: Create and manage Infinispan clusters.
          olm.skipRange: '>=2.1.x <2.2.1'
          operators.operatorframework.io/builder: operator-sdk-v1.3.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/infinispan/operator:2.2.5.Final
        version: 2.2.5
      entries:
      - name: infinispan-operator.v2.2.5
        version: 2.2.5
      - name: infinispan-operator.v2.2.4
        version: 2.2.4
      - name: infinispan-operator.v2.2.3
        version: 2.2.3
      - name: infinispan-operator.v2.2.2
        version: 2.2.2
      - name: infinispan-operator.v2.2.1
        version: 2.2.1
      - name: infinispan-operator.v2.2.0
        version: 2.2.0
      name: 2.2.x
    - currentCSV: infinispan-operator.v2.3.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.3.7.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/infinispan/operator:2.3.7.Final
        version: 2.3.8
      entries:
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: 2.3.x
    - currentCSV: infinispan-operator.v2.4.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.4.17.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/infinispan/operator:2.4.17.Final
        version: 2.4.17
      entries:
      - name: infinispan-operator.v2.4.17
        version: 2.4.17
      - name: infinispan-operator.v2.4.16
        version: 2.4.16
      - name: infinispan-operator.v2.4.15
        version: 2.4.15
      - name: infinispan-operator.v2.4.14
        version: 2.4.14
      - name: infinispan-operator.v2.4.13
        version: 2.4.13
      - name: infinispan-operator.v2.4.12
        version: 2.4.12
      - name: infinispan-operator.v2.4.11
        version: 2.4.11
      - name: infinispan-operator.v2.4.10
        version: 2.4.10
      - name: infinispan-operator.v2.4.9
        version: 2.4.9
      - name: infinispan-operator.v2.4.8
        version: 2.4.8
      - name: infinispan-operator.v2.4.7
        version: 2.4.7
      - name: infinispan-operator.v2.4.6
        version: 2.4.6
      - name: infinispan-operator.v2.4.5
        version: 2.4.5
      - name: infinispan-operator.v2.4.4
        version: 2.4.4
      - name: infinispan-operator.v2.4.3
        version: 2.4.3
      - name: infinispan-operator.v2.4.2
        version: 2.4.2
      - name: infinispan-operator.v2.4.1
        version: 2.4.1
      - name: infinispan-operator.v2.4.0
        version: 2.4.0
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: 2.4.x
    - currentCSV: infinispan-operator.v2.4.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.4.18.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/infinispan/operator:2.4.18.Final
        version: 2.4.18
      entries:
      - name: infinispan-operator.v2.4.18
        version: 2.4.18
      - name: infinispan-operator.v2.4.16
        version: 2.4.16
      - name: infinispan-operator.v2.4.15
        version: 2.4.15
      - name: infinispan-operator.v2.4.14
        version: 2.4.14
      - name: infinispan-operator.v2.4.13
        version: 2.4.13
      - name: infinispan-operator.v2.4.12
        version: 2.4.12
      - name: infinispan-operator.v2.4.11
        version: 2.4.11
      - name: infinispan-operator.v2.4.10
        version: 2.4.10
      - name: infinispan-operator.v2.4.9
        version: 2.4.9
      - name: infinispan-operator.v2.4.8
        version: 2.4.8
      - name: infinispan-operator.v2.4.7
        version: 2.4.7
      - name: infinispan-operator.v2.4.6
        version: 2.4.6
      - name: infinispan-operator.v2.4.5
        version: 2.4.5
      - name: infinispan-operator.v2.4.4
        version: 2.4.4
      - name: infinispan-operator.v2.4.3
        version: 2.4.3
      - name: infinispan-operator.v2.4.2
        version: 2.4.2
      - name: infinispan-operator.v2.4.1
        version: 2.4.1
      - name: infinispan-operator.v2.4.0
        version: 2.4.0
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: stable
    defaultChannel: stable
    packageName: infinispan
    provider:
      name: Infinispan
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent, part of Cisco
      provider-url: https://isovalent.com/
    name: clife
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clife.v1.17.8-cee.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cilium.io/v1alpha1",
                "kind": "CiliumConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "clife"
                  },
                  "name": "ciliumconfig"
                },
                "spec": {
                  "cluster": {
                    "name": "default"
                  },
                  "clusterHealthPort": 9940,
                  "cni": {
                    "binPath": "/var/lib/cni/bin",
                    "chainingMode": "portmap",
                    "confPath": "/var/run/multus/cni/net.d",
                    "exclusive": false
                  },
                  "enterprise": {
                    "featureGate": {
                      "approved": [
                        "CNIChainingMode"
                      ]
                    }
                  },
                  "hubble": {
                    "enabled": true
                  },
                  "ipam": {
                    "mode": "cluster-pool"
                  },
                  "kubeProxyReplacement": false,
                  "operator": {
                    "prometheus": {
                      "enabled": true,
                      "serviceMonitor": {
                        "enabled": true
                      }
                    }
                  },
                  "prometheus": {
                    "enabled": true,
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "securityContext": {
                    "privileged": true
                  },
                  "sessionAffinity": true,
                  "tunnelPort": 4789
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Networking,Security
          createdAt: "2025-09-23T07:55:19Z"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CiliumConfig defines the configuration of Isovalent Networking
              for Kubernetes and all its components
            displayName: CiliumConfig
            kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Networking for Kubernetes, built on Cilium, provides
          eBPF-based networking, observability, and security with optimal scale and
          performance for platform teams operating Kubernetes environments across
          cloud and on-prem infrastructure.
        displayName: Isovalent Networking for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cilium
        - networking
        - observability
        - security
        - eBPF
        - CNI
        links:
        - name: Isovalent Networking for Kubernetes
          url: https://isovalent.com/product/
        maintainers:
        - email: support@isovalent.com
          name: Isovalent, part of Cisco
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Isovalent, part of Cisco
          url: https://isovalent.com/
        relatedImages:
        - quay.io/isovalent/startup-script-ubi:v1.17.8-cee.1@sha256:9a6a413262ec59026345aefc743de847891d80e31ef2f41db954acdcee11c2ff
        - quay.io/isovalent/certgen-ubi:v1.17.8-cee.1@sha256:fb63bed641a279cc09114e93821c7b7555c61393ab594ea2b823ce34fa73335e
        - quay.io/isovalent/cilium-envoy-ubi:v1.17.8-cee.1@sha256:c3de0334f90f9b2540cdec7e2356d813f2df93f0d7d8701005eab7161c7663b5
        - quay.io/isovalent/clife@sha256:ba9bd4a5c35e1d26e395afbc2a4de89f42df1e5e61c9b23959c23194368b7b0e
        - quay.io/isovalent/operator-generic-ubi:v1.17.8-cee.1@sha256:37e4b9f65b2c67d24afeeebebe5d0cb59c09280179de57cd77da2b68e19c3c1b
        - quay.io/isovalent/cilium-ubi:v1.17.8-cee.1@sha256:e9816096e481fad5d95209f902e4a7018aeb65d1e3be1574e6b289e21229f87a
        - quay.io/isovalent/hubble-relay-ubi:v1.17.8-cee.1@sha256:88964d4836d59420dc7e0abb758654ceaf2abcfb53e19bbe2c7962aa68874e37
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.17.8-cee.1@sha256:b097c823638b4fdbe5bdd2b6040a17fef0c70a020e67a18f8231d60eeac02dd5
        version: 1.17.8-cee.1
      entries:
      - name: clife.v1.17.8-cee.1
        version: 1.17.8-cee.1
      - name: clife.v1.17.7-cee.1
        version: 1.17.7-cee.1
      - name: clife.v1.17.5-cee.1
        version: 1.17.5-cee.1
      - name: clife.v1.17.4-cee.1
        version: 1.17.4-cee.1
      - name: clife.v0.0.1
        version: 0.0.1
      name: "1.17"
    defaultChannel: "1.17"
    packageName: clife
    provider:
      name: Isovalent, part of Cisco
      url: https://isovalent.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microsafe
      provider-url: https://microsafe.com.mx/
    name: operator-aims-mobile
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operator-aims-mobile.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.microsafe.com.mx/v1",
                "kind": "AimsMobile",
                "metadata": {
                  "name": "aimsmobile-sample"
                },
                "spec": {
                  "adminApp": {
                    "autoscaling": {
                      "enabled": false
                    },
                    "config": {
                      "NODE_ENV": "production",
                      "VITE_EXPRESS_HOST": "aims-mob-service",
                      "VITE_EXPRESS_PORT": "8080",
                      "VITE_EXPRESS_PROTOCOL": "http",
                      "VITE_KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "VITE_KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "VITE_KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "VITE_VERSION": "1.0.0"
                    },
                    "image": "quay.io/microsafe_app_ban/aims-mob-admin-private@sha256:8e333f900885b0d3f6442c8ffcb43bd93540b8c38ad5bf17cbbd78929e0b1be0",
                    "ingress": {
                      "enabled": true,
                      "isOpenshiftRoute": true
                    },
                    "name": "aims-mob-admin",
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "secretName": "aims-mob-admin-secrets",
                    "service": {
                      "enabled": true,
                      "port": 8083,
                      "targetPort": 8080
                    }
                  },
                  "dbService": {
                    "autoscaling": {
                      "enabled": false
                    },
                    "config": {
                      "MSSQL_HOST": "mssql-database.prod.host",
                      "MSSQL_PORT": "mssql-database.prod.port"
                    },
                    "image": "quay.io/microsafe_app_ban/aims-db-service-private@sha256:bcb4d8624626436526549427f9f169cff714b39cbcc530dd2b3470c6712abaf9",
                    "ingress": {
                      "enabled": false
                    },
                    "name": "aims-db-service",
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "300m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "128Mi"
                      }
                    },
                    "secretName": "aims-db-service-secrets",
                    "service": {
                      "enabled": true,
                      "port": 3000,
                      "targetPort": 3000
                    }
                  },
                  "mainService": {
                    "annotations": {
                      "instrumentation.opentelemetry.io/inject-nodejs": "true"
                    },
                    "autoscaling": {
                      "enabled": true,
                      "maxReplicas": 2,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "config": {
                      "AIMS1_API_URL": "http://aims-db-service:3000",
                      "EXPRESS_PORT": "4000",
                      "FRONTEND_EXPRESS_HOST": "aims-mob-admin",
                      "FRONTEND_EXPRESS_PORT": "4000",
                      "FRONTEND_EXPRESS_PROTOCOL": "http",
                      "FRONTEND_IP": "aims-mob-admin",
                      "FRONTEND_KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "FRONTEND_KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "FRONTEND_KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "FRONTEND_PORT": "80",
                      "FRONTEND_PROTOCOL": "http",
                      "FRONTEND_VERSION": "1.0.0",
                      "KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "MSSQL_HOST": "mssql-database.prod.host",
                      "MSSQL_PORT": "mssql-database.prod.port",
                      "NODE_ENV": "production",
                      "OTLP_LOG_EXPORTER_URL": "opentelemetry-logs.prod.url",
                      "OTLP_METRIC_EXPORTER_URL": "opentelemetry-metrics.prod.url",
                      "OTLP_TRACE_EXPORTER_URL": "opentelemetry-traces.prod.url",
                      "REMEDY_API_URL": "remedy-aims.prod.svc",
                      "REMEDY_USER": "RMDYUSR"
                    },
                    "ingress": {
                      "enabled": true,
                      "isOpenshiftRoute": true
                    },
                    "name": "aims-mob-service",
                    "proxyContainer": {
                      "image": "quay.io/microsafe_app_ban/aims-mob-proxy-private@sha256:29b82de987b28c23f17d5f2176877c5aecea1a0ef182bb3a63612290ff469949",
                      "name": "aims-mob-service-proxy",
                      "resources": {
                        "limits": {
                          "cpu": "2000m",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "replicas": 2,
                    "secretName": "aims-mob-service-secrets",
                    "service": {
                      "enabled": true,
                      "port": 8085,
                      "targetPort": 8081
                    },
                    "serviceContainer": {
                      "image": "quay.io/microsafe_app_ban/aims-mob-service-private@sha256:f9586200106aa4de259f5cae2d22d13dc147a76dfd06e4f4bea09797071dad17",
                      "name": "aims-mob-service-main",
                      "resources": {
                        "limits": {
                          "cpu": "2500m",
                          "memory": "2048Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "128Mi"
                        }
                      }
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": false,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          createdAt: "2025-09-25T19:23:26Z"
          description: Este operador despliega los servicios AIMS Mobile.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/network-aware: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          instrumentation.opentelemetry.io/inject-nodejs: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: Valid Subscription Microsafe
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://quay.io/repository/microsafe_app_ban/operador-aims-mobile-private
          support: Microsafe
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AimsMobile
            name: aimsmobiles.cache.microsafe.com.mx
            version: v1
        description: Este operador despliega la aplicación AIMS Mobile.
        displayName: AIMS Mobile Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aims mobile
        - aims
        links:
        - name: AIMS Mobile Operator
          url: https://microsafe.com.mx/
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Microsafe
          url: https://microsafe.com.mx/
        relatedImages:
        - quay.io/microsafe_app_ban/aims-db-service-private@sha256:bcb4d8624626436526549427f9f169cff714b39cbcc530dd2b3470c6712abaf9
        - quay.io/microsafe_app_ban/aims-mob-admin-private@sha256:8e333f900885b0d3f6442c8ffcb43bd93540b8c38ad5bf17cbbd78929e0b1be0
        - quay.io/microsafe_app_ban/operador-aims-mobile-private@sha256:731f2d731c79ade02f4b6c39b380590833cacddfd795b703cec10ad55f730fad
        - quay.io/microsafe_app_ban/aims-mob-proxy-private@sha256:29b82de987b28c23f17d5f2176877c5aecea1a0ef182bb3a63612290ff469949
        - quay.io/microsafe_app_ban/aims-mob-service-private@sha256:f9586200106aa4de259f5cae2d22d13dc147a76dfd06e4f4bea09797071dad17
        version: 0.0.10
      entries:
      - name: operator-aims-mobile.v0.0.10
        version: 0.0.10
      - name: operator-aims-mobile.v0.0.9
        version: 0.0.9
      - name: operator-aims-mobile.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: operator-aims-mobile
    provider:
      name: Microsafe
      url: https://microsafe.com.mx/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzograph-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzograph-operator.v2.0.102
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzograph.clusters.cambridgesemantics.com/v2",
                "kind": "AnzoGraph",
                "metadata": {
                  "name": "azg01"
                },
                "spec": {
                  "db": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_data": "anzograph-data-grpc",
                            "app_mgmt": "anzograph-mgmt-grpc"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_data": "anzograph-data-grpc",
                              "app_mgmt": "anzograph-mgmt-grpc"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-db@sha256:8a5bd26551990929fdb776d6b421a9713cb84e03b19c18f81b625582f7f41af5",
                                "name": "db",
                                "resources": {
                                  "limits": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  },
                                  "requests": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  },
                  "deployFrontend": false,
                  "frontend": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_mgmt": "anzograph-frontend"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_mgmt": "anzograph-frontend"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-frontend@sha256:c1443bcdc8e998fc699f61980a665f1a083350adf2d3d1d55a7cb55e31ca50b3",
                                "name": "frontend",
                                "resources": {
                                  "limits": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-operator@sha256:a5b212558f748ddc8156541b546290320b656506e039a18903ad9fa6e638e732
          createdAt: "2021-08-11"
          description: kubernetes operator for AnzoGraph DB
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzograph-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzograph-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoGraph is the Schema for the anzographs API
            displayName: AnzoGraph Operator
            kind: AnzoGraph
            name: anzographs.anzograph.clusters.cambridgesemantics.com
            version: v2
        description: |-
          The AnzoGraph Operator provides the way to install and configure an AnzoGraph
          cluster on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v2.0.1/README_openshift_marketplace.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzograph/userdoc/ )

          ### Support

          We offer Support to our customers with the AnzoGraph db Enterprise Edition License [here]( https://customercenter.cambridgesemantics.com/ ). For AnzoGraph db Free Edition questions, get help from our Anzograph User Community at Stack Overflow. When submitting a question, include the tag 'anzograph'.
        displayName: AnzoGraph Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        - Database
        links:
        - name: Anzograph Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v2.0.1/README_openshift_marketplace.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-operator@sha256:a5b212558f748ddc8156541b546290320b656506e039a18903ad9fa6e638e732
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph@sha256:e3bdddf811b5ba69f388d023aea1ba538edd00e9f415c33ec555321d39129a36
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-db@sha256:8a5bd26551990929fdb776d6b421a9713cb84e03b19c18f81b625582f7f41af5
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-frontend@sha256:c1443bcdc8e998fc699f61980a665f1a083350adf2d3d1d55a7cb55e31ca50b3
        version: 2.0.1
      entries:
      - name: anzograph-operator.v2.0.102
        version: 2.0.1
      name: stable
    defaultChannel: stable
    packageName: anzograph-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: kubernetes-nmstate-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubernetes-nmstate-operator.4.13.0-202309261702
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nmstate.io/v1",
                "kind": "NMState",
                "metadata": {
                  "name": "nmstate"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift4/kubernetes-nmstate-rhel8-operator@sha256:bdcb80591b839ec6b4b94efa838ca41585f3044826534413b91efd2acdca4b4d
          createdAt: "2022-02-21 08:46:16"
          description: |
            Kubernetes NMState is a declaritive means of configuring NetworkManager.
          olm.skipRange: '>=4.3.0 <4.13.0'
          operatorframework.io/suggested-namespace: openshift-nmstate
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/kubernetes-nmstate
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an NMState deployment.
            displayName: NMState
            kind: NMState
            name: nmstates.nmstate.io
            version: v1
        description: A Kubernetes Operator to install Kubernetes NMState
        displayName: Kubernetes NMState Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - nmstate
        - networking
        - NetworkManager
        links:
        - name: Kubernetes Nmstate Operator
          url: https://github.com/nmstate/kubernetes-nmstate
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: GA
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/kubernetes-nmstate-rhel8-operator@sha256:bdcb80591b839ec6b4b94efa838ca41585f3044826534413b91efd2acdca4b4d
        - registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel8@sha256:a96434b959c1f61c1cfb4352284aaae453d2e41ca0824d016975a1216451b665
        version: 4.13.0-202309261702
      entries:
      - name: kubernetes-nmstate-operator.4.13.0-202309261702
        version: 4.13.0-202309261702
      name: "4.13"
    - currentCSV: kubernetes-nmstate-operator.4.16.0-202509241326
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nmstate.io/v1",
                "kind": "NMState",
                "metadata": {
                  "name": "nmstate"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:14750ac40ba2a9a92e2b37351eed94c22f2663747ea09f241a3a80607a9fca09
          createdAt: "2022-02-21 08:46:16"
          description: |
            Kubernetes NMState is a declaritive means of configuring NetworkManager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-nmstate
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/kubernetes-nmstate
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an NMState deployment.
            displayName: NMState
            kind: NMState
            name: nmstates.nmstate.io
            version: v1
        description: A Kubernetes Operator to install Kubernetes NMState
        displayName: Kubernetes NMState Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - nmstate
        - networking
        - NetworkManager
        links:
        - name: Kubernetes Nmstate Operator
          url: https://github.com/nmstate/kubernetes-nmstate
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: GA
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:14750ac40ba2a9a92e2b37351eed94c22f2663747ea09f241a3a80607a9fca09
        - registry.redhat.io/openshift4/nmstate-console-plugin-rhel9@sha256:8d03a6c414da30521e0794d3d61336bfbb1d9d4c3fec5ab5dac4bd24c897a540
        - registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel9@sha256:c592fdb65d99de2c3d274d7e72621408e578bdba39753bb425d0765ac016b11a
        version: 4.16.0-202509241326
      entries:
      - name: kubernetes-nmstate-operator.4.16.0-202509241326
        version: 4.16.0-202509241326
      - name: kubernetes-nmstate-operator.4.16.0-202509101329
        version: 4.16.0-202509101329
      - name: kubernetes-nmstate-operator.4.16.0-202508251030
        version: 4.16.0-202508251030
      - name: kubernetes-nmstate-operator.4.16.0-202507221936
        version: 4.16.0-202507221936
      - name: kubernetes-nmstate-operator.4.16.0-202507100305
        version: 4.16.0-202507100305
      - name: kubernetes-nmstate-operator.4.16.0-202506250035
        version: 4.16.0-202506250035
      - name: kubernetes-nmstate-operator.4.16.0-202506041738
        version: 4.16.0-202506041738
      - name: kubernetes-nmstate-operator.4.16.0-202505150035
        version: 4.16.0-202505150035
      - name: kubernetes-nmstate-operator.4.16.0-202505060805
        version: 4.16.0-202505060805
      - name: kubernetes-nmstate-operator.4.16.0-202504141106
        version: 4.16.0-202504141106
      - name: kubernetes-nmstate-operator.4.16.0-202503260136
        version: 4.16.0-202503260136
      - name: kubernetes-nmstate-operator.4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: kubernetes-nmstate-operator.4.16.0-202502190034
        version: 4.16.0-202502190034
      - name: kubernetes-nmstate-operator.4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: kubernetes-nmstate-operator.4.16.0-202501281135
        version: 4.16.0-202501281135
      - name: kubernetes-nmstate-operator.4.16.0-202501211505
        version: 4.16.0-202501211505
      - name: kubernetes-nmstate-operator.4.16.0-202501080105
        version: 4.16.0-202501080105
      - name: kubernetes-nmstate-operator.4.16.0-202412180105
        version: 4.16.0-202412180105
      - name: kubernetes-nmstate-operator.4.16.0-202412110604
        version: 4.16.0-202412110604
      - name: kubernetes-nmstate-operator.4.16.0-202412040032
        version: 4.16.0-202412040032
      - name: kubernetes-nmstate-operator.4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: kubernetes-nmstate-operator.4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: kubernetes-nmstate-operator.4.16.0-202410300036
        version: 4.16.0-202410300036
      - name: kubernetes-nmstate-operator.4.16.0-202410231737
        version: 4.16.0-202410231737
      - name: kubernetes-nmstate-operator.4.16.0-202410180404
        version: 4.16.0-202410180404
      - name: kubernetes-nmstate-operator.4.16.0-202410091137
        version: 4.16.0-202410091137
      - name: kubernetes-nmstate-operator.4.16.0-202410021935
        version: 4.16.0-202410021935
      - name: kubernetes-nmstate-operator.4.16.0-202409251605
        version: 4.16.0-202409251605
      - name: kubernetes-nmstate-operator.4.16.0-202409111235
        version: 4.16.0-202409111235
      - name: kubernetes-nmstate-operator.4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: kubernetes-nmstate-operator.4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: kubernetes-nmstate-operator.4.16.0-202408211407
        version: 4.16.0-202408211407
      - name: kubernetes-nmstate-operator.4.16.0-202408140143
        version: 4.16.0-202408140143
      - name: kubernetes-nmstate-operator.4.16.0-202408070540
        version: 4.16.0-202408070540
      - name: kubernetes-nmstate-operator.4.16.0-202407251436
        version: 4.16.0-202407251436
      - name: kubernetes-nmstate-operator.4.16.0-202407240806
        version: 4.16.0-202407240806
      - name: kubernetes-nmstate-operator.4.16.0-202407181806
        version: 4.16.0-202407181806
      - name: kubernetes-nmstate-operator.4.16.0-202407101306
        version: 4.16.0-202407101306
      - name: kubernetes-nmstate-operator.4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: kubernetes-nmstate-operator.4.16.0-202406260236
        version: 4.16.0-202406260236
      - name: kubernetes-nmstate-operator.4.16.0-202406140306
        version: 4.16.0-202406140306
      - name: kubernetes-nmstate-operator.4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: kubernetes-nmstate-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: istio-workspace-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: istio-workspace-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "workspace.maistra.io/v1alpha1",
                "kind": "Session",
                "metadata": {
                  "name": "session-sample"
                },
                "spec": {
                  "ref": [
                    {
                      "args": {
                        "image": "quay.io/maistra-dev/istio-workspace-test:latest"
                      },
                      "name": "ratings-v1",
                      "strategy": "prepared-image"
                    }
                  ],
                  "route": {
                    "name": "x-workspace-route",
                    "type": "header",
                    "value": "feature-x"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/maistra/istio-workspace:v0.5.3
          createdAt: "2022-11-16T13:50:35Z"
          description: Safely develop and test on any Kubernetes cluster without affecting
            others.
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/maistra/istio-workspace
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Session controls the creation of the specialized hidden routes.
            displayName: Session
            kind: Session
            name: sessions.workspace.maistra.io
            version: v1alpha1
        description: "Istio Workspace enables developers to:\n\n * run one or more
          services locally during development but make it appear like it would be
          developed in the actual cluster\n   * this way you can use your favourite
          development tools\n   * have instant feedback loop\n   * don't make your
          machine overheating trying to run the whole cluster locally\n * safely preview
          incoming changes in pull requests without affecting regular users \n * have
          confidence in testing or troubleshooting your services directly in the cluster\n\nIstio
          Workspace supports testing on multi-user environment in an unobtrusive way.\nIt
          doesn’t really matter if it is QE cluster or actual production. We give
          you confidence that your changes won’t blow up the cluster, and your users
          won’t even notice a glitch.\n\nYou will need [Kubernetes](https://k8s.io)
          or [Openshift](https://openshift.com) cluster with  [Istio](https://istio.io/)
          installed. \n\nYou can read more about our vision on [Red Hat Developer’s
          blog](https://developers.redhat.com/blog/2020/07/14/developing-and-testing-on-production-with-kubernetes-and-istio-workspace/)
          or ...\n\n## See it in action!\n\n[![https://youtu.be/XTNVadUzMCc](https://img.youtube.com/vi/XTNVadUzMCc/hqdefault.jpg)](https://youtu.be/XTNVadUzMCc)\n\nWatch
          demo: [\"How to develop on production: An introduction to Istio-Workspaces\"](https://youtu.be/XTNVadUzMCc).\n\n##
          Documentation\n\nMore details can be found on our [documentation page](https://istio-workspace-docs.netlify.com/).\n"
        displayName: Istio Workspace
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - telepresence
        links:
        - name: Istio Workspace Documentation
          url: https://istio-workspace-docs.netlify.com/
        - name: How to install CLI tool
          url: https://github.com/maistra/istio-workspace#install-in-two-easy-steps
        maintainers:
        - email: aslak@redhat.com
          name: Aslak Knutsen
        - email: bartosz@redhat.com
          name: Bartosz Majsak
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/maistra/istio-workspace:v0.5.3
        version: 0.5.3
      entries:
      - name: istio-workspace-operator.v0.5.3
        version: 0.5.3
      - name: istio-workspace-operator.v0.5.2
        version: 0.5.2
      - name: istio-workspace-operator.v0.5.1
        version: 0.5.1
      - name: istio-workspace-operator.v0.5.0
        version: 0.5.0
      - name: istio-workspace-operator.v0.4.0
        version: 0.4.0
      - name: istio-workspace-operator.v0.3.0
        version: 0.3.0
      - name: istio-workspace-operator.v0.2.0
        version: 0.2.0
      - name: istio-workspace-operator.v0.1.0
        version: 0.1.0
      - name: istio-workspace-operator.v0.0.10
        version: 0.0.10
      - name: istio-workspace-operator.v0.0.9
        version: 0.0.9
      - name: istio-workspace-operator.v0.0.8
        version: 0.0.8
      - name: istio-workspace-operator.v0.0.7
        version: 0.0.7
      - name: istio-workspace-operator.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: istio-workspace-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ThinkParQ
      provider-url: https://www.thinkparq.com
    name: beegfs-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: beegfs-csi-driver-operator.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "beegfs.csi.netapp.com/v1",
                "kind": "BeegfsDriver",
                "metadata": {
                  "name": "csi-beegfs-cr"
                },
                "spec": {
                  "nodeAffinityControllerService": {
                    "preferredDuringSchedulingIgnoredDuringExecution": [
                      {
                        "preference": {
                          "matchExpressions": [
                            {
                              "key": "node-role.kubernetes.io/master",
                              "operator": "Exists"
                            }
                          ]
                        },
                        "weight": 50
                      }
                    ],
                    "requiredDuringSchedulingIgnoredDuringExecution": {
                      "nodeSelectorTerms": [
                        {
                          "matchExpressions": [
                            {
                              "key": "node.openshift.io/os_id",
                              "operator": "NotIn",
                              "values": [
                                "rhcos"
                              ]
                            }
                          ]
                        }
                      ]
                    }
                  },
                  "nodeAffinityNodeService": {
                    "requiredDuringSchedulingIgnoredDuringExecution": {
                      "nodeSelectorTerms": [
                        {
                          "matchExpressions": [
                            {
                              "key": "node.openshift.io/os_id",
                              "operator": "NotIn",
                              "values": [
                                "rhcos"
                              ]
                            }
                          ]
                        }
                      ]
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: ghcr.io/thinkparq/beegfs-csi-driver-operator:v1.7.0
          createdAt: 2021-06-30:03:11:00
          description: Install and manage the BeeGFS CSI driver
          operatorframework.io/suggested-namespace: beegfs-csi
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ThinkParQ/beegfs-csi-driver
          support: ThinkParQ
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys the BeeGFS CSI driver
            displayName: BeeGFS Driver
            kind: BeegfsDriver
            name: beegfsdrivers.beegfs.csi.netapp.com
            version: v1
        description: |
          The BeeGFS Container Storage Interface (CSI) driver provides high performing and scalable storage for workloads
          running in container orchestrators like Kubernetes. It allows containers to access existing datasets or request
          on-demand ephemeral or persistent high speed storage backed by BeeGFS parallel file systems.

          This operator eases the installation and management of the BeeGFS CSI driver in an OpenShift or other Operator
          Lifecycle Manager enabled cluster.

          ### Documentation

          Compatibility information and a quick start guide for the BeeGFS CSI driver itself are available on our [main GitHub
          page](https://www.github.com/ThinkParQ/beegfs-csi-driver). Additional detailed documentation for the driver is
          available in the [docs directory](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/docs) and for the
          operator is available in the [operator
          README](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/operator/README.md).

          ### Important Notes

          The BeeGFS client packages and kernel module must be pre-installed on all nodes that will run the driver. See the
          driver ["Getting Started" guide](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/README.md#getting-started)
          for additional information.

          The BeeGFS CSI driver is only supported on nodes running a [BeeGFS supported Linux
          distribution](https://doc.beegfs.io/latest/release_notes.html#supported-linux-distributions-and-kernels). Red Hat
          CoreOS (RHCOS), the default distribution in OpenShift environments, is NOT supported. However, Red Hat Enterprise
          Linux (RHEL) nodes can be added to an OpenShift cluster and RHEL is supported for running BeeGFS. By default, the
          operator will install the driver with a node affinity that ensures it does not run on RHCOS nodes. See the driver
          [compatibility
          matrix](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/README.md#interoperability-and-csi-feature-matrix)
          for additional information.

          We have recently published experimental manifests that enable the BeeGFS CSI driver on OpenShift nodes running
          RHCOS. These manifests are not intended for production use. See [this
          README](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/deploy/openshift-beegfs-client/README.md) for more
          details.

          All driver configuration can be specified through the BeegfsDrivers API (in a BeegfsDriver Custom Resource)
          except for BeeGFS connAuth information, which must be placed separately into a Kubernetes Secret. See the [ConnAuth
          Configuration
          section](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/operator/README.md#connauth-configuration) of the
          operator documentation for additional information.

          Starting with version 1.4.0 of the BeeGFS CSI driver the driver and operator images will be signed using the Cosign
          tool. Refer to the [BeeGFS CSI driver operator
          README](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/operator/README.md)
          for details on how to verify the operator and driver image signatures.
        displayName: BeeGFS CSI Driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CSI
        - BeeGFS
        - ThinkParQ
        - HPC
        - AI
        - High-Performance Computing
        links:
        - name: BeeGFS introduction
          url: https://www.beegfs.io/c/
        - name: BeeGFS documentation
          url: https://doc.beegfs.io/latest/index.html
        - name: BeeGFS CSI driver source code
          url: https://github.com/ThinkParQ/beegfs-csi-driver
        - name: BeeGFS blog posts
          url: https://www.beegfs.io/c/blog/
        maintainers:
        - email: csi-driver@beegfs.io
          name: ThinkParQ
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: ThinkParQ
          url: https://www.thinkparq.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - ghcr.io/thinkparq/beegfs-csi-driver-operator:v1.7.0
        version: 1.7.0
      entries:
      - name: beegfs-csi-driver-operator.v1.7.0
        version: 1.7.0
      - name: beegfs-csi-driver-operator.v1.6.0
        version: 1.6.0
      - name: beegfs-csi-driver-operator.v1.5.0
        version: 1.5.0
      - name: beegfs-csi-driver-operator.v1.4.0
        version: 1.4.0
      - name: beegfs-csi-driver-operator.v1.3.0
        version: 1.3.0
      - name: beegfs-csi-driver-operator.v1.2.2
        version: 1.2.2
      - name: beegfs-csi-driver-operator.v1.2.1
        version: 1.2.1
      - name: beegfs-csi-driver-operator.v1.2.0
        version: 1.2.0
      name: stable
    defaultChannel: stable
    packageName: beegfs-csi-driver-operator
    provider:
      name: ThinkParQ
      url: https://www.thinkparq.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stackwatch
      provider-url: www.kubecost.com
    name: kubecost
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubecost.v2.2.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kubecost.com/v1alpha1",
                "kind": "CostAnalyzer",
                "metadata": {
                  "name": "costanalyzer-sample"
                },
                "spec": {
                  "global": {
                    "prometheus": {
                      "enabled": true,
                      "fqdn": "http://cost-analyzer-prometheus-server.default.svc"
                    },
                    "grafana": {
                      "enabled": false,
                      "domainName": "cost-analyzer-grafana.default.svc",
                      "scheme": "http",
                      "proxy": false
                    },
                    "gcpstore": {
                      "enabled": false
                    },
                    "gmp": {
                      "enabled": false,
                      "prometheusServerEndpoint": "http://localhost:8085/",
                      "gmpProxy": {
                        "enabled": false,
                        "image": "gke.gcr.io/prometheus-engine/frontend:v0.4.1-gke.0",
                        "imagePullPolicy": "Always",
                        "name": "gmp-proxy",
                        "port": 8085,
                        "projectId": "YOUR_PROJECT_ID"
                      }
                    },
                    "amp": {
                      "enabled": false,
                      "prometheusServerEndpoint": "http://localhost:8005/workspaces/<workspaceId>/",
                      "remoteWriteService": "https://aps-workspaces.us-west-2.amazonaws.com/workspaces/<workspaceId>/api/v1/remote_write",
                      "sigv4": {
                        "region": "us-west-2"
                      }
                    },
                    "mimirProxy": {
                      "enabled": false,
                      "name": "mimir-proxy",
                      "image": "nginxinc/nginx-unprivileged",
                      "port": 8085,
                      "mimirEndpoint": "$mimir_endpoint",
                      "orgIdentifier": "$your_tenant_ID"
                    },
                    "notifications": {
                      "alertmanager": {
                        "enabled": false,
                        "fqdn": "http://cost-analyzer-prometheus-server.default.svc"
                      }
                    },
                    "savedReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Saved Report 0",
                          "window": "today",
                          "aggregateBy": "namespace",
                          "chartDisplay": "category",
                          "idle": "separate",
                          "rate": "cumulative",
                          "accumulate": false,
                          "filters": [
                            {
                              "key": "cluster",
                              "operator": ":",
                              "value": "dev"
                            }
                          ]
                        },
                        {
                          "title": "Example Saved Report 1",
                          "window": "month",
                          "aggregateBy": "controllerKind",
                          "chartDisplay": "category",
                          "idle": "share",
                          "rate": "monthly",
                          "accumulate": false,
                          "filters": [
                            {
                              "key": "namespace",
                              "operator": "!:",
                              "value": "kubecost"
                            }
                          ]
                        },
                        {
                          "title": "Example Saved Report 2",
                          "window": "2020-11-11T00:00:00Z,2020-12-09T23:59:59Z",
                          "aggregateBy": "service",
                          "chartDisplay": "category",
                          "idle": "hide",
                          "rate": "daily",
                          "accumulate": true,
                          "filters": []
                        }
                      ]
                    },
                    "assetReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Asset Report 0",
                          "window": "today",
                          "aggregateBy": "type",
                          "accumulate": false,
                          "filters": [
                            {
                              "property": "cluster",
                              "value": "cluster-one"
                            }
                          ]
                        }
                      ]
                    },
                    "advancedReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Advanced Report 0",
                          "window": "7d",
                          "aggregateBy": "namespace",
                          "filters": [
                            {
                              "key": "cluster",
                              "operator": ":",
                              "value": "dev"
                            }
                          ],
                          "cloudBreakdown": "service",
                          "cloudJoin": "label:kubernetes_namespace"
                        }
                      ]
                    },
                    "cloudCostReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Cloud Cost Report 0",
                          "window": "today",
                          "aggregateBy": "service",
                          "accumulate": false
                        }
                      ]
                    },
                    "podAnnotations": {},
                    "additionalLabels": {},
                    "securityContext": {
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      },
                      "fsGroup": 1001,
                      "runAsGroup": 1001,
                      "runAsUser": 1001,
                      "fsGroupChangePolicy": "OnRootMismatch"
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "privileged": false,
                      "readOnlyRootFilesystem": true,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "platforms": {
                      "openshift": {
                        "enabled": true,
                        "route": {
                          "enabled": false,
                          "annotations": {}
                        },
                        "scc": {
                          "nodeExporter": false,
                          "networkCosts": false
                        },
                        "securityContext": {
                          "runAsNonRoot": true,
                          "seccompProfile": {
                            "type": "RuntimeDefault"
                          }
                        }
                      },
                      "cicd": {
                        "enabled": true,
                        "skipSanityChecks": true
                      }
                    }
                  },
                  "upgrade": {
                    "toV2": false
                  },
                  "kubecostToken": null,
                  "pricingCsv": {
                    "enabled": false,
                    "location": {
                      "provider": "AWS",
                      "region": "us-east-1",
                      "URI": "s3://kc-csv-test/pricing_schema.csv",
                      "csvAccessCredentials": "pricing-schema-access-secret"
                    }
                  },
                  "saml": {
                    "enabled": false,
                    "rbac": {
                      "enabled": false
                    }
                  },
                  "oidc": {
                    "enabled": false,
                    "clientID": "",
                    "clientSecret": "",
                    "existingCustomSecret": {
                      "enabled": false,
                      "name": ""
                    },
                    "skipOnlineTokenValidation": false,
                    "rbac": {
                      "enabled": false
                    }
                  },
                  "systemProxy": {
                    "enabled": false,
                    "httpProxyUrl": "",
                    "httpsProxyUrl": "",
                    "noProxy": ""
                  },
                  "kubecostFrontend": {
                    "enabled": true,
                    "deployMethod": "singlepod",
                    "haReplicas": 2,
                    "image": "gcr.io/kubecost1/frontend",
                    "imagePullPolicy": "Always",
                    "resources": {
                      "requests": {
                        "cpu": "10m",
                        "memory": "55Mi"
                      }
                    },
                    "deploymentStrategy": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 1,
                      "periodSeconds": 5,
                      "failureThreshold": 6
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 1,
                      "periodSeconds": 5,
                      "failureThreshold": 6
                    },
                    "ipv6": {
                      "enabled": true
                    },
                    "useDefaultFqdn": false
                  },
                  "kubecostMetrics": {
                    "exporter": {
                      "enabled": false,
                      "port": 9005,
                      "prometheusScrape": true,
                      "resources": {},
                      "tolerations": [],
                      "affinity": {},
                      "service": {
                        "annotations": {}
                      },
                      "serviceMonitor": {
                        "enabled": false,
                        "additionalLabels": {},
                        "metricRelabelings": [],
                        "relabelings": []
                      },
                      "priorityClassName": "",
                      "additionalLabels": {},
                      "nodeSelector": {},
                      "extraArgs": []
                    }
                  },
                  "sigV4Proxy": {
                    "image": "public.ecr.aws/aws-observability/aws-sigv4-proxy:latest",
                    "imagePullPolicy": "Always",
                    "name": "aps",
                    "port": 8005,
                    "region": "us-west-2",
                    "host": "aps-workspaces.us-west-2.amazonaws.com",
                    "extraEnv": null,
                    "resources": {}
                  },
                  "kubecostModel": {
                    "image": "gcr.io/kubecost1/cost-model",
                    "imagePullPolicy": "Always",
                    "outOfClusterPromMetricsEnabled": false,
                    "warmCache": false,
                    "etl": true,
                    "etlFileStoreEnabled": true,
                    "etlDailyStoreDurationDays": 91,
                    "etlHourlyStoreDurationHours": 49,
                    "etlWeeklyStoreDurationWeeks": 53,
                    "etlReadOnlyMode": false,
                    "plugins": {
                      "enabled": false,
                      "install": {
                        "enabled": false,
                        "fullImageName": "curlimages/curl:latest",
                        "securityContext": {
                          "allowPrivilegeEscalation": false,
                          "seccompProfile": {
                            "type": "RuntimeDefault"
                          },
                          "capabilities": {
                            "drop": [
                              "ALL"
                            ]
                          },
                          "readOnlyRootFilesystem": true,
                          "runAsNonRoot": true,
                          "runAsUser": 1001
                        }
                      },
                      "folder": "/opt/opencost/plugin",
                      "enabledPlugins": [],
                      "configSecret": "kubecost-plugin-secret"
                    },
                    "cloudCost": {
                      "labelList": {
                        "IsIncludeList": false,
                        "labels": ""
                      },
                      "topNItems": 1000
                    },
                    "allocation": null,
                    "containerStatsEnabled": true,
                    "maxQueryConcurrency": 5,
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "55Mi"
                      }
                    },
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "extraArgs": [],
                    "ingress": {
                      "enabled": false,
                      "labels": null,
                      "annotations": null,
                      "paths": [
                        "/"
                      ],
                      "pathType": "ImplementationSpecific",
                      "hosts": [
                        "cost-analyzer-model.local"
                      ],
                      "tls": []
                    },
                    "utcOffset": "+00:00",
                    "extraPorts": []
                  },
                  "etlUtils": {
                    "enabled": false,
                    "fullImageName": null,
                    "resources": {},
                    "env": {},
                    "nodeSelector": {},
                    "tolerations": {},
                    "affinity": {}
                  },
                  "ingress": {
                    "enabled": false,
                    "labels": null,
                    "annotations": null,
                    "paths": [
                      "/"
                    ],
                    "pathType": "ImplementationSpecific",
                    "hosts": [
                      "cost-analyzer.local"
                    ],
                    "tls": []
                  },
                  "nodeSelector": {},
                  "tolerations": [],
                  "affinity": {},
                  "topologySpreadConstraints": [],
                  "priority": {
                    "enabled": false,
                    "name": ""
                  },
                  "networkPolicy": {
                    "enabled": false,
                    "denyEgress": true,
                    "sameNamespace": true,
                    "costAnalyzer": {
                      "enabled": false,
                      "annotations": {},
                      "additionalLabels": {}
                    }
                  },
                  "extraVolumes": [],
                  "extraVolumeMounts": [],
                  "persistentVolume": {
                    "size": "32Gi",
                    "dbSize": "32.0Gi",
                    "enabled": true,
                    "labels": {},
                    "annotations": {},
                    "dbPVEnabled": false
                  },
                  "service": {
                    "type": "ClusterIP",
                    "port": 9090,
                    "targetPort": 9090,
                    "nodePort": {},
                    "labels": {},
                    "annotations": {},
                    "sessionAffinity": {
                      "enabled": false,
                      "timeoutSeconds": 10800
                    }
                  },
                  "prometheus": {
                    "rbac": {
                      "create": true
                    },
                    "serviceAccounts": {
                      "alertmanager": {
                        "create": true,
                        "name": null
                      },
                      "nodeExporter": {
                        "create": true,
                        "name": null
                      },
                      "pushgateway": {
                        "create": true,
                        "name": null
                      },
                      "server": {
                        "create": true,
                        "name": null,
                        "annotations": {}
                      }
                    },
                    "imagePullSecrets": null,
                    "extraScrapeConfigs": "- job_name: kubecost\n  honor_labels: true\n  scrape_interval: 1m\n  scrape_timeout: 60s\n  metrics_path: /metrics\n  scheme: http\n  dns_sd_configs:\n  - names:\n    - {{ template \"cost-analyzer.serviceName\" . }}\n    type: 'A'\n    port: 9003\n- job_name: kubecost-networking\n  kubernetes_sd_configs:\n    - role: pod\n  relabel_configs:\n  # Scrape only the the targets matching the following metadata\n    - source_labels: [__meta_kubernetes_pod_label_app_kubernetes_io_instance]\n      action: keep\n      regex:  kubecost\n    - source_labels: [__meta_kubernetes_pod_label_app_kubernetes_io_name]\n      action: keep\n      regex:  network-costs\n",
                    "server": {
                      "enabled": true,
                      "name": "server",
                      "sidecarContainers": null,
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "image": {
                        "repository": "quay.io/prometheus/prometheus",
                        "tag": "v2.50.1",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "prefixURL": "",
                      "baseURL": "",
                      "env": [],
                      "extraFlags": [
                        "web.enable-lifecycle"
                      ],
                      "configPath": "/etc/config/prometheus.yml",
                      "global": {
                        "scrape_interval": "1m",
                        "scrape_timeout": "60s",
                        "evaluation_interval": "1m",
                        "external_labels": {
                          "cluster_id": "cluster-one"
                        }
                      },
                      "remoteWrite": {},
                      "remoteRead": {},
                      "extraArgs": {
                        "query.max-concurrency": 1,
                        "query.max-samples": 100000000
                      },
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "extraHostPathMounts": [],
                      "extraConfigmapMounts": [],
                      "extraSecretMounts": [],
                      "configMapOverrideName": "",
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "extraLabels": {},
                        "hosts": [],
                        "pathType": "Prefix",
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "affinity": {},
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "32Gi",
                        "subPath": ""
                      },
                      "emptyDir": {
                        "sizeLimit": ""
                      },
                      "podAnnotations": {},
                      "deploymentAnnotations": {},
                      "podLabels": {},
                      "alertmanagers": [],
                      "replicaCount": 1,
                      "statefulSet": {
                        "enabled": false,
                        "annotations": {},
                        "labels": {},
                        "podManagementPolicy": "OrderedReady",
                        "headless": {
                          "annotations": {},
                          "labels": {},
                          "servicePort": 80
                        }
                      },
                      "readinessProbeInitialDelay": 5,
                      "readinessProbeTimeout": 3,
                      "readinessProbeFailureThreshold": 3,
                      "readinessProbeSuccessThreshold": 1,
                      "livenessProbeInitialDelay": 5,
                      "livenessProbeTimeout": 3,
                      "livenessProbeFailureThreshold": 3,
                      "livenessProbeSuccessThreshold": 1,
                      "resources": {},
                      "verticalAutoscaler": {
                        "enabled": false
                      },
                      "securityContext": {},
                      "containerSecurityContext": {},
                      "service": {
                        "annotations": {},
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 80,
                        "sessionAffinity": "None",
                        "type": "ClusterIP",
                        "gRPC": {
                          "enabled": false,
                          "servicePort": 10901
                        },
                        "statefulsetReplica": {
                          "enabled": false,
                          "replica": 0
                        }
                      },
                      "terminationGracePeriodSeconds": 300,
                      "retention": "97h"
                    },
                    "alertmanager": {
                      "enabled": false,
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "name": "alertmanager",
                      "image": {
                        "repository": "quay.io/prometheus/alertmanager",
                        "tag": "v0.27.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "extraArgs": {},
                      "prefixURL": "",
                      "baseURL": "http://localhost:9093",
                      "extraEnv": {},
                      "extraSecretMounts": [],
                      "configMapOverrideName": "",
                      "configFromSecret": "",
                      "configFileName": "alertmanager.yml",
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "extraLabels": {},
                        "hosts": [],
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "affinity": {},
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "2Gi",
                        "subPath": ""
                      },
                      "podAnnotations": {},
                      "podLabels": {},
                      "replicaCount": 1,
                      "statefulSet": {
                        "enabled": false,
                        "podManagementPolicy": "OrderedReady",
                        "headless": {
                          "annotations": {},
                          "labels": {},
                          "servicePort": 80
                        }
                      },
                      "resources": {},
                      "securityContext": {
                        "runAsUser": 1001,
                        "runAsNonRoot": true,
                        "runAsGroup": 1001,
                        "fsGroup": 1001
                      },
                      "service": {
                        "annotations": {},
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 80,
                        "sessionAffinity": "None",
                        "type": "ClusterIP"
                      }
                    },
                    "alertmanagerFiles": {
                      "alertmanager.yml": {
                        "global": {},
                        "receivers": [
                          {
                            "name": "default-receiver"
                          }
                        ],
                        "route": {
                          "group_wait": "10s",
                          "group_interval": "5m",
                          "receiver": "default-receiver",
                          "repeat_interval": "3h"
                        }
                      }
                    },
                    "configmapReload": {
                      "prometheus": {
                        "enabled": false,
                        "name": "configmap-reload",
                        "image": {
                          "repository": "quay.io/prometheus-operator/prometheus-config-reloader",
                          "tag": "v0.72.0",
                          "pullPolicy": "IfNotPresent"
                        },
                        "extraArgs": {},
                        "extraVolumeDirs": [],
                        "extraConfigmapMounts": [],
                        "resources": {},
                        "containerSecurityContext": {}
                      },
                      "alertmanager": {
                        "enabled": false,
                        "name": "configmap-reload",
                        "image": {
                          "repository": "quay.io/prometheus-operator/prometheus-config-reloader",
                          "tag": "v0.72.0",
                          "pullPolicy": "IfNotPresent"
                        },
                        "extraArgs": {},
                        "extraVolumeDirs": [],
                        "extraConfigmapMounts": [],
                        "resources": {}
                      }
                    },
                    "nodeExporter": {
                      "enabled": false,
                      "hostNetwork": true,
                      "hostPID": true,
                      "dnsPolicy": "ClusterFirstWithHostNet",
                      "name": "node-exporter",
                      "image": {
                        "repository": "prom/node-exporter",
                        "tag": "v1.7.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "extraArgs": {},
                      "extraHostPathMounts": [],
                      "extraConfigmapMounts": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "deploymentAnnotations": {},
                      "pod": {
                        "labels": {}
                      },
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "resources": {},
                      "securityContext": {},
                      "service": {
                        "annotations": {
                          "prometheus.io/scrape": "true"
                        },
                        "labels": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "hostPort": 9100,
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 9100,
                        "type": "ClusterIP"
                      }
                    },
                    "pushgateway": {
                      "enabled": false,
                      "name": "pushgateway",
                      "image": {
                        "repository": "prom/pushgateway",
                        "tag": "v1.6.2",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "extraArgs": {},
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "hosts": [],
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "replicaCount": 1,
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "resources": {},
                      "securityContext": {
                        "runAsUser": 1001,
                        "runAsNonRoot": true
                      },
                      "service": {
                        "annotations": {
                          "prometheus.io/probe": "pushgateway"
                        },
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 9091,
                        "type": "ClusterIP"
                      },
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "2Gi",
                        "subPath": ""
                      }
                    },
                    "serverFiles": {
                      "alerting_rules.yml": {},
                      "alerts": {},
                      "recording_rules.yml": {},
                      "prometheus.yml": {
                        "rule_files": [
                          "/etc/config/recording_rules.yml",
                          "/etc/config/alerting_rules.yml",
                          "/etc/config/rules",
                          "/etc/config/alerts"
                        ],
                        "scrape_configs": [
                          {
                            "job_name": "prometheus",
                            "static_configs": [
                              {
                                "targets": [
                                  "localhost:9090"
                                ]
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-nodes-cadvisor",
                            "scheme": "https",
                            "tls_config": {
                              "ca_file": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                              "insecure_skip_verify": true
                            },
                            "bearer_token_file": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                            "kubernetes_sd_configs": [
                              {
                                "role": "node"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_node_label_(.+)"
                              },
                              {
                                "target_label": "__address__",
                                "replacement": "kubernetes.default.svc:443"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_node_name"
                                ],
                                "regex": "(.+)",
                                "target_label": "__metrics_path__",
                                "replacement": "/api/v1/nodes/$1/proxy/metrics/cadvisor"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(container_cpu_usage_seconds_total|container_memory_working_set_bytes|container_network_receive_errors_total|container_network_transmit_errors_total|container_network_receive_packets_dropped_total|container_network_transmit_packets_dropped_total|container_memory_usage_bytes|container_cpu_cfs_throttled_periods_total|container_cpu_cfs_periods_total|container_fs_usage_bytes|container_fs_limit_bytes|container_cpu_cfs_periods_total|container_fs_inodes_free|container_fs_inodes_total|container_fs_usage_bytes|container_fs_limit_bytes|container_cpu_cfs_throttled_periods_total|container_cpu_cfs_periods_total|container_network_receive_bytes_total|container_network_transmit_bytes_total|container_fs_inodes_free|container_fs_inodes_total|container_fs_usage_bytes|container_fs_limit_bytes|container_spec_cpu_shares|container_spec_memory_limit_bytes|container_network_receive_bytes_total|container_network_transmit_bytes_total|container_fs_reads_bytes_total|container_network_receive_bytes_total|container_fs_writes_bytes_total|container_fs_reads_bytes_total|cadvisor_version_info|kubecost_pv_info)",
                                "action": "keep"
                              },
                              {
                                "source_labels": [
                                  "container"
                                ],
                                "target_label": "container_name",
                                "regex": "(.+)",
                                "action": "replace"
                              },
                              {
                                "source_labels": [
                                  "pod"
                                ],
                                "target_label": "pod_name",
                                "regex": "(.+)",
                                "action": "replace"
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-nodes",
                            "scheme": "https",
                            "tls_config": {
                              "ca_file": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                              "insecure_skip_verify": true
                            },
                            "bearer_token_file": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                            "kubernetes_sd_configs": [
                              {
                                "role": "node"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_node_label_(.+)"
                              },
                              {
                                "target_label": "__address__",
                                "replacement": "kubernetes.default.svc:443"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_node_name"
                                ],
                                "regex": "(.+)",
                                "target_label": "__metrics_path__",
                                "replacement": "/api/v1/nodes/$1/proxy/metrics"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(kubelet_volume_stats_used_bytes)",
                                "action": "keep"
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-service-endpoints",
                            "kubernetes_sd_configs": [
                              {
                                "role": "endpoints"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_scrape"
                                ],
                                "action": "keep",
                                "regex": true
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_endpoints_name"
                                ],
                                "action": "keep",
                                "regex": "(.*node-exporter|kubecost-network-costs)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_scheme"
                                ],
                                "action": "replace",
                                "target_label": "__scheme__",
                                "regex": "(https?)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_path"
                                ],
                                "action": "replace",
                                "target_label": "__metrics_path__",
                                "regex": "(.+)"
                              },
                              {
                                "source_labels": [
                                  "__address__",
                                  "__meta_kubernetes_service_annotation_prometheus_io_port"
                                ],
                                "action": "replace",
                                "target_label": "__address__",
                                "regex": "([^:]+)(?::\\d+)?;(\\d+)",
                                "replacement": "$1:$2"
                              },
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_service_label_(.+)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_namespace"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_namespace"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_name"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_name"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_pod_node_name"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_node"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(container_cpu_allocation|container_cpu_usage_seconds_total|container_fs_limit_bytes|container_fs_writes_bytes_total|container_gpu_allocation|container_memory_allocation_bytes|container_memory_usage_bytes|container_memory_working_set_bytes|container_network_receive_bytes_total|container_network_transmit_bytes_total|DCGM_FI_DEV_GPU_UTIL|deployment_match_labels|kube_daemonset_status_desired_number_scheduled|kube_daemonset_status_number_ready|kube_deployment_spec_replicas|kube_deployment_status_replicas|kube_deployment_status_replicas_available|kube_job_status_failed|kube_namespace_annotations|kube_namespace_labels|kube_node_info|kube_node_labels|kube_node_status_allocatable|kube_node_status_allocatable_cpu_cores|kube_node_status_allocatable_memory_bytes|kube_node_status_capacity|kube_node_status_capacity_cpu_cores|kube_node_status_capacity_memory_bytes|kube_node_status_condition|kube_persistentvolume_capacity_bytes|kube_persistentvolume_status_phase|kube_persistentvolumeclaim_info|kube_persistentvolumeclaim_resource_requests_storage_bytes|kube_pod_container_info|kube_pod_container_resource_limits|kube_pod_container_resource_limits_cpu_cores|kube_pod_container_resource_limits_memory_bytes|kube_pod_container_resource_requests|kube_pod_container_resource_requests_cpu_cores|kube_pod_container_resource_requests_memory_bytes|kube_pod_container_status_restarts_total|kube_pod_container_status_running|kube_pod_container_status_terminated_reason|kube_pod_labels|kube_pod_owner|kube_pod_status_phase|kube_replicaset_owner|kube_statefulset_replicas|kube_statefulset_status_replicas|kubecost_cluster_info|kubecost_cluster_management_cost|kubecost_cluster_memory_working_set_bytes|kubecost_load_balancer_cost|kubecost_network_internet_egress_cost|kubecost_network_region_egress_cost|kubecost_network_zone_egress_cost|kubecost_node_is_spot|kubecost_pod_network_egress_bytes_total|node_cpu_hourly_cost|node_cpu_seconds_total|node_disk_reads_completed|node_disk_reads_completed_total|node_disk_writes_completed|node_disk_writes_completed_total|node_filesystem_device_error|node_gpu_count|node_gpu_hourly_cost|node_memory_Buffers_bytes|node_memory_Cached_bytes|node_memory_MemAvailable_bytes|node_memory_MemFree_bytes|node_memory_MemTotal_bytes|node_network_transmit_bytes_total|node_ram_hourly_cost|node_total_hourly_cost|pod_pvc_allocation|pv_hourly_cost|service_selector_labels|statefulSet_match_labels|kubecost_pv_info|up)",
                                "action": "keep"
                              }
                            ]
                          }
                        ]
                      },
                      "rules": {
                        "groups": [
                          {
                            "name": "CPU",
                            "rules": [
                              {
                                "expr": "sum(rate(container_cpu_usage_seconds_total{container!=\"\"}[5m]))",
                                "record": "cluster:cpu_usage:rate5m"
                              },
                              {
                                "expr": "rate(container_cpu_usage_seconds_total{container!=\"\"}[5m])",
                                "record": "cluster:cpu_usage_nosum:rate5m"
                              },
                              {
                                "expr": "avg(irate(container_cpu_usage_seconds_total{container!=\"POD\", container!=\"\"}[5m])) by (container,pod,namespace)",
                                "record": "kubecost_container_cpu_usage_irate"
                              },
                              {
                                "expr": "sum(container_memory_working_set_bytes{container!=\"POD\",container!=\"\"}) by (container,pod,namespace)",
                                "record": "kubecost_container_memory_working_set_bytes"
                              },
                              {
                                "expr": "sum(container_memory_working_set_bytes{container!=\"POD\",container!=\"\"})",
                                "record": "kubecost_cluster_memory_working_set_bytes"
                              }
                            ]
                          },
                          {
                            "name": "Savings",
                            "rules": [
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind!=\"DaemonSet\"}) by (pod) * sum(container_cpu_allocation) by (pod))",
                                "record": "kubecost_savings_cpu_allocation",
                                "labels": {
                                  "daemonset": "false"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind=\"DaemonSet\"}) by (pod) * sum(container_cpu_allocation) by (pod)) / sum(kube_node_info)",
                                "record": "kubecost_savings_cpu_allocation",
                                "labels": {
                                  "daemonset": "true"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind!=\"DaemonSet\"}) by (pod) * sum(container_memory_allocation_bytes) by (pod))",
                                "record": "kubecost_savings_memory_allocation_bytes",
                                "labels": {
                                  "daemonset": "false"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind=\"DaemonSet\"}) by (pod) * sum(container_memory_allocation_bytes) by (pod)) / sum(kube_node_info)",
                                "record": "kubecost_savings_memory_allocation_bytes",
                                "labels": {
                                  "daemonset": "true"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    "alertRelabelConfigs": null,
                    "networkPolicy": {
                      "enabled": false
                    }
                  },
                  "networkCosts": {
                    "enabled": false,
                    "image": {
                      "repository": "gcr.io/kubecost1/kubecost-network-costs",
                      "tag": "v0.17.3"
                    },
                    "imagePullPolicy": "Always",
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "prometheusScrape": false,
                    "trafficLogging": true,
                    "logLevel": "info",
                    "port": 3001,
                    "resources": {
                      "limits": {
                        "cpu": "500m"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "20Mi"
                      }
                    },
                    "extraArgs": [],
                    "config": {
                      "destinations": {
                        "in-zone": [
                          "127.0.0.0/8",
                          "169.254.0.0/16",
                          "10.0.0.0/8",
                          "172.16.0.0/12",
                          "192.168.0.0/16"
                        ],
                        "in-region": [],
                        "cross-region": [],
                        "internet": [],
                        "direct-classification": []
                      },
                      "services": {
                        "google-cloud-services": false,
                        "amazon-web-services": false,
                        "azure-cloud-services": false
                      }
                    },
                    "tolerations": [],
                    "affinity": {},
                    "service": {
                      "annotations": {},
                      "labels": {}
                    },
                    "priorityClassName": "",
                    "podMonitor": {
                      "enabled": false,
                      "additionalLabels": {}
                    },
                    "additionalLabels": {},
                    "nodeSelector": {},
                    "annotations": {},
                    "healthCheckProbes": {},
                    "additionalSecurityContext": {}
                  },
                  "kubecostDeployment": {
                    "replicas": 1,
                    "labels": {},
                    "annotations": {}
                  },
                  "forecasting": {
                    "enabled": true,
                    "fullImageName": "gcr.io/kubecost1/kubecost-modeling:v0.1.6",
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "limits": {
                        "cpu": "1500m",
                        "memory": "1Gi"
                      }
                    },
                    "env": {
                      "GUNICORN_CMD_ARGS": "--log-level info -t 1200"
                    },
                    "priority": {
                      "enabled": false,
                      "name": ""
                    },
                    "nodeSelector": {},
                    "tolerations": {},
                    "affinity": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    }
                  },
                  "kubecostAggregator": {
                    "deployMethod": "singlepod",
                    "enabled": false,
                    "replicas": 1,
                    "stagingEmptyDirSizeLimit": "2Gi",
                    "numDBCopyPartitions": 25,
                    "env": {
                      "LOG_LEVEL": "info",
                      "DB_READ_THREADS": "1",
                      "DB_WRITE_THREADS": "1",
                      "DB_CONCURRENT_INGESTION_COUNT": "3"
                    },
                    "persistentConfigsStorage": {
                      "storageClass": "",
                      "storageRequest": "1Gi"
                    },
                    "aggregatorDbStorage": {
                      "storageClass": "",
                      "storageRequest": "128Gi"
                    },
                    "resources": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "cloudCost": {
                      "enabled": false,
                      "resources": {},
                      "readinessProbe": {
                        "enabled": true,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 10,
                        "failureThreshold": 200
                      }
                    },
                    "jaeger": {
                      "enabled": false,
                      "image": "jaegertracing/all-in-one",
                      "imageVersion": "latest"
                    }
                  },
                  "diagnostics": {
                    "enabled": true,
                    "primary": {
                      "enabled": false,
                      "retention": "7d",
                      "readonly": false
                    },
                    "pollingInterval": "300s",
                    "keepDiagnosticHistory": false,
                    "collectHelmValues": false,
                    "deployment": {
                      "enabled": false,
                      "resources": {
                        "requests": {
                          "cpu": "10m",
                          "memory": "20Mi"
                        }
                      },
                      "env": {},
                      "labels": {},
                      "securityContext": {},
                      "containerSecurityContext": {},
                      "nodeSelector": {},
                      "tolerations": {},
                      "affinity": {}
                    }
                  },
                  "clusterController": {
                    "enabled": false,
                    "image": {
                      "repository": "gcr.io/kubecost1/cluster-controller",
                      "tag": "v0.16.0"
                    },
                    "imagePullPolicy": "Always",
                    "priorityClassName": "",
                    "tolerations": [],
                    "actionConfigs": {
                      "clusterTurndown": [],
                      "namespaceTurndown": null,
                      "clusterRightsize": null,
                      "containerRightsize": null
                    },
                    "kubescaler": {
                      "defaultResizeAll": false
                    },
                    "namespaceTurndown": {
                      "rbac": {
                        "enabled": true
                      }
                    }
                  },
                  "reporting": {
                    "logCollection": true,
                    "productAnalytics": true,
                    "errorReporting": true,
                    "valuesReporting": true
                  },
                  "serviceMonitor": {
                    "enabled": false,
                    "additionalLabels": {},
                    "metricRelabelings": [],
                    "relabelings": [],
                    "networkCosts": {
                      "enabled": false,
                      "scrapeTimeout": "10s",
                      "additionalLabels": {},
                      "metricRelabelings": [],
                      "relabelings": []
                    }
                  },
                  "prometheusRule": {
                    "enabled": false,
                    "additionalLabels": {}
                  },
                  "supportNFS": false,
                  "initChownDataImage": "busybox",
                  "initChownData": {
                    "resources": {}
                  },
                  "grafana": {
                    "rbac": {
                      "create": true
                    },
                    "serviceAccount": {
                      "create": true,
                      "name": ""
                    },
                    "replicas": 1,
                    "deploymentStrategy": "RollingUpdate",
                    "readinessProbe": {
                      "httpGet": {
                        "path": "/api/health",
                        "port": 3000
                      }
                    },
                    "livenessProbe": {
                      "httpGet": {
                        "path": "/api/health",
                        "port": 3000
                      },
                      "initialDelaySeconds": 60,
                      "timeoutSeconds": 30,
                      "failureThreshold": 10
                    },
                    "image": {
                      "repository": "grafana/grafana",
                      "tag": "10.3.4",
                      "pullPolicy": "IfNotPresent"
                    },
                    "securityContext": {},
                    "priorityClassName": "",
                    "downloadDashboardsImage": {
                      "repository": "curlimages/curl",
                      "tag": "latest",
                      "pullPolicy": "IfNotPresent"
                    },
                    "podAnnotations": {},
                    "annotations": {},
                    "service": {
                      "type": "ClusterIP",
                      "port": 80,
                      "annotations": {},
                      "labels": {}
                    },
                    "ingress": {
                      "enabled": false,
                      "annotations": {},
                      "labels": {},
                      "path": "/",
                      "pathType": "Prefix",
                      "hosts": [
                        "chart-example.local"
                      ],
                      "tls": []
                    },
                    "resources": {},
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "persistence": {
                      "enabled": false
                    },
                    "adminUser": "admin",
                    "adminPassword": "strongpassword",
                    "env": {},
                    "envFromSecret": "",
                    "extraSecretMounts": [],
                    "plugins": [],
                    "dashboardProviders": {},
                    "dashboards": {},
                    "dashboardsConfigMaps": {},
                    "ldap": {
                      "existingSecret": "",
                      "config": ""
                    },
                    "smtp": {
                      "existingSecret": ""
                    },
                    "sidecar": {
                      "image": {
                        "repository": "kiwigrid/k8s-sidecar",
                        "tag": "1.26.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "resources": {},
                      "dashboards": {
                        "enabled": true,
                        "label": "grafana_dashboard",
                        "labelValue": "1",
                        "annotations": {},
                        "error_throttle_sleep": 0,
                        "folder": "/tmp/dashboards"
                      },
                      "datasources": {
                        "enabled": false,
                        "error_throttle_sleep": 0,
                        "label": "grafana_datasource"
                      }
                    },
                    "grafana.ini": {
                      "server": {
                        "serve_from_sub_path": false,
                        "root_url": "%(protocol)s://%(domain)s:%(http_port)s/grafana"
                      },
                      "paths": {
                        "data": "/var/lib/grafana/data",
                        "logs": "/var/log/grafana",
                        "plugins": "/var/lib/grafana/plugins",
                        "provisioning": "/etc/grafana/provisioning"
                      },
                      "analytics": {
                        "check_for_updates": true
                      },
                      "log": {
                        "mode": "console"
                      },
                      "grafana_net": {
                        "url": "https://grafana.net"
                      },
                      "auth.anonymous": {
                        "enabled": true,
                        "org_role": "Editor",
                        "org_name": "Main Org."
                      }
                    }
                  },
                  "serviceAccount": {
                    "create": true,
                    "annotations": {}
                  },
                  "awsstore": {
                    "useAwsStore": false,
                    "imageNameAndVersion": "gcr.io/kubecost1/awsstore:latest",
                    "createServiceAccount": false,
                    "priorityClassName": "",
                    "nodeSelector": {},
                    "annotations": {}
                  },
                  "federatedETL": {
                    "agentOnly": false,
                    "federatedCluster": false,
                    "redirectS3Backup": false,
                    "useMultiClusterDB": false
                  },
                  "kubecostAdmissionController": {
                    "enabled": false,
                    "secretName": "webhook-server-tls",
                    "caBundle": "${CA_BUNDLE}"
                  },
                  "costEventsAudit": {
                    "enabled": false
                  },
                  "extraObjects": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: gcr.io/kubecost1/operator/kubecost-operator-controller:v2.2.4
          createdAt: "2024-05-19T20:46:35Z"
          description: An operator that deploys Kubecost 2.2.4 via the Helm chart.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Stackwatch
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The CostAnalyzer API describes a deployment of Kubecost and
              its dependencies.
            displayName: CostAnalyzer
            kind: CostAnalyzer
            name: costanalyzers.charts.kubecost.com
            version: v1alpha1
        description: |-
          Kubecost provides real-time cost visibility and insights for teams using Kubernetes, helping you continuously reduce your cloud costs.

          This OpenShift operator deploys Kubecost from the official Helm chart.

          Kubecost allows you to:

          * See allocated spend across all native Kubernetes concepts, so you can provide your teams with transparent, accurate cost data reconciled with your actual cloud bill.
          * Join in-cluster costs like CPU and memory with out-of-cluster spend from cloud infrastructure services for complete cost visibility—all services from AWS, GCP, and Microsoft Azure are supported.
          * Gain insights you can use to save 30-50% or more on your infrastructure spend, without exposing your private information—your data never gets shared externally, even with us.
          * Have real-time alerts and recurring reports, empowering teams to take control of their Kubernetes-enabled infrastructure, stay within budgeted limits, and address monitoring interruptions immediately.

          Visit the Kubecost website for more details including features, pricing, documentation, and more at www.kubecost.com.
        displayName: Kubecost
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubecost
        - cost
        - kubernetes
        links:
        - name: Kubecost
          url: https://www.kubecost.com
        maintainers:
        - email: czoller@kubecost.com
          name: Chip Zoller
        maturity: alpha
        provider:
          name: Stackwatch
          url: www.kubecost.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - gcr.io/kubecost1/operator/kubecost-operator-controller:v2.2.4
        version: 2.2.4
      entries:
      - name: kubecost.v2.2.4
        version: 2.2.4
      - name: kubecost.v2.1.0
        version: 2.1.0
      - name: kubecost-operator.v1.107.1
        version: 1.107.1
      - name: kubecost-operator.v1.106.2
        version: 1.106.2
      - name: kubecost-operator.v1.106.0
        version: 1.106.0
      - name: kubecost.v1.99.0
        version: 1.99.0
      - name: kubecost.v1.97.0
        version: 1.97.0
      name: alpha
    defaultChannel: alpha
    packageName: kubecost
    provider:
      name: Stackwatch
      url: www.kubecost.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: prometurbo
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometurbo-operator.v8.7.1-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.7.1-SNAPSHOT
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.7.1-SNAPSHOT
        version: 8.7.1-beta.1
      entries:
      - name: prometurbo-operator.v8.7.1-beta.1
        version: 8.7.1-beta.1
      name: beta
    - currentCSV: prometurbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.6
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.6
        version: 8.6.0
      entries:
      - name: prometurbo-operator.v8.6.0
        version: 8.6.0
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: stable
    - currentCSV: prometurbo-operator.v8.5.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.5
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.5
        version: 8.5.0
      entries:
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: v8.5.0
    - currentCSV: prometurbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.6
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.6
        version: 8.6.0
      entries:
      - name: prometurbo-operator.v8.6.0
        version: 8.6.0
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: v8.6.0
    defaultChannel: stable
    packageName: prometurbo
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apache KIE
      provider-url: ""
    name: sonataflow-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sonataflow-operator.v10.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "docker.io/apache",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: docker.io/apache/incubator-kie-sonataflow-operator:10.0.0
          description: SonataFlow Kubernetes Operator for deploying workflow applications
            based on the CNCF Serverless Workflow specification
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/apache/incubator-kie-tools/packages/sonataflow-operator
          support: Apache Software Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: Sonata Flow Build
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: Sonata Flow Cluster Platform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: Sonata Flow Platform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: Sonata Flow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          SonataFlow Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification](https://serverlessworkflow.io/):

          * Deploy workflow applications using the [dev profile](https://sonataflow.org/serverlessworkflow/latest/cloud/operator/developing-workflows.html), suited for the your development cycle
          * Build workflow applications based on the platform you're currently working on.
        displayName: SonataFlow Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        links:
        - name: Product Page
          url: https://github.com/apache/incubator-kie-tools
        maintainers:
        - email: dev@kie.apache.org
          name: Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Apache Software Foundation
        relatedImages:
        - docker.io/apache/incubator-kie-sonataflow-operator:10.0.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 10.0.0
      entries:
      - name: sonataflow-operator.v10.0.0
        version: 10.0.0
      - name: sonataflow-operator.v1.40.1
        version: 1.40.1
      - name: sonataflow-operator.v1.40.0
        version: 1.40.0
      name: alpha
    - currentCSV: sonataflow-operator.v10.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "host/namespace",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: docker.io/apache/incubator-kie-sonataflow-operator:10.1.0
          createdAt: "2025-09-11T14:44:56Z"
          description: SonataFlow Kubernetes Operator for deploying workflow applications
            based on the CNCF Serverless Workflow specification
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/apache/incubator-kie-tools/packages/sonataflow-operator
          support: Apache KIE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: SonataFlowBuild
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: SonataFlowClusterPlatform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: SonataFlowPlatform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: SonataFlow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          SonataFlow Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification](https://serverlessworkflow.io/):

          * Deploy workflow applications using the [dev profile](https://sonataflow.org/serverlessworkflow/latest/cloud/operator/developing-workflows.html), suited for the your development cycle
          * Build workflow applications based on the platform you're currently working on.
        displayName: SonataFlow Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        links:
        - name: Product Page
          url: https://sonataflow.org/serverlessworkflow/latest/index.html
        maintainers:
        - email: dev@kie.apache.org
          name: Apache KIE
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Apache KIE
        relatedImages:
        - docker.io/apache/incubator-kie-sonataflow-operator:10.1.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 10.1.0
      entries:
      - name: sonataflow-operator.v10.1.0
        version: 10.1.0
      - name: sonataflow-operator.v10.0.0
        version: 10.0.0
      - name: sonataflow-operator.v1.40.1
        version: 1.40.1
      - name: sonataflow-operator.v1.40.0
        version: 1.40.0
      name: stable
    defaultChannel: stable
    packageName: sonataflow-operator
    provider:
      name: Apache KIE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: directpv
      provider-url: ""
    name: private-dpv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: private-dpv-operator.v5.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.quay.io/v1alpha1",
                "kind": "AistorVolumemanager",
                "metadata": {
                  "name": "aistorvolumemanager-sample"
                },
                "spec": {
                  "annotations": {},
                  "controller": {
                    "podAnnotations": {},
                    "podLabels": {},
                    "replicas": 3
                  },
                  "images": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": []
                  },
                  "labels": {},
                  "license": "",
                  "nodeserver": {
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-10-05T15:47:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AistorVolumemanager
            name: aistorvolumemanagers.charts.quay.io
            version: v1alpha1
        description: directpv
        displayName: directpv
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - directpv
        links:
        - name: Private Dpv Operator
          url: https://private-dpv-operator.domain
        maintainers:
        - email: cesarcelis@minio.io
          name: celis
        maturity: alpha
        provider:
          name: directpv
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - quay.io/cniackz4/directpv-operator@sha256:af06e57661a7e05809e44c3fac571ed949cc04fd656fdb0c8324b533778eb604
        version: 5.0.3
      entries:
      - name: private-dpv-operator.v5.0.3
        version: 5.0.3
      - name: private-dpv-operator.v5.0.2
        version: 5.0.2
      name: alpha
    defaultChannel: alpha
    packageName: private-dpv-operator
    provider:
      name: directpv
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fortanix Inc.
      provider-url: https://fortanix.com
    name: ccm-node-agent-dcap-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ccm-node-agent-dcap-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1",
                "kind": "CcmNodeAgent",
                "metadata": {
                  "name": "ccmnodeagent-sample"
                },
                "spec": {
                  "ccmJoinToken": {
                    "secretName": "ccm-join-token",
                    "token": ""
                  },
                  "fullnameOverride": "",
                  "image": {
                    "aesmd": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/aesmd-dcap",
                      "tag": "latest"
                    },
                    "ccmNodeAgent": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/ccm-node-agent-ubi7",
                      "tag": "latest"
                    },
                    "ccmNodeAgentInit": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "busybox",
                      "tag": "latest"
                    },
                    "multiPlatformRegistration": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/multi-platform-registration-ubi8",
                      "tag": "latest"
                    }
                  },
                  "log": {
                    "debug": false
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "ports": {
                    "containerPort": 9092,
                    "hostPort": 9092
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CCM Node Agent software enables registration of the compute
              nodes to Fortanix CCM when installed in the Openshift Cluster Nodes<U+200B>.
            displayName: CCM Node Agent
            kind: CcmNodeAgent
            name: ccmnodeagents.charts.operatorhub.io
            version: v1
        description: "CCM Node Agent software enables registration of the compute
          nodes to Fortanix CCM when installed in the Openshift Cluster Nodes​. \nIt
          also assists with application attestation and visibility for Fortanix CCM.\n\n##
          About this Operator\nThis Operator is based on a Helm chart for CCM Node
          Agent. It enrolls all the nodes in the openshift cluster to a particular
          CCM account. You have to provide the base64 encode of join token of CCM
          account while creating CcmNodeAgent resource.\n\n## Prerequisites\nAll the
          worker nodes in the cluster must have SGX capabilities, in-kernel driver
          installed and fsgsbase support.\n\n## Installation\n * Choose a namespace
          to install the Operator. \n Operator will be available in all namespaces.\n*
          Select Approval Strategy: Available options are Automatic or Manual. Automatic
          approval will automatically upgrade the running instance of your Operator
          without human intervention when a new version is available. In case of Manual
          approval, when a new version is available, OLM will create an update request.
          You must manually approve the request to update the Operator to a new version.\n*
          Generate your CCM Account Join Token: Please log in to https://ccm.fortanix.com,
          and in the Management Console tab, click the + ENROLL NODE button. In the
          ENROLL NODE screen, a Join Token will be generated in the text box for \"Get
          a join token to register an SGX compute node\". This Join Token is used
          by the compute node to authenticate itself.\n* Base64 encode the Join Token:\n```shell\necho
          -n <join-token> | openssl base64 -A\n```\n* Create an instance of CcmNodeAgent
          by setting the `spec.ccmJoinToken.Token` to the base64 encoded CCM account
          join token.\n* After CcmNodeAgent resource is installed successfully you
          should be able to see all the worker nodes of the cluster in your CCM account
          under Compute Node Tab.\n\n## Getting Started\nNow to run your application
          inside SGX enclaves on your cluster follow the [documentation](https://support.fortanix.com/hc/en-us/articles/360043527431-User-s-Guide-Add-and-Edit-an-Application).\n"
        displayName: CCM Node Agent DCAP Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dockerfile
        - Helm
        - CCM
        - Confidential Compute
        links:
        - name: CCM Node Agent DCAP Operator
          url: https://support.fortanix.com/hc/en-us/categories/360003107511-Confidential-Computing-Manager
        maintainers:
        - email: info@fortanix.com
          name: Fortanix Inc.
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Fortanix Inc.
          url: https://fortanix.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62575404000cb77fa8d2284c87b31e39e667a55b0fb47821263b23b3d1135a7e
        - docker.io/fortanix/ccm-node-agent-dcap-operator@sha256:4fd28bea26aee461eb9c1800cdf5f041b64ae3ea6652203f8ad9feb680aaefee
        version: 0.0.2
      entries:
      - name: ccm-node-agent-dcap-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ccm-node-agent-dcap-operator
    provider:
      name: Fortanix Inc.
      url: https://fortanix.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Infinidat
      provider-url: https://infinidat.com
    name: infinibox-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infinibox-operator-certified.v2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csidriver.infinidat.com/v1alpha1",
                "kind": "InfiniboxCsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infinibox-operator-certified",
                    "app.kubernetes.io/instance": "infiniboxcsidriver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infiniboxcsidriver",
                    "app.kubernetes.io/part-of": "infinibox-operator-certified"
                  },
                  "name": "infiniboxcsidriver-sample"
                },
                "spec": {
                  "Ibox_Creds": [
                    {
                      "SecretName": "infinibox-creds",
                      "hostname": "ibox0000",
                      "inbound_secret": "0.0000000000000",
                      "inbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonout",
                      "outbound_secret": "0.0000000000001",
                      "outbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonin",
                      "password": "my-pool-admin-passwd",
                      "username": "my-pool-admin"
                    }
                  ],
                  "Infinibox_Protocol_Secrets": [
                    {
                      "SecretName": "infinibox-protocol-secret",
                      "iscsi": {
                        "network_space": "default_iscsi_space",
                        "useCHAP": "none"
                      },
                      "nfs": {
                        "network_space": "default_nas_space",
                        "nfs_export_permissions": ""
                      },
                      "nvme": {
                        "network_space": "default_nvme_space"
                      },
                      "preferred_auto_order": "fc,nvme,iscsi",
                      "storage_protocol": "fc"
                    }
                  ],
                  "autoUpdate": false,
                  "cleanupNFSPerms": false,
                  "controllerserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "controllerservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "createEvents": true,
                  "csiDriverName": "infinibox-csi-driver",
                  "csiDriverVersion": "v2.23.0",
                  "e2etesting": false,
                  "images": {
                    "attachersidecar": "registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901",
                    "attachersidecar_pull_policy": "IfNotPresent",
                    "csidriver": "registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61",
                    "csidriver_pull_policy": "Always",
                    "livenesssidecar": "registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc",
                    "livenesssidecar_pull_policy": "IfNotPresent",
                    "provisionersidecar": "registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2",
                    "provisionersidecar_pull_policy": "IfNotPresent",
                    "registrarsidecar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24",
                    "registrarsidecar_pull_policy": "IfNotPresent",
                    "resizersidecar": "registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d",
                    "resizersidecar_pull_policy": "IfNotPresent",
                    "snapshottersidecar": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a",
                    "snapshottersidecar_pull_policy": "IfNotPresent"
                  },
                  "instanceCount": 1,
                  "logLevel": "debug",
                  "nodeSelector": {
                    "kubernetes.io/os": "linux"
                  },
                  "nodeserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "nodeservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "protocolSecret": "",
                  "removeDomainName": false,
                  "replicaCount": 1,
                  "skipCredentialsCreation": false,
                  "volumeNamePrefix": "ibox"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
          createdAt: "2025-09-27T16:12:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2.23.0
          operatorframework.io/suggested-namespace: infinidat-csi
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Infinidat/infinibox-csi-driver
          support: Infinidat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Iboxreplica
            name: iboxreplicas.csidriver.infinidat.com
            version: v1
          - kind: InfiniboxCsiDriver
            name: infiniboxcsidrivers.csidriver.infinidat.com
            version: v1alpha1
        description: "Infinidat InfiniBox Container Storage Interface (CSI) Driver
          is a\n    CNCF-compliant Kubernetes integration for InfiniBox storage systems,
          offering\n    advanced enterprise functionality for petabyte-scale Kubernetes
          deployments including\n    Red Hat OpenShift.\n \n## Features and Benefits\n
          \ \n* **Multi-protocol flexibility** - manage\n    Kubernetes Persistent
          Volumes attached via block and file protocols,\n    including Fibre Channel,
          iSCSI, and NFS, with all Kubernetes PV access modes\n    \n* **Multi-petabyte
          scalability** - support hundreds of thousands of PVs per InfiniBox\n    system
          and control multiple InfiniBox arrays within a single Kubernetes cluster\n
          \   \\ \n* **Advanced enterprise features** - manage native InfiniBox snapshots
          and\n    clones, including restoring from snapshots, and import PVs created
          outside of\n    InfiniBox CSI Driver\n  \n## Required Parameters\n  \n*
          `hostname` - IP address\n    or hostname of the InfiniBox management interface\n*
          `username` / `password` -\n    InfiniBox credentials\n* `SecretName` - secret
          name, to be used in the StorageClass\n    to define a specific InfiniBox
          for persistent volumes\n\n## Optional Parameters\n*\n    `inbound_user`
          / `inbound_secret` / `outbound_user` / `outbound_secret` - credentials \n
          \   for iSCSI CHAP authentication\n\n    \n## Installation Instructions\n
          \n1. Create *infinidat-csi* namespace to install operator into.\n    \n2.
          Delete the following *clusterrolebindings*, if they exist, before installing
          operator\n     * infinidat-csi-operator-infinidat-csi-attacher\n     * infinidat-csi-operator-infinidat-csi-controller
          \                    \n     * infinidat-csi-operator-infinidat-csi-driver
          \                    \n     * infinidat-csi-operator-infinidat-csi-node
          \                              \n     * infinidat-csi-operator-infinidat-csi-provisioner
          \                       \n     * infinidat-csi-operator-infinidat-csi-resizer
          \          \n     * infinidat-csi-operator-infinidat-csi-snapshotter \n
          \   \n3. Install Infinidat Infinibox CSI Driver - Operator (this) into the
          *infinidat-csi* namespace.\n    \n4. Apply the following Security Context
          Constraints to the created service accounts in the *infinidat-csi* \n    namespace\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    \n5. Recreate or edit any existing storage classes, pointing to the
          new namespace\n\n     \n6. If you choose a different namespace other than
          *infinidat-csi* you will need to edit the clusterrolebindings listed above
          to specify your custom namespace\n\n     \n7. Install an instance of Infinidat
          Driver into the *infinidat-csi* namespace"
        displayName: InfiniBox CSI Driver - Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - csi
        - nas
        - san
        - storage
        - infinidat
        - infinibox
        links:
        - name: Infinidat
          url: https://infinidat.com
        maintainers:
        - email: info@infinidat.com
          name: Infinidat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Infinidat
          url: https://infinidat.com
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified@sha256:42ff34a019bc67076515e89735b4bb9921fbd4610c127ac8b61e1fa6a6a08194
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24
        version: 2.23.0
      entries:
      - name: infinibox-operator-certified.v2.23.0
        version: 2.23.0
      - name: infinibox-operator-certified.v2.22.0
        version: 2.22.0
      - name: infinibox-operator-certified.v2.21.0
        version: 2.21.0
      name: alpha
    - currentCSV: infinibox-operator-certified.v2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csidriver.infinidat.com/v1alpha1",
                "kind": "InfiniboxCsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infinibox-operator-certified",
                    "app.kubernetes.io/instance": "infiniboxcsidriver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infiniboxcsidriver",
                    "app.kubernetes.io/part-of": "infinibox-operator-certified"
                  },
                  "name": "infiniboxcsidriver-sample"
                },
                "spec": {
                  "Ibox_Creds": [
                    {
                      "SecretName": "infinibox-creds",
                      "hostname": "ibox0000",
                      "inbound_secret": "0.0000000000000",
                      "inbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonout",
                      "outbound_secret": "0.0000000000001",
                      "outbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonin",
                      "password": "my-pool-admin-passwd",
                      "username": "my-pool-admin"
                    }
                  ],
                  "Infinibox_Protocol_Secrets": [
                    {
                      "SecretName": "infinibox-protocol-secret",
                      "iscsi": {
                        "network_space": "default_iscsi_space",
                        "useCHAP": "none"
                      },
                      "nfs": {
                        "network_space": "default_nas_space",
                        "nfs_export_permissions": ""
                      },
                      "nvme": {
                        "network_space": "default_nvme_space"
                      },
                      "preferred_auto_order": "fc,nvme,iscsi",
                      "storage_protocol": "fc"
                    }
                  ],
                  "autoUpdate": false,
                  "cleanupNFSPerms": false,
                  "controllerserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "controllerservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "createEvents": true,
                  "csiDriverName": "infinibox-csi-driver",
                  "csiDriverVersion": "v2.23.0",
                  "e2etesting": false,
                  "images": {
                    "attachersidecar": "registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901",
                    "attachersidecar_pull_policy": "IfNotPresent",
                    "csidriver": "registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61",
                    "csidriver_pull_policy": "Always",
                    "livenesssidecar": "registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc",
                    "livenesssidecar_pull_policy": "IfNotPresent",
                    "provisionersidecar": "registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2",
                    "provisionersidecar_pull_policy": "IfNotPresent",
                    "registrarsidecar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24",
                    "registrarsidecar_pull_policy": "IfNotPresent",
                    "resizersidecar": "registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d",
                    "resizersidecar_pull_policy": "IfNotPresent",
                    "snapshottersidecar": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a",
                    "snapshottersidecar_pull_policy": "IfNotPresent"
                  },
                  "instanceCount": 1,
                  "logLevel": "debug",
                  "nodeSelector": {
                    "kubernetes.io/os": "linux"
                  },
                  "nodeserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "nodeservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "protocolSecret": "",
                  "removeDomainName": false,
                  "replicaCount": 1,
                  "skipCredentialsCreation": false,
                  "volumeNamePrefix": "ibox"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
          createdAt: "2025-09-27T16:12:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2.23.0
          operatorframework.io/suggested-namespace: infinidat-csi
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Infinidat/infinibox-csi-driver
          support: Infinidat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Iboxreplica
            name: iboxreplicas.csidriver.infinidat.com
            version: v1
          - kind: InfiniboxCsiDriver
            name: infiniboxcsidrivers.csidriver.infinidat.com
            version: v1alpha1
        description: "Infinidat InfiniBox Container Storage Interface (CSI) Driver
          is a\n    CNCF-compliant Kubernetes integration for InfiniBox storage systems,
          offering\n    advanced enterprise functionality for petabyte-scale Kubernetes
          deployments including\n    Red Hat OpenShift.\n \n## Features and Benefits\n
          \ \n* **Multi-protocol flexibility** - manage\n    Kubernetes Persistent
          Volumes attached via block and file protocols,\n    including Fibre Channel,
          iSCSI, and NFS, with all Kubernetes PV access modes\n    \n* **Multi-petabyte
          scalability** - support hundreds of thousands of PVs per InfiniBox\n    system
          and control multiple InfiniBox arrays within a single Kubernetes cluster\n
          \   \\ \n* **Advanced enterprise features** - manage native InfiniBox snapshots
          and\n    clones, including restoring from snapshots, and import PVs created
          outside of\n    InfiniBox CSI Driver\n  \n## Required Parameters\n  \n*
          `hostname` - IP address\n    or hostname of the InfiniBox management interface\n*
          `username` / `password` -\n    InfiniBox credentials\n* `SecretName` - secret
          name, to be used in the StorageClass\n    to define a specific InfiniBox
          for persistent volumes\n\n## Optional Parameters\n*\n    `inbound_user`
          / `inbound_secret` / `outbound_user` / `outbound_secret` - credentials \n
          \   for iSCSI CHAP authentication\n\n    \n## Installation Instructions\n
          \n1. Create *infinidat-csi* namespace to install operator into.\n    \n2.
          Delete the following *clusterrolebindings*, if they exist, before installing
          operator\n     * infinidat-csi-operator-infinidat-csi-attacher\n     * infinidat-csi-operator-infinidat-csi-controller
          \                    \n     * infinidat-csi-operator-infinidat-csi-driver
          \                    \n     * infinidat-csi-operator-infinidat-csi-node
          \                              \n     * infinidat-csi-operator-infinidat-csi-provisioner
          \                       \n     * infinidat-csi-operator-infinidat-csi-resizer
          \          \n     * infinidat-csi-operator-infinidat-csi-snapshotter \n
          \   \n3. Install Infinidat Infinibox CSI Driver - Operator (this) into the
          *infinidat-csi* namespace.\n    \n4. Apply the following Security Context
          Constraints to the created service accounts in the *infinidat-csi* \n    namespace\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    \n5. Recreate or edit any existing storage classes, pointing to the
          new namespace\n\n     \n6. If you choose a different namespace other than
          *infinidat-csi* you will need to edit the clusterrolebindings listed above
          to specify your custom namespace\n\n     \n7. Install an instance of Infinidat
          Driver into the *infinidat-csi* namespace"
        displayName: InfiniBox CSI Driver - Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - csi
        - nas
        - san
        - storage
        - infinidat
        - infinibox
        links:
        - name: Infinidat
          url: https://infinidat.com
        maintainers:
        - email: info@infinidat.com
          name: Infinidat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Infinidat
          url: https://infinidat.com
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified@sha256:42ff34a019bc67076515e89735b4bb9921fbd4610c127ac8b61e1fa6a6a08194
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24
        version: 2.23.0
      entries:
      - name: infinibox-operator-certified.v2.23.0
        version: 2.23.0
      - name: infinibox-operator-certified.v2.22.0
        version: 2.22.0
      - name: infinibox-operator-certified.v2.21.0
        version: 2.21.0
      name: stable
    defaultChannel: stable
    packageName: infinibox-operator-certified
    provider:
      name: Infinidat
      url: https://infinidat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Shipwright Contributors
      provider-url: https://shipwright.io
    name: shipwright-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: shipwright-operator.v0.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "shipwright-build"
                },
                "spec": {
                  "targetNamespace": "shipwright-build"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: ghcr.io/shipwright-io/operator/operator:0.13.0@sha256:12cd5f936043c96f31cf571d7a4f8b6f2e240852cd9ad1a5423dbab0c208b542
          description: Shipwright is a framework for building container images on
            Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/shipwright-io/operator
          support: The Shipwright Contributors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
        description: |
          Shipwright is a framework for building container images on Kubernetes.

          Read more: [https://shipwright.io](https://shipwright.io)

          ## Usage

          To deploy and manage [Shipwright Builds](https://github.com/shipwright-io/build) in your cluster,
          first make sure this operator is installed and running on your cluster.

          Next, create the following:

          ```yaml
          ---
          apiVersion: operator.shipwright.io/v1alpha1
          kind: ShipwrightBuild
          metadata:
            name: shipwright-operator
          spec:
            targetNamespace: shipwright-build
          ```

          The operator will deploy Shipwright Builds in the provided `targetNamespace`.
          When `.spec.targetNamespace` is not set, the namespace will default to `shipwright-build`.
          Refer to the [ShipwrightBuild documentation](https://github.com/shipwright-io/operator/blob/main/docs/shipwrightbuild.md) for more information about this custom resource.
        displayName: Shipwright Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://shipwright.io
        - name: Shipwright Builds
          url: https://github.com/shipwright-io/build
        maintainers:
        - email: shipwright-dev@lists.shipwright.io
          name: The Shipwright Contributors
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Shipwright Contributors
          url: https://shipwright.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - ghcr.io/shipwright-io/operator/operator:0.13.0@sha256:12cd5f936043c96f31cf571d7a4f8b6f2e240852cd9ad1a5423dbab0c208b542
        version: 0.13.0
      entries:
      - name: shipwright-operator.v0.13.0
        version: 0.13.0
      - name: shipwright-operator.v0.9.0
        version: 0.9.0
      name: alpha
    defaultChannel: alpha
    packageName: shipwright-operator
    provider:
      name: The Shipwright Contributors
      url: https://shipwright.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: barbican-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: barbican-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "barbican.openstack.org/v1beta1",
                "kind": "Barbican",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "barbican-operator",
                    "app.kubernetes.io/instance": "barbican",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "barbican",
                    "app.kubernetes.io/part-of": "barbican-operator"
                  },
                  "name": "barbican"
                },
                "spec": {
                  "barbicanAPI": {
                    "customServiceConfig": "[optional]\noverrides = True\n",
                    "defautlConfigOverwrite": {
                      "optional_policy.json": "{\"some\": \"custom policy\"}\n"
                    },
                    "passwordSelectors": {
                      "database": "BarbicanDatabasePassword",
                      "service": "BarbicanPassword"
                    }
                  },
                  "barbicanKeystoneListener": {
                    "customServiceConfig": "[optional]\noverrides = True\n",
                    "defautlConfigOverwrite": {
                      "optional_policy.json": "{\"some\": \"custom policy\"}\n"
                    }
                  },
                  "barbicanWorker": {
                    "customServiceConfig": "[optional]\noverrides = True\n",
                    "defautlConfigOverwrite": {
                      "optional_policy.json": "{\"some\": \"custom policy\"}\n"
                    }
                  },
                  "customServiceConfig": "[DEFAULT]\ndebug = True\n",
                  "databaseAccount": "barbican",
                  "databaseInstance": "openstack",
                  "defaultConfigOverwrite": {
                    "policy.json": "{\"some\": \"custom policy\"}\n"
                  },
                  "passwordSelectors": {
                    "database": "BarbicanDatabasePassword",
                    "service": "BarbicanPassword",
                    "simplecryptokek": "BarbicanSimpleCryptoKEK"
                  },
                  "preserveJobs": true,
                  "rabbitMqCusterName": "barbican_rabbit",
                  "secret": "osp-secret",
                  "serviceAccount": "barbican",
                  "serviceUser": "barbican"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:32Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BarbicanAPI is the Schema for the barbicanapis API
            displayName: Barbican API
            kind: BarbicanAPI
            name: barbicanapis.barbican.openstack.org
            version: v1beta1
          - kind: BarbicanKeystoneListener
            name: barbicankeystonelisteners.barbican.openstack.org
            version: v1beta1
          - description: Barbican is the Schema for the barbicans API
            displayName: Barbican
            kind: Barbican
            name: barbicans.barbican.openstack.org
            version: v1beta1
          - description: BarbicanWorker is the Schema for the barbicanworkers API
            displayName: Barbican Worker
            kind: BarbicanWorker
            name: barbicanworkers.barbican.openstack.org
            version: v1beta1
        description: Barbican Operator
        displayName: Barbican Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - KeyManager
        - Barbican
        links:
        - name: Barbican Operator
          url: https://github.com/openstack-k8s-operators/barbican-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-barbican-worker-rhel9@sha256:b5133f539658ed5d532e0c3cf63d90bc0f76c7f0368963a8ebec49123f27f7a7
        - registry.redhat.io/rhoso/openstack-barbican-keystone-listener-rhel9@sha256:ac112444c0a12664e8a7900e72cd97fa795df167c0d95437a029644f0819f30d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/barbican-rhel9-operator@sha256:1275a3f7f09999e9dc25bc38f917a3b5d780d9d309442cddd0e7adf313d6f576
        - registry.redhat.io/rhoso/openstack-barbican-api-rhel9@sha256:0db6c1ab5b4ca302e38338a3ab9cee9bd4ae7c8e1e7bd3419e47b65df9ef1f4f
        version: 1.0.6
      entries:
      - name: barbican-operator.v1.0.6
        version: 1.0.6
      - name: barbican-operator.v1.0.5
        version: 1.0.5
      - name: barbican-operator.v1.0.4
        version: 1.0.4
      - name: barbican-operator.v1.0.3
        version: 1.0.3
      - name: barbican-operator.v1.0.2
        version: 1.0.2
      - name: barbican-operator.v1.0.1
        version: 1.0.1
      - name: barbican-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: barbican-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: ansible-cloud-addons-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aca-operator.v2.4.0-0.1737660894
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Cloud Addons Operator manages everything Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.1.1-0.1645500829 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy an AAP AWS Billing instance
            displayName: AnsibleAutomationPlatformAwsBilling
            kind: AnsibleAutomationPlatformAwsBilling
            name: ansibleautomationplatformawsbillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Azure Billing instance
            displayName: AnsibleAutomationPlatformAzureBilling
            kind: AnsibleAutomationPlatformAzureBilling
            name: ansibleautomationplatformazurebillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Metrics instance
            displayName: AnsibleAutomationPlatformMetrics
            kind: AnsibleAutomationPlatformMetrics
            name: ansibleautomationplatformmetrics.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP UI instance
            displayName: AutomationPlatform
            kind: AutomationPlatform
            name: automationplatforms.platform.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Cloud Addons is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Cloud Addons Operator provides for push-button deployments of Cloud Add-ons into your OpenShift environment. Ansible Cloud Addons deployments include:
            * Cloud Billing - Cloud billing connector for Ansible Automation Platform
            * Cloud UI - Unified welcome page UI for Ansible Automation Platform

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Cloud Addons operator by following instructions presented when you click the Install button.
        displayName: Ansible Cloud Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-platform
        - ansible-automation-platform
        - aap
        - cloud
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:38ee3484ba60bb474e7231e9d0718157386716cea7ffd6da6f32ddd551a5243d
        - registry.redhat.io/ansible-automation-platform-24/aap-cloud-billing-rhel8-operator@sha256:c461b0c4781e95d2b2e31e7bbd6d6d192806e8a8dd2e8b9225b5735da93dd6a2
        - registry.redhat.io/ansible-automation-platform-24/aap-cloud-ui-rhel8-operator@sha256:06bf313d4daf080ce6f2b1ba859fefd84431c7a2a0618feae55b92c9f5fff594
        - registry.redhat.io/ansible-automation-platform-24/aap-cloud-billing-rhel8@sha256:84fa86d312663a74502516f2ba838ac80e500c7bc17dc518940984a603da2548
        - registry.redhat.io/ansible-automation-platform-24/aap-cloud-metrics-collector-rhel8@sha256:6bd27ad06d0a65c7fa1cd2419bfc30976d53a5b89698d309cabbbd65b531a1ab
        - registry.redhat.io/ansible-automation-platform-24/aap-cloud-ui-rhel8@sha256:8489c253072186db064289f5ecc0a08ed3654a73be3c3f2130be479098e59047
        version: 2.4.0+0.1737660894
      entries:
      - name: aca-operator.v2.4.0-0.1708974547
        version: 2.4.0+0.1708974547
      - name: aca-operator.v2.4.0-0.1711587287
        version: 2.4.0+0.1711587287
      - name: aca-operator.v2.4.0-0.1713463331
        version: 2.4.0+0.1713463331
      - name: aca-operator.v2.4.0-0.1715891874
        version: 2.4.0+0.1715891874
      - name: aca-operator.v2.4.0-0.1717108266
        version: 2.4.0+0.1717108266
      - name: aca-operator.v2.4.0-0.1717558193
        version: 2.4.0+0.1717558193
      - name: aca-operator.v2.4.0-0.1718149348
        version: 2.4.0+0.1718149348
      - name: aca-operator.v2.4.0-0.1720478479
        version: 2.4.0+0.1720478479
      - name: aca-operator.v2.4.0-0.1725253934
        version: 2.4.0+0.1725253934
      - name: aca-operator.v2.4.0-0.1726108417
        version: 2.4.0+0.1726108417
      - name: aca-operator.v2.4.0-0.1727255735
        version: 2.4.0+0.1727255735
      - name: aca-operator.v2.4.0-0.1730152463
        version: 2.4.0+0.1730152463
      - name: aca-operator.v2.4.0-0.1733181569
        version: 2.4.0+0.1733181569
      - name: aca-operator.v2.4.0-0.1733937635
        version: 2.4.0+0.1733937635
      - name: aca-operator.v2.4.0-0.1737660894
        version: 2.4.0+0.1737660894
      name: stable-2.4-cluster-scoped
    - currentCSV: aca-operator.v2.5.0-0.1758143647
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Cloud Addons Operator manages everything Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.1.1-0.1645500829 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy an AAP AWS Billing instance
            displayName: AnsibleAutomationPlatformAwsBilling
            kind: AnsibleAutomationPlatformAwsBilling
            name: ansibleautomationplatformawsbillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Azure Billing instance
            displayName: AnsibleAutomationPlatformAzureBilling
            kind: AnsibleAutomationPlatformAzureBilling
            name: ansibleautomationplatformazurebillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Metrics instance
            displayName: AnsibleAutomationPlatformMetrics
            kind: AnsibleAutomationPlatformMetrics
            name: ansibleautomationplatformmetrics.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP UI instance
            displayName: AutomationPlatform
            kind: AutomationPlatform
            name: automationplatforms.platform.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Cloud Addons is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Cloud Addons Operator provides for push-button deployments of Cloud Add-ons into your OpenShift environment. Ansible Cloud Addons deployments include:
            * Cloud Billing - Cloud billing connector for Ansible Automation Platform
            * Cloud UI - Unified welcome page UI for Ansible Automation Platform

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Cloud Addons operator by following instructions presented when you click the Install button.
        displayName: Ansible Cloud Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-platform
        - ansible-automation-platform
        - aap
        - cloud
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-billing-rhel8-operator@sha256:f4b29ec59498bb6c9fdb2790ffc9d8c878efa3fd9da3416f7fe9f2a253171508
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-ui-rhel8-operator@sha256:47e2c40cdcfe218d5bf7abbf6fc35340bacdc2d02dac7f81deee7876b5570c53
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-billing-rhel8@sha256:0e50f539bfeaec91f0b18409cbc1888297a26af5114252a8ca07e1b97436001d
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-metrics-collector-rhel8@sha256:c6ab25b3839f8c3134e7d5e1f0b7716ea5fd66fe8d131261ff08aabd97c3cd2d
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-ui-rhel8@sha256:869efe830cdf23f9344263f0e9b5a209cb13b84d9d7588a1744fd3ea4cd3f83b
        version: 2.5.0+0.1758143647
      entries:
      - name: aca-operator.v2.5.0-0.1727240231
        version: 2.5.0+0.1727240231
      - name: aca-operator.v2.5.0-0.1727874815
        version: 2.5.0+0.1727874815
      - name: aca-operator.v2.5.0-0.1728516369
        version: 2.5.0+0.1728516369
      - name: aca-operator.v2.5.0-0.1729737962
        version: 2.5.0+0.1729737962
      - name: aca-operator.v2.5.0-0.1731452082
        version: 2.5.0+0.1731452082
      - name: aca-operator.v2.5.0-0.1733188778
        version: 2.5.0+0.1733188778
      - name: aca-operator.v2.5.0-0.1733964232
        version: 2.5.0+0.1733964232
      - name: aca-operator.v2.5.0-0.1736892777
        version: 2.5.0+0.1736892777
      - name: aca-operator.v2.5.0-0.1737095650
        version: 2.5.0+0.1737095650
      - name: aca-operator.v2.5.0-0.1737671171
        version: 2.5.0+0.1737671171
      - name: aca-operator.v2.5.0-0.1739244024
        version: 2.5.0+0.1739244024
      - name: aca-operator.v2.5.0-0.1740088272
        version: 2.5.0+0.1740088272
      - name: aca-operator.v2.5.0-0.1740769078
        version: 2.5.0+0.1740769078
      - name: aca-operator.v2.5.0-0.1741367998
        version: 2.5.0+0.1741367998
      - name: aca-operator.v2.5.0-0.1742427652
        version: 2.5.0+0.1742427652
      - name: aca-operator.v2.5.0-0.1743655743
        version: 2.5.0+0.1743655743
      - name: aca-operator.v2.5.0-0.1746134365
        version: 2.5.0+0.1746134365
      - name: aca-operator.v2.5.0-0.1747339736
        version: 2.5.0+0.1747339736
      - name: aca-operator.v2.5.0-0.1749072064
        version: 2.5.0+0.1749072064
      - name: aca-operator.v2.5.0-0.1750895274
        version: 2.5.0+0.1750895274
      - name: aca-operator.v2.5.0-0.1753899647
        version: 2.5.0+0.1753899647
      - name: aca-operator.v2.5.0-0.1755834478
        version: 2.5.0+0.1755834478
      - name: aca-operator.v2.5.0-0.1758143647
        version: 2.5.0+0.1758143647
      name: stable-2.5-cluster-scoped
    defaultChannel: stable-2.5-cluster-scoped
    packageName: ansible-cloud-addons-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: service-telemetry-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-telemetry-operator.v1.5.1757650583
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infra.watch/v1beta1",
                "kind": "ServiceTelemetry",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "alerting": {
                    "alertmanager": {
                      "receivers": {
                        "snmpTraps": {
                          "alertOidLabel": "oid",
                          "community": "public",
                          "enabled": false,
                          "port": 162,
                          "retries": 5,
                          "target": "192.168.24.254",
                          "timeout": 1,
                          "trapDefaultOid": "1.3.6.1.4.1.50495.15.1.2.1",
                          "trapDefaultSeverity": "",
                          "trapOidPrefix": "1.3.6.1.4.1.50495.15"
                        }
                      },
                      "storage": {
                        "persistent": {
                          "pvcStorageRequest": "20G"
                        },
                        "strategy": "persistent"
                      }
                    },
                    "enabled": true
                  },
                  "backends": {
                    "events": {
                      "elasticsearch": {
                        "certificates": {
                          "caCertDuration": "70080h",
                          "endpointCertDuration": "70080h"
                        },
                        "enabled": false,
                        "forwarding": {
                          "hostUrl": "https://elasticsearch-es-http:9200",
                          "tlsSecretName": "elasticsearch-es-cert",
                          "tlsServerName": "",
                          "useBasicAuth": true,
                          "useTls": true,
                          "userSecretName": "elasticsearch-es-elastic-user"
                        },
                        "storage": {
                          "persistent": {
                            "pvcStorageRequest": "20Gi"
                          },
                          "strategy": "persistent"
                        },
                        "version": "7.16.1"
                      }
                    },
                    "metrics": {
                      "prometheus": {
                        "enabled": true,
                        "scrapeInterval": "30s",
                        "storage": {
                          "persistent": {
                            "pvcStorageRequest": "20G"
                          },
                          "retention": "24h",
                          "strategy": "persistent"
                        }
                      }
                    }
                  },
                  "clouds": [
                    {
                      "metrics": {
                        "collectors": [
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 16384,
                              "verbose": false
                            },
                            "collectorType": "collectd",
                            "debugEnabled": false,
                            "subscriptionAddress": "collectd/cloud1-telemetry"
                          },
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 16384,
                              "verbose": false
                            },
                            "collectorType": "ceilometer",
                            "debugEnabled": false,
                            "subscriptionAddress": "anycast/ceilometer/cloud1-metering.sample"
                          },
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 65535,
                              "verbose": false
                            },
                            "collectorType": "sensubility",
                            "debugEnabled": false,
                            "subscriptionAddress": "sensubility/cloud1-telemetry"
                          }
                        ]
                      },
                      "name": "cloud1"
                    }
                  ],
                  "graphing": {
                    "enabled": false,
                    "grafana": {
                      "baseImage": "registry.redhat.io/rhel8/grafana:9",
                      "dashboards": {
                        "enabled": true
                      },
                      "disableSignoutMenu": false,
                      "ingressEnabled": true
                    }
                  },
                  "highAvailability": {
                    "enabled": false
                  },
                  "observabilityStrategy": "use_redhat",
                  "transports": {
                    "qdr": {
                      "auth": "basic",
                      "certificates": {
                        "caCertDuration": "70080h",
                        "endpointCertDuration": "70080h"
                      },
                      "enabled": true,
                      "web": {
                        "enabled": false
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/stf/service-telemetry-rhel9-operator@sha256:822becd77fa8835e753c88b3e307f62e42e090bdedc6e1fbcbe9fd6110310846
          createdAt: "2025-09-12T04:17:40Z"
          description: Service Telemetry Framework. Umbrella Operator for instantiating
            the required dependencies and configuration of various components to build
            a Service Telemetry platform for telco grade monitoring.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: service-telemetry
          operators.openshift.io/valid-subscription: '["OpenStack Platform", "Cloud
            Infrastructure", "Cloud Suite"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/infrawatch/service-telemetry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of the Service Telemetry Framework
            displayName: STF Cluster
            kind: ServiceTelemetry
            name: servicetelemetrys.infra.watch
            version: v1beta1
        description: Service Telemetry Operator for monitoring clouds
        displayName: Service Telemetry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - servicetelemetry
        - monitoring
        - telemetry
        - notifications
        - telecommunications
        links:
        - name: Source Code
          url: https://github.com/infrawatch/service-telemetry-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_openstack_platform/17.1/html/service_telemetry_framework_1.5/index
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy:latest@sha256:dc24aa382e02c4ac058d4b824a21f0e90a6259d6d817d0341b2a41809b5daa78
        - registry.redhat.io/cluster-observability-operator/coo-prometheus-rhel8:0.2.0@sha256:6bbb1aebb1cfdc3310e398ec7720ca1d4a9a89cecf7406412071bfaceb0d3cb3
        - registry.redhat.io/cluster-observability-operator/coo-prometheus-alertmanager-rhel8:0.2.0@sha256:cfd18e83d64a55ac310937eefe83fc5b9e98660962a127cd4f8d9a382764c171
        - registry.redhat.io/stf/service-telemetry-rhel9-operator@sha256:822becd77fa8835e753c88b3e307f62e42e090bdedc6e1fbcbe9fd6110310846
        - registry.redhat.io/stf/prometheus-webhook-snmp-rhel9@sha256:03199f8a3b6b22c82e01ede3adab6e98b276369d642029ae03641aac3b25581a
        version: 1.5.1757650583
      entries:
      - name: service-telemetry-operator.v1.5.1757650583
        version: 1.5.1757650583
      - name: service-telemetry-operator.v1.5.1743671601
        version: 1.5.1743671601
      - name: service-telemetry-operator.v1.5.1740051851
        version: 1.5.1740051851
      - name: service-telemetry-operator.v1.5.1731336911
        version: 1.5.1731336911
      - name: service-telemetry-operator.v1.5.1725952694
        version: 1.5.1725952694
      name: stable-1.5
    defaultChannel: stable-1.5
    packageName: service-telemetry-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: node-healthcheck-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-healthcheck-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "maxUnhealthy": "49%",
                  "remediationTemplate": {
                    "apiVersion": "poison-pill.medik8s.io/v1alpha1",
                    "kind": "PoisonPillRemediationTemplate",
                    "name": "group-x",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-healthcheck-operator:0.1.0-latest
          createdAt: "2021-04-20 12:00:00"
          description: An operator to monitor node healthcheck and remdiate via 3rd
            party providers like poison-pill.
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: This operator deploys the Node Health Check controller and its
          artifacts
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://node-healthcheck-operator.domain
        maintainers:
        - email: medik8s@googlegroups.com
          name: medik8s team
        maturity: alpha
        provider:
          name: medik8s
          url: https://github.com/medik8s
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/node-healthcheck-operator:0.1.0-latest
        version: 0.1.1
      entries:
      - name: node-healthcheck-operator.v0.1.1
        version: 0.1.1
      name: alpha
    - currentCSV: node-healthcheck-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-healthcheck-operator:v0.3.0
          createdAt: "2022-08-24 11:40:38"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.2.0 <0.3.0'
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: medik8s team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: medik8s
          url: https://github.com/medik8s
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/node-healthcheck-operator:v0.3.0
        version: 0.3.0
      entries:
      - name: node-healthcheck-operator.v0.3.0
        version: 0.3.0
      name: candidate
    - currentCSV: node-healthcheck-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: quay.io/medik8s/node-healthcheck-operator:v0.6.1
          createdAt: "2023-11-14T15:45:34Z"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.1.0 <0.6.1'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/medik8s/node-healthcheck-operator:v0.6.1
        - quay.io/medik8s/node-remediation-console:v0.6.1
        version: 0.6.1
      entries:
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      - name: node-healthcheck-operator.v0.6.0
        version: 0.6.0
      - name: node-healthcheck-operator.v0.5.0
        version: 0.5.0
      - name: node-healthcheck-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: node-healthcheck-operator
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: GitLab Inc
      provider-url: https://about.gitlab.com/
    name: gitlab-operator-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "gitlab": {
                        "gitlab-shell": {
                          "maxReplicas": 2,
                          "minReplicas": 1
                        },
                        "sidekiq": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1000M"
                            }
                          }
                        },
                        "webservice": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1500M"
                            }
                          }
                        }
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "installCertmanager": false,
                      "minio": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      },
                      "nginx-ingress": {
                        "controller": {
                          "electionID": "ingress-controller-leader",
                          "replicaCount": 1
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      },
                      "redis": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-29T12:42:23Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      name: stable
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "gitlab": {
                        "gitlab-shell": {
                          "maxReplicas": 2,
                          "minReplicas": 1
                        },
                        "sidekiq": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1000M"
                            }
                          }
                        },
                        "webservice": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1500M"
                            }
                          }
                        }
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "installCertmanager": false,
                      "minio": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      },
                      "nginx-ingress": {
                        "controller": {
                          "electionID": "ingress-controller-leader",
                          "replicaCount": 1
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      },
                      "redis": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-29T12:42:23Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      name: unstable
    defaultChannel: stable
    packageName: gitlab-operator-kubernetes
    provider:
      name: GitLab Inc
      url: https://about.gitlab.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: edb-hcp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: edb-hcp-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edbpgai.edb.com/v1alpha1",
                "kind": "HybridControlPlane",
                "metadata": {
                  "name": "placeholder"
                },
                "spec": {
                  "beaconAgent": {
                    "applianceProviderEnabled": true,
                    "beaconDBConnected": true,
                    "beaconServerAddress": "upm-beacon-server-ingestion.upm-beacon.svc.cluster.local:8080",
                    "inCluster": true,
                    "location": "devspatcher",
                    "logLevel": "debug",
                    "plaintext": true,
                    "providers": [
                      "appliance"
                    ],
                    "provisioning": {
                      "imageDiscovery": true,
                      "imagesetDiscoveryAuthenticationType": "token",
                      "imagesetDiscoveryContainerRegistryURL": "placeholder",
                      "loadBalancersEnabled": true,
                      "nodePortDomain": "nodeportdomain.com",
                      "openshift": true,
                      "provider": "aws"
                    },
                    "tlsInsecure": false,
                    "transparentDataEncryptionMethods": [
                      "passphrase",
                      "hashicorp_vault"
                    ]
                  },
                  "beaconServer": {
                    "logLevel": "debug"
                  },
                  "componentsParameters": {
                    "edb-migration-portal": {
                      "enable_pdb": "false"
                    },
                    "transporter-data-operator": {
                      "fips_enabled": "false"
                    },
                    "transporter-dp-agent": {
                      "rw_service_url": "placeholder"
                    },
                    "transporter-rw-service": {
                      "domain_name": "placeholder"
                    },
                    "upm-api-pnxt-metric": {
                      "log_level": "info",
                      "tsdb_host": "thanos-query.monitoring.svc.cluster.local",
                      "tsdb_port": "9090"
                    },
                    "upm-api-pnxt-pd": {
                      "log_level": "info"
                    },
                    "upm-beaco-ff-base": {
                      "usage_generator_interval": "24h"
                    },
                    "upm-beaco-usages": {
                      "usage_generator_interval": "24h"
                    },
                    "upm-beacon": {
                      "beacon_db_hibernation": "\"off\"",
                      "beacon_server_replicas": "1",
                      "external_metrics_replicas": "1",
                      "image_discovery_credentials_secret_name": "placeholder",
                      "server_host": "placeholder"
                    },
                    "upm-istio": {
                      "ingress_grpc_tls_node_port": "30290",
                      "ingress_http_node_port": "32542",
                      "ingress_https_node_port": "30288"
                    },
                    "upm-istio-gateway": {
                      "cookie_aeskey": "rzkutHl8NJNztPMEJYykZouHslNiA7xmIXH+58ISUVo=",
                      "ingress_grpc_tls_node_port": "30290",
                      "ingress_http_node_port": "32542",
                      "ingress_https_node_port": "30288",
                      "openshift_console_domain_name": "placeholder"
                    },
                    "upm-prometheus": {
                      "cpu": "200m",
                      "memory": "6Gi",
                      "prometheus_retention_size": "15GB",
                      "prometheus_storage_size": "20Gi"
                    },
                    "upm-thanos": {
                      "memcached_cpu": "500m",
                      "memcached_memory": "2Gi",
                      "query_frontend_cpu": "200m",
                      "query_frontend_memory": "128Mi"
                    },
                    "upm-trust-manager": {
                      "cert_manager_namespace": "cert-manager"
                    }
                  },
                  "flavour": "placeholder",
                  "globalParameters": {
                    "default_ingress_domain": "placeholder",
                    "internal_backup_folder": "a7462dbc7106",
                    "portal_certificate_secret": "beaco-istio-gw-cert",
                    "portal_domain_name": "placeholder",
                    "portal_port": "443",
                    "storage_class": "placeholder"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "placeholder",
                      "namespace": "default"
                    }
                  ],
                  "imageRegistry": "placeholder",
                  "pgai": {
                    "portal": {
                      "authentication": {
                        "clientSecret": "edb-postgres-ai-secret",
                        "idpConnectors": [],
                        "staticPasswords": [
                          {
                            "email": "owner@mycompany.com",
                            "hash": "$2y$10/JLTqiuGblH.RemIutWxkztN5m",
                            "userID": "c5998173-a605-449a-a9a5-4a9c33e26df7",
                            "username": "Owner MyCompany"
                          }
                        ]
                      }
                    }
                  },
                  "source": {
                    "useLocalKustomizations": true
                  },
                  "version": "placeholder"
                }
              }
            ]
          capabilities: Basic Install
          categories: Database,AI/Machine Learning
          containerImage: docker.enterprisedb.com/pgai-platform/edb-hcp-operator/manager@sha256:82610c624e956a67df4858929e1aabd561a7537ba68d40e6512b68f55e2649e0
          createdAt: "2025-06-17T11:00:42Z"
          description: Unify transactional, analytical, and AI workloads and bring
            cloud agility to your hybrid data infrastructure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/EnterpriseDB/edb-hcp-operator
          support: techsupport@enterprisedb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HybridControlPlane is the Schema for the hybridcontrolplanes
              API
            displayName: Hybrid Control Plane
            kind: HybridControlPlane
            name: hybridcontrolplanes.edbpgai.edb.com
            version: v1alpha1
        description: |
          EDB Postgres AI is a sovereign data and AI platform that brings analytics and AI closer to core operational data and enables cloud agility in private environments.
          Unify your structured and unstructured data in one multi-model data platform with centralized management and observability to:
          - Accelerate enterprise-grade transactional databases
          - Enable rapid analytics and powerful lakehouse integrations
          - Automate AI pipelines and semantic search for faster GenAI development
          Using EDB’s unmatched Postgres expertise, EDB Postgres AI automates database operations, ensuring expert-level execution across deployments – boosting operational efficiency by up to 30% and accelerating innovation.
        displayName: EDB Postgres AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Postgres
        - Database
        - DBaaS
        - EDB Postgres AI
        links:
        - name: EDB Postgres AI Software Deployment
          url: https://www.enterprisedb.com/products/software-deployment
        maintainers:
        - email: techsupport@enterprisedb.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - docker.enterprisedb.com/pgai-platform/edb-hcp-operator/manager@sha256:82610c624e956a67df4858929e1aabd561a7537ba68d40e6512b68f55e2649e0
        version: 1.2.0
      entries:
      - name: edb-hcp-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: edb-hcp-operator
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SAP LinuxLab - Red Hat
      provider-url: https://github.com/sap-linuxlab
    name: sap-data-intelligence-observer-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: observer-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sdi.sap-redhat.io/v1alpha1",
                "kind": "SDIObserver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "observer-operator",
                    "app.kubernetes.io/instance": "sdiobserver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sdiobserver",
                    "app.kubernetes.io/part-of": "observer-operator"
                  },
                  "name": "sdiobserver-sample"
                },
                "spec": {
                  "manageSDINodeConfig": true,
                  "sdiNamespace": "sdi",
                  "sdiVSystemRoute": {
                    "managementState": "Managed"
                  },
                  "slcbNamespace": "sap-slcbridge",
                  "slcbRoute": {
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/redhat-sap-cop/sdi-observer-operator:0.1.5
          description: Configure and enable SAP Data Intelligence to run on OpenShift
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SDIObserver is the Schema for the sdiobservers API
            displayName: SDIObserver
            kind: SDIObserver
            name: sdiobservers.sdi.sap-redhat.io
            version: v1alpha1
        description: Operator for monitoring SAP Data Intelligence (SDI) namespace
          and modifying objects in there that enable running of SDI on top of OpenShift.
          The observer shall be run in a dedicated namespace. It must be deployed
          before the SDI installation is started. This is the prerequisite for the
          successful installation of the SDI on OpenShift. Please refer to https://access.redhat.com/articles/5100521
          for detailed information.
        displayName: SAP Data Intelligence 3 - Observer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sdi
        - sap data intelligence
        links:
        - name: SAP Data Intelligence 3 - Observer Operator
          url: https://github.com/redhat-sap/sap-data-intelligence/tree/master/observer-operator
        maintainers:
        - email: mjiao@redhat.com
          name: Manjun Jiao
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/redhat-sap-cop/sdi-observer-operator:0.1.5
        version: 0.0.3
      entries:
      - name: observer-operator.v0.0.3
        version: 0.0.3
      name: alpha
    - currentCSV: observer-operator.v1.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sdi.sap-redhat.io/v1alpha1",
                "kind": "SDIObserver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/part-of": "observer-operator"
                  },
                  "name": "sdiobserver-sample",
                  "namespace": "sdi-observer"
                },
                "spec": {
                  "manageSDINodeConfig": true,
                  "sdiNamespace": "sdi",
                  "sdiVSystemRoute": {
                    "managementState": "Managed"
                  },
                  "slcbNamespace": "sap-slcbridge",
                  "slcbRoute": {
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/redhat-sap-cop/sdi-observer-operator:0.1.32
          createdAt: "20240726"
          description: Configure and enable SAP Data Intelligence 3 to run on OpenShift
            4
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mjiao/sap-data-intelligence/tree/master/observer-operator
          support: RedHat-SAPLinuxLab
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SDIObserver is the Schema for the sdiobservers API
            displayName: SDIObserver
            kind: SDIObserver
            name: sdiobservers.sdi.sap-redhat.io
            version: v1alpha1
        description: Operator for monitoring SAP Data Intelligence (SDI) namespace
          and modifying objects in there that enable running of SDI on top of OpenShift.
          It must be deployed before the SDI installation is started. This is the
          prerequisite for the successful installation of the SDI on OpenShift. Please
          refer to https://access.redhat.com/articles/7042265#ocp-install-sdi-observer-operator
          for detailed information.
        displayName: SAP Data Intelligence 3 - Observer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sdi
        - sap data intelligence
        links:
        - name: SAP Data Intelligence 3 - Observer Operator
          url: https://github.com/redhat-sap/sap-data-intelligence/tree/master/observer-operator
        maintainers:
        - email: mjiao@redhat.com
          name: Manjun Jiao
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/redhat-sap-cop/sdi-observer-operator:0.1.32
        version: 1.0.9
      entries:
      - name: observer-operator.v1.0.9
        version: 1.0.9
      - name: observer-operator.v1.0.8
        version: 1.0.8
      - name: observer-operator.v1.0.7
        version: 1.0.7
      - name: observer-operator.v1.0.6
        version: 1.0.6
      - name: observer-operator.v1.0.5
        version: 1.0.5
      - name: observer-operator.v1.0.4
        version: 1.0.4
      - name: observer-operator.v1.0.3
        version: 1.0.3
      - name: observer-operator.v1.0.2
        version: 1.0.2
      - name: observer-operator.v1.0.1
        version: 1.0.1
      - name: observer-operator.v0.0.3
        version: 0.0.3
      name: stable
    defaultChannel: stable
    packageName: sap-data-intelligence-observer-operator
    provider:
      name: SAP LinuxLab - Red Hat
      url: https://github.com/sap-linuxlab
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: k8gb.io
      provider-url: https://github.com/k8gb-io/k8gb
    name: k8gb
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k8gb.v0.11.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"k8gb.absa.oss/v1beta1","kind":"Gslb","metadata":{"name":"test-gslb-failover","namespace":"test-gslb"},"spec":{"ingress":{"rules":[{"host":"failover.test.k8gb.io","http":{"paths":[{"backend":{"service":{"name":"frontend-podinfo","port":{"name":"http"}}},"path":"/"}]}}]},"strategy":{"primaryGeoTag":"eu-west-1","type":"failover"}}}]'
          capabilities: Seamless Upgrades
          categories: Networking
          certified: "false"
          containerImage: docker.io/absaoss/k8gb:v0.11.4
          createdAt: "2021-09-24 12:00:00"
          description: A cloud native Kubernetes Global Balancer
          operatorframework.io/suggested-namespace: k8gb
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: k8gb
          operators.operatorframework.io/builder: operator-sdk-v1.12.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/k8gb-io/k8gb
          support: cncf-k8gb-maintainers@lists.cncf.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gslb resource for global load balancing strategy configuration
            displayName: Gslb
            kind: Gslb
            name: gslbs.k8gb.absa.oss
            version: v1beta1
          - description: Using ExternalDNS it synchronizes exposed Kubernetes Services
              and Ingresses with DNS providers
            displayName: DNSEndpoint
            kind: DNSEndpoint
            name: dnsendpoints.externaldns.k8s.io
            version: v1alpha1
        description: |
          **A cloud native Kubernetes Global Balancer**

          A Global Service Load Balancing solution with a focus on having cloud native qualities and work natively in a Kubernetes context.

          Key Differentiators:
           - Load balancing is based on timeproof DNS protocol which is perfect for global scope and extremely reliable
           - No dedicated management cluster and no single point of failure
           - Kubernetes native application health checks utilizing status of Liveness and Readiness probes for load balancing decisions
           - Configuration with a single Kubernetes CRD of `Gslb` kind

          Operator needs to be configured by setting couple of environment variables. This can be done by modifying
          the subscription yaml in the web ui or via kubectl. Here is an example of changing the `EDGE_DNS_ZONE`:

          ```yaml
          kind: Subscription
            ...
          spec:
            ...
            config:
              env:
              - name: CLUSTER_GEO_TAG
                value: us
              - name: EDGE_DNS_ZONE
                value: mycloud.example.com
                ...
           ```
           For more deployment variables see [helm code](https://github.com/k8gb-io/k8gb/blob/v0.8.3/chart/k8gb/templates/operator.yaml#L53:L68).

           For more information see [k8gb.io](https://k8gb.io).
        displayName: k8gb
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gslb
        - dns-lb
        - kuberneters-global-balancer
        - kubernetes-operator
        - balancer
        - multi-cluster
        links:
        - name: Home
          url: https://www.k8gb.io/
        - name: Source
          url: https://github.com/k8gb-io/k8gb
        maintainers:
        - email: dinar.valeev@absa.africa
          name: Dinar Valeev
        - email: jiri.kremser@gmail.com
          name: Jiri Kremser
        - email: kuritka@gmail.com
          name: Michal Kuritka
        - email: timofey.ilinykh@absa.africa
          name: Timofey Ilinykh
        - email: yury.tsarev@absa.africa
          name: Yury Tsarev
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: k8gb.io
          url: https://github.com/k8gb-io/k8gb
        relatedImages:
        - absaoss/k8s_crd:v0.0.11
        - docker.io/absaoss/k8gb:v0.11.4
        version: 0.11.4
      entries:
      - name: k8gb.v0.11.4
        version: 0.11.4
      name: alpha
    defaultChannel: alpha
    packageName: k8gb
    provider:
      name: k8gb.io
      url: https://github.com/k8gb-io/k8gb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Internal
      provider-url: redhat.com
    name: l5-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: l5-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pets.bestie.com/v1",
                "kind": "Bestie",
                "metadata": {
                  "name": "bestie"
                },
                "spec": {
                  "image": "quay.io/opdev/bestie",
                  "maxReplicas": 10,
                  "size": 3,
                  "version": "1.3"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Application Runtime
          certified: "false"
          containerImage: quay.io/opdev/l5-operator:v0.0.3
          createdAt: "2023-06-27T16:56:10Z"
          description: L5 Operator is a demo operator that aims to demonstrate level
            5 capabilities for web applications.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          relatedImages: quay.io/opdev/bestie@sha256:0e99ada2b1de19043845dd9076789584aeddc2646402e5f0987dd00424be1ee7
          repository: https://github.com/opdev/l5-operator-demo
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Bestie is the Schema for the besties API.
            displayName: Bestie
            kind: Bestie
            name: besties.pets.bestie.com
            version: v1
        description: |
          ## Usage
          The L5 operator is meant to demonstrate a minimum viable prototype of all capability levels outilned in the operator framework. The L5 operator is compatible with both Kubernetes and OpenShift clusters. When running on vanilla kubernetes a few prerequisites need to be kept in mind which may not be available out of the box. These include an ingress controller and a prometheus operator for level 4 capability. The L5 operator also depends on the Crunchy Data PostgreSQL operator. On openshift when installing via the Operator Lifecycle Manager, this will automatically be installed when installing the operator via the embdeded operator hub. The Operator Lifecycle Manager is also available for vanilla kubernetes however it may not be installed out of the box. In this case the Crunchy Data PostgreSQL operator must be installed manually. The Crunchy Data PostgreSQL operator can be installed on vanilla kuberentes using the [instructions provided by Crunchy Data](https://access.crunchydata.com/documentation/postgres-operator/4.7.5/installation/postgres-operator/) More documentation and examples can be found on [github](https://github.com/opdev/l5-operator-demo).
        displayName: L5 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - l5-operator
        links:
        - name: L5 Operator
          url: https://opdev.github.io/l5-operator-demo/
        maintainers:
        - email: partnereng@redhat.com
          name: OpDev Team
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat Internal
          url: redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/opdev/l5-operator:v0.0.3
        version: 0.0.3
      entries:
      - name: l5-operator.v0.0.3
        version: 0.0.3
      - name: l5-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: l5-operator
    provider:
      name: Red Hat Internal
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
    name: aikit-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aikit-operator.v2021.2.102120
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aikit.intel/v1alpha1",
                "kind": "AIKitContainer",
                "metadata": {
                  "name": "intel-aikit-container",
                  "namespace": "redhat-ods-applications"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagestream": {
                    "name": "oneapi-aikit",
                    "namespace": "redhat-ods-applications",
                    "registry": {
                      "root": "registry.connect.redhat.com",
                      "repo": "intel",
                      "name": "oneapi-aikit",
                      "version": "latest"
                    }
                  },
                  "nameOverride": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AIKitContainer
            name: aikitcontainers.aikit.intel
            version: v1alpha1
        description: |
          The Intel® oneAPI AI Analytics Toolkit gives data scientists, AI developers, and researchers familiar Python* tools and frameworks to accelerate end-to-end data science and machine learning pipelines on Intel® architectures. The components are built using oneAPI libraries for low-level compute optimizations. This toolkit maximizes performance from preprocessing through machine and deep learning phases and provides interoperability for efficient model development and deployment across single and multinodes.
        displayName: Intel® oneAPI AI Analytics Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aikit-operator
        links:
        - name: Aikit Operator
          url: https://aikit-operator.domain
        maintainers:
        - email: kam.d.kasravi@intel.com
          name: Kam Kasravi
        - email: abolfazl.shahbazi@intel.com
          name: Abolfazl Shahbazi
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Intel
          url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
        relatedImages:
        - registry.connect.redhat.com/intel/aikit-operator@sha256:7abb5e6164ffb3e4b87c39061f45d09848218fcfb74b48bf5e1035d56b41ee0c
        version: 2021.2.102120
      entries:
      - name: aikit-operator.v2021.2.102120
        version: 2021.2.102120
      - name: aikit-operator.v2021.2.100410
        version: 2021.2.100410
      - name: aikit-operator.v2021.2.92911
        version: 2021.2.92911
      - name: aikit-operator.v2021.2.92814
        version: 2021.2.92814
      - name: aikit-operator.v2021.2.92808-dev-a60e5ba
        version: 2021.2.92808-dev-a60e5ba
      name: alpha
    defaultChannel: alpha
    packageName: aikit-operator
    provider:
      name: Intel
      url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-airflow-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airflow-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:d8f8c439f33e5a80e16c7867a008d71693fd4928357ac6919c77bd0ff0bb1339
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:d8f8c439f33e5a80e16c7867a008d71693fd4928357ac6919c77bd0ff0bb1339
        version: 24.11.1
      entries:
      - name: airflow-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: airflow-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:173363db3cb0997aff09fe4b7f3a54c04f22bc56db83d2b30757e3cc745c4587
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:173363db3cb0997aff09fe4b7f3a54c04f22bc56db83d2b30757e3cc745c4587
        version: 25.3.0
      entries:
      - name: airflow-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: airflow-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
        version: 25.7.0
      entries:
      - name: airflow-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: airflow-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
        version: 25.7.0
      entries:
      - name: airflow-operator.v25.7.0
        version: 25.7.0
      - name: airflow-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-airflow-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: windows-machine-config-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: windows-machine-config-operator.v10.16.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          createdAt: "2025-06-06T22:08:07Z"
          description: An operator that enables Windows container workloads on OCP
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=10.15.0 <10.16.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.openshift.io/valid-subscription: '["Red Hat OpenShift support
            for Windows Containers"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ### Introduction
          The Windows Machine Config Operator configures Windows Machines into nodes, enabling Windows container workloads to
          be run on OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/4.16/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://docs.openshift.com/container-platform/4.16/windows_containers/byoh-windows-instance.html)
          The operator completes all the necessary steps to configure the Windows instance so that it can join the cluster as a worker node.

          Usage of this operator requires an OpenShift Container Platform for Windows Containers subscription. Users looking to
          deploy Windows containers workloads in production clusters should acquire a subscription before attempting to
          install this operator. Users without a subscription can try the community operator, a distribution which lacks
          official support.

          ### Pre-requisites
          * A Red Hat OpenShift Container Platform for Windows Containers [subscription](https://access.redhat.com/support/policy/updates/openshift#windows)
          * OCP 4.16 cluster running on Azure, AWS, GCP or vSphere configured with hybrid OVN Kubernetes networking
          * [WMCO prerequisites](https://docs.openshift.com/container-platform/4.16/windows_containers/understanding-windows-container-workloads.html#wmco-prerequisites__understanding-windows-container-workloads)

          ### Usage
          Once the `openshift-windows-machine-config-operator` namespace has been created, a secret must be created containing
          the private key that will be used to access the Windows VMs:
          ```
          # Create secret containing the private key in the openshift-windows-machine-config-operator namespace
          oc create secret generic cloud-private-key --from-file=private-key.pem=/path/to/key -n openshift-windows-machine-config-operator
          ```
          We strongly recommend not using the same
          [private key](https://docs.openshift.com/container-platform/4.16/installing/installing_azure/installing-azure-default.html#ssh-agent-using_installing-azure-default)
          used when installing the cluster

          Below is an example of a vSphere Windows MachineSet which can create Windows Machines that the WMCO can react upon.
          Please note that the windows-user-data secret will be created by the WMCO lazily when it is configuring the first
          Windows Machine. After that, the windows-user-data will be available for the subsequent MachineSets to be consumed.
          It might take around 10 minutes for the Windows VM to be configured so that it joins the cluster. Please note that
          the MachineSet should have following labels:
          * `machine.openshift.io/os-id: Windows`
          * `machine.openshift.io/cluster-api-machine-role: worker`
          * `machine.openshift.io/cluster-api-machine-type: worker`

          The following label must be added to the Machine spec within the MachineSet spec:
          * `node-role.kubernetes.io/worker: ""`

          Not having these labels will result in the Windows node not being marked as a worker.

          `<infrastructureID>` should be replaced with the output of:
          ```
          oc get -o jsonpath='{.status.infrastructureName}{"\n"}' infrastructure cluster
          ```

          The following template variables need to be replaced as follows with values from your vSphere environment:
          * `<Windows_VM_template>`: template name
          * `<VM Network Name>`: network name
          * `<vCenter DataCenter Name>`: datacenter name
          * `<Path to VM Folder in vCenter>`: path where your OpenShift cluster is running
          * `<vCenter Datastore Name>`: datastore name
          * `<vCenter Server FQDN/IP>`: IP address or FQDN of the vCenter server

          Please note that on vSphere, Windows Machine names cannot be more than 15 characters long. The MachineSet name therefore
          cannot be more than 9 characters long, due to the way Machine names are generated from it.
          ```
          apiVersion: machine.openshift.io/v1beta1
          kind: MachineSet
          metadata:
            labels:
              machine.openshift.io/cluster-api-cluster: <infrastructureID>
            name: winworker
            namespace: openshift-machine-api
          spec:
            replicas: 1
            selector:
              matchLabels:
                machine.openshift.io/cluster-api-cluster: <infrastructureID>
                machine.openshift.io/cluster-api-machineset: winworker
            template:
              metadata:
                labels:
                  machine.openshift.io/cluster-api-cluster: <infrastructureID>
                  machine.openshift.io/cluster-api-machine-role: worker
                  machine.openshift.io/cluster-api-machine-type: worker
                  machine.openshift.io/cluster-api-machineset: winworker
                  machine.openshift.io/os-id: Windows
              spec:
                metadata:
                  labels:
                    node-role.kubernetes.io/worker: ""
                providerSpec:
                  value:
                    apiVersion: vsphereprovider.openshift.io/v1beta1
                    credentialsSecret:
                      name: vsphere-cloud-credentials
                      namespace: openshift-machine-api
                    diskGiB: 128
                    kind: VSphereMachineProviderSpec
                    memoryMiB: 16384
                    metadata:
                      creationTimestamp: null
                    network:
                      devices:
                      - networkName: "<VM Network Name>"
                    numCPUs: 4
                    numCoresPerSocket: 1
                    snapshot: ""
                    template: <Windows_VM_template
                    userDataSecret:
                      name: windows-user-data
                    workspace:
                      datacenter: <vCenter DataCenter Name>
                      datastore: <vCenter Datastore Name>
                      folder: <Path to VM Folder in vCenter> # e.g. /DC/vm/ocp45-2tdrm
                      server: <vCenter Server FQDN/IP>
          ```
          Example MachineSet for other cloud providers:
          - [AWS](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-aws.html)
          - [Azure](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-azure.html)
          - [GCP](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-gcp.html)

          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.

          ### Reporting issues
          Support for this distribution of WMCO requires a Red Hat OpenShfit subscription. Support should be requested through the Red Hat Customer Portal.
          Please read through the [troubleshooting document](https://docs.openshift.com/container-platform/4.16/support/troubleshooting/troubleshooting-windows-container-workload-issues.html)
          before opening a support case.
        displayName: Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4-wincw/windows-machine-config-rhel9-operator@sha256:c9a6166158e1ccb1764007d79c6e0f879505c3546eb45cbbe79ebeefcc78a728
        version: 10.16.2
      entries:
      - name: windows-machine-config-operator.v10.16.2
        version: 10.16.2
      - name: windows-machine-config-operator.v10.16.1
        version: 10.16.1
      - name: windows-machine-config-operator.v10.16.0
        version: 10.16.0
      name: preview
    - currentCSV: windows-machine-config-operator.v10.16.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          createdAt: "2025-06-06T22:08:07Z"
          description: An operator that enables Windows container workloads on OCP
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=10.15.0 <10.16.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.openshift.io/valid-subscription: '["Red Hat OpenShift support
            for Windows Containers"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ### Introduction
          The Windows Machine Config Operator configures Windows Machines into nodes, enabling Windows container workloads to
          be run on OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/4.16/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://docs.openshift.com/container-platform/4.16/windows_containers/byoh-windows-instance.html)
          The operator completes all the necessary steps to configure the Windows instance so that it can join the cluster as a worker node.

          Usage of this operator requires an OpenShift Container Platform for Windows Containers subscription. Users looking to
          deploy Windows containers workloads in production clusters should acquire a subscription before attempting to
          install this operator. Users without a subscription can try the community operator, a distribution which lacks
          official support.

          ### Pre-requisites
          * A Red Hat OpenShift Container Platform for Windows Containers [subscription](https://access.redhat.com/support/policy/updates/openshift#windows)
          * OCP 4.16 cluster running on Azure, AWS, GCP or vSphere configured with hybrid OVN Kubernetes networking
          * [WMCO prerequisites](https://docs.openshift.com/container-platform/4.16/windows_containers/understanding-windows-container-workloads.html#wmco-prerequisites__understanding-windows-container-workloads)

          ### Usage
          Once the `openshift-windows-machine-config-operator` namespace has been created, a secret must be created containing
          the private key that will be used to access the Windows VMs:
          ```
          # Create secret containing the private key in the openshift-windows-machine-config-operator namespace
          oc create secret generic cloud-private-key --from-file=private-key.pem=/path/to/key -n openshift-windows-machine-config-operator
          ```
          We strongly recommend not using the same
          [private key](https://docs.openshift.com/container-platform/4.16/installing/installing_azure/installing-azure-default.html#ssh-agent-using_installing-azure-default)
          used when installing the cluster

          Below is an example of a vSphere Windows MachineSet which can create Windows Machines that the WMCO can react upon.
          Please note that the windows-user-data secret will be created by the WMCO lazily when it is configuring the first
          Windows Machine. After that, the windows-user-data will be available for the subsequent MachineSets to be consumed.
          It might take around 10 minutes for the Windows VM to be configured so that it joins the cluster. Please note that
          the MachineSet should have following labels:
          * `machine.openshift.io/os-id: Windows`
          * `machine.openshift.io/cluster-api-machine-role: worker`
          * `machine.openshift.io/cluster-api-machine-type: worker`

          The following label must be added to the Machine spec within the MachineSet spec:
          * `node-role.kubernetes.io/worker: ""`

          Not having these labels will result in the Windows node not being marked as a worker.

          `<infrastructureID>` should be replaced with the output of:
          ```
          oc get -o jsonpath='{.status.infrastructureName}{"\n"}' infrastructure cluster
          ```

          The following template variables need to be replaced as follows with values from your vSphere environment:
          * `<Windows_VM_template>`: template name
          * `<VM Network Name>`: network name
          * `<vCenter DataCenter Name>`: datacenter name
          * `<Path to VM Folder in vCenter>`: path where your OpenShift cluster is running
          * `<vCenter Datastore Name>`: datastore name
          * `<vCenter Server FQDN/IP>`: IP address or FQDN of the vCenter server

          Please note that on vSphere, Windows Machine names cannot be more than 15 characters long. The MachineSet name therefore
          cannot be more than 9 characters long, due to the way Machine names are generated from it.
          ```
          apiVersion: machine.openshift.io/v1beta1
          kind: MachineSet
          metadata:
            labels:
              machine.openshift.io/cluster-api-cluster: <infrastructureID>
            name: winworker
            namespace: openshift-machine-api
          spec:
            replicas: 1
            selector:
              matchLabels:
                machine.openshift.io/cluster-api-cluster: <infrastructureID>
                machine.openshift.io/cluster-api-machineset: winworker
            template:
              metadata:
                labels:
                  machine.openshift.io/cluster-api-cluster: <infrastructureID>
                  machine.openshift.io/cluster-api-machine-role: worker
                  machine.openshift.io/cluster-api-machine-type: worker
                  machine.openshift.io/cluster-api-machineset: winworker
                  machine.openshift.io/os-id: Windows
              spec:
                metadata:
                  labels:
                    node-role.kubernetes.io/worker: ""
                providerSpec:
                  value:
                    apiVersion: vsphereprovider.openshift.io/v1beta1
                    credentialsSecret:
                      name: vsphere-cloud-credentials
                      namespace: openshift-machine-api
                    diskGiB: 128
                    kind: VSphereMachineProviderSpec
                    memoryMiB: 16384
                    metadata:
                      creationTimestamp: null
                    network:
                      devices:
                      - networkName: "<VM Network Name>"
                    numCPUs: 4
                    numCoresPerSocket: 1
                    snapshot: ""
                    template: <Windows_VM_template
                    userDataSecret:
                      name: windows-user-data
                    workspace:
                      datacenter: <vCenter DataCenter Name>
                      datastore: <vCenter Datastore Name>
                      folder: <Path to VM Folder in vCenter> # e.g. /DC/vm/ocp45-2tdrm
                      server: <vCenter Server FQDN/IP>
          ```
          Example MachineSet for other cloud providers:
          - [AWS](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-aws.html)
          - [Azure](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-azure.html)
          - [GCP](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-gcp.html)

          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.

          ### Reporting issues
          Support for this distribution of WMCO requires a Red Hat OpenShfit subscription. Support should be requested through the Red Hat Customer Portal.
          Please read through the [troubleshooting document](https://docs.openshift.com/container-platform/4.16/support/troubleshooting/troubleshooting-windows-container-workload-issues.html)
          before opening a support case.
        displayName: Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4-wincw/windows-machine-config-rhel9-operator@sha256:c9a6166158e1ccb1764007d79c6e0f879505c3546eb45cbbe79ebeefcc78a728
        version: 10.16.2
      entries:
      - name: windows-machine-config-operator.v10.16.2
        version: 10.16.2
      - name: windows-machine-config-operator.v10.16.1
        version: 10.16.1
      - name: windows-machine-config-operator.v10.16.0
        version: 10.16.0
      name: stable
    defaultChannel: stable
    packageName: windows-machine-config-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-documentdb-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-documentdb-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBInstance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/documentdb-controller:1.1.1
          createdAt: "2025-09-19T22:13:37Z"
          description: AWS DocumentDB controller is a service controller for managing
            DocumentDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DBCluster represents the state of an AWS documentdb DBCluster
              resource.
            displayName: DBCluster
            kind: DBCluster
            name: dbclusters.documentdb.services.k8s.aws
            version: v1alpha1
          - description: DBInstance represents the state of an AWS documentdb DBInstance
              resource.
            displayName: DBInstance
            kind: DBInstance
            name: dbinstances.documentdb.services.k8s.aws
            version: v1alpha1
          - description: DBSubnetGroup represents the state of an AWS documentdb DBSubnetGroup
              resource.
            displayName: DBSubnetGroup
            kind: DBSubnetGroup
            name: dbsubnetgroups.documentdb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon DocumentDB resources in AWS from within your Kubernetes cluster.

          **About Amazon DocumentDB**

          Amazon DocumentDB (with MongoDB compatibility) is a fast, reliable, and fully managed database service. Amazon DocumentDB makes it easy to set up, operate, and scale MongoDB-compatible databases in the cloud. With Amazon DocumentDB, you can run the same application code and use the same drivers and tools that you use with MongoDB.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon DocumentDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - documentdb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon DocumentDB Developer Resources
          url: https://aws.amazon.com/documentdb/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: documentdb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/documentdb-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-documentdb-controller.v1.1.1
        version: 1.1.1
      - name: ack-documentdb-controller.v1.1.0
        version: 1.1.0
      - name: ack-documentdb-controller.v1.0.14
        version: 1.0.14
      - name: ack-documentdb-controller.v1.0.13
        version: 1.0.13
      - name: ack-documentdb-controller.v1.0.12
        version: 1.0.12
      - name: ack-documentdb-controller.v1.0.11
        version: 1.0.11
      - name: ack-documentdb-controller.v1.0.10
        version: 1.0.10
      - name: ack-documentdb-controller.v1.0.9
        version: 1.0.9
      - name: ack-documentdb-controller.v1.0.7
        version: 1.0.7
      - name: ack-documentdb-controller.v1.0.6
        version: 1.0.6
      - name: ack-documentdb-controller.v1.0.5
        version: 1.0.5
      - name: ack-documentdb-controller.v1.0.2
        version: 1.0.2
      - name: ack-documentdb-controller.v1.0.1
        version: 1.0.1
      - name: ack-documentdb-controller.v1.0.0
        version: 1.0.0
      - name: ack-documentdb-controller.v0.0.9
        version: 0.0.9
      - name: ack-documentdb-controller.v0.0.8
        version: 0.0.8
      - name: ack-documentdb-controller.v0.0.7
        version: 0.0.7
      - name: ack-documentdb-controller.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: ack-documentdb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NGINX Inc
      provider-url: ""
    name: nginx-ingress-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nginx-ingress-operator.v3.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.nginx.org/v1alpha1",
                "kind": "NginxIngress",
                "metadata": {
                  "name": "nginxingress-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "annotations": {},
                    "appprotect": {
                      "enable": false
                    },
                    "appprotectdos": {
                      "debug": false,
                      "enable": false,
                      "maxDaemons": 0,
                      "maxWorkers": 0,
                      "memory": 0
                    },
                    "autoscaling": {
                      "annotations": {},
                      "behavior": {},
                      "enabled": false,
                      "maxReplicas": 3,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 50,
                      "targetMemoryUtilizationPercentage": 50
                    },
                    "config": {
                      "annotations": {},
                      "entries": {}
                    },
                    "containerPort": {
                      "http": 80,
                      "https": 443
                    },
                    "customConfigMap": "",
                    "customPorts": [],
                    "defaultHTTPListenerPort": 80,
                    "defaultHTTPSListenerPort": 443,
                    "defaultTLS": {
                      "cert": "",
                      "key": "",
                      "secret": ""
                    },
                    "disableIPV6": false,
                    "dnsPolicy": "ClusterFirst",
                    "enableCertManager": false,
                    "enableCustomResources": true,
                    "enableExternalDNS": false,
                    "enableLatencyMetrics": false,
                    "enableOIDC": false,
                    "enableSSLDynamicReload": true,
                    "enableSnippets": false,
                    "enableTLSPassthrough": false,
                    "env": [],
                    "extraContainers": [],
                    "globalConfiguration": {
                      "create": false,
                      "spec": {}
                    },
                    "healthStatus": false,
                    "healthStatusURI": "/nginx-health",
                    "hostNetwork": false,
                    "hostPort": {
                      "enable": false,
                      "http": 80,
                      "https": 443
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "nginx/nginx-ingress",
                      "tag": "5.2.0-ubi"
                    },
                    "ingressClass": {
                      "create": true,
                      "name": "nginx",
                      "setAsDefaultIngress": false
                    },
                    "initContainerResources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "initContainers": [],
                    "kind": "deployment",
                    "lifecycle": {},
                    "logFormat": "glog",
                    "logLevel": "info",
                    "mgmt": {
                      "licenseTokenSecretName": "license-token"
                    },
                    "minReadySeconds": 0,
                    "name": "controller",
                    "nginxDebug": false,
                    "nginxReloadTimeout": 60000,
                    "nginxStatus": {
                      "allowCidrs": "127.0.0.1",
                      "enable": true,
                      "port": 8080
                    },
                    "nginxplus": false,
                    "pod": {
                      "annotations": {},
                      "extraLabels": {}
                    },
                    "podDisruptionBudget": {
                      "annotations": {},
                      "enabled": false
                    },
                    "readOnlyRootFilesystem": false,
                    "readyStatus": {
                      "enable": true,
                      "initialDelaySeconds": 0,
                      "port": 8081
                    },
                    "replicaCount": 1,
                    "reportIngressStatus": {
                      "annotations": {},
                      "enable": true,
                      "enableLeaderElection": true,
                      "ingressLink": "",
                      "leaderElectionLockName": "nginx-ingress-leader"
                    },
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "selectorLabels": {},
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "create": true,
                      "customPorts": [],
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraLabels": {},
                      "httpPort": {
                        "enable": true,
                        "port": 80,
                        "targetPort": 80
                      },
                      "httpsPort": {
                        "enable": true,
                        "port": 443,
                        "targetPort": 443
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "type": "LoadBalancer"
                    },
                    "serviceAccount": {
                      "annotations": {},
                      "imagePullSecretName": "",
                      "imagePullSecretsNames": []
                    },
                    "shareProcessNamespace": false,
                    "strategy": {},
                    "terminationGracePeriodSeconds": 30,
                    "tlsPassthroughPort": 443,
                    "tolerations": [],
                    "volumeMounts": [],
                    "volumes": [],
                    "watchNamespace": "",
                    "watchNamespaceLabel": "",
                    "watchSecretNamespace": "",
                    "wildcardTLS": {
                      "cert": "",
                      "key": "",
                      "secret": ""
                    }
                  },
                  "nginxServiceMesh": {
                    "enable": false,
                    "enableEgress": false
                  },
                  "prometheus": {
                    "create": true,
                    "port": 9113,
                    "scheme": "http",
                    "secret": "",
                    "service": {
                      "create": false,
                      "labels": {
                        "service": "nginx-ingress-prometheus-service"
                      }
                    },
                    "serviceMonitor": {
                      "create": false,
                      "endpoints": [
                        {
                          "port": "prometheus"
                        }
                      ],
                      "labels": {},
                      "selectorMatchLabels": {
                        "service": "nginx-ingress-prometheus-service"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "serviceInsight": {
                    "create": false,
                    "port": 9114,
                    "scheme": "http",
                    "secret": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking
          certified: "true"
          containerImage: quay.io/nginx/nginx-ingress-operator@sha256:fb372438634b4a667218e725091bd448a253e7a9c40e7dda1044a318780d995f
          createdAt: "2025-09-15T16:08:18Z"
          description: The NGINX Ingress Operator is a Kubernetes/OpenShift component
            which deploys and manages one or more NGINX/NGINX Plus Ingress Controllers
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nginx-ingress
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/nginx/nginx-ingress-helm-operator
          support: NGINX Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The `NginxIngress` Custom Resource is the definition of a
              deployment of the Ingress Controller.
            displayName: Nginx Ingress Controller
            kind: NginxIngress
            name: nginxingresses.charts.nginx.org
            version: v1alpha1
        description: The NGINX Ingress Operator is a Kubernetes/OpenShift component
          which deploys and manages one or more NGINX/NGINX Plus Ingress Controllers
        displayName: Nginx Ingress Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nginx
        - ingress-controller
        - ingress
        - controller
        - kubernetes
        - openshift
        links:
        - name: Nginx Ingress Operator
          url: https://github.com/nginx/nginx-ingress-helm-operator
        maintainers:
        - email: kubernetes@nginx.com
          name: NGINX Inc
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: NGINX Inc
        relatedImages:
        - quay.io/nginx/nginx-ingress-operator@sha256:fb372438634b4a667218e725091bd448a253e7a9c40e7dda1044a318780d995f
        - quay.io/brancz/kube-rbac-proxy@sha256:754ab2a723c8fc7ea8e09ae1224f8af5be91849b8b2f2a3735c6237f1c534187
        version: 3.3.0
      entries:
      - name: nginx-ingress-operator.v3.3.0
        version: 3.3.0
      name: alpha
    defaultChannel: alpha
    packageName: nginx-ingress-operator
    provider:
      name: NGINX Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: QuestLabs
      provider-url: https://www.questlabs.in
    name: alloy
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: alloy.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.alloys.com/v1alpha1",
                "kind": "Alloy",
                "metadata": {
                  "name": "alloy"
                },
                "spec": {
                  "acmeEmail": "",
                  "dnsFQDN": "",
                  "sslTermination": "true"
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Alloy
            name: alloys.charts.alloys.com
            version: v1alpha1
        description: |-
          Alloy is platform to build trusted data in OSDU, available on-premise or in the cloud. It is equipped with built-in data quality rules derived from PPDM and OSDU standards that control the ingestion of data.

              • Applicable for conventional and new energy streams
              • Highly Scalable

              Benefits
              • Helps to build robust applications/solutions on stable data foundation using full data lifecycle APIs & built-in data/business rules
              • Help data governance team to monitor/oversea data across organisation and intervene when required
              • Substantially low implementation time for any new data entity or new domain (solar, wind etc.) because of modular rules packages
              • Capability to deploy workflows for complex corporate specific data quality and enrichment requirements
              • Spend less time & resources to deploy fit-for-purpose data
              • Schema based checks out of the box for conventional and new energy streams.
              • Sustainable and future proof


              Prerequisites
              1. Please create alloy project before installing alloy operator

          Getting Help

          Once Alloy is successfully installed you can also browse documentation at https://your.domain.com/alloy/docs/introduction
          Please visit https://www.questlabs.in/alloy for more details and if you have any queries or suggestions, mail us at info@questlabs.in
        displayName: Alloy
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - alloy
        links:
        - name: alloy
          url: https://www.questlabs.in
        maintainers:
        - email: info@questlabs.in
          name: QuestLabs
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: QuestLabs
          url: https://www.questlabs.in
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d291e0d38bd95eb2731a10bd5ff06bfa07c6621528d3c2bbe888397d87b3feaa
        - alloy.azurecr.io/alloy@sha256:bf29f69889d408507cf1a9a940beb80d95ba9faff1da2d70a4aea82cc235d96d
        version: 1.0.4
      entries:
      - name: alloy.v1.0.4
        version: 1.0.4
      - name: alloy.v1.0.3
        version: 1.0.3
      - name: alloy.v1.0.2
        version: 1.0.2
      - name: alloy.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: alloy
    provider:
      name: QuestLabs
      url: https://www.questlabs.in
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: SynaXG Engineering Team
      provider-url: ""
    name: synaxg-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: synaxg-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "card.synaxg.app.com/v1",
                "kind": "SG3Reset",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "synaxg-operator",
                    "app.kubernetes.io/instance": "sg3reset-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sg3reset",
                    "app.kubernetes.io/part-of": "synaxg-operator"
                  },
                  "name": "sg3reset-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/synaxg/synaxg-reset-utility@sha256:4050d5731a21bbec6e26165b8dc2b7632577ffa376e7eb056aa3245cf9574753",
                  "sg3ResetList": [
                    {
                      "nodeName": "test",
                      "pciList": [
                        {
                          "enable": "false",
                          "pciAddr": "000d:01:00.0"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "card.synaxg.app.com/v1",
                "kind": "SG3Update",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "synaxg-operator",
                    "app.kubernetes.io/instance": "sg3update-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sg3update",
                    "app.kubernetes.io/part-of": "synaxg-operator"
                  },
                  "name": "sg3update-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/synaxg/synaxg-upgrade-utility@sha256:7ffec3249edd2467f2414c2ce4232e8e436278fd1b43a51b66f54b6ea10a60d1",
                  "sg3UpdateList": [
                    {
                      "nodeName": "test",
                      "pciList": [
                        {
                          "enable": "false",
                          "pciAddr": "000d:01:00.0",
                          "sdkImageAddr": "quay.io/synaxgcom/sdk-img@sha256:1ed57013f278e5a858f71f343649652dbd52b6f3cf451374e45b6ac427fa76f9"
                        }
                      ]
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/synaxg/synaxg-operator-utility@sha256:3b22a8022a97cb1ffdb47ffcc9d9d562601f48669aa14fd58cd8ba6128b39bd7
          createdAt: "2025-07-01T03:42:07Z"
          description: SynaXG Operator will work with the SuperMicro Server + SG3
            card  and Openshift Container Platform.The Marvell kernel module “octeon_ep”
            and “octeon_ep_vf” are necessary to the Operator. The Operator can send
            command of “reset” and “update” to SG3 board via the VF channel.As a result,
            users can realize the process of “reset card” and “update firmware” on
            the web GUI, and monitor the executing result.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: free text
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: SynaXG
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SG3Reset is the Schema for the sg3resets API
            displayName: SG3 Reset
            kind: SG3Reset
            name: sg3resets.card.synaxg.app.com
            version: v1
          - description: SG3Update is the Schema for the sg3updates API
            displayName: SG3 Update
            kind: SG3Update
            name: sg3updates.card.synaxg.app.com
            version: v1
        description: SynaXG Operator will work with the SuperMicro Server + SG3 card  and
          Openshift Container Platform.The Marvell kernel module “octeon_ep” and “octeon_ep_vf”
          are necessary to the Operator. The Operator can send command of “reset”
          and “update” to SG3 board via the VF channel.As a result, users can realize
          the process of “reset card” and “update firmware” on the web GUI, and monitor
          the executing result.
        displayName: SynaXG vDU card 102X Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vDU
        - 102X
        - SynaXG
        links:
        - name: Synaxg Operator
          url: https://synaxg-operator.domain
        maintainers:
        - email: notice@synaxg.com
          name: Mantosh Malhotra
        maturity: alpha
        provider:
          name: SynaXG Engineering Team
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:38ee3484ba60bb474e7231e9d0718157386716cea7ffd6da6f32ddd551a5243d
        - registry.connect.redhat.com/synaxg/synaxg-operator-utility@sha256:3b22a8022a97cb1ffdb47ffcc9d9d562601f48669aa14fd58cd8ba6128b39bd7
        - registry.connect.redhat.com/synaxg/synaxg-reset-utility@sha256:4050d5731a21bbec6e26165b8dc2b7632577ffa376e7eb056aa3245cf9574753
        - registry.connect.redhat.com/synaxg/synaxg-upgrade-utility@sha256:7ffec3249edd2467f2414c2ce4232e8e436278fd1b43a51b66f54b6ea10a60d1
        - quay.io/synaxgcom/sdk-img@sha256:1ed57013f278e5a858f71f343649652dbd52b6f3cf451374e45b6ac427fa76f9
        version: 1.2.0
      entries:
      - name: synaxg-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: synaxg-operator
    provider:
      name: SynaXG Engineering Team
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-opa-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opa-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:08f995a39db5c2c6a17008cffa5216bb9bebd784d13d2b94154474d727fc796f
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/opa-operator@sha256:08f995a39db5c2c6a17008cffa5216bb9bebd784d13d2b94154474d727fc796f
        version: 24.11.1
      entries:
      - name: opa-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: opa-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:826aa12df4531d93a337ba70bd0c0d1e2e8062efe6954fc22797f6474c7732ca
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/opa-operator@sha256:826aa12df4531d93a337ba70bd0c0d1e2e8062efe6954fc22797f6474c7732ca
        version: 25.3.0
      entries:
      - name: opa-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: opa-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaClusterSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
        version: 25.7.0
      entries:
      - name: opa-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: opa-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaClusterSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
        version: 25.7.0
      entries:
      - name: opa-operator.v25.7.0
        version: 25.7.0
      - name: opa-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-opa-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: datagrid
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datagrid-operator.v8.2.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "adminAuth": {
                    "secretName": "basic-auth"
                  },
                  "clusterName": "example-infinispan",
                  "name": "mycache"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC batch-cache"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:1435aa60e8ab4d25c86daae68dcb93d14999f50d7c1e11a3c5726a86d3d30453
          createdAt: "2021-08-24T17:39:44Z"
          description: Create and manage Red Hat Data Grid clusters.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Restore API schema.
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Interoperability:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Data Grid Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
          Red Hat Data Grid is a distributed, in-memory data store that increases application performance and delivers open-source capabilities to handle demanding use cases.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.2/html/data_grid_operator_8.2_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.2
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:1435aa60e8ab4d25c86daae68dcb93d14999f50d7c1e11a3c5726a86d3d30453
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4d8057709958ef60bcabdcec52f05eae8df121fd3f4762c3218abca38130f638
        - registry.redhat.io/datagrid/datagrid-8-openj9-11-rhel8@sha256:c8607d6538a7528e27e6ecb138e0a93a03b3c9fa1f8a9a5aad3fd9f4a6814448
        version: 8.2.8
      entries:
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.2.x
    - currentCSV: datagrid-operator.v8.3.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:4d9d7221b029866837add5f3cb52dd0b7956f57996ce2d8b76344ce2de678716
          createdAt: "2022-11-03T08:40:57Z"
          description: Create and manage Red Hat Data Grid clusters.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.3/html/data_grid_operator_8.3_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.3
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:4d9d7221b029866837add5f3cb52dd0b7956f57996ce2d8b76344ce2de678716
        version: 8.3.9
      entries:
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.3.x
    - currentCSV: datagrid-operator.v8.4.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:e5315682869de72514e3e027bfb02bf0bb64919453ca738c7ecfa1e34634ac2b
          createdAt: 2024-05-30T08:05:19 EDT
          description: Create and manage Red Hat Data Grid clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.4/html/data_grid_operator_8.4_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.4
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:985193caf9bc3698c4d85aeedd1cfea09bccda881636d67885387b2a83c65ebd
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9d9e466b7c367d0406406550c74bc790aa68f3f875ea4f652eb28f2e25a05458
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:b56269cb849afca34e7f82cc8e01a44bb32af6126f24db28019a58e556d4cb3c
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:d522a20ca6909bd464f85967d8b86fe42b83f773478a80785f9a85a0cb87dfed
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bf7760520f5bea151a531d538d525c4591c47166995aefb4d77fef138de11544
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bcb040bf640691568a93a4c71c8072c37ffb76b4296ade83186f279862b7f197
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:24cf79efa89bbfa495665b1711a27531b7ab0813695a89b2c38c96c2e792eb24
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:532dca6facb584c1261c0480c499af9f857c70dc390e0d46e46afec2561468af
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:78a6bed6fcb6962e5689b92806789b72a66db21b7c607b392a573c98593f2c09
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:ae852e4679978725bb7d24c1bdfe25b40bd883130ed04a36d4fca49dbea024ef
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:424b30dccd773f69c079b35a42e8c3e339246c5d490e14e3fcfb8bd8144a9160
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9223b4b5c1c866a91033dac40f30a4caae0f6946cdb64935a3fdfffdb9ede660
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:c15d4fd827f554291c14851f187947f4109c0e20257fa2f1e141b32669569358
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:54bc26a0fddc160a4e0d1a813aec78abb3c9cb77db5039b74c23de759faa67f5
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:554b831a0803d55a4e3b7022011b5a5f6d2b4b65009065a9b28bc8e55fc3ad46
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:e5315682869de72514e3e027bfb02bf0bb64919453ca738c7ecfa1e34634ac2b
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4dbe1b7a1a36990c0a31d5a66f3cc4ca2954c9e1ec3b87431113a4111530277d
        version: 8.4.16
      entries:
      - name: datagrid-operator.v8.4.16
        version: 8.4.16
      - name: datagrid-operator.v8.4.15
        version: 8.4.15
      - name: datagrid-operator.v8.4.14
        version: 8.4.14
      - name: datagrid-operator.v8.4.13
        version: 8.4.13
      - name: datagrid-operator.v8.4.12
        version: 8.4.12
      - name: datagrid-operator.v8.4.11
        version: 8.4.11
      - name: datagrid-operator.v8.4.10
        version: 8.4.9
      - name: datagrid-operator.v8.4.8
        version: 8.4.8
      - name: datagrid-operator.v8.4.7
        version: 8.4.7
      - name: datagrid-operator.v8.4.6
        version: 8.4.6
      - name: datagrid-operator.v8.4.5
        version: 8.4.5
      - name: datagrid-operator.v8.4.4
        version: 8.4.4
      - name: datagrid-operator.v8.4.3
        version: 8.4.3
      - name: datagrid-operator.v8.4.2
        version: 8.4.2
      - name: datagrid-operator.v8.4.1
        version: 8.4.1
      - name: datagrid-operator.v8.4.0
        version: 8.4.0
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.4.x
    - currentCSV: datagrid-operator.v8.5.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel9-operator@sha256:d719d778b0848588a7e61d3c3fa689763f51a6533af4cf9e1c738c4ed8fac760
          createdAt: 2025-08-29T05:44:37 UTC
          description: Create and manage Red Hat Data Grid clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.5/html/data_grid_operator_8.5_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.5
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:a711c50e489995fb60e5dec23ce286360743feda5066e4a0861d448aa4f2e903
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bf7760520f5bea151a531d538d525c4591c47166995aefb4d77fef138de11544
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bcb040bf640691568a93a4c71c8072c37ffb76b4296ade83186f279862b7f197
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:4b9bf39d2ff14bcfbd4628a399093703d87a2b2b9c8e74c3d7cfdff4b5167fa4
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:ae852e4679978725bb7d24c1bdfe25b40bd883130ed04a36d4fca49dbea024ef
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:d522a20ca6909bd464f85967d8b86fe42b83f773478a80785f9a85a0cb87dfed
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9223b4b5c1c866a91033dac40f30a4caae0f6946cdb64935a3fdfffdb9ede660
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:11bfe5c8109d26164cff428232b2001ab8df464f14a382661b936543336e2960
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:ad4f91f984b4b898012963d52e6ebcd166ec03e76e26eba3191ecb87f51fb18d
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:c15d4fd827f554291c14851f187947f4109c0e20257fa2f1e141b32669569358
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:5656cd188d054551489d45be996266d03e3faa8814b8bfc0f360cf0dceb83cbe
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:424b30dccd773f69c079b35a42e8c3e339246c5d490e14e3fcfb8bd8144a9160
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:879f3dcc24ffe795472779f777d955666ba4041c17e717d5e165cd7f12d7955a
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:985193caf9bc3698c4d85aeedd1cfea09bccda881636d67885387b2a83c65ebd
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:b56269cb849afca34e7f82cc8e01a44bb32af6126f24db28019a58e556d4cb3c
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:9bc0b730570ab55fbeca83075949f4caa96ed5c37637420a59cd4a7db01ea840
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:532dca6facb584c1261c0480c499af9f857c70dc390e0d46e46afec2561468af
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9d9e466b7c367d0406406550c74bc790aa68f3f875ea4f652eb28f2e25a05458
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:54bc26a0fddc160a4e0d1a813aec78abb3c9cb77db5039b74c23de759faa67f5
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:78a6bed6fcb6962e5689b92806789b72a66db21b7c607b392a573c98593f2c09
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4dbe1b7a1a36990c0a31d5a66f3cc4ca2954c9e1ec3b87431113a4111530277d
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:48cf13e71953cea6f62354289f0d8ece7a8a481ee0bb2c76eb9352d438db9ca3
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:0d368459c98c5788dd55d2add5b4b7d438b9a8c814ea8aee2222ea956699cbb6
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:c25e3152d6a6b6d5cd86d311bf567e594b9948440785209cd8d7986076a5a10d
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:f59ae21773dfc207454a709ac5f7aad2be6a6a20ac88da26c379edeed54537b8
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:24cf79efa89bbfa495665b1711a27531b7ab0813695a89b2c38c96c2e792eb24
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:6e1766784b684561666134d41dca5331b80d50e63695fe7b1a1d6d4cdb73f0bd
        - registry.redhat.io/datagrid/datagrid-8-rhel9-operator@sha256:d719d778b0848588a7e61d3c3fa689763f51a6533af4cf9e1c738c4ed8fac760
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:554b831a0803d55a4e3b7022011b5a5f6d2b4b65009065a9b28bc8e55fc3ad46
        version: 8.5.11
      entries:
      - name: datagrid-operator.v8.5.11
        version: 8.5.11
      - name: datagrid-operator.v8.5.10
        version: 8.5.10
      - name: datagrid-operator.v8.5.9
        version: 8.5.9
      - name: datagrid-operator.v8.5.8
        version: 8.5.8
      - name: datagrid-operator.v8.5.7
        version: 8.5.7
      - name: datagrid-operator.v8.5.6
        version: 8.5.6
      - name: datagrid-operator.v8.5.5
        version: 8.5.5
      - name: datagrid-operator.v8.5.4
        version: 8.5.4
      - name: datagrid-operator.v8.5.3
        version: 8.5.3
      - name: datagrid-operator.v8.5.2
        version: 8.5.2
      - name: datagrid-operator.v8.5.1
        version: 8.5.1
      - name: datagrid-operator.v8.5.0
        version: 8.5.0
      - name: datagrid-operator.v8.4.16
        version: 8.4.16
      - name: datagrid-operator.v8.4.15
        version: 8.4.15
      - name: datagrid-operator.v8.4.14
        version: 8.4.14
      - name: datagrid-operator.v8.4.13
        version: 8.4.13
      - name: datagrid-operator.v8.4.12
        version: 8.4.12
      - name: datagrid-operator.v8.4.11
        version: 8.4.11
      - name: datagrid-operator.v8.4.10
        version: 8.4.9
      - name: datagrid-operator.v8.4.8
        version: 8.4.8
      - name: datagrid-operator.v8.4.7
        version: 8.4.7
      - name: datagrid-operator.v8.4.6
        version: 8.4.6
      - name: datagrid-operator.v8.4.5
        version: 8.4.5
      - name: datagrid-operator.v8.4.4
        version: 8.4.4
      - name: datagrid-operator.v8.4.3
        version: 8.4.3
      - name: datagrid-operator.v8.4.2
        version: 8.4.2
      - name: datagrid-operator.v8.4.1
        version: 8.4.1
      - name: datagrid-operator.v8.4.0
        version: 8.4.0
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: stable
    defaultChannel: stable
    packageName: datagrid
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JohnSnowLABS
      provider-url: nlp.johnsnowlabs.com
    name: annotationlab
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: annotationlab.v5.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.johnsnowlabs.com/v1",
                "kind": "Annotationlab",
                "metadata": {
                  "name": "annotationlab"
                },
                "spec": {
                  "activelearning": {
                    "extraEnv": "",
                    "extraVolumeMounts": "",
                    "extraVolumes": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/johnsnowlabs/annotationlab",
                      "tag": "active-learning-5.7.0"
                    },
                    "lifecycleHooks": ""
                  },
                  "affinity": "podAntiAffinity:\n  requiredDuringSchedulingIgnoredDuringExecution:\n    - labelSelector:\n        matchLabels:\n          {{- include \"annotationlab.selectorLabels\" . | nindent 10 }}\n        matchExpressions:\n          - key: role\n            operator: NotIn\n            values:\n              - test\n      topologyKey: kubernetes.io/hostname\n  preferredDuringSchedulingIgnoredDuringExecution:\n    - weight: 100\n      podAffinityTerm:\n        labelSelector:\n          matchLabels:\n            {{- include \"annotationlab.selectorLabels\" . | nindent 12 }}\n          matchExpressions:\n            - key: role\n              operator: NotIn\n              values:\n                - test\n        topologyKey: failure-domain.beta.kubernetes.io/zone\n",
                  "airflow": {
                    "activelearning": {
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/johnsnowlabs/annotationlab",
                        "tag": "active-learning-5.7.0"
                      }
                    },
                    "affinity": {},
                    "airflowConfigAnnotations": {},
                    "airflowHome": "/opt/airflow",
                    "airflowLocalSettings": "{{- if semverCompare \"\u003e=2.2.0\" .Values.airflowVersion }}\n{{- if not (or .Values.webserverSecretKey .Values.webserverSecretKeySecretName) }}\nfrom airflow.www.utils import UIAlert\n\nDASHBOARD_UIALERTS = [\n  UIAlert(\n    'Usage of a dynamic webserver secret key detected. We recommend a static webserver secret key instead.'\n    ' See the \u003ca href='\n    '\"https://airflow.apache.org/docs/helm-chart/stable/production-guide.html#webserver-secret-key\"\u003e'\n    'Helm Chart Production Guide\u003c/a\u003e for more details.',\n    category=\"warning\",\n    roles=[\"Admin\"],\n    html=True,\n  )\n]\n{{- end }}\n{{- end }}",
                    "airflowPodAnnotations": {},
                    "airflowVersion": "2.5.1",
                    "allowPodLaunching": true,
                    "cleanup": {
                      "affinity": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow kubernetes cleanup-pods --namespace={{ .Release.Namespace }}"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "failedJobsHistoryLimit": null,
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "resources": {},
                      "schedule": "*/15 * * * *",
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "successfulJobsHistoryLimit": null,
                      "tolerations": [],
                      "topologySpreadConstraints": []
                    },
                    "config": {
                      "api": {
                        "auth_backend": "airflow.api.auth.backend.default"
                      },
                      "celery": {
                        "flower_url_prefix": "{{ .Values.ingress.flower.path }}",
                        "worker_concurrency": 16
                      },
                      "celery_kubernetes_executor": {
                        "kubernetes_queue": "kubernetes"
                      },
                      "core": {
                        "colored_console_log": "False",
                        "dags_are_paused_at_creation": "False",
                        "dags_folder": "{{ include \"airflow_dags\" . }}",
                        "executor": "{{ .Values.executor }}",
                        "load_examples": "False",
                        "remote_logging": "{{- ternary \"True\" \"False\" .Values.elasticsearch.enabled }}"
                      },
                      "elasticsearch": {
                        "json_format": "True",
                        "log_id_template": "{dag_id}_{task_id}_{execution_date}_{try_number}"
                      },
                      "elasticsearch_configs": {
                        "max_retries": 3,
                        "retry_timeout": "True",
                        "timeout": 30
                      },
                      "kerberos": {
                        "ccache": "{{ .Values.kerberos.ccacheMountPath }}/{{ .Values.kerberos.ccacheFileName }}",
                        "keytab": "{{ .Values.kerberos.keytabPath }}",
                        "principal": "{{ .Values.kerberos.principal }}",
                        "reinit_frequency": "{{ .Values.kerberos.reinitFrequency }}"
                      },
                      "kubernetes": {
                        "airflow_configmap": "{{ include \"airflow_config\" . }}",
                        "airflow_local_settings_configmap": "{{ include \"airflow_config\" . }}",
                        "multi_namespace_mode": "{{ ternary \"True\" \"False\" .Values.multiNamespaceMode }}",
                        "namespace": "{{ .Release.Namespace }}",
                        "pod_template_file": "{{ include \"airflow_pod_template_file\" . }}/pod_template_file.yaml",
                        "worker_container_repository": "{{ .Values.images.airflow.repository | default .Values.defaultAirflowRepository }}",
                        "worker_container_tag": "{{ .Values.images.airflow.tag | default .Values.defaultAirflowTag }}"
                      },
                      "logging": {
                        "colored_console_log": "False",
                        "remote_logging": "{{- ternary \"True\" \"False\" .Values.elasticsearch.enabled }}"
                      },
                      "metrics": {
                        "statsd_host": "{{ printf \"%s-statsd\" .Release.Name }}",
                        "statsd_on": "{{ ternary \"True\" \"False\" .Values.statsd.enabled }}",
                        "statsd_port": 9125,
                        "statsd_prefix": "airflow"
                      },
                      "scheduler": {
                        "run_duration": 41460,
                        "standalone_dag_processor": "{{ ternary \"True\" \"False\" .Values.dagProcessor.enabled }}",
                        "statsd_host": "{{ printf \"%s-statsd\" .Release.Name }}",
                        "statsd_on": "{{ ternary \"True\" \"False\" .Values.statsd.enabled }}",
                        "statsd_port": 9125,
                        "statsd_prefix": "airflow"
                      },
                      "webserver": {
                        "authenticate": "False",
                        "base_url": "http://0.0.0.0:8080",
                        "enable_proxy_fix": "True",
                        "rbac": "False"
                      }
                    },
                    "createUserJob": {
                      "affinity": {},
                      "annotations": {},
                      "applyCustomEnv": true,
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"users create\" \"create_user\" }} \"$@\"",
                        "--",
                        "-r",
                        "{{ .Values.webserver.defaultUser.role }}",
                        "-u",
                        "{{ .Values.webserver.defaultUser.username }}",
                        "-e",
                        "{{ .Values.webserver.defaultUser.email }}",
                        "-f",
                        "{{ .Values.webserver.defaultUser.firstName }}",
                        "-l",
                        "{{ .Values.webserver.defaultUser.lastName }}",
                        "-p",
                        "{{ .Values.webserver.defaultUser.password }}"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "jobAnnotations": {},
                      "labels": {},
                      "nodeSelector": {},
                      "resources": {},
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-create-user-job"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "useHelmHooks": true
                    },
                    "dagProcessor": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow dag-processor"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 60,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "dags": {
                      "gitSync": {
                        "branch": "v2-2-stable",
                        "containerName": "git-sync",
                        "depth": 1,
                        "enabled": false,
                        "env": [],
                        "extraVolumeMounts": [],
                        "maxFailures": 0,
                        "repo": "https://github.com/apache/airflow.git",
                        "resources": {},
                        "rev": "HEAD",
                        "securityContext": {},
                        "subPath": "tests/dags",
                        "uid": 65533,
                        "wait": 5
                      },
                      "persistence": {
                        "accessMode": "ReadWriteOnce",
                        "enabled": false,
                        "existingClaim": null,
                        "size": "1Gi",
                        "storageClassName": null,
                        "subPath": null
                      }
                    },
                    "data": {
                      "brokerUrl": null,
                      "brokerUrlSecretName": null,
                      "metadataConnection": {
                        "db": "airflow",
                        "pass": "airflow",
                        "port": 5432,
                        "protocol": "postgresql",
                        "sslmode": "disable",
                        "user": "airflow"
                      },
                      "metadataSecretName": null,
                      "resultBackendConnection": null,
                      "resultBackendSecretName": null
                    },
                    "defaultAirflowRepository": "apache/airflow",
                    "defaultAirflowTag": "2.5.1",
                    "elasticsearch": {
                      "connection": {},
                      "enabled": false,
                      "secretName": null
                    },
                    "enableBuiltInSecretEnvVars": {
                      "AIRFLOW_CONN_AIRFLOW_DB": true,
                      "AIRFLOW__CELERY__BROKER_URL": true,
                      "AIRFLOW__CELERY__CELERY_RESULT_BACKEND": true,
                      "AIRFLOW__CELERY__RESULT_BACKEND": true,
                      "AIRFLOW__CORE__FERNET_KEY": true,
                      "AIRFLOW__CORE__SQL_ALCHEMY_CONN": true,
                      "AIRFLOW__DATABASE__SQL_ALCHEMY_CONN": true,
                      "AIRFLOW__ELASTICSEARCH__ELASTICSEARCH_HOST": true,
                      "AIRFLOW__ELASTICSEARCH__HOST": true,
                      "AIRFLOW__WEBSERVER__SECRET_KEY": true
                    },
                    "env": [],
                    "executor": "CeleryExecutor",
                    "extraConfigMaps": {},
                    "extraEnv": "- name: ACTIVE_LEARNING_IMAGE_PULL_POLICY\n  value: \"{{ .Values.activelearning.image.pullPolicy }}\"\n- name: ACTIVE_LEARNING_IMAGE\n  value: \"{{ .Values.activelearning.image.repository }}:{{ .Values.activelearning.image.tag }}\"\n- name: NUMBER_OF_PREANNOTATOR_DAG\n  value: {{ .Values.preannotator_queues.count | quote }}\n- name: AIRFLOW__LOGGING__LOGGING_LEVEL\n  value: INFO\n- name: ANNOTATIONLAB_FULLNAME\n  value: {{ include \"annotationlab.annotationlab.fullname\" . }}\n- name: ANNOTATIONLAB_URL\n  value: http://{{ include \"annotationlab.annotationlab.fullname\" . }}:8200\n- name: AIRFLOW_NAMESPACE\n  value: {{ .Release.Namespace }}\n- name: DATABASE_CONNECTION_STRING\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"annotationlab.annotationlab.fullname\" . }}-secret\n      key: DATABASE_CONNECTION_STRING\n",
                    "extraEnvFrom": null,
                    "extraSecrets": {},
                    "fernetKey": null,
                    "fernetKeySecretName": null,
                    "flower": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"celery flower\" \"flower\" }}"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "networkPolicy": {
                        "ingress": {
                          "from": [],
                          "ports": [
                            {
                              "port": "{{ .Values.ports.flowerUI }}"
                            }
                          ]
                        }
                      },
                      "nodeSelector": {},
                      "password": null,
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "secretName": null,
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "loadBalancerIP": null,
                        "loadBalancerSourceRanges": [],
                        "ports": [
                          {
                            "name": "flower-ui",
                            "port": "{{ .Values.ports.flowerUI }}"
                          }
                        ],
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "username": null
                    },
                    "fullnameOverride": "",
                    "gid": 0,
                    "global": {},
                    "image": {
                      "repository": "docker.io/johnsnowlabs/annotationlab"
                    },
                    "images": {
                      "airflow": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": "dataflows-5.7.0"
                      },
                      "flower": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": null
                      },
                      "gitSync": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "k8s.gcr.io/git-sync/git-sync",
                        "tag": "v3.6.3"
                      },
                      "migrationsWaitTimeout": 60,
                      "pgbouncer": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "apache/airflow",
                        "tag": "airflow-pgbouncer-2021.04.28-1.14.0"
                      },
                      "pgbouncerExporter": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "apache/airflow",
                        "tag": "airflow-pgbouncer-exporter-2021.09.22-0.12.0"
                      },
                      "pod_template": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": null
                      },
                      "redis": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/redis",
                        "tag": "alpine3.16"
                      },
                      "statsd": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "quay.io/prometheus/statsd-exporter",
                        "tag": "v0.22.8"
                      },
                      "useDefaultImageForMigration": false
                    },
                    "ingress": {
                      "enabled": false,
                      "flower": {
                        "annotations": {},
                        "enabled": false,
                        "host": "",
                        "hosts": [],
                        "ingressClassName": "",
                        "path": "/",
                        "pathType": "ImplementationSpecific",
                        "tls": {
                          "enabled": false,
                          "secretName": ""
                        }
                      },
                      "web": {
                        "annotations": {},
                        "enabled": false,
                        "host": "",
                        "hosts": [],
                        "ingressClassName": "",
                        "path": "/",
                        "pathType": "ImplementationSpecific",
                        "precedingPaths": [],
                        "succeedingPaths": [],
                        "tls": {
                          "enabled": false,
                          "secretName": ""
                        }
                      }
                    },
                    "kerberos": {
                      "ccacheFileName": "cache",
                      "ccacheMountPath": "/var/kerberos-ccache",
                      "config": "# This is an example config showing how you can use templating and how \"example\" config\n# might look like. It works with the test kerberos server that we are using during integration\n# testing at Apache Airflow (see `scripts/ci/docker-compose/integration-kerberos.yml` but in\n# order to make it production-ready you must replace it with your own configuration that\n# Matches your kerberos deployment. Administrators of your Kerberos instance should\n# provide the right configuration.\n\n[logging]\ndefault = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kerberos_libs.log\"\nkdc = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kerberos_kdc.log\"\nadmin_server = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kadmind.log\"\n\n[libdefaults]\ndefault_realm = FOO.COM\nticket_lifetime = 10h\nrenew_lifetime = 7d\nforwardable = true\n\n[realms]\nFOO.COM = {\n  kdc = kdc-server.foo.com\n  admin_server = admin_server.foo.com\n}\n",
                      "configPath": "/etc/krb5.conf",
                      "enabled": false,
                      "keytabBase64Content": null,
                      "keytabPath": "/etc/airflow.keytab",
                      "principal": "airflow@FOO.COM",
                      "reinitFrequency": 3600
                    },
                    "kubeVersionOverride": "",
                    "labels": {},
                    "limits": [],
                    "logs": {
                      "persistence": {
                        "enabled": false,
                        "existingClaim": null,
                        "size": "100Gi",
                        "storageClassName": null
                      }
                    },
                    "migrateDatabaseJob": {
                      "affinity": {},
                      "annotations": {},
                      "applyCustomEnv": true,
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"db upgrade\" \"upgradedb\" }}"
                      ],
                      "command": null,
                      "enabled": true,
                      "extraContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "jobAnnotations": {},
                      "nodeSelector": {},
                      "resources": {},
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-migrate-database-job"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "useHelmHooks": true
                    },
                    "multiNamespaceMode": false,
                    "nameOverride": "",
                    "networkPolicies": {
                      "enabled": false
                    },
                    "nodeSelector": {},
                    "pgbouncer": {
                      "affinity": {},
                      "annotations": {},
                      "args": null,
                      "auth_file": "/etc/pgbouncer/users.txt",
                      "auth_type": "md5",
                      "ciphers": "normal",
                      "command": [
                        "pgbouncer",
                        "-u",
                        "nobody",
                        "/etc/pgbouncer/pgbouncer.ini"
                      ],
                      "configSecretName": null,
                      "enabled": false,
                      "extraIni": null,
                      "extraIniMetadata": null,
                      "extraIniResultBackend": null,
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "logConnections": 0,
                      "logDisconnections": 0,
                      "maxClientConn": 100,
                      "metadataPoolSize": 10,
                      "metricsExporterSidecar": {
                        "resources": {},
                        "sslmode": "disable"
                      },
                      "nodeSelector": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "resultBackendPoolSize": 5,
                      "revisionHistoryLimit": null,
                      "service": {
                        "extraAnnotations": {}
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "ssl": {
                        "ca": null,
                        "cert": null,
                        "key": null
                      },
                      "sslmode": "prefer",
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 65534,
                      "verbose": 0
                    },
                    "podTemplate": null,
                    "ports": {
                      "airflowUI": 8080,
                      "flowerUI": 5555,
                      "pgbouncer": 6543,
                      "pgbouncerScrape": 9127,
                      "redisDB": 6379,
                      "statsdIngest": 9125,
                      "statsdScrape": 9102,
                      "workerLogs": 8793
                    },
                    "postgresql": {
                      "auth": {
                        "enablePostgresUser": true,
                        "password": null,
                        "postgresPassword": "postgres",
                        "username": null
                      },
                      "common": {
                        "exampleValue": "common-chart",
                        "global": {
                          "postgresql": {}
                        }
                      },
                      "commonAnnotations": {},
                      "enabled": true,
                      "extraEnv": [],
                      "global": {
                        "postgresql": {}
                      },
                      "image": {
                        "debug": false,
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgresql",
                        "tag": "11"
                      },
                      "ldap": {
                        "baseDN": "",
                        "bindDN": "",
                        "enabled": false,
                        "port": "",
                        "prefix": "",
                        "scheme": "",
                        "search_attr": "",
                        "search_filter": "",
                        "server": "",
                        "suffix": "",
                        "tls": false,
                        "url": ""
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 30,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "master": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": [],
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "metrics": {
                        "enabled": false,
                        "extraEnvVars": {},
                        "image": {
                          "pullPolicy": "IfNotPresent",
                          "registry": "docker.io",
                          "repository": "bitnami/postgres-exporter",
                          "tag": "0.8.0-debian-10-r166"
                        },
                        "livenessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "prometheusRule": {
                          "additionalLabels": {},
                          "enabled": false,
                          "namespace": "",
                          "rules": []
                        },
                        "readinessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "securityContext": {},
                        "service": {
                          "annotations": {
                            "prometheus.io/port": "9187",
                            "prometheus.io/scrape": "true"
                          },
                          "type": "ClusterIP"
                        },
                        "serviceMonitor": {
                          "additionalLabels": {},
                          "enabled": false
                        }
                      },
                      "nameOverride": "airflow-postgresql",
                      "networkPolicy": {
                        "allowExternal": true,
                        "enabled": false,
                        "explicitNamespacesSelector": {}
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "enabled": true,
                        "mountPath": "/bitnami/postgresql",
                        "size": "8Gi",
                        "subPath": ""
                      },
                      "postgresqlDataDir": "/bitnami/postgresql/data",
                      "postgresqlDatabase": "airflow",
                      "postgresqlPassword": "airflow",
                      "postgresqlUsername": "airflow",
                      "psp": {
                        "create": false
                      },
                      "rbac": {
                        "create": false
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "replication": {
                        "applicationName": "my_application",
                        "enabled": false,
                        "numSynchronousReplicas": 0,
                        "password": "repl_password",
                        "slaveReplicas": 1,
                        "synchronousCommit": "off",
                        "user": "repl_user"
                      },
                      "resources": {
                        "requests": {
                          "cpu": "250m",
                          "memory": "256Mi"
                        }
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "port": 5432,
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "enabled": true,
                        "name": "annotationlab-airflow-postgresql"
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": true
                        },
                        "enabled": true
                      },
                      "slave": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "tls": {
                        "certFilename": "",
                        "certKeyFilename": "",
                        "certificatesSecret": "",
                        "enabled": false,
                        "preferServerCiphers": true
                      },
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "volumePermissions": {
                        "enabled": false,
                        "image": {
                          "pullPolicy": "Always",
                          "registry": "docker.io",
                          "repository": "bitnami/minideb",
                          "tag": "buster"
                        },
                        "securityContext": {}
                      }
                    },
                    "preannotator_queues": {
                      "count": 5
                    },
                    "quotas": {},
                    "rbac": {
                      "create": true,
                      "createSCCRoleBinding": false
                    },
                    "redis": {
                      "affinity": {},
                      "enabled": true,
                      "nodeSelector": {},
                      "password": "DefaulTRediS-P0ss",
                      "passwordSecretName": null,
                      "persistence": {
                        "annotations": {},
                        "enabled": true,
                        "size": "1Gi",
                        "storageClassName": null
                      },
                      "podAnnotations": {},
                      "resources": {},
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "annotationlab-airflow-redis"
                      },
                      "terminationGracePeriodSeconds": 600,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 0
                    },
                    "registry": {
                      "connection": {},
                      "secretName": "kubernetesdockerhub"
                    },
                    "registryCredentials": {
                      "pullPolicy": "IfNotPresent",
                      "registryUrl": "docker.io/annotationlab"
                    },
                    "revisionHistoryLimit": null,
                    "scheduler": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow scheduler"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "logGroomerSidecar": {
                        "args": [
                          "bash",
                          "/clean-logs"
                        ],
                        "command": null,
                        "enabled": true,
                        "resources": {},
                        "retentionDays": 15
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-scheduler"
                      },
                      "strategy": null,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "updateStrategy": null,
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "secret": [],
                    "securityContext": {},
                    "statsd": {
                      "affinity": {},
                      "args": [
                        "--statsd.mapping-config=/etc/statsd-exporter/mappings.yml"
                      ],
                      "enabled": false,
                      "extraMappings": [],
                      "extraNetworkPolicies": [],
                      "nodeSelector": {},
                      "overrideMappings": [],
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "securityContext": {},
                      "service": {
                        "extraAnnotations": {}
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-statsd"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 65534
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "triggerer": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow triggerer"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "annotationlab-airflow-triggerer"
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 60,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "uid": 50000,
                    "volumeMounts": [],
                    "volumes": [],
                    "webserver": {
                      "affinity": {},
                      "allowPodLogReading": true,
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow webserver"
                      ],
                      "command": null,
                      "defaultUser": {
                        "email": "admin@example.com",
                        "enabled": false,
                        "firstName": "admin",
                        "lastName": "user",
                        "password": "admin",
                        "role": "Admin",
                        "username": "admin"
                      },
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "failureThreshold": 20,
                        "initialDelaySeconds": 15,
                        "periodSeconds": 5,
                        "scheme": "HTTP",
                        "timeoutSeconds": 30
                      },
                      "networkPolicy": {
                        "ingress": {
                          "from": [],
                          "ports": [
                            {
                              "port": "{{ .Values.ports.airflowUI }}"
                            }
                          ]
                        }
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "readinessProbe": {
                        "failureThreshold": 20,
                        "initialDelaySeconds": 15,
                        "periodSeconds": 5,
                        "scheme": "HTTP",
                        "timeoutSeconds": 30
                      },
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "loadBalancerIP": null,
                        "loadBalancerSourceRanges": [],
                        "ports": [
                          {
                            "name": "airflow-ui",
                            "port": "{{ .Values.ports.airflowUI }}"
                          }
                        ],
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-webserver"
                      },
                      "strategy": null,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      },
                      "webserverConfig": null,
                      "webserverConfigConfigMapName": null
                    },
                    "webserverSecretKey": "f565cd64401775f93c820dc63fc3d4f6",
                    "webserverSecretKeySecretName": null,
                    "workers": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"celery worker\" \"worker\" }}"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "hostAliases": [],
                      "keda": {
                        "advanced": {},
                        "cooldownPeriod": 30,
                        "enabled": false,
                        "maxReplicaCount": 10,
                        "minReplicaCount": 0,
                        "namespaceLabels": {},
                        "pollingInterval": 5
                      },
                      "kerberosSidecar": {
                        "enabled": false,
                        "resources": {}
                      },
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "enabled": true,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "logGroomerSidecar": {
                        "args": [
                          "bash",
                          "/clean-logs"
                        ],
                        "command": null,
                        "enabled": true,
                        "resources": {},
                        "retentionDays": 15
                      },
                      "nodeSelector": {},
                      "persistence": {
                        "annotations": {},
                        "enabled": true,
                        "fixPermissions": false,
                        "size": "5Gi",
                        "storageClassName": null
                      },
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-worker"
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 600,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "updateStrategy": null,
                      "waitForMigrations": {
                        "env": []
                      }
                    }
                  },
                  "backup": {
                    "enable": false,
                    "files": true,
                    "image": {
                      "repository": null,
                      "tag": null
                    },
                    "s3_access_key": "",
                    "s3_bucket_fullpath": "",
                    "s3_secret_key": ""
                  },
                  "configuration": {
                    "FLASK_SECRET_KEY": "rpaQghA-9lmpLwB8qxW0FsW1TJyZZYuB60ppzswpTzM=",
                    "KEYCLOAK_CLIENT_ID": "annotationlab",
                    "KEYCLOAK_CLIENT_SECRET_KEY": "09a71c59-0351-4ce6-bc8f-8fd3feb9d2ff",
                    "KEYCLOAK_REALM_NAME": "master",
                    "PERSISTENT_STORAGE": "/projects"
                  },
                  "extraContainers": "",
                  "extraEnv": "",
                  "extraNetworkPolicies": "- namespaceSelector: {}\n",
                  "extraVolumeMounts": "",
                  "extraVolumes": "",
                  "github_oauth_origin": "https://github.oauth.johnsnowlabs.com",
                  "google_analytics": {
                    "api_secret": "WENfd00yeUxRRWl4N0Utbml5SWt3Zw==",
                    "measurement_id": "Ry1XODBIMDhTTFBZ"
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "docker.io/johnsnowlabs/annotationlab",
                    "tag": "5.7.0"
                  },
                  "ingress": {
                    "annotations": {},
                    "defaultBackend": false,
                    "enabled": true,
                    "hosts": [
                      {
                        "host": "annotationlab",
                        "path": "/"
                      }
                    ],
                    "ingressroute": {
                      "enabled": false
                    },
                    "servicePort": "http",
                    "tls": [],
                    "uploadLimitInMegabytes": 16
                  },
                  "installAirflow": true,
                  "installKeycloak": true,
                  "installPostgresql": true,
                  "keycloak": {
                    "affinity": "podAntiAffinity:\n  requiredDuringSchedulingIgnoredDuringExecution:\n    - labelSelector:\n        matchLabels:\n          {{- include \"keycloak.selectorLabels\" . | nindent 10 }}\n        matchExpressions:\n          - key: app.kubernetes.io/component\n            operator: NotIn\n            values:\n              - test\n      topologyKey: kubernetes.io/hostname\n  preferredDuringSchedulingIgnoredDuringExecution:\n    - weight: 100\n      podAffinityTerm:\n        labelSelector:\n          matchLabels:\n            {{- include \"keycloak.selectorLabels\" . | nindent 12 }}\n          matchExpressions:\n            - key: app.kubernetes.io/component\n              operator: NotIn\n              values:\n                - test\n        topologyKey: topology.kubernetes.io/zone\n",
                    "args": [],
                    "auththeme": {
                      "image": {
                        "repository": "docker.io/johnsnowlabs/annotationlab",
                        "tag": "auth-theme-5.7.0"
                      }
                    },
                    "autoscaling": {
                      "behavior": {
                        "scaleDown": {
                          "policies": [
                            {
                              "periodSeconds": 300,
                              "type": "Pods",
                              "value": 1
                            }
                          ],
                          "stabilizationWindowSeconds": 300
                        }
                      },
                      "enabled": false,
                      "labels": {},
                      "maxReplicas": 10,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 80,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 3
                    },
                    "cache": {
                      "stack": "default"
                    },
                    "clusterDomain": "cluster.local",
                    "command": [
                      "/opt/keycloak/bin/kc.sh",
                      "--verbose",
                      "start",
                      "--http-enabled=true",
                      "--http-port=8080",
                      "--hostname-strict=false",
                      "--hostname-strict-https=false",
                      "--spi-events-listener-jboss-logging-success-level=info",
                      "--spi-events-listener-jboss-logging-error-level=warn"
                    ],
                    "database": {
                      "enabled": true,
                      "postgresqlDatabase": "keycloak",
                      "postgresqlPassword": "keycloak",
                      "postgresqlUsername": "keycloak"
                    },
                    "dbchecker": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/busybox",
                        "tag": 1.32
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "32Mi"
                        },
                        "requests": {
                          "cpu": "20m",
                          "memory": "32Mi"
                        }
                      },
                      "securityContext": {}
                    },
                    "enableServiceLinks": true,
                    "extraContainers": "",
                    "extraEnv": "- name: KC_FEATURES\n  value: \"authorization,account-api,admin-api,admin-fine-grained-authz,ciba,impersonation,kerberos,preview,step-up-authentication,token-exchange,web-authn\"\n- name: KEYCLOAK_ADMIN\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"keycloak.fullname\" . }}-admincreds\n      key: user\n- name: KEYCLOAK_ADMIN_PASSWORD\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"keycloak.fullname\" . }}-admincreds\n      key: password\n- name: JAVA_OPTS_APPEND\n  value: \u003e-\n    -XX:+UseContainerSupport\n    -XX:MaxRAMPercentage=50.0\n    -Djava.awt.headless=true\n    -Djgroups.dns.query={{ include \"keycloak.fullname\" . }}-headless\n",
                    "extraEnvFrom": "",
                    "extraInitContainers": "- name: theme-provider\n  image: {{ .Values.auththeme.image.repository }}:{{ .Values.auththeme.image.tag }}\n  imagePullPolicy: IfNotPresent\n  command:\n    - sh\n  args:\n    - -c\n    - |\n      echo \"Copying theme...\"\n      cp -R /johnsnowlabs/* /theme\n  volumeMounts:\n    - name: theme\n      mountPath: /theme\n",
                    "extraPorts": [],
                    "extraServiceMonitor": {
                      "annotations": {},
                      "enabled": false,
                      "interval": "10s",
                      "labels": {},
                      "namespace": "",
                      "namespaceSelector": {},
                      "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/realms/master/metrics",
                      "port": "http",
                      "scrapeTimeout": "10s"
                    },
                    "extraVolumeMounts": "- name: admincreds\n  mountPath: /secrets/admincreds\n  readOnly: true\n- name: theme\n  mountPath: /opt/keycloak/themes/johnsnowlabs\n",
                    "extraVolumes": "- name: admincreds\n  secret:\n    secretName: '{{ include \"annotationlab.keycloak.fullname\" . }}-admincreds'\n- name: theme\n  emptyDir: {}\n",
                    "fullnameOverride": "",
                    "global": {},
                    "health": {
                      "enabled": true
                    },
                    "hostAliases": [],
                    "http": {
                      "relativePath": "/auth"
                    },
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/keycloak/keycloak",
                      "tag": "20.0.3"
                    },
                    "imagePullSecrets": [
                      {
                        "name": "kubernetesdockerhub"
                      }
                    ],
                    "ingress": {
                      "annotations": {
                        "nginx.ingress.kubernetes.io/proxy-buffer-size": "16k"
                      },
                      "console": {
                        "annotations": {},
                        "enabled": false,
                        "ingressClassName": "",
                        "rules": [
                          {
                            "host": "{{ .Release.Name }}.keycloak.example.com",
                            "paths": [
                              {
                                "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/admin",
                                "pathType": "Prefix"
                              }
                            ]
                          }
                        ],
                        "tls": []
                      },
                      "enabled": false,
                      "ingressClassName": "",
                      "labels": {},
                      "rules": [
                        {
                          "host": "keycloak.example.com",
                          "paths": [
                            {
                              "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/",
                              "pathType": "Prefix"
                            }
                          ]
                        }
                      ],
                      "servicePort": "http",
                      "tls": []
                    },
                    "lifecycleHooks": "",
                    "livenessProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health/live'\n  port: http\ninitialDelaySeconds: 0\ntimeoutSeconds: 5\n",
                    "metrics": {
                      "enabled": true
                    },
                    "nameOverride": "",
                    "networkPolicy": {
                      "egress": [],
                      "enabled": false,
                      "extraFrom": [],
                      "labels": {}
                    },
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podDisruptionBudget": {},
                    "podLabels": {},
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {},
                    "postgresql": {
                      "common": {
                        "exampleValue": "common-chart",
                        "global": {
                          "postgresql": {}
                        }
                      },
                      "commonAnnotations": {},
                      "enabled": true,
                      "extraEnv": [],
                      "global": {
                        "postgresql": {}
                      },
                      "image": {
                        "debug": false,
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgresql",
                        "tag": "11.20.0-debian-11-r12"
                      },
                      "ldap": {
                        "baseDN": "",
                        "bindDN": "",
                        "bind_password": null,
                        "enabled": false,
                        "port": "",
                        "prefix": "",
                        "scheme": "",
                        "search_attr": "",
                        "search_filter": "",
                        "server": "",
                        "suffix": "",
                        "tls": false,
                        "url": ""
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 30,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "master": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": [],
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "metrics": {
                        "enabled": false,
                        "extraEnvVars": {},
                        "image": {
                          "pullPolicy": "IfNotPresent",
                          "registry": "docker.io",
                          "repository": "bitnami/postgres-exporter",
                          "tag": "0.8.0-debian-10-r166"
                        },
                        "livenessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "prometheusRule": {
                          "additionalLabels": {},
                          "enabled": false,
                          "namespace": "",
                          "rules": []
                        },
                        "readinessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "securityContext": {},
                        "service": {
                          "annotations": {
                            "prometheus.io/port": "9187",
                            "prometheus.io/scrape": "true"
                          },
                          "loadBalancerIP": null,
                          "type": "ClusterIP"
                        },
                        "serviceMonitor": {
                          "additionalLabels": {},
                          "enabled": false
                        }
                      },
                      "nameOverride": "keycloak-postgres",
                      "networkPolicy": {
                        "allowExternal": true,
                        "enabled": false,
                        "explicitNamespacesSelector": {}
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "enabled": true,
                        "mountPath": "/bitnami/postgresql",
                        "size": "8Gi",
                        "subPath": ""
                      },
                      "postgresqlDataDir": "/bitnami/postgresql/data",
                      "postgresqlDatabase": "keycloak",
                      "postgresqlPassword": "keycloak",
                      "postgresqlUsername": "keycloak",
                      "psp": {
                        "create": false
                      },
                      "rbac": {
                        "create": false
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "replication": {
                        "applicationName": "my_application",
                        "enabled": false,
                        "numSynchronousReplicas": 0,
                        "password": "repl_password",
                        "slaveReplicas": 1,
                        "synchronousCommit": "off",
                        "user": "repl_user"
                      },
                      "resources": {
                        "requests": {
                          "cpu": "250m",
                          "memory": "256Mi"
                        }
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "port": 5432,
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "enabled": true,
                        "name": "annotationlab-keycloak-postgresql"
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": true
                        },
                        "enabled": true
                      },
                      "slave": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "tls": {
                        "certCAFilename": null,
                        "certFilename": "",
                        "certKeyFilename": "",
                        "certificatesSecret": "",
                        "crlFilename": null,
                        "enabled": false,
                        "preferServerCiphers": true
                      },
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "volumePermissions": {
                        "enabled": false,
                        "image": {
                          "pullPolicy": "Always",
                          "registry": "docker.io",
                          "repository": "bitnami/minideb",
                          "tag": "buster"
                        },
                        "securityContext": {}
                      }
                    },
                    "priorityClassName": "",
                    "prometheusRule": {
                      "annotations": {},
                      "enabled": false,
                      "labels": {},
                      "namespace": "",
                      "rules": []
                    },
                    "proxy": {
                      "enabled": true,
                      "mode": "edge"
                    },
                    "rbac": {
                      "create": false,
                      "rules": []
                    },
                    "readinessProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health/ready'\n  port: http\ninitialDelaySeconds: 10\ntimeoutSeconds: 1\n",
                    "replicas": 1,
                    "resources": {},
                    "restartPolicy": "Always",
                    "route": {
                      "annotations": {},
                      "enabled": false,
                      "host": "",
                      "labels": {},
                      "path": "/",
                      "tls": {
                        "enabled": true,
                        "insecureEdgeTerminationPolicy": "Redirect",
                        "termination": "edge"
                      }
                    },
                    "secrets": {
                      "admincreds": {
                        "stringData": {
                          "password": "secret",
                          "user": "admin"
                        }
                      }
                    },
                    "securityContext": {},
                    "service": {
                      "annotations": {},
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "httpNodePort": null,
                      "httpPort": 80,
                      "httpsNodePort": null,
                      "httpsPort": 8443,
                      "labels": {},
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "sessionAffinity": "",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "serviceAccount": {
                      "allowReadPods": false,
                      "annotations": {},
                      "automountServiceAccountToken": true,
                      "create": false,
                      "imagePullSecrets": [],
                      "labels": {},
                      "name": "annotationlab-keyclo"
                    },
                    "serviceHeadless": {
                      "annotations": {}
                    },
                    "serviceMonitor": {
                      "annotations": {},
                      "enabled": false,
                      "interval": "10s",
                      "labels": {},
                      "namespace": "",
                      "namespaceSelector": {},
                      "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/metrics",
                      "port": "http",
                      "scrapeTimeout": "10s"
                    },
                    "skipInitContainers": false,
                    "startupProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health'\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 1\nfailureThreshold: 60\nperiodSeconds: 5\n",
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {},
                    "terminationGracePeriodSeconds": 60,
                    "test": {
                      "deletionPolicy": "before-hook-creation",
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/seleniarm/standalone-chromium",
                        "tag": "111.0"
                      },
                      "podSecurityContext": {},
                      "securityContext": {}
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": null,
                    "updateStrategy": "RollingUpdate"
                  },
                  "licenseserver": {
                    "origin": "https://license.johnsnowlabs.com"
                  },
                  "lifecycleHooks": "# postStart:\n#   exec:\n#     command: [\"/bin/sh\", \"-c\", \"ls\"]\n",
                  "modelshub_origin": "https://modelshub.johnsnowlabs.com",
                  "myjohnsnowlabs": {
                    "client_id": "UIGgoiIL9szXW8Nnik0O7tPT4to87oVtWvYWfc6r",
                    "origin": "https://my.johnsnowlabs.com"
                  },
                  "networkPolicy": {
                    "enabled": false
                  },
                  "nodeSelector": {},
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "annotations": {},
                    "enabled": true,
                    "size": "8Gi"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "postgresql": {
                    "common": {
                      "exampleValue": "common-chart",
                      "global": {
                        "postgresql": {}
                      }
                    },
                    "commonAnnotations": {},
                    "extraEnv": [],
                    "global": {
                      "postgresql": {}
                    },
                    "image": {
                      "debug": false,
                      "pullPolicy": "IfNotPresent",
                      "registry": "docker.io",
                      "repository": "bitnami/postgresql",
                      "tag": "11.20.0-debian-11-r12"
                    },
                    "ldap": {
                      "baseDN": "",
                      "bindDN": "",
                      "bind_password": null,
                      "enabled": false,
                      "port": "",
                      "prefix": "",
                      "scheme": "",
                      "search_attr": "",
                      "search_filter": "",
                      "server": "",
                      "suffix": "",
                      "tls": false,
                      "url": ""
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "master": {
                      "affinity": {},
                      "annotations": {},
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podLabels": {},
                      "priorityClassName": "",
                      "service": {},
                      "sidecars": [],
                      "tolerations": []
                    },
                    "metrics": {
                      "enabled": false,
                      "extraEnvVars": {},
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgres-exporter",
                        "tag": "0.8.0-debian-10-r166"
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "prometheusRule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {
                          "prometheus.io/port": "9187",
                          "prometheus.io/scrape": "true"
                        },
                        "loadBalancerIP": null,
                        "type": "ClusterIP"
                      },
                      "serviceMonitor": {
                        "additionalLabels": {},
                        "enabled": false
                      }
                    },
                    "networkPolicy": {
                      "allowExternal": true,
                      "enabled": false,
                      "explicitNamespacesSelector": {}
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {},
                      "enabled": true,
                      "mountPath": "/bitnami/postgresql",
                      "size": "8Gi",
                      "subPath": ""
                    },
                    "postgresqlDataDir": "/bitnami/postgresql/data",
                    "postgresqlDatabase": "annotationlab",
                    "postgresqlPassword": "annotationlab",
                    "postgresqlUsername": "annotationlab",
                    "psp": {
                      "create": false
                    },
                    "rbac": {
                      "create": false
                    },
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 6,
                      "initialDelaySeconds": 5,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "replication": {
                      "applicationName": "my_application",
                      "enabled": false,
                      "numSynchronousReplicas": 0,
                      "password": "repl_password",
                      "slaveReplicas": 1,
                      "synchronousCommit": "off",
                      "user": "repl_user"
                    },
                    "resources": {
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {},
                    "service": {
                      "annotations": {},
                      "port": 5432,
                      "type": "ClusterIP"
                    },
                    "serviceAccount": {
                      "enabled": false
                    },
                    "shmVolume": {
                      "chmod": {
                        "enabled": true
                      },
                      "enabled": true
                    },
                    "slave": {
                      "affinity": {},
                      "annotations": {},
                      "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podLabels": {},
                      "priorityClassName": "",
                      "service": {},
                      "sidecars": [],
                      "tolerations": []
                    },
                    "tls": {
                      "certCAFilename": null,
                      "certFilename": "",
                      "certKeyFilename": "",
                      "certificatesSecret": "",
                      "crlFilename": null,
                      "enabled": false,
                      "preferServerCiphers": true
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "volumePermissions": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "Always",
                        "registry": "docker.io",
                        "repository": "bitnami/minideb",
                        "tag": "buster"
                      },
                      "securityContext": {}
                    }
                  },
                  "preannotator_queues": {
                    "count": 5,
                    "service": {
                      "port": 5000,
                      "portName": "http",
                      "type": "ClusterIP"
                    }
                  },
                  "priorityClassName": "",
                  "readinessProbe": "httpGet:\n  path: /healthz\n  port: http\ninitialDelaySeconds: 10\ntimeoutSeconds: 1\n",
                  "registryCredentials": {
                    "create": false,
                    "password": "XXXXX",
                    "secretName": "kubernetesdockerhub",
                    "username": "XXXXX"
                  },
                  "replicas": 1,
                  "resources": {},
                  "restartPolicy": "Always",
                  "securityContext": {
                    "enabled": false,
                    "fsGroup": 65534,
                    "fsGroupChangePolicy": "OnRootMismatch",
                    "runAsUser": 65534
                  },
                  "service": {
                    "port": 8200,
                    "portName": "http",
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": false,
                    "name": "annotationlab"
                  },
                  "sharedData": {
                    "enabled": true,
                    "images": {
                      "storageSize": "5Gi"
                    },
                    "storageSize": "100Gi"
                  },
                  "sqlalchemy_pool_size": 10,
                  "telemetry": {
                    "enabled": true
                  },
                  "terminationGracePeriodSeconds": 60,
                  "tolerations": [],
                  "updateStrategy": {
                    "type": "Recreate"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.johnsnowlabs.com/annotationlab/annotationlab-operator:v5.7.0
          createdAt: "2023-12-19T10:26:54Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/annotationlab/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/annotationlab/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: johnsnowlabs-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: NLP Lab
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Annotationlab
            name: annotationlabs.apps.johnsnowlabs.com
            version: v1
        description: "NLP Lab is a data labeling and annotation tool that helps you\nlabel
          different types of data using a simple interface with a standardized output\nformat.\n\nIt
          contains enterprise features like Project Management and sharing, SSO\nintegration,
          OpenAPI Specification 3.0 compliant API, task management, and active\nlearning/preannotation
          using Spark NLP.\n\nFurthermore, the production deployment of the application
          is done using a helm chart.\nThe same helm chart is also used to create
          an AWS AMI for an EC2 based single server \ndeployment using packer.\n\nMake
          sure you have defined shared storage such as efs/nfs/cephfs prior to installing
          the NLP Lab Operator.\nUpdate the host property in the YAML configuration
          to define the required domain name to use instead of the \ndefault hostname
          annotationlab.\n\nNLP Lab Operator Installtion guide [NLP Lab Operator][]:\n[NLP
          Lab Operator]: https://nlp.johnsnowlabs.com/docs/en/alab/install    \n"
        displayName: NLPLab
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nlplab
        links:
        - name: Annotationlab
          url: https://annotationlab.domain
        maintainers:
        - email: bhandari@johnsnowlabs.com
          name: bikash
        maturity: alpha
        provider:
          name: JohnSnowLABS
          url: nlp.johnsnowlabs.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - registry.johnsnowlabs.com/annotationlab/annotationlab-operator:v5.7.0
        version: 5.7.0
      entries:
      - name: annotationlab.v5.7.0
        version: 5.7.0
      - name: annotationlab.v5.4.1
        version: 5.4.1
      - name: annotationlab.v4.9.2
        version: 4.9.2
      name: alpha
    defaultChannel: alpha
    packageName: annotationlab
    provider:
      name: JohnSnowLABS
      url: nlp.johnsnowlabs.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.6.5-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6.5
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/kubeturbo-operator:8.6.5
        version: 8.6.5-beta.1
      entries:
      - name: kubeturbo-operator.v8.6.5-beta.1
        version: 8.6.5-beta.1
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: beta
    - currentCSV: kubeturbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/kubeturbo-operator:8.6
        version: 8.6.0
      entries:
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: stable
    - currentCSV: kubeturbo-operator.v8.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.4
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/kubeturbo-operator:8.4
        version: 8.4.0
      entries:
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.4.0
    - currentCSV: kubeturbo-operator.v8.5.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.5
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/kubeturbo-operator:8.5
        version: 8.5.0
      entries:
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.5.0
    - currentCSV: kubeturbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/kubeturbo-operator:8.6
        version: 8.6.0
      entries:
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.6.0
    defaultChannel: stable
    packageName: kubeturbo
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Silicom Ltd.
      provider-url: http://www.silicom-usa.com
    name: silicom-sts-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: silicom-sts-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 3,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-BC-8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/mode": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "hp-ocp",
                  "namespase": "openshift-operators"
                },
                "spec": {
                  "dummy": 0
                },
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicom-sts/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicom-sts/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/rmr/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - quay.io/silicom/gpsd@sha256:e479c542e061b5422865f2617ed0563794f265c6e6660e5a9d204c9e1cb2f346
        - quay.io/silicom/tsyncd@sha256:3f1cbd4a8b5d62d7d3e525af5f26d9f27b8c08e886b0639a1e368163e166b537
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - quay.io/silicom/sts-operator@sha256:7d474d0aa91ac57d8f5817b033b67b511c8601e0dbd83d808a69004bd24b4e60
        - quay.io/silicom/grpc-tsyncd@sha256:631bb754742e1a24376913baed60e480a6e6ae988ad9a3892254ff96996c15d1
        - quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
        - quay.io/silicom/phc2sys@sha256:b067167a1a5d93e9fe715c7774f6cdbbcd93cac27780f022e44dcb8a4d4d7bb9
        - quay.io/silicom/sts-plugin@sha256:233bc95f8617a06bd452548ec116d91551dd4933970860d8d5ccb81f75ad2220
        - quay.io/silicom/ice-driver-src@sha256:569c8333e21c7c022ea71fa5b418404899040d5b3a3a4128eeac21f62aacaf3e
        version: 0.1.3
      entries:
      - name: silicom-sts-operator.v0.1.3
        version: 0.1.3
      - name: silicom-sts-operator.v0.0.9
        version: 0.0.9
      - name: silicom-sts-operator.v0.0.8
        version: 0.0.8
      name: alpha
    - currentCSV: silicom-sts-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 3,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-BC-8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/mode": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "hp-ocp",
                  "namespase": "openshift-operators"
                },
                "spec": {
                  "dummy": 0
                },
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicom-sts/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicom-sts/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/rmr/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - quay.io/silicom/sts-plugin@sha256:233bc95f8617a06bd452548ec116d91551dd4933970860d8d5ccb81f75ad2220
        - quay.io/silicom/sts-operator@sha256:7d474d0aa91ac57d8f5817b033b67b511c8601e0dbd83d808a69004bd24b4e60
        - quay.io/silicom/tsyncd@sha256:3f1cbd4a8b5d62d7d3e525af5f26d9f27b8c08e886b0639a1e368163e166b537
        - quay.io/silicom/grpc-tsyncd@sha256:631bb754742e1a24376913baed60e480a6e6ae988ad9a3892254ff96996c15d1
        - quay.io/silicom/ice-driver-src@sha256:569c8333e21c7c022ea71fa5b418404899040d5b3a3a4128eeac21f62aacaf3e
        - quay.io/silicom/gpsd@sha256:e479c542e061b5422865f2617ed0563794f265c6e6660e5a9d204c9e1cb2f346
        - quay.io/silicom/phc2sys@sha256:b067167a1a5d93e9fe715c7774f6cdbbcd93cac27780f022e44dcb8a4d4d7bb9
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
        version: 0.1.3
      entries:
      - name: silicom-sts-operator.v0.1.3
        version: 0.1.3
      - name: silicom-sts-operator.v0.0.9
        version: 0.0.9
      - name: silicom-sts-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: silicom-sts-operator
    provider:
      name: Silicom Ltd.
      url: http://www.silicom-usa.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-flashsystemcluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "flashsystemcluster.odf.ibm.com/v1alpha1",
                  "name": "my-flashsystemcluster",
                  "namespace": "openshift-storage"
                }
              },
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-storagecluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "storagecluster.ocs.openshift.io/v1",
                  "name": "my-storagecluster",
                  "namespace": "openshift-storage"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-console"]'
          containerImage: registry.redhat.io/odf4/odf-rhel9-operator@sha256:8d377d2d8a94e461a6eca8d1c3b4f5cf1f9ec1ba6a03044304403b58bbbabaa4
          createdAt: "2025-09-12T07:34:03Z"
          description: OpenShift Data Foundation provides a common control plane for
            storage solutions on OpenShift Container Platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.16"}]'
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operator.openshift.io/uninstall-message: 'Warning: deleting the operand
            (StorageSystem) will result in data loss'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "odf.openshift.io/v1alpha1",
              "kind": "StorageSystem",
              "metadata": {
                "name": "ocs-storagecluster-storagesystem",
                "namespace": "openshift-storage"
              },
              "spec": {
                "kind": "storagecluster.ocs.openshift.io/v1",
                "name": "ocs-storagecluster",
                "namespace": "openshift-storage"
              }
            }
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/odf-operator
          support: Red Hat
          vendors.odf.openshift.io/kind: '["storagecluster.ocs.openshift.io/v1", "flashsystemcluster.odf.ibm.com/v1alpha1"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageSystem is the Schema for the storagesystems API
            displayName: Storage System
            kind: StorageSystem
            name: storagesystems.odf.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Data Foundation** deploys three operators.

          ### OpenShift Data Foundation
          The OpenShift Data Foundation operator is the primary operator for
          OpenShift Data Foundation. It serves to facilitate the other operators in
          OpenShift Data Foundation by performing administrative tasks outside their
          scope as well as watching and configuring their CustomResources.

          ### Rook
          [Rook][1] deploys and manages Ceph on OpenShift, which provides block,
          file and object storage.

          ### NooBaa operator
          The NooBaa operator deploys and manages the [NooBaa][2] Multi-Cloud Gateway
          on OpenShift, which provides object storage.

          # Core Capabilities
          * **Self-managing service:** No matter which supported storage technologies
          you choose, OpenShift Data Foundation ensures that resources can be
          deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Data Foundation you
          can either build dedicated storage clusters or hyper-converged clusters
          where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Data Foundation:**
          OpenShift Data Foundation integrates Ceph with multiple storage
          presentations including object storage (compatible with S3), block storage,
          and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Data Foundation efficiently
          distributes and replicates your data across your cluster to minimize the
          risk of data loss. With snapshots, cloning, and versioning, no more losing
          sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your
          datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data
          storage for your workloads, using a single namespace.

          # Subscriptions
          OpenShift Data Foundation requires a matching 1:1 ratio with the OpenShift
          subscription used. The Essential subscription is included in the OpenShift
          Platform Plus subscription. Disaster Recovery features, external mode, and
          KMS integration for encryption at rest require a bundle of OpenShift
          Platform Plus with OpenShift Data Foundation Advanced subscription.

          [1]: https://rook.io
          [2]: https://www.noobaa.io
        displayName: OpenShift Data Foundation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/odf4/odf-rhel9-operator@sha256:8d377d2d8a94e461a6eca8d1c3b4f5cf1f9ec1ba6a03044304403b58bbbabaa4
        - registry.redhat.io/odf4/odf-console-rhel9@sha256:5d0d27a091a74e65985d2f047d3f6322043f2ae80fa5ff3192ec84cce92f143d
        version: 4.15.20-rhodf
      entries:
      - name: odf-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odf-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odf-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odf-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odf-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odf-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odf-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odf-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odf-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odf-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odf-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odf-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odf-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odf-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odf-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odf-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odf-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odf-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odf-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odf-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odf-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-flashsystemcluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "flashsystemcluster.odf.ibm.com/v1alpha1",
                  "name": "my-flashsystemcluster",
                  "namespace": "openshift-storage"
                }
              },
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-storagecluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "storagecluster.ocs.openshift.io/v1",
                  "name": "my-storagecluster",
                  "namespace": "openshift-storage"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-console"]'
          containerImage: registry.redhat.io/odf4/odf-rhel9-operator@sha256:20cb12259667539a0244cfde37297b975e2682833fbe537caecb5107ebecf15f
          createdAt: "2025-09-12T06:33:04Z"
          description: OpenShift Data Foundation provides a common control plane for
            storage solutions on OpenShift Container Platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.17"}]'
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operator.openshift.io/uninstall-message: 'Warning: deleting the operand
            (StorageSystem) will result in data loss'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "odf.openshift.io/v1alpha1",
              "kind": "StorageSystem",
              "metadata": {
                "name": "ocs-storagecluster-storagesystem",
                "namespace": "openshift-storage"
              },
              "spec": {
                "kind": "storagecluster.ocs.openshift.io/v1",
                "name": "ocs-storagecluster",
                "namespace": "openshift-storage"
              }
            }
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/odf-operator
          support: Red Hat
          vendors.odf.openshift.io/kind: '["storagecluster.ocs.openshift.io/v1", "flashsystemcluster.odf.ibm.com/v1alpha1"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageSystem is the Schema for the storagesystems API
            displayName: Storage System
            kind: StorageSystem
            name: storagesystems.odf.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Data Foundation** deploys three operators.

          ### OpenShift Data Foundation
          The OpenShift Data Foundation operator is the primary operator for
          OpenShift Data Foundation. It serves to facilitate the other operators in
          OpenShift Data Foundation by performing administrative tasks outside their
          scope as well as watching and configuring their CustomResources.

          ### Rook
          [Rook][1] deploys and manages Ceph on OpenShift, which provides block,
          file and object storage.

          ### NooBaa operator
          The NooBaa operator deploys and manages the [NooBaa][2] Multi-Cloud Gateway
          on OpenShift, which provides object storage.

          # Core Capabilities
          * **Self-managing service:** No matter which supported storage technologies
          you choose, OpenShift Data Foundation ensures that resources can be
          deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Data Foundation you
          can either build dedicated storage clusters or hyper-converged clusters
          where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Data Foundation:**
          OpenShift Data Foundation integrates Ceph with multiple storage
          presentations including object storage (compatible with S3), block storage,
          and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Data Foundation efficiently
          distributes and replicates your data across your cluster to minimize the
          risk of data loss. With snapshots, cloning, and versioning, no more losing
          sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your
          datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data
          storage for your workloads, using a single namespace.

          # Subscriptions
          OpenShift Data Foundation requires a matching 1:1 ratio with the OpenShift
          subscription used. The Essential subscription is included in the OpenShift
          Platform Plus subscription. Disaster Recovery features, external mode, and
          KMS integration for encryption at rest require a bundle of OpenShift
          Platform Plus with OpenShift Data Foundation Advanced subscription.

          [1]: https://rook.io
          [2]: https://www.noobaa.io
        displayName: OpenShift Data Foundation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/odf4/odf-rhel9-operator@sha256:20cb12259667539a0244cfde37297b975e2682833fbe537caecb5107ebecf15f
        - registry.redhat.io/odf4/odf-console-rhel9@sha256:8bd822a3410e520fc57ac69c7599831173003e2c34b980c517f79aca97210257
        version: 4.16.16-rhodf
      entries:
      - name: odf-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odf-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odf-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odf-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odf-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odf-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odf-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odf-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odf-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odf-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odf-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odf-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odf-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odf-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odf-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odf-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odf-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odf-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: jaeger-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jaeger-operator.v1.65.0-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "jaegertracing.io/v1",
                "kind": "Jaeger",
                "metadata": {
                  "name": "my-jaeger"
                },
                "spec": {
                  "strategy": "allInOne"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/rhosdt/jaeger-rhel8-operator@sha256:51a8685e4dc70888c991996f149cd42233049d18f702ec72b066d6dad77a07c9
          createdAt: 25 Aug 2025, 13:30
          description: Provides tracing, monitoring and troubleshooting for microservices-based
            distributed systems
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.65.0-4'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-distributed-tracing
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/os-observability/konflux-jaeger
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Jaeger is the Schema for the jaegers API
            displayName: Jaeger
            kind: Jaeger
            name: jaegers.jaegertracing.io
            version: v1
        description: |2-

          ## Operator deprecation warning

          This Operator is deprecated and not recommended for new deployments. This Operator will be removed from the `redhat-operators` catalog in a future release.

          If this is not the first time you are installing it, migrate to the Red Hat build of OpenTelemetry Operator and the Tempo Operator for distributed tracing collection and storage. For more information, see "Migrating" in the Red Hat build of OpenTelemetry documentation, "Installing" in the Red Hat build of OpenTelemetry documentation, and "Installing" in the distributed tracing platform (Tempo) documentation.

          ## About

          Red Hat OpenShift distributed tracing platform based on Jaeger. Jaeger is project inspired by [Dapper](https://research.google.com/pubs/pub36356.html) and [OpenZipkin](http://zipkin.io/). It is a distributed tracing system released as open source by Uber Technologies. It is used for monitoring and troubleshooting microservices-based distributed systems.

          ### Core capabilities
          Jaeger is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed context propagation
          * Distributed transaction monitoring
          * Root cause analysis
          * Service dependency analysis
          * Performance / latency optimization
          * OpenTracing compatible data model
          * Multiple storage backends: Elasticsearch, Memory.

          ### Operator features

          * **Multiple modes** - Provides `allInOne`, `production` and `streaming` [modes of deployment](https://www.jaegertracing.io/docs/latest/operator/#deployment-strategies).

          * **Configuration** - The Operator manages [configuration information](https://www.jaegertracing.io/docs/latest/operator/#configuring-the-custom-resource) when installing Jaeger instances.

          * **Storage** - [Configure storage](https://www.jaegertracing.io/docs/latest/operator/#storage-options) used by Jaeger. By default, `memory` is used. Other options include `elasticsearch`. The operator can delegate creation of an Elasticsearch cluster to the Elasticsearch Operator if deployed.

          * **Agent** - can be deployed as [sidecar](https://www.jaegertracing.io/docs/latest/operator/#auto-injecting-jaeger-agent-sidecars) (default) and/or [daemonset](https://www.jaegertracing.io/docs/latest/operator/#installing-the-agent-as-daemonset).

          * **UI** - Optionally setup secure route to provide [access to the Jaeger UI](https://www.jaegertracing.io/docs/latest/operator/#accessing-the-jaeger-console-ui).

          ### Before you start
          1. Ensure that the appropriate storage solution, that will be used by the Jaeger instance, is available and configured.
          2. If intending to deploy an Elasticsearch cluster via the Jaeger custom resource, then the Elasticsearch Operator must first be installed.

          ### Support & Troubleshooting

          Red Hat OpenShift distributed tracing Jaeger is available and supported as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/4.17/html/distributed_tracing/distributed-tracing-platform-jaeger) is recommended when installing, configuring, and managing the Operator and instances.
          Support is provided to Red Hat OpenShift entitled customers subject to the [Production Scope for Coverage](https://access.redhat.com/support/offerings/production/soc) and the [Red Hat OpenShift distributed tracing Life Cycle](https://access.redhat.com/support/policy/updates/openshift#jaeger).
        displayName: Red Hat OpenShift distributed tracing platform
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tracing
        - monitoring
        - troubleshooting
        links:
        - name: Jaeger Operator Source Code
          url: https://github.com/jaegertracing/jaeger-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:667d6545b92aa8381f65369de3b5a6781d1de5b4bcfa58ccde4721b6b0406f3d
        - registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:d92003d6cf0b4f3dae11f849355c9179e78e768f3b7dbeaf3224a4a019ec5d5d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e77efd450e757772d0753a73062753dabd08e53f21315e9b87f4cc1efa332661
        - registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:1d586f9aafd1cf847a14f509692ce54956b0e754a9a04beaa7a75c480f86717c
        - registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:7b021ed467d35b2bcae87765dfc64f639b03bf71f7e34bbea1dfad16626847b9
        - registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:1f4a9daf152420d9ca7cf89c38ae5550b12dd274bada389c7b85ea4cbcbc4fb7
        - registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:fd8110529daa8cf3e7f9fcf71346adc5f48cb05717ebbad1c25bc61c94f31656
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        - registry.redhat.io/rhosdt/jaeger-rhel8-operator@sha256:51a8685e4dc70888c991996f149cd42233049d18f702ec72b066d6dad77a07c9
        version: 1.65.0-4
      entries:
      - name: jaeger-operator.v1.65.0-4
        version: 1.65.0-4
      - name: jaeger-operator.v1.65.0-3
        version: 1.65.0-3
      - name: jaeger-operator.v1.65.0-2
        version: 1.65.0-2
      - name: jaeger-operator.v1.65.0-1
        version: 1.65.0-1
      - name: jaeger-operator.v1.62.0-2
        version: 1.62.0-2
      - name: jaeger-operator.v1.62.0-1
        version: 1.62.0-1
      - name: jaeger-operator.v1.57.0-10
        version: 1.57.0-10
      - name: jaeger-operator.v1.57.0-10-0.1730817420.p
        version: 1.57.0-10+0.1730817420.p
      - name: jaeger-operator.v1.57.0-7
        version: 1.57.0-7
      - name: jaeger-operator.v1.57.0-6
        version: 1.57.0-6
      - name: jaeger-operator.v1.57.0-5
        version: 1.57.0-5
      - name: jaeger-operator.v1.53.0-4
        version: 1.53.0-4
      - name: jaeger-operator.v1.53.0-3
        version: 1.53.0-3
      - name: jaeger-operator.v1.51.0-1
        version: 1.51.0-1
      - name: jaeger-operator.v1.47.1-5
        version: 1.47.1-5
      - name: jaeger-operator.v1.42.0-5-0.1687199951.p
        version: 1.42.0-5+0.1687199951.p
      - name: jaeger-operator.v1.42.0-5
        version: 1.42.0-5
      - name: jaeger-operator.v1.34.1-5
        version: 1.34.1-5
      - name: jaeger-operator.v1.30.2
        version: 1.30.2
      name: stable
    defaultChannel: stable
    packageName: jaeger-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeRocketCI
      provider-url: https://docs.kuberocketci.io/
    name: sonar-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sonar-operator.v3.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Sonar",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonar-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonar",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonar-sample"
                },
                "spec": {
                  "defaultPermissionTemplate": "edp-default",
                  "secret": "sonar-secret",
                  "settings": [
                    {
                      "key": "sonar.global.exclusions",
                      "values": [
                        "**/*.jsx",
                        "**/*.tsx"
                      ]
                    },
                    {
                      "fieldValues": {
                        "beginBlockRegexp": ".*",
                        "endBlockRegexp": ".*"
                      },
                      "key": "sonar.issue.ignore.block"
                    },
                    {
                      "key": "sonar.dbcleaner.hoursBeforeKeepingOnlyOneSnapshotByDay",
                      "value": "20"
                    }
                  ],
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonargroup-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonargroup",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonargroup-sample"
                },
                "spec": {
                  "description": "Developers",
                  "name": "sonar-developers",
                  "permissions": [
                    "admin"
                  ],
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarPermissionTemplate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarpermissiontemplate-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarpermissiontemplate",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarpermissiontemplate-sample"
                },
                "spec": {
                  "default": true,
                  "description": "EDP permission templates",
                  "groupsPermissions": {
                    "sonar-developers": [
                      "scan"
                    ],
                    "sonar-users": [
                      "issueadmin"
                    ]
                  },
                  "name": "edp-default",
                  "projectKeyPattern": ".+",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityGate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarqualitygate-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualitygate",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarqualitygate-sample"
                },
                "spec": {
                  "conditions": {
                    "new_duplicated_lines_density": {
                      "error": "1",
                      "op": "GT"
                    }
                  },
                  "default": true,
                  "name": "sonarqualitygate-sample",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityProfile",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarqualityprofile-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualityprofile",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarqualityprofile-sample"
                },
                "spec": {
                  "default": true,
                  "language": "go",
                  "name": "sonarqualityprofile-sample",
                  "rules": {
                    "go:S1151": {
                      "params": "max=\"6\"",
                      "severity": "MAJOR"
                    }
                  },
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonaruser-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonaruser",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonaruser-sample"
                },
                "spec": {
                  "email": "sonaruser@mail.com",
                  "groups": [
                    "sonar-administrators"
                  ],
                  "login": "sonaruser",
                  "name": "sonar user",
                  "permissions": [
                    "scan"
                  ],
                  "secret": "sonaruser-secret",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/sonar-operator:3.1.0
          createdAt: "2023-09-08T12:53:57Z"
          description: An Operator for managing Sonar entities.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-sonar-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonarGroup is the Schema for the sonar group API.
            displayName: Sonar Group
            kind: SonarGroup
            name: sonargroups.edp.epam.com
            version: v1alpha1
          - description: SonarPermissionTemplate is the Schema for the sonar permission
              template API.
            displayName: Sonar Permission Template
            kind: SonarPermissionTemplate
            name: sonarpermissiontemplates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityGate is the Schema for the sonarqualitygates
              API
            displayName: Sonar Quality Gate
            kind: SonarQualityGate
            name: sonarqualitygates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityProfile is the Schema for the sonarqualityprofiles
              API
            displayName: Sonar Quality Profile
            kind: SonarQualityProfile
            name: sonarqualityprofiles.edp.epam.com
            version: v1alpha1
          - description: Sonar is the Schema for the sonars API.
            displayName: Sonar
            kind: Sonar
            name: sonars.edp.epam.com
            version: v1alpha1
          - description: SonarUser is the Schema for the sonarusers API
            displayName: Sonar User
            kind: SonarUser
            name: sonarusers.edp.epam.com
            version: v1alpha1
        description: |
          Sonar Operator is an operator that is responsible for establishing
          a connection to provided SonarQube server, reconciling Sonar entities (users,
          groups, permission templates, etc) according to the created CRs.
        displayName: EDP Sonar Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - analysis
        - automation
        - code
        - continuous
        - devops
        - edp
        - inspection
        - metrics
        - metrics
        - monitoring
        - operator
        - quality
        - security
        - software
        - static
        - technical
        - vulnerabilities
        links:
        - name: Sonar Operator
          url: https://github.com/epam/edp-sonar-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/sonar-operator:3.1.0
        version: 3.1.0
      entries:
      - name: sonar-operator.v3.1.0
        version: 3.1.0
      name: alpha
    - currentCSV: sonar-operator.v3.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Sonar",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonar"
                  },
                  "name": "sonar-sample"
                },
                "spec": {
                  "defaultPermissionTemplate": "edp-default",
                  "secret": "sonar-secret",
                  "settings": [
                    {
                      "key": "sonar.global.exclusions",
                      "values": [
                        "**/*.jsx",
                        "**/*.tsx"
                      ]
                    },
                    {
                      "fieldValues": {
                        "beginBlockRegexp": ".*",
                        "endBlockRegexp": ".*"
                      },
                      "key": "sonar.issue.ignore.block"
                    },
                    {
                      "key": "sonar.dbcleaner.hoursBeforeKeepingOnlyOneSnapshotByDay",
                      "value": "20"
                    }
                  ],
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonargroup"
                  },
                  "name": "sonargroup-sample"
                },
                "spec": {
                  "description": "Developers",
                  "name": "sonar-developers",
                  "permissions": [
                    "admin"
                  ],
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarPermissionTemplate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarpermissiontemplate"
                  },
                  "name": "sonarpermissiontemplate-sample"
                },
                "spec": {
                  "default": true,
                  "description": "EDP permission templates",
                  "groupsPermissions": {
                    "sonar-developers": [
                      "scan"
                    ],
                    "sonar-users": [
                      "issueadmin"
                    ]
                  },
                  "name": "edp-default",
                  "projectKeyPattern": ".+",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityGate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualitygate"
                  },
                  "name": "sonarqualitygate-sample"
                },
                "spec": {
                  "conditions": {
                    "new_duplicated_lines_density": {
                      "error": "1",
                      "op": "GT"
                    }
                  },
                  "default": true,
                  "name": "sonarqualitygate-sample",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityProfile",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualityprofile"
                  },
                  "name": "sonarqualityprofile-sample"
                },
                "spec": {
                  "default": true,
                  "language": "go",
                  "name": "sonarqualityprofile-sample",
                  "rules": {
                    "go:S1151": {
                      "params": "max=\"6\"",
                      "severity": "MAJOR"
                    }
                  },
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonaruser"
                  },
                  "name": "sonaruser-sample"
                },
                "spec": {
                  "email": "sonaruser@mail.com",
                  "groups": [
                    "sonar-administrators"
                  ],
                  "login": "sonaruser",
                  "name": "sonar user",
                  "permissions": [
                    "scan"
                  ],
                  "secret": "sonaruser-secret",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/sonar-operator:3.3.0
          createdAt: "2025-04-22T11:39:32Z"
          description: An Operator for managing Sonar entities.
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/epam/edp-sonar-operator
          support: KubeRocketCI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonarGroup is the Schema for the sonar group API.
            displayName: Sonar Group
            kind: SonarGroup
            name: sonargroups.edp.epam.com
            version: v1alpha1
          - description: SonarPermissionTemplate is the Schema for the sonar permission
              template API.
            displayName: Sonar Permission Template
            kind: SonarPermissionTemplate
            name: sonarpermissiontemplates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityGate is the Schema for the sonarqualitygates
              API
            displayName: Sonar Quality Gate
            kind: SonarQualityGate
            name: sonarqualitygates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityProfile is the Schema for the sonarqualityprofiles
              API
            displayName: Sonar Quality Profile
            kind: SonarQualityProfile
            name: sonarqualityprofiles.edp.epam.com
            version: v1alpha1
          - description: Sonar is the Schema for the sonars API.
            displayName: Sonar
            kind: Sonar
            name: sonars.edp.epam.com
            version: v1alpha1
          - description: SonarUser is the Schema for the sonarusers API.
            displayName: Sonar User
            kind: SonarUser
            name: sonarusers.edp.epam.com
            version: v1alpha1
        description: |
          Sonar Operator is an operator that is responsible for establishing
          a connection to provided SonarQube server, reconciling Sonar entities (users,
          groups, permission templates, etc) according to the created CRs.
        displayName: EDP Sonar Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - analysis
        - automation
        - code
        - continuous
        - devops
        - edp
        - inspection
        - metrics
        - metrics
        - monitoring
        - operator
        - quality
        - security
        - software
        - static
        - technical
        - vulnerabilities
        links:
        - name: Sonar Operator
          url: https://github.com/epam/edp-sonar-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: KubeRocketCI
          url: https://docs.kuberocketci.io/
        relatedImages:
        - docker.io/epamedp/sonar-operator:3.3.0
        version: 3.3.0
      entries:
      - name: sonar-operator.v3.3.0
        version: 3.3.0
      - name: sonar-operator.v3.2.0
        version: 3.2.0
      name: stable
    defaultChannel: stable
    packageName: sonar-operator
    provider:
      name: KubeRocketCI
      url: https://docs.kuberocketci.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco
      provider-url: ""
    name: aci-containers-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aci-containers-operator.v1.0.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"aci.ctrl/v1alpha1","kind":"AciContainersOperator","metadata":{"name":"acicnioperator","namespace":"aci-containers-system"},"spec":{}}]'
          capabilities: Seamless Upgrades
          categories: Networking
          certified: "false"
          containerImage: quay.io/noiro/aci-containers-operator@sha256:2b2d685e254c206be5371285852a8f95d0318af69cb85b786427969e035c125c
          createdAt: "2025-04-10T12:00:00Z"
          description: The Cisco Application Centric Infrastructure (ACI) Containers
            Operator brings the ACI CNI Plugin and Policy model to Kubernetes clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://github.com/noironetworks/aci-containers/blob/master/README.md
          repository: https://github.com/noironetworks/aci-containers
          support: acisupport@cisco.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aci Containers Operator
            displayName: AciContainersOperator
            kind: AciContainersOperator
            name: acicontainersoperators.aci.ctrl
            version: v1alpha1
        description: |
          ## ACI CNI Plugin

          The Cisco Application Centric Infrastructure (ACI) CNI plugin brings the ACI Networking and Policy model to Kubernetes clusters that reside on-prem or in the cloud. It is fully open source and relies on the Opflex Protocol to program Open vSwitch instances running on the Kubernetes nodes. It provides IP address management, L2/L3 networking, load balancing, and security functions for container workloads.

          The Cisco ACI CNI plugin for the Red Hat OpenShift Container Platform provides a single, programmable network infrastructure that eliminates bottlenecks between OpenShift workloads and the rest of the infrastructure while providing enterprise-grade security and offers flexible micro segmentation possibilities.

          Unified Networking - By extending the integrated overlay of the ACI fabric to OpenShift workloads, Pod and Service endpoints are managed as native ACI network resources at the same level as Bare Metal or Virtual Machines.

          No Egress Router Required - Applications running in OpenShift Pods that need high-bandwidth, low-latency traffic for data transfers external to the cluster suffer the bottleneck imposed by the egress router implementation. Pods can easily connect with other fabric or non-fabric endpoints leveraging the combined hardware and software distributed routing and switching provided by ACI and the ACI CNI plugin.

          Simplified Operations and Enhanced Visibility - Whether policies and networking are defined by the devops or fabric or security admins, the APIC becomes the single source of truth for operations and security governance. Kubernetes workload visibility is provided to Network Operators within their familiar Cisco tools.

          Hardware-assisted Load Balancing - Ingress connections to LoadBalancer-type services using ACI’s Policy Based Redirect technology moves load balancing into the fabric's hardware.

          Granular Security - Security can be implemented by using native NetworkPolicy or by using ACI EPGs and contracts, or both models complementing each other. Network policy of container workloads becomes a shared responsibility while ensuring global compliance and tenant constraints are enforced.
        displayName: Cisco ACI Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cisco
        - aci
        links:
        - name: Cisco ACI Homepage
          url: https://www.cisco.com/c/en/us/td/docs/switches/datacenter/aci/apic/sw/kb/b_Kubernetes_Integration_with_ACI.html
        maintainers:
        - email: abhishs9@cisco.com
          name: 'Abhishek Sharma '
        - email: sumitnaiksatam@gmail.com
          name: Sumit Naiksatam
        - email: vlella@cisco.com
          name: Vishal Lella
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Cisco
        relatedImages:
        - quay.io/noiro/aci-containers-operator@sha256:2b2d685e254c206be5371285852a8f95d0318af69cb85b786427969e035c125c
        - quay.io/noiro/acc-provision-operator@sha256:9c8fa6d705602af13d539fb919e72e7ee5064c6998fc9ca95d10108e986c68fe
        version: 1.0.10
      entries:
      - name: aci-containers-operator.v1.0.10
        version: 1.0.10
      name: stable
    defaultChannel: stable
    packageName: aci-containers-operator
    provider:
      name: Cisco
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Accuknox
      provider-url: https://accuknox.com
    name: kubearmor-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubearmor-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          com.redhat.openshift.versions: v4.12
          createdAt: "2023-05-01T05:17:07Z"
          description: The KubeArmor Operator deloys and manages KubeArmor daemonset
            and its componenets within a cluster.
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.28.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/kubearmor/KubeArmor
          support: Accuknox
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          The KubeArmor Operator deploys and manages KubeArmor and its components:
          - KubeArmor Daemonset
          - KubeArmor Relay Server
          - KubeArmor Controller to manage CRDs (KubeArmorPolicy, KubeArmorHostPolicy)

          KubeArmor is a cloud-native runtime security enforcement system that restricts the behavior (such as process
          execution, file access, and networking operations) of pods, containers, and nodes (VMs) at the system level.

          KubeArmor leverages Linux security modules (LSMs) such as AppArmor, SELinux, or BPF-LSM to enforce the user-specified policies. KubeArmor generates rich alerts/telemetry events with container/pod/namespace identities by leveraging eBPF.

          ## Documentation

          github: [https://github.com/kubearmor/KubeArmor](https://github.com/kubearmor/KubeArmor)
          website: [https://kubearmor.io/](https://kubearmor.io/)
        displayName: Kubearmor Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubearmor-operator
        - runtime-security
        - security
        - cloud-native
        links:
        - name: Kubearmor
          url: https://kubearmor.io
        maintainers:
        - email: ramakant@accuknox.com
          name: Ramakant Sharma
        maturity: alpha
        provider:
          name: Accuknox
          url: https://www.accuknox.com/
        relatedImages:
        - docker.io/kubearmor/kubearmor-init@sha256:44fa534587972c5f85e3e6f362dd8268e9088a33c5657a21b72a7e8563164fb5
        - docker.io/kubearmor/kubearmor-relay-server@sha256:695f9ade3382bcf858077e2a7b9b01e0713071517603778d72f7aca5a6d71323
        - docker.io/kubearmor/kubearmor-controller@sha256:ccec0fb8bb1e91bd48631cc89baf752b842f19c33b50edab08f8b33e57af64d4
        - docker.io/kubearmor/kubearmor-operator@sha256:8c2d56781a2adf78ee58f956e01417f38b6ad2f319feada6470e93f4f9bf0808
        - docker.io/kubearmor/kubearmor@sha256:52f7049ecdbfa779663d2cd44ddd7f36aecd8ccd2ca91e69709b628600029f13
        version: 1.0.0
      entries:
      - name: kubearmor-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: kubearmor-operator.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.kubearmor.com/v1",
                "kind": "KubeArmorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kubearmorconfig",
                    "app.kubernetes.io/part-of": "kubearmoroperator"
                  },
                    "name": "kubearmorconfig-default",
                    "namespace": "placeholder"
                },
                "spec": {
                  "adapters": {
                    "elasticsearch": {
                      "url": "",
                      "enabled": false,
                      "auth": {
                        "secretName": "elastic-secret",
                        "usernameKey": "username",
                        "passwordKey": "password",
                        "allowInsecureTLS": false,
                        "caCertSecretName": "",
                        "caCertKey": "ca.crt"
                      },
                      "alertsIndex": "kubearmor-alerts"
                    }
                  },
                  "recommendedPolicies": {
                      "enable": true,
                      "matchExpressions": [
                          {
                              "key": "namespace",
                              "operator": "NotIn",
                              "values": [
                                  "kube-system",
                                  "kubearmor",
                                  "openshift-infra",
                                  "openshift-ingress",
                                  "openshift-ingress-canary",
                                  "openshift-ingress-operator",
                                  "openshift-insights",
                                  "openshift-kni-infra",
                                  "openshift-kube-apiserver",
                                  "openshift-kube-apiserver-operator",
                                  "openshift-kube-controller-manager",
                                  "openshift-kube-controller-manager-operator",
                                  "openshift-kube-scheduler",
                                  "openshift-kube-scheduler-operator",
                                  "openshift-kube-storage-version-migrator",
                                  "openshift-kube-storage-version-migrator-operator",
                                  "openshift-logging",
                                  "openshift-machine-api",
                                  "openshift-machine-config-operator",
                                  "openshift-managed-node-metadata-operator",
                                  "openshift-managed-upgrade-operator",
                                  "openshift-marketplace",
                                  "openshift-monitoring",
                                  "openshift-multus",
                                  "openshift-must-gather-operator",
                                  "openshift-network-diagnostics",
                                  "openshift-network-node-identity",
                                  "openshift-network-operator",
                                  "openshift-node",
                                  "openshift-nutanix-infra",
                                  "openshift-oauth-apiserver",
                                  "openshift-observability-operator",
                                  "openshift-ocm-agent-operator",
                                  "openshift-openstack-infra",
                                  "openshift-operator-lifecycle-manager",
                                  "openshift-operators",
                                  "openshift-operators-redhat",
                                  "openshift-osd-metrics",
                                  "openshift-ovirt-infra",
                                  "openshift-ovn-kubernetes",
                                  "openshift-package-operator",
                                  "openshift-rbac-permissions",
                                  "openshift-route-controller-manager",
                                  "openshift-route-monitor-operator",
                                  "openshift-security",
                                  "openshift-service-ca",
                                  "openshift-service-ca-operator",
                                  "openshift-splunk-forwarder-operator",
                                  "openshift-sre-pruning",
                                  "openshift-user-workload-monitoring",
                                  "openshift-validation-webhook",
                                  "openshift-velero",
                                  "openshift-vsphere-infra",
                                  "openshift",
                                  "openshift-addon-operator",
                                  "openshift-apiserver",
                                  "openshift-apiserver-operator",
                                  "openshift-aqua",
                                  "openshift-authentication",
                                  "openshift-authentication-operator",
                                  "openshift-backplane",
                                  "openshift-backplane-cee",
                                  "openshift-backplane-csa",
                                  "openshift-backplane-cse",
                                  "openshift-backplane-csm",
                                  "openshift-backplane-managed-scripts",
                                  "openshift-backplane-mcs-tier-two",
                                  "openshift-backplane-mobb",
                                  "openshift-backplane-srep",
                                  "openshift-backplane-tam",
                                  "openshift-cloud-controller-manager",
                                  "openshift-cloud-controller-manager-operator",
                                  "openshift-cloud-credential-operator",
                                  "openshift-cloud-ingress-operator",
                                  "openshift-cloud-network-config-controller",
                                  "openshift-cloud-platform-infra",
                                  "openshift-cluster-csi-drivers",
                                  "openshift-cluster-machine-approver",
                                  "openshift-cluster-node-tuning-operator",
                                  "openshift-cluster-samples-operator",
                                  "openshift-cluster-storage-operator",
                                  "openshift-cluster-version",
                                  "openshift-codeready-workspaces",
                                  "openshift-config",
                                  "openshift-config-managed",
                                  "openshift-config-operator",
                                  "openshift-console",
                                  "openshift-console-operator",
                                  "openshift-console-user-settings",
                                  "openshift-controller-manager",
                                  "openshift-controller-manager-operator",
                                  "openshift-custom-domains-operator",
                                  "openshift-customer-monitoring",
                                  "openshift-deployment-validation-operator",
                                  "openshift-dns",
                                  "openshift-dns-operator",
                                  "openshift-etcd",
                                  "openshift-etcd-operator",
                                  "openshift-host-network",
                                  "openshift-image-registry"
                              ]
                          }
                      ]
                  },
                  "defaultCapabilitiesPosture": "audit",
                  "defaultFilePosture": "audit",
                  "defaultNetworkPosture": "audit",
                  "defaultVisibility": "process,network",
                  "enableStdOutLogs": false,
                  "enableStdOutAlerts": false,
                  "enableStdOutMsgs": false,
                  "seccompEnabled": false
                }
              }
            ]
          alm-examples-metadata: '{"kubearmorconfig-default":{"description":"Default
            KubeArmorConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Security, Monitoring
          certified: "true"
          com.redhat.openshift.versions: v4.9
          containerImage: docker.io/kubearmor/kubearmor-operator@sha256:d6c6b16e636474568835ded97cecb5ed17bc6004501227f6b6d63522ed57b01c
          createdAt: "2024-03-20T17:43:15Z"
          description: KubeArmor Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: kubearmor
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kubearmor/KubeArmor/blob/main/pkg/KubeArmorOperator/
          support: Accuknox
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeArmorConfig is the Schema for the kubearmorconfigs API
            kind: KubeArmorConfig
            name: kubearmorconfigs.operator.kubearmor.com
            version: v1
        description: |2-

          ## Prerequisites

          1. Create a new project for KubeArmor (Recommended)

             ```
             oc new-project kubearmor
             ```

           2. Create [KubeArmor SCC](https://github.com/kubearmor/KubeArmor/blob/main/pkg/KubeArmorOperator/config/rbac/kubearmor-scc.yaml)

             ```
             oc create -f https://raw.githubusercontent.com/kubearmor/KubeArmor/main/pkg/KubeArmorOperator/config/rbac/kubearmor-scc.yaml
             ```

             Note: please change the serviceaccount namespace if you are using some other namespace than the kubearmor.

          KubeArmor is a cloud-native runtime security enforcement system that restricts the behavior
          (such as process execution, file access, and networking operations) of pods, containers,
          and nodes (VMs) at the system level.
          KubeArmor leverages Linux security modules (LSMs) such as AppArmor, SELinux, or BPF-LSM to enforce the
          user-specified policies. KubeArmor generates rich alerts/telemetry events with container/pod/namespace
          identities by leveraging eBPF.

          ## Features and Benefits

          ### Harden Infrastructure

          safeguards vital paths like certificate bundles, enforces security rules based on MITRE, STIGs, and CIS guidelines,
          and restricts access to raw database tables, enhancing overall cluster security.

          ### Least Permissive Access

          controls sensitive asset access through process and network whitelisting,
          reducing security risks and maintaining system integrity.

          ### Application Behavior

          in-depth monitoring of process executions, file system access, service bindings,
          ingress and egress connections, and sensitive system call profiling

          ## Documentation

          github: [https://github.com/kubearmor/KubeArmor](https://github.com/kubearmor/KubeArmor)

          website: [https://kubearmor.io/](https://kubearmor.io/)

          ## Installing the KubeArmor Operator

          The KubeArmor Operator, available through Red Hat's Operator catalog, offers stable update channel.
          **Stable Channel**: This channel provides the latest stable version of KubeArmor along with its patches.

          The KubeArmor operator is responsible for deploying various components of the KubeArmor application, including the KubeArmor
          Daemonset, KubeArmor Controller, KubeArmor Relay Server, and KubeArmor Snitch. The role of KubeArmor Snitch is to gather information
          about the enabled Linux security modules on the host, which is utilized by KubeArmor to enforce runtime security policies.
          Additionally, KubeArmor Snitch collects details about the container runtime socket and storage volume cluster configurations.
          This deployment is executed as a short-term job and necessitates root user privileges. KubeArmor Snitch is certified by RedHat
          as a container with root privileges.

          KubeArmor Operator comes with a custom resource:

          - **KubeArmorConfig** - User will provide the initial configurations of the KubeArmor using KubeArmorConfig CR.

          ### KubeArmorConfig Customization Options:

          | Parameter                                 | Description                                          | Default                                 |
          | ---                                       | ---                                                  | ---                                     |
          | `alertThrottling`                         | Enable throttling for the KubeArmor generated alerts | False                                   |
          | `defaultFilePosture`                      | Default file posture for KubeArmor                   | audit                                   |
          | `defaultNetworkPosture`                   | Default network posture for KubeArmor                | audit                                   |
          | `defaultCapabilitiesPosture`              | Default capabilities posture for KubeArmor           | audit                                   |
          | `defaultVisibility`                       | Default visibility for KubeArmor                     | process,file,network                    |
          | `enableStdOutLogs` |  show stdout logs for relay server | false |
          | `enableStdOutAlerts` |  show stdout alerts for relay server | false |
          | `enableStdOutMsgs` |  show stdout messages for relay server | false |
          | `kubearmorImage.image`                    | Default image for KubeArmor Daemonset                | kubearmor/kubearmor-ubi:stable          |
          | `kubearmorImage.imagePullPolicy`          | Default imagePullPolicy for KubeArmor Daemonset      | Always                                  |
          | `kubearmorInitImage.image`                | Default image for KubeArmor Init Container           | kubearmor/kubearmor-init:stable         |
          | `kubearmorInitImage.imagePullPolicy`      | Default imagePullPolicy for KubeArmor Init Container | Always                                  |
          | `kubearmorRelayImage.image`               | Default image for KubeArmor Relay Server             | kubearmor/kubearmor-relay-server:latest |
          | `kubearmorRelayImage.imagePullPolicy`     | Default imagePullPolicy for KubeArmor Relay Server   | Always                                  |
          | `kubearmorControllerImage.image`          | Default image for KubeArmor Controller               | kubearmor/kubearmor-controller:latest   |
          | `kubearmorControllerImage.imagePullPolicy`| Default imagePullPolicy for KubeArmor Controller     | Always                                  |
          | `maxAlertPerSec`                          | Maximum no of alerts per second to trigger throttling| 10                                      |
          | `recommendedPolicies`                     | Recommended Policies Spec                            | -                                       |
          | `recommendedPolicies.matchExpressions`    | Target Namespaces for recommended policies           | -                                       |
          | `recommendedPolicies.excludePolicy`       | Exclude policies from recommended policies           | -                                       |
          | `seccompEnabled`                          | Deploy KubeArmor Daemon with default seccomp profile | True                                    |
          | `throttleSec`                             | Thottling duration in seconds                        | 30                                      |

          Users are not required to specify the images in the KubearmorConfig custom resource (CR). By default, the operator will utilize the images published with the operator release.

          For detailed information on Default Posture refer [this guide](https://github.com/kubearmor/KubeArmor/blob/main/getting-started/default_posture.md)

          similarly for default visibility refer [this guide](https://github.com/kubearmor/KubeArmor/blob/main/getting-started/kubearmor_visibility.md)

          ## Support

          For Support please join KubeArmor Slack [Here](https://join.slack.com/t/kubearmor/shared_invite/zt-1ltmqdbc6-rSHw~LM6MesZZasmP2hAcA)
        displayName: KubeArmor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - security
        - alerting
        - runtime-security
        - cloud native
        - container security
        links:
        - name: Website
          url: https://kubearmor.io
        - name: Github
          url: https://github.com/kubearmor/KubeArmor.git
        - name: Slack
          url: https://join.slack.com/t/kubearmor/shared_invite/zt-1ltmqdbc6-rSHw~LM6MesZZasmP2hAcA
        maintainers:
        - email: ramakant@accuknox.com
          name: Ramakant Sharma
        maturity: stable
        provider:
          name: Accuknox
          url: https://accuknox.com
        relatedImages:
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:1153a4592746b05e019bde4d818d176ff9350c013f84d49829032540de882841
        - docker.io/kubearmor/kubearmor-operator@sha256:d6c6b16e636474568835ded97cecb5ed17bc6004501227f6b6d63522ed57b01c
        - docker.io/kubearmor/kubearmor-snitch@sha256:c3d0f45590f9a0ebb50254047f12a7468844bcae7961eccb11f5ef578832c0eb
        - docker.io/kubearmor/kubearmor-ubi@sha256:629e295a2e64953bf95f3f3e842c42ff665ed6e4db32db8e43dcde4befdd3fc2
        - docker.io/kubearmor/kubearmor-init@sha256:e30ee508aa95d273f74a543e1b412b98c1714d04f093b59991c0ec2abaa450be
        - docker.io/kubearmor/kubearmor-relay-server@sha256:0a1a20ef02bd05eaa4a900948cadee6ae7650b0dde70abe1935608eaa5f931b7
        - docker.io/kubearmor/kubearmor-controller@sha256:2d369138cdbb31468c41e9ce70eda986a16bbeb9dd1192cd8a011c9df8a83857
        version: 1.5.2
      entries:
      - name: kubearmor-operator.v1.5.2
        version: 1.5.2
      name: stable
    defaultChannel: stable
    packageName: kubearmor-operator-certified
    provider:
      name: Accuknox
      url: https://accuknox.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Tigera
      provider-url: https://tigera.io/
    name: tigera-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tigera-operator.v1.20.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:bd93e66be5cfbc123fe4e8c3f6e8ddf7654217bf26136f06dcf23f28b53063d2
          createdAt: "2021-10-05T20:02:14.564507Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.20.4
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:bd93e66be5cfbc123fe4e8c3f6e8ddf7654217bf26136f06dcf23f28b53063d2
        version: 1.20.4
      entries:
      - name: tigera-operator.v1.20.4
        version: 1.20.4
      - name: tigera-operator.v1.20.3
        version: 1.20.3
      - name: tigera-operator.v1.20.2
        version: 1.20.2
      - name: tigera-operator.v1.20.1
        version: 1.20.1
      name: release-v1.20
    - currentCSV: tigera-operator.v1.23.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:f258275fe3d48c2ae16346a57f628523a6f2fb7b854b6d43b169056e158537f1
          createdAt: "2022-01-12T22:55:57.409244Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.23.4
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:f258275fe3d48c2ae16346a57f628523a6f2fb7b854b6d43b169056e158537f1
        version: 1.23.4
      entries:
      - name: tigera-operator.v1.23.4
        version: 1.23.4
      - name: tigera-operator.v1.23.3
        version: 1.23.3
      - name: tigera-operator.v1.23.2
        version: 1.23.2
      - name: tigera-operator.v1.23.1
        version: 1.23.1
      - name: tigera-operator.v1.23.0
        version: 1.23.0
      name: release-v1.23
    - currentCSV: tigera-operator.v1.25.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:e862708c700df55f99a9c9bb3c453e31f22aef1993d6079d80433eb8da831c94
          createdAt: "2022-02-17T05:39:13.720885Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.25.1
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:e862708c700df55f99a9c9bb3c453e31f22aef1993d6079d80433eb8da831c94
        version: 1.25.1
      entries:
      - name: tigera-operator.v1.25.1
        version: 1.25.1
      name: release-v1.25
    - currentCSV: tigera-operator.v1.28.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:9e8e47045fc67a2b00f9f2fd575fa526d7ddcac50af0b8b6d583b7518ccce120
          createdAt: "2023-05-17T16:13:52.148441156Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.28.13
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:9e8e47045fc67a2b00f9f2fd575fa526d7ddcac50af0b8b6d583b7518ccce120
        version: 1.28.13
      entries:
      - name: tigera-operator.v1.28.13
        version: 1.28.13
      name: release-v1.28
    - currentCSV: tigera-operator.v1.29.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:2a5acd29d8dca6adb318af04dbb15d93d899ed2b28b2ee9aaac09147395efb58
          createdAt: "2023-03-31T20:49:52.518230079Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.29.3
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:2a5acd29d8dca6adb318af04dbb15d93d899ed2b28b2ee9aaac09147395efb58
        version: 1.29.3
      entries:
      - name: tigera-operator.v1.29.3
        version: 1.29.3
      name: release-v1.29
    - currentCSV: tigera-operator.v1.30.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:431f037ff18b5c867d01312e42671effc55602421aeed25dd3f6109f70596b4a
          createdAt: "2023-11-17T00:52:31.263564294Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.30.9
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: APIServer
            name: apiservers.operator.tigera.io
            version: v1
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:431f037ff18b5c867d01312e42671effc55602421aeed25dd3f6109f70596b4a
        version: 1.30.9
      entries:
      - name: tigera-operator.v1.30.9
        version: 1.30.9
      name: release-v1.30
    - currentCSV: tigera-operator.v1.32.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:715ac9a30f8a9579e44258af20de354715429e11836b493918e9e1a696e9b028
          createdAt: "2023-12-15T23:45:03.964812504Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.32.3
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:715ac9a30f8a9579e44258af20de354715429e11836b493918e9e1a696e9b028
        version: 1.32.3
      entries:
      - name: tigera-operator.v1.32.3
        version: 1.32.3
      name: release-v1.32
    - currentCSV: tigera-operator.v1.36.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:3341fa9475c0325b86228c8726389f9bae9fd6c430c66fe5cd5dc39d7bb6ad4b
          createdAt: "2025-02-06T02:39:01.007218336Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.36.5
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:3341fa9475c0325b86228c8726389f9bae9fd6c430c66fe5cd5dc39d7bb6ad4b
        version: 1.36.5
      entries:
      - name: tigera-operator.v1.36.5
        version: 1.36.5
      name: release-v1.36
    - currentCSV: tigera-operator.v1.38.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "calicoNetwork": {
                    "linuxDataplane": "BPF"
                  },
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing, Monitoring, Networking, Security
          containerImage: quay.io/tigera/operator@sha256:e0a34b265aebce1a2db906d8dad99190706e8bf3910cae626b9c2eb6bbb21775
          createdAt: "2025-05-06T13:27:07.802897329Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          features.operators.openshift.io/valid-subscription: No subscription required
          olm.skipRange: <1.38.0
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift.
        displayName: Tigera Operator v1.38
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tigera-operator
        - networking
        - security
        - monitoring
        links:
        - name: Tigera
          url: https://tigera.io/
        - name: Calico Introduction
          url: https://docs.tigera.io/calico/latest/about/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.tigera.io/calico/latest/getting-started/kubernetes/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
          url: https://tigera.io/
        relatedImages:
        - quay.io/tigera/operator@sha256:e0a34b265aebce1a2db906d8dad99190706e8bf3910cae626b9c2eb6bbb21775
        version: 1.38.0
      entries:
      - name: tigera-operator.v1.38.0
        version: 1.38.0
      name: release-v1.38
    defaultChannel: release-v1.38
    packageName: tigera-operator
    provider:
      name: Tigera
      url: https://tigera.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solution-Soft
      provider-url: www.solution-soft.com
    name: timemachine-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: timemachine-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tm.solution-soft.com/v1alpha1",
                "kind": "TimeMachine",
                "metadata": {
                  "name": "timemachine-sample"
                },
                "spec": {
                  "cluster_provider": "ibm",
                  "aws": {
                    "shared_volume_name": ""
                  },
                  "ibm": {
                    "persistentvolume": {
                      "storageclass": "ibmc-file-silver-gid"
                    }
                  },
                  "tm_lichost": "52.152.148.40",
                  "tm_licpass": "qwaszx",
                  "tm_licport": "57777",
                  "imagePullSecrets": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
          createdAt: "2024-02-05T09:46:17Z"
          description: Time Machine Operator enables time travel of OpenShift pods,
            without the need to change the system clock on any of the cluster nodes
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/timemachine-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/timemachine-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=0.0.4 <1.0.17'
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: SolutionSoft Systems, Inc. - support@solution-soft.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Time Machine
            displayName: Time Machine
            kind: TimeMachine
            name: timemachines.tm.solution-soft.com
            version: v1alpha1
        description: |-
          Time Machine operator enables time travel of OpenShift pods/deployments.

          To time travel a pod where your application is, you just need to install Time Machine operator and create a Time Machine custom resource in the same namespace, without the need of changing the container images of the application you’re using, or adding containers to the pod with your application.

          Time travelling is done on the namespace level where all the (configured) target pods/deployments see the same virtual time.
        displayName: Time Machine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - time shift
        - time travel
        - virtual clocks
        - time machine
        - virtual time
        - solution-soft
        links:
        - name: Timemachine Operator
          url: https://catalog.redhat.com/software/operators/detail/5e98727a6c5dcb34dfbb19e3
        maintainers:
        - email: dtrajkovic@solution-soft.com
          name: Dragan Trajkovic
        maturity: alpha
        provider:
          name: Solution-Soft
          url: www.solution-soft.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0c304c11986790f446604319dfc29db494b15f5f9fa13f2c436688ec29c92dbe
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
        - registry.connect.redhat.com/rhel-timemachine-container/virtual-clock-testing-tool@sha256:a81a823328b0ee795c40668c89a08bcf545c5192d9eed0146071e3721a0a8372
        version: 1.0.17
      entries:
      - name: timemachine-operator.v1.0.17
        version: 1.0.17
      - name: timemachine-operator.v1.0.16
        version: 1.0.16
      name: alpha
    defaultChannel: alpha
    packageName: timemachine-operator-rhmp
    provider:
      name: Solution-Soft
      url: www.solution-soft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: ironic-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ironic-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "Ironic",
                "metadata": {
                  "name": "ironic"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseInstance": "openstack",
                  "ironicAPI": {},
                  "ironicConductors": [
                    {
                      "storageRequest": "10G"
                    }
                  ],
                  "ironicInspector": {},
                  "ironicNeutronAgent": {},
                  "secret": "osp-secret",
                  "serviceUser": "ironic",
                  "storageClass": "local-storage"
                }
              },
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "IronicAPI",
                "metadata": {
                  "name": "ironicapi-sample"
                },
                "spec": {
                  "databaseHostname": "openstack.ironic-kuttl-tests.svc"
                }
              },
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "IronicConductor",
                "metadata": {
                  "name": "ironicconductor-sample"
                },
                "spec": {
                  "databaseHostname": "openstack.ironic-kuttl-tests.svc",
                  "storageRequest": "10G"
                }
              },
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "IronicInspector",
                "metadata": {
                  "name": "ironicinspector-sample"
                },
                "spec": {
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "IronicNeutronAgent",
                "metadata": {
                  "name": "ironic-neutron-agent-sample",
                  "namespace": "openstack"
                },
                "spec": {
                  "secret": "osp-secret"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:02Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IronicAPI is the Schema for the ironicapis API
            displayName: Ironic API
            kind: IronicAPI
            name: ironicapis.ironic.openstack.org
            version: v1beta1
          - description: IronicConductor is the Schema for the ironicconductors Conductor
            displayName: Ironic Conductor
            kind: IronicConductor
            name: ironicconductors.ironic.openstack.org
            version: v1beta1
          - description: IronicInspector is the Schema for the ironicinspectors
            displayName: Ironic Inspector
            kind: IronicInspector
            name: ironicinspectors.ironic.openstack.org
            version: v1beta1
          - description: IronicNeutronAgent is the Schema for the ML2 networking-baremetal's
              agent
            displayName: Ironic Neutron Agent
            kind: IronicNeutronAgent
            name: ironicneutronagents.ironic.openstack.org
            version: v1beta1
          - description: Ironic is the Schema for the ironics API
            displayName: Ironic
            kind: Ironic
            name: ironics.ironic.openstack.org
            version: v1beta1
        description: Ironic Operator
        displayName: Ironic Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Identity
        - Ironic
        links:
        - name: Ironic Operator
          url: https://github.com/openstack-k8s-operators/ironic-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/ironic-rhel9-operator@sha256:184a1783a3b2a38bdb23f9811243414a2b21c16d78ca79ed83e2bcaa720d199b
        - registry.redhat.io/rhoso/openstack-ironic-api-rhel9@sha256:86c5330a3065a11f251e8c333bd855d37294c4e9d36cc68208fd16b3f95ef15a
        - registry.redhat.io/rhoso/openstack-ironic-conductor-rhel9@sha256:4950d74335eb48ed04f1a4f602730f1dfd42f139233a4edfebef0d8403c00b2d
        - registry.redhat.io/rhoso/openstack-ironic-inspector-rhel9@sha256:34ff13efc254d4390fbef0a5c088c569241d44d063ec16aed42e36d8179e0c85
        - registry.redhat.io/rhoso/openstack-ironic-pxe-rhel9@sha256:d49af96baa8d90dad3b244ed8c006a9945c41d05e5fa197cc35cc662b49cdd7c
        - registry.redhat.io/rhoso/openstack-ironic-neutron-agent-rhel9@sha256:132627eaf483353c5daefbbc84c6b06f51c50d0813a229ddc558e9df18bb5578
        - registry.redhat.io/rhoso/ironic-python-agent-rhel9@sha256:f41166809d2e7d76446c01f81d325a320df21e64f4b776d0f83cadcb9387eb7f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: ironic-operator.v1.0.6
        version: 1.0.6
      - name: ironic-operator.v1.0.5
        version: 1.0.5
      - name: ironic-operator.v1.0.4
        version: 1.0.4
      - name: ironic-operator.v1.0.3
        version: 1.0.3
      - name: ironic-operator.v1.0.2
        version: 1.0.2
      - name: ironic-operator.v1.0.1
        version: 1.0.1
      - name: ironic-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: ironic-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator.v1.6.3
        version: 1.6.3
      - name: dynatrace-operator.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator.v1.4.0
        version: 1.4.0
      - name: dynatrace-operator.v1.3.2
        version: 1.3.2
      - name: dynatrace-operator.v1.3.1
        version: 1.3.1
      - name: dynatrace-operator.v1.3.0
        version: 1.3.0
      - name: dynatrace-operator.v0.6.0
        version: 0.6.0
      - name: dynatrace-operator.v0.5.1
        version: 0.5.1
      - name: dynatrace-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Fujitsu
      provider-url: https://www.postgresql.fastware.com/
    name: fujitsu-enterprise-postgres-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fujitsu-enterprise-postgres-operator.v7.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPAction",
                "metadata": {
                  "name": "new-fep-action"
                },
                "spec": {
                  "fepAction": {
                    "args": [
                      "new-fep-sts-0",
                      "new-fep-sts-1"
                    ],
                    "type": "reload"
                  },
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "targetClusterName": "new-fep"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPExporter",
                "metadata": {
                  "name": "new-fep-exporter"
                },
                "spec": {
                  "fepExporter": {
                    "exporterLogLevel": "error",
                    "fepClusterList": [
                      "new-fep1"
                    ],
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "userCustomQueries": "usr_example:\n  query: \"SELECT EXTRACT(EPOCH FROM (now() - pg_last_xact_replay_timestamp())) as lag\"\n  master: true\n  metrics:\n    - lag:\n        usage: \"GAUGE\"\n        description: \"Replication lag behind master in seconds\""
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPLogging",
                "metadata": {
                  "name": "new-fep-logging"
                },
                "spec": {
                  "fepLogging": {
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPPgpool2",
                "metadata": {
                  "name": "new-fep-pgpool2"
                },
                "spec": {
                  "clientAuthMethod": "scram",
                  "count": 2,
                  "customhba": "local   all         all                               trust\nhost    all         all         127.0.0.1/32          trust\nhost    all         all         ::1/128               trust\n",
                  "customlogsize": "128Mi",
                  "customparams": "listen_addresses = '*'\npcp_listen_addresses = '*'\nnum_init_children = 32\nreserved_connections = 0\nenable_pool_hba = off\nallow_clear_text_frontend_auth = off\nauthentication_timeout = 80\nbackend_weight0 = 1\nbackend_weight1 = 1\nbackend_flag0 = 'ALWAYS_PRIMARY'\nbackend_flag1 = 'DISALLOW_TO_FAILOVER'\nconnection_cache = on\nmax_pool = 4\nlisten_backlog_multiplier = 2\nserialize_accept = off\nchild_life_time = 300\nclient_idle_limit = 0\nchild_max_connections = 0\nconnection_life_time = 0\nreset_query_list = 'ABORT; DISCARD ALL'\nclient_min_messages = info\nlog_min_messages = debug1\nlog_statement = on\nlog_per_node_statement = on\nlog_client_messages = on\nlog_hostname = on\nlog_connections = on\nlog_line_prefix = '%t: pid %p: '\nload_balance_mode = on\nignore_leading_white_space = on\nwhite_function_list = ''\nblack_function_list = 'currval,lastval,nextval,setval'\nblack_query_pattern_list = ''\ndatabase_redirect_preference_list = ''\napp_name_redirect_preference_list = ''\nallow_sql_comments = off\ndisable_load_balance_on_write = 'transaction'\nstatement_level_load_balance = on\nsr_check_period = 0\nsr_check_user = 'postgres'\ndelay_threshold = 0\nlog_standby_delay = 'none'\nssl = on\nssl_ciphers = 'HIGH:MEDIUM:+3DES:!aNULL'\nssl_prefer_server_ciphers = off\nssl_ecdh_curve = 'prime256v1'\nssl_dh_params_file = ''\nrelcache_expire = 0\nrelcache_size = 256\ncheck_temp_table = catalog\ncheck_unlogged_table = on\nenable_shared_relcache = off\nrelcache_query_target = primary\nwd_port0 = 9000\nfailover_on_backend_error = off\n",
                  "custompcp": "none",
                  "customsslcacert": "none",
                  "customsslcert": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                  "customsslkey": "none",
                  "fepclustername": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "limits": {
                    "cpu": "400m",
                    "memory": "512Mi"
                  },
                  "requests": {
                    "cpu": "200m",
                    "memory": "256Mi"
                  },
                  "serviceport": 9999,
                  "statusport": 9898
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPRestore",
                "metadata": {
                  "name": "new-fep-restore"
                },
                "spec": {
                  "fromFEPcluster": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "mcSpec": {
                    "limits": {
                      "cpu": "200m",
                      "memory": "300Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "restoretype": "latest",
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "toFEPcluster": "new-fep-2"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v2",
                "kind": "FEPCluster",
                "metadata": {
                  "name": "new-fep"
                },
                "spec": {
                  "fep": {
                    "customAnnotations": {
                      "allDeployments": {}
                    },
                    "forceSsl": true,
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "instances": 1,
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "podAntiAffinity": false,
                    "podDisruptionBudget": false,
                    "servicePort": 27500,
                    "syncMode": "off",
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "usePodName": true
                  },
                  "fepChildCrVal": {
                    "backup": {
                      "image": {
                        "pullPolicy": "IfNotPresent"
                      },
                      "mcSpec": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      },
                      "pgbackrestParams": "# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[global]\nrepo1-retention-full=7\nrepo1-retention-full-type=time\nlog-path=/database/log/backup\n",
                      "postScript": " ",
                      "preScript": " ",
                      "schedule": {
                        "num": 2
                      },
                      "schedule1": {
                        "schedule": "15 0 * * 0",
                        "type": "full"
                      },
                      "schedule2": {
                        "schedule": "15 0 * * 1-6",
                        "type": "incr"
                      },
                      "schedule3": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule4": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule5": {
                        "schedule": " ",
                        "type": " "
                      }
                    },
                    "customPgAudit": "# define pg audit custom params here to override defaults.\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[output]\nlogger = 'auditlog'\nlog_directory = '/database/log/audit'\nlog_truncate_on_rotation = on\nlog_filename = 'pgaudit-%a.log'\nlog_rotation_age = 1d\nlog_rotation_size = 0\n[rule]\n",
                    "customPgHba": "# define pg_hba custom rules here to be merged with default rules.\n# TYPE     DATABASE        USER        ADDRESS        METHOD\n",
                    "customPgParams": "# define custom postgresql.conf parameters below to override defaults.\n# Current values are as per default FEP deployment\n# If you add a library to shared_preload_libraries, add it after the default value.\n# fsep_operator_security is available in FEP15 and later container images.\nshared_preload_libraries='pgx_datamasking,pg_prewarm,pg_stat_statements,fsep_operator_security'\nsession_preload_libraries='pg_prewarm'\nmax_prepared_transactions = 100\nmax_worker_processes = 30\nmax_connections = 100\nwork_mem = 1MB\nmaintenance_work_mem = 12MB\nshared_buffers = 128MB\neffective_cache_size = 384MB\ncheckpoint_completion_target = 0.8\n# tcp parameters\ntcp_keepalives_idle = 30\ntcp_keepalives_interval = 10\ntcp_keepalives_count = 3\n# logging parameters in default fep installation\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\nlog_directory = '/database/log'\nlog_filename = 'logfile-%a.log'\nlog_file_mode = 0600\nlog_truncate_on_rotation = on\nlog_rotation_age = 1d\nlog_rotation_size = 0\nlog_checkpoints = on\nlog_line_prefix = '%e %t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h'\nlog_lock_waits = on\nlog_autovacuum_min_duration = 60s\nlogging_collector = on\npgaudit.config_file='/opt/app-root/src/pgaudit-cfg/pgaudit.conf'\nlog_replication_commands = on\nlog_min_messages = WARNING\nlog_destination = csvlog\n# wal_archive parameters in default fep installation\narchive_mode = on\narchive_command = 'pgbackrest --stanza=backupstanza --config=/database/userdata/pgbackrest.conf archive-push %p'\nwal_level = replica\nmax_wal_senders = 12\nwal_keep_segments = 64\ntrack_activities = on\ntrack_counts = on\npassword_encryption = 'scram-sha-256'\n",
                    "storage": {
                      "archivewalVol": {
                        "size": "1Gi"
                      },
                      "backupVol": {
                        "size": "2Gi"
                      },
                      "dataVol": {
                        "size": "2Gi"
                      },
                      "logVol": {
                        "size": "1Gi"
                      },
                      "tablespaceVol": {
                        "size": "512Mi"
                      },
                      "walVol": {
                        "size": "1200Mi"
                      }
                    },
                    "sysUsers": {
                      "pgRewindPassword": "rewind_password",
                      "pgRewindUser": "rewind_user",
                      "pgdb": "mydb",
                      "pgpassword": "mydbpassword",
                      "pgreplpassword": "repluserpwd",
                      "pgrepluser": "repluser",
                      "pguser": "mydbuser",
                      "tdepassphrase": "tde-passphrase"
                    },
                    "systemCertificates": {
                      "cacrt": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                      "crt": "-----BEGIN CERTIFICATE-----\nMIIDUTCCAjmgAwIBAgIRAMocW3qMoHrD6qRvMPppMkMwDQYJKoZIhvcNAQELBQAw\nNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9yIEt1\nYmVybmV0ZXMwHhcNMjEwMjA2MDQzMjM2WhcNMjYwMjA1MDQzMjM2WjA/MRAwDgYD\nVQQKEwdGdWppdHN1MSswKQYDVQQDEyJGVUpJVFNVIEVudGVycHJpc2UgUG9zdGdy\nZXMgU2VydmVyMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA4AI33yvH\nZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I2e4SceTKi6O3C/I1XuvWlpng\n5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4ANB5JWWqDOjrRT3o7nRPGXfila\nbP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYpmjdbfxabTz07ig0+6/cwKoRR\nxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTFYvmAH7gcdssSFBt8NPlUATHE\nsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6Wdgmu5H2pDml8CDNLDv98Aj7i\n+I5SRKKcVPlnuQIDAQABo1AwTjAdBgNVHSUEFjAUBggrBgEFBQcDAQYIKwYBBQUH\nAwIwDAYDVR0TAQH/BAIwADAfBgNVHSMEGDAWgBQcwrrUO0u+FhIUuVdrDRCQRsi6\nZjANBgkqhkiG9w0BAQsFAAOCAQEAm5dxBoI9pScOCvRAchg4CprdRDSJb9K6yB3O\nnCAxnM47iHeXnY3WlnI388kHu8DU7O4ba1tJbGs3KY9KzioPk43pU12jWkO1onoF\n+mTDjx/Ef1cYWA9r5q/LtgTa6Q2sxV4O2x67QW82aAnaxO34dV5zWCPIvAoovZBV\nHRT+BgCg3r2vD1RGKK2nl1aYJtWhO1SZubam+VttdZ/vbM9oOJctxmImsEtBXjkY\nKteePdQtLL5o03JhyXWyRshCq+HMmKf2KgyY8gvydGcP4eLQdBWcW40LcnVq6UjT\n0kJycJEKngMVademq1ZWHGaiYB7hyT6GhgIcHUJ2cKrPgbEh1Q==\n-----END CERTIFICATE-----",
                      "key": "-----BEGIN RSA PRIVATE KEY-----\nMIIEowIBAAKCAQEA4AI33yvHZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I\n2e4SceTKi6O3C/I1XuvWlpng5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4AN\nB5JWWqDOjrRT3o7nRPGXfilabP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYp\nmjdbfxabTz07ig0+6/cwKoRRxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTF\nYvmAH7gcdssSFBt8NPlUATHEsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6W\ndgmu5H2pDml8CDNLDv98Aj7i+I5SRKKcVPlnuQIDAQABAoIBAFPQYKlOzw/+BA0b\nyMIUpdctIMb/54CR/xR0mVw1DbSjigNVPjHUQvB8Y1B2FAITQObgJO06bAv0QdWN\nRb0/v/yYiNJDFjaLjaIAHlO/2+oWrXbFaZqgpVDJhB+e1xaZr2x7XGxm+p925k30\nl6pvIRY+I8JRKvZiV1VZHwL/R3JOtPr++xMZtLVjVOI+f+ySqJ+TZHuAjm49EKxj\ncEmmJ28b7QcziXsvKy00f+zbqLIBKXQdZAFU5eEr1BsDRXdRW+Kf0XIvftuy4BJZ\nvoKT+VGhEvF/qysswL4+6IAO6tpuYnnM0Y2d3sOGoWPkTcQK0MekYKzL/WmtCjNs\n9hodJtECgYEA5EWyhEOf4uOKe5TDp697UCUvXLoOR58FDe/S8XNvScn29jjOkqIg\nOMoqo9xAkJTNTzqn5UUdt1x/pgM2NxlPLFijrc0zQlX3SoOO2ryDd9WNi7YKtN16\nKJqa536WeZu2OEbuAZ+S3GALVy1RPeTNPnUOmKnF06DjDUGzLNCZy10CgYEA+zfw\n952DWuz1U0Z4wvAEqqcgUKXPKrkTXV/iUnjkDkrLYVr0ZofDNTXrdHl+UedFmaOC\ncieZn6DNhcdz5tKtyysGMH3g/qs9PfoGUngvcXsy0Egk04l3x1jc8TTCLqXZXYaQ\nHMsx51n+R58oncPtzYSUOr9qQ6PbC2CstTbFJA0CgYEAjGEsUliAB/jknfEzjXjG\nPdhQUxb8VyE864Az2lah9t/kJzFyIAziAeqZ5GE7t247AGFTBRTHHI8e1Qoemi3P\nWbc9GVIbFs1lIYbcIDpUIyrKPEP8O5QEXtoNLxXTFgAjRGKiVY87spjCAJ+W2ZhO\ne/1it5GYXfgQCYQA2yuBmOUCgYANRkR2YR1axaCk+NlSu6oTdmdPu6M5x7PNQE7O\nOtMaKjua9lppvIzFGAdMDUtueoEEAE7ZR1xnwfB6PDLUpJdIYAqgr1YfPt8qkjaZ\nTv56yZ7CwL0pbF8m6nwqRrZoDp1wwraEvvvxFKFKGY/k3kCHlpTakdjEoDjn3gDi\nRnWeVQKBgCEneMSzucei5LRppRtRaJw/Btll8qlPMlX3W7dxQ3cLwpmLOn0m51Fp\nPIZ44zYK8R6fu4+/sSrlfaIg86Ugeufp6YNxyNROKxUGza5vDIu5OftwWtBeg+UK\nZ8lLWNdX6pp7WMujmF3H1DrkBbauYMUKZ4UxUYtelgHERMePIxwb\n-----END RSA PRIVATE KEY-----"
                    }
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "false"
          containerImage: quay.io/fujitsu/fujitsu-enterprise-postgres-operator@sha256:b7dd55c0bfda84e398cbfbb87345fcec218cb866e286e18a111330321ab6ad54
          createdAt: "2025-09-10T11:50:37Z"
          description: OpenShift Operator for Fujitsu Enterprise Postgres
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.1.0 <7.3.0'
          operators.openshift.io/valid-subscription: '["Fujitsu Enterprise Postgres
            Advanced Edition Annual Subscription License per Core 17 Operator for
            Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/internal-objects: '["fepautoscales.fep.fujitsu.io","fepconfigs.fep.fujitsu.io","fepusers.fep.fujitsu.io","fepcerts.fep.fujitsu.io","fepvolumes.fep.fujitsu.io","fepbackups.fep.fujitsu.io"]'
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: fj-dbaas-tec@dl.jp.fujitsu.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: FEPAction
            name: fepactions.fep.fujitsu.io
            version: v1
          - kind: FEPAutoscale
            name: fepautoscales.fep.fujitsu.io
            version: v1
          - kind: FEPBackup
            name: fepbackups.fep.fujitsu.io
            version: v1
          - kind: FEPCert
            name: fepcerts.fep.fujitsu.io
            version: v1
          - kind: FEPCluster
            name: fepclusters.fep.fujitsu.io
            version: v2
          - kind: FEPConfig
            name: fepconfigs.fep.fujitsu.io
            version: v1
          - kind: FEPExporter
            name: fepexporters.fep.fujitsu.io
            version: v1
          - kind: FEPLogging
            name: feploggings.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2Cert
            name: feppgpool2certs.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2
            name: feppgpool2s.fep.fujitsu.io
            version: v1
          - kind: FEPRestore
            name: feprestores.fep.fujitsu.io
            version: v1
          - kind: FEPUpgrade
            name: fepupgrades.fep.fujitsu.io
            version: v1
          - kind: FEPUser
            name: fepusers.fep.fujitsu.io
            version: v1
          - kind: FEPVolume
            name: fepvolumes.fep.fujitsu.io
            version: v1
        description: "Fujitsu Enterprise Postgres 17 delivers an enterprise-grade
          PostgreSQL on OpenShift Container Platform.\n\nThis solution provides the
          flexibility of a hybrid cloud solution while delivering an enhanced distribution
          of PostgreSQL to support enterprise-level workloads and provide improved
          deployment and management, availability, performance, data governance and
          security.  \n\nAvailable as a multi-architecture container built for both
          amd64, s390x and ppc64le.\n\nThe download and Use of the Product is strictly
          subject to the terms of the End User License Agreement with Fujitsu Limited
          found at https://www.fast.fujitsu.com/fujitsu-enterprise-postgres-license-agreements.
          Where the Product that has been embedded as a whole or part into a third
          party program, only Authorised Customers may download and use the Product.\n"
        displayName: Fujitsu Enterprise Postgres Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - Fujitsu
        - fep
        links:
        - name: Fujitsu Enterprise Postgres
          url: https://www.postgresql.fastware.com/
        maintainers:
        - email: fj-dbaas-tec@dl.jp.fujitsu.com
          name: Fujitsu
        maturity: stable
        provider:
          name: Fujitsu
          url: https://www.postgresql.fastware.com/
        relatedImages:
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-upgrade@sha256:bf3b069a71202ea78c7c0c7b36315c8bef8e6a1c11cd0668dcd888329e26ecfb
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-backup@sha256:dd8b84e200c9d0fb32cd567a33968e15e15efc115ff120f54fdd8862c7ea3111
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentbit@sha256:9622d2191525da39e5100e1882dd2cfa3eda740d8094f3fbb71ca8a5734d2843
        - quay.io/fujitsu/fujitsu-enterprise-postgres-utils@sha256:57a3fb792480a0d3a8f075c48dee2d02f1d79cf459761e3305cf62799575a885
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-backup@sha256:42252444e2dcfe5435532f318cce041990d2b71efa26007924717924d30a46ca
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-utils@sha256:edeec4d00fd8282aaa7ec6fe3ec4cd84070a6402611002a18addcac87fb8ee94
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-server@sha256:fa4e8849d75daf331ba7d488bcb52c083cd92bb5f288e4b62c980754e299f618
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-pgpool2@sha256:185af2e4fd1701aded1ff5e44d4336562f5a219c30ba535c2a3097b914c24efe
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-upgrade@sha256:04af997b4f548ffec711b63a06c35715b4e7863537c83697b3ee0c1d95a6aec3
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-pgpool2@sha256:8eaace2f9c8e5e844992fe5e008cd1061af33cc79b4d442c4a83a65610bcc502
        - quay.io/fujitsu/fujitsu-enterprise-postgres-exporter@sha256:1c0e22aeab65ff3848772424fc3ea6baa6b58b0d88db5413bda160a08435bcff
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-upgrade@sha256:8a0f7ca1f59ba162424a308e8a70c4717135e2847e61210753b71876c9b0a75f
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-restore@sha256:09be02274427e3022849fb635a4c9abdc34696838cae39f3afafc8009567db9d
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-backup@sha256:78a3ad639d0c63858b9db0a5de36ce11a9ea121f4535e9ca07ac7defb689bc0d
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-server@sha256:3e93a368a6198818e1b7c5525324c8130c5ed8d7b4cb499a47d4bec45ef0347b
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-pgpool2@sha256:c878935ffc796c405d811a6e8a9194aaaaa8652ea7d7b879c82919ce09794b35
        - quay.io/fujitsu/fujitsu-enterprise-postgres-cronjob@sha256:7582964eef14f89f2ca9046298357afd0132fb236f30f44f84cee14692227b0c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-restore@sha256:a81a0ce926572b75aa3151e2d2cc45a0ca411cc0158d69b89247ef70418f3740
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-server@sha256:1c5823afc7e77d46940214693c12c0ad6ffd6c7d9a156e26399ab55cd5ea1dfa
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-restore@sha256:17301c2191c0d75be82d12c5a35c0f74b8d65304bcfb6d4ab312a10ffa286eda
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-restore@sha256:a6e90dd77634033f2009942b635078cceb0f4c6b15b031624c5dce820713b14c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-server@sha256:5c65f8e76f70341c512100760de6792a880ad0d6eebc9e68154b5bbd1332a896
        - quay.io/fujitsu/fujitsu-enterprise-postgres-operator@sha256:b7dd55c0bfda84e398cbfbb87345fcec218cb866e286e18a111330321ab6ad54
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-utils@sha256:b7ae535f5e5fa5eb51c6f4cdb406c7b7a9dabe8573939b0431dada6f7d27b87e
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentd@sha256:3c5534b62695867c5973c9106fc46895ed73a4b9136ad83b2586bf7e71c427d0
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-pgpool2@sha256:38aac8108dd3590abcfd854a6ec2c025a68d2170a1ba68d359b45fea6c83cec1
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-backup@sha256:ac268e2af652bdabb8ee5d764a2b335dc73c48b9022eee8c9f3cd422227a4d98
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-upgrade@sha256:c0e44447349f02eb9f6105001136402f849d876a288d61997c8c05625facff45
        version: 7.3.0
      entries:
      - name: fujitsu-enterprise-postgres-operator.v7.3.0
        version: 7.3.0
      - name: fujitsu-enterprise-postgres-operator.v7.2.7
        version: 7.2.7
      name: stable
    defaultChannel: stable
    packageName: fujitsu-enterprise-postgres-operator
    provider:
      name: Fujitsu
      url: https://www.postgresql.fastware.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-block-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12
    - currentCSV: ibm-block-csi-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "k8s.gcr.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
        version: 1.12.0
      entries:
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.0
    - currentCSV: ibm-block-csi-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
        version: 1.12.1
      entries:
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.1
    - currentCSV: ibm-block-csi-operator.v1.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:af9215670675fe9c4fa2dac7a14d547c249f3aa0c0e72f1e5390d237e411735b
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:af9215670675fe9c4fa2dac7a14d547c249f3aa0c0e72f1e5390d237e411735b
        version: 1.12.2
      entries:
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.2
    - currentCSV: ibm-block-csi-operator.v1.12.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:6e7e13d259f4d743524ebcd546af5c73d1a3ba90edaf4b5edd2005a3f8adcdfe
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:6e7e13d259f4d743524ebcd546af5c73d1a3ba90edaf4b5edd2005a3f8adcdfe
        version: 1.12.3
      entries:
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.3
    - currentCSV: ibm-block-csi-operator.v1.12.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:42728e7fd1d2302511ab08f51335ab188eef125870fa27f322df83ebb8e1f447
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:42728e7fd1d2302511ab08f51335ab188eef125870fa27f322df83ebb8e1f447
        version: 1.12.4
      entries:
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.4
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.5
    defaultChannel: stable
    packageName: ibm-block-csi-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cli-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cli-manager-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "CliManager",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-cli-manager-operator"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed",
                  "operatorLogLevel": "Normal"
                }
              }
            ]
          capabilities: Basic Install
          categories: Openshift Optional
          createdAt: "2024-06-04"
          description: "The CLI Manager is a comprehensive tool designed to simplify
            the management of \nOpenShift CLI plugins within the OpenShift environment.
            \nModeled after the popular krew plugin manager, it offers seamless integration,
            \neasy installation, and efficient handling of a wide array of plugins,
            \nenhancing your OpenShift command-line experience.\n"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          repository: https://github.com/openshift/cli-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: CLI Manager
            kind: CliManager
            name: climanagers.operator.openshift.io
            version: v1
          - displayName: CLI Plugin
            kind: Plugin
            name: plugins.config.openshift.io
            version: v1alpha1
        description: "The CLI Manager is a comprehensive tool designed to simplify
          the management of \nOpenShift CLI plugins within the OpenShift environment.
          \nModeled after the popular krew plugin manager, it offers seamless integration,
          \neasy installation, and efficient handling of a wide array of plugins,
          \nenhancing your OpenShift command-line experience.\n"
        displayName: CLI Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cli-manager-operator
        - cli-manager
        links:
        - name: Source Code
          url: https://github.com/openshift/cli-manager
        - name: Source Code
          url: https://github.com/openshift/cli-manager-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/cli-manager-operator/cli-manager-rhel9-operator@sha256:31973d7f43d052ddfb9554dd623e37fb541a8c1f08ea981558eaf4bb5ab22d5a
        - registry.redhat.io/cli-manager-operator/cli-manager-rhel9@sha256:07853bc33a0ce783d8d45c08c161e454a3abe09a8a8dadeca70c3a199466b58b
        version: 0.1.0
      entries:
      - name: cli-manager-operator.v0.1.0
        version: 0.1.0
      name: tech-preview
    defaultChannel: tech-preview
    packageName: cli-manager
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cryostat-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cryostat-operator.v4.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.cryostat.io/v1beta2",
                "kind": "Cryostat",
                "metadata": {
                  "name": "cryostat-sample"
                },
                "spec": {
                  "enableCertManager": true,
                  "eventTemplates": [],
                  "networkPolicies": {},
                  "reportOptions": {
                    "replicas": 0
                  },
                  "trustedCertSecrets": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Developer Tools
          containerImage: registry.redhat.io/cryostat/cryostat-rhel9-operator@sha256:867a649b54ad144295c7aceda6b90631841aa31a62062df62d4980313723f359
          createdAt: "2025-09-30T15:10:08Z"
          description: JVM monitoring and profiling tool
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.0 <4.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "operator.cryostat.io/v1beta2",
              "kind": "Cryostat",
              "metadata": {
                "name": "cryostat-sample"
              },
              "spec": {
                "enableCertManager": true,
                "reportOptions": {
                  "replicas": 0
                }
              }
            }
          operators.openshift.io/valid-subscription: '["Red Hat build of OpenJDK"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/cryostatio/cryostat-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cryostat allows you to install Cryostat for a single namespace,
              or multiple namespaces. It contains configuration options for controlling
              the Deployment of the Cryostat application and its related components.
              A Cryostat instance must be created to instruct the operator to deploy
              the Cryostat application.
            displayName: Cryostat
            kind: Cryostat
            name: cryostats.operator.cryostat.io
            version: v1beta2
          - description: Cryostat allows you to install Cryostat for a single namespace.
              It contains configuration options for controlling the Deployment of
              the Cryostat application and its related components. A Cryostat instance
              must be created to instruct the operator to deploy the Cryostat application.
            displayName: Cryostat
            kind: Cryostat
            name: cryostats.operator.cryostat.io
            version: v1beta1
        description: |
          The Red Hat build of Cryostat provides a cloud-based solution for interacting with the JDK Flight Recorder already present in OpenJDK 11+ JVMs. With Cryostat, users can remotely start, stop, retrieve, and even analyze JFR event data, providing the capability to easily take advantage of Flight Recorder's extremely low runtime cost and overhead and the flexibility to monitor applications and analyze recording data without transferring data outside of the cluster the application runs within.
          ##Prerequisites
          Cryostat requires [cert-manager](https://cert-manager.io/) to run. If not already installed in your cluster, please [install](https://cert-manager.io/docs/installation/) it using your preferred method.
          When running on Kubernetes, the operator also requires [Ingress configurations](https://github.com/cryostatio/cryostat-operator/blob/v4.0.0/docs/config.md#network-options) for each of its services to make them available outside of the cluster. The user is responsible for providing the hostnames for each Ingress.
        displayName: Red Hat build of Cryostat
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flightrecorder
        - java
        - jdk
        - jfr
        - jmc
        - missioncontrol
        - monitoring
        - profiling
        - diagnostic
        links:
        - name: Upstream Project
          url: https://github.com/cryostatio/cryostat
        - name: Website
          url: https://cryostat.io/
        maintainers:
        - email: cryostat-development@googlegroups.com
          name: The Cryostat Authors
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi9/nginx-124@sha256:7c1d7172880a3649591a0dc1ceb670c133ef88eed3d8192d8a2e80e3878d04c3
        - registry.redhat.io/cryostat/cryostat-reports-rhel9@sha256:5535a539b185fa2781621741ccd7f10a6b9947adc6170a95e930f2d9d7a9f608
        - registry.redhat.io/insights-runtimes-tech-preview/runtimes-inventory-rhel8-operator@sha256:58e01d11c3d30abc5b0ee3795511bd04e39aea21e2853b4f4cc6d1c8a2bd7bb0
        - registry.redhat.io/cryostat/cryostat-rhel9@sha256:fadae8713daf82c329573191381f8f9e4ebf3fc852ac5f1a68ce0ea1da376469
        - registry.redhat.io/cryostat/jfr-datasource-rhel9@sha256:1e5f29aa096d783fd0a18b3ba31c391dc203761ba3f4af8e5aeeef1f7cc86cee
        - registry.redhat.io/cryostat/cryostat-db-rhel9@sha256:23d61204ecaf037cf17bf9a5cc4428461a60fecce077884d9458802fb071daab
        - registry.redhat.io/cryostat/cryostat-grafana-dashboard-rhel9@sha256:5715e897f9a8969aedd0066b2ba597264495b4c85e4ae8ae7d8871f93d13f4c1
        - registry.redhat.io/cryostat/cryostat-agent-init-rhel9@sha256:273d6c3d3194a4e9ec50c1d8e147cd9bc6d277d8d66878d885a18528b9ba7dc6
        - registry.redhat.io/cryostat/cryostat-rhel9-operator@sha256:867a649b54ad144295c7aceda6b90631841aa31a62062df62d4980313723f359
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        - registry.redhat.io/cryostat/cryostat-ose-oauth-proxy-rhel9@sha256:d38ce4dcac5b073b001d886f0494e0fc9b750f8bd68b855fc103767a1ed8b7df
        - registry.redhat.io/cryostat/cryostat-storage-rhel9@sha256:f642ca8a0ca0aa130462c0f423ce4c775cb296f8a40c9270e5369176e87c7ee6
        - registry.redhat.io/cryostat/cryostat-openshift-console-plugin-rhel9@sha256:ca748f96414ad648eb1e3953aeeeb828b187f3c5c30e014513adc3619a698683
        version: 4.0.3
      entries:
      - name: cryostat-operator.v4.0.3
        version: 4.0.3
      - name: cryostat-operator.v4.0.2-4
        version: 4.0.2+4
      - name: cryostat-operator.v4.0.2
        version: 4.0.2
      - name: cryostat-operator.v4.0.1
        version: 4.0.1
      - name: cryostat-operator.v4.0.1-5
        version: 4.0.1+5
      - name: cryostat-operator.v4.0.1-4
        version: 4.0.1+4
      - name: cryostat-operator.v4.0.1-3
        version: 4.0.1+3
      - name: cryostat-operator.v4.0.1-2
        version: 4.0.1+2
      - name: cryostat-operator.v4.0.0-11
        version: 4.0.0+11
      - name: cryostat-operator.v4.0.0
        version: 4.0.0
      - name: cryostat-operator.v4.0.0-9
        version: 4.0.0+9
      - name: cryostat-operator.v4.0.0-10
        version: 4.0.0+10
      - name: cryostat-operator.v3.0.1-7
        version: 3.0.1+7
      - name: cryostat-operator.v3.0.1-9
        version: 3.0.1+9
      - name: cryostat-operator.v3.0.1-8
        version: 3.0.1+8
      - name: cryostat-operator.v3.0.1
        version: 3.0.1
      - name: cryostat-operator.v3.0.1-6
        version: 3.0.1+6
      - name: cryostat-operator.v3.0.1-5
        version: 3.0.1+5
      - name: cryostat-operator.v3.0.1-3
        version: 3.0.1+3
      - name: cryostat-operator.v3.0.1-13
        version: 3.0.1+13
      - name: cryostat-operator.v3.0.1-12
        version: 3.0.1+12
      - name: cryostat-operator.v3.0.1-11
        version: 3.0.1+11
      - name: cryostat-operator.v3.0.1-10
        version: 3.0.1+10
      - name: cryostat-operator.v3.0.0-12
        version: 3.0.0+12
      - name: cryostat-operator.v3.0.0
        version: 3.0.0
      - name: cryostat-operator.v3.0.0-6
        version: 3.0.0+6
      - name: cryostat-operator.v3.0.0-8
        version: 3.0.0+8
      - name: cryostat-operator.v2.4.0
        version: 2.4.0
      - name: cryostat-operator.v2.4.0-8
        version: 2.4.0+8
      - name: cryostat-operator.v2.4.0-7
        version: 2.4.0+7
      - name: cryostat-operator.v2.4.0-6
        version: 2.4.0+6
      - name: cryostat-operator.v2.4.0-4
        version: 2.4.0+4
      - name: cryostat-operator.v2.4.0-3
        version: 2.4.0+3
      - name: cryostat-operator.v2.3.1-6
        version: 2.3.1+6
      - name: cryostat-operator.v2.3.1-8
        version: 2.3.1+8
      - name: cryostat-operator.v2.3.1-9
        version: 2.3.1+9
      - name: cryostat-operator.v2.3.1
        version: 2.3.1
      - name: cryostat-operator.v2.3.0
        version: 2.3.0
      - name: cryostat-operator.v2.3.0-6
        version: 2.3.0+6
      - name: cryostat-operator.v2.2.1-14
        version: 2.2.1+14
      - name: cryostat-operator.v2.2.1-8
        version: 2.2.1+8
      - name: cryostat-operator.v2.2.1-5
        version: 2.2.1+5
      - name: cryostat-operator.v2.2.1-13
        version: 2.2.1+13
      - name: cryostat-operator.v2.2.1-12
        version: 2.2.1+12
      - name: cryostat-operator.v2.2.1-10
        version: 2.2.1+10
      - name: cryostat-operator.v2.2.1
        version: 2.2.1
      - name: cryostat-operator.v2.2.0
        version: 2.2.0
      - name: cryostat-operator.v2.1.1-2.1659469835.p
        version: 2.1.1+2.1659469835.p
      - name: cryostat-operator.v2.1.1-2.1655334614.p
        version: 2.1.1+2.1655334614.p
      - name: cryostat-operator.v2.1.1
        version: 2.1.1
      - name: cryostat-operator.v2.1.0
        version: 2.1.0
      name: stable
    defaultChannel: stable
    packageName: cryostat-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kove
      provider-url: https://kove.com
    name: kove-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kove-operator.v3.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "containers.kove.com/v1",
                "kind": "KoveMemoryConfig",
                "metadata": {
                  "labels": {
                    "KoveMemoryConfig": ""
                  },
                  "name": "example-memory-config"
                },
                "spec": {
                  "container": "",
                  "namespace": "",
                  "size": ""
                }
              },
              {
                "apiVersion": "containers.kove.com/v1",
                "kind": "KoveSystem",
                "metadata": {
                  "labels": {
                    "KoveSystem": ""
                  },
                  "name": "kove-system"
                },
                "spec": {
                  "autogrowInitialIncrement": "1 GiB",
                  "autogrowMaxIncrement": "1 TiB",
                  "l4MaxMemoryPercent": 75,
                  "managementConsoleAddress": "",
                  "memoryPlusLocalMemoryMaxSize": "0 MiB"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: quay.io/kove/kove-operator@sha256:0fb5fa2e6702d5e5bf692c25f190f65d306fc72cdceebf0814ff64adcb5f6994
          createdAt: "2025-09-25T16:31:36Z"
          description: Configure containers with Kove® pooled memory
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: kove-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Kove
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configure containers with Kove® pooled memory
            displayName: Kove Memory Configuration
            kind: KoveMemoryConfig
            name: kovememoryconfigs.containers.kove.com
            version: v1
          - description: Configure cluster with Kove:SDM™
            displayName: Kove System Configuration
            kind: KoveSystem
            name: kovesystems.containers.kove.com
            version: v1
        description: |
          ## About the Operator
          The Kove:SDM™ operator integrates Kove's software-defined memory
          solution with Kubernetes and OpenShift clusters. The operator enables
          containers to transparently provision any amount of memory, enabling total
          container memory capacity to exceed physical hardware limitations.

          Use the Kove:SDM™ operator to:

          * Create custom resources to automatically allocate memory to specific
            containers
          * Increase container flexibility with dynamic memory capacity that
            auto-grows as needed
          * Scale up container density by removing memory as an orchestration
            bottleneck

          ## Using the Operator
          Using the operator is as simple as creating a custom resource which
          specifies an initial Kove memory allocation size, a namespace, and a
          container name. Containers in the specified namespace with a matching
          container name will automatically be provided with a Kove memory allocation
          which will auto-grow as needed.

          ## About Kove
          Kove makes memory a provisionable resource, just like storage, CPUs, and
          networking. Kove:SDM™ provides a transparent, linearly scalable
          memory resource accessible to any application without code changes, from
          bare-metal apps to virtual machines and containers.
        displayName: Kove:SDM™ Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kove
        - Pooled Memory
        - Memory Pool
        - Memory Target
        - External Memory
        - Memory
        links:
        - name: Kove
          url: https://kove.com
        maintainers:
        - email: contact-kove@kove.com
          name: Kove
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Kove
          url: https://kove.com
        relatedImages:
        - registry.access.redhat.com/ubi8/pause@sha256:eacbaaadc932633330e799eda24e9b731b5a4ba7ed823ee70791d21029724a9a
        - quay.io/kove/kove-agent@sha256:9d026b925e1d414a531530a30ed8b5f10b421ee12c7c13949ea0ee9907d36b75
        - quay.io/kove/kove-client-drivers@sha256:1c1786e1bbf16609218154126b4340b2f7c4b71cbd9045589402f75c032ea38d
        - quay.io/kove/kove-compatibility-detection@sha256:0303830a1f785b5e4ce6824db7c50f8feee2678fd5e87ded983686ab3f60b232
        - quay.io/kove/kove-operator@sha256:0fb5fa2e6702d5e5bf692c25f190f65d306fc72cdceebf0814ff64adcb5f6994
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fa0acefda43e309e7f9f31f5247c933d0b868ff1573ece2f38e0549d7e2ec942
        - quay.io/kove/kove-container-hook@sha256:66ee33d59ced63406d9b40f3a3226fab7f52071f8bed9aa1ba7d859837038528
        - quay.io/kove/kove-xmem@sha256:41bfc536a278153b0a4460d2836b56762b56cf50b94d17e123a63be066c28fe7
        - quay.io/kove/kove-client-drivers@sha256:716ccda2d8f082a954f2042fd189fec257bd2dcb2c50e497bad42b9a9e5f0e3d
        version: 3.0.2
      entries:
      - name: kove-operator.v3.0.2
        version: 3.0.2
      - name: kove-operator.v3.0.1
        version: 3.0.1
      - name: kove-operator.v3.0.0
        version: 3.0.0
      name: stable
    defaultChannel: stable
    packageName: kove-operator
    provider:
      name: Kove
      url: https://kove.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom
      provider-url: https://www.broadcom.com/
    name: layer7-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: layer7-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "security.brcmlabs.com/v1",
                "kind": "Gateway",
                "metadata": {
                  "name": "ssg"
                },
                "spec": {
                  "app": {
                    "image": "docker.io/caapim/gateway:11.1.2",
                    "management": {
                      "cluster": {
                        "hostname": "gateway.brcmlabs.com",
                        "password": "7layer"
                      },
                      "password": "7layer",
                      "username": "admin"
                    },
                    "replicas": 1,
                    "service": {
                      "ports": [
                        {
                          "name": "https",
                          "port": 8443,
                          "protocol": "TCP",
                          "targetPort": 8443
                        },
                        {
                          "name": "management",
                          "port": 9443,
                          "protocol": "TCP",
                          "targetPort": 9443
                        }
                      ],
                      "type": "LoadBalancer"
                    }
                  },
                  "license": {
                    "accept": false,
                    "secretName": "gateway-license"
                  },
                  "version": "11.1.2"
                },
                "status": {}
              },
              {
                "apiVersion": "security.brcmlabs.com/v1",
                "kind": "Repository",
                "metadata": {
                  "name": "my-repository"
                },
                "spec": {
                  "auth": {},
                  "branch": "main",
                  "enabled": true,
                  "endpoint": "https://github.com/\u003cusername\u003e/\u003crepository\u003e",
                  "type": "git"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7Api",
                "metadata": {
                  "name": "l7api-sample"
                },
                "spec": {
                  "deploymentTags": [
                    "ssg",
                    "ssg1"
                  ],
                  "graphmanBundle": "jsongz",
                  "portalPublished": false,
                  "serviceUrl": "/v1/api"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7Portal",
                "metadata": {
                  "name": "dev-portal"
                },
                "spec": {
                  "auth": {},
                  "deploymentTags": [
                    "ssg"
                  ],
                  "enabled": true,
                  "endpoint": "dev-ssg.brcmlabs.com",
                  "enrollmentBundle": "",
                  "portalTenant": "portal"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7StateStore",
                "metadata": {
                  "name": "state-store-example"
                },
                "spec": {
                  "redis": {
                    "database": 0,
                    "groupName": "l7gw",
                    "masterPassword": "7layer",
                    "standalone": {
                      "host": "redis-standalone",
                      "port": 6379
                    },
                    "storeId": "test",
                    "type": "standalone"
                  },
                  "type": "redis"
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: docker.io/caapim/layer7-operator:v1.2.1
          createdAt: "2025-07-04T17:22:57Z"
          description: The Layer7 Operator covers all aspects of deploying, maintaining
            and upgrading Layer7 API Gateways in Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: layer7-operator-system
          operators.openshift.io/valid-subscription: Layer7 (Broadcom) API Gateway
            license
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/caapim/layer7-operator
          support: Broadcom Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gateway is the Schema for the Gateway Custom Resource
            displayName: Gateway
            kind: Gateway
            name: gateways.security.brcmlabs.com
            version: v1
          - description: L7Api is the Schema for the l7apis API
            displayName: L7 Api
            kind: L7Api
            name: l7apis.security.brcmlabs.com
            version: v1alpha1
          - description: L7Portal is the Schema for the l7portals API
            displayName: L7 Portal
            kind: L7Portal
            name: l7portals.security.brcmlabs.com
            version: v1alpha1
          - description: L7StateStore is the Schema for the l7statestores API
            displayName: L7 State Store
            kind: L7StateStore
            name: l7statestores.security.brcmlabs.com
            version: v1alpha1
          - description: Repository is the Schema for the repositories API
            displayName: Repository
            kind: Repository
            name: repositories.security.brcmlabs.com
            version: v1
        description: |
          The [Layer7 Operator](https://github.com/CAAPIM/layer7-operator) covers all aspects of deploying, maintaining and upgrading Layer7 API Gateways in Kubernetes. The Layer7 Operator is exclusive to Kubernetes and utilizes the Container Gateway with specific emphasis on running in Ephemeral mode (without an external MySQL policy database).

          ## Getting Started
          This [Getting Started](https://github.com/CAAPIM/layer7-operator/wiki/Getting-Started) guide covers a simple introduction to the Layer7 Operator. You can find more documentation [here](https://github.com/CAAPIM/layer7-operator/wiki).
        displayName: Layer7 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - layer7-operator
        links:
        - name: Layer7 Operator
          url: https://github.com/caapim/layer7-operator
        maintainers:
        - email: gary.vermeulen@broadcom.com
          name: Gary Vermeulen
        maturity: alpha
        provider:
          name: Broadcom
          url: https://www.broadcom.com/
        relatedImages:
        - docker.io/caapim/graphman-static-init:1.0.3
        - docker.io/caapim/gateway:11.1.2
        - docker.io/caapim/layer7-operator:v1.2.1
        version: 1.2.1
      entries:
      - name: layer7-operator.v1.2.1
        version: 1.2.1
      - name: layer7-operator.v1.2.0
        version: 1.2.0
      name: preview
    defaultChannel: preview
    packageName: layer7-operator
    provider:
      name: Broadcom
      url: https://www.broadcom.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: eG Innovations
      provider-url: https://www.eginnovations.com
    name: eginnovations-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eginnovations-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eginnovations.com/v1beta1",
                "kind": "EgUniversalAgent",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eginnovations-operator"
                  },
                  "name": "eguniversalagent",
                  "namespace": "egagent"
                },
                "spec": {
                  "agent": {
                    "env": [
                      {
                        "name": "EG_MANAGER",
                        "value": "Replace with eG Manager IP"
                      },
                      {
                        "name": "EG_MANAGER_PORT",
                        "value": "Replace with eG Manager Port"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "true"
          containerImage: docker.io/eginnovations/universal-agent-operator@sha256:b8e3e26dca1b94fabfd79496feb36f3f2de4434331ef9ca810125892a936dca7
          createdAt: "2025-09-11T20:08:00Z"
          description: eG Innovations provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with eG Universal Agent Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eginnovations/universal-agent-operator
          support: eG Innovations
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgUniversalAgent is the Schema for the eguniversalagents
              API
            displayName: Eg Universal Agent
            kind: EgUniversalAgent
            name: eguniversalagents.eginnovations.com
            version: v1beta1
        description: |
          eG Innovations’ converged application and infrastructure performance monitoring capabilities provide end-to-end visibility and correlated analytics for every layer, every tier of your IT landscape.

          The eG Enterprise Universal Agent Operator automatically configures a host agent on every Kubernetes worker node. The host agent auto-discovers the worker nodes and application containers running as Pods on each node and tracks their performance and utilization levels. In-depth monitoring of applications running on containers is also provided using the same host agent. No additional agents are required for the containers.

          ### Capabilities
            * Rolls out the egagent pod per node to monitor its pods and the node itself
            * Automatic BTM profiler injection using webhook based injection for deep application monitoring and transection tracing.

          ### Installation
            1. Create the namespace "egagent" for installing eG Agent on OpenShift Cluster.

               <br />``` $ oc create ns egagent ```
            2. Add "egagent" serviceaccount to scc( Security Context Constraints ) of OpenShift Cluster as privileged using below command.

               <br />``` $ oc adm policy add-scc-to-user privileged -z egagent -n egagent ```
            2. Install eG Innovations Universal Agent operator from [OpenShift Container Platform](https://www.openshift.com/) OperatorHub.
            3. Create the eG Universal Agent CRD (Custom Resource Definition)
               a. Click "View YAML Example" and copy the yaml contents and save it to (filename).yaml on your host.
               b. To apply the CRD run the following command

               <br />``` $ oc apply -f (filename).yaml ```

          ### Required Parameters
            * EG\_MANAGER, EG\_MANAGER\_PORT - provide the ip and port of your eG Manager Installation.

          ### Advanced Options
            * Disable Certificate Checking - disable any certificate validation that may interact poorly with proxies within your cluster
            * Image Override - use a copy of the eG Iniversal Agent container image from a registry other that Docker's or Red Hat's
            * Environment variables - define environment variables for the eG Universal Agent container of for APM profiler.
            * Namespace Selectors- select a subset of namesapces to enable instrumentation for applications.

          For complete list of supported parameters and the detailed documentation available in [eG Innovations Docs](https://docs.eginnovations.com)
        displayName: eG Innovations Universal Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Tracing
        links:
        - name: eG Innovations Kubernetes Monitoring Overview
          url: https://www.eginnovations.com/documentation/Kubernetes/Introduction-to-Monitoring-Kubernetes.htm
        maintainers:
        - email: devops@eginnovations.com
          name: eG Innovations
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: eG Innovations
          url: https://www.eginnovations.com
        relatedImages:
        - docker.io/eginnovations/universal-agent-operator@sha256:b8e3e26dca1b94fabfd79496feb36f3f2de4434331ef9ca810125892a936dca7
        version: 0.0.11
      entries:
      - name: eginnovations-operator.v0.0.11
        version: 0.0.11
      name: beta
    defaultChannel: beta
    packageName: eginnovations-operator
    provider:
      name: eG Innovations
      url: https://www.eginnovations.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-trino-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trino-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:904eb8ac055171c75c8c86e4923df85461b81e6d5d9269d9cddfb03fc615f9f4
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:904eb8ac055171c75c8c86e4923df85461b81e6d5d9269d9cddfb03fc615f9f4
        version: 24.11.1
      entries:
      - name: trino-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: trino-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b50b247ed8f0ad262ed59e8aafd099919e98156472e5663957f9922616140bd9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b50b247ed8f0ad262ed59e8aafd099919e98156472e5663957f9922616140bd9
        version: 25.3.0
      entries:
      - name: trino-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: trino-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
        version: 25.7.0
      entries:
      - name: trino-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: trino-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
        version: 25.7.0
      entries:
      - name: trino-operator.v25.7.0
        version: 25.7.0
      - name: trino-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-trino-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-kafka-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kafka-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:f76b5558088bead829be8018a89e3efc1bb7a84d40429a0a1df1c88552db169b
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:f76b5558088bead829be8018a89e3efc1bb7a84d40429a0a1df1c88552db169b
        version: 24.11.1
      entries:
      - name: kafka-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: kafka-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:72b390c2d382af0a9cbe2d9ae394fc61c62b9a3f6c7310cf5f325d0f17257f1f
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:72b390c2d382af0a9cbe2d9ae394fc61c62b9a3f6c7310cf5f325d0f17257f1f
        version: 25.3.0
      entries:
      - name: kafka-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: kafka-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
        version: 25.7.0
      entries:
      - name: kafka-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: kafka-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
        version: 25.7.0
      entries:
      - name: kafka-operator.v25.7.0
        version: 25.7.0
      - name: kafka-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-kafka-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: vfunction-operator
      app.kubernetes.io/managed-by: operator
      app.kubernetes.io/name: vfunction
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      name: vfunction-server-operator
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: vFunction
      provider-url: ""
    name: vfunction-server-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vfunction-server-operator.v2.2.469
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vfunction.com/v1",
                "kind": "VfunctionServer",
                "metadata": {
                  "name": "example-vfunction-server",
                  "namespace": "vfunction",
                  "labels": {
                      "name": "vfunction",
                      "app.kubernetes.io/name": "vfunction",
                      "app.kubernetes.io/instance": "vfunction-server",
                      "app.kubernetes.io/managed-by": "operator"
                  }
                },
                "spec": {
                  "admin": {
                    "email": "admin@mycompany.com",
                    "name": "Admin",
                    "password": "Password1!"
                  },
                  "host": "http://my.domain.com",
                  "org_name": "MyCompany",
                  "upgrade": "Daily",
                  "measurements": {
                    "auto_scaling": "No",
                    "min_number_of_services": "2",
                    "max_number_of_services": "10"
                  },
                  "smtp": {
                    "identity": "",
                    "password": "",
                    "url": "",
                    "user": ""
                  },
                  "tls": {
                    "crt": "",
                    "key": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, AI/Machine Learning
          certified: Red Hat certified
          containerImage: registry.marketplace.redhat.com/rhm/vfunction/vfunction-server-operator@sha256:60d1965460101d0fd90ecc9499d0a6d4c9684a54ada61a43a253d30661fab22f
          createdAt: "2021-08-03T19:08:54Z"
          description: vFunction is a cutting-edge code analysis, machine learning,
            and automation to boost your Java modernization projects.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/vfunction-server-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/vfunction-server-operator-rhmp/support?utm_source=openshift_console
          support: vFunction
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of vFunction running on your cluster.
            displayName: vFunction
            kind: VfunctionServer
            name: vfunctionservers.vfunction.com
            version: v1
        description: |2-

          ## Introduction

          [vFunction](https://www.vfunction.com/) is a cloud-native modernization platform that combines dynamic and static code analysis, machine learning, and automation to automatically identify and extract services from existing applications. vFunction is the only platform purpose-built for modernization of Java applications.

          The vFunction modernization process starts by learning the running monolithic application, and surfacing the interdependencies within it. Using AI, the platform analyzes and identifies services that can be separated from the application. This decomposition can present a range of micro, mini, or even macro services, depending on your application environment, each being an independently deployable and scalable application component.

          vFunction automates the extraction of these services, enabling you to modernize your monolith, quickly and easily.

          ## The vFunction platform

          The platform consists of 3 basic components; the server, the controller package, and a tools package. The server runs as an operand on an OpenShift environment. The controller package is installed on the machine that runs the monolithic application which can be either a Linux or a Windows machine, and the tools are run on a development machine, with access to the code of the monolithic application.

          The controller package consists of three elements: the vFunction agent, that collects data during the dynamic analysis phase; the vFunction Viper application, that performs static analysis on the binaries of the application; and the vFunction controller that handles all the communication between the agent, Viper, and the vFunction server.

          The vFunction agent is a mix of a Java and native agent, and needs to run on the JVM that is currently running your application. [Refer to the vFunction OOB Support Matrix document](https://drive.google.com/file/d/1ccq8LFab1FrYAimDUxwgjiCOdk4QuVzs/view) for a list of supported application servers and JVMs.

          This operator installs a vFunction server instance to be connected later with one or more vFunction controllers installed on your application machine(s).

          ## Before You Start

          The vFunction operator requires a Red Hat OpenShift Kubernetes Application Platform 4.x.

          The cluster recommended configuration (which is the same as the minimal one) should consist of:

          - *1 master node*

          - *1 worker node*

          - *Storage capacity enough for dynamically provision 2 PVs of 50G each*

          - *A default StorageClass configured*

          vFunction doesn’t provide any encryption mechanism for data stored on any attached storage. If required, encryption can be achieved by the client by using its own external means on the storage itself.

          vFunction supports backup of all critical data as part of the application. Nevertheless, for complete backup of entire data it is recommended that the client apply its own storage based backup mechanism.

          ## Install Prerequisites

          ###Cluster:
          The operator was certified and tested on OCP 4.6 and 4.7.
          If you encounter any issue with other OCP versions or cloud providers please contact info@vfunction.com.

          ###Storage:
          The operator creates two new PersistentVolumeClaims (PVCs) during installation time. For their creation, it relies on the default StorageClass to dynamically provision two PersistentVolumes (PVs). The system must have a default StorageClass in place.

          Both PersistentVolumes are accessed with ReadWriteOnce mode.

          ###Project:
          The operator should be installed in a new and dedicated project (namespace). If you intend to install more than one vFunction operands in the same cluster, please use different projects for each one. Installation of more than one operand on the same cluster will allow scaling by load balancing multiple applications to different vFunction servers.

          ## Installation instructions

          1. Prepare the YAML file to use in the installation:

              a. Copy the following YAML template into a text editor:

              ```yaml
                  apiVersion: vfunction.com/v1
                  kind: VfunctionServer
                  metadata:
                    name: vfunction
                    namespace: vfunction
                    labels:
                      name: vfunction
                      app.kubernetes.io/name: vfunction
                      app.kubernetes.io/instance: vfunction-server
                      app.kubernetes.io/managed-by: operator
                  spec:
                    host: "http://my.domain.com"
                    org_name: "MyCompany"
                    upgrade: "Daily"
                    admin:
                      email: "admin@mycompany.com"
                      name: "Admin"
                      password: "Password1!"
                    smtp:
                      password: ""
                      url: ""
                      identity: ""
                      user: ""
                    measurements:
                      auto_scaling: "No"
                      min_number_of_services: "2"
                      max_number_of_services: "10"
                    tls:
                      crt: |
                        -----BEGIN CERTIFICATE-----
                        ...
                        -----END CERTIFICATE-----
                      key: |
                        -----BEGIN PRIVATE KEY-----
                        ...
                        -----END PRIVATE KEY-----
              ```
              b. Customize the template fields by replacing with your information:

              - **host**: Enter the FQDN for accessing the vFunction dashboard.
                          Make sure that you use a domain name and not an IP address. Writing "https://" will mandate TLS while writing "http://" will expose the server through HTTP.
              - **org_name**: Enter your organization name.
              - **upgrade**: Choose auto-upgrade mode
                  - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically (default).
                  - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
                  - **Never**: The automatic upgrading mechanism is off.
                  - **Maintenance**: Put the operator in a maintenance mode.

              - **admin.email**: Enter the email address of your administrator.
              - **admin.name**: Enter the name of your administrator.
              - **admin.password**: Enter the password you want to use for the vFunction administrator.
                                    The password should be at least 8 characters long, and consist of at least one lowercase letter, at least one uppercase letter, at least one number, and at least one special character.
              - **smtp.user**: Enter the email address for a designated user for the SMTP server (optional).
              - **smtp.password**: Enter this user’s password (optional).
              - **smtp.identity**: Enter this user’s password (optional).
              - **smtp.url**: Enter the SMTP server URL (optional).
              - **tls.crt**: If your host FQDN starts with “https”, paste in the certifications you have for using the TLS connection.
              - **tls.key**: If your host FQDN starts with “https”, enter the key you have for using the TLS connection.
              - **measurements.auto_scaling**: Indicates if the measurement services auto scalling is active.
              - **measurements.min_num_of_services**: Minimum Number of measurement service pods.
              - **measurements.max_num_of_services**: Maximum Number of measurement service pods.
              - **measurements.S3**: Save all measurements data to S3, instead of local PV (optional).
                  - **measurements.S3.bucket**: Measurements S3 bucket name.
                  - **measurements.S3.key**: Measurements S3 key.
                  - **measurements.S3.secret**: Measurements S3 secret.
                  - **measurements.S3.region**: Measurements S3 region.
              - **backup.mysql.S3**: Backup MySQL to S3 (optional).
                  - **backup.mysql.S3.bucket**: MySQL backup S3 bucket name.
                  - **backup.mysql.S3.key**: MySQL backup S3 key.
                  - **backup.mysql.S3.secret**: MySQL backup S3 secret.
                  - **backup.mysql.S3.region**: MySQL backup S3 region.

          1. In the Installed Operators window choose and click on vFunction Operator and then Create Instance.
          1. In the Create vFunctionServer window, paste the YAML file that you created in step 1.
          1. Click Create.
          1. You can now click on your new created vFunction operand in order to see its details and installation progress.

          ## Verifying the installation

          Check that the operand installed successfully:

          1. Check that the "Operand State" property shows “Working”.
          1. Check that the "Successfully Installed" property shows "Yes".

          If "Operand State" shows "Failed", it indicates the operand wasn’t installed correctly. Check the events and logs for all pods (operator and image containers) for any issues. In the event that you cannot troubleshoot, contact vFunction support.

          ## After installation

          The vFunction site is now accessible via the newly created vFunction application custom address, for example, my.domain.com. There are two ways you can access the vFunction dashboard.

          ### Access using the router canonical hostname

          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address, and to the ”vfunction” subdomain of the  OpenShift canonical router hostname as the alias.
          For example, ***my.domain.com.   CNAME   vfunction.apps.ocp4.my-openshift.com.***
          2. Find your cluster Router Canonical Hostname address in the newly created vfunction-route-xxx under your vFunction OpenShift project -> **Networking** -> **Routes** -> **vfunction-route-xxx** route -> **Router Canonical Hostname** field.
          You can now access the dashboard using your defined "host" spec property (as above).

          ### Access using the nginx service location

          You may use this access method if your OpenShift is installed on a provider that supports exposing LoadBalancer-type services.
          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address and to the vfunction-nginx-xxx service location.
          For example, ***my.domain.com.   CNAME   a05951ed7cdf-1394239323.us-east-1.elb.amazonaws.com.***
          1. Point your custom domain to the vfunction-nginx-xxx service's external IP location, which you can find in the OpenShift project > **Networking** -> **Services** -> **vfunction-nginx-xxx** service > **Service Address** > **Location** field.

          ## Upgrade and Rollback

          The vFunction operator includes a built-in auto-upgrade mechanism.

          You can choose one of 4 upgrading modes:
          - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically (default).
          - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
          - **Never**: The automatic upgrading mechanism is off.
          - **Maintenance**: Put the operator in a maintenance mode.

          For an on-demand upgrade,  you can change anytime the upgrading mode from “Never” to “Always”, wait for the operand to be upgraded, and change again to “Never”.

          During the upgrade, the "Operand State" property will change to “Upgrading...” and back again to “Working” after a successful upgrade.

          In case of a failure upgrading the operand, an automatic rollback to the last working version will occur. The failed version will be marked as defective and the operator will not try to upgrade to it again.

          Choosing the "Maintenance" option will tell the operator to take down all pods, so you can safely fix storage issues, return volumes from snapshots, etc.
          After finishing the maintenance, return the upgrade mode to your original desired policy.
        displayName: vFunction Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - vfunction
        - analyzing
        - services
        links:
        - name: vFunction
          url: https://vfunction.com
        - name: Terms
          url: https://www.vfunction.com/eula/
        maintainers:
        - email: raviv@vfunction.com
          name: Raviv Halivni
        maturity: alpha
        provider:
          name: vFunction
        relatedImages:
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-organizations@sha256:bcc2c3d6222af95e614318daea3108fffc6bc77aa17b23031cce296a3811b79f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-users@sha256:e47cc051953976a0e075e8cfb457cd4749e9b9e4394f2fbecab2458c2ade92ca
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-mysql@sha256:a6e06ea08fffa33b709d4d1240587bb3ab5b4cc6a28f493a227555e88e645798
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-nginx@sha256:435d1489aaf29a811081ddb51f40ae692e5833f42c31dbced6bc189c0ddc40ab
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-storage@sha256:b778cb9c724a59e51b4cb9088047be471cdf338f1138c31b81f79e50a491203f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-idp@sha256:21ca95a0209f3c557a45658452d154777f928539a93129148470562c79e68d70
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-measurements@sha256:35ffbee4b0eead58706b23cf5a702b1d55d2a7913cd0019cefbff94957d4e5d6
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-server-operator@sha256:60d1965460101d0fd90ecc9499d0a6d4c9684a54ada61a43a253d30661fab22f
        version: 2.2.469
      entries:
      - name: vfunction-server-operator.v2.2.469
        version: 2.2.469
      name: stable
    defaultChannel: stable
    packageName: vfunction-server-operator-rhmp
    provider:
      name: vFunction
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-xtradb-cluster-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.17.0
        version: 1.17.0
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.16.1
        version: 1.16.1
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.16.0
        version: 1.16.0
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.15.1
        version: 1.15.1
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.15.0
        version: 1.15.0
      name: stable
    - currentCSV: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        version: 1.18.0-cw
      entries:
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0-cw
        version: 1.18.0-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.17.0-cw
        version: 1.17.0-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.16.1-cw
        version: 1.16.1-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.16.0-cw
        version: 1.16.0-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.15.0-cw
        version: 1.15.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator-certified-rhmp
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: community-windows-machine-config-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: community-windows-machine-config-operator.v6.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-windows/community-windows-machine-config-operator:community-4.11-6c3601f
          createdAt: "2022-08-31T18:57:22Z"
          description: An operator that enables Windows container workloads on OCP
          olm.skipRange: '>=5.0.0 <6.0.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ## WARNING
          Community distribution of the Windows Machine Config Operator.
          This is a preview build, and is not meant for production workloads.
          Issues with this distribution of WMCO can be opened against the [WMCO repository](https://github.com/openshift/windows-machine-config-operator).
          Please read through the [troubleshooting doc](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/docs/TROUBLESHOOTING.md)
          before opening an issue.
          Please ensure that when installing this operator the starting CSV you subscribe to is supported on the
          version of OKD/OCP you are using. This CSV is meant for OKD/OCP community-4.11.
          ## Documentation
          ### Introduction
          The Windows Machine Config Operator configures Windows instances into nodes, enabling Windows container workloads
          to be ran within OKD/OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/latest/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/README.md#configuring-byoh-bring-your-own-host-windows-instances).
          The operator will do all the necessary steps to configure the instance so that it can join the cluster as a worker node.
          ### Pre-requisites
          - [Cluster and OS pre-requisites](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/docs/wmco-prerequisites.md)
          ### Usage
          Please see the usage section of [README.md](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/README.md#usage).
          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.
        displayName: Community Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: preview
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-windows/community-windows-machine-config-operator:community-4.11-6c3601f
        version: 6.0.0
      entries:
      - name: community-windows-machine-config-operator.v6.0.0
        version: 6.0.0
      name: preview
    defaultChannel: preview
    packageName: community-windows-machine-config-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stolostron
      provider-url: ""
    name: cluster-aas-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-aas-operator.v0.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplate",
                "metadata": {
                  "name": "clustertemplate-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplateInstance",
                "metadata": {
                  "name": "clustertemplateinstance-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplateQuota",
                "metadata": {
                  "name": "clustertemplatequota-sample"
                },
                "spec": {}
              }
            ]
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/cluster-templates-operator:2.11.0-0c3131154065397d37e3b66ae635d7ea88650bf9
          createdAt: "2024-06-06T19:34:44Z"
          description: Easily install fully configured clusters with guard-rails.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/cluster-templates-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents instance of a cluster
            displayName: Cluster template instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.clustertemplate.openshift.io
            version: v1alpha1
          - description: Defines which ClusterTemplates can be used in a given namespace
            displayName: Cluster template quota
            kind: ClusterTemplateQuota
            name: clustertemplatequotas.clustertemplate.openshift.io
            version: v1alpha1
          - description: Template of a cluster - both installation and post-install
              setup are defined as ArgoCD application spec. Any application source
              is supported - typically a Helm chart
            displayName: Cluster template
            kind: ClusterTemplate
            name: clustertemplates.clustertemplate.openshift.io
            version: v1alpha1
          - kind: ClusterTemplateSetup
            name: clustertemplatesetup.clustertemplate.openshift.io
            version: v1alpha1
          - kind: Config
            name: config.clustertemplate.openshift.io
            version: v1alpha1
        description: |
          **Self-service clusters with guardrails!**
          Cluster as a service operator provides an easy way to define clusters as templates and allows creating instances of those templates even for non-privileged developer/devops engineers. Cluster templates operator also allows specifing quotas for the developer/devops engineers.

          ### User Interface
          The User interface is installed but is not enabled by default. To enable go to `Home` -> `Overview` -> in the status card click `Dynamic Plugins` -> click `View all` -> click the `pencil` icon -> pick `Enable` and hit `Save`. You can access the UI in the `All Clusters` perspective under `Infrastructure` -> `Cluster Templates`.

          ### Documentation
          Documentation can be found on our [website](https://github.com/stolostron/cluster-templates-operator).

          ### Getting help
          To report any issues or ask questions, create an issue on our [Github repo](https://github.com/stolostron/cluster-templates-operator/issues)
        displayName: Cluster as a service operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cluster
        - templates
        - hive
        - hypershift
        - selfservice
        - openshift
        - multicluster
        - quota
        - helm
        - argocd
        links:
        - name: Cluster as a service operator
          url: https://github.com/stolostron/cluster-templates-operator
        maintainers:
        - email: rawagner@redhat.com
          name: Rastislav Wagner
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Stolostron
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/stolostron/cluster-templates-operator:2.11.0-0c3131154065397d37e3b66ae635d7ea88650bf9
        version: 0.1.5
      entries:
      - name: cluster-aas-operator.v0.1.5
        version: 0.1.5
      - name: cluster-aas-operator.v0.1.4
        version: 0.1.4
      - name: cluster-aas-operator.v0.1.3
        version: 0.1.3
      - name: cluster-aas-operator.v0.1.2
        version: 0.1.2
      - name: cluster-aas-operator.v0.1.1
        version: 0.1.1
      - name: cluster-aas-operator.v0.1.0
        version: 0.1.0
      - name: cluster-aas-operator.v0.0.6
        version: 0.0.6
      - name: cluster-aas-operator.v0.0.5
        version: 0.0.5
      - name: cluster-aas-operator.v0.0.4
        version: 0.0.4
      - name: cluster-aas-operator.v0.0.3
        version: 0.0.3
      - name: cluster-aas-operator.v0.0.2
        version: 0.0.2
      - name: cluster-aas-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cluster-aas-operator
    provider:
      name: Stolostron
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: ibm-spectrum-scale-csi-operator
      app.kubernetes.io/managed-by: ibm-spectrum-scale-csi-operator
      app.kubernetes.io/name: ibm-spectrum-scale-csi-operator
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-spectrum-scale-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-spectrum-scale-csi-operator.v2.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "CSIScaleOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-spectrum-scale-csi-operator",
                    "app.kubernetes.io/managed-by": "ibm-spectrum-scale-csi-operator",
                    "app.kubernetes.io/name": "ibm-spectrum-scale-csi-operator",
                    "release": "ibm-spectrum-scale-csi-operator"
                  },
                  "name": "ibm-spectrum-scale-csi",
                  "namespace": "ibm-spectrum-scale-csi-driver"
                },
                "spec": {
                  "attacherNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "clusters": [
                    {
                      "id": "\u003cPrimary Cluster ID\u003e",
                      "primary": {
                        "primaryFs": "\u003c Primary Filesystem \u003e"
                      },
                      "restApi": [
                        {
                          "guiHost": "\u003c Primary cluster GUI IP/Hostname \u003e"
                        }
                      ],
                      "secrets": "secret1",
                      "secureSslMode": false
                    }
                  ],
                  "pluginNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "provisionerNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "resizerNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "snapshotterNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ]
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/ibm-spectrum-scale/ibm-spectrum-scale-csi-operator@sha256:b62ad588fbc338090f405740fd8bceaf59b3267cdb532c60ae965643ccae4d5d
          createdAt: Thu May 5 04:42:15 IST 2022
          description: An operator for deploying and managing the IBM Spectrum Scale
            CSI Driver.
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-spectrum-scale-csi/
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIScaleOperator is the Schema for the csiscaleoperators
              API
            displayName: IBM CSI Spectrum Scale Driver
            kind: CSIScaleOperator
            name: csiscaleoperators.csi.ibm.com
            version: v1
        description: |
          The IBM Spectrum Scale CSI Operator for Kubernetes installs, manages,
          upgrades the IBM Spectrum Scale CSI Driver on OpenShift and Kubernetes
          for provisioning of persistent volumes on IBM Spectrum Scale.

          ### Documentation Documentation for IBM Spectrum Scale CSI Driver and Operator can be found at [IBM Documentation](https://www.ibm.com/docs/en/spectrum-scale-csi).

          ### Getting help
          IBM Spectrum Scale CSI driver is part of the IBM Spectrum Scale offering.
          Please follow the [IBM support procedure](https://www.ibm.com/mysupport/s/?language=en_US) for any issues with the driver.

          For non-urgent issues, suggestions, recommendations feel free to open an
          issue in [github](https://github.com/IBM/ibm-spectrum-scale-csi). Issues will be addressed as team availability permits.

          **Note:** Custom Resource example represents the minimum settings needed to run
          the operator. If your environment needs more advanced settings (e.g. remote clusters,
          node mapping, image pull secret, Taint and Toleration etc.) please refer to  the sample [Custom Resource](https://raw.githubusercontent.com/IBM/ibm-spectrum-scale-csi/v2.6.0/operator/config/samples/csiscaleoperators.csi.ibm.com_cr.yaml).
        displayName: IBM Spectrum Scale CSI Plugin Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - GPFS
        - Spectrum Scale
        - CSI Driver
        links:
        - name: IBM CSI Spectrum Scale Documentation
          url: https://www.ibm.com/docs/en/spectrum-scale-csi
        - name: CSI Developer Documentation
          url: https://kubernetes-csi.github.io/docs/
        maintainers:
        - email: madhu.punjabi@in.ibm.com
          name: Madhu Thorat
        - email: amdabhad@in.ibm.com
          name: Amol Dabhade
        - email: Nitish.Kumar4@ibm.com
          name: Nitish Kumar
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibm-spectrum-scale/ibm-spectrum-scale-csi-operator@sha256:b62ad588fbc338090f405740fd8bceaf59b3267cdb532c60ae965643ccae4d5d
        version: 2.6.0
      entries:
      - name: ibm-spectrum-scale-csi-operator.v2.6.0
        version: 2.6.0
      - name: ibm-spectrum-scale-csi-operator.v2.5.1
        version: 2.5.1
      - name: ibm-spectrum-scale-csi-operator.v2.5.0
        version: 2.5.0
      - name: ibm-spectrum-scale-csi-operator.v2.4.0
        version: 2.4.0
      - name: ibm-spectrum-scale-csi-operator.v2.3.1
        version: 2.3.1
      - name: ibm-spectrum-scale-csi-operator.v2.3.0
        version: 2.3.0
      name: stable
    defaultChannel: stable
    packageName: ibm-spectrum-scale-csi-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Managed Openshift Black Belt
      provider-url: ""
    name: ecr-secret-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ecr-secret-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "ecr-secret-sample",
                  "namespace": "test-ecr-secret-operator"
                },
                "spec": {
                  "ecr_registry": "ACCOUNT_ID.dkr.ecr.us-east-2.amazonaws.com",
                  "frequency": "10h",
                  "generated_secret_name": "ecr-docker-secret",
                  "region": "us-east-2"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ecr-secret-operator:v0.3.2
          operatorframework.io/suggested-namespace: ecr-secret-operator
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ecr-secret-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret is the Schema for the secrets API
            displayName: Secret
            kind: Secret
            name: secrets.ecr.mobb.redhat.com
            version: v1alpha1
        description: "# ECR Secret Operator\n\nAmazon Elastic Container Registry [Private
          Registry Authentication](https://docs.aws.amazon.com/AmazonECR/latest/userguide/registry_auth.html)
          provides a temporary token that is valid only for 12 hours. It is a challenge
          for automatic container image build process to refresh the token or secret
          in a timely manner.\n\nThis operators frequently talks with AWS ECR GetAuthroization
          Token and create/update the secret, so that the service account can perform
          docker image build.\n\n\n## How to use this operator\n\n### Prerequisites\n\n*
          [Create an ECR private repository](https://docs.aws.amazon.com/AmazonECR/latest/userguide/repository-create.html)\n*
          Create An Openshift Cluster\n* Provide AWS Authentication to the operator.
          Two Options:\n  **Note: You have to provide a secret aws-ecr-cloud-credentials
          in the installed namespace**\n  * [IAM User](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_user.md)\n
          \ * [STS Assume Role](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_assume_role.md)\n\n###
          Install the operator\n\n### Create the ECR Secret CRD\n\n```\napiVersion:
          ecr.mobb.redhat.com/v1alpha1\nkind: Secret\nmetadata:\n  name: ecr-secret\n
          \ namespace: test-ecr-secret-operator\nspec:\n  generated_secret_name: ecr-docker-secret\n
          \ ecr_registry: [ACCOUNT_ID].dkr.ecr.us-east-2.amazonaws.com\n  frequency:
          10h\n  region: us-east-2\n```\n\n```\noc create -f CRD_FILE\n```\n\nA docker
          registry secret is created by the operator momentally and the token is patched
          every 10 hours\n\n```\noc get secret ecr-docker-secret   \nNAME                TYPE
          \                            DATA   AGE\necr-docker-secret   kubernetes.io/dockerconfigjson
          \  1      16h\n```\n\n### A sample build process with generated secret\n\n\nLink
          the secret to builder\n\n```\noc secrets link builder ecr-docker-secret
          \n```\n\nConfigure [build config](https://github.com/rh-mobb/ecr-secret-operator/tree/main/samples/build-config.yaml)
          to point to your ECR Container repository\n\n```\noc create imagestream
          ruby\noc tag openshift/ruby:2.5-ubi8 ruby:2.5\noc create -f samples/build-config.yaml\noc
          start-build ruby-sample-build --wait\n```\n\nBuild should succeed and push
          the image to the the private ECR Container repository\n\n"
        displayName: ECR Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - redhat
        links:
        - name: Ecr Secret Operator
          url: https://github.com/rh-mobb/ecr-secret-operator
        maintainers:
        - email: shading@redhat.com
          name: Shaozhen Ding
        maturity: alpha
        provider:
          name: Managed Openshift Black Belt
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/mobb/ecr-secret-operator:v0.3.2
        version: 0.3.2
      entries:
      - name: ecr-secret-operator.v0.3.2
        version: 0.3.2
      name: v0.3
    - currentCSV: ecr-secret-operator.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "ecr-secret-sample",
                  "namespace": "test-ecr-secret-operator"
                },
                "spec": {
                  "ecr_registry": "ACCOUNT_ID.dkr.ecr.us-east-2.amazonaws.com",
                  "frequency": "10h",
                  "generated_secret_name": "ecr-docker-secret",
                  "region": "us-east-2"
                }
              },
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "ArgoHelmRepoSecret",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ecr-secret-operator",
                    "app.kubernetes.io/instance": "argohelmreposecret-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "argohelmreposecret",
                    "app.kubernetes.io/part-of": "ecr-secret-operator"
                  },
                  "name": "argohelmreposecret-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ecr-secret-operator:v0.4.1
          createdAt: "2023-04-14T14:52:13Z"
          operatorframework.io/suggested-namespace: ecr-secret-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ecr-secret-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ArgoHelmRepoSecret is the Schema for the argohelmreposecrets
              API
            displayName: Argo Helm Repo Secret
            kind: ArgoHelmRepoSecret
            name: argohelmreposecrets.ecr.mobb.redhat.com
            version: v1alpha1
          - description: Secret is the Schema for the secrets API
            displayName: Secret
            kind: Secret
            name: secrets.ecr.mobb.redhat.com
            version: v1alpha1
        description: "# ECR Secret Operator\n\nAmazon Elastic Container Registry [Private
          Registry Authentication](https://docs.aws.amazon.com/AmazonECR/latest/userguide/registry_auth.html)
          provides a temporary token that is valid only for 12 hours. It is a challenge
          for automatic container image build process to refresh the token or secret
          in a timely manner.\n\nThis operators frequently talks with AWS ECR GetAuthroization
          Token and create/update the secret, so that the service account can perform
          docker image build.\n\n\n## How to use this operator\n\n### Prerequisites\n\n*
          [Create an ECR private repository](https://docs.aws.amazon.com/AmazonECR/latest/userguide/repository-create.html)\n*
          Create An Openshift Cluster\n* Provide AWS Authentication to the operator.
          Two Options:\n  **Note: You have to provide a secret aws-ecr-cloud-credentials
          in the installed namespace**\n  * [IAM User](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_user.md)\n
          \ * [STS Assume Role](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_assume_role.md)\n\n###
          Install the operator\n\n### Create the ECR Secret CRD\n\n```\napiVersion:
          ecr.mobb.redhat.com/v1alpha1\nkind: Secret\nmetadata:\n  name: ecr-secret\n
          \ namespace: test-ecr-secret-operator\nspec:\n  generated_secret_name: ecr-docker-secret\n
          \ ecr_registry: [ACCOUNT_ID].dkr.ecr.us-east-2.amazonaws.com\n  frequency:
          10h\n  region: us-east-2\n```\n\n```\noc create -f CRD_FILE\n```\n\nA docker
          registry secret is created by the operator momentally and the token is patched
          every 10 hours\n\n```\noc get secret ecr-docker-secret   \nNAME                TYPE
          \                            DATA   AGE\necr-docker-secret   kubernetes.io/dockerconfigjson
          \  1      16h\n```\n\n### A sample build process with generated secret\n\n\nLink
          the secret to builder\n\n```\noc secrets link builder ecr-docker-secret
          \n```\n\nConfigure [build config](https://github.com/rh-mobb/ecr-secret-operator/tree/main/samples/build-config.yaml)
          to point to your ECR Container repository\n\n```\noc create imagestream
          ruby\noc tag openshift/ruby:2.5-ubi8 ruby:2.5\noc create -f samples/build-config.yaml\noc
          start-build ruby-sample-build --wait\n```\n\nBuild should succeed and push
          the image to the the private ECR Container repository\n\n"
        displayName: ECR Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - redhat
        links:
        - name: Ecr Secret Operator
          url: https://github.com/rh-mobb/ecr-secret-operator
        maintainers:
        - email: shading@redhat.com
          name: Shaozhen Ding
        maturity: alpha
        provider:
          name: Managed Openshift Black Belt
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/mobb/ecr-secret-operator:v0.4.1
        version: 0.4.1
      entries:
      - name: ecr-secret-operator.v0.4.1
        version: 0.4.1
      - name: ecr-secret-operator.v0.4.0
        version: 0.4.0
      name: v0.4
    defaultChannel: v0.4
    packageName: ecr-secret-operator
    provider:
      name: Managed Openshift Black Belt
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-apigatewayv2-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-apigatewayv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Deployment",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "API",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Authorizer",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Route",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Stage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "VPCLink",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/apigatewayv2-controller:1.1.1
          createdAt: "2025-09-19T21:58:07Z"
          description: AWS API Gateway v2 controller is a service controller for managing
            API Gateway v2 resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIMapping represents the state of an AWS apigatewayv2 APIMapping
              resource.
            displayName: APIMapping
            kind: APIMapping
            name: apimappings.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: API represents the state of an AWS apigatewayv2 API resource.
            displayName: API
            kind: API
            name: apis.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Authorizer represents the state of an AWS apigatewayv2 Authorizer
              resource.
            displayName: Authorizer
            kind: Authorizer
            name: authorizers.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Deployment represents the state of an AWS apigatewayv2 Deployment
              resource.
            displayName: Deployment
            kind: Deployment
            name: deployments.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: DomainName represents the state of an AWS apigatewayv2 DomainName
              resource.
            displayName: DomainName
            kind: DomainName
            name: domainnames.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Integration represents the state of an AWS apigatewayv2 Integration
              resource.
            displayName: Integration
            kind: Integration
            name: integrations.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Route represents the state of an AWS apigatewayv2 Route resource.
            displayName: Route
            kind: Route
            name: routes.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Stage represents the state of an AWS apigatewayv2 Stage resource.
            displayName: Stage
            kind: Stage
            name: stages.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: VPCLink represents the state of an AWS apigatewayv2 VPCLink
              resource.
            displayName: VPCLink
            kind: VPCLink
            name: vpclinks.apigatewayv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon API Gateway v2 resources in AWS from within your Kubernetes cluster. Currently only HTTP APIs are supported. WebSocket API support will be added soon. For HTTP APIs, DomainName and APIMapping resource support will be added soon.

          **About Amazon API Gateway v2**

          Amazon API Gateway is a fully managed service that makes it easy for developers to create, publish, maintain, monitor, and secure APIs at any scale. APIs act as the "front door" for applications to access data, business logic, or functionality from your backend services. Using API Gateway, you can create RESTful APIs and WebSocket APIs that enable real-time two-way communication applications. API Gateway supports containerized and serverless workloads, as well as web applications.

          API Gateway handles all the tasks involved in accepting and processing up to hundreds of thousands of concurrent API calls, including traffic management, CORS support, authorization and access control, throttling, monitoring, and API version management. API Gateway has no minimum fees or startup costs. You pay for the API calls you receive and the amount of data transferred out and, with the API Gateway tiered pricing model, you can reduce your cost as your API usage scales.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon API Gateway v2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apigatewayv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon API Gateway v2 Developer Resources
          url: https://aws.amazon.com/api-gateway/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: api gateway maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/apigatewayv2-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-apigatewayv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-apigatewayv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-apigatewayv2-controller.v1.0.30
        version: 1.0.30
      - name: ack-apigatewayv2-controller.v1.0.29
        version: 1.0.29
      - name: ack-apigatewayv2-controller.v1.0.28
        version: 1.0.28
      - name: ack-apigatewayv2-controller.v1.0.26
        version: 1.0.26
      - name: ack-apigatewayv2-controller.v1.0.25
        version: 1.0.25
      - name: ack-apigatewayv2-controller.v1.0.23
        version: 1.0.23
      - name: ack-apigatewayv2-controller.v1.0.21
        version: 1.0.21
      - name: ack-apigatewayv2-controller.v1.0.17
        version: 1.0.17
      - name: ack-apigatewayv2-controller.v1.0.16
        version: 1.0.16
      - name: ack-apigatewayv2-controller.v1.0.15
        version: 1.0.15
      - name: ack-apigatewayv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-apigatewayv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-apigatewayv2-controller.v1.0.7
        version: 1.0.7
      - name: ack-apigatewayv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-apigatewayv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-apigatewayv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-apigatewayv2-controller.v1.0.3
        version: 1.0.3
      - name: ack-apigatewayv2-controller.v1.0.2
        version: 1.0.2
      - name: ack-apigatewayv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-apigatewayv2-controller.v0.1.6
        version: 0.1.6
      - name: ack-apigatewayv2-controller.v0.1.4
        version: 0.1.4
      - name: ack-apigatewayv2-controller.v0.1.3
        version: 0.1.3
      - name: ack-apigatewayv2-controller.v0.1.2
        version: 0.1.2
      - name: ack-apigatewayv2-controller.v0.1.1
        version: 0.1.1
      - name: ack-apigatewayv2-controller.v0.1.0
        version: 0.1.0
      - name: ack-apigatewayv2-controller.v0.0.22
        version: 0.0.22
      - name: ack-apigatewayv2-controller.v0.0.21
        version: 0.0.21
      - name: ack-apigatewayv2-controller.v0.0.20
        version: 0.0.20
      - name: ack-apigatewayv2-controller.v0.0.19
        version: 0.0.19
      - name: ack-apigatewayv2-controller.v0.0.18
        version: 0.0.18
      - name: ack-apigatewayv2-controller.v0.0.17
        version: 0.0.17
      - name: ack-apigatewayv2-controller.v0.0.16
        version: 0.0.16
      - name: ack-apigatewayv2-controller.v0.0.15
        version: 0.0.15
      - name: ack-apigatewayv2-controller.v0.0.14
        version: 0.0.14
      - name: ack-apigatewayv2-controller.v0.0.13
        version: 0.0.13
      - name: ack-apigatewayv2-controller.v0.0.12
        version: 0.0.12
      - name: ack-apigatewayv2-controller.v0.0.11
        version: 0.0.11
      - name: ack-apigatewayv2-controller.v0.0.10
        version: 0.0.10
      - name: ack-apigatewayv2-controller.v0.0.9
        version: 0.0.9
      name: alpha
    defaultChannel: alpha
    packageName: ack-apigatewayv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Backube
      provider-url: https://github.com/backube
    name: snapscheduler
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: snapscheduler.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "snapscheduler.backube/v1",
                "kind": "SnapshotSchedule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "snapscheduler",
                    "app.kubernetes.io/instance": "daily",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "snapshotschedule",
                    "app.kubernetes.io/part-of": "snapscheduler"
                  },
                  "name": "daily"
                },
                "spec": {
                  "retention": {
                    "maxCount": 7
                  },
                  "schedule": "0 0 * * *"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/backube/snapscheduler:3.5.0
          createdAt: "2025-05-14T14:46:43Z"
          description: The SnapScheduler operator takes snapshots of Kubernetes CSI-based
            persistent volumes according to user-supplied schedules.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/snapscheduler
          support: https://github.com/backube/snapscheduler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SnapshotSchedule defines a schedule for taking automated
              snapshots of PVC(s)
            displayName: Snapshot Schedule
            kind: SnapshotSchedule
            name: snapshotschedules.snapscheduler.backube
            version: v1
        description: |-
          ## About this operator

          The SnapScheduler operator takes snapshots of CSI-based PersistentVolumes
          according to a configurable
          [Cron-like](https://en.wikipedia.org/wiki/Cron#Overview) schedule. The
          schedules include configurable retention policies for snapshots as well as
          selectors to limit the volumes that are snapshotted. An example schedule
          could be:

          > *Snapshot **all volumes** in a namespace **daily at midnight**, retaining
          > the most recent **7** snapshots for each volume.*

          Multiple schedules can be combined to provide more elaborate protection
          schemes. For example, a given volume (or collection of volumes) could be
          protected with:

          - 6 hourly snapshots
          - 7 daily snapshots
          - 4 weekly snapshots
          - 12 monthly snapshots

          ### How it works

          The operator watches for `SnapshotSchedule` CRs in each namespace. When the
          current time matches the schedule's cronspec, the operator creates a
          `VolumeSnapshot` object for each `PersistentVolumeClaim` in the namespace
          (or subset thereof if a label selector is provided). The `VolumeSnapshot`
          objects are named according to the template:
          `<pvcname>-<schedulename>-<timestamp>`. After creating the new snapshots,
          the oldest snapshots are removed if necessary, according to the retention
          policy of the schedule.

          Please see the [full
          documentation](https://backube.github.io/snapscheduler/) for more
          information.

          ## Requirements

          - OpenShift >= 4.7 (Kubernetes >= 1.20)
          - CSI-based storage driver that supports snapshots (i.e. has the
            `CREATE_DELETE_SNAPSHOT` capability)

          ## Examples

          The schedule for snapshotting is controlled by the
          `snapshotschedules.snapscheduler.backube` Custom Resource. This is a
          namespaced resource that applies only to the PersistentVolumeClaims in its
          namespace.

          Below is a simple example that keeps 7 daily (taken at midnight) snapshots
          of all PVCs in a given namespace:

          ```yaml
          ---
          apiVersion: snapscheduler.backube/v1
          kind: SnapshotSchedule
          metadata:
            name: daily
          spec:
            retention:
              maxCount: 7
            schedule: "0 0 * * *"
          ```

          See the [usage
          documentation](https://backube.github.io/snapscheduler/usage.html) for full
          details, including how to:

          - add label selectors to restrict which PVCs this schedule applies to
          - set the VolumeSnapshotClass used by the schedule
          - apply custom labels to the automatically created VolumeSnapshot objects
        displayName: SnapScheduler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - snapshot
        - storage
        - scheduler
        links:
        - name: Documentation
          url: https://backube.github.io/snapscheduler/
        - name: Source
          url: https://github.com/backube/snapscheduler
        - name: Blog
          url: https://backube.github.io/
        maintainers:
        - email: jstrunk@redhat.com
          name: John Strunk
        maturity: stable
        minKubeVersion: 1.20.0-0
        provider:
          name: Backube
          url: https://github.com/backube
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.19.1
        - quay.io/backube/snapscheduler:3.5.0
        version: 3.5.0
      entries:
      - name: snapscheduler.v3.5.0
        version: 3.5.0
      - name: snapscheduler.v3.4.0
        version: 3.4.0
      - name: snapscheduler.v3.3.0
        version: 3.3.0
      - name: snapscheduler.v3.2.0
        version: 3.2.0
      - name: snapscheduler.v3.1.0
        version: 3.1.0
      - name: snapscheduler.v3.0.0
        version: 3.0.0
      - name: snapscheduler.v2.1.0
        version: 2.1.0
      - name: snapscheduler.v2.0.0
        version: 2.0.0
      name: candidate
    - currentCSV: snapscheduler.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "snapscheduler.backube/v1",
                "kind": "SnapshotSchedule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "snapscheduler",
                    "app.kubernetes.io/instance": "daily",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "snapshotschedule",
                    "app.kubernetes.io/part-of": "snapscheduler"
                  },
                  "name": "daily"
                },
                "spec": {
                  "retention": {
                    "maxCount": 7
                  },
                  "schedule": "0 0 * * *"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/backube/snapscheduler:3.5.0
          createdAt: "2025-05-14T14:46:43Z"
          description: The SnapScheduler operator takes snapshots of Kubernetes CSI-based
            persistent volumes according to user-supplied schedules.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/snapscheduler
          support: https://github.com/backube/snapscheduler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SnapshotSchedule defines a schedule for taking automated
              snapshots of PVC(s)
            displayName: Snapshot Schedule
            kind: SnapshotSchedule
            name: snapshotschedules.snapscheduler.backube
            version: v1
        description: |-
          ## About this operator

          The SnapScheduler operator takes snapshots of CSI-based PersistentVolumes
          according to a configurable
          [Cron-like](https://en.wikipedia.org/wiki/Cron#Overview) schedule. The
          schedules include configurable retention policies for snapshots as well as
          selectors to limit the volumes that are snapshotted. An example schedule
          could be:

          > *Snapshot **all volumes** in a namespace **daily at midnight**, retaining
          > the most recent **7** snapshots for each volume.*

          Multiple schedules can be combined to provide more elaborate protection
          schemes. For example, a given volume (or collection of volumes) could be
          protected with:

          - 6 hourly snapshots
          - 7 daily snapshots
          - 4 weekly snapshots
          - 12 monthly snapshots

          ### How it works

          The operator watches for `SnapshotSchedule` CRs in each namespace. When the
          current time matches the schedule's cronspec, the operator creates a
          `VolumeSnapshot` object for each `PersistentVolumeClaim` in the namespace
          (or subset thereof if a label selector is provided). The `VolumeSnapshot`
          objects are named according to the template:
          `<pvcname>-<schedulename>-<timestamp>`. After creating the new snapshots,
          the oldest snapshots are removed if necessary, according to the retention
          policy of the schedule.

          Please see the [full
          documentation](https://backube.github.io/snapscheduler/) for more
          information.

          ## Requirements

          - OpenShift >= 4.7 (Kubernetes >= 1.20)
          - CSI-based storage driver that supports snapshots (i.e. has the
            `CREATE_DELETE_SNAPSHOT` capability)

          ## Examples

          The schedule for snapshotting is controlled by the
          `snapshotschedules.snapscheduler.backube` Custom Resource. This is a
          namespaced resource that applies only to the PersistentVolumeClaims in its
          namespace.

          Below is a simple example that keeps 7 daily (taken at midnight) snapshots
          of all PVCs in a given namespace:

          ```yaml
          ---
          apiVersion: snapscheduler.backube/v1
          kind: SnapshotSchedule
          metadata:
            name: daily
          spec:
            retention:
              maxCount: 7
            schedule: "0 0 * * *"
          ```

          See the [usage
          documentation](https://backube.github.io/snapscheduler/usage.html) for full
          details, including how to:

          - add label selectors to restrict which PVCs this schedule applies to
          - set the VolumeSnapshotClass used by the schedule
          - apply custom labels to the automatically created VolumeSnapshot objects
        displayName: SnapScheduler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - snapshot
        - storage
        - scheduler
        links:
        - name: Documentation
          url: https://backube.github.io/snapscheduler/
        - name: Source
          url: https://github.com/backube/snapscheduler
        - name: Blog
          url: https://backube.github.io/
        maintainers:
        - email: jstrunk@redhat.com
          name: John Strunk
        maturity: stable
        minKubeVersion: 1.20.0-0
        provider:
          name: Backube
          url: https://github.com/backube
        relatedImages:
        - quay.io/backube/snapscheduler:3.5.0
        - quay.io/brancz/kube-rbac-proxy:v0.19.1
        version: 3.5.0
      entries:
      - name: snapscheduler.v3.5.0
        version: 3.5.0
      - name: snapscheduler.v3.4.0
        version: 3.4.0
      - name: snapscheduler.v3.3.0
        version: 3.3.0
      - name: snapscheduler.v3.2.0
        version: 3.2.0
      - name: snapscheduler.v3.1.0
        version: 3.1.0
      - name: snapscheduler.v3.0.0
        version: 3.0.0
      - name: snapscheduler.v2.1.0
        version: 2.1.0
      - name: snapscheduler.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: snapscheduler
    provider:
      name: Backube
      url: https://github.com/backube
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CNCF
      provider-url: ""
    name: jaeger
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jaeger-operator.v1.65.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "jaegertracing.io/v1",
                "kind": "Jaeger",
                "metadata": {
                  "name": "my-jaeger"
                },
                "spec": {
                  "strategy": "allInOne"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing
          certified: "false"
          containerImage: quay.io/jaegertracing/jaeger-operator:1.65.0
          createdAt: "2025-01-22T20:40:19Z"
          description: Provides tracing, monitoring and troubleshooting for microservices-based
            distributed systems
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/jaegertracing/jaeger-operator
          support: Jaeger Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Jaeger is the Schema for the jaegers API
            displayName: Jaeger
            kind: Jaeger
            name: jaegers.jaegertracing.io
            version: v1
        description: |-
          Jaeger, inspired by [Dapper](https://research.google.com/pubs/pub36356.html) and [OpenZipkin](http://zipkin.io/), is a distributed tracing system released as open source by Uber Technologies. It is used for monitoring and troubleshooting microservices-based distributed systems.
          ### Core capabilities
          Jaeger is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed context propagation
          * Distributed transaction monitoring
          * Root cause analysis
          * Service dependency analysis
          * Performance / latency optimization
          * OpenTracing compatible data model
          * Multiple storage backends: Badger, Cassandra, Elasticsearch, Memory.
          ### Operator features
          * **Multiple modes** - Supports `allInOne`, `production` and `streaming` [modes of deployment](https://www.jaegertracing.io/docs/latest/operator/#deployment-strategies).
          * **Configuration** - The Operator manages [configuration information](https://www.jaegertracing.io/docs/latest/operator/#configuring-the-custom-resource) when installing Jaeger instances.
          * **Storage** - [Configure storage](https://www.jaegertracing.io/docs/latest/operator/#storage-options) used by Jaeger. By default, `memory` is used. Other options include `badger`, `cassandra` or `elasticsearch`. On OpenShift, the operator can delegate creation of an Elasticsearch cluster to the Elasticsearch Operator if deployed.
          * **Agent** - can be deployed as [sidecar](https://www.jaegertracing.io/docs/latest/operator/#auto-injecting-jaeger-agent-sidecars) (default) and/or [daemonset](https://www.jaegertracing.io/docs/latest/operator/#installing-the-agent-as-daemonset).
          * **UI** - Optionally setup ingress (Kubernetes) or secure route (OpenShift) to provide [access to the Jaeger UI](https://www.jaegertracing.io/docs/latest/operator/#accessing-the-jaeger-console-ui).
          ### Before you start
          1. Ensure that the appropriate storage solution, that will be used by the Jaeger instance, is available and configured.
          2. If intending to deploy an Elasticsearch cluster via the Jaeger custom resource, then the Elasticsearch Operator must first be installed.
          ### Troubleshooting
          * https://www.jaegertracing.io/docs/latest/troubleshooting/
        displayName: Community Jaeger Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tracing
        - monitoring
        - troubleshooting
        links:
        - name: Jaeger Operator Source Code
          url: https://github.com/jaegertracing/jaeger-operator
        maintainers:
        - email: jaeger-tracing@googlegroups.com
          name: Jaeger Google Group
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: CNCF
        relatedImages:
        - quay.io/jaegertracing/jaeger-operator:1.65.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 1.65.0
      entries:
      - name: jaeger-operator.v1.65.0
        version: 1.65.0
      - name: jaeger-operator.v1.62.0
        version: 1.62.0
      - name: jaeger-operator.v1.61.0
        version: 1.61.0
      - name: jaeger-operator.v1.60.0
        version: 1.60.0
      - name: jaeger-operator.v1.59.0
        version: 1.59.0
      - name: jaeger-operator.v1.57.0
        version: 1.57.0
      - name: jaeger-operator.v1.56.0
        version: 1.56.0
      - name: jaeger-operator.v1.55.0
        version: 1.55.0
      - name: jaeger-operator.v1.53.0
        version: 1.53.0
      - name: jaeger-operator.v1.52.0
        version: 1.52.0
      - name: jaeger-operator.v1.51.0
        version: 1.51.0
      - name: jaeger-operator.v1.49.0
        version: 1.49.0
      - name: jaeger-operator.v1.48.1
        version: 1.48.1
      - name: jaeger-operator.v1.47.0
        version: 1.47.0
      - name: jaeger-operator.v1.46.0
        version: 1.46.0
      - name: jaeger-operator.v1.45.0
        version: 1.45.0
      - name: jaeger-operator.v1.44.0
        version: 1.44.0
      - name: jaeger-operator.v1.43.0
        version: 1.43.0
      - name: jaeger-operator.v1.42.0
        version: 1.42.0
      - name: jaeger-operator.v1.41.1
        version: 1.41.1
      - name: jaeger-operator.v1.41.0
        version: 1.41.0
      - name: jaeger-operator.v1.40.0
        version: 1.40.0
      - name: jaeger-operator.v1.39.0
        version: 1.39.0
      - name: jaeger-operator.v1.38.1
        version: 1.38.1
      - name: jaeger-operator.v1.38.0
        version: 1.38.0
      - name: jaeger-operator.v1.37.0
        version: 1.37.0
      - name: jaeger-operator.v1.36.0
        version: 1.36.0
      - name: jaeger-operator.v1.35.0
        version: 1.35.0
      - name: jaeger-operator.v1.34.1
        version: 1.34.1
      - name: jaeger-operator.v1.34.0
        version: 1.34.0
      - name: jaeger-operator.v1.33.0
        version: 1.33.0
      - name: jaeger-operator.v1.32.0
        version: 1.32.0
      - name: jaeger-operator.v1.31.0
        version: 1.31.0
      - name: jaeger-operator.v1.30.0
        version: 1.30.0
      - name: jaeger-operator.v1.29.1
        version: 1.29.1
      - name: jaeger-operator.v1.25.0
        version: 1.25.0
      - name: jaeger-operator.v1.24.0
        version: 1.24.0
      name: stable
    defaultChannel: stable
    packageName: jaeger
    provider:
      name: CNCF
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: JWS team
      provider-url: ""
    name: jws-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jws-operator.v2.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "web.servers.org/v1alpha1",
                "kind": "WebServer",
                "metadata": {
                  "name": "webserver-example"
                },
                "spec": {
                  "applicationName": "webapp",
                  "replicas": 1,
                  "useInsightsClient": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/jboss-webserver-5/jws5-rhel8-operator@sha256:ea042277bdcdb70b2dc32083ea868efd4bf474d5a8ec0f9b47f5bea35aaeae1e
          createdAt: "2025-09-12T08:31:52.558719Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat JBoss Web Server",
            "Red Hat Runtimes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/web-servers/jws-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Web Server is the schema for the webservers API
            displayName: Web Server
            kind: WebServer
            name: webservers.web.servers.org
            version: v1alpha1
        description: WebServer operator can manage JBoss Web Server images.
        displayName: WebServer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WebServer
        - JBoss
        links:
        - name: Jws Operator
          url: https://jws-operator.domain
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: JWS team
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:021ff671b0fb72eed635b05489ec86342fcb7ada3ba0353591fdb9328871d067
        - registry.redhat.io/jboss-webserver-5/jws5-rhel8-operator@sha256:ea042277bdcdb70b2dc32083ea868efd4bf474d5a8ec0f9b47f5bea35aaeae1e
        - registry.redhat.io/insights-runtimes-tech-preview/runtimes-inventory-rhel8-operator@sha256:58e01d11c3d30abc5b0ee3795511bd04e39aea21e2853b4f4cc6d1c8a2bd7bb0
        version: 2.2.0
      entries:
      - name: jws-operator.v2.2.0
        version: 2.2.0
      - name: jws-operator.v2.1.3
        version: 2.1.3
      - name: jws-operator.v2.1.2
        version: 2.1.2
      - name: jws-operator.v2.1.1
        version: 2.1.1
      - name: jws-operator.v2.1.0
        version: 2.1.0
      - name: jws-operator.v2.0.16
        version: 2.0.16
      - name: jws-operator.v2.0.15
        version: 2.0.15
      - name: jws-operator.v2.0.14
        version: 2.0.14
      - name: jws-operator.v2.0.13
        version: 2.0.13
      - name: jws-operator.v2.0.12
        version: 2.0.12
      - name: jws-operator.v2.0.11
        version: 2.0.11
      - name: jws-operator.v2.0.10
        version: 2.0.10
      - name: jws-operator.v2.0.9
        version: 2.0.9
      - name: jws-operator.v2.0.8
        version: 2.0.8
      - name: jws-operator.v2.0.7
        version: 2.0.7
      - name: jws-operator.v2.0.6
        version: 2.0.6
      - name: jws-operator.v2.0.5
        version: 2.0.5
      - name: jws-operator.v2.0.4
        version: 2.0.4
      - name: jws-operator.v2.0.3
        version: 2.0.3
      - name: jws-operator.v2.0.2
        version: 2.0.2
      - name: jws-operator.v2.0.1
        version: 2.0.1
      name: alpha
    defaultChannel: alpha
    packageName: jws-operator
    provider:
      name: JWS team
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Windup
      provider-url: ""
    name: windup-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: windup-operator.v0.0.25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
             {
               "apiVersion": "windup.jboss.org/v1",
               "kind": "Windup",
               "metadata": {
                 "name": "windup",
                 "labels": {
                   "application": "windup"
                 }
               },
               "spec": {}
             }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/windupeng/windup-operator-native:0.0.25
          createdAt: 01/22/24
          description: Windup is an analysis tool that supports the modernization
            and migration of Java applications.
          imagePullPolicy: Always
          operatorframework.io/suggested-namespace: windup
          repository: https://github.com/windup/windup-operator/
          support: https://issues.redhat.com/projects/WINDUP
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "Create a Windup Web Console application.  \n\nTo access
              the Windup Web Console application from the \"Developer\" perspective,
              please go to Topology and click on the \"Open URL\" icon on the web-console
              pod.  \n\nAlternatively from the \"Administrator\" perspective, please
              go to Networking->Routes and click on the Location hyperlink.  \n\nKnown
              issue  \nIf you want to customize the Windup Web Console instance's
              parameters and you can not see any in the `Form view`, please switch
              to the `YAML view` and change them as needed.\n"
            displayName: Windup
            kind: Windup
            name: windups.windup.jboss.org
            version: v1
        description: |
          Windup is a web console application that supports large-scale Java application modernization and migration projects across a broad range of transformations and use cases.

          It analyzes application code, supports effort estimation, accelerates code migration, and enables users to move applications to containers.

          For more information please refer to the [https://windup.github.io/](https://windup.github.io/) page.
        displayName: Windup Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - modernization
        - windup
        links:
        - name: Website
          url: https://windup.github.io
        - name: GitHub
          url: https://github.com/windup/windup-operator
        maintainers:
        - email: migrate@redhat.com
          name: Windup
        maturity: alpha
        provider:
          name: Windup
        relatedImages:
        - quay.io/windupeng/windup-operator-native:0.0.25
        version: 0.0.25
      entries:
      - name: windup-operator.v0.0.25
        version: 0.0.25
      - name: windup-operator.v0.0.24
        version: 0.0.24
      - name: windup-operator.v0.0.23
        version: 0.0.23
      - name: windup-operator.v0.0.22
        version: 0.0.22
      - name: windup-operator.v0.0.21
        version: 0.0.21
      - name: windup-operator.v0.0.20
        version: 0.0.20
      - name: windup-operator.v0.0.19
        version: 0.0.19
      - name: windup-operator.v0.0.18
        version: 0.0.18
      - name: windup-operator.v0.0.17
        version: 0.0.17
      - name: windup-operator.v0.0.16
        version: 0.0.16
      - name: windup-operator.v0.0.15
        version: 0.0.15
      - name: windup-operator.v0.0.14
        version: 0.0.14
      - name: windup-operator.v0.0.13
        version: 0.0.13
      - name: windup-operator.v0.0.12
        version: 0.0.12
      - name: windup-operator.v0.0.11
        version: 0.0.11
      name: alpha
    defaultChannel: alpha
    packageName: windup-operator
    provider:
      name: Windup
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: watcher-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: watcher-operator.v1.0.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "Watcher",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcher",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcher",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcher"
                },
                "spec": {
                  "apiServiceTemplate": {
                    "tls": {
                      "caBundleSecretName": "combined-ca-bundle"
                    }
                  },
                  "databaseInstance": "openstack"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherAPI",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcher-operator"
                  },
                  "name": "watcherapi-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherApplier",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcherapplier-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcherapplier",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcherapplier-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherDecisionEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcherdecisionengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcherdecisionengine",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcherdecisionengine-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-09-02T13:32:31Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.14
          operatorframework.io/suggested-namespace: openstack-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: WatcherAPI is the Schema for the watcherapis API
            displayName: Watcher API
            kind: WatcherAPI
            name: watcherapis.watcher.openstack.org
            version: v1beta1
          - description: WatcherApplier is the Schema for the watcherappliers API
            displayName: Watcher Applier
            kind: WatcherApplier
            name: watcherappliers.watcher.openstack.org
            version: v1beta1
          - description: WatcherDecisionEngine is the Schema for the watcherdecisionengines
              API
            displayName: Watcher Decision Engine
            kind: WatcherDecisionEngine
            name: watcherdecisionengines.watcher.openstack.org
            version: v1beta1
          - description: Watcher is the Schema for the watchers API
            displayName: Watcher
            kind: Watcher
            name: watchers.watcher.openstack.org
            version: v1beta1
        description: The Watcher Operator project
        displayName: Watcher Operator (Technology Preview)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        links:
        - name: Watcher Operator
          url: https://github.com/openstack-k8s-operators/watcher-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/watcher-rhel9-operator@sha256:63c341a3e20a522c601098502a8fd8f1164d23c773ae8f55782c43ff4b6c8c18
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:eed8234b0fe47083ab941f513709f778bfcb80e32747b71a65fd5f62328a5ca0
        - registry.redhat.io/rhoso/openstack-watcher-api-rhel9@sha256:e97369e3a47bb882f1920c6af3f6345a842af45c58f0f5c5df9ac2933cf3098a
        - registry.redhat.io/rhoso/openstack-watcher-decision-engine-rhel9@sha256:c759d66cf0986c0036be91c008cbd059b4412c94160ed82dd12cd4e00d0685a9
        - registry.redhat.io/rhoso/openstack-watcher-applier-rhel9@sha256:c92a82f901613991926367bb8aab4f875c27a3e4a5282d92847a8463d98cd9e3
        version: 1.0.14
      entries:
      - name: watcher-operator.v1.0.14
        version: 1.0.14
      - name: watcher-operator.v1.0.13
        version: 1.0.13
      - name: watcher-operator.v1.0.12
        version: 1.0.12
      - name: watcher-operator.v1.0.11
        version: 1.0.11
      - name: watcher-operator.v1.0.10
        version: 1.0.10
      - name: watcher-operator.v1.0.9
        version: 1.0.9
      - name: watcher-operator.v1.0.8
        version: 1.0.8
      - name: watcher-operator.v1.0.7
        version: 1.0.7
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: watcher-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: VMware Tanzu
      provider-url: ""
    name: rabbitmq-cluster-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-cluster-operator.v2.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "rabbitmq.com/v1beta1",
                "kind": "RabbitmqCluster",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "override": {
                    "statefulset": {
                      "spec": {
                        "template": {
                          "spec": {
                            "containers": [],
                            "securityContext": {}
                          }
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/rabbitmqoperator/cluster-operator:2.16.0
          createdAt: 2025-08-12T09:34:33
          repository: https://github.com/rabbitmq/cluster-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Rabbitmq Cluster
            displayName: RabbitmqCluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: "## About the managed application\nRabbitMQ is an open-source
          message broker that originally implemented the\nAdvanced Message Queuing
          Protocol (AMQP). Nowadays features a plug-in\narchitecture supporting protocols
          such as Streaming Text Oriented Messaging\nProtocol (STOMP), MQ Telemetry
          Transport (MQTT), and others. RabbitMQ\nsupports native streaming and AMQP
          1.0.\n\nVisit [rabbitmq.com](rabbitmq.com) for more information.\n\n## About
          this Operator\nRabbitMQ Cluster Kubernetes Operator is developed and maintened
          by the VMware \nRabbitMQ team and provides a consistent and easy way to
          deploy RabbitMQ\nclusters to Kubernetes and run them, including \"day two\"
          (continuous)\noperations. RabbitMQ clusters deployed using the Operator
          can be used by\napplications running on Kubernetes or outside of Kubernetes.\n\nRead
          [RabbitMQ operator docs](https://www.rabbitmq.com/kubernetes/operator/operator-overview)\nfor
          more information.\n"
        displayName: RabbitMQ-cluster-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        - messaging
        - amqp
        - amqp10
        - stomp
        - mqtt
        - mqtt5
        - queues
        - streaming
        links:
        - name: Documentation
          url: https://www.rabbitmq.com/kubernetes/operator/operator-overview.html
        - name: GitHub
          url: https://github.com/rabbitmq/cluster-operator
        maintainers:
        - email: rabbitmq-users@googlegroups.com
          name: VMware Tanzu
        maturity: stable
        minKubeVersion: 1.26.0
        nativeApis:
        - group: apps
          kind: Deployment
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: ReplicaSet
          version: v1
        - group: ""
          kind: Pod
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: ConfigMap
          version: v1
        provider:
          name: VMware Tanzu
        relatedImages:
        - quay.io/rabbitmqoperator/cluster-operator:2.16.0
        version: 2.16.0
      entries:
      - name: rabbitmq-cluster-operator.v2.16.0
        version: 2.16.0
      - name: rabbitmq-cluster-operator.v2.15.0
        version: 2.15.0
      - name: rabbitmq-cluster-operator.v2.14.0
        version: 2.14.0
      - name: rabbitmq-cluster-operator.v2.13.0
        version: 2.13.0
      - name: rabbitmq-cluster-operator.v2.12.1
        version: 2.12.1
      - name: rabbitmq-cluster-operator.v2.10.0
        version: 2.10.0
      - name: rabbitmq-cluster-operator.v2.9.0
        version: 2.9.0
      - name: rabbitmq-cluster-operator.v2.8.0
        version: 2.8.0
      - name: rabbitmq-cluster-operator.v2.7.0
        version: 2.7.0
      - name: rabbitmq-cluster-operator.v2.6.0
        version: 2.6.0
      - name: rabbitmq-cluster-operator.v2.5.0
        version: 2.5.0
      - name: rabbitmq-cluster-operator.v2.4.0
        version: 2.4.0
      - name: rabbitmq-cluster-operator.v2.3.0
        version: 2.3.0
      - name: rabbitmq-cluster-operator.v2.2.0
        version: 2.2.0
      - name: rabbitmq-cluster-operator.v2.1.0
        version: 2.1.0
      - name: rabbitmq-cluster-operator.v2.0.0
        version: 2.0.0
      - name: rabbitmq-cluster-operator.v1.14.0
        version: 1.14.0
      name: stable
    defaultChannel: stable
    packageName: rabbitmq-cluster-operator
    provider:
      name: VMware Tanzu
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: ""
    name: ibm-metrics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-metrics-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MarketplaceConfig",
                "metadata": {
                  "name": "marketplaceconfig"
                },
                "spec": {
                  "features": {
                    "deployment": true,
                    "meterDefinitionCatalogServer": false,
                    "registration": true
                  },
                  "installIBMCatalogSource": true,
                  "isDisconnected": false,
                  "license": {
                    "accept": false
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterBase",
                "metadata": {
                  "name": "meterbase"
                },
                "spec": {
                  "enabled": true,
                  "prometheus": {
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "1",
                        "memory": "1G"
                      }
                    },
                    "storage": {
                      "size": "20Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterReport",
                "metadata": {
                  "name": "example-meterreport"
                },
                "spec": {
                  "endTime": "2020-08-05T00:00:00Z",
                  "prometheusService": {
                    "name": "rhm-prometheus-meterbase",
                    "namespace": "openshift-redhat-marketplace",
                    "targetPort": "rbac"
                  },
                  "startTime": "2020-08-04T00:00:00Z"
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "example-meterdefinition-4"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "meterdefinition-sample"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
          createdAt: "2025-08-28T14:18:55Z"
          description: The IBM Metrics Operator provides workload metering and reporting
            for IBM and Red Hat Marketplace customers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
          - description: MeterBase is the resource that sets up Metering for Red Hat
              Marketplace. This is an internal resource not meant to be modified directly.
            displayName: Meter Base
            kind: MeterBase
            name: meterbases.marketplace.redhat.com
            version: v1alpha1
          - description: MeterDefinition defines the meter workloads used to enable
              pay for use billing.
            displayName: Meter Definition
            kind: MeterDefinition
            name: meterdefinitions.marketplace.redhat.com
            version: v1beta1
          - kind: MeterReport
            name: meterreports.marketplace.redhat.com
            version: v1alpha1
        description: "The IBM Metrics Operator provides workload metering and reporting
          for IBM Software Central and Red Hat Marketplace customers.\n### **Important
          Note**\nA set of instructions for onboarding is provided here. For more
          detailed onboarding instructions or information about what is installed
          please visit [swc.saas.ibm.com](https://swc.saas.ibm.com).\n\nUsage metrics
          may be monitored through [https://swc.saas.ibm.com](https://swc.saas.ibm.com)
          with only IBM Metrics Operator and a Red Hat Marketplace account.\n\n###
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n### Prerequisites\n1. Installations are required
          to [enable monitoring for user-defined projects](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          as the Prometheus provider.\n2. Edit the cluster-monitoring-config ConfigMap
          object:\n\n   ```sh\n   $ oc -n openshift-monitoring edit configmap cluster-monitoring-config\n
          \   ```\n\n3. Add enableUserWorkload: true under data/config.yaml:\n  \n
          \   ```sh\n    apiVersion: v1\n    kind: ConfigMap\n    metadata:\n      name:
          cluster-monitoring-config\n        namespace: openshift-monitoring\n    data:\n
          \     config.yaml: |\n        enableUserWorkload: true\n    ```\n\n4. Configure
          the user-workload-monitoring-config ConfigMap object:\n\n    ```sh\n    $
          oc -n openshift-user-workload-monitoring edit configmap user-workload-monitoring-config\n
          \   ```\n\n5. Configure a minimum retention time of 168h and minimum storage
          capacity of 40Gi\n  \n    ```sh\n    apiVersion: v1\n    kind: ConfigMap\n
          \   metadata:\n      name: user-workload-monitoring-config\n      namespace:
          openshift-user-workload-monitoring\n\n    data:\n      config.yaml: |\n
          \       prometheus:\n          retention: 168h\n          volumeClaimTemplate:\n
          \           spec:\n              resources:\n                requests:\n
          \                 storage: 40Gi\n    ```\n\n### Resources Required\n\nMinimum
          system resources required:\n\n| Operator                | Memory (MB) |
          CPU (cores) | Disk (GB) | Nodes |\n| ----------------------- | -----------
          | ----------- | --------- | ----- |\n| **Metrics**   |        750  |     0.25
          \   | 3x1       |    3  |\n| **Deployment** |        250  |     0.25    |
          -         |    1  |\n\n| Prometheus Provider  | Memory (GB) | CPU (cores)
          | Disk (GB) | Nodes |\n| --------- | ----------- | ----------- | ---------
          | ----- |\n| **[Openshift User Workload Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)**
          |          1  |     0.1       | 2x40        |   2    |\n\nMultiple nodes
          are required to provide pod scheduling for high availability for IBM Metrics
          Operator Data Service and Prometheus.\n\nThe IBM Metrics Operator automatically
          creates 3 x 1Gi PersistentVolumeClaims to store reports as part of the data
          service, with _ReadWriteOnce_ access mode. Te PersistentVolumeClaims are
          automatically created by the ibm-metrics-operator after creating a `redhat-marketplace-pull-secret`
          and accepting the license in `marketplaceconfig`.\n\n| NAME                                |
          CAPACITY | ACCESS MODES |\n| ----------------------------------- | --------
          | ------------ |\n| rhm-data-service-rhm-data-service-0 | 1Gi | RWO |\n|
          rhm-data-service-rhm-data-service-1 | 1Gi | RWO |\n| rhm-data-service-rhm-data-service-2
          | 1Gi | RWO |\n\n### Supported Storage Providers\n\n- OpenShift Container
          Storage / OpenShift Data Foundation version 4.x, from version 4.2 or higher\n-
          IBM Cloud Block storage and IBM Cloud File storage\n- IBM Storage Suite
          for IBM Cloud Paks:\n  - File storage from IBM Spectrum Fusion/Scale \n
          \ - Block storage from IBM Spectrum Virtualize, FlashSystem or DS8K\n- Portworx
          Storage, version 2.5.5 or above\n- Amazon Elastic File Storage\n\n### Access
          Modes required\n\n - ReadWriteOnce (RWO)\n\n### Provisioning Options supported\n\nChoose
          one of the following options to provision storage for the ibm-metrics-operator
          data-service\n\n#### Dynamic provisioning using a default StorageClass\n
          \  - A StorageClass is defined with a `metadata.annotations: storageclass.kubernetes.io/is-default-class:
          \"true\"`\n   - PersistentVolumes will be provisioned automatically for
          the generated PersistentVolumeClaims\n--- \n#### Manually create each PersistentVolumeClaim
          with a specific StorageClass\n   - Must be performed before creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`. Otherwise, the automatically
          generated PersistentVolumeClaims are immutable.\n```\napiVersion: v1\nkind:
          PersistentVolumeClaim\nmetadata:\n  labels:\n    app: rhm-data-service\n
          \ name: rhm-data-service-rhm-data-service-0\n  namespace: ibm-software-central\nspec:\n
          \ storageClassName: rook-cephfs\n  accessModes:\n  - ReadWriteOnce\n  resources:\n
          \   requests:\n      storage: 1Gi\n```\n---\n#### Manually provision each
          PersistentVolume for the generated PersistentVolumeClaims with a specific
          StorageClass\n  - May be performed before or after creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`.  \n```\napiVersion: v1\nkind:
          PersistentVolume\nmetadata:\n  name: rhm-data-service-rhm-data-service-0\nspec:\n
          \ csi:\n    driver: rook-ceph.cephfs.csi.ceph.com\n    volumeHandle: rhm-data-service-rhm-data-service-0\n
          \ capacity:\n    storage: 1Gi\n  accessModes:\n    - ReadWriteOnce\n  persistentVolumeReclaimPolicy:
          Delete\n  storageClassName: rook-cephfs\n  volumeMode: Filesystem\n  claimRef:\n
          \   kind: PersistentVolumeClaim\n    namespace: ibm-software-central\n    name:
          rhm-data-service-rhm-data-service-0\n```\n\n### Installation\n1. Create
          or get your [pull secret](https://swc.saas.ibm.com/en-us/documentation/clusters#get-pull-secret).\n2.
          Install the IBM Metrics Operator\n3. Create a Kubernetes secret in the installed
          namespace with the name `redhat-marketplace-pull-secret` and key `PULL_SECRET`
          with the value of the pull secret.\n    ```sh\n    # Replace ${PULL_SECRET}
          with your pull secret\n    oc create secret generic redhat-marketplace-pull-secret
          -n ibm-software-central --from-literal=PULL_SECRET=${PULL_SECRET}\n    ```\n4.
          Use of the platform is governed by the:\n\n    [IBM Cloud Services Agreement](https://www.ibm.com/support/customer/csol/terms/?id=Z126-6304_WS&_ga=2.116312197.2046730452.1684328846-812467790.1684328846)
          (or other base agreement between you and IBM such as a [Passport Advantage
          Agreement](https://www.ibm.com/software/passportadvantage/pa_agreements.html?_ga=2.116312197.2046730452.1684328846-812467790.1684328846))
          and the [Service Description for the Red Hat Marketplace](https://www.ibm.com/support/customer/csol/terms/?id=i126-8719&_ga=2.83289621.2046730452.1684328846-812467790.1684328846).\n
          \   \n5. Update MarketplaceConfig to accept the license.\n    ```\n    oc
          patch marketplaceconfig marketplaceconfig -n ibm-software-central --type='merge'
          -p '{\"spec\": {\"license\": {\"accept\": true}}}'\n    ```\n6. Install
          the pull secret as a global pull secret on the cluster.\n\n    These steps
          require `oc`, `jq`, and `base64` to be available on your machine.\n\n    ```sh\n
          \   # Create the docker pull secret file using your pull secret.\n    #
          Store it in a file called entitledregistryconfigjson.\n    oc create secret
          docker-registry entitled-registry --docker-server=registry.marketplace.redhat.com
          --docker-username \"cp\" --docker-password \"${PULL_SECRET}\" --dry-run=client
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > entitledregistryconfigjson\n
          \   # Get the current global secrets on the cluster and store it as a file
          named dockerconfigjson\n    oc get secret pull-secret -n openshift-config
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > dockerconfigjson\n
          \   # Merge the two dockerconfigs together into a file called dockerconfigjson-merged.\n
          \   jq -s '.[0] * .[1]' dockerconfigjson entitledregistryconfigjson > dockerconfigjson-merged\n
          \   # Set the cluster's dockerconfig file to the new merged version.\n    oc
          set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson=dockerconfigjson-merged\n
          \   ```\n  7. (Optional) Your IBM product may have installed IBM Metrics
          Operator using an `ibm-entitlement-key` for authentication to IBM Software
          Central. To authorize the automatic creation of an IBM Software Central
          account for the IBM account owning the `ibm-entitlement-key`\n      ```\n
          \     oc patch marketplaceconfig marketplaceconfig -n ibm-software-central
          --type='merge' -p '{\"spec\": {\"authorizeAccountCreation\": true}}'\n      ```\n\n###
          Why is a global pull secret required?\nIn order to successfully install
          the products hosted by the container image registry, you will need to make
          the pull secret available across the cluster. This can be achieved by applying
          the pull as a [global pull secret](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html#images-update-global-pull-secret_using-image-pull-secrets).
          For alternative approaches, please see the official OpenShift [documentation](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html).\n\n\n###
          SecurityContextConstraints requirements\n\nThe Operators and their components
          support running under the OpenShift Container Platform default restricted
          and restricted-v2 security context constraints.\n\n### Installation Namespace
          and ClusterRoleBinding requirements\n\nThe IBM Metrics Operator components
          require specific ClusterRoleBindings.\n- The metric-state component requires
          a ClusterRoleBinding for the the `view` ClusterRole.\n  - Ability to read
          non-sensitive CustomResources. \n  - The `view` ClusterRole dynamically
          adds new CustomResourceDefinitions.\n- The operator & reporter component
          requires a ClusterRoleBinding for the the `cluster-monitoring-view` ClusterRole.\n
          \ - Ability to view Prometheus metrics. \n  - The underlying ClusterRole
          RBAC often updated between OpenShift versions.\n\nDue to limitations of
          Operator Lifecycle Manager (OLM), this ClusterRoleBinding can not be provided
          dynamically for arbitrary installation target namespaces.\n\nA static ClusterRoleBinding
          is included for installation to the default namespace of `ibm-software-central`,
          and namespaces `openshift-redhat-marketplace`,  `redhat-marketplace`, `ibm-common-services`.\n\nTo
          create the ClusterRoleBindings for installation to an alternate namespace\n```\noc
          project INSTALL-NAMESPACE\noc adm policy add-cluster-role-to-user view -z
          ibm-metrics-operator-metric-state\noc adm policy add-cluster-role-to-user
          cluster-monitoring-view -z ibm-metrics-operator-controller-manager,ibm-metrics-operator-reporter\n```\n\n###
          Metric State scoping requirements\nThe metric-state Deployment obtains `get/list/watch`
          access to metered resources via the `view` ClusterRole. For operators deployed
          using Operator Lifecycle Manager (OLM), permissions are added to `clusterrole/view`
          dynamically via a generated and annotated `-view` ClusterRole. If you wish
          to meter an operator, and its Custom Resource Definitions (CRDs) are not
          deployed through OLM, one of two options are required\n1. Add the following
          label to a clusterrole that has get/list/watch access to your CRD: `rbac.authorization.k8s.io/aggregate-to-view:
          \"true\"`, thereby dynamically adding it to `clusterrole/view`\n2. Create
          a ClusterRole that has get/list/watch access to your CRD, and create a ClusterRoleBinding
          for the metric-state ServiceAccount\n\nAttempting to meter a resource with
          a MeterDefinition without the required permissions will log an `AccessDeniedError`
          in metric-state.\n\n### Disaster Recovery\n\nTo plan for disaster recovery,
          note the PhysicalVolumeClaims `rhm-data-service-rhm-data-service-N`. \n-
          In connected environments, MeterReport data upload attempts occur hourly,
          and are then removed from data-service. There is a low risk of losing much
          unreported data.\n- In an airgap environment, MeterReport data must be pulled
          from data-service and uploaded manually using [datactl](https://github.com/redhat-marketplace/datactl/).
          To prevent data loss in a disaster scenario, the data-service volumes should
          be considered in a recovery plan.\n\n### Subscription Config\n\nIt is possible
          to [configure](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md)
          how OLM deploys an Operator via the `config` field in the [Subscription](https://github.com/operator-framework/olm-book/blob/master/docs/subscriptions.md)
          object.\n\nThe IBM Metrics Operator will also read the `config` and append
          it to the operands. The primary use case is to control scheduling using
          [Tolerations](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/)
          and [NodeSelectors](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/).\n\nA
          limitation is that the `config` elements are only appended to the operands.
          The elements in the operands are not removed if the `config` is removed
          from the `Subscripion`. The operand must be modified manually, or deleted
          and recreated by the controller.\n\n### Cluster permission requirements\n\n|API
          group             |Resources              |Verbs                                     |\n|----------------------|-----------------------|------------------------------------------|\n|apps
          \                 |statefulsets           |get;list;watch                            |\n|authentication.k8s.io
          |tokenreviews           |create                                    |\n|authorization.k8s.io
          \ |subjectaccessreviews   |create                                    |\n|config.openshift.io
          \  |clusterversions        |get;list;watch                            |\n|marketplace.redhat.com|meterdefinitions
          \      |get;list;watch;create;update;patch;delete |\n|operators.coreos.com
          \ |clusterserviceversions |get;list;watch                            |\n|operators.coreos.com
          \ |subscriptions          |get;list;watch                            |\n|operators.coreos.com
          \ |operatorgroups         |get;list                                  |\n|storage.k8s.io
          \       |storageclasses         |get;list;watch                            |\n|
          \                     |namespaces             |get;list;watch                            |\n|
          \                     |configmaps             |get;list;watch                            |\n|
          \                     |services               |get;list;watch                            |\n|
          \                     |**clusterrole/view**   |                                          |\n\n\n###
          Documentation\nYou can find our documentation [here.](https://swc.saas.ibm.com/en-us/documentation/)\n\n###
          Getting help\nIf you encounter any issues while using the operators, you
          can create an issue on our [Github\nrepo](https://github.com/redhat-marketplace/redhat-marketplace-operator)
          for bugs, enhancements, or other requests. You can also visit our main page
          and\nreview our [support](https://swc.saas.ibm.com/en-us/support) and [documentation](https://swc.saas.ibm.com/en-us/documentation/).\n\n###
          Readme\nYou can find our readme [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/README.md)\n\n###
          License information\nYou can find our license information [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/LICENSE)\n"
        displayName: IBM Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        - buy
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/redhat-marketplace-authcheck@sha256:13a81b8aca79ae8c89ee24dcc30f3322a4cfdedf929e7bb984295c1e7e631fa9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/redhat-marketplace-data-service@sha256:30fd8d2996b921cfce51f697ee4490a6d9a836021952429b877336f5edf2749d
        - quay.io/rh-marketplace/redhat-marketplace-reporter@sha256:2bd098f057ca07f2b14be9b60ea6ad8ca6b0cae1d226b1ca3b27e23d04ccdf20
        - quay.io/rh-marketplace/redhat-marketplace-metric-state@sha256:e4014c9a546c10e84b13a09db6a18431ee9489ad1296808635e48d9ef4963cf1
        - quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
        version: 2.20.0
      entries:
      - name: ibm-metrics-operator.v2.20.0
        version: 2.20.0
      name: beta
    - currentCSV: ibm-metrics-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MarketplaceConfig",
                "metadata": {
                  "name": "marketplaceconfig"
                },
                "spec": {
                  "features": {
                    "deployment": true,
                    "meterDefinitionCatalogServer": false,
                    "registration": true
                  },
                  "installIBMCatalogSource": true,
                  "isDisconnected": false,
                  "license": {
                    "accept": false
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterBase",
                "metadata": {
                  "name": "meterbase"
                },
                "spec": {
                  "enabled": true,
                  "prometheus": {
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "1",
                        "memory": "1G"
                      }
                    },
                    "storage": {
                      "size": "20Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterReport",
                "metadata": {
                  "name": "example-meterreport"
                },
                "spec": {
                  "endTime": "2020-08-05T00:00:00Z",
                  "prometheusService": {
                    "name": "rhm-prometheus-meterbase",
                    "namespace": "openshift-redhat-marketplace",
                    "targetPort": "rbac"
                  },
                  "startTime": "2020-08-04T00:00:00Z"
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "example-meterdefinition-4"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "meterdefinition-sample"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
          createdAt: "2025-08-28T14:18:55Z"
          description: The IBM Metrics Operator provides workload metering and reporting
            for IBM and Red Hat Marketplace customers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
          - description: MeterBase is the resource that sets up Metering for Red Hat
              Marketplace. This is an internal resource not meant to be modified directly.
            displayName: Meter Base
            kind: MeterBase
            name: meterbases.marketplace.redhat.com
            version: v1alpha1
          - description: MeterDefinition defines the meter workloads used to enable
              pay for use billing.
            displayName: Meter Definition
            kind: MeterDefinition
            name: meterdefinitions.marketplace.redhat.com
            version: v1beta1
          - kind: MeterReport
            name: meterreports.marketplace.redhat.com
            version: v1alpha1
        description: "The IBM Metrics Operator provides workload metering and reporting
          for IBM Software Central and Red Hat Marketplace customers.\n### **Important
          Note**\nA set of instructions for onboarding is provided here. For more
          detailed onboarding instructions or information about what is installed
          please visit [swc.saas.ibm.com](https://swc.saas.ibm.com).\n\nUsage metrics
          may be monitored through [https://swc.saas.ibm.com](https://swc.saas.ibm.com)
          with only IBM Metrics Operator and a Red Hat Marketplace account.\n\n###
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n### Prerequisites\n1. Installations are required
          to [enable monitoring for user-defined projects](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          as the Prometheus provider.\n2. Edit the cluster-monitoring-config ConfigMap
          object:\n\n   ```sh\n   $ oc -n openshift-monitoring edit configmap cluster-monitoring-config\n
          \   ```\n\n3. Add enableUserWorkload: true under data/config.yaml:\n  \n
          \   ```sh\n    apiVersion: v1\n    kind: ConfigMap\n    metadata:\n      name:
          cluster-monitoring-config\n        namespace: openshift-monitoring\n    data:\n
          \     config.yaml: |\n        enableUserWorkload: true\n    ```\n\n4. Configure
          the user-workload-monitoring-config ConfigMap object:\n\n    ```sh\n    $
          oc -n openshift-user-workload-monitoring edit configmap user-workload-monitoring-config\n
          \   ```\n\n5. Configure a minimum retention time of 168h and minimum storage
          capacity of 40Gi\n  \n    ```sh\n    apiVersion: v1\n    kind: ConfigMap\n
          \   metadata:\n      name: user-workload-monitoring-config\n      namespace:
          openshift-user-workload-monitoring\n\n    data:\n      config.yaml: |\n
          \       prometheus:\n          retention: 168h\n          volumeClaimTemplate:\n
          \           spec:\n              resources:\n                requests:\n
          \                 storage: 40Gi\n    ```\n\n### Resources Required\n\nMinimum
          system resources required:\n\n| Operator                | Memory (MB) |
          CPU (cores) | Disk (GB) | Nodes |\n| ----------------------- | -----------
          | ----------- | --------- | ----- |\n| **Metrics**   |        750  |     0.25
          \   | 3x1       |    3  |\n| **Deployment** |        250  |     0.25    |
          -         |    1  |\n\n| Prometheus Provider  | Memory (GB) | CPU (cores)
          | Disk (GB) | Nodes |\n| --------- | ----------- | ----------- | ---------
          | ----- |\n| **[Openshift User Workload Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)**
          |          1  |     0.1       | 2x40        |   2    |\n\nMultiple nodes
          are required to provide pod scheduling for high availability for IBM Metrics
          Operator Data Service and Prometheus.\n\nThe IBM Metrics Operator automatically
          creates 3 x 1Gi PersistentVolumeClaims to store reports as part of the data
          service, with _ReadWriteOnce_ access mode. Te PersistentVolumeClaims are
          automatically created by the ibm-metrics-operator after creating a `redhat-marketplace-pull-secret`
          and accepting the license in `marketplaceconfig`.\n\n| NAME                                |
          CAPACITY | ACCESS MODES |\n| ----------------------------------- | --------
          | ------------ |\n| rhm-data-service-rhm-data-service-0 | 1Gi | RWO |\n|
          rhm-data-service-rhm-data-service-1 | 1Gi | RWO |\n| rhm-data-service-rhm-data-service-2
          | 1Gi | RWO |\n\n### Supported Storage Providers\n\n- OpenShift Container
          Storage / OpenShift Data Foundation version 4.x, from version 4.2 or higher\n-
          IBM Cloud Block storage and IBM Cloud File storage\n- IBM Storage Suite
          for IBM Cloud Paks:\n  - File storage from IBM Spectrum Fusion/Scale \n
          \ - Block storage from IBM Spectrum Virtualize, FlashSystem or DS8K\n- Portworx
          Storage, version 2.5.5 or above\n- Amazon Elastic File Storage\n\n### Access
          Modes required\n\n - ReadWriteOnce (RWO)\n\n### Provisioning Options supported\n\nChoose
          one of the following options to provision storage for the ibm-metrics-operator
          data-service\n\n#### Dynamic provisioning using a default StorageClass\n
          \  - A StorageClass is defined with a `metadata.annotations: storageclass.kubernetes.io/is-default-class:
          \"true\"`\n   - PersistentVolumes will be provisioned automatically for
          the generated PersistentVolumeClaims\n--- \n#### Manually create each PersistentVolumeClaim
          with a specific StorageClass\n   - Must be performed before creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`. Otherwise, the automatically
          generated PersistentVolumeClaims are immutable.\n```\napiVersion: v1\nkind:
          PersistentVolumeClaim\nmetadata:\n  labels:\n    app: rhm-data-service\n
          \ name: rhm-data-service-rhm-data-service-0\n  namespace: ibm-software-central\nspec:\n
          \ storageClassName: rook-cephfs\n  accessModes:\n  - ReadWriteOnce\n  resources:\n
          \   requests:\n      storage: 1Gi\n```\n---\n#### Manually provision each
          PersistentVolume for the generated PersistentVolumeClaims with a specific
          StorageClass\n  - May be performed before or after creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`.  \n```\napiVersion: v1\nkind:
          PersistentVolume\nmetadata:\n  name: rhm-data-service-rhm-data-service-0\nspec:\n
          \ csi:\n    driver: rook-ceph.cephfs.csi.ceph.com\n    volumeHandle: rhm-data-service-rhm-data-service-0\n
          \ capacity:\n    storage: 1Gi\n  accessModes:\n    - ReadWriteOnce\n  persistentVolumeReclaimPolicy:
          Delete\n  storageClassName: rook-cephfs\n  volumeMode: Filesystem\n  claimRef:\n
          \   kind: PersistentVolumeClaim\n    namespace: ibm-software-central\n    name:
          rhm-data-service-rhm-data-service-0\n```\n\n### Installation\n1. Create
          or get your [pull secret](https://swc.saas.ibm.com/en-us/documentation/clusters#get-pull-secret).\n2.
          Install the IBM Metrics Operator\n3. Create a Kubernetes secret in the installed
          namespace with the name `redhat-marketplace-pull-secret` and key `PULL_SECRET`
          with the value of the pull secret.\n    ```sh\n    # Replace ${PULL_SECRET}
          with your pull secret\n    oc create secret generic redhat-marketplace-pull-secret
          -n ibm-software-central --from-literal=PULL_SECRET=${PULL_SECRET}\n    ```\n4.
          Use of the platform is governed by the:\n\n    [IBM Cloud Services Agreement](https://www.ibm.com/support/customer/csol/terms/?id=Z126-6304_WS&_ga=2.116312197.2046730452.1684328846-812467790.1684328846)
          (or other base agreement between you and IBM such as a [Passport Advantage
          Agreement](https://www.ibm.com/software/passportadvantage/pa_agreements.html?_ga=2.116312197.2046730452.1684328846-812467790.1684328846))
          and the [Service Description for the Red Hat Marketplace](https://www.ibm.com/support/customer/csol/terms/?id=i126-8719&_ga=2.83289621.2046730452.1684328846-812467790.1684328846).\n
          \   \n5. Update MarketplaceConfig to accept the license.\n    ```\n    oc
          patch marketplaceconfig marketplaceconfig -n ibm-software-central --type='merge'
          -p '{\"spec\": {\"license\": {\"accept\": true}}}'\n    ```\n6. Install
          the pull secret as a global pull secret on the cluster.\n\n    These steps
          require `oc`, `jq`, and `base64` to be available on your machine.\n\n    ```sh\n
          \   # Create the docker pull secret file using your pull secret.\n    #
          Store it in a file called entitledregistryconfigjson.\n    oc create secret
          docker-registry entitled-registry --docker-server=registry.marketplace.redhat.com
          --docker-username \"cp\" --docker-password \"${PULL_SECRET}\" --dry-run=client
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > entitledregistryconfigjson\n
          \   # Get the current global secrets on the cluster and store it as a file
          named dockerconfigjson\n    oc get secret pull-secret -n openshift-config
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > dockerconfigjson\n
          \   # Merge the two dockerconfigs together into a file called dockerconfigjson-merged.\n
          \   jq -s '.[0] * .[1]' dockerconfigjson entitledregistryconfigjson > dockerconfigjson-merged\n
          \   # Set the cluster's dockerconfig file to the new merged version.\n    oc
          set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson=dockerconfigjson-merged\n
          \   ```\n  7. (Optional) Your IBM product may have installed IBM Metrics
          Operator using an `ibm-entitlement-key` for authentication to IBM Software
          Central. To authorize the automatic creation of an IBM Software Central
          account for the IBM account owning the `ibm-entitlement-key`\n      ```\n
          \     oc patch marketplaceconfig marketplaceconfig -n ibm-software-central
          --type='merge' -p '{\"spec\": {\"authorizeAccountCreation\": true}}'\n      ```\n\n###
          Why is a global pull secret required?\nIn order to successfully install
          the products hosted by the container image registry, you will need to make
          the pull secret available across the cluster. This can be achieved by applying
          the pull as a [global pull secret](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html#images-update-global-pull-secret_using-image-pull-secrets).
          For alternative approaches, please see the official OpenShift [documentation](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html).\n\n\n###
          SecurityContextConstraints requirements\n\nThe Operators and their components
          support running under the OpenShift Container Platform default restricted
          and restricted-v2 security context constraints.\n\n### Installation Namespace
          and ClusterRoleBinding requirements\n\nThe IBM Metrics Operator components
          require specific ClusterRoleBindings.\n- The metric-state component requires
          a ClusterRoleBinding for the the `view` ClusterRole.\n  - Ability to read
          non-sensitive CustomResources. \n  - The `view` ClusterRole dynamically
          adds new CustomResourceDefinitions.\n- The operator & reporter component
          requires a ClusterRoleBinding for the the `cluster-monitoring-view` ClusterRole.\n
          \ - Ability to view Prometheus metrics. \n  - The underlying ClusterRole
          RBAC often updated between OpenShift versions.\n\nDue to limitations of
          Operator Lifecycle Manager (OLM), this ClusterRoleBinding can not be provided
          dynamically for arbitrary installation target namespaces.\n\nA static ClusterRoleBinding
          is included for installation to the default namespace of `ibm-software-central`,
          and namespaces `openshift-redhat-marketplace`,  `redhat-marketplace`, `ibm-common-services`.\n\nTo
          create the ClusterRoleBindings for installation to an alternate namespace\n```\noc
          project INSTALL-NAMESPACE\noc adm policy add-cluster-role-to-user view -z
          ibm-metrics-operator-metric-state\noc adm policy add-cluster-role-to-user
          cluster-monitoring-view -z ibm-metrics-operator-controller-manager,ibm-metrics-operator-reporter\n```\n\n###
          Metric State scoping requirements\nThe metric-state Deployment obtains `get/list/watch`
          access to metered resources via the `view` ClusterRole. For operators deployed
          using Operator Lifecycle Manager (OLM), permissions are added to `clusterrole/view`
          dynamically via a generated and annotated `-view` ClusterRole. If you wish
          to meter an operator, and its Custom Resource Definitions (CRDs) are not
          deployed through OLM, one of two options are required\n1. Add the following
          label to a clusterrole that has get/list/watch access to your CRD: `rbac.authorization.k8s.io/aggregate-to-view:
          \"true\"`, thereby dynamically adding it to `clusterrole/view`\n2. Create
          a ClusterRole that has get/list/watch access to your CRD, and create a ClusterRoleBinding
          for the metric-state ServiceAccount\n\nAttempting to meter a resource with
          a MeterDefinition without the required permissions will log an `AccessDeniedError`
          in metric-state.\n\n### Disaster Recovery\n\nTo plan for disaster recovery,
          note the PhysicalVolumeClaims `rhm-data-service-rhm-data-service-N`. \n-
          In connected environments, MeterReport data upload attempts occur hourly,
          and are then removed from data-service. There is a low risk of losing much
          unreported data.\n- In an airgap environment, MeterReport data must be pulled
          from data-service and uploaded manually using [datactl](https://github.com/redhat-marketplace/datactl/).
          To prevent data loss in a disaster scenario, the data-service volumes should
          be considered in a recovery plan.\n\n### Subscription Config\n\nIt is possible
          to [configure](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md)
          how OLM deploys an Operator via the `config` field in the [Subscription](https://github.com/operator-framework/olm-book/blob/master/docs/subscriptions.md)
          object.\n\nThe IBM Metrics Operator will also read the `config` and append
          it to the operands. The primary use case is to control scheduling using
          [Tolerations](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/)
          and [NodeSelectors](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/).\n\nA
          limitation is that the `config` elements are only appended to the operands.
          The elements in the operands are not removed if the `config` is removed
          from the `Subscripion`. The operand must be modified manually, or deleted
          and recreated by the controller.\n\n### Cluster permission requirements\n\n|API
          group             |Resources              |Verbs                                     |\n|----------------------|-----------------------|------------------------------------------|\n|apps
          \                 |statefulsets           |get;list;watch                            |\n|authentication.k8s.io
          |tokenreviews           |create                                    |\n|authorization.k8s.io
          \ |subjectaccessreviews   |create                                    |\n|config.openshift.io
          \  |clusterversions        |get;list;watch                            |\n|marketplace.redhat.com|meterdefinitions
          \      |get;list;watch;create;update;patch;delete |\n|operators.coreos.com
          \ |clusterserviceversions |get;list;watch                            |\n|operators.coreos.com
          \ |subscriptions          |get;list;watch                            |\n|operators.coreos.com
          \ |operatorgroups         |get;list                                  |\n|storage.k8s.io
          \       |storageclasses         |get;list;watch                            |\n|
          \                     |namespaces             |get;list;watch                            |\n|
          \                     |configmaps             |get;list;watch                            |\n|
          \                     |services               |get;list;watch                            |\n|
          \                     |**clusterrole/view**   |                                          |\n\n\n###
          Documentation\nYou can find our documentation [here.](https://swc.saas.ibm.com/en-us/documentation/)\n\n###
          Getting help\nIf you encounter any issues while using the operators, you
          can create an issue on our [Github\nrepo](https://github.com/redhat-marketplace/redhat-marketplace-operator)
          for bugs, enhancements, or other requests. You can also visit our main page
          and\nreview our [support](https://swc.saas.ibm.com/en-us/support) and [documentation](https://swc.saas.ibm.com/en-us/documentation/).\n\n###
          Readme\nYou can find our readme [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/README.md)\n\n###
          License information\nYou can find our license information [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/LICENSE)\n"
        displayName: IBM Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        - buy
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/redhat-marketplace-metric-state@sha256:e4014c9a546c10e84b13a09db6a18431ee9489ad1296808635e48d9ef4963cf1
        - quay.io/rh-marketplace/redhat-marketplace-authcheck@sha256:13a81b8aca79ae8c89ee24dcc30f3322a4cfdedf929e7bb984295c1e7e631fa9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/redhat-marketplace-data-service@sha256:30fd8d2996b921cfce51f697ee4490a6d9a836021952429b877336f5edf2749d
        - quay.io/rh-marketplace/redhat-marketplace-reporter@sha256:2bd098f057ca07f2b14be9b60ea6ad8ca6b0cae1d226b1ca3b27e23d04ccdf20
        - quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
        version: 2.20.0
      entries:
      - name: ibm-metrics-operator.v2.20.0
        version: 2.20.0
      name: stable
    defaultChannel: stable
    packageName: ibm-metrics-operator
    provider:
      name: Red Hat Marketplace
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-dr-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-dr-manager.v8.0.220
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com",
                  "resRecord": {
                    "HostName_1": "100.200.30.41",
                    "HostName_2": "100.200.30.43",
                    "abc": "HostName_2",
                    "www": "HostName_1",
                    "xyz": "HostName_3"
                  },
                  "stealthMasters": [
                    "1.2.3.4"
                  ],
                  "tsigKeyFile": "/Kvcslnxdns.com.+157+10641.private"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "dataReplicationPointer": "test-datareplication",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 10,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          olm.skipRange: <8.0.220
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
            - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:a98f9cc51f2c7baeec6dbb4933a42ae8ae20ff1a42c9d48a37174cf5211bbcf8
        version: 8.0.220
      entries:
      - name: infoscale-dr-manager.v8.0.220
        version: 8.0.220
      - name: infoscale-dr-manager.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-dr-manager.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          createdAt: "2024-06-17T10:14:13Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:c46cbf78631fc790c77b6c13c1d1d4b5ab35f49164f7c2e36c7ec59220cbdce2
        version: 8.0.330
      entries:
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-dr-manager.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
          createdAt: "2024-02-22T14:14:42Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721648-168743109)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721654-168743109)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
        version: 8.0.410
      entries:
      - name: infoscale-dr-manager.v8.0.410
        version: 8.0.410
      - name: infoscale-dr-manager.v8.0.409
        version: 8.0.409
      - name: infoscale-dr-manager.v8.0.400
        version: 8.0.400
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-dr-manager.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
          createdAt: "2024-02-22T14:14:42Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721648-168743109)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721654-168743109)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
        version: 8.0.410
      entries:
      - name: infoscale-dr-manager.v8.0.410
        version: 8.0.410
      - name: infoscale-dr-manager.v8.0.409
        version: 8.0.409
      - name: infoscale-dr-manager.v8.0.400
        version: 8.0.400
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-dr-manager.v8.0.100
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com",
                  "resRecord": {
                    "abc": "r7515-054-vm9",
                    "r7515-054-vm10": "10.221.85.83",
                    "r7515-054-vm8": "10.221.85.81",
                    "r7515-054-vm9": "10.221.85.82",
                    "www": "r7515-054-vm8",
                    "xyz": "r7515-054-vm10"
                  },
                  "stealthMasters": [
                    "1.2.3.4"
                  ],
                  "tsigKeyFile": "/Kvcslnxdns.com.+157+10641.private"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "currentPrimary": "Clus1",
                  "force": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "fail",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "dataReplicationPointer": "test-datareplication",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 10,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:eb999187a1f9a9ce5566089e24adf598b1c229d31c5e57ddcfe4439ba25e97dc
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
            - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membershrip
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membershrip
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponsding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:eb999187a1f9a9ce5566089e24adf598b1c229d31c5e57ddcfe4439ba25e97dc
        version: 8.0.100
      entries:
      - name: infoscale-dr-manager.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-dr-manager
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NdmSpc project
      provider-url: ""
    name: ndmspc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ndmspc-operator.v0.20250209.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.ndmspc.io/v1alpha1",
                "kind": "NdmSpcConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ndmspc-operator",
                    "app.kubernetes.io/instance": "ndmspcconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ndmspcconfig",
                    "app.kubernetes.io/part-of": "ndmspc-operator"
                  },
                  "name": "ndmspc-sample"
                },
                "spec": {
                  "executor": {
                    "image": "registry.gitlab.com/ndmspc/api:v0.20240215.2",
                    "name": "executor"
                  },
                  "executors": 1,
                  "ingressclasstype": "nginx",
                  "ingressenabled": true,
                  "ingresshostdomain": "127.0.0.1.sslip.io",
                  "ingressnostname": "ndmspc",
                  "ingressssltraefik": false,
                  "naf": {
                    "image": "registry.gitlab.com/ndmspc/naf:v0.20250129.0",
                    "name": "naf"
                  },
                  "ndmvr": {
                    "image": "registry.gitlab.com/ndmspc/ndmvr/shared:v0.20230320.0",
                    "name": "ndmvr"
                  },
                  "salsa": {
                    "image": "registry.gitlab.com/ndmspc/user/ndmspc:v0.20240412.0"
                  },
                  "storage": {
                    "image": "registry.gitlab.com/ndmspc/storage-server:v0.20250205.0",
                    "mongodb": {
                      "image": "docker.io/library/mongo:latest",
                      "password": "admin",
                      "port": 27017,
                      "username": "admin"
                    },
                    "name": "storage",
                    "port": 8080
                  },
                  "webapp": {
                    "image": "registry.gitlab.com/ndmspc/react-ndmspc:v0.20240412.0",
                    "name": ""
                  },
                  "webappcontainerport": 8080,
                  "webapps": 1,
                  "worker": {
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "1Gi"
                      }
                    }
                  },
                  "workers": 2,
                  "workerslots": 1,
                  "zmq2ws": {
                    "image": "registry.gitlab.com/ndmspc/zmq2ws:v1.20231012.0",
                    "name": "zmq2ws"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: registry.gitlab.com/ndmspc/ndmspc-operator:0.20250209.0
          createdAt: "2025-02-09T12:49:06Z"
          description: The NdmSpc operator automates the deployment and management
            of Ndmspc suite on Kubernetes cluster.
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://gitlab.com/ndmspc/ndmspc-operator
          support: NDMSPC Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NdmSpc configuration
            displayName: NdmSpc Config
            kind: NdmSpcConfig
            name: ndmspcconfigs.apps.ndmspc.io
            version: v1alpha1
        description: "The NdmSpc operator automates the deployment and management
          of Ndmspc suite on Kubernetes cluster. This operator allows user to deploy
          SALSA cluster with ZMQ2WS and job executor service and specify custom webapp
          page.\n\n### Capabilities\n\n- SALSA cluster\n- ZMQ2WS api service\n- Job
          executor service \n- Custom WebApp page\n"
        displayName: NdmSpc operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ndmspc
        links:
        - name: Ndmspc Operator
          url: https://ndmspc.gitlab.io/docs
        maintainers:
        - email: ndmspc@gmail.com
          name: NdmSpc Support
        maturity: alpha
        provider:
          name: NdmSpc project
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - registry.gitlab.com/ndmspc/ndmspc-operator:0.20250209.0
        version: 0.20250209.0
      entries:
      - name: ndmspc-operator.v0.20250209.0
        version: 0.20250209.0
      - name: ndmspc-operator.v0.20250131.0
        version: 0.20250131.0
      - name: ndmspc-operator.v0.20240412.0
        version: 0.20240412.0
      - name: ndmspc-operator.v0.11.6
        version: 0.11.6
      - name: ndmspc-operator.v0.11.4
        version: 0.11.4
      - name: ndmspc-operator.v0.11.3
        version: 0.11.3
      - name: ndmspc-operator.v0.11.2
        version: 0.11.2
      - name: ndmspc-operator.v0.11.1
        version: 0.11.1
      - name: ndmspc-operator.v0.10.2
        version: 0.10.2
      - name: ndmspc-operator.v0.10.1
        version: 0.10.1
      - name: ndmspc-operator.v0.10.0
        version: 0.10.0
      - name: ndmspc-operator.v0.9.0
        version: 0.9.0
      - name: ndmspc-operator.v0.7.4
        version: 0.7.4
      - name: ndmspc-operator.v0.7.1
        version: 0.7.1
      - name: ndmspc-operator.v0.6.0
        version: 0.6.0
      - name: ndmspc-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: ndmspc-operator
    provider:
      name: NdmSpc project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: sosivio
      provider-url: https://www.sosiv.io
    name: sosivio
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sosivio-operator.v1.5.1-97
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sosivio.sosiv.io/v1alpha1",
                "kind": "Sosivio",
                "metadata": {
                  "name": "sosivio"
                },
                "spec": {
                  "platform": "ocp4",
                  "useSha": false
                }
              },
              {
                "apiVersion": "sosivio.sosiv.io/v1alpha1",
                "kind": "Sosivio",
                "metadata": {
                  "name": "sosivio"
                },
                "spec": {
                  "platform": "ocp4",
                  "useSha": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Monitoring,Integration & Delivery,Logging & Tracing,Developer
            Tools
          com.redhat.openshift.versions: v4.1-v4.1
          containerImage: quay.io/sosivio/sosivio-operator:v1.5.1-97
          createdAt: "2021-10-28 01:01:01"
          description: Sosivio | Predictive Troubleshooting for Kubernetes
          operatorhub.io/ui-metadata-max-k8s-version: "1.25"
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Sosivio/operator
          support: Sosivio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Sosivio is the Schema for the sosivios API
            displayName: Sosivio
            kind: Sosivio
            name: sosivios.sosivio.sosiv.io
            version: v1alpha1
        description: "# Sosivio Operator\nThe Sosivio Operator for RedHat’s Openshift
          Container Platform allows users to quickly and easily deploy Sosivio’s Kubernetes
          Predictive Troubleshooting Platform on your cluster.\n\n# Sosivio Overview\n\nSosivio
          is the first and only non-intrusive and predictive autonomous cloud-native
          platform built specifically for Kubernetes from the ground up. Sosivio provides
          predictive troubleshooting tools, root cause determination and instant remediation
          for any failures in Kubernetes environments.\n\nPredict and Prevent!\n\nThe
          Operator will deploy and manage Sosivio on Kubernetes and OpenShift. Included
          are the following features;\n-   **Root Cause Analysis and Remediation**
          - predictive troubleshooting to prevent failures before they occur. Instantly
          understand where the root cause is and how to fix it.\n-   **Application
          Profiling** - analyze container resource consumption and recommendations
          for requests and limits.\n-   **Real-time Metrics** - in-depth, granular,
          and accurate real-time data from the process level inside containers (and
          pods) with failure correlation.\n-   **Cluster Health Checks** - proactive
          health checks for cluster infrastructure and applications running on the
          cluster.\n-   **Multi-Cluster View** - management, monitoring, and maintenance
          capabilities for all clusters from one location. \n-   **Kubernetes Components
          Alerting** - notifications sent through Slack, email and other 3rd party
          platforms.\n\n# Installation\nSosivio is completely **non-intrusive** and
          **does not require any persistent storage**. There is **no data offloading**
          outside of your Kubernetes cluster. Applying our Data Swirling methodology
          enables the product to perform at a **remarkably low resource footprint**
          and requires **no persistent volume configurations**.\n\n\nSosivio is installed
          under one namespace (labeled: \"sosivio\") which creates all the required
          components for the product.\n\nAll users will start with 4 weeks of Sosivio
          Premium after which Sosivio will automatically convert to the Community
          Version (free forever).\n\n# Help\nMore information about Sosivio and our
          product usage instructions can be found in our documentation [here](https://docs.sosiv.io/)
          \ \n"
        displayName: Sosivio | Predictive Troubleshooting for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sosivio
        - kubernetes
        - observanility
        - monitoring
        - root cause
        links:
        - name: Sosivio
          url: https://www.sosiv.io
        - name: Documentation
          url: https://docs.sosiv.io
        maintainers:
        - email: info@sosiv.io
          name: Sosivio
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: sosivio
          url: https://www.sosiv.io
        relatedImages:
        - quay.io/sosivio/discovery-engine@sha256:b87b3f7c9ed3aed80d0a77788fd8925789450d9dc05d25017e89dff611cd82da
        - quay.io/sosivio/nsq@sha256:781aac23ef0a1ed0af2b8e93d625ae49bd11dbdd98787d3c79dde8d128f3810d
        - quay.io/sosivio/actuator@sha256:a9abe9730800ca2b0a1bbd7a361b2ab3a818cbf127798314e5477fb1cb16ba4c
        - quay.io/sosivio/sosivio-node-collector@sha256:61ed8dd52de5edc09d16a31b18f2f65ad9344c0b6c9cfab323c3eb3d7e7ed060
        - quay.io/sosivio/crud-manager@sha256:38289a28829fa737917d76dd4810ba6140e123061ead0a4b79fa79b2e246a293
        - quay.io/sosivio/sosivio-operator:v1.5.1-97
        - quay.io/sosivio/sosivio-dashboard@sha256:fdfca5bb2123d1008cec2b656cde73c7e9364f84c5ff4f4a6fee5751b1bfb2f8
        - quay.io/sosivio/sosivio-node-netpacket@sha256:84dd159b1dfdb52dee3eb1c4c93f323bae49ce517ff44663b69978e14bdd8390
        - quay.io/sosivio/classifier@sha256:dd8bf145524db23ff333efa1bcdff258dab116d3534f95f73775d8dd339e6963
        - quay.io/sosivio/communicator@sha256:fffec9a6c66594fde5f05e2dfb84361544ca9d53a345188c197a0ea6f53bc4b9
        - quay.io/sosivio/authentication@sha256:0ea6025f333576792f5b1fc4d96d9ab1a404749f185830410e71248683004da2
        - quay.io/sosivio/sosiviodb@sha256:a3c6c3975bc2b0316135ee5c547f1eb215f3a1f9c8b0a9c8eceff23415f93597
        - quay.io/sosivio/kube-rbac-proxy:v0.8.0
        - quay.io/sosivio/sequence-recognition@sha256:7335cf8f3e40bc46e8ca0de525d99f0655833b7c52d36dc273aa7b096eb70fa2
        - quay.io/sosivio/network-aggregator@sha256:e59d937f4d9c52acad0adadab39c6849d20b4ee72a83e239a33ea137f08323f4
        - quay.io/sosivio/correlation-entities@sha256:c00af09b2589bad4740fd44d7a81c69b2e29138c360939b2a45f65edd44b6901
        - quay.io/sosivio/contract-testing@sha256:cad6fac6ed3fdf2597ceb09ba5dbb7d81f70cf45f483a7064e3968215dd982e3
        - quay.io/sosivio/kube-rbac-proxy@sha256:34e8724e0f47e31eb2ec3279ac398b657db5f60f167426ee73138e2e84af6486
        - quay.io/sosivio/analyzer@sha256:3825608fe7c43ef059108e76665acc99fe72bea77d4d9e3852f6ad45e74379b1
        - quay.io/sosivio/notifier@sha256:35f5f3e8e5ccad914fcfe34fb4b4a1b044ee87f5a6141284cd5273f0b2b3bf22
        - quay.io/sosivio/sosivio-node-pinger@sha256:937de52ad7b76e6cff4ad5197de9705988d61f285ebadac5ebadb5856f59ee58
        - quay.io/sosivio/sosivio-operator@sha256:11bc627773bb5e08b629e0b013e7f600fcb4aaefbd97f2de337d339475616763
        version: 1.5.1
      entries:
      - name: sosivio-operator.v1.5.1-97
        version: 1.5.1
      - name: sosivio-operator.v1.5.0-95
        version: 1.5.0
      - name: sosivio-operator.v1.4.4-85
        version: 1.4.4
      - name: sosivio-operator.v1.4.3-83
        version: 1.4.3
      - name: sosivio-operator.v1.4.2-82
        version: 1.4.2
      - name: sosivio-operator.v1.4.1-81
        version: 1.4.1-81
      - name: sosivio-operator.v1.4.1.1-80
        version: 1.4.1-80
      - name: sosivio-operator.v1.4.1-78
        version: 1.4.1-79
      - name: sosivio-operator.v1.3.1-76
        version: 1.3.1-76
      - name: sosivio-operator.v1.3.0-75
        version: 1.3.0-75
      - name: sosivio-operator.v1.2.1-1
        version: 1.2.1-1
      - name: sosivio-operator.v1.2.0-1
        version: 1.2.0-1
      name: stable
    defaultChannel: stable
    packageName: sosivio
    provider:
      name: sosivio
      url: https://www.sosiv.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-keycloak-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-keycloak-operator.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2023-09-19T12:43:29Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KCOSClientRoleMapping
            name: kcosclientrolemappings.keycloak.org
            version: v1alpha1
          - kind: KCOSRealmRole
            name: kcosrealmroles.keycloak.org
            version: v1alpha1
          - kind: KCOSUserRoleMapping
            name: kcosuserrolemappings.keycloak.org
            version: v1alpha1
          - kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: Keysight Keycloak Operator
        displayName: Keysight Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Keycloak Operator
          url: https://keysight-keycloak-operator.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak@sha256:772f73cce6a77b9c46f8035faac99a28602219657d18b75b19b4d905d91942c9
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak-postgresql@sha256:b636955199065e18f9d4679835ae78214843258e5abb4ae758288d6f00df2f8e
        - prod.harbor.keysight.digital/loadcore/keysight-keycloak-operator@sha256:d74e7e98a114768a6f3327edabce9336743a078844fffee985053b6a90100da8
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak-init-container@sha256:90965beaf41c760cb8c8bf906acf08f8fe4770d313de8dfd92d18d7c25348a38
        version: 0.1.21
      entries:
      - name: keysight-keycloak-operator.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-keycloak-operator
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: fence-agents-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fence-agents-remediation.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  }
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:00cc420250d8e25d9da7c43d9c4cff35c7edeb7d390af754c48c0ce00e9e7770
          createdAt: "2023-10-30 14:06:57"
          description: Fence Agents Remediation Operator for remediating nodes using
            upstream fence-agents.
          olm.skipRange: '>=0.2.0 <0.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: "Fence Agents Remediation Operator (FAR)\n\nThe operator will
          remediate/fence a node when a FenceAgentsRemediation (far) custom resource
          exists with the node's name.\nFAR uses an existing set of upstream fencing
          agents for environments with a \ntraditional API end-point (e.g., IPMI)
          for power cycling cluster nodes.\n\nThe operator is recommended to be used
          with Node Healthcheck Operator for easier and smoother experience,\nbut
          it can also work as a standalonde remediator for the more advanced user.\n"
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:00cc420250d8e25d9da7c43d9c4cff35c7edeb7d390af754c48c0ce00e9e7770
        version: 0.2.1
      entries:
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.14-eus
    - currentCSV: fence-agents-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:a367e8fdd80938e97ac195631a996e0c537bf3666aef0199147534762dcf739f
          createdAt: "2024-08-10 22:49:23"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.2.0 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:a367e8fdd80938e97ac195631a996e0c537bf3666aef0199147534762dcf739f
        version: 0.4.1
      entries:
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.16-eus
    - currentCSV: fence-agents-remediation.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-16 10:25:57"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.5.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.5/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel9-operator@sha256:da5975557078fbfa6882df87cff5bda5671b9838b5fb0a738fea724469d03576
        version: 0.5.1
      entries:
      - name: fence-agents-remediation.v0.5.1
        version: 0.5.1
      - name: fence-agents-remediation.v0.5.0
        version: 0.5.0
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.18-eus
    - currentCSV: fence-agents-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  }
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:5c3778c7aae45f54e949fbae3d133777f7aa78fe1251ac3d5510aa28b71e2ba4
          createdAt: "2023-04-20 15:36:05"
          description: Fence Agents Remediation Operator for remediating nodes using
            upstream fence-agents.
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: "Fence Agents Remediation Operator (FAR)\n\nThe operator will
          remediate/fence a node when a FenceAgentsRemediation (far) custom resource
          exists with the node's name.\nFAR uses an existing set of upstream fencing
          agents for environments with a \ntraditional API end-point (e.g., IPMI)
          for power cycling cluster nodes.\n\nThe operator is recommended to be used
          with Node Healthcheck Operator for easier and smoother experience,\nbut
          it can also work as a standalonde remediator for the more advanced user.\n"
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3af07bfe0fb73b1a7f00487eba302d6da442e742ab4007fc50e52efd4d45dae5
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:5c3778c7aae45f54e949fbae3d133777f7aa78fe1251ac3d5510aa28b71e2ba4
        version: 0.1.0
      entries:
      - name: fence-agents-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: fence-agents-remediation.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeSecretNames": {
                    "master-0": "fence-agents-credentials-master0",
                    "master-1": "fence-agents-credentials-master1",
                    "master-2": "fence-agents-credentials-master2",
                    "worker-0": "fence-agents-credentials-worker0",
                    "worker-1": "fence-agents-credentials-worker1",
                    "worker-2": "fence-agents-credentials-worker2"
                  },
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedSecretName": "fence-agents-credentials-shared",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {
                    "spec": {
                      "agent": "fence_ipmilan",
                      "nodeSecretNames": {
                        "master-0": "fence-agents-credentials-master0",
                        "master-1": "fence-agents-credentials-master1",
                        "master-2": "fence-agents-credentials-master2",
                        "worker-0": "fence-agents-credentials-worker0",
                        "worker-1": "fence-agents-credentials-worker1",
                        "worker-2": "fence-agents-credentials-worker2"
                      },
                      "nodeparameters": {
                        "--ipport": {
                          "master-0": "6230",
                          "master-1": "6231",
                          "master-2": "6232",
                          "worker-0": "6233",
                          "worker-1": "6234",
                          "worker-2": "6235"
                        }
                      },
                      "remediationStrategy": "ResourceDeletion",
                      "retrycount": 5,
                      "retryinterval": "5s",
                      "sharedSecretName": "fence-agents-credentials-shared",
                      "sharedparameters": {
                        "--action": "reboot",
                        "--ip": "192.168.111.1",
                        "--lanplus": "",
                        "--username": "admin"
                      },
                      "timeout": "60s"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-09-08 09:19:00"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.5.0 <0.6.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.8/html/remediation_fencing_and_maintenance/about-remediation-fencing-maintenance
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel9-operator@sha256:eef1a298718650a22a3da4a07c140ce21b6f287d87af79801610317fa8f63d52
        version: 0.6.0
      entries:
      - name: fence-agents-remediation.v0.6.0
        version: 0.6.0
      - name: fence-agents-remediation.v0.5.1
        version: 0.5.1
      - name: fence-agents-remediation.v0.5.0
        version: 0.5.0
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: stable
    defaultChannel: stable
    packageName: fence-agents-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-s3control-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-s3control-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/s3control-controller:1.1.1
          createdAt: "2025-09-19T20:52:02Z"
          description: AWS S3 Control controller is a service controller for managing
            S3 Control resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessPoint represents the state of an AWS s3control AccessPoint
              resource.
            displayName: AccessPoint
            kind: AccessPoint
            name: accesspoints.s3control.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon S3 Control resources in AWS from within your Kubernetes cluster.

          **About Amazon S3 Control**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon S3 Control
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - s3control
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon S3 Control Developer Resources
          url: https://aws.amazon.com/S3 Control/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: s3control maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/s3control-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-s3control-controller.v1.1.1
        version: 1.1.1
      - name: ack-s3control-controller.v1.1.0
        version: 1.1.0
      - name: ack-s3control-controller.v1.0.11
        version: 1.0.11
      - name: ack-s3control-controller.v1.0.10
        version: 1.0.10
      name: alpha
    defaultChannel: alpha
    packageName: ack-s3control-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-efs-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-efs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "AccessPoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "FileSystem",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "MountTarget",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/efs-controller:1.1.1
          createdAt: "2025-09-19T18:38:13Z"
          description: AWS EFS controller is a service controller for managing EFS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessPoint represents the state of an AWS efs AccessPoint
              resource.
            displayName: AccessPoint
            kind: AccessPoint
            name: accesspoints.efs.services.k8s.aws
            version: v1alpha1
          - description: FileSystem represents the state of an AWS efs FileSystem
              resource.
            displayName: FileSystem
            kind: FileSystem
            name: filesystems.efs.services.k8s.aws
            version: v1alpha1
          - description: MountTarget represents the state of an AWS efs MountTarget
              resource.
            displayName: MountTarget
            kind: MountTarget
            name: mounttargets.efs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EFS resources in AWS from within your Kubernetes cluster.

          **About Amazon Elastic File System (Amazon EFS)**

          Amazon Elastic File System (Amazon EFS) provides serverless, fully elastic file storage so that you can share file data without provisioning or managing storage capacity and performance. Amazon EFS is built to scale on demand to petabytes without disrupting applications, growing and shrinking automatically as you add and remove files. Because Amazon EFS has a simple web services interface, you can create and configure file systems quickly and easily. The service manages all the file storage infrastructure for you, meaning that you can avoid the complexity of deploying, patching, and maintaining complex file system configurations.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Elastic File System (Amazon
          EFS)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - efs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EFS Developer Resources
          url: https://aws.amazon.com/efs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: efs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/efs-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-efs-controller.v1.1.1
        version: 1.1.1
      - name: ack-efs-controller.v1.1.0
        version: 1.1.0
      - name: ack-efs-controller.v1.0.19
        version: 1.0.19
      - name: ack-efs-controller.v1.0.18
        version: 1.0.18
      - name: ack-efs-controller.v1.0.17
        version: 1.0.17
      - name: ack-efs-controller.v1.0.16
        version: 1.0.16
      - name: ack-efs-controller.v1.0.15
        version: 1.0.15
      - name: ack-efs-controller.v1.0.14
        version: 1.0.14
      - name: ack-efs-controller.v1.0.13
        version: 1.0.13
      - name: ack-efs-controller.v1.0.12
        version: 1.0.12
      - name: ack-efs-controller.v1.0.11
        version: 1.0.11
      - name: ack-efs-controller.v1.0.10
        version: 1.0.10
      - name: ack-efs-controller.v1.0.8
        version: 1.0.8
      - name: ack-efs-controller.v1.0.6
        version: 1.0.6
      - name: ack-efs-controller.v1.0.5
        version: 1.0.5
      - name: ack-efs-controller.v1.0.3
        version: 1.0.3
      - name: ack-efs-controller.v1.0.2
        version: 1.0.2
      - name: ack-efs-controller.v1.0.0
        version: 1.0.0
      - name: ack-efs-controller.v0.0.9
        version: 0.0.9
      - name: ack-efs-controller.v0.0.8
        version: 0.0.8
      - name: ack-efs-controller.v0.0.7
        version: 0.0.7
      - name: ack-efs-controller.v0.0.6
        version: 0.0.6
      - name: ack-efs-controller.v0.0.5
        version: 0.0.5
      - name: ack-efs-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-efs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solace Corporation
      provider-url: www.solace.com
    name: pubsubplus-eventbroker-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pubsubplus-eventbroker-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pubsubplus.solace.com/v1beta1",
                "kind": "PubSubPlusEventBroker",
                "metadata": {
                  "name": "non-ha-standalone-example"
                },
                "spec": {
                  "redundancy": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "true"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.10
          containerImage: registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator:1.4.0
          createdAt: "2025-07-31T19:48:36Z"
          description: The Solace PubSub+ Event Broker Operator deploys and manages
            the lifecycle of PubSub+ Event Brokers
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
          support: Solace Products
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PubSub+ Event Broker
            displayName: PubSub+ Event Broker
            kind: PubSubPlusEventBroker
            name: pubsubpluseventbrokers.pubsubplus.solace.com
            version: v1beta1
        description: |
          ## Solace PubSub+ Platform
          Solace [PubSub+ Platform](https://solace.com/products/platform/) is a complete event streaming and management platform
          for real-time enterprises. The [PubSub+ Software Event Broker](https://solace.com/products/event-broker/software/) efficiently streams
          event-driven information between applications, IoT devices, and user interfaces running in the cloud, on-premises,
          and in hybrid environments using open APIs and protocols like AMQP, JMS, MQTT, REST and WebSocket.
          It can be installed into a variety of public and private clouds, PaaS, and on-premises environments.
          Event brokers in multiple locations can be linked together in an [Event Mesh](https://solace.com/what-is-an-event-mesh/)
          to dynamically share events across the distributed enterprise.

          ## Solace PubSub+ Event Broker Operator
          Solace Pubsub+ Event Broker Operator automatically deploys and manages PubSub+ Software Event Brokers on Kubernetes and OpenShift environments.

          Features include:
          * Initial deployment configuration: Production-ready HA, non-HA or for developers
          * Rolling update for configuration changes or broker upgrades
          * Prometheus monitoring support

          ## Getting started
          Refer to the [Solace PubSub+ Event Broker Operator Quick Start](https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/README.md)
        displayName: Solace PubSub+ Event Broker Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - solace
        - pubsubplus
        - pubsub+
        - pubsub
        - messaging
        - advanced event broker
        - event broker
        - event mesh
        - event streaming
        - data streaming
        - event integration
        - middleware
        links:
        - name: Operator QuickStart
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
        - name: Operator Documentation
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/docs/EventBrokerOperatorUserGuide.md
        - name: Solace.com
          url: https://solace.com
        - name: Understanding Solace Technology
          url: https://solace.com/products/platform/
        - name: Solace PubSub+ Technical Documentation
          url: https://docs.solace.com/
        - name: Solace Developer Portal
          url: https://www.solace.dev/
        - name: Solace Community
          url: https://solace.community/
        maintainers:
        - email: support@solace.com
          name: Solace R&D
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Solace Corporation
          url: www.solace.com
        relatedImages:
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator:1.4.0
        version: 1.4.0
      entries:
      - name: pubsubplus-eventbroker-operator.v1.4.0
        version: 1.4.0
      - name: pubsubplus-eventbroker-operator.v1.3.0
        version: 1.3.0
      - name: pubsubplus-eventbroker-operator.v1.2.0
        version: 1.2.0
      - name: pubsubplus-eventbroker-operator.v1.1.2
        version: 1.1.2
      - name: pubsubplus-eventbroker-operator.v1.1.1
        version: 1.1.1
      - name: pubsubplus-eventbroker-operator.v1.1.0
        version: 1.1.0
      - name: pubsubplus-eventbroker-operator.v1.0.3
        version: 1.0.3
      - name: pubsubplus-eventbroker-operator.v1.0.2
        version: 1.0.2
      - name: pubsubplus-eventbroker-operator.v1.0.1
        version: 1.0.1
      - name: pubsubplus-eventbroker-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pubsubplus-eventbroker-operator
    provider:
      name: Solace Corporation
      url: www.solace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: ibm-data-reporter-operator
      app.kubernetes.io/managed-by: ibm-data-reporter-operator
      app.kubernetes.io/name: ibm-data-reporter-operator
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: ""
    name: ibm-data-reporter-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-data-reporter-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "DataReporterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-data-reporter-operator",
                    "app.kubernetes.io/instance": "datareporterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datareporterconfig",
                    "app.kubernetes.io/part-of": "ibm-data-reporter-operator"
                  },
                  "name": "datareporterconfig"
                },
                "spec": {
                  "userConfig": [
                    {
                      "metadata": {
                        "ameta1": "ametadata1",
                        "bmeta1": "bmetadata1",
                        "cmeta1": "cmetadata1",
                        "dmeta1": "dmetadata1"
                      },
                      "userName": "system:serviceaccount:redhat-marketplace:ibm-data-reporter-operator-api"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
          createdAt: ""
          description: The IBM Data Reporter Operator accepts events and transforms
            them into reports submitted to the Data Service of the IBM Metrics Operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReporterConfig is the Schema for the datareporterconfigs
              API
            displayName: Data Reporter Config
            kind: DataReporterConfig
            name: datareporterconfigs.marketplace.redhat.com
            version: v1alpha1
          required:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
        description: "# IBM Data Reporter Operator\n\n# Introduction\n\nThe IBM Data
          Reporter Operator accepts events and transforms them into reports submitted
          to the IBM Metrics Operator Data Service.\n\n# Details\n\nThe IBM Data Reporter
          Operator deploys a service that provides an endpoint to which callers can
          send raw json event data. The event data is transformed into a report and
          is sent to the IBM Metrics Operator's Data Service. The IBM Metrics Operator
          Data Service periodically uploads the reports to IBM Software Central. It
          can also filter, transform and forward events to alternate endpoints.\n\n##
          Prerequisites\n\n- OpenShift Container Platform, major version 4 with any
          available supported minor version\n- Install IBM Metrics Operator\n  - IBM
          Data Reporter Operator prerequisties the IBM Metrics Operator's Data Service\n
          \ - Register the Cluster by creating a `redhat-marketplace-pull-secret`,
          as per the instructions\n  - Once `rhm-data-service` has started IBM Data
          Reporter Operator will accept requests\n\n## SecurityContextConstraints
          Requirements\n\n- The operator runs under Red Hat restricted SCC\n\n## Resources
          Required\n\n- The operator requires at least 85 Mi memory.\n\n## Limitations\n\n-
          Only runs on amd64, s390x and ppc64le architectures\n\n## Installing\n\n-
          A user with the cluster administrator role.\n- Install this operator in
          the same namespace as the IBM Metrics Operator\n  - default namespace: `ibm-software-central`\n\n##
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n# Configuration\n\nOptional:\n- Configure the
          Custom Resource Definition `DataReporterConfig` named `datareporterconfig`
          as per the following sample\n  - Configuration that can not be reconciled
          successfully will be reported on the DataReporterConfig Status\n  - jsonPath
          expressions are handled by: https://github.com/ohler55/ojg\n    - jsonPath
          comparison: https://cburgmer.github.io/json-path-comparison/\n  - Transformer
          type currently supported is `kazaam`\n    - https://github.com/willie68/kazaam\n\nSample
          DataReporterConfig:\n```YAML\napiVersion: marketplace.redhat.com/v1alpha1\nkind:
          DataReporterConfig\nmetadata:\n  name: datareporterconfig\nspec:\n  # true:
          report generated for 1 event and provide return code from data-service,
          confirming delivery\n  # false: events will be accumulated before building
          report, and immediately return 200, delivery unfconfirmed\n  confirmDelivery:
          false\n  dataFilters:   # dataFilters matches the first dataFilter selector,
          applies transforms and forwards to data-service and destinations\n  - altDestinations:
          \  # List of desintations to transform and foward events to\n    - authorization:
          \  # optional endpoint if an authorization token must first be requested
          \n        authDestHeader: Authorization   # http header key to be appended
          to destination request\n        authDestHeaderPrefix: 'Bearer '   # optional
          prefix for constructing destination request http header value\n        bodyData:
          \  # body data to send to authorization request, often an apiKey\n          secretKeyRef:\n
          \           name: auth-body-data-secret\n            key: bodydata\n        header:
          \  # secret map to use for authorization request headers\n          secret:\n
          \           name: auth-header-map-secret\n        tokenExpr: $.token   #
          optionally extract the returned body data using a jsonPath expression\n
          \       url: https://127.0.0.1:8443/api/2.0/accounts/my_account_id/apikeys/token
          \  # the url to make an authorization request to\n      header:   # secret
          map to use for destination request headers\n        secret:\n          name:
          dest-header-map-secret\n      transformer:   # The transformer to apply
          to event data for this destination\n        configMapKeyRef:\n          key:
          kazaam.json\n          name: kazaam-configmap\n        type: kazaam\n      url:
          https://127.0.0.1:8443/metering/api/1.0/usage   # the url to send the event
          to\n      urlSuffixExpr: $.properties.productId   # optional jsonPath expression
          on the event to extract and set a suffix to the url path\n    manifestType:
          dataReporter   # override manifestType of report sent to data-service\n
          \   selector:   # Matches against event data and a user. Empty selector
          matches all.\n      matchExpressions:   # jsonPath expression must return
          a result for all expressions to match (AND)\n      - $[?($.event == \"Account
          Contractual Usage\")]\n      - $.properties.productId\n      - $[?($.properties.source
          != null)]\n      - $[?($.properties.unit == \"AppPoints\")]\n      - $[?($.properties.quantity
          >= 0)]\n      - $[?($.timestamp != null)]\n      matchUsers:   # must match
          one of these users (OR). Omitting users matches any user\n      - system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n
          \   transformer:   # The transformer to apply to the event for data-service\n
          \     configMapKeyRef:\n        key: kazaam.json\n        name: kazaam-configmap\n
          \     type: kazaam\n  tlsConfig:   # TLS configuration for requests outbound
          to destinations\n    caCerts:\n    - name: tls-ca-certificates\n      key:
          ca.crt\n    certificates:\n    - clientCert:\n        secretKeyRef:\n          name:
          tls-ca-certificates\n          key: tls.crt\n      clientKey:\n        secretKeyRef:\n
          \         name: tls-ca-certificates\n          key: tls.key\n    cipherSuites:
          \n    - TLS_AES_128_GCM_SHA256\n    - TLS_AES_256_GCM_SHA384\n    - TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\n    - TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\n    insecureSkipVerify: false\n
          \   minVersion: VersionTLS12\n  userConfig:   # Optional metadata to apply
          to report sent to data-service for a specific user\n  - metadata:\n      ameta1:
          ametadata1\n      bmeta1: bmetadata1\n      cmeta1: cmetadata1\n      dmeta1:
          dmetadata1\n    userName: system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n```\n\nSample
          Header Secrets:\n```YAML\napiVersion: v1\nstringData:\n  accept: | \n    */*\nkind:
          Secret\nmetadata:\n  name: dest-header-map-secret\ntype: Opaque\n---\napiVersion:
          v1\nstringData:\n  accept: application/json\n  Content-Type: application/json\nkind:
          Secret\nmetadata:\n  name: auth-header-map-secret\ntype: Opaque\n```\n\nSample
          Authentication Body Secret:\n```YAML\napiVersion: v1\nstringData:\n  bodydata:
          |\n    {\"apikey\": \"myapikey\"}\nkind: Secret\nmetadata:\n  name: auth-body-data-secret\ntype:
          Opaque\n```\n\nSample TLS Secret:\n```YAML\napiVersion: v1\ndata:\n  ca.crt:
          LS0t...\n  tls.key: LS0t...\n  tls.crt: LS0t...\nkind: Secret\nmetadata:\n
          \ name: tls-ca-certificates\n```\n\nSample Kazaam ConfigMap:\n```YAML\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: kazaam-configmap\ndata:\n  kazaam.json:
          |\n    [\n        {\n            \"operation\": \"timestamp\",\n            \"spec\":
          {\n                \"timestamp\": {\n                \"inputFormat\": \"2006-01-02T15:04:05.999999+00:00\",\n
          \               \"outputFormat\": \"$unixmilli\"\n                }\n            }\n
          \       },\n        {\n            \"operation\": \"shift\",\n            \"spec\":
          {\n                \"instances[0].instanceId\": \"properties.source\",\n
          \               \"instances[0].startTime\": \"timestamp\",\n                \"instances[0].endTime\":
          \"timestamp\",\n                \"instances[0].metricUsage[0].metricId\":
          \"properties.unit\",\n                \"instances[0].metricUsage[0].quantity\":
          \"properties.quantity\"\n            }\n        },\n        {\n            \"operation\":
          \"default\",\n            \"spec\": {\n                \"meteringModel\":
          \"point-in-time\",\n                \"meteringPlan\": \"contract\"\n            }\n
          \       }\n    ]\n```\n\n### API Service User Configuration\n\n- The ClusterRole
          for api access is `clusterrole/ibm-data-reporter-operator-api`\n- The default
          ServiceAccount provided as an api user is `system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api`\n
          \ - The default ClusterRoleBinding for this user is `clusterrolebinding/ibm-data-reporter-operator-api`\n\nOptional:\n\n-
          To create an additional ServiceAccount and ClusterRoleBinding for api access\n\n```SHELL\nNAMESPACE=$(oc
          config view --minify -o jsonpath='{..namespace}') && \\\noc create serviceaccount
          my-api-service-account && \\\noc create clusterrolebinding ibm-data-reporter-operator-api
          --clusterrole=ibm-data-reporter-operator-api --serviceaccount=${NAMESPACE}:my-api-service-account\n```\n\n-
          Update datareporterconfig to attach metadata to reports associated with
          this user\n\n## API Service Usage\n\n- Get Token & Host\n\n```SHELL\noc
          project ibm-software-central\nexport DRTOKEN=$(oc create token ibm-data-reporter-operator-api
          --namespace ibm-software-central --duration 1h)\nexport DRHOST=$(oc get
          route ibm-data-reporter --template='{{ .spec.host }}')\n```\n\n- Get the
          Status\n\n```SHELL\ncurl -k -H \"Authorization: Bearer ${DRTOKEN}\" https://${DRHOST}/v1/status
          \n```\n\n- Post an Event\n\n```SHELL\ncurl -k -H \"Authorization: Bearer
          ${DRTOKEN}\" -X POST -d '{\"event\":\"myevent\"}' https://${DRHOST}/v1/event\n```\n\n##
          Storage\n\n- The operator temporarily stores events in pod memory, and writes
          to the IBM Metrics Operator data-service, which requires a PersistentVolume\n\n##
          License\n\nCopyright IBM Corporation 2023. All Rights Reserved.\n\nLicensed
          under the Apache License, Version 2.0 (the \"License\");\nyou may not use
          this file except in compliance with the License.\nYou may obtain a copy
          of the License at\n\n    http://www.apache.org/licenses/LICENSE-2.0\n\nUnless
          required by applicable law or agreed to in writing, software\ndistributed
          under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES
          OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for
          the specific language governing permissions and\nlimitations under the License.\n"
        displayName: IBM Data Reporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 2.20.0
      entries:
      - name: ibm-data-reporter-operator.v2.20.0
        version: 2.20.0
      name: beta
    - currentCSV: ibm-data-reporter-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "DataReporterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-data-reporter-operator",
                    "app.kubernetes.io/instance": "datareporterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datareporterconfig",
                    "app.kubernetes.io/part-of": "ibm-data-reporter-operator"
                  },
                  "name": "datareporterconfig"
                },
                "spec": {
                  "userConfig": [
                    {
                      "metadata": {
                        "ameta1": "ametadata1",
                        "bmeta1": "bmetadata1",
                        "cmeta1": "cmetadata1",
                        "dmeta1": "dmetadata1"
                      },
                      "userName": "system:serviceaccount:redhat-marketplace:ibm-data-reporter-operator-api"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
          createdAt: ""
          description: The IBM Data Reporter Operator accepts events and transforms
            them into reports submitted to the Data Service of the IBM Metrics Operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReporterConfig is the Schema for the datareporterconfigs
              API
            displayName: Data Reporter Config
            kind: DataReporterConfig
            name: datareporterconfigs.marketplace.redhat.com
            version: v1alpha1
          required:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
        description: "# IBM Data Reporter Operator\n\n# Introduction\n\nThe IBM Data
          Reporter Operator accepts events and transforms them into reports submitted
          to the IBM Metrics Operator Data Service.\n\n# Details\n\nThe IBM Data Reporter
          Operator deploys a service that provides an endpoint to which callers can
          send raw json event data. The event data is transformed into a report and
          is sent to the IBM Metrics Operator's Data Service. The IBM Metrics Operator
          Data Service periodically uploads the reports to IBM Software Central. It
          can also filter, transform and forward events to alternate endpoints.\n\n##
          Prerequisites\n\n- OpenShift Container Platform, major version 4 with any
          available supported minor version\n- Install IBM Metrics Operator\n  - IBM
          Data Reporter Operator prerequisties the IBM Metrics Operator's Data Service\n
          \ - Register the Cluster by creating a `redhat-marketplace-pull-secret`,
          as per the instructions\n  - Once `rhm-data-service` has started IBM Data
          Reporter Operator will accept requests\n\n## SecurityContextConstraints
          Requirements\n\n- The operator runs under Red Hat restricted SCC\n\n## Resources
          Required\n\n- The operator requires at least 85 Mi memory.\n\n## Limitations\n\n-
          Only runs on amd64, s390x and ppc64le architectures\n\n## Installing\n\n-
          A user with the cluster administrator role.\n- Install this operator in
          the same namespace as the IBM Metrics Operator\n  - default namespace: `ibm-software-central`\n\n##
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n# Configuration\n\nOptional:\n- Configure the
          Custom Resource Definition `DataReporterConfig` named `datareporterconfig`
          as per the following sample\n  - Configuration that can not be reconciled
          successfully will be reported on the DataReporterConfig Status\n  - jsonPath
          expressions are handled by: https://github.com/ohler55/ojg\n    - jsonPath
          comparison: https://cburgmer.github.io/json-path-comparison/\n  - Transformer
          type currently supported is `kazaam`\n    - https://github.com/willie68/kazaam\n\nSample
          DataReporterConfig:\n```YAML\napiVersion: marketplace.redhat.com/v1alpha1\nkind:
          DataReporterConfig\nmetadata:\n  name: datareporterconfig\nspec:\n  # true:
          report generated for 1 event and provide return code from data-service,
          confirming delivery\n  # false: events will be accumulated before building
          report, and immediately return 200, delivery unfconfirmed\n  confirmDelivery:
          false\n  dataFilters:   # dataFilters matches the first dataFilter selector,
          applies transforms and forwards to data-service and destinations\n  - altDestinations:
          \  # List of desintations to transform and foward events to\n    - authorization:
          \  # optional endpoint if an authorization token must first be requested
          \n        authDestHeader: Authorization   # http header key to be appended
          to destination request\n        authDestHeaderPrefix: 'Bearer '   # optional
          prefix for constructing destination request http header value\n        bodyData:
          \  # body data to send to authorization request, often an apiKey\n          secretKeyRef:\n
          \           name: auth-body-data-secret\n            key: bodydata\n        header:
          \  # secret map to use for authorization request headers\n          secret:\n
          \           name: auth-header-map-secret\n        tokenExpr: $.token   #
          optionally extract the returned body data using a jsonPath expression\n
          \       url: https://127.0.0.1:8443/api/2.0/accounts/my_account_id/apikeys/token
          \  # the url to make an authorization request to\n      header:   # secret
          map to use for destination request headers\n        secret:\n          name:
          dest-header-map-secret\n      transformer:   # The transformer to apply
          to event data for this destination\n        configMapKeyRef:\n          key:
          kazaam.json\n          name: kazaam-configmap\n        type: kazaam\n      url:
          https://127.0.0.1:8443/metering/api/1.0/usage   # the url to send the event
          to\n      urlSuffixExpr: $.properties.productId   # optional jsonPath expression
          on the event to extract and set a suffix to the url path\n    manifestType:
          dataReporter   # override manifestType of report sent to data-service\n
          \   selector:   # Matches against event data and a user. Empty selector
          matches all.\n      matchExpressions:   # jsonPath expression must return
          a result for all expressions to match (AND)\n      - $[?($.event == \"Account
          Contractual Usage\")]\n      - $.properties.productId\n      - $[?($.properties.source
          != null)]\n      - $[?($.properties.unit == \"AppPoints\")]\n      - $[?($.properties.quantity
          >= 0)]\n      - $[?($.timestamp != null)]\n      matchUsers:   # must match
          one of these users (OR). Omitting users matches any user\n      - system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n
          \   transformer:   # The transformer to apply to the event for data-service\n
          \     configMapKeyRef:\n        key: kazaam.json\n        name: kazaam-configmap\n
          \     type: kazaam\n  tlsConfig:   # TLS configuration for requests outbound
          to destinations\n    caCerts:\n    - name: tls-ca-certificates\n      key:
          ca.crt\n    certificates:\n    - clientCert:\n        secretKeyRef:\n          name:
          tls-ca-certificates\n          key: tls.crt\n      clientKey:\n        secretKeyRef:\n
          \         name: tls-ca-certificates\n          key: tls.key\n    cipherSuites:
          \n    - TLS_AES_128_GCM_SHA256\n    - TLS_AES_256_GCM_SHA384\n    - TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\n    - TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\n    insecureSkipVerify: false\n
          \   minVersion: VersionTLS12\n  userConfig:   # Optional metadata to apply
          to report sent to data-service for a specific user\n  - metadata:\n      ameta1:
          ametadata1\n      bmeta1: bmetadata1\n      cmeta1: cmetadata1\n      dmeta1:
          dmetadata1\n    userName: system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n```\n\nSample
          Header Secrets:\n```YAML\napiVersion: v1\nstringData:\n  accept: | \n    */*\nkind:
          Secret\nmetadata:\n  name: dest-header-map-secret\ntype: Opaque\n---\napiVersion:
          v1\nstringData:\n  accept: application/json\n  Content-Type: application/json\nkind:
          Secret\nmetadata:\n  name: auth-header-map-secret\ntype: Opaque\n```\n\nSample
          Authentication Body Secret:\n```YAML\napiVersion: v1\nstringData:\n  bodydata:
          |\n    {\"apikey\": \"myapikey\"}\nkind: Secret\nmetadata:\n  name: auth-body-data-secret\ntype:
          Opaque\n```\n\nSample TLS Secret:\n```YAML\napiVersion: v1\ndata:\n  ca.crt:
          LS0t...\n  tls.key: LS0t...\n  tls.crt: LS0t...\nkind: Secret\nmetadata:\n
          \ name: tls-ca-certificates\n```\n\nSample Kazaam ConfigMap:\n```YAML\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: kazaam-configmap\ndata:\n  kazaam.json:
          |\n    [\n        {\n            \"operation\": \"timestamp\",\n            \"spec\":
          {\n                \"timestamp\": {\n                \"inputFormat\": \"2006-01-02T15:04:05.999999+00:00\",\n
          \               \"outputFormat\": \"$unixmilli\"\n                }\n            }\n
          \       },\n        {\n            \"operation\": \"shift\",\n            \"spec\":
          {\n                \"instances[0].instanceId\": \"properties.source\",\n
          \               \"instances[0].startTime\": \"timestamp\",\n                \"instances[0].endTime\":
          \"timestamp\",\n                \"instances[0].metricUsage[0].metricId\":
          \"properties.unit\",\n                \"instances[0].metricUsage[0].quantity\":
          \"properties.quantity\"\n            }\n        },\n        {\n            \"operation\":
          \"default\",\n            \"spec\": {\n                \"meteringModel\":
          \"point-in-time\",\n                \"meteringPlan\": \"contract\"\n            }\n
          \       }\n    ]\n```\n\n### API Service User Configuration\n\n- The ClusterRole
          for api access is `clusterrole/ibm-data-reporter-operator-api`\n- The default
          ServiceAccount provided as an api user is `system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api`\n
          \ - The default ClusterRoleBinding for this user is `clusterrolebinding/ibm-data-reporter-operator-api`\n\nOptional:\n\n-
          To create an additional ServiceAccount and ClusterRoleBinding for api access\n\n```SHELL\nNAMESPACE=$(oc
          config view --minify -o jsonpath='{..namespace}') && \\\noc create serviceaccount
          my-api-service-account && \\\noc create clusterrolebinding ibm-data-reporter-operator-api
          --clusterrole=ibm-data-reporter-operator-api --serviceaccount=${NAMESPACE}:my-api-service-account\n```\n\n-
          Update datareporterconfig to attach metadata to reports associated with
          this user\n\n## API Service Usage\n\n- Get Token & Host\n\n```SHELL\noc
          project ibm-software-central\nexport DRTOKEN=$(oc create token ibm-data-reporter-operator-api
          --namespace ibm-software-central --duration 1h)\nexport DRHOST=$(oc get
          route ibm-data-reporter --template='{{ .spec.host }}')\n```\n\n- Get the
          Status\n\n```SHELL\ncurl -k -H \"Authorization: Bearer ${DRTOKEN}\" https://${DRHOST}/v1/status
          \n```\n\n- Post an Event\n\n```SHELL\ncurl -k -H \"Authorization: Bearer
          ${DRTOKEN}\" -X POST -d '{\"event\":\"myevent\"}' https://${DRHOST}/v1/event\n```\n\n##
          Storage\n\n- The operator temporarily stores events in pod memory, and writes
          to the IBM Metrics Operator data-service, which requires a PersistentVolume\n\n##
          License\n\nCopyright IBM Corporation 2023. All Rights Reserved.\n\nLicensed
          under the Apache License, Version 2.0 (the \"License\");\nyou may not use
          this file except in compliance with the License.\nYou may obtain a copy
          of the License at\n\n    http://www.apache.org/licenses/LICENSE-2.0\n\nUnless
          required by applicable law or agreed to in writing, software\ndistributed
          under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES
          OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for
          the specific language governing permissions and\nlimitations under the License.\n"
        displayName: IBM Data Reporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
        version: 2.20.0
      entries:
      - name: ibm-data-reporter-operator.v2.20.0
        version: 2.20.0
      name: stable
    defaultChannel: stable
    packageName: ibm-data-reporter-operator
    provider:
      name: Red Hat Marketplace
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ssm-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ssm-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ssm-controller:1.1.1
          createdAt: "2025-09-19T20:57:07Z"
          description: AWS SSM controller is a service controller for managing SSM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Document represents the state of an AWS ssm Document resource.
            displayName: Document
            kind: Document
            name: documents.ssm.services.k8s.aws
            version: v1alpha1
          - description: PatchBaseline represents the state of an AWS ssm PatchBaseline
              resource.
            displayName: PatchBaseline
            kind: PatchBaseline
            name: patchbaselines.ssm.services.k8s.aws
            version: v1alpha1
          - description: ResourceDataSync represents the state of an AWS ssm ResourceDataSync
              resource.
            displayName: ResourceDataSync
            kind: ResourceDataSync
            name: resourcedatasyncs.ssm.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon SSM resources in AWS from within your Kubernetes cluster.

          **About Amazon SSM**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon SSM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ssm
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SSM Developer Resources
          url: https://aws.amazon.com/SSM/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ssm maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ssm-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ssm-controller.v1.1.1
        version: 1.1.1
      - name: ack-ssm-controller.v1.1.0
        version: 1.1.0
      - name: ack-ssm-controller.v1.0.12
        version: 1.0.12
      - name: ack-ssm-controller.v1.0.10
        version: 1.0.10
      - name: ack-ssm-controller.v1.0.9
        version: 1.0.9
      - name: ack-ssm-controller.v1.0.8
        version: 1.0.8
      - name: ack-ssm-controller.v1.0.7
        version: 1.0.7
      - name: ack-ssm-controller.v1.0.6
        version: 1.0.6
      - name: ack-ssm-controller.v1.0.5
        version: 1.0.5
      - name: ack-ssm-controller.v1.0.4
        version: 1.0.4
      - name: ack-ssm-controller.v1.0.3
        version: 1.0.3
      - name: ack-ssm-controller.v1.0.2
        version: 1.0.2
      - name: ack-ssm-controller.v1.0.1
        version: 1.0.1
      - name: ack-ssm-controller.v1.0.0
        version: 1.0.0
      - name: ack-ssm-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-ssm-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: authorino-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: authorino-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta3",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/authorino-rhel9-operator@sha256:2d81613856df9725ad20928a7a9895d14482c42effce0a3a0de88f40ec35d411
          createdAt: "2025-03-25T22:39:23Z"
          description: Enables authentication and authorization for Gateways and applications
            in a Gateway API network.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta3
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: |-
          [Authorino](https://docs.kuadrant.io/authorino/) is Red Hat's Kubernetes-native lightweight external authorization service for tailor-made Zero Trust API security.

          Install this Red Hat official distribution of [Authorino Operator](https://docs.kuadrant.io/authorino-operator/) to manage instances of Authorino in this cluster.

          The current state of this distribution of the operator is: **Generally Available**.

          The Community version of Authorino Operator, based on upstream public images, is available in [OperatorHub.io](https://operatorhub.io/operator/authorino-operator).

          ## Getting started

          After installing the Operator, request an instance of the external authorization service by creating an `Authorino` custom resource.

          **Minimal example (namespaced)**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            listener:
              tls:
                enabled: false
            oidcServer:
              tls:
                enabled: false
          ```

          **Extended example**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            clusterWide: true
            authConfigLabelSelectors: environment=production
            secretLabelSelectors: authorino.kuadrant.io/component=authorino,environment=production

            replicas: 2

            evaluatorCacheSize: 2 # mb

            logLevel: info
            logMode: production

            listener:
              ports:
                grpc: 50001
                http: 5001
              tls:
                certSecretRef:
                  name: authorino-server-cert # secret must contain `tls.crt` and `tls.key` entries
              timeout: 2

            oidcServer:
              port: 8083
              tls:
                certSecretRef:
                  name: authorino-oidc-server-cert # secret must contain `tls.crt` and `tls.key` entries

            metrics:
              port: 8080
              deep: true

            healthz:
              port: 8081

            tracing:
              endpoint: rpc://otel-collector.observability.svc.cluster.local:4317
              insecure: true

            volumes:
              items:
                - name: keycloak-tls-cert
                  mountPath: /etc/ssl/certs
                  configMaps:
                    - keycloak-tls-cert
                  items: # details to mount the k8s configmap in the authorino pods
                    - key: keycloak.crt
                      path: keycloak.crt
              defaultMode: 420
          ```

          ### Cluster-wide vs Namespaced

          Namespaced instances only watch auth resources (`AuthConfig` and `Secrets`) created in the same namespace as the Authorino service. Use this mode for dedicated instances that do not require elevated privileges.

          Cluster-wide instances watch resources across the entire cluster (all namespaces.) Deploying and running Authorino in this mode requires elevated privileges.

          ### Multi-tenancy

          Use the `authConfigLabelSelectors` field of the `Authorino` custom resource to narrow the scope of the Authorino instance.

          Only `AuthConfig` custom resources whose labels match the value of this field will be handled by the Authorino instance.

          ## Protect a host

          To protect a host, create an `AuthConfig` custom resource for the host. E.g.:

          ```yaml
          apiVersion: authorino.kuadrant.io/v1beta3
          kind: AuthConfig
          metadata:
            name: my-api-protection
          spec:
            hosts:
            - my-api.io

            authentication:
              "keycloak":
                jwt:
                  issuerUrl: https://keycloak.keycloak.svc.cluster.local:8080/realms/my-realm

            authorization:
              "k8s-rbac":
                kubernetesSubjectAccessReview:
                  user:
                    selector: auth.identity.user.username
                  resourceAttributes:
                    resource:
                      value: my-api
                    verb:
                      selector: request.method
                cache:
                  key:
                    selector: auth.identity.user.username
                  ttl: 30
              "after-2am-only":
                rego: |
                  allow {
                    [hour, _, _] := time.clock(time.now_ns())
                    hour >= 2
                  }
          ```

          Make sure all requests to the host are fisrt checked with the Authorino instance, by configuring an Envoy proxy for external authz:

          ```yaml
          clusters:
          - name: my-api
            …
          - name: authorino
            connect_timeout: 0.25s
            type: STRICT_DNS
            lb_policy: ROUND_ROBIN
            http2_protocol_options: {}
            load_assignment:
              cluster_name: authorino
              endpoints:
              - lb_endpoints:
                - endpoint:
                    address:
                      socket_address:
                        address: authorino-authorino-authorization
                        port_value: 50051
          listeners:
          - filter_chains:
            - filters:
              name: envoy.http_connection_manager
              typed_config:
                "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
                use_remote_address: true
                route_config:
                  name: my-api-config
                  virtual_hosts:
                  - name: my-api-vs
                    domains:
                    - my-api.io
                    routes:
                    - match:
                        prefix: /
                      route:
                        cluster: my-api
                http_filters:
                - name: envoy.filters.http.ext_authz
                  typed_config:
                    "@type": type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz
                    transport_api_version: V3
                    failure_mode_allow: false
                    include_peer_certificate: true
                    grpc_service:
                      envoy_grpc:
                        cluster_name: authorino
                      timeout: 1s
          ```

          ...or, if using Istio, by creating an [`AuthorizationPolicy`](https://istio.io/latest/docs/reference/config/security/authorization-policy/#AuthorizationPolicy) custom resource. Use `action: CUSTOM` in the resource and the Authorino authorization service configured in the mesh extension provider settings.

          ## Features

          **Authentication**

          * JWT validation (with OpenID Connect Discovery)
          * OAuth 2.0 Token Introspection (opaque tokens)
          * Kubernetes TokenReview (ServiceAccount tokens)
          * API key authentication
          * X.509 client certificate authentication
          * Anonymous access
          * Proxy-handled (authentication performed by the proxy)

          **Authorization**

          * Built-in simple pattern matching (e.g. JWT claims, request attributes checking)
          * OPA policies (inline Rego and fetch from external registry)
          * Kubernetes SubjectAccessReview (resource and non-resource attributes)
          * Authzed SpiceDB

          **External metadata**

          * HTTP request
          * OpenID Connect User Info
          * UMA-protected resource attributes

          **Custom responses**

          * Header injection (Festival Wristbands tokens, JSON, plain text)
          * Envoy Dynamic Metadata
          * Custom HTTP response (status code, headers, messages, body, etc)

          **Callbacks**

          * HTTP webhooks

          **Caching**

          * OpenID Connect and User-Managed Access configs
          * JSON Web Keys (JWKs) and JSON Web Key Sets (JWKS)
          * Access tokens
          * External metadata
          * Precompiled Rego policies
          * Policy evaluation

          Check out the full [Feature Specification](https://docs.kuadrant.io/authorino/docs/features/) and how-to guides in the [Kuadrant Docs](https://docs.kuadrant.io) website.
        displayName: Red Hat - Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: GitHub
          url: https://github.com/kuadrant/authorino-operator
        - name: Documentation
          url: https://docs.kuadrant.io/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhcl-1/authorino-rhel9-operator@sha256:2d81613856df9725ad20928a7a9895d14482c42effce0a3a0de88f40ec35d411
        - registry.redhat.io/rhcl-1/authorino-rhel9@sha256:413d3796ad0101bf14d3341d31d9367e67fea1e73ebb3fb79124ef4c5dde0c09
        version: 1.2.2
      entries:
      - name: authorino-operator.v1.2.2
        version: 1.2.2
      - name: authorino-operator.v1.2.1
        version: 1.2.1
      - name: authorino-operator.v1.2.0
        version: 1.2.0
      - name: authorino-operator.v1.1.3
        version: 1.1.3
      - name: authorino-operator.v1.1.2
        version: 1.1.2
      - name: authorino-operator.v1.1.1
        version: 1.1.1
      - name: authorino-operator.v1.1.0
        version: 1.1.0
      - name: authorino-operator.v1.0.2
        version: 1.0.2
      - name: authorino-operator.v0.16.1
        version: 0.16.1
      - name: authorino-operator.v0.16.0
        version: 0.16.0
      name: stable
    - currentCSV: authorino-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta1",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "hosts": [
                    "my-api.io"
                  ],
                  "identity": [
                    {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "in": "authorization_header",
                        "keySelector": "APIKEY"
                      },
                      "name": "api-key-users"
                    }
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/3scale-tech-preview/authorino-rhel9-operator@sha256:eb574781cf39917f8f5d9e5c806f520534601f819ee1ebcbb751832da8938008
          createdAt: "2024-11-08T11:06:33Z"
          description: Authorino
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta1
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: |-
          [Authorino](https://docs.kuadrant.io/authorino/) is Red Hat's Kubernetes-native lightweight external authorization service for tailor-made Zero Trust API security.

          Install this Red Hat official distribution of [Authorino Operator](https://docs.kuadrant.io/authorino-operator/) to manage instances of Authorino in this cluster.

          The current state of this distribution of the operator is: **_Technical Preview_**.

          The Community version of Authorino Operator, based on upstream public images, is available in [OperatorHub.io](https://operatorhub.io/operator/authorino-operator).

          ## Getting started

          After installing the Operator, request an instance of the external authorization service by creating an `Authorino` custom resource.

          **Minimal example (namespaced)**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            listener:
              tls:
                enabled: false
            oidcServer:
              tls:
                enabled: false
          ```

          **Extended example**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            clusterWide: true
            authConfigLabelSelectors: environment=production
            secretLabelSelectors: authorino.kuadrant.io/component=authorino,environment=production

            replicas: 2

            evaluatorCacheSize: 2 # mb

            logLevel: info
            logMode: production

            listener:
              ports:
                grpc: 50001
                http: 5001
              tls:
                certSecretRef:
                  name: authorino-server-cert # secret must contain `tls.crt` and `tls.key` entries
              timeout: 2

            oidcServer:
              port: 8083
              tls:
                certSecretRef:
                  name: authorino-oidc-server-cert # secret must contain `tls.crt` and `tls.key` entries

            metrics:
              port: 8080
              deep: true

            healthz:
              port: 8081

            tracing:
              endpoint: rpc://otel-collector.observability.svc.cluster.local:4317
              insecure: true

            volumes:
              items:
                - name: keycloak-tls-cert
                  mountPath: /etc/ssl/certs
                  configMaps:
                    - keycloak-tls-cert
                  items: # details to mount the k8s configmap in the authorino pods
                    - key: keycloak.crt
                      path: keycloak.crt
              defaultMode: 420
          ```

          ### Cluster-wide vs Namespaced

          Namespaced instances only watch auth resources (`AuthConfig` and `Secrets`) created in the same namespace as the Authorino service. Use this mode for dedicated instances that do not require elevated privileges.

          Cluster-wide instances watch resources across the entire cluster (all namespaces.) Deploying and running Authorino in this mode requires elevated privileges.

          ### Multi-tenancy

          Use the `authConfigLabelSelectors` field of the `Authorino` custom resource to narrow the scope of the Authorino instance.

          Only `AuthConfig` custom resources whose labels match the value of this field will be handled by the Authorino instance.

          ## Protect a host

          To protect a host, create an `AuthConfig` custom resource for the host. E.g.:

          ```yaml
          apiVersion: authorino.kuadrant.io/v1beta2
          kind: AuthConfig
          metadata:
            name: my-api-protection
          spec:
            hosts:
            - my-api.io

            authentication:
              "keycloak":
                jwt:
                  issuerUrl: https://keycloak.keycloak.svc.cluster.local:8080/realms/my-realm

            authorization:
              "k8s-rbac":
                kubernetesSubjectAccessReview:
                  user:
                    selector: auth.identity.user.username
                  resourceAttributes:
                    resource:
                      value: my-api
                    verb:
                      selector: request.method
                cache:
                  key:
                    selector: auth.identity.user.username
                  ttl: 30
              "after-2am-only":
                rego: |
                  allow {
                    [hour, _, _] := time.clock(time.now_ns())
                    hour >= 2
                  }
          ```

          Make sure all requests to the host are fisrt checked with the Authorino instance, by configuring an Envoy proxy for external authz:

          ```yaml
          clusters:
          - name: my-api
            …
          - name: authorino
            connect_timeout: 0.25s
            type: STRICT_DNS
            lb_policy: ROUND_ROBIN
            http2_protocol_options: {}
            load_assignment:
              cluster_name: authorino
              endpoints:
              - lb_endpoints:
                - endpoint:
                    address:
                      socket_address:
                        address: authorino-authorino-authorization
                        port_value: 50051
          listeners:
          - filter_chains:
            - filters:
              name: envoy.http_connection_manager
              typed_config:
                "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
                use_remote_address: true
                route_config:
                  name: my-api-config
                  virtual_hosts:
                  - name: my-api-vs
                    domains:
                    - my-api.io
                    routes:
                    - match:
                        prefix: /
                      route:
                        cluster: my-api
                http_filters:
                - name: envoy.filters.http.ext_authz
                  typed_config:
                    "@type": type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz
                    transport_api_version: V3
                    failure_mode_allow: false
                    include_peer_certificate: true
                    grpc_service:
                      envoy_grpc:
                        cluster_name: authorino
                      timeout: 1s
          ```

          ...or, if using Istio, by creating an [`AuthorizationPolicy`](https://istio.io/latest/docs/reference/config/security/authorization-policy/#AuthorizationPolicy) custom resource. Use `action: CUSTOM` in the resource and the Authorino authorization service configured in the mesh extension provider settings.

          ## Features

          **Authentication**

          * JWT validation (with OpenID Connect Discovery)
          * OAuth 2.0 Token Introspection (opaque tokens)
          * Kubernetes TokenReview (ServiceAccount tokens)
          * API key authentication
          * X.509 client certificate authentication
          * Anonymous access
          * Proxy-handled (authentication performed by the proxy)

          **Authorization**

          * Built-in simple pattern matching (e.g. JWT claims, request attributes checking)
          * OPA policies (inline Rego and fetch from external registry)
          * Kubernetes SubjectAccessReview (resource and non-resource attributes)
          * Authzed SpiceDB

          **External metadata**

          * HTTP request
          * OpenID Connect User Info
          * UMA-protected resource attributes

          **Custom responses**

          * Header injection (Festival Wristbands tokens, JSON, plain text)
          * Envoy Dynamic Metadata
          * Custom HTTP response (status code, headers, messages, body, etc)

          **Callbacks**

          * HTTP webhooks

          **Caching**

          * OpenID Connect and User-Managed Access configs
          * JSON Web Keys (JWKs) and JSON Web Key Sets (JWKS)
          * Access tokens
          * External metadata
          * Precompiled Rego policies
          * Policy evaluation

          Check out the full [Feature Specification](https://docs.kuadrant.io/authorino/docs/features/) and how-to guides in the [Kuadrant Docs](https://docs.kuadrant.io) website.
        displayName: Red Hat - Authorino (Technical Preview)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: GitHub
          url: https://github.com/kuadrant/authorino-operator
        - name: Documentation
          url: https://docs.kuadrant.io/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-tech-preview/authorino-rhel9@sha256:a473dae20e71e3e813ac30ba978f2ab3c5e19d7d39b501ae9103dca892107c87
        - registry.redhat.io/3scale-tech-preview/authorino-rhel9-operator@sha256:eb574781cf39917f8f5d9e5c806f520534601f819ee1ebcbb751832da8938008
        - registry.redhat.io/openshift4/ose-cli-rhel9@sha256:3a398158d10018bfac5df291d6998b513cf9a30e5477bcaefddcf66d229627c7
        version: 1.1.3
      entries:
      - name: authorino-operator.v1.1.3
        version: 1.1.3
      - name: authorino-operator.v1.1.2
        version: 1.1.2
      - name: authorino-operator.v1.1.1
        version: 1.1.1
      - name: authorino-operator.v1.1.0
        version: 1.1.0
      - name: authorino-operator.v1.0.2
        version: 1.0.2
      name: tech-preview-v1
    defaultChannel: stable
    packageName: authorino-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-xtradb-cluster-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "docker.io/percona/percona-xtradb-cluster:8.0.42-33.1",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "docker.io/percona/haproxy:2.8.15",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "docker.io/percona/proxysql2:2.7.3",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "docker.io/percona/fluentbit:4.0.1",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:3.3.1",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "docker.io/percona/percona-xtrabackup:8.0.35-34.1",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.io/percona/percona-xtradb-cluster-operator:1.18.0
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <v1.18.0
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            clustercheck: Y2x1c3RlcmNoZWNrcGFzc3dvcmQ=
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            clustercheck: Y2x1c3RlcmNoZWNrcGFzc3dvcmQ=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ```
          ###  Release Highlights
          * **PMM3 support**

          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**

          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**

          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**

          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.

        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - database
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - docker.io/percona/percona-xtradb-cluster-operator:1.18.0
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator.v1.17.0
        version: 1.17.0
      - name: percona-xtradb-cluster-operator.v1.16.1
        version: 1.16.1
      - name: percona-xtradb-cluster-operator.v1.16.0
        version: 1.16.0
      - name: percona-xtradb-cluster-operator.v1.15.1
        version: 1.15.1
      - name: percona-xtradb-cluster-operator.v1.15.0
        version: 1.15.0
      - name: percona-xtradb-cluster-operator.v1.9.0
        version: 1.9.0
      name: stable
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator
    provider:
      name: Percona
      url: https://www.percona.com/
kind: List
metadata:
  resourceVersion: ""
apiVersion: v1
items:
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: ovms-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openvino-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "ModelServer",
                "metadata": {
                  "name": "model-server-sample"
                },
                "spec": {
                  "image_name": "registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0",
                  "deployment_parameters": {
                    "replicas": 1
                  },
                  "service_parameters": {
                    "grpc_port": 8080,
                    "rest_port": 8081
                  },
                  "models_settings": {
                    "single_model_mode": true,
                    "config_configmap_name": "",
                    "config_path": "",
                    "model_name": "resnet",
                    "model_path": "gs://<bucket_name>/<model_dir>",
                    "nireq": 0,
                    "plugin_config": "{\"PERFORMANCE_HINT\":\"LATENCY\"}",
                    "batch_size": "",
                    "shape": "",
                    "model_version_policy": "{\"latest\": { \"num_versions\":1 }}",
                    "layout": "",
                    "target_device": "CPU",
                    "is_stateful": false,
                    "idle_sequence_cleanup": false,
                    "low_latency_transformation": true,
                    "max_sequence_number": 0
                  },
                  "server_settings": {
                    "file_system_poll_wait_seconds": 0,
                    "sequence_cleaner_poll_wait_minutes": 0,
                    "log_level": "INFO",
                    "grpc_workers": 1,
                    "rest_workers": 0
                  },
                  "models_repository": {
                    "https_proxy": "",
                    "http_proxy": "",
                    "storage_type": "google",
                    "models_host_path": "",
                    "models_volume_claim": "",
                    "aws_secret_access_key": "",
                    "aws_access_key_id": "",
                    "aws_region": "",
                    "s3_compat_api_endpoint": "",
                    "gcp_creds_secret_name": "",
                    "azure_storage_connection_string": ""
                  }
                }
              },
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "Notebook",
                "metadata": {
                  "name": "v2024.1"
                },
                "spec": {
                  "git_ref": "2024.1",
                  "auto_update_image": false
                }
              }
            ]
          alm-examples-metadata: '{ "model-server-sample":{"description":"Example
            of the model server configuration"}, "v2024":{"description":"Default notebook
            configuration"} }'
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
          createdAt: "2024-05-14T09:29:35Z"
          description: OpenVINO Toolkit Operator manages OpenVINO components in OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/ovms-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/ovms-operator-rhmp/support?utm_source=openshift_console
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.34.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Provides inference serving through gRPC & REST
            displayName: Model Server
            kind: ModelServer
            name: modelservers.intel.com
            version: v1alpha1
          - description: Adds a collection of ready-to-run Jupyter notebooks for learning
              and experimenting with the OpenVINO™ Toolkit to Red Hat® OpenShift®
              Data Science
            displayName: Notebook
            kind: Notebook
            name: notebooks.intel.com
            version: v1alpha1
        description: |
          OpenVINO Toolkit Operator manages OpenVINO components in OpenShift.

          Currently there available components are ModelServer and Notebook.

          # Model Server
          [OpenVINO™ Model Server](https://github.com/openvinotoolkit/model_server) (OVMS) is a scalable, high-performance solution for serving machine learning models optimized for Intel® architectures. The server provides an inference service via gRPC or REST API for any models trained in a framework that is supported by [OpenVINO](https://software.intel.com/en-us/openvino-toolki://docs.openvino.ai/latest/index.html).
          Model Server configuration parameters is explained [here](https://github.com/openvinotoolkit/operator/blob/main/docs/modelserver_params.md).
          ## Using the cluster
          OpenVINO Model Server can be consumed as a `Service`. It is called like with `ModelServer` resource with `-ovms` suffix.
          The suffix is omitted when `ovms` phrase is included in the name.
          The service exposes gRPC and REST API interfaces to run inference requests.
          ```
          oc get pods
          NAME                                        READY   STATUS    RESTARTS   AGE
          model-server-sample-ovms-586f6f76df-dpps4   1/1     Running   0          8h

          oc get services
          NAME                       TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)             AGE
          model-server-sample-ovms   ClusterIP   172.25.199.210   <none>        8080/TCP,8081/TCP   8h
          ```

          # Notebook
          The Notebook resource integrates JupyterHub from OpenShift Data Science or Open Data Hub with a container image that includes [developer
          tools](https://github.com/openvinotoolkit/openvino/blob/master/docs/install_guides/pypi-openvino-dev.md)
          from the OpenVINO toolkit and a set of Jupyter notebook tutorials. It enables selecting a defined image `OpenVINO™ Toolkit` from
          the Jupyter Spawner choice list.

          Create the `Notebook` resource in the same project with JupyterHub and RedHat OpenShift Data Science operator as seen in [documentation](https://github.com/openvinotoolkit/operator/blob/main/docs/notebook_in_rhods.md).
          It builds the image in the cluster based on Dockerfile from [openvino_notebooks](https://github.com/openvinotoolkit/openvino_notebooks).

          ## References
          OpenVINO Model Server on [Github](https://github.com/openvinotoolkit/model_server)

          OpenVINO Model Server Operator on [Github](https://github.com/openvinotoolkit/operator)
        displayName: OpenVINO Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Inference OpenVINO
        links:
        - name: OpenVINO Toolkit Operator
          url: https://github.com/openvinotoolkit/operator
        - name: OpenVINO Model Server
          url: https://github.com/openvinotoolkit/model_server
        - name: OpenVINO Notebooks
          url: https://github.com/openvinotoolkit/openvino_notebooks
        maintainers:
        - email: openvino-operator-support@intel.com
          name: Operator Support
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5714b93ac39277affdc10c3ba142d8c032ed0ddb2ba61c04cbddac710464c4
        - registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
        - registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0
        version: 1.2.0
      entries:
      - name: openvino-operator.v1.2.0
        version: 1.2.0
      - name: openvino-operator.v1.1.0
        version: 1.1.0
      - name: openvino-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: ovms-operator-rhmp
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Aerospike
      provider-url: ""
    name: aerospike-kubernetes-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aerospike-kubernetes-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "memory-size": 3000000000,
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:6.3.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: quay.io/aerospike/aerospike-kubernetes-operator@sha256:a3033b50ae434b97de4de1ca03013dabcfac580d810a6ee8c0fe05b6a9301952
          createdAt: 2023-05-04T09:32:47UTC
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: <=2.4.0
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Aerospike
        relatedImages:
        - quay.io/aerospike/aerospike-kubernetes-operator@sha256:a3033b50ae434b97de4de1ca03013dabcfac580d810a6ee8c0fe05b6a9301952
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        version: 2.5.0
      entries:
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      name: alpha
    - currentCSV: aerospike-kubernetes-operator-rhmp.v4.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "data-size": 1073741824,
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:8.1.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackup",
                "metadata": {
                  "name": "aerospikebackup",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "aerospike-cluster": {
                      "aerospike-aerospikebackup-test-cluster": {
                        "credentials": {
                          "password": "admin123",
                          "user": "admin"
                        },
                        "seed-nodes": [
                          {
                            "host-name": "aerocluster.aerospike.svc.cluster.local",
                            "port": 3000
                          }
                        ]
                      }
                    },
                    "backup-routines": {
                      "aerospike-aerospikebackup-test-routine": {
                        "backup-policy": "test-policy",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "local"
                      },
                      "aerospike-aerospikebackup-test-routine1": {
                        "backup-policy": "test-policy1",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "s3Storage"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackupService",
                "metadata": {
                  "name": "aerospikebackupservice-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "config": {
                    "backup-policies": {
                      "test-policy": {
                        "parallel": 3
                      },
                      "test-policy1": {
                        "parallel": 3
                      }
                    },
                    "service": {
                      "http": {
                        "port": 8081
                      }
                    },
                    "storage": {
                      "local": {
                        "local-storage": {
                          "path": "/tmp/localStorage"
                        }
                      },
                      "s3Storage": {
                        "s3-storage": {
                          "bucket": "aerospike-kubernetes-operator-test",
                          "s3-profile": "default",
                          "s3-region": "us-east-1"
                        }
                      }
                    }
                  },
                  "image": "aerospike/aerospike-backup-service:3.2.0",
                  "secrets": [
                    {
                      "secretName": "aws-secret",
                      "volumeMount": {
                        "mountPath": "/root/.aws/credentials",
                        "name": "aws-secret",
                        "subPath": "credentials"
                      }
                    }
                  ],
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeRestore",
                "metadata": {
                  "name": "aerospikerestore-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "backup-data-path": "aerospike-aerospikebackup-test-routine/backup/1733506015827/data/test",
                    "destination": {
                      "credentials": {
                        "password": "admin123",
                        "user": "admin"
                      },
                      "label": "destinationCluster",
                      "seed-nodes": [
                        {
                          "host-name": "aerocluster.test.svc.cluster.local",
                          "port": 3000
                        }
                      ]
                    },
                    "policy": {
                      "no-generation": true,
                      "no-indexes": true,
                      "parallel": 3
                    },
                    "source": {
                      "local-storage": {
                        "path": "/tmp/localStorage"
                      }
                    }
                  },
                  "type": "Full"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/aerospike/aerospike-kubernetes-operator@sha256:e6d6bd7c58e036b3728e52475a9b764ca8c46b83cbc4751cac66294495bf2b31
          createdAt: "2025-09-15T10:14:05Z"
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
          support: Aerospike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeBackup is the Schema for the aerospikebackup API
            displayName: Aerospike Backup
            kind: AerospikeBackup
            name: aerospikebackups.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeBackupService is the Schema for the aerospikebackupservices
              API
            displayName: Aerospike Backup Service
            kind: AerospikeBackupService
            name: aerospikebackupservices.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1
          - description: AerospikeRestore is the Schema for the aerospikerestores
              API
            displayName: Aerospike Restore
            kind: AerospikeRestore
            name: aerospikerestores.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
           -   Aerospike cluster monitoring
           -   Backup and restore Aerospike clusters
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Aerospike
        relatedImages:
        - quay.io/aerospike/aerospike-kubernetes-operator@sha256:e6d6bd7c58e036b3728e52475a9b764ca8c46b83cbc4751cac66294495bf2b31
        version: 4.1.1
      entries:
      - name: aerospike-kubernetes-operator-rhmp.v4.1.1
        version: 4.1.1
      - name: aerospike-kubernetes-operator-rhmp.v4.1.0
        version: 4.1.0
      - name: aerospike-kubernetes-operator-rhmp.v4.0.2
        version: 4.0.2
      - name: aerospike-kubernetes-operator-rhmp.v4.0.1
        version: 4.0.1
      - name: aerospike-kubernetes-operator-rhmp.v4.0.0
        version: 4.0.0
      - name: aerospike-kubernetes-operator-rhmp.v3.4.1
        version: 3.4.1
      - name: aerospike-kubernetes-operator-rhmp.v3.4.0
        version: 3.4.0
      - name: aerospike-kubernetes-operator-rhmp.v3.3.1
        version: 3.3.1
      - name: aerospike-kubernetes-operator-rhmp.v3.3.0
        version: 3.3.0
      - name: aerospike-kubernetes-operator-rhmp.v3.2.2
        version: 3.2.2
      - name: aerospike-kubernetes-operator-rhmp.v3.2.1
        version: 3.2.1
      - name: aerospike-kubernetes-operator.v3.2.0
        version: 3.2.0
      - name: aerospike-kubernetes-operator.v3.1.0
        version: 3.1.0
      - name: aerospike-kubernetes-operator.v3.0.0
        version: 3.0.0
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      name: stable
    defaultChannel: stable
    packageName: aerospike-kubernetes-operator-rhmp
    provider:
      name: Aerospike
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: samsung
      provider-url: ""
    name: socmmd
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: socmmd.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cmmd.samsung.com/v1",
                "kind": "CMMD",
                "metadata": {
                  "name": "my-app",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "allocateMode": "auto",
                  "enable": true,
                  "payload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "metadata": {
                      "name": "my-app"
                    },
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "name": "my-app"
                        }
                      },
                      "template": {
                        "metadata": {
                          "labels": {
                            "name": "my-app"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "quay.io/my-account/my-app:latest",
                              "name": "my-app-01"
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "cmmd.samsung.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "machineconfiguration.openshift.io/role": "cmmd"
                  },
                  "name": "cmmd-node-label",
                  "namespace": "socmmd"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          console.openshift.io/plugins: '["socmmd-console-plugin"]'
          containerImage: quay.io/samsung_sag/socmmd@sha256:2ef361fb15f0fec99e0a8f1ef83ad4dd14f701d579c900cce084c4a62cf805f8
          createdAt: "2024-08-01T05:45:38Z"
          description: Samsung Operator For CMM-D
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "cmmd.samsung.com/v1",
              "kind": "Label",
              "metadata": {
                "name": "cmmd-node-label",
                "namespace": "socmmd",
                "labels": {}
              },
              "spec": {}
            }
          operatorframework.io/suggested-namespace: socmmd-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenMPDK/SMDK/tree/main/lib/redhat/socmmd
          support: samsung
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CMMD is the Schema for the cmmds API.
            displayName: CMMD
            kind: CMMD
            name: cmmds.cmmd.samsung.com
            version: v1
          - description: Label is the Schema for the labels API.
            displayName: Node Label
            kind: Label
            name: labels.cmmd.samsung.com
            version: v1
        description: Samsung Operator For CMM-D
        displayName: socmmd
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - samsung
        - memory
        - extend
        - cmm-d
        links:
        - name: Samsung Operator For CMM-D
          url: https://github.com/OpenMPDK/SMDK/tree/main/lib/redhat/socmmd
        maintainers:
        - email: junhyeok.im@samsung.com
          name: junhyeok-im
        maturity: stable
        provider:
          name: samsung
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/samsung_sag/socmmd@sha256:2ef361fb15f0fec99e0a8f1ef83ad4dd14f701d579c900cce084c4a62cf805f8
        version: 0.0.2
      entries:
      - name: socmmd.v0.0.2
        version: 0.0.2
      - name: socmmd.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: socmmd
    provider:
      name: samsung
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicast-community-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicast-community-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/apicast-operator:v0.8.0
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway built on top of NGINX. It is part
            of the Red Hat 3scale API Management Platform
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |
          The APIcast Operator creates and maintains the Red Hat 3scale API Gateway in several deployment configurations.

          APIcast is an API gateway built on top of [NGINX](https://www.nginx.com/). It is part of the [Red Hat 3scale API Management Platform](https://www.redhat.com/en/technologies/jboss-middleware/3scale).

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator/blob/v0.8.0/doc/operator-user-guide.md#upgrading-APIcast) for more information.

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/apicast-operator/tree/v0.8.0).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: APIcast
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://github.com/3scale/apicast-operator/tree/v0.8.0
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/apicast-operator:v0.8.0
        version: 0.8.0
      entries:
      - name: apicast-community-operator.v0.8.0
        version: 0.8.0
      - name: apicast-community-operator.v0.7.1
        version: 0.7.1
      - name: apicast-community-operator.v0.6.0
        version: 0.6.0
      - name: apicast-community-operator.v0.5.1
        version: 0.5.1
      name: alpha
    - currentCSV: apicast-community-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/apicast-operator:v0.8.0
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway built on top of NGINX. It is part
            of the Red Hat 3scale API Management Platform
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |
          The APIcast Operator creates and maintains the Red Hat 3scale API Gateway in several deployment configurations.

          APIcast is an API gateway built on top of [NGINX](https://www.nginx.com/). It is part of the [Red Hat 3scale API Management Platform](https://www.redhat.com/en/technologies/jboss-middleware/3scale).

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator/blob/v0.8.0/doc/operator-user-guide.md#upgrading-APIcast) for more information.

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/apicast-operator/tree/v0.8.0).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: APIcast
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://github.com/3scale/apicast-operator/tree/v0.8.0
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/apicast-operator:v0.8.0
        version: 0.8.0
      entries:
      - name: apicast-community-operator.v0.8.0
        version: 0.8.0
      - name: apicast-community-operator.v0.7.1
        version: 0.7.1
      - name: apicast-community-operator.v0.6.0
        version: 0.6.0
      - name: apicast-community-operator.v0.5.1
        version: 0.5.1
      name: stable
    defaultChannel: stable
    packageName: apicast-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-appsec
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-appsec.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Appsec",
                "metadata": {
                  "name": "appsec-sample"
                },
                "spec": {
                  "appsecDataModel": {
                    "port": 8080
                  },
                  "appsecResourceService": {
                    "port": 8080
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "remoteRepository": "docker.io",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "reportTemplates": {
                    "enabled": false
                  },
                  "systemMonitor": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:39:34Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Appsec
            name: appsecs.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP APPSEC Operator
        displayName: Keysight WAP APPSEC Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Appsec
          url: https://keysight-wap-appsec.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-appsec@sha256:e39ff225c26b750f86553693dda47d0dcc9834a4ab0975aea154361ee3527314
        version: 0.1.21
      entries:
      - name: keysight-wap-appsec.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-appsec
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-operator.v23.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerify",
                "metadata": {
                  "name": "ibmsecurityverify-sample"
                },
                "spec": {
                  "clientSecret": "--secret--",
                  "sessionLifetime": 3600,
                  "ssoPath": "/verify-sso"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isv-saas/verify-operator@sha256:c3bed2e5db48e54884d2278d24ca4864f389a56a667db481c627d01ab5ffaba7
          createdAt: "2023-03-22T02:49:11Z"
          description: The IBM Security Verify operator can consistently enforce policy-driven
            security, including authentication and authorization, using the Ingress
            networking capability of OpenShift.
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.9"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM-Security/verify-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerify is the Schema for the ibmsecurityverifies
              API.
            displayName: IBMSecurity Verify
            kind: IBMSecurityVerify
            name: ibmsecurityverifies.ibm.com
            version: v1
        description: "[IBM Security Verify](https://www.ibm.com/products/verify-saas)
          is an Identity-as-a-Service platform that allows IT, security and business
          leaders to protect their digital users, assets and data in a hybrid, multi-cloud
          world by enabling technical agility and operational efficiency. IBM Security
          Verify SaaS provides single sign-on (SSO), multi-factor authentication (MFA),
          AI-powered context for risk-based authentication for adaptive access decisions,
          user management, access recertification campaigns and identity analytics.\n\nAs
          part of using Red Hat OpenShift, you are entitled to Verify SaaS SSO for
          unlimited apps and users.  If you do not already have a Verify SaaS SSO
          tenant a new one can be [created](https://www.ibm.com/account/reg/signup?formid=urx-51255).\n\nFor
          a detailed description of IBM Security Verify refer to the [Official documentation](https://www.ibm.com/docs/en/security-verify).\n\nThe
          IBM Security Verify operator can consistently enforce policy-driven security
          by using the Ingress networking capability of OpenShift in conjunction with
          the [Nginx Ingress operator](https://www.nginx.com/blog/getting-started-nginx-ingress-operator-red-hat-openshift/).
          With this approach, you can enforce authentication and authorization policies
          for all of the applications in your cluster at the same time, without ever
          changing your application code. You can also dynamically register your application
          to start protecting them centrally from the cloud via Verify SaaS. \n\nSee
          the project [Readme](https://github.com/IBM-Security/verify-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Security Verify Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Product Information
          url: https://www.ibm.com/products/verify-saas
        - name: Verify Documentation
          url: https://www.ibm.com/docs/en/security-verify
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Development Team
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - icr.io/isv-saas/verify-operator@sha256:c3bed2e5db48e54884d2278d24ca4864f389a56a667db481c627d01ab5ffaba7
        version: 23.3.0
      entries:
      - name: ibm-security-verify-operator.v23.3.0
        version: 23.3.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hitachi
      provider-url: ""
    name: hspc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hspc-operator.v1.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.hitachi.com/v1",
                "kind": "HSPC",
                "metadata": {
                  "name": "hspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.hitachivantara.com/hitachicsi-oci-oss/hspc-operator@sha256:4dda4a43cefc19c8b1992756b499b1f9300920c5250dd77379f523997cfd5d54
          createdAt: "2025-06-13T10:30:39Z"
          description: An operator for managing Hitachi Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: Hitachi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HSPC is the Schema for the hspcs API
            displayName: HSPC
            kind: HSPC
            name: hspcs.csi.hitachi.com
            version: v1
        description: |-
          ## About
          Hitachi Storage Plug-in for Containers is a software plugin component for Kubernetes and OpenShift environments that is used to create and manage persistent block storage from Hitachi Virtual Storage Platform One storages.


          For full documentation, go to our [Product Documentation site](https://docs.hitachivantara.com/home) and refer to the reference guide for HSPC v3.16.0.

          ## Requirements

          ### Supported Driver Version

          * HSPC v3.16.0
        displayName: Hitachi Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hitachi
        - Storage
        - Container Storage Interface
        - CSI
        links:
        - name: Product Documentation
          url: https://docs.hitachivantara.com/home
        maintainers:
        - email: Container-info@hitachivantara.com
          name: Hitachi
        maturity: stable
        provider:
          name: Hitachi
        relatedImages:
        - registry.hitachivantara.com/hitachicsi-oci-oss/hspc-operator@sha256:4dda4a43cefc19c8b1992756b499b1f9300920c5250dd77379f523997cfd5d54
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - registry.hitachivantara.com/hitachicsi-oci-oss/hspc-csi-driver@sha256:2c580d98be55fffd8f6083379b6f23cbfe2aeb42fc4aec9e890072bc831c87c4
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:81d32545fb1b2c319bfeb36a70db4b4d3aea9c51c0574b209b8f64750b32d2f0
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        version: 1.16.0
      entries:
      - name: hspc-operator.v1.16.0
        version: 1.16.0
      name: stable
    defaultChannel: stable
    packageName: hspc-operator
    provider:
      name: Hitachi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CyberArk
      provider-url: https://www.cyberark.com/
    name: conjur-follower-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: conjur-follower-operator.v2.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:75fb5de2c111084ffcc3d973e6c63f1adfae78e29f6fd02152419fbca553ce7b",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:bee2cd31b3620a31e98f9d21326a2cdcecb36479099b06cc9660e4e83354d940",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:383776c23556a0a327a0bacc843d22c4b1516082ddd0cfb8a8275500f2eb841a",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:a9b109c7f3bfba5553aa4366338fa211896564f8e33856d75b4918acf8429f1e",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:b3096c17e7f3b9d723c906619dde293ea084e78c0da48c534335a8eee07b23c7",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:680a2044bc0e94fbc9376cd1eed9d067de91657f566c9752654480c939447841",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:ed53aea1866b1e2262386ac38cb9e1e47cc3f9047ae47f49070e2f506af6ca5a"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0 <2.5.3'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:cf46bdb4238cdb1dc51329822da55b7836958be2344d26f9e1819c0aa97fece6
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:383776c23556a0a327a0bacc843d22c4b1516082ddd0cfb8a8275500f2eb841a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:75fb5de2c111084ffcc3d973e6c63f1adfae78e29f6fd02152419fbca553ce7b
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:bee2cd31b3620a31e98f9d21326a2cdcecb36479099b06cc9660e4e83354d940
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:a9b109c7f3bfba5553aa4366338fa211896564f8e33856d75b4918acf8429f1e
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:b3096c17e7f3b9d723c906619dde293ea084e78c0da48c534335a8eee07b23c7
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:680a2044bc0e94fbc9376cd1eed9d067de91657f566c9752654480c939447841
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:ed53aea1866b1e2262386ac38cb9e1e47cc3f9047ae47f49070e2f506af6ca5a
        version: 2.5.3
      entries:
      - name: conjur-follower-operator.v2.5.3
        version: 2.5.3
      name: "2.5"
    - currentCSV: conjur-follower-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:c5337633eb3cc272f989e3e70c2e9bace9c819bba9e817537c7a691a8b6ee596
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813
        version: 2.6.3
      entries:
      - name: conjur-follower-operator.v2.6.3
        version: 2.6.3
      - name: conjur-follower-operator.v2.6.2
        version: 2.6.2
      - name: conjur-follower-operator.v2.6.1
        version: 2.6.1
      - name: conjur-follower-operator.v2.6.0
        version: 2.6.0
      - name: conjur-follower-operator.v2.5.0
        version: 2.5.0
      - name: conjur-follower-operator.v2.4.4
        version: 2.4.4
      - name: conjur-follower-operator.v2.4.3
        version: 2.4.3
      - name: conjur-follower-operator.v2.4.2
        version: 2.4.2
      - name: conjur-follower-operator.v2.3.5
        version: 2.3.5
      - name: conjur-follower-operator.v2.3.4
        version: 2.3.4
      - name: conjur-follower-operator.v2.3.1-290
        version: 2.3.1
      - name: conjur-follower-operator.v2.3.0
        version: 2.3.0
      - name: conjur-follower-operator.v2.2.4
        version: 2.2.4
      - name: conjur-follower-operator.v2.2.2
        version: 2.2.2
      - name: conjur-follower-operator.v2.2.1
        version: 2.2.1
      name: preview
    - currentCSV: conjur-follower-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:c5337633eb3cc272f989e3e70c2e9bace9c819bba9e817537c7a691a8b6ee596
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f
        version: 2.6.3
      entries:
      - name: conjur-follower-operator.v2.6.3
        version: 2.6.3
      - name: conjur-follower-operator.v2.6.2
        version: 2.6.2
      - name: conjur-follower-operator.v2.6.1
        version: 2.6.1
      - name: conjur-follower-operator.v2.6.0
        version: 2.6.0
      - name: conjur-follower-operator.v2.5.0
        version: 2.5.0
      - name: conjur-follower-operator.v2.4.4
        version: 2.4.4
      - name: conjur-follower-operator.v2.4.3
        version: 2.4.3
      - name: conjur-follower-operator.v2.4.2
        version: 2.4.2
      - name: conjur-follower-operator.v2.3.5
        version: 2.3.5
      - name: conjur-follower-operator.v2.3.4
        version: 2.3.4
      - name: conjur-follower-operator.v2.3.1-290
        version: 2.3.1
      - name: conjur-follower-operator.v2.3.0
        version: 2.3.0
      - name: conjur-follower-operator.v2.2.4
        version: 2.2.4
      - name: conjur-follower-operator.v2.2.2
        version: 2.2.2
      - name: conjur-follower-operator.v2.2.1
        version: 2.2.1
      name: stable
    defaultChannel: stable
    packageName: conjur-follower-operator
    provider:
      name: CyberArk
      url: https://www.cyberark.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/file-integrity-operator
    name: file-integrity-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: file-integrity-operator.v0.1.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <0.1.13'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator
          url: https://file-integrity-operator.domain
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/openshift4/file-integrity-rhel8-operator@sha256:a65f1244636649734763164a5d6155a6f6a7b79b3950ecc03fb3c58188583599
        version: 0.1.13
      entries:
      - name: file-integrity-operator.v0.1.13
        version: 0.1.13
      name: "4.7"
    - currentCSV: file-integrity-operator.v0.1.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <0.1.32'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:c3e34e6198eadb48c007cd949d8be38e26ef7d15d2f9047a8c3c89c03201e4ea
        version: 0.1.32
      entries:
      - name: file-integrity-operator.v0.1.32
        version: 0.1.32
      name: release-0.1
    - currentCSV: file-integrity-operator.v1.3.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          createdAt: "2024-11-21T18:02:08Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.3.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:f7039c6aaaa3e60dc4a3965cc1341c5217837e4b4030eb943771967e0b206578
        version: 1.3.6
      entries:
      - name: file-integrity-operator.v1.3.6
        version: 1.3.6
      - name: file-integrity-operator.v1.3.5
        version: 1.3.5
      - name: file-integrity-operator.v1.3.4
        version: 1.3.4
      - name: file-integrity-operator.v1.3.3
        version: 1.3.3
      name: stable
    - currentCSV: file-integrity-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <1.0.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:dd1dfeccf59214a0de653c3a4c4dbbc8c41640954318d13386cda73891660e4f
        version: 1.0.0
      entries:
      - name: file-integrity-operator.v1.0.0
        version: 1.0.0
      name: v1
    defaultChannel: stable
    packageName: file-integrity-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/file-integrity-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: resource-locker-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: resource-locker-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "ResourceLocker",
                "metadata": {
                  "name": "combined"
                },
                "spec": {
                  "patches": [
                    {
                      "id": "patch1",
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 0).metadata.name }}: {{ (index . 1).metadata.name }}\n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "kind": "Namespace",
                          "name": "resource-locker-test"
                        },
                        {
                          "apiVersion": "v1",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "resource-locker-test"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "resource-locker-test"
                      }
                    },
                    {
                      "id": "patch2",
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 0) }}: {{ (index . 1) }}\n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "fieldPath": "$.metadata.uid",
                          "kind": "Namespace",
                          "name": "resource-locker-test"
                        },
                        {
                          "apiVersion": "v1",
                          "fieldPath": "$.metadata.uid",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "resource-locker-test"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "resource-locker-test"
                      }
                    }
                  ],
                  "resources": [
                    {
                      "object": {
                        "apiVersion": "networking.k8s.io/v1",
                        "kind": "NetworkPolicy",
                        "metadata": {
                          "name": "allow-from-same-namespace",
                          "namespace": "resource-locker-test"
                        },
                        "spec": {
                          "ingress": [
                            {
                              "from": [
                                {
                                  "podSelector": {}
                                }
                              ]
                            }
                          ],
                          "podSelector": {}
                        }
                      }
                    },
                    {
                      "object": {
                        "apiVersion": "networking.k8s.io/v1",
                        "kind": "NetworkPolicy",
                        "metadata": {
                          "name": "allow-from-default-namespace",
                          "namespace": "resource-locker-test"
                        },
                        "spec": {
                          "ingress": [
                            {
                              "from": [
                                {
                                  "namespaceSelector": {
                                    "matchLabels": {
                                      "name": "default"
                                    }
                                  }
                                }
                              ]
                            }
                          ],
                          "podSelector": {}
                        }
                      }
                    }
                  ],
                  "serviceAccountRef": {
                    "name": "default"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/resource-locker-operator@sha256:2176db02f561c24012c5179495c89c3819500dff21e89a62052d465eb3314b44
          createdAt: "2021-11-30T19:38:40Z"
          description: This operator provides a facility to lock resources and / or
            patches into the state described by its custom resource preventing any
            drifts.
          operatorframework.io/suggested-namespace: resource-locker-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/resource-locker-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ResourceLocker Represents the intetion to create and enforce
              a set of resources and/or patches.
            displayName: Resource Locker
            kind: ResourceLocker
            name: resourcelockers.redhatcop.redhat.io
            version: v1alpha1
        description: |
          The resource locker operator allows you to specify a set of configurations that the operator will "keep in place" (lock) preventing any drifts.
          Two types of configurations may be specified:
          * Resources. This will instruct the operator to create and enforce the specified resource. In this case the operator "owns" the created resources.
          * Patches to resources. This will instruct the operator to patch- and enforce the change on- a pre-existing resource. In this case the operator does not "own" the resource.
          Locked resources are defined with the `ResourceLocker` CRD. Here is the high-level structure of this CRD:
          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: ResourceLocker
          metadata:
            name: test-simple-resource
          spec:
            resources:
            - object:
                apiVersion: v1
            ...
            patches:
            - targetObjectRef:
              ...
              patchTemplate: |
                metadata:
                  annotations:
                    ciao: hello
            ...
            serviceAccountRef:
              name: default
          ```
          It contains:
          * `resources`: representing an array of resources
          * `patches`: representing an array of patches
          * `serviceAccountRef`: a reference to a service account defined in the same namespace as the ResourceLocker CR, that will be used to create the resources and apply the patches. If not specified the service account will be defaulted to: `default`
          For each ResourceLocker a manager is dynamically allocated. For each resource and patch a controller with the needed watches is created and associated with the previously created manager.
          **Note**: creating a `ResourceLocker` CR will not work without due preparation (of the service account in particular). Follow these [instructions](https://github.com/redhat-cop/resource-locker-operator/blob/master/EXAMPLES.md) for a step by step walkthrough of the needed prerequisites to make this operator work.
        displayName: Resource Locker Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - resource_management
        - enforcing
        - patches
        - gitops
        links:
        - name: Resource Locker Operator
          url: https://github.com/redhat-cop/resource-locker-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/resource-locker-operator:latest
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/resource-locker-operator@sha256:2176db02f561c24012c5179495c89c3819500dff21e89a62052d465eb3314b44
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.1.3
      entries:
      - name: resource-locker-operator.v1.1.3
        version: 1.1.3
      - name: resource-locker-operator.v1.1.2
        version: 1.1.2
      - name: resource-locker-operator.v1.1.1
        version: 1.1.1
      - name: resource-locker-operator.v1.1.0
        version: 1.1.0
      - name: resource-locker-operator.v1.0.9
        version: 1.0.9
      name: alpha
    defaultChannel: alpha
    packageName: resource-locker-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Jason Skrzypek
      provider-url: https://github.com/shpwrck
    name: openshift-node-upgrade-mutex-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-node-upgrade-mutex-operator.v0.2.66
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexRule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutexrule-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutexrule",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutexrule-sample"
                },
                "spec": {
                  "jobName": "example",
                  "jobNamespace": "node-upgrade-mutex-operator",
                  "jobSpec": {
                    "selector": {},
                    "template": {
                      "metadata": {
                        "name": "mutex"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              "20"
                            ],
                            "image": "ubi9/toolbox",
                            "name": "mutex"
                          }
                        ],
                        "restartPolicy": "Never"
                      }
                    }
                  },
                  "type": "kubernetes"
                }
              },
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexRunConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutexrunconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutexrunconfig",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutexrunconfig-sample"
                },
                "spec": {
                  "canary": {
                    "name": "canary-example"
                  },
                  "mutexRules": [
                    {
                      "name": "mutexrule-sample",
                      "namespace": "node-upgrade-mutex-operator"
                    }
                  ],
                  "mutexTargets": [
                    {
                      "name": "mutextarget-sample",
                      "namespace": "node-upgrade-mutex-operator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexTarget",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutextarget-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutextarget",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutextarget-sample"
                },
                "spec": {
                  "machineConfigPools": [
                    "worker"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/jskrzypek/openshift-node-upgrade-mutex-operator
          createdAt: "2025-05-30T17:49:00Z"
          description: |
            This software injects required operations into the worker node upgrade process.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: node-upgrade-mutex-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          provider: Jason Skrzypek
          repository: https://github.com/shpwrck/openshift-upgrade-rollout-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A node mutex rule definition in the form of a kubernetes
              job spec
            displayName: Mutex Rule
            kind: MutexRule
            name: mutexrules.openshift-optional.skrzypek.io
            version: v1alpha1
          - description: Lists of Targets and Rules to apply to them, as well as an
              upgrade facilitating Machine Config Pool
            displayName: Mutex Run Config
            kind: MutexRunConfig
            name: mutexrunconfigs.openshift-optional.skrzypek.io
            version: v1alpha1
          - description: A list of Machine Config Pools to enforce upgrade procedure.
            displayName: Mutex Target
            kind: MutexTarget
            name: mutextargets.openshift-optional.skrzypek.io
            version: v1alpha1
        description: |
          ## Introduction

          This operator injects required operations in the form of a Kubernetes job into the worker node upgrade process using MachineConfigPools. In short, when provided a list of operations (`MutexRules`) and a list of MachineConfigPools (`MutexTargets`) , the OpenShift Node Upgrade Mutex Operator will execute each operation for a given target, enable updates for the given target by modifying it's MachineConfigPool membership, and wait until the node has been upgraded before proceeding onto subsequent operations and targets.

          ## Dependencies

          ### Release Version Label

          In order to confirm the successful upgrade of each target node, the NFD Operator is used to generate a version feature on every worker node. In order to proceed through given targets, each node must have this label
          `'machineconfiguration.openshift.io/release-image-version'` with a key providing the semantic version of this node (e.g. `'4.16.4'`)

          There are several ways to accomplish this, but an example MachineConfig has been provided [here](https://github.com/shpwrck/openshift-node-upgrade-mutex-operator/blob/main/machineconfig.yaml).

          ### Canary Machine Config Pool

          In addition to node labels, a MachineConfigPool that will be used to accomodate targets subject to upgrades must be provided matching the following criteria:
          * selects only worker nodes
          * is not paused

          Applying the example below will create a compatible canary pool:

          ```yaml
          ---
          apiVersion: machineconfiguration.openshift.io/v1
          kind: MachineConfigPool
          metadata:
            name: canary
          spec:
            machineConfigSelector:
              matchLabels:
                machineconfiguration.openshift.io/role: worker
            nodeSelector:
              matchLabels:
                node-role.kubernetes.io/worker: ''
              matchExpressions:
                - key: 'isCanary'
                  operator: Exists
            paused: false
          ```

          ## Usage / Custom Resources Definitions

          #### MutexRule:

          A resource storing a single rule referring to a template for a Kubernetes job to be created.
          This resource **does not** modify existing resources.

          ```yaml
          ---
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexRule
          metadata:
            name: k8s-example
          spec:
            jobName: "example" # Jobs will be run with this prefix name
            jobNamespace: "node-upgrade-mutex-operator" # Jobs will be run in this namespace
            jobSpec: # Accepts full 'batch/v1' job spec
              selector: {}
              template:
                metadata:
                  name: mutex
                spec:
                  containers:
                  - command:
                    - sleep
                    args:
                    - "30"
                    image: ubi9/toolbox
                    name: mutex
                  restartPolicy: Never
            cleanUpJobSpec: # Accepts full 'batch/v1' job spec (Optional)
              selector: {}
              template:
                metadata:
                  name: mutex
                spec:
                  containers:
                  - command:
                    - sleep
                    args:
                    - "30"
                    image: ubi9/toolbox
                    name: mutex
                  restartPolicy: Never
            type: kubernetes
          ```

          #### MutexTarget:

          A resource storing a set of MachineConfigPools to pause and incrementally upgrade.
          This resource **does not** modify existing resources.

          ```yaml
          ---
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexTarget
          metadata:
            name: worker-example
          spec:
            machineConfigPools:
              - worker # Name of existing machineconfigpool
          ```

          #### MutexRunConfig:

          A resource that references existing MutexRules and MutexTargets as well as an existing MachineConfigPool that will allow members to be upgraded. Every rule defined in the `mutexRules` list will be added as a mandatory task to every node that is a member of the each MachineConfigPool defined in the `mutexTargets` list.
          This resource **does** modify existing resources.

          ```yaml
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexRunConfig
          metadata:
            name: example
          spec:
            skipTLSVerify: false # Defaults to false if omitted
            canary:
              name: canary # Name of unpaused machineconfigpool to modify
            mutexRules: # List of existing mutexRules
              - name: k8s-example
                namespace: node-upgrade-mutex-operator
            mutexTargets: # List of existing mutexTargets
              - name: worker-example
                namespace: node-upgrade-mutex-operator
          ```

          ## Procedure

          #### MutexRule

          1. Upon creation/modification, the schema of each K8s rule is validated.
          2. Status is updated to indicate valid or invalid configuration.

          #### MutexTargets

          1. Upon creation/modification, the existence of each target is validated.
          2. Status is updated to indicate valid or invalid configuration.
          3. The nodes assigned to each target are added to the status field `nodes`.

          #### MutexRunConfigs

          1. Upon creation/modification, the existence as well as validation status of each rule, target, and canary pool is confirmed.
          2. If valid, the existence of an upgrade is queried. An upgrade is available if the following is true:
              * The `status.history` of the `cluster` ClusterVersion equals `Completed`
              * The `release-image-version` of any worker node is less than the `spec.desiredUpdate.version` of the `cluster` ClusterVersion.
          3. If an update exists, each target MachineConfigPool has `spec.paused` set to `true`.
          4. If each pool is paused, each node is updated to include a label for each rule with the following format:
              * `mutexrule.openshift-optional.skrzypek.io/{{MutexRule_Name}}.{{MutexRule_Namespace}}`
          5. If labels have been applied, the provided canary pool is modified to exclude nodes with labels provided by step (4).
          6. Assuming steps (2-5) have completed successfully, then the process for running jobs and updating nodes begins. The process per node is as follows:
              * Launch each K8S job sequentially
              * Remove the current node's rule labels
              * Wait until the `release-image-version` of the node matches the `spec.desiredUpdate.version` of the `cluster` ClusterVersion
              * Launch each K8S cleanup job sequentially
          7. If the update succeeds, all machine config pools are unpaused and all match expressions are removed.

          ## Permissions Granted

          In addition to the permissions required to operate the included CRDs, the following access is granted to the operator's service account.

          | Resource | Permissions |
          | :---: | :---: |
          | Jobs | Create & Read |
          | Nodes | Read & Update |
          | ClusterVersion | Read |
          | MachineConfigPools | Read & Update |

          ## Notes

          * This operator was built using the operator-sdk's ansible plugin and therefore has all of the inherited capabilities and restrictions.
          * This operator has undergone minimal testing and should not be run in production environments without acknowledging the risks.
          * This operator is currently alpha and subject to change based on requirements.

          ## Additional Information
          For additional information refer to the [GitHub Repository](https://github.com/shpwrck/openshift-node-upgrade-mutex-operator)
        displayName: OpenShift Node Upgrade Mutex Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - OpenShift Optional
        - Upgrade
        - Mutex
        links:
        - name: Openshift Node Upgrade Mutex Operator
          url: https://github.com/shpwrck/openshift-node-upgrade-mutex-operator
        maintainers:
        - email: jskrzype@redhat.com
          name: Jason Skrzypek
        maturity: alpha
        minKubeVersion: 1.27.0
        provider:
          name: Jason Skrzypek
          url: https://github.com/shpwrck
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - quay.io/jskrzypek/openshift-node-upgrade-mutex-operator@sha256:d4c28e8b256c00429d9999595c2386f1a25de597ee8e6c33871bb7b22b84c4af
        version: 0.2.66
      entries:
      - name: openshift-node-upgrade-mutex-operator.v0.2.66
        version: 0.2.66
      - name: openshift-node-upgrade-mutex-operator.v0.2.65
        version: 0.2.65
      - name: openshift-node-upgrade-mutex-operator.v0.2.60
        version: 0.2.60
      - name: openshift-node-upgrade-mutex-operator.v0.2.59
        version: 0.2.59
      - name: openshift-node-upgrade-mutex-operator.v0.2.53
        version: 0.2.53
      - name: openshift-node-upgrade-mutex-operator.v0.2.38
        version: 0.2.38
      - name: openshift-node-upgrade-mutex-operator.v0.2.34
        version: 0.2.34
      - name: openshift-node-upgrade-mutex-operator.v0.2.15
        version: 0.2.15
      - name: openshift-node-upgrade-mutex-operator.v0.2.14
        version: 0.2.14
      - name: openshift-node-upgrade-mutex-operator.v0.1.22
        version: 0.1.22
      name: alpha
    defaultChannel: alpha
    packageName: openshift-node-upgrade-mutex-operator
    provider:
      name: Jason Skrzypek
      url: https://github.com/shpwrck
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Digital.ai
      provider-url: https://digital.ai
    name: digitalai-deploy-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: digitalai-deploy-operator.v25.3.0-beta.813
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "quorum",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.0-debian-12-r4"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.0-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.5.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.2-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabsunsupported/deploy-operator@sha256:9ccca8ab999d4b6058c979bbb70568a63cf43a257ae65a787b2ff4d9bf470c39 '
          createdAt: "2025-09-02T05:43:37Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabsunsupported/deploy-operator@sha256:9ccca8ab999d4b6058c979bbb70568a63cf43a257ae65a787b2ff4d9bf470c39
        - docker.io/xebialabsunsupported/xl-deploy@sha256:41feb1d24d70a0c78a560b338d05c96c5383d570e3ef1ff37daff0eb8c92b386
        - docker.io/xebialabsunsupported/deploy-task-engine@sha256:15322f2263c2b9b70336177aec170e59497fa320281c2d73605918b30272b8f1
        - docker.io/xebialabsunsupported/central-configuration@sha256:3a35c2aada71a7f5ab8fe7d5c68441fe58b0b2c778b6ea85ac6f3f7b351f80ad
        version: 25.3.0-beta.813
      entries:
      - name: digitalai-deploy-operator.v25.3.0-beta.813
        version: 25.3.0-beta.813
      - name: digitalai-deploy-operator.v25.3.0-beta.711
        version: 25.3.0-beta.711
      - name: digitalai-deploy-operator.v25.3.0-beta.609
        version: 25.3.0-beta.609
      - name: digitalai-deploy-operator.v25.3.0-beta.512
        version: 25.3.0-beta.512
      name: beta
    - currentCSV: digitalai-deploy-operator.v24.1.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "ingressClass": "haproxy-daid",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "tag": "1.26.0-debian-12-r1"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "tag": "1.10.1-debian-12-r5"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "tag": "15.6.0-debian-12-r7"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "tag": "3.12.13-debian-12-r2"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator without SCC creation and with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:2b3413cd41f2b437e6027c8ae2cff12d1113e0ac1be69fc24840d68a05d3a665 '
          createdAt: "2024-10-16T05:39:21Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind:
          DigitalaiDeploy\nmetadata:\n  name: daid-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  securityContextConstraints:\n
          \   enabled: false\n  centralConfiguration:\n    replicaCount: 1\n    podSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  master:\n    replicaCount: 1\n    persistence:\n
          \     storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n      runAsUser:
          null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  worker:\n    replicaCount: 1\n    persistence:\n
          \     storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n      runAsUser:
          null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  route:\n    enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n        runAsUser: null\n        runAsGroup: null\n
          \       fsGroup: null\n      containerSecurityContext:\n        enabled:
          false\n        runAsUser: null\n        runAsGroup: null\n      securityContextConstraints:\n
          \       enabled: false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n
          \   install: true\n    persistence:\n      size: 1Gi\n      storageClass:
          ''\n    replicaCount: 1\n    podSecurityContext:\n      enabled: false\n
          \     runAsUser: null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     enabled: false\n      runAsUser: null\n      runAsGroup: null\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"daid-doc-digitalai-deploy-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xld-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Embedded PostgreSQL for
          DB management\n- Embedded RabbitMQ for message queue management\n\n### Security
          Configuration Details\n\n- `securityContextConstraints.enabled: false` -
          disables creation of SCCs;\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/deploy-operator@sha256:2b3413cd41f2b437e6027c8ae2cff12d1113e0ac1be69fc24840d68a05d3a665
        - docker.io/xebialabs/deploy-task-engine@sha256:8a381fd84483ea919fa7f8bee5cf19d9dec0d1d61cce73ec6b7c56fa504c6c93
        - docker.io/xebialabs/xl-deploy@sha256:77fdf6a3909367b060eba2ef698baddb77e4086d4d977f9c617d1e26a4427472
        - docker.io/xebialabs/central-configuration@sha256:83359e49c66e4fab502cc75663b456eb37d81d7cb512f2c48667f4d6a324e0df
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:771a9a173e033a3ad8b46f5c00a7036eaa88c8d8d1fbd89217325168998113ea
        version: 24.1.10
      entries:
      - name: digitalai-deploy-operator.v24.1.10
        version: 24.1.10
      name: stable
    - currentCSV: digitalai-deploy-operator.v24.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "image": {
                        "tag": "v0.14.7"
                      },
                      "ingressClass": "haproxy-daid",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.26.2-debian-12-r10"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "15.6.0-debian-12-r20"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.12.14-debian-12-r7"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator without SCC creation and with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:7d505a855728c761433ef7d46f0a64ffe2b4192f4819bad86c23582375099ac9 '
          createdAt: "2025-08-25T04:53:23Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind:
          DigitalaiDeploy\nmetadata:\n  name: daid-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  securityContextConstraints:\n
          \   enabled: false\n  centralConfiguration:\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  master:\n    replicaCount:
          1\n    persistence:\n      storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  worker:\n    replicaCount:
          1\n    persistence:\n      storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  route:\n    enabled: false\n
          \   annotations:\n      haproxy.router.openshift.io/cookie_name: SESSION_XLD\n
          \     haproxy.router.openshift.io/disable_cookies: \"false\"\n      haproxy.router.openshift.io/rewrite-target:
          /\n    hostname: '<mandatory-deploy-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n      securityContextConstraints:\n
          \       enabled: false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n
          \   install: true\n    persistence:\n      size: 1Gi\n      storageClass:
          ''\n    replicaCount: 1\n    podSecurityContext:\n      enabled: false\n
          \   containerSecurityContext:\n      enabled: false\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"daid-doc-digitalai-deploy-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xld-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Embedded PostgreSQL for
          DB management\n- Embedded RabbitMQ for message queue management\n\n### Security
          Configuration Details\n\n- `securityContextConstraints.enabled: false` -
          disables creation of SCCs;\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/xl-deploy@sha256:c2fe209f198fb2f42dc78a2e1544c026a6891d60682b4e47634234bbf7346540
        - docker.io/xebialabs/central-configuration@sha256:581a0bc71ae3652ceb643d81927fcf8cabea92f1d301f5d354076baff32fbcde
        - docker.io/xebialabs/deploy-operator@sha256:7d505a855728c761433ef7d46f0a64ffe2b4192f4819bad86c23582375099ac9
        - docker.io/xebialabs/deploy-task-engine@sha256:a97d67ea53e1b954ffb4d9514fd61426dcc0cab08922009eef02f32162026d6a
        version: 24.1.21
      entries:
      - name: digitalai-deploy-operator.v24.1.21
        version: 24.1.21
      - name: digitalai-deploy-operator.v24.1.20
        version: 24.1.20
      - name: digitalai-deploy-operator.v24.1.19
        version: 24.1.19
      - name: digitalai-deploy-operator.v24.1.18
        version: 24.1.18
      - name: digitalai-deploy-operator.v24.1.17
        version: 24.1.17
      - name: digitalai-deploy-operator.v24.1.14
        version: 24.1.14
      - name: digitalai-deploy-operator.v24.1.10
        version: 24.1.10
      name: stable-v24.1
    - currentCSV: digitalai-deploy-operator.v24.3.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "16.3.0-debian-12-r23"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n{{ if eq .Values.queueType \"quorum\" -}}\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n{{- end }}\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.13.7-debian-12-r5"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "queueType": "classic",
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:41ffc26c1d99a808e537f2366d92473ecf76dbffbac699dfa04785c3016f7470 '
          createdAt: "2025-09-26T18:18:41Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/central-configuration@sha256:4b83d4e9a523ae3615579907c65790478a3b22315ca5afce594793424d1d33b1
        - docker.io/xebialabs/deploy-operator@sha256:41ffc26c1d99a808e537f2366d92473ecf76dbffbac699dfa04785c3016f7470
        - docker.io/xebialabs/xl-deploy@sha256:8536efdf530dde45b84b14191ebfdc7ec5ebb1844f84871032fe2d335b13908d
        - docker.io/xebialabs/deploy-task-engine@sha256:75fe942305bf3bb403eb243b157b88844363eb702f03227a61e76ad4321bf0ee
        version: 24.3.11
      entries:
      - name: digitalai-deploy-operator.v24.3.11
        version: 24.3.11
      - name: digitalai-deploy-operator.v24.3.10
        version: 24.3.10
      - name: digitalai-deploy-operator.v24.3.9
        version: 24.3.9
      - name: digitalai-deploy-operator.v24.3.8
        version: 24.3.8
      - name: digitalai-deploy-operator.v24.3.7
        version: 24.3.7
      - name: digitalai-deploy-operator.v24.3.6
        version: 24.3.6
      - name: digitalai-deploy-operator.v24.3.5
        version: 24.3.5
      name: stable-v24.3
    - currentCSV: digitalai-deploy-operator.v25.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "quorum",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.1-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.4.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.3-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:d192ef9008d72c9792e1b7b988bdae8f1bbce45ed0e44afc48e3e268b657b38b '
          createdAt: "2025-09-26T16:30:53Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/deploy-task-engine@sha256:13bb779d6041836acb95c95f1eed75f4b5b2940d75687306183a9f8a2177cab3
        - docker.io/xebialabs/central-configuration@sha256:0c46275ffa3b251ccb76d2623e0e5a4cd1441bf7f94901845bda743f087bc6e4
        - docker.io/xebialabs/xl-deploy@sha256:e4212d20fad19f09cbe53796371bec7fc368cfe63316f1b7bccecf5faf6ea733
        - docker.io/xebialabs/deploy-operator@sha256:d192ef9008d72c9792e1b7b988bdae8f1bbce45ed0e44afc48e3e268b657b38b
        version: 25.1.5
      entries:
      - name: digitalai-deploy-operator.v25.1.5
        version: 25.1.5
      - name: digitalai-deploy-operator.v25.1.4
        version: 25.1.4
      - name: digitalai-deploy-operator.v25.1.3
        version: 25.1.3
      - name: digitalai-deploy-operator.v25.1.2
        version: 25.1.2
      - name: digitalai-deploy-operator.v25.1.1
        version: 25.1.1
      - name: digitalai-deploy-operator.v25.1.0
        version: 25.1.0
      name: stable-v25.1
    defaultChannel: stable-v24.3
    packageName: digitalai-deploy-operator
    provider:
      name: Digital.ai
      url: https://digital.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://access.redhat.com/products/red-hat-fuse
    name: fuse-online
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-online.v7.10.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:ed4faedcb97bb6a45f568586ff19da19da31ecbc304f12e0ae7978381247744a
          createdAt: 2022-01-31 15:55:46 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:ed4faedcb97bb6a45f568586ff19da19da31ecbc304f12e0ae7978381247744a
        - registry.redhat.io/rhel8/postgresql-12@sha256:44215c5b244b190c82d9d371a8ffd39b0c83576bd85c980cc72782a5cf9e6e1b
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:5048cd4fab2f98e6000bc60835c539cc40b70dbe265e927cfe185595347d8054
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:5dd1bc8f6b7e724736f4aeac86409bda3e0f64f62b736688e9572d14411232f1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:fcdd0292ebfc7e9edc36241f63b54a9d924a7605329c934ff2b10373af26dbbf
        - registry.redhat.io/openshift4/ose-prometheus@sha256:7f679d6c03535e300dac366f4bc6dd8d781d96dc47592836713f8bb76be36fa9
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:4f8849362e9ed6ff95d318a04e287b7ba33254a2cf7abdaf9b90cab7cbba6646
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:df2d55dd8092e9e94d034166890117cf5d1cfd5cf52ef0f90cfa2374cc9e2d06
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:3f0277e2c278740307c9b2d966f539ba12317e306198cb6c4c1c52a2ad9f4f40
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:96810c770f75057b8bba6b78a5ba414c8289315d8af1a387f681ae5e4fb7d6dd
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:ef229a4c8851ec495f28d0eeab83b137cd02698e283f76ea1b32271b7222621b
        version: 7.10.4
      entries:
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.10.x
    - currentCSV: fuse-online.v7.11.1-0.1687357787.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:658af3fc7819efb7cc39cd3a0504dcddc2ce7788c9b1e5178df869f7c9a742a8
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.11.1-0.1684718286.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:a2b1cd0256f30aff8fa37c332419b4483c0bfe34fb4f574ad8f18b0858969342
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:4de91e4527022e1b25df8d3835ddf778f773c20334314be9ba01a885c82287f5
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:8b9bd777bcbd20b1a1bbbc18274a5076a67bea1f928d772cb243c57fb1da7408
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:ac514973ec428d815d7a248d09d27f2a84f762aedfea63b3b48de9e633267ed5
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:658af3fc7819efb7cc39cd3a0504dcddc2ce7788c9b1e5178df869f7c9a742a8
        - registry.redhat.io/rhel8/postgresql-12@sha256:9248c4eaa8aeedacc1c06d7e3141ca1457147eef59e329273eb78e32fcd27e79
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:251c29e81f8e338f980b475b841647a977807908c388f4c86e18da7bff1b198a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:400a6ee03c3200c19b9b15c5d037a849633156a62e13922fb2e62838024ae69c
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:3d98d775bec85704c9a10c2f0d8e4f8116b7a41494f73c0b763e010262e5cd4d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        version: 7.11.1+0.1687357787.p
      entries:
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.11.x
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.12.x
    - currentCSV: fuse-online.v7.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta2",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:c70c5e76dc3db559c722bea2a767eb7dcba1eda786e1ef3a45c4cf32e1aca9ee
          createdAt: 2021-07-21 12:25:15 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta2
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to the product documentation at https://access.redhat.com/documentation/en-us/red_hat_fuse
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:d6df09b15599e776b49e1d46aa01728fd02d5629cee65e77dc277c0166554e0a
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:c668fe1f5d25baddf627acd79df4c8d2ab8b05309d3d41784ddc0a10dd5f34ad
        - registry.redhat.io/openshift4/ose-prometheus@sha256:13e9ea359cbcd3e0e94a6589b86b708dc84197f152701fd4092fffa234d11b58
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:60d5cf00411b03af3f1149a2bb3aeeb196c063d19625f876a094ae6baf46b54b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:f65600506db985689e5b7cd8b838e646d4268a0c4e272e19d7db3577a8dd7ac4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:75d5772c792a7c795281acec3d59e90021c109e9d7d1110868be0b882768aefb
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:c70c5e76dc3db559c722bea2a767eb7dcba1eda786e1ef3a45c4cf32e1aca9ee
        - registry.redhat.io/rhscl/postgresql-12-rhel7@sha256:c3e95e808eecba8ac57632452d3553509131ca783f62c68754a582cf5c55efab
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:eff9f8edf28ccbadcacab2022fff9e6bc93926af97521b49e3968eaf7b6fda65
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:d6f0c804c179fa559d3baed2e2d106c916a0439dc2215c4d34b878bc2e9bc281
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bafd85b0fd1256b8a60a2bb368a7e68f5dbad72718e5059a65660e01a425d8e2
        version: 7.9.3
      entries:
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.9.x
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: candidate
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: latest
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: stable
    defaultChannel: latest
    packageName: fuse-online
    provider:
      name: Red Hat
      url: https://access.redhat.com/products/red-hat-fuse
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: eG Innovations
      provider-url: https://www.eginnovations.com
    name: eginnovations-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eginnovations-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eginnovations.com/v1beta1",
                "kind": "EgUniversalAgent",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eginnovations-operator"
                  },
                  "name": "eguniversalagent",
                  "namespace": "egagent"
                },
                "spec": {
                  "agent": {
                    "env": [
                      {
                        "name": "EG_MANAGER",
                        "value": "Replace with eG Manager IP"
                      },
                      {
                        "name": "EG_MANAGER_PORT",
                        "value": "Replace with eG Manager Port"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "true"
          containerImage: docker.io/eginnovations/universal-agent-operator:0.0.11
          createdAt: "2025-09-11T20:08:00Z"
          description: eG Innovations provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with eG Universal Agent Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eginnovations/universal-agent-operator
          support: eG Innovations
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgUniversalAgent is the Schema for the eguniversalagents
              API
            displayName: Eg Universal Agent
            kind: EgUniversalAgent
            name: eguniversalagents.eginnovations.com
            version: v1beta1
        description: |
          eG Innovations’ converged application and infrastructure performance monitoring capabilities provide end-to-end visibility and correlated analytics for every layer, every tier of your IT landscape.

          The eG Enterprise Universal Agent Operator automatically configures a host agent on every Kubernetes worker node. The host agent auto-discovers the worker nodes and application containers running as Pods on each node and tracks their performance and utilization levels. In-depth monitoring of applications running on containers is also provided using the same host agent. No additional agents are required for the containers.

          ### Capabilities
            * Rolls out the egagent pod per node to monitor its pods and the node itself
            * Automatic BTM profiler injection using webhook based injection for deep application monitoring and transection tracing.

          ### Installation
            1. Create the namespace "egagent" for installing eG Agent on OpenShift Cluster.

               <br />``` $ oc create ns egagent ```
            2. Add "egagent" serviceaccount to scc( Security Context Constraints ) of OpenShift Cluster as privileged using below command.

               <br />``` $ oc adm policy add-scc-to-user privileged -z egagent -n egagent ```
            2. Install eG Innovations Universal Agent operator from [OpenShift Container Platform](https://www.openshift.com/) OperatorHub.
            3. Create the eG Universal Agent CRD (Custom Resource Definition)
               a. Click "View YAML Example" and copy the yaml contents and save it to (filename).yaml on your host.
               b. To apply the CRD run the following command

               <br />``` $ oc apply -f (filename).yaml ```

          ### Required Parameters
            * EG\_MANAGER, EG\_MANAGER\_PORT - provide the ip and port of your eG Manager Installation.

          ### Advanced Options
            * Disable Certificate Checking - disable any certificate validation that may interact poorly with proxies within your cluster
            * Image Override - use a copy of the eG Iniversal Agent container image from a registry other that Docker's or Red Hat's
            * Environment variables - define environment variables for the eG Universal Agent container of for APM profiler.
            * Namespace Selectors- select a subset of namesapces to enable instrumentation for applications.

          For complete list of supported parameters and the detailed documentation available in [eG Innovations Docs](https://docs.eginnovations.com)
        displayName: eG Innovations Universal Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Tracing
        links:
        - name: eG Innovations Kubernetes Monitoring Overview
          url: https://www.eginnovations.com/documentation/Kubernetes/Introduction-to-Monitoring-Kubernetes.htm
        maintainers:
        - email: devops@eginnovations.com
          name: eG Innovations
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: eG Innovations
          url: https://www.eginnovations.com
        relatedImages:
        - docker.io/eginnovations/universal-agent-operator:0.0.11
        version: 0.0.11
      entries:
      - name: eginnovations-operator.v0.0.11
        version: 0.0.11
      - name: eginnovations-operator.v0.0.10
        version: 0.0.10
      - name: eginnovations-operator.v0.0.9
        version: 0.0.9
      - name: eginnovations-operator.v0.0.8
        version: 0.0.8
      name: beta
    defaultChannel: beta
    packageName: eginnovations-operator
    provider:
      name: eG Innovations
      url: https://www.eginnovations.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: APIMatic.io
      provider-url: https://www.apimatic.io/
    name: apimatic-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apimatic-kubernetes-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicodegen.apimatic.io/v1beta1",
                "kind": "APIMatic",
                "metadata": {
                  "name": "apimatic-sample"
                },
                "spec": {
                  "licensespec": {
                    "licenseSourceName": "apimaticlicense",
                    "licenseSourceType": "ConfigMap"
                  },
                  "podspec": {
                    "apimaticContainerSpec": {
                      "image": "apimaticio/apimatic-codegen",
                      "imagePullSecret": "apimaticimagesecret"
                    }
                  },
                  "replicas": 3,
                  "servicespec": {
                    "apimaticserviceport": {
                      "nodePort": 32000,
                      "port": 8070
                    },
                    "servicetype": "NodePort"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: docker.io/apimaticio/apimatic-kubernetes-operator:v0.0.1
          createdAt: "2021-08-24 12:00:00"
          description: Generate client SDKs and interactive Documentation for your
            APIs in minutes
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/apimatic/apimatic-kubernetes-operator
          support: APIMatic.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIMatic is the Schema for the apimatics API
            displayName: APIMatic App
            kind: APIMatic
            name: apimatics.apicodegen.apimatic.io
            version: v1beta1
        description: "##About APIMatic \n\n APIMatic is a Developer Experience   platform
          for Web APIs that empowers users to create multi-language SDKs with API
          reference docs, dynamic code samples, assisted authentication and more -
          all auto-generated to provide a comprehensive developer experience beyond
          any API documentation tool.\n\n## Table of contents\n\n* [Introduction](#introduction)\n*
          [Features](#features)\n* [Running the Sample](#running-the-sample)\n  *
          [Prerequisites](#prerequisites)\n  * [Steps for Direct Deployment](#steps-for-direct-deployment)\n
          \ * [Steps for OLM Deployment](#steps-for-olm-deployment)\n* [Technical
          Support](#technical-support)\n* [Copyrights](#copyrights)\n\n## Introduction\n\nAPIMatic
          Operator simplifies the configuration and lifecycle management of the APIMatic
          code and docs generation solution on different Kubernetes distributions
          and OpenShift. The Operator encapsulates key operational knowledge on how
          to configure and upgrade the APIMatic CodeGen application, making it easy
          to get up and running.\n\n\nMore information about the underlying APIMatic
          CodeGen API that is exposed\nby this operator can be found [here](https://apimatic-core-v3-docs.netlify.app/#/http/getting-started/overview-apimatic-core).\n\n##
          Features\n\nAPIMatic Operator provides the following features:\n- Deploys
          the APIMatic CodeGen Web API service within the Kubernetes or OpenShift
          cluster.\n- Exposes the APIMatic CodeGen API external to the cluster, using
          Service type as [NodePort](https://kubernetes.io/docs/concepts/services-networking/service/#nodeport),
          [LoadBalancer](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          or [ExternalName](https://kubernetes.io/docs/concepts/services-networking/service/#externalname).\n-
          For exposing the service through an ingress resource, create an Ingress
          resource in the namespace of your APIMatic CR and set owned APIMatic service
          created by the operator as a backed service. More information can be found
          [here](https://kubernetes.io/docs/concepts/services-networking/ingress/).\n-
          Manual horizontal scaling of pods.\n  ```sh\n  kubectl scale apm apimatic-sample--replicas=2\n
          \ ```\n\n## Running the Sample \n\n### Prerequisites\n\nPlease contact APIMatic
          at [support@apimatic.io](mailto:support@apimatic.io) to register with the
          APIMatic CodeGen Docker Hub registry and acquire a valid license to run
          the APIMatic CodeGen API.\n\nFurther prerequisites for running the sample
          include:\n\n- [go](https://golang.org/) v1.16.*\n- [git](https://git-scm.com/)\n-
          [make](https://www.gnu.org/software/make/)\n- [Operator SDK](https://sdk.operatorframework.io/docs/overview/)\n-
          A running Kubernetes cluster with [kubectl](https://kubernetes.io/docs/reference/kubectl/overview/)
          on client. For testing purposes, you can use [Minikube](https://minikube.sigs.k8s.io/docs/)
          or [kind](https://kind.sigs.k8s.io/)\n- For checking the service created
          by the APIMatic operator on-prem, you can use [MetalLB](https://metallb.org/)\n\n###
          Steps for Direct Deployment\n\nTo run the sample for checking the APIMatic
          operator:\n\n- Clone the APIMatic repository into your working directory
          using the following command:\n  ```sh\n  git clone https://github.com/apimatic/apimatic-kubernetes-operator.git\n
          \ ```\n- Run `make deploy` to set up the APIMatic operator resources. This
          will deploy the '`apimatic-system`' namespace as well as the CRD and the
          RBAC manifests.\n\n- Create a secret named '`apimaticimagesecret`' to allow
          pulling the APIMatic CodeGen image using the Docker Hub username registered
          with APIMatic. If you have not done so, please contact support@apimatic.io
          for the steps required.\n  ```sh\n  kubectl create secret docker-registry
          apimaticimagesecret --docker-server=https://index.docker.io/v1/ --docker-username=<your-name>
          --docker-password=<your-pword> --docker-email=<your-email>\n  ```\n- Create
          a [configmap](https://kubernetes.io/docs/concepts/configuration/configmap/)
          resource named '`apimaticlicense`' that will serve as the volume storing
          the APIMatic license. An example of this is given below which will create
          the ConfigMap using the *License.lic* file located at */usr/local/apimatic/license/*\n
          \ ```sh\n  kubectl create configmap apimaticlicense --from-file /usr/local/apimatic/license/License.lic\n
          \ ```\n- This will deploy a ConfigMap resource with the following definition:\n
          \ ```sh\n  apiVersion: v1\n  data:\n    License.lic: \\\"<License file contents
          here>\\\"\n  kind: ConfigMap\n  metadata:\n    name: apimaticlicense\n    namespace:
          default\n  ```\n- Now run the sample using the following command:\n  ```sh\n
          \ kubectl apply -f config/samples/apicodegen_v1beta1_apimatic.yaml\n  ```\n-
          You will now see a new [Deployment](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/)
          with replica count of 3 and [Service](https://kubernetes.io/docs/concepts/services-networking/service/)
          of type `NodePort` created, both named ***apimatic-sample***. Accessing
          http://localhost:32000 from your browser (or using curl from within the
          Minikube or Kind cluster, if using that) should now show the APIMatic Titan
          page.\n- You can now use the exposed APIMatic CodeGen API to generate API
          SDKs and Docs using [curl](https://curl.se/), [Postman](https://www.postman.com/)
          or your own custom Web application that consumes the APIMatic CodeGen API
          service.\n\n- Once done, you can remove the APIMatic resources using the
          following command:\n  ```sh\n  make undeploy\n  ```\n\n### Steps for OLM
          Deployment\n\nThe following steps can be used to utilize [Operator LifeCycle
          Manager (OLM)](https://olm.operatorframework.io/docs/) to deploy the operator
          and run the sample. The steps are as follows:\n\n- If not already done so,
          clone the APIMatic repository into your working directory:\n  ```sh\n  git
          clone https://github.com/apimatic/apimatic-kubernetes-operator.git\n  ```
          \n- [Install OLM in your Kubernetes cluster](https://olm.operatorframework.io/docs/getting-started/#installing-olm-in-your-cluster).\n\n-
          Run the following script to install the resources required by OLM to deploy
          the APIMatic operator in the Kubernetes cluster within the `apimatic-system`
          namespace. Information about the different resources required can be found
          using the steps given [here](https://olm.operatorframework.io/docs/tasks/).\n
          \ ```sh\n  kubectl apply -f olm/manifests.yaml\n  ```\n\n- This should spin
          up the `ClusterServiceVersion` of the operator in the `apimatic-system`
          namespace, following which the operator pod will spin up. To ensure the
          that the operator installed successfully, check for the `ClusterServiceVersion`
          and the operator deployment in the namespace it was installed in.\n  ```sh\n
          \ kubectl get csv -n apimatic-system\n  \n  kubectl get deployment -n apimatic-system\n
          \ ```\n\n  - Create a secret named `apimaticimagesecret` to allow pulling
          the APIMatic CodeGen image using the Docker Hub username registered with
          APIMatic. If you have not done so, please contact support@apimatic.io for
          the steps required.\n  ```sh\n  kubectl create secret docker-registry apimaticimagesecret
          --docker-server=https://index.docker.io/v1/ --docker-username=<your-name>
          --docker-password=<your-pword> --docker-email=<your-email>\n  ```\n- Create
          a [configmap](https://kubernetes.io/docs/concepts/configuration/configmap/)
          resource named `apimaticlicense` that will serve as the volume storing the
          APIMatic license. An example of this is given below which will create the
          ConfigMap using the *License.lic* file located at */usr/local/apimatic/license/*\n
          \ ```sh\n  kubectl create configmap apimaticlicense --from-file /usr/local/apimatic/license/License.lic\n
          \ ```\n- This will deploy a ConfigMap resource with the following definition:\n
          \ ```sh\n  apiVersion: v1\n  data:\n    License.lic: \\\"<License file contents
          here>\\\"\n  kind: ConfigMap\n  metadata:\n    name: apimaticlicense\n    namespace:
          default\n  ```\n- Now run the sample using the following command:\n  ```sh\n
          \ kubectl apply -f config/samples/apicodegen_v1beta1_apimatic.yaml\n  ```\n\n-
          You will now see a new [Deployment](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/)
          with replica count of 3 and [Service](https://kubernetes.io/docs/concepts/services-networking/service/)
          of type `NodePort` created, both named ***apimatic-sample***. Accessing
          http://localhost:32000 from your browser (or using curl from within the
          Minikube or Kind cluster, if using that) should now show the APIMatic Titan
          page.\n- You can now use the exposed APIMatic CodeGen API to generate API
          SDKs and Docs using [curl](https://curl.se/), [Postman](https://www.postman.com/)
          or your own custom Web application that consumes the APIMatic CodeGen API
          service.\n- Once done, you can remove the APIMatic operator resources using
          the follow script:\n  ```sh\n  kubectl delete -f olm/manifests.yaml\n  ```\n\n##
          Technical Support\n\n- To request additional features in the future, or
          if you notice any discrepancy regarding this document, please drop an email
          to [support@apimatic.io](mailto:support@apimatic.io).\n\n### Copyrights\n\n&copy;
          2021 APIMatic.io\n"
        displayName: APIMatic Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Rest APIs
        - API SDK Generator
        - API Docs Generator
        - API Developer Portal
        - APIMatic
        - Developer Tools
        links:
        - name: Overview
          url: https://www.apimatic.io/
        - name: Developer Experience Portal
          url: https://www.apimatic.io/developer-experience-portal
        - name: Code-Gen as a Service
          url: https://www.apimatic.io/continuous-code-generation
        - name: API Transformer
          url: https://www.apimatic.io/transformer
        - name: Contact
          url: https://www.apimatic.io/contact
        - name: Documentation
          url: https://docs.apimatic.io/
        - name: Blogs
          url: https://www.apimatic.io/blog/
        maintainers:
        - email: support@apimatic.io
          name: Support
        maturity: alpha
        provider:
          name: APIMatic.io
          url: https://www.apimatic.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - docker.io/apimaticio/apimatic-kubernetes-operator:v0.0.1
        version: 0.0.1
      entries:
      - name: apimatic-kubernetes-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: apimatic-kubernetes-operator
    provider:
      name: APIMatic.io
      url: https://www.apimatic.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-openshift-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.14
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-06-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:154f85cc6721a70909dee755ddd88897ad51038afe2538b6dee4d26e7d8b9afb
        version: 0.0.14
      entries:
      - name: joget-dx8-openshift-operator.v0.0.14
        version: 0.0.14
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-openshift-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sumo Logic
      provider-url: https://www.sumologic.com/
    name: sumologic-kubernetes-collection-helm-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "helm-operator.sumologic.com/v1alpha1",
                    "kind": "SumologicCollection",
                    "metadata": {
                        "name": "collection"
                    },
                    "spec": {
                        "sumologic": {
                            "accessId": "",
                            "accessKey": "",
                            "clusterName": "kubernetes",
                            "scc": {
                                "create": true
                            }
                        },
                        "kube-prometheus-stack": {
                            "prometheus-node-exporter": {
                                "service": {
                                    "port": 9200,
                                    "targetPort": 9200
                                }
                            }
                        },
                        "otellogs": {
                            "daemonset": {
                                "containers": {
                                    "otelcol": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                },
                                "initContainers": {
                                    "changeowner": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                }
                            }
                        }
                    }
                }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing,Monitoring
          containerImage: registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
          createdAt: "2024-07-19T10:48:00Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/sumologic-kubernetes-collection-helm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/sumologic-kubernetes-collection-helm-operator-rhmp/support?utm_source=openshift_console
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator.git
          support: Sumo Logic
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configuration for Sumo Logic Kubernetes Collection
            displayName: SumologicCollection
            kind: SumologicCollection
            name: sumologiccollections.helm-operator.sumologic.com
            version: v1alpha1
        description: Helm Operator for the Sumo Logic Kubernetes Collection Helm Chart
          v4.13.0-2
        displayName: Sumo Logic Kubernetes Collection Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - collection
        - logging
        - metrics
        - monitoring
        links:
        - name: Sumo Logic Kubernetes Collection Helm Operator
          url: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator
        maintainers:
        - email: collection@sumologic.com
          name: Sumo Logic
        maturity: alpha
        provider:
          name: Sumo Logic
          url: https://www.sumologic.com/
        relatedImages:
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
        - registry.connect.redhat.com/sumologic/kube-state-metrics@sha256:3551ededa327a35db798ec8286cd843580e3ae395d2231e041e8c4b7621127ee
        - registry.connect.redhat.com/sumologic/metrics-server@sha256:65d2a3ddc4b4183f438fa18b655079c93c905cb9e3b8c54977336205dc1fff07
        - registry.connect.redhat.com/sumologic/prometheus@sha256:0ef6eaf570a40499dd5de92a79593e950bc771b014e0ee6ff7419b4b82a36fcc
        - registry.connect.redhat.com/sumologic/autoinstrumentation-python@sha256:f0fa45fa45e64c7df88f21181ae18ee96fe83cd3988258663d0806ea6f39e6bb
        - registry.connect.redhat.com/sumologic/kubernetes-setup@sha256:f07eaf68fa11ffc708e4485a83bd3763fc5712f7d73d29edfe5b9c9e08f05df6
        - registry.connect.redhat.com/sumologic/sumologic-otel-collector@sha256:4554d43dba204e3ec8e91b36442a438f10234fa0cb4db7fb7880b257a6485112
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:e1f49fcf5b7e86d9eb49c16dbbf9484cb486f55f999fd51d7c738c3b1f2666cf
        - registry.connect.redhat.com/sumologic/autoinstrumentation-dotnet@sha256:f4f385c4489c74b5c43f2fe6b6620a9eb8d9ecab470567236b3c45f636106640
        - registry.connect.redhat.com/sumologic/tailing-sidecar@sha256:738d653a78dc415243ab3627db9135768175fb7f51d0803566f5d9e824c8744b
        - registry.connect.redhat.com/sumologic/telegraf-operator@sha256:ff958222be01c0c3966474463e43bbd5bcf348790dd34bb761718cb1cea2d793
        - registry.connect.redhat.com/sumologic/autoinstrumentation-nodejs@sha256:c206eef4fdecaa58b4f01e5e7f0dbd74e9606eab2b4debf47461d36191e248d1
        - registry.connect.redhat.com/sumologic/node-exporter@sha256:11e9e604432a66d9b388cb1bf70eac077c7f0372d1264342ef927d76b8ef0272
        - registry.connect.redhat.com/sumologic/kubernetes-tools-kubectl@sha256:d3155b864cc74f327adedf7d013a8f0fcb11c2e73a5655b7cafa848824e80172
        - registry.connect.redhat.com/sumologic/thanos@sha256:16faf7191668e86d8b576bba055dd0cbdeb009635976b9e07bd00f926032f11c
        - registry.connect.redhat.com/sumologic/opentelemetry-operator@sha256:5587fe5b7ebfa57ad320ad2521103f908317388d1741c40e5f7be5ed9e5f9ad7
        - registry.connect.redhat.com/sumologic/nginx-unprivileged@sha256:7260555d16c94c4fe1e016626b4cb8a5c0de6b700e2eea28744331fb298f71ae
        - registry.connect.redhat.com/sumologic/prometheus-operator@sha256:58a329e6582456b32abebb470086191f8366b927af4703741613993465f116b8
        - registry.connect.redhat.com/sumologic/telegraf@sha256:75efd2e509c6017ea0fc5c9103ce068cbc89087b16506a4ef48cb5ff00eab84e
        - registry.connect.redhat.com/sumologic/tailing-sidecar-operator@sha256:c16b8b18e9b263fc3eb24059bba512bcfc3201b32105af6e5518392826ddd66c
        - registry.connect.redhat.com/sumologic/prometheus-config-reloader@sha256:0b605d36782afd8f50fd631d0e9c7b19d39ce962af89b7a737b2f09a8ceff291
        - registry.connect.redhat.com/sumologic/autoinstrumentation-java@sha256:242958b7d8905f4ac02f6ad38897ba58b9badf2a6213a389e67dcc71fee3630c
        - registry.connect.redhat.com/sumologic/busybox@sha256:209bc8e0b1d958fc699f3bb75b1248544e372ad35cdc46c991bd1698f641e1f9
        version: 4.13.0-2
      entries:
      - name: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
        version: 4.13.0-2
      name: alpha
    defaultChannel: alpha
    packageName: sumologic-kubernetes-collection-helm-operator-rhmp
    provider:
      name: Sumo Logic
      url: https://www.sumologic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-bamoe-businessautomation: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: bamoe-businessautomation-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bamoe-businessautomation-operator.8.0.8-1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.kiegroup.org/v2","kind":"KieApp","metadata":{"name":"rhpam-trial"},"spec":{"environment":"rhpam-trial"}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/ibm-bamoe/bamoe-rhel8-operator@sha256:0646da8707d10b3cc2e2344269abd8797db34127a3326bf76bbc1547aa4bbd73
          createdAt: "2025-07-22 17:40:17"
          description: Deploys and manages IBM Business Automation Manager Open Editions
            environment.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM Business Automation Manager
            Open Edition"]'
          repository: https://github.com/kiegroup/kie-cloud-operator
          support: https://www.ibm.com/mysupport
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A project prescription running an IBM BAMOE environment.
            displayName: KieApp
            kind: KieApp
            name: kieapps.app.kiegroup.org
            version: v2
        description: |-
          Deploys and manages IBM Business Automation Manager Open Editions environment.

          * **IBM Process Automation Manager** is a platform for developing containerized microservices and applications that automate business decisions and processes. It includes business process management (BPM), business rules management (BRM), and business resource optimization and complex event processing (CEP) technologies. It also includes a user experience platform to create engaging user interfaces for process and decision services with minimal coding.
        displayName: IBM Business Automation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kieapp
        - pam
        - decision
        - kie
        - cloud
        - bpm
        - process
        - automation
        - operator
        links:
        - name: Product Page
          url: https://ibm.com
        - name: Documentation
          url: https://www.ibm.com/docs/en/ibamoe
        maintainers:
        - email: mysphelp@us.ibm.com
          name: IBM
        maturity: 8.x-stable
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-monitoring-rhel8@sha256:ef1d77bb085172f927ee53d1674c85e2acaf1f580d7cad55bae8bf2c33c49c01
        - registry.redhat.io/openshift4/ose-cli@sha256:b2b93def4eb5edeccfa2d8ba22e493646325842674dff0446b45677bf641a342
        - registry.redhat.io/rhscl/postgresql-13-rhel7@sha256:338dd09681dcf860e6a976a3bc48a1474c2376d698815e4eb611ae9c5aa68966
        - registry.redhat.io/ibm-bamoe/bamoe-process-migration-rhel8@sha256:3fd45f2dabca7220772d823275a5697b42dd35f53c27671b093117933ef2d4f6
        - registry.redhat.io/rhscl/mysql-80-rhel7@sha256:e4f535c78bc2fdb58914c52fb805185035ff31515452b4df714f9ab623ba383a
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-rhel8@sha256:c3553707936cb9eee8a6a622643f72c91e1b0c11add5ec9ad5865242a9e133b4
        - registry.redhat.io/ibm-bamoe/bamoe-kieserver-rhel8@sha256:01bf97beffb3647968aae052cdd401d6d562c35f3e78ac483dbf55c44a71dc2f
        - registry.redhat.io/jboss-datagrid-7/datagrid73-openshift@sha256:4056b1cf8ff5995f56eecfdc087d9ece56a216dba6bd73acc5693f0778e8d573
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-rhel8@sha256:82d8ebde7eb3ecfbf72e5b71e7e392d3a3597fc084b6cf35e8fec23bfe41210d
        - registry.redhat.io/amq7/amq-broker@sha256:792c1cb19e62519cc47d834650fe5a6a6ca589a37d0b53d4f4549e80cabae3f9
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:98e9a6618558bc8a1b9cb91d1c9b3380968d335092f6e953d1abe6af96abf65c
        - registry.redhat.io/ibm-bamoe/bamoe-smartrouter-rhel8@sha256:2f1a7519617b5c0d082f943de0db295370649d968d699d76cfd829f4c9826a0f
        - registry.redhat.io/ibm-bamoe/bamoe-rhel8-operator@sha256:0646da8707d10b3cc2e2344269abd8797db34127a3326bf76bbc1547aa4bbd73
        - registry.redhat.io/ibm-bamoe/bamoe-process-migration-rhel8@sha256:a287a2fbaded371b8fcae5e4124b2d905bb228f94da0fc899ca2600b378dca0d
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-monitoring-rhel8@sha256:25025e489c15108edffe5270d257e1ed9be0a67aa067bcfc184647793aac2617
        - registry.redhat.io/ibm-bamoe/bamoe-dashbuilder-rhel8@sha256:efa55e17190b0ebe1618452f786876ef354b8fbec25c316aed7ed5bb1ce88116
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:457fc73423fa94591bc7e35729a25c58d3465583afccf342fb338b6e11188575
        - registry.redhat.io/ibm-bamoe/bamoe-smartrouter-rhel8@sha256:5ed8821328413ec11859ae317679f0692afff9945f134fc8909e07a3ac387794
        - registry.redhat.io/ibm-bamoe/bamoe-kieserver-rhel8@sha256:c9f08ddc6111cdd74faa0db5721d5730fc7716fccf4dc2a658595f7e0c184c10
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:f714b9c1555e7f27a3c9cf2cf687a74188f36b1283d4e09cec4db3c084d9fe64
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhscl/postgresql-12-rhel7@sha256:191237ccffc476e764714a69e35d9421dcee5ddd7184442d7a8306a06b97e6bb
        - registry.redhat.io/ibm-bamoe/bamoe-dashbuilder-rhel8@sha256:b3b7388b3b116ef37aabbb04b5d2c3e36e58d2fc95f1e59795b423010e071fc3
        - registry.redhat.io/ibm-bamoe/bamoe-controller-rhel8@sha256:3fcd9ae0bc194c198cbb5ee94272e80f302658733e467766bd1789357e9fd015
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:fb326be806b337b0ae1711f974d15e073aebcbad77aaa1123a63cf4f331da5f2
        - registry.redhat.io/ibm-bamoe/bamoe-controller-rhel8@sha256:6cd0f07856d612dca46b0a01a93b796f7a0f112b98cedfd6754f9fd770ee843d
        version: 8.0.8-1
      entries:
      - name: bamoe-businessautomation-operator.8.0.8-1
        version: 8.0.8-1
      - name: bamoe-businessautomation-operator.8.0.7-2
        version: 8.0.7-2
      - name: bamoe-businessautomation-operator.8.0.7-1
        version: 8.0.7-1
      - name: bamoe-businessautomation-operator.8.0.6-3
        version: 8.0.6-3
      - name: bamoe-businessautomation-operator.8.0.6-2
        version: 8.0.6-2
      - name: bamoe-businessautomation-operator.8.0.6-1
        version: 8.0.6-1
      - name: bamoe-businessautomation-operator.8.0.5-3
        version: 8.0.5-3
      - name: bamoe-businessautomation-operator.8.0.5-2
        version: 8.0.5-2
      - name: bamoe-businessautomation-operator.8.0.5-1
        version: 8.0.5-1
      - name: bamoe-businessautomation-operator.8.0.4-3
        version: 8.0.4-3
      - name: bamoe-businessautomation-operator.8.0.4-2
        version: 8.0.4-2
      - name: bamoe-businessautomation-operator.8.0.4-1
        version: 8.0.4-1
      - name: bamoe-businessautomation-operator.8.0.3-4
        version: 8.0.3-4
      - name: bamoe-businessautomation-operator.8.0.3-3
        version: 8.0.3-3
      - name: bamoe-businessautomation-operator.8.0.3-2
        version: 8.0.3-2
      - name: bamoe-businessautomation-operator.8.0.3-1
        version: 8.0.3-1
      - name: bamoe-businessautomation-operator.8.0.2-3
        version: 8.0.2-3
      - name: bamoe-businessautomation-operator.8.0.2-2
        version: 8.0.2-2
      - name: bamoe-businessautomation-operator.8.0.2-1
        version: 8.0.2-1
      - name: bamoe-businessautomation-operator.8.0.1-2
        version: 8.0.1-2
      - name: bamoe-businessautomation-operator.8.0.1-1
        version: 8.0.1-1
      - name: bamoe-businessautomation-operator.8.0.0-1
        version: 8.0.0-1
      name: 8.x-stable
    defaultChannel: 8.x-stable
    packageName: bamoe-businessautomation-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Odoo Community Association
      provider-url: ""
    name: odoo-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odoo-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"odoo-community.org/v1alpha1","kind":"Odoo","metadata":{"name":"example"},"spec":{"postgresqlVolumeSize":"4Gi","odooVolumeSize":"4Gi"}}]'
          capabilities: Full Lifecycle
          categories: Application Runtime
          certified: "false"
          containerImage: ursa/odoo-operator:0.0.2
          createdAt: "2021-04-26T12:00:00Z"
          description: Operator to manage Odoo instances within an OpenShift cluster
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          repository: https://github.com/ursais/odoo-operator
          support: Open Source Integrators
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Odoo
            displayName: Odoo
            kind: Odoo
            name: odoos.odoo-community.org
            version: v1alpha1
        description: |
          ## About Odoo

          [Odoo](https://odoo-community.org), is a suite of business management software tools including CRM, e-commerce, billing, accounting, manufacturing, warehouse, project management, and inventory management to name a few.
          ## About the Odoo Operator

          This is an Ansible-based operator created with operator-sdk v1.5.0.

          The upgrade is handled by the Odoo container itself using [Marabunta](https://pypi.org/project/marabunta/) and [Anthem](https://pypi.org/project/anthem/).

          ## Roadmap
          * Add horizontal pod autoscaler
          * Add cronjob for backup
          ## Prerequisites for enabling this Operator

          Please refer to the README file on the [GitHub project](https://github.com/ursais/odoo-operator).
        displayName: Odoo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ERP
        - Business Software
        links:
        - name: Odoo Community Association
          url: https://odoo-community.org
        - name: GitHub Repository
          url: https://github.com/ursais/odoo-operator
        - name: Open Source Integrators
          url: https://opensourceintegrators.com
        maintainers:
        - email: mchambreuil@opensourceintegrators.com
          name: Maxime Chambreuil
        maturity: alpha
        provider:
          name: Odoo Community Association
        relatedImages:
        - ursa/odoo-operator:0.0.2
        version: 0.0.2
      entries:
      - name: odoo-operator.v0.0.2
        version: 0.0.2
      - name: odoo-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: odoo-operator
    provider:
      name: Odoo Community Association
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fortanix Inc.
      provider-url: https://fortanix.com
    name: ccm-node-agent-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ccm-node-agent-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1",
                "kind": "CcmNodeAgent",
                "metadata": {
                  "name": "ccmnodeagent-sample"
                },
                "spec": {
                  "ccmJoinToken": {
                    "secretName": "ccm-join-token",
                    "token": ""
                  },
                  "fullnameOverride": "",
                  "image": {
                    "ccmNodeAgent": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/ccm-node-agent-ubi7",
                      "tag": "latest"
                    },
                    "ccmNodeAgentInit": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "busybox",
                      "tag": "latest"
                    },
                    "sgxSoftwareInstaller": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/sgx-software-installer-ubi7",
                      "tag": "latest"
                    }
                  },
                  "log": {
                    "debug": false
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "ports": {
                    "containerPort": 9092,
                    "hostPort": 9092
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CCM Node Agent software enables registration of the compute
              nodes to Fortanix CCM when installed in the Openshift Cluster Nodes<U+200B>.
            displayName: CCM Node Agent
            kind: CcmNodeAgent
            name: ccmnodeagents.charts.operatorhub.io
            version: v1
        description: "CCM Node Agent software enables registration of the compute
          nodes to Fortanix CCM when installed in the Openshift Cluster Nodes​. \nIt
          also assists with application attestation and visibility for Fortanix CCM.\n\n##
          About this Operator\nThis Operator is based on a Helm chart for CCM Node
          Agent. It enrolls all the nodes in the openshift cluster to a particular
          CCM account. You have to provide the base64 encode of join token of CCM
          account while creating CcmNodeAgent resource.\n\n## Prerequisites\nAll the
          worker nodes in the cluster must have SGX capabilities.\n\n## Installation\n*
          Choose a namespace to install the Operator. Operator will be available in
          all namespaces.\n* Select Approval Strategy: Available options are Automatic
          or Manual. Automatic approval will automatically upgrade the running instance
          of your Operator without human intervention when a new version is available.
          In case of Manual approval, when a new version is available, OLM will create
          an update request. You must manually approve the request to update the Operator
          to a new version.\n* Generate your CCM Account Join Token: Please log in
          to https://ccm.fortanix.com, and in the Management Console tab, click the
          + ENROLL NODE button. In the ENROLL NODE screen, a Join Token will be generated
          in the text box for \"Get a join token to register an SGX compute node\".
          This Join Token is used by the compute node to authenticate itself.\n* Base64
          encode the Join Token:\n```shell\necho -n <join-token> | openssl base64
          -A\n```\n* Create an instance of CcmNodeAgent by setting the `spec.ccmJoinToken.Token`
          to the base64 encoded CCM account join token.\n* After CcmNodeAgent resource
          is installed successfully you should be able to see all the worker nodes
          of the cluster in your CCM account under Compute Node Tab.\n\n## Getting
          Started\nNow to run your application inside SGX enclaves on your cluster
          follow the [documentation](https://support.fortanix.com/hc/en-us/articles/360043527431-User-s-Guide-Add-and-Edit-an-Application).\n"
        displayName: CCM Node Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dockerfile
        - Helm
        - CCM
        - Confidential Compute
        links:
        - name: CCM Node Agent Operator
          url: https://support.fortanix.com/hc/en-us/categories/360003107511-Confidential-Computing-Manager
        maintainers:
        - email: info@fortanix.com
          name: Fortanix Inc.
        maturity: alpha
        provider:
          name: Fortanix Inc.
          url: https://fortanix.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62575404000cb77fa8d2284c87b31e39e667a55b0fb47821263b23b3d1135a7e
        - docker.io/fortanix/ccm-node-agent-operator@sha256:e2a3d52551198102ddbe7f4ebd9621edc357c57ddeff47b7ed9f2cf56a1bed39
        version: 0.0.2
      entries:
      - name: ccm-node-agent-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ccm-node-agent-operator
    provider:
      name: Fortanix Inc.
      url: https://fortanix.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: https://intel.com
    name: intel-device-plugins-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: intel-device-plugins-operator.v0.29.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "GpuDevicePlugin",
                "metadata": {
                  "name": "gpudeviceplugin-sample"
                },
                "spec": {
                  "enableMonitoring": true,
                  "image": "registry.connect.redhat.com/intel/intel-gpu-plugin@sha256:963caf1e6e16b5e297671c536eec04dc57cf45f5cc5cf458bdcafa56bdb97972",
                  "preferredAllocationPolicy": "none",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/gpu": "true"
                  },
                  "sharedDevNum": 1
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "QatDevicePlugin",
                "metadata": {
                  "name": "qatdeviceplugin-sample"
                },
                "spec": {
                  "dpdkDriver": "vfio-pci",
                  "image": "registry.connect.redhat.com/intel/intel-qat-plugin@sha256:53617c3a72af169031cbaba1380cd6a25c9d2b50c9a2da822cf37b473c6e7a8f",
                  "initImage": "registry.connect.redhat.com/intel/intel-qat-initcontainer@sha256:14f282a0adae43262657f24ce99a620b671c9a2ce12a8216163e6b22ac6e252b",
                  "kernelVfDrivers": [
                    "c6xxvf",
                    "4xxxvf"
                  ],
                  "logLevel": 4,
                  "maxNumDevices": 128,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/qat": "true"
                  }
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "DsaDevicePlugin",
                "metadata": {
                  "name": "dsadeviceplugin-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/intel/intel-dsa-plugin@sha256:092575515739bf9a0f5982c3605b63382bc4d1250128269434c2bfe4a27ec99d",
                  "initImage": "registry.connect.redhat.com/intel/intel-idxd-config-initcontainer@sha256:649e4f4e7f362ba054fb85a67e3baa23ec9649c68df356d1e33721ca3eef2f6b",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/dsa": "true"
                  },
                  "sharedDevNum": 10
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "SgxDevicePlugin",
                "metadata": {
                  "name": "sgxdeviceplugin-sample"
                },
                "spec": {
                  "enclaveLimit": 110,
                  "image": "registry.connect.redhat.com/intel/intel-sgx-plugin@sha256:e54cdf61feb0b67e330ad24d92e7fa4000953c2ec4d4300d052e032d419d5e0a",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/sgx": "true"
                  },
                  "provisionLimit": 110
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Security, AI/Machine Learning
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/intel/intel-technology-enabling-for-openshift
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GpuDevicePlugin is the Schema for the gpudeviceplugins API.
              It represents the GPU device plugin responsible for advertising Intel
              GPU hardware resources to the kubelet.
            displayName: Intel GPU Device Plugin
            kind: GpuDevicePlugin
            name: gpudeviceplugins.deviceplugin.intel.com
            version: v1
          - description: DsaDevicePlugin is the Schema for the dsadeviceplugins API.
              It represents the DSA device plugin responsible for advertising Intel
              DSA hardware resources to the kubelet.
            displayName: Intel DSA Device Plugin
            kind: DsaDevicePlugin
            name: dsadeviceplugins.deviceplugin.intel.com
            version: v1
          - description: QatDevicePlugin is the Schema for the qatdeviceplugins API.
              It represents the QAT device plugin responsible for advertising Intel
              QuickAssist Technology hardware resources to the kubelet.
            displayName: Intel QuickAssist Technology Device Plugin
            kind: QatDevicePlugin
            name: qatdeviceplugins.deviceplugin.intel.com
            version: v1
          - description: SgxDevicePlugin is the Schema for the sgxdeviceplugins API.
              It represents the SGX device plugin responsible for advertising SGX
              device nodes to the kubelet.
            displayName: Intel Software Guard Extensions Device Plugin
            kind: SgxDevicePlugin
            name: sgxdeviceplugins.deviceplugin.intel.com
            version: v1
        description: |
          [Intel Device Plugins for Kubernetes](https://github.com/intel/intel-technology-enabling-for-openshift) is a collection of
          [device plugins](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/) advertising Intel specific hardware resources
          to the kubelet. Currently the operator has basic support for Intel SGX, Intel GPU, Intel QAT and Intel DSA. Intel IAA device plugin will be supported in a future release.
        displayName: Intel Device Plugins Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - qat
        - sgx
        - gpu
        - dsa
        - dlb
        links:
        - name: Intel Device Plugins Operator
          url: https://github.com/intel/intel-technology-enabling-for-openshift
        maintainers:
        - email: martin.xu@intel.com
          name: Martin Xu
        - email: chaitanya.kulkarni@intel.com
          name: Chaitanya Kulkarni
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Intel® Corporation
          url: https://intel.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - registry.connect.redhat.com/intel/intel-sgx-plugin@sha256:e54cdf61feb0b67e330ad24d92e7fa4000953c2ec4d4300d052e032d419d5e0a
        - registry.connect.redhat.com/intel/intel-gpu-plugin@sha256:963caf1e6e16b5e297671c536eec04dc57cf45f5cc5cf458bdcafa56bdb97972
        - registry.connect.redhat.com/intel/intel-qat-plugin@sha256:53617c3a72af169031cbaba1380cd6a25c9d2b50c9a2da822cf37b473c6e7a8f
        - registry.connect.redhat.com/intel/intel-qat-initcontainer@sha256:14f282a0adae43262657f24ce99a620b671c9a2ce12a8216163e6b22ac6e252b
        - registry.connect.redhat.com/intel/intel-dsa-plugin@sha256:092575515739bf9a0f5982c3605b63382bc4d1250128269434c2bfe4a27ec99d
        - registry.connect.redhat.com/intel/intel-idxd-initcontainer@sha256:649e4f4e7f362ba054fb85a67e3baa23ec9649c68df356d1e33721ca3eef2f6b
        - registry.connect.redhat.com/intel/intel-deviceplugin-operator@sha256:283e8b60d7ab8560f8ec8875b905e6446b59a91dd5e0d97da72e3a9aab235cc4
        version: 0.29.1
      entries:
      - name: intel-device-plugins-operator.v0.29.1
        version: 0.29.1
      - name: intel-device-plugins-operator.v0.29.0
        version: 0.29.0
      - name: intel-device-plugins-operator.v0.28.0
        version: 0.28.0
      name: alpha
    defaultChannel: alpha
    packageName: intel-device-plugins-operator
    provider:
      name: Intel® Corporation
      url: https://intel.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: designate-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: designate-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "designate.openstack.org/v1beta1",
                "kind": "Designate",
                "metadata": {
                  "name": "designate"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "designate",
                  "databaseInstance": "openstack",
                  "designateAPI": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate"
                  },
                  "designateBackendbind9": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "networkAttachments": [
                      "designate"
                    ],
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate",
                    "storageClass": "local-storage",
                    "storageRequest": "1G"
                  },
                  "designateCentral": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate"
                  },
                  "designateMdns": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "networkAttachments": [
                      "designate"
                    ],
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate"
                  },
                  "designateProducer": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate"
                  },
                  "designateUnbound": {
                    "replicas": 1
                  },
                  "designateWorker": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "networkAttachments": [
                      "designate"
                    ],
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate"
                  },
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "osp-secret",
                  "serviceUser": "designate"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:48Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DesignateAPI is the Schema for the designateapis API
            displayName: Designate API
            kind: DesignateAPI
            name: designateapis.designate.openstack.org
            version: v1beta1
          - kind: DesignateBackendbind9
            name: designatebackendbind9s.designate.openstack.org
            version: v1beta1
          - description: DesignateCentral is the Schema for the designatecentral API
            displayName: Designate Central
            kind: DesignateCentral
            name: designatecentrals.designate.openstack.org
            version: v1beta1
          - description: DesignateMdns is the Schema for the designatemdnses API
            displayName: Designate Mdns
            kind: DesignateMdns
            name: designatemdnses.designate.openstack.org
            version: v1beta1
          - description: DesignateProducer is the Schema for the designateproducer
              API
            displayName: Designate Producer
            kind: DesignateProducer
            name: designateproducers.designate.openstack.org
            version: v1beta1
          - description: Designate is the Schema for the designates API
            displayName: Designate
            kind: Designate
            name: designates.designate.openstack.org
            version: v1beta1
          - description: DesignateUnbound is the Schema for the designateunbound API
            displayName: Designate Unbound
            kind: DesignateUnbound
            name: designateunbounds.designate.openstack.org
            version: v1beta1
          - description: DesignateWorker is the Schema for the designateworker API
            displayName: Designate Worker
            kind: DesignateWorker
            name: designateworkers.designate.openstack.org
            version: v1beta1
        description: Designate Operator
        displayName: Designate Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Network
        - Designate
        links:
        - name: Designate Operator
          url: https://github.com/openstack-k8s-operators/designate-operator
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso/openstack-designate-api-rhel9@sha256:1a9baa4a0f99029eb8a367ddcce264581aa37fdf1834ba277c30a3f0ae33f400
        - registry.redhat.io/rhoso/openstack-designate-central-rhel9@sha256:ad610cdc29b31cd82a81cea22793176351574f43c0f7705273c8ffc03ad2fdac
        - registry.redhat.io/rhoso/openstack-designate-producer-rhel9@sha256:4beb652a073a65965f96f125d0fcb454255efa5e53a84bd0dd9120b57e59fcf3
        - registry.redhat.io/rhoso-operators/designate-rhel9-operator@sha256:e6c1ba511cfd0278fc8cd2836c6324d97ae3f6e9affc992600978062fa2d4549
        - registry.redhat.io/rhoso/openstack-designate-mdns-rhel9@sha256:a4109755360f340aadf6bc15d24599f882d82bd59739065f87275f9274d9eeb9
        - registry.redhat.io/rhoso/openstack-designate-worker-rhel9@sha256:94ef291ca85b06dffc93d149b09b412d9190c99f7a1b0b300d50bd70346401ad
        - registry.redhat.io/rhoso/openstack-designate-backend-bind9-rhel9@sha256:89ce30fc1eff339867ce422203e6e0c63296667e102ce85d823d44c30f7556ad
        - registry.redhat.io/rhoso/openstack-unbound-rhel9@sha256:ad95d40227a1e52f739f09a2ce8db647d6a2fa7ef63e58a2444688728fa5bdfd
        version: 1.0.6
      entries:
      - name: designate-operator.v1.0.6
        version: 1.0.6
      - name: designate-operator.v1.0.5
        version: 1.0.5
      - name: designate-operator.v1.0.4
        version: 1.0.4
      - name: designate-operator.v1.0.3
        version: 1.0.3
      - name: designate-operator.v1.0.2
        version: 1.0.2
      - name: designate-operator.v1.0.1
        version: 1.0.1
      - name: designate-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: designate-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-cert-manager-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-manager-operator.v1.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
          createdAt: 2025-08-06T08:21:40
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.16.1 <1.17.0'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.17.4](https://github.com/cert-manager/cert-manager/tree/v1.17.4), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:96d51e3a64bf30cbd92836c7cbd82f06edca16eef78ab1432757d34c16628659
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b
        version: 1.17.0
      entries:
      - name: cert-manager-operator.v1.17.0
        version: 1.17.0
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      - name: cert-manager-operator.v1.14.1
        version: 1.14.1
      - name: cert-manager-operator.v1.14.0
        version: 1.14.0
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1
    - currentCSV: cert-manager-operator.v1.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "\u003cACCESS_KEY_ID\u003e",
                            "hostedZoneID": "\u003cHOSTED_ZONE_ID\u003e",
                            "region": "\u003cAWS_REGION\u003e",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4e854c14fecbd27107b91b76164ce2e1575ce60202d835abba5962fb5d386560
          createdAt: 2024-04-30T13:29:33
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.13.0 <1.13.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/infrastructure-features: '["proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.13.6](https://github.com/cert-manager/cert-manager/tree/v1.13.6), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:afc82895c09d75e3d6a932aea2d3e3ef998f765306da92070de1dab26baf0593
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4e854c14fecbd27107b91b76164ce2e1575ce60202d835abba5962fb5d386560
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:578f74c3e96db7fc2f4e94dd9fbc7d6499d87b0bc55576930ee17de08a95022c
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:7915f8049d2451b425e3674eb60e596b9b11946ffda1497f5d5607799e2bf431
        version: 1.13.1
      entries:
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1.13
    - currentCSV: cert-manager-operator.v1.14.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:5c593bfd70309f87cd05716e233e65aa55394eafcdaf85f178698912cb29d79d
          createdAt: 2025-04-09T05:19:10
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.14.1 <1.14.2'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/infrastructure-features: '["proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.14.7](https://github.com/cert-manager/cert-manager/tree/v1.14.7), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:5c593bfd70309f87cd05716e233e65aa55394eafcdaf85f178698912cb29d79d
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:5df082f8e3787e6ac2edf395113806999ae192425ce18b4ddba9ab20c135fa5b
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:49fb8f710319d6f3479b98d2c5dd7d0c9df8c1678b2c00727314dea1a8e933a1
        version: 1.14.2
      entries:
      - name: cert-manager-operator.v1.14.2
        version: 1.14.2
      - name: cert-manager-operator.v1.14.1
        version: 1.14.1
      - name: cert-manager-operator.v1.14.0
        version: 1.14.0
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1.14
    - currentCSV: cert-manager-operator.v1.15.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:f328263e2d29e34ede65e4501f0447b2d9f84e9445a365c2fa2fbb253939e274
          createdAt: 2025-03-13T18:26:37
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.15.0 <1.15.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/infrastructure-features: '["proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.15.5](https://github.com/cert-manager/cert-manager/tree/v1.15.2), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:f328263e2d29e34ede65e4501f0447b2d9f84e9445a365c2fa2fbb253939e274
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9573d74bd2b926ec94af76f813e6358f14c5b2f4e0eedab7c1ff1070b7279a5c
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:6af3ee8b2a5a87042fb7158bda8d6cf2e6324d1e265974acf77214d4cd0ea0d3
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:20efff60a0caf5eafb38986fd21611697b5bc534c2e789da233983a9739938ed
        version: 1.15.1
      entries:
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      - name: cert-manager-operator.v1.14.1
        version: 1.14.1
      - name: cert-manager-operator.v1.14.0
        version: 1.14.0
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1.15
    - currentCSV: cert-manager-operator.v1.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:d30e755a266bb035d50e2b257e7a26c1fe6a5ccb66125576db99d537aa03ec27
          createdAt: 2025-07-10T08:52:43
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.15.1 <1.16.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.16.5](https://github.com/cert-manager/cert-manager/tree/v1.16.5), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:831e4b1e5aa6eed6fbb22f779ddbcb9c73e4bf6cc079d0c6652737a6a44e6320
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:d30e755a266bb035d50e2b257e7a26c1fe6a5ccb66125576db99d537aa03ec27
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:33308217d2226c2555eb18020710d16175295a8e349263a5435e4b158bbc735f
        version: 1.16.1
      entries:
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      - name: cert-manager-operator.v1.14.1
        version: 1.14.1
      - name: cert-manager-operator.v1.14.0
        version: 1.14.0
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1.16
    - currentCSV: cert-manager-operator.v1.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
          createdAt: 2025-08-06T08:21:40
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.16.1 <1.17.0'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.17.4](https://github.com/cert-manager/cert-manager/tree/v1.17.4), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:96d51e3a64bf30cbd92836c7cbd82f06edca16eef78ab1432757d34c16628659
        version: 1.17.0
      entries:
      - name: cert-manager-operator.v1.17.0
        version: 1.17.0
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      - name: cert-manager-operator.v1.14.1
        version: 1.14.1
      - name: cert-manager-operator.v1.14.0
        version: 1.14.0
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1.17
    defaultChannel: stable-v1
    packageName: openshift-cert-manager-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ram-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ram-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ram-controller:1.1.1
          createdAt: "2025-09-19T22:18:09Z"
          description: AWS RAM controller is a service controller for managing RAM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Permission represents the state of an AWS ram Permission
              resource.
            displayName: Permission
            kind: Permission
            name: permissions.ram.services.k8s.aws
            version: v1alpha1
          - description: ResourceShare represents the state of an AWS ram ResourceShare
              resource.
            displayName: ResourceShare
            kind: ResourceShare
            name: resourceshares.ram.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon RAM resources in AWS from within your Kubernetes cluster.

          **About Amazon RAM**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon RAM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ram
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon RAM Developer Resources
          url: https://aws.amazon.com/RAM/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ram maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ram-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ram-controller.v1.1.1
        version: 1.1.1
      - name: ack-ram-controller.v1.1.0
        version: 1.1.0
      - name: ack-ram-controller.v1.0.11
        version: 1.0.11
      - name: ack-ram-controller.v1.0.10
        version: 1.0.10
      name: alpha
    defaultChannel: alpha
    packageName: ack-ram-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operator-kogitocloud: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kogito-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kogito-operator.v1.44.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "process-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogito-infinispan-infra"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "infinispan.org/v1",
                    "kind": "Infinispan"
                  }
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/kiegroup/kogito-operator:1.44.1
          createdAt: "2019-08-22T13:12:22Z"
          description: Kogito Operator for deployment and management of Kogito services.
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.app.kiegroup.org
            version: v1beta1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.app.kiegroup.org
            version: v1beta1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.app.kiegroup.org
            version: v1beta1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.app.kiegroup.org
            version: v1beta1
        description: |-
          Kogito Operator is designed to deploy Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Creates two kinds of build configuration: Native or JVM. For more information refer to [Native build and deployment](https://docs.jboss.org/kogito/release/latest/html_single/#_native_build_and_deployment) section on our docs
          * Builds a Kogito service from a remote git repository and creates a custom image for the given runtime: Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Installs the [Kogito Data Index Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-data-index-service_kogito-deploying-on-openshift) to enable data index feature for Kogito services.
          * Installs the [Kogito Jobs Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-jobs-service_kogito-deploying-on-openshift) to enable jobs scheduling.

          For Kubernetes deployment of Kogito services, please refer to the section [Deploying Kogito services on Kubernetes](https://docs.jboss.org/kogito/release/latest/html_single/#proc-kogito-deploying-on-kubernetes_kogito-deploying-on-openshift) on our docs.

          **IMPORTANT!** Kogito Operator might need Infinispan Operator and Strimzi to implement persistence and messaging use cases if you do not provide these services beforehand.
          If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the Kogito Operator.
        displayName: Kogito
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://kogito.kie.org/
        - name: Documentation
          url: https://github.com/kiegroup/kogito-operator/blob/main/README.md
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/kiegroup/kogito-operator:1.44.1
        version: 1.44.1
      entries:
      - name: kogito-operator.v1.44.1
        version: 1.44.1
      - name: kogito-operator.v1.43.0
        version: 1.43.0
      - name: kogito-operator.v1.42.0
        version: 1.42.0
      - name: kogito-operator.v1.41.0
        version: 1.41.0
      - name: kogito-operator.v1.40.0
        version: 1.40.0
      - name: kogito-operator.v1.39.0
        version: 1.39.0
      - name: kogito-operator.v1.38.0
        version: 1.38.0
      - name: kogito-operator.v1.37.0
        version: 1.37.0
      - name: kogito-operator.v1.36.0
        version: 1.36.0
      - name: kogito-operator.v1.35.0
        version: 1.35.0
      - name: kogito-operator.v1.34.0
        version: 1.34.0
      - name: kogito-operator.v1.33.0
        version: 1.33.0
      - name: kogito-operator.v1.32.0
        version: 1.32.0
      - name: kogito-operator.v1.31.0
        version: 1.31.0
      - name: kogito-operator.v1.30.0
        version: 1.30.0
      - name: kogito-operator.v1.29.0
        version: 1.29.0
      - name: kogito-operator.v1.28.0
        version: 1.28.0
      - name: kogito-operator.v1.27.0
        version: 1.27.0
      - name: kogito-operator.v1.26.0
        version: 1.26.0
      - name: kogito-operator.v1.25.0
        version: 1.25.0
      - name: kogito-operator.v1.24.0
        version: 1.24.0
      - name: kogito-operator.v1.23.0
        version: 1.23.0
      - name: kogito-operator.v1.22.1
        version: 1.22.1
      - name: kogito-operator.v1.22.0
        version: 1.22.0
      - name: kogito-operator.v1.21.0
        version: 1.21.0
      - name: kogito-operator.v1.20.0
        version: 1.20.0
      - name: kogito-operator.v1.19.0
        version: 1.19.0
      - name: kogito-operator.v1.18.0
        version: 1.18.0
      - name: kogito-operator.v1.17.0
        version: 1.17.0
      - name: kogito-operator.v1.16.1
        version: 1.16.1
      - name: kogito-operator.v1.16.0
        version: 1.16.0
      - name: kogito-operator.v1.15.0
        version: 1.15.0
      - name: kogito-operator.v1.14.0
        version: 1.14.0
      - name: kogito-operator.v1.13.0
        version: 1.13.0
      - name: kogito-operator.v1.12.0
        version: 1.12.0
      - name: kogito-operator.v1.11.1
        version: 1.11.1
      - name: kogito-operator.v1.11.0
        version: 1.11.0
      - name: kogito-operator.v1.10.0
        version: 1.10.0
      - name: kogito-operator.v1.9.1
        version: 1.9.1
      - name: kogito-operator.v1.9.0
        version: 1.9.0
      - name: kogito-operator.v1.8.0
        version: 1.8.0
      - name: kogito-operator.v1.7.0
        version: 1.7.0
      - name: kogito-operator.v1.6.0
        version: 1.6.0
      - name: kogito-operator.v1.5.0
        version: 1.5.0
      - name: kogito-operator.v1.4.0
        version: 1.4.0
      name: 1.x
    - currentCSV: kogito-operator.v1.44.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "process-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogito-infinispan-infra"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "infinispan.org/v1",
                    "kind": "Infinispan"
                  }
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/kiegroup/kogito-operator:1.44.1
          createdAt: "2019-08-22T13:12:22Z"
          description: Kogito Operator for deployment and management of Kogito services.
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.app.kiegroup.org
            version: v1beta1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.app.kiegroup.org
            version: v1beta1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.app.kiegroup.org
            version: v1beta1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.app.kiegroup.org
            version: v1beta1
        description: |-
          Kogito Operator is designed to deploy Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Creates two kinds of build configuration: Native or JVM. For more information refer to [Native build and deployment](https://docs.jboss.org/kogito/release/latest/html_single/#_native_build_and_deployment) section on our docs
          * Builds a Kogito service from a remote git repository and creates a custom image for the given runtime: Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Installs the [Kogito Data Index Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-data-index-service_kogito-deploying-on-openshift) to enable data index feature for Kogito services.
          * Installs the [Kogito Jobs Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-jobs-service_kogito-deploying-on-openshift) to enable jobs scheduling.

          For Kubernetes deployment of Kogito services, please refer to the section [Deploying Kogito services on Kubernetes](https://docs.jboss.org/kogito/release/latest/html_single/#proc-kogito-deploying-on-kubernetes_kogito-deploying-on-openshift) on our docs.

          **IMPORTANT!** Kogito Operator might need Infinispan Operator and Strimzi to implement persistence and messaging use cases if you do not provide these services beforehand.
          If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the Kogito Operator.
        displayName: Kogito
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://kogito.kie.org/
        - name: Documentation
          url: https://github.com/kiegroup/kogito-operator/blob/main/README.md
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kiegroup/kogito-operator:1.44.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        version: 1.44.1
      entries:
      - name: kogito-operator.v1.44.1
        version: 1.44.1
      - name: kogito-operator.v1.43.0
        version: 1.43.0
      - name: kogito-operator.v1.42.0
        version: 1.42.0
      - name: kogito-operator.v1.41.0
        version: 1.41.0
      - name: kogito-operator.v1.40.0
        version: 1.40.0
      - name: kogito-operator.v1.39.0
        version: 1.39.0
      - name: kogito-operator.v1.38.0
        version: 1.38.0
      - name: kogito-operator.v1.37.0
        version: 1.37.0
      - name: kogito-operator.v1.36.0
        version: 1.36.0
      - name: kogito-operator.v1.35.0
        version: 1.35.0
      - name: kogito-operator.v1.34.0
        version: 1.34.0
      - name: kogito-operator.v1.33.0
        version: 1.33.0
      - name: kogito-operator.v1.32.0
        version: 1.32.0
      - name: kogito-operator.v1.31.0
        version: 1.31.0
      - name: kogito-operator.v1.30.0
        version: 1.30.0
      - name: kogito-operator.v1.29.0
        version: 1.29.0
      - name: kogito-operator.v1.28.0
        version: 1.28.0
      - name: kogito-operator.v1.27.0
        version: 1.27.0
      - name: kogito-operator.v1.26.0
        version: 1.26.0
      - name: kogito-operator.v1.25.0
        version: 1.25.0
      - name: kogito-operator.v1.24.0
        version: 1.24.0
      - name: kogito-operator.v1.23.0
        version: 1.23.0
      - name: kogito-operator.v1.22.1
        version: 1.22.1
      - name: kogito-operator.v1.22.0
        version: 1.22.0
      - name: kogito-operator.v1.21.0
        version: 1.21.0
      - name: kogito-operator.v1.20.0
        version: 1.20.0
      - name: kogito-operator.v1.19.0
        version: 1.19.0
      - name: kogito-operator.v1.18.0
        version: 1.18.0
      - name: kogito-operator.v1.17.0
        version: 1.17.0
      - name: kogito-operator.v1.16.1
        version: 1.16.1
      - name: kogito-operator.v1.16.0
        version: 1.16.0
      - name: kogito-operator.v1.15.0
        version: 1.15.0
      - name: kogito-operator.v1.14.0
        version: 1.14.0
      - name: kogito-operator.v1.13.0
        version: 1.13.0
      - name: kogito-operator.v1.12.0
        version: 1.12.0
      - name: kogito-operator.v1.11.1
        version: 1.11.1
      - name: kogito-operator.v1.11.0
        version: 1.11.0
      - name: kogito-operator.v1.10.0
        version: 1.10.0
      - name: kogito-operator.v1.9.1
        version: 1.9.1
      - name: kogito-operator.v1.9.0
        version: 1.9.0
      - name: kogito-operator.v1.8.0
        version: 1.8.0
      - name: kogito-operator.v1.7.0
        version: 1.7.0
      - name: kogito-operator.v1.6.0
        version: 1.6.0
      - name: kogito-operator.v1.5.0
        version: 1.5.0
      - name: kogito-operator.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: 1.x
    packageName: kogito-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: neutron-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neutron-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "neutron.openstack.org/v1beta1",
                "kind": "NeutronAPI",
                "metadata": {
                  "name": "neutron",
                  "namespace": "openstack"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "neutron",
                  "databaseInstance": "openstack",
                  "memcachedInstance": "memcached",
                  "preserveJobs": false,
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "neutron-secret",
                  "serviceUser": "neutron"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:57Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NeutronAPI is the Schema for the neutronapis API
            displayName: Neutron API
            kind: NeutronAPI
            name: neutronapis.neutron.openstack.org
            version: v1beta1
        description: Neutron Operator
        displayName: Neutron Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Network
        - Neutron
        links:
        - name: Neutron Operator
          url: https://github.com/openstack-k8s-operators/neutron-operator
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/neutron-rhel9-operator@sha256:16b1a5f83692f3d193754df664992e5b88cb357f158b69261eb802430f3a69f1
        - registry.redhat.io/rhoso/openstack-neutron-server-rhel9@sha256:806e5e009a77a29c82699487e539669e1ec2136514a0332d222b5cb3fed6f452
        version: 1.0.6
      entries:
      - name: neutron-operator.v1.0.6
        version: 1.0.6
      - name: neutron-operator.v1.0.5
        version: 1.0.5
      - name: neutron-operator.v1.0.4
        version: 1.0.4
      - name: neutron-operator.v1.0.3
        version: 1.0.3
      - name: neutron-operator.v1.0.2
        version: 1.0.2
      - name: neutron-operator.v1.0.1
        version: 1.0.1
      - name: neutron-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: neutron-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Tempo Operator SIG
      provider-url: ""
    name: tempo-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tempo-operator.v0.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoMonolithic",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "jaegerui": {
                    "enabled": true,
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    }
                  },
                  "storage": {
                    "traces": {
                      "backend": "memory"
                    }
                  }
                }
              },
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoStack",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "resources": {
                    "total": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2Gi"
                      }
                    }
                  },
                  "storage": {
                    "secret": {
                      "name": "my-storage-secret",
                      "type": "s3"
                    }
                  },
                  "storageSize": "1Gi",
                  "template": {
                    "queryFrontend": {
                      "jaegerQuery": {
                        "enabled": true,
                        "ingress": {
                          "type": "route"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          containerImage: ghcr.io/grafana/tempo-operator/tempo-operator:v0.18.0
          createdAt: "2025-09-17T14:23:28Z"
          description: Create and manage deployments of Tempo, a high-scale distributed
            tracing backend.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: tempo-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/tempo-operator
          support: Grafana Tempo Operator SIG
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TempoMonolithic manages a Tempo deployment in monolithic
              mode.
            displayName: TempoMonolithic
            kind: TempoMonolithic
            name: tempomonolithics.tempo.grafana.com
            version: v1alpha1
          - description: TempoStack manages a Tempo deployment in microservices mode.
            displayName: TempoStack
            kind: TempoStack
            name: tempostacks.tempo.grafana.com
            version: v1alpha1
        description: |-
          Tempo is an open source, easy-to-use, and high-scale distributed tracing backend.
          It can ingest common open source tracing protocols including Jaeger, Zipkin, and OpenTelemetry and requires only object storage to operate.
          Please refer to the [Tempo documentation](https://grafana.com/docs/tempo/latest/) for more information about Tempo.

          The Community Tempo Operator can manage Tempo deployments in Microservices mode (`TempoStack` CR) and Monolithic mode (`TempoMonolithic` CR).

          ### Operator features
          * **Resource Limits** - Specify overall resource requests and limits in the `TempoStack` CR; the operator assigns fractions of it to each component
          * **AuthN and AuthZ** - Supports OpenID Control (OIDC) and role-based access control (RBAC)
          * **Managed upgrades** - Updating the operator will automatically update all managed Tempo clusters
          * **Multitenancy** - Multiple tenants can send traces to the same Tempo cluster
          * **mTLS** - Communication between the Tempo components can be secured via mTLS
          * **Jaeger UI** - Traces can be visualized in Jaeger UI and exposed via Ingress or OpenShift Route
          * **Observability** - The operator and `TempoStack` operands expose telemetry (metrics, traces) and integrate with Prometheus `ServiceMonitor` and `PrometheusRule`

          ### Prerequisites
          `TempoStack` requires object storage to store its traces. `TempoMonolithic` can store traces in-memory, in a Persistent Volume and in object storage.
          Please ensure that an object storage solution is available and configured.
        displayName: Community Tempo Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tempo
        - tracing
        - observability
        - monitoring
        - database
        links:
        - name: Tempo Operator
          url: https://github.com/grafana/tempo-operator
        maintainers:
        - email: ruben.vp8510@gmail.com
          name: Ruben Vargas
        - email: p.loffay@gmail.com
          name: Pavol Loffay
        - email: bongartz@klimlive.de
          name: Benedikt Bongartz
        - email: andreas@gerstmayr.me
          name: Andreas Gerstmayr
        - email: iblancas@redhat.com
          name: Israel Blancas Alvarez
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Grafana Tempo Operator SIG
        relatedImages:
        - docker.io/jaegertracing/jaeger-query:1.68.0
        - docker.io/grafana/tempo-query:2.8.2
        - quay.io/observatorium/api:main-2025-08-25-7489eb0
        - quay.io/observatorium/opa-openshift:main-2025-06-16-ecdeca0
        - quay.io/openshift/origin-oauth-proxy:4.14
        - ghcr.io/grafana/tempo-operator/tempo-operator:v0.18.0
        - docker.io/grafana/tempo:2.8.2
        version: 0.18.0
      entries:
      - name: tempo-operator.v0.18.0
        version: 0.18.0
      - name: tempo-operator.v0.17.1
        version: 0.17.1
      - name: tempo-operator.v0.15.3
        version: 0.15.3
      - name: tempo-operator.v0.15.2
        version: 0.15.2
      - name: tempo-operator.v0.15.1
        version: 0.15.1
      - name: tempo-operator.v0.15.0
        version: 0.15.0
      - name: tempo-operator.v0.14.2
        version: 0.14.2
      - name: tempo-operator.v0.14.1
        version: 0.14.1
      - name: tempo-operator.v0.14.0
        version: 0.14.0
      - name: tempo-operator.v0.13.0
        version: 0.13.0
      - name: tempo-operator.v0.12.0
        version: 0.12.0
      - name: tempo-operator.v0.11.1
        version: 0.11.1
      - name: tempo-operator.v0.11.0
        version: 0.11.0
      - name: tempo-operator.v0.10.0
        version: 0.10.0
      - name: tempo-operator.v0.9.0
        version: 0.9.0
      - name: tempo-operator.v0.8.0
        version: 0.8.0
      - name: tempo-operator.v0.7.0
        version: 0.7.0
      - name: tempo-operator.v0.6.0
        version: 0.6.0
      - name: tempo-operator.v0.5.0
        version: 0.5.0
      - name: tempo-operator.v0.4.0
        version: 0.4.0
      - name: tempo-operator.v0.3.0
        version: 0.3.0
      - name: tempo-operator.v0.2.0
        version: 0.2.0
      - name: tempo-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: tempo-operator
    provider:
      name: Grafana Tempo Operator SIG
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: project-quay
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-operator.v3.10.0-nightly.20231014
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.10.0-nightly.20231014
          createdAt: 2023-10-14 06:26 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.10.0-nightly.20231014'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.10.0-nightly.20231014
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.10.0-nightly.20231014
        version: 3.10.0-nightly.20231014
      entries:
      - name: quay-operator.v3.10.0-nightly.20231014
        version: 3.10.0-nightly.20231014
      - name: quay-operator.v3.10.0-nightly.20231013
        version: 3.10.0-nightly.20231013
      - name: quay-operator.v3.10.0-nightly.20231012
        version: 3.10.0-nightly.20231012
      - name: quay-operator.v3.10.0-nightly.20231011
        version: 3.10.0-nightly.20231011
      - name: quay-operator.v3.10.0-nightly.20231010
        version: 3.10.0-nightly.20231010
      - name: quay-operator.v3.10.0-nightly.20231009
        version: 3.10.0-nightly.20231009
      - name: quay-operator.v3.10.0-nightly.20231008
        version: 3.10.0-nightly.20231008
      - name: quay-operator.v3.10.0-nightly.20231007
        version: 3.10.0-nightly.20231007
      - name: quay-operator.v3.10.0-nightly.20231006
        version: 3.10.0-nightly.20231006
      - name: quay-operator.v3.10.0-nightly.20231005
        version: 3.10.0-nightly.20231005
      - name: quay-operator.v3.10.0-nightly.20231002
        version: 3.10.0-nightly.20231002
      - name: quay-operator.v3.10.0-nightly.20230930
        version: 3.10.0-nightly.20230930
      - name: quay-operator.v3.10.0-nightly.20230929
        version: 3.10.0-nightly.20230929
      - name: quay-operator.v3.10.0-nightly.20230928
        version: 3.10.0-nightly.20230928
      - name: quay-operator.v3.10.0-nightly.20230926
        version: 3.10.0-nightly.20230926
      - name: quay-operator.v3.10.0-nightly.20230925
        version: 3.10.0-nightly.20230925
      - name: quay-operator.v3.10.0-nightly.20230923
        version: 3.10.0-nightly.20230923
      - name: quay-operator.v3.10.0-nightly.20230922
        version: 3.10.0-nightly.20230922
      - name: quay-operator.v3.10.0-nightly.20230921
        version: 3.10.0-nightly.20230921
      - name: quay-operator.v3.10.0-nightly.20230920
        version: 3.10.0-nightly.20230920
      - name: quay-operator.v3.10.0-nightly.20230919
        version: 3.10.0-nightly.20230919
      - name: quay-operator.v3.10.0-nightly.20230918
        version: 3.10.0-nightly.20230918
      - name: quay-operator.v3.10.0-nightly.20230917
        version: 3.10.0-nightly.20230917
      - name: quay-operator.v3.10.0-nightly.20230916
        version: 3.10.0-nightly.20230916
      - name: quay-operator.v3.10.0-nightly.20230915
        version: 3.10.0-nightly.20230915
      - name: quay-operator.v3.10.0-nightly.20230914
        version: 3.10.0-nightly.20230914
      - name: quay-operator.v3.10.0-nightly.20230913
        version: 3.10.0-nightly.20230913
      name: candidate-3.10
    - currentCSV: quay-operator.v3.9.0-nightly.20230812
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.9.0-nightly.20230812
          createdAt: 2023-08-12 06:06 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.9.0-nightly.20230812'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.9.0-nightly.20230812
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.9.0-nightly.20230812
        version: 3.9.0-nightly.20230812
      entries:
      - name: quay-operator.v3.9.0-nightly.20230812
        version: 3.9.0-nightly.20230812
      - name: quay-operator.v3.9.0-nightly.20230724
        version: 3.9.0-nightly.20230724
      - name: quay-operator.v3.9.0-nightly.20230723
        version: 3.9.0-nightly.20230723
      - name: quay-operator.v3.9.0-nightly.20230722
        version: 3.9.0-nightly.20230722
      - name: quay-operator.v3.9.0-nightly.20230721
        version: 3.9.0-nightly.20230721
      - name: quay-operator.v3.9.0-nightly.20230720
        version: 3.9.0-nightly.20230720
      - name: quay-operator.v3.9.0-nightly.20230717
        version: 3.9.0-nightly.20230717
      - name: quay-operator.v3.9.0-nightly.20230716
        version: 3.9.0-nightly.20230716
      - name: quay-operator.v3.9.0-nightly.20230715
        version: 3.9.0-nightly.20230715
      - name: quay-operator.v3.9.0-nightly.20230714
        version: 3.9.0-nightly.20230714
      - name: quay-operator.v3.9.0-nightly.20230712
        version: 3.9.0-nightly.20230712
      - name: quay-operator.v3.9.0-nightly.20230711
        version: 3.9.0-nightly.20230711
      name: candidate-3.9
    - currentCSV: quay-operator.v3.7.0-rc.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.7.0-rc.4
          createdAt: 2022-05-17 14:05 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.3.x <3.7.0-rc.4'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.7.0-rc.4
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.7.0-rc.4
        version: 3.7.0-rc.4
      entries:
      - name: quay-operator.v3.7.0-rc.4
        version: 3.7.0-rc.4
      - name: quay-operator.v3.7.0-rc.3
        version: 3.7.0-rc.3
      - name: quay-operator.v3.7.0-rc.2
        version: 3.7.0-rc.2
      - name: quay-operator.v3.7.0-rc.1
        version: 3.7.0-rc.1
      name: preview-3.7
    - currentCSV: quay-operator.v3.8.0-rc.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.8.0-rc.7
          createdAt: 2022-11-29 15:11 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.8.0-rc.7'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.8.0-rc.7
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.8.0-rc.7
        version: 3.8.0-rc.7
      entries:
      - name: quay-operator.v3.8.0-rc.7
        version: 3.8.0-rc.7
      - name: quay-operator.v3.8.0-rc.6
        version: 3.8.0-rc.6
      - name: quay-operator.v3.8.0-rc.5
        version: 3.8.0-rc.5
      - name: quay-operator.v3.8.0-rc.4
        version: 3.8.0-rc.4
      - name: quay-operator.v3.8.0-rc.3
        version: 3.8.0-rc.3
      - name: quay-operator.v3.8.0-rc.2
        version: 3.8.0-rc.2
      - name: quay-operator.v3.8.0-rc.1
        version: 3.8.0-rc.1
      name: preview-3.8
    - currentCSV: quay-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.10.15
          createdAt: "2025-09-25T14:06:05Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.10.15
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.10.15
        version: 3.10.15
      entries:
      - name: quay-operator.v3.10.15
        version: 3.10.15
      - name: quay-operator.v3.10.14
        version: 3.10.14
      - name: quay-operator.v3.10.13
        version: 3.10.13
      - name: quay-operator.v3.10.12
        version: 3.10.12
      - name: quay-operator.v3.10.11
        version: 3.10.11
      - name: quay-operator.v3.10.10
        version: 3.10.10
      - name: quay-operator.v3.10.9
        version: 3.10.9
      - name: quay-operator.v3.10.8
        version: 3.10.8
      - name: quay-operator.v3.10.7
        version: 3.10.7
      - name: quay-operator.v3.10.6
        version: 3.10.6
      - name: quay-operator.v3.10.5
        version: 3.10.5
      - name: quay-operator.v3.10.4
        version: 3.10.4
      - name: quay-operator.v3.10.3
        version: 3.10.3
      - name: quay-operator.v3.10.2
        version: 3.10.2
      - name: quay-operator.v3.10.1
        version: 3.10.1
      - name: quay-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.11.13
          createdAt: "2025-09-02T06:02:15Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.11.13
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.11.13
        version: 3.11.13
      entries:
      - name: quay-operator.v3.11.13
        version: 3.11.13
      - name: quay-operator.v3.11.12
        version: 3.11.12
      - name: quay-operator.v3.11.11
        version: 3.11.11
      - name: quay-operator.v3.11.10
        version: 3.11.10
      - name: quay-operator.v3.11.9
        version: 3.11.9
      - name: quay-operator.v3.11.8
        version: 3.11.8
      - name: quay-operator.v3.11.7
        version: 3.11.7
      - name: quay-operator.v3.11.6
        version: 3.11.6
      - name: quay-operator.v3.11.5
        version: 3.11.5
      - name: quay-operator.v3.11.4
        version: 3.11.4
      - name: quay-operator.v3.11.3
        version: 3.11.3
      - name: quay-operator.v3.11.2
        version: 3.11.2
      - name: quay-operator.v3.11.1
        version: 3.11.1
      - name: quay-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.12.12
          createdAt: "2025-09-23T09:59:30Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.12.12
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.12.12
        version: 3.12.12
      entries:
      - name: quay-operator.v3.12.12
        version: 3.12.12
      - name: quay-operator.v3.12.11
        version: 3.12.11
      - name: quay-operator.v3.12.10
        version: 3.12.10
      - name: quay-operator.v3.12.9
        version: 3.12.9
      - name: quay-operator.v3.12.8
        version: 3.12.8
      - name: quay-operator.v3.12.7
        version: 3.12.7
      - name: quay-operator.v3.12.6
        version: 3.12.6
      - name: quay-operator.v3.12.5
        version: 3.12.5
      - name: quay-operator.v3.12.4
        version: 3.12.4
      - name: quay-operator.v3.12.3
        version: 3.12.3
      - name: quay-operator.v3.12.2
        version: 3.12.2
      - name: quay-operator.v3.12.1
        version: 3.12.1
      - name: quay-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.13.8
          createdAt: "2025-08-28T03:28:40Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.13.8
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.13.8
        version: 3.13.8
      entries:
      - name: quay-operator.v3.13.8
        version: 3.13.8
      - name: quay-operator.v3.13.7
        version: 3.13.7
      - name: quay-operator.v3.13.6
        version: 3.13.6
      - name: quay-operator.v3.13.5
        version: 3.13.5
      - name: quay-operator.v3.13.4
        version: 3.13.4
      - name: quay-operator.v3.13.3
        version: 3.13.3
      - name: quay-operator.v3.13.2
        version: 3.13.2
      - name: quay-operator.v3.13.1
        version: 3.13.1
      - name: quay-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.14.5
          createdAt: "2025-09-22T14:19:49Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.14.5
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.14.5
        version: 3.14.5
      entries:
      - name: quay-operator.v3.14.5
        version: 3.14.5
      - name: quay-operator.v3.14.4
        version: 3.14.4
      - name: quay-operator.v3.14.3
        version: 3.14.3
      - name: quay-operator.v3.14.2
        version: 3.14.2
      - name: quay-operator.v3.14.1
        version: 3.14.1
      - name: quay-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.15.2
          createdAt: "2025-09-22T13:44:32Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.15.2
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.15.2
        version: 3.15.2
      entries:
      - name: quay-operator.v3.15.2
        version: 3.15.2
      - name: quay-operator.v3.15.1
        version: 3.15.1
      - name: quay-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-operator.v3.6.2
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"quay.redhat.com/v1\",\n    \"kind\":
            \"QuayRegistry\",\n    \"metadata\": {\n      \"name\": \"example-registry\"\n
            \   },\n    \"spec\": {\n      \"components\": [\n        {\"kind\": \"clair\",
            \"managed\": true},\n        {\"kind\": \"postgres\", \"managed\": true},\n
            \       {\"kind\": \"objectstorage\", \"managed\": true},\n        {\"kind\":
            \"redis\", \"managed\": true},\n        {\"kind\": \"horizontalpodautoscaler\",
            \"managed\": true},\n        {\"kind\": \"route\", \"managed\": true},\n
            \       {\"kind\": \"mirror\", \"managed\": true},\n        {\"kind\":
            \"monitoring\", \"managed\": true},\n        {\"kind\": \"tls\", \"managed\":
            true}\n      ]\n    }\n  }, \n  {\n    \"apiVersion\": \"redhatcop.redhat.io/v1alpha1\",\n
            \   \"kind\": \"QuayEcosystem\",\n    \"metadata\": {\n      \"name\":
            \"example-quayecosystem\"\n    },\n    \"spec\": {}\n  }\n]"
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:v3.6.2
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.5.x <3.6.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.6.2
          repository: https://github.com/quay/quay-operator
          support: Quay Team
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Project Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        - quay
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:v3.6.2
        version: 3.6.2
      entries:
      - name: quay-operator.v3.6.2
        version: 3.6.2
      name: stable-3.6
    - currentCSV: quay-operator.v3.7.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.7.11
          createdAt: 2023-02-22 10:02 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.3.x <3.7.11'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.7.11
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.7.11
        version: 3.7.11
      entries:
      - name: quay-operator.v3.7.11
        version: 3.7.11
      - name: quay-operator.v3.7.10
        version: 3.7.10
      - name: quay-operator.v3.7.9
        version: 3.7.9
      - name: quay-operator.v3.7.8
        version: 3.7.8
      - name: quay-operator.v3.7.7
        version: 3.7.7
      - name: quay-operator.v3.7.6
        version: 3.7.6
      - name: quay-operator.v3.7.5
        version: 3.7.5
      - name: quay-operator.v3.7.4
        version: 3.7.4
      - name: quay-operator.v3.7.3
        version: 3.7.3
      - name: quay-operator.v3.7.2
        version: 3.7.2
      - name: quay-operator.v3.7.1
        version: 3.7.1
      - name: quay-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: quay-operator.v3.8.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.8.13
          createdAt: 2023-10-17 16:13 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.8.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.8.13
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.8.13
        version: 3.8.13
      entries:
      - name: quay-operator.v3.8.13
        version: 3.8.13
      - name: quay-operator.v3.8.12
        version: 3.8.12
      - name: quay-operator.v3.8.11
        version: 3.8.11
      - name: quay-operator.v3.8.10
        version: 3.8.10
      - name: quay-operator.v3.8.8
        version: 3.8.8
      - name: quay-operator.v3.8.7
        version: 3.8.7
      - name: quay-operator.v3.8.6
        version: 3.8.6
      - name: quay-operator.v3.8.5
        version: 3.8.5
      - name: quay-operator.v3.8.4
        version: 3.8.4
      - name: quay-operator.v3.8.3
        version: 3.8.3
      - name: quay-operator.v3.8.2
        version: 3.8.2
      - name: quay-operator.v3.8.1
        version: 3.8.1
      - name: quay-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.9.16
          createdAt: "2025-09-25T14:31:58Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.9.16
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.9.16
        version: 3.9.16
      entries:
      - name: quay-operator.v3.9.16
        version: 3.9.16
      - name: quay-operator.v3.9.15
        version: 3.9.15
      - name: quay-operator.v3.9.14
        version: 3.9.14
      - name: quay-operator.v3.9.13
        version: 3.9.13
      - name: quay-operator.v3.9.12
        version: 3.9.12
      - name: quay-operator.v3.9.11
        version: 3.9.11
      - name: quay-operator.v3.9.10
        version: 3.9.10
      - name: quay-operator.v3.9.9
        version: 3.9.9
      - name: quay-operator.v3.9.6
        version: 3.9.6
      - name: quay-operator.v3.9.5
        version: 3.9.5
      - name: quay-operator.v3.9.4
        version: 3.9.4
      - name: quay-operator.v3.9.3
        version: 3.9.3
      - name: quay-operator.v3.9.2
        version: 3.9.2
      - name: quay-operator.v3.9.1
        version: 3.9.1
      - name: quay-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: project-quay
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-ui
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-ui.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Ui",
                "metadata": {
                  "name": "ui-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "volumeMounts": [
                    {
                      "mountPath": "/opt/app-root/src/assets/config.json",
                      "name": "config",
                      "subPath": "config.json"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:37:58Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ui
            name: uis.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP UI Operator
        displayName: Keysight WAP UI Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Ui
          url: https://keysight-wap-ui.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-ui@sha256:876a394f5cc2f745091ad2ee9ca0d08404cfb95284222e1f3a8607e193bd986a
        version: 0.1.21
      entries:
      - name: keysight-wap-ui.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-ui
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator-certified-rhmp.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator-certified-rhmp.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator-certified-rhmp.v2.6.1
        version: 2.6.1
      - name: percona-postgresql-operator-certified-rhmp.v2.6.0
        version: 2.6.0
      - name: percona-postgresql-operator-certified-rhmp.v2.5.0
        version: 2.5.0
      name: stable
    - currentCSV: percona-postgresql-operator-certified-rhmp.v2.7.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        version: 2.7.0-cw
      entries:
      - name: percona-postgresql-operator-certified-rhmp.v2.7.0-cw
        version: 2.7.0-cw
      - name: percona-postgresql-operator-certified-rhmp.v2.6.1-cw
        version: 2.6.1-cw
      - name: percona-postgresql-operator-certified-rhmp.v2.6.0-cw
        version: 2.6.0-cw
      - name: percona-postgresql-operator-certified-rhmp.v2.5.0-cw
        version: 2.5.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-postgresql-operator-certified-rhmp
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: aws-efs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-efs-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"aws-efs.managed.openshift.io/v1alpha1","kind":"SharedVolume","metadata":{"name":"sv1"},"spec":{"accessPointID":
            "fsap-0123456789abcdef", "fileSystemID": "fs-0123cdef"}}]'
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/app-sre/aws-efs-operator:9c87267
          createdAt: "2021-12-20T19:51:50Z"
          description: Manage read-write-many access to AWS EFS volumes in an OpenShift
            cluster.
          repository: https://github.com/openshift/aws-efs-operator
          support: https://github.com/openshift/aws-efs-operator/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Shared Volume
            displayName: SharedVolume
            kind: SharedVolume
            name: sharedvolumes.aws-efs.managed.openshift.io
            version: v1alpha1
        description: |-
          ## About the managed application
          AWS provides Elastic File System, an NFS-based shared storage solution. A couple of mechanisms exist to enable mounting such a file system in containers. Curating these options and making them work in a cluster is nontrivial.
          ## About this Operator
          The aws-efs-operator minimizes the barrier to entry for container use of EFS by:
          - Installing the [AWS EFS CSI Driver](https://github.com/kubernetes-sigs/aws-efs-csi-driver) and supporting infrastructure.
          - Reducing the identification of a shared storage resource to a simple Custom Resource Definition with only two fields.
          - Creating the necessary PersistentVolume and PersistentVolumeClaim resources that can simply be referenced from a Pod specification.

          More information can be found in the operator's [source repository](https://github.com/openshift/aws-efs-operator).
          ## Prerequisites for enabling this Operator
          An OpenShift cluster backed by AWS.
        displayName: AWS EFS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AWS
        - EFS
        - NFS
        - Shared storage
        links:
        - name: README
          url: https://github.com/openshift/aws-efs-operator/blob/master/README.md
        - name: Design
          url: https://github.com/openshift/aws-efs-operator/blob/master/DESIGN.md
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/app-sre/aws-efs-operator:9c87267
        version: 0.0.8
      entries:
      - name: aws-efs-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: aws-efs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-application-gateway-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-application-gateway-operator.v25.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMApplicationGateway",
                "metadata": {
                  "name": "iag-instance"
                },
                "spec": {
                  "configuration": [
                    {
                      "dataKey": "config",
                      "name": "test-config",
                      "type": "configmap"
                    },
                    {
                      "discoveryEndpoint": "https://isam.mmfa.ibm.com/mga/sps/oauth/oauth20/metadata/test",
                      "postData": [
                        {
                          "name": "redirect_uris",
                          "values": [
                            "https://isam.mmfa.ibm.com/pkmsoidc"
                          ]
                        },
                        {
                          "name": "client_name",
                          "value": "OperatorTest"
                        },
                        {
                          "name": "enforce_pkce",
                          "value": "false"
                        },
                        {
                          "name": "all_users_entitled",
                          "value": "true"
                        },
                        {
                          "name": "consent_action",
                          "value": "never_prompt"
                        }
                      ],
                      "secret": "oidc-client",
                      "type": "oidc_registration"
                    },
                    {
                      "headers": [
                        {
                          "name": "Authorization",
                          "secretKey": "value",
                          "type": "secret",
                          "value": "githubsecret"
                        }
                      ],
                      "type": "web",
                      "url": "https://raw.github.com/ibm-security/iag-config/master/test/sample1.yaml"
                    },
                    {
                      "type": "literal",
                      "value": "version: \"22.07\"\n\nserver:\n  local_applications:\n    cred_viewer:\n      path_segment: creds\n      enable_html: true\n\nidentity:\n  oidc:\n    discovery_endpoint: \"https://iag-dev.ibmcloudsecurity.com/oidc/endpoint/default/.well-known/openid-configuration\"\n    client_id: 1cbfe647-9e5f-4d99-8e05-8ed1c862eb47\n    client_secret: uPP8rM7N0e\n"
                    }
                  ],
                  "deployment": {
                    "image": "icr.io/ibmappgateway/ibm-application-gateway@sha256:1144b79a2930fe942ce627b0abcc3bed2f0b0770ae8313cb77ffe6772511c1f9",
                    "imagePullPolicy": "Always",
                    "imagePullSecrets": [
                      {
                        "name": "regcred"
                      }
                    ],
                    "lang": "C",
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 8,
                      "periodSeconds": 9,
                      "successThreshold": 7,
                      "timeoutSeconds": 1
                    },
                    "readinessProbe": {
                      "failureThreshold": 2,
                      "initialDelaySeconds": 7,
                      "periodSeconds": 8,
                      "successThreshold": 4,
                      "timeoutSeconds": 5
                    },
                    "serviceAccountName": "iag"
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/ibmappgateway/ibm-application-gateway-operator@sha256:6a16d4ef802a866942f504b11a948565245d303f322944200c1a1269fe7c371c
          createdAt: "2025-02-05T21:15:37Z"
          description: The IBM Application Gateway operator manages the lifecycle
            of IBM Application Gateway containers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM-Security/ibm-application-gateway-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMApplicationGateway is the Schema for the ibmapplicationgateways
              API
            displayName: IBMApplication Gateway
            kind: IBMApplicationGateway
            name: ibmapplicationgateways.ibm.com
            version: v1
        description: "The [IBM Application Gateway (IAG)](https://ibm.biz/ibm-app-gateway)
          image provides a containerized secure Web Reverse proxy which is designed
          to sit in front of your application, seamlessly adding authentication and
          authorization protection to your application.\n\nAn IAG instance deployed
          on Kubernetes can be a complex deployment.  In particular the configuration
          can be defined externally in one or more locations, and changes to this
          configuration may require all instances to be reloaded for the changes to
          take effect. \n\nThe internal Kubernetes deployment controller does not
          have any knowledge of how an IBM Application Gateway instance should behave
          when the configuration changes. As such this IBM Application Gateway specific
          Kubernetes operator is available to be deployed. Once deployed the operator
          can manage the lifecycle of the IBM Application Gateway instances.\n\nSee
          the project [Readme](https://github.com/IBM-Security/ibm-application-gateway-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Application Gateway Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Security
        - Identity and Access
        links:
        - name: IBM Application Gateway product documentation
          url: https://ibm.biz/ibm-app-gateway
        maintainers:
        - email: isamdev@au1.ibm.com
          name: IBM Application Gateway Development Team
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - icr.io/ibmappgateway/ibm-application-gateway-operator@sha256:6a16d4ef802a866942f504b11a948565245d303f322944200c1a1269fe7c371c
        - icr.io/ibmappgateway/ibm-application-gateway@sha256:1144b79a2930fe942ce627b0abcc3bed2f0b0770ae8313cb77ffe6772511c1f9
        version: 25.2.0
      entries:
      - name: ibm-application-gateway-operator.v25.2.0
        version: 25.2.0
      - name: ibm-application-gateway-operator.v24.4.0
        version: 24.4.0
      name: stable
    defaultChannel: stable
    packageName: ibm-application-gateway-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cockroach Labs
      provider-url: ""
    name: cockroachdb-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroach-operator.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v21.1.10",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cockroachdb/cockroachdb-operator@sha256:9fba7cd5fdf589317911925cb69338508bf9de928b1817c6f117c612635bbab3
          createdAt: "2021-10-19T15:47:03Z"
          description: CockroachDB Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:e67299fc0e79cff5b6caf9c0df154180dbaedc57b84f072acf68b7b2c958668d
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ec1672ce8091b677689a31cf09dbde5d1697f4caddae0151040b0fb156722e9d
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:e39bce1d9b08a7d62eb05d53e29fabc668bf6a4e4ae3eb9994eebf5b642ff6cf
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ab47ddd22328e0a1564814f049616c3e3569d24c0a290abee9308392c6a0de23
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:49421968126f9c98499bc0cf0c65729370ab566da3fbd1588c100b20052b972e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:c56f980b87a066d1e65ac8cff3d75040af9bd13093494cf4b4d30b5234d56edc
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:de94036e9cc8618e0493313512d880ae145e94d3bff54ed51b0de66a88da1319
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:3edda46eccce8350440a0ea78dad2e41a3eaa0838fe57945ad5318739e968516
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:0bd22d69db5ac570f30d3379afc67f6c2e46fc5b027c8aab42c3f562085d4672
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:bcf3f976437c4ee52e922aa5d7e382a86a031d780b711c08c6c158dbf2d26cfa
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:0ef0234b6fd42977b8a9eda2c59ca8194cc1c8cf6425b99d7bf8fde8feb826c5
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:87dcdbdc07904b217880e70484582c9d87dde7e4071e01cf9e2e6da43111190e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:2664d78fbe88c368927ddbf80704f430e6ca2ce1f72cb4b7e1906dfe72be7cd0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:fe0c615b4d1ef4ec6e089e85deae2b6bd85410be8ed525e2317c759669417f47
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1b56420753eac1126039b8ad4be82ddb59c89a4ec72fb7b6198cea88a10f3a91
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1ed18af1d3ffc65c24946c817d7d04d5f01025d3fb6c9eb78940e343b989bed0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:162a356cc8b423926634f83a1372b6321104d8de7a2c5b9b7b2681c769a88cc7
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:288ae92ebdfc848540ff80ef682b74e50809e9742cafce22b028112326d66b65
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:867c46bb4d8ca3f2fb8dc94063facc99758344c7b9d04f10e50932023ef5d262
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ba9da9aa9d662c0fa9fd16a40551eedf100869b008cf9c0c09cbb87a09fda697
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:25791b76b0d4b7220dadab3a405b3fdc330ba264ef6c9fbfdfb924c154c4bb5e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:cdab9fc62e07dd349703ae394adb48ab3d2281610df35e081f4b76e208b38c8e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:a5414e54c4c513c7877f925b09eadc4c28eb592d3e120bcf9c65ab10e07cd607
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:98b76cf63aabc39d026317905d6ba72eec3f0d35c0f770ebb8b789c3fa0709ee
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:9f5dfcde260804cd2d6976a6b3544ddabdd899365be2104210dc78ba13755ec3
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:27ef61d9f320a52ee9617f2ab8a1d1a1eacc576a0add2d85d99f9413912b651f
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:38e940144e34b69dfb7d1cf31456e7a0016c5dd40d5626baa346446c7678ceb0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1cea12847a0d236437f7fc28fba476fe512c343cd020d21bb50316e50da6ffd6
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:c1a07a8686bbadcb8c68ffba4e40e1285e581e24d43e5811f38a22d57fb9cc56
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:d290099496d070f135b5cc1e6bd59607512524d399df81698f11cfa09b8dca4a
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:b02549e5c3673b8154441990afa05a478d1ecc2bd0a70af64fd693cb6b469cb0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:5eb59093ee81f8115d82ec6a8a1c35eefad055cbb3a175978cef7afe9196e6a0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:fec6b0f56f2b94f141c8553a63b2e08b9494d4e6c5dc109d73268621ca445c56
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:d9a756df2ed2536b938d7c24fbf3de149cd1a0f44a3a1a0cc5d042fe8980362e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroachdb-operator@sha256:9fba7cd5fdf589317911925cb69338508bf9de928b1817c6f117c612635bbab3
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:6667919a41d304d5d4ade3ded4f11b42d722a995a4283e11e15320529f7f9abf
        version: 2.3.0
      entries:
      - name: cockroach-operator.v2.3.0
        version: 2.3.0
      name: beta
    - currentCSV: cockroach-operator.v2.18.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v25.2.2",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
          createdAt: "2025-07-22T11:30:56Z"
          description: CockroachDB Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9ab968e4ca62f23bf9848e5475ef4076df7318a8560f63d5b2090a32af6fe4e2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:90cf21c36d19d87b97d06edd1c0912a3bb288c769e6159651877db3206aa7355
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8f1f2fd1aca914a1ef007117c85ac944055bbfb873a468a7374d239ef20b065e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9b5a2e0c006eb4be8e1444734a98743551f278a24e81c20bebc96f0f3b631eb0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6371377718325235a2720dba0c97eb4321ee29b31a28cb893f76abd62baa339
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:68fe64fe3e112288ccfc2a7dffd69b9b2786a5da3d3bdedd6addf41659d460c9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5240f9f591263569fad7ee934e81fa8a1617de3c25da94ddd59ee5b2b3407f90
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f29c31bb267451b389608af3e8a4a28d6ca0691e67b77e8f57bb5e461b064893
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8737b67b91983817e3e4ff3f47f0b070bd36162d309b60993529426c6979cb24
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a3ac544d87a9d8e7c92897c4c23f803753c7ab31467db91b5c51244afc2c6be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf001173bbf00cdf443b8d64a3ad724389f1cf83e3628d0b609019c8cddad35a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce3d94a3fbdcde5858beb3c265987759971d991c397c6486e19904bd4706f44d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7d6320ac7fb0925f10cf6f524a13de629cbf13ccc371912e4fe8c777bc15d1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bd75aa918355e77cd671c3e11ee807861378e96245245f4fdf243cba5749f9a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d590946b2f7c7ec0c7ec057a338e81f864d6a9e3d3ac4e0769aa5756a8e13fc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:32fdca575c334822e4356aab36a7ed97b685c065925fe85f1b8ba8425c57159b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc4b75ed71845d5b18e1d9163d480fd9d0d3cb5ebbcfed00a2e4e1b174c0a5de
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f74c8a5dea4560ad59c6f6f6ab1e699f063334ee8db704460f870508dbeaeff0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4b3aa590623a38f435b075de4c09f8a343d8ece8b673e6bf7c6df5d5eb4bdf4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:46d9287ef11e3b21a36abfa31a26da3b93040bb6163b9223972a5dd1930fecff
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e0f21abea34825e55622c432a306f2e8aafd64d913eb78f2f601c51dc878a1da
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2b5c028f293c6f394e58a5495cb3dbfbedc215d0d03778d0217e045586d55a46
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08cec123813304ab03bd66d72a09949667a874a42320bedafda391ebcdb2e56c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6a2ee8d62c586eb74ebbef6bef256e06e1ba5cb51c1ae55b1654adafc2ca8859
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c2a62179e8ca293d7e30b9da599e48795345cf07a72abd3a83063107b53b03cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:72ea3d0ef72972dba93362dc31366c5946a9898f5c16780b877316a97068ef8f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1304d7f871a95c8b087bad1b548217562792edc66c8fe07c621e05242b8cd53e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9492894eaa5f43d00ac80b60029b32a0b0a85ae88ba609b46f98702c9349a5bb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bf36dab9b3257a4bba24d986f1451fc51a3a37023fbb988281f1941fd3faedd
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d467383af41aa80c26172964c34152abeba45121599804e502984655b72179f0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:022c6e1a0de875474bc953718a481a0af526c49dc910094dcdce0b5677aced18
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2d89398f03736f8ffaa82b10fed0f19fba166d3d2145b26f25cb71f3c39acf07
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e23af8f9d32c3f80a52d79ceb4b359dda1d75ba0dd4cb8ed4b66ac2f4e40c69b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:61e602b0e3582bf00460eb9328808ccf66e3ff1b1f6df07ee6c5c76a91846876
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8443c15eb1653dd9e5ec7aad54d4b8b092d41a0977e42b76e372e5e21280746e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0ec1dc184a2199736b4bd7de3241a357be7427caea5da1c931b08bc06f5c3dc0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c9a31bdf19175880d0981e21dd19a1458b6e8cde8d4acbcb554f3bea3578a64e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3f587e7b9f5a611013af2608e2151ed782859a52501b00adbc95739c56a8ff1e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7ad12df8e258b03c4099d375d02d272bd774bf188d108901d5cf6cb865e57ef3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4f753151af4c3d3f5a1df9014461316249a6d0621b8fba5c4430307b3355251
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:58de9d0b2ce005c7ad3e99152100e59b7677398557028a7a59e56e669fabb81c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1f6b1ec92251397dbcf7ed4a82f1f05619d64676aae2717f4eff1c33e4c2c27a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:266165ab87b19a05d6796e8b15ff0eba632d6ac96e6393400464c6136dd500ec
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:086d7b435993fdf21fd59b0093f52105a9028d6b769398b0033f98a65dfb7e79
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc1ab77a9537ccf236ce6dc370940dfedcfaa91b6e3db37bf2d5eade7ff84708
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4d8e804b35b73d7978c112748aea2d2bb68f65081b90bca62c39f2b6e2989378
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:744193037d6b64cb73fecab15ebe022abc2b79ab0335b76ec1415ac6379b7dfa
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5181196fff3bf8a5ec8fcfe4f1958cc69f285f0beb4f0b476b4b5e0912f3db42
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:40f00476b63cb9e370fe85f01407173693213a9242fc117d28b06378ca0d98e0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fd6c9b8ca82ff4f2d0fe29166d1fc9bca6439cb40907be60f4d8f1b9735de2f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:85a0e085901c2a73192556d5d70795d4b66cbcfe01d24c73b6fe6db8d5f2e5ca
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6bb15b36d64eebb6e4c1db5a5466e108b271d53383c58e0b6c78cec214756a9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b83eaff2e217c06098509709cead525b55b00ccd48019d8357e6231892008055
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:858dc82977b3c9e989a3a20626671b0c67c834c97132094a9212bb227ab55929
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da367cf0ac52045002e1c186f8e6964267ad87d5c25f9e72fd2c9b9a98a32702
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b77b80eed7ce149d2a3f228e5f188e76d73f32bc50ecde20bba9cfa7bdc4714c
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:53c7b960a2ed70a0998a4f8e28a4f78f7f676e01cffebe665e91e3a7629d88ed
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1558a1f5eab79f1a8c6c85d81740cc539dc23025ea457c0019d93a72dc05d57a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db4e5c2f27ba23e95a363c8174c01f0f9c6b6a27e0e581c86a38f3a1f68dc43c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:37bf9e1ec7e3eff47755fe7521223f31ec6aefc6e63901db3881b30898cf78f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:35ede65457d845e8c063ad7d9515d3a69f98e0751424ecbac484963678a0b1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:50393ad005fed19618d04f0716c6d55e2f1afce651e785c396638e852cac83b3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cad60044ad6573bd80b65e892a9ec0510dbe4fcbbfc6b51010ecf419f56f1024
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f42fa22aac3cc41ce72e144a72530ffbee765a392b24593dafd06cc098f8410b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:31b161f9bf9028a1a62cd2eef6abda01257d8889bdc3159822642549a75e3429
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9c85c25d5392817d17b68714dca6195ad53255d43e9c65b021d75b12d97e0579
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a6e2fe24ad52913e2160f85b9fe665a4e4c4d79b164b4b899f69dd9d92f999e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cc20270212f01fc231ff788e20b0476b8228c440bfc40430c0836d5eb702a042
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1374cbdc4ecf38e4b4226328f532e830b8932b34d9449bf2fbb36e60d0310521
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bd336715332f776a72aaf5c3ca52e62dd53a9e8521c1feb6e135cc868d694a39
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7481c0cf3b579c8d98d94d4aa0334ca94bf2ccfcd25ad70b57c42b22bbf4bcee
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4635550222775fc75737be5e5c0359303e3a23c4d1e1c296d7408cd15ad94aad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7daaf1368ab2c25694cdea0a598e1de59247b04a4b4b4b8348ed6edbaa194f9d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fe4a40006d1e8439423de1417cc9bc55716d790b882485074bb2a5ec68d7d1ad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5b3de46526d169fcd1318354545b3a18861843fa2022a23ed04d795493825287
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acc083f7dde1fbc557d858d03fc26b68b2056997fc1a457e4695650214d5b303
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0b0e0588c1aadf0c3826cd584bd2b3a7d6781882dcb5c5e037b47ff5cb84509b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3bf017b2b2506724500f5e29017d1b956825dd0f61a7c0049b0ff084c3554b5b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce1209ca6cc8724baee981c97c66e2852e512423bc0c0a16cdd80a6838b54031
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e82684b828256b3c2814930210031b3b70bc818514d4359a25f87ffdbf22be6e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:19f36f53f7da67755eb86da77098c5eeb84b381f9c18aa29ba573793dc498564
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f5a0ccc02dc9e938e484d5b5282ff650d1890d5f754c30a0c0d717989ed5d600
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9d0bfb642090ff55a3cc3da0f72663cae18063622a4ec785ba99bc71ef4db116
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f045fe5037d44aab1a0b2209241b3931ec7cb9897cb3c6a96a9f8d5280e55dd2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:44a0f40a6c46ba08297b482f3d96defe87b181c944803e5381c9acc42cabf97f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db07461824e73410456ef540b0bce3c7fe550aa67a73d6174300ca74b8f75276
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0754f7901feea9017bc5041cb91b7ac1cdbca10c2064f37c3648a2c395f24f83
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:28ae9661b38a72decf413c6a6c3758e61098debf5162197c27a12a070c7a096d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:743cda1eda6e273c798e3405a9a97e07dd6aa5ad4d5a52789be3b0f39a32a039
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:91f7e59f9e35a0395756eafed906963c535e886125294d671d930ce959a8f511
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d78c2e0709214f96794564f13386861f4b541e55b5bd316ad495cf104f7d40b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da3bd2e9ff2b08d5bb86b1f5faa95a865f85c5e8912df00b6b79190c1e518a2f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f3ffebd812881a830ff46c860984814dd61edcd51b478ebd3667759fb7710682
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5d6901b6b6fd6594d42bfc1842d050c7cffcb0867a5591a67469f0f860b1f21b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e8f26fd132affbaeb8ef6b1807a646c83b301e5dc334a592dfb2cf8dc3dbaf10
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c74cf927a2b2c5098ad74ad0880996f8b042186a9911d888a231ef5b9e2a0715
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5a54d564e08baefc2786f4b742b2507d31984664cd00333406aa7a3aa5b9c913
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:310310515625f099a928545865f7096997871ee4a16650a01c76c3799a18b684
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a0faa52a999289b308dc4caf050c6d9fe2f42e112aa30e4af282ed1f21c5c346
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a59cafff8d4ff2c62204ece15880117011c30978823c232902c7182d747ec827
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1e5d85f5db10fa79d09263db90daa44549bcc3b025e9f3f98cf0d4e51766394b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6ca927f137be28481e3a226e6c6bcb7c5ac27664984f17ffc6c1419cd7d08eb7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e50aab39722af22ccbf3d1db19e37972d52b0fc9a40998a64618dc966b2bac57
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d978678b3c254abd52f69f226b0aacc4b02c2aaca328c54ef10f6f9bb453582d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6408f89b0a88ec154f7ee98c314be574971c4c9aaf906b473d10124d822889be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:468b0a40a33be1dd5a2a89ce745d7e9012bc7371f2f01cc0f8f12c0ac887db43
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b06bd7f3802e8fa2af14bb0b07804cad4cc0fbbc97da18d09a9106bf074ed6d8
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f438fbbf9c87ca0d54c170e2c5105b1e265182e1c2ede48adea370e2be9dc895
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d8cfb221f67516675fdfde061f6ec5dfee1224ca7eb3d8be375c27d483f306af
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e151bded670f4cbd165a8ef864324805667d3e84ce3ace0cdf8f9aad087fccf1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cbc74889e9ca5edbd5aaa243d7649bca0c4da2719239f8d0ec9827868112a771
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a6f1c43486d044fed9f59406118f429c8037b389bd8c8d7a1c7aeb89ed2661f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2af60025ecb3bb933b61328be4c2b2bfd0e7d26f53b72430629208505e45c6d2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:79d688bebe0643ef0dd6ba7108cdb34c0442470c0c20ae0df88b5d02a3476cd0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9097e135159f46396ef4649f724d61bb5b9113c488f989d404586fbb7486fd85
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d14f77f70aea0422a906510c1ac2652ff76e131ed705cb002f80e13e887b0cd9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0fea6ed8a78ea624240ee6ffb3573d5cf1115186aa9180fcd7d1273351deaaa3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e18d73af3257c8cdf00245f2f718b56667524c1cc69d3db28e9b868a539ae16f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:dd521d7bad62da054eec54179be75c643a5492d981e11c0e66db41902cd2182c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:363d8feb560324a848c5771edd224dc19020013af7641979ee81ae8ee536d7e3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7b907d03b001b758d8a4b5e898725c73dda2ec2aa4b7425b685320b947154d11
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4b56c0311af14220a466aa11f25fd10fa38280825539bd02215c945bbafb861c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:42c9930d6861f6a2147ed6e01827eddd730279966763ec68868567a19cc164ab
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:740ba10f1ef0a6c5fa6ad31ab18855fc75f8e024ce4c23b0676c5a9c5b62cd75
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8e8c081c231b5b15eca483ceec16ea9190ba73aae6e421865162446eb584519a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:56109e57ee0379cf48644bcf8226a9238e01139cacc5499002c99f973f121911
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:742a6f69a3a209a7080a1e911fd96467d2a91aacb3f91d6be80a6a6e2e54b2ae
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0e2c7023f1f97acc20c46a4944635a9c5f57bb9bb3dd534cf41fea1a9cfdfdbe
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf7d3bea47638b763b312c7c11ca034a7ac62d993d75a13edceec74799b97e52
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:226d7f1c3f0d70ac6532ad9490bebcdef44a40959b6aa7c11b0f2d1383a46217
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6cd9f9b9fcb6e8f9a931a6ad01cb66329a74d40b22f9029fbc50e57248bad7cb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1ce71e80c0cd0e0063eba275ab91f77982f2e92a8aef8e89cab99b93feab0a88
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1b6d934a7b6c42a61e5ad51380e69af341a575b42205299fd58bc0c6ff306f6a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b046aa0a42dae992a03685ed82a20835532d6d4bf048e2366ff7382be4b3138
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acbfee4492327a720fe7da76929a86192d8af2442bbb76f0a089f6a81a890be3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:818f592c50d4ed5599faa053344f7ecd963003352f1faa3b72dc0f5e00a0e78b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4e5f7df1dc1e1db398c36d590431e7e5782897b209972d8e9e4671971c10d1b6
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5458a8fd89aa664b60e59c35bd933975cd37d18e19b11589e4ca0c2844ea0ed4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2a128b570352a5a7f82b27353f9b774caccce83e0e7bb2673e4802d96ad0bd51
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2171e1c75d0c665fe33e1ea334e605604c51de6dbc22d88a3a0ac604259c448f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ec6eb7c28c213cc83b2d7919cd87988f9a07f12276eb7351d0915f1567a5b095
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e5814748c5ab229a1dea5662a007063c0df06fbbfcfe041fd9a6df9ef67e07cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:82d8104a37349c513447e68ea3790efb256370632353176e5fc13c2e8df1bd44
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b78a4d3864e16d270979d9d7756012abf09e6cdb7f1eb4832f6497c26281099
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9f42e929f2cc7e9b6d24dd61e19c5d93dcfeabd34b9c8b935afbe78577efb18b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8fc8de4f3671a73f9aa747e5b23ec973d192fff6f2869c38985cd656063c1e3d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:02e1ef01a023da0eee653f63b314786fd66a78d184a97a314422f6f9c3ef8e7e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:65923c37fecda2d3174212fa9d9a6f34241d065389336ea69a7be99ba16cedc2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bed27f3f2c2d7627d1af4500d7ea0302da5c3fd3d875a6480fb6d98da481312e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:10b9df672260c023dfb37166e156c0d7e0e13144dc3e485af922e8151fd3ab05
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f05824bb8eaa5a93e54059449eaa236dd83d729049ad00065ef3242f5f74c91e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:87db6e079b96fb392d98ef5fc7063836bff10717efd75061016ce688be0ba38e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:29abb169ffefe36f6d69edd438a7135ee23059cedab8f87ad3b2d52361b96b35
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a4f932c3e9ff3aeb70ba1f7a92aa4642bd14cbb7d37d04ff750ed517b916cdb9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08496b84b3cc691909227be17e67bcd6f4d5185a93712f44d8a0985bd2c5fcd7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f53b07212320c6939d3bb4ddd4e6a0bce311ab5b63858eb1a8728531f0dbd510
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b0057de903654ea0315a2c96affcc8f50a92ede8ad121ed30edb3aaf0fb7b554
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d0014b1c19890b9e8ef1a4f3efe7eb5b269be5610ba6ababd4258b52d2f85e8a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:aeac9cd8e8acdbd2b0c339c48b85079c159b04f0ade24df8e3f9cae3f48b4c70
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6366ba29f301a82215c05e60d78e1de66bcee85440ce2fc3c692712d4da32685
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8a70ba12f610ac0e62066e417a0833bc1d01e91ac99d0dd794a2c0cd2a350164
        version: 2.18.2
      entries:
      - name: cockroach-operator.v2.18.2
        version: 2.18.2
      name: stable
    defaultChannel: stable
    packageName: cockroachdb-certified-rhmp
    provider:
      name: Cockroach Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      control-plane: odfmo-controller-manager
      full_version: 4.15.4-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-compose-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-compose-operator.v4.15.4-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "labels": {
                    "control-plane": "odfmo-controller-manager"
                  },
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              }
            ]
          capabilities: Basic Install
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.4-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          This is a CSV used for mirroring the entire product properly. Do not install it please.
        displayName: ODF Compose (DO NOT INSTALL)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: branto@redhat.com
          name: Boris Ranto
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-operator-bundle@sha256:3773158dbd141530cd851f36c3c8a5f54dc0892d003633118e123f5862923666
        - registry.redhat.io/odf4/odr-cluster-operator-bundle@sha256:4c7049b6137b3d49fa9b5e51e0cf7cd430d70c8790e0ed017e6bd811fcc506c8
        - registry.redhat.io/odf4/odf-csi-addons-operator-bundle@sha256:b2b8144c34df667b13987773b11df2fc8ac96d418999292303ef2c03cd3c79d6
        - registry.redhat.io/odf4/mcg-operator-bundle@sha256:9ce0ab75f56ebeeac63aefd0f48cba671da3377ac628830b3c03be9569f8c87a
        - registry.redhat.io/odf4/ocs-operator-bundle@sha256:1e91c5c6c122fdccab52c37ff2cef76317af80970dbdc06aec752115117b2fa0
        - registry.redhat.io/odf4/ocs-client-operator-bundle@sha256:8904c171ea317249e9a19c3a7638f688357a0b238582330b1bddc23a11a9925e
        - registry.redhat.io/odf4/odf-multicluster-operator-bundle@sha256:f7e1c75ee6b1a62fddbcc1e6a718a9cbd5ecf364e4ef10f29da7a15bc82d5b17
        - registry.redhat.io/odf4/odf-multicluster-rhel8-operator@sha256:861e56167698558109197b515aff93691f8d0fdba19311be18a32512f6896139
        - registry.redhat.io/odf4/odr-hub-operator-bundle@sha256:862a9d983013c16b7dfd5194ab5752c26a473ca5a4575bbcd80be59f97f8ac27
        version: 4.15.4-rhodf
      entries:
      - name: odf-compose-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      name: stable-4.15
    defaultChannel: stable-4.15
    packageName: odf-compose-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: unsupported
      operatorframework.io/arch.ppc64le: unsupported
      operatorframework.io/arch.s390x: unsupported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: mta-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mta-operator.v6.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:f5d10cd2528bebf5094c7744e47d9de6fe827fd09024c5ffc0e1b74f1a6cc3b4
          createdAt: "2023-04-26T14:04:20Z"
          description: MTA contains tools that support the modernization and migration
            of applications to Kubernetes and OpenShift
          olm.skipRange: '>=0.0.0 <6.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/tackle2-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is a collection of tools that supports large-scale application modernization and migration projects to Kubernetes and OpenShift.


          Migration Toolkit for Applications allows users to maintain their portfolio of applications with a full set of metadata and to assess their suitability for modernization leveraging a questionnaire based approach.


          ### Install

          By default, MTA requires a RWX capable storage class for caching purposes, this can be disabled via the rwx_supported parameter in Tackle CR.

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-windup-addon-rhel8@sha256:a4200fc0a39df67eae31c53ee9ae0c05769fb410d0312ebd0bbf8bf1567b2a48
        - registry.redhat.io/mta/mta-hub-rhel8@sha256:c75e1d9d3307d95be392a53440d9123601a6ada3693e3f2573e03022565e4339
        - registry.redhat.io/rhel8/postgresql-12@sha256:3a69efa85af6a6f95665eca69085b007fa1fcb645a37f356895aeb4c9527a621
        - registry.redhat.io/mta/mta-pathfinder-rhel8@sha256:a66dd030eb99207b6aaacc6b7c7cb5182901da836ebf50373c14f8e5ecb7b642
        - registry.redhat.io/mta/mta-ui-rhel8@sha256:de6cb95ce3663922ecc725b186876b043a9c7794e95510459ffbcba7d668fc26
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:f5d10cd2528bebf5094c7744e47d9de6fe827fd09024c5ffc0e1b74f1a6cc3b4
        - registry.redhat.io/mta/mta-admin-addon-rhel8@sha256:812b8dc621a06bb99cf2c59abb8257da3ceeb165c563e14d258e69803a557cc8
        version: 6.0.3
      entries:
      - name: mta-operator.v6.0.3
        version: 6.0.3
      - name: mta-operator.v6.0.2
        version: 6.0.2
      - name: mta-operator.v6.0.1
        version: 6.0.1
      - name: mta-operator.v6.0.0
        version: 6.0.0
      name: stable-v6.0
    - currentCSV: mta-operator.v6.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:6da71d4caecfbe049e63576d4325d4919002196da2e875f79703d8ff0d5a1705
          createdAt: "2023-10-26T15:29:26Z"
          description: MTA contains tools that support the modernization and migration
            of applications to Kubernetes and OpenShift
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0 <6.1.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/tackle2-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is a collection of tools that supports large-scale application modernization and migration projects to Kubernetes and OpenShift.


          Migration Toolkit for Applications allows users to maintain their portfolio of applications with a full set of metadata and to assess their suitability for modernization leveraging a questionnaire based approach.


          ### Install

          By default, MTA requires a RWX capable storage class for caching purposes, this can be disabled via the rwx_supported parameter in Tackle CR.

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:6da71d4caecfbe049e63576d4325d4919002196da2e875f79703d8ff0d5a1705
        - registry.redhat.io/mta/mta-hub-rhel8@sha256:32624f2bddfe8b6b9c6c1c3f2be541c8c61be29680d13b6d446e8874bdf8e219
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/mta/mta-pathfinder-rhel8@sha256:1df881da1e20c2501caf53e1cf59b6b366281d842f0948ca5f91694ec1db829d
        - registry.redhat.io/mta/mta-ui-rhel8@sha256:3845d6d246003ebeb49bcf1d95d6be2eb9cd05a280a5bf8320b201e878a41427
        - registry.redhat.io/mta/mta-windup-addon-rhel8@sha256:6bcaf8e8b63e169652d6568d77e91a289084e621a586105a2d429b5726569058
        version: 6.1.4
      entries:
      - name: mta-operator.v6.1.4
        version: 6.1.4
      - name: mta-operator.v6.1.3
        version: 6.1.3
      - name: mta-operator.v6.1.2
        version: 6.1.2
      - name: mta-operator.v6.1.1
        version: 6.1.1
      - name: mta-operator.v6.1.0
        version: 6.1.0
      name: stable-v6.1
    - currentCSV: mta-operator.v6.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:6938c3fc6dd8fb032e1765fc11d9586dbfd1aaba4b423b0d87da4120ee0bb60d
          createdAt: "2024-06-10T22:12:58Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0 <6.2.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-windup-addon-rhel9@sha256:d1f4939b92dd13c15cb7344ec47a571669f6d108dc2e7d795ce29350f2cb3838
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:6938c3fc6dd8fb032e1765fc11d9586dbfd1aaba4b423b0d87da4120ee0bb60d
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:e59725832547e80cab58077a1b6e42c313a4057621defa841f35c0a7627f149d
        - registry.redhat.io/rhel8/postgresql-12@sha256:c4716bee5e2cf2e43698bc99e5aa42f16a0f2d68c1fbdcd3602696bf88e408dc
        - registry.redhat.io/mta/mta-pathfinder-rhel9@sha256:a2a8b6d204d99494a6a005e8cb1fbd92afdc84dc01e4c1de4ad7d8798bd83741
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:c35f3ed429a7d351cb8c2d07600b4ad35615a5ba08e5cc9f8b9bd504aea3f7ee
        version: 6.2.3
      entries:
      - name: mta-operator.v6.2.3
        version: 6.2.3
      - name: mta-operator.v6.2.2
        version: 6.2.2
      - name: mta-operator.v6.2.1
        version: 6.2.1
      - name: mta-operator.v6.2.0
        version: 6.2.0
      name: stable-v6.2
    - currentCSV: mta-operator.v7.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:268ecf5efc51af71f9b2bc2d01b410f8cfcf81a948b82d2d476fef6c0cd3e033
          createdAt: "2024-05-10T19:37:30Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:268ecf5efc51af71f9b2bc2d01b410f8cfcf81a948b82d2d476fef6c0cd3e033
        - registry.redhat.io/mta/mta-windup-shim-rhel9@sha256:e2ea89055936d7ad229b5e890c6ba234c6cb22d8b688515ed4d6a34848efea83
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:d554cc36748fc05e30eb10e6fe816ff6aab2fe07b61cc60e9d93de5c3cf17748
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:0e1ec99af5666a060176adc52612f20209e3398a9a5a2aa1dcff11ef260b2006
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:1b5e016046d93b6a2674d63f08e0a46f8050e785bd58fd94ba7a70487efe7e2a
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:18c611415589a093e16e9c2f2fba9f3c83d9514f79c8e4fde9704ef0679ea4d5
        - registry.redhat.io/mta/mta-analyzer-lsp-rhel9@sha256:0c8f59f35a950ea8b45f16cf71362128ccc335b58a996a099384fd737dd6ff90
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:af879d8b9b5e396a5308fb675a1d6dba6d70151b9bdbe7829362e081a6adab1c
        - registry.redhat.io/rhel8/postgresql-12@sha256:da619f6617d99790b4fe00fdb23ab4aea0cdb0592590162d51b3ef94aca01a7b
        version: 7.0.3
      entries:
      - name: mta-operator.v7.0.3
        version: 7.0.3
      - name: mta-operator.v7.0.2
        version: 7.0.2
      - name: mta-operator.v7.0.1
        version: 7.0.1
      - name: mta-operator.v7.0.0
        version: 7.0.0
      name: stable-v7.0
    - currentCSV: mta-operator.v7.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:08b33b44140397785f8aa1b736796e09cf6e49ac30cf7c386a2fbd8b678d47b1
          createdAt: "2024-09-16T23:45:42Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.1.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:16cb28d683159ba71a5d7c233404ac1d6b7a07a374b46d24effe551a13a596c3
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:23c97ac122eb7cd1e953df8da968da420f6db8708baf772759517aca4929748b
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:e4e70201df23c9e1168646d3c6013a8ffa22481a85b07df491f240310c13ebe1
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:08b33b44140397785f8aa1b736796e09cf6e49ac30cf7c386a2fbd8b678d47b1
        - registry.redhat.io/rhel8/postgresql-15@sha256:16faf1ee82382263408f713ee4002f47ff13c82d4545e8a69d5b0ebd1d289330
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:30a3b1d5a70dec74a9253ca34f18bd8f2cdfaa68b85b0b25dc395158c2596330
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel8@sha256:582f28e609c1a0466d30c5586d6b6023f1b1de3c28f76cabb498322914ab24d4
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:ac38f1c3d34bc718928eb20743835c99783d846f63a3197d13839ea22f1634c8
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:4ad19b50cdd6be21e010155f2026abf0dc865e72aac1353e35bbb5eeb8320a4e
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:ee2c8ce2af0be097353758d9ee7eaa7e6b3fcd5f2bf5e04280e450e2b619d2d8
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        version: 7.1.1
      entries:
      - name: mta-operator.v7.1.1
        version: 7.1.1
      - name: mta-operator.v7.1.0
        version: 7.1.0
      name: stable-v7.1
    - currentCSV: mta-operator.v7.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:917d6cd534c158f6c4a7e9c83f0c66fa85ba81037c352d8a039e198581126a7f
          createdAt: "2025-06-17T16:04:15Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.2.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:917d6cd534c158f6c4a7e9c83f0c66fa85ba81037c352d8a039e198581126a7f
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel8@sha256:c0643c05833dd1b70ed361bf56c46f56ed6536aac3106b92f0d85b83da1b9d15
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:10ef4b218d603432508b5e1f3685fb32a80cbce6c327e35952cdcccb0465468b
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:9d6c02f6463771e747a30126de95fc0728fd95fff4ddfa6c70d828c55e0b6790
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:2dfc51d5398a90e73273b3e332f70d7ed94bac05e12b142fba22104ecc929436
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:b10fdc78c6f453a34a3d306313dc29fcb3152688a6a1d54cd236d00d9fab88af
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:4cde2c2df482293bd4d6b86a344336e13d2f9f81b29323698f603e656b14b190
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:70ff3b5e429a8fe0c68752ee05c47ead885b59bf9c8d7235337801270e7fc1de
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:c9447659195b9eebb8f8a265fd5052ca9b2611d57080017448297a33409863c6
        - registry.redhat.io/rhel8/postgresql-15@sha256:dd4aec2775da1cf187c0839384cb16d69268e9781d4330874bbc865f98153a7a
        version: 7.2.3
      entries:
      - name: mta-operator.v7.2.3
        version: 7.2.3
      - name: mta-operator.v7.2.2
        version: 7.2.2
      - name: mta-operator.v7.2.1
        version: 7.2.1
      - name: mta-operator.v7.2.0
        version: 7.2.0
      name: stable-v7.2
    - currentCSV: mta-operator.v7.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel9-operator@sha256:6abd6541a955543207c1ff4b903928debcc89dd6ea3d7aeaa3e47465849f7b37
          createdAt: "2025-08-15T16:32:16Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.3.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-15@sha256:854710b428c4c1132c448cbe9ad30087a8a4b3c33ae9c0c3a4de21a15c0ae642
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel9@sha256:9748d954ad8654ac203257286497ccf19cfdfa211fee40449cbe732ff48c1176
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:32328fc0459099ef1de0f92e064c15d29a46d88dfe2f54c88f822d44ab82e3f4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        - registry.redhat.io/mta/mta-rhel9-operator@sha256:6abd6541a955543207c1ff4b903928debcc89dd6ea3d7aeaa3e47465849f7b37
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:b2ff154a4b4ead22311005b7115c02346928b5ebf77d776ef909ad4af1f350bc
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:6118c1db15667851f26eb2acf6060e045ed9065459e3e649bbdd0e2625d825bd
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:072f676323e9abcb2ff22eb5e7a70f6681e9d885d8666ddd81060fbad54c443e
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:8ca00460b2fb64583a247f5b906c19743b7732b60ff8215f1baec12a766c61d1
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:80aa13813fe121da49e166d81d09d63749393f3375c3b96d762d0b09c6685ca9
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:ef57a58e609e244333d35681a164a1d3bf382c0eee37d33c993b018439814198
        version: 7.3.2
      entries:
      - name: mta-operator.v7.3.2
        version: 7.3.2
      - name: mta-operator.v7.3.1
        version: 7.3.1
      - name: mta-operator.v7.3.0
        version: 7.3.0
      name: stable-v7.3
    - currentCSV: mta-operator.v8.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel9-operator@sha256:83c2b804667d28f782d64132a0c393194083d10c820b88df49c2503d3351fd8b
          createdAt: "2025-09-26T03:03:31Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.3.0 <8.0.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Schema
            displayName: Schema
            kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-platform-addon-rhel9@sha256:dcb35178810b7eddc2f2713434d377003f8f4fb7022a40f6c1926b74b13fe23a
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:a77ee9b76225b6cca667fa3402492b0845587fe5e400fa8d141b5ab636c56811
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel9@sha256:2a6f0b392b86fe16244d5bf9fcb84ed9770a22329a96e9861f03872b7209fb04
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ef10b2ecbfa0c4ac8568af9ee64ee4a554e3fdc9089ee5683b6c2c3a74caafbe
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:616d04cfbab92341157c42eee0683f31867d81c5a289723da8da8208d1054774
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:aaf7bf10b1af107ad55289caab2cf9dd052e69211968d2e8d05c1d25f21ed2b3
        - registry.redhat.io/mta/mta-rhel9-operator@sha256:83c2b804667d28f782d64132a0c393194083d10c820b88df49c2503d3351fd8b
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:b1b38e218ea37b529f344eaa1f0f3b55b716ad7953e852c19f34f4a25ab6c5a7
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:af2c1790f2b98ebcfadb5d6bdd43cb9e7e96cdabd5b653d48d0fb67901781c49
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:460d1d45ed44ae9328ed93c5fed240e154549bbec662dc1fc55569ff7256aca9
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:8ae0887fa27d6f4281e5bb9b629bf6b64a9c460948a002839ba42da9b25b5521
        - registry.redhat.io/mta/mta-solution-server-rhel9@sha256:e9d48d11ae203108229d67bc96721daa1c3bb10160b9d6d957cd2e0ba68a2ef7
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        version: 8.0.0
      entries:
      - name: mta-operator.v8.0.0
        version: 8.0.0
      name: stable-v8.0
    defaultChannel: stable-v8.0
    packageName: mta-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubemq.io
      provider-url: ""
    name: kubemq-operator-marketplace-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubemq-operator.v1.7.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
              "apiVersion": "core.k8s.kubemq.io/v1beta1",
              "kind": "KubemqCluster",
              "metadata": {
                "name": "kubemq-cluster"
              },
              "spec": {
                "replicas": 3
              }
            },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqDashboard",
                "metadata": {
                  "name": "kubemq-dashboard"
                }
              },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqConnector",
                "metadata": {
                  "name": "kubemq-bridges"
                },
                "spec": {
                  "type": "bridges",
                  "config": "bindings: null"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/kubemq/kubemq-operator:v1.7.2
          createdAt: 2020-03-11 12:00:00+00:00
          description: Kubernetes Messaging Platform
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/kubemq-operator-marketplace-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/kubemq-operator-marketplace-rhmp/support?utm_source=openshift_console
          support: support@kubemq.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kubemq Cluster
            displayName: Kubemq Cluster
            kind: KubemqCluster
            name: kubemqclusters.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Cluster Dashboard
            displayName: Kubemq Dashboard
            kind: KubemqDashboard
            name: kubemqdashboards.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Connector
            displayName: Kubemq Connector
            kind: KubemqConnector
            name: kubemqconnectors.core.k8s.kubemq.io
            version: v1beta1
        description: |2-

          ## Kubernetes Messaging Platform
          A message-based architecture for microservices providing flexible connectivity for hybrid environments: multi-cloud, on-premises, and at the edge. Powered by enterprise-grade message broker and message queue with ready to use connectors, bridges, and control center.

          Messaging platform components:
            - **KubeMQ Cluster** - Enterprise-grade message broker and message queue, scalable, high available, and secured.
            - **KubeMQ Targets Connector** - connects a KubeMQ cluster with external systems and cloud services.
            - **KubeMQ Bridges Connector** - bridge, replicate, aggregate, and transform messages between KubeMQ clusters.
            - **KubeMQ Sources Connector** - transform external systems and cloud services data to KubeMQ cluster messages.
            - **KubeMQ Dashboard** - monitor KubeMQ cluster metrics.

          ### Kubernetes Native
          Innovative and modern message queue and message broker in a lightweight container developed to run in Kubernetes, certified in the CNCF landscape, and connect natively to the cloud-native ecosystem.

          ### Ease of use

          Simple deployment in Kubernetes in less than 1 minute. Developer friendly by simple to use SDKs and elimination of the many developers and DevOps-centered challenges to define exchanges, brokers, channels, routes, and predefined topics

          ### All messaging patterns

          KubeMQ is available with all messaging patterns such as Queue, Stream, Pub/Sub, and RPC, Saving the need to maintain multiple messaging platforms for different use cases.

          ### Hybrid Environments

          KubeMQ enables the microservices from multiple environments to communicate and build one hybrid infrastructure solution across clouds, on-prem, and at the edge.

          ### KubeMQ Cluster key features

          - All-batteries included Messaging Queue Broker for the Kubernetes environment
          - Deployed with Operator for full life cycle operation
          - Blazing fast (written in Go), small and lightweight Docker container
          - Asynchronous and Synchronous messaging with support for  `At Most Once Delivery` and `At Least Once Delivery` models
          - Supports durable FIFO based Queue, Publish-Subscribe Events, Publish-Subscribe with Persistence (Events Store), RPC Command and Query messaging patterns
          - Supports gRPC, Rest and WebSocket Transport protocols with TLS support (both RPC and Stream modes)
          - Supports Access control Authorization and Authentication
          - Supports message masticating and smart routing
          - No Message broker configuration needed (i.e., queues, exchanges)
          - .Net, Java, Python, Go, and NodeJS SDKs

          ### KubeMQ Targets Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major services types (databases, cache, messaging, serverless, HTTP, etc.)
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-targets)

          ### KubeMQ Bridges Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Build Any Topology - connects KubeMQ clusters in 1:1, 1:n , n:1, n:n
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-bridges)

          ### KubeMQ Sources Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major messaging services such as Kafka, ActiveMQ, SQS, and more.
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-sources)

          ## Getting Started

          [Start with OpenShift](https://docs.kubemq.io/getting-started/create-cluster/openshift)

          ## Get Enterprise Trial License

          [Register](https://account.kubemq.io/login/register)

          ## Request a Demo

          [See KubeMQ in Action](https://kubemq.io/request-a-demo/)
        displayName: Kubemq Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - messaging
        - queue
        - broker
        - pubsub
        - rpc
        links:
        - name: Learn more about Kubemq
          url: https://kubemq.io/
        - name: Register for free license
          url: https://account.kubemq.io/login/register
        - name: Contact Sales
          url: https://kubemq.io/contact-sales/
        - name: Documentation
          url: https://docs.kubemq.io/
        - name: Get Help
          url: https://kubemq.io/contact-support/
        maintainers:
        - email: support@kubemq.io
          name: kubemq
        maturity: alpha
        provider:
          name: Kubemq.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-grafana:v4.3
        - registry.marketplace.redhat.com/rhm/kubemq/kubemq-operator:v1.7.2
        - registry.marketplace.redhat.com/rhm/kubemq/kubemq-uni:v2.2.12
        - registry.redhat.io/openshift4/ose-prometheus:v4.3
        version: 1.7.8
      entries:
      - name: kubemq-operator.v1.7.8
        version: 1.7.8
      name: alpha
    defaultChannel: alpha
    packageName: kubemq-operator-marketplace-rhmp
    provider:
      name: Kubemq.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: crunchy-postgres-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/crunchy-postgres-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/crunchy-postgres-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-exporter@sha256:3b231902b2116cd38de05c7378202716c95813b9961fb8cee811b61d36cddf86
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:0e283616c2d32e53258d62d02702bc4e179ef8c03933592154babed1a70bf666
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:a7f62f367817fc08294a0b8b82ac9826a42c0ae668efdfacaa158a4287b8a688
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:c559c896e174674d151f93543ac92a1b5647eb4676313f7dc6cbb8faee786f18
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:56a5e3a5758982e086c9f78201a7137e0d934316f058a31cf30df44639a569a2
        - registry.connect.redhat.com/crunchydata/crunchy-pgadmin4@sha256:0d84acb1ce6a9b046e3f8b222a583ddcc91dda0a117c9f637ef173d0ac992f57
        - registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:b2c3c74fdcc0e08f555e50c527f0d322292333c5d3dce061a0dee3680a1fbc56
        - registry.connect.redhat.com/crunchydata/crunchy-pgbackrest@sha256:b70da0035b74379f2507cfabdd9d1cc1e3c3eecb9066c22f292981c9a27916e1
        - registry.connect.redhat.com/crunchydata/crunchy-pgbouncer@sha256:17cb58dec78447c2f70a414a2aeff6facbf921b42a0f366d0dbf7be95aa8987a
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:4bdd0588ffbb3f628e7df1c72670e7a5087269438343b68779012bc3a13dea31
        - registry.connect.redhat.com/crunchydata/crunchy-upgrade@sha256:68bc93cce61d9b1e44822ae3e3ac8f7a9c2b712eddfaea090eabefacd60c67f6
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: crunchy-postgres-operator-rhmp
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-networkfirewall-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-networkfirewall-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "Firewall",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "FirewallPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "RuleGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/networkfirewall-controller:1.1.1
          createdAt: "2025-09-19T21:03:12Z"
          description: AWS Network Firewall controller is a service controller for
            managing Network Firewall resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FirewallPolicy represents the state of an AWS networkfirewall
              FirewallPolicy resource.
            displayName: FirewallPolicy
            kind: FirewallPolicy
            name: firewallpolicies.networkfirewall.services.k8s.aws
            version: v1alpha1
          - description: Firewall represents the state of an AWS networkfirewall Firewall
              resource.
            displayName: Firewall
            kind: Firewall
            name: firewalls.networkfirewall.services.k8s.aws
            version: v1alpha1
          - description: RuleGroup represents the state of an AWS networkfirewall
              RuleGroup resource.
            displayName: RuleGroup
            kind: RuleGroup
            name: rulegroups.networkfirewall.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Network Firewall resources in AWS from within your Kubernetes cluster.

          **About Amazon Network Firewall**

          AWS Network Firewall is a stateful, managed, network firewall and intrusion detection and prevention service for your virtual private cloud (VPC) that you create in Amazon Virtual Private Cloud (Amazon VPC).
          With Network Firewall, you can filter traffic at the perimeter of your VPC. This includes filtering traffic going to and coming from an internet gateway, NAT gateway, or over VPN or AWS Direct Connect. Network Firewall uses the open source intrusion prevention system (IPS), Suricata, for stateful inspection. Network Firewall supports Suricata compatible rules. For more information, see [Working with stateful rule groups in AWS Network Firewall](https://docs.aws.amazon.com/network-firewall/latest/developerguide/stateful-rule-groups-ips.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Network Firewall
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networkfirewall
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Network Firewall Developer Resources
          url: https://aws.amazon.com/network-firewall/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: network firewall maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/networkfirewall-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-networkfirewall-controller.v1.1.1
        version: 1.1.1
      - name: ack-networkfirewall-controller.v1.1.0
        version: 1.1.0
      - name: ack-networkfirewall-controller.v1.0.13
        version: 1.0.13
      - name: ack-networkfirewall-controller.v1.0.12
        version: 1.0.12
      - name: ack-networkfirewall-controller.v1.0.11
        version: 1.0.11
      - name: ack-networkfirewall-controller.v1.0.10
        version: 1.0.10
      - name: ack-networkfirewall-controller.v1.0.9
        version: 1.0.9
      - name: ack-networkfirewall-controller.v1.0.8
        version: 1.0.8
      - name: ack-networkfirewall-controller.v1.0.7
        version: 1.0.7
      - name: ack-networkfirewall-controller.v1.0.6
        version: 1.0.6
      - name: ack-networkfirewall-controller.v1.0.5
        version: 1.0.5
      - name: ack-networkfirewall-controller.v1.0.4
        version: 1.0.4
      - name: ack-networkfirewall-controller.v1.0.1
        version: 1.0.1
      - name: ack-networkfirewall-controller.v1.0.0
        version: 1.0.0
      - name: ack-networkfirewall-controller.v0.0.10
        version: 0.0.10
      - name: ack-networkfirewall-controller.v0.0.9
        version: 0.0.9
      - name: ack-networkfirewall-controller.v0.0.8
        version: 0.0.8
      - name: ack-networkfirewall-controller.v0.0.7
        version: 0.0.7
      - name: ack-networkfirewall-controller.v0.0.6
        version: 0.0.6
      - name: ack-networkfirewall-controller.v0.0.5
        version: 0.0.5
      - name: ack-networkfirewall-controller.v0.0.2
        version: 0.0.2
      - name: ack-networkfirewall-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-networkfirewall-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: opdev.github.io
    name: spark-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spark-helm-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apachespark.apache.org/v1alpha1",
                "kind": "Spark",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/install-disable-crds": "true"
                  },
                  "name": "spark-operator"
                },
                "spec": {
                  "affinity": {},
                  "batchScheduler": {
                    "enable": false
                  },
                  "controllerThreads": 10,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/googlecloudplatform/spark-operator",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "ingressUrlFormat": "",
                  "istio": {
                    "enabled": false
                  },
                  "labelSelectorFilter": "",
                  "leaderElection": {
                    "lockName": "spark-operator-lock",
                    "lockNamespace": ""
                  },
                  "logLevel": 2,
                  "metrics": {
                    "enable": true,
                    "endpoint": "/metrics",
                    "port": 10254,
                    "portName": "metrics",
                    "prefix": ""
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podMonitor": {
                    "enable": false,
                    "jobLabel": "spark-operator-podmonitor",
                    "labels": {},
                    "podMetricsEndpoint": {
                      "interval": "5s",
                      "scheme": "http"
                    }
                  },
                  "podSecurityContext": {},
                  "rbac": {
                    "create": false,
                    "createClusterRole": true,
                    "createRole": true
                  },
                  "replicaCount": 1,
                  "resourceQuotaEnforcement": {
                    "enable": false
                  },
                  "resources": {},
                  "resyncInterval": 30,
                  "securityContext": {},
                  "serviceAccounts": {
                    "spark": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "sparkoperator": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    }
                  },
                  "sparkJobNamespace": "",
                  "tolerations": [],
                  "uiService": {
                    "enable": true
                  },
                  "volumeMounts": [],
                  "volumes": [],
                  "webhook": {
                    "cleanupAnnotations": {
                      "helm.sh/hook": "pre-delete, pre-upgrade",
                      "helm.sh/hook-delete-policy": "hook-succeeded"
                    },
                    "enable": false,
                    "initAnnotations": {
                      "helm.sh/hook": "pre-install, pre-upgrade",
                      "helm.sh/hook-weight": "50"
                    },
                    "namespaceSelector": "",
                    "port": 8080,
                    "timeout": 30
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "ScheduledSparkApplication",
                "metadata": {
                  "name": "spark-pi-scheduled",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "concurrencyPolicy": "Allow",
                  "schedule": "@every 5m",
                  "template": {
                    "driver": {
                      "coreLimit": "1200m",
                      "cores": 1,
                      "labels": {
                        "version": "3.1.1"
                      },
                      "memory": "512m",
                      "serviceAccount": "spark-operator-spark",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "executor": {
                      "cores": 1,
                      "instances": 1,
                      "labels": {
                        "version": "3.1.1"
                      },
                      "memory": "512m",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "image": "gcr.io/spark-operator/spark:v3.1.1",
                    "imagePullPolicy": "Always",
                    "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.12-3.1.1.jar",
                    "mainClass": "org.apache.spark.examples.SparkPi",
                    "mode": "cluster",
                    "restartPolicy": {
                      "type": "Never"
                    },
                    "sparkVersion": "3.1.1",
                    "type": "Scala",
                    "volumes": [
                      {
                        "hostPath": {
                          "path": "/tmp",
                          "type": "Directory"
                        },
                        "name": "test-volume"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "pyspark-pi",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "driver": {
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "serviceAccount": "spark-operator-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "gcr.io/spark-operator/spark-py:v3.1.1",
                  "imagePullPolicy": "Always",
                  "mainApplicationFile": "local:///opt/spark/examples/src/main/python/pi.py",
                  "mode": "cluster",
                  "pythonVersion": "3",
                  "restartPolicy": {
                    "onFailureRetries": 3,
                    "onFailureRetryInterval": 10,
                    "onSubmissionFailureRetries": 5,
                    "onSubmissionFailureRetryInterval": 20,
                    "type": "OnFailure"
                  },
                  "sparkVersion": "3.1.1",
                  "type": "Python",
                  "volumes": [
                    {
                      "hostPath": {
                        "path": "/tmp",
                        "type": "Directory"
                      },
                      "name": "test-volume"
                    }
                  ]
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "spark-pi",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "driver": {
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "serviceAccount": "spark-operator-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "gcr.io/spark-operator/spark:v3.1.1",
                  "imagePullPolicy": "Always",
                  "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.12-3.1.1.jar",
                  "mainClass": "org.apache.spark.examples.SparkPi",
                  "mode": "cluster",
                  "restartPolicy": {
                    "type": "Never"
                  },
                  "sparkVersion": "3.1.1",
                  "type": "Scala",
                  "volumes": [
                    {
                      "hostPath": {
                        "path": "/tmp",
                        "type": "Directory"
                      },
                      "name": "test-volume"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/spark-helm-operator-controller:0.0.2
          createdAt: "2023-07-14T17:57:44Z"
          description: An experimental operator that installs the spark on k8s operator
            from its helm chart.
          operatorframework.io/initialization-resource: "{\n  \"apiVersion\": \"apachespark.apache.org/v1alpha1\",\n
            \ \"kind\": \"Spark\",\n  \"metadata\": {\n    \"name\": \"spark-operator\"\n
            \ },\n  \"spec\": {\n    \"affinity\": {},\n    \"batchScheduler\": {\n
            \     \"enable\": false\n    },\n    \"controllerThreads\": 10,\n    \"fullnameOverride\":
            \"\",\n    \"image\": {\n      \"pullPolicy\": \"IfNotPresent\",\n      \"repository\":
            \"ghcr.io/googlecloudplatform/spark-operator\",\n      \"tag\": \"\"\n
            \   },\n    \"imagePullSecrets\": [],\n    \"ingressUrlFormat\": \"\",\n
            \   \"istio\": {\n      \"enabled\": false\n    },\n    \"labelSelectorFilter\":
            \"\",\n    \"leaderElection\": {\n      \"lockName\": \"spark-operator-lock\",\n
            \     \"lockNamespace\": \"\"\n    },\n    \"logLevel\": 2,\n    \"metrics\":
            {\n      \"enable\": true,\n      \"endpoint\": \"/metrics\",\n      \"port\":
            10254,\n      \"portName\": \"metrics\",\n      \"prefix\": \"\"\n    },\n
            \   \"nameOverride\": \"\",\n    \"nodeSelector\": {},\n    \"podAnnotations\":
            {},\n    \"podLabels\": {},\n    \"podMonitor\": {\n      \"enable\":
            false,\n      \"jobLabel\": \"spark-operator-podmonitor\",\n      \"labels\":
            {},\n      \"podMetricsEndpoint\": {\n        \"interval\": \"5s\",\n
            \       \"scheme\": \"http\"\n      }\n    },\n    \"podSecurityContext\":
            {},\n    \"rbac\": {\n      \"create\": false,\n      \"createClusterRole\":
            true,\n      \"createRole\": true\n    },\n    \"replicaCount\": 1,\n
            \   \"resourceQuotaEnforcement\": {\n      \"enable\": false\n    },\n
            \   \"resources\": {},\n    \"resyncInterval\": 30,\n    \"securityContext\":
            {},\n    \"serviceAccounts\": {\n      \"spark\": {\n        \"annotations\":
            {},\n        \"create\": true,\n        \"name\": \"\"\n      },\n      \"sparkoperator\":
            {\n        \"annotations\": {},\n        \"create\": true,\n        \"name\":
            \"\"\n      }\n    },\n    \"sparkJobNamespace\": \"\",\n    \"tolerations\":
            [],\n    \"uiService\": {\n      \"enable\": true\n    },\n    \"volumeMounts\":
            [],\n    \"volumes\": [],\n    \"webhook\": {\n      \"cleanupAnnotations\":
            {\n        \"helm.sh/hook\": \"pre-delete, pre-upgrade\",\n        \"helm.sh/hook-delete-policy\":
            \"hook-succeeded\"\n      },\n      \"enable\": false,\n      \"initAnnotations\":
            {\n        \"helm.sh/hook\": \"pre-install, pre-upgrade\",\n        \"helm.sh/hook-weight\":
            \"50\"\n      },\n      \"namespaceSelector\": \"\",\n      \"port\":
            8080,\n      \"timeout\": 30\n    }\n  }\n}    "
          operatorframework.io/suggested-namespace: spark-operator
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys a scheduled spark application based on the spec provided
            displayName: Scheduled Spark Application
            kind: ScheduledSparkApplication
            name: scheduledsparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - description: Deploys a spark application based on the spec provided
            displayName: Spark Application
            kind: SparkApplication
            name: sparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - description: Deploys the spark operator based on the google cloud platform
              helm chart
            displayName: Spark Operator
            kind: Spark
            name: sparks.apachespark.apache.org
            version: v1alpha1
        description: An experimental helm based operator that wraps the spark on k8s
          helm chart from google cloud platform enabling installation and workload
          management of Apache Spark
        displayName: Spark Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - big data
        - map reduce
        - data science
        links:
        - name: Spark Helm Operator
          url: https://github.com/skattoju/spark-helm-operator
        maintainers:
        - email: skattoju@redhat.com
          name: Sid Kattoju
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: opdev.github.io
        relatedImages:
        - quay.io/opdev/spark-helm-operator-controller:0.0.2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 0.0.2
      entries:
      - name: spark-helm-operator.v0.0.2
        version: 0.0.2
      - name: spark-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: spark-helm-operator
    provider:
      name: opdev
      url: opdev.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-test-core
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-test-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Testcore",
                "metadata": {
                  "name": "testcore-sample"
                },
                "spec": {
                  "agentController": {
                    "port": 8080
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "dataseqService": {
                    "ingressEnabled": true,
                    "port": 8080,
                    "providers": {
                      "loadCore": {
                        "enabled": true
                      }
                    }
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "ingress": {
                    "backwardsCompatibilityEnabled": true
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "systemMonitor": {
                    "enabled": false
                  },
                  "trafficController": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:35:34Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Testcore
            name: testcores.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Test Core Operator
        displayName: Keysight WAP Test Core Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Test Core
          url: https://keysight-wap-test-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-test-core@sha256:9e2d81061abff49112f908e71dace03a748e283f32597aa22ad58471fff30caf
        version: 0.1.21
      entries:
      - name: keysight-wap-test-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-test-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp
      provider-url: https://www.netapp.com/
    name: netapp-spark-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netapp-spark-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.netapp.com/v1alpha1",
                "kind": "SparkOperator",
                "metadata": {
                  "name": "sparkoperator-sample"
                },
                "spec": {
                  "spark-operator": {
                    "affinity": {},
                    "controllerThreads": 20,
                    "disableExecutorReporting": false,
                    "hostNetwork": false,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "public.ecr.aws/f4k1p1n4/spark-operator",
                      "tag": "108810d-ubi9-rel20240916"
                    },
                    "imagePullSecrets": [],
                    "istio": {
                      "enabled": true
                    },
                    "jobAffinity": {},
                    "podLabels": {},
                    "pprof": {
                      "enable": true,
                      "port": 6060,
                      "portName": "pprof"
                    },
                    "rbac": {
                      "createClusterRole": true,
                      "createRole": true
                    },
                    "resources": {
                      "requests": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      }
                    },
                    "securityContext": {},
                    "serviceAccounts": {
                      "spark": {
                        "create": true
                      }
                    },
                    "sparkJobNamespace": "",
                    "tolerations": [],
                    "uiService": {
                      "enable": false
                    },
                    "webhook": {
                      "enable": true,
                      "jobTTLSecondsAfterFinished": 60,
                      "port": 9443
                    },
                    "workqueue": {
                      "maxDelay": {
                        "enabled": false,
                        "value": "6h"
                      },
                      "tokenBucketSize": 500,
                      "tokenRefillRate": 50
                    }
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "ScheduledSparkApplication",
                "metadata": {
                  "name": "spark-pi-scheduled"
                },
                "spec": {
                  "concurrencyPolicy": "Allow",
                  "schedule": "@every 5m",
                  "template": {
                    "arguments": [
                      "1000"
                    ],
                    "driver": {
                      "annotations": {
                        "openshift.io/required-scc": "sparkoperator-sample-spark"
                      },
                      "coreLimit": "1200m",
                      "cores": 1,
                      "labels": {
                        "version": "3.5.0"
                      },
                      "memory": "512m",
                      "serviceAccount": "sparkoperator-sample-spark",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "executor": {
                      "annotations": {
                        "openshift.io/required-scc": "sparkoperator-sample-spark"
                      },
                      "cores": 1,
                      "instances": 1,
                      "labels": {
                        "version": "3.5.0"
                      },
                      "memory": "512m",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "image": "public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce",
                    "imagePullPolicy": "IfNotPresent",
                    "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.13-3.5.0.jar",
                    "mainClass": "org.apache.spark.examples.SparkPi",
                    "mode": "cluster",
                    "restartPolicy": {
                      "type": "Never"
                    },
                    "sparkVersion": "3.5.0",
                    "type": "Scala",
                    "volumes": [
                      {
                        "emptyDir": {},
                        "name": "test-volume"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "spark-pi"
                },
                "spec": {
                  "arguments": [
                    "1000"
                  ],
                  "driver": {
                    "annotations": {
                      "openshift.io/required-scc": "sparkoperator-sample-spark"
                    },
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.5.0"
                    },
                    "memory": "512m",
                    "serviceAccount": "sparkoperator-sample-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "annotations": {
                      "openshift.io/required-scc": "sparkoperator-sample-spark"
                    },
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.5.0"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce",
                  "imagePullPolicy": "IfNotPresent",
                  "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.13-3.5.0.jar",
                  "mainClass": "org.apache.spark.examples.SparkPi",
                  "mode": "cluster",
                  "restartPolicy": {
                    "type": "Never"
                  },
                  "sparkVersion": "3.5.0",
                  "type": "Scala",
                  "volumes": [
                    {
                      "emptyDir": {},
                      "name": "test-volume"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Big Data, Integration & Delivery, AI/Machine Learning, Streaming
            & Messaging
          com.redhat.openshift.versions: v4.14
          containerImage: public.ecr.aws/f4k1p1n4/netapp-spark-operator-controller@sha256:9fe1d0793f11b1c525a9a92cbcde6f4bc1bf3c887617a8496bed3a5a0008a67e
          createdAt: "2024-10-21T13:47:47Z"
          description: Red Hat and NetApp have partnered to offer enterprise-grade
            supported Apache Spark on OpenShift.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ScheduledSparkApplication
            name: scheduledsparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - kind: SparkApplication
            name: sparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - kind: SparkOperator
            name: sparkoperators.charts.netapp.com
            version: v1alpha1
        description: |-
          # NetApp Supported Apache Spark on OpenShift

          Red Hat and NetApp have partnered to offer enterprise-grade supported Apache Spark on OpenShift.
          Use this certified operator to launch Apache Spark applications on OpenShift.

          ## Documentation

          Please find the official documentation at https://www.instaclustr.com/support/documentation/apache-spark/apache-spark-on-red-hat-openshift/getting-started/

          ## Support

          With questions about Apache Spark, contact NetApp's Instaclustr support team on the [web](https://support.instaclustr.com/) or by email at support@instaclustr.com

          With questions about Operators in the OpenShift OperatorHub, contact Red Hat's support team on the [web](https://www.redhat.com/en/services/support) or by email at customerservice@redhat.com

          With questions about pricing, email ng-netapp-apache-spark-support-redhat-openshift@netapp.com
        displayName: NetApp Supported Apache Spark on OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Spark
        - Big Data
        - Apache Spark
        - AI
        - ML
        - Data Engineering
        links:
        - name: NetApp Supported Apache Spark on OpenShift
          url: https://www.instaclustr.com/support-solutions
        maintainers:
        - email: thorsteinn.sigurdsson@netapp.com
          name: Thorsteinn Sigurdsson
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: NetApp
          url: https://www.netapp.com/
        relatedImages:
        - public.ecr.aws/f4k1p1n4/netapp-spark-operator-controller@sha256:9fe1d0793f11b1c525a9a92cbcde6f4bc1bf3c887617a8496bed3a5a0008a67e
        - public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce
        version: 1.1.3
      entries:
      - name: netapp-spark-operator.v1.1.3
        version: 1.1.3
      name: alpha
    defaultChannel: alpha
    packageName: netapp-spark-operator
    provider:
      name: NetApp
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Trilio
      provider-url: ""
    name: k8s-triliovault
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k8s-triliovault-stable.3.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/trilio/k8s-triliovault-operator@sha256:0e3094945427231ecbb884b903dc4d347dfaa73e5a04eb4c5cbaed6fce1cbbb8
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          olm.skipRange: '>=3.0.0 <3.1.3'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          TrilioVault for Kubernetes is an enterprise-grade, cloud-native platform purpose built for data protection and management of Kubernetes applications for IT managers, administrators and developers. TrilioVault supports upstream Kubernetes and OpenShift environments and offers the following features:
            * Application-Centric - protects both data and metadata for Helm, Operator or custom Label based applications.
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any Storage (CSI, NFS, S3), or any Cloud (Private or Public).
            * Application Deployment & Tooling: Helm, Operators, Labels, Prometheus, Fluentd.
          Tutorials
          ------------
          Please click the link below to access the TrilioVault for Kubernetes "How-To" series for deployment, best practice and use-case videos
          <a href="https://www.trilio.io/triliovault-for-kubernetes">TrilioVault for Kubernetes - Tutorials and How-To </a>”
          Licensing
          --------
          Customers can download a 30-day Free Trial or a 10-node Basic Edition at no cost.  You can also connect with the Trilio team for an Enterprise Edition license with no limitations and Premium Support.
          For more information on license plans please vist:
          <a href="www.trilio.io/plans"> Trilio Vault for Kubernetes licensing and plans </a>
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our TrilioVault technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. www.trilio.io and @triliodata on Twitter.
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://www.trilio.io/
        - name: Product Licensing
          url: https://www.trilio.io/plans
        - name: Tutorials
          url: https://www.trilio.io/triliovault-for-kubernetes/
        - name: Support for Free/Basic Users
          url: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Trilio
        relatedImages:
        - registry.connect.redhat.com/trilio/k8s-triliovault-operator@sha256:0e3094945427231ecbb884b903dc4d347dfaa73e5a04eb4c5cbaed6fce1cbbb8
        - registry.connect.redhat.com/trilio/trilio-webhook-init@sha256:c45e9b6b75862451d873d3eba29d6bd11cafc64823855f82ccb8c0b3d779ebb0
        - registry.connect.redhat.com/trilio/web@sha256:76865d8afa8059d40810cb94df29420950a7593a40e87d142d771c57b3319e1d
        - registry.connect.redhat.com/trilio/minio@sha256:cee0f976edc79ff9e7a3f3574a7115dd960042241f3ed5b4636240ad609f9bca
        - registry.connect.redhat.com/trilio/control-plane@sha256:01e36ed5b07104865f922d2996d2211b2b6000adc90266197d9bac74d908db97
        - registry.connect.redhat.com/trilio/datamover@sha256:c7bc15f8217899ed73a812d278eb118633edbee1bf284132054b6f9ca2afd801
        - registry.connect.redhat.com/trilio/dex@sha256:b6331cf3cb3a6a66d9b9ea639cda3c2559fbf1f92f7174032352c1c7bb5176d9
        - registry.connect.redhat.com/trilio/nats@sha256:47396bfcff3e65efb3410fa8369f11f20d794ab543f9ada87b451448b1813178
        - registry.connect.redhat.com/trilio/event-stack@sha256:37c33ae1adddfe739c98fd2b2183c49ec7cfdafc88dec2970ad7c670016ebaac
        version: 3.1.3
      entries:
      - name: k8s-triliovault-stable.3.1.3
        version: 3.1.3
      - name: k8s-triliovault-stable.3.1.2
        version: 3.1.2
      - name: k8s-triliovault-stable.3.1.1
        version: 3.1.1
      - name: k8s-triliovault-stable.3.1.0
        version: 3.1.0
      name: 3.1.x
    - currentCSV: k8s-triliovault-stable.4.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:69871613ceecefff7f691f21bb0b9147236af5a5dc2c7b36ad5e0cd8551f0e6a
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <4.0.5'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/event-stack@sha256:c793418df5a1fc1a49f629d335f4684d9cde1a5f2b462f7c9483ffba904456d2
        - quay.io/triliodata/control-plane@sha256:53368a72a60a38d3ce4293236a841346da6da2380ffd2525e5ffdb8dbcf11c7f
        - quay.io/triliodata/web@sha256:20de1a2f9477ad238bf24f48c8c16a38521c3a86164f2eb6b10d20f09839016e
        - quay.io/triliodata/datamover@sha256:e6848d94ba9f81843bde882a030b0f2556aae4bf16d79e2c5b199417cf626c85
        - quay.io/triliodata/k8s-triliovault-operator@sha256:69871613ceecefff7f691f21bb0b9147236af5a5dc2c7b36ad5e0cd8551f0e6a
        - quay.io/triliodata/operator-webhook-init@sha256:aa85b787f7f37a71e909f049d2abdda746cdd0627e2f74aaf7476d5e5a7f4401
        - quay.io/triliodata/dex@sha256:b4b2372fd03be586b9ff8cd269b9001797d813a34155ab37de31a44fe3cab1ea
        - quay.io/triliodata/nats@sha256:e6723595bcc5674b6201091d45d711d29a5cd1421c8b97c76cf00e468379fddf
        version: 4.0.5
      entries:
      - name: k8s-triliovault-stable.4.0.5
        version: 4.0.5
      - name: k8s-triliovault-stable.4.0.4
        version: 4.0.4
      - name: k8s-triliovault-stable.4.0.3
        version: 4.0.3
      - name: k8s-triliovault-stable.4.0.2
        version: 4.0.2
      - name: k8s-triliovault-stable.4.0.1
        version: 4.0.1
      - name: k8s-triliovault-stable.4.0.0
        version: 4.0.0
      name: 4.0.x
    - currentCSV: k8s-triliovault-stable.5.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:8d582842dbef1c64bbe3b0ce5eeb64790910bdc0416c5e5561557642dede7022
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <5.0.4'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/web@sha256:04c4baadae9c05f4add098ff62774d00102d56581e261626840d17f74effe438
        - quay.io/triliodata/datamover@sha256:104bdca9513c78f81c4340e9a3219df0a3bc8a5b026e4a252363a7e2c10977ba
        - quay.io/triliodata/dex@sha256:b4b2372fd03be586b9ff8cd269b9001797d813a34155ab37de31a44fe3cab1ea
        - quay.io/triliodata/nats@sha256:e6723595bcc5674b6201091d45d711d29a5cd1421c8b97c76cf00e468379fddf
        - quay.io/triliodata/event-stack@sha256:859535e19e0ebebc8261c649c2446f8ca5fe4b4e5b6236e69b780dcc0d0bb5b9
        - quay.io/triliodata/k8s-triliovault-operator@sha256:8d582842dbef1c64bbe3b0ce5eeb64790910bdc0416c5e5561557642dede7022
        - quay.io/triliodata/operator-webhook-init@sha256:ffcc55e85f884b884947b73fa7d2fb257704ff70ccd7f2dc37d07e02292abed6
        - quay.io/triliodata/control-plane@sha256:e60ef0a2c7a5b7879c9718e9a14c223f5290b4710b598fae19466a7558484824
        version: 5.0.4
      entries:
      - name: k8s-triliovault-stable.5.0.4
        version: 5.0.4
      - name: k8s-triliovault-stable.5.0.3
        version: 5.0.3
      - name: k8s-triliovault-stable.5.0.2
        version: 5.0.2
      - name: k8s-triliovault-stable.5.0.1
        version: 5.0.1
      - name: k8s-triliovault-stable.5.0.0
        version: 5.0.0
      name: 5.0.x
    - currentCSV: k8s-triliovault-stable.5.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "2Gi",
                      "cpu": "1"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:f0cf3f586561414671668210c59ac44071fde6abc954917e6ca04df5777a9a2f
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <5.1.1'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/k8s-triliovault-operator@sha256:f0cf3f586561414671668210c59ac44071fde6abc954917e6ca04df5777a9a2f
        - quay.io/triliodata/event-stack@sha256:548ae1a748919f518e811774ec87f128aa8857be234837e09308c295de165726
        - quay.io/triliodata/control-plane@sha256:2e7e44cf6bc70dd9b978dce2cb2d9a36b8886c0f55286d5a6f5eb0a5182c4350
        - quay.io/triliodata/datamover@sha256:8a1350c8975610f23dd0b40d604d218f59dca79db6beb81ffe78a2682e40864f
        - quay.io/triliodata/dex@sha256:6b9d1667916968fb1f187ad5945fcc08db702dcab4c947e309dd9d2881a7b9cf
        - quay.io/triliodata/nats@sha256:72283871b6b9343aeabc3ff090305979422f192d30a6cbc4b0f20b2de7533996
        - quay.io/triliodata/filerecovery-vm-ssh-server@sha256:09186be228bd0b59e83119b28f3239c6173e410a3cf36957fe2d9b56d25200f8
        - quay.io/triliodata/web@sha256:3908a39eed44d702c6b6804d392d08a570b3f5562590d050f17ddc7a5c87b37f
        - quay.io/triliodata/filerecovery@sha256:42559ac5d949d5ecf819cd4768a71fa688c72230502ff0b56539a8bb5e2f5838
        version: 5.1.1
      entries:
      - name: k8s-triliovault-stable.5.1.1
        version: 5.1.1
      - name: k8s-triliovault-stable.5.1.0
        version: 5.1.0
      name: 5.1.x
    defaultChannel: 5.1.x
    packageName: k8s-triliovault
    provider:
      name: Trilio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Couchbase
      provider-url: ""
    name: couchbase-enterprise-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: couchbase-operator.v2.6.4-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator@sha256:1f0a6865c360a14ba98706fd6fe8dff3bdbdede32e3622f93f19c82e377ca33f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        version: 2.6.4-7
      entries:
      - name: couchbase-operator.v2.6.4-7
        version: 2.6.4-7
      - name: couchbase-operator.v2.6.4-6
        version: 2.6.4-6
      name: 2.6.4
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:707571a7881bc12d5f1b0b1a94eb3c675d5c0ea0114c994a47ceb5aa79df4013
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        version: 2.7.0-5
      entries:
      - name: couchbase-operator.v2.7.0-5
        version: 2.7.0-5
      name: 2.7.0
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:a79e1b6781d5f8badf664e7efea6e2889e43fae6e0d37169d766685dcd6ff394
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        version: 2.7.1-1
      entries:
      - name: couchbase-operator.v2.7.1-1
        version: 2.7.1-1
      name: 2.7.1
    - currentCSV: couchbase-operator.v2.8.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator@sha256:df61928b064014d851c6e7dff6dcd63a433a27fa5623254b6e389c3eea170e36
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        version: 2.8.0-2
      entries:
      - name: couchbase-operator.v2.8.0-2
        version: 2.8.0-2
      - name: couchbase-operator.v2.8.0-1
        version: 2.8.0-1
      name: 2.8.0
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        version: 2.8.1-2
      entries:
      - name: couchbase-operator.v2.8.1-2
        version: 2.8.1-2
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: 2.8.1
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        version: 2.8.1-2
      entries:
      - name: couchbase-operator.v2.8.1-2
        version: 2.8.1-2
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: stable
    defaultChannel: stable
    packageName: couchbase-enterprise-certified
    provider:
      name: Couchbase
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: aws-efs-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-efs-csi-driver-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-aws-efs-csi-driver-rhel9-operator@sha256:35ef24108df5f24d24fefe915f501965c98edea934fa6c2895e375ed006e5e7b
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure AWS EFS CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0-0 <4.16.0-202509240123'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/aws-efs-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          AWS EFS CSI Driver Operator provides AWS Elastic File Service (EFS) CSI Driver that enables you to create and mount AWS EFS PersistentVolumes.

          Please follow [AWS EFS CSI Driver Operator documentation](https://docs.openshift.com/container-platform/4.16/storage/container_storage_interface/persistent-storage-csi-aws-efs.html) to:

          1. [Install the operator](https://docs.openshift.com/container-platform/4.16/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#persistent-storage-csi-olm-operator-install_persistent-storage-csi-aws-efs). This section also covers details about configuring an AWS role in OCP cluster that uses Secure Token Services (STS).

          2. [Install the AWS EFS CSI driver using the operator](https://docs.openshift.com/container-platform/4.16/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#persistent-storage-csi-efs-driver-install_persistent-storage-csi-aws-efs)

          3. Finally, [create a StorageClass](https://docs.openshift.com/container-platform/4.16/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#storage-create-storage-class_persistent-storage-csi-aws-efs) that enables dynamic provisioning of PersistentVolumes.
        displayName: AWS EFS CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - efs
        links:
        - name: Documentation
          url: https://github.com/openshift/aws-efs-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/aws-efs-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-aws-efs-csi-driver-rhel9-operator@sha256:35ef24108df5f24d24fefe915f501965c98edea934fa6c2895e375ed006e5e7b
        - registry.redhat.io/openshift4/ose-aws-efs-csi-driver-container-rhel9@sha256:c857dda8f399fa104bd4f392a354b74365657acea6b8fc5b023fdb4bf55057c2
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:9c3c3d2b857ce289324aee56df1ba4ab5e5ea1e76cd86ac988233463915834b4
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0627c09fe3f5a015f117fb30af1f4d17340aefa48b23994737d3901a804b63d3
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:a5704d26d0b562dba9633d1c107f34e05c750c0e3a64dbcace4dab7b964fff8d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        version: 4.16.0-202509240123
      entries:
      - name: aws-efs-csi-driver-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: aws-efs-csi-driver-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: aws-efs-csi-driver-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: aws-efs-csi-driver-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: aws-efs-csi-driver-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: aws-efs-csi-driver-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: aws-efs-csi-driver-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: aws-efs-csi-driver-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: aws-efs-csi-driver-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: aws-efs-csi-driver-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: aws-efs-csi-driver-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: aws-efs-csi-driver-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: aws-efs-csi-driver-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: aws-efs-csi-driver-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: aws-efs-csi-driver-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: aws-efs-csi-driver-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: aws-efs-csi-driver-operator.v4.16.0-202410280205
        version: 4.16.0-202410280205
      - name: aws-efs-csi-driver-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: aws-efs-csi-driver-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: aws-efs-csi-driver-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: aws-efs-csi-driver-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: aws-efs-csi-driver-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: aws-efs-csi-driver-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: aws-efs-csi-driver-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: aws-efs-csi-driver-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: aws-efs-csi-driver-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: aws-efs-csi-driver-operator.v4.16.0-202408061508
        version: 4.16.0-202408061508
      - name: aws-efs-csi-driver-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: aws-efs-csi-driver-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: aws-efs-csi-driver-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: aws-efs-csi-driver-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: aws-efs-csi-driver-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: aws-efs-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: self-node-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: self-node-remediation.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:cb03286b9d9cda4a47e5b1c0058ca315bf082bcc8697506ea88d97ba77452bea
          createdAt: "2023-10-20 16:38:06"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          olm.skipRange: '>=0.4.0 <0.5.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the heath detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-self-node-remediation-operator.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:cb03286b9d9cda4a47e5b1c0058ca315bf082bcc8697506ea88d97ba77452bea
        version: 0.5.1
      entries:
      - name: self-node-remediation.v0.5.1
        version: 0.5.1
      name: 4.12-eus
    - currentCSV: self-node-remediation.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:e2048a8a2ec37e10b57dafeba2e131311e61982242cda4beb989e5e30c1a5537
          createdAt: "2023-10-20 15:32:20"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          olm.skipRange: '>=0.5.0 <0.7.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:e2048a8a2ec37e10b57dafeba2e131311e61982242cda4beb989e5e30c1a5537
        version: 0.7.1
      entries:
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.14-eus
    - currentCSV: self-node-remediation.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:6bd86a47d69f7662e24a93ce205a8179018152a97696a9223d41ed1887c35257
          createdAt: "2025-07-02 16:40:20"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.9.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:6bd86a47d69f7662e24a93ce205a8179018152a97696a9223d41ed1887c35257
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ef91935704bb9f0af4cdad07333e802587e3d458b2cd911e04d557b1defe3412
        version: 0.9.1
      entries:
      - name: self-node-remediation.v0.9.1
        version: 0.9.1
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.16-eus
    - currentCSV: self-node-remediation.v0.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
          createdAt: "2025-07-02 18:34:25"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        version: 0.10.2
      entries:
      - name: self-node-remediation.v0.10.2
        version: 0.10.2
      - name: self-node-remediation.v0.10.1
        version: 0.10.1
      - name: self-node-remediation.v0.10.0
        version: 0.10.0
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.18-eus
    - currentCSV: self-node-remediation.v0.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
          createdAt: "2025-07-02 18:34:25"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        version: 0.10.2
      entries:
      - name: self-node-remediation.v0.10.2
        version: 0.10.2
      - name: self-node-remediation.v0.10.1
        version: 0.10.1
      - name: self-node-remediation.v0.10.0
        version: 0.10.0
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: stable
    defaultChannel: stable
    packageName: self-node-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC Corporation
      provider-url: https://jpn.nec.com/webotx/
    name: webotx-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: webotx-operator.v11.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "webotx.nec.com/v3",
                "kind": "ApplicationServer",
                "metadata": {
                  "name": "webotx-express"
                },
                "spec": {
                  "deployment": {
                    "image": "registry.connect.redhat.com/nec/webotx-express:11.11.00.00@sha256:73a5dc48b4e3340951dbe207d744f5ac68474701dd134fa7aa05e65e3510c8ad",
                    "replicas": 1
                  },
                  "service": {
                    "externalTrafficPolicy": "Cluster",
                    "ports": [
                      {
                        "name": "http",
                        "nodePort": 30080,
                        "port": 8080,
                        "protocol": "TCP",
                        "targetPort": 8080
                      },
                      {
                        "name": "https",
                        "nodePort": 30443,
                        "port": 8443,
                        "protocol": "TCP",
                        "targetPort": 8443
                      }
                    ],
                    "sessionAffinity": "ClientIP",
                    "sessionAffinityConfig": {
                      "clientIP": {
                        "timeoutSeconds": 10800
                      }
                    },
                    "type": "NodePort"
                  },
                  "webotx": {
                    "metering": {
                      "enabled": false,
                      "port": 8443,
                      "scheme": "https"
                    },
                    "sessionReplication": {
                      "enabled": false,
                      "useAllProcessGroups": false,
                      "useJmxAgent": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApplicationServer is the Schema for the applicationservers
              API
            displayName: Application Server
            kind: ApplicationServer
            name: applicationservers.webotx.nec.com
            version: v3
        description: WebOTX Operator is the Operator for WebOTX Application Server.
          WebOTX Operator enables to deploy and administrate WebOTX Application Server
          automatically.
        displayName: WebOTX Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - webotx
        - WebOTX
        - nec
        - NEC
        links:
        - name: WebOTX Operator for Kubernetes
          url: https://github.com/webotx/operator
        maintainers:
        - email: info-webotx@isd.jp.nec.com
          name: NEC Corporation
        maturity: stable
        provider:
          name: NEC Corporation
          url: https://jpn.nec.com/webotx/
        relatedImages:
        - registry.connect.redhat.com/nec/webotx-operator:11.11.00.00@sha256:df0d0315ed88be875ca7a6474500f43f17855f763c5c7d0f4d3ec58a2e46f6fb
        version: 11.11.0
      entries:
      - name: webotx-operator.v11.11.0
        version: 11.11.0
      name: stable
    defaultChannel: stable
    packageName: webotx-operator
    provider:
      name: NEC Corporation
      url: https://jpn.nec.com/webotx/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator-certified-rhmp.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a",
                  "imagePullPolicy": "Always",
                  "initImage": "registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-server-mongodb-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-server-mongodb-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:3c03fb18a5d73876576b80aca2b61e7c84ca052839ccb467d14e15f6d1f9d834
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a57926b53eef862b70a53818a7ccb0e9756f878f30d4128b608090b109bdc585
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator-certified-rhmp.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator-certified-rhmp.v1.20.0
        version: 1.20.0
      - name: percona-server-mongodb-operator-certified-rhmp.v1.19.1
        version: 1.19.1
      - name: percona-server-mongodb-operator-certified-rhmp.v1.19.0
        version: 1.19.0
      - name: percona-server-mongodb-operator-certified-rhmp.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator-certified-rhmp
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kms-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kms-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Alias",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Key",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kms-controller:1.1.1
          createdAt: "2025-09-19T22:02:02Z"
          description: AWS KMS controller is a service controller for managing KMS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alias represents the state of an AWS kms Alias resource.
            displayName: Alias
            kind: Alias
            name: aliases.kms.services.k8s.aws
            version: v1alpha1
          - description: Grant represents the state of an AWS kms Grant resource.
            displayName: Grant
            kind: Grant
            name: grants.kms.services.k8s.aws
            version: v1alpha1
          - description: Key represents the state of an AWS kms Key resource.
            displayName: Key
            kind: Key
            name: keys.kms.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Key Management Service (KMS) resources in AWS from within your Kubernetes cluster.

          **About Amazon KMS**

          AWS Key Management Service (AWS KMS) is a managed service that makes it easy for you to create and control the cryptographic keys that are used to protect your data. AWS KMS uses hardware security modules (HSM) to protect and validate your AWS KMS keys under the [FIPS 140-2 Cryptographic Module Validation Program](https://csrc.nist.gov/projects/cryptographic-module-validation-program/Certificate/3139), except in the China (Beijing) and China (Ningxia) Regions.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon KMS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kms
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon KMS Developer Resources
          url: https://aws.amazon.com/kms/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kms maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/kms-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-kms-controller.v1.1.1
        version: 1.1.1
      - name: ack-kms-controller.v1.1.0
        version: 1.1.0
      - name: ack-kms-controller.v1.0.31
        version: 1.0.31
      - name: ack-kms-controller.v1.0.30
        version: 1.0.30
      - name: ack-kms-controller.v1.0.29
        version: 1.0.29
      - name: ack-kms-controller.v1.0.28
        version: 1.0.28
      - name: ack-kms-controller.v1.0.27
        version: 1.0.27
      - name: ack-kms-controller.v1.0.26
        version: 1.0.26
      - name: ack-kms-controller.v1.0.25
        version: 1.0.25
      - name: ack-kms-controller.v1.0.24
        version: 1.0.24
      - name: ack-kms-controller.v1.0.23
        version: 1.0.23
      - name: ack-kms-controller.v1.0.22
        version: 1.0.22
      - name: ack-kms-controller.v1.0.19
        version: 1.0.19
      - name: ack-kms-controller.v1.0.18
        version: 1.0.18
      - name: ack-kms-controller.v1.0.17
        version: 1.0.17
      - name: ack-kms-controller.v1.0.16
        version: 1.0.16
      - name: ack-kms-controller.v1.0.15
        version: 1.0.15
      - name: ack-kms-controller.v1.0.14
        version: 1.0.14
      - name: ack-kms-controller.v1.0.13
        version: 1.0.13
      - name: ack-kms-controller.v1.0.12
        version: 1.0.12
      - name: ack-kms-controller.v1.0.8
        version: 1.0.8
      - name: ack-kms-controller.v1.0.7
        version: 1.0.7
      - name: ack-kms-controller.v1.0.5
        version: 1.0.5
      - name: ack-kms-controller.v1.0.4
        version: 1.0.4
      - name: ack-kms-controller.v1.0.3
        version: 1.0.3
      - name: ack-kms-controller.v1.0.2
        version: 1.0.2
      - name: ack-kms-controller.v1.0.0
        version: 1.0.0
      - name: ack-kms-controller.v0.1.5
        version: 0.1.5
      - name: ack-kms-controller.v0.1.4
        version: 0.1.4
      - name: ack-kms-controller.v0.1.3
        version: 0.1.3
      - name: ack-kms-controller.v0.1.2
        version: 0.1.2
      - name: ack-kms-controller.v0.1.1
        version: 0.1.1
      - name: ack-kms-controller.v0.1.0
        version: 0.1.0
      - name: ack-kms-controller.v0.0.20
        version: 0.0.20
      - name: ack-kms-controller.v0.0.19
        version: 0.0.19
      - name: ack-kms-controller.v0.0.18
        version: 0.0.18
      - name: ack-kms-controller.v0.0.17
        version: 0.0.17
      - name: ack-kms-controller.v0.0.16
        version: 0.0.16
      - name: ack-kms-controller.v0.0.15
        version: 0.0.15
      - name: ack-kms-controller.v0.0.14
        version: 0.0.14
      - name: ack-kms-controller.v0.0.13
        version: 0.0.13
      - name: ack-kms-controller.v0.0.12
        version: 0.0.12
      - name: ack-kms-controller.v0.0.11
        version: 0.0.11
      - name: ack-kms-controller.v0.0.10
        version: 0.0.10
      - name: ack-kms-controller.v0.0.9
        version: 0.0.9
      - name: ack-kms-controller.v0.0.8
        version: 0.0.8
      - name: ack-kms-controller.v0.0.7
        version: 0.0.7
      - name: ack-kms-controller.v0.0.6
        version: 0.0.6
      - name: ack-kms-controller.v0.0.5
        version: 0.0.5
      - name: ack-kms-controller.v0.0.4
        version: 0.0.4
      - name: ack-kms-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-kms-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: proactive-node-scaling-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: proactive-node-scaling-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NodeScalingWatermark",
                "metadata": {
                  "name": "ai-ml"
                },
                "spec": {
                  "nodeSelector": {
                    "machine.openshift.io/cluster-api-machine-type": "ai-ml"
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "workload",
                      "operator": "Equal",
                      "value": "ai-ml"
                    }
                  ],
                  "watermarkPercentage": 20
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NodeScalingWatermark",
                "metadata": {
                  "name": "us-west-2a"
                },
                "spec": {
                  "nodeSelector": {
                    "topology.kubernetes.io/zone": "us-west-2a"
                  },
                  "watermarkPercentage": 20
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "false"
          containerImage: quay.io/redhat-cop/proactive-node-scaling-operator@sha256:892303babb45090a6458cd4b1c61a01364012c43ea0bd6ed4d10bdcb8041e2ab
          createdAt: "2022-01-22T14:14:23Z"
          description: This operator provides automation proactively scaling nodes
            before pods get stuck waiting for the cluster autoscaler.
          operatorframework.io/suggested-namespace: proactive-node-scaling-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/proactive-node-scaling-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeScalingWatermark is the Schema for the nodescalingwatermarks
              API
            displayName: Node Scaling Watermark
            kind: NodeScalingWatermark
            name: nodescalingwatermarks.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator makes the [cluster autoscaler](https://github.com/kubernetes/autoscaler/tree/master/cluster-autoscaler) more proactive. As of now the cluster auto scaler will create new nodes only when a pod is pending because it cannot be allocated due to lack of capacity. This is not a goos user experience as the pending workload has to wait for several minutes as the new node is create and joins the cluster.

          The Proactive Node Scaling Operator improves the user experience by allocating low priority pods that don't do anything. When the cluster is full and a new user pod is created the following happens:

          1. some of the low priority pods are de-scheduled to make room for the user pod, which can then be scheduled. The user workload does not have to wait in this case.

          2. the de-scheduled low priority pods are rescheduled and in doing so the trigger the cluster autoscaler to add new nodes.

          Essentially this operator allows you to trade wasted resources for faster response time.

          In order for this operator to work correctly [pod priorities](https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/) must be defined. The default name for the priority class used by this operator is "proactive-node-autoscaling-pods" and it should have the lowest possible priority, 0. To ensure your regular workloads get a normal priority you should also define a PriorityClass for those with a higher priority than 0 and set globalDefault to true.

          For example:

          ```yaml
          apiVersion: scheduling.k8s.io/v1
          kind: PriorityClass
          metadata:
            name: proactive-node-autoscaling-pods
          value: 0
          globalDefault: false
          description: "This priority class is the priority class used for Proactive Node Scaling Pods."
          ---
          apiVersion: scheduling.k8s.io/v1
          kind: PriorityClass
          metadata:
            name: normal-workload
          value: 1000
          globalDefault: true
          description: "This priority classis the cluster default and should be used for normal workloads."
          ```

          Also for this operator to work the [cluster autoscaler](https://github.com/kubernetes/autoscaler/tree/master/cluster-autoscaler) must be active, see OpenShift instructions [here](https://docs.openshift.com/container-platform/4.6/machine_management/applying-autoscaling.html) on how to turn it on.

          To activate the proactive autoscaling, a CR must be defined, here is an example:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: NodeScalingWatermark
          metadata:
            name: us-west-2a
          spec:
            priorityClassName: proactive-node-autoscaling-pods
            watermarkPercentage: 20
            nodeSelector:
              topology.kubernetes.io/zone: us-west-2a
          ```

          The `nodeSelector` selects the nodes observed by this operator, which are also the nodes on which the low priority pods will be scheduled. The nodes observed by the cluster autoscaler should coincide with the nodes selected by this operator CR.

          The `watermarkPercentage` define the percentage of capacity of user workload that will be allocated to low priority pods. So in this example 20% of the user allocated capacity will be allocated via low priority pods. This also means that when the user workload reaches 80% capacity of the nodes selected by this CR (and the autoscaler), the cluster will start to scale.
        displayName: Proactive Node Scaling Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Capacity Management
        - Node Autoscaling
        - Proactive Autoscaling
        - Elastic Capacity
        links:
        - name: Proactive Node Scaling Operator
          url: https://github.com/redhat-cop/proactive-node-scaling-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/proactive-node-scaling-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/proactive-node-scaling-operator@sha256:892303babb45090a6458cd4b1c61a01364012c43ea0bd6ed4d10bdcb8041e2ab
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 0.4.2
      entries:
      - name: proactive-node-scaling-operator.v0.4.2
        version: 0.4.2
      - name: proactive-node-scaling-operator.v0.4.0
        version: 0.4.0
      - name: proactive-node-scaling-operator.v0.3.10
        version: 0.3.10
      - name: proactive-node-scaling-operator.v0.3.9
        version: 0.3.9
      - name: proactive-node-scaling-operator.v0.3.8
        version: 0.3.8
      - name: proactive-node-scaling-operator.v0.3.7
        version: 0.3.7
      - name: proactive-node-scaling-operator.v0.3.6
        version: 0.3.6
      - name: proactive-node-scaling-operator.v0.3.5
        version: 0.3.5
      - name: proactive-node-scaling-operator.v0.3.4
        version: 0.3.4
      - name: proactive-node-scaling-operator.v0.3.3
        version: 0.3.3
      - name: proactive-node-scaling-operator.v0.3.2
        version: 0.3.2
      - name: proactive-node-scaling-operator.v0.3.1
        version: 0.3.1
      - name: proactive-node-scaling-operator.v0.3.0
        version: 0.3.0
      - name: proactive-node-scaling-operator.v0.2.0
        version: 0.2.0
      - name: proactive-node-scaling-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: proactive-node-scaling-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Debezium Authors
      provider-url: https://debezium.io/
    name: debezium-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: debezium-operator.v2.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.4.2.Final
          createdAt: "2023-12-13T12:20:56Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.4.2.Final
        version: 2.4.2
      entries:
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.4.x
    - currentCSV: debezium-operator.v2.5.3-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.5.3.Final
          createdAt: "2024-03-19T08:17:17Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.5.3.Final
        version: 2.5.3-final
      entries:
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.5.x
    - currentCSV: debezium-operator.v2.6.1-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.6.1.Final
          createdAt: "2024-04-12T17:18:16Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.5.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.6.1.Final
        version: 2.6.1-final
      entries:
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.6.x
    - currentCSV: debezium-operator.v2.7.3-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.7.3.Final
          createdAt: "2024-09-20T10:31:16Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.7.3.Final
        version: 2.7.3-final
      entries:
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.7.x
    - currentCSV: debezium-operator.v3.0.0-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:3.0.0.Final
          createdAt: "2024-10-02T09:55:38Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.7.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:3.0.0.Final
        version: 3.0.0-final
      entries:
      - name: debezium-operator.v3.0.0-final
        version: 3.0.0-final
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-3.0.x
    - currentCSV: debezium-operator.v3.0.0-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:3.0.0.Final
          createdAt: "2024-10-02T09:55:38Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.7.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:3.0.0.Final
        version: 3.0.0-final
      entries:
      - name: debezium-operator.v3.0.0-final
        version: 3.0.0-final
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-latest
    defaultChannel: debezium-latest
    packageName: debezium-operator
    provider:
      name: Debezium Authors
      url: https://debezium.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-notifications
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-notifications.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Notifications",
                "metadata": {
                  "name": "notifications-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configuration": {
                    "port": 8080
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp",
                    "test": "wap-rhel8-toolbox:8.7.0-3747-36-ocp",
                    "testRepository": "prod.harbor.keysight.digital/loadcore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:34:47Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Notifications
            name: notifications.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Notifications Operator
        displayName: Keysight WAP Notifications Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Notifications
          url: https://keysight-wap-notifications.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-notifications@sha256:50c24180b3c16cf3eb82f6f4eeb61b91a62a198c122f7e59aa680a25bc488c2d
        version: 0.1.21
      entries:
      - name: keysight-wap-notifications.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-notifications
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-stats-dashboards
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-stats-dashboards.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Statsdashboard",
                "metadata": {
                  "name": "statsdashboard-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "statsDashboardService": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:41:08Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Statsdashboard
            name: statsdashboards.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Stats Dashboard Operator
        displayName: Keysight WAP Stats Dashboard Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Stats Dashboards
          url: https://keysight-wap-stats-dashboards.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-stats-dashboards@sha256:bcdc68bf0b9e9c22d775d704320829d258a9ebc77dfe41e2414061740f87f236
        version: 0.1.21
      entries:
      - name: keysight-wap-stats-dashboards.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-stats-dashboards
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Saison Technology Co., Ltd.
      provider-url: https://www.hulft.com/
    name: hulft
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hulft.v10.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hulft.saison-technology.com/v1",
                "kind": "Hulft",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "hulft",
                    "app.kubernetes.io/instance": "sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hulft",
                    "app.kubernetes.io/part-of": "hulft"
                  },
                  "name": "sample",
                  "namespace": "user-apps"
                },
                "spec": {
                  "applicationName": "hulft-apps",
                  "control": {
                    "allowExternalAccess": true,
                    "replicas": {
                      "max": 1,
                      "min": 1
                    }
                  },
                  "db": {
                    "authorizationSecret": "secret-hulft-db-authorization",
                    "connectionRetryCount": 30,
                    "hostname": "database.example.com",
                    "name": "hulft",
                    "port": 3306,
                    "tablePrefix": "hulft"
                  },
                  "licenseSecret": "secret-hulft-license",
                  "resourceIdentifier": "sample",
                  "transfer": {
                    "control": {
                      "retryCount": 10,
                      "retryInterval": 10000,
                      "timeout": 60000
                    },
                    "gracefulShutdownWait": 3600,
                    "observe": {
                      "websocket": {
                        "allowExternalAccess": false
                      }
                    },
                    "replicas": {
                      "max": 5,
                      "min": 1
                    },
                    "s3": {
                      "defaultRegion": "ap-northeast-1"
                    },
                    "websocket": {
                      "allowExternalAccess": false
                    }
                  }
                },
                "status": {
                  "applicationName": "hulft-apps",
                  "control": {
                    "allowExternalAccess": true,
                    "apiClientTlsSecret": "secret-hulft-sample-transfer-control-server-tls",
                    "consoleLoginSecret": "secret-hulft-sample-console-login",
                    "consoleTlsSecret": "secret-hulft-sample-control-console-tls",
                    "hostname": "",
                    "httpPort": 30080,
                    "httpsPort": 30443,
                    "replicas": {
                      "max": 1,
                      "min": 1
                    },
                    "samlSecret": ""
                  },
                  "db": {
                    "authorizationSecret": "secret-hulft-db-authorization",
                    "connectionRetryCount": 30,
                    "encryptionKeySecret": "secret-hulft-sample-db-encryption-key",
                    "hostname": "database.example.com",
                    "name": "hulft",
                    "port": 3306,
                    "tablePrefix": "hulft"
                  },
                  "licenseSecret": "secret-hulft-license",
                  "resourceIdentifier": "sample",
                  "transfer": {
                    "control": {
                      "clientTlsSecret": "secret-hulft-sample-control-console-tls",
                      "port": 30443,
                      "retryCount": 10,
                      "retryInterval": 10000,
                      "serverTlsSecret": "secret-hulft-sample-transfer-control-server-tls",
                      "timeout": 60000
                    },
                    "gracefulShutdownWait": 3600,
                    "hostname": "hulft-sample-transfer-wss-ci.apps.openshift.example.com",
                    "hulft": {
                      "port": 30000
                    },
                    "observe": {
                      "port": 31000,
                      "websocket": {
                        "allowExternalAccess": true,
                        "port": 33443,
                        "serverTlsSecret": "secret-hulft-sample-transfer-observe-wss-server-tls"
                      }
                    },
                    "replicas": {
                      "max": 5,
                      "min": 1
                    },
                    "s3": {
                      "authenticationSecret": "",
                      "defaultRegion": "ap-northeast-1"
                    },
                    "websocket": {
                      "allowExternalAccess": true,
                      "clientTlsSecret": "",
                      "port": 31443,
                      "serverTlsSecret": "secret-hulft-sample-transfer-wss-server-tls"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/saison-technology/hulft-operator@sha256:8ee0a29d0ce96f4ec230f387467b241dccbb2bd350cd642b275ef5f5d3b64972
          createdAt: "2025-09-30T07:54:26Z"
          description: HULFT10 for Container Platform is file integration middleware
            that provides transfer functions for the data accumulated in your IT system
            during the daily operations of your business, such as sales data, customer
            information, and various logs. File transfer from/to on-premises HULFT
            is also available.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: To use HULFT10 for Container
            Platform, you must purchase the product and have a key issued. Please
            refer to [Start Up Guide](https://www.hulft.com/help/ja-jp/HULFT-V10/CONP-SUGOS)
            for the deployment procedure.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Saison Technology Co., Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hulft is the Schema for the hulfts API
            displayName: Hulft
            kind: Hulft
            name: hulfts.hulft.saison-technology.com
            version: v1
        description: |
          HULFT10 for Container Platform is file integration middleware that provides transfer functions for the data accumulated in your IT system during the daily operations of your business, such as sales data, customer information, and various logs. File transfer from/to on-premises HULFT is also available.
          To use HULFT10 for Container Platform, you must purchase the product and have a key issued. Please refer to [Start Up Guide](https://www.hulft.com/help/ja-jp/HULFT-V10/CONP-SUGOS) for the deployment procedure.
        displayName: HULFT10 for Container Platform
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backup & Recovery
        - Storage
        - Migration
        links:
        - name: Product page
          url: https://www.hulft.com/software/hulft-10
        - name: Product manual
          url: https://www.hulft.com/help/ja-jp/HULFT-V10/index.htm
        maintainers:
        - email: salessupport_mp@hulft.com
          name: Saison Technology Co., Ltd.
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Saison Technology Co., Ltd.
          url: https://www.hulft.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fa0acefda43e309e7f9f31f5247c933d0b868ff1573ece2f38e0549d7e2ec942
        - registry.connect.redhat.com/saison-technology/hulft-operator@sha256:8ee0a29d0ce96f4ec230f387467b241dccbb2bd350cd642b275ef5f5d3b64972
        version: 10.5.0
      entries:
      - name: hulft.v10.5.0
        version: 10.5.0
      - name: hulft.v10.4.0
        version: 10.4.0
      - name: hulft.v10.3.0
        version: 10.3.0
      name: stable
    defaultChannel: stable
    packageName: hulft
    provider:
      name: Saison Technology Co., Ltd.
      url: https://www.hulft.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: local-storage-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: local-storage-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "local.storage.openshift.io/v1",
                "kind": "LocalVolume",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "storageClassDevices": [
                    {
                      "devicePaths": [
                        "/dev/disk/by-id/ata-crucial"
                      ],
                      "fsType": "ext4",
                      "storageClassName": "foobar",
                      "volumeMode": "Filesystem"
                    }
                  ]
                }
              },
              {
                "apiVersion": "local.storage.openshift.io/v1alpha1",
                "kind": "LocalVolumeSet",
                "metadata": {
                  "name": "example-localvolumeset"
                },
                "spec": {
                  "deviceInclusionSpec": {
                    "deviceMechanicalProperties": [
                      "Rotational",
                      "NonRotational"
                    ],
                    "deviceTypes": [
                      "disk"
                    ],
                    "maxSize": "100G",
                    "minSize": "10G"
                  },
                  "maxDeviceCount": 10,
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "storageClassName": "example-storageclass",
                  "volumeMode": "Block"
                }
              },
              {
                "apiVersion": "local.storage.openshift.io/v1alpha1",
                "kind": "LocalVolumeDiscovery",
                "metadata": {
                  "name": "auto-discover-devices"
                },
                "spec": {
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-local-storage-rhel9-operator@sha256:6f6491f023ffa21eff6a9b938cb1869f362d6b4a5b04c27dc5f13f49622ca178
          createdAt: "2019-08-14T00:00:00Z"
          description: Configure and use local storage volumes.
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.17"}]'
          olm.skipRange: '>=4.3.0 <4.16.0-202509240123'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-local-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["localvolumediscoveryresults.local.storage.openshift.io"]'
          repository: https://github.com/openshift/local-storage-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manage local storage volumes for OpenShift
            displayName: Local Volume
            kind: LocalVolume
            name: localvolumes.local.storage.openshift.io
            version: v1
          - description: A Local Volume set allows you to filter a set of storage
              volumes, group them and create a dedicated storage class to consume
              storage from the set of volumes.
            displayName: Local Volume Set
            kind: LocalVolumeSet
            name: localvolumesets.local.storage.openshift.io
            version: v1alpha1
          - description: Discover list of potentially usable disks on the chosen set
              of nodes
            displayName: Local Volume Discovery
            kind: LocalVolumeDiscovery
            name: localvolumediscoveries.local.storage.openshift.io
            version: v1alpha1
          - description: Disc inventory of available disks from selected nodes
            displayName: Local Volume Discovery Result
            kind: LocalVolumeDiscoveryResult
            name: localvolumediscoveryresults.local.storage.openshift.io
            version: v1alpha1
        description: |
          Operator that configures local storage volumes for use in Kubernetes and OpenShift. OpenShift 4.2 and above are the only supported OpenShift versions.
        displayName: Local Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - local storage
        links:
        - name: Documentation
          url: https://github.com/openshift/local-storage-operator/tree/master/docs
        - name: Source Repository
          url: https://github.com/openshift/local-storage-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-local-storage-rhel9-operator@sha256:6f6491f023ffa21eff6a9b938cb1869f362d6b4a5b04c27dc5f13f49622ca178
        - registry.redhat.io/openshift4/ose-local-storage-diskmaker-rhel9@sha256:d8d8b21fce5930a36c9755d82037fce1ab81f5ae2d5fb1bfd643d500c4ec7fba
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel9@sha256:ec5847d5006d66a014315decd579e10104b086a06bfefa33430545b4ffcece71
        version: 4.16.0-202509240123
      entries:
      - name: local-storage-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: local-storage-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: local-storage-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: local-storage-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: local-storage-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: local-storage-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: local-storage-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: local-storage-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: local-storage-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: local-storage-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: local-storage-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: local-storage-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: local-storage-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: local-storage-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: local-storage-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: local-storage-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: local-storage-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: local-storage-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: local-storage-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: local-storage-operator.v4.16.0-202410031007
        version: 4.16.0-202410031007
      - name: local-storage-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: local-storage-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: local-storage-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: local-storage-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: local-storage-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: local-storage-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: local-storage-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: local-storage-operator.v4.16.0-202408021139
        version: 4.16.0-202408021139
      - name: local-storage-operator.v4.16.0-202407190036
        version: 4.16.0-202407190036
      - name: local-storage-operator.v4.16.0-202407111006
        version: 4.16.0-202407111006
      - name: local-storage-operator.v4.16.0-202407081937
        version: 4.16.0-202407081937
      - name: local-storage-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: local-storage-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: local-storage-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: local-storage-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: local-storage-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: security-profiles-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: security-profiles-operator.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "AppArmorProfile",
                "metadata": {
                  "annotations": {
                    "description": "Block writing to any files in the disk."
                  },
                  "name": "test-profile"
                },
                "spec": {
                  "abstract": {
                    "filesystem": {
                      "readOnlyPaths": [],
                      "readWritePaths": [],
                      "writeOnlyPaths": []
                    }
                  }
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "SecurityProfilesOperatorDaemon",
                "metadata": {
                  "name": "spod",
                  "namespace": "security-profiles-operator"
                },
                "spec": {
                  "enableAppArmor": false,
                  "enableLogEnricher": false,
                  "enableSelinux": false
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "ProfileRecording",
                "metadata": {
                  "name": "test-recording"
                },
                "spec": {
                  "kind": "SeccompProfile",
                  "podSelector": {
                    "matchLabels": {
                      "app": "alpine"
                    }
                  },
                  "recorder": "bpf"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "ProfileBinding",
                "metadata": {
                  "name": "profile-binding"
                },
                "spec": {
                  "image": "nginx:1.19.1",
                  "profileRef": {
                    "kind": "SeccompProfile",
                    "name": "profile-allow-unsafe"
                  }
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha2",
                "kind": "RawSelinuxProfile",
                "metadata": {
                  "name": "errorlogger"
                },
                "spec": {
                  "policy": "(blockinherit container)\n(allow process var_log_t ( dir ( open read getattr lock search ioctl add_name remove_name write ))) \n(allow process var_log_t ( file ( getattr read write append ioctl lock map open create  ))) \n(allow process var_log_t ( sock_file ( getattr read write append open  ))) \n"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "Blocks all syscalls."
                  },
                  "name": "profile-block-all"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_ERRNO"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "UNSAFE: Allows all syscalls whilst logging their use. Similar to running as unconfined in terms of enforcement."
                  },
                  "name": "profile-complain-unsafe"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_LOG"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "UNSAFE: Allows all syscalls. Similar to running as unconfined as it provides no enforcement."
                  },
                  "name": "profile-allow-unsafe"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_ALLOW"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "Enables complain mode whilst blocking high-risk syscalls. Some essential syscalls are allowed to decrease log noise."
                  },
                  "name": "profile-complain-block-high-risk"
                },
                "spec": {
                  "architectures": [
                    "SCMP_ARCH_X86_64",
                    "SCMP_ARCH_PPC64LE"
                  ],
                  "defaultAction": "SCMP_ACT_LOG",
                  "syscalls": [
                    {
                      "action": "SCMP_ACT_ALLOW",
                      "names": [
                        "exit",
                        "exit_group",
                        "futex",
                        "nanosleep"
                      ]
                    },
                    {
                      "action": "SCMP_ACT_ERRNO",
                      "names": [
                        "acct",
                        "add_key",
                        "bpf",
                        "clock_adjtime",
                        "clock_settime",
                        "create_module",
                        "delete_module",
                        "finit_module",
                        "get_kernel_syms",
                        "get_mempolicy",
                        "init_module",
                        "ioperm",
                        "iopl",
                        "kcmp",
                        "kexec_file_load",
                        "kexec_load",
                        "keyctl",
                        "lookup_dcookie",
                        "mbind",
                        "mount",
                        "move_pages",
                        "name_to_handle_at",
                        "nfsservctl",
                        "open_by_handle_at",
                        "perf_event_open",
                        "personality",
                        "pivot_root",
                        "process_vm_readv",
                        "process_vm_writev",
                        "ptrace",
                        "query_module",
                        "quotactl",
                        "reboot",
                        "request_key",
                        "set_mempolicy",
                        "setns",
                        "settimeofday",
                        "stime",
                        "swapoff",
                        "swapon",
                        "_sysctl",
                        "sysfs",
                        "umount2",
                        "umount",
                        "unshare",
                        "uselib",
                        "userfaultfd",
                        "ustat",
                        "vm86old",
                        "vm86"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha2",
                "kind": "SelinuxProfile",
                "metadata": {
                  "name": "errorlogger"
                },
                "spec": {
                  "allow": {
                    "var_log_t": {
                      "dir": [
                        "open",
                        "read",
                        "getattr",
                        "lock",
                        "search",
                        "ioctl",
                        "add_name",
                        "remove_name",
                        "write"
                      ],
                      "file": [
                        "getattr",
                        "read",
                        "write",
                        "append",
                        "ioctl",
                        "lock",
                        "map",
                        "open",
                        "create"
                      ],
                      "sock_file": [
                        "getattr",
                        "read",
                        "write",
                        "append",
                        "open"
                      ]
                    }
                  },
                  "inherit": [
                    {
                      "name": "container"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          createdAt: "2025-02-25T13:34:54Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.1 <0.9.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-security-profiles
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ProfileBinding is the Schema for the profilebindings API.
            displayName: Profile Binding
            kind: ProfileBinding
            name: profilebindings.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: ProfileRecording is the Schema for the profilerecordings
              API.
            displayName: Profile Recording
            kind: ProfileRecording
            name: profilerecordings.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: SeccompProfile is a cluster level specification for a seccomp
              profile. See https://github.com/opencontainers/runtime-spec/blob/master/config-linux.md#seccomp
            displayName: Seccomp Profile
            kind: SeccompProfile
            name: seccompprofiles.security-profiles-operator.x-k8s.io
            version: v1beta1
          - description: SecurityProfileNodeStatus is a per-node status of a security
              profile
            displayName: Security Profile Node Status
            kind: SecurityProfileNodeStatus
            name: securityprofilenodestatuses.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: SecurityProfilesOperatorDaemon is the Schema to configure
              the spod deployment.
            displayName: Security Profiles Operator Daemon
            kind: SecurityProfilesOperatorDaemon
            name: securityprofilesoperatordaemons.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - kind: RawSelinuxProfile
            name: rawselinuxprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha2
          - description: SelinuxProfile is the Schema for the selinuxprofiles API.
            displayName: Selinux Profile
            kind: SelinuxProfile
            name: selinuxprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha2
          - description: AppArmorProfile is a cluster level specification for an AppArmor
              profile.
            displayName: App Armor Profile
            kind: AppArmorProfile
            name: apparmorprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha1
        description: SPO is an operator which aims to make it easier for users to
          use SELinux, seccomp and AppArmor in Kubernetes clusters
        displayName: Security Profiles Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - selinux
        - seccomp
        - apparmor
        - ebpf
        links:
        - name: Security Profiles Operator
          url: https://github.com/openshift/security-profiles-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-security-profiles-rhel8-operator@sha256:828274d71c65ef820e6a8b652e3f102624d09a0609fb00bae7b32dc62ff37fd1
        - registry.redhat.io/compliance/openshift-selinuxd-rhel9@sha256:3278d69814a8d0d1e31dcdef454c24828aadbb468a3d5d3e15910c76481765bd
        - registry.redhat.io/compliance/openshift-selinuxd-rhel8@sha256:81273ff149482109417064b40d333edb131481c05547e4ffa6485475460a48ba
        version: 0.9.0
      entries:
      - name: security-profiles-operator.v0.9.0
        version: 0.9.0
      - name: security-profiles-operator.v0.8.6
        version: 0.8.6
      - name: security-profiles-operator.v0.8.5
        version: 0.8.5
      - name: security-profiles-operator.v0.8.4
        version: 0.8.4
      - name: security-profiles-operator.v0.8.2
        version: 0.8.2
      name: release-alpha-rhel-8
    defaultChannel: release-alpha-rhel-8
    packageName: security-profiles-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: prometheus-exporter-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometheus-exporter-operator.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.3scale.net/v1alpha1",
                "kind": "PrometheusExporter",
                "metadata": {
                  "name": "example-memcached"
                },
                "spec": {
                  "dbHost": "your-memcached-host",
                  "dbPort": 11211,
                  "grafanaDashboard": {
                    "label": {
                      "key": "autodiscovery",
                      "value": "enabled"
                    }
                  },
                  "type": "memcached"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/3scale/prometheus-exporter-operator:v0.7.0
          createdAt: "2023-09-07 00:00:00"
          description: Operator to setup 3rd party prometheus exporters, with a collection
            of grafana dashboards
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/3scale-ops/prometheus-exporter-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configures a prometheus exporter to monitor a memcached instance
            displayName: PrometheusExporter
            kind: PrometheusExporter
            name: prometheusexporters.monitoring.3scale.net
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK to centralize the setup of 3rd party prometheus exporters on **Kubernetes/OpenShift**, with a collection of grafana dashboards.

          By just providing a few parameters like *dbHost* or *dbPort* (operator manage the container image, port, argument, command, volumes... and also prometheus `ServiceMonitor` and `GrafanaDashboard` k8s objects), you can setup different prometheus exporters to monitor:
          * The **internals from different databases**
          * **HTTP/TCP endpoints** (availability, latency, SSL/TLS certificate expiration...)
          * Any available **cloudwatch metric from any AWS Service**
          * Sendgrid email statistics (delivered, bounces, errors, spam...)

          Current prometheus exporters types supported, managed by same prometheus-exporter-operator:
          * memcached
          * redis
          * mysql
          * postgresql
          * sphinx
          * es (elasticsearch)
          * cloudwatch
          * probe (blackbox)
          * sendgrid

          The operator manages the lifecycle of the following objects:
          * Deployment (one per CR)
          * Service (one per CR)
          * ServiceMonitor (optional, one per CR)
          * GrafanaDashboard (optional, one per Namespace)

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale-ops/prometheus-exporter-operator#documentation).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale-ops/prometheus-exporter-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:
          * Raising any issues you find using Prometheus Exporter Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale-ops/prometheus-exporter-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving [documentation](https://github.com/3scale-ops/prometheus-exporter-operator)
          * Talking about Prometheus Exporter Operator
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale-ops/prometheus-exporter-operator/issues).

          ### License
          Prometheus Exporter Operator is licensed under the [Apache 2.0 license](https://github.com/3scale-ops/prometheus-exporter-operator/blob/main/LICENSE)
        displayName: Prometheus Exporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - metrics
        - observability
        - prometheus
        - exporter
        - grafana
        - mysql
        - postgresql
        - memcached
        - redis
        - sphinx
        - elasticsearch
        - cloudwatch
        - blackbox
        - sendgrid
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/prometheus-exporter-operator
        maintainers:
        - email: 3scale-operations@redhat.com
          name: 3scale Ops
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/3scale/prometheus-exporter-operator:v0.7.0
        version: 0.7.0
      entries:
      - name: prometheus-exporter-operator.v0.7.0
        version: 0.7.0
      - name: prometheus-exporter-operator.v0.3.4
        version: 0.3.4
      name: alpha
    defaultChannel: alpha
    packageName: prometheus-exporter-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: boxboat-github-practice
      provider-url: https://github.com/boxboat-github-practice
    name: github-arc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: github-arc-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "github-practice.boxboat.com/v1alpha1",
                "kind": "ActionsRunnerController",
                "metadata": {
                  "name": "arc-sample"
                },
                "spec": {
                  "authSecret": {
                    "name": "ghauth"
                  },
                  "createRunnerNamespaces": true,
                  "openshift": true,
                  "runnerNamespaces": [
                    "test1",
                    "test2"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery, OpenShift Optional
          containerImage: ghcr.io/boxboat-github-practice/github-arc-operator:1.1.0
          createdAt: "2023-03-03T20:49:16Z"
          description: Operator for deploying the github actions runner controller
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/boxboat-github-practice/github-arc-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActionsRunnerController is the schema for the ActionsRunnerController
              API
            displayName: Actions Runner Controller
            kind: ActionsRunnerController
            name: actionsrunnercontrollers.github-practice.boxboat.com
            version: v1alpha1
        description: helm based operator for github actions runner controller
        displayName: github-arc-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - github
        - actions
        - boxboat
        - ibm
        - self-hosted
        - runners
        links:
        - name: Github Arc Operator
          url: https://github.com/boxboat-github-practice/github-arc-operator
        - name: Actions Runner Controller
          url: https://github.com/actions/actions-runner-controller
        - name: Boxboat
          url: https://boxboat.com/
        maintainers:
        - email: Michael.McMullen@ibm.com
          name: mbmcmullen27
        maturity: alpha
        provider:
          name: boxboat-github-practice
          url: https://github.com/boxboat-github-practice
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - ghcr.io/boxboat-github-practice/github-arc-operator:1.1.0
        version: 1.1.0
      entries:
      - name: github-arc-operator.v1.1.0
        version: 1.1.0
      - name: github-arc-operator.v1.0.4
        version: 1.0.4
      name: alpha
    defaultChannel: alpha
    packageName: github-arc-operator
    provider:
      name: boxboat-github-practice
      url: https://github.com/boxboat-github-practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: special-resource-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: special-resource-operator.v4.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sro.openshift.io/v1beta1",
                "kind": "SpecialResource",
                "metadata": {
                  "name": "simple-kmod"
                },
                "spec": {
                  "chart": {
                    "name": "simple-kmod",
                    "repository": {
                      "name": "example",
                      "url": "file:///charts/example"
                    },
                    "version": "0.0.1"
                  },
                  "driverContainer": {
                    "source": {
                      "git": {
                        "ref": "master",
                        "uri": "https://github.com/openshift-psap/kvc-simple-kmod.git"
                      }
                    }
                  },
                  "namespace": "simple-kmod",
                  "set": {
                    "apiVersion": "sro.openshift.io/v1beta1",
                    "buildArgs": [
                      {
                        "name": "KMODVER",
                        "value": "SRO"
                      }
                    ],
                    "kind": "Values",
                    "kmodNames": [
                      "simple-kmod",
                      "simple-procfs-kmod"
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional, Storage
          containerImage: quay.io/openshift-psap/special-resource-operator:release-4.9
          description: The special resource operator is an orchestrator for resources
            in a cluster required to deploy special hardware and software. Examples
            include hardware accelerators or software defined storage, where extra
            management or kernel modules are needed.
          olm.skipRange: '>=4.6.0 <4.9.0'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpecialResource is the Schema for the specialresources API
            displayName: Special Resource
            kind: SpecialResource
            name: specialresources.sro.openshift.io
            version: v1beta1
          required:
          - description: Discovers special resources and labels the nodes with the
              PCI vendor ID.
            displayName: Node Feature Discovery
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
        description: "## About the managed application \nThe SRO will deploy the complete
          stack for enabling the accelerator in a cluster. \nIt manages the drivers,
          device-plugin, node-exporter, grafana and feature discovery.\n## About this
          Operator\n README: [https://github.com/openshift-psap/special-resource-operator](https://github.com/openshift-psap/special-resource-operator)\n\n##
          Prerequisites for enabling this Operator\nThe SRO has a dependency on NFD
          (Node Feature Discovery)\n"
        displayName: Special Resource Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GPI
        links:
        - name: README
          url: https://github.com/openshift-psap/special-resource-operator
        maintainers:
        - email: zkaiser@redhat.com
          name: Zvonko Kaiser
        - email: dagray@redhat.com
          name: David Gray
        maturity: alpha
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy
        - quay.io/openshift-psap/special-resource-operator:release-4.9
        version: 4.9.0
      entries:
      - name: special-resource-operator.v4.9.0
        version: 4.9.0
      - name: special-resource-operator.v4.8.2
        version: 4.8.2
      - name: special-resource-operator.v4.8.1
        version: 4.8.1
      - name: special-resource-operator.v4.8.0
        version: 4.8.0
      name: alpha
    defaultChannel: alpha
    packageName: special-resource-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Digital.ai
      provider-url: https://digital.ai
    name: digitalai-release-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: digitalai-release-operator.v25.3.0-beta.813
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 5,
                    "metrics": [],
                    "minReplicas": 3
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "cluster": {
                    "management": {
                      "ssl": {
                        "enabled": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    },
                    "manager": "xlr-legacy",
                    "mode": "full",
                    "name": "xl-release_cluster",
                    "node": {
                      "datacenter": "default"
                    },
                    "remote": {
                      "sslEngine": {
                        "enabled": false,
                        "enabledAlgorithms": [
                          "TLS_DHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_DHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_RSA_WITH_AES_128_CBC_SHA"
                        ],
                        "hostnameVerification": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "requireMutualAuthentication": false,
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"${XLR_CLUSTER_NAME}\"\n    manager = \"${XLR_CLUSTER_MANAGER}\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n      {{- if .Values.storage.reportingEngine.downingProviderClass }}\n      cluster {\n        downing-provider-class = {{.Values.storage.reportingEngine.downingProviderClass | quote}}\n      }\n      {{- end }}\n\n      {{- if .Values.cluster.remote.sslEngine.enabled }}\n      remote {\n        artery {  \n          ssl {  \n            config-ssl-engine {  \n              enabled-algorithms = [\n                {{- $last := sub (len .Values.cluster.remote.sslEngine.enabledAlgorithms) 1 }}\n                {{ range $index, $v := .Values.cluster.remote.sslEngine.enabledAlgorithms }}\n                {{ $v | quote }}{{ if ne $index $last }},{{ end }}\n                {{- end }}\n              ]  \n              key-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEY_PASSWORD}\"  \n              key-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.keystore \"default\" (printf \"cluster_remote_keystore.%s\" .Values.cluster.remote.sslEngine.keystoreType) \"context\" $) }}\"  \n              key-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEYSTORE_PASSWORD}\"  \n              trust-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.truststore \"default\" (printf \"cluster_remote_truststore.%s\" .Values.cluster.remote.sslEngine.truststoreType) \"context\" $) }}\"  \n              trust-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_TRUSTSTORE_PASSWORD}\"  \n              hostname-verification = {{ if .Values.cluster.remote.sslEngine.hostnameVerification }}on{{ else }}off{{ end }}\n              require-mutual-authentication = {{ if .Values.cluster.remote.sslEngine.requireMutualAuthentication }}on{{ else }}off{{ end }}  \n            }  \n          }  \n          transport = tls-tcp  \n        }  \n      }\n      {{- end }}\n    }\n    {{- if .Values.cluster.management.ssl.enabled }}\n    xlr-pekko-native {\n      pekko {\n        management {\n          http {  \n            route-providers-read-only = true  \n          }  \n          ssl {  \n            enabled = true\n            keyPassword = \"${?CLUSTER_MANAGEMENT_SSL_KEY_PASSWORD}\"  \n            keyStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.keystore \"default\" (printf \"cluster_management_keystore.%s\" .Values.cluster.management.ssl.keystoreType) \"context\" $) }}\"  \n            keyStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_KEYSTORE_PASSWORD}\"  \n            keystoreType = {{ .Values.cluster.management.ssl.keystoreType }}  \n            trustStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.truststore \"default\" (printf \"cluster_management_truststore.%s\" .Values.cluster.management.ssl.truststoreType) \"context\" $) }}\"  \n            trustStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_TRUSTSTORE_PASSWORD}\"  \n            trustStoreType = {{ .Values.cluster.management.ssl.truststoreType }}  \n          }  \n        }  \n      }\n    }\n    {{- end }}\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\nmanagement.endpoint.prometheus.enabled = {{ .Values.metrics.prometheus.enabled | default false }}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      datacenter = ${XLR_CLUSTER_DATACENTER}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "cd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "livenessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/liveness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "jmx": {
                      "enabled": false
                    },
                    "prometheus": {
                      "enabled": false,
                      "podAnnotations": {
                        "prometheus.io/port": "{{ .Values.service.ports.metrics }}",
                        "prometheus.io/scrape": "true"
                      },
                      "rule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "serviceMonitor": {
                        "annotations": {},
                        "basicAuth": {
                          "password": {
                            "key": "releasePassword",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          },
                          "username": {
                            "key": "releaseUsername",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          }
                        },
                        "enabled": false,
                        "honorLabels": false,
                        "interval": "30s",
                        "jobLabel": "",
                        "labels": {},
                        "metricRelabelings": [],
                        "namespace": "",
                        "params": {},
                        "path": "/s/actuator/prometheus",
                        "podTargetLabels": {},
                        "relabelings": [],
                        "scrapeTimeout": "",
                        "selector": {
                          "service-type": "normal"
                        },
                        "targetLabels": {}
                      }
                    }
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.0-debian-12-r4"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.0-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/extra/lib",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.5.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rbac": {
                    "create": true
                  },
                  "readinessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {
                      "service-type": "normal"
                    },
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "startupProbe": {
                    "enabled": true,
                    "failureThreshold": 12,
                    "initialDelaySeconds": 60,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 450,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnamilegacy/os-shell",
                      "tag": "12-debian-12-r49"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabsunsupported/release-operator@sha256:52914f17dc55dfaef55aed382e7269408fb676a213d12117eafa906d81506824 '
          createdAt: "2025-09-02T05:48:19Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql.\n\n```yaml\napiVersion:
          xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n
          \ global:\n    security:\n      allowInsecureImages: true\n  k8sSetup:\n
          \   platform: Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula:
          true\n  keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n```\n\n#### Configuration Details\n\nThe sample configuration
          uses:\n\n- Autogenerated limited license\n- Simple administrator initial
          password\n- Default keystore setup\n- Only one replica is set for all pods\n-
          All Pods are missing resource limit setup\n- Default storage class and minimal
          size for persistent storage\n- Release pods are using persistence access
          mode `ReadWriteOnce`, for Release pods the requirement is to have `ReadWriteMany`\n-
          Embedded PostgreSQL for DB management\n\n#### Security Configuration Details\n\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabsunsupported/release-operator@sha256:52914f17dc55dfaef55aed382e7269408fb676a213d12117eafa906d81506824
        - docker.io/xebialabsunsupported/xl-release@sha256:0548a1ba63da5ae09c59b67a30240bf48877ab8950e09b0f0f6fd7c88ba34c13
        version: 25.3.0-beta.813
      entries:
      - name: digitalai-release-operator.v25.3.0-beta.813
        version: 25.3.0-beta.813
      - name: digitalai-release-operator.v25.3.0-beta.711
        version: 25.3.0-beta.711
      - name: digitalai-release-operator.v25.3.0-beta.609
        version: 25.3.0-beta.609
      - name: digitalai-release-operator.v25.3.0-beta.512
        version: 25.3.0-beta.512
      name: beta
    - currentCSV: digitalai-release-operator.v24.1.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "fsGroup": null,
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "fsGroup": null,
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  jvm {\n      override-jvm-options = {{ .Values.jvm.overrideOptions }}\n      networkaddress.cache.ttl = {{ index .Values.jvm.options \"networkaddress.cache.ttl\" }}\n  }\n}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "ingressClass": "haproxy-dair",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-client/xl",
                        "kube",
                        "upgrade",
                        "--skip-context-check",
                        "--local-repo",
                        "/opt/xebialabs/xl-op-blueprints",
                        "--answers",
                        "/opt/xebialabs/xl-client/generated_answers.yaml"
                      ],
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    },
                    "overrideOptions": true
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "tag": "1.26.0-debian-12-r1"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "tag": "1.10.1-debian-12-r5"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "tag": "15.6.0-debian-12-r7"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "tag": "3.12.13-debian-12-r2"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnami/os-shell",
                      "tag": "12-debian-12-r16"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator without SCC creation and with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: docker.io/xebialabs/release-operator@sha256:47d6e31f66fce52c6566240b85fb028890409ad3addeea134120d1afa4814883
          createdAt: "2024-10-16T04:25:19Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind:
          DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    runAsUser: null\n    runAsGroup: null\n
          \   fsGroup: null\n  containerSecurityContext:\n    runAsUser: null\n    runAsGroup:
          null\n  securityContextConstraints:\n    enabled: false\n  volumePermissions:\n
          \   enabled: false\n  route:\n    enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n        runAsUser: null\n        runAsGroup: null\n        fsGroup:
          null\n      containerSecurityContext:\n        enabled: false\n        runAsUser:
          null\n        runAsGroup: null\n      securityContextConstraints:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n      runAsUser: null\n
          \     runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     enabled: false\n      runAsUser: null\n      runAsGroup: null\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"dair-doc-digitalai-release-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xlr-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Release pods are using
          persistence access mode `ReadWriteOnce`, for Release pods the requirement
          is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB management\n- Embedded
          RabbitMQ for message queue management\n\n#### Security Configuration Details\n\n-
          `securityContextConstraints.enabled: false` - disables creation of SCCs;\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:771a9a173e033a3ad8b46f5c00a7036eaa88c8d8d1fbd89217325168998113ea
        - docker.io/xebialabs/release-operator@sha256:47d6e31f66fce52c6566240b85fb028890409ad3addeea134120d1afa4814883
        - docker.io/xebialabs/xl-release@sha256:d1b65493ed524d37c60483f44b8bcc877ff398c81d36c7e767f2254dadec98cc
        version: 24.1.10
      entries:
      - name: digitalai-release-operator.v24.1.10
        version: 24.1.10
      name: stable
    - currentCSV: digitalai-release-operator.v24.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  jvm {\n      override-jvm-options = {{ .Values.jvm.overrideOptions }}\n      networkaddress.cache.ttl = {{ index .Values.jvm.options \"networkaddress.cache.ttl\" }}\n  }\n}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "image": {
                        "tag": "v0.14.7"
                      },
                      "ingressClass": "haproxy-dair",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-client/xl",
                        "kube",
                        "upgrade",
                        "--skip-context-check",
                        "--local-repo",
                        "/opt/xebialabs/xl-op-blueprints",
                        "--answers",
                        "/opt/xebialabs/xl-client/generated_answers.yaml"
                      ],
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    },
                    "overrideOptions": true
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.26.2-debian-12-r10"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "15.6.0-debian-12-r20"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.12.14-debian-12-r7"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnamilegacy/os-shell",
                      "tag": "12-debian-12-r49"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator without SCC creation and with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:aca76f0f5db170ddfc6ad692ea2874b9efb83fe732540946d88a0c078907ec8c '
          createdAt: "2025-08-22T08:54:25Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind:
          DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  securityContextConstraints:\n    enabled: false\n
          \ volumePermissions:\n    enabled: false\n  route:\n    enabled: false\n
          \   annotations:\n      haproxy.router.openshift.io/cookie_name: JSESSIONID\n
          \     haproxy.router.openshift.io/disable_cookies: 'false'\n      haproxy.router.openshift.io/rewrite-target:
          /\n      haproxy.router.openshift.io/timeout: 120s\n    hostname: '<mandatory-release-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n      securityContextConstraints:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   securityContextConstraints:\n      enabled: false\n    volumePermissions:\n
          \     enabled: false\n```\n\nIf SCC is required, it must be managed manually.
          The necessary permissions for accessing SecurityContextConstraints must
          be granted when SCC is enabled; otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"dair-doc-digitalai-release-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xlr-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Release pods are using
          persistence access mode `ReadWriteOnce`, for Release pods the requirement
          is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB management\n- Embedded
          RabbitMQ for message queue management\n\n#### Security Configuration Details\n\n-
          `securityContextConstraints.enabled: false` - disables creation of SCCs;\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/release-operator@sha256:aca76f0f5db170ddfc6ad692ea2874b9efb83fe732540946d88a0c078907ec8c
        - docker.io/xebialabs/xl-release@sha256:1cadd1d6ad7389ac9fe398ffba246b78238e54c180045b9dd44fdeefa3cf2af5
        version: 24.1.21
      entries:
      - name: digitalai-release-operator.v24.1.21
        version: 24.1.21
      - name: digitalai-release-operator.v24.1.20
        version: 24.1.20
      - name: digitalai-release-operator.v24.1.19
        version: 24.1.19
      - name: digitalai-release-operator.v24.1.18
        version: 24.1.18
      - name: digitalai-release-operator.v24.1.17
        version: 24.1.17
      - name: digitalai-release-operator.v24.1.14
        version: 24.1.14
      - name: digitalai-release-operator.v24.1.10
        version: 24.1.10
      name: stable-v24.1
    - currentCSV: digitalai-release-operator.v24.3.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "#!/bin/sh\n\ncd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "16.3.0-debian-12-r23"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n{{ if eq .Values.queueType \"quorum\" -}}\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n{{- end }}\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.13.7-debian-12-r5"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "queueType": "classic",
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-release",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:725fb5f7cfa4098c4110a40f54dc646f7663267bcba53603a7c1303c44051803 '
          createdAt: "2025-09-26T18:24:33Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n
          \ name: dair-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n```\n\n#### Configuration
          Details\n\nThe sample configuration uses:\n\n- Autogenerated limited license\n-
          Simple administrator initial password\n- Default keystore setup\n- Only
          one replica is set for all pods\n- All Pods are missing resource limit setup\n-
          Default storage class and minimal size for persistent storage\n- Release
          pods are using persistence access mode `ReadWriteOnce`, for Release pods
          the requirement is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB
          management\n- Embedded RabbitMQ for message queue management\n\n#### Security
          Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined ranges (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/release-operator@sha256:725fb5f7cfa4098c4110a40f54dc646f7663267bcba53603a7c1303c44051803
        - docker.io/xebialabs/xl-release@sha256:e9e87975888fcd4c8c164db90cb0a5c7abde88bf272c59d97e465af7b49bc6be
        version: 24.3.11
      entries:
      - name: digitalai-release-operator.v24.3.11
        version: 24.3.11
      - name: digitalai-release-operator.v24.3.10
        version: 24.3.10
      - name: digitalai-release-operator.v24.3.9
        version: 24.3.9
      - name: digitalai-release-operator.v24.3.8
        version: 24.3.8
      - name: digitalai-release-operator.v24.3.7
        version: 24.3.7
      - name: digitalai-release-operator.v24.3.6
        version: 24.3.6
      - name: digitalai-release-operator.v24.3.5
        version: 24.3.5
      name: stable-v24.3
    - currentCSV: digitalai-release-operator.v25.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "cluster": {
                    "management": {
                      "ssl": {
                        "enabled": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    },
                    "manager": "xlr-legacy",
                    "mode": "full",
                    "name": "xl-release_cluster",
                    "node": {
                      "datacenter": "default"
                    },
                    "remote": {
                      "sslEngine": {
                        "enabled": false,
                        "enabledAlgorithms": [
                          "TLS_DHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_DHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_RSA_WITH_AES_128_CBC_SHA"
                        ],
                        "hostnameVerification": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "requireMutualAuthentication": false,
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"${XLR_CLUSTER_NAME}\"\n    manager = \"${XLR_CLUSTER_MANAGER}\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n      {{- if .Values.storage.reportingEngine.downingProviderClass }}\n      cluster {\n        downing-provider-class = {{.Values.storage.reportingEngine.downingProviderClass | quote}}\n      }\n      {{- end }}\n\n      {{- if .Values.cluster.remote.sslEngine.enabled }}\n      remote {\n        artery {  \n          ssl {  \n            config-ssl-engine {  \n              enabled-algorithms = [\n                {{- $last := sub (len .Values.cluster.remote.sslEngine.enabledAlgorithms) 1 }}\n                {{ range $index, $v := .Values.cluster.remote.sslEngine.enabledAlgorithms }}\n                {{ $v | quote }}{{ if ne $index $last }},{{ end }}\n                {{- end }}\n              ]  \n              key-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEY_PASSWORD}\"  \n              key-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.keystore \"default\" (printf \"cluster_remote_keystore.%s\" .Values.cluster.remote.sslEngine.keystoreType) \"context\" $) }}\"  \n              key-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEYSTORE_PASSWORD}\"  \n              trust-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.truststore \"default\" (printf \"cluster_remote_truststore.%s\" .Values.cluster.remote.sslEngine.truststoreType) \"context\" $) }}\"  \n              trust-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_TRUSTSTORE_PASSWORD}\"  \n              hostname-verification = {{ if .Values.cluster.remote.sslEngine.hostnameVerification }}on{{ else }}off{{ end }}\n              require-mutual-authentication = {{ if .Values.cluster.remote.sslEngine.requireMutualAuthentication }}on{{ else }}off{{ end }}  \n            }  \n          }  \n          transport = tls-tcp  \n        }  \n      }\n      {{- end }}\n    }\n    {{- if .Values.cluster.management.ssl.enabled }}\n    xlr-pekko-native {\n      pekko {\n        management {\n          http {  \n            route-providers-read-only = true  \n          }  \n          ssl {  \n            enabled = true\n            keyPassword = \"${?CLUSTER_MANAGEMENT_SSL_KEY_PASSWORD}\"  \n            keyStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.keystore \"default\" (printf \"cluster_management_keystore.%s\" .Values.cluster.management.ssl.keystoreType) \"context\" $) }}\"  \n            keyStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_KEYSTORE_PASSWORD}\"  \n            keystoreType = {{ .Values.cluster.management.ssl.keystoreType }}  \n            trustStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.truststore \"default\" (printf \"cluster_management_truststore.%s\" .Values.cluster.management.ssl.truststoreType) \"context\" $) }}\"  \n            trustStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_TRUSTSTORE_PASSWORD}\"  \n            trustStoreType = {{ .Values.cluster.management.ssl.truststoreType }}  \n          }  \n        }  \n      }\n    }\n    {{- end }}\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\nmanagement.endpoint.prometheus.enabled = {{ .Values.metrics.prometheus.enabled | default false }}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      datacenter = ${XLR_CLUSTER_DATACENTER}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "#!/bin/sh\n\ncd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "livenessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/liveness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "jmx": {
                      "enabled": false
                    },
                    "prometheus": {
                      "enabled": false,
                      "podAnnotations": {
                        "prometheus.io/port": "{{ .Values.service.ports.metrics }}",
                        "prometheus.io/scrape": "true"
                      },
                      "rule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "serviceMonitor": {
                        "annotations": {},
                        "basicAuth": {
                          "password": {
                            "key": "releasePassword",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          },
                          "username": {
                            "key": "releaseUsername",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          }
                        },
                        "enabled": false,
                        "honorLabels": false,
                        "interval": "30s",
                        "jobLabel": "",
                        "labels": {},
                        "metricRelabelings": [],
                        "namespace": "",
                        "params": {},
                        "path": "/s/actuator/prometheus",
                        "podTargetLabels": {},
                        "relabelings": [],
                        "scrapeTimeout": "",
                        "selector": {
                          "service-type": "normal"
                        },
                        "targetLabels": {}
                      }
                    }
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.1-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.4.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.3-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "readinessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {
                      "service-type": "normal"
                    },
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "startupProbe": {
                    "enabled": true,
                    "failureThreshold": 12,
                    "initialDelaySeconds": 60,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 450,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-release",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:838443baf39514c5e0ac0a09cab411942a3cc4a13e261c8c60138788fe8d96e4 '
          createdAt: "2025-09-26T17:28:37Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n
          \ name: dair-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n```\n\n#### Configuration
          Details\n\nThe sample configuration uses:\n\n- Autogenerated limited license\n-
          Simple administrator initial password\n- Default keystore setup\n- Only
          one replica is set for all pods\n- All Pods are missing resource limit setup\n-
          Default storage class and minimal size for persistent storage\n- Release
          pods are using persistence access mode `ReadWriteOnce`, for Release pods
          the requirement is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB
          management\n- Embedded RabbitMQ for message queue management\n\n#### Security
          Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined ranges (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/release-operator@sha256:838443baf39514c5e0ac0a09cab411942a3cc4a13e261c8c60138788fe8d96e4
        - docker.io/xebialabs/xl-release@sha256:9b63789cdb87745832f1283d3f0e28ac5a4514d767895e0914a6be6672f2d14f
        version: 25.1.5
      entries:
      - name: digitalai-release-operator.v25.1.5
        version: 25.1.5
      - name: digitalai-release-operator.v25.1.4
        version: 25.1.4
      - name: digitalai-release-operator.v25.1.3
        version: 25.1.3
      - name: digitalai-release-operator.v25.1.2
        version: 25.1.2
      - name: digitalai-release-operator.v25.1.1
        version: 25.1.1
      - name: digitalai-release-operator.v25.1.0
        version: 25.1.0
      name: stable-v25.1
    defaultChannel: stable-v24.3
    packageName: digitalai-release-operator
    provider:
      name: Digital.ai
      url: https://digital.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Confluent, Inc.
      provider-url: https://confluent.io
    name: confluent-for-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: confluent-for-kubernetes.v2.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.8.4",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:0ca8769b04d6134ee231af84dfe98fb65161148d0d32e85f18dad6a3a345ed33
          createdAt: "2025-10-01T13:12:32Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0 <2.10.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - confluentinc/confluent-operator@sha256:0ca8769b04d6134ee231af84dfe98fb65161148d0d32e85f18dad6a3a345ed33
        - docker.io/confluentinc/confluent-operator@sha256:0ca8769b04d6134ee231af84dfe98fb65161148d0d32e85f18dad6a3a345ed33
        - docker.io/confluentinc/cp-server@sha256:45eb6e4f9c1201202d90216fbbdfcd8655eb378c937f09ac0129cd259956337d
        - docker.io/confluentinc/cp-server-connect@sha256:24aaa1cc8eb0df643f1efa37999b7ac33d844bfd1e4502da0d678996e4ac5ddd
        - docker.io/confluentinc/confluent-init-container@sha256:175ba68c876eca5804597b1591fdf551b2091d2f1c50e4daed2c32958af5d649
        - docker.io/confluentinc/cp-zookeeper@sha256:3e0af6ac6d34dcb8b9ed8f7a08109a48b2b2acd331c1dc4ea875222685c9afbd
        - docker.io/confluentinc/cp-schema-registry@sha256:bba639a3f9e2bf5f99b5d056ed520d2c268a125bee4f546d6cbca7a2c58f7dbb
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:2024d117dd413e207e49f37558d79dc882fd29f6b39e2d722708f18076813361
        - docker.io/confluentinc/cp-kafka-rest@sha256:8218e8db4731219ceaf9dbc378ead9c49aa86f98bc5106f62fbf70958dfd1ee4
        - docker.io/confluentinc/cp-ksqldb-server@sha256:093e87359c176d13d81317e972888a835cd01092d061bfc03fd60f0d1b4c8428
        version: 2.10.3
      entries:
      - name: confluent-for-kubernetes.v2.10.3
        version: 2.10.3
      name: "2.10"
    - currentCSV: confluent-for-kubernetes.v2.11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.9.3",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:85b512d3f20dc6e4a4da1759ee5ed306a4da0604592563a9af35be2cdf9ba1b6
          createdAt: "2025-10-01T13:10:39Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.10.0 <2.11.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/confluent-operator@sha256:85b512d3f20dc6e4a4da1759ee5ed306a4da0604592563a9af35be2cdf9ba1b6
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:351b68b062824c15db7ef999ba04062defdd4ea1ebe9665932147f99cf74879d
        - docker.io/confluentinc/cp-ksqldb-server@sha256:14a44cbbc13094974e7a5e14b33a829480ec3da75868130fdc7abaeda707dc5e
        - confluentinc/confluent-operator@sha256:85b512d3f20dc6e4a4da1759ee5ed306a4da0604592563a9af35be2cdf9ba1b6
        - docker.io/confluentinc/cp-server@sha256:e61e6c122b50ed345bc3e4e39cda6c93430e669535f8a70957e221563a7af1bd
        - docker.io/confluentinc/confluent-init-container@sha256:bccc7b723afdbe4e04c668720cdb1570835f8fd58ab7dd9a06ecf807f1176019
        - docker.io/confluentinc/cp-kafka-rest@sha256:cc72b3585ef9e7557910ab3efe3082444abe1fee191cdfb8747907fd2643971f
        - docker.io/confluentinc/cp-schema-registry@sha256:bd974dda5abf35fcad66c25fad92ee684e82305435b291b953d221f74d2bc119
        - docker.io/confluentinc/cp-zookeeper@sha256:03eadde2afa390f7339707e6baef2ecf7affb898aaa432378f61da54f5636ba3
        - docker.io/confluentinc/cp-server-connect@sha256:0d3b81e511cdf820daac4a2458b3cada40b4a7c3d142062a48880d38186bbcc4
        version: 2.11.3
      entries:
      - name: confluent-for-kubernetes.v2.11.3
        version: 2.11.3
      name: "2.11"
    - currentCSV: confluent-for-kubernetes.v2.8.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.6.4",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:551b14777b121b1e083524a0d5cce03cbe2fd2833537751bcd6be6cf9801f95e
          createdAt: "2024-12-02T06:51:43Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.0 <2.8.5'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-server@sha256:1b2ae325f87047b83d37d506066f073c27b0c42a0fd3c3409aa079ba736d1975
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:01bcde2aa35c0ad062ece49c63d41bf487618ea3c33e42ea4c23fde7194b26c9
        - docker.io/confluentinc/cp-schema-registry@sha256:dec88d30c8f21bc7fb38d28603c97c81c666ccaee63ff0136ecff48b3b3ec775
        - docker.io/confluentinc/confluent-init-container@sha256:b1ad1aa4daa81967a93b02323172f96db6a50c7135866f016344b2eb1a736a55
        - docker.io/confluentinc/cp-server-connect@sha256:255875bc57d81c322d47d5f6f11bc217e99ae709d1fdc5aec8e11829dd34279a
        - confluentinc/confluent-operator@sha256:551b14777b121b1e083524a0d5cce03cbe2fd2833537751bcd6be6cf9801f95e
        - docker.io/confluentinc/cp-zookeeper@sha256:7e728feebfc90a693d028216da68a5d79e5e8212855c63bac08e5b10b141d143
        - docker.io/confluentinc/cp-kafka-rest@sha256:2a5bc0d39693682e25d79f5036e18b71376df0941fedc2cb962a4b4d5a5eff82
        - docker.io/confluentinc/cp-ksqldb-server@sha256:50e4834c44759ab4a2a0bbab0f20545e124325b7ee12ba1a275af73b57084104
        - docker.io/confluentinc/confluent-operator@sha256:551b14777b121b1e083524a0d5cce03cbe2fd2833537751bcd6be6cf9801f95e
        version: 2.8.5
      entries:
      - name: confluent-for-kubernetes.v2.8.5
        version: 2.8.5
      name: "2.8"
    - currentCSV: confluent-for-kubernetes.v2.9.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.7.5",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:20648758fdb98d0f8913e75fb919ad0b622de8ecda9ab394912103e5e17d4007
          createdAt: "2025-10-03T06:52:55Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0 <2.9.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-server-connect@sha256:560833472dbdf23cf85b780e8306ae1bfe16ce6e2c0a9b326c64a513117525f1
        - docker.io/confluentinc/cp-zookeeper@sha256:2bd46d1f5d1eec0675bb5d1c08634d6458bbac7cacfb233aa61c68137fd4a7ce
        - docker.io/confluentinc/cp-kafka-rest@sha256:b6ac629a880474d81182eaf8459b4d83fe8126db796f1f285798beecda55a50a
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:257c38a75b3f4241780e20f16a7adc026dcca590f2e9c8da7c4e86b961e7611f
        - confluentinc/confluent-operator@sha256:20648758fdb98d0f8913e75fb919ad0b622de8ecda9ab394912103e5e17d4007
        - docker.io/confluentinc/confluent-operator@sha256:20648758fdb98d0f8913e75fb919ad0b622de8ecda9ab394912103e5e17d4007
        - docker.io/confluentinc/cp-ksqldb-server@sha256:6817a555ae4c041869f9b422a115c22894b2cd3988e82dc601694ce713b3ccd4
        - docker.io/confluentinc/confluent-init-container@sha256:b138d954318cc98767c10b4564a1412a22766317a83b9a452285bb0b3d893c63
        - docker.io/confluentinc/cp-server@sha256:79243e7e8b721f9fdf26fc57330ebb27d69c3b4686dbdf5319f6f39c2c7be836
        - docker.io/confluentinc/cp-schema-registry@sha256:6506a8c659d4154ae3e2d702f08b4ccac63513392ba454cbc9cf384d070c46eb
        version: 2.9.7
      entries:
      - name: confluent-for-kubernetes.v2.9.7
        version: 2.9.7
      name: "2.9"
    - currentCSV: confluent-for-kubernetes.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:8.0.1",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-ng-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "alertManagerClient": {
                      "url": "http://controlcenter-next-gen.confluent.svc.cluster.local:9093"
                    },
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    },
                    "prometheusClient": {
                      "url": "http://controlcenter-next-gen.confluent.svc.cluster.local:9090"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center-next-gen:2.2.0",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "replicas": 1,
                  "services": {
                    "alertmanager": {
                      "image": "confluentinc/cp-enterprise-alertmanager:2.2.0"
                    },
                    "prometheus": {
                      "image": "confluentinc/cp-enterprise-prometheus:2.2.0",
                      "pvc": {
                        "dataVolumeCapacity": "10Gi"
                      }
                    }
                  },
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:8.0.1",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:8.0.1",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:8.0.1",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:8.0.1",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:8.0.1",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:55b61717690fe334b5eebf5e7a55590537abddbd404f2eede81d4dd8cd29eda4
          createdAt: "2025-10-03T07:59:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <3.0.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-kafka-rest@sha256:f45f23e891ae073037ed7670659ecb6c3f72f9eea33e109f16243eec24a287a3
        - docker.io/confluentinc/cp-server-connect@sha256:4739f3a07d9f8eadd58f448a8e05b1392ef9b1dc9e9085d25a71f32b5b806a6e
        - docker.io/confluentinc/cp-ksqldb-server@sha256:df94bb805cff4f92315135a042f650c42f887bca70a40df8ecef2603532f11a7
        - docker.io/confluentinc/cp-enterprise-prometheus@sha256:db0989ef9859dbc88bc59f9b8ade86f2ba13e9dc36533e654566218328e5ff57
        - docker.io/confluentinc/cp-enterprise-control-center-next-gen@sha256:6378a0174433685ee52a7fe2e48424d3f598827ed2a5570ec74b916102485e0d
        - confluentinc/confluent-operator@sha256:55b61717690fe334b5eebf5e7a55590537abddbd404f2eede81d4dd8cd29eda4
        - docker.io/confluentinc/confluent-operator@sha256:55b61717690fe334b5eebf5e7a55590537abddbd404f2eede81d4dd8cd29eda4
        - docker.io/confluentinc/cp-schema-registry@sha256:2152975dfb62afbdf1d438dd0eb1f7fc3c929adf5804fe4790eccda0c96b9208
        - docker.io/confluentinc/cp-server@sha256:207cabc12e0d122199da875b2b0e39d1cf6f7fe435d5f184a9c8d7dc57ea81d9
        - docker.io/confluentinc/cp-enterprise-alertmanager@sha256:fcce6f5311c26cd58427a039530a9f8dfd009249ff5a519694aa945d34502252
        - docker.io/confluentinc/confluent-init-container@sha256:bf38b3bd9851bfd3ef0635e93b9e8b9ada570d393f650f60d8ae8943537730dc
        version: 3.0.1
      entries:
      - name: confluent-for-kubernetes.v3.0.1
        version: 3.0.1
      name: "3.0"
    defaultChannel: "3.0"
    packageName: confluent-for-kubernetes
    provider:
      name: Confluent, Inc.
      url: https://confluent.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pachyderm, Inc.
      provider-url: https://pachyderm.com
    name: pachyderm-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pachyderm-operator.v0.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "Pachyderm",
                "metadata": {
                  "name": "pachyderm-sample"
                },
                "spec": {
                  "console": {
                    "disable": true
                  },
                  "pachd": {
                    "metrics": {
                      "disable": false
                    },
                    "storage": {
                      "amazon": {
                        "credentialSecretName": "pachyderm-aws-secret"
                      },
                      "backend": "AMAZON"
                    }
                  },
                  "version": "v2.1.6"
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermExport",
                "metadata": {
                  "name": "pachydermexport-sample"
                },
                "spec": {
                  "backup": {
                    "target": "pachyderm-sample"
                  }
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermImport",
                "metadata": {
                  "name": "pachydermimport-sample"
                },
                "spec": {
                  "backup": "backups/pachyderm-backup-20220610.tar.gz",
                  "destination": {
                    "name": "pachyderm-restore",
                    "namespace": "pachyderm-workloads"
                  },
                  "storageSecret": "s3-credentials-secret"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
          createdAt: "2021-05-20T08:05:00Z"
          description: Pachyderm provides the data layer that allows data science
            teams to productionize and scale their machine learning lifecycle with
            data driven automation, petabyte scalability and end-to-end reproducibility.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/pachyderm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/pachyderm-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pachyderm/openshift-operator
          support: Pachyderm, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PachydermExport is the Schema for the pachydermexports API
            displayName: Pachyderm Export
            kind: PachydermExport
            name: pachydermexports.aiml.pachyderm.com
            version: v1beta1
          - description: PachydermImport is the Schema for the pachydermimports API
            displayName: Pachyderm Import
            kind: PachydermImport
            name: pachydermimports.aiml.pachyderm.com
            version: v1beta1
          - description: Pachyderm is the Schema for the pachyderms API
            displayName: Pachyderm
            kind: Pachyderm
            name: pachyderms.aiml.pachyderm.com
            version: v1beta1
        description: "The Pachyderm operator aims to make deploying and managing Pachyderm
          instances on Openshift container platform easier.\n\nPachyderm is the data
          foundation for machine learning.  Pachyderm provides industry leading data
          versioning, pipelines and lineage that allow data science teams to automate
          the machine learning lifecycle and optimize their machine learning operations
          (MLOps). \n\n## Getting started\n\n1. Pachyderm uses an object store to
          store your data. Set the pachd storage backend at `pachyderm.spec.pachd.storage.backend`
          to `AMAZON`, `MICROSOFT`, `GOOGLE` or `MINIO` depending on your storage
          provider. Additionally, create a user with read/write access to the object
          store.\n\n2. Setup the user credentials in the section, `pachyderm.spec.pachd.storage`.
          Required fields depend on the backend type set in step 1 above.\n\n3. Finally,
          deploy the pachyderm cluster in its own namespace.\n\nFor more information
          on getting started, reference the documentation.\n"
        displayName: Pachyderm
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pachyderm
        - versioning
        - machine learning
        - artificial intelligence
        links:
        - name: Documentation
          url: https://docs.pachyderm.com/latest/
        maintainers:
        - email: kyle@pachyderm.io
          name: Kyle White
        - email: eochieng@redhat.com
          name: Edmund Ochieng
        maturity: beta
        minKubeVersion: v1.19.0
        provider:
          name: Pachyderm, Inc.
          url: https://pachyderm.com
        relatedImages:
        - registry.connect.redhat.com/pachyderm/worker@sha256:374f83d92899707bc87b0318fda6bffa78d2e9e523b827bb3d40693a2d1bdecd
        - registry.connect.redhat.com/pachyderm/etcd@sha256:7f330b51874379b1d4b71d94e37e0ecfa52b86d628296136970688cf204f53b5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
        - registry.connect.redhat.com/pachyderm/backup-handler@sha256:e2ef6bd8191e8ad58c765c0db8f5bb21ea7123caed0c0f58fe681a024201306d
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/pachd@sha256:116477fefdf413ea02c6ac3d013f1f91d3f6e4b16aed2408c9890832cdb41430
        - registry.redhat.io/rhel8/postgresql-13@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/init-utils@sha256:5c0258bfa0203079925648c0a7dda6d05a29b6f461854821a1eec541777ff359
        - registry.connect.redhat.com/pachyderm/pgbouncer@sha256:c3b6949227df24abb8934d111a34b44e87fb16194ac34a5d0fe284cff02e9ac0
        version: 0.1.4
      entries:
      - name: pachyderm-operator.v0.1.4
        version: 0.1.4
      name: stable
    defaultChannel: stable
    packageName: pachyderm-operator-rhmp
    provider:
      name: Pachyderm, Inc.
      url: https://pachyderm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-community-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.5.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "registry": "docker.io",
                  "tag": "5.4.6",
                  "oem": "",
                  "imagePullSecrets": "",
                  "psp": false,
                  "rbac": true,
                  "serviceAccount": "default",
                  "leastPrivilege": false,
                  "global": {
                    "cattle": {
                      "url": ""
                    },
                    "azure": {
                      "enabled": false,
                      "identity": {
                        "clientId": "DONOTMODIFY"
                      },
                      "marketplace": {
                        "planId": "DONOTMODIFY"
                      },
                      "extension": {
                        "resourceId": "DONOTMODIFY"
                      },
                      "serviceAccount": "csp",
                      "imagePullSecrets": "",
                      "images": {
                        "neuvector_csp_pod": {
                          "tag": "latest",
                          "image": "neuvector-billing-azure-by-suse-llc",
                          "registry": "registry.suse.de/suse/sle-15-sp5/update/pubclouds/images",
                          "imagePullPolicy": "IfNotPresent"
                        },
                        "controller": {
                          "tag": "5.2.4",
                          "image": "controller",
                          "registry": "docker.io/neuvector"
                        },
                        "manager": {
                          "tag": "5.2.4",
                          "image": "manager",
                          "registry": "docker.io/neuvector"
                        },
                        "enforcer": {
                          "tag": "5.2.4",
                          "image": "enforcer",
                          "registry": "docker.io/neuvector"
                        }
                      }
                    },
                    "aws": {
                      "enabled": false,
                      "accountNumber": "",
                      "roleName": "",
                      "serviceAccount": "csp",
                      "annotations": {},
                      "imagePullSecrets": "",
                      "image": {
                        "digest": "",
                        "repository": "neuvector/neuvector-csp-adapter",
                        "tag": "latest",
                        "imagePullPolicy": "IfNotPresent"
                      }
                    }
                  },
                  "bootstrapPassword": "",
                  "autoGenerateCert": true,
                  "defaultValidityPeriod": 365,
                  "internal": {
                    "certmanager": {
                      "enabled": false,
                      "secretname": "neuvector-internal"
                    },
                    "autoGenerateCert": true,
                    "autoRotateCert": true
                  },
                  "controller": {
                    "enabled": true,
                    "annotations": {},
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": {
                      "repository": "neuvector/controller",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "searchRegistries": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "nodePort": "",
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      }
                    },
                    "ranchersso": {
                      "enabled": false
                    },
                    "pvc": {
                      "enabled": false,
                      "existingClaim": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {
                        "userinitcfg.yaml": {
                          "users": [
                            {
                              "Fullname": "admin",
                              "Password": "",
                              "Role": "admin"
                            }
                          ]
                        }
                      }
                    },
                    "certupgrader": {
                      "env": [],
                      "schedule": "",
                      "imagePullPolicy": "IfNotPresent",
                      "timeout": 3600,
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "prime": {
                      "enabled": false,
                      "image": {
                        "repository": "neuvector/compliance-config",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "1.0.7",
                        "hash": ""
                      }
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": {
                      "repository": "neuvector/enforcer",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "env": [],
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/etcd"
                      }
                    ],
                    "resources": {},
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "manager": {
                    "enabled": true,
                    "image": {
                      "repository": "neuvector/manager",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "priorityClassName": "",
                    "env": {
                      "ssl": true,
                      "envs": []
                    },
                    "svc": {
                      "type": "ClusterIP",
                      "nodePort": "",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": "",
                      "tls": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "topologySpreadConstraints": [],
                    "affinity": {},
                    "podLabels": {},
                    "podAnnotations": {},
                    "tolerations": [],
                    "nodeSelector": {},
                    "runAsUser": "",
                    "probes": {
                      "enabled": false,
                      "timeout": 1,
                      "periodSeconds": 10,
                      "startupFailureThreshold": 30
                    }
                  },
                  "cve": {
                    "adapter": {
                      "enabled": false,
                      "image": {
                        "repository": "neuvector/registry-adapter",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "0.1.9",
                        "hash": ""
                      },
                      "priorityClassName": "",
                      "resources": {},
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt"
                      },
                      "harbor": {
                        "protocol": "https",
                        "secretName": ""
                      },
                      "svc": {
                        "type": "ClusterIP",
                        "loadBalancerIP": "",
                        "annotations": {}
                      },
                      "route": {
                        "enabled": true,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      },
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ingressClassName": "",
                        "path": "/",
                        "annotations": {
                          "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                        },
                        "tls": false,
                        "secretName": ""
                      },
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      }
                    },
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "cacert": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                      "image": {
                        "registry": "",
                        "repository": "neuvector/updater",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "0.0.5",
                        "hash": ""
                      },
                      "schedule": "0 0 * * *",
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": {
                        "registry": "",
                        "repository": "neuvector/scanner",
                        "imagePullPolicy": "Always",
                        "tag": "6",
                        "hash": ""
                      },
                      "priorityClassName": "",
                      "resources": {},
                      "topologySpreadConstraints": [],
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      },
                      "volumes": "",
                      "volumeMounts": ""
                    }
                  },
                  "resources": {},
                  "runtimePath": "",
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhooksvc": {
                    "enabled": true
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  },
                  "lease": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "false"
          containerImage: docker.io/neuvector/neuvector-operator@sha256:6815df6435eafaf7b2bc9d5e3f0c9b65f0a50ad4e05d2862e56b227e19aeca27
          createdAt: "2019-11-18T02:09:59Z"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: NeuVector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart. NeuVector Operator version 1.5.5 deploys NeuVector app version 5.4.6.

          **Complete below steps to create namespace and Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Privileged SCC is added to Service Account specified in the deploment yaml by Operator version 1.3.4 and above in new deployment. In case of upgrading NeuVector Operator from previous version to 1.3.4, please delete Privileged SCC before upgrading.

                   oc delete rolebinding -nneuvector system:openshift:scc:privileged

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s  neuvector/default


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - wavefront
        - wavefront collector
        - wavefront proxy
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - docker.io/neuvector/neuvector-operator@sha256:6815df6435eafaf7b2bc9d5e3f0c9b65f0a50ad4e05d2862e56b227e19aeca27
        version: 1.5.5
      entries:
      - name: neuvector-operator.v1.5.5
        version: 1.5.5
      - name: neuvector-operator.v1.5.4
        version: 1.5.4
      - name: neuvector-operator.v1.5.3
        version: 1.5.3
      - name: neuvector-operator.v1.5.2
        version: 1.5.2
      - name: neuvector-operator.v1.5.1
        version: 1.5.1
      - name: neuvector-operator.v1.5.0
        version: 1.5.0
      - name: neuvector-operator.v1.4.9
        version: 1.4.9
      - name: neuvector-operator.v1.4.8
        version: 1.4.8
      - name: neuvector-operator.v1.4.7
        version: 1.4.7
      - name: neuvector-operator.v1.4.6
        version: 1.4.6
      - name: neuvector-operator.v1.4.5
        version: 1.4.5
      - name: neuvector-operator.v1.4.4
        version: 1.4.4
      - name: neuvector-operator.v1.4.3
        version: 1.4.3
      - name: neuvector-operator.v1.4.2
        version: 1.4.2
      - name: neuvector-operator.v1.4.1
        version: 1.4.1
      - name: neuvector-operator.v1.4.0
        version: 1.4.0
      - name: neuvector-operator.v1.3.9
        version: 1.3.9
      - name: neuvector-operator.v1.3.8
        version: 1.3.8
      - name: neuvector-operator.v1.3.7
        version: 1.3.7
      - name: neuvector-operator.v1.3.6
        version: 1.3.6
      - name: neuvector-operator.v1.3.5
        version: 1.3.5
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-community-operator
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: mariadb-operator
      provider-url: https://github.com/mariadb-operator/mariadb-operator
    name: mariadb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-operator.v25.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.mariadb.mmontes.io/v1alpha1",
                "kind": "MariadbOperator",
                "metadata": {
                  "name": "mariadb-operator"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "caValidity": "35064h",
                    "certValidity": "8766h",
                    "enabled": true,
                    "extrArgs": [],
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "ha": {
                      "enabled": false,
                      "replicas": 3
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "lookaheadValidity": "2160h",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podSecurityContext": {},
                    "requeueDuration": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "automount": true,
                      "enabled": true,
                      "extraLabels": {},
                      "name": ""
                    },
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    },
                    "tolerations": []
                  },
                  "clusterName": "cluster.local",
                  "extrArgs": [],
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "fullnameOverride": "",
                  "ha": {
                    "enabled": false,
                    "replicas": 3
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "logLevel": "INFO",
                  "metrics": {
                    "enabled": false,
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "rbac": {
                    "enabled": true
                  },
                  "resources": {},
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "automount": true,
                    "enabled": true,
                    "extraLabels": {},
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "annotations": {},
                    "cert": {
                      "caPath": "/tmp/k8s-webhook-server/certificate-authority",
                      "certManager": {
                        "duration": "",
                        "enabled": false,
                        "issuerRef": {},
                        "renewBefore": ""
                      },
                      "path": "/tmp/k8s-webhook-server/serving-certs",
                      "secretAnnotations": {}
                    },
                    "extrArgs": [],
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "ha": {
                      "enabled": false,
                      "replicas": 3
                    },
                    "hostNetwork": false,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podSecurityContext": {},
                    "port": 10250,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "automount": true,
                      "enabled": true,
                      "extraLabels": {},
                      "name": ""
                    },
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    },
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup"
                },
                "spec": {
                  "args": [
                    "--single-transaction",
                    "--all-databases"
                  ],
                  "logLevel": "info",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "backups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Connection",
                "metadata": {
                  "name": "connection"
                },
                "spec": {
                  "database": "mariadb",
                  "healthCheck": {
                    "interval": "10s",
                    "retryInterval": "3s"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "params": {
                    "parseTime": "true"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "secretName": "connection",
                  "secretTemplate": {
                    "annotations": {
                      "k8s.mariadb.com/connection": "sample"
                    },
                    "databaseKey": "database",
                    "hostKey": "host",
                    "key": "dsn",
                    "labels": {
                      "k8s.mariadb.com/connection": "sample"
                    },
                    "passwordKey": "password",
                    "portKey": "port",
                    "usernameKey": "username"
                  },
                  "serviceName": "mariadb",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Database",
                "metadata": {
                  "name": "data-test"
                },
                "spec": {
                  "characterSet": "utf8",
                  "collate": "utf8_general_ci",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "retryInterval": "5s"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "ExternalMariaDB",
                "metadata": {
                  "name": "external-mariadb"
                },
                "spec": {
                  "connection": {
                    "healthCheck": {
                      "interval": "5s"
                    },
                    "secretName": "external-mariadb"
                  },
                  "host": "mariadb.default.svc.cluster.local",
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "port": 3306,
                  "username": "root"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "grant"
                },
                "spec": {
                  "database": "*",
                  "grantOption": true,
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "privileges": [
                    "SELECT",
                    "INSERT",
                    "UPDATE"
                  ],
                  "requeueInterval": "30s",
                  "retryInterval": "5s",
                  "table": "*",
                  "username": "user"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "MariaDB",
                "metadata": {
                  "name": "mariadb"
                },
                "spec": {
                  "connection": {
                    "secretName": "mariadb-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "database": "mariadb",
                  "galera": {
                    "enabled": true
                  },
                  "metrics": {
                    "enabled": true,
                    "passwordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "mariadb-metrics"
                    }
                  },
                  "myCnf": "[mariadb]\nbind-address=*\ndefault_storage_engine=InnoDB\nbinlog_format=row\ninnodb_autoinc_lock_mode=2\ninnodb_buffer_pool_size=1024M\nmax_allowed_packet=256M\n",
                  "passwordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-password"
                  },
                  "primaryConnection": {
                    "secretName": "mariadb-conn-primary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "primaryService": {
                    "type": "ClusterIP"
                  },
                  "replicas": 3,
                  "rootPasswordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-root"
                  },
                  "secondaryConnection": {
                    "secretName": "mariadb-conn-secondary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "secondaryService": {
                    "type": "ClusterIP"
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "storage": {
                    "size": "1Gi"
                  },
                  "updateStrategy": {
                    "type": "ReplicasFirstPrimaryLast"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "MaxScale",
                "metadata": {
                  "name": "maxscale-galera"
                },
                "spec": {
                  "admin": {
                    "guiEnabled": true,
                    "port": 8989
                  },
                  "auth": {
                    "generate": true
                  },
                  "config": {
                    "sync": {
                      "database": "mysql",
                      "interval": "5s",
                      "timeout": "10s"
                    }
                  },
                  "connection": {
                    "port": 3306,
                    "secretName": "mxs-galera-conn"
                  },
                  "guiKubernetesService": {
                    "metadata": {
                      "annotations": {
                        "metallb.universe.tf/loadBalancerIPs": "172.18.0.231"
                      }
                    },
                    "type": "LoadBalancer"
                  },
                  "kubernetesService": {
                    "annotations": {
                      "metallb.universe.tf/loadBalancerIPs": "172.18.0.224"
                    },
                    "type": "LoadBalancer"
                  },
                  "mariaDbRef": {
                    "name": "mariadb-galera"
                  },
                  "monitor": {
                    "cooperativeMonitoring": "majority_of_all",
                    "interval": "2s",
                    "params": {
                      "available_when_donor": "false",
                      "disable_master_failback": "false",
                      "disable_master_role_setting": "false"
                    }
                  },
                  "replicas": 3,
                  "requeueInterval": "10s",
                  "services": [
                    {
                      "listener": {
                        "params": {
                          "connection_metadata": "tx_isolation=auto"
                        },
                        "port": 3306,
                        "protocol": "MariaDBProtocol"
                      },
                      "name": "rw-router",
                      "params": {
                        "master_accept_reads": "true",
                        "max_replication_lag": "3s",
                        "max_slave_connections": "255",
                        "transaction_replay": "true",
                        "transaction_replay_attempts": "10",
                        "transaction_replay_timeout": "5s"
                      },
                      "router": "readwritesplit"
                    },
                    {
                      "listener": {
                        "port": 3307
                      },
                      "name": "rconn-master-router",
                      "params": {
                        "master_accept_reads": "true",
                        "max_replication_lag": "3s",
                        "router_options": "master"
                      },
                      "router": "readconnroute"
                    },
                    {
                      "listener": {
                        "port": 3308
                      },
                      "name": "rconn-slave-router",
                      "params": {
                        "max_replication_lag": "3s",
                        "router_options": "slave"
                      },
                      "router": "readconnroute"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "PhysicalBackup",
                "metadata": {
                  "name": "physicalbackup"
                },
                "spec": {
                  "compression": "bzip2",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "podAffinity": true,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "immediate": true,
                    "suspend": false
                  },
                  "serviceAccountName": "backup",
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "physicalbackups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "region": "us-east-1",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  },
                  "timeout": "1h"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore"
                },
                "spec": {
                  "backupRef": {
                    "name": "backup"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "targetRecoveryTime": "2023-12-19T09:00:00Z"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "SqlJob",
                "metadata": {
                  "name": "03-stars"
                },
                "spec": {
                  "database": "mariadb",
                  "dependsOn": [
                    {
                      "name": "01-users"
                    },
                    {
                      "name": "02-repos"
                    }
                  ],
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "sql": "CREATE TABLE IF NOT EXISTS stars (\n  id bigint PRIMARY KEY AUTO_INCREMENT,\n  user_id bigint NOT NULL,\n  repo_id bigint NOT NULL,\n  FOREIGN KEY (user_id) REFERENCES users(id) ON DELETE CASCADE,\n  FOREIGN KEY (repo_id) REFERENCES repos(id) ON DELETE CASCADE,\n  UNIQUE KEY (user_id, repo_id)\n);\nINSERT INTO stars(user_id, repo_id) \n  VALUES((SELECT id FROM users ORDER BY RAND() LIMIT 1), (SELECT id FROM repos ORDER BY RAND() LIMIT 1))\n  ON DUPLICATE KEY UPDATE id=id;\nDELETE FROM stars WHERE id = (SELECT id FROM stars ORDER BY RAND() LIMIT 1);\nSELECT r.name AS repo, COUNT(*) AS stars\nFROM stars s\nJOIN repos r\nON s.repo_id = r.id\nGROUP BY r.id\nORDER BY stars DESC;",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "user"
                },
                "spec": {
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxUserConnections": 20,
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "retryInterval": "5s"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: ghcr.io/mariadb-operator/mariadb-operator-helm:25.8.4
          createdAt: "2025-09-19T13:24:35Z"
          description: Run and operate MariaDB in a cloud native way
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mariadb-operator/mariadb-operator
          support: mariadb-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configures a backup
            displayName: Backup
            kind: Backup
            name: backups.k8s.mariadb.com
            version: v1alpha1
          - description: Configures a connection
            displayName: Connection
            kind: Connection
            name: connections.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a logical database
            displayName: Database
            kind: Database
            name: databases.k8s.mariadb.com
            version: v1alpha1
          - description: Configures an external MariaDB resources
            displayName: ExternalMariaDB
            kind: ExternalMariaDB
            name: externalmariadbs.k8s.mariadb.com
            version: v1alpha1
          - description: Grants permissions to an user in a database
            displayName: Grant
            kind: Grant
            name: grants.k8s.mariadb.com
            version: v1alpha1
          - description: Configures MariaDB helm chart based operator
            displayName: MariadbOperator
            kind: MariadbOperator
            name: mariadboperators.helm.mariadb.mmontes.io
            version: v1alpha1
          - description: Provisions a MariaDB instance
            displayName: MariaDB
            kind: MariaDB
            name: mariadbs.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a MaxScale database proxy
            displayName: MaxScale
            kind: MaxScale
            name: maxscales.k8s.mariadb.com
            version: v1alpha1
          - description: Configures a physical backup
            displayName: PhysicalBackup
            kind: PhysicalBackup
            name: physicalbackups.k8s.mariadb.com
            version: v1alpha1
          - description: Restores a backup
            displayName: Restore
            kind: Restore
            name: restores.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a SQL job
            displayName: SqlJob
            kind: SqlJob
            name: sqljobs.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a user
            displayName: User
            kind: User
            name: users.k8s.mariadb.com
            version: v1alpha1
        description: |
          Install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) via [OLM](https://olm.operatorframework.io/) using the [helm chart](https://artifacthub.io/packages/helm/mariadb-operator/mariadb-operator).

          This helm operator provides provides a 1:1 mapping between the official helm chart and the [`MariadbOperator`](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/config/samples/helm_v1alpha1_mariadboperator.yaml) CRD, allowing to install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) via OLM without having to do any change in the helm chart.

          Normally, you would install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) providing this `values.yaml` to the helm chart:
          ```yaml
          image:
            repository: ghcr.io/mariadb-operator/mariadb-operator
            pullPolicy: IfNotPresent
          logLevel: INFO
          ha:
            enabled: true
          metrics:
            enabled: true
            serviceMonitor:
              enabled: true
          webhook:
            cert:
              certManager:
                enabled: true
          ```

          This helm chart installation is abstracted in the [`MariadbOperator`](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/config/samples/helm_v1alpha1_mariadboperator.yaml) CRD, which will be reconciled by the helm operator:
          ```yaml
          apiVersion: helm.k8s.mariadb.com/v1alpha1
          kind: MariadbOperator
          metadata:
            name: mariadb-operator
          spec:
            image:
              repository: ghcr.io/mariadb-operator/mariadb-operator
              pullPolicy: IfNotPresent
            logLevel: INFO
            ha:
              enabled: true
            metrics:
              enabled: true
              serviceMonitor:
                enabled: true
            webhook:
              cert:
                certManager:
                  enabled: true
          ```

          Once you have installed the operator, you will able to install a [`MariaDB`](https://github.com/mariadb-operator/mariadb-operator/blob/main/examples/manifests/mariadb_v1alpha1_mariadb.yaml) instance. Refer to the main [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) documentation for getting started with the rest of CRDs.

          ## Documentation
          * [mariadb-operator](https://github.com/mariadb-operator/mariadb-operator/blob/main/README.md)
          * [mariadb-operator-helm](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/README.md)

          ## Releases
          This operator is automatically published in the following repositories whenever a new version of the [helm chart](https://artifacthub.io/packages/helm/mariadb-operator/mariadb-operator) is released:
          - [k8s-operatorhub/community-operators](https://github.com/k8s-operatorhub/community-operators)
          - [redhat-openshift-ecosystem/community-operators-prod](https://github.com/redhat-openshift-ecosystem/community-operators-prod)

          ## Roadmap
          Take a look at our [roadmap](https://github.com/mariadb-operator/mariadb-operator/blob/main/ROADMAP.md) and feel free to open an issue to suggest new features.

          ## Contributing
          We welcome and encourage contributions to this project! Please check our [contributing](https://github.com/mariadb-operator/mariadb-operator/blob/main/CONTRIBUTING.md) and [development](https://github.com/mariadb-operator/mariadb-operator/blob/main/docs/DEVELOPMENT.md) guides. PRs welcome!

          ## Get in touch
          Join us on Slack: **[MariaDB Community Slack](https://r.mariadb.com/join-community-slack)**.
        displayName: MariaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mariadb
        - mysql
        - operator
        - mariadb-operator
        - database
        - maxscale
        links:
        - name: GitHub
          url: https://github.com/mariadb-operator/mariadb-operator
        maintainers:
        - email: mariadb-operator@proton.me
          name: mmontes11
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: mariadb-operator
          url: https://github.com/mariadb-operator/mariadb-operator
        relatedImages:
        - ghcr.io/mariadb-operator/mariadb-operator-helm:25.8.4
        version: 25.8.4
      entries:
      - name: mariadb-operator.v25.8.4
        version: 25.8.4
      - name: mariadb-operator.v25.8.3
        version: 25.8.3
      - name: mariadb-operator.v25.8.2
        version: 25.8.2
      - name: mariadb-operator.v25.8.1
        version: 25.8.1
      - name: mariadb-operator.v25.8.0
        version: 25.8.0
      - name: mariadb-operator.v0.38.1
        version: 0.38.1
      - name: mariadb-operator.v0.38.0
        version: 0.38.0
      - name: mariadb-operator.v0.37.1
        version: 0.37.1
      - name: mariadb-operator.v0.37.0
        version: 0.37.0
      - name: mariadb-operator.v0.36.0
        version: 0.36.0
      - name: mariadb-operator.v0.35.1
        version: 0.35.1
      - name: mariadb-operator.v0.35.0
        version: 0.35.0
      - name: mariadb-operator.v0.34.0
        version: 0.34.0
      - name: mariadb-operator.v0.33.0
        version: 0.33.0
      - name: mariadb-operator.v0.32.0
        version: 0.32.0
      - name: mariadb-operator.v0.31.0
        version: 0.31.0
      - name: mariadb-operator.v0.30.0
        version: 0.30.0
      - name: mariadb-operator.v0.29.0
        version: 0.29.0
      - name: mariadb-operator.v0.28.1
        version: 0.28.1
      - name: mariadb-operator.v0.27.0
        version: 0.27.0
      - name: mariadb-operator.v0.26.0
        version: 0.26.0
      - name: mariadb-operator.v0.24.0
        version: 0.24.0
      - name: mariadb-operator.v0.23.1
        version: 0.23.1
      - name: mariadb-operator.v0.23.0
        version: 0.23.0
      - name: mariadb-operator.v0.22.0
        version: 0.22.0
      - name: mariadb-operator.v0.21.0
        version: 0.21.0
      - name: mariadb-operator.v0.20.0
        version: 0.20.0
      - name: mariadb-operator.v0.19.0
        version: 0.19.0
      - name: mariadb-operator.v0.18.0
        version: 0.18.0
      - name: mariadb-operator.v0.17.0
        version: 0.17.0
      - name: mariadb-operator.v0.16.0
        version: 0.16.0
      - name: mariadb-operator.v0.15.1
        version: 0.15.1
      - name: mariadb-operator.v0.15.0
        version: 0.15.0
      - name: mariadb-operator.v0.14.0
        version: 0.14.0
      - name: mariadb-operator.v0.13.0
        version: 0.13.0
      - name: mariadb-operator.v0.11.0
        version: 0.11.0
      - name: mariadb-operator.v0.10.0
        version: 0.10.0
      - name: mariadb-operator.v0.9.1
        version: 0.9.1
      - name: mariadb-operator.v0.9.0
        version: 0.9.0
      - name: mariadb-operator.v0.8.0
        version: 0.8.0
      - name: mariadb-operator.v0.7.1
        version: 0.7.1
      - name: mariadb-operator.v0.7.0
        version: 0.7.0
      - name: mariadb-operator.v0.6.7
        version: 0.6.7
      - name: mariadb-operator.v0.6.6
        version: 0.6.6
      - name: mariadb-operator.v0.6.5
        version: 0.6.5
      - name: mariadb-operator.v0.6.4
        version: 0.6.4
      name: alpha
    defaultChannel: alpha
    packageName: mariadb-operator
    provider:
      name: mariadb-operator
      url: https://github.com/mariadb-operator/mariadb-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Anchore Inc.
      provider-url: ""
    name: anchore-engine
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anchore-engine.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.anchore.io/v1alpha1",
                "kind": "AnchoreEngine",
                "metadata": {
                  "name": "anchoreengine-sample"
                },
                "spec": {
                  "anchoreGlobal": {
                    "defaultAdminEmail": "example@email.com"
                  },
                  "postgresql": {
                    "extraEnv": [
                      {
                        "name": "POSTGRESQL_USER",
                        "value": "anchoreengine"
                      },
                      {
                        "name": "POSTGRESQL_PASSWORD",
                        "value": "anchore-postgres,123"
                      },
                      {
                        "name": "POSTGRESQL_DATABASE",
                        "value": "anchore"
                      },
                      {
                        "name": "PGUSER",
                        "value": "postgres"
                      }
                    ],
                    "postgresDatabase": "anchore",
                    "postgresPassword": "anchore-postgres,123",
                    "postgresUser": "anchoreengine"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/anchore/engine-operator@sha256:baf0f5acd319bc4d0132c88922ee3f357fdcdc278d931f243d0cbec3f6d416c6
          createdAt: "2021-10-06T23:53:51Z"
          description: Anchore Engine - container image scanning service for policy-based
            security, best-practice and compliance enforcement.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/anchore/engine-operator
          support: anchore.slack.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys Anchore Engine
            displayName: Anchore Engine Operator
            kind: AnchoreEngine
            name: anchoreengines.charts.anchore.io
            version: v1alpha1
        description: "Anchore Engine is an open source software system that provides
          a centralized service for analyzing container images, \n scanning for security
          vulnerabilities, and enforcing deployment policies. Anchore Engine is provided
          as a Docker \n container image that can be run standalone or with an orchestration
          platform such as Kubernetes. Anchore Engine allows \n users to perform detailed
          analysis on their container images, run queries, produce reports, and define
          policies that \n can be used in CI/CD pipelines. Users can extend Anchore
          Engine with plugins that add new queries, image analysis, and \n policies.
          Anchore Engine can be accessed directly through a RESTful API or via the
          Anchore CLI.\n\n The Anchore Engine Operator is based on the official stable
          [Helm Chart](https://github.com/anchore/anchore-charts/tree/master/stable/anchore-engine)."
        displayName: Anchore Engine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - analysis
        - docker
        - anchore
        - anchore-engine
        - image
        - security
        - vulnerability
        - scanner
        links:
        - name: Anchore Inc.
          url: https://www.anchore.com
        - name: Anchore Engine
          url: https://github.com/anchore/anchore-engine
        maintainers:
        - email: bradyt@anchore.com
          name: Brady Todhunter
        - email: zach@anchore.com
          name: Zach Hill
        maturity: alpha
        provider:
          name: Anchore Inc.
        relatedImages:
        - registry.connect.redhat.com/anchore/engine-operator@sha256:baf0f5acd319bc4d0132c88922ee3f357fdcdc278d931f243d0cbec3f6d416c6
        - registry.connect.redhat.com/anchore/engine0@sha256:68e09d5ff695b46f24de3d7610f88dfc14b77b38758bbd68d03350dbae3deb0f
        - registry.redhat.io/rhel8/postgresql-96@sha256:e8177c5de05ccdd6d12e360b65c63c889b2de725d573ffdedd8914230b118639
        version: 1.0.1
      entries:
      - name: anchore-engine.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: anchore-engine
    provider:
      name: Anchore Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: https://github.com/Mellanox/maintenance-operator/
    name: nvidia-maintenance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nvidia-maintenance-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "MaintenanceOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "maintenanceoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "maintenanceoperatorconfig",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "maintenanceoperatorconfig-sample"
                },
                "spec": {
                  "logLevel": "info",
                  "maxNodeMaintenanceTimeSeconds": 1600,
                  "maxParallelOperations": 1,
                  "maxUnavailable": "30%"
                }
              },
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "nodemaintenance-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodemaintenance",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "cordon": true,
                  "drainSpec": {
                    "deleteEmptyDir": true,
                    "force": true,
                    "podEvictionFilters": [
                      {
                        "byResourceNameRegex": "nvidia.com/gpu-*"
                      },
                      {
                        "byResourceNameRegex": "nvidia.com/rdma*"
                      }
                    ],
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  },
                  "nodeName": "worker-node-01",
                  "requestorID": "some-cluster-component.vendor.com",
                  "waitForPodCompletion": {
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
          createdAt: "2025-07-13T12:33:06Z"
          description: Node maintenance in K8s cluster in a coordinated manner
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-maintenance-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: NVIDIA
          repository: https://github.com/Mellanox/maintenance-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MaintenanceOperatorConfig is the Schema for the maintenanceoperatorconfigs
              API
            displayName: Maintenance Operator Config
            kind: MaintenanceOperatorConfig
            name: maintenanceoperatorconfigs.maintenance.nvidia.com
            version: v1alpha1
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.maintenance.nvidia.com
            version: v1alpha1
        description: Coordinates And Performs Common Node Preparations For Maintenance
        displayName: NVIDIA Maintenance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node
        - maintenance
        - drain
        links:
        - name: NVIDIA Maintenance Operator
          url: https://github.com/Mellanox/maintenance-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/maintenance-operator/
        relatedImages:
        - ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
        version: 0.2.3
      entries:
      - name: nvidia-maintenance-operator.v0.2.3
        version: 0.2.3
      name: stable
    - currentCSV: nvidia-maintenance-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "MaintenanceOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "maintenanceoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "maintenanceoperatorconfig",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "maintenanceoperatorconfig-sample"
                },
                "spec": {
                  "logLevel": "info",
                  "maxNodeMaintenanceTimeSeconds": 1600,
                  "maxParallelOperations": 1,
                  "maxUnavailable": "30%"
                }
              },
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "nodemaintenance-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodemaintenance",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "cordon": true,
                  "drainSpec": {
                    "deleteEmptyDir": true,
                    "force": true,
                    "podEvictionFilters": [
                      {
                        "byResourceNameRegex": "nvidia.com/gpu-*"
                      },
                      {
                        "byResourceNameRegex": "nvidia.com/rdma*"
                      }
                    ],
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  },
                  "nodeName": "worker-node-01",
                  "requestorID": "some-cluster-component.vendor.com",
                  "waitForPodCompletion": {
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
          createdAt: "2025-07-13T12:33:06Z"
          description: Node maintenance in K8s cluster in a coordinated manner
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-maintenance-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: NVIDIA
          repository: https://github.com/Mellanox/maintenance-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MaintenanceOperatorConfig is the Schema for the maintenanceoperatorconfigs
              API
            displayName: Maintenance Operator Config
            kind: MaintenanceOperatorConfig
            name: maintenanceoperatorconfigs.maintenance.nvidia.com
            version: v1alpha1
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.maintenance.nvidia.com
            version: v1alpha1
        description: Coordinates And Performs Common Node Preparations For Maintenance
        displayName: NVIDIA Maintenance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node
        - maintenance
        - drain
        links:
        - name: NVIDIA Maintenance Operator
          url: https://github.com/Mellanox/maintenance-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/maintenance-operator/
        relatedImages:
        - ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
        version: 0.2.3
      entries:
      - name: nvidia-maintenance-operator.v0.2.3
        version: 0.2.3
      name: v0.2
    defaultChannel: v0.2
    packageName: nvidia-maintenance-operator
    provider:
      name: NVIDIA
      url: https://github.com/Mellanox/maintenance-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: mariadb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mariadb.openstack.org/v1beta1",
                "kind": "Galera",
                "metadata": {
                  "name": "openstack"
                },
                "spec": {
                  "replicas": 3,
                  "secret": "osp-secret",
                  "storageClass": "local-storage",
                  "storageRequest": "500M"
                }
              },
              {
                "apiVersion": "mariadb.openstack.org/v1beta1",
                "kind": "MariaDBAccount",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "mariadb-operator",
                    "app.kubernetes.io/instance": "mariadbaccount-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mariadbaccount",
                    "app.kubernetes.io/part-of": "mariadb-operator",
                    "mariaDBDatabaseName": "neutron"
                  },
                  "name": "neutron1"
                },
                "spec": {
                  "secret": "neutrondb-secret",
                  "userName": "neutron1"
                }
              },
              {
                "apiVersion": "mariadb.openstack.org/v1beta1",
                "kind": "MariaDBDatabase",
                "metadata": {
                  "labels": {
                    "dbName": "mariadb"
                  },
                  "name": "scott"
                },
                "spec": {
                  "name": "scott",
                  "secret": "tiger"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:39Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Galera is the Schema for the galeras API
            displayName: Galera
            kind: Galera
            name: galeras.mariadb.openstack.org
            version: v1beta1
          - description: MariaDBAccount is the Schema for the mariadbaccounts API
            displayName: Maria DBAccount
            kind: MariaDBAccount
            name: mariadbaccounts.mariadb.openstack.org
            version: v1beta1
          - description: MariaDBDatabase is the Schema for the mariadbdatabases API
            displayName: Maria DBDatabase
            kind: MariaDBDatabase
            name: mariadbdatabases.mariadb.openstack.org
            version: v1beta1
        description: MariaDB Operator
        displayName: MariaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Identity
        - MariaDB
        links:
        - name: MariaDB Operator
          url: https://github.com/openstack-k8s-operators/mariadb-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/mariadb-rhel9-operator@sha256:5ae64a3cf9bf5eaf244548c3bf2e8ef89c9436cb676721c1f1d1c2ff8285e7e5
        - registry.redhat.io/rhoso/openstack-mariadb-rhel9@sha256:13119fe0ec56263a2bff3fc9c3892ea6386b837955280d8488f05d8ea6c4e44d
        version: 1.0.6
      entries:
      - name: mariadb-operator.v1.0.6
        version: 1.0.6
      - name: mariadb-operator.v1.0.5
        version: 1.0.5
      - name: mariadb-operator.v1.0.4
        version: 1.0.4
      - name: mariadb-operator.v1.0.3
        version: 1.0.3
      - name: mariadb-operator.v1.0.2
        version: 1.0.2
      - name: mariadb-operator.v1.0.1
        version: 1.0.1
      - name: mariadb-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: mariadb-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: GitLab, Inc.
      provider-url: https://gitlab.com
    name: gitlab-runner-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-runner-operator.v1.41.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta2",
                "kind": "Runner",
                "metadata": {
                  "name": "gitlab-runner-example",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "gitlabUrl": "https://gitlab.example.com",
                  "token": "runner-token-secret",
                  "tags": "kubernetes,example",
                  "concurrent": 10,
                  "imagePullPolicy": "IfNotPresent"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery, Developer Tools
          containerImage: registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
          createdAt: "2025-09-25T15:08:40Z"
          description: GitLab Runner operator manages lifecycle of GitLab Runner instances
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.11.0 <1.41.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: none
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
          support: GitLab, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Runner is the open source project used to run your jobs and
              send the results back to GitLab
            displayName: GitLab Runner
            kind: Runner
            name: runners.apps.gitlab.com
            version: v1beta2
        description: |-
          GitLab Runner is the lightweight, highly-scalable agent that runs your build jobs and sends the results back to a GitLab instance. GitLab Runner works in conjunction with GitLab CI/CD, the open-source continuous integration service included with GitLab.

              The GitLab Runner operator manages the lifecycle of GitLab Runner in Kubernetes or Openshift clusters. The operator aims to automate the tasks needed to run your CI/CD jobs in your container orchestration platform.

              ## Prerequisites

                For Kubernetes cluster, install cert-manager:

                ```shell
                kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.7.1/cert-manager.yaml
                ```

              ## GitLab Runner version

              This version of **GitLab Runner Operator** ships with **GitLab Runner v18.4.0**.

              To use a different version of **GitLab Runner** change the [`runnerImage` and `helperImage` properties](https://docs.gitlab.com/runner/configuration/configuring_runner_operator.html#operator-properties).

              ## Usage

               To link a GitLab Runner instance to a self-hosted GitLab instance or the hosted [GitLab](https://gitlab.com), you first need to:

               1. Create a secret containing the `runner-token` from your GitLab project.

                ```
                cat > gitlab-runner-secret.yml << EOF
                apiVersion: v1
                kind: Secret
                metadata:
                  name: gitlab-runner-secret
                type: Opaque
                # Only one of the following fields must be set. The Operator will fail to register Runner if both are provided.
                # # NOTE: runner-registration-token is deprecated and will be removed in GitLab 18.0. You should use runner-token instead.
                stringData:
                  runner-token: REPLACE_ME # your project runner token
                  # runner-registration-token: "" # your project runner secret
                EOF
                ```

                ```
                oc apply -f gitlab-runner-secret.yml
                ```

               2. Create the Custom Resource Definition (CRD) file and include the following information. The tags value must be openshift for the job to run.

                 ```
                 cat > gitlab-runner.yml << EOF
                 apiVersion: apps.gitlab.com/v1beta2
                 kind: Runner
                 metadata:
                   name: gitlab-runner
                 spec:
                   gitlabUrl: https://gitlab.example.com
                   buildImage: alpine
                   token: gitlab-runner-secret
                   tags: openshift
                 EOF
                 ```

                ```
                oc apply -f gitlab-runner.yml
                ```

              ## Full documentation

              Visit [Install GitLab Runner Operator](https://docs.gitlab.com/runner/install/operator.html)
        displayName: GitLab Runner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GitLab
        - CI/CD
        - DevOps
        - SAST
        - DAST
        links:
        - name: Gitlab Runner Operator
          url: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
        - name: GitLab Docs
          url: https://docs.gitlab.com
        - name: GitLab Runner
          url: https://docs.gitlab.com/runner/
        - name: GitLab CI/CD
          url: https://docs.gitlab.com/ee/ci/quick_start/
        maintainers:
        - email: support@gitlab.com
          name: Georgi N. Georgiev
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: GitLab, Inc.
          url: https://gitlab.com
        relatedImages:
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-ocp@sha256:0f2a5a3b6779d719eecbf079b99cf06b652603d74b563f57d753e32966d253dd
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-helper-ocp@sha256:29727fae83a226bcf57607240d4046a0b8c208b5ff42ee1f8ca0f115bb188809
        version: 1.41.1
      entries:
      - name: gitlab-runner-operator.v1.41.1
        version: 1.41.1
      - name: gitlab-runner-operator.v1.41.0
        version: 1.41.0
      - name: gitlab-runner-operator.v1.40.1
        version: 1.40.1
      - name: gitlab-runner-operator.v1.40.0
        version: 1.40.0
      - name: gitlab-runner-operator.v1.39.0
        version: 1.39.0
      name: stable
    defaultChannel: stable
    packageName: gitlab-runner-operator
    provider:
      name: GitLab, Inc.
      url: https://gitlab.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "docker.io/percona/percona-postgresql-operator:2.7.0-ppg17.5.2-postgres",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "docker.io/percona/percona-pgbouncer:1.24.1",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "docker.io/percona/percona-pgbackrest:2.55.0",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:3.2.0",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: docker.io/percona/percona-postgresql-operator:2.7.0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          olm.skipRange: <v2.7.0
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native

          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - docker.io/percona/percona-postgresql-operator:2.7.0
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator.v2.6.1
        version: 2.6.1
      - name: percona-postgresql-operator.v2.6.0
        version: 2.6.0
      - name: percona-postgresql-operator.v2.5.0
        version: 2.5.0
      name: stable
    defaultChannel: stable
    packageName: percona-postgresql-operator
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: VMware Tanzu
      provider-url: ""
    name: rabbitmq-messaging-topology-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-messaging-topology-operator.v1.17.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [{"apiVersion":"rabbitmq.com/v1beta1","kind":"Binding","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Exchange","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Federation","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Permission","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Policy","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"OperatorPolicy","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Queue","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"SchemaReplication","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Shovel","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1alpha1","kind":"SuperStream","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"User","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Vhost","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"TopicPermission","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}}]
          capabilities: Full Lifecycle
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/rabbitmqoperator/messaging-topology-operator:1.17.4
          createdAt: 2025-08-21T11:15:42
          description: Kubernetes operator to allow developers to create and manage
            RabbitMQ messaging topologies within a RabbitMQ cluster using a declarative
            Kubernetes API.
          repository: https://github.com/rabbitmq/messaging-topology-operator
          support: VMware Tanzu
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Binding
            displayName: Binding
            kind: Binding
            name: bindings.rabbitmq.com
            version: v1beta1
          - description: Exchange
            displayName: Exchange
            kind: Exchange
            name: exchanges.rabbitmq.com
            version: v1beta1
          - description: Federation
            displayName: Federation
            kind: Federation
            name: federations.rabbitmq.com
            version: v1beta1
          - description: Permission
            displayName: Permission
            kind: Permission
            name: permissions.rabbitmq.com
            version: v1beta1
          - description: TopicPermission
            displayName: TopicPermission
            kind: TopicPermission
            name: topicpermissions.rabbitmq.com
            version: v1beta1
          - description: Policy
            displayName: Policy
            kind: Policy
            name: policies.rabbitmq.com
            version: v1beta1
          - description: OperatorPolicy
            displayName: OperatorPolicy
            kind: OperatorPolicy
            name: operatorpolicies.rabbitmq.com
            version: v1beta1
          - description: Queue
            displayName: Queue
            kind: Queue
            name: queues.rabbitmq.com
            version: v1beta1
          - description: Schema Replication
            displayName: SchemaReplication
            kind: SchemaReplication
            name: schemareplications.rabbitmq.com
            version: v1beta1
          - description: Shovel
            displayName: Shovel
            kind: Shovel
            name: shovels.rabbitmq.com
            version: v1beta1
          - description: Super Stream
            displayName: SuperStream
            kind: SuperStream
            name: superstreams.rabbitmq.com
            version: v1alpha1
          - description: User
            displayName: User
            kind: User
            name: users.rabbitmq.com
            version: v1beta1
          - description: Vhost
            displayName: Vhost
            kind: Vhost
            name: vhosts.rabbitmq.com
            version: v1beta1
          required:
          - description: Rabbitmq Cluster
            displayName: RabbitmqCluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: |
          ## About the managed application
          RabbitMQ is an open-source message-broker software that originally implemented the Advanced Message Queuing Protocol and has since been extended with a plug-in architecture to support Streaming Text Oriented Messaging Protocol, MQ Telemetry Transport, and other protocols.
          ## About this Operator
          Kubernetes operator to allow developers to create and manage RabbitMQ messaging topologies within a RabbitMQ cluster using a declarative Kubernetes API. A Messaging topology is the collection of objects such as exchanges, queues, bindings and policies that provides specific messaging or streaming scenario. This operator is used with RabbitMQ clusters deployed via the RabbitMQ Cluster Kubernetes Operator. This repository contains custom controllers and custom resource definitions (CRDs) enabling a declarative API for RabbitMQ messaging topologies.
          ## Prerequisites for enabling this Operator
          The operator works on top of the RabbitMQ cluster operator.
          Messaging Topology Operator is tested with the latest release of RabbitMQ Cluster Operator. It uses the generated default user secret from RabbitmqCluster (set in rabbitmqcluster.status.binding) to authenticate with RabbitMQ server.
        displayName: rabbitmq-messaging-topology-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        - messaging
        - amqp
        - amqp10
        - stomp
        - mqtt
        - mqtt5
        - queues
        - streaming
        links:
        - name: GitHub
          url: https://github.com/rabbitmq/messaging-topology-operator
        - name: Documentation
          url: https://www.rabbitmq.com/kubernetes/operator/install-topology-operator.html
        maintainers:
        - email: rabbitmq-users@googlegroups.com
          name: VMware Tanzu
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        provider:
          name: VMware Tanzu
        relatedImages:
        - quay.io/rabbitmqoperator/messaging-topology-operator:1.17.4
        version: 1.17.4
      entries:
      - name: rabbitmq-messaging-topology-operator.v1.17.4
        version: 1.17.4
      - name: rabbitmq-messaging-topology-operator.v1.16.0
        version: 1.16.0
      - name: rabbitmq-messaging-topology-operator.v1.15.0
        version: 1.15.0
      - name: rabbitmq-messaging-topology-operator.v1.14.2
        version: 1.14.2
      - name: rabbitmq-messaging-topology-operator.v1.14.1
        version: 1.14.1
      - name: rabbitmq-messaging-topology-operator.v1.13.0
        version: 1.13.0
      - name: rabbitmq-messaging-topology-operator.v1.12.2
        version: 1.12.2
      - name: rabbitmq-messaging-topology-operator.v1.12.1
        version: 1.12.1
      name: stable
    defaultChannel: stable
    packageName: rabbitmq-messaging-topology-operator
    provider:
      name: VMware Tanzu
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Omar Gaye IBM
      provider-url: omarg.net
    name: memcached-operator-ogaye
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: memcached-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.omarg.net/v1alpha1",
                "kind": "Memcached",
                "metadata": {
                  "name": "memcached-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.6.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Memcached is the Schema for the memcacheds API
            displayName: Memcached
            kind: Memcached
            name: memcacheds.cache.omarg.net
            version: v1alpha1
        description: Omar memcached operator
        displayName: memcached-operator-ogaye
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Omar
        - Memcached
        links:
        - name: Memcached Operator OGaye
          url: https://memcached-operator.domain
        maturity: alpha
        provider:
          name: Omar Gaye IBM
          url: omarg.net
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9211b70edbb93d718428e1c9a992e5fc74a20cb213f8a0592ffe340c362890fe
        - quay.io/ogaye/memcached-operator@sha256:5340ac11ce676dd3204bf3807a4c2333a8dddcbcec918db0aeafbc977163c1c6
        version: 0.0.1
      entries:
      - name: memcached-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: memcached-operator-ogaye
    provider:
      name: Omar Gaye IBM
      url: omarg.net
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Aqua Security, Inc.
      provider-url: ""
    name: aqua-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aqua-operator.2022.4.836
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaCsp",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "namespace": "aqua",
                    "version": "2022.4",
                    "requirements": true
                  },
                  "common": {
                    "dbDiskSize": 10,
                    "databaseSecret": {
                      "key": "db-password",
                      "name": "aqua-database-password"
                    }
                  },
                  "database": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "gateway": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "server": {
                    "replicas": 1,
                    "service": "LoadBalancer"
                  },
                  "route": true,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaDatabase",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "splitDB": false
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "diskSize": 10,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "runAsNonRoot": false,
                  "gateway": {
                    "host": "aqua-gateway",
                    "port": 8443
                  },
                  "token": "<<your-token>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaKubeEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "config": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "aqua-secure",
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "service": "ClusterIP"
                  },
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaGateway",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  }
                },
                "runAsNonRoot": false
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaScanner",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "deploy": {
                    "replicas": 1
                  },
                  "runAsNonRoot": false,
                  "login": {
                    "username": "<<YOUR AQUA USER NAME>>",
                    "password": "<<YOUR AQUA USER PASSWORD>>",
                    "host": "http://aqua-server:8080"
                  }
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaServer",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "LoadBalancer"
                  }
                },
                "runAsNonRoot": false
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaLightning",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "global": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "Default-cluster-name"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "kubeEnforcer": {
                    "infra": {
                      "serviceAccount": "aqua-kube-enforcer-sa"
                    },
                    "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                    "deploy": {
                      "service": "ClusterIP"
                    },
                    "starboard": {
                      "infra": {
                        "serviceAccount": "aqua-kube-enforcer-sa"
                      },
                      "config": {
                        "imagePullSecret": "aqua-registry"
                      },
                      "deploy": {
                        "replicas": 1
                      }
                    },
                    "env": []
                  },
                  "enforcer": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "deploy": null,
                    "token": "<<ENFORCER_GROUP_TOKEN>>",
                    "env": [
                    ],
                    "secret": {
                      "name": null,
                      "key": null
                    },
                    "runAsNonRoot": null,
                    "rhcosVersion": null
                  }
                }
              },
            {
                  "apiVersion": "operator.aquasec.com/v1alpha1",
                  "kind": "AquaCloudConnector",
                  "metadata": {
                    "name": "aqua"
                  },
                  "spec": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "common": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    },
                    "runAsNonRoot": false,
                    "gateway": {
                      "host": "aqua-gateway",
                      "port": "8443"
                    },
                    "login": {
                      "host": "aqua-gateway:8443",
                      "password": "<<YOUR AQUA USER PASSWORD>>",
                      "token": "<<YOUR AQUA CLOUD CONNECTOR TOKEN>>",
                      "username": "<<YOUR AQUA USER NAME>>"
                    },
                    "tunnels": [
                      {
                        "host": "<account>.azurecr.io",
                        "port": "443"
                      },
                      {
                        "host": "gke.gcr.io",
                        "port": "443"
                      },
                      {
                        "host": "<account>.jfrog.io",
                        "port": "443"
                      },
                      {
                        "host": "1.1.1.1",
                        "port": "443"
                      },
                      {
                        "region": "us-east1",
                        "type": "ecr"
                      }
                    ]
                  }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/aquasec/aquasec@sha256:c9f2c770a51e9664006c12d124ca18aab76bea8c32a6a253fe31d44c02633642
          createdAt: ""
          description: The Aqua Security Operator runs within a Openshift cluster
            and provides a means to deploy and manage Aqua Security cluster and components.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2022.4.836'
          operators.operatorframework.io/internal-objects: '["aquastarboards.aquasecurity.github.io","configauditreports.aquasecurity.github.io",
            "clusterconfigauditreports.aquasecurity.github.io"]'
          repository: https://github.com/aquasecurity/aqua-operator
          support: Aqua Security, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aqua Security CSP Deployment with Aqua Operator
            displayName: AquaCsp
            kind: AquaCsp
            name: aquacsps.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Lightning Enforcers Deployment with Aqua Operator
            displayName: AquaLightningEnforcers
            kind: AquaLightning
            name: aqualightnings.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Database Deployment with Aqua Operator
            displayName: AquaDatabase
            kind: AquaDatabase
            name: aquadatabases.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Enforcer Deployment with Aqua Operator
            displayName: AquaEnforcer
            kind: AquaEnforcer
            name: aquaenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security KubeEnforcer Deployment with Aqua Operator
            displayName: AquaKubeEnforcer
            kind: AquaKubeEnforcer
            name: aquakubeenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Gateway Deployment with Aqua Operator
            displayName: AquaGateway
            kind: AquaGateway
            name: aquagateways.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Scanner Deployment with Aqua Operator
            displayName: AquaScanner
            kind: AquaScanner
            name: aquascanners.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Server Deployment with Aqua Operator
            displayName: AquaServer
            kind: AquaServer
            name: aquaservers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Starboard for kube-enforcer.
            displayName: AquaStarboard
            kind: AquaStarboard
            name: aquastarboards.aquasecurity.github.io
            version: v1alpha1
          - description: Aqua Cloud Connector Deployment with Aqua Operator
            displayName: AquaCloudConnector
            kind: AquaCloudConnector
            name: aquacloudconnectors.operator.aquasec.com
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ClusterConfigAuditReport
            kind: ClusterConfigAuditReport
            name: clusterconfigauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ConfigAuditReport
            kind: ConfigAuditReport
            name: configauditreports.aquasecurity.github.io
            version: v1alpha1
        description: |-
          The Aqua Security Operator runs within an OpenShift cluster and provides a means to deploy and manage the Aqua Security cluster and components :
          * Server (aka “console”)
          * Database (for production environments we recommend to use an external database and not the Aqua default database)
          * Gateway
          * Enforcer (aka “agent”)
          * KubeEnforcer
          * Scanner
          * CSP (package that contains the Server, Database, and Gateway)
          * LightningEnforcer (package that contains Enforcer and Kube Enforcer) - Do not deploy using Certified Operator.
          * AquaCloudConnector.

          Use the Aqua-Operator to
          * Deploy Aqua Security components on OpenShift
          * Scale up Aqua Security components with extra replicas
          * Assign metadata tags to Aqua Security components

          The Aqua Operator provides a few [Custom Resources](https://github.com/aquasecurity/aqua-operator/tree/master/config/crd) for managing the Aqua CSP platform.

          ## Prerequisites
          There are only 4 prerequisites:
          1. Verify that the OpenShift version is officially supported [Verify Here](https://docs.aquasec.com/saas/workload-protection/product-information/system-requirements/)
          2. Make sure you have a license and to obtain one, please contact Aqua Security at [Contact Us](mailto:cloudsales@aquasec.com).
          3. Create a new project for aqua

              ```oc new-project aqua```

          4. Create the secret for Aqua Database password. If you are using an External Database, make sure you use those credentials.
          NOTE: This step is optional and you can specify the Database password when creating the CRs.

              ```oc create secret generic aqua-database-password --from-literal=db-password=<password> -n aqua```

          Please note that for the certified operator, the Docker registry secret is NOT needed. For the Red Hat certified operator, the Aqua application images are stored in the Red Hat Connect registry and will be automatically pulled by the Aqua operator.

          ## Choosing the right channel

          Channels are a flexible publishing mechanism that offers you different Aqua versions to deploy, at any point. These channels map to their respective Aqua versions, describing the maturity based on where they are in their lifecycle:
          * 4.6: This maps to the older version of Aqua 4.6.0
          * 5.0: This maps to the Aqua version 5.0.0
          * 5.3: This maps to the Aqua version 5.3.0
          * 6.0: This maps to the Aqua version 6.0.0
          * 6.2: This maps to the Aqua version 6.2.0
          * 6.5: This maps to the Aqua version 6.5.0
          * 2022.4: This is the latest and greatest version of Aqua (default)

          ## Deploying the Aqua Operator

          Aqua Operator follows an easy push button deployment that installs the operator in its own aqua namespace.
          Please refer to the instructions [here](https://github.com/aquasecurity/aqua-operator/blob/2022.4.0/docs/DeployOpenShiftOperator.md).

          ## Configuring the Aqua Operator

          Operators leverage CRDs as a configuration mechanism for the application. Aqua offers a number of CRDs that can be tweaked and configured as per your needs. Please refer to this [link](https://github.com/aquasecurity/aqua-operator/blob/6.2.0/docs/DeployOpenShiftOperator.md#aquacsp-crds) for Custom Resource examples for popular scenarios.

          ## Support

          For support please contact support@aquasec.com.
        displayName: Aqua Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aqua-security
        - scanning
        - security
        - runtime-security
        links:
        - name: Aqua Security
          url: https://www.aquasec.com/
        - name: Aqua Operator Github
          url: https://github.com/aquasecurity/aqua-operator
        maintainers:
        - email: support@aquasec.com
          name: Aqua Support, Aqua Security
        maturity: alpha
        provider:
          name: Aqua Security, Inc.
        relatedImages:
        - registry.connect.redhat.com/aquasec/console@sha256:ed371053e46e89ad68ed48282ddf28abc668a613a9c32e3aa45e5111512b8466
        - registry.connect.redhat.com/aquasec/kube-bench@sha256:bbb4e8a61033b303525af6764f685e62a2f441453c3f0eaaf4cbab6997bf93ee
        - registry.connect.redhat.com/aquasec/gateway@sha256:11aa01e681986f6b5bd53b41a07db4456023080a753970d8c872e000f68f3b9b
        - registry.connect.redhat.com/aquasec/kube-enforcer@sha256:c456b65a3f1dc2de7209f3d9cbcf1bb523130021433b8fd740c41d9bd6f7db9f
        - registry.connect.redhat.com/aquasec/aquasec@sha256:c9f2c770a51e9664006c12d124ca18aab76bea8c32a6a253fe31d44c02633642
        - registry.connect.redhat.com/aquasec/enforcer@sha256:50128b750e0a094de266c9a2cb42ce06a757f441a68d206a8a9c17f7b7119066
        - registry.connect.redhat.com/aquasec/scanner@sha256:c086d735c7e6ea842ad1aaeb1eb19d3a951e6e88d397a20f2cd559c6bc042d26
        - registry.connect.redhat.com/aquasec/cloud-connector@sha256:f9da1936ac5e4b8e4537dca69f920c682fb9fac98d091680e01d2122f2b0f85f
        - registry.connect.redhat.com/aquasec/database@sha256:c91ae82b0179c3ec4e156835a9f23b66767ba15514ff33fb268b35d8f04d6fde
        - registry.connect.redhat.com/aquasec/starboard@sha256:61673f7ba1d983ff4bb4bd76036ac55e5b81696c9a3e3aa4ab1a49a27a7a2106
        version: 2022.4.836
      entries:
      - name: aqua-operator.2022.4.836
        version: 2022.4.836
      - name: aqua-operator.2022.4.829
        version: 2022.4.829
      - name: aqua-operator.2022.4.818
        version: 2022.4.818
      - name: aqua-operator.2022.4.805
        version: 2022.4.805
      - name: aqua-operator.2022.4.760
        version: 2022.4.760
      - name: aqua-operator.2022.4.759
        version: 2022.4.759
      - name: aqua-operator.2022.4.720
        version: 2022.4.720
      - name: aqua-operator.2022.4.683
        version: 2022.4.683
      - name: aqua-operator.2022.4.662
        version: 2022.4.662
      - name: aqua-operator.2022.4.588
        version: 2022.4.588
      - name: aqua-operator.2022.4.587
        version: 2022.4.587
      - name: aqua-operator.2022.4.519
        version: 2022.4.519
      - name: aqua-operator.2022.4.518
        version: 2022.4.518
      name: "2022.4"
    defaultChannel: "2022.4"
    packageName: aqua-operator-certified
    provider:
      name: Aqua Security, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: netobserv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: network-observability-operator.v1.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowMetric",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netobserv-operator",
                    "app.kubernetes.io/instance": "flowmetric-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "flowmetric",
                    "app.kubernetes.io/part-of": "netobserv-operator"
                  },
                  "name": "flowmetric-sample"
                },
                "spec": {
                  "charts": [
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "",
                          "promQL": "sum(rate($METRIC[2m]))"
                        }
                      ],
                      "title": "External ingress traffic",
                      "type": "SingleStat",
                      "unit": "Bps"
                    },
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "{{DstK8S_Namespace}} / {{DstK8S_OwnerName}}",
                          "promQL": "sum(rate($METRIC{DstK8S_Namespace!=\"\"}[2m])) by (DstK8S_Namespace, DstK8S_OwnerName)"
                        }
                      ],
                      "sectionName": "External",
                      "title": "Top external ingress traffic per workload",
                      "type": "StackArea",
                      "unit": "Bps"
                    }
                  ],
                  "direction": "Ingress",
                  "filters": [
                    {
                      "field": "SrcSubnetLabel",
                      "matchType": "Absence"
                    }
                  ],
                  "labels": [
                    "DstK8S_HostName",
                    "DstK8S_Namespace",
                    "DstK8S_OwnerName",
                    "DstK8S_OwnerType"
                  ],
                  "metricName": "cluster_external_ingress_bytes_total",
                  "type": "Counter",
                  "valueField": "Bytes"
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "app"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "infra"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "enable": true,
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "includeList": [
                        "node_ingress_bytes_total",
                        "workload_ingress_bytes_total",
                        "namespace_flows_total"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta2",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "metrics": {
                        "server": {
                          "port": 9400
                        }
                      },
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "eBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "Disabled"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "\"app\""
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "\"infra\""
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "\"Pod\"",
                          "src_kind": "\"Pod\""
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "\"Service\""
                        },
                        "name": "Services network"
                      }
                    ],
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    }
                  },
                  "deploymentModel": "Direct",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "enable": true,
                    "lokiStack": {
                      "name": "loki"
                    },
                    "mode": "Monolithic",
                    "monolithic": {
                      "tenantID": "netobserv",
                      "tls": {
                        "caCert": {
                          "certFile": "service-ca.crt",
                          "name": "loki-gateway-ca-bundle",
                          "type": "configmap"
                        },
                        "enable": false
                      },
                      "url": "http://loki.netobserv.svc:3100/"
                    },
                    "readTimeout": "30s",
                    "writeBatchSize": 10485760,
                    "writeBatchWait": "1s",
                    "writeTimeout": "10s"
                  },
                  "namespace": "netobserv",
                  "networkPolicy": {
                    "additionalNamespaces": [],
                    "enable": false
                  },
                  "processor": {
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "Flows",
                    "metrics": {
                      "disableAlerts": [],
                      "server": {
                        "port": 9401
                      }
                    },
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "prometheus": {
                    "querier": {
                      "enable": true,
                      "mode": "Auto",
                      "timeout": "30s"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking, Observability, OpenShift Optional
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:13b6dd0ddf4bdd2836b0d3ec9c79119a902456dcee100cfac6898768e1f22aab
          createdAt: 2025-09-08T21:48:09
          description: Network flows collector and monitoring solution
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: '{"apiVersion":"flows.netobserv.io/v1beta2",
            "kind":"FlowCollector","metadata":{"name":"cluster"},"spec": {}}'
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
          support: NetObserv team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta2
          - description: '`FlowMetric` is the schema for the custom metrics API, which
              allows to generate more metrics out of flow logs. You can find examples
              here: https://github.com/netobserv/network-observability-operator/tree/main/config/samples/flowmetrics'
            displayName: Flow Metric
            kind: FlowMetric
            name: flowmetrics.flows.netobserv.io
            version: v1alpha1
        description: |
          Network Observability is an OpenShift operator that deploys a monitoring pipeline consisting in:
          - an eBPF agent, that generates network flows from captured packets
          - flowlogs-pipeline, a component that collects, enriches and exports these flows
          - a Console plugin for flows visualization with powerful filtering options, a topology representation and more

          Flow data is then available in multiple ways, each optional:

          - As Cluster Monitoring metrics
          - As raw flow logs stored in Grafana Loki
          - As raw flow logs exported to a collector

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, can optionally be used as the backend to store all collected flows. The Network Observability operator does not install Loki directly, however we provide some guidance to help you there.

          - For a production or production-like environment usage, refer to [the operator documentation](https://docs.openshift.com/container-platform/latest/network_observability/installing-operators.html).

          - For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 10GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          oc create namespace netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/1-storage.yaml) -n netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          If you prefer to not use Loki, you must set `spec.loki.enable` to `false` in `FlowCollector`.
          In that case, you can still get the Prometheus metrics or export raw flows to a custom collector. But be aware that some of the Console plugin features will be disabled. For instance, you will not be able to view raw flows there, and the metrics / topology will have a more limited level of details, missing information such as pods or IPs.

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for example [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/config/samples/flows_v1beta2_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          oc edit flowcollector cluster
          ```

          As it operates cluster-wide on every node, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you might have to configure differently if you used another installation method. Make sure to disable it (`spec.loki.enable`) if you don't want to use Loki.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`) an optional list of exporters to which to send enriched flows. KAFKA and IPFIX exporters are supported. This allows you to define any custom storage or processing that can read from Kafka or use the IPFIX standard.

          - To enable availability zones awareness, set `spec.processor.addZone` to `true`.

          ## Resource considerations

          The following table outlines examples of resource considerations for clusters with certain workload sizes.
          The examples outlined in the table demonstrate scenarios that are tailored to specific workloads. Consider each example only as a baseline from which adjustments can be made to accommodate your workload needs.


          | Resource recommendations                        | Extra small (10 nodes) | Small (25 nodes)       | Medium (65 nodes) **    | Large (120 nodes) **          |
          | ----------------------------------------------- | ---------------------- | ---------------------- | ----------------------- | ----------------------------- |
          | *Worker Node vCPU and memory*                   | 4 vCPUs\| 16GiB mem *  | 16 vCPUs\| 64GiB mem * | 16 vCPUs\| 64GiB mem  * |16 vCPUs\| 64GiB Mem *         |
          | *LokiStack size*                                | `1x.extra-small`       | `1x.small`             | `1x.small`              | `1x.medium`                   |
          | *Network Observability controller memory limit* | 400Mi (default)        | 400Mi (default)        | 400Mi (default)         | 800Mi                         |
          | *eBPF sampling rate*                            | 50 (default)           | 50 (default)           | 50 (default)            | 50 (default)                  |
          | *eBPF memory limit*                             | 800Mi (default)        | 800Mi (default)        | 2000Mi                  | 800Mi (default)               |
          | *FLP memory limit*                              | 800Mi (default)        | 800Mi (default)        | 800Mi (default)         | 800Mi (default)               |
          | *FLP Kafka partitions*                          | N/A                    | 48                     | 48                      | 48                            |
          | *Kafka consumer replicas*                       | N/A                    | 24                     | 24                      | 24                            |
          | *Kafka brokers*                                 | N/A                    | 3 (default)            | 3 (default)             | 3 (default)                   |

          *. Tested with AWS M6i instances.
          **. In addition to this worker and its controller, 3 infra nodes (size `M6i.12xlarge`) and 1 workload node (size `M6i.8xlarge`) were tested.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - More information on [configuring metrics](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/Metrics.md).
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: Network Observability
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/network-observability/network-observability-console-plugin-compat-rhel9@sha256:d130f94cdd3cc863a31c59f395a6c4f663e8e0da23747304353c6674b20d81be
        - registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:13b6dd0ddf4bdd2836b0d3ec9c79119a902456dcee100cfac6898768e1f22aab
        - registry.redhat.io/network-observability/network-observability-ebpf-agent-rhel9@sha256:0b4337f4db1e74a2bea1021a617cb42690251036f583bb771cc5ec888f38c97f
        - registry.redhat.io/network-observability/network-observability-flowlogs-pipeline-rhel9@sha256:09454618709a5c83beb4252f1c14c149716d1c09b6567c0a571d5390e8a639a2
        - registry.redhat.io/network-observability/network-observability-console-plugin-rhel9@sha256:7d92eadb3cd08a08672a65a196a5d3181e764c68e8c20cab2554933cdb10fd0c
        version: 1.9.3
      entries:
      - name: network-observability-operator.v1.9.3
        version: 1.9.3
      - name: network-observability-operator.v1.9.2
        version: 1.9.2
      - name: network-observability-operator.v1.9.1
        version: 1.9.1
      - name: network-observability-operator.v1.9.0
        version: 1.9.0
      - name: network-observability-operator.v1.8.1
        version: 1.8.1
      - name: network-observability-operator.v1.8.0
        version: 1.8.0
      - name: network-observability-operator.v1.7.0
        version: 1.7.0
      - name: network-observability-operator.v1.6.2
        version: 1.6.2
      - name: network-observability-operator.v1.6.1
        version: 1.6.1
      - name: network-observability-operator.v1.6.0
        version: 1.6.0
      - name: network-observability-operator.v1.5.0
        version: 1.5.0
      - name: network-observability-operator.v1.4.2
        version: 1.4.2
      - name: network-observability-operator.v1.4.1
        version: 1.4.1
      - name: network-observability-operator.v1.4.0
        version: 1.4.0
      - name: network-observability-operator.v1.3.0
        version: 1.3.0
      - name: network-observability-operator.v1.2.0
        version: 1.2.0
      - name: network-observability-operator.v1.1.0-202302110050
        version: 1.1.0
      - name: network-observability-operator.v1.0.0-202301070345
        version: 1.0.0
      name: stable
    - currentCSV: network-observability-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:fd943746840cba26996dabb47b1beb3445571fa43cb1b859fb073e4aa114b376
          createdAt: 2023-06-22T09:31:50
          description: Network flows collector and monitoring solution
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
          - description: FlowCollector is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
        description: |
          Network Observability is an OpenShift operator that deploys a monitoring pipeline to collect and enrich network flows that are produced by the Network Observability eBPF agent.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When a `FlowCollector` instance is created, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For normal usage, we recommend two options:

          - Installing the [Loki Operator](https://docs.openshift.com/container-platform/4.11//logging/cluster-logging-loki.html). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          oc create namespace netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.0.2/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.0.2/config/samples/flows_v1beta1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          oc edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.0.2/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: Network Observability
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:fd943746840cba26996dabb47b1beb3445571fa43cb1b859fb073e4aa114b376
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:37b3c9b2eed351d9d1a39c99045bbc7993ccb819f67f7b390deaa34930f45946
        - registry.redhat.io/network-observability/network-observability-console-plugin-rhel9@sha256:880c0812a2cd4b7194f417e31fbdcee951ce76776fa06e33c5c9cd479a854daa
        - registry.redhat.io/network-observability/network-observability-ebpf-agent-rhel9@sha256:e31acc12d8690c339e13b5dd22d33e707701362bdd31086c9b4c0f48cad62cc3
        - registry.redhat.io/network-observability/network-observability-flowlogs-pipeline-rhel9@sha256:7c7c42fff63e0599bb943a9e9617b43181360d5a5254b27f368010af275cf0c3
        version: 1.3.0
      entries:
      - name: network-observability-operator.v1.3.0
        version: 1.3.0
      - name: network-observability-operator.v1.2.0
        version: 1.2.0
      - name: network-observability-operator.v1.1.0-202302110050
        version: 1.1.0
      - name: network-observability-operator.v1.0.0-202301070345
        version: 1.0.0
      name: v1.0.x
    defaultChannel: stable
    packageName: netobserv-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: openshift-builds-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-builds-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "openshift-builds"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-rhel8-operator@sha256:7100a83aa053a976ceb77ace79f090ec8601bc9b2152ff6f7d00f0b1deea7f0b
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n\n## Usage\n\nTo deploy and manage [builds
          for Red Hat OpenShift](https://github.com/shipwright-io/build) in your cluster,\nfirst
          make sure this operator is installed and running on your cluster.\n\nNext,
          create the following:\n\n```yaml\n---\napiVersion: operator.shipwright.io/v1alpha1\nkind:
          ShipwrightBuild\nmetadata:\n  name: openshift-builds-operator\nspec:\n  targetNamespace:
          openshift-builds\n```\n\nThe operator will deploy Shipwright Build in the
          provided `targetNamespace`.\nWhen `.spec.targetNamespace` is not set, the
          namespace will default to ` openshift-builds`.\nRefer to the [ShipwrightBuild
          operator documentation](https://github.com/shipwright-io/operator/blob/main/docs/shipwrightbuild.md)
          for more information about this custom resource.\n"
        displayName: builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://shipwright.io
        - name: Shipwright Builds
          url: https://github.com/shipwright-io/build
        maintainers:
        - email: pipeline-integrations-team@redhat.com
          name: Red Hat Pipelines Integration Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-rhel8-operator@sha256:7100a83aa053a976ceb77ace79f090ec8601bc9b2152ff6f7d00f0b1deea7f0b
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel8@sha256:73efcb6ad82f698d7463718b06c86f2b2b3b5326689a62dc39c9f5d68f8943f2
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel8@sha256:0cf079a24b8b3a4b3e1f30710fea85039d9cf323cfedd06e7d1e33e6336fb848
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel8@sha256:22930726a25b11de2571da0395b53352939d142a16f4db6fd5460abd4cc37ba5
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel8@sha256:fd903804672ec08acf9a2b49aa7ecc5f9b543c12f7462637ef756733171b3817
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel8@sha256:f95348e5fc48676d93ebc919af2c27a041dfd632ec285296246d886daab11576
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel8@sha256:d01c1749ce76dd4bf5fefaa6c553725d33533d6e70e884e916f12217b674c3cf
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:97cade2c1ee468261aec5400728c8d44de387b459134aec7a4c3b5ec5a335d2c
        version: 1.0.2
      entries:
      - name: openshift-builds-operator.v1.0.2
        version: 1.0.2
      - name: openshift-builds-operator.v1.0.1
        version: 1.0.1
      - name: openshift-builds-operator.v1.0.0
        version: 1.0.0
      name: builds-1.0
    - currentCSV: openshift-builds-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          createdAt: "2025-01-23T14:48:43Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.openshift.com/builds/latest/about/overview-openshift-builds.html
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:778ad1946bdad68e1430a2fb366cee25cd98d87e44cf5ad34b87603c2f4171d0
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:e8582857566020c71653228355cbadc2c1cda41e36fc767c204a4edb05e7f42a
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:fa039b62ebd1fc6e86db8ec1c7ae5fe0dd3a6f6f642b63b4dc89af32a78cea6b
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:1ec9a1dcfc08b707df48175bba76f8a6cbfcc8396d9e9940cc3d7100e6731701
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:0e139db4370c37e91ef7eb3e0cf9a2587cb81f37709009a37f2724d4461ed2f3
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:949a1a5150642167b6c7d2b8f6d92a2885afe453319c02495f351b2e79eff3dd
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:7cfffdd43fdd33a20d115f35c2ba492da3c1d0e3ab970587bc888c68c47dfacf
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:97cade2c1ee468261aec5400728c8d44de387b459134aec7a4c3b5ec5a335d2c
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:e9f7b8741f265a332008424a2c5b3dae9ea154e623c469085c2abafb53fc62d4
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:5babdf590730b217deb59acbccf142c0500bc692542131810714cc00724ee02c
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:5ec474ab9e62c92e44c1a6635913032aead6246faec0895dcfbdb28ab0dc8a1a
        version: 1.1.1
      entries:
      - name: openshift-builds-operator.v1.1.1
        version: 1.1.1
      - name: openshift-builds-operator.v1.1.0
        version: 1.1.0
      name: builds-1.1
    - currentCSV: openshift-builds-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          createdAt: "2025-01-23T13:47:38Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.2
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:ee2792972dc52d81814603e26103e0d1117194db5ab3b5577736b0fc3285a8ab
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3fa22124916523b958c67af8ad652e73a2c3d68bb5579da1cba1ade537f3b7ae
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:8ace52981058ed51cec5b893e6d773de963b12ab715258ac1f2cbc6430ad52e7
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:775c15b93b19ba855791d34c4a95ef203dd9f8e51110c67ebf131e38eda49422
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:2b288e36e3606cf8a0d7723772edb8d4ac11772ec450539ef3448ad517155ecb
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:4bee8594d17172e5b9893a865a9922a1c7869027633ad64b3f684ec44755bac6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3fcd8e2bf0bcb8ff8c93a87af2c59a3bcae7be8792f9d3236c9b5bbd9b6db3b2
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:e8582857566020c71653228355cbadc2c1cda41e36fc767c204a4edb05e7f42a
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:78ee35ccc70b78638c667152b7b6247417c0aa1bb6d3f817d59b22eb179899ca
        - registry.redhat.io/ubi8/buildah@sha256:56c71ecd450f284325ba8434cbf5b3006078b11f0625c6258689699fa811c95d
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:ebfe4f8a97f6f0f9afe38953ae8cc2fdf2ec0b76401c819bb12c1df7cd671e9e
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:016c93578908f1dbdc3052333a1d5a9ab0c03104070ac36a6ddf99350d7510f4
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:87c4c0e9fe4d9baf8ae4f191c51d90c46729273e182fb5b3dfd951632e7bfad3
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:a91c2bb1239ef7fbce8c1ab9308da03b533df9cea75c7d978fd76f38087bc912
        version: 1.2.2
      entries:
      - name: openshift-builds-operator.v1.2.2
        version: 1.2.2
      - name: openshift-builds-operator.v1.2.1
        version: 1.2.1
      - name: openshift-builds-operator.v1.2.0
        version: 1.2.0
      name: builds-1.2
    - currentCSV: openshift-builds-operator.v1.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-01-29T08:58:54Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.2
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:6e44758fd4a7b34814ac7c71e52d93633781d39f7897ecab98a090859b487b4f
        - registry.redhat.io/ubi9/buildah@sha256:27d837f9bc69ad3c3651cf3315e2501b137c11baa553d9d46140e5cf7fa7873a
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:0783132b14ed8050664810f37e7ad6388fefbc586c45ecc7ba69dd5041a55c9f
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:e8582857566020c71653228355cbadc2c1cda41e36fc767c204a4edb05e7f42a
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:04b8bfc0dc6f260ef367ca1cce2bcc84032bc8be17b4a595be12a344813a2632
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:1e9a54212690f7930d2b681db87a6f79f5a2acc5c982734d4b283a2472b59db5
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:3523e7c2215e156081e5f05443884d65efad9b7b9ca81a35416695020231916a
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:d4b3f451becb7a42341c98dd5b9fc5d2b582c93cac2728070ef4bb83b64abb4c
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:a4ba591c57bbc815c93023f4a080213be32581ba6a1c2d94b0ecbc5c90b07a1e
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:0b1dd35f55181f34775188bc71b3ff8dff66c70acfcfabad19353a02e19351f7
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:6214708f3b517cb901af9f9bd0502ba54f8e5183d0d909a3b559deee6394026e
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:ed3fdd1e85f5c5434de6ec0feaec8c7fe4e680ca6c64258287f18ec2886bb71f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3fcd8e2bf0bcb8ff8c93a87af2c59a3bcae7be8792f9d3236c9b5bbd9b6db3b2
        version: 1.3.2
      entries:
      - name: openshift-builds-operator.v1.3.2
        version: 1.3.2
      - name: openshift-builds-operator.v1.3.1
        version: 1.3.1
      - name: openshift-builds-operator.v1.3.0
        version: 1.3.0
      name: builds-1.3
    - currentCSV: openshift-builds-operator.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2024-11-15T22:28:01Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:9d12e6cbf26e31a1e4458c99a972ed36125d759c5ec57773f203cd8dd0985de7
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:ed3fdd1e85f5c5434de6ec0feaec8c7fe4e680ca6c64258287f18ec2886bb71f
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:538416bc1384eb987e8a7e80b57eda3feddaf4260b196709a6e9a9125e3af39e
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:1a6ca51c2c68950af605ff367356ab6c0ef7d1b2fd6fb8f00aa665196ff7af84
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:f4b935c16dcbbc0393733437747497c1d24f2cc782c40c49738e5f37c037b3e7
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:c60ca44275297d2ee8c499d9d7fb4056d104f4bd5dec8a02e7388241ac466c44
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:1c7c24111f1709d7648156873ced09462638aa4b3cf7e7aee13297e51bcdfb3e
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:6251cb303774e29d137a37d6527875c371fe1d3f6a5fa0383e80f0e9ef5c02be
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:f87c2069105121927fb0c0a2e37f8f72bc6cb5e035fa8d22f7b41ce0e86c0631
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:4e9c152023b2db04ee5d55f102490294b393bf78baa3aa8917f85f8547fbe3c8
        - registry.redhat.io/ubi9/buildah@sha256:6886997da8c92214e99e3e8da47a0f08239172d2544b4edae058d5cb018effab
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:adc4a31d63b7044db7094fddcb1a4b235d848e842257f5baf22b6ead03fd6c91
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:5e87ba403f9e9fff3002a8275aba0d6c36aed54fb4754d90fa0516ad989f187a
        version: 1.4.1
      entries:
      - name: openshift-builds-operator.v1.4.1
        version: 1.4.1
      - name: openshift-builds-operator.v1.4.0
        version: 1.4.0
      name: builds-1.4
    - currentCSV: openshift-builds-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-08-14T12:12:15Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              OpenShiftBuild describes the desired state of Builds for OpenShift, and the status of
              all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I), Buildah, and Buildpacks. You can create and apply
          build resources, view logs of build runs, \nand manage builds in your OpenShift
          Container Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:38560dea7627632d8b24fa670c1d7ac1218cf86975329908786a460f76de558d
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ce7050192d9850ed4c35cb3295ccdeec1aa593e27e3a3da6d6aca97cde782abe
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c1a84feb88d53e93bdcf2cd5f76e2cbb18541c8b0e2979c132a888d6c280b664
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:ddcfcebd5d788d11f756939f07215b2e3507a1d8fe704a1052e15bb4ebe592e8
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:753e24a1cd57ca22552447363aed48f064b525b49e52f774f3e25f41d8dcb329
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:e2a7a42cd5916c05d95d3fb0decc34d0840420532270422aa7f6b7bdd94de1f2
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:bcae84267f582278591d93aecf94c8d7fb1376521d9b5208201c959a0fe56774
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:f4a75c115b5dc8853845f666f3bd063b2736e01e428a605fb5cd0f02736ecce8
        - registry.redhat.io/ubi9/buildah@sha256:8ea2b81b32f70b9502d88722a5f69de8c0cf5efa6f30df041873546e9f9438cb
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:4dc123fa703006662efcc5569a843a043bb31cf69b6868d6a9b6321c66bfca18
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:1d5a0bfe0013ab7602efe95be8a814be1e5586728fdf0e198c97739cc12f1d87
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:504179e8ee9f88a23f1d4c28c7683ab7395f4c1c4868765bf3c1406669fd52fc
        version: 1.5.1
      entries:
      - name: openshift-builds-operator.v1.5.1
        version: 1.5.1
      - name: openshift-builds-operator.v1.5.0
        version: 1.5.0
      name: builds-1.5
    - currentCSV: openshift-builds-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-09-23T13:38:59Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I), Buildah, and Buildpacks. You can create and apply
          build resources, view logs of build runs, \nand manage builds in your OpenShift
          Container Platform namespaces.\n\n## Prerequisites\n\n* OpenShift Pipelines
          operator must be installed before installing this operator\n\nRead more:
          [https://shipwright.io](https://shipwright.io)\n\n## Features\n\n* Standard
          Kubernetes-native API for building container images from source code and
          Dockerfile\n\n* Support for Source-to-Image (S2I) and Buildah build strategies\n\n*
          Extensibility with your own custom build strategies\n\n* Execution of builds
          from source code in a local directory\n\n* Shipwright CLI for creating and
          viewing logs, and managing builds on the cluster\n\n* Integrated user experience
          with the Developer perspective of the OpenShift Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:c6805e85ad4dc29f3a0582a1fbf99647a90c1af0b3924647839ad881cbb199fa
        - registry.redhat.io/ubi9/buildah@sha256:8ea2b81b32f70b9502d88722a5f69de8c0cf5efa6f30df041873546e9f9438cb
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:f9f43795c7ed1e416b92497cf2fe647beb5e45d755f6a6d89c3915b7a9c76570
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:1a61d4a8f69fa845ee678dd96f934685821828e7575852f97b6e7b1dd79a8297
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c1a84feb88d53e93bdcf2cd5f76e2cbb18541c8b0e2979c132a888d6c280b664
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:3d2e501abe1aa757e4e6ba03fe8e7a45bbfdb4303360ed7130c6344051887ed8
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:9f38cd42591b6a6b65e78a3cc5c165f8498ad3c79460a6c47acfb8744d2dd883
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:0ea2b0153ae07ee15972abbbe034ef7feb0dcaa403ac993afcb171906d9cc796
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:23e3f29ae63546db018701abb29b444b9b825fb3a2bda5084542649d9827d4ba
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:dd96e8c857bc52097d4d194db8fa2cb368bbdcbb6594722660d4e906d0bdc420
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ce7050192d9850ed4c35cb3295ccdeec1aa593e27e3a3da6d6aca97cde782abe
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:fd7c805c6d0638f2814503c6779719279119c7268b1eb0eda9b7b58ac775efa6
        version: 1.6.0
      entries:
      - name: openshift-builds-operator.v1.6.0
        version: 1.6.0
      name: builds-1.6
    - currentCSV: openshift-builds-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-09-23T13:38:59Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I), Buildah, and Buildpacks. You can create and apply
          build resources, view logs of build runs, \nand manage builds in your OpenShift
          Container Platform namespaces.\n\n## Prerequisites\n\n* OpenShift Pipelines
          operator must be installed before installing this operator\n\nRead more:
          [https://shipwright.io](https://shipwright.io)\n\n## Features\n\n* Standard
          Kubernetes-native API for building container images from source code and
          Dockerfile\n\n* Support for Source-to-Image (S2I) and Buildah build strategies\n\n*
          Extensibility with your own custom build strategies\n\n* Execution of builds
          from source code in a local directory\n\n* Shipwright CLI for creating and
          viewing logs, and managing builds on the cluster\n\n* Integrated user experience
          with the Developer perspective of the OpenShift Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:0ea2b0153ae07ee15972abbbe034ef7feb0dcaa403ac993afcb171906d9cc796
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:c6805e85ad4dc29f3a0582a1fbf99647a90c1af0b3924647839ad881cbb199fa
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:3d2e501abe1aa757e4e6ba03fe8e7a45bbfdb4303360ed7130c6344051887ed8
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:fd7c805c6d0638f2814503c6779719279119c7268b1eb0eda9b7b58ac775efa6
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:23e3f29ae63546db018701abb29b444b9b825fb3a2bda5084542649d9827d4ba
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ce7050192d9850ed4c35cb3295ccdeec1aa593e27e3a3da6d6aca97cde782abe
        - registry.redhat.io/ubi9/buildah@sha256:8ea2b81b32f70b9502d88722a5f69de8c0cf5efa6f30df041873546e9f9438cb
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:f9f43795c7ed1e416b92497cf2fe647beb5e45d755f6a6d89c3915b7a9c76570
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:1a61d4a8f69fa845ee678dd96f934685821828e7575852f97b6e7b1dd79a8297
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:9f38cd42591b6a6b65e78a3cc5c165f8498ad3c79460a6c47acfb8744d2dd883
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c1a84feb88d53e93bdcf2cd5f76e2cbb18541c8b0e2979c132a888d6c280b664
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:dd96e8c857bc52097d4d194db8fa2cb368bbdcbb6594722660d4e906d0bdc420
        version: 1.6.0
      entries:
      - name: openshift-builds-operator.v1.6.0
        version: 1.6.0
      - name: openshift-builds-operator.v1.5.1
        version: 1.5.1
      - name: openshift-builds-operator.v1.5.0
        version: 1.5.0
      - name: openshift-builds-operator.v1.4.1
        version: 1.4.1
      - name: openshift-builds-operator.v1.4.0
        version: 1.4.0
      - name: openshift-builds-operator.v1.3.2
        version: 1.3.2
      - name: openshift-builds-operator.v1.3.1
        version: 1.3.1
      - name: openshift-builds-operator.v1.3.0
        version: 1.3.0
      - name: openshift-builds-operator.v1.2.2
        version: 1.2.2
      - name: openshift-builds-operator.v1.2.1
        version: 1.2.1
      - name: openshift-builds-operator.v1.2.0
        version: 1.2.0
      - name: openshift-builds-operator.v1.1.1
        version: 1.1.1
      - name: openshift-builds-operator.v1.1.0
        version: 1.1.0
      - name: openshift-builds-operator.v1.0.2
        version: 1.0.2
      - name: openshift-builds-operator.v1.0.1
        version: 1.0.1
      - name: openshift-builds-operator.v1.0.0
        version: 1.0.0
      name: latest
    defaultChannel: latest
    packageName: openshift-builds-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: kaoto.io
      provider-url: https://kaoto.io
    name: kaoto-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kaoto-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "designer.kaoto.io/v1alpha1",
                "kind": "Kaoto",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "kaoto-operator",
                    "app.kubernetes.io/instance": "kaoto-demo",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kaoto",
                    "app.kubernetes.io/part-of": "kaoto-operator"
                  },
                  "name": "kaoto-demo"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          containerImage: quay.io/kaotoio/kaoto-operator@sha256:590d4958f148954ea1edb9825fb7fb4935d041ddcdb4a99302da14a605380071
          createdAt: "2025-09-03T12:48:58Z"
          description: Integration Editor for Apache Camel
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/KaotoIO/kaoto-operator
          support: contact@kaoto.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kaoto is the Schema for the kaotoes API.
            displayName: Kaoto
            kind: Kaoto
            name: kaotoes.designer.kaoto.io
            version: v1alpha1
        description: Kaoto is a low code and no code integration designer to edit
          and deploy based on Apache Camel . It is designed to be extendable and flexible
          and adaptable to different usecases.
        displayName: Kaoto Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - kaoto
        - camel-k
        - designer
        links:
        - name: Kaoto Operator
          url: https://kaoto-operator.domain
        maintainers:
        - email: info@kaoto.io
          name: kaoto
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: kaoto.io
          url: https://kaoto.io
        relatedImages:
        - quay.io/kaotoio/kaoto-operator@sha256:590d4958f148954ea1edb9825fb7fb4935d041ddcdb4a99302da14a605380071
        version: 2.7.0
      entries:
      - name: kaoto-operator.v2.7.0
        version: 2.7.0
      - name: kaoto-operator.v2.6.0
        version: 2.6.0
      - name: kaoto-operator.v2.5.0
        version: 2.5.0
      - name: kaoto-operator.v2.4.0
        version: 2.4.0
      - name: kaoto-operator.v2.3.0
        version: 2.3.0
      - name: kaoto-operator.v2.1.0
        version: 2.1.0
      - name: kaoto-operator.v2.0.1
        version: 2.0.1
      - name: kaoto-operator.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: kaoto-operator
    provider:
      name: kaoto.io
      url: https://kaoto.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/securesign/secure-sign-operator
    name: rhtas-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhtas-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": {
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname-intermediate",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname-leaf",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname-root",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname-intermediate",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname-leaf",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname-root",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": {
                  "database": {
                    "create": true
                  },
                  "server": {
                    "replicas": 1
                  },
                  "signer": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:0ebc6b3bd992309a914cc0da8ad8fc628224f583b66f1ca2cd10d2281b781cde
          createdAt: "2025-09-30T11:33:48Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:e7f5a9a2ea76fdfde34f6aac264b51c596cf4ab53f6922b58514c77a1977677e
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:f2798c21dbcb3334b256bf1370a42c013f8f3eefb0641e1c2f250428a82ddb82
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:2dd696771c0188a119d746711551b1643a14a579d0fc66d70918cca9de0c0107
        - registry.redhat.io/rhtas/createtree-rhel9@sha256:244758721aa1c63adfffbae31df58059ff8a06311f6cb60515eeae14b20e78ce
        - registry.redhat.io/rhtas/rekor-monitor-rhel9@sha256:da3aa5c441653f00c4b558ef6ebf21fef518731e70f152a1fd3a750b1145d901
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:b311fb891203fef071d53c9dab67a45961638af98bdfd8e02026dc2dbc25fd00
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:8f8a5d4abf7878ba34151f8632aa7db25218300dbb9f17b121c5f61e0e5c5e03
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:eaae082559c81c5e1253f1b5cf9e298b14235470c67e9111c18ece46078228d5
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:2d91dceb216d8cb8dacefcfe86f357f4ff4eb5e5e998f264d1ad231821a92fad
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:043103f7c888d150053b873276b92b2ee1907c8a7238652ad0770e36d44c10e2
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:0ebc6b3bd992309a914cc0da8ad8fc628224f583b66f1ca2cd10d2281b781cde
        - registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:08302a0af012c1a1f04ef15de86558c9eef55620ab7b179fbda170025f0584be
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:497203991dcbca71275d691ec970ba480cb8e80c9ada6af876f5559abe0d6155
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:110509f5f24240d79afff7cb0d8a4ead0ed038c6100df8e435b01f270b7d83b2
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:abaa247777eb05bf8c97ba1ffc30c00d0d84590fac2fdfbffa06ba374b924888
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:fcb225019f865dd4b7af1f5134a9aed9c5c8a425f3ee43b74e0517ce602e209c
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:e6a5bf7804f1c4371e4de06805d6e1ddfc98e356e355ac3b55a3be25c5636b4d
        version: 1.3.0
      entries:
      - name: rhtas-operator.v1.3.0
        version: 1.3.0
      - name: rhtas-operator.v1.2.1
        version: 1.2.1
      - name: rhtas-operator.v1.2.0
        version: 1.2.0
      - name: rhtas-operator.v1.1.2
        version: 1.1.2
      - name: rhtas-operator.v1.1.1
        version: 1.1.1
      - name: rhtas-operator.v1.1.0
        version: 1.1.0
      - name: rhtas-operator.v1.0.2
        version: 1.0.2
      - name: rhtas-operator.v1.0.1
        version: 1.0.1
      - name: rhtas-operator.v1.0.0
        version: 1.0.0
      name: stable
    - currentCSV: rhtas-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e1a78636cd7cc8e7227b6b0a5ddfd93d002db0464b15e966e88337d38989ea5a
          createdAt: "2024-06-27T09:45:39Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e1a78636cd7cc8e7227b6b0a5ddfd93d002db0464b15e966e88337d38989ea5a
        version: 1.0.2
      entries:
      - name: rhtas-operator.v1.0.2
        version: 1.0.2
      - name: rhtas-operator.v1.0.1
        version: 1.0.1
      - name: rhtas-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: rhtas-operator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e52d07e2cd8cbdc2ccd0336a4fd29cbb2e5b8a1be628350294a54579cc2636ed
          createdAt: "2025-04-08T09:29:59Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e52d07e2cd8cbdc2ccd0336a4fd29cbb2e5b8a1be628350294a54579cc2636ed
        version: 1.1.2
      entries:
      - name: rhtas-operator.v1.1.2
        version: 1.1.2
      - name: rhtas-operator.v1.1.1
        version: 1.1.1
      - name: rhtas-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: rhtas-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
          createdAt: "2025-08-17T21:52:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:02fbd4772c185edf38fcee4e5c5e1c9dad82b77c9da824addaebc445b67d35ac
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
        - registry.redhat.io/ubi9/httpd-24@sha256:badeee0aa3477e61de64f6a2b93e8a3673edd38fc860125f7536506f3d592132
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:c22cdb9a17f85e13c00412cba83aa644e90a0faea0236029d0f1c53fe212e227
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:12b438ea70f83ee6edb12c96b507f92510f4afba824eac34c8f50793c4199db6
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:fc018a45b7eb48690c46c83f52cac55fa40089e6b02686bc21265dc7c2205b8f
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:1d782a127d2f8bdab633f287cb2df2e400ae91bf5d4b913473539d8512dfaab6
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:09b4aeeb607c88c72f69e6f87cb840c82ef5752c64a58fa13551db749fac2530
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:eb38e98dbb9828fe033a5388609132f7246fe77d2f4d258d015a94ea30752b24
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:6270ba29adf779de5d3916d5e82f1cd9491ba89d4ae2042347b6ad6d1819dd26
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:1c2201d50469d70ec6e21546c7c74bd52251d420dc1dcfa5375c1cf61dd3a9fd
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:030986ea26f33db3a192c67a93cd2bde352f23c68a17b062fc062955675c3c51
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:531603563c69aac6905cd6b8157d30781040527404ac2313acba5a9310e0b713
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:fe31830df2d99359749cb86cd1ac2f32fcc3de07aae427d2f2cd3998d31d2833
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:a93df3206409de75613d398e7dcd8b30ff7cc36eb00349c435784f73f00a2335
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:640633823a54c11fe2a8b3d05e571399fda415b6f0b3adf1ba9806882a94bdd4
        version: 1.2.1
      entries:
      - name: rhtas-operator.v1.2.1
        version: 1.2.1
      - name: rhtas-operator.v1.2.0
        version: 1.2.0
      name: stable-v1.2
    - currentCSV: rhtas-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": {
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname-intermediate",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname-leaf",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname-root",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname-intermediate",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname-leaf",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname-root",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": {
                  "database": {
                    "create": true
                  },
                  "server": {
                    "replicas": 1
                  },
                  "signer": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:0ebc6b3bd992309a914cc0da8ad8fc628224f583b66f1ca2cd10d2281b781cde
          createdAt: "2025-09-30T11:33:48Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:abaa247777eb05bf8c97ba1ffc30c00d0d84590fac2fdfbffa06ba374b924888
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:eaae082559c81c5e1253f1b5cf9e298b14235470c67e9111c18ece46078228d5
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:2d91dceb216d8cb8dacefcfe86f357f4ff4eb5e5e998f264d1ad231821a92fad
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:e6a5bf7804f1c4371e4de06805d6e1ddfc98e356e355ac3b55a3be25c5636b4d
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:8f8a5d4abf7878ba34151f8632aa7db25218300dbb9f17b121c5f61e0e5c5e03
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:2dd696771c0188a119d746711551b1643a14a579d0fc66d70918cca9de0c0107
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:110509f5f24240d79afff7cb0d8a4ead0ed038c6100df8e435b01f270b7d83b2
        - registry.redhat.io/rhtas/createtree-rhel9@sha256:244758721aa1c63adfffbae31df58059ff8a06311f6cb60515eeae14b20e78ce
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:b311fb891203fef071d53c9dab67a45961638af98bdfd8e02026dc2dbc25fd00
        - registry.redhat.io/rhtas/rekor-monitor-rhel9@sha256:da3aa5c441653f00c4b558ef6ebf21fef518731e70f152a1fd3a750b1145d901
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:08302a0af012c1a1f04ef15de86558c9eef55620ab7b179fbda170025f0584be
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:043103f7c888d150053b873276b92b2ee1907c8a7238652ad0770e36d44c10e2
        - registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:e7f5a9a2ea76fdfde34f6aac264b51c596cf4ab53f6922b58514c77a1977677e
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:fcb225019f865dd4b7af1f5134a9aed9c5c8a425f3ee43b74e0517ce602e209c
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:0ebc6b3bd992309a914cc0da8ad8fc628224f583b66f1ca2cd10d2281b781cde
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:497203991dcbca71275d691ec970ba480cb8e80c9ada6af876f5559abe0d6155
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:f2798c21dbcb3334b256bf1370a42c013f8f3eefb0641e1c2f250428a82ddb82
        version: 1.3.0
      entries:
      - name: rhtas-operator.v1.3.0
        version: 1.3.0
      name: stable-v1.3
    defaultChannel: stable
    packageName: rhtas-operator
    provider:
      name: Red Hat
      url: https://github.com/securesign/secure-sign-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-lambda-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-lambda-controller.v1.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "Alias",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "CodeSigningConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "EventSourceMapping",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/lambda-controller:1.9.1
          createdAt: "2025-09-19T19:23:07Z"
          description: AWS Lambda controller is a service controller for managing
            Lambda resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alias represents the state of an AWS lambda Alias resource.
            displayName: Alias
            kind: Alias
            name: aliases.lambda.services.k8s.aws
            version: v1alpha1
          - description: CodeSigningConfig represents the state of an AWS lambda CodeSigningConfig
              resource.
            displayName: CodeSigningConfig
            kind: CodeSigningConfig
            name: codesigningconfigs.lambda.services.k8s.aws
            version: v1alpha1
          - description: EventSourceMapping represents the state of an AWS lambda
              EventSourceMapping resource.
            displayName: EventSourceMapping
            kind: EventSourceMapping
            name: eventsourcemappings.lambda.services.k8s.aws
            version: v1alpha1
          - description: Function represents the state of an AWS lambda Function resource.
            displayName: Function
            kind: Function
            name: functions.lambda.services.k8s.aws
            version: v1alpha1
          - description: FunctionURLConfig represents the state of an AWS lambda FunctionURLConfig
              resource.
            displayName: FunctionURLConfig
            kind: FunctionURLConfig
            name: functionurlconfigs.lambda.services.k8s.aws
            version: v1alpha1
          - description: LayerVersion represents the state of an AWS lambda LayerVersion
              resource.
            displayName: LayerVersion
            kind: LayerVersion
            name: layerversions.lambda.services.k8s.aws
            version: v1alpha1
          - description: Version represents the state of an AWS lambda Version resource.
            displayName: Version
            kind: Version
            name: versions.lambda.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Lambda resources in AWS from within your Kubernetes cluster.

          **About Amazon Lambda**

          Lambda is a compute service that lets you run code without provisioning or managing servers. Lambda runs your code on a high-availability compute infrastructure and performs all of the administration of the compute resources, including server and operating system maintenance, capacity provisioning and automatic scaling, code monitoring and logging. With Lambda, you can run code for virtually any type of application or backend service. All you need to do is supply your code in one of the [languages that Lambda supports](https://docs.aws.amazon.com/lambda/latest/dg/lambda-runtimes.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Lambda
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lambda
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Lambda Developer Resources
          url: https://aws.amazon.com/lambda/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: lambda maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/lambda-controller:1.9.1
        version: 1.9.1
      entries:
      - name: ack-lambda-controller.v1.9.1
        version: 1.9.1
      - name: ack-lambda-controller.v1.9.0
        version: 1.9.0
      - name: ack-lambda-controller.v1.8.0
        version: 1.8.0
      - name: ack-lambda-controller.v1.7.7
        version: 1.7.7
      - name: ack-lambda-controller.v1.7.6
        version: 1.7.6
      - name: ack-lambda-controller.v1.7.5
        version: 1.7.5
      - name: ack-lambda-controller.v1.7.4
        version: 1.7.4
      - name: ack-lambda-controller.v1.7.2
        version: 1.7.2
      - name: ack-lambda-controller.v1.7.1
        version: 1.7.1
      - name: ack-lambda-controller.v1.7.0
        version: 1.7.0
      - name: ack-lambda-controller.v1.6.3
        version: 1.6.3
      - name: ack-lambda-controller.v1.6.2
        version: 1.6.2
      - name: ack-lambda-controller.v1.6.1
        version: 1.6.1
      - name: ack-lambda-controller.v1.6.0
        version: 1.6.0
      - name: ack-lambda-controller.v1.5.8
        version: 1.5.8
      - name: ack-lambda-controller.v1.5.7
        version: 1.5.7
      - name: ack-lambda-controller.v1.5.6
        version: 1.5.6
      - name: ack-lambda-controller.v1.5.5
        version: 1.5.5
      - name: ack-lambda-controller.v1.5.4
        version: 1.5.4
      - name: ack-lambda-controller.v1.5.3
        version: 1.5.3
      - name: ack-lambda-controller.v1.5.2
        version: 1.5.2
      - name: ack-lambda-controller.v1.5.1
        version: 1.5.1
      - name: ack-lambda-controller.v1.5.0
        version: 1.5.0
      - name: ack-lambda-controller.v1.4.2
        version: 1.4.2
      - name: ack-lambda-controller.v1.4.0
        version: 1.4.0
      - name: ack-lambda-controller.v1.3.4
        version: 1.3.4
      - name: ack-lambda-controller.v1.3.2
        version: 1.3.2
      - name: ack-lambda-controller.v1.3.1
        version: 1.3.1
      - name: ack-lambda-controller.v1.3.0
        version: 1.3.0
      - name: ack-lambda-controller.v1.2.1
        version: 1.2.1
      - name: ack-lambda-controller.v1.2.0
        version: 1.2.0
      - name: ack-lambda-controller.v1.1.1
        version: 1.1.1
      - name: ack-lambda-controller.v1.1.0
        version: 1.1.0
      - name: ack-lambda-controller.v1.0.1
        version: 1.0.1
      - name: ack-lambda-controller.v1.0.0
        version: 1.0.0
      - name: ack-lambda-controller.v0.1.7
        version: 0.1.7
      - name: ack-lambda-controller.v0.1.6
        version: 0.1.6
      - name: ack-lambda-controller.v0.1.5
        version: 0.1.5
      - name: ack-lambda-controller.v0.1.3
        version: 0.1.3
      - name: ack-lambda-controller.v0.1.2
        version: 0.1.2
      - name: ack-lambda-controller.v0.1.1
        version: 0.1.1
      - name: ack-lambda-controller.v0.1.0
        version: 0.1.0
      - name: ack-lambda-controller.v0.0.17
        version: 0.0.17
      - name: ack-lambda-controller.v0.0.16
        version: 0.0.16
      - name: ack-lambda-controller.v0.0.15
        version: 0.0.15
      - name: ack-lambda-controller.v0.0.14
        version: 0.0.14
      - name: ack-lambda-controller.v0.0.13
        version: 0.0.13
      - name: ack-lambda-controller.v0.0.12
        version: 0.0.12
      - name: ack-lambda-controller.v0.0.11
        version: 0.0.11
      - name: ack-lambda-controller.v0.0.10
        version: 0.0.10
      - name: ack-lambda-controller.v0.0.9
        version: 0.0.9
      - name: ack-lambda-controller.v0.0.8
        version: 0.0.8
      - name: ack-lambda-controller.v0.0.7
        version: 0.0.7
      - name: ack-lambda-controller.v0.0.6
        version: 0.0.6
      - name: ack-lambda-controller.v0.0.5
        version: 0.0.5
      - name: ack-lambda-controller.v0.0.4
        version: 0.0.4
      - name: ack-lambda-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-lambda-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: fence-agents-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fence-agents-remediation.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/fence-agents-remediation-operator:v0.4.0
          createdAt: "2024-05-23 19:57:04"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://www.medik8s.io/remediation/fence-agents-remediation/fence-agents-remediation/
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/medik8s/fence-agents-remediation-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      - name: fence-agents-remediation.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: fence-agents-remediation
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: hpe-filex-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-filex-csi-operator.v1.0.0-beta3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.hpe.com/v1",
                "kind": "HPEGreenLakeFileCSIDriver",
                "metadata": {
                  "name": "hpegreenlakefilecsidriver-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "disableNodeConformance": false,
                  "imagePullPolicy": "IfNotPresent",
                  "images": {
                    "csiAttacher": "registry.k8s.io/sig-storage/csi-attacher@sha256:6e54dae32284f60a2de1645c527715e4c934e5ce7899560c0d845bac957118dc",
                    "csiControllerDriver": "quay.io/hpestorage/filex-csi-driver@sha256:8e96df8f21db969dcd507c6fef8ecc82b7f4d9b33a9159f16345ebe7e02d8161",
                    "csiNodeDriver": "quay.io/hpestorage/filex-csi-driver@sha256:8e96df8f21db969dcd507c6fef8ecc82b7f4d9b33a9159f16345ebe7e02d8161",
                    "csiNodeDriverRegistrar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f",
                    "csiNodeInit": "quay.io/hpestorage/filex-csi-init@sha256:65a014b344d262f628100dcd96fecd228c42162bbfa67037952a5a67f16a1437",
                    "csiProvisioner": "registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f",
                    "csiResizer": "registry.k8s.io/sig-storage/csi-resizer@sha256:ab774734705a906561e15b67f6a96538f3ad520335d636f793aaafe87a3b5200",
                    "csiSnapshotter": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:dd788d79cf4c1b8edee6d9b80b8a1ebfc51a38a365c5be656986b129be9ac784"
                  },
                  "kubeletRootDir": "/var/lib/kubelet",
                  "node": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/hpestorage/filex-csi-driver-operator@sha256:e00066fccf2c673547ccafffb0a280130b6825ab6a82a52bd43d752535c63775
          createdAt: "2025-01-10T18:20:35Z"
          description: A Container Storage Interface (CSI) driver for HPE GreenLake
            File Storage.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.0.0-beta3'
          operatorhub.io/ui-metadata-max-k8s-version: "1.32"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/hpe-storage/co-deployments
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE GreenLake File Storage CSI Driver
            displayName: HPEGreenLakeFileCSIDriver
            kind: HPEGreenLakeFileCSIDriver
            name: hpegreenlakefilecsidrivers.storage.hpe.com
            version: v1
        description: |
          The HPE GreenLake File Storage CSI Operator deploys the  HPE GreenLake File Storage CSI Driver on OpenShift for dynamic  provisioning of persistent volumes.
          Please refer to our CSI driver [documentation](https://scod.hpedev.io/filex_csi_driver/index.html) for supported HPE platforms and full list of supported CSI features.
          ## Installation
          Refer to the HPE GreenLake File Storage CSI Operator [documentation](https://scod.hpedev.io/filex_csi_driver/deployment.html#operator).
        displayName: HPE GreenLake File Storage CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE CSI
        - HPE Storage
        - CSI
        links:
        - name: Documentation
          url: https://scod.hpedev.io/filex_csi_driver/index.html
        maintainers:
        - email: dev-hi-containers@hpe.com
          name: HPE Storage Containers Team
        maturity: beta
        minKubeVersion: 1.25.0
        provider:
          name: HPE
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - quay.io/hpestorage/filex-csi-init@sha256:65a014b344d262f628100dcd96fecd228c42162bbfa67037952a5a67f16a1437
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:dd788d79cf4c1b8edee6d9b80b8a1ebfc51a38a365c5be656986b129be9ac784
        - registry.k8s.io/sig-storage/csi-resizer@sha256:ab774734705a906561e15b67f6a96538f3ad520335d636f793aaafe87a3b5200
        - quay.io/hpestorage/filex-csi-driver@sha256:8e96df8f21db969dcd507c6fef8ecc82b7f4d9b33a9159f16345ebe7e02d8161
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3fcd8e2bf0bcb8ff8c93a87af2c59a3bcae7be8792f9d3236c9b5bbd9b6db3b2
        - registry.connect.redhat.com/hpestorage/filex-csi-driver-operator@sha256:e00066fccf2c673547ccafffb0a280130b6825ab6a82a52bd43d752535c63775
        - registry.k8s.io/sig-storage/csi-attacher@sha256:6e54dae32284f60a2de1645c527715e4c934e5ce7899560c0d845bac957118dc
        version: 1.0.0-beta3
      entries:
      - name: hpe-filex-csi-operator.v1.0.0-beta3
        version: 1.0.0-beta3
      name: beta
    defaultChannel: beta
    packageName: hpe-filex-csi-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SmartX
      provider-url: https://www.smartx.com
    name: iomesh-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: iomesh-operator.0.10.1-rc4
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
          createdAt: 2021-08-17 00:00:00+00:00
          repository: https://github.com/iomesh/docs
          support: SmartX, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: iomesh
            displayName: iomesh
            kind: IOMeshCluster
            name: iomeshclusters.iomesh.com
            version: v1alpha1
        description: "IOMesh (https://www.iomesh.com/) is a cloud-native storage product
          specifically designed and developed for Kubernetes. It takes ZBS as its
          core, a “production-ready” distributed storage system developed independently
          by SmartX, to provide production-level high performance and reliable persistent
          storage for business-critical applications such as MySQL, Cassandra, MongoDB,
          etc., contributing to the containerized transformation of stateful applications.
          \     \n## Pre-Requisites\n* According to https://docs.iomesh.com/deploy/prerequisites,
          make sure that your openshift cluster meets the \"Installation Requirements\"
          and set up the worker nodes correctly according to the \"Setup Worker Node\"
          chapter.\n* Run IOMesh Operator pre-installation script in an environment
          where oc or kubectl can be used to access the openshift cluster, the script
          will install the dependencies of IOMesh Operator and specific IOMesh settings
          for the openshift cluster:\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-pre-install-openshift.sh
          | sh -\n```\n## Install IOMesh Operator and IOMesh Cluster\n* Click install
          button on the current web page to install IOMesh Operator\n* On the **Installed
          Operators** -> **IOMesh Operator** -> **Create instance** -> **YAML view**,
          fill in an IOMesh Custom Resources according https://iomesh.run/iomesh.yaml,
          change the `spec.*.dataCIDR` to your own data network CIDR:\n## Post Install\n*
          Run IOMesh Operator post-installation script in an environment where oc
          or kubectl can be used to access the openshift cluster, the script will
          install IOMesh CSI Driver\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-post-install-openshift.sh
          | sh -\n```\n## Getting Started\n* Try to quickstart setup and use IOMesh:
          https://docs.iomesh.com/deploy/setup-iomesh\n"
        displayName: IOMesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iomesh-operator
        links:
        - name: IOMesh Operator
          url: https://iomesh.com
        maintainers:
        - email: na.wang@smartx.com
          name: Na Wang
        maturity: alpha
        provider:
          name: SmartX
          url: https://www.smartx.com
        relatedImages:
        - iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
        version: 0.10.1-rc4
      entries:
      - name: iomesh-operator.0.10.1-rc4
        version: 0.10.1-rc4
      name: stable
    defaultChannel: stable
    packageName: iomesh-operator
    provider:
      name: SmartX
      url: https://www.smartx.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-activemqartemis: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/
    name: amq-broker-rhel9
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-broker-operator.v7.13.1-opr-1-0.1758027574.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 9 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-broker-operator.v7.13.1-opr-1-0.1756381365.p
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 9 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel9-operator@sha256:f34672152017a94ed51056435e218341e22787788ceed52439c1ff066480c715
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9cc48eecf1442ae04b8543fa5d4381a13bc2831390850828834d387006d1342b
        - registry.redhat.io/amq7/amq-broker-rhel9@sha256:755d0f7421efba76f52818bb3a8667cd7d2dcff2c7054dac1a4637b74ecac51d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:47fd4ce2533496828aba37bd1f9715e2164d5c90bd0fc6b25e7e0786d723bf01
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b2402179b1d0ff9c950defe5d8808195ef9a4c3e13e3c469cae31ab51d5f5014
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:bb6fbd68475a7852b4d99eea6c4ab313f9267da7963162f0d75375d7063409e7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:40920687402fd747ca02d384e63692a5e1db625723a2bf3056a607c92ee8eee1
        - registry.redhat.io/amq7/amq-broker-rhel9@sha256:4304a5414911fe55ce4c0133723be234ff288463f4916bd18e743d6359489138
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:7f9b921d89b6d4d91b365659afff127d99f1455b94b48a10a6a7f4c38dd8b700
        - registry.redhat.io/amq7/amq-broker-init-rhel9@sha256:ff80e0a7accda6825bf834d9aec27f48205dd6329991890dd017b4d21d0da5c8
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:16253895bc113f395dc6ab5cd511901e43b6ee676369ec80d598a8741c3c9bf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d42d713da0ce6806fdc6492b6342586783e6865a82a8647d3c4288439b1751ee
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:dc13871dfed52402c40fd192900fc3cb5a49da4fa08e4ec4d0e23a879e83db25
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ffffa9875f0379e9373f89f05eb06e5a193273bb04bc3aa5f85b044357b79098
        - registry.redhat.io/amq7/amq-broker-init-rhel9@sha256:e2cb996e0bf131e71fb8e80c128af18ab2b9330c9fca7a5437842796158e2870
        version: 7.13.1-opr-1+0.1758027574.p
      entries:
      - name: amq-broker-operator.v7.13.1-opr-1
        version: 7.13.1-opr-1
      - name: amq-broker-operator.v7.13.1-opr-1-0.1754564924.p
        version: 7.13.1-opr-1+0.1754564924.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1756374447.p
        version: 7.13.1-opr-1+0.1756374447.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1756381365.p
        version: 7.13.1-opr-1+0.1756381365.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1758024989.p
        version: 7.13.1-opr-1+0.1758024989.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1758027574.p
        version: 7.13.1-opr-1+0.1758027574.p
      - name: amq-broker-operator.v7.13.0-opr-1
        version: 7.13.0-opr-1
      - name: amq-broker-operator.v7.13.0-opr-1-0.1747315188.p
        version: 7.13.0-opr-1+0.1747315188.p
      - name: amq-broker-operator.v7.13.0-opr-1-0.1747658456.p
        version: 7.13.0-opr-1+0.1747658456.p
      name: 7.13.x
    defaultChannel: 7.13.x
    packageName: amq-broker-rhel9
    provider:
      name: Red Hat
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: quay-bridge-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-bridge-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "my-quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecretName": "<NAMESPACE>/<SECRET>",
                  "quayHostname": "https://<QUAY_URL>",
                  "insecureRegistry": false,
                  "whitelistNamespaces": ["default"]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:501fe9b003b5066b3936def0ef4de05abd1a12e795e718d6c9a62d2c6c804c2e
          createdAt: 2021-11-02 11:11 UTC
          description: Enhance OCP using Red Hat Quay container registry
          repository: https://github.com/redhat-cop/quay-openshift-registry-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a connected Quay instance.
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.redhatcop.redhat.io
            version: v1alpha1
        description: "## Overview\n\nThe OpenShift Container Platform contains an
          in cluster image registry for pushing and pulling images that either produced
          within the cluster or are references from external sources. Quay is a highly
          available, enterprise grade image registry with enhanced role based access
          control (RBAC) features. The goal of this operator is to support replicating
          the features found within OpenShift intgrated image registry with Quay.\n\n##
          Functionality\n\nThe following capabilities are enabled through this operator:\n\n*
          Synchronization of OpenShift namespaces as Quay organizations\n    * Creation
          of Robot accounts for each default namespace service account\n    * Creation
          of Secrets for each created Robot Account\n        * Associate Robot Secret
          to Service Account as Mountable and Image Pull Secret\n    * Synchronization
          of OpenShift ImageStreams as Quay Repositories\n* Automatically rewrite
          new Builds making use of ImageStreams to output to Quay\n* Automatically
          import ImageStream tag once build completes\n\n## Prerequisites\n\nThe following
          requirements must be satisfied prior to setting and configuring the integration
          between OpenShift and Quay:\n\n* A Red Hat Quay environment with a user
          with superuser permissions\n* A Red Hat OpenShift Container Platform environment
          (at least version 4.0) with cluster-admin permissions on an OpenShift cluster
          \n*\tOpenShift Command Line tool (CLI)\n\n## Setup and Configuration\n\nThe
          setup process involves steps within both Quay and OpenShift as described
          below:\n\n### Quay Setup\n\nThe integration between OpenShift and Quay makes
          extensive use of the Quay RESTFul API. To allow for other methods aside
          from user credentials to be used for authenticating against the API, an
          application is required to be created. When an application is created, an
          OAuth token is generated which can be used by the operator in OpenShift.
          Applications are created within a Quay organization. While an existing organization
          can be utilized, it is recommended that a dedicated organization be used.\n\n1.\tLogin
          to Quay as a user with superuser access and select the organization for
          which the external application will be configured within.\n2.\tOn the left
          hand side of the page, select **Applications**\n3.\tCreate a new application
          by selecting the **Create New Application** button and entering a name for
          the application\n4.\tAfter the new application has been created, select
          the application that was created\n5.\tOn the lefthand navigation bar, select
          the **Generate Token** button to create a new OAuth2 token\n6.\tSelect all
          of the checkboxes to grant the necessary access that is needed for the integration\n7.\tReview
          the permissions that will be assigned and then select the **Authorize Application**
          button\n8.\tTake note of the generated Access Token as it will be needed
          in the following section\n\n### OpenShift Setup\n\nTo complete the process
          to begin to integrate OpenShift and Quay, several steps are required. Before
          beginning, ensure that you have the OpenShift Command Line tool installed
          and that you are logged into OpenShift as a cluster administrator. \n\n####
          Create A Secret for the Quay OAuth Token\n\nThe Operator will use the previously
          obtained Access Token to communicate with Quay. Store this token within
          OpenShift as a secret.\n\nExecute the following command to create a secret
          called `quay-integration` with a key called `token` containing the access
          token:\n\n```\n$ oc create secret generic quay-integration --from-literal=token=<access_token>\n```\n\nThis
          token will be referenced in the final section.\n\n#### MutatingWebhookConfiguration
          Support\n\nSupport for dynamic interception of API requests that are performed
          as part of OpenShift’s typical build process is facilitated through a [MutatingWebhookConfiguration](https://docs.openshift.com/container-platform/3.11/architecture/additional_concepts/dynamic_admission_controllers.html).
          A MutatingWebhookConfiguration allows for invoking an API running within
          a project on OpenShift when certain API requests are received. In particular,
          to support Quay’s integration with OpenShift, any new Build requests should
          be intercepted so that the output can be modified to target Quay instead
          of OpenShift’s integrated registry. \n\nKubernetes requires that the webhook
          endpoint is secured via SSL using a certificate that makes use of the certificate
          authority for the cluster. Fortunately, OpenShift provides support for generating
          a certificate signed by the cluster. A script is available in the documentation
          called `webhook-create-signed-cert.sh` that automates the process for requesting
          a new certificate.\n\nFirst, create a `Service` for the Operator's webhook
          endpoint:\n```yaml\napiVersion: v1\nkind: Service\nmetadata:\n  labels:\n
          \   name: quay-bridge-operator\n  name: quay-bridge-operator\n  namespace:
          openshift-operators\nspec:\n  ports:\n    - name: https\n      port: 443\n
          \     protocol: TCP\n      targetPort: 8443\n  selector:\n    name: quay-bridge-operator\n
          \ sessionAffinity: None\n  type: ClusterIP\n```\n\nExecute the following
          command to request the certificate:\n\n```\n./webhook-create-signed-cert.sh
          --namespace openshift-operators --secret quay-bridge-operator-webhook-certs
          --service quay-bridge-operator\n```\n\nThe result will place the newly created
          private key and certificate within a secret within the secret specified.
          The secret will be mounted into the appropriate located within the operator
          as declared in the Deployment of the Operator. \n\nOnce the secret has been
          created, focus can shift to the `MutatingWebhookConfiguration` (see below).
          The majority of the content should be in place to apply to the cluster.
          The one section that requires configuration is the caBundle property. This
          refers to the Certificate Authority (CA) for the OpenShift environment.
          The OpenShift CA is available as a _ConfigMap_ within the `kube-system`
          namespace.\n\nExecute the following command to retrieve the CA and format
          the result as a single line so that it can be entered into the MutatingWebhookConfiguration
          resource:\n\n```\noc get configmap -n kube-system extension-apiserver-authentication
          -o=jsonpath='{.data.client-ca-file}' | base64 | tr -d '\\n'\n```\n\nReplace
          the `${CA_BUNDLE}` variable in the following `MutatingWebhookConfiguration`
          YAML and create it:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\nkind:
          MutatingWebhookConfiguration \nmetadata:\n  name: quay-bridge-operator\nwebhooks:\n
          \ - name: quayintegration.redhatcop.redhat.io\n    clientConfig: \n      service:\n
          \       namespace: openshift-operators\n        name: quay-bridge-operator\n
          \       path: \"/admissionwebhook\"\n      caBundle: \"${CA_BUNDLE}\"\n
          \   rules: \n    - operations:  [ \"CREATE\" ]\n      apiGroups: [ \"build.openshift.io\"
          ]\n      apiVersions: [\"v1\" ]\n      resources: [ \"builds\" ]\n    failurePolicy:
          Fail\n```\n\n_Note: Until the operator is running, new requests for builds
          will fail since the webserver the MutatingWebhookConfiguration invokes is
          not available and a proper is response is required in order for the object
          to be persisted in etcd._\n\n#### Create the QuayIntegration Custom Resource\n\nFinally,
          to complete the integration between OpenShift and Quay, a `QuayIntegration`
          custom resource needs to be created.\n\nThe following is an example of a
          basic definition of a `QuayIntegration` resource associated from the associated
          CRD.\n\n```\napiVersion: redhatcop.redhat.io/v1alpha1\nkind: QuayIntegration\nmetadata:\n
          \ name: example-quayintegration\nspec:\n  clusterID: openshift\n  credentialsSecretName:
          <NAMESPACE>/<SECRET>\n  quayHostname: https://<QUAY_URL>\n```\n\nThe _clusterID_
          is a value which should be unique across the entire ecosystem. This value
          is optional and defaults to `openshift`.\n\nThe _credentialsSecretName_
          is a NamespacedName value of the secret containing the token that was previously
          created.\n\nNote: If Quay is using self signed certificates, the property
          `insecureRegistry: true`\n\nOrganizations within Quay should be created
          for the related namespaces in OpenShift."
        displayName: Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Source Code
          url: https://github.com/redhat-cop/quay-openshift-registry-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:501fe9b003b5066b3936def0ef4de05abd1a12e795e718d6c9a62d2c6c804c2e
        version: 3.4.7
      entries:
      - name: quay-bridge-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: quay-bridge-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "my-quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecretName": "<NAMESPACE>/<SECRET>",
                  "quayHostname": "https://<QUAY_URL>",
                  "insecureRegistry": false,
                  "whitelistNamespaces": ["default"]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:9149df5ffc83dce6593072b7f72805e07e700b84527e45b4b7486b9c38f70731
          createdAt: 2021-09-15 23:09 UTC
          description: Enhance OCP using Red Hat Quay container registry
          olm.skipRange: '>=3.4.x <3.5.7'
          repository: https://github.com/redhat-cop/quay-openshift-registry-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a connected Quay instance.
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.redhatcop.redhat.io
            version: v1alpha1
        description: "## Overview\n\nThe OpenShift Container Platform contains an
          in cluster image registry for pushing and pulling images that either produced
          within the cluster or are references from external sources. Quay is a highly
          available, enterprise grade image registry with enhanced role based access
          control (RBAC) features. The goal of this operator is to support replicating
          the features found within OpenShift intgrated image registry with Quay.\n\n##
          Functionality\n\nThe following capabilities are enabled through this operator:\n\n*
          Synchronization of OpenShift namespaces as Quay organizations\n    * Creation
          of Robot accounts for each default namespace service account\n    * Creation
          of Secrets for each created Robot Account\n        * Associate Robot Secret
          to Service Account as Mountable and Image Pull Secret\n    * Synchronization
          of OpenShift ImageStreams as Quay Repositories\n* Automatically rewrite
          new Builds making use of ImageStreams to output to Quay\n* Automatically
          import ImageStream tag once build completes\n\n## Prerequisites\n\nThe following
          requirements must be satisfied prior to setting and configuring the integration
          between OpenShift and Quay:\n\n* A Red Hat Quay environment with a user
          with superuser permissions\n* A Red Hat OpenShift Container Platform environment
          (at least version 4.0) with cluster-admin permissions on an OpenShift cluster
          \n*\tOpenShift Command Line tool (CLI)\n\n## Setup and Configuration\n\nThe
          setup process involves steps within both Quay and OpenShift as described
          below:\n\n### Quay Setup\n\nThe integration between OpenShift and Quay makes
          extensive use of the Quay RESTFul API. To allow for other methods aside
          from user credentials to be used for authenticating against the API, an
          application is required to be created. When an application is created, an
          OAuth token is generated which can be used by the operator in OpenShift.
          Applications are created within a Quay organization. While an existing organization
          can be utilized, it is recommended that a dedicated organization be used.\n\n1.\tLogin
          to Quay as a user with superuser access and select the organization for
          which the external application will be configured within.\n2.\tOn the left
          hand side of the page, select **Applications**\n3.\tCreate a new application
          by selecting the **Create New Application** button and entering a name for
          the application\n4.\tAfter the new application has been created, select
          the application that was created\n5.\tOn the lefthand navigation bar, select
          the **Generate Token** button to create a new OAuth2 token\n6.\tSelect all
          of the checkboxes to grant the necessary access that is needed for the integration\n7.\tReview
          the permissions that will be assigned and then select the **Authorize Application**
          button\n8.\tTake note of the generated Access Token as it will be needed
          in the following section\n\n### OpenShift Setup\n\nTo complete the process
          to begin to integrate OpenShift and Quay, several steps are required. Before
          beginning, ensure that you have the OpenShift Command Line tool installed
          and that you are logged into OpenShift as a cluster administrator. \n\n####
          Create A Secret for the Quay OAuth Token\n\nThe Operator will use the previously
          obtained Access Token to communicate with Quay. Store this token within
          OpenShift as a secret.\n\nExecute the following command to create a secret
          called `quay-integration` with a key called `token` containing the access
          token:\n\n```\n$ oc create secret generic quay-integration --from-literal=token=<access_token>\n```\n\nThis
          token will be referenced in the final section.\n\n#### MutatingWebhookConfiguration
          Support\n\nSupport for dynamic interception of API requests that are performed
          as part of OpenShift’s typical build process is facilitated through a [MutatingWebhookConfiguration](https://docs.openshift.com/container-platform/3.11/architecture/additional_concepts/dynamic_admission_controllers.html).
          A MutatingWebhookConfiguration allows for invoking an API running within
          a project on OpenShift when certain API requests are received. In particular,
          to support Quay’s integration with OpenShift, any new Build requests should
          be intercepted so that the output can be modified to target Quay instead
          of OpenShift’s integrated registry. \n\nKubernetes requires that the webhook
          endpoint is secured via SSL using a certificate that makes use of the certificate
          authority for the cluster. Fortunately, OpenShift provides support for generating
          a certificate signed by the cluster. A script is available in the documentation
          called `webhook-create-signed-cert.sh` that automates the process for requesting
          a new certificate.\n\nFirst, create a `Service` for the Operator's webhook
          endpoint:\n```yaml\napiVersion: v1\nkind: Service\nmetadata:\n  labels:\n
          \   name: quay-bridge-operator\n  name: quay-bridge-operator\n  namespace:
          openshift-operators\nspec:\n  ports:\n    - name: https\n      port: 443\n
          \     protocol: TCP\n      targetPort: 8443\n  selector:\n    name: quay-bridge-operator\n
          \ sessionAffinity: None\n  type: ClusterIP\n```\n\nExecute the following
          command to request the certificate:\n\n```\n./webhook-create-signed-cert.sh
          --namespace openshift-operators --secret quay-bridge-operator-webhook-certs
          --service quay-bridge-operator\n```\n\nThe result will place the newly created
          private key and certificate within a secret within the secret specified.
          The secret will be mounted into the appropriate located within the operator
          as declared in the Deployment of the Operator. \n\nOnce the secret has been
          created, focus can shift to the `MutatingWebhookConfiguration` (see below).
          The majority of the content should be in place to apply to the cluster.
          The one section that requires configuration is the caBundle property. This
          refers to the Certificate Authority (CA) for the OpenShift environment.
          The OpenShift CA is available as a _ConfigMap_ within the `kube-system`
          namespace.\n\nExecute the following command to retrieve the CA and format
          the result as a single line so that it can be entered into the MutatingWebhookConfiguration
          resource:\n\n```\noc get configmap -n kube-system extension-apiserver-authentication
          -o=jsonpath='{.data.client-ca-file}' | base64 | tr -d '\\n'\n```\n\nReplace
          the `${CA_BUNDLE}` variable in the following `MutatingWebhookConfiguration`
          YAML and create it:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\nkind:
          MutatingWebhookConfiguration \nmetadata:\n  name: quay-bridge-operator\nwebhooks:\n
          \ - name: quayintegration.redhatcop.redhat.io\n    clientConfig: \n      service:\n
          \       namespace: openshift-operators\n        name: quay-bridge-operator\n
          \       path: \"/admissionwebhook\"\n      caBundle: \"${CA_BUNDLE}\"\n
          \   rules: \n    - operations:  [ \"CREATE\" ]\n      apiGroups: [ \"build.openshift.io\"
          ]\n      apiVersions: [\"v1\" ]\n      resources: [ \"builds\" ]\n    failurePolicy:
          Fail\n    sideEffects: None\n    admissionReviewVersions: [v1beta1]\n    matchPolicy:
          Exact\n    timeoutSeconds: 30  \n```\n\n_Note: Until the operator is running,
          new requests for builds will fail since the webserver the MutatingWebhookConfiguration
          invokes is not available and a proper is response is required in order for
          the object to be persisted in etcd._\n\n#### Create the QuayIntegration
          Custom Resource\n\nFinally, to complete the integration between OpenShift
          and Quay, a `QuayIntegration` custom resource needs to be created.\n\nThe
          following is an example of a basic definition of a `QuayIntegration` resource
          associated from the associated CRD.\n\n```\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          QuayIntegration\nmetadata:\n  name: example-quayintegration\nspec:\n  clusterID:
          openshift\n  credentialsSecretName: <NAMESPACE>/<SECRET>\n  quayHostname:
          https://<QUAY_URL>\n```\n\nThe _clusterID_ is a value which should be unique
          across the entire ecosystem. This value is optional and defaults to `openshift`.\n\nThe
          _credentialsSecretName_ is a NamespacedName value of the secret containing
          the token that was previously created.\n\nNote: If Quay is using self signed
          certificates, the property `insecureRegistry: true`\n\nOrganizations within
          Quay should be created for the related namespaces in OpenShift."
        displayName: Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Source Code
          url: https://github.com/redhat-cop/quay-openshift-registry-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:9149df5ffc83dce6593072b7f72805e07e700b84527e45b4b7486b9c38f70731
        version: 3.5.7
      entries:
      - name: quay-bridge-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: quay-bridge-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:33dc7673155f01728225221768f6b90d87cc7609dcda46b394dba2421c136a23
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.10.15
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:33dc7673155f01728225221768f6b90d87cc7609dcda46b394dba2421c136a23
        version: 3.10.15
      entries:
      - name: quay-bridge-operator.v3.10.15
        version: 3.10.15
      - name: quay-bridge-operator.v3.10.14
        version: 3.10.14
      - name: quay-bridge-operator.v3.10.13
        version: 3.10.13
      - name: quay-bridge-operator.v3.10.12
        version: 3.10.12
      - name: quay-bridge-operator.v3.10.11
        version: 3.10.11
      - name: quay-bridge-operator.v3.10.10
        version: 3.10.10
      - name: quay-bridge-operator.v3.10.9
        version: 3.10.9
      - name: quay-bridge-operator.v3.10.8
        version: 3.10.8
      - name: quay-bridge-operator.v3.10.7
        version: 3.10.7
      - name: quay-bridge-operator.v3.10.6
        version: 3.10.6
      - name: quay-bridge-operator.v3.10.5
        version: 3.10.5
      - name: quay-bridge-operator.v3.10.4
        version: 3.10.4
      - name: quay-bridge-operator.v3.10.3
        version: 3.10.3
      - name: quay-bridge-operator.v3.10.2
        version: 3.10.2
      - name: quay-bridge-operator.v3.10.1
        version: 3.10.1
      - name: quay-bridge-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-bridge-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:511f8471db70e8d5fd9d0b547c73f1416bbac6f0fe6cdeb6c59a9ef8bd845ac2
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.11.13
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:511f8471db70e8d5fd9d0b547c73f1416bbac6f0fe6cdeb6c59a9ef8bd845ac2
        version: 3.11.13
      entries:
      - name: quay-bridge-operator.v3.11.13
        version: 3.11.13
      - name: quay-bridge-operator.v3.11.12
        version: 3.11.12
      - name: quay-bridge-operator.v3.11.11
        version: 3.11.11
      - name: quay-bridge-operator.v3.11.10
        version: 3.11.10
      - name: quay-bridge-operator.v3.11.9
        version: 3.11.9
      - name: quay-bridge-operator.v3.11.8
        version: 3.11.8
      - name: quay-bridge-operator.v3.11.7
        version: 3.11.7
      - name: quay-bridge-operator.v3.11.6
        version: 3.11.6
      - name: quay-bridge-operator.v3.11.5
        version: 3.11.5
      - name: quay-bridge-operator.v3.11.4
        version: 3.11.4
      - name: quay-bridge-operator.v3.11.3
        version: 3.11.3
      - name: quay-bridge-operator.v3.11.2
        version: 3.11.2
      - name: quay-bridge-operator.v3.11.1
        version: 3.11.1
      - name: quay-bridge-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-bridge-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:c6e04de3ad3b8c0ded95146d090e120bffed7f9dc5e74332a183fe1c2e45d16d
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.12.12
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:c6e04de3ad3b8c0ded95146d090e120bffed7f9dc5e74332a183fe1c2e45d16d
        version: 3.12.12
      entries:
      - name: quay-bridge-operator.v3.12.12
        version: 3.12.12
      - name: quay-bridge-operator.v3.12.11
        version: 3.12.11
      - name: quay-bridge-operator.v3.12.10
        version: 3.12.10
      - name: quay-bridge-operator.v3.12.9
        version: 3.12.9
      - name: quay-bridge-operator.v3.12.8
        version: 3.12.8
      - name: quay-bridge-operator.v3.12.7
        version: 3.12.7
      - name: quay-bridge-operator.v3.12.6
        version: 3.12.6
      - name: quay-bridge-operator.v3.12.5
        version: 3.12.5
      - name: quay-bridge-operator.v3.12.4
        version: 3.12.4
      - name: quay-bridge-operator.v3.12.3
        version: 3.12.3
      - name: quay-bridge-operator.v3.12.2
        version: 3.12.2
      - name: quay-bridge-operator.v3.12.1
        version: 3.12.1
      - name: quay-bridge-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-bridge-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:e8ee1944d8e0e9d515a6e63a3e39901c304abd421fcad2abbf9a3c43c7ca6243
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.13.8
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:e8ee1944d8e0e9d515a6e63a3e39901c304abd421fcad2abbf9a3c43c7ca6243
        version: 3.13.8
      entries:
      - name: quay-bridge-operator.v3.13.8
        version: 3.13.8
      - name: quay-bridge-operator.v3.13.7
        version: 3.13.7
      - name: quay-bridge-operator.v3.13.6
        version: 3.13.6
      - name: quay-bridge-operator.v3.13.5
        version: 3.13.5
      - name: quay-bridge-operator.v3.13.4
        version: 3.13.4
      - name: quay-bridge-operator.v3.13.3
        version: 3.13.3
      - name: quay-bridge-operator.v3.13.2
        version: 3.13.2
      - name: quay-bridge-operator.v3.13.1
        version: 3.13.1
      - name: quay-bridge-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-bridge-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b12f0dfb8fdeae3f521e41fcf60ef118bdfd08ad7f68bd87c5494b067b7c7c58
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.14.5
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b12f0dfb8fdeae3f521e41fcf60ef118bdfd08ad7f68bd87c5494b067b7c7c58
        version: 3.14.5
      entries:
      - name: quay-bridge-operator.v3.14.5
        version: 3.14.5
      - name: quay-bridge-operator.v3.14.4
        version: 3.14.4
      - name: quay-bridge-operator.v3.14.3
        version: 3.14.3
      - name: quay-bridge-operator.v3.14.2
        version: 3.14.2
      - name: quay-bridge-operator.v3.14.1
        version: 3.14.1
      - name: quay-bridge-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-bridge-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b2ac4a48c80209d7bebe8b1b2d066571f99a9f47b78604e47a8300d027941b66
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.15.2
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b2ac4a48c80209d7bebe8b1b2d066571f99a9f47b78604e47a8300d027941b66
        version: 3.15.2
      entries:
      - name: quay-bridge-operator.v3.15.2
        version: 3.15.2
      - name: quay-bridge-operator.v3.15.1
        version: 3.15.1
      - name: quay-bridge-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-bridge-operator.v3.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:0ad7af440c6e00fb578ef40e6d600964324fd94ce7835a28d0ec355b7415093f
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.6.10
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:99dc6b5af811bb33892c17334dd4e1a7eddc6982d49a793b1789f505e00f49b1
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:0ad7af440c6e00fb578ef40e6d600964324fd94ce7835a28d0ec355b7415093f
        version: 3.6.10
      entries:
      - name: quay-bridge-operator.v3.6.10
        version: 3.6.10
      - name: quay-bridge-operator.v3.6.9
        version: 3.6.9
      - name: quay-bridge-operator.v3.6.8
        version: 3.6.8
      - name: quay-bridge-operator.v3.6.7
        version: 3.6.7
      - name: quay-bridge-operator.v3.6.6
        version: 3.6.6
      - name: quay-bridge-operator.v3.6.5
        version: 3.6.5
      - name: quay-bridge-operator.v3.6.4
        version: 3.6.4
      - name: quay-bridge-operator.v3.6.2
        version: 3.6.2
      - name: quay-bridge-operator.v3.6.1
        version: 3.6.1
      - name: quay-bridge-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: quay-bridge-operator.v3.7.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:6a0dd80dc48ad0aaa576b3de78b2b265e29a3c10e4996aa6b8f129592fe20f09
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.7.14
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:975f287513039b2244894b9031e590f665d48f9af480dfdc4b85d4f96dd4bc37
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:6a0dd80dc48ad0aaa576b3de78b2b265e29a3c10e4996aa6b8f129592fe20f09
        version: 3.7.14
      entries:
      - name: quay-bridge-operator.v3.7.14
        version: 3.7.14
      - name: quay-bridge-operator.v3.7.13
        version: 3.7.13
      - name: quay-bridge-operator.v3.7.12
        version: 3.7.12
      - name: quay-bridge-operator.v3.7.11
        version: 3.7.11
      - name: quay-bridge-operator.v3.7.10
        version: 3.7.10
      - name: quay-bridge-operator.v3.7.9
        version: 3.7.9
      - name: quay-bridge-operator.v3.7.8
        version: 3.7.8
      - name: quay-bridge-operator.v3.7.7
        version: 3.7.7
      - name: quay-bridge-operator.v3.7.6
        version: 3.7.6
      - name: quay-bridge-operator.v3.7.5
        version: 3.7.5
      - name: quay-bridge-operator.v3.7.4
        version: 3.7.4
      - name: quay-bridge-operator.v3.7.3
        version: 3.7.3
      - name: quay-bridge-operator.v3.7.2
        version: 3.7.2
      - name: quay-bridge-operator.v3.7.1
        version: 3.7.1
      - name: quay-bridge-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: quay-bridge-operator.v3.8.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:f8544a546dd6d209355c3a1cbe35138f6910491fc4584b401634431da4414ec0
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.8.15
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:975f287513039b2244894b9031e590f665d48f9af480dfdc4b85d4f96dd4bc37
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:f8544a546dd6d209355c3a1cbe35138f6910491fc4584b401634431da4414ec0
        version: 3.8.15
      entries:
      - name: quay-bridge-operator.v3.8.15
        version: 3.8.15
      - name: quay-bridge-operator.v3.8.14
        version: 3.8.14
      - name: quay-bridge-operator.v3.8.13
        version: 3.8.13
      - name: quay-bridge-operator.v3.8.12
        version: 3.8.12
      - name: quay-bridge-operator.v3.8.11
        version: 3.8.11
      - name: quay-bridge-operator.v3.8.10
        version: 3.8.10
      - name: quay-bridge-operator.v3.8.9
        version: 3.8.9
      - name: quay-bridge-operator.v3.8.8
        version: 3.8.8
      - name: quay-bridge-operator.v3.8.7
        version: 3.8.7
      - name: quay-bridge-operator.v3.8.6
        version: 3.8.6
      - name: quay-bridge-operator.v3.8.5
        version: 3.8.5
      - name: quay-bridge-operator.v3.8.4
        version: 3.8.4
      - name: quay-bridge-operator.v3.8.3
        version: 3.8.3
      - name: quay-bridge-operator.v3.8.2
        version: 3.8.2
      - name: quay-bridge-operator.v3.8.1
        version: 3.8.1
      - name: quay-bridge-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-bridge-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:875e0b63fdb7d44879650eb2f788ea705d753520841a5d875428f0b3ca70a36c
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.9.16
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:875e0b63fdb7d44879650eb2f788ea705d753520841a5d875428f0b3ca70a36c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 3.9.16
      entries:
      - name: quay-bridge-operator.v3.9.16
        version: 3.9.16
      - name: quay-bridge-operator.v3.9.15
        version: 3.9.15
      - name: quay-bridge-operator.v3.9.14
        version: 3.9.14
      - name: quay-bridge-operator.v3.9.13
        version: 3.9.13
      - name: quay-bridge-operator.v3.9.12
        version: 3.9.12
      - name: quay-bridge-operator.v3.9.11
        version: 3.9.11
      - name: quay-bridge-operator.v3.9.10
        version: 3.9.10
      - name: quay-bridge-operator.v3.9.9
        version: 3.9.9
      - name: quay-bridge-operator.v3.9.8
        version: 3.9.8
      - name: quay-bridge-operator.v3.9.7
        version: 3.9.7
      - name: quay-bridge-operator.v3.9.6
        version: 3.9.6
      - name: quay-bridge-operator.v3.9.5
        version: 3.9.5
      - name: quay-bridge-operator.v3.9.4
        version: 3.9.4
      - name: quay-bridge-operator.v3.9.3
        version: 3.9.3
      - name: quay-bridge-operator.v3.9.2
        version: 3.9.2
      - name: quay-bridge-operator.v3.9.1
        version: 3.9.1
      - name: quay-bridge-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: quay-bridge-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: machine-deletion-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: machine-deletion-remediation.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:420a4d4a8610a3694fd3ef9809c55a4f140ec8cea7f09219647f82a84d1f411b
          createdAt: "2023-10-30 13:59:28"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          olm.skipRange: '>=0.1.0 <0.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:420a4d4a8610a3694fd3ef9809c55a4f140ec8cea7f09219647f82a84d1f411b
        version: 0.2.1
      entries:
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.14-eus
    - currentCSV: machine-deletion-remediation.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:1445e6df8f49672aa729806f0e257043ecc22e19e666bb5a799286c6b878b1dd
          createdAt: "2024-08-11 06:47:20"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.2.0 <0.3.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f4381aaf002b761a2ba614332a2af973f4de5f6a4c84dd96b8808725e561fa89
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:1445e6df8f49672aa729806f0e257043ecc22e19e666bb5a799286c6b878b1dd
        version: 0.3.1
      entries:
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.16-eus
    - currentCSV: machine-deletion-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-29 09:35:10"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.3.1 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.6/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel9-operator@sha256:dbe9095d494ebc7b56982cb1f74329d7a7fd0cbb7544e9f90207eb35de7997dc
        version: 0.4.1
      entries:
      - name: machine-deletion-remediation.v0.4.1
        version: 0.4.1
      - name: machine-deletion-remediation.v0.4.0
        version: 0.4.0
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.18-eus
    - currentCSV: machine-deletion-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:7e1cad02f1e74a1355c168fc53925bd0fbcc55f2535a46f734c8de082eb49400
          createdAt: "2023-04-24 08:22:17"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3af07bfe0fb73b1a7f00487eba302d6da442e742ab4007fc50e52efd4d45dae5
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:7e1cad02f1e74a1355c168fc53925bd0fbcc55f2535a46f734c8de082eb49400
        version: 0.1.0
      entries:
      - name: machine-deletion-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: machine-deletion-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-29 09:35:10"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.3.1 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.6/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel9-operator@sha256:dbe9095d494ebc7b56982cb1f74329d7a7fd0cbb7544e9f90207eb35de7997dc
        version: 0.4.1
      entries:
      - name: machine-deletion-remediation.v0.4.1
        version: 0.4.1
      - name: machine-deletion-remediation.v0.4.0
        version: 0.4.0
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: stable
    defaultChannel: stable
    packageName: machine-deletion-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: sailoperator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sailoperator.v1.28.0-nightly-2025-10-06
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:07ea72de6ff9184f346ee2948228e73cfe4180c8d05fe7ee972c22378ad450ef
          createdAt: "2025-10-06T03:29:39Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0
          - master
          - v1.28-alpha.8de61b9a

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-testing/ztunnel@sha256:38b5902bf36f5163274ce0963dfb1e106000ac639d429f05846448de8454492d
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - quay.io/sail-dev/sail-operator@sha256:07ea72de6ff9184f346ee2948228e73cfe4180c8d05fe7ee972c22378ad450ef
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - gcr.io/istio-testing/install-cni@sha256:9525a3711201bcca04ce2ca9b1576d4094a03779d60db0a36b15a956020bbec1
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-testing/proxyv2@sha256:11cf6a3083a3c1ffa539a21cee30ddc0abe47fb44bb16520cdc2f4141dcfee03
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-testing/pilot@sha256:13e2125870f59f752dc5834b702c2dab28f36edc00dccdeeaf872023018e3df0
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        version: 1.28.0-nightly-2025-10-06
      entries:
      - name: sailoperator.v1.28.0-nightly-2025-10-06
        version: 1.28.0-nightly-2025-10-06
      - name: sailoperator.v1.28.0-nightly-2025-10-05
        version: 1.28.0-nightly-2025-10-05
      - name: sailoperator.v1.28.0-nightly-2025-10-04
        version: 1.28.0-nightly-2025-10-04
      - name: sailoperator.v1.28.0-nightly-2025-10-03
        version: 1.28.0-nightly-2025-10-03
      - name: sailoperator.v1.28.0-nightly-2025-10-02
        version: 1.28.0-nightly-2025-10-02
      - name: sailoperator.v1.28.0-nightly-2025-10-01
        version: 1.28.0-nightly-2025-10-01
      - name: sailoperator.v1.28.0-nightly-2025-09-30
        version: 1.28.0-nightly-2025-09-30
      - name: sailoperator.v1.28.0-nightly-2025-09-29
        version: 1.28.0-nightly-2025-09-29
      - name: sailoperator.v1.28.0-nightly-2025-09-28
        version: 1.28.0-nightly-2025-09-28
      - name: sailoperator.v1.28.0-nightly-2025-09-27
        version: 1.28.0-nightly-2025-09-27
      - name: sailoperator.v1.28.0-nightly-2025-09-26
        version: 1.28.0-nightly-2025-09-26
      - name: sailoperator.v1.28.0-nightly-2025-09-25
        version: 1.28.0-nightly-2025-09-25
      - name: sailoperator.v1.28.0-nightly-2025-09-23
        version: 1.28.0-nightly-2025-09-23
      - name: sailoperator.v1.28.0-nightly-2025-09-22
        version: 1.28.0-nightly-2025-09-22
      name: 1.28-nightly
    - currentCSV: sailoperator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.23.0"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.23.0"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/maistra-dev/sail-operator:0.1-latest
          createdAt: "2024-08-20T11:54:30Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1alpha1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1alpha1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1alpha1
          - description: |-
              RemoteIstio represents a remote Istio Service Mesh deployment consisting of one or more
              remote control plane instances (represented by one or more IstioRevision objects).
            displayName: Remote Istio
            kind: RemoteIstio
            name: remoteistios.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.23.0
          - v1.22.4

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - docker.io/istio/proxyv2:1.22.4
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.16.0
        - quay.io/maistra-dev/sail-operator:0.1-latest
        - docker.io/istio/pilot:1.23.0
        - docker.io/istio/proxyv2:1.23.0
        - docker.io/istio/install-cni:1.22.4
        - docker.io/istio/pilot:1.22.4
        - docker.io/istio/ztunnel:1.23.0
        - docker.io/istio/ztunnel:1.22.4
        - docker.io/istio/install-cni:1.23.0
        version: 0.1.0
      entries:
      - name: sailoperator.v0.1.0
        version: 0.1.0
      - name: sailoperator.v0.1.0-rc.1
        version: 0.1.0-rc.1
      name: candidates
    - currentCSV: sailoperator.v1.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
          createdAt: "2025-09-05T10:12:19Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        version: 1.27.1
      entries:
      - name: sailoperator.v1.27.1
        version: 1.27.1
      - name: sailoperator.v1.27.0
        version: 1.27.0
      - name: sailoperator.v1.26.3
        version: 1.26.3
      - name: sailoperator.v1.26.2
        version: 1.26.2
      - name: sailoperator.v1.26.1
        version: 1.26.1
      - name: sailoperator.v1.26.0
        version: 1.26.0
      - name: sailoperator.v1.25.2
        version: 1.25.2
      - name: sailoperator.v1.25.1
        version: 1.25.1
      - name: sailoperator.v1.25.0
        version: 1.25.0
      name: stable
    - currentCSV: sailoperator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.3"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.3"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator:1.0-latest
          createdAt: "2025-03-07T10:17:30Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.24-latest
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.23-latest
          - v1.23.5
          - v1.23.4
          - v1.23.3
          - v1.23.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/sail-dev/sail-operator:1.0-latest
        - gcr.io/istio-release/ztunnel:1.24.1
        - gcr.io/istio-release/proxyv2:1.24.2
        - gcr.io/istio-release/install-cni:1.24.3
        - gcr.io/istio-release/install-cni:1.23.0
        - gcr.io/istio-release/proxyv2:1.23.5
        - gcr.io/istio-release/proxyv2:1.23.4
        - gcr.io/istio-release/pilot:1.24.2
        - gcr.io/istio-release/pilot:1.23.0
        - gcr.io/istio-release/install-cni:1.24.2
        - gcr.io/istio-release/pilot:1.23.4
        - gcr.io/istio-release/proxyv2:1.23.3
        - gcr.io/istio-release/proxyv2:1.23.0
        - gcr.io/istio-release/pilot:1.23.3
        - gcr.io/istio-release/pilot:1.24.1
        - gcr.io/istio-release/pilot:1.23.5
        - gcr.io/istio-release/proxyv2:1.24.3
        - gcr.io/istio-release/ztunnel:1.24.3
        - gcr.io/istio-release/ztunnel:1.23.4
        - gcr.io/istio-release/install-cni:1.23.5
        - gcr.io/istio-release/ztunnel:1.23.3
        - gcr.io/istio-release/install-cni:1.24.1
        - gcr.io/istio-release/proxyv2:1.24.1
        - gcr.io/istio-release/ztunnel:1.23.0
        - gcr.io/istio-release/install-cni:1.23.4
        - gcr.io/istio-release/ztunnel:1.23.5
        - gcr.io/istio-release/ztunnel:1.24.2
        - gcr.io/istio-release/install-cni:1.23.3
        - gcr.io/istio-release/pilot:1.24.3
        version: 1.0.0
      entries:
      - name: sailoperator.v1.0.0
        version: 1.0.0
      name: stable-1.0
    - currentCSV: sailoperator.v1.25.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.25.2"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.25.2"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator:1.25.2
          createdAt: "2025-05-07T19:44:51Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.25.2'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.25-latest
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0
          - v1.23-latest
          - v1.23.5
          - v1.23.4
          - v1.23.3
          - v1.23.2
          - v1.22-latest
          - v1.22.8
          - v1.22.7
          - v1.22.6
          - v1.22.5
          - v1.21.6

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/ztunnel:1.22.6
        - gcr.io/istio-release/pilot:1.22.8
        - gcr.io/istio-release/install-cni:1.23.5
        - quay.io/sail-dev/sail-operator:1.25.2
        - gcr.io/istio-release/pilot:1.22.6
        - gcr.io/istio-release/install-cni:1.24.2
        - gcr.io/istio-release/install-cni:1.24.3
        - gcr.io/istio-release/install-cni:1.23.3
        - gcr.io/istio-release/ztunnel:1.22.5
        - gcr.io/istio-release/proxyv2:1.23.3
        - gcr.io/istio-release/pilot:1.23.2
        - gcr.io/istio-release/install-cni:1.25.1
        - gcr.io/istio-release/pilot:1.24.3
        - gcr.io/istio-release/ztunnel:1.25.1
        - gcr.io/istio-release/install-cni:1.22.8
        - gcr.io/istio-release/proxyv2:1.22.8
        - gcr.io/istio-release/ztunnel:1.21.6
        - gcr.io/istio-release/pilot:1.23.5
        - gcr.io/istio-release/pilot:1.24.1
        - gcr.io/istio-release/proxyv2:1.21.6
        - gcr.io/istio-release/proxyv2:1.25.1
        - gcr.io/istio-release/ztunnel:1.23.4
        - gcr.io/istio-release/proxyv2:1.23.5
        - gcr.io/istio-release/pilot:1.25.2
        - gcr.io/istio-release/install-cni:1.21.6
        - gcr.io/istio-release/ztunnel:1.23.3
        - gcr.io/istio-release/pilot:1.22.5
        - gcr.io/istio-release/proxyv2:1.25.2
        - gcr.io/istio-release/pilot:1.24.4
        - gcr.io/istio-release/pilot:1.23.3
        - gcr.io/istio-release/pilot:1.25.1
        - gcr.io/istio-release/proxyv2:1.23.4
        - gcr.io/istio-release/install-cni:1.22.7
        - gcr.io/istio-release/ztunnel:1.24.3
        - gcr.io/istio-release/ztunnel:1.24.4
        - gcr.io/istio-release/proxyv2:1.22.6
        - gcr.io/istio-release/proxyv2:1.24.2
        - gcr.io/istio-release/ztunnel:1.23.2
        - gcr.io/istio-release/ztunnel:1.24.2
        - gcr.io/istio-release/ztunnel:1.25.2
        - gcr.io/istio-release/install-cni:1.24.4
        - gcr.io/istio-release/pilot:1.22.7
        - gcr.io/istio-release/install-cni:1.25.2
        - gcr.io/istio-release/ztunnel:1.22.7
        - gcr.io/istio-release/proxyv2:1.22.7
        - gcr.io/istio-release/ztunnel:1.22.8
        - gcr.io/istio-release/pilot:1.24.0
        - gcr.io/istio-release/pilot:1.21.6
        - gcr.io/istio-release/install-cni:1.24.1
        - gcr.io/istio-release/install-cni:1.22.5
        - gcr.io/istio-release/install-cni:1.23.2
        - gcr.io/istio-release/install-cni:1.24.0
        - gcr.io/istio-release/ztunnel:1.23.5
        - gcr.io/istio-release/proxyv2:1.24.0
        - gcr.io/istio-release/install-cni:1.22.6
        - gcr.io/istio-release/proxyv2:1.23.2
        - gcr.io/istio-release/proxyv2:1.24.1
        - gcr.io/istio-release/pilot:1.24.2
        - gcr.io/istio-release/proxyv2:1.22.5
        - gcr.io/istio-release/pilot:1.23.4
        - gcr.io/istio-release/ztunnel:1.24.0
        - gcr.io/istio-release/install-cni:1.23.4
        - gcr.io/istio-release/ztunnel:1.24.1
        - gcr.io/istio-release/proxyv2:1.24.4
        - gcr.io/istio-release/proxyv2:1.24.3
        version: 1.25.2
      entries:
      - name: sailoperator.v1.25.2
        version: 1.25.2
      - name: sailoperator.v1.25.1
        version: 1.25.1
      - name: sailoperator.v1.25.0
        version: 1.25.0
      name: stable-1.25
    - currentCSV: sailoperator.v1.26.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.3"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.3"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:d37258646d76a9f708a69fbaa081a0a95908a2b0c8837a14d7ecbfcf505f9bd4
          createdAt: "2025-08-04T06:23:45Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.26-latest
          - v1.26.3
          - v1.26.2
          - v1.26.1
          - v1.26.0
          - v1.25-latest
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/proxyv2@sha256:fd734e6031566b4fb92be38f0f6bb02fdba6c199c45c2db5dc988bbc4fdee026
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/pilot@sha256:c4e33921e030bb22d3b8fabc693851d52c07a7af79a9534addc2f50005fa462a
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - quay.io/sail-dev/sail-operator@sha256:d37258646d76a9f708a69fbaa081a0a95908a2b0c8837a14d7ecbfcf505f9bd4
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/ztunnel@sha256:2d567893e5b452d5c0b9c43a43efd4f8e28e639a344e28d90498f13858f13d46
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/install-cni@sha256:8e782f5077f63c0a15fd4b9a4fa75f8543033c28ccf2457e25afbf1aefb6a72a
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        version: 1.26.3
      entries:
      - name: sailoperator.v1.26.3
        version: 1.26.3
      - name: sailoperator.v1.26.2
        version: 1.26.2
      - name: sailoperator.v1.26.1
        version: 1.26.1
      - name: sailoperator.v1.26.0
        version: 1.26.0
      name: stable-1.26
    - currentCSV: sailoperator.v1.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
          createdAt: "2025-09-05T10:12:19Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        version: 1.27.1
      entries:
      - name: sailoperator.v1.27.1
        version: 1.27.1
      - name: sailoperator.v1.27.0
        version: 1.27.0
      name: stable-1.27
    defaultChannel: stable
    packageName: sailoperator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-websocket-service
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-websocket-service.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "websocket.loadcore.keysight.com/v1alpha1",
                "kind": "Websocket",
                "metadata": {
                  "name": "websocket-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "websocket": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:26:39Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Websocket
            name: websockets.websocket.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Websocket Operator
        displayName: Keysight Websocket Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Websocket Service
          url: https://keysight-websocket-service.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-websocket-service@sha256:c91da3557ddcb8f052c9caa514d2dbf6c22b4963c080117a9d1859f1458fc404
        version: 0.1.21
      entries:
      - name: keysight-websocket-service.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-websocket-service
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/in-en/products/cloud-pak-for-data
    name: ibm-cpd-datastage-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-datastage-operator.v7.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ds.cpd.ibm.com/v1",
                "kind": "DataStage",
                "metadata": {
                  "name": "datastage"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "fileStorageClass": "managed-nfs-storage",
                  "scaleConfig": "small"
                }
              }
            ]
          capabilities: Seamless Upgrades
          cloudpakId: 49a42b864bb94569bef0188ead948f11
          cloudpakName: IBM DataStage Enterprise Plus Cartridge for IBM Cloud Pak
            for Data
          containerImage: registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
          description: The IBM DataStage Operator is used to deploy IBM DataStage
            services
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-datastage-operator
          olm.skipRange: '>=1.0.0 <7.3.0'
          operatorframework.io/arch.amd64: supported
          operatorframework.io/arch.s390x: supported
          operatorframework.io/os.linux: supported
          operatorframework.io/suggested-namespace: ibm-common-services
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation: [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)
              License: [IBM DataStage Enterprise Cartridge for Cloud Pak for Data](https://ibm.biz/Bdahic),
              [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)'
            kind: DataStage
            name: datastages.ds.cpd.ibm.com
            version: v1
          - description: Documentation [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)
              License [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)
            kind: PXRuntime
            name: pxruntimes.ds.cpd.ibm.com
            version: v1
        description: |-
          # Introduction
          Industry-leading batch and real-time data integration service for data engineers to build trusted pipelines (ETL/ELT) for AI and operational use cases across on-premises and hybrid cloud environments.
          With DataStage, you can design and run data flows that move and transform data anywhere, at any scale.
          No matter how complex your data landscape, DataStage can streamline your data movement costs and increase productivity. DataStage offers:
          <ul><li>A best-in-breed parallel processing engine that enables you to process your data where it resides</li><li>Automated job design</li><li>Simple integration with cloud data lakes, real-time data sources, relational databases, big data, and NoSQL data stores</li></ul>
          The DataStage service uses Cloud Pak for Data platform connections and integration points, with services like Data Virtualization, to simplify the process of connecting to and accessing your data.

          With DataStage, Data Engineers can use the simple user interface to build no-code/low-code data pipelines. The interface offers hundreds of functions and connectors that reduce development time and inconsistencies across pipelines. The interface also makes it easy to collaborate with your peers and control access to specific analytics projects.
          The service also provides automatic workload balancing to provide high performance pipelines that make efficient use of available compute resources.

          ## Details
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=SSQNUZ_4.0/svc-welcome/dsent.html)

          ## Documentation

          [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)

          ## Installing
          To install the operator follow the the installation and configuration instructions within the [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=datastage-installing).

          ## Prerequisites

          - OpenShift Version >= 4.6
          - IBM Cloud Pak for Data >= 4.0.2
          - IBM Cloud Pak for Data Common Core Services (CCS) >= 2.0.2 (The DataStage service automatically installs the CCS if they are not already installed.)

          ### Resources Required

          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=requirements-hardware)

          ## Configuration

          To scale DataStage services, you can specify spec.scaleConfig in the DataStage custom resource to small, medium, or large.

          ## License
          [IBM DataStage Enterprise Cartridge for Cloud Pak for Data](https://ibm.biz/Bdahic)
          [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)

          ## Limitations
          - Only the `x86-64` platform is supported by this operator.
          - See [Limitations and known issues in DataStage](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=issues-datastage)

          ## SecurityContextConstraints Requirements
          Custom SecurityContextConstraints definition:
          ```
          apiVersion: security.openshift.io/v1
          kind: SecurityContextConstraints
          allowHostDirVolumePlugin: false
          allowHostIPC: false
          allowHostNetwork: false
          allowHostPID: false
          allowHostPorts: false
          allowPrivilegeEscalation: true
          allowPrivilegedContainer: false
          allowedCapabilities: null
          defaultAddCapabilities: null
          fsGroup:
            type: MustRunAs
          metadata:
            annotations:
              include.release.openshift.io/self-managed-high-availability: "true"
              kubernetes.io/description: restrict denies access to all host features and requires pods to be run with a UID, and SELinux context that are allocated to the namespace.  This is the most restrictive SCC and it is used by default for authenticated users.
              release.openshift.io/create-only: "true"
            name: restrict
          priority: null
          readOnlyRootFilesystem: false
          requiredDropCapabilities:
          - KILL
          - MKNOD
          - SETUID
          - SETGID
          runAsUser:
            type: MustRunAsRange
          seLinuxContext:
            type: MustRunAs
          supplementalGroups:
            type: RunAsAny
          users: []
          volumes:
          - configMap
          - downwardAPI
          - emptyDir
          - persistentVolumeClaim
          - projected
          - secret
          ```
        displayName: IBM DataStage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DataStage
        - Analytics
        - Cloud Pak
        - Data
        - Cloud Pak for Data
        - IBM
        maintainers:
        - email: cp4d@ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/in-en/products/cloud-pak-for-data
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:046eb99c8f1f37dd4cffa1c6c9a93aea50bb5d10a549aa12c3c90134d03c6d1c
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:4604a076c837600cd4b3e29b31e22470915fe5b6b1191ae9401407f8620f2472
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:5f5faf7d15f46d79eea873c12ff631ed59bc55bc57f1f7aee9fd0a8f35fdd3cf
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:6c569c4687546a03a5f4e127acc13b219ce3efa057fbdfd55e96868e549c70cb
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:8b72cafe9de355634bc3218cc1dc701ffb376a29aecb5cd12a5719d43a1304bd
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:3f8133fcd904befa579b99d3a828fe737faf13ae14e842ed0408eeeee72de5fb
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:7015b410b947338dcafde98b68b95cb4199c08511ae46e733691e8fa835ad988
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:3eecc54f26cd87870c201b5ade1d037d2733ccfda40fc1aedd9a2164e582275d
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:001bf35df08cd397397c3fdd1ddbd7953cdf4bde06965427ae640e5608e60ddf
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:1b39e1d5d7c13451252250ceeb29b91c1e7ef3739fc17baf9958e1bf07c4c2fd
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:ef03bb6ed647d92b9c84d34810376da7049aad88589eb2e35798d82a22f5b994
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:10492010b28f57f8badbc7cadbe15d1447269052b387d878c6d9ecbd2f62923b
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:99fc418fb1a09ee6c45e90dfba47f73ea8e1da1b19f9f570d2a749b68118b7a3
        version: 7.3.0
      entries:
      - name: ibm-cpd-datastage-operator.v7.3.0
        version: 7.3.0
      name: v7.3
    defaultChannel: v7.3
    packageName: ibm-cpd-datastage-operator
    provider:
      name: IBM
      url: https://www.ibm.com/in-en/products/cloud-pak-for-data
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Advanced Micro Devices, Inc.
      provider-url: ""
    name: amd-gpu-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amd-gpu-operator.v1.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "amd.com/v1alpha1",
                "kind": "DeviceConfig",
                "metadata": {
                  "name": "test-deviceconfig",
                  "namespace": "openshift-amd-gpu"
                },
                "spec": {
                  "devicePlugin": {
                    "devicePluginImage": "rocm/k8s-device-plugin:rhubi-latest",
                    "nodeLabellerImage": "rocm/k8s-node-labeller:rhubi-latest"
                  },
                  "driver": {
                    "enable": "true",
                    "version": "6.4.2"
                  },
                  "selector": {
                    "feature.node.kubernetes.io/amd-gpu": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning,Monitoring
          containerImage: docker.io/rocm/gpu-operator@sha256:1583d4ec172d50d4416f4030eb85910ec54633b7e8e061d4bd79b14dbf98d28c
          createdAt: "2025-08-01T01:55:48Z"
          description: |-
            Operator responsible for deploying AMD GPU kernel drivers, device plugin, device test runner and device metrics exporter
            For more information, visit [documentation](https://instinct.docs.amd.com/projects/gpu-operator/en/latest/)
          deviceConfigManagerImage: docker.io/rocm/device-config-manager@sha256:d2896271eacfc664ac7651425f46a85af4f4ef98d713988c2a4fee4a846548ee
          devicePluginImage: docker.io/rocm/k8s-device-plugin@sha256:840c4ea1bf58df60ef3e80e3bf27bbf620a393854e42f58b2224b288767e3dd4
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          metricsExporterImage: docker.io/rocm/device-metrics-exporter@sha256:6c946c6ca10d5c0f40e67b5759759f12d51c45e3b934cb1af912821d5fd76e6e
          nodelabellerImage: docker.io/rocm/k8s-node-labeller@sha256:75f916c5a407e40e9ffdd26cdfc65b7da855710568937e740a969094e2991065
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-amd-gpu
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ROCm/gpu-operator
          support: Advanced Micro Devices, Inc.
          testRunnerImage: docker.io/rocm/test-runner@sha256:98e698263b4df26959341b14b3ddded9df51e504e7c6819d5c7c56c434855e9d
          utilsContainerImage: docker.io/rocm/gpu-operator-utils@sha256:9cbe60dce5b3a2ffd93f7148a363ed5f072a320377d5e5bcbbfb1e05d4959287
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DeviceConfig describes how to enable AMD GPU device
            displayName: DeviceConfig
            kind: DeviceConfig
            name: deviceconfigs.amd.com
            version: v1alpha1
        description: |-
          Operator responsible for deploying AMD GPU kernel drivers, device plugin, device test runner and device metrics exporter
          For more information, visit [documentation](https://instinct.docs.amd.com/projects/gpu-operator/en/latest/)
        displayName: amd-gpu-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AMD
        - GPU
        - AI
        - Deep Learning
        - Hardware
        - Driver
        - Monitoring
        links:
        - name: AMD GPU Operator
          url: https://github.com/ROCm/gpu-operator
        maintainers:
        - email: Yan.Sun3@amd.com
          name: Yan Sun
        - email: farshad.ghodsian@amd.com
          name: Farshad Ghodsian
        - email: shrey.ajmera@amd.com
          name: Shrey Ajmera
        maturity: stable
        provider:
          name: Advanced Micro Devices, Inc.
        relatedImages:
        - docker.io/rocm/gpu-operator@sha256:1583d4ec172d50d4416f4030eb85910ec54633b7e8e061d4bd79b14dbf98d28c
        - docker.io/rocm/device-metrics-exporter@sha256:6c946c6ca10d5c0f40e67b5759759f12d51c45e3b934cb1af912821d5fd76e6e
        - docker.io/rocm/k8s-device-plugin@sha256:840c4ea1bf58df60ef3e80e3bf27bbf620a393854e42f58b2224b288767e3dd4
        - docker.io/rocm/k8s-node-labeller@sha256:75f916c5a407e40e9ffdd26cdfc65b7da855710568937e740a969094e2991065
        - docker.io/rocm/gpu-operator-utils@sha256:9cbe60dce5b3a2ffd93f7148a363ed5f072a320377d5e5bcbbfb1e05d4959287
        - docker.io/rocm/test-runner@sha256:98e698263b4df26959341b14b3ddded9df51e504e7c6819d5c7c56c434855e9d
        - docker.io/rocm/device-config-manager@sha256:d2896271eacfc664ac7651425f46a85af4f4ef98d713988c2a4fee4a846548ee
        version: 1.3.2
      entries:
      - name: amd-gpu-operator.v1.3.2
        version: 1.3.2
      - name: amd-gpu-operator.v1.3.1
        version: 1.3.1
      - name: amd-gpu-operator.v1.2.1
        version: 1.2.1
      - name: amd-gpu-operator.v1.1.1
        version: 1.1.1
      name: alpha
    defaultChannel: alpha
    packageName: amd-gpu-operator
    provider:
      name: Advanced Micro Devices, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-zookeeper-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zookeeper-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:1d36865fe1cac88c9edc63c0e35e42a80456486668de5a575f10cc4b445d47db
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:1d36865fe1cac88c9edc63c0e35e42a80456486668de5a575f10cc4b445d47db
        version: 24.11.1
      entries:
      - name: zookeeper-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: zookeeper-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:6a4729289275e473907a3792c001dbe7dd2dc0498aabc719cf4175d956e3c87f
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:6a4729289275e473907a3792c001dbe7dd2dc0498aabc719cf4175d956e3c87f
        version: 25.3.0
      entries:
      - name: zookeeper-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: zookeeper-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
        version: 25.7.0
      entries:
      - name: zookeeper-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: zookeeper-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
        version: 25.7.0
      entries:
      - name: zookeeper-operator.v25.7.0
        version: 25.7.0
      - name: zookeeper-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-zookeeper-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cluster-kube-descheduler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clusterkubedescheduleroperator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "KubeDescheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-kube-descheduler-operator"
                },
                "spec": {
                  "deschedulingIntervalSeconds": 3600,
                  "managementState": "Managed",
                  "mode": "Predictive",
                  "profiles": ["AffinityAndTaints"]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-cluster-kube-descheduler-operator@sha256:ba0b71ff2a30a069b4a8a8f3c1e0898aaadc6db112e4cc12aff7c77ced7a0405
          createdAt: 2021/10/13
          description: An operator to run descheduler in Openshift cluster.
          olm.skipRange: '>=4.10.0 <4.13.0-202309181427'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-kube-descheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeDescheduler is the Schema for the deschedulers API
            displayName: Kube Descheduler
            kind: KubeDescheduler
            name: kubedeschedulers.operator.openshift.io
            version: v1
        description: |
          The Kube Descheduler Operator provides the ability to evict a running pod so that the pod can be rescheduled onto a more suitable node.

          There are several situations where descheduling can benefit your cluster:

          * Nodes are underutilized or overutilized.
          * Pod and node affinity requirements, such as taints or labels, have changed and the original scheduling decisions are no longer appropriate for certain nodes.
          * Node failure requires pods to be moved.
          * New nodes are added to clusters.

          ## Descheduler Profiles

          Once the operator is installed, you can configure one or more profiles to identify pods to evict. The scheduler will schedule the replacement of the evicted pods.

          The following profiles are available:

          * AffinityAndTaints
          * TopologyAndDuplicates
          * SoftTopologyAndDuplicates
          * LifecycleAndUtilization
          * EvictPodsWithPVC
          * EvictPodsWithLocalStorage

          These profiles are documented in detail in the [descheduler operator README](https://github.com/openshift/cluster-kube-descheduler-operator#profiles).

          ## Additional Parameters

          In addition to the profiles, the following parameters can be configured:

          * `deschedulingIntervalSeconds` - Set the number of seconds between descheduler runs. A value of `0` in this field runs the descheduler once and exits.
          * `profileCustomizations` - Allows certain profile parameters to be tweaked, such as `podLifetime` (see [README](https://github.com/openshift/cluster-kube-descheduler-operator#profile-customizations) for more info).
        displayName: Kube Descheduler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - scheduling
        - descheduler
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/cluster-kube-descheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-cluster-kube-descheduler-operator@sha256:ba0b71ff2a30a069b4a8a8f3c1e0898aaadc6db112e4cc12aff7c77ced7a0405
        - registry.redhat.io/openshift4/ose-descheduler@sha256:45dc69ad93ab50bdf9ce1bb79f6d98f849e320db68af30475b10b7f5497a1b13
        version: 4.13.0-202309181427
      entries:
      - name: clusterkubedescheduleroperator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: "4.13"
    - currentCSV: clusterkubedescheduleroperator.v5.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "KubeDescheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-kube-descheduler-operator"
                },
                "spec": {
                  "deschedulingIntervalSeconds": 3600,
                  "managementState": "Managed",
                  "mode": "Predictive",
                  "profiles": ["AffinityAndTaints"]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/kube-descheduler-operator/kube-descheduler-rhel9-operator@sha256:03404d74da227fdf5805f61657272af2c915ce0014d6a902ef384b7a70bd10e2
          createdAt: 2024/10/30
          description: An operator to run descheduler in Openshift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.12.0 <5.0.2'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-kube-descheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeDescheduler is the Schema for the deschedulers API
            displayName: Kube Descheduler
            kind: KubeDescheduler
            name: kubedeschedulers.operator.openshift.io
            version: v1
        description: |
          The Kube Descheduler Operator provides the ability to evict a running pod so that the pod can be rescheduled onto a more suitable node.

          There are several situations where descheduling can benefit your cluster:

          * Nodes are underutilized or overutilized.
          * Pod and node affinity requirements, such as taints or labels, have changed and the original scheduling decisions are no longer appropriate for certain nodes.
          * Node failure requires pods to be moved.
          * New nodes are added to clusters.

          ## Descheduler Profiles

          Once the operator is installed, you can configure one or more profiles to identify pods to evict. The scheduler will schedule the replacement of the evicted pods.

          The following profiles are available:

          * AffinityAndTaints
          * TopologyAndDuplicates
          * SoftTopologyAndDuplicates
          * LifecycleAndUtilization
          * EvictPodsWithPVC
          * EvictPodsWithLocalStorage

          These profiles are documented in detail in the [descheduler operator README](https://github.com/openshift/cluster-kube-descheduler-operator#profiles).

          ## Additional Parameters

          In addition to the profiles, the following parameters can be configured:

          * `deschedulingIntervalSeconds` - Set the number of seconds between descheduler runs. A value of `0` in this field runs the descheduler once and exits.
          * `profileCustomizations` - Allows certain profile parameters to be tweaked, such as `podLifetime` (see [README](https://github.com/openshift/cluster-kube-descheduler-operator#profile-customizations) for more info).
        displayName: Kube Descheduler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - scheduling
        - descheduler
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/cluster-kube-descheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/kube-descheduler-operator/kube-descheduler-rhel9-operator@sha256:03404d74da227fdf5805f61657272af2c915ce0014d6a902ef384b7a70bd10e2
        - registry.redhat.io/kube-descheduler-operator/descheduler-rhel9@sha256:7ba3f46cbbec2e47edcc255fae405bdbf2b9df3a232fc6809ce65fb1827a3b45
        version: 5.0.2
      entries:
      - name: clusterkubedescheduleroperator.v5.0.2
        version: 5.0.2
      - name: clusterkubedescheduleroperator.4.14.0-202311021650
        version: 4.14.0-202311021650
      - name: clusterkubedescheduleroperator.4.14.0-202310201027
        version: 4.14.0-202310201027
      name: stable
    defaultChannel: stable
    packageName: cluster-kube-descheduler-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Weights & Biases
      provider-url: https://wandb.ai
    name: wandb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: wandb-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.wandb.com/v1",
                "kind": "WeightsAndBiases",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "wandb-operator",
                    "app.kubernetes.io/instance": "weightsandbiases-sample",
                    "app.kubernetes.io/managed-by": "olm",
                    "app.kubernetes.io/name": "weightsandbiases",
                    "app.kubernetes.io/part-of": "wandb-operator"
                  },
                  "name": "weightsandbiases-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning,Developer Tools
          certified: "false"
          containerImage: quay.io/wandb_tools/wandb-k8s-operator:1.0.0
          createdAt: "2024-10-10T14:46:54Z"
          description: Operator for Weights and Biases, enabling seamless integration
            and management of machine learning experiments in Kubernetes environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: wandb-system
          repository: https://github.com/wandb/operator
          support: WeightsAndBiases
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: WeightsAndBiases is a custom resource for managing W&B configurations.
            displayName: Weights And Biases
            kind: WeightsAndBiases
            name: weightsandbiases.apps.wandb.com
            version: v1
        description: |-
          The Weights and Biases Operator simplifies the management of machine learning experiments, tracking, and infrastructure. It integrates with Kubernetes to deploy and manage Weights and Biases instances.
          Features:
          - Experiment Tracking
          - Hyperparameter Tuning
          - Model Management
          - Integration with popular ML frameworks
        displayName: Weights & Biases Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - machine learning
        - experiment tracking
        - ML infrastructure
        - weights and biases
        links:
        - name: Weights & Biases Documentation
          url: https://docs.wandb.ai/
        maintainers:
        - email: abhinav.garg@domain.com
          name: Abhinav Garg
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Weights & Biases
          url: https://wandb.ai
        relatedImages:
        - quay.io/wandb_tools/wandb-k8s-operator:1.0.0
        version: 1.0.0
      entries:
      - name: wandb-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: wandb-operator
    provider:
      name: Weights & Biases
      url: https://wandb.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: keepalived-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keepalived-operator.v1.5.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KeepalivedGroup",
                "metadata": {
                  "name": "keepalivedgroup-workers"
                },
                "spec": {
                  "blacklistRouterIDs": [
                    1,
                    2
                  ],
                  "image": "registry.redhat.io/openshift4/ose-keepalived-ipfailover",
                  "interface": "ens3",
                  "nodeSelector": {
                    "node-role.kubernetes.io/loadbalancer": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Networking, Cloud Provider
          certified: "false"
          containerImage: quay.io/redhat-cop/keepalived-operator@sha256:cd0c2f7df82df1178795e4a18a0942da33adb9670b192a7b7a2b1483a5b4ccf8
          createdAt: "2023-03-11T18:48:22Z"
          description: This operator provides a simple way to create self-hosted load
            balancers using keepalived.
          operatorframework.io/suggested-namespace: keepalived-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/keepalived-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KeepalivedGroup
            name: keepalivedgroups.redhatcop.redhat.io
            version: v1alpha1
        description: "The objective of the keepalived operator is to allow for a way
          to create self-hosted load balancers in an automated way. From a user experience
          point of view the behavior is the same as of when creating [`LoadBalancer`](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          services with a cloud provider able to manage them.\n\nThe keepalived operator
          can be used in all environments that allows nodes to advertise additional
          IPs on their NICs (and at least for now, in networks that allow multicast),
          however it's mainly aimed at supporting LoadBalancer services and ExternalIPs
          on bare metal installations (or other installation environments where a
          cloud provider is not available).\n\nOne possible use of the keepalived
          operator is also to support [OpenShift Ingresses](https://docs.openshift.com/container-platform/4.5/networking/configuring_ingress_cluster_traffic/overview-traffic.html)
          in environments where an external load balancer cannot be provisioned.\n\n##
          How it works\n\nThe keepalived operator will create one or more [VIPs](https://en.wikipedia.org/wiki/Virtual_IP_address)
          (an HA IP that floats between multiple nodes), based on the [`LoadBalancer`](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          services and/or services requesting [`ExternalIPs`](https://kubernetes.io/docs/concepts/services-networking/service/#external-ips).\n\nFor
          `LoadBalancer` services the IPs found at `.Status.LoadBalancer.Ingress[].IP`
          will become VIPs.\n\nFor services requesting a `ExternalIPs`, the IPs found
          at `.Spec.ExternalIPs[]` will become VIPs.\n\nNote that a service can be
          of `LoadBalancer` type and also request `ExternalIPs`, it this case both
          sets of IPs will become VIPs.\n\nDue to a [keepalived](https://www.keepalived.org/manpage.html)
          limitation a single keepalived cluster can manage up to 256 VIP configurations.
          Multiple keepalived clusters can coexists in the same network as long as
          they use different multicast ports [TODO verify this statement].\n\nTo address
          this limitation the `KeepalivedGroup` [CRD](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/)
          has been introduced. This CRD is supposed to be configured by an administrator
          and allows you to specify a node selector to pick on which nodes the keepalived
          pods should be deployed. Here is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          KeepalivedGroup\nmetadata:\n  name: keepalivedgroup-router\nspec:\n  image:
          registry.redhat.io/openshift4/ose-keepalived-ipfailover\n  interface: ens3\n
          \ nodeSelector:\n    node-role.kubernetes.io/loadbalancer: \"\"\n  blacklistRouterIDs:\n
          \ - 1\n  - 2  \n```\n\nThis KeepalivedGroup will be deployed on all the
          nodes with role `loadbalancer`. One must also specify the network device
          on which the VIPs will be exposed, it is assumed that all the nodes have
          the same network device configuration.\n\nServices must be annotated to
          opt-in to being observed by the keepalived operator and to specify which
          KeepalivedGroup they refer to. The annotation looks like this:\n\n`keepalived-operator.redhat-cop.io/keepalivedgroup:
          <keepalivedgroup namespace>/<keepalivedgroup-name>`\n\nThe image used for
          the keepalived containers can be specified with `.Spec.Image` it will default
          to `registry.redhat.io/openshift4/ose-keepalived-ipfailover` if undefined.
          \n\n## Requirements\n\n### Security Context Constraints\n\nEach KeepalivedGroup
          deploys a [daemonset](https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/)
          that requires the [privileged scc](https://docs.openshift.com/container-platform/4.5/authentication/managing-security-context-constraints.html),
          this permission must be given to the `default` service account in the namespace
          where the keepalived group is created by and administrator.\n\n```shell\noc
          adm policy add-scc-to-user privileged -z default -n keepalived-operator\n```\n\n###
          Cluster Network Operator\n\nIn Openshift, use of an external IP address
          is governed by the following fields in the `Network.config.openshift.io`
          CR named `cluster`\n\n* `spec.externalIP.autoAssignCIDRs` defines an IP
          address block used by the load balancer when choosing an external IP address
          for the service. OpenShift supports only a single IP address block for automatic
          assignment.\n\n* `spec.externalIP.policy` defines the permissible IP address
          blocks when manually specifying an IP address. OpenShift does not apply
          policy rules to IP address blocks defined by `spec.externalIP.autoAssignCIDRs`\n\nThe
          following patch can be used to configure the Cluster Network Operator:\n\n```yaml\nspec:\n
          \ externalIP:\n    policy:\n      allowedCIDRs:\n      - ${ALLOWED_CIDR}\n
          \   autoAssignCIDRs:\n      - \"${AUTOASSIGNED_CIDR}\"\n```\n\nHere is an
          example of how to apply the patch:\n\n```shell\nexport ALLOWED_CIDR=\"192.168.131.128/26\"\nexport
          AUTOASSIGNED_CIDR=\"192.168.131.192/26\"\noc patch network cluster -p \"$(envsubst
          < ./network-patch.yaml | yq r -j -)\" --type=merge\n```\nAdditionally, the
          fields can be edited manually via `oc edit Network.config.openshift.io cluster`\n\n##
          Blacklisting router IDs\n\nIf the Keepalived pods are deployed on nodes
          which are in the same network (same broadcast domain to be precise) with
          other keepalived the process, it's necessary to ensure that there is no
          collision between the used routers it.\nFor this purpose it is possible
          to provide a `blacklistRouterIDs` field with a list of black-listed IDs
          that will not be used.\n\n## OpenShift RHV, vSphere, OSP and bare metal
          IPI instructions\n\nWhen IPI is used for RHV, vSphere, OSP or bare metal
          platforms, three keepalived VIPs are deployed. To make sure that keepalived-operator
          can work in these environment we need to discover and blacklist the corresponding
          VRRP router IDs.\n\nTo discover the VRRP router IDs being used, run the
          following command, you can run this command from you laptop:\n\n```shell\npodman
          run quay.io/openshift/origin-baremetal-runtimecfg:4.5 vr-ids <cluster_name>\n```\n\nIf
          you don't know your cluster name, run this command:\n\n```shell\npodman
          run quay.io/openshift/origin-baremetal-runtimecfg:4.5 vr-ids $(oc get cm
          cluster-config-v1 -n kube-system -o jsonpath='{.data.install-config}'| yq
          -r .metadata.name)\n```\n\nThen use these [instructions](#Blacklisting-router-IDs)
          to blacklist those VRRP router IDs.\n\n## Verbatim Configurations\n\nKeepalived
          has dozens of [configurations](https://www.keepalived.org/manpage.html).
          At the early stage of this project it's difficult to tell which one should
          be modeled in the API. Yet, users of this project may still need to use
          them. To account for that there is a way to pass verbatim options both at
          the keepalived group level (which maps to the keepalived config `global_defs`
          section) and at the service level (which maps to the keepalived config `vrrp_instance`
          section).\n\nKeepalivedGroup-level verbatim configurations can be passed
          as in the following example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          KeepalivedGroup\nmetadata:\n  name: keepalivedgroup-router\nspec:\n  interface:
          ens3\n  nodeSelector:\n    node-role.kubernetes.io/loadbalancer: \"\"\n
          \ verbatimConfig:  \n    vrrp_iptables: my-keepalived\n```\n\nthis will
          map to the following `global_defs`:\n\n```\n    global_defs {\n        router_id
          keepalivedgroup-router\n        vrrp_iptables my-keepalived\n    }\n```\n\nService-level
          verbatim configurations can be passed as in the following example:\n\n```yaml\napiVersion:
          v1\nkind: Service\nmetadata:\n  annotations:\n    keepalived-operator.redhat-cop.io/keepalivedgroup:
          keepalived-operator/keepalivedgroup-router\n    keepalived-operator.redhat-cop.io/verbatimconfig:
          '{ \"track_src_ip\": \"\" }'\n```\n\nthis will map to the following `vrrp_instance`
          section\n\n```\n    vrrp_instance openshift-ingress/router-default {\n        interface
          ens3\n        virtual_router_id 1  \n        virtual_ipaddress {\n          192.168.131.129\n
          \       }\n        track_src_ip\n    }\n```\n\n## Metrics collection\n\nEach
          keepalived pod exposes a [Prometheus](https://prometheus.io/) metrics port
          at `9650`. Metrics are collected with [keepalived_exporter](github.com/gen2brain/keepalived_exporter),
          the available metrics are described in the project documentation.\n\nWhen
          a keepalived group is created a [`PodMonitor`](https://github.com/coreos/prometheus-operator/blob/master/Documentation/api.md#podmonitor)
          rule to collect those metrics. All PodMonitor resources created that way
          have the label: `metrics: keepalived`. It is up to you to make sure your
          Prometheus instance watches for those `PodMonitor` rules. Here is an example
          of a fragment of a `Prometheus` CR configured to collect the keepalived
          pod metrics:\n\n```yaml\n  podMonitorSelector:\n    matchLabels:\n      metrics:
          keepalived\n```\n"
        displayName: Keepalived Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadbalancer
        - keepalived
        - ipfailover
        - externalIP
        - self-hosted
        links:
        - name: Keepalived Operator
          url: https://github.com/redhat-cop/keepalived-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/keepalived-operator
        maintainers:
        - email: rspazzol@redhat.com'
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/keepalived-operator@sha256:cd0c2f7df82df1178795e4a18a0942da33adb9670b192a7b7a2b1483a5b4ccf8
        version: 1.5.4
      entries:
      - name: keepalived-operator.v1.5.4
        version: 1.5.4
      - name: keepalived-operator.v1.5.3
        version: 1.5.3
      - name: keepalived-operator.v1.5.2
        version: 1.5.2
      - name: keepalived-operator.v1.5.1
        version: 1.5.1
      - name: keepalived-operator.v1.5.0
        version: 1.5.0
      - name: keepalived-operator.v1.4.2
        version: 1.4.2
      - name: keepalived-operator.v1.4.1
        version: 1.4.1
      - name: keepalived-operator.v1.4.0
        version: 1.4.0
      - name: keepalived-operator.v1.3.4
        version: 1.3.4
      - name: keepalived-operator.v1.3.3
        version: 1.3.3
      - name: keepalived-operator.v1.3.2
        version: 1.3.2
      - name: keepalived-operator.v1.3.1
        version: 1.3.1
      - name: keepalived-operator.v1.3.0
        version: 1.3.0
      - name: keepalived-operator.v1.2.2
        version: 1.2.2
      - name: keepalived-operator.v1.2.1
        version: 1.2.1
      - name: keepalived-operator.v1.2.0
        version: 1.2.0
      - name: keepalived-operator.v1.1.0
        version: 1.1.0
      - name: keepalived-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: keepalived-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel Corporation
      provider-url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
    name: sriov-fec
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sriov-fec.v2.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecClusterConfig",
                "metadata": {
                  "name": "acc100SampleConfig",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "acceleratorSelector": {
                    "deviceID": "someDevice",
                    "driver": "someDriver",
                    "maxVirtualFunctions": 2,
                    "pciAddress": "somePciAddress",
                    "vendorID": "someVendor"
                  },
                  "drainSkip": false,
                  "nodeSelector": {
                    "expectedLabel1": "valueOfExpectedLabel1",
                    "expectedLabelN": "valueOfExpectedLabelN"
                  },
                  "physicalFunction": {
                    "bbDevConfig": {
                      "acc100": {
                        "downlink4G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 0
                        },
                        "downlink5G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 4
                        },
                        "maxQueueSize": 1024,
                        "numVfBundles": 16,
                        "pfMode": true,
                        "uplink4G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 0
                        },
                        "uplink5G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 4
                        }
                      }
                    },
                    "pfDriver": "pci-pf-stub",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  },
                  "priority": 100
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecClusterConfig",
                "metadata": {
                  "name": "n3000SampleConfig",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "acceleratorSelector": {
                    "deviceID": "someDevice",
                    "driver": "someDriver",
                    "maxVirtualFunctions": 2,
                    "pciAddress": "somePciAddress",
                    "vendorID": "someVendor"
                  },
                  "drainSkip": false,
                  "nodeSelector": {
                    "expectedLabel1": "valueOfExpectedLabel1",
                    "expectedLabelN": "valueOfExpectedLabelN"
                  },
                  "physicalFunction": {
                    "bbDevConfig": {
                      "n3000": {
                        "downlink": {
                          "bandwidth": 3,
                          "loadBalance": 128,
                          "queues": {
                            "vf0": 16,
                            "vf1": 16,
                            "vf2": 0,
                            "vf3": 0,
                            "vf4": 0,
                            "vf5": 0,
                            "vf6": 0,
                            "vf7": 0
                          }
                        },
                        "flrTimeout": 610,
                        "networkType": "FPGA_5GNR",
                        "pfMode": true,
                        "uplink": {
                          "bandwidth": 3,
                          "loadBalance": 128,
                          "queues": {
                            "vf0": 16,
                            "vf1": 16,
                            "vf2": 0,
                            "vf3": 0,
                            "vf4": 0,
                            "vf5": 0,
                            "vf6": 0,
                            "vf7": 0
                          }
                        }
                      }
                    },
                    "pfDriver": "pci-pf-stub",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  },
                  "priority": 100
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecNodeConfig",
                "metadata": {
                  "name": "acc100-worker",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "drainSkip": false,
                  "physicalFunctions": [
                    {
                      "bbDevConfig": {
                        "acc100": {
                          "downlink4G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 0
                          },
                          "downlink5G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 4
                          },
                          "maxQueueSize": 1024,
                          "numVfBundles": 16,
                          "pfMode": true,
                          "uplink4G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 0
                          },
                          "uplink5G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 4
                          }
                        }
                      },
                      "pci_addr": "somePciAddress",
                      "pf_driver": "pci-pf-stub",
                      "vf_amount": 2,
                      "vf_driver": "vfio-pci"
                    }
                  ]
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecNodeConfig",
                "metadata": {
                  "name": "n3000-worker",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "drainSkip": false,
                  "physicalFunctions": [
                    {
                      "bbDevConfig": {
                        "n3000": {
                          "downlink": {
                            "bandwidth": 3,
                            "loadBalance": 128,
                            "queues": {
                              "vf0": 16,
                              "vf1": 16,
                              "vf2": 0,
                              "vf3": 0,
                              "vf4": 0,
                              "vf5": 0,
                              "vf6": 0,
                              "vf7": 0
                            }
                          },
                          "flrTimeout": 610,
                          "networkType": "FPGA_5GNR",
                          "pfMode": true,
                          "uplink": {
                            "bandwidth": 3,
                            "loadBalance": 128,
                            "queues": {
                              "vf0": 16,
                              "vf1": 16,
                              "vf2": 0,
                              "vf3": 0,
                              "vf4": 0,
                              "vf5": 0,
                              "vf6": 0,
                              "vf7": 0
                            }
                          }
                        }
                      },
                      "pci_addr": "somePciAddress",
                      "pf_driver": "pci-pf-stub",
                      "vf_amount": 2,
                      "vf_driver": "vfio-pci"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/intel/sriov-fec-operator@sha256:adb6a4c2da6399c8ed40c906f790bce13bfcefae3326a047f6a50592b3f5e485
          createdAt: "2025-01-15"
          description: An operator for Intel Wireless FEC Accelerator to orchestrate
            and manage the resources/devices exposed by a range of Intel's vRAN FEC
            acceleration devices/hardware within the OpenShift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/intel/sriov-fec-operator
          support: Intel Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovFecClusterConfig is the Schema for the sriovfecclusterconfigs
              API
            displayName: SriovFecClusterConfig
            kind: SriovFecClusterConfig
            name: sriovfecclusterconfigs.sriovfec.intel.com
            version: v2
          - description: SriovFecNodeConfig is the Schema for the sriovfecnodeconfigs
              API
            displayName: SriovFecNodeConfig
            kind: SriovFecNodeConfig
            name: sriovfecnodeconfigs.sriovfec.intel.com
            version: v2
          - description: SriovVrbClusterConfig is the Schema for the sriovvrbclusterconfigs
              API
            displayName: SriovVrbClusterConfig
            kind: SriovVrbClusterConfig
            name: sriovvrbclusterconfigs.sriovvrb.intel.com
            version: v1
          - description: SriovVrbNodeConfig is the Schema for the sriovvrbnodeconfigs
              API
            displayName: SriovVrbNodeConfig
            kind: SriovVrbNodeConfig
            name: sriovvrbnodeconfigs.sriovvrb.intel.com
            version: v1
        description: The Intel vRAN Boost family of accelerators ACC100/VRB1/VRB2,
          to offload and accelerate the computing-intensive process of forward error
          correction (FEC) and more for 4G/LTE and 5G technology, freeing up processing
          power. It allows the optimization of data plane performance to reduce total
          cost of ownership while maintaining a high degree of flexibility.  The Intel
          vRAN Boost accelerators ACC100/VRB1/VRB2 plays a key role in accelerating
          5G and network functions virtualization (NFV) adoption for ecosystem partners
          such as telecommunications equipment manufacturers (TEMs) virtual network
          functions (VNF) vendors, system integrators and telcos, to bring scalable
          and high-performance solutions to market. The Intel ACC100 includes a variant
          that is design to be Network Equipment Building System (NEBS)-friendly,
          and features a Root-of-Trust device that helps protect systems from FPGA
          host security exploits.
        displayName: SRIOV-FEC Operator for Intel vRAN Boost Accelerators
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ACC100
        - vRAN
        - ORAN
        links:
        - name: SRIOV-FEC source code
          url: https://github.com/intel/sriov-fec-operator
        maturity: alpha
        provider:
          name: Intel Corporation
          url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
        relatedImages:
        - registry.connect.redhat.com/intel/sriov-fec-daemon@sha256:5b0c60b67bcf5d7b2797deab0d5e3fa079577179f0acc6b0062533584ea0bc23
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1ed3b514bf9af6ac0d386e3739874d1d964e667d54e041c2d67edd769b0009fb
        - registry.connect.redhat.com/intel/sriov-fec-operator@sha256:adb6a4c2da6399c8ed40c906f790bce13bfcefae3326a047f6a50592b3f5e485
        - registry.connect.redhat.com/intel/n3000-labeler@sha256:b417cfcedc998a35491784005cc7eee3d9c647b84eac5ca41cb459076870b8f6
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin@sha256:d0fcb6e9135d9a19ea26851bcdb4ac72c04bd747706fc5123020460e9dbfc21b
        version: 2.11.1
      entries:
      - name: sriov-fec.v2.11.1
        version: 2.11.1
      - name: sriov-fec.v2.11.0
        version: 2.11.0
      - name: sriov-fec.v2.10.0
        version: 2.10.0
      - name: sriov-fec.v2.9.0
        version: 2.9.0
      - name: sriov-fec.v2.8.0
        version: 2.8.0
      - name: sriov-fec.v2.7.2
        version: 2.7.2
      - name: sriov-fec.v2.7.1
        version: 2.7.1
      name: stable
    defaultChannel: stable
    packageName: sriov-fec
    provider:
      name: Intel Corporation
      url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: spyre-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spyre-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreClusterPolicy",
                "metadata": {
                  "name": "spyreclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2",
                      "spyreFilter": ".",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "latest"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "spyre-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "spyre-device-plugin-init",
                      "repository": "icr.io/ibmaiu_internal",
                      "version": "0.1.1"
                    },
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "spyre-exporter",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.2"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "spyre-webhook-validator",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  },
                  "scheduler": {
                    "image": "spyre-scheduler",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  }
                }
              },
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu_internal/spyre-operator@sha256:0f657c98668a8358c37b3fc2ed3f31ec78b7ef7f9a9678e7cd35d383d22ce2da
          createdAt: "2025-07-31T10:32:48Z"
          description: Operator for IBM Spyre devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM Spyre
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["spyrenodestates.spyre.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpyreClusterPolicy is the Schema for the Spyre API
            displayName: Spyre Cluster Policy
            kind: SpyreClusterPolicy
            name: spyreclusterpolicies.spyre.ibm.com
            version: v1alpha1
          - description: SpyreNodeState is the Schema for the SpyreNodeState API
            displayName: Spyre Node State
            kind: SpyreNodeState
            name: spyrenodestates.spyre.ibm.com
            version: v1alpha1
        description: Operator for IBM Spyre devices
        displayName: IBM Spyre Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: IBM Spyre Operator
          url: https://spyre-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: Satyam.Gupta.3@ibm.com
          name: Satyam Gupta
        - email: Rehan.Khan7@ibm.com
          name: Rehan Khan
        - email: Nishan.Acharya@ibm.com
          name: Nishan Acharya
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu_internal/spyre-operator@sha256:0f657c98668a8358c37b3fc2ed3f31ec78b7ef7f9a9678e7cd35d383d22ce2da
        - icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2
        version: 0.1.2
      entries:
      - name: spyre-operator.v0.1.2
        version: 0.1.2
      name: stable-v0.1
    - currentCSV: spyre-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreClusterPolicy",
                "metadata": {
                  "name": "spyreclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb",
                      "spyreFilter": ".",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "latest"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "spyre-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "spyre-device-plugin-init",
                      "repository": "quay.io/ibm-aiu",
                      "version": "0.2.0"
                    },
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "spyre-exporter",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "spyre-webhook-validator",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "scheduler": {
                    "image": "spyre-scheduler",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.1"
                  }
                }
              },
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/ibm-aiu/spyre-operator@sha256:d0cdd449f1d8f0dd9281c4d9a809333db6c4ed201bb14187b6c70ae686a30e99
          createdAt: "2025-09-02T13:25:34Z"
          description: Operator for IBM Spyre devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM Spyre
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["spyrenodestates.spyre.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpyreClusterPolicy is the Schema for the Spyre API
            displayName: Spyre Cluster Policy
            kind: SpyreClusterPolicy
            name: spyreclusterpolicies.spyre.ibm.com
            version: v1alpha1
          - description: SpyreNodeState is the Schema for the SpyreNodeState API
            displayName: Spyre Node State
            kind: SpyreNodeState
            name: spyrenodestates.spyre.ibm.com
            version: v1alpha1
        description: Operator for IBM Spyre devices
        displayName: IBM Spyre Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: IBM Spyre Operator
          url: https://spyre-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: Satyam.Gupta.3@ibm.com
          name: Satyam Gupta
        - email: Rehan.Khan7@ibm.com
          name: Rehan Khan
        - email: Nishan.Acharya@ibm.com
          name: Nishan Acharya
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - quay.io/ibm-aiu/spyre-operator@sha256:d0cdd449f1d8f0dd9281c4d9a809333db6c4ed201bb14187b6c70ae686a30e99
        version: 0.2.1
      entries:
      - name: spyre-operator.v0.2.1
        version: 0.2.1
      name: stable-v0.2
    - currentCSV: spyre-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreClusterPolicy",
                "metadata": {
                  "name": "spyreclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "config": {
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb",
                      "spyreFilter": ".",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "latest"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "spyre-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "spyre-device-plugin-init",
                      "repository": "quay.io/ibm-aiu",
                      "version": "1.0.0"
                    },
                    "repository": "quay.io/ibm-aiu",
                    "version": "1.0.0"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "spyre-exporter",
                    "repository": "quay.io/ibm-aiu",
                    "version": "1.0.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "spyre-webhook-validator",
                    "repository": "quay.io/ibm-aiu",
                    "version": "1.0.0"
                  },
                  "scheduler": {
                    "image": "spyre-scheduler",
                    "repository": "quay.io/ibm-aiu",
                    "version": "1.0.0"
                  }
                }
              },
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/ibm-aiu/spyre-operator@sha256:dd262e8c65f35b12c5d79571e8b5ddd54dd0ee2693f28bea05a6fc02ac1fe6f0
          createdAt: "2025-10-01T00:40:23Z"
          description: Operator for IBM Spyre devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: spyre-operator
          operators.openshift.io/valid-subscription: use with IBM Spyre
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["spyrenodestates.spyre.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpyreClusterPolicy is the Schema for the Spyre API
            displayName: Spyre Cluster Policy
            kind: SpyreClusterPolicy
            name: spyreclusterpolicies.spyre.ibm.com
            version: v1alpha1
          - description: SpyreNodeState is the Schema for the SpyreNodeState API
            displayName: Spyre Node State
            kind: SpyreNodeState
            name: spyrenodestates.spyre.ibm.com
            version: v1alpha1
        description: Operator for IBM Spyre devices
        displayName: IBM Spyre Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: IBM Spyre Operator
          url: https://spyre-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: Satyam.Gupta.3@ibm.com
          name: Satyam Gupta
        - email: Rehan.Khan7@ibm.com
          name: Rehan Khan
        - email: Nishan.Acharya@ibm.com
          name: Nishan Acharya
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - quay.io/ibm-aiu/spyre-operator@sha256:dd262e8c65f35b12c5d79571e8b5ddd54dd0ee2693f28bea05a6fc02ac1fe6f0
        version: 1.0.0
      entries:
      - name: spyre-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: spyre-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hive-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hive-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:e71fc728632622f7ea97357ec361dd2bb73e63c0707e8014ca8ecf8ec4ab93f1
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:e71fc728632622f7ea97357ec361dd2bb73e63c0707e8014ca8ecf8ec4ab93f1
        version: 24.11.1
      entries:
      - name: hive-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: hive-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:5530f5034cd83c2e8b860fce737ac491a2adc6d22246356bee0c79768ed64706
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:5530f5034cd83c2e8b860fce737ac491a2adc6d22246356bee0c79768ed64706
        version: 25.3.0
      entries:
      - name: hive-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: hive-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
        version: 25.7.0
      entries:
      - name: hive-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hive-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
        version: 25.7.0
      entries:
      - name: hive-operator.v25.7.0
        version: 25.7.0
      - name: hive-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hive-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CSI Addons Community
      provider-url: ""
    name: odf-csi-addons-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-csi-addons-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          createdAt: "2025-09-12T07:33:01Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIAddonsNode is the Schema for the csiaddonsnode API
            displayName: CSIAddons Node
            kind: CSIAddonsNode
            name: csiaddonsnodes.csiaddons.openshift.io
            version: v1alpha1
          - description: NetworkFence is the Schema for the networkfences API
            displayName: Network Fence
            kind: NetworkFence
            name: networkfences.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceCronJob is the Schema for the reclaimspacecronjobs
              API
            displayName: Reclaim Space CronJob
            kind: ReclaimSpaceCronJob
            name: reclaimspacecronjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceJob is the Schema for the reclaimspacejobs API
            displayName: Reclaim Space Job
            kind: ReclaimSpaceJob
            name: reclaimspacejobs.csiaddons.openshift.io
            version: v1alpha1
          - description: VolumeReplicationClass is the Schema for the volumereplicationclasses
              API
            displayName: Volume Replication Class
            kind: VolumeReplicationClass
            name: volumereplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplication is the Schema for the volumereplications
              API
            displayName: Volume Replication
            kind: VolumeReplication
            name: volumereplications.replication.storage.openshift.io
            version: v1alpha1
        description: CSI Addons provides the CSI Addons Controller that enables advanced
          storage operations for CSI-drivers.
        displayName: CSI Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: CSI Addons
          url: https://csi-addons.github.io
        maintainers:
        - email: csi-addons@redhat.com
          name: CSI Addons Community
        maturity: alpha
        provider:
          name: CSI Addons Community
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/odf4/odf-csi-addons-rhel9-operator@sha256:8b0c16c21fccca9b674c4d980a793b6cde482b86363b793e173126b030bdcc55
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:9a0008f662b6afa9e2677a8a83bd662d806ad6f4d03fb49bccb78ed487ce4c7a
        version: 4.15.20-rhodf
      entries:
      - name: odf-csi-addons-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odf-csi-addons-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odf-csi-addons-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odf-csi-addons-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odf-csi-addons-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odf-csi-addons-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odf-csi-addons-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odf-csi-addons-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odf-csi-addons-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odf-csi-addons-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odf-csi-addons-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odf-csi-addons-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odf-csi-addons-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odf-csi-addons-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odf-csi-addons-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odf-csi-addons-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odf-csi-addons-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odf-csi-addons-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odf-csi-addons-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odf-csi-addons-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odf-csi-addons-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          createdAt: "2025-09-12T06:32:07Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIAddonsNode is the Schema for the csiaddonsnode API
            displayName: CSIAddons Node
            kind: CSIAddonsNode
            name: csiaddonsnodes.csiaddons.openshift.io
            version: v1alpha1
          - description: NetworkFence is the Schema for the networkfences API
            displayName: Network Fence
            kind: NetworkFence
            name: networkfences.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceCronJob is the Schema for the reclaimspacecronjobs
              API
            displayName: Reclaim Space CronJob
            kind: ReclaimSpaceCronJob
            name: reclaimspacecronjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceJob is the Schema for the reclaimspacejobs API
            displayName: Reclaim Space Job
            kind: ReclaimSpaceJob
            name: reclaimspacejobs.csiaddons.openshift.io
            version: v1alpha1
          - description: VolumeReplicationClass is the Schema for the volumereplicationclasses
              API
            displayName: Volume Replication Class
            kind: VolumeReplicationClass
            name: volumereplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplication is the Schema for the volumereplications
              API
            displayName: Volume Replication
            kind: VolumeReplication
            name: volumereplications.replication.storage.openshift.io
            version: v1alpha1
        description: CSI Addons provides the CSI Addons Controller that enables advanced
          storage operations for CSI-drivers.
        displayName: CSI Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: CSI Addons
          url: https://csi-addons.github.io
        maintainers:
        - email: csi-addons@redhat.com
          name: CSI Addons Community
        maturity: alpha
        provider:
          name: CSI Addons Community
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/odf4/odf-csi-addons-rhel9-operator@sha256:d370f19578d1c466d39d00da82f1f3fe3d9f13d7021ed0c39f759b24b0ed7573
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:0afcb516eb528857887c6ee1a9cba081dc8f6b7d47d9f26a0dbb2dca0cefe07c
        version: 4.16.16-rhodf
      entries:
      - name: odf-csi-addons-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odf-csi-addons-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odf-csi-addons-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odf-csi-addons-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odf-csi-addons-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odf-csi-addons-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odf-csi-addons-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odf-csi-addons-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odf-csi-addons-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odf-csi-addons-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odf-csi-addons-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odf-csi-addons-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odf-csi-addons-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odf-csi-addons-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odf-csi-addons-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odf-csi-addons-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odf-csi-addons-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odf-csi-addons-operator
    provider:
      name: CSI Addons Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp, Inc.
      provider-url: https://www.netapp.com/
    name: trident-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trident-operator.v25.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "trident.netapp.io/v1",
                "kind": "TridentOrchestrator",
                "metadata": {
                  "name": "trident"
                },
                "spec": {
                  "IPv6": false,
                  "debug": true,
                  "nodePrep": [],
                  "imagePullSecrets": [],
                  "imageRegistry": "",
                  "namespace": "trident",
                  "silenceAutosupport": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: docker.io/netapp/trident-operator:25.06.2
          createdAt: "2025-08-03"
          description: Trident Operator, to manage NetApp Trident installations
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          repository: https://github.com/netapp/trident
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Used to deploy Astra Trident.
            displayName: Trident Orchestrator
            kind: TridentOrchestrator
            name: tridentorchestrators.trident.netapp.io
            version: v1
          - description: Automates AWS FSxN backend configuration
            displayName: Trident Configurator
            kind: TridentConfigurator
            name: tridentconfigurators.trident.netapp.io
            version: v1
        description: |
          NetApp Trident is an open source storage provisioner and orchestrator maintained by NetApp. It enables you to create storage volumes for containerized applications managed by Docker and Kubernetes. For full release information, including patch release changes, see https://docs.netapp.com/us-en/trident/trident-rn.html.
        displayName: NetApp Trident
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - NetApp
        - ONTAP
        - Trident
        - storage
        - performance
        - open source
        - Astra
        - Cloud Native
        - AWS
        - Azure
        - GCP
        - Cloud Volumes
        links:
        - name: GitHub Repository
          url: https://github.com/NetApp/trident
        - name: Trident documentation
          url: https://docs.netapp.com/us-en/trident/index.html
        - name: Support policy
          url: https://mysupport.netapp.com/site/info/version-support
        - name: Release Notes
          url: https://docs.netapp.com/us-en/trident/trident-rn.html
        maintainers:
        - email: ng-astra-product@netapp.com
          name: NetApp
        provider:
          name: NetApp, Inc.
          url: https://www.netapp.com/
        relatedImages:
        - docker.io/netapp/trident-operator:25.06.2
        version: 25.6.2
      entries:
      - name: trident-operator.v25.6.2
        version: 25.6.2
      - name: trident-operator.v25.6.1
        version: 25.6.1
      - name: trident-operator.v25.6.0
        version: 25.6.0
      - name: trident-operator.v25.2.1
        version: 25.2.1
      - name: trident-operator.v25.2.0
        version: 25.2.0
      - name: trident-operator.v24.10.1
        version: 24.10.1
      - name: trident-operator.v24.10.0
        version: 24.10.0
      - name: trident-operator.v24.6.0
        version: 24.6.0
      - name: trident-operator.v24.2.0
        version: 24.2.0
      name: stable
    defaultChannel: stable
    packageName: trident-operator
    provider:
      name: NetApp, Inc.
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: multicluster-operators-subscription
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-operators-subscription.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.3.0
          createdAt: "2021-07-19T18:00:00Z"
          description: An operator to subscribe resources from object Bucket, helm
            and github types of channels
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-application:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-channel:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-placementrule:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-deployable:2.3.0
        version: 0.3.0
      entries:
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.3
    - currentCSV: multicluster-operators-subscription.v0.3.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.4.0
          createdAt: "2021-09-29T10:23:00Z"
          description: An operator to subscribe resources from object Bucket, helm
            and github types of channels
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-placementrule:2.4.0
        - quay.io/open-cluster-management/multicloud-integrations:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-deployable:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-application:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-channel:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.4.0
        version: 0.3.6
      entries:
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.4
    - currentCSV: multicluster-operators-subscription.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.5.0
          createdAt: "2022-03-23T18:00:00Z"
          description: An operator for deploying resources from object Bucket, helm
            repo or github repo to multiple managed clusters
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management-io/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: subscription status per cluster per app
            displayName: App Subscription Report
            kind: SubscriptionReport
            name: subscriptionreports.apps.open-cluster-management.io
            version: v1alpha1
          - description: subscription status per package
            displayName: App Subscription Status
            kind: SubscriptionStatus
            name: subscriptionstatuses.apps.open-cluster-management.io
            version: v1alpha1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.5.0
        - quay.io/open-cluster-management/multicloud-integrations:2.5.0
        - quay.io/open-cluster-management/multicluster-operators-application:2.5.0
        - quay.io/open-cluster-management/multicluster-operators-channel:2.5.0
        version: 0.4.0
      entries:
      - name: multicluster-operators-subscription.v0.4.0
        version: 0.4.0
      - name: multicluster-operators-subscription.v0.3.9
        version: 0.3.9
      - name: multicluster-operators-subscription.v0.3.8
        version: 0.3.8
      - name: multicluster-operators-subscription.v0.3.7
        version: 0.3.7
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.5
    - currentCSV: multicluster-operators-subscription.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-operators-subscription:2.6.0
          createdAt: "2022-06-06T18:00:00Z"
          description: An operator for deploying resources from object Bucket, helm
            repo or github repo to multiple managed clusters
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management-io/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: subscription status per cluster per app
            displayName: App Subscription Report
            kind: SubscriptionReport
            name: subscriptionreports.apps.open-cluster-management.io
            version: v1alpha1
          - description: subscription status per package
            displayName: App Subscription Status
            kind: SubscriptionStatus
            name: subscriptionstatuses.apps.open-cluster-management.io
            version: v1alpha1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/multicluster-operators-subscription:2.6.0
        - quay.io/stolostron/multicloud-integrations:2.6.0
        - quay.io/stolostron/multicluster-operators-application:2.6.0
        - quay.io/stolostron/multicluster-operators-channel:2.6.0
        version: 0.4.1
      entries:
      - name: multicluster-operators-subscription.v0.4.1
        version: 0.4.1
      - name: multicluster-operators-subscription.v0.4.0
        version: 0.4.0
      - name: multicluster-operators-subscription.v0.3.9
        version: 0.3.9
      - name: multicluster-operators-subscription.v0.3.8
        version: 0.3.8
      - name: multicluster-operators-subscription.v0.3.7
        version: 0.3.7
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.6
    defaultChannel: release-2.5
    packageName: multicluster-operators-subscription
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.90.0-beta.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"externalDbIP":
            "10.0.2.15","customImageNames": false,"repository": "icr.io/cpopen/turbonomic","tag":"8.17.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
          createdAt: "2023-03-24T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=42.53.0-beta.1 <42.90.0-beta.2'
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints. Turbonomic platform install and configuration
              documentation can be found online here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How do I start?

          Turbonomic platform installation and configuration documentation can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
          Carefully read through the prerequisites before beginning the installation.  Once ready, on the `Details` tab for the Turbonomic Platform Operator, under `Provided APIs`, click `Create instance`.
          For samples of common configurations, refer to sample custom resource YAMLs provided here: https://github.com/turbonomic/t8c-install/tree/master/samples
          Looking to upgrade your current deployment? Refer to documentation here: https://www.ibm.com/docs/en/tarm/latest?topic=version-updating-using-red-hat-openshift-operatorhub

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
        version: 42.90.0-beta.2
      entries:
      - name: t8c-operator.v42.90.0-beta.2
        version: 42.90.0-beta.2
      name: beta
    - currentCSV: t8c-operator.v42.90.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"externalDbIP":
            "10.0.2.15","customImageNames": false,"repository": "icr.io/cpopen/turbonomic","tag":"8.17.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
          createdAt: "2023-03-24T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=42.53.0 <42.90.0'
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints. Turbonomic platform install and configuration
              documentation can be found online here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How do I start?

          Turbonomic platform installation and configuration documentation can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
          Carefully read through the prerequisites before beginning the installation.  Once ready, on the `Details` tab for the Turbonomic Platform Operator, under `Provided APIs`, click `Create instance`.
          For samples of common configurations, refer to sample custom resource YAMLs provided here: https://github.com/turbonomic/t8c-install/tree/master/samples
          Looking to upgrade your current deployment? Refer to documentation here: https://www.ibm.com/docs/en/tarm/latest?topic=version-updating-using-red-hat-openshift-operatorhub

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
        version: 42.90.0
      entries:
      - name: t8c-operator.v42.90.0
        version: 42.90.0
      name: stable
    defaultChannel: stable
    packageName: t8c-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: https://stakater.com
    name: tenant-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tenant-operator.v0.12.66
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: =0.12.65
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:9f806a065b25229741dac8437a32ec45a99acecf4e6bcbe6e7d1fa4192bc560d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a35b96eacb4e46114b39354f66429f8f05df09c55c98c9416b4ffa2768310ef7
        version: 0.12.66
      entries:
      - name: tenant-operator.v0.12.66
        version: 0.12.66
      name: release-0.12
    - currentCSV: tenant-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2024-10-16T08:15:54Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.0
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:feb71a5045e621f8a8cbfe7e41831100b6b85af9812138071b6217cc4396860a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a35b96eacb4e46114b39354f66429f8f05df09c55c98c9416b4ffa2768310ef7
        version: 1.0.0
      entries:
      - name: tenant-operator.v1.0.0
        version: 1.0.0
      name: release-1.0
    - currentCSV: tenant-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-01-27T13:23:27Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.0
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:f412b6d05038a185f5fe35f9c917388194a5ea66afb538a513423c24390983ce
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a35b96eacb4e46114b39354f66429f8f05df09c55c98c9416b4ffa2768310ef7
        version: 1.1.0
      entries:
      - name: tenant-operator.v1.1.0
        version: 1.1.0
      name: release-1.1
    - currentCSV: tenant-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-03-19T13:30:50Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.2.1
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a35b96eacb4e46114b39354f66429f8f05df09c55c98c9416b4ffa2768310ef7
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:a899d60b31efd5cf2358efcd48737b1c4f4a0356ab6c0abe7a5f62ea825a52cd
        version: 1.2.1
      entries:
      - name: tenant-operator.v1.2.1
        version: 1.2.1
      name: release-1.2
    - currentCSV: tenant-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-06-11T15:25:04Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.3.0
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:e88aa5b008ece406bca35ec22fdc87c842544049d613d4d81c8bc5dd2fb50812
        version: 1.3.0
      entries:
      - name: tenant-operator.v1.3.0
        version: 1.3.0
      name: release-1.3
    - currentCSV: tenant-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-09-08T08:14:28Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.4.2
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:7ee14c0974d1eda849d2c199a56786b0ae777ef6569085be917449f1998b1158
        version: 1.4.2
      entries:
      - name: tenant-operator.v1.4.2
        version: 1.4.2
      name: release-1.4
    defaultChannel: release-1.4
    packageName: tenant-operator
    provider:
      name: Stakater
      url: https://stakater.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.0.4
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
        version: 1.0.4
      entries:
      - name: datadog-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      name: stable
    defaultChannel: stable
    packageName: datadog-operator-certified-rhmp
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sfn-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sfn-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sfn.services.k8s.aws/v1alpha1",
                "kind": "Activity",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sfn.services.k8s.aws/v1alpha1",
                "kind": "StateMachine",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sfn-controller:1.1.1
          createdAt: "2025-09-19T18:41:33Z"
          description: AWS Step Functions controller is a service controller for managing
            Step Functions resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Activity represents the state of an AWS sfn Activity resource.
            displayName: Activity
            kind: Activity
            name: activities.sfn.services.k8s.aws
            version: v1alpha1
          - description: StateMachine represents the state of an AWS sfn StateMachine
              resource.
            displayName: StateMachine
            kind: StateMachine
            name: statemachines.sfn.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Step Functions resources in AWS from within your Kubernetes cluster.

          **About Amazon Step FUnctions**

          Step Functions is a serverless orchestration service that lets you combine AWS Lambda functions and other AWS services to build business-critical applications. Through Step Functions' graphical console, you see your application’s workflow as a series of event-driven steps.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Step Functions
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sfn
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Step Functions Developer Resources
          url: https://aws.amazon.com/step-functions/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: step functions maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sfn-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-sfn-controller.v1.1.1
        version: 1.1.1
      - name: ack-sfn-controller.v1.1.0
        version: 1.1.0
      - name: ack-sfn-controller.v1.0.29
        version: 1.0.29
      - name: ack-sfn-controller.v1.0.28
        version: 1.0.28
      - name: ack-sfn-controller.v1.0.27
        version: 1.0.27
      - name: ack-sfn-controller.v1.0.26
        version: 1.0.26
      - name: ack-sfn-controller.v1.0.25
        version: 1.0.25
      - name: ack-sfn-controller.v1.0.24
        version: 1.0.24
      - name: ack-sfn-controller.v1.0.23
        version: 1.0.23
      - name: ack-sfn-controller.v1.0.22
        version: 1.0.22
      - name: ack-sfn-controller.v1.0.21
        version: 1.0.21
      - name: ack-sfn-controller.v1.0.20
        version: 1.0.20
      - name: ack-sfn-controller.v1.0.19
        version: 1.0.19
      - name: ack-sfn-controller.v1.0.18
        version: 1.0.18
      - name: ack-sfn-controller.v1.0.17
        version: 1.0.17
      - name: ack-sfn-controller.v1.0.16
        version: 1.0.16
      - name: ack-sfn-controller.v1.0.15
        version: 1.0.15
      - name: ack-sfn-controller.v1.0.14
        version: 1.0.14
      - name: ack-sfn-controller.v1.0.13
        version: 1.0.13
      - name: ack-sfn-controller.v1.0.11
        version: 1.0.11
      - name: ack-sfn-controller.v1.0.10
        version: 1.0.10
      - name: ack-sfn-controller.v1.0.9
        version: 1.0.9
      - name: ack-sfn-controller.v1.0.8
        version: 1.0.8
      - name: ack-sfn-controller.v1.0.6
        version: 1.0.6
      - name: ack-sfn-controller.v1.0.5
        version: 1.0.5
      - name: ack-sfn-controller.v1.0.2
        version: 1.0.2
      - name: ack-sfn-controller.v1.0.1
        version: 1.0.1
      - name: ack-sfn-controller.v1.0.0
        version: 1.0.0
      - name: ack-sfn-controller.v0.1.3
        version: 0.1.3
      - name: ack-sfn-controller.v0.1.2
        version: 0.1.2
      - name: ack-sfn-controller.v0.1.1
        version: 0.1.1
      - name: ack-sfn-controller.v0.1.0
        version: 0.1.0
      - name: ack-sfn-controller.v0.0.19
        version: 0.0.19
      - name: ack-sfn-controller.v0.0.18
        version: 0.0.18
      - name: ack-sfn-controller.v0.0.17
        version: 0.0.17
      - name: ack-sfn-controller.v0.0.16
        version: 0.0.16
      - name: ack-sfn-controller.v0.0.15
        version: 0.0.15
      - name: ack-sfn-controller.v0.0.14
        version: 0.0.14
      - name: ack-sfn-controller.v0.0.13
        version: 0.0.13
      - name: ack-sfn-controller.v0.0.12
        version: 0.0.12
      - name: ack-sfn-controller.v0.0.11
        version: 0.0.11
      - name: ack-sfn-controller.v0.0.10
        version: 0.0.10
      - name: ack-sfn-controller.v0.0.9
        version: 0.0.9
      - name: ack-sfn-controller.v0.0.8
        version: 0.0.8
      - name: ack-sfn-controller.v0.0.7
        version: 0.0.7
      - name: ack-sfn-controller.v0.0.6
        version: 0.0.6
      - name: ack-sfn-controller.v0.0.5
        version: 0.0.5
      name: alpha
    defaultChannel: alpha
    packageName: ack-sfn-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeVirt project
      provider-url: ""
    name: community-kubevirt-hyperconverged
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubevirt-hyperconverged-operator.v1.10.7
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployTektonTaskResources":true,"deployVmConsoleProxy":true},"tektonPipelines":{"namespace":"kubevirt"},"tektonTasks":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:0ee30557d5d5a42dec19121e70ed77fbdacd8932f55b11e2c7c1a3d5ad5df4aa
          createdAt: "2024-02-09 14:34:30"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:38293a1050bce9c7cb786cbe8719be1edb91a922a6c583eedb61e56729ed4c72
        - quay.io/kubevirt/virt-controller@sha256:0789fafed2913b35a771e3db882748502b3250be04ece86d97f30201779b4e54
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:e79027973b09aac75860f267c7d7f830599978c38a081001f404b9a1c3d2990f
        - quay.io/kubevirt/mtq-lock-server@sha256:9614df0d5af7c998bcceaf73b548f2e3b29ff8ee4ca64c32613bfbdd6e452f6c
        - quay.io/kubevirt/cdi-uploadproxy@sha256:551221d79902a5053d1c734b81163d69f087217e2ac13c49bdf6900336ef0786
        - quay.io/kubevirt/virt-operator@sha256:cbca7f93afbb18a840f7e763e4c03aeeb1fd97967f381c12c443cc179b51da29
        - quay.io/kubevirt/hostpath-provisioner@sha256:eae30f37b3e837478a4500144b7bc8eb4f5ddf15f49fa7354bc9d5e240ea8aea
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5ff5392cb762a41ee9bc4767e3a693249ae08b46a4322f623f24ad6b63dd06b5
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:57573a24923e5588bca6bc337a8b2b08406c5b77583974365d2cf063c0dd5d06
        - quay.io/kubevirt/cni-default-plugins@sha256:825e3f9fec1996c54a52cec806154945b38f76476b160d554c36e38dfffe5e61
        - quay.io/kubevirt/kubevirt-template-validator@sha256:a148b6e812b70f326b48d0439b1e44a967a6e1f8cf2f98b1a08e9294a93674eb
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:f64ab5981d9ddef1662dd0a05888783f5dbadda8ede596585ba80bfc97a0f51e
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:f7567bea7f24184da06c3ed87b396c804342fcfd89f54989622780ce0b7b0724
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:3fbcc32bd4e4d15bd93c96def784a229cd84cca27942bf4858b581f31c97ee02
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4dc0b87ccd69f9865b89234d8555d3a614ab0a16ed94a3016ffd27f8106132ce
        - quay.io/kubevirt/virtio-container-disk@sha256:f8e71a3a930ce95d646d80862d5ff6f9d3ed7ef2541dd2e2acec0bce26274734
        - quay.io/kubevirt/virt-api@sha256:707003b221496b4432da2f507d1e36e528b45888b5d321e06d460f0678da44ae
        - quay.io/kubevirt/virt-artifacts-server:1.10.7-unstable
        - quay.io/kubevirt/virt-handler@sha256:138dfda5fea8622f3da0d6413fe214fef80c2fd6a6f9533592a0dbfa7e1865b5
        - quay.io/kubevirt/hostpath-csi-driver@sha256:0ca8b50b0adb20afd4e20da9974200f3290b6d51aa7c1129de6e1b953bf26594
        - quay.io/kubevirt/mtq-controller@sha256:96dd73ea0d7847d9f6322192559747bd85655e25ee25482d112e6e3c6fef21b2
        - quay.io/kubevirt/libguestfs-tools@sha256:b7e248be783691daad1326cf41641d5042be80c5c057ebc5315ca7a1619b0581
        - quay.io/kubevirt/cdi-apiserver@sha256:e9e39408413b1478d2e98eba68913f9e20c93000558b190b47de73bdfd1d9ac4
        - quay.io/kubevirt/ovs-cni-plugin@sha256:08f72edf2bef876bba0b0f5513d30225304ad5e7ad6912a61c083664acdb99ff
        - quay.io/kubevirt/macvtap-cni@sha256:434420511e09b2b5ede785a2c9062b6658ffbc26fbdd4629ce06110f9039c600
        - quay.io/kubevirt/cdi-importer@sha256:3143bbc67cdc6267eb48b7eaac664b8551ac4c11401dfbf4921efd3f233e6ce9
        - ghcr.io/kubevirt/kubesecondarydns@sha256:e87e829380a1e576384145f78ccaa885ba1d5690d5de7d0b73d40cfb804ea24d
        - quay.io/kubevirt/virt-exportproxy@sha256:f14444b0200a85efb4b3c176caefe70aabba7d295e33d2af14027a59ce297e24
        - quay.io/kubevirt/cdi-operator@sha256:6c63521d835578b0fbb77d3145b648f95ab0932d5d1b36ff2068ed8fcd91bc5a
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:16ecabb1d733aa17084cd16d88963c5255d04f389e0f4d41941d67e8be973594
        - quay.io/kubevirt/virt-launcher@sha256:4c5fce3de2e2589197de72fb0c9436490ea318aca952c05a622c43e067023f35
        - quay.io/kubevirt/bridge-marker@sha256:bba066e3b5ff3fb8c5e20861fe8abe51e3c9b50ad6ce3b2616af9cb5479a06d0
        - quay.io/kubevirt/cdi-uploadserver@sha256:30f1827d3696cf996b081c22c3267ca78e7219c872fdb54950198fa54359f6ee
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - quay.io/kubevirt/cdi-controller@sha256:27c47883a08226f83757971d3adafb0cd9bcb26e58fbcf7208236070e0adf37e
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:0ee30557d5d5a42dec19121e70ed77fbdacd8932f55b11e2c7c1a3d5ad5df4aa
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:86c10c122767661e31ee26a9e9de7ea606c2c1168b32793b4ea567ede2ae3b21
        - quay.io/kubevirt/virt-exportserver@sha256:73311f79a9c71007f8572b3cc40cd6f6da404c7ef0a9c6509fb717d979546582
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f6717ce72a2615c7fbc746b4068f788e78579c54c43b8716e5ce650d97af2df1
        - quay.io/kubevirt/kubemacpool@sha256:3214020c21204c189cdf2d767beb8efb61ee3ef65dd8de7f0f1edc61e13e8adf
        - quay.io/kubevirt/ssp-operator@sha256:49cac00844f091d6e80301573e00235d685415185ec29c89b4bd7361f938711e
        - quay.io/kubevirt/mtq-operator@sha256:ef5c4bb1fa72f9e6e75d5585ca352f7563720389e6c9876aa8b07d12d1be6194
        - quay.io/kubevirt/cdi-cloner@sha256:9d31b14f23259398c5bac636f5ead13ad0afd6fe8eeab4499e8e047b4d85074f
        version: 1.10.7
      entries:
      - name: kubevirt-hyperconverged-operator.v1.10.7
        version: 1.10.7
      name: 1.10.7
    - currentCSV: kubevirt-hyperconverged-operator.v1.11.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployTektonTaskResources":true,"deployVmConsoleProxy":true},"tektonPipelines":{"namespace":"kubevirt"},"tektonTasks":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:dd6538e720f18b723b0807037ca7cc1014cba09e57b20706e10e6eb641ac6acc
          createdAt: "2024-03-06 15:11:00"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: quay.io/kubevirt/must-gather
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - quay.io/kubevirt/aaq-server@sha256:0c6b01c48030e03e99b3ff2b319658c2bf2b2001f4e3e030ba70d1143b25717a
        - quay.io/kubevirt/virt-exportserver@sha256:f8473fc824f8bac5f21c64b71cdb4227de28f9bb25c049932f1259216b874a86
        - quay.io/kubevirt/cdi-importer@sha256:7320dce01f4628363992997b592200b4804daf15a9f237cd50a4a6acab7798e8
        - quay.io/kubevirt/virt-artifacts-server@sha256:22fa3218deca38dc98fa0e905df16173a47ccec85707eb91404a367da6480817
        - quay.io/kubevirt/virt-operator@sha256:0e6d56ea0ebdf9bb832f29ab0dc4603f0f619474bb8bc02a01e0a06e9d8d6f4a
        - quay.io/kubevirt/mtq-operator@sha256:90c7608b0ba085582a65ccdc11d089049c5e6b0438d018dd9dc76eb3490383c7
        - quay.io/kubevirt/cni-default-plugins@sha256:c884d6d08f8c0db98964f1eb3877b44ade41fa106083802a9914775df17d5291
        - quay.io/kubevirt/bridge-marker@sha256:bba066e3b5ff3fb8c5e20861fe8abe51e3c9b50ad6ce3b2616af9cb5479a06d0
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:dd6538e720f18b723b0807037ca7cc1014cba09e57b20706e10e6eb641ac6acc
        - quay.io/kubevirt/virt-handler@sha256:238550bb263608ba334b44a15647cbf875482ccfb46049713c14af5d5f0cc788
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:f1bd6ee18c4021c1c94f29edfab89b49b6a4d1b800936c19dbef2d75f8202f2d
        - quay.io/kubevirt/kubemacpool@sha256:cf8daa57ae6603b776d3af512331b143fa03bc2f4b72f28420fddcf5e4156d0a
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:4ed1f37116cbaf883edd8783201d091fbbe121fcb84e7184ee531d3b1abd77ab
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - registry.k8s.io/sig-storage/livenessprobe@sha256:5baeb4a6d7d517434292758928bb33efc6397368cbb48c8a4cf29496abf4e987
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:bf17da674ed6dac95aa4fd7637173fd4e026ebb4523b66f529c0eff394a25bcc
        - quay.io/kubevirt/cdi-uploadserver@sha256:16d257e575b2bd504f148f63b57dc181fa6058d33a11ada64888183c3e4d4863
        - quay.io/kubevirt/virt-exportproxy@sha256:942c548d4848a87945eba51683002797eb4e1f5c463a1873115b2ec0e1e6e169
        - quay.io/kubevirt/ovs-cni-plugin@sha256:e16ac74343da21abb8fb668ce71e728053d00503a992dae2164b9e94a280113e
        - quay.io/kubevirt/hostpath-csi-driver@sha256:8d84b9a33e0ec3749aece7c5f9879d1e0d248e8028ab3f776528a044d387033c
        - ghcr.io/kubevirt/kubesecondarydns@sha256:e87e829380a1e576384145f78ccaa885ba1d5690d5de7d0b73d40cfb804ea24d
        - quay.io/kubevirt/mtq-controller@sha256:ae7dd07e4018c63414b021eba34aa4776a6c7d6ff4fcdf9fd2d7ca30aae98a29
        - quay.io/kubevirt/cdi-apiserver@sha256:95f6e71b36883abcb9ec463cfea2332f5490ec9eb3e4868bdf0a4da769279094
        - quay.io/kubevirt/cdi-controller@sha256:606de89e68be2ac911466e6713345932bda496c1e4f20e34ef5efa270456c9fd
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:dd2948ca8e62d17eafba40bdcaad8367a97f092fd3b8d9849b2d62aec3152a72
        - quay.io/kubevirt/aaq-operator@sha256:590d24742401eeff3fb6a9c8ed977c63442f8f0a0505509fae69aa1bf8aa84ae
        - quay.io/kubevirt/macvtap-cni@sha256:850b89343ace7c7ea6b18dd8e11964613974e9d1f7377af03854d407fb15230a
        - quay.io/kubevirt/virtio-container-disk@sha256:f8e71a3a930ce95d646d80862d5ff6f9d3ed7ef2541dd2e2acec0bce26274734
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:3fbcc32bd4e4d15bd93c96def784a229cd84cca27942bf4858b581f31c97ee02
        - quay.io/kubevirt/virt-controller@sha256:4f576e30b83db7b9f96dde77896898bde87db3517963c4227634d0c3e96bb86d
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:70efc16ab2bfec04f73ccc292064d080870659cfd47c2e0c9017ff7b6fd78863
        - quay.io/kubevirt/cdi-operator@sha256:8a27e1450c3bf8c1d8924e74470a62419d02ed2202924b082a74e2414b1fc21a
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:c53535af8a7f7e3164609838c4b191b42b2d81238d75c1b2a2b582ada62a9780
        - quay.io/kubevirt/libguestfs-tools@sha256:4971af1ee42959d366a3db6129ab684d11064e00bbd3c2b36324ad14e90cb903
        - quay.io/kubevirt/cdi-cloner@sha256:b22985af0ee7b1ce4e1dedded08c853d90d9e7209f98c5d470897bbefce2a063
        - quay.io/kubevirt/ssp-operator@sha256:d57ef05fea09c4690dbe728ebe6a39d9df1a3574dfa7d18346bf3fe66afed702
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:57573a24923e5588bca6bc337a8b2b08406c5b77583974365d2cf063c0dd5d06
        - quay.io/kubevirt/virt-launcher@sha256:4fbfd76561667e98fafc75c0b804f77df251f96471d1121d69f149cf8fb9a4d8
        - quay.io/kubevirt/kubevirt-template-validator@sha256:f6db539fe4d4b752644c6bc8a308f5a80ae09a71da5c5d535114dc4c3cf18e0a
        - quay.io/kubevirt/aaq-controller@sha256:cb2fa52ae3f6a1a99b2de77b58c5db54c43ba8ee86c6540b4844c901180ce611
        - quay.io/kubevirt/virt-api@sha256:735662c824da3d6069624d40674163d758c2256f8043cf6f02f505c3a724afa3
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:e2def4213ec0657e72eb790ae8a115511d5b8f164a62d3568d2f1bff189917e8
        - quay.io/kubevirt/mtq-lock-server@sha256:8e686107dd9990dab1a3d90ee1068541938f673bf54ea9e1308b2ed7218546b0
        - quay.io/kubevirt/hostpath-provisioner@sha256:adb1d33e95e8991dc47107926172469b5ecbeace509391ce021ca42d305a41d3
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:38293a1050bce9c7cb786cbe8719be1edb91a922a6c583eedb61e56729ed4c72
        - quay.io/kubevirt/cdi-uploadproxy@sha256:b89104ef5513d63cab43b34ac52f752c3d935cb986d5527d920c2c31d3b6f239
        version: 1.11.0
      entries:
      - name: kubevirt-hyperconverged-operator.v1.11.0
        version: 1.11.0
      name: 1.11.0
    - currentCSV: kubevirt-hyperconverged-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:7afeef450521fb7bbc918e874d9297dc3762cdeb28db0bef96d66d48b8752a9d
          createdAt: "2024-08-07 15:33:27"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: quay.io/kubevirt/must-gather
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:7afeef450521fb7bbc918e874d9297dc3762cdeb28db0bef96d66d48b8752a9d
        - quay.io/kubevirt/virt-artifacts-server@sha256:cc3e1e17bc3fa2a895f934210c9106bee01c3f9bf001f988d920eb58628e347b
        - quay.io/kubevirt/virt-operator@sha256:1ecc62c25f869e17027c930150d5c9b52b72d7e1fb434376bc0be0d886952b2e
        - quay.io/kubevirt/aaq-operator@sha256:002513b78dfd795f5195534b651eace136ead3d73c8ac8cae145607e79c8a8ba
        - quay.io/kubevirt/hostpath-csi-driver@sha256:b7ccc469ad7e640fed2e98b50eae57a64e81c7b9862a799b7a9892d14ade8937
        - quay.io/kubevirt/virt-controller@sha256:3471b00eb9f4774ae33867c7d0b83e1d26a065a2bd19d305b5d3946ab665b5e9
        - quay.io/kubevirt/aaq-controller@sha256:170abd7a75397ddce07ba0d64886ead25183c49cacbccf0d191a4dd37ca013fd
        - quay.io/kubevirt/ssp-operator@sha256:a15a53d0f3ef530fe469acf3b7673486011605d0021d4dbe29af94096b3b65ee
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:413a6c229ebb608868e8b2c7c06b251c6b1d43c07af12a76b893292cbed8bef9
        - quay.io/kubevirt/ovs-cni-plugin@sha256:e16ac74343da21abb8fb668ce71e728053d00503a992dae2164b9e94a280113e
        - quay.io/kubevirt/kubevirt-template-validator@sha256:8f7c389a7cd7ef34f2edc84245bc1d0ebcf13e86da5e0b8893f56733c482b10d
        - quay.io/kubevirt/hostpath-provisioner@sha256:2b93b6c4ce0307a8f0dbdae2226da4acb1bae781788fbde7f0c32d6ad6421aa0
        - quay.io/kubevirt/mtq-lock-server@sha256:8b5189518ac6f68161d6a259e2a6f70e7902855ea1510aa260e79093520a7803
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac
        - quay.io/kubevirt/libguestfs-tools@sha256:0de3de89720d4e5433a33944f00dac590daed6ab12e4d26a657ef082d0573747
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:e2def4213ec0657e72eb790ae8a115511d5b8f164a62d3568d2f1bff189917e8
        - quay.io/kubevirt/mtq-operator@sha256:859524721ae1a3b1e5dd6e29036b4c80acff13f74a4ee6769d867283e9ccbba5
        - registry.k8s.io/sig-storage/livenessprobe@sha256:5baeb4a6d7d517434292758928bb33efc6397368cbb48c8a4cf29496abf4e987
        - quay.io/kubevirt/virtio-container-disk@sha256:f8e71a3a930ce95d646d80862d5ff6f9d3ed7ef2541dd2e2acec0bce26274734
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:935475c2850466aa5ac57e4de627fb177515cb2c402a95842ead095d82b6df5f
        - quay.io/kubevirt/mtq-controller@sha256:04b38720b65479046270c8132cff5ab0f2dabc9cae91de1cc95fecf167efc356
        - quay.io/kubevirt/virt-exportserver@sha256:acac33a19528592cc3ef0cd3a25fa4c72bcf992e4c2c247437007fcb9de0aa6d
        - quay.io/kubevirt/cdi-uploadproxy@sha256:86285df5cf52533732ddf4fdb74674cbce877ba69833cb8ed058c2a157cb5573
        - quay.io/kubevirt/cdi-cloner@sha256:05da8df0b64024f9e2a48a1a54f06c821d79a3efe7828fad2c4fda6a01d0cd86
        - quay.io/kubevirt/cni-default-plugins@sha256:0c354fa9d695b8cab97b459e8afea2f7662407a987e83f6f6f1a8af4b45726be
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:3c70cb833819d8346e5531b10ac516283bab7098e782708768a6fe00ae0ddfac
        - quay.io/kubevirt/cdi-operator@sha256:f68e4f2a082fc72d92bc1df0340285af03cd12d8c7721b38c619d1de0f4c175a
        - ghcr.io/kubevirt/kubesecondarydns@sha256:6268d84154e2483fbce8c1adacbdaf6f0839117b2d48d9fa4687cc8f76bd5130
        - quay.io/kubevirt/virt-api@sha256:aeeae94c9dd3b352ff2fbcdba77a6b084fb4ebe0fed35fe7af9481e9e2b1f24c
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:83b460502671fb4f34116363a1a39b2ddfc9d14a920ee0a6413bfc3bd0580404
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - quay.io/kubevirt/cdi-apiserver@sha256:4ba004f57e7d589ddb5372a9729360460a97effecc6e5ecaccd7fbf16c6fcf13
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:d51821f001c959a79943cb22f5ea47589d7cf5c032e11ec1d571e8db7bf29b78
        - quay.io/kubevirt/virt-exportproxy@sha256:dd3b5c12314fa76b88d55d978210de81562fb888351748ea3997551c8c96a877
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:c4b6b02737bc24906fcce57fe6626d1a36cb2b91baa971af2a5e5a919093c34e
        - quay.io/kubevirt/cdi-importer@sha256:de5b89c879a531165482d540deffdc2dec0c188fd0f7b5fbd49c0298cf9a5b1d
        - quay.io/kubevirt/kubemacpool@sha256:20e156be33e6d3692c456081acbb91a4349f94de448f7d1f1cddd0228931b31f
        - quay.io/kubevirt/virt-launcher@sha256:cd37543ed2bc6baadbcbffa6951d740e4c65dc0ea366dec0a8593193c384e6de
        - quay.io/kubevirt/virt-handler@sha256:2a7d96da4de3ea51339b3a047abb4b05e088e3d1080110b42a85ea5baff4a0cc
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:246b3819cfa3833f9a8cb01e8cb859aa61a1924a85ea4a112e43e8414f32c429
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:3fbcc32bd4e4d15bd93c96def784a229cd84cca27942bf4858b581f31c97ee02
        - quay.io/kubevirt/cdi-controller@sha256:392ead2fdc5050c584765ebf99706cc5c0ab9c2b424b692ad35005e77d110b58
        - quay.io/kubevirt/cdi-uploadserver@sha256:d7fb653bb33dade84583fff6048092346639532d6675aaf9f88070171cb0fe7c
        - quay.io/kubevirt/bridge-marker@sha256:bba066e3b5ff3fb8c5e20861fe8abe51e3c9b50ad6ce3b2616af9cb5479a06d0
        - quay.io/kubevirt/aaq-server@sha256:9db6ca9147b1d75ba4d6bb1326c4ec289327eb9f8aadbcd951fcc689e584cd0d
        - quay.io/kubevirt/macvtap-cni@sha256:850b89343ace7c7ea6b18dd8e11964613974e9d1f7377af03854d407fb15230a
        version: 1.12.0
      entries:
      - name: kubevirt-hyperconverged-operator.v1.12.0
        version: 1.12.0
      - name: kubevirt-hyperconverged-operator.v1.11.1
        version: 1.11.1
      - name: kubevirt-hyperconverged-operator.v1.10.0
        version: 1.10.0
      - name: kubevirt-hyperconverged-operator.v1.9.0
        version: 1.9.0
      - name: kubevirt-hyperconverged-operator.v1.8.2
        version: 1.8.2
      - name: kubevirt-hyperconverged-operator.v1.8.1
        version: 1.8.1
      - name: kubevirt-hyperconverged-operator.v1.8.0
        version: 1.8.0
      - name: kubevirt-hyperconverged-operator.v1.7.0
        version: 1.7.0
      - name: kubevirt-hyperconverged-operator.v1.6.0
        version: 1.6.0
      - name: kubevirt-hyperconverged-operator.v1.5.2
        version: 1.5.2
      - name: kubevirt-hyperconverged-operator.v1.5.1
        version: 1.5.1
      - name: kubevirt-hyperconverged-operator.v1.5.0
        version: 1.5.0
      - name: kubevirt-hyperconverged-operator.v1.4.4
        version: 1.4.4
      - name: kubevirt-hyperconverged-operator.v1.4.3
        version: 1.4.3
      - name: kubevirt-hyperconverged-operator.v1.4.2
        version: 1.4.2
      - name: kubevirt-hyperconverged-operator.v1.4.1
        version: 1.4.1
      - name: kubevirt-hyperconverged-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: community-kubevirt-hyperconverged
    provider:
      name: KubeVirt project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: StreamNative
      provider-url: https://streamnative.io
    name: sn-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
        - docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21
        - docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      name: alpha
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
        - docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21
        - docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      name: beta
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21
        - quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
        - docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      name: stable
    defaultChannel: alpha
    packageName: sn-operator
    provider:
      name: StreamNative
      url: https://streamnative.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzo-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzo-operator.v2.0.101
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzo.cambridgesemantics.com/v1",
                "kind": "Anzo",
                "metadata": {
                  "name": "agent01"
                },
                "spec": {
                  "nodeConfig": {
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "app": "anzo"
                        }
                      },
                      "serviceName": "anzo-agent01",
                      "template": {
                        "metadata": {
                          "labels": {
                            "app": "anzo"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo@sha256:d1c2d54ca205d669c5ea1bb6aae105afe2385cc5695cecd78423ffa1d98929f8",
                              "name": "anzo",
                              "resources": {
                                "limits": {
                                  "cpu": "4000m",
                                  "memory": "12Gi"
                                },
                                "requests": {
                                  "cpu": "4000m",
                                  "memory": "12Gi"
                                }
                              }
                            }
                          ],
                          "serviceAccountName": "anzo-operator"
                        }
                      }
                    }
                  },
                  "role": "AnzoAgent"
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-operator@sha256:563e964f5771b9bfd860ea993b6e0936a54504908f5b2d1dff6f99ade74bb7a0
          createdAt: "2021-08-02"
          description: kubernetes operator for Anzo
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzo-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzo-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Anzo is the Schema for the anzos API
            displayName: Anzo Operator
            kind: Anzo
            name: anzos.anzo.cambridgesemantics.com
            version: v1
        description: |-
          The Anzo Operator provides the way to install and configure an anzo agent setup on Red Hat K8S environment.
          Currently, this is supported only through when deployed as an Anzo Agent as part of an Anzo Unstructured deployment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v2.0.0/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers through [ Cambridge Semantics Customer Center ]( https://customercenter.cambridgesemantics.com/ ).
        displayName: Anzo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        links:
        - name: Anzo Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v2.0.0/README_openshift.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-operator@sha256:563e964f5771b9bfd860ea993b6e0936a54504908f5b2d1dff6f99ade74bb7a0
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo@sha256:d1c2d54ca205d669c5ea1bb6aae105afe2385cc5695cecd78423ffa1d98929f8
        version: 2.0.0
      entries:
      - name: anzo-operator.v2.0.101
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anzo-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: odr-hub-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odr-hub-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east"
                },
                "spec": {
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east1"
                },
                "spec": {
                  "cidrs": [
                    "192.168.39.45/32"
                  ],
                  "clusterFence": "ManuallyFenced",
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east1"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPlacementControl",
                "metadata": {
                  "name": "drplacementcontrol-sample",
                  "namespace": "application-namespace"
                },
                "spec": {
                  "drPolicyRef": {
                    "name": "drpolicy-sample"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "excludedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "includedResources": [
                          "*"
                        ]
                      }
                    ]
                  },
                  "placementRef": {
                    "kind": "PlacementRule",
                    "name": "application-placement-rule"
                  },
                  "preferredCluster": "east",
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east",
                    "west"
                  ],
                  "replicationClassSelector": {
                    "matchLabels": {
                      "class": "ramen"
                    }
                  },
                  "schedulingInterval": "1h"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "metrodr-drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east1",
                    "east2"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DRCluster is the Schema for the drclusters API
            displayName: DRCluster
            kind: DRCluster
            name: drclusters.ramendr.openshift.io
            version: v1alpha1
          - description: DRPlacementControl is the Schema for the drplacementcontrols
              API
            displayName: DRPlacement Control
            kind: DRPlacementControl
            name: drplacementcontrols.ramendr.openshift.io
            version: v1alpha1
          - description: DRPolicy is the Schema for the drpolicies API
            displayName: DRPolicy
            kind: DRPolicy
            name: drpolicies.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Hub is a disaster-recovery orchestrator for stateful
          applications. It operates from an Advanced Cluster Management (ACM) hub
          cluster to orchestrate the recovery of application state, and scheduling
          of ACM PlacementRule for disaster recovery operations.
        displayName: Openshift DR Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:769acf15d016d4992793c9c527d7c268d2a67502dd7e0fadc240e4896398195b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 4.15.20-rhodf
      entries:
      - name: odr-hub-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odr-hub-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odr-hub-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odr-hub-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odr-hub-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odr-hub-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odr-hub-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odr-hub-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odr-hub-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odr-hub-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odr-hub-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odr-hub-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odr-hub-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odr-hub-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odr-hub-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odr-hub-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odr-hub-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odr-hub-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odr-hub-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odr-hub-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odr-hub-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east"
                },
                "spec": {
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east1"
                },
                "spec": {
                  "cidrs": [
                    "192.168.39.45/32"
                  ],
                  "clusterFence": "ManuallyFenced",
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east1"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPlacementControl",
                "metadata": {
                  "name": "drplacementcontrol-sample",
                  "namespace": "application-namespace"
                },
                "spec": {
                  "drPolicyRef": {
                    "name": "drpolicy-sample"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "excludedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "includedResources": [
                          "*"
                        ]
                      }
                    ]
                  },
                  "placementRef": {
                    "kind": "PlacementRule",
                    "name": "application-placement-rule"
                  },
                  "preferredCluster": "east",
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east",
                    "west"
                  ],
                  "replicationClassSelector": {
                    "matchLabels": {
                      "class": "ramen-replication-class"
                    }
                  },
                  "schedulingInterval": "1h",
                  "volumeSnapshotClassSelector": {
                    "matchLabels": {
                      "class": "ramen-snapshot-class"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "metrodr-drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east1",
                    "east2"
                  ],
                  "replicationClassSelector": {},
                  "schedulingInterval": {},
                  "volumeSnapshotClassSelector": {}
                }
              }
            ]
          capabilities: Deep Insights
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DRCluster is the Schema for the drclusters API
            displayName: DRCluster
            kind: DRCluster
            name: drclusters.ramendr.openshift.io
            version: v1alpha1
          - description: DRPlacementControl is the Schema for the drplacementcontrols
              API
            displayName: DRPlacement Control
            kind: DRPlacementControl
            name: drplacementcontrols.ramendr.openshift.io
            version: v1alpha1
          - description: DRPolicy is the Schema for the drpolicies API
            displayName: DRPolicy
            kind: DRPolicy
            name: drpolicies.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Hub is a disaster-recovery orchestrator for stateful
          applications. It operates from an Advanced Cluster Management (ACM) hub
          cluster to orchestrate the recovery of application state, and scheduling
          of ACM PlacementRule for disaster recovery operations.
        displayName: Openshift DR Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:af262b0890300183db1766bc392017b0bda723a9ba841dc25e3d928aa413db9a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        version: 4.16.16-rhodf
      entries:
      - name: odr-hub-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odr-hub-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odr-hub-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odr-hub-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odr-hub-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odr-hub-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odr-hub-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odr-hub-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odr-hub-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odr-hub-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odr-hub-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odr-hub-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odr-hub-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odr-hub-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odr-hub-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odr-hub-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odr-hub-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odr-hub-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: container-security-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: container-security-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secscan.quay.redhat.com/v1alpha1",
                "kind": "ImageManifestVuln",
                "metadata": {
                  "selfLink": "/apis/secscan.quay.redhat.com/v1alpha1/namespaces/openshift-cluster-version/imagemanifestvulns/sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18",
                  "resourceVersion": "14565395",
                  "name": "sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18",
                  "uid": "3f16a188-e69a-406c-aa06-df340195409c",
                  "creationTimestamp": "2021-03-17T12:07:16Z",
                  "generation": 2,
                  "managedFields": [
                    {
                      "apiVersion": "secscan.quay.redhat.com/v1alpha1",
                      "fieldsType": "FieldsV1",
                      "fieldsV1": {
                        "f:metadata": {
                          "f:labels": {
                            ".": {},
                            "f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": {}
                          }
                        },
                        "f:spec": {
                          ".": {},
                          "f:features": {},
                          "f:image": {},
                          "f:manifest": {}
                        },
                        "f:status": {
                          ".": {},
                          "f:affectedPods": {
                            ".": {},
                            "f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": {}
                          },
                          "f:fixableCount": {},
                          "f:highCount": {},
                          "f:highestSeverity": {},
                          "f:lastUpdate": {},
                          "f:lowCount": {},
                          "f:mediumCount": {}
                        }
                      },
                      "manager": "security-labeller",
                      "operation": "Update",
                      "time": "2021-03-17T13:07:20Z"
                    }
                  ],
                  "namespace": "openshift-cluster-version",
                  "labels": {
                    "openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": "true"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:447e05859134a0455ca55f8aae16c8a27ce06af4c18c191b41f6339df553f1bf
          createdAt: 2021-11-02 11:11 UTC
          description: Identify image vulnerabilities in Kubernetes pods
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Quay Container Security Operator (CSO) brings Quay and Clair
          metadata to Kubernetes / OpenShift. Starting with vulnerability information
          the scope will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Quay Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Quay Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:447e05859134a0455ca55f8aae16c8a27ce06af4c18c191b41f6339df553f1bf
        version: 3.4.7
      entries:
      - name: container-security-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: container-security-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"secscan.quay.redhat.com/v1alpha1\",\n
            \   \"kind\": \"ImageManifestVuln\",\n    \"metadata\": {\n      \"selfLink\":
            \"/apis/secscan.quay.redhat.com/v1alpha1/namespaces/openshift-cluster-version/imagemanifestvulns/sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18\",\n
            \     \"resourceVersion\": \"14565395\",\n      \"name\": \"sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18\",\n
            \     \"uid\": \"3f16a188-e69a-406c-aa06-df340195409c\",\n      \"creationTimestamp\":
            \"2021-03-17T12:07:16Z\",\n      \"generation\": 2,\n      \"managedFields\":
            [\n        {\n          \"apiVersion\": \"secscan.quay.redhat.com/v1alpha1\",\n
            \         \"fieldsType\": \"FieldsV1\",\n          \"fieldsV1\": {\n            \"f:metadata\":
            {\n              \"f:labels\": {\n                \".\": {},\n                \"f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            {}\n              }\n            },\n            \"f:spec\": {\n              \".\":
            {},\n              \"f:features\": {},\n              \"f:image\": {},\n
            \             \"f:manifest\": {}\n            },\n            \"f:status\":
            {\n              \".\": {},\n              \"f:affectedPods\": {\n                \".\":
            {},\n                \"f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            {}\n              },\n              \"f:fixableCount\": {},\n              \"f:highCount\":
            {},\n              \"f:highestSeverity\": {},\n              \"f:lastUpdate\":
            {},\n              \"f:lowCount\": {},\n              \"f:mediumCount\":
            {}\n            }\n          },\n          \"manager\": \"security-labeller\",\n
            \         \"operation\": \"Update\",\n          \"time\": \"2021-03-17T13:07:20Z\"\n
            \       }\n      ],\n      \"namespace\": \"openshift-cluster-version\",\n
            \     \"labels\": {\n        \"openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            \"true\"\n      }\n    },\n    \"spec\": {\n      \"features\": [\n        {\n
            \         \"name\": \"musl\",\n          \"namespaceName\": \"alpine:v3.14\",\n
            \         \"version\": \"1.2.2-r3\",\n          \"versionformat\": \"dpkg\",\n
            \         \"vulnerabilities\": [\n            {\n              \"fixedby\":
            \"1.2.2_pre2-r0\",\n              \"link\": \"https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28928\",\n
            \             \"metadata\": \"{\\\"NVD\\\": {\\\"CVSSv3\\\": {\\\"Vectors\\\":
            \\\"CVSS:3.0/AV:L/AC:L/PR:L/UI:N/S:U/C:N/I:N/A:H\\\", \\\"ImpactScore\\\":
            3.6, \\\"Score\\\": 5.5, \\\"ExploitabilityScore\\\": 1.8}, \\\"CVSSv2\\\":
            {\\\"Score\\\": 2.1, \\\"PublishedDateTime\\\": \\\"2020-11-24T18:15Z\\\",
            \\\"Vectors\\\": \\\"AV:L/AC:L/Au:N/C:N/I:N/A:P\\\"}}}\",\n              \"name\":
            \"CVE-2020-28928\",\n              \"namespaceName\": \"alpine:v3.14\",\n
            \             \"severity\": \"Low\"\n            }\n          ]\n        }\n
            \     ],\n      \"image\": \"quay.io/harishg/pruned-redhat-operators-index\",\n
            \     \"manifest\": \"sha256:19631c86642609254777ff59fc9e128aa5d0a94a7dbb504ef1fcfb2bbc6e486e\"\n
            \   }\n  }\n]    "
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:6b3581e62550d6a615d36eee6eaf4c89b9c35ba53cab8433bd93c68393a9d115
          createdAt: 2021-09-10 23:09 UTC
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.4.x <3.5.7'
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Quay Container Security Operator (CSO) brings Quay and Clair
          metadata to Kubernetes / OpenShift. Starting with vulnerability information
          the scope will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Quay Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Quay Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:6b3581e62550d6a615d36eee6eaf4c89b9c35ba53cab8433bd93c68393a9d115
        version: 3.5.7
      entries:
      - name: container-security-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: container-security-operator.v3.10.15
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:de7c66a0004f3c0643c09d4c3bbb4a0f1f8f2cbf264250802cdc58028fc72b50
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.10.15
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:de7c66a0004f3c0643c09d4c3bbb4a0f1f8f2cbf264250802cdc58028fc72b50
        version: 3.10.15
      entries:
      - name: container-security-operator.v3.10.15
        version: 3.10.15
      - name: container-security-operator.v3.10.14
        version: 3.10.14
      - name: container-security-operator.v3.10.13
        version: 3.10.13
      - name: container-security-operator.v3.10.12
        version: 3.10.12
      - name: container-security-operator.v3.10.11
        version: 3.10.11
      - name: container-security-operator.v3.10.10
        version: 3.10.10
      - name: container-security-operator.v3.10.9
        version: 3.10.9
      - name: container-security-operator.v3.10.8
        version: 3.10.8
      - name: container-security-operator.v3.10.7
        version: 3.10.7
      - name: container-security-operator.v3.10.6
        version: 3.10.6
      - name: container-security-operator.v3.10.5
        version: 3.10.5
      - name: container-security-operator.v3.10.4
        version: 3.10.4
      - name: container-security-operator.v3.10.3
        version: 3.10.3
      - name: container-security-operator.v3.10.2
        version: 3.10.2
      - name: container-security-operator.v3.10.1
        version: 3.10.1
      - name: container-security-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: container-security-operator.v3.11.13
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:3914389fd53c371627921298c5dc3564999d92bc720f27a240f74c034ebbf5ec
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.11.13
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:3914389fd53c371627921298c5dc3564999d92bc720f27a240f74c034ebbf5ec
        version: 3.11.13
      entries:
      - name: container-security-operator.v3.11.13
        version: 3.11.13
      - name: container-security-operator.v3.11.12
        version: 3.11.12
      - name: container-security-operator.v3.11.11
        version: 3.11.11
      - name: container-security-operator.v3.11.10
        version: 3.11.10
      - name: container-security-operator.v3.11.9
        version: 3.11.9
      - name: container-security-operator.v3.11.8
        version: 3.11.8
      - name: container-security-operator.v3.11.7
        version: 3.11.7
      - name: container-security-operator.v3.11.6
        version: 3.11.6
      - name: container-security-operator.v3.11.5
        version: 3.11.5
      - name: container-security-operator.v3.11.4
        version: 3.11.4
      - name: container-security-operator.v3.11.3
        version: 3.11.3
      - name: container-security-operator.v3.11.2
        version: 3.11.2
      - name: container-security-operator.v3.11.1
        version: 3.11.1
      - name: container-security-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: container-security-operator.v3.12.12
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:1a15328f0b2c64d3a857f2f155ad4def095027e809610de98545e303eec63225
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.12.12
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:1a15328f0b2c64d3a857f2f155ad4def095027e809610de98545e303eec63225
        version: 3.12.12
      entries:
      - name: container-security-operator.v3.12.12
        version: 3.12.12
      - name: container-security-operator.v3.12.11
        version: 3.12.11
      - name: container-security-operator.v3.12.10
        version: 3.12.10
      - name: container-security-operator.v3.12.9
        version: 3.12.9
      - name: container-security-operator.v3.12.8
        version: 3.12.8
      - name: container-security-operator.v3.12.7
        version: 3.12.7
      - name: container-security-operator.v3.12.6
        version: 3.12.6
      - name: container-security-operator.v3.12.5
        version: 3.12.5
      - name: container-security-operator.v3.12.4
        version: 3.12.4
      - name: container-security-operator.v3.12.3
        version: 3.12.3
      - name: container-security-operator.v3.12.2
        version: 3.12.2
      - name: container-security-operator.v3.12.1
        version: 3.12.1
      - name: container-security-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: container-security-operator.v3.13.8
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:ba50621739581da9a661c2c2e355849c328cfa33b7311b06abb148e1183a6321
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.13.8
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:ba50621739581da9a661c2c2e355849c328cfa33b7311b06abb148e1183a6321
        version: 3.13.8
      entries:
      - name: container-security-operator.v3.13.8
        version: 3.13.8
      - name: container-security-operator.v3.13.7
        version: 3.13.7
      - name: container-security-operator.v3.13.6
        version: 3.13.6
      - name: container-security-operator.v3.13.5
        version: 3.13.5
      - name: container-security-operator.v3.13.4
        version: 3.13.4
      - name: container-security-operator.v3.13.3
        version: 3.13.3
      - name: container-security-operator.v3.13.2
        version: 3.13.2
      - name: container-security-operator.v3.13.1
        version: 3.13.1
      - name: container-security-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: container-security-operator.v3.14.5
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:c42487ccf098df3ec6dbeb5d409346bdd30e92781cf3ffea5fdb6dd0b196d0cf
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.14.5
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:c42487ccf098df3ec6dbeb5d409346bdd30e92781cf3ffea5fdb6dd0b196d0cf
        version: 3.14.5
      entries:
      - name: container-security-operator.v3.14.5
        version: 3.14.5
      - name: container-security-operator.v3.14.4
        version: 3.14.4
      - name: container-security-operator.v3.14.3
        version: 3.14.3
      - name: container-security-operator.v3.14.2
        version: 3.14.2
      - name: container-security-operator.v3.14.1
        version: 3.14.1
      - name: container-security-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: container-security-operator.v3.15.2
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:81e929ec095f791a83310ee7fcda239d5c348ba69a3cf5adcf1b3a77c349ce77
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.15.2
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:81e929ec095f791a83310ee7fcda239d5c348ba69a3cf5adcf1b3a77c349ce77
        version: 3.15.2
      entries:
      - name: container-security-operator.v3.15.2
        version: 3.15.2
      - name: container-security-operator.v3.15.1
        version: 3.15.1
      - name: container-security-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: container-security-operator.v3.6.10
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:641b80df8d7102d49a5dfd5b99245669f60393880a6f1bc8ba8450b88a640d96
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.6.10
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:641b80df8d7102d49a5dfd5b99245669f60393880a6f1bc8ba8450b88a640d96
        version: 3.6.10
      entries:
      - name: container-security-operator.v3.6.10
        version: 3.6.10
      - name: container-security-operator.v3.6.9
        version: 3.6.9
      - name: container-security-operator.v3.6.8
        version: 3.6.8
      - name: container-security-operator.v3.6.7
        version: 3.6.7
      - name: container-security-operator.v3.6.6
        version: 3.6.6
      - name: container-security-operator.v3.6.5
        version: 3.6.5
      - name: container-security-operator.v3.6.4
        version: 3.6.4
      - name: container-security-operator.v3.6.2
        version: 3.6.2
      - name: container-security-operator.v3.6.1
        version: 3.6.1
      - name: container-security-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: container-security-operator.v3.7.14
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:10736a06beb503b4eb73260a87a2cde9bd83353374ee0ac9e8dd0a6260cf2665
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.7.14
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:10736a06beb503b4eb73260a87a2cde9bd83353374ee0ac9e8dd0a6260cf2665
        version: 3.7.14
      entries:
      - name: container-security-operator.v3.7.14
        version: 3.7.14
      - name: container-security-operator.v3.7.13
        version: 3.7.13
      - name: container-security-operator.v3.7.12
        version: 3.7.12
      - name: container-security-operator.v3.7.11
        version: 3.7.11
      - name: container-security-operator.v3.7.10
        version: 3.7.10
      - name: container-security-operator.v3.7.9
        version: 3.7.9
      - name: container-security-operator.v3.7.8
        version: 3.7.8
      - name: container-security-operator.v3.7.7
        version: 3.7.7
      - name: container-security-operator.v3.7.6
        version: 3.7.6
      - name: container-security-operator.v3.7.5
        version: 3.7.5
      - name: container-security-operator.v3.7.4
        version: 3.7.4
      - name: container-security-operator.v3.7.3
        version: 3.7.3
      - name: container-security-operator.v3.7.2
        version: 3.7.2
      - name: container-security-operator.v3.7.1
        version: 3.7.1
      - name: container-security-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: container-security-operator.v3.8.15
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:108bd033543beb7cce077130c96f992c202f7a258229c239795f8f19e5fae5e5
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.8.15
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:108bd033543beb7cce077130c96f992c202f7a258229c239795f8f19e5fae5e5
        version: 3.8.15
      entries:
      - name: container-security-operator.v3.8.15
        version: 3.8.15
      - name: container-security-operator.v3.8.14
        version: 3.8.14
      - name: container-security-operator.v3.8.13
        version: 3.8.13
      - name: container-security-operator.v3.8.12
        version: 3.8.12
      - name: container-security-operator.v3.8.11
        version: 3.8.11
      - name: container-security-operator.v3.8.10
        version: 3.8.10
      - name: container-security-operator.v3.8.9
        version: 3.8.9
      - name: container-security-operator.v3.8.8
        version: 3.8.8
      - name: container-security-operator.v3.8.7
        version: 3.8.7
      - name: container-security-operator.v3.8.6
        version: 3.8.6
      - name: container-security-operator.v3.8.5
        version: 3.8.5
      - name: container-security-operator.v3.8.4
        version: 3.8.4
      - name: container-security-operator.v3.8.3
        version: 3.8.3
      - name: container-security-operator.v3.8.2
        version: 3.8.2
      - name: container-security-operator.v3.8.1
        version: 3.8.1
      - name: container-security-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: container-security-operator.v3.9.16
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:df8e1d246273d21049c6bb8169a2e5f3fc8d69c85c564359abea7364969c6acc
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.9.16
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:df8e1d246273d21049c6bb8169a2e5f3fc8d69c85c564359abea7364969c6acc
        version: 3.9.16
      entries:
      - name: container-security-operator.v3.9.16
        version: 3.9.16
      - name: container-security-operator.v3.9.15
        version: 3.9.15
      - name: container-security-operator.v3.9.14
        version: 3.9.14
      - name: container-security-operator.v3.9.13
        version: 3.9.13
      - name: container-security-operator.v3.9.12
        version: 3.9.12
      - name: container-security-operator.v3.9.11
        version: 3.9.11
      - name: container-security-operator.v3.9.10
        version: 3.9.10
      - name: container-security-operator.v3.9.9
        version: 3.9.9
      - name: container-security-operator.v3.9.8
        version: 3.9.8
      - name: container-security-operator.v3.9.7
        version: 3.9.7
      - name: container-security-operator.v3.9.6
        version: 3.9.6
      - name: container-security-operator.v3.9.5
        version: 3.9.5
      - name: container-security-operator.v3.9.4
        version: 3.9.4
      - name: container-security-operator.v3.9.3
        version: 3.9.3
      - name: container-security-operator.v3.9.2
        version: 3.9.2
      - name: container-security-operator.v3.9.1
        version: 3.9.1
      - name: container-security-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: container-security-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: prometurbo-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometurbo-operator.v8.17.40-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5-beta.1 <8.17.40-beta.1'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
          - kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.

          ### Get Started
          Please follow [this](https://www.ibm.com/docs/en/tarm/8.15.x?topic=configuration-prometheus) for an overview of Prometurbo and deployment instructions.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/prometurbo
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/prometurbo-operator@sha256:a1605506884f6bce7471d79c6243728bce991ceb381d7fe0c51f1cdcbeb96e22
        version: 8.17.40-beta.1
      entries:
      - name: prometurbo-operator.v8.17.40-beta.1
        version: 8.17.40-beta.1
      name: beta
    - currentCSV: prometurbo-operator.v8.17.40
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5 <8.17.40'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
          - kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.

          ### Get Started
          Please follow [this](https://www.ibm.com/docs/en/tarm/8.15.x?topic=configuration-prometheus) for an overview of Prometurbo and deployment instructions.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/prometurbo
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/prometurbo-operator@sha256:cbcf90e44ffcedec8e3f161eff4c95b581924a1db195bf55d13da3714346b26f
        version: 8.17.40
      entries:
      - name: prometurbo-operator.v8.17.40
        version: 8.17.40
      name: stable
    defaultChannel: stable
    packageName: prometurbo-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: AppDynamics LLC
      provider-url: https://www.appdynamics.com/
    name: appdynamics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appdynamics-operator.v24.7.0-861
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clusteragent",
                "metadata": {
                  "name": "k8s-cluster-agent",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "appName": "<app-name>",
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/cluster-agent:tag"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "InfraViz",
                "metadata": {
                  "name": "appdynamics-infraviz",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/machine-agent:latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3c420cce8b626d8f5cf424410527c6a7a17ca3348e903dd8cdfcd74819b8de41
          createdAt: "2023-02-06T14:26:07Z"
          description: AppDynamics Operator to install AppDynamics Cluster Agent &
            AppDynamics Machine Agent for K8s Monitoring for various Cloud Providers
            like EKS, AKS etc.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: support@appdynamics.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clusteragent is the Schema for the clusteragents API
            displayName: Clusteragent
            kind: Clusteragent
            name: clusteragents.cluster.appdynamics.com
            version: v1alpha1
          - description: InfraViz is the Schema for the infravizs API
            displayName: Infra Viz
            kind: InfraViz
            name: infravizs.cluster.appdynamics.com
            version: v1alpha1
        description: |-
          AppDynamics Operator to install AppDynamics Cluster Agent & Machine Agent for K8s Monitoring on various Cloud Providers like EKS, AKS, GCP, Openshift etc.
          **About AppDynamics Operator**
          AppDynamics Operator simplifies the installation and lifecycle management of the AppDynamics ClusterAgent and the AppDynamics Machine Agent for K8s Monitoring on different Cloud Providers like EKS, AKS, GCP, OpenShift etc.
          **Installation Steps**
          Please refer below links for installation guide:
          [Install AppDynamics Cluster Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli)
          [Install AppDynamics Machine Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli)
          [Install AppDynamics Cluster Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts)
          [Install Appdynamics Machine Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts)
        displayName: AppDynamics Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AppDynamics
        - Operator
        - AppDynamics Cluster Agent
        - AppDynamics Machine Agent
        - AppDynamics K8s Monitoring
        links:
        - name: AppDynamics Kubernetes Monitoring Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent
        - name: AppDynamics Cluster Agent Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/overview-of-cluster-monitoring
        - name: AppDynamics Cluster Agent Requirements
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/cluster-agent-requirements-and-supported-environments
        - name: Install AppDynamics Cluster Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli
        - name: Install AppDynamics Cluster Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts
        - name: Install AppDynamics Machine Agent & NetViz Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli
        - name: Install AppDynamics Machine Agent & NetViz Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: AppDynamics LLC
          url: https://www.appdynamics.com/
        relatedImages:
        - registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3c420cce8b626d8f5cf424410527c6a7a17ca3348e903dd8cdfcd74819b8de41
        version: 24.7.0-861
      entries:
      - name: appdynamics-operator.v24.7.0-861
        version: 24.7.0-861
      name: alpha
    - currentCSV: appdynamics-operator.v25.9.0-1218
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clusteragent",
                "metadata": {
                  "name": "k8s-cluster-agent",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "appName": "<app-name>",
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/cluster-agent:tag"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "InfraViz",
                "metadata": {
                  "name": "appdynamics-infraviz",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/machine-agent:latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3320c2c8d5f60adc541befd099dac1ef96257b16e0df28b727b0e589f78d981a
          createdAt: "2023-02-06T14:26:07Z"
          description: AppDynamics Operator to install AppDynamics Cluster Agent &
            AppDynamics Machine Agent for K8s Monitoring for various Cloud Providers
            like EKS, AKS etc.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: support@appdynamics.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clusteragent is the Schema for the clusteragents API
            displayName: Clusteragent
            kind: Clusteragent
            name: clusteragents.cluster.appdynamics.com
            version: v1alpha1
          - description: InfraViz is the Schema for the infravizs API
            displayName: Infra Viz
            kind: InfraViz
            name: infravizs.cluster.appdynamics.com
            version: v1alpha1
        description: |-
          AppDynamics Operator to install AppDynamics Cluster Agent & Machine Agent for K8s Monitoring on various Cloud Providers like EKS, AKS, GCP, Openshift etc.
          **About AppDynamics Operator**
          AppDynamics Operator simplifies the installation and lifecycle management of the AppDynamics ClusterAgent and the AppDynamics Machine Agent for K8s Monitoring on different Cloud Providers like EKS, AKS, GCP, OpenShift etc.
          **Installation Steps**
          Please refer below links for installation guide:
          [Install AppDynamics Cluster Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli)
          [Install AppDynamics Machine Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli)
          [Install AppDynamics Cluster Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts)
          [Install Appdynamics Machine Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts)
        displayName: AppDynamics Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - AppDynamics
        - Operator
        - AppDynamics Cluster Agent
        - AppDynamics Machine Agent
        - AppDynamics K8s Monitoring
        links:
        - name: AppDynamics Kubernetes Monitoring Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent
        - name: AppDynamics Cluster Agent Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/overview-of-cluster-monitoring
        - name: AppDynamics Cluster Agent Requirements
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/cluster-agent-requirements-and-supported-environments
        - name: Install AppDynamics Cluster Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli
        - name: Install AppDynamics Cluster Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts
        - name: Install AppDynamics Machine Agent & NetViz Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli
        - name: Install AppDynamics Machine Agent & NetViz Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: AppDynamics LLC
          url: https://www.appdynamics.com/
        relatedImages:
        - registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3320c2c8d5f60adc541befd099dac1ef96257b16e0df28b727b0e589f78d981a
        version: 25.9.0-1218
      entries:
      - name: appdynamics-operator.v25.9.0-1218
        version: 25.9.0-1218
      name: stable
    defaultChannel: alpha
    packageName: appdynamics-operator
    provider:
      name: AppDynamics LLC
      url: https://www.appdynamics.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: www.redhat.com
    name: odf-node-recovery-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-node-recovery-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "NodeRecovery",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "odf-node-recovery-operator",
                    "app.kubernetes.io/instance": "noderecovery-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "noderecovery",
                    "app.kubernetes.io/part-of": "odf-node-recovery-operator"
                  },
                  "name": "noderecovery-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Storage,OpenShift Optional
          description: ODF Node Recovery is an operator that assist in the recovery
            of and ODF cluster that has a device or a node replaced
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/jordigilh/odf-node-recovery-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeRecovery is the Schema for the noderecoveries API
            displayName: Node Recovery
            kind: NodeRecovery
            name: noderecoveries.odf.openshift.io
            version: v1alpha1
        description: |
          ODF Node Recovery is an operator that assist in the recovery of an
          ODF cluster in these 2 scenarios:
          * <a href="https://docs.redhat.com/en/documentation/red_hat_openshift_data_foundation/4.14/html-single/replacing_devices/index"> Device replacement<a/>
          * <a href="https://docs.redhat.com/en/documentation/red_hat_openshift_data_foundation/4.14/html-single/replacing_nodes/index"> Node replacement</a>
        displayName: ODF Node Recovery Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - node recovery
        links:
        - name: GitHub Repository
          url: https://github.com/jordigilh/odf-node-recovery-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
          url: www.redhat.com
        relatedImages:
        - quay.io/jordigilh/odf-node-recovery-controller@sha256:4164fafa9bb1c4d6121f4c5fc700fa2827c3994ca98a65d641471a2a46d4f59c
        version: 1.1.0
      entries:
      - name: odf-node-recovery-operator.v1.1.0
        version: 1.1.0
      - name: odf-node-recovery-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: odf-node-recovery-operator
    provider:
      name: Red Hat, Inc.
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: authorino-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: authorino-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/authorino-operator:v0.8.0
          createdAt: 2021-12-08T10-00-00Z
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta1
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: The operator to manage instances of Authorino
        displayName: Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: Authorino Operator
          url: https://github.com/Kuadrant/authorino-operator
        - name: Authorino
          url: https://github.com/Kuadrant/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/authorino-operator:v0.8.0
        version: 0.8.0
      entries:
      - name: authorino-operator.v0.8.0
        version: 0.8.0
      - name: authorino-operator.v0.7.0
        version: 0.7.0
      - name: authorino-operator.v0.6.0
        version: 0.6.0
      - name: authorino-operator.v0.5.0
        version: 0.5.0
      - name: authorino-operator.v0.4.1
        version: 0.4.1
      - name: authorino-operator.v0.4.0
        version: 0.4.0
      - name: authorino-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: authorino-operator.v0.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta3",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/authorino-operator:v0.16.0
          createdAt: "2024-11-25T13:22:24Z"
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta3
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: The operator to manage instances of Authorino
        displayName: Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: Authorino Operator
          url: https://github.com/Kuadrant/authorino-operator
        - name: Authorino
          url: https://github.com/Kuadrant/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/authorino-operator:v0.16.0
        - quay.io/kuadrant/authorino:v0.20.0
        version: 0.16.0
      entries:
      - name: authorino-operator.v0.16.0
        version: 0.16.0
      - name: authorino-operator.v0.15.1
        version: 0.15.1
      - name: authorino-operator.v0.15.0
        version: 0.15.0
      - name: authorino-operator.v0.14.0
        version: 0.14.0
      - name: authorino-operator.v0.13.1
        version: 0.13.1
      - name: authorino-operator.v0.13.0
        version: 0.13.0
      - name: authorino-operator.v0.12.0
        version: 0.12.0
      - name: authorino-operator.v0.11.1
        version: 0.11.1
      - name: authorino-operator.v0.10.0
        version: 0.10.0
      - name: authorino-operator.v0.9.0
        version: 0.9.0
      - name: authorino-operator.v0.8.0
        version: 0.8.0
      - name: authorino-operator.v0.7.0
        version: 0.7.0
      - name: authorino-operator.v0.6.0
        version: 0.6.0
      - name: authorino-operator.v0.5.0
        version: 0.5.0
      - name: authorino-operator.v0.4.1
        version: 0.4.1
      - name: authorino-operator.v0.4.0
        version: 0.4.0
      - name: authorino-operator.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: authorino-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Eclipse Foundation
      provider-url: ""
    name: eclipse-che
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eclipse-che.v7.110.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "org.eclipse.che/v1",
                "kind": "CheCluster",
                "metadata": {
                  "name": "eclipse-che",
                  "namespace": "eclipse-che"
                },
                "spec": {
                  "auth": {
                    "identityProviderURL": "",
                    "oAuthClientName": "",
                    "oAuthSecret": ""
                  },
                  "database": {
                    "externalDb": false
                  },
                  "k8s": {
                    "ingressDomain": null,
                    "tlsSecretName": null
                  },
                  "metrics": {
                    "enable": true
                  },
                  "server": {
                    "externalDevfileRegistries": [
                      {
                        "url": "https://registry.devfile.io"
                      }
                    ],
                    "externalDevfileRegistry": true,
                    "externalPluginRegistry": true,
                    "workspaceNamespaceDefault": "<username>-che"
                  },
                  "storage": {
                    "pvcStrategy": "common"
                  }
                }
              },
              {
                "apiVersion": "org.eclipse.che/v2",
                "kind": "CheCluster",
                "metadata": {
                  "name": "eclipse-che",
                  "namespace": "eclipse-che"
                },
                "spec": {
                  "components": {
                    "devfileRegistry": {
                      "disableInternalRegistry": true,
                      "externalDevfileRegistries": [
                        {
                          "url": "https://registry.devfile.io"
                        }
                      ]
                    },
                    "pluginRegistry": {
                      "disableInternalRegistry": true
                    }
                  },
                  "containerRegistry": {},
                  "devEnvironments": {},
                  "gitServices": {},
                  "networking": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/eclipse/che-operator@sha256:10c4d05c2cb848d04b97caf752c7f4d181e6973683263e45adec417342164e51
          createdAt: "2025-10-02T14:44:20Z"
          description: A Kube-native development solution that delivers portable and
            collaborative developer workspaces.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eclipse-che/che-operator
          support: Eclipse Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              The `CheCluster` custom resource allows defining and managing Eclipse Che server installation.
              Based on these settings, the  Operator automatically creates and maintains several ConfigMaps:
              `che`, `plugin-registry` that will contain the appropriate environment variables
              of the various components of the installation. These generated ConfigMaps must NOT be updated manually.
            displayName: Eclipse Che instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v2
          - description: The `CheCluster` custom resource allows defining and managing
              a Che server installation
            displayName: Eclipse Che instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v1
        description: |
          A collaborative Kubernetes-native development solution that delivers OpenShift workspaces and in-browser IDE for rapid cloud application development.
          This operator installs the Plugin and Devfile registries, Dashboard, Gateway and the Eclipse Che server, and configures these services.
          OpenShift OAuth is used directly for authentication. TLS mode is on.

          ## How to Install
          Press the **Install** button, choose the channel and the upgrade strategy, and wait for the **Installed** Operator status.
          When the operator is installed, create a new Custom Resource (CR) of Kind CheCluster (click the **Create New** button).
          The CR spec contains all defaults. You can start using Eclipse Che when the CR status is set to **Available**, and you see a URL to Eclipse Che.

          ## Defaults
          By default, the operator deploys Eclipse Che with:
          * 10Gi storage
          * Auto-generated passwords
          * Bundled Plugin and Devfile registries

          Use `oc edit checluster/eclipse-che -n eclipse-che` to update Eclipse Che default installation options.
          See more in the [Installation guide](https://www.eclipse.org/che/docs/stable/administration-guide/configuring-che/).

          ### Certificates
          Operator uses a default router certificate to secure Eclipse Che routes.
          Follow the [guide](https://www.eclipse.org/che/docs/stable/administration-guide/importing-untrusted-tls-certificates/)
          to import certificates into Eclipse Che.
        displayName: Eclipse Che
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eclipse che
        - workspaces
        - devtools
        - developer
        - ide
        - java
        links:
        - name: Product Page
          url: https://www.eclipse.org/che
        - name: Documentation
          url: https://www.eclipse.org/che/docs
        - name: Operator GitHub Repo
          url: https://github.com/eclipse-che/che-operator
        maintainers:
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Eclipse Foundation
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy@sha256:738c854322f56d63ebab75de5210abcdd5e0782ce2d30c0ecd4620f63b24694d
        - quay.io/devfile/universal-developer-image@sha256:3cf53fb9eab60ee3156e9edd4ec5af72b70e55d5ffb28f41ecb5b26e01dc9838
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:354fc75eb7a21a934381e93d03ef9d42bc2c8ae8989fdcacecfb39b863b96ced
        - quay.io/eclipse/che-operator@sha256:10c4d05c2cb848d04b97caf752c7f4d181e6973683263e45adec417342164e51
        - quay.io/eclipse/che--traefik@sha256:4cf907247939b5d20bf4eff73abd21cb413c339600dde76dbc94a874b2578a27
        - quay.io/che-incubator/che-code@sha256:cfcfbe4d665b18b11401c3c35b23b04d292c31e7f4be6d78a5e6d60fce776ed0
        - quay.io/che-incubator/che-code-server@sha256:357c95297ce2182ef0a792038ea122075fdeb6c8e835099be09e8d4f105eac2a
        - quay.io/eclipse/che-tls-secret-creator@sha256:cbb2fd8ce903d822377f2093af1aa2543672ed8c516d607a7a9dd857379dac5d
        - quay.io/che-incubator/configbump@sha256:396714521307ae9f79453c492fb8064cf940cd6b4e6183a825f1056045623874
        - quay.io/che-incubator/che-idea-dev-server@sha256:87c8b53cc295e3d7eb35b326e221a949553be3345135a4b6c1999d66f345a8b2
        - quay.io/che-incubator/che-code-sshd@sha256:7eb5381fba8878b8f252523caf70c9957b9f9e699db4564061017fb0a023cc1a
        - quay.io/che-incubator/che-idea@sha256:8aae69dc4b0c122491a75400639af0fe92b5e214c6e68ac97cda29fb58b44151
        - quay.io/eclipse/che-plugin-registry@sha256:a70a0d98505ba5aeed691fc70c3c59aed3304155fba0ab3ad70a6d0453b22093
        - quay.io/che-incubator/che-code-server@sha256:2093baab1a73e9682f5892f80582481abfbef99d744472ee6098e2d82fb2899f
        - quay.io/che-incubator/header-rewrite-proxy@sha256:bd7873b8feef35f218f54c6251ea224bea2c8bf202a328230019a0ba2941245d
        - quay.io/oauth2-proxy/oauth2-proxy@sha256:dcb6ff8dd21bf3058f6a22c6fa385fa5b897a9cd3914c88a2cc2bb0a85f8065d
        - quay.io/che-incubator/che-code@sha256:ac62a2f42ae6d6f6628620a00133afece954fcbed2df309fa3de79e4c28cc029
        - quay.io/openshift/origin-oauth-proxy@sha256:870bfe92a4663720775c0dfe5728ecbb10a17f0644eef5f57276ec135034c6a1
        - quay.io/eclipse/che-dashboard@sha256:2db3255752aebd75dceea578cd27b7e8ecb41c2290827ff7104ed77e1b61feb9
        - quay.io/eclipse/che-server@sha256:b83d536cdd3cb2b86def5599495c40870dde20c194a003736a2b47a5e5e17a06
        version: 7.110.0
      entries:
      - name: eclipse-che.v7.110.0
        version: 7.110.0
      - name: eclipse-che.v7.109.0
        version: 7.109.0
      - name: eclipse-che.v7.108.0
        version: 7.108.0
      - name: eclipse-che.v7.107.0
        version: 7.107.0
      - name: eclipse-che.v7.106.0
        version: 7.106.0
      - name: eclipse-che.v7.105.0
        version: 7.105.0
      - name: eclipse-che.v7.104.0
        version: 7.104.0
      - name: eclipse-che.v7.103.0
        version: 7.103.0
      - name: eclipse-che.v7.102.0
        version: 7.102.0
      - name: eclipse-che.v7.101.0
        version: 7.101.0
      - name: eclipse-che.v7.100.0
        version: 7.100.0
      - name: eclipse-che.v7.99.0
        version: 7.99.0
      - name: eclipse-che.v7.98.0
        version: 7.98.0
      - name: eclipse-che.v7.97.0
        version: 7.97.0
      - name: eclipse-che.v7.96.0
        version: 7.96.0
      - name: eclipse-che.v7.95.0
        version: 7.95.0
      - name: eclipse-che.v7.94.0
        version: 7.94.0
      - name: eclipse-che.v7.93.0
        version: 7.93.0
      - name: eclipse-che.v7.92.0
        version: 7.92.0
      - name: eclipse-che.v7.91.0
        version: 7.91.0
      - name: eclipse-che.v7.90.0
        version: 7.90.0
      - name: eclipse-che.v7.89.0
        version: 7.89.0
      - name: eclipse-che.v7.88.0
        version: 7.88.0
      - name: eclipse-che.v7.87.0
        version: 7.87.0
      - name: eclipse-che.v7.86.0
        version: 7.86.0
      - name: eclipse-che.v7.85.0
        version: 7.85.0
      - name: eclipse-che.v7.84.0
        version: 7.84.0
      - name: eclipse-che.v7.83.0
        version: 7.83.0
      - name: eclipse-che.v7.82.0
        version: 7.82.0
      - name: eclipse-che.v7.81.0
        version: 7.81.0
      - name: eclipse-che.v7.80.0
        version: 7.80.0
      - name: eclipse-che.v7.79.0
        version: 7.79.0
      - name: eclipse-che.v7.78.0
        version: 7.78.0
      - name: eclipse-che.v7.77.0
        version: 7.77.0
      - name: eclipse-che.v7.76.0
        version: 7.76.0
      - name: eclipse-che.v7.75.0
        version: 7.75.0
      - name: eclipse-che.v7.74.0
        version: 7.74.0
      - name: eclipse-che.v7.73.0
        version: 7.73.0
      - name: eclipse-che.v7.72.0
        version: 7.72.0
      - name: eclipse-che.v7.71.0
        version: 7.71.0
      - name: eclipse-che.v7.70.0
        version: 7.70.0
      - name: eclipse-che.v7.69.0
        version: 7.69.0
      - name: eclipse-che.v7.68.0
        version: 7.68.0
      - name: eclipse-che.v7.67.0
        version: 7.67.0
      - name: eclipse-che.v7.66.0
        version: 7.66.0
      - name: eclipse-che.v7.65.0
        version: 7.65.0
      - name: eclipse-che.v7.64.0
        version: 7.64.0
      - name: eclipse-che.v7.63.0
        version: 7.63.0
      - name: eclipse-che.v7.62.0
        version: 7.62.0
      - name: eclipse-che.v7.61.0
        version: 7.61.0
      - name: eclipse-che.v7.60.1
        version: 7.60.1
      - name: eclipse-che.v7.60.0
        version: 7.60.0
      - name: eclipse-che.v7.59.0
        version: 7.59.0
      - name: eclipse-che.v7.58.0
        version: 7.58.0
      - name: eclipse-che.v7.57.0
        version: 7.57.0
      - name: eclipse-che.v7.56.0
        version: 7.56.0
      - name: eclipse-che.v7.55.0
        version: 7.55.0
      - name: eclipse-che.v7.54.0
        version: 7.54.0
      - name: eclipse-che.v7.53.0
        version: 7.53.0
      - name: eclipse-che.v7.52.0
        version: 7.52.0
      - name: eclipse-che.v7.51.0
        version: 7.51.0
      - name: eclipse-che.v7.50.0
        version: 7.50.0
      - name: eclipse-che.v7.49.0
        version: 7.49.0
      - name: eclipse-che.v7.48.1
        version: 7.48.1
      - name: eclipse-che.v7.47.0
        version: 7.47.0
      - name: eclipse-che.v7.46.0
        version: 7.46.0
      - name: eclipse-che.v7.45.0
        version: 7.45.0
      - name: eclipse-che.v7.44.0
        version: 7.44.0
      - name: eclipse-che.v7.42.0
        version: 7.42.0
      - name: eclipse-che.v7.41.2
        version: 7.41.2
      - name: eclipse-che.v7.41.1
        version: 7.41.1
      - name: eclipse-che.v7.41.0
        version: 7.41.0
      - name: eclipse-che.v7.40.2
        version: 7.40.2
      - name: eclipse-che.v7.40.1
        version: 7.40.1
      name: stable
    defaultChannel: stable
    packageName: eclipse-che
    provider:
      name: Eclipse Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solution-Soft
      provider-url: www.solution-soft.com
    name: timemachine-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: timemachine-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tm.solution-soft.com/v1alpha1",
                "kind": "TimeMachine",
                "metadata": {
                  "name": "timemachine-sample"
                },
                "spec": {
                  "cluster_provider": "ibm",
                  "aws": {
                    "shared_volume_name": ""
                  },
                  "ibm": {
                    "persistentvolume": {
                      "storageclass": "ibmc-file-silver-gid"
                    }
                  },
                  "tm_lichost": "52.152.148.40",
                  "tm_licpass": "qwaszx",
                  "tm_licport": "57777",
                  "imagePullSecrets": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
          createdAt: "2024-01-30T10:20:30Z"
          description: Time Machine Operator enables time travel of OpenShift pods,
            without the need to change the system clock on any of the cluster nodes
          olm.skipRange: '>=0.0.4 <1.0.17'
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: SolutionSoft Systems, Inc. - support@solution-soft.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Time Machine
            displayName: Time Machine
            kind: TimeMachine
            name: timemachines.tm.solution-soft.com
            version: v1alpha1
        description: |-
          Time Machine operator enables time travel of OpenShift pods/deployments.

          To time travel a pod where your application is, you just need to install the Time Machine operator and create a Time Machine custom resource in the same namespace, without the need of changing the container images of the application you’re using, or adding containers to the pod with your application.

          Time travelling is done on the namespace level where all the (configured) target pods/deployments see the same virtual time.
        displayName: Time Machine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - time shift
        - time travel
        - virtual clocks
        - time machine
        - virtual time
        - solution-soft
        links:
        - name: Timemachine Operator
          url: https://catalog.redhat.com/software/operators/detail/5e98727a6c5dcb34dfbb19e3
        maintainers:
        - email: dtrajkovic@solution-soft.com
          name: Dragan Trajkovic
        maturity: alpha
        provider:
          name: Solution-Soft
          url: www.solution-soft.com
        relatedImages:
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
        - registry.connect.redhat.com/rhel-timemachine-container/virtual-clock-testing-tool@sha256:a81a823328b0ee795c40668c89a08bcf545c5192d9eed0146071e3721a0a8372
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0c304c11986790f446604319dfc29db494b15f5f9fa13f2c436688ec29c92dbe
        version: 1.0.17
      entries:
      - name: timemachine-operator.v1.0.17
        version: 1.0.17
      - name: timemachine-operator.v1.0.16
        version: 1.0.16
      name: alpha
    defaultChannel: alpha
    packageName: timemachine-operator
    provider:
      name: Solution-Soft
      url: www.solution-soft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: antrea.io
      provider-url: ""
    name: antrea-by-broadcom-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: antrea-by-broadcom-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.antrea.vmware.com/v1",
                "kind": "AntreaInstall",
                "metadata": {
                  "name": "antrea-install",
                  "namespace": "antrea-operator"
                },
                "spec": {
                  "antreaAgentConfig": "# FeatureGates is a map of feature names to bools that enable or disable experimental features.\nfeatureGates:\n# AllAlpha is a global toggle for alpha features. Per-feature key values override the default set by AllAlpha.\n#  AllAlpha: false\n\n# AllBeta is a global toggle for beta features. Per-feature key values override the default set by AllBeta.\n#  AllBeta: false\n\n# Enable AntreaProxy which provides ServiceLB for in-cluster Services in antrea-agent.\n# It should be enabled on Windows, otherwise NetworkPolicy will not take effect on\n# Service traffic.\n#  AntreaProxy: true\n\n# Enable EndpointSlice support in AntreaProxy. Don't enable this feature unless that EndpointSlice\n# API version v1 is supported and set as enabled in Kubernetes. If AntreaProxy is not enabled,\n# this flag will not take effect.\n#  EndpointSlice: false\n\n# Enable TopologyAwareHints in AntreaProxy. This requires AntreaProxy and EndpointSlice to be\n# enabled, otherwise this flag will not take effect.\n#  TopologyAwareHints: false\n\n# Enable traceflow which provides packet tracing feature to diagnose network issue.\n#  Traceflow: true\n\n# Enable NodePortLocal feature to make the Pods reachable externally through NodePort\n#  NodePortLocal: true\n\n# Enable Antrea ClusterNetworkPolicy feature to complement K8s NetworkPolicy for cluster admins\n# to define security policies which apply to the entire cluster, and Antrea NetworkPolicy\n# feature that supports priorities, rule actions and externalEntities in the future.\n#  AntreaPolicy: true\n\n# Enable flowexporter which exports polled conntrack connections as IPFIX flow records from each\n# agent to a configured collector.\n#  FlowExporter: false\n\n# Enable collecting and exposing NetworkPolicy statistics.\n#  NetworkPolicyStats: true\n\n# Enable controlling SNAT IPs of Pod egress traffic.\n#  Egress: true\n\n# Enable AntreaIPAM, which can allocate IP addresses from IPPools. AntreaIPAM is required by the\n# bridging mode and allocates IPs to Pods in bridging mode. It is also required to use Antrea for\n# IPAM when configuring secondary network interfaces with Multus.\n#  AntreaIPAM: false\n\n# Enable multicast traffic.\n#  Multicast: false\n\n# Enable Antrea Multi-cluster Gateway to support cross-cluster traffic.\n# This feature is supported only with encap mode.\n#  Multicluster: false\n\n# Enable support for provisioning secondary network interfaces for Pods (using\n# Pod annotations). At the moment, Antrea can only create secondary network\n# interfaces using SR-IOV VFs on baremetal Nodes.\n#  SecondaryNetwork: false\n\n# Enable managing external IPs of Services of LoadBalancer type.\n#  ServiceExternalIP: false\n\n# Enable mirroring or redirecting the traffic Pods send or receive.\n#  TrafficControl: false\n\n# Enable certificated-based authentication for IPsec.\n#  IPsecCertAuth: false\n\n# Enable collecting support bundle files with SupportBundleCollection CRD.\n#  SupportBundleCollection: false\n\n# Enable users to protect their applications by specifying how they are allowed to communicate with others, taking\n# into account application context.\n#  L7NetworkPolicy: false\n\n# Name of the OpenVSwitch bridge antrea-agent will create and use.\n# Make sure it doesn't conflict with your existing OpenVSwitch bridges.\novsBridge: \"br-int\"\n\n# Datapath type to use for the OpenVSwitch bridge created by Antrea. At the moment, the only\n# supported value is 'system', which corresponds to the kernel datapath.\n#ovsDatapathType: system\n\n# Name of the interface antrea-agent will create and use for host \u003c--\u003e pod communication.\n# Make sure it doesn't conflict with your existing interfaces.\nhostGateway: \"antrea-gw0\"\n\n# Determines how traffic is encapsulated. It has the following options:\n# encap(default):    Inter-node Pod traffic is always encapsulated and Pod to external network\n#                    traffic is SNAT'd.\n# noEncap:           Inter-node Pod traffic is not encapsulated; Pod to external network traffic is\n#                    SNAT'd if noSNAT is not set to true. Underlying network must be capable of\n#                    supporting Pod traffic across IP subnets.\n# hybrid:            noEncap if source and destination Nodes are on the same subnet, otherwise encap.\n# networkPolicyOnly: Antrea enforces NetworkPolicy only, and utilizes CNI chaining and delegates Pod\n#                    IPAM and connectivity to the primary CNI.\n#\ntrafficEncapMode: \"encap\"\n\n# Whether or not to SNAT (using the Node IP) the egress traffic from a Pod to the external network.\n# This option is for the noEncap traffic mode only, and the default value is false. In the noEncap\n# mode, if the cluster's Pod CIDR is reachable from the external network, then the Pod traffic to\n# the external network needs not be SNAT'd. In the networkPolicyOnly mode, antrea-agent never\n# performs SNAT and this option will be ignored; for other modes it must be set to false.\nnoSNAT: false\n\n# Tunnel protocols used for encapsulating traffic across Nodes. If WireGuard is enabled in trafficEncryptionMode,\n# this option will not take effect. Supported values:\n# - geneve (default)\n# - vxlan\n# - gre\n# - stt\n# Note that \"gre\" is not supported for IPv6 clusters (IPv6-only or dual-stack clusters).\ntunnelType: \"geneve\"\n\n# TunnelPort is the destination port for UDP and TCP based tunnel protocols (Geneve, VXLAN, and STT).\n# If zero, it will use the assigned IANA port for the protocol, i.e. 6081 for Geneve, 4789 for VXLAN,\n# and 7471 for STT.\ntunnelPort: 0\n\n# TunnelCsum determines whether to compute UDP encapsulation header (Geneve or VXLAN) checksums on outgoing\n# packets. For Linux kernel before Mar 2021, UDP checksum must be present to trigger GRO on the receiver for better\n# performance of Geneve and VXLAN tunnels. The issue has been fixed by\n# https://github.com/torvalds/linux/commit/89e5c58fc1e2857ccdaae506fb8bc5fed57ee063, thus computing UDP checksum is\n# no longer necessary.\n# It should only be set to true when you are using an unpatched Linux kernel and observing poor transfer performance.\ntunnelCsum: false\n\n# Determines how tunnel traffic is encrypted. Currently encryption only works with encap mode.\n# It has the following options:\n# - none (default):  Inter-node Pod traffic will not be encrypted.\n# - ipsec:           Enable IPsec (ESP) encryption for Pod traffic across Nodes. Antrea uses\n#                    Preshared Key (PSK) for IKE authentication. When IPsec tunnel is enabled,\n#                    the PSK value must be passed to Antrea Agent through an environment\n#                    variable: ANTREA_IPSEC_PSK.\n# - wireGuard:       Enable WireGuard for tunnel traffic encryption.\ntrafficEncryptionMode: \"none\"\n\n# Enable bridging mode of Pod network on Nodes, in which the Node's transport interface is connected\n# to the OVS bridge, and cross-Node/VLAN traffic of AntreaIPAM Pods (Pods whose IP addresses are\n# allocated by AntreaIPAM from IPPools) is sent to the underlay network, and forwarded/routed by the\n# underlay network.\n# This option requires the `AntreaIPAM` feature gate to be enabled. At this moment, it supports only\n# IPv4 and Linux Nodes, and can be enabled only when `ovsDatapathType` is `system`,\n# `trafficEncapMode` is `noEncap`, and `noSNAT` is true.\nenableBridgingMode: false\n\n# Disable TX checksum offloading for container network interfaces. It's supposed to be set to true when the\n# datapath doesn't support TX checksum offloading, which causes packets to be dropped due to bad checksum.\n# It affects Pods running on Linux Nodes only.\ndisableTXChecksumOffload: false\n\n# Default MTU to use for the host gateway interface and the network interface of each Pod.\n# If omitted, antrea-agent will discover the MTU of the Node's primary interface and\n# also adjust MTU to accommodate for tunnel encapsulation overhead (if applicable).\ndefaultMTU: 0\n\n# wireGuard specifies WireGuard related configurations.\nwireGuard:\n  # The port for WireGuard to receive traffic.\n  port: 51820\n\negress:\n  # exceptCIDRs is the CIDR ranges to which outbound Pod traffic will not be SNAT'd by Egresses.\n  exceptCIDRs:\n\n# ClusterIP CIDR range for Services. It's required when AntreaProxy is not enabled, and should be\n# set to the same value as the one specified by --service-cluster-ip-range for kube-apiserver. When\n# AntreaProxy is enabled, this parameter is not needed and will be ignored if provided.\nserviceCIDR: \"\"\n\n# ClusterIP CIDR range for IPv6 Services. It's required when using kube-proxy to provide IPv6 Service in a Dual-Stack\n# cluster or an IPv6 only cluster. The value should be the same as the configuration for kube-apiserver specified by\n# --service-cluster-ip-range. When AntreaProxy is enabled, this parameter is not needed.\n# No default value for this field.\nserviceCIDRv6: \"\"\n\n# The port for the antrea-agent APIServer to serve on.\n# Note that if it's set to another value, the `containerPort` of the `api` port of the\n# `antrea-agent` container must be set to the same value.\napiPort: 10350\n\n# Enable metrics exposure via Prometheus. Initializes Prometheus metrics listener.\nenablePrometheusMetrics: true\n\n# Provide the IPFIX collector address as a string with format \u003cHOST\u003e:[\u003cPORT\u003e][:\u003cPROTO\u003e].\n# HOST can either be the DNS name or the IP of the Flow Collector. For example,\n# \"flow-aggregator.flow-aggregator.svc\" can be provided as DNS name to connect\n# to the Antrea Flow Aggregator service. If IP, it can be either IPv4 or IPv6.\n# However, IPv6 address should be wrapped with [].\n# If PORT is empty, we default to 4739, the standard IPFIX port.\n# If no PROTO is given, we consider \"tls\" as default. We support \"tls\", \"tcp\" and\n# \"udp\" protocols. \"tls\" is used for securing communication between flow exporter and\n# flow aggregator.\nflowCollectorAddr: \"flow-aggregator.flow-aggregator.svc:4739:tls\"\n\n# Provide flow poll interval as a duration string. This determines how often the\n# flow exporter dumps connections from the conntrack module. Flow poll interval\n# should be greater than or equal to 1s (one second).\n# Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\", \"m\", \"h\".\nflowPollInterval: \"5s\"\n\n# Provide the active flow export timeout, which is the timeout after which a flow\n# record is sent to the collector for active flows. Thus, for flows with a continuous\n# stream of packets, a flow record will be exported to the collector once the elapsed\n# time since the last export event is equal to the value of this timeout.\n# Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\", \"m\", \"h\".\nactiveFlowExportTimeout: \"5s\"\n\n# Provide the idle flow export timeout, which is the timeout after which a flow\n# record is sent to the collector for idle flows. A flow is considered idle if no\n# packet matching this flow has been observed since the last export event.\n# Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\", \"m\", \"h\".\nidleFlowExportTimeout: \"15s\"\n\nnodePortLocal:\n# Enable NodePortLocal, a feature used to make Pods reachable using port forwarding on the host. To\n# enable this feature, you need to set \"enable\" to true, and ensure that the NodePortLocal feature\n# gate is also enabled (which is the default).\n  enable: false\n# Provide the port range used by NodePortLocal. When the NodePortLocal feature is enabled, a port\n# from that range will be assigned whenever a Pod's container defines a specific port to be exposed\n# (each container can define a list of ports as pod.spec.containers[].ports), and all Node traffic\n# directed to that port will be forwarded to the Pod.\n  portRange: \"61000-62000\"\n\n# Provide the address of Kubernetes apiserver, to override any value provided in kubeconfig or InClusterConfig.\n# Defaults to \"\". It must be a host string, a host:port pair, or a URL to the base of the apiserver.\nkubeAPIServerOverride: \"\"\n\n# Provide the address of DNS server, to override the kube-dns service. It's used to resolve hostname in FQDN policy.\n# Defaults to \"\". It must be a host string or a host:port pair of the DNS server (e.g. 10.96.0.10, 10.96.0.10:53,\n# [fd00:10:96::a]:53).\ndnsServerOverride: \"\"\n\n# Comma-separated list of Cipher Suites. If omitted, the default Go Cipher Suites will be used.\n# https://golang.org/pkg/crypto/tls/#pkg-constants\n# Note that TLS1.3 Cipher Suites cannot be added to the list. But the apiserver will always\n# prefer TLS1.3 Cipher Suites whenever possible.\ntlsCipherSuites: \"\"\n\n# TLS min version from: VersionTLS10, VersionTLS11, VersionTLS12, VersionTLS13.\ntlsMinVersion: \"\"\n\n# The name of the interface on Node which is used for tunneling or routing the traffic across Nodes.\n# If there are multiple IP addresses configured on the interface, the first one is used. The IP\n# address used for tunneling or routing traffic to remote Nodes is decided in the following order of\n# preference (from highest to lowest):\n# 1. transportInterface\n# 2. transportInterfaceCIDRs\n# 3. The Node IP\ntransportInterface: \"\"\n\nmulticast:\n# The names of the interfaces on Nodes that are used to forward multicast traffic.\n# Defaults to transport interface if not set.\n  multicastInterfaces:\n\n# The interval at which the antrea-agent sends IGMP queries to Pods.\n# Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\", \"m\", \"h\".\n  igmpQueryInterval: \"125s\"\n\n# The network CIDRs of the interface on Node which is used for tunneling or routing the traffic across\n# Nodes. If there are multiple interfaces configured the same network CIDR, the first one is used. The\n# IP address used for tunneling or routing traffic to remote Nodes is decided in the following order of\n# preference (from highest to lowest):\n# 1. transportInterface\n# 2. transportInterfaceCIDRs\n# 3. The Node IP\ntransportInterfaceCIDRs:\n\n# Option antreaProxy contains AntreaProxy related configuration options.\nantreaProxy:\n  # ProxyAll tells antrea-agent to proxy all Service traffic, including NodePort, LoadBalancer, and ClusterIP traffic,\n  # regardless of where they come from. Therefore, running kube-proxy is no longer required. This requires the AntreaProxy\n  # feature to be enabled.\n  # Note that this option is experimental. If kube-proxy is removed, option kubeAPIServerOverride must be used to access\n  # apiserver directly.\n  proxyAll: false\n  # A string array of values which specifies the host IPv4/IPv6 addresses for NodePort. Values can be valid IP blocks.\n  # (e.g. 1.2.3.0/24, 1.2.3.4/32). An empty string slice is meant to select all host IPv4/IPv6 addresses.\n  # Note that the option is only valid when proxyAll is true.\n  nodePortAddresses:\n  # An array of string values to specify a list of Services which should be ignored by AntreaProxy (traffic to these\n  # Services will not be load-balanced). Values can be a valid ClusterIP (e.g. 10.11.1.2) or a Service name\n  # with Namespace (e.g. kube-system/kube-dns)\n  skipServices:\n  # When ProxyLoadBalancerIPs is set to false, AntreaProxy no longer load-balances traffic destined to the\n  # External IPs of LoadBalancer Services. This is useful when the external LoadBalancer provides additional\n  # capabilities (e.g. TLS termination) and it is desirable for Pod-to-ExternalIP traffic to be sent to the\n  # external LoadBalancer instead of being load-balanced to an Endpoint directly by AntreaProxy.\n  # Note that setting ProxyLoadBalancerIPs to false usually only makes sense when ProxyAll is set to true and\n  # kube-proxy is removed from the cluser, otherwise kube-proxy will still load-balance this traffic.\n  proxyLoadBalancerIPs: true\n\n# IPsec tunnel related configurations.\nipsec:\n  # The authentication mode of IPsec tunnel. It has the following options:\n  # - psk (default): Use pre-shared key (PSK) for IKE authentication.\n  # - cert:          Use CA-signed certificates for IKE authentication. This option requires the `IPsecCertAuth`\n  #                  feature gate to be enabled.\n  authenticationMode: \"psk\"\n\nmulticluster:\n# Enable Antrea Multi-cluster Gateway to support cross-cluster traffic.\n# This feature is supported only with encap mode.\n  enable: false\n# The Namespace where Antrea Multi-cluster Controller is running.\n# The default is antrea-agent's Namespace.\n  namespace: \"\"\n# Enable StretchedNetworkPolicy which could be enforced on cross-cluster traffic.\n# This feature is supported only with encap mode.\n  enableStretchedNetworkPolicy: false\n",
                  "antreaCNIConfig": "{\n    \"cniVersion\":\"0.3.0\",\n    \"name\": \"antrea\",\n    \"plugins\": [\n        {\n            \"type\": \"antrea\",\n            \"ipam\": {\n                \"type\": \"host-local\"\n            }\n        }\n        ,\n        {\n            \"type\": \"portmap\",\n            \"capabilities\": {\"portMappings\": true}\n        }\n        ,\n        {\n            \"type\": \"bandwidth\",\n            \"capabilities\": {\"bandwidth\": true}\n        }\n    ]\n}\n",
                  "antreaControllerConfig": "# FeatureGates is a map of feature names to bools that enable or disable experimental features.\nfeatureGates:\n# AllAlpha is a global toggle for alpha features. Per-feature key values override the default set by AllAlpha.\n#  AllAlpha: false\n\n# AllBeta is a global toggle for beta features. Per-feature key values override the default set by AllBeta.\n#  AllBeta: false\n\n# Enable traceflow which provides packet tracing feature to diagnose network issue.\n#  Traceflow: true\n\n# Enable Antrea ClusterNetworkPolicy feature to complement K8s NetworkPolicy for cluster admins\n# to define security policies which apply to the entire cluster, and Antrea NetworkPolicy\n# feature that supports priorities, rule actions and externalEntities in the future.\n#  AntreaPolicy: true\n\n# Enable collecting and exposing NetworkPolicy statistics.\n#  NetworkPolicyStats: true\n\n# Enable multicast traffic.\n#  Multicast: false\n\n# Enable controlling SNAT IPs of Pod egress traffic.\n#  Egress: true\n\n# Run Kubernetes NodeIPAMController with Antrea.\n#  NodeIPAM: false\n\n# Enable AntreaIPAM, which can allocate IP addresses from IPPools. AntreaIPAM is required by the\n# bridging mode and allocates IPs to Pods in bridging mode. It is also required to use Antrea for\n# IPAM when configuring secondary network interfaces with Multus.\n#  AntreaIPAM: false\n\n# Enable managing external IPs of Services of LoadBalancer type.\n#  ServiceExternalIP: false\n\n# Enable certificated-based authentication for IPsec.\n#  IPsecCertAuth: false\n\n# Enable managing ExternalNode for unmanaged VM/BM.\n#  ExternalNode: false\n\n# Enable collecting support bundle files with SupportBundleCollection CRD.\n#  SupportBundleCollection: false\n\n# Enable multi-cluster features.\n#  Multicluster: false\n\n# Enable users to protect their applications by specifying how they are allowed to communicate with others, taking\n# into account application context.\n#  L7NetworkPolicy: false\n\n# The port for the antrea-controller APIServer to serve on.\n# Note that if it's set to another value, the `containerPort` of the `api` port of the\n# `antrea-controller` container must be set to the same value.\napiPort: 10349\n\n# Enable metrics exposure via Prometheus. Initializes Prometheus metrics listener.\nenablePrometheusMetrics: true\n\n# Indicates whether to use auto-generated self-signed TLS certificate.\n# If false, a Secret named \"antrea-controller-tls\" must be provided with the following keys:\n#   ca.crt: \u003cCA certificate\u003e\n#   tls.crt: \u003cTLS certificate\u003e\n#   tls.key: \u003cTLS private key\u003e\nselfSignedCert: true\n\n# Comma-separated list of Cipher Suites. If omitted, the default Go Cipher Suites will be used.\n# https://golang.org/pkg/crypto/tls/#pkg-constants\n# Note that TLS1.3 Cipher Suites cannot be added to the list. But the apiserver will always\n# prefer TLS1.3 Cipher Suites whenever possible.\ntlsCipherSuites: \"\"\n\n# TLS min version from: VersionTLS10, VersionTLS11, VersionTLS12, VersionTLS13.\ntlsMinVersion: \"\"\n\nnodeIPAM:\n  # Enable the integrated Node IPAM controller within the Antrea controller.\n  enableNodeIPAM: false\n  # CIDR ranges for Pods in cluster. String array containing single CIDR range, or multiple ranges.\n  # The CIDRs could be either IPv4 or IPv6. At most one CIDR may be specified for each IP family.\n  # Value ignored when enableNodeIPAM is false.\n  clusterCIDRs:\n  # CIDR ranges for Services in cluster. It is not necessary to specify it when there is no overlap with clusterCIDRs.\n  # Value ignored when enableNodeIPAM is false.\n  serviceCIDR: \"\"\n  serviceCIDRv6: \"\"\n  # Mask size for IPv4 Node CIDR in IPv4 or dual-stack cluster. Value ignored when enableNodeIPAM is false\n  # or when IPv4 Pod CIDR is not configured. Valid range is 16 to 30.\n  nodeCIDRMaskSizeIPv4: 24\n  # Mask size for IPv6 Node CIDR in IPv6 or dual-stack cluster. Value ignored when enableNodeIPAM is false\n  # or when IPv6 Pod CIDR is not configured. Valid range is 64 to 126.\n  nodeCIDRMaskSizeIPv6: 64\n\nipsecCSRSigner:\n  # Determines the auto-approve policy of Antrea CSR signer for IPsec certificates management.\n  # If enabled, Antrea will auto-approve the CertificateSingingRequest (CSR) if its subject and x509 extensions\n  # are permitted, and the requestor can be validated. If K8s `BoundServiceAccountTokenVolume` feature is enabled,\n  # the Pod identity will also be validated to provide maximum security.\n  # If set to false, Antrea will not auto-approve CertificateSingingRequests and they need to be approved\n  # manually by `kubectl certificate approve`.\n  autoApprove: true\n  # Indicates whether to use auto-generated self-signed CA certificate.\n  # If false, a Secret named \"antrea-ipsec-ca\" must be provided with the following keys:\n  #   tls.crt: \u003cCA certificate\u003e\n  #   tls.key: \u003cCA private key\u003e\n  selfSignedCA: true\n\nmulticluster:\n  # Enable StretchedNetworkPolicy which allow Antrea-native policies to select peers\n  # from other clusters in a ClusterSet.\n  enableStretchedNetworkPolicy: false\n",
                  "antreaImage": "antrea/antrea-ubi:v2.1.0",
                  "antreaPlatform": "openshift"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: projects.packages.broadcom.com/antreainterworking/antrea-operator:v2.1.0_vmware.3@sha256:70f129499c801f164adf8772e7ccd5d1fc659df5f298d7851e6bf02ee760f2a9
          createdAt: "2025-04-16T12:00:00Z"
          description: An operator which installs Antrea network CNI plugin on the
            Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          support: projectantrea-maintainers@googlegroups.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AntreaInstall is the Schema for the antreainstalls API
            displayName: Antrea Install
            kind: AntreaInstall
            name: antreainstalls.operator.antrea.vmware.com
            version: v1
        description: An operator which installs Antrea network CNI plugin on the Kubernetes
          cluster.
        displayName: Antrea-by-Broadcom Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        links:
        - name: Antrea Operator For Kubernetes
          url: https://github.com/vmware/antrea-operator-for-kubernetes
        maintainers:
        - email: projectantrea-maintainers@googlegroups.com
          name: Project Antrea Maintainers
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: antrea.io
        relatedImages:
        - projects.packages.broadcom.com/antreainterworking/antrea-operator:v2.1.0_vmware.3@sha256:70f129499c801f164adf8772e7ccd5d1fc659df5f298d7851e6bf02ee760f2a9
        version: 1.10.0
      entries:
      - name: antrea-by-broadcom-operator.v1.10.0
        version: 1.10.0
      name: alpha
    defaultChannel: alpha
    packageName: antrea-by-broadcom-operator
    provider:
      name: antrea.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Nokia
      provider-url: https://docs.eda.dev
    name: eda-k8s-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eda-openshift-operator.v4.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.eda.nokia.com/v1",
                "kind": "ConnectControllerConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "operator"
                  },
                  "name": "connectcontrollerconfig-sample",
                  "namespace": "eda-connect-k8s"
                },
                "spec": {
                  "connectPluginName": "",
                  "controllerImagePullSecret": "",
                  "edaNamespace": "eda",
                  "heartBeat": "",
                  "logLevel": "",
                  "skipTLSVerify": false,
                  "tlsCertificateData": "",
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:c43a7fa7475db83011cc08b728982511fc0279a259e1aacc8ddde9e15f947725
          createdAt: "2025-06-10T15:47:27Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: eda-connect-k8s
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["connectnetworkattachmentstatuses.connect.eda.nokia.com",
            connectnetworkdefinitions.connect.eda.nokia.com]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConnectNetworkAttachmentStatus is the Schema for the connectnetworkattachmentstatuses
              API
            displayName: Connect NetworkAttachment Status
            kind: ConnectNetworkAttachmentStatus
            name: connectnetworkattachmentstatuses.connect.eda.nokia.com
            version: v1
          - description: ConnectNetworkDefinition is the Schema for the connectnetworkdefinitions
              API
            displayName: Connect Network Definition
            kind: ConnectNetworkDefinition
            name: connectnetworkdefinitions.connect.eda.nokia.com
            version: v1
          - description: ConnectControllerConfig is the Schema for the connectcontrollerconfigs
              API
            displayName: Connect Controller Config
            kind: ConnectControllerConfig
            name: connectcontrollerconfigs.config.eda.nokia.com
            version: v1
        description: "K8s operator to deploy eda connect controller to configure EDA
          Connect.\eb"
        displayName: eda-openshift-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - EDA Connect
        - K8s Controller
        - EDA K8s Controller Operator
        links:
        - name: EDA Connect K8s Controller Operator
          url: https://docs.eda.dev/connect/openshift-plugin/
        maintainers:
        - email: support@nokia.com
          name: nokia
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Nokia
          url: https://docs.eda.dev
        relatedImages:
        - ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:c43a7fa7475db83011cc08b728982511fc0279a259e1aacc8ddde9e15f947725
        version: 4.0.0
      entries:
      - name: eda-openshift-operator.v4.0.0
        version: 4.0.0
      name: alpha
    - currentCSV: eda-openshift-operator.v4.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.eda.nokia.com/v1",
                "kind": "ConnectControllerConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "operator"
                  },
                  "name": "connectcontrollerconfig-sample",
                  "namespace": "eda-connect-k8s"
                },
                "spec": {
                  "connectPluginName": "",
                  "controllerImagePullSecret": "",
                  "edaNamespace": "eda",
                  "heartBeat": "",
                  "logLevel": "",
                  "skipTLSVerify": false,
                  "tlsCertificateData": "",
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:ec264b52d581b8fd76e737b9811ab894957a4edf2f627f90bd9fe707890e6bf5
          createdAt: "2025-06-10T15:47:27Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: eda-connect-k8s
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["connectnetworkattachmentstatuses.connect.eda.nokia.com",
            connectnetworkdefinitions.connect.eda.nokia.com]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConnectNetworkAttachmentStatus is the Schema for the connectnetworkattachmentstatuses
              API
            displayName: Connect NetworkAttachment Status
            kind: ConnectNetworkAttachmentStatus
            name: connectnetworkattachmentstatuses.connect.eda.nokia.com
            version: v1
          - description: ConnectNetworkDefinition is the Schema for the connectnetworkdefinitions
              API
            displayName: Connect Network Definition
            kind: ConnectNetworkDefinition
            name: connectnetworkdefinitions.connect.eda.nokia.com
            version: v1
          - description: ConnectControllerConfig is the Schema for the connectcontrollerconfigs
              API
            displayName: Connect Controller Config
            kind: ConnectControllerConfig
            name: connectcontrollerconfigs.config.eda.nokia.com
            version: v1
        description: "K8s operator to deploy eda connect controller to configure EDA
          Connect.\eb"
        displayName: eda-openshift-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - EDA Connect
        - K8s Controller
        - EDA K8s Controller Operator
        links:
        - name: EDA Connect K8s Controller Operator
          url: https://docs.eda.dev/connect/openshift-plugin/
        maintainers:
        - email: support@nokia.com
          name: nokia
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Nokia
          url: https://docs.eda.dev
        relatedImages:
        - ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:ec264b52d581b8fd76e737b9811ab894957a4edf2f627f90bd9fe707890e6bf5
        version: 4.0.1
      entries:
      - name: eda-openshift-operator.v4.0.1
        version: 4.0.1
      name: stable
    defaultChannel: alpha
    packageName: eda-k8s-operator
    provider:
      name: Nokia
      url: https://docs.eda.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-prometheus-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-prometheus-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:b6071d6a1ec7cdd5f65b6d7296c1e9cd5660972a8ed614da4277e3a4ab5d1df1
          createdAt: "2024-03-01T23:43:00Z"
          description: Manage the full lifecycle of configuring and managing Prometheus
            and Alertmanager servers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/prometheus-operator/prometheus-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Prometheus Operator for Kubernetes provides easy monitoring definitions for Kubernetes services and deployment and management of Prometheus instances.

          Once installed, the Prometheus Operator provides the following features:

          * **Create/Destroy**: Easily launch a Prometheus instance for your Kubernetes namespace, a specific application or team easily using the Operator.
          * **Simple Configuration**: Configure the fundamentals of Prometheus like versions, persistence, retention policies, and replicas from a native Kubernetes resource.
          * **Target Services via Labels**: Automatically generate monitoring target configurations based on familiar Kubernetes label queries; no need to learn a Prometheus specific configuration language.

          ### Other Supported Features

          **High availability**

          Multiple instances are run across failure zones and data is replicated. This keeps your monitoring available during an outage, when you need it most.

          **Updates via automated operations**

          New Prometheus versions are deployed using a rolling update with no downtime, making it easy to stay up to date.

          **Handles the dynamic nature of containers**

          Alerting rules are attached to groups of containers instead of individual instances, which is ideal for the highly dynamic nature of container deployment.
        displayName: Prometheus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - prometheus
        - monitoring
        - tsdb
        - alerting
        - observability
        links:
        - name: Prometheus
          url: https://www.prometheus.io/
        - name: Documentation
          url: https://github.com/prometheus-operator/prometheus-operator/tree/master/Documentation
        - name: Prometheus Operator
          url: https://github.com/prometheus-operator/prometheus-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:b6071d6a1ec7cdd5f65b6d7296c1e9cd5660972a8ed614da4277e3a4ab5d1df1
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:413f7da5cb2e78abf33eb4e43f5fd483c19e791416eec7f5d4c3dbeb54db33a0
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader-rhel9@sha256:eb7614528884f2eefe1ae5e33a09624590c23e0e75510b4d88a274880cf090bc
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager-rhel9@sha256:7d1a3392872fd3990ae3acce79da659ba2238b30421824b7be533dcf8b93e89c
        version: 4.16.16-rhodf
      entries:
      - name: odf-prometheus-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odf-prometheus-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odf-prometheus-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odf-prometheus-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odf-prometheus-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odf-prometheus-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odf-prometheus-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odf-prometheus-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odf-prometheus-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odf-prometheus-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odf-prometheus-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odf-prometheus-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odf-prometheus-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odf-prometheus-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odf-prometheus-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odf-prometheus-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odf-prometheus-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odf-prometheus-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Alvearie
      provider-url: https://alvearie.github.io
    name: alvearie-imaging-ingestion
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: imaging-ingestion-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomEventBridge",
                "metadata": {
                  "name": "events"
                },
                "spec": {
                  "dicomEventDrivenIngestionName": "core",
                  "eventBridge": {},
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "events",
                  "natsTokenSecret": "nats-events-secure-bound-token",
                  "natsUrl": "jetstream.imaging-ingestion.svc.cluster.local:4222",
                  "role": "hub"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomEventDrivenIngestion",
                "metadata": {
                  "name": "core"
                },
                "spec": {
                  "databaseConfigName": "db-config",
                  "databaseSecretName": "db-secret",
                  "eventProcessor": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  },
                  "imagePullPolicy": "Always"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomInstanceBinding",
                "metadata": {
                  "name": "stow"
                },
                "spec": {
                  "bindingConfigName": "instance-binding-config",
                  "bindingSecretName": "instance-binding-secret",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "instanceBinding": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomStudyBinding",
                "metadata": {
                  "name": "fhir"
                },
                "spec": {
                  "bindingConfigName": "study-binding-config",
                  "bindingSecretName": "study-binding-secret",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "studyBinding": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomwebIngestionService",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "bucketConfigName": "imaging-ingestion",
                  "bucketSecretName": "imaging-ingestion",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "providerName": "provider",
                  "stowService": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  },
                  "wadoService": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DimseIngestionService",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "applicationEntityTitle": "DICOM-INGEST",
                  "bucketConfigName": "imaging-ingestion",
                  "bucketSecretName": "imaging-ingestion",
                  "dicomEventDrivenIngestionName": "core",
                  "dimseService": {},
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "DIMSE",
                  "natsTokenSecret": "ingestion-nats-secure-bound-token",
                  "natsUrl": "nats-secure.imaging-ingestion.svc.cluster.local:4222",
                  "providerName": "provider"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DimseProxy",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "applicationEntityTitle": "DCM4CHEE",
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "DIMSE",
                  "natsTokenSecret": "ingestion-nats-secure-bound-token",
                  "natsUrl": "nats-secure.imaging-ingestion.svc.cluster.local:4222",
                  "proxy": {},
                  "targetDimseHost": "arc.dcm4chee.svc.cluster.local",
                  "targetDimsePort": 11112
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: alvearie/imaging-ingestion-operator:0.0.3
          createdAt: "2021-04-08T15:29:55Z"
          description: The Alvearie Imaging Ingestion provides a collection of components
            for extending enterprise medical imaging informatics systems to Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Alvearie/imaging-ingestion
          support: alvearie.slack.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DicomEventBridge is the Schema for the dicomeventbridges
              API
            displayName: Dicom Event Bridge
            kind: DicomEventBridge
            name: dicomeventbridges.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Event driven manifest of all DICOM data across all associated
              S3 buckets
            displayName: Dicom Event Driven Ingestion
            kind: DicomEventDrivenIngestion
            name: dicomeventdriveningestions.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Fan-out notification of DICOM instance data (STOW-RS)
            displayName: Dicom Instance Binding
            kind: DicomInstanceBinding
            name: dicominstancebindings.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Fan-out notification of DICOM studies (FHIRv4 ImagingStudy)
            displayName: Dicom Study Binding
            kind: DicomStudyBinding
            name: dicomstudybindings.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides DICOMweb WADO-RS and STOW-RS access to a S3 bucket
            displayName: Dicomweb Ingestion Service
            kind: DicomwebIngestionService
            name: dicomwebingestionservices.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides a proxied DIMSE Application Entity (AE) in the cluster
              for C-STORE operations to a storage space
            displayName: Dimse Ingestion Service
            kind: DimseIngestionService
            name: dimseingestionservices.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides a bidirectional proxied DIMSE Application Entity
              (AE) in the cluster
            displayName: Dimse Proxy
            kind: DimseProxy
            name: dimseproxies.imaging-ingestion.alvearie.org
            version: v1alpha1
        description: |
          The Alvearie Imaging Ingestion provides a collection of components for extending enterprise medical imaging informatics systems to Kubernetes.

          ## About this Operator
          This operator enables standards-based medical imaging application interfaces such as DICOM Message Service Element (DIMSE) and DICOMweb.
          The API's provided allow for extending enterprise imaging systems to the cloud with custom, fit for purpose deployments.
          This is possible at speed, low cost, and little or zero integration effort.or extending enterprise medical imaging informatics systems to a modern cloud.


          Some potential usages:

          - Perform lightweight secure bridging of DICOM from the enterprise imaging zone to a cloud availability zone.
          - Utilize S3 buckets for segregating DICOM data to be used for clinical or scientific purposes.
          - Fan-out DICOM data to a number of different DICOM subscribers
          - Generate FHIRv4 *ImagingStudy* resources from ingested content and publish to a FHIRv4 based patient logitudinal record.
          - Distribute imaging study arrival and update notifications to imaging study subscribers


          ## Prerequisites

          - [Knative](https://operatorhub.io/operator/knative-operator)
          - ProgresSQL such as [Crunchy](https://operatorhub.io/operator/postgresql)
          - S3 storage such as [MinIO](https://operatorhub.io/operator/minio-operator)
          - [NATS operator](https://github.com/nats-io/nats-operator/) (Optional:  only required for DIMSE Proxy component)



          ## Documentation

          Visit the [official github repository](https://github.com/Alvearie/imaging-ingestion) for documentation and examples.
        displayName: Alvearie Imaging Ingestion Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - dicom
        - medical imaging
        - fhir
        - dimse-c
        - dimse
        - imaging study
        links:
        - name: Alvearie Imaging Ingestion Operator
          url: https://github.com/Alvearie/imaging-ingestion
        maintainers:
        - email: rduggan@ca.ibm.com
          name: Richard Duggan
        - email: jjacob@us.ibm.com
          name: Jeesmon Jacob
        maturity: alpha
        provider:
          name: Alvearie
          url: https://alvearie.github.io
        relatedImages:
        - alvearie/imaging-ingestion-operator:0.0.3
        version: 0.0.3
      entries:
      - name: imaging-ingestion-operator.v0.0.3
        version: 0.0.3
      - name: imaging-ingestion-operator.v0.0.2
        version: 0.0.2
      - name: imaging-ingestion-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: alvearie-imaging-ingestion
    provider:
      name: Alvearie
      url: https://alvearie.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solace Corporation
      provider-url: www.solace.com
    name: pubsubplus-eventbroker-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pubsubplus-eventbroker-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pubsubplus.solace.com/v1beta1",
                "kind": "PubSubPlusEventBroker",
                "metadata": {
                  "name": "non-ha-standalone-example"
                },
                "spec": {
                  "redundancy": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "true"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.10
          containerImage: registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator@sha256:d68db92bcb217d94dc0ad0a1cea625a0a9c9299d886d325a903163f9f887dcd5
          createdAt: "2025-07-31T19:48:36Z"
          description: The Solace PubSub+ Event Broker Operator deploys and manages
            the lifecycle of PubSub+ Event Brokers
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
          support: Solace Products
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PubSub+ Event Broker
            displayName: PubSub+ Event Broker
            kind: PubSubPlusEventBroker
            name: pubsubpluseventbrokers.pubsubplus.solace.com
            version: v1beta1
        description: |
          ## Solace PubSub+ Platform
          Solace [PubSub+ Platform](https://solace.com/products/platform/) is a complete event streaming and management platform
          for real-time enterprises. The [PubSub+ Software Event Broker](https://solace.com/products/event-broker/software/) efficiently streams
          event-driven information between applications, IoT devices, and user interfaces running in the cloud, on-premises,
          and in hybrid environments using open APIs and protocols like AMQP, JMS, MQTT, REST and WebSocket.
          It can be installed into a variety of public and private clouds, PaaS, and on-premises environments.
          Event brokers in multiple locations can be linked together in an [Event Mesh](https://solace.com/what-is-an-event-mesh/)
          to dynamically share events across the distributed enterprise.

          ## Solace PubSub+ Event Broker Operator
          Solace Pubsub+ Event Broker Operator automatically deploys and manages PubSub+ Software Event Brokers on Kubernetes and OpenShift environments.

          Features include:
          * Initial deployment configuration: Production-ready HA, non-HA or for developers
          * Rolling update for configuration changes or broker upgrades
          * Prometheus monitoring support

          ## Getting started
          Refer to the [Solace PubSub+ Event Broker Operator Quick Start](https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/README.md)
        displayName: Solace PubSub+ Event Broker Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - solace
        - pubsubplus
        - pubsub+
        - pubsub
        - messaging
        - advanced event broker
        - event broker
        - event mesh
        - event streaming
        - data streaming
        - event integration
        - middleware
        links:
        - name: Operator QuickStart
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
        - name: Operator Documentation
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/docs/EventBrokerOperatorUserGuide.md
        - name: Solace.com
          url: https://solace.com
        - name: Understanding Solace Technology
          url: https://solace.com/products/platform/
        - name: Solace PubSub+ Technical Documentation
          url: https://docs.solace.com/
        - name: Solace Developer Portal
          url: https://www.solace.dev/
        - name: Solace Community
          url: https://solace.community/
        maintainers:
        - email: support@solace.com
          name: Solace R&D
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Solace Corporation
          url: www.solace.com
        relatedImages:
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator@sha256:d68db92bcb217d94dc0ad0a1cea625a0a9c9299d886d325a903163f9f887dcd5
        version: 1.4.0
      entries:
      - name: pubsubplus-eventbroker-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: pubsubplus-eventbroker-operator
    provider:
      name: Solace Corporation
      url: www.solace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: StreamNative
      provider-url: https://streamnative.io
    name: function-mesh
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: function-mesh.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "java-function-sample",
                  "namespace": "default"
                },
                "spec": {
                  "autoAck": true,
                  "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                  "clusterName": "test-pulsar",
                  "forwardSourceMessageProperty": true,
                  "input": {
                    "topics": [
                      "persistent://public/default/java-function-input-topic"
                    ],
                    "typeClassName": "java.lang.String"
                  },
                  "java": {
                    "extraDependenciesDir": "random-dir/",
                    "jar": "pulsar-functions-api-examples.jar",
                    "jarLocation": "public/default/nlu-test-java-function"
                  },
                  "logTopic": "persistent://public/default/logging-function-logs",
                  "maxPendingAsyncRequests": 1000,
                  "maxReplicas": 5,
                  "output": {
                    "topic": "persistent://public/default/java-function-output-topic",
                    "typeClassName": "java.lang.String"
                  },
                  "pod": {
                    "annotations": {
                      "managed-function": "true"
                    },
                    "env": [
                      {
                        "name": "EXAMPLE_VARIABLE",
                        "value": "exampleValue"
                      }
                    ],
                    "imagePullSecrets": [
                      {
                        "name": "regcred"
                      }
                    ],
                    "initContainers": [
                      {
                        "command": [
                          "sh",
                          "-c",
                          "echo The app is running! \u0026\u0026 sleep 30"
                        ],
                        "image": "busybox:1.28",
                        "name": "init-function"
                      }
                    ],
                    "labels": {
                      "locaction": "mtv"
                    },
                    "sidecars": [
                      {
                        "command": [
                          "sh",
                          "-c",
                          "echo The app is running! \u0026\u0026 sleep 30000"
                        ],
                        "image": "busybox:1.28",
                        "name": "sidecar-function"
                      }
                    ],
                    "volumes": [
                      {
                        "emptyDir": {},
                        "name": "cache-volume"
                      }
                    ]
                  },
                  "pulsar": {
                    "authSecret": "test-auth",
                    "pulsarConfig": "test-pulsar",
                    "tlsSecret": "test-tls"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "secretsMap": {
                    "name": {
                      "key": "username",
                      "path": "test-secret"
                    },
                    "pwd": {
                      "key": "password",
                      "path": "test-secret"
                    }
                  },
                  "volumeMounts": [
                    {
                      "mountPath": "/cache",
                      "name": "cache-volume"
                    }
                  ]
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "FunctionMesh",
                "metadata": {
                  "name": "functionmesh-sample"
                },
                "spec": {
                  "functions": [
                    {
                      "autoAck": true,
                      "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                      "clusterName": "test-pulsar",
                      "forwardSourceMessageProperty": true,
                      "input": {
                        "topics": [
                          "persistent://public/default/functionmesh-input-topic"
                        ],
                        "typeClassName": "java.lang.String"
                      },
                      "java": {
                        "jar": "pulsar-functions-api-examples.jar",
                        "jarLocation": "public/default/nlu-test-functionmesh-ex1"
                      },
                      "logTopic": "persistent://public/default/logging-function-log",
                      "maxReplicas": 1,
                      "name": "ex1",
                      "output": {
                        "topic": "persistent://public/default/mid-topic",
                        "typeClassName": "java.lang.String"
                      },
                      "pulsar": {
                        "pulsarConfig": "mesh-test-pulsar"
                      },
                      "replicas": 1,
                      "resources": {
                        "limits": {
                          "cpu": "0.2",
                          "memory": "1.1G"
                        },
                        "requests": {
                          "cpu": "0.1",
                          "memory": "1G"
                        }
                      }
                    },
                    {
                      "autoAck": true,
                      "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                      "clusterName": "test-pulsar",
                      "forwardSourceMessageProperty": true,
                      "input": {
                        "topics": [
                          "persistent://public/default/mid-topic"
                        ],
                        "typeClassName": "java.lang.String"
                      },
                      "java": {
                        "jar": "pulsar-functions-api-examples.jar",
                        "jarLocation": "public/default/nlu-test-functionmesh-ex2"
                      },
                      "logTopic": "persistent://public/default/logging-function-logs",
                      "maxReplicas": 1,
                      "name": "ex2",
                      "output": {
                        "topic": "persistent://public/default/functionmesh-output-topic",
                        "typeClassName": "java.lang.String"
                      },
                      "pulsar": {
                        "pulsarConfig": "mesh-test-pulsar"
                      },
                      "replicas": 1,
                      "resources": {
                        "limits": {
                          "cpu": "0.2",
                          "memory": "1.1G"
                        },
                        "requests": {
                          "cpu": "0.1",
                          "memory": "1G"
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Sink",
                "metadata": {
                  "name": "sink-sample"
                },
                "spec": {
                  "autoAck": true,
                  "className": "org.apache.pulsar.io.elasticsearch.ElasticSearchSink",
                  "clusterName": "test-pulsar",
                  "image": "streamnative/pulsar-io-elastic-search:2.10.0.0-rc10",
                  "input": {
                    "topics": [
                      "persistent://public/default/input"
                    ],
                    "typeClassName": "[B"
                  },
                  "java": {
                    "jar": "connectors/pulsar-io-elastic-search-2.10.0.0-rc10.nar",
                    "jarLocation": ""
                  },
                  "maxReplicas": 1,
                  "pulsar": {
                    "pulsarConfig": "test-sink"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "sinkConfig": {
                    "elasticSearchUrl": "http://quickstart-es-http.default.svc.cluster.local:9200",
                    "indexName": "my_index",
                    "password": "wJ757TmoXEd941kXm07Z2GW3",
                    "typeName": "doc",
                    "username": "elastic"
                  }
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Source",
                "metadata": {
                  "name": "source-sample"
                },
                "spec": {
                  "className": "org.apache.pulsar.io.debezium.mongodb.DebeziumMongoDbSource",
                  "clusterName": "test-pulsar",
                  "forwardSourceMessageProperty": true,
                  "image": "streamnative/pulsar-io-debezium-mongodb:2.10.0.0-rc10",
                  "java": {
                    "jar": "connectors/pulsar-io-debezium-mongodb-2.10.0.0-rc10.nar",
                    "jarLocation": ""
                  },
                  "maxReplicas": 1,
                  "output": {
                    "producerConf": {
                      "maxPendingMessages": 1000,
                      "maxPendingMessagesAcrossPartitions": 50000,
                      "useThreadLocalProducers": true
                    },
                    "topic": "persistent://public/default/destination",
                    "typeClassName": "org.apache.pulsar.common.schema.KeyValue"
                  },
                  "pulsar": {
                    "pulsarConfig": "test-source"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "sourceConfig": {
                    "database.whitelist": "inventory",
                    "mongodb.hosts": "rs0/mongo-dbz-0.mongo.default.svc.cluster.local:27017,rs0/mongo-dbz-1.mongo.default.svc.cluster.local:27017,rs0/mongo-dbz-2.mongo.default.svc.cluster.local:27017",
                    "mongodb.name": "dbserver1",
                    "mongodb.password": "dbz",
                    "mongodb.task.id": "1",
                    "mongodb.user": "debezium",
                    "pulsar.service.url": "pulsar://test-pulsar-broker.default.svc.cluster.local:6650"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/function-mesh@sha256:3e13e5cc3ac4d7387ee0adc02cb5183aad6cf2301f76ab230482609c4da1c0b7
          createdAt: "2022-10-27T07:49:28Z"
          description: The Function Mesh Operator manages the Pulsar Functions and
            Connectors deployed on a Kubernetes cluster.
          olm.properties: '[{"type":"olm.maxOpenShiftVersion","value":"4.8"}]'
          olm.skipRange: <0.7.0
          operatorhub.io/ui-metadata-max-k8s-version: "1.21"
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/streamnative/function-mesh
          support: StreamNative
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FunctionMesh is the Schema for the functionmeshes API
            displayName: Mesh
            kind: FunctionMesh
            name: functionmeshes.compute.functionmesh.io
            version: v1alpha1
          - description: Function is the Schema for the functions API
            displayName: Function
            kind: Function
            name: functions.compute.functionmesh.io
            version: v1alpha1
          - description: Sink is the Schema for the sinks API
            displayName: Sink
            kind: Sink
            name: sinks.compute.functionmesh.io
            version: v1alpha1
          - description: Source is the Schema for the sources API
            displayName: Source
            kind: Source
            name: sources.compute.functionmesh.io
            version: v1alpha1
        description: |
          [Function Mesh](https://functionmesh.io/) is a serverless framework purpose-built for stream processing applications. It brings powerful event-streaming capabilities to your applications by orchestrating multiple [Pulsar Functions](http://pulsar.apache.org/docs/en/next/functions-overview/) and [Pulsar IO connectors](http://pulsar.apache.org/docs/en/next/io-overview/) for complex stream processing jobs on Kubernetes.
          Prerequisites:
          - Install cert-manager operator first.
        displayName: Function Mesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pulsar
        - function-mesh
        - serverless
        - pulsar-functions
        - pulsar-io-connectors
        - StreamNative
        links:
        - name: Function Mesh
          url: https://function-mesh.io
        maintainers:
        - email: function-mesh@streamnative.io
          name: Function Mesh
        maturity: alpha
        minKubeVersion: v1.17.0
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/function-mesh@sha256:3e13e5cc3ac4d7387ee0adc02cb5183aad6cf2301f76ab230482609c4da1c0b7
        - docker.cloudsmith.io/streamnative/mirrors/gcr.io/kubebuilder/kube-rbac-proxy@sha256:67ecb332573384515406ebd71816781366b70adb0eb66345e5980e92603373e1
        version: 0.7.0
      entries:
      - name: function-mesh.v0.7.0
        version: 0.7.0
      name: alpha
    defaultChannel: alpha
    packageName: function-mesh
    provider:
      name: StreamNative
      url: https://streamnative.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.3.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo-operator@sha256:32178a16dd2115e0a11a2e3dfbfc130ae1d0afd8e293131e76b6fc1a79b36c12
          createdAt: "2019-05-01T00:00:00Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/kubeturbo-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/kubeturbo-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo-operator@sha256:32178a16dd2115e0a11a2e3dfbfc130ae1d0afd8e293131e76b6fc1a79b36c12
        - registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo@sha256:671be694176d9cc31b8a22c11bbfefc2504ae72d27fd010b315b6839f9ab59d5
        version: 8.3.3
      entries:
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: stable
    defaultChannel: stable
    packageName: kubeturbo-certified-rhmp
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: devspaces
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devspacesoperator.v3.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "org.eclipse.che/v1",
                "kind": "CheCluster",
                "metadata": {
                  "name": "devspaces",
                  "namespace": "openshift-devspaces"
                },
                "spec": {
                  "auth": {
                    "identityProviderURL": "",
                    "oAuthClientName": "",
                    "oAuthSecret": ""
                  },
                  "database": {
                    "externalDb": false
                  },
                  "k8s": {
                    "ingressDomain": null,
                    "tlsSecretName": null
                  },
                  "metrics": {
                    "enable": true
                  },
                  "server": {
                    "workspaceNamespaceDefault": "<username>-devspaces"
                  },
                  "storage": {
                    "pvcStrategy": "common"
                  }
                }
              },
              {
                "apiVersion": "org.eclipse.che/v2",
                "kind": "CheCluster",
                "metadata": {
                  "name": "devspaces",
                  "namespace": "openshift-devspaces"
                },
                "spec": {
                  "components": {},
                  "containerRegistry": {},
                  "devEnvironments": {},
                  "gitServices": {},
                  "networking": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/devspaces/devspaces-rhel9-operator@sha256:39c95279bced68928d028ef026f82eafe53eddb3dbe57703c8007cb5a170a70a
          createdAt: "2025-09-11T22:30:49+00:00"
          description: Devfile v2 and v1 development solution, 1 instance per cluster,
            for portable, collaborative k8s workspaces.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/devspaces-images/
          skipRange: '>=3.0.0 <3.23.0'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              The `CheCluster` custom resource allows defining and managing Red Hat OpenShift Dev Spaces server installation.
              Based on these settings, the  Operator automatically creates and maintains several ConfigMaps:
              `che`, `plugin-registry` that will contain the appropriate environment variables
              of the various components of the installation. These generated ConfigMaps must NOT be updated manually.
            displayName: Red Hat OpenShift Dev Spaces instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v2
          - description: The `CheCluster` custom resource allows defining and managing
              a Che server installation
            displayName: Red Hat OpenShift Dev Spaces instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v1
        description: |
          A collaborative Kubernetes-native development solution that delivers OpenShift workspaces and in-browser IDE for rapid cloud application development.
          This operator installs the Plugin and Devfile registries, Dashboard, Gateway and the Red Hat OpenShift Dev Spaces server, and configures these services.
          OpenShift OAuth is used directly for authentication. TLS mode is on.

          ## How to Install
          Press the **Install** button, choose the channel and the upgrade strategy, and wait for the **Installed** Operator status.
          When the operator is installed, create a new Custom Resource (CR) of Kind CheCluster (click the **Create New** button).
          The CR spec contains all defaults. You can start using Red Hat OpenShift Dev Spaces when the CR status is set to **Available**, and you see a URL to Red Hat OpenShift Dev Spaces.

          ## Defaults
          By default, the operator deploys Red Hat OpenShift Dev Spaces with:
          * 10Gi storage
          * Auto-generated passwords
          * Bundled Plugin and Devfile registries

          Use `oc edit checluster/devspaces -n openshift-devspaces` to update Red Hat OpenShift Dev Spaces default installation options.
          See more in the [Installation guide](https://docs.redhat.com/en/documentation/red_hat_openshift_dev_spaces/3.23/html/administration_guide/configuring-devspaces).

          ### Certificates
          Operator uses a default router certificate to secure Red Hat OpenShift Dev Spaces routes.
          Follow the [guide](https://docs.redhat.com/en/documentation/red_hat_openshift_dev_spaces/3.23/html/administration_guide/configuring-devspaces#importing-untrusted-tls-certificates)
          to import certificates into Red Hat OpenShift Dev Spaces.
        displayName: Red Hat OpenShift Dev Spaces
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eclipse che
        - workspaces
        - devtools
        - developer
        - ide
        - java
        - devspaces
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/openshift-dev-spaces/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_openshift_dev_spaces
        - name: Operator GitHub Repo
          url: https://github.com/redhat-developer/devspaces-images/
        maintainers:
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/devspaces/pluginregistry-rhel9@sha256:3ee131041761f5250b29d497043ecf7820e02e03e73ed0f929c17e2247c0d7a3
        - registry.redhat.io/devspaces/traefik-rhel9@sha256:95cf6a28028ec7a546a3d8d78377f204bb559c3d1022a73f08c646e76b85ce85
        - registry.redhat.io/devspaces-tech-preview/jetbrains-ide-rhel9@sha256:46bf7bea716df74f14d2629fb7e0e2ab692564cec348ffb8a6f4fea1739b23c2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5a6edfecf1636805e6f0bcad02df10bceec764f0580f3ac8fe62aba978d50006
        - registry.redhat.io/devspaces/server-rhel9@sha256:c0ec8c792f88aacf6ee31a4b6d2d847d864707a084148b48e4e402388989f868
        - registry.redhat.io/devspaces/devspaces-rhel9-operator@sha256:39c95279bced68928d028ef026f82eafe53eddb3dbe57703c8007cb5a170a70a
        - registry.redhat.io/devspaces/code-rhel9@sha256:abc9b12131c8cdca0719e0404d354310f53769990b089f1a3adea31193798a8e
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6bcfa981a7a75c242a16afa6dab36bf6e3b8a1442fa863875919ad551f48eceb
        - registry.redhat.io/devspaces/configbump-rhel9@sha256:cecbc3796058618da9ad0473a6f937aae5aca25776bb434894286f22baa4677c
        - registry.redhat.io/devspaces/dashboard-rhel9@sha256:c34cf5ac9bf95c1c5082afd5a2941b5e412be8ca43aef6a72e8fe5a279a43271
        - registry.redhat.io/devspaces/udi-rhel9@sha256:f7325ee7259a0585e8017d4442d778a37f4c8a67d81fe70c241a014001e37539
        - registry.redhat.io/devspaces-tech-preview/idea-rhel9@sha256:df24ebb239a2cc0d4543309f51a7268c7d524d428a536e0b3443c51089c3f87e
        version: 3.23.0
      entries:
      - name: devspacesoperator.v3.23.0
        version: 3.23.0
      - name: devspacesoperator.v3.22.0
        version: 3.22.0
      - name: devspacesoperator.v3.21.0
        version: 3.21.0
      - name: devspacesoperator.v3.20.0
        version: 3.20.0
      - name: devspacesoperator.v3.19.0
        version: 3.19.0
      - name: devspacesoperator.v3.18.1-0.1739475915.p
        version: 3.18.1+0.1739475915.p
      - name: devspacesoperator.v3.18.1
        version: 3.18.1
      - name: devspacesoperator.v3.18.0
        version: 3.18.0
      - name: devspacesoperator.v3.17.0
        version: 3.17.0
      - name: devspacesoperator.v3.16.1
        version: 3.16.1
      - name: devspacesoperator.v3.16.0
        version: 3.16.0
      - name: devspacesoperator.v3.15.0
        version: 3.15.0
      - name: devspacesoperator.v3.14.0
        version: 3.14.0
      - name: devspacesoperator.v3.13.0
        version: 3.13.0
      - name: devspacesoperator.v3.12.0-0.1712003563.p
        version: 3.12.0+0.1712003563.p
      - name: devspacesoperator.v3.12.0
        version: 3.12.0
      - name: devspacesoperator.v3.11.0
        version: 3.11.0
      - name: devspacesoperator.v3.10.1
        version: 3.10.1
      - name: devspacesoperator.v3.10.0
        version: 3.10.0
      - name: devspacesoperator.v3.9.1
        version: 3.9.1
      - name: devspacesoperator.v3.9.0
        version: 3.9.0
      - name: devspacesoperator.v3.8.0
        version: 3.8.0
      - name: devspacesoperator.v3.7.2
        version: 3.7.2
      - name: devspacesoperator.v3.7.1-0.1691407091.p
        version: 3.7.1+0.1691407091.p
      - name: devspacesoperator.v3.7.1
        version: 3.7.1
      - name: devspacesoperator.v3.7.1-0.1691548593.p
        version: 3.7.1+0.1691548593.p
      - name: devspacesoperator.v3.7.0
        version: 3.7.0
      - name: devspacesoperator.v3.6.0
        version: 3.6.0
      - name: devspacesoperator.v3.6.0-0.1685320856.p
        version: 3.6.0+0.1685320856.p
      - name: devspacesoperator.v3.6.0-0.1685322634.p
        version: 3.6.0+0.1685322634.p
      - name: devspacesoperator.v3.5.0
        version: 3.5.0
      - name: devspacesoperator.v3.4.0-0.1675914643.p
        version: 3.4.0+0.1675914643.p
      - name: devspacesoperator.v3.4.0
        version: 3.4.0
      - name: devspacesoperator.v3.3.0-0.1673528130.p
        version: 3.3.0+0.1673528130.p
      - name: devspacesoperator.v3.3.0
        version: 3.3.0
      name: stable
    defaultChannel: stable
    packageName: devspaces
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Jooho Lee
      provider-url: https://github.com/jooho/nfs-provisioner-operator
    name: nfs-provisioner-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfs-provisioner-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.jhouse.com/v1alpha1",
                "kind": "NFSProvisioner",
                "metadata": {
                  "name": "nfsprovisioner-sample"
                },
                "spec": {
                  "nfsImageConfiguration": {
                    "image": "k8s.gcr.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439",
                    "imagePullPolicy": "IfNotPresent"
                  },
                  "scForNFS": "nfs",
                  "scForNFSPvc": "local-sc",
                  "storageSize": "1G"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: quay.io/jooholee/nfs-provisioner-operator@sha256:d9c013967421ec72644a588a155975bf856a856f1ef38ba71dfea306fdb47acd
          createdAt: "2025-08-08T20:23:50Z"
          description: Create and manage NFS Server and Provisioner
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/jooho/nfs-provisioner-operator
          support: NFS
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NFSProvisioner is the Schema for the nfsprovisioners API
            displayName: NFS Provisioner App
            kind: NFSProvisioner
            name: nfsprovisioners.cache.jhouse.com
            version: v1alpha1
        description: "This operator deploy NFS server with local storage and also
          provide provisioner for storageClass.\n### Core Capabilities\n* **NFS Server:**
          Deployed\n* **NFS Provisioner:** Help customers to create PV using StorageClass\n*
          **StorageClass:** Dynamically create PV for requested PVC\n### NFS Provisioner
          Operator Features\n* NFS Server can use localStorage PVC or HostPath on
          the node\n### Must-Gather \n> isv-cli must-gather --image=quay.io/jooholee/nfs-provisioner-operator-must-gather:0.0.1\n"
        displayName: NFS Provisioner Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nfs
        - storage
        links:
        - name: Nfs Provisioner Operator
          url: https://github.com/jooho/nfs-provisioner-operator
        maintainers:
        - email: ljhiyh@gmail.com
          name: Jooho
        maturity: alpha
        provider:
          name: Jooho Lee
          url: https://github.com/jooho/nfs-provisioner-operator
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - quay.io/jooholee/nfs-provisioner-operator@sha256:d9c013967421ec72644a588a155975bf856a856f1ef38ba71dfea306fdb47acd
        version: 0.0.8
      entries:
      - name: nfs-provisioner-operator.v0.0.8
        version: 0.0.8
      - name: nfs-provisioner-operator.v0.0.7
        version: 0.0.7
      - name: nfs-provisioner-operator.v0.0.6
        version: 0.0.6
      - name: nfs-provisioner-operator.v0.0.5
        version: 0.0.5
      - name: nfs-provisioner-operator.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: nfs-provisioner-operator
    provider:
      name: Jooho Lee
      url: https://github.com/jooho/nfs-provisioner-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-licensing
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-licensing.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Licensing",
                "metadata": {
                  "name": "licensing-sample"
                },
                "spec": {
                  "environment": {
                    "deploymentType": "openshift"
                  },
                  "hostid": {
                    "enabled": true,
                    "environment": {},
                    "hostMounts": {},
                    "image": "prod.harbor.keysight.digital/loadcore/kcos-licensing-hostid-vmuuid@sha256:076a2c6aeec875470414af7905613a25b56bcb2aca5edcb1c4c53037cd01b332",
                    "imagePullPolicy": "IfNotPresent"
                  },
                  "image": {
                    "dbimage": "prod.harbor.keysight.digital/loadcore/postgresql-ubi-image@sha256:17f6a8f775d06650939d2c73e82b45d73d50190cc735ca70695d082f7790834b",
                    "initOwnership": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf",
                    "repository": "prod.harbor.keysight.digital/loadcore"
                  },
                  "ingress": {
                    "annotations": {
                      "nginx.ingress.kubernetes.io/auth-method": "POST",
                      "nginx.ingress.kubernetes.io/auth-response-headers": "X-AUTH-USER",
                      "nginx.ingress.kubernetes.io/auth-url": "https://127.0.0.1/token/validate"
                    },
                    "enabled": true
                  },
                  "nodeSelector": {
                    "node-role.kubernetes.io/kcos-licensing": "kcos-licensing"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:29:04Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Licensing
            name: licensings.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Licensing Operator
        displayName: Keysight KCOS Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Licensing
          url: https://keysight-kcos-licensing.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-licensing@sha256:a09291923ad8d7533b33e40e1a3c4e735c2a4ae3b2b87be7c9d1a1a9244b3dad
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - prod.harbor.keysight.digital/loadcore/postgresql-ubi-image@sha256:17f6a8f775d06650939d2c73e82b45d73d50190cc735ca70695d082f7790834b
        - prod.harbor.keysight.digital/loadcore/kcos-licensing-hostid-vmuuid@sha256:076a2c6aeec875470414af7905613a25b56bcb2aca5edcb1c4c53037cd01b332
        version: 0.1.21
      entries:
      - name: keysight-kcos-licensing.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-licensing
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: multicluster-engine
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-engine.v2.5.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-08-06T12:21:10Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0 <2.5.10'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '[]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine is the Schema for the multiclusterengines
              API
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:80f33a762535ff69f57ecb04de043b8fe249a63767712f4942ac6685450e8f28
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:1af88c3703b668020137f90657e6262b73e407aa858bbc6d9c12ba3a1cab8043
        - registry.redhat.io/multicluster-engine/agent-service-rhel8@sha256:92464c57700dcb13ccf5431c104a3fba6845f651776aa5beb1ce7171087dc1ea
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:cb2c57c51edc8a2320a10ac217cca6a401db09de55c0c9d0c9a7b4545e2636e9
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:7f32ed997e03dba1746b3627d5a33fc69282b2e831edd21ec74020fe4fa1ca5d
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:90e792c92ef8c75ffb97545394206c9197997d474493d93f09fca273f65133e4
        - registry.redhat.io/multicluster-engine/hive-rhel8@sha256:27eefe6ad7d93e752f4d76b776884129615c2e87ff24116c401a886e98abfcc5
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:1e9eff48ea837a1d19c595986d765d74810d6d1397602ad79e2e820b1cbb5c6e
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:297d8d0d5855db90ad40af702d6ad5f3f32a4aab486686919d86bbd21c367ca9
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:b0a70eee404ba1267c0ae1a8eb140251d054e722305d680685223cfda360aa37
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:53b8d48e540c3d9bb5d948af0375c26377271e595fffdb3d5f921e3a358fa298
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:36abf06619cd3971da8f42f4193621a1384970365fce0f48e86a65e2840bca38
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel8@sha256:360d3dfb823d8e199574cf1c3fbf890fa74c71fc48644e22b47339363e941b3e
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:c8d48e7ccdce43ce72154e464e74c32477efb0094aab327ed0eaab6c1244774a
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:515eda1e4e1ced222709103789af14f7b4e99d4da037deddd8499aad3b45451b
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:3e3f7ee898e0f6b0f74026399753d7fde41013ade76f7963d5c25e578eb1906e
        - registry.redhat.io/rhel8/postgresql-12@sha256:da619f6617d99790b4fe00fdb23ab4aea0cdb0592590162d51b3ef94aca01a7b
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel8@sha256:6bf850d8fb250740113bfb004b19edbd5cf560c8175b81e2f3766280b3482e27
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:43cc171404571bb3c87275a86e9ab4409861c1c01b5f1441c455565d56e1921d
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:d71c1441577b1cd360d074581ad19c2b0cfa99cdfe1869af1777ff7de6821735
        - registry.redhat.io/multicluster-engine/cluster-api-provider-aws-rhel9@sha256:7f65119dc5d81b8ba485b651efe1d9013b62ec57e4725dbf8bdc2e3b7748f677
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:1a5ff852d96bd15c631a6f89e2b0047cd8dde6264646f51106f2b523decc64ba
        - registry.redhat.io/multicluster-engine/aws-encryption-provider-rhel9@sha256:22cefc43e205e4bdd7874c7e1fccba60fe556df4b2bab055d01c834d62f194bd
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:53ad9402a0a13c894b485f418dd1353c8870792893a5ce33f9d6c3cc9839ebc4
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:2497e077a00dfa94b913359be0c696126e932db1a45e066ead65dee76a1db8fd
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel8@sha256:1e755b64358c10309804a3bf44ca640a7d5bb6e0474d33774a8d76095f2b8c3d
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:25b30d4f4ccc23fda707db88969c0ae036ecb4a2baf15b82221e3351c23292f9
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:a4f58b94d45c1ae1c4ba69b12030b9945b37f3ce6e96c27919f256ae887ec28b
        - registry.redhat.io/multicluster-engine/cluster-api-rhel9@sha256:0e329ca39b1f18520da093a5194ef6e1d5527270c74bf260517620577a21b6c7
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:19dba8be2295f54a93853dbe566132011f0120b2c4b48de2816509f5d201e420
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:bd6d0189ea2bbc87b5745f9d7f9c82118f8dcc9ee63fe0a36bb5e7dfc0e56021
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:424ca238d9c8b20dc434a07aa6a27fe6e7c63b3e2480eead284570864d682175
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:3d420ba8bc5e1dbe629c0d4aa65c0a79a7914ac349828046994e245f4f2adb2b
        - registry.redhat.io/multicluster-engine/assisted-installer-reporter-rhel8@sha256:5d41ca13f4ae102760b65375578c2f632aa6ce41092732300d6727f195da3d55
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:460d37a0032813aa85a3b18bd6f152e4ed6b89f802d17d787c3365082420d379
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:d7ccc8ddbeb11ac8a3a0d4946920a0ab87d9e9595d7c61740bd57f3f4a19a711
        version: 2.5.10
      entries:
      - name: multicluster-engine.v2.5.10
        version: 2.5.10
      - name: multicluster-engine.v2.5.9
        version: 2.5.9
      - name: multicluster-engine.v2.5.8
        version: 2.5.8
      - name: multicluster-engine.v2.5.7
        version: 2.5.7
      - name: multicluster-engine.v2.5.6
        version: 2.5.6
      - name: multicluster-engine.v2.5.5
        version: 2.5.5
      - name: multicluster-engine.v2.5.4
        version: 2.5.4
      - name: multicluster-engine.v2.5.3
        version: 2.5.3
      - name: multicluster-engine.v2.5.2
        version: 2.5.2
      - name: multicluster-engine.v2.5.1
        version: 2.5.1
      - name: multicluster-engine.v2.5.0-103
        version: 2.5.0-103
      name: stable-2.5
    - currentCSV: multicluster-engine.v2.6.8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-08-04T11:16:16Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0 <2.6.8'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:69a40339bf3b3dd3c3b6359721895313147a5f682b62faf5151288ba63b65812
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:f6b15f827aeabc33074de1ff9ed20edaf04271b4e4d2eb16abbe786c78e033ca
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:9e1ea96b998b0c7854ca3c7c8a505fc96cbc1b35e6ce9f49b1dbec703fdc8070
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:7dde47f7fe16a26676a6e5b1739954d76cad1cddbefce2d6869ffb2af399e403
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:f563ff0acb4ea9d1bb112765b67bee535063304a98a2aa2cb7f21b0155d49262
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:54b8e17cacae2d355d82677ed334b4981b50e825579ff30406039933833e828b
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:a8117c99b33eedc702baeab32f0347892b587dfc1860f467cb1900565c3fdb3b
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:00ea8fe4945e93d2cbeceda4440acd6a53a48b4dd420dfe2cbd5287e20f0362f
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel8@sha256:42fee1066ff00066a172f82754ad9f08e8c29029e75f76712cb283d9fe5d1d9a
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:1d098eab26421f99ef71ffce82e4fa59ef0f9043de04ee74ca43486cbdb82006
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:b39819aa916974f4a408d23177e99cfb71e21af8ff9b0eacf6c37159c354ba70
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:0737e57ae8af6a62ac580710276d2c1620ac1b1283d3b84369db43e76e518610
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:5f9d5bdfd45f931989d069dcb1dea91a69020d83011fe0846ddfe22bba14b001
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel8@sha256:a630a20a633d057e475b4c39b342512119a71021e199aa0becf80899d8840ad7
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:c3c654be0b20cac501af1e31631464c32aecdbd16cbfb88f865900c70ccf77c2
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:7c5f339e762db1714b3f6cf1fc1bc8f08892cac433ab5a46e73b530fa08486fb
        - registry.redhat.io/multicluster-engine/cluster-api-rhel9@sha256:bd4eb05427ba2ea35f01227600cebd0fe8598d4de456348c24bb5e88d9ed8720
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:aff06db527e4ddc713a2fb8885824f0482647cbdcd844b311ccc3880a0dd2c3f
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:5ef8acc612817d478cede3b1bda55abafac00d0debebd68d948d157ee5b09286
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:294928c58b33ae8e776f86df211dac3b63a2cd9fc00aa995a0de037a29cb2c7d
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:2d807348774b8c6c0255c4bc931a956fc987c961fae6be689835410e8e301fd0
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel8@sha256:4d96379ad99d299d2a89eb1b0eb8a22db8bdada9a16248503cd8a9a5900eab35
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:d2c1d90a4e829b181fdf3d2838764c8f2b61c22b761279f53cb22316ad46ca03
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:93d65bfa0102c7ad7ec6ebfaad9ab062a31098b7ad7026313bcd8c6a495d5ca7
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:cd23f237c91dd8585946b31e17e38193f2e5ac8993d65459bb703a68f6fa8ca0
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:80a6615500949cc8c7a5b2a1e8c8c5f4564e9cdb9f995787a44f462e5bbd62cb
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel8@sha256:a703f2b1f4eb538aa02dbc0f4ff793506076480e94905c367b0d4ae352b87b65
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:5d488c8a644f2bfce97444d770516de10ceecc988cb3e9fbf0072d0b0a2eeb83
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:be1aa4cf5e09914774dcf69ab7ccdaf67504309d40c7e8cf4d91e9c1080e706f
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:82c0af2d285ba943dfa1da9f0856b26ec7e4c2c431577fcd4ab031ae040cfaea
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:87f5db53ad18f6d84501a5c2294cff73a5425276cdbde92225909826091b5ffa
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:46c4ac0cacef587b6e678b6cb0f68c29f5ac63bd59209d42a3ae0b7807e99972
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:d7c8bef4c3dec710ccb299482a10bfdf6efacd4e2ed2e6e323d448ed505000b3
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:d71ef80913d4e34389bce56c4a282cf7d2be705f003575dab87c07ca71f8c2d6
        version: 2.6.8
      entries:
      - name: multicluster-engine.v2.6.8
        version: 2.6.8
      - name: multicluster-engine.v2.6.7
        version: 2.6.7
      - name: multicluster-engine.v2.6.6
        version: 2.6.6
      - name: multicluster-engine.v2.6.5
        version: 2.6.5
      - name: multicluster-engine.v2.6.4
        version: 2.6.4
      - name: multicluster-engine.v2.6.3
        version: 2.6.3
      - name: multicluster-engine.v2.6.2
        version: 2.6.2
      - name: multicluster-engine.v2.6.1
        version: 2.6.1
      - name: multicluster-engine.v2.6.0
        version: 2.6.0
      name: stable-2.6
    - currentCSV: multicluster-engine.v2.7.5
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-21T22:32:21Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0 <2.7.5'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:44b07a18116ad827c23ca5a98b9fe8ef74a99abe4787cd01d848de42df44178d
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:15c2873703afcc74e41ae27dbb4761ad19c4bd2594ca6de82d3b6c02752e0ce3
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:816712bb3ccb58b705a85e5f4a55e519316fd8b9d7d9295d1042420ce44a4bad
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:e6a5e71f0aee8a682179cfd8f85b45a2ee04a1de680fc03dabdb659a2d2c0396
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:2a78ab192a554e78f3fc367c47e571f96b650ee9fb5ba0d06674f340f4320d94
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:9d1be5eb4fa357fe6bd97a52a53b3eeb537c431daf7af8a30efc3e0b6c5d5d66
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:2e3fe5e3b39721cc77b7d32b5682300eb60b1bc6940b8e83662b458d88489140
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:a24e9855a11d879d5308e29af6e3b2947365b7e466e7fa12dc57997140753604
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:b124242a55b1b2a2258cb12094617a0e009cf1cf58a60ab1b42d48d4ea2514ac
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:b66f2f4b5e0d77d501f9cc9da483f976f3c0eaf0ca72a5e9c9435c92d58e1388
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:00d415071f4c10575f9e5068868a7f80b0c0fb9c8614da86bc515923bed136a3
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:e2e16ff6213c0274126c9d64d495c0abbe17379e841ab730de66520908f1a190
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:4a0f1d3e696310f7a5bccbb358c538023a28f5bbc23b5743d10fd61b452d45d0
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:041044c08a1d526b349b5ad5fc4c4457f3632b1353530ed925c65ef2cce8e9c5
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:2025e424c11c6fe4a4395945c2e7d124d07cde43bb8d8cbe70286fe6ae345754
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:8c420720a48ea7c3aa096ae1929dcaa979c39e7865c5d43cba899e294745fc94
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:081151ca906b0f1afbcd3ee9c6d9e5bfe234494223c660ef376aa4004b0508f4
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:3defab1abc7395c9f9c22912ec6825be11b120a1416404eb0599e91f7f9c679a
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:c0fb7985620d5bd0dfadd57cb75664ee6777b8dac5a3463967e9b649e9eb836c
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:9e5c66ffdf39ae923ecccf79be538b883f2bdc8019169d55057565f077997d67
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:4fa34ab0aa8854eb5e942f2ed315041bf2026ce0bed86f57125878af0c73dc8b
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:7369bc79a283ff8b34c3189ea83e55f88a00d619cc78f8dad03d302dcb34a550
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:fedc699c0f786fb303d3956769f9c5fca1473ec4bf014ec87c5a4d40e34dcdb3
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:b82b1837dfc24a367f68741dece6d19c89c576706f006ee8d5123a0c6782c492
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:90c19d2aaa50e7cc9184f2f220b568c34c9fdcfdcba314da555885c999078057
        - registry.redhat.io/multicluster-engine/cluster-api-rhel9@sha256:cd8d6e0056fb0436275c12cc6b42f11c0ccd06729c7108e964b03d6c48d063bd
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:0684bcde2c018ae893bd71be8df3448afa09d199415e83285d2c740cff66192d
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:2216b015943eb8d9f25bf4ac37fbad2b7215a5c0003bd7c3b3d97b3b0a5ba708
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:555fcfce226e3b059f8f911d44a48578b898e085b6ad2575479feee71ec8baaf
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:a6736c7cac00f78c0e8877549ae2b74b7207cc348f24f175133db49fd38f240a
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:8e41d489e1def9b540ddd6387a1cc151b24e91a5fa44a74fbbaa0d1978588d57
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:45571dbb55bd213400eec6d2065b40615185bf3e52efdf495ab73201ec94faff
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:1c176688f13052783662115d5cabe8d614b8d9f79507dc2bdd4be1604e2eecc0
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:2cbd8c470dc4b263fd11952aa792094a7550f4b99e10f22186d51621bc8d3b3f
        version: 2.7.5
      entries:
      - name: multicluster-engine.v2.7.5
        version: 2.7.5
      - name: multicluster-engine.v2.7.4
        version: 2.7.4
      - name: multicluster-engine.v2.7.3
        version: 2.7.3
      - name: multicluster-engine.v2.7.2
        version: 2.7.2
      - name: multicluster-engine.v2.7.1
        version: 2.7.1
      - name: multicluster-engine.v2.7.0
        version: 2.7.0
      name: stable-2.7
    - currentCSV: multicluster-engine.v2.8.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-09-08T13:08:30Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0 <2.8.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:f869555beec0875c74f30a88428ac9641029dfed0365d45b23d5cae469cf7d0f
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:7ca7e87615e12b9d7fa75befd05f4f45bdc4bf8ef1b576c69f36a23c66c90cd6
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:988eb4aa01ba0d7e6728b23d043e4c0babe872ac23b6520f4daac714619a360b
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:f809a8774b0e3f891af46931dd4d2c11b95da4e7f5a8aee1814152d9a468d357
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:c5ac848e3b7eb88cdbdbf0b5f4aa96698d852b3a8118cf7146eb7e1948afaeba
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:a7f63077799a62fe6ece76a12346a87b5cbae966ab69df6ec6d951259dbf0e4a
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:c3b62ab1234257d30f41fa1ab33cb319321777a18de8ed751be1fcd48b204f2e
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:318ca0c7e9cd9a7336178519cadbf21eaeb474cd4a8eb5cf411a640d1c77372e
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:5ea86c20383a6932c593e020fd3b7b1a2a8e666bc374bff8b4bca0ee2d4ecc33
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:aa2754d35c8460a2879d4d523ebf32351d36e8fd7886335b5c0d2ca4befb8e8e
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:bee8f0c047cf8dbfa3f28712ce91bea55e926b2edb26c394f18a2d780e142083
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:7c86c89ff57091d4112783d82d1add47f2d56dba584418d5664c198ba048e0c2
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:f708f3e1007ed2b0e420e51d8b7ff9dd9bccaea9234755f4375f5514018211b4
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:3788b2674dda0b2cb4e9ca21bdce0b9e44702d72458b2396ef0acdad37fa866b
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:18b0e86305afe2eae3c26e637755fa9e8ab2abf30bc15e5428467a9a81ec99b0
        - registry.redhat.io/openshift4/ose-aws-cluster-api-controllers-rhel9@sha256:ada8b480a86d0a9e3ecc58f8b1cb53dbd7452b2b9183b3ba2b15a6a58c4436d1
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:f694b6e9e76941b11cc8a15153d9005cffe9bbbf58d0962f8ff546e90a63d93c
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:806b6c956f7bad824f0fe9acc3b534964e8f5f3399a1813f75511bc0f3904414
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:cf457b5a13cdd54c45a7b88f0f150a04e756bfac727c4365cb12f113d904937d
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:5461643a61850d95409d989e81c304053372f1dc5dbba291f3f3386de9ecab1c
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:261b99b93d98825ade0a0d3b9c4e465a785f35d2a310dedd4bbd6e4aa1f33ef9
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:0392c81a404be343d658dbc64153cbc46a64d0adf69aa8024c3f727351102d75
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:8379827635dfc4b51e04145d8048f7dcf9c6956ea1a22371bd22f948ae01b183
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:ee014c576263e19798df98460ebb5f1a21cb6249f09fe8a267b2dfa98b07d1eb
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:d4d11d0aeb3e1c9ff2a80e769d92dce509a0202a2ef1c9d2430030acdbf85bce
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:83a82f9558c406899e573308947c15b1f36de454006bb0e2ac44709f0c2c28d7
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:ea2ba066abd170e1087558c7b701faeb1be2e1667943dc14440f19c2ee5b0c30
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:fb2b8cef9916f387d26e65d9b53c7567ae1433d43d16e879ddb582ddc248eaa3
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:5ad96dad74c6f4967bd5fd397d6697e01b3b25bb3b8bae123a38e96964ddfa17
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:b9afa4d5328f17c7c3ce1dc105d177ecacf3447092d51a50f250a4b0ea692e2c
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:c6cd578123c1c86fcb3305ecca117543106f943cbcc4baf9065683988b2e10ff
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:e5e459e665afc14a1540e300015222b7c1ac896e2bee889283d67c3e19ca01ac
        - registry.redhat.io/openshift4/ose-cluster-api-rhel9@sha256:ec58463239eaa9bd15b31cbcfea0a34fd650201bac6120aa7ddaf72734451e06
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:ab0898d513b9b5ca2f5ca5879adc4e73da3232d515c0647a2a0ec28792264db1
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:0d1dfcb06eaa26de74994891371a5657aded7a59972f006f8c4a3859369877f3
        version: 2.8.3
      entries:
      - name: multicluster-engine.v2.8.3
        version: 2.8.3
      - name: multicluster-engine.v2.8.2
        version: 2.8.2
      - name: multicluster-engine.v2.8.1
        version: 2.8.1
      - name: multicluster-engine.v2.8.0
        version: 2.8.0
      name: stable-2.8
    - currentCSV: multicluster-engine.v2.9.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-30T19:22:43Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0 <2.9.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/open-cluster-management/backplane-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational components for managing multiple Kubernetes-based clusters. The deployment of the multicluster engine components is
              determined based on the configuration that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:8901e984b3b47f0650cddc1c36bdd89af22390f3d4a562d562fde7774d8ba6db
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:f3d65f68538e94260fad54f145433a36366f8c88f9c9f80b7fc5545ba62e1773
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:0a9c0ccb6a75273474fb04b785aa8fc58aec2f9ffbff8c232d359e14892bbae2
        - registry.redhat.io/multicluster-engine/capoa-control-plane-rhel9@sha256:b510a25a6370ca6b8d2f74165814ebfdae0ce9898cc343e73fab6344eebf2bed
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:8db1386d12d40dfbc80054c023897e5bfc0a987f6d07792b0f9f50f59f1a68d8
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:57699ecb529cb46843de847deef3e3757ca252910249d8deac16fcb00c01b636
        - registry.redhat.io/multicluster-engine/capoa-bootstrap-rhel9@sha256:a8b1278a2df9b17cee66813715995385bb3539af05e4df9b5bf7d9ccd508c313
        - registry.redhat.io/multicluster-engine/mce-capi-webhook-config-rhel9@sha256:7f1f70813e6d753b4acb04c7e9d7acb0047ad91f8a9a10e0fe2fff6c4411c8b4
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:a91276fcf5a589ef11f01b39df7d33ceb58ebcb47f95dfab772114bce7c1b4a8
        - registry.redhat.io/openshift4/ose-cluster-api-rhel9@sha256:9a72b8081a57846d942981ced07da86ad8e197309a1bc3c9d53876528b58f8ec
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:6a75df76fb0d9516b13fa6b079bc6c4d49dc6774f0d4f31343abfb8d70edcc23
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:ce0fd7bdc73cd2c84d1a9c5af528916488ca8116dd05d83dacf72f2af4248e65
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:ac43feaf079fa0d250d53cd9685fb68677d888dd091730faa32757dcdf2c95e4
        - registry.redhat.io/rhel8/postgresql-12@sha256:4773a29755ea826c0d3b58a8b8297c1d7388019516195cae9bd815f28ec4bca0
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:5e6594464bf7ddc2a66f127f645d93ad35d8cd95dd70544fa7fcec791f67f45c
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:36a6000b6482966f8bbf86e45c25335c066736ae1f8695a8b68e6b76b7c491ac
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:1b218f0aa0ea1f494094e3bc783ef40968cbff780e22de186d578a1d64c33bf8
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:3c450cb6a3ba6f2e3028856f14967803c1a691d95b8f54bd9d29d93691d6e795
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:8076db8ccd5578cb5aba2d66f53790563d3f42b93a6cfd14c7094d3a25707982
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:ad382d5eb347c448c46495a0a17503d789db4eb1096ad71a588b316b235e437f
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:6b95560a6a6dc1ddd41c76f275f83baec52ec9d79de04bfe54feaed02db123a3
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:ee75e805f0f9d73f073ce870f5d46962c0033d63158017841d17410febfc2c57
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:25ec9a0cc4ac92d4bed75c38fa37ef8d5ff9478cc475049b303d74eb58687a81
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:f8c3898e29f8c0a20c3be92bc4ee5a9443b9fc8218db95ba541fe3e57a89c40d
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:bf343592bf2f4399b5ac61709f56528435990e903812e264cbef845ff4433231
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:9555fd5f577f32a1d8ebe25f0d03e453c8ea925e4a12dbbe1015c58d3c8528b0
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:1cc6fddaf9e55637e90d5c3e598467df8967b8a362f8b8d46286b256bad6dd3f
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:a41cafc65e30deb073e051542702e258901b552ad00e42eacab05d149d6f0565
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:ea57247cf11f6ab966893ce4e6d2948232a4314588af2ca33c1d9689da7482dd
        - registry.redhat.io/openshift4/ose-baremetal-cluster-api-controllers-rhel9@sha256:f855d5ce17d3ac4560afc73c35d03f045f84551fdccc00312c890288a889b9ce
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:72a4ac73bb5e7797329e1856f785751d48815a64e82350a5d694bb5a8e121956
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:cd9dc89b933b394b08fe05ab5f0bbecea727bfb2c29770f11ec1db4aa5740b3c
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:37087ad8678f3e8ec8e2b2d035d3a12087ce9e0c993194a04a9874fd07e80d91
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:cecec1f7d7f0d37cf17fb25ccf34602728f2a09c61af7e212967b4ee28d2dec3
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:d068d73f41d8f8ecbec36752d0a432913bfbe2580f489abefe513c06233e80c9
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:edbb174cb69dd31339a53542d24c4fd00552e41fa09d7e2553baa7a1ad8dd810
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:1bbd134d44839258172e013fe84bfd174d36d6b7232e2cc8a3d8556998181d9d
        - registry.redhat.io/openshift4/ose-aws-cluster-api-controllers-rhel9@sha256:cf13b0658271cd0dbd1ce6dcc350e64c5459ef02ba4dc88a0031de9d6f26ffd1
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:8d189b565e4c3c434988aa4ff0ecad4c5f86b2a94120325c142ff111b25facde
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:902c5b3b4a4e8cb0a14fa5aba8451ec02830d5167ff1b91323ded62255983cd9
        version: 2.9.0
      entries:
      - name: multicluster-engine.v2.9.0
        version: 2.9.0
      name: stable-2.9
    defaultChannel: stable-2.9
    packageName: multicluster-engine
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JohnSnowLabs
      provider-url: ""
    name: nlp-server
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nlp-server.v0.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.johnsnowlabs.com/v1",
                "kind": "Nlpserver",
                "metadata": {
                  "name": "nlpserver-sample"
                },
                "spec": {
                  "ExtraEnv": {
                    "johnsnowlabs_client_id": "sI4MKSmLHOX2Pg7XhM3McJS2oyKG5PHcp0BlANEW",
                    "nb_uid": "1000",
                    "nb_user": "johnsnowlabs",
                    "pyspark_version": "3.0.2",
                    "spark_nlp_version": "3.4.4"
                  },
                  "affinity": {},
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "johnsnowlabs/nlp-server",
                    "tag": "0.8.0rc1"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {
                      "kubernetes.io/ingress.class": "nginx",
                      "nginx.ingress.kubernetes.io/ssl-redirect": "true"
                    },
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "nlp-server.jsl.aiplatform.eks.test.johnsnowlabs.com",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": [
                      {
                        "hosts": [
                          "nlp-server.jsl.aiplatform.eks.test.johnsnowlabs.com"
                        ],
                        "secretName": "annotationlab-tls-secret"
                      }
                    ]
                  },
                  "livenessProbe": "httpGet:\n  path: /\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 3\n",
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "readinessProbe": "httpGet:\n  path: /\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 3\n",
                  "replicaCount": 1,
                  "resources": {},
                  "securityContext": {},
                  "service": {
                    "port": 5000,
                    "portName": "http",
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": "nlp-server-sa"
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.johnsnowlabs.com/nlp-server/nlp-server-operator:0.8.6
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nlpserver
            name: nlpservers.apps.johnsnowlabs.com
            version: v1
        description: NLPServer from JohnSnow labs
        displayName: nlp-server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nlp-server
        links:
        - name: Nlp Server
          url: https://johnsnowlabs.com
        maturity: alpha
        provider:
          name: JohnSnowLabs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - registry.johnsnowlabs.com/nlp-server/nlp-server-operator:0.8.6
        version: 0.8.6
      entries:
      - name: nlp-server.v0.8.6
        version: 0.8.6
      name: alpha
    defaultChannel: alpha
    packageName: nlp-server
    provider:
      name: JohnSnowLabs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator-certified.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator-certified.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator-certified.v2.6.1
        version: 2.6.1
      - name: percona-postgresql-operator-certified.v2.6.0
        version: 2.6.0
      - name: percona-postgresql-operator-certified.v2.5.0
        version: 2.5.0
      name: stable
    - currentCSV: percona-postgresql-operator-certified.v2.7.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        version: 2.7.0-cw
      entries:
      - name: percona-postgresql-operator-certified.v2.7.0-cw
        version: 2.7.0-cw
      - name: percona-postgresql-operator-certified.v2.6.1-cw
        version: 2.6.1-cw
      - name: percona-postgresql-operator-certified.v2.6.0-cw
        version: 2.6.0-cw
      - name: percona-postgresql-operator-certified.v2.5.0-cw
        version: 2.5.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-postgresql-operator-certified
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Starburst Data, Inc.
      provider-url: https://www.starburstdata.com
    name: starburst-enterprise-helm-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: starburst-enterprise-helm-operator.v402.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstEnterprise",
                "metadata": {
                  "name": "starburstenterprise-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "cache": {
                    "diskUsagePercentage": 80,
                    "enabled": false,
                    "ttl": "7d",
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "catalogs": {
                    "tpch": "connector.name=tpch\ntpch.splits-per-node=4"
                  },
                  "commonLabels": {},
                  "coordinator": {
                    "additionalProperties": "",
                    "affinity": {},
                    "deploymentAnnotations": {},
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=true\nnode-scheduler.include-coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://localhost:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "resources": {
                      "cpu": 16,
                      "memory": "60Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "tolerations": []
                  },
                  "environment": null,
                  "expose": {
                    "clusterIp": {
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      }
                    },
                    "ingress": {
                      "annotations": {},
                      "host": null,
                      "ingressClassName": null,
                      "ingressName": "coordinator-ingress",
                      "path": "/",
                      "pathType": "ImplementationSpecific",
                      "serviceName": "starburst",
                      "servicePort": 8080,
                      "tls": {
                        "enabled": true,
                        "secretName": null
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "nodePort": 30080,
                          "port": 8080
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "externalSecrets": {
                    "enabled": false,
                    "goDaddy": {
                      "backendType": "secretsManager"
                    },
                    "secretPrefix": "external0",
                    "type": "goDaddy"
                  },
                  "extraArguments": [],
                  "extraSecret": {
                    "file": null,
                    "name": null
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise",
                    "tag": "402-e.1"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "initImage": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise-init",
                    "tag": "1.3.2"
                  },
                  "internal": {
                    "ports": {
                      "http": {
                        "port": 8080
                      },
                      "https": {
                        "port": 8443
                      }
                    }
                  },
                  "internalTls": false,
                  "licenseManager": {
                    "cronjob": {
                      "affinity": {},
                      "nodeSelector": {},
                      "priorityClassName": null,
                      "securityContext": {},
                      "tolerations": []
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "registry.connect.redhat.com/starburst/license-manager",
                      "tag": "0.0.2"
                    },
                    "url": "https://license-api.dogfood.eng.starburstdata.net/test/license/checkout"
                  },
                  "livenessProbe": null,
                  "prometheus": {
                    "agent": {
                      "config": "/etc/starburst/telemetry/prometheus.yaml",
                      "port": 8081,
                      "version": "0.16.1"
                    },
                    "enabled": true,
                    "rules": [
                      {
                        "attrNameSnakeCase": true,
                        "name": "$1",
                        "pattern": "trino.execution<name=QueryManager><>(running_queries|queued_queries)",
                        "type": "GAUGE"
                      },
                      {
                        "name": "failed_queries",
                        "pattern": "trino.execution<name=QueryManager><>FailedQueries\\.TotalCount",
                        "type": "COUNTER"
                      }
                    ]
                  },
                  "query": {
                    "maxConcurrentQueries": 3
                  },
                  "readinessProbe": null,
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "securityContext": {},
                  "sharedSecret": null,
                  "siac": {
                    "additionalHeapSizePercentage": 15,
                    "enabled": false
                  },
                  "spilling": {
                    "enabled": false,
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "startupProbe": null,
                  "usageMetrics": {
                    "enabled": false,
                    "usageClient": {
                      "initialDelay": "1m",
                      "interval": "1m"
                    }
                  },
                  "userDatabase": {
                    "enabled": false,
                    "users": [
                      {
                        "password": "46991b33f7a75ff79213c0dc0e610610",
                        "username": "admin"
                      }
                    ]
                  },
                  "worker": {
                    "additionalProperties": "",
                    "affinity": {},
                    "autoscaling": {
                      "enabled": false,
                      "maxReplicas": 100,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "deploymentAnnotations": {},
                    "deploymentTerminationGracePeriodSeconds": 300,
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://coordinator:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "kedaScaler": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.connect.redhat.com/starburst/keda-trino-scaler",
                        "tag": "0.1.7"
                      },
                      "port": 8021,
                      "scaledObjectSpec": {
                        "advanced": {},
                        "cooldownPeriod": 300,
                        "idleReplicaCount": 0,
                        "maxReplicaCount": 100,
                        "minReplicaCount": 1,
                        "pollingInterval": 30,
                        "scaleTargetRef": {
                          "name": "worker"
                        },
                        "triggers": [
                          {
                            "metadata": {
                              "numberOfQueriesPerWorker": "10",
                              "scaleInToIdleReplicaCountIfNoQueuedQueriesLeft": "true",
                              "scaleMethod": "query_queue"
                            },
                            "type": "external"
                          }
                        ]
                      }
                    },
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "replicas": 2,
                    "resources": {
                      "cpu": 16,
                      "memory": "100Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "starburstWorkerShutdownGracePeriodSeconds": 120,
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstHive",
                "metadata": {
                  "name": "starbursthive-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "affinity": {},
                  "commonLabels": {},
                  "database": {
                    "external": {
                      "driver": null,
                      "jdbcUrl": null,
                      "password": null,
                      "user": null
                    },
                    "internal": {
                      "databaseName": "hive",
                      "databasePassword": "HivePass1234",
                      "databaseUser": "hive",
                      "driver": "org.postgresql.Driver",
                      "env": [],
                      "envFrom": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.redhat.io/rhel8/postgresql-10",
                        "tag": "1"
                      },
                      "port": 5432,
                      "resources": {
                        "limits": {
                          "cpu": 2,
                          "memory": "1Gi"
                        },
                        "requests": {
                          "cpu": 2,
                          "memory": "1Gi"
                        }
                      },
                      "volume": {
                        "emptyDir": {}
                      }
                    },
                    "type": "internal"
                  },
                  "deploymentAnnotations": {},
                  "env": [],
                  "envFrom": [],
                  "expose": {
                    "clusterIp": {
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "nodePort": 30083,
                          "port": 9083
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "extraArguments": [],
                  "hdfs": {
                    "hadoopUserName": null
                  },
                  "heapSizePercentage": 85,
                  "hiveMetastoreStorageAuthorization": true,
                  "hiveMetastoreWarehouseDir": null,
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/hive",
                    "tag": "3.1.2-e.22"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "nodeSelector": {},
                  "objectStorage": {
                    "adl": {
                      "oauth2": {
                        "clientId": null,
                        "credential": null,
                        "refreshUrl": null
                      }
                    },
                    "awsS3": {
                      "accessKey": null,
                      "endpoint": null,
                      "pathStyleAccess": false,
                      "region": null,
                      "secretKey": null
                    },
                    "azure": {
                      "abfs": {
                        "accessKey": {
                          "accessKey": null,
                          "storageAccount": null
                        },
                        "authType": "accessKey",
                        "oauth": {
                          "clientId": null,
                          "endpoint": null,
                          "secret": null
                        }
                      },
                      "wasb": {
                        "accessKey": null,
                        "storageAccount": null
                      }
                    },
                    "gs": {
                      "cloudKeyFileSecret": null
                    }
                  },
                  "podAnnotations": {},
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "resources": {
                    "limits": {
                      "cpu": 1,
                      "memory": "1Gi"
                    },
                    "requests": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "securityContext": {},
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
          description: Deploys StarburstEnterprise Helm Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Starburst Data
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: StarburstEnterprise
            name: starburstenterprises.charts.starburstdata.com
            version: v1alpha1
          - kind: StarburstHive
            name: starbursthives.charts.starburstdata.com
            version: v1alpha1
        description: CRs for deploying starburst enterprise and hive
        displayName: Starburst Enterprise Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - starburst
        - trino
        - sep
        - bigdata
        - data
        links:
        - name: Starburst Enterprise Helm Operator
          url: https://www.starburstdata.com
        maintainers:
        - email: hello@starburstdata.com
          name: Starburst Data, Inc.
        maturity: alpha
        provider:
          name: Starburst Data, Inc.
          url: https://www.starburstdata.com
        relatedImages:
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
        version: 402.1.0
      entries:
      - name: starburst-enterprise-helm-operator.v402.1.0
        version: 402.1.0
      name: alpha
    defaultChannel: alpha
    packageName: starburst-enterprise-helm-operator-rhmp
    provider:
      name: Starburst Data, Inc.
      url: https://www.starburstdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: keystone-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keystone-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keystone.openstack.org/v1beta1",
                "kind": "KeystoneAPI",
                "metadata": {
                  "name": "keystone"
                },
                "spec": {
                  "adminProject": "admin",
                  "adminUser": "admin",
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "keystone",
                  "databaseInstance": "openstack",
                  "preserveJobs": false,
                  "region": "regionOne",
                  "resources": {
                    "requests": {
                      "cpu": "1.0",
                      "memory": "500Mi"
                    }
                  },
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "keystone.openstack.org/v1beta1",
                "kind": "KeystoneEndpoint",
                "metadata": {
                  "name": "placement"
                },
                "spec": {
                  "endpoints": {
                    "admin": "http://placement-admin-openstack.apps-crc.testing",
                    "internal": "http://placement-internal-openstack.apps-crc.testing",
                    "public": "http://placement-public-openstack.apps-crc.testing"
                  },
                  "serviceName": "placement"
                }
              },
              {
                "apiVersion": "keystone.openstack.org/v1beta1",
                "kind": "KeystoneService",
                "metadata": {
                  "name": "placement"
                },
                "spec": {
                  "enabled": true,
                  "passwordSelector": "PlacementPassword",
                  "secret": "osp-secret",
                  "serviceDescription": "Placement service",
                  "serviceName": "placement",
                  "serviceType": "placement",
                  "serviceUser": "placement"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:01Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KeystoneAPI is the Schema for the keystoneapis API
            displayName: Keystone API
            kind: KeystoneAPI
            name: keystoneapis.keystone.openstack.org
            version: v1beta1
          - description: KeystoneEndpoint is the Schema for the keystoneendpoints
              API
            displayName: Keystone Endpoint
            kind: KeystoneEndpoint
            name: keystoneendpoints.keystone.openstack.org
            version: v1beta1
          - description: KeystoneService is the Schema for the keystoneservices API
            displayName: Keystone Service
            kind: KeystoneService
            name: keystoneservices.keystone.openstack.org
            version: v1beta1
        description: Keystone Operator
        displayName: Keystone Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Identity
        - Keystone
        links:
        - name: Keystone Operator
          url: https://github.com/openstack-k8s-operators/keystone-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-keystone-rhel9@sha256:740e92a78d1e3874d2a40e37c970c6163d9e295dbc236914f1204baed070c595
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/keystone-rhel9-operator@sha256:8b055945a3808f7397265a10e6a6463af0ee3ef80fdff9d66402b488e5b70523
        version: 1.0.6
      entries:
      - name: keystone-operator.v1.0.6
        version: 1.0.6
      - name: keystone-operator.v1.0.5
        version: 1.0.5
      - name: keystone-operator.v1.0.4
        version: 1.0.4
      - name: keystone-operator.v1.0.3
        version: 1.0.3
      - name: keystone-operator.v1.0.2
        version: 1.0.2
      - name: keystone-operator.v1.0.1
        version: 1.0.1
      - name: keystone-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: keystone-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: ""
    name: ingress-node-firewall
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ingress-node-firewall.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ingressnodefirewall.openshift.io/v1alpha1",
                "kind": "IngressNodeFirewall",
                "metadata": {
                  "name": "ingressnodefirewall-demo2"
                },
                "spec": {
                  "ingress": [
                    {
                      "rules": [
                        {
                          "action": "Deny",
                          "order": 10,
                          "protocolConfig": {
                            "icmp": {
                              "icmpType": 8
                            },
                            "protocol": "ICMP"
                          }
                        },
                        {
                          "action": "Deny",
                          "order": 20,
                          "protocolConfig": {
                            "protocol": "TCP",
                            "tcp": {
                              "ports": "8000-9000"
                            }
                          }
                        }
                      ],
                      "sourceCIDRs": [
                        "172.16.0.0/12"
                      ]
                    },
                    {
                      "rules": [
                        {
                          "action": "Deny",
                          "order": 10,
                          "protocolConfig": {
                            "icmpv6": {
                              "icmpType": 128
                            },
                            "protocol": "ICMPv6"
                          }
                        }
                      ],
                      "sourceCIDRs": [
                        "fc00:f853:ccd:e793::0/64"
                      ]
                    }
                  ],
                  "interfaces": [
                    "eth0"
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "do-node-ingress-firewall": "true"
                    }
                  }
                }
              },
              {
                "apiVersion": "ingressnodefirewall.openshift.io/v1alpha1",
                "kind": "IngressNodeFirewallConfig",
                "metadata": {
                  "name": "ingressnodefirewallconfig",
                  "namespace": "ingress-node-firewall-system"
                },
                "spec": {
                  "nodeSelector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ingress-node-firewall-rhel9-operator@sha256:144403b009d1b5f2b2ee25ed3cebc525f3588f21bfa671b25c662cf800798839
          createdAt: "2024-03-13T12:36:01Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.11.0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-ingress-node-firewall
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/ingress-node-firewall
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: IngressNodeFirewallConfig
            name: ingressnodefirewallconfigs.ingressnodefirewall.openshift.io
            version: v1alpha1
          - kind: IngressNodeFirewallNodeState
            name: ingressnodefirewallnodestates.ingressnodefirewall.openshift.io
            version: v1alpha1
          - kind: IngressNodeFirewall
            name: ingressnodefirewalls.ingressnodefirewall.openshift.io
            version: v1alpha1
        description: The Ingress node firewall operator, manages the configuration
          of cluster scoped Ingress Node firewall rules for Kubernetes.
        displayName: Ingress Node Firewall Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ingressnodefirewall
        links:
        - name: Ingress Node Firewall
          url: https://github.com/openshift/ingress-node-firewall
        maintainers:
        - email: mmahmoud@redhat.com
          name: msherif1234
        maturity: alpha
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ingress-node-firewall-rhel9@sha256:4e3d69793f9280f6910a6749e80e8f6fe5d20c7a2c160d5682f310309c497e64
        - registry.redhat.io/openshift4/ingress-node-firewall-rhel9-operator@sha256:144403b009d1b5f2b2ee25ed3cebc525f3588f21bfa671b25c662cf800798839
        version: 4.16.0-202509240123
      entries:
      - name: ingress-node-firewall.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: ingress-node-firewall.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: ingress-node-firewall.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: ingress-node-firewall.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: ingress-node-firewall.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: ingress-node-firewall.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: ingress-node-firewall.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: ingress-node-firewall.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: ingress-node-firewall.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: ingress-node-firewall.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: ingress-node-firewall.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: ingress-node-firewall.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: ingress-node-firewall.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: ingress-node-firewall.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: ingress-node-firewall.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: ingress-node-firewall.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: ingress-node-firewall.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: ingress-node-firewall.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: ingress-node-firewall.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: ingress-node-firewall.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: ingress-node-firewall.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: ingress-node-firewall.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: ingress-node-firewall.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: ingress-node-firewall.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: ingress-node-firewall.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: ingress-node-firewall.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: ingress-node-firewall.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: ingress-node-firewall.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: ingress-node-firewall.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: ingress-node-firewall.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: ingress-node-firewall.v4.16.0-202405281506
        version: 4.16.0-202405281506
      name: stable
    defaultChannel: stable
    packageName: ingress-node-firewall
    provider:
      name: Red Hat, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cluster-observability-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-observability-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e
          createdAt: "2025-06-24T11:30:26Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=..0 <1.2.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-cluster-observability-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs",
              "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs",
              "prometheusagents.monitoring.rhobs",
              "perses.perses.dev"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
          support: Cluster Observability (https://issues.redhat.com/projects/COO/)
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: Perses is the Schema for the perses API
            displayName: Perses
            kind: Perses
            name: perses.perses.dev
            version: v1alpha1
          - description: A Perses Dashboard
            displayName: Perses Dashboard
            kind: PersesDashboard
            name: persesdashboards.perses.dev
            version: v1alpha1
          - description: A Perses Datasource
            displayName: Perses Datasource
            kind: PersesDatasource
            name: persesdatasources.perses.dev
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
          - description: UIPlugin defines a console plugin for observability
            displayName: UIPlugin
            kind: UIPlugin
            name: uiplugins.observability.openshift.io
            version: v1alpha1
        description: |-
          Cluster Observability Operator is a Go based Kubernetes operator to easily setup and manage various observability tools.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          - Setup console plugins
          - Setup korrel8r
          - Setup Perses
          - Setup Cluster Health Analyzer
          ### Documentation
          - **[Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/cluster_observability_operator/index)**
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)
        displayName: Cluster Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad
        - registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf
        - registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd
        - registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a
        - registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552
        - registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:e2681bce57dc9c15701f5591532c2dfe8f19778606661339553a28dc003dbca5
        - registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e
        - registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:e54c1e1301be66933f3ecb01d5a0ca27f58aabfd905b18b7d057bbf23bdb7b0d
        - registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:8597c48fc71fc6ec8e87dbe40dace4dbb7b817c1039db608af76a0d90f7ac2d0
        - registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404
        - registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223
        - registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00
        - registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1
        - registry.redhat.io/cluster-observability-operator/perses-0-1-rhel9-operator@sha256:bfed9f442aea6e8165644f1dc615beea06ec7fd84ea3f8ca393a63d3627c6a7c
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed
        version: 1.2.2
      entries:
      - name: cluster-observability-operator.v1.2.2
        version: 1.2.2
      - name: cluster-observability-operator.v1.2.1
        version: 1.2.1
      - name: cluster-observability-operator.v1.2.0
        version: 1.2.0
      - name: cluster-observability-operator.v1.1.1
        version: 1.1.1
      - name: cluster-observability-operator.v1.1.0
        version: 1.1.0
      - name: cluster-observability-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: cluster-observability-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kinesis-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kinesis-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kinesis.services.k8s.aws/v1alpha1",
                "kind": "Stream",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kinesis-controller:1.1.1
          createdAt: "2025-09-19T20:56:43Z"
          description: AWS Kinesis controller is a service controller for managing
            Kinesis resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Stream represents the state of an AWS kinesis Stream resource.
            displayName: Stream
            kind: Stream
            name: streams.kinesis.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Kinesis resources in AWS from within your Kubernetes cluster.

          **About Amazon Kinesis**

          You can use Amazon Kinesis Data Streams to collect and process large streams of data records in real time. You can create data-processing applications, known as Kinesis Data Streams applications. A typical Kinesis Data Streams application reads data from a data stream as data records. These applications can use the Kinesis Client Library, and they can run on Amazon EC2 instances.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Kinesis
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kinesis
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Kinesis Developer Resources
          url: https://aws.amazon.com/kinesis/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kinesis maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/kinesis-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-kinesis-controller.v1.1.1
        version: 1.1.1
      - name: ack-kinesis-controller.v1.1.0
        version: 1.1.0
      - name: ack-kinesis-controller.v1.0.16
        version: 1.0.16
      - name: ack-kinesis-controller.v1.0.15
        version: 1.0.15
      - name: ack-kinesis-controller.v1.0.14
        version: 1.0.14
      - name: ack-kinesis-controller.v1.0.13
        version: 1.0.13
      - name: ack-kinesis-controller.v1.0.12
        version: 1.0.12
      - name: ack-kinesis-controller.v1.0.11
        version: 1.0.11
      - name: ack-kinesis-controller.v1.0.10
        version: 1.0.10
      - name: ack-kinesis-controller.v1.0.9
        version: 1.0.9
      - name: ack-kinesis-controller.v1.0.8
        version: 1.0.8
      - name: ack-kinesis-controller.v1.0.7
        version: 1.0.7
      - name: ack-kinesis-controller.v1.0.6
        version: 1.0.6
      - name: ack-kinesis-controller.v1.0.5
        version: 1.0.5
      - name: ack-kinesis-controller.v1.0.4
        version: 1.0.4
      - name: ack-kinesis-controller.v1.0.3
        version: 1.0.3
      - name: ack-kinesis-controller.v1.0.2
        version: 1.0.2
      - name: ack-kinesis-controller.v1.0.1
        version: 1.0.1
      - name: ack-kinesis-controller.v1.0.0
        version: 1.0.0
      - name: ack-kinesis-controller.v0.0.19
        version: 0.0.19
      - name: ack-kinesis-controller.v0.0.18
        version: 0.0.18
      - name: ack-kinesis-controller.v0.0.17
        version: 0.0.17
      - name: ack-kinesis-controller.v0.0.16
        version: 0.0.16
      - name: ack-kinesis-controller.v0.0.15
        version: 0.0.15
      - name: ack-kinesis-controller.v0.0.10
        version: 0.0.10
      - name: ack-kinesis-controller.v0.0.9
        version: 0.0.9
      - name: ack-kinesis-controller.v0.0.8
        version: 0.0.8
      - name: ack-kinesis-controller.v0.0.6
        version: 0.0.6
      - name: ack-kinesis-controller.v0.0.5
        version: 0.0.5
      - name: ack-kinesis-controller.v0.0.4
        version: 0.0.4
      - name: ack-kinesis-controller.v0.0.3
        version: 0.0.3
      - name: ack-kinesis-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-kinesis-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Seldon Technologies
      provider-url: ""
    name: seldon-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: seldon-operator.v1.14.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1",
                "kind": "SeldonDeployment",
                "metadata": {
                  "labels": {
                    "app": "seldon",
                    "app.kubernetes.io/instance": "seldon1",
                    "app.kubernetes.io/name": "seldon",
                    "app.kubernetes.io/version": "v0.5"
                  },
                  "name": "seldon-model"
                },
                "spec": {
                  "name": "test-deployment",
                  "predictors": [
                    {
                      "componentSpecs": [
                        {
                          "spec": {
                            "containers": [
                              {
                                "image": "seldonio/mock_classifier:1.6.0",
                                "name": "classifier"
                              }
                            ]
                          }
                        }
                      ],
                      "graph": {
                        "children": [],
                        "name": "classifier",
                        "type": "MODEL"
                      },
                      "name": "example",
                      "replicas": 1
                    }
                  ]
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: docker.io/seldonio/seldon-core-operator:1.14.1
          createdAt: "2019-05-21 15:00:00"
          description: The Seldon operator for management, monitoring and operations
            of machine learning systems through the Seldon Engine. Once installed,
            the Seldon Operator provides multiple functions which facilitate the productisation,
            monitoring and maintenance of machine learning systems at scale.
          olm.skipRange: <1.13.1
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/SeldonIO/seldon-core
          support: Clive Cox
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A seldon engine deployment
            displayName: Seldon Deployment
            kind: SeldonDeployment
            name: seldondeployments.machinelearning.seldon.io
            version: v1
        description: "The Seldon operator enables for native operation of production
          machine learning workloads, including monitoring and operations of language-agnostic
          models with the benefits of real-time metrics and log analysis.\n   \n##
          Overview\nSeldon Core is an open source platform for deploying machine learning
          models on a Kubernetes cluster.\n\n* Deploy machine learning models in the
          cloud or on-premise.\n* Get metrics and ensure proper governance and compliance
          for your running machine learning models.\n* Create powerful inference graphs
          made up of multiple components.\n* Provide a consistent serving layer for
          models built using heterogeneous ML toolkits.\n\nYou can get started by
          following the guides in our documentation at https://docs.seldon.io/projects/seldon-core/en/latest/workflow/README.html\n"
        displayName: Seldon Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - aiops
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://docs.seldon.io/projects/seldon-core/en/latest/
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Seldon Technologies
        relatedImages:
        - docker.io/seldonio/seldon-core-operator:1.14.1
        - ""
        version: 1.14.1
      entries:
      - name: seldon-operator.v1.14.1
        version: 1.14.1
      name: stable
    defaultChannel: stable
    packageName: seldon-operator
    provider:
      name: Seldon Technologies
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: keycloak-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keycloak-operator.v19.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "False"
          containerImage: quay.io/keycloak/keycloak-operator:19.0.3-legacy
          createdAt: "2022-10-07 17:00:00"
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
          * Install Extensions
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/docs/latest/server_installation/index.html#_operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@lists.jboss.org
          name: Keycloak DEV mailing list
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/keycloak/keycloak-operator:19.0.3-legacy
        version: 19.0.3
      entries:
      - name: keycloak-operator.v19.0.3
        version: 19.0.3
      - name: keycloak-operator.v19.0.2
        version: 19.0.2
      - name: keycloak-operator.v19.0.1
        version: 19.0.1
      - name: keycloak-operator.v19.0.0
        version: 19.0.0
      - name: keycloak-operator.v18.0.2
        version: 18.0.2
      - name: keycloak-operator.v18.0.1
        version: 18.0.1
      - name: keycloak-operator.v18.0.0
        version: 18.0.0
      - name: keycloak-operator.v17.0.1
        version: 17.0.1
      - name: keycloak-operator.v17.0.0
        version: 17.0.0
      - name: keycloak-operator.v16.1.0
        version: 16.1.0
      - name: keycloak-operator.v16.0.0
        version: 16.0.0
      - name: keycloak-operator.v15.1.1
        version: 15.1.1
      - name: keycloak-operator.v15.1.0
        version: 15.1.0
      - name: keycloak-operator.v15.0.2
        version: 15.0.2
      - name: keycloak-operator.v15.0.1
        version: 15.0.1
      - name: keycloak-operator.v15.0.0
        version: 15.0.0
      - name: keycloak-operator.v14.0.0
        version: 14.0.0
      name: alpha
    - currentCSV: keycloak-operator.v20.0.0-alpha.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "False"
          containerImage: quay.io/keycloak/keycloak-operator:19.0.3
          createdAt: 10/07/22 17:00:00
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/docs/latest/server_installation/index.html#_operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@lists.jboss.org
          name: Keycloak DEV mailing list
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/keycloak/keycloak-operator:19.0.3
        version: 20.0.0-alpha.7
      entries:
      - name: keycloak-operator.v20.0.0-alpha.7
        version: 20.0.0-alpha.7
      - name: keycloak-operator.v20.0.0-alpha.6
        version: 20.0.0-alpha.6
      - name: keycloak-operator.v20.0.0-alpha.5
        version: 20.0.0-alpha.5
      - name: keycloak-operator.v20.0.0-alpha.4
        version: 20.0.0-alpha.4
      - name: keycloak-operator.v20.0.0-alpha.3
        version: 20.0.0-alpha.3
      - name: keycloak-operator.v20.0.0-alpha.2
        version: 20.0.0-alpha.2
      - name: keycloak-operator.v20.0.0-alpha.1
        version: 20.0.0-alpha.1
      name: candidate
    - currentCSV: keycloak-operator.v26.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/keycloak/keycloak-operator:26.4.0
          createdAt: "2025-09-30T11:58:06Z"
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/keycloak/keycloak-operator:26.4.0
        version: 26.4.0
      entries:
      - name: keycloak-operator.v26.4.0
        version: 26.4.0
      - name: keycloak-operator.v26.3.5
        version: 26.3.5
      - name: keycloak-operator.v26.3.4
        version: 26.3.4
      - name: keycloak-operator.v26.3.3
        version: 26.3.3
      - name: keycloak-operator.v26.3.2
        version: 26.3.2
      - name: keycloak-operator.v26.3.1
        version: 26.3.1
      - name: keycloak-operator.v26.3.0
        version: 26.3.0
      - name: keycloak-operator.v26.2.5
        version: 26.2.5
      - name: keycloak-operator.v26.2.4
        version: 26.2.4
      - name: keycloak-operator.v26.2.3
        version: 26.2.3
      - name: keycloak-operator.v26.2.2
        version: 26.2.2
      - name: keycloak-operator.v26.2.1
        version: 26.2.1
      - name: keycloak-operator.v26.2.0
        version: 26.2.0
      - name: keycloak-operator.v26.1.5
        version: 26.1.5
      - name: keycloak-operator.v26.1.4
        version: 26.1.4
      - name: keycloak-operator.v26.1.3
        version: 26.1.3
      - name: keycloak-operator.v26.1.2
        version: 26.1.2
      - name: keycloak-operator.v26.1.1
        version: 26.1.1
      - name: keycloak-operator.v26.1.0
        version: 26.1.0
      - name: keycloak-operator.v26.0.8
        version: 26.0.8
      - name: keycloak-operator.v26.0.7
        version: 26.0.7
      - name: keycloak-operator.v26.0.6
        version: 26.0.6
      - name: keycloak-operator.v26.0.5
        version: 26.0.5
      - name: keycloak-operator.v26.0.4
        version: 26.0.4
      - name: keycloak-operator.v26.0.2
        version: 26.0.2
      - name: keycloak-operator.v26.0.1
        version: 26.0.1
      - name: keycloak-operator.v26.0.0
        version: 26.0.0
      - name: keycloak-operator.v25.0.6
        version: 25.0.6
      - name: keycloak-operator.v25.0.5
        version: 25.0.5
      - name: keycloak-operator.v25.0.4
        version: 25.0.4
      - name: keycloak-operator.v25.0.2
        version: 25.0.2
      - name: keycloak-operator.v25.0.1
        version: 25.0.1
      - name: keycloak-operator.v25.0.0
        version: 25.0.0
      - name: keycloak-operator.v24.0.5
        version: 24.0.5
      - name: keycloak-operator.v24.0.4
        version: 24.0.4
      - name: keycloak-operator.v24.0.3
        version: 24.0.3
      - name: keycloak-operator.v24.0.2
        version: 24.0.2
      - name: keycloak-operator.v24.0.1
        version: 24.0.1
      - name: keycloak-operator.v24.0.0
        version: 24.0.0
      - name: keycloak-operator.v23.0.7
        version: 23.0.7
      - name: keycloak-operator.v23.0.6
        version: 23.0.6
      - name: keycloak-operator.v23.0.5
        version: 23.0.5
      - name: keycloak-operator.v23.0.4
        version: 23.0.4
      - name: keycloak-operator.v23.0.3
        version: 23.0.3
      - name: keycloak-operator.v23.0.2
        version: 23.0.2
      - name: keycloak-operator.v23.0.1
        version: 23.0.1
      - name: keycloak-operator.v23.0.0
        version: 23.0.0
      - name: keycloak-operator.v22.0.5
        version: 22.0.5
      - name: keycloak-operator.v22.0.4
        version: 22.0.4
      - name: keycloak-operator.v22.0.3
        version: 22.0.3
      - name: keycloak-operator.v22.0.2
        version: 22.0.2
      - name: keycloak-operator.v22.0.1
        version: 22.0.1
      - name: keycloak-operator.v22.0.0
        version: 22.0.0
      - name: keycloak-operator.v21.1.2
        version: 21.1.2
      - name: keycloak-operator.v21.1.1
        version: 21.1.1
      - name: keycloak-operator.v21.1.0
        version: 21.1.0
      - name: keycloak-operator.v21.0.2
        version: 21.0.2
      - name: keycloak-operator.v21.0.1
        version: 21.0.1
      - name: keycloak-operator.v21.0.0
        version: 21.0.0
      - name: keycloak-operator.v20.0.3
        version: 20.0.3
      - name: keycloak-operator.v20.0.2
        version: 20.0.2
      - name: keycloak-operator.v20.0.1
        version: 20.0.1
      - name: keycloak-operator.v20.0.0
        version: 20.0.0
      name: fast
    defaultChannel: fast
    packageName: keycloak-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: redis-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"Redis","metadata":{"name":"redis"},"spec":{"mode":"cluster","size":3,"global":{"image":"quay.io/opstree/redis:v6.2","imagePullPolicy":"IfNotPresent","password":"Opstree@1234","resources":{"requests":{"cpu":"100m","memory":"128Mi"},"limits":{"cpu":"100m","memory":"128Mi"}}},"master":{"service":{"type":"ClusterIP"},"redisConfig":{}},"slave":{"service":{"type":"ClusterIP"},"redisConfig":{}},"service":{"type":"ClusterIP"},"redisConfig":{},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:1.0","imagePullPolicy":"Always","resources":{"requests":{"cpu":"100m","memory":"128Mi"},"limits":{"cpu":"100m","memory":"128Mi"}}},"storage":{"volumeClaimTemplate":{"spec":{"storageClassName":"buildpiper-storage","accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}}}}]'
          capabilities: Seamless Upgrades
          categories: Database, Storage
          certified: "false"
          containerImage: quay.io/opstree/redis-operator:v0.4.0
          createdAt: 06-02-2021
          description: A Golang based redis operator that will make/oversee Redis
            standalone/cluster mode setup on top of the Kubernetes.
          repository: https://github.com/ot-container-kit/redis-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Redis
            displayName: Redis
            kind: Redis
            name: redis.redis.redis.opstreelabs.in
            version: v1beta1
        description: |+
          A Golang based redis operator that will make/oversee Redis standalone/cluster mode setup on top of the Kubernetes. It can create a redis cluster setup with best practices on Cloud as well as the Bare metal environment. Also, it provides an in-built monitoring capability using redis-exporter.

          ### Supported Features

          Here the features which are supported by this operator:-
          - Redis cluster/standalone mode setup
          - Inbuilt monitoring with prometheus exporter
          - Dynamic storage provisioning with pvc template
          - Resources restrictions with k8s requests and limits
          - Password/Password-less setup
          - Node selector and affinity
          - Priority class to manage setup priority
          - SecurityContext to manipulate kernel parameters
          ### Documentation

          [Documentation](https://docs.opstreelabs.in/redis-operator/)

          ### Getting Help

          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.

          ### Contributing

          You can contribute to this project by:-
          - Raising Bugs and Feature Requests
          - Fixing issue by opening a Pull Request
          - Improving Documentation

          ### License

          Logging Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/redis-operator/blob/master/LICENSE)


        displayName: Redis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - Redis
        - Opstree
        - CRD
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://docs.opstreelabs.in/redis-operator/
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: beta
        minKubeVersion: 1.11.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/opstree/redis-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: redis-operator.v0.4.0
        version: 0.4.0
      - name: redis-operator.v0.3.0
        version: 0.3.0
      name: preview
    - currentCSV: redis-operator.v0.15.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisCluster","metadata":{"name":"redis-cluster"},"spec":{"clusterSize":3,"clusterVersion":"v7","securityContext":{"runAsUser":1000,"fsGroup":1000},"persistenceEnabled":true,"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"Redis","metadata":{"name":"redis-standalone"},"spec":{"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"IfNotPresent"},"securityContext":{"runAsUser":1000,"fsGroup":1000}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisSentinel","metadata":{"name":"redis-sentinel"},"spec":{"clusterSize":3,"podSecurityContext":{"runAsUser":1000,"fsGroup":1000},"redisSentinelConfig":{"redisReplicationName":"redis-replication"},"kubernetesConfig":{"image":"quay.io/opstree/redis-sentinel:v7.0.7","imagePullPolicy":"IfNotPresent","resources":{"requests":{"cpu":"101m","memory":"128Mi"},"limits":{"cpu":"101m","memory":"128Mi"}}}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisReplication","metadata":{"name":"redis-replication"},"spec":{"clusterSize":3,"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"Always"},"podSecurityContext":{"runAsUser":1000,"fsGroup":1000}}}]'
          capabilities: Seamless Upgrades
          categories: Database, Storage
          certified: "false"
          containerImage: quay.io/opstree/redis-operator:v0.15.1
          createdAt: 24-09-2023
          description: A Golang based redis operator that will make/oversee Redis
            standalone/cluster mode setup on top of the Kubernetes.
          repository: https://github.com/ot-container-kit/redis-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Redis Cluster
            displayName: RedisCluster
            kind: RedisCluster
            name: redisclusters.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis Replication
            displayName: RedisReplication
            kind: RedisReplication
            name: redisreplications.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis Sentinel
            displayName: RedisSentinel
            kind: RedisSentinel
            name: redissentinels.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis
            displayName: Redis
            kind: Redis
            name: redis.redis.redis.opstreelabs.in
            version: v1beta1
        description: "A Golang based redis operator that will make/oversee Redis\nstandalone/cluster
          mode setup on top of the Kubernetes. It can create a\nredis cluster setup
          with best practices on Cloud as well as the Bare metal\nenvironment. Also,
          it provides an in-built monitoring capability using\nredis-exporter.\n\nOrganizations
          that are using Redis Operator to manage their redis workload \ncan be found
          here. If your organization is also using Redis Operator, please \nfree to
          add by creating a PR https://github.com/OT-CONTAINER-KIT/redis-operator/pulls.\n\nThis
          operator only supports versions of redis =>6.\n\n### Supported Features\n\nHere
          the features which are supported by this operator:-\n- Redis cluster and
          standalone mode setup\n- Redis cluster failover and recovery\n- Redis leader
          and follower setup with Sentinel for High Availability\n- Inbuilt monitoring
          with redis exporter\n- Password and password-less setup of redis\n- TLS
          support for additional security layer\n- Ipv4 and Ipv6 support for redis
          setup\n- External configuration support for leader and follower\n- Detailed
          monitoring grafana dashboard\n\n### Documentation\n\n[Documentation](https://ot-redis-operator.netlify.app/)\n\n###
          Getting Help\nThis project is managed by OpsTree Solutions. If you have
          any queries or\nsuggestions, mail us at opensource@opstree.com.\n\n### Contributing\n\nYou
          can contribute to this project by:-\n- Raising Bugs and Feature Requests\n-
          Fixing issue by opening a Pull Request\n- Improving Documentation\n\n###
          License\n\nLogging Operator is licensed under [Apache License, Version\n2.0](https://github.com/OT-CONTAINER-KIT/redis-operator/blob/master/LICENSE)\n"
        displayName: Redis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - Redis
        - Opstree
        - CRD
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-redis-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/opstree/redis-operator:v0.15.1
        version: 0.15.1
      entries:
      - name: redis-operator.v0.15.1
        version: 0.15.1
      - name: redis-operator.v0.15.0
        version: 0.15.0
      - name: redis-operator.v0.14.0
        version: 0.14.0
      - name: redis-operator.v0.13.0
        version: 0.13.0
      - name: redis-operator.v0.8.0
        version: 0.8.0
      - name: redis-operator.v0.6.0
        version: 0.6.0
      - name: redis-operator.v0.5.0
        version: 0.5.0
      - name: redis-operator.v0.4.0
        version: 0.4.0
      - name: redis-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: redis-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Veeam Kasten
      provider-url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
    name: k10-kasten-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k10-kasten-operator-rhmp.v8.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apik10.kasten.io/v1alpha1",
                "kind": "K10",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/reconcile-period": "2m",
                    "helm.sdk.operatorframework.io/rollback-force": "false"
                  },
                  "name": "k10"
                },
                "spec": {
                  "auth": {
                    "basicAuth": {
                      "enabled": false,
                      "htpasswd": "",
                      "secretName": ""
                    },
                    "tokenAuth": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "persistence": {
                      "catalog": {
                        "size": ""
                      },
                      "storageClass": ""
                    }
                  },
                  "metering": {
                    "mode": ""
                  },
                  "route": {
                    "enabled": false,
                    "host": "",
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "apik10.kasten.io/v1alpha1",
                "kind": "K10restore",
                "metadata": {
                  "name": "k10restore"
                },
                "spec": {
                  "pointInTime": null,
                  "profileName": "",
                  "skipResource": null,
                  "sourceClusterID": ""
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          console.openshift.io/plugins: '["veeam-kasten-console-plugin"]'
          containerImage: registry.connect.redhat.com/kasten/k10-operator@sha256:4e8ccf38e47f92d2a308587591f85fb4f15aea1e7fd5c248803764c94eaecd4a
          createdAt: Wed Sep 24 02:11:59 PM UTC 2025
          description: Kasten data management platform for OpenShift, The Market Leader
            for Kubernetes Backup and DR
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/k10-kasten-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/k10-kasten-operator-rhmp/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: kasten-io
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: contact@kasten.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              k10restore can recover the K10 platform in case of a variety of disasters such as
              the accidental deletion of K10, failure of underlying storage that K10 uses for its catalog,
              or even the accidental destruction of the Kubernetes cluster on which K10 is deployed.

              **Before creating k10restore**: [K10 Disaster Recovery documentation](https://docs.kasten.io/latest/operating/dr.html#recovering-with-the-operator)
            kind: K10restore
            name: k10restores.apik10.kasten.io
            version: v1alpha1
          - description: |-
              [Kasten's k10](https://docs.kasten.io/) is a data lifecycle management system for all your persistence enabled container-based applications.

              **Note: The complete list of values supported by the K10 spec field in the "YAML View" are documented [here](https://docs.kasten.io/latest/install/advanced.html#complete-list-of-k10-helm-options).**
            kind: K10
            name: k10s.apik10.kasten.io
            version: v1alpha1
        description: |-
          Kasten data management platform for OpenShift, The Market Leader for Kubernetes Backup and DR.

          The supported Kubernetes versions are listed here - https://docs.kasten.io/latest/operating/support.html
          Please review release notes before an upgrade - https://docs.kasten.io/latest/releasenotes.html
        displayName: Veeam Kasten (Free)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kubernetes
        - storage
        - data-management
        links:
        - name: Kasten Documentation
          url: https://docs.kasten.io
        - name: Veeam Kasten
          url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
        maintainers:
        - email: contact@kasten.io
          name: Veeam Kasten
        maturity: stable
        provider:
          name: Veeam Kasten
          url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:503d5ee0fbcd5bd3492785f4246e19cbdb92960be0d05f3c6f123a32fe0fe84d
        - registry.connect.redhat.com/kasten/bloblifecyclemanager@sha256:7b209314b96bc8ebaa77e97426bb34176d05844cc3e8d4b462cfba5c4a8d3798
        - registry.connect.redhat.com/kasten/kanister-tools@sha256:7af2e37c9027f46c565901d35a3a7da4d9c7ddd2812628b67275d4acf006b80c
        - registry.connect.redhat.com/kasten/kanister@sha256:1d69ff91cec8e7c131dfbe1ad3e282aab854d3412f5f6e3b808d25eb7b8c38bc
        - registry.connect.redhat.com/kasten/metric-sidecar@sha256:20c142258de56cf8c60d4b0afa5e72b9317023ebf4ab109fb84e0d57fae854a4
        - registry.connect.redhat.com/kasten/controllermanager@sha256:ab3b815bb12cd3c263cbfdba3939a95097450f1864431cca7380827986d704ed
        - registry.connect.redhat.com/kasten/configmap-reload@sha256:01010f2bb580eb131d63d3271efe40eadd8269b84cdb02bdea36750d1359a43f
        - registry.connect.redhat.com/kasten/aggregatedapis@sha256:db868afd5c6d4a8d2e7c8376ed6943963c4be299fca6b26931f907430763d244
        - registry.connect.redhat.com/kasten/jobs@sha256:c6358878e981baee4ac72abcc1269c206ee014af2c16584d1b41fe18e0f0e57f
        - registry.connect.redhat.com/kasten/upgrade@sha256:bf2200697d38f3ee95ba2863dfbae4d982273d2fd8450c9ffbc3a85be7d6d87d
        - registry.connect.redhat.com/kasten/prometheus@sha256:29d913bb06c3245cca5798772dba7d7d5dc9b76ed2e0700f5eb4d20416e24fa2
        - registry.connect.redhat.com/kasten/k10tools@sha256:59caa1920c76a1bc86e7733731602e3f0fc08026d9a16000bb748503a4d9cd9b
        - registry.connect.redhat.com/kasten/garbagecollector@sha256:6f2972bb255f5a2de33ace5ce805b15d5e04d3c259c9c8d018c1257c74562878
        - registry.connect.redhat.com/kasten/events@sha256:7cae1c0909ea7a468e96484d99300677bdad49f34cc57c412ebaf342cb176264
        - registry.connect.redhat.com/kasten/logging@sha256:014bfdfc3620cc2fe6540b9993b9b9e7d70c9b969714477db4de91a90bdb4a38
        - registry.connect.redhat.com/kasten/k10-operator@sha256:4e8ccf38e47f92d2a308587591f85fb4f15aea1e7fd5c248803764c94eaecd4a
        - registry.connect.redhat.com/kasten/auth@sha256:f8543e67de3d5276c597c26f16053bf9bbcbd9cbdad4652214935a1e40a26817
        - registry.connect.redhat.com/kasten/frontend@sha256:a021964ed85133897e4956014b5c5013da3332a36a7e96a0de2e2e1b67888491
        - registry.connect.redhat.com/kasten/init@sha256:cc7fbbda1b6946e93dd083fb7092f38789f626c26ac68c226b61fe57064a8f11
        - registry.connect.redhat.com/kasten/dex@sha256:4c8522a27c0d4f0977baa074800bba96a9bd9aefc57473de283b1d95586f7517
        - registry.connect.redhat.com/kasten/catalog@sha256:11a77ac968634850c72cd10f2bde55afc632c68cdc0881aeae83c100288c6aec
        - registry.connect.redhat.com/kasten/dashboardbff@sha256:e1208c60146faa8c3c7ef73b2290663840276ed8d3b3c990f495cee2967104ee
        - registry.connect.redhat.com/kasten/state@sha256:191383c76c4ff22805c06217e9f32a2f41a1206afab5f16acec4a5dd56e4a6b4
        - registry.connect.redhat.com/kasten/repositories@sha256:924a607874e1807a877cb454e00539d64f4487e7435d4bfb524f3e569ad1f942
        - registry.connect.redhat.com/kasten/metering@sha256:f809696ac1309fc54fde87e5d2030267d266386fb73a56287868bbac433756d8
        - registry.connect.redhat.com/kasten/restorectl@sha256:9dc3c834bd3d4543115ae1ad586ee1248fc2b637dbc2d1e2c66ecc2a459db67e
        - registry.connect.redhat.com/kasten/datamover@sha256:c9e2dc5b646aa0c465a7b2daf011462273a6f5b53fdff772d879293208e2d57c
        - registry.connect.redhat.com/kasten/vbrintegrationapi@sha256:1a20b0da4efb7e802516224675b37751bd56d406e57b16216388d2b5167c648d
        - registry.connect.redhat.com/kasten/executor@sha256:32c4ac1aa2ca802ff77e913991949e6841149963863ad527a67f6bab6c94a5af
        - registry.connect.redhat.com/kasten/gateway@sha256:dc2d7bdeb8abaf9454470afd53af2cd5430505351a97b6501ff65a60dd2a6db1
        - registry.connect.redhat.com/kasten/ocpconsoleplugin@sha256:69f6bfb9421f4264dd7d066d809311022c0db57c41319a6242532f6acfbd2145
        - registry.connect.redhat.com/kasten/crypto@sha256:86bfbbf4c4f06bab1e4b091f9f312d536813e2c11aa899690b3277ff73110a09
        version: 8.0.9
      entries:
      - name: k10-kasten-operator-rhmp.v8.0.9
        version: 8.0.9
      - name: k10-kasten-operator-rhmp.v8.0.8
        version: 8.0.8
      - name: k10-kasten-operator-rhmp.v8.0.7
        version: 8.0.7
      - name: k10-kasten-operator-rhmp.v8.0.6
        version: 8.0.6
      - name: k10-kasten-operator-rhmp.v8.0.5
        version: 8.0.5
      - name: k10-kasten-operator-rhmp.v8.0.4
        version: 8.0.4
      - name: k10-kasten-operator-rhmp.v8.0.3
        version: 8.0.3
      - name: k10-kasten-operator-rhmp.v8.0.2
        version: 8.0.2
      - name: k10-kasten-operator-rhmp.v8.0.1
        version: 8.0.1
      - name: k10-kasten-operator-rhmp.v8.0.0
        version: 8.0.0
      - name: k10-kasten-operator-rhmp.v7.5.10
        version: 7.5.10
      - name: k10-kasten-operator-rhmp.v7.5.9
        version: 7.5.9
      - name: k10-kasten-operator-rhmp.v7.5.8
        version: 7.5.8
      - name: k10-kasten-operator-rhmp.v7.5.7
        version: 7.5.7
      - name: k10-kasten-operator-rhmp.v7.5.4
        version: 7.5.4
      - name: k10-kasten-operator-rhmp.v7.5.3
        version: 7.5.3
      - name: k10-kasten-operator-rhmp.v7.5.2
        version: 7.5.2
      - name: k10-kasten-operator-rhmp.v7.5.1
        version: 7.5.1
      - name: k10-kasten-operator-rhmp.v7.5.0
        version: 7.5.0
      - name: k10-kasten-operator-rhmp.v7.0.14
        version: 7.0.14
      - name: k10-kasten-operator-rhmp.v7.0.13
        version: 7.0.13
      - name: k10-kasten-operator-rhmp.v7.0.12
        version: 7.0.12
      - name: k10-kasten-operator-rhmp.v7.0.11
        version: 7.0.11
      - name: k10-kasten-operator-rhmp.v7.0.9
        version: 7.0.9
      - name: k10-kasten-operator-rhmp.v7.0.8
        version: 7.0.8
      - name: k10-kasten-operator-rhmp.v7.0.7
        version: 7.0.7
      - name: k10-kasten-operator-rhmp.v7.0.6
        version: 7.0.6
      - name: k10-kasten-operator-rhmp.v7.0.5
        version: 7.0.5
      - name: k10-kasten-operator-rhmp.v7.0.4
        version: 7.0.4
      - name: k10-kasten-operator-rhmp.v7.0.3
        version: 7.0.3
      - name: k10-kasten-operator-rhmp.v7.0.1
        version: 7.0.1
      - name: k10-kasten-operator-rhmp.v6.5.14
        version: 6.5.14
      - name: k10-kasten-operator-rhmp.v6.5.13
        version: 6.5.13
      - name: k10-kasten-operator-rhmp.v6.5.12
        version: 6.5.12
      - name: k10-kasten-operator-rhmp.v6.5.11
        version: 6.5.11
      - name: k10-kasten-operator-rhmp.v6.5.10
        version: 6.5.10
      - name: k10-kasten-operator-rhmp.v6.5.8
        version: 6.5.8
      - name: k10-kasten-operator-rhmp.v6.5.6
        version: 6.5.6
      - name: k10-kasten-operator-rhmp.v6.5.5
        version: 6.5.5
      - name: k10-kasten-operator-rhmp.v6.5.4
        version: 6.5.4
      - name: k10-kasten-operator-rhmp.v6.5.3
        version: 6.5.3
      - name: k10-kasten-operator-rhmp.v6.5.1
        version: 6.5.1
      - name: k10-kasten-operator-rhmp.v6.5.0
        version: 6.5.0
      - name: k10-kasten-operator-rhmp.v6.0.12
        version: 6.0.12
      - name: k10-kasten-operator-rhmp.v6.0.11
        version: 6.0.11
      - name: k10-kasten-operator-rhmp.v6.0.9
        version: 6.0.9
      - name: k10-kasten-operator-rhmp.v6.0.8
        version: 6.0.8
      - name: k10-kasten-operator-rhmp.v6.0.7
        version: 6.0.7
      - name: k10-kasten-operator-rhmp.v6.0.6
        version: 6.0.6
      - name: k10-kasten-operator-rhmp.v6.0.5
        version: 6.0.5
      - name: k10-kasten-operator-rhmp.v6.0.4
        version: 6.0.4
      - name: k10-kasten-operator-rhmp.v6.0.3
        version: 6.0.3
      - name: k10-kasten-operator-rhmp.v6.0.2
        version: 6.0.2
      - name: k10-kasten-operator-rhmp.v6.0.1
        version: 6.0.1
      - name: k10-kasten-operator-rhmp.v6.0.0
        version: 6.0.0
      name: stable
    defaultChannel: stable
    packageName: k10-kasten-operator-rhmp
    provider:
      name: Veeam Kasten
      url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SusQL Operator Contributors
      provider-url: https://github.com/sustainable-computing-io/susql-operator
    name: susql-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: susql-operator.v0.0.35
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "susql.ibm.com/v1",
                "kind": "LabelGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "susql-operator",
                    "app.kubernetes.io/instance": "labelgroup-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "susql-operator",
                    "app.kubernetes.io/part-of": "susql-operator"
                  },
                  "name": "labelgroup-sample"
                },
                "spec": {
                  "labels": [
                    "labelgroup-sample-1",
                    "labelgroup-sample-2"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: quay.io/sustainable_computing_io/susql_operator:0.0.34
          createdAt: "2025-07-10T01:20:37Z"
          description: 'Aggregates energy and CO2 emission data for pods tagged with
            SusQL labels '
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/sustainable-computing-io/susql-operator
          support: https://github.com/sustainable-computing-io/susql-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LabelGroup is the Schema for the LabelGroups API
            displayName: Label Group
            kind: LabelGroup
            name: labelgroups.susql.ibm.com
            version: v1
        description: |-
          ### About this Operator

          SusQL is a Kubernetes operator that aggregates energy and estimated
          carbon dioxide emission data for pods tagged with SusQL specific
          labels. The energy measurements are taken from Kepler which should
          be deployed in the cluster before using SusQL.

          ### Prerequisites

          1. Deployment of [Kepler](https://sustainable-computing.io/) on the cluster
          2. Ensure that [User Project Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          is enabled to monitor energy consumed in user projects.
          3. Creation of a SusQL Service Monitor:
                `oc apply -n <NAMESPACE> -f https://raw.githubusercontent.com/sustainable-computing-io/susql-operator/main/hack/susql-servicemonitor.yaml`

          ### API Backward Compatibility

          Until Operator supports **seamless upgrades**, a new version of the operator may
          introduce a change that is **NOT** backwards compatible with the previous version.
          Thus, to upgrade the operator, uninstall the already installed version completely
          (including CRDs) and install the new version. For example to upgrade from 0.1.0
          to 0.2.0, you must first uninstall 0.1.0.

          ### LabelGroup spec Example

          ```
          apiVersion: susql.ibm.com/v1
          kind: LabelGroup
          metadata:
            name: labelgroup-name
            namespace: susql-operator
          spec:
            labels:
              - my-label-1
              - my-label-2
          ```

          ### Documentation

          Documentation can be found at [SusQL Operator](https://github.com/sustainable-computing-io/susql-operator#readme).

          ### Getting help

          If you encounter any issues while using the SusQL Operator, you can create an
          [issue](https://github.com/sustainable-computing-io/susql-operator/issues)
          on our GitHub repo for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by raising [issues](https://github.com/sustainable-computing-io/susql-operator/issues)
          related to susql-operator.

          ### License
          susql-operator is licensed under the Apache 2.0 license
        displayName: SusQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - aggregation
        - sustainability
        - energy
        - kepler
        - susql
        - carbon dioxide (CO2) emissions
        - green computing
        links:
        - name: SusQL Operator
          url: https://github.com/sustainable-computing-io/susql-operator
        - name: Sustainable Computing
          url: https://github.com/sustainable-computing-io
        maintainers:
        - email: trent@jp.ibm.com
          name: Scott Trent
        - email: mohammed.munir.abdi@ibm.com
          name: Mohammed Abdi
        maturity: alpha
        provider:
          name: SusQL Operator Contributors
          url: https://github.com/sustainable-computing-io/susql-operator
        relatedImages:
        - quay.io/sustainable_computing_io/susql_operator:0.0.34
        version: 0.0.35
      entries:
      - name: susql-operator.v0.0.35
        version: 0.0.35
      - name: susql-operator.v0.0.34
        version: 0.0.34
      - name: susql-operator.v0.0.33
        version: 0.0.33
      - name: susql-operator.v0.0.32
        version: 0.0.32
      - name: susql-operator.v0.0.30
        version: 0.0.30
      - name: susql-operator.v0.0.29
        version: 0.0.29
      - name: susql-operator.v0.0.28
        version: 0.0.28
      - name: susql-operator.v0.0.27
        version: 0.0.27
      name: alpha
    defaultChannel: alpha
    packageName: susql-operator
    provider:
      name: SusQL Operator Contributors
      url: https://github.com/sustainable-computing-io/susql-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cockroach Labs
      provider-url: ""
    name: cockroachdb-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroach-operator.v2.18.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v25.2.2",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
          createdAt: "2025-07-22T11:30:55Z"
          description: CockroachDB Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1304d7f871a95c8b087bad1b548217562792edc66c8fe07c621e05242b8cd53e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6408f89b0a88ec154f7ee98c314be574971c4c9aaf906b473d10124d822889be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:65923c37fecda2d3174212fa9d9a6f34241d065389336ea69a7be99ba16cedc2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3bf017b2b2506724500f5e29017d1b956825dd0f61a7c0049b0ff084c3554b5b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:85a0e085901c2a73192556d5d70795d4b66cbcfe01d24c73b6fe6db8d5f2e5ca
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bd75aa918355e77cd671c3e11ee807861378e96245245f4fdf243cba5749f9a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e8f26fd132affbaeb8ef6b1807a646c83b301e5dc334a592dfb2cf8dc3dbaf10
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6ca927f137be28481e3a226e6c6bcb7c5ac27664984f17ffc6c1419cd7d08eb7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da367cf0ac52045002e1c186f8e6964267ad87d5c25f9e72fd2c9b9a98a32702
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc1ab77a9537ccf236ce6dc370940dfedcfaa91b6e3db37bf2d5eade7ff84708
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e0f21abea34825e55622c432a306f2e8aafd64d913eb78f2f601c51dc878a1da
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1b6d934a7b6c42a61e5ad51380e69af341a575b42205299fd58bc0c6ff306f6a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a6f1c43486d044fed9f59406118f429c8037b389bd8c8d7a1c7aeb89ed2661f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b06bd7f3802e8fa2af14bb0b07804cad4cc0fbbc97da18d09a9106bf074ed6d8
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:740ba10f1ef0a6c5fa6ad31ab18855fc75f8e024ce4c23b0676c5a9c5b62cd75
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c9a31bdf19175880d0981e21dd19a1458b6e8cde8d4acbcb554f3bea3578a64e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf7d3bea47638b763b312c7c11ca034a7ac62d993d75a13edceec74799b97e52
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f438fbbf9c87ca0d54c170e2c5105b1e265182e1c2ede48adea370e2be9dc895
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9ab968e4ca62f23bf9848e5475ef4076df7318a8560f63d5b2090a32af6fe4e2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d978678b3c254abd52f69f226b0aacc4b02c2aaca328c54ef10f6f9bb453582d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:dd521d7bad62da054eec54179be75c643a5492d981e11c0e66db41902cd2182c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d0014b1c19890b9e8ef1a4f3efe7eb5b269be5610ba6ababd4258b52d2f85e8a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5a54d564e08baefc2786f4b742b2507d31984664cd00333406aa7a3aa5b9c913
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5181196fff3bf8a5ec8fcfe4f1958cc69f285f0beb4f0b476b4b5e0912f3db42
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9f42e929f2cc7e9b6d24dd61e19c5d93dcfeabd34b9c8b935afbe78577efb18b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:266165ab87b19a05d6796e8b15ff0eba632d6ac96e6393400464c6136dd500ec
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:46d9287ef11e3b21a36abfa31a26da3b93040bb6163b9223972a5dd1930fecff
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acbfee4492327a720fe7da76929a86192d8af2442bbb76f0a089f6a81a890be3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6371377718325235a2720dba0c97eb4321ee29b31a28cb893f76abd62baa339
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:79d688bebe0643ef0dd6ba7108cdb34c0442470c0c20ae0df88b5d02a3476cd0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e23af8f9d32c3f80a52d79ceb4b359dda1d75ba0dd4cb8ed4b66ac2f4e40c69b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:58de9d0b2ce005c7ad3e99152100e59b7677398557028a7a59e56e669fabb81c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7daaf1368ab2c25694cdea0a598e1de59247b04a4b4b4b8348ed6edbaa194f9d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:35ede65457d845e8c063ad7d9515d3a69f98e0751424ecbac484963678a0b1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f42fa22aac3cc41ce72e144a72530ffbee765a392b24593dafd06cc098f8410b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a4f932c3e9ff3aeb70ba1f7a92aa4642bd14cbb7d37d04ff750ed517b916cdb9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:19f36f53f7da67755eb86da77098c5eeb84b381f9c18aa29ba573793dc498564
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4635550222775fc75737be5e5c0359303e3a23c4d1e1c296d7408cd15ad94aad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f3ffebd812881a830ff46c860984814dd61edcd51b478ebd3667759fb7710682
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2a128b570352a5a7f82b27353f9b774caccce83e0e7bb2673e4802d96ad0bd51
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f74c8a5dea4560ad59c6f6f6ab1e699f063334ee8db704460f870508dbeaeff0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bd336715332f776a72aaf5c3ca52e62dd53a9e8521c1feb6e135cc868d694a39
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5240f9f591263569fad7ee934e81fa8a1617de3c25da94ddd59ee5b2b3407f90
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d14f77f70aea0422a906510c1ac2652ff76e131ed705cb002f80e13e887b0cd9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:10b9df672260c023dfb37166e156c0d7e0e13144dc3e485af922e8151fd3ab05
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2af60025ecb3bb933b61328be4c2b2bfd0e7d26f53b72430629208505e45c6d2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:32fdca575c334822e4356aab36a7ed97b685c065925fe85f1b8ba8425c57159b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:42c9930d6861f6a2147ed6e01827eddd730279966763ec68868567a19cc164ab
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4b3aa590623a38f435b075de4c09f8a343d8ece8b673e6bf7c6df5d5eb4bdf4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:02e1ef01a023da0eee653f63b314786fd66a78d184a97a314422f6f9c3ef8e7e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:310310515625f099a928545865f7096997871ee4a16650a01c76c3799a18b684
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6a2ee8d62c586eb74ebbef6bef256e06e1ba5cb51c1ae55b1654adafc2ca8859
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acc083f7dde1fbc557d858d03fc26b68b2056997fc1a457e4695650214d5b303
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b78a4d3864e16d270979d9d7756012abf09e6cdb7f1eb4832f6497c26281099
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:61e602b0e3582bf00460eb9328808ccf66e3ff1b1f6df07ee6c5c76a91846876
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2d89398f03736f8ffaa82b10fed0f19fba166d3d2145b26f25cb71f3c39acf07
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5458a8fd89aa664b60e59c35bd933975cd37d18e19b11589e4ca0c2844ea0ed4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:50393ad005fed19618d04f0716c6d55e2f1afce651e785c396638e852cac83b3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ec6eb7c28c213cc83b2d7919cd87988f9a07f12276eb7351d0915f1567a5b095
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6366ba29f301a82215c05e60d78e1de66bcee85440ce2fc3c692712d4da32685
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5d6901b6b6fd6594d42bfc1842d050c7cffcb0867a5591a67469f0f860b1f21b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4d8e804b35b73d7978c112748aea2d2bb68f65081b90bca62c39f2b6e2989378
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:743cda1eda6e273c798e3405a9a97e07dd6aa5ad4d5a52789be3b0f39a32a039
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4f753151af4c3d3f5a1df9014461316249a6d0621b8fba5c4430307b3355251
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce1209ca6cc8724baee981c97c66e2852e512423bc0c0a16cdd80a6838b54031
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f5a0ccc02dc9e938e484d5b5282ff650d1890d5f754c30a0c0d717989ed5d600
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f045fe5037d44aab1a0b2209241b3931ec7cb9897cb3c6a96a9f8d5280e55dd2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08cec123813304ab03bd66d72a09949667a874a42320bedafda391ebcdb2e56c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5b3de46526d169fcd1318354545b3a18861843fa2022a23ed04d795493825287
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d590946b2f7c7ec0c7ec057a338e81f864d6a9e3d3ac4e0769aa5756a8e13fc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0754f7901feea9017bc5041cb91b7ac1cdbca10c2064f37c3648a2c395f24f83
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db07461824e73410456ef540b0bce3c7fe550aa67a73d6174300ca74b8f75276
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c74cf927a2b2c5098ad74ad0880996f8b042186a9911d888a231ef5b9e2a0715
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7d6320ac7fb0925f10cf6f524a13de629cbf13ccc371912e4fe8c777bc15d1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:28ae9661b38a72decf413c6a6c3758e61098debf5162197c27a12a070c7a096d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:226d7f1c3f0d70ac6532ad9490bebcdef44a40959b6aa7c11b0f2d1383a46217
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:44a0f40a6c46ba08297b482f3d96defe87b181c944803e5381c9acc42cabf97f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b046aa0a42dae992a03685ed82a20835532d6d4bf048e2366ff7382be4b3138
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1ce71e80c0cd0e0063eba275ab91f77982f2e92a8aef8e89cab99b93feab0a88
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:29abb169ffefe36f6d69edd438a7135ee23059cedab8f87ad3b2d52361b96b35
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:68fe64fe3e112288ccfc2a7dffd69b9b2786a5da3d3bdedd6addf41659d460c9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8443c15eb1653dd9e5ec7aad54d4b8b092d41a0977e42b76e372e5e21280746e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b83eaff2e217c06098509709cead525b55b00ccd48019d8357e6231892008055
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bed27f3f2c2d7627d1af4500d7ea0302da5c3fd3d875a6480fb6d98da481312e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b0057de903654ea0315a2c96affcc8f50a92ede8ad121ed30edb3aaf0fb7b554
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cbc74889e9ca5edbd5aaa243d7649bca0c4da2719239f8d0ec9827868112a771
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1558a1f5eab79f1a8c6c85d81740cc539dc23025ea457c0019d93a72dc05d57a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf001173bbf00cdf443b8d64a3ad724389f1cf83e3628d0b609019c8cddad35a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:87db6e079b96fb392d98ef5fc7063836bff10717efd75061016ce688be0ba38e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:72ea3d0ef72972dba93362dc31366c5946a9898f5c16780b877316a97068ef8f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1374cbdc4ecf38e4b4226328f532e830b8932b34d9449bf2fbb36e60d0310521
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2171e1c75d0c665fe33e1ea334e605604c51de6dbc22d88a3a0ac604259c448f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db4e5c2f27ba23e95a363c8174c01f0f9c6b6a27e0e581c86a38f3a1f68dc43c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d8cfb221f67516675fdfde061f6ec5dfee1224ca7eb3d8be375c27d483f306af
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:aeac9cd8e8acdbd2b0c339c48b85079c159b04f0ade24df8e3f9cae3f48b4c70
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:086d7b435993fdf21fd59b0093f52105a9028d6b769398b0033f98a65dfb7e79
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9d0bfb642090ff55a3cc3da0f72663cae18063622a4ec785ba99bc71ef4db116
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3f587e7b9f5a611013af2608e2151ed782859a52501b00adbc95739c56a8ff1e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:82d8104a37349c513447e68ea3790efb256370632353176e5fc13c2e8df1bd44
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c2a62179e8ca293d7e30b9da599e48795345cf07a72abd3a83063107b53b03cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d467383af41aa80c26172964c34152abeba45121599804e502984655b72179f0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e5814748c5ab229a1dea5662a007063c0df06fbbfcfe041fd9a6df9ef67e07cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a6e2fe24ad52913e2160f85b9fe665a4e4c4d79b164b4b899f69dd9d92f999e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6cd9f9b9fcb6e8f9a931a6ad01cb66329a74d40b22f9029fbc50e57248bad7cb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:56109e57ee0379cf48644bcf8226a9238e01139cacc5499002c99f973f121911
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f53b07212320c6939d3bb4ddd4e6a0bce311ab5b63858eb1a8728531f0dbd510
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce3d94a3fbdcde5858beb3c265987759971d991c397c6486e19904bd4706f44d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e151bded670f4cbd165a8ef864324805667d3e84ce3ace0cdf8f9aad087fccf1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6bb15b36d64eebb6e4c1db5a5466e108b271d53383c58e0b6c78cec214756a9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0e2c7023f1f97acc20c46a4944635a9c5f57bb9bb3dd534cf41fea1a9cfdfdbe
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:91f7e59f9e35a0395756eafed906963c535e886125294d671d930ce959a8f511
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f05824bb8eaa5a93e54059449eaa236dd83d729049ad00065ef3242f5f74c91e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b77b80eed7ce149d2a3f228e5f188e76d73f32bc50ecde20bba9cfa7bdc4714c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cad60044ad6573bd80b65e892a9ec0510dbe4fcbbfc6b51010ecf419f56f1024
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8fc8de4f3671a73f9aa747e5b23ec973d192fff6f2869c38985cd656063c1e3d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:022c6e1a0de875474bc953718a481a0af526c49dc910094dcdce0b5677aced18
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:468b0a40a33be1dd5a2a89ce745d7e9012bc7371f2f01cc0f8f12c0ac887db43
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4e5f7df1dc1e1db398c36d590431e7e5782897b209972d8e9e4671971c10d1b6
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7481c0cf3b579c8d98d94d4aa0334ca94bf2ccfcd25ad70b57c42b22bbf4bcee
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08496b84b3cc691909227be17e67bcd6f4d5185a93712f44d8a0985bd2c5fcd7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fd6c9b8ca82ff4f2d0fe29166d1fc9bca6439cb40907be60f4d8f1b9735de2f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:742a6f69a3a209a7080a1e911fd96467d2a91aacb3f91d6be80a6a6e2e54b2ae
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e82684b828256b3c2814930210031b3b70bc818514d4359a25f87ffdbf22be6e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8737b67b91983817e3e4ff3f47f0b070bd36162d309b60993529426c6979cb24
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4b56c0311af14220a466aa11f25fd10fa38280825539bd02215c945bbafb861c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9097e135159f46396ef4649f724d61bb5b9113c488f989d404586fbb7486fd85
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1e5d85f5db10fa79d09263db90daa44549bcc3b025e9f3f98cf0d4e51766394b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a0faa52a999289b308dc4caf050c6d9fe2f42e112aa30e4af282ed1f21c5c346
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1f6b1ec92251397dbcf7ed4a82f1f05619d64676aae2717f4eff1c33e4c2c27a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a59cafff8d4ff2c62204ece15880117011c30978823c232902c7182d747ec827
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:858dc82977b3c9e989a3a20626671b0c67c834c97132094a9212bb227ab55929
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:744193037d6b64cb73fecab15ebe022abc2b79ab0335b76ec1415ac6379b7dfa
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d78c2e0709214f96794564f13386861f4b541e55b5bd316ad495cf104f7d40b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9b5a2e0c006eb4be8e1444734a98743551f278a24e81c20bebc96f0f3b631eb0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e50aab39722af22ccbf3d1db19e37972d52b0fc9a40998a64618dc966b2bac57
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:818f592c50d4ed5599faa053344f7ecd963003352f1faa3b72dc0f5e00a0e78b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0b0e0588c1aadf0c3826cd584bd2b3a7d6781882dcb5c5e037b47ff5cb84509b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:53c7b960a2ed70a0998a4f8e28a4f78f7f676e01cffebe665e91e3a7629d88ed
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f29c31bb267451b389608af3e8a4a28d6ca0691e67b77e8f57bb5e461b064893
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bf36dab9b3257a4bba24d986f1451fc51a3a37023fbb988281f1941fd3faedd
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7b907d03b001b758d8a4b5e898725c73dda2ec2aa4b7425b685320b947154d11
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8f1f2fd1aca914a1ef007117c85ac944055bbfb873a468a7374d239ef20b065e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:37bf9e1ec7e3eff47755fe7521223f31ec6aefc6e63901db3881b30898cf78f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:40f00476b63cb9e370fe85f01407173693213a9242fc117d28b06378ca0d98e0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8a70ba12f610ac0e62066e417a0833bc1d01e91ac99d0dd794a2c0cd2a350164
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2b5c028f293c6f394e58a5495cb3dbfbedc215d0d03778d0217e045586d55a46
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da3bd2e9ff2b08d5bb86b1f5faa95a865f85c5e8912df00b6b79190c1e518a2f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7ad12df8e258b03c4099d375d02d272bd774bf188d108901d5cf6cb865e57ef3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cc20270212f01fc231ff788e20b0476b8228c440bfc40430c0836d5eb702a042
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:31b161f9bf9028a1a62cd2eef6abda01257d8889bdc3159822642549a75e3429
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0fea6ed8a78ea624240ee6ffb3573d5cf1115186aa9180fcd7d1273351deaaa3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9c85c25d5392817d17b68714dca6195ad53255d43e9c65b021d75b12d97e0579
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a3ac544d87a9d8e7c92897c4c23f803753c7ab31467db91b5c51244afc2c6be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e18d73af3257c8cdf00245f2f718b56667524c1cc69d3db28e9b868a539ae16f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:363d8feb560324a848c5771edd224dc19020013af7641979ee81ae8ee536d7e3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fe4a40006d1e8439423de1417cc9bc55716d790b882485074bb2a5ec68d7d1ad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:90cf21c36d19d87b97d06edd1c0912a3bb288c769e6159651877db3206aa7355
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0ec1dc184a2199736b4bd7de3241a357be7427caea5da1c931b08bc06f5c3dc0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8e8c081c231b5b15eca483ceec16ea9190ba73aae6e421865162446eb584519a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc4b75ed71845d5b18e1d9163d480fd9d0d3cb5ebbcfed00a2e4e1b174c0a5de
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9492894eaa5f43d00ac80b60029b32a0b0a85ae88ba609b46f98702c9349a5bb
        version: 2.18.2
      entries:
      - name: cockroach-operator.v2.18.2
        version: 2.18.2
      name: stable
    defaultChannel: stable
    packageName: cockroachdb-certified
    provider:
      name: Cockroach Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Elastic
      provider-url: ""
    name: elasticsearch-eck-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: elasticsearch-eck-operator-certified.v3.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "elasticsearch.k8s.elastic.co/v1",
                    "kind": "Elasticsearch",
                    "metadata": {
                        "name": "elasticsearch-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "nodeSets": [
                            {
                                "name": "default",
                                "config": {
                                    "node.roles": ["master", "data"],
                                    "node.attr.attr_name": "attr_value",
                                    "node.store.allow_mmap": false
                                },
                                "podTemplate": {
                                    "metadata": {
                                        "labels": {
                                            "foo": "bar"
                                        }
                                    },
                                    "spec": {
                                        "containers": [
                                            {
                                                "name": "elasticsearch",
                                                "resources": {
                                                    "requests": {
                                                        "memory": "4Gi",
                                                        "cpu": 1
                                                    },
                                                    "limits": {
                                                        "memory": "4Gi",
                                                        "cpu": 2
                                                    }
                                                }
                                            }
                                        ]
                                    }
                                },
                                "count": 3
                            }
                        ]
                    }
                },
                {
                    "apiVersion": "kibana.k8s.elastic.co/v1",
                    "kind": "Kibana",
                    "metadata": {
                        "name": "kibana-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        },
                        "podTemplate": {
                            "metadata": {
                                "labels": {
                                    "foo": "bar"
                                }
                            },
                            "spec": {
                                "containers": [
                                    {
                                        "name": "kibana",
                                        "resources": {
                                            "requests": {
                                                "memory": "1Gi",
                                                "cpu": 0.5
                                            },
                                            "limits": {
                                                "memory": "2Gi",
                                                "cpu": 2
                                            }
                                        }
                                    }
                                ]
                            }
                        }
                    }
                },
                {
                    "apiVersion": "apm.k8s.elastic.co/v1",
                    "kind": "ApmServer",
                    "metadata": {
                        "name": "apmserver-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                    "apiVersion": "enterprisesearch.k8s.elastic.co/v1",
                    "kind": "EnterpriseSearch",
                    "metadata": {
                        "name": "ent-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "config": {
                            "ent_search.external_url": "https://localhost:3002"
                        },
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                  "apiVersion": "beat.k8s.elastic.co/v1beta1",
                  "kind": "Beat",
                  "metadata": {
                    "name": "heartbeat-sample"
                  },
                  "spec": {
                    "type": "heartbeat",
                    "version": "9.1.0",
                    "elasticsearchRef": {
                      "name": "elasticsearch-sample"
                    },
                    "config": {
                      "heartbeat.monitors": [
                        {
                          "type": "tcp",
                          "schedule": "@every 5s",
                          "hosts": [
                            "elasticsearch-sample-es-http.default.svc:9200"
                          ]
                        }
                      ]
                    },
                    "deployment": {
                      "replicas": 1,
                      "podTemplate": {
                        "spec": {
                          "securityContext": {
                            "runAsUser": 0
                          }
                        }
                      }
                    }
                  }
                },
                {
                  "apiVersion": "agent.k8s.elastic.co/v1alpha1",
                  "kind": "Agent",
                  "metadata": {
                    "name": "agent-sample"
                  },
                  "spec": {
                    "version": "9.1.0",
                    "elasticsearchRefs": [
                      {
                        "name": "elasticsearch-sample"
                      }
                    ],
                    "daemonSet": {},
                    "config": {
                      "inputs": [
                        {
                          "name": "system-1",
                          "revision": 1,
                          "type": "system/metrics",
                          "use_output": "default",
                          "meta": {
                            "package": {
                              "name": "system",
                              "version": "0.9.1"
                            }
                          },
                          "data_stream": {
                            "namespace": "default"
                          },
                          "streams": [
                            {
                              "id": "system/metrics-system.cpu",
                              "data_stream": {
                                "dataset": "system.cpu",
                                "type": "metrics"
                              },
                              "metricsets": [
                                "cpu"
                              ],
                              "cpu.metrics": [
                                "percentages",
                                "normalized_percentages"
                              ],
                              "period": "10s"
                            }
                          ]
                        }
                      ]
                    }
                  }
                },
                {
                    "apiVersion": "maps.k8s.elastic.co/v1alpha1",
                    "kind": "ElasticMapsServer",
                    "metadata": {
                        "name": "ems-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                    "apiVersion": "logstash.k8s.elastic.co/v1alpha1",
                    "kind": "Logstash",
                    "metadata" : {
                        "name": "logstash-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.connect.redhat.com/elastic/eck-operator@sha256:1750178c5dd716c031155f2e7058e19d4b7f1d3fdbb3e8a806b177a71be88500
          createdAt: "2025-07-30 11:10:34"
          description: Run Elasticsearch, Kibana, APM Server, Beats, Enterprise Search,
            Elastic Agent, Elastic Maps Server and Logstash on Kubernetes and OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Elastic Basic license
          repository: https://github.com/elastic/cloud-on-k8s
          support: elastic.co
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Elastic Agent instance
            displayName: Elastic Agent
            kind: Agent
            name: agents.agent.k8s.elastic.co
            version: v1alpha1
          - description: APM Server instance
            displayName: APM Server
            kind: ApmServer
            name: apmservers.apm.k8s.elastic.co
            version: v1
          - description: Beats instance
            displayName: Beats
            kind: Beat
            name: beats.beat.k8s.elastic.co
            version: v1beta1
          - description: Elastic Maps Server instance
            displayName: Elastic Maps Server
            kind: ElasticMapsServer
            name: elasticmapsservers.maps.k8s.elastic.co
            version: v1alpha1
          - description: Instance of an Elasticsearch autoscaler
            displayName: Elasticsearch Autoscaler
            kind: ElasticsearchAutoscaler
            name: elasticsearchautoscalers.autoscaling.k8s.elastic.co
            version: v1alpha1
          - description: Instance of an Elasticsearch cluster
            displayName: Elasticsearch Cluster
            kind: Elasticsearch
            name: elasticsearches.elasticsearch.k8s.elastic.co
            version: v1
          - description: Enterprise Search instance
            displayName: Enterprise Search
            kind: EnterpriseSearch
            name: enterprisesearches.enterprisesearch.k8s.elastic.co
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.kibana.k8s.elastic.co
            version: v1
          - description: Logstash instance
            displayName: Logstash
            kind: Logstash
            name: logstashes.logstash.k8s.elastic.co
            version: v1alpha1
          - description: Elastic Stack Config Policy
            displayName: Elastic Stack Config Policy
            kind: StackConfigPolicy
            name: stackconfigpolicies.stackconfigpolicy.k8s.elastic.co
            version: v1alpha1
        description: |-
          Elastic Cloud on Kubernetes (ECK) is the official operator by Elastic for automating the deployment, provisioning, management, and orchestration of Elasticsearch, Kibana, APM Server, Beats, Enterprise Search, Elastic Agent, Elastic Maps Server, and Logstash on Kubernetes.

          Current features:

          *  Elasticsearch, Kibana, APM Server, Enterprise Search, Beats, Elastic Agent, Elastic Maps Server, and Logstash deployments
          *  TLS Certificates management
          *  Safe Elasticsearch cluster configuration and topology changes
          *  Persistent volumes usage
          *  Custom node configuration and attributes
          *  Secure settings keystore updates

          Supported versions:

          * Kubernetes 1.29-1.33
          * OpenShift 4.15-4.19
          * Google Kubernetes Engine (GKE), Azure Kubernetes Service (AKS), and Amazon Elastic Kubernetes Service (EKS)
          * Elasticsearch, Kibana, APM Server: 7.17+, 8+, 9+
          * Enterprise Search: 7.17+, 8+
          * Beats: 7.17+, 8+, 9+
          * Elastic Agent: 7.17+, 8+, 9+
          * Elastic Maps Server: 7.17+, 8+, 9+
          * Logstash 8.12+, 9+

          ECK should work with all conformant installers as listed in these [FAQs](https://github.com/cncf/k8s-conformance/blob/master/faq.md#what-is-a-distribution-hosted-platform-and-an-installer). Distributions include source patches and so may not work as-is with ECK.
          Alpha, beta, and stable API versions follow the same [conventions used by Kubernetes](https://kubernetes.io/docs/concepts/overview/kubernetes-api/#api-versioning).
          See the full [Elastic support matrix](https://www.elastic.co/support/matrix#matrix_kubernetes) for more information.
          See the [Quickstart](https://www.elastic.co/guide/en/cloud-on-k8s/3.1/k8s-quickstart.html) to get started with ECK.
        displayName: Elasticsearch (ECK) Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - analytics
        - search
        - database
        - apm
        - logstash
        links:
        - name: Documentation
          url: https://www.elastic.co/guide/en/cloud-on-k8s/3.1/index.html
        maintainers:
        - email: eck@elastic.co
          name: Elastic
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Elastic
        relatedImages:
        - registry.connect.redhat.com/elastic/eck-operator@sha256:1750178c5dd716c031155f2e7058e19d4b7f1d3fdbb3e8a806b177a71be88500
        version: 3.1.0
      entries:
      - name: elasticsearch-eck-operator-certified.v3.1.0
        version: 3.1.0
      - name: elasticsearch-eck-operator-certified.v3.0.0
        version: 3.0.0
      - name: elasticsearch-eck-operator-certified.v2.16.1
        version: 2.16.1
      - name: elasticsearch-eck-operator-certified.v2.16.0
        version: 2.16.0
      - name: elasticsearch-eck-operator-certified.v2.15.0
        version: 2.15.0
      - name: elasticsearch-eck-operator-certified.v2.14.0
        version: 2.14.0
      - name: elasticsearch-eck-operator-certified.v2.13.0
        version: 2.13.0
      - name: elasticsearch-eck-operator-certified.v2.12.1
        version: 2.12.1
      - name: elasticsearch-eck-operator-certified.v2.12.0
        version: 2.12.0
      - name: elasticsearch-eck-operator-certified.v2.11.1
        version: 2.11.1
      - name: elasticsearch-eck-operator-certified.v2.11.0
        version: 2.11.0
      - name: elasticsearch-eck-operator-certified.v2.10.0
        version: 2.10.0
      - name: elasticsearch-eck-operator-certified.v2.9.0
        version: 2.9.0
      - name: elasticsearch-eck-operator-certified.v2.8.0
        version: 2.8.0
      - name: elasticsearch-eck-operator-certified.v2.7.0
        version: 2.7.0
      - name: elasticsearch-eck-operator-certified.v2.6.2
        version: 2.6.2
      - name: elasticsearch-eck-operator-certified.v2.6.1
        version: 2.6.1
      - name: elasticsearch-eck-operator-certified.v2.5.0
        version: 2.5.0
      - name: elasticsearch-eck-operator-certified.v2.4.0
        version: 2.4.0
      - name: elasticsearch-eck-operator-certified.v2.3.0
        version: 2.3.0
      - name: elasticsearch-eck-operator-certified.v2.2.0
        version: 2.2.0
      - name: elasticsearch-eck-operator-certified.v2.1.0
        version: 2.1.0
      - name: elasticsearch-eck-operator-certified.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: elasticsearch-eck-operator-certified
    provider:
      name: Elastic
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Portworx by PureStorage
      provider-url: ""
    name: portworx-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: portworx-operator.v25.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageCluster",
                "metadata": {
                  "name": "portworx",
                  "namespace": "test-operator",
                  "annotations": {
                    "portworx.io/is-openshift": "true"
                  }
                },
                "spec": {}
              },
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageNode",
                "metadata": {
                  "name": "example",
                  "namespace": "test-operator"
                },
                "spec": {}
              }
            ]
          capabilities: Auto Pilot
          categories: Storage
          certified: "true"
          console.openshift.io/plugins: '["portworx"]'
          containerImage: registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
          createdAt: "2025-09-06T19:38:43Z"
          description: Cloud native storage solution for production workloads
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "core.libopenstorage.org/v1",
              "kind": "StorageCluster",
              "metadata": {
                "name": "portworx",
                "annotations": {
                  "portworx.io/is-openshift": "true"
                }
              }
            }
          support: Portworx, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Storage Cluster installs Portworx in the cluster. It has
              all the necessary configurations to setup and update a Portworx cluster.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.core.libopenstorage.org
            version: v1
          - description: Do not create Storage Node as it is internally created by
              the operator. It represents the status of a Portworx node.
            displayName: Storage Node
            kind: StorageNode
            name: storagenodes.core.libopenstorage.org
            version: v1
        description: |
          Portworx-Enterprise is the most widely-used and reliable cloud-native
          storage solution for production workloads and provides high-availability,
          data protection and security for containerized applications.

          Portworx Enterprise enables you to migrate entire applications, including
          data, between clusters in a single data center or cloud, or between clouds,
          with a single kubectl command.

          The cloud native storage and data management platform that enterprises trust
          to manage data in containers now has an operator which simplifies the install,
          configuration, upgrades and manages the Portworx Enterprise cluster lifecycle.

          Learn more about the Portworx Enterprise
          [the data platform for Kubernetes](https://portworx.com/products/introduction)

          To learn more about the platform features, please visit our
          [product features page](https://portworx.com/products/features)

          ### About Portworx

          Portworx is the solution for running stateful containers in production,
          designed with DevOps in mind. With Portworx, users can manage any database
          or stateful service on any infrastructure using any container scheduler,
          including Kubernetes, Mesosphere DC/OS, and Docker Swarm. Portworx solves
          the five most common problems DevOps teams encounter when running stateful
          services in production: persistence, high availability, data automation,
          security, and support for multiple data stores and infrastructure.

          ### How to install StorageCluster

          To customize your cluster's configuration (specification), use the
          [Spec Generator](https://central.portworx.com/) from PX-Central.

          ### Prerequisite

          Ensure ports 17001-17020 on worker nodes are reachable from master and other worker nodes.

          ### Tutorials

          * [Portworx Enterprise on Openshift](https://portworx.com/openshift)

          * [Stateful applications on Kubernetes](https://docs.portworx.com/portworx-enterprise/concepts/kubernetes-storage-101/applications)

          * [Portworx Enterprise on Kubernetes](https://docs.portworx.com/portworx-install-with-kubernetes)

          * [Kafka on Kubernetes](https://portworx.com/kafka-kubernetes)

          * [Elastisearch on Kubernetes](https://portworx.com/elasticsearch-kubernetes)

          * [PostgreSQL on Kubernetes](https://portworx.com/postgres-kubernetes/)

          * [MongoDB on Kubernetes](https://portworx.com/mongodb-kubernetes/)

          * [Cassandra on Kubernetes](https://portworx.com/cassandra-kubernetes/)

          * [Kubernetes backup and recovery](https://portworx.com/kubernetes-backup/)

          * [Disaster Recovery for Kubernetes](https://portworx.com/kubernetes-disaster-recovery/)

          ### Uninstall

          Deleting the StorageCluster object for Portworx cluster does not stop Portworx
          service running on the nodes, to avoid application downtime.

          To uninstall Portworx completely without wiping the data, you should add the
          following delete strategy to the StorageCluster spec:
          ```
          spec:
            deleteStrategy:
              type: Uninstall
          ```
          **Caution:** To uninstall Portworx and **wipe all the data**, you should use the following
          delete strategy:
          ```
          spec:
            deleteStrategy:
              type: UninstallAndWipe
          ```
        displayName: Portworx Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - portworx
        - persistent storage
        - storage
        - cloud native
        - open source
        links:
        - name: Product Features
          url: https://portworx.com/products/features
        - name: Documentation
          url: https://docs.portworx.com/portworx-enterprise/platform/openshift
        - name: Support
          url: https://docs.portworx.com/portworx-enterprise/support/contact-support
        - name: Source Code
          url: https://github.com/libopenstorage/operator
        maintainers:
        - email: support@portworx.com
          name: Portworx
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Portworx by PureStorage
        relatedImages:
        - registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
        version: 25.3.1
      entries:
      - name: portworx-operator.v25.3.1
        version: 25.3.1
      - name: portworx-operator.v25.3.0
        version: 25.3.0
      - name: portworx-operator.v25.2.2
        version: 25.2.2
      - name: portworx-operator.v25.2.1
        version: 25.2.1
      - name: portworx-operator.v25.2.0
        version: 25.2.0
      - name: portworx-operator.v24.2.4
        version: 24.2.4
      - name: portworx-operator.v24.2.3
        version: 24.2.3
      - name: portworx-operator.v24.2.2
        version: 24.2.2
      - name: portworx-operator.v24.2.1
        version: 24.2.1
      - name: portworx-operator.v24.2.0
        version: 24.2.0
      - name: portworx-operator.v24.1.3
        version: 24.1.3
      - name: portworx-operator.v24.1.2
        version: 24.1.2
      - name: portworx-operator.v24.1.1
        version: 24.1.1
      - name: portworx-operator.v24.1.0
        version: 24.1.0
      - name: portworx-operator.v23.10.5
        version: 23.10.5
      - name: portworx-operator.v23.10.4
        version: 23.10.4
      - name: portworx-operator.v23.10.3
        version: 23.10.3
      - name: portworx-operator.v23.10.2
        version: 23.10.2
      - name: portworx-operator.v23.10.1
        version: 23.10.1
      - name: portworx-operator.v23.7.0
        version: 23.7.0
      - name: portworx-operator.v23.5.1
        version: 23.5.1
      - name: portworx-operator.v23.5.0
        version: 23.5.0
      - name: portworx-operator.v23.4.0
        version: 23.4.0
      - name: portworx-operator.v23.3.1
        version: 23.3.1
      - name: portworx-operator.v23.3.0
        version: 23.3.0
      - name: portworx-operator.v1.10.5
        version: 1.10.5
      - name: portworx-operator.v1.10.4
        version: 1.10.4
      - name: portworx-operator.v1.10.3
        version: 1.10.3
      - name: portworx-operator.v1.10.2
        version: 1.10.2
      - name: portworx-operator.v1.10.1
        version: 1.10.1
      - name: portworx-operator.v1.10.0
        version: 1.10.0
      - name: portworx-operator.v1.9.1
        version: 1.9.1
      - name: portworx-operator.v1.9.0
        version: 1.9.0
      - name: portworx-operator.v1.8.1
        version: 1.8.1
      - name: portworx-operator.v1.8.0
        version: 1.8.0
      - name: portworx-operator.v1.7.1
        version: 1.7.1
      - name: portworx-operator.v1.6.1
        version: 1.6.1
      name: alpha
    - currentCSV: portworx-operator.v25.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageCluster",
                "metadata": {
                  "name": "portworx",
                  "namespace": "test-operator",
                  "annotations": {
                    "portworx.io/is-openshift": "true"
                  }
                },
                "spec": {}
              },
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageNode",
                "metadata": {
                  "name": "example",
                  "namespace": "test-operator"
                },
                "spec": {}
              }
            ]
          capabilities: Auto Pilot
          categories: Storage
          certified: "true"
          console.openshift.io/plugins: '["portworx"]'
          containerImage: registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
          createdAt: "2025-09-06T19:38:43Z"
          description: Cloud native storage solution for production workloads
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "core.libopenstorage.org/v1",
              "kind": "StorageCluster",
              "metadata": {
                "name": "portworx",
                "annotations": {
                  "portworx.io/is-openshift": "true"
                }
              }
            }
          support: Portworx, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Storage Cluster installs Portworx in the cluster. It has
              all the necessary configurations to setup and update a Portworx cluster.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.core.libopenstorage.org
            version: v1
          - description: Do not create Storage Node as it is internally created by
              the operator. It represents the status of a Portworx node.
            displayName: Storage Node
            kind: StorageNode
            name: storagenodes.core.libopenstorage.org
            version: v1
        description: |
          Portworx-Enterprise is the most widely-used and reliable cloud-native
          storage solution for production workloads and provides high-availability,
          data protection and security for containerized applications.

          Portworx Enterprise enables you to migrate entire applications, including
          data, between clusters in a single data center or cloud, or between clouds,
          with a single kubectl command.

          The cloud native storage and data management platform that enterprises trust
          to manage data in containers now has an operator which simplifies the install,
          configuration, upgrades and manages the Portworx Enterprise cluster lifecycle.

          Learn more about the Portworx Enterprise
          [the data platform for Kubernetes](https://portworx.com/products/introduction)

          To learn more about the platform features, please visit our
          [product features page](https://portworx.com/products/features)

          ### About Portworx

          Portworx is the solution for running stateful containers in production,
          designed with DevOps in mind. With Portworx, users can manage any database
          or stateful service on any infrastructure using any container scheduler,
          including Kubernetes, Mesosphere DC/OS, and Docker Swarm. Portworx solves
          the five most common problems DevOps teams encounter when running stateful
          services in production: persistence, high availability, data automation,
          security, and support for multiple data stores and infrastructure.

          ### How to install StorageCluster

          To customize your cluster's configuration (specification), use the
          [Spec Generator](https://central.portworx.com/) from PX-Central.

          ### Prerequisite

          Ensure ports 17001-17020 on worker nodes are reachable from master and other worker nodes.

          ### Tutorials

          * [Portworx Enterprise on Openshift](https://portworx.com/openshift)

          * [Stateful applications on Kubernetes](https://docs.portworx.com/portworx-enterprise/concepts/kubernetes-storage-101/applications)

          * [Portworx Enterprise on Kubernetes](https://docs.portworx.com/portworx-install-with-kubernetes)

          * [Kafka on Kubernetes](https://portworx.com/kafka-kubernetes)

          * [Elastisearch on Kubernetes](https://portworx.com/elasticsearch-kubernetes)

          * [PostgreSQL on Kubernetes](https://portworx.com/postgres-kubernetes/)

          * [MongoDB on Kubernetes](https://portworx.com/mongodb-kubernetes/)

          * [Cassandra on Kubernetes](https://portworx.com/cassandra-kubernetes/)

          * [Kubernetes backup and recovery](https://portworx.com/kubernetes-backup/)

          * [Disaster Recovery for Kubernetes](https://portworx.com/kubernetes-disaster-recovery/)

          ### Uninstall

          Deleting the StorageCluster object for Portworx cluster does not stop Portworx
          service running on the nodes, to avoid application downtime.

          To uninstall Portworx completely without wiping the data, you should add the
          following delete strategy to the StorageCluster spec:
          ```
          spec:
            deleteStrategy:
              type: Uninstall
          ```
          **Caution:** To uninstall Portworx and **wipe all the data**, you should use the following
          delete strategy:
          ```
          spec:
            deleteStrategy:
              type: UninstallAndWipe
          ```
        displayName: Portworx Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - portworx
        - persistent storage
        - storage
        - cloud native
        - open source
        links:
        - name: Product Features
          url: https://portworx.com/products/features
        - name: Documentation
          url: https://docs.portworx.com/portworx-enterprise/platform/openshift
        - name: Support
          url: https://docs.portworx.com/portworx-enterprise/support/contact-support
        - name: Source Code
          url: https://github.com/libopenstorage/operator
        maintainers:
        - email: support@portworx.com
          name: Portworx
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Portworx by PureStorage
        relatedImages:
        - registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
        version: 25.3.1
      entries:
      - name: portworx-operator.v25.3.1
        version: 25.3.1
      - name: portworx-operator.v25.3.0
        version: 25.3.0
      - name: portworx-operator.v25.2.2
        version: 25.2.2
      - name: portworx-operator.v25.2.1
        version: 25.2.1
      - name: portworx-operator.v25.2.0
        version: 25.2.0
      - name: portworx-operator.v24.2.4
        version: 24.2.4
      - name: portworx-operator.v24.2.3
        version: 24.2.3
      - name: portworx-operator.v24.2.2
        version: 24.2.2
      - name: portworx-operator.v24.2.1
        version: 24.2.1
      - name: portworx-operator.v24.2.0
        version: 24.2.0
      - name: portworx-operator.v24.1.3
        version: 24.1.3
      - name: portworx-operator.v24.1.2
        version: 24.1.2
      - name: portworx-operator.v24.1.1
        version: 24.1.1
      - name: portworx-operator.v24.1.0
        version: 24.1.0
      - name: portworx-operator.v23.10.5
        version: 23.10.5
      - name: portworx-operator.v23.10.4
        version: 23.10.4
      - name: portworx-operator.v23.10.3
        version: 23.10.3
      - name: portworx-operator.v23.10.2
        version: 23.10.2
      - name: portworx-operator.v23.10.1
        version: 23.10.1
      - name: portworx-operator.v23.7.0
        version: 23.7.0
      - name: portworx-operator.v23.5.1
        version: 23.5.1
      - name: portworx-operator.v23.5.0
        version: 23.5.0
      - name: portworx-operator.v23.4.0
        version: 23.4.0
      - name: portworx-operator.v23.3.1
        version: 23.3.1
      - name: portworx-operator.v23.3.0
        version: 23.3.0
      - name: portworx-operator.v1.10.5
        version: 1.10.5
      - name: portworx-operator.v1.10.4
        version: 1.10.4
      - name: portworx-operator.v1.10.3
        version: 1.10.3
      - name: portworx-operator.v1.10.2
        version: 1.10.2
      - name: portworx-operator.v1.10.1
        version: 1.10.1
      - name: portworx-operator.v1.10.0
        version: 1.10.0
      - name: portworx-operator.v1.9.1
        version: 1.9.1
      - name: portworx-operator.v1.9.0
        version: 1.9.0
      - name: portworx-operator.v1.8.1
        version: 1.8.1
      - name: portworx-operator.v1.8.0
        version: 1.8.0
      - name: portworx-operator.v1.7.1
        version: 1.7.1
      - name: portworx-operator.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: portworx-certified
    provider:
      name: Portworx by PureStorage
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: clusterresourceoverride
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clusterresourceoverride-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.autoscaling.openshift.io/v1",
                "kind": "ClusterResourceOverride",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "podResourceOverride": {
                    "spec": {
                      "memoryRequestToLimitPercent": 50,
                      "cpuRequestToLimitPercent": 25,
                      "limitCPUToMemoryPercent": 200
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certifiedLevel: Primed
          containerImage: registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9-operator@sha256:40b19c94f688c3a58189fcb404df35a40a8acbbdbf73fadb8818a86fd14a9ab9
          createdAt: 2019/11/15
          description: An operator to manage the OpenShift ClusterResourceOverride
            Mutating Admission Webhook Server
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          healthIndex: B
          olm.skipRange: '>=4.3.0 <4.16.0-202509240123'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-resource-override-admission-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of ClusterResourceOverride Admission
              Webhook
            displayName: ClusterResourceOverride
            kind: ClusterResourceOverride
            name: clusterresourceoverrides.operator.autoscaling.openshift.io
            version: v1
        description: "ClusterResourceOverride\n==============\n\nContainers can specify
          compute resource requests and limits. Requests are used for scheduling your
          container and provide a minimum service guarantee. Limits constrain the
          amount of compute resource that may be consumed on your node.\n\nThe scheduler
          attempts to optimize the compute resource use across all nodes in your cluster.
          It places pods onto specific nodes, taking the pods' compute resource requests
          and nodes' available capacity into consideration.\n\nRequests and limits
          enable administrators to allow and manage the overcommitment of resources
          on a node, which may be desirable in development environments where a trade
          off of guaranteed performance for capacity is acceptable.\n\n### Requests
          and Limits\n\nFor each compute resource, a container may specify a resource
          request and limit. Scheduling decisions are made based on the request to
          ensure that a node has enough capacity available to meet the requested value.
          If a container specifies limits, but omits requests, the requests are defaulted
          to the limits. A container is not able to exceed the specified limit on
          the node.\n\nThe enforcement of limits is dependent upon the compute resource
          type. If a container makes no request or limit, the container is scheduled
          to a node with no resource guarantees. In practice, the container is able
          to consume as much of the specified resource as is available with the lowest
          local priority. In low resource situations, containers that specify no resource
          requests are given the lowest quality of service.\n\n### Compute Resources\n\nThe
          node-enforced behavior for compute resources is specific to the resource
          type.\n\n#### CPU\n\nA container is guaranteed the amount of CPU it requests
          and is additionally able to consume excess CPU available on the node, up
          to any limit specified by the container. If multiple containers are attempting
          to use excess CPU, CPU time is distributed based on the amount of CPU requested
          by each container.\n\nFor example, if one container requested 500m of CPU
          time and another container requested 250m of CPU time, then any extra CPU
          time available on the node is distributed among the containers in a 2:1
          ratio. If a container specified a limit, it will be throttled not to use
          more CPU than the specified limit.\n\nCPU requests are enforced using the
          CFS shares support in the Linux kernel. By default, CPU limits are enforced
          using the CFS quota support in the Linux kernel over a 100ms measuring interval,
          though this can be disabled.\n\n#### Memory\n\nA container is guaranteed
          the amount of memory it requests. A container may use more memory than requested,
          but once it exceeds its requested amount, it could be killed in a low memory
          situation on the node.\n\nIf a container uses less memory than requested,
          it will not be killed unless system tasks or daemons need more memory than
          was accounted for in the node’s resource reservation. If a container specifies
          a limit on memory, it is immediately killed if it exceeds the limit amount.\n\n###
          Configuring the Cluster for Overcommitment\n\nScheduling is based on resources
          requested, while quota and hard limits refer to resource limits, which can
          be set higher than requested resources. The difference between request and
          limit determines the level of overcommit; for instance, if a container is
          given a memory request of 1Gi and a memory limit of 2Gi, it is scheduled
          based on the 1Gi request being available on the node, but could use up to
          2Gi; so it is 200% overcommitted. If OpenShift Container Platform administrators
          would like to control the level of overcommit and manage container density
          on nodes, ClusterResourceOverride Admission Webhook can be configured to
          override the ratio between request and limit set on developer containers.
          In conjunction with a per-project LimitRange specifying limits and defaults,
          this adjusts the container limit and request to achieve the desired level
          of overcommit.\n\nThis requires creating a custom resource of `ClusterResourceOverride`
          type as in the following example:\n\n    \n    - apiVersion: operator.autoscaling.openshift.io/v1\n
          \   - kind: ClusterResourceOverride\n    - metadata:\n    -   name: cluster\n
          \   - spec:\n    -   podResourceOverride:\n    -     spec:\n    -       memoryRequestToLimitPercent:
          25\n    -       cpuRequestToLimitPercent: 25\n    -       limitCPUToMemoryPercent:
          200\n    \n **memoryRequestToLimitPercent**: (optional, 1-100) If a container
          memory limit has been specified or defaulted, the memory request is overridden
          to this percentage of the limit.\n\n **cpuRequestToLimitPercent**: (optional,
          1-100) If a container CPU limit has been specified or defaulted, the CPU
          request is overridden to this percentage of the limit.\n\n **limitCPUToMemoryPercent**:
          (optional, positive integer) If a container memory limit has been specified
          or defaulted, the CPU limit is overridden to a percentage of the memory
          limit, with a 100 percentage scaling 1Gi of RAM to equal 1 CPU core. This
          is processed prior to overriding CPU request (if configured).\n\n Note that
          these overrides have no effect if no limits have been set on containers.
          [Create a LimitRange object] (https://docs.openshift.com/container-platform/3.3/admin_guide/limits.html#admin-guide-limits)
          with default limits (per individual project, or in the [project template](https://docs.openshift.com/container-platform/3.3/admin_guide/managing_projects.html#modifying-the-template-for-new-projects))
          in order to ensure that the overrides apply.\n\nWhen configured, overrides
          can be enabled per-project by applying the following label.\n    ```\n    clusterresourceoverrides.admission.autoscaling.openshift.io/enabled:
          \"true\"\n    ```\n\n"
        displayName: ClusterResourceOverride Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - deschedule
        - scale
        - binpack
        - efficiency
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9-operator@sha256:40b19c94f688c3a58189fcb404df35a40a8acbbdbf73fadb8818a86fd14a9ab9
        - registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9@sha256:e4992c556fe6be50fa30a45b05cffd6029ebd8914eca7c1c683564ff869e0e7d
        version: 4.16.0-202509240123
      entries:
      - name: clusterresourceoverride-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: clusterresourceoverride-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: clusterresourceoverride-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: clusterresourceoverride-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: clusterresourceoverride-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: clusterresourceoverride-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: clusterresourceoverride-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: clusterresourceoverride-operator.v4.16.0-202505092110
        version: 4.16.0-202505092110
      - name: clusterresourceoverride-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: clusterresourceoverride-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: clusterresourceoverride-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: clusterresourceoverride-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: clusterresourceoverride-operator.v4.16.0-202501241104
        version: 4.16.0-202501241104
      - name: clusterresourceoverride-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: clusterresourceoverride-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: clusterresourceoverride-operator.v4.16.0-202412021735
        version: 4.16.0-202412021735
      - name: clusterresourceoverride-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: clusterresourceoverride-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: clusterresourceoverride-operator.v4.16.0-202411081636
        version: 4.16.0-202411081636
      - name: clusterresourceoverride-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: clusterresourceoverride-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: clusterresourceoverride-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: clusterresourceoverride-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: clusterresourceoverride-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: clusterresourceoverride-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: clusterresourceoverride-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: clusterresourceoverride-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: clusterresourceoverride-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: clusterresourceoverride-operator.v4.16.0-202407040206
        version: 4.16.0-202407040206
      - name: clusterresourceoverride-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: clusterresourceoverride-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: clusterresourceoverride-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: clusterresourceoverride-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: clusterresourceoverride
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicurio-registry-3
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-3.v3.0.15-r1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "simple"
                },
                "spec": {
                  "app": {
                    "ingress": {
                      "host": "simple-app.apps.cluster.example"
                    }
                  },
                  "ui": {
                    "ingress": {
                      "host": "simple-ui.apps.cluster.example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/apicurio/apicurio-registry-rhel9-operator@sha256:e6d2ceab8accdb6e1b9766733039e24df620d7128eb1e13e60a61339130d52a7
          createdAt: "2025-09-29"
          description: Deploy and manage Red Hat build of Apicurio Registry 3 on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations",
            "Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Red Hat build
              of Apicurio Registry version 3.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Red Hat build of Apicurio Registry 3

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Allows you manage event schemas and API designs using the Apicurio Registry web console, REST API, Maven plug-in, or Java client
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, MySQL\*, SQL Server\*, in-memory\*
          - Compatible with Confluent APIs
          - Runs on a lightweight Quarkus platform

          _Note: Storage options marked with * have not yet been tested by Red Hat for production deployments._

          ## Red Hat build of Apicurio Registry 3 Operator

          Provides a quick and easy way to deploy and manage Red Hat build of Apicurio Registry 3 on OpenShift.

          **Features**
          - Supports installation and configuration of Red Hat build of Apicurio Registry 3
          - Perform a rolling upgrade of Red Hat build of Apicurio Registry 3

          ## Prerequisites

          This operator does not deploy storage for Red Hat build of Apicurio Registry 3. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Red Hat build of Apicurio Registry 3 and Apicurio Registry 3 Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Red Hat build of Apicurio Registry 3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/apicurio/apicurio-registry-ui-rhel9@sha256:e4791546fd371b34c7580c6aca2faed8527c985c8029fa8087fb46fd94ab8feb
        - registry.redhat.io/apicurio/apicurio-registry-rhel9-operator@sha256:e6d2ceab8accdb6e1b9766733039e24df620d7128eb1e13e60a61339130d52a7
        - registry.redhat.io/apicurio/apicurio-registry-rhel9@sha256:c86d73d5e2b4146f016ce7ec67265fe80bd596f472247eb9f5590cc0c89dba75
        version: 3.0.15-r1
      entries:
      - name: apicurio-registry-3.v3.0.15-r1
        version: 3.0.15-r1
      - name: apicurio-registry-3.v3.0.9-r3
        version: 3.0.9-r3
      - name: apicurio-registry-3.v3.0.9-r2
        version: 3.0.9-r2
      - name: apicurio-registry-3.v3.0.9-r1
        version: 3.0.9-r1
      - name: apicurio-registry-3.v3.0.7-r2
        version: 3.0.7-r2
      - name: apicurio-registry-3.v3.0.7-r1
        version: 3.0.7-r1
      - name: apicurio-registry-3.v3.0.6-beta.r1
        version: 3.0.6-beta.r1
      - name: apicurio-registry-3.v3.0.6-beta.r1-0.1740495889.p
        version: 3.0.6-beta.r1+0.1740495889.p
      - name: apicurio-registry-3.v3.0.6-beta
        version: 3.0.6-beta
      name: 3.x
    defaultChannel: 3.x
    packageName: apicurio-registry-3
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: MariaDB Corporation
      provider-url: ""
    name: mariadb-enterprise-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-enterprise-operator.v25.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup"
                },
                "spec": {
                  "args": [
                    "--single-transaction",
                    "--all-databases"
                  ],
                  "logLevel": "info",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "0 0 * * *",
                    "suspend": false
                  },
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "backups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Connection",
                "metadata": {
                  "name": "connection"
                },
                "spec": {
                  "database": "mariadb",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "secretName": "connection",
                  "secretTemplate": {
                    "key": "dsn"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Database",
                "metadata": {
                  "name": "database"
                },
                "spec": {
                  "characterSet": "utf8",
                  "collate": "utf8_general_ci",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "name": "my-logical-database"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "grant"
                },
                "spec": {
                  "database": "*",
                  "grantOption": true,
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "privileges": [
                    "SELECT",
                    "INSERT",
                    "UPDATE"
                  ],
                  "table": "*",
                  "username": "my-user"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "MariaDB",
                "metadata": {
                  "name": "mariadb"
                },
                "spec": {
                  "connection": {
                    "secretName": "mariadb-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "database": "mariadb",
                  "galera": {
                    "enabled": true
                  },
                  "metrics": {
                    "enabled": true,
                    "passwordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "mariadb-metrics"
                    }
                  },
                  "myCnf": "[mariadb]\nbind-address=*\ndefault_storage_engine=InnoDB\nbinlog_format=row\ninnodb_autoinc_lock_mode=2\ninnodb_buffer_pool_size=1024M\nmax_allowed_packet=256M\n",
                  "passwordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-password"
                  },
                  "primaryConnection": {
                    "secretName": "mariadb-conn-primary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "primaryService": {
                    "type": "ClusterIP"
                  },
                  "replicas": 3,
                  "rootPasswordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-root"
                  },
                  "secondaryConnection": {
                    "secretName": "mariadb-conn-secondary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "secondaryService": {
                    "type": "ClusterIP"
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "storage": {
                    "size": "1Gi"
                  },
                  "updateStrategy": {
                    "type": "ReplicasFirstPrimaryLast"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "MaxScale",
                "metadata": {
                  "name": "maxscale"
                },
                "spec": {
                  "admin": {
                    "guiEnabled": true,
                    "port": 8989
                  },
                  "auth": {
                    "adminPasswordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "maxscale-admin"
                    },
                    "adminUsername": "mariadb-enterprise-operator",
                    "generate": true
                  },
                  "config": {
                    "sync": {
                      "database": "mysql",
                      "interval": "5s",
                      "timeout": "10s"
                    }
                  },
                  "connection": {
                    "port": 3306,
                    "secretName": "maxscale-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "guiKubernetesService": {
                    "type": "ClusterIP"
                  },
                  "kubernetesService": {
                    "type": "ClusterIP"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "metrics": {
                    "enabled": true
                  },
                  "monitor": {
                    "cooperativeMonitoring": "majority_of_all",
                    "interval": "2s",
                    "params": {
                      "available_when_donor": "false",
                      "disable_master_failback": "false",
                      "disable_master_role_setting": "false"
                    }
                  },
                  "replicas": 2,
                  "requeueInterval": "10s",
                  "services": [
                    {
                      "listener": {
                        "port": 3306,
                        "protocol": "MariaDBProtocol"
                      },
                      "name": "rw-router",
                      "params": {
                        "transaction_replay": "true",
                        "transaction_replay_attempts": "10",
                        "transaction_replay_timeout": "5s"
                      },
                      "router": "readwritesplit"
                    }
                  ]
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "PhysicalBackup",
                "metadata": {
                  "name": "physicalbackup"
                },
                "spec": {
                  "compression": "bzip2",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "podAffinity": true,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "0 0 * * *",
                    "immediate": true,
                    "suspend": false
                  },
                  "serviceAccountName": "backup",
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "physicalbackups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "region": "us-east-1",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  },
                  "timeout": "1h"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore"
                },
                "spec": {
                  "backupRef": {
                    "name": "backup"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "targetRecoveryTime": "2023-12-19T09:00:00Z"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "SqlJob",
                "metadata": {
                  "name": "sqljob"
                },
                "spec": {
                  "database": "mariadb",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "sql": "CREATE TABLE IF NOT EXISTS orders (\n  id bigint PRIMARY KEY AUTO_INCREMENT,\n  customer_id INT,\n  order_date DATE,\n  total_amount_cents INT\n);\nINSERT INTO orders (customer_id, order_date, total_amount_cents) VALUES\n  (101, '2024-05-15', 5000),\n  (102, '2024-05-14', 7525),\n  (103, '2024-05-13', 10050),\n  (104, '2024-05-12', 2575);",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "user"
                },
                "spec": {
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxUserConnections": 20,
                  "name": "my-user",
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.mariadb.com/mariadb-enterprise-operator@sha256:1effbbfa7632be92c17ef61c18a25f010b187840beec42c31d2410d0209e6985
          createdAt: "2025-08-28T11:35:03Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["MariaDB Enterprise"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: MariaDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API. It is used to define
              backup jobs and its storage.
            displayName: Backup
            kind: Backup
            name: backups.enterprise.mariadb.com
            version: v1alpha1
          - description: Connection is the Schema for the connections API. It is used
              to configure connection strings for the applications connecting to MariaDB.
            displayName: Connection
            kind: Connection
            name: connections.enterprise.mariadb.com
            version: v1alpha1
          - description: Database is the Schema for the databases API. It is used
              to define a logical database as if you were running a 'CREATE DATABASE'
              statement.
            displayName: Database
            kind: Database
            name: databases.enterprise.mariadb.com
            version: v1alpha1
          - description: Grant is the Schema for the grants API. It is used to define
              grants as if you were running a 'GRANT' statement.
            displayName: Grant
            kind: Grant
            name: grants.enterprise.mariadb.com
            version: v1alpha1
          - description: MariaDB is the Schema for the mariadbs API. It is used to
              define MariaDB clusters.
            displayName: Maria DB
            kind: MariaDB
            name: mariadbs.enterprise.mariadb.com
            version: v1alpha1
          - description: MaxScale is the Schema for the maxscales API. It is used
              to define MaxScale clusters.
            displayName: Max Scale
            kind: MaxScale
            name: maxscales.enterprise.mariadb.com
            version: v1alpha1
          - description: PhysicalBackup is the Schema for the physicalbackups API.
              It is used to define physical backup jobs and its storage.
            displayName: Physical Backup
            kind: PhysicalBackup
            name: physicalbackups.enterprise.mariadb.com
            version: v1alpha1
          - description: Restore is the Schema for the restores API. It is used to
              define restore jobs and its restoration source.
            displayName: Restore
            kind: Restore
            name: restores.enterprise.mariadb.com
            version: v1alpha1
          - description: SqlJob is the Schema for the sqljobs API. It is used to run
              sql scripts as jobs.
            displayName: Sql Job
            kind: SqlJob
            name: sqljobs.enterprise.mariadb.com
            version: v1alpha1
          - description: User is the Schema for the users API.  It is used to define
              grants as if you were running a 'CREATE USER' statement.
            displayName: User
            kind: User
            name: users.enterprise.mariadb.com
            version: v1alpha1
        description: |
          MariaDB Enterprise Operator for Openshift is an operator that manages MariaDB Enterprise operands declaratively via Custom Resources, with support for the following features:

          * Provision and Configure MariaDB and MaxScale Declaratively: Define MariaDB Enterprise Server and MaxScale clusters in YAML manifests and deploy them with ease in Kubernetes.
          * High Availability with Galera: Ensure availability with MariaDB Enterprise Cluster, providing synchronous multi-master replication.
          * Query and Connection-Based Routing with MaxScale: MaxScale provides query routing and connection load balancing for improved application performance.
          * Cluster-Aware Rolling Updates: Perform rolling updates on MariaDB and MaxScale clusters, ensuring zero-downtime upgrades with no disruptions to your applications.
          * Flexible Storage Configuration and Volume Expansion: Easily configure storage for MariaDB instances, including the ability to expand volumes as needed.
          * Physical Backups based on [mariadb-backup](https://mariadb.com/docs/server/server-usage/backup-and-restore/mariadb-backup/full-backup-and-restore-with-mariadb-backup) and [Kubernetes VolumeSnapshots](https://kubernetes.io/docs/concepts/storage/volume-snapshots/). By leveraging the [BACKUP STAGE](https://mariadb.com/docs/server/reference/sql-statements/administrative-sql-statements/backup-commands/backup-stage) feature, backups are taken without long read locks or service interruptions.
          * Logical Backups based on [mariadb-dump](https://mariadb.com/docs/server/clients-and-utilities/backup-restore-and-import-clients/mariadb-dump).
          * Backup Management: Take, restore, and schedule backups with multiple storage types supported: S3, PVCs, Kubernetes volumes and VolumeSnapshots..
          * Policy-Driven Backup Retention: Implement backup retention policies with bzip2 and gzip compression.
          * Target Recovery Time: Restore your database to the closest available backup based on a specified recovery time.
          * Bootstrap New Instances: Initialize new MariaDB instances from backups, S3, PVCs or VolumeSnapshots to quickly spin up new clusters.
          * TLS Certificate Management: Issue, configure, and rotate TLS certificates and Certificate Authorities (CAs) for secure connections.
          * Advanced TLS Support: customize certificate lifetime, private key algorithm and TLS version.
          * Native Integration with cert-manager: Leverage [cert-manager](https://cert-manager.io/docs/), the de-facto standard for managing certificates in Kubernetes, to enable issuance with private CAs, public CAs and HashiCorp Vault.
          * Prometheus Metrics: Expose metrics using the MariaDB and MaxScale Prometheus exporters.
          * Native Integration with prometheus-operator: Leverage [prometheus-operator](https://github.com/prometheus-operator/prometheus-operator) to scrape metrics from MariaDB and MaxScale instances.
          * Declarative User and Database Management: Manage users, grants, and logical databases in a declarative manner using Kubernetes resources.
          * Secure, immutable and lightweight images based on Red Hat UBI, available for multiple architectires (amd64, arm64 and ppc64le).
          * [Operator certified ](https://catalog.redhat.com/en/software/container-stacks/detail/65789bcbe17f1b31944acb1d#overview) by Red Hat.

          ## Installation

          In order to use this product, you will need a [MariaDB Enterprise subscription](https://mariadb.com/products/enterprise/).

          To be able to pull MariaDB Enterprise images, you will need to configure the MariaDB registry credentials in Openshift using the [MariaDB customer download token](https://customers.mariadb.com/downloads/token/) as described below:

          - Extract your [Openshift global pull secret](https://docs.redhat.com/en/documentation/openshift_container_platform/4.18/html/images/managing-images#images-update-global-pull-secret_using-image-pull-secrets):
            ```bash
            oc extract secret/pull-secret -n openshift-config --confirm
            ```

          - Login in the MariaDB registry providing the customer download token as password:
            ```bash
            oc registry login \
              --registry="docker.mariadb.com" \
              --auth-basic="<email>:<customer-download-token>" \
              --to=.dockerconfigjson
            ```

          - Update the global pull secret:
            ```bash
            oc set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson
            ```

          Alternatively, you can also create a `Secret` named `mariadb-enterprise` in the namespace where are you installing the operator by using the following command:
            ```bash
            oc create secret docker-registry mariadb-enterprise \
              --docker-server=docker.mariadb.com \
              --docker-username=<email> \
              --docker-password=<customer-download-token> \
              -n <operator-namespace>
            ```

          At this point, you can proceed to install the MariaDB Enterprise Operator, either via the Openshift console or by creating a `Subscription` object as described in the [Openshift installation docs](https://mariadb.com/kb/en/openshift/).

          Refer to the [quickstart guide](https://mariadb.com/kb/en/quickstart/), the [examples catalog](https://mariadb.com/kb/en/examples-catalog/) and the [API reference](https://mariadb.com/kb/en/api-reference/) for getting started with the Custom Resources offered by this operator.
        displayName: MariaDB Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mariadb
        - mysql
        - database
        - operator
        - mariadb-enterprise-operator
        - database
        links:
        - name: MariaDB Enterprise
          url: https://mariadb.com/products/enterprise/
        - name: Documentation
          url: https://mariadb.com/kb/en/mariadb-enterprise-operator/
        maintainers:
        - email: operator@mariadb.com
          name: MariaDB Corporation
        - email: support@mariadb.com
          name: Support
        - email: sales-EMEA@mariadb.com
          name: Sales EMEA
        - email: sales-APAC@mariadb.com
          name: Sales APAC
        - email: sales-nam@mariadb.com
          name: Sales Americas
        maturity: beta
        minKubeVersion: 1.29.0
        provider:
          name: MariaDB Corporation
        relatedImages:
        - docker.mariadb.com/enterprise-server@sha256:9872721d35ebe8d820b1aa9c6dd83b363a7b2df9af11a65bff6b80ff4144460a
        - docker.mariadb.com/maxscale@sha256:d9e04a7a406353dab2051e190be0c32ed20b780ed71e56e55cedd4ccb61af416
        - mariadb/mariadb-prometheus-exporter-ubi@sha256:f072175198cad050e80617d13e63f415bacdfe47e2dd177fde7283a96ee56c07
        - mariadb/maxscale-prometheus-exporter-ubi@sha256:cc17c1554795ad01b4d8db799f6aa90e00e41ede431f09d30eec64d9dd4479d9
        - docker.mariadb.com/mariadb-enterprise-operator@sha256:1effbbfa7632be92c17ef61c18a25f010b187840beec42c31d2410d0209e6985
        version: 25.8.1-1
      entries:
      - name: mariadb-enterprise-operator.v25.8.1-1
        version: 25.8.1-1
      - name: mariadb-enterprise-operator.v25.8.0
        version: 25.8.0
      - name: mariadb-enterprise-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: mariadb-enterprise-operator
    provider:
      name: MariaDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Instana
      provider-url: ""
    name: instana-agent-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: instana-agent-operator.v2.1.36
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instana.io/v1",
                "kind": "InstanaAgent",
                "metadata": {
                  "name": "instana-agent",
                  "namespace": "instana-agent"
                },
                "spec": {
                  "agent": {
                    "configuration_yaml": "# You can leave this empty, or use this to configure your instana agent.\n# See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.\n",
                    "endpointHost": "ingress-red-saas.instana.io",
                    "endpointPort": "443",
                    "env": {},
                    "key": "replace-key"
                  },
                  "cluster": {
                    "name": "my-cluster"
                  },
                  "opentelemetry": {
                    "grpc": {
                      "enabled": true
                    },
                    "http": {
                      "enabled": true
                    }
                  },
                  "zone": {
                    "name": "edited-zone"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring,Logging & Tracing,OpenShift Optional
          certified: "false"
          containerImage: icr.io/instana/instana-agent-operator@sha256:ab8b35f0b45c68f69210d1b17c5dd609799f6bb0bc9f12d0d4e5302c910607a1
          description: Fully automated Application Performance Monitoring (APM) for
            microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/instana-agent-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/instana-agent-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/instana/instana-agent-operator
          support: Instana
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InstanaAgent is the Schema for the agents API
            displayName: Instana Agent
            kind: InstanaAgent
            name: agents.instana.io
            version: v1
          - description: InstanaAgentRemote is the Schema for the agents API
            displayName: Instana Agent Remote
            kind: InstanaAgentRemote
            name: agentsremote.instana.io
            version: v1
        description: |
          # Instana

          Instana is an [APM solution](https://www.instana.com/product-overview/) built for microservices that enables IT Ops to build applications faster and deliver higher quality services by automating monitoring, tracing and root cause analysis. The solution is optimized for [Kubernetes](https://www.instana.com/automatic-kubernetes-monitoring/) and [OpenShift](https://www.instana.com/blog/automatic-root-cause-analysis-for-openshift-applications/).

          ## Instana Agent Operator

          This is the Kubernetes Operator for installing the Instana Agent on Kubernetes or OpenShift.

          ## Prerequisites for OpenShift

          Before the agent will be able to run in OpenShift, you need to perform a couple of extra configuration steps.

          You need to set up a project for the Instana Agent and configure it's permissions.

          The project you create here needs to be the namespace where you create the Instana Agent custom resource that the operator will use to deploy the agent.

          For example, create the `instana-agent` project:

              oc new-project instana-agent

          Then, ensure the `instana-agent` service account is in the privileged security context:

              oc adm policy add-scc-to-user privileged -z instana-agent

          The `instana-agent-remote` service account must be allowed to use any userid:

              oc adm policy add-scc-to-user anyuid -z instana-agent-remote -n instana-agent

          This service account will be created by the operator.

          Now you can proceed with installing the operator for the Instana agent.

          ## Installation and Configuration

          First, install this operator from [OperatorHub.io](https://operatorhub.io/), [OpenShift Container Platform](https://www.openshift.com/), or [OKD](https://www.okd.io/).

          Second, create a custom resource with the agent configuration in the target namespace (for now, this MUST always be the `instana-agent` namespace). The operator will pick up the custom resource and install the Instana agent accordingly.

          The following is a minimal template of the custom resource:

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgent
          metadata:
            name: instana-agent
            namespace: instana-agent
          spec:
            zone:
              name: my-zone # (optional) name of the zone of the host
            cluster:
              name: replace-me # replace with the name of your Kubernetes cluster
            agent:
              key: replace-me # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io # the monitoring ingress endpoint
              endpointPort: "443" # the monitoring ingress endpoint port, wrapped in quotes
              env:
                INSTANA_AGENT_TAGS: example
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://ibm.biz/monitoring-k8s
          ```

          Save the template in a file `instana-agent.yaml` and edit the following values:

          * If your target namespace is not `instana-agent`, replace the `namespace:` accordingly.
          * `agent.key` must be set with your Instana agent key.
          * `agent.endpointHost` must be set with the monitoring ingress endpoint, generally either `saas-us-west-2.instana.io` or `saas-eu-west-1.instana.io`.
          * `agent.endpointPort` must be set with the monitoring ingress port, generally "443" (wrapped in quotes).
          * `zone.name` should be set with the name of the Kubernetes cluster that is be displayed in Instana.

          For advanced configuration, you can edit the contents of the `configuration.yaml` file. View documentation [here](https://ibm.biz/monitoring-k8s).

          Apply the custom resource with `kubectl apply -f instana-agent.yaml`. After some time, you should see `instana-agent` Pods being created on each node of your cluster, and your cluster should show on the infrastructure map on your Instana Web interface.

          In addition to the InstanaAgent custom resource, a remote agent can be deployed. This agent will not monitor workload running inside the cluster, but can deploy a single replica of an agent which can get an independent configuration for running remote monitoring for certain sensors.

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgentRemote
          metadata:
            name: remote-agent
            namespace: "instana-agent"
          spec:
            zone:
              name: edited-zone # name of the zone of the host
            hostname:
              name: use-custom-hostname-for-remote-agent # (optional) give your remote agent pod a custom hostname
            agent:
              key: replace-key # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io
              endpointPort: "443"
              env: {}
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.
          ```

          ## Uninstalling

          In order to uninstall the Instana agent, simply remove the custom resource with `kubectl delete -f instana-agent.yaml`.

          ## Source Code

          The Instana agent operator is an open source project hosted on [https://github.com/instana/instana-agent-operator](https://github.com/instana/instana-agent-operator/).
        displayName: Instana Agent Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - APM
        - Instana
        links:
        - name: GitHub Repository
          url: https://github.com/instana/instana-agent-operator
        maintainers:
        - email: support@instana.com
          name: Instana
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: Instana
        relatedImages:
        - icr.io/instana/instana-agent-operator@sha256:ab8b35f0b45c68f69210d1b17c5dd609799f6bb0bc9f12d0d4e5302c910607a1
        - icr.io/instana/agent@sha256:ed71e6c3e5bdb155853a2db8c916460e6a867fb02627e3a332186188cbbd98dd
        version: 2.1.36
      entries:
      - name: instana-agent-operator.v2.1.36
        version: 2.1.36
      - name: instana-agent-operator.v2.1.35
        version: 2.1.35
      - name: instana-agent-operator.v2.1.34
        version: 2.1.34
      - name: instana-agent-operator.v2.1.33
        version: 2.1.33
      - name: instana-agent-operator.v2.1.32
        version: 2.1.32
      - name: instana-agent-operator.v2.1.31
        version: 2.1.31
      - name: instana-agent-operator.v2.1.30
        version: 2.1.30
      - name: instana-agent-operator.v2.1.28
        version: 2.1.28
      - name: instana-agent-operator.v2.1.27
        version: 2.1.27
      - name: instana-agent-operator.v2.1.24
        version: 2.1.24
      - name: instana-agent-operator.v2.1.23
        version: 2.1.23
      - name: instana-agent-operator.v2.1.22
        version: 2.1.22
      - name: instana-agent-operator.v2.1.21
        version: 2.1.21
      - name: instana-agent-operator.v2.1.20
        version: 2.1.20
      - name: instana-agent-operator.v2.1.19
        version: 2.1.19
      - name: instana-agent-operator.v2.1.18
        version: 2.1.18
      - name: instana-agent-operator.v2.1.17
        version: 2.1.17
      - name: instana-agent-operator.v2.1.16
        version: 2.1.16
      - name: instana-agent-operator.v2.1.15
        version: 2.1.15
      - name: instana-agent-operator.v2.1.14
        version: 2.1.14
      - name: instana-agent-operator.v2.1.13
        version: 2.1.13
      - name: instana-agent-operator.v2.1.12
        version: 2.1.12
      - name: instana-agent-operator.v2.1.11
        version: 2.1.11
      - name: instana-agent-operator.v2.1.10
        version: 2.1.10
      - name: instana-agent-operator.v2.1.9
        version: 2.1.9
      - name: instana-agent-operator.v2.1.8
        version: 2.1.8
      - name: instana-agent-operator.v2.1.7
        version: 2.1.7
      - name: instana-agent-operator.v2.1.6
        version: 2.1.6
      - name: instana-agent-operator.v2.1.5
        version: 2.1.5
      - name: instana-agent-operator.v2.1.4
        version: 2.1.4
      - name: instana-agent-operator.v2.1.3
        version: 2.1.3
      - name: instana-agent-operator.v2.1.2
        version: 2.1.2
      - name: instana-agent-operator.v2.1.1
        version: 2.1.1
      - name: instana-agent-operator.v2.1.0
        version: 2.1.0
      - name: instana-agent-operator.v2.0.32
        version: 2.0.32
      - name: instana-agent-operator.v2.0.31
        version: 2.0.31
      - name: instana-agent-operator-rhmp.v2.0.30
        version: 2.0.30
      - name: instana-agent-operator.v2.0.27
        version: 2.0.27
      - name: instana-agent-operator.v2.0.19
        version: 2.0.19
      - name: instana-agent-operator.v2.0.18
        version: 2.0.18
      - name: instana-agent-operator.v2.0.17
        version: 2.0.17
      - name: instana-agent-operator.v2.0.13
        version: 2.0.13
      - name: instana-agent-operator.v2.0.12
        version: 2.0.12
      - name: instana-agent-operator.v2.0.11
        version: 2.0.11
      - name: instana-agent-operator.v2.0.10
        version: 2.0.10
      - name: instana-agent-operator.v2.0.9
        version: 2.0.9
      - name: instana-agent-operator.v2.0.8
        version: 2.0.8
      - name: instana-agent-operator.v2.0.7
        version: 2.0.7
      - name: instana-agent-operator.v2.0.5
        version: 2.0.5
      - name: instana-agent-operator.v2.0.4
        version: 2.0.4
      name: stable
    defaultChannel: stable
    packageName: instana-agent-operator-rhmp
    provider:
      name: Instana
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: pelorus-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pelorus-operator.v0.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.pelorus.dora-metrics.io/v1alpha1",
                "kind": "Pelorus",
                "metadata": {
                  "name": "pelorus-sample"
                },
                "spec": {
                  "exporters": {
                    "global": {},
                    "instances": [
                      {
                        "app_name": "deploytime-exporter",
                        "exporter_type": "deploytime"
                      },
                      {
                        "app_name": "failuretime-exporter",
                        "enabled": false,
                        "exporter_type": "failure"
                      },
                      {
                        "app_name": "committime-exporter",
                        "exporter_type": "committime"
                      },
                      {
                        "app_name": "webhook-exporter",
                        "exporter_type": "webhook"
                      }
                    ]
                  },
                  "federate_openshift_monitoring": {
                    "enabled": false
                  },
                  "openshift_prometheus_basic_auth_pass": "changeme",
                  "openshift_prometheus_htpasswd_auth": "internal:{SHA}+pvrmeQCmtWmYVOZ57uuITVghrM=",
                  "prometheus_retention": "1y",
                  "prometheus_retention_size": "1GB",
                  "prometheus_storage": false,
                  "prometheus_storage_pvc_capacity": "2Gi",
                  "prometheus_storage_pvc_storageclass": "gp2"
                }
              }
            ]
          capabilities: Basic Install
          categories: |
            Modernization & Migration,Developer Tools,Monitoring,Integration & Delivery
          containerImage: quay.io/pelorus/pelorus-operator:0.0.9
          createdAt: "2024-05-27T20:04:01Z"
          description: |
            Tool that helps IT organizations measure their impact on the overall performance of their organization
          operatorframework.io/suggested-namespace: pelorus
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/dora-metrics/pelorus/
          support: Pelorus Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pelorus is the Schema for pelorus API to adopt instance to
              the requested workflow
            displayName: Pelorus
            kind: Pelorus
            name: peloruses.charts.pelorus.dora-metrics.io
            version: v1alpha1
        description: |
          Pelorus is a tool that helps IT organizations measure their impact on the overall performance of their organization. It does this by gathering metrics about team and organizational behaviors over time in some key areas of IT that have been shown to impact the value they deliver to the organization as a whole. Some of the key outcomes Pelorus can focus on are:

          - Software Delivery Performance
          - Product Quality and Sustainability
          - Customer experience

          For more background on the project you can read [@trevorquinn](https://github.com/trevorquinn)'s blog post on [Metrics Driven Transformation](https://www.openshift.com/blog/exploring-a-metrics-driven-approach-to-transformation).

          ## Software Delivery Performance as an outcome

          Currently, Pelorus functionality can capture proven metrics that measure Software Delivery Performance -- a significant outcome that IT organizations aim to deliver.

          Pelorus is a Grafana dashboard that can easily be deployed to an OpenShift cluster, and provides an organizational-level view of the [four critical measures of software delivery performance](https://blog.openshift.com/exploring-a-metrics-driven-approach-to-transformation/).

          ## Software Delivery Metrics Dashboard

          A short video describing each of metrics provided by Pelorus is available [here](https://www.youtube.com/watch?v=7-iB_KhUaQg).

          ## Demo

          [YouTube Video](https://www.youtube.com/watch?v=VPCOIfDcgso) with the Pelorus in action recorded during online Konveyor Community event.

          ## Documentation

          Pelorus documentation is available at [pelorus.readthedocs.io](https://pelorus.readthedocs.io/en/latest/).

          ## Contributing to Pelorus

          If you are interested in contributing to the Pelorus project, please review our Contribution guide which can be found in the [contribution guide](https://github.com/dora-metrics/pelorus/blob/master/CONTRIBUTING.md)

          ## Statement of Support

          Our support policy can be found in the [Upstream Support statement](https://github.com/dora-metrics/pelorus/blob/master/docs/UpstreamSupport.md)

          ## Code of Conduct
          Refer to dora-metrics's Code of Conduct [here](https://github.com/dora-metrics/pelorus/blob/master/CODE_OF_CONDUCT.md).

          ## License

          This repository is licensed under the terms of [Apache-2.0 License](https://raw.githubusercontent.com/dora-metrics/pelorus/master/LICENSE).
        displayName: Pelorus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pelorus-operator
        - dora
        - dora metrics
        - pelorus
        - metrics
        - transformation
        - devops
        links:
        - name: Pelorus Docs
          url: https://pelorus.readthedocs.io/en/latest/
        - name: Pelorus Getting Started
          url: https://pelorus.readthedocs.io/en/latest/GettingStarted/Overview/
        - name: Pelorus GIT repository
          url: https://github.com/dora-metrics/pelorus/
        maintainers:
        - email: kgranger@redhat.com
          name: Kevin Granger
        - email: whayutin@redhat.com
          name: Wesley Hayutin
        - email: mpryc@redhat.com
          name: Michal Pryc
        - email: msouzaol@redhat.com
          name: Mateus Souza Oliveira
        - email: esauer@redhat.com
          name: Eric Sauer
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - quay.io/pelorus/pelorus-operator:0.0.9
        version: 0.0.9
      entries:
      - name: pelorus-operator.v0.0.9
        version: 0.0.9
      - name: pelorus-operator.v0.0.8
        version: 0.0.8
      - name: pelorus-operator.v0.0.7
        version: 0.0.7
      - name: pelorus-operator.v0.0.6
        version: 0.0.6
      - name: pelorus-operator.v0.0.5
        version: 0.0.5
      - name: pelorus-operator.v0.0.4
        version: 0.0.4
      - name: pelorus-operator.v0.0.3
        version: 0.0.3
      - name: pelorus-operator.v0.0.2
        version: 0.0.2
      - name: pelorus-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: pelorus-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CodeFlare
      provider-url: https://github.com/project-codeflare
    name: codeflare-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: codeflare-operator.v1.16.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: quay.io/project-codeflare/codeflare-operator:v1.16.0
          createdAt: "2025-06-16T10:53:51Z"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/project-codeflare/codeflare-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AppWrapper
            name: appwrappers.workload.codeflare.dev
            version: v1beta2
        description: CodeFlare allows you to scale complex pipelines anywhere
        displayName: CodeFlare Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Pipelines
        - Scaling
        - App
        - Jobs
        links:
        - name: Codeflare Operator
          url: https://codeflare.dev/
        maintainers:
        - email: aasthana@redhat.com
          name: Anish Asthana
        - email: kpostlet@redhat.com
          name: Kevin Postlethwait
        - email: meyceoz@redhat.com
          name: Mustafa Eyceoz
        maturity: alpha
        provider:
          name: CodeFlare
          url: https://github.com/project-codeflare
        relatedImages:
        - quay.io/project-codeflare/codeflare-operator:v1.16.0
        version: 1.16.0
      entries:
      - name: codeflare-operator.v1.16.0
        version: 1.16.0
      - name: codeflare-operator.v1.15.0
        version: 1.15.0
      - name: codeflare-operator.v1.14.0
        version: 1.14.0
      - name: codeflare-operator.v1.13.0
        version: 1.13.0
      - name: codeflare-operator.v1.12.0
        version: 1.12.0
      - name: codeflare-operator.v1.11.0
        version: 1.11.0
      - name: codeflare-operator.v1.10.0
        version: 1.10.0
      - name: codeflare-operator.v1.9.0
        version: 1.9.0
      - name: codeflare-operator.v1.8.0
        version: 1.8.0
      - name: codeflare-operator.v1.7.0
        version: 1.7.0
      - name: codeflare-operator.v1.6.0
        version: 1.6.0
      - name: codeflare-operator.v1.5.0
        version: 1.5.0
      - name: codeflare-operator.v1.4.4
        version: 1.4.4
      - name: codeflare-operator.v1.4.3
        version: 1.4.3
      - name: codeflare-operator.v1.4.2
        version: 1.4.2
      - name: codeflare-operator.v1.4.1
        version: 1.4.1
      - name: codeflare-operator.v1.4.0
        version: 1.4.0
      - name: codeflare-operator.v1.3.1
        version: 1.3.1
      - name: codeflare-operator.v1.3.0
        version: 1.3.0
      - name: codeflare-operator.v1.2.0
        version: 1.2.0
      - name: codeflare-operator.v1.1.0
        version: 1.1.0
      - name: codeflare-operator.v1.0.1
        version: 1.0.1
      - name: codeflare-operator.v1.0.0
        version: 1.0.0
      - name: codeflare-operator.v1.0.0-rc.4
        version: 1.0.0-rc.4
      - name: codeflare-operator.v1.0.0-rc.3
        version: 1.0.0-rc.3
      - name: codeflare-operator.v1.0.0-rc.2
        version: 1.0.0-rc.2
      - name: codeflare-operator.v1.0.0-rc.1
        version: 1.0.0-rc.1
      - name: codeflare-operator.v0.2.3
        version: 0.2.3
      - name: codeflare-operator.v0.2.2
        version: 0.2.2
      - name: codeflare-operator.v0.2.1
        version: 0.2.1
      - name: codeflare-operator.v0.2.0
        version: 0.2.0
      - name: codeflare-operator.v0.1.0
        version: 0.1.0
      - name: codeflare-operator.v0.0.6
        version: 0.0.6
      - name: codeflare-operator.v0.0.5
        version: 0.0.5
      - name: codeflare-operator.v0.0.4
        version: 0.0.4
      - name: codeflare-operator.v0.0.3
        version: 0.0.3
      - name: codeflare-operator.v0.0.2
        version: 0.0.2
      - name: codeflare-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: codeflare-operator
    provider:
      name: CodeFlare
      url: https://github.com/project-codeflare
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: opdev.github.io
    name: airflow-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airflow-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "workflow.apache.org/v1alpha1",
                "kind": "AirFlow",
                "metadata": {
                  "name": "airflow-helm"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/airflow-helm-operator:0.0.1
          createdAt: "2023-06-28T16:57:21Z"
          description: An experimental operator that installs Apache Airflow.
          operatorframework.io/suggested-namespace: airflow-helm
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys Apache Airflow based on the official helm chart in
              openshift
            kind: AirFlow
            name: airflows.workflow.apache.org
            version: v1alpha1
        description: Deploys Apache Airflow based on the official helm chart with
          openshift specific modifications applied by default
        displayName: Airflow Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        - data
        - batch
        - scheduler
        - data workflow
        links:
        - name: Airflow Helm Operator
          url: https://airflow-helm-operator.domain
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: opdev.github.io
        relatedImages:
        - quay.io/opdev/airflow-helm-operator:0.0.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 0.0.1
      entries:
      - name: airflow-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: airflow-helm-operator
    provider:
      name: opdev
      url: opdev.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-sso
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-sso.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Sso",
                "metadata": {
                  "name": "sso-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "kcos-keycloak": {
                    "auth": {
                      "baseURL": "https://keycloak.kcos-sso.svc.cluster.local:8443",
                      "enabled": true,
                      "sessionCookie": "AT"
                    },
                    "authproxy": {
                      "image": "prod.harbor.keysight.digital/loadcore/kcos-auth-proxy@sha256:a777ee5116280fbeb1e1342445fef5135f9ca61e983a3e2821f9ca4b2038b044",
                      "initContainer": {
                        "image": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf"
                      }
                    },
                    "crsCleanupJob": {
                      "clientExec": "oc",
                      "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "realm": {
                      "accessTokenLifespan": 266400,
                      "accessTokenLifespanForImplicitFlow": 266400,
                      "admin": {
                        "password": "admin"
                      },
                      "offlineSessionIdleTimeout": 2592000,
                      "offlineSessionMaxLifespan": 5184000,
                      "registrationAllowed": true,
                      "ssoSessionIdleTimeout": 266400,
                      "ssoSessionMaxLifespan": 266400
                    }
                  },
                  "subcharts": {
                    "enable": {
                      "kcos-keycloak": true,
                      "keycloak-operator": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:29:57Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Sso
            name: ssoes.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS SSO Operator
        displayName: Keysight KCOS SSO Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Sso
          url: https://keysight-kcos-sso.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-sso@sha256:298417fac2fc053bdc352f77b59b14e39eadd66c871cfb547304a517489f294e
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - prod.harbor.keysight.digital/loadcore/kcos-auth-proxy@sha256:a777ee5116280fbeb1e1342445fef5135f9ca61e983a3e2821f9ca4b2038b044
        version: 0.1.21
      entries:
      - name: keysight-kcos-sso.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-sso
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-cpd-wspipelines
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-wspipelines.v10.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wspipelines.cpd.ibm.com/v1beta1",
                "kind": "WSPipelines",
                "metadata": {
                  "name": "wspipelines-cr",
                  "labels": {
                    "app.kubernetes.io/name": "wspipelines",
                    "app.kubernetes.io/instance": "wspipelines.wspipelines.cpd.ibm.com",
                    "app.kubernetes.io/managed-by": "wspipelines-operator",
                    "wspipelinesBuildNumber": "2017"
                }
                },
                "spec": {
                  "version": "5.1.3",
                  "scaleConfig": "small",
                  "license": {
                    "accept": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:f9f616d17f816468957b69bbb09e6c92d678acb4ba7a40f1cb493d08fb98deb8
          createdAt: "2025-02-11"
          description: WS Pipelines CSV file
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-wspipelines
          olm.skipRange: '>=1.0.0 <10.3.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM watsonx.ai"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation For additional details regarding install parameters
              check: https://ibm.biz/BdaaL9 or https://ibm.biz/BdaaLC. License By
              installing this product you accept the license terms https://ibm.biz/BdaaL9
              or https://ibm.biz/BdaaLC.'
            kind: WSPipelines
            name: wspipelines.wspipelines.cpd.ibm.com
            version: v1beta1
        description: |-
          # Name
          IBM&reg; Watson Pipelines
          # Introduction
          ## Summary
          The Watson Pipelines editor provides a graphical interface for orchestrating an end-to-end flow of assets from creation through deployment. It enables to automate an end-to-end flow to prepare data, then create, train, deploy, monitor and update machine learning models and Python scripts. To design a pipeline you drag nodes onto the canvas, specify objects and parameters, then run and monitor the pipeline.
          ## Features
          Watson Pipelines are built off of Kubeflow pipelines on Tekton runtime and are fully integrated into the Watson Studio platform, allowing users to combine tools including:
          - Notebooks
          - Data refinery flows
          - AutoAI experiments
          - Web service / online deployments
          - Batch deployments
          - Import and export of Project / Space assets
          into repeatable and automated pipelines.
          # Details
          ## Prerequisites
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-pre-installation-tasks)
          ### Resources Required
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=planning-system-requirements)
          # Installing
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing)
          ## Configuration
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing)
          ## Storage
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=planning-storage-considerations)
          ## Limitations
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=issues-watson-studio)
          ## Documentation
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=overview)
          ## PodSecurityPolicy Requirements
          Custom PodSecurityPolicy definition:
          ```
          Not Applicable
          ```
          ## SecurityContextConstraints Requirements
          Custom SecurityContextConstraints definition:
          ```
          apiVersion: security.openshift.io/v1
          kind: SecurityContextConstraints
          allowHostDirVolumePlugin: false
          allowHostIPC: false
          allowHostNetwork: false
          allowHostPID: false
          allowHostPorts: false
          allowPrivilegeEscalation: false
          allowPrivilegedContainer: false
          allowedCapabilities: null
          defaultAddCapabilities: null
          fsGroup:
            type: MustRunAs
          metadata:
            annotations:
              include.release.openshift.io/self-managed-high-availability: "true"
              kubernetes.io/description: restrict denies access to all host features and requires pods to be run with a UID, and SELinux context that are allocated to the namespace.  This is the most restrictive SCC and it is used by default for authenticated users.
              release.openshift.io/create-only: "true"
            name: restrict
          priority: null
          readOnlyRootFilesystem: false
          requiredDropCapabilities:
          - ALL
          runAsUser:
            type: MustRunAsRange
          seLinuxContext:
            type: MustRunAs
          supplementalGroups:
            type: RunAsAny
          users: []
          volumes:
          - configMap
          - downwardAPI
          - emptyDir
          - persistentVolumeClaim
          - projected
          - secret
          ```
        displayName: Watson Pipelines
        installModes:
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: OwnNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - wspipelines
        - WSPipelines
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: v1
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:a477d47eaf8f8e569601d3c23c1def66f31c098ac3f5a1ae350618c460917dd8
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:0164d8fbea591bbd9d67d84603162c3781b08551eddcc2a8a842aafc828af18c
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:164f6063f3a50a36e28a240ae7c060c4f90266b7ed799bfdc075f90628bea7ff
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:ebcfc8941330da6991480dc18ac34d6df4fe879e562ee537e5c36df73f4bc536
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a24f4cb635e8b5759a5e2bda20ddefc6d5ea9b16fb1081592885c54aefb5480c
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:29cd5c9cd60bbcfe211337786616622b911e2319db9870eb9d38d4f630c03e75
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:f0c68715f78bd20c352c97dd48ac5a912cad58e19a00159b5e6878f6beeb00fb
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:a2949c3868c2c91bf35e50c64b51fcc24372986744b55520e21e9ce31fde59f0
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-bundle@sha256:842f2ac568db4db79cf0450b413b4b09b5f2180ef3541385b46fb650ac7066c0
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:f9f616d17f816468957b69bbb09e6c92d678acb4ba7a40f1cb493d08fb98deb8
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:61155acdb8982648ae5748be9f5944ba913db1590ec3e773206d6fbb1d8952cb
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:8d7625a0ba656425f601b6ec9a398a6d30511fe1de119cf0e29f680c704ea225
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:e4888fadfadb3462e3a9ad0395a526a2631872f70caf291dc18b60269ded199f
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:48b7cfa31b9a185176c20afc119e37346a6f9380d74000edf6c729b4677ca44e
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:19bb32e0389361e8314e627bcad2e9bd93984289485cf632466834353d9c3449
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:e29885c1ee083a73b952b0de8d1cd675b2b2830abf44d69a8577f56972a601fb
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:00f8bcee4ba7e0c5cf5f0c52e9e7707df367041c9f4dd5508dbefe0dbdba8877
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:99d0b3192e6e46167423505f48655e7d49af1a461ae00b95c9c7c9d4209e5b65
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:25a4bb9cea1c3bdbeeeb205622512bcba98163f9629eba35625756346ce9d652
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:6fc976bc622c2165dd8bf93c9244e1c29b7c973a4da7a5ce4d0a097a1c1ea4fb
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:ad2215adfb8dd2dc600bf5097b7f88c257d5fbaf541409c01b36cdb67c130236
        version: 10.3.0
      entries:
      - name: ibm-cpd-wspipelines.v10.3.0
        version: 10.3.0
      name: v10.3
    defaultChannel: v10.3
    packageName: ibm-cpd-wspipelines
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-enterprise
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-enterprise.v1.33.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "my-replica-set"
                },
                "spec": {
                  "credentials": "my-credentials",
                  "members": 3,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "sample-sharded-cluster"
                },
                "spec": {
                  "configServerCount": 3,
                  "credentials": "my-credentials",
                  "mongodsPerShardCount": 3,
                  "mongosCount": 2,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "shardCount": 1,
                  "type": "ShardedCluster",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBMultiCluster",
                "metadata": {
                  "name": "multi-replica-set"
                },
                "spec": {
                  "clusterSpecList": [
                    {
                      "clusterName": "e2e.cluster1.mongokubernetes.com",
                      "members": 2
                    },
                    {
                      "clusterName": "e2e.cluster2.mongokubernetes.com",
                      "members": 1
                    },
                    {
                      "clusterName": "e2e.cluster3.mongokubernetes.com",
                      "members": 2
                    }
                  ],
                  "credentials": "my-credentials",
                  "duplicateServiceObjects": false,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBOpsManager",
                "metadata": {
                  "name": "ops-manager"
                },
                "spec": {
                  "adminCredentials": "ops-manager-admin",
                  "applicationDatabase": {
                    "members": 3,
                    "podSpec": {
                      "cpu": 1
                    }
                  },
                  "configuration": {
                    "mms.fromEmailAddr": "admin@thecompany.com"
                  },
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "version": "6.0.3"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBUser",
                "metadata": {
                  "name": "my-replica-set-x509-user"
                },
                "spec": {
                  "db": "$external",
                  "mongodbResourceRef": {
                    "name": "my-replica-set"
                  },
                  "roles": [
                    {
                      "db": "admin",
                      "name": "dbOwner"
                    }
                  ],
                  "username": "CN=my-replica-set-x509-user,OU=cloud,O=MongoDB,L=New York,ST=New York,C=US"
                },
                "status": {
                  "phase": "Updated"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/mongodb/mongodb-enterprise-operator-ubi@sha256:6d394a5247cc346c4421bd9b37948c0378970dd3f4fc5d6eaac681f4c9655957
          createdAt: "2025-04-15T11:32:47Z"
          description: The MongoDB Enterprise Kubernetes Operator enables easy deploys
            of MongoDB into Kubernetes clusters, using our management, monitoring
            and backup platforms, Ops Manager and Cloud Manager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mongodb/mongodb-enterprise-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MongoDB Deployment
            displayName: MongoDB Deployment
            kind: MongoDB
            name: mongodb.mongodb.com
            version: v1
          - description: MongoDB Multi Deployment
            displayName: MongoDB Multi Deployment
            kind: MongoDBMultiCluster
            name: mongodbmulticluster.mongodb.com
            version: v1
          - description: MongoDB x509 User
            displayName: MongoDB User
            kind: MongoDBUser
            name: mongodbusers.mongodb.com
            version: v1
          - description: MongoDB Ops Manager
            displayName: MongoDB Ops Manager
            kind: MongoDBOpsManager
            name: opsmanagers.mongodb.com
            version: v1
        description: |
          The MongoDB Enterprise Kubernetes Operator enables easy deploys of MongoDB
          into Kubernetes clusters, using our management, monitoring and backup
          platforms, Ops Manager and Cloud Manager.

          ## Before You Start

          To start using the operator you''ll need an account in MongoDB Cloud Manager or
          a MongoDB Ops Manager deployment.

          * [Create a Secret with your OpsManager API key](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-operator-credentials/#procedure)

          * [Create a ConfigMap with your OpsManager project ID and URL](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-project-using-configmap/)

          By installing this integration, you will be able to deploy MongoDB instances
          with a single simple command.

          ## Required Parameters

          * `opsManager` or `cloudManager` - Enter the name of the ConfigMap containing project information
          * `credentials` - Enter the name of the Secret containing your OpsManager credentials
          * `type` - Enter MongoDB Deployment Types ("Standalone", "ReplicaSet", "ShardedCluster"

          ## Supported MongoDB Deployment Types ##

          * Standalone: An instance of mongod that is running as a single server and
          not as part of a replica set, this is, it does not do any kind of
          replication.

          * Replica Set: A replica set in MongoDB is a group of mongod processes that
          maintain the same data set. Replica sets provide redundancy and high
          availability, and are the basis for all production deployments. This section
          introduces replication in MongoDB as well as the components and architecture
          of replica sets. The section also provides tutorials for common tasks
          related to replica sets.

          * Sharded Cluster: The set of nodes comprising a sharded MongoDB deployment.
          A sharded cluster consists of config servers, shards, and one or more mongos
          routing processes. Sharding is a A database architecture that partitions
          data by key ranges and distributes the data among two or more database
          instances. Sharding enables horizontal scaling.

          ## Requirements for deploying MongoDB OpsManager

          * In order to deploy resources of type MongoDB OpsManager, you will need to
          create a secret containing the [credentials](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/plan-om-resource/#om-rsrc-prereqs)
          for the Global Owner user

          ## Security ##

          The operator can enable TLS for all traffic between servers and also between
          clients and servers. Before enabling `security.tls.enabled` to `true` you
          should create your certificates.  or you can leave the operator to create all
          the certificates for you. The operator ability to create certs is been
          deprecated due to Kubernetes API changes.

          For more information, please read the official MongoDB
          Kubernetes Operator  [docs](https://docs.mongodb.com/kubernetes-operator/stable/).
        displayName: MongoDB Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mongodb
        - databases
        - nosql
        links:
        - name: Documentation
          url: https://docs.opsmanager.mongodb.com/current/tutorial/install-k8s-operator/index.html
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: stable
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-enterprise-server@sha256:fb8397cbcdbf7c33f6ad142492792771127f0d9a8c46388f350b6927405ee590
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa8683500e8d3632c4588bbcdddb4e3186a01f48aefe1b6d95f8860dae8b6c76
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:7d93aee462c60cacc4e5e66729f5131c5c0fa434e6e6400c8e2ccac4bd5e20ba
        - quay.io/mongodb/mongodb-agent-ubi@sha256:85f883076bc8edc516800704e2290c54b1bee14a9a533df84928c7ce1ae807d2
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0b2d01d543d38feddcc9d351e711df33a0a5187b22d95cdb926f6015ee1093bd
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ca2aaf57d70a0b9d743a6533c4e1b88324056197453de540d3352e679a3c9c92
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7e5361c2a1604dec3fc671d629806fe6ddc6df524b69a441b86933a4b30093
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b2bd8eb7f8b96adb874468ff1b0ebae2db4ebe490b6e1da4c9a6c5108f813ff5
        - quay.io/mongodb/mongodb-agent-ubi@sha256:766e6cbeec27f995211aa509ff7f118b6945f3e728141f28f9c4ea0dd2297af5
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3add8b68bfbaa42596fe5dbbee4623ba5ff1a826b18bb71301525adf574dcb94
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0fde236fe26d5b20210d6f246aadacd7845874712def516cbe2eec11d79d5181
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:dfec259209995bb040a1eff1fcafb1d6c3fdd145129f12c04a56ec74834b80a6
        - quay.io/mongodb/mongodb-enterprise-server@sha256:110a560f5a0ebe1fde1b2c20ae5ff5638682229d1c8656b993bdaf7b71eb94bc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:19925e1297ab652db59c35f53e7b7c8b2b277f8ecdc4215682961aaf2b3e924e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa300abf739e26350a258f643fa8e57aedacd041055c7b923bd22b21ef554cdb
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:83af86f8098e3c81b6e0879eabd582592ef6cbb1d6592f5e175c0794b9df2a0e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5421e00c0935b7a9ced51a11e935773ef5df4d259f3e1b1930fbf2f1b9d46586
        - quay.io/mongodb/mongodb-agent-ubi@sha256:38175af108178d89e4d3359ee2754534e5e9b24e843f12711fc14fe1dead516a
        - quay.io/mongodb/mongodb-agent-ubi@sha256:6ebe232ed0ea87045792e578ecb11dd46411333ad0ea61383c247704b4a0efe9
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:d3248506b6bd6b01c5973ab7bdac51ee41efa0ec8dd6211be4cc0b76a53a40c1
        - quay.io/mongodb/mongodb-enterprise-init-database-ubi@sha256:ab12a330e6b40735ee516168073887a4f37b58694bd10026a451eb9972e55500
        - quay.io/mongodb/mongodb-agent-ubi@sha256:5119a1ceeae6440dbe88b89550a7ae18de8e85c9a2059f467ad66adf7ec5d75f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:11ab324698f919f84a5dab5a964d93fe63f98a48146bb3b86974cba7606d30ec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c752a7f0726d21d55f204a359b85fa29cd727a5fe571b37289ae3ea52d6206b0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8f8735d377d214f64961381b5b1ed7cd3c306e6354dbc6b45c51134a25444b93
        - quay.io/mongodb/mongodb-enterprise-server@sha256:db63f6d93e03a3e644ef6ef4846ed7d24576c688ffb584226c78e1d1fba38878
        - quay.io/mongodb/mongodb-agent-ubi@sha256:575a43be4c667f3249f46f1fdcc06dd2e68cacc9eed989da0f6edd7e214d51b7
        - quay.io/mongodb/mongodb-agent-ubi@sha256:23d962d038888c6aba5540c8f5d81d7e86c4ccfaa07b9e159825d98676609f8e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0ee80032abe873b330537d4b64511ee76c5efb512af95287f28ddeffab13bb6d
        - quay.io/mongodb/mongodb-agent-ubi@sha256:8c11348eac180f5faa113db2804817302ec78847472e1891b063744adba4aa60
        - quay.io/mongodb/mongodb-agent-ubi@sha256:6d007271faed24bdf06eb64f2b54e5cdc3e11ce71840a7f586dabe2ee02c640e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:83db493ac6158d654d3ce55619f312d53cfbac258e0faf22e9110ca004e174bd
        - quay.io/mongodb/mongodb-agent-ubi@sha256:2395fe9d38a1a9160f7cb2f400b776ea3fc2e72187d23b6dee7fd1e4fb9ca721
        - quay.io/mongodb/mongodb-agent-ubi@sha256:66b49d9dce414dcd657c3550b26f71964b82edfb67a41a3ce8d2386b04ef07e8
        - quay.io/mongodb/mongodb-agent-ubi@sha256:d927afa9d3d569d943518f0a684e19afd4adb0875f55df7c904b9dd019589e5e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:c96ad084edbdc6a6d907f57795fa38f39199b730328dbc64997d56ea50d55dfc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:67238129df09d33ee4746ff1acb5515d699a29d9f42350eae3f3f69ccb7fbf54
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:f64d99a77d6a6b9ed10de9a28b1b8b8155414eca2e561f3fdb04bd66914fb3c5
        - quay.io/mongodb/mongodb-enterprise-server@sha256:15e31285c8949a4666f7702df53416b0905d76945a28575a6028a876cc8a6f0f
        - quay.io/mongodb/mongodb-agent-ubi@sha256:31cfefd2660da3efe1b6d63e39090464c28ae006d6ed1dffb2189a09183b9c61
        - quay.io/mongodb/mongodb-agent-ubi@sha256:97bb6bb5ecd5a802e8a91e69aa10d8bde0f648b106dc5591d49ee94209822daa
        - quay.io/mongodb/mongodb-agent-ubi@sha256:f5f3b14ee6533b21d89d9ed6d4af47dbdee9e071f909dc686dfcdf6742179e46
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:985f770667e6c4c083e0c4bb79a948b83efa0cdc4ed3012b1dbf72fe48b38986
        - quay.io/mongodb/mongodb-agent-ubi@sha256:9d1f21df008b14ca721c9ed0ff905af23e1a56b64cd4eddab4df1e8e9b63df8c
        - quay.io/mongodb/mongodb-enterprise-server@sha256:523566f44cbb6bbf0cd12b1adae939a883f2254c78dc4b8c9d0098581337b1db
        - quay.io/mongodb/mongodb-agent-ubi@sha256:2dbaaece4dd48087370e43d9bcf6b0a0312f86cb5dfe916e8725323bd5401992
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9a3a6496be0d2d446e6d654eb56386c7726e0cd2d7232f6dcec579923edf0014
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1d418c65d82ccad205b7889be4ba161d3cbfa8b1fb75b21c8fcf02de38fad9ca
        - quay.io/mongodb/mongodb-enterprise-server@sha256:172e338a43df01ee5408d8d1e3cb038d2c55e77761c1a2ac6ed37959d1525140
        - quay.io/mongodb/mongodb-enterprise-server@sha256:655896bac9e82de8e9729d59a36c3a31f66235f22b93449260a2ccb7709de084
        - quay.io/mongodb/mongodb-agent-ubi@sha256:695074fbbfcadf454230b80fbf1094260e631f6108cefb2050473edaee2a217a
        - quay.io/mongodb/mongodb-agent-ubi@sha256:b175972792aa1be268d5fb3cf3ad0156f4aecffa323e1b473360fc2a3a8d8891
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d876f57ea45582e84748b02459748e440af443a7eb0882585638973409c08f6a
        - quay.io/mongodb/mongodb-agent-ubi@sha256:411f85772e7a4b2c76e21d26d320b7743612080bf0cb4c7900d9a626cfeea33b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a48a4d5c275fd669a27469649fb39ea38c728dbcf9f0a9da4e7cf27a2ce912ed
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:2be0c6c89898b18618932b6a6303aeefdde60f48ec37d5c1fca5b8d04a00fd78
        - quay.io/mongodb/mongodb-agent-ubi@sha256:bd7edb8d0931dd952524f1f0ef99f61b0f69385d9e61efd034a4f45fd315a40b
        - quay.io/mongodb/mongodb-agent-ubi@sha256:ee150e969c0bb786969c7e0a45fb1812766b752aee539072e28648817826433b
        - quay.io/mongodb/mongodb-agent-ubi@sha256:7144ac5e5038e92a6d142bf39eba46554c1b5fafeb8ea59f0390896a0d35ed17
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:2127d7f2684ae36b123422f55237055aad02fdb71c4c21a5b8f0b0f6d8baedfc
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:c9a211a2c31b97eed3a25a814198547f2aa3470df93d661ca3700b828100d8de
        - quay.io/mongodb/mongodb-agent-ubi@sha256:c03e4a7663206f54c5ef39974502816f66d8d64936c19849f0337ea4547e614e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:35b263861d6ea66b541abc190ce1aaf4d4e36f128e5d8ba2a5e9fa8e25d84446
        - quay.io/mongodb/mongodb-enterprise-server@sha256:85c1d669c073caf9cbaac815e4c8ea2ca18b71deefa03bf84cd6d8b8ebd53e12
        - quay.io/mongodb/mongodb-enterprise-server@sha256:37af0913742b6de339c7e664b30f35dcef04506529e4a09a9bbbdd57643f7149
        - quay.io/mongodb/mongodb-agent-ubi@sha256:442f5d99d02674099525abaa69daa42cec758ce5554198f978927aa5c49006c8
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f1bae6375f846af6fca7faeeff4aedf9fab41a9d45b515ba74e9fd7336408617
        - quay.io/mongodb/mongodb-enterprise-server@sha256:605dcca06a45e5ea127edb9ff262a1c3ddafdab2f30b6d3a18f73bfb8a6acdb0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ef732e3eea91cc86caf2eabedc0c9115ed235a8d204f33e74bbf090a0b0259fb
        - quay.io/mongodb/mongodb-agent-ubi@sha256:0f8d102780a9d2ce54b7453f9575d0e24b15f014cd9b74816b1e6a4090b979b2
        - quay.io/mongodb/mongodb-enterprise-init-appdb-ubi@sha256:be4295843d92ce0f325b8daaeb3a5490c11772d0348cbd9ea35a6126d038b7fb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:33a30d11969eb21952681108092fe7f916e25c8e1de43e6a4e40e126025a2bec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7aa00dca44559318356627e88440765576dd6706187044edd1f21d0d0dcb0bf6
        - quay.io/mongodb/mongodb-enterprise-server@sha256:df6081325ea2fe01d1b27d0c70005dec8232aaf51a43d91661b722b8a1761263
        - quay.io/mongodb/mongodb-enterprise-server@sha256:dfd4277f0c16ed476ac7de51894739e8c8d13a65ea5024b9a97b3dce0f6c6975
        - quay.io/mongodb/mongodb-enterprise-server@sha256:584255582db31d0bb1f9593feb3b675d934d8750766fc98d3a1b3340feae4995
        - quay.io/mongodb/mongodb-enterprise-database-ubi@sha256:e2ca129f04ad6bf4e0dbb988badb083187725f28bac01bb7080cc80cad9ac316
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7885ab6e88214f9289d8babf5189136dde0dd4c84a15a5481ccd3ef512b2e66f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f2d9a3264a8814f33143b73c56624572fb80afeab184133d3a901713a31c1a87
        - quay.io/mongodb/mongodb-agent-ubi@sha256:03763de5018fd5fa14843444d24adfc5e387605970b531801ef2d0916233a010
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d21c65780cadc9d8668d2c28935d492625d6993f13ba562722200a819281cbcc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b00ac773623e1bf422b47e6d783263d2d08737865abe9abbe106d86d1ea25165
        - quay.io/mongodb/mongodb-enterprise-operator-ubi@sha256:6d394a5247cc346c4421bd9b37948c0378970dd3f4fc5d6eaac681f4c9655957
        - quay.io/mongodb/mongodb-agent-ubi@sha256:b23cfa639c5a4d20381245772dfe396ec915af3a3046df5d9b10016b900431f5
        - quay.io/mongodb/mongodb-enterprise-server@sha256:cd0cdf1b2a58f4c3b514e602cc1d96543410f631825045bbf65bd3fb84113f1f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8445d59d491a6d3890ea9c8b421214bcd0fdcfbc420ed40497a790b79f0bc89e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c54eeb910855594ed4bf5dcbbbcc14aa6f733e97313896a8381fb4be6aeac583
        - quay.io/mongodb/mongodb-agent-ubi@sha256:36c84115c6c645fd7d65d1004d2507ef3592aa1f07176645b1ee8e8b2a4f5f73
        - quay.io/mongodb/mongodb-enterprise-init-ops-manager-ubi@sha256:6ccc8148c266a9037ea744e8d19875146b1c231909f29a8f01c7ce415ca15936
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d49417ea85ee1ddb84429552a5aac87a82c2843904d8c0300a43b657b7a49e0e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:bd868284d9dfcb7c5a20e1160218d0172358384229b556c37e0789af0030eb24
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c07f20f8e7a5479e43cac7c74fc9f18cb4d4a1bd8b6b46f7dcc6f7ec7525fe28
        - quay.io/mongodb/mongodb-enterprise-server@sha256:13186ad45a2ba3ab55ed5c2a008eca494ee3937c8519475379476cc8e6f1d055
        - quay.io/mongodb/mongodb-agent-ubi@sha256:3b57a03dc17a301fda86c8afed1917c10d9b98947680116399c975ece944f5de
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d323c1bcd14118cc537ac282996ea5a58756d20b6b73f1eec0efa8710d246077
        - quay.io/mongodb/mongodb-agent-ubi@sha256:8467b1c2924a19929fc101ef2ff8d7e53cb0115a0cdc2e371dc0733bd0262395
        version: 1.33.0
      entries:
      - name: mongodb-enterprise.v1.33.0
        version: 1.33.0
      - name: mongodb-enterprise.v1.32.0
        version: 1.32.0
      - name: mongodb-enterprise.v1.31.0
        version: 1.31.0
      - name: mongodb-enterprise.v1.30.0
        version: 1.30.0
      - name: mongodb-enterprise.v1.29.0
        version: 1.29.0
      - name: mongodb-enterprise.v1.28.0
        version: 1.28.0
      - name: mongodb-enterprise.v1.27.0
        version: 1.27.0
      - name: mongodb-enterprise.v1.26.0
        version: 1.26.0
      - name: mongodb-enterprise.v1.25.0
        version: 1.25.0
      - name: mongodb-enterprise.v1.24.0
        version: 1.24.0
      - name: mongodb-enterprise.v1.23.0
        version: 1.23.0
      - name: mongodb-enterprise.v1.22.0
        version: 1.22.0
      - name: mongodb-enterprise.v1.21.0
        version: 1.21.0
      - name: mongodb-enterprise.v1.20.1
        version: 1.20.1
      - name: mongodb-enterprise.v1.19.1
        version: 1.19.1
      - name: mongodb-enterprise.v1.18.0
        version: 1.18.0
      - name: mongodb-enterprise.v1.17.2
        version: 1.17.2
      - name: mongodb-enterprise.v1.17.1
        version: 1.17.1
      - name: mongodb-enterprise.v1.17.0
        version: 1.17.0
      - name: mongodb-enterprise.v1.16.4
        version: 1.16.4
      - name: mongodb-enterprise.v1.16.3
        version: 1.16.3
      - name: mongodb-enterprise.v1.16.2
        version: 1.16.2
      - name: mongodb-enterprise.v1.16.1
        version: 1.16.1
      name: stable
    defaultChannel: stable
    packageName: mongodb-enterprise
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: manila-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: manila-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "manila.openstack.org/v1beta1",
                "kind": "Manila",
                "metadata": {
                  "name": "manila",
                  "namespace": "openstack"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "manila",
                  "databaseInstance": "openstack",
                  "manilaAPI": {},
                  "manilaScheduler": {},
                  "manilaShares": {
                    "share1": {}
                  },
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "osp-secret",
                  "serviceUser": "manila"
                }
              },
              {
                "apiVersion": "manila.openstack.org/v1beta1",
                "kind": "ManilaAPI",
                "metadata": {
                  "name": "manilaapi-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "manila.openstack.org/v1beta1",
                "kind": "ManilaScheduler",
                "metadata": {
                  "name": "manilascheduler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "manila.openstack.org/v1beta1",
                "kind": "ManilaShare",
                "metadata": {
                  "name": "manilashare-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:57Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManilaAPI is the Schema for the manilaapis API
            displayName: Manila API
            kind: ManilaAPI
            name: manilaapis.manila.openstack.org
            version: v1beta1
          - description: Manila is the Schema for the manilas API
            displayName: Manila
            kind: Manila
            name: manilas.manila.openstack.org
            version: v1beta1
          - description: ManilaScheduler is the Schema for the manilaschedulers API
            displayName: Manila Scheduler
            kind: ManilaScheduler
            name: manilaschedulers.manila.openstack.org
            version: v1beta1
          - description: ManilaShare is the Schema for the manilashares API
            displayName: Manila Share
            kind: ManilaShare
            name: manilashares.manila.openstack.org
            version: v1beta1
        description: Manila Operator
        displayName: Manila Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Share
        - Manila
        links:
        - name: Manila Operator
          url: https://github.com/openstack-k8s-operators/manila-operator
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/manila-rhel9-operator@sha256:1208676b94d8a726c277088591d898e08a3a8331a0125ba7bff101152afba934
        - registry.redhat.io/rhoso/openstack-manila-scheduler-rhel9@sha256:7a8f645126c073213ce066cdcac2164df51276f8aa4a3f89954664d3010e6fae
        - registry.redhat.io/rhoso/openstack-manila-share-rhel9@sha256:83a6d292785805a0ac2895810ea466cfb6e1cb7aa1a53ae8f4e22671381490df
        - registry.redhat.io/rhoso/openstack-manila-api-rhel9@sha256:4cbf725169614c2974d6d1a683cfc0179bfe06ca29cf30f7dca88caeb368c51a
        version: 1.0.6
      entries:
      - name: manila-operator.v1.0.6
        version: 1.0.6
      - name: manila-operator.v1.0.5
        version: 1.0.5
      - name: manila-operator.v1.0.4
        version: 1.0.4
      - name: manila-operator.v1.0.3
        version: 1.0.3
      - name: manila-operator.v1.0.2
        version: 1.0.2
      - name: manila-operator.v1.0.1
        version: 1.0.1
      - name: manila-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: manila-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: horizon-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: horizon-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "horizon.openstack.org/v1beta1",
                "kind": "Horizon",
                "metadata": {
                  "name": "horizon"
                },
                "spec": {
                  "customServiceConfig": "DEBUG = True\nLOGGING['handlers']['console']['level'] = 'DEBUG'\n",
                  "replicas": 1,
                  "secret": "osp-secret"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:00Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Horizon is the Schema for the horizons API
            displayName: Horizon
            kind: Horizon
            name: horizons.horizon.openstack.org
            version: v1beta1
        description: Horizon Operator
        displayName: Horizon Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - UI
        - Horizon
        links:
        - name: Horizon Operator
          url: https://github.com/openstack-k8s-operators/horizon-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-horizon-rhel9@sha256:b02c115d40aacc30e6763e295da90d69681b24fba20ebbc53133be852393b399
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/horizon-rhel9-operator@sha256:b76038b96a064c275721c1d45b95d7b0dfdb9fc626a2b3250d0adb1f97f89319
        version: 1.0.6
      entries:
      - name: horizon-operator.v1.0.6
        version: 1.0.6
      - name: horizon-operator.v1.0.5
        version: 1.0.5
      - name: horizon-operator.v1.0.4
        version: 1.0.4
      - name: horizon-operator.v1.0.3
        version: 1.0.3
      - name: horizon-operator.v1.0.2
        version: 1.0.2
      - name: horizon-operator.v1.0.1
        version: 1.0.1
      - name: horizon-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: horizon-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq-streams
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amqstreams.v2.2.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.3",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.3",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:f908d848b5e67de4b162597b190ef18d449ffd4b6540457b24b491445d5c5ea3
          createdAt: "2022-07-01 10:33:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.1.0-0 <2.2.2-2'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ", "Red Hat Application Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from AMQ Streams 1.7, migration to `v1beta2` needs to be completed for all AMQ Streams CRDs and CRs before the upgrade to 2.2 is done! If upgrading from AMQ Streams version earlier than 1.7, you need to first install the CRDs from AMQ Streams 1.7 and complete the migration to `v1beta2` for all AMQ Streams CRDs and CRs before the upgrade to 2.2 is done! For more details about the CRD upgrades, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.2/html/deploying_and_upgrading_amq_streams_on_openshift/assembly-upgrade-str#assembly-upgrade-resources-str).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.2/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: Red Hat Integration - AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        - registry.redhat.io/amq7/amq-streams-kafka-31-rhel8@sha256:d0607285fae95365b9d34b9474466824fb5ad012871204e8b13e52e438b66bf5
        - registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:f908d848b5e67de4b162597b190ef18d449ffd4b6540457b24b491445d5c5ea3
        - registry.redhat.io/amq7/amq-streams-kafka-32-rhel8@sha256:e3bbfff505f12c0360e175f4640c2dd9a7adcf3dce1ae2f586f2bde805ed8b1f
        - registry.redhat.io/amq7/amq-streams-bridge-rhel8@sha256:5d49bd23da40590e74422424b9457398bad89da8276718ff6a8a088e901c8345
        version: 2.2.2-2
      entries:
      - name: amqstreams.v2.2.2-2
        version: 2.2.2-2
      - name: amqstreams.v2.2.2-1
        version: 2.2.2-1
      - name: amqstreams.v2.2.2-0
        version: 2.2.2-0
      - name: amqstreams.v2.2.1-7
        version: 2.2.1-7
      - name: amqstreams.v2.2.1-6
        version: 2.2.1-6
      - name: amqstreams.v2.2.1-5
        version: 2.2.1-5
      - name: amqstreams.v2.2.1-4
        version: 2.2.1-4
      - name: amqstreams.v2.2.1-3
        version: 2.2.1-3
      - name: amqstreams.v2.2.1-2
        version: 2.2.1-2
      - name: amqstreams.v2.2.1-1
        version: 2.2.1-1
      - name: amqstreams.v2.2.1-0
        version: 2.2.1-0
      - name: amqstreams.v2.2.0-5
        version: 2.2.0-5
      name: amq-streams-2.2.x
    - currentCSV: amqstreams.v2.3.0-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.3.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.3.1",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:2e46a25144d697761f0119deaf528c6f1c2ccd04e2acd95c661b0df4f32e1a9f
          createdAt: "2022-10-31 17:05:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.2.0-0 <2.3.0-4'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.3.0 release, AMQ Streams supports only OCP 4.8 to 4.12!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 2.2 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more details, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.3/html/deploying_and_upgrading_amq_streams_on_openshift/assembly-upgrade-str#assembly-upgrade-resources-str

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.3/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-streams-kafka-32-rhel8@sha256:0152a85f2d0bc85881dacaee62c6e8b0b8b04f57a542b9b82dc93ad5303d6bf9
        - registry.redhat.io/amq7/amq-streams-kafka-33-rhel8@sha256:046c2539fef32bd928fc1f4f4c41e63f9a143b371fcc201d78c66f48fc34ea32
        - registry.redhat.io/amq7/amq-streams-bridge-rhel8@sha256:c822b3d87f7dea233fc523238d8d0516e42f9a3d252f50f4d427c934995f9b22
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        - registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:2e46a25144d697761f0119deaf528c6f1c2ccd04e2acd95c661b0df4f32e1a9f
        version: 2.3.0-4
      entries:
      - name: amqstreams.v2.3.0-4
        version: 2.3.0-4
      name: amq-streams-2.3.x
    - currentCSV: amqstreams.v2.4.0-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:cd1fefcd199400a504b4b9942c24ba6becafcecc952a9e39fa2729228be79e77
          createdAt: "2022-10-31 17:05:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.3.0-0 <2.4.0-3'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.4.0 release, AMQ Streams supports only OCP 4.8 to 4.12!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more details, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.4/html/deploying_and_managing_amq_streams_on_openshift/assembly-upgrade-str)

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.4/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:8d467fc746b78440df06f6d88d266f6ac9b5a4574a2b17fd870c173a107f8d65
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:cd1fefcd199400a504b4b9942c24ba6becafcecc952a9e39fa2729228be79e77
        - registry.redhat.io/amq-streams/kafka-33-rhel8@sha256:aa780a250444c2426be79e95366050a33f4a4c9d68c08b58743413ea67a2d6d7
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:50b3be789993cfeedc4923a8e760c1cf52498b2178be3abfa3719fc2d15ce723
        version: 2.4.0-3
      entries:
      - name: amqstreams.v2.4.0-3
        version: 2.4.0-3
      - name: amqstreams.v2.4.0-2
        version: 2.4.0-2
      - name: amqstreams.v2.4.0-1
        version: 2.4.0-1
      name: amq-streams-2.4.x
    - currentCSV: amqstreams.v2.5.2-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
          createdAt: "2023-07-10 21:47:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.2.0-0 <2.5.2-8'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.5.2 release, AMQ Streams supports only OCP 4.12 to 4.16!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more information, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:8c7e8ccb6c2d5da32b70d9d0c3919c303ad73a25447f6993d964f7b19cca1072
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:4ede5eeed159bddacfe830a768bab4996003d4c57a496fc8ec9b41df916d4acc
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:95bc35951eb545b918114af1e92a7139e2e3323b7d054dcb2a060843c7326d28
        version: 2.5.2-8
      entries:
      - name: amqstreams.v2.5.2-8
        version: 2.5.2-8
      - name: amqstreams.v2.5.2-7
        version: 2.5.2-7
      - name: amqstreams.v2.5.2-6
        version: 2.5.2-6
      - name: amqstreams.v2.5.2-5
        version: 2.5.2-5
      - name: amqstreams.v2.5.2-4
        version: 2.5.2-4
      - name: amqstreams.v2.5.2-3
        version: 2.5.2-3
      - name: amqstreams.v2.5.2-2
        version: 2.5.2-2
      - name: amqstreams.v2.5.2-1
        version: 2.5.2-1
      - name: amqstreams.v2.5.2-0
        version: 2.5.2-0
      - name: amqstreams.v2.5.1-5
        version: 2.5.1-5
      - name: amqstreams.v2.5.1-4
        version: 2.5.1-4
      - name: amqstreams.v2.5.1-3
        version: 2.5.1-3
      - name: amqstreams.v2.5.1-2
        version: 2.5.1-2
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.5.x
    - currentCSV: amqstreams.v2.6.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.6.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:dacf7ddffe7cbe2dc35e7d3290637b3334117e09dde986e6eaeaceded7069fbf
          createdAt: "2023-10-14 18:18:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.6.0-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: "**Red Hat AMQ Streams** is a massively scalable, distributed,
          and high performance data streaming platform based on the Apache Kafka®
          project. AMQ Streams provides an event streaming backbone that allows microservices
          and other application components to exchange data with extremely high throughput
          and low latency.\n**From the 2.6.0 release, AMQ Streams supports only OCP
          4.11 to 4.14! \n**The core capabilities include:** * A pub/sub messaging
          model, similar to a traditional enterprise messaging system, in which application
          components publish and consume events to/from an ordered stream\n* The long
          term, fault-tolerant storage of events\n* The ability for a consumer to
          replay streams of events\n* The ability to partition topics for horizontal
          scalability\n### Upgrades\n**!!! IMPORTANT !!!** **Direct upgrade from AMQ
          Stream 1.7 or earlier is not supported!** You have to upgrade first to one
          of the previous versions of AMQ Streams. You will also need to convert the
          CRD resources. For more information, see the [documentation for AMQ Streams
          on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).\n###
          Supported Features\n* **Manages the Kafka Cluster** - Deploys and manages
          all of the components of this complex application, including dependencies
          like Apache ZooKeeper® that are traditionally hard to administer.\n* **Includes
          Kafka Connect** - Allows for configuration of common data sources and sinks
          to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for mirroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashboards\n### Upgrading your Clusters\nThe
          AMQ Streams Operator understands how to run and upgrade between a set of
          Kafka versions. When specifying a new version in your config, check to make
          sure you aren't using any features that may have been removed. For more
          information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).\n###
          Storage\nAn efficient data storage infrastructure is essential to the optimal
          performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires
          block storage. The use of file storage (for example, NFS) is not recommended.\nThe
          AMQ Streams Operator supports three types of data storage:\n* Ephemeral
          (Recommended for development only!)\n* Persistent\n* JBOD (Just a Bunch
          of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nAMQ Streams
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation for the current release
          can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).\n"
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:dacf7ddffe7cbe2dc35e7d3290637b3334117e09dde986e6eaeaceded7069fbf
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:be7466ffd7f624f769eedca2e4106c597d2bf1d95cc9679f235e6dcdec66cfa9
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:fa8361e5089ab4449261f5a0ad57bd0bd059dc254b94bd9ab68340faae386947
        - registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:8b10bfb697b48ba3ed246fb846384f4cc67f05f670e9521edf0f47f829869404
        version: 2.6.0-2
      entries:
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.6.x
    - currentCSV: amqstreams.v2.7.0-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:671c1a2fcb0856cb208c0ead539fff9571fa2281a1e3d45b3df9ebd3df8c787e
          createdAt: "2024-03-09 21:27:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0-0 <2.7.0-7'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.6.0 release, Streams supports only OCP 4.12 to 4.14!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:671c1a2fcb0856cb208c0ead539fff9571fa2281a1e3d45b3df9ebd3df8c787e
        - registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:66362c800b78003d26df9471b7f21f5c9ad7fca8e8fd2511428dfc549cdf46c2
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:be9b82af13cbc120bb415abf4fcc97420e543172d6450b836c47344e4dcacd00
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:a7620559f3e5a0a4d5ecee44538619f698158847d9d0a06c7168404ed6f7cd36
        - registry.redhat.io/amq-streams/kafka-36-rhel9@sha256:8890c3217343d9aa3dd2ab9c7fa1742f960b5a26e09906581ce6075e61dc2888
        version: 2.7.0-7
      entries:
      - name: amqstreams.v2.7.0-7
        version: 2.7.0-7
      - name: amqstreams.v2.7.0-6
        version: 2.7.0-6
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.7.x
    - currentCSV: amqstreams.v2.8.0-0-0.1738265624.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:675d4acc375fb20e1a3fd5755184a318341c15f9337a11ea41e902fcda6d0bec
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.0-0 <2.8.0-0'
          olm.substitutesFor: amqstreams.v2.8.0-0
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.8.0 release, Streams supports only OCP 4.12 to 4.14!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:675d4acc375fb20e1a3fd5755184a318341c15f9337a11ea41e902fcda6d0bec
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:b46457a0a7f63518bc12b2789f75afb0e44496e62de918b9b584129c7078de86
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:4df906cf1490922a4587d96250f3be6ff605fe2caf87dfba7b812c144c5c5774
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:e771a84a458388c07ecf970ec625e4203c18e5d2bd91e7571d9c44ebb09313de
        - registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:17b2d255439f38c6ef54c0a56929ae636743b12c688d42223ac3fb58729e51ad
        version: 2.8.0-0+0.1738265624.p
      entries:
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.8.x
    - currentCSV: amqstreams.v2.9.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.9.2-1'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.9.0 release, Streams supports only OCP 4.14 to 4.18!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:f5985592b7f05c9b3c64caa0ac90f34f26b9b4277b342b6bd4a6415d2c563f9a
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:86ea2e823cd83c476521f2a1e9da5619c4810661c60d28295d0f5482c8aa519d
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:ddfb50ec2f1cf45ccccc632182bf0ff549c14969da35b7248802eab67d9c2e9d
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:67647bff1a76adb5b51f43c99d5f88a8aa38055c1e87190add8972bb621ddc15
        version: 2.9.2-1
      entries:
      - name: amqstreams.v2.9.2-1
        version: 2.9.2-1
      - name: amqstreams.v2.9.2-0
        version: 2.9.2-0
      - name: amqstreams.v2.9.1-2
        version: 2.9.1-2
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.9.x
    - currentCSV: amqstreams.v2.9.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.9.2-1'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.9.0 release, Streams supports only OCP 4.14 to 4.18!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:86ea2e823cd83c476521f2a1e9da5619c4810661c60d28295d0f5482c8aa519d
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:ddfb50ec2f1cf45ccccc632182bf0ff549c14969da35b7248802eab67d9c2e9d
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:67647bff1a76adb5b51f43c99d5f88a8aa38055c1e87190add8972bb621ddc15
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:f5985592b7f05c9b3c64caa0ac90f34f26b9b4277b342b6bd4a6415d2c563f9a
        version: 2.9.2-1
      entries:
      - name: amqstreams.v2.9.2-1
        version: 2.9.2-1
      - name: amqstreams.v2.9.2-0
        version: 2.9.2-0
      - name: amqstreams.v2.9.1-2
        version: 2.9.1-2
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.x
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-3.0.x
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-3.x
    - currentCSV: amqstreams.v2.5.2-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
          createdAt: "2023-07-10 21:47:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.2.0-0 <2.5.2-8'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.5.2 release, AMQ Streams supports only OCP 4.12 to 4.16!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more information, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:4ede5eeed159bddacfe830a768bab4996003d4c57a496fc8ec9b41df916d4acc
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:95bc35951eb545b918114af1e92a7139e2e3323b7d054dcb2a060843c7326d28
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:8c7e8ccb6c2d5da32b70d9d0c3919c303ad73a25447f6993d964f7b19cca1072
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
        version: 2.5.2-8
      entries:
      - name: amqstreams.v2.5.2-8
        version: 2.5.2-8
      - name: amqstreams.v2.5.2-7
        version: 2.5.2-7
      - name: amqstreams.v2.5.2-6
        version: 2.5.2-6
      - name: amqstreams.v2.5.2-5
        version: 2.5.2-5
      - name: amqstreams.v2.5.2-4
        version: 2.5.2-4
      - name: amqstreams.v2.5.2-3
        version: 2.5.2-3
      - name: amqstreams.v2.5.2-2
        version: 2.5.2-2
      - name: amqstreams.v2.5.2-1
        version: 2.5.2-1
      - name: amqstreams.v2.5.2-0
        version: 2.5.2-0
      - name: amqstreams.v2.5.1-5
        version: 2.5.1-5
      - name: amqstreams.v2.5.1-4
        version: 2.5.1-4
      - name: amqstreams.v2.5.1-3
        version: 2.5.1-3
      - name: amqstreams.v2.5.1-2
        version: 2.5.1-2
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-lts
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: stable
    defaultChannel: stable
    packageName: amq-streams
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JFrog
      provider-url: ""
    name: openshiftxray-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: xray-operator.v2.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftXray",
                "metadata": {
                  "name": "openshiftxray"
                },
                "spec": {
                  "xray": {
                    "common": {
                      "xrayUserId": "1000721035",
                      "xrayGroupId": "1000721035"
                    },
                    "analysis": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-analysis"
                      },
                      "name": "xray-analysis",
                      "podManagementPolicy": "Parallel",
                      "preStartCommand": null,
                      "updateStrategy": "RollingUpdate"
                    },
                    "database": {
                      "password": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "global": {
                      "postgresqlTlsSecret": null
                    },
                    "indexer": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-indexer"
                      },
                      "name": "xray-indexer",
                      "podManagementPolicy": "Parallel",
                      "updateStrategy": "RollingUpdate"
                    },
                    "persist": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-persist"
                      },
                      "name": "xray-persist",
                      "persistence": {
                        "size": "10Gi"
                      },
                      "podManagementPolicy": "Parallel",
                      "preStartCommand": null,
                      "updateStrategy": "RollingUpdate"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "rabbitmq": {
                      "enabled": true,
                      "replicaCount": 1,
                      "rbac": {
                        "create": true
                      },
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-rabbitmq"
                      },
                      "podSecurityContext": {
                        "fsGroup": 1000721001,
                        "runAsUser": 1000721001
                      },
                      "auth": {
                        "username": "xray",
                        "password": "xray",
                        "erlangCookie": "XRAYRABBITMQCLUSTER",
                        "tls": {
                          "enabled": false,
                          "failIfNoPeerCert": true,
                          "sslOptionsVerify": "verify_peer",
                          "caCertificate": "",
                          "serverCertificate": "",
                          "serverKey": ""
                        }
                      }
                    },
                    "rabbitmq-ha": {
                      "enabled": false,
                      "image": {
                        "repository": "registry.connect.redhat.com/jfrog/xray-rabbitmq"
                      },
                      "rabbitmqUsername": "guest",
                      "rabbitmqPassword": "guest",
                      "managementUsername": "guest",
                      "managementPassword": "guest",
                      "initContainer": {
                        "enabled": false
                      },
                      "securityContext": {
                        "fsGroup": 1000721035,
                        "runAsUser": 1000721035,
                        "runAsGroup": 1000721035
                      },
                      "replicaCount": 1
                    },
                    "replicaCount": 1,
                    "unifiedUpgradeAllowed":  "true",
                    "router": {
                      "image": {
                        "imagePullPolicy": "IfNotPresent",
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-router"
                      },
                      "name": "router"
                    },
                    "server": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-server"
                      },
                      "name": "xray-server",
                      "podManagementPolicy": "Parallel",
                      "replicaCount": 1,
                      "updateStrategy": "RollingUpdate"
                    },
                    "xray": {
                      "consoleLog": false,
                      "jfrogUrl": "OVERRIDE",
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/xray-operator@sha256:da7f6977170876b335875f4074ddfc0266b665351249295ce3c53f0e95891cc2
          createdAt: 2020-05-22 00:00:00+00:00
          description: JFrog Xray Enterprise deploys Xray continuous security scanner
            into Openshift (Requires Jfrog Artifactory)
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Xray Instances
            displayName: Xray
            kind: OpenshiftXray
            name: openshiftxrays.charts.helm.k8s.io
            version: v1
        description: "## Overview Openshift Operator to deploy JFrog Xray Continuous
          Security scanner into your Openshift cluster. NOTE: Artifactory is required
          for JFrog Xray to connect with and scan artifacts against.\n### Major version
          v2.0.4 Breaking change notification!!\nThis release has major bug fix for
          Rabbitmq Pod. Please update to version 2.0.4 of the operator.\n### Major
          version v2.0.0 Breaking change notification!! Xray is dependent upon Rabbitmq.
          The version of Rabbitmq that this chart uses has significantly changed from
          version 1.1.8 of the operator. For full details on how to migrate please
          visit: `https://github.com/jfrog/charts/blob/master/stable/xray/RABBITMQ_MIGRATION_NOTES.md`\nIf
          you are using operator version < 2.1.0 there is a issue with upgrade where
          xray version is not updated if you update operator version.As workaround
          you can update the Custom Resource OpenshiftXray which has hardcored image
          tag value.After upgrading to version 2.1.0 you can remove the image tags
          parameters from your CR so that future upgrade works without workaround.\n###
          Security Context Constraints To deploy this helm chart you will need to
          be a cluster admin w/ access to the anyuid scc and add the operator service
          account to the anyuid scc. ``` oc adm policy add-scc-to-user anyuid -z xray-operator
          ``` ``` oc adm policy add-scc-to-user anyuid -z openshiftxray ``` ``` oc
          adm policy add-scc-to-user anyuid -z openshiftxray-rabbitmq-ha ``` ## Usage
          An external DB is required. The operator will not deploy a DB but will require
          you to specify the configuration values to connect to it. Search for JFrog
          and click JFrog Xray Operator to install. Go to the Installed Operators.
          Wait for the JFrog Xray Operator to complete the installation. Open the
          Operator and click on the provided API: Xray Click Create New Instance and
          provide the following parameters for your DB configuration: ``` JFROG_URL
          DATABASE_URL DATABASE_USER DATABASE_PASSWORD ``` JFROG_URL is the external
          ip or DNS of your Artifactory to connect Xray to. Artifactory is required
          to use this operator. You can set your JFROG_URL to the service name of
          your Artifactory Nginx: ``` oc get svc -n my_namespace | grep nginx ```
          DATABASE_URL must be a Postgresql URL in the format: `` postgres://postgres-postgresql:5432/xraydb?sslmode=disable
          `` DATABASE_USER and DATABASE_PASSWORD must supply a valid user on Postgresql.
          Click Create for Xray to deploy into OpenShift. Open Artifactory in a web
          browser to complete the onboarding wizard for Xray! ### Air gap environments
          To use Xray Operator in an air gap environment you will need to download
          the images as image streams into your Openshift air gap cluster manually.
          Use the image overrides to then specify the image stream locations that
          are local to your cluster. Next you will need to setup the database of index
          data for xray to use when scanning artifacts. Follow the link below for
          instructions on setup: https://www.jfrog.com/confluence/display/JFROG/Configuring+Xray#ConfiguringXray-SynchronizingtheDatabase "
        displayName: JFrog Xray Continuous Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Xray Continuous Security
          url: https://jfrog.com/xray/
        - name: JFrog Xray Continuous Security Video
          url: https://youtu.be/wGzyKlQklKQ
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: alpha
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/xray-indexer@sha256:7f40d2dfe4a35b25d16dfcbfa04b4a58f227c132d4c3c86345e7b2bd19a97311
        - registry.connect.redhat.com/jfrog/xray-router@sha256:019235289e17d0e22dcd23a1505f71a78710827dfab338d8fa980f64c14f162f
        - registry.connect.redhat.com/jfrog/xray-rabbitmq@sha256:459d7d5e78192a6f45a5527e8fbaf1d8b30299b10ad4c42515d3d2bb5fcc75f4
        - registry.connect.redhat.com/jfrog/init@sha256:3f27e44b44c3556c50318110dbd542e34ef1061e28d7f0e0e7b8baa73c5f3e87
        - registry.connect.redhat.com/jfrog/xray-server@sha256:c5584c4e9528f0199e840b2b1ad1144220850ca9595ebf3e270bc6f86de4a71a
        - registry.connect.redhat.com/jfrog/xray-analysis@sha256:bcd8211f1bc753bfd7735165ce39b476e4efff198bffa877883620b41fcb33b5
        - registry.connect.redhat.com/jfrog/xray-persist@sha256:93fae0c124919a57d5162e32e88d3456ef6254889e454f89c21c89e69e6ae323
        - registry.connect.redhat.com/jfrog/xray-operator@sha256:da7f6977170876b335875f4074ddfc0266b665351249295ce3c53f0e95891cc2
        version: 2.1.6
      entries:
      - name: xray-operator.v2.1.6
        version: 2.1.6
      name: alpha
    defaultChannel: alpha
    packageName: openshiftxray-operator
    provider:
      name: JFrog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: eap
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eap-operator.v3.2.2
      currentCSVDesc:
        annotations:
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "false"
          containerImage: registry.redhat.io/jboss-eap/eap-rhel9-operator@sha256:2361e73b8c517177e80f60d204996ebf65b166e4e8b60b329323a2b409c54f6c
          createdAt: "2025-07-10T14:38:00Z"
          description: Operator that creates and manages Java applications running
            on JBoss EAP.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat JBoss Enterprise Application
            Platform", "Red Hat Runtimes", "Red Hat Integration", "Red Hat Process
            Automation"]'
          repository: https://github.com/wildfly/wildfly-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An application running on EAP application runtime.
            displayName: WildFlyServer
            kind: WildFlyServer
            name: wildflyservers.wildfly.org
            version: v1alpha1
        description: "# Red Hat JBoss Enterprise Application Platform (JBoss EAP)\n\nRed
          Hat® JBoss® Enterprise Application Platform (JBoss EAP) delivers enterprise-grade
          security, performance, and scalability in any environment.  Some features
          of JBoss EAP are:\n\n**Optimized for cloud and containers**: with simplified
          deployment and full Jakarta EE performance for applications in any environment.\n\n**Lightweight,
          flexible architecture**: which is well-suited for microservices and traditional
          applications.\n\n**Support and standardize microservices development**:
          With the JBoss Enterprise Application Platform expansion pack that supports
          Eclipse MicroProfile application programming interfaces (APIs).\n\n# JBoss
          EAP Operator\nOperators are software components that extend the Kubernetes
          API to customize how to manage applications and their components.\n\nThe
          JBoss EAP Operator facilitates the deployment and management of JBoss EAP
          on container orchestration platforms such as Red Hat OpenShift. The operator
          acts as an automated management tool, simplifying the process of provisioning,
          scaling, and updating JBoss EAP applications.\n\n# JBoss EAP Operator capabilities\nThe
          JBoss EAP Operator deploys and manages multiple instances of the JBoss EAP
          Java application in a Red Hat OpenShift cluster. \n\n**Simplifies management**:
          Creates, configures, manages, and seamlessly upgrades instances of complex
          stateful applications. No more manual configuration, making management a
          breeze.\n\n**Efficient scaling**: Effortlessly scales the instances of your
          JBoss EAP Java application based on demand.\n\n**Faster updates**: Seamlessly
          updates, enabling you to roll out new operator features and bug fixes without
          downtime\n\n**Safe transaction recovery**: Provides safe transaction recovery
          in your application cluster by verifying all transactions are completed
          before scaling down the replicas and marking a pod as clean for termination.\n\n#
          Operator installation and configuration\nLet's walk through the steps to
          install the JBoss EAP Operator on your OpenShift Cluster and deploy your
          first example application:\n\n**Step 1: Install JBoss EAP from OperatorHub**\n1.
          Click Install in the JBoss EAP operator window.\n2. In the Install Operator
          window, for Installation mode, select All namespaces on the cluster The
          operator is available in openshift-operators namespace.\n\n**Step 2: Build
          an application image**\n\nThe JBoss EAP Operator deploys an application
          image. If you do not have your JBoss EAP application image built yet, you
          can use the following example to build the image using Helm Charts.\n\n1.
          In the main navigation, click the dropdown menu (perspective switcher) and
          select Developer.\n2. Click Add > Helm Chart.\n3. Search for \"JBoss EAP
          8\" (Provided by Red Hat) and create a new Helm Chart release.\n4. On the
          Release name text box, type “eap-application”. \n5. Switch to the YAML view
          and replace the existing code with the following:\n\n    ```\n    build:\n
          \     uri: https://github.com/jboss-eap-up-and-running/eap8-getting-started\n
          \   deploy:\n      enabled: false\n    ```\n\n6. Click Create to build the
          application image.\n\n    The application image is created with the name
          eap-application:latest.\n\n**Step 3: Deploy your application with JBoss
          EAP Operator**\n\n1. In the main navigation, click the dropdown menu (perspective
          switcher) and select Developer.\n2. Click Add > Operator Backed > WildFlyServer
          > Create.\n3. Switch to the YAML view and replace the existing code with
          the following:\n\n    ```\n    apiVersion: wildfly.org/v1alpha1\n    kind:
          WildFlyServer\n    metadata:\n      name: eap-application\n    spec:\n      applicationImage:
          eap-application:latest\n      replicas: 1\n    ```\n\n 4. Click Create to
          deploy the application with JBoss EAP Operator.\n"
        displayName: JBoss EAP
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eap
        - java
        - open source
        - application runtime
        links:
        - name: JBoss EAP Homepage
          url: https://www.redhat.com/en/technologies/jboss-middleware/application-platform
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_jboss_enterprise_application_platform/
        - name: Blog
          url: https://developers.redhat.com/blog/
        - name: Quickstart
          url: https://github.com/jboss-developer/jboss-eap-quickstarts
        - name: Red Hat Customer Portal
          url: https://access.redhat.com/support
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat, Inc.
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/jboss-eap/eap-rhel9-operator@sha256:2361e73b8c517177e80f60d204996ebf65b166e4e8b60b329323a2b409c54f6c
        version: 3.2.2
      entries:
      - name: eap-operator.v3.2.2
        version: 3.2.2
      - name: eap-operator.v3.2.1
        version: 3.2.1
      - name: eap-operator.v3.2.0
        version: 3.2.0
      - name: eap-operator.v3.1.9
        version: 3.1.9
      - name: eap-operator.v3.1.8
        version: 3.1.8
      - name: eap-operator.v3.1.7
        version: 3.1.7
      - name: eap-operator.v3.1.6
        version: 3.1.6
      - name: eap-operator.v3.1.5
        version: 3.1.5
      - name: eap-operator.v3.1.4
        version: 3.1.4
      - name: eap-operator.v3.1.3
        version: 3.1.3
      - name: eap-operator.v3.1.2
        version: 3.1.2
      - name: eap-operator.v3.1.1
        version: 3.1.1
      - name: eap-operator.v3.1.0
        version: 3.1.0
      - name: eap-operator.v3.0.2
        version: 3.0.2
      - name: eap-operator.v3.0.1
        version: 3.0.1
      - name: eap-operator.v3.0.0
        version: 3.0.0
      - name: eap-operator.v2.4.3
        version: 2.4.3
      - name: eap-operator.v2.4.2
        version: 2.4.2
      - name: eap-operator.v2.4.1
        version: 2.4.1
      - name: eap-operator.v2.4.0
        version: 2.4.0
      - name: eap-operator.v2.3.10
        version: 2.3.10
      - name: eap-operator.v2.3.9
        version: 2.3.9
      - name: eap-operator.v2.3.8
        version: 2.3.8
      - name: eap-operator.v2.3.7
        version: 2.3.7
      - name: eap-operator.v2.3.6
        version: 2.3.6
      - name: eap-operator.v2.3.5
        version: 2.3.5
      - name: eap-operator.v2.3.4
        version: 2.3.4
      - name: eap-operator.v2.3.3
        version: 2.3.3
      - name: eap-operator.v2.3.2
        version: 2.3.2
      - name: eap-operator.v2.3.1
        version: 2.3.1
      - name: eap-operator.v2.3.0
        version: 2.3.0
      - name: eap-operator.v2.2.2
        version: 2.2.2
      - name: eap-operator.v2.2.1
        version: 2.2.1
      - name: eap-operator.v2.2.0
        version: 2.2.0
      name: stable
    defaultChannel: stable
    packageName: eap
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: m88i Labs
      provider-url: ""
    name: nexus-operator-m88i
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.m88i.io/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "name": "nexus3"
                },
                "spec": {
                  "networking": {
                    "expose": false
                  },
                  "persistence": {
                    "persistent": false
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    },
                    "requests": {
                      "cpu": "1",
                      "memory": "2Gi"
                    }
                  },
                  "useRedHatImage": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/m88i/nexus-operator:0.6.0
          createdAt: "2019-11-16T13:12:22Z"
          description: Nexus Operator to deploy and manage Nexus 3.x servers
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/m88i/nexus-operator
          support: m88i Labs
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: Nexus
            kind: Nexus
            name: nexus.apps.m88i.io
            version: v1alpha1
        description: |-
          Creates a new Nexus 3.x deployment in a Kubernetes cluster. Will help DevOps to have a quick Nexus application exposed to the world that can be used in a CI/CD process:

          * Deploys a new Nexus 3.x server based on either Community or Red Hat images
          * Creates an [Ingress](https://kubernetes.io/docs/concepts/services-networking/ingress/) in Kubernetes (1.14+) environments to expose the application to the world
          * On OpenShift, creates a Route to expose the service outside the cluster
          * Automatically creates Apache, Red Hat and JBoss Maven repositories
          * Automatically updates Nexus within the same minor version

          [See our documentation](https://github.com/m88i/nexus-operator/blob/main/README.md) for more installation and usage scenarios.

          If you experience any issues or have any ideas for new features, please [file an issue in our Github repository](https://github.com/m88i/nexus-operator/issues) or send an email to our maillist: [nexus-operator@googlegroups.com](mailto:nexus-operator@googlegroups.com)

          *Please note that the operator is an individual work and it's not provided nor supported by Sonatype.*
        displayName: Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nexus
        - sonatype
        - maven
        - docker
        - ci
        - continuous integration
        - continuous delivery
        - repository
        - repository manager
        - dev tools
        links:
        - name: Documentation
          url: https://github.com/m88i/nexus-operator/blob/main/README.md
        - name: Nexus Operator source code repository
          url: https://github.com/m88i/nexus-operator
        maintainers:
        - email: nexus-operator@googlegroups.com
          name: m88i Labs
        maturity: alpha
        provider:
          name: m88i Labs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/m88i/nexus-operator:0.6.0
        version: 0.6.0
      entries:
      - name: nexus-operator.v0.6.0
        version: 0.6.0
      name: alpha
    defaultChannel: alpha
    packageName: nexus-operator-m88i
    provider:
      name: m88i Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Aerospike
      provider-url: ""
    name: aerospike-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aerospike-kubernetes-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "memory-size": 3000000000,
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:6.3.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: docker.io/aerospike/aerospike-kubernetes-operator:2.5.0
          createdAt: 2023-05-04T09:32:47UTC
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Aerospike
        relatedImages:
        - docker.io/aerospike/aerospike-kubernetes-operator:2.5.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 2.5.0
      entries:
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      - name: aerospike-kubernetes-operator.v2.4.0
        version: 2.4.0
      - name: aerospike-kubernetes-operator.v2.3.0
        version: 2.3.0
      - name: aerospike-kubernetes-operator.v2.2.1
        version: 2.2.1
      - name: aerospike-kubernetes-operator.v2.2.0
        version: 2.2.0
      - name: aerospike-kubernetes-operator.v2.1.0
        version: 2.1.0
      name: alpha
    - currentCSV: aerospike-kubernetes-operator.v4.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "data-size": 1073741824,
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:8.1.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackup",
                "metadata": {
                  "name": "aerospikebackup",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "aerospike-cluster": {
                      "aerospike-aerospikebackup-test-cluster": {
                        "credentials": {
                          "password": "admin123",
                          "user": "admin"
                        },
                        "seed-nodes": [
                          {
                            "host-name": "aerocluster.aerospike.svc.cluster.local",
                            "port": 3000
                          }
                        ]
                      }
                    },
                    "backup-routines": {
                      "aerospike-aerospikebackup-test-routine": {
                        "backup-policy": "test-policy",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "local"
                      },
                      "aerospike-aerospikebackup-test-routine1": {
                        "backup-policy": "test-policy1",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "s3Storage"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackupService",
                "metadata": {
                  "name": "aerospikebackupservice-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "config": {
                    "backup-policies": {
                      "test-policy": {
                        "parallel": 3
                      },
                      "test-policy1": {
                        "parallel": 3
                      }
                    },
                    "service": {
                      "http": {
                        "port": 8081
                      }
                    },
                    "storage": {
                      "local": {
                        "local-storage": {
                          "path": "/tmp/localStorage"
                        }
                      },
                      "s3Storage": {
                        "s3-storage": {
                          "bucket": "aerospike-kubernetes-operator-test",
                          "s3-profile": "default",
                          "s3-region": "us-east-1"
                        }
                      }
                    }
                  },
                  "image": "aerospike/aerospike-backup-service:3.2.0",
                  "secrets": [
                    {
                      "secretName": "aws-secret",
                      "volumeMount": {
                        "mountPath": "/root/.aws/credentials",
                        "name": "aws-secret",
                        "subPath": "credentials"
                      }
                    }
                  ],
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeRestore",
                "metadata": {
                  "name": "aerospikerestore-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "backup-data-path": "aerospike-aerospikebackup-test-routine/backup/1733506015827/data/test",
                    "destination": {
                      "credentials": {
                        "password": "admin123",
                        "user": "admin"
                      },
                      "label": "destinationCluster",
                      "seed-nodes": [
                        {
                          "host-name": "aerocluster.test.svc.cluster.local",
                          "port": 3000
                        }
                      ]
                    },
                    "policy": {
                      "no-generation": true,
                      "no-indexes": true,
                      "parallel": 3
                    },
                    "source": {
                      "local-storage": {
                        "path": "/tmp/localStorage"
                      }
                    }
                  },
                  "type": "Full"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: docker.io/aerospike/aerospike-kubernetes-operator:4.1.1
          createdAt: "2025-09-15T10:14:05Z"
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          olm.skipRange: '>=4.0.0 <4.1.1'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
          support: Aerospike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeBackup is the Schema for the aerospikebackup API
            displayName: Aerospike Backup
            kind: AerospikeBackup
            name: aerospikebackups.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeBackupService is the Schema for the aerospikebackupservices
              API
            displayName: Aerospike Backup Service
            kind: AerospikeBackupService
            name: aerospikebackupservices.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1
          - description: AerospikeRestore is the Schema for the aerospikerestores
              API
            displayName: Aerospike Restore
            kind: AerospikeRestore
            name: aerospikerestores.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
           -   Aerospike cluster monitoring
           -   Backup and restore Aerospike clusters
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Aerospike
        relatedImages:
        - docker.io/aerospike/aerospike-kubernetes-operator:4.1.1
        version: 4.1.1
      entries:
      - name: aerospike-kubernetes-operator.v4.1.1
        version: 4.1.1
      - name: aerospike-kubernetes-operator.v4.1.0
        version: 4.1.0
      - name: aerospike-kubernetes-operator.v4.0.2
        version: 4.0.2
      - name: aerospike-kubernetes-operator.v4.0.1
        version: 4.0.1
      - name: aerospike-kubernetes-operator.v4.0.0
        version: 4.0.0
      - name: aerospike-kubernetes-operator.v3.4.1
        version: 3.4.1
      - name: aerospike-kubernetes-operator.v3.4.0
        version: 3.4.0
      - name: aerospike-kubernetes-operator.v3.3.1
        version: 3.3.1
      - name: aerospike-kubernetes-operator.v3.3.0
        version: 3.3.0
      - name: aerospike-kubernetes-operator.v3.2.2
        version: 3.2.2
      - name: aerospike-kubernetes-operator.v3.2.1
        version: 3.2.1
      - name: aerospike-kubernetes-operator.v3.2.0
        version: 3.2.0
      - name: aerospike-kubernetes-operator.v3.1.0
        version: 3.1.0
      - name: aerospike-kubernetes-operator.v3.0.0
        version: 3.0.0
      name: stable
    defaultChannel: stable
    packageName: aerospike-kubernetes-operator
    provider:
      name: Aerospike
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: 3scale-community-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: 3scale-community-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.8.2
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Improving [documentation](https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.8.2
        version: 0.8.2
      entries:
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.11
    - currentCSV: 3scale-community-operator.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.9.0
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.9.0/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.9.0/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.9.0
        version: 0.9.0
      entries:
      - name: 3scale-community-operator.v0.9.0
        version: 0.9.0
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.12
    - currentCSV: 3scale-community-operator.v0.10.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "ProductCRName": "product1-sample",
                  "Production": true
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.10.1
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.10.1/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Improving [documentation](https://github.com/3scale/3scale-operator/blob/master/doc/operator-user-guide.md)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.10.1/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.10.1
        version: 0.10.1
      entries:
      - name: 3scale-community-operator.v0.10.1
        version: 0.10.1
      - name: 3scale-community-operator.v0.9.0
        version: 0.9.0
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.13
    defaultChannel: threescale-2.13
    packageName: 3scale-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: https://www.intel.com
    name: ipm-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ipm-controller.v25.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ipm.intel.com/v1",
                "kind": "CPUPowerProv",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ipm-controller",
                    "node-label": "ipm-provisioning"
                  },
                  "name": "cpupowerprov-sample"
                },
                "spec": {
                  "cstates": {
                    "cores": {
                      "5": {
                        "C1": true,
                        "C1E": true,
                        "C6": false
                      },
                      "9,12": {
                        "C1": true,
                        "C6": true
                      }
                    },
                    "idleGovernor": "ladder"
                  },
                  "pstates": {
                    "cores": {
                      "5": {
                        "energyPerformanceBias": "normal",
                        "energyPerformancePreference": "performance",
                        "freqGovernor": "performance",
                        "maxFreqMHz": "3500",
                        "minFreqMHz": "800"
                      },
                      "9,12": {
                        "freqGovernor": "powersave",
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    },
                    "global": {
                      "conservative": {
                        "downThreshold": 20,
                        "freqStep": 5,
                        "ignoreNiceLoad": 0,
                        "samplingDownFactor": 1,
                        "samplingRate": 8888,
                        "upThreshold": 88
                      },
                      "maxPerfPct": 95,
                      "minPerfPct": 45,
                      "ondemand": {
                        "ignoreNiceLoad": 0,
                        "ioIsBusy": 1,
                        "powersaveBias": 0,
                        "samplingDownFactor": 1,
                        "samplingRate": 8888,
                        "upThreshold": 88
                      },
                      "schedutil": {
                        "rateLimitUs": 5000
                      },
                      "turbo": true
                    }
                  },
                  "uncoreFreqCPU": {
                    "cpu0": [
                      {
                        "maxFreqMHz": "1600",
                        "minFreqMHz": "800"
                      }
                    ],
                    "cpu1": [
                      {
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    ]
                  },
                  "uncoreFreqIO": {
                    "cpu0": [
                      {
                        "maxFreqMHz": "1600",
                        "minFreqMHz": "800"
                      }
                    ],
                    "cpu1": [
                      {
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/ipm-controller@sha256:32135c888f86a2c725d093ada9a59d222229bee5640d37f6e19a9e9812b4b67a
          createdAt: "2025-08-07T13:13:25Z"
          description: IPM Controller manages resources via REST API interactions.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.operatorframework.io/required-sccs: '["privileged"]'
          olm.securityContextConstraints.required: privileged
          operators.openshift.io/infrastructure-features: RestrictedNetwork
          operators.openshift.io/valid-subscription: Intel OBL Limited Distribution
            (Commercial Use) License Agreement [v2022.12.20]
          operators.operatorframework.io.bundle.package.v1: ipm-controller
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Intel® Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CPUPowerProv is the Schema for the cpupowerprovs API
            displayName: CPUPower Prov
            kind: CPUPowerProv
            name: cpupowerprovs.ipm.intel.com
            version: v1
        description: Infrastructure Power Manager (IPM) governs platform power technologies
          based on input from application workloads, the platform itself and customer
          input
        displayName: Intel® Infrastructure Power Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - intel
        - infrastructure
        - power
        - manager
        - management
        links:
        - name: IPM
          url: https://builders.intel.com/intel-technologies/power-management
        maintainers:
        - email: infrastructure.power.manager.operator@intel.com
          name: Infrastructure Power Manager Operator
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Intel® Corporation
          url: https://www.intel.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.13.0@sha256:471549d9cc07fcec0f285a05745b3d39a48e9a503b01b4047156a86ec1a30c00
        - registry.connect.redhat.com/intel/ipm-controller@sha256:32135c888f86a2c725d093ada9a59d222229bee5640d37f6e19a9e9812b4b67a
        - registry.connect.redhat.com/intel/ipm-agent@sha256:54de45cdc2e34768f6cdbf1cdee219406684143dddc4d236877709dc2536fe9c
        - registry.connect.redhat.com/intel/ipm-watcher@sha256:51eb773d5a4d1f401d4cbf2b42b8df4d0359beb85d5e274b8392dd7c5165e6ff
        version: 25.7.0
      entries:
      - name: ipm-controller.v25.7.0
        version: 25.7.0
      name: stable
    defaultChannel: stable
    packageName: ipm-controller
    provider:
      name: Intel® Corporation
      url: https://www.intel.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: ""
    name: nim-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nim-operator-certified.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMCache",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "source": {
                    "ngc": {
                      "modelPuller": "nvcr.io/nim/meta/llama-3.2-1b-instruct@sha256:1ff21ad32124d2382f8c733879394e7849cc3df5172651a24fcb671ac4fe77c6",
                      "pullSecret": "ngc-secret",
                      "authSecret": "ngc-api-secret",
                      "model": {
                        "lora": false,
                        "precision": "fp16",
                        "engine": "tensorrt_llm",
                        "qosProfile": "throughput",
                        "gpus": [
                          {
                            "product": "a100",
                            "ids": []
                          }
                        ]
                      }
                    }
                  },
                  "storage": {
                    "pvc": {
                      "create": true,
                      "storageClass": "local-path",
                      "size": "50Gi",
                      "volumeAccessMode": "ReadWriteOnce"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMService",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "image": {
                    "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                    "tag": "1.12.0",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [
                      "ngc-secret"
                    ]
                  },
                  "authSecret": "ngc-api-secret",
                  "storage": {
                    "nimCache": {
                      "name": "meta-llama3-1b-instruct",
                      "profile": ""
                    }
                  },
                  "scale": {
                    "enabled": false,
                    "hpa": {
                      "minReplicas": 1,
                      "maxReplicas": 10
                    }
                  },
                  "resources": {
                    "limits": {
                      "nvidia.com/gpu": 1
                    }
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMPipeline",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "k8s-nim-operator",
                    "app.kubernetes.io/managed-by": "k8s-nim-operator"
                  },
                  "name": "nimpipeline-sample"
                },
                "spec": {
                  "services": [
                    {
                      "name": "meta-llama3-1b-instruct",
                      "enabled": true,
                      "spec": {
                        "image": {
                          "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                          "tag": "1.12.0",
                          "pullPolicy": "IfNotPresent",
                          "pullSecrets": [
                            "ngc-secret"
                          ]
                        },
                        "authSecret": "ngc-api-secret",
                        "storage": {
                          "nimCache": {
                            "name": "meta-llama3-1b-instruct",
                            "profile": ""
                          }
                        },
                        "replicas": 1,
                        "resources": {
                          "limits": {
                            "nvidia.com/gpu": 1
                          }
                        },
                        "expose": {
                          "service": {
                            "type": "ClusterIP",
                            "port": 8000
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMBuild",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "nimCache": {
                    "name": "meta-llama3-1b-instruct",
                    "profile": "7cc8597690a35aba19a3636f35e7f1c7e7dbc005fe88ce9394cad4a4adeed414"
                  },
                  "image": {
                    "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                    "tag": "1.12.0",
                    "pullSecrets": [
                     "ngc-secret"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoCustomizer",
                "metadata": {
                  "name": "nemocustomizer-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "scheduler": {
                    "type": "volcano"
                  },
                  "wandb": {
                    "secretName": "wandb-secret",
                    "apiKeyKey": "apiKey",
                    "encryptionKey": "encryptionKey"
                  },
                  "otel": {
                    "enabled": true,
                    "exporterOtlpEndpoint": "http://customizer-otel-opentelemetry-collector.nemo.svc.cluster.local:4317"
                  },
                  "databaseConfig": {
                    "credentials": {
                      "user": "ncsuser",
                      "secretName": "customizer-pg-existing-secret",
                      "passwordKey": "password"
                    },
                    "host": "customizer-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "ncsdb"
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/customizer-api",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "entitystore": {
                    "endpoint": "http://nemoentitystore-sample.nemo.svc.cluster.local:8000"
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000"
                  },
                  "mlflow": {
                    "endpoint": "http://mlflow-tracking.nemo.svc.cluster.local:80"
                  },
                  "nemoDatastoreTools": {
                    "image": "nvcr.io/nvidia/nemo-microservices/nds-v2-huggingface-cli@sha256:7958e5244217de8af0ed723dfe7ea8e291b3afbb2a0aa73f5c27b218a1149b11"
                  },
                  "modelDownloadJobs": {
                    "image": "nvcr.io/nvidia/nemo-microservices/customizer-api@sha256:c590fa19121af2aac0379933deace3991f27c3ef88a8043560d6e3a172273b0b",
                    "ngcAPISecret": {
                      "name": "ngc-api-secret",
                      "key": "NGC_API_KEY"
                    },
                    "securityContext": {
                      "fsGroup": 1000,
                      "runAsNonRoot": true,
                      "runAsUser": 1000,
                      "runAsGroup": 1000
                    },
                    "ttlSecondsAfterFinished": 600,
                    "pollIntervalSeconds": 15
                  },
                  "modelConfig": {
                    "name": "nemo-model-config"
                  },
                  "trainingConfig": {
                    "configMap": {
                      "name": "nemo-training-config"
                    },
                    "modelPVC": {
                      "create": true,
                      "name": "finetuning-ms-models-pvc",
                      "storageClass": "",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": 50
                    },
                    "workspacePVC": {
                      "storageClass": "local-path",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": 10,
                      "mountPath": "/pvc/workspace"
                    },
                    "image": {
                      "repository": "nvcr.io/nvidia/nemo-microservices/customizer",
                      "tag": "25.08"
                    },
                    "env": [
                      { "name": "LOG_LEVEL", "value": "INFO" }
                    ],
                    "networkConfig": [
                      { "name": "NCCL_IB_SL", "value": "0" },
                      { "name": "NCCL_IB_TC", "value": "41" },
                      { "name": "NCCL_IB_QPS_PER_CONNECTION", "value": "4" },
                      { "name": "UCX_TLS", "value": "TCP" },
                      { "name": "UCX_NET_DEVICES", "value": "eth0" },
                      { "name": "HCOLL_ENABLE_MCAST_ALL", "value": "0" },
                      { "name": "NCCL_IB_GID_INDEX", "value": "3" }
                    ],
                    "ttlSecondsAfterFinished": 3600,
                    "timeout": 3600,
                    "tolerations": [
                      {
                        "key": "nvidia.com/gpu",
                        "operator": "Exists",
                        "effect": "NoSchedule"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoGuardrail",
                "metadata": {
                  "name": "nemoguardrails-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "configStore": {
                    "pvc": {
                      "name": "pvc-guardrail-config",
                      "create": true,
                      "storageClass": "",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": "1Gi"
                    }
                  },
                  "nimEndpoint": {
                    "baseURL": "http://meta-llama3-1b-instruct.nemo.svc.cluster.local:8000/v1"
                  },
                  "databaseConfig": {
                    "host": "guardrail-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "guardraildb",
                    "credentials": {
                      "user": "guardrailuser",
                      "secretName": "guardrail-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/guardrails",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [
                      "ngc-secret"
                    ]
                  },
                  "metrics": {
                    "serviceMonitor": {}
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "ephemeral-storage": "10Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoEvaluator",
                "metadata": {
                  "name": "nemoevaluator-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "evaluationImages": {
                    "bigcodeEvalHarness": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-bigcode@sha256:62f74c32d1ed07925ddb906de8975025bf63d73e82e7faefb840339aaf21a9dc",
                    "lmEvalHarness": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-lm-eval-harness@sha256:58e7c3da7f1edfed2ff508fa8abab7660366c2c17bb55d17aa68159d6e16be6d",
                    "similarityMetrics": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-custom-eval@sha256:eb51edf668481d1f5d2b100c850958dadf184fd8779433e45da40fa2a35ec4b9",
                    "llmAsJudge": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055",
                    "mtBench": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055",
                    "retriever": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-retriever@sha256:cb99fb3ea5b4e9989832679d4c6028573c28b99e43744e008ac5c9748b7dd250",
                    "rag": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-rag@sha256:b5e35cf277185e0901a4bbba35780b3673be71ae2c5e8efcdbab7710a8a6f783",
                    "bfcl": "nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-bfcl@sha256:95f45f66140d481d544b36b16bc3de7ad5e090d76225fd73ed70461b9bd9907c",
                    "agenticEval": "nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-agentic-eval@sha256:b38bbc5e61df0930204efa096cf4f40db4655b92d03741190513b019d67fc7dc"
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/evaluator",
                    "tag": "25.06",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "argoWorkflows": {
                    "endpoint": "https://argo-workflows-server.nemo.svc.cluster.local:2746",
                    "serviceAccount": "argo-workflows-executor"
                  },
                  "vectorDB": {
                    "endpoint": "http://milvus.nemo.svc.cluster.local:19530"
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000/v1/hf"
                  },
                  "entitystore": {
                    "endpoint": "http://nemoentitystore-sample.nemo.svc.cluster.local:8000"
                  },
                  "databaseConfig": {
                    "host": "evaluator-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "evaldb",
                    "credentials": {
                      "user": "evaluser",
                      "secretName": "evaluator-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "otel": {
                    "enabled": true,
                    "exporterOtlpEndpoint": "http://evaluator-otel-opentelemetry-collector.nemo.svc.cluster.local:4317"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoEntitystore",
                "metadata": {
                  "name": "nemoentitystore-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/entity-store",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "databaseConfig": {
                    "databaseName": "nesdb",
                    "host": "entity-store-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "credentials": {
                      "user": "nesuser",
                      "secretName": "entity-store-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000"
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoDatastore",
                "metadata": {
                  "name": "nemodatastore-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "secrets": {
                    "datastoreConfigSecret": "nemo-ms-nemo-datastore",
                    "datastoreInitSecret": "nemo-ms-nemo-datastore-init",
                    "datastoreInlineConfigSecret": "nemo-ms-nemo-datastore-inline-config",
                    "giteaAdminSecret": "gitea-admin-credentials",
                    "lfsJwtSecret": "nemo-ms-nemo-datastore--lfs-jwt"
                  },
                  "databaseConfig": {
                    "credentials": {
                      "user": "ndsuser",
                      "secretName": "datastore-pg-existing-secret",
                      "passwordKey": "password"
                    },
                    "host": "datastore-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "ndsdb"
                  },
                  "pvc": {
                    "name": "pvc-shared-data",
                    "create": true,
                    "storageClass": "",
                    "volumeAccessMode": "ReadWriteOnce",
                    "size": "10Gi"
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/datastore",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "replicas": 1,
                  "resources": {
                    "requests": {
                      "memory": "256Mi",
                      "cpu": "500m"
                    },
                    "limits": {
                      "memory": "512Mi",
                      "cpu": "1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: nvcr.io/nvidia/cloud-native/k8s-nim-operator@sha256:f42dd5c3dc269553ee46725ed4b1ab2d2f253ef06117f1d9d52faaa22d8fe60d
          createdAt: Wed Sep 10 05:49:48 PDT 2025
          description: An Operator for deployment and maintenance of NVIDIA NIMs and
            NeMo microservices in a Kubernetes environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          k8sMinVersion: 1.26.0
          repository: github.com/NVIDIA/k8s-nim-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NIM Cache
            displayName: NIMCache
            kind: NIMCache
            name: nimcaches.apps.nvidia.com
            version: v1alpha1
          - description: NIM Service
            displayName: NIMService
            kind: NIMService
            name: nimservices.apps.nvidia.com
            version: v1alpha1
          - description: NIM Pipeline
            displayName: NIMPipeline
            kind: NIMPipeline
            name: nimpipelines.apps.nvidia.com
            version: v1alpha1
          - description: NIM Build
            displayName: NIMBuild
            kind: NIMBuild
            name: nimbuilds.apps.nvidia.com
            version: v1alpha1
          - description: NEMO guardrail Service
            displayName: NemoGuardrail
            kind: NemoGuardrail
            name: nemoguardrails.apps.nvidia.com
            version: v1alpha1
          - description: NEMO datastore Service
            displayName: NemoDatastore
            kind: NemoDatastore
            name: nemodatastores.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Eval Service
            displayName: NemoEvaluator
            kind: NemoEvaluator
            name: nemoevaluators.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Customizer Service
            displayName: NemoCustomizer
            kind: NemoCustomizer
            name: nemocustomizers.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Entitystore Service
            displayName: NemoEntitystore
            kind: NemoEntitystore
            name: nemoentitystores.apps.nvidia.com
            version: v1alpha1
        description: |-
          ## NVIDIA NIM Operator
          The NVIDIA NIM Operator enables Kubernetes cluster administrators to operate the software components and services necessary to run NVIDIA NIMs in various domains such as reasoning, retrieval, speech, and biology. Additionally, it allows the use of NeMo Microservices to fine-tune, evaluate, or apply guardrails to your models.
          The Operator manages the life cycle of the following microservices and the models they use:
          ### NVIDIA NIM models, such as:
          * Reasoning LLMs
          * Retrieval - Embedding, Reranking, etc.
          * Speech
          * Biology
          ### NeMo core microservices:
          * NeMo Customizer
          * NeMo Evaluator
          * NeMo Guardrails
          ### NeMo platform component microservices:
          * NeMo Data Store
          * NeMo Entity Store
        displayName: The NVIDIA NIM Operator for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        links:
        - name: NVIDIA NIM
          url: https://docs.nvidia.com/nim/index.html
        - name: Documentation
          url: https://docs.nvidia.com/nim-operator/latest
        maintainers:
        - email: operator_feedback@nvidia.com
          name: shivamerla
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: NVIDIA
        relatedImages:
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-retriever@sha256:cb99fb3ea5b4e9989832679d4c6028573c28b99e43744e008ac5c9748b7dd250
        - nvcr.io/nvidia/nemo-microservices/customizer-api@sha256:c590fa19121af2aac0379933deace3991f27c3ef88a8043560d6e3a172273b0b
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-lm-eval-harness@sha256:58e7c3da7f1edfed2ff508fa8abab7660366c2c17bb55d17aa68159d6e16be6d
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-custom-eval@sha256:eb51edf668481d1f5d2b100c850958dadf184fd8779433e45da40fa2a35ec4b9
        - nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-bfcl@sha256:95f45f66140d481d544b36b16bc3de7ad5e090d76225fd73ed70461b9bd9907c
        - nvcr.io/nvidia/cloud-native/k8s-nim-operator@sha256:f42dd5c3dc269553ee46725ed4b1ab2d2f253ef06117f1d9d52faaa22d8fe60d
        - nvcr.io/nvidia/nemo-microservices/nds-v2-huggingface-cli@sha256:7958e5244217de8af0ed723dfe7ea8e291b3afbb2a0aa73f5c27b218a1149b11
        - nvcr.io/nim/meta/llama-3.2-1b-instruct@sha256:1ff21ad32124d2382f8c733879394e7849cc3df5172651a24fcb671ac4fe77c6
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-bigcode@sha256:62f74c32d1ed07925ddb906de8975025bf63d73e82e7faefb840339aaf21a9dc
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-rag@sha256:b5e35cf277185e0901a4bbba35780b3673be71ae2c5e8efcdbab7710a8a6f783
        - nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-agentic-eval@sha256:b38bbc5e61df0930204efa096cf4f40db4655b92d03741190513b019d67fc7dc
        version: 3.0.0
      entries:
      - name: nim-operator-certified.v3.0.0
        version: 3.0.0
      - name: nim-operator-certified.v2.0.2
        version: 2.0.2
      - name: nim-operator-certified.v2.0.1
        version: 2.0.1
      - name: nim-operator-certified.v2.0.0
        version: 2.0.0
      - name: nim-operator-certified.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: stable
    packageName: nim-operator-certified
    provider:
      name: NVIDIA
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-druid-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: druid-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:909a03d95f8173c418f6547092421a5531ff8957aa4319ce5fc890f32fe1f7be
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:909a03d95f8173c418f6547092421a5531ff8957aa4319ce5fc890f32fe1f7be
        version: 24.11.1
      entries:
      - name: druid-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: druid-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:253239268b8d32dbea10233fe6ab202bbe8a074c369a929630681dea37935383
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:253239268b8d32dbea10233fe6ab202bbe8a074c369a929630681dea37935383
        version: 25.3.0
      entries:
      - name: druid-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: druid-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
        version: 25.7.0
      entries:
      - name: druid-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: druid-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
        version: 25.7.0
      entries:
      - name: druid-operator.v25.7.0
        version: 25.7.0
      - name: druid-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-druid-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: aiu-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aiu-operator.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "labels": {
                    "operator-version": "2.1.0-dev"
                  },
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "aiu": {
                    "strategy": "single"
                  },
                  "devicePlugin": {
                    "aiuSelector": {
                      "deviceID": "06a7",
                      "vendor": "1014"
                    },
                    "image": "aiu-device-plugin",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "deviceType": "vfio-pci",
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation"
                  ],
                  "featureDiscovery": {
                    "image": "aiu-feature-discovery",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "isRdma": false,
                  "loglevel": "info",
                  "metricsExporter": {
                    "enabled": true,
                    "image": "aiu-exporter",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "numVfs": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
          createdAt: "2024-10-07T13:07:51Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 2.1.1
      entries:
      - name: aiu-operator.v2.1.1
        version: 2.1.1
      name: candidate-2.1
    - currentCSV: aiu-operator.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "labels": {
                    "operator-version": "2.1.0-dev"
                  },
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "aiu": {
                    "strategy": "single"
                  },
                  "devicePlugin": {
                    "aiuSelector": {
                      "deviceID": "06a7",
                      "vendor": "1014"
                    },
                    "image": "aiu-device-plugin",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "deviceType": "vfio-pci",
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation"
                  ],
                  "featureDiscovery": {
                    "image": "aiu-feature-discovery",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "isRdma": false,
                  "loglevel": "info",
                  "metricsExporter": {
                    "enabled": true,
                    "image": "aiu-exporter",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "numVfs": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
          createdAt: "2024-10-07T13:07:51Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
        version: 2.1.1
      entries:
      - name: aiu-operator.v2.1.1
        version: 2.1.1
      name: fast-2.1
    - currentCSV: aiu-operator.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "labels": {
                    "operator-version": "2.1.0-dev"
                  },
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "aiu": {
                    "strategy": "single"
                  },
                  "devicePlugin": {
                    "aiuSelector": {
                      "deviceID": "06a7",
                      "vendor": "1014"
                    },
                    "image": "aiu-device-plugin",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "deviceType": "vfio-pci",
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation"
                  ],
                  "featureDiscovery": {
                    "image": "aiu-feature-discovery",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "isRdma": false,
                  "loglevel": "info",
                  "metricsExporter": {
                    "enabled": true,
                    "image": "aiu-exporter",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "numVfs": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
          createdAt: "2024-10-07T13:07:51Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 2.1.1
      entries:
      - name: aiu-operator.v2.1.1
        version: 2.1.1
      name: stable-2.1
    - currentCSV: aiu-operator.v2.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "labels": {
                    "operator-version": "2.3.3"
                  },
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "aiu-device-plugin",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.1"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "featureDiscovery": {
                    "image": "aiu-feature-discovery",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "aiu-exporter",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "aiu-webhook-validator",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "scheduler": {
                    "image": "aiu-kube-scheduler-ocp-4.16",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.2"
                  }
                }
              },
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuNodeState",
                "metadata": {
                  "labels": {
                    "operator-version": "2.3.3"
                  },
                  "name": "aiunodestate"
                },
                "spec": {
                  "nodeName": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2025_03/aiu-operator@sha256:2ea8268e94d44581b3490cd4bbbd5525359073129bae42c7f007c05727afe1c3
          createdAt: "2025-04-03T01:19:38Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["aiunodestates.aiu.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu/release_2025_03/aiu-operator@sha256:2ea8268e94d44581b3490cd4bbbd5525359073129bae42c7f007c05727afe1c3
        version: 2.3.3
      entries:
      - name: aiu-operator.v2.3.3
        version: 2.3.3
      name: stable-v2.3
    - currentCSV: aiu-operator.v2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "aiuFilter": ".",
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "release_2025_05"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "aiu-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "e2e_stable",
                      "repository": "icr.io/ibmaiu_internal/x86_64/dd2",
                      "version": "release_2025_05"
                    },
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "aiu-exporter",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "aiu-webhook-validator",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "scheduler": {
                    "image": "aiu-kube-scheduler-ocp-4.16",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  }
                }
              },
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2025_05/aiu-operator@sha256:9796ab1c39d1dd55f8aa79c9f585be0f988b5026745184bdc890dfd1846909df
          createdAt: "2025-05-28T05:53:45Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["aiunodestates.aiu.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu/release_2025_05/aiu-operator@sha256:9796ab1c39d1dd55f8aa79c9f585be0f988b5026745184bdc890dfd1846909df
        - icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453
        version: 2.4.0
      entries:
      - name: aiu-operator.v2.4.0
        version: 2.4.0
      name: stable-v2.4
    defaultChannel: stable-v2.4
    packageName: aiu-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lifecycle-agent
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lifecycle-agent.v4.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "ImageBasedUpgrade",
                "metadata": {
                  "name": "upgrade"
                },
                "spec": {
                  "autoRollbackOnFailure": {},
                  "extraManifests": [
                    {
                      "name": "sno-extramanifests",
                      "namespace": "openshift-lifecycle-agent"
                    }
                  ],
                  "oadpContent": [
                    {
                      "name": "oadp-cm-sno-backup",
                      "namespace": "openshift-adp"
                    }
                  ],
                  "seedImageRef": {
                    "image": "quay.io/xyz",
                    "version": "4.16.0"
                  },
                  "stage": "Idle"
                }
              },
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "SeedGenerator",
                "metadata": {
                  "name": "seedimage"
                },
                "spec": {
                  "seedImage": "quay.io/xyz"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:6c603c89401823086d7a02d7ad221da44d8c84000ba63dd30675ae0cbcacc95f
          description: The Lifecycle Agent for OpenShift provides local lifecycle
            management services for Single Node Openshift (SNO) clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.3'
          operatorframework.io/suggested-namespace: openshift-lifecycle-agent
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-lifecycle-agent",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/lifecycle-agent
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ImageBasedUpgrade is the Schema for the ImageBasedUpgrades
              API
            displayName: Image-based Cluster Upgrade
            kind: ImageBasedUpgrade
            name: imagebasedupgrades.lca.openshift.io
            version: v1
          - description: SeedGenerator is the Schema for the seedgenerators API
            displayName: Seed Generator
            kind: SeedGenerator
            name: seedgenerators.lca.openshift.io
            version: v1
        description: "# Lifecycle Agent for OpenShift\nThe Lifecycle Agent for OpenShift
          provides local lifecycle management services \nfor Single Node Openshift
          (SNO) clusters.\n\n## Where to find more information\nYou can find additional
          guidance in the [agent repository](https://github.com/openshift-kni/lifecycle-agent).\n"
        displayName: Lifecycle Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - upgrade
        links:
        - name: Lifecycle Agent
          url: https://lifecycle-agent.domain
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift4/recert-rhel9@sha256:c91def5d3da4e777d2555c354e39bf021d9f499993f86614969dee5ecc00fc49
        - registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:6c603c89401823086d7a02d7ad221da44d8c84000ba63dd30675ae0cbcacc95f
        version: 4.16.3
      entries:
      - name: lifecycle-agent.v4.16.3
        version: 4.16.3
      - name: lifecycle-agent.v4.16.2
        version: 4.16.2
      - name: lifecycle-agent.v4.16.1
        version: 4.16.1
      - name: lifecycle-agent.v4.16.0
        version: 4.16.0
      name: "4.16"
    - currentCSV: lifecycle-agent.v4.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "ImageBasedUpgrade",
                "metadata": {
                  "name": "upgrade"
                },
                "spec": {
                  "autoRollbackOnFailure": {},
                  "extraManifests": [
                    {
                      "name": "sno-extramanifests",
                      "namespace": "openshift-lifecycle-agent"
                    }
                  ],
                  "oadpContent": [
                    {
                      "name": "oadp-cm-sno-backup",
                      "namespace": "openshift-adp"
                    }
                  ],
                  "seedImageRef": {
                    "image": "quay.io/xyz",
                    "version": "4.16.0"
                  },
                  "stage": "Idle"
                }
              },
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "SeedGenerator",
                "metadata": {
                  "name": "seedimage"
                },
                "spec": {
                  "seedImage": "quay.io/xyz"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:6c603c89401823086d7a02d7ad221da44d8c84000ba63dd30675ae0cbcacc95f
          description: The Lifecycle Agent for OpenShift provides local lifecycle
            management services for Single Node Openshift (SNO) clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.3'
          operatorframework.io/suggested-namespace: openshift-lifecycle-agent
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-lifecycle-agent",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/lifecycle-agent
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ImageBasedUpgrade is the Schema for the ImageBasedUpgrades
              API
            displayName: Image-based Cluster Upgrade
            kind: ImageBasedUpgrade
            name: imagebasedupgrades.lca.openshift.io
            version: v1
          - description: SeedGenerator is the Schema for the seedgenerators API
            displayName: Seed Generator
            kind: SeedGenerator
            name: seedgenerators.lca.openshift.io
            version: v1
        description: "# Lifecycle Agent for OpenShift\nThe Lifecycle Agent for OpenShift
          provides local lifecycle management services \nfor Single Node Openshift
          (SNO) clusters.\n\n## Where to find more information\nYou can find additional
          guidance in the [agent repository](https://github.com/openshift-kni/lifecycle-agent).\n"
        displayName: Lifecycle Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - upgrade
        links:
        - name: Lifecycle Agent
          url: https://lifecycle-agent.domain
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:6c603c89401823086d7a02d7ad221da44d8c84000ba63dd30675ae0cbcacc95f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift4/recert-rhel9@sha256:c91def5d3da4e777d2555c354e39bf021d9f499993f86614969dee5ecc00fc49
        version: 4.16.3
      entries:
      - name: lifecycle-agent.v4.16.3
        version: 4.16.3
      - name: lifecycle-agent.v4.16.2
        version: 4.16.2
      - name: lifecycle-agent.v4.16.1
        version: 4.16.1
      - name: lifecycle-agent.v4.16.0
        version: 4.16.0
      name: stable
    defaultChannel: stable
    packageName: lifecycle-agent
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-opensearchservice-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-opensearchservice-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opensearchservice.services.k8s.aws/v1alpha1",
                "kind": "Domain",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/opensearchservice-controller:1.1.1
          createdAt: "2025-09-19T18:05:50Z"
          description: AWS OpenSearch Service controller is a service controller for
            managing OpenSearch Service resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Domain represents the state of an AWS opensearchservice Domain
              resource.
            displayName: Domain
            kind: Domain
            name: domains.opensearchservice.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage OpenSearch Service resources in AWS from within your Kubernetes cluster.

          **About Amazon OpenSearch Service**

          Amazon OpenSearch Service (successor to Amazon Elasticsearch Service) is a managed service that makes it easy to deploy, operate, and scale OpenSearch clusters in the AWS Cloud. Amazon OpenSearch Service supports OpenSearch and legacy Elasticsearch OSS. When you create a cluster, you have the option of which search engine to use. OpenSearch Service offers broad compatibility with Elasticsearch OSS 7.10, the final open source version of the software. For information about what changed with the recent service renaming, and the actions you might need to take, see [Amazon OpenSearch Service - Summary of changes](https://docs.aws.amazon.com/opensearch-service/latest/developerguide/rename.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon OpenSearch Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opensearchservice
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon OpenSearch Service Developer Resources
          url: https://aws.amazon.com/opensearch-service/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: opensearch service maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/opensearchservice-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-opensearchservice-controller.v1.1.1
        version: 1.1.1
      - name: ack-opensearchservice-controller.v1.1.0
        version: 1.1.0
      - name: ack-opensearchservice-controller.v1.0.15
        version: 1.0.15
      - name: ack-opensearchservice-controller.v1.0.12
        version: 1.0.12
      - name: ack-opensearchservice-controller.v1.0.10
        version: 1.0.10
      - name: ack-opensearchservice-controller.v1.0.9
        version: 1.0.9
      - name: ack-opensearchservice-controller.v1.0.8
        version: 1.0.8
      - name: ack-opensearchservice-controller.v1.0.7
        version: 1.0.7
      - name: ack-opensearchservice-controller.v1.0.6
        version: 1.0.6
      - name: ack-opensearchservice-controller.v1.0.5
        version: 1.0.5
      - name: ack-opensearchservice-controller.v1.0.4
        version: 1.0.4
      - name: ack-opensearchservice-controller.v1.0.2
        version: 1.0.2
      - name: ack-opensearchservice-controller.v1.0.1
        version: 1.0.1
      - name: ack-opensearchservice-controller.v1.0.0
        version: 1.0.0
      - name: ack-opensearchservice-controller.v0.0.28
        version: 0.0.28
      - name: ack-opensearchservice-controller.v0.0.27
        version: 0.0.27
      - name: ack-opensearchservice-controller.v0.0.26
        version: 0.0.26
      - name: ack-opensearchservice-controller.v0.0.24
        version: 0.0.24
      - name: ack-opensearchservice-controller.v0.0.23
        version: 0.0.23
      - name: ack-opensearchservice-controller.v0.0.19
        version: 0.0.19
      - name: ack-opensearchservice-controller.v0.0.17
        version: 0.0.17
      - name: ack-opensearchservice-controller.v0.0.15
        version: 0.0.15
      - name: ack-opensearchservice-controller.v0.0.14
        version: 0.0.14
      - name: ack-opensearchservice-controller.v0.0.13
        version: 0.0.13
      - name: ack-opensearchservice-controller.v0.0.12
        version: 0.0.12
      - name: ack-opensearchservice-controller.v0.0.11
        version: 0.0.11
      - name: ack-opensearchservice-controller.v0.0.10
        version: 0.0.10
      - name: ack-opensearchservice-controller.v0.0.9
        version: 0.0.9
      - name: ack-opensearchservice-controller.v0.0.8
        version: 0.0.8
      - name: ack-opensearchservice-controller.v0.0.7
        version: 0.0.7
      - name: ack-opensearchservice-controller.v0.0.6
        version: 0.0.6
      - name: ack-opensearchservice-controller.v0.0.5
        version: 0.0.5
      - name: ack-opensearchservice-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-opensearchservice-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
    name: flink-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flink-kubernetes-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"flink.apache.org/v1beta1",
                "kind":"FlinkDeployment",
                "metadata":{
                    "name":"basic-example"
                },
                "spec":{
                    "image":"flink:1.16",
                    "flinkVersion":"v1_16",
                    "flinkConfiguration":{
                        "taskmanager.numberOfTaskSlots":"2"
                    },
                    "serviceAccount":"flink",
                    "jobManager":{
                        "resource":{
                            "memory":"2048m",
                            "cpu":1
                        }
                    },
                    "taskManager":{
                        "resource":{
                            "memory":"2048m",
                            "cpu":1
                        }
                    },
                    "job":{
                        "jarURI":"local:///opt/flink/examples/streaming/StateMachineExample.jar",
                        "parallelism":2,
                        "upgradeMode":"stateless"
                    }
                }
              },
              {
                "apiVersion":"flink.apache.org/v1beta1",
                "kind":"FlinkSessionJob",
                "metadata":{
                  "name":"basic-session-job-example2"
                },
                "spec":{
                  "deploymentName":"basic-session-cluster",
                  "job":{
                      "jarURI":"https://repo1.maven.org/maven2/org/apache/flink/flink-examples-streaming_2.12/1.16.1/flink-examples-streaming_2.12-1.16.1.jar",
                      "parallelism":2,
                      "upgradeMode":"stateless",
                      "entryClass":"org.apache.flink.streaming.examples.statemachine.StateMachineExample"
                  }
                }
              },
              {
                "apiVersion": "flink.apache.org/v1beta1",
                "kind": "FlinkStateSnapshot",
                "metadata": {
                  "name": "example-savepoint"
                },
                "spec": {
                  "savepoint": {
                    "alreadyExists": true,
                    "disposeOnDelete": false,
                    "path": "/flink-data/savepoints/savepoint-98682a-21931bf10c39"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "basic-example": {
                "description": "Example flink deployment"
              },
              "basic-session-job-example2": {
                "description": "Example flink session job"
              }
            }
          capabilities: Deep Insights
          categories: Big Data, Streaming & Messaging
          containerImage: ghcr.io/apache/flink-kubernetes-operator:1.12.0
          createdAt: "2025-06-23 18:42:54"
          description: Flink Kubernetes Operator manages the complete deployment lifecycle
            of Apache Flink applications.
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/apache/flink-kubernetes-operator.git
          support: Apache Software Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The FlinkDeployment CR defines Flink Application and Session
              cluster deployments
            displayName: Flink Deployment
            kind: FlinkDeployment
            name: flinkdeployments.flink.apache.org
            version: v1beta1
          - description: The FlinkSessionJob CR defines the session job on the Session
              cluster and each Session cluster can run multiple FlinkSessionJob.
            displayName: Flink Session Job
            kind: FlinkSessionJob
            name: flinksessionjobs.flink.apache.org
            version: v1beta1
          - description: The FlinkStateSnapshots CR is used to describe savepoint
              or checkpoint for a Flink job.
            displayName: Flink State Snapshot
            kind: FlinkStateSnapshot
            name: flinkstatesnapshots.flink.apache.org
            version: v1beta1
        description: |-
          Flink Kubernetes Operator acts as a control plane to manage the complete deployment lifecycle of Apache Flink applications.
          ## Installation
          This operator can be installed from [OperatorHub.io](https://operatorhub.io/operator/flink-kubernetes-operator).

          By default, the Flink operator monitors and run Flink applications in the same namespace as the operator.
          To run Flink jobs in another namespace, users are responsible for setting up the RBAC in that namespace.

          For example, users need to do the following for each additional namespace that runs the Flink applications:

          1. Switch to the namespace by running:

              ```sh
              kubectl config set-context --current --namespace=CHANGEIT
              ```

          2. Create the service account, role, and role binding in the namespace using the commands below:

              ```sh
              kubectl apply -f - <<EOF
              apiVersion: v1
              kind: ServiceAccount
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink
              ---
              apiVersion: rbac.authorization.k8s.io/v1
              kind: Role
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink
              rules:
              - apiGroups:
                - ""
                resources:
                - pods
                - configmaps
                verbs:
                - '*'
              - apiGroups:
                - apps
                resources:
                - deployments
                - deployments/finalizers
                verbs:
                - '*'
              ---
              apiVersion: rbac.authorization.k8s.io/v1
              kind: RoleBinding
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink-role-binding
              roleRef:
                apiGroup: rbac.authorization.k8s.io
                kind: Role
                name: flink
              subjects:
              - kind: ServiceAccount
                name: flink
              EOF
              ```

          3. Proceed deploying Flink applications using the the Custom Resource Definition below.

              ```sh
              kubectl create -f https://raw.githubusercontent.com/apache/flink-kubernetes-operator/release-1.12.0/examples/basic.yaml
              ```

          See [Flink custom resources](https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/custom-resource/overview/#overview) for more detail.
        displayName: Flink Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flink
        - streaming
        links:
        - name: Website
          url: https://flink.apache.org/
        - name: Documentation
          url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/
        - name: Mailing list
          url: https://lists.apache.org/list.html?dev@flink.apache.org
        - name: Slack
          url: https://apache-flink.slack.com/join/shared_invite/zt-1llkzbgyt-K2nNGGg88rfsDGLkT09Qzg#/shared-invite/email
        - name: GitHub
          url: https://github.com/apache/flink-kubernetes-operator
        maintainers:
        - email: mbalassi@apache.org
          name: Marton Balassi
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: jbusche@us.ibm.com
          name: James Busche
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Community
          url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
        relatedImages:
        - ghcr.io/apache/flink-kubernetes-operator:1.12.0
        version: 1.12.0
      entries:
      - name: flink-kubernetes-operator.v1.12.0
        version: 1.12.0
      - name: flink-kubernetes-operator.v1.11.0
        version: 1.11.0
      - name: flink-kubernetes-operator.v1.10.0
        version: 1.10.0
      - name: flink-kubernetes-operator.v1.6.0
        version: 1.6.0
      - name: flink-kubernetes-operator.v1.5.0
        version: 1.5.0
      - name: flink-kubernetes-operator.v1.4.0
        version: 1.4.0
      - name: flink-kubernetes-operator.v1.3.1
        version: 1.3.1
      - name: flink-kubernetes-operator.v1.3.0
        version: 1.3.0
      name: alpha
    defaultChannel: alpha
    packageName: flink-kubernetes-operator
    provider:
      name: Community
      url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeRocketCI
      provider-url: https://docs.kuberocketci.io/
    name: edp-keycloak-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: edp-keycloak-operator.v1.19.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "keycloak-sample"
                },
                "spec": {
                  "secret": "my-keycloak-secret",
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakAuthFlow",
                "metadata": {
                  "name": "keycloakauthflow-sample"
                },
                "spec": {
                  "alias": "MyBrowser",
                  "authenticationExecutions": [
                    {
                      "authenticator": "auth-cookie",
                      "priority": 0,
                      "requirement": "ALTERNATIVE"
                    },
                    {
                      "authenticator": "identity-provider-redirector",
                      "authenticatorConfig": {
                        "alias": "my-alias",
                        "config": {
                          "defaultProvider": "my-alias"
                        }
                      },
                      "priority": 1,
                      "requirement": "REQUIRED"
                    }
                  ],
                  "builtIn": false,
                  "description": "browser with idp",
                  "providerId": "basic-flow",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "topLevel": true
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-sample"
                },
                "spec": {
                  "advancedProtocolMappers": true,
                  "clientId": "agocd",
                  "defaultClientScopes": [
                    "argocd_groups"
                  ],
                  "directAccess": true,
                  "public": false,
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-name:client-secret-key",
                  "webUrl": "https://argocd.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClientScope",
                "metadata": {
                  "name": "keycloakclientscope-sample"
                },
                "spec": {
                  "description": "Group Membership",
                  "name": "groups",
                  "protocol": "openid-connect",
                  "protocolMappers": [
                    {
                      "config": {
                        "access.token.claim": "true",
                        "claim.name": "groups",
                        "full.path": "false",
                        "id.token.claim": "true",
                        "userinfo.token.claim": "true"
                      },
                      "name": "groups",
                      "protocol": "openid-connect",
                      "protocolMapper": "oidc-group-membership-mapper"
                    }
                  ],
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "keycloakrealm-sample"
                },
                "spec": {
                  "id": "d1-id-kc-realm-name",
                  "keycloakRef": {
                    "kind": "Keycloak",
                    "name": "keycloak-sample"
                  },
                  "passwordPolicy": [
                    {
                      "type": "forceExpiredPasswordChange",
                      "value": "365"
                    },
                    {
                      "type": "length",
                      "value": "8"
                    }
                  ],
                  "realmEventConfig": {
                    "adminEventsDetailsEnabled": false,
                    "adminEventsEnabled": true,
                    "enabledEventTypes": [
                      "UPDATE_CONSENT_ERROR",
                      "CLIENT_LOGIN"
                    ],
                    "eventsEnabled": true,
                    "eventsExpiration": 15000,
                    "eventsListeners": [
                      "jboss-logging"
                    ]
                  },
                  "realmName": "d2-id-kc-realm-name",
                  "ssoAutoRedirectEnabled": false,
                  "ssoRealmEnabled": false,
                  "ssoRealmName": "openshift"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmComponent",
                "metadata": {
                  "name": "keycloakrealmcomponent-sample"
                },
                "spec": {
                  "config": {
                    "allowPasswordAuthentication": [
                      "true"
                    ],
                    "cachePolicy": [
                      "EVICT_WEEKLY"
                    ],
                    "debug": [
                      "true"
                    ],
                    "editMode": [
                      "READ_ONLY"
                    ],
                    "enabled": [
                      "true"
                    ],
                    "evictionDay": [
                      "3"
                    ],
                    "evictionHour": [
                      "5"
                    ],
                    "evictionMinute": [
                      "7"
                    ],
                    "kerberosRealm": [
                      "test-realm"
                    ],
                    "keyTab": [
                      "test-key-tab"
                    ],
                    "priority": [
                      "0"
                    ],
                    "serverPrincipal": [
                      "srv-principal-test"
                    ],
                    "updateProfileFirstLogin": [
                      "true"
                    ]
                  },
                  "name": "cr-kerb-test",
                  "providerId": "kerberos",
                  "providerType": "org.keycloak.storage.UserStorageProvider",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmGroup",
                "metadata": {
                  "name": "keycloakrealmgroup-sample"
                },
                "spec": {
                  "name": "ArgoCDAdmins",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "keycloakrealmidentityprovider-sample"
                },
                "spec": {
                  "alias": "instagram",
                  "authenticateByDefault": false,
                  "config": {
                    "clientId": "foo",
                    "clientSecret": "$secretName:secretKey",
                    "hideOnLoginPage": "true",
                    "syncMode": "IMPORT",
                    "useJwksUrl": "true"
                  },
                  "enabled": true,
                  "firstBrokerLoginFlowAlias": "first broker login",
                  "mappers": [
                    {
                      "config": {
                        "role": "role-tr",
                        "syncMode": "INHERIT"
                      },
                      "identityProviderAlias": "instagram",
                      "identityProviderMapper": "oidc-hardcoded-role-idp-mapper",
                      "name": "test3212"
                    },
                    {
                      "config": {
                        "attribute": "foo",
                        "attribute.value": "bar",
                        "syncMode": "IMPORT"
                      },
                      "identityProviderAlias": "instagram",
                      "identityProviderMapper": "hardcoded-attribute-idp-mapper",
                      "name": "test-33221"
                    }
                  ],
                  "providerId": "instagram",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRole",
                "metadata": {
                  "name": "keycloakrealmrole-sample"
                },
                "spec": {
                  "composite": true,
                  "description": "default developer role",
                  "name": "developer",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRoleBatch",
                "metadata": {
                  "name": "keycloakrealmrolebatch-sample"
                },
                "spec": {
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "roles": [
                    {
                      "composite": true,
                      "description": "default developer role",
                      "isDefault": false,
                      "name": "developer"
                    },
                    {
                      "composite": true,
                      "description": "default administrator role",
                      "isDefault": false,
                      "name": "administrator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmUser",
                "metadata": {
                  "name": "keycloakrealmuser-sample"
                },
                "spec": {
                  "attributes": {
                    "baz": "jazz",
                    "foo": "bar"
                  },
                  "email": "john.snow13@example.com",
                  "emailVerified": true,
                  "enabled": true,
                  "firstName": "John",
                  "keepResource": true,
                  "lastName": "Snow",
                  "password": "12345678",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "requiredUserActions": [
                    "UPDATE_PASSWORD"
                  ],
                  "username": "john.snow13"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloak",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloak-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloak",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloak-sample"
                },
                "spec": {
                  "secret": "keycloak-access",
                  "url": "https://keycloak.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloakRealm",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloakrealm-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloakrealm",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloakrealm-sample"
                },
                "spec": {
                  "clusterKeycloakRef": "clusterkeycloak-sample",
                  "realmName": "realm-sample"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          containerImage: docker.io/epamedp/keycloak-operator:1.19.0
          createdAt: "2023-11-15T16:04:45Z"
          description: An Operator for managing Keycloak
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-keycloak-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterKeycloakRealm is the Schema for the clusterkeycloakrealms
              API.
            displayName: Cluster Keycloak Realm
            kind: ClusterKeycloakRealm
            name: clusterkeycloakrealms.v1.edp.epam.com
            version: v1alpha1
          - description: ClusterKeycloak is the Schema for the clusterkeycloaks API.
            displayName: Cluster Keycloak
            kind: ClusterKeycloak
            name: clusterkeycloaks.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakAuthFlow is the Schema for the keycloak authentication
              flow API.
            displayName: Keycloak Auth Flow
            kind: KeycloakAuthFlow
            name: keycloakauthflows.v1.edp.epam.com
            version: v1
          - description: KeycloakClient is the Schema for the keycloak clients API.
            displayName: Keycloak Client
            kind: KeycloakClient
            name: keycloakclients.v1.edp.epam.com
            version: v1
          - description: KeycloakClientScope is the Schema for the keycloakclientscopes
              API.
            displayName: Keycloak Client Scope
            kind: KeycloakClientScope
            name: keycloakclientscopes.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmComponent is the Schema for the keycloak component
              API.
            displayName: Keycloak Realm Component
            kind: KeycloakRealmComponent
            name: keycloakrealmcomponents.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmGroup is the Schema for the keycloak group API.
            displayName: Keycloak Realm Group
            kind: KeycloakRealmGroup
            name: keycloakrealmgroups.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmIdentityProvider is the Schema for the keycloak
              realm identity provider API.
            displayName: Keycloak Realm Identity Provider
            kind: KeycloakRealmIdentityProvider
            name: keycloakrealmidentityproviders.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRoleBatch is the Schema for the keycloak roles
              API.
            displayName: Keycloak Realm Role Batch
            kind: KeycloakRealmRoleBatch
            name: keycloakrealmrolebatches.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRole is the Schema for the keycloak group API.
            displayName: Keycloak Realm Role
            kind: KeycloakRealmRole
            name: keycloakrealmroles.v1.edp.epam.com
            version: v1
          - description: KeycloakRealm is the Schema for the keycloak realms API.
            displayName: Keycloak Realm
            kind: KeycloakRealm
            name: keycloakrealms.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmUser is the Schema for the keycloak user API.
            displayName: Keycloak Realm User
            kind: KeycloakRealmUser
            name: keycloakrealmusers.v1.edp.epam.com
            version: v1
          - description: Keycloak is the Schema for the keycloaks API.
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.v1.edp.epam.com
            version: v1
        description: |
          Keycloak Operator is an operator that is responsible for establishing
          a connection to provided Keycloak Server, reconciling Keycloak entities (realms,
          roles, groups, users, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Keycloak `Master` realm, and assign a `create-realm` role.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  keycloak-access
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Keycloak` with Keycloak instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: Keycloak
              metadata:
                name: keycloak-sample
              spec:
                secret: keycloak-access             # Secret name
                url: https://keycloak.example.com   # Keycloak URL
              ```

              Wait for the `.status` field with  `status.connected: true`

          4. Create Keycloak realm and group using Custom Resources:

            ```yaml
            apiVersion: v1.edp.epam.com/v1
            kind: KeycloakRealm
            metadata:
              name: keycloakrealm-sample
            spec:
              realmName: realm-sample
              keycloakOwner: keycloak-sample   # the name of `kind: Keycloak`
              ```

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: KeycloakRealmGroup
              metadata:
                name: argocd-admins
              spec:
                name: ArgoCDAdmins
                realm: keycloakrealm-sample   # the name of `kind: KeycloakRealm`
              ```
        displayName: EDP Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - authentication
        - authorization
        - edp
        - idp
        - keycloak
        - oauth
        - oidc
        - operator
        - saml
        - sso
        links:
        - name: Edp Keycloak Operator
          url: https://github.com/epam/edp-keycloak-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/keycloak-operator:1.19.0
        version: 1.19.0
      entries:
      - name: edp-keycloak-operator.v1.19.0
        version: 1.19.0
      - name: edp-keycloak-operator.v1.18.2
        version: 1.18.2
      - name: edp-keycloak-operator.v1.17.1
        version: 1.17.1
      - name: edp-keycloak-operator.v1.13.0
        version: 1.13.0
      name: alpha
    - currentCSV: edp-keycloak-operator.v1.29.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "keycloak-sample"
                },
                "spec": {
                  "secret": "my-keycloak-secret",
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakAuthFlow",
                "metadata": {
                  "name": "keycloakauthflow-sample"
                },
                "spec": {
                  "alias": "MyBrowser",
                  "authenticationExecutions": [
                    {
                      "authenticator": "auth-cookie",
                      "priority": 0,
                      "requirement": "ALTERNATIVE"
                    },
                    {
                      "authenticator": "identity-provider-redirector",
                      "authenticatorConfig": {
                        "alias": "my-alias",
                        "config": {
                          "defaultProvider": "my-alias"
                        }
                      },
                      "priority": 1,
                      "requirement": "REQUIRED"
                    }
                  ],
                  "builtIn": false,
                  "description": "browser with idp",
                  "providerId": "basic-flow",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "topLevel": true
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-authorization-sample"
                },
                "spec": {
                  "adminFineGrainedPermissionsEnabled": true,
                  "authorization": {
                    "permissions": [
                      {
                        "decisionStrategy": "AFFIRMATIVE",
                        "description": "Resource permission",
                        "logic": "POSITIVE",
                        "name": "resource-permission",
                        "policies": [
                          "role-policy"
                        ],
                        "resources": [
                          "resource1"
                        ],
                        "type": "resource"
                      },
                      {
                        "decisionStrategy": "CONSENSUS",
                        "description": "Scope permission",
                        "logic": "POSITIVE",
                        "name": "scope-permission",
                        "policies": [
                          "role-policy"
                        ],
                        "scopes": [
                          "scope1"
                        ],
                        "type": "scope"
                      }
                    ],
                    "policies": [
                      {
                        "decisionStrategy": "AFFIRMATIVE",
                        "description": "Role policy",
                        "logic": "POSITIVE",
                        "name": "role-policy",
                        "rolePolicy": {
                          "roles": [
                            {
                              "name": "developer",
                              "required": true
                            }
                          ]
                        },
                        "type": "role"
                      },
                      {
                        "aggregatedPolicy": {
                          "policies": [
                            "policy1",
                            "policy2"
                          ]
                        },
                        "description": "Aggregate policy",
                        "name": "aggregate-policy",
                        "type": "aggregate"
                      },
                      {
                        "clientPolicy": {
                          "clients": [
                            "client1",
                            "client2"
                          ]
                        },
                        "description": "Client policy",
                        "name": "client-policy",
                        "type": "client"
                      },
                      {
                        "description": "Group policy",
                        "groupPolicy": {
                          "groups": [
                            {
                              "extendChildren": true,
                              "name": "group1"
                            }
                          ]
                        },
                        "name": "group-policy",
                        "type": "group"
                      },
                      {
                        "description": "Role policy",
                        "name": "role-policy",
                        "rolePolicy": {
                          "roles": [
                            {
                              "name": "developer",
                              "required": true
                            }
                          ]
                        },
                        "type": "role"
                      },
                      {
                        "description": "Time policy",
                        "name": "time-policy",
                        "timePolicy": {
                          "notBefore": "2021-01-01T00:00:00Z",
                          "notOnOrAfter": "2021-12-31T23:59:59Z"
                        },
                        "type": "time"
                      },
                      {
                        "description": "User policy",
                        "name": "user-policy",
                        "type": "user",
                        "userPolicy": {
                          "users": [
                            "user1",
                            "user2"
                          ]
                        }
                      }
                    ],
                    "resources": [
                      {
                        "displayName": "Resource 1",
                        "iconUri": "https://example.com/icon.png",
                        "name": "resource1",
                        "scopes": [
                          "scope1"
                        ],
                        "type": "test"
                      }
                    ],
                    "scopes": [
                      "scope1"
                    ]
                  },
                  "authorizationServicesEnabled": true,
                  "clientId": "authorization-sample",
                  "directAccess": true,
                  "permission": {
                    "scopePermissions": [
                      {
                        "name": "token-exchange",
                        "policies": [
                          "policy1"
                        ]
                      }
                    ]
                  },
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-authorization-sample:client-secret-key",
                  "serviceAccount": {
                    "enabled": true
                  },
                  "webUrl": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-sample"
                },
                "spec": {
                  "advancedProtocolMappers": true,
                  "authenticationFlowBindingOverrides": {
                    "browser": "browser",
                    "directGrant": "direct grant"
                  },
                  "clientId": "agocd",
                  "defaultClientScopes": [
                    "argocd_groups"
                  ],
                  "directAccess": true,
                  "public": false,
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-name:client-secret-key",
                  "webUrl": "https://argocd.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClientScope",
                "metadata": {
                  "name": "keycloakclientscope-sample"
                },
                "spec": {
                  "description": "Group Membership",
                  "name": "groups",
                  "protocol": "openid-connect",
                  "protocolMappers": [
                    {
                      "config": {
                        "access.token.claim": "true",
                        "claim.name": "groups",
                        "full.path": "false",
                        "id.token.claim": "true",
                        "userinfo.token.claim": "true"
                      },
                      "name": "groups",
                      "protocol": "openid-connect",
                      "protocolMapper": "oidc-group-membership-mapper"
                    }
                  ],
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "keycloakrealm-sample"
                },
                "spec": {
                  "id": "d1-id-kc-realm-name",
                  "keycloakRef": {
                    "kind": "Keycloak",
                    "name": "keycloak-sample"
                  },
                  "organizationsEnabled": true,
                  "passwordPolicy": [
                    {
                      "type": "forceExpiredPasswordChange",
                      "value": "365"
                    },
                    {
                      "type": "length",
                      "value": "8"
                    }
                  ],
                  "realmEventConfig": {
                    "adminEventsDetailsEnabled": false,
                    "adminEventsEnabled": true,
                    "enabledEventTypes": [
                      "UPDATE_CONSENT_ERROR",
                      "CLIENT_LOGIN"
                    ],
                    "eventsEnabled": true,
                    "eventsExpiration": 15000,
                    "eventsListeners": [
                      "jboss-logging"
                    ]
                  },
                  "realmName": "d2-id-kc-realm-name",
                  "userProfileConfig": {
                    "attributes": [
                      {
                        "annotations": {
                          "inputType": "text"
                        },
                        "displayName": "Test Attribute",
                        "group": "test-group",
                        "multivalued": true,
                        "name": "test-attribute",
                        "permissions": {
                          "edit": [
                            "admin"
                          ],
                          "view": [
                            "admin",
                            "user"
                          ]
                        },
                        "required": {
                          "roles": [
                            "admin"
                          ],
                          "scopes": [
                            "profile"
                          ]
                        },
                        "selector": {
                          "scopes": [
                            "profile"
                          ]
                        },
                        "validations": {
                          "email": {
                            "max-local-length": {
                              "intVal": 64
                            }
                          },
                          "local-date": {},
                          "multivalued": {
                            "max": {
                              "stringVal": "10"
                            },
                            "min": {
                              "stringVal": "1"
                            }
                          },
                          "options": {
                            "options": {
                              "sliceVal": [
                                "option1",
                                "option2"
                              ]
                            }
                          }
                        }
                      }
                    ],
                    "groups": [
                      {
                        "annotations": {
                          "groupAnnotation": "groupAnnotation"
                        },
                        "displayDescription": "Test Group",
                        "displayHeader": "Test Group",
                        "name": "test-group"
                      }
                    ],
                    "unmanagedAttributePolicy": "ENABLED"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmComponent",
                "metadata": {
                  "name": "keycloakrealmcomponent-sample"
                },
                "spec": {
                  "config": {
                    "allowPasswordAuthentication": [
                      "true"
                    ],
                    "cachePolicy": [
                      "EVICT_WEEKLY"
                    ],
                    "debug": [
                      "true"
                    ],
                    "editMode": [
                      "READ_ONLY"
                    ],
                    "enabled": [
                      "true"
                    ],
                    "evictionDay": [
                      "3"
                    ],
                    "evictionHour": [
                      "5"
                    ],
                    "evictionMinute": [
                      "7"
                    ],
                    "kerberosRealm": [
                      "test-realm"
                    ],
                    "keyTab": [
                      "test-key-tab"
                    ],
                    "priority": [
                      "0"
                    ],
                    "serverPrincipal": [
                      "srv-principal-test"
                    ],
                    "updateProfileFirstLogin": [
                      "true"
                    ]
                  },
                  "name": "cr-kerb-test",
                  "providerId": "kerberos",
                  "providerType": "org.keycloak.storage.UserStorageProvider",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmGroup",
                "metadata": {
                  "name": "keycloakrealmgroup-sample"
                },
                "spec": {
                  "name": "ArgoCDAdmins",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "test-org-idp",
                  "namespace": "default"
                },
                "spec": {
                  "alias": "test-org-idp",
                  "config": {
                    "clientId": "test-org-client-id",
                    "clientSecret": "test-org-client-secret",
                    "kc.org.broker.redirect.mode.email-matches": "true",
                    "kc.org.domain": "example.com"
                  },
                  "enabled": true,
                  "providerId": "github",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "test-org-realm"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "keycloakrealmidentityprovider-sample"
                },
                "spec": {
                  "alias": "github",
                  "authenticateByDefault": false,
                  "config": {
                    "clientId": "foo",
                    "clientSecret": "$secretName:secretKey",
                    "syncMode": "IMPORT",
                    "useJwksUrl": "true"
                  },
                  "enabled": true,
                  "firstBrokerLoginFlowAlias": "first broker login",
                  "mappers": [
                    {
                      "config": {
                        "role": "role-tr",
                        "syncMode": "INHERIT"
                      },
                      "identityProviderAlias": "github",
                      "identityProviderMapper": "oidc-hardcoded-role-idp-mapper",
                      "name": "test3212"
                    },
                    {
                      "config": {
                        "attribute": "foo",
                        "attribute.value": "bar",
                        "syncMode": "IMPORT"
                      },
                      "identityProviderAlias": "github",
                      "identityProviderMapper": "hardcoded-attribute-idp-mapper",
                      "name": "test-33221"
                    }
                  ],
                  "providerId": "github",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRole",
                "metadata": {
                  "name": "keycloakrealmrole-sample"
                },
                "spec": {
                  "composite": true,
                  "description": "default developer role",
                  "name": "developer",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRoleBatch",
                "metadata": {
                  "name": "keycloakrealmrolebatch-sample"
                },
                "spec": {
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "roles": [
                    {
                      "composite": true,
                      "description": "default developer role",
                      "isDefault": false,
                      "name": "developer"
                    },
                    {
                      "composite": true,
                      "description": "default administrator role",
                      "isDefault": false,
                      "name": "administrator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmUser",
                "metadata": {
                  "name": "keycloakrealmuser-sample"
                },
                "spec": {
                  "attributes": {
                    "baz": "jazz",
                    "foo": "bar"
                  },
                  "email": "john.snow13@example.com",
                  "emailVerified": true,
                  "enabled": true,
                  "firstName": "John",
                  "keepResource": true,
                  "lastName": "Snow",
                  "password": "12345678",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "requiredUserActions": [
                    "UPDATE_PASSWORD"
                  ],
                  "username": "john.snow13"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloak",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloak-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloak",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloak-sample"
                },
                "spec": {
                  "secret": "keycloak-access",
                  "url": "https://keycloak.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloakRealm",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloakrealm-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloakrealm",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloakrealm-sample"
                },
                "spec": {
                  "authenticationFlows": {
                    "browserFlow": "browserFlow-sample"
                  },
                  "clusterKeycloakRef": "clusterkeycloak-sample",
                  "organizationsEnabled": false,
                  "realmName": "realm-sample"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "KeycloakOrganization",
                "metadata": {
                  "name": "test-keycloak-organization",
                  "namespace": "default"
                },
                "spec": {
                  "alias": "test-org",
                  "attributes": {
                    "department": [
                      "engineering",
                      "qa"
                    ],
                    "location": [
                      "us-east"
                    ]
                  },
                  "description": "Test organization",
                  "domains": [
                    "example.com",
                    "test.com"
                  ],
                  "identityProviders": [
                    {
                      "alias": "test-org-idp"
                    }
                  ],
                  "name": "Test Organization",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "test-org-realm"
                  },
                  "redirectUrl": "https://example.com/redirect"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          containerImage: docker.io/epamedp/keycloak-operator:1.29.0
          createdAt: "2025-09-26T12:16:55Z"
          description: An Operator for managing Keycloak
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/epam/edp-keycloak-operator
          support: KubeRocketCI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterKeycloakRealm is the Schema for the clusterkeycloakrealms
              API.
            displayName: Cluster Keycloak Realm
            kind: ClusterKeycloakRealm
            name: clusterkeycloakrealms.v1.edp.epam.com
            version: v1alpha1
          - description: ClusterKeycloak is the Schema for the clusterkeycloaks API.
            displayName: Cluster Keycloak
            kind: ClusterKeycloak
            name: clusterkeycloaks.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakAuthFlow is the Schema for the keycloak authentication
              flow API.
            displayName: Keycloak Auth Flow
            kind: KeycloakAuthFlow
            name: keycloakauthflows.v1.edp.epam.com
            version: v1
          - description: KeycloakClient is the Schema for the keycloak clients API.
            displayName: Keycloak Client
            kind: KeycloakClient
            name: keycloakclients.v1.edp.epam.com
            version: v1
          - description: KeycloakClientScope is the Schema for the keycloakclientscopes
              API.
            displayName: Keycloak Client Scope
            kind: KeycloakClientScope
            name: keycloakclientscopes.v1.edp.epam.com
            version: v1
          - description: KeycloakOrganization is the Schema for the organizations
              API.
            displayName: Keycloak Organization
            kind: KeycloakOrganization
            name: keycloakorganizations.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakRealmComponent is the Schema for the keycloak component
              API.
            displayName: Keycloak Realm Component
            kind: KeycloakRealmComponent
            name: keycloakrealmcomponents.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmGroup is the Schema for the keycloak group API.
            displayName: Keycloak Realm Group
            kind: KeycloakRealmGroup
            name: keycloakrealmgroups.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmIdentityProvider is the Schema for the keycloak
              realm identity provider API.
            displayName: Keycloak Realm Identity Provider
            kind: KeycloakRealmIdentityProvider
            name: keycloakrealmidentityproviders.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRoleBatch is the Schema for the keycloak roles
              API.
            displayName: Keycloak Realm Role Batch
            kind: KeycloakRealmRoleBatch
            name: keycloakrealmrolebatches.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRole is the Schema for the keycloak group API.
            displayName: Keycloak Realm Role
            kind: KeycloakRealmRole
            name: keycloakrealmroles.v1.edp.epam.com
            version: v1
          - description: KeycloakRealm is the Schema for the keycloak realms API.
            displayName: Keycloak Realm
            kind: KeycloakRealm
            name: keycloakrealms.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmUser is the Schema for the keycloak user API.
            displayName: Keycloak Realm User
            kind: KeycloakRealmUser
            name: keycloakrealmusers.v1.edp.epam.com
            version: v1
          - description: Keycloak is the Schema for the keycloaks API.
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.v1.edp.epam.com
            version: v1
        description: |
          Keycloak Operator is an operator that is responsible for establishing
          a connection to provided Keycloak Server, reconciling Keycloak entities (realms,
          roles, groups, users, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Keycloak `Master` realm, and assign a `create-realm` role.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  keycloak-access
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Keycloak` with Keycloak instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: Keycloak
              metadata:
                name: keycloak-sample
              spec:
                secret: keycloak-access             # Secret name
                url: https://keycloak.example.com   # Keycloak URL
              ```

              Wait for the `.status` field with  `status.connected: true`

          4. Create Keycloak realm and group using Custom Resources:

            ```yaml
            apiVersion: v1.edp.epam.com/v1
            kind: KeycloakRealm
            metadata:
              name: keycloakrealm-sample
            spec:
              realmName: realm-sample
              keycloakOwner: keycloak-sample   # the name of `kind: Keycloak`
              ```

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: KeycloakRealmGroup
              metadata:
                name: argocd-admins
              spec:
                name: ArgoCDAdmins
                realm: keycloakrealm-sample   # the name of `kind: KeycloakRealm`
              ```
        displayName: EDP Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - authentication
        - authorization
        - edp
        - idp
        - keycloak
        - kuberocketci
        - oauth
        - oidc
        - operator
        - saml
        - sso
        links:
        - name: Edp Keycloak Operator
          url: https://github.com/epam/edp-keycloak-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: KubeRocketCI
          url: https://docs.kuberocketci.io/
        relatedImages:
        - docker.io/epamedp/keycloak-operator:1.29.0
        version: 1.29.0
      entries:
      - name: edp-keycloak-operator.v1.29.0
        version: 1.29.0
      - name: edp-keycloak-operator.v1.28.0
        version: 1.28.0
      - name: edp-keycloak-operator.v1.27.1
        version: 1.27.1
      - name: edp-keycloak-operator.v1.26.0
        version: 1.26.0
      - name: edp-keycloak-operator.v1.25.0
        version: 1.25.0
      - name: edp-keycloak-operator.v1.24.0
        version: 1.24.0
      - name: edp-keycloak-operator.v1.23.0
        version: 1.23.0
      name: stable
    defaultChannel: stable
    packageName: edp-keycloak-operator
    provider:
      name: KubeRocketCI
      url: https://docs.kuberocketci.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: minhokim
      provider-url: ""
    name: node-discovery-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-discovery-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.wunicorns.com/v1alpha1",
                "kind": "NodeDiscovery",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "node-discovery-operator",
                    "app.kubernetes.io/instance": "nodediscovery-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodediscovery",
                    "app.kubernetes.io/part-of": "node-discovery-operator"
                  },
                  "name": "nodediscovery-sample"
                },
                "spec": {
                  "containers": [
                    {
                      "args": [
                        "tail",
                        "-f",
                        "/dev/null"
                      ],
                      "image": "alicek106/stress",
                      "imagePullPolicy": "IfNotPresent",
                      "name": "stress",
                      "resources": {
                        "limits": {
                          "cpu": "100m",
                          "memory": "1000Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "1000Mi"
                        }
                      },
                      "terminationGracePeriodSeconds": 10
                    }
                  ],
                  "cpus": "0",
                  "enabled": "true",
                  "mems": "0",
                  "nodeName": "worker3"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-18T01:32:51Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeDiscovery is the Schema for the nodediscoveries API
            displayName: Node Discovery
            kind: NodeDiscovery
            name: nodediscoveries.cache.wunicorns.com
            version: v1alpha1
        description: find cgroup infor and set value.
        displayName: node-discovery-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cgroup
        - cpulist
        - mems
        - cpus
        - numa
        links:
        - name: Node Discovery Operator
          url: https://node-discovery-operator.domain
        maintainers:
        - email: wunicorns1436@gmail.com
          name: minhokim
        maturity: alpha
        provider:
          name: minhokim
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/caspell/cgroup-discovery@sha256:b437dcde0107124e43a4e2fb8026ec191f44b5e761bedc2c295f47ff489cf343
        version: 0.0.1
      entries:
      - name: node-discovery-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: node-discovery-operator
    provider:
      name: minhokim
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HashiCorp
      provider-url: https://www.hashicorp.com/
    name: hcp-terraform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hcp-terraform-operator.v2.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "AgentPool",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "agentTokens": [
                    {
                      "name": "token"
                    }
                  ],
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Module",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "module": {
                    "source": "redeux/terraform-cloud-agent/kubernetes",
                    "version": "1.0.1"
                  },
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  },
                  "workspace": {
                    "id": "WORKSPACE_ID"
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Project",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Workspace",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.connect.redhat.com/hashicorp/hcp-terraform-operator@sha256:8885a6a5eae092117a670ec5541a3e4a24c6b23807fd17b0446d91b01eb6e421
          createdAt: "2025-05-28T13:58:52Z"
          description: HCP Terraform Operator for Kubernetes allows managing HCP Terraform
            / Terraform Enterprise resources via Kubernetes Custom Resources.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Valid HCP Terraform / Terraform
            Enterprise subscription is required
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hashicorp/hcp-terraform-operator
          support: HashiCorp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              AgentPool manages HCP Terraform Agent Pools, HCP Terraform Agent Tokens and can perform HCP Terraform Agent scaling.
              More infromation:
                - https://developer.hashicorp.com/terraform/cloud-docs/agents/agent-pools
                - https://developer.hashicorp.com/terraform/cloud-docs/users-teams-organizations/api-tokens#agent-api-tokens
                - https://developer.hashicorp.com/terraform/cloud-docs/agents
            displayName: Agent Pool
            kind: AgentPool
            name: agentpools.app.terraform.io
            version: v1alpha2
          - description: |-
              Module implements API-driven Run Workflows.
              More information:
                - https://developer.hashicorp.com/terraform/cloud-docs/run/api
            displayName: Module
            kind: Module
            name: modules.app.terraform.io
            version: v1alpha2
          - description: |-
              Project manages HCP Terraform Projects.
              More information:
              - https://developer.hashicorp.com/terraform/cloud-docs/projects/manage
            displayName: Project
            kind: Project
            name: projects.app.terraform.io
            version: v1alpha2
          - description: |-
              Workspace manages HCP Terraform Workspaces.
              More information:
                - https://developer.hashicorp.com/terraform/cloud-docs/workspaces
            displayName: Workspace
            kind: Workspace
            name: workspaces.app.terraform.io
            version: v1alpha2
        description: HCP Terraform Operator for Kubernetes allows managing HCP Terraform
          / Terraform Enterprise resources via Kubernetes Custom Resources.
        displayName: HCP Terraform Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - terraform
        links:
        - name: HCP Terraform Operator
          url: https://github.com/hashicorp/hcp-terraform-operator
        maintainers:
        - email: hcp-tf-operator@hashicorp.com
          name: HashiCorp
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: HashiCorp
          url: https://www.hashicorp.com/
        relatedImages:
        - registry.connect.redhat.com/hashicorp/hcp-terraform-operator@sha256:8885a6a5eae092117a670ec5541a3e4a24c6b23807fd17b0446d91b01eb6e421
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        version: 2.9.2
      entries:
      - name: hcp-terraform-operator.v2.9.2
        version: 2.9.2
      - name: hcp-terraform-operator.v2.9.1
        version: 2.9.1
      - name: hcp-terraform-operator.v2.9.0
        version: 2.9.0
      - name: hcp-terraform-operator.v2.8.1
        version: 2.8.1
      - name: hcp-terraform-operator.v2.8.0
        version: 2.8.0
      name: alpha
    defaultChannel: alpha
    packageName: hcp-terraform-operator
    provider:
      name: HashiCorp
      url: https://www.hashicorp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Appian Corporation
      provider-url: https://appian.com/
    name: appian-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appian-operator.v0.184.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: AI/Machine Learning, Application Runtime, Big Data, Integration
            & Delivery, Modernization & Migration
          createdAt: "2025-08-07T15:50:26Z"
          description: The Appian operator facilitates managing instances of Appian
            on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appian-operator
          operators.openshift.io/valid-subscription: '["Appian License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Appian Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Appian
            name: appians.crd.k8s.appian.com
            version: v1beta1
        description: The Appian operator facilitates managing instances of Appian
          on Kubernetes.
        displayName: Appian operator
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - Appian
        links:
        - name: Appian
          url: https://appian.com/
        - name: Documentation
          url: https://docs.appian.com/suite/help/latest/k8s
        maintainers:
        - email: support@appian.com
          name: Appian Corporation
        maturity: planning
        provider:
          name: Appian Corporation
          url: https://appian.com/
        relatedImages:
        - 937620572175.dkr.ecr.us-east-1.amazonaws.com/appian/appian-operator@sha256:26c2ba27f9360226f34f289109cb0d10e9bf44777146ccfd5f367adf2367a400
        version: 0.184.0
      entries:
      - name: appian-operator.v0.184.0
        version: 0.184.0
      - name: appian-operator.v0.171.0
        version: 0.171.0
      name: alpha
    defaultChannel: alpha
    packageName: appian-operator
    provider:
      name: Appian Corporation
      url: https://appian.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-ingress
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-ingress.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Ingress",
                "metadata": {
                  "name": "ingress-sample"
                },
                "spec": {
                  "admissionWebhooks": {
                    "patch": {
                      "image": {
                        "repository": "docker.io/jettech/kube-webhook-certgen"
                      }
                    }
                  },
                  "configMapMerger": {
                    "registry": "prod.harbor.keysight.digital/loadcore",
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "environment": "openshift",
                  "ingress-nginx": {
                    "controller": {
                      "config": {
                        "global-auth-signin": "https://$host/eula/static/",
                        "global-auth-url": "http://kcos-framework-v1-kcos-eula.kcos-framework.svc.cluster.local/v1/check"
                      },
                      "daemonset": {
                        "useHostPort": false
                      },
                      "hostNetwork": false,
                      "image": {
                        "registry": "prod.harbor.keysight.digital/loadcore"
                      },
                      "ingressClassResource": {
                        "default": true
                      },
                      "nodeSelector": "",
                      "service": {
                        "nodePorts": {
                          "http": 80,
                          "https": 443
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "defaultBackend": {
                      "image": {
                        "repository": "k8s.gcr.io/defaultbackend-amd64"
                      }
                    }
                  },
                  "jobDelWebhook": {
                    "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:28:18Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ingress
            name: ingresses.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Ingress Operator
        displayName: Keysight KCOS Ingress Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Ingress
          url: https://keysight-kcos-ingress.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-ingress@sha256:c6f6f8b44276efd8f5ffaa099b393e7e97d9cfa627952c0d049b98d92dc3ab59
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        version: 0.1.21
      entries:
      - name: keysight-kcos-ingress.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-ingress
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-migration
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-migration.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Migration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {
                  "auth": {
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configuration": {
                    "port": 8080
                  },
                  "database": {
                    "image": {
                      "name": "postgresql-ubi-image",
                      "tag": "14.2.0-3747-32"
                    }
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "esinfo": {
                    "batchSize": 10000,
                    "inputIndex": "wireless_stats"
                  },
                  "image": {
                    "remoteRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "systemMonitor": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:36:21Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Migration
            name: migrations.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Migration Operator
        displayName: Keysight WAP Migration Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Migration
          url: https://keysight-wap-migration.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-migration@sha256:118b466f1c59832a3620fa9b153507cc66ba38ee08b337b3a1a4225b158bebc5
        version: 0.1.21
      entries:
      - name: keysight-wap-migration.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-migration
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-xtradb-cluster-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator-certified.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator-certified.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator-certified.v1.17.0
        version: 1.17.0
      - name: percona-xtradb-cluster-operator-certified.v1.16.1
        version: 1.16.1
      - name: percona-xtradb-cluster-operator-certified.v1.16.0
        version: 1.16.0
      - name: percona-xtradb-cluster-operator-certified.v1.15.1
        version: 1.15.1
      - name: percona-xtradb-cluster-operator-certified.v1.15.0
        version: 1.15.0
      name: stable
    - currentCSV: percona-xtradb-cluster-operator-certified.v1.18.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        version: 1.18.0-cw
      entries:
      - name: percona-xtradb-cluster-operator-certified.v1.18.0-cw
        version: 1.18.0-cw
      - name: percona-xtradb-cluster-operator-certified.v1.17.0-cw
        version: 1.17.0-cw
      - name: percona-xtradb-cluster-operator-certified.v1.16.1-cw
        version: 1.16.1-cw
      - name: percona-xtradb-cluster-operator-certified.v1.16.0-cw
        version: 1.16.0-cw
      - name: percona-xtradb-cluster-operator-certified.v1.15.0-cw
        version: 1.15.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator-certified
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-storage-odf-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-storage-odf-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:962ca15393dd2c48f8a7189c1801b43e9fd1cd073a1971286d485f4b96060426
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.6.0'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:962ca15393dd2c48f8a7189c1801b43e9fd1cd073a1971286d485f4b96060426
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:2f785096aaa63b042e415aa0b64ace2a085cb7c219853f089d803bc37d8295ce
        version: 1.6.0
      entries:
      - name: ibm-storage-odf-operator.v1.6.0
        version: 1.6.0
      name: stable-v1.6
    - currentCSV: ibm-storage-odf-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:2dd8608645b13f81330466e25664b132d46af480a3ab2b0dd48fb1feba38870a
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.7.1'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:2dd8608645b13f81330466e25664b132d46af480a3ab2b0dd48fb1feba38870a
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:24eae9d1502ce98bb024d3eba154ddec07fdac0d78ecad0c3a64b55f5b3600f7
        version: 1.7.1
      entries:
      - name: ibm-storage-odf-operator.v1.7.1
        version: 1.7.1
      name: stable-v1.7
    - currentCSV: ibm-storage-odf-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:7432340ac156e620240eea73a19321a871fa89cc513c340a91d97d3f3fd4ded2
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.8.0'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:7432340ac156e620240eea73a19321a871fa89cc513c340a91d97d3f3fd4ded2
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:3344c7e987674fb98173bbc6ae1b189613e3ccefb468f1af9866734ddbf0408f
        version: 1.8.0
      entries:
      - name: ibm-storage-odf-operator.v1.8.0
        version: 1.8.0
      name: stable-v1.8
    defaultChannel: stable-v1.8
    packageName: ibm-storage-odf-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Strimzi
      provider-url: ""
    name: strimzi-kafka-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: strimzi-cluster-operator.v0.48.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.1.0",
                      "metadataVersion": "4.1-IV1",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.1.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
          createdAt: "2025-09-25 10:32:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.48.0 supports only Kubernetes 1.27 and newer.**
            Kubernetes 1.25 and 1.26 are not supported anymore

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.48.0
          * Support for Kafka 4.1.0
          * Add monitoring of custom resources using kubernetes-state-metrics (KSM)
          * Support for type: custom client authentication
          * Added support for Strimzi Metrics Reporter to Kafka Connect, Mirror Maker 2 and Kafka Bridge.
          * Added support for KIP-745 in Kafka Connector and Mirror Maker 2, allowing the usage of includeTasks and onlyFailed arguments in Kafka connectors restart.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
        - quay.io/strimzi/kafka@sha256:34450afc1a3399a9026fc3c6ec90e9f6adfec9faf6520545b4d84e90cda34964
        - quay.io/strimzi/kafka-bridge@sha256:53034f64f0b672f10b5bacea1c7a25132f118df7fd5c9032c4dbf702ed93796a
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:64380efcb8da87fdc0e0de238a20456be9abb830174c2f2331882b7a84657e57
        - quay.io/strimzi/kafka@sha256:48db05ad2e9fa39eb388089e7e25654bcbf4da1ddc3a5635d675056c28bbcb96
        version: 0.48.0
      entries:
      - name: strimzi-cluster-operator.v0.48.0
        version: 0.48.0
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: stable
    - currentCSV: strimzi-cluster-operator.v0.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnectS2I",
                "metadata":{
                  "name":"my-connect-s2i-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "whitelist":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 2,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": 1,
                        "offset.storage.replication.factor": 1,
                        "status.storage.replication.factor": 1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "NetworkOutboundCapacityGoal",
                    "CpuCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:59d939f342f96d0e8a54c96fd187832c4b19cd4137bf55ace53d486728ae6c5f
          createdAt: "2021-05-14 11:37:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster with Source 2 Image support
            displayName: Kafka Connect Source to Image
            kind: KafkaConnectS2I
            name: kafkaconnects2is.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  Starting with Strimzi 0.23, we support only the `v1beta2` API version of all Strimzi custom resources.  This is a required for migration to `apiextensions/v1` which is needed because Kubernetes 1.22 will remove support for `apiextensions/v1beta1`.  Migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs **before upgrading to Strimzi 0.23**. For more details about the CRD upgrades, see the [documentation](https://strimzi.io/docs/operators/0.23.0/deploying.html#assembly-upgrade-resources-str).
          ### New in 0.23
          * Migrate to CRD v1 (required by Kubernetes 1.22+)
          * Strimzi API versions `v1alpha1` and `v1beta1` were removed from all Strimzi custom resources apart from `KafkaTopic` and `KafkaUser` (use `v1beta2` versions instead)
          * Add support for Kafka 2.8.0 and 2.6.2, remove support for Kafka 2.5.x
          * Make it possible to configure maximum number of connections and maximum connection creation rate in listener configuration
          * Add support for configuring finalizers for `loadbalancer` type listeners
          * Use dedicated Service Account for Kafka Connect Build on Kubernetes
          * Remove direct ZooKeeper access for handling user quotas in the User Operator. Add usage of Admin Client API instead.
          * Support for configuring custom Authorizer implementation
          * Changed Reconciliation interval for Topic Operator from 90 to 120 seconds (to keep it the same as for other operators)
          * Changed Zookeeper session timeout default value to 18 seconds for Topic and User Operators (for improved resiliency)
          * Removed requirement for replicas and partitions KafkaTopic spec making these parameters optional
          * Support to configure a custom filter for parent CR's labels propagation into subresources
          * Allow disabling service links (environment variables describing Kubernetes services) in Pod template
          * Add support for separate control plane listener (disabled by default, available via the `ControlPlaneListener` feature gate)
          * Support for Dual Stack networking
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _master_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/master/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/master/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:59d939f342f96d0e8a54c96fd187832c4b19cd4137bf55ace53d486728ae6c5f
        - quay.io/strimzi/kafka@sha256:699bff48ee445d8ba381a791cd0340015e3ecede8d169f41bc151095ab7fb5cc
        - quay.io/strimzi/kafka@sha256:d2745f48ed262290ce66c3362ffb9b9a8cfeb9c76beac40c995f72f3f06ad10e
        - quay.io/strimzi/kafka@sha256:dbc53bb374d2cb9513a5bb8f80fea751c91b249cabd36d8b3c04af5e5c961738
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/kafka@sha256:e90695b35d6dd5db20c9900dd0dab95d1ef908f3145b4f9f436b903303c88549
        - quay.io/strimzi/kafka-bridge@sha256:bed07186355a4c04ee50e90048e3bd3a023264093a9ef28f0278cce44640e93f
        - quay.io/strimzi/kafka@sha256:641f46114c59d61577bcb3f4ef5af16aea3ed42c2a575f8d3641c82848df85e8
        - quay.io/strimzi/jmxtrans@sha256:b4ec1844f4cc5eb74cd0aac83d74b81b977e178bf670d1136d1af0b68663d0a3
        version: 0.23.0
      entries:
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.23.x
    - currentCSV: strimzi-cluster-operator.v0.24.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnectS2I",
                "metadata":{
                  "name":"my-connect-s2i-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 2,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:786ea69c77e0651f009a73a8b3ce0b3d41b3cd71b917dc0d9bdc61c725f494a6
          createdAt: "2021-06-24 13:13:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster with Source 2 Image support
            displayName: Kafka Connect Source to Image
            kind: KafkaConnectS2I
            name: kafkaconnects2is.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: "Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org)
          cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/)
          in various deployment configurations. See our [website](https://strimzi.io)
          for more details about the project.\n### CRD Upgrades\n**!!! IMPORTANT !!!**
          \ This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.
          \ If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed
          for all Strimzi CRDs and CRs before the upgrade to 0.24 is done!  If upgrading
          from Strimzi version earlier than 0.22, you need to first install the CRDs
          from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi
          CRDs and CRs before the upgrade to 0.24 is done!  For more details about
          the CRD upgrades, see the documentation.\n### New in 0.24\n* Add support
          for Kafka 2.7.1 and remove support for 2.6.0, 2.6.1, and 2.6.2\n* Add support
          for [Kubernetes Configuration Provider for Apache Kafka](https://github.com/strimzi/kafka-kubernetes-config-provider)\n*
          Use Red Hat UBI8 base image\n* Support for patching of service accounts
          and configuring their labels and annotations. The feature is disabled by
          default and enabled using the new `ServiceAccountPatching` feature gate.\n*
          Added support for configuring cluster-operator's worker thread pool size
          that is used for various sync and async tasks\n* Add Kafka Quotas plugin
          with produce, consume, and storage quotas\n* Support pausing reconciliation
          of KafkaTopic CR with annotation `strimzi.io/pause-reconciliation`\n* Update
          cruise control to 2.5.57\n* Update to Strimzi Kafka Bridge to 0.20.1\n*
          Support for broker load information added to the rebalance optimization
          proposal. Information on the load difference, before and after a rebalance
          is stored in a ConfigMap\n* Add support for selectively changing the verbosity
          of logging for individual CRs, using markers.\n* Added support for `controller_mutation_rate'
          quota. Creation/Deletion of topics and creation of partitions can be configured
          through this.\n* Use newer version of Kafka Exporter with different bugfixes
          \n### Supported Features\n* **Manages the Kafka Cluster** - Deploys and
          manages all of the components of this complex application, including dependencies
          like Apache ZooKeeper® that are traditionally hard to administer.\n* **Includes
          Kafka Connect** - Allows for configuration of common data sources and sinks
          to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for morroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashabords\n### Upgrading your Clusters\nThe
          Strimzi Operator understands how to run and upgrade between a set of Kafka
          versions. When specifying a new version in your config, check to make sure
          you aren't using any features that may have been removed. See [the upgrade
          guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str)
          for more information.\n### Storage\nAn efficient data storage infrastructure
          is essential to the optimal performance of Apache Kafka®. Apache Kafka®
          deployed via Strimzi requires block storage. The use of file storage (for
          example, NFS) is not recommended.\nThe Strimzi Operator supports three types
          of data storage:\n* Ephemeral (Recommended for development only!)\n* Persistent\n*
          JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nStrimzi
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation to the current _main_
          branch as well as all releases can be found on our [website](https://strimzi.io/documentation).\n###
          Getting help\nIf you encounter any issues while using Strimzi, you can get
          help using:\n* [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          Contributing\nYou can contribute by:\n* Raising any issues you find using
          Strimzi\n* Fixing issues by opening Pull Requests\n* Improving documentation\n*
          Talking about Strimzi\n\nAll bugs, tasks or enhancements are tracked as
          [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues).
          Issues which  might be a good start for new contributors are marked with
          [\"good-start\"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start)
          label.\n\nThe [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md)
          describes how to build Strimzi and how to  test your changes before submitting
          a patch or opening a PR.\n\nThe [Documentation Contributor Guide](https://strimzi.io/contributing/guide/)
          describes how to contribute to Strimzi documentation.\n\nIf you want to
          get in touch with us first before contributing, you can use:\n* [Strimzi
          mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          License\nStrimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).\n"
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/kafka@sha256:95cfe9000afda2f7def269ca46472d3803ee85146c521884884d8505a7187daf
        - quay.io/strimzi/kafka@sha256:8959b7968ab8b3306906cdbff2ebb8d63329af37e58124a601843795c4ef5bd6
        - quay.io/strimzi/kafka@sha256:fbb08410d9595029bc4a02ed859971264e6ce2dc85dd6a9855eaa7bb58b52a25
        - quay.io/strimzi/kafka-bridge@sha256:537ea10e87619251204747ebc1005c632a6c6ed270f5ef06f86c0c3ae4a7553b
        - quay.io/strimzi/operator@sha256:786ea69c77e0651f009a73a8b3ce0b3d41b3cd71b917dc0d9bdc61c725f494a6
        - quay.io/strimzi/jmxtrans@sha256:2c54adabf6f31c087dbb898057f9c5e200aaf4c5b39fb3a47bbaf7f6587df72d
        version: 0.24.0
      entries:
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.24.x
    - currentCSV: strimzi-cluster-operator.v0.25.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:afaf1dd1f2f696ae80674d74159661ea239ae609f78a73a1423c7a17a39e5044
          createdAt: "2021-06-24 13:13:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.25 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.25 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.25
          * Move from Scala 2.12 to Scala 2.13
          * Open Policy Agent authorizer updated to a new version supporting Scala 2.13.
            The new version of the authroizer uses different format of the input data sent to the OPA server and might require you to update your policies.
            See [release notes](https://github.com/strimzi/strimzi-kafka-operator/releases/tag/0.25.0) for more details.

          * Allow a custom password to be set for SCRAM-SHA-512 users by referencing a secret in the `KafkaUser` resource
          * Add support for [EnvVar Configuration Provider for Apache Kafka](https://github.com/strimzi/kafka-env-var-config-provider)
          * Add support for `tls-external` authentication to User Operator to allow management of ACLs and Quotas for TLS users with user certificates generated externally
          * Support for disabling the automatic generation of network policies by the Cluster Operator
          * Configure fixed size limit for `emptyDir` volumes used for temporary files
          * Update Strimzi Kafka Bridge to 0.20.2
          * The `KafkaConnectS2I` resource has been removed and is no longer supported by the operator.
            Please use the [migration guide](https://strimzi.io/docs/operators/0.24.0/full/using.html#proc-migrating-kafka-connect-s2i-str) to migrate your `KafkaConnectS2I` deployments to [`KafkaConnect` Build](https://strimzi.io/docs/operators/latest/full/deploying.html#creating-new-image-using-kafka-connect-build-str) instead.

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:d8991608ec71f17e850c33714ee02ca87ba94bc0f077be62ad1ed127ca244823
        - quay.io/strimzi/kafka@sha256:ee17eb3611019b0cfa5e9e196610edc1dbd7084471fb20990f055c89da8eef45
        - quay.io/strimzi/kafka@sha256:56a33815f7d50a97ea9bb4cce1e3a6d04d918b841b5aa285f16b792d8930059e
        - quay.io/strimzi/kafka-bridge@sha256:a005e20d7fbe04d6f587f2093e21a00556ec62b850df757b9f7d9d57e3141c40
        - quay.io/strimzi/jmxtrans@sha256:af5eb8459ac9a0acbca51e67c6e0217f62ba404cba9f2251968c927dcef75252
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/operator@sha256:afaf1dd1f2f696ae80674d74159661ea239ae609f78a73a1423c7a17a39e5044
        version: 0.25.0
      entries:
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.25.x
    - currentCSV: strimzi-cluster-operator.v0.26.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.0.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"3.0",
                        "inter.broker.protocol.version":"3.0"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.0.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:5d90caa66f57939ba62a9a50efa00fce319ba596e719a3f8a42170d7635ee3f4
          createdAt: "2021-12-14 00:16:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.26 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.26 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.26.1
          * **Updated Log4j2 to fix CVE-2021-44228**
          ### New in 0.26.0
          * Add support for Kafka 2.8.1 and 3.0.0; remove Kafka 2.7.0 and 2.7.1
          * Add support for exposing JMX in Zookeeper
          * Allow configuring labels and annotations for JMX authentication secrets
          * Enable Cruise Control anomaly.detection configurations
          * Add support for building connector images from the Maven coordinates
          * Allow Kafka Connect Build artifacts to be downloaded from insecure servers
          * Add option to specify pull secret in Kafka Connect Build on OpenShift
          * Configurable authentication, authorization, and SSL for Cruise Control API
          * Allow to configure `/tmp` volume size via Pod template. By default `1Mi` is used.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:d3899c79747e4cb3a6df5a9fcd80b81f038ca0634a97329c1d160499469876dd
        - quay.io/strimzi/jmxtrans@sha256:bff5d3bab575d0b9ed95dd79d2d22518cc00c4d1c9dbb5614ed16149d4920a1b
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/maven-builder@sha256:663eb81388ae8f824e7920c272f6d2e2274cf6c140d61416607261cdce9d50e2
        - quay.io/strimzi/kafka@sha256:8c022a8535742b2eea691737f863af5ddef1e8df08b415b8139ecd02a79052a5
        - quay.io/strimzi/kafka@sha256:bd22a134b0c351ddfc78a3263850c2f16733c52f57ac92984cbe0ffff477c724
        - quay.io/strimzi/operator@sha256:5d90caa66f57939ba62a9a50efa00fce319ba596e719a3f8a42170d7635ee3f4
        - quay.io/strimzi/kafka@sha256:6367b0fbba9593e3ac736040ea2d7ef71296b1ec144d3b54d9370352141b90e4
        version: 0.26.1
      entries:
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.26.x
    - currentCSV: strimzi-cluster-operator.v0.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.0.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"3.0",
                        "inter.broker.protocol.version":"3.0"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.0.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:9ea49407a04006d0bddfaaee822adfb95f5857af32ffcee81ba60ce891b07ba8
          createdAt: "2022-01-14 20:00:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.27 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.27 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.27.1
          * Updated Log4j2 to version 2.17.1 and Fabric8 Kubernetes Client to 5.10.2
          ### New in 0.27.0
          * Updated Log4j2 to version 2.17.0 (adresses CVE-2021-44228, CVE-2021-45046, CVE-2021-45105)
          * (Experimental) support for AArch64 / ARM64 architecture
          * `ControlPlaneListener` and `ServiceAccountPatching` feature gates are now in the beta phase and are enabled by default
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a07ebe6600d6baba09980abdde2931b3b6abd2f75aeb2cac5c6a2d22a37aa807
        - quay.io/strimzi/maven-builder@sha256:9acd178cc504be8acaa70a39f3c17830aae1f0e93a656110790f9476b09260d5
        - quay.io/strimzi/kafka@sha256:f4d68bb94447c6612f70de7f6587e9e7ef712c83769ea1f11a8dbda0d241a059
        - quay.io/strimzi/kafka@sha256:a3ec567ca529df2fa41f8986d21e86dde66be94aa4baa56faa66db4dcb3b258e
        - quay.io/strimzi/kafka@sha256:5682a6bfcc4ee13d53a30d587c8f96bc40abe3618880989336c5af274ea3cefa
        - quay.io/strimzi/kafka-bridge@sha256:55a52316aead02efa2004174e11ad4728bb8b6f1c1415f725f597f1d2f2da188
        - quay.io/strimzi/operator@sha256:9ea49407a04006d0bddfaaee822adfb95f5857af32ffcee81ba60ce891b07ba8
        - quay.io/strimzi/jmxtrans@sha256:d2d2477517eb1ac98b51ad6d64d8c1c578141637afa8d46037076f0013183c0b
        version: 0.27.1
      entries:
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.27.x
    - currentCSV: strimzi-cluster-operator.v0.28.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.1.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.1"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.1.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.1.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:8b010c5fb8244b742a37f264feefc116d75bb2abd2d6895375cbeea250727b39
          createdAt: "2022-02-12 23:02:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.28 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.28 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.28.0
          * Support for Kafka 3.1.0
          * Support for `StrimziPodSet` resources (disabled by default through the `UseStrimziPodSets` feature gate)
          * Support custom authentication mechanisms in Kafka listeners
          * Intra-broker disk balancing using Cruise Control
          * Add support for disabling the FIPS mode in OpenJDK
          * Fix renewing your own CA certificates
          * Added the option `createBootstrapService` to disable the creation of the bootstrap service for the Load Balancer Type Listener
          * Dependency updates (Strimzi Kafka Bridge, Strimzi OAuth, Open Policy Agent Authroizer plugin and others)
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a07ebe6600d6baba09980abdde2931b3b6abd2f75aeb2cac5c6a2d22a37aa807
        - quay.io/strimzi/maven-builder@sha256:3d5d664e089651944712361a6d8ba42b578dbd9bb3dfd9972f614520c9cfab75
        - quay.io/strimzi/kafka@sha256:7c04f65e216bb941458aeadffb85e9eeee43d5ca67da566017c2c93a0664c98e
        - quay.io/strimzi/kafka@sha256:06e9bda15495feb2b92d9950e834cdeed6ef20064f021a879610f1b59dd46950
        - quay.io/strimzi/kafka-bridge@sha256:c345c783d8fbbee4edeca2ebc47e7c454fcc2e26686820b8037de97369280ba0
        - quay.io/strimzi/jmxtrans@sha256:3a55a81790dc88b90d1bf858e2cfd7e591ad0e79f763c6d1667b2d3eb0d0c79d
        - quay.io/strimzi/operator@sha256:8b010c5fb8244b742a37f264feefc116d75bb2abd2d6895375cbeea250727b39
        version: 0.28.0
      entries:
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.28.x
    - currentCSV: strimzi-cluster-operator.v0.29.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:afde94c6d2544207cb7f3c94b54455643558a9d5daaa13b9f736cc8da20e319c
          createdAt: "2022-05-15 00:01:07"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.29 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.29 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.29.0
          * Add support for Apache Kafka 3.0.1, 3.1.1 and 3.2.0
          * Renew user certificates in User Operator only during maintenance windows
          * Remove TLS sidecar from Cruise Control pod. Cruise Control is now configured to not using ZooKeeper, so the TLS sidecar is not needed anymore.
          * Allow Cruise Control topic names to be configured
          * Add support for `spec.rack.topologyKey` property in Mirror Maker 2 to enable "fetch from the closest replica" feature.
          * Experiemntal support for the s390x platform
          * Added new rebalancing modes on the `KafkaRebalance` custom resource to add or remove brokers
          * Experimental KRaft mode (ZooKeeper-less Kafka). Use it for development and testing only!
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:544c4987f7f3ca558b04e9ef4357be204f088212b7b5b9ed0f909401a8bc5238
        - quay.io/strimzi/maven-builder@sha256:0c096a59faf8fd224298c01415a895043db4647534c060cb7e53207e9448bf65
        - quay.io/strimzi/operator@sha256:afde94c6d2544207cb7f3c94b54455643558a9d5daaa13b9f736cc8da20e319c
        - quay.io/strimzi/kafka@sha256:95678e843bc87cbe571f3f299045dd075725816b59e1b87821cd6025876ec559
        - quay.io/strimzi/kafka@sha256:935caab03f1d0365b47ddefec0c1d8fa449fc3de715ccdd298aedef5ce968c73
        - quay.io/strimzi/kafka@sha256:ac190141120c1800cc7131d3a1f701e64efdcd4a5d87f0b2e0e2cadfdef8b438
        - quay.io/strimzi/kafka@sha256:9dfc09273ad8fc4bdc75cbe93b4a061a05a5defb5884a221ae1e178ba602787f
        - quay.io/strimzi/kafka@sha256:284cde40784b6ff7bfb81b6dd28928e1e8c598fe6a72504f025592e5b43fa94c
        - quay.io/strimzi/kafka-bridge@sha256:b7bc3dd90e87223766f608a0087b521f0bba79d1c21eaa833bc2708c71148b7f
        - quay.io/strimzi/jmxtrans@sha256:f89e3c9618430908a8b390f2490f5aa43687b12889b434472a962ce0170c6463
        version: 0.29.0
      entries:
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.29.x
    - currentCSV: strimzi-cluster-operator.v0.30.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:587db6f4389f56209f48853beedbfe249f6d590a4cb0000be2c42b57e43c2adb
          createdAt: "2022-07-16 00:00:10"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: "Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org)
          cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/)
          in various deployment configurations. See our [website](https://strimzi.io)
          for more details about the project.\n### CRD Upgrades\n**!!! IMPORTANT !!!**
          \ This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.
          \ If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed
          for all Strimzi CRDs and CRs before the upgrade to 0.30 is done!  If upgrading
          from Strimzi version earlier than 0.22, you need to first install the CRDs
          from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi
          CRDs and CRs before the upgrade to 0.30 is done!  For more details about
          the CRD upgrades, see the documentation.\n### New in 0.30.0\n* The `UseStrimziPodSets`
          feature gate moves to beta stage. By default, StrimziPodSets are used instead
          of StatefulSets. If needed, `UseStrimziPodSets` can be disabled in the feature
          gates configuration in the Cluster Operator.\n* The `ServiceAccountPatching`
          feature gate moves to GA. It cannot be disabled anymore and will be permanently
          enabled.\n* Remove support for Kafka 3.0.0 and 3.0.1\n* Add support for
          `simple` authorization and for the User Operator to the experimental `UseKRaft`
          feature gate. _(Note: Due to [KAFKA-13909](https://issues.apache.org/jira/browse/KAFKA-13909),
          broker restarts currently don't work when authorization is enabled.)_\n*
          Add CPU and network capacity overrides for Cruise Control capacity config\n*
          Use better encryption and digest algorithms when creating the PKCS12 stores.
          For existing clusters, the certificates will not be updated during upgrade
          but only next time the PKCS12 store is created. \n* Operator emits Kafka
          events to explain why it restarted a Kafka broker\n* Better configurability
          of the Kafka Admin client in the User Operator\n* Update Strimzi Kafka Bridge
          to 0.21.6\n### Supported Features\n* **Manages the Kafka Cluster** - Deploys
          and manages all of the components of this complex application, including
          dependencies like Apache ZooKeeper® that are traditionally hard to administer.\n*
          **Includes Kafka Connect** - Allows for configuration of common data sources
          and sinks to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for mirroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashboards\n### Upgrading your Clusters\nThe
          Strimzi Operator understands how to run and upgrade between a set of Kafka
          versions. When specifying a new version in your config, check to make sure
          you aren't using any features that may have been removed. See [the upgrade
          guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str)
          for more information.\n### Storage\nAn efficient data storage infrastructure
          is essential to the optimal performance of Apache Kafka®. Apache Kafka®
          deployed via Strimzi requires block storage. The use of file storage (for
          example, NFS) is not recommended.\nThe Strimzi Operator supports three types
          of data storage:\n* Ephemeral (Recommended for development only!)\n* Persistent\n*
          JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nStrimzi
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation to the current _main_
          branch as well as all releases can be found on our [website](https://strimzi.io/documentation).\n###
          Getting help\nIf you encounter any issues while using Strimzi, you can get
          help using:\n* [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          Contributing\nYou can contribute by:\n* Raising any issues you find using
          Strimzi\n* Fixing issues by opening Pull Requests\n* Improving documentation\n*
          Talking about Strimzi\nAll bugs, tasks or enhancements are tracked as [GitHub
          issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues
          which  might be a good start for new contributors are marked with [\"good-start\"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start)
          label.\nThe [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md)
          describes how to build Strimzi and how to  test your changes before submitting
          a patch or opening a PR.\nThe [Documentation Contributor Guide](https://strimzi.io/contributing/guide/)
          describes how to contribute to Strimzi documentation.\nIf you want to get
          in touch with us first before contributing, you can use:\n* [Strimzi mailing
          list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n* [Strimzi
          Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          License\nStrimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).\n"
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:544c4987f7f3ca558b04e9ef4357be204f088212b7b5b9ed0f909401a8bc5238
        - quay.io/strimzi/maven-builder@sha256:f2cb45234f3a7a010824a370d03032b706271cd0d1d260da929f046f9259d7e0
        - quay.io/strimzi/kafka@sha256:68d6f4cb5cc5ccad44de799bd9cf417c469a09e1024dbe42298942f25a09e004
        - quay.io/strimzi/kafka@sha256:d9258ddd453c3a53ca8e737b989d27725818da95f7a4fc879c9f07fe743c0cef
        - quay.io/strimzi/kafka@sha256:5d881605425ab0b37d0b9cac6957d00ba904df0db185ca6181c810151bdab8e7
        - quay.io/strimzi/operator@sha256:587db6f4389f56209f48853beedbfe249f6d590a4cb0000be2c42b57e43c2adb
        - quay.io/strimzi/kafka-bridge@sha256:77fc68da171c74d3a4b75b1030eaaa5eae12e200002f7aa2ee1969385eba1d66
        - quay.io/strimzi/jmxtrans@sha256:de1ca88b713b926b72df535aca2cfad24fb1b2bf40481065d6fa0c421b872320
        version: 0.30.0
      entries:
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.30.x
    - currentCSV: strimzi-cluster-operator.v0.31.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.3",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.3",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:6e5e207c70c25c595d3ce60268c0214c7518e95836c9b1212ef1eb2f008c7a76
          createdAt: "2022-09-20 21:07:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          **This is the last version of Strimzi with support for Kubernetes 1.16, 1.17 and 1.18!**
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.31 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.31 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.31.1
          * Kafka 3.1.2 and 3.2.3 (fixes CVE-2022-34917)
          * Make `sasl.server.max.receive.size` broker option user configurable
          * Update Strimzi Kafka Bridge to 0.22.1
          ### New in 0.31.0
          * Support for Apache Kafka 3.2.1
          * Pluggable Pod Security Profiles with built-in support for _restricted_ Kubernetes Security Profile
          * Support for leader election and running multiple operator replicas (1 active leader replicas and one or more stand-by replicas)
          * Support for IPv6 addresses being used in Strimzi issued certificates
          * `StrimziPodSet` reconciliation metrics
          * Updated Strimzi Kafka Bridge to 0.22.0 and Kafka Exporter to 1.6.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:03cc48bc5cf367f898338665b253855bed7d6096fe577adfe344da2f2cddcfe4
        - quay.io/strimzi/operator@sha256:6e5e207c70c25c595d3ce60268c0214c7518e95836c9b1212ef1eb2f008c7a76
        - quay.io/strimzi/kaniko-executor@sha256:06581fb615f0b13ddfe27fae26ba5836b12c6e28afd55a81cba7685c9ca4bf79
        - quay.io/strimzi/maven-builder@sha256:670ca70900f2229ef140a592b77f2107934601f8b050b79250eab17ff62099b6
        - quay.io/strimzi/kafka-bridge@sha256:3ddac62a4f8c8a3ebb72d5841d3826f7fd3fbb0e00eb75fb40bced0a576b86a9
        - quay.io/strimzi/kafka@sha256:b7c059bd72d8b063a96dde788d91d1ad87498bad3e73acba600929d595aeb7d6
        - quay.io/strimzi/kafka@sha256:430b3eb70e47a91d7b5b6f230d5be024c2de853099cc0437e219b443e0ead9f5
        - quay.io/strimzi/kafka@sha256:9553aae22d35448681cbbf7a831a28149fb159f3df58ccf8c8ec4ca7eb9c1fdf
        - quay.io/strimzi/kafka@sha256:83acc1e2cc1e64a88bdbb54b9818a6ad3658c3da60e72c4be365f824c5fcf57c
        - quay.io/strimzi/jmxtrans@sha256:52dd107062d78c29bb58c905ac654c39df72bbe3c25613023aa33e834930606d
        - quay.io/strimzi/kafka@sha256:652924a1dba46aa716caba8381f1e791c09035e31ce62fdcc81f9b69af36cd02
        version: 0.31.1
      entries:
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.31.x
    - currentCSV: strimzi-cluster-operator.v0.32.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.3.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.3.1",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:c5e0e5dca7503a8371b8d366db0f49e3fc68f62ea480262e39b6cc277991c97c
          createdAt: "2022-10-31 17:05:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **From the 0.32.0 release, Strimzi supports only Kubernetes 1.19 and newer! Kubernetes 1.16, 1.17 and 1.18 are not supported anymore.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.32.0/deploying.html#assembly-upgrade-str).
          ### New in 0.32.0
          * Add support for Kafka 3.3.1 and remove support for Kafka 3.1.0, 3.1.1, and 3.1.2
          * Update KafkaConnector CR status so the 'NotReady' condition is added if the connector or any tasks are reporting a 'FAILED' state.
          * Add auto-approval mechanism on KafkaRebalance resource when an optimization proposal is ready
          * The ControlPlaneListener feature gate moves to GA
          * Add client rack-awareness support to Strimzi Bridge pods
          * Add support for OpenTelemetry for distributed tracing
          * ZookeeperRoller considers unready pods
          * Support multiple operations per ACLRule
          * Add new listener type "cluster-ip" which is using per-broker services instead of the pod DNS names
          * Update Strimzi OAuth library to 0.11.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:8c9f38b45dd814e617c9c51cc90fd51e23d7dd23e8efe7e9eff013522bd6a5a0
        - quay.io/strimzi/kafka@sha256:bb97008b7f928ed98939b2008a7baf8909c88e726a59aaf607f2eeba55375cc0
        - quay.io/strimzi/kafka@sha256:680ae1958dbcb9da8ee4128a67c1163a6ee2744221f7d29f73d8bcc237fd0173
        - quay.io/strimzi/maven-builder@sha256:593e9980b786ca8ad3223151c3e17f16c7652ee0e86d722af4c34b023038ecc2
        - quay.io/strimzi/operator@sha256:c5e0e5dca7503a8371b8d366db0f49e3fc68f62ea480262e39b6cc277991c97c
        - quay.io/strimzi/jmxtrans@sha256:208928133b2854f60bb34912e07561a3511de895b815960e73d565be2d553918
        - quay.io/strimzi/kafka@sha256:4bacf5b54041bd0ab9199c8c5d51e8058758139db4b7abc380cf0400737716ef
        - quay.io/strimzi/kafka-bridge@sha256:aad0cee4381cc1c8c6638ab5732b135acd271b5889486ca9a33312cce67e4a02
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        version: 0.32.0
      entries:
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.32.x
    - currentCSV: strimzi-cluster-operator.v0.33.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:cf0d41fdb99b20e91063e23eb1d664817c18866c9a2f5c4430eac897e78d5d53
          createdAt: "2023-02-18 21:21:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.33 supports only Kubernetes 1.19 and newer! Kubernetes versions 1.16, 1.17 and 1.18 are not supported anymore since Strimzi 0.32.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.33.2/deploying.html#assembly-upgrade-str).
          ### New in 0.33
          _Note: Strimzi 0.33.2 add support for Kafka 3.4.0 which fixes CVE-2023-25194._
          * Add support for Kafka 3.3.2
          * Automatic restarts of failed Connect or Mirror Maker 2 connectors
          * Redesign of Strimzi User Operator to improve its scalability
          * Move to Java 17
          * Improved FIPS support
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/maven-builder@sha256:de9e51365e7acc63feae745618804e6286fd730f41c3b31582cbb7319a93b483
        - quay.io/strimzi/kafka@sha256:116cf543fbadadd0b45007ce24d44a0c79a62280e8faa1fc80df40291fbdf05d
        - quay.io/strimzi/kafka@sha256:9eca149cfc436580e3cdd6cee0ca318366b8d0ec53c2cf007e4e2c614b988921
        - quay.io/strimzi/kafka@sha256:301dc52ed5891ed4fa33e79ca3c344b1ac0eec7469dd46a5a7ac22064e053fc2
        - quay.io/strimzi/kafka-bridge@sha256:3a380d5f3ea509f90cdf95280d83b42f5de4f86ee3a512e02234e7639f50c9eb
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/kafka@sha256:30945aecb9d63cf8863c672222f130469c89aefa07078c95aaed52c40fd16146
        - quay.io/strimzi/jmxtrans@sha256:5ebdafa7a5e3718ed3274d6254c9e655613c272340094b997b1c97fde2598b4a
        - quay.io/strimzi/operator@sha256:cf0d41fdb99b20e91063e23eb1d664817c18866c9a2f5c4430eac897e78d5d53
        - quay.io/strimzi/kafka@sha256:fe9d3f8b029ff32d5d725e57384592914957314d63e92fda3abf4028fe4e66f6
        - quay.io/strimzi/kafka@sha256:40a847f8455ec6ea9ff9fe1fff20b1ed01f848326dc13977d13ac911f0cb40b7
        version: 0.33.2
      entries:
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.33.x
    - currentCSV: strimzi-cluster-operator.v0.34.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:f4ee3ca482a2dd19785f59634578a0850273444c1173a5699a99cf02f3a018fe
          createdAt: "2023-03-17 13:02:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.34 supports only Kubernetes 1.19 and newer! Kubernetes versions 1.16, 1.17 and 1.18 are not supported anymore since Strimzi 0.32.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.34.0/deploying.html#assembly-upgrade-str).
          ### New in 0.34
          * Add support for Apache Kafka 3.4.0 (fixes CVE-2023-25194) and remove support for Apache Kafka 3.2.x
          * Stable Pod identities for Kafka Connect and MirrorMaker 2 (disabled by default)
          * Use JDK HTTP client in the Kubernetes client instead of the OkHttp client
          * Add truststore configuration for HTTPS connections to Open Policy Agent server
          * Update Strimzi HTTP Bridge to 0.25.0 and strimzi Oauth to 0.12.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:37cf8d0fea1d2078511920190a72b49977ff4047b6ecbfed6b394c6f9e1e2ff0
        - quay.io/strimzi/kafka-bridge@sha256:51babca0f7a328f4db005c6a16daa8ac4a4306c46fe3ca89382919134aa09179
        - quay.io/strimzi/jmxtrans@sha256:62272e8a03da2144340bd7267f7a824d1ed70f4f3e30dc7aadfba9ee1b9baf76
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/maven-builder@sha256:f781b6b857d437c2de62585462e97b9fae357912c9f6711c1df39f431278878d
        - quay.io/strimzi/operator@sha256:f4ee3ca482a2dd19785f59634578a0850273444c1173a5699a99cf02f3a018fe
        - quay.io/strimzi/kafka@sha256:1da35fd7b36b113d1343e254fdc6efd97ca8660cc7977d8985aa3a3547d5ebb1
        - quay.io/strimzi/kafka@sha256:1ace2cc0db9653ac1b388477fa56c3c1cd97c7b3846eba3c7824dc58ea179476
        version: 0.34.0
      entries:
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.34.x
    - currentCSV: strimzi-cluster-operator.v0.35.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:06a94a3021cf028ccc1a49271f35f79216029e344536e664f196c1725ff2c663
          createdAt: "2023-06-07 16:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.35 is the last version with support for Kubernetes 1.19 and 1.20. From Strimzi 0.36, only Kubernetes 1.21 and newer will be supported.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.35.1/deploying.html#assembly-upgrade-str).
          ### New in 0.35
          * Move feature gate `UseStrimziPodSets` to GA and remove support for StatefulSets
          * Remove support for JMX Trans
          * Redesigned the Cluster and User Operator configuration to make it more efficient and flexible
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:39778b90c2b2afc30261e4ad5135805e1a10a2b60e2e53108fb9f80487f1208a
        - quay.io/strimzi/maven-builder@sha256:88a79eff3b3a386880a630658964b7754caed9e99dd6e645a4c0d23d0fdb47ee
        - quay.io/strimzi/operator@sha256:06a94a3021cf028ccc1a49271f35f79216029e344536e664f196c1725ff2c663
        - quay.io/strimzi/kafka@sha256:4de4874a7b722ad813f4dcc58acf509527bca0609999b81e70d81e3b38534d9d
        - quay.io/strimzi/kafka@sha256:0d910e7138cb49e1cd8cd84cef88bce35698b93ddd683a3398f1d485a3162693
        - quay.io/strimzi/kafka@sha256:54c6b25b31f51ef401c1b6e2a1b27432911e819bf3e502e05186f01be3f798e5
        - quay.io/strimzi/kafka-bridge@sha256:d6be183e492f8f88157ab9fe0af53950df8b6711a8a8c33da465de6064f6f86e
        version: 0.35.1
      entries:
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.35.x
    - currentCSV: strimzi-cluster-operator.v0.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:e9e03b31007c0f9b67a6a33e6b644e8c1ba35bd9976804ee3d22a04e4bf511b1
          createdAt: "2023-07-10 21:47:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!! Strimzi 0.36.1 supports only Kubernetes 1.21 and newer! Kubernetes versions 1.19 and 1.20 are not supported anymore since Strimzi 0.36.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.35.1/deploying.html#assembly-upgrade-str).
          ### New in 0.36
          * Add support for Apache Kafka 3.4.1, 3.5.0, and 3.5.1
          * Remove support for 3.3.1 and 3.3.2
          * Add support for _Kafka node pools_
          * Add support for _Unidirectional Topic Operator_
          * Enable SCRAM-SHA authentication in KRaft mode (supported in Apache Kafka 3.5.0 and newer)
          * Add support for insecure flag in Maven artifacts in Kafka Connect Build
          * Improve Kafka rolling update to avoid rolling broker in log recovery
          * Added support for Kafka Exporter topic exclude and consumer group exclude parameters
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:9f35f40be3d47b0cb1f6628f7ffd9c6a687ab9f564829f341dd66fd0759c064f
        - quay.io/strimzi/kafka@sha256:67c2b563a65c85c9393608f9af7eaf12ea48dcc0ef5f726d51f27273a3845186
        - quay.io/strimzi/operator@sha256:e9e03b31007c0f9b67a6a33e6b644e8c1ba35bd9976804ee3d22a04e4bf511b1
        - quay.io/strimzi/kafka@sha256:ea35817ff1ef295c682d70562f209ef658fcd87c61f467a4549d7a2f39e71700
        - quay.io/strimzi/kafka-bridge@sha256:fefee252895e9f94756ed5c9ea71749e3fa89281df01bc55878a244ca7d17697
        - quay.io/strimzi/kaniko-executor@sha256:5ff9fc7d0b7b694875ad0150fc787c582d42da3b4864d807ee8a6524e47b0e4e
        - quay.io/strimzi/maven-builder@sha256:2a469b40122027b981914db5b5c102be48280e039c3089963422c181174af226
        - quay.io/strimzi/kafka@sha256:a0da6a48b5bcbff355365f07df2d99ef6443d22c5690495f10a6417e761b3799
        version: 0.36.1
      entries:
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.36.x
    - currentCSV: strimzi-cluster-operator.v0.37.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:52f376e64b9b31d74162e26248c278e9bd301d2d8b4240ee5bcf780dfe2f8ef8
          createdAt: "2023-09-03 17:45:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.37.0/deploying.html#assembly-upgrade-str).
          ### New in 0.37
          * The StableConnectIdentites feature gate moves to a beta stage.
          * Support for the ppc64le platform
          * Version fields to the Kafka custom resource status to track installation and upgrade state
          * Infinite auto-restarts of Kafka Connect and Kafka Mirror Maker 2 connectors
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:1de6e073e8625a193fcfd11dd1a6717cc17596911c10bcebd792c00b08a7ebaf
        - quay.io/strimzi/kafka@sha256:2997ba1c169eb792a6004d7a41b674560515f3ea6e62ab8b9a1fdb747eab40c1
        - quay.io/strimzi/kafka@sha256:dfeb1eb7b94107c443b6bc82fcfee7fcf782516c83c6fc64451fb6a7e2de74a6
        - quay.io/strimzi/kafka-bridge@sha256:5b821f6f3396002ffa7c5804391b627ff9da3b494b2777dde518c55c2b73cefc
        - quay.io/strimzi/kaniko-executor@sha256:5729f1f0511da95e64e358cfabb3980ad9a660af2968118a87d6755f5d652a67
        - quay.io/strimzi/maven-builder@sha256:995be06a50b86c992fcee9f00e44f6a21c3406975e4e35802f10e88b72ffb9df
        - quay.io/strimzi/operator@sha256:52f376e64b9b31d74162e26248c278e9bd301d2d8b4240ee5bcf780dfe2f8ef8
        - quay.io/strimzi/kafka@sha256:38bbadcdb196a373e1698be890caa1dc01134ce91763d8f95c71768a117e1932
        version: 0.37.0
      entries:
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.37.x
    - currentCSV: strimzi-cluster-operator.v0.38.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:aee45bbe439335e45a0423e6565bae0d0727d85c7b807812949bfa127d467fe6
          createdAt: "2023-10-14 18:18:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.38.0/deploying.html#assembly-upgrade-str).
          ### New in 0.38
          * Add support for Apache Kafka 3.6.0 and remove support for 3.4.0 and 3.4.1
          * Container signing and SBOMs
          * Stopping Connect and MirrorMaker 2 connectors
          * Manual rolling updates of Kafka Connect and Kafka Mirror Maker 2 pods using annotation
          * Prevent scale-down of brokers that still contain any replicas
          * Run ZooKeeper- and KRaft-based clusters in parallel
          * Unidirectional Topic Operator improvements
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:0cc01438e8596643407b6b8ff05fafcea7ef54fa9410d4ae2492a9dbf41118b1
        - quay.io/strimzi/kafka@sha256:20d451fb32ec82eab3c16891198ace46bb9b9193f1f5511bb60ce249080dd7d4
        - quay.io/strimzi/kafka-bridge@sha256:13838889cc20a3a382ea64e749cad94f997c0b928840588f2d1befcf19247bea
        - quay.io/strimzi/kaniko-executor@sha256:4fe7a3f3f78e93369284a5ab456873c911230926ca071d1d88167e63d8ce376c
        - quay.io/strimzi/maven-builder@sha256:922d91dab9e2b8e14b6db25d8b48d7f4e3521bedd435468cc22039599ac4f888
        - quay.io/strimzi/operator@sha256:aee45bbe439335e45a0423e6565bae0d0727d85c7b807812949bfa127d467fe6
        - quay.io/strimzi/kafka@sha256:50e24c797455f9cdbb3cdde8e4f26bb8e1f0503469456da5d179017fd3e03ffe
        version: 0.38.0
      entries:
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.38.x
    - currentCSV: strimzi-cluster-operator.v0.39.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.6.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:50b49f9b5e554cca517fc2794e11319d5aaba1676494b285baa3a74fc79eb501
          createdAt: "2023-12-14 20:28:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.39.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.39 is the last minor release with support for Kubernetes 1.21 and 1.22.** From Strimzi 0.40 on, only Kubernetes 1.23 and newer will be supported.
          ### New in 0.39
          * Add support for Apache Kafka 3.5.2 and 3.6.1
          * The `StableConnectIdentities` feature gate moves to GA stage and is now permanently enabled without the possibility to disable it.
            All Connect and Mirror Maker 2 operands will now use StrimziPodSets.

          * The `KafkaNodePools` feature gate moves to the beta stage and is enabled by default.
            If needed, `KafkaNodePools` can be disabled in the feature gates configuration in the Cluster Operator.

          * The `UnidirectionalTopicOperator` feature gate moves to the beta stage and is enabled by default.
            If needed, `UnidirectionalTopicOperator` can be disabled in the feature gates configuration in the Cluster Operator.

          * Improved Kafka Connect metrics and dashboard example files
          * Allow specifying and managing KRaft metadata version
          * Add support for KRaft to KRaft upgrades (Apache Kafka upgrades for the KRaft-based clusters)
          * Improved Kafka Mirror Maker 2 dashboard example file
          * Support for rolling updates of KRaft controller nodes
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:dd1706f36031c96dfcd0da7b3f804168e5f076220caa302973455fdcd76bd03e
        - quay.io/strimzi/kaniko-executor@sha256:723b7add2473e73c45c81361e1708af4971c321b0391c1298a0557fb8ca4f7f6
        - quay.io/strimzi/maven-builder@sha256:6ce3c16284379d5db9e42a6f3e2b765c981091164a28fdd6bf927b6c5d4ab557
        - quay.io/strimzi/kafka@sha256:49b7530ada1175750ad5a706e545ba3594d85e1e3a90f98b0200a73fa9875057
        - quay.io/strimzi/kafka@sha256:84691b3afb67b312400e01572ff08b64c1e1b1d32a4276e0c21bba10c71c36fe
        - quay.io/strimzi/kafka-bridge@sha256:13838889cc20a3a382ea64e749cad94f997c0b928840588f2d1befcf19247bea
        - quay.io/strimzi/kafka@sha256:53b960411e852f5cb182ecc5d395a3562865f83e15e6251f3a9af0eb413feeed
        - quay.io/strimzi/kafka@sha256:2badc65532b483bba6c1f6a8b0c9895e993ce24e9ebd3cd953d0d17fbe36d27a
        - quay.io/strimzi/operator@sha256:50b49f9b5e554cca517fc2794e11319d5aaba1676494b285baa3a74fc79eb501
        version: 0.39.0
      entries:
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.39.x
    - currentCSV: strimzi-cluster-operator.v0.40.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:2228d89b53e3524e69d1bcf5c2cd373884ce696d9005389a6c78dad23310e837
          createdAt: "2024-03-09 21:27:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.40.0/deploying.html#assembly-upgrade-str).

          **From Strimzi 0.40 on, we support only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.40
          * Add support for Apache Kafka 3.7.0. Remove support for Apache Kafka 3.5.0, 3.5.1, and 3.5.2.
          * The `UseKRaft` feature gate moves to beta stage and is enabled by default. If needed, `UseKRaft` can be disabled in the feature gates configuration in the Cluster Operator.
          * Add support for ZooKeeper to KRaft migration by enabling the users to move from using ZooKeeper to store metadata to KRaft.
          * Add support for moving from dedicated controller-only KRaft nodes to mixed KRaft nodes
          * Added support for Kafka Exporter `offset.show-all` parameter
          * Prevent removal of the `broker` process role from KRaft mixed-nodes that have assigned partition-replicas
          * Improve broker scale-down prevention to continue in reconciliation when scale-down cannot be executed
          * Added support for Tiered Storage by enabling the configuration of custom storage plugins through the Kafka custom resource.
          * Update HTTP bridge to the latest 0.28.0 release
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:73deb1af0d93e9fa8955a5a412d0e7138384bdc80e7574e815d21796e446b636
        - quay.io/strimzi/kaniko-executor@sha256:3fb4ace22589fcf8d3f01b74c37b10b32d87e15d103ea59e0933b61c4e46ff4a
        - quay.io/strimzi/maven-builder@sha256:d02a9b48c77a44eeab6426c9af4adf75405217994bcb01ce0196e616041350d8
        - quay.io/strimzi/operator@sha256:2228d89b53e3524e69d1bcf5c2cd373884ce696d9005389a6c78dad23310e837
        - quay.io/strimzi/kafka@sha256:2c143eb10b66037ef877311eb767457e4b877d5ff2c47d393c76babec155caff
        - quay.io/strimzi/kafka@sha256:e2b9f2835498831bbf30ac2d651de143d57088bcb23943101b066e941db92aca
        - quay.io/strimzi/kafka@sha256:c491a95f9b9b58f406461faf222ef612ad8d2f4d7b53f7f3645b20c99d7d1583
        version: 0.40.0
      entries:
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.40.x
    - currentCSV: strimzi-cluster-operator.v0.41.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:4c687a4fac96ffc22433ca6d00fb1bd31722767b44c9d87c465f95b57fb70089
          createdAt: "2024-05-09 22:23:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.41.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.41 supports only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.41
          * Support for Apache Kafka 3.6.2
          * Metrics to monitor CA expiration
          * Support for JBOD storage in KRaft mode. (Note: JBOD support in KRaft mode is considered early access in Apache Kafka 3.7.x)
          * Support for changing topic replication factor by changing the `KafkaTopic` resources.
          * The `KafkaNodePools` feature gate moves to GA stage and is permanently enabled without the possibility to disable it. To use the Kafka Node Pool resources, you still need to use the `strimzi.io/node-pools: enabled` annotation on the `Kafka` custom resources.
          * Support for configuring the `externalIPs` field in node port type services.
          * The `UnidirectionalTopicOperator` feature gate moves to GA stage and is permanently enabled without the possibility to disable it. If the topics whose names start with `strimzi-store-topic` and `strimzi-topic-operator` still exist, you can delete them.
          * Improve validation of `KafkaMirrorMaker2` resource to make sure the Connect cluster is set to the target cluster
          * Continue reconciliation after failed manual rolling update using the `strimzi.io/manual-rolling-update` annotation (when the `ContinueReconciliationOnManualRollingUpdateFailure` feature gate is enabled)
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/maven-builder@sha256:17f09b07fc41a059e6a048cb766046d43a072cfd3a8266c97388a37543dadee8
        - quay.io/strimzi/kafka@sha256:c3a346235ce4dea42986648ca07f2482d8fc0bf654c1d441b1301a23999304db
        - quay.io/strimzi/kafka@sha256:bf49499980bbe5d4dd399de0dd396089b0fd9dd29bc46e14d39b20f1e826aec1
        - quay.io/strimzi/kafka@sha256:41811bb106d6182978a889eaedffeadffaa5ab1599c4cd8e4f5015b131c7ac30
        - quay.io/strimzi/kafka@sha256:6243d98890c665ef65b5f5fa2ffd046412281ffef31bc7a07cac37bd77b2579d
        - quay.io/strimzi/kafka-bridge@sha256:c833f930ead2d1e9054912a20689ee979d335f5e5ca8769e3c6b8a5514a0df96
        - quay.io/strimzi/operator@sha256:4c687a4fac96ffc22433ca6d00fb1bd31722767b44c9d87c465f95b57fb70089
        - quay.io/strimzi/kaniko-executor@sha256:9404037bd5431b012c8c633ec3345907667453c5532e5bfaaedc78286de22c62
        version: 0.41.0
      entries:
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.41.x
    - currentCSV: strimzi-cluster-operator.v0.42.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:ca5c1b70e1196deec4bf519921bfa55a967aa1ea15bd4c182bc56bea08d2e114
          createdAt: "2024-07-03 12:34:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.42.0/deploying.html#assembly-upgrade-str).

          **From version 0.40.0, Strimzi supports only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.42
          * Support for Apache Kafka 3.7.1
          * The UseKRaft feature gate moves to GA stage and is permanently enabled without the possibility of disabling it. To use KRaft (ZooKeeper-less Apache Kafka), you still need to use the strimzi.io/kraft: enabled annotation on the Kafka custom resources or migrate from an existing ZooKeeper-based cluster.
          * Update the base image used by Strimzi containers from UBI8 to UBI9
          * Add support for filename patterns when configuring trusted certificates
          * Enhance KafkaBridge resource with consumer inactivity timeout and HTTP consumer/producer enablement.
          * Add support for setting publishNotReadyAddresses on services for listener types other than internal.
          * Added support for configuring the quotas plugin with type strimzi or kafka in the Kafka custom resource. The Strimzi Quotas plugin version was updated to 0.3.1.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:6527c426f753521c07f3a3df2ee432273a41b109594096b4c6cae2e8e609d669
        - quay.io/strimzi/kafka@sha256:6c519531cdd5f40f326b295d8827344140efe2607a1826f7f1cab8c3427f43f7
        - quay.io/strimzi/kafka-bridge@sha256:fdddffb3de59ba48c05964bb0e760593088d24d1ee35b090112fe6cc6b553cd8
        - quay.io/strimzi/kafka@sha256:bc49c10b3e839e161e515c3a1e411d1f8df9de5e2c8d71aa7b63ccdbbae9e8d2
        - quay.io/strimzi/kaniko-executor@sha256:11821b0078a270612f1bcc37f5dc716ab273c974648c40f95fa4804352e647a7
        - quay.io/strimzi/operator@sha256:ca5c1b70e1196deec4bf519921bfa55a967aa1ea15bd4c182bc56bea08d2e114
        - quay.io/strimzi/maven-builder@sha256:342e8d0086c53cf7334127798e5bd3764b28f5e4909ba67d65f6d937dc3dba05
        - quay.io/strimzi/kafka@sha256:841a46e9c10481df3ef702412caf362b55ae07c30a3a5927073356cd36dc049f
        - quay.io/strimzi/kafka@sha256:c56c5abf94897f20f976603cba0678a33cdfbbcd6f9030a193f7684cb9894a6c
        version: 0.42.0
      entries:
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.42.x
    - currentCSV: strimzi-cluster-operator.v0.43.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:d3732245f34afaa6b571f1f912e1cd017891b22003d6367ae8ca059b2a5cb13c
          createdAt: "2024-08-22 15:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.43.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.43 is the last version with support for Kubernetes 1.23 and 1.24.** From Strimzi 0.44, only Kubernetes versions 1.25 and newer will be supported.
          ### New in 0.43
          * Support for Apache Kafka 3.8.0
          * Removed support for Apache Kafka 3.6.0, 3.6.1, and 3.6.2
          * Specifying additional volumes and volume mounts in Strimzi custom resources
          * Support for custom Cruise Control API users
          * Unregistration of KRaft nodes after scale-down
          * Kafka Exporter updated to 1.8.0 (and the Grafana dashboard updated to work with it)
          * Service account based authentication
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:80f329ecd3ba96fac12e75be874dce0debc8395337737581de34238ce11c8fb1
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:6414f8f3985f221b0eda53867c00494140108d17579c4dcbbc93934bb00e5d49
        - quay.io/strimzi/kafka@sha256:d27cf5bf3624e18ba705246b14f92d12b8a15eb8551b13abf027cf71b549baf0
        - quay.io/strimzi/kafka@sha256:ad06ead5125bce87498b1582a76e78143a6d21408ac4b12bb7e6007aa4cabd84
        - quay.io/strimzi/operator@sha256:d3732245f34afaa6b571f1f912e1cd017891b22003d6367ae8ca059b2a5cb13c
        - quay.io/strimzi/kafka@sha256:b2f4e8895e9ed66228512ce14aec685101fa75dc329128d47cb1c15e3aef3c9c
        version: 0.43.0
      entries:
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.43.x
    - currentCSV: strimzi-cluster-operator.v0.44.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:b07b81f7e282dea2e4e29a7c93cfdd911d4715a5c32fe4c4e3d7c71cba3091e8
          createdAt: "2024-08-22 15:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.44.0/deploying.html#assembly-upgrade-str).

          **From version 0.44, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**

          **ZooKeeper support will be soon removed from Apache Kafka and Strimzi.** Currently, the last Strimzi version with ZooKeeper support is expected to be Strimzi 0.45. Please plan your migration to KRaft (ZooKeeper-less Apache Kafka) accordingly. Follow the [documentation](https://strimzi.io/docs/operators/latest/full/deploying.html#assembly-kraft-mode-str) for more details.

          **Kafka Mirror Maker 1 support will be soon removed from Apache Kafka and Strimzi.** Currently, the last Strimzi version with Mirror Maker 1 support is expected to be Strimzi 0.45. Please plan your migration to Mirror Maker 2 or another mirroring tool.

          ### New in 0.44
          * Managing offsets in Connect and Mirror Maker 2 connectors through custom resources
          * Templating `advertisedHost` and `host` fields in listener configuration
          * Configuring environment variables in operands based on Secrets and ConfigMaps
          * Disabling Pod disruption Budget management
          * Auto-rebalancing at scale-up/scale-down
          * `ContinueReconciliationOnManualRollingUpdateFailure` feature gate moves to the beta stage and is enabled by default
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:9a4224635e09ece2c1d3ffe9d715813cf63037461c9f06c1b5f53554aed5ecdc
        - quay.io/strimzi/kafka-bridge@sha256:cc95ce9f496057f3b18d5e8297b40a410cc66f9fd4a2227615971458a26bac25
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:7b55f971a3b62923a9a61a61c82f542e07deeb6ffdbaede68244ba358a0853f2
        - quay.io/strimzi/operator@sha256:b07b81f7e282dea2e4e29a7c93cfdd911d4715a5c32fe4c4e3d7c71cba3091e8
        - quay.io/strimzi/kafka@sha256:238bb080825b7ecebaa819be09788bc3d056ae8162ebcd41444076dc22f85740
        - quay.io/strimzi/kafka@sha256:98d332b7fa5c9e48f60b516e00ed5812bd9268e014fd2337284e7cc3e664f1eb
        version: 0.44.0
      entries:
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.44.x
    - currentCSV: strimzi-cluster-operator.v0.45.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:58e8722216b0b0e8b1dd79109725238fa4016a16cc6a9fca0e327bccb1506810
          createdAt: "2025-06-27 01:38:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.45.1/deploying.html#assembly-upgrade-str).

          **From version 0.45, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**

          **ZooKeeper and Kafka Mirror Maker 1 support will be removed from Apache Kafka and Strimzi.** Strimzi 0.45 is the last minor Strimzi version with support for ZooKeeper-based Apache Kafka clusters and MirrorMaker 1 deployments. Please make sure to [migrate to KRaft](https://strimzi.io/docs/operators/latest/full/deploying.html#assembly-kraft-mode-str) and MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          **Strimzi EnvVar Configuration Provider and Strimzi MirrorMaker 2 Extensions will be removed from Strimzi.** Strimzi 0.45 is the last Strimzi version to include the Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0). Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.
          ### New in 0.45.1
          * Add support for Kafka 3.9.1
          * Update HTTP bridge to 0.31.2
          * Update Strimzi Kafka OAuth library to 0.15.1
          * Fixed bug which may lead to metadata loss within the cluster when restarting a KRaft migration after a previous rollback due to missing `/migration` znode deletion.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:e98fbe49cadbc5d3afbd8f17abea218f216c95d52f4fa1fe514a7a988624da24
        - quay.io/strimzi/kafka@sha256:ba52ed046b1dccdbd96f4e68057ce014d862a7c9c1fc670760c023b9aa09f23f
        - quay.io/strimzi/kafka-bridge@sha256:c462441a966f0732855e684b4840563bfad45391345c9511fa49e1630ba9c3e3
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:770e410f1e7e0e92c31bae6c19831ba36d59dd11b44e90c159aef16a965aa998
        - quay.io/strimzi/kafka@sha256:173cb5793756f75739c51a83bbe06d7f0094fd003a2982160a5cc6e216cd6d28
        - quay.io/strimzi/kafka@sha256:6a05924839a2c7da0431408e6d6223a055e75d88b3f057ac5caa7f9097227f56
        - quay.io/strimzi/operator@sha256:58e8722216b0b0e8b1dd79109725238fa4016a16cc6a9fca0e327bccb1506810
        version: 0.45.1
      entries:
      - name: strimzi-cluster-operator.v0.45.1
        version: 0.45.1
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.45.x
    - currentCSV: strimzi-cluster-operator.v0.46.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.0.0",
                      "metadataVersion": "4.0-IV3",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:6a22e4968e9d409405423d12c04a0e3fecaea2ee2794bd27de7cf21f59453265
          createdAt: "2025-06-26 11:58:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.46.0/deploying.html#assembly-upgrade-str).

          **From version 0.46, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**
          **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed. Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!
          **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images. Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.
          **Support for MirrorMaker 1 has been removed. Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.
          ### New in 0.46.1
          * Update CSV file for OLM installation to add missing permission for broker Pods to "get" nodes Kubernetes resource.
          * Allow the injection of annotations on the deployment object in Helm charts.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:3e1815aed3f74b8573cebfced66b2c78ca0cf7c746c26db2002380cd39d09ae6
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:b532234c234fc30fecec992ad1a6154c5ef08ebf9a20ec8bc62e0c88d791ff36
        - quay.io/strimzi/kafka@sha256:f3837f8cb2a60e906e84486128c41c398424f4085df2348e9e0486bfc12089fd
        - quay.io/strimzi/kafka@sha256:82273d873ce21a73f8f3118bf61627cb15734b43ea7ca92d43b290340b44e485
        - quay.io/strimzi/operator@sha256:6a22e4968e9d409405423d12c04a0e3fecaea2ee2794bd27de7cf21f59453265
        version: 0.46.1
      entries:
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.46.x
    - currentCSV: strimzi-cluster-operator.v0.47.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.0.0",
                      "metadataVersion": "4.0-IV3",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:589c4d63641d9a944462dd682f6e5febe8b38ff55b9253949b061aca16feb154
          createdAt: "2025-07-12 17:21:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.47 is the last Strimzi version with support for Kubernetes 1.25 and 1.26.**
            From Strimzi 0.48.0 on, we will support only Kubernetes 1.27 and newer.

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.47.0
          * Support for Kafka 3.9.1
          * Progress tracking for Cruise Control rebalances
          * Support for Kubernetes Image Volumes to mount custom plugins
          * Support for [JsonTemplateLayout](https://logging.apache.org/log4j/2.x/manual/json-template-layout.html) in Operators, Cruise Control, and Kafka 4.0.0
          * Support for Strimzi Metrics Reporter for the Kafka brokers / controllers
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/operator@sha256:589c4d63641d9a944462dd682f6e5febe8b38ff55b9253949b061aca16feb154
        - quay.io/strimzi/maven-builder@sha256:c67a3ad3d92420ac8b7f8435eeec53ebffaea4e894e65414a7c47c20e3faba27
        - quay.io/strimzi/kafka@sha256:978e2ed5f0d315e8bf0ab0503ee839acbd9da01ca427e6581b59ac25d36bc9c9
        - quay.io/strimzi/kafka@sha256:8a793cdf5084cdb72c6663419530985158b16330a518720e9e9d1d96babbaf7e
        - quay.io/strimzi/kafka@sha256:551882f47a6d2dc8300bcfca251fffffc3379acb2724a7efe16f17c0c4f529fb
        - quay.io/strimzi/kafka-bridge@sha256:428c5587f547dabfc995a73745569d113380c7e6b8bf0b2504c6a9a00d6c4608
        version: 0.47.0
      entries:
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.47.x
    - currentCSV: strimzi-cluster-operator.v0.48.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.1.0",
                      "metadataVersion": "4.1-IV1",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.1.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
          createdAt: "2025-09-25 10:32:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.48.0 supports only Kubernetes 1.27 and newer.**
            Kubernetes 1.25 and 1.26 are not supported anymore

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.48.0
          * Support for Kafka 4.1.0
          * Add monitoring of custom resources using kubernetes-state-metrics (KSM)
          * Support for type: custom client authentication
          * Added support for Strimzi Metrics Reporter to Kafka Connect, Mirror Maker 2 and Kafka Bridge.
          * Added support for KIP-745 in Kafka Connector and Mirror Maker 2, allowing the usage of includeTasks and onlyFailed arguments in Kafka connectors restart.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:64380efcb8da87fdc0e0de238a20456be9abb830174c2f2331882b7a84657e57
        - quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
        - quay.io/strimzi/kafka@sha256:48db05ad2e9fa39eb388089e7e25654bcbf4da1ddc3a5635d675056c28bbcb96
        - quay.io/strimzi/kafka@sha256:34450afc1a3399a9026fc3c6ec90e9f6adfec9faf6520545b4d84e90cda34964
        - quay.io/strimzi/kafka-bridge@sha256:53034f64f0b672f10b5bacea1c7a25132f118df7fd5c9032c4dbf702ed93796a
        version: 0.48.0
      entries:
      - name: strimzi-cluster-operator.v0.48.0
        version: 0.48.0
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.48.x
    defaultChannel: stable
    packageName: strimzi-kafka-operator
    provider:
      name: Strimzi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-nifi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nifi-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:05946a5e8f040e8ee46d56f2d4cca3d97e02d086e6d8224200204de4d76851fa
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:05946a5e8f040e8ee46d56f2d4cca3d97e02d086e6d8224200204de4d76851fa
        version: 24.11.1
      entries:
      - name: nifi-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: nifi-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:e5c3c4e2ad1c455fbb04a60658936bfb44c2bc9d2b616737ce1a0eba201b7a3f
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiClusterSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:e5c3c4e2ad1c455fbb04a60658936bfb44c2bc9d2b616737ce1a0eba201b7a3f
        version: 25.3.0
      entries:
      - name: nifi-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: nifi-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiClusterSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
        version: 25.7.0
      entries:
      - name: nifi-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: nifi-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiClusterSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
        version: 25.7.0
      entries:
      - name: nifi-operator.v25.7.0
        version: 25.7.0
      - name: nifi-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-nifi-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/en/products/edge/validated-patterns
    name: patterns-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: patterns-operator.v0.0.62
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gitops.hybrid-cloud-patterns.io/v1alpha1",
                "kind": "Pattern",
                "metadata": {
                  "name": "pattern-sample"
                },
                "spec": {
                  "clusterGroupName": "hub",
                  "gitSpec": {
                    "targetRepo": "https://github.com/validatedpatterns/multicloud-gitops",
                    "targetRevision": "main"
                  },
                  "multiSourceConfig": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/hybridcloudpatterns/patterns-operator:0.0.62
          createdAt: "2025-09-02 16:14:40"
          description: "An operator to deploy and manage architecture patterns from
            https://validatedpatterns.io. \nThis operator collects some Analytics
            like cluster version, cloud type, etc. To disable this\nadd \"ANALYTICS:
            false\" to the environment variables in the subscription."
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/validatedpatterns/patterns-operator
          support: validatedpatterns@googlegroups.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pattern is the Schema for the patterns API
            displayName: Pattern
            kind: Pattern
            name: patterns.gitops.hybrid-cloud-patterns.io
            version: v1alpha1
        description: Deploys and manages architecture patterns from https://validatedpatterns.io
        displayName: Validated Patterns Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Validated
        - Pattern
        - Blueprint
        - Architecture
        links:
        - name: Hybrid Cloud Patterns
          url: https://validatedpatterns.io
        maintainers:
        - email: validatedpatterns@googlegroups.com
          name: patterns-team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/en/products/edge/validated-patterns
        relatedImages:
        - quay.io/hybridcloudpatterns/kube-rbac-proxy@sha256:409a464692c7db3cf68893376479dca1b53802412237637725927f7e78aa0844
        - quay.io/hybridcloudpatterns/patterns-operator@sha256:15ed44938b112bcf9b3a9a566749b39ba2b9053a6a8b9b499ef247977c18496c
        version: 0.0.62
      entries:
      - name: patterns-operator.v0.0.62
        version: 0.0.62
      - name: patterns-operator.v0.0.61
        version: 0.0.61
      - name: patterns-operator.v0.0.60
        version: 0.0.60
      - name: patterns-operator.v0.0.59
        version: 0.0.59
      - name: patterns-operator.v0.0.58
        version: 0.0.58
      - name: patterns-operator.v0.0.57
        version: 0.0.57
      - name: patterns-operator.v0.0.56
        version: 0.0.56
      - name: patterns-operator.v0.0.55
        version: 0.0.55
      - name: patterns-operator.v0.0.54
        version: 0.0.54
      - name: patterns-operator.v0.0.53
        version: 0.0.53
      - name: patterns-operator.v0.0.52
        version: 0.0.52
      - name: patterns-operator.v0.0.51
        version: 0.0.51
      - name: patterns-operator.v0.0.50
        version: 0.0.50
      - name: patterns-operator.v0.0.49
        version: 0.0.49
      - name: patterns-operator.v0.0.48
        version: 0.0.48
      - name: patterns-operator.v0.0.47
        version: 0.0.47
      - name: patterns-operator.v0.0.46
        version: 0.0.46
      - name: patterns-operator.v0.0.45
        version: 0.0.45
      - name: patterns-operator.v0.0.44
        version: 0.0.44
      - name: patterns-operator.v0.0.43
        version: 0.0.43
      - name: patterns-operator.v0.0.42
        version: 0.0.42
      - name: patterns-operator.v0.0.41
        version: 0.0.41
      - name: patterns-operator.v0.0.40
        version: 0.0.40
      - name: patterns-operator.v0.0.39
        version: 0.0.39
      - name: patterns-operator.v0.0.38
        version: 0.0.38
      - name: patterns-operator.v0.0.37
        version: 0.0.37
      - name: patterns-operator.v0.0.36
        version: 0.0.36
      - name: patterns-operator.v0.0.35
        version: 0.0.35
      - name: patterns-operator.v0.0.34
        version: 0.0.34
      - name: patterns-operator.v0.0.33
        version: 0.0.33
      - name: patterns-operator.v0.0.32
        version: 0.0.32
      - name: patterns-operator.v0.0.31
        version: 0.0.31
      - name: patterns-operator.v0.0.30
        version: 0.0.30
      - name: patterns-operator.v0.0.29
        version: 0.0.29
      - name: patterns-operator.v0.0.28
        version: 0.0.28
      - name: patterns-operator.v0.0.27
        version: 0.0.27
      - name: patterns-operator.v0.0.26
        version: 0.0.26
      - name: patterns-operator.v0.0.25
        version: 0.0.25
      - name: patterns-operator.v0.0.24
        version: 0.0.24
      - name: patterns-operator.v0.0.23
        version: 0.0.23
      - name: patterns-operator.v0.0.22
        version: 0.0.22
      - name: patterns-operator.v0.0.21
        version: 0.0.21
      - name: patterns-operator.v0.0.20
        version: 0.0.20
      - name: patterns-operator.v0.0.19
        version: 0.0.19
      - name: patterns-operator.v0.0.18
        version: 0.0.18
      - name: patterns-operator.v0.0.17
        version: 0.0.17
      - name: patterns-operator.v0.0.16
        version: 0.0.16
      - name: patterns-operator.v0.0.15
        version: 0.0.15
      - name: patterns-operator.v0.0.14
        version: 0.0.14
      - name: patterns-operator.v0.0.13
        version: 0.0.13
      - name: patterns-operator.v0.0.12
        version: 0.0.12
      - name: patterns-operator.v0.0.11
        version: 0.0.11
      - name: patterns-operator.v0.0.10
        version: 0.0.10
      - name: patterns-operator.v0.0.9
        version: 0.0.9
      - name: patterns-operator.v0.0.8
        version: 0.0.8
      - name: patterns-operator.v0.0.7
        version: 0.0.7
      - name: patterns-operator.v0.0.6
        version: 0.0.6
      - name: patterns-operator.v0.0.5
        version: 0.0.5
      - name: patterns-operator.v0.0.4
        version: 0.0.4
      - name: patterns-operator.v0.0.3
        version: 0.0.3
      - name: patterns-operator.v0.0.2
        version: 0.0.2
      - name: patterns-operator.v0.0.1
        version: 0.0.1
      name: fast
    defaultChannel: fast
    packageName: patterns-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com/en/products/edge/validated-patterns
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sns-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sns-controller.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "PlatformApplication",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "PlatformEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sns-controller:1.2.2
          createdAt: "2025-09-30T21:19:34Z"
          description: AWS SNS controller is a service controller for managing SNS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PlatformApplication represents the state of an AWS sns PlatformApplication
              resource.
            displayName: PlatformApplication
            kind: PlatformApplication
            name: platformapplications.sns.services.k8s.aws
            version: v1alpha1
          - description: PlatformEndpoint represents the state of an AWS sns PlatformEndpoint
              resource.
            displayName: PlatformEndpoint
            kind: PlatformEndpoint
            name: platformendpoints.sns.services.k8s.aws
            version: v1alpha1
          - description: Subscription represents the state of an AWS sns Subscription
              resource.
            displayName: Subscription
            kind: Subscription
            name: subscriptions.sns.services.k8s.aws
            version: v1alpha1
          - description: Topic represents the state of an AWS sns Topic resource.
            displayName: Topic
            kind: Topic
            name: topics.sns.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Simple Notification Service (SNS) resources in AWS from within your Kubernetes cluster.

          **About Amazon SNS**

          Amazon Simple Notification Service (Amazon SNS) is a managed service that provides message delivery from publishers to subscribers (also known as producers and consumers). Publishers communicate asynchronously with subscribers by sending messages to a topic, which is a logical access point and communication channel. Clients can subscribe to the SNS topic and receive published messages using a supported endpoint type, such as Amazon Kinesis Data Firehose, Amazon SQS, AWS Lambda, HTTP, email, mobile push notifications, and mobile text messages (SMS).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SNS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sns
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SNS Developer Resources
          url: https://aws.amazon.com/sns/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sns maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sns-controller:1.2.2
        version: 1.2.2
      entries:
      - name: ack-sns-controller.v1.2.2
        version: 1.2.2
      - name: ack-sns-controller.v1.2.1
        version: 1.2.1
      - name: ack-sns-controller.v1.2.0
        version: 1.2.0
      - name: ack-sns-controller.v1.1.16
        version: 1.1.16
      - name: ack-sns-controller.v1.1.15
        version: 1.1.15
      - name: ack-sns-controller.v1.1.14
        version: 1.1.14
      - name: ack-sns-controller.v1.1.13
        version: 1.1.13
      - name: ack-sns-controller.v1.1.12
        version: 1.1.12
      - name: ack-sns-controller.v1.1.11
        version: 1.1.11
      - name: ack-sns-controller.v1.1.10
        version: 1.1.10
      - name: ack-sns-controller.v1.1.9
        version: 1.1.9
      - name: ack-sns-controller.v1.1.8
        version: 1.1.8
      - name: ack-sns-controller.v1.1.7
        version: 1.1.7
      - name: ack-sns-controller.v1.1.6
        version: 1.1.6
      - name: ack-sns-controller.v1.1.5
        version: 1.1.5
      - name: ack-sns-controller.v1.1.4
        version: 1.1.4
      - name: ack-sns-controller.v1.1.3
        version: 1.1.3
      - name: ack-sns-controller.v1.1.2
        version: 1.1.2
      - name: ack-sns-controller.v1.1.1
        version: 1.1.1
      - name: ack-sns-controller.v1.1.0
        version: 1.1.0
      - name: ack-sns-controller.v1.0.14
        version: 1.0.14
      - name: ack-sns-controller.v1.0.13
        version: 1.0.13
      - name: ack-sns-controller.v1.0.12
        version: 1.0.12
      - name: ack-sns-controller.v1.0.11
        version: 1.0.11
      - name: ack-sns-controller.v1.0.9
        version: 1.0.9
      - name: ack-sns-controller.v1.0.6
        version: 1.0.6
      - name: ack-sns-controller.v1.0.5
        version: 1.0.5
      - name: ack-sns-controller.v1.0.4
        version: 1.0.4
      - name: ack-sns-controller.v1.0.3
        version: 1.0.3
      - name: ack-sns-controller.v1.0.2
        version: 1.0.2
      - name: ack-sns-controller.v1.0.1
        version: 1.0.1
      - name: ack-sns-controller.v1.0.0
        version: 1.0.0
      - name: ack-sns-controller.v0.0.12
        version: 0.0.12
      - name: ack-sns-controller.v0.0.11
        version: 0.0.11
      - name: ack-sns-controller.v0.0.10
        version: 0.0.10
      - name: ack-sns-controller.v0.0.9
        version: 0.0.9
      - name: ack-sns-controller.v0.0.8
        version: 0.0.8
      - name: ack-sns-controller.v0.0.7
        version: 0.0.7
      - name: ack-sns-controller.v0.0.5
        version: 0.0.5
      - name: ack-sns-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-sns-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Helm Community
      provider-url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
    name: cockroachdb
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroachdb.v5.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1alpha1",
                "kind": "Cockroachdb",
                "metadata": {
                  "name": "cockroachdb-sample"
                },
                "spec": {
                  "clusterDomain": "cluster.local",
                  "conf": {
                    "attrs": [],
                    "cache": "25%",
                    "cluster-name": "",
                    "disable-cluster-name-verification": false,
                    "http-port": 8080,
                    "join": [],
                    "locality": "",
                    "logtostderr": "INFO",
                    "max-sql-memory": "25%",
                    "port": 26257,
                    "single-node": false,
                    "sql-audit-dir": ""
                  },
                  "image": {
                    "credentials": {},
                    "pullPolicy": "IfNotPresent",
                    "repository": "cockroachdb/cockroach",
                    "tag": "v20.2.4"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [],
                    "labels": {},
                    "paths": [
                      "/"
                    ],
                    "tls": []
                  },
                  "init": {
                    "affinity": {},
                    "annotations": {},
                    "labels": {
                      "app.kubernetes.io/component": "init"
                    },
                    "nodeSelector": {},
                    "resources": {},
                    "tolerations": []
                  },
                  "labels": {},
                  "networkPolicy": {
                    "enabled": false,
                    "ingress": {
                      "grpc": [],
                      "http": []
                    }
                  },
                  "service": {
                    "discovery": {
                      "annotations": {},
                      "labels": {
                        "app.kubernetes.io/component": "cockroachdb"
                      }
                    },
                    "ports": {
                      "grpc": {
                        "external": {
                          "name": "grpc",
                          "port": 26257
                        },
                        "internal": {
                          "name": "grpc-internal",
                          "port": 26257
                        }
                      },
                      "http": {
                        "name": "http",
                        "port": 8080
                      }
                    },
                    "public": {
                      "annotations": {},
                      "labels": {
                        "app.kubernetes.io/component": "cockroachdb"
                      },
                      "type": "ClusterIP"
                    }
                  },
                  "serviceMonitor": {
                    "annotations": {},
                    "enabled": false,
                    "interval": "10s",
                    "labels": {}
                  },
                  "statefulset": {
                    "annotations": {},
                    "args": [],
                    "budget": {
                      "maxUnavailable": 1
                    },
                    "env": [],
                    "labels": {
                      "app.kubernetes.io/component": "cockroachdb"
                    },
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "podAffinity": {},
                    "podAntiAffinity": {
                      "topologyKey": "kubernetes.io/hostname",
                      "type": "soft",
                      "weight": 100
                    },
                    "podManagementPolicy": "Parallel",
                    "priorityClassName": "",
                    "replicas": 3,
                    "resources": {},
                    "secretMounts": [],
                    "tolerations": [],
                    "topologySpreadConstraints": {
                      "maxSkew": 1,
                      "topologyKey": "topology.kubernetes.io/zone",
                      "whenUnsatisfiable": "ScheduleAnyway"
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    }
                  },
                  "storage": {
                    "hostPath": "",
                    "persistentVolume": {
                      "annotations": {},
                      "enabled": true,
                      "labels": {},
                      "size": "100Gi",
                      "storageClass": ""
                    }
                  },
                  "tls": {
                    "certs": {
                      "clientRootSecret": "cockroachdb-root",
                      "nodeSecret": "cockroachdb-node",
                      "provided": false,
                      "tlsSecret": false
                    },
                    "enabled": false,
                    "init": {
                      "image": {
                        "credentials": {},
                        "pullPolicy": "IfNotPresent",
                        "repository": "cockroachdb/cockroach-k8s-request-cert",
                        "tag": "0.4"
                      }
                    },
                    "serviceAccount": {
                      "create": true,
                      "name": ""
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          certified: "false"
          containerImage: quay.io/helmoperators/cockroachdb:v5.0.4
          createdAt: 2021-01-23T10-00-00Z
          description: CockroachDB Operator based on the CockroachDB helm chart
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/dmesser/cockroachdb-operator
          support: dmesser
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a CockroachDB cluster
            displayName: CockroachDB
            kind: Cockroachdb
            name: cockroachdbs.charts.operatorhub.io
            version: v1alpha1
        description: |
          CockroachDB is a scalable, survivable, strongly-consistent SQL database.
          ## About this Operator
          This Operator is based on a Helm chart for CockroachDB. It supports the same parameters like the official [CockroachDB Helm Chart](https://github.com/cockroachdb/helm-charts/tree/master/cockroachdb).

          ## Core capabilities
          * **StatefulSet** - Sets up a dynamically scalable CockroachDB cluster using a Kubernetes StatefulSet
          * **Expand Replicas** - Supports expanding the set of replicas by simply editing your object
          * **Dashboard** - Installs the CockroachDB user interface to administer your cluster. Easily expose it via an Ingress rule.
          Review all of the [confiuguration options](https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration) to best run your database instance. The example configuration is derived from the chart's [`values.yaml`](https://github.com/helm/charts/blob/master/stable/cockroachdb/values.yaml).
          ## Using the cluster
          The resulting cluster endpoint can be consumed from a `Service` that follows the pattern: `<StatefulSet-name>-public`. For example to connect using the command line client, use something like the following to obtain the name of the service:
          ```
          kubectl get service -l chart=cockroachdb-5.0.4
          NAME                                           TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)              AGE
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb          ClusterIP   None             <none>        26257/TCP,8080/TCP   24m
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public   ClusterIP   10.106.249.134   <none>        26257/TCP,8080/TCP   24m
          ```
          Then you can use the CockroachDB command line client to connect to the database cluster:
          ```
          kubectl run -it --rm cockroach-client --image=cockroachdb/cockroach --restart=Never --command -- ./cockroach sql --insecure --host example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public
          ```
          ## Before you start
          This Operator requires a cluster with PV support in order to run correctly.

          Note that for a production cluster, you will likely want to override the following parameters in [`values.yaml`](values.yaml) with your own values.

          - `statefulset.resources.requests.memory` and `statefulset.resources.limits.memory` allocate memory resources to CockroachDB pods in your cluster.
          - `conf.cache` and `conf.max-sql-memory` are memory limits that we recommend setting to 1/4 of the above resource allocation. When running CockroachDB, you must set these limits explicitly to avoid running out of memory.
          - `storage.persistentVolume.size` defaults to `100Gi` of disk space per pod, which you may increase or decrease for your use case.
          - `storage.persistentVolume.storageClass` uses the default storage class for your environment. We strongly recommend that you specify a storage class which uses an SSD.
          - `tls.enabled` must be set to `yes`/`true` to deploy in secure mode.

          For more information on overriding the `values.yaml` parameters, please see:
          > <https://www.cockroachlabs.com/docs/stable/orchestrate-cockroachdb-with-kubernetes.html#step-2-start-cockroachdb>

          If you are running in secure mode (with configuration parameter `tls.enabled` set to `yes`/`true`) and `tls.certs.provided` set to `no`/`false`), then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          When `tls.certs.provided` is set to `yes`/`true`, this chart will use certificates created outside of Kubernetes. You may want to use this if you want to use a different certificate authority from the one being used by Kubernetes or if your Kubernetes cluster doesn't fully support certificate-signing requests. To use this, first set up your certificates and load them into your Kubernetes cluster as Secrets using the commands below:

          ```
          mkdir certs
          mkdir my-safe-directory
          cockroach cert create-ca --certs-dir=certs --ca-key=my-safe-directory/ca.key
          cockroach cert create-client root --certs-dir=certs --ca-key=my-safe-directory/ca.key
          kubectl create secret generic cockroachdb-root --from-file=certs
          cockroach cert create-node --certs-dir=certs --ca-key=my-safe-directory/ca.key localhost 127.0.0.1 eerie-horse-cockroachdb-public eerie-horse-cockroachdb-public.default eerie-horse-cockroachdb-public.default.svc.cluster.local *.eerie-horse-cockroachdb *.eerie-horse-cockroachdb.default *.eerie-horse-cockroachdb.default.svc.cluster.local
          kubectl create secret generic cockroachdb-node --from-file=certs
          ```

          Set `tls.certs.tlsSecret` to `yes/true` if you make use of [cert-manager][3] in your cluster.

          [cert-manager][3] stores generated certificates in dedicated TLS secrets. Thus, they are always named:

          * `ca.crt`
          * `tls.crt`
          * `tls.key`

          On the other hand, CockroachDB also demands dedicated certificate filenames:

          * `ca.crt`
          * `node.crt`
          * `node.key`
          * `client.root.crt`
          * `client.root.key`

          By activating `tls.certs.tlsSecret` we benefit from projected secrets and convert the TLS secret filenames to their according CockroachDB filenames.

          If you are running in secure mode, then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          Confirm that all pods are `Running` successfully and init has been completed:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS      RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running     0          1m
          my-release-cockroachdb-1            1/1       Running     0          1m
          my-release-cockroachdb-2            1/1       Running     0          1m
          my-release-cockroachdb-init-k6jcr   0/1       Completed   0          1m
          ```

          Confirm that persistent volumes are created and claimed for each pod:

          ```shell
          kubectl get pv
          ```

          ```
          NAME                                       CAPACITY   ACCESS MODES   RECLAIM POLICY   STATUS    CLAIM                                      STORAGECLASS   REASON    AGE
          pvc-64878ebf-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-0   standard                 51s
          pvc-64945b4f-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-1   standard                 51s
          pvc-649d920d-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-2   standard                 51s
          ```

          ## Upgrading the cluster

          ### Chart version 3.0.0 and after

          Launch a temporary interactive pod and start the built-in SQL client:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          Set `cluster.preserve_downgrade_option`, where `$current_version` is the CockroachDB version currently running (e.g., `19.2`):

          ```sql
          > SET CLUSTER SETTING cluster.preserve_downgrade_option = '$current_version';
          ```

          Exit the shell and delete the temporary pod:

          ```sql
          > \q
          ```

          Kick off the upgrade process by changing the new Docker image, where `$new_version` is the CockroachDB version to which you are upgrading:

          ```shell
          kubectl delete job my-release-cockroachdb-init
          ```

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"image":{"tag":"<new version tag here>"}}}' --type merge
          ```

          Kubernetes will carry out a safe [rolling upgrade](https://kubernetes.io/docs/tutorials/stateful-application/basic-stateful-set/#updating-statefulsets) of your CockroachDB nodes one-by-one. Monitor the cluster's pods until all have been successfully restarted:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS              RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running             0          2m
          my-release-cockroachdb-1            1/1       Running             0          3m
          my-release-cockroachdb-2            1/1       Running             0          3m
          my-release-cockroachdb-3            0/1       ContainerCreating   0          25s
          my-release-cockroachdb-init-nwjkh   0/1       ContainerCreating   0          6s
          ```

          ```shell
          kubectl get pods \
          -o jsonpath='{range .items[*]}{.metadata.name}{"\t"}{.spec.containers[0].image}{"\n"}'
          ```

          ```
          my-release-cockroachdb-0    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-1    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-2    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-3    cockroachdb/cockroach:v20.2.4
          ```

          Resume normal operations. Once you are comfortable that the stability and performance of the cluster is what you'd expect post-upgrade, finalize the upgrade:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          ```sql
          > RESET CLUSTER SETTING cluster.preserve_downgrade_option;
          > \q
          ```

          ## Configuration

          The following table lists the configurable parameters of the CockroachDB chart and their default values.
          For details see the [`values.yaml`](values.yaml) file.

          | Parameter                                                 | Description                                                     | Default                                          |
          | ---------                                                 | -----------                                                     | -------                                          |
          | `clusterDomain`                                           | Cluster's default DNS domain                                    | `cluster.local`                                  |
          | `conf.attrs`                                              | CockroachDB node attributes                                     | `[]`                                             |
          | `conf.cache`                                              | Size of CockroachDB's in-memory cache                           | `25%`                                            |
          | `conf.cluster-name`                                       | Name of CockroachDB cluster                                     | `""`                                             |
          | `conf.disable-cluster-name-verification`                  | Disable CockroachDB cluster name verification                   | `no`                                             |
          | `conf.join`                                               | List of already-existing CockroachDB instances                  | `[]`                                             |
          | `conf.max-disk-temp-storage`                              | Max storage capacity for temp data                              | `0`                                              |
          | `conf.max-offset`                                         | Max allowed clock offset for CockroachDB cluster                | `500ms`                                          |
          | `conf.max-sql-memory`                                     | Max memory to use processing SQL querie                         | `25%`                                            |
          | `conf.locality`                                           | Locality attribute for this deployment                          | `""`                                             |
          | `conf.single-node`                                        | Disable CockroachDB clustering (standalone mode)                | `no`                                             |
          | `conf.sql-audit-dir`                                      | Directory for SQL audit log                                     | `""`                                             |
          | `conf.port`                                               | CockroachDB primary serving port in Pods                        | `26257`                                          |
          | `conf.http-port`                                          | CockroachDB HTTP port in Pods                                   | `8080`                                           |
          | `image.repository`                                        | Container image name                                            | `cockroachdb/cockroach`                          |
          | `image.tag`                                               | Container image tag                                             | `v20.2.4`                                        |
          | `image.pullPolicy`                                        | Container pull policy                                           | `IfNotPresent`                                   |
          | `image.credentials`                                       | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `statefulset.replicas`                                    | StatefulSet replicas number                                     | `3`                                              |
          | `statefulset.updateStrategy`                              | Update strategy for StatefulSet Pods                            | `{"type": "RollingUpdate"}`                      |
          | `statefulset.podManagementPolicy`                         | `OrderedReady`/`Parallel` Pods creation/deletion order          | `Parallel`                                       |
          | `statefulset.budget.maxUnavailable`                       | k8s PodDisruptionBudget parameter                               | `1`                                              |
          | `statefulset.args`                                        | Extra command-line arguments                                    | `[]`                                             |
          | `statefulset.env`                                         | Extra env vars                                                  | `[]`                                             |
          | `statefulset.secretMounts`                                | Additional Secrets to mount at cluster members                  | `[]`                                             |
          | `statefulset.labels`                                      | Additional labels of StatefulSet and its Pods                   | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `statefulset.annotations`                                 | Additional annotations of StatefulSet Pods                      | `{}`                                             |
          | `statefulset.nodeAffinity`                                | [Node affinity rules][2] of StatefulSet Pods                    | `{}`                                             |
          | `statefulset.podAffinity`                                 | [Inter-Pod affinity rules][1] of StatefulSet Pods               | `{}`                                             |
          | `statefulset.podAntiAffinity`                             | [Anti-affinity rules][1] of StatefulSet Pods                    | auto                                             |
          | `statefulset.podAntiAffinity.topologyKey`                 | The topologyKey for auto [anti-affinity rules][1]               | `kubernetes.io/hostname`                         |
          | `statefulset.podAntiAffinity.type`                        | Type of auto [anti-affinity rules][1]                           | `soft`                                           |
          | `statefulset.podAntiAffinity.weight`                      | Weight for `soft` auto [anti-affinity rules][1]                 | `100`                                            |
          | `statefulset.nodeSelector`                                | Node labels for StatefulSet Pods assignment                     | `{}`                                             |
          | `statefulset.priorityClassName`                           | [PriorityClassName][4] for StatefulSet Pods                     | `""`                                             |
          | `statefulset.tolerations`                                 | Node taints to tolerate by StatefulSet Pods                     | `[]`                                             |
          | `statefulset.topologySpreadConstraints`                   | [Topology Spread Constraints rules][5] of StatefulSet Pods      | auto                                             |
          | `statefulset.topologySpreadConstraints.maxSkew`           | Degree to which Pods may be unevenly distributed                | `1`                                              |
          | `statefulset.topologySpreadConstraints.topologyKey`       | The key of node labels                                          | `topology.kubernetes.io/zone`                    |
          | `statefulset.topologySpreadConstraints.whenUnsatisfiable` | `ScheduleAnyway`/`DoNotSchedule` for unsatisfiable constraints  | `ScheduleAnyway`                                 |
          | `statefulset.resources`                                   | Resource requests and limits for StatefulSet Pods               | `{}`                                             |
          | `service.ports.grpc.external.port`                        | CockroachDB primary serving port in Services                    | `26257`                                          |
          | `service.ports.grpc.external.name`                        | CockroachDB primary serving port name in Services               | `grpc`                                           |
          | `service.ports.grpc.internal.port`                        | CockroachDB inter-communication port in Services                | `26257`                                          |
          | `service.ports.grpc.internal.name`                        | CockroachDB inter-communication port name in Services           | `grpc-internal`                                  |
          | `service.ports.http.port`                                 | CockroachDB HTTP port in Services                               | `8080`                                           |
          | `service.ports.http.name`                                 | CockroachDB HTTP port name in Services                          | `http`                                           |
          | `service.public.type`                                     | Public Service type                                             | `ClusterIP`                                      |
          | `service.public.labels`                                   | Additional labels of public Service                             | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.public.annotations`                              | Additional annotations of public Service                        | `{}`                                             |
          | `service.discovery.labels`                                | Additional labels of discovery Service                          | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.discovery.annotations`                           | Additional annotations of discovery Service                     | `{}`                                             |
          | `ingress.enabled`                                         | Enable ingress resource for CockroachDB                         | `false`                                          |
          | `ingress.labels`                                          | Additional labels of Ingress                                    | `{}`                                             |
          | `ingress.annotations`                                     | Additional annotations of Ingress                               | `{}`                                             |
          | `ingress.paths`                                           | Paths for the default host                                      | `[/]`                                            |
          | `ingress.hosts`                                           | CockroachDB Ingress hostnames                                   | `[]`                                             |
          | `ingress.tls[0].hosts`                                    | CockroachDB Ingress tls hostnames                               | `nil`                                            |
          | `ingress.tls[0].secretName`                               | CockroachDB Ingress tls secret name                             | `nil`                                            |
          | `serviceMonitor.enabled`                                  | Create [ServiceMonitor](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/design.md#servicemonitor) Resource for scraping metrics using [PrometheusOperator](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/user-guides/getting-started.md#prometheus-operator)                     | `false`                                             |
          | `serviceMonitor.labels`                                   | Additional labels of ServiceMonitor                             | `{}`                                             |
          | `serviceMonitor.annotations`                              | Additional annotations of ServiceMonitor                        | `{}`                                             |
          | `serviceMonitor.interval`                                 | ServiceMonitor scrape metrics interval                          | `10s`                                            |
          | `serviceMonitor.scrapeTimeout`                            | ServiceMonitor scrape timeout                                   | `nil`                                            |
          | `storage.hostPath`                                        | Absolute path on host to store data                             | `""`                                             |
          | `storage.persistentVolume.enabled`                        | Whether to use PersistentVolume to store data                   | `yes`                                            |
          | `storage.persistentVolume.size`                           | PersistentVolume size                                           | `100Gi`                                          |
          | `storage.persistentVolume.storageClass`                   | PersistentVolume class                                          | `""`                                             |
          | `storage.persistentVolume.labels`                         | Additional labels of PersistentVolumeClaim                      | `{}`                                             |
          | `storage.persistentVolume.annotations`                    | Additional annotations of PersistentVolumeClaim                 | `{}`                                             |
          | `init.labels`                                             | Additional labels of init Job and its Pod                       | `{"app.kubernetes.io/component": "init"}`        |
          | `init.annotations`                                        | Additional labels of the Pod of init Job                        | `{}`                                             |
          | `init.affinity`                                           | [Affinity rules][2] of init Job Pod                             | `{}`                                             |
          | `init.nodeSelector`                                       | Node labels for init Job Pod assignment                         | `{}`                                             |
          | `init.tolerations`                                        | Node taints to tolerate by init Job Pod                         | `[]`                                             |
          | `init.resources`                                          | Resource requests and limits for the Pod of init Job            | `{}`                                             |
          | `tls.enabled`                                             | Whether to run securely using TLS certificates                  | `no`                                             |
          | `tls.serviceAccount.create`                               | Whether to create a new RBAC service account                    | `yes`                                            |
          | `tls.serviceAccount.name`                                 | Name of RBAC service account to use                             | `""`                                             |
          | `tls.certs.provided`                                      | Bring your own certs scenario, i.e certificates are provided    | `no`                                             |
          | `tls.certs.clientRootSecret`                              | If certs are provided, secret name for client root cert         | `cockroachdb-root`                               |
          | `tls.certs.nodeSecret`                                    | If certs are provided, secret name for node cert                | `cockroachdb-node`                               |
          | `tls.certs.tlsSecret`                                     | Own certs are stored in TLS secret                              | `no`                                             |
          | `tls.init.image.repository`                               | Image to use for requesting TLS certificates                    | `cockroachdb/cockroach-k8s-request-cert`         |
          | `tls.init.image.tag`                                      | Image tag to use for requesting TLS certificates                | `0.4`                                            |
          | `tls.init.image.pullPolicy`                               | Requesting TLS certificates container pull policy               | `IfNotPresent`                                   |
          | `tls.init.image.credentials`                              | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `networkPolicy.enabled`                                   | Enable NetworkPolicy for CockroachDB's Pods                     | `no`                                             |
          | `networkPolicy.ingress.grpc`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |
          | `networkPolicy.ingress.http`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |


          Override the default parameters using the `CockroachDB` custom resource.

          ## Deep dive

          ### Connecting to the CockroachDB cluster

          Once you've created the cluster, you can start talking to it by connecting to its `-public` Service. CockroachDB is PostgreSQL wire protocol compatible, so there's a [wide variety of supported clients](https://www.cockroachlabs.com/docs/install-client-drivers.html). As an example, we'll open up a SQL shell using CockroachDB's built-in shell and play around with it a bit, like this (likely needing to replace `my-release-cockroachdb-public` with the name of the `-public` Service that was created with your installed chart):

          ```shell
          kubectl run cockroach-client --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host my-release-cockroachdb-public
          ```

          ```
          Waiting for pod default/cockroach-client to be running, status is Pending,
          pod ready: false
          If you don't see a command prompt, try pressing enter.
          root@my-release-cockroachdb-public:26257> SHOW DATABASES;
          +--------------------+
          |      Database      |
          +--------------------+
          | information_schema |
          | pg_catalog         |
          | system             |
          +--------------------+
          (3 rows)
          root@my-release-cockroachdb-public:26257> CREATE DATABASE bank;
          CREATE DATABASE
          root@my-release-cockroachdb-public:26257> CREATE TABLE bank.accounts (id INT
          PRIMARY KEY, balance DECIMAL);
          CREATE TABLE
          root@my-release-cockroachdb-public:26257> INSERT INTO bank.accounts VALUES
          (1234, 10000.50);
          INSERT 1
          root@my-release-cockroachdb-public:26257> SELECT * FROM bank.accounts;
          +------+---------+
          |  id  | balance |
          +------+---------+
          | 1234 | 10000.5 |
          +------+---------+
          (1 row)
          root@my-release-cockroachdb-public:26257> \q
          Waiting for pod default/cockroach-client to terminate, status is Running
          pod "cockroach-client" deleted
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          ### Cluster health

          Because our pod spec includes regular health checks of the CockroachDB processes, simply running `kubectl get pods` and looking at the `STATUS` column is sufficient to determine the health of each instance in the cluster.

          If you want more detailed information about the cluster, the best place to look is the Admin UI.

          ### Accessing the Admin UI

          If you want to see information about how the cluster is doing, you can try pulling up the CockroachDB Admin UI by port-forwarding from your local machine to one of the pods (replacing `my-release-cockroachdb-0` with the name of one of your pods:

          ```shell
          kubectl port-forward my-release-cockroachdb-0 8080
          ```

          You should then be able to access the Admin UI by visiting <http://localhost:8080/> in your web browser.

          ### Failover

          If any CockroachDB member fails, it is restarted or recreated automatically by the Kubernetes infrastructure, and will re-join the cluster automatically when it comes back up. You can test this scenario by killing any of the CockroachDB pods:

          ```shell
          kubectl delete pod my-release-cockroachdb-1
          ```

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          After a while:

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-1  1/1       Running       0          20s
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          You can check the state of re-joining from the new pod's logs:

          ```shell
          kubectl logs my-release-cockroachdb-1
          ```

          ```
          [...]
          I161028 19:32:09.754026 1 server/node.go:586  [n1] node connected via gossip and
          verified as part of cluster {"35ecbc27-3f67-4e7d-9b8f-27c31aae17d6"}
          [...]
          cockroachdb-0.my-release-cockroachdb.default.svc.cluster.local:26257
          build:      beta-20161027-55-gd2d3c7f @ 2016/10/28 19:27:25 (go1.7.3)
          admin:      http://0.0.0.0:8080
          sql:
          postgresql://root@my-release-cockroachdb-1.my-release-cockroachdb.default.svc.cluster.local:26257?sslmode=disable
          logs:       cockroach-data/logs
          store[0]:   path=cockroach-data
          status:     restarted pre-existing node
          clusterID:  {35ecbc27-3f67-4e7d-9b8f-27c31aae17d6}
          nodeID:     2
          [...]
          ```

          ### NetworkPolicy

          To enable NetworkPolicy for CockroachDB, install [a networking plugin that implements the Kubernetes NetworkPolicy spec](https://kubernetes.io/docs/tasks/administer-cluster/declare-network-policy#before-you-begin), and set `networkPolicy.enabled` to `yes`/`true`.

          For Kubernetes v1.5 & v1.6, you must also turn on NetworkPolicy by setting the `DefaultDeny` Namespace annotation. Note: this will enforce policy for _all_ pods in the Namespace:

          ```shell
          kubectl annotate namespace default "net.beta.kubernetes.io/network-policy={\"ingress\":{\"isolation\":\"DefaultDeny\"}}"
          ```

          For more precise policy, set `networkPolicy.ingress.grpc` and `networkPolicy.ingress.http` rules. This will only allow pods that match the provided rules to connect to CockroachDB.

          ### Scaling

          Scaling should be managed via the `helm upgrade` command. After resizing your cluster on your cloud environment (e.g., GKE or EKS), run the following command to add a pod. This assumes you scaled from 3 to 4 nodes:

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"statefulset":{"replicas":"4"}}}' --type merge
          ```

          Note, that if you are running in secure mode (`tls.enabled` is `yes`/`true`) and increase the size of your cluster, you will also have to approve the CSR (certificate-signing request) of each new node (using `kubectl get csr` and `kubectl certificate approve`).

          [1]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [2]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#node-affinity
          [3]: https://cert-manager.io/
          [4]: https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass
          [5]: https://kubernetes.io/docs/concepts/workloads/pods/pod-topology-spread-constraints/
        displayName: CockroachDB Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cockroach
        - cockroachdb
        links:
        - name: Helm Chart Source
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb
        - name: Configuration Options
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration
        - name: CockroachDB Source
          url: https://github.com/cockroachdb/cockroach
        maintainers:
        - email: dmesser@redhat.com
          name: Daniel Messer
        maturity: stable
        minKubeVersion: 1.8.0
        provider:
          name: Helm Community
          url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/helmoperators/cockroachdb:v5.0.4
        version: 5.0.4
      entries:
      - name: cockroachdb.v5.0.4
        version: 5.0.4
      - name: cockroachdb.v5.0.3
        version: 5.0.3
      name: stable-5.x
    - currentCSV: cockroachdb.v6.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1alpha1",
                "kind": "Cockroachdb",
                "metadata": {
                  "name": "cockroachdb-sample"
                },
                "spec": {"image":{"repository":"cockroachdb/cockroach","tag":"v22.2.4","pullPolicy":"IfNotPresent","credentials":{}},"labels":{},"clusterDomain":"cluster.local","conf":{"attrs":[],"cache":"25%","cluster-name":"","disable-cluster-name-verification":false,"join":[],"log":{"enabled":false,"config":{}},"logtostderr":"INFO","max-sql-memory":"25%","locality":"","single-node":false,"sql-audit-dir":"","port":26257,"http-port":8080,"path":"cockroach-data","store":{"enabled":false,"type":null,"size":null,"attrs":null}},"statefulset":{"replicas":3,"updateStrategy":{"type":"RollingUpdate"},"podManagementPolicy":"Parallel","budget":{"maxUnavailable":1},"args":[],"env":[],"secretMounts":[],"labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{},"nodeAffinity":{},"podAffinity":{},"podAntiAffinity":{"topologyKey":"kubernetes.io/hostname","type":"soft","weight":100},"nodeSelector":{},"priorityClassName":"","tolerations":[],"topologySpreadConstraints":{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"},"resources":{},"customLivenessProbe":{},"customReadinessProbe":{}},"service":{"ports":{"grpc":{"external":{"port":26257,"name":"grpc"},"internal":{"port":26257,"name":"grpc-internal"}},"http":{"port":8080,"name":"http"}},"public":{"type":"ClusterIP","labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{}},"discovery":{"labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{}}},"ingress":{"enabled":false,"labels":{},"annotations":{},"paths":["/"],"hosts":[],"tls":[]},"prometheus":{"enabled":true},"serviceMonitor":{"enabled":false,"labels":{},"annotations":{},"interval":"10s","namespaced":false},"storage":{"hostPath":"","persistentVolume":{"enabled":true,"size":"100Gi","storageClass":"","labels":{},"annotations":{}}},"init":{"labels":{"app.kubernetes.io/component":"init"},"jobAnnotations":{},"annotations":{},"affinity":{},"nodeSelector":{},"tolerations":[],"resources":{},"provisioning":{"enabled":false,"clusterSettings":null,"users":[],"databases":[]}},"tls":{"enabled":true,"serviceAccount":{"create":true,"name":""},"copyCerts":{"image":"busybox"},"certs":{"provided":false,"clientRootSecret":"cockroachdb-root","nodeSecret":"cockroachdb-node","tlsSecret":false,"selfSigner":{"enabled":true,"caProvided":false,"caSecret":"","minimumCertDuration":"624h","caCertDuration":"43800h","caCertExpiryWindow":"648h","clientCertDuration":"672h","clientCertExpiryWindow":"48h","nodeCertDuration":"8760h","nodeCertExpiryWindow":"168h","rotateCerts":true,"readinessWait":"30s","podUpdateTimeout":"2m"},"certManager":false,"certManagerIssuer":{"group":"cert-manager.io","kind":"Issuer","name":"cockroachdb","clientCertDuration":"672h","clientCertExpiryWindow":"48h","nodeCertDuration":"8760h","nodeCertExpiryWindow":"168h"},"useCertManagerV1CRDs":false},"selfSigner":{"image":{"repository":"cockroachlabs-helm-charts/cockroach-self-signer-cert","tag":"1.3","pullPolicy":"IfNotPresent","credentials":{},"registry":"gcr.io"}}},"networkPolicy":{"enabled":false,"ingress":{"grpc":[],"http":[]}},"iap":{"enabled":false}}
              }
            ]
          capabilities: Basic Install
          categories: Database
          certified: "false"
          containerImage: quay.io/cockroachdb/cockroach-helm-operator:6.0.0
          createdAt: 2021-01-23T10-00-00Z
          description: CockroachDB Operator based on the CockroachDB helm chart
          olm.skipRange: <6.0.0
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/cockroachdb/helm-charts
          support: CockroachDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a CockroachDB cluster
            displayName: CockroachDB
            kind: Cockroachdb
            name: cockroachdbs.charts.operatorhub.io
            version: v1alpha1
        description: |
          CockroachDB is a scalable, survivable, strongly-consistent SQL database.
          ## About this Operator
          This Operator is based on a Helm chart for CockroachDB. It supports reconfiguration for some parameters, but notably does not handle scale down of the replica count in a seamless manner. Scale up works great.

          ## Core capabilities
          * **StatefulSet** - Sets up a dynamically scalable CockroachDB cluster using a Kubernetes StatefulSet
          * **Expand Replicas** - Supports expanding the set of replicas by simply editing your object
          * **Dashboard** - Installs the CockroachDB user interface to administer your cluster. Easily expose it via an Ingress rule.
          Review all of the [confiuguration options](https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration) to best run your database instance. The example configuration is derived from the chart's [`values.yaml`](https://github.com/helm/charts/blob/master/stable/cockroachdb/values.yaml).
          ## Using the cluster
          The resulting cluster endpoint can be consumed from a `Service` that follows the pattern: `<StatefulSet-name>-public`. For example to connect using the command line client, use something like the following to obtain the name of the service:
          ```
          kubectl get service -l chart=cockroachdb-6.0.0
          NAME                                           TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)              AGE
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb          ClusterIP   None             <none>        26257/TCP,8080/TCP   24m
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public   ClusterIP   10.106.249.134   <none>        26257/TCP,8080/TCP   24m
          ```
          Then you can use the CockroachDB command line client to connect to the database cluster:
          ```
          kubectl run -it --rm cockroach-client --image=cockroachdb/cockroach --restart=Never --command -- ./cockroach sql --insecure --host example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public
          ```
          ## Before you start
          This Operator requires a cluster with PV support in order to run correctly.

          Note that for a production cluster, you will likely want to override the following parameters in [`values.yaml`](values.yaml) with your own values.

          - `statefulset.resources.requests.memory` and `statefulset.resources.limits.memory` allocate memory resources to CockroachDB pods in your cluster.
          - `conf.cache` and `conf.max-sql-memory` are memory limits that we recommend setting to 1/4 of the above resource allocation. When running CockroachDB, you must set these limits explicitly to avoid running out of memory.
          - `storage.persistentVolume.size` defaults to `100Gi` of disk space per pod, which you may increase or decrease for your use case.
          - `storage.persistentVolume.storageClass` uses the default storage class for your environment. We strongly recommend that you specify a storage class which uses an SSD.
          - `tls.enabled` must be set to `yes`/`true` to deploy in secure mode.

          For more information on overriding the `values.yaml` parameters, please see:
          > <https://www.cockroachlabs.com/docs/stable/orchestrate-cockroachdb-with-kubernetes.html#step-2-start-cockroachdb>

          If you are running in secure mode (with configuration parameter `tls.enabled` set to `yes`/`true`) and `tls.certs.provided` set to `no`/`false`), then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          When `tls.certs.provided` is set to `yes`/`true`, this chart will use certificates created outside of Kubernetes. You may want to use this if you want to use a different certificate authority from the one being used by Kubernetes or if your Kubernetes cluster doesn't fully support certificate-signing requests. To use this, first set up your certificates and load them into your Kubernetes cluster as Secrets using the commands below:

          ```
          mkdir certs
          mkdir my-safe-directory
          cockroach cert create-ca --certs-dir=certs --ca-key=my-safe-directory/ca.key
          cockroach cert create-client root --certs-dir=certs --ca-key=my-safe-directory/ca.key
          kubectl create secret generic cockroachdb-root --from-file=certs
          cockroach cert create-node --certs-dir=certs --ca-key=my-safe-directory/ca.key localhost 127.0.0.1 eerie-horse-cockroachdb-public eerie-horse-cockroachdb-public.default eerie-horse-cockroachdb-public.default.svc.cluster.local *.eerie-horse-cockroachdb *.eerie-horse-cockroachdb.default *.eerie-horse-cockroachdb.default.svc.cluster.local
          kubectl create secret generic cockroachdb-node --from-file=certs
          ```

          Set `tls.certs.tlsSecret` to `yes/true` if you make use of [cert-manager][3] in your cluster.

          [cert-manager][3] stores generated certificates in dedicated TLS secrets. Thus, they are always named:

          * `ca.crt`
          * `tls.crt`
          * `tls.key`

          On the other hand, CockroachDB also demands dedicated certificate filenames:

          * `ca.crt`
          * `node.crt`
          * `node.key`
          * `client.root.crt`
          * `client.root.key`

          By activating `tls.certs.tlsSecret` we benefit from projected secrets and convert the TLS secret filenames to their according CockroachDB filenames.

          If you are running in secure mode, then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          Confirm that all pods are `Running` successfully and init has been completed:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS      RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running     0          1m
          my-release-cockroachdb-1            1/1       Running     0          1m
          my-release-cockroachdb-2            1/1       Running     0          1m
          my-release-cockroachdb-init-k6jcr   0/1       Completed   0          1m
          ```

          Confirm that persistent volumes are created and claimed for each pod:

          ```shell
          kubectl get pv
          ```

          ```
          NAME                                       CAPACITY   ACCESS MODES   RECLAIM POLICY   STATUS    CLAIM                                      STORAGECLASS   REASON    AGE
          pvc-64878ebf-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-0   standard                 51s
          pvc-64945b4f-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-1   standard                 51s
          pvc-649d920d-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-2   standard                 51s
          ```

          ## Upgrading the cluster

          ### Chart version 3.0.0 and after

          Launch a temporary interactive pod and start the built-in SQL client:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          Set `cluster.preserve_downgrade_option`, where `$current_version` is the CockroachDB version currently running (e.g., `19.2`):

          ```sql
          > SET CLUSTER SETTING cluster.preserve_downgrade_option = '$current_version';
          ```

          Exit the shell and delete the temporary pod:

          ```sql
          > \q
          ```

          Kick off the upgrade process by changing the new Docker image, where `$new_version` is the CockroachDB version to which you are upgrading:

          ```shell
          kubectl delete job my-release-cockroachdb-init
          ```

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"image":{"tag":"<new version tag here>"}}}' --type merge
          ```

          Kubernetes will carry out a safe [rolling upgrade](https://kubernetes.io/docs/tutorials/stateful-application/basic-stateful-set/#updating-statefulsets) of your CockroachDB nodes one-by-one. Monitor the cluster's pods until all have been successfully restarted:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS              RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running             0          2m
          my-release-cockroachdb-1            1/1       Running             0          3m
          my-release-cockroachdb-2            1/1       Running             0          3m
          my-release-cockroachdb-3            0/1       ContainerCreating   0          25s
          my-release-cockroachdb-init-nwjkh   0/1       ContainerCreating   0          6s
          ```

          ```shell
          kubectl get pods \
          -o jsonpath='{range .items[*]}{.metadata.name}{"\t"}{.spec.containers[0].image}{"\n"}'
          ```

          ```
          my-release-cockroachdb-0    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-1    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-2    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-3    cockroachdb/cockroach:v20.2.4
          ```

          Resume normal operations. Once you are comfortable that the stability and performance of the cluster is what you'd expect post-upgrade, finalize the upgrade:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          ```sql
          > RESET CLUSTER SETTING cluster.preserve_downgrade_option;
          > \q
          ```

          ## Configuration

          The following table lists the configurable parameters of the CockroachDB chart and their default values.
          For details see the [`values.yaml`](values.yaml) file.

          | Parameter                                                 | Description                                                     | Default                                          |
          | ---------                                                 | -----------                                                     | -------                                          |
          | `clusterDomain`                                           | Cluster's default DNS domain                                    | `cluster.local`                                  |
          | `conf.attrs`                                              | CockroachDB node attributes                                     | `[]`                                             |
          | `conf.cache`                                              | Size of CockroachDB's in-memory cache                           | `25%`                                            |
          | `conf.cluster-name`                                       | Name of CockroachDB cluster                                     | `""`                                             |
          | `conf.disable-cluster-name-verification`                  | Disable CockroachDB cluster name verification                   | `no`                                             |
          | `conf.join`                                               | List of already-existing CockroachDB instances                  | `[]`                                             |
          | `conf.max-disk-temp-storage`                              | Max storage capacity for temp data                              | `0`                                              |
          | `conf.max-offset`                                         | Max allowed clock offset for CockroachDB cluster                | `500ms`                                          |
          | `conf.max-sql-memory`                                     | Max memory to use processing SQL querie                         | `25%`                                            |
          | `conf.locality`                                           | Locality attribute for this deployment                          | `""`                                             |
          | `conf.single-node`                                        | Disable CockroachDB clustering (standalone mode)                | `no`                                             |
          | `conf.sql-audit-dir`                                      | Directory for SQL audit log                                     | `""`                                             |
          | `conf.port`                                               | CockroachDB primary serving port in Pods                        | `26257`                                          |
          | `conf.http-port`                                          | CockroachDB HTTP port in Pods                                   | `8080`                                           |
          | `image.repository`                                        | Container image name                                            | `cockroachdb/cockroach`                          |
          | `image.tag`                                               | Container image tag                                             | `v20.2.4`                                        |
          | `image.pullPolicy`                                        | Container pull policy                                           | `IfNotPresent`                                   |
          | `image.credentials`                                       | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `statefulset.replicas`                                    | StatefulSet replicas number                                     | `3`                                              |
          | `statefulset.updateStrategy`                              | Update strategy for StatefulSet Pods                            | `{"type": "RollingUpdate"}`                      |
          | `statefulset.podManagementPolicy`                         | `OrderedReady`/`Parallel` Pods creation/deletion order          | `Parallel`                                       |
          | `statefulset.budget.maxUnavailable`                       | k8s PodDisruptionBudget parameter                               | `1`                                              |
          | `statefulset.args`                                        | Extra command-line arguments                                    | `[]`                                             |
          | `statefulset.env`                                         | Extra env vars                                                  | `[]`                                             |
          | `statefulset.secretMounts`                                | Additional Secrets to mount at cluster members                  | `[]`                                             |
          | `statefulset.labels`                                      | Additional labels of StatefulSet and its Pods                   | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `statefulset.annotations`                                 | Additional annotations of StatefulSet Pods                      | `{}`                                             |
          | `statefulset.nodeAffinity`                                | [Node affinity rules][2] of StatefulSet Pods                    | `{}`                                             |
          | `statefulset.podAffinity`                                 | [Inter-Pod affinity rules][1] of StatefulSet Pods               | `{}`                                             |
          | `statefulset.podAntiAffinity`                             | [Anti-affinity rules][1] of StatefulSet Pods                    | auto                                             |
          | `statefulset.podAntiAffinity.topologyKey`                 | The topologyKey for auto [anti-affinity rules][1]               | `kubernetes.io/hostname`                         |
          | `statefulset.podAntiAffinity.type`                        | Type of auto [anti-affinity rules][1]                           | `soft`                                           |
          | `statefulset.podAntiAffinity.weight`                      | Weight for `soft` auto [anti-affinity rules][1]                 | `100`                                            |
          | `statefulset.nodeSelector`                                | Node labels for StatefulSet Pods assignment                     | `{}`                                             |
          | `statefulset.priorityClassName`                           | [PriorityClassName][4] for StatefulSet Pods                     | `""`                                             |
          | `statefulset.tolerations`                                 | Node taints to tolerate by StatefulSet Pods                     | `[]`                                             |
          | `statefulset.topologySpreadConstraints`                   | [Topology Spread Constraints rules][5] of StatefulSet Pods      | auto                                             |
          | `statefulset.topologySpreadConstraints.maxSkew`           | Degree to which Pods may be unevenly distributed                | `1`                                              |
          | `statefulset.topologySpreadConstraints.topologyKey`       | The key of node labels                                          | `topology.kubernetes.io/zone`                    |
          | `statefulset.topologySpreadConstraints.whenUnsatisfiable` | `ScheduleAnyway`/`DoNotSchedule` for unsatisfiable constraints  | `ScheduleAnyway`                                 |
          | `statefulset.resources`                                   | Resource requests and limits for StatefulSet Pods               | `{}`                                             |
          | `service.ports.grpc.external.port`                        | CockroachDB primary serving port in Services                    | `26257`                                          |
          | `service.ports.grpc.external.name`                        | CockroachDB primary serving port name in Services               | `grpc`                                           |
          | `service.ports.grpc.internal.port`                        | CockroachDB inter-communication port in Services                | `26257`                                          |
          | `service.ports.grpc.internal.name`                        | CockroachDB inter-communication port name in Services           | `grpc-internal`                                  |
          | `service.ports.http.port`                                 | CockroachDB HTTP port in Services                               | `8080`                                           |
          | `service.ports.http.name`                                 | CockroachDB HTTP port name in Services                          | `http`                                           |
          | `service.public.type`                                     | Public Service type                                             | `ClusterIP`                                      |
          | `service.public.labels`                                   | Additional labels of public Service                             | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.public.annotations`                              | Additional annotations of public Service                        | `{}`                                             |
          | `service.discovery.labels`                                | Additional labels of discovery Service                          | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.discovery.annotations`                           | Additional annotations of discovery Service                     | `{}`                                             |
          | `ingress.enabled`                                         | Enable ingress resource for CockroachDB                         | `false`                                          |
          | `ingress.labels`                                          | Additional labels of Ingress                                    | `{}`                                             |
          | `ingress.annotations`                                     | Additional annotations of Ingress                               | `{}`                                             |
          | `ingress.paths`                                           | Paths for the default host                                      | `[/]`                                            |
          | `ingress.hosts`                                           | CockroachDB Ingress hostnames                                   | `[]`                                             |
          | `ingress.tls[0].hosts`                                    | CockroachDB Ingress tls hostnames                               | `nil`                                            |
          | `ingress.tls[0].secretName`                               | CockroachDB Ingress tls secret name                             | `nil`                                            |
          | `serviceMonitor.enabled`                                  | Create [ServiceMonitor](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/design.md#servicemonitor) Resource for scraping metrics using [PrometheusOperator](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/user-guides/getting-started.md#prometheus-operator)                     | `false`                                             |
          | `serviceMonitor.labels`                                   | Additional labels of ServiceMonitor                             | `{}`                                             |
          | `serviceMonitor.annotations`                              | Additional annotations of ServiceMonitor                        | `{}`                                             |
          | `serviceMonitor.interval`                                 | ServiceMonitor scrape metrics interval                          | `10s`                                            |
          | `serviceMonitor.scrapeTimeout`                            | ServiceMonitor scrape timeout                                   | `nil`                                            |
          | `storage.hostPath`                                        | Absolute path on host to store data                             | `""`                                             |
          | `storage.persistentVolume.enabled`                        | Whether to use PersistentVolume to store data                   | `yes`                                            |
          | `storage.persistentVolume.size`                           | PersistentVolume size                                           | `100Gi`                                          |
          | `storage.persistentVolume.storageClass`                   | PersistentVolume class                                          | `""`                                             |
          | `storage.persistentVolume.labels`                         | Additional labels of PersistentVolumeClaim                      | `{}`                                             |
          | `storage.persistentVolume.annotations`                    | Additional annotations of PersistentVolumeClaim                 | `{}`                                             |
          | `init.labels`                                             | Additional labels of init Job and its Pod                       | `{"app.kubernetes.io/component": "init"}`        |
          | `init.annotations`                                        | Additional labels of the Pod of init Job                        | `{}`                                             |
          | `init.affinity`                                           | [Affinity rules][2] of init Job Pod                             | `{}`                                             |
          | `init.nodeSelector`                                       | Node labels for init Job Pod assignment                         | `{}`                                             |
          | `init.tolerations`                                        | Node taints to tolerate by init Job Pod                         | `[]`                                             |
          | `init.resources`                                          | Resource requests and limits for the Pod of init Job            | `{}`                                             |
          | `tls.enabled`                                             | Whether to run securely using TLS certificates                  | `no`                                             |
          | `tls.serviceAccount.create`                               | Whether to create a new RBAC service account                    | `yes`                                            |
          | `tls.serviceAccount.name`                                 | Name of RBAC service account to use                             | `""`                                             |
          | `tls.certs.provided`                                      | Bring your own certs scenario, i.e certificates are provided    | `no`                                             |
          | `tls.certs.clientRootSecret`                              | If certs are provided, secret name for client root cert         | `cockroachdb-root`                               |
          | `tls.certs.nodeSecret`                                    | If certs are provided, secret name for node cert                | `cockroachdb-node`                               |
          | `tls.certs.tlsSecret`                                     | Own certs are stored in TLS secret                              | `no`                                             |
          | `tls.init.image.repository`                               | Image to use for requesting TLS certificates                    | `cockroachdb/cockroach-k8s-request-cert`         |
          | `tls.init.image.tag`                                      | Image tag to use for requesting TLS certificates                | `0.4`                                            |
          | `tls.init.image.pullPolicy`                               | Requesting TLS certificates container pull policy               | `IfNotPresent`                                   |
          | `tls.init.image.credentials`                              | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `networkPolicy.enabled`                                   | Enable NetworkPolicy for CockroachDB's Pods                     | `no`                                             |
          | `networkPolicy.ingress.grpc`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |
          | `networkPolicy.ingress.http`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |


          Override the default parameters using the `CockroachDB` custom resource.

          ## Deep dive

          ### Connecting to the CockroachDB cluster

          Once you've created the cluster, you can start talking to it by connecting to its `-public` Service. CockroachDB is PostgreSQL wire protocol compatible, so there's a [wide variety of supported clients](https://www.cockroachlabs.com/docs/install-client-drivers.html). As an example, we'll open up a SQL shell using CockroachDB's built-in shell and play around with it a bit, like this (likely needing to replace `my-release-cockroachdb-public` with the name of the `-public` Service that was created with your installed chart):

          ```shell
          kubectl run cockroach-client --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host my-release-cockroachdb-public
          ```

          ```
          Waiting for pod default/cockroach-client to be running, status is Pending,
          pod ready: false
          If you don't see a command prompt, try pressing enter.
          root@my-release-cockroachdb-public:26257> SHOW DATABASES;
          +--------------------+
          |      Database      |
          +--------------------+
          | information_schema |
          | pg_catalog         |
          | system             |
          +--------------------+
          (3 rows)
          root@my-release-cockroachdb-public:26257> CREATE DATABASE bank;
          CREATE DATABASE
          root@my-release-cockroachdb-public:26257> CREATE TABLE bank.accounts (id INT
          PRIMARY KEY, balance DECIMAL);
          CREATE TABLE
          root@my-release-cockroachdb-public:26257> INSERT INTO bank.accounts VALUES
          (1234, 10000.50);
          INSERT 1
          root@my-release-cockroachdb-public:26257> SELECT * FROM bank.accounts;
          +------+---------+
          |  id  | balance |
          +------+---------+
          | 1234 | 10000.5 |
          +------+---------+
          (1 row)
          root@my-release-cockroachdb-public:26257> \q
          Waiting for pod default/cockroach-client to terminate, status is Running
          pod "cockroach-client" deleted
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          ### Cluster health

          Because our pod spec includes regular health checks of the CockroachDB processes, simply running `kubectl get pods` and looking at the `STATUS` column is sufficient to determine the health of each instance in the cluster.

          If you want more detailed information about the cluster, the best place to look is the Admin UI.

          ### Accessing the Admin UI

          If you want to see information about how the cluster is doing, you can try pulling up the CockroachDB Admin UI by port-forwarding from your local machine to one of the pods (replacing `my-release-cockroachdb-0` with the name of one of your pods:

          ```shell
          kubectl port-forward my-release-cockroachdb-0 8080
          ```

          You should then be able to access the Admin UI by visiting <http://localhost:8080/> in your web browser.

          ### Failover

          If any CockroachDB member fails, it is restarted or recreated automatically by the Kubernetes infrastructure, and will re-join the cluster automatically when it comes back up. You can test this scenario by killing any of the CockroachDB pods:

          ```shell
          kubectl delete pod my-release-cockroachdb-1
          ```

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          After a while:

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-1  1/1       Running       0          20s
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          You can check the state of re-joining from the new pod's logs:

          ```shell
          kubectl logs my-release-cockroachdb-1
          ```

          ```
          [...]
          I161028 19:32:09.754026 1 server/node.go:586  [n1] node connected via gossip and
          verified as part of cluster {"35ecbc27-3f67-4e7d-9b8f-27c31aae17d6"}
          [...]
          cockroachdb-0.my-release-cockroachdb.default.svc.cluster.local:26257
          build:      beta-20161027-55-gd2d3c7f @ 2016/10/28 19:27:25 (go1.7.3)
          admin:      http://0.0.0.0:8080
          sql:
          postgresql://root@my-release-cockroachdb-1.my-release-cockroachdb.default.svc.cluster.local:26257?sslmode=disable
          logs:       cockroach-data/logs
          store[0]:   path=cockroach-data
          status:     restarted pre-existing node
          clusterID:  {35ecbc27-3f67-4e7d-9b8f-27c31aae17d6}
          nodeID:     2
          [...]
          ```

          ### NetworkPolicy

          To enable NetworkPolicy for CockroachDB, install [a networking plugin that implements the Kubernetes NetworkPolicy spec](https://kubernetes.io/docs/tasks/administer-cluster/declare-network-policy#before-you-begin), and set `networkPolicy.enabled` to `yes`/`true`.

          For Kubernetes v1.5 & v1.6, you must also turn on NetworkPolicy by setting the `DefaultDeny` Namespace annotation. Note: this will enforce policy for _all_ pods in the Namespace:

          ```shell
          kubectl annotate namespace default "net.beta.kubernetes.io/network-policy={\"ingress\":{\"isolation\":\"DefaultDeny\"}}"
          ```

          For more precise policy, set `networkPolicy.ingress.grpc` and `networkPolicy.ingress.http` rules. This will only allow pods that match the provided rules to connect to CockroachDB.

          ### Scaling

          Scaling should be managed via the `helm upgrade` command. After resizing your cluster on your cloud environment (e.g., GKE or EKS), run the following command to add a pod. This assumes you scaled from 3 to 4 nodes:

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"statefulset":{"replicas":"4"}}}' --type merge
          ```

          Note, that if you are running in secure mode (`tls.enabled` is `yes`/`true`) and increase the size of your cluster, you will also have to approve the CSR (certificate-signing request) of each new node (using `kubectl get csr` and `kubectl certificate approve`).

          [1]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [2]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#node-affinity
          [3]: https://cert-manager.io/
          [4]: https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass
          [5]: https://kubernetes.io/docs/concepts/workloads/pods/pod-topology-spread-constraints/
        displayName: CockroachDB Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cockroach
        - cockroachdb
        links:
        - name: Helm Chart Source
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb
        - name: Configuration Options
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration
        - name: CockroachDB Source
          url: https://github.com/cockroachdb/cockroach
        maintainers:
        - email: yandu@cockroachlabs.com
          name: Yandu Oppacher
        maturity: stable
        minKubeVersion: 1.8.0
        provider:
          name: Helm Community
          url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/cockroachdb/cockroach-helm-operator:6.0.0
        version: 6.0.0
      entries:
      - name: cockroachdb.v6.0.0
        version: 6.0.0
      name: stable-v6.x
    defaultChannel: stable-v6.x
    packageName: cockroachdb
    provider:
      name: Helm Community
      url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CrowdStrike
      provider-url: https://crowdStrike.com
    name: falcon-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: falcon-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconContainer",
                "metadata": {
                  "name": "falcon-sidecar-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "sidecar"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconNodeSensor",
                "metadata": {
                  "name": "falcon-node-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "daemonset"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security,Monitoring
          containerImage: quay.io/crowdstrike/falcon-operator:0.8.1
          createdAt: "2024-06-06T10:00:00Z"
          description: '[DEPRECATED] Use the CrowdStrike Falcon Operator from the
            certified channel instead'
          operatorframework.io/suggested-namespace: falcon-operator
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/CrowdStrike/falcon-operator
          support: Community Only
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FalconContainer is the Schema for the falconcontainers API
            displayName: Falcon Container
            kind: FalconContainer
            name: falconcontainers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconNodeSensor is the Schema for the falconnodesensors
              API
            displayName: Falcon Node Sensor
            kind: FalconNodeSensor
            name: falconnodesensors.falcon.crowdstrike.com
            version: v1alpha1
        description: |-
          ## Deprecated
          This CrowdStrike community operator is deprecated, use the CrowdStrike Falcon Operator from the certified channel instead.

          ## About the CrowdStrike Falcon Operator
          The CrowdStrike Falcon Operator deploys CrowdStrike Falcon Workload Protection to the cluster. The operator exposes 2 custom resources that allows you to deploy either the Falcon Container Sensor or Falcon Node Sensor.

          ## About Custom Resources

          | Custom Resource                                       | Description                                                      |
          | :--------                                             | :------------                                                    |
          | [FalconContainer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/container/README.md) | Manages installation of Falcon Container Sensor on the cluster   |
          | [FalconNodeSensor](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/node/README.md)     | Manages installation of Falcon Linux Sensor on the cluster nodes |

          ## Installation and Deployment

          For installation and deployment of the CrowdStrike Falcon Operator and its Custom Resources, please read the [Installation and Deployment Guide](./docs/install_guide.md) and choose the deployment method that is right for your target environment.

          ## Getting Help
          If you encounter any issues while using the Falcon Operator, you can create an issue on our [Github repo](https://github.com/CrowdStrike/falcon-operator) for bugs, enhancements, or other requests.

          ## Contributing
          You can contribute by:

          * Raising any issues you find using Falcon Operator
          * Fixing issues by opening [Pull Requests](https://github.com/CrowdStrike/falcon-operator/pulls)
          * Improving documentation
          * Talking about the CrowdStrike Falcon Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/CrowdStrike/falcon-operator/issues).

          ## Additional Resources
           - CrowdStrike Container Security: [Product Page](https://www.crowdstrike.com/products/cloud-security/falcon-cloud-workload-protection/container-security/)
           - So You Think Your Containers Are Secure? Four Steps to Ensure a Secure Container Deployment: [Blog Post](https://www.crowdstrike.com/blog/four-steps-to-ensure-a-secure-containter-deployment/)
           - Container Security With CrowdStrike: [Blog Post](https://www.crowdstrike.com/blog/tech-center/container-security/)
           - To learn more about Falcon Container Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/support/documentation/146/falcon-container-sensor-for-linux), [Release Notes](https://falcon.crowdstrike.com/support/news/release-notes-falcon-container-sensor-for-linux)
           - To learn more about Falcon Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/documentation/20/falcon-sensor-for-linux#kubernetes-support)
           - [Developer Documentation](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/developer_guide.md)
        displayName: '[DEPRECATED] CrowdStrike Operator'
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - runtime-protection
        links:
        - name: Falcon Operator
          url: https://github.com/CrowdStrike/falcon-operator
        maintainers:
        - email: integrations@crowdstrike.com
          name: CrowdStrike Solutions Architecture
        maturity: alpha
        minKubeVersion: 1.17.0
        provider:
          name: CrowdStrike
          url: https://crowdStrike.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/crowdstrike/falcon-operator:0.8.1
        version: 0.8.2
      entries:
      - name: falcon-operator.v0.8.2
        version: 0.8.2
      - name: falcon-operator.v0.8.1
        version: 0.8.1
      name: alpha
    defaultChannel: alpha
    packageName: falcon-operator
    provider:
      name: CrowdStrike
      url: https://crowdStrike.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Foundation Model Stack
      provider-url: ""
    name: multi-nic-cni-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multi-nic-cni-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.3.0",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.3.0",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.3.0
          createdAt: "2024-04-17T04:51:40Z"
          description: Automate container multi-network configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: http://github.com/foundation-model-stack/multi-nic-cni
          support: Foundation Model Stack Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Foundation Model Stack
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.3.0
        version: 1.3.0
      entries:
      - name: multi-nic-cni-operator.v1.3.0
        version: 1.3.0
      - name: multi-nic-cni-operator.v1.2.9
        version: 1.2.9
      - name: multi-nic-cni-operator.v1.2.8
        version: 1.2.8
      - name: multi-nic-cni-operator.v1.2.2
        version: 1.2.2
      - name: multi-nic-cni-operator.v1.2.1
        version: 1.2.1
      - name: multi-nic-cni-operator.v1.0.2
        version: 1.0.2
      name: alpha
    - currentCSV: multi-nic-cni-operator.v1.2.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.7",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.7",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.7
          createdAt: "2025-05-09T12:45:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: minzhang@redhat.com
          name: Min Zhang
        - email: ebendror@redhat.com
          name: Eran Ben Dror
        maturity: alpha
        provider:
          name: Foundation Model Stack
        relatedImages:
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.7
        version: 1.2.7
      entries:
      - name: multi-nic-cni-operator.v1.2.7
        version: 1.2.7
      - name: multi-nic-cni-operator.v1.2.3
        version: 1.2.3
      - name: multi-nic-cni-operator.v1.2.0
        version: 1.2.0
      - name: multi-nic-cni-operator.v1.1.0
        version: 1.1.0
      name: beta
    - currentCSV: multi-nic-cni-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.6",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.6",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.6
          createdAt: "2025-03-24T14:49:54Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: minzhang@redhat.com
          name: Min Zhang
        - email: ebendror@redhat.com
          name: Eran Ben Dror
        maturity: alpha
        provider:
          name: Foundation Model Stack
        relatedImages:
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.6
        version: 1.2.6
      entries:
      - name: multi-nic-cni-operator.v1.2.6
        version: 1.2.6
      - name: multi-nic-cni-operator.v1.2.5
        version: 1.2.5
      - name: multi-nic-cni-operator.v1.2.4
        version: 1.2.4
      - name: multi-nic-cni-operator.v1.0.5
        version: 1.0.5
      - name: multi-nic-cni-operator.v1.0.4
        version: 1.0.4
      - name: multi-nic-cni-operator.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: multi-nic-cni-operator
    provider:
      name: Foundation Model Stack
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: assisted-service-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: assisted-service-operator.v0.7.74
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-x86_64-live.x86_64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-aarch64-live.aarch64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-s390x-live.s390x.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-x86_64-live.x86_64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-aarch64-live.aarch64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-s390x-live.s390x.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-x86_64-live.x86_64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-aarch64-live.aarch64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-ppc64le-live.ppc64le.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-s390x-live.s390x.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-x86_64-live.x86_64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-aarch64-live.aarch64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-ppc64le-live.ppc64le.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-s390x-live.s390x.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-s390x-live.s390x.iso",
                      "version": "418.94.202409162337-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service:latest
          createdAt: "2024-10-16T17:43:34Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.74'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AgentClassification is the Schema for the AgentClassifications
              API
            displayName: Agent Classification
            kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - description: Agent is the Schema for the hosts API
            displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: Infra Env
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service:latest
        version: 0.7.74
      entries:
      - name: assisted-service-operator.v0.7.74
        version: 0.7.74
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: assisted-service-operator.v0.7.71
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-x86_64-live.x86_64.iso",
                      "version": "414.92.202310170514-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-aarch64-live.aarch64.iso",
                      "version": "414.92.202310170514-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202309201615-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-s390x-live.s390x.iso",
                      "version": "414.92.202309201615-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-s390x-live.s390x.iso",
                      "version": "415.92.202310310037-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:cffbdc5fd51c08f16ec2aa484cf226e3b29510919a8aa9668d766007de8f3d79
          createdAt: "2023-12-15T15:26:08Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.71'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service@sha256:cffbdc5fd51c08f16ec2aa484cf226e3b29510919a8aa9668d766007de8f3d79
        - quay.io/edge-infrastructure/assisted-image-service@sha256:027fa9c5a881848786745aafeac5e48575eefe22417abe29d9e8eb17940b8f80
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/edge-infrastructure/assisted-installer@sha256:0f4ecf9a75b6755d02a45f9b70357ac45b6b4db66f24f0e83f888269d64f4293
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:40bafed94547a5fe96deb5492f10225501b965198227c9634869931692bda492
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:98740646fbfc0520f9d85ac5b483d1a898f22e164c5234901be51970bef7fcf3
        version: 0.7.71
      entries:
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.10
    - currentCSV: assisted-service-operator.v0.7.73
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-x86_64-live.x86_64.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-aarch64-live.aarch64.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-s390x-live.s390x.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-x86_64-live.x86_64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-aarch64-live.aarch64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-s390x-live.s390x.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.16.0-ec.4/rhcos-4.16.0-ec.4-x86_64-live.x86_64.iso",
                      "version": "416.94.202403071059-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-aarch64-live.aarch64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-s390x-live.s390x.iso",
                      "version": "415.92.202402130021-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:641fb65d8f0117cc6007567337ae38d5e912add454a76905617b60db99b60a64
          createdAt: "2024-03-20T10:08:49Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.73'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:2cc14304ba27c51d61b86db05445f3f0acc68593a4d76081da8eb9d6bae41757
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:d0da07ec81877302fc376e8049dca69c1acdbf56a37aa3e320cd75f2b85a2ed0
        - quay.io/edge-infrastructure/assisted-service@sha256:641fb65d8f0117cc6007567337ae38d5e912add454a76905617b60db99b60a64
        - quay.io/edge-infrastructure/assisted-image-service@sha256:b33068bbe88e38c0444fb158e233e5644fb127416198d52fdc809a3f2e684fde
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/edge-infrastructure/assisted-installer@sha256:234e4680b0ff4376f2c11d6f94a75a5dd6ae16c3218d6b63d2003cf43c65e185
        version: 0.7.73
      entries:
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.11
    - currentCSV: assisted-service-operator.v0.7.74
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-x86_64-live.x86_64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-aarch64-live.aarch64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-s390x-live.s390x.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-x86_64-live.x86_64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-aarch64-live.aarch64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-s390x-live.s390x.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-x86_64-live.x86_64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-aarch64-live.aarch64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-ppc64le-live.ppc64le.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-s390x-live.s390x.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-x86_64-live.x86_64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-aarch64-live.aarch64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-ppc64le-live.ppc64le.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-s390x-live.s390x.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-s390x-live.s390x.iso",
                      "version": "418.94.202409162337-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service:latest
          createdAt: "2024-10-16T17:43:34Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.74'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AgentClassification is the Schema for the AgentClassifications
              API
            displayName: Agent Classification
            kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - description: Agent is the Schema for the hosts API
            displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: Infra Env
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service:latest
        version: 0.7.74
      entries:
      - name: assisted-service-operator.v0.7.74
        version: 0.7.74
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.13
    - currentCSV: assisted-service-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "osImages": [
                    {
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202105062123-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/ocpmetal/assisted-service@sha256:64e88c5aa41c606bc2db2b0991acdbb1166559306474e13f5e2810f325616d64
          createdAt: "2021-09-23T11:11:20Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.0 <0.0.8'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled, for
              all other names a "Warning" event will be emitted + message logged and
              it will be ignored forever.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent".
          No other name will be accepted.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/ocpmetal/assisted-service@sha256:64e88c5aa41c606bc2db2b0991acdbb1166559306474e13f5e2810f325616d64
        - quay.io/ocpmetal/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/ocpmetal/assisted-installer@sha256:db97f7a565a8d5af61f3876d6465996e90c7f79dad6ec71c5bb5cce837444974
        - quay.io/ocpmetal/assisted-installer-agent@sha256:a391892766d1ce81300bf82943318e7c3be0d2945fc19065527ab9f64fc24cf8
        - quay.io/ocpmetal/assisted-installer-controller@sha256:43b65c310eee985ee0fc20ad8ee684f7557ffe704fd1a2fa53831e3dbc41e849
        version: 0.0.8
      entries:
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.3
    - currentCSV: assisted-service-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.2/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.2/rhcos-4.8.2-x86_64-live.x86_64.iso",
                      "version": "48.84.202107202156-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-x86_64-live.x86_64.iso",
                      "version": "49.84.202110081407-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:c3a852842b88f47c51cf1c62d3a19b102729fb355df19ced0e03e8265739c39c
          createdAt: "2022-04-07T11:23:06Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.1.3'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled, for
              all other names a "Warning" event will be emitted + message logged and
              it will be ignored forever.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent".
          No other name will be accepted.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-image-service@sha256:add19e251cafdfc4f34264d0aec5b660f0ea2dcc13326f4a2c8d50a2e465c0a2
        - quay.io/edge-infrastructure/assisted-service@sha256:c3a852842b88f47c51cf1c62d3a19b102729fb355df19ced0e03e8265739c39c
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/edge-infrastructure/assisted-installer@sha256:461d60b21d1d3ccb15a170f937f4871088cd0f83108b2271e84bdf2883a66870
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:979650f2e03e457d71958e8813285fcb7185fe0035f21a6ba956b4277cb29b70
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:658787d1e6c1f733c5f03863655aff315258c0b38e1972573c3878df46e4e463
        version: 0.1.3
      entries:
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.4
    - currentCSV: assisted-service-operator.v0.3.30
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-x86_64-live.x86_64.iso",
                      "version": "49.84.202110081407-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-aarch64-live-rootfs.aarch64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-aarch64-live.aarch64.iso",
                      "version": "49.84.202110080947-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.3/rhcos-4.10.3-x86_64-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.3/rhcos-4.10.3-x86_64-live.x86_64.iso",
                      "version": "410.84.202201251210-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:7c13ce52d82f761aec47bc5a6639d7baabe99d0f4f2e2106c4341018c606b458
          createdAt: "2022-06-13T23:16:28Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.3.30'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service@sha256:7c13ce52d82f761aec47bc5a6639d7baabe99d0f4f2e2106c4341018c606b458
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/edge-infrastructure/assisted-installer@sha256:57cf8d5fe49067a8c1345bedaf90f2b63aededdd7d615878496182c35a13892e
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:6271d639e4c0647bd28fbc85cea54f31143a876c3237376b9c670ec254a9c2fc
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:2267d2d7be38f81d530fed977b0c46f6c976369cd89c123c8e036c3b1dd988d5
        - quay.io/edge-infrastructure/assisted-image-service@sha256:d6396f9360b877a4f57139ba5b4969f243278db709be4de3573f2fba75e50018
        version: 0.3.30
      entries:
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.5
    - currentCSV: assisted-service-operator.v0.4.41
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.40/rhcos-4.9.40-x86_64-live.x86_64.iso",
                      "version": "49.84.202206171736-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.16/rhcos-4.10.16-x86_64-live.x86_64.iso",
                      "version": "410.84.202205191234-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.16/rhcos-4.10.16-aarch64-live.aarch64.iso",
                      "version": "410.84.202205191023-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.11.0-rc.2/rhcos-4.11.0-rc.2-x86_64-live.x86_64.iso",
                      "version": "411.86.202207090519-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.11.0-rc.2/rhcos-4.11.0-rc.2-aarch64-live.aarch64.iso",
                      "version": "411.86.202207090159-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:0fa869095185759c26c391f0e353f293c903ea739576f160d6668f5f20404bdb
          createdAt: "2022-09-11T14:56:07Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.4.41'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:56484f99799d22f345215bb65c6b81b6ee14e07d26b625c3f283a2ddfdc37238
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:951b9d625bc300b5b72a3916cb382015527ea786f10b7acbeebf83ae854ca51a
        - quay.io/edge-infrastructure/assisted-service@sha256:0fa869095185759c26c391f0e353f293c903ea739576f160d6668f5f20404bdb
        - quay.io/edge-infrastructure/assisted-image-service@sha256:104d2aae313c724fb10d3ef3bba78b2370fc0597de31335453894b679a450ab8
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:4ca00219efbc0d9dcce415dadc1039e9495f3a3b00ab3bd575b41554d7934899
        - quay.io/edge-infrastructure/assisted-installer@sha256:803168aeb7e01db77e3b2ff7418ceea3d93daccaceeb95f8fc02027144ebac75
        version: 0.4.41
      entries:
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.6
    - currentCSV: assisted-service-operator.v0.4.50
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-x86_64-live.x86_64.iso",
                      "version": "411.86.202210041459-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-aarch64-live.aarch64.iso",
                      "version": "411.86.202210032347-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.12.0-ec.2/rhcos-4.12.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "412.86.202208101039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.12.0-ec.2/rhcos-4.12.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "412.86.202208101040-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:0ca1cfe04f088cd4c7318cb91978754d6472ada29e3436d05e22335740bfff49
          createdAt: "2022-11-23T19:17:47Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.4.50'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service@sha256:0ca1cfe04f088cd4c7318cb91978754d6472ada29e3436d05e22335740bfff49
        - quay.io/edge-infrastructure/assisted-image-service@sha256:71e8c9c76caa8fe01255fdb4a1172cda372969f7fe1399a3a54fb93e967e68cf
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:54b39c6cafefb70d258ec1d491b6a0f1dd888cd85bd442c8f40a4805cf4721ca
        - quay.io/edge-infrastructure/assisted-installer@sha256:f225c86242b0bfbf58753123ef33b7c07f367a5099641d19cfda5a7121382586
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:34e0341c0471b1772db11535107db06e511c8d5d9816edc3971e745928164ecd
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:c1c7281b9f06e56330cc55ac8d74185d42ee82dff105ae747db677b8eb059aea
        version: 0.4.50
      entries:
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.7
    - currentCSV: assisted-service-operator.v0.6.65
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-x86_64-live.x86_64.iso",
                      "version": "411.86.202210041459-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-aarch64-live.aarch64.iso",
                      "version": "411.86.202210032347-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-s390x-live.s390x.iso",
                      "version": "411.86.202210171842-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202210171844-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-x86_64-live.x86_64.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-aarch64-live.aarch64.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-s390x-live.s390x.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-x86_64-live.x86_64.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-aarch64-live.aarch64.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-s390x-live.s390x.iso",
                      "version": "413.92.202303281804-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:2dfaf2c3ac58b2987b0088df302c7cf717d0b6640cdb8b2327370aa110120784
          createdAt: "2023-05-24T11:15:57Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.6.65'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-image-service@sha256:b1ba1076fe16b2f868ed31a660b16583a957a83634c7ec3cf03574b56e02144c
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:1287d947464fbf5275feb31d106d2feb856c6117f0eaec8d947f2e03bd8618c5
        - quay.io/edge-infrastructure/assisted-installer@sha256:f8a7e7cb6ff1a4a7daa0cb26f0fab8da17f109653af5979c9d53c35b7ca2c1bb
        - quay.io/edge-infrastructure/assisted-service@sha256:2dfaf2c3ac58b2987b0088df302c7cf717d0b6640cdb8b2327370aa110120784
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:1ad91b13c763437863480df68f5b65576e95c558d4537886432a81c0b8d3cedc
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:a6f99547564cc2e047df74003019817b2dbd3dad992a0c8e53aaaba4f5b02ab6
        version: 0.6.65
      entries:
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.8
    - currentCSV: assisted-service-operator.v0.7.67
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-s390x-live.s390x.iso",
                      "version": "414.92.202305090606-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:fb0397a1402d42a64e3128e66a0817aa9466e67eb407cb792dda0d380bd5a189
          createdAt: "2023-10-08T07:14:20Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.67'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service@sha256:fb0397a1402d42a64e3128e66a0817aa9466e67eb407cb792dda0d380bd5a189
        - quay.io/edge-infrastructure/assisted-image-service@sha256:69ff087be8e5d8458a97f49d88b8b2025ee341d362dcf5618988c011b66688e0
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/edge-infrastructure/assisted-installer@sha256:d85be59bc379927f46d0a88ef04172c5c6733e31e1e4a8188de2d434d5bedb66
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:f2aeb4f7d193ea2c5702cae0f6039c02991b9f417d091468d46974408c97efe4
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:35c118d8c07dd72fe3307424cc051d9e0426e916412171449a05e41445b9f599
        version: 0.7.67
      entries:
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.9
    defaultChannel: alpha
    packageName: assisted-service-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-commons-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: commons-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:6768e4189fa8deebf337e4732e5f12bb2ad4acdc112d9d05f5dc1a4ed81f08de
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for S3ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for S3BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/commons-operator@sha256:6768e4189fa8deebf337e4732e5f12bb2ad4acdc112d9d05f5dc1a4ed81f08de
        version: 24.11.1
      entries:
      - name: commons-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: commons-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:adc2b25a676f7357ca3971c13ecf5c9ed3cb1af24ffc902b03733cc814e8946f
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for S3ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for S3BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/commons-operator@sha256:adc2b25a676f7357ca3971c13ecf5c9ed3cb1af24ffc902b03733cc814e8946f
        version: 25.3.0
      entries:
      - name: commons-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: commons-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
        version: 25.7.0
      entries:
      - name: commons-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: commons-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
        version: 25.7.0
      entries:
      - name: commons-operator.v25.7.0
        version: 25.7.0
      - name: commons-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-commons-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: web-terminal
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: web-terminal.v1.11.1-0.1747215995.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: registry.redhat.io/web-terminal/web-terminal-rhel9-operator@sha256:0478d14e92df84fdfbf4584384d34dc9a71427a6487c2564d2eb7815ba1ac12b
          createdAt: "2021-10-26T07:24:32Z"
          description: Start a web terminal in your browser with common CLI tools
            for interacting with the cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.substitutesFor: web-terminal.v1.11.1
          operatorframework.io/suggested-namespace: openshift-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/redhat-developer/web-terminal-operator/
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - kind: DevWorkspaceRouting
            name: devworkspaceroutings.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha1
        description: |
          Start a web terminal in your browser with common CLI tools for interacting with
          the cluster.

          **Note:** The Web Terminal Operator integrates with the OpenShift Console in
          OpenShift 4.5.3 and higher to simplify web terminal instance creation and
          automate OpenShift login. In earlier versions of OpenShift, the operator can
          be installed but web terminals will have to be created and accessed manually.

          ## Description
          The Web Terminal Operator leverages the
          [DevWorkspace Operator](https://github.com/devfile/devworkspace-operator)
          to provision enviroments which support common cloud CLI tools. When this
          operator is installed, the DevWorkspace Operator will be installed as a
          dependency.

          ## How to Install
          Press the **Install** button, choose the upgrade strategy, and wait for the
          **Installed** Operator status.

          When the operator is installed, you will see a terminal button appear on the
          top right of the console after refreshing the OpenShift console window.

          ## How to Uninstall
          The Web Terminal Operator requires manual steps to fully uninstall the operator.
          As the Web Terminal Operator is designed as a way to access the OpenShift
          cluster, web terminal instances store user credentials. To avoid exposing these
          credentials to unwanted parties, the operator deploys webhooks and finalizers
          that aren't removed when the operator is uninstalled. See the
          [uninstall guide](https://docs.openshift.com/container-platform/latest/web_console/web_terminal/uninstalling-web-terminal.html)
          for more details.

          ## Documentation
          Documentation for this Operator is available at https://docs.openshift.com/container-platform/latest/web\_console/web\_terminal/installing-web-terminal.html
        displayName: Web Terminal
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - workspace
        - devtools
        - developer
        - ide
        - terminal
        links:
        - name: Web Terminal Repo
          url: https://github.com/redhat-developer/web-terminal-operator/
        maintainers:
        - email: aobuchow@redhat.com
          name: Andrew Obuchowicz
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/web-terminal/web-terminal-tooling-rhel9@sha256:5c24220f884dcdf1b1e5ac1e20dc6b7c8c4300bb89e8f118c0b11c331a56ab3f
        - registry.redhat.io/web-terminal/web-terminal-exec-rhel9@sha256:cfc8200340655a045f45d02fa327538f87e98d6369bdef2b46cf447053c44426
        - registry.redhat.io/web-terminal/web-terminal-rhel9-operator@sha256:0478d14e92df84fdfbf4584384d34dc9a71427a6487c2564d2eb7815ba1ac12b
        version: 1.11.1+0.1747215995.p
      entries:
      - name: web-terminal.v1.11.1-0.1747215995.p
        version: 1.11.1+0.1747215995.p
      - name: web-terminal.v1.11.1
        version: 1.11.1
      - name: web-terminal.v1.11.0
        version: 1.11.0
      - name: web-terminal.v1.10.1
        version: 1.10.1
      - name: web-terminal.v1.10.1-0.1740684238.p
        version: 1.10.1+0.1740684238.p
      - name: web-terminal.v1.10.0-0.1731481377.p
        version: 1.10.0+0.1731481377.p
      - name: web-terminal.v1.10.0-0.1732652667.p
        version: 1.10.0+0.1732652667.p
      - name: web-terminal.v1.10.0
        version: 1.10.0
      - name: web-terminal.v1.10.0-0.1727169028.p
        version: 1.10.0+0.1727169028.p
      - name: web-terminal.v1.10.0-0.1720435222.p
        version: 1.10.0+0.1720435222.p
      - name: web-terminal.v1.10.0-0.1720402943.p
        version: 1.10.0+0.1720402943.p
      - name: web-terminal.v1.9.0-0.1708477317.p
        version: 1.9.0+0.1708477317.p
      - name: web-terminal.v1.9.0
        version: 1.9.0
      - name: web-terminal.v1.8.0-0.1708477299.p
        version: 1.8.0+0.1708477299.p
      - name: web-terminal.v1.8.0-0.1701199376.p
        version: 1.8.0+0.1701199376.p
      - name: web-terminal.v1.8.0-0.1692219820.p
        version: 1.8.0+0.1692219820.p
      - name: web-terminal.v1.8.0
        version: 1.8.0
      - name: web-terminal.v1.7.0-0.1682321121.p
        version: 1.7.0+0.1682321121.p
      - name: web-terminal.v1.7.0-0.1692219820.p
        version: 1.7.0+0.1692219820.p
      - name: web-terminal.v1.7.0
        version: 1.7.0
      - name: web-terminal.v1.7.0-0.1681197295.p
        version: 1.7.0+0.1681197295.p
      - name: web-terminal.v1.7.0-0.1708477265.p
        version: 1.7.0+0.1708477265.p
      - name: web-terminal.v1.7.0-0.1684429884.p
        version: 1.7.0+0.1684429884.p
      - name: web-terminal.v1.6.0
        version: 1.6.0
      - name: web-terminal.v1.6.0-0.1692219820.p
        version: 1.6.0+0.1692219820.p
      - name: web-terminal.v1.5.1-0.1661829403.p
        version: 1.5.1+0.1661829403.p
      - name: web-terminal.v1.5.1
        version: 1.5.1
      - name: web-terminal.v1.5.0-0.1657220207.p
        version: 1.5.0+0.1657220207.p
      - name: web-terminal.v1.5.0
        version: 1.5.0
      - name: web-terminal.v1.4.0
        version: 1.4.0
      - name: web-terminal.v1.3.0
        version: 1.3.0
      name: fast
    defaultChannel: fast
    packageName: web-terminal
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM Quantum
      provider-url: ""
    name: openshift-qiskit-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-qiskit-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "qiskit.ibm.com/v1alpha1",
                "kind": "QiskitPlayground",
                "metadata": {
                  "name": "qiskitplayground-sample"
                },
                "spec": {
                  "image": "qiskit/jupyter:0.1",
                  "imagePullPolicy": "Always",
                  "resources": {
                    "requests": {
                      "memory": "2G",
                      "cpu": "1"
                    },
                    "limits": {
                      "memory": "3G",
                      "cpu": "1.5"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: docker.io/qiskit/operator:0.1
          createdAt: 2021-004-23 22:25:53
          description: |-
            Launches a developement environment Jupyter notebok development environment for
            developing quantum applications using Qiskit.
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator
          support: https://github.com/qiskit-community/openshift-quantum-operators/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QiskitPlayground is the Schema for the qiskitplaygrounds
              API
            displayName: Qiskit Playground
            kind: QiskitPlayground
            name: qiskitplaygrounds.qiskit.ibm.com
            version: v1alpha1
        description: |-
          ### Introduction
          Launches a developement Jupiter lab environment with Qiskit.

          QiskitPlayground is a Jupyter notebook with the pre installed qiskit libraries for implementing quantum application development.
          It comes with pre installed python packages for visualising results. QiskitPlayground can run quantum
          applications on simulator or on real quantum machines using IBM Q account.

          ### Documentation
          For detailed installation guide [follow](https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator)

          ### Contributing
          You can contribute by
            * Raising any [issues](https://github.com/qiskit-community/openshift-quantum-operators/issues) you find using ibm-quantum-operator community operator
            * Fixing issues by opening [Pull Requests](https://github.com/qiskit-community/openshift-quantum-operators/pulls)
            * Talking about openshift-qiskit-operator

          ### License
          kubernetes-qiskit-operator is licensed under the [Apache 2.0 license](https://github.com/qiskit-community/openshift-quantum-operators/blob/master/LICENSE)

          #### Note
          Any api could be changed any time without any notice. That said, your feedback is very important and appreciated to make this project more stable and useful.
        displayName: QiskitPlayground
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Qiskit
        - Open Source
        - Quantum
        links:
        - name: GitHub
          url: https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator
        maintainers:
        - email: blublinsky@ibm.com
          name: boris
        maturity: alpha
        provider:
          name: IBM Quantum
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - docker.io/qiskit/operator:0.1
        version: 0.2.1
      entries:
      - name: openshift-qiskit-operator.v0.2.1
        version: 0.2.1
      - name: openshift-qiskit-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: openshift-qiskit-operator
    provider:
      name: IBM Quantum
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: ""
    name: hawtio-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hawtio-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.14.0",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/hawtio/operator:0.5.0
          createdAt: "2022-04-08T00:00:00Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Hawtio Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - docker.io/hawtio/operator:0.5.0
        version: 0.5.0
      entries:
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: alpha
    - currentCSV: hawtio-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/hawtio/operator:1.2.0
          createdAt: "2024-12-02T17:22:48Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          olm.skipRange: '>=1.0.0 <1.0.2'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Hawtio console that eases the discovery and management
              of hawtio-enabled applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#jvm-agent).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/hawtio/operator:1.2.0
        version: 1.2.0
      entries:
      - name: hawtio-operator.v1.2.0
        version: 1.2.0
      - name: hawtio-operator.v1.1.1
        version: 1.1.1
      - name: hawtio-operator.v1.1.0
        version: 1.1.0
      - name: hawtio-operator.v1.0.1
        version: 1.0.1
      - name: hawtio-operator.v1.0.0
        version: 1.0.0
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: latest
    - currentCSV: hawtio-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/hawtio/operator:1.2.0
          createdAt: "2024-12-02T17:22:48Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          olm.skipRange: '>=1.0.0 <1.0.2'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Hawtio console that eases the discovery and management
              of hawtio-enabled applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#jvm-agent).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/hawtio/operator:1.2.0
        version: 1.2.0
      entries:
      - name: hawtio-operator.v1.2.0
        version: 1.2.0
      - name: hawtio-operator.v1.1.1
        version: 1.1.1
      - name: hawtio-operator.v1.1.0
        version: 1.1.0
      - name: hawtio-operator.v1.0.1
        version: 1.0.1
      - name: hawtio-operator.v1.0.0
        version: 1.0.0
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: stable-v1
    defaultChannel: stable-v1
    packageName: hawtio-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "docker.io/percona/percona-server-mongodb:7.0.18-11",
                  "imagePullPolicy": "Always",
                  "initImage": "docker.io/percona/percona-server-mongodb-operator:1.20.1",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:2.44.1",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "docker.io/percona/percona-backup-mongodb:2.9.1",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.io/percona/percona-server-mongodb-operator:1.20.1
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.20.1
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - docker.io/percona/percona-server-mongodb-operator:1.20.1
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator.v1.20.0
        version: 1.20.0
      - name: percona-server-mongodb-operator.v1.19.1
        version: 1.19.1
      - name: percona-server-mongodb-operator.v1.19.0
        version: 1.19.0
      - name: percona-server-mongodb-operator.v1.18.0
        version: 1.18.0
      - name: percona-server-mongodb-operator.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: directpv
      provider-url: ""
    name: minio-volume-manager-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-volume-manager-operator.v5.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.quay.io/v1alpha1",
                "kind": "AistorVolumemanager",
                "metadata": {
                  "name": "aistorvolumemanager-sample"
                },
                "spec": {
                  "annotations": {},
                  "controller": {
                    "podAnnotations": {},
                    "podLabels": {},
                    "replicas": 3
                  },
                  "images": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": []
                  },
                  "labels": {},
                  "license": "",
                  "nodeserver": {
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-10-05T23:18:24Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AistorVolumemanager
            name: aistorvolumemanagers.charts.quay.io
            version: v1alpha1
        description: directpv
        displayName: directpv
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - directpv
        links:
        - name: Private Dpv Operator
          url: https://private-dpv-operator.domain
        maintainers:
        - email: cesarcelis@minio.io
          name: celis
        maturity: alpha
        provider:
          name: directpv
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - quay.io/cniackz4/directpv-operator@sha256:af06e57661a7e05809e44c3fac571ed949cc04fd656fdb0c8324b533778eb604
        version: 5.0.3
      entries:
      - name: minio-volume-manager-operator.v5.0.3
        version: 5.0.3
      - name: private-dpv-operator.v5.0.0
        version: 5.0.0
      name: stable
    defaultChannel: stable
    packageName: minio-volume-manager-operator
    provider:
      name: directpv
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-commonui-operator-app
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-commonui-operator.v4.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operators.ibm.com/v1alpha1",
                "kind": "CommonWebUI",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "example-commonwebui",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator"
                  },
                  "name": "example-commonwebui"
                },
                "spec": {
                  "commonWebUIConfig": {
                    "cpuLimits": "300",
                    "cpuMemory": "256",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.8.0",
                    "ingressPath": "/common-nav",
                    "landingPage": "",
                    "requestLimits": "300",
                    "requestMemory": "256",
                    "serviceName": "common-web-ui"
                  },
                  "globalUIConfig": {
                    "cloudPakVersion": "3.6.0",
                    "defaultAdminUser": "admin",
                    "defaultAuth": "",
                    "enterpriseLDAP": "",
                    "enterpriseSAML": "",
                    "osAuth": "",
                    "sessionPollingInterval": 5000
                  },
                  "operatorVersion": "4.8.0",
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "256Mi"
                    },
                    "requests": {
                      "cpu": "300m",
                      "memory": "256Mi",
                      "ephemeral-storage": "251Mi"
                    }
                  },
                  "version": "4.8.0"
                }
              },
              {
                "apiVersion": "foundation.ibm.com/v1",
                "kind": "NavConfiguration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "common-web-ui-config",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator",
                    "default": "true",
                    "name": "common-web-ui-config"
                  },
                  "name": "common-web-ui-config"
                },
                "spec": {
                  "about": {
                    "copyright": "© 2018, 2020 IBM. All rights reserved.",
                    "licenses": [
                      "yq, version 3.3.0, MIT+GPL",
                      "MongoDB, version 4.0.16 Community Edition, SSPL",
                      "Ansible: © 2017 Red Hat, Inc., http://www.redhat.com; © Henry Graham (hzgraham) \u003cHenry.Graham@mail.wvu.edu\u003e",
                      "calico-bird: © 1998–2008, Martin Mares \u003cmj@ucw.cz\u003e; © 1998–2000 Pavel Machek \u003cpavel@ucw.cz\u003e;",
                      "© 1998–2008, Ondrej Filip \u003cfeela@network.cz\u003e; © 2009–2013,  CZ.NIC z.s.p.o.",
                      "chrony: © Richard P. Curnow  1997-2003, GPL v2",
                      "collectd, © 2017-2018, version 5.7.2, GPL v2, \u003chttps://github.com/collectd/collectd/tree/collectd-5.7.2\u003e",
                      "crudini: © Pádraig Brady \u003cP@draigBrady.com\u003e",
                      "Galera-3: © 2007–2014 Codership Oy \u003cinfo@codership.com\u003e",
                      "glusterfs: © 2010–2013+ James Shubin \u003chttps://ttboj.wordpress.com/\u003e",
                      "haproxy: © 2000–2013  Willy Tarreau \u003cw@1wt.eu\u003e",
                      "heketi v6.0.0: © 2015 The heketi Authors, GPL v2",
                      "heketi v8.0.0: © 2015 The heketi Authors, GPL v2",
                      "heketi-master/apps.app.go: © 2015 The heketi Authors",
                      "heketi-master/client/api/go-client/backup.go: © 2016 The heketi Authors",
                      "heketi-master/doc/man/heketi-cli.8: © 2016 The heketi Authors",
                      "heketi-master/extras/docker/gluster/gluster-setup.sh: © 2016 Red Hat, Inc. \u003chttp://www.redhat.com\u003e",
                      "ieee-data: © 2013 Luciano Bello \u003cluciano@debian.org\u003e",
                      "javax.mail: © 2017 Oracle and/or its affiliates. All rights reserved.",
                      "keepalived: © 2001-2017 Alexandre Cassen \u003cacassen@gmail.com\u003e",
                      "libonig2: © 2006–2008 Max Kellermann \u003cmax@duempel.org\u003e; © 2014–2015 Jörg Frings-Fürst \u003cdebian@jff-webhosting.net\u003e",
                      "libtomcrypt: © 2004 Sam Hocevar \u003csam@hocevar.net\u003e, GPL v2",
                      "mariadb-common: © 2018 MariaDB. All rights reserved.",
                      "mariaDB: © 2018 MariaDB. All rights reserved. \u003chttps://mariadb.com/\u003e",
                      "mariadb-server: © 2018 MariaDB. All rights reserved.",
                      "minitar: © 2004 Mauricio Julio Fernandez Pradier and Austin Ziegler",
                      "MongoDB: © 2007 Free Software Foundation, Inc. \u003chttp://fsf.org/\u003e",
                      "nvmi-cli: © 1989, 1991 Free Software Foundation, Inc., GPL v2",
                      "OpenJDK: © 2018 Oracle Corporation and/or its affiliates",
                      "openshift-mariadb-galera: © 2007 Free Software Foundation, Inc. \u003chttp://fsf.org/\u003e",
                      "percona-xtrabackup: © 2006–2018 Percona LLC.",
                      "pwgen: © Christian Thöing \u003cc.thoeing@web.de\u003e",
                      "rdoc: © 2001–2003 Dave Thomas, The Pragmatic Programmers",
                      "readline: © Chet Ramey \u003cchet.ramey@case.edu\u003e",
                      "John the Ripper password cracker: © 1996–2013 by Solar Designer \u003csolar@openwall.com\u003e",
                      "spdx-exceptions: © 2018 SPDX Workgroup a Linux Foundation Project. All rights reserved.",
                      "socat: © 2001–2010 Gerhard Rieger",
                      "sshpass: © 2006, 2008 Lingnu Open Source Consulting Ltd. \u003chttp://www.lingnu.com\u003e",
                      "timelimit: © 2001, 2007 - 2010  Peter Pentchev, GPL v2",
                      "ua-parser-js: © 2012-2018 Faisal Salman \u003cf@faisalman.com\u003e, GPL v2",
                      "ubuntu-cloud-keyring: © 2010 Michael Vogt \u003cmichael.vogt@canonical.com\u003e",
                      "unboundid-ldapsdk: © 2015 UnboundID. The LDAP SDK for Java is developed by UnboundID. \u003cinfo@unboundid.com\u003e",
                      "xmpp4r: © Lucas Nussbaum \u003clucas@lucas-nussbaum.net\u003e, Stephan Maka \u003cstephan@spaceboyz.net\u003e, and others.",
                      "module-assistant: © 2003-2008 Eduard Bloch \u003cblade@debian.org\u003e, version 0.11.8, GPL v2;",
                      "© 2009 Cyril Brulebois \u003ckibi@debian.org\u003e, version 0.11.8, GPL v2; © 2013-2018 Andreas Beckmann \u003canbe@debian.org\u003e, version 0.11.8, GPL v2",
                      "module-init-tools: © 2011 ProFUSION embedded systems, version 22, GPL v2",
                      "thin: © 2017 Marc-Andre Cournoyer \u003cmacournoyer@gmail.com\u003e, version 1.7.2, GPL v2",
                      "gosu, © 1999-2014, version 1.1, GPL v3",
                      "mercurial (Python), © 2006-2018 ,version v4.5.3, GPL v2",
                      "garden-runc, © 2015-Present CloudFoundry.org Foundation, Inc. All Rights Reserved, version 1.17.0, GPLv2",
                      "libtomcrypt0, © 2003-2007 Tom St Denis \u003ctomstdenis@gmail.com\u003e, version 1.17-7, GPLv2",
                      "console-setup-min, © 1999,2000,2001,2002,2003,2006,2007,2008,2009,2010,2011 Anton Zinoviev, \u003canton@lml.bas.bg\u003e,version 1.108, GPLv2",
                      "dracut, © 2009 Harald Hoyer \u003charald@redhat.com\u003e, version 044+3-3, GPLv2",
                      "dracut-core, © 2009 Harald Hoyer \u003charald@redhat.com\u003e, version 044+3-3, GPLv2",
                      "g++, version 5.4.0-6ubuntu, GPL v2",
                      "libstdc++6, version 5.4.0-6ubuntu, GPL v3",
                      "libstdc++-5-dev, version 5.4.0-6ubuntu, GPL v3",
                      "docker-engine-selinux, version 3b5fac4, GPLv2",
                      "unorm, version 1.5.0, GPL v2",
                      "psmisc, version 22.20, GPL v2",
                      "lvm2-devel, version 2.0.2, GPL v2",
                      "nfs-utils, version 1.3, GPL v2",
                      "popt-static, version 1.13, GPL v2",
                      "sysvinit-tools, version 2.88, GPL v2",
                      "stunnel, version 5.53, GPL v2",
                      "stunnel, version 5.39, GPL v2",
                      "LVM2, version 2.02.180-10.el7_6.2, GPL v2",
                      "sysdig, version 2c43237, GPL",
                      "chisels, version 9722dbc, GPL",
                      "MongoDB, version 4.0.12, SSPL",
                      "ffi (Ruby Gem), 1.11.1, GPL",
                      "inotify-tools, v3.14, GPL v2",
                      "logrotate, v3.8.6, GPL v2",
                      "checker-qual, version 2.0.0, GPLv2",
                      "ocrad-bower, v1.0.0, GPL v3",
                      "Grafana, v7.5.12, AGPL",
                      "IBM Semeru Runtime Open Edition binaries, GPL v2",
                      "checker-compat-qual 2.5.5, GPL-2.0",
                      "javax.annotation-api 1.3.1, GPL-2.0"
                    ],
                    "logoUrl": "IBM Cloud Pak | Administration Hub",
                    "logoAltText": "IBM Cloud Pak | Administration Hub"
                  },
                  "header": {
                    "disabledItems": [
                      "createResource",
                      "catalog"
                    ],
                    "docUrlMapping": "http://ibm.biz/cpcs_adminui",
                    "logoAltText": "IBM Cloud Pak Administration Hub",
                    "logoHeight": "47px",
                    "logoUrl": "/common-nav/graphics/ibm-cloudpack-logo.svg",
                    "logoWidth": "190px"
                  },
                  "login": {
                    "loginDialog": {
                      "acceptText": "Your acceptance text here",
                      "dialogText": "You must set your dialog for this environment",
                      "enable": false,
                      "headerText": "Header text here"
                    },
                    "logoAltText": "Cloud Pak",
                    "logoHeight": "47px",
                    "logoUrl": "/common-nav/api/graphics/logincloudpak.svg",
                    "logoWidth": "190px"
                  },
                  "navItems": [
                    {
                      "id": "providers",
                      "label": "Identity providers",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/realms"
                    }
                  ]
                }
              },
              {
                "apiVersion": "foundation.ibm.com/v1",
                "kind": "NavConfiguration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "icp4i",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator",
                    "name": "icp4i"
                  },
                  "name": "icp4i"
                },
                "spec": {
                  "header": {
                    "disabledItems": [
                      "createResource",
                      "catalog",
                      "bookmark"
                    ],
                    "logoAltText": "Cloud Pak for Integration",
                    "logoUrl": "/common-nav/graphics/ibm-cloudpak-integration.svg"
                  },
                  "navItems": [
                    {
                      "detectionServiceName": true,
                      "id": "metering",
                      "label": "Metering",
                      "serviceId": "metering-ui",
                      "serviceName": "metering-ui",
                      "url": "/metering/dashboard?ace_config={ 'showClusterData': false }\u0026dashboard=cpi.icp.main"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "monitoring",
                      "isAuthorized": [
                        "Administrator",
                        "ClusterAdministrator",
                        "Operator"
                      ],
                      "label": "Monitoring",
                      "serviceId": "monitoring-ui",
                      "serviceName": "ibm-monitoring-grafana",
                      "target": "_blank",
                      "url": "/grafana"
                    },
                    {
                      "id": "id-access",
                      "label": "Identity and access",
                      "serviceId": "webui-nav"
                    },
                    {
                      "id": "providers",
                      "parentId": "id-access",
                      "label": "Identity providers",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/realms?useNav=icp4i"
                    },
                    {
                      "id": "teams-ids",
                      "parentId": "id-access",
                      "label": "Teams and service IDs",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/teams?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "logging",
                      "label": "Logging",
                      "serviceId": "kibana",
                      "serviceName": "kibana",
                      "target": "_blank",
                      "url": "/kibana"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "releases",
                      "label": "Helm Releases",
                      "serviceId": "catalog-ui",
                      "serviceName": "catalog-ui",
                      "url": "/catalog/instances?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "repos",
                      "label": "Helm Repositories",
                      "serviceId": "catalog-ui",
                      "serviceName": "catalog-ui",
                      "url": "/catalog/repositories?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "licensing",
                      "label": "Licensing",
                      "serviceId": "ibm-license-service-reporter",
                      "url": "/license-service-reporter",
                      "isAuthorized": [
                        "ClusterAdministrator",
                        "CloudPakAdministrator"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operators.ibm.com/v1alpha1",
                "kind": "SwitcherItem",
                "metadata": {
                  "name": "example-switcheritem"
                },
                "spec": {
                  "cloudPakInfo": {
                    "display": "Administration",
                    "label": "administration-hub",
                    "landingPage": "/common-nav/dashboard",
                    "logoURL": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
          description: The IBM Common Web UI delivers the common header API and the
            identity and access pages for IBM Cloud Platform Common Services.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.8.0
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-commonui-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            displayName: CommonWebUI service
            kind: CommonWebUI
            name: commonwebuis.operators.ibm.com
            version: v1alpha1
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            kind: NavConfiguration
            name: navconfigurations.foundation.ibm.com
            version: v1
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            kind: SwitcherItem
            name: switcheritems.operators.ibm.com
            version: v1alpha1
        description: "**Important:** Do not install this operator directly. Only install
          this operator using the IBM Common Services Operator. For more information
          about installing this operator and other Common Services operators, see
          [Installer documentation](http://ibm.biz/cpcs_opinstall). Additionally,
          you can exit this panel and navigate to the IBM Common Services tile in
          OperatorHub to learn more about the operator.\n\n If you are using this
          operator as part of an IBM Cloud Pak, see the documentation for that IBM
          Cloud Pak to learn more about how to install and use the operator service.
          For more information about IBM Cloud Paks, see [IBM Cloud Paks that use
          Common Services](http://ibm.biz/cpcs_cloudpaks).\n\n You can use the ibm-commonui-operator
          to install the Common Web UI service for the IBM Cloud Platform Common Services
          and access the Common Web UI console. You can use the Common Web UI console
          to access information and features from other IBM Cloud Platform Common
          Services or IBM Cloud Paks that you install. \n\nFor more information about
          the available IBM Cloud Platform Common Services, see the [IBM Knowledge
          Center](http://ibm.biz/cpcsdocs). \n## Supported platforms \n\n Red Hat
          OpenShift Container Platform 4.2 or newer installed on one of the following
          platforms: \n\n- Linux x86_64 \n- Linux on Power (ppc64le) \n- Linux on
          IBM Z and LinuxONE \n## Prerequisites\n\n The Common Web UI service has
          dependencies on other IBM Cloud Platform Common Services. Before you install
          this operator, you need to first install the operator dependencies and prerequisites:
          \n For the list of operator dependencies, see the IBM Knowledge Center [Common
          Services dependencies documentation](http://ibm.biz/cpcs_opdependencies).
          \n For the list of prerequisites for installing the operator, see the IBM
          Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).
          \n## Documentation \n\n To install the operator with the IBM Common Services
          Operator follow the the installation and configuration instructions within
          the IBM Knowledge Center. \n- If you are using the operator as part of an
          IBM Cloud Pak, see the documentation for that IBM Cloud Pak, for a list
          of IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).
          \n- If you are using the operator with an IBM Containerized Software, see
          the IBM Cloud Platform Common Services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall)."
        displayName: Ibm Common UI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        - Web Console
        - Common Services
        links:
        - name: GitHub
          url: https://github.com/IBM/ibm-commonui-operator
        maintainers:
        - email: siddarth.patel@ibm.com
          name: Siddarth Patel
        - email: sgrube@us.ibm.com
          name: Steve Grube
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:e4eb6a75e0d207232e479693d9c99ba863f8e62830f3191ca4d1e7ceaf3586d1
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator-bundle@sha256:e4340bd9481da998a6b89e94c884c46e17cfb50d9d0c1099f39a3c54e500d2b2
        version: 4.8.0
      entries:
      - name: ibm-commonui-operator.v4.8.0
        version: 4.8.0
      name: v4.7
    defaultChannel: v4.7
    packageName: ibm-commonui-operator-app
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent
      provider-url: ""
    name: cilium-enterprise
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cilium.v1.10.13-xed86eae
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"prometheus":{"serviceMonitor":{"enabled":false}},"mode":"cluster-pool"},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium Enterprise - eBPF-based Networking, Security, and Observability
        displayName: Cilium Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Enterprise Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Cilium Enterprise
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-ui-enterprise@sha256:c9921c65892a5293a4f1bef787636d6ae58cec1af65da06862a47e3450afdc6e
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/isovalent/hubble-relay@sha256:9145b03acc954efde54e1ab9da878462526c5e820b83a92e32cc24291f784d33
        - quay.io/isovalent/clustermesh-apiserver@sha256:eb3ee3d2cb07e6c5966e4bc2e50615949303be5d5128e156d5c0dbe8404dc093
        - quay.io/isovalent/hubble-enterprise-metadata@sha256:72c7aa3961f14df413569cd156820a45553bb24d393b11246a18bf8b6553e022
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        - quay.io/isovalent/cilium@sha256:6887537789603946543aaa370a3ec592287a235231bf16aa386e091823285515
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:cc1f0e6b27ccda8a843dfb7fa4a8fef052ba29716646557815677c5d1bea2a1a
        - quay.io/isovalent/hubble-export-fluentd@sha256:05cb13f533f2c89b98a47cd58badda07add54077cd22283ca5028cb39f2e4486
        - quay.io/isovalent/hubble-export-s3@sha256:ce4f71f7f13301133b7d471b87fb81108b2c2da7fc29865ef96bc8431b2be72b
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/oauth2-proxy/oauth2-proxy@sha256:ecd26b74a01f2b547ddaed4d32d35f8f5e09c378d5c1fc6cfa63f0adf659ac2b
        - quay.io/isovalent/hubble-enterprise-operator@sha256:0d5e6a5d9c07780c18e5aa81d8f1c95c399bff2d36a44fa5f61ac89a788ca83b
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/hubble-ui-enterprise-backend@sha256:e892cba6dafcdf6704df3135a0b38d11894a5f276ace40683e74cb62b9c311d8
        - quay.io/isovalent/operator-generic@sha256:1cc0b0d2e71564ad8759be326c66ea13b4955b9441fb67e906da534e1b38aefa
        - quay.io/cilium/hubble-export-stdout@sha256:f1b1475a3c89950481264f5f8ca63eea3b911a45dc83fdf15122d695a0fba9b3
        - quay.io/isovalent/hubble-enterprise@sha256:1e1d3e6c199546b2398f4166ff64210c04075bb407c8f07c5be8d3db08560d4a
        - quay.io/isovalent/hubble-rbac@sha256:03f3304c157024356a6de13c3d91b4506b8e54d587920d80ebcc76cbcfeaeda6
        version: 1.10.13+xed86eae
      entries:
      - name: cilium.v1.10.13-xed86eae
        version: 1.10.13+xed86eae
      - name: cilium.v1.10.12-xb203fae
        version: 1.10.12+xb203fae
      - name: cilium.v1.10.11-x10ae342
        version: 1.10.11+x10ae342
      name: "1.10"
    - currentCSV: cilium.v1.11.21-xd0d07b1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium Enterprise - eBPF-based Networking, Security, and Observability
        displayName: Cilium Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Enterprise Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Cilium Enterprise
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - cgr.dev/chainguard/oauth2-proxy:7.4.0@sha256:54857c70321004d3e0b8b8d7208ce5422c1f71c18db5be02d2543f187ea8e8e2
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:372511f6fa3523dadfe9128caf77520cf74ce24df4a19e1583ee197375e06014
        - quay.io/isovalent/hubble-ui-enterprise:v0.23.0@sha256:02635d1392cca7e09ab7797f69147880068ba7e3b0d865be97eeab33edd646a0
        - quay.io/cilium/certgen:v0.1.5@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/startup-script:62093c5c233ea914bfa26a10ba41f8780d9b737f@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/isovalent/clustermesh-apiserver:v1.11.21-cee.1@sha256:e16d8375ed8a1c3a872047b3aacf3469bde43552f1a51ee3e159ef2e9fea3e97
        - quay.io/coreos/etcd:v3.4.13@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/isovalent/hubble-relay:v1.11.21-cee.1@sha256:5933dd1f8111d6bca0c68db9247385b5d215d5b613095bafab8244098f2c5a26
        - quay.io/isovalent/hubble-ui-enterprise-backend:v0.23.0@sha256:2fbdad86bb5244075f31487262d790f6fa75cdd993f868b952252905c118621c
        - quay.io/isovalent/hubble-rbac:v1.3.0@sha256:4c106dd133861a678eb24cf98b35d5023cccca0834c954c49b68514c366aedf0
        - quay.io/isovalent/hubble-enterprise-operator:v1.10.5@sha256:31b18a1ea7002936f04dd070549e0c9a02f8a33394666fc5b0523ab1350c6120
        - quay.io/isovalent/hubble-export-stdout:v1.0.3@sha256:bef600089c1671fce11bb4bcb98b81f25e65152a1a2f13b64ea637cfc5779c11
        - quay.io/isovalent/hubble-export-s3:v1.0.0@sha256:89c178c0a010166c68819ab5de32ab59b64a88aa23a05d8e4f78adb1a4b6c3d3
        - quay.io/isovalent/hubble-export-fluentd:v1.5.1@sha256:eaf42a84821509b69e2ef2b658796062e79a90e63045a98689da3da99a12e492
        - quay.io/isovalent/hubble-enterprise:v1.10.5@sha256:340700d4e28ded2aa9c50ff9f938996bb43faa84e0b2ba38ec79452648d7e5d7
        - quay.io/isovalent/operator-generic:v1.11.21-cee.1@sha256:25b668c1febe1d1a70b63dfe926be4e55fbf77fd186e9df7581e41d7e9813469
        - quay.io/isovalent/cilium:v1.11.21-cee.1@sha256:f41241a15cd96d26727aa1ec413ace643728ddbe42f7fb7b234dbda22bf3950b
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:d3148bf3b9765467936ee067b25b5ef3683d3e55260f517c7fe271ac0e19cb47
        version: 1.11.21+xd0d07b1
      entries:
      - name: cilium.v1.11.21-xd0d07b1
        version: 1.11.21+xd0d07b1
      - name: cilium.v1.11.20-x0f6d966
        version: 1.11.20+x0f6d966
      name: "1.11"
    - currentCSV: cilium.v1.12.21-x3861686
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.12.14 <1.12.21'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/cilium:v1.12.21-cee.1@sha256:3a462308cf12b572db987c03d190ee33691ccb04775d01beb192703a52f8d47b
        - quay.io/isovalent/hubble-relay:v1.12.21-cee.1@sha256:9d8dadc3f330acf0fa169c428d5fdf1d138d0294065eec5c628f6647106fe8b6
        - quay.io/coreos/etcd:v3.5.4@sha256:795d8660c48c439a7c3764c2330ed9222ab5db5bb524d8d0607cac76f7ba82a3
        - quay.io/isovalent/hubble-export-s3:v1.0.0@sha256:89c178c0a010166c68819ab5de32ab59b64a88aa23a05d8e4f78adb1a4b6c3d3
        - quay.io/isovalent/hubble-enterprise-operator:v1.10.5@sha256:31b18a1ea7002936f04dd070549e0c9a02f8a33394666fc5b0523ab1350c6120
        - quay.io/isovalent/clustermesh-apiserver:v1.12.21-cee.1@sha256:3f6549409966086a49e338c7829b9141a73ef2ebeebaac91db430aeb3ccd2e22
        - quay.io/isovalent/hubble-ui-enterprise:v0.28.1@sha256:11dfba0e4fcf43415aa6e8b3b6e98a5c4f468ec67b377328ad48b92037bbfbb0
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:f95c9e8002d8e381d604752b1d59b75595badbdb0b11237ac1cac4ecd04f4800
        - quay.io/isovalent/hubble-ui-enterprise-backend:v0.28.1@sha256:515aeb8166dd6995b04d4cd3333c94a740182943ba4e90a38871ee89c0b64e92
        - quay.io/isovalent/hubble-export-fluentd:v1.5.1@sha256:eaf42a84821509b69e2ef2b658796062e79a90e63045a98689da3da99a12e492
        - quay.io/cilium/certgen:v0.1.8@sha256:4a456552a5f192992a6edcec2febb1c54870d665173a33dc7d876129b199ddbd
        - quay.io/isovalent/hubble-enterprise:v1.10.5@sha256:340700d4e28ded2aa9c50ff9f938996bb43faa84e0b2ba38ec79452648d7e5d7
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/isovalent/hubble-rbac:v1.3.0@sha256:4c106dd133861a678eb24cf98b35d5023cccca0834c954c49b68514c366aedf0
        - quay.io/isovalent/operator-generic:v1.12.21-cee.1@sha256:4e19cda799a747b69f53b5e5a860ba256399d18221d5a4bbf243142ba1c69bb8
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/hubble-export-stdout:v1.0.3@sha256:bef600089c1671fce11bb4bcb98b81f25e65152a1a2f13b64ea637cfc5779c11
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:b6e40d76f0925e76f0a5b181be3f6c62728fa3e4aaef979caf9ea245f4ed03da
        - quay.io/cilium/startup-script:62093c5c233ea914bfa26a10ba41f8780d9b737f@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        version: 1.12.21+x3861686
      entries:
      - name: cilium.v1.12.21-x3861686
        version: 1.12.21+x3861686
      - name: cilium.v1.12.19-xca684e0
        version: 1.12.19+xca684e0
      - name: cilium.v1.12.18-xcb1adcf
        version: 1.12.18+xcb1adcf
      - name: cilium.v1.12.17-xae21c14
        version: 1.12.17+xae21c14
      - name: cilium.v1.12.16-x38e0b62
        version: 1.12.16+x38e0b62
      - name: cilium.v1.12.15-x8ca64a3
        version: 1.12.15+x8ca64a3
      - name: cilium.v1.12.14-xc86b260
        version: 1.12.14+xc86b260
      - name: cilium.v1.12.13-x2b5cec9
        version: 1.12.13+x2b5cec9
      - name: cilium.v1.12.12-xbd7b835
        version: 1.12.12+xbd7b835
      - name: cilium.v1.12.11-xffdea76
        version: 1.12.11+xffdea76
      - name: cilium.v1.12.10-x4a0e9e7
        version: 1.12.10+x4a0e9e7
      - name: cilium.v1.12.9-x4be66e4
        version: 1.12.9+x4be66e4
      - name: cilium.v1.12.8-x87b90a9
        version: 1.12.8+x87b90a9
      - name: cilium.v1.12.7-x53a2aab
        version: 1.12.7+x53a2aab
      name: "1.12"
    - currentCSV: cilium-enterprise.v1.13.21-x96449fb
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.7 <1.13.21+x96449fb'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/startup-script:c54c7edeab7fde4da68e59acd319ab24af242c3f@sha256:a9d0f451353f1b2f29ed3868e230ecfae10150eab9fd521de67adb3c8ca9c33a
        - quay.io/cilium/certgen:v0.1.13@sha256:01802e6a153a9473b06ebade7ee5730f8f2c6cc8db8768508161da3cdd778641
        - quay.io/isovalent/hubble-export-fluentd:v1.11.0@sha256:50cd2e4c2d04e8befcc2ae1a1463f07cba01edb9b546fb455fe7976311165385
        - quay.io/isovalent/hubble-enterprise:v1.11.6@sha256:eb60d59d1fb1b136855829274b31805c4206f8d050c3bf068e63a552576c1b38
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:b04e75456359b50b18e75464da5d7b467fdfeb058fc08ce3a10f057ba5f60107
        - quay.io/isovalent/cilium:v1.13.21-cee.2@sha256:f760cd85fa64eb8a197ec282644a8db630745a8a2854ee08aa2662cf67188db5
        - quay.io/isovalent/hubble-export-s3:v1.0.2@sha256:518daea1074d260d824c3e03a93ab6b2a7a5f4c22b8e07938770add8257f4a52
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:fc7046dee1a204163e68a6d9a42fec200b02129493a7d166f899dd12e2763364
        - quay.io/isovalent/hubble-rbac:v1.5.6@sha256:739ca45ea09829b0fa8a6198901b34eb5254df55be071bc854bbd37cae12a557
        - quay.io/isovalent/hubble-relay:v1.13.21-cee.2@sha256:aaf5f959a76d6b517e627d2d0b09dab86ae942616819a9a95f057ed4f45ea480
        - quay.io/isovalent/operator-generic:v1.13.21-cee.2@sha256:6f2d0d575e3f74395c959833992cd5fbb6971f4e292df1b7cf45be8b07be5e7a
        - quay.io/isovalent/clustermesh-apiserver:v1.13.21-cee.2@sha256:980b0235fd16817ecc4457ad4317a4601993b76df2e81b07a53f6f437c2705af
        - quay.io/isovalent/hubble-export-stdout:v1.0.4@sha256:a0ff77efa890478f18f5deeba2cd1e2dfd6c5c77981adac57b14d7734de71ff1
        - quay.io/isovalent/hubble-ui-enterprise:v1.0.3@sha256:55bf787c96561f4a460e2bfe5966421e2d85f3f9a38ac2487c6f6486c97cd5af
        - quay.io/isovalent/hubble-ui-enterprise-backend:v1.0.3@sha256:6756d2b808c38efce4765f90641d26aa2e0e79329ecc378a123cd3dc2c439640
        - quay.io/isovalent/hubble-enterprise-operator:v1.11.6@sha256:38253c2e17a8b2e88a226f0976c9cf4a00445212e2886e6f07cbb3a52757324f
        - quay.io/coreos/etcd:v3.5.4@sha256:795d8660c48c439a7c3764c2330ed9222ab5db5bb524d8d0607cac76f7ba82a3
        version: 1.13.21+x96449fb
      entries:
      - name: cilium-enterprise.v1.13.21-x96449fb
        version: 1.13.21+x96449fb
      - name: cilium-enterprise.v1.13.20-x5b70650
        version: 1.13.20+x5b70650
      - name: cilium-enterprise.v1.13.19-xdb27142
        version: 1.13.19+xdb27142
      - name: cilium-enterprise.v1.13.18-x0c21a94
        version: 1.13.18+x0c21a94
      - name: cilium-enterprise.v1.13.17-xc8f905f
        version: 1.13.17+xc8f905f
      - name: cilium-enterprise.v1.13.16-xcfe62ed
        version: 1.13.16+xcfe62ed
      - name: cilium-enterprise.v1.13.15-x1dfd26f
        version: 1.13.15+x1dfd26f
      - name: cilium.v1.13.14-xb70933f
        version: 1.13.14+xb70933f
      - name: cilium.v1.13.12-x453f6ad
        version: 1.13.12+x453f6ad
      - name: cilium.v1.13.11-xf6d6935
        version: 1.13.11+xf6d6935
      - name: cilium.v1.13.10-x9d57fc7
        version: 1.13.10+x9d57fc7
      - name: cilium.v1.13.9-xe8a359b
        version: 1.13.9+xe8a359b
      - name: cilium.v1.13.8-x96e1ca9
        version: 1.13.8+x96e1ca9
      - name: cilium.v1.13.7-x7320e7d
        version: 1.13.7+x7320e7d
      - name: cilium.v1.13.6-xd3714ef
        version: 1.13.6+xd3714ef
      - name: cilium.v1.13.5-x8386cae
        version: 1.13.5+x8386cae
      - name: cilium.v1.13.4-xc8b22ea
        version: 1.13.4+xc8b22ea
      - name: cilium.v1.13.3-x69e9561
        version: 1.13.3+x69e9561
      - name: cilium.v1.13.2-x4b1c6c8
        version: 1.13.2+x4b1c6c8
      name: "1.13"
    - currentCSV: cilium-enterprise.v1.14.22-x6d86c88
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.14.22+x6d86c88'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-enterprise:v1.11.6@sha256:eb60d59d1fb1b136855829274b31805c4206f8d050c3bf068e63a552576c1b38
        - quay.io/isovalent/hubble-rbac:v1.5.2@sha256:29309837e2246a1b6849d7a09d01ceebfdb43fa0ade973d14584f0e52b856b95
        - quay.io/coreos/etcd:v3.5.4@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:148b1029eac14049e2ba84d4d75be1ca43ef2d89951adbdeddec392ad7c9fe1e
        - quay.io/isovalent/hubble-ui-enterprise:v1.2.2@sha256:3e03a58f6eea5d2df195f42ee2817f7b28fa20bbe140f5086937eee8da1b8d56
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/isovalent/hubble-enterprise-operator:v1.11.6@sha256:38253c2e17a8b2e88a226f0976c9cf4a00445212e2886e6f07cbb3a52757324f
        - quay.io/cilium/startup-script:c54c7edeab7fde4da68e59acd319ab24af242c3f@sha256:a9d0f451353f1b2f29ed3868e230ecfae10150eab9fd521de67adb3c8ca9c33a
        - quay.io/isovalent/operator-generic:v1.14.22-cee.1@sha256:a997fe24ff0b7f7df3ee12748413ae53d6e07c803162b0db4d812b916b0cfc74
        - quay.io/isovalent/hubble-export-fluentd:v1.8.1@sha256:ce0cce63a8473d5ae73872fc65d82c54f2875511bb03f0bc8ee8480148c43ee1
        - quay.io/cilium/certgen:v0.1.17@sha256:07959f3125dae659ab9036c14df046e2536f7a4f6c92dabda837388b0344dc74
        - quay.io/isovalent/hubble-relay:v1.14.22-cee.1@sha256:d7ffc880d2e42e0876417ec26115b0e9155ad9a2e7f47c02c434746afa42b29a
        - quay.io/isovalent/clustermesh-apiserver:v1.14.22-cee.1@sha256:588e2b173ab409288eda2d404571184562d3d9fd12e2d9afe7282a5a7d79113a
        - quay.io/cilium/cilium-envoy:v1.32.5-1743506100-0821ef0acdf9f824d47d34e02932be522b3e7233@sha256:305c8b7da56ef1b64e4deb0642acf93a3944f2712e40c99a40bb005e6dc9761b
        - quay.io/isovalent/hubble-ui-enterprise-backend:v1.2.2@sha256:4258f7e0bff24ff1e0b214f62b37097032b6426ba28edd3dfc4f451a34de0110
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:fc7046dee1a204163e68a6d9a42fec200b02129493a7d166f899dd12e2763364
        - quay.io/isovalent/hubble-export-stdout:v1.0.4@sha256:a0ff77efa890478f18f5deeba2cd1e2dfd6c5c77981adac57b14d7734de71ff1
        - quay.io/isovalent/cilium:v1.14.22-cee.1@sha256:0e664a5a9e0edba38ab1a25eb7fa50861bb15f69ed0bbe3f3c792bd909240da3
        - quay.io/isovalent/hubble-export-s3:v1.0.2@sha256:518daea1074d260d824c3e03a93ab6b2a7a5f4c22b8e07938770add8257f4a52
        version: 1.14.22+x6d86c88
      entries:
      - name: cilium-enterprise.v1.14.22-x6d86c88
        version: 1.14.22+x6d86c88
      - name: cilium-enterprise.v1.14.21-xe0c9553
        version: 1.14.21+xe0c9553
      - name: cilium-enterprise.v1.14.20-xe01e743
        version: 1.14.20+xe01e743
      - name: cilium-enterprise.v1.14.19-xc1b9aff
        version: 1.14.19+xc1b9aff
      name: "1.14"
    - currentCSV: cilium-enterprise.v1.15.21-x2661a01
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.15.21+x2661a01'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:bd1a4d47ef48a6eb84c49d979e8f5ce867af5932828dfd3233fa9f78c2eac1f4
        - quay.io/isovalent/cilium-envoy-ubi:v1.33.4-1752151664-7c2edb0b44cf95f326d628b837fcdd845102ba68@sha256:e81fdde48c3d5cd93111e1d51d9f8554adb32ee937601a9124354e0894933b99
        - quay.io/isovalent/operator-generic-ubi:v1.15.21-cee.1@sha256:97daac94735e497e1e897beaa01eb50866abace1aa9a3919fc57971e45f538fe
        - quay.io/isovalent/hubble-ui-frontend-ubi:v1.3.7@sha256:f2b4e935abe3eb94f82dab41c9d3d929fca71d82fa39a6c8002683375b5303a4
        - quay.io/isovalent/oauth2-proxy:7.12.0-20250819@sha256:fe8f76728c36536d498b2a5c11172cb18dc30bbb555df2a6d14c036b29cb9f10
        - quay.io/isovalent/cilium:v1.15.21-cee.1@sha256:39ef22d467783481eccb376fcb5d6bc254ed8f4b3b25032121498db40278f0b8
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.15.21-cee.1@sha256:36d00eb9f8ed45a3d8892942cfae097782dd6509889ef7d1a451f9b0334d17e1
        - quay.io/isovalent/certgen-ubi:v0.1.19@sha256:c32e96b0972c9bcd3896945c46ea23e330c9dea407afee3a637daae637296f4c
        - quay.io/isovalent/hubble-ui-backend-ubi:v1.3.7@sha256:a4b5e7a032c1bbcd969a6873bc12935f8b6aa4c2d3a2ed87f5494fcfcdfd02d4
        - quay.io/isovalent/hubble-relay-ubi:v1.15.21-cee.1@sha256:8461f720ca95a9c543a298bee1c1e56241f56d5426030c13afed8262f2f5c5ff
        - quay.io/isovalent/startup-script-ubi:1755531540-60ee83e@sha256:2e1d5048cbfcd3c77f732da8203a6c792cbf19897b5cf137b9ad23a6a267f6b2
        version: 1.15.21+x2661a01
      entries:
      - name: cilium-enterprise.v1.15.21-x2661a01
        version: 1.15.21+x2661a01
      - name: cilium-enterprise.v1.15.20-x6df37bb
        version: 1.15.20+x6df37bb
      - name: cilium-enterprise.v1.15.19-x0be49ab
        version: 1.15.19+x0be49ab
      - name: cilium-enterprise.v1.15.18-xeeaa70b
        version: 1.15.18+xeeaa70b
      - name: cilium-enterprise.v1.15.17-xdfd9482
        version: 1.15.17+xdfd9482
      - name: cilium-enterprise.v1.15.16-xc7e31ca
        version: 1.15.16+xc7e31ca
      - name: cilium-enterprise.v1.15.15-xf37f22c
        version: 1.15.15+xf37f22c
      - name: cilium-enterprise.v1.15.14-x0e512e3
        version: 1.15.14+x0e512e3
      - name: cilium-enterprise.v1.15.13-x134fa29
        version: 1.15.13+x134fa29
      name: "1.15"
    - currentCSV: cilium-enterprise.v1.16.15-x6419e26
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.16.0 <1.16.15+x6419e26'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/startup-script-ubi:v1.16.15-cee.1@sha256:5d59a46be44c4d899576e69e0f20e148245e17d905a0769a73fd01f2432d1bab
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:ec8a17956e75e91b60e2999d10e67ef76f9b4cfd754db415aaf387d21690da87
        - quay.io/isovalent/cilium-envoy-ubi:v1.16.15-cee.1@sha256:843e6fdf6845f3d5f841dbd2dce75224b16c9d34c2b0ce7f329d9666466bfc6b
        - quay.io/isovalent/cilium-ubi:v1.16.15-cee.1@sha256:1d7bbbcb613ef20cd1719bd77473e71c198960ee5d8e563aec9c65cfd3c2efe2
        - quay.io/isovalent/hubble-relay-ubi:v1.16.15-cee.1@sha256:c992d549627e8f9180c272e7d96ff3783de184796633ed5d0b6fe2cf7e74055d
        - quay.io/isovalent/operator-generic-ubi:v1.16.15-cee.1@sha256:970aa156b6f5c390b6835a626190e18656f02db6a5cf27aaa1ac9c7872f4b15a
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.16.15-cee.1@sha256:f7573ab4145d306051b5d6fa7ff1af395fc71935f5204839ab1c9f574aee17ad
        - quay.io/isovalent/certgen-ubi:v1.16.15-cee.1@sha256:e3f1fa5ee59887d3a15c43947a853c8a3f4fc650ed72bcd891b277037ea16a9d
        version: 1.16.15+x6419e26
      entries:
      - name: cilium-enterprise.v1.16.15-x6419e26
        version: 1.16.15+x6419e26
      - name: cilium-enterprise.v1.16.13-xb2de075
        version: 1.16.13+xb2de075
      - name: cilium-enterprise.v1.16.12-xd373871
        version: 1.16.12+xd373871
      - name: cilium-enterprise.v1.16.11-xbd1cd23
        version: 1.16.11+xbd1cd23
      - name: cilium-enterprise.v1.16.10-x89ab4b0
        version: 1.16.10+x89ab4b0
      - name: cilium-enterprise.v1.16.9-x8ec4137
        version: 1.16.9+x8ec4137
      - name: cilium-enterprise.v1.16.8-yd2b50b4
        version: 1.16.8+yd2b50b4
      - name: cilium-enterprise.v1.16.7-x4867421
        version: 1.16.7+x4867421
      - name: cilium-enterprise.v1.16.6-x7e6eb1c
        version: 1.16.6+x7e6eb1c
      name: "1.16"
    defaultChannel: "1.16"
    packageName: cilium-enterprise
    provider:
      name: Isovalent
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: heat-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: heat-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "heat.openstack.org/v1beta1",
                "kind": "Heat",
                "metadata": {
                  "name": "heat"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = True\n",
                  "databaseInstance": "openstack",
                  "heatAPI": {},
                  "heatCfnAPI": {},
                  "heatEngine": {},
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "heat.openstack.org/v1beta1",
                "kind": "HeatAPI",
                "metadata": {
                  "name": "heat-api"
                },
                "spec": {
                  "databaseAccount": "heat",
                  "databaseHostname": "openstack",
                  "passwordSelectors": {
                    "authEncryptionKey": "HeatAuthEncryptionKey",
                    "service": "HeatPassword"
                  },
                  "replicas": 1,
                  "resources": {},
                  "secret": "osp-secret",
                  "serviceAccount": "heat-heat",
                  "serviceUser": "heat",
                  "transportURLSecret": "rabbitmq-transport-url-heat-heat-transport"
                }
              },
              {
                "apiVersion": "heat.openstack.org/v1beta1",
                "kind": "HeatCfnAPI",
                "metadata": {
                  "name": "heat-cfnapi"
                },
                "spec": {
                  "databaseAccount": "heat",
                  "databaseHostname": "openstack",
                  "passwordSelectors": {
                    "authEncryptionKey": "HeatAuthEncryptionKey",
                    "service": "HeatPassword"
                  },
                  "replicas": 1,
                  "resources": {},
                  "secret": "osp-secret",
                  "serviceAccount": "heat-heat",
                  "serviceUser": "heat",
                  "transportURLSecret": "rabbitmq-transport-url-heat-heat-transport"
                }
              },
              {
                "apiVersion": "heat.openstack.org/v1beta1",
                "kind": "HeatEngine",
                "metadata": {
                  "name": "heat-engine"
                },
                "spec": {
                  "databaseAccount": "heat",
                  "databaseHostname": "openstack",
                  "passwordSelectors": {
                    "authEncryptionKey": "HeatAuthEncryptionKey",
                    "service": "HeatPassword"
                  },
                  "replicas": 1,
                  "resources": {},
                  "secret": "osp-secret",
                  "serviceAccount": "heat-heat",
                  "serviceUser": "heat",
                  "transportURLSecret": "rabbitmq-transport-url-heat-heat-transport"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:16Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HeatAPI is the Schema for the heatapis API
            displayName: Heat API
            kind: HeatAPI
            name: heatapis.heat.openstack.org
            version: v1beta1
          - description: HeatCfnAPI is the Schema for the heatcfnapis API
            displayName: Heat Cfn API
            kind: HeatCfnAPI
            name: heatcfnapis.heat.openstack.org
            version: v1beta1
          - description: HeatEngine is the Schema for the heatengines API
            displayName: Heat Engine
            kind: HeatEngine
            name: heatengines.heat.openstack.org
            version: v1beta1
          - description: Heat is the Schema for the heats API
            displayName: Heat
            kind: Heat
            name: heats.heat.openstack.org
            version: v1beta1
        description: Heat Operator
        displayName: Heat Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Orchestration
        - Heat
        links:
        - name: Heat Operator
          url: https://github.com/openstack-k8s-operators/heat-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/heat-rhel9-operator@sha256:fa61aeee8d63087d150a7641d1e8d4b67dd067bcfed6a140401fe396f1e32403
        - registry.redhat.io/rhoso/openstack-heat-api-rhel9@sha256:dba236e5fed89fbe46397b6ab01de5795e66e9194f5cd99fdf137c37a7b96135
        - registry.redhat.io/rhoso/openstack-heat-api-cfn-rhel9@sha256:95548184e68325a5fe1604d5c4d35d1e9cca7f39b84303fce50479e0e35f8280
        - registry.redhat.io/rhoso/openstack-heat-engine-rhel9@sha256:0b6fc782986bf11b26c0839fd81c52257f4a31aa5e5a6728ce251f3670537aba
        version: 1.0.6
      entries:
      - name: heat-operator.v1.0.6
        version: 1.0.6
      - name: heat-operator.v1.0.5
        version: 1.0.5
      - name: heat-operator.v1.0.4
        version: 1.0.4
      - name: heat-operator.v1.0.3
        version: 1.0.3
      - name: heat-operator.v1.0.2
        version: 1.0.2
      - name: heat-operator.v1.0.1
        version: 1.0.1
      - name: heat-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: heat-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: servicemeshoperator3
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: servicemeshoperator3.v3.0.0-tp.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh-tech-preview/istio-rhel9-operator@sha256:f5dfe495ba92ecb880485b66cfafbf6451f8142b8e0d5fd7f29db94171c51336
          createdAt: "2024-12-12T10:28:38Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1alpha1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1alpha1
          - description: IstioRevisionTag references a Istio or IstioRevision object
              and serves as an alias for sidecar injection.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1alpha1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-rhel9-operator@sha256:f5dfe495ba92ecb880485b66cfafbf6451f8142b8e0d5fd7f29db94171c51336
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-pilot-rhel9@sha256:376f64eab0ebeb06035983a60d1e7647c5d41354f31a49cb34f4deb7b6708e3e
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-pilot-rhel9@sha256:ea9372fabe06cd8e65ea7b4567ce3c8fbddd9d673dc34a12743fd95c548059d1
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-proxyv2-rhel9@sha256:c21481d6dae58b82dac93b1b47b9d9f53a30617fa9d930a0b5d075620cffa74b
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-must-gather-rhel9@sha256:2bf3704ba8d5945973ccf6bd13418457e929836352acf0ea6a47976ffa246616
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-proxyv2-rhel9@sha256:cbd6bc47341ab8618609f89d063552e17232541c727429855c01deab73e4d06c
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-must-gather-rhel9@sha256:e3978aecaafe2678355e21a7303f8620f53d8b774f6592f10527c15bae957224
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-cni-rhel9@sha256:d0ab36246c556249b6d67346eb0c72ebd22483a6502caa0fc75112a8d327b053
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-cni-rhel9@sha256:f16cd8a29ddcbb3de77cbfcf7e1da78e42cc0c1c502132d477a01c52689cdf23
        version: 3.0.0-tp.2
      entries:
      - name: servicemeshoperator3.v3.0.0-tp.2
        version: 3.0.0-tp.2
      - name: servicemeshoperator3.v3.0.0-tp.1
        version: 3.0.0-tp.1
      name: candidates
    - currentCSV: servicemeshoperator3.v3.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.4"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.4"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
          createdAt: "2025-09-10T06:22:42Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:366266f10e658c4cea86bddf6ee847e1908ea4dd780cb5f7fb0e466bac8995f1
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:0d6930075b8906bee0ab1aeac82f6e2431c16163cf87f7f0c68c10b2c40552a1
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:26747627ae22bbdffcf9de58077454fc0c890cda83659d7834b15dea2b5aaaf2
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2ea112ab90b8540f11e9949d77c3e7e3b3ef57ac3bf23f6cf1e883a88430e1f9
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:028e10651db0d1ddb769a27c9483c6d41be6ac597f253afd9d599f395d9c82d8
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:ecc6a22a471f4f4b836648f3a72ea7f57a4ea960ebcbc7dbe3fe729500902d0b
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d518f3d1539f45e1253c5c9fa22062802804601d4998cd50344e476a3cc388fe
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:33feac0c9c888b375e207e86b3f9a1a289126ca46b64632a7f693e2191cfbbfd
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:2ca96716812d339775d3dfd396c863834dd3810720e98bfcf87bc67c5fbd29b5
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:6bb44c84363a4113c5bce2ae0d6528b345fddd12bc559bbfaacf6222060ddcb0
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:6881fb9f5a4b447be4c468b8585fff86171bd3d82671a48559f3ceabcb9dad24
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:14c5a52faf20267baa43d9a72ee6416f56fbc41dd640adc2aba3c4043802a0e9
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:b757b81d34711cb8abfcb0e8896c023d39d4778f237dd80b637ba4bf613df893
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:f7ab868455b5e887e2d89b7078678531290365f9166cae99ce56119f31fa2ba4
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d757e700a124d7583b9456ec9e08e5e7e249c1de202bc4e973102922b9f69191
        version: 3.1.2
      entries:
      - name: servicemeshoperator3.v3.1.2
        version: 3.1.2
      - name: servicemeshoperator3.v3.1.1
        version: 3.1.1
      - name: servicemeshoperator3.v3.1.0
        version: 3.1.0
      - name: servicemeshoperator3.v3.0.5
        version: 3.0.5
      - name: servicemeshoperator3.v3.0.4
        version: 3.0.4
      - name: servicemeshoperator3.v3.0.3
        version: 3.0.3
      - name: servicemeshoperator3.v3.0.2
        version: 3.0.2
      - name: servicemeshoperator3.v3.0.1
        version: 3.0.1
      - name: servicemeshoperator3.v3.0.0
        version: 3.0.0
      name: stable
    - currentCSV: servicemeshoperator3.v3.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.6"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.6"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:d716f3558fbdf8514f225de708135ba87fdcac4c718ef101f394daf546514afb
          createdAt: "2025-09-03T07:53:34Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:d716f3558fbdf8514f225de708135ba87fdcac4c718ef101f394daf546514afb
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        version: 3.0.5
      entries:
      - name: servicemeshoperator3.v3.0.5
        version: 3.0.5
      - name: servicemeshoperator3.v3.0.4
        version: 3.0.4
      - name: servicemeshoperator3.v3.0.3
        version: 3.0.3
      - name: servicemeshoperator3.v3.0.2
        version: 3.0.2
      - name: servicemeshoperator3.v3.0.1
        version: 3.0.1
      - name: servicemeshoperator3.v3.0.0
        version: 3.0.0
      name: stable-3.0
    - currentCSV: servicemeshoperator3.v3.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.4"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.4"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
          createdAt: "2025-09-10T06:22:42Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:028e10651db0d1ddb769a27c9483c6d41be6ac597f253afd9d599f395d9c82d8
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d518f3d1539f45e1253c5c9fa22062802804601d4998cd50344e476a3cc388fe
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:366266f10e658c4cea86bddf6ee847e1908ea4dd780cb5f7fb0e466bac8995f1
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:33feac0c9c888b375e207e86b3f9a1a289126ca46b64632a7f693e2191cfbbfd
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:ecc6a22a471f4f4b836648f3a72ea7f57a4ea960ebcbc7dbe3fe729500902d0b
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:b757b81d34711cb8abfcb0e8896c023d39d4778f237dd80b637ba4bf613df893
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:14c5a52faf20267baa43d9a72ee6416f56fbc41dd640adc2aba3c4043802a0e9
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:6bb44c84363a4113c5bce2ae0d6528b345fddd12bc559bbfaacf6222060ddcb0
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:0d6930075b8906bee0ab1aeac82f6e2431c16163cf87f7f0c68c10b2c40552a1
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:6881fb9f5a4b447be4c468b8585fff86171bd3d82671a48559f3ceabcb9dad24
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d757e700a124d7583b9456ec9e08e5e7e249c1de202bc4e973102922b9f69191
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2ea112ab90b8540f11e9949d77c3e7e3b3ef57ac3bf23f6cf1e883a88430e1f9
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:26747627ae22bbdffcf9de58077454fc0c890cda83659d7834b15dea2b5aaaf2
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:f7ab868455b5e887e2d89b7078678531290365f9166cae99ce56119f31fa2ba4
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:2ca96716812d339775d3dfd396c863834dd3810720e98bfcf87bc67c5fbd29b5
        version: 3.1.2
      entries:
      - name: servicemeshoperator3.v3.1.2
        version: 3.1.2
      - name: servicemeshoperator3.v3.1.1
        version: 3.1.1
      - name: servicemeshoperator3.v3.1.0
        version: 3.1.0
      name: stable-3.1
    defaultChannel: stable
    packageName: servicemeshoperator3
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: must-gather-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: must-gather-operator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "MustGather",
                "metadata": {
                  "name": "example-mustgather"
                },
                "spec": {
                  "caseID": "02527285",
                  "caseManagementAccountSecretRef": {
                    "name": "case-management-creds"
                  },
                  "serviceAccountRef": {
                    "name": "must-gather-admin"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security, Monitoring
          certified: "false"
          containerImage: quay.io/redhat-cop/must-gather-operator@sha256:aebef122cdb8409f5a5d70d3cdc585e02272704d57e103553a84bc6fc238e646
          createdAt: "2021-12-02T17:13:53Z"
          description: This operator provides a facility to easily upload must-gather
            reports to a Red Hat case.
          operatorframework.io/suggested-namespace: global-load-balancer-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/must-gather-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: MustGather
            name: mustgathers.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The must gather operator helps collecting must gather information on a cluster and uploading it to a case.
          To use the operator a cluster administrator can create the following must gather CR:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: MustGather
          metadata:
            name: example-mustgather
          spec:
            caseID: '02527285'
            caseManagementAccountSecretRef:
              name: case-management-creds
          ```

          this request will collect the standard must gather info and upload it to case `#02527285` using the credentials found in the `caseManagementCreds` secret.

          A more complex example:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: MustGather
          metadata:
            name: full-mustgather
          spec:
            caseID: '02527285'
            caseManagementAccountSecretRef:
              name: case-management-creds
            serviceAccountRef:
              name: must-gather-admin
            mustGatherImages:
            - quay.io/kubevirt/must-gather:latest
            - quay.io/ocs-dev/ocs-must-gather
          ```

          in this example we are using a specific service account (which must have cluster admin permissions as per must-gather requirements) and we are specifying a couple of additional must gather images to be run for the `kubevirt` and `ocs` subsystem. If not specified serviceAccountRef.Name will default to `default`. Also the standard must gather image: `quay.io/openshift/origin-must-gather:latest` is always added by default.
        displayName: Must Gather Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - case_management
        - support
        - must_gather
        links:
        - name: Must Gather Operator
          url: https://github.com/redhat-cop/must-gather-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/must-gather-operator
        maintainers:
        - email: rspazzol@redhat.com'
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/must-gather-operator@sha256:aebef122cdb8409f5a5d70d3cdc585e02272704d57e103553a84bc6fc238e646
        version: 1.1.2
      entries:
      - name: must-gather-operator.v1.1.2
        version: 1.1.2
      - name: must-gather-operator.v1.1.1
        version: 1.1.1
      - name: must-gather-operator.v1.1.0
        version: 1.1.0
      - name: must-gather-operator.v1.0.6
        version: 1.0.6
      - name: must-gather-operator.v1.0.5
        version: 1.0.5
      - name: must-gather-operator.v1.0.4
        version: 1.0.4
      - name: must-gather-operator.v1.0.3
        version: 1.0.3
      - name: must-gather-operator.v1.0.2
        version: 1.0.2
      - name: must-gather-operator.v1.0.1
        version: 1.0.1
      - name: must-gather-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: must-gather-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CrowdStrike
      provider-url: https://crowdStrike.com
    name: falcon-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: falcon-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconAdmission",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-admission",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconadmission",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-kac"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "admission_controller"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconContainer",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-sidecar-sensor",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconcontainer",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-container-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "sidecar"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconDeployment",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcondeployment-sample",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falcon-deployment",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-deployment"
                },
                "spec": {
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconImageAnalyzer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "falcon-operator",
                    "app.kubernetes.io/instance": "falconimageanalyzer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "falconimageanalyzer",
                    "app.kubernetes.io/part-of": "falcon-operator"
                  },
                  "name": "falcon-image-analyzer"
                },
                "spec": {
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconNodeSensor",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-node-sensor",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconnodesensor",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-node-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "daemonset"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security,Monitoring
          containerImage: registry.connect.redhat.com/crowdstrike/falcon-operator@sha256:ffe00e9e22e209d1e08f29c3d257bbd51937fcbde68a1ed93b542a26dd436dee
          createdAt: "2025-06-03T22:34:20Z"
          description: Falcon Operator installs CrowdStrike Falcon Sensors on the
            cluster
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: falcon-operator
          operators.openshift.io/valid-subscription: Not required to use this operator
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: registry.connect.redhat.com/crowdstrike/falcon-operator
          support: CrowdStrike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FalconAdmission is the Schema for the falconadmissions API
            displayName: Falcon Admission
            kind: FalconAdmission
            name: falconadmissions.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconContainer is the Schema for the falconcontainers API
            displayName: Falcon Container
            kind: FalconContainer
            name: falconcontainers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconDeployment is the Schema for the falcondeployments
              API
            displayName: Falcon Deployment
            kind: FalconDeployment
            name: falcondeployments.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconImageAnalyzer is the Schema for the falconImageAnalyzers
              API
            displayName: Falcon Image Analyzer
            kind: FalconImageAnalyzer
            name: falconimageanalyzers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconNodeSensor is the Schema for the falconnodesensors
              API
            displayName: Falcon Node Sensor
            kind: FalconNodeSensor
            name: falconnodesensors.falcon.crowdstrike.com
            version: v1alpha1
        description: |-
          The CrowdStrike Falcon Operator installs CrowdStrike Falcon custom resources on a Kubernetes cluster.

          ## About the CrowdStrike Falcon Operator
          The CrowdStrike Falcon Operator deploys CrowdStrike Falcon to the cluster. The operator exposes custom resources that allow you to protect your Kubernetes clusters when deployed.

          ## About Custom Resources

          | Custom Resource                                       | Description                                                      |
          | :--                                             | :--                                                    |
          | [FalconAdmission](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/admission/README.md) | Manages installation of Falcon Admission Controller on the cluster |
          | [FalconImageAnalyzer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/imageanalyzer/README.md) | Manages installation of Falcon Image Assessment at Runtime on the cluster |
          | [FalconContainer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/container/README.md) | Manages installation of Falcon Container Sensor on the cluster   |
          | [FalconNodeSensor](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/node/README.md)     | Manages installation of Falcon Linux Sensor on the cluster nodes |
          | [FalconDeployment](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/falcondeployment/README.md)       | Deploys FalconAdmission, FalconImageAnalyzer, FalconContainer, and FalconNodeSensor CRs from a single manifest |

          ## Installation and Deployment

          For installation and deployment of the CrowdStrike Falcon Operator and its Custom Resources, please read the [Installation and Deployment Guide](./docs/install_guide.md) and choose the deployment method that is right for your target environment.

          ## Getting Help
          If you encounter any issues while using the Falcon Operator, you can create an issue on our [Github repo](https://github.com/CrowdStrike/falcon-operator) for bugs, enhancements, or other requests.

          ## Contributing
          You can contribute by:

          * Raising any issues you find using Falcon Operator
          * Fixing issues by opening [Pull Requests](https://github.com/CrowdStrike/falcon-operator/pulls)
          * Improving documentation
          * Talking about the CrowdStrike Falcon Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/CrowdStrike/falcon-operator/issues).

          ## Additional Resources
           - CrowdStrike Container Security: [Product Page](https://www.crowdstrike.com/products/cloud-security/falcon-cloud-workload-protection/container-security/)
           - So You Think Your Containers Are Secure? Four Steps to Ensure a Secure Container Deployment: [Blog Post](https://www.crowdstrike.com/blog/four-steps-to-ensure-a-secure-containter-deployment/)
           - Container Security With CrowdStrike: [Blog Post](https://www.crowdstrike.com/blog/tech-center/container-security/)
           - To learn more about Falcon Container Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/support/documentation/146/falcon-container-sensor-for-linux), [Release Notes](https://falcon.crowdstrike.com/support/news/release-notes-falcon-container-sensor-for-linux)
           - To learn more about Falcon Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/documentation/20/falcon-sensor-for-linux#kubernetes-support)
           - [Developer Documentation](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/developer_guide.md)
        displayName: CrowdStrike Falcon Platform - Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - runtime-protection
        links:
        - name: Falcon Operator
          url: https://github.com/CrowdStrike/falcon-operator
        maintainers:
        - email: integrations@crowdstrike.com
          name: CrowdStrike Solutions Architecture
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: CrowdStrike
          url: https://crowdStrike.com
        relatedImages:
        - registry.connect.redhat.com/crowdstrike/falcon-operator@sha256:ffe00e9e22e209d1e08f29c3d257bbd51937fcbde68a1ed93b542a26dd436dee
        - registry.crowdstrike.com/falcon-container/us-1/release/falcon-sensor@sha256:92d222cd604c9f2c9e9f2c9b402b9fb0858fcb04dac38c15e25125978f43c99b
        - registry.crowdstrike.com/falcon-kac/us-1/release/falcon-kac@sha256:5049128dcfe9565f39fc6e28fb94f5bf05fb7c51cf787fc6fe5c63a2b04f800d
        - registry.crowdstrike.com/falcon-imageanalyzer/us-1/release/falcon-imageanalyzer@sha256:f08603648658041ad52ccf0cfd058da40b385904830030aa82dc50609aed7ceb
        - registry.crowdstrike.com/falcon-sensor/us-1/release/falcon-sensor@sha256:ed4f3fe8d9f5f491600f18d4440dc1d8e3a07d4b9c08a1f17382a23a3a5cf523
        version: 1.8.0
      entries:
      - name: falcon-operator.v1.8.0
        version: 1.8.0
      - name: falcon-operator.v1.7.0
        version: 1.7.0
      - name: falcon-operator.v1.6.0
        version: 1.6.0
      - name: falcon-operator.v1.5.0
        version: 1.5.0
      - name: falcon-operator.v1.4.0
        version: 1.4.0
      - name: falcon-operator.v1.3.1
        version: 1.3.1
      - name: falcon-operator.v1.3.0
        version: 1.3.0
      - name: falcon-operator.v1.2.0
        version: 1.2.0
      - name: falcon-operator.v1.1.0
        version: 1.1.0
      - name: falcon-operator.v1.0.1
        version: 1.0.1
      - name: falcon-operator.v1.0.0
        version: 1.0.0
      name: certified-1.0
    defaultChannel: certified-1.0
    packageName: falcon-operator
    provider:
      name: CrowdStrike
      url: https://crowdStrike.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Airlock
      provider-url: ""
    name: airlock-microgateway
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airlock-microgateway.v4.5.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControl","metadata":{"name":"access-control-example"},"spec":{"policies":[{"requestConditions":{"path":{"matcher":{"exact":"/secure"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}},{"oidc":{"claim":{"name":"roles","value":{"matcher":{"contains":"\"admin\""}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"oidc-example"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"identity-propagation-example"}}],"onFailure":"Pass"}},{"authorization":{}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurity","metadata":{"name":"content-security-example"},"spec":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1"},"spec":{"value":{"name":"ext-authz","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}}}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"lifetime":"16h"},"prefix":"session_prefix_example"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SidecarGateway","metadata":{"name":"sidecar-gateway-example"},"spec":{"podSelector":{"matchLabels":{"app":"example-app"}},"sessionHandlingRef":{"name":"session-handling-example"},"applications":[{"containerPort":8443,"routes":[{"pathPrefix":"/metrics","unsecured":{}},{"pathPrefix":"/","secured":{"accessControlRef":{"name":"access-control-example"},"contentSecurityRef":{"name":"content-security-example"}}}],"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}},"restrictions":{"http":{"headersLength":"80Ki"}},"protocol":{"auto":{}},"tls":{"enable":true,"secretRef":{"name":"example-tls-secret"},"protocol":{"minimum":"TLSv1_3"},"clientCertificate":{"ignored":{}},"xfcc":"AlwaysForwardOnly"}},"upstream":{"tls":{"enable":true,"protocol":{"minimum":"TLSv1_2"}}},"envoyHTTPFilterRefs":{"prepend":[{"name":"envoy-http-filter-example-1"}]},"telemetryRef":{"name":"telemetry-example"}}],"envoyClusterRefs":[{"name":"envoy-cluster-example-1"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-06-13T17:33:31Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: AccessControl
            name: accesscontrols.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurity
            name: contentsecurities.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: SidecarGateway
            name: sidecargateways.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056)
          * Gateway API
          * (Optional) Airlock Microgateway CNI (Required for [data plane mode sidecar](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000137)).

          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ### (Optional) Install Airlock Microgateway CNI
          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          Refer to [our documentation](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000140) for installation instructions

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD_MONITOR\_CREATE | "false" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.5/
        - name: Airlock Microgateway Labs
          url: https://play.instruqt.com/airlock/invite/hyi9fy4b4jzc?icp_referrer=openshift
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:806366a0b3f78844fe14775fc290fbfbde8d2a701d34a6478b019a0e7140b699
        version: 4.5.4
      entries:
      - name: airlock-microgateway.v4.5.4
        version: 4.5.4
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.5"
    - currentCSV: airlock-microgateway.v4.6.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControl","metadata":{"name":"access-control-example"},"spec":{"policies":[{"requestConditions":{"path":{"matcher":{"exact":"/secure"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}},{"oidc":{"claim":{"name":"roles","value":{"matcher":{"contains":"\"admin\""}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"oidc-example"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"token-exchange-example"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"identity-propagation-example"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"jwt-example"}}}}},{"authorization":{}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurity","metadata":{"name":"content-security-example"},"spec":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1"},"spec":{"value":{"name":"ext-authz","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}}}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"lifetime":"16h"},"prefix":"session_prefix_example"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SidecarGateway","metadata":{"name":"sidecar-gateway-example"},"spec":{"podSelector":{"matchLabels":{"app":"example-app"}},"sessionHandlingRef":{"name":"session-handling-example"},"applications":[{"containerPort":8443,"routes":[{"pathPrefix":"/metrics","unsecured":{}},{"pathPrefix":"/","secured":{"accessControlRef":{"name":"access-control-example"},"contentSecurityRef":{"name":"content-security-example"}}}],"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}},"restrictions":{"http":{"headersLength":"80Ki"}},"protocol":{"auto":{}},"tls":{"enable":true,"secretRef":{"name":"example-tls-secret"},"protocol":{"minimum":"TLSv1_3"},"clientCertificate":{"ignored":{}},"xfcc":"AlwaysForwardOnly"}},"upstream":{"tls":{"enable":true,"protocol":{"minimum":"TLSv1_2"}}},"envoyHTTPFilterRefs":{"prepend":[{"name":"envoy-http-filter-example-1"}]},"telemetryRef":{"name":"telemetry-example"}}],"envoyClusterRefs":[{"name":"envoy-cluster-example-1"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-23T06:27:21Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: AccessControl
            name: accesscontrols.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurity
            name: contentsecurities.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: SidecarGateway
            name: sidecargateways.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056)
          * Gateway API
          * (Optional) Airlock Microgateway CNI (Required for [data plane mode sidecar](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000137)).

          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.3.0/standard-install.yaml
          ```

          ### (Optional) Install Airlock Microgateway CNI
          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          Refer to [our documentation](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000140) for installation instructions

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD_MONITOR\_CREATE | "false" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.6/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:8fa300ced5b69acbad824c037d2b344fd1926bcfb39e264e9f0f055ceb9bd68d
        version: 4.6.4
      entries:
      - name: airlock-microgateway.v4.6.4
        version: 4.6.4
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.6"
    - currentCSV: airlock-microgateway.v4.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1","namespace":"default"},"spec":{"value":{"name":"envoycluster/default/envoy-cluster-example-1","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyExtensionPolicy","metadata":{"name":"envoy-extension-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"envoyHTTPFilterRefs":{"prepend":[{"name":"geoip-blocker"}],"append":[{"name":"lua-transform-request"}]},"envoyClusterRefs":[{"name":"geoip-server"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"volumes":[{"name":"vol-1","emptyDir":{}}],"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}},"volumeMounts":[{"name":"vol-1","mountPath":"/tmp/data"}]}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"telemetryRef":{"name":"telemetry-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"mode":"Enforce","persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"idle":"30s","lifetime":"16h"},"prefix":"session_prefix_example","cookie":{"name":"SESS","attributes":{"domain":"example.org","path":"/","sameSite":"Strict","secure":true}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"},"session":{"id":"%SESSION_ID%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-03T13:02:23Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyExtensionPolicy
            name: envoyextensionpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ***Note:*** Sidecar-based data plane mode is *not* supported by this OLM Bundle.
          If you want to use the Sidecar-based data plane mode, use Helm instead. Please note that the Airlock Microgateway CNI and the Helm charts are not certified for RedHat OpenShift.
          Refer to https://github.com/airlock/microgateway for installation instruction with Helm.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056)
          * [Gateway API](https://gateway-api.sigs.k8s.io/)

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ***Note:*** *This step is only required for OpenShift versions below 4.19, as Gateway API is included by default in OpenShift 4.19+.*

          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD\_MONITOR\_CREATE | "true" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gateway API
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.7/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:d4ce35184922953c209a9ebb234ca3370dbfc5054de411b5503c26ac2730b7ae
        version: 4.7.1
      entries:
      - name: airlock-microgateway.v4.7.1
        version: 4.7.1
      - name: airlock-microgateway.v4.7.0
        version: 4.7.0
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.7"
    - currentCSV: airlock-microgateway.v4.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1","namespace":"default"},"spec":{"value":{"name":"envoycluster/default/envoy-cluster-example-1","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyExtensionPolicy","metadata":{"name":"envoy-extension-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"envoyHTTPFilterRefs":{"prepend":[{"name":"geoip-blocker"}],"append":[{"name":"lua-transform-request"}]},"envoyClusterRefs":[{"name":"geoip-server"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"volumes":[{"name":"vol-1","emptyDir":{}}],"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}},"volumeMounts":[{"name":"vol-1","mountPath":"/tmp/data"}]}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"telemetryRef":{"name":"telemetry-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"mode":"Enforce","persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"idle":"30s","lifetime":"16h"},"prefix":"session_prefix_example","cookie":{"name":"SESS","attributes":{"domain":"example.org","path":"/","sameSite":"Strict","secure":true}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"},"session":{"id":"%SESSION_ID%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-03T13:02:23Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyExtensionPolicy
            name: envoyextensionpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ***Note:*** Sidecar-based data plane mode is *not* supported by this OLM Bundle.
          If you want to use the Sidecar-based data plane mode, use Helm instead. Please note that the Airlock Microgateway CNI and the Helm charts are not certified for RedHat OpenShift.
          Refer to https://github.com/airlock/microgateway for installation instruction with Helm.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056)
          * [Gateway API](https://gateway-api.sigs.k8s.io/)

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ***Note:*** *This step is only required for OpenShift versions below 4.19, as Gateway API is included by default in OpenShift 4.19+.*

          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD\_MONITOR\_CREATE | "true" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gateway API
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.7/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:d4ce35184922953c209a9ebb234ca3370dbfc5054de411b5503c26ac2730b7ae
        version: 4.7.1
      entries:
      - name: airlock-microgateway.v4.7.1
        version: 4.7.1
      - name: airlock-microgateway.v4.7.0
        version: 4.7.0
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: stable
    defaultChannel: "4.7"
    packageName: airlock-microgateway
    provider:
      name: Airlock
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: marin3r
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marin3r.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyConfig",
                "metadata": {
                  "name": "envoyconfig-example",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "nodeID": "example",
                  "resources": [
                    {
                      "type": "cluster",
                      "value": {
                        "connect_timeout": "0.01s",
                        "dns_lookup_family": "V4_ONLY",
                        "lb_policy": "ROUND_ROBIN",
                        "load_assignment": {
                          "cluster_name": "example",
                          "endpoints": [
                            {
                              "lb_endpoints": [
                                {
                                  "endpoint": {
                                    "address": {
                                      "socket_address": {
                                        "address": "example",
                                        "port_value": 8080
                                      }
                                    }
                                  }
                                }
                              ]
                            }
                          ]
                        },
                        "name": "example",
                        "type": "STRICT_DNS"
                      }
                    },
                    {
                      "type": "route",
                      "value": {
                        "name": "local",
                        "virtual_hosts": [
                          {
                            "domains": [
                              "*"
                            ],
                            "name": "all",
                            "routes": [
                              {
                                "match": {
                                  "prefix": "/"
                                },
                                "route": {
                                  "cluster": "example"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    {
                      "type": "listener",
                      "value": {
                        "address": {
                          "socket_address": {
                            "address": "0.0.0.0",
                            "port_value": 8443
                          }
                        },
                        "filter_chains": [
                          {
                            "filters": [
                              {
                                "name": "envoy.filters.network.http_connection_manager",
                                "typed_config": {
                                  "@type": "type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager",
                                  "access_log": [
                                    {
                                      "name": "envoy.access_loggers.file",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.access_loggers.file.v3.FileAccessLog",
                                        "path": "/dev/stdout"
                                      }
                                    }
                                  ],
                                  "http_filters": [
                                    {
                                      "name": "envoy.filters.http.router",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.filters.http.router.v3.Router"
                                      }
                                    }
                                  ],
                                  "rds": {
                                    "config_source": {
                                      "ads": {},
                                      "resource_api_version": "V3"
                                    },
                                    "route_config_name": "local"
                                  },
                                  "stat_prefix": "https"
                                }
                              }
                            ],
                            "transport_socket": {
                              "name": "envoy.transport_sockets.tls",
                              "typed_config": {
                                "@type": "type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext",
                                "common_tls_context": {
                                  "tls_certificate_sds_secret_configs": [
                                    {
                                      "name": "example.default.svc",
                                      "sds_config": {
                                        "ads": {},
                                        "resource_api_version": "V3"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          }
                        ],
                        "name": "https"
                      }
                    },
                    {
                      "blueprint": "tlsCertificate",
                      "generateFromTlsSecret": "example.default.svc",
                      "type": "secret"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "DiscoveryService",
                "metadata": {
                  "name": "discoveryservice-example",
                  "namespace": "my-namespace"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyDeployment",
                "metadata": {
                  "name": "envoydeployment-sample",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "discoveryServiceRef": "discoveryservice-example",
                  "envoyConfigRef": "envoyconfig-example",
                  "ports": [
                    {
                      "name": "https",
                      "port": 8443
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: quay.io/3scale/marin3r
          createdAt: "2024-08-05T16:08:25Z"
          description: Lighweight, CRD based Envoy control plane for Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["envoyconfigrevisions.marin3r.3scale.net","discoveryservicecertificates.operator.marin3r.3scale.net"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/3scale-ops/marin3r
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DiscoveryServiceCertificate is an internal resource used
              to create certificates. This resource is used by the DiscoveryService
              controller to create the required certificates for the different components.
              Direct use of DiscoveryServiceCertificate objects is discouraged.
            displayName: DiscoveryServiceCertificate
            kind: DiscoveryServiceCertificate
            name: discoveryservicecertificates.operator.marin3r.3scale.net
            version: v1alpha1
          - description: DiscoveryService represents an envoy discovery service server.
              Only one instance per namespace is currently supported.
            displayName: DiscoveryService
            kind: DiscoveryService
            name: discoveryservices.operator.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfigRevision is an internal resource that stores a
              specific version of an EnvoyConfig resource. EnvoyConfigRevisions are
              automatically created and deleted by the EnvoyConfig controller and
              are not intended to be directly used. Use EnvoyConfig objects instead.
            displayName: EnvoyConfigRevision
            kind: EnvoyConfigRevision
            name: envoyconfigrevisions.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfig holds the configuration for a given envoy nodeID.
              The spec of an EnvoyConfig object holds the Envoy resources that conform
              the desired configuration for the given nodeID and that the discovery
              service will send to any envoy client that identifies itself with that
              nodeID.
            displayName: EnvoyConfig
            kind: EnvoyConfig
            name: envoyconfigs.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyDeployment is a resource to deploy and manage a Kubernetes
              Deployment of Envoy Pods.
            displayName: EnvoyDeployment
            kind: EnvoyDeployment
            name: envoydeployments.operator.marin3r.3scale.net
            version: v1alpha1
        description: |
          MARIN3R implements a control plane to deploy, configure and operate a fleet of envoy instances within a Kubernetes cluster.
          It operates a group of gateways and Pod sidecar containers to provide general purpose functionality required to operate production systems.

          Features:

          * Deploy and manage an Envoy xDS server using the DiscoveryService custom resource.
          * Inject Envoy sidecar containers based on Pod annotations.
          * Deploy Envoy as a Kubernetes Deployment using the EnvoyDeployment custom resource.
          * Dynamic Envoy configuration and re-configuration using the EnvoyConfig custom resource.
          * Use Kubernetes Secrets as certificate sources.
          * Syntactic validation of Envoy configurations.
          * Lifecycle management of Envoy containers (graceful shutdown and connection draining).
          * Self-healing.

          Check the [docs](https://github.com/3scale-ops/marin3r#readme) for more information.

          ## License
          MARIN3R is licensed under the [Apache 2.0 license](https://github.com/3scale/prometheus-exporter-operator/blob/master/LICENSE)
        displayName: MARIN3R
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - envoy
        - discovery
        - controlplane
        - proxy
        - sidecar
        - xds
        - network
        - gateway
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/marin3r
        maintainers:
        - email: 3scale-operations+marin3r@redhat.com
          name: 3scale Operations Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/marin3r:v0.13.1
        version: 0.13.1
      entries:
      - name: marin3r.v0.13.1
        version: 0.13.1
      - name: marin3r.v0.12.3
        version: 0.12.3
      - name: marin3r.v0.11.1
        version: 0.11.1
      - name: marin3r.v0.10.0
        version: 0.10.0
      name: alpha
    - currentCSV: marin3r.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyConfig",
                "metadata": {
                  "name": "envoyconfig-example",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "nodeID": "example",
                  "resources": [
                    {
                      "type": "cluster",
                      "value": {
                        "connect_timeout": "0.01s",
                        "dns_lookup_family": "V4_ONLY",
                        "lb_policy": "ROUND_ROBIN",
                        "load_assignment": {
                          "cluster_name": "example",
                          "endpoints": [
                            {
                              "lb_endpoints": [
                                {
                                  "endpoint": {
                                    "address": {
                                      "socket_address": {
                                        "address": "example",
                                        "port_value": 8080
                                      }
                                    }
                                  }
                                }
                              ]
                            }
                          ]
                        },
                        "name": "example",
                        "type": "STRICT_DNS"
                      }
                    },
                    {
                      "type": "route",
                      "value": {
                        "name": "local",
                        "virtual_hosts": [
                          {
                            "domains": [
                              "*"
                            ],
                            "name": "all",
                            "routes": [
                              {
                                "match": {
                                  "prefix": "/"
                                },
                                "route": {
                                  "cluster": "example"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    {
                      "type": "listener",
                      "value": {
                        "address": {
                          "socket_address": {
                            "address": "0.0.0.0",
                            "port_value": 8443
                          }
                        },
                        "filter_chains": [
                          {
                            "filters": [
                              {
                                "name": "envoy.filters.network.http_connection_manager",
                                "typed_config": {
                                  "@type": "type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager",
                                  "access_log": [
                                    {
                                      "name": "envoy.access_loggers.file",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.access_loggers.file.v3.FileAccessLog",
                                        "path": "/dev/stdout"
                                      }
                                    }
                                  ],
                                  "http_filters": [
                                    {
                                      "name": "envoy.filters.http.router",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.filters.http.router.v3.Router"
                                      }
                                    }
                                  ],
                                  "rds": {
                                    "config_source": {
                                      "ads": {},
                                      "resource_api_version": "V3"
                                    },
                                    "route_config_name": "local"
                                  },
                                  "stat_prefix": "https"
                                }
                              }
                            ],
                            "transport_socket": {
                              "name": "envoy.transport_sockets.tls",
                              "typed_config": {
                                "@type": "type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext",
                                "common_tls_context": {
                                  "tls_certificate_sds_secret_configs": [
                                    {
                                      "name": "example.default.svc",
                                      "sds_config": {
                                        "ads": {},
                                        "resource_api_version": "V3"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          }
                        ],
                        "name": "https"
                      }
                    },
                    {
                      "blueprint": "tlsCertificate",
                      "generateFromTlsSecret": "example.default.svc",
                      "type": "secret"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "DiscoveryService",
                "metadata": {
                  "name": "discoveryservice-example",
                  "namespace": "my-namespace"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyDeployment",
                "metadata": {
                  "name": "envoydeployment-sample",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "discoveryServiceRef": "discoveryservice-example",
                  "envoyConfigRef": "envoyconfig-example",
                  "ports": [
                    {
                      "name": "https",
                      "port": 8443
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: quay.io/3scale/marin3r
          createdAt: "2024-08-05T16:08:25Z"
          description: Lighweight, CRD based Envoy control plane for Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["envoyconfigrevisions.marin3r.3scale.net","discoveryservicecertificates.operator.marin3r.3scale.net"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/3scale-ops/marin3r
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DiscoveryServiceCertificate is an internal resource used
              to create certificates. This resource is used by the DiscoveryService
              controller to create the required certificates for the different components.
              Direct use of DiscoveryServiceCertificate objects is discouraged.
            displayName: DiscoveryServiceCertificate
            kind: DiscoveryServiceCertificate
            name: discoveryservicecertificates.operator.marin3r.3scale.net
            version: v1alpha1
          - description: DiscoveryService represents an envoy discovery service server.
              Only one instance per namespace is currently supported.
            displayName: DiscoveryService
            kind: DiscoveryService
            name: discoveryservices.operator.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfigRevision is an internal resource that stores a
              specific version of an EnvoyConfig resource. EnvoyConfigRevisions are
              automatically created and deleted by the EnvoyConfig controller and
              are not intended to be directly used. Use EnvoyConfig objects instead.
            displayName: EnvoyConfigRevision
            kind: EnvoyConfigRevision
            name: envoyconfigrevisions.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfig holds the configuration for a given envoy nodeID.
              The spec of an EnvoyConfig object holds the Envoy resources that conform
              the desired configuration for the given nodeID and that the discovery
              service will send to any envoy client that identifies itself with that
              nodeID.
            displayName: EnvoyConfig
            kind: EnvoyConfig
            name: envoyconfigs.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyDeployment is a resource to deploy and manage a Kubernetes
              Deployment of Envoy Pods.
            displayName: EnvoyDeployment
            kind: EnvoyDeployment
            name: envoydeployments.operator.marin3r.3scale.net
            version: v1alpha1
        description: |
          MARIN3R implements a control plane to deploy, configure and operate a fleet of envoy instances within a Kubernetes cluster.
          It operates a group of gateways and Pod sidecar containers to provide general purpose functionality required to operate production systems.

          Features:

          * Deploy and manage an Envoy xDS server using the DiscoveryService custom resource.
          * Inject Envoy sidecar containers based on Pod annotations.
          * Deploy Envoy as a Kubernetes Deployment using the EnvoyDeployment custom resource.
          * Dynamic Envoy configuration and re-configuration using the EnvoyConfig custom resource.
          * Use Kubernetes Secrets as certificate sources.
          * Syntactic validation of Envoy configurations.
          * Lifecycle management of Envoy containers (graceful shutdown and connection draining).
          * Self-healing.

          Check the [docs](https://github.com/3scale-ops/marin3r#readme) for more information.

          ## License
          MARIN3R is licensed under the [Apache 2.0 license](https://github.com/3scale/prometheus-exporter-operator/blob/master/LICENSE)
        displayName: MARIN3R
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - envoy
        - discovery
        - controlplane
        - proxy
        - sidecar
        - xds
        - network
        - gateway
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/marin3r
        maintainers:
        - email: 3scale-operations+marin3r@redhat.com
          name: 3scale Operations Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/marin3r:v0.13.1
        version: 0.13.1
      entries:
      - name: marin3r.v0.13.1
        version: 0.13.1
      - name: marin3r.v0.12.3
        version: 0.12.3
      - name: marin3r.v0.11.1
        version: 0.11.1
      - name: marin3r.v0.10.0
        version: 0.10.0
      name: stable
    defaultChannel: stable
    packageName: marin3r
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: machine-deletion-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: machine-deletion-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/machine-deletion-remediation-operator:v0.1.0
          createdAt: "2023-05-11 10:46:12"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/machine-deletion-remediation-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: machine-deletion-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: machine-deletion-remediation.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/machine-deletion-remediation-operator:v0.3.0
          createdAt: "2024-03-05 18:45:16"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/medik8s/machine-deletion-remediation-operator:v0.3.0
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        version: 0.3.0
      entries:
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      - name: machine-deletion-remediation.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: machine-deletion-remediation
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-nimbusmosaic-dusim
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-nimbusmosaic-dusim.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Nimbusmosaicdusim",
                "metadata": {
                  "name": "nimbusmosaicdusim-sample"
                },
                "spec": {
                  "development": false,
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "microservices": {
                    "jvm": "-Xmx300M -Xms64M -Xss256k"
                  },
                  "service": {
                    "type": "ClusterIP"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:41:56Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nimbusmosaicdusim
            name: nimbusmosaicdusims.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Nimbus Mosaic DUSim Operator
        displayName: Keysight Nimbus Mosaic DUSim Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Nimbusmosaic Dusim
          url: https://keysight-nimbusmosaic-dusim.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-nimbusmosaic-dusim@sha256:655dd6595a9c8e82947bed24991ac52f8fa55be99ca410e7d9f0026c51ef4520
        version: 0.1.21
      entries:
      - name: keysight-nimbusmosaic-dusim.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-nimbusmosaic-dusim
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-namespace-scope-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-namespace-scope-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1",
                "kind": "NamespaceScope",
                "metadata": {
                  "name": "namespacescope"
                },
                "spec": {
                  "namespaceMembers": [
                    "ibm-common-services",
                    "default"
                  ],
                  "restartLabels": {
                    "intent": "projected"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
          createdAt: "2025-02-19T00:54:24Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.2.14
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/IBM/ibm-namespace-scope-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NamespaceScope is the Schema for the namespacescopes API
            displayName: Namespace Scope
            kind: NamespaceScope
            name: namespacescopes.operator.ibm.com
            version: v1
        description: This operator automates the extension of operator watch and service
          account permission scope to other namespaces in an openshift cluster.
        displayName: IBM NamespaceScope Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Namespace Scope Operator
          url: https://github.com/IBM/ibm-namespace-scope-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:0774b1d916ebc354050612474800dda3a501c66a7a5aab6d18698c8c4e81ac4e
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-bundle@sha256:b6cb9e5976b749dd7818ac19d1cfd5e24673f3c052a41d6e2fe1d2d7094694b4
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:7a2cd72736bf9d30341763fd1652ff4b9ce3c785da5123645ee7664ee38c7283
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-catalog@sha256:c016904e8c5cfcf8270241ad5e28b52f528859984ed894cdd5e1283fa90cd6e8
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator-bundle@sha256:c743dd7de2edbd08c6c73e20cf15f3b147491ad77a81677339e88caffb93e84b
        version: 4.2.14
      entries:
      - name: ibm-namespace-scope-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-namespace-scope-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hitachi
      provider-url: ""
    name: hnas-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hnas-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.csi.hitachi.com/v1",
                "kind": "HNAS",
                "metadata": {
                  "name": "hnas",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-operator@sha256:9d3f9f10fc7172ee9cebbd9ee318906756d7e96909f2e66e0534e98bb53e0a7b
          createdAt: "2025-07-30T15:12:23Z"
          description: An operator for managing Hitachi NAS CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HNAS is the Schema for the hnas API
            displayName: HNAS
            kind: HNAS
            name: hnas.charts.csi.hitachi.com
            version: v1
        description: |-
          The Hitachi NAS CSI driver operator deploys the Hitachi NAS CSI driver on Kubernetes and OpenShift environments.  The driver can be used to create and manage persistent storage hosted on Hitachi NAS products, including those in the Hitachi VSP One File 3x range and the HNAS range of products.

          For the complete documentation, go to our [Product Documentation site](https://docs.hitachivantara.com/home) and refer to the Hitachi NAS CSI Driver User Guide - MK-92HNAS091.  The guide covers prerequisites for the NAS system and covers the driver functionality and all configuration options available for the driver.
        displayName: Hitachi NAS CSI Driver Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hitachi
        - Storage
        - Container Storage Interface
        - CSI
        - NAS
        links:
        - name: Product Documentation
          url: https://docs.hitachivantara.com/home
        maturity: stable
        minKubeVersion: 1.28.0
        provider:
          name: Hitachi
        relatedImages:
        - registry.k8s.io/sig-storage/csi-resizer@sha256:ab774734705a906561e15b67f6a96538f3ad520335d636f793aaafe87a3b5200
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:dd788d79cf4c1b8edee6d9b80b8a1ebfc51a38a365c5be656986b129be9ac784
        - registry.k8s.io/sig-storage/livenessprobe@sha256:2c5f9dc4ea5ac5509d93c664ae7982d4ecdec40ca7b0638c24e5b16243b8360f
        - registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-operator@sha256:9d3f9f10fc7172ee9cebbd9ee318906756d7e96909f2e66e0534e98bb53e0a7b
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:d7138bcc3aa5f267403d45ad4292c95397e421ea17a0035888850f424c7de25d
        - registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-driver@sha256:6b6f6816c9f4e608e93e7fbc0347255415426a0ff0c9a53b7750e5ba54d91262
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d5e46da8aff7d73d6f00c761dae94472bcda6e78f4f17b3802dc89d44de0111b
        - registry.k8s.io/sig-storage/csi-attacher@sha256:a399393ff5bd156277c56bae0c08389b1a1b95b7fd6ea44a316ce55e0dd559d7
        version: 1.0.0
      entries:
      - name: hnas-csi-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: hnas-csi-operator
    provider:
      name: Hitachi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-memorydb-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-memorydb-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "ACL",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "ParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "Snapshot",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "SubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/memorydb-controller:1.1.1
          createdAt: "2025-09-19T19:36:49Z"
          description: Amazon MemoryDB for Redis controller is a service controller
            for managing MemoryDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ACL represents the state of an AWS memorydb ACL resource.
            displayName: ACL
            kind: ACL
            name: acls.memorydb.services.k8s.aws
            version: v1alpha1
          - description: Cluster represents the state of an AWS memorydb Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.memorydb.services.k8s.aws
            version: v1alpha1
          - description: ParameterGroup represents the state of an AWS memorydb ParameterGroup
              resource.
            displayName: ParameterGroup
            kind: ParameterGroup
            name: parametergroups.memorydb.services.k8s.aws
            version: v1alpha1
          - description: Snapshot represents the state of an AWS memorydb Snapshot
              resource.
            displayName: Snapshot
            kind: Snapshot
            name: snapshots.memorydb.services.k8s.aws
            version: v1alpha1
          - description: SubnetGroup represents the state of an AWS memorydb SubnetGroup
              resource.
            displayName: SubnetGroup
            kind: SubnetGroup
            name: subnetgroups.memorydb.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS memorydb User resource.
            displayName: User
            kind: User
            name: users.memorydb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon MemoryDB for Redis resources in AWS from within your Kubernetes cluster.

          **About Amazon MemoryDB for Redis**

          MemoryDB for Redis is a durable, in-memory database service that delivers ultra-fast performance. It is purpose-built for modern applications with microservices architectures.
          MemoryDB is compatible with Redis, a popular open source data store, enabling you to quickly build applications using the same flexible and friendly Redis data structures, APIs, and commands that they already use today. With MemoryDB, all of your data is stored in memory, which enables you to achieve microsecond read and single-digit millisecond write latency and high throughput. MemoryDB also stores data durably across multiple Availability Zones (AZs) using a Multi-AZ transactional log to enable fast failover, database recovery, and node restarts.
          Delivering both in-memory performance and Multi-AZ durability, MemoryDB can be used as a high-performance primary database for your microservices applications, eliminating the need to separately manage both a cache and durable database.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - MemoryDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - memorydb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon MemoryDB for Redis Developer Resources
          url: https://docs.aws.amazon.com/memorydb/?id=docs_gateway
        maintainers:
        - email: ack-maintainers@amazon.com
          name: memorydb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/memorydb-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-memorydb-controller.v1.1.1
        version: 1.1.1
      - name: ack-memorydb-controller.v1.1.0
        version: 1.1.0
      - name: ack-memorydb-controller.v1.0.15
        version: 1.0.15
      - name: ack-memorydb-controller.v1.0.13
        version: 1.0.13
      - name: ack-memorydb-controller.v1.0.12
        version: 1.0.12
      - name: ack-memorydb-controller.v1.0.11
        version: 1.0.11
      - name: ack-memorydb-controller.v1.0.10
        version: 1.0.10
      - name: ack-memorydb-controller.v1.0.9
        version: 1.0.9
      - name: ack-memorydb-controller.v1.0.8
        version: 1.0.8
      - name: ack-memorydb-controller.v1.0.7
        version: 1.0.7
      - name: ack-memorydb-controller.v1.0.6
        version: 1.0.6
      - name: ack-memorydb-controller.v1.0.5
        version: 1.0.5
      - name: ack-memorydb-controller.v1.0.4
        version: 1.0.4
      - name: ack-memorydb-controller.v1.0.3
        version: 1.0.3
      - name: ack-memorydb-controller.v1.0.2
        version: 1.0.2
      - name: ack-memorydb-controller.v1.0.1
        version: 1.0.1
      - name: ack-memorydb-controller.v1.0.0
        version: 1.0.0
      - name: ack-memorydb-controller.v0.0.12
        version: 0.0.12
      - name: ack-memorydb-controller.v0.0.11
        version: 0.0.11
      - name: ack-memorydb-controller.v0.0.10
        version: 0.0.10
      - name: ack-memorydb-controller.v0.0.9
        version: 0.0.9
      - name: ack-memorydb-controller.v0.0.8
        version: 0.0.8
      - name: ack-memorydb-controller.v0.0.7
        version: 0.0.7
      - name: ack-memorydb-controller.v0.0.6
        version: 0.0.6
      - name: ack-memorydb-controller.v0.0.5
        version: 0.0.5
      - name: ack-memorydb-controller.v0.0.4
        version: 0.0.4
      - name: ack-memorydb-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-memorydb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: GitLab, Inc.
      provider-url: https://gitlab.com
    name: gitlab-runner-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-runner-operator.v1.41.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta2",
                "kind": "Runner",
                "metadata": {
                  "name": "gitlab-runner-example",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "gitlabUrl": "https://gitlab.example.com",
                  "token": "runner-token-secret",
                  "tags": "kubernetes,example",
                  "concurrent": 10,
                  "imagePullPolicy": "IfNotPresent"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery, Developer Tools
          containerImage: registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
          createdAt: "2025-09-25T15:08:40Z"
          description: GitLab Runner operator manages lifecycle of GitLab Runner instances
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.11.0 <1.41.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: none
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
          support: GitLab, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Runner is the open source project used to run your jobs and
              send the results back to GitLab
            displayName: GitLab Runner
            kind: Runner
            name: runners.apps.gitlab.com
            version: v1beta2
        description: |-
          GitLab Runner is the lightweight, highly-scalable agent that runs your build jobs and sends the results back to a GitLab instance. GitLab Runner works in conjunction with GitLab CI/CD, the open-source continuous integration service included with GitLab.

              The GitLab Runner operator manages the lifecycle of GitLab Runner in Kubernetes or Openshift clusters. The operator aims to automate the tasks needed to run your CI/CD jobs in your container orchestration platform.

              ## Prerequisites

                For Kubernetes cluster, install cert-manager:

                ```shell
                kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.7.1/cert-manager.yaml
                ```

              ## GitLab Runner version

              This version of **GitLab Runner Operator** ships with **GitLab Runner v18.4.0**.

              To use a different version of **GitLab Runner** change the [`runnerImage` and `helperImage` properties](https://docs.gitlab.com/runner/configuration/configuring_runner_operator.html#operator-properties).

              ## Usage

               To link a GitLab Runner instance to a self-hosted GitLab instance or the hosted [GitLab](https://gitlab.com), you first need to:

               1. Create a secret containing the `runner-token` from your GitLab project.

                ```
                cat > gitlab-runner-secret.yml << EOF
                apiVersion: v1
                kind: Secret
                metadata:
                  name: gitlab-runner-secret
                type: Opaque
                # Only one of the following fields must be set. The Operator will fail to register Runner if both are provided.
                # # NOTE: runner-registration-token is deprecated and will be removed in GitLab 18.0. You should use runner-token instead.
                stringData:
                  runner-token: REPLACE_ME # your project runner token
                  # runner-registration-token: "" # your project runner secret
                EOF
                ```

                ```
                oc apply -f gitlab-runner-secret.yml
                ```

               2. Create the Custom Resource Definition (CRD) file and include the following information. The tags value must be openshift for the job to run.

                 ```
                 cat > gitlab-runner.yml << EOF
                 apiVersion: apps.gitlab.com/v1beta2
                 kind: Runner
                 metadata:
                   name: gitlab-runner
                 spec:
                   gitlabUrl: https://gitlab.example.com
                   buildImage: alpine
                   token: gitlab-runner-secret
                   tags: openshift
                 EOF
                 ```

                ```
                oc apply -f gitlab-runner.yml
                ```

              ## Full documentation

              Visit [Install GitLab Runner Operator](https://docs.gitlab.com/runner/install/operator.html)
        displayName: GitLab Runner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GitLab
        - CI/CD
        - DevOps
        - SAST
        - DAST
        links:
        - name: Gitlab Runner Operator
          url: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
        - name: GitLab Docs
          url: https://docs.gitlab.com
        - name: GitLab Runner
          url: https://docs.gitlab.com/runner/
        - name: GitLab CI/CD
          url: https://docs.gitlab.com/ee/ci/quick_start/
        maintainers:
        - email: support@gitlab.com
          name: Georgi N. Georgiev
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: GitLab, Inc.
          url: https://gitlab.com
        relatedImages:
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-ocp@sha256:0f2a5a3b6779d719eecbf079b99cf06b652603d74b563f57d753e32966d253dd
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-helper-ocp@sha256:29727fae83a226bcf57607240d4046a0b8c208b5ff42ee1f8ca0f115bb188809
        version: 1.41.1
      entries:
      - name: gitlab-runner-operator.v1.41.1
        version: 1.41.1
      - name: gitlab-runner-operator.v1.41.0
        version: 1.41.0
      - name: gitlab-runner-operator.v1.40.1
        version: 1.40.1
      - name: gitlab-runner-operator.v1.40.0
        version: 1.40.0
      - name: gitlab-runner-operator.v1.39.0
        version: 1.39.0
      name: stable
    defaultChannel: stable
    packageName: gitlab-runner-operator
    provider:
      name: GitLab, Inc.
      url: https://gitlab.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Anjuna Security
      provider-url: https://www.anjuna.io
    name: anjuna-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anjuna-operator.v2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anjuna.io/v1",
                "kind": "AnjunaConfig",
                "metadata": {
                  "name": "anjuna-config-remote"
                },
                "spec": {
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "runtimeImage": "registry.connect.redhat.com/anjuna/anjuna-k8s-sev-runtime@sha256:1531b1ba955b2d3ea5d9238d2e20617a534bb1bab73f5fadadf8c9d3aee410b4"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-02-19T01:11:18Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          operatorframework.io/suggested-namespace: anjuna-operator
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnjunaConfig operand instructs the Anjuna Seaglass Operator
              to know (1) what needs to be installed into the cluster and (2) what
              nodes will be affected.
            displayName: Anjuna Config
            kind: AnjunaConfig
            name: anjunaconfigs.anjuna.io
            version: v1
        description: "The Anjuna Seaglass Operator for OpenShift allows running any
          Pod in\nOpenShift in a Confidential Pod leveraging the full power of the
          Confidential\nComputing technology.\n\n# Anjuna Seaglass\nAnjuna Seaglass™
          is the Universal Confidential Computing Platform to run applications in
          any cloud, and on-premises, with complete data security and privacy. \nIt
          isolates workloads in a hardware-backed environment that creates a zero-trust
          model for code and data.\n\nThe Anjuna Seaglass platform allows users to
          leverage the power of Confidential Computing:\n\nFor any Confidential Computing
          technology\nIn any Cloud Service Provider, as well as on-premises\nWith
          no changes to the application\n\nRunning a Confidential Pod with Anjuna
          Seaglass protects the Pod's memory with hardware-enforced encryption and
          integrity protection, and enables you to authenticate and trust the workload
          using Remote Attestation.\n\n# Data-in-use protection\nWhen protecting sensitive
          data managed by a Pod in OpenShift, one can protect the data at rest by
          encrypting the data written to the persistent storage, and protecting the
          data in transit when sending data over the network.\n\nHowever, when the
          data is used in the Pod's memory, it cannot be both encrypted and used at
          the same time.\nThis means that if a malicious insider (either human or
          software) had access to the OpenShift Node, they could dump the Pod's memory
          and access all of its sensitive data. This includes accessing the secrets
          used to encrypt the data at rest and in transit.\n\nRunning the Pod in a
          Confidential Pod with the Anjuna Seaglass platform solves this problem.
          Anjuna Seaglass runs the Pod inside a Trusted Execution Environment, where
          the Pod memory is encrypted by the hardware so that no privileged human
          or software on the Pod can access this memory in the clear.\n\n# Enabling
          zero trust for the Pod\nWhen running a Pod in OpenShift, the Pod may need
          to create trust with other Pods or with software running outside of the
          OpenShift cluster.\nThe standard way to create this trust is to identify
          with a token or credentials that could be easily read and stolen from the
          Pod disk or by providing a machine identity, a process name, or an IP that
          could be spoofed or reused.\n\nConfidential Computing technology provides
          a mechanism called Remote Attestation, which provides an identity to the
          software via the Hardware—the Software Measurement.\n\nWhen running a Pod
          as a Confidential Pod with the Anjuna Seaglass platform, a user can leverage
          the Remote Attestation to perform trusted authentication with the Confidential
          Pod, in a way that cannot be spoofed.\nThe Anjuna Policy Manager allows
          you to protect the first secret of a Confidential Pod and remove any secret
          from the Pod image or Kubernetes control plane that could be read by malicious
          insiders.\n\n# Quick start and configuration instructions\nFollow the directions
          in Installing the [Anjuna Kubernetes Toolset to OpenShift on GCP](https://docs.anjuna.io/k8s-sev/latest/installation/openshift_gcp.html)
          \nto learn how to deploy the Operator into your OpenShift cluster.\n\nFollow
          the directions in [Quickstart guides for OpenShift on GCP](https://docs.anjuna.io/k8s-sev/latest/applications/gcp.html)
          \nto learn how to deploy your first Confidential Pod in your OpenShift cluster.\n\nFollow
          the directions in [Attestation with Anjuna Policy Manager](https://docs.anjuna.io/k8s-sev/latest/attestation.html)
          \nto learn how to leverage Remote Attestation to protect your Pod's first
          secret."
        displayName: Anjuna Seaglass™ Operator for OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Security
        links:
        - name: Anjuna Operator Documentation
          url: https://docs.anjuna.io/k8s-sev/latest/index.html
        - name: Anjuna Seaglass
          url: https://www.anjuna.io/essence-of-anjuna-seaglass
        - name: Anjuna Security
          url: https://www.anjuna.io
        maintainers:
        - email: support@anjuna.io
          name: Anjuna Security
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Anjuna Security
          url: https://www.anjuna.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - registry.connect.redhat.com/anjuna/anjuna-k8s-sev-tools@sha256:e5ba31c240aaff8ca8f6addf12577fd6045bf6619a1d2cb2987b2ff5d75f4028
        - registry.connect.redhat.com/anjuna/anjuna-k8s-sev-runtime@sha256:1531b1ba955b2d3ea5d9238d2e20617a534bb1bab73f5fadadf8c9d3aee410b4
        version: 2.0.0
      entries:
      - name: anjuna-operator.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anjuna-operator
    provider:
      name: Anjuna Security
      url: https://www.anjuna.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzograph-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzograph-operator.v3.2.500
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzograph.clusters.cambridgesemantics.com/v2",
                "kind": "AnzoGraph",
                "metadata": {
                  "name": "azg01"
                },
                "spec": {
                  "db": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_data": "anzograph-data-grpc",
                            "app_mgmt": "anzograph-mgmt-grpc"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_data": "anzograph-data-grpc",
                              "app_mgmt": "anzograph-mgmt-grpc"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzograph-db@sha256:30ed2ec337135701ab8b54e75bb48ad8928a2c48818643c23b9dba4484255a2e",
                                "name": "db",
                                "resources": {
                                  "limits": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  },
                                  "requests": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  }
                                }
                              }
                            ],
                            "securityContext": {
                              "fsGroup": 1000,
                              "runAsNonRoot": true,
                              "runAsUser": 1000,
                              "seccompProfile": {
                                "type": "RuntimeDefault"
                              }
                            },
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  },
                  "deployFrontend": false,
                  "frontend": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_mgmt": "anzograph-frontend"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_mgmt": "anzograph-frontend"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzograph-frontend@sha256:e5446b3c04889f2189e955702c4165eb243eb89f47ac797eb417b60b1459647f",
                                "name": "frontend",
                                "resources": {
                                  "limits": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "securityContext": {
                              "fsGroup": 1000,
                              "runAsNonRoot": true,
                              "runAsUser": 1000,
                              "seccompProfile": {
                                "type": "RuntimeDefault"
                              }
                            },
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/anzograph-operator@sha256:2333154a14ab7f3a03a601f00c017feeafb9d9814aff5e59c95cef615ab2abd6
          createdAt: "2025-09-22T12:03:18Z"
          description: kubernetes operator for Altair Graph Lakehouse DB
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Altair Engineering Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoGraph is the Schema for the anzographs API
            displayName: Altair Graph Lakehouse Operator
            kind: AnzoGraph
            name: anzographs.anzograph.clusters.cambridgesemantics.com
            version: v2
        description: |-
          The Altair Graph Lakehouse Operator provides the way to install and configure an Altair Graph Lakehouse
          cluster on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v3.2.5/README_openshift_marketplace.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzograph/userdoc/ )

          ### Support

          We offer Support to our customers with the Altair Graph Lakehouse db Enterprise Edition License [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Lakehouse Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        - Database
        links:
        - name: Anzograph Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v3.2.5/README_openshift_marketplace.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzograph-operator@sha256:2333154a14ab7f3a03a601f00c017feeafb9d9814aff5e59c95cef615ab2abd6
        - registry.connect.redhat.com/cambridgesemantics/anzograph-db@sha256:30ed2ec337135701ab8b54e75bb48ad8928a2c48818643c23b9dba4484255a2e
        - registry.connect.redhat.com/cambridgesemantics/anzograph-frontend@sha256:e5446b3c04889f2189e955702c4165eb243eb89f47ac797eb417b60b1459647f
        version: 3.2.500
      entries:
      - name: anzograph-operator.v3.2.500
        version: 3.2.500
      - name: anzograph-operator.v3.2.300
        version: 3.2.300
      - name: anzograph-operator.v3.2.200
        version: 3.2.200
      - name: anzograph-operator.v3.2.104
        version: 3.2.104
      name: stable
    defaultChannel: stable
    packageName: anzograph-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: kubernetes-nmstate-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubernetes-nmstate-operator.4.13.0-202309261702
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nmstate.io/v1",
                "kind": "NMState",
                "metadata": {
                  "name": "nmstate"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift4/kubernetes-nmstate-rhel8-operator@sha256:bdcb80591b839ec6b4b94efa838ca41585f3044826534413b91efd2acdca4b4d
          createdAt: "2022-02-21 08:46:16"
          description: |
            Kubernetes NMState is a declaritive means of configuring NetworkManager.
          olm.skipRange: '>=4.3.0 <4.13.0'
          operatorframework.io/suggested-namespace: openshift-nmstate
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/kubernetes-nmstate
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an NMState deployment.
            displayName: NMState
            kind: NMState
            name: nmstates.nmstate.io
            version: v1
        description: A Kubernetes Operator to install Kubernetes NMState
        displayName: Kubernetes NMState Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - nmstate
        - networking
        - NetworkManager
        links:
        - name: Kubernetes Nmstate Operator
          url: https://github.com/nmstate/kubernetes-nmstate
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: GA
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/kubernetes-nmstate-rhel8-operator@sha256:bdcb80591b839ec6b4b94efa838ca41585f3044826534413b91efd2acdca4b4d
        - registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel8@sha256:a96434b959c1f61c1cfb4352284aaae453d2e41ca0824d016975a1216451b665
        version: 4.13.0-202309261702
      entries:
      - name: kubernetes-nmstate-operator.4.13.0-202309261702
        version: 4.13.0-202309261702
      name: "4.13"
    - currentCSV: kubernetes-nmstate-operator.4.16.0-202509241326
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nmstate.io/v1",
                "kind": "NMState",
                "metadata": {
                  "name": "nmstate"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:14750ac40ba2a9a92e2b37351eed94c22f2663747ea09f241a3a80607a9fca09
          createdAt: "2022-02-21 08:46:16"
          description: |
            Kubernetes NMState is a declaritive means of configuring NetworkManager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-nmstate
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/kubernetes-nmstate
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an NMState deployment.
            displayName: NMState
            kind: NMState
            name: nmstates.nmstate.io
            version: v1
        description: A Kubernetes Operator to install Kubernetes NMState
        displayName: Kubernetes NMState Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - nmstate
        - networking
        - NetworkManager
        links:
        - name: Kubernetes Nmstate Operator
          url: https://github.com/nmstate/kubernetes-nmstate
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: GA
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:14750ac40ba2a9a92e2b37351eed94c22f2663747ea09f241a3a80607a9fca09
        - registry.redhat.io/openshift4/nmstate-console-plugin-rhel9@sha256:8d03a6c414da30521e0794d3d61336bfbb1d9d4c3fec5ab5dac4bd24c897a540
        - registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel9@sha256:c592fdb65d99de2c3d274d7e72621408e578bdba39753bb425d0765ac016b11a
        version: 4.16.0-202509241326
      entries:
      - name: kubernetes-nmstate-operator.4.16.0-202509241326
        version: 4.16.0-202509241326
      - name: kubernetes-nmstate-operator.4.16.0-202509101329
        version: 4.16.0-202509101329
      - name: kubernetes-nmstate-operator.4.16.0-202508251030
        version: 4.16.0-202508251030
      - name: kubernetes-nmstate-operator.4.16.0-202507221936
        version: 4.16.0-202507221936
      - name: kubernetes-nmstate-operator.4.16.0-202507100305
        version: 4.16.0-202507100305
      - name: kubernetes-nmstate-operator.4.16.0-202506250035
        version: 4.16.0-202506250035
      - name: kubernetes-nmstate-operator.4.16.0-202506041738
        version: 4.16.0-202506041738
      - name: kubernetes-nmstate-operator.4.16.0-202505150035
        version: 4.16.0-202505150035
      - name: kubernetes-nmstate-operator.4.16.0-202505060805
        version: 4.16.0-202505060805
      - name: kubernetes-nmstate-operator.4.16.0-202504141106
        version: 4.16.0-202504141106
      - name: kubernetes-nmstate-operator.4.16.0-202503260136
        version: 4.16.0-202503260136
      - name: kubernetes-nmstate-operator.4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: kubernetes-nmstate-operator.4.16.0-202502190034
        version: 4.16.0-202502190034
      - name: kubernetes-nmstate-operator.4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: kubernetes-nmstate-operator.4.16.0-202501281135
        version: 4.16.0-202501281135
      - name: kubernetes-nmstate-operator.4.16.0-202501211505
        version: 4.16.0-202501211505
      - name: kubernetes-nmstate-operator.4.16.0-202501080105
        version: 4.16.0-202501080105
      - name: kubernetes-nmstate-operator.4.16.0-202412180105
        version: 4.16.0-202412180105
      - name: kubernetes-nmstate-operator.4.16.0-202412110604
        version: 4.16.0-202412110604
      - name: kubernetes-nmstate-operator.4.16.0-202412040032
        version: 4.16.0-202412040032
      - name: kubernetes-nmstate-operator.4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: kubernetes-nmstate-operator.4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: kubernetes-nmstate-operator.4.16.0-202410300036
        version: 4.16.0-202410300036
      - name: kubernetes-nmstate-operator.4.16.0-202410231737
        version: 4.16.0-202410231737
      - name: kubernetes-nmstate-operator.4.16.0-202410180404
        version: 4.16.0-202410180404
      - name: kubernetes-nmstate-operator.4.16.0-202410091137
        version: 4.16.0-202410091137
      - name: kubernetes-nmstate-operator.4.16.0-202410021935
        version: 4.16.0-202410021935
      - name: kubernetes-nmstate-operator.4.16.0-202409251605
        version: 4.16.0-202409251605
      - name: kubernetes-nmstate-operator.4.16.0-202409111235
        version: 4.16.0-202409111235
      - name: kubernetes-nmstate-operator.4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: kubernetes-nmstate-operator.4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: kubernetes-nmstate-operator.4.16.0-202408211407
        version: 4.16.0-202408211407
      - name: kubernetes-nmstate-operator.4.16.0-202408140143
        version: 4.16.0-202408140143
      - name: kubernetes-nmstate-operator.4.16.0-202408070540
        version: 4.16.0-202408070540
      - name: kubernetes-nmstate-operator.4.16.0-202407251436
        version: 4.16.0-202407251436
      - name: kubernetes-nmstate-operator.4.16.0-202407240806
        version: 4.16.0-202407240806
      - name: kubernetes-nmstate-operator.4.16.0-202407181806
        version: 4.16.0-202407181806
      - name: kubernetes-nmstate-operator.4.16.0-202407101306
        version: 4.16.0-202407101306
      - name: kubernetes-nmstate-operator.4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: kubernetes-nmstate-operator.4.16.0-202406260236
        version: 4.16.0-202406260236
      - name: kubernetes-nmstate-operator.4.16.0-202406140306
        version: 4.16.0-202406140306
      - name: kubernetes-nmstate-operator.4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: kubernetes-nmstate-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: istio-workspace-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: istio-workspace-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "workspace.maistra.io/v1alpha1",
                "kind": "Session",
                "metadata": {
                  "name": "session-sample"
                },
                "spec": {
                  "ref": [
                    {
                      "args": {
                        "image": "quay.io/maistra-dev/istio-workspace-test:latest"
                      },
                      "name": "ratings-v1",
                      "strategy": "prepared-image"
                    }
                  ],
                  "route": {
                    "name": "x-workspace-route",
                    "type": "header",
                    "value": "feature-x"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/maistra/istio-workspace:v0.5.3
          createdAt: "2022-11-16T13:50:35Z"
          description: Safely develop and test on any Kubernetes cluster without affecting
            others.
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/maistra/istio-workspace
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Session controls the creation of the specialized hidden routes.
            displayName: Session
            kind: Session
            name: sessions.workspace.maistra.io
            version: v1alpha1
        description: "Istio Workspace enables developers to:\n\n * run one or more
          services locally during development but make it appear like it would be
          developed in the actual cluster\n   * this way you can use your favourite
          development tools\n   * have instant feedback loop\n   * don't make your
          machine overheating trying to run the whole cluster locally\n * safely preview
          incoming changes in pull requests without affecting regular users \n * have
          confidence in testing or troubleshooting your services directly in the cluster\n\nIstio
          Workspace supports testing on multi-user environment in an unobtrusive way.\nIt
          doesn’t really matter if it is QE cluster or actual production. We give
          you confidence that your changes won’t blow up the cluster, and your users
          won’t even notice a glitch.\n\nYou will need [Kubernetes](https://k8s.io)
          or [Openshift](https://openshift.com) cluster with  [Istio](https://istio.io/)
          installed. \n\nYou can read more about our vision on [Red Hat Developer’s
          blog](https://developers.redhat.com/blog/2020/07/14/developing-and-testing-on-production-with-kubernetes-and-istio-workspace/)
          or ...\n\n## See it in action!\n\n[![https://youtu.be/XTNVadUzMCc](https://img.youtube.com/vi/XTNVadUzMCc/hqdefault.jpg)](https://youtu.be/XTNVadUzMCc)\n\nWatch
          demo: [\"How to develop on production: An introduction to Istio-Workspaces\"](https://youtu.be/XTNVadUzMCc).\n\n##
          Documentation\n\nMore details can be found on our [documentation page](https://istio-workspace-docs.netlify.com/).\n"
        displayName: Istio Workspace
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - telepresence
        links:
        - name: Istio Workspace Documentation
          url: https://istio-workspace-docs.netlify.com/
        - name: How to install CLI tool
          url: https://github.com/maistra/istio-workspace#install-in-two-easy-steps
        maintainers:
        - email: aslak@redhat.com
          name: Aslak Knutsen
        - email: bartosz@redhat.com
          name: Bartosz Majsak
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/maistra/istio-workspace:v0.5.3
        version: 0.5.3
      entries:
      - name: istio-workspace-operator.v0.5.3
        version: 0.5.3
      - name: istio-workspace-operator.v0.5.2
        version: 0.5.2
      - name: istio-workspace-operator.v0.5.1
        version: 0.5.1
      - name: istio-workspace-operator.v0.5.0
        version: 0.5.0
      - name: istio-workspace-operator.v0.4.0
        version: 0.4.0
      - name: istio-workspace-operator.v0.3.0
        version: 0.3.0
      - name: istio-workspace-operator.v0.2.0
        version: 0.2.0
      - name: istio-workspace-operator.v0.1.0
        version: 0.1.0
      - name: istio-workspace-operator.v0.0.10
        version: 0.0.10
      - name: istio-workspace-operator.v0.0.9
        version: 0.0.9
      - name: istio-workspace-operator.v0.0.8
        version: 0.0.8
      - name: istio-workspace-operator.v0.0.7
        version: 0.0.7
      - name: istio-workspace-operator.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: istio-workspace-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibmcloud-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibmcloud-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"ibmcloud.ibm.com/v1","kind":"Service","metadata":{"annotations":{"ibmcloud.ibm.com/self-healing":"enabled"},"name":"mytranslator"},"spec":{"plan":"lite","serviceClass":"language-translator"}},{"apiVersion":"ibmcloud.ibm.com/v1","kind":"Binding","metadata":{"name":"binding-translator"},"spec":{"parameters":[{"name":"serviceid_crn","value":"crn:v1:bluemix:public:iam-identity::a/ACCOUNT_ID::serviceid:SERVICE_ID"}],"secretName":"translator-secret","serviceName":"mytranslator"}}]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: cloudoperators/ibmcloud-operator:1.1.0
          createdAt: "2022-03-17T04:17:40Z"
          description: The IBM Cloud Operator provides a Kubernetes CRD-Based API
            to manage the lifecycle of IBM public cloud services.
          repository: https://github.com/IBM/cloud-operators
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Binding is an instance of a service binding resource on IBM
              Cloud. A Binding creates a secret with the service instance credentials.
            displayName: Binding
            kind: Binding
            name: bindings.ibmcloud.ibm.com
            version: v1
          - description: Service is an instance of a Service resource on IBM Cloud.
            displayName: Service
            kind: Service
            name: services.ibmcloud.ibm.com
            version: v1
        description: |-
          With the IBM Cloud Operator, you can provision and bind [IBM public cloud services](https://cloud.ibm.com/catalog#services) to your Kubernetes cluster in a Kubernetes-native way. The IBM Cloud Operator is based on the [Kubernetes custom resource definition (CRD) API](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/) so that your applications can create, update, and delete IBM Cloud services from within the cluster by calling Kubnernetes APIs, instead of needing to use several IBM Cloud APIs in addition to configuring your app for Kubernetes.


          ## Features

          * **Service provisioning**: Create any service with any plan that is available in the [IBM Cloud catalog](https://cloud.ibm.com/catalog#services).

          * **Bindings management**: Automatically create Kubernetes secrets in your Kubernetes cluster with the credentials to bind to any provisioned service to the cluster.


          ## Upgrading the operator


          To upgrade an OperatorHub installation, use the OpenShift web console to perform the upgrade. If upgrading from the `alpha` channel subscription to `stable`, follow the additional upgrade instructions below.

          ### Upgrading to version v0.3 or v1.0
          **IMPORTANT NOTICE:** v0.1 and v0.2 used a different naming scheme for secrets and configmaps. Before you update the IBM Cloud Operator, create secret and configmap resources with these names and copy the contents of your previous resources to the new resources. Then, the upgraded operator recognizes and continues to update the resources.

          Existing `Service` and `Binding` resources **do not** need modification for the upgrade.

          | Previous names (v0.1 and v0.2)         | **Current names (v0.3 and v1.0)**             | Description                                                                           |
          |:---------------------------------------|:--------------------------------------------|:--------------------------------------------------------------------------------------|
          | secret-ibm-cloud-operator              | **ibmcloud-operator-secret**                | Secret with the API key, scoped to the namespace.                                     |
          | config-ibm-cloud-operator              | **ibmcloud-operator-defaults**              | ConfigMap with the default values for new resources.                                  |
          | ibm-cloud-operator                     | **ibmcloud-operator-config**                | ConfigMap with the management namespace configuration.                                |
          | ${namespace}-secret-ibm-cloud-operator | **${namespace}-ibmcloud-operator-secret**   | Management namespace Secret with the API key for ${namespace}.                        |
          | ${namespace}-config-ibm-cloud-operator | **${namespace}-ibmcloud-operator-defaults** | Management namespace ConfigMap with default values for new resources in ${namespace}. |

          **Tip:** Forgot to update the secret and configmap names before upgrading? The operator will not take action on Service and Binding resources until the new secrets and configmaps have been created. Creating these after the upgrade will also work.


          ### OperatorHub stable channel
          Continuous updates and bug fixes are provided in the latest stable channel. Subscribing to the `stable` channel in OperatorHub means you automatically get the latest updates without breaking backward compatibility.

          ### Changelog

          For a list of recent changes, see the [releases page](https://github.com/IBM/cloud-operators/releases).


          ## Prerequisites

          1.  Have an [IBM Cloud account](https://cloud.ibm.com/registration).
          2.  Have a cluster that runs Kubernetes version 1.11 or later (OpenShift 3.11 or later).
          3.  Install the required command line tools.
              *   [IBM Cloud CLI](https://cloud.ibm.com/docs/cli?topic=cloud-cli-getting-started) (`ibmcloud`)
              *   [Kubernetes CLI](https://kubernetes.io/docs/tasks/tools/install-kubectl/) (`kubectl`)
          4.  Log in to your IBM Cloud account from the CLI.

              ```bash
              ibmcloud login
              ```

          5.  Target the appropriate resource group (`-g`) and default region (`-r`) for provisioning services:

              ```bash
              ibmcloud target -g default -r us-south
              ```

          6.  If you also want to provision Cloud Foundry services, target the appropriate `org` and `space` (`--cf`).

              ```bash
              ibmcloud target --cf
              ```

          7.  Set the Kubernetes context of your CLI to your cluster so that you can run `kubectl` commands. For example, if your cluster runs OpenShift, use the `oc login` command.


          ## Setting up the operator

          You can use an installation script to set up the IBM Cloud Operator.
          By default, the installation script stores an API key in a Kubernetes secret in your cluster that can be accessed by the IBM Cloud Operator.
          Next, the script sets default values that are used to provision IBM Cloud services, like the resource group and region to provision the services in. Later, you can override any default value in the `Service` custom resource. Finally, the script deploys the operator in your cluster.

          To use your own API key, set the `IBMCLOUD_API_KEY` environment variable to the key before running the installation script:

              export IBMCLOUD_API_KEY="CUSTOM_API_KEY"

          If installed with Operator Hub, the operator will run in the `openshift-operators` namespace. Otherwise, it will run in the `ibmcloud-operator-system` namespace.

          Prefer to create the secrets and defaults yourself? See the [IBM Cloud Operator installation guide](https://github.com/IBM/cloud-operators/blob/master/docs/install.md).

          ### Using a service ID

          By default, the installation script creates an IBM Cloud API key that impersonates your user credentials, to use to set up the IBM Cloud Operator. However, you might want to create a service ID in IBM Cloud Identity and Access Managment (IAM). By using a service ID, you can control access for the IBM Cloud Operator without having the permissions tied to a particular user, such as if that user leaves the company. For more information, see the [IBM Cloud docs](https://cloud.ibm.com/docs/account?topic=account-serviceids).

          1.  Create a service ID in IBM Cloud IAM. If possible, do not use spaces in the names for your IAM credentials. When you use the operator binding feature, any spaces are replaced with underscores.

              ```bash
              ibmcloud iam service-id-create serviceid-ico -d service-ID-for-ibm-cloud-operator
              ```

          2.  Assign the service ID access to the required permissions to work with the IBM Cloud services that you want the operator to manage. The required permissions vary with each IBM Cloud service. You can also scope an access policy to different regions or resource groups. For example, the following command grants the service ID the **Administrator** platform role in the default resource group in the US South (Dallas) region. For more information, see the [IBM Cloud docs](https://cloud.ibm.com/docs/account?topic=account-userroles).

              ```bash
              ibmcloud iam service-policy-create serviceid-ico --roles Administrator --resource-group-name default --region us-south
              ```

          3.  Create an API key for the service ID.

              ```bash
              ibmcloud iam service-api-key-create apikey-ico serviceid-ico -d api-key-for-ibm-cloud-operator
              ```

          4.  Set the API key of the service ID as your CLI environment variable. Now, when you run the installation script, the script uses the service ID's API key. The following command is an example for macOS.

              ```bash
              export IBMCLOUD_API_KEY=<apikey-ico-value>
              ```

          5.  Confirm that the API key environment variable is set in your CLI.

              ```bash
              echo $IBMCLOUD_API_KEY
              ```

          6.  Follow the [prerequisite steps](https://github.com/IBM/cloud-operators/blob/master/README.md#prerequisites) to log in to the IBM Cloud account that owns the service ID.


          ### Installing the operator for OpenShift clusters

          Before you begin, complete the [prerequisite steps](https://github.com/IBM/cloud-operators/blob/master/README.md#prerequisites) to log in to IBM Cloud and your cluster, and optionally set up a [service ID API key](https://github.com/IBM/cloud-operators/blob/master/README.md#using-a-service-id).

          To configure the latest release for OpenShift before installing via the OperatorHub, run the following script:

          *   **Latest release**:

              ```bash
              curl -sL https://raw.githubusercontent.com/IBM/cloud-operators/master/hack/configure-operator.sh | bash
              ```

          *   **Specific release**: Replace `-v 0.0.0` with the specific version that you want to install.

              ```bash
              curl -sL https://raw.githubusercontent.com/IBM/cloud-operators/master/hack/configure-operator.sh | bash -s -- -v 0.0.0 store-creds
              ```


          ## Using the IBM Cloud Operator

          To use the IBM Cloud Operator, create a service instance and then bind the service to your cluster. For more information, see the [sample configuration files](https://github.com/IBM/cloud-operators/blob/master/config/samples), [user guide](https://github.com/IBM/cloud-operators/blob/master/docs/user-guide.md), and [reference documentation](https://github.com/IBM/cloud-operators/blob/master/README.md#reference).

          #### Step 1: Creating a service instance

          1.  To create an instance of an IBM public cloud service, first create a `Service` custom resource file. For more options, see the [Service properties](https://github.com/IBM/cloud-operators/blob/master/README.md#service-properties) reference doc.
              *   `<SERVICE_CLASS>` is the IBM Cloud service that you want to create. To list IBM Cloud services, run `ibmcloud catalog service-marketplace` and use the **Name** value from the output.
              *   `<PLAN>` is the plan for the IBM Cloud service that you want to create, such as `free` or `standard`. To list available plans, run `ibmcloud catalog service <SERVICE_CLASS> | grep plan`.

              ```yaml
              apiVersion: ibmcloud.ibm.com/v1
              kind: Service
              metadata:
                  name: myservice
              spec:
                  plan: <PLAN>
                  serviceClass: <SERVICE_CLASS>
              ```

          2.  Create the service instance in your cluster.

              ```bash
              kubectl apply -f filepath/myservice.yaml
              ```

          3.  Check that your service status is **Online** in your cluster.

              ```bash
              kubectl get services.ibmcloud
              NAME           STATUS   AGE
              myservice      Online   12s
              ```

          4.  Verify that your service instance is created in IBM Cloud.

              ```bash
              ibmcloud resource service-instances | grep myservice
              ```

          #### Step 2: Binding the service instance

          1.  To bind your service to the cluster so that your apps can use the service, create a `Binding` custom resource, where the `serviceName` field is the name of the `Service` custom resource that you previously created. For more options, see [Binding properties](https://github.com/IBM/cloud-operators/blob/master/README.md#binding-properties).

              ```yaml
              apiVersion: ibmcloud.ibm.com/v1
              kind: Binding
              metadata:
                  name: mybinding
              spec:
                  serviceName: myservice
              ```

          2.  Create the binding in your cluster.

              ```bash
              kubectl apply -f filepath/mybinding.yaml
              ```

          3.  Check that your service status is **Online**.

              ```bash
              kubectl get bindings.ibmcloud
              NAME         STATUS   AGE
              mybinding    Online   25s
              ```

          4.  Check that a secret of the same name as your binding is created. The secret contains the service credentials that apps in your cluster can use to access the service.

              ```bash
              kubectl get secrets
              NAME         TYPE      DATA   AGE
              mybinding    Opaque    6      102s
              ```
        displayName: IBM Cloud Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: Source on GitHub
          url: https://github.com/ibm/cloud-operators
        maintainers:
        - email: johnstarich@johnstarich.com
          name: John Starich
        - email: mvaziri@us.ibm.com
          name: vazirim
        - email: greg.lanthier@ca.ibm.com
          name: Greg Lanthier
        - email: halacs@users.noreply.github.com
          name: Halacs
        - email: artberger@users.noreply.github.com
          name: Art
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: IBM
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - cloudoperators/ibmcloud-operator:1.1.0
        version: 1.1.0
      entries:
      - name: ibmcloud-operator.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: ibmcloud-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Infinispan
      provider-url: ""
    name: infinispan
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infinispan-operator.v2.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.2.5.Final
          description: Create and manage Infinispan clusters.
          olm.skipRange: '>=2.1.x <2.2.1'
          operators.operatorframework.io/builder: operator-sdk-v1.3.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/infinispan/operator:2.2.5.Final
        version: 2.2.5
      entries:
      - name: infinispan-operator.v2.2.5
        version: 2.2.5
      - name: infinispan-operator.v2.2.4
        version: 2.2.4
      - name: infinispan-operator.v2.2.3
        version: 2.2.3
      - name: infinispan-operator.v2.2.2
        version: 2.2.2
      - name: infinispan-operator.v2.2.1
        version: 2.2.1
      - name: infinispan-operator.v2.2.0
        version: 2.2.0
      name: 2.2.x
    - currentCSV: infinispan-operator.v2.3.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.3.7.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/infinispan/operator:2.3.7.Final
        version: 2.3.8
      entries:
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: 2.3.x
    - currentCSV: infinispan-operator.v2.4.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.4.17.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/infinispan/operator:2.4.17.Final
        version: 2.4.17
      entries:
      - name: infinispan-operator.v2.4.17
        version: 2.4.17
      - name: infinispan-operator.v2.4.16
        version: 2.4.16
      - name: infinispan-operator.v2.4.15
        version: 2.4.15
      - name: infinispan-operator.v2.4.14
        version: 2.4.14
      - name: infinispan-operator.v2.4.13
        version: 2.4.13
      - name: infinispan-operator.v2.4.12
        version: 2.4.12
      - name: infinispan-operator.v2.4.11
        version: 2.4.11
      - name: infinispan-operator.v2.4.10
        version: 2.4.10
      - name: infinispan-operator.v2.4.9
        version: 2.4.9
      - name: infinispan-operator.v2.4.8
        version: 2.4.8
      - name: infinispan-operator.v2.4.7
        version: 2.4.7
      - name: infinispan-operator.v2.4.6
        version: 2.4.6
      - name: infinispan-operator.v2.4.5
        version: 2.4.5
      - name: infinispan-operator.v2.4.4
        version: 2.4.4
      - name: infinispan-operator.v2.4.3
        version: 2.4.3
      - name: infinispan-operator.v2.4.2
        version: 2.4.2
      - name: infinispan-operator.v2.4.1
        version: 2.4.1
      - name: infinispan-operator.v2.4.0
        version: 2.4.0
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: 2.4.x
    - currentCSV: infinispan-operator.v2.4.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.4.18.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/infinispan/operator:2.4.18.Final
        version: 2.4.18
      entries:
      - name: infinispan-operator.v2.4.18
        version: 2.4.18
      - name: infinispan-operator.v2.4.16
        version: 2.4.16
      - name: infinispan-operator.v2.4.15
        version: 2.4.15
      - name: infinispan-operator.v2.4.14
        version: 2.4.14
      - name: infinispan-operator.v2.4.13
        version: 2.4.13
      - name: infinispan-operator.v2.4.12
        version: 2.4.12
      - name: infinispan-operator.v2.4.11
        version: 2.4.11
      - name: infinispan-operator.v2.4.10
        version: 2.4.10
      - name: infinispan-operator.v2.4.9
        version: 2.4.9
      - name: infinispan-operator.v2.4.8
        version: 2.4.8
      - name: infinispan-operator.v2.4.7
        version: 2.4.7
      - name: infinispan-operator.v2.4.6
        version: 2.4.6
      - name: infinispan-operator.v2.4.5
        version: 2.4.5
      - name: infinispan-operator.v2.4.4
        version: 2.4.4
      - name: infinispan-operator.v2.4.3
        version: 2.4.3
      - name: infinispan-operator.v2.4.2
        version: 2.4.2
      - name: infinispan-operator.v2.4.1
        version: 2.4.1
      - name: infinispan-operator.v2.4.0
        version: 2.4.0
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: stable
    defaultChannel: stable
    packageName: infinispan
    provider:
      name: Infinispan
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent, part of Cisco
      provider-url: https://isovalent.com/
    name: clife
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clife.v1.17.8-cee.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cilium.io/v1alpha1",
                "kind": "CiliumConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "clife"
                  },
                  "name": "ciliumconfig"
                },
                "spec": {
                  "cluster": {
                    "name": "default"
                  },
                  "clusterHealthPort": 9940,
                  "cni": {
                    "binPath": "/var/lib/cni/bin",
                    "chainingMode": "portmap",
                    "confPath": "/var/run/multus/cni/net.d",
                    "exclusive": false
                  },
                  "enterprise": {
                    "featureGate": {
                      "approved": [
                        "CNIChainingMode"
                      ]
                    }
                  },
                  "hubble": {
                    "enabled": true
                  },
                  "ipam": {
                    "mode": "cluster-pool"
                  },
                  "kubeProxyReplacement": false,
                  "operator": {
                    "prometheus": {
                      "enabled": true,
                      "serviceMonitor": {
                        "enabled": true
                      }
                    }
                  },
                  "prometheus": {
                    "enabled": true,
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "securityContext": {
                    "privileged": true
                  },
                  "sessionAffinity": true,
                  "tunnelPort": 4789
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Networking,Security
          createdAt: "2025-09-23T07:55:19Z"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CiliumConfig defines the configuration of Isovalent Networking
              for Kubernetes and all its components
            displayName: CiliumConfig
            kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Networking for Kubernetes, built on Cilium, provides
          eBPF-based networking, observability, and security with optimal scale and
          performance for platform teams operating Kubernetes environments across
          cloud and on-prem infrastructure.
        displayName: Isovalent Networking for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cilium
        - networking
        - observability
        - security
        - eBPF
        - CNI
        links:
        - name: Isovalent Networking for Kubernetes
          url: https://isovalent.com/product/
        maintainers:
        - email: support@isovalent.com
          name: Isovalent, part of Cisco
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Isovalent, part of Cisco
          url: https://isovalent.com/
        relatedImages:
        - quay.io/isovalent/startup-script-ubi:v1.17.8-cee.1@sha256:9a6a413262ec59026345aefc743de847891d80e31ef2f41db954acdcee11c2ff
        - quay.io/isovalent/certgen-ubi:v1.17.8-cee.1@sha256:fb63bed641a279cc09114e93821c7b7555c61393ab594ea2b823ce34fa73335e
        - quay.io/isovalent/cilium-envoy-ubi:v1.17.8-cee.1@sha256:c3de0334f90f9b2540cdec7e2356d813f2df93f0d7d8701005eab7161c7663b5
        - quay.io/isovalent/clife@sha256:ba9bd4a5c35e1d26e395afbc2a4de89f42df1e5e61c9b23959c23194368b7b0e
        - quay.io/isovalent/operator-generic-ubi:v1.17.8-cee.1@sha256:37e4b9f65b2c67d24afeeebebe5d0cb59c09280179de57cd77da2b68e19c3c1b
        - quay.io/isovalent/cilium-ubi:v1.17.8-cee.1@sha256:e9816096e481fad5d95209f902e4a7018aeb65d1e3be1574e6b289e21229f87a
        - quay.io/isovalent/hubble-relay-ubi:v1.17.8-cee.1@sha256:88964d4836d59420dc7e0abb758654ceaf2abcfb53e19bbe2c7962aa68874e37
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.17.8-cee.1@sha256:b097c823638b4fdbe5bdd2b6040a17fef0c70a020e67a18f8231d60eeac02dd5
        version: 1.17.8-cee.1
      entries:
      - name: clife.v1.17.8-cee.1
        version: 1.17.8-cee.1
      - name: clife.v1.17.7-cee.1
        version: 1.17.7-cee.1
      - name: clife.v1.17.5-cee.1
        version: 1.17.5-cee.1
      - name: clife.v1.17.4-cee.1
        version: 1.17.4-cee.1
      - name: clife.v0.0.1
        version: 0.0.1
      name: "1.17"
    defaultChannel: "1.17"
    packageName: clife
    provider:
      name: Isovalent, part of Cisco
      url: https://isovalent.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microsafe
      provider-url: https://microsafe.com.mx/
    name: operator-aims-mobile
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operator-aims-mobile.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.microsafe.com.mx/v1",
                "kind": "AimsMobile",
                "metadata": {
                  "name": "aimsmobile-sample"
                },
                "spec": {
                  "adminApp": {
                    "autoscaling": {
                      "enabled": false
                    },
                    "config": {
                      "NODE_ENV": "production",
                      "VITE_EXPRESS_HOST": "aims-mob-service",
                      "VITE_EXPRESS_PORT": "8080",
                      "VITE_EXPRESS_PROTOCOL": "http",
                      "VITE_KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "VITE_KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "VITE_KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "VITE_VERSION": "1.0.0"
                    },
                    "image": "quay.io/microsafe_app_ban/aims-mob-admin-private@sha256:8e333f900885b0d3f6442c8ffcb43bd93540b8c38ad5bf17cbbd78929e0b1be0",
                    "ingress": {
                      "enabled": true,
                      "isOpenshiftRoute": true
                    },
                    "name": "aims-mob-admin",
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "secretName": "aims-mob-admin-secrets",
                    "service": {
                      "enabled": true,
                      "port": 8083,
                      "targetPort": 8080
                    }
                  },
                  "dbService": {
                    "autoscaling": {
                      "enabled": false
                    },
                    "config": {
                      "MSSQL_HOST": "mssql-database.prod.host",
                      "MSSQL_PORT": "mssql-database.prod.port"
                    },
                    "image": "quay.io/microsafe_app_ban/aims-db-service-private@sha256:bcb4d8624626436526549427f9f169cff714b39cbcc530dd2b3470c6712abaf9",
                    "ingress": {
                      "enabled": false
                    },
                    "name": "aims-db-service",
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "300m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "128Mi"
                      }
                    },
                    "secretName": "aims-db-service-secrets",
                    "service": {
                      "enabled": true,
                      "port": 3000,
                      "targetPort": 3000
                    }
                  },
                  "mainService": {
                    "annotations": {
                      "instrumentation.opentelemetry.io/inject-nodejs": "true"
                    },
                    "autoscaling": {
                      "enabled": true,
                      "maxReplicas": 2,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "config": {
                      "AIMS1_API_URL": "http://aims-db-service:3000",
                      "EXPRESS_PORT": "4000",
                      "FRONTEND_EXPRESS_HOST": "aims-mob-admin",
                      "FRONTEND_EXPRESS_PORT": "4000",
                      "FRONTEND_EXPRESS_PROTOCOL": "http",
                      "FRONTEND_IP": "aims-mob-admin",
                      "FRONTEND_KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "FRONTEND_KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "FRONTEND_KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "FRONTEND_PORT": "80",
                      "FRONTEND_PROTOCOL": "http",
                      "FRONTEND_VERSION": "1.0.0",
                      "KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "MSSQL_HOST": "mssql-database.prod.host",
                      "MSSQL_PORT": "mssql-database.prod.port",
                      "NODE_ENV": "production",
                      "OTLP_LOG_EXPORTER_URL": "opentelemetry-logs.prod.url",
                      "OTLP_METRIC_EXPORTER_URL": "opentelemetry-metrics.prod.url",
                      "OTLP_TRACE_EXPORTER_URL": "opentelemetry-traces.prod.url",
                      "REMEDY_API_URL": "remedy-aims.prod.svc",
                      "REMEDY_USER": "RMDYUSR"
                    },
                    "ingress": {
                      "enabled": true,
                      "isOpenshiftRoute": true
                    },
                    "name": "aims-mob-service",
                    "proxyContainer": {
                      "image": "quay.io/microsafe_app_ban/aims-mob-proxy-private@sha256:29b82de987b28c23f17d5f2176877c5aecea1a0ef182bb3a63612290ff469949",
                      "name": "aims-mob-service-proxy",
                      "resources": {
                        "limits": {
                          "cpu": "2000m",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "replicas": 2,
                    "secretName": "aims-mob-service-secrets",
                    "service": {
                      "enabled": true,
                      "port": 8085,
                      "targetPort": 8081
                    },
                    "serviceContainer": {
                      "image": "quay.io/microsafe_app_ban/aims-mob-service-private@sha256:f9586200106aa4de259f5cae2d22d13dc147a76dfd06e4f4bea09797071dad17",
                      "name": "aims-mob-service-main",
                      "resources": {
                        "limits": {
                          "cpu": "2500m",
                          "memory": "2048Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "128Mi"
                        }
                      }
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": false,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          createdAt: "2025-09-25T19:23:26Z"
          description: Este operador despliega los servicios AIMS Mobile.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/network-aware: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          instrumentation.opentelemetry.io/inject-nodejs: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: Valid Subscription Microsafe
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://quay.io/repository/microsafe_app_ban/operador-aims-mobile-private
          support: Microsafe
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AimsMobile
            name: aimsmobiles.cache.microsafe.com.mx
            version: v1
        description: Este operador despliega la aplicación AIMS Mobile.
        displayName: AIMS Mobile Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aims mobile
        - aims
        links:
        - name: AIMS Mobile Operator
          url: https://microsafe.com.mx/
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Microsafe
          url: https://microsafe.com.mx/
        relatedImages:
        - quay.io/microsafe_app_ban/aims-db-service-private@sha256:bcb4d8624626436526549427f9f169cff714b39cbcc530dd2b3470c6712abaf9
        - quay.io/microsafe_app_ban/aims-mob-admin-private@sha256:8e333f900885b0d3f6442c8ffcb43bd93540b8c38ad5bf17cbbd78929e0b1be0
        - quay.io/microsafe_app_ban/operador-aims-mobile-private@sha256:731f2d731c79ade02f4b6c39b380590833cacddfd795b703cec10ad55f730fad
        - quay.io/microsafe_app_ban/aims-mob-proxy-private@sha256:29b82de987b28c23f17d5f2176877c5aecea1a0ef182bb3a63612290ff469949
        - quay.io/microsafe_app_ban/aims-mob-service-private@sha256:f9586200106aa4de259f5cae2d22d13dc147a76dfd06e4f4bea09797071dad17
        version: 0.0.10
      entries:
      - name: operator-aims-mobile.v0.0.10
        version: 0.0.10
      - name: operator-aims-mobile.v0.0.9
        version: 0.0.9
      - name: operator-aims-mobile.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: operator-aims-mobile
    provider:
      name: Microsafe
      url: https://microsafe.com.mx/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzograph-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzograph-operator.v2.0.102
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzograph.clusters.cambridgesemantics.com/v2",
                "kind": "AnzoGraph",
                "metadata": {
                  "name": "azg01"
                },
                "spec": {
                  "db": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_data": "anzograph-data-grpc",
                            "app_mgmt": "anzograph-mgmt-grpc"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_data": "anzograph-data-grpc",
                              "app_mgmt": "anzograph-mgmt-grpc"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-db@sha256:8a5bd26551990929fdb776d6b421a9713cb84e03b19c18f81b625582f7f41af5",
                                "name": "db",
                                "resources": {
                                  "limits": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  },
                                  "requests": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  },
                  "deployFrontend": false,
                  "frontend": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_mgmt": "anzograph-frontend"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_mgmt": "anzograph-frontend"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-frontend@sha256:c1443bcdc8e998fc699f61980a665f1a083350adf2d3d1d55a7cb55e31ca50b3",
                                "name": "frontend",
                                "resources": {
                                  "limits": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-operator@sha256:a5b212558f748ddc8156541b546290320b656506e039a18903ad9fa6e638e732
          createdAt: "2021-08-11"
          description: kubernetes operator for AnzoGraph DB
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzograph-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzograph-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoGraph is the Schema for the anzographs API
            displayName: AnzoGraph Operator
            kind: AnzoGraph
            name: anzographs.anzograph.clusters.cambridgesemantics.com
            version: v2
        description: |-
          The AnzoGraph Operator provides the way to install and configure an AnzoGraph
          cluster on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v2.0.1/README_openshift_marketplace.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzograph/userdoc/ )

          ### Support

          We offer Support to our customers with the AnzoGraph db Enterprise Edition License [here]( https://customercenter.cambridgesemantics.com/ ). For AnzoGraph db Free Edition questions, get help from our Anzograph User Community at Stack Overflow. When submitting a question, include the tag 'anzograph'.
        displayName: AnzoGraph Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        - Database
        links:
        - name: Anzograph Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v2.0.1/README_openshift_marketplace.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-operator@sha256:a5b212558f748ddc8156541b546290320b656506e039a18903ad9fa6e638e732
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph@sha256:e3bdddf811b5ba69f388d023aea1ba538edd00e9f415c33ec555321d39129a36
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-db@sha256:8a5bd26551990929fdb776d6b421a9713cb84e03b19c18f81b625582f7f41af5
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-frontend@sha256:c1443bcdc8e998fc699f61980a665f1a083350adf2d3d1d55a7cb55e31ca50b3
        version: 2.0.1
      entries:
      - name: anzograph-operator.v2.0.102
        version: 2.0.1
      name: stable
    defaultChannel: stable
    packageName: anzograph-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fortanix Inc.
      provider-url: https://fortanix.com
    name: ccm-node-agent-dcap-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ccm-node-agent-dcap-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1",
                "kind": "CcmNodeAgent",
                "metadata": {
                  "name": "ccmnodeagent-sample"
                },
                "spec": {
                  "ccmJoinToken": {
                    "secretName": "ccm-join-token",
                    "token": ""
                  },
                  "fullnameOverride": "",
                  "image": {
                    "aesmd": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/aesmd-dcap",
                      "tag": "latest"
                    },
                    "ccmNodeAgent": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/ccm-node-agent-ubi7",
                      "tag": "latest"
                    },
                    "ccmNodeAgentInit": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "busybox",
                      "tag": "latest"
                    },
                    "multiPlatformRegistration": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/multi-platform-registration-ubi8",
                      "tag": "latest"
                    }
                  },
                  "log": {
                    "debug": false
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "ports": {
                    "containerPort": 9092,
                    "hostPort": 9092
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CCM Node Agent software enables registration of the compute
              nodes to Fortanix CCM when installed in the Openshift Cluster Nodes<U+200B>.
            displayName: CCM Node Agent
            kind: CcmNodeAgent
            name: ccmnodeagents.charts.operatorhub.io
            version: v1
        description: "CCM Node Agent software enables registration of the compute
          nodes to Fortanix CCM when installed in the Openshift Cluster Nodes​. \nIt
          also assists with application attestation and visibility for Fortanix CCM.\n\n##
          About this Operator\nThis Operator is based on a Helm chart for CCM Node
          Agent. It enrolls all the nodes in the openshift cluster to a particular
          CCM account. You have to provide the base64 encode of join token of CCM
          account while creating CcmNodeAgent resource.\n\n## Prerequisites\nAll the
          worker nodes in the cluster must have SGX capabilities, in-kernel driver
          installed and fsgsbase support.\n\n## Installation\n * Choose a namespace
          to install the Operator. \n Operator will be available in all namespaces.\n*
          Select Approval Strategy: Available options are Automatic or Manual. Automatic
          approval will automatically upgrade the running instance of your Operator
          without human intervention when a new version is available. In case of Manual
          approval, when a new version is available, OLM will create an update request.
          You must manually approve the request to update the Operator to a new version.\n*
          Generate your CCM Account Join Token: Please log in to https://ccm.fortanix.com,
          and in the Management Console tab, click the + ENROLL NODE button. In the
          ENROLL NODE screen, a Join Token will be generated in the text box for \"Get
          a join token to register an SGX compute node\". This Join Token is used
          by the compute node to authenticate itself.\n* Base64 encode the Join Token:\n```shell\necho
          -n <join-token> | openssl base64 -A\n```\n* Create an instance of CcmNodeAgent
          by setting the `spec.ccmJoinToken.Token` to the base64 encoded CCM account
          join token.\n* After CcmNodeAgent resource is installed successfully you
          should be able to see all the worker nodes of the cluster in your CCM account
          under Compute Node Tab.\n\n## Getting Started\nNow to run your application
          inside SGX enclaves on your cluster follow the [documentation](https://support.fortanix.com/hc/en-us/articles/360043527431-User-s-Guide-Add-and-Edit-an-Application).\n"
        displayName: CCM Node Agent DCAP Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dockerfile
        - Helm
        - CCM
        - Confidential Compute
        links:
        - name: CCM Node Agent DCAP Operator
          url: https://support.fortanix.com/hc/en-us/categories/360003107511-Confidential-Computing-Manager
        maintainers:
        - email: info@fortanix.com
          name: Fortanix Inc.
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Fortanix Inc.
          url: https://fortanix.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62575404000cb77fa8d2284c87b31e39e667a55b0fb47821263b23b3d1135a7e
        - docker.io/fortanix/ccm-node-agent-dcap-operator@sha256:4fd28bea26aee461eb9c1800cdf5f041b64ae3ea6652203f8ad9feb680aaefee
        version: 0.0.2
      entries:
      - name: ccm-node-agent-dcap-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ccm-node-agent-dcap-operator
    provider:
      name: Fortanix Inc.
      url: https://fortanix.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Infinidat
      provider-url: https://infinidat.com
    name: infinibox-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infinibox-operator-certified.v2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csidriver.infinidat.com/v1alpha1",
                "kind": "InfiniboxCsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infinibox-operator-certified",
                    "app.kubernetes.io/instance": "infiniboxcsidriver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infiniboxcsidriver",
                    "app.kubernetes.io/part-of": "infinibox-operator-certified"
                  },
                  "name": "infiniboxcsidriver-sample"
                },
                "spec": {
                  "Ibox_Creds": [
                    {
                      "SecretName": "infinibox-creds",
                      "hostname": "ibox0000",
                      "inbound_secret": "0.0000000000000",
                      "inbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonout",
                      "outbound_secret": "0.0000000000001",
                      "outbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonin",
                      "password": "my-pool-admin-passwd",
                      "username": "my-pool-admin"
                    }
                  ],
                  "Infinibox_Protocol_Secrets": [
                    {
                      "SecretName": "infinibox-protocol-secret",
                      "iscsi": {
                        "network_space": "default_iscsi_space",
                        "useCHAP": "none"
                      },
                      "nfs": {
                        "network_space": "default_nas_space",
                        "nfs_export_permissions": ""
                      },
                      "nvme": {
                        "network_space": "default_nvme_space"
                      },
                      "preferred_auto_order": "fc,nvme,iscsi",
                      "storage_protocol": "fc"
                    }
                  ],
                  "autoUpdate": false,
                  "cleanupNFSPerms": false,
                  "controllerserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "controllerservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "createEvents": true,
                  "csiDriverName": "infinibox-csi-driver",
                  "csiDriverVersion": "v2.23.0",
                  "e2etesting": false,
                  "images": {
                    "attachersidecar": "registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901",
                    "attachersidecar_pull_policy": "IfNotPresent",
                    "csidriver": "registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61",
                    "csidriver_pull_policy": "Always",
                    "livenesssidecar": "registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc",
                    "livenesssidecar_pull_policy": "IfNotPresent",
                    "provisionersidecar": "registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2",
                    "provisionersidecar_pull_policy": "IfNotPresent",
                    "registrarsidecar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24",
                    "registrarsidecar_pull_policy": "IfNotPresent",
                    "resizersidecar": "registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d",
                    "resizersidecar_pull_policy": "IfNotPresent",
                    "snapshottersidecar": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a",
                    "snapshottersidecar_pull_policy": "IfNotPresent"
                  },
                  "instanceCount": 1,
                  "logLevel": "debug",
                  "nodeSelector": {
                    "kubernetes.io/os": "linux"
                  },
                  "nodeserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "nodeservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "protocolSecret": "",
                  "removeDomainName": false,
                  "replicaCount": 1,
                  "skipCredentialsCreation": false,
                  "volumeNamePrefix": "ibox"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
          createdAt: "2025-09-27T16:12:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2.23.0
          operatorframework.io/suggested-namespace: infinidat-csi
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Infinidat/infinibox-csi-driver
          support: Infinidat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Iboxreplica
            name: iboxreplicas.csidriver.infinidat.com
            version: v1
          - kind: InfiniboxCsiDriver
            name: infiniboxcsidrivers.csidriver.infinidat.com
            version: v1alpha1
        description: "Infinidat InfiniBox Container Storage Interface (CSI) Driver
          is a\n    CNCF-compliant Kubernetes integration for InfiniBox storage systems,
          offering\n    advanced enterprise functionality for petabyte-scale Kubernetes
          deployments including\n    Red Hat OpenShift.\n \n## Features and Benefits\n
          \ \n* **Multi-protocol flexibility** - manage\n    Kubernetes Persistent
          Volumes attached via block and file protocols,\n    including Fibre Channel,
          iSCSI, and NFS, with all Kubernetes PV access modes\n    \n* **Multi-petabyte
          scalability** - support hundreds of thousands of PVs per InfiniBox\n    system
          and control multiple InfiniBox arrays within a single Kubernetes cluster\n
          \   \\ \n* **Advanced enterprise features** - manage native InfiniBox snapshots
          and\n    clones, including restoring from snapshots, and import PVs created
          outside of\n    InfiniBox CSI Driver\n  \n## Required Parameters\n  \n*
          `hostname` - IP address\n    or hostname of the InfiniBox management interface\n*
          `username` / `password` -\n    InfiniBox credentials\n* `SecretName` - secret
          name, to be used in the StorageClass\n    to define a specific InfiniBox
          for persistent volumes\n\n## Optional Parameters\n*\n    `inbound_user`
          / `inbound_secret` / `outbound_user` / `outbound_secret` - credentials \n
          \   for iSCSI CHAP authentication\n\n    \n## Installation Instructions\n
          \n1. Create *infinidat-csi* namespace to install operator into.\n    \n2.
          Delete the following *clusterrolebindings*, if they exist, before installing
          operator\n     * infinidat-csi-operator-infinidat-csi-attacher\n     * infinidat-csi-operator-infinidat-csi-controller
          \                    \n     * infinidat-csi-operator-infinidat-csi-driver
          \                    \n     * infinidat-csi-operator-infinidat-csi-node
          \                              \n     * infinidat-csi-operator-infinidat-csi-provisioner
          \                       \n     * infinidat-csi-operator-infinidat-csi-resizer
          \          \n     * infinidat-csi-operator-infinidat-csi-snapshotter \n
          \   \n3. Install Infinidat Infinibox CSI Driver - Operator (this) into the
          *infinidat-csi* namespace.\n    \n4. Apply the following Security Context
          Constraints to the created service accounts in the *infinidat-csi* \n    namespace\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    \n5. Recreate or edit any existing storage classes, pointing to the
          new namespace\n\n     \n6. If you choose a different namespace other than
          *infinidat-csi* you will need to edit the clusterrolebindings listed above
          to specify your custom namespace\n\n     \n7. Install an instance of Infinidat
          Driver into the *infinidat-csi* namespace"
        displayName: InfiniBox CSI Driver - Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - csi
        - nas
        - san
        - storage
        - infinidat
        - infinibox
        links:
        - name: Infinidat
          url: https://infinidat.com
        maintainers:
        - email: info@infinidat.com
          name: Infinidat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Infinidat
          url: https://infinidat.com
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified@sha256:42ff34a019bc67076515e89735b4bb9921fbd4610c127ac8b61e1fa6a6a08194
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24
        version: 2.23.0
      entries:
      - name: infinibox-operator-certified.v2.23.0
        version: 2.23.0
      - name: infinibox-operator-certified.v2.22.0
        version: 2.22.0
      - name: infinibox-operator-certified.v2.21.0
        version: 2.21.0
      name: alpha
    - currentCSV: infinibox-operator-certified.v2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csidriver.infinidat.com/v1alpha1",
                "kind": "InfiniboxCsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infinibox-operator-certified",
                    "app.kubernetes.io/instance": "infiniboxcsidriver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infiniboxcsidriver",
                    "app.kubernetes.io/part-of": "infinibox-operator-certified"
                  },
                  "name": "infiniboxcsidriver-sample"
                },
                "spec": {
                  "Ibox_Creds": [
                    {
                      "SecretName": "infinibox-creds",
                      "hostname": "ibox0000",
                      "inbound_secret": "0.0000000000000",
                      "inbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonout",
                      "outbound_secret": "0.0000000000001",
                      "outbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonin",
                      "password": "my-pool-admin-passwd",
                      "username": "my-pool-admin"
                    }
                  ],
                  "Infinibox_Protocol_Secrets": [
                    {
                      "SecretName": "infinibox-protocol-secret",
                      "iscsi": {
                        "network_space": "default_iscsi_space",
                        "useCHAP": "none"
                      },
                      "nfs": {
                        "network_space": "default_nas_space",
                        "nfs_export_permissions": ""
                      },
                      "nvme": {
                        "network_space": "default_nvme_space"
                      },
                      "preferred_auto_order": "fc,nvme,iscsi",
                      "storage_protocol": "fc"
                    }
                  ],
                  "autoUpdate": false,
                  "cleanupNFSPerms": false,
                  "controllerserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "controllerservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "createEvents": true,
                  "csiDriverName": "infinibox-csi-driver",
                  "csiDriverVersion": "v2.23.0",
                  "e2etesting": false,
                  "images": {
                    "attachersidecar": "registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901",
                    "attachersidecar_pull_policy": "IfNotPresent",
                    "csidriver": "registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61",
                    "csidriver_pull_policy": "Always",
                    "livenesssidecar": "registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc",
                    "livenesssidecar_pull_policy": "IfNotPresent",
                    "provisionersidecar": "registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2",
                    "provisionersidecar_pull_policy": "IfNotPresent",
                    "registrarsidecar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24",
                    "registrarsidecar_pull_policy": "IfNotPresent",
                    "resizersidecar": "registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d",
                    "resizersidecar_pull_policy": "IfNotPresent",
                    "snapshottersidecar": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a",
                    "snapshottersidecar_pull_policy": "IfNotPresent"
                  },
                  "instanceCount": 1,
                  "logLevel": "debug",
                  "nodeSelector": {
                    "kubernetes.io/os": "linux"
                  },
                  "nodeserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "nodeservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "protocolSecret": "",
                  "removeDomainName": false,
                  "replicaCount": 1,
                  "skipCredentialsCreation": false,
                  "volumeNamePrefix": "ibox"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
          createdAt: "2025-09-27T16:12:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2.23.0
          operatorframework.io/suggested-namespace: infinidat-csi
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Infinidat/infinibox-csi-driver
          support: Infinidat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Iboxreplica
            name: iboxreplicas.csidriver.infinidat.com
            version: v1
          - kind: InfiniboxCsiDriver
            name: infiniboxcsidrivers.csidriver.infinidat.com
            version: v1alpha1
        description: "Infinidat InfiniBox Container Storage Interface (CSI) Driver
          is a\n    CNCF-compliant Kubernetes integration for InfiniBox storage systems,
          offering\n    advanced enterprise functionality for petabyte-scale Kubernetes
          deployments including\n    Red Hat OpenShift.\n \n## Features and Benefits\n
          \ \n* **Multi-protocol flexibility** - manage\n    Kubernetes Persistent
          Volumes attached via block and file protocols,\n    including Fibre Channel,
          iSCSI, and NFS, with all Kubernetes PV access modes\n    \n* **Multi-petabyte
          scalability** - support hundreds of thousands of PVs per InfiniBox\n    system
          and control multiple InfiniBox arrays within a single Kubernetes cluster\n
          \   \\ \n* **Advanced enterprise features** - manage native InfiniBox snapshots
          and\n    clones, including restoring from snapshots, and import PVs created
          outside of\n    InfiniBox CSI Driver\n  \n## Required Parameters\n  \n*
          `hostname` - IP address\n    or hostname of the InfiniBox management interface\n*
          `username` / `password` -\n    InfiniBox credentials\n* `SecretName` - secret
          name, to be used in the StorageClass\n    to define a specific InfiniBox
          for persistent volumes\n\n## Optional Parameters\n*\n    `inbound_user`
          / `inbound_secret` / `outbound_user` / `outbound_secret` - credentials \n
          \   for iSCSI CHAP authentication\n\n    \n## Installation Instructions\n
          \n1. Create *infinidat-csi* namespace to install operator into.\n    \n2.
          Delete the following *clusterrolebindings*, if they exist, before installing
          operator\n     * infinidat-csi-operator-infinidat-csi-attacher\n     * infinidat-csi-operator-infinidat-csi-controller
          \                    \n     * infinidat-csi-operator-infinidat-csi-driver
          \                    \n     * infinidat-csi-operator-infinidat-csi-node
          \                              \n     * infinidat-csi-operator-infinidat-csi-provisioner
          \                       \n     * infinidat-csi-operator-infinidat-csi-resizer
          \          \n     * infinidat-csi-operator-infinidat-csi-snapshotter \n
          \   \n3. Install Infinidat Infinibox CSI Driver - Operator (this) into the
          *infinidat-csi* namespace.\n    \n4. Apply the following Security Context
          Constraints to the created service accounts in the *infinidat-csi* \n    namespace\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    \n5. Recreate or edit any existing storage classes, pointing to the
          new namespace\n\n     \n6. If you choose a different namespace other than
          *infinidat-csi* you will need to edit the clusterrolebindings listed above
          to specify your custom namespace\n\n     \n7. Install an instance of Infinidat
          Driver into the *infinidat-csi* namespace"
        displayName: InfiniBox CSI Driver - Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - csi
        - nas
        - san
        - storage
        - infinidat
        - infinibox
        links:
        - name: Infinidat
          url: https://infinidat.com
        maintainers:
        - email: info@infinidat.com
          name: Infinidat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Infinidat
          url: https://infinidat.com
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified@sha256:42ff34a019bc67076515e89735b4bb9921fbd4610c127ac8b61e1fa6a6a08194
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24
        version: 2.23.0
      entries:
      - name: infinibox-operator-certified.v2.23.0
        version: 2.23.0
      - name: infinibox-operator-certified.v2.22.0
        version: 2.22.0
      - name: infinibox-operator-certified.v2.21.0
        version: 2.21.0
      name: stable
    defaultChannel: stable
    packageName: infinibox-operator-certified
    provider:
      name: Infinidat
      url: https://infinidat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Shipwright Contributors
      provider-url: https://shipwright.io
    name: shipwright-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: shipwright-operator.v0.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "shipwright-build"
                },
                "spec": {
                  "targetNamespace": "shipwright-build"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: ghcr.io/shipwright-io/operator/operator:0.13.0@sha256:12cd5f936043c96f31cf571d7a4f8b6f2e240852cd9ad1a5423dbab0c208b542
          description: Shipwright is a framework for building container images on
            Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/shipwright-io/operator
          support: The Shipwright Contributors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
        description: |
          Shipwright is a framework for building container images on Kubernetes.

          Read more: [https://shipwright.io](https://shipwright.io)

          ## Usage

          To deploy and manage [Shipwright Builds](https://github.com/shipwright-io/build) in your cluster,
          first make sure this operator is installed and running on your cluster.

          Next, create the following:

          ```yaml
          ---
          apiVersion: operator.shipwright.io/v1alpha1
          kind: ShipwrightBuild
          metadata:
            name: shipwright-operator
          spec:
            targetNamespace: shipwright-build
          ```

          The operator will deploy Shipwright Builds in the provided `targetNamespace`.
          When `.spec.targetNamespace` is not set, the namespace will default to `shipwright-build`.
          Refer to the [ShipwrightBuild documentation](https://github.com/shipwright-io/operator/blob/main/docs/shipwrightbuild.md) for more information about this custom resource.
        displayName: Shipwright Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://shipwright.io
        - name: Shipwright Builds
          url: https://github.com/shipwright-io/build
        maintainers:
        - email: shipwright-dev@lists.shipwright.io
          name: The Shipwright Contributors
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Shipwright Contributors
          url: https://shipwright.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - ghcr.io/shipwright-io/operator/operator:0.13.0@sha256:12cd5f936043c96f31cf571d7a4f8b6f2e240852cd9ad1a5423dbab0c208b542
        version: 0.13.0
      entries:
      - name: shipwright-operator.v0.13.0
        version: 0.13.0
      - name: shipwright-operator.v0.9.0
        version: 0.9.0
      name: alpha
    defaultChannel: alpha
    packageName: shipwright-operator
    provider:
      name: The Shipwright Contributors
      url: https://shipwright.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ThinkParQ
      provider-url: https://www.thinkparq.com
    name: beegfs-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: beegfs-csi-driver-operator.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "beegfs.csi.netapp.com/v1",
                "kind": "BeegfsDriver",
                "metadata": {
                  "name": "csi-beegfs-cr"
                },
                "spec": {
                  "nodeAffinityControllerService": {
                    "preferredDuringSchedulingIgnoredDuringExecution": [
                      {
                        "preference": {
                          "matchExpressions": [
                            {
                              "key": "node-role.kubernetes.io/master",
                              "operator": "Exists"
                            }
                          ]
                        },
                        "weight": 50
                      }
                    ],
                    "requiredDuringSchedulingIgnoredDuringExecution": {
                      "nodeSelectorTerms": [
                        {
                          "matchExpressions": [
                            {
                              "key": "node.openshift.io/os_id",
                              "operator": "NotIn",
                              "values": [
                                "rhcos"
                              ]
                            }
                          ]
                        }
                      ]
                    }
                  },
                  "nodeAffinityNodeService": {
                    "requiredDuringSchedulingIgnoredDuringExecution": {
                      "nodeSelectorTerms": [
                        {
                          "matchExpressions": [
                            {
                              "key": "node.openshift.io/os_id",
                              "operator": "NotIn",
                              "values": [
                                "rhcos"
                              ]
                            }
                          ]
                        }
                      ]
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: ghcr.io/thinkparq/beegfs-csi-driver-operator:v1.7.0
          createdAt: 2021-06-30:03:11:00
          description: Install and manage the BeeGFS CSI driver
          operatorframework.io/suggested-namespace: beegfs-csi
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ThinkParQ/beegfs-csi-driver
          support: ThinkParQ
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys the BeeGFS CSI driver
            displayName: BeeGFS Driver
            kind: BeegfsDriver
            name: beegfsdrivers.beegfs.csi.netapp.com
            version: v1
        description: |
          The BeeGFS Container Storage Interface (CSI) driver provides high performing and scalable storage for workloads
          running in container orchestrators like Kubernetes. It allows containers to access existing datasets or request
          on-demand ephemeral or persistent high speed storage backed by BeeGFS parallel file systems.

          This operator eases the installation and management of the BeeGFS CSI driver in an OpenShift or other Operator
          Lifecycle Manager enabled cluster.

          ### Documentation

          Compatibility information and a quick start guide for the BeeGFS CSI driver itself are available on our [main GitHub
          page](https://www.github.com/ThinkParQ/beegfs-csi-driver). Additional detailed documentation for the driver is
          available in the [docs directory](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/docs) and for the
          operator is available in the [operator
          README](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/operator/README.md).

          ### Important Notes

          The BeeGFS client packages and kernel module must be pre-installed on all nodes that will run the driver. See the
          driver ["Getting Started" guide](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/README.md#getting-started)
          for additional information.

          The BeeGFS CSI driver is only supported on nodes running a [BeeGFS supported Linux
          distribution](https://doc.beegfs.io/latest/release_notes.html#supported-linux-distributions-and-kernels). Red Hat
          CoreOS (RHCOS), the default distribution in OpenShift environments, is NOT supported. However, Red Hat Enterprise
          Linux (RHEL) nodes can be added to an OpenShift cluster and RHEL is supported for running BeeGFS. By default, the
          operator will install the driver with a node affinity that ensures it does not run on RHCOS nodes. See the driver
          [compatibility
          matrix](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/README.md#interoperability-and-csi-feature-matrix)
          for additional information.

          We have recently published experimental manifests that enable the BeeGFS CSI driver on OpenShift nodes running
          RHCOS. These manifests are not intended for production use. See [this
          README](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/deploy/openshift-beegfs-client/README.md) for more
          details.

          All driver configuration can be specified through the BeegfsDrivers API (in a BeegfsDriver Custom Resource)
          except for BeeGFS connAuth information, which must be placed separately into a Kubernetes Secret. See the [ConnAuth
          Configuration
          section](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/operator/README.md#connauth-configuration) of the
          operator documentation for additional information.

          Starting with version 1.4.0 of the BeeGFS CSI driver the driver and operator images will be signed using the Cosign
          tool. Refer to the [BeeGFS CSI driver operator
          README](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/operator/README.md)
          for details on how to verify the operator and driver image signatures.
        displayName: BeeGFS CSI Driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CSI
        - BeeGFS
        - ThinkParQ
        - HPC
        - AI
        - High-Performance Computing
        links:
        - name: BeeGFS introduction
          url: https://www.beegfs.io/c/
        - name: BeeGFS documentation
          url: https://doc.beegfs.io/latest/index.html
        - name: BeeGFS CSI driver source code
          url: https://github.com/ThinkParQ/beegfs-csi-driver
        - name: BeeGFS blog posts
          url: https://www.beegfs.io/c/blog/
        maintainers:
        - email: csi-driver@beegfs.io
          name: ThinkParQ
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: ThinkParQ
          url: https://www.thinkparq.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - ghcr.io/thinkparq/beegfs-csi-driver-operator:v1.7.0
        version: 1.7.0
      entries:
      - name: beegfs-csi-driver-operator.v1.7.0
        version: 1.7.0
      - name: beegfs-csi-driver-operator.v1.6.0
        version: 1.6.0
      - name: beegfs-csi-driver-operator.v1.5.0
        version: 1.5.0
      - name: beegfs-csi-driver-operator.v1.4.0
        version: 1.4.0
      - name: beegfs-csi-driver-operator.v1.3.0
        version: 1.3.0
      - name: beegfs-csi-driver-operator.v1.2.2
        version: 1.2.2
      - name: beegfs-csi-driver-operator.v1.2.1
        version: 1.2.1
      - name: beegfs-csi-driver-operator.v1.2.0
        version: 1.2.0
      name: stable
    defaultChannel: stable
    packageName: beegfs-csi-driver-operator
    provider:
      name: ThinkParQ
      url: https://www.thinkparq.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stackwatch
      provider-url: www.kubecost.com
    name: kubecost
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubecost.v2.2.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kubecost.com/v1alpha1",
                "kind": "CostAnalyzer",
                "metadata": {
                  "name": "costanalyzer-sample"
                },
                "spec": {
                  "global": {
                    "prometheus": {
                      "enabled": true,
                      "fqdn": "http://cost-analyzer-prometheus-server.default.svc"
                    },
                    "grafana": {
                      "enabled": false,
                      "domainName": "cost-analyzer-grafana.default.svc",
                      "scheme": "http",
                      "proxy": false
                    },
                    "gcpstore": {
                      "enabled": false
                    },
                    "gmp": {
                      "enabled": false,
                      "prometheusServerEndpoint": "http://localhost:8085/",
                      "gmpProxy": {
                        "enabled": false,
                        "image": "gke.gcr.io/prometheus-engine/frontend:v0.4.1-gke.0",
                        "imagePullPolicy": "Always",
                        "name": "gmp-proxy",
                        "port": 8085,
                        "projectId": "YOUR_PROJECT_ID"
                      }
                    },
                    "amp": {
                      "enabled": false,
                      "prometheusServerEndpoint": "http://localhost:8005/workspaces/<workspaceId>/",
                      "remoteWriteService": "https://aps-workspaces.us-west-2.amazonaws.com/workspaces/<workspaceId>/api/v1/remote_write",
                      "sigv4": {
                        "region": "us-west-2"
                      }
                    },
                    "mimirProxy": {
                      "enabled": false,
                      "name": "mimir-proxy",
                      "image": "nginxinc/nginx-unprivileged",
                      "port": 8085,
                      "mimirEndpoint": "$mimir_endpoint",
                      "orgIdentifier": "$your_tenant_ID"
                    },
                    "notifications": {
                      "alertmanager": {
                        "enabled": false,
                        "fqdn": "http://cost-analyzer-prometheus-server.default.svc"
                      }
                    },
                    "savedReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Saved Report 0",
                          "window": "today",
                          "aggregateBy": "namespace",
                          "chartDisplay": "category",
                          "idle": "separate",
                          "rate": "cumulative",
                          "accumulate": false,
                          "filters": [
                            {
                              "key": "cluster",
                              "operator": ":",
                              "value": "dev"
                            }
                          ]
                        },
                        {
                          "title": "Example Saved Report 1",
                          "window": "month",
                          "aggregateBy": "controllerKind",
                          "chartDisplay": "category",
                          "idle": "share",
                          "rate": "monthly",
                          "accumulate": false,
                          "filters": [
                            {
                              "key": "namespace",
                              "operator": "!:",
                              "value": "kubecost"
                            }
                          ]
                        },
                        {
                          "title": "Example Saved Report 2",
                          "window": "2020-11-11T00:00:00Z,2020-12-09T23:59:59Z",
                          "aggregateBy": "service",
                          "chartDisplay": "category",
                          "idle": "hide",
                          "rate": "daily",
                          "accumulate": true,
                          "filters": []
                        }
                      ]
                    },
                    "assetReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Asset Report 0",
                          "window": "today",
                          "aggregateBy": "type",
                          "accumulate": false,
                          "filters": [
                            {
                              "property": "cluster",
                              "value": "cluster-one"
                            }
                          ]
                        }
                      ]
                    },
                    "advancedReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Advanced Report 0",
                          "window": "7d",
                          "aggregateBy": "namespace",
                          "filters": [
                            {
                              "key": "cluster",
                              "operator": ":",
                              "value": "dev"
                            }
                          ],
                          "cloudBreakdown": "service",
                          "cloudJoin": "label:kubernetes_namespace"
                        }
                      ]
                    },
                    "cloudCostReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Cloud Cost Report 0",
                          "window": "today",
                          "aggregateBy": "service",
                          "accumulate": false
                        }
                      ]
                    },
                    "podAnnotations": {},
                    "additionalLabels": {},
                    "securityContext": {
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      },
                      "fsGroup": 1001,
                      "runAsGroup": 1001,
                      "runAsUser": 1001,
                      "fsGroupChangePolicy": "OnRootMismatch"
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "privileged": false,
                      "readOnlyRootFilesystem": true,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "platforms": {
                      "openshift": {
                        "enabled": true,
                        "route": {
                          "enabled": false,
                          "annotations": {}
                        },
                        "scc": {
                          "nodeExporter": false,
                          "networkCosts": false
                        },
                        "securityContext": {
                          "runAsNonRoot": true,
                          "seccompProfile": {
                            "type": "RuntimeDefault"
                          }
                        }
                      },
                      "cicd": {
                        "enabled": true,
                        "skipSanityChecks": true
                      }
                    }
                  },
                  "upgrade": {
                    "toV2": false
                  },
                  "kubecostToken": null,
                  "pricingCsv": {
                    "enabled": false,
                    "location": {
                      "provider": "AWS",
                      "region": "us-east-1",
                      "URI": "s3://kc-csv-test/pricing_schema.csv",
                      "csvAccessCredentials": "pricing-schema-access-secret"
                    }
                  },
                  "saml": {
                    "enabled": false,
                    "rbac": {
                      "enabled": false
                    }
                  },
                  "oidc": {
                    "enabled": false,
                    "clientID": "",
                    "clientSecret": "",
                    "existingCustomSecret": {
                      "enabled": false,
                      "name": ""
                    },
                    "skipOnlineTokenValidation": false,
                    "rbac": {
                      "enabled": false
                    }
                  },
                  "systemProxy": {
                    "enabled": false,
                    "httpProxyUrl": "",
                    "httpsProxyUrl": "",
                    "noProxy": ""
                  },
                  "kubecostFrontend": {
                    "enabled": true,
                    "deployMethod": "singlepod",
                    "haReplicas": 2,
                    "image": "gcr.io/kubecost1/frontend",
                    "imagePullPolicy": "Always",
                    "resources": {
                      "requests": {
                        "cpu": "10m",
                        "memory": "55Mi"
                      }
                    },
                    "deploymentStrategy": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 1,
                      "periodSeconds": 5,
                      "failureThreshold": 6
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 1,
                      "periodSeconds": 5,
                      "failureThreshold": 6
                    },
                    "ipv6": {
                      "enabled": true
                    },
                    "useDefaultFqdn": false
                  },
                  "kubecostMetrics": {
                    "exporter": {
                      "enabled": false,
                      "port": 9005,
                      "prometheusScrape": true,
                      "resources": {},
                      "tolerations": [],
                      "affinity": {},
                      "service": {
                        "annotations": {}
                      },
                      "serviceMonitor": {
                        "enabled": false,
                        "additionalLabels": {},
                        "metricRelabelings": [],
                        "relabelings": []
                      },
                      "priorityClassName": "",
                      "additionalLabels": {},
                      "nodeSelector": {},
                      "extraArgs": []
                    }
                  },
                  "sigV4Proxy": {
                    "image": "public.ecr.aws/aws-observability/aws-sigv4-proxy:latest",
                    "imagePullPolicy": "Always",
                    "name": "aps",
                    "port": 8005,
                    "region": "us-west-2",
                    "host": "aps-workspaces.us-west-2.amazonaws.com",
                    "extraEnv": null,
                    "resources": {}
                  },
                  "kubecostModel": {
                    "image": "gcr.io/kubecost1/cost-model",
                    "imagePullPolicy": "Always",
                    "outOfClusterPromMetricsEnabled": false,
                    "warmCache": false,
                    "etl": true,
                    "etlFileStoreEnabled": true,
                    "etlDailyStoreDurationDays": 91,
                    "etlHourlyStoreDurationHours": 49,
                    "etlWeeklyStoreDurationWeeks": 53,
                    "etlReadOnlyMode": false,
                    "plugins": {
                      "enabled": false,
                      "install": {
                        "enabled": false,
                        "fullImageName": "curlimages/curl:latest",
                        "securityContext": {
                          "allowPrivilegeEscalation": false,
                          "seccompProfile": {
                            "type": "RuntimeDefault"
                          },
                          "capabilities": {
                            "drop": [
                              "ALL"
                            ]
                          },
                          "readOnlyRootFilesystem": true,
                          "runAsNonRoot": true,
                          "runAsUser": 1001
                        }
                      },
                      "folder": "/opt/opencost/plugin",
                      "enabledPlugins": [],
                      "configSecret": "kubecost-plugin-secret"
                    },
                    "cloudCost": {
                      "labelList": {
                        "IsIncludeList": false,
                        "labels": ""
                      },
                      "topNItems": 1000
                    },
                    "allocation": null,
                    "containerStatsEnabled": true,
                    "maxQueryConcurrency": 5,
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "55Mi"
                      }
                    },
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "extraArgs": [],
                    "ingress": {
                      "enabled": false,
                      "labels": null,
                      "annotations": null,
                      "paths": [
                        "/"
                      ],
                      "pathType": "ImplementationSpecific",
                      "hosts": [
                        "cost-analyzer-model.local"
                      ],
                      "tls": []
                    },
                    "utcOffset": "+00:00",
                    "extraPorts": []
                  },
                  "etlUtils": {
                    "enabled": false,
                    "fullImageName": null,
                    "resources": {},
                    "env": {},
                    "nodeSelector": {},
                    "tolerations": {},
                    "affinity": {}
                  },
                  "ingress": {
                    "enabled": false,
                    "labels": null,
                    "annotations": null,
                    "paths": [
                      "/"
                    ],
                    "pathType": "ImplementationSpecific",
                    "hosts": [
                      "cost-analyzer.local"
                    ],
                    "tls": []
                  },
                  "nodeSelector": {},
                  "tolerations": [],
                  "affinity": {},
                  "topologySpreadConstraints": [],
                  "priority": {
                    "enabled": false,
                    "name": ""
                  },
                  "networkPolicy": {
                    "enabled": false,
                    "denyEgress": true,
                    "sameNamespace": true,
                    "costAnalyzer": {
                      "enabled": false,
                      "annotations": {},
                      "additionalLabels": {}
                    }
                  },
                  "extraVolumes": [],
                  "extraVolumeMounts": [],
                  "persistentVolume": {
                    "size": "32Gi",
                    "dbSize": "32.0Gi",
                    "enabled": true,
                    "labels": {},
                    "annotations": {},
                    "dbPVEnabled": false
                  },
                  "service": {
                    "type": "ClusterIP",
                    "port": 9090,
                    "targetPort": 9090,
                    "nodePort": {},
                    "labels": {},
                    "annotations": {},
                    "sessionAffinity": {
                      "enabled": false,
                      "timeoutSeconds": 10800
                    }
                  },
                  "prometheus": {
                    "rbac": {
                      "create": true
                    },
                    "serviceAccounts": {
                      "alertmanager": {
                        "create": true,
                        "name": null
                      },
                      "nodeExporter": {
                        "create": true,
                        "name": null
                      },
                      "pushgateway": {
                        "create": true,
                        "name": null
                      },
                      "server": {
                        "create": true,
                        "name": null,
                        "annotations": {}
                      }
                    },
                    "imagePullSecrets": null,
                    "extraScrapeConfigs": "- job_name: kubecost\n  honor_labels: true\n  scrape_interval: 1m\n  scrape_timeout: 60s\n  metrics_path: /metrics\n  scheme: http\n  dns_sd_configs:\n  - names:\n    - {{ template \"cost-analyzer.serviceName\" . }}\n    type: 'A'\n    port: 9003\n- job_name: kubecost-networking\n  kubernetes_sd_configs:\n    - role: pod\n  relabel_configs:\n  # Scrape only the the targets matching the following metadata\n    - source_labels: [__meta_kubernetes_pod_label_app_kubernetes_io_instance]\n      action: keep\n      regex:  kubecost\n    - source_labels: [__meta_kubernetes_pod_label_app_kubernetes_io_name]\n      action: keep\n      regex:  network-costs\n",
                    "server": {
                      "enabled": true,
                      "name": "server",
                      "sidecarContainers": null,
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "image": {
                        "repository": "quay.io/prometheus/prometheus",
                        "tag": "v2.50.1",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "prefixURL": "",
                      "baseURL": "",
                      "env": [],
                      "extraFlags": [
                        "web.enable-lifecycle"
                      ],
                      "configPath": "/etc/config/prometheus.yml",
                      "global": {
                        "scrape_interval": "1m",
                        "scrape_timeout": "60s",
                        "evaluation_interval": "1m",
                        "external_labels": {
                          "cluster_id": "cluster-one"
                        }
                      },
                      "remoteWrite": {},
                      "remoteRead": {},
                      "extraArgs": {
                        "query.max-concurrency": 1,
                        "query.max-samples": 100000000
                      },
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "extraHostPathMounts": [],
                      "extraConfigmapMounts": [],
                      "extraSecretMounts": [],
                      "configMapOverrideName": "",
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "extraLabels": {},
                        "hosts": [],
                        "pathType": "Prefix",
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "affinity": {},
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "32Gi",
                        "subPath": ""
                      },
                      "emptyDir": {
                        "sizeLimit": ""
                      },
                      "podAnnotations": {},
                      "deploymentAnnotations": {},
                      "podLabels": {},
                      "alertmanagers": [],
                      "replicaCount": 1,
                      "statefulSet": {
                        "enabled": false,
                        "annotations": {},
                        "labels": {},
                        "podManagementPolicy": "OrderedReady",
                        "headless": {
                          "annotations": {},
                          "labels": {},
                          "servicePort": 80
                        }
                      },
                      "readinessProbeInitialDelay": 5,
                      "readinessProbeTimeout": 3,
                      "readinessProbeFailureThreshold": 3,
                      "readinessProbeSuccessThreshold": 1,
                      "livenessProbeInitialDelay": 5,
                      "livenessProbeTimeout": 3,
                      "livenessProbeFailureThreshold": 3,
                      "livenessProbeSuccessThreshold": 1,
                      "resources": {},
                      "verticalAutoscaler": {
                        "enabled": false
                      },
                      "securityContext": {},
                      "containerSecurityContext": {},
                      "service": {
                        "annotations": {},
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 80,
                        "sessionAffinity": "None",
                        "type": "ClusterIP",
                        "gRPC": {
                          "enabled": false,
                          "servicePort": 10901
                        },
                        "statefulsetReplica": {
                          "enabled": false,
                          "replica": 0
                        }
                      },
                      "terminationGracePeriodSeconds": 300,
                      "retention": "97h"
                    },
                    "alertmanager": {
                      "enabled": false,
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "name": "alertmanager",
                      "image": {
                        "repository": "quay.io/prometheus/alertmanager",
                        "tag": "v0.27.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "extraArgs": {},
                      "prefixURL": "",
                      "baseURL": "http://localhost:9093",
                      "extraEnv": {},
                      "extraSecretMounts": [],
                      "configMapOverrideName": "",
                      "configFromSecret": "",
                      "configFileName": "alertmanager.yml",
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "extraLabels": {},
                        "hosts": [],
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "affinity": {},
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "2Gi",
                        "subPath": ""
                      },
                      "podAnnotations": {},
                      "podLabels": {},
                      "replicaCount": 1,
                      "statefulSet": {
                        "enabled": false,
                        "podManagementPolicy": "OrderedReady",
                        "headless": {
                          "annotations": {},
                          "labels": {},
                          "servicePort": 80
                        }
                      },
                      "resources": {},
                      "securityContext": {
                        "runAsUser": 1001,
                        "runAsNonRoot": true,
                        "runAsGroup": 1001,
                        "fsGroup": 1001
                      },
                      "service": {
                        "annotations": {},
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 80,
                        "sessionAffinity": "None",
                        "type": "ClusterIP"
                      }
                    },
                    "alertmanagerFiles": {
                      "alertmanager.yml": {
                        "global": {},
                        "receivers": [
                          {
                            "name": "default-receiver"
                          }
                        ],
                        "route": {
                          "group_wait": "10s",
                          "group_interval": "5m",
                          "receiver": "default-receiver",
                          "repeat_interval": "3h"
                        }
                      }
                    },
                    "configmapReload": {
                      "prometheus": {
                        "enabled": false,
                        "name": "configmap-reload",
                        "image": {
                          "repository": "quay.io/prometheus-operator/prometheus-config-reloader",
                          "tag": "v0.72.0",
                          "pullPolicy": "IfNotPresent"
                        },
                        "extraArgs": {},
                        "extraVolumeDirs": [],
                        "extraConfigmapMounts": [],
                        "resources": {},
                        "containerSecurityContext": {}
                      },
                      "alertmanager": {
                        "enabled": false,
                        "name": "configmap-reload",
                        "image": {
                          "repository": "quay.io/prometheus-operator/prometheus-config-reloader",
                          "tag": "v0.72.0",
                          "pullPolicy": "IfNotPresent"
                        },
                        "extraArgs": {},
                        "extraVolumeDirs": [],
                        "extraConfigmapMounts": [],
                        "resources": {}
                      }
                    },
                    "nodeExporter": {
                      "enabled": false,
                      "hostNetwork": true,
                      "hostPID": true,
                      "dnsPolicy": "ClusterFirstWithHostNet",
                      "name": "node-exporter",
                      "image": {
                        "repository": "prom/node-exporter",
                        "tag": "v1.7.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "extraArgs": {},
                      "extraHostPathMounts": [],
                      "extraConfigmapMounts": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "deploymentAnnotations": {},
                      "pod": {
                        "labels": {}
                      },
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "resources": {},
                      "securityContext": {},
                      "service": {
                        "annotations": {
                          "prometheus.io/scrape": "true"
                        },
                        "labels": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "hostPort": 9100,
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 9100,
                        "type": "ClusterIP"
                      }
                    },
                    "pushgateway": {
                      "enabled": false,
                      "name": "pushgateway",
                      "image": {
                        "repository": "prom/pushgateway",
                        "tag": "v1.6.2",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "extraArgs": {},
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "hosts": [],
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "replicaCount": 1,
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "resources": {},
                      "securityContext": {
                        "runAsUser": 1001,
                        "runAsNonRoot": true
                      },
                      "service": {
                        "annotations": {
                          "prometheus.io/probe": "pushgateway"
                        },
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 9091,
                        "type": "ClusterIP"
                      },
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "2Gi",
                        "subPath": ""
                      }
                    },
                    "serverFiles": {
                      "alerting_rules.yml": {},
                      "alerts": {},
                      "recording_rules.yml": {},
                      "prometheus.yml": {
                        "rule_files": [
                          "/etc/config/recording_rules.yml",
                          "/etc/config/alerting_rules.yml",
                          "/etc/config/rules",
                          "/etc/config/alerts"
                        ],
                        "scrape_configs": [
                          {
                            "job_name": "prometheus",
                            "static_configs": [
                              {
                                "targets": [
                                  "localhost:9090"
                                ]
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-nodes-cadvisor",
                            "scheme": "https",
                            "tls_config": {
                              "ca_file": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                              "insecure_skip_verify": true
                            },
                            "bearer_token_file": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                            "kubernetes_sd_configs": [
                              {
                                "role": "node"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_node_label_(.+)"
                              },
                              {
                                "target_label": "__address__",
                                "replacement": "kubernetes.default.svc:443"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_node_name"
                                ],
                                "regex": "(.+)",
                                "target_label": "__metrics_path__",
                                "replacement": "/api/v1/nodes/$1/proxy/metrics/cadvisor"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(container_cpu_usage_seconds_total|container_memory_working_set_bytes|container_network_receive_errors_total|container_network_transmit_errors_total|container_network_receive_packets_dropped_total|container_network_transmit_packets_dropped_total|container_memory_usage_bytes|container_cpu_cfs_throttled_periods_total|container_cpu_cfs_periods_total|container_fs_usage_bytes|container_fs_limit_bytes|container_cpu_cfs_periods_total|container_fs_inodes_free|container_fs_inodes_total|container_fs_usage_bytes|container_fs_limit_bytes|container_cpu_cfs_throttled_periods_total|container_cpu_cfs_periods_total|container_network_receive_bytes_total|container_network_transmit_bytes_total|container_fs_inodes_free|container_fs_inodes_total|container_fs_usage_bytes|container_fs_limit_bytes|container_spec_cpu_shares|container_spec_memory_limit_bytes|container_network_receive_bytes_total|container_network_transmit_bytes_total|container_fs_reads_bytes_total|container_network_receive_bytes_total|container_fs_writes_bytes_total|container_fs_reads_bytes_total|cadvisor_version_info|kubecost_pv_info)",
                                "action": "keep"
                              },
                              {
                                "source_labels": [
                                  "container"
                                ],
                                "target_label": "container_name",
                                "regex": "(.+)",
                                "action": "replace"
                              },
                              {
                                "source_labels": [
                                  "pod"
                                ],
                                "target_label": "pod_name",
                                "regex": "(.+)",
                                "action": "replace"
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-nodes",
                            "scheme": "https",
                            "tls_config": {
                              "ca_file": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                              "insecure_skip_verify": true
                            },
                            "bearer_token_file": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                            "kubernetes_sd_configs": [
                              {
                                "role": "node"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_node_label_(.+)"
                              },
                              {
                                "target_label": "__address__",
                                "replacement": "kubernetes.default.svc:443"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_node_name"
                                ],
                                "regex": "(.+)",
                                "target_label": "__metrics_path__",
                                "replacement": "/api/v1/nodes/$1/proxy/metrics"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(kubelet_volume_stats_used_bytes)",
                                "action": "keep"
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-service-endpoints",
                            "kubernetes_sd_configs": [
                              {
                                "role": "endpoints"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_scrape"
                                ],
                                "action": "keep",
                                "regex": true
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_endpoints_name"
                                ],
                                "action": "keep",
                                "regex": "(.*node-exporter|kubecost-network-costs)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_scheme"
                                ],
                                "action": "replace",
                                "target_label": "__scheme__",
                                "regex": "(https?)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_path"
                                ],
                                "action": "replace",
                                "target_label": "__metrics_path__",
                                "regex": "(.+)"
                              },
                              {
                                "source_labels": [
                                  "__address__",
                                  "__meta_kubernetes_service_annotation_prometheus_io_port"
                                ],
                                "action": "replace",
                                "target_label": "__address__",
                                "regex": "([^:]+)(?::\\d+)?;(\\d+)",
                                "replacement": "$1:$2"
                              },
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_service_label_(.+)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_namespace"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_namespace"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_name"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_name"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_pod_node_name"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_node"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(container_cpu_allocation|container_cpu_usage_seconds_total|container_fs_limit_bytes|container_fs_writes_bytes_total|container_gpu_allocation|container_memory_allocation_bytes|container_memory_usage_bytes|container_memory_working_set_bytes|container_network_receive_bytes_total|container_network_transmit_bytes_total|DCGM_FI_DEV_GPU_UTIL|deployment_match_labels|kube_daemonset_status_desired_number_scheduled|kube_daemonset_status_number_ready|kube_deployment_spec_replicas|kube_deployment_status_replicas|kube_deployment_status_replicas_available|kube_job_status_failed|kube_namespace_annotations|kube_namespace_labels|kube_node_info|kube_node_labels|kube_node_status_allocatable|kube_node_status_allocatable_cpu_cores|kube_node_status_allocatable_memory_bytes|kube_node_status_capacity|kube_node_status_capacity_cpu_cores|kube_node_status_capacity_memory_bytes|kube_node_status_condition|kube_persistentvolume_capacity_bytes|kube_persistentvolume_status_phase|kube_persistentvolumeclaim_info|kube_persistentvolumeclaim_resource_requests_storage_bytes|kube_pod_container_info|kube_pod_container_resource_limits|kube_pod_container_resource_limits_cpu_cores|kube_pod_container_resource_limits_memory_bytes|kube_pod_container_resource_requests|kube_pod_container_resource_requests_cpu_cores|kube_pod_container_resource_requests_memory_bytes|kube_pod_container_status_restarts_total|kube_pod_container_status_running|kube_pod_container_status_terminated_reason|kube_pod_labels|kube_pod_owner|kube_pod_status_phase|kube_replicaset_owner|kube_statefulset_replicas|kube_statefulset_status_replicas|kubecost_cluster_info|kubecost_cluster_management_cost|kubecost_cluster_memory_working_set_bytes|kubecost_load_balancer_cost|kubecost_network_internet_egress_cost|kubecost_network_region_egress_cost|kubecost_network_zone_egress_cost|kubecost_node_is_spot|kubecost_pod_network_egress_bytes_total|node_cpu_hourly_cost|node_cpu_seconds_total|node_disk_reads_completed|node_disk_reads_completed_total|node_disk_writes_completed|node_disk_writes_completed_total|node_filesystem_device_error|node_gpu_count|node_gpu_hourly_cost|node_memory_Buffers_bytes|node_memory_Cached_bytes|node_memory_MemAvailable_bytes|node_memory_MemFree_bytes|node_memory_MemTotal_bytes|node_network_transmit_bytes_total|node_ram_hourly_cost|node_total_hourly_cost|pod_pvc_allocation|pv_hourly_cost|service_selector_labels|statefulSet_match_labels|kubecost_pv_info|up)",
                                "action": "keep"
                              }
                            ]
                          }
                        ]
                      },
                      "rules": {
                        "groups": [
                          {
                            "name": "CPU",
                            "rules": [
                              {
                                "expr": "sum(rate(container_cpu_usage_seconds_total{container!=\"\"}[5m]))",
                                "record": "cluster:cpu_usage:rate5m"
                              },
                              {
                                "expr": "rate(container_cpu_usage_seconds_total{container!=\"\"}[5m])",
                                "record": "cluster:cpu_usage_nosum:rate5m"
                              },
                              {
                                "expr": "avg(irate(container_cpu_usage_seconds_total{container!=\"POD\", container!=\"\"}[5m])) by (container,pod,namespace)",
                                "record": "kubecost_container_cpu_usage_irate"
                              },
                              {
                                "expr": "sum(container_memory_working_set_bytes{container!=\"POD\",container!=\"\"}) by (container,pod,namespace)",
                                "record": "kubecost_container_memory_working_set_bytes"
                              },
                              {
                                "expr": "sum(container_memory_working_set_bytes{container!=\"POD\",container!=\"\"})",
                                "record": "kubecost_cluster_memory_working_set_bytes"
                              }
                            ]
                          },
                          {
                            "name": "Savings",
                            "rules": [
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind!=\"DaemonSet\"}) by (pod) * sum(container_cpu_allocation) by (pod))",
                                "record": "kubecost_savings_cpu_allocation",
                                "labels": {
                                  "daemonset": "false"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind=\"DaemonSet\"}) by (pod) * sum(container_cpu_allocation) by (pod)) / sum(kube_node_info)",
                                "record": "kubecost_savings_cpu_allocation",
                                "labels": {
                                  "daemonset": "true"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind!=\"DaemonSet\"}) by (pod) * sum(container_memory_allocation_bytes) by (pod))",
                                "record": "kubecost_savings_memory_allocation_bytes",
                                "labels": {
                                  "daemonset": "false"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind=\"DaemonSet\"}) by (pod) * sum(container_memory_allocation_bytes) by (pod)) / sum(kube_node_info)",
                                "record": "kubecost_savings_memory_allocation_bytes",
                                "labels": {
                                  "daemonset": "true"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    "alertRelabelConfigs": null,
                    "networkPolicy": {
                      "enabled": false
                    }
                  },
                  "networkCosts": {
                    "enabled": false,
                    "image": {
                      "repository": "gcr.io/kubecost1/kubecost-network-costs",
                      "tag": "v0.17.3"
                    },
                    "imagePullPolicy": "Always",
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "prometheusScrape": false,
                    "trafficLogging": true,
                    "logLevel": "info",
                    "port": 3001,
                    "resources": {
                      "limits": {
                        "cpu": "500m"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "20Mi"
                      }
                    },
                    "extraArgs": [],
                    "config": {
                      "destinations": {
                        "in-zone": [
                          "127.0.0.0/8",
                          "169.254.0.0/16",
                          "10.0.0.0/8",
                          "172.16.0.0/12",
                          "192.168.0.0/16"
                        ],
                        "in-region": [],
                        "cross-region": [],
                        "internet": [],
                        "direct-classification": []
                      },
                      "services": {
                        "google-cloud-services": false,
                        "amazon-web-services": false,
                        "azure-cloud-services": false
                      }
                    },
                    "tolerations": [],
                    "affinity": {},
                    "service": {
                      "annotations": {},
                      "labels": {}
                    },
                    "priorityClassName": "",
                    "podMonitor": {
                      "enabled": false,
                      "additionalLabels": {}
                    },
                    "additionalLabels": {},
                    "nodeSelector": {},
                    "annotations": {},
                    "healthCheckProbes": {},
                    "additionalSecurityContext": {}
                  },
                  "kubecostDeployment": {
                    "replicas": 1,
                    "labels": {},
                    "annotations": {}
                  },
                  "forecasting": {
                    "enabled": true,
                    "fullImageName": "gcr.io/kubecost1/kubecost-modeling:v0.1.6",
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "limits": {
                        "cpu": "1500m",
                        "memory": "1Gi"
                      }
                    },
                    "env": {
                      "GUNICORN_CMD_ARGS": "--log-level info -t 1200"
                    },
                    "priority": {
                      "enabled": false,
                      "name": ""
                    },
                    "nodeSelector": {},
                    "tolerations": {},
                    "affinity": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    }
                  },
                  "kubecostAggregator": {
                    "deployMethod": "singlepod",
                    "enabled": false,
                    "replicas": 1,
                    "stagingEmptyDirSizeLimit": "2Gi",
                    "numDBCopyPartitions": 25,
                    "env": {
                      "LOG_LEVEL": "info",
                      "DB_READ_THREADS": "1",
                      "DB_WRITE_THREADS": "1",
                      "DB_CONCURRENT_INGESTION_COUNT": "3"
                    },
                    "persistentConfigsStorage": {
                      "storageClass": "",
                      "storageRequest": "1Gi"
                    },
                    "aggregatorDbStorage": {
                      "storageClass": "",
                      "storageRequest": "128Gi"
                    },
                    "resources": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "cloudCost": {
                      "enabled": false,
                      "resources": {},
                      "readinessProbe": {
                        "enabled": true,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 10,
                        "failureThreshold": 200
                      }
                    },
                    "jaeger": {
                      "enabled": false,
                      "image": "jaegertracing/all-in-one",
                      "imageVersion": "latest"
                    }
                  },
                  "diagnostics": {
                    "enabled": true,
                    "primary": {
                      "enabled": false,
                      "retention": "7d",
                      "readonly": false
                    },
                    "pollingInterval": "300s",
                    "keepDiagnosticHistory": false,
                    "collectHelmValues": false,
                    "deployment": {
                      "enabled": false,
                      "resources": {
                        "requests": {
                          "cpu": "10m",
                          "memory": "20Mi"
                        }
                      },
                      "env": {},
                      "labels": {},
                      "securityContext": {},
                      "containerSecurityContext": {},
                      "nodeSelector": {},
                      "tolerations": {},
                      "affinity": {}
                    }
                  },
                  "clusterController": {
                    "enabled": false,
                    "image": {
                      "repository": "gcr.io/kubecost1/cluster-controller",
                      "tag": "v0.16.0"
                    },
                    "imagePullPolicy": "Always",
                    "priorityClassName": "",
                    "tolerations": [],
                    "actionConfigs": {
                      "clusterTurndown": [],
                      "namespaceTurndown": null,
                      "clusterRightsize": null,
                      "containerRightsize": null
                    },
                    "kubescaler": {
                      "defaultResizeAll": false
                    },
                    "namespaceTurndown": {
                      "rbac": {
                        "enabled": true
                      }
                    }
                  },
                  "reporting": {
                    "logCollection": true,
                    "productAnalytics": true,
                    "errorReporting": true,
                    "valuesReporting": true
                  },
                  "serviceMonitor": {
                    "enabled": false,
                    "additionalLabels": {},
                    "metricRelabelings": [],
                    "relabelings": [],
                    "networkCosts": {
                      "enabled": false,
                      "scrapeTimeout": "10s",
                      "additionalLabels": {},
                      "metricRelabelings": [],
                      "relabelings": []
                    }
                  },
                  "prometheusRule": {
                    "enabled": false,
                    "additionalLabels": {}
                  },
                  "supportNFS": false,
                  "initChownDataImage": "busybox",
                  "initChownData": {
                    "resources": {}
                  },
                  "grafana": {
                    "rbac": {
                      "create": true
                    },
                    "serviceAccount": {
                      "create": true,
                      "name": ""
                    },
                    "replicas": 1,
                    "deploymentStrategy": "RollingUpdate",
                    "readinessProbe": {
                      "httpGet": {
                        "path": "/api/health",
                        "port": 3000
                      }
                    },
                    "livenessProbe": {
                      "httpGet": {
                        "path": "/api/health",
                        "port": 3000
                      },
                      "initialDelaySeconds": 60,
                      "timeoutSeconds": 30,
                      "failureThreshold": 10
                    },
                    "image": {
                      "repository": "grafana/grafana",
                      "tag": "10.3.4",
                      "pullPolicy": "IfNotPresent"
                    },
                    "securityContext": {},
                    "priorityClassName": "",
                    "downloadDashboardsImage": {
                      "repository": "curlimages/curl",
                      "tag": "latest",
                      "pullPolicy": "IfNotPresent"
                    },
                    "podAnnotations": {},
                    "annotations": {},
                    "service": {
                      "type": "ClusterIP",
                      "port": 80,
                      "annotations": {},
                      "labels": {}
                    },
                    "ingress": {
                      "enabled": false,
                      "annotations": {},
                      "labels": {},
                      "path": "/",
                      "pathType": "Prefix",
                      "hosts": [
                        "chart-example.local"
                      ],
                      "tls": []
                    },
                    "resources": {},
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "persistence": {
                      "enabled": false
                    },
                    "adminUser": "admin",
                    "adminPassword": "strongpassword",
                    "env": {},
                    "envFromSecret": "",
                    "extraSecretMounts": [],
                    "plugins": [],
                    "dashboardProviders": {},
                    "dashboards": {},
                    "dashboardsConfigMaps": {},
                    "ldap": {
                      "existingSecret": "",
                      "config": ""
                    },
                    "smtp": {
                      "existingSecret": ""
                    },
                    "sidecar": {
                      "image": {
                        "repository": "kiwigrid/k8s-sidecar",
                        "tag": "1.26.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "resources": {},
                      "dashboards": {
                        "enabled": true,
                        "label": "grafana_dashboard",
                        "labelValue": "1",
                        "annotations": {},
                        "error_throttle_sleep": 0,
                        "folder": "/tmp/dashboards"
                      },
                      "datasources": {
                        "enabled": false,
                        "error_throttle_sleep": 0,
                        "label": "grafana_datasource"
                      }
                    },
                    "grafana.ini": {
                      "server": {
                        "serve_from_sub_path": false,
                        "root_url": "%(protocol)s://%(domain)s:%(http_port)s/grafana"
                      },
                      "paths": {
                        "data": "/var/lib/grafana/data",
                        "logs": "/var/log/grafana",
                        "plugins": "/var/lib/grafana/plugins",
                        "provisioning": "/etc/grafana/provisioning"
                      },
                      "analytics": {
                        "check_for_updates": true
                      },
                      "log": {
                        "mode": "console"
                      },
                      "grafana_net": {
                        "url": "https://grafana.net"
                      },
                      "auth.anonymous": {
                        "enabled": true,
                        "org_role": "Editor",
                        "org_name": "Main Org."
                      }
                    }
                  },
                  "serviceAccount": {
                    "create": true,
                    "annotations": {}
                  },
                  "awsstore": {
                    "useAwsStore": false,
                    "imageNameAndVersion": "gcr.io/kubecost1/awsstore:latest",
                    "createServiceAccount": false,
                    "priorityClassName": "",
                    "nodeSelector": {},
                    "annotations": {}
                  },
                  "federatedETL": {
                    "agentOnly": false,
                    "federatedCluster": false,
                    "redirectS3Backup": false,
                    "useMultiClusterDB": false
                  },
                  "kubecostAdmissionController": {
                    "enabled": false,
                    "secretName": "webhook-server-tls",
                    "caBundle": "${CA_BUNDLE}"
                  },
                  "costEventsAudit": {
                    "enabled": false
                  },
                  "extraObjects": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: gcr.io/kubecost1/operator/kubecost-operator-controller:v2.2.4
          createdAt: "2024-05-19T20:46:35Z"
          description: An operator that deploys Kubecost 2.2.4 via the Helm chart.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Stackwatch
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The CostAnalyzer API describes a deployment of Kubecost and
              its dependencies.
            displayName: CostAnalyzer
            kind: CostAnalyzer
            name: costanalyzers.charts.kubecost.com
            version: v1alpha1
        description: |-
          Kubecost provides real-time cost visibility and insights for teams using Kubernetes, helping you continuously reduce your cloud costs.

          This OpenShift operator deploys Kubecost from the official Helm chart.

          Kubecost allows you to:

          * See allocated spend across all native Kubernetes concepts, so you can provide your teams with transparent, accurate cost data reconciled with your actual cloud bill.
          * Join in-cluster costs like CPU and memory with out-of-cluster spend from cloud infrastructure services for complete cost visibility—all services from AWS, GCP, and Microsoft Azure are supported.
          * Gain insights you can use to save 30-50% or more on your infrastructure spend, without exposing your private information—your data never gets shared externally, even with us.
          * Have real-time alerts and recurring reports, empowering teams to take control of their Kubernetes-enabled infrastructure, stay within budgeted limits, and address monitoring interruptions immediately.

          Visit the Kubecost website for more details including features, pricing, documentation, and more at www.kubecost.com.
        displayName: Kubecost
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubecost
        - cost
        - kubernetes
        links:
        - name: Kubecost
          url: https://www.kubecost.com
        maintainers:
        - email: czoller@kubecost.com
          name: Chip Zoller
        maturity: alpha
        provider:
          name: Stackwatch
          url: www.kubecost.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - gcr.io/kubecost1/operator/kubecost-operator-controller:v2.2.4
        version: 2.2.4
      entries:
      - name: kubecost.v2.2.4
        version: 2.2.4
      - name: kubecost.v2.1.0
        version: 2.1.0
      - name: kubecost-operator.v1.107.1
        version: 1.107.1
      - name: kubecost-operator.v1.106.2
        version: 1.106.2
      - name: kubecost-operator.v1.106.0
        version: 1.106.0
      - name: kubecost.v1.99.0
        version: 1.99.0
      - name: kubecost.v1.97.0
        version: 1.97.0
      name: alpha
    defaultChannel: alpha
    packageName: kubecost
    provider:
      name: Stackwatch
      url: www.kubecost.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: prometurbo
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometurbo-operator.v8.7.1-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.7.1-SNAPSHOT
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.7.1-SNAPSHOT
        version: 8.7.1-beta.1
      entries:
      - name: prometurbo-operator.v8.7.1-beta.1
        version: 8.7.1-beta.1
      name: beta
    - currentCSV: prometurbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.6
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.6
        version: 8.6.0
      entries:
      - name: prometurbo-operator.v8.6.0
        version: 8.6.0
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: stable
    - currentCSV: prometurbo-operator.v8.5.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.5
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.5
        version: 8.5.0
      entries:
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: v8.5.0
    - currentCSV: prometurbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.6
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.6
        version: 8.6.0
      entries:
      - name: prometurbo-operator.v8.6.0
        version: 8.6.0
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: v8.6.0
    defaultChannel: stable
    packageName: prometurbo
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apache KIE
      provider-url: ""
    name: sonataflow-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sonataflow-operator.v10.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "docker.io/apache",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: docker.io/apache/incubator-kie-sonataflow-operator:10.0.0
          description: SonataFlow Kubernetes Operator for deploying workflow applications
            based on the CNCF Serverless Workflow specification
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/apache/incubator-kie-tools/packages/sonataflow-operator
          support: Apache Software Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: Sonata Flow Build
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: Sonata Flow Cluster Platform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: Sonata Flow Platform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: Sonata Flow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          SonataFlow Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification](https://serverlessworkflow.io/):

          * Deploy workflow applications using the [dev profile](https://sonataflow.org/serverlessworkflow/latest/cloud/operator/developing-workflows.html), suited for the your development cycle
          * Build workflow applications based on the platform you're currently working on.
        displayName: SonataFlow Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        links:
        - name: Product Page
          url: https://github.com/apache/incubator-kie-tools
        maintainers:
        - email: dev@kie.apache.org
          name: Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Apache Software Foundation
        relatedImages:
        - docker.io/apache/incubator-kie-sonataflow-operator:10.0.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 10.0.0
      entries:
      - name: sonataflow-operator.v10.0.0
        version: 10.0.0
      - name: sonataflow-operator.v1.40.1
        version: 1.40.1
      - name: sonataflow-operator.v1.40.0
        version: 1.40.0
      name: alpha
    - currentCSV: sonataflow-operator.v10.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "host/namespace",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: docker.io/apache/incubator-kie-sonataflow-operator:10.1.0
          createdAt: "2025-09-11T14:44:56Z"
          description: SonataFlow Kubernetes Operator for deploying workflow applications
            based on the CNCF Serverless Workflow specification
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/apache/incubator-kie-tools/packages/sonataflow-operator
          support: Apache KIE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: SonataFlowBuild
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: SonataFlowClusterPlatform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: SonataFlowPlatform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: SonataFlow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          SonataFlow Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification](https://serverlessworkflow.io/):

          * Deploy workflow applications using the [dev profile](https://sonataflow.org/serverlessworkflow/latest/cloud/operator/developing-workflows.html), suited for the your development cycle
          * Build workflow applications based on the platform you're currently working on.
        displayName: SonataFlow Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        links:
        - name: Product Page
          url: https://sonataflow.org/serverlessworkflow/latest/index.html
        maintainers:
        - email: dev@kie.apache.org
          name: Apache KIE
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Apache KIE
        relatedImages:
        - docker.io/apache/incubator-kie-sonataflow-operator:10.1.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 10.1.0
      entries:
      - name: sonataflow-operator.v10.1.0
        version: 10.1.0
      - name: sonataflow-operator.v10.0.0
        version: 10.0.0
      - name: sonataflow-operator.v1.40.1
        version: 1.40.1
      - name: sonataflow-operator.v1.40.0
        version: 1.40.0
      name: stable
    defaultChannel: stable
    packageName: sonataflow-operator
    provider:
      name: Apache KIE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: directpv
      provider-url: ""
    name: private-dpv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: private-dpv-operator.v5.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.quay.io/v1alpha1",
                "kind": "AistorVolumemanager",
                "metadata": {
                  "name": "aistorvolumemanager-sample"
                },
                "spec": {
                  "annotations": {},
                  "controller": {
                    "podAnnotations": {},
                    "podLabels": {},
                    "replicas": 3
                  },
                  "images": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": []
                  },
                  "labels": {},
                  "license": "",
                  "nodeserver": {
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-10-05T15:47:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AistorVolumemanager
            name: aistorvolumemanagers.charts.quay.io
            version: v1alpha1
        description: directpv
        displayName: directpv
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - directpv
        links:
        - name: Private Dpv Operator
          url: https://private-dpv-operator.domain
        maintainers:
        - email: cesarcelis@minio.io
          name: celis
        maturity: alpha
        provider:
          name: directpv
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - quay.io/cniackz4/directpv-operator@sha256:af06e57661a7e05809e44c3fac571ed949cc04fd656fdb0c8324b533778eb604
        version: 5.0.3
      entries:
      - name: private-dpv-operator.v5.0.3
        version: 5.0.3
      - name: private-dpv-operator.v5.0.2
        version: 5.0.2
      name: alpha
    defaultChannel: alpha
    packageName: private-dpv-operator
    provider:
      name: directpv
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: service-telemetry-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-telemetry-operator.v1.5.1757650583
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infra.watch/v1beta1",
                "kind": "ServiceTelemetry",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "alerting": {
                    "alertmanager": {
                      "receivers": {
                        "snmpTraps": {
                          "alertOidLabel": "oid",
                          "community": "public",
                          "enabled": false,
                          "port": 162,
                          "retries": 5,
                          "target": "192.168.24.254",
                          "timeout": 1,
                          "trapDefaultOid": "1.3.6.1.4.1.50495.15.1.2.1",
                          "trapDefaultSeverity": "",
                          "trapOidPrefix": "1.3.6.1.4.1.50495.15"
                        }
                      },
                      "storage": {
                        "persistent": {
                          "pvcStorageRequest": "20G"
                        },
                        "strategy": "persistent"
                      }
                    },
                    "enabled": true
                  },
                  "backends": {
                    "events": {
                      "elasticsearch": {
                        "certificates": {
                          "caCertDuration": "70080h",
                          "endpointCertDuration": "70080h"
                        },
                        "enabled": false,
                        "forwarding": {
                          "hostUrl": "https://elasticsearch-es-http:9200",
                          "tlsSecretName": "elasticsearch-es-cert",
                          "tlsServerName": "",
                          "useBasicAuth": true,
                          "useTls": true,
                          "userSecretName": "elasticsearch-es-elastic-user"
                        },
                        "storage": {
                          "persistent": {
                            "pvcStorageRequest": "20Gi"
                          },
                          "strategy": "persistent"
                        },
                        "version": "7.16.1"
                      }
                    },
                    "metrics": {
                      "prometheus": {
                        "enabled": true,
                        "scrapeInterval": "30s",
                        "storage": {
                          "persistent": {
                            "pvcStorageRequest": "20G"
                          },
                          "retention": "24h",
                          "strategy": "persistent"
                        }
                      }
                    }
                  },
                  "clouds": [
                    {
                      "metrics": {
                        "collectors": [
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 16384,
                              "verbose": false
                            },
                            "collectorType": "collectd",
                            "debugEnabled": false,
                            "subscriptionAddress": "collectd/cloud1-telemetry"
                          },
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 16384,
                              "verbose": false
                            },
                            "collectorType": "ceilometer",
                            "debugEnabled": false,
                            "subscriptionAddress": "anycast/ceilometer/cloud1-metering.sample"
                          },
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 65535,
                              "verbose": false
                            },
                            "collectorType": "sensubility",
                            "debugEnabled": false,
                            "subscriptionAddress": "sensubility/cloud1-telemetry"
                          }
                        ]
                      },
                      "name": "cloud1"
                    }
                  ],
                  "graphing": {
                    "enabled": false,
                    "grafana": {
                      "baseImage": "registry.redhat.io/rhel8/grafana:9",
                      "dashboards": {
                        "enabled": true
                      },
                      "disableSignoutMenu": false,
                      "ingressEnabled": true
                    }
                  },
                  "highAvailability": {
                    "enabled": false
                  },
                  "observabilityStrategy": "use_redhat",
                  "transports": {
                    "qdr": {
                      "auth": "basic",
                      "certificates": {
                        "caCertDuration": "70080h",
                        "endpointCertDuration": "70080h"
                      },
                      "enabled": true,
                      "web": {
                        "enabled": false
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/stf/service-telemetry-rhel9-operator@sha256:822becd77fa8835e753c88b3e307f62e42e090bdedc6e1fbcbe9fd6110310846
          createdAt: "2025-09-12T04:17:40Z"
          description: Service Telemetry Framework. Umbrella Operator for instantiating
            the required dependencies and configuration of various components to build
            a Service Telemetry platform for telco grade monitoring.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: service-telemetry
          operators.openshift.io/valid-subscription: '["OpenStack Platform", "Cloud
            Infrastructure", "Cloud Suite"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/infrawatch/service-telemetry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of the Service Telemetry Framework
            displayName: STF Cluster
            kind: ServiceTelemetry
            name: servicetelemetrys.infra.watch
            version: v1beta1
        description: Service Telemetry Operator for monitoring clouds
        displayName: Service Telemetry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - servicetelemetry
        - monitoring
        - telemetry
        - notifications
        - telecommunications
        links:
        - name: Source Code
          url: https://github.com/infrawatch/service-telemetry-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_openstack_platform/17.1/html/service_telemetry_framework_1.5/index
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy:latest@sha256:dc24aa382e02c4ac058d4b824a21f0e90a6259d6d817d0341b2a41809b5daa78
        - registry.redhat.io/cluster-observability-operator/coo-prometheus-rhel8:0.2.0@sha256:6bbb1aebb1cfdc3310e398ec7720ca1d4a9a89cecf7406412071bfaceb0d3cb3
        - registry.redhat.io/cluster-observability-operator/coo-prometheus-alertmanager-rhel8:0.2.0@sha256:cfd18e83d64a55ac310937eefe83fc5b9e98660962a127cd4f8d9a382764c171
        - registry.redhat.io/stf/service-telemetry-rhel9-operator@sha256:822becd77fa8835e753c88b3e307f62e42e090bdedc6e1fbcbe9fd6110310846
        - registry.redhat.io/stf/prometheus-webhook-snmp-rhel9@sha256:03199f8a3b6b22c82e01ede3adab6e98b276369d642029ae03641aac3b25581a
        version: 1.5.1757650583
      entries:
      - name: service-telemetry-operator.v1.5.1757650583
        version: 1.5.1757650583
      - name: service-telemetry-operator.v1.5.1743671601
        version: 1.5.1743671601
      - name: service-telemetry-operator.v1.5.1740051851
        version: 1.5.1740051851
      - name: service-telemetry-operator.v1.5.1731336911
        version: 1.5.1731336911
      - name: service-telemetry-operator.v1.5.1725952694
        version: 1.5.1725952694
      name: stable-1.5
    defaultChannel: stable-1.5
    packageName: service-telemetry-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: barbican-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: barbican-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "barbican.openstack.org/v1beta1",
                "kind": "Barbican",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "barbican-operator",
                    "app.kubernetes.io/instance": "barbican",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "barbican",
                    "app.kubernetes.io/part-of": "barbican-operator"
                  },
                  "name": "barbican"
                },
                "spec": {
                  "barbicanAPI": {
                    "customServiceConfig": "[optional]\noverrides = True\n",
                    "defautlConfigOverwrite": {
                      "optional_policy.json": "{\"some\": \"custom policy\"}\n"
                    },
                    "passwordSelectors": {
                      "database": "BarbicanDatabasePassword",
                      "service": "BarbicanPassword"
                    }
                  },
                  "barbicanKeystoneListener": {
                    "customServiceConfig": "[optional]\noverrides = True\n",
                    "defautlConfigOverwrite": {
                      "optional_policy.json": "{\"some\": \"custom policy\"}\n"
                    }
                  },
                  "barbicanWorker": {
                    "customServiceConfig": "[optional]\noverrides = True\n",
                    "defautlConfigOverwrite": {
                      "optional_policy.json": "{\"some\": \"custom policy\"}\n"
                    }
                  },
                  "customServiceConfig": "[DEFAULT]\ndebug = True\n",
                  "databaseAccount": "barbican",
                  "databaseInstance": "openstack",
                  "defaultConfigOverwrite": {
                    "policy.json": "{\"some\": \"custom policy\"}\n"
                  },
                  "passwordSelectors": {
                    "database": "BarbicanDatabasePassword",
                    "service": "BarbicanPassword",
                    "simplecryptokek": "BarbicanSimpleCryptoKEK"
                  },
                  "preserveJobs": true,
                  "rabbitMqCusterName": "barbican_rabbit",
                  "secret": "osp-secret",
                  "serviceAccount": "barbican",
                  "serviceUser": "barbican"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:32Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BarbicanAPI is the Schema for the barbicanapis API
            displayName: Barbican API
            kind: BarbicanAPI
            name: barbicanapis.barbican.openstack.org
            version: v1beta1
          - kind: BarbicanKeystoneListener
            name: barbicankeystonelisteners.barbican.openstack.org
            version: v1beta1
          - description: Barbican is the Schema for the barbicans API
            displayName: Barbican
            kind: Barbican
            name: barbicans.barbican.openstack.org
            version: v1beta1
          - description: BarbicanWorker is the Schema for the barbicanworkers API
            displayName: Barbican Worker
            kind: BarbicanWorker
            name: barbicanworkers.barbican.openstack.org
            version: v1beta1
        description: Barbican Operator
        displayName: Barbican Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - KeyManager
        - Barbican
        links:
        - name: Barbican Operator
          url: https://github.com/openstack-k8s-operators/barbican-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-barbican-worker-rhel9@sha256:b5133f539658ed5d532e0c3cf63d90bc0f76c7f0368963a8ebec49123f27f7a7
        - registry.redhat.io/rhoso/openstack-barbican-keystone-listener-rhel9@sha256:ac112444c0a12664e8a7900e72cd97fa795df167c0d95437a029644f0819f30d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/barbican-rhel9-operator@sha256:1275a3f7f09999e9dc25bc38f917a3b5d780d9d309442cddd0e7adf313d6f576
        - registry.redhat.io/rhoso/openstack-barbican-api-rhel9@sha256:0db6c1ab5b4ca302e38338a3ab9cee9bd4ae7c8e1e7bd3419e47b65df9ef1f4f
        version: 1.0.6
      entries:
      - name: barbican-operator.v1.0.6
        version: 1.0.6
      - name: barbican-operator.v1.0.5
        version: 1.0.5
      - name: barbican-operator.v1.0.4
        version: 1.0.4
      - name: barbican-operator.v1.0.3
        version: 1.0.3
      - name: barbican-operator.v1.0.2
        version: 1.0.2
      - name: barbican-operator.v1.0.1
        version: 1.0.1
      - name: barbican-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: barbican-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: ansible-cloud-addons-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aca-operator.v2.4.0-0.1737660894
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Cloud Addons Operator manages everything Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.1.1-0.1645500829 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy an AAP AWS Billing instance
            displayName: AnsibleAutomationPlatformAwsBilling
            kind: AnsibleAutomationPlatformAwsBilling
            name: ansibleautomationplatformawsbillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Azure Billing instance
            displayName: AnsibleAutomationPlatformAzureBilling
            kind: AnsibleAutomationPlatformAzureBilling
            name: ansibleautomationplatformazurebillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Metrics instance
            displayName: AnsibleAutomationPlatformMetrics
            kind: AnsibleAutomationPlatformMetrics
            name: ansibleautomationplatformmetrics.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP UI instance
            displayName: AutomationPlatform
            kind: AutomationPlatform
            name: automationplatforms.platform.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Cloud Addons is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Cloud Addons Operator provides for push-button deployments of Cloud Add-ons into your OpenShift environment. Ansible Cloud Addons deployments include:
            * Cloud Billing - Cloud billing connector for Ansible Automation Platform
            * Cloud UI - Unified welcome page UI for Ansible Automation Platform

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Cloud Addons operator by following instructions presented when you click the Install button.
        displayName: Ansible Cloud Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-platform
        - ansible-automation-platform
        - aap
        - cloud
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:38ee3484ba60bb474e7231e9d0718157386716cea7ffd6da6f32ddd551a5243d
        - registry.redhat.io/ansible-automation-platform-24/aap-cloud-billing-rhel8-operator@sha256:c461b0c4781e95d2b2e31e7bbd6d6d192806e8a8dd2e8b9225b5735da93dd6a2
        - registry.redhat.io/ansible-automation-platform-24/aap-cloud-ui-rhel8-operator@sha256:06bf313d4daf080ce6f2b1ba859fefd84431c7a2a0618feae55b92c9f5fff594
        - registry.redhat.io/ansible-automation-platform-24/aap-cloud-billing-rhel8@sha256:84fa86d312663a74502516f2ba838ac80e500c7bc17dc518940984a603da2548
        - registry.redhat.io/ansible-automation-platform-24/aap-cloud-metrics-collector-rhel8@sha256:6bd27ad06d0a65c7fa1cd2419bfc30976d53a5b89698d309cabbbd65b531a1ab
        - registry.redhat.io/ansible-automation-platform-24/aap-cloud-ui-rhel8@sha256:8489c253072186db064289f5ecc0a08ed3654a73be3c3f2130be479098e59047
        version: 2.4.0+0.1737660894
      entries:
      - name: aca-operator.v2.4.0-0.1708974547
        version: 2.4.0+0.1708974547
      - name: aca-operator.v2.4.0-0.1711587287
        version: 2.4.0+0.1711587287
      - name: aca-operator.v2.4.0-0.1713463331
        version: 2.4.0+0.1713463331
      - name: aca-operator.v2.4.0-0.1715891874
        version: 2.4.0+0.1715891874
      - name: aca-operator.v2.4.0-0.1717108266
        version: 2.4.0+0.1717108266
      - name: aca-operator.v2.4.0-0.1717558193
        version: 2.4.0+0.1717558193
      - name: aca-operator.v2.4.0-0.1718149348
        version: 2.4.0+0.1718149348
      - name: aca-operator.v2.4.0-0.1720478479
        version: 2.4.0+0.1720478479
      - name: aca-operator.v2.4.0-0.1725253934
        version: 2.4.0+0.1725253934
      - name: aca-operator.v2.4.0-0.1726108417
        version: 2.4.0+0.1726108417
      - name: aca-operator.v2.4.0-0.1727255735
        version: 2.4.0+0.1727255735
      - name: aca-operator.v2.4.0-0.1730152463
        version: 2.4.0+0.1730152463
      - name: aca-operator.v2.4.0-0.1733181569
        version: 2.4.0+0.1733181569
      - name: aca-operator.v2.4.0-0.1733937635
        version: 2.4.0+0.1733937635
      - name: aca-operator.v2.4.0-0.1737660894
        version: 2.4.0+0.1737660894
      name: stable-2.4-cluster-scoped
    - currentCSV: aca-operator.v2.5.0-0.1758143647
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Cloud Addons Operator manages everything Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.1.1-0.1645500829 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy an AAP AWS Billing instance
            displayName: AnsibleAutomationPlatformAwsBilling
            kind: AnsibleAutomationPlatformAwsBilling
            name: ansibleautomationplatformawsbillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Azure Billing instance
            displayName: AnsibleAutomationPlatformAzureBilling
            kind: AnsibleAutomationPlatformAzureBilling
            name: ansibleautomationplatformazurebillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Metrics instance
            displayName: AnsibleAutomationPlatformMetrics
            kind: AnsibleAutomationPlatformMetrics
            name: ansibleautomationplatformmetrics.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP UI instance
            displayName: AutomationPlatform
            kind: AutomationPlatform
            name: automationplatforms.platform.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Cloud Addons is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Cloud Addons Operator provides for push-button deployments of Cloud Add-ons into your OpenShift environment. Ansible Cloud Addons deployments include:
            * Cloud Billing - Cloud billing connector for Ansible Automation Platform
            * Cloud UI - Unified welcome page UI for Ansible Automation Platform

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Cloud Addons operator by following instructions presented when you click the Install button.
        displayName: Ansible Cloud Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-platform
        - ansible-automation-platform
        - aap
        - cloud
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-billing-rhel8-operator@sha256:f4b29ec59498bb6c9fdb2790ffc9d8c878efa3fd9da3416f7fe9f2a253171508
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-ui-rhel8-operator@sha256:47e2c40cdcfe218d5bf7abbf6fc35340bacdc2d02dac7f81deee7876b5570c53
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-billing-rhel8@sha256:0e50f539bfeaec91f0b18409cbc1888297a26af5114252a8ca07e1b97436001d
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-metrics-collector-rhel8@sha256:c6ab25b3839f8c3134e7d5e1f0b7716ea5fd66fe8d131261ff08aabd97c3cd2d
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-ui-rhel8@sha256:869efe830cdf23f9344263f0e9b5a209cb13b84d9d7588a1744fd3ea4cd3f83b
        version: 2.5.0+0.1758143647
      entries:
      - name: aca-operator.v2.5.0-0.1727240231
        version: 2.5.0+0.1727240231
      - name: aca-operator.v2.5.0-0.1727874815
        version: 2.5.0+0.1727874815
      - name: aca-operator.v2.5.0-0.1728516369
        version: 2.5.0+0.1728516369
      - name: aca-operator.v2.5.0-0.1729737962
        version: 2.5.0+0.1729737962
      - name: aca-operator.v2.5.0-0.1731452082
        version: 2.5.0+0.1731452082
      - name: aca-operator.v2.5.0-0.1733188778
        version: 2.5.0+0.1733188778
      - name: aca-operator.v2.5.0-0.1733964232
        version: 2.5.0+0.1733964232
      - name: aca-operator.v2.5.0-0.1736892777
        version: 2.5.0+0.1736892777
      - name: aca-operator.v2.5.0-0.1737095650
        version: 2.5.0+0.1737095650
      - name: aca-operator.v2.5.0-0.1737671171
        version: 2.5.0+0.1737671171
      - name: aca-operator.v2.5.0-0.1739244024
        version: 2.5.0+0.1739244024
      - name: aca-operator.v2.5.0-0.1740088272
        version: 2.5.0+0.1740088272
      - name: aca-operator.v2.5.0-0.1740769078
        version: 2.5.0+0.1740769078
      - name: aca-operator.v2.5.0-0.1741367998
        version: 2.5.0+0.1741367998
      - name: aca-operator.v2.5.0-0.1742427652
        version: 2.5.0+0.1742427652
      - name: aca-operator.v2.5.0-0.1743655743
        version: 2.5.0+0.1743655743
      - name: aca-operator.v2.5.0-0.1746134365
        version: 2.5.0+0.1746134365
      - name: aca-operator.v2.5.0-0.1747339736
        version: 2.5.0+0.1747339736
      - name: aca-operator.v2.5.0-0.1749072064
        version: 2.5.0+0.1749072064
      - name: aca-operator.v2.5.0-0.1750895274
        version: 2.5.0+0.1750895274
      - name: aca-operator.v2.5.0-0.1753899647
        version: 2.5.0+0.1753899647
      - name: aca-operator.v2.5.0-0.1755834478
        version: 2.5.0+0.1755834478
      - name: aca-operator.v2.5.0-0.1758143647
        version: 2.5.0+0.1758143647
      name: stable-2.5-cluster-scoped
    defaultChannel: stable-2.5-cluster-scoped
    packageName: ansible-cloud-addons-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SAP LinuxLab - Red Hat
      provider-url: https://github.com/sap-linuxlab
    name: sap-data-intelligence-observer-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: observer-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sdi.sap-redhat.io/v1alpha1",
                "kind": "SDIObserver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "observer-operator",
                    "app.kubernetes.io/instance": "sdiobserver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sdiobserver",
                    "app.kubernetes.io/part-of": "observer-operator"
                  },
                  "name": "sdiobserver-sample"
                },
                "spec": {
                  "manageSDINodeConfig": true,
                  "sdiNamespace": "sdi",
                  "sdiVSystemRoute": {
                    "managementState": "Managed"
                  },
                  "slcbNamespace": "sap-slcbridge",
                  "slcbRoute": {
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/redhat-sap-cop/sdi-observer-operator:0.1.5
          description: Configure and enable SAP Data Intelligence to run on OpenShift
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SDIObserver is the Schema for the sdiobservers API
            displayName: SDIObserver
            kind: SDIObserver
            name: sdiobservers.sdi.sap-redhat.io
            version: v1alpha1
        description: Operator for monitoring SAP Data Intelligence (SDI) namespace
          and modifying objects in there that enable running of SDI on top of OpenShift.
          The observer shall be run in a dedicated namespace. It must be deployed
          before the SDI installation is started. This is the prerequisite for the
          successful installation of the SDI on OpenShift. Please refer to https://access.redhat.com/articles/5100521
          for detailed information.
        displayName: SAP Data Intelligence 3 - Observer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sdi
        - sap data intelligence
        links:
        - name: SAP Data Intelligence 3 - Observer Operator
          url: https://github.com/redhat-sap/sap-data-intelligence/tree/master/observer-operator
        maintainers:
        - email: mjiao@redhat.com
          name: Manjun Jiao
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/redhat-sap-cop/sdi-observer-operator:0.1.5
        version: 0.0.3
      entries:
      - name: observer-operator.v0.0.3
        version: 0.0.3
      name: alpha
    - currentCSV: observer-operator.v1.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sdi.sap-redhat.io/v1alpha1",
                "kind": "SDIObserver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/part-of": "observer-operator"
                  },
                  "name": "sdiobserver-sample",
                  "namespace": "sdi-observer"
                },
                "spec": {
                  "manageSDINodeConfig": true,
                  "sdiNamespace": "sdi",
                  "sdiVSystemRoute": {
                    "managementState": "Managed"
                  },
                  "slcbNamespace": "sap-slcbridge",
                  "slcbRoute": {
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/redhat-sap-cop/sdi-observer-operator:0.1.32
          createdAt: "20240726"
          description: Configure and enable SAP Data Intelligence 3 to run on OpenShift
            4
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mjiao/sap-data-intelligence/tree/master/observer-operator
          support: RedHat-SAPLinuxLab
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SDIObserver is the Schema for the sdiobservers API
            displayName: SDIObserver
            kind: SDIObserver
            name: sdiobservers.sdi.sap-redhat.io
            version: v1alpha1
        description: Operator for monitoring SAP Data Intelligence (SDI) namespace
          and modifying objects in there that enable running of SDI on top of OpenShift.
          It must be deployed before the SDI installation is started. This is the
          prerequisite for the successful installation of the SDI on OpenShift. Please
          refer to https://access.redhat.com/articles/7042265#ocp-install-sdi-observer-operator
          for detailed information.
        displayName: SAP Data Intelligence 3 - Observer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sdi
        - sap data intelligence
        links:
        - name: SAP Data Intelligence 3 - Observer Operator
          url: https://github.com/redhat-sap/sap-data-intelligence/tree/master/observer-operator
        maintainers:
        - email: mjiao@redhat.com
          name: Manjun Jiao
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/redhat-sap-cop/sdi-observer-operator:0.1.32
        version: 1.0.9
      entries:
      - name: observer-operator.v1.0.9
        version: 1.0.9
      - name: observer-operator.v1.0.8
        version: 1.0.8
      - name: observer-operator.v1.0.7
        version: 1.0.7
      - name: observer-operator.v1.0.6
        version: 1.0.6
      - name: observer-operator.v1.0.5
        version: 1.0.5
      - name: observer-operator.v1.0.4
        version: 1.0.4
      - name: observer-operator.v1.0.3
        version: 1.0.3
      - name: observer-operator.v1.0.2
        version: 1.0.2
      - name: observer-operator.v1.0.1
        version: 1.0.1
      - name: observer-operator.v0.0.3
        version: 0.0.3
      name: stable
    defaultChannel: stable
    packageName: sap-data-intelligence-observer-operator
    provider:
      name: SAP LinuxLab - Red Hat
      url: https://github.com/sap-linuxlab
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: node-healthcheck-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-healthcheck-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "maxUnhealthy": "49%",
                  "remediationTemplate": {
                    "apiVersion": "poison-pill.medik8s.io/v1alpha1",
                    "kind": "PoisonPillRemediationTemplate",
                    "name": "group-x",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-healthcheck-operator:0.1.0-latest
          createdAt: "2021-04-20 12:00:00"
          description: An operator to monitor node healthcheck and remdiate via 3rd
            party providers like poison-pill.
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: This operator deploys the Node Health Check controller and its
          artifacts
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://node-healthcheck-operator.domain
        maintainers:
        - email: medik8s@googlegroups.com
          name: medik8s team
        maturity: alpha
        provider:
          name: medik8s
          url: https://github.com/medik8s
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/node-healthcheck-operator:0.1.0-latest
        version: 0.1.1
      entries:
      - name: node-healthcheck-operator.v0.1.1
        version: 0.1.1
      name: alpha
    - currentCSV: node-healthcheck-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-healthcheck-operator:v0.3.0
          createdAt: "2022-08-24 11:40:38"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.2.0 <0.3.0'
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: medik8s team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: medik8s
          url: https://github.com/medik8s
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/node-healthcheck-operator:v0.3.0
        version: 0.3.0
      entries:
      - name: node-healthcheck-operator.v0.3.0
        version: 0.3.0
      name: candidate
    - currentCSV: node-healthcheck-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: quay.io/medik8s/node-healthcheck-operator:v0.6.1
          createdAt: "2023-11-14T15:45:34Z"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.1.0 <0.6.1'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/medik8s/node-healthcheck-operator:v0.6.1
        - quay.io/medik8s/node-remediation-console:v0.6.1
        version: 0.6.1
      entries:
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      - name: node-healthcheck-operator.v0.6.0
        version: 0.6.0
      - name: node-healthcheck-operator.v0.5.0
        version: 0.5.0
      - name: node-healthcheck-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: node-healthcheck-operator
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: GitLab Inc
      provider-url: https://about.gitlab.com/
    name: gitlab-operator-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "gitlab": {
                        "gitlab-shell": {
                          "maxReplicas": 2,
                          "minReplicas": 1
                        },
                        "sidekiq": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1000M"
                            }
                          }
                        },
                        "webservice": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1500M"
                            }
                          }
                        }
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "installCertmanager": false,
                      "minio": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      },
                      "nginx-ingress": {
                        "controller": {
                          "electionID": "ingress-controller-leader",
                          "replicaCount": 1
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      },
                      "redis": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-29T12:42:23Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      name: stable
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "gitlab": {
                        "gitlab-shell": {
                          "maxReplicas": 2,
                          "minReplicas": 1
                        },
                        "sidekiq": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1000M"
                            }
                          }
                        },
                        "webservice": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1500M"
                            }
                          }
                        }
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "installCertmanager": false,
                      "minio": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      },
                      "nginx-ingress": {
                        "controller": {
                          "electionID": "ingress-controller-leader",
                          "replicaCount": 1
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      },
                      "redis": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-29T12:42:23Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      name: unstable
    defaultChannel: stable
    packageName: gitlab-operator-kubernetes
    provider:
      name: GitLab Inc
      url: https://about.gitlab.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: edb-hcp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: edb-hcp-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edbpgai.edb.com/v1alpha1",
                "kind": "HybridControlPlane",
                "metadata": {
                  "name": "placeholder"
                },
                "spec": {
                  "beaconAgent": {
                    "applianceProviderEnabled": true,
                    "beaconDBConnected": true,
                    "beaconServerAddress": "upm-beacon-server-ingestion.upm-beacon.svc.cluster.local:8080",
                    "inCluster": true,
                    "location": "devspatcher",
                    "logLevel": "debug",
                    "plaintext": true,
                    "providers": [
                      "appliance"
                    ],
                    "provisioning": {
                      "imageDiscovery": true,
                      "imagesetDiscoveryAuthenticationType": "token",
                      "imagesetDiscoveryContainerRegistryURL": "placeholder",
                      "loadBalancersEnabled": true,
                      "nodePortDomain": "nodeportdomain.com",
                      "openshift": true,
                      "provider": "aws"
                    },
                    "tlsInsecure": false,
                    "transparentDataEncryptionMethods": [
                      "passphrase",
                      "hashicorp_vault"
                    ]
                  },
                  "beaconServer": {
                    "logLevel": "debug"
                  },
                  "componentsParameters": {
                    "edb-migration-portal": {
                      "enable_pdb": "false"
                    },
                    "transporter-data-operator": {
                      "fips_enabled": "false"
                    },
                    "transporter-dp-agent": {
                      "rw_service_url": "placeholder"
                    },
                    "transporter-rw-service": {
                      "domain_name": "placeholder"
                    },
                    "upm-api-pnxt-metric": {
                      "log_level": "info",
                      "tsdb_host": "thanos-query.monitoring.svc.cluster.local",
                      "tsdb_port": "9090"
                    },
                    "upm-api-pnxt-pd": {
                      "log_level": "info"
                    },
                    "upm-beaco-ff-base": {
                      "usage_generator_interval": "24h"
                    },
                    "upm-beaco-usages": {
                      "usage_generator_interval": "24h"
                    },
                    "upm-beacon": {
                      "beacon_db_hibernation": "\"off\"",
                      "beacon_server_replicas": "1",
                      "external_metrics_replicas": "1",
                      "image_discovery_credentials_secret_name": "placeholder",
                      "server_host": "placeholder"
                    },
                    "upm-istio": {
                      "ingress_grpc_tls_node_port": "30290",
                      "ingress_http_node_port": "32542",
                      "ingress_https_node_port": "30288"
                    },
                    "upm-istio-gateway": {
                      "cookie_aeskey": "rzkutHl8NJNztPMEJYykZouHslNiA7xmIXH+58ISUVo=",
                      "ingress_grpc_tls_node_port": "30290",
                      "ingress_http_node_port": "32542",
                      "ingress_https_node_port": "30288",
                      "openshift_console_domain_name": "placeholder"
                    },
                    "upm-prometheus": {
                      "cpu": "200m",
                      "memory": "6Gi",
                      "prometheus_retention_size": "15GB",
                      "prometheus_storage_size": "20Gi"
                    },
                    "upm-thanos": {
                      "memcached_cpu": "500m",
                      "memcached_memory": "2Gi",
                      "query_frontend_cpu": "200m",
                      "query_frontend_memory": "128Mi"
                    },
                    "upm-trust-manager": {
                      "cert_manager_namespace": "cert-manager"
                    }
                  },
                  "flavour": "placeholder",
                  "globalParameters": {
                    "default_ingress_domain": "placeholder",
                    "internal_backup_folder": "a7462dbc7106",
                    "portal_certificate_secret": "beaco-istio-gw-cert",
                    "portal_domain_name": "placeholder",
                    "portal_port": "443",
                    "storage_class": "placeholder"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "placeholder",
                      "namespace": "default"
                    }
                  ],
                  "imageRegistry": "placeholder",
                  "pgai": {
                    "portal": {
                      "authentication": {
                        "clientSecret": "edb-postgres-ai-secret",
                        "idpConnectors": [],
                        "staticPasswords": [
                          {
                            "email": "owner@mycompany.com",
                            "hash": "$2y$10/JLTqiuGblH.RemIutWxkztN5m",
                            "userID": "c5998173-a605-449a-a9a5-4a9c33e26df7",
                            "username": "Owner MyCompany"
                          }
                        ]
                      }
                    }
                  },
                  "source": {
                    "useLocalKustomizations": true
                  },
                  "version": "placeholder"
                }
              }
            ]
          capabilities: Basic Install
          categories: Database,AI/Machine Learning
          containerImage: docker.enterprisedb.com/pgai-platform/edb-hcp-operator/manager@sha256:82610c624e956a67df4858929e1aabd561a7537ba68d40e6512b68f55e2649e0
          createdAt: "2025-06-17T11:00:42Z"
          description: Unify transactional, analytical, and AI workloads and bring
            cloud agility to your hybrid data infrastructure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/EnterpriseDB/edb-hcp-operator
          support: techsupport@enterprisedb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HybridControlPlane is the Schema for the hybridcontrolplanes
              API
            displayName: Hybrid Control Plane
            kind: HybridControlPlane
            name: hybridcontrolplanes.edbpgai.edb.com
            version: v1alpha1
        description: |
          EDB Postgres AI is a sovereign data and AI platform that brings analytics and AI closer to core operational data and enables cloud agility in private environments.
          Unify your structured and unstructured data in one multi-model data platform with centralized management and observability to:
          - Accelerate enterprise-grade transactional databases
          - Enable rapid analytics and powerful lakehouse integrations
          - Automate AI pipelines and semantic search for faster GenAI development
          Using EDB’s unmatched Postgres expertise, EDB Postgres AI automates database operations, ensuring expert-level execution across deployments – boosting operational efficiency by up to 30% and accelerating innovation.
        displayName: EDB Postgres AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Postgres
        - Database
        - DBaaS
        - EDB Postgres AI
        links:
        - name: EDB Postgres AI Software Deployment
          url: https://www.enterprisedb.com/products/software-deployment
        maintainers:
        - email: techsupport@enterprisedb.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - docker.enterprisedb.com/pgai-platform/edb-hcp-operator/manager@sha256:82610c624e956a67df4858929e1aabd561a7537ba68d40e6512b68f55e2649e0
        version: 1.2.0
      entries:
      - name: edb-hcp-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: edb-hcp-operator
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-documentdb-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-documentdb-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBInstance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/documentdb-controller:1.1.1
          createdAt: "2025-09-19T22:13:37Z"
          description: AWS DocumentDB controller is a service controller for managing
            DocumentDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DBCluster represents the state of an AWS documentdb DBCluster
              resource.
            displayName: DBCluster
            kind: DBCluster
            name: dbclusters.documentdb.services.k8s.aws
            version: v1alpha1
          - description: DBInstance represents the state of an AWS documentdb DBInstance
              resource.
            displayName: DBInstance
            kind: DBInstance
            name: dbinstances.documentdb.services.k8s.aws
            version: v1alpha1
          - description: DBSubnetGroup represents the state of an AWS documentdb DBSubnetGroup
              resource.
            displayName: DBSubnetGroup
            kind: DBSubnetGroup
            name: dbsubnetgroups.documentdb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon DocumentDB resources in AWS from within your Kubernetes cluster.

          **About Amazon DocumentDB**

          Amazon DocumentDB (with MongoDB compatibility) is a fast, reliable, and fully managed database service. Amazon DocumentDB makes it easy to set up, operate, and scale MongoDB-compatible databases in the cloud. With Amazon DocumentDB, you can run the same application code and use the same drivers and tools that you use with MongoDB.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon DocumentDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - documentdb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon DocumentDB Developer Resources
          url: https://aws.amazon.com/documentdb/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: documentdb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/documentdb-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-documentdb-controller.v1.1.1
        version: 1.1.1
      - name: ack-documentdb-controller.v1.1.0
        version: 1.1.0
      - name: ack-documentdb-controller.v1.0.14
        version: 1.0.14
      - name: ack-documentdb-controller.v1.0.13
        version: 1.0.13
      - name: ack-documentdb-controller.v1.0.12
        version: 1.0.12
      - name: ack-documentdb-controller.v1.0.11
        version: 1.0.11
      - name: ack-documentdb-controller.v1.0.10
        version: 1.0.10
      - name: ack-documentdb-controller.v1.0.9
        version: 1.0.9
      - name: ack-documentdb-controller.v1.0.7
        version: 1.0.7
      - name: ack-documentdb-controller.v1.0.6
        version: 1.0.6
      - name: ack-documentdb-controller.v1.0.5
        version: 1.0.5
      - name: ack-documentdb-controller.v1.0.2
        version: 1.0.2
      - name: ack-documentdb-controller.v1.0.1
        version: 1.0.1
      - name: ack-documentdb-controller.v1.0.0
        version: 1.0.0
      - name: ack-documentdb-controller.v0.0.9
        version: 0.0.9
      - name: ack-documentdb-controller.v0.0.8
        version: 0.0.8
      - name: ack-documentdb-controller.v0.0.7
        version: 0.0.7
      - name: ack-documentdb-controller.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: ack-documentdb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: k8gb.io
      provider-url: https://github.com/k8gb-io/k8gb
    name: k8gb
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k8gb.v0.11.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"k8gb.absa.oss/v1beta1","kind":"Gslb","metadata":{"name":"test-gslb-failover","namespace":"test-gslb"},"spec":{"ingress":{"rules":[{"host":"failover.test.k8gb.io","http":{"paths":[{"backend":{"service":{"name":"frontend-podinfo","port":{"name":"http"}}},"path":"/"}]}}]},"strategy":{"primaryGeoTag":"eu-west-1","type":"failover"}}}]'
          capabilities: Seamless Upgrades
          categories: Networking
          certified: "false"
          containerImage: docker.io/absaoss/k8gb:v0.11.4
          createdAt: "2021-09-24 12:00:00"
          description: A cloud native Kubernetes Global Balancer
          operatorframework.io/suggested-namespace: k8gb
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: k8gb
          operators.operatorframework.io/builder: operator-sdk-v1.12.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/k8gb-io/k8gb
          support: cncf-k8gb-maintainers@lists.cncf.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gslb resource for global load balancing strategy configuration
            displayName: Gslb
            kind: Gslb
            name: gslbs.k8gb.absa.oss
            version: v1beta1
          - description: Using ExternalDNS it synchronizes exposed Kubernetes Services
              and Ingresses with DNS providers
            displayName: DNSEndpoint
            kind: DNSEndpoint
            name: dnsendpoints.externaldns.k8s.io
            version: v1alpha1
        description: |
          **A cloud native Kubernetes Global Balancer**

          A Global Service Load Balancing solution with a focus on having cloud native qualities and work natively in a Kubernetes context.

          Key Differentiators:
           - Load balancing is based on timeproof DNS protocol which is perfect for global scope and extremely reliable
           - No dedicated management cluster and no single point of failure
           - Kubernetes native application health checks utilizing status of Liveness and Readiness probes for load balancing decisions
           - Configuration with a single Kubernetes CRD of `Gslb` kind

          Operator needs to be configured by setting couple of environment variables. This can be done by modifying
          the subscription yaml in the web ui or via kubectl. Here is an example of changing the `EDGE_DNS_ZONE`:

          ```yaml
          kind: Subscription
            ...
          spec:
            ...
            config:
              env:
              - name: CLUSTER_GEO_TAG
                value: us
              - name: EDGE_DNS_ZONE
                value: mycloud.example.com
                ...
           ```
           For more deployment variables see [helm code](https://github.com/k8gb-io/k8gb/blob/v0.8.3/chart/k8gb/templates/operator.yaml#L53:L68).

           For more information see [k8gb.io](https://k8gb.io).
        displayName: k8gb
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gslb
        - dns-lb
        - kuberneters-global-balancer
        - kubernetes-operator
        - balancer
        - multi-cluster
        links:
        - name: Home
          url: https://www.k8gb.io/
        - name: Source
          url: https://github.com/k8gb-io/k8gb
        maintainers:
        - email: dinar.valeev@absa.africa
          name: Dinar Valeev
        - email: jiri.kremser@gmail.com
          name: Jiri Kremser
        - email: kuritka@gmail.com
          name: Michal Kuritka
        - email: timofey.ilinykh@absa.africa
          name: Timofey Ilinykh
        - email: yury.tsarev@absa.africa
          name: Yury Tsarev
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: k8gb.io
          url: https://github.com/k8gb-io/k8gb
        relatedImages:
        - absaoss/k8s_crd:v0.0.11
        - docker.io/absaoss/k8gb:v0.11.4
        version: 0.11.4
      entries:
      - name: k8gb.v0.11.4
        version: 0.11.4
      name: alpha
    defaultChannel: alpha
    packageName: k8gb
    provider:
      name: k8gb.io
      url: https://github.com/k8gb-io/k8gb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Internal
      provider-url: redhat.com
    name: l5-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: l5-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pets.bestie.com/v1",
                "kind": "Bestie",
                "metadata": {
                  "name": "bestie"
                },
                "spec": {
                  "image": "quay.io/opdev/bestie",
                  "maxReplicas": 10,
                  "size": 3,
                  "version": "1.3"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Application Runtime
          certified: "false"
          containerImage: quay.io/opdev/l5-operator:v0.0.3
          createdAt: "2023-06-27T16:56:10Z"
          description: L5 Operator is a demo operator that aims to demonstrate level
            5 capabilities for web applications.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          relatedImages: quay.io/opdev/bestie@sha256:0e99ada2b1de19043845dd9076789584aeddc2646402e5f0987dd00424be1ee7
          repository: https://github.com/opdev/l5-operator-demo
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Bestie is the Schema for the besties API.
            displayName: Bestie
            kind: Bestie
            name: besties.pets.bestie.com
            version: v1
        description: |
          ## Usage
          The L5 operator is meant to demonstrate a minimum viable prototype of all capability levels outilned in the operator framework. The L5 operator is compatible with both Kubernetes and OpenShift clusters. When running on vanilla kubernetes a few prerequisites need to be kept in mind which may not be available out of the box. These include an ingress controller and a prometheus operator for level 4 capability. The L5 operator also depends on the Crunchy Data PostgreSQL operator. On openshift when installing via the Operator Lifecycle Manager, this will automatically be installed when installing the operator via the embdeded operator hub. The Operator Lifecycle Manager is also available for vanilla kubernetes however it may not be installed out of the box. In this case the Crunchy Data PostgreSQL operator must be installed manually. The Crunchy Data PostgreSQL operator can be installed on vanilla kuberentes using the [instructions provided by Crunchy Data](https://access.crunchydata.com/documentation/postgres-operator/4.7.5/installation/postgres-operator/) More documentation and examples can be found on [github](https://github.com/opdev/l5-operator-demo).
        displayName: L5 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - l5-operator
        links:
        - name: L5 Operator
          url: https://opdev.github.io/l5-operator-demo/
        maintainers:
        - email: partnereng@redhat.com
          name: OpDev Team
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat Internal
          url: redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/opdev/l5-operator:v0.0.3
        version: 0.0.3
      entries:
      - name: l5-operator.v0.0.3
        version: 0.0.3
      - name: l5-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: l5-operator
    provider:
      name: Red Hat Internal
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
    name: aikit-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aikit-operator.v2021.2.102120
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aikit.intel/v1alpha1",
                "kind": "AIKitContainer",
                "metadata": {
                  "name": "intel-aikit-container",
                  "namespace": "redhat-ods-applications"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagestream": {
                    "name": "oneapi-aikit",
                    "namespace": "redhat-ods-applications",
                    "registry": {
                      "root": "registry.connect.redhat.com",
                      "repo": "intel",
                      "name": "oneapi-aikit",
                      "version": "latest"
                    }
                  },
                  "nameOverride": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AIKitContainer
            name: aikitcontainers.aikit.intel
            version: v1alpha1
        description: |
          The Intel® oneAPI AI Analytics Toolkit gives data scientists, AI developers, and researchers familiar Python* tools and frameworks to accelerate end-to-end data science and machine learning pipelines on Intel® architectures. The components are built using oneAPI libraries for low-level compute optimizations. This toolkit maximizes performance from preprocessing through machine and deep learning phases and provides interoperability for efficient model development and deployment across single and multinodes.
        displayName: Intel® oneAPI AI Analytics Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aikit-operator
        links:
        - name: Aikit Operator
          url: https://aikit-operator.domain
        maintainers:
        - email: kam.d.kasravi@intel.com
          name: Kam Kasravi
        - email: abolfazl.shahbazi@intel.com
          name: Abolfazl Shahbazi
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Intel
          url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
        relatedImages:
        - registry.connect.redhat.com/intel/aikit-operator@sha256:7abb5e6164ffb3e4b87c39061f45d09848218fcfb74b48bf5e1035d56b41ee0c
        version: 2021.2.102120
      entries:
      - name: aikit-operator.v2021.2.102120
        version: 2021.2.102120
      - name: aikit-operator.v2021.2.100410
        version: 2021.2.100410
      - name: aikit-operator.v2021.2.92911
        version: 2021.2.92911
      - name: aikit-operator.v2021.2.92814
        version: 2021.2.92814
      - name: aikit-operator.v2021.2.92808-dev-a60e5ba
        version: 2021.2.92808-dev-a60e5ba
      name: alpha
    defaultChannel: alpha
    packageName: aikit-operator
    provider:
      name: Intel
      url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-airflow-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airflow-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:d8f8c439f33e5a80e16c7867a008d71693fd4928357ac6919c77bd0ff0bb1339
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:d8f8c439f33e5a80e16c7867a008d71693fd4928357ac6919c77bd0ff0bb1339
        version: 24.11.1
      entries:
      - name: airflow-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: airflow-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:173363db3cb0997aff09fe4b7f3a54c04f22bc56db83d2b30757e3cc745c4587
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:173363db3cb0997aff09fe4b7f3a54c04f22bc56db83d2b30757e3cc745c4587
        version: 25.3.0
      entries:
      - name: airflow-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: airflow-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
        version: 25.7.0
      entries:
      - name: airflow-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: airflow-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
        version: 25.7.0
      entries:
      - name: airflow-operator.v25.7.0
        version: 25.7.0
      - name: airflow-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-airflow-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: windows-machine-config-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: windows-machine-config-operator.v10.16.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          createdAt: "2025-06-06T22:08:07Z"
          description: An operator that enables Windows container workloads on OCP
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=10.15.0 <10.16.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.openshift.io/valid-subscription: '["Red Hat OpenShift support
            for Windows Containers"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ### Introduction
          The Windows Machine Config Operator configures Windows Machines into nodes, enabling Windows container workloads to
          be run on OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/4.16/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://docs.openshift.com/container-platform/4.16/windows_containers/byoh-windows-instance.html)
          The operator completes all the necessary steps to configure the Windows instance so that it can join the cluster as a worker node.

          Usage of this operator requires an OpenShift Container Platform for Windows Containers subscription. Users looking to
          deploy Windows containers workloads in production clusters should acquire a subscription before attempting to
          install this operator. Users without a subscription can try the community operator, a distribution which lacks
          official support.

          ### Pre-requisites
          * A Red Hat OpenShift Container Platform for Windows Containers [subscription](https://access.redhat.com/support/policy/updates/openshift#windows)
          * OCP 4.16 cluster running on Azure, AWS, GCP or vSphere configured with hybrid OVN Kubernetes networking
          * [WMCO prerequisites](https://docs.openshift.com/container-platform/4.16/windows_containers/understanding-windows-container-workloads.html#wmco-prerequisites__understanding-windows-container-workloads)

          ### Usage
          Once the `openshift-windows-machine-config-operator` namespace has been created, a secret must be created containing
          the private key that will be used to access the Windows VMs:
          ```
          # Create secret containing the private key in the openshift-windows-machine-config-operator namespace
          oc create secret generic cloud-private-key --from-file=private-key.pem=/path/to/key -n openshift-windows-machine-config-operator
          ```
          We strongly recommend not using the same
          [private key](https://docs.openshift.com/container-platform/4.16/installing/installing_azure/installing-azure-default.html#ssh-agent-using_installing-azure-default)
          used when installing the cluster

          Below is an example of a vSphere Windows MachineSet which can create Windows Machines that the WMCO can react upon.
          Please note that the windows-user-data secret will be created by the WMCO lazily when it is configuring the first
          Windows Machine. After that, the windows-user-data will be available for the subsequent MachineSets to be consumed.
          It might take around 10 minutes for the Windows VM to be configured so that it joins the cluster. Please note that
          the MachineSet should have following labels:
          * `machine.openshift.io/os-id: Windows`
          * `machine.openshift.io/cluster-api-machine-role: worker`
          * `machine.openshift.io/cluster-api-machine-type: worker`

          The following label must be added to the Machine spec within the MachineSet spec:
          * `node-role.kubernetes.io/worker: ""`

          Not having these labels will result in the Windows node not being marked as a worker.

          `<infrastructureID>` should be replaced with the output of:
          ```
          oc get -o jsonpath='{.status.infrastructureName}{"\n"}' infrastructure cluster
          ```

          The following template variables need to be replaced as follows with values from your vSphere environment:
          * `<Windows_VM_template>`: template name
          * `<VM Network Name>`: network name
          * `<vCenter DataCenter Name>`: datacenter name
          * `<Path to VM Folder in vCenter>`: path where your OpenShift cluster is running
          * `<vCenter Datastore Name>`: datastore name
          * `<vCenter Server FQDN/IP>`: IP address or FQDN of the vCenter server

          Please note that on vSphere, Windows Machine names cannot be more than 15 characters long. The MachineSet name therefore
          cannot be more than 9 characters long, due to the way Machine names are generated from it.
          ```
          apiVersion: machine.openshift.io/v1beta1
          kind: MachineSet
          metadata:
            labels:
              machine.openshift.io/cluster-api-cluster: <infrastructureID>
            name: winworker
            namespace: openshift-machine-api
          spec:
            replicas: 1
            selector:
              matchLabels:
                machine.openshift.io/cluster-api-cluster: <infrastructureID>
                machine.openshift.io/cluster-api-machineset: winworker
            template:
              metadata:
                labels:
                  machine.openshift.io/cluster-api-cluster: <infrastructureID>
                  machine.openshift.io/cluster-api-machine-role: worker
                  machine.openshift.io/cluster-api-machine-type: worker
                  machine.openshift.io/cluster-api-machineset: winworker
                  machine.openshift.io/os-id: Windows
              spec:
                metadata:
                  labels:
                    node-role.kubernetes.io/worker: ""
                providerSpec:
                  value:
                    apiVersion: vsphereprovider.openshift.io/v1beta1
                    credentialsSecret:
                      name: vsphere-cloud-credentials
                      namespace: openshift-machine-api
                    diskGiB: 128
                    kind: VSphereMachineProviderSpec
                    memoryMiB: 16384
                    metadata:
                      creationTimestamp: null
                    network:
                      devices:
                      - networkName: "<VM Network Name>"
                    numCPUs: 4
                    numCoresPerSocket: 1
                    snapshot: ""
                    template: <Windows_VM_template
                    userDataSecret:
                      name: windows-user-data
                    workspace:
                      datacenter: <vCenter DataCenter Name>
                      datastore: <vCenter Datastore Name>
                      folder: <Path to VM Folder in vCenter> # e.g. /DC/vm/ocp45-2tdrm
                      server: <vCenter Server FQDN/IP>
          ```
          Example MachineSet for other cloud providers:
          - [AWS](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-aws.html)
          - [Azure](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-azure.html)
          - [GCP](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-gcp.html)

          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.

          ### Reporting issues
          Support for this distribution of WMCO requires a Red Hat OpenShfit subscription. Support should be requested through the Red Hat Customer Portal.
          Please read through the [troubleshooting document](https://docs.openshift.com/container-platform/4.16/support/troubleshooting/troubleshooting-windows-container-workload-issues.html)
          before opening a support case.
        displayName: Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4-wincw/windows-machine-config-rhel9-operator@sha256:c9a6166158e1ccb1764007d79c6e0f879505c3546eb45cbbe79ebeefcc78a728
        version: 10.16.2
      entries:
      - name: windows-machine-config-operator.v10.16.2
        version: 10.16.2
      - name: windows-machine-config-operator.v10.16.1
        version: 10.16.1
      - name: windows-machine-config-operator.v10.16.0
        version: 10.16.0
      name: preview
    - currentCSV: windows-machine-config-operator.v10.16.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          createdAt: "2025-06-06T22:08:07Z"
          description: An operator that enables Windows container workloads on OCP
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=10.15.0 <10.16.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.openshift.io/valid-subscription: '["Red Hat OpenShift support
            for Windows Containers"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ### Introduction
          The Windows Machine Config Operator configures Windows Machines into nodes, enabling Windows container workloads to
          be run on OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/4.16/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://docs.openshift.com/container-platform/4.16/windows_containers/byoh-windows-instance.html)
          The operator completes all the necessary steps to configure the Windows instance so that it can join the cluster as a worker node.

          Usage of this operator requires an OpenShift Container Platform for Windows Containers subscription. Users looking to
          deploy Windows containers workloads in production clusters should acquire a subscription before attempting to
          install this operator. Users without a subscription can try the community operator, a distribution which lacks
          official support.

          ### Pre-requisites
          * A Red Hat OpenShift Container Platform for Windows Containers [subscription](https://access.redhat.com/support/policy/updates/openshift#windows)
          * OCP 4.16 cluster running on Azure, AWS, GCP or vSphere configured with hybrid OVN Kubernetes networking
          * [WMCO prerequisites](https://docs.openshift.com/container-platform/4.16/windows_containers/understanding-windows-container-workloads.html#wmco-prerequisites__understanding-windows-container-workloads)

          ### Usage
          Once the `openshift-windows-machine-config-operator` namespace has been created, a secret must be created containing
          the private key that will be used to access the Windows VMs:
          ```
          # Create secret containing the private key in the openshift-windows-machine-config-operator namespace
          oc create secret generic cloud-private-key --from-file=private-key.pem=/path/to/key -n openshift-windows-machine-config-operator
          ```
          We strongly recommend not using the same
          [private key](https://docs.openshift.com/container-platform/4.16/installing/installing_azure/installing-azure-default.html#ssh-agent-using_installing-azure-default)
          used when installing the cluster

          Below is an example of a vSphere Windows MachineSet which can create Windows Machines that the WMCO can react upon.
          Please note that the windows-user-data secret will be created by the WMCO lazily when it is configuring the first
          Windows Machine. After that, the windows-user-data will be available for the subsequent MachineSets to be consumed.
          It might take around 10 minutes for the Windows VM to be configured so that it joins the cluster. Please note that
          the MachineSet should have following labels:
          * `machine.openshift.io/os-id: Windows`
          * `machine.openshift.io/cluster-api-machine-role: worker`
          * `machine.openshift.io/cluster-api-machine-type: worker`

          The following label must be added to the Machine spec within the MachineSet spec:
          * `node-role.kubernetes.io/worker: ""`

          Not having these labels will result in the Windows node not being marked as a worker.

          `<infrastructureID>` should be replaced with the output of:
          ```
          oc get -o jsonpath='{.status.infrastructureName}{"\n"}' infrastructure cluster
          ```

          The following template variables need to be replaced as follows with values from your vSphere environment:
          * `<Windows_VM_template>`: template name
          * `<VM Network Name>`: network name
          * `<vCenter DataCenter Name>`: datacenter name
          * `<Path to VM Folder in vCenter>`: path where your OpenShift cluster is running
          * `<vCenter Datastore Name>`: datastore name
          * `<vCenter Server FQDN/IP>`: IP address or FQDN of the vCenter server

          Please note that on vSphere, Windows Machine names cannot be more than 15 characters long. The MachineSet name therefore
          cannot be more than 9 characters long, due to the way Machine names are generated from it.
          ```
          apiVersion: machine.openshift.io/v1beta1
          kind: MachineSet
          metadata:
            labels:
              machine.openshift.io/cluster-api-cluster: <infrastructureID>
            name: winworker
            namespace: openshift-machine-api
          spec:
            replicas: 1
            selector:
              matchLabels:
                machine.openshift.io/cluster-api-cluster: <infrastructureID>
                machine.openshift.io/cluster-api-machineset: winworker
            template:
              metadata:
                labels:
                  machine.openshift.io/cluster-api-cluster: <infrastructureID>
                  machine.openshift.io/cluster-api-machine-role: worker
                  machine.openshift.io/cluster-api-machine-type: worker
                  machine.openshift.io/cluster-api-machineset: winworker
                  machine.openshift.io/os-id: Windows
              spec:
                metadata:
                  labels:
                    node-role.kubernetes.io/worker: ""
                providerSpec:
                  value:
                    apiVersion: vsphereprovider.openshift.io/v1beta1
                    credentialsSecret:
                      name: vsphere-cloud-credentials
                      namespace: openshift-machine-api
                    diskGiB: 128
                    kind: VSphereMachineProviderSpec
                    memoryMiB: 16384
                    metadata:
                      creationTimestamp: null
                    network:
                      devices:
                      - networkName: "<VM Network Name>"
                    numCPUs: 4
                    numCoresPerSocket: 1
                    snapshot: ""
                    template: <Windows_VM_template
                    userDataSecret:
                      name: windows-user-data
                    workspace:
                      datacenter: <vCenter DataCenter Name>
                      datastore: <vCenter Datastore Name>
                      folder: <Path to VM Folder in vCenter> # e.g. /DC/vm/ocp45-2tdrm
                      server: <vCenter Server FQDN/IP>
          ```
          Example MachineSet for other cloud providers:
          - [AWS](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-aws.html)
          - [Azure](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-azure.html)
          - [GCP](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-gcp.html)

          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.

          ### Reporting issues
          Support for this distribution of WMCO requires a Red Hat OpenShfit subscription. Support should be requested through the Red Hat Customer Portal.
          Please read through the [troubleshooting document](https://docs.openshift.com/container-platform/4.16/support/troubleshooting/troubleshooting-windows-container-workload-issues.html)
          before opening a support case.
        displayName: Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4-wincw/windows-machine-config-rhel9-operator@sha256:c9a6166158e1ccb1764007d79c6e0f879505c3546eb45cbbe79ebeefcc78a728
        version: 10.16.2
      entries:
      - name: windows-machine-config-operator.v10.16.2
        version: 10.16.2
      - name: windows-machine-config-operator.v10.16.1
        version: 10.16.1
      - name: windows-machine-config-operator.v10.16.0
        version: 10.16.0
      name: stable
    defaultChannel: stable
    packageName: windows-machine-config-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JohnSnowLABS
      provider-url: nlp.johnsnowlabs.com
    name: annotationlab
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: annotationlab.v5.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.johnsnowlabs.com/v1",
                "kind": "Annotationlab",
                "metadata": {
                  "name": "annotationlab"
                },
                "spec": {
                  "activelearning": {
                    "extraEnv": "",
                    "extraVolumeMounts": "",
                    "extraVolumes": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/johnsnowlabs/annotationlab",
                      "tag": "active-learning-5.7.0"
                    },
                    "lifecycleHooks": ""
                  },
                  "affinity": "podAntiAffinity:\n  requiredDuringSchedulingIgnoredDuringExecution:\n    - labelSelector:\n        matchLabels:\n          {{- include \"annotationlab.selectorLabels\" . | nindent 10 }}\n        matchExpressions:\n          - key: role\n            operator: NotIn\n            values:\n              - test\n      topologyKey: kubernetes.io/hostname\n  preferredDuringSchedulingIgnoredDuringExecution:\n    - weight: 100\n      podAffinityTerm:\n        labelSelector:\n          matchLabels:\n            {{- include \"annotationlab.selectorLabels\" . | nindent 12 }}\n          matchExpressions:\n            - key: role\n              operator: NotIn\n              values:\n                - test\n        topologyKey: failure-domain.beta.kubernetes.io/zone\n",
                  "airflow": {
                    "activelearning": {
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/johnsnowlabs/annotationlab",
                        "tag": "active-learning-5.7.0"
                      }
                    },
                    "affinity": {},
                    "airflowConfigAnnotations": {},
                    "airflowHome": "/opt/airflow",
                    "airflowLocalSettings": "{{- if semverCompare \"\u003e=2.2.0\" .Values.airflowVersion }}\n{{- if not (or .Values.webserverSecretKey .Values.webserverSecretKeySecretName) }}\nfrom airflow.www.utils import UIAlert\n\nDASHBOARD_UIALERTS = [\n  UIAlert(\n    'Usage of a dynamic webserver secret key detected. We recommend a static webserver secret key instead.'\n    ' See the \u003ca href='\n    '\"https://airflow.apache.org/docs/helm-chart/stable/production-guide.html#webserver-secret-key\"\u003e'\n    'Helm Chart Production Guide\u003c/a\u003e for more details.',\n    category=\"warning\",\n    roles=[\"Admin\"],\n    html=True,\n  )\n]\n{{- end }}\n{{- end }}",
                    "airflowPodAnnotations": {},
                    "airflowVersion": "2.5.1",
                    "allowPodLaunching": true,
                    "cleanup": {
                      "affinity": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow kubernetes cleanup-pods --namespace={{ .Release.Namespace }}"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "failedJobsHistoryLimit": null,
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "resources": {},
                      "schedule": "*/15 * * * *",
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "successfulJobsHistoryLimit": null,
                      "tolerations": [],
                      "topologySpreadConstraints": []
                    },
                    "config": {
                      "api": {
                        "auth_backend": "airflow.api.auth.backend.default"
                      },
                      "celery": {
                        "flower_url_prefix": "{{ .Values.ingress.flower.path }}",
                        "worker_concurrency": 16
                      },
                      "celery_kubernetes_executor": {
                        "kubernetes_queue": "kubernetes"
                      },
                      "core": {
                        "colored_console_log": "False",
                        "dags_are_paused_at_creation": "False",
                        "dags_folder": "{{ include \"airflow_dags\" . }}",
                        "executor": "{{ .Values.executor }}",
                        "load_examples": "False",
                        "remote_logging": "{{- ternary \"True\" \"False\" .Values.elasticsearch.enabled }}"
                      },
                      "elasticsearch": {
                        "json_format": "True",
                        "log_id_template": "{dag_id}_{task_id}_{execution_date}_{try_number}"
                      },
                      "elasticsearch_configs": {
                        "max_retries": 3,
                        "retry_timeout": "True",
                        "timeout": 30
                      },
                      "kerberos": {
                        "ccache": "{{ .Values.kerberos.ccacheMountPath }}/{{ .Values.kerberos.ccacheFileName }}",
                        "keytab": "{{ .Values.kerberos.keytabPath }}",
                        "principal": "{{ .Values.kerberos.principal }}",
                        "reinit_frequency": "{{ .Values.kerberos.reinitFrequency }}"
                      },
                      "kubernetes": {
                        "airflow_configmap": "{{ include \"airflow_config\" . }}",
                        "airflow_local_settings_configmap": "{{ include \"airflow_config\" . }}",
                        "multi_namespace_mode": "{{ ternary \"True\" \"False\" .Values.multiNamespaceMode }}",
                        "namespace": "{{ .Release.Namespace }}",
                        "pod_template_file": "{{ include \"airflow_pod_template_file\" . }}/pod_template_file.yaml",
                        "worker_container_repository": "{{ .Values.images.airflow.repository | default .Values.defaultAirflowRepository }}",
                        "worker_container_tag": "{{ .Values.images.airflow.tag | default .Values.defaultAirflowTag }}"
                      },
                      "logging": {
                        "colored_console_log": "False",
                        "remote_logging": "{{- ternary \"True\" \"False\" .Values.elasticsearch.enabled }}"
                      },
                      "metrics": {
                        "statsd_host": "{{ printf \"%s-statsd\" .Release.Name }}",
                        "statsd_on": "{{ ternary \"True\" \"False\" .Values.statsd.enabled }}",
                        "statsd_port": 9125,
                        "statsd_prefix": "airflow"
                      },
                      "scheduler": {
                        "run_duration": 41460,
                        "standalone_dag_processor": "{{ ternary \"True\" \"False\" .Values.dagProcessor.enabled }}",
                        "statsd_host": "{{ printf \"%s-statsd\" .Release.Name }}",
                        "statsd_on": "{{ ternary \"True\" \"False\" .Values.statsd.enabled }}",
                        "statsd_port": 9125,
                        "statsd_prefix": "airflow"
                      },
                      "webserver": {
                        "authenticate": "False",
                        "base_url": "http://0.0.0.0:8080",
                        "enable_proxy_fix": "True",
                        "rbac": "False"
                      }
                    },
                    "createUserJob": {
                      "affinity": {},
                      "annotations": {},
                      "applyCustomEnv": true,
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"users create\" \"create_user\" }} \"$@\"",
                        "--",
                        "-r",
                        "{{ .Values.webserver.defaultUser.role }}",
                        "-u",
                        "{{ .Values.webserver.defaultUser.username }}",
                        "-e",
                        "{{ .Values.webserver.defaultUser.email }}",
                        "-f",
                        "{{ .Values.webserver.defaultUser.firstName }}",
                        "-l",
                        "{{ .Values.webserver.defaultUser.lastName }}",
                        "-p",
                        "{{ .Values.webserver.defaultUser.password }}"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "jobAnnotations": {},
                      "labels": {},
                      "nodeSelector": {},
                      "resources": {},
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-create-user-job"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "useHelmHooks": true
                    },
                    "dagProcessor": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow dag-processor"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 60,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "dags": {
                      "gitSync": {
                        "branch": "v2-2-stable",
                        "containerName": "git-sync",
                        "depth": 1,
                        "enabled": false,
                        "env": [],
                        "extraVolumeMounts": [],
                        "maxFailures": 0,
                        "repo": "https://github.com/apache/airflow.git",
                        "resources": {},
                        "rev": "HEAD",
                        "securityContext": {},
                        "subPath": "tests/dags",
                        "uid": 65533,
                        "wait": 5
                      },
                      "persistence": {
                        "accessMode": "ReadWriteOnce",
                        "enabled": false,
                        "existingClaim": null,
                        "size": "1Gi",
                        "storageClassName": null,
                        "subPath": null
                      }
                    },
                    "data": {
                      "brokerUrl": null,
                      "brokerUrlSecretName": null,
                      "metadataConnection": {
                        "db": "airflow",
                        "pass": "airflow",
                        "port": 5432,
                        "protocol": "postgresql",
                        "sslmode": "disable",
                        "user": "airflow"
                      },
                      "metadataSecretName": null,
                      "resultBackendConnection": null,
                      "resultBackendSecretName": null
                    },
                    "defaultAirflowRepository": "apache/airflow",
                    "defaultAirflowTag": "2.5.1",
                    "elasticsearch": {
                      "connection": {},
                      "enabled": false,
                      "secretName": null
                    },
                    "enableBuiltInSecretEnvVars": {
                      "AIRFLOW_CONN_AIRFLOW_DB": true,
                      "AIRFLOW__CELERY__BROKER_URL": true,
                      "AIRFLOW__CELERY__CELERY_RESULT_BACKEND": true,
                      "AIRFLOW__CELERY__RESULT_BACKEND": true,
                      "AIRFLOW__CORE__FERNET_KEY": true,
                      "AIRFLOW__CORE__SQL_ALCHEMY_CONN": true,
                      "AIRFLOW__DATABASE__SQL_ALCHEMY_CONN": true,
                      "AIRFLOW__ELASTICSEARCH__ELASTICSEARCH_HOST": true,
                      "AIRFLOW__ELASTICSEARCH__HOST": true,
                      "AIRFLOW__WEBSERVER__SECRET_KEY": true
                    },
                    "env": [],
                    "executor": "CeleryExecutor",
                    "extraConfigMaps": {},
                    "extraEnv": "- name: ACTIVE_LEARNING_IMAGE_PULL_POLICY\n  value: \"{{ .Values.activelearning.image.pullPolicy }}\"\n- name: ACTIVE_LEARNING_IMAGE\n  value: \"{{ .Values.activelearning.image.repository }}:{{ .Values.activelearning.image.tag }}\"\n- name: NUMBER_OF_PREANNOTATOR_DAG\n  value: {{ .Values.preannotator_queues.count | quote }}\n- name: AIRFLOW__LOGGING__LOGGING_LEVEL\n  value: INFO\n- name: ANNOTATIONLAB_FULLNAME\n  value: {{ include \"annotationlab.annotationlab.fullname\" . }}\n- name: ANNOTATIONLAB_URL\n  value: http://{{ include \"annotationlab.annotationlab.fullname\" . }}:8200\n- name: AIRFLOW_NAMESPACE\n  value: {{ .Release.Namespace }}\n- name: DATABASE_CONNECTION_STRING\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"annotationlab.annotationlab.fullname\" . }}-secret\n      key: DATABASE_CONNECTION_STRING\n",
                    "extraEnvFrom": null,
                    "extraSecrets": {},
                    "fernetKey": null,
                    "fernetKeySecretName": null,
                    "flower": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"celery flower\" \"flower\" }}"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "networkPolicy": {
                        "ingress": {
                          "from": [],
                          "ports": [
                            {
                              "port": "{{ .Values.ports.flowerUI }}"
                            }
                          ]
                        }
                      },
                      "nodeSelector": {},
                      "password": null,
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "secretName": null,
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "loadBalancerIP": null,
                        "loadBalancerSourceRanges": [],
                        "ports": [
                          {
                            "name": "flower-ui",
                            "port": "{{ .Values.ports.flowerUI }}"
                          }
                        ],
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "username": null
                    },
                    "fullnameOverride": "",
                    "gid": 0,
                    "global": {},
                    "image": {
                      "repository": "docker.io/johnsnowlabs/annotationlab"
                    },
                    "images": {
                      "airflow": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": "dataflows-5.7.0"
                      },
                      "flower": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": null
                      },
                      "gitSync": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "k8s.gcr.io/git-sync/git-sync",
                        "tag": "v3.6.3"
                      },
                      "migrationsWaitTimeout": 60,
                      "pgbouncer": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "apache/airflow",
                        "tag": "airflow-pgbouncer-2021.04.28-1.14.0"
                      },
                      "pgbouncerExporter": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "apache/airflow",
                        "tag": "airflow-pgbouncer-exporter-2021.09.22-0.12.0"
                      },
                      "pod_template": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": null
                      },
                      "redis": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/redis",
                        "tag": "alpine3.16"
                      },
                      "statsd": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "quay.io/prometheus/statsd-exporter",
                        "tag": "v0.22.8"
                      },
                      "useDefaultImageForMigration": false
                    },
                    "ingress": {
                      "enabled": false,
                      "flower": {
                        "annotations": {},
                        "enabled": false,
                        "host": "",
                        "hosts": [],
                        "ingressClassName": "",
                        "path": "/",
                        "pathType": "ImplementationSpecific",
                        "tls": {
                          "enabled": false,
                          "secretName": ""
                        }
                      },
                      "web": {
                        "annotations": {},
                        "enabled": false,
                        "host": "",
                        "hosts": [],
                        "ingressClassName": "",
                        "path": "/",
                        "pathType": "ImplementationSpecific",
                        "precedingPaths": [],
                        "succeedingPaths": [],
                        "tls": {
                          "enabled": false,
                          "secretName": ""
                        }
                      }
                    },
                    "kerberos": {
                      "ccacheFileName": "cache",
                      "ccacheMountPath": "/var/kerberos-ccache",
                      "config": "# This is an example config showing how you can use templating and how \"example\" config\n# might look like. It works with the test kerberos server that we are using during integration\n# testing at Apache Airflow (see `scripts/ci/docker-compose/integration-kerberos.yml` but in\n# order to make it production-ready you must replace it with your own configuration that\n# Matches your kerberos deployment. Administrators of your Kerberos instance should\n# provide the right configuration.\n\n[logging]\ndefault = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kerberos_libs.log\"\nkdc = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kerberos_kdc.log\"\nadmin_server = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kadmind.log\"\n\n[libdefaults]\ndefault_realm = FOO.COM\nticket_lifetime = 10h\nrenew_lifetime = 7d\nforwardable = true\n\n[realms]\nFOO.COM = {\n  kdc = kdc-server.foo.com\n  admin_server = admin_server.foo.com\n}\n",
                      "configPath": "/etc/krb5.conf",
                      "enabled": false,
                      "keytabBase64Content": null,
                      "keytabPath": "/etc/airflow.keytab",
                      "principal": "airflow@FOO.COM",
                      "reinitFrequency": 3600
                    },
                    "kubeVersionOverride": "",
                    "labels": {},
                    "limits": [],
                    "logs": {
                      "persistence": {
                        "enabled": false,
                        "existingClaim": null,
                        "size": "100Gi",
                        "storageClassName": null
                      }
                    },
                    "migrateDatabaseJob": {
                      "affinity": {},
                      "annotations": {},
                      "applyCustomEnv": true,
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"db upgrade\" \"upgradedb\" }}"
                      ],
                      "command": null,
                      "enabled": true,
                      "extraContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "jobAnnotations": {},
                      "nodeSelector": {},
                      "resources": {},
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-migrate-database-job"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "useHelmHooks": true
                    },
                    "multiNamespaceMode": false,
                    "nameOverride": "",
                    "networkPolicies": {
                      "enabled": false
                    },
                    "nodeSelector": {},
                    "pgbouncer": {
                      "affinity": {},
                      "annotations": {},
                      "args": null,
                      "auth_file": "/etc/pgbouncer/users.txt",
                      "auth_type": "md5",
                      "ciphers": "normal",
                      "command": [
                        "pgbouncer",
                        "-u",
                        "nobody",
                        "/etc/pgbouncer/pgbouncer.ini"
                      ],
                      "configSecretName": null,
                      "enabled": false,
                      "extraIni": null,
                      "extraIniMetadata": null,
                      "extraIniResultBackend": null,
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "logConnections": 0,
                      "logDisconnections": 0,
                      "maxClientConn": 100,
                      "metadataPoolSize": 10,
                      "metricsExporterSidecar": {
                        "resources": {},
                        "sslmode": "disable"
                      },
                      "nodeSelector": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "resultBackendPoolSize": 5,
                      "revisionHistoryLimit": null,
                      "service": {
                        "extraAnnotations": {}
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "ssl": {
                        "ca": null,
                        "cert": null,
                        "key": null
                      },
                      "sslmode": "prefer",
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 65534,
                      "verbose": 0
                    },
                    "podTemplate": null,
                    "ports": {
                      "airflowUI": 8080,
                      "flowerUI": 5555,
                      "pgbouncer": 6543,
                      "pgbouncerScrape": 9127,
                      "redisDB": 6379,
                      "statsdIngest": 9125,
                      "statsdScrape": 9102,
                      "workerLogs": 8793
                    },
                    "postgresql": {
                      "auth": {
                        "enablePostgresUser": true,
                        "password": null,
                        "postgresPassword": "postgres",
                        "username": null
                      },
                      "common": {
                        "exampleValue": "common-chart",
                        "global": {
                          "postgresql": {}
                        }
                      },
                      "commonAnnotations": {},
                      "enabled": true,
                      "extraEnv": [],
                      "global": {
                        "postgresql": {}
                      },
                      "image": {
                        "debug": false,
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgresql",
                        "tag": "11"
                      },
                      "ldap": {
                        "baseDN": "",
                        "bindDN": "",
                        "enabled": false,
                        "port": "",
                        "prefix": "",
                        "scheme": "",
                        "search_attr": "",
                        "search_filter": "",
                        "server": "",
                        "suffix": "",
                        "tls": false,
                        "url": ""
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 30,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "master": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": [],
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "metrics": {
                        "enabled": false,
                        "extraEnvVars": {},
                        "image": {
                          "pullPolicy": "IfNotPresent",
                          "registry": "docker.io",
                          "repository": "bitnami/postgres-exporter",
                          "tag": "0.8.0-debian-10-r166"
                        },
                        "livenessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "prometheusRule": {
                          "additionalLabels": {},
                          "enabled": false,
                          "namespace": "",
                          "rules": []
                        },
                        "readinessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "securityContext": {},
                        "service": {
                          "annotations": {
                            "prometheus.io/port": "9187",
                            "prometheus.io/scrape": "true"
                          },
                          "type": "ClusterIP"
                        },
                        "serviceMonitor": {
                          "additionalLabels": {},
                          "enabled": false
                        }
                      },
                      "nameOverride": "airflow-postgresql",
                      "networkPolicy": {
                        "allowExternal": true,
                        "enabled": false,
                        "explicitNamespacesSelector": {}
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "enabled": true,
                        "mountPath": "/bitnami/postgresql",
                        "size": "8Gi",
                        "subPath": ""
                      },
                      "postgresqlDataDir": "/bitnami/postgresql/data",
                      "postgresqlDatabase": "airflow",
                      "postgresqlPassword": "airflow",
                      "postgresqlUsername": "airflow",
                      "psp": {
                        "create": false
                      },
                      "rbac": {
                        "create": false
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "replication": {
                        "applicationName": "my_application",
                        "enabled": false,
                        "numSynchronousReplicas": 0,
                        "password": "repl_password",
                        "slaveReplicas": 1,
                        "synchronousCommit": "off",
                        "user": "repl_user"
                      },
                      "resources": {
                        "requests": {
                          "cpu": "250m",
                          "memory": "256Mi"
                        }
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "port": 5432,
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "enabled": true,
                        "name": "annotationlab-airflow-postgresql"
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": true
                        },
                        "enabled": true
                      },
                      "slave": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "tls": {
                        "certFilename": "",
                        "certKeyFilename": "",
                        "certificatesSecret": "",
                        "enabled": false,
                        "preferServerCiphers": true
                      },
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "volumePermissions": {
                        "enabled": false,
                        "image": {
                          "pullPolicy": "Always",
                          "registry": "docker.io",
                          "repository": "bitnami/minideb",
                          "tag": "buster"
                        },
                        "securityContext": {}
                      }
                    },
                    "preannotator_queues": {
                      "count": 5
                    },
                    "quotas": {},
                    "rbac": {
                      "create": true,
                      "createSCCRoleBinding": false
                    },
                    "redis": {
                      "affinity": {},
                      "enabled": true,
                      "nodeSelector": {},
                      "password": "DefaulTRediS-P0ss",
                      "passwordSecretName": null,
                      "persistence": {
                        "annotations": {},
                        "enabled": true,
                        "size": "1Gi",
                        "storageClassName": null
                      },
                      "podAnnotations": {},
                      "resources": {},
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "annotationlab-airflow-redis"
                      },
                      "terminationGracePeriodSeconds": 600,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 0
                    },
                    "registry": {
                      "connection": {},
                      "secretName": "kubernetesdockerhub"
                    },
                    "registryCredentials": {
                      "pullPolicy": "IfNotPresent",
                      "registryUrl": "docker.io/annotationlab"
                    },
                    "revisionHistoryLimit": null,
                    "scheduler": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow scheduler"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "logGroomerSidecar": {
                        "args": [
                          "bash",
                          "/clean-logs"
                        ],
                        "command": null,
                        "enabled": true,
                        "resources": {},
                        "retentionDays": 15
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-scheduler"
                      },
                      "strategy": null,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "updateStrategy": null,
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "secret": [],
                    "securityContext": {},
                    "statsd": {
                      "affinity": {},
                      "args": [
                        "--statsd.mapping-config=/etc/statsd-exporter/mappings.yml"
                      ],
                      "enabled": false,
                      "extraMappings": [],
                      "extraNetworkPolicies": [],
                      "nodeSelector": {},
                      "overrideMappings": [],
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "securityContext": {},
                      "service": {
                        "extraAnnotations": {}
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-statsd"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 65534
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "triggerer": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow triggerer"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "annotationlab-airflow-triggerer"
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 60,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "uid": 50000,
                    "volumeMounts": [],
                    "volumes": [],
                    "webserver": {
                      "affinity": {},
                      "allowPodLogReading": true,
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow webserver"
                      ],
                      "command": null,
                      "defaultUser": {
                        "email": "admin@example.com",
                        "enabled": false,
                        "firstName": "admin",
                        "lastName": "user",
                        "password": "admin",
                        "role": "Admin",
                        "username": "admin"
                      },
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "failureThreshold": 20,
                        "initialDelaySeconds": 15,
                        "periodSeconds": 5,
                        "scheme": "HTTP",
                        "timeoutSeconds": 30
                      },
                      "networkPolicy": {
                        "ingress": {
                          "from": [],
                          "ports": [
                            {
                              "port": "{{ .Values.ports.airflowUI }}"
                            }
                          ]
                        }
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "readinessProbe": {
                        "failureThreshold": 20,
                        "initialDelaySeconds": 15,
                        "periodSeconds": 5,
                        "scheme": "HTTP",
                        "timeoutSeconds": 30
                      },
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "loadBalancerIP": null,
                        "loadBalancerSourceRanges": [],
                        "ports": [
                          {
                            "name": "airflow-ui",
                            "port": "{{ .Values.ports.airflowUI }}"
                          }
                        ],
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-webserver"
                      },
                      "strategy": null,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      },
                      "webserverConfig": null,
                      "webserverConfigConfigMapName": null
                    },
                    "webserverSecretKey": "f565cd64401775f93c820dc63fc3d4f6",
                    "webserverSecretKeySecretName": null,
                    "workers": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"celery worker\" \"worker\" }}"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "hostAliases": [],
                      "keda": {
                        "advanced": {},
                        "cooldownPeriod": 30,
                        "enabled": false,
                        "maxReplicaCount": 10,
                        "minReplicaCount": 0,
                        "namespaceLabels": {},
                        "pollingInterval": 5
                      },
                      "kerberosSidecar": {
                        "enabled": false,
                        "resources": {}
                      },
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "enabled": true,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "logGroomerSidecar": {
                        "args": [
                          "bash",
                          "/clean-logs"
                        ],
                        "command": null,
                        "enabled": true,
                        "resources": {},
                        "retentionDays": 15
                      },
                      "nodeSelector": {},
                      "persistence": {
                        "annotations": {},
                        "enabled": true,
                        "fixPermissions": false,
                        "size": "5Gi",
                        "storageClassName": null
                      },
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-worker"
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 600,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "updateStrategy": null,
                      "waitForMigrations": {
                        "env": []
                      }
                    }
                  },
                  "backup": {
                    "enable": false,
                    "files": true,
                    "image": {
                      "repository": null,
                      "tag": null
                    },
                    "s3_access_key": "",
                    "s3_bucket_fullpath": "",
                    "s3_secret_key": ""
                  },
                  "configuration": {
                    "FLASK_SECRET_KEY": "rpaQghA-9lmpLwB8qxW0FsW1TJyZZYuB60ppzswpTzM=",
                    "KEYCLOAK_CLIENT_ID": "annotationlab",
                    "KEYCLOAK_CLIENT_SECRET_KEY": "09a71c59-0351-4ce6-bc8f-8fd3feb9d2ff",
                    "KEYCLOAK_REALM_NAME": "master",
                    "PERSISTENT_STORAGE": "/projects"
                  },
                  "extraContainers": "",
                  "extraEnv": "",
                  "extraNetworkPolicies": "- namespaceSelector: {}\n",
                  "extraVolumeMounts": "",
                  "extraVolumes": "",
                  "github_oauth_origin": "https://github.oauth.johnsnowlabs.com",
                  "google_analytics": {
                    "api_secret": "WENfd00yeUxRRWl4N0Utbml5SWt3Zw==",
                    "measurement_id": "Ry1XODBIMDhTTFBZ"
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "docker.io/johnsnowlabs/annotationlab",
                    "tag": "5.7.0"
                  },
                  "ingress": {
                    "annotations": {},
                    "defaultBackend": false,
                    "enabled": true,
                    "hosts": [
                      {
                        "host": "annotationlab",
                        "path": "/"
                      }
                    ],
                    "ingressroute": {
                      "enabled": false
                    },
                    "servicePort": "http",
                    "tls": [],
                    "uploadLimitInMegabytes": 16
                  },
                  "installAirflow": true,
                  "installKeycloak": true,
                  "installPostgresql": true,
                  "keycloak": {
                    "affinity": "podAntiAffinity:\n  requiredDuringSchedulingIgnoredDuringExecution:\n    - labelSelector:\n        matchLabels:\n          {{- include \"keycloak.selectorLabels\" . | nindent 10 }}\n        matchExpressions:\n          - key: app.kubernetes.io/component\n            operator: NotIn\n            values:\n              - test\n      topologyKey: kubernetes.io/hostname\n  preferredDuringSchedulingIgnoredDuringExecution:\n    - weight: 100\n      podAffinityTerm:\n        labelSelector:\n          matchLabels:\n            {{- include \"keycloak.selectorLabels\" . | nindent 12 }}\n          matchExpressions:\n            - key: app.kubernetes.io/component\n              operator: NotIn\n              values:\n                - test\n        topologyKey: topology.kubernetes.io/zone\n",
                    "args": [],
                    "auththeme": {
                      "image": {
                        "repository": "docker.io/johnsnowlabs/annotationlab",
                        "tag": "auth-theme-5.7.0"
                      }
                    },
                    "autoscaling": {
                      "behavior": {
                        "scaleDown": {
                          "policies": [
                            {
                              "periodSeconds": 300,
                              "type": "Pods",
                              "value": 1
                            }
                          ],
                          "stabilizationWindowSeconds": 300
                        }
                      },
                      "enabled": false,
                      "labels": {},
                      "maxReplicas": 10,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 80,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 3
                    },
                    "cache": {
                      "stack": "default"
                    },
                    "clusterDomain": "cluster.local",
                    "command": [
                      "/opt/keycloak/bin/kc.sh",
                      "--verbose",
                      "start",
                      "--http-enabled=true",
                      "--http-port=8080",
                      "--hostname-strict=false",
                      "--hostname-strict-https=false",
                      "--spi-events-listener-jboss-logging-success-level=info",
                      "--spi-events-listener-jboss-logging-error-level=warn"
                    ],
                    "database": {
                      "enabled": true,
                      "postgresqlDatabase": "keycloak",
                      "postgresqlPassword": "keycloak",
                      "postgresqlUsername": "keycloak"
                    },
                    "dbchecker": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/busybox",
                        "tag": 1.32
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "32Mi"
                        },
                        "requests": {
                          "cpu": "20m",
                          "memory": "32Mi"
                        }
                      },
                      "securityContext": {}
                    },
                    "enableServiceLinks": true,
                    "extraContainers": "",
                    "extraEnv": "- name: KC_FEATURES\n  value: \"authorization,account-api,admin-api,admin-fine-grained-authz,ciba,impersonation,kerberos,preview,step-up-authentication,token-exchange,web-authn\"\n- name: KEYCLOAK_ADMIN\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"keycloak.fullname\" . }}-admincreds\n      key: user\n- name: KEYCLOAK_ADMIN_PASSWORD\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"keycloak.fullname\" . }}-admincreds\n      key: password\n- name: JAVA_OPTS_APPEND\n  value: \u003e-\n    -XX:+UseContainerSupport\n    -XX:MaxRAMPercentage=50.0\n    -Djava.awt.headless=true\n    -Djgroups.dns.query={{ include \"keycloak.fullname\" . }}-headless\n",
                    "extraEnvFrom": "",
                    "extraInitContainers": "- name: theme-provider\n  image: {{ .Values.auththeme.image.repository }}:{{ .Values.auththeme.image.tag }}\n  imagePullPolicy: IfNotPresent\n  command:\n    - sh\n  args:\n    - -c\n    - |\n      echo \"Copying theme...\"\n      cp -R /johnsnowlabs/* /theme\n  volumeMounts:\n    - name: theme\n      mountPath: /theme\n",
                    "extraPorts": [],
                    "extraServiceMonitor": {
                      "annotations": {},
                      "enabled": false,
                      "interval": "10s",
                      "labels": {},
                      "namespace": "",
                      "namespaceSelector": {},
                      "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/realms/master/metrics",
                      "port": "http",
                      "scrapeTimeout": "10s"
                    },
                    "extraVolumeMounts": "- name: admincreds\n  mountPath: /secrets/admincreds\n  readOnly: true\n- name: theme\n  mountPath: /opt/keycloak/themes/johnsnowlabs\n",
                    "extraVolumes": "- name: admincreds\n  secret:\n    secretName: '{{ include \"annotationlab.keycloak.fullname\" . }}-admincreds'\n- name: theme\n  emptyDir: {}\n",
                    "fullnameOverride": "",
                    "global": {},
                    "health": {
                      "enabled": true
                    },
                    "hostAliases": [],
                    "http": {
                      "relativePath": "/auth"
                    },
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/keycloak/keycloak",
                      "tag": "20.0.3"
                    },
                    "imagePullSecrets": [
                      {
                        "name": "kubernetesdockerhub"
                      }
                    ],
                    "ingress": {
                      "annotations": {
                        "nginx.ingress.kubernetes.io/proxy-buffer-size": "16k"
                      },
                      "console": {
                        "annotations": {},
                        "enabled": false,
                        "ingressClassName": "",
                        "rules": [
                          {
                            "host": "{{ .Release.Name }}.keycloak.example.com",
                            "paths": [
                              {
                                "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/admin",
                                "pathType": "Prefix"
                              }
                            ]
                          }
                        ],
                        "tls": []
                      },
                      "enabled": false,
                      "ingressClassName": "",
                      "labels": {},
                      "rules": [
                        {
                          "host": "keycloak.example.com",
                          "paths": [
                            {
                              "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/",
                              "pathType": "Prefix"
                            }
                          ]
                        }
                      ],
                      "servicePort": "http",
                      "tls": []
                    },
                    "lifecycleHooks": "",
                    "livenessProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health/live'\n  port: http\ninitialDelaySeconds: 0\ntimeoutSeconds: 5\n",
                    "metrics": {
                      "enabled": true
                    },
                    "nameOverride": "",
                    "networkPolicy": {
                      "egress": [],
                      "enabled": false,
                      "extraFrom": [],
                      "labels": {}
                    },
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podDisruptionBudget": {},
                    "podLabels": {},
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {},
                    "postgresql": {
                      "common": {
                        "exampleValue": "common-chart",
                        "global": {
                          "postgresql": {}
                        }
                      },
                      "commonAnnotations": {},
                      "enabled": true,
                      "extraEnv": [],
                      "global": {
                        "postgresql": {}
                      },
                      "image": {
                        "debug": false,
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgresql",
                        "tag": "11.20.0-debian-11-r12"
                      },
                      "ldap": {
                        "baseDN": "",
                        "bindDN": "",
                        "bind_password": null,
                        "enabled": false,
                        "port": "",
                        "prefix": "",
                        "scheme": "",
                        "search_attr": "",
                        "search_filter": "",
                        "server": "",
                        "suffix": "",
                        "tls": false,
                        "url": ""
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 30,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "master": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": [],
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "metrics": {
                        "enabled": false,
                        "extraEnvVars": {},
                        "image": {
                          "pullPolicy": "IfNotPresent",
                          "registry": "docker.io",
                          "repository": "bitnami/postgres-exporter",
                          "tag": "0.8.0-debian-10-r166"
                        },
                        "livenessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "prometheusRule": {
                          "additionalLabels": {},
                          "enabled": false,
                          "namespace": "",
                          "rules": []
                        },
                        "readinessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "securityContext": {},
                        "service": {
                          "annotations": {
                            "prometheus.io/port": "9187",
                            "prometheus.io/scrape": "true"
                          },
                          "loadBalancerIP": null,
                          "type": "ClusterIP"
                        },
                        "serviceMonitor": {
                          "additionalLabels": {},
                          "enabled": false
                        }
                      },
                      "nameOverride": "keycloak-postgres",
                      "networkPolicy": {
                        "allowExternal": true,
                        "enabled": false,
                        "explicitNamespacesSelector": {}
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "enabled": true,
                        "mountPath": "/bitnami/postgresql",
                        "size": "8Gi",
                        "subPath": ""
                      },
                      "postgresqlDataDir": "/bitnami/postgresql/data",
                      "postgresqlDatabase": "keycloak",
                      "postgresqlPassword": "keycloak",
                      "postgresqlUsername": "keycloak",
                      "psp": {
                        "create": false
                      },
                      "rbac": {
                        "create": false
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "replication": {
                        "applicationName": "my_application",
                        "enabled": false,
                        "numSynchronousReplicas": 0,
                        "password": "repl_password",
                        "slaveReplicas": 1,
                        "synchronousCommit": "off",
                        "user": "repl_user"
                      },
                      "resources": {
                        "requests": {
                          "cpu": "250m",
                          "memory": "256Mi"
                        }
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "port": 5432,
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "enabled": true,
                        "name": "annotationlab-keycloak-postgresql"
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": true
                        },
                        "enabled": true
                      },
                      "slave": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "tls": {
                        "certCAFilename": null,
                        "certFilename": "",
                        "certKeyFilename": "",
                        "certificatesSecret": "",
                        "crlFilename": null,
                        "enabled": false,
                        "preferServerCiphers": true
                      },
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "volumePermissions": {
                        "enabled": false,
                        "image": {
                          "pullPolicy": "Always",
                          "registry": "docker.io",
                          "repository": "bitnami/minideb",
                          "tag": "buster"
                        },
                        "securityContext": {}
                      }
                    },
                    "priorityClassName": "",
                    "prometheusRule": {
                      "annotations": {},
                      "enabled": false,
                      "labels": {},
                      "namespace": "",
                      "rules": []
                    },
                    "proxy": {
                      "enabled": true,
                      "mode": "edge"
                    },
                    "rbac": {
                      "create": false,
                      "rules": []
                    },
                    "readinessProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health/ready'\n  port: http\ninitialDelaySeconds: 10\ntimeoutSeconds: 1\n",
                    "replicas": 1,
                    "resources": {},
                    "restartPolicy": "Always",
                    "route": {
                      "annotations": {},
                      "enabled": false,
                      "host": "",
                      "labels": {},
                      "path": "/",
                      "tls": {
                        "enabled": true,
                        "insecureEdgeTerminationPolicy": "Redirect",
                        "termination": "edge"
                      }
                    },
                    "secrets": {
                      "admincreds": {
                        "stringData": {
                          "password": "secret",
                          "user": "admin"
                        }
                      }
                    },
                    "securityContext": {},
                    "service": {
                      "annotations": {},
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "httpNodePort": null,
                      "httpPort": 80,
                      "httpsNodePort": null,
                      "httpsPort": 8443,
                      "labels": {},
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "sessionAffinity": "",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "serviceAccount": {
                      "allowReadPods": false,
                      "annotations": {},
                      "automountServiceAccountToken": true,
                      "create": false,
                      "imagePullSecrets": [],
                      "labels": {},
                      "name": "annotationlab-keyclo"
                    },
                    "serviceHeadless": {
                      "annotations": {}
                    },
                    "serviceMonitor": {
                      "annotations": {},
                      "enabled": false,
                      "interval": "10s",
                      "labels": {},
                      "namespace": "",
                      "namespaceSelector": {},
                      "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/metrics",
                      "port": "http",
                      "scrapeTimeout": "10s"
                    },
                    "skipInitContainers": false,
                    "startupProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health'\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 1\nfailureThreshold: 60\nperiodSeconds: 5\n",
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {},
                    "terminationGracePeriodSeconds": 60,
                    "test": {
                      "deletionPolicy": "before-hook-creation",
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/seleniarm/standalone-chromium",
                        "tag": "111.0"
                      },
                      "podSecurityContext": {},
                      "securityContext": {}
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": null,
                    "updateStrategy": "RollingUpdate"
                  },
                  "licenseserver": {
                    "origin": "https://license.johnsnowlabs.com"
                  },
                  "lifecycleHooks": "# postStart:\n#   exec:\n#     command: [\"/bin/sh\", \"-c\", \"ls\"]\n",
                  "modelshub_origin": "https://modelshub.johnsnowlabs.com",
                  "myjohnsnowlabs": {
                    "client_id": "UIGgoiIL9szXW8Nnik0O7tPT4to87oVtWvYWfc6r",
                    "origin": "https://my.johnsnowlabs.com"
                  },
                  "networkPolicy": {
                    "enabled": false
                  },
                  "nodeSelector": {},
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "annotations": {},
                    "enabled": true,
                    "size": "8Gi"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "postgresql": {
                    "common": {
                      "exampleValue": "common-chart",
                      "global": {
                        "postgresql": {}
                      }
                    },
                    "commonAnnotations": {},
                    "extraEnv": [],
                    "global": {
                      "postgresql": {}
                    },
                    "image": {
                      "debug": false,
                      "pullPolicy": "IfNotPresent",
                      "registry": "docker.io",
                      "repository": "bitnami/postgresql",
                      "tag": "11.20.0-debian-11-r12"
                    },
                    "ldap": {
                      "baseDN": "",
                      "bindDN": "",
                      "bind_password": null,
                      "enabled": false,
                      "port": "",
                      "prefix": "",
                      "scheme": "",
                      "search_attr": "",
                      "search_filter": "",
                      "server": "",
                      "suffix": "",
                      "tls": false,
                      "url": ""
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "master": {
                      "affinity": {},
                      "annotations": {},
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podLabels": {},
                      "priorityClassName": "",
                      "service": {},
                      "sidecars": [],
                      "tolerations": []
                    },
                    "metrics": {
                      "enabled": false,
                      "extraEnvVars": {},
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgres-exporter",
                        "tag": "0.8.0-debian-10-r166"
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "prometheusRule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {
                          "prometheus.io/port": "9187",
                          "prometheus.io/scrape": "true"
                        },
                        "loadBalancerIP": null,
                        "type": "ClusterIP"
                      },
                      "serviceMonitor": {
                        "additionalLabels": {},
                        "enabled": false
                      }
                    },
                    "networkPolicy": {
                      "allowExternal": true,
                      "enabled": false,
                      "explicitNamespacesSelector": {}
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {},
                      "enabled": true,
                      "mountPath": "/bitnami/postgresql",
                      "size": "8Gi",
                      "subPath": ""
                    },
                    "postgresqlDataDir": "/bitnami/postgresql/data",
                    "postgresqlDatabase": "annotationlab",
                    "postgresqlPassword": "annotationlab",
                    "postgresqlUsername": "annotationlab",
                    "psp": {
                      "create": false
                    },
                    "rbac": {
                      "create": false
                    },
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 6,
                      "initialDelaySeconds": 5,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "replication": {
                      "applicationName": "my_application",
                      "enabled": false,
                      "numSynchronousReplicas": 0,
                      "password": "repl_password",
                      "slaveReplicas": 1,
                      "synchronousCommit": "off",
                      "user": "repl_user"
                    },
                    "resources": {
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {},
                    "service": {
                      "annotations": {},
                      "port": 5432,
                      "type": "ClusterIP"
                    },
                    "serviceAccount": {
                      "enabled": false
                    },
                    "shmVolume": {
                      "chmod": {
                        "enabled": true
                      },
                      "enabled": true
                    },
                    "slave": {
                      "affinity": {},
                      "annotations": {},
                      "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podLabels": {},
                      "priorityClassName": "",
                      "service": {},
                      "sidecars": [],
                      "tolerations": []
                    },
                    "tls": {
                      "certCAFilename": null,
                      "certFilename": "",
                      "certKeyFilename": "",
                      "certificatesSecret": "",
                      "crlFilename": null,
                      "enabled": false,
                      "preferServerCiphers": true
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "volumePermissions": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "Always",
                        "registry": "docker.io",
                        "repository": "bitnami/minideb",
                        "tag": "buster"
                      },
                      "securityContext": {}
                    }
                  },
                  "preannotator_queues": {
                    "count": 5,
                    "service": {
                      "port": 5000,
                      "portName": "http",
                      "type": "ClusterIP"
                    }
                  },
                  "priorityClassName": "",
                  "readinessProbe": "httpGet:\n  path: /healthz\n  port: http\ninitialDelaySeconds: 10\ntimeoutSeconds: 1\n",
                  "registryCredentials": {
                    "create": false,
                    "password": "XXXXX",
                    "secretName": "kubernetesdockerhub",
                    "username": "XXXXX"
                  },
                  "replicas": 1,
                  "resources": {},
                  "restartPolicy": "Always",
                  "securityContext": {
                    "enabled": false,
                    "fsGroup": 65534,
                    "fsGroupChangePolicy": "OnRootMismatch",
                    "runAsUser": 65534
                  },
                  "service": {
                    "port": 8200,
                    "portName": "http",
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": false,
                    "name": "annotationlab"
                  },
                  "sharedData": {
                    "enabled": true,
                    "images": {
                      "storageSize": "5Gi"
                    },
                    "storageSize": "100Gi"
                  },
                  "sqlalchemy_pool_size": 10,
                  "telemetry": {
                    "enabled": true
                  },
                  "terminationGracePeriodSeconds": 60,
                  "tolerations": [],
                  "updateStrategy": {
                    "type": "Recreate"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.johnsnowlabs.com/annotationlab/annotationlab-operator:v5.7.0
          createdAt: "2023-12-19T10:26:54Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/annotationlab/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/annotationlab/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: johnsnowlabs-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: NLP Lab
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Annotationlab
            name: annotationlabs.apps.johnsnowlabs.com
            version: v1
        description: "NLP Lab is a data labeling and annotation tool that helps you\nlabel
          different types of data using a simple interface with a standardized output\nformat.\n\nIt
          contains enterprise features like Project Management and sharing, SSO\nintegration,
          OpenAPI Specification 3.0 compliant API, task management, and active\nlearning/preannotation
          using Spark NLP.\n\nFurthermore, the production deployment of the application
          is done using a helm chart.\nThe same helm chart is also used to create
          an AWS AMI for an EC2 based single server \ndeployment using packer.\n\nMake
          sure you have defined shared storage such as efs/nfs/cephfs prior to installing
          the NLP Lab Operator.\nUpdate the host property in the YAML configuration
          to define the required domain name to use instead of the \ndefault hostname
          annotationlab.\n\nNLP Lab Operator Installtion guide [NLP Lab Operator][]:\n[NLP
          Lab Operator]: https://nlp.johnsnowlabs.com/docs/en/alab/install    \n"
        displayName: NLPLab
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nlplab
        links:
        - name: Annotationlab
          url: https://annotationlab.domain
        maintainers:
        - email: bhandari@johnsnowlabs.com
          name: bikash
        maturity: alpha
        provider:
          name: JohnSnowLABS
          url: nlp.johnsnowlabs.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - registry.johnsnowlabs.com/annotationlab/annotationlab-operator:v5.7.0
        version: 5.7.0
      entries:
      - name: annotationlab.v5.7.0
        version: 5.7.0
      - name: annotationlab.v5.4.1
        version: 5.4.1
      - name: annotationlab.v4.9.2
        version: 4.9.2
      name: alpha
    defaultChannel: alpha
    packageName: annotationlab
    provider:
      name: JohnSnowLABS
      url: nlp.johnsnowlabs.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NGINX Inc
      provider-url: ""
    name: nginx-ingress-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nginx-ingress-operator.v3.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.nginx.org/v1alpha1",
                "kind": "NginxIngress",
                "metadata": {
                  "name": "nginxingress-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "annotations": {},
                    "appprotect": {
                      "enable": false
                    },
                    "appprotectdos": {
                      "debug": false,
                      "enable": false,
                      "maxDaemons": 0,
                      "maxWorkers": 0,
                      "memory": 0
                    },
                    "autoscaling": {
                      "annotations": {},
                      "behavior": {},
                      "enabled": false,
                      "maxReplicas": 3,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 50,
                      "targetMemoryUtilizationPercentage": 50
                    },
                    "config": {
                      "annotations": {},
                      "entries": {}
                    },
                    "containerPort": {
                      "http": 80,
                      "https": 443
                    },
                    "customConfigMap": "",
                    "customPorts": [],
                    "defaultHTTPListenerPort": 80,
                    "defaultHTTPSListenerPort": 443,
                    "defaultTLS": {
                      "cert": "",
                      "key": "",
                      "secret": ""
                    },
                    "disableIPV6": false,
                    "dnsPolicy": "ClusterFirst",
                    "enableCertManager": false,
                    "enableCustomResources": true,
                    "enableExternalDNS": false,
                    "enableLatencyMetrics": false,
                    "enableOIDC": false,
                    "enableSSLDynamicReload": true,
                    "enableSnippets": false,
                    "enableTLSPassthrough": false,
                    "env": [],
                    "extraContainers": [],
                    "globalConfiguration": {
                      "create": false,
                      "spec": {}
                    },
                    "healthStatus": false,
                    "healthStatusURI": "/nginx-health",
                    "hostNetwork": false,
                    "hostPort": {
                      "enable": false,
                      "http": 80,
                      "https": 443
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "nginx/nginx-ingress",
                      "tag": "5.2.0-ubi"
                    },
                    "ingressClass": {
                      "create": true,
                      "name": "nginx",
                      "setAsDefaultIngress": false
                    },
                    "initContainerResources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "initContainers": [],
                    "kind": "deployment",
                    "lifecycle": {},
                    "logFormat": "glog",
                    "logLevel": "info",
                    "mgmt": {
                      "licenseTokenSecretName": "license-token"
                    },
                    "minReadySeconds": 0,
                    "name": "controller",
                    "nginxDebug": false,
                    "nginxReloadTimeout": 60000,
                    "nginxStatus": {
                      "allowCidrs": "127.0.0.1",
                      "enable": true,
                      "port": 8080
                    },
                    "nginxplus": false,
                    "pod": {
                      "annotations": {},
                      "extraLabels": {}
                    },
                    "podDisruptionBudget": {
                      "annotations": {},
                      "enabled": false
                    },
                    "readOnlyRootFilesystem": false,
                    "readyStatus": {
                      "enable": true,
                      "initialDelaySeconds": 0,
                      "port": 8081
                    },
                    "replicaCount": 1,
                    "reportIngressStatus": {
                      "annotations": {},
                      "enable": true,
                      "enableLeaderElection": true,
                      "ingressLink": "",
                      "leaderElectionLockName": "nginx-ingress-leader"
                    },
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "selectorLabels": {},
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "create": true,
                      "customPorts": [],
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraLabels": {},
                      "httpPort": {
                        "enable": true,
                        "port": 80,
                        "targetPort": 80
                      },
                      "httpsPort": {
                        "enable": true,
                        "port": 443,
                        "targetPort": 443
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "type": "LoadBalancer"
                    },
                    "serviceAccount": {
                      "annotations": {},
                      "imagePullSecretName": "",
                      "imagePullSecretsNames": []
                    },
                    "shareProcessNamespace": false,
                    "strategy": {},
                    "terminationGracePeriodSeconds": 30,
                    "tlsPassthroughPort": 443,
                    "tolerations": [],
                    "volumeMounts": [],
                    "volumes": [],
                    "watchNamespace": "",
                    "watchNamespaceLabel": "",
                    "watchSecretNamespace": "",
                    "wildcardTLS": {
                      "cert": "",
                      "key": "",
                      "secret": ""
                    }
                  },
                  "nginxServiceMesh": {
                    "enable": false,
                    "enableEgress": false
                  },
                  "prometheus": {
                    "create": true,
                    "port": 9113,
                    "scheme": "http",
                    "secret": "",
                    "service": {
                      "create": false,
                      "labels": {
                        "service": "nginx-ingress-prometheus-service"
                      }
                    },
                    "serviceMonitor": {
                      "create": false,
                      "endpoints": [
                        {
                          "port": "prometheus"
                        }
                      ],
                      "labels": {},
                      "selectorMatchLabels": {
                        "service": "nginx-ingress-prometheus-service"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "serviceInsight": {
                    "create": false,
                    "port": 9114,
                    "scheme": "http",
                    "secret": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking
          certified: "true"
          containerImage: quay.io/nginx/nginx-ingress-operator@sha256:fb372438634b4a667218e725091bd448a253e7a9c40e7dda1044a318780d995f
          createdAt: "2025-09-15T16:08:18Z"
          description: The NGINX Ingress Operator is a Kubernetes/OpenShift component
            which deploys and manages one or more NGINX/NGINX Plus Ingress Controllers
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nginx-ingress
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/nginx/nginx-ingress-helm-operator
          support: NGINX Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The `NginxIngress` Custom Resource is the definition of a
              deployment of the Ingress Controller.
            displayName: Nginx Ingress Controller
            kind: NginxIngress
            name: nginxingresses.charts.nginx.org
            version: v1alpha1
        description: The NGINX Ingress Operator is a Kubernetes/OpenShift component
          which deploys and manages one or more NGINX/NGINX Plus Ingress Controllers
        displayName: Nginx Ingress Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nginx
        - ingress-controller
        - ingress
        - controller
        - kubernetes
        - openshift
        links:
        - name: Nginx Ingress Operator
          url: https://github.com/nginx/nginx-ingress-helm-operator
        maintainers:
        - email: kubernetes@nginx.com
          name: NGINX Inc
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: NGINX Inc
        relatedImages:
        - quay.io/nginx/nginx-ingress-operator@sha256:fb372438634b4a667218e725091bd448a253e7a9c40e7dda1044a318780d995f
        - quay.io/brancz/kube-rbac-proxy@sha256:754ab2a723c8fc7ea8e09ae1224f8af5be91849b8b2f2a3735c6237f1c534187
        version: 3.3.0
      entries:
      - name: nginx-ingress-operator.v3.3.0
        version: 3.3.0
      name: alpha
    defaultChannel: alpha
    packageName: nginx-ingress-operator
    provider:
      name: NGINX Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: QuestLabs
      provider-url: https://www.questlabs.in
    name: alloy
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: alloy.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.alloys.com/v1alpha1",
                "kind": "Alloy",
                "metadata": {
                  "name": "alloy"
                },
                "spec": {
                  "acmeEmail": "",
                  "dnsFQDN": "",
                  "sslTermination": "true"
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Alloy
            name: alloys.charts.alloys.com
            version: v1alpha1
        description: |-
          Alloy is platform to build trusted data in OSDU, available on-premise or in the cloud. It is equipped with built-in data quality rules derived from PPDM and OSDU standards that control the ingestion of data.

              • Applicable for conventional and new energy streams
              • Highly Scalable

              Benefits
              • Helps to build robust applications/solutions on stable data foundation using full data lifecycle APIs & built-in data/business rules
              • Help data governance team to monitor/oversea data across organisation and intervene when required
              • Substantially low implementation time for any new data entity or new domain (solar, wind etc.) because of modular rules packages
              • Capability to deploy workflows for complex corporate specific data quality and enrichment requirements
              • Spend less time & resources to deploy fit-for-purpose data
              • Schema based checks out of the box for conventional and new energy streams.
              • Sustainable and future proof


              Prerequisites
              1. Please create alloy project before installing alloy operator

          Getting Help

          Once Alloy is successfully installed you can also browse documentation at https://your.domain.com/alloy/docs/introduction
          Please visit https://www.questlabs.in/alloy for more details and if you have any queries or suggestions, mail us at info@questlabs.in
        displayName: Alloy
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - alloy
        links:
        - name: alloy
          url: https://www.questlabs.in
        maintainers:
        - email: info@questlabs.in
          name: QuestLabs
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: QuestLabs
          url: https://www.questlabs.in
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d291e0d38bd95eb2731a10bd5ff06bfa07c6621528d3c2bbe888397d87b3feaa
        - alloy.azurecr.io/alloy@sha256:bf29f69889d408507cf1a9a940beb80d95ba9faff1da2d70a4aea82cc235d96d
        version: 1.0.4
      entries:
      - name: alloy.v1.0.4
        version: 1.0.4
      - name: alloy.v1.0.3
        version: 1.0.3
      - name: alloy.v1.0.2
        version: 1.0.2
      - name: alloy.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: alloy
    provider:
      name: QuestLabs
      url: https://www.questlabs.in
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: SynaXG Engineering Team
      provider-url: ""
    name: synaxg-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: synaxg-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "card.synaxg.app.com/v1",
                "kind": "SG3Reset",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "synaxg-operator",
                    "app.kubernetes.io/instance": "sg3reset-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sg3reset",
                    "app.kubernetes.io/part-of": "synaxg-operator"
                  },
                  "name": "sg3reset-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/synaxg/synaxg-reset-utility@sha256:4050d5731a21bbec6e26165b8dc2b7632577ffa376e7eb056aa3245cf9574753",
                  "sg3ResetList": [
                    {
                      "nodeName": "test",
                      "pciList": [
                        {
                          "enable": "false",
                          "pciAddr": "000d:01:00.0"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "card.synaxg.app.com/v1",
                "kind": "SG3Update",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "synaxg-operator",
                    "app.kubernetes.io/instance": "sg3update-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sg3update",
                    "app.kubernetes.io/part-of": "synaxg-operator"
                  },
                  "name": "sg3update-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/synaxg/synaxg-upgrade-utility@sha256:7ffec3249edd2467f2414c2ce4232e8e436278fd1b43a51b66f54b6ea10a60d1",
                  "sg3UpdateList": [
                    {
                      "nodeName": "test",
                      "pciList": [
                        {
                          "enable": "false",
                          "pciAddr": "000d:01:00.0",
                          "sdkImageAddr": "quay.io/synaxgcom/sdk-img@sha256:1ed57013f278e5a858f71f343649652dbd52b6f3cf451374e45b6ac427fa76f9"
                        }
                      ]
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/synaxg/synaxg-operator-utility@sha256:3b22a8022a97cb1ffdb47ffcc9d9d562601f48669aa14fd58cd8ba6128b39bd7
          createdAt: "2025-07-01T03:42:07Z"
          description: SynaXG Operator will work with the SuperMicro Server + SG3
            card  and Openshift Container Platform.The Marvell kernel module “octeon_ep”
            and “octeon_ep_vf” are necessary to the Operator. The Operator can send
            command of “reset” and “update” to SG3 board via the VF channel.As a result,
            users can realize the process of “reset card” and “update firmware” on
            the web GUI, and monitor the executing result.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: free text
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: SynaXG
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SG3Reset is the Schema for the sg3resets API
            displayName: SG3 Reset
            kind: SG3Reset
            name: sg3resets.card.synaxg.app.com
            version: v1
          - description: SG3Update is the Schema for the sg3updates API
            displayName: SG3 Update
            kind: SG3Update
            name: sg3updates.card.synaxg.app.com
            version: v1
        description: SynaXG Operator will work with the SuperMicro Server + SG3 card  and
          Openshift Container Platform.The Marvell kernel module “octeon_ep” and “octeon_ep_vf”
          are necessary to the Operator. The Operator can send command of “reset”
          and “update” to SG3 board via the VF channel.As a result, users can realize
          the process of “reset card” and “update firmware” on the web GUI, and monitor
          the executing result.
        displayName: SynaXG vDU card 102X Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vDU
        - 102X
        - SynaXG
        links:
        - name: Synaxg Operator
          url: https://synaxg-operator.domain
        maintainers:
        - email: notice@synaxg.com
          name: Mantosh Malhotra
        maturity: alpha
        provider:
          name: SynaXG Engineering Team
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:38ee3484ba60bb474e7231e9d0718157386716cea7ffd6da6f32ddd551a5243d
        - registry.connect.redhat.com/synaxg/synaxg-operator-utility@sha256:3b22a8022a97cb1ffdb47ffcc9d9d562601f48669aa14fd58cd8ba6128b39bd7
        - registry.connect.redhat.com/synaxg/synaxg-reset-utility@sha256:4050d5731a21bbec6e26165b8dc2b7632577ffa376e7eb056aa3245cf9574753
        - registry.connect.redhat.com/synaxg/synaxg-upgrade-utility@sha256:7ffec3249edd2467f2414c2ce4232e8e436278fd1b43a51b66f54b6ea10a60d1
        - quay.io/synaxgcom/sdk-img@sha256:1ed57013f278e5a858f71f343649652dbd52b6f3cf451374e45b6ac427fa76f9
        version: 1.2.0
      entries:
      - name: synaxg-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: synaxg-operator
    provider:
      name: SynaXG Engineering Team
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-opa-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opa-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:08f995a39db5c2c6a17008cffa5216bb9bebd784d13d2b94154474d727fc796f
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/opa-operator@sha256:08f995a39db5c2c6a17008cffa5216bb9bebd784d13d2b94154474d727fc796f
        version: 24.11.1
      entries:
      - name: opa-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: opa-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:826aa12df4531d93a337ba70bd0c0d1e2e8062efe6954fc22797f6474c7732ca
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/opa-operator@sha256:826aa12df4531d93a337ba70bd0c0d1e2e8062efe6954fc22797f6474c7732ca
        version: 25.3.0
      entries:
      - name: opa-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: opa-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaClusterSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
        version: 25.7.0
      entries:
      - name: opa-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: opa-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaClusterSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
        version: 25.7.0
      entries:
      - name: opa-operator.v25.7.0
        version: 25.7.0
      - name: opa-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-opa-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: datagrid
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datagrid-operator.v8.2.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "adminAuth": {
                    "secretName": "basic-auth"
                  },
                  "clusterName": "example-infinispan",
                  "name": "mycache"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC batch-cache"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:1435aa60e8ab4d25c86daae68dcb93d14999f50d7c1e11a3c5726a86d3d30453
          createdAt: "2021-08-24T17:39:44Z"
          description: Create and manage Red Hat Data Grid clusters.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Restore API schema.
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Interoperability:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Data Grid Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
          Red Hat Data Grid is a distributed, in-memory data store that increases application performance and delivers open-source capabilities to handle demanding use cases.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.2/html/data_grid_operator_8.2_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.2
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:1435aa60e8ab4d25c86daae68dcb93d14999f50d7c1e11a3c5726a86d3d30453
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4d8057709958ef60bcabdcec52f05eae8df121fd3f4762c3218abca38130f638
        - registry.redhat.io/datagrid/datagrid-8-openj9-11-rhel8@sha256:c8607d6538a7528e27e6ecb138e0a93a03b3c9fa1f8a9a5aad3fd9f4a6814448
        version: 8.2.8
      entries:
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.2.x
    - currentCSV: datagrid-operator.v8.3.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:4d9d7221b029866837add5f3cb52dd0b7956f57996ce2d8b76344ce2de678716
          createdAt: "2022-11-03T08:40:57Z"
          description: Create and manage Red Hat Data Grid clusters.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.3/html/data_grid_operator_8.3_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.3
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:4d9d7221b029866837add5f3cb52dd0b7956f57996ce2d8b76344ce2de678716
        version: 8.3.9
      entries:
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.3.x
    - currentCSV: datagrid-operator.v8.4.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:e5315682869de72514e3e027bfb02bf0bb64919453ca738c7ecfa1e34634ac2b
          createdAt: 2024-05-30T08:05:19 EDT
          description: Create and manage Red Hat Data Grid clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.4/html/data_grid_operator_8.4_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.4
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:985193caf9bc3698c4d85aeedd1cfea09bccda881636d67885387b2a83c65ebd
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9d9e466b7c367d0406406550c74bc790aa68f3f875ea4f652eb28f2e25a05458
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:b56269cb849afca34e7f82cc8e01a44bb32af6126f24db28019a58e556d4cb3c
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:d522a20ca6909bd464f85967d8b86fe42b83f773478a80785f9a85a0cb87dfed
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bf7760520f5bea151a531d538d525c4591c47166995aefb4d77fef138de11544
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bcb040bf640691568a93a4c71c8072c37ffb76b4296ade83186f279862b7f197
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:24cf79efa89bbfa495665b1711a27531b7ab0813695a89b2c38c96c2e792eb24
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:532dca6facb584c1261c0480c499af9f857c70dc390e0d46e46afec2561468af
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:78a6bed6fcb6962e5689b92806789b72a66db21b7c607b392a573c98593f2c09
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:ae852e4679978725bb7d24c1bdfe25b40bd883130ed04a36d4fca49dbea024ef
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:424b30dccd773f69c079b35a42e8c3e339246c5d490e14e3fcfb8bd8144a9160
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9223b4b5c1c866a91033dac40f30a4caae0f6946cdb64935a3fdfffdb9ede660
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:c15d4fd827f554291c14851f187947f4109c0e20257fa2f1e141b32669569358
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:54bc26a0fddc160a4e0d1a813aec78abb3c9cb77db5039b74c23de759faa67f5
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:554b831a0803d55a4e3b7022011b5a5f6d2b4b65009065a9b28bc8e55fc3ad46
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:e5315682869de72514e3e027bfb02bf0bb64919453ca738c7ecfa1e34634ac2b
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4dbe1b7a1a36990c0a31d5a66f3cc4ca2954c9e1ec3b87431113a4111530277d
        version: 8.4.16
      entries:
      - name: datagrid-operator.v8.4.16
        version: 8.4.16
      - name: datagrid-operator.v8.4.15
        version: 8.4.15
      - name: datagrid-operator.v8.4.14
        version: 8.4.14
      - name: datagrid-operator.v8.4.13
        version: 8.4.13
      - name: datagrid-operator.v8.4.12
        version: 8.4.12
      - name: datagrid-operator.v8.4.11
        version: 8.4.11
      - name: datagrid-operator.v8.4.10
        version: 8.4.9
      - name: datagrid-operator.v8.4.8
        version: 8.4.8
      - name: datagrid-operator.v8.4.7
        version: 8.4.7
      - name: datagrid-operator.v8.4.6
        version: 8.4.6
      - name: datagrid-operator.v8.4.5
        version: 8.4.5
      - name: datagrid-operator.v8.4.4
        version: 8.4.4
      - name: datagrid-operator.v8.4.3
        version: 8.4.3
      - name: datagrid-operator.v8.4.2
        version: 8.4.2
      - name: datagrid-operator.v8.4.1
        version: 8.4.1
      - name: datagrid-operator.v8.4.0
        version: 8.4.0
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.4.x
    - currentCSV: datagrid-operator.v8.5.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel9-operator@sha256:d719d778b0848588a7e61d3c3fa689763f51a6533af4cf9e1c738c4ed8fac760
          createdAt: 2025-08-29T05:44:37 UTC
          description: Create and manage Red Hat Data Grid clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.5/html/data_grid_operator_8.5_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.5
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:a711c50e489995fb60e5dec23ce286360743feda5066e4a0861d448aa4f2e903
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bf7760520f5bea151a531d538d525c4591c47166995aefb4d77fef138de11544
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bcb040bf640691568a93a4c71c8072c37ffb76b4296ade83186f279862b7f197
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:4b9bf39d2ff14bcfbd4628a399093703d87a2b2b9c8e74c3d7cfdff4b5167fa4
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:ae852e4679978725bb7d24c1bdfe25b40bd883130ed04a36d4fca49dbea024ef
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:d522a20ca6909bd464f85967d8b86fe42b83f773478a80785f9a85a0cb87dfed
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9223b4b5c1c866a91033dac40f30a4caae0f6946cdb64935a3fdfffdb9ede660
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:11bfe5c8109d26164cff428232b2001ab8df464f14a382661b936543336e2960
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:ad4f91f984b4b898012963d52e6ebcd166ec03e76e26eba3191ecb87f51fb18d
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:c15d4fd827f554291c14851f187947f4109c0e20257fa2f1e141b32669569358
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:5656cd188d054551489d45be996266d03e3faa8814b8bfc0f360cf0dceb83cbe
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:424b30dccd773f69c079b35a42e8c3e339246c5d490e14e3fcfb8bd8144a9160
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:879f3dcc24ffe795472779f777d955666ba4041c17e717d5e165cd7f12d7955a
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:985193caf9bc3698c4d85aeedd1cfea09bccda881636d67885387b2a83c65ebd
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:b56269cb849afca34e7f82cc8e01a44bb32af6126f24db28019a58e556d4cb3c
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:9bc0b730570ab55fbeca83075949f4caa96ed5c37637420a59cd4a7db01ea840
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:532dca6facb584c1261c0480c499af9f857c70dc390e0d46e46afec2561468af
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9d9e466b7c367d0406406550c74bc790aa68f3f875ea4f652eb28f2e25a05458
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:54bc26a0fddc160a4e0d1a813aec78abb3c9cb77db5039b74c23de759faa67f5
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:78a6bed6fcb6962e5689b92806789b72a66db21b7c607b392a573c98593f2c09
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4dbe1b7a1a36990c0a31d5a66f3cc4ca2954c9e1ec3b87431113a4111530277d
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:48cf13e71953cea6f62354289f0d8ece7a8a481ee0bb2c76eb9352d438db9ca3
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:0d368459c98c5788dd55d2add5b4b7d438b9a8c814ea8aee2222ea956699cbb6
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:c25e3152d6a6b6d5cd86d311bf567e594b9948440785209cd8d7986076a5a10d
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:f59ae21773dfc207454a709ac5f7aad2be6a6a20ac88da26c379edeed54537b8
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:24cf79efa89bbfa495665b1711a27531b7ab0813695a89b2c38c96c2e792eb24
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:6e1766784b684561666134d41dca5331b80d50e63695fe7b1a1d6d4cdb73f0bd
        - registry.redhat.io/datagrid/datagrid-8-rhel9-operator@sha256:d719d778b0848588a7e61d3c3fa689763f51a6533af4cf9e1c738c4ed8fac760
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:554b831a0803d55a4e3b7022011b5a5f6d2b4b65009065a9b28bc8e55fc3ad46
        version: 8.5.11
      entries:
      - name: datagrid-operator.v8.5.11
        version: 8.5.11
      - name: datagrid-operator.v8.5.10
        version: 8.5.10
      - name: datagrid-operator.v8.5.9
        version: 8.5.9
      - name: datagrid-operator.v8.5.8
        version: 8.5.8
      - name: datagrid-operator.v8.5.7
        version: 8.5.7
      - name: datagrid-operator.v8.5.6
        version: 8.5.6
      - name: datagrid-operator.v8.5.5
        version: 8.5.5
      - name: datagrid-operator.v8.5.4
        version: 8.5.4
      - name: datagrid-operator.v8.5.3
        version: 8.5.3
      - name: datagrid-operator.v8.5.2
        version: 8.5.2
      - name: datagrid-operator.v8.5.1
        version: 8.5.1
      - name: datagrid-operator.v8.5.0
        version: 8.5.0
      - name: datagrid-operator.v8.4.16
        version: 8.4.16
      - name: datagrid-operator.v8.4.15
        version: 8.4.15
      - name: datagrid-operator.v8.4.14
        version: 8.4.14
      - name: datagrid-operator.v8.4.13
        version: 8.4.13
      - name: datagrid-operator.v8.4.12
        version: 8.4.12
      - name: datagrid-operator.v8.4.11
        version: 8.4.11
      - name: datagrid-operator.v8.4.10
        version: 8.4.9
      - name: datagrid-operator.v8.4.8
        version: 8.4.8
      - name: datagrid-operator.v8.4.7
        version: 8.4.7
      - name: datagrid-operator.v8.4.6
        version: 8.4.6
      - name: datagrid-operator.v8.4.5
        version: 8.4.5
      - name: datagrid-operator.v8.4.4
        version: 8.4.4
      - name: datagrid-operator.v8.4.3
        version: 8.4.3
      - name: datagrid-operator.v8.4.2
        version: 8.4.2
      - name: datagrid-operator.v8.4.1
        version: 8.4.1
      - name: datagrid-operator.v8.4.0
        version: 8.4.0
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: stable
    defaultChannel: stable
    packageName: datagrid
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeRocketCI
      provider-url: https://docs.kuberocketci.io/
    name: sonar-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sonar-operator.v3.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Sonar",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonar-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonar",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonar-sample"
                },
                "spec": {
                  "defaultPermissionTemplate": "edp-default",
                  "secret": "sonar-secret",
                  "settings": [
                    {
                      "key": "sonar.global.exclusions",
                      "values": [
                        "**/*.jsx",
                        "**/*.tsx"
                      ]
                    },
                    {
                      "fieldValues": {
                        "beginBlockRegexp": ".*",
                        "endBlockRegexp": ".*"
                      },
                      "key": "sonar.issue.ignore.block"
                    },
                    {
                      "key": "sonar.dbcleaner.hoursBeforeKeepingOnlyOneSnapshotByDay",
                      "value": "20"
                    }
                  ],
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonargroup-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonargroup",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonargroup-sample"
                },
                "spec": {
                  "description": "Developers",
                  "name": "sonar-developers",
                  "permissions": [
                    "admin"
                  ],
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarPermissionTemplate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarpermissiontemplate-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarpermissiontemplate",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarpermissiontemplate-sample"
                },
                "spec": {
                  "default": true,
                  "description": "EDP permission templates",
                  "groupsPermissions": {
                    "sonar-developers": [
                      "scan"
                    ],
                    "sonar-users": [
                      "issueadmin"
                    ]
                  },
                  "name": "edp-default",
                  "projectKeyPattern": ".+",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityGate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarqualitygate-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualitygate",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarqualitygate-sample"
                },
                "spec": {
                  "conditions": {
                    "new_duplicated_lines_density": {
                      "error": "1",
                      "op": "GT"
                    }
                  },
                  "default": true,
                  "name": "sonarqualitygate-sample",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityProfile",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarqualityprofile-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualityprofile",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarqualityprofile-sample"
                },
                "spec": {
                  "default": true,
                  "language": "go",
                  "name": "sonarqualityprofile-sample",
                  "rules": {
                    "go:S1151": {
                      "params": "max=\"6\"",
                      "severity": "MAJOR"
                    }
                  },
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonaruser-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonaruser",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonaruser-sample"
                },
                "spec": {
                  "email": "sonaruser@mail.com",
                  "groups": [
                    "sonar-administrators"
                  ],
                  "login": "sonaruser",
                  "name": "sonar user",
                  "permissions": [
                    "scan"
                  ],
                  "secret": "sonaruser-secret",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/sonar-operator:3.1.0
          createdAt: "2023-09-08T12:53:57Z"
          description: An Operator for managing Sonar entities.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-sonar-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonarGroup is the Schema for the sonar group API.
            displayName: Sonar Group
            kind: SonarGroup
            name: sonargroups.edp.epam.com
            version: v1alpha1
          - description: SonarPermissionTemplate is the Schema for the sonar permission
              template API.
            displayName: Sonar Permission Template
            kind: SonarPermissionTemplate
            name: sonarpermissiontemplates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityGate is the Schema for the sonarqualitygates
              API
            displayName: Sonar Quality Gate
            kind: SonarQualityGate
            name: sonarqualitygates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityProfile is the Schema for the sonarqualityprofiles
              API
            displayName: Sonar Quality Profile
            kind: SonarQualityProfile
            name: sonarqualityprofiles.edp.epam.com
            version: v1alpha1
          - description: Sonar is the Schema for the sonars API.
            displayName: Sonar
            kind: Sonar
            name: sonars.edp.epam.com
            version: v1alpha1
          - description: SonarUser is the Schema for the sonarusers API
            displayName: Sonar User
            kind: SonarUser
            name: sonarusers.edp.epam.com
            version: v1alpha1
        description: |
          Sonar Operator is an operator that is responsible for establishing
          a connection to provided SonarQube server, reconciling Sonar entities (users,
          groups, permission templates, etc) according to the created CRs.
        displayName: EDP Sonar Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - analysis
        - automation
        - code
        - continuous
        - devops
        - edp
        - inspection
        - metrics
        - metrics
        - monitoring
        - operator
        - quality
        - security
        - software
        - static
        - technical
        - vulnerabilities
        links:
        - name: Sonar Operator
          url: https://github.com/epam/edp-sonar-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/sonar-operator:3.1.0
        version: 3.1.0
      entries:
      - name: sonar-operator.v3.1.0
        version: 3.1.0
      name: alpha
    - currentCSV: sonar-operator.v3.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Sonar",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonar"
                  },
                  "name": "sonar-sample"
                },
                "spec": {
                  "defaultPermissionTemplate": "edp-default",
                  "secret": "sonar-secret",
                  "settings": [
                    {
                      "key": "sonar.global.exclusions",
                      "values": [
                        "**/*.jsx",
                        "**/*.tsx"
                      ]
                    },
                    {
                      "fieldValues": {
                        "beginBlockRegexp": ".*",
                        "endBlockRegexp": ".*"
                      },
                      "key": "sonar.issue.ignore.block"
                    },
                    {
                      "key": "sonar.dbcleaner.hoursBeforeKeepingOnlyOneSnapshotByDay",
                      "value": "20"
                    }
                  ],
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonargroup"
                  },
                  "name": "sonargroup-sample"
                },
                "spec": {
                  "description": "Developers",
                  "name": "sonar-developers",
                  "permissions": [
                    "admin"
                  ],
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarPermissionTemplate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarpermissiontemplate"
                  },
                  "name": "sonarpermissiontemplate-sample"
                },
                "spec": {
                  "default": true,
                  "description": "EDP permission templates",
                  "groupsPermissions": {
                    "sonar-developers": [
                      "scan"
                    ],
                    "sonar-users": [
                      "issueadmin"
                    ]
                  },
                  "name": "edp-default",
                  "projectKeyPattern": ".+",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityGate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualitygate"
                  },
                  "name": "sonarqualitygate-sample"
                },
                "spec": {
                  "conditions": {
                    "new_duplicated_lines_density": {
                      "error": "1",
                      "op": "GT"
                    }
                  },
                  "default": true,
                  "name": "sonarqualitygate-sample",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityProfile",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualityprofile"
                  },
                  "name": "sonarqualityprofile-sample"
                },
                "spec": {
                  "default": true,
                  "language": "go",
                  "name": "sonarqualityprofile-sample",
                  "rules": {
                    "go:S1151": {
                      "params": "max=\"6\"",
                      "severity": "MAJOR"
                    }
                  },
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonaruser"
                  },
                  "name": "sonaruser-sample"
                },
                "spec": {
                  "email": "sonaruser@mail.com",
                  "groups": [
                    "sonar-administrators"
                  ],
                  "login": "sonaruser",
                  "name": "sonar user",
                  "permissions": [
                    "scan"
                  ],
                  "secret": "sonaruser-secret",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/sonar-operator:3.3.0
          createdAt: "2025-04-22T11:39:32Z"
          description: An Operator for managing Sonar entities.
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/epam/edp-sonar-operator
          support: KubeRocketCI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonarGroup is the Schema for the sonar group API.
            displayName: Sonar Group
            kind: SonarGroup
            name: sonargroups.edp.epam.com
            version: v1alpha1
          - description: SonarPermissionTemplate is the Schema for the sonar permission
              template API.
            displayName: Sonar Permission Template
            kind: SonarPermissionTemplate
            name: sonarpermissiontemplates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityGate is the Schema for the sonarqualitygates
              API
            displayName: Sonar Quality Gate
            kind: SonarQualityGate
            name: sonarqualitygates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityProfile is the Schema for the sonarqualityprofiles
              API
            displayName: Sonar Quality Profile
            kind: SonarQualityProfile
            name: sonarqualityprofiles.edp.epam.com
            version: v1alpha1
          - description: Sonar is the Schema for the sonars API.
            displayName: Sonar
            kind: Sonar
            name: sonars.edp.epam.com
            version: v1alpha1
          - description: SonarUser is the Schema for the sonarusers API.
            displayName: Sonar User
            kind: SonarUser
            name: sonarusers.edp.epam.com
            version: v1alpha1
        description: |
          Sonar Operator is an operator that is responsible for establishing
          a connection to provided SonarQube server, reconciling Sonar entities (users,
          groups, permission templates, etc) according to the created CRs.
        displayName: EDP Sonar Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - analysis
        - automation
        - code
        - continuous
        - devops
        - edp
        - inspection
        - metrics
        - metrics
        - monitoring
        - operator
        - quality
        - security
        - software
        - static
        - technical
        - vulnerabilities
        links:
        - name: Sonar Operator
          url: https://github.com/epam/edp-sonar-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: KubeRocketCI
          url: https://docs.kuberocketci.io/
        relatedImages:
        - docker.io/epamedp/sonar-operator:3.3.0
        version: 3.3.0
      entries:
      - name: sonar-operator.v3.3.0
        version: 3.3.0
      - name: sonar-operator.v3.2.0
        version: 3.2.0
      name: stable
    defaultChannel: stable
    packageName: sonar-operator
    provider:
      name: KubeRocketCI
      url: https://docs.kuberocketci.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.6.5-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6.5
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/kubeturbo-operator:8.6.5
        version: 8.6.5-beta.1
      entries:
      - name: kubeturbo-operator.v8.6.5-beta.1
        version: 8.6.5-beta.1
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: beta
    - currentCSV: kubeturbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/kubeturbo-operator:8.6
        version: 8.6.0
      entries:
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: stable
    - currentCSV: kubeturbo-operator.v8.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.4
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/kubeturbo-operator:8.4
        version: 8.4.0
      entries:
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.4.0
    - currentCSV: kubeturbo-operator.v8.5.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.5
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/kubeturbo-operator:8.5
        version: 8.5.0
      entries:
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.5.0
    - currentCSV: kubeturbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/kubeturbo-operator:8.6
        version: 8.6.0
      entries:
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.6.0
    defaultChannel: stable
    packageName: kubeturbo
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Silicom Ltd.
      provider-url: http://www.silicom-usa.com
    name: silicom-sts-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: silicom-sts-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 3,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-BC-8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/mode": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "hp-ocp",
                  "namespase": "openshift-operators"
                },
                "spec": {
                  "dummy": 0
                },
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicom-sts/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicom-sts/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/rmr/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - quay.io/silicom/gpsd@sha256:e479c542e061b5422865f2617ed0563794f265c6e6660e5a9d204c9e1cb2f346
        - quay.io/silicom/tsyncd@sha256:3f1cbd4a8b5d62d7d3e525af5f26d9f27b8c08e886b0639a1e368163e166b537
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - quay.io/silicom/sts-operator@sha256:7d474d0aa91ac57d8f5817b033b67b511c8601e0dbd83d808a69004bd24b4e60
        - quay.io/silicom/grpc-tsyncd@sha256:631bb754742e1a24376913baed60e480a6e6ae988ad9a3892254ff96996c15d1
        - quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
        - quay.io/silicom/phc2sys@sha256:b067167a1a5d93e9fe715c7774f6cdbbcd93cac27780f022e44dcb8a4d4d7bb9
        - quay.io/silicom/sts-plugin@sha256:233bc95f8617a06bd452548ec116d91551dd4933970860d8d5ccb81f75ad2220
        - quay.io/silicom/ice-driver-src@sha256:569c8333e21c7c022ea71fa5b418404899040d5b3a3a4128eeac21f62aacaf3e
        version: 0.1.3
      entries:
      - name: silicom-sts-operator.v0.1.3
        version: 0.1.3
      - name: silicom-sts-operator.v0.0.9
        version: 0.0.9
      - name: silicom-sts-operator.v0.0.8
        version: 0.0.8
      name: alpha
    - currentCSV: silicom-sts-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 3,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-BC-8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/mode": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "hp-ocp",
                  "namespase": "openshift-operators"
                },
                "spec": {
                  "dummy": 0
                },
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicom-sts/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicom-sts/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/rmr/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - quay.io/silicom/sts-plugin@sha256:233bc95f8617a06bd452548ec116d91551dd4933970860d8d5ccb81f75ad2220
        - quay.io/silicom/sts-operator@sha256:7d474d0aa91ac57d8f5817b033b67b511c8601e0dbd83d808a69004bd24b4e60
        - quay.io/silicom/tsyncd@sha256:3f1cbd4a8b5d62d7d3e525af5f26d9f27b8c08e886b0639a1e368163e166b537
        - quay.io/silicom/grpc-tsyncd@sha256:631bb754742e1a24376913baed60e480a6e6ae988ad9a3892254ff96996c15d1
        - quay.io/silicom/ice-driver-src@sha256:569c8333e21c7c022ea71fa5b418404899040d5b3a3a4128eeac21f62aacaf3e
        - quay.io/silicom/gpsd@sha256:e479c542e061b5422865f2617ed0563794f265c6e6660e5a9d204c9e1cb2f346
        - quay.io/silicom/phc2sys@sha256:b067167a1a5d93e9fe715c7774f6cdbbcd93cac27780f022e44dcb8a4d4d7bb9
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
        version: 0.1.3
      entries:
      - name: silicom-sts-operator.v0.1.3
        version: 0.1.3
      - name: silicom-sts-operator.v0.0.9
        version: 0.0.9
      - name: silicom-sts-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: silicom-sts-operator
    provider:
      name: Silicom Ltd.
      url: http://www.silicom-usa.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-flashsystemcluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "flashsystemcluster.odf.ibm.com/v1alpha1",
                  "name": "my-flashsystemcluster",
                  "namespace": "openshift-storage"
                }
              },
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-storagecluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "storagecluster.ocs.openshift.io/v1",
                  "name": "my-storagecluster",
                  "namespace": "openshift-storage"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-console"]'
          containerImage: registry.redhat.io/odf4/odf-rhel9-operator@sha256:8d377d2d8a94e461a6eca8d1c3b4f5cf1f9ec1ba6a03044304403b58bbbabaa4
          createdAt: "2025-09-12T07:34:03Z"
          description: OpenShift Data Foundation provides a common control plane for
            storage solutions on OpenShift Container Platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.16"}]'
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operator.openshift.io/uninstall-message: 'Warning: deleting the operand
            (StorageSystem) will result in data loss'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "odf.openshift.io/v1alpha1",
              "kind": "StorageSystem",
              "metadata": {
                "name": "ocs-storagecluster-storagesystem",
                "namespace": "openshift-storage"
              },
              "spec": {
                "kind": "storagecluster.ocs.openshift.io/v1",
                "name": "ocs-storagecluster",
                "namespace": "openshift-storage"
              }
            }
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/odf-operator
          support: Red Hat
          vendors.odf.openshift.io/kind: '["storagecluster.ocs.openshift.io/v1", "flashsystemcluster.odf.ibm.com/v1alpha1"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageSystem is the Schema for the storagesystems API
            displayName: Storage System
            kind: StorageSystem
            name: storagesystems.odf.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Data Foundation** deploys three operators.

          ### OpenShift Data Foundation
          The OpenShift Data Foundation operator is the primary operator for
          OpenShift Data Foundation. It serves to facilitate the other operators in
          OpenShift Data Foundation by performing administrative tasks outside their
          scope as well as watching and configuring their CustomResources.

          ### Rook
          [Rook][1] deploys and manages Ceph on OpenShift, which provides block,
          file and object storage.

          ### NooBaa operator
          The NooBaa operator deploys and manages the [NooBaa][2] Multi-Cloud Gateway
          on OpenShift, which provides object storage.

          # Core Capabilities
          * **Self-managing service:** No matter which supported storage technologies
          you choose, OpenShift Data Foundation ensures that resources can be
          deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Data Foundation you
          can either build dedicated storage clusters or hyper-converged clusters
          where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Data Foundation:**
          OpenShift Data Foundation integrates Ceph with multiple storage
          presentations including object storage (compatible with S3), block storage,
          and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Data Foundation efficiently
          distributes and replicates your data across your cluster to minimize the
          risk of data loss. With snapshots, cloning, and versioning, no more losing
          sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your
          datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data
          storage for your workloads, using a single namespace.

          # Subscriptions
          OpenShift Data Foundation requires a matching 1:1 ratio with the OpenShift
          subscription used. The Essential subscription is included in the OpenShift
          Platform Plus subscription. Disaster Recovery features, external mode, and
          KMS integration for encryption at rest require a bundle of OpenShift
          Platform Plus with OpenShift Data Foundation Advanced subscription.

          [1]: https://rook.io
          [2]: https://www.noobaa.io
        displayName: OpenShift Data Foundation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/odf4/odf-rhel9-operator@sha256:8d377d2d8a94e461a6eca8d1c3b4f5cf1f9ec1ba6a03044304403b58bbbabaa4
        - registry.redhat.io/odf4/odf-console-rhel9@sha256:5d0d27a091a74e65985d2f047d3f6322043f2ae80fa5ff3192ec84cce92f143d
        version: 4.15.20-rhodf
      entries:
      - name: odf-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odf-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odf-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odf-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odf-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odf-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odf-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odf-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odf-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odf-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odf-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odf-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odf-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odf-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odf-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odf-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odf-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odf-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odf-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odf-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odf-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-flashsystemcluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "flashsystemcluster.odf.ibm.com/v1alpha1",
                  "name": "my-flashsystemcluster",
                  "namespace": "openshift-storage"
                }
              },
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-storagecluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "storagecluster.ocs.openshift.io/v1",
                  "name": "my-storagecluster",
                  "namespace": "openshift-storage"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-console"]'
          containerImage: registry.redhat.io/odf4/odf-rhel9-operator@sha256:20cb12259667539a0244cfde37297b975e2682833fbe537caecb5107ebecf15f
          createdAt: "2025-09-12T06:33:04Z"
          description: OpenShift Data Foundation provides a common control plane for
            storage solutions on OpenShift Container Platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.17"}]'
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operator.openshift.io/uninstall-message: 'Warning: deleting the operand
            (StorageSystem) will result in data loss'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "odf.openshift.io/v1alpha1",
              "kind": "StorageSystem",
              "metadata": {
                "name": "ocs-storagecluster-storagesystem",
                "namespace": "openshift-storage"
              },
              "spec": {
                "kind": "storagecluster.ocs.openshift.io/v1",
                "name": "ocs-storagecluster",
                "namespace": "openshift-storage"
              }
            }
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/odf-operator
          support: Red Hat
          vendors.odf.openshift.io/kind: '["storagecluster.ocs.openshift.io/v1", "flashsystemcluster.odf.ibm.com/v1alpha1"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageSystem is the Schema for the storagesystems API
            displayName: Storage System
            kind: StorageSystem
            name: storagesystems.odf.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Data Foundation** deploys three operators.

          ### OpenShift Data Foundation
          The OpenShift Data Foundation operator is the primary operator for
          OpenShift Data Foundation. It serves to facilitate the other operators in
          OpenShift Data Foundation by performing administrative tasks outside their
          scope as well as watching and configuring their CustomResources.

          ### Rook
          [Rook][1] deploys and manages Ceph on OpenShift, which provides block,
          file and object storage.

          ### NooBaa operator
          The NooBaa operator deploys and manages the [NooBaa][2] Multi-Cloud Gateway
          on OpenShift, which provides object storage.

          # Core Capabilities
          * **Self-managing service:** No matter which supported storage technologies
          you choose, OpenShift Data Foundation ensures that resources can be
          deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Data Foundation you
          can either build dedicated storage clusters or hyper-converged clusters
          where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Data Foundation:**
          OpenShift Data Foundation integrates Ceph with multiple storage
          presentations including object storage (compatible with S3), block storage,
          and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Data Foundation efficiently
          distributes and replicates your data across your cluster to minimize the
          risk of data loss. With snapshots, cloning, and versioning, no more losing
          sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your
          datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data
          storage for your workloads, using a single namespace.

          # Subscriptions
          OpenShift Data Foundation requires a matching 1:1 ratio with the OpenShift
          subscription used. The Essential subscription is included in the OpenShift
          Platform Plus subscription. Disaster Recovery features, external mode, and
          KMS integration for encryption at rest require a bundle of OpenShift
          Platform Plus with OpenShift Data Foundation Advanced subscription.

          [1]: https://rook.io
          [2]: https://www.noobaa.io
        displayName: OpenShift Data Foundation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/odf4/odf-rhel9-operator@sha256:20cb12259667539a0244cfde37297b975e2682833fbe537caecb5107ebecf15f
        - registry.redhat.io/odf4/odf-console-rhel9@sha256:8bd822a3410e520fc57ac69c7599831173003e2c34b980c517f79aca97210257
        version: 4.16.16-rhodf
      entries:
      - name: odf-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odf-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odf-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odf-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odf-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odf-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odf-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odf-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odf-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odf-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odf-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odf-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odf-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odf-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odf-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odf-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odf-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odf-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: jaeger-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jaeger-operator.v1.65.0-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "jaegertracing.io/v1",
                "kind": "Jaeger",
                "metadata": {
                  "name": "my-jaeger"
                },
                "spec": {
                  "strategy": "allInOne"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/rhosdt/jaeger-rhel8-operator@sha256:51a8685e4dc70888c991996f149cd42233049d18f702ec72b066d6dad77a07c9
          createdAt: 25 Aug 2025, 13:30
          description: Provides tracing, monitoring and troubleshooting for microservices-based
            distributed systems
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.65.0-4'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-distributed-tracing
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/os-observability/konflux-jaeger
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Jaeger is the Schema for the jaegers API
            displayName: Jaeger
            kind: Jaeger
            name: jaegers.jaegertracing.io
            version: v1
        description: |2-

          ## Operator deprecation warning

          This Operator is deprecated and not recommended for new deployments. This Operator will be removed from the `redhat-operators` catalog in a future release.

          If this is not the first time you are installing it, migrate to the Red Hat build of OpenTelemetry Operator and the Tempo Operator for distributed tracing collection and storage. For more information, see "Migrating" in the Red Hat build of OpenTelemetry documentation, "Installing" in the Red Hat build of OpenTelemetry documentation, and "Installing" in the distributed tracing platform (Tempo) documentation.

          ## About

          Red Hat OpenShift distributed tracing platform based on Jaeger. Jaeger is project inspired by [Dapper](https://research.google.com/pubs/pub36356.html) and [OpenZipkin](http://zipkin.io/). It is a distributed tracing system released as open source by Uber Technologies. It is used for monitoring and troubleshooting microservices-based distributed systems.

          ### Core capabilities
          Jaeger is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed context propagation
          * Distributed transaction monitoring
          * Root cause analysis
          * Service dependency analysis
          * Performance / latency optimization
          * OpenTracing compatible data model
          * Multiple storage backends: Elasticsearch, Memory.

          ### Operator features

          * **Multiple modes** - Provides `allInOne`, `production` and `streaming` [modes of deployment](https://www.jaegertracing.io/docs/latest/operator/#deployment-strategies).

          * **Configuration** - The Operator manages [configuration information](https://www.jaegertracing.io/docs/latest/operator/#configuring-the-custom-resource) when installing Jaeger instances.

          * **Storage** - [Configure storage](https://www.jaegertracing.io/docs/latest/operator/#storage-options) used by Jaeger. By default, `memory` is used. Other options include `elasticsearch`. The operator can delegate creation of an Elasticsearch cluster to the Elasticsearch Operator if deployed.

          * **Agent** - can be deployed as [sidecar](https://www.jaegertracing.io/docs/latest/operator/#auto-injecting-jaeger-agent-sidecars) (default) and/or [daemonset](https://www.jaegertracing.io/docs/latest/operator/#installing-the-agent-as-daemonset).

          * **UI** - Optionally setup secure route to provide [access to the Jaeger UI](https://www.jaegertracing.io/docs/latest/operator/#accessing-the-jaeger-console-ui).

          ### Before you start
          1. Ensure that the appropriate storage solution, that will be used by the Jaeger instance, is available and configured.
          2. If intending to deploy an Elasticsearch cluster via the Jaeger custom resource, then the Elasticsearch Operator must first be installed.

          ### Support & Troubleshooting

          Red Hat OpenShift distributed tracing Jaeger is available and supported as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/4.17/html/distributed_tracing/distributed-tracing-platform-jaeger) is recommended when installing, configuring, and managing the Operator and instances.
          Support is provided to Red Hat OpenShift entitled customers subject to the [Production Scope for Coverage](https://access.redhat.com/support/offerings/production/soc) and the [Red Hat OpenShift distributed tracing Life Cycle](https://access.redhat.com/support/policy/updates/openshift#jaeger).
        displayName: Red Hat OpenShift distributed tracing platform
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tracing
        - monitoring
        - troubleshooting
        links:
        - name: Jaeger Operator Source Code
          url: https://github.com/jaegertracing/jaeger-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:667d6545b92aa8381f65369de3b5a6781d1de5b4bcfa58ccde4721b6b0406f3d
        - registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:d92003d6cf0b4f3dae11f849355c9179e78e768f3b7dbeaf3224a4a019ec5d5d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e77efd450e757772d0753a73062753dabd08e53f21315e9b87f4cc1efa332661
        - registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:1d586f9aafd1cf847a14f509692ce54956b0e754a9a04beaa7a75c480f86717c
        - registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:7b021ed467d35b2bcae87765dfc64f639b03bf71f7e34bbea1dfad16626847b9
        - registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:1f4a9daf152420d9ca7cf89c38ae5550b12dd274bada389c7b85ea4cbcbc4fb7
        - registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:fd8110529daa8cf3e7f9fcf71346adc5f48cb05717ebbad1c25bc61c94f31656
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        - registry.redhat.io/rhosdt/jaeger-rhel8-operator@sha256:51a8685e4dc70888c991996f149cd42233049d18f702ec72b066d6dad77a07c9
        version: 1.65.0-4
      entries:
      - name: jaeger-operator.v1.65.0-4
        version: 1.65.0-4
      - name: jaeger-operator.v1.65.0-3
        version: 1.65.0-3
      - name: jaeger-operator.v1.65.0-2
        version: 1.65.0-2
      - name: jaeger-operator.v1.65.0-1
        version: 1.65.0-1
      - name: jaeger-operator.v1.62.0-2
        version: 1.62.0-2
      - name: jaeger-operator.v1.62.0-1
        version: 1.62.0-1
      - name: jaeger-operator.v1.57.0-10
        version: 1.57.0-10
      - name: jaeger-operator.v1.57.0-10-0.1730817420.p
        version: 1.57.0-10+0.1730817420.p
      - name: jaeger-operator.v1.57.0-7
        version: 1.57.0-7
      - name: jaeger-operator.v1.57.0-6
        version: 1.57.0-6
      - name: jaeger-operator.v1.57.0-5
        version: 1.57.0-5
      - name: jaeger-operator.v1.53.0-4
        version: 1.53.0-4
      - name: jaeger-operator.v1.53.0-3
        version: 1.53.0-3
      - name: jaeger-operator.v1.51.0-1
        version: 1.51.0-1
      - name: jaeger-operator.v1.47.1-5
        version: 1.47.1-5
      - name: jaeger-operator.v1.42.0-5-0.1687199951.p
        version: 1.42.0-5+0.1687199951.p
      - name: jaeger-operator.v1.42.0-5
        version: 1.42.0-5
      - name: jaeger-operator.v1.34.1-5
        version: 1.34.1-5
      - name: jaeger-operator.v1.30.2
        version: 1.30.2
      name: stable
    defaultChannel: stable
    packageName: jaeger-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator.v1.6.3
        version: 1.6.3
      - name: dynatrace-operator.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator.v1.4.0
        version: 1.4.0
      - name: dynatrace-operator.v1.3.2
        version: 1.3.2
      - name: dynatrace-operator.v1.3.1
        version: 1.3.1
      - name: dynatrace-operator.v1.3.0
        version: 1.3.0
      - name: dynatrace-operator.v0.6.0
        version: 0.6.0
      - name: dynatrace-operator.v0.5.1
        version: 0.5.1
      - name: dynatrace-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco
      provider-url: ""
    name: aci-containers-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aci-containers-operator.v1.0.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"aci.ctrl/v1alpha1","kind":"AciContainersOperator","metadata":{"name":"acicnioperator","namespace":"aci-containers-system"},"spec":{}}]'
          capabilities: Seamless Upgrades
          categories: Networking
          certified: "false"
          containerImage: quay.io/noiro/aci-containers-operator@sha256:2b2d685e254c206be5371285852a8f95d0318af69cb85b786427969e035c125c
          createdAt: "2025-04-10T12:00:00Z"
          description: The Cisco Application Centric Infrastructure (ACI) Containers
            Operator brings the ACI CNI Plugin and Policy model to Kubernetes clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://github.com/noironetworks/aci-containers/blob/master/README.md
          repository: https://github.com/noironetworks/aci-containers
          support: acisupport@cisco.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aci Containers Operator
            displayName: AciContainersOperator
            kind: AciContainersOperator
            name: acicontainersoperators.aci.ctrl
            version: v1alpha1
        description: |
          ## ACI CNI Plugin

          The Cisco Application Centric Infrastructure (ACI) CNI plugin brings the ACI Networking and Policy model to Kubernetes clusters that reside on-prem or in the cloud. It is fully open source and relies on the Opflex Protocol to program Open vSwitch instances running on the Kubernetes nodes. It provides IP address management, L2/L3 networking, load balancing, and security functions for container workloads.

          The Cisco ACI CNI plugin for the Red Hat OpenShift Container Platform provides a single, programmable network infrastructure that eliminates bottlenecks between OpenShift workloads and the rest of the infrastructure while providing enterprise-grade security and offers flexible micro segmentation possibilities.

          Unified Networking - By extending the integrated overlay of the ACI fabric to OpenShift workloads, Pod and Service endpoints are managed as native ACI network resources at the same level as Bare Metal or Virtual Machines.

          No Egress Router Required - Applications running in OpenShift Pods that need high-bandwidth, low-latency traffic for data transfers external to the cluster suffer the bottleneck imposed by the egress router implementation. Pods can easily connect with other fabric or non-fabric endpoints leveraging the combined hardware and software distributed routing and switching provided by ACI and the ACI CNI plugin.

          Simplified Operations and Enhanced Visibility - Whether policies and networking are defined by the devops or fabric or security admins, the APIC becomes the single source of truth for operations and security governance. Kubernetes workload visibility is provided to Network Operators within their familiar Cisco tools.

          Hardware-assisted Load Balancing - Ingress connections to LoadBalancer-type services using ACI’s Policy Based Redirect technology moves load balancing into the fabric's hardware.

          Granular Security - Security can be implemented by using native NetworkPolicy or by using ACI EPGs and contracts, or both models complementing each other. Network policy of container workloads becomes a shared responsibility while ensuring global compliance and tenant constraints are enforced.
        displayName: Cisco ACI Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cisco
        - aci
        links:
        - name: Cisco ACI Homepage
          url: https://www.cisco.com/c/en/us/td/docs/switches/datacenter/aci/apic/sw/kb/b_Kubernetes_Integration_with_ACI.html
        maintainers:
        - email: abhishs9@cisco.com
          name: 'Abhishek Sharma '
        - email: sumitnaiksatam@gmail.com
          name: Sumit Naiksatam
        - email: vlella@cisco.com
          name: Vishal Lella
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Cisco
        relatedImages:
        - quay.io/noiro/aci-containers-operator@sha256:2b2d685e254c206be5371285852a8f95d0318af69cb85b786427969e035c125c
        - quay.io/noiro/acc-provision-operator@sha256:9c8fa6d705602af13d539fb919e72e7ee5064c6998fc9ca95d10108e986c68fe
        version: 1.0.10
      entries:
      - name: aci-containers-operator.v1.0.10
        version: 1.0.10
      name: stable
    defaultChannel: stable
    packageName: aci-containers-operator
    provider:
      name: Cisco
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Accuknox
      provider-url: https://accuknox.com
    name: kubearmor-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubearmor-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          com.redhat.openshift.versions: v4.12
          createdAt: "2023-05-01T05:17:07Z"
          description: The KubeArmor Operator deloys and manages KubeArmor daemonset
            and its componenets within a cluster.
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.28.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/kubearmor/KubeArmor
          support: Accuknox
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          The KubeArmor Operator deploys and manages KubeArmor and its components:
          - KubeArmor Daemonset
          - KubeArmor Relay Server
          - KubeArmor Controller to manage CRDs (KubeArmorPolicy, KubeArmorHostPolicy)

          KubeArmor is a cloud-native runtime security enforcement system that restricts the behavior (such as process
          execution, file access, and networking operations) of pods, containers, and nodes (VMs) at the system level.

          KubeArmor leverages Linux security modules (LSMs) such as AppArmor, SELinux, or BPF-LSM to enforce the user-specified policies. KubeArmor generates rich alerts/telemetry events with container/pod/namespace identities by leveraging eBPF.

          ## Documentation

          github: [https://github.com/kubearmor/KubeArmor](https://github.com/kubearmor/KubeArmor)
          website: [https://kubearmor.io/](https://kubearmor.io/)
        displayName: Kubearmor Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubearmor-operator
        - runtime-security
        - security
        - cloud-native
        links:
        - name: Kubearmor
          url: https://kubearmor.io
        maintainers:
        - email: ramakant@accuknox.com
          name: Ramakant Sharma
        maturity: alpha
        provider:
          name: Accuknox
          url: https://www.accuknox.com/
        relatedImages:
        - docker.io/kubearmor/kubearmor-init@sha256:44fa534587972c5f85e3e6f362dd8268e9088a33c5657a21b72a7e8563164fb5
        - docker.io/kubearmor/kubearmor-relay-server@sha256:695f9ade3382bcf858077e2a7b9b01e0713071517603778d72f7aca5a6d71323
        - docker.io/kubearmor/kubearmor-controller@sha256:ccec0fb8bb1e91bd48631cc89baf752b842f19c33b50edab08f8b33e57af64d4
        - docker.io/kubearmor/kubearmor-operator@sha256:8c2d56781a2adf78ee58f956e01417f38b6ad2f319feada6470e93f4f9bf0808
        - docker.io/kubearmor/kubearmor@sha256:52f7049ecdbfa779663d2cd44ddd7f36aecd8ccd2ca91e69709b628600029f13
        version: 1.0.0
      entries:
      - name: kubearmor-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: kubearmor-operator.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.kubearmor.com/v1",
                "kind": "KubeArmorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kubearmorconfig",
                    "app.kubernetes.io/part-of": "kubearmoroperator"
                  },
                    "name": "kubearmorconfig-default",
                    "namespace": "placeholder"
                },
                "spec": {
                  "adapters": {
                    "elasticsearch": {
                      "url": "",
                      "enabled": false,
                      "auth": {
                        "secretName": "elastic-secret",
                        "usernameKey": "username",
                        "passwordKey": "password",
                        "allowInsecureTLS": false,
                        "caCertSecretName": "",
                        "caCertKey": "ca.crt"
                      },
                      "alertsIndex": "kubearmor-alerts"
                    }
                  },
                  "recommendedPolicies": {
                      "enable": true,
                      "matchExpressions": [
                          {
                              "key": "namespace",
                              "operator": "NotIn",
                              "values": [
                                  "kube-system",
                                  "kubearmor",
                                  "openshift-infra",
                                  "openshift-ingress",
                                  "openshift-ingress-canary",
                                  "openshift-ingress-operator",
                                  "openshift-insights",
                                  "openshift-kni-infra",
                                  "openshift-kube-apiserver",
                                  "openshift-kube-apiserver-operator",
                                  "openshift-kube-controller-manager",
                                  "openshift-kube-controller-manager-operator",
                                  "openshift-kube-scheduler",
                                  "openshift-kube-scheduler-operator",
                                  "openshift-kube-storage-version-migrator",
                                  "openshift-kube-storage-version-migrator-operator",
                                  "openshift-logging",
                                  "openshift-machine-api",
                                  "openshift-machine-config-operator",
                                  "openshift-managed-node-metadata-operator",
                                  "openshift-managed-upgrade-operator",
                                  "openshift-marketplace",
                                  "openshift-monitoring",
                                  "openshift-multus",
                                  "openshift-must-gather-operator",
                                  "openshift-network-diagnostics",
                                  "openshift-network-node-identity",
                                  "openshift-network-operator",
                                  "openshift-node",
                                  "openshift-nutanix-infra",
                                  "openshift-oauth-apiserver",
                                  "openshift-observability-operator",
                                  "openshift-ocm-agent-operator",
                                  "openshift-openstack-infra",
                                  "openshift-operator-lifecycle-manager",
                                  "openshift-operators",
                                  "openshift-operators-redhat",
                                  "openshift-osd-metrics",
                                  "openshift-ovirt-infra",
                                  "openshift-ovn-kubernetes",
                                  "openshift-package-operator",
                                  "openshift-rbac-permissions",
                                  "openshift-route-controller-manager",
                                  "openshift-route-monitor-operator",
                                  "openshift-security",
                                  "openshift-service-ca",
                                  "openshift-service-ca-operator",
                                  "openshift-splunk-forwarder-operator",
                                  "openshift-sre-pruning",
                                  "openshift-user-workload-monitoring",
                                  "openshift-validation-webhook",
                                  "openshift-velero",
                                  "openshift-vsphere-infra",
                                  "openshift",
                                  "openshift-addon-operator",
                                  "openshift-apiserver",
                                  "openshift-apiserver-operator",
                                  "openshift-aqua",
                                  "openshift-authentication",
                                  "openshift-authentication-operator",
                                  "openshift-backplane",
                                  "openshift-backplane-cee",
                                  "openshift-backplane-csa",
                                  "openshift-backplane-cse",
                                  "openshift-backplane-csm",
                                  "openshift-backplane-managed-scripts",
                                  "openshift-backplane-mcs-tier-two",
                                  "openshift-backplane-mobb",
                                  "openshift-backplane-srep",
                                  "openshift-backplane-tam",
                                  "openshift-cloud-controller-manager",
                                  "openshift-cloud-controller-manager-operator",
                                  "openshift-cloud-credential-operator",
                                  "openshift-cloud-ingress-operator",
                                  "openshift-cloud-network-config-controller",
                                  "openshift-cloud-platform-infra",
                                  "openshift-cluster-csi-drivers",
                                  "openshift-cluster-machine-approver",
                                  "openshift-cluster-node-tuning-operator",
                                  "openshift-cluster-samples-operator",
                                  "openshift-cluster-storage-operator",
                                  "openshift-cluster-version",
                                  "openshift-codeready-workspaces",
                                  "openshift-config",
                                  "openshift-config-managed",
                                  "openshift-config-operator",
                                  "openshift-console",
                                  "openshift-console-operator",
                                  "openshift-console-user-settings",
                                  "openshift-controller-manager",
                                  "openshift-controller-manager-operator",
                                  "openshift-custom-domains-operator",
                                  "openshift-customer-monitoring",
                                  "openshift-deployment-validation-operator",
                                  "openshift-dns",
                                  "openshift-dns-operator",
                                  "openshift-etcd",
                                  "openshift-etcd-operator",
                                  "openshift-host-network",
                                  "openshift-image-registry"
                              ]
                          }
                      ]
                  },
                  "defaultCapabilitiesPosture": "audit",
                  "defaultFilePosture": "audit",
                  "defaultNetworkPosture": "audit",
                  "defaultVisibility": "process,network",
                  "enableStdOutLogs": false,
                  "enableStdOutAlerts": false,
                  "enableStdOutMsgs": false,
                  "seccompEnabled": false
                }
              }
            ]
          alm-examples-metadata: '{"kubearmorconfig-default":{"description":"Default
            KubeArmorConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Security, Monitoring
          certified: "true"
          com.redhat.openshift.versions: v4.9
          containerImage: docker.io/kubearmor/kubearmor-operator@sha256:d6c6b16e636474568835ded97cecb5ed17bc6004501227f6b6d63522ed57b01c
          createdAt: "2024-03-20T17:43:15Z"
          description: KubeArmor Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: kubearmor
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kubearmor/KubeArmor/blob/main/pkg/KubeArmorOperator/
          support: Accuknox
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeArmorConfig is the Schema for the kubearmorconfigs API
            kind: KubeArmorConfig
            name: kubearmorconfigs.operator.kubearmor.com
            version: v1
        description: |2-

          ## Prerequisites

          1. Create a new project for KubeArmor (Recommended)

             ```
             oc new-project kubearmor
             ```

           2. Create [KubeArmor SCC](https://github.com/kubearmor/KubeArmor/blob/main/pkg/KubeArmorOperator/config/rbac/kubearmor-scc.yaml)

             ```
             oc create -f https://raw.githubusercontent.com/kubearmor/KubeArmor/main/pkg/KubeArmorOperator/config/rbac/kubearmor-scc.yaml
             ```

             Note: please change the serviceaccount namespace if you are using some other namespace than the kubearmor.

          KubeArmor is a cloud-native runtime security enforcement system that restricts the behavior
          (such as process execution, file access, and networking operations) of pods, containers,
          and nodes (VMs) at the system level.
          KubeArmor leverages Linux security modules (LSMs) such as AppArmor, SELinux, or BPF-LSM to enforce the
          user-specified policies. KubeArmor generates rich alerts/telemetry events with container/pod/namespace
          identities by leveraging eBPF.

          ## Features and Benefits

          ### Harden Infrastructure

          safeguards vital paths like certificate bundles, enforces security rules based on MITRE, STIGs, and CIS guidelines,
          and restricts access to raw database tables, enhancing overall cluster security.

          ### Least Permissive Access

          controls sensitive asset access through process and network whitelisting,
          reducing security risks and maintaining system integrity.

          ### Application Behavior

          in-depth monitoring of process executions, file system access, service bindings,
          ingress and egress connections, and sensitive system call profiling

          ## Documentation

          github: [https://github.com/kubearmor/KubeArmor](https://github.com/kubearmor/KubeArmor)

          website: [https://kubearmor.io/](https://kubearmor.io/)

          ## Installing the KubeArmor Operator

          The KubeArmor Operator, available through Red Hat's Operator catalog, offers stable update channel.
          **Stable Channel**: This channel provides the latest stable version of KubeArmor along with its patches.

          The KubeArmor operator is responsible for deploying various components of the KubeArmor application, including the KubeArmor
          Daemonset, KubeArmor Controller, KubeArmor Relay Server, and KubeArmor Snitch. The role of KubeArmor Snitch is to gather information
          about the enabled Linux security modules on the host, which is utilized by KubeArmor to enforce runtime security policies.
          Additionally, KubeArmor Snitch collects details about the container runtime socket and storage volume cluster configurations.
          This deployment is executed as a short-term job and necessitates root user privileges. KubeArmor Snitch is certified by RedHat
          as a container with root privileges.

          KubeArmor Operator comes with a custom resource:

          - **KubeArmorConfig** - User will provide the initial configurations of the KubeArmor using KubeArmorConfig CR.

          ### KubeArmorConfig Customization Options:

          | Parameter                                 | Description                                          | Default                                 |
          | ---                                       | ---                                                  | ---                                     |
          | `alertThrottling`                         | Enable throttling for the KubeArmor generated alerts | False                                   |
          | `defaultFilePosture`                      | Default file posture for KubeArmor                   | audit                                   |
          | `defaultNetworkPosture`                   | Default network posture for KubeArmor                | audit                                   |
          | `defaultCapabilitiesPosture`              | Default capabilities posture for KubeArmor           | audit                                   |
          | `defaultVisibility`                       | Default visibility for KubeArmor                     | process,file,network                    |
          | `enableStdOutLogs` |  show stdout logs for relay server | false |
          | `enableStdOutAlerts` |  show stdout alerts for relay server | false |
          | `enableStdOutMsgs` |  show stdout messages for relay server | false |
          | `kubearmorImage.image`                    | Default image for KubeArmor Daemonset                | kubearmor/kubearmor-ubi:stable          |
          | `kubearmorImage.imagePullPolicy`          | Default imagePullPolicy for KubeArmor Daemonset      | Always                                  |
          | `kubearmorInitImage.image`                | Default image for KubeArmor Init Container           | kubearmor/kubearmor-init:stable         |
          | `kubearmorInitImage.imagePullPolicy`      | Default imagePullPolicy for KubeArmor Init Container | Always                                  |
          | `kubearmorRelayImage.image`               | Default image for KubeArmor Relay Server             | kubearmor/kubearmor-relay-server:latest |
          | `kubearmorRelayImage.imagePullPolicy`     | Default imagePullPolicy for KubeArmor Relay Server   | Always                                  |
          | `kubearmorControllerImage.image`          | Default image for KubeArmor Controller               | kubearmor/kubearmor-controller:latest   |
          | `kubearmorControllerImage.imagePullPolicy`| Default imagePullPolicy for KubeArmor Controller     | Always                                  |
          | `maxAlertPerSec`                          | Maximum no of alerts per second to trigger throttling| 10                                      |
          | `recommendedPolicies`                     | Recommended Policies Spec                            | -                                       |
          | `recommendedPolicies.matchExpressions`    | Target Namespaces for recommended policies           | -                                       |
          | `recommendedPolicies.excludePolicy`       | Exclude policies from recommended policies           | -                                       |
          | `seccompEnabled`                          | Deploy KubeArmor Daemon with default seccomp profile | True                                    |
          | `throttleSec`                             | Thottling duration in seconds                        | 30                                      |

          Users are not required to specify the images in the KubearmorConfig custom resource (CR). By default, the operator will utilize the images published with the operator release.

          For detailed information on Default Posture refer [this guide](https://github.com/kubearmor/KubeArmor/blob/main/getting-started/default_posture.md)

          similarly for default visibility refer [this guide](https://github.com/kubearmor/KubeArmor/blob/main/getting-started/kubearmor_visibility.md)

          ## Support

          For Support please join KubeArmor Slack [Here](https://join.slack.com/t/kubearmor/shared_invite/zt-1ltmqdbc6-rSHw~LM6MesZZasmP2hAcA)
        displayName: KubeArmor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - security
        - alerting
        - runtime-security
        - cloud native
        - container security
        links:
        - name: Website
          url: https://kubearmor.io
        - name: Github
          url: https://github.com/kubearmor/KubeArmor.git
        - name: Slack
          url: https://join.slack.com/t/kubearmor/shared_invite/zt-1ltmqdbc6-rSHw~LM6MesZZasmP2hAcA
        maintainers:
        - email: ramakant@accuknox.com
          name: Ramakant Sharma
        maturity: stable
        provider:
          name: Accuknox
          url: https://accuknox.com
        relatedImages:
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:1153a4592746b05e019bde4d818d176ff9350c013f84d49829032540de882841
        - docker.io/kubearmor/kubearmor-operator@sha256:d6c6b16e636474568835ded97cecb5ed17bc6004501227f6b6d63522ed57b01c
        - docker.io/kubearmor/kubearmor-snitch@sha256:c3d0f45590f9a0ebb50254047f12a7468844bcae7961eccb11f5ef578832c0eb
        - docker.io/kubearmor/kubearmor-ubi@sha256:629e295a2e64953bf95f3f3e842c42ff665ed6e4db32db8e43dcde4befdd3fc2
        - docker.io/kubearmor/kubearmor-init@sha256:e30ee508aa95d273f74a543e1b412b98c1714d04f093b59991c0ec2abaa450be
        - docker.io/kubearmor/kubearmor-relay-server@sha256:0a1a20ef02bd05eaa4a900948cadee6ae7650b0dde70abe1935608eaa5f931b7
        - docker.io/kubearmor/kubearmor-controller@sha256:2d369138cdbb31468c41e9ce70eda986a16bbeb9dd1192cd8a011c9df8a83857
        version: 1.5.2
      entries:
      - name: kubearmor-operator.v1.5.2
        version: 1.5.2
      name: stable
    defaultChannel: stable
    packageName: kubearmor-operator-certified
    provider:
      name: Accuknox
      url: https://accuknox.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Tigera
      provider-url: https://tigera.io/
    name: tigera-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tigera-operator.v1.20.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:bd93e66be5cfbc123fe4e8c3f6e8ddf7654217bf26136f06dcf23f28b53063d2
          createdAt: "2021-10-05T20:02:14.564507Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.20.4
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:bd93e66be5cfbc123fe4e8c3f6e8ddf7654217bf26136f06dcf23f28b53063d2
        version: 1.20.4
      entries:
      - name: tigera-operator.v1.20.4
        version: 1.20.4
      - name: tigera-operator.v1.20.3
        version: 1.20.3
      - name: tigera-operator.v1.20.2
        version: 1.20.2
      - name: tigera-operator.v1.20.1
        version: 1.20.1
      name: release-v1.20
    - currentCSV: tigera-operator.v1.23.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:f258275fe3d48c2ae16346a57f628523a6f2fb7b854b6d43b169056e158537f1
          createdAt: "2022-01-12T22:55:57.409244Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.23.4
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:f258275fe3d48c2ae16346a57f628523a6f2fb7b854b6d43b169056e158537f1
        version: 1.23.4
      entries:
      - name: tigera-operator.v1.23.4
        version: 1.23.4
      - name: tigera-operator.v1.23.3
        version: 1.23.3
      - name: tigera-operator.v1.23.2
        version: 1.23.2
      - name: tigera-operator.v1.23.1
        version: 1.23.1
      - name: tigera-operator.v1.23.0
        version: 1.23.0
      name: release-v1.23
    - currentCSV: tigera-operator.v1.25.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:e862708c700df55f99a9c9bb3c453e31f22aef1993d6079d80433eb8da831c94
          createdAt: "2022-02-17T05:39:13.720885Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.25.1
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:e862708c700df55f99a9c9bb3c453e31f22aef1993d6079d80433eb8da831c94
        version: 1.25.1
      entries:
      - name: tigera-operator.v1.25.1
        version: 1.25.1
      name: release-v1.25
    - currentCSV: tigera-operator.v1.28.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:9e8e47045fc67a2b00f9f2fd575fa526d7ddcac50af0b8b6d583b7518ccce120
          createdAt: "2023-05-17T16:13:52.148441156Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.28.13
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:9e8e47045fc67a2b00f9f2fd575fa526d7ddcac50af0b8b6d583b7518ccce120
        version: 1.28.13
      entries:
      - name: tigera-operator.v1.28.13
        version: 1.28.13
      name: release-v1.28
    - currentCSV: tigera-operator.v1.29.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:2a5acd29d8dca6adb318af04dbb15d93d899ed2b28b2ee9aaac09147395efb58
          createdAt: "2023-03-31T20:49:52.518230079Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.29.3
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:2a5acd29d8dca6adb318af04dbb15d93d899ed2b28b2ee9aaac09147395efb58
        version: 1.29.3
      entries:
      - name: tigera-operator.v1.29.3
        version: 1.29.3
      name: release-v1.29
    - currentCSV: tigera-operator.v1.30.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:431f037ff18b5c867d01312e42671effc55602421aeed25dd3f6109f70596b4a
          createdAt: "2023-11-17T00:52:31.263564294Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.30.9
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: APIServer
            name: apiservers.operator.tigera.io
            version: v1
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:431f037ff18b5c867d01312e42671effc55602421aeed25dd3f6109f70596b4a
        version: 1.30.9
      entries:
      - name: tigera-operator.v1.30.9
        version: 1.30.9
      name: release-v1.30
    - currentCSV: tigera-operator.v1.32.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:715ac9a30f8a9579e44258af20de354715429e11836b493918e9e1a696e9b028
          createdAt: "2023-12-15T23:45:03.964812504Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.32.3
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:715ac9a30f8a9579e44258af20de354715429e11836b493918e9e1a696e9b028
        version: 1.32.3
      entries:
      - name: tigera-operator.v1.32.3
        version: 1.32.3
      name: release-v1.32
    - currentCSV: tigera-operator.v1.36.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:3341fa9475c0325b86228c8726389f9bae9fd6c430c66fe5cd5dc39d7bb6ad4b
          createdAt: "2025-02-06T02:39:01.007218336Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.36.5
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:3341fa9475c0325b86228c8726389f9bae9fd6c430c66fe5cd5dc39d7bb6ad4b
        version: 1.36.5
      entries:
      - name: tigera-operator.v1.36.5
        version: 1.36.5
      name: release-v1.36
    - currentCSV: tigera-operator.v1.38.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "calicoNetwork": {
                    "linuxDataplane": "BPF"
                  },
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing, Monitoring, Networking, Security
          containerImage: quay.io/tigera/operator@sha256:e0a34b265aebce1a2db906d8dad99190706e8bf3910cae626b9c2eb6bbb21775
          createdAt: "2025-05-06T13:27:07.802897329Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          features.operators.openshift.io/valid-subscription: No subscription required
          olm.skipRange: <1.38.0
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift.
        displayName: Tigera Operator v1.38
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tigera-operator
        - networking
        - security
        - monitoring
        links:
        - name: Tigera
          url: https://tigera.io/
        - name: Calico Introduction
          url: https://docs.tigera.io/calico/latest/about/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.tigera.io/calico/latest/getting-started/kubernetes/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
          url: https://tigera.io/
        relatedImages:
        - quay.io/tigera/operator@sha256:e0a34b265aebce1a2db906d8dad99190706e8bf3910cae626b9c2eb6bbb21775
        version: 1.38.0
      entries:
      - name: tigera-operator.v1.38.0
        version: 1.38.0
      name: release-v1.38
    defaultChannel: release-v1.38
    packageName: tigera-operator
    provider:
      name: Tigera
      url: https://tigera.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solution-Soft
      provider-url: www.solution-soft.com
    name: timemachine-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: timemachine-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tm.solution-soft.com/v1alpha1",
                "kind": "TimeMachine",
                "metadata": {
                  "name": "timemachine-sample"
                },
                "spec": {
                  "cluster_provider": "ibm",
                  "aws": {
                    "shared_volume_name": ""
                  },
                  "ibm": {
                    "persistentvolume": {
                      "storageclass": "ibmc-file-silver-gid"
                    }
                  },
                  "tm_lichost": "52.152.148.40",
                  "tm_licpass": "qwaszx",
                  "tm_licport": "57777",
                  "imagePullSecrets": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
          createdAt: "2024-02-05T09:46:17Z"
          description: Time Machine Operator enables time travel of OpenShift pods,
            without the need to change the system clock on any of the cluster nodes
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/timemachine-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/timemachine-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=0.0.4 <1.0.17'
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: SolutionSoft Systems, Inc. - support@solution-soft.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Time Machine
            displayName: Time Machine
            kind: TimeMachine
            name: timemachines.tm.solution-soft.com
            version: v1alpha1
        description: |-
          Time Machine operator enables time travel of OpenShift pods/deployments.

          To time travel a pod where your application is, you just need to install Time Machine operator and create a Time Machine custom resource in the same namespace, without the need of changing the container images of the application you’re using, or adding containers to the pod with your application.

          Time travelling is done on the namespace level where all the (configured) target pods/deployments see the same virtual time.
        displayName: Time Machine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - time shift
        - time travel
        - virtual clocks
        - time machine
        - virtual time
        - solution-soft
        links:
        - name: Timemachine Operator
          url: https://catalog.redhat.com/software/operators/detail/5e98727a6c5dcb34dfbb19e3
        maintainers:
        - email: dtrajkovic@solution-soft.com
          name: Dragan Trajkovic
        maturity: alpha
        provider:
          name: Solution-Soft
          url: www.solution-soft.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0c304c11986790f446604319dfc29db494b15f5f9fa13f2c436688ec29c92dbe
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
        - registry.connect.redhat.com/rhel-timemachine-container/virtual-clock-testing-tool@sha256:a81a823328b0ee795c40668c89a08bcf545c5192d9eed0146071e3721a0a8372
        version: 1.0.17
      entries:
      - name: timemachine-operator.v1.0.17
        version: 1.0.17
      - name: timemachine-operator.v1.0.16
        version: 1.0.16
      name: alpha
    defaultChannel: alpha
    packageName: timemachine-operator-rhmp
    provider:
      name: Solution-Soft
      url: www.solution-soft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: ironic-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ironic-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "Ironic",
                "metadata": {
                  "name": "ironic"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseInstance": "openstack",
                  "ironicAPI": {},
                  "ironicConductors": [
                    {
                      "storageRequest": "10G"
                    }
                  ],
                  "ironicInspector": {},
                  "ironicNeutronAgent": {},
                  "secret": "osp-secret",
                  "serviceUser": "ironic",
                  "storageClass": "local-storage"
                }
              },
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "IronicAPI",
                "metadata": {
                  "name": "ironicapi-sample"
                },
                "spec": {
                  "databaseHostname": "openstack.ironic-kuttl-tests.svc"
                }
              },
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "IronicConductor",
                "metadata": {
                  "name": "ironicconductor-sample"
                },
                "spec": {
                  "databaseHostname": "openstack.ironic-kuttl-tests.svc",
                  "storageRequest": "10G"
                }
              },
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "IronicInspector",
                "metadata": {
                  "name": "ironicinspector-sample"
                },
                "spec": {
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "IronicNeutronAgent",
                "metadata": {
                  "name": "ironic-neutron-agent-sample",
                  "namespace": "openstack"
                },
                "spec": {
                  "secret": "osp-secret"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:02Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IronicAPI is the Schema for the ironicapis API
            displayName: Ironic API
            kind: IronicAPI
            name: ironicapis.ironic.openstack.org
            version: v1beta1
          - description: IronicConductor is the Schema for the ironicconductors Conductor
            displayName: Ironic Conductor
            kind: IronicConductor
            name: ironicconductors.ironic.openstack.org
            version: v1beta1
          - description: IronicInspector is the Schema for the ironicinspectors
            displayName: Ironic Inspector
            kind: IronicInspector
            name: ironicinspectors.ironic.openstack.org
            version: v1beta1
          - description: IronicNeutronAgent is the Schema for the ML2 networking-baremetal's
              agent
            displayName: Ironic Neutron Agent
            kind: IronicNeutronAgent
            name: ironicneutronagents.ironic.openstack.org
            version: v1beta1
          - description: Ironic is the Schema for the ironics API
            displayName: Ironic
            kind: Ironic
            name: ironics.ironic.openstack.org
            version: v1beta1
        description: Ironic Operator
        displayName: Ironic Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Identity
        - Ironic
        links:
        - name: Ironic Operator
          url: https://github.com/openstack-k8s-operators/ironic-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/ironic-rhel9-operator@sha256:184a1783a3b2a38bdb23f9811243414a2b21c16d78ca79ed83e2bcaa720d199b
        - registry.redhat.io/rhoso/openstack-ironic-api-rhel9@sha256:86c5330a3065a11f251e8c333bd855d37294c4e9d36cc68208fd16b3f95ef15a
        - registry.redhat.io/rhoso/openstack-ironic-conductor-rhel9@sha256:4950d74335eb48ed04f1a4f602730f1dfd42f139233a4edfebef0d8403c00b2d
        - registry.redhat.io/rhoso/openstack-ironic-inspector-rhel9@sha256:34ff13efc254d4390fbef0a5c088c569241d44d063ec16aed42e36d8179e0c85
        - registry.redhat.io/rhoso/openstack-ironic-pxe-rhel9@sha256:d49af96baa8d90dad3b244ed8c006a9945c41d05e5fa197cc35cc662b49cdd7c
        - registry.redhat.io/rhoso/openstack-ironic-neutron-agent-rhel9@sha256:132627eaf483353c5daefbbc84c6b06f51c50d0813a229ddc558e9df18bb5578
        - registry.redhat.io/rhoso/ironic-python-agent-rhel9@sha256:f41166809d2e7d76446c01f81d325a320df21e64f4b776d0f83cadcb9387eb7f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: ironic-operator.v1.0.6
        version: 1.0.6
      - name: ironic-operator.v1.0.5
        version: 1.0.5
      - name: ironic-operator.v1.0.4
        version: 1.0.4
      - name: ironic-operator.v1.0.3
        version: 1.0.3
      - name: ironic-operator.v1.0.2
        version: 1.0.2
      - name: ironic-operator.v1.0.1
        version: 1.0.1
      - name: ironic-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: ironic-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kove
      provider-url: https://kove.com
    name: kove-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kove-operator.v3.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "containers.kove.com/v1",
                "kind": "KoveMemoryConfig",
                "metadata": {
                  "labels": {
                    "KoveMemoryConfig": ""
                  },
                  "name": "example-memory-config"
                },
                "spec": {
                  "container": "",
                  "namespace": "",
                  "size": ""
                }
              },
              {
                "apiVersion": "containers.kove.com/v1",
                "kind": "KoveSystem",
                "metadata": {
                  "labels": {
                    "KoveSystem": ""
                  },
                  "name": "kove-system"
                },
                "spec": {
                  "autogrowInitialIncrement": "1 GiB",
                  "autogrowMaxIncrement": "1 TiB",
                  "l4MaxMemoryPercent": 75,
                  "managementConsoleAddress": "",
                  "memoryPlusLocalMemoryMaxSize": "0 MiB"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: quay.io/kove/kove-operator@sha256:0fb5fa2e6702d5e5bf692c25f190f65d306fc72cdceebf0814ff64adcb5f6994
          createdAt: "2025-09-25T16:31:36Z"
          description: Configure containers with Kove® pooled memory
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: kove-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Kove
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configure containers with Kove® pooled memory
            displayName: Kove Memory Configuration
            kind: KoveMemoryConfig
            name: kovememoryconfigs.containers.kove.com
            version: v1
          - description: Configure cluster with Kove:SDM™
            displayName: Kove System Configuration
            kind: KoveSystem
            name: kovesystems.containers.kove.com
            version: v1
        description: |
          ## About the Operator
          The Kove:SDM™ operator integrates Kove's software-defined memory
          solution with Kubernetes and OpenShift clusters. The operator enables
          containers to transparently provision any amount of memory, enabling total
          container memory capacity to exceed physical hardware limitations.

          Use the Kove:SDM™ operator to:

          * Create custom resources to automatically allocate memory to specific
            containers
          * Increase container flexibility with dynamic memory capacity that
            auto-grows as needed
          * Scale up container density by removing memory as an orchestration
            bottleneck

          ## Using the Operator
          Using the operator is as simple as creating a custom resource which
          specifies an initial Kove memory allocation size, a namespace, and a
          container name. Containers in the specified namespace with a matching
          container name will automatically be provided with a Kove memory allocation
          which will auto-grow as needed.

          ## About Kove
          Kove makes memory a provisionable resource, just like storage, CPUs, and
          networking. Kove:SDM™ provides a transparent, linearly scalable
          memory resource accessible to any application without code changes, from
          bare-metal apps to virtual machines and containers.
        displayName: Kove:SDM™ Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kove
        - Pooled Memory
        - Memory Pool
        - Memory Target
        - External Memory
        - Memory
        links:
        - name: Kove
          url: https://kove.com
        maintainers:
        - email: contact-kove@kove.com
          name: Kove
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Kove
          url: https://kove.com
        relatedImages:
        - registry.access.redhat.com/ubi8/pause@sha256:eacbaaadc932633330e799eda24e9b731b5a4ba7ed823ee70791d21029724a9a
        - quay.io/kove/kove-agent@sha256:9d026b925e1d414a531530a30ed8b5f10b421ee12c7c13949ea0ee9907d36b75
        - quay.io/kove/kove-client-drivers@sha256:1c1786e1bbf16609218154126b4340b2f7c4b71cbd9045589402f75c032ea38d
        - quay.io/kove/kove-compatibility-detection@sha256:0303830a1f785b5e4ce6824db7c50f8feee2678fd5e87ded983686ab3f60b232
        - quay.io/kove/kove-operator@sha256:0fb5fa2e6702d5e5bf692c25f190f65d306fc72cdceebf0814ff64adcb5f6994
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fa0acefda43e309e7f9f31f5247c933d0b868ff1573ece2f38e0549d7e2ec942
        - quay.io/kove/kove-container-hook@sha256:66ee33d59ced63406d9b40f3a3226fab7f52071f8bed9aa1ba7d859837038528
        - quay.io/kove/kove-xmem@sha256:41bfc536a278153b0a4460d2836b56762b56cf50b94d17e123a63be066c28fe7
        - quay.io/kove/kove-client-drivers@sha256:716ccda2d8f082a954f2042fd189fec257bd2dcb2c50e497bad42b9a9e5f0e3d
        version: 3.0.2
      entries:
      - name: kove-operator.v3.0.2
        version: 3.0.2
      - name: kove-operator.v3.0.1
        version: 3.0.1
      - name: kove-operator.v3.0.0
        version: 3.0.0
      name: stable
    defaultChannel: stable
    packageName: kove-operator
    provider:
      name: Kove
      url: https://kove.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Fujitsu
      provider-url: https://www.postgresql.fastware.com/
    name: fujitsu-enterprise-postgres-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fujitsu-enterprise-postgres-operator.v7.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPAction",
                "metadata": {
                  "name": "new-fep-action"
                },
                "spec": {
                  "fepAction": {
                    "args": [
                      "new-fep-sts-0",
                      "new-fep-sts-1"
                    ],
                    "type": "reload"
                  },
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "targetClusterName": "new-fep"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPExporter",
                "metadata": {
                  "name": "new-fep-exporter"
                },
                "spec": {
                  "fepExporter": {
                    "exporterLogLevel": "error",
                    "fepClusterList": [
                      "new-fep1"
                    ],
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "userCustomQueries": "usr_example:\n  query: \"SELECT EXTRACT(EPOCH FROM (now() - pg_last_xact_replay_timestamp())) as lag\"\n  master: true\n  metrics:\n    - lag:\n        usage: \"GAUGE\"\n        description: \"Replication lag behind master in seconds\""
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPLogging",
                "metadata": {
                  "name": "new-fep-logging"
                },
                "spec": {
                  "fepLogging": {
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPPgpool2",
                "metadata": {
                  "name": "new-fep-pgpool2"
                },
                "spec": {
                  "clientAuthMethod": "scram",
                  "count": 2,
                  "customhba": "local   all         all                               trust\nhost    all         all         127.0.0.1/32          trust\nhost    all         all         ::1/128               trust\n",
                  "customlogsize": "128Mi",
                  "customparams": "listen_addresses = '*'\npcp_listen_addresses = '*'\nnum_init_children = 32\nreserved_connections = 0\nenable_pool_hba = off\nallow_clear_text_frontend_auth = off\nauthentication_timeout = 80\nbackend_weight0 = 1\nbackend_weight1 = 1\nbackend_flag0 = 'ALWAYS_PRIMARY'\nbackend_flag1 = 'DISALLOW_TO_FAILOVER'\nconnection_cache = on\nmax_pool = 4\nlisten_backlog_multiplier = 2\nserialize_accept = off\nchild_life_time = 300\nclient_idle_limit = 0\nchild_max_connections = 0\nconnection_life_time = 0\nreset_query_list = 'ABORT; DISCARD ALL'\nclient_min_messages = info\nlog_min_messages = debug1\nlog_statement = on\nlog_per_node_statement = on\nlog_client_messages = on\nlog_hostname = on\nlog_connections = on\nlog_line_prefix = '%t: pid %p: '\nload_balance_mode = on\nignore_leading_white_space = on\nwhite_function_list = ''\nblack_function_list = 'currval,lastval,nextval,setval'\nblack_query_pattern_list = ''\ndatabase_redirect_preference_list = ''\napp_name_redirect_preference_list = ''\nallow_sql_comments = off\ndisable_load_balance_on_write = 'transaction'\nstatement_level_load_balance = on\nsr_check_period = 0\nsr_check_user = 'postgres'\ndelay_threshold = 0\nlog_standby_delay = 'none'\nssl = on\nssl_ciphers = 'HIGH:MEDIUM:+3DES:!aNULL'\nssl_prefer_server_ciphers = off\nssl_ecdh_curve = 'prime256v1'\nssl_dh_params_file = ''\nrelcache_expire = 0\nrelcache_size = 256\ncheck_temp_table = catalog\ncheck_unlogged_table = on\nenable_shared_relcache = off\nrelcache_query_target = primary\nwd_port0 = 9000\nfailover_on_backend_error = off\n",
                  "custompcp": "none",
                  "customsslcacert": "none",
                  "customsslcert": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                  "customsslkey": "none",
                  "fepclustername": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "limits": {
                    "cpu": "400m",
                    "memory": "512Mi"
                  },
                  "requests": {
                    "cpu": "200m",
                    "memory": "256Mi"
                  },
                  "serviceport": 9999,
                  "statusport": 9898
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPRestore",
                "metadata": {
                  "name": "new-fep-restore"
                },
                "spec": {
                  "fromFEPcluster": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "mcSpec": {
                    "limits": {
                      "cpu": "200m",
                      "memory": "300Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "restoretype": "latest",
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "toFEPcluster": "new-fep-2"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v2",
                "kind": "FEPCluster",
                "metadata": {
                  "name": "new-fep"
                },
                "spec": {
                  "fep": {
                    "customAnnotations": {
                      "allDeployments": {}
                    },
                    "forceSsl": true,
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "instances": 1,
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "podAntiAffinity": false,
                    "podDisruptionBudget": false,
                    "servicePort": 27500,
                    "syncMode": "off",
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "usePodName": true
                  },
                  "fepChildCrVal": {
                    "backup": {
                      "image": {
                        "pullPolicy": "IfNotPresent"
                      },
                      "mcSpec": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      },
                      "pgbackrestParams": "# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[global]\nrepo1-retention-full=7\nrepo1-retention-full-type=time\nlog-path=/database/log/backup\n",
                      "postScript": " ",
                      "preScript": " ",
                      "schedule": {
                        "num": 2
                      },
                      "schedule1": {
                        "schedule": "15 0 * * 0",
                        "type": "full"
                      },
                      "schedule2": {
                        "schedule": "15 0 * * 1-6",
                        "type": "incr"
                      },
                      "schedule3": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule4": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule5": {
                        "schedule": " ",
                        "type": " "
                      }
                    },
                    "customPgAudit": "# define pg audit custom params here to override defaults.\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[output]\nlogger = 'auditlog'\nlog_directory = '/database/log/audit'\nlog_truncate_on_rotation = on\nlog_filename = 'pgaudit-%a.log'\nlog_rotation_age = 1d\nlog_rotation_size = 0\n[rule]\n",
                    "customPgHba": "# define pg_hba custom rules here to be merged with default rules.\n# TYPE     DATABASE        USER        ADDRESS        METHOD\n",
                    "customPgParams": "# define custom postgresql.conf parameters below to override defaults.\n# Current values are as per default FEP deployment\n# If you add a library to shared_preload_libraries, add it after the default value.\n# fsep_operator_security is available in FEP15 and later container images.\nshared_preload_libraries='pgx_datamasking,pg_prewarm,pg_stat_statements,fsep_operator_security'\nsession_preload_libraries='pg_prewarm'\nmax_prepared_transactions = 100\nmax_worker_processes = 30\nmax_connections = 100\nwork_mem = 1MB\nmaintenance_work_mem = 12MB\nshared_buffers = 128MB\neffective_cache_size = 384MB\ncheckpoint_completion_target = 0.8\n# tcp parameters\ntcp_keepalives_idle = 30\ntcp_keepalives_interval = 10\ntcp_keepalives_count = 3\n# logging parameters in default fep installation\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\nlog_directory = '/database/log'\nlog_filename = 'logfile-%a.log'\nlog_file_mode = 0600\nlog_truncate_on_rotation = on\nlog_rotation_age = 1d\nlog_rotation_size = 0\nlog_checkpoints = on\nlog_line_prefix = '%e %t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h'\nlog_lock_waits = on\nlog_autovacuum_min_duration = 60s\nlogging_collector = on\npgaudit.config_file='/opt/app-root/src/pgaudit-cfg/pgaudit.conf'\nlog_replication_commands = on\nlog_min_messages = WARNING\nlog_destination = csvlog\n# wal_archive parameters in default fep installation\narchive_mode = on\narchive_command = 'pgbackrest --stanza=backupstanza --config=/database/userdata/pgbackrest.conf archive-push %p'\nwal_level = replica\nmax_wal_senders = 12\nwal_keep_segments = 64\ntrack_activities = on\ntrack_counts = on\npassword_encryption = 'scram-sha-256'\n",
                    "storage": {
                      "archivewalVol": {
                        "size": "1Gi"
                      },
                      "backupVol": {
                        "size": "2Gi"
                      },
                      "dataVol": {
                        "size": "2Gi"
                      },
                      "logVol": {
                        "size": "1Gi"
                      },
                      "tablespaceVol": {
                        "size": "512Mi"
                      },
                      "walVol": {
                        "size": "1200Mi"
                      }
                    },
                    "sysUsers": {
                      "pgRewindPassword": "rewind_password",
                      "pgRewindUser": "rewind_user",
                      "pgdb": "mydb",
                      "pgpassword": "mydbpassword",
                      "pgreplpassword": "repluserpwd",
                      "pgrepluser": "repluser",
                      "pguser": "mydbuser",
                      "tdepassphrase": "tde-passphrase"
                    },
                    "systemCertificates": {
                      "cacrt": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                      "crt": "-----BEGIN CERTIFICATE-----\nMIIDUTCCAjmgAwIBAgIRAMocW3qMoHrD6qRvMPppMkMwDQYJKoZIhvcNAQELBQAw\nNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9yIEt1\nYmVybmV0ZXMwHhcNMjEwMjA2MDQzMjM2WhcNMjYwMjA1MDQzMjM2WjA/MRAwDgYD\nVQQKEwdGdWppdHN1MSswKQYDVQQDEyJGVUpJVFNVIEVudGVycHJpc2UgUG9zdGdy\nZXMgU2VydmVyMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA4AI33yvH\nZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I2e4SceTKi6O3C/I1XuvWlpng\n5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4ANB5JWWqDOjrRT3o7nRPGXfila\nbP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYpmjdbfxabTz07ig0+6/cwKoRR\nxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTFYvmAH7gcdssSFBt8NPlUATHE\nsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6Wdgmu5H2pDml8CDNLDv98Aj7i\n+I5SRKKcVPlnuQIDAQABo1AwTjAdBgNVHSUEFjAUBggrBgEFBQcDAQYIKwYBBQUH\nAwIwDAYDVR0TAQH/BAIwADAfBgNVHSMEGDAWgBQcwrrUO0u+FhIUuVdrDRCQRsi6\nZjANBgkqhkiG9w0BAQsFAAOCAQEAm5dxBoI9pScOCvRAchg4CprdRDSJb9K6yB3O\nnCAxnM47iHeXnY3WlnI388kHu8DU7O4ba1tJbGs3KY9KzioPk43pU12jWkO1onoF\n+mTDjx/Ef1cYWA9r5q/LtgTa6Q2sxV4O2x67QW82aAnaxO34dV5zWCPIvAoovZBV\nHRT+BgCg3r2vD1RGKK2nl1aYJtWhO1SZubam+VttdZ/vbM9oOJctxmImsEtBXjkY\nKteePdQtLL5o03JhyXWyRshCq+HMmKf2KgyY8gvydGcP4eLQdBWcW40LcnVq6UjT\n0kJycJEKngMVademq1ZWHGaiYB7hyT6GhgIcHUJ2cKrPgbEh1Q==\n-----END CERTIFICATE-----",
                      "key": "-----BEGIN RSA PRIVATE KEY-----\nMIIEowIBAAKCAQEA4AI33yvHZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I\n2e4SceTKi6O3C/I1XuvWlpng5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4AN\nB5JWWqDOjrRT3o7nRPGXfilabP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYp\nmjdbfxabTz07ig0+6/cwKoRRxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTF\nYvmAH7gcdssSFBt8NPlUATHEsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6W\ndgmu5H2pDml8CDNLDv98Aj7i+I5SRKKcVPlnuQIDAQABAoIBAFPQYKlOzw/+BA0b\nyMIUpdctIMb/54CR/xR0mVw1DbSjigNVPjHUQvB8Y1B2FAITQObgJO06bAv0QdWN\nRb0/v/yYiNJDFjaLjaIAHlO/2+oWrXbFaZqgpVDJhB+e1xaZr2x7XGxm+p925k30\nl6pvIRY+I8JRKvZiV1VZHwL/R3JOtPr++xMZtLVjVOI+f+ySqJ+TZHuAjm49EKxj\ncEmmJ28b7QcziXsvKy00f+zbqLIBKXQdZAFU5eEr1BsDRXdRW+Kf0XIvftuy4BJZ\nvoKT+VGhEvF/qysswL4+6IAO6tpuYnnM0Y2d3sOGoWPkTcQK0MekYKzL/WmtCjNs\n9hodJtECgYEA5EWyhEOf4uOKe5TDp697UCUvXLoOR58FDe/S8XNvScn29jjOkqIg\nOMoqo9xAkJTNTzqn5UUdt1x/pgM2NxlPLFijrc0zQlX3SoOO2ryDd9WNi7YKtN16\nKJqa536WeZu2OEbuAZ+S3GALVy1RPeTNPnUOmKnF06DjDUGzLNCZy10CgYEA+zfw\n952DWuz1U0Z4wvAEqqcgUKXPKrkTXV/iUnjkDkrLYVr0ZofDNTXrdHl+UedFmaOC\ncieZn6DNhcdz5tKtyysGMH3g/qs9PfoGUngvcXsy0Egk04l3x1jc8TTCLqXZXYaQ\nHMsx51n+R58oncPtzYSUOr9qQ6PbC2CstTbFJA0CgYEAjGEsUliAB/jknfEzjXjG\nPdhQUxb8VyE864Az2lah9t/kJzFyIAziAeqZ5GE7t247AGFTBRTHHI8e1Qoemi3P\nWbc9GVIbFs1lIYbcIDpUIyrKPEP8O5QEXtoNLxXTFgAjRGKiVY87spjCAJ+W2ZhO\ne/1it5GYXfgQCYQA2yuBmOUCgYANRkR2YR1axaCk+NlSu6oTdmdPu6M5x7PNQE7O\nOtMaKjua9lppvIzFGAdMDUtueoEEAE7ZR1xnwfB6PDLUpJdIYAqgr1YfPt8qkjaZ\nTv56yZ7CwL0pbF8m6nwqRrZoDp1wwraEvvvxFKFKGY/k3kCHlpTakdjEoDjn3gDi\nRnWeVQKBgCEneMSzucei5LRppRtRaJw/Btll8qlPMlX3W7dxQ3cLwpmLOn0m51Fp\nPIZ44zYK8R6fu4+/sSrlfaIg86Ugeufp6YNxyNROKxUGza5vDIu5OftwWtBeg+UK\nZ8lLWNdX6pp7WMujmF3H1DrkBbauYMUKZ4UxUYtelgHERMePIxwb\n-----END RSA PRIVATE KEY-----"
                    }
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "false"
          containerImage: quay.io/fujitsu/fujitsu-enterprise-postgres-operator@sha256:b7dd55c0bfda84e398cbfbb87345fcec218cb866e286e18a111330321ab6ad54
          createdAt: "2025-09-10T11:50:37Z"
          description: OpenShift Operator for Fujitsu Enterprise Postgres
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.1.0 <7.3.0'
          operators.openshift.io/valid-subscription: '["Fujitsu Enterprise Postgres
            Advanced Edition Annual Subscription License per Core 17 Operator for
            Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/internal-objects: '["fepautoscales.fep.fujitsu.io","fepconfigs.fep.fujitsu.io","fepusers.fep.fujitsu.io","fepcerts.fep.fujitsu.io","fepvolumes.fep.fujitsu.io","fepbackups.fep.fujitsu.io"]'
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: fj-dbaas-tec@dl.jp.fujitsu.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: FEPAction
            name: fepactions.fep.fujitsu.io
            version: v1
          - kind: FEPAutoscale
            name: fepautoscales.fep.fujitsu.io
            version: v1
          - kind: FEPBackup
            name: fepbackups.fep.fujitsu.io
            version: v1
          - kind: FEPCert
            name: fepcerts.fep.fujitsu.io
            version: v1
          - kind: FEPCluster
            name: fepclusters.fep.fujitsu.io
            version: v2
          - kind: FEPConfig
            name: fepconfigs.fep.fujitsu.io
            version: v1
          - kind: FEPExporter
            name: fepexporters.fep.fujitsu.io
            version: v1
          - kind: FEPLogging
            name: feploggings.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2Cert
            name: feppgpool2certs.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2
            name: feppgpool2s.fep.fujitsu.io
            version: v1
          - kind: FEPRestore
            name: feprestores.fep.fujitsu.io
            version: v1
          - kind: FEPUpgrade
            name: fepupgrades.fep.fujitsu.io
            version: v1
          - kind: FEPUser
            name: fepusers.fep.fujitsu.io
            version: v1
          - kind: FEPVolume
            name: fepvolumes.fep.fujitsu.io
            version: v1
        description: "Fujitsu Enterprise Postgres 17 delivers an enterprise-grade
          PostgreSQL on OpenShift Container Platform.\n\nThis solution provides the
          flexibility of a hybrid cloud solution while delivering an enhanced distribution
          of PostgreSQL to support enterprise-level workloads and provide improved
          deployment and management, availability, performance, data governance and
          security.  \n\nAvailable as a multi-architecture container built for both
          amd64, s390x and ppc64le.\n\nThe download and Use of the Product is strictly
          subject to the terms of the End User License Agreement with Fujitsu Limited
          found at https://www.fast.fujitsu.com/fujitsu-enterprise-postgres-license-agreements.
          Where the Product that has been embedded as a whole or part into a third
          party program, only Authorised Customers may download and use the Product.\n"
        displayName: Fujitsu Enterprise Postgres Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - Fujitsu
        - fep
        links:
        - name: Fujitsu Enterprise Postgres
          url: https://www.postgresql.fastware.com/
        maintainers:
        - email: fj-dbaas-tec@dl.jp.fujitsu.com
          name: Fujitsu
        maturity: stable
        provider:
          name: Fujitsu
          url: https://www.postgresql.fastware.com/
        relatedImages:
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-upgrade@sha256:bf3b069a71202ea78c7c0c7b36315c8bef8e6a1c11cd0668dcd888329e26ecfb
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-backup@sha256:dd8b84e200c9d0fb32cd567a33968e15e15efc115ff120f54fdd8862c7ea3111
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentbit@sha256:9622d2191525da39e5100e1882dd2cfa3eda740d8094f3fbb71ca8a5734d2843
        - quay.io/fujitsu/fujitsu-enterprise-postgres-utils@sha256:57a3fb792480a0d3a8f075c48dee2d02f1d79cf459761e3305cf62799575a885
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-backup@sha256:42252444e2dcfe5435532f318cce041990d2b71efa26007924717924d30a46ca
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-utils@sha256:edeec4d00fd8282aaa7ec6fe3ec4cd84070a6402611002a18addcac87fb8ee94
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-server@sha256:fa4e8849d75daf331ba7d488bcb52c083cd92bb5f288e4b62c980754e299f618
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-pgpool2@sha256:185af2e4fd1701aded1ff5e44d4336562f5a219c30ba535c2a3097b914c24efe
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-upgrade@sha256:04af997b4f548ffec711b63a06c35715b4e7863537c83697b3ee0c1d95a6aec3
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-pgpool2@sha256:8eaace2f9c8e5e844992fe5e008cd1061af33cc79b4d442c4a83a65610bcc502
        - quay.io/fujitsu/fujitsu-enterprise-postgres-exporter@sha256:1c0e22aeab65ff3848772424fc3ea6baa6b58b0d88db5413bda160a08435bcff
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-upgrade@sha256:8a0f7ca1f59ba162424a308e8a70c4717135e2847e61210753b71876c9b0a75f
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-restore@sha256:09be02274427e3022849fb635a4c9abdc34696838cae39f3afafc8009567db9d
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-backup@sha256:78a3ad639d0c63858b9db0a5de36ce11a9ea121f4535e9ca07ac7defb689bc0d
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-server@sha256:3e93a368a6198818e1b7c5525324c8130c5ed8d7b4cb499a47d4bec45ef0347b
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-pgpool2@sha256:c878935ffc796c405d811a6e8a9194aaaaa8652ea7d7b879c82919ce09794b35
        - quay.io/fujitsu/fujitsu-enterprise-postgres-cronjob@sha256:7582964eef14f89f2ca9046298357afd0132fb236f30f44f84cee14692227b0c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-restore@sha256:a81a0ce926572b75aa3151e2d2cc45a0ca411cc0158d69b89247ef70418f3740
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-server@sha256:1c5823afc7e77d46940214693c12c0ad6ffd6c7d9a156e26399ab55cd5ea1dfa
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-restore@sha256:17301c2191c0d75be82d12c5a35c0f74b8d65304bcfb6d4ab312a10ffa286eda
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-restore@sha256:a6e90dd77634033f2009942b635078cceb0f4c6b15b031624c5dce820713b14c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-server@sha256:5c65f8e76f70341c512100760de6792a880ad0d6eebc9e68154b5bbd1332a896
        - quay.io/fujitsu/fujitsu-enterprise-postgres-operator@sha256:b7dd55c0bfda84e398cbfbb87345fcec218cb866e286e18a111330321ab6ad54
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-utils@sha256:b7ae535f5e5fa5eb51c6f4cdb406c7b7a9dabe8573939b0431dada6f7d27b87e
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentd@sha256:3c5534b62695867c5973c9106fc46895ed73a4b9136ad83b2586bf7e71c427d0
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-pgpool2@sha256:38aac8108dd3590abcfd854a6ec2c025a68d2170a1ba68d359b45fea6c83cec1
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-backup@sha256:ac268e2af652bdabb8ee5d764a2b335dc73c48b9022eee8c9f3cd422227a4d98
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-upgrade@sha256:c0e44447349f02eb9f6105001136402f849d876a288d61997c8c05625facff45
        version: 7.3.0
      entries:
      - name: fujitsu-enterprise-postgres-operator.v7.3.0
        version: 7.3.0
      - name: fujitsu-enterprise-postgres-operator.v7.2.7
        version: 7.2.7
      name: stable
    defaultChannel: stable
    packageName: fujitsu-enterprise-postgres-operator
    provider:
      name: Fujitsu
      url: https://www.postgresql.fastware.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-block-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12
    - currentCSV: ibm-block-csi-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "k8s.gcr.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
        version: 1.12.0
      entries:
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.0
    - currentCSV: ibm-block-csi-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
        version: 1.12.1
      entries:
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.1
    - currentCSV: ibm-block-csi-operator.v1.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:af9215670675fe9c4fa2dac7a14d547c249f3aa0c0e72f1e5390d237e411735b
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:af9215670675fe9c4fa2dac7a14d547c249f3aa0c0e72f1e5390d237e411735b
        version: 1.12.2
      entries:
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.2
    - currentCSV: ibm-block-csi-operator.v1.12.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:6e7e13d259f4d743524ebcd546af5c73d1a3ba90edaf4b5edd2005a3f8adcdfe
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:6e7e13d259f4d743524ebcd546af5c73d1a3ba90edaf4b5edd2005a3f8adcdfe
        version: 1.12.3
      entries:
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.3
    - currentCSV: ibm-block-csi-operator.v1.12.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:42728e7fd1d2302511ab08f51335ab188eef125870fa27f322df83ebb8e1f447
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:42728e7fd1d2302511ab08f51335ab188eef125870fa27f322df83ebb8e1f447
        version: 1.12.4
      entries:
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.4
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.5
    defaultChannel: stable
    packageName: ibm-block-csi-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cli-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cli-manager-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "CliManager",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-cli-manager-operator"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed",
                  "operatorLogLevel": "Normal"
                }
              }
            ]
          capabilities: Basic Install
          categories: Openshift Optional
          createdAt: "2024-06-04"
          description: "The CLI Manager is a comprehensive tool designed to simplify
            the management of \nOpenShift CLI plugins within the OpenShift environment.
            \nModeled after the popular krew plugin manager, it offers seamless integration,
            \neasy installation, and efficient handling of a wide array of plugins,
            \nenhancing your OpenShift command-line experience.\n"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          repository: https://github.com/openshift/cli-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: CLI Manager
            kind: CliManager
            name: climanagers.operator.openshift.io
            version: v1
          - displayName: CLI Plugin
            kind: Plugin
            name: plugins.config.openshift.io
            version: v1alpha1
        description: "The CLI Manager is a comprehensive tool designed to simplify
          the management of \nOpenShift CLI plugins within the OpenShift environment.
          \nModeled after the popular krew plugin manager, it offers seamless integration,
          \neasy installation, and efficient handling of a wide array of plugins,
          \nenhancing your OpenShift command-line experience.\n"
        displayName: CLI Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cli-manager-operator
        - cli-manager
        links:
        - name: Source Code
          url: https://github.com/openshift/cli-manager
        - name: Source Code
          url: https://github.com/openshift/cli-manager-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/cli-manager-operator/cli-manager-rhel9-operator@sha256:31973d7f43d052ddfb9554dd623e37fb541a8c1f08ea981558eaf4bb5ab22d5a
        - registry.redhat.io/cli-manager-operator/cli-manager-rhel9@sha256:07853bc33a0ce783d8d45c08c161e454a3abe09a8a8dadeca70c3a199466b58b
        version: 0.1.0
      entries:
      - name: cli-manager-operator.v0.1.0
        version: 0.1.0
      name: tech-preview
    defaultChannel: tech-preview
    packageName: cli-manager
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cryostat-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cryostat-operator.v4.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.cryostat.io/v1beta2",
                "kind": "Cryostat",
                "metadata": {
                  "name": "cryostat-sample"
                },
                "spec": {
                  "enableCertManager": true,
                  "eventTemplates": [],
                  "networkPolicies": {},
                  "reportOptions": {
                    "replicas": 0
                  },
                  "trustedCertSecrets": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Developer Tools
          containerImage: registry.redhat.io/cryostat/cryostat-rhel9-operator@sha256:867a649b54ad144295c7aceda6b90631841aa31a62062df62d4980313723f359
          createdAt: "2025-09-30T15:10:08Z"
          description: JVM monitoring and profiling tool
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.0 <4.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "operator.cryostat.io/v1beta2",
              "kind": "Cryostat",
              "metadata": {
                "name": "cryostat-sample"
              },
              "spec": {
                "enableCertManager": true,
                "reportOptions": {
                  "replicas": 0
                }
              }
            }
          operators.openshift.io/valid-subscription: '["Red Hat build of OpenJDK"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/cryostatio/cryostat-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cryostat allows you to install Cryostat for a single namespace,
              or multiple namespaces. It contains configuration options for controlling
              the Deployment of the Cryostat application and its related components.
              A Cryostat instance must be created to instruct the operator to deploy
              the Cryostat application.
            displayName: Cryostat
            kind: Cryostat
            name: cryostats.operator.cryostat.io
            version: v1beta2
          - description: Cryostat allows you to install Cryostat for a single namespace.
              It contains configuration options for controlling the Deployment of
              the Cryostat application and its related components. A Cryostat instance
              must be created to instruct the operator to deploy the Cryostat application.
            displayName: Cryostat
            kind: Cryostat
            name: cryostats.operator.cryostat.io
            version: v1beta1
        description: |
          The Red Hat build of Cryostat provides a cloud-based solution for interacting with the JDK Flight Recorder already present in OpenJDK 11+ JVMs. With Cryostat, users can remotely start, stop, retrieve, and even analyze JFR event data, providing the capability to easily take advantage of Flight Recorder's extremely low runtime cost and overhead and the flexibility to monitor applications and analyze recording data without transferring data outside of the cluster the application runs within.
          ##Prerequisites
          Cryostat requires [cert-manager](https://cert-manager.io/) to run. If not already installed in your cluster, please [install](https://cert-manager.io/docs/installation/) it using your preferred method.
          When running on Kubernetes, the operator also requires [Ingress configurations](https://github.com/cryostatio/cryostat-operator/blob/v4.0.0/docs/config.md#network-options) for each of its services to make them available outside of the cluster. The user is responsible for providing the hostnames for each Ingress.
        displayName: Red Hat build of Cryostat
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flightrecorder
        - java
        - jdk
        - jfr
        - jmc
        - missioncontrol
        - monitoring
        - profiling
        - diagnostic
        links:
        - name: Upstream Project
          url: https://github.com/cryostatio/cryostat
        - name: Website
          url: https://cryostat.io/
        maintainers:
        - email: cryostat-development@googlegroups.com
          name: The Cryostat Authors
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi9/nginx-124@sha256:7c1d7172880a3649591a0dc1ceb670c133ef88eed3d8192d8a2e80e3878d04c3
        - registry.redhat.io/cryostat/cryostat-reports-rhel9@sha256:5535a539b185fa2781621741ccd7f10a6b9947adc6170a95e930f2d9d7a9f608
        - registry.redhat.io/insights-runtimes-tech-preview/runtimes-inventory-rhel8-operator@sha256:58e01d11c3d30abc5b0ee3795511bd04e39aea21e2853b4f4cc6d1c8a2bd7bb0
        - registry.redhat.io/cryostat/cryostat-rhel9@sha256:fadae8713daf82c329573191381f8f9e4ebf3fc852ac5f1a68ce0ea1da376469
        - registry.redhat.io/cryostat/jfr-datasource-rhel9@sha256:1e5f29aa096d783fd0a18b3ba31c391dc203761ba3f4af8e5aeeef1f7cc86cee
        - registry.redhat.io/cryostat/cryostat-db-rhel9@sha256:23d61204ecaf037cf17bf9a5cc4428461a60fecce077884d9458802fb071daab
        - registry.redhat.io/cryostat/cryostat-grafana-dashboard-rhel9@sha256:5715e897f9a8969aedd0066b2ba597264495b4c85e4ae8ae7d8871f93d13f4c1
        - registry.redhat.io/cryostat/cryostat-agent-init-rhel9@sha256:273d6c3d3194a4e9ec50c1d8e147cd9bc6d277d8d66878d885a18528b9ba7dc6
        - registry.redhat.io/cryostat/cryostat-rhel9-operator@sha256:867a649b54ad144295c7aceda6b90631841aa31a62062df62d4980313723f359
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        - registry.redhat.io/cryostat/cryostat-ose-oauth-proxy-rhel9@sha256:d38ce4dcac5b073b001d886f0494e0fc9b750f8bd68b855fc103767a1ed8b7df
        - registry.redhat.io/cryostat/cryostat-storage-rhel9@sha256:f642ca8a0ca0aa130462c0f423ce4c775cb296f8a40c9270e5369176e87c7ee6
        - registry.redhat.io/cryostat/cryostat-openshift-console-plugin-rhel9@sha256:ca748f96414ad648eb1e3953aeeeb828b187f3c5c30e014513adc3619a698683
        version: 4.0.3
      entries:
      - name: cryostat-operator.v4.0.3
        version: 4.0.3
      - name: cryostat-operator.v4.0.2-4
        version: 4.0.2+4
      - name: cryostat-operator.v4.0.2
        version: 4.0.2
      - name: cryostat-operator.v4.0.1
        version: 4.0.1
      - name: cryostat-operator.v4.0.1-5
        version: 4.0.1+5
      - name: cryostat-operator.v4.0.1-4
        version: 4.0.1+4
      - name: cryostat-operator.v4.0.1-3
        version: 4.0.1+3
      - name: cryostat-operator.v4.0.1-2
        version: 4.0.1+2
      - name: cryostat-operator.v4.0.0-11
        version: 4.0.0+11
      - name: cryostat-operator.v4.0.0
        version: 4.0.0
      - name: cryostat-operator.v4.0.0-9
        version: 4.0.0+9
      - name: cryostat-operator.v4.0.0-10
        version: 4.0.0+10
      - name: cryostat-operator.v3.0.1-7
        version: 3.0.1+7
      - name: cryostat-operator.v3.0.1-9
        version: 3.0.1+9
      - name: cryostat-operator.v3.0.1-8
        version: 3.0.1+8
      - name: cryostat-operator.v3.0.1
        version: 3.0.1
      - name: cryostat-operator.v3.0.1-6
        version: 3.0.1+6
      - name: cryostat-operator.v3.0.1-5
        version: 3.0.1+5
      - name: cryostat-operator.v3.0.1-3
        version: 3.0.1+3
      - name: cryostat-operator.v3.0.1-13
        version: 3.0.1+13
      - name: cryostat-operator.v3.0.1-12
        version: 3.0.1+12
      - name: cryostat-operator.v3.0.1-11
        version: 3.0.1+11
      - name: cryostat-operator.v3.0.1-10
        version: 3.0.1+10
      - name: cryostat-operator.v3.0.0-12
        version: 3.0.0+12
      - name: cryostat-operator.v3.0.0
        version: 3.0.0
      - name: cryostat-operator.v3.0.0-6
        version: 3.0.0+6
      - name: cryostat-operator.v3.0.0-8
        version: 3.0.0+8
      - name: cryostat-operator.v2.4.0
        version: 2.4.0
      - name: cryostat-operator.v2.4.0-8
        version: 2.4.0+8
      - name: cryostat-operator.v2.4.0-7
        version: 2.4.0+7
      - name: cryostat-operator.v2.4.0-6
        version: 2.4.0+6
      - name: cryostat-operator.v2.4.0-4
        version: 2.4.0+4
      - name: cryostat-operator.v2.4.0-3
        version: 2.4.0+3
      - name: cryostat-operator.v2.3.1-6
        version: 2.3.1+6
      - name: cryostat-operator.v2.3.1-8
        version: 2.3.1+8
      - name: cryostat-operator.v2.3.1-9
        version: 2.3.1+9
      - name: cryostat-operator.v2.3.1
        version: 2.3.1
      - name: cryostat-operator.v2.3.0
        version: 2.3.0
      - name: cryostat-operator.v2.3.0-6
        version: 2.3.0+6
      - name: cryostat-operator.v2.2.1-14
        version: 2.2.1+14
      - name: cryostat-operator.v2.2.1-8
        version: 2.2.1+8
      - name: cryostat-operator.v2.2.1-5
        version: 2.2.1+5
      - name: cryostat-operator.v2.2.1-13
        version: 2.2.1+13
      - name: cryostat-operator.v2.2.1-12
        version: 2.2.1+12
      - name: cryostat-operator.v2.2.1-10
        version: 2.2.1+10
      - name: cryostat-operator.v2.2.1
        version: 2.2.1
      - name: cryostat-operator.v2.2.0
        version: 2.2.0
      - name: cryostat-operator.v2.1.1-2.1659469835.p
        version: 2.1.1+2.1659469835.p
      - name: cryostat-operator.v2.1.1-2.1655334614.p
        version: 2.1.1+2.1655334614.p
      - name: cryostat-operator.v2.1.1
        version: 2.1.1
      - name: cryostat-operator.v2.1.0
        version: 2.1.0
      name: stable
    defaultChannel: stable
    packageName: cryostat-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-xtradb-cluster-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.17.0
        version: 1.17.0
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.16.1
        version: 1.16.1
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.16.0
        version: 1.16.0
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.15.1
        version: 1.15.1
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.15.0
        version: 1.15.0
      name: stable
    - currentCSV: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        version: 1.18.0-cw
      entries:
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0-cw
        version: 1.18.0-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.17.0-cw
        version: 1.17.0-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.16.1-cw
        version: 1.16.1-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.16.0-cw
        version: 1.16.0-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.15.0-cw
        version: 1.15.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator-certified-rhmp
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: community-windows-machine-config-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: community-windows-machine-config-operator.v6.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-windows/community-windows-machine-config-operator:community-4.11-6c3601f
          createdAt: "2022-08-31T18:57:22Z"
          description: An operator that enables Windows container workloads on OCP
          olm.skipRange: '>=5.0.0 <6.0.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ## WARNING
          Community distribution of the Windows Machine Config Operator.
          This is a preview build, and is not meant for production workloads.
          Issues with this distribution of WMCO can be opened against the [WMCO repository](https://github.com/openshift/windows-machine-config-operator).
          Please read through the [troubleshooting doc](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/docs/TROUBLESHOOTING.md)
          before opening an issue.
          Please ensure that when installing this operator the starting CSV you subscribe to is supported on the
          version of OKD/OCP you are using. This CSV is meant for OKD/OCP community-4.11.
          ## Documentation
          ### Introduction
          The Windows Machine Config Operator configures Windows instances into nodes, enabling Windows container workloads
          to be ran within OKD/OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/latest/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/README.md#configuring-byoh-bring-your-own-host-windows-instances).
          The operator will do all the necessary steps to configure the instance so that it can join the cluster as a worker node.
          ### Pre-requisites
          - [Cluster and OS pre-requisites](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/docs/wmco-prerequisites.md)
          ### Usage
          Please see the usage section of [README.md](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/README.md#usage).
          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.
        displayName: Community Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: preview
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-windows/community-windows-machine-config-operator:community-4.11-6c3601f
        version: 6.0.0
      entries:
      - name: community-windows-machine-config-operator.v6.0.0
        version: 6.0.0
      name: preview
    defaultChannel: preview
    packageName: community-windows-machine-config-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom
      provider-url: https://www.broadcom.com/
    name: layer7-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: layer7-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "security.brcmlabs.com/v1",
                "kind": "Gateway",
                "metadata": {
                  "name": "ssg"
                },
                "spec": {
                  "app": {
                    "image": "docker.io/caapim/gateway:11.1.2",
                    "management": {
                      "cluster": {
                        "hostname": "gateway.brcmlabs.com",
                        "password": "7layer"
                      },
                      "password": "7layer",
                      "username": "admin"
                    },
                    "replicas": 1,
                    "service": {
                      "ports": [
                        {
                          "name": "https",
                          "port": 8443,
                          "protocol": "TCP",
                          "targetPort": 8443
                        },
                        {
                          "name": "management",
                          "port": 9443,
                          "protocol": "TCP",
                          "targetPort": 9443
                        }
                      ],
                      "type": "LoadBalancer"
                    }
                  },
                  "license": {
                    "accept": false,
                    "secretName": "gateway-license"
                  },
                  "version": "11.1.2"
                },
                "status": {}
              },
              {
                "apiVersion": "security.brcmlabs.com/v1",
                "kind": "Repository",
                "metadata": {
                  "name": "my-repository"
                },
                "spec": {
                  "auth": {},
                  "branch": "main",
                  "enabled": true,
                  "endpoint": "https://github.com/\u003cusername\u003e/\u003crepository\u003e",
                  "type": "git"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7Api",
                "metadata": {
                  "name": "l7api-sample"
                },
                "spec": {
                  "deploymentTags": [
                    "ssg",
                    "ssg1"
                  ],
                  "graphmanBundle": "jsongz",
                  "portalPublished": false,
                  "serviceUrl": "/v1/api"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7Portal",
                "metadata": {
                  "name": "dev-portal"
                },
                "spec": {
                  "auth": {},
                  "deploymentTags": [
                    "ssg"
                  ],
                  "enabled": true,
                  "endpoint": "dev-ssg.brcmlabs.com",
                  "enrollmentBundle": "",
                  "portalTenant": "portal"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7StateStore",
                "metadata": {
                  "name": "state-store-example"
                },
                "spec": {
                  "redis": {
                    "database": 0,
                    "groupName": "l7gw",
                    "masterPassword": "7layer",
                    "standalone": {
                      "host": "redis-standalone",
                      "port": 6379
                    },
                    "storeId": "test",
                    "type": "standalone"
                  },
                  "type": "redis"
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: docker.io/caapim/layer7-operator:v1.2.1
          createdAt: "2025-07-04T17:22:57Z"
          description: The Layer7 Operator covers all aspects of deploying, maintaining
            and upgrading Layer7 API Gateways in Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: layer7-operator-system
          operators.openshift.io/valid-subscription: Layer7 (Broadcom) API Gateway
            license
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/caapim/layer7-operator
          support: Broadcom Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gateway is the Schema for the Gateway Custom Resource
            displayName: Gateway
            kind: Gateway
            name: gateways.security.brcmlabs.com
            version: v1
          - description: L7Api is the Schema for the l7apis API
            displayName: L7 Api
            kind: L7Api
            name: l7apis.security.brcmlabs.com
            version: v1alpha1
          - description: L7Portal is the Schema for the l7portals API
            displayName: L7 Portal
            kind: L7Portal
            name: l7portals.security.brcmlabs.com
            version: v1alpha1
          - description: L7StateStore is the Schema for the l7statestores API
            displayName: L7 State Store
            kind: L7StateStore
            name: l7statestores.security.brcmlabs.com
            version: v1alpha1
          - description: Repository is the Schema for the repositories API
            displayName: Repository
            kind: Repository
            name: repositories.security.brcmlabs.com
            version: v1
        description: |
          The [Layer7 Operator](https://github.com/CAAPIM/layer7-operator) covers all aspects of deploying, maintaining and upgrading Layer7 API Gateways in Kubernetes. The Layer7 Operator is exclusive to Kubernetes and utilizes the Container Gateway with specific emphasis on running in Ephemeral mode (without an external MySQL policy database).

          ## Getting Started
          This [Getting Started](https://github.com/CAAPIM/layer7-operator/wiki/Getting-Started) guide covers a simple introduction to the Layer7 Operator. You can find more documentation [here](https://github.com/CAAPIM/layer7-operator/wiki).
        displayName: Layer7 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - layer7-operator
        links:
        - name: Layer7 Operator
          url: https://github.com/caapim/layer7-operator
        maintainers:
        - email: gary.vermeulen@broadcom.com
          name: Gary Vermeulen
        maturity: alpha
        provider:
          name: Broadcom
          url: https://www.broadcom.com/
        relatedImages:
        - docker.io/caapim/graphman-static-init:1.0.3
        - docker.io/caapim/gateway:11.1.2
        - docker.io/caapim/layer7-operator:v1.2.1
        version: 1.2.1
      entries:
      - name: layer7-operator.v1.2.1
        version: 1.2.1
      - name: layer7-operator.v1.2.0
        version: 1.2.0
      name: preview
    defaultChannel: preview
    packageName: layer7-operator
    provider:
      name: Broadcom
      url: https://www.broadcom.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: eG Innovations
      provider-url: https://www.eginnovations.com
    name: eginnovations-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eginnovations-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eginnovations.com/v1beta1",
                "kind": "EgUniversalAgent",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eginnovations-operator"
                  },
                  "name": "eguniversalagent",
                  "namespace": "egagent"
                },
                "spec": {
                  "agent": {
                    "env": [
                      {
                        "name": "EG_MANAGER",
                        "value": "Replace with eG Manager IP"
                      },
                      {
                        "name": "EG_MANAGER_PORT",
                        "value": "Replace with eG Manager Port"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "true"
          containerImage: docker.io/eginnovations/universal-agent-operator@sha256:b8e3e26dca1b94fabfd79496feb36f3f2de4434331ef9ca810125892a936dca7
          createdAt: "2025-09-11T20:08:00Z"
          description: eG Innovations provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with eG Universal Agent Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eginnovations/universal-agent-operator
          support: eG Innovations
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgUniversalAgent is the Schema for the eguniversalagents
              API
            displayName: Eg Universal Agent
            kind: EgUniversalAgent
            name: eguniversalagents.eginnovations.com
            version: v1beta1
        description: |
          eG Innovations’ converged application and infrastructure performance monitoring capabilities provide end-to-end visibility and correlated analytics for every layer, every tier of your IT landscape.

          The eG Enterprise Universal Agent Operator automatically configures a host agent on every Kubernetes worker node. The host agent auto-discovers the worker nodes and application containers running as Pods on each node and tracks their performance and utilization levels. In-depth monitoring of applications running on containers is also provided using the same host agent. No additional agents are required for the containers.

          ### Capabilities
            * Rolls out the egagent pod per node to monitor its pods and the node itself
            * Automatic BTM profiler injection using webhook based injection for deep application monitoring and transection tracing.

          ### Installation
            1. Create the namespace "egagent" for installing eG Agent on OpenShift Cluster.

               <br />``` $ oc create ns egagent ```
            2. Add "egagent" serviceaccount to scc( Security Context Constraints ) of OpenShift Cluster as privileged using below command.

               <br />``` $ oc adm policy add-scc-to-user privileged -z egagent -n egagent ```
            2. Install eG Innovations Universal Agent operator from [OpenShift Container Platform](https://www.openshift.com/) OperatorHub.
            3. Create the eG Universal Agent CRD (Custom Resource Definition)
               a. Click "View YAML Example" and copy the yaml contents and save it to (filename).yaml on your host.
               b. To apply the CRD run the following command

               <br />``` $ oc apply -f (filename).yaml ```

          ### Required Parameters
            * EG\_MANAGER, EG\_MANAGER\_PORT - provide the ip and port of your eG Manager Installation.

          ### Advanced Options
            * Disable Certificate Checking - disable any certificate validation that may interact poorly with proxies within your cluster
            * Image Override - use a copy of the eG Iniversal Agent container image from a registry other that Docker's or Red Hat's
            * Environment variables - define environment variables for the eG Universal Agent container of for APM profiler.
            * Namespace Selectors- select a subset of namesapces to enable instrumentation for applications.

          For complete list of supported parameters and the detailed documentation available in [eG Innovations Docs](https://docs.eginnovations.com)
        displayName: eG Innovations Universal Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Tracing
        links:
        - name: eG Innovations Kubernetes Monitoring Overview
          url: https://www.eginnovations.com/documentation/Kubernetes/Introduction-to-Monitoring-Kubernetes.htm
        maintainers:
        - email: devops@eginnovations.com
          name: eG Innovations
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: eG Innovations
          url: https://www.eginnovations.com
        relatedImages:
        - docker.io/eginnovations/universal-agent-operator@sha256:b8e3e26dca1b94fabfd79496feb36f3f2de4434331ef9ca810125892a936dca7
        version: 0.0.11
      entries:
      - name: eginnovations-operator.v0.0.11
        version: 0.0.11
      name: beta
    defaultChannel: beta
    packageName: eginnovations-operator
    provider:
      name: eG Innovations
      url: https://www.eginnovations.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-trino-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trino-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:904eb8ac055171c75c8c86e4923df85461b81e6d5d9269d9cddfb03fc615f9f4
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:904eb8ac055171c75c8c86e4923df85461b81e6d5d9269d9cddfb03fc615f9f4
        version: 24.11.1
      entries:
      - name: trino-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: trino-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b50b247ed8f0ad262ed59e8aafd099919e98156472e5663957f9922616140bd9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b50b247ed8f0ad262ed59e8aafd099919e98156472e5663957f9922616140bd9
        version: 25.3.0
      entries:
      - name: trino-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: trino-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
        version: 25.7.0
      entries:
      - name: trino-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: trino-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
        version: 25.7.0
      entries:
      - name: trino-operator.v25.7.0
        version: 25.7.0
      - name: trino-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-trino-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-kafka-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kafka-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:f76b5558088bead829be8018a89e3efc1bb7a84d40429a0a1df1c88552db169b
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:f76b5558088bead829be8018a89e3efc1bb7a84d40429a0a1df1c88552db169b
        version: 24.11.1
      entries:
      - name: kafka-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: kafka-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:72b390c2d382af0a9cbe2d9ae394fc61c62b9a3f6c7310cf5f325d0f17257f1f
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:72b390c2d382af0a9cbe2d9ae394fc61c62b9a3f6c7310cf5f325d0f17257f1f
        version: 25.3.0
      entries:
      - name: kafka-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: kafka-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
        version: 25.7.0
      entries:
      - name: kafka-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: kafka-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
        version: 25.7.0
      entries:
      - name: kafka-operator.v25.7.0
        version: 25.7.0
      - name: kafka-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-kafka-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: vfunction-operator
      app.kubernetes.io/managed-by: operator
      app.kubernetes.io/name: vfunction
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      name: vfunction-server-operator
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: vFunction
      provider-url: ""
    name: vfunction-server-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vfunction-server-operator.v2.2.469
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vfunction.com/v1",
                "kind": "VfunctionServer",
                "metadata": {
                  "name": "example-vfunction-server",
                  "namespace": "vfunction",
                  "labels": {
                      "name": "vfunction",
                      "app.kubernetes.io/name": "vfunction",
                      "app.kubernetes.io/instance": "vfunction-server",
                      "app.kubernetes.io/managed-by": "operator"
                  }
                },
                "spec": {
                  "admin": {
                    "email": "admin@mycompany.com",
                    "name": "Admin",
                    "password": "Password1!"
                  },
                  "host": "http://my.domain.com",
                  "org_name": "MyCompany",
                  "upgrade": "Daily",
                  "measurements": {
                    "auto_scaling": "No",
                    "min_number_of_services": "2",
                    "max_number_of_services": "10"
                  },
                  "smtp": {
                    "identity": "",
                    "password": "",
                    "url": "",
                    "user": ""
                  },
                  "tls": {
                    "crt": "",
                    "key": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, AI/Machine Learning
          certified: Red Hat certified
          containerImage: registry.marketplace.redhat.com/rhm/vfunction/vfunction-server-operator@sha256:60d1965460101d0fd90ecc9499d0a6d4c9684a54ada61a43a253d30661fab22f
          createdAt: "2021-08-03T19:08:54Z"
          description: vFunction is a cutting-edge code analysis, machine learning,
            and automation to boost your Java modernization projects.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/vfunction-server-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/vfunction-server-operator-rhmp/support?utm_source=openshift_console
          support: vFunction
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of vFunction running on your cluster.
            displayName: vFunction
            kind: VfunctionServer
            name: vfunctionservers.vfunction.com
            version: v1
        description: |2-

          ## Introduction

          [vFunction](https://www.vfunction.com/) is a cloud-native modernization platform that combines dynamic and static code analysis, machine learning, and automation to automatically identify and extract services from existing applications. vFunction is the only platform purpose-built for modernization of Java applications.

          The vFunction modernization process starts by learning the running monolithic application, and surfacing the interdependencies within it. Using AI, the platform analyzes and identifies services that can be separated from the application. This decomposition can present a range of micro, mini, or even macro services, depending on your application environment, each being an independently deployable and scalable application component.

          vFunction automates the extraction of these services, enabling you to modernize your monolith, quickly and easily.

          ## The vFunction platform

          The platform consists of 3 basic components; the server, the controller package, and a tools package. The server runs as an operand on an OpenShift environment. The controller package is installed on the machine that runs the monolithic application which can be either a Linux or a Windows machine, and the tools are run on a development machine, with access to the code of the monolithic application.

          The controller package consists of three elements: the vFunction agent, that collects data during the dynamic analysis phase; the vFunction Viper application, that performs static analysis on the binaries of the application; and the vFunction controller that handles all the communication between the agent, Viper, and the vFunction server.

          The vFunction agent is a mix of a Java and native agent, and needs to run on the JVM that is currently running your application. [Refer to the vFunction OOB Support Matrix document](https://drive.google.com/file/d/1ccq8LFab1FrYAimDUxwgjiCOdk4QuVzs/view) for a list of supported application servers and JVMs.

          This operator installs a vFunction server instance to be connected later with one or more vFunction controllers installed on your application machine(s).

          ## Before You Start

          The vFunction operator requires a Red Hat OpenShift Kubernetes Application Platform 4.x.

          The cluster recommended configuration (which is the same as the minimal one) should consist of:

          - *1 master node*

          - *1 worker node*

          - *Storage capacity enough for dynamically provision 2 PVs of 50G each*

          - *A default StorageClass configured*

          vFunction doesn’t provide any encryption mechanism for data stored on any attached storage. If required, encryption can be achieved by the client by using its own external means on the storage itself.

          vFunction supports backup of all critical data as part of the application. Nevertheless, for complete backup of entire data it is recommended that the client apply its own storage based backup mechanism.

          ## Install Prerequisites

          ###Cluster:
          The operator was certified and tested on OCP 4.6 and 4.7.
          If you encounter any issue with other OCP versions or cloud providers please contact info@vfunction.com.

          ###Storage:
          The operator creates two new PersistentVolumeClaims (PVCs) during installation time. For their creation, it relies on the default StorageClass to dynamically provision two PersistentVolumes (PVs). The system must have a default StorageClass in place.

          Both PersistentVolumes are accessed with ReadWriteOnce mode.

          ###Project:
          The operator should be installed in a new and dedicated project (namespace). If you intend to install more than one vFunction operands in the same cluster, please use different projects for each one. Installation of more than one operand on the same cluster will allow scaling by load balancing multiple applications to different vFunction servers.

          ## Installation instructions

          1. Prepare the YAML file to use in the installation:

              a. Copy the following YAML template into a text editor:

              ```yaml
                  apiVersion: vfunction.com/v1
                  kind: VfunctionServer
                  metadata:
                    name: vfunction
                    namespace: vfunction
                    labels:
                      name: vfunction
                      app.kubernetes.io/name: vfunction
                      app.kubernetes.io/instance: vfunction-server
                      app.kubernetes.io/managed-by: operator
                  spec:
                    host: "http://my.domain.com"
                    org_name: "MyCompany"
                    upgrade: "Daily"
                    admin:
                      email: "admin@mycompany.com"
                      name: "Admin"
                      password: "Password1!"
                    smtp:
                      password: ""
                      url: ""
                      identity: ""
                      user: ""
                    measurements:
                      auto_scaling: "No"
                      min_number_of_services: "2"
                      max_number_of_services: "10"
                    tls:
                      crt: |
                        -----BEGIN CERTIFICATE-----
                        ...
                        -----END CERTIFICATE-----
                      key: |
                        -----BEGIN PRIVATE KEY-----
                        ...
                        -----END PRIVATE KEY-----
              ```
              b. Customize the template fields by replacing with your information:

              - **host**: Enter the FQDN for accessing the vFunction dashboard.
                          Make sure that you use a domain name and not an IP address. Writing "https://" will mandate TLS while writing "http://" will expose the server through HTTP.
              - **org_name**: Enter your organization name.
              - **upgrade**: Choose auto-upgrade mode
                  - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically (default).
                  - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
                  - **Never**: The automatic upgrading mechanism is off.
                  - **Maintenance**: Put the operator in a maintenance mode.

              - **admin.email**: Enter the email address of your administrator.
              - **admin.name**: Enter the name of your administrator.
              - **admin.password**: Enter the password you want to use for the vFunction administrator.
                                    The password should be at least 8 characters long, and consist of at least one lowercase letter, at least one uppercase letter, at least one number, and at least one special character.
              - **smtp.user**: Enter the email address for a designated user for the SMTP server (optional).
              - **smtp.password**: Enter this user’s password (optional).
              - **smtp.identity**: Enter this user’s password (optional).
              - **smtp.url**: Enter the SMTP server URL (optional).
              - **tls.crt**: If your host FQDN starts with “https”, paste in the certifications you have for using the TLS connection.
              - **tls.key**: If your host FQDN starts with “https”, enter the key you have for using the TLS connection.
              - **measurements.auto_scaling**: Indicates if the measurement services auto scalling is active.
              - **measurements.min_num_of_services**: Minimum Number of measurement service pods.
              - **measurements.max_num_of_services**: Maximum Number of measurement service pods.
              - **measurements.S3**: Save all measurements data to S3, instead of local PV (optional).
                  - **measurements.S3.bucket**: Measurements S3 bucket name.
                  - **measurements.S3.key**: Measurements S3 key.
                  - **measurements.S3.secret**: Measurements S3 secret.
                  - **measurements.S3.region**: Measurements S3 region.
              - **backup.mysql.S3**: Backup MySQL to S3 (optional).
                  - **backup.mysql.S3.bucket**: MySQL backup S3 bucket name.
                  - **backup.mysql.S3.key**: MySQL backup S3 key.
                  - **backup.mysql.S3.secret**: MySQL backup S3 secret.
                  - **backup.mysql.S3.region**: MySQL backup S3 region.

          1. In the Installed Operators window choose and click on vFunction Operator and then Create Instance.
          1. In the Create vFunctionServer window, paste the YAML file that you created in step 1.
          1. Click Create.
          1. You can now click on your new created vFunction operand in order to see its details and installation progress.

          ## Verifying the installation

          Check that the operand installed successfully:

          1. Check that the "Operand State" property shows “Working”.
          1. Check that the "Successfully Installed" property shows "Yes".

          If "Operand State" shows "Failed", it indicates the operand wasn’t installed correctly. Check the events and logs for all pods (operator and image containers) for any issues. In the event that you cannot troubleshoot, contact vFunction support.

          ## After installation

          The vFunction site is now accessible via the newly created vFunction application custom address, for example, my.domain.com. There are two ways you can access the vFunction dashboard.

          ### Access using the router canonical hostname

          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address, and to the ”vfunction” subdomain of the  OpenShift canonical router hostname as the alias.
          For example, ***my.domain.com.   CNAME   vfunction.apps.ocp4.my-openshift.com.***
          2. Find your cluster Router Canonical Hostname address in the newly created vfunction-route-xxx under your vFunction OpenShift project -> **Networking** -> **Routes** -> **vfunction-route-xxx** route -> **Router Canonical Hostname** field.
          You can now access the dashboard using your defined "host" spec property (as above).

          ### Access using the nginx service location

          You may use this access method if your OpenShift is installed on a provider that supports exposing LoadBalancer-type services.
          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address and to the vfunction-nginx-xxx service location.
          For example, ***my.domain.com.   CNAME   a05951ed7cdf-1394239323.us-east-1.elb.amazonaws.com.***
          1. Point your custom domain to the vfunction-nginx-xxx service's external IP location, which you can find in the OpenShift project > **Networking** -> **Services** -> **vfunction-nginx-xxx** service > **Service Address** > **Location** field.

          ## Upgrade and Rollback

          The vFunction operator includes a built-in auto-upgrade mechanism.

          You can choose one of 4 upgrading modes:
          - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically (default).
          - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
          - **Never**: The automatic upgrading mechanism is off.
          - **Maintenance**: Put the operator in a maintenance mode.

          For an on-demand upgrade,  you can change anytime the upgrading mode from “Never” to “Always”, wait for the operand to be upgraded, and change again to “Never”.

          During the upgrade, the "Operand State" property will change to “Upgrading...” and back again to “Working” after a successful upgrade.

          In case of a failure upgrading the operand, an automatic rollback to the last working version will occur. The failed version will be marked as defective and the operator will not try to upgrade to it again.

          Choosing the "Maintenance" option will tell the operator to take down all pods, so you can safely fix storage issues, return volumes from snapshots, etc.
          After finishing the maintenance, return the upgrade mode to your original desired policy.
        displayName: vFunction Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - vfunction
        - analyzing
        - services
        links:
        - name: vFunction
          url: https://vfunction.com
        - name: Terms
          url: https://www.vfunction.com/eula/
        maintainers:
        - email: raviv@vfunction.com
          name: Raviv Halivni
        maturity: alpha
        provider:
          name: vFunction
        relatedImages:
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-organizations@sha256:bcc2c3d6222af95e614318daea3108fffc6bc77aa17b23031cce296a3811b79f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-users@sha256:e47cc051953976a0e075e8cfb457cd4749e9b9e4394f2fbecab2458c2ade92ca
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-mysql@sha256:a6e06ea08fffa33b709d4d1240587bb3ab5b4cc6a28f493a227555e88e645798
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-nginx@sha256:435d1489aaf29a811081ddb51f40ae692e5833f42c31dbced6bc189c0ddc40ab
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-storage@sha256:b778cb9c724a59e51b4cb9088047be471cdf338f1138c31b81f79e50a491203f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-idp@sha256:21ca95a0209f3c557a45658452d154777f928539a93129148470562c79e68d70
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-measurements@sha256:35ffbee4b0eead58706b23cf5a702b1d55d2a7913cd0019cefbff94957d4e5d6
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-server-operator@sha256:60d1965460101d0fd90ecc9499d0a6d4c9684a54ada61a43a253d30661fab22f
        version: 2.2.469
      entries:
      - name: vfunction-server-operator.v2.2.469
        version: 2.2.469
      name: stable
    defaultChannel: stable
    packageName: vfunction-server-operator-rhmp
    provider:
      name: vFunction
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CNCF
      provider-url: ""
    name: jaeger
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jaeger-operator.v1.65.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "jaegertracing.io/v1",
                "kind": "Jaeger",
                "metadata": {
                  "name": "my-jaeger"
                },
                "spec": {
                  "strategy": "allInOne"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing
          certified: "false"
          containerImage: quay.io/jaegertracing/jaeger-operator:1.65.0
          createdAt: "2025-01-22T20:40:19Z"
          description: Provides tracing, monitoring and troubleshooting for microservices-based
            distributed systems
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/jaegertracing/jaeger-operator
          support: Jaeger Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Jaeger is the Schema for the jaegers API
            displayName: Jaeger
            kind: Jaeger
            name: jaegers.jaegertracing.io
            version: v1
        description: |-
          Jaeger, inspired by [Dapper](https://research.google.com/pubs/pub36356.html) and [OpenZipkin](http://zipkin.io/), is a distributed tracing system released as open source by Uber Technologies. It is used for monitoring and troubleshooting microservices-based distributed systems.
          ### Core capabilities
          Jaeger is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed context propagation
          * Distributed transaction monitoring
          * Root cause analysis
          * Service dependency analysis
          * Performance / latency optimization
          * OpenTracing compatible data model
          * Multiple storage backends: Badger, Cassandra, Elasticsearch, Memory.
          ### Operator features
          * **Multiple modes** - Supports `allInOne`, `production` and `streaming` [modes of deployment](https://www.jaegertracing.io/docs/latest/operator/#deployment-strategies).
          * **Configuration** - The Operator manages [configuration information](https://www.jaegertracing.io/docs/latest/operator/#configuring-the-custom-resource) when installing Jaeger instances.
          * **Storage** - [Configure storage](https://www.jaegertracing.io/docs/latest/operator/#storage-options) used by Jaeger. By default, `memory` is used. Other options include `badger`, `cassandra` or `elasticsearch`. On OpenShift, the operator can delegate creation of an Elasticsearch cluster to the Elasticsearch Operator if deployed.
          * **Agent** - can be deployed as [sidecar](https://www.jaegertracing.io/docs/latest/operator/#auto-injecting-jaeger-agent-sidecars) (default) and/or [daemonset](https://www.jaegertracing.io/docs/latest/operator/#installing-the-agent-as-daemonset).
          * **UI** - Optionally setup ingress (Kubernetes) or secure route (OpenShift) to provide [access to the Jaeger UI](https://www.jaegertracing.io/docs/latest/operator/#accessing-the-jaeger-console-ui).
          ### Before you start
          1. Ensure that the appropriate storage solution, that will be used by the Jaeger instance, is available and configured.
          2. If intending to deploy an Elasticsearch cluster via the Jaeger custom resource, then the Elasticsearch Operator must first be installed.
          ### Troubleshooting
          * https://www.jaegertracing.io/docs/latest/troubleshooting/
        displayName: Community Jaeger Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tracing
        - monitoring
        - troubleshooting
        links:
        - name: Jaeger Operator Source Code
          url: https://github.com/jaegertracing/jaeger-operator
        maintainers:
        - email: jaeger-tracing@googlegroups.com
          name: Jaeger Google Group
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: CNCF
        relatedImages:
        - quay.io/jaegertracing/jaeger-operator:1.65.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 1.65.0
      entries:
      - name: jaeger-operator.v1.65.0
        version: 1.65.0
      - name: jaeger-operator.v1.62.0
        version: 1.62.0
      - name: jaeger-operator.v1.61.0
        version: 1.61.0
      - name: jaeger-operator.v1.60.0
        version: 1.60.0
      - name: jaeger-operator.v1.59.0
        version: 1.59.0
      - name: jaeger-operator.v1.57.0
        version: 1.57.0
      - name: jaeger-operator.v1.56.0
        version: 1.56.0
      - name: jaeger-operator.v1.55.0
        version: 1.55.0
      - name: jaeger-operator.v1.53.0
        version: 1.53.0
      - name: jaeger-operator.v1.52.0
        version: 1.52.0
      - name: jaeger-operator.v1.51.0
        version: 1.51.0
      - name: jaeger-operator.v1.49.0
        version: 1.49.0
      - name: jaeger-operator.v1.48.1
        version: 1.48.1
      - name: jaeger-operator.v1.47.0
        version: 1.47.0
      - name: jaeger-operator.v1.46.0
        version: 1.46.0
      - name: jaeger-operator.v1.45.0
        version: 1.45.0
      - name: jaeger-operator.v1.44.0
        version: 1.44.0
      - name: jaeger-operator.v1.43.0
        version: 1.43.0
      - name: jaeger-operator.v1.42.0
        version: 1.42.0
      - name: jaeger-operator.v1.41.1
        version: 1.41.1
      - name: jaeger-operator.v1.41.0
        version: 1.41.0
      - name: jaeger-operator.v1.40.0
        version: 1.40.0
      - name: jaeger-operator.v1.39.0
        version: 1.39.0
      - name: jaeger-operator.v1.38.1
        version: 1.38.1
      - name: jaeger-operator.v1.38.0
        version: 1.38.0
      - name: jaeger-operator.v1.37.0
        version: 1.37.0
      - name: jaeger-operator.v1.36.0
        version: 1.36.0
      - name: jaeger-operator.v1.35.0
        version: 1.35.0
      - name: jaeger-operator.v1.34.1
        version: 1.34.1
      - name: jaeger-operator.v1.34.0
        version: 1.34.0
      - name: jaeger-operator.v1.33.0
        version: 1.33.0
      - name: jaeger-operator.v1.32.0
        version: 1.32.0
      - name: jaeger-operator.v1.31.0
        version: 1.31.0
      - name: jaeger-operator.v1.30.0
        version: 1.30.0
      - name: jaeger-operator.v1.29.1
        version: 1.29.1
      - name: jaeger-operator.v1.25.0
        version: 1.25.0
      - name: jaeger-operator.v1.24.0
        version: 1.24.0
      name: stable
    defaultChannel: stable
    packageName: jaeger
    provider:
      name: CNCF
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: JWS team
      provider-url: ""
    name: jws-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jws-operator.v2.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "web.servers.org/v1alpha1",
                "kind": "WebServer",
                "metadata": {
                  "name": "webserver-example"
                },
                "spec": {
                  "applicationName": "webapp",
                  "replicas": 1,
                  "useInsightsClient": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/jboss-webserver-5/jws5-rhel8-operator@sha256:ea042277bdcdb70b2dc32083ea868efd4bf474d5a8ec0f9b47f5bea35aaeae1e
          createdAt: "2025-09-12T08:31:52.558719Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat JBoss Web Server",
            "Red Hat Runtimes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/web-servers/jws-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Web Server is the schema for the webservers API
            displayName: Web Server
            kind: WebServer
            name: webservers.web.servers.org
            version: v1alpha1
        description: WebServer operator can manage JBoss Web Server images.
        displayName: WebServer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WebServer
        - JBoss
        links:
        - name: Jws Operator
          url: https://jws-operator.domain
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: JWS team
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:021ff671b0fb72eed635b05489ec86342fcb7ada3ba0353591fdb9328871d067
        - registry.redhat.io/jboss-webserver-5/jws5-rhel8-operator@sha256:ea042277bdcdb70b2dc32083ea868efd4bf474d5a8ec0f9b47f5bea35aaeae1e
        - registry.redhat.io/insights-runtimes-tech-preview/runtimes-inventory-rhel8-operator@sha256:58e01d11c3d30abc5b0ee3795511bd04e39aea21e2853b4f4cc6d1c8a2bd7bb0
        version: 2.2.0
      entries:
      - name: jws-operator.v2.2.0
        version: 2.2.0
      - name: jws-operator.v2.1.3
        version: 2.1.3
      - name: jws-operator.v2.1.2
        version: 2.1.2
      - name: jws-operator.v2.1.1
        version: 2.1.1
      - name: jws-operator.v2.1.0
        version: 2.1.0
      - name: jws-operator.v2.0.16
        version: 2.0.16
      - name: jws-operator.v2.0.15
        version: 2.0.15
      - name: jws-operator.v2.0.14
        version: 2.0.14
      - name: jws-operator.v2.0.13
        version: 2.0.13
      - name: jws-operator.v2.0.12
        version: 2.0.12
      - name: jws-operator.v2.0.11
        version: 2.0.11
      - name: jws-operator.v2.0.10
        version: 2.0.10
      - name: jws-operator.v2.0.9
        version: 2.0.9
      - name: jws-operator.v2.0.8
        version: 2.0.8
      - name: jws-operator.v2.0.7
        version: 2.0.7
      - name: jws-operator.v2.0.6
        version: 2.0.6
      - name: jws-operator.v2.0.5
        version: 2.0.5
      - name: jws-operator.v2.0.4
        version: 2.0.4
      - name: jws-operator.v2.0.3
        version: 2.0.3
      - name: jws-operator.v2.0.2
        version: 2.0.2
      - name: jws-operator.v2.0.1
        version: 2.0.1
      name: alpha
    defaultChannel: alpha
    packageName: jws-operator
    provider:
      name: JWS team
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Windup
      provider-url: ""
    name: windup-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: windup-operator.v0.0.25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
             {
               "apiVersion": "windup.jboss.org/v1",
               "kind": "Windup",
               "metadata": {
                 "name": "windup",
                 "labels": {
                   "application": "windup"
                 }
               },
               "spec": {}
             }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/windupeng/windup-operator-native:0.0.25
          createdAt: 01/22/24
          description: Windup is an analysis tool that supports the modernization
            and migration of Java applications.
          imagePullPolicy: Always
          operatorframework.io/suggested-namespace: windup
          repository: https://github.com/windup/windup-operator/
          support: https://issues.redhat.com/projects/WINDUP
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "Create a Windup Web Console application.  \n\nTo access
              the Windup Web Console application from the \"Developer\" perspective,
              please go to Topology and click on the \"Open URL\" icon on the web-console
              pod.  \n\nAlternatively from the \"Administrator\" perspective, please
              go to Networking->Routes and click on the Location hyperlink.  \n\nKnown
              issue  \nIf you want to customize the Windup Web Console instance's
              parameters and you can not see any in the `Form view`, please switch
              to the `YAML view` and change them as needed.\n"
            displayName: Windup
            kind: Windup
            name: windups.windup.jboss.org
            version: v1
        description: |
          Windup is a web console application that supports large-scale Java application modernization and migration projects across a broad range of transformations and use cases.

          It analyzes application code, supports effort estimation, accelerates code migration, and enables users to move applications to containers.

          For more information please refer to the [https://windup.github.io/](https://windup.github.io/) page.
        displayName: Windup Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - modernization
        - windup
        links:
        - name: Website
          url: https://windup.github.io
        - name: GitHub
          url: https://github.com/windup/windup-operator
        maintainers:
        - email: migrate@redhat.com
          name: Windup
        maturity: alpha
        provider:
          name: Windup
        relatedImages:
        - quay.io/windupeng/windup-operator-native:0.0.25
        version: 0.0.25
      entries:
      - name: windup-operator.v0.0.25
        version: 0.0.25
      - name: windup-operator.v0.0.24
        version: 0.0.24
      - name: windup-operator.v0.0.23
        version: 0.0.23
      - name: windup-operator.v0.0.22
        version: 0.0.22
      - name: windup-operator.v0.0.21
        version: 0.0.21
      - name: windup-operator.v0.0.20
        version: 0.0.20
      - name: windup-operator.v0.0.19
        version: 0.0.19
      - name: windup-operator.v0.0.18
        version: 0.0.18
      - name: windup-operator.v0.0.17
        version: 0.0.17
      - name: windup-operator.v0.0.16
        version: 0.0.16
      - name: windup-operator.v0.0.15
        version: 0.0.15
      - name: windup-operator.v0.0.14
        version: 0.0.14
      - name: windup-operator.v0.0.13
        version: 0.0.13
      - name: windup-operator.v0.0.12
        version: 0.0.12
      - name: windup-operator.v0.0.11
        version: 0.0.11
      name: alpha
    defaultChannel: alpha
    packageName: windup-operator
    provider:
      name: Windup
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stolostron
      provider-url: ""
    name: cluster-aas-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-aas-operator.v0.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplate",
                "metadata": {
                  "name": "clustertemplate-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplateInstance",
                "metadata": {
                  "name": "clustertemplateinstance-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplateQuota",
                "metadata": {
                  "name": "clustertemplatequota-sample"
                },
                "spec": {}
              }
            ]
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/cluster-templates-operator:2.11.0-0c3131154065397d37e3b66ae635d7ea88650bf9
          createdAt: "2024-06-06T19:34:44Z"
          description: Easily install fully configured clusters with guard-rails.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/cluster-templates-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents instance of a cluster
            displayName: Cluster template instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.clustertemplate.openshift.io
            version: v1alpha1
          - description: Defines which ClusterTemplates can be used in a given namespace
            displayName: Cluster template quota
            kind: ClusterTemplateQuota
            name: clustertemplatequotas.clustertemplate.openshift.io
            version: v1alpha1
          - description: Template of a cluster - both installation and post-install
              setup are defined as ArgoCD application spec. Any application source
              is supported - typically a Helm chart
            displayName: Cluster template
            kind: ClusterTemplate
            name: clustertemplates.clustertemplate.openshift.io
            version: v1alpha1
          - kind: ClusterTemplateSetup
            name: clustertemplatesetup.clustertemplate.openshift.io
            version: v1alpha1
          - kind: Config
            name: config.clustertemplate.openshift.io
            version: v1alpha1
        description: |
          **Self-service clusters with guardrails!**
          Cluster as a service operator provides an easy way to define clusters as templates and allows creating instances of those templates even for non-privileged developer/devops engineers. Cluster templates operator also allows specifing quotas for the developer/devops engineers.

          ### User Interface
          The User interface is installed but is not enabled by default. To enable go to `Home` -> `Overview` -> in the status card click `Dynamic Plugins` -> click `View all` -> click the `pencil` icon -> pick `Enable` and hit `Save`. You can access the UI in the `All Clusters` perspective under `Infrastructure` -> `Cluster Templates`.

          ### Documentation
          Documentation can be found on our [website](https://github.com/stolostron/cluster-templates-operator).

          ### Getting help
          To report any issues or ask questions, create an issue on our [Github repo](https://github.com/stolostron/cluster-templates-operator/issues)
        displayName: Cluster as a service operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cluster
        - templates
        - hive
        - hypershift
        - selfservice
        - openshift
        - multicluster
        - quota
        - helm
        - argocd
        links:
        - name: Cluster as a service operator
          url: https://github.com/stolostron/cluster-templates-operator
        maintainers:
        - email: rawagner@redhat.com
          name: Rastislav Wagner
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Stolostron
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/stolostron/cluster-templates-operator:2.11.0-0c3131154065397d37e3b66ae635d7ea88650bf9
        version: 0.1.5
      entries:
      - name: cluster-aas-operator.v0.1.5
        version: 0.1.5
      - name: cluster-aas-operator.v0.1.4
        version: 0.1.4
      - name: cluster-aas-operator.v0.1.3
        version: 0.1.3
      - name: cluster-aas-operator.v0.1.2
        version: 0.1.2
      - name: cluster-aas-operator.v0.1.1
        version: 0.1.1
      - name: cluster-aas-operator.v0.1.0
        version: 0.1.0
      - name: cluster-aas-operator.v0.0.6
        version: 0.0.6
      - name: cluster-aas-operator.v0.0.5
        version: 0.0.5
      - name: cluster-aas-operator.v0.0.4
        version: 0.0.4
      - name: cluster-aas-operator.v0.0.3
        version: 0.0.3
      - name: cluster-aas-operator.v0.0.2
        version: 0.0.2
      - name: cluster-aas-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cluster-aas-operator
    provider:
      name: Stolostron
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: ibm-spectrum-scale-csi-operator
      app.kubernetes.io/managed-by: ibm-spectrum-scale-csi-operator
      app.kubernetes.io/name: ibm-spectrum-scale-csi-operator
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-spectrum-scale-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-spectrum-scale-csi-operator.v2.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "CSIScaleOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-spectrum-scale-csi-operator",
                    "app.kubernetes.io/managed-by": "ibm-spectrum-scale-csi-operator",
                    "app.kubernetes.io/name": "ibm-spectrum-scale-csi-operator",
                    "release": "ibm-spectrum-scale-csi-operator"
                  },
                  "name": "ibm-spectrum-scale-csi",
                  "namespace": "ibm-spectrum-scale-csi-driver"
                },
                "spec": {
                  "attacherNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "clusters": [
                    {
                      "id": "\u003cPrimary Cluster ID\u003e",
                      "primary": {
                        "primaryFs": "\u003c Primary Filesystem \u003e"
                      },
                      "restApi": [
                        {
                          "guiHost": "\u003c Primary cluster GUI IP/Hostname \u003e"
                        }
                      ],
                      "secrets": "secret1",
                      "secureSslMode": false
                    }
                  ],
                  "pluginNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "provisionerNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "resizerNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "snapshotterNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ]
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/ibm-spectrum-scale/ibm-spectrum-scale-csi-operator@sha256:b62ad588fbc338090f405740fd8bceaf59b3267cdb532c60ae965643ccae4d5d
          createdAt: Thu May 5 04:42:15 IST 2022
          description: An operator for deploying and managing the IBM Spectrum Scale
            CSI Driver.
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-spectrum-scale-csi/
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIScaleOperator is the Schema for the csiscaleoperators
              API
            displayName: IBM CSI Spectrum Scale Driver
            kind: CSIScaleOperator
            name: csiscaleoperators.csi.ibm.com
            version: v1
        description: |
          The IBM Spectrum Scale CSI Operator for Kubernetes installs, manages,
          upgrades the IBM Spectrum Scale CSI Driver on OpenShift and Kubernetes
          for provisioning of persistent volumes on IBM Spectrum Scale.

          ### Documentation Documentation for IBM Spectrum Scale CSI Driver and Operator can be found at [IBM Documentation](https://www.ibm.com/docs/en/spectrum-scale-csi).

          ### Getting help
          IBM Spectrum Scale CSI driver is part of the IBM Spectrum Scale offering.
          Please follow the [IBM support procedure](https://www.ibm.com/mysupport/s/?language=en_US) for any issues with the driver.

          For non-urgent issues, suggestions, recommendations feel free to open an
          issue in [github](https://github.com/IBM/ibm-spectrum-scale-csi). Issues will be addressed as team availability permits.

          **Note:** Custom Resource example represents the minimum settings needed to run
          the operator. If your environment needs more advanced settings (e.g. remote clusters,
          node mapping, image pull secret, Taint and Toleration etc.) please refer to  the sample [Custom Resource](https://raw.githubusercontent.com/IBM/ibm-spectrum-scale-csi/v2.6.0/operator/config/samples/csiscaleoperators.csi.ibm.com_cr.yaml).
        displayName: IBM Spectrum Scale CSI Plugin Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - GPFS
        - Spectrum Scale
        - CSI Driver
        links:
        - name: IBM CSI Spectrum Scale Documentation
          url: https://www.ibm.com/docs/en/spectrum-scale-csi
        - name: CSI Developer Documentation
          url: https://kubernetes-csi.github.io/docs/
        maintainers:
        - email: madhu.punjabi@in.ibm.com
          name: Madhu Thorat
        - email: amdabhad@in.ibm.com
          name: Amol Dabhade
        - email: Nitish.Kumar4@ibm.com
          name: Nitish Kumar
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibm-spectrum-scale/ibm-spectrum-scale-csi-operator@sha256:b62ad588fbc338090f405740fd8bceaf59b3267cdb532c60ae965643ccae4d5d
        version: 2.6.0
      entries:
      - name: ibm-spectrum-scale-csi-operator.v2.6.0
        version: 2.6.0
      - name: ibm-spectrum-scale-csi-operator.v2.5.1
        version: 2.5.1
      - name: ibm-spectrum-scale-csi-operator.v2.5.0
        version: 2.5.0
      - name: ibm-spectrum-scale-csi-operator.v2.4.0
        version: 2.4.0
      - name: ibm-spectrum-scale-csi-operator.v2.3.1
        version: 2.3.1
      - name: ibm-spectrum-scale-csi-operator.v2.3.0
        version: 2.3.0
      name: stable
    defaultChannel: stable
    packageName: ibm-spectrum-scale-csi-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Managed Openshift Black Belt
      provider-url: ""
    name: ecr-secret-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ecr-secret-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "ecr-secret-sample",
                  "namespace": "test-ecr-secret-operator"
                },
                "spec": {
                  "ecr_registry": "ACCOUNT_ID.dkr.ecr.us-east-2.amazonaws.com",
                  "frequency": "10h",
                  "generated_secret_name": "ecr-docker-secret",
                  "region": "us-east-2"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ecr-secret-operator:v0.3.2
          operatorframework.io/suggested-namespace: ecr-secret-operator
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ecr-secret-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret is the Schema for the secrets API
            displayName: Secret
            kind: Secret
            name: secrets.ecr.mobb.redhat.com
            version: v1alpha1
        description: "# ECR Secret Operator\n\nAmazon Elastic Container Registry [Private
          Registry Authentication](https://docs.aws.amazon.com/AmazonECR/latest/userguide/registry_auth.html)
          provides a temporary token that is valid only for 12 hours. It is a challenge
          for automatic container image build process to refresh the token or secret
          in a timely manner.\n\nThis operators frequently talks with AWS ECR GetAuthroization
          Token and create/update the secret, so that the service account can perform
          docker image build.\n\n\n## How to use this operator\n\n### Prerequisites\n\n*
          [Create an ECR private repository](https://docs.aws.amazon.com/AmazonECR/latest/userguide/repository-create.html)\n*
          Create An Openshift Cluster\n* Provide AWS Authentication to the operator.
          Two Options:\n  **Note: You have to provide a secret aws-ecr-cloud-credentials
          in the installed namespace**\n  * [IAM User](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_user.md)\n
          \ * [STS Assume Role](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_assume_role.md)\n\n###
          Install the operator\n\n### Create the ECR Secret CRD\n\n```\napiVersion:
          ecr.mobb.redhat.com/v1alpha1\nkind: Secret\nmetadata:\n  name: ecr-secret\n
          \ namespace: test-ecr-secret-operator\nspec:\n  generated_secret_name: ecr-docker-secret\n
          \ ecr_registry: [ACCOUNT_ID].dkr.ecr.us-east-2.amazonaws.com\n  frequency:
          10h\n  region: us-east-2\n```\n\n```\noc create -f CRD_FILE\n```\n\nA docker
          registry secret is created by the operator momentally and the token is patched
          every 10 hours\n\n```\noc get secret ecr-docker-secret   \nNAME                TYPE
          \                            DATA   AGE\necr-docker-secret   kubernetes.io/dockerconfigjson
          \  1      16h\n```\n\n### A sample build process with generated secret\n\n\nLink
          the secret to builder\n\n```\noc secrets link builder ecr-docker-secret
          \n```\n\nConfigure [build config](https://github.com/rh-mobb/ecr-secret-operator/tree/main/samples/build-config.yaml)
          to point to your ECR Container repository\n\n```\noc create imagestream
          ruby\noc tag openshift/ruby:2.5-ubi8 ruby:2.5\noc create -f samples/build-config.yaml\noc
          start-build ruby-sample-build --wait\n```\n\nBuild should succeed and push
          the image to the the private ECR Container repository\n\n"
        displayName: ECR Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - redhat
        links:
        - name: Ecr Secret Operator
          url: https://github.com/rh-mobb/ecr-secret-operator
        maintainers:
        - email: shading@redhat.com
          name: Shaozhen Ding
        maturity: alpha
        provider:
          name: Managed Openshift Black Belt
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/mobb/ecr-secret-operator:v0.3.2
        version: 0.3.2
      entries:
      - name: ecr-secret-operator.v0.3.2
        version: 0.3.2
      name: v0.3
    - currentCSV: ecr-secret-operator.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "ecr-secret-sample",
                  "namespace": "test-ecr-secret-operator"
                },
                "spec": {
                  "ecr_registry": "ACCOUNT_ID.dkr.ecr.us-east-2.amazonaws.com",
                  "frequency": "10h",
                  "generated_secret_name": "ecr-docker-secret",
                  "region": "us-east-2"
                }
              },
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "ArgoHelmRepoSecret",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ecr-secret-operator",
                    "app.kubernetes.io/instance": "argohelmreposecret-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "argohelmreposecret",
                    "app.kubernetes.io/part-of": "ecr-secret-operator"
                  },
                  "name": "argohelmreposecret-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ecr-secret-operator:v0.4.1
          createdAt: "2023-04-14T14:52:13Z"
          operatorframework.io/suggested-namespace: ecr-secret-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ecr-secret-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ArgoHelmRepoSecret is the Schema for the argohelmreposecrets
              API
            displayName: Argo Helm Repo Secret
            kind: ArgoHelmRepoSecret
            name: argohelmreposecrets.ecr.mobb.redhat.com
            version: v1alpha1
          - description: Secret is the Schema for the secrets API
            displayName: Secret
            kind: Secret
            name: secrets.ecr.mobb.redhat.com
            version: v1alpha1
        description: "# ECR Secret Operator\n\nAmazon Elastic Container Registry [Private
          Registry Authentication](https://docs.aws.amazon.com/AmazonECR/latest/userguide/registry_auth.html)
          provides a temporary token that is valid only for 12 hours. It is a challenge
          for automatic container image build process to refresh the token or secret
          in a timely manner.\n\nThis operators frequently talks with AWS ECR GetAuthroization
          Token and create/update the secret, so that the service account can perform
          docker image build.\n\n\n## How to use this operator\n\n### Prerequisites\n\n*
          [Create an ECR private repository](https://docs.aws.amazon.com/AmazonECR/latest/userguide/repository-create.html)\n*
          Create An Openshift Cluster\n* Provide AWS Authentication to the operator.
          Two Options:\n  **Note: You have to provide a secret aws-ecr-cloud-credentials
          in the installed namespace**\n  * [IAM User](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_user.md)\n
          \ * [STS Assume Role](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_assume_role.md)\n\n###
          Install the operator\n\n### Create the ECR Secret CRD\n\n```\napiVersion:
          ecr.mobb.redhat.com/v1alpha1\nkind: Secret\nmetadata:\n  name: ecr-secret\n
          \ namespace: test-ecr-secret-operator\nspec:\n  generated_secret_name: ecr-docker-secret\n
          \ ecr_registry: [ACCOUNT_ID].dkr.ecr.us-east-2.amazonaws.com\n  frequency:
          10h\n  region: us-east-2\n```\n\n```\noc create -f CRD_FILE\n```\n\nA docker
          registry secret is created by the operator momentally and the token is patched
          every 10 hours\n\n```\noc get secret ecr-docker-secret   \nNAME                TYPE
          \                            DATA   AGE\necr-docker-secret   kubernetes.io/dockerconfigjson
          \  1      16h\n```\n\n### A sample build process with generated secret\n\n\nLink
          the secret to builder\n\n```\noc secrets link builder ecr-docker-secret
          \n```\n\nConfigure [build config](https://github.com/rh-mobb/ecr-secret-operator/tree/main/samples/build-config.yaml)
          to point to your ECR Container repository\n\n```\noc create imagestream
          ruby\noc tag openshift/ruby:2.5-ubi8 ruby:2.5\noc create -f samples/build-config.yaml\noc
          start-build ruby-sample-build --wait\n```\n\nBuild should succeed and push
          the image to the the private ECR Container repository\n\n"
        displayName: ECR Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - redhat
        links:
        - name: Ecr Secret Operator
          url: https://github.com/rh-mobb/ecr-secret-operator
        maintainers:
        - email: shading@redhat.com
          name: Shaozhen Ding
        maturity: alpha
        provider:
          name: Managed Openshift Black Belt
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/mobb/ecr-secret-operator:v0.4.1
        version: 0.4.1
      entries:
      - name: ecr-secret-operator.v0.4.1
        version: 0.4.1
      - name: ecr-secret-operator.v0.4.0
        version: 0.4.0
      name: v0.4
    defaultChannel: v0.4
    packageName: ecr-secret-operator
    provider:
      name: Managed Openshift Black Belt
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-apigatewayv2-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-apigatewayv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Deployment",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "API",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Authorizer",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Route",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Stage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "VPCLink",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/apigatewayv2-controller:1.1.1
          createdAt: "2025-09-19T21:58:07Z"
          description: AWS API Gateway v2 controller is a service controller for managing
            API Gateway v2 resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIMapping represents the state of an AWS apigatewayv2 APIMapping
              resource.
            displayName: APIMapping
            kind: APIMapping
            name: apimappings.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: API represents the state of an AWS apigatewayv2 API resource.
            displayName: API
            kind: API
            name: apis.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Authorizer represents the state of an AWS apigatewayv2 Authorizer
              resource.
            displayName: Authorizer
            kind: Authorizer
            name: authorizers.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Deployment represents the state of an AWS apigatewayv2 Deployment
              resource.
            displayName: Deployment
            kind: Deployment
            name: deployments.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: DomainName represents the state of an AWS apigatewayv2 DomainName
              resource.
            displayName: DomainName
            kind: DomainName
            name: domainnames.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Integration represents the state of an AWS apigatewayv2 Integration
              resource.
            displayName: Integration
            kind: Integration
            name: integrations.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Route represents the state of an AWS apigatewayv2 Route resource.
            displayName: Route
            kind: Route
            name: routes.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Stage represents the state of an AWS apigatewayv2 Stage resource.
            displayName: Stage
            kind: Stage
            name: stages.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: VPCLink represents the state of an AWS apigatewayv2 VPCLink
              resource.
            displayName: VPCLink
            kind: VPCLink
            name: vpclinks.apigatewayv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon API Gateway v2 resources in AWS from within your Kubernetes cluster. Currently only HTTP APIs are supported. WebSocket API support will be added soon. For HTTP APIs, DomainName and APIMapping resource support will be added soon.

          **About Amazon API Gateway v2**

          Amazon API Gateway is a fully managed service that makes it easy for developers to create, publish, maintain, monitor, and secure APIs at any scale. APIs act as the "front door" for applications to access data, business logic, or functionality from your backend services. Using API Gateway, you can create RESTful APIs and WebSocket APIs that enable real-time two-way communication applications. API Gateway supports containerized and serverless workloads, as well as web applications.

          API Gateway handles all the tasks involved in accepting and processing up to hundreds of thousands of concurrent API calls, including traffic management, CORS support, authorization and access control, throttling, monitoring, and API version management. API Gateway has no minimum fees or startup costs. You pay for the API calls you receive and the amount of data transferred out and, with the API Gateway tiered pricing model, you can reduce your cost as your API usage scales.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon API Gateway v2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apigatewayv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon API Gateway v2 Developer Resources
          url: https://aws.amazon.com/api-gateway/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: api gateway maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/apigatewayv2-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-apigatewayv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-apigatewayv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-apigatewayv2-controller.v1.0.30
        version: 1.0.30
      - name: ack-apigatewayv2-controller.v1.0.29
        version: 1.0.29
      - name: ack-apigatewayv2-controller.v1.0.28
        version: 1.0.28
      - name: ack-apigatewayv2-controller.v1.0.26
        version: 1.0.26
      - name: ack-apigatewayv2-controller.v1.0.25
        version: 1.0.25
      - name: ack-apigatewayv2-controller.v1.0.23
        version: 1.0.23
      - name: ack-apigatewayv2-controller.v1.0.21
        version: 1.0.21
      - name: ack-apigatewayv2-controller.v1.0.17
        version: 1.0.17
      - name: ack-apigatewayv2-controller.v1.0.16
        version: 1.0.16
      - name: ack-apigatewayv2-controller.v1.0.15
        version: 1.0.15
      - name: ack-apigatewayv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-apigatewayv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-apigatewayv2-controller.v1.0.7
        version: 1.0.7
      - name: ack-apigatewayv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-apigatewayv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-apigatewayv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-apigatewayv2-controller.v1.0.3
        version: 1.0.3
      - name: ack-apigatewayv2-controller.v1.0.2
        version: 1.0.2
      - name: ack-apigatewayv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-apigatewayv2-controller.v0.1.6
        version: 0.1.6
      - name: ack-apigatewayv2-controller.v0.1.4
        version: 0.1.4
      - name: ack-apigatewayv2-controller.v0.1.3
        version: 0.1.3
      - name: ack-apigatewayv2-controller.v0.1.2
        version: 0.1.2
      - name: ack-apigatewayv2-controller.v0.1.1
        version: 0.1.1
      - name: ack-apigatewayv2-controller.v0.1.0
        version: 0.1.0
      - name: ack-apigatewayv2-controller.v0.0.22
        version: 0.0.22
      - name: ack-apigatewayv2-controller.v0.0.21
        version: 0.0.21
      - name: ack-apigatewayv2-controller.v0.0.20
        version: 0.0.20
      - name: ack-apigatewayv2-controller.v0.0.19
        version: 0.0.19
      - name: ack-apigatewayv2-controller.v0.0.18
        version: 0.0.18
      - name: ack-apigatewayv2-controller.v0.0.17
        version: 0.0.17
      - name: ack-apigatewayv2-controller.v0.0.16
        version: 0.0.16
      - name: ack-apigatewayv2-controller.v0.0.15
        version: 0.0.15
      - name: ack-apigatewayv2-controller.v0.0.14
        version: 0.0.14
      - name: ack-apigatewayv2-controller.v0.0.13
        version: 0.0.13
      - name: ack-apigatewayv2-controller.v0.0.12
        version: 0.0.12
      - name: ack-apigatewayv2-controller.v0.0.11
        version: 0.0.11
      - name: ack-apigatewayv2-controller.v0.0.10
        version: 0.0.10
      - name: ack-apigatewayv2-controller.v0.0.9
        version: 0.0.9
      name: alpha
    defaultChannel: alpha
    packageName: ack-apigatewayv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Backube
      provider-url: https://github.com/backube
    name: snapscheduler
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: snapscheduler.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "snapscheduler.backube/v1",
                "kind": "SnapshotSchedule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "snapscheduler",
                    "app.kubernetes.io/instance": "daily",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "snapshotschedule",
                    "app.kubernetes.io/part-of": "snapscheduler"
                  },
                  "name": "daily"
                },
                "spec": {
                  "retention": {
                    "maxCount": 7
                  },
                  "schedule": "0 0 * * *"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/backube/snapscheduler:3.5.0
          createdAt: "2025-05-14T14:46:43Z"
          description: The SnapScheduler operator takes snapshots of Kubernetes CSI-based
            persistent volumes according to user-supplied schedules.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/snapscheduler
          support: https://github.com/backube/snapscheduler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SnapshotSchedule defines a schedule for taking automated
              snapshots of PVC(s)
            displayName: Snapshot Schedule
            kind: SnapshotSchedule
            name: snapshotschedules.snapscheduler.backube
            version: v1
        description: |-
          ## About this operator

          The SnapScheduler operator takes snapshots of CSI-based PersistentVolumes
          according to a configurable
          [Cron-like](https://en.wikipedia.org/wiki/Cron#Overview) schedule. The
          schedules include configurable retention policies for snapshots as well as
          selectors to limit the volumes that are snapshotted. An example schedule
          could be:

          > *Snapshot **all volumes** in a namespace **daily at midnight**, retaining
          > the most recent **7** snapshots for each volume.*

          Multiple schedules can be combined to provide more elaborate protection
          schemes. For example, a given volume (or collection of volumes) could be
          protected with:

          - 6 hourly snapshots
          - 7 daily snapshots
          - 4 weekly snapshots
          - 12 monthly snapshots

          ### How it works

          The operator watches for `SnapshotSchedule` CRs in each namespace. When the
          current time matches the schedule's cronspec, the operator creates a
          `VolumeSnapshot` object for each `PersistentVolumeClaim` in the namespace
          (or subset thereof if a label selector is provided). The `VolumeSnapshot`
          objects are named according to the template:
          `<pvcname>-<schedulename>-<timestamp>`. After creating the new snapshots,
          the oldest snapshots are removed if necessary, according to the retention
          policy of the schedule.

          Please see the [full
          documentation](https://backube.github.io/snapscheduler/) for more
          information.

          ## Requirements

          - OpenShift >= 4.7 (Kubernetes >= 1.20)
          - CSI-based storage driver that supports snapshots (i.e. has the
            `CREATE_DELETE_SNAPSHOT` capability)

          ## Examples

          The schedule for snapshotting is controlled by the
          `snapshotschedules.snapscheduler.backube` Custom Resource. This is a
          namespaced resource that applies only to the PersistentVolumeClaims in its
          namespace.

          Below is a simple example that keeps 7 daily (taken at midnight) snapshots
          of all PVCs in a given namespace:

          ```yaml
          ---
          apiVersion: snapscheduler.backube/v1
          kind: SnapshotSchedule
          metadata:
            name: daily
          spec:
            retention:
              maxCount: 7
            schedule: "0 0 * * *"
          ```

          See the [usage
          documentation](https://backube.github.io/snapscheduler/usage.html) for full
          details, including how to:

          - add label selectors to restrict which PVCs this schedule applies to
          - set the VolumeSnapshotClass used by the schedule
          - apply custom labels to the automatically created VolumeSnapshot objects
        displayName: SnapScheduler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - snapshot
        - storage
        - scheduler
        links:
        - name: Documentation
          url: https://backube.github.io/snapscheduler/
        - name: Source
          url: https://github.com/backube/snapscheduler
        - name: Blog
          url: https://backube.github.io/
        maintainers:
        - email: jstrunk@redhat.com
          name: John Strunk
        maturity: stable
        minKubeVersion: 1.20.0-0
        provider:
          name: Backube
          url: https://github.com/backube
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.19.1
        - quay.io/backube/snapscheduler:3.5.0
        version: 3.5.0
      entries:
      - name: snapscheduler.v3.5.0
        version: 3.5.0
      - name: snapscheduler.v3.4.0
        version: 3.4.0
      - name: snapscheduler.v3.3.0
        version: 3.3.0
      - name: snapscheduler.v3.2.0
        version: 3.2.0
      - name: snapscheduler.v3.1.0
        version: 3.1.0
      - name: snapscheduler.v3.0.0
        version: 3.0.0
      - name: snapscheduler.v2.1.0
        version: 2.1.0
      - name: snapscheduler.v2.0.0
        version: 2.0.0
      name: candidate
    - currentCSV: snapscheduler.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "snapscheduler.backube/v1",
                "kind": "SnapshotSchedule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "snapscheduler",
                    "app.kubernetes.io/instance": "daily",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "snapshotschedule",
                    "app.kubernetes.io/part-of": "snapscheduler"
                  },
                  "name": "daily"
                },
                "spec": {
                  "retention": {
                    "maxCount": 7
                  },
                  "schedule": "0 0 * * *"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/backube/snapscheduler:3.5.0
          createdAt: "2025-05-14T14:46:43Z"
          description: The SnapScheduler operator takes snapshots of Kubernetes CSI-based
            persistent volumes according to user-supplied schedules.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/snapscheduler
          support: https://github.com/backube/snapscheduler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SnapshotSchedule defines a schedule for taking automated
              snapshots of PVC(s)
            displayName: Snapshot Schedule
            kind: SnapshotSchedule
            name: snapshotschedules.snapscheduler.backube
            version: v1
        description: |-
          ## About this operator

          The SnapScheduler operator takes snapshots of CSI-based PersistentVolumes
          according to a configurable
          [Cron-like](https://en.wikipedia.org/wiki/Cron#Overview) schedule. The
          schedules include configurable retention policies for snapshots as well as
          selectors to limit the volumes that are snapshotted. An example schedule
          could be:

          > *Snapshot **all volumes** in a namespace **daily at midnight**, retaining
          > the most recent **7** snapshots for each volume.*

          Multiple schedules can be combined to provide more elaborate protection
          schemes. For example, a given volume (or collection of volumes) could be
          protected with:

          - 6 hourly snapshots
          - 7 daily snapshots
          - 4 weekly snapshots
          - 12 monthly snapshots

          ### How it works

          The operator watches for `SnapshotSchedule` CRs in each namespace. When the
          current time matches the schedule's cronspec, the operator creates a
          `VolumeSnapshot` object for each `PersistentVolumeClaim` in the namespace
          (or subset thereof if a label selector is provided). The `VolumeSnapshot`
          objects are named according to the template:
          `<pvcname>-<schedulename>-<timestamp>`. After creating the new snapshots,
          the oldest snapshots are removed if necessary, according to the retention
          policy of the schedule.

          Please see the [full
          documentation](https://backube.github.io/snapscheduler/) for more
          information.

          ## Requirements

          - OpenShift >= 4.7 (Kubernetes >= 1.20)
          - CSI-based storage driver that supports snapshots (i.e. has the
            `CREATE_DELETE_SNAPSHOT` capability)

          ## Examples

          The schedule for snapshotting is controlled by the
          `snapshotschedules.snapscheduler.backube` Custom Resource. This is a
          namespaced resource that applies only to the PersistentVolumeClaims in its
          namespace.

          Below is a simple example that keeps 7 daily (taken at midnight) snapshots
          of all PVCs in a given namespace:

          ```yaml
          ---
          apiVersion: snapscheduler.backube/v1
          kind: SnapshotSchedule
          metadata:
            name: daily
          spec:
            retention:
              maxCount: 7
            schedule: "0 0 * * *"
          ```

          See the [usage
          documentation](https://backube.github.io/snapscheduler/usage.html) for full
          details, including how to:

          - add label selectors to restrict which PVCs this schedule applies to
          - set the VolumeSnapshotClass used by the schedule
          - apply custom labels to the automatically created VolumeSnapshot objects
        displayName: SnapScheduler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - snapshot
        - storage
        - scheduler
        links:
        - name: Documentation
          url: https://backube.github.io/snapscheduler/
        - name: Source
          url: https://github.com/backube/snapscheduler
        - name: Blog
          url: https://backube.github.io/
        maintainers:
        - email: jstrunk@redhat.com
          name: John Strunk
        maturity: stable
        minKubeVersion: 1.20.0-0
        provider:
          name: Backube
          url: https://github.com/backube
        relatedImages:
        - quay.io/backube/snapscheduler:3.5.0
        - quay.io/brancz/kube-rbac-proxy:v0.19.1
        version: 3.5.0
      entries:
      - name: snapscheduler.v3.5.0
        version: 3.5.0
      - name: snapscheduler.v3.4.0
        version: 3.4.0
      - name: snapscheduler.v3.3.0
        version: 3.3.0
      - name: snapscheduler.v3.2.0
        version: 3.2.0
      - name: snapscheduler.v3.1.0
        version: 3.1.0
      - name: snapscheduler.v3.0.0
        version: 3.0.0
      - name: snapscheduler.v2.1.0
        version: 2.1.0
      - name: snapscheduler.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: snapscheduler
    provider:
      name: Backube
      url: https://github.com/backube
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: watcher-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: watcher-operator.v1.0.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "Watcher",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcher",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcher",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcher"
                },
                "spec": {
                  "apiServiceTemplate": {
                    "tls": {
                      "caBundleSecretName": "combined-ca-bundle"
                    }
                  },
                  "databaseInstance": "openstack"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherAPI",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcher-operator"
                  },
                  "name": "watcherapi-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherApplier",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcherapplier-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcherapplier",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcherapplier-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherDecisionEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcherdecisionengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcherdecisionengine",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcherdecisionengine-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-09-02T13:32:31Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.14
          operatorframework.io/suggested-namespace: openstack-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: WatcherAPI is the Schema for the watcherapis API
            displayName: Watcher API
            kind: WatcherAPI
            name: watcherapis.watcher.openstack.org
            version: v1beta1
          - description: WatcherApplier is the Schema for the watcherappliers API
            displayName: Watcher Applier
            kind: WatcherApplier
            name: watcherappliers.watcher.openstack.org
            version: v1beta1
          - description: WatcherDecisionEngine is the Schema for the watcherdecisionengines
              API
            displayName: Watcher Decision Engine
            kind: WatcherDecisionEngine
            name: watcherdecisionengines.watcher.openstack.org
            version: v1beta1
          - description: Watcher is the Schema for the watchers API
            displayName: Watcher
            kind: Watcher
            name: watchers.watcher.openstack.org
            version: v1beta1
        description: The Watcher Operator project
        displayName: Watcher Operator (Technology Preview)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        links:
        - name: Watcher Operator
          url: https://github.com/openstack-k8s-operators/watcher-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/watcher-rhel9-operator@sha256:63c341a3e20a522c601098502a8fd8f1164d23c773ae8f55782c43ff4b6c8c18
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:eed8234b0fe47083ab941f513709f778bfcb80e32747b71a65fd5f62328a5ca0
        - registry.redhat.io/rhoso/openstack-watcher-api-rhel9@sha256:e97369e3a47bb882f1920c6af3f6345a842af45c58f0f5c5df9ac2933cf3098a
        - registry.redhat.io/rhoso/openstack-watcher-decision-engine-rhel9@sha256:c759d66cf0986c0036be91c008cbd059b4412c94160ed82dd12cd4e00d0685a9
        - registry.redhat.io/rhoso/openstack-watcher-applier-rhel9@sha256:c92a82f901613991926367bb8aab4f875c27a3e4a5282d92847a8463d98cd9e3
        version: 1.0.14
      entries:
      - name: watcher-operator.v1.0.14
        version: 1.0.14
      - name: watcher-operator.v1.0.13
        version: 1.0.13
      - name: watcher-operator.v1.0.12
        version: 1.0.12
      - name: watcher-operator.v1.0.11
        version: 1.0.11
      - name: watcher-operator.v1.0.10
        version: 1.0.10
      - name: watcher-operator.v1.0.9
        version: 1.0.9
      - name: watcher-operator.v1.0.8
        version: 1.0.8
      - name: watcher-operator.v1.0.7
        version: 1.0.7
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: watcher-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NdmSpc project
      provider-url: ""
    name: ndmspc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ndmspc-operator.v0.20250209.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.ndmspc.io/v1alpha1",
                "kind": "NdmSpcConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ndmspc-operator",
                    "app.kubernetes.io/instance": "ndmspcconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ndmspcconfig",
                    "app.kubernetes.io/part-of": "ndmspc-operator"
                  },
                  "name": "ndmspc-sample"
                },
                "spec": {
                  "executor": {
                    "image": "registry.gitlab.com/ndmspc/api:v0.20240215.2",
                    "name": "executor"
                  },
                  "executors": 1,
                  "ingressclasstype": "nginx",
                  "ingressenabled": true,
                  "ingresshostdomain": "127.0.0.1.sslip.io",
                  "ingressnostname": "ndmspc",
                  "ingressssltraefik": false,
                  "naf": {
                    "image": "registry.gitlab.com/ndmspc/naf:v0.20250129.0",
                    "name": "naf"
                  },
                  "ndmvr": {
                    "image": "registry.gitlab.com/ndmspc/ndmvr/shared:v0.20230320.0",
                    "name": "ndmvr"
                  },
                  "salsa": {
                    "image": "registry.gitlab.com/ndmspc/user/ndmspc:v0.20240412.0"
                  },
                  "storage": {
                    "image": "registry.gitlab.com/ndmspc/storage-server:v0.20250205.0",
                    "mongodb": {
                      "image": "docker.io/library/mongo:latest",
                      "password": "admin",
                      "port": 27017,
                      "username": "admin"
                    },
                    "name": "storage",
                    "port": 8080
                  },
                  "webapp": {
                    "image": "registry.gitlab.com/ndmspc/react-ndmspc:v0.20240412.0",
                    "name": ""
                  },
                  "webappcontainerport": 8080,
                  "webapps": 1,
                  "worker": {
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "1Gi"
                      }
                    }
                  },
                  "workers": 2,
                  "workerslots": 1,
                  "zmq2ws": {
                    "image": "registry.gitlab.com/ndmspc/zmq2ws:v1.20231012.0",
                    "name": "zmq2ws"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: registry.gitlab.com/ndmspc/ndmspc-operator:0.20250209.0
          createdAt: "2025-02-09T12:49:06Z"
          description: The NdmSpc operator automates the deployment and management
            of Ndmspc suite on Kubernetes cluster.
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://gitlab.com/ndmspc/ndmspc-operator
          support: NDMSPC Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NdmSpc configuration
            displayName: NdmSpc Config
            kind: NdmSpcConfig
            name: ndmspcconfigs.apps.ndmspc.io
            version: v1alpha1
        description: "The NdmSpc operator automates the deployment and management
          of Ndmspc suite on Kubernetes cluster. This operator allows user to deploy
          SALSA cluster with ZMQ2WS and job executor service and specify custom webapp
          page.\n\n### Capabilities\n\n- SALSA cluster\n- ZMQ2WS api service\n- Job
          executor service \n- Custom WebApp page\n"
        displayName: NdmSpc operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ndmspc
        links:
        - name: Ndmspc Operator
          url: https://ndmspc.gitlab.io/docs
        maintainers:
        - email: ndmspc@gmail.com
          name: NdmSpc Support
        maturity: alpha
        provider:
          name: NdmSpc project
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - registry.gitlab.com/ndmspc/ndmspc-operator:0.20250209.0
        version: 0.20250209.0
      entries:
      - name: ndmspc-operator.v0.20250209.0
        version: 0.20250209.0
      - name: ndmspc-operator.v0.20250131.0
        version: 0.20250131.0
      - name: ndmspc-operator.v0.20240412.0
        version: 0.20240412.0
      - name: ndmspc-operator.v0.11.6
        version: 0.11.6
      - name: ndmspc-operator.v0.11.4
        version: 0.11.4
      - name: ndmspc-operator.v0.11.3
        version: 0.11.3
      - name: ndmspc-operator.v0.11.2
        version: 0.11.2
      - name: ndmspc-operator.v0.11.1
        version: 0.11.1
      - name: ndmspc-operator.v0.10.2
        version: 0.10.2
      - name: ndmspc-operator.v0.10.1
        version: 0.10.1
      - name: ndmspc-operator.v0.10.0
        version: 0.10.0
      - name: ndmspc-operator.v0.9.0
        version: 0.9.0
      - name: ndmspc-operator.v0.7.4
        version: 0.7.4
      - name: ndmspc-operator.v0.7.1
        version: 0.7.1
      - name: ndmspc-operator.v0.6.0
        version: 0.6.0
      - name: ndmspc-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: ndmspc-operator
    provider:
      name: NdmSpc project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: VMware Tanzu
      provider-url: ""
    name: rabbitmq-cluster-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-cluster-operator.v2.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "rabbitmq.com/v1beta1",
                "kind": "RabbitmqCluster",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "override": {
                    "statefulset": {
                      "spec": {
                        "template": {
                          "spec": {
                            "containers": [],
                            "securityContext": {}
                          }
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/rabbitmqoperator/cluster-operator:2.16.0
          createdAt: 2025-08-12T09:34:33
          repository: https://github.com/rabbitmq/cluster-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Rabbitmq Cluster
            displayName: RabbitmqCluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: "## About the managed application\nRabbitMQ is an open-source
          message broker that originally implemented the\nAdvanced Message Queuing
          Protocol (AMQP). Nowadays features a plug-in\narchitecture supporting protocols
          such as Streaming Text Oriented Messaging\nProtocol (STOMP), MQ Telemetry
          Transport (MQTT), and others. RabbitMQ\nsupports native streaming and AMQP
          1.0.\n\nVisit [rabbitmq.com](rabbitmq.com) for more information.\n\n## About
          this Operator\nRabbitMQ Cluster Kubernetes Operator is developed and maintened
          by the VMware \nRabbitMQ team and provides a consistent and easy way to
          deploy RabbitMQ\nclusters to Kubernetes and run them, including \"day two\"
          (continuous)\noperations. RabbitMQ clusters deployed using the Operator
          can be used by\napplications running on Kubernetes or outside of Kubernetes.\n\nRead
          [RabbitMQ operator docs](https://www.rabbitmq.com/kubernetes/operator/operator-overview)\nfor
          more information.\n"
        displayName: RabbitMQ-cluster-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        - messaging
        - amqp
        - amqp10
        - stomp
        - mqtt
        - mqtt5
        - queues
        - streaming
        links:
        - name: Documentation
          url: https://www.rabbitmq.com/kubernetes/operator/operator-overview.html
        - name: GitHub
          url: https://github.com/rabbitmq/cluster-operator
        maintainers:
        - email: rabbitmq-users@googlegroups.com
          name: VMware Tanzu
        maturity: stable
        minKubeVersion: 1.26.0
        nativeApis:
        - group: apps
          kind: Deployment
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: ReplicaSet
          version: v1
        - group: ""
          kind: Pod
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: ConfigMap
          version: v1
        provider:
          name: VMware Tanzu
        relatedImages:
        - quay.io/rabbitmqoperator/cluster-operator:2.16.0
        version: 2.16.0
      entries:
      - name: rabbitmq-cluster-operator.v2.16.0
        version: 2.16.0
      - name: rabbitmq-cluster-operator.v2.15.0
        version: 2.15.0
      - name: rabbitmq-cluster-operator.v2.14.0
        version: 2.14.0
      - name: rabbitmq-cluster-operator.v2.13.0
        version: 2.13.0
      - name: rabbitmq-cluster-operator.v2.12.1
        version: 2.12.1
      - name: rabbitmq-cluster-operator.v2.10.0
        version: 2.10.0
      - name: rabbitmq-cluster-operator.v2.9.0
        version: 2.9.0
      - name: rabbitmq-cluster-operator.v2.8.0
        version: 2.8.0
      - name: rabbitmq-cluster-operator.v2.7.0
        version: 2.7.0
      - name: rabbitmq-cluster-operator.v2.6.0
        version: 2.6.0
      - name: rabbitmq-cluster-operator.v2.5.0
        version: 2.5.0
      - name: rabbitmq-cluster-operator.v2.4.0
        version: 2.4.0
      - name: rabbitmq-cluster-operator.v2.3.0
        version: 2.3.0
      - name: rabbitmq-cluster-operator.v2.2.0
        version: 2.2.0
      - name: rabbitmq-cluster-operator.v2.1.0
        version: 2.1.0
      - name: rabbitmq-cluster-operator.v2.0.0
        version: 2.0.0
      - name: rabbitmq-cluster-operator.v1.14.0
        version: 1.14.0
      name: stable
    defaultChannel: stable
    packageName: rabbitmq-cluster-operator
    provider:
      name: VMware Tanzu
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: ""
    name: ibm-metrics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-metrics-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MarketplaceConfig",
                "metadata": {
                  "name": "marketplaceconfig"
                },
                "spec": {
                  "features": {
                    "deployment": true,
                    "meterDefinitionCatalogServer": false,
                    "registration": true
                  },
                  "installIBMCatalogSource": true,
                  "isDisconnected": false,
                  "license": {
                    "accept": false
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterBase",
                "metadata": {
                  "name": "meterbase"
                },
                "spec": {
                  "enabled": true,
                  "prometheus": {
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "1",
                        "memory": "1G"
                      }
                    },
                    "storage": {
                      "size": "20Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterReport",
                "metadata": {
                  "name": "example-meterreport"
                },
                "spec": {
                  "endTime": "2020-08-05T00:00:00Z",
                  "prometheusService": {
                    "name": "rhm-prometheus-meterbase",
                    "namespace": "openshift-redhat-marketplace",
                    "targetPort": "rbac"
                  },
                  "startTime": "2020-08-04T00:00:00Z"
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "example-meterdefinition-4"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "meterdefinition-sample"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
          createdAt: "2025-08-28T14:18:55Z"
          description: The IBM Metrics Operator provides workload metering and reporting
            for IBM and Red Hat Marketplace customers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
          - description: MeterBase is the resource that sets up Metering for Red Hat
              Marketplace. This is an internal resource not meant to be modified directly.
            displayName: Meter Base
            kind: MeterBase
            name: meterbases.marketplace.redhat.com
            version: v1alpha1
          - description: MeterDefinition defines the meter workloads used to enable
              pay for use billing.
            displayName: Meter Definition
            kind: MeterDefinition
            name: meterdefinitions.marketplace.redhat.com
            version: v1beta1
          - kind: MeterReport
            name: meterreports.marketplace.redhat.com
            version: v1alpha1
        description: "The IBM Metrics Operator provides workload metering and reporting
          for IBM Software Central and Red Hat Marketplace customers.\n### **Important
          Note**\nA set of instructions for onboarding is provided here. For more
          detailed onboarding instructions or information about what is installed
          please visit [swc.saas.ibm.com](https://swc.saas.ibm.com).\n\nUsage metrics
          may be monitored through [https://swc.saas.ibm.com](https://swc.saas.ibm.com)
          with only IBM Metrics Operator and a Red Hat Marketplace account.\n\n###
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n### Prerequisites\n1. Installations are required
          to [enable monitoring for user-defined projects](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          as the Prometheus provider.\n2. Edit the cluster-monitoring-config ConfigMap
          object:\n\n   ```sh\n   $ oc -n openshift-monitoring edit configmap cluster-monitoring-config\n
          \   ```\n\n3. Add enableUserWorkload: true under data/config.yaml:\n  \n
          \   ```sh\n    apiVersion: v1\n    kind: ConfigMap\n    metadata:\n      name:
          cluster-monitoring-config\n        namespace: openshift-monitoring\n    data:\n
          \     config.yaml: |\n        enableUserWorkload: true\n    ```\n\n4. Configure
          the user-workload-monitoring-config ConfigMap object:\n\n    ```sh\n    $
          oc -n openshift-user-workload-monitoring edit configmap user-workload-monitoring-config\n
          \   ```\n\n5. Configure a minimum retention time of 168h and minimum storage
          capacity of 40Gi\n  \n    ```sh\n    apiVersion: v1\n    kind: ConfigMap\n
          \   metadata:\n      name: user-workload-monitoring-config\n      namespace:
          openshift-user-workload-monitoring\n\n    data:\n      config.yaml: |\n
          \       prometheus:\n          retention: 168h\n          volumeClaimTemplate:\n
          \           spec:\n              resources:\n                requests:\n
          \                 storage: 40Gi\n    ```\n\n### Resources Required\n\nMinimum
          system resources required:\n\n| Operator                | Memory (MB) |
          CPU (cores) | Disk (GB) | Nodes |\n| ----------------------- | -----------
          | ----------- | --------- | ----- |\n| **Metrics**   |        750  |     0.25
          \   | 3x1       |    3  |\n| **Deployment** |        250  |     0.25    |
          -         |    1  |\n\n| Prometheus Provider  | Memory (GB) | CPU (cores)
          | Disk (GB) | Nodes |\n| --------- | ----------- | ----------- | ---------
          | ----- |\n| **[Openshift User Workload Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)**
          |          1  |     0.1       | 2x40        |   2    |\n\nMultiple nodes
          are required to provide pod scheduling for high availability for IBM Metrics
          Operator Data Service and Prometheus.\n\nThe IBM Metrics Operator automatically
          creates 3 x 1Gi PersistentVolumeClaims to store reports as part of the data
          service, with _ReadWriteOnce_ access mode. Te PersistentVolumeClaims are
          automatically created by the ibm-metrics-operator after creating a `redhat-marketplace-pull-secret`
          and accepting the license in `marketplaceconfig`.\n\n| NAME                                |
          CAPACITY | ACCESS MODES |\n| ----------------------------------- | --------
          | ------------ |\n| rhm-data-service-rhm-data-service-0 | 1Gi | RWO |\n|
          rhm-data-service-rhm-data-service-1 | 1Gi | RWO |\n| rhm-data-service-rhm-data-service-2
          | 1Gi | RWO |\n\n### Supported Storage Providers\n\n- OpenShift Container
          Storage / OpenShift Data Foundation version 4.x, from version 4.2 or higher\n-
          IBM Cloud Block storage and IBM Cloud File storage\n- IBM Storage Suite
          for IBM Cloud Paks:\n  - File storage from IBM Spectrum Fusion/Scale \n
          \ - Block storage from IBM Spectrum Virtualize, FlashSystem or DS8K\n- Portworx
          Storage, version 2.5.5 or above\n- Amazon Elastic File Storage\n\n### Access
          Modes required\n\n - ReadWriteOnce (RWO)\n\n### Provisioning Options supported\n\nChoose
          one of the following options to provision storage for the ibm-metrics-operator
          data-service\n\n#### Dynamic provisioning using a default StorageClass\n
          \  - A StorageClass is defined with a `metadata.annotations: storageclass.kubernetes.io/is-default-class:
          \"true\"`\n   - PersistentVolumes will be provisioned automatically for
          the generated PersistentVolumeClaims\n--- \n#### Manually create each PersistentVolumeClaim
          with a specific StorageClass\n   - Must be performed before creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`. Otherwise, the automatically
          generated PersistentVolumeClaims are immutable.\n```\napiVersion: v1\nkind:
          PersistentVolumeClaim\nmetadata:\n  labels:\n    app: rhm-data-service\n
          \ name: rhm-data-service-rhm-data-service-0\n  namespace: ibm-software-central\nspec:\n
          \ storageClassName: rook-cephfs\n  accessModes:\n  - ReadWriteOnce\n  resources:\n
          \   requests:\n      storage: 1Gi\n```\n---\n#### Manually provision each
          PersistentVolume for the generated PersistentVolumeClaims with a specific
          StorageClass\n  - May be performed before or after creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`.  \n```\napiVersion: v1\nkind:
          PersistentVolume\nmetadata:\n  name: rhm-data-service-rhm-data-service-0\nspec:\n
          \ csi:\n    driver: rook-ceph.cephfs.csi.ceph.com\n    volumeHandle: rhm-data-service-rhm-data-service-0\n
          \ capacity:\n    storage: 1Gi\n  accessModes:\n    - ReadWriteOnce\n  persistentVolumeReclaimPolicy:
          Delete\n  storageClassName: rook-cephfs\n  volumeMode: Filesystem\n  claimRef:\n
          \   kind: PersistentVolumeClaim\n    namespace: ibm-software-central\n    name:
          rhm-data-service-rhm-data-service-0\n```\n\n### Installation\n1. Create
          or get your [pull secret](https://swc.saas.ibm.com/en-us/documentation/clusters#get-pull-secret).\n2.
          Install the IBM Metrics Operator\n3. Create a Kubernetes secret in the installed
          namespace with the name `redhat-marketplace-pull-secret` and key `PULL_SECRET`
          with the value of the pull secret.\n    ```sh\n    # Replace ${PULL_SECRET}
          with your pull secret\n    oc create secret generic redhat-marketplace-pull-secret
          -n ibm-software-central --from-literal=PULL_SECRET=${PULL_SECRET}\n    ```\n4.
          Use of the platform is governed by the:\n\n    [IBM Cloud Services Agreement](https://www.ibm.com/support/customer/csol/terms/?id=Z126-6304_WS&_ga=2.116312197.2046730452.1684328846-812467790.1684328846)
          (or other base agreement between you and IBM such as a [Passport Advantage
          Agreement](https://www.ibm.com/software/passportadvantage/pa_agreements.html?_ga=2.116312197.2046730452.1684328846-812467790.1684328846))
          and the [Service Description for the Red Hat Marketplace](https://www.ibm.com/support/customer/csol/terms/?id=i126-8719&_ga=2.83289621.2046730452.1684328846-812467790.1684328846).\n
          \   \n5. Update MarketplaceConfig to accept the license.\n    ```\n    oc
          patch marketplaceconfig marketplaceconfig -n ibm-software-central --type='merge'
          -p '{\"spec\": {\"license\": {\"accept\": true}}}'\n    ```\n6. Install
          the pull secret as a global pull secret on the cluster.\n\n    These steps
          require `oc`, `jq`, and `base64` to be available on your machine.\n\n    ```sh\n
          \   # Create the docker pull secret file using your pull secret.\n    #
          Store it in a file called entitledregistryconfigjson.\n    oc create secret
          docker-registry entitled-registry --docker-server=registry.marketplace.redhat.com
          --docker-username \"cp\" --docker-password \"${PULL_SECRET}\" --dry-run=client
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > entitledregistryconfigjson\n
          \   # Get the current global secrets on the cluster and store it as a file
          named dockerconfigjson\n    oc get secret pull-secret -n openshift-config
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > dockerconfigjson\n
          \   # Merge the two dockerconfigs together into a file called dockerconfigjson-merged.\n
          \   jq -s '.[0] * .[1]' dockerconfigjson entitledregistryconfigjson > dockerconfigjson-merged\n
          \   # Set the cluster's dockerconfig file to the new merged version.\n    oc
          set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson=dockerconfigjson-merged\n
          \   ```\n  7. (Optional) Your IBM product may have installed IBM Metrics
          Operator using an `ibm-entitlement-key` for authentication to IBM Software
          Central. To authorize the automatic creation of an IBM Software Central
          account for the IBM account owning the `ibm-entitlement-key`\n      ```\n
          \     oc patch marketplaceconfig marketplaceconfig -n ibm-software-central
          --type='merge' -p '{\"spec\": {\"authorizeAccountCreation\": true}}'\n      ```\n\n###
          Why is a global pull secret required?\nIn order to successfully install
          the products hosted by the container image registry, you will need to make
          the pull secret available across the cluster. This can be achieved by applying
          the pull as a [global pull secret](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html#images-update-global-pull-secret_using-image-pull-secrets).
          For alternative approaches, please see the official OpenShift [documentation](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html).\n\n\n###
          SecurityContextConstraints requirements\n\nThe Operators and their components
          support running under the OpenShift Container Platform default restricted
          and restricted-v2 security context constraints.\n\n### Installation Namespace
          and ClusterRoleBinding requirements\n\nThe IBM Metrics Operator components
          require specific ClusterRoleBindings.\n- The metric-state component requires
          a ClusterRoleBinding for the the `view` ClusterRole.\n  - Ability to read
          non-sensitive CustomResources. \n  - The `view` ClusterRole dynamically
          adds new CustomResourceDefinitions.\n- The operator & reporter component
          requires a ClusterRoleBinding for the the `cluster-monitoring-view` ClusterRole.\n
          \ - Ability to view Prometheus metrics. \n  - The underlying ClusterRole
          RBAC often updated between OpenShift versions.\n\nDue to limitations of
          Operator Lifecycle Manager (OLM), this ClusterRoleBinding can not be provided
          dynamically for arbitrary installation target namespaces.\n\nA static ClusterRoleBinding
          is included for installation to the default namespace of `ibm-software-central`,
          and namespaces `openshift-redhat-marketplace`,  `redhat-marketplace`, `ibm-common-services`.\n\nTo
          create the ClusterRoleBindings for installation to an alternate namespace\n```\noc
          project INSTALL-NAMESPACE\noc adm policy add-cluster-role-to-user view -z
          ibm-metrics-operator-metric-state\noc adm policy add-cluster-role-to-user
          cluster-monitoring-view -z ibm-metrics-operator-controller-manager,ibm-metrics-operator-reporter\n```\n\n###
          Metric State scoping requirements\nThe metric-state Deployment obtains `get/list/watch`
          access to metered resources via the `view` ClusterRole. For operators deployed
          using Operator Lifecycle Manager (OLM), permissions are added to `clusterrole/view`
          dynamically via a generated and annotated `-view` ClusterRole. If you wish
          to meter an operator, and its Custom Resource Definitions (CRDs) are not
          deployed through OLM, one of two options are required\n1. Add the following
          label to a clusterrole that has get/list/watch access to your CRD: `rbac.authorization.k8s.io/aggregate-to-view:
          \"true\"`, thereby dynamically adding it to `clusterrole/view`\n2. Create
          a ClusterRole that has get/list/watch access to your CRD, and create a ClusterRoleBinding
          for the metric-state ServiceAccount\n\nAttempting to meter a resource with
          a MeterDefinition without the required permissions will log an `AccessDeniedError`
          in metric-state.\n\n### Disaster Recovery\n\nTo plan for disaster recovery,
          note the PhysicalVolumeClaims `rhm-data-service-rhm-data-service-N`. \n-
          In connected environments, MeterReport data upload attempts occur hourly,
          and are then removed from data-service. There is a low risk of losing much
          unreported data.\n- In an airgap environment, MeterReport data must be pulled
          from data-service and uploaded manually using [datactl](https://github.com/redhat-marketplace/datactl/).
          To prevent data loss in a disaster scenario, the data-service volumes should
          be considered in a recovery plan.\n\n### Subscription Config\n\nIt is possible
          to [configure](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md)
          how OLM deploys an Operator via the `config` field in the [Subscription](https://github.com/operator-framework/olm-book/blob/master/docs/subscriptions.md)
          object.\n\nThe IBM Metrics Operator will also read the `config` and append
          it to the operands. The primary use case is to control scheduling using
          [Tolerations](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/)
          and [NodeSelectors](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/).\n\nA
          limitation is that the `config` elements are only appended to the operands.
          The elements in the operands are not removed if the `config` is removed
          from the `Subscripion`. The operand must be modified manually, or deleted
          and recreated by the controller.\n\n### Cluster permission requirements\n\n|API
          group             |Resources              |Verbs                                     |\n|----------------------|-----------------------|------------------------------------------|\n|apps
          \                 |statefulsets           |get;list;watch                            |\n|authentication.k8s.io
          |tokenreviews           |create                                    |\n|authorization.k8s.io
          \ |subjectaccessreviews   |create                                    |\n|config.openshift.io
          \  |clusterversions        |get;list;watch                            |\n|marketplace.redhat.com|meterdefinitions
          \      |get;list;watch;create;update;patch;delete |\n|operators.coreos.com
          \ |clusterserviceversions |get;list;watch                            |\n|operators.coreos.com
          \ |subscriptions          |get;list;watch                            |\n|operators.coreos.com
          \ |operatorgroups         |get;list                                  |\n|storage.k8s.io
          \       |storageclasses         |get;list;watch                            |\n|
          \                     |namespaces             |get;list;watch                            |\n|
          \                     |configmaps             |get;list;watch                            |\n|
          \                     |services               |get;list;watch                            |\n|
          \                     |**clusterrole/view**   |                                          |\n\n\n###
          Documentation\nYou can find our documentation [here.](https://swc.saas.ibm.com/en-us/documentation/)\n\n###
          Getting help\nIf you encounter any issues while using the operators, you
          can create an issue on our [Github\nrepo](https://github.com/redhat-marketplace/redhat-marketplace-operator)
          for bugs, enhancements, or other requests. You can also visit our main page
          and\nreview our [support](https://swc.saas.ibm.com/en-us/support) and [documentation](https://swc.saas.ibm.com/en-us/documentation/).\n\n###
          Readme\nYou can find our readme [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/README.md)\n\n###
          License information\nYou can find our license information [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/LICENSE)\n"
        displayName: IBM Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        - buy
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/redhat-marketplace-authcheck@sha256:13a81b8aca79ae8c89ee24dcc30f3322a4cfdedf929e7bb984295c1e7e631fa9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/redhat-marketplace-data-service@sha256:30fd8d2996b921cfce51f697ee4490a6d9a836021952429b877336f5edf2749d
        - quay.io/rh-marketplace/redhat-marketplace-reporter@sha256:2bd098f057ca07f2b14be9b60ea6ad8ca6b0cae1d226b1ca3b27e23d04ccdf20
        - quay.io/rh-marketplace/redhat-marketplace-metric-state@sha256:e4014c9a546c10e84b13a09db6a18431ee9489ad1296808635e48d9ef4963cf1
        - quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
        version: 2.20.0
      entries:
      - name: ibm-metrics-operator.v2.20.0
        version: 2.20.0
      name: beta
    - currentCSV: ibm-metrics-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MarketplaceConfig",
                "metadata": {
                  "name": "marketplaceconfig"
                },
                "spec": {
                  "features": {
                    "deployment": true,
                    "meterDefinitionCatalogServer": false,
                    "registration": true
                  },
                  "installIBMCatalogSource": true,
                  "isDisconnected": false,
                  "license": {
                    "accept": false
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterBase",
                "metadata": {
                  "name": "meterbase"
                },
                "spec": {
                  "enabled": true,
                  "prometheus": {
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "1",
                        "memory": "1G"
                      }
                    },
                    "storage": {
                      "size": "20Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterReport",
                "metadata": {
                  "name": "example-meterreport"
                },
                "spec": {
                  "endTime": "2020-08-05T00:00:00Z",
                  "prometheusService": {
                    "name": "rhm-prometheus-meterbase",
                    "namespace": "openshift-redhat-marketplace",
                    "targetPort": "rbac"
                  },
                  "startTime": "2020-08-04T00:00:00Z"
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "example-meterdefinition-4"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "meterdefinition-sample"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
          createdAt: "2025-08-28T14:18:55Z"
          description: The IBM Metrics Operator provides workload metering and reporting
            for IBM and Red Hat Marketplace customers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
          - description: MeterBase is the resource that sets up Metering for Red Hat
              Marketplace. This is an internal resource not meant to be modified directly.
            displayName: Meter Base
            kind: MeterBase
            name: meterbases.marketplace.redhat.com
            version: v1alpha1
          - description: MeterDefinition defines the meter workloads used to enable
              pay for use billing.
            displayName: Meter Definition
            kind: MeterDefinition
            name: meterdefinitions.marketplace.redhat.com
            version: v1beta1
          - kind: MeterReport
            name: meterreports.marketplace.redhat.com
            version: v1alpha1
        description: "The IBM Metrics Operator provides workload metering and reporting
          for IBM Software Central and Red Hat Marketplace customers.\n### **Important
          Note**\nA set of instructions for onboarding is provided here. For more
          detailed onboarding instructions or information about what is installed
          please visit [swc.saas.ibm.com](https://swc.saas.ibm.com).\n\nUsage metrics
          may be monitored through [https://swc.saas.ibm.com](https://swc.saas.ibm.com)
          with only IBM Metrics Operator and a Red Hat Marketplace account.\n\n###
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n### Prerequisites\n1. Installations are required
          to [enable monitoring for user-defined projects](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          as the Prometheus provider.\n2. Edit the cluster-monitoring-config ConfigMap
          object:\n\n   ```sh\n   $ oc -n openshift-monitoring edit configmap cluster-monitoring-config\n
          \   ```\n\n3. Add enableUserWorkload: true under data/config.yaml:\n  \n
          \   ```sh\n    apiVersion: v1\n    kind: ConfigMap\n    metadata:\n      name:
          cluster-monitoring-config\n        namespace: openshift-monitoring\n    data:\n
          \     config.yaml: |\n        enableUserWorkload: true\n    ```\n\n4. Configure
          the user-workload-monitoring-config ConfigMap object:\n\n    ```sh\n    $
          oc -n openshift-user-workload-monitoring edit configmap user-workload-monitoring-config\n
          \   ```\n\n5. Configure a minimum retention time of 168h and minimum storage
          capacity of 40Gi\n  \n    ```sh\n    apiVersion: v1\n    kind: ConfigMap\n
          \   metadata:\n      name: user-workload-monitoring-config\n      namespace:
          openshift-user-workload-monitoring\n\n    data:\n      config.yaml: |\n
          \       prometheus:\n          retention: 168h\n          volumeClaimTemplate:\n
          \           spec:\n              resources:\n                requests:\n
          \                 storage: 40Gi\n    ```\n\n### Resources Required\n\nMinimum
          system resources required:\n\n| Operator                | Memory (MB) |
          CPU (cores) | Disk (GB) | Nodes |\n| ----------------------- | -----------
          | ----------- | --------- | ----- |\n| **Metrics**   |        750  |     0.25
          \   | 3x1       |    3  |\n| **Deployment** |        250  |     0.25    |
          -         |    1  |\n\n| Prometheus Provider  | Memory (GB) | CPU (cores)
          | Disk (GB) | Nodes |\n| --------- | ----------- | ----------- | ---------
          | ----- |\n| **[Openshift User Workload Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)**
          |          1  |     0.1       | 2x40        |   2    |\n\nMultiple nodes
          are required to provide pod scheduling for high availability for IBM Metrics
          Operator Data Service and Prometheus.\n\nThe IBM Metrics Operator automatically
          creates 3 x 1Gi PersistentVolumeClaims to store reports as part of the data
          service, with _ReadWriteOnce_ access mode. Te PersistentVolumeClaims are
          automatically created by the ibm-metrics-operator after creating a `redhat-marketplace-pull-secret`
          and accepting the license in `marketplaceconfig`.\n\n| NAME                                |
          CAPACITY | ACCESS MODES |\n| ----------------------------------- | --------
          | ------------ |\n| rhm-data-service-rhm-data-service-0 | 1Gi | RWO |\n|
          rhm-data-service-rhm-data-service-1 | 1Gi | RWO |\n| rhm-data-service-rhm-data-service-2
          | 1Gi | RWO |\n\n### Supported Storage Providers\n\n- OpenShift Container
          Storage / OpenShift Data Foundation version 4.x, from version 4.2 or higher\n-
          IBM Cloud Block storage and IBM Cloud File storage\n- IBM Storage Suite
          for IBM Cloud Paks:\n  - File storage from IBM Spectrum Fusion/Scale \n
          \ - Block storage from IBM Spectrum Virtualize, FlashSystem or DS8K\n- Portworx
          Storage, version 2.5.5 or above\n- Amazon Elastic File Storage\n\n### Access
          Modes required\n\n - ReadWriteOnce (RWO)\n\n### Provisioning Options supported\n\nChoose
          one of the following options to provision storage for the ibm-metrics-operator
          data-service\n\n#### Dynamic provisioning using a default StorageClass\n
          \  - A StorageClass is defined with a `metadata.annotations: storageclass.kubernetes.io/is-default-class:
          \"true\"`\n   - PersistentVolumes will be provisioned automatically for
          the generated PersistentVolumeClaims\n--- \n#### Manually create each PersistentVolumeClaim
          with a specific StorageClass\n   - Must be performed before creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`. Otherwise, the automatically
          generated PersistentVolumeClaims are immutable.\n```\napiVersion: v1\nkind:
          PersistentVolumeClaim\nmetadata:\n  labels:\n    app: rhm-data-service\n
          \ name: rhm-data-service-rhm-data-service-0\n  namespace: ibm-software-central\nspec:\n
          \ storageClassName: rook-cephfs\n  accessModes:\n  - ReadWriteOnce\n  resources:\n
          \   requests:\n      storage: 1Gi\n```\n---\n#### Manually provision each
          PersistentVolume for the generated PersistentVolumeClaims with a specific
          StorageClass\n  - May be performed before or after creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`.  \n```\napiVersion: v1\nkind:
          PersistentVolume\nmetadata:\n  name: rhm-data-service-rhm-data-service-0\nspec:\n
          \ csi:\n    driver: rook-ceph.cephfs.csi.ceph.com\n    volumeHandle: rhm-data-service-rhm-data-service-0\n
          \ capacity:\n    storage: 1Gi\n  accessModes:\n    - ReadWriteOnce\n  persistentVolumeReclaimPolicy:
          Delete\n  storageClassName: rook-cephfs\n  volumeMode: Filesystem\n  claimRef:\n
          \   kind: PersistentVolumeClaim\n    namespace: ibm-software-central\n    name:
          rhm-data-service-rhm-data-service-0\n```\n\n### Installation\n1. Create
          or get your [pull secret](https://swc.saas.ibm.com/en-us/documentation/clusters#get-pull-secret).\n2.
          Install the IBM Metrics Operator\n3. Create a Kubernetes secret in the installed
          namespace with the name `redhat-marketplace-pull-secret` and key `PULL_SECRET`
          with the value of the pull secret.\n    ```sh\n    # Replace ${PULL_SECRET}
          with your pull secret\n    oc create secret generic redhat-marketplace-pull-secret
          -n ibm-software-central --from-literal=PULL_SECRET=${PULL_SECRET}\n    ```\n4.
          Use of the platform is governed by the:\n\n    [IBM Cloud Services Agreement](https://www.ibm.com/support/customer/csol/terms/?id=Z126-6304_WS&_ga=2.116312197.2046730452.1684328846-812467790.1684328846)
          (or other base agreement between you and IBM such as a [Passport Advantage
          Agreement](https://www.ibm.com/software/passportadvantage/pa_agreements.html?_ga=2.116312197.2046730452.1684328846-812467790.1684328846))
          and the [Service Description for the Red Hat Marketplace](https://www.ibm.com/support/customer/csol/terms/?id=i126-8719&_ga=2.83289621.2046730452.1684328846-812467790.1684328846).\n
          \   \n5. Update MarketplaceConfig to accept the license.\n    ```\n    oc
          patch marketplaceconfig marketplaceconfig -n ibm-software-central --type='merge'
          -p '{\"spec\": {\"license\": {\"accept\": true}}}'\n    ```\n6. Install
          the pull secret as a global pull secret on the cluster.\n\n    These steps
          require `oc`, `jq`, and `base64` to be available on your machine.\n\n    ```sh\n
          \   # Create the docker pull secret file using your pull secret.\n    #
          Store it in a file called entitledregistryconfigjson.\n    oc create secret
          docker-registry entitled-registry --docker-server=registry.marketplace.redhat.com
          --docker-username \"cp\" --docker-password \"${PULL_SECRET}\" --dry-run=client
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > entitledregistryconfigjson\n
          \   # Get the current global secrets on the cluster and store it as a file
          named dockerconfigjson\n    oc get secret pull-secret -n openshift-config
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > dockerconfigjson\n
          \   # Merge the two dockerconfigs together into a file called dockerconfigjson-merged.\n
          \   jq -s '.[0] * .[1]' dockerconfigjson entitledregistryconfigjson > dockerconfigjson-merged\n
          \   # Set the cluster's dockerconfig file to the new merged version.\n    oc
          set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson=dockerconfigjson-merged\n
          \   ```\n  7. (Optional) Your IBM product may have installed IBM Metrics
          Operator using an `ibm-entitlement-key` for authentication to IBM Software
          Central. To authorize the automatic creation of an IBM Software Central
          account for the IBM account owning the `ibm-entitlement-key`\n      ```\n
          \     oc patch marketplaceconfig marketplaceconfig -n ibm-software-central
          --type='merge' -p '{\"spec\": {\"authorizeAccountCreation\": true}}'\n      ```\n\n###
          Why is a global pull secret required?\nIn order to successfully install
          the products hosted by the container image registry, you will need to make
          the pull secret available across the cluster. This can be achieved by applying
          the pull as a [global pull secret](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html#images-update-global-pull-secret_using-image-pull-secrets).
          For alternative approaches, please see the official OpenShift [documentation](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html).\n\n\n###
          SecurityContextConstraints requirements\n\nThe Operators and their components
          support running under the OpenShift Container Platform default restricted
          and restricted-v2 security context constraints.\n\n### Installation Namespace
          and ClusterRoleBinding requirements\n\nThe IBM Metrics Operator components
          require specific ClusterRoleBindings.\n- The metric-state component requires
          a ClusterRoleBinding for the the `view` ClusterRole.\n  - Ability to read
          non-sensitive CustomResources. \n  - The `view` ClusterRole dynamically
          adds new CustomResourceDefinitions.\n- The operator & reporter component
          requires a ClusterRoleBinding for the the `cluster-monitoring-view` ClusterRole.\n
          \ - Ability to view Prometheus metrics. \n  - The underlying ClusterRole
          RBAC often updated between OpenShift versions.\n\nDue to limitations of
          Operator Lifecycle Manager (OLM), this ClusterRoleBinding can not be provided
          dynamically for arbitrary installation target namespaces.\n\nA static ClusterRoleBinding
          is included for installation to the default namespace of `ibm-software-central`,
          and namespaces `openshift-redhat-marketplace`,  `redhat-marketplace`, `ibm-common-services`.\n\nTo
          create the ClusterRoleBindings for installation to an alternate namespace\n```\noc
          project INSTALL-NAMESPACE\noc adm policy add-cluster-role-to-user view -z
          ibm-metrics-operator-metric-state\noc adm policy add-cluster-role-to-user
          cluster-monitoring-view -z ibm-metrics-operator-controller-manager,ibm-metrics-operator-reporter\n```\n\n###
          Metric State scoping requirements\nThe metric-state Deployment obtains `get/list/watch`
          access to metered resources via the `view` ClusterRole. For operators deployed
          using Operator Lifecycle Manager (OLM), permissions are added to `clusterrole/view`
          dynamically via a generated and annotated `-view` ClusterRole. If you wish
          to meter an operator, and its Custom Resource Definitions (CRDs) are not
          deployed through OLM, one of two options are required\n1. Add the following
          label to a clusterrole that has get/list/watch access to your CRD: `rbac.authorization.k8s.io/aggregate-to-view:
          \"true\"`, thereby dynamically adding it to `clusterrole/view`\n2. Create
          a ClusterRole that has get/list/watch access to your CRD, and create a ClusterRoleBinding
          for the metric-state ServiceAccount\n\nAttempting to meter a resource with
          a MeterDefinition without the required permissions will log an `AccessDeniedError`
          in metric-state.\n\n### Disaster Recovery\n\nTo plan for disaster recovery,
          note the PhysicalVolumeClaims `rhm-data-service-rhm-data-service-N`. \n-
          In connected environments, MeterReport data upload attempts occur hourly,
          and are then removed from data-service. There is a low risk of losing much
          unreported data.\n- In an airgap environment, MeterReport data must be pulled
          from data-service and uploaded manually using [datactl](https://github.com/redhat-marketplace/datactl/).
          To prevent data loss in a disaster scenario, the data-service volumes should
          be considered in a recovery plan.\n\n### Subscription Config\n\nIt is possible
          to [configure](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md)
          how OLM deploys an Operator via the `config` field in the [Subscription](https://github.com/operator-framework/olm-book/blob/master/docs/subscriptions.md)
          object.\n\nThe IBM Metrics Operator will also read the `config` and append
          it to the operands. The primary use case is to control scheduling using
          [Tolerations](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/)
          and [NodeSelectors](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/).\n\nA
          limitation is that the `config` elements are only appended to the operands.
          The elements in the operands are not removed if the `config` is removed
          from the `Subscripion`. The operand must be modified manually, or deleted
          and recreated by the controller.\n\n### Cluster permission requirements\n\n|API
          group             |Resources              |Verbs                                     |\n|----------------------|-----------------------|------------------------------------------|\n|apps
          \                 |statefulsets           |get;list;watch                            |\n|authentication.k8s.io
          |tokenreviews           |create                                    |\n|authorization.k8s.io
          \ |subjectaccessreviews   |create                                    |\n|config.openshift.io
          \  |clusterversions        |get;list;watch                            |\n|marketplace.redhat.com|meterdefinitions
          \      |get;list;watch;create;update;patch;delete |\n|operators.coreos.com
          \ |clusterserviceversions |get;list;watch                            |\n|operators.coreos.com
          \ |subscriptions          |get;list;watch                            |\n|operators.coreos.com
          \ |operatorgroups         |get;list                                  |\n|storage.k8s.io
          \       |storageclasses         |get;list;watch                            |\n|
          \                     |namespaces             |get;list;watch                            |\n|
          \                     |configmaps             |get;list;watch                            |\n|
          \                     |services               |get;list;watch                            |\n|
          \                     |**clusterrole/view**   |                                          |\n\n\n###
          Documentation\nYou can find our documentation [here.](https://swc.saas.ibm.com/en-us/documentation/)\n\n###
          Getting help\nIf you encounter any issues while using the operators, you
          can create an issue on our [Github\nrepo](https://github.com/redhat-marketplace/redhat-marketplace-operator)
          for bugs, enhancements, or other requests. You can also visit our main page
          and\nreview our [support](https://swc.saas.ibm.com/en-us/support) and [documentation](https://swc.saas.ibm.com/en-us/documentation/).\n\n###
          Readme\nYou can find our readme [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/README.md)\n\n###
          License information\nYou can find our license information [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/LICENSE)\n"
        displayName: IBM Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        - buy
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/redhat-marketplace-metric-state@sha256:e4014c9a546c10e84b13a09db6a18431ee9489ad1296808635e48d9ef4963cf1
        - quay.io/rh-marketplace/redhat-marketplace-authcheck@sha256:13a81b8aca79ae8c89ee24dcc30f3322a4cfdedf929e7bb984295c1e7e631fa9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/redhat-marketplace-data-service@sha256:30fd8d2996b921cfce51f697ee4490a6d9a836021952429b877336f5edf2749d
        - quay.io/rh-marketplace/redhat-marketplace-reporter@sha256:2bd098f057ca07f2b14be9b60ea6ad8ca6b0cae1d226b1ca3b27e23d04ccdf20
        - quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
        version: 2.20.0
      entries:
      - name: ibm-metrics-operator.v2.20.0
        version: 2.20.0
      name: stable
    defaultChannel: stable
    packageName: ibm-metrics-operator
    provider:
      name: Red Hat Marketplace
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-dr-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-dr-manager.v8.0.220
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com",
                  "resRecord": {
                    "HostName_1": "100.200.30.41",
                    "HostName_2": "100.200.30.43",
                    "abc": "HostName_2",
                    "www": "HostName_1",
                    "xyz": "HostName_3"
                  },
                  "stealthMasters": [
                    "1.2.3.4"
                  ],
                  "tsigKeyFile": "/Kvcslnxdns.com.+157+10641.private"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "dataReplicationPointer": "test-datareplication",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 10,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          olm.skipRange: <8.0.220
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
            - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:a98f9cc51f2c7baeec6dbb4933a42ae8ae20ff1a42c9d48a37174cf5211bbcf8
        version: 8.0.220
      entries:
      - name: infoscale-dr-manager.v8.0.220
        version: 8.0.220
      - name: infoscale-dr-manager.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-dr-manager.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          createdAt: "2024-06-17T10:14:13Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:c46cbf78631fc790c77b6c13c1d1d4b5ab35f49164f7c2e36c7ec59220cbdce2
        version: 8.0.330
      entries:
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-dr-manager.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
          createdAt: "2024-02-22T14:14:42Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721648-168743109)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721654-168743109)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
        version: 8.0.410
      entries:
      - name: infoscale-dr-manager.v8.0.410
        version: 8.0.410
      - name: infoscale-dr-manager.v8.0.409
        version: 8.0.409
      - name: infoscale-dr-manager.v8.0.400
        version: 8.0.400
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-dr-manager.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
          createdAt: "2024-02-22T14:14:42Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721648-168743109)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721654-168743109)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
        version: 8.0.410
      entries:
      - name: infoscale-dr-manager.v8.0.410
        version: 8.0.410
      - name: infoscale-dr-manager.v8.0.409
        version: 8.0.409
      - name: infoscale-dr-manager.v8.0.400
        version: 8.0.400
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-dr-manager.v8.0.100
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com",
                  "resRecord": {
                    "abc": "r7515-054-vm9",
                    "r7515-054-vm10": "10.221.85.83",
                    "r7515-054-vm8": "10.221.85.81",
                    "r7515-054-vm9": "10.221.85.82",
                    "www": "r7515-054-vm8",
                    "xyz": "r7515-054-vm10"
                  },
                  "stealthMasters": [
                    "1.2.3.4"
                  ],
                  "tsigKeyFile": "/Kvcslnxdns.com.+157+10641.private"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "currentPrimary": "Clus1",
                  "force": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "fail",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "dataReplicationPointer": "test-datareplication",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 10,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:eb999187a1f9a9ce5566089e24adf598b1c229d31c5e57ddcfe4439ba25e97dc
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
            - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membershrip
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membershrip
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponsding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:eb999187a1f9a9ce5566089e24adf598b1c229d31c5e57ddcfe4439ba25e97dc
        version: 8.0.100
      entries:
      - name: infoscale-dr-manager.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-dr-manager
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-s3control-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-s3control-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/s3control-controller:1.1.1
          createdAt: "2025-09-19T20:52:02Z"
          description: AWS S3 Control controller is a service controller for managing
            S3 Control resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessPoint represents the state of an AWS s3control AccessPoint
              resource.
            displayName: AccessPoint
            kind: AccessPoint
            name: accesspoints.s3control.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon S3 Control resources in AWS from within your Kubernetes cluster.

          **About Amazon S3 Control**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon S3 Control
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - s3control
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon S3 Control Developer Resources
          url: https://aws.amazon.com/S3 Control/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: s3control maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/s3control-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-s3control-controller.v1.1.1
        version: 1.1.1
      - name: ack-s3control-controller.v1.1.0
        version: 1.1.0
      - name: ack-s3control-controller.v1.0.11
        version: 1.0.11
      - name: ack-s3control-controller.v1.0.10
        version: 1.0.10
      name: alpha
    defaultChannel: alpha
    packageName: ack-s3control-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: sosivio
      provider-url: https://www.sosiv.io
    name: sosivio
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sosivio-operator.v1.5.1-97
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sosivio.sosiv.io/v1alpha1",
                "kind": "Sosivio",
                "metadata": {
                  "name": "sosivio"
                },
                "spec": {
                  "platform": "ocp4",
                  "useSha": false
                }
              },
              {
                "apiVersion": "sosivio.sosiv.io/v1alpha1",
                "kind": "Sosivio",
                "metadata": {
                  "name": "sosivio"
                },
                "spec": {
                  "platform": "ocp4",
                  "useSha": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Monitoring,Integration & Delivery,Logging & Tracing,Developer
            Tools
          com.redhat.openshift.versions: v4.1-v4.1
          containerImage: quay.io/sosivio/sosivio-operator:v1.5.1-97
          createdAt: "2021-10-28 01:01:01"
          description: Sosivio | Predictive Troubleshooting for Kubernetes
          operatorhub.io/ui-metadata-max-k8s-version: "1.25"
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Sosivio/operator
          support: Sosivio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Sosivio is the Schema for the sosivios API
            displayName: Sosivio
            kind: Sosivio
            name: sosivios.sosivio.sosiv.io
            version: v1alpha1
        description: "# Sosivio Operator\nThe Sosivio Operator for RedHat’s Openshift
          Container Platform allows users to quickly and easily deploy Sosivio’s Kubernetes
          Predictive Troubleshooting Platform on your cluster.\n\n# Sosivio Overview\n\nSosivio
          is the first and only non-intrusive and predictive autonomous cloud-native
          platform built specifically for Kubernetes from the ground up. Sosivio provides
          predictive troubleshooting tools, root cause determination and instant remediation
          for any failures in Kubernetes environments.\n\nPredict and Prevent!\n\nThe
          Operator will deploy and manage Sosivio on Kubernetes and OpenShift. Included
          are the following features;\n-   **Root Cause Analysis and Remediation**
          - predictive troubleshooting to prevent failures before they occur. Instantly
          understand where the root cause is and how to fix it.\n-   **Application
          Profiling** - analyze container resource consumption and recommendations
          for requests and limits.\n-   **Real-time Metrics** - in-depth, granular,
          and accurate real-time data from the process level inside containers (and
          pods) with failure correlation.\n-   **Cluster Health Checks** - proactive
          health checks for cluster infrastructure and applications running on the
          cluster.\n-   **Multi-Cluster View** - management, monitoring, and maintenance
          capabilities for all clusters from one location. \n-   **Kubernetes Components
          Alerting** - notifications sent through Slack, email and other 3rd party
          platforms.\n\n# Installation\nSosivio is completely **non-intrusive** and
          **does not require any persistent storage**. There is **no data offloading**
          outside of your Kubernetes cluster. Applying our Data Swirling methodology
          enables the product to perform at a **remarkably low resource footprint**
          and requires **no persistent volume configurations**.\n\n\nSosivio is installed
          under one namespace (labeled: \"sosivio\") which creates all the required
          components for the product.\n\nAll users will start with 4 weeks of Sosivio
          Premium after which Sosivio will automatically convert to the Community
          Version (free forever).\n\n# Help\nMore information about Sosivio and our
          product usage instructions can be found in our documentation [here](https://docs.sosiv.io/)
          \ \n"
        displayName: Sosivio | Predictive Troubleshooting for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sosivio
        - kubernetes
        - observanility
        - monitoring
        - root cause
        links:
        - name: Sosivio
          url: https://www.sosiv.io
        - name: Documentation
          url: https://docs.sosiv.io
        maintainers:
        - email: info@sosiv.io
          name: Sosivio
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: sosivio
          url: https://www.sosiv.io
        relatedImages:
        - quay.io/sosivio/discovery-engine@sha256:b87b3f7c9ed3aed80d0a77788fd8925789450d9dc05d25017e89dff611cd82da
        - quay.io/sosivio/nsq@sha256:781aac23ef0a1ed0af2b8e93d625ae49bd11dbdd98787d3c79dde8d128f3810d
        - quay.io/sosivio/actuator@sha256:a9abe9730800ca2b0a1bbd7a361b2ab3a818cbf127798314e5477fb1cb16ba4c
        - quay.io/sosivio/sosivio-node-collector@sha256:61ed8dd52de5edc09d16a31b18f2f65ad9344c0b6c9cfab323c3eb3d7e7ed060
        - quay.io/sosivio/crud-manager@sha256:38289a28829fa737917d76dd4810ba6140e123061ead0a4b79fa79b2e246a293
        - quay.io/sosivio/sosivio-operator:v1.5.1-97
        - quay.io/sosivio/sosivio-dashboard@sha256:fdfca5bb2123d1008cec2b656cde73c7e9364f84c5ff4f4a6fee5751b1bfb2f8
        - quay.io/sosivio/sosivio-node-netpacket@sha256:84dd159b1dfdb52dee3eb1c4c93f323bae49ce517ff44663b69978e14bdd8390
        - quay.io/sosivio/classifier@sha256:dd8bf145524db23ff333efa1bcdff258dab116d3534f95f73775d8dd339e6963
        - quay.io/sosivio/communicator@sha256:fffec9a6c66594fde5f05e2dfb84361544ca9d53a345188c197a0ea6f53bc4b9
        - quay.io/sosivio/authentication@sha256:0ea6025f333576792f5b1fc4d96d9ab1a404749f185830410e71248683004da2
        - quay.io/sosivio/sosiviodb@sha256:a3c6c3975bc2b0316135ee5c547f1eb215f3a1f9c8b0a9c8eceff23415f93597
        - quay.io/sosivio/kube-rbac-proxy:v0.8.0
        - quay.io/sosivio/sequence-recognition@sha256:7335cf8f3e40bc46e8ca0de525d99f0655833b7c52d36dc273aa7b096eb70fa2
        - quay.io/sosivio/network-aggregator@sha256:e59d937f4d9c52acad0adadab39c6849d20b4ee72a83e239a33ea137f08323f4
        - quay.io/sosivio/correlation-entities@sha256:c00af09b2589bad4740fd44d7a81c69b2e29138c360939b2a45f65edd44b6901
        - quay.io/sosivio/contract-testing@sha256:cad6fac6ed3fdf2597ceb09ba5dbb7d81f70cf45f483a7064e3968215dd982e3
        - quay.io/sosivio/kube-rbac-proxy@sha256:34e8724e0f47e31eb2ec3279ac398b657db5f60f167426ee73138e2e84af6486
        - quay.io/sosivio/analyzer@sha256:3825608fe7c43ef059108e76665acc99fe72bea77d4d9e3852f6ad45e74379b1
        - quay.io/sosivio/notifier@sha256:35f5f3e8e5ccad914fcfe34fb4b4a1b044ee87f5a6141284cd5273f0b2b3bf22
        - quay.io/sosivio/sosivio-node-pinger@sha256:937de52ad7b76e6cff4ad5197de9705988d61f285ebadac5ebadb5856f59ee58
        - quay.io/sosivio/sosivio-operator@sha256:11bc627773bb5e08b629e0b013e7f600fcb4aaefbd97f2de337d339475616763
        version: 1.5.1
      entries:
      - name: sosivio-operator.v1.5.1-97
        version: 1.5.1
      - name: sosivio-operator.v1.5.0-95
        version: 1.5.0
      - name: sosivio-operator.v1.4.4-85
        version: 1.4.4
      - name: sosivio-operator.v1.4.3-83
        version: 1.4.3
      - name: sosivio-operator.v1.4.2-82
        version: 1.4.2
      - name: sosivio-operator.v1.4.1-81
        version: 1.4.1-81
      - name: sosivio-operator.v1.4.1.1-80
        version: 1.4.1-80
      - name: sosivio-operator.v1.4.1-78
        version: 1.4.1-79
      - name: sosivio-operator.v1.3.1-76
        version: 1.3.1-76
      - name: sosivio-operator.v1.3.0-75
        version: 1.3.0-75
      - name: sosivio-operator.v1.2.1-1
        version: 1.2.1-1
      - name: sosivio-operator.v1.2.0-1
        version: 1.2.0-1
      name: stable
    defaultChannel: stable
    packageName: sosivio
    provider:
      name: sosivio
      url: https://www.sosiv.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-keycloak-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-keycloak-operator.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2023-09-19T12:43:29Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KCOSClientRoleMapping
            name: kcosclientrolemappings.keycloak.org
            version: v1alpha1
          - kind: KCOSRealmRole
            name: kcosrealmroles.keycloak.org
            version: v1alpha1
          - kind: KCOSUserRoleMapping
            name: kcosuserrolemappings.keycloak.org
            version: v1alpha1
          - kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: Keysight Keycloak Operator
        displayName: Keysight Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Keycloak Operator
          url: https://keysight-keycloak-operator.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak@sha256:772f73cce6a77b9c46f8035faac99a28602219657d18b75b19b4d905d91942c9
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak-postgresql@sha256:b636955199065e18f9d4679835ae78214843258e5abb4ae758288d6f00df2f8e
        - prod.harbor.keysight.digital/loadcore/keysight-keycloak-operator@sha256:d74e7e98a114768a6f3327edabce9336743a078844fffee985053b6a90100da8
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak-init-container@sha256:90965beaf41c760cb8c8bf906acf08f8fe4770d313de8dfd92d18d7c25348a38
        version: 0.1.21
      entries:
      - name: keysight-keycloak-operator.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-keycloak-operator
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: fence-agents-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fence-agents-remediation.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  }
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:00cc420250d8e25d9da7c43d9c4cff35c7edeb7d390af754c48c0ce00e9e7770
          createdAt: "2023-10-30 14:06:57"
          description: Fence Agents Remediation Operator for remediating nodes using
            upstream fence-agents.
          olm.skipRange: '>=0.2.0 <0.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: "Fence Agents Remediation Operator (FAR)\n\nThe operator will
          remediate/fence a node when a FenceAgentsRemediation (far) custom resource
          exists with the node's name.\nFAR uses an existing set of upstream fencing
          agents for environments with a \ntraditional API end-point (e.g., IPMI)
          for power cycling cluster nodes.\n\nThe operator is recommended to be used
          with Node Healthcheck Operator for easier and smoother experience,\nbut
          it can also work as a standalonde remediator for the more advanced user.\n"
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:00cc420250d8e25d9da7c43d9c4cff35c7edeb7d390af754c48c0ce00e9e7770
        version: 0.2.1
      entries:
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.14-eus
    - currentCSV: fence-agents-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:a367e8fdd80938e97ac195631a996e0c537bf3666aef0199147534762dcf739f
          createdAt: "2024-08-10 22:49:23"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.2.0 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:a367e8fdd80938e97ac195631a996e0c537bf3666aef0199147534762dcf739f
        version: 0.4.1
      entries:
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.16-eus
    - currentCSV: fence-agents-remediation.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-16 10:25:57"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.5.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.5/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel9-operator@sha256:da5975557078fbfa6882df87cff5bda5671b9838b5fb0a738fea724469d03576
        version: 0.5.1
      entries:
      - name: fence-agents-remediation.v0.5.1
        version: 0.5.1
      - name: fence-agents-remediation.v0.5.0
        version: 0.5.0
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.18-eus
    - currentCSV: fence-agents-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  }
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:5c3778c7aae45f54e949fbae3d133777f7aa78fe1251ac3d5510aa28b71e2ba4
          createdAt: "2023-04-20 15:36:05"
          description: Fence Agents Remediation Operator for remediating nodes using
            upstream fence-agents.
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: "Fence Agents Remediation Operator (FAR)\n\nThe operator will
          remediate/fence a node when a FenceAgentsRemediation (far) custom resource
          exists with the node's name.\nFAR uses an existing set of upstream fencing
          agents for environments with a \ntraditional API end-point (e.g., IPMI)
          for power cycling cluster nodes.\n\nThe operator is recommended to be used
          with Node Healthcheck Operator for easier and smoother experience,\nbut
          it can also work as a standalonde remediator for the more advanced user.\n"
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3af07bfe0fb73b1a7f00487eba302d6da442e742ab4007fc50e52efd4d45dae5
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:5c3778c7aae45f54e949fbae3d133777f7aa78fe1251ac3d5510aa28b71e2ba4
        version: 0.1.0
      entries:
      - name: fence-agents-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: fence-agents-remediation.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeSecretNames": {
                    "master-0": "fence-agents-credentials-master0",
                    "master-1": "fence-agents-credentials-master1",
                    "master-2": "fence-agents-credentials-master2",
                    "worker-0": "fence-agents-credentials-worker0",
                    "worker-1": "fence-agents-credentials-worker1",
                    "worker-2": "fence-agents-credentials-worker2"
                  },
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedSecretName": "fence-agents-credentials-shared",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {
                    "spec": {
                      "agent": "fence_ipmilan",
                      "nodeSecretNames": {
                        "master-0": "fence-agents-credentials-master0",
                        "master-1": "fence-agents-credentials-master1",
                        "master-2": "fence-agents-credentials-master2",
                        "worker-0": "fence-agents-credentials-worker0",
                        "worker-1": "fence-agents-credentials-worker1",
                        "worker-2": "fence-agents-credentials-worker2"
                      },
                      "nodeparameters": {
                        "--ipport": {
                          "master-0": "6230",
                          "master-1": "6231",
                          "master-2": "6232",
                          "worker-0": "6233",
                          "worker-1": "6234",
                          "worker-2": "6235"
                        }
                      },
                      "remediationStrategy": "ResourceDeletion",
                      "retrycount": 5,
                      "retryinterval": "5s",
                      "sharedSecretName": "fence-agents-credentials-shared",
                      "sharedparameters": {
                        "--action": "reboot",
                        "--ip": "192.168.111.1",
                        "--lanplus": "",
                        "--username": "admin"
                      },
                      "timeout": "60s"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-09-08 09:19:00"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.5.0 <0.6.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.8/html/remediation_fencing_and_maintenance/about-remediation-fencing-maintenance
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel9-operator@sha256:eef1a298718650a22a3da4a07c140ce21b6f287d87af79801610317fa8f63d52
        version: 0.6.0
      entries:
      - name: fence-agents-remediation.v0.6.0
        version: 0.6.0
      - name: fence-agents-remediation.v0.5.1
        version: 0.5.1
      - name: fence-agents-remediation.v0.5.0
        version: 0.5.0
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: stable
    defaultChannel: stable
    packageName: fence-agents-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ssm-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ssm-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ssm-controller:1.1.1
          createdAt: "2025-09-19T20:57:07Z"
          description: AWS SSM controller is a service controller for managing SSM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Document represents the state of an AWS ssm Document resource.
            displayName: Document
            kind: Document
            name: documents.ssm.services.k8s.aws
            version: v1alpha1
          - description: PatchBaseline represents the state of an AWS ssm PatchBaseline
              resource.
            displayName: PatchBaseline
            kind: PatchBaseline
            name: patchbaselines.ssm.services.k8s.aws
            version: v1alpha1
          - description: ResourceDataSync represents the state of an AWS ssm ResourceDataSync
              resource.
            displayName: ResourceDataSync
            kind: ResourceDataSync
            name: resourcedatasyncs.ssm.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon SSM resources in AWS from within your Kubernetes cluster.

          **About Amazon SSM**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon SSM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ssm
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SSM Developer Resources
          url: https://aws.amazon.com/SSM/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ssm maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ssm-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ssm-controller.v1.1.1
        version: 1.1.1
      - name: ack-ssm-controller.v1.1.0
        version: 1.1.0
      - name: ack-ssm-controller.v1.0.12
        version: 1.0.12
      - name: ack-ssm-controller.v1.0.10
        version: 1.0.10
      - name: ack-ssm-controller.v1.0.9
        version: 1.0.9
      - name: ack-ssm-controller.v1.0.8
        version: 1.0.8
      - name: ack-ssm-controller.v1.0.7
        version: 1.0.7
      - name: ack-ssm-controller.v1.0.6
        version: 1.0.6
      - name: ack-ssm-controller.v1.0.5
        version: 1.0.5
      - name: ack-ssm-controller.v1.0.4
        version: 1.0.4
      - name: ack-ssm-controller.v1.0.3
        version: 1.0.3
      - name: ack-ssm-controller.v1.0.2
        version: 1.0.2
      - name: ack-ssm-controller.v1.0.1
        version: 1.0.1
      - name: ack-ssm-controller.v1.0.0
        version: 1.0.0
      - name: ack-ssm-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-ssm-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-efs-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-efs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "AccessPoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "FileSystem",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "MountTarget",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/efs-controller:1.1.1
          createdAt: "2025-09-19T18:38:13Z"
          description: AWS EFS controller is a service controller for managing EFS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessPoint represents the state of an AWS efs AccessPoint
              resource.
            displayName: AccessPoint
            kind: AccessPoint
            name: accesspoints.efs.services.k8s.aws
            version: v1alpha1
          - description: FileSystem represents the state of an AWS efs FileSystem
              resource.
            displayName: FileSystem
            kind: FileSystem
            name: filesystems.efs.services.k8s.aws
            version: v1alpha1
          - description: MountTarget represents the state of an AWS efs MountTarget
              resource.
            displayName: MountTarget
            kind: MountTarget
            name: mounttargets.efs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EFS resources in AWS from within your Kubernetes cluster.

          **About Amazon Elastic File System (Amazon EFS)**

          Amazon Elastic File System (Amazon EFS) provides serverless, fully elastic file storage so that you can share file data without provisioning or managing storage capacity and performance. Amazon EFS is built to scale on demand to petabytes without disrupting applications, growing and shrinking automatically as you add and remove files. Because Amazon EFS has a simple web services interface, you can create and configure file systems quickly and easily. The service manages all the file storage infrastructure for you, meaning that you can avoid the complexity of deploying, patching, and maintaining complex file system configurations.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Elastic File System (Amazon
          EFS)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - efs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EFS Developer Resources
          url: https://aws.amazon.com/efs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: efs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/efs-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-efs-controller.v1.1.1
        version: 1.1.1
      - name: ack-efs-controller.v1.1.0
        version: 1.1.0
      - name: ack-efs-controller.v1.0.19
        version: 1.0.19
      - name: ack-efs-controller.v1.0.18
        version: 1.0.18
      - name: ack-efs-controller.v1.0.17
        version: 1.0.17
      - name: ack-efs-controller.v1.0.16
        version: 1.0.16
      - name: ack-efs-controller.v1.0.15
        version: 1.0.15
      - name: ack-efs-controller.v1.0.14
        version: 1.0.14
      - name: ack-efs-controller.v1.0.13
        version: 1.0.13
      - name: ack-efs-controller.v1.0.12
        version: 1.0.12
      - name: ack-efs-controller.v1.0.11
        version: 1.0.11
      - name: ack-efs-controller.v1.0.10
        version: 1.0.10
      - name: ack-efs-controller.v1.0.8
        version: 1.0.8
      - name: ack-efs-controller.v1.0.6
        version: 1.0.6
      - name: ack-efs-controller.v1.0.5
        version: 1.0.5
      - name: ack-efs-controller.v1.0.3
        version: 1.0.3
      - name: ack-efs-controller.v1.0.2
        version: 1.0.2
      - name: ack-efs-controller.v1.0.0
        version: 1.0.0
      - name: ack-efs-controller.v0.0.9
        version: 0.0.9
      - name: ack-efs-controller.v0.0.8
        version: 0.0.8
      - name: ack-efs-controller.v0.0.7
        version: 0.0.7
      - name: ack-efs-controller.v0.0.6
        version: 0.0.6
      - name: ack-efs-controller.v0.0.5
        version: 0.0.5
      - name: ack-efs-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-efs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solace Corporation
      provider-url: www.solace.com
    name: pubsubplus-eventbroker-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pubsubplus-eventbroker-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pubsubplus.solace.com/v1beta1",
                "kind": "PubSubPlusEventBroker",
                "metadata": {
                  "name": "non-ha-standalone-example"
                },
                "spec": {
                  "redundancy": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "true"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.10
          containerImage: registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator:1.4.0
          createdAt: "2025-07-31T19:48:36Z"
          description: The Solace PubSub+ Event Broker Operator deploys and manages
            the lifecycle of PubSub+ Event Brokers
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
          support: Solace Products
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PubSub+ Event Broker
            displayName: PubSub+ Event Broker
            kind: PubSubPlusEventBroker
            name: pubsubpluseventbrokers.pubsubplus.solace.com
            version: v1beta1
        description: |
          ## Solace PubSub+ Platform
          Solace [PubSub+ Platform](https://solace.com/products/platform/) is a complete event streaming and management platform
          for real-time enterprises. The [PubSub+ Software Event Broker](https://solace.com/products/event-broker/software/) efficiently streams
          event-driven information between applications, IoT devices, and user interfaces running in the cloud, on-premises,
          and in hybrid environments using open APIs and protocols like AMQP, JMS, MQTT, REST and WebSocket.
          It can be installed into a variety of public and private clouds, PaaS, and on-premises environments.
          Event brokers in multiple locations can be linked together in an [Event Mesh](https://solace.com/what-is-an-event-mesh/)
          to dynamically share events across the distributed enterprise.

          ## Solace PubSub+ Event Broker Operator
          Solace Pubsub+ Event Broker Operator automatically deploys and manages PubSub+ Software Event Brokers on Kubernetes and OpenShift environments.

          Features include:
          * Initial deployment configuration: Production-ready HA, non-HA or for developers
          * Rolling update for configuration changes or broker upgrades
          * Prometheus monitoring support

          ## Getting started
          Refer to the [Solace PubSub+ Event Broker Operator Quick Start](https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/README.md)
        displayName: Solace PubSub+ Event Broker Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - solace
        - pubsubplus
        - pubsub+
        - pubsub
        - messaging
        - advanced event broker
        - event broker
        - event mesh
        - event streaming
        - data streaming
        - event integration
        - middleware
        links:
        - name: Operator QuickStart
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
        - name: Operator Documentation
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/docs/EventBrokerOperatorUserGuide.md
        - name: Solace.com
          url: https://solace.com
        - name: Understanding Solace Technology
          url: https://solace.com/products/platform/
        - name: Solace PubSub+ Technical Documentation
          url: https://docs.solace.com/
        - name: Solace Developer Portal
          url: https://www.solace.dev/
        - name: Solace Community
          url: https://solace.community/
        maintainers:
        - email: support@solace.com
          name: Solace R&D
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Solace Corporation
          url: www.solace.com
        relatedImages:
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator:1.4.0
        version: 1.4.0
      entries:
      - name: pubsubplus-eventbroker-operator.v1.4.0
        version: 1.4.0
      - name: pubsubplus-eventbroker-operator.v1.3.0
        version: 1.3.0
      - name: pubsubplus-eventbroker-operator.v1.2.0
        version: 1.2.0
      - name: pubsubplus-eventbroker-operator.v1.1.2
        version: 1.1.2
      - name: pubsubplus-eventbroker-operator.v1.1.1
        version: 1.1.1
      - name: pubsubplus-eventbroker-operator.v1.1.0
        version: 1.1.0
      - name: pubsubplus-eventbroker-operator.v1.0.3
        version: 1.0.3
      - name: pubsubplus-eventbroker-operator.v1.0.2
        version: 1.0.2
      - name: pubsubplus-eventbroker-operator.v1.0.1
        version: 1.0.1
      - name: pubsubplus-eventbroker-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pubsubplus-eventbroker-operator
    provider:
      name: Solace Corporation
      url: www.solace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: ibm-data-reporter-operator
      app.kubernetes.io/managed-by: ibm-data-reporter-operator
      app.kubernetes.io/name: ibm-data-reporter-operator
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: ""
    name: ibm-data-reporter-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-data-reporter-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "DataReporterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-data-reporter-operator",
                    "app.kubernetes.io/instance": "datareporterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datareporterconfig",
                    "app.kubernetes.io/part-of": "ibm-data-reporter-operator"
                  },
                  "name": "datareporterconfig"
                },
                "spec": {
                  "userConfig": [
                    {
                      "metadata": {
                        "ameta1": "ametadata1",
                        "bmeta1": "bmetadata1",
                        "cmeta1": "cmetadata1",
                        "dmeta1": "dmetadata1"
                      },
                      "userName": "system:serviceaccount:redhat-marketplace:ibm-data-reporter-operator-api"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
          createdAt: ""
          description: The IBM Data Reporter Operator accepts events and transforms
            them into reports submitted to the Data Service of the IBM Metrics Operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReporterConfig is the Schema for the datareporterconfigs
              API
            displayName: Data Reporter Config
            kind: DataReporterConfig
            name: datareporterconfigs.marketplace.redhat.com
            version: v1alpha1
          required:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
        description: "# IBM Data Reporter Operator\n\n# Introduction\n\nThe IBM Data
          Reporter Operator accepts events and transforms them into reports submitted
          to the IBM Metrics Operator Data Service.\n\n# Details\n\nThe IBM Data Reporter
          Operator deploys a service that provides an endpoint to which callers can
          send raw json event data. The event data is transformed into a report and
          is sent to the IBM Metrics Operator's Data Service. The IBM Metrics Operator
          Data Service periodically uploads the reports to IBM Software Central. It
          can also filter, transform and forward events to alternate endpoints.\n\n##
          Prerequisites\n\n- OpenShift Container Platform, major version 4 with any
          available supported minor version\n- Install IBM Metrics Operator\n  - IBM
          Data Reporter Operator prerequisties the IBM Metrics Operator's Data Service\n
          \ - Register the Cluster by creating a `redhat-marketplace-pull-secret`,
          as per the instructions\n  - Once `rhm-data-service` has started IBM Data
          Reporter Operator will accept requests\n\n## SecurityContextConstraints
          Requirements\n\n- The operator runs under Red Hat restricted SCC\n\n## Resources
          Required\n\n- The operator requires at least 85 Mi memory.\n\n## Limitations\n\n-
          Only runs on amd64, s390x and ppc64le architectures\n\n## Installing\n\n-
          A user with the cluster administrator role.\n- Install this operator in
          the same namespace as the IBM Metrics Operator\n  - default namespace: `ibm-software-central`\n\n##
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n# Configuration\n\nOptional:\n- Configure the
          Custom Resource Definition `DataReporterConfig` named `datareporterconfig`
          as per the following sample\n  - Configuration that can not be reconciled
          successfully will be reported on the DataReporterConfig Status\n  - jsonPath
          expressions are handled by: https://github.com/ohler55/ojg\n    - jsonPath
          comparison: https://cburgmer.github.io/json-path-comparison/\n  - Transformer
          type currently supported is `kazaam`\n    - https://github.com/willie68/kazaam\n\nSample
          DataReporterConfig:\n```YAML\napiVersion: marketplace.redhat.com/v1alpha1\nkind:
          DataReporterConfig\nmetadata:\n  name: datareporterconfig\nspec:\n  # true:
          report generated for 1 event and provide return code from data-service,
          confirming delivery\n  # false: events will be accumulated before building
          report, and immediately return 200, delivery unfconfirmed\n  confirmDelivery:
          false\n  dataFilters:   # dataFilters matches the first dataFilter selector,
          applies transforms and forwards to data-service and destinations\n  - altDestinations:
          \  # List of desintations to transform and foward events to\n    - authorization:
          \  # optional endpoint if an authorization token must first be requested
          \n        authDestHeader: Authorization   # http header key to be appended
          to destination request\n        authDestHeaderPrefix: 'Bearer '   # optional
          prefix for constructing destination request http header value\n        bodyData:
          \  # body data to send to authorization request, often an apiKey\n          secretKeyRef:\n
          \           name: auth-body-data-secret\n            key: bodydata\n        header:
          \  # secret map to use for authorization request headers\n          secret:\n
          \           name: auth-header-map-secret\n        tokenExpr: $.token   #
          optionally extract the returned body data using a jsonPath expression\n
          \       url: https://127.0.0.1:8443/api/2.0/accounts/my_account_id/apikeys/token
          \  # the url to make an authorization request to\n      header:   # secret
          map to use for destination request headers\n        secret:\n          name:
          dest-header-map-secret\n      transformer:   # The transformer to apply
          to event data for this destination\n        configMapKeyRef:\n          key:
          kazaam.json\n          name: kazaam-configmap\n        type: kazaam\n      url:
          https://127.0.0.1:8443/metering/api/1.0/usage   # the url to send the event
          to\n      urlSuffixExpr: $.properties.productId   # optional jsonPath expression
          on the event to extract and set a suffix to the url path\n    manifestType:
          dataReporter   # override manifestType of report sent to data-service\n
          \   selector:   # Matches against event data and a user. Empty selector
          matches all.\n      matchExpressions:   # jsonPath expression must return
          a result for all expressions to match (AND)\n      - $[?($.event == \"Account
          Contractual Usage\")]\n      - $.properties.productId\n      - $[?($.properties.source
          != null)]\n      - $[?($.properties.unit == \"AppPoints\")]\n      - $[?($.properties.quantity
          >= 0)]\n      - $[?($.timestamp != null)]\n      matchUsers:   # must match
          one of these users (OR). Omitting users matches any user\n      - system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n
          \   transformer:   # The transformer to apply to the event for data-service\n
          \     configMapKeyRef:\n        key: kazaam.json\n        name: kazaam-configmap\n
          \     type: kazaam\n  tlsConfig:   # TLS configuration for requests outbound
          to destinations\n    caCerts:\n    - name: tls-ca-certificates\n      key:
          ca.crt\n    certificates:\n    - clientCert:\n        secretKeyRef:\n          name:
          tls-ca-certificates\n          key: tls.crt\n      clientKey:\n        secretKeyRef:\n
          \         name: tls-ca-certificates\n          key: tls.key\n    cipherSuites:
          \n    - TLS_AES_128_GCM_SHA256\n    - TLS_AES_256_GCM_SHA384\n    - TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\n    - TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\n    insecureSkipVerify: false\n
          \   minVersion: VersionTLS12\n  userConfig:   # Optional metadata to apply
          to report sent to data-service for a specific user\n  - metadata:\n      ameta1:
          ametadata1\n      bmeta1: bmetadata1\n      cmeta1: cmetadata1\n      dmeta1:
          dmetadata1\n    userName: system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n```\n\nSample
          Header Secrets:\n```YAML\napiVersion: v1\nstringData:\n  accept: | \n    */*\nkind:
          Secret\nmetadata:\n  name: dest-header-map-secret\ntype: Opaque\n---\napiVersion:
          v1\nstringData:\n  accept: application/json\n  Content-Type: application/json\nkind:
          Secret\nmetadata:\n  name: auth-header-map-secret\ntype: Opaque\n```\n\nSample
          Authentication Body Secret:\n```YAML\napiVersion: v1\nstringData:\n  bodydata:
          |\n    {\"apikey\": \"myapikey\"}\nkind: Secret\nmetadata:\n  name: auth-body-data-secret\ntype:
          Opaque\n```\n\nSample TLS Secret:\n```YAML\napiVersion: v1\ndata:\n  ca.crt:
          LS0t...\n  tls.key: LS0t...\n  tls.crt: LS0t...\nkind: Secret\nmetadata:\n
          \ name: tls-ca-certificates\n```\n\nSample Kazaam ConfigMap:\n```YAML\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: kazaam-configmap\ndata:\n  kazaam.json:
          |\n    [\n        {\n            \"operation\": \"timestamp\",\n            \"spec\":
          {\n                \"timestamp\": {\n                \"inputFormat\": \"2006-01-02T15:04:05.999999+00:00\",\n
          \               \"outputFormat\": \"$unixmilli\"\n                }\n            }\n
          \       },\n        {\n            \"operation\": \"shift\",\n            \"spec\":
          {\n                \"instances[0].instanceId\": \"properties.source\",\n
          \               \"instances[0].startTime\": \"timestamp\",\n                \"instances[0].endTime\":
          \"timestamp\",\n                \"instances[0].metricUsage[0].metricId\":
          \"properties.unit\",\n                \"instances[0].metricUsage[0].quantity\":
          \"properties.quantity\"\n            }\n        },\n        {\n            \"operation\":
          \"default\",\n            \"spec\": {\n                \"meteringModel\":
          \"point-in-time\",\n                \"meteringPlan\": \"contract\"\n            }\n
          \       }\n    ]\n```\n\n### API Service User Configuration\n\n- The ClusterRole
          for api access is `clusterrole/ibm-data-reporter-operator-api`\n- The default
          ServiceAccount provided as an api user is `system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api`\n
          \ - The default ClusterRoleBinding for this user is `clusterrolebinding/ibm-data-reporter-operator-api`\n\nOptional:\n\n-
          To create an additional ServiceAccount and ClusterRoleBinding for api access\n\n```SHELL\nNAMESPACE=$(oc
          config view --minify -o jsonpath='{..namespace}') && \\\noc create serviceaccount
          my-api-service-account && \\\noc create clusterrolebinding ibm-data-reporter-operator-api
          --clusterrole=ibm-data-reporter-operator-api --serviceaccount=${NAMESPACE}:my-api-service-account\n```\n\n-
          Update datareporterconfig to attach metadata to reports associated with
          this user\n\n## API Service Usage\n\n- Get Token & Host\n\n```SHELL\noc
          project ibm-software-central\nexport DRTOKEN=$(oc create token ibm-data-reporter-operator-api
          --namespace ibm-software-central --duration 1h)\nexport DRHOST=$(oc get
          route ibm-data-reporter --template='{{ .spec.host }}')\n```\n\n- Get the
          Status\n\n```SHELL\ncurl -k -H \"Authorization: Bearer ${DRTOKEN}\" https://${DRHOST}/v1/status
          \n```\n\n- Post an Event\n\n```SHELL\ncurl -k -H \"Authorization: Bearer
          ${DRTOKEN}\" -X POST -d '{\"event\":\"myevent\"}' https://${DRHOST}/v1/event\n```\n\n##
          Storage\n\n- The operator temporarily stores events in pod memory, and writes
          to the IBM Metrics Operator data-service, which requires a PersistentVolume\n\n##
          License\n\nCopyright IBM Corporation 2023. All Rights Reserved.\n\nLicensed
          under the Apache License, Version 2.0 (the \"License\");\nyou may not use
          this file except in compliance with the License.\nYou may obtain a copy
          of the License at\n\n    http://www.apache.org/licenses/LICENSE-2.0\n\nUnless
          required by applicable law or agreed to in writing, software\ndistributed
          under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES
          OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for
          the specific language governing permissions and\nlimitations under the License.\n"
        displayName: IBM Data Reporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 2.20.0
      entries:
      - name: ibm-data-reporter-operator.v2.20.0
        version: 2.20.0
      name: beta
    - currentCSV: ibm-data-reporter-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "DataReporterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-data-reporter-operator",
                    "app.kubernetes.io/instance": "datareporterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datareporterconfig",
                    "app.kubernetes.io/part-of": "ibm-data-reporter-operator"
                  },
                  "name": "datareporterconfig"
                },
                "spec": {
                  "userConfig": [
                    {
                      "metadata": {
                        "ameta1": "ametadata1",
                        "bmeta1": "bmetadata1",
                        "cmeta1": "cmetadata1",
                        "dmeta1": "dmetadata1"
                      },
                      "userName": "system:serviceaccount:redhat-marketplace:ibm-data-reporter-operator-api"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
          createdAt: ""
          description: The IBM Data Reporter Operator accepts events and transforms
            them into reports submitted to the Data Service of the IBM Metrics Operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReporterConfig is the Schema for the datareporterconfigs
              API
            displayName: Data Reporter Config
            kind: DataReporterConfig
            name: datareporterconfigs.marketplace.redhat.com
            version: v1alpha1
          required:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
        description: "# IBM Data Reporter Operator\n\n# Introduction\n\nThe IBM Data
          Reporter Operator accepts events and transforms them into reports submitted
          to the IBM Metrics Operator Data Service.\n\n# Details\n\nThe IBM Data Reporter
          Operator deploys a service that provides an endpoint to which callers can
          send raw json event data. The event data is transformed into a report and
          is sent to the IBM Metrics Operator's Data Service. The IBM Metrics Operator
          Data Service periodically uploads the reports to IBM Software Central. It
          can also filter, transform and forward events to alternate endpoints.\n\n##
          Prerequisites\n\n- OpenShift Container Platform, major version 4 with any
          available supported minor version\n- Install IBM Metrics Operator\n  - IBM
          Data Reporter Operator prerequisties the IBM Metrics Operator's Data Service\n
          \ - Register the Cluster by creating a `redhat-marketplace-pull-secret`,
          as per the instructions\n  - Once `rhm-data-service` has started IBM Data
          Reporter Operator will accept requests\n\n## SecurityContextConstraints
          Requirements\n\n- The operator runs under Red Hat restricted SCC\n\n## Resources
          Required\n\n- The operator requires at least 85 Mi memory.\n\n## Limitations\n\n-
          Only runs on amd64, s390x and ppc64le architectures\n\n## Installing\n\n-
          A user with the cluster administrator role.\n- Install this operator in
          the same namespace as the IBM Metrics Operator\n  - default namespace: `ibm-software-central`\n\n##
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n# Configuration\n\nOptional:\n- Configure the
          Custom Resource Definition `DataReporterConfig` named `datareporterconfig`
          as per the following sample\n  - Configuration that can not be reconciled
          successfully will be reported on the DataReporterConfig Status\n  - jsonPath
          expressions are handled by: https://github.com/ohler55/ojg\n    - jsonPath
          comparison: https://cburgmer.github.io/json-path-comparison/\n  - Transformer
          type currently supported is `kazaam`\n    - https://github.com/willie68/kazaam\n\nSample
          DataReporterConfig:\n```YAML\napiVersion: marketplace.redhat.com/v1alpha1\nkind:
          DataReporterConfig\nmetadata:\n  name: datareporterconfig\nspec:\n  # true:
          report generated for 1 event and provide return code from data-service,
          confirming delivery\n  # false: events will be accumulated before building
          report, and immediately return 200, delivery unfconfirmed\n  confirmDelivery:
          false\n  dataFilters:   # dataFilters matches the first dataFilter selector,
          applies transforms and forwards to data-service and destinations\n  - altDestinations:
          \  # List of desintations to transform and foward events to\n    - authorization:
          \  # optional endpoint if an authorization token must first be requested
          \n        authDestHeader: Authorization   # http header key to be appended
          to destination request\n        authDestHeaderPrefix: 'Bearer '   # optional
          prefix for constructing destination request http header value\n        bodyData:
          \  # body data to send to authorization request, often an apiKey\n          secretKeyRef:\n
          \           name: auth-body-data-secret\n            key: bodydata\n        header:
          \  # secret map to use for authorization request headers\n          secret:\n
          \           name: auth-header-map-secret\n        tokenExpr: $.token   #
          optionally extract the returned body data using a jsonPath expression\n
          \       url: https://127.0.0.1:8443/api/2.0/accounts/my_account_id/apikeys/token
          \  # the url to make an authorization request to\n      header:   # secret
          map to use for destination request headers\n        secret:\n          name:
          dest-header-map-secret\n      transformer:   # The transformer to apply
          to event data for this destination\n        configMapKeyRef:\n          key:
          kazaam.json\n          name: kazaam-configmap\n        type: kazaam\n      url:
          https://127.0.0.1:8443/metering/api/1.0/usage   # the url to send the event
          to\n      urlSuffixExpr: $.properties.productId   # optional jsonPath expression
          on the event to extract and set a suffix to the url path\n    manifestType:
          dataReporter   # override manifestType of report sent to data-service\n
          \   selector:   # Matches against event data and a user. Empty selector
          matches all.\n      matchExpressions:   # jsonPath expression must return
          a result for all expressions to match (AND)\n      - $[?($.event == \"Account
          Contractual Usage\")]\n      - $.properties.productId\n      - $[?($.properties.source
          != null)]\n      - $[?($.properties.unit == \"AppPoints\")]\n      - $[?($.properties.quantity
          >= 0)]\n      - $[?($.timestamp != null)]\n      matchUsers:   # must match
          one of these users (OR). Omitting users matches any user\n      - system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n
          \   transformer:   # The transformer to apply to the event for data-service\n
          \     configMapKeyRef:\n        key: kazaam.json\n        name: kazaam-configmap\n
          \     type: kazaam\n  tlsConfig:   # TLS configuration for requests outbound
          to destinations\n    caCerts:\n    - name: tls-ca-certificates\n      key:
          ca.crt\n    certificates:\n    - clientCert:\n        secretKeyRef:\n          name:
          tls-ca-certificates\n          key: tls.crt\n      clientKey:\n        secretKeyRef:\n
          \         name: tls-ca-certificates\n          key: tls.key\n    cipherSuites:
          \n    - TLS_AES_128_GCM_SHA256\n    - TLS_AES_256_GCM_SHA384\n    - TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\n    - TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\n    insecureSkipVerify: false\n
          \   minVersion: VersionTLS12\n  userConfig:   # Optional metadata to apply
          to report sent to data-service for a specific user\n  - metadata:\n      ameta1:
          ametadata1\n      bmeta1: bmetadata1\n      cmeta1: cmetadata1\n      dmeta1:
          dmetadata1\n    userName: system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n```\n\nSample
          Header Secrets:\n```YAML\napiVersion: v1\nstringData:\n  accept: | \n    */*\nkind:
          Secret\nmetadata:\n  name: dest-header-map-secret\ntype: Opaque\n---\napiVersion:
          v1\nstringData:\n  accept: application/json\n  Content-Type: application/json\nkind:
          Secret\nmetadata:\n  name: auth-header-map-secret\ntype: Opaque\n```\n\nSample
          Authentication Body Secret:\n```YAML\napiVersion: v1\nstringData:\n  bodydata:
          |\n    {\"apikey\": \"myapikey\"}\nkind: Secret\nmetadata:\n  name: auth-body-data-secret\ntype:
          Opaque\n```\n\nSample TLS Secret:\n```YAML\napiVersion: v1\ndata:\n  ca.crt:
          LS0t...\n  tls.key: LS0t...\n  tls.crt: LS0t...\nkind: Secret\nmetadata:\n
          \ name: tls-ca-certificates\n```\n\nSample Kazaam ConfigMap:\n```YAML\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: kazaam-configmap\ndata:\n  kazaam.json:
          |\n    [\n        {\n            \"operation\": \"timestamp\",\n            \"spec\":
          {\n                \"timestamp\": {\n                \"inputFormat\": \"2006-01-02T15:04:05.999999+00:00\",\n
          \               \"outputFormat\": \"$unixmilli\"\n                }\n            }\n
          \       },\n        {\n            \"operation\": \"shift\",\n            \"spec\":
          {\n                \"instances[0].instanceId\": \"properties.source\",\n
          \               \"instances[0].startTime\": \"timestamp\",\n                \"instances[0].endTime\":
          \"timestamp\",\n                \"instances[0].metricUsage[0].metricId\":
          \"properties.unit\",\n                \"instances[0].metricUsage[0].quantity\":
          \"properties.quantity\"\n            }\n        },\n        {\n            \"operation\":
          \"default\",\n            \"spec\": {\n                \"meteringModel\":
          \"point-in-time\",\n                \"meteringPlan\": \"contract\"\n            }\n
          \       }\n    ]\n```\n\n### API Service User Configuration\n\n- The ClusterRole
          for api access is `clusterrole/ibm-data-reporter-operator-api`\n- The default
          ServiceAccount provided as an api user is `system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api`\n
          \ - The default ClusterRoleBinding for this user is `clusterrolebinding/ibm-data-reporter-operator-api`\n\nOptional:\n\n-
          To create an additional ServiceAccount and ClusterRoleBinding for api access\n\n```SHELL\nNAMESPACE=$(oc
          config view --minify -o jsonpath='{..namespace}') && \\\noc create serviceaccount
          my-api-service-account && \\\noc create clusterrolebinding ibm-data-reporter-operator-api
          --clusterrole=ibm-data-reporter-operator-api --serviceaccount=${NAMESPACE}:my-api-service-account\n```\n\n-
          Update datareporterconfig to attach metadata to reports associated with
          this user\n\n## API Service Usage\n\n- Get Token & Host\n\n```SHELL\noc
          project ibm-software-central\nexport DRTOKEN=$(oc create token ibm-data-reporter-operator-api
          --namespace ibm-software-central --duration 1h)\nexport DRHOST=$(oc get
          route ibm-data-reporter --template='{{ .spec.host }}')\n```\n\n- Get the
          Status\n\n```SHELL\ncurl -k -H \"Authorization: Bearer ${DRTOKEN}\" https://${DRHOST}/v1/status
          \n```\n\n- Post an Event\n\n```SHELL\ncurl -k -H \"Authorization: Bearer
          ${DRTOKEN}\" -X POST -d '{\"event\":\"myevent\"}' https://${DRHOST}/v1/event\n```\n\n##
          Storage\n\n- The operator temporarily stores events in pod memory, and writes
          to the IBM Metrics Operator data-service, which requires a PersistentVolume\n\n##
          License\n\nCopyright IBM Corporation 2023. All Rights Reserved.\n\nLicensed
          under the Apache License, Version 2.0 (the \"License\");\nyou may not use
          this file except in compliance with the License.\nYou may obtain a copy
          of the License at\n\n    http://www.apache.org/licenses/LICENSE-2.0\n\nUnless
          required by applicable law or agreed to in writing, software\ndistributed
          under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES
          OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for
          the specific language governing permissions and\nlimitations under the License.\n"
        displayName: IBM Data Reporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
        version: 2.20.0
      entries:
      - name: ibm-data-reporter-operator.v2.20.0
        version: 2.20.0
      name: stable
    defaultChannel: stable
    packageName: ibm-data-reporter-operator
    provider:
      name: Red Hat Marketplace
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-xtradb-cluster-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "docker.io/percona/percona-xtradb-cluster:8.0.42-33.1",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "docker.io/percona/haproxy:2.8.15",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "docker.io/percona/proxysql2:2.7.3",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "docker.io/percona/fluentbit:4.0.1",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:3.3.1",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "docker.io/percona/percona-xtrabackup:8.0.35-34.1",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.io/percona/percona-xtradb-cluster-operator:1.18.0
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <v1.18.0
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            clustercheck: Y2x1c3RlcmNoZWNrcGFzc3dvcmQ=
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            clustercheck: Y2x1c3RlcmNoZWNrcGFzc3dvcmQ=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ```
          ###  Release Highlights
          * **PMM3 support**

          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**

          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**

          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**

          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.

        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - database
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - docker.io/percona/percona-xtradb-cluster-operator:1.18.0
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator.v1.17.0
        version: 1.17.0
      - name: percona-xtradb-cluster-operator.v1.16.1
        version: 1.16.1
      - name: percona-xtradb-cluster-operator.v1.16.0
        version: 1.16.0
      - name: percona-xtradb-cluster-operator.v1.15.1
        version: 1.15.1
      - name: percona-xtradb-cluster-operator.v1.15.0
        version: 1.15.0
      - name: percona-xtradb-cluster-operator.v1.9.0
        version: 1.9.0
      name: stable
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: authorino-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: authorino-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta3",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/authorino-rhel9-operator@sha256:2d81613856df9725ad20928a7a9895d14482c42effce0a3a0de88f40ec35d411
          createdAt: "2025-03-25T22:39:23Z"
          description: Enables authentication and authorization for Gateways and applications
            in a Gateway API network.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta3
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: |-
          [Authorino](https://docs.kuadrant.io/authorino/) is Red Hat's Kubernetes-native lightweight external authorization service for tailor-made Zero Trust API security.

          Install this Red Hat official distribution of [Authorino Operator](https://docs.kuadrant.io/authorino-operator/) to manage instances of Authorino in this cluster.

          The current state of this distribution of the operator is: **Generally Available**.

          The Community version of Authorino Operator, based on upstream public images, is available in [OperatorHub.io](https://operatorhub.io/operator/authorino-operator).

          ## Getting started

          After installing the Operator, request an instance of the external authorization service by creating an `Authorino` custom resource.

          **Minimal example (namespaced)**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            listener:
              tls:
                enabled: false
            oidcServer:
              tls:
                enabled: false
          ```

          **Extended example**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            clusterWide: true
            authConfigLabelSelectors: environment=production
            secretLabelSelectors: authorino.kuadrant.io/component=authorino,environment=production

            replicas: 2

            evaluatorCacheSize: 2 # mb

            logLevel: info
            logMode: production

            listener:
              ports:
                grpc: 50001
                http: 5001
              tls:
                certSecretRef:
                  name: authorino-server-cert # secret must contain `tls.crt` and `tls.key` entries
              timeout: 2

            oidcServer:
              port: 8083
              tls:
                certSecretRef:
                  name: authorino-oidc-server-cert # secret must contain `tls.crt` and `tls.key` entries

            metrics:
              port: 8080
              deep: true

            healthz:
              port: 8081

            tracing:
              endpoint: rpc://otel-collector.observability.svc.cluster.local:4317
              insecure: true

            volumes:
              items:
                - name: keycloak-tls-cert
                  mountPath: /etc/ssl/certs
                  configMaps:
                    - keycloak-tls-cert
                  items: # details to mount the k8s configmap in the authorino pods
                    - key: keycloak.crt
                      path: keycloak.crt
              defaultMode: 420
          ```

          ### Cluster-wide vs Namespaced

          Namespaced instances only watch auth resources (`AuthConfig` and `Secrets`) created in the same namespace as the Authorino service. Use this mode for dedicated instances that do not require elevated privileges.

          Cluster-wide instances watch resources across the entire cluster (all namespaces.) Deploying and running Authorino in this mode requires elevated privileges.

          ### Multi-tenancy

          Use the `authConfigLabelSelectors` field of the `Authorino` custom resource to narrow the scope of the Authorino instance.

          Only `AuthConfig` custom resources whose labels match the value of this field will be handled by the Authorino instance.

          ## Protect a host

          To protect a host, create an `AuthConfig` custom resource for the host. E.g.:

          ```yaml
          apiVersion: authorino.kuadrant.io/v1beta3
          kind: AuthConfig
          metadata:
            name: my-api-protection
          spec:
            hosts:
            - my-api.io

            authentication:
              "keycloak":
                jwt:
                  issuerUrl: https://keycloak.keycloak.svc.cluster.local:8080/realms/my-realm

            authorization:
              "k8s-rbac":
                kubernetesSubjectAccessReview:
                  user:
                    selector: auth.identity.user.username
                  resourceAttributes:
                    resource:
                      value: my-api
                    verb:
                      selector: request.method
                cache:
                  key:
                    selector: auth.identity.user.username
                  ttl: 30
              "after-2am-only":
                rego: |
                  allow {
                    [hour, _, _] := time.clock(time.now_ns())
                    hour >= 2
                  }
          ```

          Make sure all requests to the host are fisrt checked with the Authorino instance, by configuring an Envoy proxy for external authz:

          ```yaml
          clusters:
          - name: my-api
            …
          - name: authorino
            connect_timeout: 0.25s
            type: STRICT_DNS
            lb_policy: ROUND_ROBIN
            http2_protocol_options: {}
            load_assignment:
              cluster_name: authorino
              endpoints:
              - lb_endpoints:
                - endpoint:
                    address:
                      socket_address:
                        address: authorino-authorino-authorization
                        port_value: 50051
          listeners:
          - filter_chains:
            - filters:
              name: envoy.http_connection_manager
              typed_config:
                "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
                use_remote_address: true
                route_config:
                  name: my-api-config
                  virtual_hosts:
                  - name: my-api-vs
                    domains:
                    - my-api.io
                    routes:
                    - match:
                        prefix: /
                      route:
                        cluster: my-api
                http_filters:
                - name: envoy.filters.http.ext_authz
                  typed_config:
                    "@type": type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz
                    transport_api_version: V3
                    failure_mode_allow: false
                    include_peer_certificate: true
                    grpc_service:
                      envoy_grpc:
                        cluster_name: authorino
                      timeout: 1s
          ```

          ...or, if using Istio, by creating an [`AuthorizationPolicy`](https://istio.io/latest/docs/reference/config/security/authorization-policy/#AuthorizationPolicy) custom resource. Use `action: CUSTOM` in the resource and the Authorino authorization service configured in the mesh extension provider settings.

          ## Features

          **Authentication**

          * JWT validation (with OpenID Connect Discovery)
          * OAuth 2.0 Token Introspection (opaque tokens)
          * Kubernetes TokenReview (ServiceAccount tokens)
          * API key authentication
          * X.509 client certificate authentication
          * Anonymous access
          * Proxy-handled (authentication performed by the proxy)

          **Authorization**

          * Built-in simple pattern matching (e.g. JWT claims, request attributes checking)
          * OPA policies (inline Rego and fetch from external registry)
          * Kubernetes SubjectAccessReview (resource and non-resource attributes)
          * Authzed SpiceDB

          **External metadata**

          * HTTP request
          * OpenID Connect User Info
          * UMA-protected resource attributes

          **Custom responses**

          * Header injection (Festival Wristbands tokens, JSON, plain text)
          * Envoy Dynamic Metadata
          * Custom HTTP response (status code, headers, messages, body, etc)

          **Callbacks**

          * HTTP webhooks

          **Caching**

          * OpenID Connect and User-Managed Access configs
          * JSON Web Keys (JWKs) and JSON Web Key Sets (JWKS)
          * Access tokens
          * External metadata
          * Precompiled Rego policies
          * Policy evaluation

          Check out the full [Feature Specification](https://docs.kuadrant.io/authorino/docs/features/) and how-to guides in the [Kuadrant Docs](https://docs.kuadrant.io) website.
        displayName: Red Hat - Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: GitHub
          url: https://github.com/kuadrant/authorino-operator
        - name: Documentation
          url: https://docs.kuadrant.io/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhcl-1/authorino-rhel9-operator@sha256:2d81613856df9725ad20928a7a9895d14482c42effce0a3a0de88f40ec35d411
        - registry.redhat.io/rhcl-1/authorino-rhel9@sha256:413d3796ad0101bf14d3341d31d9367e67fea1e73ebb3fb79124ef4c5dde0c09
        version: 1.2.2
      entries:
      - name: authorino-operator.v1.2.2
        version: 1.2.2
      - name: authorino-operator.v1.2.1
        version: 1.2.1
      - name: authorino-operator.v1.2.0
        version: 1.2.0
      - name: authorino-operator.v1.1.3
        version: 1.1.3
      - name: authorino-operator.v1.1.2
        version: 1.1.2
      - name: authorino-operator.v1.1.1
        version: 1.1.1
      - name: authorino-operator.v1.1.0
        version: 1.1.0
      - name: authorino-operator.v1.0.2
        version: 1.0.2
      - name: authorino-operator.v0.16.1
        version: 0.16.1
      - name: authorino-operator.v0.16.0
        version: 0.16.0
      name: stable
    - currentCSV: authorino-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta1",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "hosts": [
                    "my-api.io"
                  ],
                  "identity": [
                    {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "in": "authorization_header",
                        "keySelector": "APIKEY"
                      },
                      "name": "api-key-users"
                    }
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/3scale-tech-preview/authorino-rhel9-operator@sha256:eb574781cf39917f8f5d9e5c806f520534601f819ee1ebcbb751832da8938008
          createdAt: "2024-11-08T11:06:33Z"
          description: Authorino
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta1
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: |-
          [Authorino](https://docs.kuadrant.io/authorino/) is Red Hat's Kubernetes-native lightweight external authorization service for tailor-made Zero Trust API security.

          Install this Red Hat official distribution of [Authorino Operator](https://docs.kuadrant.io/authorino-operator/) to manage instances of Authorino in this cluster.

          The current state of this distribution of the operator is: **_Technical Preview_**.

          The Community version of Authorino Operator, based on upstream public images, is available in [OperatorHub.io](https://operatorhub.io/operator/authorino-operator).

          ## Getting started

          After installing the Operator, request an instance of the external authorization service by creating an `Authorino` custom resource.

          **Minimal example (namespaced)**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            listener:
              tls:
                enabled: false
            oidcServer:
              tls:
                enabled: false
          ```

          **Extended example**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            clusterWide: true
            authConfigLabelSelectors: environment=production
            secretLabelSelectors: authorino.kuadrant.io/component=authorino,environment=production

            replicas: 2

            evaluatorCacheSize: 2 # mb

            logLevel: info
            logMode: production

            listener:
              ports:
                grpc: 50001
                http: 5001
              tls:
                certSecretRef:
                  name: authorino-server-cert # secret must contain `tls.crt` and `tls.key` entries
              timeout: 2

            oidcServer:
              port: 8083
              tls:
                certSecretRef:
                  name: authorino-oidc-server-cert # secret must contain `tls.crt` and `tls.key` entries

            metrics:
              port: 8080
              deep: true

            healthz:
              port: 8081

            tracing:
              endpoint: rpc://otel-collector.observability.svc.cluster.local:4317
              insecure: true

            volumes:
              items:
                - name: keycloak-tls-cert
                  mountPath: /etc/ssl/certs
                  configMaps:
                    - keycloak-tls-cert
                  items: # details to mount the k8s configmap in the authorino pods
                    - key: keycloak.crt
                      path: keycloak.crt
              defaultMode: 420
          ```

          ### Cluster-wide vs Namespaced

          Namespaced instances only watch auth resources (`AuthConfig` and `Secrets`) created in the same namespace as the Authorino service. Use this mode for dedicated instances that do not require elevated privileges.

          Cluster-wide instances watch resources across the entire cluster (all namespaces.) Deploying and running Authorino in this mode requires elevated privileges.

          ### Multi-tenancy

          Use the `authConfigLabelSelectors` field of the `Authorino` custom resource to narrow the scope of the Authorino instance.

          Only `AuthConfig` custom resources whose labels match the value of this field will be handled by the Authorino instance.

          ## Protect a host

          To protect a host, create an `AuthConfig` custom resource for the host. E.g.:

          ```yaml
          apiVersion: authorino.kuadrant.io/v1beta2
          kind: AuthConfig
          metadata:
            name: my-api-protection
          spec:
            hosts:
            - my-api.io

            authentication:
              "keycloak":
                jwt:
                  issuerUrl: https://keycloak.keycloak.svc.cluster.local:8080/realms/my-realm

            authorization:
              "k8s-rbac":
                kubernetesSubjectAccessReview:
                  user:
                    selector: auth.identity.user.username
                  resourceAttributes:
                    resource:
                      value: my-api
                    verb:
                      selector: request.method
                cache:
                  key:
                    selector: auth.identity.user.username
                  ttl: 30
              "after-2am-only":
                rego: |
                  allow {
                    [hour, _, _] := time.clock(time.now_ns())
                    hour >= 2
                  }
          ```

          Make sure all requests to the host are fisrt checked with the Authorino instance, by configuring an Envoy proxy for external authz:

          ```yaml
          clusters:
          - name: my-api
            …
          - name: authorino
            connect_timeout: 0.25s
            type: STRICT_DNS
            lb_policy: ROUND_ROBIN
            http2_protocol_options: {}
            load_assignment:
              cluster_name: authorino
              endpoints:
              - lb_endpoints:
                - endpoint:
                    address:
                      socket_address:
                        address: authorino-authorino-authorization
                        port_value: 50051
          listeners:
          - filter_chains:
            - filters:
              name: envoy.http_connection_manager
              typed_config:
                "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
                use_remote_address: true
                route_config:
                  name: my-api-config
                  virtual_hosts:
                  - name: my-api-vs
                    domains:
                    - my-api.io
                    routes:
                    - match:
                        prefix: /
                      route:
                        cluster: my-api
                http_filters:
                - name: envoy.filters.http.ext_authz
                  typed_config:
                    "@type": type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz
                    transport_api_version: V3
                    failure_mode_allow: false
                    include_peer_certificate: true
                    grpc_service:
                      envoy_grpc:
                        cluster_name: authorino
                      timeout: 1s
          ```

          ...or, if using Istio, by creating an [`AuthorizationPolicy`](https://istio.io/latest/docs/reference/config/security/authorization-policy/#AuthorizationPolicy) custom resource. Use `action: CUSTOM` in the resource and the Authorino authorization service configured in the mesh extension provider settings.

          ## Features

          **Authentication**

          * JWT validation (with OpenID Connect Discovery)
          * OAuth 2.0 Token Introspection (opaque tokens)
          * Kubernetes TokenReview (ServiceAccount tokens)
          * API key authentication
          * X.509 client certificate authentication
          * Anonymous access
          * Proxy-handled (authentication performed by the proxy)

          **Authorization**

          * Built-in simple pattern matching (e.g. JWT claims, request attributes checking)
          * OPA policies (inline Rego and fetch from external registry)
          * Kubernetes SubjectAccessReview (resource and non-resource attributes)
          * Authzed SpiceDB

          **External metadata**

          * HTTP request
          * OpenID Connect User Info
          * UMA-protected resource attributes

          **Custom responses**

          * Header injection (Festival Wristbands tokens, JSON, plain text)
          * Envoy Dynamic Metadata
          * Custom HTTP response (status code, headers, messages, body, etc)

          **Callbacks**

          * HTTP webhooks

          **Caching**

          * OpenID Connect and User-Managed Access configs
          * JSON Web Keys (JWKs) and JSON Web Key Sets (JWKS)
          * Access tokens
          * External metadata
          * Precompiled Rego policies
          * Policy evaluation

          Check out the full [Feature Specification](https://docs.kuadrant.io/authorino/docs/features/) and how-to guides in the [Kuadrant Docs](https://docs.kuadrant.io) website.
        displayName: Red Hat - Authorino (Technical Preview)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: GitHub
          url: https://github.com/kuadrant/authorino-operator
        - name: Documentation
          url: https://docs.kuadrant.io/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-tech-preview/authorino-rhel9@sha256:a473dae20e71e3e813ac30ba978f2ab3c5e19d7d39b501ae9103dca892107c87
        - registry.redhat.io/3scale-tech-preview/authorino-rhel9-operator@sha256:eb574781cf39917f8f5d9e5c806f520534601f819ee1ebcbb751832da8938008
        - registry.redhat.io/openshift4/ose-cli-rhel9@sha256:3a398158d10018bfac5df291d6998b513cf9a30e5477bcaefddcf66d229627c7
        version: 1.1.3
      entries:
      - name: authorino-operator.v1.1.3
        version: 1.1.3
      - name: authorino-operator.v1.1.2
        version: 1.1.2
      - name: authorino-operator.v1.1.1
        version: 1.1.1
      - name: authorino-operator.v1.1.0
        version: 1.1.0
      - name: authorino-operator.v1.0.2
        version: 1.0.2
      name: tech-preview-v1
    defaultChannel: stable
    packageName: authorino-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-athena-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-athena-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "athena.services.k8s.aws/v1alpha1",
                "kind": "WorkGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/athena-controller:1.1.1
          createdAt: "2025-09-19T21:06:02Z"
          description: AWS Athena controller is a service controller for managing
            Athena resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PreparedStatement represents the state of an AWS athena PreparedStatement
              resource.
            displayName: PreparedStatement
            kind: PreparedStatement
            name: preparedstatements.athena.services.k8s.aws
            version: v1alpha1
          - description: WorkGroup represents the state of an AWS athena WorkGroup
              resource.
            displayName: WorkGroup
            kind: WorkGroup
            name: workgroups.athena.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Athena resources in AWS from within your Kubernetes cluster.

          **About Amazon Athena**

          Amazon Athena is an interactive query service that makes it easy to analyze data directly in Amazon Simple Storage Service (Amazon S3) using standard SQL. With a few actions in the AWS Management Console, you can point Athena at your data stored in Amazon S3 and begin using standard SQL to run ad-hoc queries and get results in seconds.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Athena
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - athena
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Athena Developer Resources
          url: https://aws.amazon.com/athena/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: athena maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/athena-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-athena-controller.v1.1.1
        version: 1.1.1
      - name: ack-athena-controller.v1.1.0
        version: 1.1.0
      - name: ack-athena-controller.v1.0.14
        version: 1.0.14
      - name: ack-athena-controller.v1.0.13
        version: 1.0.13
      - name: ack-athena-controller.v1.0.12
        version: 1.0.12
      - name: ack-athena-controller.v1.0.11
        version: 1.0.11
      - name: ack-athena-controller.v1.0.10
        version: 1.0.10
      - name: ack-athena-controller.v1.0.9
        version: 1.0.9
      - name: ack-athena-controller.v1.0.8
        version: 1.0.8
      - name: ack-athena-controller.v1.0.7
        version: 1.0.7
      - name: ack-athena-controller.v1.0.6
        version: 1.0.6
      - name: ack-athena-controller.v1.0.5
        version: 1.0.5
      - name: ack-athena-controller.v1.0.4
        version: 1.0.4
      - name: ack-athena-controller.v1.0.1
        version: 1.0.1
      - name: ack-athena-controller.v1.0.0
        version: 1.0.0
      - name: ack-athena-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-athena-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Flux project
      provider-url: ""
    name: flux
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.toolkit.fluxcd.io/v2beta1",
                "kind": "HelmRelease",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "5m",
                    "chart": {
                      "spec": {
                          "chart": "podinfo",
                          "version": "4.0.x",
                          "sourceRef": {
                            "kind": "HelmRepository",
                            "name": "podinfo",
                            "namespace": "flux-system"
                          },
                          "interval": "1m"
                      }
                    },
                    "values": {
                      "replicaCount": 2
                    }
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "GitRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "1m",
                    "ref": {
                      "branch": "master"
                    },
                    "url": "https://github.com/stefanprodan/podinfo"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmRepository",
                "metadata": {
                    "name": "stable",
                    "namespace": "default"
                },
                "spec": {
                    "url": "https://kubernetes-charts.storage.googleapis.com/",
                    "interval": "10m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmChart",
                "metadata": {
                    "name": "redis",
                    "namespace": "default"
                },
                "spec": {
                    "chart": "redis",
                    "version": "10.5.7",
                    "sourceRef": {
                      "name": "stable",
                      "kind": "HelmRepository"
                    },
                    "interval": "5m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "Bucket",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "1m",
                    "provider": "generic",
                    "bucketName": "podinfo",
                    "endpoint": "minio.minio.svc.cluster.local:9000",
                    "insecure": true,
                    "secretRef": {
                      "name": "minio-credentials"
                    }
                }
              },
              {
                "apiVersion": "kustomize.toolkit.fluxcd.io/v1beta2",
                "kind": "Kustomization",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "5m",
                    "path": "./deploy/overlays/dev",
                    "prune": true,
                    "sourceRef": {
                      "kind": "GitRepository",
                      "name": "podinfo"
                    },
                    "timeout": "2m",
                    "wait": true
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "address": "https://github.com/stefanprodan/podinfo",
                    "secretRef": {
                      "name": "api-token"
                    }
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Alert",
                "metadata": {
                    "name": "webapp",
                    "namespace": "default"
                },
                "spec": {
                    "providerRef": {
                      "name": "on-call-slack"
                    },
                    "eventSeverity": "info",
                    "eventSources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "Bucket",
                          "name": "secrets"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-backend"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-frontend"
                      }
                    ]
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Receiver",
                "metadata": {
                    "name": "github-receiver",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "events": [
                      "ping",
                      "push"
                    ],
                    "secretRef": {
                      "name": "webhook-token"
                    },
                    "resources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "HelmRepository",
                          "name": "webapp"
                      }
                    ]
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "image": "ghcr.io/stefanprodan/podinfo",
                    "interval": "1m0s"
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImagePolicy",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "imageRepositoryRef": {
                      "name": "podinfo"
                    },
                    "policy": {
                      "semver": {
                          "range": "5.0.x"
                      }
                    }
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageUpdateAutomation",
                "metadata": {
                  "name": "flux-system",
                  "namespace": "flux-system"
                },
                "spec": {
                  "interval": "1m0s",
                  "sourceRef": {
                    "kind": "GitRepository",
                    "name": "flux-system"
                  },
                  "git": {
                    "checkout": {
                      "ref": {
                        "branch": "main"
                      }
                    },
                    "commit": {
                      "author": {
                        "email": "fluxcdbot@users.noreply.github.com",
                        "name": "fluxcdbot"
                      },
                      "messageTemplate": "{{range .Updated.Images}}{{println .}}{{end}}"
                    },
                    "push": {
                      "branch": "main"
                    }
                  },
                  "update": {
                    "path": "./clusters/my-cluster",
                    "strategy": "Setters"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/fluxcd/source-controller:v1.3.0
          description: Flux is a Continuous Delivery solution for Kubernetes.
          operatorframework.io/suggested-namespace: flux-system
          support: The Flux community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta3
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta1
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta2
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta1
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta2
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta2
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta2
          - description: OCIRepository
            displayName: OCIRepository
            kind: OCIRepository
            name: ocirepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta3
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta2
        description: |-
          [Flux](https://fluxcd.io/) is a Continuous Delivery solution for Kubernetes.
          Flux is constructed with the [GitOps Toolkit](https://toolkit.fluxcd.io/components/), a set of composable APIs and specialized tools for keeping Kubernetes clusters in sync with sources of configuration (like Git & Helm repositories), and automating updates to configuration when there is new code to deploy.

          Flux is a CNCF incubation project made for:

          - **cluster operators** who automate provision and configuration of clusters
          - **platform engineers** who build continuous delivery for developer teams
          - **app developers** who rely on continuous delivery to get their code live

          Features:

          - Source configuration from Git and Helm repositories, and S3-compatible buckets (e.g., Minio)
          - Kustomize and Helm support
          - Event-triggered and periodic reconciliation
          - Integration with Kubernetes RBAC
          - Health assessment (clusters and workloads)
          - Dependency management (infrastructure and workloads)
          - Secrets management (builtin support for Mozilla SOPS and Cloud KMS)
          - Alerting to external systems (webhook senders)
          - External events handling (webhook receivers)
          - Automated container image updates to Git (image scanning and patching)
          - Policy-driven validation (OPA, Kyverno, admission controllers)
          - Seamless integration with Git providers (GitHub, GitLab, Bitbucket)
          - Interoperability with Cluster API (CAPI) providers

          ### Steps to support bootstraping with Flux CLI

          SecComp Profile should be removed as it requires a special privilege to get the access.
          Also, it requires to enforce `runUserAs` to the same UID provided by the images to prevent OpenShift
          to alter the user id expected by our controllers.

          To do so, please specifiy a patch using the following Kustomization file, for example.

          ```yaml
          apiVersion: kustomize.config.k8s.io/v1beta1
          kind: Kustomization
          resources:
            - gotk-components.yaml
            - gotk-sync.yaml
          patches:
            - patch: |
                apiVersion: apps/v1
                kind: Deployment
                metadata:
                  name: all
                spec:
                  template:
                    spec:
                      containers:
                        - name: manager
                          securityContext:
                            runAsUser: 65534
                            seccompProfile:
                              $patch: delete
              target:
                kind: Deployment
                labelSelector: app.kubernetes.io/part-of=flux
          ```

          Then before installing Flux, please also perform the following steps to bind Flux to the `nonroot` SecurityContextConstraint (SCC). This applies to installation via OC's OLM or via Flux CLI.

          ```sh
          # replace default "flux-system" with the project/namespace name for your flux installation
          namespace="flux-system"
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:source-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:kustomize-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:helm-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:notification-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-automation-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-reflector-controller

          ```

          To see complete steps please refer to this [documentation](https://github.com/fluxcd/website/blob/main/content/en/docs/use-cases/openshift.md).

          If you've already installed Flux, you might need to uninstall it after applying the above pre-requisite steps.

          ### Network Policies

          This version of Flux is shipped without the normally provided network Policies.
          By default all traffic is enabled, and anyone can access the source-controller
          artifacts.

          It is not recommended to run it in production as is.

          Please visit [Flux's policies](https://github.com/fluxcd/flux2/tree/main/manifests/policies)
          to deploy the needed network policies and secure your Flux setup.

          ### Breaking Changes

          - Flux 0.18.x ships with v1beta2 of the kustomize.toolkit.fluxcd.io/Kustomization CRD.
          - Flux 0.25.x is the last Flux release where Kubernetes 1.19 is supported.

          ### Flux on OpenShift Use Cases

          - [The GitOps with Flux + OpenShift](https://www.youtube.com/watch?v=8ybCtNVP5To) talk by Andrew Block (Red Hat) showing how to setup multi-tenacy on OpenShift using Flux.
          - [Flux OpenShift Use Case](https://fluxcd.io/docs/use-cases/openshift/) on FluxCD.io
        displayName: Flux
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - gitops
        - cd
        - continuous delivery
        - fluxcd
        - helm
        - kustomize
        links:
        - name: Homepage
          url: https://fluxcd.io/
        - name: Documentation
          url: https://fluxcd.io/docs/
        - name: Slack
          url: https://cloud-native.slack.com/messages/flux
        - name: Mailing List
          url: https://lists.cncf.io/g/cncf-flux-dev
        - name: Source Code
          url: https://github.com/fluxcd/flux2
        maintainers:
        - email: support@weave.works
          name: The Flux maintainers
        maturity: stable
        provider:
          name: The Flux project
        relatedImages:
        - ghcr.io/fluxcd/helm-controller:v1.0.1
        - ghcr.io/fluxcd/image-automation-controller:v0.38.0
        - ghcr.io/fluxcd/image-reflector-controller:v0.32.0
        - ghcr.io/fluxcd/kustomize-controller:v1.3.0
        - ghcr.io/fluxcd/notification-controller:v1.3.0
        - ghcr.io/fluxcd/source-controller:v1.3.0
        version: 2.3.0
      entries:
      - name: flux.v2.3.0
        version: 2.3.0
      - name: flux.v2.2.2
        version: 2.2.2
      - name: flux.v2.2.0
        version: 2.2.0
      - name: flux.v2.0.1
        version: 2.0.1
      - name: flux.v0.41.2
        version: 0.41.2
      - name: flux.v0.40.0
        version: 0.40.0
      - name: flux.v0.39.0
        version: 0.39.0
      - name: flux.v0.38.3
        version: 0.38.3
      - name: flux.v0.37.0
        version: 0.37.0
      - name: flux.v0.36.0
        version: 0.36.0
      - name: flux.v0.35.0
        version: 0.35.0
      - name: flux.v0.31.4
        version: 0.31.4
      - name: flux.v0.31.3
        version: 0.31.3
      - name: flux.v0.30.2
        version: 0.30.2
      - name: flux.v0.28.5
        version: 0.28.5
      - name: flux.v0.28.4
        version: 0.28.4
      - name: flux.v0.28.3
        version: 0.28.3
      - name: flux.v0.27.4
        version: 0.27.4
      - name: flux.v0.27.3
        version: 0.27.3
      - name: flux.v0.27.2
        version: 0.27.2
      - name: flux.v0.27.0
        version: 0.27.0
      - name: flux.v0.26.1
        version: 0.26.1
      - name: flux.v0.26.0
        version: 0.26.0
      - name: flux.v0.25.3
        version: 0.25.3
      - name: flux.v0.25.2
        version: 0.25.2
      - name: flux.v0.24.1
        version: 0.24.1
      - name: flux.v0.24.0
        version: 0.24.0
      - name: flux.v0.23.0
        version: 0.23.0
      - name: flux.v0.22.1
        version: 0.22.1
      - name: flux.v0.22.0
        version: 0.22.0
      - name: flux.v0.21.1
        version: 0.21.1
      - name: flux.v0.21.0
        version: 0.21.0
      - name: flux.v0.20.1
        version: 0.20.1
      - name: flux.v0.20.0
        version: 0.20.0
      - name: flux.v0.19.1
        version: 0.19.1
      - name: flux.v0.18.3
        version: 0.18.3
      - name: flux.v0.18.1
        version: 0.18.1
      - name: flux.v0.17.2
        version: 0.17.2
      - name: flux.v0.17.1
        version: 0.17.1
      - name: flux.v0.17.0
        version: 0.17.0
      - name: flux.v0.16.2
        version: 0.16.2
      - name: flux.v0.16.1
        version: 0.16.1
      - name: flux.v0.16.0
        version: 0.16.0
      - name: flux.v0.15.3
        version: 0.15.3
      - name: flux.v0.15.2
        version: 0.15.2
      - name: flux.v0.14.2
        version: 0.14.2
      - name: flux.v0.14.1
        version: 0.14.1
      - name: flux.v0.14.0
        version: 0.14.0
      - name: flux.v0.13.4
        version: 0.13.4
      name: stable
    defaultChannel: stable
    packageName: flux
    provider:
      name: The Flux project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          olm.skipRange: <1.0.4
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
        version: 1.0.4
      entries:
      - name: datadog-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      name: stable
    defaultChannel: stable
    packageName: datadog-operator-certified
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kiali-ossm
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kiali-operator.v2.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "Kiali",
                "metadata": {
                  "name": "kiali"
                },
                "spec": {
                  "version": "default",
                  "istio_namespace": "istio-system",
                  "deployment": {
                    "logger": {
                      "log_level": "info"
                    },
                    "view_only_mode": false
                  }
                }
              },
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "OSSMConsole",
                "metadata": {
                  "name": "ossmconsole"
                },
                "spec": {
                  "version": "default",
                  "kiali": {
                    "serviceName": "",
                    "serviceNamespace": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9-operator@sha256:0612dfa52f2bf39a1d87e640409313585fedc7cfcff6d00e77bb14809d27d646
          createdAt: "2024-12-12T10:44:15Z"
          description: This productized operator provides Kiali and OSSMC. Kiali is
            the Istio observability and management Console. OSSMC is the OpenShift
            Service Mesh Console plugin, powered by Kiali.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.1.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kiali/kiali
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A configuration file for a Kiali installation.
            displayName: Kiali
            kind: Kiali
            name: kialis.kiali.io
            version: v1alpha1
          - description: A configuration file for a OpenShift Service Mesh Console
              installation.
            displayName: OpenShift Service Mesh Console
            kind: OSSMConsole
            name: ossmconsoles.kiali.io
            version: v1alpha1
        description: |-
          ## About the managed application

          A Microservice Architecture breaks up the monolith into many smaller pieces
          that are composed together. Patterns to secure the communication between
          services like fault tolerance (via timeout, retry, circuit breaking, etc.)
          have come up as well as distributed tracing to be able to see where calls
          are going.

          A service mesh can now provide these services on a platform level and frees
          the application writers from those tasks. Routing decisions are done at the
          mesh level.

          Kiali works with OpenShift Service Mesh to visualize the service
          mesh topology, to provide visibility into features like circuit breakers,
          request rates and more. It offers insights about the mesh components at
          different levels, from abstract Applications to Services and Workloads.

          See [https://www.kiali.io](https://www.kiali.io) to read more.

          ### Accessing the UI

          By default, the Kiali operator exposes the Kiali UI as an OpenShift Route.

          If on OpenShift, you can create an OSSMConsole CR to have the operator
          install the OpenShift ServiceMesh Console plugin to the OpenShift Console
          thus providing an interface directly integrated with the OpenShift Console.

          ## About this Operator

          ### Kiali Custom Resource Configuration Settings

          For quick descriptions of all the settings you can configure in the Kiali
          Custom Resource (CR), see
          [the kiali.io docs](https://kiali.io/docs/configuration/kialis.kiali.io/).

          ## Prerequisites for enabling this Operator

          Kiali is a companion tool for OpenShift Service Mesh. So before you install Kiali, you must have
          already installed Service Mesh.
        displayName: Kiali Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - service-mesh
        - observability
        - monitoring
        - istio
        links:
        - name: Installation Guide
          url: https://www.kiali.io/docs/installation
        - name: Features
          url: https://www.kiali.io/docs/features
        - name: Documentation Home
          url: https://www.kiali.io
        - name: Blogs and Articles
          url: https://medium.com/kialiproject
        - name: Server and UI Source Code
          url: https://github.com/kiali/kiali
        - name: Operator Source Code
          url: https://github.com/kiali/kiali-operator
        - name: OpenShift Service Mesh Console Code
          url: https://github.com/kiali/openshift-servicemesh-plugin
        maintainers:
        - email: kiali-dev@googlegroups.com
          name: Kiali Developers Google Group
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel8@sha256:e37d76dd7c7a2db79ff1e96afef19b07c3beaa8ebf0c7bd85b808908b8f0ce77
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9-operator@sha256:0612dfa52f2bf39a1d87e640409313585fedc7cfcff6d00e77bb14809d27d646
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9@sha256:06fde6c696a33d17b0d2f7620ec46422c82277fd17a3fe6d7319de05c732aace
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:976a31ca8cf17ac0887b8a586edfcc29eb30175802461b3114758af525e7b800
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:66e4d933a9d69158bf87dd45835183bf2308b77c48908ab64496c5d22c816293
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:27a6aebffa3d049ee48050126e6107db2e07c49e9ce9297175facb6368e1e792
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-ossmc-rhel9@sha256:99643f294c3f92a6fc82f321cab7c756e4879fefb48fa824f1bf49ff7182658d
        version: 2.1.2
      entries:
      - name: kiali-operator.v2.1.2
        version: 2.1.2
      - name: kiali-operator.v1.89.8
        version: 1.89.8
      - name: kiali-operator.v1.89.7
        version: 1.89.7
      - name: kiali-operator.v1.89.6
        version: 1.89.6
      - name: kiali-operator.v1.89.1
        version: 1.89.1
      - name: kiali-operator.v1.73.10
        version: 1.73.10
      - name: kiali-operator.v1.73.8
        version: 1.73.8
      - name: kiali-operator.v1.73.7
        version: 1.73.7
      - name: kiali-operator.v1.73.4
        version: 1.73.4
      - name: kiali-operator.v1.65.11
        version: 1.65.11
      - name: kiali-operator.v1.65.10
        version: 1.65.10
      - name: kiali-operator.v1.65.9
        version: 1.65.9
      - name: kiali-operator.v1.65.8
        version: 1.65.8
      - name: kiali-operator.v1.65.7
        version: 1.65.7
      - name: kiali-operator.v1.65.6
        version: 1.65.6
      - name: kiali-operator.v1.57.7
        version: 1.57.7
      name: candidate
    - currentCSV: kiali-operator.v2.11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "Kiali",
                "metadata": {
                  "name": "kiali"
                },
                "spec": {
                  "version": "default",
                  "istio_namespace": "istio-system",
                  "deployment": {
                    "logger": {
                      "log_level": "info"
                    },
                    "view_only_mode": false
                  }
                }
              },
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "OSSMConsole",
                "metadata": {
                  "name": "ossmconsole"
                },
                "spec": {
                  "version": "default",
                  "kiali": {
                    "serviceName": "",
                    "serviceNamespace": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/kiali-rhel9-operator@sha256:2970861bd8e65e36d5080dc4d56340d08a52dbb1217be352998d135eb854732f
          createdAt: "2025-09-10T00:28:49Z"
          description: This productized operator provides Kiali and OSSMC. Kiali is
            the Istio observability and management Console. OSSMC is the OpenShift
            Service Mesh Console plugin, powered by Kiali.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kiali/kiali
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A configuration file for a Kiali installation.
            displayName: Kiali
            kind: Kiali
            name: kialis.kiali.io
            version: v1alpha1
          - description: A configuration file for a OpenShift Service Mesh Console
              installation.
            displayName: OpenShift Service Mesh Console
            kind: OSSMConsole
            name: ossmconsoles.kiali.io
            version: v1alpha1
        description: |-
          ## About the managed application

          A Microservice Architecture breaks up the monolith into many smaller pieces
          that are composed together. Patterns to secure the communication between
          services like fault tolerance (via timeout, retry, circuit breaking, etc.)
          have come up as well as distributed tracing to be able to see where calls
          are going.

          A service mesh can now provide these services on a platform level and frees
          the application writers from those tasks. Routing decisions are done at the
          mesh level.

          Kiali works with OpenShift Service Mesh to visualize the service
          mesh topology, to provide visibility into features like circuit breakers,
          request rates and more. It offers insights about the mesh components at
          different levels, from abstract Applications to Services and Workloads.

          See [https://www.kiali.io](https://www.kiali.io) to read more.

          ### Accessing the UI

          By default, the Kiali operator exposes the Kiali UI as an OpenShift Route.

          If on OpenShift, you can create an OSSMConsole CR to have the operator
          install the OpenShift ServiceMesh Console plugin to the OpenShift Console
          thus providing an interface directly integrated with the OpenShift Console.

          ## About this Operator

          ### Kiali Custom Resource Configuration Settings

          For quick descriptions of all the settings you can configure in the Kiali
          Custom Resource (CR), see
          [the kiali.io docs](https://kiali.io/docs/configuration/kialis.kiali.io/).

          ## Prerequisites for enabling this Operator

          Kiali is a companion tool for OpenShift Service Mesh. So before you install Kiali, you must have
          already installed Service Mesh.
        displayName: Kiali Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - service-mesh
        - observability
        - monitoring
        - istio
        links:
        - name: Installation Guide
          url: https://www.kiali.io/docs/installation
        - name: Features
          url: https://www.kiali.io/docs/features
        - name: Documentation Home
          url: https://www.kiali.io
        - name: Blogs and Articles
          url: https://medium.com/kialiproject
        - name: Server and UI Source Code
          url: https://github.com/kiali/kiali
        - name: Operator Source Code
          url: https://github.com/kiali/kiali-operator
        - name: OpenShift Service Mesh Console Code
          url: https://github.com/kiali/openshift-servicemesh-plugin
        maintainers:
        - email: kiali-dev@googlegroups.com
          name: Kiali Developers Google Group
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9-operator@sha256:2970861bd8e65e36d5080dc4d56340d08a52dbb1217be352998d135eb854732f
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:7487fbb4b776dfca28ad09311ba857e07ad98bc91ca5ee6fcaaa5b7783d04fbd
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel9@sha256:9a0d4862a25c2999a3f85b9a7a0332dab4ec1744548191d7a00c838b06ba26d2
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel9@sha256:fe25be563d50e5ad4e5475f153e03d4b24c7db094d7e824c51ebaae79418b1f7
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel8@sha256:81db94df331799f4087596378914b712d85d117dde3df2d4902079a82d4a9c55
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9@sha256:a00ec00e466e080592dca5c1cc153698b1412a54e1979647f74281f79b5691cf
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9@sha256:85eeb972e46f26ff30e04991e76ec3050d98b334e7b7f201a094f8d50b06854e
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:d729ef8d74c2f6e02d8c68ca761b1fb3c7c3706f1290b684468f0acb0ef313e0
        version: 2.11.3
      entries:
      - name: kiali-operator.v2.11.3
        version: 2.11.3
      - name: kiali-operator.v2.11.2
        version: 2.11.2
      - name: kiali-operator.v2.11.1
        version: 2.11.1
      - name: kiali-operator.v2.4.7
        version: 2.4.7
      - name: kiali-operator.v2.4.6
        version: 2.4.6
      - name: kiali-operator.v2.4.5
        version: 2.4.5
      - name: kiali-operator.v2.4.4
        version: 2.4.4
      - name: kiali-operator.v2.4.3
        version: 2.4.3
      - name: kiali-operator.v1.89.10
        version: 1.89.10
      - name: kiali-operator.v1.89.9
        version: 1.89.9
      - name: kiali-operator.v1.89.8
        version: 1.89.8
      - name: kiali-operator.v1.89.7
        version: 1.89.7
      - name: kiali-operator.v1.89.6
        version: 1.89.6
      - name: kiali-operator.v1.89.1
        version: 1.89.1
      - name: kiali-operator.v1.73.10
        version: 1.73.10
      - name: kiali-operator.v1.73.8
        version: 1.73.8
      - name: kiali-operator.v1.73.7
        version: 1.73.7
      - name: kiali-operator.v1.73.4
        version: 1.73.4
      - name: kiali-operator.v1.65.11
        version: 1.65.11
      - name: kiali-operator.v1.65.10
        version: 1.65.10
      - name: kiali-operator.v1.65.9
        version: 1.65.9
      - name: kiali-operator.v1.65.8
        version: 1.65.8
      - name: kiali-operator.v1.65.7
        version: 1.65.7
      - name: kiali-operator.v1.65.6
        version: 1.65.6
      - name: kiali-operator.v1.57.7
        version: 1.57.7
      name: stable
    defaultChannel: stable
    packageName: kiali-ossm
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: test-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: test-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "test.openstack.org/v1beta1",
                "kind": "HorizonTest",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "test-operator",
                    "app.kubernetes.io/instance": "horizontest-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "horizontest",
                    "app.kubernetes.io/part-of": "test-operator"
                  },
                  "name": "horizontest-sample"
                },
                "spec": {
                  "adminPassword": "12345678",
                  "adminUsername": "admin",
                  "authUrl": "https://keystone-public-openstack.apps.ocp.openstack.lab",
                  "backoffLimit": 0,
                  "containerImage": "",
                  "dashboardUrl": "https://horizon-openstack.apps.ocp.openstack.lab/",
                  "flavorName": "m1.tiny",
                  "horizonRepoBranch": "master",
                  "horizonTestDir": "/var/lib/horizontest",
                  "imageUrl": "http://download.cirros-cloud.net/0.6.2/cirros-0.6.2-x86_64-disk.img",
                  "logsDirectoryName": "horizon",
                  "parallel": false,
                  "password": "horizontest",
                  "projectName": "horizontest",
                  "repoUrl": "https://review.opendev.org/openstack/horizon",
                  "storageClass": "local-storage",
                  "user": "horizontest"
                }
              },
              {
                "apiVersion": "test.openstack.org/v1beta1",
                "kind": "Tempest",
                "metadata": {
                  "name": "tempest-tests",
                  "namespace": "openstack"
                },
                "spec": {
                  "containerImage": "",
                  "tempestRun": {
                    "concurrency": 8,
                    "includeList": "tempest.api.identity.v3.*\n"
                  },
                  "tempestconfRun": null
                }
              },
              {
                "apiVersion": "test.openstack.org/v1beta1",
                "kind": "Tobiko",
                "metadata": {
                  "name": "tobiko-tests",
                  "namespace": "openstack"
                },
                "spec": {
                  "config": "[DEFAULT]\nlog_file = value\n\n[testcase]\ntimeout = value\ntest_runner_timeout = value\n\n[ubuntu]\ninterface_name = value\n\n[keystone]\ninterface_name = value\n",
                  "containerImage": "",
                  "testenv": "py3",
                  "version": "master",
                  "workflow": [
                    {
                      "stepName": "unit-tests",
                      "testenv": "py3"
                    },
                    {
                      "stepName": "functional-tests",
                      "testenv": "functional"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:43Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: Ansible Test
            kind: AnsibleTest
            name: ansibletests.test.openstack.org
            version: v1beta1
          - displayName: Horizon Test
            kind: HorizonTest
            name: horizontests.test.openstack.org
            version: v1beta1
          - displayName: Tempest
            kind: Tempest
            name: tempests.test.openstack.org
            version: v1beta1
          - kind: Tobiko
            name: tobikoes.test.openstack.org
            version: v1beta1
        description: Test Operator
        displayName: OpenStack Test Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Tempest
        - Tobiko
        - Tests
        links:
        - name: Test Operator
          url: https://github.com/openstack-k8s-operators/test-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/test-rhel9-operator@sha256:17949ad6a69a14ea38870d471fd4f5a51122927f795976ac7e2b0189e2d0da5a
        - registry.redhat.io/rhoso/openstack-tempest-all-rhel9@sha256:2f07239d932a67a6589844f4380f613cc16614ff4de7699f64c17359bbdc5534
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: test-operator.v1.0.6
        version: 1.0.6
      - name: test-operator.v1.0.5
        version: 1.0.5
      - name: test-operator.v1.0.4
        version: 1.0.4
      - name: test-operator.v1.0.3
        version: 1.0.3
      - name: test-operator.v1.0.2
        version: 1.0.2
      - name: test-operator.v1.0.1
        version: 1.0.1
      - name: test-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: test-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhsso-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhsso-operator.7.5.1-opr-005
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm"
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "False"
          containerImage: registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:66577ec32a9a975f304e52b8ab49b9f7602158dcab1ef5d8c28a5d3455491336
          createdAt: "2021-08-09 00:00:00"
          description: An Operator for installing and managing Red Hat Single Sign-On
          olm.skipRange: <7.5.0
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Red Hat Single Sign-On Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Red Hat Single Sign-On.

          Red Hat Single Sign-On lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
        displayName: Red Hat Single Sign-On Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://access.redhat.com/products/red-hat-single-sign-on
        - name: Keycloak
          url: https://www.keycloak.org/
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat Customer Service
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso7-rhel8-init-container@sha256:614a9b1d7975769e8559e0b8c1578d16f6e3f7fee0778283797f4d62f5fc2100
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:5f0840a69642d13016430151892ec987fd67e711563b1dfb1a57353037c3bf79
        - registry.redhat.io/rh-sso-7/sso75-openj9-openshift-rhel8@sha256:2d4ea7226f43e7e8e8d3fcef6d425eb5104a15acf07791355033300b80db2901
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:66577ec32a9a975f304e52b8ab49b9f7602158dcab1ef5d8c28a5d3455491336
        version: 7.5.1-opr-005
      entries:
      - name: rhsso-operator.7.5.1-opr-005
        version: 7.5.1-opr-005
      - name: rhsso-operator.7.5.1-opr-004
        version: 7.5.1-opr-004
      - name: rhsso-operator.7.5.1-opr-003
        version: 7.5.1-opr-003
      - name: rhsso-operator.7.5.1-opr-002
        version: 7.5.1-opr-002
      - name: rhsso-operator.7.5.1-opr-001
        version: 7.5.1-opr-001
      - name: rhsso-operator.7.5.0
        version: 7.5.0
      name: alpha
    - currentCSV: rhsso-operator.7.6.12-opr-004
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm"
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "False"
          containerImage: registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:229265579ec77f2bd18c936ba19e1f9dbc6b2f0a13d963d5eae5b52c260010d1
          createdAt: "2021-08-09 00:00:00"
          description: An Operator for installing and managing Red Hat Single Sign-On
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Red Hat Single Sign-On Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Red Hat Single Sign-On.

          Red Hat Single Sign-On lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
        displayName: Red Hat Single Sign-On Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://access.redhat.com/products/red-hat-single-sign-on
        - name: Keycloak
          url: https://www.keycloak.org/
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat Customer Service
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:229265579ec77f2bd18c936ba19e1f9dbc6b2f0a13d963d5eae5b52c260010d1
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:923206112ed0873ee58edd0817dbce7bb4c9a89dc75d9b3ba2affe9aed4811de
        - registry.redhat.io/rh-sso-7/sso7-rhel8-init-container@sha256:282daf78934ca495ad4051371a58e3af0917bb9241d043a66e23fc8b12fbd1ce
        version: 7.6.12-opr-004
      entries:
      - name: rhsso-operator.7.6.12-opr-004
        version: 7.6.12-opr-004
      - name: rhsso-operator.7.6.12-opr-003
        version: 7.6.12-opr-003
      - name: rhsso-operator.7.6.12-opr-002
        version: 7.6.12-opr-002
      - name: rhsso-operator.7.6.12-opr-001
        version: 7.6.12-opr-001
      - name: rhsso-operator.7.6.11-opr-006
        version: 7.6.11-opr-006
      - name: rhsso-operator.7.6.11-opr-005
        version: 7.6.11-opr-005
      - name: rhsso-operator.7.6.11-opr-004
        version: 7.6.11-opr-004
      - name: rhsso-operator.7.6.11-opr-003
        version: 7.6.11-opr-003
      - name: rhsso-operator.7.6.11-opr-002
        version: 7.6.11-opr-002
      - name: rhsso-operator.7.6.11-opr-001
        version: 7.6.11-opr-001
      - name: rhsso-operator.7.6.10-opr-001
        version: 7.6.10-opr-001
      - name: rhsso-operator.7.6.9-opr-003
        version: 7.6.9-opr-003
      - name: rhsso-operator.7.6.9-opr-002
        version: 7.6.9-opr-002
      - name: rhsso-operator.7.6.9-opr-001
        version: 7.6.9-opr-001
      - name: rhsso-operator.7.6.8-opr-002
        version: 7.6.8-opr-002
      - name: rhsso-operator.7.6.8-opr-001
        version: 7.6.8-opr-001
      - name: rhsso-operator.7.6.7-opr-002
        version: 7.6.7-opr-002
      - name: rhsso-operator.7.6.7-opr-001
        version: 7.6.7-opr-001
      - name: rhsso-operator.7.6.6-opr-004
        version: 7.6.6-opr-004
      - name: rhsso-operator.7.6.6-opr-003
        version: 7.6.6-opr-003
      - name: rhsso-operator.7.6.6-opr-002
        version: 7.6.6-opr-002
      - name: rhsso-operator.7.6.6-opr-001
        version: 7.6.6-opr-001
      - name: rhsso-operator.7.6.5-opr-004
        version: 7.6.5-opr-004
      - name: rhsso-operator.7.6.5-opr-003
        version: 7.6.5-opr-003
      - name: rhsso-operator.7.6.5-opr-002
        version: 7.6.5-opr-002
      - name: rhsso-operator.7.6.5-opr-001
        version: 7.6.5-opr-001
      - name: rhsso-operator.7.6.4-opr-003
        version: 7.6.4-opr-003
      - name: rhsso-operator.7.6.4-opr-002
        version: 7.6.4-opr-002
      - name: rhsso-operator.7.6.4-opr-001
        version: 7.6.4-opr-001
      - name: rhsso-operator.7.6.3-opr-002
        version: 7.6.3-opr-002
      - name: rhsso-operator.7.6.3-opr-001
        version: 7.6.3-opr-001
      - name: rhsso-operator.7.6.2-opr-003
        version: 7.6.2-opr-003
      - name: rhsso-operator.7.6.2-opr-002
        version: 7.6.2-opr-002
      - name: rhsso-operator.7.6.2-opr-001
        version: 7.6.2-opr-001
      - name: rhsso-operator.7.6.1-opr-005
        version: 7.6.1-opr-005
      - name: rhsso-operator.7.6.1-opr-004
        version: 7.6.1-opr-004
      - name: rhsso-operator.7.6.1-opr-003
        version: 7.6.1-opr-003
      - name: rhsso-operator.7.6.1-opr-002
        version: 7.6.1-opr-002
      - name: rhsso-operator.7.6.1-opr-001
        version: 7.6.1-opr-001
      - name: rhsso-operator.7.6.0-opr-005
        version: 7.6.0-opr-005
      - name: rhsso-operator.7.6.0-opr-004
        version: 7.6.0-opr-004
      - name: rhsso-operator.7.6.0-opr-003
        version: 7.6.0-opr-003
      - name: rhsso-operator.7.6.0-opr-002
        version: 7.6.0-opr-002
      - name: rhsso-operator.7.6.0-opr-001
        version: 7.6.0-opr-001
      - name: rhsso-operator.7.5.2-opr-003
        version: 7.5.2-opr-003
      - name: rhsso-operator.7.5.2-opr-002
        version: 7.5.2-opr-002
      - name: rhsso-operator.7.5.2-opr-001
        version: 7.5.2-opr-001
      - name: rhsso-operator.7.5.1-opr-011
        version: 7.5.1-opr-011
      - name: rhsso-operator.7.5.1-opr-010
        version: 7.5.1-opr-010
      - name: rhsso-operator.7.5.1-opr-009
        version: 7.5.1-opr-009
      - name: rhsso-operator.7.5.1-opr-009-0.1646836600.p
        version: 7.5.1-opr-009+0.1646836600.p
      - name: rhsso-operator.7.5.1-opr-008
        version: 7.5.1-opr-008
      - name: rhsso-operator.7.5.1-opr-007
        version: 7.5.1-opr-007
      - name: rhsso-operator.7.5.1-opr-006
        version: 7.5.1-opr-006
      - name: rhsso-operator.7.5.1-opr-005
        version: 7.5.1-opr-005
      - name: rhsso-operator.7.5.1-opr-004
        version: 7.5.1-opr-004
      - name: rhsso-operator.7.5.1-opr-003
        version: 7.5.1-opr-003
      - name: rhsso-operator.7.5.1-opr-002
        version: 7.5.1-opr-002
      - name: rhsso-operator.7.5.1-opr-001
        version: 7.5.1-opr-001
      - name: rhsso-operator.7.5.0
        version: 7.5.0
      name: stable
    defaultChannel: stable
    packageName: rhsso-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-glue-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-glue-controller.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/glue-controller:0.2.1
          createdAt: "2025-09-19T20:55:49Z"
          description: AWS Glue controller is a service controller for managing Glue
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Job represents the state of an AWS glue Job resource.
            displayName: Job
            kind: Job
            name: jobs.glue.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Glue resources in AWS from within your Kubernetes cluster.

          **About Amazon Glue**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Glue
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - glue
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Glue Developer Resources
          url: https://aws.amazon.com/Glue/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: glue maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/glue-controller:0.2.1
        version: 0.2.1
      entries:
      - name: ack-glue-controller.v0.2.1
        version: 0.2.1
      - name: ack-glue-controller.v0.2.0
        version: 0.2.0
      - name: ack-glue-controller.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ack-glue-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Protocol Labs
      provider-url: ""
    name: ipfs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ipfs-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.ipfs.io/v1alpha1",
                "kind": "CircuitRelay",
                "metadata": {
                  "name": "circuitrelay-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cluster.ipfs.io/v1alpha1",
                "kind": "IpfsCluster",
                "metadata": {
                  "name": "ipfscluster-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Storage, Networking, Big Data
          containerImage: quay.io/redhat-et-ipfs/ipfs-operator:v0.0.1
          createdAt: 2022-04-05T15:50:37-0400
          operatorframework.io/suggested-namespace: ipfs-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ipfs-cluster/ipfs-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CircuitRelay is the Schema for the circuitrelays API.
            displayName: Circuit Relay
            kind: CircuitRelay
            name: circuitrelays.cluster.ipfs.io
            version: v1alpha1
          - description: IpfsCluster is the Schema for the ipfs API.
            displayName: Ipfs Cluster
            kind: IpfsCluster
            name: ipfsclusters.cluster.ipfs.io
            version: v1alpha1
        description: "The IPFS Operator enables users to easily create IPFS Cluster
          instances\nwithin their Kubernetes or OpenShift clusters. \n\n### Supported
          Features\n* Horizontal scaling\n* Joining collaborative clusters\n* Controlling
          IPFS node resource usage\n* Specifying reprovider settings  \n\n### Documentation
          \nDocumentation can be found on our [read-the-docs page](https://ipfs-operator.readthedocs.io/en/latest/),
          as well as our [GitHub repository](https://github.com/ipfs-cluster/ipfs-operator).\n\n###
          Getting Help\nIn case you run into issues while using the operator, please
          create an issue in our [GitHub\nrepository](https://github.com/ipfs-cluster/ipfs-operator).\n\n###
          Contributing\nWe welcome contributions of all types and sizes! You can find
          open issues\nin our [GitHub Repository](https://github.com/ipfs-cluster/ipfs-operator/issues),\nas
          well as make your own issue if you feel that something is missing.\n\n###
          License\nIPFS Operator is licensed under the [Apache 2.0 License](https://github.com/ipfs-cluster/ipfs-operator/blob/main/LICENSE)"
        displayName: IPFS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipfs
        - ipfs cluster
        - decentralized storage
        - peer-to-peer
        - distributed hash table
        - content-addressed
        links:
        - name: IPFS Operator GitHub page
          url: https://github.com/ipfs-cluster/ipfs-operator
        - name: IPFS Cluster
          url: https://ipfscluster.io/
        - name: IPFS
          url: https://ipfs.io/
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Protocol Labs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/redhat-et-ipfs/ipfs-operator:v0.0.1
        version: 0.0.2
      entries:
      - name: ipfs-operator.v0.0.2
        version: 0.0.2
      - name: ipfs-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ipfs-operator
    provider:
      name: Protocol Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: gaudi-ai-sw-tools-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gaudi-ai-sw-tools-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aitools.intel/v1",
                "kind": "GaudiAIToolsContainer",
                "metadata": {
                  "name": "intel-gaudi-1.17.0-495"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagespec": {
                    "registry": "registry.connect.redhat.com",
                    "repo": "intel/gaudi-notebooks",
                    "tags": [
                      {
                        "gaudi_software": "1.17.0-495",
                        "rhel_os": "9.2"
                      }
                    ]
                  },
                  "nameOverride": "",
                  "namespace": "redhat-ods-applications"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: registry.connect.redhat.com/intel/gaudi-ai-sw-tools-operator@sha256:0c286f1e4cec57a189d749501f82345afbe0a1dbbbf92b906ad3c549e47b89cd
          createdAt: "2024-09-11T21:27:02Z"
          description: Operator to enable Gaudi AI SW Tools into OpenShift AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: GaudiAIToolsContainer
            name: gaudiaitoolscontainers.aitools.intel
            version: v1
        description: Operator to enable Gaudi AI SW Tools into OpenShift AI
        displayName: Intel® Gaudi AI SW Tools Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gaudi
        - opendatahub
        - openshiftai
        - notebooks
        links:
        - name: Intel® Gaudi AI SW Tools Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/66d7aa630eb66a02febc8103
        maintainers:
        - email: sharvil.shah@intel.com
          name: sharvil
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.connect.redhat.com/intel/gaudi-ai-sw-tools-operator@sha256:0c286f1e4cec57a189d749501f82345afbe0a1dbbbf92b906ad3c549e47b89cd
        version: 0.0.2
      entries:
      - name: gaudi-ai-sw-tools-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: gaudi-ai-sw-tools-operator
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-certified-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "psp": false,
                  "serviceAccount": "default",
                  "controller": {
                    "enabled": true,
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": "registry.connect.redhat.com/neuvector/controller",
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": ""
                      }
                    },
                    "pvc": {
                      "enabled": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "path": "/",
                          "annotations": {
                            "ingress.kubernetes.io/protocol": "https"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "path": "/",
                          "annotations": {
                            "ingress.kubernetes.io/protocol": "https"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "path": "/",
                      "annotations": {
                        "ingress.kubernetes.io/protocol": "https"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {}
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/enforcer",
                    "priorityClassName": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      }
                    ],
                    "resources": {}
                  },
                  "manager": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/manager",
                    "priorityClassName": "",
                    "env": {
                      "ssl": true
                    },
                    "svc": {
                      "type": "NodePort",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "path": "/",
                      "annotations": {},
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "affinity": {},
                    "tolerations": [],
                    "nodeSelector": {}
                  },
                  "cve": {
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "image": "registry.access.redhat.com/ubi8@sha256:228824aa581f3b31bf79411f8448b798291c667a37155bdea61cfa128b2833f2",
                      "schedule": "0 0 * * *",
                      "priorityClassName": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": "registry.connect.redhat.com/neuvector/scanner@sha256:42004e13b348bb43b58ae8538c983a70b237e3aca870758ac629fb4fc43e8130",
                      "priorityClassName": "",
                      "resources": {},
                      "affinity": {},
                      "tolerations": [],
                      "nodeSelector": {}
                    }
                  },
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "resources": {},
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "true"
          containerImage: registry.connect.redhat.com/neuvector/neuvector-operator@sha256:6155ac3c95aa6471c9857e0a6f5e5d19e6fc1d8cc7d014108d7d4a10d14f96bb
          createdAt: "2020-12-02 14:19:21"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/neuvector-certified-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/neuvector-certified-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: Neuvector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm/tree/master/charts/core) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart.  NeuVector Operator versions are tied to NeuVector product versions. Version 1.3.4 of the NeuVector Certified Operator deploys version 4.4.3 of NeuVector.

          **Complete below steps to Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Grant Service Account Access to the Privileged SCC

                   oc -n neuvector adm policy add-scc-to-user privileged -z default

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s                    neuvector/default


          **Add NeuVector license from NeuVector WebUI->setting**


          #Deploying the NeuVector Operator#


          Please refer to the instructions [here](https://github.com/neuvector/neuvector-operator/blob/master/README.md)


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Networking
        - Security
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - registry.connect.redhat.com/neuvector/scanner@sha256:42004e13b348bb43b58ae8538c983a70b237e3aca870758ac629fb4fc43e8130
        - registry.access.redhat.com/ubi8@sha256:228824aa581f3b31bf79411f8448b798291c667a37155bdea61cfa128b2833f2
        - registry.connect.redhat.com/neuvector/neuvector-operator@sha256:6155ac3c95aa6471c9857e0a6f5e5d19e6fc1d8cc7d014108d7d4a10d14f96bb
        - registry.connect.redhat.com/neuvector/controller@sha256:acc943568520282061547b310c50080dc32924a531bbb543d959b281eb6664f4
        - registry.connect.redhat.com/neuvector/enforcer@sha256:be1d299ec3532d55444681a882e9fd1238949da6598987308e2800ed26291b7a
        - registry.connect.redhat.com/neuvector/manager@sha256:8245cd179d11fff279690c5a422f46b4ff252ed72b9b998f0c9b5453bc7d8fc0
        version: 1.3.4
      entries:
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-certified-operator-rhmp
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dpu-network-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dpu-network-operator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dpu.openshift.io/v1alpha1",
                "kind": "OVNKubeConfig",
                "metadata": {
                  "name": "ovnkubeconfig-sample"
                },
                "spec": {
                  "kubeConfigFile": "tenant-cluster-1-kubeconf",
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/dpu-worker": ""
                    }
                  },
                  "poolName": "dpu"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional, Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/dpu-network-rhel8-operator@sha256:d07b8ecb5d1625599878b5c1e34b0de94188b81d90c829787a2d13969f6d5971
          createdAt: "2023-07-13T18:03:11Z"
          description: The operator is responsible for the life-cycle management of
            the ovn-kube components and the necessary host network initialization
            on DPU cards.
          olm.skipRange: '>=4.10.0-0 <4.13.0-202309181427'
          operatorframework.io/suggested-namespace: openshift-dpu-network-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/dpu-network-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OVNKubeConfig is the Schema for the ovnkubeconfigs API
            displayName: OVNKube Config
            kind: OVNKubeConfig
            name: ovnkubeconfigs.dpu.openshift.io
            version: v1alpha1
        description: The operator to be responsible for the life-cycle management
          of the ovn-kube components and the necessary host network initialization
          on DPU cards.
        displayName: DPU Network Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dpu
        - network
        links:
        - name: Dpu Network Operator
          url: https://dpu-network-operator.domain
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:767682dd3bd65651a8204c9fb732b9c48b99127189992b43abc6ecce027f4589
        - registry.redhat.io/openshift4/dpu-network-rhel8-operator@sha256:d07b8ecb5d1625599878b5c1e34b0de94188b81d90c829787a2d13969f6d5971
        version: 4.13.0-202309181427
      entries:
      - name: dpu-network-operator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: stable
    defaultChannel: stable
    packageName: dpu-network-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-odlm
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operand-deployment-lifecycle-manager.v4.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandBindInfo",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "bindings": {
                    "public": {
                      "configmap": "mongodb-configmap",
                      "secret": "mongodb-secret"
                    }
                  },
                  "description": "Binding information that should be accessible to MongoDB adopters",
                  "operand": "mongodb-atlas-kubernetes",
                  "registry": "example-service"
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "services": [
                    {
                      "name": "jaeger",
                      "spec": {
                        "jaeger": {
                          "strategy": "allinone"
                        }
                      }
                    },
                    {
                      "name": "mongodb-atlas-kubernetes",
                      "spec": {
                        "atlasDeployment": {
                          "deploymentSpec": {
                            "name": "test-deployment"
                          },
                          "projectRef": {
                            "name": "my-project"
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "operators": [
                    {
                      "channel": "stable",
                      "installMode": "cluster",
                      "name": "jaeger",
                      "namespace": "default",
                      "packageName": "jaeger",
                      "sourceName": "community-operators",
                      "sourceNamespace": "openshift-marketplace"
                    },
                    {
                      "channel": "stable",
                      "name": "mongodb-atlas-kubernetes",
                      "namespace": "default",
                      "packageName": "mongodb-atlas-kubernetes",
                      "sourceName": "community-operators",
                      "sourceNamespace": "openshift-marketplace"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandRequest",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "requests": [
                    {
                      "operands": [
                        {
                          "name": "jaeger"
                        },
                        {
                          "name": "mongodb-atlas-kubernetes"
                        }
                      ],
                      "registry": "example-service"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, Monitoring, Logging & Tracing, Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
          createdAt: "2025-02-14T21:16:49Z"
          description: The Operand Deployment Lifecycle Manager provides a Kubernetes
            CRD-based API to manage the lifecycle of operands.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-odlm
          olm.skipRange: '>=1.2.0 <4.5.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/operand-deployment-lifecycle-manager
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OperandBindInfo is the Schema for the operandbindinfoes API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandBindInfo
            kind: OperandBindInfo
            name: operandbindinfos.operator.ibm.com
            version: v1alpha1
          - description: OperandConfig is the Schema for the operandconfigs API. Documentation
              For additional details regarding install parameters check https://ibm.biz/icpfs39install.
              License By installing this product you accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandConfig
            kind: OperandConfig
            name: operandconfigs.operator.ibm.com
            version: v1alpha1
          - description: OperandRegistry is the Schema for the operandregistries API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandRegistry
            kind: OperandRegistry
            name: operandregistries.operator.ibm.com
            version: v1alpha1
          - description: OperandRequest is the Schema for the operandrequests API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandRequest
            kind: OperandRequest
            name: operandrequests.operator.ibm.com
            version: v1alpha1
          - description: OperatorConfig is the Schema for the operatorconfigs API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.ibm.com
            version: v1alpha1
        description: |-
          # Introduction

          The operand is the instance managed by the operator. ODLM is used to manage the lifecycle of for a group of operands, compared with operator lifecycle manager, ODLM focus on the management of operands but not operators.

          # Details

          ### What are operands?

          Operator is a method of packaging, deploying and managing a Kubernetes application.
          Operands are the services and applications that Operator manage.

          For example, cert-manager operator deploys a cert-manager deployment, then the cert-manager-operator is an operator and the cert-manager deployment is an operand.

          The ODLM will have four CRDs:

          | Resource                 | Short Name | Description                                                                                |
          |--------------------------|------------|--------------------------------------------------------------------------------------------|
          | OperandRequest | opreq | It defines which operator/operand want to be installed in the cluster |
          | OperandRegistry | opreg | It defines the OLM information, like channel and catalog source, for each operator |
          | OperandConfig | opcon | It defines the parameters that should be used to install the operator's operand |
          | OperandBindInfo | opbi | It identifies secrets and/or configmaps that should be shared with requests |

          ### Goal

          1. A single entry point to manage a group of operands
          1. User can select a set of operands to install
          1. The install can be invoked through either OCP UI or CLI

          ## Prerequisites
          - [operator-sdk] version v1.3.0.
          - [go] version 1.15.7+
          - [oc] version v3.11+ or [kubectl] v1.11.3+
          - Access to an Openshift v4.3+ cluster

          ### Resources Required

          #### Minimum scheduling capacity

          | Software                   | Memory (MB) | CPU (cores) |
          | -------------------------- | ----------- | ----------- |
          | Operand Deployment Lifecycle Manager | 200          | 0.2        |
          | **Total**                  | 200         | 0.2         |

          ## Installing

          ### Create Operator NS, Group, Subscription

          ```yaml
          apiVersion: v1
          kind: Namespace
          metadata:
            name: odlm

          ---
          apiVersion: operators.coreos.com/v1alpha2
          kind: OperatorGroup
          metadata:
            name: operatorgroup
            namespace: odlm
          spec:
            targetNamespaces:
            - odlm

          ---
          apiVersion: v1
          data:
            namespaces: odlm
          kind: ConfigMap
          metadata:
            name: namespace-scope
            namespace: odlm

          ---
          apiVersion: operators.coreos.com/v1alpha1
          kind: Subscription
          metadata:
            name: operand-deployment-lifecycle-manager
            namespace: odlm
          spec:
            channel: v3.20
            name: ibm-odlm
            source: opencloud-operators
            sourceNamespace: olm
            config:
              env:
              - name: INSTALL_SCOPE
                value: namespaced
          END
          ```

          ## Configuration

          OperandRegistry defines the OLM information used for installation, like package name and catalog source, for each operator.

          Following is an example of the OperandRegistry CR:

          **NOTE:** The "spec.operators[*].name" parameter must be unique for each entry.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRegistry
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            operators:
            - name: jenkins [3]
              namespace: default [4]
              channel: alpha [5]
              packageName: jenkins-operator [6]
              scope: public [7]
              sourceName: community-operators [8]
              sourceNamespace: openshift-marketplace [9]
              installMode: cluster [10]
              installPlanApproval: Manual [11]
          ```

          The OperandRegistry Custom Resource (CR) lists OLM Operator information for operands that may be requested for installation and/or access by an application that runs in a namespace. The registry CR specifies:

          1. `name` of the OperandRegistry
          2. `namespace` of the OperandRegistry
          3. `name` is the name of the operator, which should be the same as the services name in the OperandConfig and OperandRequest.
          4. `namespace` defines the namespace where the operator and its CR will be deployed. (1) When InstallMode is `cluster`, the operator will be deployed into the `openshift-operators` namespace and the operator CRs will be deployed into the namespace this parameter defines. (2) When InstallMode is empty or set to `namespace`, it is the namespace where both operator and operator CR will be deployed.
          5. `channel` is the name of OLM channel that is subscribed for the operator.
          6. `packageName` is the name of the package in CatalogSource that is subscribed for the operator.
          7. (optional) `scope` is an indicator, either public or private, that dictates whether deployment can be requested from other namespaces (public) or only from the namespace of this OperandRegistry (private). The default value is private.
          8. `sourceName` is the name of the CatalogSource.
          9. `sourceNamespace` is the namespace of the CatalogSource.
          10. (optional) `installMode` is the install mode of the operator, can be either `namespace` (OLM one namespace) or `cluster` (OLM all namespaces). The default value is `namespace`. Operator is deployed in `openshift-operators` namespace when InstallMode is set to `cluster`.
          11. (optional) `installPlanApproval` is the approval mode for emitted installplan. The default value is `Automatic`.

          ### OperandConfig Spec

          OperandConfig defines the individual operand configuration. The OperandConfig Custom Resource (CR) defines the parameters for each operator that is listed in the OperandRegistry that should be used to install the operator instance by specifying an installation CR.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          Kind: OperandConfigs
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            services:
            - name: jenkins [3]
              spec: [4]
                jenkins:
                  port: 8081
          ```

          OperandConfig defines the individual operand deployment config:

          1. `name` of the OperandConfig
          2. `namespace` of the OperandConfig
          3. `name` is the name of the operator, which should be the same as the services name in the OperandRegistry and OperandRequest.
          4. `spec` defines a map. Its key is the kind name of the custom resource. Its value is merged to the spec field of custom resource. For more details, you can check the following topic **How does ODLM create the individual operator CR?**

          #### How does Operator create the individual operator CR

          Jenkins Operator has one CRD: Jenkins:

          The OperandConfig CR has

          ```yaml
          - name: jenkins
            spec:
              jenkins:
                service:
                  port: 8081
          ```

          The Jenkins Operator CSV has

          ```yaml
          apiVersion: operators.coreos.com/v1alpha1
          kind: ClusterServiceVersion
          metadata:
            annotations:
            alm-examples: |-
              [
              {
                "apiVersion":"jenkins.io/v1alpha2",
                "kind":"Jenkins",
                "metadata": {
                  "name":"example"
                },
                "spec":{
                  ...
                  "service":{"port":8080,"type":"ClusterIP"},
                  ...
                }
              }
            ]
          ```

          The ODLM will deep merge the OperandConfig CR spec and Jenkins Operator CSV alm-examples to create the Jenkins CR.

          ```yaml
          apiVersion: jenkins.io/v1alpha2
          kind: Jenkins
          metadata:
            name: example
          spec:
            ...
            service:
              port: 8081
              type: ClusterIP
            ...
          ```

          For day2 operations, the ODLM will patch the OperandConfigs CR spec to the existing Jenkins CR.

          ### OperandRequest Spec

          OperandRequest defines which operator/operand you want to install in the cluster.

          **NOTE:** OperandRequest custom resource is used to trigger a deployment for Operators and Operands.

          #### OperandRequest sample to create custom resource via OperandConfig

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRequest
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            requests:
            - registry: example-service [3]
              registryNamespace: example-service-ns [4]
              operands: [5]
              - name: jenkins [6]
                bindings: [7]
                  public:
                    secret: jenkins-operator-credential [8]
                    configmap: jenkins-operator-base-configuration [9]
          ```

          1. `name` of the OperandRequest
          2. `namespace` of the OperandRequest
          3. `registry` identifies the name of the OperandRegistry CR from which this operand deployment is being requested.
          4. (optional) `registryNamespace` identifies the namespace in which the OperandRegistry CR is defined. **Note:** If the `registryNamespace` is not specified then it is assumed that the OperandRegistry CR is in the current (OperandRequest's) namespace.
          5. `operands` in the CR is a list of operands.
          6. `name` of operands in the CR must match a name specification in an OperandRegistry's CR.
          7. (optional) The `bindings` of the operands is a map to get and rename the secret and/or configmap from the provider and create them in the requester's namespace. If the requester wants to rename the secret and/or configmap, they need to know the key of the binding in the OperandBindInfo. If the key of the bindings map is prefixed with public, it means the secret and/or configmap can be shared with the requester in the other namespace. If the key of the bindings map is prefixed with private, it means the secret and/or configmap can only be shared within its own namespace.
          8. (optional) `secret` names a secret that should be created in the requester's namespace with formatted data that can be used to interact with the service.
          9. (optional) `configmap` names a configmap that should be created in the requester's namespace with formatted data that can be used to interact with the service.

          #### OperandRequest sample to create custom resource via OperandRequest

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRequest
          metadata:
            name: example-service
            namespace: example-service-ns
          spec:
            requests:
            - registry: example-service
              registryNamespace: example-service-ns
              operands:
              - name: jenkins
                kind: Jenkins [1]
                apiVersion: "jenkins.io/v1alpha2" [2]
                instanceName: "example" [3]
                spec: [4]
                  service:
                    port: 8081
          ```

          1. `kind` of the target custom resource. If it is set in the operand item, ODLM will create CR according to OperandRequest only and get rid of the alm-example and OperandConfig.
          2. `apiVersion` of the target custom resource.
          3. `instanceName` is the name of the custom resource. If `instanceName` is not set, the name of the custom resource will be created with the name of the OperandRequest as a prefix.
          4. `spec` is the spec field of the target CR.

          ### OperandBindInfo Spec

          The ODLM will use the OperandBindInfo to copy the generated secret and/or configmap to a requester's namespace when a service is requested with the OperandRequest CR. An example specification for an OperandBindInfo CR is shown below.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandBindInfo
          metadata:
            name: publicjenkinsbinding [1]
            namespace: example-service-ns [2]
          spec:
            operand: jenkins [3]
            registry: example-service [4]
            description: "Binding information that should be accessible to jenkins adopters" [5]
            bindings: [6]
              public:
                secret: jenkins-operator-credentials-example [7]
                configmap: jenkins-operator-base-configuration-example [8]
          ```

          Fields in this CR are described below.

          1. `name` of the OperandBindInfo
          2. `namespace` of the OperandBindInfo
          3. The `operand` should be the the individual operator name.
          4. The `registry` section must match the name in the OperandRegistry CR in the current namespace.
          5. `description` is used to add a detailed description of a service.
          6. The `bindings` section is used to specify information about the access/configuration data that is to be shared. If the key of the bindings map is prefixed with public, it means the secret and/or configmap can be shared with the requester in the other namespace. If the key of the bindings map is prefixed with private, it means the secret and/or configmap can only be shared within its own namespace. If the key of the bindings map is prefixed with protected, it means the secret and/or configmap can only be shared if it is explicitly declared in the OperandRequest.
          7. The `secret` field names an existing secret, if any, that has been created and holds information that is to be shared with the requester.
          8. The `configmap` field identifies a configmap object, if any, that should be shared with the requester

          ODLM will use the OperandBindInfo CR to pass information to an adopter when they create a OperandRequest to access the service, assuming that both have compatible scopes. ODLM will copy the information from the shared service's "OperandBindInfo.bindinfo[].secret" and/or "OperandBindInfo.bindinfo[].configmap" to the requester namespace.

          **NOTE:** If in the OperandRequest, there is no secret and/or configmap name specified in the bindings or no bindings field in the element of operands, ODLM will copy the secret and/or configmap to the requester's namespace and rename them to the name of the OperandBindInfo + secret/configmap name.

          ### E2E Use Case

          1. User installs ODLM from OLM

              The ODLM will automatically generate two default CRD CRs, since OperandRegistry and OperandConfigs CRs don't define the state, so it should be fine.

          2. **Optional:** Users can update the OperandConfigs CR with their own parameter values

              Users can skip this step if the default configurations match the requirement.

          3. User creates the OperandRequest CR from OLM

              This tells ODLM that users want to install some of the individual operator/operand.

          4. The rest works will be done by OLM and ODLM

              Finally, users will get what they want.

          ### Operator/Operand Upgrade

          - For operator/operand upgrade, you only need to publish your operator OLM to your operator channel, and OLM will handle the upgrade automatically.
          - If there are major version, then you may want to update `channel` in `OperandRegistry` to trigger upgrade.

          ### Limitations

          ODLM depends on NamespaceScope Operator

          ### SecurityContextConstraints Requirements

          The Operand Deployment Lifecycle Manager supports running under the OpenShift Container Platform default restricted security context constraints.
        displayName: Operand Deployment Lifecycle Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Operand Deployment Lifecycle Manager Project
          url: https://github.com/IBM/operand-deployment-lifecycle-manager
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d4dff1a8968c18d984e16c4c9f1517c3a10e23a11bf166c64d734773324d4c9d
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
        - registry.marketplace.redhat.com/cpopen/operand-deployment-lifecycle-manager-bundle@sha256:68f71f2767a40ca9333f0a670e07493e8a8c13a5c7ca877ea5bd8a544b352d9a
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:27195e9758afdf8521971866fd60ea3dafed91a146c3c4165876e5ea476a4622
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bb074bdf89cbc7c703fadb17639608599ab4f66db9be5617e5f79a0550676790
        version: 4.5.0
      entries:
      - name: operand-deployment-lifecycle-manager.v4.5.0
        version: 4.5.0
      name: v4.5
    defaultChannel: v4.5
    packageName: ibm-odlm
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fujitsu
      provider-url: https://www.postgresql.fastware.com/
    name: fujitsu-enterprise-postgres-operator-with-cg-module
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPAction",
                "metadata": {
                  "name": "new-fep-action"
                },
                "spec": {
                  "fepAction": {
                    "args": [
                      "new-fep-sts-0",
                      "new-fep-sts-1"
                    ],
                    "type": "reload"
                  },
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "targetClusterName": "new-fep"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPExporter",
                "metadata": {
                  "name": "new-fep-exporter"
                },
                "spec": {
                  "fepExporter": {
                    "exporterLogLevel": "error",
                    "fepClusterList": [
                      "new-fep1"
                    ],
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "userCustomQueries": "usr_example:\n  query: \"SELECT EXTRACT(EPOCH FROM (now() - pg_last_xact_replay_timestamp())) as lag\"\n  master: true\n  metrics:\n    - lag:\n        usage: \"GAUGE\"\n        description: \"Replication lag behind master in seconds\""
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPLogging",
                "metadata": {
                  "name": "new-fep-logging"
                },
                "spec": {
                  "fepLogging": {
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPPgpool2",
                "metadata": {
                  "name": "new-fep-pgpool2"
                },
                "spec": {
                  "clientAuthMethod": "scram",
                  "count": 2,
                  "customhba": "local   all         all                               trust\nhost    all         all         127.0.0.1/32          trust\nhost    all         all         ::1/128               trust\n",
                  "customlogsize": "128Mi",
                  "customparams": "listen_addresses = '*'\npcp_listen_addresses = '*'\nnum_init_children = 32\nreserved_connections = 0\nenable_pool_hba = off\nallow_clear_text_frontend_auth = off\nauthentication_timeout = 80\nbackend_weight0 = 1\nbackend_weight1 = 1\nbackend_flag0 = 'ALWAYS_PRIMARY'\nbackend_flag1 = 'DISALLOW_TO_FAILOVER'\nconnection_cache = on\nmax_pool = 4\nlisten_backlog_multiplier = 2\nserialize_accept = off\nchild_life_time = 300\nclient_idle_limit = 0\nchild_max_connections = 0\nconnection_life_time = 0\nreset_query_list = 'ABORT; DISCARD ALL'\nclient_min_messages = info\nlog_min_messages = debug1\nlog_statement = on\nlog_per_node_statement = on\nlog_client_messages = on\nlog_hostname = on\nlog_connections = on\nlog_line_prefix = '%t: pid %p: '\nload_balance_mode = on\nignore_leading_white_space = on\nwhite_function_list = ''\nblack_function_list = 'currval,lastval,nextval,setval'\nblack_query_pattern_list = ''\ndatabase_redirect_preference_list = ''\napp_name_redirect_preference_list = ''\nallow_sql_comments = off\ndisable_load_balance_on_write = 'transaction'\nstatement_level_load_balance = on\nsr_check_period = 0\nsr_check_user = 'postgres'\ndelay_threshold = 0\nlog_standby_delay = 'none'\nssl = on\nssl_ciphers = 'HIGH:MEDIUM:+3DES:!aNULL'\nssl_prefer_server_ciphers = off\nssl_ecdh_curve = 'prime256v1'\nssl_dh_params_file = ''\nrelcache_expire = 0\nrelcache_size = 256\ncheck_temp_table = catalog\ncheck_unlogged_table = on\nenable_shared_relcache = off\nrelcache_query_target = primary\nwd_port0 = 9000\nfailover_on_backend_error = off\n",
                  "custompcp": "none",
                  "customsslcacert": "none",
                  "customsslcert": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                  "customsslkey": "none",
                  "fepclustername": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "limits": {
                    "cpu": "400m",
                    "memory": "512Mi"
                  },
                  "requests": {
                    "cpu": "200m",
                    "memory": "256Mi"
                  },
                  "serviceport": 9999,
                  "statusport": 9898
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPRestore",
                "metadata": {
                  "name": "new-fep-restore"
                },
                "spec": {
                  "fromFEPcluster": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "mcSpec": {
                    "limits": {
                      "cpu": "200m",
                      "memory": "300Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "restoretype": "latest",
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "toFEPcluster": "new-fep-2"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v2",
                "kind": "FEPCluster",
                "metadata": {
                  "name": "new-fep"
                },
                "spec": {
                  "fep": {
                    "customAnnotations": {
                      "allDeployments": {}
                    },
                    "forceSsl": true,
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "instances": 1,
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "podAntiAffinity": false,
                    "podDisruptionBudget": false,
                    "servicePort": 27500,
                    "syncMode": "off",
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "usePodName": true
                  },
                  "fepChildCrVal": {
                    "backup": {
                      "image": {
                        "pullPolicy": "IfNotPresent"
                      },
                      "mcSpec": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      },
                      "pgbackrestParams": "# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[global]\nrepo1-retention-full=7\nrepo1-retention-full-type=time\nlog-path=/database/log/backup\n",
                      "postScript": " ",
                      "preScript": " ",
                      "schedule": {
                        "num": 2
                      },
                      "schedule1": {
                        "schedule": "15 0 * * 0",
                        "type": "full"
                      },
                      "schedule2": {
                        "schedule": "15 0 * * 1-6",
                        "type": "incr"
                      },
                      "schedule3": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule4": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule5": {
                        "schedule": " ",
                        "type": " "
                      }
                    },
                    "customPgAudit": "# define pg audit custom params here to override defaults.\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[output]\nlogger = 'auditlog'\nlog_directory = '/database/log/audit'\nlog_truncate_on_rotation = on\nlog_filename = 'pgaudit-%a.log'\nlog_rotation_age = 1d\nlog_rotation_size = 0\n[rule]\n",
                    "customPgHba": "# define pg_hba custom rules here to be merged with default rules.\n# TYPE     DATABASE        USER        ADDRESS        METHOD\n",
                    "customPgParams": "# define custom postgresql.conf parameters below to override defaults.\n# Current values are as per default FEP deployment\n# If you add a library to shared_preload_libraries, add it after the default value.\n# fsep_operator_security is available in FEP15 and later container images.\nshared_preload_libraries='pgx_datamasking,pg_prewarm,pg_stat_statements,fsep_operator_security'\nsession_preload_libraries='pg_prewarm'\nmax_prepared_transactions = 100\nmax_worker_processes = 30\nmax_connections = 100\nwork_mem = 1MB\nmaintenance_work_mem = 12MB\nshared_buffers = 128MB\neffective_cache_size = 384MB\ncheckpoint_completion_target = 0.8\n# tcp parameters\ntcp_keepalives_idle = 30\ntcp_keepalives_interval = 10\ntcp_keepalives_count = 3\n# logging parameters in default fep installation\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\nlog_directory = '/database/log'\nlog_filename = 'logfile-%a.log'\nlog_file_mode = 0600\nlog_truncate_on_rotation = on\nlog_rotation_age = 1d\nlog_rotation_size = 0\nlog_checkpoints = on\nlog_line_prefix = '%e %t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h'\nlog_lock_waits = on\nlog_autovacuum_min_duration = 60s\nlogging_collector = on\npgaudit.config_file='/opt/app-root/src/pgaudit-cfg/pgaudit.conf'\nlog_replication_commands = on\nlog_min_messages = WARNING\nlog_destination = csvlog\n# wal_archive parameters in default fep installation\narchive_mode = on\narchive_command = 'pgbackrest --stanza=backupstanza --config=/database/userdata/pgbackrest.conf archive-push %p'\nwal_level = replica\nmax_wal_senders = 12\nwal_keep_segments = 64\ntrack_activities = on\ntrack_counts = on\npassword_encryption = 'scram-sha-256'\n",
                    "storage": {
                      "archivewalVol": {
                        "size": "1Gi"
                      },
                      "backupVol": {
                        "size": "2Gi"
                      },
                      "dataVol": {
                        "size": "2Gi"
                      },
                      "logVol": {
                        "size": "1Gi"
                      },
                      "tablespaceVol": {
                        "size": "512Mi"
                      },
                      "walVol": {
                        "size": "1200Mi"
                      }
                    },
                    "sysUsers": {
                      "pgRewindPassword": "rewind_password",
                      "pgRewindUser": "rewind_user",
                      "pgdb": "mydb",
                      "pgpassword": "mydbpassword",
                      "pgreplpassword": "repluserpwd",
                      "pgrepluser": "repluser",
                      "pguser": "mydbuser",
                      "tdepassphrase": "tde-passphrase"
                    },
                    "systemCertificates": {
                      "cacrt": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                      "crt": "-----BEGIN CERTIFICATE-----\nMIIDUTCCAjmgAwIBAgIRAMocW3qMoHrD6qRvMPppMkMwDQYJKoZIhvcNAQELBQAw\nNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9yIEt1\nYmVybmV0ZXMwHhcNMjEwMjA2MDQzMjM2WhcNMjYwMjA1MDQzMjM2WjA/MRAwDgYD\nVQQKEwdGdWppdHN1MSswKQYDVQQDEyJGVUpJVFNVIEVudGVycHJpc2UgUG9zdGdy\nZXMgU2VydmVyMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA4AI33yvH\nZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I2e4SceTKi6O3C/I1XuvWlpng\n5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4ANB5JWWqDOjrRT3o7nRPGXfila\nbP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYpmjdbfxabTz07ig0+6/cwKoRR\nxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTFYvmAH7gcdssSFBt8NPlUATHE\nsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6Wdgmu5H2pDml8CDNLDv98Aj7i\n+I5SRKKcVPlnuQIDAQABo1AwTjAdBgNVHSUEFjAUBggrBgEFBQcDAQYIKwYBBQUH\nAwIwDAYDVR0TAQH/BAIwADAfBgNVHSMEGDAWgBQcwrrUO0u+FhIUuVdrDRCQRsi6\nZjANBgkqhkiG9w0BAQsFAAOCAQEAm5dxBoI9pScOCvRAchg4CprdRDSJb9K6yB3O\nnCAxnM47iHeXnY3WlnI388kHu8DU7O4ba1tJbGs3KY9KzioPk43pU12jWkO1onoF\n+mTDjx/Ef1cYWA9r5q/LtgTa6Q2sxV4O2x67QW82aAnaxO34dV5zWCPIvAoovZBV\nHRT+BgCg3r2vD1RGKK2nl1aYJtWhO1SZubam+VttdZ/vbM9oOJctxmImsEtBXjkY\nKteePdQtLL5o03JhyXWyRshCq+HMmKf2KgyY8gvydGcP4eLQdBWcW40LcnVq6UjT\n0kJycJEKngMVademq1ZWHGaiYB7hyT6GhgIcHUJ2cKrPgbEh1Q==\n-----END CERTIFICATE-----",
                      "key": "-----BEGIN RSA PRIVATE KEY-----\nMIIEowIBAAKCAQEA4AI33yvHZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I\n2e4SceTKi6O3C/I1XuvWlpng5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4AN\nB5JWWqDOjrRT3o7nRPGXfilabP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYp\nmjdbfxabTz07ig0+6/cwKoRRxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTF\nYvmAH7gcdssSFBt8NPlUATHEsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6W\ndgmu5H2pDml8CDNLDv98Aj7i+I5SRKKcVPlnuQIDAQABAoIBAFPQYKlOzw/+BA0b\nyMIUpdctIMb/54CR/xR0mVw1DbSjigNVPjHUQvB8Y1B2FAITQObgJO06bAv0QdWN\nRb0/v/yYiNJDFjaLjaIAHlO/2+oWrXbFaZqgpVDJhB+e1xaZr2x7XGxm+p925k30\nl6pvIRY+I8JRKvZiV1VZHwL/R3JOtPr++xMZtLVjVOI+f+ySqJ+TZHuAjm49EKxj\ncEmmJ28b7QcziXsvKy00f+zbqLIBKXQdZAFU5eEr1BsDRXdRW+Kf0XIvftuy4BJZ\nvoKT+VGhEvF/qysswL4+6IAO6tpuYnnM0Y2d3sOGoWPkTcQK0MekYKzL/WmtCjNs\n9hodJtECgYEA5EWyhEOf4uOKe5TDp697UCUvXLoOR58FDe/S8XNvScn29jjOkqIg\nOMoqo9xAkJTNTzqn5UUdt1x/pgM2NxlPLFijrc0zQlX3SoOO2ryDd9WNi7YKtN16\nKJqa536WeZu2OEbuAZ+S3GALVy1RPeTNPnUOmKnF06DjDUGzLNCZy10CgYEA+zfw\n952DWuz1U0Z4wvAEqqcgUKXPKrkTXV/iUnjkDkrLYVr0ZofDNTXrdHl+UedFmaOC\ncieZn6DNhcdz5tKtyysGMH3g/qs9PfoGUngvcXsy0Egk04l3x1jc8TTCLqXZXYaQ\nHMsx51n+R58oncPtzYSUOr9qQ6PbC2CstTbFJA0CgYEAjGEsUliAB/jknfEzjXjG\nPdhQUxb8VyE864Az2lah9t/kJzFyIAziAeqZ5GE7t247AGFTBRTHHI8e1Qoemi3P\nWbc9GVIbFs1lIYbcIDpUIyrKPEP8O5QEXtoNLxXTFgAjRGKiVY87spjCAJ+W2ZhO\ne/1it5GYXfgQCYQA2yuBmOUCgYANRkR2YR1axaCk+NlSu6oTdmdPu6M5x7PNQE7O\nOtMaKjua9lppvIzFGAdMDUtueoEEAE7ZR1xnwfB6PDLUpJdIYAqgr1YfPt8qkjaZ\nTv56yZ7CwL0pbF8m6nwqRrZoDp1wwraEvvvxFKFKGY/k3kCHlpTakdjEoDjn3gDi\nRnWeVQKBgCEneMSzucei5LRppRtRaJw/Btll8qlPMlX3W7dxQ3cLwpmLOn0m51Fp\nPIZ44zYK8R6fu4+/sSrlfaIg86Ugeufp6YNxyNROKxUGza5vDIu5OftwWtBeg+UK\nZ8lLWNdX6pp7WMujmF3H1DrkBbauYMUKZ4UxUYtelgHERMePIxwb\n-----END RSA PRIVATE KEY-----"
                    }
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "false"
          containerImage: quay.io/fujitsu/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:eaf601e7aef525ebcd4b769c859ed74c7a8077f843b3a75d76713eabf0295745
          createdAt: "2025-09-11T09:32:52Z"
          description: OpenShift Operator for Fujitsu Enterprise Postgres
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.1.0 <7.1.6'
          operators.openshift.io/valid-subscription: '["Fujitsu Enterprise Postgres
            Advanced Edition with Cryptographic Module Annual Subscription License
            per Core 17 Operator for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/internal-objects: '["fepautoscales.fep.fujitsu.io","fepconfigs.fep.fujitsu.io","fepusers.fep.fujitsu.io","fepcerts.fep.fujitsu.io","fepvolumes.fep.fujitsu.io","fepbackups.fep.fujitsu.io"]'
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: fj-dbaas-tec@dl.jp.fujitsu.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: FEPAction
            name: fepactions.fep.fujitsu.io
            version: v1
          - kind: FEPAutoscale
            name: fepautoscales.fep.fujitsu.io
            version: v1
          - kind: FEPBackup
            name: fepbackups.fep.fujitsu.io
            version: v1
          - kind: FEPCert
            name: fepcerts.fep.fujitsu.io
            version: v1
          - kind: FEPCluster
            name: fepclusters.fep.fujitsu.io
            version: v2
          - kind: FEPConfig
            name: fepconfigs.fep.fujitsu.io
            version: v1
          - kind: FEPExporter
            name: fepexporters.fep.fujitsu.io
            version: v1
          - kind: FEPLogging
            name: feploggings.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2Cert
            name: feppgpool2certs.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2
            name: feppgpool2s.fep.fujitsu.io
            version: v1
          - kind: FEPRestore
            name: feprestores.fep.fujitsu.io
            version: v1
          - kind: FEPUpgrade
            name: fepupgrades.fep.fujitsu.io
            version: v1
          - kind: FEPUser
            name: fepusers.fep.fujitsu.io
            version: v1
          - kind: FEPVolume
            name: fepvolumes.fep.fujitsu.io
            version: v1
        description: "Fujitsu Enterprise Postgres 17 delivers an enterprise-grade
          PostgreSQL on OpenShift Container Platform.\n\nThis solution provides the
          flexibility of a hybrid cloud solution while delivering an enhanced distribution
          of PostgreSQL to support enterprise-level workloads and provide improved
          deployment and management, availability, performance, data governance and
          security.  \n\nAvailable as a multi-architecture container built for both
          amd64.\n\nThe download and Use of the Product is strictly subject to the
          terms of the End User License Agreement with Fujitsu Limited found at https://www.fast.fujitsu.com/fujitsu-enterprise-postgres-license-agreements.
          Where the Product that has been embedded as a whole or part into a third
          party program, only Authorised Customers may download and use the Product.\n"
        displayName: Fujitsu Enterprise Postgres Operator with Cryptographic Module
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - Fujitsu
        - fep
        links:
        - name: Fujitsu Enterprise Postgres
          url: https://www.postgresql.fastware.com/
        maintainers:
        - email: fj-dbaas-tec@dl.jp.fujitsu.com
          name: Fujitsu
        maturity: stable
        provider:
          name: Fujitsu
          url: https://www.postgresql.fastware.com/
        relatedImages:
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-restore-with-cryptographic-module@sha256:edeb7b72a89e05995661e6a17e75b5b477a3ca351e0db146a939a33e88e16c97
        - quay.io/fujitsu/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:eaf601e7aef525ebcd4b769c859ed74c7a8077f843b3a75d76713eabf0295745
        - quay.io/fujitsu/fujitsu-enterprise-postgres-utils-with-cryptographic-module@sha256:7d169df0b8fe47000891afde79be8dd7978da3f893d1e92f01f57fd8c0b3b4c6
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-pgpool2-with-cryptographic-module@sha256:e510bc2d30df33e4c9d3babb898ff9cbd13704fa3fa69a284bf672765fb03083
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-server-with-cryptographic-module@sha256:7778f4b03861b56371d18f7726b68ad21d0faa7c1a25cb38fe7cac483d81c095
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-restore-with-cryptographic-module@sha256:6c0d6e10cfe1868be2986fdde14c6f55f86b12cea01a82199b1e2e0c3154ef75
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-upgrade-with-cryptographic-module@sha256:b8be0344b6cf99b2c5de2b16f04f6f2b05bde861d9b38afd319bbbc4dd72a620
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-upgrade-with-cryptographic-module@sha256:caf6e1d2cc31d01958e4adcb4c3c307d761f350d5b0963ce19214f041c14184a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-restore-with-cryptographic-module@sha256:98a3b0bfa6ec7a9a4a08fb524c331c3de6a9b580d6386a7b5430955ed84b2fc5
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-backup-with-cryptographic-module@sha256:dd15cf3b6b4a5c959efef8b602abe5d64b4f5a424fb6a3e55079a2040e67cd7a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-backup-with-cryptographic-module@sha256:03b7f3c76e3eb775e4ec7f66f7506ee1feeab56dca96e868ac3f38b9c1ca5567
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentd-with-cryptographic-module@sha256:f743f35012e6fd3d54a6b54baeeee98bd7c65fe52923acd87546066fa9b54d7e
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-upgrade-with-cryptographic-module@sha256:1fe690a9c04a55213062ad3bef8c0cff73849f8777c3945b9a8ae2dcff6330fa
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-utils-with-cryptographic-module@sha256:8b2b6a981c863f7075ced1235c614317b6526626b42405295a9c664736f0a80f
        - quay.io/fujitsu/fujitsu-enterprise-postgres-cronjob-with-cryptographic-module@sha256:809a64600396378ff8c825595d40f72aedbc2b57f394420548262b085a3f2ea5
        - quay.io/fujitsu/fujitsu-enterprise-postgres-exporter-with-cryptographic-module@sha256:71e1d1e715e2d89757e80d90b8d5b11baa729110cd5f7898df920a4c85f2bf8a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentbit-with-cryptographic-module@sha256:bb661b3aad0b825d56a9696d6d2d9f5c04add2a83e793fc3d4a73980f4a7a0ac
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-utils-with-cryptographic-module@sha256:4caf1dc5211ac03893ce9a62b417f412ffd411f13569f3ac75e12ca5128af9e4
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-pgpool2-with-cryptographic-module@sha256:a4c5a59ca18f6fa295b78d570f03991f1a4b4bd6df52f8526a1254ac1a567990
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-pgpool2-with-cryptographic-module@sha256:e8aa27cbb57503ec5f4f37beeb2e833ace08d4c583d4361b7ce945c7461a904c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-server-with-cryptographic-module@sha256:cad4e36dc2b1e5ca6bf9308d9f93958d4c8f2bbdc14e000ca0b5700fa335df44
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-backup-with-cryptographic-module@sha256:5b49aa46386ca954ebcbea28fbabd15929cb64fc2a4d2dd2796c9d2e15b212c6
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-server-with-cryptographic-module@sha256:a012f21f63729ba667af097375e08874156e2cf1b94ae203985d45f11031fea9
        version: 7.1.6
      entries:
      - name: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.6
        version: 7.1.6
      - name: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.5
        version: 7.1.5
      name: stable
    defaultChannel: stable
    packageName: fujitsu-enterprise-postgres-operator-with-cg-module
    provider:
      name: Fujitsu
      url: https://www.postgresql.fastware.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-05-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx8-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx8-operator-rhmp/support?utm_source=openshift_console
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:154f85cc6721a70909dee755ddd88897ad51038afe2538b6dee4d26e7d8b9afb
        version: 0.0.10
      entries:
      - name: joget-dx8-openshift-operator.v0.0.10
        version: 0.0.10
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-operator-rhmp
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://www.redhat.com/
    name: rhdh
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhdh-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
          createdAt: "2025-09-12T06:25:11Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.7.1'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.7/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:62f32e50727c5006766a34daa5b6472b649cd9894f3fe9543b8ecc67e6760e8e
        version: 1.7.1
      entries:
      - name: rhdh-operator.v1.7.1
        version: 1.7.1
      - name: rhdh-operator.v1.7.0
        version: 1.7.0
      - name: rhdh-operator.v1.6.5
        version: 1.6.5
      - name: rhdh-operator.v1.6.4
        version: 1.6.4
      - name: rhdh-operator.v1.6.3
        version: 1.6.3
      - name: rhdh-operator.v1.6.2
        version: 1.6.2
      - name: rhdh-operator.v1.6.1
        version: 1.6.1
      - name: rhdh-operator.v1.6.0
        version: 1.6.0
      - name: rhdh-operator.v1.5.3
        version: 1.5.3
      - name: rhdh-operator.v1.5.2
        version: 1.5.2
      - name: rhdh-operator.v1.5.1
        version: 1.5.1
      - name: rhdh-operator.v1.4.3
        version: 1.4.3
      - name: rhdh-operator.v1.4.2
        version: 1.4.2
      - name: rhdh-operator.v1.4.1
        version: 1.4.1
      - name: rhdh-operator.v1.4.0
        version: 1.4.0
      - name: rhdh-operator.v1.3.5
        version: 1.3.5
      - name: rhdh-operator.v1.3.4
        version: 1.3.4
      - name: rhdh-operator.v1.3.3
        version: 1.3.3
      - name: rhdh-operator.v1.3.1
        version: 1.3.1
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast
    - currentCSV: rhdh-operator.v1.1.2-0.1714688890.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:9bea7eabdea44342248dfd6091a8f5c3b6e65884c916b3c2a073a4a64481aa6f
          createdAt: "2023-03-20T16:11:34Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins, configuration settings, and deployment
            mechanisms, which can help streamline the process of setting up a self-managed
            internal developer portal for adopters who are just starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: rhdh-operator.v1.1.2
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.1.0 <1.2.0'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/janus-idp/operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: asoro@redhat.com
          name: Armel Soro
        - email: gazarenk@redhat.com
          name: Gennady Azarenkov
        - email: jianrzha@redhat.com
          name: Jianrong Zhang
        - email: nboldt@redhat.com
          name: Nick Boldt
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:95ed239eb427c7b1e422ece28c5277b5018efb751c2b472bf502b8216219c047
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:9bea7eabdea44342248dfd6091a8f5c3b6e65884c916b3c2a073a4a64481aa6f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:d7ae2456857cab09a9e9acf69c9990ef18deb7f920046b8c2483fbfd0d42bbfc
        - registry.redhat.io/rhel9/postgresql-15@sha256:d5f3d496ecd55688e384cc0627700a97af6924582123bcf00749c375724af3de
        version: 1.1.2+0.1714688890.p
      entries:
      - name: rhdh-operator.v1.1.2
        version: 1.1.2
      - name: rhdh-operator.v1.1.2-0.1714608881.p
        version: 1.1.2+0.1714608881.p
      - name: rhdh-operator.v1.1.2-0.1714688618.p
        version: 1.1.2+0.1714688618.p
      - name: rhdh-operator.v1.1.2-0.1714688890.p
        version: 1.1.2+0.1714688890.p
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.1
    - currentCSV: rhdh-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:c9051a6c339cd40cbc7474a33b32a6b94913af964e12ca5da0c0b7e1c6f444d5
          createdAt: "2024-11-28T17:55:19Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.2.6'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.2/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.2/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: asoro@redhat.com
          name: Armel Soro
        - email: cdaley@redhat.com
          name: Corey Daley
        - email: gazarenk@redhat.com
          name: Gennady Azarenkov
        - email: nboldt@redhat.com
          name: Nick Boldt
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:ef29f2b340a6dc52c5014239876d54a54f7c0f0bf4ea917391438a69c94a8532
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:c9051a6c339cd40cbc7474a33b32a6b94913af964e12ca5da0c0b7e1c6f444d5
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:025eb58c69b9cf5992ecf2fccb29a3d36061cbcf62652a1c6878f0c4d767c1fb
        - registry.redhat.io/rhel9/postgresql-15@sha256:24fb4e7914a6e1464d015be9e5582cc4b9da224137408bd429e7ea4f391aa198
        version: 1.2.6
      entries:
      - name: rhdh-operator.v1.2.6
        version: 1.2.6
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.2
    - currentCSV: rhdh-operator.v1.3.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:5ce579edc10bde5c1c400104f7e0d9f86ca277cd565aaf3dd84b8636f5b949a8
          createdAt: "2025-02-10T18:12:52Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.3.5'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.3/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:5ce579edc10bde5c1c400104f7e0d9f86ca277cd565aaf3dd84b8636f5b949a8
        - registry.redhat.io/rhel9/postgresql-15@sha256:24fb4e7914a6e1464d015be9e5582cc4b9da224137408bd429e7ea4f391aa198
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:c044fed8485a62a5051ac726525da546936be22f268db05491bf81582e0e6194
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b4934c45ba595b36e65c73b734c58d1517e7f43b7cfc309ca8bc853733582f76
        version: 1.3.5
      entries:
      - name: rhdh-operator.v1.3.5
        version: 1.3.5
      - name: rhdh-operator.v1.3.4
        version: 1.3.4
      - name: rhdh-operator.v1.3.3
        version: 1.3.3
      - name: rhdh-operator.v1.3.1
        version: 1.3.1
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.3
    - currentCSV: rhdh-operator.v1.4.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:ae5b52c1b35e82f3c564402e367a2a2028e16806867bba0868643089ba7ed223
          createdAt: "2025-04-03T05:42:32Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.4.3'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.4/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:ae5b52c1b35e82f3c564402e367a2a2028e16806867bba0868643089ba7ed223
        version: 1.4.3
      entries:
      - name: rhdh-operator.v1.4.3
        version: 1.4.3
      - name: rhdh-operator.v1.4.2
        version: 1.4.2
      - name: rhdh-operator.v1.4.1
        version: 1.4.1
      - name: rhdh-operator.v1.4.0
        version: 1.4.0
      name: fast-1.4
    - currentCSV: rhdh-operator.v1.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:8fe167cc649a1c53994c376f3cf3eed8c6e318741a0c08e5aaf07cb0b74e7f8a
          createdAt: "2025-06-05T14:31:21Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.5.3'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.5/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:8fe167cc649a1c53994c376f3cf3eed8c6e318741a0c08e5aaf07cb0b74e7f8a
        - registry.redhat.io/rhel9/postgresql-15@sha256:74377f5a61dd7e3e0ba5d2fa880b7107025433a026396567c79eac927b4212bc
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:2ad8c87e82df27c2b6ceba94179d1f0a40fe775368663c1b5572611888e0e02b
        version: 1.5.3
      entries:
      - name: rhdh-operator.v1.5.3
        version: 1.5.3
      - name: rhdh-operator.v1.5.2
        version: 1.5.2
      - name: rhdh-operator.v1.5.1
        version: 1.5.1
      name: fast-1.5
    - currentCSV: rhdh-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:50f20300c81621221260d03d22b1442750e3f1d2e62f3e511d63925d84c4d1f9
          createdAt: "2025-08-27T18:43:12Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.6.5'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.6/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:50f20300c81621221260d03d22b1442750e3f1d2e62f3e511d63925d84c4d1f9
        - registry.redhat.io/rhel9/postgresql-15@sha256:5e225c1646251d2509b3a6e45fad23abfb11dd5c45d3224b59178f90deb38a70
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:177a7308f61a6b1ba21bc0f7bf550a93d4478e52479e2c94a2437f2152e3d8ff
        version: 1.6.5
      entries:
      - name: rhdh-operator.v1.6.5
        version: 1.6.5
      - name: rhdh-operator.v1.6.4
        version: 1.6.4
      - name: rhdh-operator.v1.6.3
        version: 1.6.3
      - name: rhdh-operator.v1.6.2
        version: 1.6.2
      - name: rhdh-operator.v1.6.1
        version: 1.6.1
      - name: rhdh-operator.v1.6.0
        version: 1.6.0
      name: fast-1.6
    - currentCSV: rhdh-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
          createdAt: "2025-09-12T06:25:11Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.7.1'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.7/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:62f32e50727c5006766a34daa5b6472b649cd9894f3fe9543b8ecc67e6760e8e
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
        version: 1.7.1
      entries:
      - name: rhdh-operator.v1.7.1
        version: 1.7.1
      - name: rhdh-operator.v1.7.0
        version: 1.7.0
      name: fast-1.7
    defaultChannel: fast
    packageName: rhdh
    provider:
      name: Red Hat Inc.
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-activemqartemis: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/
    name: amq-broker-rhel8
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-broker-operator.v7.10.7-opr-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                    "name": "ex-aao",
                    "application": "ex-aao-app"
                },
                "spec": {
                    "deploymentPlan": {
                        "image": "placeholder",
                        "size": 2,
                        "requireLogin": false,
                        "persistenceEnabled": false,
                        "journalType": "nio",
                        "messageMigration": false,
                        "jolokiaAgentEnabled": false,
                        "managementRBACEnabled": true
                    }
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                    "name": "ex-aaoaddress"
                },
                "spec": {
                    "addressName": "myAddress0",
                    "queueName": "myQueue0",
                    "routingType": "anycast"
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                    "name": "ex-aaoscaledown"
                },
                "spec": {
                    "localOnly": "true"
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                    "name": "ex-keycloak"
                },
                "spec": {
                    "loginModules": {
                        "propertiesLoginModules": [
                            {
                                "name": "prop-module",
                                "users": [
                                    {
                                        "name": "superman",
                                        "roles": [
                                            "root"
                                        ]
                                    }
                                ]
                            }
                        ]
                    },
                    "securityDomains": {
                        "brokerDomain": {
                            "name": "activemq",
                            "loginModules": [
                                {
                                    "name": "prop-module",
                                    "flag": "sufficient"
                                }
                            ]
                        }
                    }
                }
            }]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:1a7aa54d2799d238eb5f49f7a95a78a896f6bf8d222567e9118e0e3963cc9aad
          createdAt: "2021-06-18T00:02:04Z"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A stateful deployment of one or more brokers
            displayName: AMQ Broker
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: Adding and removing addresses via custom resource definitions
            displayName: AMQ Broker Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: Provides message migration on clustered broker scaledown
            displayName: AMQ Broker Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: Security configuration for the broker
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
          to deploy and manage stateful AMQ Broker broker clusters
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: rkieley@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:1a7aa54d2799d238eb5f49f7a95a78a896f6bf8d222567e9118e0e3963cc9aad
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ba21779c29b431a22c824e91820705c9501fb0b7d6f7e1ae22fcd3becc8c8a2f
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:d4010af53429502f41d722c39d69a7bb147bb06d19e3d4a69de70dbe88e3990e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b575d6eebf44dc9049445e357488efb17a690953852419b534df30e0105b9ca4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ff6fefad518a6c997d4c5a6e475ba89640260167f0bc27715daf3cc30116fad1
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:56282c6a461e945a83bffb13d3f3d8d9621800f0fc9cd755a75d21eacd8f5fbf
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:979b59325aa0f34eb05625201beba53fccbb83bd5eb80a89dcb5261ae358138f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:83faefa67b33d93836a1be3cb438bb147242d602565dcbe59b4ed42b26072191
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:967d840b831b310abd3a468a8652d1fa48b2ce28b342bfab53999245deb004f1
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:2cb14c917246ffcbd5d861ca6f3d58cb1146a4e55715ee08160e115d3dfc1c3e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f37f98c809c6f29a83e3d5a3ac4494e28efe9b25d33c54f533c6a08662244622
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:4e09ee3bd8ff3479fdd00e221b7b1505ae04ca6c64fd6bc9e2f1a34b8e98d137
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0d857b95084e39c011651be89482d4dcc77c2d1c3ac3c6a819f3c9d5a57a425f
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:9465e398a55c89c121b3ae50374414190a991839fa5e1018fc02551f6aa3b9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3c00b524722b500ff0922fe7a8722868557512183a8f69248f37a68c45c6a770
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a94fa8135d3765363e442bda45df6e02a3e22bee7ea198bcb0d3584331a4320c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c5f4c08e068b9721967cf7c7cbd9a9e93fb5e39b264dd13b653e99d8f3fa9e0e
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:aa7af7e038a12b5202108b362c8af243954603e84034a16c0192684c86d5d28b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b74d03ed852a3731467ffda95266ce49f2065972f1c37bf254f3d52b34c11991
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:7abfdcfd7b2d3d12b9234f7b057bd199bceefa47bdf02f62236f82a13fdbf5b7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:869cda365447b0b7b53ed21d2496884c15dc9c07295a9dfa0337c2c9974fe018
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:5ff28e3062d3eabb9f52d8271b837fdab06cafbd8b8f732b1bc00de49ad69aad
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f58678f03a250dbb296ebeb112c9150340155799ba4dad9e90bf347b9ed0ca7b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3219f90789690a3c5cfcc86e5f602c443e75f1fdb3a19660f43ee3996bfa8f62
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:f2651cb097eb631fc6af90e4a8404d36516aa2205c546125bf0c37739ab525fa
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:35b7d96d5065f280a528626abd2c3df84e79ccea93ab6bd8b8837af66c330f30
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:45a5e054546a2c1ee338518f8477f728f5f7108c9dd8b516b4c79d4a7d6f5ba7
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:1a9f13fd565113055b2aa2022bd0709a63c1dc058a8570e0fd7eba5c7b101334
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:97c3aa5673610875a53571449f7d5210ae8d675072d044f17b416725ed0dba88
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:ba0f5df7da08bc030ff5e0e68ecb52a931c25f6b67c6556aa8a9661c82d69d42
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:83ed89fa942ccbf4451954956db42138c83f9b015e7de07fc159ddc4bb6dc843
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9a02decbc460ddfc501d8a4a9b41cd084776090951906b027838c0829e1fc15a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ff819e913406911735da39c51699405f747bd697116a273a00897ff176bf05ff
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:5d39f26f06769fe5c989ea0e0ee8f9166120e152ce2c9bc3336fc8d7cab2ff92
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:e6f71001dccfbbe1cd7cf4fded69e181c10f46cc5c8bd3a8d3045a3f3ac41e2f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:61e5a8c1e46f1c18648e0406e2a57c52f911979079020cc57c7a4b06a56940bb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:e03945756d86bd189e8c33027fa668d2065c168510085dfd06defd9f40d3e0ab
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:34739366153b825dcb113e0ed6115ef5656085cf8d47492789ff17464042b576
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:34ab1f66d8daa544e3823464383e9788e307665c875499838fbed95f3a74a4dd
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0b2126cfb6054fdf428c1f43b69e36e93a09a49ce15350e9273c98cc08c6598b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a8b2a9364fd06c8e1c29b8f8f28d3e4844bf7840a62b67137b663c558982d4a3
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:6e9e0e3fddc44fbf87591d8904176407f1ec409d578a7d108c13cdd48d85131b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:67ae74e72927557c55611d0248347b2099b117ac1aa245dac05856fe782fb59c
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:c3ac4602b48867f3d4f7cc0efd5c9f133098e409626cfb402f80386ff460dbf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c9eba0d1f9fba588711774f4db75221f057578407065127ed60c33eef564f499
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:86ff145bd68e797a0507495570063b54cbe4186fcb4acd83c900406619ef9524
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:d4a9176f49ffd9db6051fbda15d47a1aa32c282fb5010d0caf5b9243e80bf055
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:4440b20349032bb5c0bb2a8deb89832b37767af6f0b3208b067867a85641913e
        version: 7.10.7-opr-1
      entries:
      - name: amq-broker-operator.v7.10.7-opr-1
        version: 7.10.7-opr-1
      - name: amq-broker-operator.v7.10.6-opr-1
        version: 7.10.6-opr-1
      - name: amq-broker-operator.v7.10.5-opr-1
        version: 7.10.5-opr-1
      - name: amq-broker-operator.v7.10.4-opr-1
        version: 7.10.4-opr-1
      - name: amq-broker-operator.v7.10.3-opr-3
        version: 7.10.3-opr-3
      - name: amq-broker-operator.v7.10.3-opr-2
        version: 7.10.3-opr-2
      - name: amq-broker-operator.v7.10.3-opr-1
        version: 7.10.3-opr-1
      - name: amq-broker-operator.v7.10.2-opr-2-0.1676475747.p
        version: 7.10.2-opr-2+0.1676475747.p
      - name: amq-broker-operator.v7.10.2-opr-2-0.1680622941.p
        version: 7.10.2-opr-2+0.1680622941.p
      - name: amq-broker-operator.v7.10.2-opr-2
        version: 7.10.2-opr-2
      - name: amq-broker-operator.v7.10.2-opr-1
        version: 7.10.2-opr-1
      - name: amq-broker-operator.v7.10.1-opr-2
        version: 7.10.1-opr-2
      - name: amq-broker-operator.v7.10.1-opr-1
        version: 7.10.1-opr-1
      - name: amq-broker-operator.v7.10.0-opr-4
        version: 7.10.0-opr-4
      - name: amq-broker-operator.v7.10.0-opr-3
        version: 7.10.0-opr-3
      - name: amq-broker-operator.v7.10.0-opr-2
        version: 7.10.0-opr-2
      - name: amq-broker-operator.v7.10.0-opr-1
        version: 7.10.0-opr-1
      name: 7.10.x
    - currentCSV: amq-broker-operator.v7.11.8-opr-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisSecurity is the Schema for the activemqartemissecurities
              API
            displayName: Active MQArtemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:b9aecae6b0ce0b241c7a5b9e6e096395940bb6a384373a4d121822cf10c26d99
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b74d03ed852a3731467ffda95266ce49f2065972f1c37bf254f3d52b34c11991
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:442b6638133d42fa7faa41d86e47be95cdca56c2e16efcd6e7e22991c3812a36
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f37f98c809c6f29a83e3d5a3ac4494e28efe9b25d33c54f533c6a08662244622
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0b2126cfb6054fdf428c1f43b69e36e93a09a49ce15350e9273c98cc08c6598b
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:4440b20349032bb5c0bb2a8deb89832b37767af6f0b3208b067867a85641913e
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:7abfdcfd7b2d3d12b9234f7b057bd199bceefa47bdf02f62236f82a13fdbf5b7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b9f93c243ead118651083880786a81c624f390fcc3596c7ec5ebd55f34f95ea3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:35b7d96d5065f280a528626abd2c3df84e79ccea93ab6bd8b8837af66c330f30
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b575d6eebf44dc9049445e357488efb17a690953852419b534df30e0105b9ca4
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:f2651cb097eb631fc6af90e4a8404d36516aa2205c546125bf0c37739ab525fa
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:83ed89fa942ccbf4451954956db42138c83f9b015e7de07fc159ddc4bb6dc843
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:967d840b831b310abd3a468a8652d1fa48b2ce28b342bfab53999245deb004f1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0d857b95084e39c011651be89482d4dcc77c2d1c3ac3c6a819f3c9d5a57a425f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:e03945756d86bd189e8c33027fa668d2065c168510085dfd06defd9f40d3e0ab
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:56282c6a461e945a83bffb13d3f3d8d9621800f0fc9cd755a75d21eacd8f5fbf
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5e5d8b551ab8ca377e0bc939037970a2f31ee14e9f555853ebea2759d673895a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9a02decbc460ddfc501d8a4a9b41cd084776090951906b027838c0829e1fc15a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0f443f5aa818dd85bfa93e51e37bdac3909628a283aa52a89125b527dd75d9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:5d39f26f06769fe5c989ea0e0ee8f9166120e152ce2c9bc3336fc8d7cab2ff92
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:9465e398a55c89c121b3ae50374414190a991839fa5e1018fc02551f6aa3b9aa
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:1a9f13fd565113055b2aa2022bd0709a63c1dc058a8570e0fd7eba5c7b101334
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:45a5e054546a2c1ee338518f8477f728f5f7108c9dd8b516b4c79d4a7d6f5ba7
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c9eba0d1f9fba588711774f4db75221f057578407065127ed60c33eef564f499
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a94fa8135d3765363e442bda45df6e02a3e22bee7ea198bcb0d3584331a4320c
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:34ab1f66d8daa544e3823464383e9788e307665c875499838fbed95f3a74a4dd
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:86ff145bd68e797a0507495570063b54cbe4186fcb4acd83c900406619ef9524
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c5f4c08e068b9721967cf7c7cbd9a9e93fb5e39b264dd13b653e99d8f3fa9e0e
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:d4a9176f49ffd9db6051fbda15d47a1aa32c282fb5010d0caf5b9243e80bf055
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:29f07006ad62023c20d34971cb3ae900349efae5ab2918aaa8e5647d788032be
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ff819e913406911735da39c51699405f747bd697116a273a00897ff176bf05ff
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b0b3684ef1ebf8b44d3a4c46391984173792dde1bd7363ea06d43a8f95a717d0
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:cc438b4bc2b14f6bff4e5d32e7dc0fad6d2ca391c4e84b5ae925b7b40af76729
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ba21779c29b431a22c824e91820705c9501fb0b7d6f7e1ae22fcd3becc8c8a2f
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:ba0f5df7da08bc030ff5e0e68ecb52a931c25f6b67c6556aa8a9661c82d69d42
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:e6f71001dccfbbe1cd7cf4fded69e181c10f46cc5c8bd3a8d3045a3f3ac41e2f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:67ae74e72927557c55611d0248347b2099b117ac1aa245dac05856fe782fb59c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:979b59325aa0f34eb05625201beba53fccbb83bd5eb80a89dcb5261ae358138f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ade37271426ab8f543d72f8da3e9f663d665f9427379a2539fc9323f6897cc44
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:34739366153b825dcb113e0ed6115ef5656085cf8d47492789ff17464042b576
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:869cda365447b0b7b53ed21d2496884c15dc9c07295a9dfa0337c2c9974fe018
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ff6fefad518a6c997d4c5a6e475ba89640260167f0bc27715daf3cc30116fad1
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:61e5a8c1e46f1c18648e0406e2a57c52f911979079020cc57c7a4b06a56940bb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0dde839c2dce7cb684094bf26523c8e16677de03149a0fff468b8c3f106e1f4f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5f842ebe9a6a4294a056ad10189a5f277cfd3fb4e6e0b3225e194bd5c131151d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3c00b524722b500ff0922fe7a8722868557512183a8f69248f37a68c45c6a770
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:4e09ee3bd8ff3479fdd00e221b7b1505ae04ca6c64fd6bc9e2f1a34b8e98d137
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:d4010af53429502f41d722c39d69a7bb147bb06d19e3d4a69de70dbe88e3990e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a8b2a9364fd06c8e1c29b8f8f28d3e4844bf7840a62b67137b663c558982d4a3
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:5ff28e3062d3eabb9f52d8271b837fdab06cafbd8b8f732b1bc00de49ad69aad
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:d2ae3defac36f00112c6b773e13626336ba15d6f50f905ad1c33f122bfe1a6a9
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:97c3aa5673610875a53571449f7d5210ae8d675072d044f17b416725ed0dba88
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f58678f03a250dbb296ebeb112c9150340155799ba4dad9e90bf347b9ed0ca7b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:40ccaf2bb2e0f8132f2541b39cb835f17b38050b3ca9232f0861fd5fcaf4324a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:6e9e0e3fddc44fbf87591d8904176407f1ec409d578a7d108c13cdd48d85131b
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:c3ac4602b48867f3d4f7cc0efd5c9f133098e409626cfb402f80386ff460dbf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:137b893ad15e0cd5353bae134c8fae833257a12641b12d0f774f1565ce66f63c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:89bb769b4c33eb931cc7866eeabec9a315b913851cfa84bc6e11ae8d1e8b2cbb
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:2cb14c917246ffcbd5d861ca6f3d58cb1146a4e55715ee08160e115d3dfc1c3e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:25c3732e945b4bc3e9aca721ce709d0217339b065d9dac9afdb6202e9fd8a799
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:8b4c4400a02c6cacfd978147b158e2ccbe5da955578b4b970a5ee7edac865326
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3219f90789690a3c5cfcc86e5f602c443e75f1fdb3a19660f43ee3996bfa8f62
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:7d53744ae8c77aea09c59165486c9ed0eaa131140f6b346042293bce0bd8a745
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:aa7af7e038a12b5202108b362c8af243954603e84034a16c0192684c86d5d28b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9c1ac38bf8eec67a3da57c3cc61105513776a749408cb287fe6448a17889de4e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:83faefa67b33d93836a1be3cb438bb147242d602565dcbe59b4ed42b26072191
        version: 7.11.8-opr-1
      entries:
      - name: amq-broker-operator.v7.11.8-opr-1
        version: 7.11.8-opr-1
      - name: amq-broker-operator.v7.11.7-opr-2-1742200979
        version: 7.11.7-opr-2-1742200979
      - name: amq-broker-operator.v7.11.7-opr-2-1740583644
        version: 7.11.7-opr-2-1740583644
      - name: amq-broker-operator.v7.11.7-opr-2-1736956309
        version: 7.11.7-opr-2-1736956309
      - name: amq-broker-operator.v7.11.7-opr-2-1731330819
        version: 7.11.7-opr-2-1731330819
      - name: amq-broker-operator.v7.11.7-opr-2
        version: 7.11.7-opr-2
      - name: amq-broker-operator.v7.11.7-opr-1-1726157430
        version: 7.11.7-opr-1-1726157430
      - name: amq-broker-operator.v7.11.7-opr-1-1723204800
        version: 7.11.7-opr-1-1723204800
      - name: amq-broker-operator.v7.11.7-opr-1
        version: 7.11.7-opr-1
      - name: amq-broker-operator.v7.11.6-opr-2
        version: 7.11.6-opr-2
      - name: amq-broker-operator.v7.11.5-opr-1
        version: 7.11.5-opr-1
      - name: amq-broker-operator.v7.11.4-opr-1
        version: 7.11.4-opr-1
      - name: amq-broker-operator.v7.11.3-opr-1
        version: 7.11.3-opr-1
      - name: amq-broker-operator.v7.11.2-opr-1
        version: 7.11.2-opr-1
      - name: amq-broker-operator.v7.11.1-opr-2
        version: 7.11.1-opr-2
      - name: amq-broker-operator.v7.11.1-opr-1
        version: 7.11.1-opr-1
      - name: amq-broker-operator.v7.11.0-opr-3
        version: 7.11.0-opr-3
      - name: amq-broker-operator.v7.11.0-opr-2
        version: 7.11.0-opr-2
      name: 7.11.x
    - currentCSV: amq-broker-operator.v7.12.5-opr-1-0.1757928550.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-broker-operator.v7.12.5-opr-1-0.1756216162.p
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:7d53744ae8c77aea09c59165486c9ed0eaa131140f6b346042293bce0bd8a745
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9c1ac38bf8eec67a3da57c3cc61105513776a749408cb287fe6448a17889de4e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:29f07006ad62023c20d34971cb3ae900349efae5ab2918aaa8e5647d788032be
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:dc13871dfed52402c40fd192900fc3cb5a49da4fa08e4ec4d0e23a879e83db25
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5e5d8b551ab8ca377e0bc939037970a2f31ee14e9f555853ebea2759d673895a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:cc438b4bc2b14f6bff4e5d32e7dc0fad6d2ca391c4e84b5ae925b7b40af76729
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d71506472636d940b5dbd3e708d3e0448d777eeecfd38de61ba49e281d2979c6
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ffffa9875f0379e9373f89f05eb06e5a193273bb04bc3aa5f85b044357b79098
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:b9aecae6b0ce0b241c7a5b9e6e096395940bb6a384373a4d121822cf10c26d99
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:442b6638133d42fa7faa41d86e47be95cdca56c2e16efcd6e7e22991c3812a36
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:25c3732e945b4bc3e9aca721ce709d0217339b065d9dac9afdb6202e9fd8a799
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ade37271426ab8f543d72f8da3e9f663d665f9427379a2539fc9323f6897cc44
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:16253895bc113f395dc6ab5cd511901e43b6ee676369ec80d598a8741c3c9bf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:89bb769b4c33eb931cc7866eeabec9a315b913851cfa84bc6e11ae8d1e8b2cbb
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:bb6fbd68475a7852b4d99eea6c4ab313f9267da7963162f0d75375d7063409e7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b9f93c243ead118651083880786a81c624f390fcc3596c7ec5ebd55f34f95ea3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9cc48eecf1442ae04b8543fa5d4381a13bc2831390850828834d387006d1342b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:47fd4ce2533496828aba37bd1f9715e2164d5c90bd0fc6b25e7e0786d723bf01
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:023478b773a0c42739b86e822a4d435e01ef341b8136566ea52463f1b7e447ed
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:eacc4fb55e8c4c30894a51944960d536508001a49fda62782c6cb2ca8767961d
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0dde839c2dce7cb684094bf26523c8e16677de03149a0fff468b8c3f106e1f4f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:8b4c4400a02c6cacfd978147b158e2ccbe5da955578b4b970a5ee7edac865326
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d42d713da0ce6806fdc6492b6342586783e6865a82a8647d3c4288439b1751ee
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:7f9b921d89b6d4d91b365659afff127d99f1455b94b48a10a6a7f4c38dd8b700
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:90f1678c19c74454b77aed14fa6b3074a1c6f80fe884f1c6136e7e907f402aec
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b2402179b1d0ff9c950defe5d8808195ef9a4c3e13e3c469cae31ab51d5f5014
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:8329acdff170c5a8ae6ae16f6c37d3f7ab68d106a6e8be50c3fe131ec5ecceeb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5f842ebe9a6a4294a056ad10189a5f277cfd3fb4e6e0b3225e194bd5c131151d
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:40920687402fd747ca02d384e63692a5e1db625723a2bf3056a607c92ee8eee1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0f443f5aa818dd85bfa93e51e37bdac3909628a283aa52a89125b527dd75d9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:40ccaf2bb2e0f8132f2541b39cb835f17b38050b3ca9232f0861fd5fcaf4324a
        version: 7.12.5-opr-1+0.1757928550.p
      entries:
      - name: amq-broker-operator.v7.12.5-opr-1-0.1757928550.p
        version: 7.12.5-opr-1+0.1757928550.p
      - name: amq-broker-operator.v7.12.5-opr-1-0.1756216162.p
        version: 7.12.5-opr-1+0.1756216162.p
      - name: amq-broker-operator.v7.12.5-opr-1-0.1754272113.p
        version: 7.12.5-opr-1+0.1754272113.p
      - name: amq-broker-operator.v7.12.5-opr-1
        version: 7.12.5-opr-1
      - name: amq-broker-operator.v7.12.4-opr-1
        version: 7.12.4-opr-1
      - name: amq-broker-operator.v7.12.4-opr-1-0.1747217191.p
        version: 7.12.4-opr-1+0.1747217191.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741881887.p
        version: 7.12.3-opr-1+0.1741881887.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739991235.p
        version: 7.12.3-opr-1+0.1739991235.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1730142550.p
        version: 7.12.3-opr-1+0.1730142550.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1730813578.p
        version: 7.12.3-opr-1+0.1730813578.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1731483798.p
        version: 7.12.3-opr-1+0.1731483798.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1732654016.p
        version: 7.12.3-opr-1+0.1732654016.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1733315503.p
        version: 7.12.3-opr-1+0.1733315503.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1737643631.p
        version: 7.12.3-opr-1+0.1737643631.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1738596517.p
        version: 7.12.3-opr-1+0.1738596517.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739390944.p
        version: 7.12.3-opr-1+0.1739390944.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739889884.p
        version: 7.12.3-opr-1+0.1739889884.p
      - name: amq-broker-operator.v7.12.3-opr-1
        version: 7.12.3-opr-1
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741884937.p
        version: 7.12.3-opr-1+0.1741884937.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741883943.p
        version: 7.12.3-opr-1+0.1741883943.p
      - name: amq-broker-operator.v7.12.2-opr-1
        version: 7.12.2-opr-1
      - name: amq-broker-operator.v7.12.1-opr-1-0.1724054375.p
        version: 7.12.1-opr-1+0.1724054375.p
      - name: amq-broker-operator.v7.12.1-opr-1-0.1723669176.p
        version: 7.12.1-opr-1+0.1723669176.p
      - name: amq-broker-operator.v7.12.1-opr-1-0.1721247907.p
        version: 7.12.1-opr-1+0.1721247907.p
      - name: amq-broker-operator.v7.12.1-opr-1
        version: 7.12.1-opr-1
      - name: amq-broker-operator.v7.12.0-opr-2
        version: 7.12.0-opr-2
      - name: amq-broker-operator.v7.12.0-opr-1
        version: 7.12.0-opr-1
      name: 7.12.x
    defaultChannel: 7.12.x
    packageName: amq-broker-rhel8
    provider:
      name: Red Hat
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecrpublic-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecrpublic-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecrpublic-controller:1.1.1
          createdAt: "2025-09-20T01:21:52Z"
          description: AWS ECR PUBLIC controller is a service controller for managing
            ECR PUBLIC resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Repository represents the state of an AWS ecrpublic Repository
              resource.
            displayName: Repository
            kind: Repository
            name: repositories.ecrpublic.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ECR PUBLIC resources in AWS from within your Kubernetes cluster.

          **About Amazon ECR PUBLIC**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECR PUBLIC
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecrpublic
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECR PUBLIC Developer Resources
          url: https://aws.amazon.com/ECR PUBLIC/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecrpublic maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecrpublic-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ecrpublic-controller.v1.1.1
        version: 1.1.1
      - name: ack-ecrpublic-controller.v1.1.0
        version: 1.1.0
      - name: ack-ecrpublic-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecrpublic-controller.v1.0.12
        version: 1.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecrpublic-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: egressip-ipam-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: egressip-ipam-operator.v1.2.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "EgressIPAM",
                "metadata": {
                  "name": "egressipam-aws"
                },
                "spec": {
                  "cidrAssignments": [
                    {
                      "CIDR": "10.0.128.0/19",
                      "labelValue": "us-west-2a"
                    },
                    {
                      "CIDR": "10.0.160.0/19",
                      "labelValue": "us-west-2b",
                      "reservedIPs": [
                        "10.0.160.126"
                      ]
                    },
                    {
                      "CIDR": "10.0.192.0/19",
                      "labelValue": "us-west-2c"
                    }
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "topologyLabel": "failure-domain.beta.kubernetes.io/zone"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "EgressIPAM",
                "metadata": {
                  "name": "egressipam-baremetal"
                },
                "spec": {
                  "cidrAssignments": [
                    {
                      "CIDR": "192.169.0.0/24",
                      "labelValue": "true"
                    }
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "topologyLabel": "egressGateway"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Networking, Cloud Provider, Security
          certified: "false"
          containerImage: quay.io/redhat-cop/egressip-ipam-operator@sha256:d6129bd287b87696a23e0da64dcd598958f779dc88ae33b46b3c909dbab80adf
          createdAt: "2022-07-25T15:07:22Z"
          description: This operator provides automation around managing egressIPs.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: egressip-ipam-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/egressip-ipam-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgressIPAM is the Schema for the egressipams API
            displayName: Egress IPAM
            kind: EgressIPAM
            name: egressipams.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator automates the assignment of egressIPs to namespaces.
          Namespaces can opt in to receiving one or more egressIPs with the following annotation `egressip-ipam-operator.redhat-cop.io/egressipam:<egressIPAM>`, where `egressIPAM` is the CRD that controls how egressIPs are assigned.
          IPs assigned to the namespace can be looked up in the following annotation: `egressip-ipam-operator.redhat-cop.io/egressips`.

          EgressIP assignments is managed by the EgressIPAM CRD. Here is an example of it:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: EgressIPAM
          metadata:
            name: example-egressipam
          spec:
            cidrAssignments:
              - labelValue: "true"
                CIDR: 192.169.0.0/24
                reservedIPs:
                - "192.159.0.5"
            topologyLabel: egressGateway
            nodeSelector:
              node-role.kubernetes.io/worker: ""
          ```

          This EgressIPAM specifies that all nodes that comply with the specified node selector and that also have labels `egressGateway: "true"` will be assigned egressIP from the specified CIDR.

          Note that the `cidrAssigments` field is an array and therefore, multiple groups of nodes can be identified with the labelValue and different CIDRs can be assigned to them. This is usually not necessary on a bare metal deployment.

          In the bare metal scenario, when this egressCRD is created, all the `hostsubnet` relative to the nodes selected by this EgressIPAM will be update to have the EgressCIDRs field equal to the specified CIDR (see below for cloud provider scenarios).

          When a namespace is created with the opt-in annotation: `egressip-ipam-operator.redhat-cop.io/egressipam=<egressIPAM>`, an available egressIP is selected from the CIDR and assigned to the namespace.The `netnamespace` associated with this namespace is updated to use that egressIP.

          It is possible to specify a set of reserved IPs. These IPs must belong to the CIDR and will never be assigned.

          ## Passing EgressIPs as input

          The normal mode of operation of this operator is to pick a random IP from the configured CIDR. However, it also supports a scenario where egressIPs are picked by an external process and passed as input.

          In this case IPs must me passed as an annotation to the namespace, like this: `egressip-ipam-operator.redhat-cop.io/egressips=IP1,IP2...`. The value of the annotation is a comma separated array of ip with no spaces.

          There must be exactly one IP per CIDR defined in the referenced egressIPAM. Moreover, each IP must belong to the corresponding CIDR. Because this situation can lead to inconsistencies, failing to have correspondence between IPs in the namespace annotation and CIDRs in the egressIPAM CR will cause the operator to error out and stop processing all the namespaces associated with the given EgressIPAM CR.

          It also also a responsibility of the progress picking the IPs to ensure that those IPs are actually available.

          ## Assumptions

          The following assumptions apply when using this operator:

          1. If multiple EgressIPAMs are defined, their selected nodes MUST NOT overlap.
          2. When using a cloud provider the topology label MUST be `failure-domain.beta.kubernetes.io/zone`.

          ## Support for AWS

          In AWS as well as other cloud providers, one cannot freely assign IPs to machines. Additional steps need to be performed in this case. Considering this EgressIPAM

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: EgressIPAM
          metadata:
            name: egressipam-aws
          spec:
            cidrAssignments:
              - labelValue: "eu-central-1a"
                CIDR: 10.0.128.0/20
              - labelValue: "eu-central-1b"
                CIDR: 10.0.144.0/20
              - labelValue: "eu-central-1c"
                CIDR: 10.0.160.0/20
            topologyLabel: failure-domain.beta.kubernetes.io/zone
            nodeSelector:
              matchLabels:
                node-role.kubernetes.io/worker: ""
          ```

          When a namespace with the opt-in annotation is created, the following happens:

          1. for each of the CIDRs, an available IP is selected and assigned to the namespace
          2. the relative `netnamespace` is update to reflect the assignment (multiple IPs will be assigned in this case).
          3. one node per zone is selected to carry the egressIP
          4. the relative aws machines are assigned the additional IP on the main interface (support for secondary interfaces in not available)
          5. the relative `hostsubnets` are updated to reflect the assigned IP, the `egressIP` field is updated.
        displayName: Egressip Ipam Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - egressIPs
        - high availability
        - firewall support
        - traffic management
        - ipam
        links:
        - name: Egressip Ipam Operator
          url: https://github.com/redhat-cop/egressip-ipam-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/egressip-ipam-operator:v0.1.2
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/egressip-ipam-operator@sha256:d6129bd287b87696a23e0da64dcd598958f779dc88ae33b46b3c909dbab80adf
        version: 1.2.4
      entries:
      - name: egressip-ipam-operator.v1.2.4
        version: 1.2.4
      - name: egressip-ipam-operator.v1.2.3
        version: 1.2.3
      - name: egressip-ipam-operator.v1.2.2
        version: 1.2.2
      - name: egressip-ipam-operator.v1.2.1
        version: 1.2.1
      - name: egressip-ipam-operator.v1.2.0
        version: 1.2.0
      - name: egressip-ipam-operator.v1.1.0
        version: 1.1.0
      - name: egressip-ipam-operator.v1.0.8
        version: 1.0.8
      - name: egressip-ipam-operator.v1.0.7
        version: 1.0.7
      - name: egressip-ipam-operator.v1.0.6
        version: 1.0.6
      - name: egressip-ipam-operator.v1.0.5
        version: 1.0.5
      - name: egressip-ipam-operator.v1.0.4
        version: 1.0.4
      - name: egressip-ipam-operator.v1.0.3
        version: 1.0.3
      - name: egressip-ipam-operator.v1.0.2
        version: 1.0.2
      - name: egressip-ipam-operator.v1.0.1
        version: 1.0.1
      - name: egressip-ipam-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: egressip-ipam-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: hive-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hive-operator.v1.2.4953-c248a51
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:c248a51ae1
          createdAt: "2025-09-22T13:41:19Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeploymentCustomization is the Schema for clusterdeploymentcustomizations
              API.
            displayName: ClusterDeploymentCustomization
            kind: ClusterDeploymentCustomization
            name: clusterdeploymentcustomizations.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: |-
              ClusterPool represents a pool of clusters that should be kept ready to be given out to users. Clusters are removed
              from the pool once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: |-
              MachinePoolNameLease is the Schema for the MachinePoolNameLeases API. This resource is mostly empty
              as we're primarily relying on the name to determine if a lease is available.
              Note that not all cloud providers require the use of a lease for naming, at present this
              is only required for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * IBM Cloud
          * Red Hat OpenStack
          * vSphere

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Troubleshooting](https://github.com/openshift/hive/blob/master/docs/troubleshooting.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
              logLevel: debug
              targetNamespace: hive
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:c248a51ae1
        version: 1.2.4953-c248a51
      entries:
      - name: hive-operator.v1.2.4953-c248a51
        version: 1.2.4953-c248a51
      - name: hive-operator.v1.2.4928-76a65b6
        version: 1.2.4928-76a65b6
      - name: hive-operator.v1.2.4902-4929f16
        version: 1.2.4902-4929f16
      - name: hive-operator.v1.2.4896-a29a238
        version: 1.2.4896-a29a238
      - name: hive-operator.v1.2.4868-45ae912
        version: 1.2.4868-45ae912
      - name: hive-operator.v1.2.4863-e9ad0fa
        version: 1.2.4863-e9ad0fa
      - name: hive-operator.v1.2.4774-87bff59
        version: 1.2.4774-87bff59
      - name: hive-operator.v1.2.4766-6a82857
        version: 1.2.4766-6a82857
      - name: hive-operator.v1.2.4746-d25e0db
        version: 1.2.4746-d25e0db
      - name: hive-operator.v1.2.4738-b19773e
        version: 1.2.4738-b19773e
      - name: hive-operator.v1.2.4731-41f153f
        version: 1.2.4731-41f153f
      - name: hive-operator.v1.2.4703-02f3f00
        version: 1.2.4703-02f3f00
      - name: hive-operator.v1.2.4693-ec3bc72
        version: 1.2.4693-ec3bc72
      - name: hive-operator.v1.2.4686-553c6c2
        version: 1.2.4686-553c6c2
      - name: hive-operator.v1.2.4682-406d6ca
        version: 1.2.4682-406d6ca
      - name: hive-operator.v1.2.4672-af54e2f
        version: 1.2.4672-af54e2f
      - name: hive-operator.v1.2.4668-3ccd44d
        version: 1.2.4668-3ccd44d
      - name: hive-operator.v1.2.4664-00f84cc
        version: 1.2.4664-00f84cc
      - name: hive-operator.v1.2.4658-c13e207
        version: 1.2.4658-c13e207
      - name: hive-operator.v1.2.4654-3268a18
        version: 1.2.4654-3268a18
      - name: hive-operator.v1.2.4650-531d549
        version: 1.2.4650-531d549
      - name: hive-operator.v1.2.4644-365f074
        version: 1.2.4644-365f074
      - name: hive-operator.v1.2.4638-56d6d64
        version: 1.2.4638-56d6d64
      - name: hive-operator.v1.2.4625-389d717
        version: 1.2.4625-389d717
      - name: hive-operator.v1.2.4615-52100dd
        version: 1.2.4615-52100dd
      - name: hive-operator.v1.2.4605-b41ca3f
        version: 1.2.4605-b41ca3f
      - name: hive-operator.v1.2.4584-95efcb7
        version: 1.2.4584-95efcb7
      - name: hive-operator.v1.2.4574-1ec27ad
        version: 1.2.4574-1ec27ad
      - name: hive-operator.v1.2.4568-b3eca52
        version: 1.2.4568-b3eca52
      - name: hive-operator.v1.2.4564-bdf9d08
        version: 1.2.4564-bdf9d08
      - name: hive-operator.v1.2.4552-bf2d689
        version: 1.2.4552-bf2d689
      - name: hive-operator.v1.2.4532-cddd4f9
        version: 1.2.4532-cddd4f9
      - name: hive-operator.v1.2.4511-1fce770
        version: 1.2.4511-1fce770
      - name: hive-operator.v1.2.4500-b1ac27b
        version: 1.2.4500-b1ac27b
      - name: hive-operator.v1.2.4485-47df0d9
        version: 1.2.4485-47df0d9
      - name: hive-operator.v1.2.4469-7be6433
        version: 1.2.4469-7be6433
      - name: hive-operator.v1.2.4463-d7ead60
        version: 1.2.4463-d7ead60
      - name: hive-operator.v1.2.4461-97686bb
        version: 1.2.4461-97686bb
      - name: hive-operator.v1.2.4459-f910c9b
        version: 1.2.4459-f910c9b
      - name: hive-operator.v1.2.4451-83aedb9
        version: 1.2.4451-83aedb9
      - name: hive-operator.v1.2.4446-e11a23b
        version: 1.2.4446-e11a23b
      - name: hive-operator.v1.2.4441-fdbfaf0
        version: 1.2.4441-fdbfaf0
      - name: hive-operator.v1.2.4436-44323bc
        version: 1.2.4436-44323bc
      - name: hive-operator.v1.2.4428-ceaafa8
        version: 1.2.4428-ceaafa8
      - name: hive-operator.v1.2.4426-fb27114
        version: 1.2.4426-fb27114
      - name: hive-operator.v1.2.4420-fe0e302
        version: 1.2.4420-fe0e302
      - name: hive-operator.v1.2.4418-eb5403f
        version: 1.2.4418-eb5403f
      - name: hive-operator.v1.2.4416-fba8d52
        version: 1.2.4416-fba8d52
      - name: hive-operator.v1.2.4408-9c5806a
        version: 1.2.4408-9c5806a
      - name: hive-operator.v1.2.4404-410fed9
        version: 1.2.4404-410fed9
      - name: hive-operator.v1.2.4400-8c54fc9
        version: 1.2.4400-8c54fc9
      - name: hive-operator.v1.2.4386-6f1fd3e
        version: 1.2.4386-6f1fd3e
      - name: hive-operator.v1.2.4382-0e95d5b
        version: 1.2.4382-0e95d5b
      - name: hive-operator.v1.2.4366-df65e7e
        version: 1.2.4366-df65e7e
      - name: hive-operator.v1.2.4360-de169dd
        version: 1.2.4360-de169dd
      - name: hive-operator.v1.2.4355-e5f809f
        version: 1.2.4355-e5f809f
      name: alpha
    - currentCSV: hive-operator.v2.5.3516-a2ed9b3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.5.3516-a2ed9b3
          createdAt: "2022-06-03T13:24:12Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:v2.5.3516-a2ed9b3
        version: 2.5.3516-a2ed9b3
      entries:
      - name: hive-operator.v2.5.3516-a2ed9b3
        version: 2.5.3516-a2ed9b3
      - name: hive-operator.v2.5.3514-4052426
        version: 2.5.3514-4052426
      - name: hive-operator.v2.5.3508-6cb94c6
        version: 2.5.3508-6cb94c6
      - name: hive-operator.v2.5.3506-f42617c
        version: 2.5.3506-f42617c
      - name: hive-operator.v2.5.3495-5cfdd44
        version: 2.5.3495-5cfdd44
      - name: hive-operator.v2.5.3491-6feded1
        version: 2.5.3491-6feded1
      - name: hive-operator.v2.5.3489-6dec9c3
        version: 2.5.3489-6dec9c3
      name: mce-2.0
    - currentCSV: hive-operator.v2.3.3039-60b8a9a
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.3.3039-60b8a9a
          createdAt: "2022-08-15T17:23:39Z"
          description: OpenShift cluster provisioning and management at scale.
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:v2.3.3039-60b8a9a
        version: 2.3.3039-60b8a9a
      entries:
      - name: hive-operator.v2.3.3039-60b8a9a
        version: 2.3.3039-60b8a9a
      name: ocm-2.3
    - currentCSV: hive-operator.v2.4.3231-ba51985
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.4.3231-ba51985
          createdAt: "2022-08-15T11:53:28Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:v2.4.3231-ba51985
        version: 2.4.3231-ba51985
      entries:
      - name: hive-operator.v2.4.3231-ba51985
        version: 2.4.3231-ba51985
      - name: hive-operator.v2.4.3227-74831db
        version: 2.4.3227-74831db
      - name: hive-operator.v2.4.3222-d68346c
        version: 2.4.3222-d68346c
      - name: hive-operator.v2.4.3206-b0714a7
        version: 2.4.3206-b0714a7
      - name: hive-operator.v2.4.3203-672cfe7
        version: 2.4.3203-672cfe7
      - name: hive-operator.v2.4.3201-8c50440
        version: 2.4.3201-8c50440
      - name: hive-operator.v2.4.3199-cce4cb9
        version: 2.4.3199-cce4cb9
      - name: hive-operator.v2.4.3197-6c9c8e4
        version: 2.4.3197-6c9c8e4
      - name: hive-operator.v2.4.3194-f82b7d8
        version: 2.4.3194-f82b7d8
      - name: hive-operator.v2.4.3192-fae0111
        version: 2.4.3192-fae0111
      - name: hive-operator.v2.4.3189-026033b
        version: 2.4.3189-026033b
      - name: hive-operator.v2.4.3180-7299056
        version: 2.4.3180-7299056
      - name: hive-operator.v1.2.3179-0756b70e6
        version: 1.2.3179-0756b70e6
      - name: hive-operator.v1.1.16
        version: 1.1.16
      - name: hive-operator.v1.1.15
        version: 1.1.15
      - name: hive-operator.v1.1.14
        version: 1.1.14
      - name: hive-operator.v1.1.13
        version: 1.1.13
      - name: hive-operator.v1.1.12
        version: 1.1.12
      - name: hive-operator.v1.1.11
        version: 1.1.11
      - name: hive-operator.v1.1.10
        version: 1.1.10
      - name: hive-operator.v1.1.9
        version: 1.1.9
      - name: hive-operator.v1.1.8
        version: 1.1.8
      - name: hive-operator.v1.1.7
        version: 1.1.7
      - name: hive-operator.v1.1.6
        version: 1.1.6
      name: ocm-2.4
    defaultChannel: alpha
    packageName: hive-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-admin-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-admin-operator.v2024.10.2-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": null
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:ed279ff5416cce855118f66a73d55385aba26af19735eed052010ba653f64b75
          createdAt: "2024-10-25T01:47:18Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2024.10.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2023.9.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:7c4868db021beaffa762eb0de43739bbc865a4bf5900f1d17f78caed8612c073
        - registry.connect.redhat.com/ca/uma-operator@sha256:ed279ff5416cce855118f66a73d55385aba26af19735eed052010ba653f64b75
        version: 2024.10.2-12
      entries:
      - name: uma-admin-operator.v2024.10.2-12
        version: 2024.10.2-12
      name: "2024.10"
    - currentCSV: uma-admin-operator.v2025.1.3-33
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.1.1.33,990033",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": null
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:6a932a065add9f36f9615a93037315603f2d9cd8a2e3c64b1b1fc390bb4f1dcf
          createdAt: "2025-02-18T05:34:34Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.1.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2023.9.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.1.1.26,990026"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:6a932a065add9f36f9615a93037315603f2d9cd8a2e3c64b1b1fc390bb4f1dcf
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:ae91a28d57c8a359f81757e4bf79472f2ac189ac0ae3ee0d87e8ff9c433f1ccc
        version: 2025.1.3-33
      entries:
      - name: uma-admin-operator.v2025.1.3-33
        version: 2025.1.3-33
      name: "2025.1"
    - currentCSV: uma-admin-operator.v2025.2.1-25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.2.1.25,990025",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverCollectorCert": "",
                    "serverCollectorKey": "",
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:d201790c2d9ded6bc7a1c5f68a267574e7b5d39c39ce2f98e20ef630b0f7365f
          createdAt: "2025-03-05T05:51:02Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.2.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2023.9.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.1.1.26,990026"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:d201790c2d9ded6bc7a1c5f68a267574e7b5d39c39ce2f98e20ef630b0f7365f
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:8bcf0dac889f38a0bf77eb9babe3f726e3b2564d5988c05bdc5c2d91ae189c61
        version: 2025.2.1-25
      entries:
      - name: uma-admin-operator.v2025.2.1-25
        version: 2025.2.1-25
      name: "2025.2"
    - currentCSV: uma-admin-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:4c2a7f68a47428f3e386107254f74fa8c34e477eb40fc17c173df92f3c66c559
          createdAt: "2025-04-16T06:55:39Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.4.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.4.2.5,990005"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        - registry.connect.redhat.com/ca/uma-operator@sha256:4c2a7f68a47428f3e386107254f74fa8c34e477eb40fc17c173df92f3c66c559
        version: 2025.4.5-5
      entries:
      - name: uma-admin-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-admin-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:514b953eaf0959bc120f80bdc0a2e19d8f059d3fe2883a4f6843a398470a56b1
          createdAt: "2025-07-24T14:54:21Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.5.2.12,990012"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:514b953eaf0959bc120f80bdc0a2e19d8f059d3fe2883a4f6843a398470a56b1
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        version: 2025.5.4-12
      entries:
      - name: uma-admin-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-admin-operator.v2025.7.3-21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:188a2b8d2255ded6141a4b795f1b6d7bd98017b4fd57c9c1a1d38fc701358057
          createdAt: "2025-08-23T09:00:37Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.3
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.7.1.23,990023"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:188a2b8d2255ded6141a4b795f1b6d7bd98017b4fd57c9c1a1d38fc701358057
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.3-21
      entries:
      - name: uma-admin-operator.v2025.7.3-21
        version: 2025.7.3-21
      name: "2025.7"
    - currentCSV: uma-admin-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:0cfb63c62258e54d67e7baca874003a04477f30bcf7f56861f565ad068070580
          createdAt: "2025-08-15T03:20:28Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-admin-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.8.1.9,990009"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:0cfb63c62258e54d67e7baca874003a04477f30bcf7f56861f565ad068070580
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        version: 2025.8.2-9
      entries:
      - name: uma-admin-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-admin-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
          createdAt: "2025-09-24T12:23:56Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.12,990012"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-admin-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-admin-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:7b93de818fa5160e72f878b0350d13fc1d5943eac95b0794d3d4f930cf21b59a
          createdAt: "2025-07-30T08:07:39Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-admin-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "99.99.0.serp_F152813,19"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:7b93de818fa5160e72f878b0350d13fc1d5943eac95b0794d3d4f930cf21b59a
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        version: 2025.7.2-19
      entries:
      - name: uma-admin-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-admin-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1cf3909dd2caeabc53209d6f8565dcb6c3562b9c6eb715dda8ba1953f264668e
          createdAt: "2025-09-11T11:16:06Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.8,990008"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:1cf3909dd2caeabc53209d6f8565dcb6c3562b9c6eb715dda8ba1953f264668e
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-admin-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-admin-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
          createdAt: "2025-09-24T12:23:56Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.12,990012"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-admin-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-admin-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: octavia-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: octavia-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "octavia.openstack.org/v1beta1",
                "kind": "Octavia",
                "metadata": {
                  "name": "octavia"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "octavia",
                  "databaseInstance": "openstack",
                  "lbMgmtNetwork": {
                    "availabilityZones": [
                      "zone-1"
                    ]
                  },
                  "octaviaAPI": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "octavia",
                    "databaseInstance": "openstack",
                    "persistenceDatabaseAccount": "octavia-persistence",
                    "preserveJobs": false,
                    "secret": "osp-secret",
                    "serviceAccount": "octavia",
                    "serviceUser": "octavia"
                  },
                  "octaviaHealthManager": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "octavia",
                    "databaseInstance": "openstack",
                    "networkAttachments": [
                      "octavia"
                    ],
                    "persistenceDatabaseAccount": "octavia-persistence",
                    "preserveJobs": false,
                    "role": "healthmanager",
                    "secret": "osp-secret",
                    "serviceAccount": "octavia",
                    "serviceUser": "octavia"
                  },
                  "octaviaHousekeeping": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "octavia",
                    "databaseInstance": "openstack",
                    "networkAttachments": [
                      "octavia"
                    ],
                    "persistenceDatabaseAccount": "octavia-persistence",
                    "preserveJobs": false,
                    "role": "housekeeping",
                    "secret": "osp-secret",
                    "serviceAccount": "octavia",
                    "serviceUser": "octavia"
                  },
                  "octaviaWorker": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "octavia",
                    "databaseInstance": "openstack",
                    "networkAttachments": [
                      "octavia"
                    ],
                    "persistenceDatabaseAccount": "octavia-persistence",
                    "preserveJobs": false,
                    "role": "worker",
                    "secret": "osp-secret",
                    "serviceAccount": "octavia",
                    "serviceUser": "octavia"
                  },
                  "persistenceDatabaseAccount": "octavia-persistence",
                  "preserveJobs": false,
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "osp-secret",
                  "serviceUser": "octavia"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:01Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: Octavia Amphora Controller
            kind: OctaviaAmphoraController
            name: octaviaamphoracontrollers.octavia.openstack.org
            version: v1beta1
          - description: OctaviaAPI is the Schema for the octaviaapis API
            displayName: Octavia API
            kind: OctaviaAPI
            name: octaviaapis.octavia.openstack.org
            version: v1beta1
          - kind: OctaviaRsyslog
            name: octaviarsyslogs.octavia.openstack.org
            version: v1beta1
          - description: Octavia is the Schema for the octavia API
            displayName: Octavia
            kind: Octavia
            name: octavias.octavia.openstack.org
            version: v1beta1
        description: Octavia Operator
        displayName: Octavia Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Load-Balancer
        - Octavia
        links:
        - name: Octavia Operator
          url: https://github.com/openstack-k8s-operators/octavia-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/octavia-rhel9-operator@sha256:ddd280ec099cde63e38bd964d7f0608c1c34e5c9f6b140f86dde2126b0357ef5
        - registry.redhat.io/rhoso/openstack-octavia-health-manager-rhel9@sha256:7d305af6e9e6c636510186dd80037849d19bb1e6d2b2507c025d1d0d9254874a
        - registry.redhat.io/rhoso/openstack-octavia-worker-rhel9@sha256:0919ddff6421430f8ad24cb21a89c65f02f98df424267f7e436c97d555112e2a
        - registry.redhat.io/ubi9/httpd-24@sha256:dbfbe3383e12d63df7178e4cdd5b1e6e51a84c781041b9192026ea6edb72fbf3
        - registry.redhat.io/rhoso/openstack-octavia-api-rhel9@sha256:ae9aae94c2523933a346bdade16158b4302b3a0bee93dec89bae5d15606c8dfb
        - registry.redhat.io/rhoso/openstack-octavia-housekeeping-rhel9@sha256:d54868077feb205199b76e83fc9220d9b4215dbeba037cf13fb68dc6577a07dc
        version: 1.0.6
      entries:
      - name: octavia-operator.v1.0.6
        version: 1.0.6
      - name: octavia-operator.v1.0.5
        version: 1.0.5
      - name: octavia-operator.v1.0.4
        version: 1.0.4
      - name: octavia-operator.v1.0.3
        version: 1.0.3
      - name: octavia-operator.v1.0.2
        version: 1.0.2
      - name: octavia-operator.v1.0.1
        version: 1.0.1
      - name: octavia-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: octavia-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: gatekeeper-operator-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gatekeeper-operator-product.v3.11.2-0.1725401426.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel8-operator@sha256:c8894e4d02fb2c822ab2e02e489dd9b67c953d96bd3b96a8305219a2245dfd99
          createdAt: 21 Mar 2024, 01:59
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.11.0
          olm.substitutesFor: gatekeeper-operator-product.v3.11.2-0.1721233953.p
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/gatekeeper/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the Schema for the gatekeepers API
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |-
          Gatekeeper allows administrators to detect and reject non-compliant commits to an infrastructure-as-code system's source-of-truth. This strengthens compliance efforts and prevents a bad state from slowing down the organization.

          ## Support & Troubleshooting
          Gatekeeper is available and supported as part of a valid Red Hat software subscription when it is used with Red Hat Advanced Cluster Management for Kubernetes. Troubleshooting information is available in the Gatekeeper documentation.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/gatekeeper/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - registry.redhat.io/gatekeeper/gatekeeper-rhel8-operator@sha256:c8894e4d02fb2c822ab2e02e489dd9b67c953d96bd3b96a8305219a2245dfd99
        - registry.redhat.io/gatekeeper/gatekeeper-rhel8@sha256:701b165def9e4b92d91b8eac4917721e76e634d2159b8efc69bef23e10b51b9d
        version: 3.11.2+0.1725401426.p
      entries:
      - name: gatekeeper-operator-product.v3.11.2-0.1725401426.p
        version: 3.11.2+0.1725401426.p
      - name: gatekeeper-operator-product.v3.11.2-0.1721233953.p
        version: 3.11.2+0.1721233953.p
      - name: gatekeeper-operator-product.v3.11.2-0.1718224960.p
        version: 3.11.2+0.1718224960.p
      - name: gatekeeper-operator-product.v3.11.2
        version: 3.11.2
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.11"
    - currentCSV: gatekeeper-operator-product.v3.14.3-0.1746550072.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e47f0e9da800eb0cf4a0da6af98986508a35798a41581158ee7dbc977cb651dc
          createdAt: 16 Jan 2025, 06:57
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.14.3
          olm.substitutesFor: gatekeeper-operator-product.v3.14.3-0.1744033158.p
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the Schema for the gatekeepers API
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |-
          Gatekeeper allows administrators to detect and reject non-compliant commits to an infrastructure-as-code system's source-of-truth. This strengthens compliance efforts and prevents a bad state from slowing down the organization.

          ## Support & Troubleshooting
          Gatekeeper is available and supported as part of a valid Red Hat software subscription when it is used with Red Hat Advanced Cluster Management for Kubernetes. Troubleshooting information is available in the Gatekeeper documentation.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e47f0e9da800eb0cf4a0da6af98986508a35798a41581158ee7dbc977cb651dc
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:6255f7a673e6989617176842d8f0cc897e4630b5283e812a7a2d52dbb2a1f335
        version: 3.14.3+0.1746550072.p
      entries:
      - name: gatekeeper-operator-product.v3.14.3-0.1746550072.p
        version: 3.14.3+0.1746550072.p
      - name: gatekeeper-operator-product.v3.14.3-0.1744033158.p
        version: 3.14.3+0.1744033158.p
      - name: gatekeeper-operator-product.v3.14.3-0.1742934403.p
        version: 3.14.3+0.1742934403.p
      - name: gatekeeper-operator-product.v3.14.3-0.1740676608.p
        version: 3.14.3+0.1740676608.p
      - name: gatekeeper-operator-product.v3.14.3
        version: 3.14.3
      - name: gatekeeper-operator-product.v3.14.2
        version: 3.14.2
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.14"
    - currentCSV: gatekeeper-operator-product.v3.15.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:8b43248aed06563e06f7a0347ff1c45fa5892e3a9adb2b5c363962a0b152eff0
          createdAt: 13 Mar 2025, 17:07
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.15.4
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: Gatekeeper allows administrators to detect and reject non-compliant
          commits to an infrastructure-as-code system's source-of-truth. This strengthens
          compliance efforts and prevents a bad state from slowing down the organization.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:8b43248aed06563e06f7a0347ff1c45fa5892e3a9adb2b5c363962a0b152eff0
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:29690144667b3753a7e76095e136d6fc42325270e7b9f22f97dfd411439e9f54
        version: 3.15.4
      entries:
      - name: gatekeeper-operator-product.v3.15.4
        version: 3.15.4
      - name: gatekeeper-operator-product.v3.15.3
        version: 3.15.3
      - name: gatekeeper-operator-product.v3.15.2
        version: 3.15.2
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.15"
    - currentCSV: gatekeeper-operator-product.v3.17.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:3c027ec78398df0a8fc09eb79e222a9508b4f18154d1e1b477d4f156181ae51a
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.17.3
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e44078fc6073a79168ed52922b370773b451d69bf335aeb91d26a1462df21a9b
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:3c027ec78398df0a8fc09eb79e222a9508b4f18154d1e1b477d4f156181ae51a
        version: 3.17.3
      entries:
      - name: gatekeeper-operator-product.v3.17.3
        version: 3.17.3
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.17"
    - currentCSV: gatekeeper-operator-product.v3.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:0803b7219ceff7aa676488eeba58607a19c9c18a2bb8eae8b3b16cfec411cb6e
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.18.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:49726cb2702441305e8047b24054575170fb595cfe5ebf815cdbfdaa957b02f4
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:0803b7219ceff7aa676488eeba58607a19c9c18a2bb8eae8b3b16cfec411cb6e
        version: 3.18.1
      entries:
      - name: gatekeeper-operator-product.v3.18.1
        version: 3.18.1
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.18"
    - currentCSV: gatekeeper-operator-product.v3.19.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.19.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:60ee347425c6cfb36f0e1e834b33cbb3d77aa2d6c89ac513216d0b9fc55d4b52
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
        version: 3.19.1
      entries:
      - name: gatekeeper-operator-product.v3.19.1
        version: 3.19.1
      - name: gatekeeper-operator-product.v3.19.0
        version: 3.19.0
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.19"
    - currentCSV: gatekeeper-operator-product.v3.19.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.19.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:60ee347425c6cfb36f0e1e834b33cbb3d77aa2d6c89ac513216d0b9fc55d4b52
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
        version: 3.19.1
      entries:
      - name: gatekeeper-operator-product.v3.19.1
        version: 3.19.1
      - name: gatekeeper-operator-product.v3.19.0
        version: 3.19.0
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: stable
    defaultChannel: stable
    packageName: gatekeeper-operator-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://github.com/redhat-developer/app-services-operator/
    name: rhoas-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhoas-operator.0.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "ServiceRegistryConnection",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace",
                  "labels": {
                    "app.kubernetes.io/component": "external-service",
                    "app.kubernetes.io/managed-by": "rhoas"
                  }
                },
                "spec": {
                  "accessTokenSecretName": "rh-managed-services-api-accesstoken",
                  "serviceRegistryId": "exampleId",
                  "credentials": {
                    "serviceAccountSecretName": "service-account-secret"
                  }
                }
              },
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "CloudServiceAccountRequest",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace"
                },
                "spec": {
                  "serviceAccountName": "rhoas-sa",
                  "serviceAccountDescription": "Operator created service account",
                  "serviceAccountSecretName": "service-account-credentials",
                  "accessTokenSecretName": "rh-managed-services-api-accesstoken"
                }
              },
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "CloudServicesRequest",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace",
                  "labels": {
                    "app.kubernetes.io/component": "external-service",
                    "app.kubernetes.io/managed-by": "rhoas"
                  }
                },
                "spec": {
                  "accessTokenSecretName": "rh-cloud-services-api-accesstoken"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: quay.io/rhoas/service-operator:0.13.0
          createdAt: "2023-01-20 16:09:37"
          description: Red Hat OpenShift Application Services Operator
          olm.skipRange: '>=0.0.1 <0.13.0'
          repository: https://github.com/redhat-developer/app-services-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Creates RHOAS Service Accounts
            displayName: Cloud Service Account Request
            kind: CloudServiceAccountRequest
            name: cloudserviceaccountrequests.rhoas.redhat.com
            version: v1alpha1
          - description: Cloud Services Request
            displayName: Cloud Services Request
            kind: CloudServicesRequest
            name: cloudservicesrequests.rhoas.redhat.com
            version: v1alpha1
          - description: Binds Kafka Instance
            displayName: Kafka Connection
            kind: KafkaConnection
            name: kafkaconnections.rhoas.redhat.com
            version: v1alpha1
          - description: Binds Service Registry Instance
            displayName: Service Registry Connection
            kind: ServiceRegistryConnection
            name: serviceregistryconnections.rhoas.redhat.com
            version: v1alpha1
        description: "**OpenShift Application Services Operator (RHOAS-Operator)**
          creates instances of Red Hat OpenShift Application Services in your cluster.
          Operator is currently available as limited Beta and it cannot be consumed
          without invitation.\n\nRed Hat OpenShift Application Services, when added
          to Red Hat OpenShift Dedicated,  provide a streamlined developer experience
          for building, deploying,  and scaling cloud-native applications in open
          hybrid-cloud environments.   \n\nRHOAS-Operator enables developers to connect
          with their Application Services directly in their own cluster by using [RHOAS
          cli](https://github.com/redhat-developer/app-services-cli) and OpenShift
          Console UI.  \n\nRHOAS-Operator requires developers to connect with their
          Red Hat OpenShift Streams for Apache Kafka (RHOSAK) instances.  "
        displayName: OpenShift Application Services (RHOAS)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - rhoas
        - kafka
        - serviceregistry
        - app-services
        - application
        - services
        links:
        - name: RHOAS Operator Documentation
          url: https://github.com/redhat-developer/app-services-operator/tree/main/docs
        maintainers:
        - email: rhosak-eval-support@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://github.com/redhat-developer/app-services-operator/
        relatedImages:
        - quay.io/rhoas/service-operator:0.13.0
        version: 0.13.0
      entries:
      - name: rhoas-operator.0.13.0
        version: 0.13.0
      - name: rhoas-operator.0.12.0
        version: 0.12.0
      - name: rhoas-operator.0.11.0
        version: 0.11.0
      - name: rhoas-operator.0.9.6
        version: 0.9.6
      - name: rhoas-operator.0.9.5
        version: 0.9.5
      - name: rhoas-operator.0.9.3
        version: 0.9.3
      - name: rhoas-operator.0.9.2
        version: 0.9.2
      - name: rhoas-operator.0.9.1
        version: 0.9.1
      - name: rhoas-operator.0.9.0
        version: 0.9.0
      - name: rhoas-operator.0.8.2
        version: 0.8.2
      - name: rhoas-operator.0.7.10
        version: 0.7.10
      - name: rhoas-operator.0.7.9
        version: 0.7.9
      - name: rhoas-operator.0.7.8
        version: 0.7.8
      - name: rhoas-operator.0.7.6
        version: 0.7.6
      - name: rhoas-operator.0.7.2
        version: 0.7.2
      - name: rhoas-operator.0.7.1
        version: 0.7.1
      - name: rhoas-operator.0.6.8
        version: 0.6.8
      name: beta
    defaultChannel: beta
    packageName: rhoas-operator
    provider:
      name: Red Hat Inc.
      url: https://github.com/redhat-developer/app-services-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-s3-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-s3-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "s3.services.k8s.aws/v1alpha1",
                "kind": "Bucket",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/s3-controller:1.1.1
          createdAt: "2025-09-19T19:23:45Z"
          description: AWS S3 controller is a service controller for managing S3 resources
            in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Bucket represents the state of an AWS s3 Bucket resource.
            displayName: Bucket
            kind: Bucket
            name: buckets.s3.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Simple Storage Service (S3) resources in AWS from within your Kubernetes cluster.

          **About Amazon S3**

          Amazon Simple Storage Service (Amazon S3) is an object storage service that offers industry-leading scalability, data availability, security, and performance. This means customers of all sizes and industries can use it to store and protect any amount of data for a range of use cases, such as data lakes, websites, mobile applications, backup and restore, archive, enterprise applications, IoT devices, and big data analytics. Amazon S3 provides easy-to-use management features so you can organize your data and configure finely-tuned access controls to meet your specific business, organizational, and compliance requirements. Amazon S3 is designed for 99.999999999% (11 9s) of durability, and stores data for millions of applications for companies all around the world.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon S3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - s3
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon S3 Developer Resources
          url: https://aws.amazon.com/s3/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: s3 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/s3-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-s3-controller.v1.1.1
        version: 1.1.1
      - name: ack-s3-controller.v1.1.0
        version: 1.1.0
      - name: ack-s3-controller.v1.0.35
        version: 1.0.35
      - name: ack-s3-controller.v1.0.33
        version: 1.0.33
      - name: ack-s3-controller.v1.0.32
        version: 1.0.32
      - name: ack-s3-controller.v1.0.31
        version: 1.0.31
      - name: ack-s3-controller.v1.0.30
        version: 1.0.30
      - name: ack-s3-controller.v1.0.29
        version: 1.0.29
      - name: ack-s3-controller.v1.0.28
        version: 1.0.28
      - name: ack-s3-controller.v1.0.27
        version: 1.0.27
      - name: ack-s3-controller.v1.0.26
        version: 1.0.26
      - name: ack-s3-controller.v1.0.25
        version: 1.0.25
      - name: ack-s3-controller.v1.0.24
        version: 1.0.24
      - name: ack-s3-controller.v1.0.23
        version: 1.0.23
      - name: ack-s3-controller.v1.0.21
        version: 1.0.21
      - name: ack-s3-controller.v1.0.20
        version: 1.0.20
      - name: ack-s3-controller.v1.0.19
        version: 1.0.19
      - name: ack-s3-controller.v1.0.18
        version: 1.0.18
      - name: ack-s3-controller.v1.0.17
        version: 1.0.17
      - name: ack-s3-controller.v1.0.16
        version: 1.0.16
      - name: ack-s3-controller.v1.0.15
        version: 1.0.15
      - name: ack-s3-controller.v1.0.14
        version: 1.0.14
      - name: ack-s3-controller.v1.0.13
        version: 1.0.13
      - name: ack-s3-controller.v1.0.12
        version: 1.0.12
      - name: ack-s3-controller.v1.0.11
        version: 1.0.11
      - name: ack-s3-controller.v1.0.9
        version: 1.0.9
      - name: ack-s3-controller.v1.0.7
        version: 1.0.7
      - name: ack-s3-controller.v1.0.6
        version: 1.0.6
      - name: ack-s3-controller.v1.0.5
        version: 1.0.5
      - name: ack-s3-controller.v1.0.4
        version: 1.0.4
      - name: ack-s3-controller.v1.0.3
        version: 1.0.3
      - name: ack-s3-controller.v1.0.2
        version: 1.0.2
      - name: ack-s3-controller.v1.0.1
        version: 1.0.1
      - name: ack-s3-controller.v1.0.0
        version: 1.0.0
      - name: ack-s3-controller.v0.1.8
        version: 0.1.8
      - name: ack-s3-controller.v0.1.7
        version: 0.1.7
      - name: ack-s3-controller.v0.1.6
        version: 0.1.6
      - name: ack-s3-controller.v0.1.5
        version: 0.1.5
      - name: ack-s3-controller.v0.1.4
        version: 0.1.4
      - name: ack-s3-controller.v0.1.3
        version: 0.1.3
      - name: ack-s3-controller.v0.1.2
        version: 0.1.2
      - name: ack-s3-controller.v0.1.1
        version: 0.1.1
      - name: ack-s3-controller.v0.1.0
        version: 0.1.0
      - name: ack-s3-controller.v0.0.20
        version: 0.0.20
      - name: ack-s3-controller.v0.0.19
        version: 0.0.19
      - name: ack-s3-controller.v0.0.18
        version: 0.0.18
      - name: ack-s3-controller.v0.0.17
        version: 0.0.17
      - name: ack-s3-controller.v0.0.16
        version: 0.0.16
      - name: ack-s3-controller.v0.0.15
        version: 0.0.15
      - name: ack-s3-controller.v0.0.14
        version: 0.0.14
      - name: ack-s3-controller.v0.0.13
        version: 0.0.13
      - name: ack-s3-controller.v0.0.12
        version: 0.0.12
      - name: ack-s3-controller.v0.0.11
        version: 0.0.11
      - name: ack-s3-controller.v0.0.10
        version: 0.0.10
      - name: ack-s3-controller.v0.0.9
        version: 0.0.9
      - name: ack-s3-controller.v0.0.8
        version: 0.0.8
      - name: ack-s3-controller.v0.0.7
        version: 0.0.7
      name: alpha
    defaultChannel: alpha
    packageName: ack-s3-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: numaresources-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: numaresources-operator.v4.16.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodetopology.openshift.io/v1alpha1",
                "kind": "NUMAResourcesOperator",
                "metadata": {
                  "name": "numaresourcesoperator"
                },
                "spec": {
                  "nodeGroups": [
                    {
                      "machineConfigPoolSelector": {
                        "matchLabels": {
                          "pools.operator.machineconfiguration.openshift.io/worker": ""
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "nodetopology.openshift.io/v1alpha1",
                "kind": "NUMAResourcesScheduler",
                "metadata": {
                  "name": "numaresourcesscheduler"
                },
                "spec": {
                  "imageSpec": "URL_OF_SCHEDULER_IMAGE_FROM_REDHAT_REGISTRY",
                  "logLevel": "Normal",
                  "schedulerName": "topo-aware-scheduler"
                }
              }
            ]
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "true"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.15.0 <4.16.6'
          operators.openshift.io/valid-subscription: "[\n\t\"OpenShift Kubernetes
            Engine\",\n\t\"OpenShift Container Platform\",\n\t\"OpenShift Platform
            Plus\"\n]"
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NUMAResourcesOperator is the Schema for the numaresourcesoperators
              API
            displayName: NUMA Resources Operator
            kind: NUMAResourcesOperator
            name: numaresourcesoperators.nodetopology.openshift.io
            version: v1
          - description: NUMAResourcesOperator is the Schema for the numaresourcesoperators
              API
            displayName: NUMA Resources Operator
            kind: NUMAResourcesOperator
            name: numaresourcesoperators.nodetopology.openshift.io
            version: v1alpha1
          - description: NUMAResourcesScheduler is the Schema for the numaresourcesschedulers
              API
            displayName: NUMA Aware Scheduler
            kind: NUMAResourcesScheduler
            name: numaresourcesschedulers.nodetopology.openshift.io
            version: v1
          - description: NUMAResourcesScheduler is the Schema for the numaresourcesschedulers
              API
            displayName: NUMA Aware Scheduler
            kind: NUMAResourcesScheduler
            name: numaresourcesschedulers.nodetopology.openshift.io
            version: v1alpha1
        description: NUMA resources exporter operator
        displayName: numaresources-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - RTE
        - NUMA
        links:
        - name: Numaresources Operator
          url: https://github.com/openshift-kni/numaresources-operator
        maintainers:
        - email: fromani@redhat.com
          name: fromani
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/numaresources-rhel9-operator@sha256:aca8232e89172ca9fd9ce4a66cae9e8bb027b6d16482ce6b8e27fb7779c398e2
        version: 4.16.6
      entries:
      - name: numaresources-operator.v4.16.6
        version: 4.16.6
      - name: numaresources-operator.v4.16.5
        version: 4.16.5
      - name: numaresources-operator.v4.16.4
        version: 4.16.4
      - name: numaresources-operator.v4.16.3
        version: 4.16.3
      - name: numaresources-operator.v4.16.2
        version: 4.16.2
      - name: numaresources-operator.v4.16.1
        version: 4.16.1
      - name: numaresources-operator.v4.16.0
        version: 4.16.0
      name: "4.16"
    defaultChannel: "4.16"
    packageName: numaresources-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.9-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lvms-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lvms-operator.v4.15.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "my-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "default": true,
                        "fstype": "xfs",
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:696edeb4c168646b49b8b0ca5b3d06071ab66fdb010e7cd46888d8b2f3a0c7a8
          description: Logical volume manager storage provides dynamically provisioned
            local storage for container workloads
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.11'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "test-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
          operatorframework.io/suggested-namespace: openshift-storage
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-storage",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/internal-objects: '["logicalvolumes.topolvm.io",
            "lvmvolumegroups.lvm.topolvm.io", "lvmvolumegroupnodestatuses.lvm.topolvm.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/lvm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LogicalVolume
            name: logicalvolumes.topolvm.io
            version: v1
          - description: LVMCluster is the Schema for the lvmclusters API
            displayName: LVMCluster
            kind: LVMCluster
            name: lvmclusters.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroupNodeStatus
            name: lvmvolumegroupnodestatuses.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroup
            name: lvmvolumegroups.lvm.topolvm.io
            version: v1alpha1
        description: Logical volume manager storage provides dynamically provisioned
          local storage.
        displayName: LVM Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - local storage
        - operator
        - LVM
        links:
        - name: Source Repository
          url: https://github.com/openshift/lvm-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:771d513c4d4e5d77d797b4c18f86e23162a20c6d39955e93d24f62b94e62e785
        - registry.redhat.io/lvms4/lvms-must-gather-rhel9@sha256:382ceea1151da56d0931403207c9f85721d58a66b39b1cd79790eca6219807d0
        - registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:696edeb4c168646b49b8b0ca5b3d06071ab66fdb010e7cd46888d8b2f3a0c7a8
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:82c01635b32821f04a7cfcf80da8b3f7e097786f220e0c85aab47bda4536a1d6
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:2db4f32824a8f9f6780dc3d236f73684cb20db41466ffa8f9a89fd3f5eb62fdd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:7ce50398cd29df6e07ccd89790a53a711a7e5b484698c48f6581f5d4e9efb05a
        - registry.redhat.io/lvms4/topolvm-rhel9@sha256:074a206d36c9e4198d90baed86f3046abd2ec12f9b36a715b60900011366b3d4
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:05dceaff7f0b2b769ddfd9261cc66d42aab9158ce99ed3dbd6c229e784072635
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:479454d0d4de6609d8a4696b9c7f0097115afb45db3bdd46631d58afd77dcc58
        version: 4.15.11
      entries:
      - name: lvms-operator.v4.15.11
        version: 4.15.11
      - name: lvms-operator.v4.15.10
        version: 4.15.10
      - name: lvms-operator.v4.15.9
        version: 4.15.9
      - name: lvms-operator.v4.15.8
        version: 4.15.8
      - name: lvms-operator.v4.15.7
        version: 4.15.7
      - name: lvms-operator.v4.15.6
        version: 4.15.6
      - name: lvms-operator.v4.15.5
        version: 4.15.5
      - name: lvms-operator.v4.15.4
        version: 4.15.4
      - name: lvms-operator.v4.15.3
        version: 4.15.3
      - name: lvms-operator.v4.15.2
        version: 4.15.2
      - name: lvms-operator.v4.15.1
        version: 4.15.1
      name: stable-4.15
    - currentCSV: lvms-operator.v4.16.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "my-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "default": true,
                        "fstype": "xfs",
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:b9d80bb263a54192501d5ba9366a4f0c19b34875b51bad4fa75ef22c60bb3ee7
          description: Logical volume manager storage provides dynamically provisioned
            local storage for container workloads
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.9'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "test-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
          operatorframework.io/suggested-namespace: openshift-storage
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-storage",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["logicalvolumes.topolvm.io",
            "lvmvolumegroups.lvm.topolvm.io", "lvmvolumegroupnodestatuses.lvm.topolvm.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lvm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LogicalVolume
            name: logicalvolumes.topolvm.io
            version: v1
          - description: LVMCluster is the Schema for the lvmclusters API
            displayName: LVMCluster
            kind: LVMCluster
            name: lvmclusters.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroupNodeStatus
            name: lvmvolumegroupnodestatuses.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroup
            name: lvmvolumegroups.lvm.topolvm.io
            version: v1alpha1
        description: Logical volume manager storage provides dynamically provisioned
          local storage.
        displayName: LVM Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - local storage
        - operator
        - LVM
        links:
        - name: Source Repository
          url: https://github.com/openshift/lvm-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/lvms4/topolvm-rhel9@sha256:227748103049d2cced921ba9fc4e2d3671f08c897cc2a5627d755fe5f9d2767d
        - registry.redhat.io/lvms4/lvms-must-gather-rhel9@sha256:20b28b87b3fe6d6bf739181376231d6f8b7454c34b59eaf2e06c04a9fca8f41d
        - registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:b9d80bb263a54192501d5ba9366a4f0c19b34875b51bad4fa75ef22c60bb3ee7
        version: 4.16.9
      entries:
      - name: lvms-operator.v4.16.9
        version: 4.16.9
      - name: lvms-operator.v4.16.8
        version: 4.16.8
      - name: lvms-operator.v4.16.7
        version: 4.16.7
      - name: lvms-operator.v4.16.6
        version: 4.16.6
      - name: lvms-operator.v4.16.5
        version: 4.16.5
      - name: lvms-operator.v4.16.4
        version: 4.16.4
      - name: lvms-operator.v4.16.3
        version: 4.16.3
      - name: lvms-operator.v4.16.2
        version: 4.16.2
      - name: lvms-operator.v4.16.1
        version: 4.16.1
      - name: lvms-operator.v4.16.0
        version: 4.16.0
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: lvms-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-custom-metrics-autoscaler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: custom-metrics-autoscaler.v2.17.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eventing.keda.sh/v1alpha1",
                "kind": "CloudEventSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keda",
                    "app.kubernetes.io/instance": "cloudeventsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cloudeventsource",
                    "app.kubernetes.io/part-of": "keda"
                  },
                  "name": "cloudeventsource-sample"
                },
                "spec": {
                  "clusterName": "cluster-sample",
                  "destination": {
                    "http": {
                      "uri": "http://foo.bar"
                    }
                  }
                }
              },
              {
                "apiVersion": "eventing.keda.sh/v1alpha1",
                "kind": "ClusterCloudEventSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keda",
                    "app.kubernetes.io/instance": "clustercloudeventsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clustercloudeventsource",
                    "app.kubernetes.io/part-of": "keda"
                  },
                  "name": "clustercloudeventsource-sample"
                },
                "spec": {
                  "clusterName": "clustercluster-sample",
                  "destination": {
                    "http": {
                      "uri": "http://foo.bar"
                    }
                  }
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ClusterTriggerAuthentication",
                "metadata": {
                  "name": "example-clustertriggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "example-secret-parameter"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "KedaController",
                "metadata": {
                  "name": "keda",
                  "namespace": "openshift-keda"
                },
                "spec": {
                  "admissionWebhooks": {
                    "logEncoder": "console",
                    "logLevel": "info"
                  },
                  "metricsServer": {
                    "logLevel": "0"
                  },
                  "operator": {
                    "logEncoder": "console",
                    "logLevel": "info"
                  },
                  "serviceAccount": null,
                  "watchNamespace": ""
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledJob",
                "metadata": {
                  "name": "scaledjob-sample"
                },
                "spec": {
                  "failedJobsHistoryLimit": 5,
                  "jobTargetRef": {
                    "activeDeadlineSeconds": 600,
                    "backoffLimit": 6,
                    "completions": 1,
                    "parallelism": 1,
                    "template": null
                  },
                  "maxReplicaCount": 100,
                  "pollingInterval": 30,
                  "successfulJobsHistoryLimit": 5,
                  "triggers": [
                    {
                      "metadata": {
                        "property": "example-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledObject",
                "metadata": {
                  "name": "example-scaledobject"
                },
                "spec": {
                  "cooldownPeriod": 300,
                  "maxReplicaCount": 100,
                  "minReplicaCount": 0,
                  "pollingInterval": 30,
                  "scaleTargetRef": {
                    "name": "example-deployment"
                  },
                  "triggers": [
                    {
                      "metadata": {
                        "property": "example-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "TriggerAuthentication",
                "metadata": {
                  "name": "example-triggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "example-secret-parameter"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Cloud Provider
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9-operator@sha256:9a359ae6612d2f22fe983653418143873cb9939f83dca6f008ec3cb8f3d53990
          createdAt: 20 Aug 2025, 21:31
          description: Custom Metrics Autoscaler Operator, an event-driven autoscaler
            based upon KEDA
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.1 <2.17.2-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-keda
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/custom-metrics-autoscaler-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudEventSource
            name: cloudeventsources.eventing.keda.sh
            version: v1alpha1
          - kind: ClusterCloudEventSource
            name: clustercloudeventsources.eventing.keda.sh
            version: v1alpha1
          - description: Describes cluster wide authentication parameters
            displayName: ClusterTriggerAuthentication
            kind: ClusterTriggerAuthentication
            name: clustertriggerauthentications.keda.sh
            version: v1alpha1
          - description: |
              Represents an installation of a particular version of KEDA Controller.
            displayName: KedaController
            kind: KedaController
            name: kedacontrollers.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your jobs and what the triggers
            displayName: ScaledJob
            kind: ScaledJob
            name: scaledjobs.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your application and what the
              triggers are
            displayName: ScaledObject
            kind: ScaledObject
            name: scaledobjects.keda.sh
            version: v1alpha1
          - description: Describes authentication parameters
            displayName: TriggerAuthentication
            kind: TriggerAuthentication
            name: triggerauthentications.keda.sh
            version: v1alpha1
        description: |
          ## About the managed application
          Custom Metrics Autoscaler for OpenShift is an event driven autoscaler based upon KEDA.  Custom Metrics Autoscaler can monitor event sources like Kafka, RabbitMQ, or cloud event sources and feed the metrics from those sources into the Kubernetes horizontal pod autoscaler.  With Custom Metrics Autoscaler, you can have event driven and serverless scale of deployments within any Kubernetes cluster.
          ## About this Operator
          The Custom Metrics Autoscaler Operator deploys and manages installation of KEDA Controller in the cluster. Install this operator and follow installation instructions on how to install Custom Metrics Autoscaler in your cluster.

          ## Prerequisites for enabling this Operator
          ## How to install Custom Metrics Autoscaler in the cluster
          The installation of Custom Metrics Autoscaler is triggered by the creation of `KedaController` resource. Please refer to the [KedaController Spec](https://github.com/openshift/custom-metrics-autoscaler-operator/blob/main/README.md#the-kedacontroller-custom-resource) for more details on available options.

          Only a resource named `keda` in the namespace where the operator is installed will trigger the installation, reconfiguration or removal of the KEDA Controller resource.

          There should be only one KEDA Controller in the cluster.

          The operator will behave in this manner whether it is installed with the `AllNamespaces` or `OwnNamespace` install mode. While the operator more closely matches the `OwnNamespace` semantics, `AllNamespaces` is a supported installation mode to allow it to be installed to namespaces with existing `OperatorGroups` which require that installation mode, such as `openshift-operators`.
        displayName: Custom Metrics Autoscaler
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keda
        - scaling
        - event
        links:
        - name: Custom Metrics Autoscaler Documentation
          url: https://docs.openshift.com/container-platform/latest/nodes/cma/nodes-cma-autoscaling-custom.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9-operator@sha256:9a359ae6612d2f22fe983653418143873cb9939f83dca6f008ec3cb8f3d53990
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9@sha256:ace9358fbd2d81a6e16bfa049c763ca16acf6cc2a59cdc0b6282e5e0dfc902cc
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-adapter-rhel9@sha256:0c8ae9323f927346e168ba829f6c7679af6b2dd359c8d0a020dd0ae3954481ce
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-admission-webhooks-rhel9@sha256:75631a21de057844e9ae45ae2496a22e16f3a32a6fd5f6cb5f90e173e3be2180
        version: 2.17.2-1
      entries:
      - name: custom-metrics-autoscaler.v2.17.2-1
        version: 2.17.2-1
      - name: custom-metrics-autoscaler.v2.15.1-6
        version: 2.15.1-6
      - name: custom-metrics-autoscaler.v2.15.1-4
        version: 2.15.1-4
      - name: custom-metrics-autoscaler.v2.14.1-467
        version: 2.14.1-467
      - name: custom-metrics-autoscaler.v2.14.1-454
        version: 2.14.1-454
      - name: custom-metrics-autoscaler.v2.13.1-427
        version: 2.13.1-427
      - name: custom-metrics-autoscaler.v2.12.1-394
        version: 2.12.1-394
      - name: custom-metrics-autoscaler.v2.11.2-322
        version: 2.11.2-322
      - name: custom-metrics-autoscaler.v2.11.2-311
        version: 2.11.2-311
      - name: custom-metrics-autoscaler.v2.10.1-267
        version: 2.10.1-267
      name: stable
    defaultChannel: stable
    packageName: openshift-custom-metrics-autoscaler-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com
    name: aws-load-balancer-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-load-balancer-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
          createdAt: 2022-06-28T17:56:41
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:d7bc364512178c36671d8a4b5a76cf7cb10f8e56997106187b0fe1f032670ece
        version: 0.0.1
      entries:
      - name: aws-load-balancer-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: aws-load-balancer-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
          createdAt: 2022-12-09T15:32:32
          olm.skipRange: <0.2.0
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:cfd210a2cb705156883fda126d052e05eb84345cc27ba04688067f8ed9a802ea
        version: 0.2.0
      entries:
      - name: aws-load-balancer-operator.v0.2.0
        version: 0.2.0
      name: stable-v0
    - currentCSV: aws-load-balancer-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
          createdAt: 2022-06-28T17:56:41
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:d7bc364512178c36671d8a4b5a76cf7cb10f8e56997106187b0fe1f032670ece
        version: 0.0.1
      entries:
      - name: aws-load-balancer-operator.v0.0.1
        version: 0.0.1
      name: stable-v0.1
    - currentCSV: aws-load-balancer-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
          createdAt: 2022-12-09T15:32:32
          olm.skipRange: <0.2.0
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:cfd210a2cb705156883fda126d052e05eb84345cc27ba04688067f8ed9a802ea
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        version: 0.2.0
      entries:
      - name: aws-load-balancer-operator.v0.2.0
        version: 0.2.0
      name: stable-v0.2
    - currentCSV: aws-load-balancer-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
          createdAt: 2024-01-23T11:21:00
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:2e0b9332a44d8d9c23e19c7accab0813a651f39210257820db508cac28876595
        version: 1.1.1
      entries:
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1
    - currentCSV: aws-load-balancer-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:6e15df39a3d25755a79888f739e1af8b052c3fba57bf8f6b3434ef58fd4288cf
          createdAt: 2024-01-24T09:13:21
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat <<EOF | oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
          EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note about UPI:__ For UPI based installs, additional documentation can be found in
          [VPC and subnets](https://github.com/openshift/aws-load-balancer-operator/blob/release-1.0/docs/prerequisites.md#vpc-and-subnets)

          __Note about STS:__ For clusters using STS, use the [STS documentation](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#for-sts-clusters) instead.
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9f9f8f69ab98c1f25cc301572339e0cc5c5b76f2e29ed26e10ac91ad20e0090c
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:6e15df39a3d25755a79888f739e1af8b052c3fba57bf8f6b3434ef58fd4288cf
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:73db97d20ee9415e9a8941cc6a36ed81d901d760d9949e95e10c90866ef8bf64
        version: 1.0.1
      entries:
      - name: aws-load-balancer-operator.v1.0.1
        version: 1.0.1
      - name: aws-load-balancer-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: aws-load-balancer-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
          createdAt: 2024-01-23T11:21:00
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:2e0b9332a44d8d9c23e19c7accab0813a651f39210257820db508cac28876595
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
        version: 1.1.1
      entries:
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    defaultChannel: stable-v1
    packageName: aws-load-balancer-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres-community
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.17.0-rc1
        version: 1.17.0-rc1
      - name: stackgres-community.v1.16.3
        version: 1.16.3
      - name: stackgres-community.v1.16.2
        version: 1.16.2
      - name: stackgres-community.v1.16.1
        version: 1.16.1
      - name: stackgres-community.v1.16.0
        version: 1.16.0
      - name: stackgres-community.v1.16.0-rc3
        version: 1.16.0-rc3
      - name: stackgres-community.v1.16.0-rc2
        version: 1.16.0-rc2
      - name: stackgres-community.v1.16.0-rc1
        version: 1.16.0-rc1
      - name: stackgres-community.v1.15.2
        version: 1.15.2
      - name: stackgres-community.v1.15.1
        version: 1.15.1
      - name: stackgres-community.v1.15.0
        version: 1.15.0
      - name: stackgres-community.v1.15.0-rc4
        version: 1.15.0-rc4
      - name: stackgres-community.v1.15.0-rc3
        version: 1.15.0-rc3
      - name: stackgres-community.v1.15.0-rc2
        version: 1.15.0-rc2
      - name: stackgres-community.v1.15.0-rc1
        version: 1.15.0-rc1
      - name: stackgres-community.v1.14.3
        version: 1.14.3
      - name: stackgres-community.v1.14.2
        version: 1.14.2
      - name: stackgres-community.v1.14.1
        version: 1.14.1
      - name: stackgres-community.v1.14.0
        version: 1.14.0
      - name: stackgres-community.v1.14.0-rc2
        version: 1.14.0-rc2
      - name: stackgres-community.v1.14.0-rc1
        version: 1.14.0-rc1
      - name: stackgres-community.v1.13.0
        version: 1.13.0
      - name: stackgres-community.v1.13.0-rc1
        version: 1.13.0-rc1
      - name: stackgres-community.v1.12.0
        version: 1.12.0
      - name: stackgres-community.v1.12.0-rc1
        version: 1.12.0-rc1
      name: candidate
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.17.0-rc1
        version: 1.17.0-rc1
      - name: stackgres-community.v1.16.3
        version: 1.16.3
      - name: stackgres-community.v1.16.2
        version: 1.16.2
      - name: stackgres-community.v1.16.1
        version: 1.16.1
      - name: stackgres-community.v1.16.0
        version: 1.16.0
      - name: stackgres-community.v1.16.0-rc3
        version: 1.16.0-rc3
      - name: stackgres-community.v1.16.0-rc2
        version: 1.16.0-rc2
      - name: stackgres-community.v1.16.0-rc1
        version: 1.16.0-rc1
      - name: stackgres-community.v1.15.2
        version: 1.15.2
      - name: stackgres-community.v1.15.1
        version: 1.15.1
      - name: stackgres-community.v1.15.0
        version: 1.15.0
      - name: stackgres-community.v1.15.0-rc4
        version: 1.15.0-rc4
      - name: stackgres-community.v1.15.0-rc3
        version: 1.15.0-rc3
      - name: stackgres-community.v1.15.0-rc2
        version: 1.15.0-rc2
      - name: stackgres-community.v1.15.0-rc1
        version: 1.15.0-rc1
      - name: stackgres-community.v1.14.3
        version: 1.14.3
      - name: stackgres-community.v1.14.2
        version: 1.14.2
      - name: stackgres-community.v1.14.1
        version: 1.14.1
      - name: stackgres-community.v1.14.0
        version: 1.14.0
      - name: stackgres-community.v1.14.0-rc2
        version: 1.14.0-rc2
      - name: stackgres-community.v1.14.0-rc1
        version: 1.14.0-rc1
      - name: stackgres-community.v1.13.0
        version: 1.13.0
      - name: stackgres-community.v1.13.0-rc1
        version: 1.13.0-rc1
      - name: stackgres-community.v1.12.0
        version: 1.12.0
      - name: stackgres-community.v1.12.0-rc1
        version: 1.12.0-rc1
      name: fast
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.16.3
        version: 1.16.3
      - name: stackgres-community.v1.16.2
        version: 1.16.2
      - name: stackgres-community.v1.16.1
        version: 1.16.1
      - name: stackgres-community.v1.16.0
        version: 1.16.0
      - name: stackgres-community.v1.15.2
        version: 1.15.2
      - name: stackgres-community.v1.15.1
        version: 1.15.1
      - name: stackgres-community.v1.15.0
        version: 1.15.0
      - name: stackgres-community.v1.14.3
        version: 1.14.3
      - name: stackgres-community.v1.14.2
        version: 1.14.2
      - name: stackgres-community.v1.14.1
        version: 1.14.1
      - name: stackgres-community.v1.14.0
        version: 1.14.0
      - name: stackgres-community.v1.13.0
        version: 1.13.0
      - name: stackgres-community.v1.12.0
        version: 1.12.0
      name: stable
    defaultChannel: stable
    packageName: stackgres-community
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Rebaca Technologies Pvt Ltd
      provider-url: https://www.rebaca.com/telecom/solutions
    name: abot-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: abot-operator-rhmp.3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "abot.com/v3",
                "kind": "Abot",
                "metadata": {
                  "name": "abot"
                },
                "spec": {
                  "abotanalytics": {
                    "image": "registry.connect.redhat.com/rebaca/abot-analytics@sha256:c854d9eafbe26f834e10f5a97a587b2a2bf05558554897817bb8e2556124b75d",
                    "resources": {
                      "limits": {
                        "cpu": "4000m",
                        "ephemeral-storage": "8Gi",
                        "memory": "6000Mi"
                      },
                      "requests": {
                        "cpu": "2000m",
                        "ephemeral-storage": "4Gi",
                        "memory": "4000Mi"
                      }
                    }
                  },
                  "abotanalyticsbase": {
                    "image": "registry.connect.redhat.com/rebaca/abot-analyticsbase@sha256:dcc2a417a889acd6ec63538a6dfa393d4c1942dac178a04e5d65189a1ae054dc",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotbaseimage": "registry.connect.redhat.com/rebaca/abot-baseimage@sha256:2cbe38298d26be5896e0dc071aae111dd921d0dbd05c3525290f6ff31c73e926",
                  "abotbdd": {
                    "image": "registry.connect.redhat.com/rebaca/abot-bdd@sha256:d5c6bfe12a548cbdb4e0d83d5f18e5ab0c17c3af5340e437b2d0c23740c9c9e6",
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "ephemeral-storage": "8Gi",
                        "memory": "2000Mi"
                      },
                      "requests": {
                        "cpu": "1000m",
                        "ephemeral-storage": "4Gi",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotgrafana": {
                    "image": "registry.redhat.io/rhel8/grafana@sha256:d894cd34f2d8cac30e0fab69bd869f4b0f258b5086a8fd7ead0c5886ce2c5f7a",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "1500Mi"
                      },
                      "requests": {
                        "cpu": "300m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotgui": {
                    "image": "registry.connect.redhat.com/rebaca/abot-gui@sha256:78b512d6bd223a1f5d35f4cbd221f6637e2c9267726ce8899ee5d86aaac82b6f",
                    "resources": {
                      "limits": {
                        "cpu": "4000m",
                        "ephemeral-storage": "8Gi",
                        "memory": "6000Mi"
                      },
                      "requests": {
                        "cpu": "2000m",
                        "ephemeral-storage": "4Gi",
                        "memory": "4000Mi"
                      }
                    }
                  },
                  "abotkibana": {
                    "image": "registry.connect.redhat.com/elastic/kibana@sha256:2be3542d50f51ad786ead958a399f26852339e9417279688aad194c6794336ce",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "abotlogstash": {
                    "image": "registry.connect.redhat.com/elastic/logstash@sha256:dbc25dc5c31dfee08e86221af9f549b50075b2ccf7f255779c5f022c54cd892e",
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "2000Mi"
                      }
                    }
                  },
                  "abotmongo": {
                    "image": "registry.connect.redhat.com/rebaca/abot-mongodb@sha256:be9059e7e8bc1636a84575416b823c20463228785598662ee81c27e4c3c987dc",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "abotnginx": {
                    "image": "registry.connect.redhat.com/rebaca/abot-nginx@sha256:3edc22b40cd30230768115efdceec3d054d71f78214b4100b7425d8484154378",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "abotrabbitmq": {
                    "image": "registry.connect.redhat.com/rebaca/abot-rabbitmq@sha256:e609bec067fe7740590bb272db27d1a7770db47a0fc802e2288c022e101c9e25",
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "abotsidecar": {
                    "alertmanagerimage": "registry.redhat.io/rhacm2/prometheus-alertmanager-rhel8@sha256:ecaee8ad24a9c7cb3b945ee6d2c1275b9de2a676e1868abc2687e687385b3f6a",
                    "kubestatemetricsimage": "registry.redhat.io/rhacm2/kube-state-metrics-rhel8@sha256:e6bb6fa0c155bebf64749368dd951c15f1dbfa3559e2ba6734d3661d3a95c00c",
                    "prometheusimage": "registry.connect.redhat.com/kasten/prometheus@sha256:1e490fcccdd50c1026150ba4cad7c29297d2b73b3d5c80240ff31509e666363d",
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "8000Mi"
                      },
                      "requests": {
                        "cpu": "1000m",
                        "memory": "6000Mi"
                      }
                    }
                  },
                  "elasticsearch": {
                    "image": "registry.connect.redhat.com/elastic/elasticsearch@sha256:f47fd7df80ab05635e31d733f6969d38205f9f81cf95da8594a047f3d07c59ae",
                    "replicas": 3,
                    "resources": {
                      "limits": {
                        "cpu": "1200m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "800m",
                        "memory": "1000Mi"
                      }
                    },
                    "volumeClaim": {
                      "storage": "50Gi"
                    }
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "initContainers": {
                    "image": "registry.access.redhat.com/ubi8/ubi@sha256:a965f33ee4ee57dc8e40a1f9350ddf28ed0727b6cf80db46cdad0486a7580f9d",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "protocol": "https",
                  "pvc": {
                    "abotvol": "100Gi",
                    "analyticsvol": "100Gi",
                    "databak": "10Gi",
                    "mongovol": "10Gi",
                    "prometheusvol": "10Gi"
                  },
                  "replicas": 1,
                  "security": {
                    "fsGroup": 1018450000,
                    "seLinuxOptions": {
                      "level": "s0:c123,c456"
                    }
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "name": "abot"
                  },
                  "storageclass": "ocs-storagecluster-cephfs"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/rebaca/abot-operator@sha256:d3e8e1b6988827b3c636bdfcb57cda26870089d90cf2a732be522f5474c0a734
          createdAt: "2024-10-03T20:32:24Z"
          description: The ABot Test Bench is a test automation tool for 4G, 5G and
            ORAN Telecom networks that follow specifications and use standards driven
            design and compliance as specified by the 3GPP consortium.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/abot-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/abot-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=1.0.0 <3.0.0'
          operatorframework.io/suggested-namespace: abot
          operators.openshift.io/valid-subscription: '["ABot Basic License for 15
            days"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: https://abot.rebaca.com/support/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Abot
            name: abots.abot.com
            version: v3
        description: A Helm based operator for deploying Abot application in Openshift
          cluster. The ABot Test Bench is a test automation tool for 4G, 5G and ORAN
          Telecom networks that follow specifications and use standards driven design
          and compliance as specified by the 3GPP consortium. It provides a platform
          for a user to quickly check for protocol compliance and test for load characteristics
          of a particular system under test.
        displayName: Abot Operator-v3.0.0
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 5G
        - 4G
        - ORAN
        - test tool
        - compliance
        - interop
        links:
        - name: Abot Operator
          url: https://abot-operator.domain
        maintainers:
        - email: suchismita.nag@rebaca.com
          name: Suchismita Nag
        - email: samir@rebaca.com
          name: Samir Chatterjee
        maturity: alpha
        provider:
          name: Rebaca Technologies Pvt Ltd
          url: https://www.rebaca.com/telecom/solutions
        relatedImages:
        - registry.connect.redhat.com/rebaca/abot-rabbitmq@sha256:e609bec067fe7740590bb272db27d1a7770db47a0fc802e2288c022e101c9e25
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel8@sha256:e6bb6fa0c155bebf64749368dd951c15f1dbfa3559e2ba6734d3661d3a95c00c
        - registry.connect.redhat.com/rebaca/abot-bdd@sha256:d5c6bfe12a548cbdb4e0d83d5f18e5ab0c17c3af5340e437b2d0c23740c9c9e6
        - registry.connect.redhat.com/rebaca/abot-baseimage@sha256:2cbe38298d26be5896e0dc071aae111dd921d0dbd05c3525290f6ff31c73e926
        - registry.connect.redhat.com/elastic/elasticsearch@sha256:f47fd7df80ab05635e31d733f6969d38205f9f81cf95da8594a047f3d07c59ae
        - registry.connect.redhat.com/rebaca/abot-analyticsbase@sha256:dcc2a417a889acd6ec63538a6dfa393d4c1942dac178a04e5d65189a1ae054dc
        - registry.connect.redhat.com/rebaca/abot-mongodb@sha256:be9059e7e8bc1636a84575416b823c20463228785598662ee81c27e4c3c987dc
        - registry.connect.redhat.com/rebaca/abot-nginx@sha256:3edc22b40cd30230768115efdceec3d054d71f78214b4100b7425d8484154378
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e34f912158985525f6efb3c4b675f20854fc677c867d12fe3cd2e971309428dc
        - registry.connect.redhat.com/rebaca/abot-operator@sha256:d3e8e1b6988827b3c636bdfcb57cda26870089d90cf2a732be522f5474c0a734
        - registry.redhat.io/rhel8/grafana@sha256:d894cd34f2d8cac30e0fab69bd869f4b0f258b5086a8fd7ead0c5886ce2c5f7a
        - registry.connect.redhat.com/elastic/logstash@sha256:dbc25dc5c31dfee08e86221af9f549b50075b2ccf7f255779c5f022c54cd892e
        - registry.connect.redhat.com/kasten/prometheus@sha256:1e490fcccdd50c1026150ba4cad7c29297d2b73b3d5c80240ff31509e666363d
        - registry.connect.redhat.com/rebaca/abot-analytics@sha256:c854d9eafbe26f834e10f5a97a587b2a2bf05558554897817bb8e2556124b75d
        - registry.connect.redhat.com/rebaca/abot-gui@sha256:78b512d6bd223a1f5d35f4cbd221f6637e2c9267726ce8899ee5d86aaac82b6f
        - registry.connect.redhat.com/elastic/kibana@sha256:2be3542d50f51ad786ead958a399f26852339e9417279688aad194c6794336ce
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel8@sha256:ecaee8ad24a9c7cb3b945ee6d2c1275b9de2a676e1868abc2687e687385b3f6a
        - registry.access.redhat.com/ubi8/ubi@sha256:a965f33ee4ee57dc8e40a1f9350ddf28ed0727b6cf80db46cdad0486a7580f9d
        version: 3.0.0
      entries:
      - name: abot-operator-rhmp.3.0.0
        version: 3.0.0
      - name: abot-operator.2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: abot-operator-rhmp
    provider:
      name: Rebaca Technologies Pvt Ltd
      url: https://www.rebaca.com/telecom/solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: secrets-store-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secrets-store-csi-driver-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9-operator@sha256:96f3a8df9ac5ff3828b0add052af00f7d9df0ebb622a6adfd9772658d55baeb8
          createdAt: "2023-06-12T00:00:00Z"
          description: Install and configure Secrets Store CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.13.0-0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/secrets-store-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SecretProviderClass is the Schema for the secretproviderclasses
              API
            displayName: Secret Provider Class
            kind: SecretProviderClass
            name: secretproviderclasses.secrets-store.csi.x-k8s.io
            version: v1
          - description: SecretProviderClassPodStatus is the Schema for the secretproviderclassespodstatus
              API
            displayName: Secret Provider Class Pod Status
            kind: SecretProviderClassPodStatus
            name: secretproviderclasspodstatuses.secrets-store.csi.x-k8s.io
            version: v1
        description: |
          Operator that installs and configures the CSI driver for Secrets Store. The Secrets Store CSI Driver Operator is a Technology Preview feature. For more information on Red Hat Technology Preview features support scope, see https://access.redhat.com/support/offerings/techpreview/.
        displayName: Secrets Store CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: Documentation
          url: https://github.com/openshift/secrets-store-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/secrets-store-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: preview
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0627c09fe3f5a015f117fb30af1f4d17340aefa48b23994737d3901a804b63d3
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:a5704d26d0b562dba9633d1c107f34e05c750c0e3a64dbcace4dab7b964fff8d
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9-operator@sha256:96f3a8df9ac5ff3828b0add052af00f7d9df0ebb622a6adfd9772658d55baeb8
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9@sha256:148672ccd882a43c5c1d52763da465558b5ed54310418ec40c7ca0b56e408b1a
        version: 4.16.0-202509240123
      entries:
      - name: secrets-store-csi-driver-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: secrets-store-csi-driver-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: secrets-store-csi-driver-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: secrets-store-csi-driver-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: secrets-store-csi-driver-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: secrets-store-csi-driver-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: secrets-store-csi-driver-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: secrets-store-csi-driver-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: secrets-store-csi-driver-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: secrets-store-csi-driver-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: secrets-store-csi-driver-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: secrets-store-csi-driver-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: secrets-store-csi-driver-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: secrets-store-csi-driver-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: secrets-store-csi-driver-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: secrets-store-csi-driver-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: secrets-store-csi-driver-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: secrets-store-csi-driver-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: secrets-store-csi-driver-operator.v4.16.0-202410031007
        version: 4.16.0-202410031007
      - name: secrets-store-csi-driver-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: secrets-store-csi-driver-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: secrets-store-csi-driver-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: secrets-store-csi-driver-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: secrets-store-csi-driver-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: secrets-store-csi-driver-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: secrets-store-csi-driver-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: secrets-store-csi-driver-operator.v4.16.0-202408021139
        version: 4.16.0-202408021139
      - name: secrets-store-csi-driver-operator.v4.16.0-202407190036
        version: 4.16.0-202407190036
      - name: secrets-store-csi-driver-operator.v4.16.0-202407111006
        version: 4.16.0-202407111006
      - name: secrets-store-csi-driver-operator.v4.16.0-202407081937
        version: 4.16.0-202407081937
      - name: secrets-store-csi-driver-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: secrets-store-csi-driver-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: secrets-store-csi-driver-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: secrets-store-csi-driver-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: preview
    defaultChannel: preview
    packageName: secrets-store-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/limitador-operator
    name: limitador-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: limitador-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "name": "toy_get_route",
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/limitador-rhel9-operator@sha256:edb28f112a0d3536f2f52467f38df685d953f46f1999cbd61f4fd4fa4805cd34
          createdAt: "2025-04-07T15:52:04Z"
          description: Enables rate limiting for Gateways and applications in a Gateway
            API network
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Limitador is the Schema for the limitadors API
            displayName: Limitador
            kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: Enables rate limiting for Gateways and applications in a Gateway
          API network. - Part of Red Hat Connectivity Link
        displayName: Red Hat - Limitador Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        - kuadrant
        links:
        - name: GitHub
          url: https://github.com/kuadrant/limitador-operator
        - name: Documentation
          url: https://github.com/Kuadrant/limitador/tree/main/docs
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - registry.redhat.io/rhcl-1/limitador-rhel9-operator@sha256:edb28f112a0d3536f2f52467f38df685d953f46f1999cbd61f4fd4fa4805cd34
        - registry.redhat.io/rhcl-1/limitador-rhel9@sha256:98f85a05da53fbf7f20845d1969441620407cc13a185fd52714308679f116425
        version: 1.1.0
      entries:
      - name: limitador-operator.v1.1.0
        version: 1.1.0
      - name: limitador-operator.v1.0.2
        version: 1.0.2
      - name: limitador-operator.v1.0.1
        version: 1.0.1
      - name: limitador-operator.v0.12.1
        version: 0.12.1
      name: stable
    defaultChannel: stable
    packageName: limitador-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/limitador-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: run-once-duration-override-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runoncedurationoverrideoperator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "RunOnceDurationOverride",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "runOnceDurationOverride": {
                    "spec": {
                      "activeDeadlineSeconds": 3600
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certifiedLevel: "false"
          containerImage: registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9-operator@sha256:54993506a3a2fc1e88c9b5aac97ae90d2a04a06ae6e3a97bf487b5e8996b95fb
          createdAt: 2024/07/01
          description: An operator to manage the OpenShift RunOnceDurationOverride
            Mutating Admission Webhook Server
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.1.2'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/run-once-duration-override-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RunOnceDurationOverride is the schema for the runoncedurationoverrides
              API
            displayName: Run Once Duration Override
            kind: RunOnceDurationOverride
            name: runoncedurationoverrides.operator.openshift.io
            version: v1
        displayName: Run Once Duration Override Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mutating
        - webhook
        - workload
        - run-once
        links:
        - name: Source Code
          url: https://github.com/openshift/run-once-duration-override
        - name: Source Code
          url: https://github.com/openshift/run-once-duration-override-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9-operator@sha256:54993506a3a2fc1e88c9b5aac97ae90d2a04a06ae6e3a97bf487b5e8996b95fb
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9@sha256:3dc6e2a5900472e3e083abcb659a13fea3d08f7a9ff167df524b6f890a16dc8e
        version: 1.1.2
      entries:
      - name: runoncedurationoverrideoperator.v1.1.2
        version: 1.1.2
      - name: runoncedurationoverrideoperator.v1.1.1
        version: 1.1.1
      - name: runoncedurationoverrideoperator.v1.1.0
        version: 1.1.0
      - name: runoncedurationoverrideoperator.v1.0.2
        version: 1.0.2
      - name: runoncedurationoverrideoperator.v1.0.1
        version: 1.0.1
      - name: runoncedurationoverrideoperator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: run-once-duration-override-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicast-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicast-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:ccd64090b85536ddaf1b81ca07d40279c9f277b1c5911628f11d678b98dc3bad
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:ccd64090b85536ddaf1b81ca07d40279c9f277b1c5911628f11d678b98dc3bad
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:f0e9f64c353fdbee3bf7ec0eaf070c3b75330ef659a881ec4ebc874ba584d085
        version: 0.5.3
      entries:
      - name: apicast-operator.v0.5.3
        version: 0.5.3
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      name: threescale-2.11
    - currentCSV: apicast-operator.v0.6.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:8c7fe66b8d4044ad965e7b3d149e9d8e3846ac4e55465873f7095a25b297b3df
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:8c7fe66b8d4044ad965e7b3d149e9d8e3846ac4e55465873f7095a25b297b3df
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:12f7a8e69a1369a51bbcd2b34d4ec8c66dd1fc9fb515349a04e7b202953423e3
        version: 0.6.4
      entries:
      - name: apicast-operator.v0.6.4
        version: 0.6.4
      - name: apicast-operator.v0.6.3
        version: 0.6.3
      - name: apicast-operator.v0.6.2
        version: 0.6.2
      - name: apicast-operator.v0.6.1
        version: 0.6.1
      - name: apicast-operator.v0.6.0
        version: 0.6.0
      - name: apicast-operator.v0.5.3
        version: 0.5.3
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      name: threescale-2.12
    - currentCSV: apicast-operator.v0.7.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:bde44c0fb9f13f212184b24e0bef45db734a8439a92179a2089e5aebd6b87b63
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          olm.skipRange: '>=0.6.2 <0.7.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.13/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:bde44c0fb9f13f212184b24e0bef45db734a8439a92179a2089e5aebd6b87b63
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:88dc7a289c803248efbe151999707325b1bc0aac2d26827ea1707e46e5c71652
        version: 0.7.5
      entries:
      - name: apicast-operator.v0.7.5
        version: 0.7.5
      name: threescale-2.13
    - currentCSV: apicast-operator.v0.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:c5e49be7b43a0058f974a5496976255ab3fbd0677bb7bcf9c3438341cf55965a
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.2 <0.8.3'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.14/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:c5e49be7b43a0058f974a5496976255ab3fbd0677bb7bcf9c3438341cf55965a
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:f8cd25658fa6a154e7ea6e5adf0f9dae3fc596ec9989021de6a021ec94eef226
        version: 0.8.3
      entries:
      - name: apicast-operator.v0.8.3
        version: 0.8.3
      - name: apicast-operator.v0.8.2
        version: 0.8.2
      - name: apicast-operator.v0.8.1
        version: 0.8.1
      name: threescale-2.14
    - currentCSV: apicast-operator.v0.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel9-operator@sha256:c4d6acae4e4cc3236c5b8d8dcfb339b072688ab114785d8598bbe68b1f3571a9
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=0.8.0 <0.12.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.14/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel9-operator@sha256:c4d6acae4e4cc3236c5b8d8dcfb339b072688ab114785d8598bbe68b1f3571a9
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        version: 0.12.5
      entries:
      - name: apicast-operator.v0.12.5
        version: 0.12.5
      - name: apicast-operator.v0.12.4
        version: 0.12.4
      - name: apicast-operator.v0.12.3
        version: 0.12.3
      - name: apicast-operator.v0.12.2
        version: 0.12.2
      - name: apicast-operator.v0.12.1
        version: 0.12.1
      - name: apicast-operator.v0.12.0
        version: 0.12.0
      name: threescale-2.15
    - currentCSV: apicast-operator.v0.12.1-mas
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-mas/apicast-rhel9-operator@sha256:e9f22a8af53f881f1948f5d6a033c35a530895633f2eae522b2fb67f9dc6088b
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.17"}]'
          olm.skipRange: '>=0.7.9 <0.12.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-mas/apicast-gateway-rhel8@sha256:f174636a9ad43a2419fff5a9e6797e3185c540a9509188fd1606aef4acea7bdf
        - registry.redhat.io/3scale-mas/apicast-rhel9-operator@sha256:e9f22a8af53f881f1948f5d6a033c35a530895633f2eae522b2fb67f9dc6088b
        version: 0.12.1-mas
      entries:
      - name: apicast-operator.v0.12.1-mas
        version: 0.12.1-mas
      - name: apicast-operator.v0.7.9-mas
        version: 0.7.9-mas
      - name: apicast-operator.v0.7.8-mas
        version: 0.7.8-mas
      - name: apicast-operator.v0.7.7-mas
        version: 0.7.7-mas
      - name: apicast-operator.v0.7.6-mas
        version: 0.7.6-mas
      - name: apicast-operator.v0.7.5-mas
        version: 0.7.5-mas
      - name: apicast-operator.v0.7.4-mas
        version: 0.7.4-mas
      - name: apicast-operator.v0.7.3-mas
        version: 0.7.3-mas
      - name: apicast-operator.v0.7.2-mas
        version: 0.7.2-mas
      - name: apicast-operator.v0.7.1-mas
        version: 0.7.1-mas
      - name: apicast-operator.v0.7.0-mas
        version: 0.7.0-mas
      - name: apicast-operator.v0.6.0-mas
        version: 0.6.0-mas
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      name: threescale-mas
    defaultChannel: threescale-2.15
    packageName: apicast-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: cloud-bulldozer
      provider-url: ""
    name: ripsaw
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ripsaw.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery, Cloud Provider
          containerImage: quay.io/cloud-bulldozer/benchmark-operator:latest
          description: '[Benchmark-operator](https://github.com/cloud-bulldozer/ripsaw/)
            is a benchmarking tool to benchmark your kubernetes cluster. The intent
            of this Operator is to deploy common workloads to establish a performance
            baseline of Kubernetes cluster'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/cloud-bulldozer/benchmark-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Benchmark
            name: benchmarks.ripsaw.cloudbulldozer.io
            version: v1alpha1
        description: '[Benchmark-operator](https://github.com/cloud-bulldozer/benchmark-operator/)
          is a benchmarking tool to benchmark your kubernetes cluster. The intent
          of this Operator is to deploy common workloads to establish a performance
          baseline of Kubernetes cluster'
        displayName: benchmark-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - performance
        - benchmark
        links:
        - name: Site
          url: https://github.com/cloud-bulldozer/benchmark-operator
        - name: Documentation
          url: https://github.com/cloud-bulldozer/benchmark-operator/tree/master/docs
        maintainers:
        - email: rsevilla@redhat.com
          name: rsevilla87
        - email: jtaleric@redhat.com
          name: jtaleric
        - email: rzaleski@redhat.com
          name: dry939
        - email: kwhitley@redhat.com
          name: whitleykeith
        maturity: alpha
        provider:
          name: cloud-bulldozer
        relatedImages:
        - quay.io/cloud-bulldozer/benchmark-operator:latest
        - k8s.gcr.io/redis:v1
        version: 0.1.0
      entries:
      - name: ripsaw.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ripsaw
    provider:
      name: cloud-bulldozer
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: stolostron-engine
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stolostron-engine.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:b7d0c8413a80c251779fe3c1c9200f9fcfbbe085ccee3ab968ebd4a39209946c
          createdAt: "2024-06-28T02:42:26Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/work@sha256:4ef2518aa107ebd28d5fb1a98d63d82a8e0610544600169e0dd5f345599e518b
        - quay.io/stolostron/assisted-installer@sha256:a31ed7b8e4ff2f9e5ade2e61191cc759e1d47ddbba95a2af41bfa53fa6ddf855
        - quay.io/stolostron/cluster-api-provider-agent@sha256:c074ca1f28d03e86efe4f09200a38af4ac947247a19c46c5300eefcd4f337006
        - quay.io/stolostron/placement@sha256:0dbcc7c62569920afcad1443529f10e38d651a5fe5ae14ade89d6e42fb522e60
        - quay.io/stolostron/managed-serviceaccount@sha256:4c3f1e0b9c4f2937ceb46e6b3f63e3a07061f9b29fdda7dcfdf9ebd74c3bb30b
        - quay.io/stolostron/backplane-operator@sha256:b7d0c8413a80c251779fe3c1c9200f9fcfbbe085ccee3ab968ebd4a39209946c
        - quay.io/stolostron/assisted-installer-controller@sha256:213dc6fde247a5b9862f5076d40394164dd596cde7ca2061e58b73f4d2dbc89e
        - quay.io/stolostron/hive@sha256:fd39b8340cfaeb9849e83bd910e8a3575791fcc58dbbdeb3e66528f009b33d65
        - quay.io/stolostron/cluster-image-set-controller@sha256:632661c6cbfce7df0545bc2fccf1ef87257abb3cc92dcede054806613a3540e3
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:f0d6d46d3a4d65fba6b74bdc9eab99e8149133f83a5c8486f22a73ab92f78ed6
        - quay.io/stolostron/assisted-service@sha256:653728f209070c6adad846a4780449f56985b19006024cfe8137a3a1854fdc85
        - quay.io/stolostron/registration@sha256:8dca354dd63739d7578c24d57b2a72c0676c979344e05c6f2634a8a9fae3218d
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:26f8bb3c9b531e96e6538a5bff96569628f45c06eb3b991ebb421d4e90d7c7e3
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/registration-operator@sha256:4b6c31592c405db11375f4694a7f15bb13c5c9a770786a1607ea0be615396435
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/discovery-testserver@sha256:83d7d6cfaef629a4e3d5d0e9fe35456aa220f1eb68f6fe611d85232acf46bc73
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/discovery-operator-tests@sha256:c98c42b0c16306de86d7b365096cb809a0c797b0a4b7dc04a7de877d0ba5733a
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:b76a201edcad4805e32c40d3f0ad3f61584a879aa00c9e49aae3b2985bfb5fbc
        - quay.io/stolostron/addon-manager@sha256:87c829911438061a53aff2b297416bf43080fdd8d93558caefc198e6f9cb382f
        - quay.io/stolostron/backplane-operator-test@sha256:961d8c40aa3526ca368fd6833202800ba2eb344c97788a646d6fb350a5338397
        - quay.io/stolostron/hypershift-operator@sha256:7dd5eaf195e6b1ca4cecd50514f5b7045948184f9ee617341bc90edc5f9914f2
        - quay.io/stolostron/backplane-must-gather@sha256:2676ebb7438ca42f0d852100612715c429fe9fbe072297fc178d51c34378d1af
        - quay.io/stolostron/discovery-operator@sha256:acb2c4033da62055feb49ff3023b2850a2768996e1fa80f52588b3e8e458e8f5
        - quay.io/stolostron/assisted-image-service@sha256:0c29a74e0f82d1782d5170ff674f5bfcd89ecc5d180fe5b583a5e0f388b9504b
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:a39c52097faf7a21cc717768401851466f8c9341f08103e4d1fe8459ae7f5ea2
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/helloworld@sha256:abc0d8002c5fe6d95a1f24fce1b7899ca24fac576c894ea7443cf7e15e6a42c2
        - quay.io/stolostron/image-based-install-operator@sha256:b9af0ddb7ae8e6fd29b953dfb917546dfd676206e2ad143e892172b6e1cea5a1
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:b269d88e30551636e024aa1472ea4dd71b873b5a2778a199b32732488fd776fb
        - quay.io/stolostron/multicloud-manager@sha256:05da692b2780453115ef6c195658cadb2f28eb0f3a86dc0a586e459dd7c55afe
        - quay.io/stolostron/cluster-curator-controller@sha256:e3a91a4b06dd73fc4e401ddf034e61d4d5b2d324d2c5cea92941e1a56c4023c1
        - quay.io/stolostron/hypershift-addon-operator@sha256:47ed6a7c00036599051718d6015bee35818ea5bbaa14f1c01bedb338dce90668
        - quay.io/stolostron/assisted-installer-agent@sha256:a2efc62ca0080779acfea2e413e524608bc0e8f72b2b917e6c7e76864d08b2b6
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/provider-credential-controller@sha256:d3bd513c06591786216f2baab46c9f57e373d1939f038edacecb2180f673d888
        - quay.io/stolostron/cluster-proxy@sha256:6cd53c588bb9c5064ba0eb8eab03cd12c86976b353209102059f7d76ec59d7c3
        - quay.io/stolostron/cluster-proxy-addon@sha256:f70a26b276c67ad791d337e35cd102afa82f1005bc6ecbc60309932d2d163a6b
        - quay.io/stolostron/managedcluster-import-controller@sha256:925bbc6173501bdb60d329a1f5a134ea3e9230034509320f93382f1c570be275
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/console-mce@sha256:4e49f2b2fdb0491c225e4d5dc143a6d7c0a25cd855057b3792c7ee94a868ac85
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/assisted-service-el8@sha256:31422ffce8b94d662131986bd084fcba503809cb939389096f9c9dbb27cb58db
        - quay.io/stolostron/clusterclaims-controller@sha256:a0613147b0ec010b8a4442a3c515f7ffc48ec7d8137436a0e8b935443cac2f96
        version: 0.4.0
      entries:
      - name: stolostron-engine.v0.4.0
        version: 0.4.0
      name: community-0.4
    - currentCSV: stolostron-engine.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:30d5bcccc542988ac788dbca305e6f29fb678470b19e48adada9ca6f5803f04a
          createdAt: "2024-10-02T05:51:59Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.5.0 <0.5.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/assisted-installer@sha256:dcc4e28c7eea55dd2ac22dd75c2b7b96ee3552fcbeba8128ad8528fda235d879
        - quay.io/stolostron/cluster-curator-controller@sha256:fc2ad8e0b6701638ab89bb62a49700fdaadfe1149228cd0b3cd4a41810020c6d
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/work@sha256:4467a4bd00586ab837cb7e8bf56ea1faff82ed5ce2fffdeb7e51ee5fafb433b0
        - quay.io/stolostron/cluster-image-set-controller@sha256:654e218ecf0a3f67520c0521b544eea9d4552e92f54bf424e88ccab24e7c151e
        - quay.io/stolostron/assisted-installer-agent@sha256:16b040dd0690b62bc051aa6369272d004d2157a8e5230d318ba827adba8e8d63
        - quay.io/stolostron/hypershift-addon-operator@sha256:ba199ebfe0eea3ef6b07123f51116831e7b0a73860b0410ece723bcc9a3230af
        - quay.io/stolostron/placement@sha256:dc0ad451b085b3a747ee648475f8e8102b645df69b898fc88f1cb190d1b95a83
        - quay.io/stolostron/cluster-proxy@sha256:be09bb293a990600ec48aba2f9bf8b457d5bdbc6b136be319e5feebadb232a8c
        - quay.io/stolostron/discovery-operator@sha256:db56e2a78ed27a46b47b519e29db66636b9abb499fd507d99451bae47438e32b
        - quay.io/stolostron/hive@sha256:707ac9a2ad18dbc4c4686f1b72946077c7232afd05505d5e50e13f806a452392
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:4b5b2bea2aad6c7e16fe1520a2fb8454b0312f2749d2f398ca54e644e813055c
        - quay.io/stolostron/cluster-api-provider-agent@sha256:1458de75311675a5acec7361e507edb24b970ffa3403e916de3413f7061bfe36
        - quay.io/stolostron/clusterclaims-controller@sha256:1c0b3c4730ef1ab628bf2c1b7cf12b1b93b149abdefc4d908dbdc90e4b6c1c19
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:2fe53994fa5bde93112a2cc75758e0f68be013a34c8206066d0b315c2e6a6f47
        - quay.io/stolostron/assisted-service-el8@sha256:c8248f9a7ac02f47373d097953cee38b66125e98b98922f18c578a591f04bcca
        - quay.io/stolostron/console-mce@sha256:10a54c5012ef6439539297919c71d9b53aaaddc6342d90677113d6ce7b42e3fe
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/backplane-operator-test@sha256:048f21cf89fb5ed2557ffc90ae5d7104cdd5ce0e6a2a865ce6f4230d09400115
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:b76a201edcad4805e32c40d3f0ad3f61584a879aa00c9e49aae3b2985bfb5fbc
        - quay.io/stolostron/addon-manager@sha256:9384ae0e9bcc9f85d3fe36e44ab7b3a61777197846559b5add0c56f513e35333
        - quay.io/stolostron/assisted-image-service@sha256:703c12a918d3c66b62c29a5abd0e8c5032299289cfed459671d6ca82c4f35c73
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/assisted-service@sha256:6383ba508a0cac030a46df96ac37daaeddc27ad1fb41d8b4dca5a87cf4ab9c31
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:14a0648b76301de182154a8d7d6b7e7eca0bfbed226e6dcebddb75af5152451f
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/registration-operator@sha256:b42341143b1e1693748f4b7d69e97c56c8a5f9408abc5db1de96ba246b7b98a9
        - quay.io/stolostron/backplane-operator@sha256:30d5bcccc542988ac788dbca305e6f29fb678470b19e48adada9ca6f5803f04a
        - quay.io/stolostron/discovery-operator-tests@sha256:e8d3d1675242272cc139b0dba15605d00e25b17b4b1cc1b5b4f3ea8913ffb2da
        - quay.io/stolostron/discovery-testserver@sha256:0b23021d6c0c7d6c7aa97a92fd0c40f31fdff6363a5b10a1edc864b5d8d54108
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/multicloud-manager@sha256:6dd212840ea7fadd427b503ce69e6e32464881a0047e2af86f9f2c7562bd4703
        - quay.io/stolostron/managedcluster-import-controller@sha256:01ab0c8b664692d69c2a9b3363e007a5611c332ac0e3491234dd0cbfba31c88a
        - quay.io/stolostron/registration@sha256:4779cf430d479806efd2f3cb7e2a9ed3dce749287609c39ad45f1b06e220d023
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/managed-serviceaccount@sha256:667f823799984404c0c1589df7b857a6a35218c2653254905dc22ca90ce1542b
        - quay.io/stolostron/backplane-must-gather@sha256:92e992e23855375a8bae542480d789d1bb25a48959896bfb75f659df2a37dd98
        - quay.io/stolostron/image-based-install-operator@sha256:bd76503406e6d20693e636461411e158c8dde3d03c4904ecb8a1850deeefa4f5
        - quay.io/stolostron/hypershift-operator@sha256:a30fb6cf4d5e103820813019ce452d8134b0fee8ecd3c2166d5681199d4a0cad
        - quay.io/stolostron/cluster-proxy-addon@sha256:7aa09b0d53fdc0f5cec9c8a13c160277a9274466025f11b5b04da95c9a10e59c
        - quay.io/stolostron/provider-credential-controller@sha256:d240ad43c23de822f91a33849448ae70a4cf27570da3387bdecd101ca2f0b65d
        - quay.io/stolostron/assisted-installer-controller@sha256:01dd7c7fb9cb335215afefde204309d3127847b900658ee9744b9d31df847565
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/helloworld@sha256:96a9b8855fb8800a85aed6fdc4911295dc2c12629bbdcc34ed8dc8061827b17e
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        version: 0.5.3
      entries:
      - name: stolostron-engine.v0.5.3
        version: 0.5.3
      - name: stolostron-engine.v0.5.2
        version: 0.5.2
      - name: stolostron-engine.v0.5.1
        version: 0.5.1
      - name: stolostron-engine.v0.5.0
        version: 0.5.0
      name: community-0.5
    - currentCSV: stolostron-engine.v0.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:fa1fa6ef3958fd9d50bfb38197f033a635a05b47c32e653822246e708d8598cc
          createdAt: "2025-02-24T01:37:53Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.6.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/discovery-operator-tests@sha256:21449ecd291cf7203cc8e27a2a00ac60e74f5993708f9e7465e8efdfdca1c89f
        - quay.io/stolostron/assisted-installer@sha256:c88976ab4b35d458a39044e522d2d5ab1916d8485af07bbef373af1bcbbd6af0
        - quay.io/stolostron/assisted-service-el8@sha256:f50f4dece91100f45ef9e20b1e52557a55db45b17a380f32d365ee013d695d5c
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:b016a001a53bc37ae534b74cc832552759800746c27e16adcf0626ff9ad2e4c0
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:2781d87c1568f2e40409f5431bfd1a9a4c9ca35655537e7d4db65d9b881edd08
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        - quay.io/stolostron/placement@sha256:543f09446652bff8c4149945d94d715eff7a76d82783f2bad89e712798542c73
        - quay.io/stolostron/provider-credential-controller@sha256:a5ac3999d0aefb1dd9a239f9394485f93e4252a45eb4a43a43e603aaff80b3bb
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:44f8bb72b975456d4151a88a560fc525ad6f4da7968a38e94a86b6dd77d3c543
        - quay.io/stolostron/registration@sha256:9c221a40ffa119ac154d76f578e0919b5a63137c908afb0583519a9b8a399b2b
        - quay.io/stolostron/hypershift-operator@sha256:2d1809223ce5e75b4133a39b78a269703c6b5234f7ec2a7c0072124c345fae0a
        - quay.io/stolostron/clusterclaims-controller@sha256:b796cb2edd3d4c22b29fc6d2b45e5328a59b7ec1bdc52d5500e1a4ee74725d6c
        - quay.io/stolostron/cluster-api-provider-agent@sha256:dda3e9aef611c744c4c89c2689a12d0649a173c1a64954c3037de55ec8eacec9
        - quay.io/stolostron/multicloud-manager@sha256:0293f0852206562b891375b167cfe17843e2e17585f04c9bb6482d1fe78089d0
        - quay.io/stolostron/backplane-operator@sha256:fa1fa6ef3958fd9d50bfb38197f033a635a05b47c32e653822246e708d8598cc
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/discovery-testserver@sha256:f757a7a11108a88de4bc0ecc5e16f74b3bd9b5110ec94bfeef2074121ec1e47f
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/cluster-proxy-addon@sha256:7da9da3e4ecd1e0df636c9f72f9f30cc7ff24b81e247ba9825cef39f18e86182
        - quay.io/stolostron/image-based-install-operator@sha256:fecbff73eb97b28bedd7c17d5adbef4826f181a6838ad4cc72e7478acf9b4254
        - quay.io/stolostron/hive@sha256:1e1f60492c6a3b43b69fe8a0640f9b81f5a1cfafc0afcd6854c3cfe362ed063a
        - quay.io/stolostron/cluster-proxy@sha256:b341dd7058cdffccb00a026908031fa123fec5ba58ba9307dbb873b75d250e4d
        - quay.io/stolostron/cluster-image-set-controller@sha256:629361c1c0ec50fe6410fc6fe6417cba89957c49ab89201c6470c0f97c8fc91d
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/assisted-image-service@sha256:6846719f91dd954ce4b4c6e7c9654c7d79db72990270acd3c34bdd2a89af34fc
        - quay.io/stolostron/console-mce@sha256:bfaf9180cd1e5af4d34daf4f1dcdee89951067a605f21bfe6e1ce6bda85b0334
        - quay.io/stolostron/helloworld@sha256:4838f699ea2461f8358dcda32b3ac96ed05b55d4e1958eac2bdacae4807501d1
        - quay.io/stolostron/ose_cluster_api_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/assisted-installer-controller@sha256:3b0b43fdbbbbd5c1e7588fc449b002e1ed9ad0ea6daa774925a962bd499a4195
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/work@sha256:6ec081826f32467ea1e1a91f6624df349a382fb8a1cbbda906141281ac2bad5c
        - quay.io/stolostron/ose_aws_cluster_api_controllers_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/backplane-operator-test@sha256:b04661ad658423c9c7557b722adb2159d988c10445de1c59f39c44583262cb42
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/assisted-installer-agent@sha256:cd74c8b8888c100bf567dd07cfff40097c458d18fcc29a11ed3830f697156267
        - quay.io/stolostron/backplane-must-gather@sha256:99b558d687e1c17036b164d819cb1c9b96a814be79e3c87960ef7cdb521eee7d
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/discovery-operator@sha256:43f40b9c20b112248fcdb42d114fe4af73fdd45b05d9ff7855d827cdd5e971b4
        - quay.io/stolostron/managed-serviceaccount@sha256:0e4edd9ab95c1558fba40a0c41c5cc326996e9e19e296324ea10b5a442a7ef63
        - quay.io/stolostron/hypershift-addon-operator@sha256:65f2b7a356c22413307ce7b7de38ce00c31e4c43b6ee3b4b484c0bc61dc483e9
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:65180432218d035ffc7996a7643ddb5300eb8b09e47e09cb270cb9b961aecf7a
        - quay.io/stolostron/managedcluster-import-controller@sha256:ecfe7c1f2d24a75bfb8db2a0a2765fe5e9136f81f341882de896431570fd10e9
        - quay.io/stolostron/addon-manager@sha256:7e347a177d8c5b80a78411a3e8dbf04e49d157a4ca67e7fd5170970c2ba62da9
        - quay.io/stolostron/registration-operator@sha256:52289e6cf5f159d344717b49e6226ff9d4ee33942f63cc543d1b820d7ee24fc5
        - quay.io/stolostron/assisted-service@sha256:58271e54eda979c43fb474c3bb255f7328e380a7af9ba833e5f7558f323bd239
        - quay.io/stolostron/cluster-curator-controller@sha256:5d05462c2c9d93b21656a436c01fbda4256757bd057e5de7014eb6f320310a1e
        version: 0.6.3
      entries:
      - name: stolostron-engine.v0.6.3
        version: 0.6.3
      - name: stolostron-engine.v0.6.2
        version: 0.6.2
      - name: stolostron-engine.v0.6.1
        version: 0.6.1
      - name: stolostron-engine.v0.6.0
        version: 0.6.0
      name: community-0.6
    - currentCSV: stolostron-engine.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:9b39a5b6a502778d7884e276b276a9b44e6dbc88e7e82470e26d91ded3cb3839
          createdAt: "2025-06-09T15:04:58Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.7.1'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational components for managing multiple Kubernetes-based clusters. The deployment of the multicluster engine components is
              determined based on the configuration that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/cluster-image-set-controller@sha256:7f05ad13ffaa11daec066c5889b87df1b631e6eb3dc1c2913a2766c8e65218aa
        - quay.io/stolostron/cluster-proxy@sha256:1528e36dfa4fa2c9063d84c0f2b9aa8d2b549a750e812d4a2e2b8c7515bfe5cb
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:1e95f55d6918263e79de1c1e2be8a96cd3d35a346dde6d3e2fb484f48e567811
        - quay.io/stolostron/hive@sha256:86ff94d3e7c174ac4d6818183c66791edc7638965dd09e2d41172609c81f451f
        - quay.io/stolostron/cluster-api-provider-openshift-assisted-bootstrap@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/console-mce@sha256:7f538a21eb22ccd57765f47692da23be9cecfd3b7b65986f1413234bdf709530
        - quay.io/stolostron/assisted-installer-agent@sha256:ffceb06ce038013291e092324d3c49091cd1b2ff1569abe77c29484075e28107
        - quay.io/stolostron/hypershift-addon-operator@sha256:ed1c1580fee28310fc1b137c5c70271448e58ed29af4371515fe9c74512a2c7e
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/placement@sha256:6276e411c9f7cf7072d3d364569b6c3cd472742ab4ad48912f18cb4091d27a6f
        - quay.io/stolostron/ose_cluster_api_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/cluster-proxy-addon@sha256:c776405c4b1b623500c69a12b7d76563646b27cf23bab74dc8f1361534cfb81e
        - quay.io/stolostron/addon-manager@sha256:a245847f135f96c4b4fcaa673fc32013c8e407252b3259175c796e3c16111c71
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/discovery-testserver@sha256:faef24208a65283b4157521079a7bf23bc4ebd553d70a83a78c324b5cb0e24e4
        - quay.io/stolostron/discovery-operator@sha256:a6a37237bc06f0635a936b803d3c53f0f0caf5426c76b37951c4e157b181ecb3
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:1d4221cfc364743b58e15bad3afe9548f4151f06b0dbf0aec7c00715ba75abba
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/ip-address-manager@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/assisted-installer@sha256:17b78d73775a93b31ca872db79520ddc49d2be042fdf1599f419fbcbfd9a98ac
        - quay.io/stolostron/assisted-service-el8@sha256:f52c55691850a1f2f254b0e9377ba97493be337e012223e5a2b3190f03e1a349
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:44f8bb72b975456d4151a88a560fc525ad6f4da7968a38e94a86b6dd77d3c543
        - quay.io/stolostron/discovery-operator-tests@sha256:d96fe1f4f49ecfd4a27bedd87d7070a3c7f3f82b6ee8bb59e98a262d8cedcfc0
        - quay.io/stolostron/clusterclaims-controller@sha256:afab145b12a52f81e1f2eaa189731e3a6287b54f2f1f1ad2976ad378fa7dfa3a
        - quay.io/stolostron/ose_aws_cluster_api_controllers_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/backplane-operator@sha256:9b39a5b6a502778d7884e276b276a9b44e6dbc88e7e82470e26d91ded3cb3839
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        - quay.io/stolostron/assisted-image-service@sha256:b6df9c9fc8bbfdc58e2169b9e199e1eed188d3d847511eba01eb4920fabe3992
        - quay.io/stolostron/multicloud-manager@sha256:5a95a5d1dbb17456d7ce62f1a517ae49ca780e25f3ebb74b32346de1df23b4c9
        - quay.io/stolostron/helloworld@sha256:a790795409f6e781f506b6a63f39f7d344f01efa30c3de87b1727dc2010d5bbf
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/backplane-must-gather@sha256:f4f9596eade78333628199f62f8e71e4c2b3957c20da5bad807c12a2a02560db
        - quay.io/stolostron/managedcluster-import-controller@sha256:4d0be693db0576eb2e9e3d409bf1479cc60784f5747ed1b25b6a2443c5e91e36
        - quay.io/stolostron/cluster-api-provider-agent@sha256:2b22890b948ad1d597c424d73d4fe34c7c4d265ad6e54fa0ed87a7dec920ac57
        - quay.io/stolostron/ose-baremetal-cluster-api-controllers-rhel9@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/cluster-curator-controller@sha256:d667bd0892812eb4e1f724679759469db335db4ab801f9d57ea8818d8cd13188
        - quay.io/stolostron/image-based-install-operator@sha256:9d6796007b616de60560f55f14546295c4e48b3768d059098fce9367674a3368
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/registration@sha256:2db2c28af25007c77979cc456d929f5f077cd671cc4e0562c43efce2c42c4cd8
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/provider-credential-controller@sha256:3484cb0c2102f4a522efe31bc4c77910a202b42e1ad082a0a8709cc1a7698649
        - quay.io/stolostron/work@sha256:709c41d1a7a1eaf10b334dd0c0105a261d6db811693d5ae67eb55b9a26daed87
        - quay.io/stolostron/assisted-installer-controller@sha256:8ea9d12fa30825fe4f965a7d506af687e8de709f8d07dd75204747da2a6e09a8
        - quay.io/stolostron/assisted-service@sha256:af3532d487e03a077ce23e4178abbcb41cf45726b646d7660840c036677715e8
        - quay.io/stolostron/hypershift-operator@sha256:c9a8fc943ef35126434cc767f152abd09943a4d2d194a9966495fdc0870b417c
        - quay.io/stolostron/registration-operator@sha256:a9fdf837e149c8659fc99deab6a284975c8eb8785549381fa082303cdb172e44
        - quay.io/stolostron/managed-serviceaccount@sha256:4bcd5c9b2044254c28215fd78bf1e3fcd919578238986831fcea8055b2764696
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:1757b9ad635839087c4bef198d5e81f4abb7aee38bc74690021f9e8887366bcf
        - quay.io/stolostron/backplane-operator-test@sha256:8d6514fc0db5ee79c346ddc31bb094ad87fa375519da67e32398300d70d72f6e
        - quay.io/stolostron/cluster-api-provider-openshift-assisted-control-plane@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        version: 0.7.1
      entries:
      - name: stolostron-engine.v0.7.1
        version: 0.7.1
      - name: stolostron-engine.v0.7.0
        version: 0.7.0
      name: community-0.7
    defaultChannel: community-0.7
    packageName: stolostron-engine
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pulp Community
      provider-url: https://github.com/pulp/pulp-operator
    name: pulp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pulp-operator.v1.0.0-alpha.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache_enabled": true,
                  "content": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "ingress_type": "Route",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/"
                  },
                  "storage_type": "File",
                  "worker": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example-pulp"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true,
                    "redis_storage_class": "standard"
                  },
                  "content": {
                    "replicas": 1
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "pulp",
                  "file_storage_access_mode": "ReadWriteOnce",
                  "file_storage_size": "2Gi",
                  "file_storage_storage_class": "standard",
                  "image_version": "nightly",
                  "image_web_version": "nightly",
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/",
                    "telemetry": false
                  },
                  "web": {
                    "replicas": 1
                  },
                  "worker": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "galaxy-example"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "api": {
                    "replicas": 1
                  },
                  "content": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "galaxy",
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "file_storage_storage_class": "standard",
                  "image": "quay.io/pulp/galaxy-minimal",
                  "image_version": "nightly",
                  "image_web": "quay.io/pulp/galaxy-web",
                  "image_web_version": "nightly",
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/api/galaxy/pulp/",
                    "telemetry": false
                  },
                  "signing_scripts_configmap": "signing-scripts",
                  "signing_secret": "signing-galaxy",
                  "storage_type": "File",
                  "web": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "worker": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "galaxy-example",
                  "deployment_type": "galaxy",
                  "instance_name": "pulp",
                  "postgres_configuration_secret": "galaxy-example-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "galaxy-example"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.0.0-alpha.1
          createdAt: "2021-03-26 16:57:40"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1alpha1
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1alpha1
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1alpha1
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://docs.pulpproject.org/](https://docs.pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          Note that Pulp operator works with three different types of service containers (the operator itself, the main service and the web service):

          |           | Operator | Main | Web |
          | --------- | -------- | ---- | --- |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[pulp-minimal](https://quay.io/repository/pulp/pulp-minimal?tab=tags) | [pulp-web](https://quay.io/repository/pulp/pulp-web?tab=tags) |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[galaxy-minimal](https://quay.io/repository/pulp/galaxy-minimal?tab=tags) | [galaxy-web](https://quay.io/repository/pulp/galaxy-web?tab=tags) |

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://docs.pulpproject.org/pulp_operator/container/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://docs.pulpproject.org/pulp_operator/](https://docs.pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://docs.pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: beta
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - docker.io/library/postgres:13
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/pulp/pulp-operator:v1.0.0-alpha.1
        - quay.io/pulp/pulp-minimal:stable
        - quay.io/pulp/pulp-web:stable
        - docker.io/library/redis:latest
        version: 1.0.0-alpha.1
      entries:
      - name: pulp-operator.v1.0.0-alpha.1
        version: 1.0.0-alpha.1
      name: alpha
    - currentCSV: pulp-operator.v1.0.0-beta.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "Pulp",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true
                  },
                  "content": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/"
                  },
                  "worker": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "Pulp",
                "metadata": {
                  "name": "galaxy-example"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "api": {
                    "replicas": 1
                  },
                  "content": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "galaxy",
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "file_storage_storage_class": "standard",
                  "image": "quay.io/pulp/galaxy-minimal",
                  "image_version": "4.6.3",
                  "image_web": "quay.io/pulp/galaxy-web",
                  "image_web_version": "4.6.3",
                  "ingress_type": "nodeport",
                  "nodeport_port": 30000,
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "analytics": false,
                    "ansible_api_hostname": "http://nodeport.local:30000",
                    "api_root": "/api/galaxy/pulp/",
                    "content_origin": "http://nodeport.local:30000",
                    "galaxy_feature_flags": {
                      "ai_deny_index": true,
                      "execution_environments": true
                    },
                    "token_server": "http://nodeport.local:30000/token/"
                  },
                  "signing_scripts": "signing-scripts",
                  "signing_secret": "signing-galaxy",
                  "web": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "worker": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "galaxy-example",
                  "deployment_type": "galaxy",
                  "postgres_configuration_secret": "galaxy-example-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "galaxy-example"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.0.0-beta.4
          createdAt: "2024-01-18T20:33:37Z"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1beta2
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1beta2
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1beta2
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://docs.pulpproject.org/](https://docs.pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          Note that Pulp operator works with three different types of service containers (the operator itself, the main service and the web service):

          |           | Operator | Main | Web |
          | --------- | -------- | ---- | --- |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[pulp-minimal](https://quay.io/repository/pulp/pulp-minimal?tab=tags) | [pulp-web](https://quay.io/repository/pulp/pulp-web?tab=tags) |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[galaxy-minimal](https://quay.io/repository/pulp/galaxy-minimal?tab=tags) | [galaxy-web](https://quay.io/repository/pulp/galaxy-web?tab=tags) |

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://docs.pulpproject.org/pulp_operator/container/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://docs.pulpproject.org/pulp_operator/](https://docs.pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://docs.pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: beta
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.4
        - quay.io/pulp/pulp-operator:v1.0.0-beta.4
        - quay.io/pulp/pulp-minimal:stable
        - quay.io/pulp/pulp-web:stable
        - docker.io/library/redis:latest
        - docker.io/library/postgres:13
        version: 1.0.0-beta.4
      entries:
      - name: pulp-operator.v1.0.0-beta.4
        version: 1.0.0-beta.4
      - name: pulp-operator.v1.0.0-beta.3
        version: 1.0.0-beta.3
      - name: pulp-operator.v1.0.0-beta.2
        version: 1.0.0-beta.2
      - name: pulp-operator.v1.0.0-beta.1
        version: 1.0.0-beta.1
      - name: pulp-operator.v1.0.0-alpha.9
        version: 1.0.0-alpha.9
      - name: pulp-operator.v1.0.0-alpha.8
        version: 1.0.0-alpha.8
      - name: pulp-operator.v1.0.0-alpha.7
        version: 1.0.0-alpha.7
      - name: pulp-operator.v1.0.0-alpha.6
        version: 1.0.0-alpha.6
      - name: pulp-operator.v1.0.0-alpha.5
        version: 1.0.0-alpha.5
      - name: pulp-operator.v1.0.0-alpha.4
        version: 1.0.0-alpha.4
      - name: pulp-operator.v1.0.0-alpha.3
        version: 1.0.0-alpha.3
      - name: pulp-operator.v1.0.0-alpha.2
        version: 1.0.0-alpha.2
      - name: pulp-operator.v1.0.0-alpha.1
        version: 1.0.0-alpha.1
      name: beta
    defaultChannel: beta
    packageName: pulp-operator
    provider:
      name: Pulp Community
      url: https://github.com/pulp/pulp-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-common-service-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-common-service-operator.v4.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v3",
                "kind": "CommonService",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-common-service-operator",
                    "app.kubernetes.io/managed-by": "ibm-common-service-operator",
                    "app.kubernetes.io/name": "ibm-common-service-operator"
                  },
                  "name": "example-commonservice"
                },
                "spec": {
                  "size": "starterset"
                }
              }
            ]
          capabilities: Seamless Upgrades
          cloudPakThemesVersion: styles4100.css
          containerImage: registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
          createdAt: "2025-03-18T21:27:31Z"
          description: The IBM Cloud Pak foundational services operator is used to
            deploy IBM foundational services.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-common-service-operator
          nss.operator.ibm.com/managed-webhooks: ""
          olm.skipRange: '>=3.3.0 <4.12.0'
          operatorChannel: v4.12
          operatorVersion: 4.12.0
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/ibm-common-service-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CommonService is the Schema for the commonservices API. Documentation
              For additional details regarding install parameters check https://ibm.biz/icpfs39install.
              License By installing this product you accept the license terms https://ibm.biz/icpfs39license
            displayName: CommonService
            kind: CommonService
            name: commonservices.operator.ibm.com
            version: v3
        description: |-
          # Introduction

          The `IBM Cloud Pak foundational services` is a bridge to connect IBM Cloud Paks and Operand Deployment Lifecycle Manager (ODLM) with IBM Cloud Pak foundational services. You can also install the `IBM Cloud Pak foundational services` in stand-alone mode.

          When you install this operator, the operator completes the following tasks:

            - Installs ODLM operator namespace
            - Creates the Common Services `OperandRegistry` and `OperandConfig` in the service namespace

          # Details
          For more information about installing this operator and other Common Services operators, see [Installer documentation](http://ibm.biz/cpcs_opinstall). If you are using this operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For more information about IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).

          For more information about the available IBM Cloud Pak foundational services, see the [IBM Knowledge Center](http://ibm.biz/cpcsdocs).

          ## Supported platforms

          Red Hat OpenShift Container Platform 4.5 or newer installed on one of the following platforms:

           - Linux x86_64
           - Linux on Power (ppc64le)
           - Linux on IBM Z and LinuxONE

          ## Operator versions

           - 4.12.0

          ## Prerequisites

          Before you install this operator, you need to first install the operator prerequisites:

          - For the list of prerequisites for installing the operator, see the IBM Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).

          ### Resources Required

          #### Minimum scheduling capacity

          | Software                   | Memory (MB) | CPU (cores) | Disk (GB) | Nodes  |
          | -------------------------- | ----------- | ----------- | --------- | ------ |
          | ibm common service operator | 200          | 0.2        | 1          | worker |
          | **Total**                  | 200         | 0.2         | 1         |        |

          ## Documentation

          - If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak. For a list of IBM Cloud Paks, see [IBM Cloud Paks that use IBM Cloud Pak foundational services](http://ibm.biz/cpcs_cloudpaks).
          - If you are using the operator in stand-alone mode or with an IBM Containerized Software, see the IBM Cloud Pak foundational services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall).

          ## Installing

          For installation, see the [IBM Cloud Platform Common Services documentation](http://ibm.biz/cpcsdocs).

          ## Configuration

          For configuration, see the [IBM Cloud Platform Common Services documentation](http://ibm.biz/cpcsdocs).

          ### SecurityContextConstraints Requirements

          The IBM Common Service Operator supports running with the OpenShift Container Platform 4.3 default restricted Security Context Constraints (SCCs) and IBM Cloud Pak Security Context Constraints (SCCs).

          For more information about the OpenShift Container Platform Security Context Constraints, see [Managing Security Context Constraints](https://docs.openshift.com/container-platform/4.3/authentication/managing-security-context-constraints.html).

          For more information about the IBM Cloud Pak Security Context Constraints, see [Managing Security Context Constraints](https://ibm.biz/cpkspec-scc).

          ### Limitations

          No
        displayName: IBM Cloud Pak foundational services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Common Service Operator
          url: https://github.com/IBM/ibm-common-service-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90046eb6170e12439ccc5aa343f19a4d4fc99a10d98b383cf1dc0e020ebeed55
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:18f44ab15289625fa2c6b72ece6e06afb51058dcba30b8f38b4fdeedf8c9e51b
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:0fdf1beeb4888f648f6dbd17a29dcc53a715d2c4301bc714a4172b2daf8d13be
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:4209e78f12120b9c9f320496b780c3993a9317825ea17dc7d554452b28c99014
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:9a537eae7e581322e2f0764ed38a797ba7cd260724a285f24d5095058b7b47a0
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-operator-bundle@sha256:086eaeaaaf440a1d4a254984afacd8d04583aafb9ec11ac072feb42a87b986f7
        version: 4.12.0
      entries:
      - name: ibm-common-service-operator.v4.12.0
        version: 4.12.0
      name: v4.12
    defaultChannel: v4.12
    packageName: ibm-common-service-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: rebellions
      provider-url: https://rebellions.ai/
    name: rbln-npu-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: candidate
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: fast
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: rbln-npu-operator
    provider:
      name: rebellions
      url: https://rebellions.ai/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/component: operator
      app.kubernetes.io/managed-by: kustomize
      app.kubernetes.io/name: zabbix-operator-certified
      app.kubernetes.io/part-of: zabbix-operator-certified
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zabbix LLC
      provider-url: ""
    name: zabbix-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zabbix-operator-certified-rhmp.v6.0.40
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent",
                "metadata": {
                  "labels": {
                    "app": "agent",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "log_remote_commands": true,
                  "max_lines_per_second": 20,
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent2",
                "metadata": {
                  "labels": {
                    "app": "agent2",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent2"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "enable_persistent_buffer": false,
                  "enable_status_port": false,
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "persistent_buffer_claim": "",
                  "persistent_buffer_period": "1h",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixFull",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-full"
                },
                "spec": {
                  "builtin_report_service": true,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysql_volumeclaim": "mysql-full-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxyMysql",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-mysql"
                },
                "spec": {
                  "db_server_port": 3306,
                  "enable_snmp_traps": false,
                  "internal_db": true,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix_proxy",
                  "proxy": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "zabbix_mysql_volumeclaim": "mysql-proxy-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxySqlite",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-sqlite"
                },
                "spec": {
                  "db_data_volumeclaim": "zabbix-proxy-sqlite",
                  "enable_snmp_traps": false,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "proxy": {
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "node_name_as_db_name": true,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "proxy_size": 1,
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  }
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixServer",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-server"
                },
                "spec": {
                  "builtin_report_service": true,
                  "db_server_host": "mysql-server",
                  "db_server_port": 3306,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix",
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              }
            ]
          ansible.operator-sdk/reconcile-period: 5s
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          certified: "true"
          containerImage: registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
          createdAt: "2025-05-26T07:41:22Z"
          description: Zabbix operator with multiple deployment variants and different
            components
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/zabbix-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/zabbix-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <6.0.40
          operatorframework.io/suggested-namespace: zabbix
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: ""
          support: Zabbix
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Zabbix agent 2 is a new generation of Zabbix agent and may
              be used in place of Zabbix agent
            displayName: Zabbix agent 2
            kind: ZabbixAgent2
            name: zabbixagent2s.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix agent is deployed on a monitoring nodes to actively
              monitor local resources and applications
            displayName: Zabbix agent
            kind: ZabbixAgent
            name: zabbixagents.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database server, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Full
            kind: ZabbixFull
            name: zabbixfulls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with MySQL database server
            displayName: Zabbix proxy (MySQL)
            kind: ZabbixProxyMysql
            name: zabbixproxymysqls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with SQLite database
            displayName: Zabbix proxy (SQLite3)
            kind: ZabbixProxySqlite
            name: zabbixproxysqlites.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database support, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Server
            kind: ZabbixServer
            name: zabbixservers.kubernetes.zabbix.com
            version: v1alpha1
        description: "## About this Operator\n\nZabbix helps you to real-time monitoring
          of millions of metrics collected from tens of thousands of servers, virtual
          machines and network devices.\nThe Zabbix Operator allows users to easily
          deploy, manage, and maintain Zabbix deployments on OpenShift. By installing
          this integration you will be able to deploy Zabbix server / proxies and
          other components with a single command.\n## Supported Features\n* **Zabbix
          Server** - Simple Zabbix installation with included Zabbix server, Zabbix
          web-interface and Zabbix Java Gateway with MySQL database support. The feature
          does not provide MySQL service and requires an external MySQL database.
          \n* **Zabbix Server (Full)** - Zabbix installation with included Zabbix
          server, Zabbix web-interface, Zabbix Java Gateway and MySQL server instance.\n*
          **Zabbix proxy (SQLite3)** - Very simple way to gain power of Zabbix proxy.
          The feature has  SQLite3 support for Zabbix proxies and allow to specify
          amount of proxies. \n* **Zabbix proxy (MySQL)** - Another option of Zabbix
          proxy. The option support and deliver MySQL database.\n* **Zabbix agent**
          - Zabbix agent can be deployed on each available node for stability and
          performance monitoring on remote nodes. It allows to gather metrics with
          full automation!* \n## Prerequisites\nAll deployment options are require
          additional information during deployment. Please, check the following instructions
          and provide required configuration:\n* **Zabbix Server** - MySQL database
          host information and MySQL database credentials in specially formatted *Secret*.
          Additionally it is possible to specify SSL certificates for HTTPS support
          in *Secret*.\n* **Zabbix Server (Full)** - MySQL database credentials in
          specially formatted *Secret*. MySQL database volume name information. Additionally
          it is possible to specify SSL certificates for HTTPS support in *Secret*.\n*
          **Zabbix proxy (SQLite3)** - Zabbix server host information only.\n* **Zabbix
          proxy (MySQL)** - MySQL database credentials in specially formatted *Secrets*
          and Zabbix server host information.\n* **Zabbix agent** - Zabbix server
          host information only for outgoing and incoming connections. Zabbix agent
          uses \"privileged\" mode to monitor node resources! For example, running
          processes.* \n### MySQL credentials\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-full-secrets\ndata:\n  mysql_root_password: emFiYml4X3N1cGVyX3Jvb3Q=
          [1]\n  mysql_zabbix_password: emFiYml4X3VzZXJfcGFzc3dvcmQ= [2]\n  mysql_zabbix_username:
          emFiYml4 [3]\ntype: Opaque\n```\nUsing MySQL root password (*mysqlrootpassword*
          [1]) Zabbix server / proxy will try to create MySQL database schema with
          grant permissions to *mysqlzabbixusername* [2] and *mysqlzabbixpassword*
          [3].\n### SSL certificates for HTTPS\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-web-sslsecret\ndata:\n  ssl.crt: >-\n   < ssl.crt data>\n
          \ ssl.key: >-\n\t < ssl.key data >\n  dhparam.pem: >-\n   <  dhparam.pem
          data >\n```\nFiles *ssl.crt*, *ssl.key* and *dhparam.perm* are required
          for Zabbix web-interface for SSL support.\n"
        displayName: Zabbix Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - zabbix
        - appliance
        - agent
        - metrics
        links:
        - name: Zabbix
          url: https://www.zabbix.com
        - name: Zabbix Official Documentation
          url: https://www.zabbix.com/documentation/6.0/en/manual/quickstart
        - name: Zabbix Operator Installation
          url: https://www.zabbix.com/documentation/6.0/en/manual/installation/containers/openshift
        - name: Downloads
          url: https://www.zabbix.com/download
        maintainers:
        - email: alexey.pustovalov@zabbix.com
          name: Alexey Pustovalov
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Zabbix LLC
        relatedImages:
        - registry.connect.redhat.com/zabbix/zabbix-agent2-60@sha256:424003d52f3cb4360adf2b3619d8b9f8de9196c630ee7e4ceb5d177c1c5a36e4
        - registry.connect.redhat.com/zabbix/zabbix-server-mysql-60@sha256:637484faec6e9a55049c13dbd1f100c03d757c3a21749388c0605040f02d05ce
        - registry.connect.redhat.com/zabbix/zabbix-proxy-mysql-60@sha256:c2ae74d97c083ff348dc9c55932a33da4779cd9efdd9d7c5a79acfffae5b9ed0
        - registry.connect.redhat.com/zabbix/zabbix-java-gateway-60@sha256:a4feb725c12c4ac90eb9244c6fafdb6708679f2e0e7f6aa683c765127306412e
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
        - registry.connect.redhat.com/zabbix/zabbix-web-mysql-60@sha256:611d2670e1bf95b81226976c0a1f64434b2a7766f196dadbd646a7d8319f4cb1
        - registry.connect.redhat.com/zabbix/zabbix-web-service-60@sha256:4a333094e110713470a897b6406c5e340e962caf32a429dddf19865041fcc448
        - registry.redhat.io/rhel9/mysql-80@sha256:c4b5246d0c0d089a766ea07768c22d9d6be8049ac1379bd345db99390ffe2a36
        - registry.connect.redhat.com/zabbix/zabbix-snmptraps-60@sha256:e61e6b1c551dec8488cce0998f94d83d6d10b5d89a465d7119b088313032cb81
        - registry.connect.redhat.com/zabbix/zabbix-proxy-sqlite-60@sha256:62e62ebc8c1bc13fb610f7a83412e65d852f74b3c2c212baa75ea27fc7fa03eb
        - registry.connect.redhat.com/zabbix/zabbix-agent-60@sha256:2895b5e45785aaf528d4a6f78c8af0a0d494dbe9362403fdd470673d911d5004
        - quay.io/zabbix/zabbix-operator@sha256:43d26488a19967bdb2a8aafde5f43820be95011811da9cc5ec623c9029801675
        version: 6.0.40
      entries:
      - name: zabbix-operator-certified-rhmp.v6.0.40
        version: 6.0.40
      - name: zabbix-operator-certified-rhmp.v6.0.38
        version: 6.0.38
      - name: zabbix-operator-certified-rhmp.v6.0.36
        version: 6.0.36
      - name: zabbix-operator-certified-rhmp.v6.0.33
        version: 6.0.33
      - name: zabbix-operator-certified-rhmp.v6.0.29
        version: 6.0.29
      - name: zabbix-operator-certified-rhmp.v6.0.27
        version: 6.0.27
      - name: zabbix-operator-certified-rhmp.v6.0.26
        version: 6.0.26
      - name: zabbix-operator-certified-rhmp.v6.0.25
        version: 6.0.25
      - name: zabbix-operator-certified-rhmp.v6.0.17
        version: 6.0.17
      - name: zabbix-operator-certified-rhmp.v6.0.14
        version: 6.0.14
      - name: zabbix-operator-certified-rhmp.v6.0.13
        version: 6.0.13
      name: lts
    defaultChannel: lts
    packageName: zabbix-operator-certified-rhmp
    provider:
      name: Zabbix LLC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/streamshub
    name: amq-streams-console
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: alpha
    - currentCSV: amq-streams-console.v2.8.0-13-0.1738265629.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:1b1b80710885b43600e22ff2a24930b5cab06cd91fad439ad8a282c00b4564b9
          createdAt: "2024-11-12T15:34:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-streams-console.v2.8.0-13
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.8
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.23.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:1b1b80710885b43600e22ff2a24930b5cab06cd91fad439ad8a282c00b4564b9
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:38b80d81e54d59667c1f7dfa4da4125d5c7430816da130ddc84678a4e3cff6cf
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:366d60ff48fb9ba32dc73d92a775046d7b1eb1df21c93e31605372c7858f16ca
        version: 2.8.0-13+0.1738265629.p
      entries:
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.8.x
    - currentCSV: amq-streams-console.v2.9.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
          createdAt: "2025-09-22T10:41:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0-0 <2.9.2-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.9
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:839368767bec735cd12f002874917c23f58550fe21df976858bdc92946985ec6
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:b7444b70a5dc4543a9976c08a619ff6978c9ff86ac7085d300b4870b7c6b7851
        version: 2.9.2-2
      entries:
      - name: amq-streams-console.v2.9.2-2
        version: 2.9.2-2
      - name: amq-streams-console.v2.9.2-1
        version: 2.9.2-1
      - name: amq-streams-console.v2.9.1-7
        version: 2.9.1-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.9.x
    - currentCSV: amq-streams-console.v2.9.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
          createdAt: "2025-09-22T10:41:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0-0 <2.9.2-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.9
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:839368767bec735cd12f002874917c23f58550fe21df976858bdc92946985ec6
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:b7444b70a5dc4543a9976c08a619ff6978c9ff86ac7085d300b4870b7c6b7851
        version: 2.9.2-2
      entries:
      - name: amq-streams-console.v2.9.2-2
        version: 2.9.2-2
      - name: amq-streams-console.v2.9.2-1
        version: 2.9.2-1
      - name: amq-streams-console.v2.9.1-7
        version: 2.9.1-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.x
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-3.0.x
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-3.x
    defaultChannel: alpha
    packageName: amq-streams-console
    provider:
      name: Red Hat
      url: https://github.com/streamshub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: 3scale-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: 3scale-operator.v0.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "ProductCRName": "product1-sample",
                  "Production": true
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/3scale-rhel7-operator@sha256:0751f9f5343a1264224c36db3db292e5a9e28c40cd96649ef48b5399c34668dd
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.16"}]'
          olm.skipRange: '>=0.10.0 <0.11.13'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/3scale-rhel7-operator@sha256:0751f9f5343a1264224c36db3db292e5a9e28c40cd96649ef48b5399c34668dd
        - registry.redhat.io/3scale-amp2/backend-rhel8@sha256:cdc53799ff9d022462a9a4afcbcd8043497cd40ccdc6fafdcb9ccf7ce69f43d1
        - registry.redhat.io/3scale-amp2/system-rhel7@sha256:61dcc91e5d3c91a0ea85c02f2599cadcb5496b1c09a8838b4e1c8e4efe126e8f
        - registry.redhat.io/3scale-amp2/memcached-rhel7@sha256:7e2ab1e24646fbfef878eb2b5a74ed865ae812b60aa0ba505045f838f26ebbfc
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:a95d09fc3b224f550dec3de3d23fd2dbfc0a220dc869b4ad9559ee2f85327daa
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:f8cd25658fa6a154e7ea6e5adf0f9dae3fc596ec9989021de6a021ec94eef226
        - registry.redhat.io/rhel8/redis-6@sha256:5250d644e80aea794b0f732e9285883a4348dace83557f177f64d78fe4dcabd0
        - registry.redhat.io/openshift4/ose-cli@sha256:1a8caae40be21f23ba3cb48b75809a73333907033a6041927aa568b5c9290f3c
        - registry.redhat.io/3scale-amp2/searchd-rhel7@sha256:8e2001c021fad554cb8a96bb34e9973010fe4206526f19aa040fa196709fb688
        - registry.redhat.io/3scale-amp2/zync-rhel9@sha256:0d2205d9f61a58414c4ba472ce8f19b12d2df722772847d5fb7442e664759627
        - registry.redhat.io/rhel8/mysql-80@sha256:08547a93659ad804d637b91e8e1eb3464b75783d494e6c40e1c921bd74f31645
        version: 0.11.13
      entries:
      - name: 3scale-operator.v0.11.13
        version: 0.11.13
      - name: 3scale-operator.v0.11.12
        version: 0.11.12
      name: threescale-2.14
    - currentCSV: 3scale-operator.v0.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                },
                "status": {
                  "deployments": {
                    "ready": [
                      "deployment x ready"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                },
                "status": {
                  "adminId": 1,
                  "tenantId": 2
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ApplicationAuth",
                "metadata": {
                  "name": "applicationauth-sample"
                },
                "spec": {
                  "applicationCRName": "application-cr-name",
                  "authSecretRef": {
                    "name": "auth-secret-reference"
                  },
                  "generateSecret": true
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "productCRName": "product1-sample",
                  "production": true
                },
                "status": {}
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/3scale-rhel9-operator@sha256:fcea6a77d0a48f35f4c3ce63e72dde104aac6b580f340950b311eb418fff5ec9
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=0.11.0 <0.12.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          rht_backend_redis_requirements: 6.2.0
          rht_mysql_requirements: 8.0.0
          rht_postgres_requirements: 10.0.0
          rht_system_redis_requirements: 6.2.0
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: ApplicationAuth is the Schema for the applicationauths API
            displayName: Application Auth
            kind: ApplicationAuth
            name: applicationauths.capabilities.3scale.net
            version: v1beta1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/3scale-rhel9-operator@sha256:fcea6a77d0a48f35f4c3ce63e72dde104aac6b580f340950b311eb418fff5ec9
        - registry.redhat.io/3scale-amp2/manticore-rhel9@sha256:ccf6baf5f2e9494a34739a79abe0f46e36d8ab9dc81c426f7a3a8964e8f39859
        - registry.redhat.io/3scale-amp2/backend-rhel8@sha256:eb2288de900dc42ead755b28cdb2509b25ecc8dd256781beaa4a2a86e7cbe532
        - registry.redhat.io/rhel9/memcached@sha256:8738110ef16c9e35deeb07c0577ffe09d6ef029761b504a8d20253fa715da394
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:a95d09fc3b224f550dec3de3d23fd2dbfc0a220dc869b4ad9559ee2f85327daa
        - registry.redhat.io/3scale-amp2/zync-rhel9@sha256:df9d62897ff286165772150a8b5e6952d2ce3b52a8afa07bd693a2d95c2c67ff
        - registry.redhat.io/openshift4/ose-cli@sha256:1a8caae40be21f23ba3cb48b75809a73333907033a6041927aa568b5c9290f3c
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/rhel8/mysql-80@sha256:ee9aca9096473da004f9140305742eb7b9a77ed81ce55d5a6503388cccab030f
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        - registry.redhat.io/3scale-amp2/system-rhel8@sha256:4c72dfeb1919a1de4fbc47f4611d0c830ef89cf9f17255ad55a9b370618e0e8f
        version: 0.12.5
      entries:
      - name: 3scale-operator.v0.12.5
        version: 0.12.5
      - name: 3scale-operator.v0.12.4
        version: 0.12.4
      - name: 3scale-operator.v0.12.3
        version: 0.12.3
      - name: 3scale-operator.v0.12.2
        version: 0.12.2
      - name: 3scale-operator.v0.12.1
        version: 0.12.1
      - name: 3scale-operator.v0.12.0
        version: 0.12.0
      name: threescale-2.15
    - currentCSV: 3scale-operator.v0.12.1-mas
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                },
                "status": {
                  "deployments": {
                    "ready": [
                      "deployment x ready"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                },
                "status": {
                  "adminId": 1,
                  "tenantId": 2
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ApplicationAuth",
                "metadata": {
                  "name": "applicationauth-sample"
                },
                "spec": {
                  "applicationCRName": "application-cr-name",
                  "authSecretRef": {
                    "name": "auth-secret-reference"
                  },
                  "generateSecret": true
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "productCRName": "product1-sample",
                  "production": true
                },
                "status": {}
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-mas/3scale-rhel9-operator@sha256:8f41ab6027616e4bb65b1fddbcf44cc8a25a79cdda19d5cc354ac4228a15141e
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.17"}]'
          olm.skipRange: '>=0.11.8 <0.12.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          rht_backend_redis_requirements: 6.2.0
          rht_mysql_requirements: 8.0.0
          rht_postgres_requirements: 10.0.0
          rht_system_redis_requirements: 6.2.0
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: ApplicationAuth is the Schema for the applicationauths API
            displayName: Application Auth
            kind: ApplicationAuth
            name: applicationauths.capabilities.3scale.net
            version: v1beta1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-mas/3scale-manticore-search-rhel9@sha256:7379186175c9c0f0a6d922cd2a0754b429c9728dfbe8979a9d6025789ebae0da
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:a95d09fc3b224f550dec3de3d23fd2dbfc0a220dc869b4ad9559ee2f85327daa
        - registry.redhat.io/3scale-mas/backend-rhel8@sha256:d0c561c6b9dfe9741722ba4a257fbd2df4d2a709457a0b0657ef7bcca1f2a5ca
        - registry.redhat.io/3scale-mas/zync-rhel9@sha256:78e5223a1a878f6527bd88d382948c125b7cb4a81540648244553af426d1a944
        - registry.redhat.io/3scale-mas/memcached-rhel7@sha256:7a1abee11b4120f399485b438bda1e5262f04de28d3e788b207eecc3a9de9ed2
        - registry.redhat.io/3scale-mas/3scale-rhel9-operator@sha256:8f41ab6027616e4bb65b1fddbcf44cc8a25a79cdda19d5cc354ac4228a15141e
        - registry.redhat.io/rhel8/mysql-80@sha256:754f6830e0bc4aadb973b1f613f2bc7278ec13ad0f0e1e87655a2bb2f76cb9b3
        - registry.redhat.io/openshift4/ose-cli@sha256:a688994ea404ce6ce4bb5ec1249f0e779de0d6c187b5078e2ff067e60b50937d
        - registry.redhat.io/3scale-mas/apicast-gateway-rhel8@sha256:f174636a9ad43a2419fff5a9e6797e3185c540a9509188fd1606aef4acea7bdf
        - registry.redhat.io/rhel8/redis-6@sha256:218f7fa2458c274cc34f79190423e0aba03027f55ef6b3334d916951a668b7b7
        - registry.redhat.io/3scale-mas/system-rhel8@sha256:0fa28aaf3467fbd22fc1335dc6a712043e4dead694c3f14b60e89740bb7eb13e
        version: 0.12.1-mas
      entries:
      - name: 3scale-operator.v0.12.1-mas
        version: 0.12.1-mas
      - name: 3scale-operator.v0.11.8-mas
        version: 0.11.8-mas
      - name: 3scale-operator.v0.11.7-mas
        version: 0.11.7-mas
      - name: 3scale-operator.v0.11.6-mas
        version: 0.11.6-mas
      - name: 3scale-operator.v0.11.5-mas
        version: 0.11.5-mas
      - name: 3scale-operator.v0.11.3-mas
        version: 0.11.3-mas
      - name: 3scale-operator.v0.11.2-mas
        version: 0.11.2-mas
      - name: 3scale-operator.v0.11.1-mas
        version: 0.11.1-mas
      - name: 3scale-operator.v0.11.0-mas
        version: 0.11.0-mas
      - name: 3scale-operator.v0.10.0-mas
        version: 0.10.0-mas
      name: threescale-mas
    defaultChannel: threescale-2.15
    packageName: 3scale-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: operator-certification-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operator-certification-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "certification.redhat.com/v1alpha1",
                "kind": "OperatorPipeline",
                "metadata": {
                  "name": "operatorpipeline-sample"
                },
                "spec": {
                  "applyCIPipeline": true,
                  "applyHostedPipeline": false,
                  "applyReleasePipeline": false,
                  "gitHubSecretName": "github-api-token",
                  "kubeconfigSecretName": "kubeconfig",
                  "operatorPipelinesRelease": "main",
                  "pyxisSecretName": "pyxis-api-secret"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-06-23T16:02:00Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io.bundle.channel.default.v1: alpha
          operators.operatorframework.io.bundle.channels: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-openshift-ecosystem/operator-certification-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OperatorPipeline is the Schema for the operatorpipelines
              API
            displayName: Operator Pipeline
            kind: OperatorPipeline
            name: operatorpipelines.certification.redhat.com
            version: v1alpha1
        description: |-
          A Kubernetes operator to provision resources for the operator certification pipeline. This operator is installed in all namespaces which can support multi-tenant scenarios. **Note:** This operator should only be used by Red Hat partners attempting to certify their operator(s).


          **Requirements**


          The certification operator requires that you have the following tools installed, functional, and in your path.
          - [Install](https://docs.openshift.com/container-platform/4.8/cli_reference/openshift_cli/getting-started-cli.html#installing-openshift-cli) oc, the OpenShift CLI tool (tested with version 4.7.13)
          - [Install](https://tekton.dev/docs/cli/) tkn, the Tekton CLI tool (tested with version 0.19.1)
          - [Install](https://git-scm.com/downloads) git, the Git CLI tool (tested with 2.32.0)
          - The certification pipeline expects you to have the source files of your Operator bundle


          **Pre - Installation**


          The below steps exist for simplicity and to make the installation clearer. The operator watches all namespaces, so if secrets/configs/etc already exist in another namespace, feel free to use the existing namespace when following the operator installation steps.


          **Create a new namespace where the following secrets will be applied.**
          ```
          oc new-project oco
          ```


          **Add the kubeconfig secret which will be used to deploy the operator under test and run the certification checks.**
          * Open a terminal window
          * Set the `KUBECONFIG` environment variable
          ```
          export KUBECONFIG=/path/to/your/cluster/kubeconfig
          ```
          > *This kubeconfig will be used to deploy the Operator under test and run the certification checks.*
          ```
          oc create secret generic kubeconfig --from-file=kubeconfig=$KUBECONFIG
          ```


          **Configuring steps for submitting the results**
          - Add the github API token to the repo where the PR will be created
          ```
          oc create secret generic github-api-token --from-literal GITHUB_TOKEN=<github token>
          ```
          - Add RedHat Container API access key

            This API access key is specifically related to your unique partner account for Red Hat Connect portal. Instructions to obtain your API key can be found: [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/operator-cert-workflow.md#step-b---get-api-key)
          ```
          oc create secret generic pyxis-api-secret --from-literal pyxis_api_key=< API KEY >
          ```

          - Optional pipeline configurations can be found [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md#optional-configuration)


          **Execute the Pipeline (Development Iterations)**


          A pre-requisite to running a pipeline is that a `workspace-template.yaml` exists in the directory you want to execute the `tkn` commands from.

          To create a workspace-template.yaml
          ```
          cat <<EOF > workspace-template.yaml
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 5Gi
          EOF
          ```

          There are multiple ways to execute the Pipeline which can be found [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md#execute-the-pipeline-development-iterations)
        displayName: Operator Certification Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - certification
        - pipeline
        - openshift
        links:
        - name: Operator Certification CI Pipeline
          url: https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md
        - name: Red Hat OpenShift Pipelines
          url: https://github.com/openshift/tektoncd-operator
        maintainers:
        - email: certification-operator-devel@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/opdev/operator-certification-operator@sha256:7c84772f96650ff757c2552ba421355d54da41a5ed2155f72a14911b8a3f18ff
        version: 1.3.1
      entries:
      - name: operator-certification-operator.v1.3.1
        version: 1.3.1
      name: alpha
    defaultChannel: alpha
    packageName: operator-certification-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: LINBIT
      provider-url: https://linbit.com
    name: linstor-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: linstor-operator.v2.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorCluster",
                "metadata": {
                  "name": "linstorcluster"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorNodeConnection",
                "metadata": {
                  "name": "example-connection"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatellite",
                "metadata": {
                  "name": "example-satellite"
                },
                "spec": {
                  "clusterRef": {
                    "name": "linstorcluster"
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatelliteConfiguration",
                "metadata": {
                  "name": "all-satellites"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          com.redhat.openshift.versions: v4.7
          containerImage: drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
          createdAt: "2025-09-25T13:44:22Z"
          description: LINSTOR Operator deploys LINBIT SDS in the cluster.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: linbit-sds
          operators.openshift.io/valid-subscription: '[''LINBIT SDS Subscription'']'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/internal-objects: '["linstorsatellites.piraeus.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster-wide configuration for LINSTOR.
            displayName: Linstor Cluster
            kind: LinstorCluster
            name: linstorclusters.piraeus.io
            version: v1
          - description: Configuration for connections between specific nodes.
            displayName: Linstor Node Connection
            kind: LinstorNodeConnection
            name: linstornodeconnections.piraeus.io
            version: v1
          - description: Configuration for specific Satellites/Nodes in the LINSTOR
              Cluster.
            displayName: Linstor Satellite Configuration
            kind: LinstorSatelliteConfiguration
            name: linstorsatelliteconfigurations.piraeus.io
            version: v1
          - description: Resource to configure a single satellite, managed by the
              Operator itself.
            displayName: Linstor Satellite
            kind: LinstorSatellite
            name: linstorsatellites.piraeus.io
            version: v1
          required:
          - kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
        description: |
          The LINSTOR Operator manages [LINSTOR](https://github.com/LINBIT/linstor-server) clusters in Kubernetes.

          All components of the LINSTOR software stack can be managed by the operator:
          * DRBD
          * LINSTOR
          * LINSTOR CSI driver
          * LINSTOR HA Controller
        displayName: LINSTOR Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: User Guide
          url: https://linbit.com/drbd-user-guide/linstor-guide-1_0-en/
        - name: LINSTOR Operator
          url: https://charts.linstor.io/
        - name: Source Code
          url: https://github.com/LINBIT/linstor-operator-builder
        maintainers:
        - email: support@linbit.com
          name: LINBIT
        maturity: stable
        minKubeVersion: 1.20.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: DaemonSet
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: storage.k8s.io
          kind: CSIDriver
          version: v1
        provider:
          name: LINBIT
          url: https://linbit.com
        relatedImages:
        - drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:f8cde5b5c289f7b3d233773f4e79990ba392fe6e14e0177de80714e7d3e98862
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ac3963bfd4246da386b82d810a348e9d42a9e4b746acaf4dbcbe0766643cc2fa
        - drbd.io/drbd-shutdown-guard@sha256:7ba049ecb205a0086e425d2ff37285b00449549d89e179b4b943ace6e59a9156
        - drbd.io/drbd9-rhel8@sha256:7cddfa770de66cefcb42c2b33da57d0f1adfcc5f1cc12522977aa6993722bb2a
        - drbd.io/linstor-k8s-ha-controller@sha256:ed0f0f86dad88feadf8c3c32d7d143eaa52b1130af3576315a562909b39d2afa
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:2669cde3db2268d37bc671f494eaa06480478a90b8943054af14cbffc050debf
        - drbd.io/drbd-reactor@sha256:77c51aeeefdcd788cb1583738f93778032fcc03c36fb289b692d3cf195c8b184
        - drbd.io/linstor-csi@sha256:9c7ef452b0997d3373803201614c3b8075687e75c397359738a9e4ef5530cab3
        - drbd.io/linstor-controller@sha256:841b3aa84cdf13aaaf1243bc5da8ac2205ca3fe0186fdfc7a04b69397f0020d6
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:624f61b1ae9d32f5a074ee8c12839d975fba9efcd3bfa7981cfe38febf5c4495
        - drbd.io/drbd9-rhel9@sha256:4eb553336ed64e127068a4566370b040166bf11301e21ac47ac5b92237ec9d2a
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:5a650dc21a6a038de6267f117d01adfd904ff3f5c94f428f6a13735f18bf94d6
        - drbd.io/drbd9-rhel7@sha256:0c96b216e55a128a9944276af9080aced8bd72fdf1fd424f024ed9ace2c1feb5
        - drbd.io/ktls-utils@sha256:3351a6e0fc048fe1544499095cdb8fdd221e832cecfa8a9a01f046c271991138
        - drbd.io/linstor-satellite@sha256:ce0a0b05494c6e6d5f16121e5762bf58e7b78f40c00b84780e2abba2a690af1b
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:700be620923fc99ed03f9f1c4a911e4084beb30685361270c49165b27ad1d374
        - drbd.io/drbd9-rhel10@sha256:da0a7015cf15724a7c73100d505891e62262a726141347df7f2d51ec79a0a1e1
        - drbd.io/linstor-affinity-controller@sha256:78e5b299ef1702114528678387b5e74821d7658bc02f05f0bd7edd9a176fba39
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:cd58f1789027e32eae08f00dac7b2275496c41020fb7cdebc10e9986c625d4f4
        version: 2.9.1
      entries:
      - name: linstor-operator.v2.9.1
        version: 2.9.1
      name: dev
    - currentCSV: linstor-operator.v2.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorCluster",
                "metadata": {
                  "name": "linstorcluster"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorNodeConnection",
                "metadata": {
                  "name": "example-connection"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatellite",
                "metadata": {
                  "name": "example-satellite"
                },
                "spec": {
                  "clusterRef": {
                    "name": "linstorcluster"
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatelliteConfiguration",
                "metadata": {
                  "name": "all-satellites"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          com.redhat.openshift.versions: v4.7
          containerImage: drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
          createdAt: "2025-09-25T13:44:22Z"
          description: LINSTOR Operator deploys LINBIT SDS in the cluster.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: linbit-sds
          operators.openshift.io/valid-subscription: '[''LINBIT SDS Subscription'']'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/internal-objects: '["linstorsatellites.piraeus.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster-wide configuration for LINSTOR.
            displayName: Linstor Cluster
            kind: LinstorCluster
            name: linstorclusters.piraeus.io
            version: v1
          - description: Configuration for connections between specific nodes.
            displayName: Linstor Node Connection
            kind: LinstorNodeConnection
            name: linstornodeconnections.piraeus.io
            version: v1
          - description: Configuration for specific Satellites/Nodes in the LINSTOR
              Cluster.
            displayName: Linstor Satellite Configuration
            kind: LinstorSatelliteConfiguration
            name: linstorsatelliteconfigurations.piraeus.io
            version: v1
          - description: Resource to configure a single satellite, managed by the
              Operator itself.
            displayName: Linstor Satellite
            kind: LinstorSatellite
            name: linstorsatellites.piraeus.io
            version: v1
          required:
          - kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
        description: |
          The LINSTOR Operator manages [LINSTOR](https://github.com/LINBIT/linstor-server) clusters in Kubernetes.

          All components of the LINSTOR software stack can be managed by the operator:
          * DRBD
          * LINSTOR
          * LINSTOR CSI driver
          * LINSTOR HA Controller
        displayName: LINSTOR Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: User Guide
          url: https://linbit.com/drbd-user-guide/linstor-guide-1_0-en/
        - name: LINSTOR Operator
          url: https://charts.linstor.io/
        - name: Source Code
          url: https://github.com/LINBIT/linstor-operator-builder
        maintainers:
        - email: support@linbit.com
          name: LINBIT
        maturity: stable
        minKubeVersion: 1.20.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: DaemonSet
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: storage.k8s.io
          kind: CSIDriver
          version: v1
        provider:
          name: LINBIT
          url: https://linbit.com
        relatedImages:
        - drbd.io/drbd9-rhel7@sha256:0c96b216e55a128a9944276af9080aced8bd72fdf1fd424f024ed9ace2c1feb5
        - drbd.io/linstor-controller@sha256:841b3aa84cdf13aaaf1243bc5da8ac2205ca3fe0186fdfc7a04b69397f0020d6
        - drbd.io/drbd9-rhel9@sha256:4eb553336ed64e127068a4566370b040166bf11301e21ac47ac5b92237ec9d2a
        - drbd.io/linstor-k8s-ha-controller@sha256:ed0f0f86dad88feadf8c3c32d7d143eaa52b1130af3576315a562909b39d2afa
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:f8cde5b5c289f7b3d233773f4e79990ba392fe6e14e0177de80714e7d3e98862
        - drbd.io/drbd-shutdown-guard@sha256:7ba049ecb205a0086e425d2ff37285b00449549d89e179b4b943ace6e59a9156
        - drbd.io/linstor-affinity-controller@sha256:78e5b299ef1702114528678387b5e74821d7658bc02f05f0bd7edd9a176fba39
        - drbd.io/linstor-satellite@sha256:ce0a0b05494c6e6d5f16121e5762bf58e7b78f40c00b84780e2abba2a690af1b
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:5a650dc21a6a038de6267f117d01adfd904ff3f5c94f428f6a13735f18bf94d6
        - drbd.io/linstor-csi@sha256:9c7ef452b0997d3373803201614c3b8075687e75c397359738a9e4ef5530cab3
        - drbd.io/ktls-utils@sha256:3351a6e0fc048fe1544499095cdb8fdd221e832cecfa8a9a01f046c271991138
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ac3963bfd4246da386b82d810a348e9d42a9e4b746acaf4dbcbe0766643cc2fa
        - drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:cd58f1789027e32eae08f00dac7b2275496c41020fb7cdebc10e9986c625d4f4
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:624f61b1ae9d32f5a074ee8c12839d975fba9efcd3bfa7981cfe38febf5c4495
        - drbd.io/drbd9-rhel8@sha256:7cddfa770de66cefcb42c2b33da57d0f1adfcc5f1cc12522977aa6993722bb2a
        - drbd.io/drbd-reactor@sha256:77c51aeeefdcd788cb1583738f93778032fcc03c36fb289b692d3cf195c8b184
        - drbd.io/drbd9-rhel10@sha256:da0a7015cf15724a7c73100d505891e62262a726141347df7f2d51ec79a0a1e1
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:700be620923fc99ed03f9f1c4a911e4084beb30685361270c49165b27ad1d374
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:2669cde3db2268d37bc671f494eaa06480478a90b8943054af14cbffc050debf
        version: 2.9.1
      entries:
      - name: linstor-operator.v2.9.1
        version: 2.9.1
      name: stable
    defaultChannel: dev
    packageName: linstor-operator
    provider:
      name: LINBIT
      url: https://linbit.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: https://opdev.github.io/
    name: loki-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.example.com/v1alpha1",
                "kind": "Loki",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/install-disable-crds": "true"
                  },
                  "name": "loki-sample"
                },
                "spec": {
                  "global": {
                    "clusterDomain": "cluster.local",
                    "dnsNamespace": "openshift-dns",
                    "dnsService": "dns-default"
                  },
                  "loki": {
                    "auth_enabled": false,
                    "commonConfig": {
                      "replication_factor": 1
                    },
                    "storage": {
                      "type": "filesystem"
                    }
                  },
                  "monitoring": {
                    "dashboards": {
                      "enabled": false
                    },
                    "lokiCanary": {
                      "enabled": false
                    },
                    "rules": {
                      "alerting": false,
                      "enabled": false
                    },
                    "selfMonitoring": {
                      "enabled": false,
                      "grafanaAgent": {
                        "installOperator": false
                      }
                    },
                    "serviceMonitor": {
                      "enabled": false,
                      "metricsInstance": {
                        "enabled": false
                      }
                    }
                  },
                  "rbac": {
                    "sccEnabled": true
                  },
                  "singleBinary": {
                    "replicas": 1
                  },
                  "test": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/loki-helm-operator-controller:0.0.1
          createdAt: "2023-07-14T15:06:17Z"
          description: An experimental operator that installs loki from its official
            helmchart.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys the loki operator helm chart
            displayName: Loki
            kind: Loki
            name: lokis.charts.example.com
            version: v1alpha1
        description: Loki operator based on the helm chart with single binary mode
          deployment optionality
        displayName: Loki Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loki
        - single-binary
        - demo
        - helm
        links:
        - name: Loki Helm Operator
          url: https://github.com/skattoju/loki-helm-operator
        maintainers:
        - email: skattoju@redhat.com
          name: Sid Kattoju
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: https://opdev.github.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/opdev/loki-helm-operator-controller:0.0.1
        version: 0.0.1
      entries:
      - name: loki-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: loki-helm-operator
    provider:
      name: opdev
      url: https://opdev.github.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat CCIT
      provider-url: ""
    name: reportportal-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: reportportal-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ReportPortal",
                "metadata": {
                  "name": "reportportal-sample"
                },
                "spec": {
                  "accesskey": "AKIAIOSFODNN7EXAMPLE",
                  "analyzer_image": "quay.io/waynesun09/service-auto-analyzer:5.3.5",
                  "api_image": "quay.io/waynesun09/service-api:5.3.5-rootless",
                  "api_replicas": 1,
                  "app_domain": "apps.test-example.com",
                  "bucket_prefix": null,
                  "default_bucket_name": null,
                  "enable_pg_restic_backup": "yes",
                  "es_backup_schedule": "@daily",
                  "es_s3_backup_dir": "s3_backup",
                  "es_snapshot_bucket": "es-snapshot-123123",
                  "gateway_replicas": 1,
                  "index_image": "quay.io/waynesun09/service-index:5.0.10",
                  "index_replicas": 1,
                  "integration_salt_path": null,
                  "migration_image": "quay.io/waynesun09/migrations:5.3.5",
                  "pg_restic_password": "rp_user",
                  "pg_restic_s3_bucket_init": "yes",
                  "pg_s3_bucket_name": "pgbackup-123123",
                  "s3_enabled": false,
                  "s3_endpoint": null,
                  "secretkey": "wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY",
                  "uat_image": "quay.io/waynesun09/service-authorization:5.3.5",
                  "ui_image": "quay.io/waynesun09/service-ui:5.3.5",
                  "ui_replicas": 1
                }
              },
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ReportPortalRestore",
                "metadata": {
                  "name": "reportportalrestore-sample"
                },
                "spec": {
                  "accesskey": "AKIAIOSFODNN7EXAMPLE",
                  "bucket_prefix": null,
                  "es_restore": "yes",
                  "es_restore_date": "2020-08-07",
                  "es_s3_backup_dir": "s3_backup",
                  "es_snapshot_bucket": "es-snapshot-123123",
                  "pg_recovery": "yes",
                  "pg_recovery_target_time": "2020-08-07 00:00:00",
                  "pg_restic_password": "rp_user",
                  "pg_s3_bucket_name": "pgbackup-123123",
                  "s3_endpoint": null,
                  "secretkey": "wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY"
                }
              },
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ServiceApiServiceMonitor",
                "metadata": {
                  "name": "serviceapiservicemonitor-sample"
                },
                "spec": {
                  "bearer_token_secret_key": "token",
                  "bearer_token_secret_name": "rp-user-auth"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/waynesun09/rp5-operator:v0.0.8
          createdAt: "2021-04-22 14:39:00"
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/waynesun09/reportportal-operator
          support: waynesun09
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ReportPortal Restore from s3 backup
            kind: ReportPortalRestore
            name: reportportalrestores.rp5.reportportal.io
            version: v1
          - description: ReportPortal v5 Deployment with ReportPortal Operator
            kind: ReportPortal
            name: reportportals.rp5.reportportal.io
            version: v1
          - description: ReportPortal service-api ServiceMonitor for metrics
            kind: ServiceApiServiceMonitor
            name: serviceapiservicemonitors.rp5.reportportal.io
            version: v1
        description: |-
          ReportPortal is a test automation analytics platform. ReportPortal Operator support instantiating the ReportPortal services and required dependencies.

          What services are included:

              - Report Portal v5.3
              - PostgreSQL
              - Elasticsearch
              - RabbitMQ
              - Minio

          The Operator also support data backup & restore and metrics monitoring.

          ## Install

          Install the Operator to a namespace, then create ReportPortal v5 instance under Provided APIs on the UI, all Report Portal v5 services and dependencies will be installed.

          ## Data backup and restore

          The Report Portal v5 data backup and restore in the operator include two parts, PostgreSQL database data and Elasticsearch index data.

          The Operator support automatically data backup and restore for both.

          ### PostgreSQL database WAL archiving and recovery

          The Operator supports PostgreSQL WAL archiving and recovery point-in-time via Minio S3 using restic.

          #### WAL archiving

          When deploy Report Portal v5 instance, make sure `enable_pg_restic_backup` option is set as `yes`, then the WAL archiving will be enabled when initiate the database, the WAL archive data will be automatically backup to S3 storage via Minio using restic command.

          Parameter `pg_restic_s3_bucket_init` is for init the bucket if not exist, `pg_s3_bucket_name` is the bucket name and `pg_restic_password` is the restic password for encryption.

          #### WAL recovery

          WAL archive recovery will download all WAL archive files to the Postgres container and restore point-in-time.

          To restore PostgreSQL database point-in-time, on Operator UI select `ReportPortalRestore` under Provided APIs to create ReportPortalRestore operand.

          After the ReportPortalRestore operand created, the PG WAL restore will get started, as restore time will base on the database size, the process will last in minutes, hours or days.

          **Note:** Current Postgres container liveless probe timeout is set as one hour in the postgresql statefulset, it might cause restore fail if restore time is longer than one hour, so it might need be adjusted to be longer.

          ### Elasticsearch snapshot and restore

          The operator supports elasticsearch snapshot and restore with data saved in S3 bucket via Minio, the elasticsearch container have installed the s3 repository plugin.

          #### Snapshot

          When deploy Report Portal v5 instance, make sure following elasticsearch parameters are set:

              es_s3_backup_dir: s3_backup
              es_snapshot_bucket: es-snapshot-123123
              es_backup_schedule: '@daily'

          #### Restore

          Elasticsearch restore support restore to certain date as default snapshot schedule is daily.

          To restore elasticsearch from snapshot, on Operator UI select `ReportPortalRestore` under Provided APIs to create ReportPortalRestore operand.

          The backup dir and bucket name should be same as when create the RP instance, `es_restore` need set to `yes` to enable the restore and `es_restore_date` is the restore target date.

          The restore time will also be depend on the snapshot size.

          ## Metrics

          The operator by default have enabled metrics, the default ansible operator port 8080 is used for the operator container.

          All Report Portal v5 depended services, rabbitmq, minio, postgresql and elasticsearch have metrics enabled in this operator.

          The Report Portal gatway service have metrics enabled.

          By default several ServiceMonitors have been defined when create ReportPortal v5 operand:

              $ oc get servicemonitor
              NAME                    AGE
              elasticsearch-metrics   1d21h
              gateway-metrics         1d21h
              minio-metrics           1d21h
              postgres-metrics        1d21h
              rabbitmq-metrics        1d21h

          To export Report Portal v5 service-api metrics with ServiceMonitor, create a `secret` with RP user api token. The api access token could be found with login to Report Portal -> Click user icon -> Profile.

          Then select `ServiceApiServiceMonitor` under provided APIs to Create Instance on operator UI.

          Input the secret name and key name that have been created, and click create, a api-metrics servicemonitor will be created.

          To check the metrics, if [user workload monitoring](https://docs.openshift.com/container-platform/4.5/monitoring/monitoring-your-own-services.html) is enabled on the cluster, click to Monitoring on OpenShift UI and navigate to Metrics, where you could query the metrics.

          You could also deploy Promethues Operator in the current project and query via it.

          ## Doc

          More detail with installation, data backup & restore and metrics could be found at https://github.com/waynesun09/reportportal-operator#reportportal-operator
        displayName: reportportal-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - reportportal
        - failure analysis
        - test automation
        links:
        - name: Documentation
          url: https://github.com/waynesun09/reportportal-operator/blob/master/README.md
        - name: ReportPortal.io
          url: https://reportportal.io
        maintainers:
        - email: gsun@redhat.com
          name: Wayne Sun
        maturity: alpha
        provider:
          name: Red Hat CCIT
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.7.0
        - quay.io/waynesun09/rp5-operator:v0.0.8
        version: 0.0.8
      entries:
      - name: reportportal-operator.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: reportportal-operator
    provider:
      name: Red Hat CCIT
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://www.redhat.com/
    name: analytics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: analytics-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyData",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalydata-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalydata",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalydata-sample"
                },
                "spec": {
                  "anomalyname": "etcd_object_namespaces",
                  "config": {
                    "max": 100,
                    "min": 60,
                    "query": "max(apiserver_storage_objects{resource=~\"namespaces\"}) by (resource)"
                  },
                  "method": "min_max",
                  "metricdata": {
                    "groupeddata": "{ \"resource\": \"namespaces\" }",
                    "latestvalue": 121,
                    "timestamp": 1695804362
                  }
                }
              },
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalyengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalyengine",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalyengine-sample"
                },
                "spec": {
                  "anomalyqueryconfiguration": "etcd_objects_all:\n  method: percentage_change\n  query: max(apiserver_storage_objects{resource!~\"certificatesigningrequests.certificates.k8s.io|installplans.operators.coreos.com|operators.operators.coreos.com|subscriptions.operators.coreos.com|anomalydata.observability-analytics.redhat.com|anomalyengines.observability-analytics.redhat.com\"}) by (resource)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: True\nfailed_pods:\n  method: min_max\n  query: count(kube_pod_status_phase{phase!~\"Succeeded|Pending|Running|Unknown\"}) by (phase, namespace)\n  min: 0\n  max: 10\n  have_multi_result_data: True\nworkload_memory_usage_bytes:\n  method: percentage_change\n  query: sum(container_memory_working_set_bytes{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nworkload_cpu_usage_cores:\n  method: percentage_change\n  query: sum(container_cpu_usage_seconds_total{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nrest_client_requests_total:\n  method: percentage_change\n  query: sum(rest_client_requests_total)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\n",
                  "cronjobconfig": {
                    "anomalyqueries": "etcd_objects_all,failed_pods,workload_memory_usage_bytes,workload_cpu_usage_cores,rest_client_requests_total",
                    "name": "osa-anomaly-detection",
                    "resource": {
                      "cpulimit": "256m",
                      "cpurequest": "128m",
                      "memorylimit": "512Mi",
                      "memoryrequest": "256Mi"
                    },
                    "schedule": "*/5 * * * *"
                  },
                  "namespace": "osa-anomaly-detection",
                  "serviceaccountrolebinding": {
                    "clusterrolename": "cluster-monitoring-view",
                    "name": "osa-in-cluster-thanos-querier"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
          createdAt: "2024-03-12T08:02:25Z"
          description: Deploys and Manages Observability Analytics Operator
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4-alpha
          repository: https://github.com/openshift/analytics-operator
          support: RedHat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnomalyData is the Schema for the anomalydata API
            displayName: Anomaly Data
            kind: AnomalyData
            name: anomalydata.observability-analytics.redhat.com
            version: v1alpha1
          - description: AnomalyEngine is the Schema for the anomalyengines API
            displayName: Anomaly Engine
            kind: AnomalyEngine
            name: anomalyengines.observability-analytics.redhat.com
            version: v1alpha1
        description: "### Introduction\nObservability Analytics for Red Hat OpenShift
          runs OpenShift analytics processes like anomaly detection.\n\n### API Backward
          Compatibility\n**NOTE:** Until the Operator supports **seamless upgrades**,
          a new version of the operator may introduce a change that is **NOT** backward-compatible
          with the previous version. Thus, to upgrade the operator, uninstall the
          already installed version completely (including CRDs) and install the new
          version.\n\n### Installtion Steps for the Anomaly Engine\n Create namespace
          if you would like to install this operator into separate namespace (You
          can utilize existing namespace as well). ```bash kubectl create ns <Namespace>
          ``` Install the Observability Analytics operator to specific `<Namespace>`,
          wait until you see the operator in the **Installed Operator** list. Click
          on the Operator name and go to Operator details.\n\nHere you should see
          a tab called **Anomaly Engine**. Click on it, and then click on the **Create
          AnomalyEngine** button, which will redirect you to a form where you need
          to provide the required inputs for the Anomaly Engine. Once successfully
          created, you should see a cronjob in the given namespace. This cronjob will
          identify anomalies and store them into **anomalydata**.\n\n### How to See
          Anomaly Data Captured by Anomaly Engine?\nWe are storing anomaly data into
          CR.\n ####Option 1:Through `kubectl`` command\nYou can check detected anomalies
          using the following command from your terminal. \n\n```bash\nkubectl get
          anomalydata -n <Namespace>\nkubectl describe anomalydata <NAME of the anomalydata>
          -n <Namespace> \n```\n #### Option 2: From Installed Operator UI Go to installed
          operator and select **Observability Analytics** operator, now go to **Anomaly
          Data** tab where you can able to see the detected anomaly, click on specific
          anomaly and you should be able to see details of the same. ### Documentation\nDocumentation
          around how to test Operator codebase as a developer can be found [here](https://github.com/openshift/analytics-operator#readme)\n\n###
          Contributing\nYou can contribute by:\n  * Raising [issues](https://github.com/openshift/analytics-operator/issues)\n
          \   related to observability-analytics-operator\n  * Fixing issues by opening
          [Pull Requests](https://github.com/openshift/analytics-operator/pulls)\n
          \ * Discussing observability-analytics-operator\n\n### License\nobservability-analytics-operator
          is licensed under the Apache 2.0 license.\n"
        displayName: Observability Analytics
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - observability-analytics-operator
        - analytics-operator
        - analytics
        - anomaly
        links:
        - name: Analytics Operator
          url: https://analytics-operator.domain
        maintainers:
        - email: rzalavad@redhat.com
          name: Raj
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
        version: 0.1.1
      entries:
      - name: analytics-operator.v0.1.1
        version: 0.1.1
      - name: analytics-operator.v0.1.0
        version: 0.1.0
      name: alpha
    - currentCSV: analytics-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyData",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalydata-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalydata",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalydata-sample"
                },
                "spec": {
                  "anomalyname": "etcd_object_namespaces",
                  "config": {
                    "max": 100,
                    "min": 60,
                    "query": "max(apiserver_storage_objects{resource=~\"namespaces\"}) by (resource)"
                  },
                  "method": "min_max",
                  "metricdata": {
                    "groupeddata": "{ \"resource\": \"namespaces\" }",
                    "latestvalue": 121,
                    "timestamp": 1695804362
                  }
                }
              },
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalyengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalyengine",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalyengine-sample"
                },
                "spec": {
                  "anomalyqueryconfiguration": "etcd_objects_all:\n  method: percentage_change\n  query: max(apiserver_storage_objects{resource!~\"certificatesigningrequests.certificates.k8s.io|installplans.operators.coreos.com|operators.operators.coreos.com|subscriptions.operators.coreos.com|anomalydata.observability-analytics.redhat.com|anomalyengines.observability-analytics.redhat.com\"}) by (resource)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: True\nfailed_pods:\n  method: min_max\n  query: count(kube_pod_status_phase{phase!~\"Succeeded|Pending|Running|Unknown\"}) by (phase, namespace)\n  min: 0\n  max: 10\n  have_multi_result_data: True\nworkload_memory_usage_bytes:\n  method: percentage_change\n  query: sum(container_memory_working_set_bytes{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nworkload_cpu_usage_cores:\n  method: percentage_change\n  query: sum(container_cpu_usage_seconds_total{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nrest_client_requests_total:\n  method: percentage_change\n  query: sum(rest_client_requests_total)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\n",
                  "cronjobconfig": {
                    "anomalyqueries": "etcd_objects_all,failed_pods,workload_memory_usage_bytes,workload_cpu_usage_cores,rest_client_requests_total",
                    "name": "osa-anomaly-detection",
                    "resource": {
                      "cpulimit": "256m",
                      "cpurequest": "128m",
                      "memorylimit": "512Mi",
                      "memoryrequest": "256Mi"
                    },
                    "schedule": "*/5 * * * *"
                  },
                  "namespace": "osa-anomaly-detection",
                  "serviceaccountrolebinding": {
                    "clusterrolename": "cluster-monitoring-view",
                    "name": "osa-in-cluster-thanos-querier"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
          createdAt: "2024-03-12T08:02:25Z"
          description: Deploys and Manages Observability Analytics Operator
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4-alpha
          repository: https://github.com/openshift/analytics-operator
          support: RedHat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnomalyData is the Schema for the anomalydata API
            displayName: Anomaly Data
            kind: AnomalyData
            name: anomalydata.observability-analytics.redhat.com
            version: v1alpha1
          - description: AnomalyEngine is the Schema for the anomalyengines API
            displayName: Anomaly Engine
            kind: AnomalyEngine
            name: anomalyengines.observability-analytics.redhat.com
            version: v1alpha1
        description: "### Introduction\nObservability Analytics for Red Hat OpenShift
          runs OpenShift analytics processes like anomaly detection.\n\n### API Backward
          Compatibility\n**NOTE:** Until the Operator supports **seamless upgrades**,
          a new version of the operator may introduce a change that is **NOT** backward-compatible
          with the previous version. Thus, to upgrade the operator, uninstall the
          already installed version completely (including CRDs) and install the new
          version.\n\n### Installtion Steps for the Anomaly Engine\n Create namespace
          if you would like to install this operator into separate namespace (You
          can utilize existing namespace as well). ```bash kubectl create ns <Namespace>
          ``` Install the Observability Analytics operator to specific `<Namespace>`,
          wait until you see the operator in the **Installed Operator** list. Click
          on the Operator name and go to Operator details.\n\nHere you should see
          a tab called **Anomaly Engine**. Click on it, and then click on the **Create
          AnomalyEngine** button, which will redirect you to a form where you need
          to provide the required inputs for the Anomaly Engine. Once successfully
          created, you should see a cronjob in the given namespace. This cronjob will
          identify anomalies and store them into **anomalydata**.\n\n### How to See
          Anomaly Data Captured by Anomaly Engine?\nWe are storing anomaly data into
          CR.\n ####Option 1:Through `kubectl`` command\nYou can check detected anomalies
          using the following command from your terminal. \n\n```bash\nkubectl get
          anomalydata -n <Namespace>\nkubectl describe anomalydata <NAME of the anomalydata>
          -n <Namespace> \n```\n #### Option 2: From Installed Operator UI Go to installed
          operator and select **Observability Analytics** operator, now go to **Anomaly
          Data** tab where you can able to see the detected anomaly, click on specific
          anomaly and you should be able to see details of the same. ### Documentation\nDocumentation
          around how to test Operator codebase as a developer can be found [here](https://github.com/openshift/analytics-operator#readme)\n\n###
          Contributing\nYou can contribute by:\n  * Raising [issues](https://github.com/openshift/analytics-operator/issues)\n
          \   related to observability-analytics-operator\n  * Fixing issues by opening
          [Pull Requests](https://github.com/openshift/analytics-operator/pulls)\n
          \ * Discussing observability-analytics-operator\n\n### License\nobservability-analytics-operator
          is licensed under the Apache 2.0 license.\n"
        displayName: Observability Analytics
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - observability-analytics-operator
        - analytics-operator
        - analytics
        - anomaly
        links:
        - name: Analytics Operator
          url: https://analytics-operator.domain
        maintainers:
        - email: rzalavad@redhat.com
          name: Raj
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
        version: 0.1.1
      entries:
      - name: analytics-operator.v0.1.1
        version: 0.1.1
      - name: analytics-operator.v0.1.0
        version: 0.1.0
      name: dev-preview
    defaultChannel: alpha
    packageName: analytics-operator
    provider:
      name: Red Hat Inc.
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-atlas-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-atlas-kubernetes.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupCompliancePolicy",
                "metadata": {
                  "name": "my-backup-compliance-policy"
                },
                "spec": {
                  "authorizedEmail": "john.doe@example.com",
                  "authorizedUserFirstName": "John",
                  "authorizedUserLastName": "Doe",
                  "copyProtectionEnabled": false,
                  "encryptionAtRestEnabled": false,
                  "onDemandPolicy": {
                    "retentionUnit": "weeks",
                    "retentionValue": 3
                  },
                  "overwriteBackupPolicies": false,
                  "pointInTimeEnabled": true,
                  "restoreWindowDays": 42,
                  "scheduledPolicyItems": [
                    {
                      "frequencyInterval": 2,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "hourly",
                      "retentionUnit": "days",
                      "retentionValue": 2
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "weekly",
                      "retentionUnit": "weeks",
                      "retentionValue": 4
                    },
                    {
                      "frequencyInterval": 15,
                      "frequencyType": "monthly",
                      "retentionUnit": "months",
                      "retentionValue": 12
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": false,
                  "copySettings": [
                    {
                      "cloudProvider": "AWS",
                      "frequencies": [
                        "WEEKLY"
                      ],
                      "regionName": "US_WEST_1",
                      "shouldCopyOplogs": false
                    }
                  ],
                  "policy": {
                    "name": "atlasbackuppolicy-sample"
                  },
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2,
                  "useOrgAndGroupNamesInExportPrefix": true
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasCustomRole",
                "metadata": {
                  "name": "shard-operator-role"
                },
                "spec": {
                  "projectRef": {
                    "name": "my-project"
                  },
                  "role": {
                    "actions": [
                      {
                        "name": "getShardMap",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "shardingState",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "connPoolStats",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "getLog",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      }
                    ],
                    "inheritedRoles": [
                      {
                        "database": "admin",
                        "name": "operator-role-1"
                      }
                    ],
                    "name": "my-role"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDataFederation",
                "metadata": {
                  "name": "my-data-federation",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "name": "my-data-federation",
                  "privateEndpoints": [
                    {
                      "endpointId": "vpce-03f9eeaa764e32454",
                      "provider": "AWS",
                      "type": "DATA_LAKE"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project",
                    "namespace": "mongodb-atlas-system"
                  },
                  "storage": {
                    "databases": [
                      {
                        "collections": [
                          {
                            "dataSources": [
                              {
                                "storeName": "http-test",
                                "urls": [
                                  "https://data.cityofnewyork.us/api/views/vfnx-vebw/rows.csv"
                                ]
                              }
                            ],
                            "name": "test-collection-1"
                          }
                        ],
                        "name": "test-db-1"
                      }
                    ],
                    "stores": [
                      {
                        "name": "http-test",
                        "provider": "http"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "the-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "theuser"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-deployment"
                },
                "spec": {
                  "backupRef": {
                    "name": "atlasbackupschedule-sample"
                  },
                  "deploymentSpec": {
                    "backupEnabled": true,
                    "clusterType": "REPLICASET",
                    "mongoDBMajorVersion": "7.0",
                    "name": "test-deployment",
                    "replicationSpecs": [
                      {
                        "regionConfigs": [
                          {
                            "electableSpecs": {
                              "instanceSize": "M10",
                              "nodeCount": 3
                            },
                            "priority": 7,
                            "providerName": "AWS",
                            "regionName": "US_EAST_1"
                          }
                        ],
                        "zoneName": "Zone 1"
                      }
                    ],
                    "tags": [
                      {
                        "key": "team",
                        "value": "atlas-operator"
                      }
                    ]
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasFederatedAuth",
                "metadata": {
                  "name": "atlas-default-federated-auth",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "connectionSecretRef": {
                    "name": "my-org-secret",
                    "namespace": "mongodb-atlas-system"
                  },
                  "domainAllowList": [
                    "my-org-domain.com"
                  ],
                  "domainRestrictionEnabled": true,
                  "enabled": true,
                  "postAuthRoleGrants": [
                    "ORG_MEMBER"
                  ],
                  "roleMappings": [
                    {
                      "externalGroupName": "org-admin",
                      "roleAssignments": [
                        {
                          "role": "ORG_OWNER"
                        }
                      ]
                    },
                    {
                      "externalGroupName": "dev-team",
                      "roleAssignments": [
                        {
                          "role": "ORG_GROUP_CREATOR"
                        },
                        {
                          "projectName": "dev-project",
                          "role": "GROUP_OWNER"
                        }
                      ]
                    }
                  ],
                  "ssoDebugEnabled": false
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasIPAccessList",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasipaccesslist-sample"
                },
                "spec": {
                  "entries": [
                    {
                      "cidrBlock": "192.168.1.0/24"
                    },
                    {
                      "deleteAfterDate": "2025-01-20T23:59:59+02:00",
                      "ipAddress": "10.0.0.1"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkContainer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasnetworkcontainer-sample"
                },
                "spec": {
                  "cidrBlock": "10.11.0.0/16",
                  "connectionSecret": {
                    "name": "atlas-connection-secret"
                  },
                  "externalProjectRef": {
                    "projectId": "66e2f2b621571b7e69a89b66"
                  },
                  "provider": "AWS",
                  "region": "US_EAST_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkPeering",
                "metadata": {
                  "name": "atlasnetworkpeering-sample"
                },
                "spec": {
                  "atlasCidrBlock": "10.12.0.0/16",
                  "awsConfiguration": {
                    "accepterRegionName": "US_EAST_1",
                    "awsAccountId": "23214235145",
                    "routeTableCidrBlock": "10.11.0.0/16",
                    "vpcId": "vpcid-7123782134"
                  },
                  "containerId": "623412394512350",
                  "containerRegion": "US_EAST_1",
                  "projectRef": {
                    "name": "atlas-project",
                    "namespace": "namespace"
                  },
                  "provider": "AWS"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasPrivateEndpoint",
                "metadata": {
                  "name": "atlasprivateendpoint-sample"
                },
                "spec": {
                  "awsConfiguration": {
                    "id": "vpce-f4k34w51d"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "provider": "AWS",
                  "region": "EU_CENTRAL_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    },
                    {
                      "cidrBlock": "203.0.113.0/24",
                      "comment": "CIDR block for Application Server B - D"
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasSearchIndexConfig",
                "metadata": {
                  "name": "atlassearchindexconfig-sample"
                },
                "spec": {
                  "analyzer": "lucene.standard",
                  "analyzers": [
                    {
                      "name": "test-analyzer",
                      "tokenFilters": [
                        {
                          "asciiFolding": {
                            "originalTokens": "omit"
                          },
                          "type": "asciiFolding"
                        }
                      ],
                      "tokenizer": {
                        "type": "whiteSpace",
                        "whitespace": {
                          "maxTokenLenght": 200
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamConnection",
                "metadata": {
                  "name": "my-streamconnection-kafka"
                },
                "spec": {
                  "kafkaConfig": {
                    "authentication": {
                      "credentials": {
                        "name": "kafka-userpass"
                      },
                      "mechanism": "SCRAM-512"
                    },
                    "bootstrapServers": "kafka.server1:9001,kafka.server2:9002,kafka.server3:9003",
                    "security": {
                      "certificate": {
                        "name": "kafka-cert"
                      },
                      "protocol": "SSL"
                    }
                  },
                  "name": "kafka-config",
                  "type": "Kafka"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamInstance",
                "metadata": {
                  "name": "my-streaminstance-sample"
                },
                "spec": {
                  "clusterConfig": {
                    "provider": "AWS",
                    "region": "VIRGINIA_USA",
                    "tier": "SP30"
                  },
                  "connectionRegistry": [
                    {
                      "name": "my-streamconnection-sample"
                    }
                  ],
                  "name": "my-instance",
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasTeam",
                "metadata": {
                  "name": "atlasteam-sample"
                },
                "spec": {
                  "name": "my-team",
                  "usernames": [
                    "username@mongodb.com"
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasThirdPartyIntegration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasintegration-sample"
                },
                "spec": {
                  "datadog": {
                    "apiKeySecretRef": {
                      "name": "secret-ref"
                    },
                    "region": "string",
                    "sendCollectionLatencyMetrics": "string",
                    "sendDatabaseMetrics": "string"
                  },
                  "projectRef": {
                    "name": "my-atlas-project"
                  },
                  "type": "DATADOG"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/mongodb/mongodb-atlas-kubernetes-operator:2.11.0
          createdAt: "2025-09-25T14:34:34Z"
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/mongodb/mongodb-atlas-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The AtlasBackupCompliancePolicy is a configuration that enforces
              specific backup and retention requirements
            displayName: Atlas Backup Compliance Policy
            kind: AtlasBackupCompliancePolicy
            name: atlasbackupcompliancepolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasCustomRole is the Schema for the AtlasCustomRole API
            displayName: Atlas Custom Role
            kind: AtlasCustomRole
            name: atlascustomroles.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDataFederation is the Schema for the Atlas Data Federation
              API
            displayName: Atlas Data Federation
            kind: AtlasDataFederation
            name: atlasdatafederations.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasdeployments API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasFederatedAuth is the Schema for the Atlasfederatedauth
              API
            displayName: Atlas Federated Auth
            kind: AtlasFederatedAuth
            name: atlasfederatedauths.atlas.mongodb.com
            version: v1
          - description: AtlasIPAccessList is the Schema for the atlasipaccesslists
              API.
            displayName: Atlas IPAccess List
            kind: AtlasIPAccessList
            name: atlasipaccesslists.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkContainer is the Schema for the AtlasNetworkContainer
              API
            displayName: Atlas Network Container
            kind: AtlasNetworkContainer
            name: atlasnetworkcontainers.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkPeering is the Schema for the AtlasNetworkPeering
              API
            displayName: Atlas Network Peering
            kind: AtlasNetworkPeering
            name: atlasnetworkpeerings.atlas.mongodb.com
            version: v1
          - displayName: Atlas Org Settings
            kind: AtlasOrgSettings
            name: atlasorgsettings.atlas.mongodb.com
            version: v1
          - description: "The AtlasPrivateEndpoint custom resource definition (CRD)
              defines a desired [Private Endpoint](https://www.mongodb.com/docs/atlas/security-private-endpoint/#std-label-private-endpoint-overview)
              configuration for an Atlas project. It allows a private connection between
              your cloud provider and Atlas that doesn't send information through
              a public network. \n You can use private endpoints to create a unidirectional
              connection to Atlas clusters from your virtual network."
            displayName: Atlas Private Endpoint
            kind: AtlasPrivateEndpoint
            name: atlasprivateendpoints.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
          - description: Atlas Search Index Config is the Schema for the Atlas Search
              API
            displayName: Atlas Search Index Config
            kind: AtlasSearchIndexConfig
            name: atlassearchindexconfigs.atlas.mongodb.com
            version: v1
          - description: AtlasStreamConnection is the Schema for the atlasstreamconnections
              API
            displayName: Atlas Stream Connection
            kind: AtlasStreamConnection
            name: atlasstreamconnections.atlas.mongodb.com
            version: v1
          - description: AtlasStreamInstance is the Schema for the atlasstreaminstances
              API
            displayName: Atlas Stream Instance
            kind: AtlasStreamInstance
            name: atlasstreaminstances.atlas.mongodb.com
            version: v1
          - description: AtlasTeam is the Schema for the Atlas Teams API
            displayName: Atlas Team
            kind: AtlasTeam
            name: atlasteams.atlas.mongodb.com
            version: v1
          - description: AtlasThirdPartyIntegration is the Schema for the atlas 3rd
              party integrations API.
            displayName: Atlas Third Party Integration
            kind: AtlasThirdPartyIntegration
            name: atlasthirdpartyintegrations.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB
          Atlas — the only multi-cloud document database service that gives you the versatility you need to build sophisticated
          and resilient applications that can adapt to changing customer demands and market trends.

          The full documentation for the Operator can be found [here](https://docs.atlas.mongodb.com/atlas-operator/)

          ## Getting Started

          ### Supported features

          * Create and configure a project, or connect to an existing one.
          * Deploy, manage, scale, and tear down clusters.
          * Support for serverless instances.
          * Create and edit database users.
          * Manage IP Access Lists, network peering and private endpoints.
          * Configure and control Atlas’s fully managed cloud backup.
          * Configure federated authentication for your Atlas organization
          * Integrate Atlas monitoring with Prometheus.

          ... and more.

          To view the list of custom resources and each CR's schema, visit our [reference](https://www.mongodb.com/docs/atlas/operator/stable/custom-resources/)
          documentation. See the [Quickstart](https://www.mongodb.com/docs/atlas/operator/stable/ak8so-quick-start/) to get
          started with Atlas Kubernetes Operator.
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator:2.11.0
        version: 2.11.0
      entries:
      - name: mongodb-atlas-kubernetes.v2.11.0
        version: 2.11.0
      - name: mongodb-atlas-kubernetes.v2.10.0
        version: 2.10.0
      - name: mongodb-atlas-kubernetes.v2.9.1
        version: 2.9.1
      - name: mongodb-atlas-kubernetes.v2.9.0
        version: 2.9.0
      - name: mongodb-atlas-kubernetes.v2.8.2
        version: 2.8.2
      - name: mongodb-atlas-kubernetes.v2.8.1
        version: 2.8.1
      - name: mongodb-atlas-kubernetes.v2.8.0
        version: 2.8.0
      - name: mongodb-atlas-kubernetes.v2.7.1
        version: 2.7.1
      - name: mongodb-atlas-kubernetes.v2.7.0
        version: 2.7.0
      - name: mongodb-atlas-kubernetes.v2.6.1
        version: 2.6.1
      - name: mongodb-atlas-kubernetes.v2.6.0
        version: 2.6.0
      - name: mongodb-atlas-kubernetes.v2.5.0
        version: 2.5.0
      - name: mongodb-atlas-kubernetes.v2.4.1
        version: 2.4.1
      - name: mongodb-atlas-kubernetes.v2.4.0
        version: 2.4.0
      - name: mongodb-atlas-kubernetes.v2.3.1
        version: 2.3.1
      - name: mongodb-atlas-kubernetes.v2.3.0
        version: 2.3.0
      - name: mongodb-atlas-kubernetes.v2.2.2
        version: 2.2.2
      - name: mongodb-atlas-kubernetes.v2.2.1
        version: 2.2.1
      - name: mongodb-atlas-kubernetes.v2.2.0
        version: 2.2.0
      - name: mongodb-atlas-kubernetes.v2.1.0
        version: 2.1.0
      - name: mongodb-atlas-kubernetes.v2.0.1
        version: 2.0.1
      - name: mongodb-atlas-kubernetes.v2.0.0
        version: 2.0.0
      - name: mongodb-atlas-kubernetes.v1.9.1
        version: 1.9.1
      - name: mongodb-atlas-kubernetes.v1.9.0
        version: 1.9.0
      - name: mongodb-atlas-kubernetes.v1.8.2
        version: 1.8.2
      - name: mongodb-atlas-kubernetes.v1.8.1
        version: 1.8.1
      - name: mongodb-atlas-kubernetes.v1.8.0
        version: 1.8.0
      - name: mongodb-atlas-kubernetes.v1.7.3
        version: 1.7.3
      - name: mongodb-atlas-kubernetes.v1.7.2
        version: 1.7.2
      - name: mongodb-atlas-kubernetes.v1.7.1
        version: 1.7.1
      - name: mongodb-atlas-kubernetes.v1.7.0
        version: 1.7.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-atlas-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OpenTelemetry Community
      provider-url: ""
    name: opentelemetry-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opentelemetry-operator.v0.136.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "Instrumentation",
                "metadata": {
                  "name": "instrumentation"
                },
                "spec": {
                  "exporter": {
                    "endpoint": "http://otel-collector-headless:4317"
                  }
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpAMPBridge",
                "metadata": {
                  "name": "opampbridge-sample"
                },
                "spec": {
                  "capabilities": {
                    "AcceptsOpAMPConnectionSettings": true,
                    "AcceptsOtherConnectionSettings": true,
                    "AcceptsRemoteConfig": true,
                    "AcceptsRestartCommand": true,
                    "ReportsEffectiveConfig": true,
                    "ReportsHealth": true,
                    "ReportsOwnLogs": true,
                    "ReportsOwnMetrics": true,
                    "ReportsOwnTraces": true,
                    "ReportsRemoteConfig": true,
                    "ReportsStatus": true
                  },
                  "componentsAllowed": {
                    "exporters": [
                      "debug"
                    ],
                    "processors": [
                      "memory_limiter"
                    ],
                    "receivers": [
                      "otlp"
                    ]
                  },
                  "endpoint": "ws://opamp-server:4320/v1/opamp"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": "receivers:\n  otlp:\n    protocols: \n      grpc: {}\n      http: {}\n\nexporters:\n  debug: {}\n\nservice:\n  pipelines:\n    traces:\n      receivers: [otlp]\n      exporters: [debug]\n"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1beta1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": {
                    "exporters": {
                      "debug": {}
                    },
                    "receivers": {
                      "otlp": {
                        "protocols": {
                          "grpc": {},
                          "http": {}
                        }
                      }
                    },
                    "service": {
                      "pipelines": {
                        "traces": {
                          "exporters": [
                            "debug"
                          ],
                          "receivers": [
                            "otlp"
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          certified: "false"
          containerImage: ghcr.io/open-telemetry/opentelemetry-operator/opentelemetry-operator
          createdAt: "2025-09-29T13:00:19Z"
          description: Provides the OpenTelemetry components, including the Collector
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/open-telemetry/opentelemetry-operator
          support: OpenTelemetry Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instrumentation is the spec for OpenTelemetry instrumentation.
            displayName: OpenTelemetry Instrumentation
            kind: Instrumentation
            name: instrumentations.opentelemetry.io
            version: v1alpha1
          - description: OpAMPBridge is the Schema for the opampbridges API.
            displayName: OpAMP Bridge
            kind: OpAMPBridge
            name: opampbridges.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1beta1
          - description: TargetAllocator is the Schema for the targetallocators API.
            displayName: Target Allocator
            kind: TargetAllocator
            name: targetallocators.opentelemetry.io
            version: v1alpha1
        description: |-
          OpenTelemetry is a collection of tools, APIs, and SDKs. You use it to instrument, generate, collect, and export telemetry data (metrics, logs, and traces) for analysis in order to understand your software's performance and behavior.

          ### Operator features

          * **Sidecar injection** - annotate your pods and let the operator inject a sidecar.
          * **Managed upgrades** - updating the operator will automatically update your OpenTelemetry collectors.
          * **Deployment modes** - your collector can be deployed as sidecar, daemon set, or regular deployment.
          * **Service port management** - the operator detects which ports need to be exposed based on the provided configuration.
        displayName: Community OpenTelemetry Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opentelemetry
        - tracing
        - logging
        - metrics
        - monitoring
        - troubleshooting
        links:
        - name: OpenTelemetry Operator
          url: https://github.com/open-telemetry/opentelemetry-operator
        maintainers:
        - email: jpkroehling@redhat.com
          name: Juraci Paixão Kröhling
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: OpenTelemetry Community
        relatedImages:
        - ghcr.io/open-telemetry/opentelemetry-operator/opentelemetry-operator:0.136.0
        - quay.io/brancz/kube-rbac-proxy:v0.13.1
        version: 0.136.0
      entries:
      - name: opentelemetry-operator.v0.136.0
        version: 0.136.0
      - name: opentelemetry-operator.v0.135.0
        version: 0.135.0
      - name: opentelemetry-operator.v0.134.0
        version: 0.134.0
      - name: opentelemetry-operator.v0.132.0
        version: 0.132.0
      - name: opentelemetry-operator.v0.131.0
        version: 0.131.0
      - name: opentelemetry-operator.v0.129.1
        version: 0.129.1
      - name: opentelemetry-operator.v0.127.0
        version: 0.127.0
      - name: opentelemetry-operator.v0.126.0
        version: 0.126.0
      - name: opentelemetry-operator.v0.125.0
        version: 0.125.0
      - name: opentelemetry-operator.v0.124.0
        version: 0.124.0
      - name: opentelemetry-operator.v0.123.0
        version: 0.123.0
      - name: opentelemetry-operator.v0.122.0
        version: 0.122.0
      - name: opentelemetry-operator.v0.121.0
        version: 0.121.0
      - name: opentelemetry-operator.v0.120.0
        version: 0.120.0
      - name: opentelemetry-operator.v0.119.0
        version: 0.119.0
      - name: opentelemetry-operator.v0.118.0
        version: 0.118.0
      - name: opentelemetry-operator.v0.117.0
        version: 0.117.0
      - name: opentelemetry-operator.v0.116.0
        version: 0.116.0
      - name: opentelemetry-operator.v0.115.0
        version: 0.115.0
      - name: opentelemetry-operator.v0.114.1
        version: 0.114.1
      - name: opentelemetry-operator.v0.112.0
        version: 0.112.0
      - name: opentelemetry-operator.v0.111.0
        version: 0.111.0
      - name: opentelemetry-operator.v0.110.0
        version: 0.110.0
      - name: opentelemetry-operator.v0.109.0
        version: 0.109.0
      - name: opentelemetry-operator.v0.108.0
        version: 0.108.0
      - name: opentelemetry-operator.v0.107.0
        version: 0.107.0
      - name: opentelemetry-operator.v0.106.0
        version: 0.106.0
      - name: opentelemetry-operator.v0.105.0
        version: 0.105.0
      name: alpha
    defaultChannel: alpha
    packageName: opentelemetry-operator
    provider:
      name: OpenTelemetry Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: StreamNative
      provider-url: https://streamnative.io
    name: sn-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc:latest",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway:0.10.0"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator:v0.10.0
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator:v0.10.0
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      - name: sn-operator.v0.9.0
        version: 0.9.0
      - name: sn-operator.v0.8.5
        version: 0.8.5
      - name: sn-operator.v0.7.10
        version: 0.7.10
      name: alpha
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc:latest",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway:0.10.0"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator:v0.10.0
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator:v0.10.0
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      - name: sn-operator.v0.9.0
        version: 0.9.0
      - name: sn-operator.v0.8.5
        version: 0.8.5
      - name: sn-operator.v0.7.10
        version: 0.7.10
      name: beta
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc:latest",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway:0.10.0"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator:v0.10.0
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator:v0.10.0
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      - name: sn-operator.v0.9.0
        version: 0.9.0
      - name: sn-operator.v0.8.5
        version: 0.8.5
      - name: sn-operator.v0.7.10
        version: 0.7.10
      name: stable
    defaultChannel: alpha
    packageName: sn-operator
    provider:
      name: StreamNative
      url: https://streamnative.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Instana
      provider-url: ""
    name: instana-agent-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: instana-agent-operator.v2.1.35
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instana.io/v1",
                "kind": "InstanaAgent",
                "metadata": {
                  "name": "instana-agent",
                  "namespace": "instana-agent"
                },
                "spec": {
                  "agent": {
                    "configuration_yaml": "# You can leave this empty, or use this to configure your instana agent.\n# See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.\n",
                    "endpointHost": "ingress-red-saas.instana.io",
                    "endpointPort": "443",
                    "env": {},
                    "key": "replace-key"
                  },
                  "cluster": {
                    "name": "my-cluster"
                  },
                  "opentelemetry": {
                    "grpc": {
                      "enabled": true
                    },
                    "http": {
                      "enabled": true
                    }
                  },
                  "zone": {
                    "name": "edited-zone"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring,Logging & Tracing,OpenShift Optional
          certified: "false"
          containerImage: icr.io/instana/instana-agent-operator@sha256:fe949ae0ca9cc05ffe3569cd7a6f4fde48f5e3ee64f27e20e0b3102733a6e809
          description: Fully automated Application Performance Monitoring (APM) for
            microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/instana/instana-agent-operator
          support: Instana
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InstanaAgent is the Schema for the agents API
            displayName: Instana Agent
            kind: InstanaAgent
            name: agents.instana.io
            version: v1
          - description: InstanaAgentRemote is the Schema for the agents API
            displayName: Instana Agent Remote
            kind: InstanaAgentRemote
            name: agentsremote.instana.io
            version: v1
        description: |
          # Instana

          Instana is an [APM solution](https://www.instana.com/product-overview/) built for microservices that enables IT Ops to build applications faster and deliver higher quality services by automating monitoring, tracing and root cause analysis. The solution is optimized for [Kubernetes](https://www.instana.com/automatic-kubernetes-monitoring/) and [OpenShift](https://www.instana.com/blog/automatic-root-cause-analysis-for-openshift-applications/).

          ## Instana Agent Operator

          This is the Kubernetes Operator for installing the Instana Agent on Kubernetes or OpenShift.

          ## Prerequisites for OpenShift

          Before the agent will be able to run in OpenShift, you need to perform a couple of extra configuration steps.

          You need to set up a project for the Instana Agent and configure it's permissions.

          The project you create here needs to be the namespace where you create the Instana Agent custom resource that the operator will use to deploy the agent.

          For example, create the `instana-agent` project:

              oc new-project instana-agent

          Then, ensure the `instana-agent` service account is in the privileged security context:

              oc adm policy add-scc-to-user privileged -z instana-agent

          The `instana-agent-remote` service account must be allowed to use any userid:

              oc adm policy add-scc-to-user anyuid -z instana-agent-remote -n instana-agent

          This service account will be created by the operator.

          Now you can proceed with installing the operator for the Instana agent.

          ## Installation and Configuration

          First, install this operator from [OperatorHub.io](https://operatorhub.io/), [OpenShift Container Platform](https://www.openshift.com/), or [OKD](https://www.okd.io/).

          Second, create a custom resource with the agent configuration in the target namespace (for now, this MUST always be the `instana-agent` namespace). The operator will pick up the custom resource and install the Instana agent accordingly.

          The following is a minimal template of the custom resource:

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgent
          metadata:
            name: instana-agent
            namespace: instana-agent
          spec:
            zone:
              name: my-zone # (optional) name of the zone of the host
            cluster:
              name: replace-me # replace with the name of your Kubernetes cluster
            agent:
              key: replace-me # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io # the monitoring ingress endpoint
              endpointPort: "443" # the monitoring ingress endpoint port, wrapped in quotes
              env:
                INSTANA_AGENT_TAGS: example
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://ibm.biz/monitoring-k8s
          ```

          Save the template in a file `instana-agent.yaml` and edit the following values:

          * If your target namespace is not `instana-agent`, replace the `namespace:` accordingly.
          * `agent.key` must be set with your Instana agent key.
          * `agent.endpointHost` must be set with the monitoring ingress endpoint, generally either `saas-us-west-2.instana.io` or `saas-eu-west-1.instana.io`.
          * `agent.endpointPort` must be set with the monitoring ingress port, generally "443" (wrapped in quotes).
          * `zone.name` should be set with the name of the Kubernetes cluster that is be displayed in Instana.

          For advanced configuration, you can edit the contents of the `configuration.yaml` file. View documentation [here](https://ibm.biz/monitoring-k8s).

          Apply the custom resource with `kubectl apply -f instana-agent.yaml`. After some time, you should see `instana-agent` Pods being created on each node of your cluster, and your cluster should show on the infrastructure map on your Instana Web interface.

          In addition to the InstanaAgent custom resource, a remote agent can be deployed. This agent will not monitor workload running inside the cluster, but can deploy a single replica of an agent which can get an independent configuration for running remote monitoring for certain sensors.

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgentRemote
          metadata:
            name: remote-agent
            namespace: "instana-agent"
          spec:
            zone:
              name: edited-zone # name of the zone of the host
            hostname:
              name: use-custom-hostname-for-remote-agent # (optional) give your remote agent pod a custom hostname
            agent:
              key: replace-key # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io
              endpointPort: "443"
              env: {}
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.
          ```

          ## Uninstalling

          In order to uninstall the Instana agent, simply remove the custom resource with `kubectl delete -f instana-agent.yaml`.

          ## Source Code

          The Instana agent operator is an open source project hosted on [https://github.com/instana/instana-agent-operator](https://github.com/instana/instana-agent-operator/).
        displayName: Instana Agent Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - APM
        - Instana
        links:
        - name: GitHub Repository
          url: https://github.com/instana/instana-agent-operator
        maintainers:
        - email: support@instana.com
          name: Instana
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: Instana
        relatedImages:
        - icr.io/instana/instana-agent-operator@sha256:fe949ae0ca9cc05ffe3569cd7a6f4fde48f5e3ee64f27e20e0b3102733a6e809
        - icr.io/instana/agent@sha256:ed71e6c3e5bdb155853a2db8c916460e6a867fb02627e3a332186188cbbd98dd
        version: 2.1.35
      entries:
      - name: instana-agent-operator.v2.1.35
        version: 2.1.35
      - name: instana-agent-operator.v2.1.34
        version: 2.1.34
      - name: instana-agent-operator.v2.1.33
        version: 2.1.33
      - name: instana-agent-operator.v2.1.32
        version: 2.1.32
      - name: instana-agent-operator.v2.1.31
        version: 2.1.31
      - name: instana-agent-operator.v2.1.30
        version: 2.1.30
      - name: instana-agent-operator.v2.1.29
        version: 2.1.29
      - name: instana-agent-operator.v2.1.28
        version: 2.1.28
      - name: instana-agent-operator.v2.1.27
        version: 2.1.27
      - name: instana-agent-operator.v2.1.25
        version: 2.1.25
      - name: instana-agent-operator.v2.1.24
        version: 2.1.24
      - name: instana-agent-operator.v2.1.23
        version: 2.1.23
      - name: instana-agent-operator.v2.1.22
        version: 2.1.22
      - name: instana-agent-operator.v2.1.21
        version: 2.1.21
      - name: instana-agent-operator.v2.1.20
        version: 2.1.20
      - name: instana-agent-operator.v2.1.19
        version: 2.1.19
      - name: instana-agent-operator.v2.1.18
        version: 2.1.18
      - name: instana-agent-operator.v2.1.17
        version: 2.1.17
      - name: instana-agent-operator.v2.1.16
        version: 2.1.16
      - name: instana-agent-operator.v2.1.15
        version: 2.1.15
      - name: instana-agent-operator.v2.1.14
        version: 2.1.14
      - name: instana-agent-operator.v2.1.13
        version: 2.1.13
      - name: instana-agent-operator.v2.1.12
        version: 2.1.12
      - name: instana-agent-operator.v2.1.11
        version: 2.1.11
      - name: instana-agent-operator.v2.1.10
        version: 2.1.10
      - name: instana-agent-operator.v2.1.9
        version: 2.1.9
      - name: instana-agent-operator.v2.1.8
        version: 2.1.8
      - name: instana-agent-operator.v2.1.7
        version: 2.1.7
      - name: instana-agent-operator.v2.1.6
        version: 2.1.6
      - name: instana-agent-operator.v2.1.5
        version: 2.1.5
      - name: instana-agent-operator.v2.1.4
        version: 2.1.4
      - name: instana-agent-operator.v2.1.3
        version: 2.1.3
      - name: instana-agent-operator.v2.1.2
        version: 2.1.2
      - name: instana-agent-operator.v2.1.1
        version: 2.1.1
      - name: instana-agent-operator.v2.1.0
        version: 2.1.0
      - name: instana-agent-operator.v2.0.32
        version: 2.0.32
      - name: instana-agent-operator.v2.0.31
        version: 2.0.31
      - name: instana-agent-operator.v2.0.30
        version: 2.0.30
      - name: instana-agent-operator.v2.0.27
        version: 2.0.27
      - name: instana-agent-operator.v2.0.19
        version: 2.0.19
      - name: instana-agent-operator.v2.0.18
        version: 2.0.18
      - name: instana-agent-operator.v2.0.17
        version: 2.0.17
      - name: instana-agent-operator.v2.0.13
        version: 2.0.13
      - name: instana-agent-operator.v2.0.12
        version: 2.0.12
      - name: instana-agent-operator.v2.0.11
        version: 2.0.11
      - name: instana-agent-operator.v2.0.10
        version: 2.0.10
      - name: instana-agent-operator.v2.0.9
        version: 2.0.9
      - name: instana-agent-operator.v2.0.8
        version: 2.0.8
      - name: instana-agent-operator.v2.0.7
        version: 2.0.7
      - name: instana-agent-operator.v2.0.5
        version: 2.0.5
      - name: instana-agent-operator.v2.0.4
        version: 2.0.4
      name: stable
    defaultChannel: stable
    packageName: instana-agent-operator
    provider:
      name: Instana
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hdfs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hdfs-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:e8b4ad2c3b127537239b003ae19a3f58b541d599fe3d5a74a6718612fe645d50
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:e8b4ad2c3b127537239b003ae19a3f58b541d599fe3d5a74a6718612fe645d50
        version: 24.11.1
      entries:
      - name: hdfs-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: hdfs-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:00c79036eed6dd6749736b964787acb638f1a6e0156c0bba5651b93738534377
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:00c79036eed6dd6749736b964787acb638f1a6e0156c0bba5651b93738534377
        version: 25.3.0
      entries:
      - name: hdfs-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: hdfs-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
        version: 25.7.0
      entries:
      - name: hdfs-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hdfs-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
        version: 25.7.0
      entries:
      - name: hdfs-operator.v25.7.0
        version: 25.7.0
      - name: hdfs-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hdfs-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: seldon
      provider-url: ""
    name: seldon-deploy-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: seldon-deploy-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1alpha1",
                "kind": "SeldonDeploy",
                "metadata": {
                  "name": "seldondeploy-sample"
                },
                "spec": {
                  "affinity": {},
                  "alibidetect": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306"
                  },
                  "batchjobs": {
                    "mc": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48"
                    },
                    "processor": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d"
                    },
                    "pvc": {
                      "defaultSize": "1Gi"
                    },
                    "serviceAccount": "workflow"
                  },
                  "defaultUserID": "",
                  "docker": {
                    "user": "unknown"
                  },
                  "elasticsearch": {
                    "basicAuth": true,
                    "secret": {
                      "name": "elastic-credentials",
                      "passwordKey": "password",
                      "userKey": "username"
                    },
                    "url": "https://elasticsearch-seldon-es-http.seldon-logs:9200"
                  },
                  "enableAppAnalytics": false,
                  "enableAppAuth": false,
                  "env": {
                    "USERID_CLAIM_KEY": "name"
                  },
                  "external": {
                    "protocol": "http"
                  },
                  "fullnameOverride": "",
                  "gitops": {
                    "argocd": {
                      "enabled": false,
                      "namespace": "argocd"
                    },
                    "fileFormat": "json",
                    "git": {
                      "email": "",
                      "secret": "git-creds",
                      "skipVerifyGit": true,
                      "token": "",
                      "user": "",
                      "webhook": {
                        "service": {
                          "create": true,
                          "loadBalancerSourceRanges": {}
                        }
                      }
                    }
                  },
                  "image": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e",
                    "pullPolicy": "Always"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": []
                      }
                    ],
                    "tls": []
                  },
                  "ingressGateway": {
                    "ingressNamespace": "istio-system",
                    "kfServingIngressService": "istio-ingressgateway",
                    "seldonIngressService": "istio-ingressgateway"
                  },
                  "kfserving": {
                    "curlForm": "MODEL_NAME={{ .ModelName }}\u003cbr\u003e\nCLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\nSERVICE_HOSTNAME=$(kubectl get inferenceservice {{ .ModelName }} -o jsonpath='{.status.url}' | cut -d \"/\" -f 3)\u003cbr\u003e\ncurl -v -H \"Host: ${SERVICE_HOSTNAME}\" {{ .KfServingProtocol }}://$CLUSTER_IP/v1/models/$MODEL_NAME:predict -d '{{ .Payload }}'\n",
                    "enabled": false,
                    "explainForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:explain",
                    "protocol": "http",
                    "requestForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:predict"
                  },
                  "loadtest": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c"
                  },
                  "metadata": {
                    "pg": {
                      "enabled": false,
                      "secret": "metadata-postgres"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "openshiftMarketplace": {
                    "cleanupClusterServiceVersions": false,
                    "kubectlCleanupImage": "registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382",
                    "prometheus": {
                      "monitorSpecs": {
                        "create": true
                      }
                    },
                    "seldonCore": {
                      "istioGateway": {
                        "create": true,
                        "name": "seldon-gateway",
                        "namespace": "istio-system"
                      },
                      "subscription": {
                        "apiVersion": "operators.coreos.com/v1alpha1",
                        "channel": "stable",
                        "create": true,
                        "istioEnabled": true,
                        "metricsPath": "/metrics",
                        "requestLoggerEndpoint": "http://broker-ingress.knative-eventing.svc.cluster.local/seldon-logs/default"
                      }
                    }
                  },
                  "prometheus": {
                    "knative": {
                      "url": "http://prometheus-system-np.knative-monitoring.svc.cluster.local:8080/api/v1/"
                    },
                    "seldon": {
                      "jwtSecretKey": "jwt-seldon.txt",
                      "jwtSecretName": "jwt-seldon",
                      "namespaceMetricName": "namespace",
                      "resourceMetricsUrl": "https://prometheus-k8s.openshift-monitoring:9091/api/v1/",
                      "serviceMetricName": "exported_service",
                      "url": "https://thanos-querier.openshift-monitoring.svc:9091/api/v1/"
                    }
                  },
                  "rbac": {
                    "clusterWide": true,
                    "create": true,
                    "readNamespaces": true
                  },
                  "replicaCount": 1,
                  "requestLogger": {
                    "create": true,
                    "elasticsearch": {
                      "host": "elasticsearch-seldon-es-http.seldon-logs",
                      "port": "9200",
                      "protocol": "https"
                    },
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398",
                    "imagePullPolicy": "IfNotPresent",
                    "namespace": {
                      "create": false,
                      "name": "seldon-logs"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "600m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "200Mi"
                      }
                    },
                    "trigger": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "broker": "default"
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "800m",
                      "memory": "800Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "seldon": {
                    "curlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions -d '{{ .Payload }}'\n",
                    "enabled": true,
                    "explainForm": "http://{{ .ModelName }}-{{ .Predictor }}-explainer.{{ .Namespace }}:9000/v1/models/{{ .ModelName }}:explain",
                    "kfservingV2CurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .GraphModelName }}/infer -d '{{ .Payload }}'\n",
                    "protocol": "http",
                    "seldonRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions",
                    "tensorFlowCurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict -d '{{ .Payload }}'\n",
                    "tensorflowRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict",
                    "v2RequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .ModelName }}/infer"
                  },
                  "service": {
                    "port": 80,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "create": true
                  },
                  "serviceAccountName": "seldon-deploy",
                  "skipVerifyHttpCalls": true,
                  "tolerations": [],
                  "virtualService": {
                    "create": true,
                    "gateways": [
                      "istio-system/seldon-gateway"
                    ],
                    "prefix": "/seldon-deploy/"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
          createdAt: "2020-12-11 15:00:00"
          description: The Seldon deploy operator allows the creation of Seldon Deploy
            applications to provide management, monitoring and updating of machine
            learning models from a range of technologies. It includes the ability
            to run explanations as well as outlier, drift and adversarial detection
            against running models. There is a full metrics and logging integration
            via Prometheus and ELK stacks as well as full GitOps operations to allow
            full auditing and recovery of updated models.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SeldonIO/seldon-deploy-operator
          support: Seldon Dev
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: SeldonDeploy
            name: seldondeploys.machinelearning.seldon.io
            version: v1alpha1
        description: The Seldon deploy operator allows the creation of Seldon Deploy
          applications to provide management, monitoring and updating of machine learning
          models from a range of technologies. It includes the ability to run explanations
          as well as outlier, drift and adversarial detection against running models.
          There is a full metrics and logging integration via Prometheus and ELK stacks
          as well as full GitOps operations to allow full auditing and recovery of
          updated models.
        displayName: Seldon Deploy Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - machine learning
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://deploy.seldon.io
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        provider:
          name: seldon
        relatedImages:
        - registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398
        - registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
        - registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c
        version: 1.2.1
      entries:
      - name: seldon-deploy-operator.v1.2.1
        version: 1.2.1
      - name: seldon-deploy-operator.v1.2.0
        version: 1.2.0
      - name: seldon-deploy-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: seldon-deploy-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1alpha1",
                "kind": "SeldonDeploy",
                "metadata": {
                  "name": "seldondeploy-sample"
                },
                "spec": {
                  "affinity": {},
                  "alibidetect": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306"
                  },
                  "batchjobs": {
                    "mc": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48"
                    },
                    "processor": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d"
                    },
                    "pvc": {
                      "defaultSize": "1Gi"
                    },
                    "serviceAccount": "workflow"
                  },
                  "defaultUserID": "",
                  "docker": {
                    "user": "unknown"
                  },
                  "elasticsearch": {
                    "basicAuth": true,
                    "secret": {
                      "name": "elastic-credentials",
                      "passwordKey": "password",
                      "userKey": "username"
                    },
                    "url": "https://elasticsearch-seldon-es-http.seldon-logs:9200"
                  },
                  "enableAppAnalytics": false,
                  "enableAppAuth": false,
                  "env": {
                    "USERID_CLAIM_KEY": "name"
                  },
                  "external": {
                    "protocol": "http"
                  },
                  "fullnameOverride": "",
                  "gitops": {
                    "argocd": {
                      "enabled": false,
                      "namespace": "argocd"
                    },
                    "fileFormat": "json",
                    "git": {
                      "email": "",
                      "secret": "git-creds",
                      "skipVerifyGit": true,
                      "token": "",
                      "user": "",
                      "webhook": {
                        "service": {
                          "create": true,
                          "loadBalancerSourceRanges": {}
                        }
                      }
                    }
                  },
                  "image": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e",
                    "pullPolicy": "Always"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": []
                      }
                    ],
                    "tls": []
                  },
                  "ingressGateway": {
                    "ingressNamespace": "istio-system",
                    "kfServingIngressService": "istio-ingressgateway",
                    "seldonIngressService": "istio-ingressgateway"
                  },
                  "kfserving": {
                    "curlForm": "MODEL_NAME={{ .ModelName }}\u003cbr\u003e\nCLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\nSERVICE_HOSTNAME=$(kubectl get inferenceservice {{ .ModelName }} -o jsonpath='{.status.url}' | cut -d \"/\" -f 3)\u003cbr\u003e\ncurl -v -H \"Host: ${SERVICE_HOSTNAME}\" {{ .KfServingProtocol }}://$CLUSTER_IP/v1/models/$MODEL_NAME:predict -d '{{ .Payload }}'\n",
                    "enabled": false,
                    "explainForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:explain",
                    "protocol": "http",
                    "requestForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:predict"
                  },
                  "loadtest": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c"
                  },
                  "metadata": {
                    "pg": {
                      "enabled": false,
                      "secret": "metadata-postgres"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "openshiftMarketplace": {
                    "cleanupClusterServiceVersions": false,
                    "kubectlCleanupImage": "registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382",
                    "prometheus": {
                      "monitorSpecs": {
                        "create": true
                      }
                    },
                    "seldonCore": {
                      "istioGateway": {
                        "create": true,
                        "name": "seldon-gateway",
                        "namespace": "istio-system"
                      },
                      "subscription": {
                        "apiVersion": "operators.coreos.com/v1alpha1",
                        "channel": "stable",
                        "create": true,
                        "istioEnabled": true,
                        "metricsPath": "/metrics",
                        "requestLoggerEndpoint": "http://broker-ingress.knative-eventing.svc.cluster.local/seldon-logs/default"
                      }
                    }
                  },
                  "prometheus": {
                    "knative": {
                      "url": "http://prometheus-system-np.knative-monitoring.svc.cluster.local:8080/api/v1/"
                    },
                    "seldon": {
                      "jwtSecretKey": "jwt-seldon.txt",
                      "jwtSecretName": "jwt-seldon",
                      "namespaceMetricName": "namespace",
                      "resourceMetricsUrl": "https://prometheus-k8s.openshift-monitoring:9091/api/v1/",
                      "serviceMetricName": "exported_service",
                      "url": "https://thanos-querier.openshift-monitoring.svc:9091/api/v1/"
                    }
                  },
                  "rbac": {
                    "clusterWide": true,
                    "create": true,
                    "readNamespaces": true
                  },
                  "replicaCount": 1,
                  "requestLogger": {
                    "create": true,
                    "elasticsearch": {
                      "host": "elasticsearch-seldon-es-http.seldon-logs",
                      "port": "9200",
                      "protocol": "https"
                    },
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398",
                    "imagePullPolicy": "IfNotPresent",
                    "namespace": {
                      "create": false,
                      "name": "seldon-logs"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "600m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "200Mi"
                      }
                    },
                    "trigger": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "broker": "default"
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "800m",
                      "memory": "800Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "seldon": {
                    "curlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions -d '{{ .Payload }}'\n",
                    "enabled": true,
                    "explainForm": "http://{{ .ModelName }}-{{ .Predictor }}-explainer.{{ .Namespace }}:9000/v1/models/{{ .ModelName }}:explain",
                    "kfservingV2CurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .GraphModelName }}/infer -d '{{ .Payload }}'\n",
                    "protocol": "http",
                    "seldonRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions",
                    "tensorFlowCurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict -d '{{ .Payload }}'\n",
                    "tensorflowRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict",
                    "v2RequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .ModelName }}/infer"
                  },
                  "service": {
                    "port": 80,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "create": true
                  },
                  "serviceAccountName": "seldon-deploy",
                  "skipVerifyHttpCalls": true,
                  "tolerations": [],
                  "virtualService": {
                    "create": true,
                    "gateways": [
                      "istio-system/seldon-gateway"
                    ],
                    "prefix": "/seldon-deploy/"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
          createdAt: "2020-12-11 15:00:00"
          description: The Seldon deploy operator allows the creation of Seldon Deploy
            applications to provide management, monitoring and updating of machine
            learning models from a range of technologies. It includes the ability
            to run explanations as well as outlier, drift and adversarial detection
            against running models. There is a full metrics and logging integration
            via Prometheus and ELK stacks as well as full GitOps operations to allow
            full auditing and recovery of updated models.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SeldonIO/seldon-deploy-operator
          support: Seldon Dev
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: SeldonDeploy
            name: seldondeploys.machinelearning.seldon.io
            version: v1alpha1
        description: The Seldon deploy operator allows the creation of Seldon Deploy
          applications to provide management, monitoring and updating of machine learning
          models from a range of technologies. It includes the ability to run explanations
          as well as outlier, drift and adversarial detection against running models.
          There is a full metrics and logging integration via Prometheus and ELK stacks
          as well as full GitOps operations to allow full auditing and recovery of
          updated models.
        displayName: Seldon Deploy Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - machine learning
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://deploy.seldon.io
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        provider:
          name: seldon
        relatedImages:
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d
        - registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48
        - registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c
        - registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e
        version: 1.2.1
      entries:
      - name: seldon-deploy-operator.v1.2.1
        version: 1.2.1
      - name: seldon-deploy-operator.v1.2.0
        version: 1.2.0
      - name: seldon-deploy-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: seldon-deploy-operator-rhmp
    provider:
      name: seldon
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: rhods-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhods-operator.2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
          createdAt: "2025-08-12T18:04:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "llamastackoperator": {
                      "managementState": "Removed"
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (including PyTorch, TensorFlow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack – unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.23
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fbe10282ada2fa6bf07ae97b497f46ebe735de130b4ce0182bf34301429999cb
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:2d2e7803b2b1e15017ca04961f9dcd25e89c47b07b101ca54cb9e793e6027202
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:e50adae13e9d773e7be8bf81b7525dac3f5e3864c9cdf4c287342bd568ee1d6b
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:bb7c352ccc10a8c60ef8d573368d0ada59e73cd131653d11bcebdf9f564acc10
        - registry.redhat.io/rhel9/mariadb-105@sha256:050dd5a7a32395b73b8680570e967e55050b152727412fdd73a25d8816e62d53
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:ccfc84028a95f1cc6a16e06ff6e9ed2ace74f74e5e9963a4fb33544c2ce005c2
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:2886e40f94fde72eb5959ad564bbdd466657208c09bac731b3860b0b54f104c6
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:b038d20def504c140e4365f2eb117a979534629332bec75a2ba750ebafdefe3d
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5fb7e6c3ee1e3b170cd146ea9cd172fa953326b08163897605ca74a3644f2819
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:795ab1842ef675e3d8b47deb8a7485a69064428e823dfeb930a6a7e7017cde0f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:80d284cac2ef6638fe028102b71a70f26041576c11a49414e77430954a3c1b99
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f23f8051886d0f0240937507471c818f440bca8703b8209a836d5dbe3ea2f7c8
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:16ff53f4f1a1e1e66826c8355c654038da83df545427a931485ec7a2d3764b16
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:23c098f7e791441cfcfa6bf37de51763511bd8bf4d543747b7d37ed591d68886
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:edcce7cddfa4c250108532e56541ddb95099612c180b2297e5934bc89d2bb46e
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:ff740fc3615edaa90dae976da1aa6455d64811e239eb66b422836a3aae791b41
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:240692b86bc8a4fe0b545e6e4cd7cf4ff0ca990406bc55f0446d565fd67e37f6
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:a4649e8ad4970bc3d5e967811c9694fa82d78ee9f8f025b8b2fbd37237eb14fc
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:68e74bdf518fc624cc25a59a18b2a87d0aa2a9b8aad30750ace7115d297a21ca
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:ca21e218e26c46e3c63d926241846f8f307fd4a586cc4b04147da49af6018ef5
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:857560240bb45bfc2db6d7743abf11a47f929a0d3035940256a6527afd313412
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:84c3213e11299ec032aa0e5a4dfa9b3055c4246b04b029e888d8d9f1caf5c30c
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:4847ac1be55284815670c2048c1edfc56c7227494280d3c6094aa23f1ed1a9d2
        - registry.redhat.io/ubi9/toolbox@sha256:4213516b6fe7957a01072720efd9ecbc7d4443238e151b5fe0263bd0c36792b1
        - registry.redhat.io/rhoai/odh-llama-stack-core-rhel9@sha256:43b60b1ee6f66fec38fe2ffbbe08dca8541ef162332e4bd8e422ecd24ee02646
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:467b04f4ee4118bbe0934d1ab8b80f72df8703dcdac61c2b53fdb5bd1b404c14
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:a1497e64349634b2567d28f685052de118f64f6b32e7d3ea469212e2004d7eb3
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:e0c9ac73891781ef421e0a020cfd5c962f7632d38020897183cf048b1819326a
        - registry.redhat.io/rhoai/odh-llama-stack-k8s-operator-rhel9@sha256:626b9ef09ee82c8d425ddb7ed3f2ffbb3d1019f541e31604db891787c881d5c9
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:3243caa4c8aac75069d19225dcee361bf9c012782328a6266b7016b3026f6d56
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:756b7bdaa7f6e7c4a18488e690188dd91153a33913d225bab001b0085382e514
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:da8debeb2b84a42aba0b98785ccc20e02ed520e125428fe012df3ca473cbb755
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:208005da389b7d4176ad8bffc703312011a1299954df151e2abbe8f3dec01d19
        version: 2.23.0
      entries:
      - name: rhods-operator.2.23.0
        version: 2.23.0
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.20.0
        version: 2.20.0
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.18.0
        version: 2.18.0
      - name: rhods-operator.2.17.0
        version: 2.17.0
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.15.0
        version: 2.15.0
      - name: rhods-operator.2.14.0
        version: 2.14.0
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: alpha
    - currentCSV: rhods-operator.2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhods/odh-rhel8-operator@sha256:a4d7f020509b9720063e87e1c2f19af04e4c9b4986b199be4dc617831b971434
          createdAt: "2023-11-21T08:37:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            Data Science
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:08c90a8f133b4cdb1477a742bc44be4d8fc4e88153b8bc2a2a148cbdf19f4e7d
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:6a56010de933f172b195a1a575855d37b70a4968be8edb35157f6ca193969ad2
          olm.skipRange: '>=1.33.0 <2.4.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "rhods",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "rhods",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Data Science"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:eb96f4b753606528750b3c28f7f9256582f81fa030450b05afdef684634903be
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift Data Science
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift Data Science (RHODS) is a complete platform for the entire lifecycle of your AI/ML projects. It is the flagship product of OpenShift AI.

          When using RHODS, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. RHODS can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * RHODS dashboard
          * Curated Notebook Images (incl CUDA, PyTorch, Tensorflow)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh (and Kserve) with a provided OpenVino Model Serving Runtime
          * Ability to use other runtimes for serving
          * Model Monitoring
        displayName: Red Hat OpenShift Data Science
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift Data Science
        - RHODS
        - rhods
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift Data Science
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-data-science
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhods/odh-ml-pipelines-api-server-rhel8@sha256:be57f237df156474a709667fa4fb4aa142118e3459963fed56648c1cb2e38cd6
        - registry.redhat.io/rhods/odh-kuberay-operator-controller-rhel8@sha256:bfc805aa1689a4bd6f7e617acf9eb1c2c307624dd49956e9e7442afac328151a
        - registry.redhat.io/rhods/odh-data-science-pipelines-operator-controller-rhel8@sha256:250db47cb873207cfb75c7e8a2f0de08b75b056a72db73362d2d79c08cb7b185
        - registry.redhat.io/rhods/odh-trustyai-service-operator-rhel8@sha256:68471555d0b2bb8150ebfae8d4f3d1dbdff9c5b961ff2305ef23ab7514cbbea1
        - registry.redhat.io/rhods/odh-codeflare-operator-rhel8@sha256:2841fd3051a50a1b0d9e3caafa20a8084b5c472d49393b372828960dbf156674
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhods/odh-model-controller-rhel8@sha256:fde989817cd1a6ec8eb408a7f6c73e28ba043ab2d88035586b1b7e55fb3405cb
        - registry.redhat.io/rhods/odh-modelmesh-runtime-adapter-rhel8@sha256:08478a1647d0413783bcbcd966a73b444b580464de46efedb39504420f22053e
        - registry.redhat.io/rhel8/mariadb-103@sha256:08c90a8f133b4cdb1477a742bc44be4d8fc4e88153b8bc2a2a148cbdf19f4e7d
        - registry.redhat.io/rhods/odh-trustyai-service-rhel8@sha256:fe17d6e07ebaa088f274dd8b5e83892dacc11cb336065b06582a3c055095dba5
        - registry.redhat.io/rhods/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:c5758102609821a6fdabec0277e03c4c728f82176093d1ade0570aba5f3cd8f9
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhods/odh-ml-pipelines-persistenceagent-rhel8@sha256:857a8eb444375505ccb34a11cce69c1794ceabbfb00df06933d7f12cd187319b
        - registry.redhat.io/ubi8/ubi-micro@sha256:6a56010de933f172b195a1a575855d37b70a4968be8edb35157f6ca193969ad2
        - registry.redhat.io/rhods/odh-rhel8-operator@sha256:a4d7f020509b9720063e87e1c2f19af04e4c9b4986b199be4dc617831b971434
        - registry.redhat.io/rhods/odh-mm-rest-proxy-rhel8@sha256:8d989f54009484d7f43fe79d7c37757418fc3c15cd4d7e80a2076677a9645bdf
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhods/odh-ml-pipelines-cache-rhel8@sha256:33aae32e3d348d382906f6e926281852f54006a94d7bdac188bb33debadec059
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhods/odh-modelmesh-rhel8@sha256:3827221f84d3aa22440d6b364b990a4ec0666dfb5963f174f7ed518b3650d223
        - registry.redhat.io/rhods/odh-kf-notebook-controller-rhel8@sha256:cd0ab8d70559be6074fbaf1caa1d50f8eb75e1bff1a22d6d69380f19190df0b6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:eb96f4b753606528750b3c28f7f9256582f81fa030450b05afdef684634903be
        - registry.redhat.io/rhods/odh-ml-pipelines-artifact-manager-rhel8@sha256:76c67d7015a2db6965219db6533e0cddb5b7cf699c88df3ef81155f9c192a3cc
        - registry.redhat.io/rhods/odh-modelmesh-serving-controller-rhel8@sha256:d94395fb251e45aaa73d09bc6106a2947c16b90d87fc3b64bfe4f2f600be5841
        - registry.redhat.io/rhods/odh-dashboard-rhel8@sha256:b98171c5543815abf7c095fba92796ceed9ea44bd8b78db143d2815c6bcfdcf6
        - registry.redhat.io/rhods/odh-notebook-controller-rhel8@sha256:753994fce63802f1e1121990a5c1dd722216446abee54cc1a7c50ab195134414
        version: 2.4.0
      entries:
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: beta
    - currentCSV: rhods-operator.2.21.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:7d0b8d69b6defb5f232da066573b0067ec810a8ccfb2f3cb91d472d2979ab7a7
          createdAt: "2025-09-24T14:42:33Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.21
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:687c8eeed55f021ecaab1307f0e88b5b16d91f72d63b3d7168d7bbee90e8947b
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:04859c878ffcc8c42550a98be625ac52212f9b79a6fd5b3baeafb3592926cbe9
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:2a16c353fd13dc1f3f5c4266f1b9786a8e7a100dc05f2d554bdaef082271063f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:239f226328e5ac77d6dee498217d55103a6d562f4829e38abf65ba55fda0aca2
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:7da004e158006dcbdcf0c3b78c40411b82c9d56ebf54b1c30c245b6550b8ee7d
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:2b4136569e10ad7d47c82bedf55c1705a833dc67c8f833656996bfe1fddbef3e
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:52e1ba8ebf4d55c64c94c9076185c5612b537e6e5ed4944fcd6ea87ccb9cd182
        - registry.redhat.io/ubi9/toolbox@sha256:7656f815a0d13ed37f56d11635559a87a3d00840253755c71b644b858f67d556
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:7d0b8d69b6defb5f232da066573b0067ec810a8ccfb2f3cb91d472d2979ab7a7
        - registry.redhat.io/rhel9/mariadb-105@sha256:c390b12f5c2c3ea7cd897b59eba7127ce22f474130cd8c1bbbbaadae870baca4
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:1b5452b87b36de2b62febc777efdc6233a168ffcc09e2067e482cc444463a57d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:cb47e165935f0d9d3c88088689c37e68a69519357a20d3c52ab6f71240dab507
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:003fdc6cd1821d620e1ce4050ee608e2d2a32f2b08c8fa9c147e2e14418a8d91
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:80fd53a580ce7bc4d02023d38a04912e84c2cd50679fc5ea034afba935623157
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:fe6593bc73e77d315d82b7921bb1975c19fda1e7630ef5b30b69f5901ce30178
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:e68492a6a75cb83be0a968b832c1478b97f14e661376328085fd5416a08cbc31
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:c3e283c0536266195dd461042ca0036ed98cc33650d806be132db4fba45239e4
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:152b425725228953f43b57a4a034a6725993a5ca556af328ae5e06aa3f1acbfc
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:a88696cff8cd2cc2e295c9ec082b7f25f61dfc9ac12f1c208529ff7036d5af29
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fc98ed817a4e9ab84e2a66063987304f0cdbe6e5620a58a7f25cccd61b34ebf9
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:5c0d56f434298f6e8bf6cf7061cd1ae6df9915225cbf04b010b6b59a64775cc9
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:cfaf4eb86ddb9c717340ab17fa4b83323e5b40977ac70265376c87329a3f6f20
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:4c57dcf4df09fe0d03578cc8b9a3ca2ea485b6af0ca0203bfb85e640058b7daf
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:4c192208c618b49f31cf31c9de1a33c622ce3f120204c2eec8ff7c439ea6785c
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:a936d87646843edea2a84eb429bb44eb730bf3e04e8900fecef66721426dba89
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:9c2d6ce05775ac34ee5695b3367d12bf2a9c14e318a9544d00a56653ea5f2007
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:dc0db686b7688f98b5f8595661022cdffa5d26a078289cc7aed528f0ee1d08a4
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:6a388ee3cbe3813ac955d319f0ab65a81da321d982c9881c6c81fe135311fe1d
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5f581957e4f8deebf6104a394272e6fbb1c54d813bf069e95875d865eac209ad
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:2638e3a90bbb242dd2483fefe9fc8b472f334c510d0139b9b7c875c30dab7984
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:85202cb75250a1ba2691d65b52b306fd77566660dd29cdf56bf8962e8bb4939c
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:1b7befcd29c9c8f783bdd3c5654b0412ea45cd88cd18eb5f8f5a411f1323c6bd
        version: 2.21.1
      entries:
      - name: rhods-operator.2.21.1
        version: 2.21.1
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: embedded
    - currentCSV: rhods-operator.2.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
          createdAt: "2025-09-24T13:27:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.16
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:22557a6fa52d2f311750a9ba253860f423ba697d26efa02ef8524a8258d2a909
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:d7d4fa406e0fcf0507894a7676532b27f45be742467e603a86f98ea5d2615df8
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:2ba2b2c4db8bb334c50f4dfb54059f060361186900a44c06eed00b7a3c43977e
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:7b70af8847b0806d8b43c399e2b3109f016fd864f5e9d30c44e2baca5d1359dc
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a0f72ffefb2b74b488dd949493f5d295a39bb9c97f578bf219d0138601f65468
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:6cf74044ae8d5308a2dfe03fa5d81086c89302db7f4cdbe2f4174a1c48b77869
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:de5f91180ead7d73a1825fe8b032fde9b8b01392569c9789f5ce1b4b9c08a98f
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:ee01e89f98feb185f6cd59c564e590a13e4d8d9ea760cca8de51426eb71b83a1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:04f305c44413ae7dcb7017e53570ee49a509701792c5f50efadd64f47395730b
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:c8456d98e90b6505957ab3686e9fd2f156e29f123c5558e581c206daf1e7d93a
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:f738aab1eab25854c93e1d8d4d98100a8ae7bb45a6b83f0326774e4220b1183b
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:5dcdcc2424602a69451f16d31dbfa1d43cb72c095ba561eb9076f0cd1e8182ed
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:efd039012559786deb3c521a130886c265e88d635e08baace2e573a9df270134
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:27b113d96453c2054d4c965a963d2badba6daac235eddaf23234c76e87bcb069
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:4c5ff3496b2a2a739939d94ee9dafc02b682100785d228dc2fde480fb597b7a5
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:4a7599f8a866eb169c9a62885906adbf6df0417c0be15857df1eef20cd9b1be2
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:633984b538f027c93c5886ea1045dcbe81eeda74acf80001d5fc5f765bdbe0be
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:23d307a36b69e0df04f72a7d3b35e28d8417a8bbe23dba31e8e977569785c078
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:11a301728573adf2b64ea72d0cb2d83ea5d4dbebea759f346e99f18c3d368c6e
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:5192f1dbd2a9ab92ae390c4ae506efbed0970545b6122e95b014728ac937e777
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:2eedd1e2465d2e78fafb39e234dbb413a82713f187c64d70a80340ec94807d30
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f37e4048f3a152798286793f3abfc6ed814453fcbe2667255a7e78eee483c5a3
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:97e2bd9b587f08e135a9aeb9b3e0dc6eafa1a9bdacbb5ecb681ce9bd5aa37fc9
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:65482864055021272a18b5b26792ab00cfa5fc9cc005d8d3a884cc82991506f3
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:4a212ec634225c14beac09be24ddf336e562f2aa9a13555fb1196f366ddae23c
        version: 2.16.3
      entries:
      - name: rhods-operator.2.16.3
        version: 2.16.3
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.8.5
        version: 2.8.5
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: eus-2.16
    - currentCSV: rhods-operator.2.8.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:3b9e8a71ba795a2ea91f4129253ccb4be3a79c7c1c6a87082f96ccc67eabbbac
          createdAt: "2024-11-07T07:17:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.8.0 <2.8.5'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:eeb8bdc94e7fa2b454fedf72eb75756052f8bfb9d92b116a8267a5584e2528c1
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:f9c474d454d0b006e06859601ec9a64b5e5c91e9a47bfab6ad6d8cce5aa6b89d
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:45498198fadfbc1457bf2a6084ae7ea515d454d86d148783aae3abdaca6e999d
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:60507a3dddbd88560fd85204699147facd2b6371fb25e0bf6701dad02c0b4dd4
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:df34e9d2d121a3def2dbb86f108b3fb6948bc7ba7cb44c70801dfd2e81d3fbb1
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:13dff890ec67fa56864c244a40fc404e9eac05f33573dd16d088b37f9340ec41
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:28ec01e514c15cae145499840597d1be1b750123bb6014de2ecc3ef0d60fa134
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:9aa5a93831992588766ac8594ae3c4dc7111d08a90fdbc8751c40484339e46d4
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:038ce76571217d6c624d614652e8b96ef6484562cf52ec1be88392b8ef8a3bbd
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:1a1074263ae26e4c8b0e3863228c35e50133169f377ece6c47b58a12e0fcf453
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:81ae2b95f2da740cb5264b9ba8b85f0f58bb4ab80d6431f66d52848ed0c21afb
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:1dbf09cb1624c9f961f8ac8e371c97aa7c3fbfcad4c4d02ab4fad0eb86d2aa95
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:61a506fbf263cd72951ac8bd151127e0c74fc74f0a05f7aaecaa54fde47ee490
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:3b9e8a71ba795a2ea91f4129253ccb4be3a79c7c1c6a87082f96ccc67eabbbac
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:13b850a5a3f8776c6c8db887c666d88df72901895b5ef1be9a8bd372cc02fcb8
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:8ffc5da798e783c65d5be7f344219b2951d8c2233d6494a4ae2bb5c86eccd699
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:d87e4b1cba72eb4f50b3f2d59863799929c22ca37b01a1ff4828a9ce0be31f25
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:2f6db789ceb3f90f9c9911c6392e62d0a8e4d61be68a7f7cbd781f47faab6389
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:4b1b9f05a94713292a7a2f6782fe1d3eea6f36214c36b38f402aa91a554ab18a
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:5c48474563d98df32e7243e33302428ea8836be5e1a2e30d20b4d796a62cbf6b
        version: 2.8.5
      entries:
      - name: rhods-operator.2.8.5
        version: 2.8.5
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: eus-2.8
    - currentCSV: rhods-operator.2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
          createdAt: "2025-08-12T18:04:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "llamastackoperator": {
                      "managementState": "Removed"
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (including PyTorch, TensorFlow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack – unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.23
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:edcce7cddfa4c250108532e56541ddb95099612c180b2297e5934bc89d2bb46e
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-llama-stack-k8s-operator-rhel9@sha256:626b9ef09ee82c8d425ddb7ed3f2ffbb3d1019f541e31604db891787c881d5c9
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:ff740fc3615edaa90dae976da1aa6455d64811e239eb66b422836a3aae791b41
        - registry.redhat.io/ubi9/toolbox@sha256:4213516b6fe7957a01072720efd9ecbc7d4443238e151b5fe0263bd0c36792b1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhel9/mariadb-105@sha256:050dd5a7a32395b73b8680570e967e55050b152727412fdd73a25d8816e62d53
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:68e74bdf518fc624cc25a59a18b2a87d0aa2a9b8aad30750ace7115d297a21ca
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:a1497e64349634b2567d28f685052de118f64f6b32e7d3ea469212e2004d7eb3
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:e50adae13e9d773e7be8bf81b7525dac3f5e3864c9cdf4c287342bd568ee1d6b
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:84c3213e11299ec032aa0e5a4dfa9b3055c4246b04b029e888d8d9f1caf5c30c
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:ccfc84028a95f1cc6a16e06ff6e9ed2ace74f74e5e9963a4fb33544c2ce005c2
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:bb7c352ccc10a8c60ef8d573368d0ada59e73cd131653d11bcebdf9f564acc10
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:2d2e7803b2b1e15017ca04961f9dcd25e89c47b07b101ca54cb9e793e6027202
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:b038d20def504c140e4365f2eb117a979534629332bec75a2ba750ebafdefe3d
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:3243caa4c8aac75069d19225dcee361bf9c012782328a6266b7016b3026f6d56
        - registry.redhat.io/rhoai/odh-llama-stack-core-rhel9@sha256:43b60b1ee6f66fec38fe2ffbbe08dca8541ef162332e4bd8e422ecd24ee02646
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:23c098f7e791441cfcfa6bf37de51763511bd8bf4d543747b7d37ed591d68886
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5fb7e6c3ee1e3b170cd146ea9cd172fa953326b08163897605ca74a3644f2819
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:857560240bb45bfc2db6d7743abf11a47f929a0d3035940256a6527afd313412
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:795ab1842ef675e3d8b47deb8a7485a69064428e823dfeb930a6a7e7017cde0f
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f23f8051886d0f0240937507471c818f440bca8703b8209a836d5dbe3ea2f7c8
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fbe10282ada2fa6bf07ae97b497f46ebe735de130b4ce0182bf34301429999cb
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:80d284cac2ef6638fe028102b71a70f26041576c11a49414e77430954a3c1b99
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:467b04f4ee4118bbe0934d1ab8b80f72df8703dcdac61c2b53fdb5bd1b404c14
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:ca21e218e26c46e3c63d926241846f8f307fd4a586cc4b04147da49af6018ef5
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:e0c9ac73891781ef421e0a020cfd5c962f7632d38020897183cf048b1819326a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:a4649e8ad4970bc3d5e967811c9694fa82d78ee9f8f025b8b2fbd37237eb14fc
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:208005da389b7d4176ad8bffc703312011a1299954df151e2abbe8f3dec01d19
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:240692b86bc8a4fe0b545e6e4cd7cf4ff0ca990406bc55f0446d565fd67e37f6
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:756b7bdaa7f6e7c4a18488e690188dd91153a33913d225bab001b0085382e514
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:2886e40f94fde72eb5959ad564bbdd466657208c09bac731b3860b0b54f104c6
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:16ff53f4f1a1e1e66826c8355c654038da83df545427a931485ec7a2d3764b16
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:4847ac1be55284815670c2048c1edfc56c7227494280d3c6094aa23f1ed1a9d2
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:da8debeb2b84a42aba0b98785ccc20e02ed520e125428fe012df3ca473cbb755
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        version: 2.23.0
      entries:
      - name: rhods-operator.2.23.0
        version: 2.23.0
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.20.0
        version: 2.20.0
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.18.0
        version: 2.18.0
      - name: rhods-operator.2.17.0
        version: 2.17.0
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.15.0
        version: 2.15.0
      - name: rhods-operator.2.14.0
        version: 2.14.0
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: fast
    - currentCSV: rhods-operator.2.22.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
          createdAt: "2025-09-25T11:28:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.22
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:7b3d7787ba95156de9c284ad78f2d63a0d3bcb525f4d8d11ebf0203d4c5c1fe0
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:7fd540586a1ca111370100a473923b3dc6cec4255e9e4f3654bfbabd9158a2bb
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:a44d32a18b1e906c1411cafc82102881573fcd04e040a4e2ba6bd1bd617b70fa
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:6234e15f001122ab222481321153fb0eba7f4d86bf3dfd64edf3f6cecbe7d62f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:abe1f9dd11c7313fdd55b678be511a5cdc3546654a68e2b53d42c1013fec8d00
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:b2ca229bb4c4dbf04b96766b8af8d9f55f2edef59ec4110de0d20f8c857e7d27
        - registry.redhat.io/rhel9/mariadb-105@sha256:5e77b33ce04c1eafc2e8840586ad6a5b51148dea8a46dda151071e28858331f0
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:2f3f305c97e50057781553416fb4026cde306ead0fbe0213627b9b5bbcc8a0c3
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:e6fbe6723b4801f6854303b3fd7d4a4a36727bdd28dc28002b7de4929739ccc0
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:2f1a0a0803b2f5118b167e9d9c2e86660add5f79ee6f47044f63b0e8abe0b75f
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:cb5be3dd7098f4c8d66f9e85ad04656460df6102652deddc6cccb884a12be733
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:9104cbc2d595933f0a2acb7313e41ed69905e76a9fae6bd9820e7717b56a632b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e3b90530c6e38f6940d7f9408e9fa2c81eb06a857ec2b0bab5d006b1e9a63a4f
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:32c9511846077ef6035eb19be7653f6a78e0f207eb420ab93b18f742ab5779b6
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:27415b1c763cf89b79391f0b039936c3e078a078bda4420f13bf10ef9be31d5e
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:463e9be9c4409ba1e5a42713bd65844a023e7a3e9032d4e862edbaa5fe35300a
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:1709fa3c79aad4ba7eb9be8299949396092c8e20210124e0c0936385bc04e839
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:b2c2ea9d35edbcd16e1ae559c88a453129b877fc049ab7a7cd6c2ebb8f0bcb59
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:bd6d1599b4ff9eb890be9808014715b28ca19247222d65f40df024db4dd3525a
        - registry.redhat.io/ubi9/toolbox@sha256:ea4ce8cb8f59cf1b0709f1606d7cab074c3806d4016964eedc0555410cb5dfeb
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:d59ed40de694bc0e0564bab159a5b586cffa0721b6d761fb2d81dd42e6225d1c
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:74d4e0174655cf6f6273857325a1e714a1531c5a06ad38bf139bee9c7d8f5f19
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:5d79f0bcca0a213b53aee6d88834e503d343821596b857b13d3a6362de8ebeab
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:5914f71e37c815ce2567826896e051303bdef52fae1acf8d02e672fc24c7eb9d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:2065b96bd6f0a830eab66e36217a927448cfc5a7dabf2a5e7d907aa6c40657a4
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:2b3aace0d80474470e71c2c8291b25738e096109229fe6d5d3e06e9885ee58b6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:13f3679ab289eea9de54fe129a7c83f6c8ec5d26b61b430637a7c4cfcd31a74a
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:ccba67a6a5aa3403e0018f29dcdfef483220dea8065d8b572e80c653ca992add
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:2fe9473cd1d6c445d2ca7cc357ee4bb009aa1838f9cb69ea418296b2a044816b
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:d7c4ac937add511460a36375d9b6b3c475a3a9a7cc96c00df2fbad6219aee0e9
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f9965177de24f3165ec4c001c14cf1e2c2c01ffe847839af837c442e3c96beae
        version: 2.22.2
      entries:
      - name: rhods-operator.2.22.2
        version: 2.22.2
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable
    - currentCSV: rhods-operator.2.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:8643f7ee8afacfe6ae08cbbf81da76ca97032dc3ce3da4fddf16e330e9e33a08
          createdAt: "2024-11-14T21:18:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.10.0 <2.10.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:a1d590d0e2b3fae9ecbd843a664dab083061e797f66a9f0a021b0c2c487d7e45
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:0af6a91553f40ea40e86d85edc204658af3e5a45e8ab94cdb6b767e545df59b9
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:214ca0608fcbc0fd585bf01768d28c7ba17082d1043349ec8445afd663095383
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:d2514bf4b4b2d9abc19facac3f1a55405eb3ad0d9e9203e8816bfcfa08c9a6c5
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:4876f14ca9cff0e2b998a56f2f56d0775edf3963de610f4b810b4d94fb12d719
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:24238abd46633709feb75f2a3e41cf6c9e8528d2c0e4be2fac6f56de44e3fb9b
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:d9b55e8f25b81050380b4aab040243dc18a7432c48912c9b660c97dcc5eea008
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:8643f7ee8afacfe6ae08cbbf81da76ca97032dc3ce3da4fddf16e330e9e33a08
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:9f7b1cb8623de8b0550b0ba124d3b50f5e960ed011b33a3262c6346dbe2dff95
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:807ec05e9e107afc969e718d0c2c980959e2fd462e777a5ff63e55156ea150a4
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:fdc2bc06b0dfa79ebc6451ac3793b5e8d9f165171633831dafe46ebc18c10c35
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:4da77afafde99b15fd357a4ab6609644bf6330a1d9c0c03e64726ed1b9633230
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:cba9f25ac6bb2734bb990a622b515c838b881edb688bf44ff196ebe7684ab74b
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:8911afbb65c8c4f10c3597624485be54b9103e87a5020b718b35ec3df2c9bf00
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:d709f65e1194bc877fa4f9ec681b598467703d94139d57b58c24490d3a59fb82
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:220cad7a22d40998063c6b4fb2f63e69330e64ca7118490a6680b05fc99b7733
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:8fed8f58732bff512289fff738d0a94be67286c41b24eaf5928bbcb30d88c88a
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:232ebac25084db056fdde49477b5c1e6a03a0228fea3758bcf6234b3d0eaa582
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:2007da5abf840712cd2edb50462b2d8d571a0193c3ae1550d4a279eddf2b335f
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:a4b9553fde02d9c67a9c98f063eaf82fb1d2689f4b1269e57bf5f12c8d4adecf
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:5d15175f934017c04d6a6057200ff55b2172a1f4bcbeb0f2cd9e9e545631d93f
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:f4d7d3ba4a0d3322bc9f327e095e9ec56073758f6dcd8da8c7bb185cf952ead5
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:7df2f8bc2c5c0f4f347d009b62a058b5107413a2abd789f6010eb4d6c9e4f8ee
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:a9c6e3afdf605465766fa93334cc0f6bb5e46a5fdfd7ec64d648e2ed72b05ccf
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:267d6795bf53f9a1955d73445776a428dd042af0cd3cb8bb5b91692a1bc1e5d2
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:65db5e8cb2fa46d020ed4837d4479d133b52167c373e65f81cb590eec5391af2
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:0ff5664c2eaadf3d3073b3d7fb711da47fb78d21f67710d8b8aad5452cb0355d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:12741379ab24d2286b5fae77f4383fb65713af7b3356fb43d8254ea85d4b0ee3
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:52e8deeeeaa21b17370a5d5cd644745da745b6e1cee8c6295d6df75c8d2aec04
        version: 2.10.2
      entries:
      - name: rhods-operator.2.10.2
        version: 2.10.2
      - name: rhods-operator.2.10.1
        version: 2.10.1
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.10
    - currentCSV: rhods-operator.2.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:28b5fff602ad6fb36d462e1905f453509ec22fc251a1e0cf64fedba86ff420d7
          createdAt: "2024-10-17T10:53:18Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.13.0 <2.13.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:cd6b535d95eb1b44fe261a5336270fd898b966f431c14e173944484381c26d25
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:c2bd493e9345eccd124bbe0440e45cce8cf57588827d8eb9533f8bbadaa3ebbe
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:74e6cfa0bc481588cb8dc61c206ad567fbd2b28f43c2a86bee4a8e0eff55447c
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:28b5fff602ad6fb36d462e1905f453509ec22fc251a1e0cf64fedba86ff420d7
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:77bca19180f30ef08f4d1598ac6c365e190bb7909ac4a1fb35d9ba9ba8d6ff7a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a37529eff3656a570d796083404e5a3f63603afecfe9e738da1a243c0cb3d7c2
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:5d6354e65106460ab9285b42124faf288298387a8b16317565bd591f009c39a6
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:fb630d6f5817512733b8ea05f7147121928917a24ac8d98239f43d920880c359
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:6b2c0d1dd40e2c276684e17818b14375bc778e6c88df0bdc3cc24bfc9294804f
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:8883bf71abf39b4a6985940c287eb6ac7dc137d491264805d48bb112b65bd008
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:9ccdd7fe8223057a2234a3a2eb3b492e7b69e6ce714cc26a805fe2dd82935195
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:850901511bde065cf73e4c18d787dc10b6e91ca6139fc3ba68da28c41dd28236
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:688389195aebf3f43e74e40d591d2fd5f9ae41994d8ac2db661f292f6eda012a
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:2566b73e98b783871fb500bb1a8cc4e3d58a513c1bb18abfb525da0818e6138b
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:fd4ce524acc1769475079c7231f5ea79ac29551c513855ef34e5694d2c9fbbd3
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:a55f73d72ad2b07f01c49949cdb00969bfe4789375b3d7cdf9814b969fcc808d
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:7d085942e4da73c2ff4885c62f10b904dc70fd08e819e62f925eb1b4ddfa1554
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:8faab013568ee023968a5ec5b3a79977d7ddae3deafff6328e6fd56a27198e24
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:71abac4fb3d613edef7f3ba117a73bbeea61f341d6182e8310360a76576c789e
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:030c53690fe25e15bc9907f51c3f04fab39b4d085ff2098d26316ba32cc8b09f
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:8c19c7fa535f36bdce3b73cf3300e321909d44954cbc63a8aef6321c5f2fa693
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:59e9cc9cd57ba48ca7e87ae1e797a902f826bb890cb198ccf549f76ea581ba6e
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:bf25fce160dd7b8cce11e68a501006e4fa81a7858645038cdefdbba03df7ecdb
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:a3048eaa8924ce3571c6e78014438183fae70d9e8c833a7c07774c675104e3d4
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:5c2ee4136ee0014d1ef395b5e3bcab125c846f5e258ece3631d912f431c8d208
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:7c895490d49beb6bd724f47ad1b17ce501650557b0889314799c4072632f2b99
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f5b17f44c5ab72d6cf658f72f24f15e06a5ac9be1790336f7228ab8e7cfc15c3
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:cf3779c64bbaef1988886ed629c516e7d6ef7bdc73b9c95ad4c087b62e75750d
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:363a35c332fc80b068157951d5e50e6b7209947649cc64803ffc02726e7581ab
        version: 2.13.1
      entries:
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.13
    - currentCSV: rhods-operator.2.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
          createdAt: "2025-09-24T13:27:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.16
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:97e2bd9b587f08e135a9aeb9b3e0dc6eafa1a9bdacbb5ecb681ce9bd5aa37fc9
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:efd039012559786deb3c521a130886c265e88d635e08baace2e573a9df270134
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:d7d4fa406e0fcf0507894a7676532b27f45be742467e603a86f98ea5d2615df8
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:7b70af8847b0806d8b43c399e2b3109f016fd864f5e9d30c44e2baca5d1359dc
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f37e4048f3a152798286793f3abfc6ed814453fcbe2667255a7e78eee483c5a3
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:4a7599f8a866eb169c9a62885906adbf6df0417c0be15857df1eef20cd9b1be2
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:4c5ff3496b2a2a739939d94ee9dafc02b682100785d228dc2fde480fb597b7a5
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:4a212ec634225c14beac09be24ddf336e562f2aa9a13555fb1196f366ddae23c
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:2ba2b2c4db8bb334c50f4dfb54059f060361186900a44c06eed00b7a3c43977e
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:f738aab1eab25854c93e1d8d4d98100a8ae7bb45a6b83f0326774e4220b1183b
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:23d307a36b69e0df04f72a7d3b35e28d8417a8bbe23dba31e8e977569785c078
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:04f305c44413ae7dcb7017e53570ee49a509701792c5f50efadd64f47395730b
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:2eedd1e2465d2e78fafb39e234dbb413a82713f187c64d70a80340ec94807d30
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:5dcdcc2424602a69451f16d31dbfa1d43cb72c095ba561eb9076f0cd1e8182ed
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:11a301728573adf2b64ea72d0cb2d83ea5d4dbebea759f346e99f18c3d368c6e
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:27b113d96453c2054d4c965a963d2badba6daac235eddaf23234c76e87bcb069
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:c8456d98e90b6505957ab3686e9fd2f156e29f123c5558e581c206daf1e7d93a
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:de5f91180ead7d73a1825fe8b032fde9b8b01392569c9789f5ce1b4b9c08a98f
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a0f72ffefb2b74b488dd949493f5d295a39bb9c97f578bf219d0138601f65468
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:6cf74044ae8d5308a2dfe03fa5d81086c89302db7f4cdbe2f4174a1c48b77869
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:633984b538f027c93c5886ea1045dcbe81eeda74acf80001d5fc5f765bdbe0be
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:5192f1dbd2a9ab92ae390c4ae506efbed0970545b6122e95b014728ac937e777
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:65482864055021272a18b5b26792ab00cfa5fc9cc005d8d3a884cc82991506f3
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:22557a6fa52d2f311750a9ba253860f423ba697d26efa02ef8524a8258d2a909
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:ee01e89f98feb185f6cd59c564e590a13e4d8d9ea760cca8de51426eb71b83a1
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        version: 2.16.3
      entries:
      - name: rhods-operator.2.16.3
        version: 2.16.3
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.16
    - currentCSV: rhods-operator.2.19.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:c605e5fdb3d66302781146347df6da5a91dee97e4f8f5d9ac33f105b8cdee383
          createdAt: "2025-09-24T13:27:42Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:b30d60cd458133430d4c92bf84911e03cecd02f60e88a58d1c6c003543cf833a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:f0ee0d27bb784e289f7d88cc8ee0e085ca70e88a5d126562105542f259a1ac01
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:8ce44de8c683f198bf24ba36cd17e89708153d11f5b42c0a27e77f8fdb233551
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.19
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:8a6732ac3ca60f13dcb1fd8f186d2fc5dc601b9b5981a58f039af9e237f20044
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:b20ee16b776871d7c49b23f2bc91726812998561129f5dbd14eddbfe57964c93
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:c605e5fdb3d66302781146347df6da5a91dee97e4f8f5d9ac33f105b8cdee383
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:b30d60cd458133430d4c92bf84911e03cecd02f60e88a58d1c6c003543cf833a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:f49483df1d342d8e39f9a8085920aad79313a475df02a2292a136ecb784402b5
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:c8c23c07038f59db848dea0d4247b62840dbe58e1c581dc0722fb9d7f9be3b75
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:4e8f6c6ca0ed516b8310af0332637683b5bfc8f2946eb935c647698dffbbcc79
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhel8/mariadb-103@sha256:f0ee0d27bb784e289f7d88cc8ee0e085ca70e88a5d126562105542f259a1ac01
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:d24e277f468b0a48f38801e8e0bcfeb8264c3f7c128073ddaa17934b8572cc3c
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:520ed21b80e5487b3a0a13b7e5c7d791d4dd28e0b6799d57abb65d53fca5a03f
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:6275a19ef06114b33d79833a94ef10cc8203368eee98e5d415e73d20cae1db85
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:87685c5cab105d876ff0d14b0e24d8dc2a1ff5b55bebe36028c5f3376b2a5271
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:6ad35cf5c4c62ed5ad4e192afe8cebead9796cb24a964401c76ff8656c47742f
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:b6039c3c33a5dc345ceb8e757437bb13023d2bf85cc2b4c10b9ab3887e78e0c0
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:6f01ac7cc62ab961b9c8a630a558338a61f6d7902301c4a1ad337495ac7ce85a
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:4313cae03c041d461bc276abc23d706b582b5d16feab316796ca43478b9dc705
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:b24c25dfdcefb4ec26e7f311c00798df108f7685def2b1762b7624fdfb160ebf
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:bbef4839f24facce111d21aca43898dcf4e6cf57adb16200e3f1590bd5f89eae
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:49c02285fcef6949892cecca20eeee49d03a9bb4d17bb9ceff7c7a0dbbb8ab40
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel8@sha256:6166f525948a018c41d735014d97fdfa52024f7be31e18976d85d1a0d88901a0
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:baefbe4e7dd5aa74c71cb81da4a7a2b888a02d3eb1398ec4419bd8003b77f0c2
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:8ce44de8c683f198bf24ba36cd17e89708153d11f5b42c0a27e77f8fdb233551
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:af4721d61dfb61c75488a1bd6a5f62b23e20d7b7586a9cf51ba062ece2837715
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:bf6231bfdbe52b597d86e03e4730ec105ebe7a639fcd529268b7c0423f2a890c
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:71cd15f0b1acba3a48ebedff037bf8a846b426c18363a4de8896a376b6b7b3d5
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:53ac36baa374159b9065c718a9ede821bbb61d9ebe9502b2243e0a9f7aca0d16
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:d47514cbceb3cc96e9d33b7db6fc1df60893e5c4118bf6086cdd27f04c472773
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:876ca7ce7b0b975688fb4a77d6cd1a6f87cf9eb40c2a9a5b84d8cffdc80eeebd
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:de0e4b7fb822101198817e83cd7e9f4f7406e75ec91d07a833a27a417cffe66b
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:c1c97b6beb759d3f6f2dd53c4258e4ccbb3bf35cf11873bd446b135defaa69d6
        version: 2.19.3
      entries:
      - name: rhods-operator.2.19.3
        version: 2.19.3
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.19
    - currentCSV: rhods-operator.2.22.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
          createdAt: "2025-09-25T11:28:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.22
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:13f3679ab289eea9de54fe129a7c83f6c8ec5d26b61b430637a7c4cfcd31a74a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:d59ed40de694bc0e0564bab159a5b586cffa0721b6d761fb2d81dd42e6225d1c
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:d7c4ac937add511460a36375d9b6b3c475a3a9a7cc96c00df2fbad6219aee0e9
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:b2ca229bb4c4dbf04b96766b8af8d9f55f2edef59ec4110de0d20f8c857e7d27
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:9104cbc2d595933f0a2acb7313e41ed69905e76a9fae6bd9820e7717b56a632b
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f9965177de24f3165ec4c001c14cf1e2c2c01ffe847839af837c442e3c96beae
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:2f3f305c97e50057781553416fb4026cde306ead0fbe0213627b9b5bbcc8a0c3
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:b2c2ea9d35edbcd16e1ae559c88a453129b877fc049ab7a7cd6c2ebb8f0bcb59
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:6234e15f001122ab222481321153fb0eba7f4d86bf3dfd64edf3f6cecbe7d62f
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:5d79f0bcca0a213b53aee6d88834e503d343821596b857b13d3a6362de8ebeab
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:a44d32a18b1e906c1411cafc82102881573fcd04e040a4e2ba6bd1bd617b70fa
        - registry.redhat.io/rhel9/mariadb-105@sha256:5e77b33ce04c1eafc2e8840586ad6a5b51148dea8a46dda151071e28858331f0
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:7fd540586a1ca111370100a473923b3dc6cec4255e9e4f3654bfbabd9158a2bb
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:2fe9473cd1d6c445d2ca7cc357ee4bb009aa1838f9cb69ea418296b2a044816b
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:27415b1c763cf89b79391f0b039936c3e078a078bda4420f13bf10ef9be31d5e
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:7b3d7787ba95156de9c284ad78f2d63a0d3bcb525f4d8d11ebf0203d4c5c1fe0
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:2065b96bd6f0a830eab66e36217a927448cfc5a7dabf2a5e7d907aa6c40657a4
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e3b90530c6e38f6940d7f9408e9fa2c81eb06a857ec2b0bab5d006b1e9a63a4f
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:e6fbe6723b4801f6854303b3fd7d4a4a36727bdd28dc28002b7de4929739ccc0
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/ubi9/toolbox@sha256:ea4ce8cb8f59cf1b0709f1606d7cab074c3806d4016964eedc0555410cb5dfeb
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:5914f71e37c815ce2567826896e051303bdef52fae1acf8d02e672fc24c7eb9d
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:1709fa3c79aad4ba7eb9be8299949396092c8e20210124e0c0936385bc04e839
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:cb5be3dd7098f4c8d66f9e85ad04656460df6102652deddc6cccb884a12be733
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:2b3aace0d80474470e71c2c8291b25738e096109229fe6d5d3e06e9885ee58b6
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:2f1a0a0803b2f5118b167e9d9c2e86660add5f79ee6f47044f63b0e8abe0b75f
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:ccba67a6a5aa3403e0018f29dcdfef483220dea8065d8b572e80c653ca992add
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:463e9be9c4409ba1e5a42713bd65844a023e7a3e9032d4e862edbaa5fe35300a
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:abe1f9dd11c7313fdd55b678be511a5cdc3546654a68e2b53d42c1013fec8d00
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:bd6d1599b4ff9eb890be9808014715b28ca19247222d65f40df024db4dd3525a
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:74d4e0174655cf6f6273857325a1e714a1531c5a06ad38bf139bee9c7d8f5f19
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:32c9511846077ef6035eb19be7653f6a78e0f207eb420ab93b18f742ab5779b6
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        version: 2.22.2
      entries:
      - name: rhods-operator.2.22.2
        version: 2.22.2
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.22
    - currentCSV: rhods-operator.2.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:b3563e201bf2720f60e73c0b04cb70566ea52eda7d09a28753f3a1487993c2e8
          createdAt: "2024-08-01T06:23:03Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.8.0 <2.8.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:eb4da3a645b53acfd6ce8758c043ae19b4126e62bcca09165114a141c2c8d014
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:92d9e4b53d580df6cc1f14a849b477749a1306941a2e8f9d1b6160f4911c3cff
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:b3563e201bf2720f60e73c0b04cb70566ea52eda7d09a28753f3a1487993c2e8
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:5f39b11f0cd6db2d7d6e51ee666c030e2643aee92a090b49656c1e7dfd8a5430
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:90833e85e0de3d7de870074c8b04319cd11b913b9652af6a5727aae41a91458c
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:91525fac5fadece75fbdafc40b94445fb92c88cfc8c2f06b596f8c6377e59c9e
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:ab8792fcd16b1aa0065444ba54335cb3e747379740c3ec2d56beb46ff80cb044
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:987c768c0688431a53e48a80591f2a41e074994169f9e691b28b89528d8e736f
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:1a6e29214e06a2665baadbed598ba758d6141bfa522984eb05f05a80558a403c
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:850b78a201effcafe33705dcd2670f1c470d8b46ba6e307032403627b55a7cfe
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:080377d42d782bf525b4072384d0cf8409b48019b3748ca7c855f340bcb94251
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:b6ecb3858c201498024e3bfc4a93e630699be4b1fdc3a65fd2d70e4b375eb9c2
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:df2edcaf4402b11ce24439263956e3b0cc43efbac2fdf152afacb1a398dcc59a
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:889f046b255b13d99801108588d3abf577611b828657b5240a18ea21f0e3cbbe
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:1c49888ba2e9449c04d5588040c099290fce4793c2ba8874c1037316e550d6e1
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:0b8665e8978deb73be2cc65830f08a74a5a48f3e3ee7b2b817f7bbacba257d22
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:68e1331f13dfaf6641a1e7d54f4cf29c0ef42663abc98b6dc69167892d2d8adb
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:06a6d5cbbe80d49bca9e4db9929571a69e35ae7951fcae84dcd6460f7ddf622c
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:9799003449bfb61ae8f59efd2a611b0da3bccadd9b20cc1519f6c9279d085337
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:9c90873fbd5af8fac8efc3c7c27fba742a06fad2d47f336319dd05d2b116867b
        version: 2.8.4
      entries:
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.8
    defaultChannel: stable
    packageName: rhods-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: ansible-automation-platform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aap-operator.v2.4.0-0.1758079821
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.1-0.1635279521 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the hub, including all hosted Ansible
              content, secrets and the database. By default, a persistent volume claim
              will be created using the default StorageClass on your cluster to store
              the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous hub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Automation Hub
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-24

          ## How to Upgrade

          Starting with AAP 2.4 in the May 2025 async release, the managed database was upgraded to PostgreSQL 15. It is recommended to take a backup of your deployment before upgrading.

          To take a backup, create AutomationControllerBackup and AutomationHubBackup objects for your deployments.
          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel.

          Note: Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8-operator@sha256:5b957031609c8aabf7575e675bc636c1152babc9564d503802301aff4642a404
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-runner-rhel8@sha256:e14286dc9605be3a8c6cb144d0b2e3f3bb3211129005009e7566ea6c16c91f52
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8@sha256:7ae8db2b5d12b619f40db1e581db1e982a695bc1a29a2628222fff09e7c25249
        - registry.redhat.io/ansible-automation-platform-24/ee-minimal-rhel8@sha256:7670e41037f023e3fc6ee95d2342787a80077273ab8d30ba6804029ff05b4edc
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-ui-rhel8@sha256:d229d4e09b2b788d1ba3413ea7ec15374bd1613d507c285ac1ddc10570b9473f
        - registry.redhat.io/ansible-automation-platform-24/hub-web-rhel8@sha256:6ff1f602f0d9cad201b3addb65acc570a9d527de005ca85fc861d3cb98138710
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8@sha256:bfd13d6b39c5ec3a67adbafb05a0482328f0b9b97a517b45c093f6c35a3a92d3
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8@sha256:04fb5cb6b0d7c79cf5d4b754fe917224ddfcb2b9a4994462db404c833ed69483
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-rhel8-operator@sha256:b120e6a5ef376662b4324f0458eae31627087dc91fc613788143bc3954a15318
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-24/ee-supported-rhel8@sha256:f1384eb558f7da1e4a35d51408b3cf5a5a9be08f4d50503f473b26909bc246fc
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8@sha256:cccb4db08735551c697a7b2cbf75812f34500e691f9cdbaea02b30edea292e47
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8-operator@sha256:5253c9fd587ff2ca94d251f2b5fc7f284080adc5eb64643df5299beb2015592f
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8-operator@sha256:3c28a1326ecb7495952a5bb8be13b0972a4378af82611b0b9d990c5fa9fa82c3
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8-operator@sha256:e20106a8a04abd568a4c56861cea39ae444c0bd204bc807ac01f1705e12937d3
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        version: 2.4.0+0.1758079821
      entries:
      - name: aap-operator.v2.4.0-0.1702352195
        version: 2.4.0+0.1702352195
      - name: aap-operator.v2.4.0-0.1707506910
        version: 2.4.0+0.1707506910
      - name: aap-operator.v2.4.0-0.1708974895
        version: 2.4.0+0.1708974895
      - name: aap-operator.v2.4.0-0.1711590540
        version: 2.4.0+0.1711590540
      - name: aap-operator.v2.4.0-0.1713466670
        version: 2.4.0+0.1713466670
      - name: aap-operator.v2.4.0-0.1715898994
        version: 2.4.0+0.1715898994
      - name: aap-operator.v2.4.0-0.1717558474
        version: 2.4.0+0.1717558474
      - name: aap-operator.v2.4.0-0.1718152210
        version: 2.4.0+0.1718152210
      - name: aap-operator.v2.4.0-0.1720482179
        version: 2.4.0+0.1720482179
      - name: aap-operator.v2.4.0-0.1725256739
        version: 2.4.0+0.1725256739
      - name: aap-operator.v2.4.0-0.1726111413
        version: 2.4.0+0.1726111413
      - name: aap-operator.v2.4.0-0.1727257740
        version: 2.4.0+0.1727257740
      - name: aap-operator.v2.4.0-0.1730152858
        version: 2.4.0+0.1730152858
      - name: aap-operator.v2.4.0-0.1733185647
        version: 2.4.0+0.1733185647
      - name: aap-operator.v2.4.0-0.1733943951
        version: 2.4.0+0.1733943951
      - name: aap-operator.v2.4.0-0.1737661496
        version: 2.4.0+0.1737661496
      - name: aap-operator.v2.4.0-0.1741194076
        version: 2.4.0+0.1741194076
      - name: aap-operator.v2.4.0-0.1742444794
        version: 2.4.0+0.1742444794
      - name: aap-operator.v2.4.0-0.1743629049
        version: 2.4.0+0.1743629049
      - name: aap-operator.v2.4.0-0.1746132523
        version: 2.4.0+0.1746132523
      - name: aap-operator.v2.4.0-0.1749069319
        version: 2.4.0+0.1749069319
      - name: aap-operator.v2.4.0-0.1750877268
        version: 2.4.0+0.1750877268
      - name: aap-operator.v2.4.0-0.1753232084
        version: 2.4.0+0.1753232084
      - name: aap-operator.v2.4.0-0.1755833431
        version: 2.4.0+0.1755833431
      - name: aap-operator.v2.4.0-0.1758079821
        version: 2.4.0+0.1758079821
      name: stable-2.4
    - currentCSV: aap-operator.v2.4.0-0.1758080293
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.1-0.1635283332 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the hub, including all hosted Ansible
              content, secrets and the database. By default, a persistent volume claim
              will be created using the default StorageClass on your cluster to store
              the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous hub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Automation Hub
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-24

          ## How to Upgrade

          Starting with AAP 2.4 in the May 2025 async release, the managed database was upgraded to PostgreSQL 15. It is recommended to take a backup of your deployment before upgrading.

          To take a backup, create AutomationControllerBackup and AutomationHubBackup objects for your deployments.
          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel.

          Note: Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8@sha256:bfd13d6b39c5ec3a67adbafb05a0482328f0b9b97a517b45c093f6c35a3a92d3
        - registry.redhat.io/ansible-automation-platform-24/ee-minimal-rhel8@sha256:7670e41037f023e3fc6ee95d2342787a80077273ab8d30ba6804029ff05b4edc
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8@sha256:04fb5cb6b0d7c79cf5d4b754fe917224ddfcb2b9a4994462db404c833ed69483
        - registry.redhat.io/ansible-automation-platform-24/ee-supported-rhel8@sha256:f1384eb558f7da1e4a35d51408b3cf5a5a9be08f4d50503f473b26909bc246fc
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-rhel8-operator@sha256:b120e6a5ef376662b4324f0458eae31627087dc91fc613788143bc3954a15318
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8-operator@sha256:5253c9fd587ff2ca94d251f2b5fc7f284080adc5eb64643df5299beb2015592f
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8-operator@sha256:3c28a1326ecb7495952a5bb8be13b0972a4378af82611b0b9d990c5fa9fa82c3
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8@sha256:7ae8db2b5d12b619f40db1e581db1e982a695bc1a29a2628222fff09e7c25249
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8@sha256:cccb4db08735551c697a7b2cbf75812f34500e691f9cdbaea02b30edea292e47
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8-operator@sha256:5b957031609c8aabf7575e675bc636c1152babc9564d503802301aff4642a404
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8-operator@sha256:e20106a8a04abd568a4c56861cea39ae444c0bd204bc807ac01f1705e12937d3
        - registry.redhat.io/ansible-automation-platform-24/hub-web-rhel8@sha256:6ff1f602f0d9cad201b3addb65acc570a9d527de005ca85fc861d3cb98138710
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-runner-rhel8@sha256:e14286dc9605be3a8c6cb144d0b2e3f3bb3211129005009e7566ea6c16c91f52
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-ui-rhel8@sha256:d229d4e09b2b788d1ba3413ea7ec15374bd1613d507c285ac1ddc10570b9473f
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        version: 2.4.0+0.1758080293
      entries:
      - name: aap-operator.v2.4.0-0.1702352614
        version: 2.4.0+0.1702352614
      - name: aap-operator.v2.4.0-0.1707507304
        version: 2.4.0+0.1707507304
      - name: aap-operator.v2.4.0-0.1708975279
        version: 2.4.0+0.1708975279
      - name: aap-operator.v2.4.0-0.1711591330
        version: 2.4.0+0.1711591330
      - name: aap-operator.v2.4.0-0.1713467078
        version: 2.4.0+0.1713467078
      - name: aap-operator.v2.4.0-0.1715899379
        version: 2.4.0+0.1715899379
      - name: aap-operator.v2.4.0-0.1717558885
        version: 2.4.0+0.1717558885
      - name: aap-operator.v2.4.0-0.1718152680
        version: 2.4.0+0.1718152680
      - name: aap-operator.v2.4.0-0.1720482740
        version: 2.4.0+0.1720482740
      - name: aap-operator.v2.4.0-0.1725257213
        version: 2.4.0+0.1725257213
      - name: aap-operator.v2.4.0-0.1726111804
        version: 2.4.0+0.1726111804
      - name: aap-operator.v2.4.0-0.1727259094
        version: 2.4.0+0.1727259094
      - name: aap-operator.v2.4.0-0.1730153340
        version: 2.4.0+0.1730153340
      - name: aap-operator.v2.4.0-0.1733186325
        version: 2.4.0+0.1733186325
      - name: aap-operator.v2.4.0-0.1733945743
        version: 2.4.0+0.1733945743
      - name: aap-operator.v2.4.0-0.1737662040
        version: 2.4.0+0.1737662040
      - name: aap-operator.v2.4.0-0.1741194557
        version: 2.4.0+0.1741194557
      - name: aap-operator.v2.4.0-0.1742445409
        version: 2.4.0+0.1742445409
      - name: aap-operator.v2.4.0-0.1743630487
        version: 2.4.0+0.1743630487
      - name: aap-operator.v2.4.0-0.1746133932
        version: 2.4.0+0.1746133932
      - name: aap-operator.v2.4.0-0.1749069693
        version: 2.4.0+0.1749069693
      - name: aap-operator.v2.4.0-0.1750878254
        version: 2.4.0+0.1750878254
      - name: aap-operator.v2.4.0-0.1753232791
        version: 2.4.0+0.1753232791
      - name: aap-operator.v2.4.0-0.1755833968
        version: 2.4.0+0.1755833968
      - name: aap-operator.v2.4.0-0.1758080293
        version: 2.4.0+0.1758080293
      name: stable-2.4-cluster-scoped
    - currentCSV: aap-operator.v2.5.0-0.1758147230
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727257740 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation gateway - Unify the management of AAP from a central location.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner


          ## New in AAP 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-25

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.5

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console. For example, change the channel from `stable-2.4` to `stable-2.5`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.

          After changing to the new 2.5 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place.

          #### Create the AnsibleAutomationPlatform custom resource

          Once that is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: myaap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8@sha256:3a989f148648a5a62e687058ec3ca772fa22aa10faec89453626d2b17b3fd9c2
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8-operator@sha256:43e731b9a435b3a412108170b4b39e6b6215be51d362a9f03c33fa412e97da53
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8-operator@sha256:8d8182e3a87c55eacc222e71df4a9bf7a2e51102d913a5824cc91e36056d4249
        - registry.redhat.io/ansible-automation-platform-25/hub-web-rhel8@sha256:e96832efcc03c6754031cb70f56c808135f62d1b80a882dd6d6fc9e1a4cfc69b
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8@sha256:697343aaea153ab22d7efa8f1bb5aed37a55cbb353fc206f4a355e8b878ae418
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8@sha256:ac0de2b7d354eb0c15a6c12861a4de3577a982c84af661fbab25517d8105c085
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8@sha256:aed3973eb953dda3571260dcf1b33bc7dcbc18c41e3350b8c361b7e635fb9527
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-runner-rhel8@sha256:ace1d2ab49ba5ff9f11ab18e8edce27c9e73ae7298bf7335569fa1740a44899b
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8-operator@sha256:acc0fcd509eeac46db307ab744dd0a0c26770a93e495cf2ec0ccb61a91c6fbc5
        - registry.redhat.io/ansible-automation-platform-25/ee-minimal-rhel8@sha256:41f944eb1450db56c3b14fc7baece8f148d5f10886a57bad730bec60fb28ac04
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8-operator@sha256:07aa7e257d04ae0eda6fcaa4af9121f1aece439231fce4f924e6467da91f0022
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-ui-rhel8@sha256:9ab35e80a6172018ed3f5cd179e0bcb6befc1378583d9e73c3a64a7cf9e8050f
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8@sha256:9f2867603b9747c2fe37862c2c5efac4a4530902f323a8bf7cc0717de4cdb5aa
        - registry.redhat.io/ansible-automation-platform-25/gateway-proxy-rhel8@sha256:789d55f355fa0d90f9509a912d199123bfc5a4041930a72a8fd4ed1d819f0d84
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8-operator@sha256:503fc7a20e2dd7de2cb55b9be5f97efa0d2b31f054ce326f23f3a11b4d683e64
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-rhel8-operator@sha256:74be9182723b41f87b3d13b3d86b6a9049bba88e2902cc9453264aa8a0f584e4
        - registry.redhat.io/ansible-automation-platform-25/ee-supported-rhel8@sha256:52d2c73fa2015af5f1b9a88ee48085c7c7bab44330fc9f95914f48b58e4b48bc
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-chatbot-rhel8@sha256:9a37c9739c488ccf211ec5efc6debb045223d508c9cb3fa5a8256eca98ff2f0c
        version: 2.5.0+0.1758147230
      entries:
      - name: aap-operator.v2.5.0-0.1727240614
        version: 2.5.0+0.1727240614
      - name: aap-operator.v2.5.0-0.1727875185
        version: 2.5.0+0.1727875185
      - name: aap-operator.v2.5.0-0.1728520175
        version: 2.5.0+0.1728520175
      - name: aap-operator.v2.5.0-0.1729741547
        version: 2.5.0+0.1729741547
      - name: aap-operator.v2.5.0-0.1731456196
        version: 2.5.0+0.1731456196
      - name: aap-operator.v2.5.0-0.1733193164
        version: 2.5.0+0.1733193164
      - name: aap-operator.v2.5.0-0.1733969526
        version: 2.5.0+0.1733969526
      - name: aap-operator.v2.5.0-0.1736893446
        version: 2.5.0+0.1736893446
      - name: aap-operator.v2.5.0-0.1737101025
        version: 2.5.0+0.1737101025
      - name: aap-operator.v2.5.0-0.1737675968
        version: 2.5.0+0.1737675968
      - name: aap-operator.v2.5.0-0.1739248042
        version: 2.5.0+0.1739248042
      - name: aap-operator.v2.5.0-0.1740093573
        version: 2.5.0+0.1740093573
      - name: aap-operator.v2.5.0-0.1740773472
        version: 2.5.0+0.1740773472
      - name: aap-operator.v2.5.0-0.1741368605
        version: 2.5.0+0.1741368605
      - name: aap-operator.v2.5.0-0.1742434024
        version: 2.5.0+0.1742434024
      - name: aap-operator.v2.5.0-0.1743660124
        version: 2.5.0+0.1743660124
      - name: aap-operator.v2.5.0-0.1746137767
        version: 2.5.0+0.1746137767
      - name: aap-operator.v2.5.0-0.1747343762
        version: 2.5.0+0.1747343762
      - name: aap-operator.v2.5.0-0.1749074128
        version: 2.5.0+0.1749074128
      - name: aap-operator.v2.5.0-0.1749604727
        version: 2.5.0+0.1749604727
      - name: aap-operator.v2.5.0-0.1750901111
        version: 2.5.0+0.1750901111
      - name: aap-operator.v2.5.0-0.1753900156
        version: 2.5.0+0.1753900156
      - name: aap-operator.v2.5.0-0.1755835086
        version: 2.5.0+0.1755835086
      - name: aap-operator.v2.5.0-0.1758147230
        version: 2.5.0+0.1758147230
      name: stable-2.5
    - currentCSV: aap-operator.v2.5.0-0.1758147817
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727259094 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation gateway - Unify the management of AAP from a central location.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner


          ## New in AAP 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-25

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.5

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console. For example, change the channel from `stable-2.4` to `stable-2.5`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.

          After changing to the new 2.5 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place.

          #### Create the AnsibleAutomationPlatform custom resource

          Once that is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: myaap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8-operator@sha256:acc0fcd509eeac46db307ab744dd0a0c26770a93e495cf2ec0ccb61a91c6fbc5
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-ui-rhel8@sha256:9ab35e80a6172018ed3f5cd179e0bcb6befc1378583d9e73c3a64a7cf9e8050f
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8-operator@sha256:43e731b9a435b3a412108170b4b39e6b6215be51d362a9f03c33fa412e97da53
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8-operator@sha256:8d8182e3a87c55eacc222e71df4a9bf7a2e51102d913a5824cc91e36056d4249
        - registry.redhat.io/ansible-automation-platform-25/hub-web-rhel8@sha256:e96832efcc03c6754031cb70f56c808135f62d1b80a882dd6d6fc9e1a4cfc69b
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-chatbot-rhel8@sha256:9a37c9739c488ccf211ec5efc6debb045223d508c9cb3fa5a8256eca98ff2f0c
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8@sha256:3a989f148648a5a62e687058ec3ca772fa22aa10faec89453626d2b17b3fd9c2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/ee-minimal-rhel8@sha256:41f944eb1450db56c3b14fc7baece8f148d5f10886a57bad730bec60fb28ac04
        - registry.redhat.io/ansible-automation-platform-25/gateway-proxy-rhel8@sha256:789d55f355fa0d90f9509a912d199123bfc5a4041930a72a8fd4ed1d819f0d84
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8@sha256:aed3973eb953dda3571260dcf1b33bc7dcbc18c41e3350b8c361b7e635fb9527
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8@sha256:697343aaea153ab22d7efa8f1bb5aed37a55cbb353fc206f4a355e8b878ae418
        - registry.redhat.io/ansible-automation-platform-25/ee-supported-rhel8@sha256:52d2c73fa2015af5f1b9a88ee48085c7c7bab44330fc9f95914f48b58e4b48bc
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8@sha256:ac0de2b7d354eb0c15a6c12861a4de3577a982c84af661fbab25517d8105c085
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-rhel8-operator@sha256:74be9182723b41f87b3d13b3d86b6a9049bba88e2902cc9453264aa8a0f584e4
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-runner-rhel8@sha256:ace1d2ab49ba5ff9f11ab18e8edce27c9e73ae7298bf7335569fa1740a44899b
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8@sha256:9f2867603b9747c2fe37862c2c5efac4a4530902f323a8bf7cc0717de4cdb5aa
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8-operator@sha256:07aa7e257d04ae0eda6fcaa4af9121f1aece439231fce4f924e6467da91f0022
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8-operator@sha256:503fc7a20e2dd7de2cb55b9be5f97efa0d2b31f054ce326f23f3a11b4d683e64
        version: 2.5.0+0.1758147817
      entries:
      - name: aap-operator.v2.5.0-0.1727241223
        version: 2.5.0+0.1727241223
      - name: aap-operator.v2.5.0-0.1727875896
        version: 2.5.0+0.1727875896
      - name: aap-operator.v2.5.0-0.1728520687
        version: 2.5.0+0.1728520687
      - name: aap-operator.v2.5.0-0.1729742145
        version: 2.5.0+0.1729742145
      - name: aap-operator.v2.5.0-0.1731456854
        version: 2.5.0+0.1731456854
      - name: aap-operator.v2.5.0-0.1733193761
        version: 2.5.0+0.1733193761
      - name: aap-operator.v2.5.0-0.1733970210
        version: 2.5.0+0.1733970210
      - name: aap-operator.v2.5.0-0.1736894220
        version: 2.5.0+0.1736894220
      - name: aap-operator.v2.5.0-0.1737101781
        version: 2.5.0+0.1737101781
      - name: aap-operator.v2.5.0-0.1737676792
        version: 2.5.0+0.1737676792
      - name: aap-operator.v2.5.0-0.1739248649
        version: 2.5.0+0.1739248649
      - name: aap-operator.v2.5.0-0.1740094176
        version: 2.5.0+0.1740094176
      - name: aap-operator.v2.5.0-0.1740774104
        version: 2.5.0+0.1740774104
      - name: aap-operator.v2.5.0-0.1741369251
        version: 2.5.0+0.1741369251
      - name: aap-operator.v2.5.0-0.1742434756
        version: 2.5.0+0.1742434756
      - name: aap-operator.v2.5.0-0.1743660958
        version: 2.5.0+0.1743660958
      - name: aap-operator.v2.5.0-0.1746138413
        version: 2.5.0+0.1746138413
      - name: aap-operator.v2.5.0-0.1747345055
        version: 2.5.0+0.1747345055
      - name: aap-operator.v2.5.0-0.1749074612
        version: 2.5.0+0.1749074612
      - name: aap-operator.v2.5.0-0.1749607543
        version: 2.5.0+0.1749607543
      - name: aap-operator.v2.5.0-0.1750901870
        version: 2.5.0+0.1750901870
      - name: aap-operator.v2.5.0-0.1753900674
        version: 2.5.0+0.1753900674
      - name: aap-operator.v2.5.0-0.1755835623
        version: 2.5.0+0.1755835623
      - name: aap-operator.v2.5.0-0.1758147817
        version: 2.5.0+0.1758147817
      name: stable-2.5-cluster-scoped
    - currentCSV: aap-operator.v2.6.0-0.1758684783
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727257740 <2.6.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Automation Gateway - Unify the management of AAP from a central location.
            * Automation Controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private Automation Hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Ansible Lightspeed: Deploy and manage the lifecycle of an Ansible Lightspeed application.
            * Platform Resource Operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## New in AAP 2.6 and 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.


          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-26

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5 and above, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.6

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console.

          **For customers on AAP 2.4:** Change the channel directly from `stable-2.4` to `stable-2.6` to upgrade directly to AAP 2.6, which will be released in October.

          **For customers on AAP 2.5:** Change the channel from `stable-2.5` to `stable-2.6`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.

          After changing to the new 2.6 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place. When upgrading from AAP 2.4 to 2.6, your external authentication configuration in Automation Controller will automatically be migrated to the new unified Platform UI.

          **Database:** AAP 2.6 continues to use PostgreSQL 15.

          #### Create the AnsibleAutomationPlatform custom resource (Required for upgrades from 2.4 only)

          **If upgrading from AAP 2.4:** Once the operator upgrade is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created.

          **If upgrading from AAP 2.5:** This step is not required as your AnsibleAutomationPlatform custom resource already exists.

          Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          Next, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda (lightspeed is disabled by default):


          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: aap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-26/gateway-rhel9-operator@sha256:fa1b102a0690767a1e08935a0e164c1eabb2d4cf9afccd310262ee9185544de7
        - registry.redhat.io/ansible-automation-platform-26/hub-rhel9@sha256:8041b141cd8921aaaa20cce15eab4c6189da9f449d99057189284891c071ac3b
        - registry.redhat.io/ansible-automation-platform-26/gateway-proxy-rhel9@sha256:13284d20c6fb4f2de24ec1715e7abdb5c4b0e4298c5034a7795f494b93634a02
        - registry.redhat.io/rhel9/postgresql-15@sha256:5e225c1646251d2509b3a6e45fad23abfb11dd5c45d3224b59178f90deb38a70
        - registry.redhat.io/ansible-automation-platform-26/controller-rhel9-operator@sha256:863e77e52ab01cec2a71672a224080a59d963d4fc08b9ed6f3cc7950729b1746
        - registry.redhat.io/ansible-automation-platform-26/hub-rhel9-operator@sha256:18114aab6db0d00780fdeaf9efa74ee44cab94824c2fd35f4b9998b6378ad8b3
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-rhel9-operator@sha256:cfbb98e17297134c187673f355342ac5ad60a56d73f4b0c918be2ff244286cc0
        - registry.redhat.io/ansible-automation-platform-26/platform-resource-rhel9-operator@sha256:969bbd033fbba33624329ee9347b3b89fb5a1f64e71892f5c6c7d74ed14a6d39
        - registry.redhat.io/ansible-automation-platform-26/controller-rhel9@sha256:64f5e77a7842965e9c6f7bc7990354b71bb19ffb06ab8ef7793dde956cb6c0f0
        - registry.redhat.io/ansible-automation-platform-26/ee-minimal-rhel9@sha256:261c324cac081e47d559c5fccbd48e7ec8f5bf64e3f27287fbe4fb8aff6b7d99
        - registry.redhat.io/ansible-automation-platform-26/hub-web-rhel9@sha256:aa4041672542a7f828d8a51c79e9c4dcea2a185388e4f20f828104db37e03a9b
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-rhel9@sha256:01e56bee772d15b120d6e4b02a59024b21eb8b266127b1a9b7f2c535ecdb2855
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-ui-rhel9@sha256:5b5f38a5f76781a042eec9cbf70b8be026878e4fa61c7fb90d1b0ad64bd21592
        - registry.redhat.io/ubi9/ubi-minimal@sha256:7c5495d5fad59aaee12abc3cbbd2b283818ee1e814b00dbc7f25bf2d14fa4f0c
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-chatbot-rhel9@sha256:8901fba14b21569b02108d2ab9050d558dee29f8177fae9d1bfcd5de3f4ce7c1
        - registry.redhat.io/ansible-automation-platform-26/mcp-tools-rhel9@sha256:0c1f6f6958f2570c8cd073287555d28c70ef9dded702ec014d3ead8ce572463e
        - registry.redhat.io/ansible-automation-platform-26/platform-resource-runner-rhel9@sha256:9983043386df38a17ff6fffef64dc187691e0b1fefec01eef3af39aa08e532ef
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-rhel9-operator@sha256:bfd55bffd8239e224d4ebba500d8517e8e66915078d6b59244f5e910fc5e10c8
        - registry.redhat.io/ansible-automation-platform-26/gateway-rhel9@sha256:c85dbbf6beaa264afec480dc1f06576aada531cdfd15f247f4db37dc9702be33
        - registry.redhat.io/rhel9/redis-6@sha256:0962d18bb5292c0e524354f319b63287c2298a4a2fd34556c68fa285df381427
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-rhel9@sha256:c5d4e2e76944e0991729e5c3b99f43b7f29ea08ad112e5e6513feb8fb1d899a1
        - registry.redhat.io/ansible-automation-platform-26/ee-supported-rhel9@sha256:2534d07f1ae41af32adb681e283cbaea465cbe0f9d234ed5478a7bfad7676e7f
        version: 2.6.0+0.1758684783
      entries:
      - name: aap-operator.v2.6.0-0.1758684783
        version: 2.6.0+0.1758684783
      name: stable-2.6
    - currentCSV: aap-operator.v2.6.0-0.1758685241
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727259094 <2.6.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Automation Gateway - Unify the management of AAP from a central location.
            * Automation Controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private Automation Hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Ansible Lightspeed: Deploy and manage the lifecycle of an Ansible Lightspeed application.
            * Platform Resource Operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## New in AAP 2.6 and 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.


          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-26

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5 and above, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.6

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console.

          **For customers on AAP 2.4:** Change the channel directly from `stable-2.4` to `stable-2.6` to upgrade directly to AAP 2.6, which will be released in October.

          **For customers on AAP 2.5:** Change the channel from `stable-2.5` to `stable-2.6`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.

          After changing to the new 2.6 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place. When upgrading from AAP 2.4 to 2.6, your external authentication configuration in Automation Controller will automatically be migrated to the new unified Platform UI.

          **Database:** AAP 2.6 continues to use PostgreSQL 15.

          #### Create the AnsibleAutomationPlatform custom resource (Required for upgrades from 2.4 only)

          **If upgrading from AAP 2.4:** Once the operator upgrade is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created.

          **If upgrading from AAP 2.5:** This step is not required as your AnsibleAutomationPlatform custom resource already exists.

          Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          Next, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda (lightspeed is disabled by default):


          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: aap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-26/gateway-rhel9@sha256:c85dbbf6beaa264afec480dc1f06576aada531cdfd15f247f4db37dc9702be33
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-ui-rhel9@sha256:5b5f38a5f76781a042eec9cbf70b8be026878e4fa61c7fb90d1b0ad64bd21592
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-rhel9-operator@sha256:cfbb98e17297134c187673f355342ac5ad60a56d73f4b0c918be2ff244286cc0
        - registry.redhat.io/ansible-automation-platform-26/controller-rhel9@sha256:64f5e77a7842965e9c6f7bc7990354b71bb19ffb06ab8ef7793dde956cb6c0f0
        - registry.redhat.io/ansible-automation-platform-26/ee-supported-rhel9@sha256:2534d07f1ae41af32adb681e283cbaea465cbe0f9d234ed5478a7bfad7676e7f
        - registry.redhat.io/ansible-automation-platform-26/ee-minimal-rhel9@sha256:261c324cac081e47d559c5fccbd48e7ec8f5bf64e3f27287fbe4fb8aff6b7d99
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-rhel9@sha256:c5d4e2e76944e0991729e5c3b99f43b7f29ea08ad112e5e6513feb8fb1d899a1
        - registry.redhat.io/ansible-automation-platform-26/hub-web-rhel9@sha256:aa4041672542a7f828d8a51c79e9c4dcea2a185388e4f20f828104db37e03a9b
        - registry.redhat.io/ansible-automation-platform-26/platform-resource-runner-rhel9@sha256:9983043386df38a17ff6fffef64dc187691e0b1fefec01eef3af39aa08e532ef
        - registry.redhat.io/ansible-automation-platform-26/gateway-rhel9-operator@sha256:fa1b102a0690767a1e08935a0e164c1eabb2d4cf9afccd310262ee9185544de7
        - registry.redhat.io/ansible-automation-platform-26/platform-resource-rhel9-operator@sha256:969bbd033fbba33624329ee9347b3b89fb5a1f64e71892f5c6c7d74ed14a6d39
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-chatbot-rhel9@sha256:8901fba14b21569b02108d2ab9050d558dee29f8177fae9d1bfcd5de3f4ce7c1
        - registry.redhat.io/ansible-automation-platform-26/mcp-tools-rhel9@sha256:0c1f6f6958f2570c8cd073287555d28c70ef9dded702ec014d3ead8ce572463e
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-rhel9-operator@sha256:bfd55bffd8239e224d4ebba500d8517e8e66915078d6b59244f5e910fc5e10c8
        - registry.redhat.io/ansible-automation-platform-26/hub-rhel9-operator@sha256:18114aab6db0d00780fdeaf9efa74ee44cab94824c2fd35f4b9998b6378ad8b3
        - registry.redhat.io/rhel9/postgresql-15@sha256:5e225c1646251d2509b3a6e45fad23abfb11dd5c45d3224b59178f90deb38a70
        - registry.redhat.io/ubi9/ubi-minimal@sha256:7c5495d5fad59aaee12abc3cbbd2b283818ee1e814b00dbc7f25bf2d14fa4f0c
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-rhel9@sha256:01e56bee772d15b120d6e4b02a59024b21eb8b266127b1a9b7f2c535ecdb2855
        - registry.redhat.io/ansible-automation-platform-26/gateway-proxy-rhel9@sha256:13284d20c6fb4f2de24ec1715e7abdb5c4b0e4298c5034a7795f494b93634a02
        - registry.redhat.io/ansible-automation-platform-26/hub-rhel9@sha256:8041b141cd8921aaaa20cce15eab4c6189da9f449d99057189284891c071ac3b
        - registry.redhat.io/rhel9/redis-6@sha256:0962d18bb5292c0e524354f319b63287c2298a4a2fd34556c68fa285df381427
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/ansible-automation-platform-26/controller-rhel9-operator@sha256:863e77e52ab01cec2a71672a224080a59d963d4fc08b9ed6f3cc7950729b1746
        version: 2.6.0+0.1758685241
      entries:
      - name: aap-operator.v2.6.0-0.1758685241
        version: 2.6.0+0.1758685241
      name: stable-2.6-cluster-scoped
    defaultChannel: stable-2.6-cluster-scoped
    packageName: ansible-automation-platform-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/component: kmm-hub
      app.kubernetes.io/name: kmm-hub
      app.kubernetes.io/part-of: kmm
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: kernel-module-management-hub
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kernel-module-management-hub.v2.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:888a3acad339cf38a324c718e0cbe1237ca5bf2f4636b01d001752658e4cca1b
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.0.2'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:888a3acad339cf38a324c718e0cbe1237ca5bf2f4636b01d001752658e4cca1b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62c44057a27abd41cf3daf60b8736da59a82dadb65fa6c32bef65d19bd49ea49
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:192e08b121a04ae69aea60c35199e8f5e3c42305e1ebd024a605a371de4ab472
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:94b4ddc45cb34d4ffe794a51b99c868cb709f90cc4eb8b28284d70c526a0a226
        version: 2.0.2
      entries:
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.0
    - currentCSV: kernel-module-management-hub.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:d1c1490d4c20dc2ec31effa1af067676c4d264e3b7d4670fc11c2baf8a400001
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.1.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:14e100e5853444114029116255ebfb7a54e5ed51503ba969740ad48b5ca80f70
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:1a9a19df17b464d1708b2905d03b8726e4480ca1f11aa97a15ff4c049a24ea4b
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:d1c1490d4c20dc2ec31effa1af067676c4d264e3b7d4670fc11c2baf8a400001
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:611556d999ea97f9dbb78fec5963baa02de7a26626d7096ba63b177e3bcf9707
        version: 2.1.1
      entries:
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.1
    - currentCSV: kernel-module-management-hub.v2.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:90c1946b3f577cde0435402b04baa52dc54e5aab3bb84bd834ac026a57453296
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.2.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:b1d98c0f14448ae8eb80821cbec236150c338674aae2a2b37f8d0ab0d5f1430c
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:53fddf0b226eb579c71c86112a8d9b6b1d90a1498d6f6985932b16b8be532f35
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:c3dc9ad4ebc68cd85a074b520903a9310e72a90a73bdf4aa7c42b7ed66b7bcba
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:90c1946b3f577cde0435402b04baa52dc54e5aab3bb84bd834ac026a57453296
        version: 2.2.1
      entries:
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.2
    - currentCSV: kernel-module-management-hub.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:f1d8f3c95f621c3faf41867c5dd4bc13fa2b9b291422d6184e6e42f1d1460259
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.3.0'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:f1d8f3c95f621c3faf41867c5dd4bc13fa2b9b291422d6184e6e42f1d1460259
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:c5005b74402f680d11c9ea682a9a0d3f74b6ab0d01d25298d232f910f1a29c34
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:f9fde334af68220943017a7491130e6a82f15e63522f5bb3931aa7d9b00ccc40
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:06ac089cc89b303e525fd71d6439b958f34ab944a72209f1cc8cb562ecfd1112
        version: 2.3.0
      entries:
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.3
    - currentCSV: kernel-module-management-hub.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm-hub",
                    "app.kubernetes.io/name": "kmm-hub",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:2efb81137ad2981b994c62d460c917749308133a7dcad5b9a2b251e18f3052ec
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        version: 2.4.1
      entries:
      - name: kernel-module-management-hub.v2.4.1
        version: 2.4.1
      - name: kernel-module-management-hub.v2.4.0
        version: 2.4.0
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.4
    - currentCSV: kernel-module-management-hub.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm-hub",
                    "app.kubernetes.io/name": "kmm-hub",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:2efb81137ad2981b994c62d460c917749308133a7dcad5b9a2b251e18f3052ec
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        version: 2.4.1
      entries:
      - name: kernel-module-management-hub.v2.4.1
        version: 2.4.1
      - name: kernel-module-management-hub.v2.4.0
        version: 2.4.0
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: kernel-module-management-hub
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: cert-utils-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-utils-operator.v1.3.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/cert-utils-operator@sha256:0290e7b2800a5aa8eb57b0268908f2a20a93b16403c76df4c61021048161d6ff
          createdAt: "2024-01-23T12:29:04Z"
          description: Set of utilities for TLS certificates
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-utils-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/cert-utils-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Cert utils operator is a set of functionalities around certificates packaged in a [Kubernetes operator](https://github.com/operator-framework/operator-sdk).

          Certificates are assumed to be available in a [secret](https://kubernetes.io/docs/concepts/configuration/secret/) of type `kubernetes.io/tls` (other types of secrets are *ignored* by this operator).
          By convention this type of secrets have three optional entries:

          1. `tls.key`: the private key of the certificate.
          2. `tls.crt`: the actual certificate.
          3. `ca.crt`: the CA bundle that validates the certificate.

          The functionalities are the following:

          1. [Ability to populate route certificates](https://github.com/redhat-cop/cert-utils-operator#Populating-route-certificates)
          2. [Ability to create java keystore and truststore from the certificates](https://github.com/redhat-cop/cert-utils-operator#Creating-java-keystore-and-truststore)
          3. [Ability to show info regarding the certificates](https://github.com/redhat-cop/cert-utils-operator#Showing-info-on-the-certificates)
          4. [Ability to alert when a certificate is about to expire](https://github.com/redhat-cop/cert-utils-operator#Alerting-when-a-certificate-is-about-to-expire)
          5. [Ability to inject ca bundles in ValidatingWebhookConfiguration, MutatingWebhookConfiguration, CustomResourceDefinition object](https://github.com/redhat-cop/cert-utils-operator#CA-injection)

          All these feature are activated via opt-in annotations.
        displayName: Cert Utils Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Cert Utils Operator
          url: https://github.com/redhat-cop/cert-utils-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/cert-utils-operator
        maintainers:
        - email: rspazzoli@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/cert-utils-operator@sha256:0290e7b2800a5aa8eb57b0268908f2a20a93b16403c76df4c61021048161d6ff
        version: 1.3.12
      entries:
      - name: cert-utils-operator.v1.3.12
        version: 1.3.12
      - name: cert-utils-operator.v1.3.11
        version: 1.3.11
      - name: cert-utils-operator.v1.3.10
        version: 1.3.10
      - name: cert-utils-operator.v1.3.9
        version: 1.3.9
      - name: cert-utils-operator.v1.3.8
        version: 1.3.8
      - name: cert-utils-operator.v1.3.7
        version: 1.3.7
      - name: cert-utils-operator.v1.3.6
        version: 1.3.6
      - name: cert-utils-operator.v1.3.5
        version: 1.3.5
      - name: cert-utils-operator.v1.3.4
        version: 1.3.4
      - name: cert-utils-operator.v1.3.3
        version: 1.3.3
      - name: cert-utils-operator.v1.3.2
        version: 1.3.2
      - name: cert-utils-operator.v1.3.1
        version: 1.3.1
      - name: cert-utils-operator.v1.3.0
        version: 1.3.0
      - name: cert-utils-operator.v1.2.0
        version: 1.2.0
      - name: cert-utils-operator.v1.1.0
        version: 1.1.0
      - name: cert-utils-operator.v1.0.7
        version: 1.0.7
      - name: cert-utils-operator.v1.0.6
        version: 1.0.6
      - name: cert-utils-operator.v1.0.5
        version: 1.0.5
      - name: cert-utils-operator.v1.0.4
        version: 1.0.4
      - name: cert-utils-operator.v1.0.3
        version: 1.0.3
      - name: cert-utils-operator.v1.0.2
        version: 1.0.2
      - name: cert-utils-operator.v1.0.1
        version: 1.0.1
      - name: cert-utils-operator.v1.0.0
        version: 1.0.0
      - name: cert-utils-operator.v0.2.2
        version: 0.2.2
      - name: cert-utils-operator.v0.2.1
        version: 0.2.1
      - name: cert-utils-operator.v0.2.0
        version: 0.2.0
      - name: cert-utils-operator.v0.1.1
        version: 0.1.1
      - name: cert-utils-operator.v0.1.0
        version: 0.1.0
      - name: cert-utils-operator.v0.0.3
        version: 0.0.3
      - name: cert-utils-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cert-utils-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ses-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ses-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ses-controller:1.1.1
          createdAt: "2025-09-19T19:37:25Z"
          description: AWS SES controller is a service controller for managing SES
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConfigurationSet represents the state of an AWS ses ConfigurationSet
              resource.
            displayName: ConfigurationSet
            kind: ConfigurationSet
            name: configurationsets.ses.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Simple Email Service (SES) resources in AWS from within your Kubernetes cluster.

          **About Amazon SES**

          Amazon Simple Email Service (SES) is an email platform that provides an easy, cost-effective way for you to send and receive email using your own email addresses and domains.
          For example, you can send marketing emails such as special offers, transactional emails such as order confirmations, and other types of correspondence such as newsletters. When you use Amazon SES to receive mail, you can develop software solutions such as email autoresponders, email unsubscribe systems, and applications that generate customer support tickets from incoming emails.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SES
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ses
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SES Developer Resources
          url: https://aws.amazon.com/ses/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ses maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ses-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ses-controller.v1.1.1
        version: 1.1.1
      - name: ack-ses-controller.v1.1.0
        version: 1.1.0
      - name: ack-ses-controller.v1.0.13
        version: 1.0.13
      - name: ack-ses-controller.v1.0.12
        version: 1.0.12
      - name: ack-ses-controller.v1.0.11
        version: 1.0.11
      - name: ack-ses-controller.v1.0.10
        version: 1.0.10
      - name: ack-ses-controller.v1.0.9
        version: 1.0.9
      - name: ack-ses-controller.v1.0.8
        version: 1.0.8
      - name: ack-ses-controller.v1.0.7
        version: 1.0.7
      - name: ack-ses-controller.v1.0.6
        version: 1.0.6
      - name: ack-ses-controller.v1.0.5
        version: 1.0.5
      - name: ack-ses-controller.v1.0.3
        version: 1.0.3
      - name: ack-ses-controller.v1.0.2
        version: 1.0.2
      - name: ack-ses-controller.v1.0.1
        version: 1.0.1
      - name: ack-ses-controller.v1.0.0
        version: 1.0.0
      - name: ack-ses-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-ses-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco Networking - Compute Solutions
      provider-url: ""
    name: ucs-ci-solutions-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ucs-ci-solutions-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intersight.cisco.com/v1alpha1",
                "kind": "UCSCISolutions",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ucs-ci-solutions-operator",
                    "app.kubernetes.io/instance": "ucscisolutions-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ucscisolutions",
                    "app.kubernetes.io/part-of": "ucs-ci-solutions-operator"
                  },
                  "name": "ucscisolutions-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: docker.io/pkoppa/ucs-ci-solutions-operator:latest
          createdAt: "2023-07-15T15:19:39Z"
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: UCSCISolutions
            name: ucscisolutions.intersight.cisco.com
            version: v1alpha1
        description: The operator is used to automate the creation of  Pools, Policies
          and Server Profile Templates required to deploy Cisco UCS Converged Infrastructure
          Solution in Intersight.
        displayName: UCS-CI-Solutions-Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ucscisolutions
        - intersight
        - ucs
        - ci
        - flexpod
        - flashstack
        links:
        - name: Ucs Ci Solutions Operator
          url: https://ucs-ci-solutions-operator.domain
        maintainers:
        - email: pkoppa@cisco.com
          name: Paniraj Koppa
        - email: chrobrie@cisco.com
          name: Chris O'Brien
        maturity: alpha
        provider:
          name: Cisco Networking - Compute Solutions
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - docker.io/pkoppa/ucs-ci-solutions-operator:latest
        version: 0.0.1
      entries:
      - name: ucs-ci-solutions-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ucs-ci-solutions-operator
    provider:
      name: Cisco Networking - Compute Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-diagnostics
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-diagnostics.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Diagnostics",
                "metadata": {
                  "name": "diagnostics-sample"
                },
                "spec": {
                  "agentDiagnosticsService": {
                    "port": 8080
                  },
                  "auth": {
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "wapdbDiagnosticsService": {
                    "dumpedTables": "[ \"--regexpr cluster_logs_week_*\", \"current_week_cluster_logs\", \"config_metadata_dbs\", \"config_metadata_dbs_id_seq\", \"config_metadata_tags\", \"config_tags\", \"csv_reports_meta\", \"custom_import_handler_dbs\", \"data_point_db_models\", \"db_test_entries\", \"license_server_metadata\", \"operations\", \"result_file_metadata_dbs\", \"result_metadata_dbs\", \"result_metadata_tags\", \"result_tags\", \"session_meta\", \"session_types\", \"sessions\", \"stat_view_query_data\", \"test_run_configs\", \"test_run_infos\", \"test_runs\", \"wireless_globalsettings\", \"wireless_sessionconfigs\", \"wireless_templates\"]",
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:37:10Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Diagnostics
            name: diagnostics.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Diagnostics Operator
        displayName: Keysight WAP Diagnostics Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Diagnostics
          url: https://keysight-wap-diagnostics.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-diagnostics@sha256:97bf8affa7bf1f3a391f5f036313e7cd8196be2eb8e95debeab06c1b6c964c5e
        version: 0.1.21
      entries:
      - name: keysight-wap-diagnostics.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-diagnostics
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JFrog
      provider-url: ""
    name: openshiftartifactoryha-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: artifactory-ha-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftArtifactoryHa",
                "metadata": {
                  "name": "openshiftartifactoryha"
                },
                "spec": {
                  "artifactory-ha": {
                    "artifactory": {
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE",
                      "node": {
                        "replicaCount": 2,
                        "waitForPrimaryStartup": {
                          "enabled": false
                        }
                      }
                    },
                    "databaseUpgradeReady": true,
                    "database": {
                      "driver": "OVERRIDE",
                      "password": "OVERRIDE",
                      "type": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "nginx": {
                      "http": {
                        "externalPort": 80,
                        "internalPort": 8080
                      },
                      "https": {
                        "externalPort": 443,
                        "internalPort": 8443
                      },
                      "service": {
                        "ssloffload": false
                      },
                      "tlsSecretName": "OVERRIDE"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "waitForDatabase": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/artifactory-operator@sha256:de69bb16052d5ac6fed9882eb65d49446b6fa1cc5ba830dd2a39e882712b1347
          createdAt: 2024-03-23 00:00:00+00:00
          description: JFrog Artifactory Enterprise deploys Artifactory in a high
            availability environment across multiple pods
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Artifactory HA Instances
            displayName: Artifactory HA
            kind: OpenshiftArtifactoryHa
            name: openshiftartifactoryhas.charts.helm.k8s.io
            version: v1
        description: |2+
           ## Breaking change
          Please update to version 1.1.10 of the operator.
          Version 1.1.8 and 1.1.9 have issues. Please skip these versions.
          If you are using operator version < 1.2.0 there is a issue with upgrade where artifactory version is not updated if you update operator version. As workaround you can update the Custom Resource OpenshiftArtifactoryHa which has hardcored image value.After upgrading to version 1.2.0 you can remove the below parameters from your CR so that future upgrade works without workaround. ``` artifactory-ha.artifactory.image.registry artifactory-ha.artifactory.image.repository artifactory-ha.artifactory.image.tag ```
          ## Overview
          Openshift Operator to deploy JFrog Artifactory Enterprise into your Openshift cluster.
          ## Security Context Constraints
          To deploy this helm chart you will need to be a cluster admin w/ access to the anyuid scc and add the operator service account to the anyuid scc.
          ```
          oc adm policy add-scc-to-user anyuid -z artifactory-ha-operator -n openshift-operators
          ```
          Add the service account for the Artifactory chart to deploy successfully:
          ``` oc adm policy add-scc-to-user anyuid -z openshiftartifactoryha-artifactory-ha -n artifactory ```
          ### Usage

          An external DB is required. The operator will not deploy a DB but will require you to specify the configuration values to connect to it.

          Search for JFrog and click JFrog Artifactory Enterprise Operator to install.

          Go to the Installed Operators.

          Wait for the JFrog Artifactory Enterprise Operator to complete the installation.

          Open the Operator and click on the provided API: Artifactory HA.

          Click Create New Instance and provide the following parameters for your DB configuration:

          ```
          DATABASE_TYPE
          DATABASE_DRIVER
          DATABASE_URL
          DATABASE_USER
          DATABASE_PASSWORD
          ```
          Master key and Join key must be supplied. To generate a new key for each run the command below:

          ```
          # Create a key
          export JOIN_KEY=$(openssl rand -hex 32)
          echo ${JOIN_KEY}
          ```

          To use TLS you will need to first create a k8s tls secret to store your .crt and .key file into.
          Then supply the value of this k8s secret into the TLS_SECRET field.
          ``` oc create secret tls my_tls_secret --cert=tls.crt --key=tls.key --namespace=my_namespace ```
          Click Create for Artifactory Enterprise to deploy into OpenShift and connect to it on the external IP exposed by the load balancer.
          ### Create a route
          To expose Artifactory from Openshift we recommend you create a new route in Openshift.
          You can either use the oc command line tool or the Openshift web console to generate a new route.
          Depending upon where you terminate TLS you will need to either specify pass through or edge.
          Command Line (Edge):
          ``` oc create route edge --service=openshiftartifactory-ha --cert=tls.crt --key=tls.key --ca-cert=ca.crt --hostname=www.example.com ```
          Or for more information visit the official Openshift documentation on routes here:
          https://docs.openshift.com/container-platform/4.6/networking/routes/route-configuration.html

        displayName: JFrog Artifactory Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Artifact Repository
          url: https://jfrog.com/artifactory/
        - name: Artifactory Video
          url: https://www.youtube.com/watch?v=r2_A5CPo43U
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: alpha
        provider:
          name: JFrog
        relatedImages:
        - registry.redhat.io/rhel8/nginx-120@sha256:87f658744af695273c434fac8246b716728b691a17d5d0c5b89247362e61e45c
        - registry.connect.redhat.com/jfrog/artifactory-operator@sha256:de69bb16052d5ac6fed9882eb65d49446b6fa1cc5ba830dd2a39e882712b1347
        - registry.connect.redhat.com/jfrog/artifactory-pro@sha256:0986ca71419ae628892db11e7ca7da5ac1c60219c144ce9aeb2de6e683a45262
        version: 1.3.3
      entries:
      - name: artifactory-ha-operator.v1.3.3
        version: 1.3.3
      name: alpha
    - currentCSV: artifactory-ha-operator.v1.1.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftArtifactoryHa",
                "metadata": {
                  "name": "openshiftartifactoryha"
                },
                "spec": {
                  "artifactory-ha": {
                    "artifactory": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/artifactory-pro",
                        "tag": "7.27.3-1"
                      },
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE",
                      "uid": "1000721030",
                      "node": {
                        "replicaCount": 2,
                        "waitForPrimaryStartup": {
                          "enabled": false
                        }
                      }
                    },
                    "databaseUpgradeReady": true,
                    "database": {
                      "driver": "OVERRIDE",
                      "password": "OVERRIDE",
                      "type": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "initContainerImage": "registry.connect.redhat.com/jfrog/init@sha256:164b41720a37f75c6129c8a28d05cc56f821022313d44ff918cce4a2146d93e8",
                    "nginx": {
                      "uid": "1000720104",
                      "gid": "1000720107",
                      "http": {
                        "externalPort": 80,
                        "internalPort": 8080
                      },
                      "https": {
                        "externalPort": 443,
                        "internalPort": 8443
                      },
                      "image": {
                        "registry": "registry.redhat.io",
                        "repository": "rhel8/nginx-116",
                        "tag": "latest"
                      },
                      "service": {
                        "ssloffload": false
                      },
                      "tlsSecretName": "OVERRIDE"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "waitForDatabase": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/artifactory-operator@sha256:e292553e1b6e792ff19276c7f6565d7ed61dd95289eb176066329954bc93e371
          createdAt: 2020-03-25 00:00:00+00:00
          description: JFrog Artifactory Enterprise deploys Artifactory in a high
            availability environment across multiple pods
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Artifactory HA Instances
            displayName: Artifactory HA
            kind: OpenshiftArtifactoryHa
            name: openshiftartifactoryhas.charts.helm.k8s.io
            version: v1
        description: |2+
           ## Breaking change
          Please update to version 1.1.10 of the operator.
          Version 1.1.8 and 1.1.9 have issues. Please skip these versions.

          ## Overview
          Openshift Operator to deploy JFrog Artifactory Enterprise into your Openshift cluster.
          ## Security Context Constraints
          To deploy this helm chart you will need to be a cluster admin w/ access to the anyuid scc and add the operator service account to the anyuid scc.
          ```
          oc adm policy add-scc-to-user anyuid -z artifactory-ha-operator -n openshift-operators
          ```
          Add the service account for the Artifactory chart to deploy successfully:
          ``` oc adm policy add-scc-to-user anyuid -z openshiftartifactoryha-artifactory-ha -n artifactory ```
          ### Usage

          An external DB is required. The operator will not deploy a DB but will require you to specify the configuration values to connect to it.

          Search for JFrog and click JFrog Artifactory Enterprise Operator to install.

          Go to the Installed Operators.

          Wait for the JFrog Artifactory Enterprise Operator to complete the installation.

          Open the Operator and click on the provided API: Artifactory HA.

          Click Create New Instance and provide the following parameters for your DB configuration:

          ```
          DATABASE_TYPE
          DATABASE_DRIVER
          DATABASE_URL
          DATABASE_USER
          DATABASE_PASSWORD
          ```
          Master key and Join key must be supplied. To generate a new key for each run the command below:

          ```
          # Create a key
          export JOIN_KEY=$(openssl rand -hex 32)
          echo ${JOIN_KEY}
          ```

          To use TLS you will need to first create a k8s tls secret to store your .crt and .key file into.
          Then supply the value of this k8s secret into the TLS_SECRET field.
          ``` oc create secret tls my_tls_secret --cert=tls.crt --key=tls.key --namespace=my_namespace ```
          Click Create for Artifactory Enterprise to deploy into OpenShift and connect to it on the external IP exposed by the load balancer.
          ### Create a route
          To expose Artifactory from Openshift we recommend you create a new route in Openshift.
          You can either use the oc command line tool or the Openshift web console to generate a new route.
          Depending upon where you terminate TLS you will need to either specify pass through or edge.
          Command Line (Edge):
          ``` oc create route edge --service=openshiftartifactory-ha --cert=tls.crt --key=tls.key --ca-cert=ca.crt --hostname=www.example.com ```
          Or for more information visit the official Openshift documentation on routes here:
          https://docs.openshift.com/container-platform/4.6/networking/routes/route-configuration.html

        displayName: JFrog Artifactory Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Artifact Repository
          url: https://jfrog.com/artifactory/
        - name: Artifactory Video
          url: https://www.youtube.com/watch?v=r2_A5CPo43U
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: stable
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/init@sha256:164b41720a37f75c6129c8a28d05cc56f821022313d44ff918cce4a2146d93e8
        - registry.connect.redhat.com/jfrog/artifactory-operator@sha256:e292553e1b6e792ff19276c7f6565d7ed61dd95289eb176066329954bc93e371
        - registry.connect.redhat.com/jfrog/artifactory-pro@sha256:7289dde95d1e4d8e161a28cffd9c4a0cc96f92f7536d069ab63ec304a92254fd
        - registry.redhat.io/rhel8/nginx-116@sha256:d71d368d6094a5777000a8d99164959ccb97fca355e855e24bd92c13a8df04f6
        version: 1.1.18
      entries:
      - name: artifactory-ha-operator.v1.1.18
        version: 1.1.18
      name: stable
    defaultChannel: alpha
    packageName: openshiftartifactoryha-operator
    provider:
      name: JFrog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: advanced-cluster-management
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: advanced-cluster-management.v2.10.9
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-08-06T03:17:11Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0 <2.10.9'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/internal-objects: '[]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterHub defines the configuration for an instance
              of the MultiCluster Hub
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:95df18f4479b7f555a41508486624bde714778730921c20c665e9f23baadbabf
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:8f1bb3a8422d29960d8b90dd24a813a56875abf1f0393f78b862eaa48ddda783
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:50331cdd61332c8f4128b3dab237255463b307c5c40b0607730a7d3b33b9c6b5
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:5b162aeab4aa51185c880696d1472e152a20ada926a3baea0fb8494793ad238e
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:0ba2471b7a67532bdb90e02904671f0e724cdb39f74a45392ae6c693c25b354d
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:4b16da3070e30d879421aaa0d4a73aa2a1b0b852f62b8faabd801c03f22d5319
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:ced03229f208739867a43b9ed9bd158f19f4242dd1f5f045d3e5db81b2b1d171
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:4ad5ccf9ce682024b6e3025a9bcadd6f32692aa18d892afd670eacb7b4031860
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:1c2dbcf179508ea1c602f7817f85b0d3f58d6d85ae93ccc523e88a3936c9c690
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:23713ebb12d510d37de7e060d5ce4ff2198990f83cebbd2302e74bd83a55cecd
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:1aaba9ca65e363c7f9c07ef2f66f23cb8fff677204a56373634d319113d2c84b
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:e3c80e8e3af5c02c202dfb288c63b20c44c08f2a0474d2568cfb0b878f7fe888
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:9e78db48ff0b0fc0ac0347ca055a4b66e2f917073499adc6e2db2a9c16a70e01
        - registry.redhat.io/rhacm2/console-rhel9@sha256:1538a3a935aa01d8800d8145c45037eaab6987ad6e6e927017ca1cfea4bb111c
        - registry.redhat.io/rhel9/postgresql-13@sha256:31c77c19480ab82cb33525d32419b68cf59daccf0b9936e88e8238e48a2d47e5
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:fbc5b1456a37951c9f70f64357919222739f4a7e533fb6dde5b93672535fc87e
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:95890bb1aad544d97ae1b80e801a800c95ef51e1034b60e680b67bdb2f5aeb69
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:9e931cab5fcbf3df0ffe6fff8d12303800423ace3d749ee0e33b55300c816b19
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:dcd6a420467929b9af8c52c77c8717d22f3977e6b6b392b5dc90cbb04005b9f1
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:14094f09dfb166a6b342bd91e2db9162f0de88d55655ccce44d345047c700187
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:23b4d7fbc776c63dcd7007442a557124581939ce110257822e8233261d31fb7d
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:bbe57064cb1c40c35929526a082f11b2da6fd3746b6eef8af51ca63a4fd7d98b
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:46e3b8ccf3e7069b44eb3ad407329c03048b5219ef2886e237e1ab0dc3f7bae4
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:bb4b72c1510da84aad2d67516c58f37495d7682c51c8f2088e2c4a350fa36c50
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:dcbc4c5e2d46ec2b4911d8cd9f96a4c141bf7dd18b79aa4c25b77141e4f60b48
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:69a8faa2f4da2281c6c124508634628510d91ad40f725a9b41790e0821d64c14
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:500718f834e05c7a4531eb4f9d8ce560053c5125f76f6921c38c9cf150046a23
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:9db39230c528bfc675bc422a7acd40c394169b9632b088ca16358751994c9cac
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:335c903dde6f21e9c54f025c45666f6b4b6e7d455dfd5f85a6373b3947069f0a
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:3ccc45b0e9483f33be0d3b32a6d46fa794976131ad2f9861d96e319c8e35b9be
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:2f68a16a14f192b2cceaa07ed4840c08fbf413718a39bd584cacde866be7f2b1
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:9c4eb0ef8758d943796e13f65bc92b190db614eabd0fe7375ebfecc90e96e32c
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:da09087ba7d558611d733551b6ebc27ccf441581fc264efe7a5933db3816ef89
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:e02992cec8edcc9374c4cd30dfd7bccbda77d6d5cdd309862ab5265494f1278e
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:f4474ed21a71ad490778a6a378a856bec1884b4b9dc2c0e6c6140cc6c9d03b61
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:805e443fbfca4f816dc2dcea71f4cf723fd753ccea6d13f04d25945f29cb47dc
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:d8b8d130f75647cfeb9036db50d66fc4f1fca56fead6add4ba24c52f9213543e
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:13c33a997be0ddef59d1fb31835751367d4f901a1a6eac41284931bf7988035c
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:814f34f9e55b59d96021d3160b1dad7a68ffe208b3f4a999840b228aa7b3300b
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:47a8fa9541275f849a24e2ef1df87b7e9051dedd0ee8bd833d60defbd8724b46
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:3803d82d2b829c0e0883c0504176a96d2012cb7096cfcace316831a2b55bca3f
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:de68d598587a967d4b3f5e0e28c4675f74fd79cae7a5bb197c1b4b44bb4b31d9
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:35ff861165fccd848b8d328fae46cb8b3ac0c853c63062ec8672a6d41a23f333
        - registry.redhat.io/rhacm2/iam-policy-controller-rhel9@sha256:8c2b356bd9ff4ebfba9059a5f20cad8bd336af237d84cdc21f5c0aa35fabeabe
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:d989e9d2078f8a1e4b18de64f917e6d5e290c5cac769899f829b5f6bbfe6f0f2
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:87093f85c6235c99d4cca9d503b0e4198e3731e5fc3171506485091570e70bb5
        version: 2.10.9
      entries:
      - name: advanced-cluster-management.v2.10.9
        version: 2.10.9
      - name: advanced-cluster-management.v2.10.8
        version: 2.10.8
      - name: advanced-cluster-management.v2.10.7
        version: 2.10.7
      - name: advanced-cluster-management.v2.10.6
        version: 2.10.6
      - name: advanced-cluster-management.v2.10.5
        version: 2.10.5
      - name: advanced-cluster-management.v2.10.4
        version: 2.10.4
      - name: advanced-cluster-management.v2.10.3
        version: 2.10.3
      - name: advanced-cluster-management.v2.10.2
        version: 2.10.2
      - name: advanced-cluster-management.v2.10.1
        version: 2.10.1
      - name: advanced-cluster-management.v2.10.0
        version: 2.10.0
      name: release-2.10
    - currentCSV: advanced-cluster-management.v2.11.8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-08-01T05:25:37Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.10.0 <2.11.8'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:956d549e64c412f86f2083c8ddf97fea269d7806d8826aab93178bec9ed5d7af
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:9d821f4373722d9cb2b679cf5793ef22cb0b9ebcbb5a38060504c93c05c492c4
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:aef5ffbaa578d006bc7caec8fb35446dc263a51813f84c756aa57903b428329b
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:d248a5fad4ae3ee3210eb6dc4f58406d2f1e8f5c390f3c491040b437fe1adeba
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:447b1f6067a82476e14dfbb74a8f324fb86956420e55575f0017b91784f303a5
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:a081f531efb5b54d7d14d9ba0519552ecfc87c3a80d9d3ec4e598d2b445fc689
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:ffd3d9c363d8e15232bd6d12c5f1a77f7cb1ca18f94b7aea79d226568be4586e
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:95890bb1aad544d97ae1b80e801a800c95ef51e1034b60e680b67bdb2f5aeb69
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:f5c56a7381eb80c02a7c140f7c3f0ed08bb93697cd51732ee57498c080d8d3e5
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:8d15be5f40d135fd7069ac768dcf9188d8d8299c65db56df04e66f8a28941966
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:94c89ebd9257797ddc264511399689feb70d56e53332d9894c6d95fbdc51de40
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:6d4e3114494e051b3b875eeeae4bad4d48665fbcbbd3574a98f87b5a5742b41d
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:9be2d97fa3c1140db7d86157503a96ca2d3bdfb57125ab2230c8cf0649ea1e90
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:d4feeb3d3322210765d916ef1f49b47a1527f1c61041f94834c4395319a9ec32
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:49f0ec1aced72e6a2f0c862b0c6c0cd430d997a09ceb9eb0a7d25aaa1d9798b7
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:a9e0533af92ebadca27c9fe93fe4db9c68a1a8bc0060f4d13f67f26eb5691cb3
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:b3e0427341e3b1cbb1fe9166c56184eb0b0ec526d8a3992306e4861f65e4f0c0
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:ded47519fcf30e60cf38d44a101494b0318757969675782365de88b51cab24f0
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:968eb25c5cfcbdcea87b92dbff734d01b5954c13c13f5e794ad1db7bd886e2b4
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:d7552b2a3c864a3fc0d279a2519535e170fd8608195fe1f9abb4b538dfe2a3df
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:d6666c572294c7c85990304d21eaae3040b9c4e47fcb7b446289cb09aabc2d26
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:4a62d8706772ede443cffa005b984c26d137c35a33aafedaa7f01050b618c2a2
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:d4bfef313ff9044ae36aa5f832651a4dd09228d4b9b186803009aafe9774ee2c
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:1e8a26ebfcc33e60f60da4ba8d58cf7c152a99752eac6ebc9e114187e4cc2267
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:4dce3cebeb4ea0f1d3fc1c7e036dc6a3fd09e02f8f2ae4f3a753d9f4f97defe7
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:b713264826124df2c2b3a99dc268a3f3bd7eb24a9010b93c9876d83656e70e1a
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:e36832aa4852ea807d44f2b1f2c402bdaac6b7e2246bf9dd811b41f0b5e8e647
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:eaa719f9d9db8f91eced81e0e164bbc12bb502bf4760fc8c673cc542b4414771
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:45d4404abe02596611b29316bd10985f7a0684b1d305a57d3b24068ac776cf6e
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:251d3ee881c3e902e87f1752648ab084e8780aef79eb88a253ba79f984a01bcf
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:e347cd0f57e0691c61263e6554495c76ce56c7443bce3fd010cf76d068083fee
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:ad10d15efb6833b2daaf153210a080420395a7debfcb873dd757bb924197a654
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:36c90a52b12af24faaf28301d0d1a0c522723bd6cadd7391e5ee1f43f3019eef
        - registry.redhat.io/rhacm2/console-rhel9@sha256:2ed2f4add387caf790b2b41a65a3986e0046e08d0b8406ad17e4f0440ba0ac4a
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:5ba07381111651906ac3c93344e705152eb006826cb36bf5c05bd88723c4c999
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:7d39d0c29b02c399ffdcd83278693a8dd9c5ca3931e660af056030b126d10055
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:eec0e9bd6ea6f146939516da6c7ca9c3755888d96e81954f2f282fae63c1847c
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:827ca1beb707c02908914c517c8ab14b77b1c65ca7690dbbe1bcff600b7551d5
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:4e0c2cf5229fd68102a9e6d0f3155c651b1159b7af7b057151ce52990f747b08
        - registry.redhat.io/rhel9/postgresql-13@sha256:31c77c19480ab82cb33525d32419b68cf59daccf0b9936e88e8238e48a2d47e5
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:ef66846a3f178234618fd1e265067f863bbb56f14cbca0dfc96b0d741885dc9b
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:4b16da3070e30d879421aaa0d4a73aa2a1b0b852f62b8faabd801c03f22d5319
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:028b7258ec88d04397798552ed19a459110660ddbdf3ab375a7a1f74049c455a
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:147c567ce95673ba7151bcd290474756e13018ebc9f7275d61677ffe586a8a76
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:2878b5c41f8d58e986e86803bfd7eb3a05ccf23f7e49ebba56ad30245609b23d
        version: 2.11.8
      entries:
      - name: advanced-cluster-management.v2.11.8
        version: 2.11.8
      - name: advanced-cluster-management.v2.11.7
        version: 2.11.7
      - name: advanced-cluster-management.v2.11.6
        version: 2.11.6
      - name: advanced-cluster-management.v2.11.5
        version: 2.11.5
      - name: advanced-cluster-management.v2.11.4
        version: 2.11.4
      - name: advanced-cluster-management.v2.11.3
        version: 2.11.3
      - name: advanced-cluster-management.v2.11.2
        version: 2.11.2
      - name: advanced-cluster-management.v2.11.1
        version: 2.11.1
      - name: advanced-cluster-management.v2.11.0
        version: 2.11.0
      name: release-2.11
    - currentCSV: advanced-cluster-management.v2.12.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-21T21:16:17Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <2.12.4'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:b83e630e77bd011c374d73d801f10a3230f6798ede10d7f39bf2e8ff6437cf2c
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:3d3260ff206c5ee249443b1f8da49562584f7ee1abd284691b42f1728b6ed5dc
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:b85f3de5231f2f1bc3b4e9c95917f562f39e09dc9b7db35690024163e3a5afe5
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:6f2c4f1abc4388e76414544e65f08238c96fc4d6fbf71fa79330d9dca117d221
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:ec65a5744945ca20f804d94999a4002efab9ddad360a58f9498800bffa9fa0f4
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:187879345174e515863bc1b29a961a0a0d9ad44006a6e12de715b4508dec38a8
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:93df0586f9281fa2879eed0fdf2fb115be355d572cc24dc5c5c9a92137044877
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:58f08a93cd35bd2ac19f94a1779205c81a2fc3a552d1e77b7c659fd51c9dd34f
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:bd24cfabb81b464a31c3d8c21fa6371e46b83094d5c750acc33346203fe15e84
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:c2259b90decf24f8fb1cb058d4751c75fd1558e3c9fd1079857ad9c47ca9e8d2
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:22e262c0353c6f0d5d1b916074836c53437e22697f67c85a43e6d717daf651fa
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:c402a4ac61e11dc9e45fdc7bc7fcb5bf5ca3475a66d782aea60d9f6524e51a24
        - registry.redhat.io/rhel8/postgresql-13@sha256:826a1a75f4186f36217b5c1f1aa270838c4b55b1021d33a4876566bfc3b8e629
        - registry.redhat.io/rhacm2/console-rhel9@sha256:d239df44dafac3bde3eb7ae8c80cb342fa4bb4eabe22f27d2c92d283324e6e10
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:9c297b4c1f9ba17908594f1843ea6de6c5a5e0f3500258c23fbb8b5684249c9e
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:cdaf7be69a5d66cfa4777feb8266421dafd55353d6703e7e591bb9fb23017abc
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:7c548dd869f3458fb70a536708306cb615b9166c7e866cc1215e82f183b39529
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:1419e274b4e2e4c37985208c3b1b9ecd0cdef493427ace3c71ff35ed8fcc394b
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:a5c65dee1f7e4177050244109c10e384899e01d1924b41a540373aa7daf4f274
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:c46657831915ed5f42f4a55e725bc9d17e0671191f4509e610468fda160710c0
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:1e5c12ed0de090198e097cc5deb133ad075fba224f239b8b892e6e1672ec9883
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:fff4fb0924f866b305ad0ecc5d2682da6cb20de15448a2a8c640df182719b4e2
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:28730c8f2344e51b0b9ca603ac63b60885b39be86582943bd1cc200779eee0c5
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:1ccd17b482df844ec42936993ff162f665b5a473ad344270ff2425cd63d428fc
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:6a7c54cb7e6c3cc506ffb8a4f3ec7443ef80798e2c917e20b5f06d39bde02bc4
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:7375991ce3b9e64cb7542dc176414f0974047667d08beac184e2adadd9f83532
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:3ac71b363aef8db127b88e0a63ee616d0d06342142f83522e34c9cac4dd77b2e
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:c4c3bc4f45ae194b884d89319625f2a0970dada823877a35feecd8c733cb2c8d
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:7050f71c3a038fec16db8d0eee69d7d5a4294c043c13e9cf69c3cb2e28897f4c
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:70ddf96707d398d5444bb7bbde6c684275cff62b8a83c7cf6143ff529daf0dee
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:f3a00f55dee152e0345e2616ac14f652dcb4f37f3b9dfca6f049ac5522874c6e
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:f1c371fef66e96199c4a96e21192b47bca423b282d6f99e927644840aaedfd80
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:8ccbbc058d324429aef0c31c99de8c7b3a5abc02ec6f7a14cc60eef63a8f5c69
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:7b64803ddd3b7886ad3ec3de943a38cec5aeb9604732ab7a5ea5061748bf1411
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:5156129064dbe4e9989b41018b1bb250b2eb3f02e0c0afd75a874da9b54ad3c2
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:875e4bf42ad192cfb954e38c8474a529aa938855d69dc0d83c4ee4a4777c1028
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:3faa2318f159d264a9194fc01239d3a2098c184b365b153e3c56e555d5d4992e
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:7968e96b48b2ed63e7657793cf6cc2564055809df19cb43b7e284ac03f2cc26c
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:dca845ad05b02619b81f231eb57dd826720c221d22282fe7422ceebe57191371
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:794ef15591f217be629d6c680c56e84962923993f6dd77514ca5d2bdbd96217b
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:c10a72bf8c0d76b9b4d9d7c0bfca4d83a3f7dd7f6a37b4cc24ee7591638a2482
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:08dd3f6805fa02478671927982a2c77bdec0a5b7e7fd1c9faf20c8af2bb7b428
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:fbfade039739179995828b2ce4f4411b1315ac9a0e4a98e96a115012676a859d
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:b782190b3a285003d319321be09a25ee1376c951cb0cd146cbb634c9b1ca97b3
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:8e62fa4decf01a3beae7080f19469f33822f5010a688ada5db8d2d23b008e498
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:3f286e3c24fe73b4b42e6cf4b6769f91e333014450e1ab135b1e5b4a271cbaec
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:c2ef90b4704251dc9f7ef5288f5bad7aeb4cd2b3706ec531c60f6f4346073a77
        version: 2.12.4
      entries:
      - name: advanced-cluster-management.v2.12.4
        version: 2.12.4
      - name: advanced-cluster-management.v2.12.3
        version: 2.12.3
      - name: advanced-cluster-management.v2.12.2
        version: 2.12.2
      - name: advanced-cluster-management.v2.12.1
        version: 2.12.1
      - name: advanced-cluster-management.v2.12.0
        version: 2.12.0
      name: release-2.12
    - currentCSV: advanced-cluster-management.v2.13.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-09-08T18:37:07Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <2.13.4'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:dc271885425ab74e0554b531a2078133ce02006d0bb9688f3b54a850cfefd8f8
        - registry.redhat.io/rhel8/postgresql-16@sha256:89dfeaefc7e9e974e47a3365e22f354314f8a5e906e93a0152afe4e42f6a82d0
        - registry.redhat.io/rhacm2/console-rhel9@sha256:5526edf70b784bddfc2b7e1c154a44201c346fc20117635500bb54d325921c25
        - registry.redhat.io/rhacm2/acm-flightctl-worker-rhel9@sha256:57f69874b2df77c185374ddcea530af6783410a0f35528ee1a44cc2dee6f7fa2
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:0040791500337f88af7d922895487549915faa3b69dacba1a23540e7d6f2e580
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:ba51f7d077858a0abe32a254970390b50cd7d75433b5661e17f887c48c56a30f
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:65e1f0eb4e6c797c6dc6518674b28370db3fdaa6d7e17f4233c07207224dcb48
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:3028d011e7b88dd9597821128275f606fdba639e5b83e4e6a8783fa2616cba6f
        - registry.redhat.io/rhacm2/acm-flightctl-api-rhel9@sha256:8302b536c896290533f3f50084cb9047cb55af3a4404887b70704d23862619fa
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:d29ceda561f36faa581d76d8b889f7e01d253d33c88709078855ae26cada4917
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:0272215ab43b5688295d94cce0de4490a2fbabc1dd09ed41645be7acef1e3c9b
        - registry.redhat.io/rhacm2/acm-flightctl-ocp-ui-rhel9@sha256:89863ce0eb5fb73272bda8762d74461552f8502fbe262b16cbde01064cc533b3
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:84109be1814719c343b8b4676981cdbcea0356f012b6c854221c271bdb518fea
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:fc16203de469dd1ba16e428a9d7efbb16a5f511d866ef6fcc253530de1f032a6
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:ab1e4f8685896e2016a0b4c4479378058314f365e4f09ef42cba67f3649bb306
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:b3535378919524df259b85587ca83f5620433914418c9941374951c36e0530ae
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:1b7e6e6a4ae36209f98b1539c37c5975a82363b33fe5ac1e3a320e58b99be239
        - registry.redhat.io/rhacm2/acm-flightctl-periodic-rhel9@sha256:2c9b3c3d6cd7e435da337447580dfbfde8c4932308405eeae5d97ed5ec57fbfc
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:588678bea97237d87cd2e55a0e8000e73461cfc421cc91d554a2f1e352b68d14
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:7f742f024c787932932b8a4100a5329e70aad39537e0c20c2a4e245fc6245e25
        - registry.redhat.io/rhel9/redis-7@sha256:848f4298a9465dafb7ce9790e991bd8a11de2558e3a6685e1d7c4a6e0fc5f371
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:ccc9c0524149820d701c39355af89287ad749832053ea55d1585e77347fa0ec0
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:7b15c6b492e5cb2725b9b345aa3bfae4108e2815bc233ac063b81a67c6550fb6
        - registry.redhat.io/rhel8/postgresql-13@sha256:826a1a75f4186f36217b5c1f1aa270838c4b55b1021d33a4876566bfc3b8e629
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:b1620cebc62240f93b12e6937d15760440be119c421c127214974bdfa9d08ebd
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:0d8202cd10170bb0a138d52a1d71e630ff74df466541897ba86ed324a75ea1d8
        - registry.redhat.io/openshift4/ose-cli-artifacts-rhel9@sha256:67ff9f19edac8d623ce9080a8c05cde3da73afd68e7352e4ee5b22a9ccb596bc
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:3490daccc203fade26e7df348c74a907ab0ffc8901f4965606c8227849087c44
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:f74fb6b4d42f1d7431bfc9d2424f953732fed06d2abd8d1fc27b878b4d1322d5
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:ce342242c50c103fc5e4b0126551231977b35bd477eb627fb8f5d2a6510ddaa0
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:e23115b10a92c323b4cdaf0412919c273dbca33f19910120498bd49cb5810ff5
        - registry.redhat.io/rhacm2/acm-flightctl-ui-rhel9@sha256:86800d67ec8ed63e5408947ba6a441e5afc1050b72a86487b9357674ed495a7e
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:f1db9b10481319dc142ca6046e4e3141addb4c0afab482eeaf80643fad9ad039
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:96bdb2e25ae0f381de0c94a09a3cd11f3c0bdc1d8cfb4fd5ced2cd7b24fbe2c2
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:4d068813e3eb704ea7b2d342d5a2ce8fa2ac71793b1b37634c18704ef0a5e8d7
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:5fc35a944895cf7ea6afa5929c45304aaa9f17397b4f1e42cb8c799d9aead9d5
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:f9ba60c2a415a863c6bd6030fb4663913368a9c271bb40b8acfdc294dcf3a7eb
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:8b152051ef2c4c077379e72aa510c0bd613b219254cad66fab0cd3ec0d1d4bf1
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:04bcddae590e2debf35090055dd52ae4bcaac4b369727266be53ac64670203d0
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:49e99f41cbea3faefc3ae93385a10b544ce91df0d6639cfee65760be6ae0df00
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:81cc71faaf20a986d233fee799a7d2b954cd3f43875a2689f76c53920ccf4df0
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:67939d6e4342372820d7af6777f0ffd123e86e4e8186132c3d4892cb717a5e7c
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:6f95308a508499d2d995653d3fd2def94ac5170b4c3f391d32a1e0594250c33a
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:669e9e2178476bc7a4688f3ebf4c20402608bb4b10e4df6863ec821c7e572299
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:ae1f175ffbd3698f2b3036bf9b62b5c9b24729cfb8248345b8bdd1525ccc5bb4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:ab0192a51ee37b7508f3e4db66ffa53b904dcb1f76d9aa1ac0e80447d8ff9e6a
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:5aec572fc18c8ed533620553aba2c676453297ad2440336a5228a5a4f50f5782
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:a0218b9d5f1803f60d9b295d1504a870780895cffe8f1c5b3ce4c3553da5e4f9
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:a2ab651af1c97e54d2c14de78ff554c0f25e8d000af93a74fc7b7ed913f4e338
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:72c6896a5a4dbb7717d76d9eeade96130d97909eee07e8d3cc17af0658b306ab
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:9ef9f1e2010d77211126482c03e8c72dfb7458940dc95ac952ee8dd1dedd2991
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:6ed6f62f5c05a24938335287b785b80c253424e1ee20d83311cd87c6a2090895
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:2da264ca59278f63990e579939371d45afec3bc7d9922a608520e7647a0377ac
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:1393744757874e79d1fb058cb5d3fe7dcbbcd2f187bdcdcdf34e67227c61fbb9
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:296a31e1aad9f6288bdf80fb3191642e3dd404e9e8fcad198016bedbd3abe9b4
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:3cbb9d271c2f23d430a623c83606bbc8a80379c24d13e610886ad51fd21ce1e4
        version: 2.13.4
      entries:
      - name: advanced-cluster-management.v2.13.4
        version: 2.13.4
      - name: advanced-cluster-management.v2.13.3
        version: 2.13.3
      - name: advanced-cluster-management.v2.13.2
        version: 2.13.2
      - name: advanced-cluster-management.v2.13.1
        version: 2.13.1
      - name: advanced-cluster-management.v2.13.0
        version: 2.13.0
      name: release-2.13
    - currentCSV: advanced-cluster-management.v2.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-25T02:10:33Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.13.0 <2.14.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          repository: https://github.com/open-cluster-management/multiclusterhub-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MulticlusterHub defines the configuration
              for an instance of a multicluster hub, a central point for managing multiple
              Kubernetes-based clusters. The deployment of multicluster hub components
              is determined based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n### Where to find more installation
          information\nYou can find additional installation guidance in the [install
          documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:364c7416a9642dfb37f05e2eac4427918959b12b159ac90a9ca310c584d6434a
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:69f0b9a7fb494572adf2795be865d6d6ba4b67db0eed0e9d61f5dcb5e9bf6efe
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:c2ef90b4704251dc9f7ef5288f5bad7aeb4cd2b3706ec531c60f6f4346073a77
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:0e1eba23e0e8454f2dec240704e7ccc922351d58880f72514d3b2bc30c9e995e
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:c6ff6cae5bc6447d0ed24e458fc5d5fc709b2106fe93b143f9546bd962ea7f87
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:ad7f30c61bb11c0d32686e75b01f3b286a1fb06edd8d2ef6a9e2d924fba2b565
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:b489229190f9a3da66bd154e65490c3a35d1999d2db6707759f06eadb02b968d
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:8e6ee423acfafe28c002911e832d47dffc00a5227325bd6a5b1ab359c7eee5f4
        - registry.redhat.io/rhacm2/console-rhel9@sha256:d4c42a203110050a3834be8d5a0eee384ad7e9fbb991da589d75acd15055b8b1
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:0cecc3d87afcd0d63bdeae3ecae37f8a38534fbdc8bad5c548e0d4f4fb0156a4
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:8f4f59e237f366f08367cabbea3eb51e766ad52db0d35c5efc56182c13fc5c09
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:8253a9225f30b0f0a405e2c3293924acbfbf81f3d9c50bc4431fe4052a9fe908
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:55ef2c163b0c0deb3402e39acf37ff6bb05e2a5d268dcae5f8a400128d67c584
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:b7b7581e7743148d19283250d3300d292b96c48ecae96f6e6202a7ca8204e2ab
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:c329cfd004c58a7fcc40f4fbb986399dc2fe08d7922d2b4f210664f58751b1a6
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:143f49d2ce1f58282e276396722ba37165cd1bf2051d537bb4e6208786e5aed6
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:53deacec64238607afda29c27463d72ac3b582c9670e2e77a9c7e48851214795
        - registry.redhat.io/openshift4/ose-cli-rhel9@sha256:a48758b69b9a7451b7e0caea465a181f854b41f1e74ce86c15216ee1bb31fcb9
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:f9a9e10fde9065d650a7dc8eba83349e5db0a21727027464cf144724d0de8cfa
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:935baf67627d5e76c0766d5d2a9e36b398a12535f097749a987dd0c5de788570
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:0252cf95019717f03c7ac72fb034b5bd9ad7aef2da0a92dd21c56cf47f03b812
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:d56b997a9084d56c63f539147052e64ba30258afe91a9ff832f3673cbccbe467
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ee742e7baa8bcbbfccf79abcf15d22347d14f280eafaf22c42cef71051b10fba
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:0ab21d4189669901fe085c7f3a4b4c9778c853f8abd706bbfffcb59942fbc8e6
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:9c79b1153fc927e7582c5346ea0408e03f794eb983597f02a5e816ae961ba86b
        - registry.redhat.io/rhem/flightctl-periodic-rhel9@sha256:ee59f307e468e2b7b38f7c040898d587d441f8fb3d59cb022ae23e4717739e3f
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:2b1b1c707e0b5217c66a96fb82070de47fc0342747081d4be0049c4542944c8a
        - registry.redhat.io/rhel8/postgresql-16@sha256:d4abbd9492347e0aa7e1b0ff967e74d4b6a4edd6fd20d896f2f2ae03c623181c
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:46c5835a4bfc0e4767d0959bc47e939c195be197e9715aae17788d4d0e799d56
        - registry.redhat.io/rhel9/redis-7@sha256:ff31822cba556e40b6b00720839cc56f474fdb9257fb39b7604fafc11761d379
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:0ea80a1039d217e899ea012b77e99a7c5832e5b7e292dca473759c449bf2d1ea
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:e063c54c7182b70d78d24b5026f174436885383980c9b1e3e7664fa2adfad883
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:d550013d385552d6242b56c5b60f15b3ebb488c2033f808de54bcc639abc2f30
        - registry.redhat.io/rhem/flightctl-ui-ocp-rhel9@sha256:973eb3ca3da32c87ba308928a21fbd25514b2f6a6662300d4928479cb554cff3
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:bcb5949ebc31bbf6b45dfd34aae9c80c82a08531e8a3d9f6eef3354f942a5a5d
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:cee29bdb03927dc1866c267232d391ea5ee762b23c6b04a777e7d61d14a61bba
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:0cc23b5063c0d3264ed345dad8453901e4db18cc9bed1656f2799f72de75bd53
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:8151a89ebfebf2e6d336849fdd329b123b7f486938808ee80bf6bd8d44f6985c
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:7ee0aea84bf00b7b42027fa3e970c1f4b279e345a698b4050c9f413eb0383bb1
        - registry.redhat.io/rhem/flightctl-ui-rhel9@sha256:940c0564ced75b7912d3ffc923a1f6bab9aed0f4a52962f7f894d08937b7d37e
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:ce5ff191023b5e14c339b2a03aa1be05a12d024efa5302652065597e630996b6
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:90d399736ca81bb244415efaf05738beff806912a6ac7444a7d4839089b7a081
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:556d737195a4878877c650a93619af1528e52694f2a8e47e13463a603d2f4900
        - registry.redhat.io/rhem/flightctl-cli-artifacts-rhel9@sha256:3491ec932d90112e89f3a4cebe5bdd9ed6a7811f116df43bab8bb5f827129361
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:d9fa60b5b1d6b6c94c5d8b610c9a7b979202de8b46a344e0a65d06f6ef21a7ef
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:40b3e063435eea6e526a57d6bd1ae489ca703e81eeca82dda1c2e6aedec1262c
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:95085fc33a0150a39b6ef4e616fd23f0bddd9dcad0e75c4a89ed27113626e6eb
        - registry.redhat.io/rhel9/memcached@sha256:955138db9a7a23ee31198aabebd0811438fee11bb0b624c38ddb90efe52bd07c
        - registry.redhat.io/rhem/flightctl-worker-rhel9@sha256:1b1f6f85af0910f1acd029f54a608b25d3a2977c3d8b323266391543324305b4
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:f3e34b200e9480363c56da1cd88b2de37fc6a490691e4596a132217f8ea286d8
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:c70242a037e36690d8dd1030254dde778917356b4f1f61db2cecb03b6c2d7380
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:cb5346afe1c7b58eb758cdced94f2c1161c1621033b52f29757efe3a50320a65
        - registry.redhat.io/rhem/flightctl-api-rhel9@sha256:83c6c7819e04aa58d961518367ce6ad3159a28300ca097967f630044d181be34
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:8e57af77cc5dee3aaa83f6ae6e62e232c5fc4fbe4dca8eb755226351963058e3
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:44e305ca15e7c9b50bd7fa8bb391f5ee66906aa26ecc5f612d029aaee5bb4cb8
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:d65aef370a7ed01c0c7cc0ad7f0800359db7863db4ec8845de5b450992f26933
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:ad534f7b0efea45b9f4a0a026c76a2a030cb608acdb835dfea3852332acfcae1
        version: 2.14.0
      entries:
      - name: advanced-cluster-management.v2.14.0
        version: 2.14.0
      name: release-2.14
    defaultChannel: release-2.14
    packageName: advanced-cluster-management
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.17.40-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5-beta.1 <8.17.40-beta.1'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBM Turbonomic software automates and simultaneously optimizes
              performance, compliance, and cost in real time. Turbonomic discovers
              and automates resource optimization for the complete application stack.
              Applications are continually resourced to perform while satisfying technical
              and business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
          - kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Optimization for Kubernetes
          IBM Turbonomic software automates and simultaneously optimizes performance, compliance, and cost in real time.
          Turbonomic discovers and automates resource optimization for the complete application stack. Applications are continually resourced to perform while satisfying technical and business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (proactive rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          Turbonomic assures application performance by giving workloads the resources they need when they need them.

          ### How Does It Work?
          Turbonomic uses a client agent called — KubeTurbo — that runs in your cluster to discover, monitor and optimize resource usage within your environment.
          KubeTurbo runs together with the default Kubernetes scheduler and sends data back to the Turbonomic for analysis.
          Turbonomic determines the right actions that drive continuous health for your workloads and your cluster.

          Turbonomic discovers Red Hat OpenShift Virtualization resources through a new or existing target configuration with Red Hat OpenShift. For more information refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=targets-red-hat-openshift-virtualization)

          ### Get Started
          Please follow the [Kubeturbo Documentation Page](https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets) for all deployment options and an overview of typical use cases.

          ### Known Issues
          For known issues refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=notes-known-issues)
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Kubeturbo Documentation Page
          url: https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets
        - name: Kubeturbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
        - name: IBM Turbonomic Application Resource Management
          url: https://www.ibm.com/products/turbonomic
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/kubeturbo-operator@sha256:7111c7c4002a447dc6977dbb73e96507bae56766fc0552fb2d036b20583722ba
        version: 8.17.40-beta.1
      entries:
      - name: kubeturbo-operator.v8.17.40-beta.1
        version: 8.17.40-beta.1
      name: beta
    - currentCSV: kubeturbo-operator.v8.17.40
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5 <8.17.40'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBM Turbonomic software automates and simultaneously optimizes
              performance, compliance, and cost in real time. Turbonomic discovers
              and automates resource optimization for the complete application stack.
              Applications are continually resourced to perform while satisfying technical
              and business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
          - kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Optimization for Kubernetes
          IBM Turbonomic software automates and simultaneously optimizes performance, compliance, and cost in real time.
          Turbonomic discovers and automates resource optimization for the complete application stack. Applications are continually resourced to perform while satisfying technical and business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (proactive rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          Turbonomic assures application performance by giving workloads the resources they need when they need them.

          ### How Does It Work?
          Turbonomic uses a client agent called — KubeTurbo — that runs in your cluster to discover, monitor and optimize resource usage within your environment.
          KubeTurbo runs together with the default Kubernetes scheduler and sends data back to the Turbonomic for analysis.
          Turbonomic determines the right actions that drive continuous health for your workloads and your cluster.

          Turbonomic discovers Red Hat OpenShift Virtualization resources through a new or existing target configuration with Red Hat OpenShift. For more information refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=targets-red-hat-openshift-virtualization)

          ### Get Started
          Please follow the [Kubeturbo Documentation Page](https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets) for all deployment options and an overview of typical use cases.

          ### Known Issues
          For known issues refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=notes-known-issues)
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Kubeturbo Documentation Page
          url: https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets
        - name: Kubeturbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
        - name: IBM Turbonomic Application Resource Management
          url: https://www.ibm.com/products/turbonomic
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/kubeturbo-operator@sha256:829dc6b9800efd3fd77d07f955a53610c7b2888b93495b2db32f1f48e93a05a6
        version: 8.17.40
      entries:
      - name: kubeturbo-operator.v8.17.40
        version: 8.17.40
      name: stable
    defaultChannel: stable
    packageName: kubeturbo-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://access.redhat.com/products/red-hat-integration
    name: red-hat-camel-k
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: red-hat-camel-k-operator.v1.10.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
          createdAt: "2025-01-08T17:00:43Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
        version: 1.10.10
      entries:
      - name: red-hat-camel-k-operator.v1.10.10
        version: 1.10.10
      - name: red-hat-camel-k-operator.v1.10.9
        version: 1.10.9
      - name: red-hat-camel-k-operator.v1.10.9-0.1743522834.p
        version: 1.10.9+0.1743522834.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1742828111.p
        version: 1.10.9+0.1742828111.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741884247.p
        version: 1.10.9+0.1741884247.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741875993.p
        version: 1.10.9+0.1741875993.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741580860.p
        version: 1.10.9+0.1741580860.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739988175.p
        version: 1.10.9+0.1739988175.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739388437.p
        version: 1.10.9+0.1739388437.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739387844.p
        version: 1.10.9+0.1739387844.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.10.x
    - currentCSV: red-hat-camel-k-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1beta1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:9e4fb5369e0479d4a68f60725e7d12abdc66c552babe59f1358f6e6b0f26357d
          createdAt: "2020-10-06T02:45:00Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:9e4fb5369e0479d4a68f60725e7d12abdc66c552babe59f1358f6e6b0f26357d
        version: 1.4.2
      entries:
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.4.x
    - currentCSV: red-hat-camel-k-operator.v1.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
          createdAt: "2022-08-15T20:17:25Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <=1.6.8'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
        version: 1.6.10
      entries:
      - name: red-hat-camel-k-operator.v1.6.10
        version: 1.6.10
      - name: red-hat-camel-k-operator.v1.6.9
        version: 1.6.9
      name: 1.6.x
    - currentCSV: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:e916f9040d3f4feb26cc1534d3ab2722890b34c7fe146cdce8b6085b47206e22
          createdAt: "2022-11-28T10:43:18Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <1.8.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:e916f9040d3f4feb26cc1534d3ab2722890b34c7fe146cdce8b6085b47206e22
        version: 1.8.2+0.1675913507.p
      entries:
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.8.x
    - currentCSV: red-hat-camel-k-operator.v1.10.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
          createdAt: "2025-01-08T17:00:43Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
        version: 1.10.10
      entries:
      - name: red-hat-camel-k-operator.v1.10.10
        version: 1.10.10
      - name: red-hat-camel-k-operator.v1.10.9
        version: 1.10.9
      - name: red-hat-camel-k-operator.v1.10.9-0.1743522834.p
        version: 1.10.9+0.1743522834.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1742828111.p
        version: 1.10.9+0.1742828111.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741884247.p
        version: 1.10.9+0.1741884247.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741875993.p
        version: 1.10.9+0.1741875993.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741580860.p
        version: 1.10.9+0.1741580860.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739988175.p
        version: 1.10.9+0.1739988175.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739388437.p
        version: 1.10.9+0.1739388437.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739387844.p
        version: 1.10.9+0.1739387844.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: candidate
    - currentCSV: red-hat-camel-k-operator.v1.10.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
          createdAt: "2025-01-08T17:00:43Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
        version: 1.10.10
      entries:
      - name: red-hat-camel-k-operator.v1.10.10
        version: 1.10.10
      - name: red-hat-camel-k-operator.v1.10.9
        version: 1.10.9
      - name: red-hat-camel-k-operator.v1.10.9-0.1743522834.p
        version: 1.10.9+0.1743522834.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1742828111.p
        version: 1.10.9+0.1742828111.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741884247.p
        version: 1.10.9+0.1741884247.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741875993.p
        version: 1.10.9+0.1741875993.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741580860.p
        version: 1.10.9+0.1741580860.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739988175.p
        version: 1.10.9+0.1739988175.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739388437.p
        version: 1.10.9+0.1739388437.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739387844.p
        version: 1.10.9+0.1739387844.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: latest
    - currentCSV: red-hat-camel-k-operator.v1.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
          createdAt: "2022-08-15T20:17:25Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <=1.6.8'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
        version: 1.6.10
      entries:
      - name: red-hat-camel-k-operator.v1.6.10
        version: 1.6.10
      - name: red-hat-camel-k-operator.v1.6.9
        version: 1.6.9
      name: stable
    defaultChannel: 1.10.x
    packageName: red-hat-camel-k
    provider:
      name: Red Hat
      url: https://access.redhat.com/products/red-hat-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MySQL
      provider-url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
    name: ndb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ndb-operator.v1.6.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "mysql.oracle.com/v1","kind": "NdbCluster","metadata":
            {"name": "example-ndb"},"spec": {"redundancyLevel": 2,"dataNode": {"nodeCount":
            2},"mysqlNode": {"nodeCount":2}}}]'
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: container-registry.oracle.com/mysql/community-ndb-operator@sha256:b562dc3303ffae9f4dab1d73fd0fe05833bdaa8b731af4227da81af617c75c6b
          createdAt: "2024-12-18T15:05:46.000Z"
          description: The MySQL NDB Operator is a Kubernetes operator for managing
            a MySQL NDB Cluster setup inside a Kubernetes Cluster.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Ndb Cluster
            displayName: NdbCluster
            kind: NdbCluster
            name: ndbclusters.mysql.oracle.com
            version: v1
        description: |
          ## NdbCluster Custom Resource
          The NdbCluster custom resource is defined with custom fields that map to the configuration of a MySQL Cluster. To deploy a MySQL Cluster inside the K8s Cluster, an NdbCluster object with the desired configuration of the MySQL Cluster has to be created in the K8s Cluster. The NDB Operator watches for any changes made to any NdbCluster objects and will receive an event when the NdbCluster object is created in the K8s Cluster. It will then start deploying the MySQL Cluster, with the specified configuration, inside the K8s Cluster using native Kubernetes workloads and resources. Once the MySQL Cluster is ready, the NdbCluster resource can be further modified to make changes to the MySQL Cluster configuration. The Operator will receive this event as well and will update the configuration of the already running MySQL Cluster.

          The documentation at [NdbCluster-CRD.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md) explains the various custom fields of the NdbCluster CRD and the effect they have on the MySQL Cluster configuration. The [examples](https://github.com/mysql/mysql-ndb-operator/tree/main/docs/examples) folder has a few examples of the NdbCluster resource objects which can be used to deploy MySQL Cluster inside the K8s Cluster.
          ## Deploy a simple MySQL Cluster

          The [examples/example-ndb.yaml](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/examples/example-ndb.yaml) file has a NdbCluster object example-ndb that defines a simple MySQL Cluster with 2 data nodes and 2 MySQL Servers. To create this object in the default namespace of the K8s Cluster, run this command from the root of the repo :
          ```
          kubectl apply -f docs/examples/example-ndb.yaml
          ```

          Various NdbCluster status fields report back the actual status of the MySQL Cluster running inside the K8s Cluster. Wait for the MySQL Cluster nodes to be started and become ready using the [UpToDate](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md#ndbclusterconditiontypestring-alias) Condition defined in the NdbCluster object status :
          ```
          kubectl wait --for=condition=UpToDate ndb example-ndb --timeout=10m
          ```
          The MySQL Cluster is ready for transactions once the NdbCluster is UpToDate, i.e. the UpToDate condition becomes true.

          At any point, a brief status of an NdbCluster object can be viewed using kubectl get ndb command.
          ```
          kubectl get ndb example-ndb
          ```

          The output of the command gives a short summary the MySQL Cluster nodes controlled by the respective NdbCluster object.
          ```
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE   UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:2/2    Ready:2/2 3m    True
          ```

          To list all the pods created by the NDB Operator, run :
          ```
          kubectl get pods -l mysql.oracle.com/v1=example-ndb
          ```

          For example-ndb NdbCluster object, two management nodes (*-mgmd-* pods), two multi-threaded data nodes (*-ndbmtd-* pods) and two MySQL Servers (*-mysqld-*) would have been created by the NDB operator :
          ```
          NAME                   READY   STATUS    RESTARTS   AGE
          example-ndb-mgmd-0     1/1     Running   0          3m53s
          example-ndb-mgmd-1     1/1     Running   0          2m5s
          example-ndb-mysqld-0   1/1     Running   0          78s
          example-ndb-mysqld-1   1/1     Running   0          78s
          example-ndb-ndbmtd-0   1/1     Running   0          3m53s
          example-ndb-ndbmtd-1   1/1     Running   0          3m53s
          ```
          If running inside a openshift cluster, pods from statefulset "example-ndb-mysqld" will not start up. By default openshift runs all the pod with "restricted" scc, but the statefulset "example-ndb-mysqld" required higher privilege. Use the below command to check the exact scc required
          ```
          oc get pods example-ndb-mysqld-0 -o yaml | oc adm policy scc-subject-review -f -
          ```
          Create a user account with the scc and set the user to the statefulset "example-ndb-mysqld". Below commands creates an user with "anyuid" and assign it to the statefulset.
          ```
          oc create sa any-uid-user
          oc adm policy add-scc-to-user anyuid -z any-uid-user
          oc set sa sts example-ndb-mysqld any-uid-user
          ```
          ## Connect to the MySQL Cluster

          The NDB Operator, by default, creates few Services to expose the services offered by the MySQL Cluster nodes within the Kubernetes Cluster.
          ```
          kubectl get services -l mysql.oracle.com/v1=example-ndb
          ```

          The above command will generate an output similar to :
          ```
          NAME                 TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S) AGE
          example-ndb-mgmd     ClusterIP   10.100.207.255   <none>        1186/TCP 5m
          example-ndb-mysqld   ClusterIP   10.103.182.157   <none>        3306/TCP 5m
          example-ndb-ndbmtd   ClusterIP   None             <none>        1186/TCP 5m
          ```

          The NDB Operator will also create a root user with a random password on all the MySQL Servers. The root password will be stored in a K8s Secret which can be retrieved from the NdbCluster object status :
          ```
          kubectl get ndb example-ndb -o jsonpath={.status.generatedRootPasswordSecretName}
          ```

          To extract and decode the password from the Secret :
          ```
          base64 -d <<< \
          $(kubectl get secret example-ndb-mysqld-root-password \
          -o jsonpath={.data.password})
          ```

          One can also specify the password and host to be used with the root account via the NdbCluster spec. See the [CRD documentation](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md#ndbmysqldspec) for more details on this.
          ## Access MySQL Cluster from inside K8s

          To connect to the Management and MySQL Servers from within the Kubernetes Cluster, you can use the example-ndb-mgmd and example-ndb-mysqld services as the MySQL Cluster connectstring and MySQL host respectively. You can also use the mgmd and mysqld pods' hostnames to directly connect to that particular pod. The mysql, ndb_mgm clients and any other ndb tools will work as expected.

          A demonstration is available at [connect-from-inside-k8s-demo.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/connect-from-inside-k8s-demo.md).
          ## Access MySQL Cluster from outside K8s

          By default, the Management and MySQL services created by the NDB Operator are of type [ClusterIP](https://kubernetes.io/docs/concepts/services-networking/service/#publishing-services-service-types) and are accessible only from within the K8s Cluster. To expose them outside the K8s Cluster, the services have to be upgraded into a LoadBalancer type. This can be done by setting the spec.managementNode.enableLoadBalancer and spec.mysqlNode.enableLoadBalancer fields of the NdbCluster resource to true. These options can be enabled when the NdbCluster resource object is created or via an update to the object when the MySQL Cluster is already running.

          Once enabled, the previously ClusterIP type example-ndb-mgmd and example-ndb-mysqld services will be upgraded to LoadBalancer type, and they will be available at the external IP assigned to them by the cloud provider.

          Another way to access these services without enabling the LoadBalancers support is to use the kubectl port-forward command.

          In both ways, only the mysql and ndb_mgm clients work. Any NDB tool which uses the NDBAPI to connect to the MySQL data nodes will not work as expected from outside the K8s Cluster.

          A demonstration is available at [connect-from-outside-k8s-demo.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/connect-from-outside-k8s-demo.md).
          ## Updating MySQL Cluster Configuration

          Once a MySQL Cluster has been deployed by the NDB Operator inside the K8s Cluster, its configuration can be updated by updating the spec of the NdbCluster resource object that represents it. The NDB Operator will pick up any changes made to the NdbCluster resource object and will apply the updated configuration to the respective MySQL Cluster. The NDB Operator takes care of updating the MySQL Cluster config file and restarting all the management/data nodes if required. When the update is being handled by the NdbOperator, it will set the UpToDate condition of the NdbCluster resource object to false and will not accept any further updates until it completes the current one.

          The kubectl get ndb example-ndb command will report the UpToDate status of an NdbCluster. It can be used to check if the current update has been completed.
          ```
          $ kubectl get ndb example-ndb
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE   UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:1/2    Ready:2/2 10m   False
          ```

          The UpToDate condition can also be used to wait for the operator to complete the update.
          ```
          kubectl wait --for=condition=UpToDate ndb example-ndb --timeout=10m
          ```

          Once the update is complete, the UpToDate condition will be set back to true by the operator.
          ```
          $ kubectl get ndb example-ndb
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE      UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:2/2    Ready:2/2 10m50s   True
          ```

          ## Delete a MySQL Cluster

          To stop and remove the MySQL Cluster running inside the K8s Cluster, delete the NdbCluster resource object.
          ```
          kubectl delete -f docs/examples/example-ndb.yaml
          ```

          (or)
          ```
          kubectl delete ndb example-ndb
          ```

          This will delete all the data, the pods running the MySQL Cluster nodes and also delete all other associated K8s resources created by the NDB Operator.
        displayName: Ndb Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - ndb
        - in-memory
        links:
        - name: Document
          url: https://dev.mysql.com/doc/ndb-operator/en/
        maintainers:
        - email: paulo.p.silva@oracle.com
          name: paulofasilva
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: MySQL
          url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
        relatedImages:
        - container-registry.oracle.com/mysql/community-ndb-operator@sha256:b562dc3303ffae9f4dab1d73fd0fe05833bdaa8b731af4227da81af617c75c6b
        version: 1.6.2
      entries:
      - name: ndb-operator.v1.6.2
        version: 1.6.2
      - name: ndb-operator.v1.6.1
        version: 1.6.1
      - name: ndb-operator.v1.6.0
        version: 1.6.0
      name: alpha
    defaultChannel: alpha
    packageName: ndb-operator
    provider:
      name: MySQL
      url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: VASTData
      provider-url: https://www.vastdata.com
    name: vast-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vast-csi-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"storage.vastdata.com/v1","kind":"VastCSIDriver","metadata":{"name":"csi.vastdata.com"},"spec":{"applySecurityContextConstraints":true,"attachRequired":true,"controller":{"dnsPolicy":"Default","extraArgs":{"csiAttacher":[],"csiProvisioner":[],"csiResizer":[],"csiSnapshotter":[]},"nodeAffinity":{},"nodeSelector":{},"podAffinity":{},"podAntiAffinity":{},"resources":{"csiAttacher":{"limits":{"memory":"200Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiProvisioner":{"limits":{"memory":"400Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiResizer":{"limits":{"memory":"300Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiSnapshotter":{"limits":{"memory":"200Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiVastPlugin":{"limits":{"memory":"400Mi"},"requests":{"cpu":"100m","memory":"50Mi"}}},"runOnControlPlane":false,"runOnMaster":false,"tolerations":[]},"deletionViewPolicy":null,"deletionVipPool":null,"dontUseTrashApi":false,"driverType":"nfs","endpoint":null,"image":{"csiAttacher":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-attacher@sha256:b4d611100ece2f9bc980d1cb19c2285b8868da261e3b1ee8f45448ab5512ab94"},"csiNodeDriverRegistrar":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac"},"csiProvisioner":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-provisioner@sha256:405a14e1aa702f7ea133cea459e8395fe40a6125c088c55569e696d48e1bd385"},"csiResizer":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-resizer@sha256:a541e6cc2d8b011bb21b1d4ffec6b090e85270cce6276ee302d86153eec0af43"},"csiSnapshotter":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-snapshotter@sha256:2e04046334baf9be425bb0fa1d04c2d1720d770825eedbdbcdb10d430da4ad8c"},"csiVastPlugin":{"imagePullPolicy":"IfNotPresent","repository":"registry.connect.redhat.com/vastdataorg/csi@sha256:7169f279d0dd7130535d74df5b938ed6b7944bfd4fe52d242ceee5178f2ea156"}},"imagePullSecrets":[],"kubeletPath":"/var/lib/kubelet","logLevel":5,"node":{"dnsPolicy":"Default","nodeAffinity":{},"nodeSelector":{},"podAffinity":{},"podAntiAffinity":{},"propagateHostMountOptions":true,"resources":{"csiVastPlugin":{"limits":{"memory":"400Mi"},"requests":{"cpu":"100m","memory":"50Mi"}},"nodeDriverRegistrar":{"limits":{"memory":"100Mi"},"requests":{"cpu":"10m","memory":"20Mi"}}},"tolerations":[]},"numWorkers":10,"operationRetryIntervalMax":60,"operationRetryIntervalStart":10,"operationTimeout":15,"secretName":null,"sslCertsSecretName":null,"truncateVolumeName":64,"useLocalIpForMount":null,"verifySsl":false}},{"apiVersion":"storage.vastdata.com/v1","kind":"VastStorage","metadata":{"name":"vastdata-filesystem"},"spec":{"allowVolumeExpansion":true,"blockingClones":false,"clusterName":"","createSnapshotClass":true,"driverType":"nfs","ephemeralVolumeNameFormat":"csi:{namespace}:{name}:{id}","fsType":"","hostEncryption":null,"mountOptions":[],"provisioner":"","qosPolicy":"","reclaimPolicy":"Delete","secretName":"","secretNamespace":"","setDefaultStorageClass":false,"snapshotClass":{"deletionPolicy":"Delete","setDefaultSnapshotClass":false,"snapshotNameFormat":"csi:{namespace}:{name}:{id}"},"storagePath":"","subsystem":"","tenantName":"","viewPolicy":"","vipPool":"","vipPoolFQDN":null,"volumeGroup":"","volumeNameFormat":"csi:{namespace}:{name}:{id}"}},{"apiVersion":"storage.vastdata.com/v1","kind":"VastCluster","metadata":{"name":"vastcluster-sample"},"spec":{"endpoint":null,"passphrase":null,"password":null,"sslCert":null,"tenant":null,"token":null,"username":null}}]'
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/vastdataorg/csi-operator@sha256:ba7f30a42e327004ecbc81ff5ccc453bf3046a29bab5785b9f8a9746e5c6a3fd
          createdAt: "2024-07-01T11:59:59Z"
          description: |-
            VAST CSI Operator supports seamless installation and upgrade of VAST Data's CSI Driver.
            The VAST CSI Driver allows container orchestration frameworks such as Kubernetes to dynamically provision, manage and scale volumes on the [VAST Data Platform](https://www.vastdata.com/platform/overview). VAST provides scalable, reliable, and fast persistent storage that can be accessed remotely by any Kubernetes application container.
            ### Installation
            Refer to the VASTData Operator for Kubernetes [official documentation](https://support.vastdata.com/s/document-item?bundleId=vast-csi-driver-2.6-administrator-s-guide&topicId=deploying-vast-csi-driver/deploying-vast-csi-driver-on-openshift-with-vast-csi-operator.html&_LANG=enus).
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: vast-csi
          operators.operatorframework.io/builder: operator-sdk-v1.3.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/vast-data/vast-csi
          support: VastData Infra Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VastCSIDriver is a common specification for VAST CSI Controller
              and VAST CSI Node plugins, typically intended for creation in a single
              instance.
            displayName: VastCSIDriver
            kind: VastCSIDriver
            name: vastcsidrivers.storage.vastdata.com
            version: v1
          - description: VastCluster encapsulates the connection details required
              for accessing a VAST cluster. This information is ultimately stored
              in a Kubernetes Secret.
            displayName: VastCluster
            kind: VastCluster
            name: vastclusters.storage.vastdata.com
            version: v1
          - description: Represents VAST storage. Each VastStorage instance generates
              a corresponding storage class that can be used in PVCs to create workloads.
            displayName: VastStorage
            kind: VastStorage
            name: vaststorages.storage.vastdata.com
            version: v1
        description: |-
          VAST CSI Operator supports seamless installation and upgrade of VAST Data's CSI Driver.
          The VAST CSI Driver allows container orchestration frameworks such as Kubernetes to dynamically provision, manage and scale volumes on the [VAST Data Platform](https://www.vastdata.com/platform/overview). VAST provides scalable, reliable, and fast persistent storage that can be accessed remotely by any Kubernetes application container.
          ### Installation
          Refer to the VASTData Operator for Kubernetes [official documentation](https://support.vastdata.com/s/document-item?bundleId=vast-csi-driver-2.6-administrator-s-guide&topicId=deploying-vast-csi-driver/deploying-vast-csi-driver-on-openshift-with-vast-csi-operator.html&_LANG=enus).
        displayName: VAST CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - VAST
        - VASTData
        - CSI
        - Multi Cluster
        - AI
        - High-Performance Computing
        links:
        - name: VAST CSI Plugin
          url: https://github.com/vast-data/vast-csi
        maintainers:
        - email: volodymyr.boiko@vastdata.com
          name: VAST Infrastructure Team
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: VASTData
          url: https://www.vastdata.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.connect.redhat.com/vastdataorg/csi-operator@sha256:ba7f30a42e327004ecbc81ff5ccc453bf3046a29bab5785b9f8a9746e5c6a3fd
        - registry.connect.redhat.com/vastdataorg/csi@sha256:7169f279d0dd7130535d74df5b938ed6b7944bfd4fe52d242ceee5178f2ea156
        - registry.k8s.io/sig-storage/csi-resizer@sha256:a541e6cc2d8b011bb21b1d4ffec6b090e85270cce6276ee302d86153eec0af43
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:2e04046334baf9be425bb0fa1d04c2d1720d770825eedbdbcdb10d430da4ad8c
        - registry.k8s.io/sig-storage/csi-attacher@sha256:b4d611100ece2f9bc980d1cb19c2285b8868da261e3b1ee8f45448ab5512ab94
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:405a14e1aa702f7ea133cea459e8395fe40a6125c088c55569e696d48e1bd385
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac
        version: 2.6.3
      entries:
      - name: vast-csi-operator.v2.6.3
        version: 2.6.3
      name: stable
    defaultChannel: stable
    packageName: vast-csi-operator
    provider:
      name: VASTData
      url: https://www.vastdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: ""
    name: power-monitoring-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: power-monitoring-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "Kepler",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "kepler",
                    "app.kubernetes.io/name": "kepler",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "kepler"
                },
                "spec": {
                  "exporter": {
                    "deployment": {
                      "port": 9103
                    },
                    "redfish": null
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: registry.redhat.io/openshift-power-monitoring/power-monitoring-rhel9-operator@sha256:5eec15f6fe9bd4bdb1f5fb8057dc67c448c3711bcc7a0a4760f0d90c07688d1e
          createdAt: "2025-02-14T06:23:56Z"
          description: Deploys and Manages Kepler on OpenShift
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/internal-objects: |-
            [
              "keplerinternals.kepler.system.sustainable.computing.io"
            ]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift-power-monitoring/power-monitoring-operator
          support: Team Power Monitoring (team-powermon@redhat.com)
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KeplerInternal is the Schema for internal/unsupported API
            displayName: KeplerInternal
            kind: KeplerInternal
            name: keplerinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: Kepler is the Schema for the keplers API
            displayName: Kepler
            kind: Kepler
            name: keplers.kepler.system.sustainable.computing.io
            version: v1alpha1
        description: |
          ### Introduction
          Kepler (Kubernetes Efficient Power Level Exporter) uses eBPF to probe
          energy-related system stats and exports them as Prometheus metrics.

          ### NOTE
          To view the metrics exposed by Kepler, ensure that
          [User Project Monitoring](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/monitoring/configuring-user-workload-monitoring)
          is enabled in your cluster.

          ### Documentation
          Documentation and installation guide can be found below:
            * [Installation Guide](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/power_monitoring/installing-power-monitoring)
        displayName: Power monitoring for Red Hat OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kepler
        - power
        - energy
        - monitoring
        links:
        - name: Installation Guide
          url: https://docs.redhat.com/documentation/openshift_container_platform/latest/html/power_monitoring/installing-power-monitoring
        maintainers:
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: vimalkum@redhat.com
          name: Vimal Kumar
        - email: vprashar@redhat.com
          name: Vibhu Prashar
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift-power-monitoring/power-monitoring-rhel9-operator@sha256:5eec15f6fe9bd4bdb1f5fb8057dc67c448c3711bcc7a0a4760f0d90c07688d1e
        - registry.redhat.io/openshift-power-monitoring/kepler-rhel9@sha256:27d3cb0a98f1cffd649745e3886ad92995d83a0de6823ddc2d770bf036e2eeaf
        version: 0.4.0
      entries:
      - name: power-monitoring-operator.v0.4.0
        version: 0.4.0
      - name: power-monitoring-operator.v0.3.0
        version: 0.3.0
      - name: power-monitoring-operator.v0.2.0
        version: 0.2.0
      - name: power-monitoring-operator.v0.1.0
        version: 0.1.0
      name: tech-preview
    defaultChannel: tech-preview
    packageName: power-monitoring-operator
    provider:
      name: Red Hat, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtc-operator.v1.7.19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigrationController",
                  "metadata": {
                    "name": "migration-controller",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "azure_resource_group": "",
                    "cluster_name": "host",
                    "migration_velero": true,
                    "migration_controller": true,
                    "migration_ui": true,
                    "migration_log_reader": true,
                    "olm_managed": true,
                    "restic_timeout": "1h",
                    "mig_pv_limit": "100",
                    "mig_pod_limit": "100",
                    "mig_namespace_limit": "10"
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageMigration",
                  "metadata": {
                    "name": "directimagemigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageStreamMigration",
                  "metadata": {
                    "name": "directimagestreammigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigAnalytic",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "plan": {
                      "name": "foo",
                      "namespace": "openshift-migration"
                    },
                    "analyzeK8SResources": true,
                    "analyzeImageCount": true,
                    "analyzePVCapacity": true
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigCluster",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectVolumeMigration",
                  "metadata": {
                    "name": "direct",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                      "kind": "DirectVolumeMigrationProgress",
                      "metadata": {
                        "name": "direct-progress",
                        "namespace": "openshift-migration"
                      },
                      "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigPlan",
                  "metadata": {
                    "name": "migplan",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigMigration",
                  "metadata": {
                    "name": "migmigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigStorage",
                  "metadata": {
                    "name": "migstorage",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:89ba5ccb1cd9c44b8c8d2189f1fc6b04e780d46b19ee8a22e20c0b8c9d94d91f
          createdAt: "2025-04-28T19:55:38Z"
          description: Facilitates migration of container workloads from OpenShift
            3.x to OpenShift 4.x
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.5.0 <1.7.19'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-migration
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:3b6168c6df0d86d76ae4de91755c631b7f6a1f84286257d18c97d5a6af312291
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/mig-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: a request to migrate PVCs from a set of namespaces directly
              to a target
            displayName: DirectVolumeMigration
            kind: DirectVolumeMigration
            name: directvolumemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate watch the progress of rsync pod migrating
              a PVC to a target
            displayName: DirectVolumeMigrationProgress
            kind: DirectVolumeMigrationProgress
            name: directvolumemigrationprogresses.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a set of namespaces
              directly to a target
            displayName: DirectImageMigration
            kind: DirectImageMigration
            name: directimagemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a single imagestream
              directly to a target
            displayName: DirectImageStreamMigration
            kind: DirectImageStreamMigration
            name: directimagestreammigrations.migration.openshift.io
            version: v1alpha1
          - description: A request to Analyze a MigPlan and its namespaces
            displayName: MigAnalytic
            kind: MigAnalytic
            name: miganalytics.migration.openshift.io
            version: v1alpha1
          - description: A cluster defined for migration
            displayName: MigCluster
            kind: MigCluster
            name: migclusters.migration.openshift.io
            version: v1alpha1
          - description: A migration process
            displayName: MigMigration
            kind: MigMigration
            name: migmigrations.migration.openshift.io
            version: v1alpha1
          - description: A migration plan
            displayName: MigPlan
            kind: MigPlan
            name: migplans.migration.openshift.io
            version: v1alpha1
          - description: A migration contoller deployment
            displayName: MigrationController
            kind: MigrationController
            name: migrationcontrollers.migration.openshift.io
            version: v1alpha1
          - description: A migration hook
            displayName: MigHook
            kind: MigHook
            name: mighooks.migration.openshift.io
            version: v1alpha1
          - description: A migration storage location
            displayName: MigStorage
            kind: MigStorage
            name: migstorages.migration.openshift.io
            version: v1alpha1
        description: |-
          The Migration Toolkit for Containers Operator enables installation of the application migration tool components.

          After you have installed the Operator, you can install components by creating a MigrationController CR.

          By default, the Operator installs the following components on an OpenShift Container Platform 4 target cluster:

          * velero and restic, to back up data on the source cluster and restore it on the target cluster
          * mig-controller, to coordinate the backup and restore processes
          * mig-ui, the web console for managing migrations

          If you experience any issues or have feature requests, please file an [issue](https://github.com/konveyor/mig-operator/issues)

          For more information, see the [OpenShift Container Platform Migration Guide](https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html).
        displayName: Migration Toolkit for Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - crane
        - mtc
        links:
        - name: Migrations
          url: https://github.com/konveyor/mig-controller
        - name: Documentation
          url: https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html
        - name: Crane Operator
          url: https://github.com/konveyor/mig-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhmtc/openshift-migration-hook-runner-rhel8@sha256:c8c932a486d9b83e5a93290058d9e84691b87cd4161f8132a1bb6de9a360605e
        - registry.redhat.io/rhmtc/openshift-migration-velero-restic-restore-helper-rhel8@sha256:932216fd08d8ad97baf48ac7b8c989806d30a191337c650df37da537c153dc17
        - registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:3b6168c6df0d86d76ae4de91755c631b7f6a1f84286257d18c97d5a6af312291
        - registry.redhat.io/rhmtc/openshift-migration-velero-rhel8@sha256:bc010bc31e8b553f86aacbca39e28a95e74994c56e85c9fc1b8b80d0791ecb16
        - registry.redhat.io/rhmtc/openshift-migration-log-reader-rhel8@sha256:a0c8ab3a5bbc78c158c2ccccb1d06430a20f40358649196aa4e7311e54f9bdf8
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-aws-rhel8@sha256:023465788745625ad3b4dc0574e35b93d6ac8eea17be2ad1fc88959149a0df2c
        - registry.redhat.io/rhmtc/openshift-migration-ui-rhel8@sha256:66b288170dca8007382d13dff3f2e1868e409fb8002e04ec1fc5c480f868d9b6
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-gcp-rhel8@sha256:cc70c5a63af77948760bc42bf695a5e4eb0bce6285d6b350bba89a53f4b4e7e9
        - registry.redhat.io/rhmtc/openshift-migration-registry-rhel8@sha256:49485dfe708fe1294dec3f0bf34fedc38845a22f91915bba523d728a586b0c30
        - registry.redhat.io/rhmtc/openshift-migration-controller-rhel8@sha256:9b1490b04fc235191515fcd07d7af2e81f846d865016831e17872bb82bf4eaf5
        - registry.redhat.io/rhmtc/openshift-migration-rsync-transfer-rhel8@sha256:23a2bc5ad0f99fc0e7517a0172fdf3597fa619543b0192757983219a578f4ee2
        - registry.redhat.io/rhmtc/openshift-velero-plugin-rhel8@sha256:02698c655a01df2237aee6c570db96f13c0018cee2ae58b96b9570d93c354617
        - registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:89ba5ccb1cd9c44b8c8d2189f1fc6b04e780d46b19ee8a22e20c0b8c9d94d91f
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-microsoft-azure-rhel8@sha256:dad8fa4dbce3b4562bf06756fd74c2007812ec4edbc308e4517ff123c5519767
        version: 1.7.19
      entries:
      - name: mtc-operator.v1.7.19
        version: 1.7.19
      - name: mtc-operator.v1.7.18
        version: 1.7.18
      - name: mtc-operator.v1.7.17
        version: 1.7.17
      - name: mtc-operator.v1.7.16
        version: 1.7.16
      - name: mtc-operator.v1.7.15
        version: 1.7.15
      - name: mtc-operator.v1.7.14
        version: 1.7.14
      - name: mtc-operator.v1.7.13
        version: 1.7.13
      - name: mtc-operator.v1.7.12
        version: 1.7.12
      - name: mtc-operator.v1.7.11
        version: 1.7.11
      - name: mtc-operator.v1.7.10
        version: 1.7.10
      - name: mtc-operator.v1.7.9
        version: 1.7.9
      - name: mtc-operator.v1.7.8
        version: 1.7.8
      - name: mtc-operator.v1.7.7
        version: 1.7.7
      - name: mtc-operator.v1.7.6
        version: 1.7.6
      - name: mtc-operator.v1.7.5
        version: 1.7.5
      - name: mtc-operator.v1.7.4
        version: 1.7.4
      - name: mtc-operator.v1.7.3
        version: 1.7.3
      - name: mtc-operator.v1.7.2
        version: 1.7.2
      - name: mtc-operator.v1.7.1
        version: 1.7.1
      - name: mtc-operator.v1.7.0
        version: 1.7.0
      name: release-v1.7
    - currentCSV: mtc-operator.v1.8.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigrationController",
                  "metadata": {
                    "name": "migration-controller",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "azure_resource_group": "",
                    "cluster_name": "host",
                    "migration_velero": true,
                    "migration_controller": true,
                    "migration_ui": true,
                    "migration_log_reader": true,
                    "olm_managed": true,
                    "restic_timeout": "1h",
                    "mig_pv_limit": "100",
                    "mig_pod_limit": "100",
                    "mig_namespace_limit": "10"
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageMigration",
                  "metadata": {
                    "name": "directimagemigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageStreamMigration",
                  "metadata": {
                    "name": "directimagestreammigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigAnalytic",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "plan": {
                      "name": "foo",
                      "namespace": "openshift-migration"
                    },
                    "analyzeK8SResources": true,
                    "analyzeImageCount": true,
                    "analyzePVCapacity": true
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigCluster",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectVolumeMigration",
                  "metadata": {
                    "name": "direct",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                      "kind": "DirectVolumeMigrationProgress",
                      "metadata": {
                        "name": "direct-progress",
                        "namespace": "openshift-migration"
                      },
                      "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigPlan",
                  "metadata": {
                    "name": "migplan",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigMigration",
                  "metadata": {
                    "name": "migmigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigStorage",
                  "metadata": {
                    "name": "migstorage",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:8a44fe2aea3e842dcccb1e20df4818f4ca573fe13f51427ae8b603d0000a9411
          createdAt: "2025-08-22T15:27:29Z"
          description: Facilitates migration of container workloads from OpenShift
            3.x to OpenShift 4.x
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.5.0 <1.8.10'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-migration
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:f9492269b75608057fbc5edd53603807f1f80cd8581dea12032c88e04cf8f4cb
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/mig-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: a request to migrate PVCs from a set of namespaces directly
              to a target
            displayName: DirectVolumeMigration
            kind: DirectVolumeMigration
            name: directvolumemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate watch the progress of rsync pod migrating
              a PVC to a target
            displayName: DirectVolumeMigrationProgress
            kind: DirectVolumeMigrationProgress
            name: directvolumemigrationprogresses.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a set of namespaces
              directly to a target
            displayName: DirectImageMigration
            kind: DirectImageMigration
            name: directimagemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a single imagestream
              directly to a target
            displayName: DirectImageStreamMigration
            kind: DirectImageStreamMigration
            name: directimagestreammigrations.migration.openshift.io
            version: v1alpha1
          - description: A request to Analyze a MigPlan and its namespaces
            displayName: MigAnalytic
            kind: MigAnalytic
            name: miganalytics.migration.openshift.io
            version: v1alpha1
          - description: A cluster defined for migration
            displayName: MigCluster
            kind: MigCluster
            name: migclusters.migration.openshift.io
            version: v1alpha1
          - description: A migration process
            displayName: MigMigration
            kind: MigMigration
            name: migmigrations.migration.openshift.io
            version: v1alpha1
          - description: A migration plan
            displayName: MigPlan
            kind: MigPlan
            name: migplans.migration.openshift.io
            version: v1alpha1
          - description: A migration contoller deployment
            displayName: MigrationController
            kind: MigrationController
            name: migrationcontrollers.migration.openshift.io
            version: v1alpha1
          - description: A migration hook
            displayName: MigHook
            kind: MigHook
            name: mighooks.migration.openshift.io
            version: v1alpha1
          - description: A migration storage location
            displayName: MigStorage
            kind: MigStorage
            name: migstorages.migration.openshift.io
            version: v1alpha1
        description: |-
          The Migration Toolkit for Containers Operator enables installation of the application migration tool components.

          After you have installed the Operator, you can install components by creating a MigrationController CR.

          By default, the Operator installs the following components on an OpenShift Container Platform 4 target cluster:

          * velero and restic, to back up data on the source cluster and restore it on the target cluster
          * mig-controller, to coordinate the backup and restore processes
          * mig-ui, the web console for managing migrations

          If you experience any issues or have feature requests, please file an [issue](https://github.com/konveyor/mig-operator/issues)

          For more information, see the [OpenShift Container Platform Migration Guide](https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html).
        displayName: Migration Toolkit for Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - crane
        - mtc
        links:
        - name: Migrations
          url: https://github.com/konveyor/mig-controller
        - name: Documentation
          url: https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html
        - name: Crane Operator
          url: https://github.com/konveyor/mig-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhmtc/openshift-migration-registry-rhel8@sha256:accef2a0dd69b254bd120232e5dc86a56aa8f25bcfdba3f9bab7e4127c157762
        - registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:f9492269b75608057fbc5edd53603807f1f80cd8581dea12032c88e04cf8f4cb
        - registry.redhat.io/rhmtc/openshift-migration-controller-rhel8@sha256:c776387c662e21aa61378d0136b2495826e1c157548f9ca9c95d044af38ee169
        - registry.redhat.io/rhmtc/openshift-migration-hook-runner-rhel8@sha256:adaa4f1afefb4d5853713821cb6e7de0d9f5a640f0243f20b41a44be84707caf
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-mtc-rhel8@sha256:5f46ba04146438f2abd6905468d2ecbeb19374a83771d364904ac2ee6afdbefa
        - registry.redhat.io/rhmtc/openshift-migration-openvpn-rhel8@sha256:d21b99d8edefa7a1bab30c85969ec7d7560d06be308bd787e698e409d54cbc19
        - registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:8a44fe2aea3e842dcccb1e20df4818f4ca573fe13f51427ae8b603d0000a9411
        - registry.redhat.io/rhmtc/openshift-migration-log-reader-rhel8@sha256:4f76bcf3df3bb7073b8aabd3da072cacfec05ecf4bfb47e23340074756f7fc30
        - registry.redhat.io/rhmtc/openshift-migration-rsync-transfer-rhel8@sha256:442e7c7f2cc075bf468403ace8846c96ad2fe43eb7f6562d566caa24e4f86873
        - registry.redhat.io/rhmtc/openshift-migration-ui-rhel8@sha256:86e747e0330b0156379d273d0c29c13a65dafbf4f536656723e2bc746872286c
        version: 1.8.10
      entries:
      - name: mtc-operator.v1.8.10
        version: 1.8.10
      - name: mtc-operator.v1.8.9
        version: 1.8.9
      - name: mtc-operator.v1.8.8
        version: 1.8.8
      - name: mtc-operator.v1.8.7
        version: 1.8.7
      - name: mtc-operator.v1.8.6
        version: 1.8.6
      - name: mtc-operator.v1.8.5
        version: 1.8.5
      - name: mtc-operator.v1.8.4
        version: 1.8.4
      - name: mtc-operator.v1.8.3
        version: 1.8.3
      - name: mtc-operator.v1.8.2
        version: 1.8.2
      - name: mtc-operator.v1.8.1
        version: 1.8.1
      - name: mtc-operator.v1.8.0
        version: 1.8.0
      name: release-v1.8
    defaultChannel: release-v1.8
    packageName: mtc-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Performance
      provider-url: ""
    name: cluster-impairment-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-impairment-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "all-impairments"
                },
                "spec": {
                  "duration": 480,
                  "egress": {
                    "bandwidth": 10000,
                    "latency": 50,
                    "loss": 0.02
                  },
                  "ingress": {
                    "bandwidth": 10000,
                    "latency": 50,
                    "loss": 0.02
                  },
                  "interfaces": [
                    "ens2f0"
                  ],
                  "linkFlapping": {
                    "downTime": 30,
                    "enable": true,
                    "upTime": 120
                  },
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "duplication"
                },
                "spec": {
                  "duration": 60,
                  "egress": {
                    "duplication": 25
                  },
                  "ingress": {
                    "duplication": 25
                  },
                  "interfaces": [
                    "ens3"
                  ],
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "two-min-flap"
                },
                "spec": {
                  "duration": 480,
                  "interfaces": [
                    "ens2f0"
                  ],
                  "linkFlapping": {
                    "downTime": 120,
                    "enable": true,
                    "upTime": 120
                  },
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "typical-scenario"
                },
                "spec": {
                  "duration": 30,
                  "egress": {
                    "latency": 50
                  },
                  "ingress": {
                    "latency": 50
                  },
                  "interfaces": [
                    "ens2f0",
                    "eno1"
                  ],
                  "loss": 0.02,
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "uneven-latency"
                },
                "spec": {
                  "duration": 60,
                  "egress": {
                    "latency": 100
                  },
                  "ingress": {
                    "latency": 10
                  },
                  "interfaces": [
                    "ens2f0"
                  ],
                  "nodeSelector": {
                    "key": "node-role.kubernetes.io/worker",
                    "value": ""
                  },
                  "startDelay": 5
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Developer Tools
          containerImage: quay.io/redhat-performance/cluster-impairment-operator:1.0.4
          createdAt: "2021-08-11 12:15:00"
          description: Run Impairments on Kubernetes/Openshift Nodes
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/redhat-performance/cluster-impairment-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterImpairment
            name: clusterimpairments.apps.redhat.com
            version: v1alpha1
        description: "cluster-impairment-operator is an operator designed to automate
          the application of impairments to the nodes in a cluster.\n\n## Features\n\nTraffic
          Direction:\n\n* Egress impairments affect the traffic going out.    \n*
          Ingress impairments affect the traffic coming in. (Ingress impairments require
          the kernel module IFB.)\n\n### Impairments\n\n| Impairment    | Description
          \                               | Unit    | Uses Netem | Correlation Supported
          |\n|---------------|--------------------------------------------|---------|------------|-----------------------|\n|
          Bandwidth     | The bandwidth limit                        | kbit/s  | Yes
          \       | No                    |\n| Latency       | The delay of the packets
          \                  | ms      | Yes        | Yes                   |\n| Packet
          Loss   | The percent of packets that are dropped    | percent | Yes        |
          Yes                   |\n| Corruption    | The percent of packets that are
          corrupted  | percent | Yes        | Yes                   |\n| Duplication
          \  | The percent of packets that are duplicated | percent | Yes        |
          Yes                   |\n| Link Flapping | Turns the interface on and off
          \            | bool    | No         | No                    |\n\n\nOn the
          tested environment (RHEL CoreOS 48.84), the impairments can be used alongside
          link flapping.\n\n\n#### Latency Options\n\nIn addition to simply delaying
          the packets, there are advanced latency options.\n\n##### Jitter\n\nSemi-randomly
          adds or subtracts from the latency according to the distribution up to the
          amount specified.\nIf latency is 100ms, and jitter is 10ms, the actual latency
          will vary 100±10ms.\n\n\n##### Distribution\n\nThe distribution of the jitter.
          The options are:\n* Normal\n* Uniform\n* Pareto\n* Paretonormal\n\n#####
          Jitter correlation\n\nThe percent chance that the next latency's jitter
          value will correlate with the preceeding latency.\n\n##### Reorder\n\nThe
          percentage of packets that are not delayed, causing a reorder due to them
          being sent before the delayed ones.\nApplying jitter itself has the potential
          to also cause reordering.\n\n##### Reorder correlation\n\nThe percet chance
          that the value for the next reorder will correlate with the preceeding value.\n\n####
          Loss options\n\n##### Correlation\n\nThe percent chance that the previous
          loss value for a packet correlates with the loss value for the next packet.\n\n####
          Corruption options\n\n###### Corruption\n\nThe percent chance that each
          packet will be corrupted.\n\n##### Corruption correlation\n\nThe percet
          chance that the value for the next corruption will correlate with the preceeding
          value.\n\n#### Duplication options\n\n##### Duplication\n\nThe percent chance
          that each packet will be duplicated.\n\n##### Duplication correlation\n\nThe
          percet chance that the value for the next duplication will correlate with
          the preceeding value.\n\n\n\n## Configuration\n\nHere is an example of the
          ClusterImpairment custom resource.\n\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: test-impairment-cr\nspec:\n  duration:
          30 # seconds\n  startDelay: 5 # seconds. It typically takes about 2-3 seconds
          for the Daemonset to run\n  interfaces:\n  - \"ens2f0\"\n  ingress: # uses
          ifb\n    bandwidth: 0 # kbit\n    latency: 10 # ms\n    latencyOptions:\n
          \     jitter: 5 # ms\n      jitterCorrelation: 25 # percent\n      distribution:
          normal\n      reorder: 25 # percent of packets that will skip the delay\n
          \     reorderCorrelation: 25 # percent\n    loss: 0 # percent\n    lossOptions:\n
          \     correlation: 25 # percent\n    corruption: 0.1 # percent\n    corruptionOptions:\n
          \     correlation: 25 # percent\n    duplication: 1 # percent\n    duplicationOptions:\n
          \     correlation: 25 # percent\n  egress:\n    bandwidth: 0 # kbit\n    latency:
          100 # ms\n    loss: 0 # percent\n  linkFlapping:\n    enable: false\n    downTime:
          3 # Seconds\n    upTime: 3 # Seconds\n  nodeSelector:\n    key: \"node-role.kubernetes.io/worker\"\n
          \   value: \"\"\n    \n```\n\n#### Interfaces\n\nThe interfaces configuration
          option is a list of all interfaces that the impairments should be applied
          to. The current implementation will error out once it hits an invalid interface.\n\nIf
          an invalid interface is found, it will print out the list of interfaces.\n\n####
          Node Selector\n\nThe default node selector is all worker nodes, but you
          can change it to whatever node selector you want by setting the key and
          value.\nThere is a toleration to allow it to run on master nodes if the
          node selector includes master nodes.\n\n#### Duration\n\nThe number of seconds
          the impairments should run. It will try to sync the start and end time between
          all pods.\nIf set to -1, it will run indefinitely (a year), until you remove
          the ClusterImpairment custom resource. This is good for testing that requires
          steady impairments.\n\nIf the script is link flapping, set this value to
          be short enough so that if the link flap interrupts communication between
          the nodes, the daemonset will remove itself.\n\n#### Start Delay\n\nThe
          delay before starting the script. If you want the pods to be in sync, a
          minimum of a few seconds should be used because the pods take time to start
          up.\n\nYou can also utilize this feature to run an separate impairment after
          another. Just apply two resources at the same time, but add the duration
          and start delay of the first to the start delay of the second.\n\n#### Ingress
          and Egress\n\nThe configuration sections \"ingress\" and \"egress\" apply
          to each direction.\n\n##### Examples:\n\n##### Example 1\n\n\nIn this example,
          egress latency is set to 100ms, but the ingress latency is set to 10ms.
          So the latency to the interface will end up being 10ms, but 100ms going
          out. When pinging, this will result in 110ms of latency.\n\n```\n\napiVersion:
          apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n  name: uneven-latency\nspec:\n
          \ duration: 60\n  startDelay: 5\n  interfaces:\n  - \"ens2f0\"\n  ingress:\n
          \   latency: 10 # ms\n  egress:\n    latency: 100 # ms\n\n```\n\n##### Example
          2\n\n\nIn this example, link flapping will be enabled, and it will turn
          the interface on and off every 2 minutes. That is long enough for kubernetes
          to determine that a node is unavailable.\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: two-min-flap\nspec:\n  duration: 480\n
          \ startDelay: 5\n  interfaces:\n  - \"ens2f0\"\n  linkFlapping:\n    enable:
          true\n    downTime: 120 # Seconds\n    upTime: 120 # Seconds\n\n```\n\n#####
          Example 3\n\n\nIn this example, a realistic set of impairments is applied
          to `ens2f0` and `eno1` for 30 seconds:\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: typical-scenario\nspec:\n  duration:
          30 # seconds\n  startDelay: 5 # seconds\n  interfaces:\n  - \"ens2f0\"\n
          \ - \"eno1\"\n  egress:\n    latency: 50 # ms. Bidirectional, so total of
          100ms\n  ingress:\n    latency: 50 # ms. Bidirectional, so total of 100ms\n
          \ loss: 0.02 # percent\n\n```\n\n##### Example 4\n\n\nIn this example, a
          combination of latency, packet loss, bandwidth, and link flapping is applied.\n\n```\n\napiVersion:
          apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n  name: all-impairments\nspec:\n
          \ duration: 480 # seconds\n  startDelay: 5 # seconds\n  interfaces:\n  -
          \"ens2f0\"\n  egress:\n    latency: 50 # ms. Bidirectional, so total of
          100ms\n    loss: 0.02 # percent\n    bandwidth: 1000 # 1000 kbit/s, about
          1 mbit/s\n  ingress:\n    latency: 50 # ms. Bidirectional, so total of 100ms\n
          \   loss: 0.02 # percent\n    bandwidth: 1000 # 1000 kbit/s, about 1 mbit/s\n
          \ linkFlapping:\n    enable: true\n    downTime: 30 # Seconds\n    upTime:
          120 # Seconds\n\n```\n\n##### Example 5\n\n\nIn this example, packet duplicaiton
          is applied.\n\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n
          \ name: duplication\nspec:\n  duration: 60\n  startDelay: 5\n  interfaces:\n
          \ - \"ens3\"\n  ingress:\n    duplication: 25 # percent\n  egress:\n    duplication:
          25 # percent\n\n```\n\n## Setup\n\n### Requirements\n\n1. You need `make`
          installed.\n2. You need access to the kubernetes cluster with a kubeconfig.\n\n###
          Installation\n\nTo run using the current latest image:\n1. Clone the repository.\n2.
          Run `make deploy` with KUBECONFIG environment variable set to the location
          of the kubeconfig.\n\nTo run with your own code, there are more steps.\n1.
          Fork the repository.\n2. Clone to a machine that has access to the Kubernetes
          cluster and the kubeconfig.\n3. Modify the makefile to change the `IMG`
          variable to your image repository. If you do not have podman installed,
          also change podman to docker.\n4. Run `make docker-build` then `make docker-push`.\n5.
          Then run `make deploy`\n\n### Deploying from operator-hub\n\n*Important*
          You must manually apply the serviceaccount!\n\nFirst, Follow the instructions
          on this site to deploy from here.\n\nNext, apply the manifest \"serviceaccount.yaml\"
          from the manifests folder.\n\nThen you can apply your ClusterImpairment
          resources.\n\n## Running impairments\n\nFirst, configure a ClusterImpairment
          type's spec with the information for the impairment you want to run. Most
          importantly, set the interface(s). You can list the interfaces with `ifconfig`.
          If an invalid interface is set, the pod's logs will show `ifconfig` output.\n\n\nOnce
          the clusterimpairment type is set, apply it and it will work.\n\n\nTo re-run
          an impairment, delete the impairment with `oc delete -f path/to/manifest.yaml`,
          then apply it again.\n\n\n## Limitations\n\n\n### Multiple Impairments\n\n\nYou
          should avoid any configurations that apply impairments to the same interface
          on the same node. There are potential conflicts. The worker pod will attempt
          to remove all impairments on that interface before applying new ones.\n\n\nInstead,
          take advantage of the full control of both ingress and egress impairments
          from within the same ClusterImpairment resource.\n\n\nThere is currently
          a configured limit of 20 ClusterImpairment resources working at the same
          time. If that is a limitation for your use case, you can change it in `config/manager/manager.yaml`
          or you can try to find a way to make each ClusterImpairment do more.\n\n###
          Traffic Control (TC)\n\nTraffic control is how cluster-impairment-operator
          applies the latency, bandwidth, and packet loss impairments. The limitation
          is due to Linux not being realtime operating system, so the impairment will
          not be perfectly consistent.\n\n### Link Flapping\n\nWhen link flapping,
          if you flap the link that Kubernetes uses to communicate with the pods,
          you may be unable to remove the pod until the link is on long enough for
          Kubernetes to communicate with the impaired node.\n\n\nIn this case, it
          is helpful to set the duration properly instead of running for an indefinite
          or large amount of time, because the node will properly go back to the unimpaired
          state at that time.\n\n\nIf the cluster becomes offline due to the link
          flapping when you do not want it to be offline, soft restarting the nodes
          after removing the custom resource should remove all impairments.\n"
        displayName: cluster-impairment-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network
        - impairment
        - kubernetes
        - openshift
        - tc
        - trafficcontrol
        links:
        - name: Cluster Impairment Operator
          url: https://github.com/redhat-performance/cluster-impairment-operator
        maintainers:
        - email: joconnel@redhat.com
          name: Jared O'Connell
        - email: akrzos@redhat.com
          name: Alex Krzos
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat Performance
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/redhat-performance/cluster-impairment-operator:1.0.4
        version: 1.0.4
      entries:
      - name: cluster-impairment-operator.v1.0.4
        version: 1.0.4
      - name: cluster-impairment-operator.v1.0.3
        version: 1.0.3
      - name: cluster-impairment-operator.v1.0.2
        version: 1.0.2
      - name: cluster-impairment-operator.v1.0.1
        version: 1.0.1
      name: beta
    defaultChannel: beta
    packageName: cluster-impairment-operator
    provider:
      name: Red Hat Performance
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: mercury-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mercury-operator.1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:1.0.1",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CustomerCreditRating",
                  "expose": ["http"]
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:1.0.1
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:1.0.1
        version: 1.0.1
      entries:
      - name: mercury-operator.1.0.1
        version: 1.0.1
      name: 1.0.x
    - currentCSV: mercury-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_CREDIT_RATING",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-offer",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-offer-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_OFFER",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-party-routing-profile",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/party-routing-profile-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "PARTY_ROUTING_PROFILE",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-example-sdi": {
                "description": "Example Service Domain Infrastructure"
            },
            {
             "example-customer-credit-rating": {
               "description": "Example Customer Credit Rating Service Domain"
            },
            {
             "example-customer-offer": {
               "description": "Example Customer Offer Service Domain"
            },
            {
             "example-party-routing-profile": {
               "description": "Example Party Routing Profile Service Domain"
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:v1.0.2
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - description: Represents a Kafka cluster
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Integration is the Schema for the Camel K integrations API
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:v1.0.2
        version: 1.0.2
      entries:
      - name: mercury-operator.v1.0.2
        version: 1.0.2
      name: dev
    - currentCSV: mercury-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_CREDIT_RATING",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-offer",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-offer-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_OFFER",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-party-routing-profile",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/party-routing-profile-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "PARTY_ROUTING_PROFILE",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-example-sdi": {
                "description": "Example Service Domain Infrastructure"
            },
            {
             "example-customer-credit-rating": {
               "description": "Example Customer Credit Rating Service Domain"
            },
            {
             "example-customer-offer": {
               "description": "Example Customer Offer Service Domain"
            },
            {
             "example-party-routing-profile": {
               "description": "Example Party Routing Profile Service Domain"
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:v1.0.2
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - description: Represents a Kafka cluster
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Integration is the Schema for the Camel K integrations API
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:v1.0.2
        version: 1.0.2
      entries:
      - name: mercury-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: 1.0.x
    packageName: mercury-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-ati-updates
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-ati-updates.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ati.loadcore.keysight.com/v1alpha1",
                "kind": "Updates",
                "metadata": {
                  "name": "updates-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "installWait": {
                    "enabled": true,
                    "image": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:40:21Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Updates
            name: updates.ati.loadcore.keysight.com
            version: v1alpha1
        description: Keysight ATI Updates Operator
        displayName: Keysight ATI Updates Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Ati Updates
          url: https://keysight-ati-updates.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-ati-updates@sha256:abda76eb50d77e5a95191c9dbb21eefe64af5b794511ceafdfda007c15449c6e
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        version: 0.1.21
      entries:
      - name: keysight-ati-updates.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-ati-updates
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: CyberArk
      provider-url: https://www.cyberark.com/
    name: vcp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vcp-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "installer.venafi.com/v1alpha1",
                "kind": "VenafiInstall",
                "metadata": {
                  "name": "install-venafi-enhanced-issuer"
                },
                "spec": {
                  "certManager": {
                    "install": true
                  },
                  "globals": {
                    "imagePullSecretNames": [
                      "venafi"
                    ]
                  },
                  "venafiConnection": {
                    "install": true
                  },
                  "venafiEnhancedIssuer": {
                    "install": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:9da79fd0941c21e433f22aaaccc43ebbd4646e685aea0050194954d8b4019c05
          createdAt: "2024-05-30T11:35:41Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              VenafiInstall is an object for installing Venafi components onto a Kubernetes
              cluster
            displayName: VenafiInstall
            kind: VenafiInstall
            name: venafiinstalls.installer.venafi.com
            version: v1alpha1
        description: |
          Venafi Control Plane Operator for Red Hat OpenShift simplifies deployment of Venafi's enterprise components and dependencies.

          The enterprise components include:
          - Enterprise cert-manager: An enhanced version of cert-manager designed for application certificate management within your clusters.
          - Venafi Enhanced Issuer: Enables seamless certificate enrollment for your clusters directly from the Venafi Control Plane.
          - Venafi Firefly: A high-performance, lightweight issuer for certificates, centrally managed by the Venafi Control Plane.
          - Venafi Kubernetes Agent: Provides real-time visibility into certificates and Kubernetes resources directly within the Venafi Control Plane.
        displayName: Venafi Control Plane Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - venafi
        - vcp-operator
        links:
        - name: Documentation
          url: https://docs.venafi.cloud/
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Venafi
          url: https://venafi.com
        relatedImages:
        - registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:9da79fd0941c21e433f22aaaccc43ebbd4646e685aea0050194954d8b4019c05
        version: 1.0.0
      entries:
      - name: vcp-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: vcp-operator.v1.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "installer.venafi.com/v1alpha1",
                "kind": "VenafiInstall",
                "metadata": {
                  "name": "install-venafi-enhanced-issuer"
                },
                "spec": {
                  "certManager": {
                    "install": true
                  },
                  "globals": {
                    "imagePullSecretNames": [
                      "venafi"
                    ]
                  },
                  "venafiConnection": {
                    "install": true
                  },
                  "venafiEnhancedIssuer": {
                    "install": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:eb2b06d191dbfbb02332375a7bca6bc010d5410e53b3bc5fb8c320101292408c
          createdAt: "2025-07-22T06:44:30Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              VenafiInstall is an object for installing Venafi components onto a Kubernetes
              cluster
            displayName: VenafiInstall
            kind: VenafiInstall
            name: venafiinstalls.installer.venafi.com
            version: v1alpha1
        description: |
          The operator, formerly known as the Venafi Control Plane Operator, simplifies the deployment of enterprise components part of the CyberArk Certificate Manager for Kubernetes.

          The components include:
          - cert-manager: An enterprise distribution of cert-manager designed for application certificate management within your clusters.
          - Enterprise Issuer: Enables seamless certificate enrollment directly from the Certificate Manager.
          - Workload Identity Manager: A high-performance issuer for certificates, centrally managed by the Certificate Manager.
          - Agent: Provides real-time visibility into cluster resources directly within the Certificate Manager.
        displayName: CyberArk Certificate Manager Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cyberark
        - vcp-operator
        links:
        - name: Documentation
          url: https://docs.venafi.cloud/vaas/k8s-components/c-vcpo-overview/
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:eb2b06d191dbfbb02332375a7bca6bc010d5410e53b3bc5fb8c320101292408c
        version: 1.8.1
      entries:
      - name: vcp-operator.v1.8.1
        version: 1.8.1
      - name: vcp-operator.v1.8.0
        version: 1.8.0
      - name: vcp-operator.v1.7.0
        version: 1.7.0
      - name: vcp-operator.v1.6.0
        version: 1.6.0
      - name: vcp-operator.v1.5.0
        version: 1.5.0
      - name: vcp-operator.v1.4.0
        version: 1.4.0
      - name: vcp-operator.v1.3.0
        version: 1.3.0
      - name: vcp-operator.v1.2.1
        version: 1.2.1
      - name: vcp-operator.v1.2.0
        version: 1.2.0
      - name: vcp-operator.v1.1.1
        version: 1.1.1
      name: stable
    defaultChannel: stable
    packageName: vcp-operator
    provider:
      name: CyberArk
      url: https://www.cyberark.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE Ezmeral
      provider-url: ""
    name: hpe-ezmeral-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-ezmeral-csi-operator.v1.0.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ezmeral.hpe.com/v1",
                "kind": "HPEEzmeralCSIDriver",
                "metadata": {
                  "name": "hpeezmeralcsidriver"
                },
                "spec": {
                  "controllerImage": "registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03",
                  "nodeImage": "registry.connect.redhat.com/maprtech/csi-kdfplugin@sha256:ffd273ee2e608d2ef64fd3b2bcbb02d5f0a6c173ed96a15eaf5549ea45fea8d1",
                  "pullPolicy": "Always"
                }
              },
              {
                "apiVersion": "ezmeral.hpe.com/v1",
                "kind": "HPEEzmeralNFSCSIDriver",
                "metadata": {
                  "name": "hpeezmeralnfscsidriver"
                },
                "spec": {
                  "controllerImage": "registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03",
                  "nodeImage": "registry.connect.redhat.com/maprtech/csi-nfsplugin@sha256:a5c954f62f1395bafe2d68454ac604bca4be97573a3482fcbd8e76b6361cc08a",
                  "pullPolicy": "Always"
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator@sha256:c9c243c51cd760c2c2a66b489807a5e22053fcbeb94fa29afd10d9c91e2d3908
          createdAt: "2025-06-25T11:51:05Z"
          description: A Container Storage Interface (CSI) driver for HPE Ezmeral
            Data Fabric platform. The CSI driver allows you to use HPE Ezmeral Data
            Fabric with your preferred container orchestrator.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mapr/mapr-csi
          support: HPE Ezmeral Data Fabric
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE Ezmeral CSI Driver (FUSE)
            displayName: HPEEzmeralCSIDriver
            kind: HPEEzmeralCSIDriver
            name: hpeezmeralcsidrivers.ezmeral.hpe.com
            version: v1
          - description: HPE Ezmeral CSI Driver (NFS)
            displayName: HPEEzmeralNFSCSIDriver
            kind: HPEEzmeralNFSCSIDriver
            name: hpeezmeralnfscsidrivers.ezmeral.hpe.com
            version: v1
        description: |
          The HPE Ezmeral CSI Operator for Kubernetes packages, deploys, manages, upgrades the HPE Ezmeral CSI Driver on Kubernetes and OpenShift for dynamic provisioning of persistent volumes on HPE Ezmeral Data Fabric platform.
          The HPE Ezmeral CSI Driver for Kubernetes leverages HPE Ezmeral Data Fabric platform to provide scalable and persistent storage for stateful applications. Please refer to our CSI driver [documentation](https://docs.datafabric.hpe.com/home/CSIdriver/csi_overview.html) for full list of supported CSI features.
          ## Installation
          Refer to the HPE Ezmeral CSI Operator for Kubernetes [documentation](https://docs.datafabric.hpe.com/home/CSIdriver/csi_installing_operator.html)
        displayName: HPE Ezmeral Data Fabric CSI Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE Ezmeral
        - HPE Ezmeral Data Fabric
        - Storage
        - FileSystem
        - BlockVolume
        links:
        - name: Documentation
          url: https://docs.datafabric.hpe.com/home/CSIdriver/csi_overview.html
        maintainers:
        - email: ezmeral-csi-support@hpe.com
          name: HPE Ezmeral Data Fabric
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: HPE Ezmeral
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6ba961c2c2a29750c0132fe6dd6fa9f6001010afbc5f19b98add87b31b54bcf6
        - registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator@sha256:c9c243c51cd760c2c2a66b489807a5e22053fcbeb94fa29afd10d9c91e2d3908
        - registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03
        - registry.connect.redhat.com/maprtech/csi-kdfplugin@sha256:ffd273ee2e608d2ef64fd3b2bcbb02d5f0a6c173ed96a15eaf5549ea45fea8d1
        - registry.connect.redhat.com/maprtech/csi-nfsplugin@sha256:a5c954f62f1395bafe2d68454ac604bca4be97573a3482fcbd8e76b6361cc08a
        version: 1.0.18
      entries:
      - name: hpe-ezmeral-csi-operator.v1.0.18
        version: 1.0.18
      name: stable
    defaultChannel: stable
    packageName: hpe-ezmeral-csi-operator
    provider:
      name: HPE Ezmeral
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzo-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzo-operator.v3.1.200
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzo.cambridgesemantics.com/v2",
                "kind": "Anzo",
                "metadata": {
                  "name": "no-persistence"
                },
                "spec": {
                  "jsonActivation": true,
                  "jvmMemory": "4096m",
                  "nodeConfig": {
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "app": "anzo"
                        }
                      },
                      "serviceName": "anzo-no-persistence",
                      "template": {
                        "metadata": {
                          "annotations": {
                            "cluster-autoscaler.kubernetes.io/safe-to-evict": "false"
                          },
                          "labels": {
                            "app": "anzo"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "registry.connect.redhat.com/cambridgesemantics/anzo@sha256:5b192604338f602fdabe55fecefd9eb82400a77c5aa7f04bd8e54aead8644c6d",
                              "imagePullPolicy": "Always",
                              "name": "anzo",
                              "resources": {
                                "limits": {
                                  "cpu": "2.0",
                                  "memory": "8.0Gi"
                                },
                                "requests": {
                                  "cpu": "2.0",
                                  "memory": "8.0Gi"
                                }
                              }
                            }
                          ],
                          "initContainers": [
                            {
                              "name": "anzo-init",
                              "resources": {
                                "limits": {
                                  "cpu": "0.5",
                                  "memory": "2Gi"
                                },
                                "requests": {
                                  "cpu": "0.5",
                                  "memory": "2Gi"
                                }
                              }
                            }
                          ],
                          "securityContext": {
                            "fsGroup": 1001
                          },
                          "serviceAccountName": "anzo-operator"
                        }
                      }
                    }
                  },
                  "role": "AnzoServer"
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/anzo-operator@sha256:4c0c8bf7acbfbb96f15459a20364608e5d0a0933c6326a89d219666ee0d88ca2
          createdAt: "2025-07-09T06:52:12Z"
          description: kubernetes operator for Altair Graph Studio
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Altair Customer Support
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoBackup is the Schema for the anzobackups API
            displayName: Altair Graph Studio Backup
            kind: AnzoBackup
            name: anzobackups.anzobackup.cambridgesemantics.com
            version: v1beta1
          - description: Altair Graph Studio is the Schema for the anzos API
            displayName: Altair Graph Studio Operator
            kind: Anzo
            name: anzos.anzo.cambridgesemantics.com
            version: v2
        description: |-
          The Altair Graph Studio Operator provides the way to install and configure an altair graph studio server and agent setup on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v3.1.2/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/deploy-k8s.htm )

          ### Support

          We offer Support to our customers [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Studio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        links:
        - name: Altair Graph Studio Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v3.1.2/README_openshift.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzo-operator@sha256:4c0c8bf7acbfbb96f15459a20364608e5d0a0933c6326a89d219666ee0d88ca2
        - registry.connect.redhat.com/cambridgesemantics/anzo@sha256:5b192604338f602fdabe55fecefd9eb82400a77c5aa7f04bd8e54aead8644c6d
        version: 3.1.200
      entries:
      - name: anzo-operator.v3.1.200
        version: 3.1.200
      - name: anzo-operator.v3.1.106
        version: 3.1.106
      - name: anzo-operator.v3.1.105
        version: 3.1.105
      name: stable
    defaultChannel: stable
    packageName: anzo-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq7-interconnect-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: 1.10.x
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: 1.2.0
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: latest
    defaultChannel: 1.10.x
    packageName: amq7-interconnect-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-registry
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-operator.v1.1.3-v2.6.4.final
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-mem"
                },
                "spec": {
                  "configuration": {
                    "persistence": "mem"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-operator:1.1.3
          createdAt: "2024-09-17"
          description: Deploy and manage Apicurio Registry on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents an Apicurio Registry instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Apicurio Registry

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Provides a REST API and web UI to manage the artifacts and artifact meta-data
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, in-memory
          - Compatible with Confluent and IBM APIs
          - Runs on a lightweight Quarkus platform
          - Includes Maven plugin to integrate with Maven based builds

          ## Apicurio Registry Operator

          Provides a quick and easy way to deploy and manage Apicurio Registry on Kubernetes.

          **Features**
          - Supports installation and configuration of the Registry for all storage options
          - Easily perform a rolling upgrade of Apicurio Registry

          ## Prerequisites

          This operator does not deploy storage for Apicurio Registry. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/main/LICENSE)
        displayName: Apicurio Registry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-operator:1.1.3
        version: 1.1.3-v2.6.4.final
      entries:
      - name: apicurio-registry-operator.v1.1.3-v2.6.4.final
        version: 1.1.3-v2.6.4.final
      - name: apicurio-registry-operator.v1.1.2-v2.5.11.final
        version: 1.1.2-v2.5.11.final
      - name: apicurio-registry-operator.v1.1.1-v2.5.9.final
        version: 1.1.1-v2.5.9.final
      - name: apicurio-registry-operator.v1.1.0-v2.4.12.final
        version: 1.1.0-v2.4.12.final
      - name: apicurio-registry-operator.v1.0.0-v2.0.0.final
        version: 1.0.0-v2.0.0.final
      name: 2.x
    defaultChannel: 2.x
    packageName: apicurio-registry
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zettaset
      provider-url: ""
    name: xcrypt-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zts-xcrypt-operator.v0.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xcrypt.zettaset.com/v1alpha1",
                "kind": "XCrypt",
                "metadata": {
                  "name": "zts-masterset"
                },
                "spec": {
                  "status": {
                    "podNames": "zts-ca",
                    "replicas": "1,"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.connect.redhat.com/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
          createdAt: "2021-06-16T12:29:13.97Z"
          description: XCrypt Operator deploys Zettaset XCrypt Container Encryption  -
            the only software-defined solution that offers transparent high performance
            data-at-rest encryption of critical data in container environments.
          support: Zettaset
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XCrypt is the Schema for the xcrypts API
            kind: XCrypt
            name: xcrypts.xcrypt.zettaset.com
            version: v1alpha1
        description: "Zettaset’s RedHat Certified XCrypt Operator allows users to
          deploy multiple XCrypt components on OpenShift with just a few simple commands.
          The XCrypt Operator deployment provides software-defined encryption that
          transparently protects container data throughout your OpenShift environment.\n\n##
          Core Features\n* **Automated XCrypt provisioning** - Deploying a Zettaset
          XCrypt has never been easier. Adjust cluster specific configuration and
          let the Zettaset Operator take care of provisioning XCrypt services and
          setting up the XCrypt encryption to your exact specification. \n* **CSI
          Driver deployment**: Zettaset XCrypt Operator automatically deploys Zettaset
          XCrypt CSI Driver in order to provide automated and transparent data-at-rest
          \ encryption functionality compatible with Openshift and any other container
          orchestrator. Once the Operator is provisioned, the driver runs fine tuned
          and ready to work with other XCrypt components.\n* **Multi-node topology**:
          Encrypt as many nodes and volumes as you need to. XCrypt components can
          be installed on multiple nodes across the cluster. This is defined by two
          labels: zts-master and zts-worker. Normally zts-master node is a single
          node that runs Zettaset XCrypt major services required for the product operation.
          Multiple nodes can be marked as zts-worker and those are the nodes that
          can have encrypted volumes.\n* **Seamless Encrypted Volume mount**: Volumes
          are not only being encrypted, but also are mounted/unmounted to containers
          with Persistent Volume Claims. This happens as seamlessly as regular volumes
          provision and mount/unmount.\n* **Robust failover**: A node failure doesn’t
          affect XCrypt operations on other nodes of the cluster. Encryption on a
          restarted/recovered node gets back to normal functioning if the node stays
          healthy and data hasn’t been corrupted during the restart\n\n## About Zettaset
          XCrypt Kubernetes Encryption for OpenShift\nXCrypt Kubernetes Encryption
          for OpenShift is a software-only solution that enables encryption of data
          at rest stored in Red Hat OpenShift container environments without slowing
          down the DevSecOps process. In addition to providing a transparent and high-performance
          layer of security that flexibly protects container data across on-premise,
          in hybrid, and multi-cloud environments, it also helps your organization
          efficiently transition from DevOps to DevSecOps.  \n\nMore information about
          Zettaset XCrypt Kubernetes Encryption for OpenShift can be found in the
          [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)\n\n##
          RedHat OpenShift + Zettaset: The next generation of secure enterprise infrastructure
          for DevSecOps  \nBy combining RedHat’s Container Platform with Zettaset’s
          certified software-defined encryption, you can create the next generation
          of secure enterprise infrastructure that DevSecOps requires. Together they
          create the stable, consistent, and supported base that your organization
          needs to develop applications faster, simplify updates, drive innovation,
          and reduce the risk of potential security breaches and data theft.\n###
          Core Features\n* Software-only for simple and scalable deployments \n* Negligible
          performance impact\n* Transparent to developers and administrators \n* Works
          with AWS, Google, and Azure \n* Unique encryption key per container volume
          \n* Encrypted volumes are only available when in use\n* Direct integration
          with OpenShift \n* Automated encryption policy management Secure erase of
          volumes\n* Ability to securely decommission a node\n\n## How does XCrypt
          Kubernetes Encryption fit into OpenShift? \nXCrypt XCrypt Kubernetes Encryption
          for OpenShift Encryption makes it incredibly easy to protect container data
          in your OpenShift environment. \n### Use cases include:\n* Transitioning
          from DevOps to DevSecOps \n* Ensuring data protection in OpenShift container
          environments with negligible performance impact \n* Achieving compliance
          in regulated industries like healthcare and finance\n\n## Other Information\nRole-based
          access control needs to be configured prior to the Operator installation.
          Please find details in the [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)
          \   "
        displayName: Zettaset XCrypt Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - encryption
        - storage
        links:
        - name: Zettaset
          url: https://www.zettaset.com/
        maintainers:
        - email: info@zettaset.com
          name: Zettaset
        maturity: alpha
        provider:
          name: Zettaset
        relatedImages:
        - registry.connect.redhat.com/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
        - registry.connect.redhat.com/zts/xcrypt-license-server-1@sha256:bc5486892a41e70f5dc4c4e1d82fca8db2149598fb8b5950ba5da1ee2d9d245d
        - registry.connect.redhat.com/zts/xcrypt-host-manager-1@sha256:9faa4b597eb74df412761dffa057b64fb69dbf9fa016414962eb2099a343e734
        - registry.connect.redhat.com/zts/zts-csi-driver-1@sha256:c15fa9a7cc40891103f426206663dcebe1210c8ee3ab6037cbf64e1f86b9c7e5
        - registry.connect.redhat.com/zts/xcrypt-ca-1@sha256:9f1015dfb459b69f5dbf449efa0517452095a6cf4f20826ebae278f9f800ffe7
        - registry.connect.redhat.com/zts/xcrypt-kmip-server-1@sha256:9af5d21dd748dfbf4e107f27a5f10fa3c0ba928ade9b989a744b620ce34ec757
        version: 0.0.17
      entries:
      - name: zts-xcrypt-operator.v0.0.17
        version: 0.0.17
      name: alpha
    defaultChannel: alpha
    packageName: xcrypt-operator
    provider:
      name: Zettaset
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-spark-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spark-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:a6c82a6d6ebaa0d2572d644ac9013451e020f134c6fe46222fef37b1e9aa3cbb
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:a6c82a6d6ebaa0d2572d644ac9013451e020f134c6fe46222fef37b1e9aa3cbb
        version: 24.11.1
      entries:
      - name: spark-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: spark-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:8ad9948bf8264924a0eedd92c2e58269117487b9d57816384447bb6c999be681
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:8ad9948bf8264924a0eedd92c2e58269117487b9d57816384447bb6c999be681
        version: 25.3.0
      entries:
      - name: spark-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: spark-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkConnectServerSpec via
              `CustomResource`
            displayName: sparkconnectservers.spark.stackable.tech
            kind: SparkConnectServer
            name: sparkconnectservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
        version: 25.7.0
      entries:
      - name: spark-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: spark-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkConnectServerSpec via
              `CustomResource`
            displayName: sparkconnectservers.spark.stackable.tech
            kind: SparkConnectServer
            name: sparkconnectservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
        version: 25.7.0
      entries:
      - name: spark-operator.v25.7.0
        version: 25.7.0
      - name: spark-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-spark-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://skupper.io
    name: skupper-netobs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-netobs-operator.v2.1.1-rh-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability.skupper.io/v2alpha1",
                "kind": "NetworkObserver",
                "metadata": {
                  "name": "networkobserver-sample"
                },
                "spec": {
                  "auth": {
                    "basic": {
                      "create": true,
                      "secretName": ""
                    },
                    "openshift": {
                      "cookieSecretName": "",
                      "createCookieSecret": true,
                      "serviceAccount": {
                        "create": true,
                        "nameOverride": ""
                      }
                    },
                    "strategy": "openshift"
                  },
                  "extraArgs": null,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "registry.redhat.io/service-interconnect/skupper-network-observer-rhel9",
                    "tag": "2.1.1"
                  },
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nginx": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/ubi9/nginx-124",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": 9.6
                  },
                  "openshiftOauthProxy": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-oauth-proxy-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.19.0"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "prometheus": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-prometheus-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.16.0"
                  },
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "router": {
                    "certificate": {
                      "create": true,
                      "nameOverride": ""
                    },
                    "endpoint": "amqps://skupper-router-local"
                  },
                  "securityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    }
                  },
                  "service": {
                    "port": 443,
                    "type": "ClusterIP"
                  },
                  "skipManagementLabels": false,
                  "tls": {
                    "openshiftIssued": true,
                    "secretName": "",
                    "skupperIssued": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
          createdAt: "2025-08-27T20:21:18Z"
          description: Red Hat Service Interconnect Network Observer Operator provides
            the ability to monitor and observe a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkObserver
            name: networkobservers.observability.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect Network Observer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/ubi9/nginx-124@sha256:cbcca4743015b4ef167383a080f4711a6049e275e0b66f80a74f8b087b0dfc7f
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:d3056b35d9a205b9f2c48d924f199c5ac23904eb18d526e4bff229e7c7181415
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
        version: 2.1.1-rh-8
      entries:
      - name: skupper-netobs-operator.v2.1.1-rh-8
        version: 2.1.1-rh-8
      name: stable-2
    - currentCSV: skupper-netobs-operator.v2.1.1-rh-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability.skupper.io/v2alpha1",
                "kind": "NetworkObserver",
                "metadata": {
                  "name": "networkobserver-sample"
                },
                "spec": {
                  "auth": {
                    "basic": {
                      "create": true,
                      "secretName": ""
                    },
                    "openshift": {
                      "cookieSecretName": "",
                      "createCookieSecret": true,
                      "serviceAccount": {
                        "create": true,
                        "nameOverride": ""
                      }
                    },
                    "strategy": "openshift"
                  },
                  "extraArgs": null,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "registry.redhat.io/service-interconnect/skupper-network-observer-rhel9",
                    "tag": "2.1.1"
                  },
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nginx": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/ubi9/nginx-124",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": 9.6
                  },
                  "openshiftOauthProxy": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-oauth-proxy-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.19.0"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "prometheus": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-prometheus-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.16.0"
                  },
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "router": {
                    "certificate": {
                      "create": true,
                      "nameOverride": ""
                    },
                    "endpoint": "amqps://skupper-router-local"
                  },
                  "securityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    }
                  },
                  "service": {
                    "port": 443,
                    "type": "ClusterIP"
                  },
                  "skipManagementLabels": false,
                  "tls": {
                    "openshiftIssued": true,
                    "secretName": "",
                    "skupperIssued": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
          createdAt: "2025-08-27T20:21:18Z"
          description: Red Hat Service Interconnect Network Observer Operator provides
            the ability to monitor and observe a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkObserver
            name: networkobservers.observability.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect Network Observer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:d3056b35d9a205b9f2c48d924f199c5ac23904eb18d526e4bff229e7c7181415
        - registry.redhat.io/ubi9/nginx-124@sha256:cbcca4743015b4ef167383a080f4711a6049e275e0b66f80a74f8b087b0dfc7f
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        version: 2.1.1-rh-8
      entries:
      - name: skupper-netobs-operator.v2.1.1-rh-8
        version: 2.1.1-rh-8
      name: stable-2.1
    defaultChannel: stable-2
    packageName: skupper-netobs-operator
    provider:
      name: Red Hat
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/trustification/trusted-profile-analyzer-operator
    name: rhtpa-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhtpa-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtpa.io/v1",
                "kind": "TrustedProfileAnalyzer",
                "metadata": {
                  "name": "trustedprofileanalyzer-sample"
                },
                "spec": {
                  "appDomain": "change-me",
                  "collector": {},
                  "database": {},
                  "image": {
                    "fullName": "registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e",
                    "pullPolicy": "IfNotPresent"
                  },
                  "infrastructure": {
                    "port": 9010
                  },
                  "ingress": {},
                  "metrics": {
                    "enabled": false
                  },
                  "modules": {
                    "createDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "createImporters": {
                      "enabled": false,
                      "image": {},
                      "importers": {}
                    },
                    "importer": {
                      "concurrency": 1,
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {},
                      "workingDirectory": {
                        "size": "32Gi"
                      }
                    },
                    "migrateDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "server": {
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "ingress": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {}
                    }
                  },
                  "oidc": {
                    "clients": {
                      "cli": {
                        "clientSecret": null
                      },
                      "frontend": {}
                    }
                  },
                  "openshift": {
                    "useServiceCa": true
                  },
                  "partOf": "trustify",
                  "replicas": 1,
                  "rust": {},
                  "storage": {},
                  "tls": {},
                  "tracing": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-09-19T15:04:16Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Profile Analyzer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trustification/trusted-profile-analyzer-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: TrustedProfileAnalyzer
            name: trustedprofileanalyzers.rhtpa.io
            version: v1
        description: |
          [Tech Preview]
          Red Hat Trusted Profile Analyzer (RHTPA) provides a single source of truth for your software inventory across all your environments. You can verify the integrity of your in-house and third-party applications against their respective Software Bill of Materials (SBOM) document.  Trusted Profile Analyzer provides vulnerability, component and license transparency, strengthens your software supply chain security, and enables an accurate, up-to-date understanding of your software inventory.

          Features and Benefits:

          Enhanced Software Supply Chain Security: Offers critical transparency and integrity verification across the entire software lifecycle.

          Reduced Operational Risk: Proactively identifies and helps remediate vulnerabilities minimizing risk and compliance violations.

          Improved Compliance and Auditability: Simplifies the process of meeting regulatory requirements for software transparency and integrity by providing verifiable documentation and continuous monitoring.

          Faster Remediation and Incident Response: Provides targeted, actionable insights to prioritize and accelerate the patching and remediation of vulnerabilities.

          Increased Developer Productivity & Trust: Integrates security into existing workflows without significant overhead, allowing developers to build and deploy with greater confidence in the integrity of their code and components.

          Centralized Visibility and Control: Consolidates disparate security information into a single pane of glass, offering a holistic view of your application security posture.

          Red Hat Trusted Profile Analyzer is a crucial tool for any organization running applications on OpenShift that needs to confidently understand, verify, and secure the software they deploy, especially in an era of complex software supply chains and increasing regulatory scrutiny.

          [Documentation](https://docs.redhat.com/en/documentation/red_hat_trusted_profile_analyzer/2.1).
        displayName: Red Hat Trusted Profile Analyzer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rhtpa
        - trusted-profile-analyzer
        links:
        - name: Trusted Profile Analyzer Operator
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        maintainers:
        - email: rhtpa-support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        relatedImages:
        - registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
        - registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e
        version: 1.0.2
      entries:
      - name: rhtpa-operator.v1.0.2
        version: 1.0.2
      - name: rhtpa-operator.v1.0.1
        version: 1.0.1
      - name: rhtpa-operator.v1.0.0
        version: 1.0.0
      name: stable
    - currentCSV: rhtpa-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtpa.io/v1",
                "kind": "TrustedProfileAnalyzer",
                "metadata": {
                  "name": "trustedprofileanalyzer-sample"
                },
                "spec": {
                  "appDomain": "change-me",
                  "collector": {},
                  "database": {},
                  "image": {
                    "fullName": "registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e",
                    "pullPolicy": "IfNotPresent"
                  },
                  "infrastructure": {
                    "port": 9010
                  },
                  "ingress": {},
                  "metrics": {
                    "enabled": false
                  },
                  "modules": {
                    "createDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "createImporters": {
                      "enabled": false,
                      "image": {},
                      "importers": {}
                    },
                    "importer": {
                      "concurrency": 1,
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {},
                      "workingDirectory": {
                        "size": "32Gi"
                      }
                    },
                    "migrateDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "server": {
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "ingress": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {}
                    }
                  },
                  "oidc": {
                    "clients": {
                      "cli": {
                        "clientSecret": null
                      },
                      "frontend": {}
                    }
                  },
                  "openshift": {
                    "useServiceCa": true
                  },
                  "partOf": "trustify",
                  "replicas": 1,
                  "rust": {},
                  "storage": {},
                  "tls": {},
                  "tracing": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-09-19T15:04:16Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Profile Analyzer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trustification/trusted-profile-analyzer-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: TrustedProfileAnalyzer
            name: trustedprofileanalyzers.rhtpa.io
            version: v1
        description: |
          [Tech Preview]
          Red Hat Trusted Profile Analyzer (RHTPA) provides a single source of truth for your software inventory across all your environments. You can verify the integrity of your in-house and third-party applications against their respective Software Bill of Materials (SBOM) document.  Trusted Profile Analyzer provides vulnerability, component and license transparency, strengthens your software supply chain security, and enables an accurate, up-to-date understanding of your software inventory.

          Features and Benefits:

          Enhanced Software Supply Chain Security: Offers critical transparency and integrity verification across the entire software lifecycle.

          Reduced Operational Risk: Proactively identifies and helps remediate vulnerabilities minimizing risk and compliance violations.

          Improved Compliance and Auditability: Simplifies the process of meeting regulatory requirements for software transparency and integrity by providing verifiable documentation and continuous monitoring.

          Faster Remediation and Incident Response: Provides targeted, actionable insights to prioritize and accelerate the patching and remediation of vulnerabilities.

          Increased Developer Productivity & Trust: Integrates security into existing workflows without significant overhead, allowing developers to build and deploy with greater confidence in the integrity of their code and components.

          Centralized Visibility and Control: Consolidates disparate security information into a single pane of glass, offering a holistic view of your application security posture.

          Red Hat Trusted Profile Analyzer is a crucial tool for any organization running applications on OpenShift that needs to confidently understand, verify, and secure the software they deploy, especially in an era of complex software supply chains and increasing regulatory scrutiny.

          [Documentation](https://docs.redhat.com/en/documentation/red_hat_trusted_profile_analyzer/2.1).
        displayName: Red Hat Trusted Profile Analyzer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rhtpa
        - trusted-profile-analyzer
        links:
        - name: Trusted Profile Analyzer Operator
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        maintainers:
        - email: rhtpa-support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        relatedImages:
        - registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
        - registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e
        version: 1.0.2
      entries:
      - name: rhtpa-operator.v1.0.2
        version: 1.0.2
      - name: rhtpa-operator.v1.0.1
        version: 1.0.1
      - name: rhtpa-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable
    packageName: rhtpa-operator
    provider:
      name: Red Hat
      url: https://github.com/trustification/trusted-profile-analyzer-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: nova-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nova-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "Nova",
                "metadata": {
                  "name": "nova"
                },
                "spec": {
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaAPI",
                "metadata": {
                  "name": "novaapi-sample"
                },
                "spec": {
                  "apiDatabaseHostname": "openstack",
                  "cell0DatabaseHostname": "openstack",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "registeredCells": {},
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaCell",
                "metadata": {
                  "name": "novacell1-with-upcall-sample"
                },
                "spec": {
                  "apiDatabaseHostname": "openstack",
                  "cellDatabaseHostname": "openstack",
                  "cellName": "cell1",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaCompute",
                "metadata": {
                  "name": "nova-compute-ironic"
                },
                "spec": {
                  "novaComputeTemplates": {
                    "compute-ironic": {
                      "computeDriver": "ironic.IronicDriver",
                      "customServiceConfig": "",
                      "networkAttachments": [
                        "internalapi"
                      ],
                      "replicas": 1,
                      "resources": {}
                    }
                  }
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaConductor",
                "metadata": {
                  "name": "novaconductor-sample"
                },
                "spec": {
                  "cellName": "cell1",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaMetadata",
                "metadata": {
                  "name": "novametadata-sample"
                },
                "spec": {
                  "apiDatabaseAccount": "nova-api",
                  "apiDatabaseHostname": "openstack",
                  "cellDatabaseAccount": "nova-cell0",
                  "cellDatabaseHostname": "openstack",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "registeredCells": {},
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaNoVNCProxy",
                "metadata": {
                  "name": "novanovncproxy-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaScheduler",
                "metadata": {
                  "name": "novascheduler-sample"
                },
                "spec": {
                  "apiDatabaseHostname": "openstack",
                  "cell0DatabaseHostname": "openstack",
                  "customServiceConfig": "[DEFAULT]\ndebug=true\n",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "registeredCells": {},
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:21Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nova
            name: nova.nova.openstack.org
            version: v1beta1
          - description: NovaAPI is the Schema for the novaapis API
            displayName: Nova API
            kind: NovaAPI
            name: novaapis.nova.openstack.org
            version: v1beta1
          - description: NovaCell is the Schema for the novacells API
            displayName: Nova Cell
            kind: NovaCell
            name: novacells.nova.openstack.org
            version: v1beta1
          - description: NovaCompute is the Schema for the NovaCompute
            displayName: Nova Compute
            kind: NovaCompute
            name: novacomputes.nova.openstack.org
            version: v1beta1
          - description: NovaConductor is the Schema for the novaconductors API
            displayName: Nova Conductor
            kind: NovaConductor
            name: novaconductors.nova.openstack.org
            version: v1beta1
          - description: NovaMetadata is the Schema for the novametadata API
            displayName: Nova Metadata
            kind: NovaMetadata
            name: novametadata.nova.openstack.org
            version: v1beta1
          - description: NovaNoVNCProxy is the Schema for the novanovncproxies API
            displayName: Nova No VNCProxy
            kind: NovaNoVNCProxy
            name: novanovncproxies.nova.openstack.org
            version: v1beta1
          - description: NovaScheduler is the Schema for the novaschedulers API
            displayName: Nova Scheduler
            kind: NovaScheduler
            name: novaschedulers.nova.openstack.org
            version: v1beta1
        description: Nova Operator
        displayName: Nova Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Nova
        links:
        - name: Nova Operator
          url: https://github.com/openstack-k8s-operators/nova-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-nova-scheduler-rhel9@sha256:ba91196cd4f9b6e6f49868c91343b8eca7cc0d5016c36e83e2cc76f45cbe777b
        - registry.redhat.io/rhoso/openstack-nova-compute-rhel9@sha256:43cbe46ab263403fc303b6bc2388863029aad33465b237db18da7cd6d42078e1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/nova-rhel9-operator@sha256:f058c56df02a108505d51b9d19973894c2ee59055776c9f3fd7f7afe23a5106a
        - registry.redhat.io/rhoso/openstack-nova-api-rhel9@sha256:bbb8ebceb6f30b6c8beb120429e69f323bd231c09aaaddc5be9f15a527e928ed
        - registry.redhat.io/rhoso/openstack-nova-conductor-rhel9@sha256:500c36fd519611be8662e97ca31b1e59269c14d8aa94286dad9925e957462155
        - registry.redhat.io/rhoso/openstack-nova-novncproxy-rhel9@sha256:bf534f0c72f188d5f51e40b0392ace182db2055dff9ad14b01eeef13056febbf
        version: 1.0.6
      entries:
      - name: nova-operator.v1.0.6
        version: 1.0.6
      - name: nova-operator.v1.0.5
        version: 1.0.5
      - name: nova-operator.v1.0.4
        version: 1.0.4
      - name: nova-operator.v1.0.3
        version: 1.0.3
      - name: nova-operator.v1.0.2
        version: 1.0.2
      - name: nova-operator.v1.0.1
        version: 1.0.1
      - name: nova-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: nova-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Cloudera
      provider-url: https://cloudera.com
    name: cfm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cfm-operator.v2.9.0-b96
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "Nifi",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifi-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifi",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifi-sample"
                },
                "spec": {
                  "hostName": "nifi.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifi-k8s",
                    "tag": "2.8.0-b94-nifi_1.25.0.2.3.13.0-36"
                  },
                  "replicas": 1,
                  "stateManagement": {
                    "clusterProvider": {
                      "class": "org.apache.nifi.kubernetes.state.provider.KubernetesConfigMapStateProvider",
                      "id": "kubernetes-provider"
                    }
                  },
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.8.0-b94"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              },
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "NifiRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifiregistry-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifiregistry",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifiregistry-sample"
                },
                "spec": {
                  "hostName": "registry.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s",
                    "tag": "2.8.0-b94-nifi_1.25.0.2.3.13.0-36"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.8.0-b94"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning,Big Data
          com.redhat.openshift.versions: v4.10
          containerImage: container.repository.cloudera.com/cloudera/cfm-operator@sha256:890d81188143ccf9ebe2c2fe5e039cdff3a2df53f7512555673f0468147e991e
          createdAt: "2024-12-06T19:03:24Z"
          description: Cloudera Flow Management - Kubernetes Operator for Apache NiFideploys
            and operates Apache NiFi
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: A valid Cloudera license is required.
            Contact a Cloudera account team to learn more.
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NifiRegistry is the Schema for the nifiregistries API.
            displayName: Nifi Registry
            kind: NifiRegistry
            name: nifiregistries.cfm.cloudera.com
            version: v1alpha1
          - description: Nifi is the Schema for the nifis API.
            displayName: Nifi
            kind: Nifi
            name: nifis.cfm.cloudera.com
            version: v1alpha1
        description: |-
          Cloudera Flow Management - Kubernetes Operator for Apache NiFi brings enterprise-grade NiFi and NiFi Registry deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of NiFi and NiFi Registry for enterprise deployments.
          To learn more about Cloudera Flow Management - Kubernetes Operator for Apache NiFi read this
          [deep dive blog post](https://community.cloudera.com/t5/Community-Articles/Cloudera-Flow-Management-Operator-A-technical-deep-dive/ta-p/390473).

          Cloudera Flow Management - Kubernetes Operator for Apache NiFi consists of and ships multiple components including Apache NiFi, Apache NiFi Registry,
          and a custom sidecar image for logging aggregation.

          ## Installation prerequisites

          Installing Cloudera Flow Management - Kubernetes Operator for Apache NiFi requires the a valid license as well as Cloudera credentials
          (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly.
          The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) from which installation artifacts are pulled.

          ### Obtaining a license
          Contact [Cloudera sales](https://cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets
          Before installation, create two `Secrets`. One containing your Cloudera credentials and one containing your license. Cloudera credentials are provided to you
          together with your license.

          1. Create a `Secret` containing your Cloudera credentials.
              `oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***]
               --docker-server=container.repository.cloudera.com`

          2. Create a `Secret` containing your license.
              `oc create secret generic cfm-operator-license --from-file=license.txt=[***PATH TO LICENSE FILE***]`

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces in which you
            plan to deploy NiFi clusters or NiFi Registry deployments. Cloudera recommends that you create the `Secret` in all required namespaces before installation
            if you know which namespaces you will be using to deploy NiFi clusters or NiFi Registry deployments.

          ## Learn more
          To learn more, visit the [Cloudera Flow Management - Kubernetes Operator for Apache NiFi documentation library](https://docs.cloudera.com/cfm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Flow Management - Kubernetes Operator for Apache NiFi, visit the following resources:
          * [Apache NiFi](https://nifi.apache.org/)
          * [Apache NiFi Registry](https://nifi.apache.org/registry.html)
        displayName: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        - nifiregistry
        - cloudera
        - cfm
        links:
        - name: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
          url: https://docs.cloudera.com/cfm-operator
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s@sha256:914dcde9ee97894fa6f5deec285e53bf81d196e071a997cabb40fedd52deccc3
        - container.repository.cloudera.com/cloudera/cfm-tini@sha256:6493d23e20fca8daf31823403fca3dd9b42da983464fbe788a6863e141e8b66b
        - container.repository.cloudera.com/cloudera/cfm-nifi-k8s@sha256:0d700b7eff096dfc362868345826c3a6460fa43a700273b1422e95015b7727ee
        - container.repository.cloudera.com/cloudera/cfm-operator@sha256:890d81188143ccf9ebe2c2fe5e039cdff3a2df53f7512555673f0468147e991e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86af06fb3b6c17b22be47118b0a8c797fa7fdd3c00dbe99a2d2a593ca8fc3d63
        version: 2.9.0-b96
      entries:
      - name: cfm-operator.v2.9.0-b96
        version: 2.9.0-b96
      name: alpha
    - currentCSV: cfm-operator.v2.10.0-b135
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "Nifi",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifi-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifi",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifi-sample"
                },
                "spec": {
                  "hostName": "nifi.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifi-k8s",
                    "tag": "2.9.0-b96-nifi_1.27.0.2.3.14.0-14"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.9.0-b96"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              },
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "NifiRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifiregistry-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifiregistry",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifiregistry-sample"
                },
                "spec": {
                  "hostName": "registry.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s",
                    "tag": "2.9.0-b96-nifi_1.27.0.2.3.14.0-14"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.9.0-b96"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning,Big Data
          com.redhat.openshift.versions: v4.10
          containerImage: container.repository.cloudera.com/cloudera/cfm-operator@sha256:0312b6dadd5904ea9507bb857ab89d7266bf244c93b86855e0528e72813b7a85
          createdAt: "2025-05-05T14:39:51Z"
          description: Cloudera Flow Management - Kubernetes Operator for Apache NiFideploys
            and operates Apache NiFi
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: A valid Cloudera license is required.
            Contact a Cloudera account team to learn more.
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NifiRegistry is the Schema for the nifiregistries API.
            displayName: Nifi Registry
            kind: NifiRegistry
            name: nifiregistries.cfm.cloudera.com
            version: v1alpha1
          - description: Nifi is the Schema for the nifis API.
            displayName: Nifi
            kind: Nifi
            name: nifis.cfm.cloudera.com
            version: v1alpha1
        description: |-
          Cloudera Flow Management - Kubernetes Operator for Apache NiFi brings enterprise-grade NiFi and NiFi Registry deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of NiFi and NiFi Registry for enterprise deployments.
          To learn more about Cloudera Flow Management - Kubernetes Operator for Apache NiFi read this
          [deep dive blog post](https://community.cloudera.com/t5/Community-Articles/Cloudera-Flow-Management-Operator-A-technical-deep-dive/ta-p/390473).

          Cloudera Flow Management - Kubernetes Operator for Apache NiFi consists of and ships multiple components including Apache NiFi, Apache NiFi Registry,
          and a custom sidecar image for logging aggregation.

          ## Installation prerequisites

          Installing Cloudera Flow Management - Kubernetes Operator for Apache NiFi requires the a valid license as well as Cloudera credentials
          (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly.
          The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) from which installation artifacts are pulled.

          ### Obtaining a license
          Contact [Cloudera sales](https://cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets
          Before installation, create two `Secrets`. One containing your Cloudera credentials and one containing your license. Cloudera credentials are provided to you
          together with your license.

          1. Create a `Secret` containing your Cloudera credentials.
              `oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***]
               --docker-server=container.repository.cloudera.com --namespace [***OPERATOR INSTALL NAMESPACE***]`

          2. Create a `Secret` containing your license.
              `oc create secret generic cfm-operator-license --from-file=license.txt=[***PATH TO LICENSE FILE***] --namespace [***OPERATOR INSTALL NAMESPACE***]`

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces in which you
            plan to deploy NiFi clusters or NiFi Registry deployments. Cloudera recommends that you create the `Secret` in all required namespaces before installation
            if you know which namespaces you will be using to deploy NiFi clusters or NiFi Registry deployments.

          ## Learn more
          To learn more, visit the [Cloudera Flow Management - Kubernetes Operator for Apache NiFi documentation library](https://docs.cloudera.com/cfm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Flow Management - Kubernetes Operator for Apache NiFi, visit the following resources:
          * [Apache NiFi](https://nifi.apache.org/)
          * [Apache NiFi Registry](https://nifi.apache.org/registry.html)
        displayName: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        - nifiregistry
        - cloudera
        - cfm
        links:
        - name: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
          url: https://docs.cloudera.com/cfm-operator
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/cfm-nifi-k8s@sha256:8607c81906422d65784c6dca74a37af3b7c87012bca65ce28a995366014aec86
        - container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s@sha256:52e00f102579b972500eb187e64a7662a6a46caa883df743e454b1d5111b34e5
        - container.repository.cloudera.com/cloudera/cfm-operator@sha256:0312b6dadd5904ea9507bb857ab89d7266bf244c93b86855e0528e72813b7a85
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86af06fb3b6c17b22be47118b0a8c797fa7fdd3c00dbe99a2d2a593ca8fc3d63
        - container.repository.cloudera.com/cloudera/cfm-tini@sha256:d42ad5f05a9bb283fed6dbbd820855714a31662f5b851e691f13703153e90d13
        version: 2.10.0-b135
      entries:
      - name: cfm-operator.v2.10.0-b135
        version: 2.10.0-b135
      name: stable
    defaultChannel: stable
    packageName: cfm-operator
    provider:
      name: Cloudera
      url: https://cloudera.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Splunk Inc.
      provider-url: www.splunk.com
    name: splunk-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: splunk-operator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "IndexerCluster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "indexercluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "LicenseMaster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "licensemaster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "MonitoringConsole",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "monitoringconsole-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "SearchHeadCluster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "searchheadcluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "Standalone",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "standalone-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "clustermanager-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "IndexerCluster",
                "metadata": {
                  "name": "indexercluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "LicenseManager",
                "metadata": {
                  "name": "licensemanager-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "MonitoringConsole",
                "metadata": {
                  "name": "monitoringconsole-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "SearchHeadCluster",
                "metadata": {
                  "name": "searchheadcluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "Standalone",
                "metadata": {
                  "name": "standalone-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Big Data, Logging & Tracing, Monitoring, Security, AI/Machine
            Learning
          containerImage: splunk/splunk-operator@sha256:ed02561995d46ff8eea143fa78f5da118ee6fdb500832e9a33b42d059a67e229
          createdAt: "2025-09-08T10:17:59Z"
          description: The Splunk Operator for Kubernetes enables you to quickly and
            easily deploy Splunk Enterprise on your choice of private or public cloud
            provider. The Operator simplifies scaling and management of Splunk Enterprise
            by automating administrative workflows using Kubernetes best practices.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=2.4.0 <3.0.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/splunk/splunk-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterManager is the Schema for the cluster manager API
            displayName: Cluster Manager
            kind: ClusterManager
            name: clustermanagers.enterprise.splunk.com
            version: v4
          - description: ClusterMaster is the Schema for the cluster manager API
            displayName: Cluster Master
            kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v3
          - kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v1
          - kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v2
          - description: IndexerCluster is the Schema for a Splunk Enterprise indexer
              cluster
            displayName: Indexer Cluster
            kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v3
          - description: IndexerCluster is the Schema for a Splunk Enterprise indexer
              cluster
            displayName: Indexer Cluster
            kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v4
          - kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v1
          - kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v2
          - description: LicenseManager is the Schema for a Splunk Enterprise license
              manager.
            displayName: License Manager
            kind: LicenseManager
            name: licensemanagers.enterprise.splunk.com
            version: v4
          - description: LicenseMaster is the Schema for a Splunk Enterprise license
              master.
            displayName: License Master
            kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v3
          - kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v1
          - kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v2
          - description: MonitoringConsole is the Schema for the monitoringconsole
              API
            displayName: Monitoring Console
            kind: MonitoringConsole
            name: monitoringconsoles.enterprise.splunk.com
            version: v3
          - description: MonitoringConsole is the Schema for the monitoringconsole
              API
            displayName: Monitoring Console
            kind: MonitoringConsole
            name: monitoringconsoles.enterprise.splunk.com
            version: v4
          - description: SearchHeadCluster is the Schema for a Splunk Enterprise search
              head cluster
            displayName: Search Head Cluster
            kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v3
          - description: SearchHeadCluster is the Schema for a Splunk Enterprise search
              head cluster
            displayName: Search Head Cluster
            kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v4
          - kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v1
          - kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v2
          - description: Standalone is the Schema for a Splunk Enterprise standalone
              instances.
            displayName: Standalone
            kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v3
          - description: Standalone is the Schema for a Splunk Enterprise standalone
              instances.
            displayName: Standalone
            kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v4
          - kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v1
          - kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v2
        description: The Splunk Operator for Kubernetes enables you to quickly and
          easily deploy Splunk Enterprise on your choice of private or public cloud
          provider. The Operator simplifies scaling and management of Splunk Enterprise
          by automating administrative workflows using Kubernetes best practices.
        displayName: Splunk Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Splunk
        - Enterprise
        - Operator
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://github.com/splunk/splunk-operator/tree/main/docs
        maintainers:
        - email: support@splunk.com
          name: Splunk Operator Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Splunk Inc.
          url: www.splunk.com
        relatedImages:
        - splunk/splunk-operator@sha256:ed02561995d46ff8eea143fa78f5da118ee6fdb500832e9a33b42d059a67e229
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - splunk/splunk@sha256:4a5ce5a7b4e84a87396f0664895095c6510a0f2bf52e8747a373c7357f3ee313
        version: 3.0.0
      entries:
      - name: splunk-operator.v3.0.0
        version: 3.0.0
      - name: splunk-operator.v2.8.1
        version: 2.8.1
      - name: splunk-operator.v2.8.0
        version: 2.8.0
      - name: splunk-operator.v2.7.1
        version: 2.7.1
      - name: splunk-operator.v2.7.0
        version: 2.7.0
      name: stable
    defaultChannel: stable
    packageName: splunk-operator
    provider:
      name: Splunk Inc.
      url: www.splunk.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-emrcontainers-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-emrcontainers-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "emrcontainers.services.k8s.aws/v1alpha1",
                "kind": "VirtualCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/emrcontainers-controller:1.1.1
          createdAt: "2025-09-19T19:36:26Z"
          description: AWS EMR on EKS controller is a service controller for managing
            EMR on EKS resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: JobRun represents the state of an AWS emrcontainers JobRun
              resource.
            displayName: JobRun
            kind: JobRun
            name: jobruns.emrcontainers.services.k8s.aws
            version: v1alpha1
          - description: VirtualCluster represents the state of an AWS emrcontainers
              VirtualCluster resource.
            displayName: VirtualCluster
            kind: VirtualCluster
            name: virtualclusters.emrcontainers.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EMR on EKS resources in AWS from within your Kubernetes cluster.

          **About Amazon EMR on EKS**

          Amazon EMR on EKS provides a deployment option for Amazon EMR that allows you to run open-source big data frameworks on Amazon Elastic Kubernetes Service (Amazon EKS). With this deployment option, you can focus on running analytics workloads while Amazon EMR on EKS builds, configures, and manages containers for open-source applications.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EMR on EKS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - emrcontainers
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EMR on EKS Developer Resources
          url: https://docs.aws.amazon.com/emr/latest/EMR-on-EKS-DevelopmentGuide/emr-eks.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: emr on eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/emrcontainers-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-emrcontainers-controller.v1.1.1
        version: 1.1.1
      - name: ack-emrcontainers-controller.v1.1.0
        version: 1.1.0
      - name: ack-emrcontainers-controller.v1.0.27
        version: 1.0.27
      - name: ack-emrcontainers-controller.v1.0.25
        version: 1.0.25
      - name: ack-emrcontainers-controller.v1.0.24
        version: 1.0.24
      - name: ack-emrcontainers-controller.v1.0.23
        version: 1.0.23
      - name: ack-emrcontainers-controller.v1.0.22
        version: 1.0.22
      - name: ack-emrcontainers-controller.v1.0.21
        version: 1.0.21
      - name: ack-emrcontainers-controller.v1.0.18
        version: 1.0.18
      - name: ack-emrcontainers-controller.v1.0.16
        version: 1.0.16
      - name: ack-emrcontainers-controller.v1.0.15
        version: 1.0.15
      - name: ack-emrcontainers-controller.v1.0.14
        version: 1.0.14
      - name: ack-emrcontainers-controller.v1.0.13
        version: 1.0.13
      - name: ack-emrcontainers-controller.v1.0.12
        version: 1.0.12
      - name: ack-emrcontainers-controller.v1.0.11
        version: 1.0.11
      - name: ack-emrcontainers-controller.v1.0.9
        version: 1.0.9
      - name: ack-emrcontainers-controller.v1.0.8
        version: 1.0.8
      - name: ack-emrcontainers-controller.v1.0.4
        version: 1.0.4
      - name: ack-emrcontainers-controller.v1.0.3
        version: 1.0.3
      - name: ack-emrcontainers-controller.v1.0.1
        version: 1.0.1
      - name: ack-emrcontainers-controller.v1.0.0
        version: 1.0.0
      - name: ack-emrcontainers-controller.v0.1.2
        version: 0.1.2
      - name: ack-emrcontainers-controller.v0.1.0
        version: 0.1.0
      - name: ack-emrcontainers-controller.v0.0.10
        version: 0.0.10
      - name: ack-emrcontainers-controller.v0.0.9
        version: 0.0.9
      - name: ack-emrcontainers-controller.v0.0.8
        version: 0.0.8
      - name: ack-emrcontainers-controller.v0.0.7
        version: 0.0.7
      - name: ack-emrcontainers-controller.v0.0.6
        version: 0.0.6
      - name: ack-emrcontainers-controller.v0.0.5
        version: 0.0.5
      - name: ack-emrcontainers-controller.v0.0.4
        version: 0.0.4
      - name: ack-emrcontainers-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-emrcontainers-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: volume-expander-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: volume-expander-operator.v0.3.6
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage,Monitoring
          certified: "false"
          containerImage: quay.io/redhat-cop/volume-expander-operator@sha256:e4059a66a9af7c786c8c651abb8a57a55a0e528342e3a59f8c264b40f87d6e25
          createdAt: "2022-05-05T13:25:53Z"
          description: This operator provides automation expanding pod's volumes when
            they are about to run out of space.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: volume-expander-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/volume-expander-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The purpose of the volume-expander-operator is to expand volumes when they are running out of space.
          This is achieved by using the [volume expansion feature](https://kubernetes.io/docs/concepts/storage/persistent-volumes/#expanding-persistent-volumes-claims).

          The operator periodically checks the `kubelet_volume_stats_used_bytes` and `kubelet_volume_stats_capacity_bytes published` by the kubelets to decide when to expand a volume.
          Notice that these metrics are generated only when a volume is mounted to a pod. Also the kubelet takes a minute or two to start generating accurate values for these metrics. The operator accounts for that.

          This operator works based on the following annotations to PersistentVolumeClaim resources:

          | Annotation | Default &nbsp;  &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;  | Description  |
          | :--- | :--- | :--- |
          | `volume-expander-operator.redhat-cop.io/autoexpand`  | N/A  | if set to "true" enables the volume-expander-operator to watch on this PVC  |
          | `volume-expander-operator.redhat-cop.io/polling-frequency`  | `"30s"` | How frequently to poll the volume metrics. Express this value as a valid golang [Duration](https://golang.org/pkg/time/#ParseDuration)  |
          | `volume-expander-operator.redhat-cop.io/expand-threshold-percent` | `"80"` | the percentage of used storage after which the volume will be expanded. This must be a positive integer. |
          | `volume-expander-operator.redhat-cop.io/expand-by-percent` | `"25"` | the percentage by which the volume will be expanded, relative to the current size. This must be an integer between 0 and 100 |
          | `volume-expander-operator.redhat-cop.io/expand-up-to` | MaxInt64 | the upper bound for this volume to be expanded to. The default value is the largest quantity representable and is intended to be interpreted as infinite. If the default is used it is recommend to ensure the namespace has a quota on the used storage class. |

          Note that not all of the storage driver implementations support volume expansion. It is a responsibility of the user/platform administrator to ensure that storage class and the persistent volume claim meet all the requirements needed for the volume expansion feature to work properly.

          This operator was tested with [OCS](https://www.redhat.com/en/technologies/cloud-computing/openshift-container-storage), but should work with any other storage driver that supports volume expansion.
        displayName: Volume Expander Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - volume
        - storage
        - csi
        - expansion
        links:
        - name: Volume Expander Operator
          url: https://github.com/redhat-cop/volume-expander-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/volume-expander-operator@sha256:e4059a66a9af7c786c8c651abb8a57a55a0e528342e3a59f8c264b40f87d6e25
        version: 0.3.6
      entries:
      - name: volume-expander-operator.v0.3.6
        version: 0.3.6
      - name: volume-expander-operator.v0.3.5
        version: 0.3.5
      - name: volume-expander-operator.v0.3.4
        version: 0.3.4
      - name: volume-expander-operator.v0.3.3
        version: 0.3.3
      - name: volume-expander-operator.v0.3.2
        version: 0.3.2
      - name: volume-expander-operator.v0.3.1
        version: 0.3.1
      - name: volume-expander-operator.v0.3.0
        version: 0.3.0
      - name: volume-expander-operator.v0.2.0
        version: 0.2.0
      - name: volume-expander-operator.v0.1.8
        version: 0.1.8
      - name: volume-expander-operator.v0.1.7
        version: 0.1.7
      - name: volume-expander-operator.v0.1.6
        version: 0.1.6
      - name: volume-expander-operator.v0.1.5
        version: 0.1.5
      - name: volume-expander-operator.v0.1.4
        version: 0.1.4
      - name: volume-expander-operator.v0.1.3
        version: 0.1.3
      - name: volume-expander-operator.v0.1.2
        version: 0.1.2
      - name: volume-expander-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: volume-expander-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-eventbridge-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-eventbridge-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Archive",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Endpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "EventBus",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/eventbridge-controller:1.1.1
          createdAt: "2025-09-19T20:51:55Z"
          description: AWS EventBridge controller is a service controller for managing
            EventBridge resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Archive represents the state of an AWS eventbridge Archive
              resource.
            displayName: Archive
            kind: Archive
            name: archives.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: Endpoint represents the state of an AWS eventbridge Endpoint
              resource.
            displayName: Endpoint
            kind: Endpoint
            name: endpoints.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: EventBus represents the state of an AWS eventbridge EventBus
              resource.
            displayName: EventBus
            kind: EventBus
            name: eventbuses.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: Rule represents the state of an AWS eventbridge Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.eventbridge.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EventBridge resources in AWS from within your Kubernetes cluster.

          **About Amazon EventBridge**

          EventBridge is a serverless service that uses events to connect application components together, making it easier for you to build scalable event-driven applications. Use it to route events from sources such as home-grown applications, AWS services, and third- party software to consumer applications across your organization. EventBridge provides a simple and consistent way to ingest, filter, transform, and deliver events so you can build new applications quickly.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EventBridge
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eventbridge
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EventBridge Developer Resources
          url: https://aws.amazon.com/eventbridge/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eventbridge maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/eventbridge-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-eventbridge-controller.v1.1.1
        version: 1.1.1
      - name: ack-eventbridge-controller.v1.1.0
        version: 1.1.0
      - name: ack-eventbridge-controller.v1.0.28
        version: 1.0.28
      - name: ack-eventbridge-controller.v1.0.27
        version: 1.0.27
      - name: ack-eventbridge-controller.v1.0.26
        version: 1.0.26
      - name: ack-eventbridge-controller.v1.0.25
        version: 1.0.25
      - name: ack-eventbridge-controller.v1.0.24
        version: 1.0.24
      - name: ack-eventbridge-controller.v1.0.23
        version: 1.0.23
      - name: ack-eventbridge-controller.v1.0.22
        version: 1.0.22
      - name: ack-eventbridge-controller.v1.0.21
        version: 1.0.21
      - name: ack-eventbridge-controller.v1.0.20
        version: 1.0.20
      - name: ack-eventbridge-controller.v1.0.18
        version: 1.0.18
      - name: ack-eventbridge-controller.v1.0.17
        version: 1.0.17
      - name: ack-eventbridge-controller.v1.0.16
        version: 1.0.16
      - name: ack-eventbridge-controller.v1.0.15
        version: 1.0.15
      - name: ack-eventbridge-controller.v1.0.14
        version: 1.0.14
      - name: ack-eventbridge-controller.v1.0.13
        version: 1.0.13
      - name: ack-eventbridge-controller.v1.0.12
        version: 1.0.12
      - name: ack-eventbridge-controller.v1.0.11
        version: 1.0.11
      - name: ack-eventbridge-controller.v1.0.6
        version: 1.0.6
      - name: ack-eventbridge-controller.v1.0.5
        version: 1.0.5
      - name: ack-eventbridge-controller.v1.0.4
        version: 1.0.4
      - name: ack-eventbridge-controller.v1.0.3
        version: 1.0.3
      - name: ack-eventbridge-controller.v1.0.1
        version: 1.0.1
      - name: ack-eventbridge-controller.v1.0.0
        version: 1.0.0
      - name: ack-eventbridge-controller.v0.0.7
        version: 0.0.7
      - name: ack-eventbridge-controller.v0.0.6
        version: 0.0.6
      - name: ack-eventbridge-controller.v0.0.5
        version: 0.0.5
      - name: ack-eventbridge-controller.v0.0.4
        version: 0.0.4
      - name: ack-eventbridge-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-eventbridge-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Oracle Corporation
      provider-url: https://github.com/oracle/coherence-operator
    name: coherence-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: coherence-operator.v3.5.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "coherence.oracle.com/v1",
                "kind": "Coherence",
                "metadata": {
                  "name": "coherence-sample"
                },
                "spec": {
                  "replicas": 3
                }
              },
              {
                "apiVersion": "coherence.oracle.com/v1",
                "kind": "CoherenceJob",
                "metadata": {
                  "name": "coherence-job-sample"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "false"
          containerImage: container-registry.oracle.com/middleware/coherence-operator:3.5.4
          createdAt: "2025-08-28T16:15:09Z"
          description: The Oracle Coherence Kubernetes Operator enables easy management
            of Coherence clusters in a Kubernetes environment.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/arch.amd64: supported
          operatorframework.io/arch.arm64: supported
          operatorframework.io/os.linux: supported
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/oracle/coherence-operator
          support: Oracle Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Coherence is the top level schema for the Coherence API and
              custom resource definition (CRD).
            displayName: Coherence
            kind: Coherence
            name: coherence.coherence.oracle.com
            version: v1
          - description: |-
              CoherenceJob is the top level schema for the CoherenceJob API and custom resource definition (CRD)
              for configuring Coherence Job workloads.
            displayName: Coherence Job
            kind: CoherenceJob
            name: coherencejob.coherence.oracle.com
            version: v1
        description: |
          The Oracle Coherence Kubernetes Operator enables easy management of Coherence clusters in a Kubernetes environment.

          ### Installation Guide & Configuration

          For the complete installation guide and all configuration options please refer to
          [Coherence Operator Documentation](https://oracle.github.io/coherence-operator/docs/latest/#/about/01_overview).
        displayName: Coherence Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Coherence
        - Oracle
        - In Memory Data
        - Caching
        links:
        - name: Coherence Kubernetes Operator
          url: https://oracle.github.io/coherence-operator/docs/latest
        - name: Oracle Coherence
          url: https://coherence.community
        maintainers:
        - email: jonathan.knight@oracle.com
          name: Jonathan Knight
        maturity: mature
        minKubeVersion: 1.26.0
        provider:
          name: Oracle Corporation
          url: https://github.com/oracle/coherence-operator
        relatedImages:
        - container-registry.oracle.com/middleware/coherence-operator:3.5.4
        version: 3.5.4
      entries:
      - name: coherence-operator.v3.5.4
        version: 3.5.4
      - name: coherence-operator.v3.5.2
        version: 3.5.2
      - name: coherence-operator.v3.5.0
        version: 3.5.0
      - name: coherence-operator.v3.4.3
        version: 3.4.3
      name: stable
    defaultChannel: stable
    packageName: coherence-operator
    provider:
      name: Oracle Corporation
      url: https://github.com/oracle/coherence-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nxrm-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nxrm-operator-certified.v3.58.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusRepo",
                "metadata": {
                  "name": "example-nexusrepo"
                },
                "spec": {
                  "config": {
                    "data": null,
                    "enabled": false,
                    "mountPath": "/sonatype-nexus-conf"
                  },
                  "deployment": {
                    "additionalContainers": null,
                    "additionalVolumeMounts": null,
                    "additionalVolumes": null,
                    "annotations": {},
                    "initContainers": null,
                    "postStart": {
                      "command": null
                    },
                    "preStart": {
                      "command": null
                    },
                    "terminationGracePeriodSeconds": 120
                  },
                  "deploymentStrategy": {
                    "type": "Recreate"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "path": "/",
                    "rules": null,
                    "tls": {
                      "enabled": true,
                      "secretName": "nexus-tls"
                    }
                  },
                  "nexus": {
                    "dockerPort": 5003,
                    "env": [
                      {
                        "name": "INSTALL4J_ADD_VM_PARAMS",
                        "value": "-Xms2703M -Xmx2703M -XX:MaxDirectMemorySize=2703M -XX:+UnlockExperimentalVMOptions -XX:+UseCGroupMemoryLimitForHeap"
                      }
                    ],
                    "hostAliases": [],
                    "imageName": "registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:06fe31f40ff3fd7556fc7b8f6b2efe80ada1d9079ec9aaad60d1fdc0ee0677f0",
                    "imagePullPolicy": "IfNotPresent",
                    "imagePullSecret": "",
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "path": "/",
                      "periodSeconds": 30
                    },
                    "nexusPort": 8081,
                    "podAnnotations": {},
                    "readinessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "path": "/",
                      "periodSeconds": 30
                    },
                    "resources": {},
                    "securityContext": {
                    },
                    "service": {
                      "type": "NodePort"
                    }
                  },
                  "nexusProxyRoute": {
                    "annotations": null,
                    "enabled": false,
                    "labels": null
                  },
                  "persistence": {
                    "accessMode": "ReadWriteOnce",
                    "enabled": true,
                    "storageSize": "8Gi"
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": null,
                    "enabled": false,
                    "labels": null,
                    "name": "docker",
                    "portName": "docker"
                  },
                  "secret": {
                    "data": null,
                    "enabled": false,
                    "mountPath": "/etc/secret-volume",
                    "readOnly": true
                  },
                  "service": {
                    "annotations": {},
                    "enabled": false,
                    "labels": {},
                    "ports": [
                      {
                        "name": "nexus-service",
                        "port": 80,
                        "targetPort": 80
                      }
                    ]
                  },
                  "statefulset": {
                    "enabled": false
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nxrm-operator-certified@sha256:4945ad993a10e27b358b2d9869d05f23ca761162ec088e5a833939022bc23192
          createdAt: "2023-07-14"
          description: |-
            Nexus Repository is the central source of control to efficiently manage all binaries
            and build artifacts across your DevOps pipeline.
          olm.skipRange: <3.58.1-2
          repository: https://github.com/sonatype/operator-nxrm3
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus Repository
            displayName: NexusRepo
            kind: NexusRepo
            name: nexusrepos.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Repository is the central source of control to efficiently manage all binaries
          and build artifacts across your DevOps pipeline.
          The flow of open source and third-party components into and through an organization
          creates a complex software supply chain.
          Nexus Repository delivers speed, efficiency, and quality to the governance
          and management of all dependencies, libraries, and applications for your DevOps teams.

          ## Core Capabilities

          * **Dependency Management**:
            Improves reliability with repeatable, fast access to secure dependencies
          * **Developer Productivity**:
            Streamline developer workflows by enabling the sharing of components and applications across teams
          * **Supply Chain Performance**:
            Improve speed-to-market and reduced build times with release advanced staging and component tagging
          * **CI/CD Integrations**:
            Increase DevOps scalability with integrations to the most popular build and deployment tools

          Version control systems and package registries do not scale when managing proprietary,
          open source, and third-party components.
          Organizations need a central binary and build artifact repository to manage dependencies
          across the entire software supply chain.

          ## Limitations

          High Availability Clustering (HA-C) is not supported for Nexus Repository Pro for OpenShift.

          This operator will be released on a quarterly basis.

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus Repository Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage

          Once the server instance is created by the operator and running,
          you'll want to expose the service as you see fit:
          1. Create a Route to that service for nexus.port (8081).

          By default, the Nexus Repository starts up in OSS mode until a license is installed.

          The Nexus Repository can be further configured via the NexusRepo custom resource definition:

          | Parameter                                   | Description                         | Default                                 |
          | ------------------------------------------  | ----------------------------------  | ----------------------------------------|
          | `statefulset.enabled`                       | Use statefulset instead of deployment | `false` |
          | `deploymentStrategy.type`                   | Deployment Strategy     |  `Recreate` |
          | `nexus.env`                                 | Nexus environment variables         | `See example.` |
          | `nexus.resources`                           | Nexus resource requests and limits  | `{}`                                    |
          | `nexus.dockerPort`                          | Port to access docker               | `5003`                                  |
          | `nexus.nexusPort`                           | Internal port for Nexus service     | `8081`                                  |
          | `nexus.service.type`                        | Service for Nexus                   |`NodePort`                                |
          | `nexus.service.clusterIp`                   | Specific cluster IP when service type is cluster IP. Use None for headless service |`nil`   |
          | `nexus.securityContext`                     | Security Context |
          | `nexus.labels`                              | Service labels                      | `{}`                                    |
          | `nexus.podAnnotations`                      | Pod Annotations                     | `{}`
          | `nexus.livenessProbe.initialDelaySeconds`   | LivenessProbe initial delay         | 30                                      |
          | `nexus.livenessProbe.periodSeconds`         | Seconds between polls               | 30                                      |
          | `nexus.livenessProbe.failureThreshold`      | Number of attempts before failure   | 6                                       |
          | `nexus.livenessProbe.timeoutSeconds`        | Time in seconds after liveness probe times out    | `nil`                     |
          | `nexus.livenessProbe.path`                  | Path for LivenessProbe              | /                                       |
          | `nexus.readinessProbe.initialDelaySeconds`  | ReadinessProbe initial delay        | 30                                      |
          | `nexus.readinessProbe.periodSeconds`        | Seconds between polls               | 30                                      |
          | `nexus.readinessProbe.failureThreshold`     | Number of attempts before failure   | 6                                       |
          | `nexus.readinessProbe.timeoutSeconds`       | Time in seconds after readiness probe times out    | `nil`                    |
          | `nexus.readinessProbe.path`                 | Path for ReadinessProbe             | /                                       |
          | `nexus.hostAliases`                         | Aliases for IPs in /etc/hosts       | []                                      |
          | `ingress.enabled`                           | Create an ingress for Nexus         | `true`                                  |
          | `ingress.annotations`                       | Annotations to enhance ingress configuration  | `{}`                          |
          | `ingress.tls.enabled`                       | Enable TLS                          | `true`                                 |
          | `ingress.tls.secretName`                    | Name of the secret storing TLS cert, `false` to use the Ingress' default certificate | `nexus-tls`                             |
          | `ingress.path`                              | Path for ingress rules. GCP users should set to `/*` | `/`                    |
          | `tolerations`                               | tolerations list                    | `[]`                                    |
          | `config.enabled`                            | Enable configmap                    | `false`                                 |
          | `config.mountPath`                          | Path to mount the config            | `/sonatype-nexus-conf`                  |
          | `config.data`                               | Configmap data                      | `nil`                                   |
          | `deployment.terminationGracePeriodSeconds`  | Time to allow for clean shutdown    | 120                                     |
          | `deployment.annotations`                    | Annotations to enhance deployment configuration  | `{}`                       |
          | `deployment.initContainers`                 | Init containers to run before main containers  | `nil`                        |
          | `deployment.postStart.command`              | Command to run after starting the nexus container  | `nil`                    |
          | `deployment.preStart.command`               | Command to run before starting the nexus container  | `nil`                   |
          | `deployment.additionalContainers`           | Add additional Container         | `nil`                                      |
          | `deployment.additionalVolumes`              | Add additional Volumes           | `nil`                                      |
          | `deployment.additionalVolumeMounts`         | Add additional Volume mounts     | `nil`                                      |
          | `secret.enabled`                            | Enable secret                    | `false`                                    |
          | `secret.mountPath`                          | Path to mount the secret         | `/etc/secret-volume`                       |
          | `secret.readOnly`                           | Secret readonly state            | `true`                                     |
          | `secret.data`                               | Secret data                      | `nil`                                      |
          | `service.enabled`                           | Enable additional service        | `nil`                                      |
          | `service.name`                              | Service name                     | `nil`                                      |
          | `service.portName`                          | Service port name                | `nil`                                      |
          | `service.labels`                            | Service labels                   | `nil`                                      |
          | `service.annotations`                       | Service annotations              | `nil`                                      |
          | `service.loadBalancerSourceRanges`          | Service LoadBalancer source IP whitelist | `nil`                              |
          | `service.targetPort`                        | Service port                     | `nil`                                      |
          | `service.port`                              | Port for exposing service        | `nil`                                      |
          | `route.enabled`         | Set to true to create route for additional service | `false` |
          | `route.name`            | Name of route                                      | `docker` |
          | `route.portName`        | Target port name of service                        | `docker` |
          | `route.labels`          | Labels to be added to route                        | `{}` |
          | `route.annotations`     | Annotations to be added to route                   | `{}` |
          | `route.path`            | Host name of Route e.g jenkins.example.com         | nil |
        displayName: Nexus Repository Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - repository
        - sonatype
        links:
        - name: Nexus Repository Manager
          url: https://www.sonatype.com/product-nexus-repository
        maintainers:
        - email: support@sonatype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:06fe31f40ff3fd7556fc7b8f6b2efe80ada1d9079ec9aaad60d1fdc0ee0677f0
        - registry.connect.redhat.com/sonatype/nxrm-operator-certified@sha256:4945ad993a10e27b358b2d9869d05f23ca761162ec088e5a833939022bc23192
        version: 3.58.1-2
      entries:
      - name: nxrm-operator-certified.v3.58.1-2
        version: 3.58.1-2
      name: stable
    defaultChannel: stable
    packageName: nxrm-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: AMD
      provider-url: ""
    name: amd-t2-january
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amd-t2-january.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovfect2.amd.com/v1",
                "kind": "SriovT2Card",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "amd-t2-january",
                    "app.kubernetes.io/instance": "sriovt2card-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sriovt2card",
                    "app.kubernetes.io/part-of": "amd-t2-january"
                  },
                  "name": "sriovt2card",
                  "namespace": "amd-t2"
                },
                "spec": {
                  "acceleratorSelector": {
                    "pciAddress": "0000:01:00.0"
                  },
                  "nodeSelector": {
                    "kubernetes.io/hostname": "7c-c2-55-85-ba-fe.amd.com"
                  },
                  "physicalFunction": {
                    "pfDriver": "qdma-pf",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-07-14T09:46:54Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovT2Card is the Schema for the sriovt2cards API
            displayName: Sriov T2 Card
            kind: SriovT2Card
            name: sriovt2cards.sriovfect2.amd.com
            version: v1
        description: The AMD-T2 Operator for accelerator enables the seamless integration
          and management of AMD T2 accelerator within Openshift/Kubernetes envirnoment.
        displayName: AMD-T2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - SRIOV
        - Xilinx
        - FPGA
        - Telco
        - T2
        - MultiArch
        - AMD
        - TelcoAccelerator
        links:
        - name: Amd T2 January
          url: https://amd-t2-january.domain
        maintainers:
        - email: dhirajs.lakhane@amd.com
          name: DhirajLakhane
        maturity: alpha
        provider:
          name: AMD
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:928e64203edad8f1bba23593c7be04f0f8410c6e4feb98d9e9c2d00a8ff59048
        - quay.io/amdt2operator/amd-t2-january@sha256:318f56e2670582c217b8973f69f0675e7fff7dc19d07875294f2ff2fc5403902
        version: 1.0.1
      entries:
      - name: amd-t2-january.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: amd-t2-january
    provider:
      name: AMD
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: placement-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: placement-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "placement.openstack.org/v1beta1",
                "kind": "PlacementAPI",
                "metadata": {
                  "name": "placement"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "placement",
                  "databaseInstance": "openstack",
                  "preserveJobs": false,
                  "replicas": 1,
                  "secret": "placement-secret",
                  "serviceUser": "placement"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:43Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PlacementAPI is the Schema for the placementapis API
            displayName: Placement API
            kind: PlacementAPI
            name: placementapis.placement.openstack.org
            version: v1beta1
        description: Placement Operator
        displayName: Placement Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Placement
        links:
        - name: Placement Operator
          url: https://github.com/openstack-k8s-operators/placement-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/placement-rhel9-operator@sha256:d3be48618b1f0f717aea162ae41876137fee8b35329d2e3e66e9d4eaccf62376
        - registry.redhat.io/rhoso/openstack-placement-api-rhel9@sha256:3308f2eccfbe7416a85443696273417bfb816df36b810ebbefa108dda12a5b19
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: placement-operator.v1.0.6
        version: 1.0.6
      - name: placement-operator.v1.0.5
        version: 1.0.5
      - name: placement-operator.v1.0.4
        version: 1.0.4
      - name: placement-operator.v1.0.3
        version: 1.0.3
      - name: placement-operator.v1.0.2
        version: 1.0.2
      - name: placement-operator.v1.0.1
        version: 1.0.1
      - name: placement-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: placement-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/cluster-nfd-operator
    name: nfd
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfd.4.16.0-202507211806
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nfd.openshift.io/v1",
                "kind": "NodeFeatureDiscovery",
                "metadata": {
                  "name": "nfd-instance",
                  "namespace": "openshift-nfd"
                },
                "spec": {
                  "customConfig": {
                    "configData": "#    - name: \"more.kernel.features\"\n#      matchOn:\n#      - loadedKMod: [\"example_kmod3\"]\n#    - name: \"more.features.by.nodename\"\n#      value: customValue\n#      matchOn:\n#      - nodename: [\"special-.*-node-.*\"]\n"
                  },
                  "operand": {
                    "image": "registry.redhat.io/openshift4/ose-node-feature-discovery-rhel9@sha256:bb65ec6a6f88688630305e081cb4bcb88aa538b5f6a3385ee28c61522dcd1662",
                    "imagePullPolicy": "IfNotPresent",
                    "servicePort": 12000
                  },
                  "workerConfig": {
                    "configData": "core:\n#  labelWhiteList:\n#  noPublish: false\n  sleepInterval: 60s\n#  sources: [all]\n#  klog:\n#    addDirHeader: false\n#    alsologtostderr: false\n#    logBacktraceAt:\n#    logtostderr: true\n#    skipHeaders: false\n#    stderrthreshold: 2\n#    v: 0\n#    vmodule:\n##   NOTE: the following options are not dynamically run-time \n##          configurable and require a nfd-worker restart to take effect\n##          after being changed\n#    logDir:\n#    logFile:\n#    logFileMaxSize: 1800\n#    skipLogHeaders: false\nsources:\n#  cpu:\n#    cpuid:\n##     NOTE: whitelist has priority over blacklist\n#      attributeBlacklist:\n#        - \"BMI1\"\n#        - \"BMI2\"\n#        - \"CLMUL\"\n#        - \"CMOV\"\n#        - \"CX16\"\n#        - \"ERMS\"\n#        - \"F16C\"\n#        - \"HTT\"\n#        - \"LZCNT\"\n#        - \"MMX\"\n#        - \"MMXEXT\"\n#        - \"NX\"\n#        - \"POPCNT\"\n#        - \"RDRAND\"\n#        - \"RDSEED\"\n#        - \"RDTSCP\"\n#        - \"SGX\"\n#        - \"SSE\"\n#        - \"SSE2\"\n#        - \"SSE3\"\n#        - \"SSE4.1\"\n#        - \"SSE4.2\"\n#        - \"SSSE3\"\n#      attributeWhitelist:\n#  kernel:\n#    kconfigFile: \"/path/to/kconfig\"\n#    configOpts:\n#      - \"NO_HZ\"\n#      - \"X86\"\n#      - \"DMI\"\n  pci:\n    deviceClassWhitelist:\n      - \"0200\"\n      - \"03\"\n      - \"12\"\n    deviceLabelFields:\n#      - \"class\"\n      - \"vendor\"\n#      - \"device\"\n#      - \"subsystem_vendor\"\n#      - \"subsystem_device\"\n#  usb:\n#    deviceClassWhitelist:\n#      - \"0e\"\n#      - \"ef\"\n#      - \"fe\"\n#      - \"ff\"\n#    deviceLabelFields:\n#      - \"class\"\n#      - \"vendor\"\n#      - \"device\"\n#  custom:\n#    - name: \"my.kernel.feature\"\n#      matchOn:\n#        - loadedKMod: [\"example_kmod1\", \"example_kmod2\"]\n#    - name: \"my.pci.feature\"\n#      matchOn:\n#        - pciId:\n#            class: [\"0200\"]\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#        - pciId :\n#            vendor: [\"8086\"]\n#            device: [\"1000\", \"1100\"]\n#    - name: \"my.usb.feature\"\n#      matchOn:\n#        - usbId:\n#          class: [\"ff\"]\n#          vendor: [\"03e7\"]\n#          device: [\"2485\"]\n#        - usbId:\n#          class: [\"fe\"]\n#          vendor: [\"1a6e\"]\n#          device: [\"089a\"]\n#    - name: \"my.combined.feature\"\n#      matchOn:\n#        - pciId:\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#          loadedKMod : [\"vendor_kmod1\", \"vendor_kmod2\"]\n"
                  }
                }
              },
              {
                "apiVersion": "nfd.openshift.io/v1alpha1",
                "kind": "NodeFeatureRule",
                "metadata": {
                  "name": "my-sample-rule-object"
                },
                "spec": {
                  "rules": [
                    {
                      "labels": {
                        "my-sample-feature": "true"
                      },
                      "matchFeatures": [
                        {
                          "feature": "kernel.loadedmodule",
                          "matchExpressions": {
                            "dummy": {
                              "op": "Exists"
                            }
                          }
                        },
                        {
                          "feature": "kernel.config",
                          "matchExpressions": {
                            "X86": {
                              "op": "In",
                              "value": [
                                "y"
                              ]
                            }
                          }
                        }
                      ],
                      "name": "my sample rule"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery,OpenShift Optional
          containerImage: registry.redhat.io/openshift4/ose-cluster-nfd-rhel9-operator@sha256:c1b6a5b8a59c08458d57c0fc2bb2bc9b056f006d15edb4d3161a3d1c0bc97ae6
          description: |-
            The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

            NFD consists  of the following software components:

            The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

            ##NFD-Master
            NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

            ##NFD-Worker
            NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

            ##NFD-Topology-Updater
            NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.0-202507211806'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-nfd
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/internal-objects: '["noderesourcetopologies.topology.node.k8s.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/cluster-nfd-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              The NodeFeatureDiscovery instance is the CustomResource being watched by the NFD-Operator, and holds all the needed information to setup the behaviour of the master and worker pods
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
          - description: |
              NodeFeatureRule resource specifies a configuration for feature-based customization of node objects, such as node labeling.
            kind: NodeFeatureRule
            name: nodefeaturerules.nfd.openshift.io
            version: v1alpha1
          - kind: NodeFeature
            name: nodefeatures.nfd.openshift.io
            version: v1alpha1
        description: |-
          The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

          NFD consists  of the following software components:

          The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

          ##NFD-Master
          NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

          ##NFD-Worker
          NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

          ##NFD-Topology-Updater
          NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
        displayName: Node Feature Discovery Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - feature-discovery
        - feature-detection
        - node-labels
        links:
        - name: Node Feature Discovery Operator
          url: https://docs.openshift.com/container-platform/4.16/hardware_enablement/psap-node-feature-discovery-operator.html
        - name: Node Feature Discovery Documentation
          url: https://kubernetes-sigs.github.io/node-feature-discovery/stable/get-started/index.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/cluster-nfd-operator
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:bcca89fb000b0fe8722ad7807b62a43ef3ae8fa76db1d790c285a975bfecb1aa
        - registry.redhat.io/openshift4/ose-cluster-nfd-rhel9-operator@sha256:c1b6a5b8a59c08458d57c0fc2bb2bc9b056f006d15edb4d3161a3d1c0bc97ae6
        - registry.redhat.io/openshift4/ose-node-feature-discovery-rhel9@sha256:bb65ec6a6f88688630305e081cb4bcb88aa538b5f6a3385ee28c61522dcd1662
        version: 4.16.0-202507211806
      entries:
      - name: nfd.4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: nfd.4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: nfd.4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: nfd.4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: nfd.4.16.0-202504271139
        version: 4.16.0-202504271139
      - name: nfd.4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: nfd.4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: nfd.4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: nfd.4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: nfd.4.16.0-202501271512
        version: 4.16.0-202501271512
      - name: nfd.4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: nfd.4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: nfd.4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: nfd.4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: nfd.4.16.0-202410251436
        version: 4.16.0-202410251436
      - name: nfd.4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: nfd.4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: nfd.4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: nfd.4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: nfd.4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: nfd.4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: nfd.4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: nfd.4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: nfd.4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: nfd.4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: nfd.4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: nfd.4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: nfd.4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: nfd.4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: nfd
    provider:
      name: Red Hat
      url: https://github.com/openshift/cluster-nfd-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhpam-kogito-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhpam-kogito-operator.7.13.5-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "ruleunit-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogitoinfra-sample"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "kafka.strimzi.io/v1beta2",
                    "kind": "Kafka",
                    "name": "kogito-kafka"
                  }
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator@sha256:19497d4aff0ea3ea114dccb0d0e265458bb14ba6d325353eb82b554547dbd2e4
          description: RHPAM Kogito Operator for deployment and management of Kogito
            services.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Process Automation
            Manager", "Red Hat Decision Manager"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/rhpam-kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.rhpam.kiegroup.org
            version: v1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.rhpam.kiegroup.org
            version: v1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.rhpam.kiegroup.org
            version: v1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.rhpam.kiegroup.org
            version: v1
        description: |-
          RHPAM Kogito Operator is designed for deployment and management of Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Builds a RHPAM Kogito service from a remote git repository and creates a custom image for the given runtime: Red Hat build of Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Deploy RHPAM Kogito service from user provided image.
          * Setup Grafana Dashboard when Grafana Operator is available in cluster.
          * Create Prometheus monitor object when Prometheus operator is available in cluster.

          **IMPORTANT!** RHPAM Kogito Operator may need Grafana and Prometheus Operator to configure monitoring in cases if you do not provide these services beforehand. If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the RHPAM Kogito Operator.
        displayName: RHPAM Kogito Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-process-automation-manager
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_process_automation_manager/7.13/html/getting_started_with_red_hat_build_of_kogito_in_red_hat_process_automation_manager/assembly-deploying-kogito-microservices-on-openshift
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator@sha256:19497d4aff0ea3ea114dccb0d0e265458bb14ba6d325353eb82b554547dbd2e4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9492c67eaa7ab37df6f5fa623c4aac955b7d7ffaab9ae2e57dea745e82907e40
        version: 7.13.5-4
      entries:
      - name: rhpam-kogito-operator.7.13.5-4
        version: 7.13.5-4
      - name: rhpam-kogito-operator.7.13.5-3
        version: 7.13.5-3
      - name: rhpam-kogito-operator.7.13.5-2
        version: 7.13.5-2
      - name: rhpam-kogito-operator.7.13.5-1
        version: 7.13.5-1
      - name: rhpam-kogito-operator.7.13.4-3
        version: 7.13.4-3
      - name: rhpam-kogito-operator.7.13.4-2
        version: 7.13.4-2
      - name: rhpam-kogito-operator.7.13.4-1
        version: 7.13.4-1
      - name: rhpam-kogito-operator.7.13.3-1
        version: 7.13.3-1
      - name: rhpam-kogito-operator.7.13.2-3
        version: 7.13.2-3
      - name: rhpam-kogito-operator.7.13.2-2
        version: 7.13.2-2
      - name: rhpam-kogito-operator.7.13.2-1
        version: 7.13.2-1
      - name: rhpam-kogito-operator.7.13.1-2
        version: 7.13.1-2
      - name: rhpam-kogito-operator.7.13.1-1
        version: 7.13.1-1
      - name: rhpam-kogito-operator.7.13.0-1
        version: 7.13.0-1
      - name: rhpam-kogito-operator.v7.12.0-4
        version: 7.12.0-4
      - name: rhpam-kogito-operator.v7.12.0-3
        version: 7.12.0-3
      - name: rhpam-kogito-operator.v7.12.0-2
        version: 7.12.0-2
      name: 7.x
    defaultChannel: 7.x
    packageName: rhpam-kogito-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://skupper.io
    name: skupper-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-operator.v1.4.8-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus@sha256:65602221cdff7f718d4149f4dead167d05dbbe0cea1eaa8bf4b98be588135583
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:c665677c639699d6a5ef24aed5a1a3d9c0a5ee25f077a6db42a9fc6247d93132
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:eedde9f2019d4e919612bcc18f427c0d88b4846e773751af231ba0e560d48869
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:1fd908e80ef21fe7d7051c400cb099f2f3ca51c8cdd698705ef9784d34688911
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:4c9cc9df83baa1ef37cfb429babe5b86a962b8bb096209b3fb74291c67c94000
        version: 1.4.8-rh-1
      entries:
      - name: skupper-operator.v1.4.8-rh-1
        version: 1.4.8-rh-1
      - name: skupper-operator.v1.4.7-rh-3
        version: 1.4.7-rh-3
      - name: skupper-operator.v1.4.7-rh-2
        version: 1.4.7-rh-2
      - name: skupper-operator.v1.4.7-rh-1
        version: 1.4.7-rh-1
      - name: skupper-operator.v1.4.5-rh-4
        version: 1.4.5-rh-4
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1
    - currentCSV: skupper-operator.v1.4.8-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:eedde9f2019d4e919612bcc18f427c0d88b4846e773751af231ba0e560d48869
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:1fd908e80ef21fe7d7051c400cb099f2f3ca51c8cdd698705ef9784d34688911
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:4c9cc9df83baa1ef37cfb429babe5b86a962b8bb096209b3fb74291c67c94000
        - registry.redhat.io/openshift4/ose-prometheus@sha256:65602221cdff7f718d4149f4dead167d05dbbe0cea1eaa8bf4b98be588135583
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:c665677c639699d6a5ef24aed5a1a3d9c0a5ee25f077a6db42a9fc6247d93132
        version: 1.4.8-rh-1
      entries:
      - name: skupper-operator.v1.4.8-rh-1
        version: 1.4.8-rh-1
      - name: skupper-operator.v1.4.7-rh-3
        version: 1.4.7-rh-3
      - name: skupper-operator.v1.4.7-rh-2
        version: 1.4.7-rh-2
      - name: skupper-operator.v1.4.7-rh-1
        version: 1.4.7-rh-1
      - name: skupper-operator.v1.4.5-rh-4
        version: 1.4.5-rh-4
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.4
    - currentCSV: skupper-operator.v1.5.5-rh-6
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:9e9e1f2288f0218e3ba2abf4bebac9731d7cf9f9c567e0f6de6767b0b66cbee4
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:a5d7c1b2b9db672548126c0873d062a55667d93291fd3e32809d3c9c900c548e
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:9e9e1f2288f0218e3ba2abf4bebac9731d7cf9f9c567e0f6de6767b0b66cbee4
        - registry.redhat.io/openshift4/ose-prometheus@sha256:53afaf53975807765fc3777314edd218c789c17680c1f57a8e33a8f9a76f38bf
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:3e78ec8cdb729a704d5a26e54697f285ba7c0b9afee6160b7ecc703876120dea
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:be9dbcd2a3a1e1e536921390dcfd283fc45dc4423d2705376ef37a9fb53f7cf1
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:1de071dabf620265252f3922c632f17f676ba716ce4c3df931a0ccb3871dbb6c
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:5f8621d0013febcecbed4d7c82db11a4dded99ad8ce6b6d83858b9bdbe79ee5a
        version: 1.5.5-rh-6
      entries:
      - name: skupper-operator.v1.5.5-rh-6
        version: 1.5.5-rh-6
      - name: skupper-operator.v1.5.5-rh-5
        version: 1.5.5-rh-5
      - name: skupper-operator.v1.5.5-rh-4
        version: 1.5.5-rh-4
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.5
    - currentCSV: skupper-operator.v1.8.7-rh-2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:b0d43628b779a101a24e4b271318a34c21f3b71bcba2bf20804f941a93d27455
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.8.7-rh-2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:56b8fde75a6e7df235c560a0f4b9114398da5ba8b20f75f1ea7338f48ce19506
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:d7d252cff08ba1589afd8707ed349f87d3b6214906a4fb49932bda0a4d19fee4
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:5ea36e96f878577104ab7601e57e7a2a233f11dbdc533454f95e2e14a066d21f
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:3adeef827eca0382cd410f7c3e99a0fee9b50791cb0e0c6fc130b72a2550684f
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:b0d43628b779a101a24e4b271318a34c21f3b71bcba2bf20804f941a93d27455
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:f798745e799515065af099a816057b9c98a741e951d30a0f45ed47ca50e6724e
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:61542c5f0989f51ac5d6c96f5d0e29aca75614b79e7626299c77c5a74636b564
        version: 1.8.7-rh-2
      entries:
      - name: skupper-operator.v1.8.7-rh-2
        version: 1.8.7-rh-2
      - name: skupper-operator.v1.8.7-rh-1
        version: 1.8.7-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.8
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.9
    - currentCSV: skupper-operator.v2.1.1-rh-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
          createdAt: "2025-08-19T20:43:19Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:36872037066fae806cbf88cab673229f25e592295972aa18227827233362eec5
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:c01d0eab0841b1a5fddb58a0ae110beb464b91390f02662590bebc7b2a7213b6
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e8df53454cf12a14338790af6ed209798010990ca3320096e8b44acf901c4123
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:1a290a5ac8b9d1b082e740054675fe432aa424c2f55dd662caf67c740044f229
        version: 2.1.1-rh-3
      entries:
      - name: skupper-operator.v2.1.1-rh-3
        version: 2.1.1-rh-3
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2
    - currentCSV: skupper-operator.v2.0.1-rh-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:b5c73dc59a9184cd120c2ee35256cc097c790b8ef488af0c732d5efe3192b768
          createdAt: "2025-05-29T13:35:40Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.0.1-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:92f44cff160874efe57a985609e599b13c49c0c4cef846dccc80833af2ae74f3
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:e239c859a1f49816ae3068455e703b8a12919f11158b0f6ee0e6bfb31e8a8082
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:f2a41674ed196c089748d6386c18c14580e502de33235ef23c37eddeb6350f41
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:95062ff1256135cfa17907e82b09b2a211371ea0e76ccc711b7c737e040d87e9
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:498b5bb648e5bbaad07663941600173abac7d60a10744305f619509ddf7563e0
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:bd3f72544841c91c2d573a65ac3c988756f3523e17e0988027793b0f5f3ec3d1
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:b5c73dc59a9184cd120c2ee35256cc097c790b8ef488af0c732d5efe3192b768
        version: 2.0.1-rh-2
      entries:
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2.0
    - currentCSV: skupper-operator.v2.1.1-rh-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
          createdAt: "2025-08-19T20:43:19Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:c01d0eab0841b1a5fddb58a0ae110beb464b91390f02662590bebc7b2a7213b6
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e8df53454cf12a14338790af6ed209798010990ca3320096e8b44acf901c4123
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:1a290a5ac8b9d1b082e740054675fe432aa424c2f55dd662caf67c740044f229
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:36872037066fae806cbf88cab673229f25e592295972aa18227827233362eec5
        version: 2.1.1-rh-3
      entries:
      - name: skupper-operator.v2.1.1-rh-3
        version: 2.1.1-rh-3
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2.1
    defaultChannel: stable-2
    packageName: skupper-operator
    provider:
      name: Red Hat
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://horreum.hyperfoil.io
    name: horreum-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: horreum-operator.v0.7.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hyperfoil.io/v1alpha1",
                "kind": "Horreum",
                "metadata": {
                  "name": "horreum"
                },
                "spec": {
                  "nodeHost": "127.0.0.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/hyperfoil/horreum-operator:0.7.9
          createdAt: "2023-06-13T10:42:39Z"
          description: Performance results repository
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/hyperfoil/horreum-operator
          support: Red Hat Application Runtimes Performance Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Horreum is the object configuring Horreum performance results
              repository
            displayName: Horreum
            kind: Horreum
            name: horreums.hyperfoil.io
            version: v1alpha1
        description: Performance results repository
        displayName: Horreum
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - performance
        - repository
        - database
        - results
        - hyperfoil
        - benchmark
        links:
        - name: Horreum Operator
          url: https://horreum-operator.domain
        maintainers:
        - email: rvansa@redhat.com
          name: Radim Vansa
        - email: spederse@redhat.com
          name: Ståle W. Pedersen
        - email: wreicher@redhat.com
          name: Will Reichert
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat, Inc.
          url: https://horreum.hyperfoil.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/hyperfoil/horreum-operator:0.7.9
        version: 0.7.9
      entries:
      - name: horreum-operator.v0.7.9
        version: 0.7.9
      - name: horreum-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: horreum-operator
    provider:
      name: Red Hat, Inc.
      url: https://horreum.hyperfoil.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/stolostron/multicluster-global-hub
    name: multicluster-global-hub-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-global-hub-operator.v1.4.0-dev.20250302
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2025-02-14T08:24:47Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: planning
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/postgresql-16:9.5-1732622748
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-inventory/inventory-api:latest
        - quay.io/prometheuscommunity/postgres-exporter:v0.15.0
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-relations/spicedb-operator:latest
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-relations/spicedb:latest
        - quay.io/stolostron/grafana:2.12.0-SNAPSHOT-2024-09-03-21-11-25
        - quay.io/stolostron/multicluster-global-hub-operator:v1.4.0
        - quay.io/stolostron/multicluster-global-hub-agent:v1.4.0
        - quay.io/stolostron/multicluster-global-hub-manager:v1.4.0
        version: 1.4.0-dev.20250302
      entries:
      - name: multicluster-global-hub-operator.v1.4.0-dev.20250302
        version: 1.4.0-dev.20250302
      - name: multicluster-global-hub-operator.v1.4.0-dev.20250225
        version: 1.4.0-dev.20250225
      - name: multicluster-global-hub-operator.v1.3.0-dev.20240911
        version: 1.3.0-dev.20240911
      - name: multicluster-global-hub-operator.v1.3.0-dev.20240724
        version: 1.3.0-dev.20240724
      - name: multicluster-global-hub-operator.v1.2.0-dev.20240516
        version: 1.2.0-dev.20240516
      - name: multicluster-global-hub-operator.v1.2.0-dev
        version: 1.2.0-dev
      - name: multicluster-global-hub-operator.v1.1.0-dev
        version: 1.1.0-dev
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: multicluster-global-hub-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-operator-globalhub-1-5@sha256:4db1251323bd64930461cb5025da41baffcf5c00c51ea4e49940b7130971b510
          createdAt: "2025-06-10T06:58:35Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-operator-globalhub-1-5@sha256:4db1251323bd64930461cb5025da41baffcf5c00c51ea4e49940b7130971b510
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-manager-globalhub-1-5@sha256:01253426d65f1297a1955a8393051e6a4b1df5e5249774fc5189a75f35e7d644
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-agent-globalhub-1-5@sha256:6f7520eac83abec7780f831d8473544d0b8db43fc6e77d59e59317b09eb1f6c1
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/glo-grafana-globalhub-1-5@sha256:73fdf00538e63595b7cc92d484e087ff815579a95d7259c579c6e7dc1beacff6
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/postgres-exporter-globalhub-1-5@sha256:1108572a8bdafde926254f1ff94c2b22e048706be660b07c793e9f9e038a0f4a
        version: 1.5.0
      entries:
      - name: multicluster-global-hub-operator.v1.5.0
        version: 1.5.0
      - name: multicluster-global-hub-operator.v1.4.0
        version: 1.4.0
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-1.5
    - currentCSV: multicluster-global-hub-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2024-03-05T02:13:54Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.32.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:latest
        version: 1.1.0
      entries:
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.10
    - currentCSV: multicluster-global-hub-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2024-07-03T06:52:02Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v1.2.0
        version: 1.2.1
      entries:
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.11
    - currentCSV: multicluster-global-hub-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator@sha256:636aea861da80fc36fb27a419a18c5b3fbb1bc072e8ae33c75e1cc47e2c05f75
          createdAt: "2024-10-10T05:42:48Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          kafka-zookeeper-0                                       1/1     Running   0          21m
          kafka-zookeeper-1                                       1/1     Running   0          21m
          kafka-zookeeper-2                                       1/1     Running   0          21m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgres-0                      2/2     Running   0          22m
          ```
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:636aea861da80fc36fb27a419a18c5b3fbb1bc072e8ae33c75e1cc47e2c05f75
        - quay.io/prometheuscommunity/postgres-exporter@sha256:386b12d19eab2a37d7cd8ca8b4c7491cc7a830d9581f49af6c98a393da9605e6
        - quay.io/stolostron/inventory-api@sha256:a15e62ee070dba5d74cf336b3b89b467593c5a33cfe3623d922f0af4fd7df392
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/release-globalhub-1-3/multicluster-global-hub-manager-globalhub-1-3@sha256:7656d01a4d7b39c61c3249695d68011d9e0df308627b9efd117686e901281ab1
        - quay.io/stolostron/multicluster-global-hub-manager@sha256:715cb5d0c6c407ec5c0303fc7e6d83fc77ce7e06775e169a475f7c075f1e7000
        - quay.io/stolostron/grafana@sha256:719b320a05d10158fadce5ae2fc6e4254a18957a6608b6f81e876cf46d3a6897
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        version: 1.3.0
      entries:
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.12
    - currentCSV: multicluster-global-hub-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator@sha256:085e9fc6cdc6a42df10a38856c8ccff25e5acffaba67ef95da72027e90e507fc
          createdAt: "2025-02-14T08:24:47Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:085e9fc6cdc6a42df10a38856c8ccff25e5acffaba67ef95da72027e90e507fc
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-inventory/inventory-api@sha256:6a55a1919ecd92d94e05c12be1b6450956f5fda2ce17af7027cac3806c4970a4
        - quay.io/prometheuscommunity/postgres-exporter@sha256:386b12d19eab2a37d7cd8ca8b4c7491cc7a830d9581f49af6c98a393da9605e6
        - quay.io/stolostron/multicluster-global-hub-manager@sha256:4cd602ff6f8dd951848dc8a95c856004295c9f4710f77d4dafcc0735c1492395
        - quay.io/stolostron/multicluster-global-hub-agent@sha256:8dfc33dfed33a85733ee5e02dcd4f05662c75263791e0961ba17cbb3d0c340c9
        - quay.io/stolostron/grafana@sha256:719b320a05d10158fadce5ae2fc6e4254a18957a6608b6f81e876cf46d3a6897
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        version: 1.4.0
      entries:
      - name: multicluster-global-hub-operator.v1.4.0
        version: 1.4.0
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.13
    - currentCSV: multicluster-global-hub-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "storage": {
                    "name": "storage-secret"
                  },
                  "transport": {
                    "name": "transport-secret"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.5.0
          createdAt: "2022-08-23T20:00:31Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha1
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.5 or later needs to be installed.\n- PostgreSQL needs to be installed
          and you must create database for the multicluster global hub. You must create
          a secret named postgresql-secret that contains the database access credential
          in `open-cluster-management` namespace. You can run the following command
          to create the secret: \n\n```\nkubectl create secret generic storage-secret
          -n open-cluster-management \\\n  --from-literal=database_uri=<postgresql-uri>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)
          to install postgres in `hoh-postgres` namespace and create the secret `storage-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n- Kafka needs to be installed,
          and you must create two sections, `spec` and `status`. You need to create
          a secret named kafka-secret that contains the kafka access information should
          be created in `open-cluster-management` namespace. You can run the following
          command to create the secret: \n\n```\nkubectl create secret generic transport-secret
          -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n  --from-literal=CA=<CA-for-kafka-server>\n```\n >_Note:_ There is a
          sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `transport-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n## How to Install\n\nInstall the
          Multicluster Global Hub Operator by following the instructions that are
          displayed after you select the Install button. The operator must be installed
          in the same namespace as Red Hat Advanced Cluster Management for Kubernetes
          (RHACM), by default, in `open-cluster-management` namespace.\n\nA pod will
          be created in `open-cluster-management` namespace\n\n```\n$ kubectl get
          pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          19s\n```\n\nThe operator is
          now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installing the operator,
          create an instance of the MulticlusterGlobalHub resource to install the
          Multicluster Global Hub.\n\nThe pods of Multicluster Global Hub are created
          in the `open-cluster-management` namespace:\n\n```\n$ kubectl get pods -n
          open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\nmulticluster-global-hub-rbac-77888b7dc4-z9vbp
          \                   1/1     Running   0          10s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.5.0
        version: 0.5.0
      entries:
      - name: multicluster-global-hub-operator.v0.5.0
        version: 0.5.0
      name: release-2.6
    - currentCSV: multicluster-global-hub-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha2",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "largeScale": {
                      "kafka": {
                        "name": "transport-secret"
                      },
                      "postgres": {
                        "name": "storage-secret"
                      }
                    },
                    "type": "largeScale"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.6.0
          createdAt: "2023-02-09T20:00:31Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha2
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.7 or later needs to be installed.\n- PostgreSQL needs to be installed
          and you must create database for the multicluster global hub. You must create
          a secret named postgresql-secret that contains the database access credential
          in `open-cluster-management` namespace. You can run the following command
          to create the secret: \n\n```\nkubectl create secret generic storage-secret
          -n open-cluster-management \\\n  --from-literal=database_uri=<postgresql-uri>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)
          to install postgres in `hoh-postgres` namespace and create the secret `storage-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n- Kafka needs to be installed,
          and you must create two sections, `spec` and `status`. You need to create
          a secret named kafka-secret that contains the kafka access information should
          be created in `open-cluster-management` namespace. You can run the following
          command to create the secret: \n\n```\nkubectl create secret generic transport-secret
          -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n  --from-literal=CA=<CA-for-kafka-server>\n```\n >_Note:_ There is a
          sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `transport-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n## How to Install\n\nInstall the
          Multicluster Global Hub Operator by following the instructions that are
          displayed after you select the Install button. The operator must be installed
          in the same namespace as Red Hat Advanced Cluster Management for Kubernetes
          (RHACM), by default, in `open-cluster-management` namespace.\n\nA pod will
          be created in `open-cluster-management` namespace\n\n```\n$ kubectl get
          pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          19s\n```\n\nThe operator is
          now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installing the operator,
          create an instance of the MulticlusterGlobalHub resource to install the
          Multicluster Global Hub.\n\nThe pods of Multicluster Global Hub are created
          in the `open-cluster-management` namespace:\n\n```\n$ kubectl get pods -n
          open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: multicluster-global-hub-operator.v0.6.0
        version: 0.6.0
      name: release-2.7
    - currentCSV: multicluster-global-hub-operator.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha3",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "largeScale": {
                      "kafka": {
                        "transportFormat": "cloudEvents"
                      }
                    },
                    "type": "largeScale"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.7.0
          createdAt: "2023-07-19T07:53:07Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha3
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.7 or later needs to be installed.\n## How to Install\n\nInstall the Multicluster
          Global Hub Operator by following the instructions that are displayed after
          you select the Install button. The operator must be installed in the same
          namespace as Red Hat Advanced Cluster Management for Kubernetes (RHACM),
          by default, in `open-cluster-management` namespace.\n\nA pod will be created
          in `open-cluster-management` namespace\n\n```\n$ kubectl get pods -n open-cluster-management\nNAME
          \                                                            READY   STATUS
          \   RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp                1/1
          \    Running   0          19s\n```\n\nThe operator is now providing new
          Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installed the operator,
          create an instance of the MulticlusterGlobalHub resource to instantiate
          the Multicluster Global Hub.\n\nYou need to prepare:\n\n- PostgreSQL needs
          to be installed and you must create database for the multicluster global
          hub. You must create a secret named multicluster-global-hub-storage that
          contains the database access credential in `open-cluster-management` namespace.
          You can run the following command to create the secret: \n\n```\nkubectl
          create secret generic multicluster-global-hub-storage -n open-cluster-management
          \\\n  --from-literal=database_uri=<postgresql-uri> \\\n --from-file=ca.crt=<CA-for-postgres-server>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)(Note:
          the client version of kubectl must be v1.21+) to install postgres in `hoh-postgres`
          namespace and create the secret `multicluster-global-hub-storage` in namespace
          `open-cluster-management` automatically. To override the secret namespace,
          set `TARGET_NAMESPACE` environment variable to the RHACM installation namespace
          before executing the script.\n\n- Kafka needs to be installed, and you must
          create three topics, `spec` `status` and `event`. You need to create a secret
          named multicluster-global-hub-transport that contains the kafka access information
          should be created in `open-cluster-management` namespace. You can run the
          following command to create the secret: \n\n```\nkubectl create secret generic
          multicluster-global-hub-transport -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n      --from-file=ca.crt=<CA-cert-for-kafka-server> --from-file=client.crt=<Client-cert-for-kafka-server>
          --from-file=client.key=<Client-key-for-kafka-server> \n```\n >_Note:_ There
          is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `multicluster-global-hub-transport`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\nOnce an instance of the MulticlusterGlobalHub
          is created, the following pods are created in the `open-cluster-management`
          namespace:\n\n```\n$ kubectl get pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-grafana-79f8d59f58-qwcjl
          \                1/1     Running   0          22s\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: beta
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.7.0
        version: 0.7.0
      entries:
      - name: multicluster-global-hub-operator.v0.7.0
        version: 0.7.0
      name: release-2.8
    - currentCSV: multicluster-global-hub-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v1.0.0
          createdAt: "2023-10-16T03:07:33Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the MultiCluster Global Hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.7 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```

          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v1.0.0
        version: 1.0.0
      entries:
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.9
    defaultChannel: release-1.5
    packageName: multicluster-global-hub-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/stolostron/multicluster-global-hub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhbk-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhbk-operator.v22.0.13-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
          createdAt: "2024-09-17T11:20:22Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:884afe0ca1698d85a02aa47b6d60b8222e50d70ad280db6f98e1a4bbdbe13c40
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
        version: 22.0.13-opr.1
      entries:
      - name: rhbk-operator.v22.0.13-opr.1
        version: 22.0.13-opr.1
      - name: rhbk-operator.v22.0.12-opr.1
        version: 22.0.12-opr.1
      - name: rhbk-operator.v22.0.11-opr.2
        version: 22.0.11-opr.2
      - name: rhbk-operator.v22.0.11-opr.1
        version: 22.0.11-opr.1
      - name: rhbk-operator.v22.0.10-opr.1
        version: 22.0.10-opr.1
      - name: rhbk-operator.v22.0.9-opr.1
        version: 22.0.9-opr.1
      - name: rhbk-operator.v22.0.8-opr.2
        version: 22.0.8-opr.2
      - name: rhbk-operator.v22.0.8-opr.1
        version: 22.0.8-opr.1
      - name: rhbk-operator.v22.0.7-opr.1
        version: 22.0.7-opr.1
      - name: rhbk-operator.v22.0.6-opr.1
        version: 22.0.6-opr.1
      name: stable-v22
    - currentCSV: rhbk-operator.v22.0.13-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
          createdAt: "2024-09-17T11:20:22Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:884afe0ca1698d85a02aa47b6d60b8222e50d70ad280db6f98e1a4bbdbe13c40
        version: 22.0.13-opr.1
      entries:
      - name: rhbk-operator.v22.0.13-opr.1
        version: 22.0.13-opr.1
      - name: rhbk-operator.v22.0.12-opr.1
        version: 22.0.12-opr.1
      - name: rhbk-operator.v22.0.11-opr.2
        version: 22.0.11-opr.2
      - name: rhbk-operator.v22.0.11-opr.1
        version: 22.0.11-opr.1
      - name: rhbk-operator.v22.0.10-opr.1
        version: 22.0.10-opr.1
      - name: rhbk-operator.v22.0.9-opr.1
        version: 22.0.9-opr.1
      - name: rhbk-operator.v22.0.8-opr.2
        version: 22.0.8-opr.2
      - name: rhbk-operator.v22.0.8-opr.1
        version: 22.0.8-opr.1
      - name: rhbk-operator.v22.0.7-opr.1
        version: 22.0.7-opr.1
      - name: rhbk-operator.v22.0.6-opr.1
        version: 22.0.6-opr.1
      name: stable-v22.0
    - currentCSV: rhbk-operator.v24.0.11-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
          createdAt: "2025-03-26T02:27:51Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:d339ecf788d1681ba9945a2ea63ca675caec863ec75f497ef916103f4c898003
        version: 24.0.11-opr.1
      entries:
      - name: rhbk-operator.v24.0.11-opr.1
        version: 24.0.11-opr.1
      - name: rhbk-operator.v24.0.10-opr.1
        version: 24.0.10-opr.1
      - name: rhbk-operator.v24.0.9-opr.1
        version: 24.0.9-opr.1
      - name: rhbk-operator.v24.0.8-opr.1
        version: 24.0.8-opr.1
      - name: rhbk-operator.v24.0.7-opr.1
        version: 24.0.7-opr.1
      - name: rhbk-operator.v24.0.6-opr.1
        version: 24.0.6-opr.1
      - name: rhbk-operator.v24.0.5-opr.1
        version: 24.0.5-opr.1
      - name: rhbk-operator.v24.0.4-opr.1
        version: 24.0.4-opr.1
      - name: rhbk-operator.v24.0.3-opr.1
        version: 24.0.3-opr.1
      name: stable-v24
    - currentCSV: rhbk-operator.v24.0.11-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
          createdAt: "2025-03-26T02:27:51Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:d339ecf788d1681ba9945a2ea63ca675caec863ec75f497ef916103f4c898003
        version: 24.0.11-opr.1
      entries:
      - name: rhbk-operator.v24.0.11-opr.1
        version: 24.0.11-opr.1
      - name: rhbk-operator.v24.0.10-opr.1
        version: 24.0.10-opr.1
      - name: rhbk-operator.v24.0.9-opr.1
        version: 24.0.9-opr.1
      - name: rhbk-operator.v24.0.8-opr.1
        version: 24.0.8-opr.1
      - name: rhbk-operator.v24.0.7-opr.1
        version: 24.0.7-opr.1
      - name: rhbk-operator.v24.0.6-opr.1
        version: 24.0.6-opr.1
      - name: rhbk-operator.v24.0.5-opr.1
        version: 24.0.5-opr.1
      - name: rhbk-operator.v24.0.4-opr.1
        version: 24.0.4-opr.1
      - name: rhbk-operator.v24.0.3-opr.1
        version: 24.0.3-opr.1
      name: stable-v24.0
    - currentCSV: rhbk-operator.v26.2.9-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
          createdAt: "2025-09-18T15:03:55Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:50a62b46af44ce0ff7e6c3a15e1bfb2a9611ba4a8b65b8d0ab65817064692c55
        version: 26.2.9-opr.1
      entries:
      - name: rhbk-operator.v26.2.9-opr.1
        version: 26.2.9-opr.1
      - name: rhbk-operator.v26.2.8-opr.1
        version: 26.2.8-opr.1
      - name: rhbk-operator.v26.2.7-opr.1
        version: 26.2.7-opr.1
      - name: rhbk-operator.v26.2.6-opr.1
        version: 26.2.6-opr.1
      - name: rhbk-operator.v26.2.5-opr.1
        version: 26.2.5-opr.1
      - name: rhbk-operator.v26.2.4-opr.2
        version: 26.2.4-opr.2
      - name: rhbk-operator.v26.2.4-opr.1
        version: 26.2.4-opr.1
      - name: rhbk-operator.v26.0.15-opr.1
        version: 26.0.15-opr.1
      - name: rhbk-operator.v26.0.14-opr.1
        version: 26.0.14-opr.1
      - name: rhbk-operator.v26.0.13-opr.1
        version: 26.0.13-opr.1
      - name: rhbk-operator.v26.0.12-opr.1
        version: 26.0.12-opr.1
      - name: rhbk-operator.v26.0.11-opr.1
        version: 26.0.11-opr.1
      - name: rhbk-operator.v26.0.10-opr.1
        version: 26.0.10-opr.1
      - name: rhbk-operator.v26.0.9-opr.1
        version: 26.0.9-opr.1
      - name: rhbk-operator.v26.0.8-opr.1
        version: 26.0.8-opr.1
      - name: rhbk-operator.v26.0.7-opr.1
        version: 26.0.7-opr.1
      - name: rhbk-operator.v26.0.6-opr.1
        version: 26.0.6-opr.1
      - name: rhbk-operator.v26.0.5-opr.1
        version: 26.0.5-opr.1
      name: stable-v26
    - currentCSV: rhbk-operator.v26.0.15-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:fc5fc8cfb2c100c5b3a33fc11a78dfca569f5f34f45d88439340392d2c60afb2
          createdAt: "2025-08-28T19:34:30Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:fc5fc8cfb2c100c5b3a33fc11a78dfca569f5f34f45d88439340392d2c60afb2
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:4bb4d3070fe118bdd7de24e8b68133b8b6e158cea48c9f1abad2b4297be0dfdb
        version: 26.0.15-opr.1
      entries:
      - name: rhbk-operator.v26.0.15-opr.1
        version: 26.0.15-opr.1
      - name: rhbk-operator.v26.0.14-opr.1
        version: 26.0.14-opr.1
      - name: rhbk-operator.v26.0.13-opr.1
        version: 26.0.13-opr.1
      - name: rhbk-operator.v26.0.12-opr.1
        version: 26.0.12-opr.1
      - name: rhbk-operator.v26.0.11-opr.1
        version: 26.0.11-opr.1
      - name: rhbk-operator.v26.0.10-opr.1
        version: 26.0.10-opr.1
      - name: rhbk-operator.v26.0.9-opr.1
        version: 26.0.9-opr.1
      - name: rhbk-operator.v26.0.8-opr.1
        version: 26.0.8-opr.1
      - name: rhbk-operator.v26.0.7-opr.1
        version: 26.0.7-opr.1
      - name: rhbk-operator.v26.0.6-opr.1
        version: 26.0.6-opr.1
      - name: rhbk-operator.v26.0.5-opr.1
        version: 26.0.5-opr.1
      name: stable-v26.0
    - currentCSV: rhbk-operator.v26.2.9-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
          createdAt: "2025-09-18T15:03:55Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:50a62b46af44ce0ff7e6c3a15e1bfb2a9611ba4a8b65b8d0ab65817064692c55
        version: 26.2.9-opr.1
      entries:
      - name: rhbk-operator.v26.2.9-opr.1
        version: 26.2.9-opr.1
      - name: rhbk-operator.v26.2.8-opr.1
        version: 26.2.8-opr.1
      - name: rhbk-operator.v26.2.7-opr.1
        version: 26.2.7-opr.1
      - name: rhbk-operator.v26.2.6-opr.1
        version: 26.2.6-opr.1
      - name: rhbk-operator.v26.2.5-opr.1
        version: 26.2.5-opr.1
      - name: rhbk-operator.v26.2.4-opr.2
        version: 26.2.4-opr.2
      - name: rhbk-operator.v26.2.4-opr.1
        version: 26.2.4-opr.1
      name: stable-v26.2
    defaultChannel: stable-v26.2
    packageName: rhbk-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-opensearchserverless-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-opensearchserverless-controller.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/opensearchserverless-controller:0.2.1
          createdAt: "2025-09-20T21:43:17Z"
          description: AWS OpenSearchServerless controller is a service controller
            for managing OpenSearchServerless resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Collection represents the state of an AWS opensearchserverless
              Collection resource.
            displayName: Collection
            kind: Collection
            name: collections.opensearchserverless.services.k8s.aws
            version: v1alpha1
          - description: SecurityPolicy represents the state of an AWS opensearchserverless
              SecurityPolicy resource.
            displayName: SecurityPolicy
            kind: SecurityPolicy
            name: securitypolicies.opensearchserverless.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon OpenSearchServerless resources in AWS from within your Kubernetes cluster.

          **About Amazon OpenSearchServerless**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon OpenSearchServerless
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opensearchserverless
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon OpenSearchServerless Developer Resources
          url: https://aws.amazon.com/OpenSearchServerless/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: opensearchserverless maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/opensearchserverless-controller:0.2.1
        version: 0.2.1
      entries:
      - name: ack-opensearchserverless-controller.v0.2.1
        version: 0.2.1
      - name: ack-opensearchserverless-controller.v0.2.0
        version: 0.2.0
      - name: ack-opensearchserverless-controller.v0.1.2
        version: 0.1.2
      - name: ack-opensearchserverless-controller.v0.1.1
        version: 0.1.1
      - name: ack-opensearchserverless-controller.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ack-opensearchserverless-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Automotive SIG CentOS
      provider-url: ""
    name: automotive-infra
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: automotive-infra.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "ContainerBuild",
                "metadata": {
                  "name": "cb-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "ImageBuild",
                "metadata": {
                  "name": "imagebuild-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "OperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "automotive-infra"
                  },
                  "name": "operatorconfig-sample",
                  "namespace": "automotive-infra-operator"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/fedora-sig-robotics/infra-ci-operator:v0.0.10
          createdAt: "2025-03-31T14:06:33Z"
          description: Streamline CI/CD for RHIVOS and its ecosystem using OpenShift
            Pipelines
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: automotive-infra-operator
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator
          support: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Client
            name: clients.jumpstarter.dev
            version: v1alpha1
          - kind: ContainerBuild
            name: containerbuilds.automotive.sig.centos.org
            version: v1
          - kind: ExporterAccessPolicy
            name: exporteraccesspolicies.jumpstarter.dev
            version: v1alpha1
          - kind: Exporter
            name: exporters.jumpstarter.dev
            version: v1alpha1
          - kind: ImageBuild
            name: imagebuilds.automotive.sig.centos.org
            version: v1
          - kind: Lease
            name: leases.jumpstarter.dev
            version: v1alpha1
          - kind: OperatorConfig
            name: operatorconfigs.automotive.sig.centos.org
            version: v1
        description: This operator is designed to streamline CI/CD for RHIVOS and
          its ecosystem using OpenShift Pipelines
        displayName: Automotive-CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Automotive
        - RHIVOS
        - Pipelines
        - CI
        links:
        - name: Automotive infra Operator
          url: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator
        maintainers:
        - email: automotive-sig@lists.centos.org
          name: Automotive SIG CentOS Community
        maturity: alpha
        provider:
          name: Automotive SIG CentOS
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/fedora-sig-robotics/infra-ci-operator:v0.0.10
        version: 0.0.10
      entries:
      - name: automotive-infra.v0.0.10
        version: 0.0.10
      - name: automotive-infra.v0.0.9
        version: 0.0.9
      - name: automotive-infra.v0.0.8
        version: 0.0.8
      - name: automotive-infra.v0.0.7
        version: 0.0.7
      - name: automotive-infra.v0.0.6
        version: 0.0.6
      - name: automotive-infra.v0.0.5
        version: 0.0.5
      - name: automotive-infra.v0.0.4
        version: 0.0.4
      - name: automotive-infra.v0.0.3
        version: 0.0.3
      - name: automotive-infra.v0.0.2
        version: 0.0.2
      - name: automotive-infra.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: automotive-infra
    provider:
      name: Automotive SIG CentOS
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: ibm-licensing-operator
      app.kubernetes.io/managed-by: ibm-licensing-operator
      app.kubernetes.io/name: ibm-licensing
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      intent: projected
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-licensing-operator-app
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-licensing-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "IBMLicensing",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-licensing-operator",
                    "app.kubernetes.io/managed-by": "ibm-licensing-operator",
                    "app.kubernetes.io/name": "ibm-licensing"
                  },
                  "name": "instance"
                },
                "spec": {
                  "version": "4.2.14",
                  "apiSecretToken": "ibm-licensing-token",
                  "datasource": "datacollector",
                  "httpsEnable": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
          createdAt: "2025-02-25T12:16:11Z"
          description: The IBM Licensing Operator provides a Kubernetes CRD-Based
            API to monitor the license usage of products.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-licensing-operator-app
          olm.skipRange: '>=1.0.0 <4.2.14'
          operatorframework.io/suggested-namespace: ibm-licensing
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/internal-objects: '["ibmlicensingmetadatas.operator.ibm.com",
            "ibmlicensingdefinitions.operator.ibm.com", "ibmlicensingquerysources.operator.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'IBMLicensing custom resource is used to create an instance
              of the License Service, used to collect information about license usage
              of IBM containerized products and IBM Cloud Paks per cluster. You can
              retrieve license usage data through a dedicated API call and generate
              an audit snapshot on demand. Documentation: For additional details regarding
              install parameters check: https://ibm.biz/icpfs39install. License: Please
              refer to the IBM Terms website (ibm.biz/lsvc-lic) to find the license
              terms for the particular IBM product for which you are deploying this
              component.'
            displayName: IBM License Service
            kind: IBMLicensing
            name: ibmlicensings.operator.ibm.com
            version: v1alpha1
          - kind: IBMLicensingDefinition
            name: ibmlicensingdefinitions.operator.ibm.com
            version: v1
          - description: 'IBMLicensingMetadata is the schema for IBM License Service.
              Thanks to IBMLicensingMetadata, you can track the license usage of Virtual
              Processor Core (VPC) metric by Pods that are managed by automation,
              for which licensing annotations are not defined at the time of deployment,
              such as the open source OpenLiberty. / For more information, see documentation:
              https://ibm.biz/icpfs39install. By installing the operator for IBM License
              Service, you accept the license terms: https://ibm.biz/icpfs39license.'
            displayName: IBMLicensing Metadata
            kind: IBMLicensingMetadata
            name: ibmlicensingmetadatas.operator.ibm.com
            version: v1alpha1
          - kind: IBMLicensingQuerySource
            name: ibmlicensingquerysources.operator.ibm.com
            version: v1
        description: |-
          **Important:**
          - If you are using the IBM Licensing Operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For the link to your IBM Cloud Pak documentation, see [IBM Cloud Paks that use Common Services](https://ibm.biz/BdyGwb).
          - If you are using a stand-alone IBM Container Software, you can use the IBM Licensing Operator directly. For more information, see [ibm-licensing-operator for stand-alone IBM Containerized Software](https://ibm.biz/BdyGwh).

          **IBM Licensing Operator overview**

          IBM Licensing Operator installs License Service. You can use License Service to collect information about license usage of IBM Containerized products and IBM Cloud Paks per cluster. You can retrieve license usage data through a dedicated API call and generate an audit snapshot on demand.

          **Supported platforms**

           Red Hat OpenShift Container Platform 4.10 or newer installed on Linux x86_64, Linux on Power (ppc64le), Linux on IBM Z and LinuxONE.

          **Prerequisites**

           Prerequisites depend on the integration of the License Service with an IBM Cloud Pak or IBM Containerized Software. For more information, see the applicable IBM Cloud Pak documentation or [ibm-licensing-operator for stand-alone IBM Containerized Software](https://ibm.biz/BdyGwh).
        displayName: IBM Licensing
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBMLicensing
        - IBM
        - Cloud
        - License
        - Service
        - License Service
        links:
        - name: GitHub
          url: https://github.com/IBM/ibm-licensing-operator
        maintainers:
        - email: talk2sam@us.ibm.com
          name: talk2sam
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9186c350fa54e8b792d72e47be6b54a67b610f726175c9641795010d45ad51f9
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:ec137115210dae9b69cb2899ed8a5437d94c8a658ae1ae5eb295232b34b0925e
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:96e6db93b2d16ba590252d6568b68f254548aaed73ec17e4bdd8d530ea668be6
        version: 4.2.14
      entries:
      - name: ibm-licensing-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-licensing-operator-app
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator-rhmp.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/dynatrace-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/dynatrace-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator-rhmp.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator-rhmp.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator-rhmp.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator-rhmp.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator-rhmp.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator-rhmp.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator-rhmp.v1.5.0
        version: 1.5.0
      - name: dynatrace-operator-rhmp.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator-rhmp.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator-rhmp.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator-rhmp
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Argo CD Community
      provider-url: ""
    name: argocd-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: argocd-operator.v0.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd-sample"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCDExport",
                "metadata": {
                  "name": "argocdexport-sample"
                },
                "spec": {
                  "argocd": "argocd-sample"
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd-sample"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/argoprojlabs/argocd-operator@sha256:223ad82fb697471ede79201c4a042115c9ca5c00a172d5bdcd5a780e58c46c79
          createdAt: "2025-09-11T06:29:33Z"
          description: Argo CD is a declarative, GitOps continuous delivery tool for
            Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/argoproj-labs/argocd-operator
          support: Argo CD
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: An ApplicationSet is a group or set of Application resources.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: ArgoCDExport is the Schema for the argocdexports API
            displayName: Argo CDExport
            kind: ArgoCDExport
            name: argocdexports.argoproj.io
            version: v1alpha1
          - description: ArgoCD is the Schema for the argocds API
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: ArgoCD is the Schema for the argocds API
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: NotificationsConfiguration is the Schema for the notificationsconfiguration
              API
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
        description: |
          ## Overview

          The Argo CD Operator manages the full lifecycle for [Argo CD](https://argoproj.github.io/argo-cd/) and it's
          components. The operator's goal is to automate the tasks required when operating an Argo CD cluster.

          Beyond installation, the operator helps to automate the process of upgrading, backing up and restoring as needed and
          remove the human as much as possible. In addition, the operator aims to provide deep insights into the Argo CD
          environment by configuring Prometheus and Grafana to aggregate, visualize and expose the metrics already exported by
          Argo CD.

          The operator aims to provide the following, and is a work in progress.

          * Easy configuration and installation of the Argo CD components with sane defaults to get up and running quickly.
          * Provide seamless upgrades to the Argo CD components.
          * Ability to back up and restore an Argo CD cluster from a point in time or on a recurring schedule.
          * Aggregate and expose the metrics for Argo CD and the operator itself using Prometheus and Grafana.
          * Autoscale the Argo CD components as necessary to handle variability in demand.

          ## Usage

          Deploy a basic Argo CD cluster by creating a new ArgoCD resource in the namespace where the operator is installed.

          ```
          apiVersion: argoproj.io/v1alpha1
          kind: ArgoCD
          metadata:
            name: example-argocd
          spec: {}
          ```

          ## Backup

          Backup the cluster above by creating a new ArgoCDExport resource in the namespace where the operator is installed.

          ```
          apiVersion: argoproj.io/v1alpha1
          kind: ArgoCDExport
          metadata:
            name: example-argocdexport
          spec:
            argocd: example-argocd
          ```

          See the [documentation](https://argocd-operator.readthedocs.io) and examples on
          [GitHub](https://github.com/argoproj-labs/argocd-operator) for more information.
        displayName: Argo CD
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - kubernetes
        links:
        - name: Argo CD Project
          url: https://argoproj.github.io/argo-cd/
        - name: Operator Documentation
          url: https://argocd-operator.readthedocs.io
        - name: Operator Source Code
          url: https://github.com/argoproj-labs/argocd-operator
        maintainers:
        - email: aveerama@redhat.com
          name: Abhishek Veeramalla
        maturity: alpha
        provider:
          name: Argo CD Community
        relatedImages:
        - quay.io/argoprojlabs/argocd-operator@sha256:223ad82fb697471ede79201c4a042115c9ca5c00a172d5bdcd5a780e58c46c79
        version: 0.15.0
      entries:
      - name: argocd-operator.v0.15.0
        version: 0.15.0
      - name: argocd-operator.v0.14.1
        version: 0.14.1
      - name: argocd-operator.v0.14.0
        version: 0.14.0
      - name: argocd-operator.v0.13.0
        version: 0.13.0
      - name: argocd-operator.v0.12.0
        version: 0.12.0
      - name: argocd-operator.v0.11.0
        version: 0.11.0
      - name: argocd-operator.v0.10.1
        version: 0.10.1
      - name: argocd-operator.v0.10.0
        version: 0.10.0
      - name: argocd-operator.v0.9.1
        version: 0.9.1
      - name: argocd-operator.v0.9.0
        version: 0.9.0
      - name: argocd-operator.v0.8.0
        version: 0.8.0
      - name: argocd-operator.v0.7.0
        version: 0.7.0
      - name: argocd-operator.v0.5.0
        version: 0.5.0
      - name: argocd-operator.v0.4.0
        version: 0.4.0
      - name: argocd-operator.v0.3.0
        version: 0.3.0
      - name: argocd-operator.v0.2.1
        version: 0.2.1
      - name: argocd-operator.v0.2.0
        version: 0.2.0
      - name: argocd-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: argocd-operator
    provider:
      name: Argo CD Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-framework
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-framework.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Framework",
                "metadata": {
                  "name": "framework-sample"
                },
                "spec": {
                  "kcos-eula": {
                    "image": {
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    },
                    "service": {
                      "nodePort": 0,
                      "type": "ClusterIP"
                    }
                  },
                  "kcos-logging-framework": {
                    "image": {
                      "remoteRepository": "docker.io",
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    },
                    "ingress": {
                      "annotations": {
                        "nginx.ingress.kubernetes.io/auth-method": "POST",
                        "nginx.ingress.kubernetes.io/auth-response-headers": "X-AUTH-USER",
                        "nginx.ingress.kubernetes.io/auth-url": "https://127.0.0.1/token/validate",
                        "nginx.ingress.kubernetes.io/cors-allow-credentials": "true",
                        "nginx.ingress.kubernetes.io/cors-allow-methods": "PUT, GET, POST, OPTIONS, PATCH",
                        "nginx.ingress.kubernetes.io/cors-allow-origin": "https://admin.example.com,https://localhost:4200",
                        "nginx.ingress.kubernetes.io/enable-cors": "true",
                        "nginx.ingress.kubernetes.io/proxy-body-size": "0",
                        "nginx.ingress.kubernetes.io/ssl-redirect": "true"
                      },
                      "enabled": true
                    }
                  },
                  "kcos-postgresql": {
                    "jobCopySecret": {
                      "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                    },
                    "postgresql": {
                      "image": {
                        "registry": "prod.harbor.keysight.digital/loadcore"
                      },
                      "initdbScripts": {
                        "init_script.sh": "#!/bin/sh\nexport PGUSER=postgres\npsql -c \"SELECT 1 FROM pg_database WHERE datname = 'db'\" | grep -q 1 || psql -c \"CREATE DATABASE db\"\npsql -c \"SELECT 1 FROM pg_database WHERE datname = 'kcoslicensingdb'\" | grep -q 1 || psql -c \"CREATE DATABASE kcoslicensingdb\"\n"
                      },
                      "metrics": {
                        "image": {
                          "registry": "docker.io"
                        }
                      },
                      "persistence": {
                        "storageClass": "managed-loadcore-storage"
                      },
                      "pgHbaConfiguration": "local all all trust\nhost all all 0.0.0.0/0 trust",
                      "securityContext": {
                        "enabled": false
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": false
                        }
                      },
                      "volumePermissions": {
                        "image": {
                          "registry": "docker.io"
                        },
                        "securityContext": {
                          "runAsUser": "auto"
                        }
                      }
                    }
                  },
                  "kcos-system-diagnostics": {
                    "environment": {
                      "ELASTICSEARCH_ENDPOINT": "elasticsearch.keysight-wap.svc.cluster.local.",
                      "IGNORE_DATA_PREFIX": "wireless_stats"
                    },
                    "image": {
                      "initContainer": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf",
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    }
                  },
                  "subcharts": {
                    "enable": {
                      "kcos-coredns-fixup": false,
                      "kcos-eula": true,
                      "kcos-local-storage": false,
                      "kcos-logging-framework": true,
                      "kcos-logging-rotate": false,
                      "kcos-naas": false,
                      "kcos-postgresql": true,
                      "kcos-shell": false,
                      "kcos-system-diagnostics": true,
                      "kcos-vital-service": false,
                      "nfs-server-provisioner": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:27:28Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Framework
            name: frameworks.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Framework Operator
        displayName: Keysight KCOS Framework Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Framework
          url: https://keysight-kcos-framework.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-framework@sha256:531f278def85e33d453479fce74127e12e3ed62a251872eab421c6561bedd0c5
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        version: 0.1.21
      entries:
      - name: keysight-kcos-framework.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-framework
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx-openshift-operator.v0.0.36
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
          createdAt: "2019-07-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX on JBoss EAP 7
            displayName: Joget DX on JBoss EAP 7
            kind: JogetDX
            name: jogetdx.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX7/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
        - registry.connect.redhat.com/joget/joget-dx7-eap7@sha256:7245501f5fccf011d75c034b7d55542cfe3bd2c64cc7e48b3224a48b623af282
        version: 0.0.36
      entries:
      - name: joget-dx-openshift-operator.v0.0.36
        version: 0.0.36
      - name: joget-dx-openshift-operator.v0.0.35
        version: 0.0.35
      - name: joget-dx-openshift-operator.v0.0.34
        version: 0.0.34
      - name: joget-dx-openshift-operator.v0.0.33
        version: 0.0.33
      - name: joget-dx-openshift-operator.v0.0.32
        version: 0.0.32
      - name: joget-dx-openshift-operator.v0.0.31
        version: 0.0.31
      - name: joget-dx-openshift-operator.v0.0.30
        version: 0.0.30
      - name: joget-dx-openshift-operator.v0.0.29
        version: 0.0.29
      - name: joget-dx-openshift-operator.v0.0.28
        version: 0.0.28
      - name: joget-dx-openshift-operator.v0.0.27
        version: 0.0.27
      - name: joget-dx-openshift-operator.v0.0.26
        version: 0.0.26
      - name: joget-dx-openshift-operator.v0.0.25
        version: 0.0.25
      - name: joget-dx-openshift-operator.v0.0.24
        version: 0.0.24
      - name: joget-dx-openshift-operator.v0.0.23
        version: 0.0.23
      - name: joget-dx-openshift-operator.v0.0.22
        version: 0.0.22
      - name: joget-dx-openshift-operator.v0.0.21
        version: 0.0.21
      - name: joget-dx-openshift-operator.v0.0.20
        version: 0.0.20
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Habana Labs Ltd.
      provider-url: https://habana.ai
    name: habana-ai-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: habana-ai-operator.v1.19.2-32-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.19.2-32"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.19.2"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.19.2-32"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.19.2-32"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.19.2-32"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.19.2-32"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.19.2-32"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.19.2-32"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.19.2-32"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.19.2-32"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.19.2-32"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:975ed2135061d3bdc145c83abaa15f99e4c9f6efb15c708de41aa00fbcb72333
          createdAt: "2025-02-10T07:33:30Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.19.2-32-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi AI accelerator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:ac16752b53df69573da257d3a3a717a214b2eb1656779a7ef94084d74363fb46
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:b12802e7ef05273fceaef3a9987c3c4ae760714e6f333363f665eb26be9175e9
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:f05cd78d7593b877ec8a75175a9faa72d9b991d81c74af9ecd92208fde296eca
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:170a15041fc3e3901b4e8e0b9a32d53fde11026776c17a6560d500dacb976c56
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:d728a91a36503f07dfacd49fdbfefe8083607ca33babb5f2dfc1e9de05386454
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:e667695b845cd2be887f117dac66861d2c5e74def009d2939559ae7c30574dc0
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:975ed2135061d3bdc145c83abaa15f99e4c9f6efb15c708de41aa00fbcb72333
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:eeb4573de20d3341a085c75e669b7d37c424e348421d7916440fbbfff12ea482
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:71d4c7181930c71c0ca4e6b696ea416a687eec6d798a43e854fcd1bfdc224a0c
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:72cc75353615369d12a07f5488bcdc041a8435906ca6e16c53b001016108e099
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:373b9caa27a96bb7979f7a6aec246795142ce3a28c6b97a8902035ec20814fa2
        version: 1.19.2-32-1
      entries:
      - name: habana-ai-operator.v1.19.2-32-1
        version: 1.19.2-32-1
      name: "1.19"
    - currentCSV: habana-ai-operator.v1.20.0-543-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.20.0-543"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.20.0-4"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.20.0-543-2"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.20.0-543"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.20.0-543"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.20.0-543-4"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.20.0-543"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.20.0-543-4"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:d5b37eafe48446e1ffa492e195f4519b77d4a11f6750e5c71f27c1536bdfd2cc
          createdAt: "2025-02-26T09:03:44Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.20.0-543-5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:9ab54a84a00e8afc078eec2a7d643f09b1bd674e0e230733fbce7108e392e009
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:d5b37eafe48446e1ffa492e195f4519b77d4a11f6750e5c71f27c1536bdfd2cc
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:317765eac4e0332a0c4f6bead5563b343a8c22e6e0eaa62276ff66be08608c30
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:c5beac6209d0671d3a49b63f1cf95401055b042430eafaefc3859b5c633f0e84
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:68d0f44170cbf3ea15b793e0503fb07ce7e32488011fc85ccb6c9234b3b9302f
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:3ea80aa8994e8b19fbbdcd18ba2c35800774da3ca4e27a2514d0f5f9676786fa
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:957702b58ee66d7339a20b2a5ece783606d6ccebde627c58951d5ecd6bb1b13f
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:cba03259b33c4cb9bcd22622ab0dda962846ea0a80c727779365cf55129ad308
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:d14c457106b6133fc41ef727ac40a6bf2f0e5c97f21659969d1daebdc1310b43
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:c662de244c5900c89b918b16982298676d9a6d553360a5716174b28d4d64a635
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:a673900f08c38c9bb40c90b3cddea580afcc6a99195d816cd0b5cae9831ca5c9
        version: 1.20.0-543-5
      entries:
      - name: habana-ai-operator.v1.20.0-543-5
        version: 1.20.0-543-5
      name: "1.20"
    - currentCSV: habana-ai-operator.v1.21.5-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.21.5-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.21.5-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.21.5-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:0741da99140d137474b9e7edba921672eab8ff6277e1dd8bf3a4179e87f01dd4
          createdAt: "2025-09-17T11:55:32Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.21.5-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:0741da99140d137474b9e7edba921672eab8ff6277e1dd8bf3a4179e87f01dd4
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:a216ec6a6e7df50e067c557e9ed5871a3308b2ba42a4e41c9642f0c2c4503d88
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:9692d079ef18a364ed235c187482f329b3a5764a773e363acb1238044ad4a884
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:c95d3ffd6c91be3418415740258d33a8dbcedad6fe7d6bfe68d4dd122027447b
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:bc3749ff153fff71e252820cbf22fcd0753b195b2204f9731e42af29fb7ddaa0
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:bd57e7059467f0ca1667d19a70bd9a6ad96b83bbcb53c09fdad5a262fee76af9
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:c84e997c7d8c22c4d4e902a29e45e5364f0ad306f17d03a7d8ddb4391142e167
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:256cd8bebec3ed835e187e925a14a6b605ce9b3a17fe7d8886e0b32cde163d90
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:1e488c8ad31784dab0fe70144f5146d5809e2fc7c86424d1b512ab887787e28f
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:c7147359dcd6a2e5adee95259d37743ff302b1deaeb13b184044f4029d33f44a
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:d7f64d15c379524597aa2b2c5b4586c890442fa32356692013cf665b01b0f092
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:a0b45ccfb6af54b1043ccc2336ad3d7990cc447af5ba3621f3eb1953cd40ffc5
        version: 1.21.5-6
      entries:
      - name: habana-ai-operator.v1.21.5-6
        version: 1.21.5-6
      name: "1.21"
    - currentCSV: habana-ai-operator.v1.22.1-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.22.1-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
          createdAt: "2025-09-17T10:15:21Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.22.1-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:1e15f3de21c2b1e20f54d1f8a973290c6d940624506880c5bc3eb880011bcc02
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:d1974d72774c626e77707e8f44bcc631ac142bfe2a5395714969a964c963e36e
        - vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer@sha256:2cf151c2a518e3e1c7c09efb8301e469d266d1f030fff2e6180b086cc2e04ee7
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:704b4bebdfe6b85693df4b213f67862859a554c21ce4e614736172f64b541ea7
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:c9ff6f3fe24ef87c7bfc71bdb1cd7900941df52f9ad1293badd7e14783ce7ef6
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:e34a6820a58e7e8090b521781c76acc3867c94f8291634de7b138f928e0bdb74
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:53c1844b5e8eb08d0cc29865a732100208c5dfc0fe55df8b32e589e319bcab6c
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:38c21eae6249e5571921998660418e59870bc0880e7b8c9b95a4a8db30ef2e7d
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:1925586fda34c6fa444a59b1fa7be20b3cf16e845bf56566ad2a41c276b0cfe9
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:392fcd2a4730299e6e3bab9f7e96fc3b710de8adb374a2c64b7316e097ed05f9
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:7140877b23291a5db3e5e28def3c9d5e7cb7be62f046e2fb7ac45eff369cea84
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:8f413edfff1978b35ba3998ed5f1c6293e6764b1d2cd9faa08f0bd81dc6e3e10
        version: 1.22.1-6
      entries:
      - name: habana-ai-operator.v1.22.1-6
        version: 1.22.1-6
      name: "1.22"
    - currentCSV: habana-ai-operator.v1.22.1-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.22.1-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
          createdAt: "2025-09-17T10:15:21Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.22.1-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:c9ff6f3fe24ef87c7bfc71bdb1cd7900941df52f9ad1293badd7e14783ce7ef6
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:392fcd2a4730299e6e3bab9f7e96fc3b710de8adb374a2c64b7316e097ed05f9
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:38c21eae6249e5571921998660418e59870bc0880e7b8c9b95a4a8db30ef2e7d
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:1925586fda34c6fa444a59b1fa7be20b3cf16e845bf56566ad2a41c276b0cfe9
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:704b4bebdfe6b85693df4b213f67862859a554c21ce4e614736172f64b541ea7
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:7140877b23291a5db3e5e28def3c9d5e7cb7be62f046e2fb7ac45eff369cea84
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:8f413edfff1978b35ba3998ed5f1c6293e6764b1d2cd9faa08f0bd81dc6e3e10
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:1e15f3de21c2b1e20f54d1f8a973290c6d940624506880c5bc3eb880011bcc02
        - vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer@sha256:2cf151c2a518e3e1c7c09efb8301e469d266d1f030fff2e6180b086cc2e04ee7
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:e34a6820a58e7e8090b521781c76acc3867c94f8291634de7b138f928e0bdb74
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:53c1844b5e8eb08d0cc29865a732100208c5dfc0fe55df8b32e589e319bcab6c
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:d1974d72774c626e77707e8f44bcc631ac142bfe2a5395714969a964c963e36e
        version: 1.22.1-6
      entries:
      - name: habana-ai-operator.v1.22.1-6
        version: 1.22.1-6
      name: stable
    defaultChannel: stable
    packageName: habana-ai-operator
    provider:
      name: Habana Labs Ltd.
      url: https://habana.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-multicluster-orchestrator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-multicluster-orchestrator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "items": [
                    {
                      "clusterName": "local-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    },
                    {
                      "clusterName": "spoke-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    }
                  ],
                  "manageS3": false,
                  "schedulingIntervals": [
                    "5m"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["odf-multicluster-console"]'
          createdAt: "2025-09-12T07:33:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          Orchestrator for OpenShift Data Foundation clusters running across multiple OpenShift clusters.
          It uses Red Hat Advanced Cluster Management for Kubernetes as the multicluster control plane.
        displayName: ODF Multicluster Orchestrator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: ODF Multicluster Orchestrator
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: uchapaga@redhat.com
          name: Umanga Chapagain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-multicluster-rhel9-operator@sha256:09e2c7a20ab92ddc5d589072bb0ab9ce2ed8706ed8364f1b11192d7a7396e138
        - registry.redhat.io/odf4/odf-multicluster-console-rhel9@sha256:7527f4a14aed0386b8f160fa8a92b7ff1f58114970e07c3ec6c433d2a53459b5
        version: 4.15.20-rhodf
      entries:
      - name: odf-multicluster-orchestrator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odf-multicluster-orchestrator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odf-multicluster-orchestrator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odf-multicluster-orchestrator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odf-multicluster-orchestrator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odf-multicluster-orchestrator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odf-multicluster-orchestrator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odf-multicluster-orchestrator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odf-multicluster-orchestrator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odf-multicluster-orchestrator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odf-multicluster-orchestrator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odf-multicluster-orchestrator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odf-multicluster-orchestrator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odf-multicluster-orchestrator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odf-multicluster-orchestrator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odf-multicluster-orchestrator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odf-multicluster-orchestrator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odf-multicluster-orchestrator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odf-multicluster-orchestrator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odf-multicluster-orchestrator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odf-multicluster-orchestrator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "items": [
                    {
                      "clusterName": "local-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    },
                    {
                      "clusterName": "spoke-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    }
                  ],
                  "manageS3": false,
                  "schedulingIntervals": [
                    "5m"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["odf-multicluster-console"]'
          createdAt: "2025-09-12T06:32:35Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          Orchestrator for OpenShift Data Foundation clusters running across multiple OpenShift clusters.
          It uses Red Hat Advanced Cluster Management for Kubernetes as the multicluster control plane.
        displayName: ODF Multicluster Orchestrator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: ODF Multicluster Orchestrator
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: uchapaga@redhat.com
          name: Umanga Chapagain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-multicluster-rhel9-operator@sha256:78d6bc14b95d1ee91ed529422062ec191afc489d6ab4e75906ae3c7fe58749cb
        - registry.redhat.io/odf4/odf-multicluster-console-rhel9@sha256:97d3645560388f6ae08fc12cafb58b6c766e38d83de04b645b3e6fb03fbba231
        version: 4.16.16-rhodf
      entries:
      - name: odf-multicluster-orchestrator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odf-multicluster-orchestrator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odf-multicluster-orchestrator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odf-multicluster-orchestrator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odf-multicluster-orchestrator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odf-multicluster-orchestrator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odf-multicluster-orchestrator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odf-multicluster-orchestrator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odf-multicluster-orchestrator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odf-multicluster-orchestrator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odf-multicluster-orchestrator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odf-multicluster-orchestrator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odf-multicluster-orchestrator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odf-multicluster-orchestrator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odf-multicluster-orchestrator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odf-multicluster-orchestrator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odf-multicluster-orchestrator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odf-multicluster-orchestrator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-access-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-access-operator.v23.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyAccess",
                "metadata": {
                  "name": "isva-sample"
                },
                "spec": {
                  "image": "icr.io/isva/verify-access-wrp:10.0.5.0",
                  "instance": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isva/verify-access-operator:23.3.0
          createdAt: Mon Mar  20 00:59:34 UTC 2023
          description: The IBM Security Verify Access Operator manages the lifecycle
            of IBM Security Verify Access worker containers.
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM-Security/verify-access-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyAccess is the Schema for the ibmsecurityverifyaccesses
              API.
            displayName: IBMSecurity Verify Access
            kind: IBMSecurityVerifyAccess
            name: ibmsecurityverifyaccesses.ibm.com
            version: v1
        description: |+
          In a world of highly fragmented access management environments, [IBM Security Verify Access](https://www.ibm.com/au-en/products/verify-access) helps you simplify your users' access while more securely adopting web, mobile and cloud technologies. This solution helps you strike a balance between usability and security through the use of risk-based access, single sign-on, integrated access management control, identity federation and its mobile multi-factor authentication capability, IBM Verify. Take back control of your access management with IBM Security Verify Access.

          The IBM Security Verify Access operator provides lifecycle management of the lightweight containers which are used to protect an environment, namely:
          * [Web Reverse Proxy](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-web-reverse-proxy)
          * [Runtime](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-runtime)
          * [Distributed Session Cache](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-distributed-session-cache)

          The operator will manage the deployment of these lightweight IBM Security Verify Access worker containers, and also control the rolling restart of these containers when a configuration snapshot is updated.

          See the project [Readme](https://github.com/IBM-Security/verify-access-operator/blob/master/README.md) for further information and details.

        displayName: IBM Security Verify Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Access Product Information
          url: https://www.ibm.com/au-en/products/verify-access
        - name: Verify Access Documentation
          url: https://www.ibm.com/docs/en/sva
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Access Development Team
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - icr.io/isva/verify-access-operator:23.3.0
        version: 23.3.0
      entries:
      - name: ibm-security-verify-access-operator.v23.3.0
        version: 23.3.0
      - name: ibm-security-verify-access-operator.v22.10.0
        version: 22.10.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-access-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: oadp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oadp-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "Velero",
                "metadata": {
                  "name": "example-velero",
                  "namespace":"openshift-adp"
                },
                "spec": {
                  "backupStorageLocations": [
                    {
                      "config": {
                        "profile": "default",
                        "region": "us-east-1"
                      },
                      "name": "default",
                      "objectStorage": {
                        "bucket": "myBucket",
                        "prefix": "velero"
                      },
                      "provider": "aws"
                    }
                  ],
                  "defaultVeleroPlugins": [
                    "aws",
                    "csi",
                    "openshift"
                  ],
                  "veleroFeatureFlags": [
                    "EnableCSI"
                  ],
                  "enableRestic": true,
                  "volumeSnapshotLocations": [
                    {
                      "config": {
                        "profile": "default",
                        "region": "us-west-1"
                      },
                      "name": "default",
                      "provider": "aws"
                    }
                  ]
                }
               },
                {
                  "apiVersion": "velero.io/v1",
                  "kind": "Backup",
                  "metadata": {
                    "name": "backup",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "BackupStorageLocation",
                  "metadata": {
                    "name": "backupstoragelocation",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "DeleteBackupRequest",
                  "metadata": {
                    "name": "deletebackuprequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "DownloadRequest",
                  "metadata": {
                    "name": "downloadrequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "PodVolumeBackup",
                  "metadata": {
                    "name": "podvolumebackup",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "PodVolumeRestore",
                  "metadata": {
                    "name": "podvolumerestore",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "ResticRepository",
                  "metadata": {
                    "name": "resticrepository",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "Restore",
                  "metadata": {
                    "name": "restore",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "Schedule",
                  "metadata": {
                    "name": "schedule",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "ServerStatusRequest",
                  "metadata": {
                    "name": "serverstatusrequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "VolumeSnapshotLocation",
                  "metadata": {
                    "name": "volumesnapshotlocation",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/konveyor/oadp-operator@sha256:ecb8534fbc245e890b9c7766946d01e1a8b4ce9be359c4ada4c95fd908b679b3
          createdAt: "2020-09-08T12:21:00Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          olm.skipRange: '>=0.0.0 <0.4.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Velero is a tool to safely backup and restore, perform disaster
              recovery and migrate Kubernetes cluster resources and persistent volumes.
            displayName: Velero
            kind: Velero
            name: veleroes.oadp.openshift.io
            version: v1alpha1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - description: Backup is a Velero resource that respresents the capture
              of Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Velero on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://oadp.konveyor.io/
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/konveyor/velero-restic-restore-helper@sha256:c4d12829e42cabe70b662a659875900677dfb415f912f931c3bf4c09a26aed5b
        - quay.io/konveyor/openshift-velero-plugin@sha256:aec8b4fc66cfcebd5a35ce99d8c985453ebf25c11453d30256089a8b425990b5
        - quay.io/konveyor/velero-plugin-for-aws@sha256:2197c6fef32910747526d33c1462b75cb7940aeace65791d0a5ddaa038480e52
        - quay.io/konveyor/velero-plugin-for-csi@sha256:5c2261effc902d8baa16e38265f9ad1fbe5284a3e6abd2bbc97761c161fa92cf
        - quay.io/konveyor/registry@sha256:b74ab3278b79eb89f99b6970c6aa0f5306221a8495b5496faad6cab2f74ff2cd
        - quay.io/konveyor/oadp-operator@sha256:ecb8534fbc245e890b9c7766946d01e1a8b4ce9be359c4ada4c95fd908b679b3
        - quay.io/konveyor/velero-plugin-for-microsoft-azure@sha256:62a344e5cf196e55dae92e537648987d758ccc75985bc45e9be6d0646a9e64d0
        - quay.io/konveyor/velero-plugin-for-gcp@sha256:3300458b1d6c1514dca7d24353ad03bd92db396b2a29f7d3522563160a5f5219
        - quay.io/konveyor/velero@sha256:e7865cc707664eb6ee767e77ccd933420604c2562736b53e16c8043a51df9d86
        version: 0.4.2
      entries:
      - name: oadp-operator.v0.4.2
        version: 0.4.2
      - name: oadp-operator.v0.4.1
        version: 0.4.1
      - name: oadp-operator.v0.4.0
        version: 0.4.0
      - name: oadp-operator.v0.3.0
        version: 0.3.0
      name: beta
    - currentCSV: oadp-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "restic": {
                      "enable": true
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ResticRepository",
                "metadata": {
                  "name": "resticrepository",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/konveyor/oadp-operator@sha256:90609044d52173619ea1e63ddef0ef3e7c1c939ac737b64bd427cbed5374dd27
          createdAt: "2020-09-08T12:21:00Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          olm.skipRange: '>=0.0.0 <0.5.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is a Velero resource that respresents the capture
              of Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Data Protection Applications on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://oadp.konveyor.io/
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/konveyor/registry@sha256:1be970879206258335357e3cf630138c24b8a5ccab35aae05eccb260e5e8ffb4
        - quay.io/konveyor/openshift-velero-plugin@sha256:221ab0bcd1d95fc166ef7319d187e196b59d2fbf9eefd29d91bd6435d45b5c3c
        - quay.io/konveyor/oadp-operator@sha256:90609044d52173619ea1e63ddef0ef3e7c1c939ac737b64bd427cbed5374dd27
        - quay.io/konveyor/velero@sha256:0ea5336dcbf0a1fab561697b24da01a0e5b533ee23553fbfd5f8e002f5a31c3e
        - quay.io/konveyor/velero-restic-restore-helper@sha256:c16248d027c09cd5b7360ae38484df88ca1cc01f9e470e3eade95076ee1a6b84
        - quay.io/konveyor/velero-plugin-for-aws@sha256:1e8bc27d6b60000e1406085d582910cb530af3fb901ef9bc780628df6c766288
        - quay.io/konveyor/velero-plugin-for-microsoft-azure@sha256:2d0089cbe715b482fb42e1acd9add70b975a8575636d8b143e6f90cef317ea13
        - quay.io/konveyor/velero-plugin-for-gcp@sha256:ee463ed2a8019c560961b118f8a87b587ed302f24fd25480c5d4ddad2557e92f
        - quay.io/konveyor/velero-plugin-for-csi@sha256:8c36ccfd294dafc415a7f113770cbbd66dbf6ecf8fef3bc4e73adc9878c606b0
        version: 0.5.6
      entries:
      - name: oadp-operator.v0.5.6
        version: 0.5.6
      - name: oadp-operator.v0.5.5
        version: 0.5.5
      - name: oadp-operator.v0.5.4
        version: 0.5.4
      - name: oadp-operator.v0.5.3
        version: 0.5.3
      - name: oadp-operator.v0.5.2
        version: 0.5.2
      - name: oadp-operator.v0.5.1
        version: 0.5.1
      - name: oadp-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: oadp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Trustification
      provider-url: ""
    name: trustify-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trustify-operator.v0.1.0-alpha.9
      currentCSVDesc:
        annotations:
          alm-examples: |
            [{
              "apiVersion": "org.trustify/v1alpha1",
              "kind": "Trustify",
              "metadata": {
                "name": "myapp"
              },
              "spec": { }
            }]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          containerImage: ghcr.io/trustification/trustify-operator:v0.1.0-alpha.9
          createdAt: "2025-03-07T11:57:13+00:00"
          description: An Operator for installing and managing Trustify
          repository: https://github.com/trustification/trustify-operator
          support: https://github.com/trustification/trustify-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Trustify instance
            displayName: Trustify
            kind: Trustify
            name: trustifies.org.trustify
            version: v1alpha1
        description: |+
          Trustify is vendor-neutral, thought-leadering, mostly informational collection of resources devoted to making Software Supply Chains easier to create, manage, consume and ultimately... to trust!


          You can use Trustify for:

          - Store and Manage all the SBOM (Software Bill of Materials) files of your company.
          - Understand which are the Vulnerabilities that affect each of your SBOM files
          - Understand exactly which Packages are included/shipped within each SBOM

          Trustify is a project within the [Trustification community](https://trustification.io/).


          ### Install
          Once you have successfully installed the Operator, proceed to deploy components by creating the required CR.

          By default, the Operator installs the following components on a target cluster:

          * Server
          * UI

          ### Documentation
          Documentation can be found on our [website](https://trustification.io/).

          ### Getting help
          If you encounter any issues while using Trustify, you can create an issue on our [Github repo](https://github.com/trustification/trustify/issues), for bugs, enhancements or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find
          * Fixing issues by opening [Pull Requests](https://github.com/trustification/trustify/pulls)

        displayName: Trustify Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trust
        links:
        - name: Website
          url: https://trustification.io/
        - name: Github
          url: https://github.com/trustification/trustify
        maintainers:
        - email: trustification@googlegroups.com
          name: Trustification
        maturity: alpha
        minKubeVersion: 1.23.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Trustification
        relatedImages:
        - ghcr.io/trustification/trustify-operator:v0.1.0-alpha.9
        version: 0.1.0-alpha.9
      entries:
      - name: trustify-operator.v0.1.0-alpha.9
        version: 0.1.0-alpha.9
      - name: trustify-operator.v0.1.0-alpha.8
        version: 0.1.0-alpha.8
      - name: trustify-operator.v0.1.0-alpha.7
        version: 0.1.0-alpha.7
      - name: trustify-operator.v0.1.0-alpha.6
        version: 0.1.0-alpha.6
      - name: trustify-operator.v0.1.0-alpha.5
        version: 0.1.0-alpha.5
      - name: trustify-operator.v0.1.0-alpha.4
        version: 0.1.0-alpha.4
      - name: trustify-operator.v0.1.0-alpha.3
        version: 0.1.0-alpha.3
      - name: trustify-operator.v0.1.0-alpha.2
        version: 0.1.0-alpha.2
      name: alpha
    defaultChannel: alpha
    packageName: trustify-operator
    provider:
      name: Trustification
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-directory-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-directory-operator.v23.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyDirectory",
                "metadata": {
                  "name": "ibmsecurityverifydirectory-sample"
                },
                "spec": {
                  "pods": {
                    "configMap": {
                      "proxy": {
                        "key": "config.yaml",
                        "name": "isvd-proxy-config"
                      },
                      "server": {
                        "key": "config.yaml",
                        "name": "isvd-server-config"
                      }
                    },
                    "image": {
                      "label": "latest",
                      "repo": "icr.io/isvd"
                    },
                    "proxy": {
                      "replicas": 1
                    },
                    "serviceAccountName": "isvd"
                  },
                  "replicas": {
                    "pvcs": [
                      "replica-1",
                      "replica-2"
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: icr.io/isvd/verify-directory-operator@sha256:d1dd7037584c6818dcf8d5aa69472278833bb8e37fba2aec5a69185ba88dcc34
          createdAt: "2023-04-18T22:10:32Z"
          description: The IBM Security Verify Directory operator is used to scale
            an IBM Security Verify Directory Server environment.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ibm-security/verify-directory-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyDirectory is the Schema for the ibmsecurityverifydirectories
              API
            displayName: IBMSecurity Verify Directory
            kind: IBMSecurityVerifyDirectory
            name: ibmsecurityverifydirectories.ibm.com
            version: v1
        description: "IBM Security Verify Directory is a scalable, standards-based
          identity directory that helps simplify identity and directory management.
          Verify Directory helps consolidate identity silos into a single identity
          source. Verify Directory is purpose-built to provide a directory foundation
          that can help provide a trusted identity data infrastructure that assists
          in enabling mission-critical security and authentication. It is designed
          to deliver a reliable, scalable, standards-based identity data platform
          that interoperates with a broad range of operating systems and applications.
          Verify Directory supports Lightweight Directory Access Protocol (LDAP) V3,
          offering a flexible and highly scalable LDAP infrastructure. \n\nThe operator
          can be used to manage deployed instances of the IBM Security Verify Directory
          server, providing an easy mechanism to scale the number of servers which
          are used to handle the workload.\n\nSee the project [Readme](https://github.com/ibm-security/verify-directory-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Security Verify Directory
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Security
        - Identity and Access
        links:
        - name: IBM Security Verify Directory Documentation
          url: https://www.ibm.com/docs/en/svd
        maintainers:
        - email: iamdev@us.ibm.com
          name: IBM Security Verify Directory Development Team
        maturity: stable
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - icr.io/isvd/verify-directory-operator@sha256:d1dd7037584c6818dcf8d5aa69472278833bb8e37fba2aec5a69185ba88dcc34
        version: 23.4.1
      entries:
      - name: ibm-security-verify-directory-operator.v23.4.1
        version: 23.4.1
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-directory-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Flux project
      provider-url: ""
    name: flux
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.toolkit.fluxcd.io/v2beta1",
                "kind": "HelmRelease",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "5m",
                    "chart": {
                      "spec": {
                          "chart": "podinfo",
                          "version": "4.0.x",
                          "sourceRef": {
                            "kind": "HelmRepository",
                            "name": "podinfo",
                            "namespace": "flux-system"
                          },
                          "interval": "1m"
                      }
                    },
                    "values": {
                      "replicaCount": 2
                    }
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "GitRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "1m",
                    "ref": {
                      "branch": "master"
                    },
                    "url": "https://github.com/stefanprodan/podinfo"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmRepository",
                "metadata": {
                    "name": "stable",
                    "namespace": "default"
                },
                "spec": {
                    "url": "https://kubernetes-charts.storage.googleapis.com/",
                    "interval": "10m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmChart",
                "metadata": {
                    "name": "redis",
                    "namespace": "default"
                },
                "spec": {
                    "chart": "redis",
                    "version": "10.5.7",
                    "sourceRef": {
                      "name": "stable",
                      "kind": "HelmRepository"
                    },
                    "interval": "5m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "Bucket",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "1m",
                    "provider": "generic",
                    "bucketName": "podinfo",
                    "endpoint": "minio.minio.svc.cluster.local:9000",
                    "insecure": true,
                    "secretRef": {
                      "name": "minio-credentials"
                    }
                }
              },
              {
                "apiVersion": "kustomize.toolkit.fluxcd.io/v1beta2",
                "kind": "Kustomization",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "5m",
                    "path": "./deploy/overlays/dev",
                    "prune": true,
                    "sourceRef": {
                      "kind": "GitRepository",
                      "name": "podinfo"
                    },
                    "timeout": "2m",
                    "wait": true
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "address": "https://github.com/stefanprodan/podinfo",
                    "secretRef": {
                      "name": "api-token"
                    }
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Alert",
                "metadata": {
                    "name": "webapp",
                    "namespace": "default"
                },
                "spec": {
                    "providerRef": {
                      "name": "on-call-slack"
                    },
                    "eventSeverity": "info",
                    "eventSources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "Bucket",
                          "name": "secrets"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-backend"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-frontend"
                      }
                    ]
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Receiver",
                "metadata": {
                    "name": "github-receiver",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "events": [
                      "ping",
                      "push"
                    ],
                    "secretRef": {
                      "name": "webhook-token"
                    },
                    "resources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "HelmRepository",
                          "name": "webapp"
                      }
                    ]
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "image": "ghcr.io/stefanprodan/podinfo",
                    "interval": "1m0s"
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImagePolicy",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "imageRepositoryRef": {
                      "name": "podinfo"
                    },
                    "policy": {
                      "semver": {
                          "range": "5.0.x"
                      }
                    }
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageUpdateAutomation",
                "metadata": {
                  "name": "flux-system",
                  "namespace": "flux-system"
                },
                "spec": {
                  "interval": "1m0s",
                  "sourceRef": {
                    "kind": "GitRepository",
                    "name": "flux-system"
                  },
                  "git": {
                    "checkout": {
                      "ref": {
                        "branch": "main"
                      }
                    },
                    "commit": {
                      "author": {
                        "email": "fluxcdbot@users.noreply.github.com",
                        "name": "fluxcdbot"
                      },
                      "messageTemplate": "{{range .Updated.Images}}{{println .}}{{end}}"
                    },
                    "push": {
                      "branch": "main"
                    }
                  },
                  "update": {
                    "path": "./clusters/my-cluster",
                    "strategy": "Setters"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.connect.redhat.com/flux/source-controller@sha256:b576e3f010e3ff07e716064ab0dad042aba93f0ea5c250535dd18bd678c45337
          description: Flux is a Continuous Delivery solution for Kubernetes.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/flux/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/flux/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: flux-system
          support: The Flux community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta1
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta2
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta2
          - description: OCIRepository
            displayName: OCIRepository
            kind: OCIRepository
            name: ocirepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta2
        description: |-
          [Flux](https://fluxcd.io/) is a Continuous Delivery solution for Kubernetes.
          Flux is constructed with the [GitOps Toolkit](https://toolkit.fluxcd.io/components/), a set of composable APIs and specialized tools for keeping Kubernetes clusters in sync with sources of configuration (like Git & Helm repositories), and automating updates to configuration when there is new code to deploy.

          Flux is a CNCF incubation project made for:

          - **cluster operators** who automate provision and configuration of clusters
          - **platform engineers** who build continuous delivery for developer teams
          - **app developers** who rely on continuous delivery to get their code live

          Features:

          - Source configuration from Git and Helm repositories, and S3-compatible buckets (e.g., Minio)
          - Kustomize and Helm support
          - Event-triggered and periodic reconciliation
          - Integration with Kubernetes RBAC
          - Health assessment (clusters and workloads)
          - Dependency management (infrastructure and workloads)
          - Secrets management (builtin support for Mozilla SOPS and Cloud KMS)
          - Alerting to external systems (webhook senders)
          - External events handling (webhook receivers)
          - Automated container image updates to Git (image scanning and patching)
          - Policy-driven validation (OPA, Kyverno, admission controllers)
          - Seamless integration with Git providers (GitHub, GitLab, Bitbucket)
          - Interoperability with Cluster API (CAPI) providers

          ### Steps to support bootstraping with Flux CLI

          SecComp Profile should be removed as it requires a special privilege to get the access.
          Also, it requires to enforce `runUserAs` to the same UID provided by the images to prevent OpenShift
          to alter the user id expected by our controllers.

          To do so, please specifiy a patch using the following Kustomization file, for example.

          ```yaml
          apiVersion: kustomize.config.k8s.io/v1beta1
          kind: Kustomization
          resources:
            - gotk-components.yaml
            - gotk-sync.yaml
          patches:
            - patch: |
                apiVersion: apps/v1
                kind: Deployment
                metadata:
                  name: all
                spec:
                  template:
                    spec:
                      containers:
                        - name: manager
                          securityContext:
                            runAsUser: 65534
                            seccompProfile:
                              $patch: delete
              target:
                kind: Deployment
                labelSelector: app.kubernetes.io/part-of=flux
          ```

          Then before installing Flux, please also perform the following steps to bind Flux to the `nonroot` SecurityContextConstraint (SCC). This applies to installation via OC's OLM or via Flux CLI.

          ```sh
          # replace default "flux-system" with the project/namespace name for your flux installation
          namespace="flux-system"
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:source-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:kustomize-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:helm-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:notification-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-automation-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-reflector-controller

          ```

          To see complete steps please refer to this [documentation](https://github.com/fluxcd/website/blob/main/content/en/docs/use-cases/openshift.md).

          If you've already installed Flux, you might need to uninstall it after applying the above pre-requisite steps.

          ### Network Policies

          This version of Flux is shipped without the normally provided network Policies.
          By default all traffic is enabled, and anyone can access the source-controller
          artifacts.

          It is not recommended to run it in production as is.

          Please visit [Flux's policies](https://github.com/fluxcd/flux2/tree/main/manifests/policies)
          to deploy the needed network policies and secure your Flux setup.

          ### Breaking Changes

          - Flux 0.18.x ships with v1beta2 of the kustomize.toolkit.fluxcd.io/Kustomization CRD.
          - Flux 0.25.x is the last Flux release where Kubernetes 1.19 is supported.

          ### Flux on OpenShift Use Cases

          - [The GitOps with Flux + OpenShift](https://www.youtube.com/watch?v=8ybCtNVP5To) talk by Andrew Block (Red Hat) showing how to setup multi-tenacy on OpenShift using Flux.
          - [Flux OpenShift Use Case](https://fluxcd.io/docs/use-cases/openshift/) on FluxCD.io
        displayName: Flux
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - gitops
        - cd
        - continuous delivery
        - fluxcd
        - helm
        - kustomize
        links:
        - name: Homepage
          url: https://fluxcd.io/
        - name: Documentation
          url: https://fluxcd.io/docs/
        - name: Slack
          url: https://cloud-native.slack.com/messages/flux
        - name: Mailing List
          url: https://lists.cncf.io/g/cncf-flux-dev
        - name: Source Code
          url: https://github.com/fluxcd/flux2
        maintainers:
        - email: support@weave.works
          name: The Flux maintainers
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: The Flux project
        relatedImages:
        - registry.connect.redhat.com/flux/helm-controller@sha256:cc24a7b73b023ad86a6c3a845088d153ea290dda7cab1ec03eb85aaf814e6049
        - registry.connect.redhat.com/flux/image-automation-controller@sha256:0c51595307e14acc15f22d50b414a87513ebb03530ef4a8f7d6cd9f8b757243e
        - registry.connect.redhat.com/flux/image-reflector-controller@sha256:3bef7fde22be641304e5ca69f77150e416eba5141f8aa13f68be7f810e55aa59
        - registry.connect.redhat.com/flux/kustomize-controller@sha256:246290908d899b9f6feb920f88913bd2d4673cf852a3020714fd4152c7f78237
        - registry.connect.redhat.com/flux/notification-controller@sha256:b3d81234ba3312e0106f2f7cabc73fbe1d7bf683eda36b9ed5bbdaa43789be40
        - registry.connect.redhat.com/flux/source-controller@sha256:b576e3f010e3ff07e716064ab0dad042aba93f0ea5c250535dd18bd678c45337
        version: 2.0.1
      entries:
      - name: flux.v2.0.1
        version: 2.0.1
      name: stable
    defaultChannel: stable
    packageName: flux
    provider:
      name: The Flux project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: telemetry-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: telemetry-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "Autoscaling",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "telemetry-operator",
                    "app.kubernetes.io/instance": "autoscaling-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "autoscaling",
                    "app.kubernetes.io/part-of": "telemetry-operator"
                  },
                  "name": "autoscaling"
                },
                "spec": {
                  "aodh": {
                    "databaseAccount": "aodh",
                    "databaseInstance": "openstack",
                    "memcachedInstance": "memcached",
                    "passwordSelectors": null,
                    "secret": "osp-secret"
                  }
                }
              },
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "Ceilometer",
                "metadata": {
                  "name": "ceilometer",
                  "namespace": "openstack"
                },
                "spec": {
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "Logging",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "telemetry-operator",
                    "app.kubernetes.io/instance": "logging-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "logging",
                    "app.kubernetes.io/part-of": "telemetry-operator"
                  },
                  "name": "logging"
                },
                "spec": {
                  "cloNamespace": "openshift-logging",
                  "port": 10514
                }
              },
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "MetricStorage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "telemetry-operator",
                    "app.kubernetes.io/instance": "metricstorage-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "metricstorage",
                    "app.kubernetes.io/part-of": "telemetry-operator"
                  },
                  "name": "metricstorage-sample"
                },
                "spec": {
                  "monitoringStack": {
                    "alertingEnabled": true,
                    "scrapeInterval": "30s",
                    "storage": {
                      "persistent": {
                        "pvcStorageRequest": "20G"
                      },
                      "retention": "24h",
                      "strategy": "persistent"
                    }
                  }
                }
              },
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "Telemetry",
                "metadata": {
                  "name": "telemetry"
                },
                "spec": {
                  "autoscaling": {
                    "aodh": {
                      "databaseAccount": "aodh",
                      "databaseInstance": "openstack",
                      "memcachedInstance": "memcached",
                      "passwordSelectors": null,
                      "secret": "osp-secret"
                    },
                    "enabled": false,
                    "heatInstance": "heat"
                  },
                  "ceilometer": {
                    "enabled": true,
                    "secret": "osp-secret"
                  },
                  "logging": {
                    "enabled": false,
                    "port": 10514
                  },
                  "metricStorage": {
                    "enabled": false,
                    "monitoringStack": {
                      "alertingEnabled": true,
                      "scrapeInterval": "30s",
                      "storage": {
                        "persistent": {
                          "pvcStorageRequest": "20G"
                        },
                        "retention": "24h",
                        "strategy": "persistent"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:47Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Autoscaling is the Schema for the autoscalings API
            displayName: Autoscaling
            kind: Autoscaling
            name: autoscalings.telemetry.openstack.org
            version: v1beta1
          - description: Ceilometer is the Schema for the ceilometers API
            displayName: Ceilometer
            kind: Ceilometer
            name: ceilometers.telemetry.openstack.org
            version: v1beta1
          - description: Logging is the Schema for the loggings API
            displayName: Logging
            kind: Logging
            name: loggings.telemetry.openstack.org
            version: v1beta1
          - description: MetricStorage is the Schema for the metricstorages API
            displayName: Metric Storage
            kind: MetricStorage
            name: metricstorages.telemetry.openstack.org
            version: v1beta1
          - description: Telemetry is the Schema for the telemetry API
            displayName: Telemetry
            kind: Telemetry
            name: telemetries.telemetry.openstack.org
            version: v1beta1
        description: Telemetry Operator
        displayName: Telemetry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Telemetry
        - Ceilometer
        links:
        - name: Telemetry Operator
          url: https://github.com/openstack-k8s-operators/telemetry-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-ceilometer-notification-rhel9@sha256:239e5376f65682fd644e8d9b10b27ad879747a056c8fd26ff19882e34c5bb056
        - registry.redhat.io/rhoso/openstack-ceilometer-ipmi-rhel9@sha256:86618d8f54df09210349e382188f3b6563ab96ee511fb4da6eb6dc80e4027aae
        - registry.redhat.io/rhoso/openstack-aodh-listener-rhel9@sha256:3599a25515a96b3ff4768f85fecb4bf955789fb9b5d8abc3181678062020ae64
        - registry.redhat.io/rhoso/openstack-ceilometer-compute-rhel9@sha256:b8a0412041c50e312bef39c17a9c2600e9dbb42ea380d55e58243cda641cef4f
        - registry.redhat.io/rhoso/openstack-aodh-evaluator-rhel9@sha256:16027b7d4f07385c87f97e1581d4454d4ce7ed654499720d9ba5bc6fc74fc246
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/telemetry-rhel9-operator@sha256:83962f19193f98b87c54136f02d18797a8047f65e0d75c11ce9518e50aabc050
        - registry.redhat.io/rhoso/openstack-ceilometer-central-rhel9@sha256:1c79d1edb2404a8614ea29f24e2a869695d75b1695ea0937eebc9c57250fc4ab
        - registry.redhat.io/rhoso/openstack-aodh-api-rhel9@sha256:5ab39a468ab014f2fec4cb9cd210b430ae1c1262f52b9e0f22881d6d8e9def97
        - registry.redhat.io/rhoso/openstack-aodh-notifier-rhel9@sha256:f729548ac3d718f2f5b6c85045e7c676e1656207ee54f43540d246bd97426894
        - registry.redhat.io/rhoso-operators/sg-core-rhel9@sha256:a6e4edc24409361cb770bb44fe5b475dc58748426ccc4003abad7e70f01f1b97
        - registry.redhat.io/ubi9/httpd-24@sha256:dbfbe3383e12d63df7178e4cdd5b1e6e51a84c781041b9192026ea6edb72fbf3
        version: 1.0.6
      entries:
      - name: telemetry-operator.v1.0.6
        version: 1.0.6
      - name: telemetry-operator.v1.0.5
        version: 1.0.5
      - name: telemetry-operator.v1.0.4
        version: 1.0.4
      - name: telemetry-operator.v1.0.3
        version: 1.0.3
      - name: telemetry-operator.v1.0.2
        version: 1.0.2
      - name: telemetry-operator.v1.0.1
        version: 1.0.1
      - name: telemetry-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: telemetry-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Snyk Ltd.
      provider-url: ""
    name: snyk-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: snyk-operator.v1.92.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1alpha1",
                "kind": "SnykMonitor",
                "metadata": {
                  "name": "snyk-monitor"
                },
                "spec": {
                  "clusterName": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "snyk/kubernetes-monitor",
                    "tag": "1.92.0-ubi8"
                  },
                  "integrationApi": "",
                  "monitorSecrets": "snyk-monitor",
                  "scope": "Cluster",
                  "temporaryStorageSize": "50Gi",
                  "pvc": {
                    "enabled": false,
                    "create": false,
                    "name": "snyk-monitor-pvc",
                    "storageClassName": null
                  },
                  "initContainerImage": {
                    "repository": "busybox",
                    "tag": "latest"
                  },
                  "requests": {
                    "memory": "400Mi"
                  },
                  "limits": {
                    "memory": "2Gi"
                  },
                  "nodeAffinity": {
                    "disableBetaArchNodeSelector": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Security
          certified: "false"
          containerImage: docker.io/snyk/kubernetes-operator:1.92.0
          createdAt: "2022-06-09T17:49:36Z"
          description: A Kubernetes Operator for creating and managing Snyk controller
            instances.
          olm.skipRange: <1.68.2
          repository: https://github.com/snyk/kubernetes-monitor
          support: Snyk Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a deployment of the Snyk controller, which scans
              container images in the cluster for vulnerabilities.
            displayName: Snyk Monitor
            kind: SnykMonitor
            name: snykmonitors.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          A Kubernetes Operator for creating and managing Snyk Kubernetes controller instances.

          Snyk integrates with Kubernetes, enabling you to import and test your running workloads and identify vulnerabilities in their associated images and configurations that might make those workloads less secure. Once imported, Snyk continues to monitor those workloads, identifying additional security issues as new images are deployed and the workload configuration changes.

          ## Prerequisites

          To start using the Operator and the Snyk controller that it manages, you will need to create a secret containing your Snyk integrationId and Docker config file:

          ```
          kubectl create namespace snyk-monitor
          kubectl create secret generic snyk-monitor -n snyk-monitor --from-literal=dockercfg.json={} --from-literal=integrationId=${SNYK_INTEGRATION_ID}
          ```

          Refer to the [Snyk documentation](https://support.snyk.io/hc/en-us/articles/360003916138-Kubernetes-integration-overview) for more details.
        displayName: Snyk Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - snyk
        - controller
        - monitor
        - operator
        - security
        - cluster
        - container
        - image
        - security
        - vulnerabilities
        - advisories
        - audit
        - scan
        - scanning
        links:
        - name: Website
          url: https://snyk.io
        - name: Documentation
          url: https://support.snyk.io/hc/en-us/articles/360003916138-Kubernetes-integration-overview
        maintainers:
        - email: support@snyk.io
          name: Snyk Ltd.
        maturity: stable
        provider:
          name: Snyk Ltd.
        relatedImages:
        - docker.io/snyk/kubernetes-operator:1.92.0
        version: 1.92.0
      entries:
      - name: snyk-operator.v1.92.0
        version: 1.92.0
      - name: snyk-operator.v1.90.2
        version: 1.90.2
      - name: snyk-operator.v1.90.1
        version: 1.90.1
      - name: snyk-operator.v1.88.4
        version: 1.88.4
      - name: snyk-operator.v1.85.2
        version: 1.85.2
      - name: snyk-operator.v1.82.1
        version: 1.82.1
      - name: snyk-operator.v1.81.1
        version: 1.81.1
      - name: snyk-operator.v1.80.0
        version: 1.80.0
      - name: snyk-operator.v1.73.0
        version: 1.73.0
      - name: snyk-operator.v1.72.0
        version: 1.72.0
      - name: snyk-operator.v1.70.4
        version: 1.70.4
      - name: snyk-operator.v1.70.1
        version: 1.70.1
      name: stable
    defaultChannel: stable
    packageName: snyk-operator
    provider:
      name: Snyk Ltd.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Aqua Security, Inc.
      provider-url: ""
    name: aqua
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aqua-operator.v2022.4.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaCsp",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "namespace": "aqua",
                    "version": "2022.4",
                    "requirements": true
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry",
                    "dbDiskSize": 10,
                    "databaseSecret": {
                      "key": "db-password",
                      "name": "aqua-database-password"
                    }
                  },
                  "database": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "database",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "gateway": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "gateway",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "server": {
                    "replicas": 1,
                    "service": "LoadBalancer",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "console",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "route": true,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaDatabase",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "database",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "diskSize": 10,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "image": {
                      "repository": "enforcer",
                      "registry": "registry.aquasec.com",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "IfNotPresent"
                    }
                  },
                  "runAsNonRoot": false,
                  "gateway": {
                    "host": "aqua-gateway",
                    "port": 8443
                  },
                  "token": "<<your-token>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaKubeEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "config": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "aqua-secure",
                    "imagePullSecret": "aqua-registry",
                    "kubebench_image": "docker.io/aquasec/kube-bench:v0.7.1"
                  },
                  "deploy": {
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "tag": "<<KUBE_ENFORCER_TAG>>",
                      "repository": "kube-enforcer",
                      "pullPolicy": "Always"
                    }
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "starboard-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  }
                }
              },
              {
              "apiVersion": "operator.aquasec.com/v1alpha1",
              "kind": "AquaLightning",
              "metadata": {
                "name": "aqua"
              },
              "spec": {
                "global": {
                  "gateway_address": "aqua-gateway.aqua:8443",
                  "cluster_name": "Default-cluster-name"
                },
                "common": {
                  "imagePullSecret": "aqua-registry"
                },
                "kubeEnforcer": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                  "allowAnyVersion": null,
                  "deploy": {
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "tag": "2022.4",
                      "repository": "kube-enforcer",
                      "pullPolicy": "Always"
                    }
                  },
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "starboard-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  },
                  "env": []
                },
                "enforcer": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "deploy": {
                    "image": {
                      "repository": "enforcer",
                      "registry": "registry.aquasec.com",
                      "tag": "2022.4",
                      "pullPolicy": "IfNotPresent"
                    }
                  },
                  "token": "<<ENFORCER_GROUP_TOKEN>>",
                  "env": [],
                  "secret": {
                    "name": null,
                    "key": null
                  },
                  "runAsNonRoot": null,
                  "rhcosVersion": null
                }
              }
            },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaGateway",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "gateway",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaScanner",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "replicas": 1,
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "scanner",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "runAsNonRoot": false,
                  "login": {
                    "username": "<<YOUR AQUA USER NAME>>",
                    "password": "<<YOUR AQUA USER PASSWORD>>",
                    "host": "http://aqua-server:8080",
                    "token": ""
                  }
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaServer",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "LoadBalancer",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "console",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "runAsNonRoot": false
                }
              },
              {
                  "apiVersion": "operator.aquasec.com/v1alpha1",
                  "kind": "AquaCloudConnector",
                  "metadata": {
                    "name": "aqua"
                  },
                  "spec": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "common": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "image": {
                        "repository": "aqua-cloud-connector",
                        "registry": "registry.aquasec.com",
                        "tag": "<<IMAGE TAG>>"
                      },
                      "replicas": 1
                    },
                    "runAsNonRoot": false,
                    "gateway": {
                      "host": "aqua-gateway",
                      "port": "8443"
                    },
                    "login": {
                      "host": "aqua-gateway:8443",
                      "password": "<<YOUR AQUA USER PASSWORD>>",
                      "token": "<<YOUR AQUA CLOUD CONNECTOR TOKEN>>",
                      "username": "<<YOUR AQUA USER NAME>>"
                    },
                    "tunnels": [
                      {
                        "host": "<account>.azurecr.io",
                        "port": "443"
                      },
                      {
                        "host": "gke.gcr.io",
                        "port": "443"
                      },
                      {
                        "host": "<account>.jfrog.io",
                        "port": "443"
                      },
                      {
                        "host": "1.1.1.1",
                        "port": "443"
                      },
                      {
                        "region": "us-east1",
                        "type": "ecr"
                      }
                    ]
                  }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: docker.io/aquasec/aqua-operator:2022.4.18
          createdAt: "2020-03-25T08:00:00Z"
          description: The Aqua Security Operator runs within a Openshift cluster
            and provides a means to deploy and manage Aqua Security cluster and components.
          operators.operatorframework.io/internal-objects: '["aquastarboards.aquasecurity.github.io","configauditreports.aquasecurity.github.io",
            "clusterconfigauditreports.aquasecurity.github.io"]'
          repository: https://github.com/aquasecurity/aqua-operator
          support: Aqua Security, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aqua Security CSP Deployment with Aqua Operator
            displayName: AquaCsp
            kind: AquaCsp
            name: aquacsps.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Lightning Enforcers Deployment with Aqua Operator
            displayName: AquaLightningEnforcers
            kind: AquaLightning
            name: aqualightnings.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Database Deployment with Aqua Operator
            displayName: AquaDatabase
            kind: AquaDatabase
            name: aquadatabases.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Enforcer Deployment with Aqua Operator
            displayName: AquaEnforcer
            kind: AquaEnforcer
            name: aquaenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security KubeEnforcer Deployment with Aqua Operator
            displayName: AquaKubeEnforcer
            kind: AquaKubeEnforcer
            name: aquakubeenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Gateway Deployment with Aqua Operator
            displayName: AquaGateway
            kind: AquaGateway
            name: aquagateways.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Scanner Deployment with Aqua Operator
            displayName: AquaScanner
            kind: AquaScanner
            name: aquascanners.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Server Deployment with Aqua Operator
            displayName: AquaServer
            kind: AquaServer
            name: aquaservers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Starboard for kube-enforcer.
            displayName: AquaStarboard
            kind: AquaStarboard
            name: aquastarboards.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ClusterConfigAuditReport
            kind: ClusterConfigAuditReport
            name: clusterconfigauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ConfigAuditReport
            kind: ConfigAuditReport
            name: configauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: Aqua Security Cloud Connector Deployment with Aqua Operator
            displayName: AquaCloudConnector
            kind: AquaCloudConnector
            name: aquacloudconnector.operator.aquasec.com
            version: v1alpha1
        description: |-
          The Aqua Security Operator runs within an OpenShift cluster and provides a means to deploy and manage the Aqua Security cluster and components :
          * Server (aka “console”)
          * Database (for production environments we recommend to use an external database and not the Aqua default database)
          * Gateway
          * Enforcer (aka “agent”)
          * KubeEnforcer
          * Scanner
          * CSP (package that contains the Server, Database, and Gateway)
          * LightningEnforcer (package that contains Enforcer and Kube Enforcer)

          Use the Aqua-Operator to
          * Deploy Aqua Security components on OpenShift
          * Scale up Aqua Security components with extra replicas
          * Assign metadata tags to Aqua Security components

          ## Before You Begin Using the Operator CRDs

          Obtain access to the Aqua registry - https://www.aquasec.com/about-us/contact-us/


          ## Instructions and exmaples

          Follow instructions and exmaples
          [here](https://github.com/aquasecurity/aqua-operator/blob/2022.4.0/docs/DeployOpenShiftOperator.md)
        displayName: Aqua Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aqua-security
        - scanning
        - security
        - runtime-security
        links:
        - name: Aqua Security
          url: https://www.aquasec.com/
        - name: Aqua Operator Github
          url: https://github.com/aquasecurity/aqua-operator
        maintainers:
        - email: support@aquasec.com
          name: Aqua Support, Aqua Security
        maturity: alpha
        provider:
          name: Aqua Security, Inc.
        relatedImages:
        - docker.io/aquasec/aqua-operator:2022.4.18
        version: 2022.4.18
      entries:
      - name: aqua-operator.v2022.4.18
        version: 2022.4.18
      - name: aqua-operator.v2022.4.17
        version: 2022.4.17
      name: 2022.4.0
    defaultChannel: 2022.4.0
    packageName: aqua
    provider:
      name: Aqua Security, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: The bpfman Community
      provider-url: https://bpfman.io/
    name: bpfman-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bpfman-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfapplication"
                  },
                  "name": "bpfapplication-sample"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:v0.5.6"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "kprobe": {
                        "bpffunctionname": "kprobe_counter",
                        "func_name": "try_to_wake_up",
                        "offset": 0,
                        "retprobe": false
                      },
                      "type": "Kprobe"
                    },
                    {
                      "tracepoint": {
                        "bpffunctionname": "tracepoint_kill_recorder",
                        "names": [
                          "syscalls/sys_enter_kill"
                        ]
                      },
                      "type": "Tracepoint"
                    },
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "containernames": [
                            "bpfman",
                            "bpfman-agent"
                          ],
                          "namespace": "bpfman",
                          "pods": {
                            "matchLabels": {
                              "name": "bpfman-daemon"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfNsApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfnsapplication"
                  },
                  "name": "bpfapplication-sample",
                  "namespace": "acme"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:v0.5.6"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FentryProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fentryprogram"
                  },
                  "name": "fentry-example"
                },
                "spec": {
                  "bpffunctionname": "test_fentry",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fentry:v0.5.6"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FexitProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fexitprogram"
                  },
                  "name": "fexit-example"
                },
                "spec": {
                  "bpffunctionname": "test_fexit",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fexit:v0.5.6"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "KprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kprobeprogram"
                  },
                  "name": "kprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_kprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/kprobe:v0.5.6"
                    }
                  },
                  "func_name": "try_to_wake_up",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "offset": 0,
                  "retprobe": false
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcnsprogram"
                  },
                  "name": "tc-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcprogram"
                  },
                  "name": "tc-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:v0.5.6"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxnsprogram"
                  },
                  "name": "tcx-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "tcx_next",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxprogram"
                  },
                  "name": "tcx-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "tcx_pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:v0.5.6"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TracepointProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tracepointprogram"
                  },
                  "name": "tracepoint-example"
                },
                "spec": {
                  "bpffunctionname": "enter_openat",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tracepoint:v0.5.6"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "names": [
                    "syscalls/sys_enter_openat"
                  ],
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobensprogram"
                  },
                  "name": "uprobe-example",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobeprogram"
                  },
                  "name": "uprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:v0.5.6"
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpnsprogram"
                  },
                  "name": "xdp-ns-pass-all-nodes",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpprogram"
                  },
                  "name": "xdp-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:v0.5.6"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/bpfman/bpfman-operator:v0.5.6
          createdAt: "2025-02-18T13:44:53Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: bpfman
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "bpfman",
                "labels": {
                  "pod-security.kubernetes.io/enforce": "privileged",
                  "pod-security.kubernetes.io/audit": "privileged",
                  "pod-security.kubernetes.io/warn": "privileged",
                },
                "annotations": {
                  "openshift.io/node-selector": ""
                },
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/bpfman/bpfman
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BpfApplication is the Schema for the bpfapplications API
            displayName: Bpf Application
            kind: BpfApplication
            name: bpfapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsApplication is the Schema for the BpfNsApplications
              API
            displayName: Bpf Namespaced Application
            kind: BpfNsApplication
            name: bpfnsapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsProgram is the Schema for the BpfNsProgram API
            displayName: Bpf Namespaced Program
            kind: BpfNsProgram
            name: bpfnsprograms.bpfman.io
            version: v1alpha1
          - description: BpfProgram is the Schema for the BpfProgram API
            displayName: Bpf Program
            kind: BpfProgram
            name: bpfprograms.bpfman.io
            version: v1alpha1
          - description: FentryProgram is the Schema for the Fentryprograms API
            displayName: Fentry Program
            kind: FentryProgram
            name: fentryprograms.bpfman.io
            version: v1alpha1
          - description: FexitProgram is the Schema for the Fexitprograms API
            displayName: Fexit Program
            kind: FexitProgram
            name: fexitprograms.bpfman.io
            version: v1alpha1
          - description: KprobeProgram is the Schema for the Kprobeprograms API
            displayName: Kprobe Program
            kind: KprobeProgram
            name: kprobeprograms.bpfman.io
            version: v1alpha1
          - description: TcNsProgram is the Schema for the Tcnsprograms API
            displayName: TC Namespaced Program
            kind: TcNsProgram
            name: tcnsprograms.bpfman.io
            version: v1alpha1
          - description: TcProgram is the Schema for the Tcprograms API
            displayName: TC Program
            kind: TcProgram
            name: tcprograms.bpfman.io
            version: v1alpha1
          - description: TcxNsProgram is the Schema for the Tcxnsprograms API
            displayName: TCX Namespaced Program
            kind: TcxNsProgram
            name: tcxnsprograms.bpfman.io
            version: v1alpha1
          - description: TcxProgram is the Schema for the Tcxprograms API
            displayName: TCX Program
            kind: TcxProgram
            name: tcxprograms.bpfman.io
            version: v1alpha1
          - description: TracepointProgram is the Schema for the Tracepointprograms
              API
            displayName: Tracepoint Program
            kind: TracepointProgram
            name: tracepointprograms.bpfman.io
            version: v1alpha1
          - description: UprobeNsProgram is the Schema for the Uprobensprograms API
            displayName: Uprobe Namespaced Program
            kind: UprobeNsProgram
            name: uprobensprograms.bpfman.io
            version: v1alpha1
          - description: UprobeProgram is the Schema for the Uprobeprograms API
            displayName: Uprobe Program
            kind: UprobeProgram
            name: uprobeprograms.bpfman.io
            version: v1alpha1
          - description: XdpNsProgram is the Schema for the Xdpnsprograms API
            displayName: Xdp Namespaced Program
            kind: XdpNsProgram
            name: xdpnsprograms.bpfman.io
            version: v1alpha1
          - description: XdpProgram is the Schema for the Xdpprograms API
            displayName: Xdp Program
            kind: XdpProgram
            name: xdpprograms.bpfman.io
            version: v1alpha1
        description: "The bpfman Operator is a Kubernetes Operator for deploying [bpfman](https://bpfman.netlify.app/),
          a system daemon\nfor managing eBPF programs. It deploys bpfman itself along
          with CRDs to make deploying\neBPF programs in Kubernetes much easier.\n\n##
          Quick Start\n\nTo get bpfman up and running quickly simply click 'install'
          to deploy the bpfman-operator in the bpfman namespace via operator-hub.\n##
          Configuration\n\nThe `bpfman-config` configmap is automatically created
          in the `bpfman` namespace and used to configure the bpfman deployment.\n\nTo
          edit the config simply run\n\n```bash\nkubectl edit cm bpfman-config\n```\n\nThe
          following fields are adjustable\n\n- `bpfman.agent.image`: The image used
          for the bpfman-agent`\n- `bpfman.image`: The image used for bpfman`\n- `bpfman.log.level`:
          the log level for bpfman, currently supports `debug`, `info`, `warn`, `error`,
          and `fatal`, defaults to `info`\n- `bpfman.agent.log.level`: the log level
          for the bpfman-agent currently supports `info`, `debug`, and `trace` \n\nThe
          bpfman operator deploys eBPF programs via CRDs. The following CRDs are currently
          available, \n\n- BpfApplication\n- FentryProgram\n- FexitProgram\n - KprobeProgram\n-
          TcProgram\n- TcxProgram\n- TracepointProgram\n- UprobeProgram\n- XdpProgram\n
          - BpfNsApplication\n- TcNsProgram\n- TcxNsProgram\n- UprobeNsProgram\n-
          XdpNsProgram\n- BpfProgram\n - BpfNsProgram\n\n ## More information\n\nPlease
          checkout the [bpfman community website](https://bpfman.io/) for more information."
        displayName: Bpfman Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ebpf
        - kubernetes
        links:
        - name: bpfman website
          url: https://bpfman.io/
        maintainers:
        - email: astoycos@redhat.com
          name: Andrew Stoycos
        - email: afredette@redhat.com
          name: Andre Fredette
        maturity: alpha
        provider:
          name: The bpfman Community
          url: https://bpfman.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/bpfman/bpfman-operator:v0.5.6
        version: 0.5.6
      entries:
      - name: bpfman-operator.v0.5.6
        version: 0.5.6
      - name: bpfman-operator.v0.5.5
        version: 0.5.5
      - name: bpfman-operator.v0.5.4
        version: 0.5.4
      - name: bpfman-operator.v0.5.2
        version: 0.5.2
      - name: bpfman-operator.v0.5.1
        version: 0.5.1
      - name: bpfman-operator.v0.5.0
        version: 0.5.0
      - name: bpfman-operator.v0.4.2
        version: 0.4.2
      - name: bpfman-operator.v0.4.1
        version: 0.4.1
      name: alpha
    defaultChannel: alpha
    packageName: bpfman-operator
    provider:
      name: The bpfman Community
      url: https://bpfman.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-organizations-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-organizations-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "organizations.services.k8s.aws/v1alpha1",
                "kind": "OrganizationalUnit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/organizations-controller:1.1.1
          createdAt: "2025-09-19T19:35:20Z"
          description: AWS Organizations controller is a service controller for managing
            Organizations resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Account represents the state of an AWS organizations Account
              resource.
            displayName: Account
            kind: Account
            name: accounts.organizations.services.k8s.aws
            version: v1alpha1
          - description: OrganizationalUnit represents the state of an AWS organizations
              OrganizationalUnit resource.
            displayName: OrganizationalUnit
            kind: OrganizationalUnit
            name: organizationalunits.organizations.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Organizations resources in AWS from within your Kubernetes cluster.

          **About Amazon Organizations**

          With AWS Organizations you can perform account management activities at scale by consolidating multiple AWS accounts into a single organization. Consolidating accounts simplifies how you use other AWS services. You can leverage the multi-account management services available in AWS Organizations with select AWS services to perform tasks on all accounts that are members of your organization.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Organizations
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - organizations
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Organizations Developer Resources
          url: https://aws.amazon.com/organizations/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: organizations maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/organizations-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-organizations-controller.v1.1.1
        version: 1.1.1
      - name: ack-organizations-controller.v1.0.13
        version: 1.0.13
      - name: ack-organizations-controller.v1.0.11
        version: 1.0.11
      - name: ack-organizations-controller.v1.0.10
        version: 1.0.10
      - name: ack-organizations-controller.v1.0.9
        version: 1.0.9
      - name: ack-organizations-controller.v1.0.8
        version: 1.0.8
      - name: ack-organizations-controller.v1.0.7
        version: 1.0.7
      - name: ack-organizations-controller.v1.0.6
        version: 1.0.6
      - name: ack-organizations-controller.v1.0.5
        version: 1.0.5
      - name: ack-organizations-controller.v1.0.4
        version: 1.0.4
      - name: ack-organizations-controller.v1.0.1
        version: 1.0.1
      - name: ack-organizations-controller.v1.0.0
        version: 1.0.0
      - name: ack-organizations-controller.v0.0.10
        version: 0.0.10
      - name: ack-organizations-controller.v0.0.9
        version: 0.0.9
      - name: ack-organizations-controller.v0.0.8
        version: 0.0.8
      - name: ack-organizations-controller.v0.0.7
        version: 0.0.7
      - name: ack-organizations-controller.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: ack-organizations-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-sds-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-sds-operator.v8.0.230
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.230"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleStoragePool",
                "metadata": {
                  "name": "infoscalestoragepool-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:f4947eab2b31fccc86b5d5d0d3fd4eb1f7513fd8325e379cc259f5ccd1dfc3e4
          createdAt: "2023-10-26T20:36:59Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          olm.skipRange: <8.0.230
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleStoragePool is the Schema for the infoscalestoragepools
              API
            displayName: InfoScale Storage Pool
            kind: InfoScaleStoragePool
            name: infoscalestoragepools.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.230"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:f4947eab2b31fccc86b5d5d0d3fd4eb1f7513fd8325e379cc259f5ccd1dfc3e4
        version: 8.0.230
      entries:
      - name: infoscale-sds-operator.v8.0.230
        version: 8.0.230
      - name: infoscale-sds-operator.v8.0.220
        version: 8.0.220
      - name: infoscale-sds-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-sds-operator.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-sds-operator.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.330"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:dab501b99373c840c7e8e3ab479fa273b39f9bf8a896f3f1e4e945f4ab81f64e
          createdAt: "2024-06-24T09:03:32Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.330"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:dab501b99373c840c7e8e3ab479fa273b39f9bf8a896f3f1e4e945f4ab81f64e
        version: 8.0.330
      entries:
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-sds-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.410"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
          createdAt: "2024-07-30T09:00:57Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Arctera InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/168721626-168743109-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.410"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Arctera recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
        version: 8.0.410
      entries:
      - name: infoscale-sds-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-sds-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-sds-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-sds-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.410"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
          createdAt: "2024-07-30T09:00:57Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Arctera InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/168721626-168743109-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.410"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Arctera recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
        version: 8.0.410
      entries:
      - name: infoscale-sds-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-sds-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-sds-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-sds-operator.v8.0.210
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.210"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:26573b98743245c0b607abbe0d981abce6ce9fb1c5c0cd9fe735351b7136f4e8
          createdAt: "2022-11-26T16:36:59Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleStoragePool is the Schema for the infoscalestoragepools
              API
            displayName: InfoScale Storage Pool
            kind: InfoScaleStoragePool
            name: infoscalestoragepools.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.210"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:1667227b5111fe52ed304ab19ff9fadfc677239e56b2c93a3b89b1bfbc31354d
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:26573b98743245c0b607abbe0d981abce6ce9fb1c5c0cd9fe735351b7136f4e8
        version: 8.0.210
      entries:
      - name: infoscale-sds-operator.v8.0.210
        version: 8.0.210
      - name: infoscale-sds-operator.v8.0.201
        version: 8.0.201
      - name: infoscale-sds-operator.v8.0.200
        version: 8.0.200
      - name: infoscale-sds-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-sds-operator.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-sds-operator
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: ovn-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ovn-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ovn.openstack.org/v1beta1",
                "kind": "OVNController",
                "metadata": {
                  "name": "ovncontroller-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "ovn.openstack.org/v1beta1",
                "kind": "OVNDBCluster",
                "metadata": {
                  "name": "ovndbcluster-nb-sample"
                },
                "spec": {
                  "dbType": "NB",
                  "logLevel": "info",
                  "storageClass": "local-storage",
                  "storageRequest": "10G"
                }
              },
              {
                "apiVersion": "ovn.openstack.org/v1beta1",
                "kind": "OVNDBCluster",
                "metadata": {
                  "name": "ovndbcluster-sb-sample"
                },
                "spec": {
                  "dbType": "SB",
                  "logLevel": "info",
                  "storageClass": "local-storage",
                  "storageRequest": "10G"
                }
              },
              {
                "apiVersion": "ovn.openstack.org/v1beta1",
                "kind": "OVNNorthd",
                "metadata": {
                  "name": "ovnnorthd-sample"
                },
                "spec": {
                  "logLevel": "info"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:45Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OVNController is the Schema for the ovncontrollers API
            displayName: OVNController
            kind: OVNController
            name: ovncontrollers.ovn.openstack.org
            version: v1beta1
          - description: OVNDBCluster is the Schema for the ovndbclusters API
            displayName: OVNDBCluster
            kind: OVNDBCluster
            name: ovndbclusters.ovn.openstack.org
            version: v1beta1
          - description: OVNNorthd is the Schema for the ovnnorthds API
            displayName: OVNNorthd
            kind: OVNNorthd
            name: ovnnorthds.ovn.openstack.org
            version: v1beta1
        description: OVN Operator
        displayName: OVN Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Network
        - Ovn
        links:
        - name: Ovn Operator
          url: https://github.com/openstack-k8s-operators/ovn-operator
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/ovn-rhel9-operator@sha256:bce19011b3623c66a59efe6ea874e2403440813bdc970cd6170853a8093076f3
        - registry.redhat.io/rhoso/openstack-ovn-base-rhel9@sha256:b0d424f6ea6de8192382541a95aa56fa32b9b95e8d433c41c4a6dadbe4127fae
        - registry.redhat.io/rhoso/openstack-ovn-nb-db-server-rhel9@sha256:e71c8aac00a78e3d786332aaada3657276f69483df1e441b70589601ca038a5b
        - registry.redhat.io/rhoso/openstack-ovn-sb-db-server-rhel9@sha256:7aa2832f80846d5ac71733a1f186d39dce3654ebb661d32b8e4580d92be8e338
        - registry.redhat.io/rhoso/openstack-ovn-northd-rhel9@sha256:f1e4ce04b67d888fa950b83d1f00459c67c602397336da81de3ff71d4052a5d8
        - registry.redhat.io/rhoso/openstack-ovn-controller-rhel9@sha256:6c7827dbe25c168d28250f197d5ea6f3f60cfbab4c974267ec5ea0f8d631c3f5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: ovn-operator.v1.0.6
        version: 1.0.6
      - name: ovn-operator.v1.0.5
        version: 1.0.5
      - name: ovn-operator.v1.0.4
        version: 1.0.4
      - name: ovn-operator.v1.0.3
        version: 1.0.3
      - name: ovn-operator.v1.0.2
        version: 1.0.2
      - name: ovn-operator.v1.0.1
        version: 1.0.1
      - name: ovn-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: ovn-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: fuse-apicurito
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-apicurito.v7.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1alpha1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:8d19ad579f874d6b9a08f127c8900906913a5a7c1e69cbe4b5cb0333b8d86152
          createdAt: 2022-02-28 09:35:38 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.10.3'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:8d19ad579f874d6b9a08f127c8900906913a5a7c1e69cbe4b5cb0333b8d86152
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:8dbeeb8b04c81f41d896f71db90813de57e67493fa85389b63ac23f390785043
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:c02c218ccfbb385ea29af98409995a3d9c26cb0b4c76caf2e0779fa233b5ff8c
        version: 7.10.3
      entries:
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.10.x
    - currentCSV: fuse-apicurito.v7.11.1-0.1687357785.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:2f8d4450187839f0e0c761e5ddcf61b03f5996acfe3d75b60120e76e5b0167c1
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.11.2'
          olm.substitutesFor: fuse-apicurito.v7.11.1-0.1682509045.p
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:2f8d4450187839f0e0c761e5ddcf61b03f5996acfe3d75b60120e76e5b0167c1
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:df61aafc4023b9068a8636c35eabb0d5fda400f523b64cb1fc5d8c480272c435
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:4c493a2d690d0d7c1684c53d15c06b47d691aeffdde6dd471b6e75fed2fec6a9
        version: 7.11.1+0.1687357785.p
      entries:
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.11.x
    - currentCSV: fuse-apicurito.v7.12.1-0.1706187670.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:7ccefee63165e7eaf9a609e8a9f978a06f27deff69a318fe399a29292ad99106
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.12.1'
          olm.substitutesFor: fuse-apicurito.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:7ccefee63165e7eaf9a609e8a9f978a06f27deff69a318fe399a29292ad99106
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:4cd612283e8136839f1f48a3a220dac7c6c38adf14b61bc03a981e514baefe9b
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:bee0eeaa7898af835f80c5217059d73361daa42197fa6cbbaf416e1ba373eb54
        version: 7.12.1+0.1706187670.p
      entries:
      - name: fuse-apicurito.v7.12.1-0.1706187670.p
        version: 7.12.1+0.1706187670.p
      - name: fuse-apicurito.v7.12.1
        version: 7.12.1
      - name: fuse-apicurito.v7.12.0
        version: 7.12.0
      - name: fuse-apicurito.v7.12.0-0.1697748407.p
        version: 7.12.0+0.1697748407.p
      - name: fuse-apicurito.v7.12.0-0.1697419531.p
        version: 7.12.0+0.1697419531.p
      - name: fuse-apicurito.v7.12.0-0.1694087055.p
        version: 7.12.0+0.1694087055.p
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.12.x
    - currentCSV: fuse-apicurito.v7.13.0-0.1753785079.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:d58d345d74059242d57d5c1a75f31517769230d3014a42243fda20aa93ebd5c8
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.7.0 <7.13.0'
          olm.substitutesFor: fuse-apicurito.v7.13.0-0.1751291353.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:d58d345d74059242d57d5c1a75f31517769230d3014a42243fda20aa93ebd5c8
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:f5bf9240b28b3664bdb11b1f57f611e6194ff8643ae939b6ab3b2b176705f487
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:4e64a5930a10b76147ac128a216cd7c0346da2ff3e9f31ed86311941ed193cc1
        version: 7.13.0+0.1753785079.p
      entries:
      - name: fuse-apicurito.v7.13.0-0.1739885035.p
        version: 7.13.0+0.1739885035.p
      - name: fuse-apicurito.v7.13.0-0.1741884705.p
        version: 7.13.0+0.1741884705.p
      - name: fuse-apicurito.v7.13.0-0.1753785079.p
        version: 7.13.0+0.1753785079.p
      - name: fuse-apicurito.v7.13.0-0.1721224330.p
        version: 7.13.0+0.1721224330.p
      - name: fuse-apicurito.v7.13.0-0.1753778625.p
        version: 7.13.0+0.1753778625.p
      - name: fuse-apicurito.v7.13.0-0.1753702605.p
        version: 7.13.0+0.1753702605.p
      - name: fuse-apicurito.v7.13.0-0.1721607497.p
        version: 7.13.0+0.1721607497.p
      - name: fuse-apicurito.v7.13.0-0.1751291353.p
        version: 7.13.0+0.1751291353.p
      - name: fuse-apicurito.v7.13.0-0.1748960166.p
        version: 7.13.0+0.1748960166.p
      - name: fuse-apicurito.v7.13.0-0.1745523192.p
        version: 7.13.0+0.1745523192.p
      - name: fuse-apicurito.v7.13.0-0.1744737027.p
        version: 7.13.0+0.1744737027.p
      - name: fuse-apicurito.v7.13.0-0.1744280419.p
        version: 7.13.0+0.1744280419.p
      - name: fuse-apicurito.v7.13.0-0.1741875975.p
        version: 7.13.0+0.1741875975.p
      - name: fuse-apicurito.v7.13.0-0.1739473863.p
        version: 7.13.0+0.1739473863.p
      - name: fuse-apicurito.v7.13.0-0.1738056248.p
        version: 7.13.0+0.1738056248.p
      - name: fuse-apicurito.v7.13.0-0.1731460815.p
        version: 7.13.0+0.1731460815.p
      - name: fuse-apicurito.v7.13.0-0.1729189341.p
        version: 7.13.0+0.1729189341.p
      - name: fuse-apicurito.v7.13.0-0.1727957847.p
        version: 7.13.0+0.1727957847.p
      - name: fuse-apicurito.v7.13.0-0.1727193594.p
        version: 7.13.0+0.1727193594.p
      - name: fuse-apicurito.v7.13.0-0.1725022690.p
        version: 7.13.0+0.1725022690.p
      - name: fuse-apicurito.v7.13.0-0.1724659732.p
        version: 7.13.0+0.1724659732.p
      - name: fuse-apicurito.v7.13.0-0.1724053718.p
        version: 7.13.0+0.1724053718.p
      - name: fuse-apicurito.v7.13.0-0.1723709434.p
        version: 7.13.0+0.1723709434.p
      - name: fuse-apicurito.v7.13.0
        version: 7.13.0
      - name: fuse-apicurito.v7.13.0-0.1720707249.p
        version: 7.13.0+0.1720707249.p
      - name: fuse-apicurito.v7.13.0-0.1718285973.p
        version: 7.13.0+0.1718285973.p
      - name: fuse-apicurito.v7.13.0-0.1717672390.p
        version: 7.13.0+0.1717672390.p
      - name: fuse-apicurito.v7.13.0-0.1753778936.p
        version: 7.13.0+0.1753778936.p
      - name: fuse-apicurito.v7.13.0-0.1753701210.p
        version: 7.13.0+0.1753701210.p
      - name: fuse-apicurito.v7.12.1-0.1706187670.p
        version: 7.12.1+0.1706187670.p
      - name: fuse-apicurito.v7.12.1
        version: 7.12.1
      - name: fuse-apicurito.v7.12.0-0.1697419531.p
        version: 7.12.0+0.1697419531.p
      - name: fuse-apicurito.v7.12.0-0.1697748407.p
        version: 7.12.0+0.1697748407.p
      - name: fuse-apicurito.v7.12.0-0.1694087055.p
        version: 7.12.0+0.1694087055.p
      - name: fuse-apicurito.v7.12.0
        version: 7.12.0
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.13.x
    - currentCSV: fuse-apicurito.v7.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
                    "apiVersion": "apicur.io/v1alpha1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
            }]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:05c9676aada2d34a985a59abcf55105a3765d75b0d3764089d2b32e9ca5d367c
          createdAt: "2020-10-06 12:48:47"
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.9.3'
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito Project
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CRD for Apicurito
            displayName: Apicurito CRD
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:05c9676aada2d34a985a59abcf55105a3765d75b0d3764089d2b32e9ca5d367c
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:5535864ce393ae0ae5ee52ed0c7661e0351fb7df2f9235da70608245519530b2
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:dc39b79c360511ff999eba5b2c299db3e08c9c244a0a1f3da8c90c02695e506d
        version: 7.9.3
      entries:
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.9.x
    defaultChannel: fuse-apicurito-7.13.x
    packageName: fuse-apicurito
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app: kube-green
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: kube-green Community
      provider-url: https://kube-green.dev
    name: kube-green
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kube-green.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kube-green.com/v1alpha1",
                "kind": "SleepInfo",
                "metadata": {
                  "labels": {
                    "app": "kube-green"
                  },
                  "name": "sleepinfo-sample"
                },
                "spec": {
                  "excludeRef": [
                    {
                      "apiVersion": "apps/v1",
                      "kind": "Deployment",
                      "name": "api-gateway"
                    }
                  ],
                  "sleepAt": "20:00",
                  "suspendCronJobs": true,
                  "timeZone": "Europe/Rome",
                  "wakeUpAt": "08:00",
                  "weekdays": "1-5"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          containerImage: docker.io/kubegreen/kube-green:0.7.1
          createdAt: "2025-06-29T15:45:17Z"
          description: Suspend your pods when no-one's using them to save energy and
            restore when necessary
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kube-green/kube-green
          support: kube-green Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SleepInfo is the Schema for the sleepinfos API
            displayName: SleepInfo
            kind: SleepInfo
            name: sleepinfos.kube-green.com
            version: v1alpha1
        description: |
          ## About this Operator

          [kube-green](https://kube-green.dev) shuts down (some of) your resources when you don't need them, reducing the consumption of resources of the cluster.

          Install this operator and apply the SleepInfo CRD to stop your dev namespaces from running during weekends, nights or when it is not necessary.
          See the [docs](https://kube-green.dev/docs/getting-started/) for more information.
          This operator is responsible to suspend and restart pods at specific times, defined by the SleepInfo CRD.

          An example of SleepInfo:

          ```yaml
          apiVersion: kube-green.com/v1alpha1
          kind: SleepInfo
          metadata:
            name: working-hours
          spec:
            weekdays: "1-5"
            sleepAt: "20:00"
            wakeUpAt: "08:00"
            timeZone: "Europe/Rome"
          ```

          This configuration stop pods everyday from monday to friday at 20:00 and restore them at 8:00.
          In this way, your cluster nodes could downscale, producing less CO2 and saving money.
        displayName: kube-green
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - green
        - green software
        - resources
        - downscale
        - autoscale
        - climate change
        links:
        - name: Homepage
          url: https://kube-green.dev
        - name: Source Code
          url: https://github.com/kube-green/kube-green
        maintainers:
        - email: bianchidavide12+kubegreen@gmail.com
          name: Davide
        maturity: alpha
        minKubeVersion: 1.27.0
        provider:
          name: kube-green Community
          url: https://kube-green.dev
        relatedImages:
        - docker.io/kubegreen/kube-green:0.7.1
        version: 0.7.1
      entries:
      - name: kube-green.v0.7.1
        version: 0.7.1
      - name: kube-green.v0.7.0
        version: 0.7.0
      - name: kube-green.v0.6.0
        version: 0.6.0
      - name: kube-green.v0.5.2
        version: 0.5.2
      - name: kube-green.v0.5.1
        version: 0.5.1
      - name: kube-green.v0.5.0
        version: 0.5.0
      - name: kube-green.v0.4.1
        version: 0.4.1
      - name: kube-green.v0.4.0
        version: 0.4.0
      - name: kube-green.v0.3.1
        version: 0.3.1
      - name: kube-green.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: kube-green
    provider:
      name: kube-green Community
      url: https://kube-green.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeRocketCI
      provider-url: https://docs.kuberocketci.io
    name: nexus-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-operator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexus-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexus",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexus-sample"
                },
                "spec": {
                  "secret": "nexus-secret",
                  "url": "https://nexus-example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRepository",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrepository-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrepository",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrepository-sample"
                },
                "spec": {
                  "go": {
                    "proxy": {
                      "cleanup": {
                        "policyNames": [
                          "cleanup-policy"
                        ]
                      },
                      "httpClient": {
                        "autoBlock": true,
                        "blocked": true,
                        "connection": {
                          "retries": 1440,
                          "timeout": 1440,
                          "userAgentSuffix": "user-agent-suffix"
                        }
                      },
                      "name": "go-proxy",
                      "negativeCache": {
                        "enabled": true,
                        "timeToLive": 1440
                      },
                      "online": true,
                      "proxy": {
                        "contentMaxAge": 1440,
                        "metadataMaxAge": 1440,
                        "remoteUrl": "https://test.com"
                      },
                      "routingRule": "routing-rule",
                      "storage": {
                        "blobStoreName": "blob-store-name",
                        "strictContentTypeValidation": true
                      }
                    }
                  },
                  "nexusRef": {
                    "name": "nexus-sample"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrole",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrole-sample"
                },
                "spec": {
                  "description": "test-role",
                  "id": "test-role",
                  "name": "test-role",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "privileges": [
                    "nx-blobstores-all"
                  ]
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexususer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexususer",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexususer-sample"
                },
                "spec": {
                  "email": "test@gmail.com",
                  "firstName": "John1",
                  "id": "test-user",
                  "lastName": "Doe",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "roles": [
                    "nx-admin"
                  ],
                  "secret": "$user-secret:password"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/nexus-operator:3.0.0
          createdAt: "2023-12-11T12:54:50Z"
          description: An Operator for managing Nexus entities.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-nexus-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus is the Schema for the nexus API.
            displayName: Nexus
            kind: Nexus
            name: nexuses.edp.epam.com
            version: v1alpha1
          - description: NexusRepository is the Schema for the nexusrepositories API.
            displayName: Nexus Repository
            kind: NexusRepository
            name: nexusrepositories.edp.epam.com
            version: v1alpha1
          - description: NexusRole is the Schema for the nexusroles API.
            displayName: Nexus Role
            kind: NexusRole
            name: nexusroles.edp.epam.com
            version: v1alpha1
          - description: NexusUser is the Schema for the nexususers API.
            displayName: Nexus User
            kind: NexusUser
            name: nexususers.edp.epam.com
            version: v1alpha1
        description: |
          Nexus Operator is an operator that is responsible for establishing
          a connection to provided NexusRepository Server, reconciling Nexus entities (groups,
          users, repositories, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Nexus Repository.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  nexus-secret
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Nexus` with Nexus instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1alpha1
              kind: Nexus
              metadata:
                name: nexus-sample
              spec:
                secret: nexus-secret                # Secret name
                url: https://nexus.example.com      # Nexus URL
              ```

              Wait for the `.status` field with  `status.connected: true`
        displayName: EDP Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - edp
        - nexus
        - operator
        - artifact
        - storage
        links:
        - name: Nexus Operator
          url: https://github.com/epam/edp-nexus-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/nexus-operator:3.0.0
        version: 3.0.0
      entries:
      - name: nexus-operator.v3.0.0
        version: 3.0.0
      name: alpha
    - currentCSV: nexus-operator.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexus-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexus",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexus-sample"
                },
                "spec": {
                  "secret": "nexus-secret",
                  "url": "https://nexus-example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusBlobStore",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusblobstore-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusblobstore",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusblobstore-sample"
                },
                "spec": {
                  "file": {
                    "path": "storage-blobstore/nexusblobstore-sample"
                  },
                  "name": "nexusblobstore-sample",
                  "nexusRef": {
                    "name": "nexus-sample"
                  },
                  "softQuota": {
                    "limit": 1000,
                    "type": "spaceUsedQuota"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusCleanupPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexuscleanuppolicy-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexuscleanuppolicy",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexuscleanuppolicy-sample"
                },
                "spec": {
                  "criteria": {
                    "lastBlobUpdated": 30
                  },
                  "description": "Cleanup policy for go",
                  "format": "go",
                  "name": "go-cleanup-policy",
                  "nexusRef": {
                    "name": "nexus-sample"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRepository",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrepository-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrepository",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrepository-sample"
                },
                "spec": {
                  "go": {
                    "proxy": {
                      "cleanup": {
                        "policyNames": [
                          "cleanup-policy"
                        ]
                      },
                      "httpClient": {
                        "autoBlock": true,
                        "blocked": true,
                        "connection": {
                          "retries": 1440,
                          "timeout": 1440,
                          "userAgentSuffix": "user-agent-suffix"
                        }
                      },
                      "name": "go-proxy",
                      "negativeCache": {
                        "enabled": true,
                        "timeToLive": 1440
                      },
                      "online": true,
                      "proxy": {
                        "contentMaxAge": 1440,
                        "metadataMaxAge": 1440,
                        "remoteUrl": "https://test.com"
                      },
                      "routingRule": "routing-rule",
                      "storage": {
                        "blobStoreName": "blob-store-name",
                        "strictContentTypeValidation": true
                      }
                    }
                  },
                  "nexusRef": {
                    "name": "nexus-sample"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrole",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrole-sample"
                },
                "spec": {
                  "description": "test-role",
                  "id": "test-role",
                  "name": "test-role",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "privileges": [
                    "nx-blobstores-all"
                  ]
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusScript",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusscript-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusscript",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusscript-sample"
                },
                "spec": {
                  "content": "security.setAnonymousAccess(Boolean.valueOf(args))\n",
                  "name": "anonymous",
                  "nexusRef": {
                    "name": "nexus-sample"
                  },
                  "payload": "true"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexususer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexususer",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexususer-sample"
                },
                "spec": {
                  "email": "test@gmail.com",
                  "firstName": "John1",
                  "id": "test-user",
                  "lastName": "Doe",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "roles": [
                    "nx-admin"
                  ],
                  "secret": "$user-secret:password"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/nexus-operator:3.5.0
          createdAt: "2025-04-22T11:21:24Z"
          description: An Operator for managing Nexus entities.
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/epam/edp-nexus-operator
          support: KubeRocketCI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NexusBlobStore is the Schema for the nexusblobstores API.
            displayName: Nexus Blob Store
            kind: NexusBlobStore
            name: nexusblobstores.edp.epam.com
            version: v1alpha1
          - description: NexusCleanupPolicy is the Schema for the cleanuppolicies
              API.
            displayName: Nexus Cleanup Policy
            kind: NexusCleanupPolicy
            name: nexuscleanuppolicies.edp.epam.com
            version: v1alpha1
          - description: Nexus is the Schema for the nexus API.
            displayName: Nexus
            kind: Nexus
            name: nexuses.edp.epam.com
            version: v1alpha1
          - description: NexusRepository is the Schema for the nexusrepositories API.
            displayName: Nexus Repository
            kind: NexusRepository
            name: nexusrepositories.edp.epam.com
            version: v1alpha1
          - description: NexusRole is the Schema for the nexusroles API.
            displayName: Nexus Role
            kind: NexusRole
            name: nexusroles.edp.epam.com
            version: v1alpha1
          - description: NexusScript is the Schema for the nexusscripts API.
            displayName: Nexus Script
            kind: NexusScript
            name: nexusscripts.edp.epam.com
            version: v1alpha1
          - description: NexusUser is the Schema for the nexususers API.
            displayName: Nexus User
            kind: NexusUser
            name: nexususers.edp.epam.com
            version: v1alpha1
        description: |
          Nexus Operator is an operator that is responsible for establishing
          a connection to provided NexusRepository Server, reconciling Nexus entities (groups,
          users, repositories, etc) according to the created CRs.

          ## Quick Start

          1. Login into Nexus and create user. Attach permissions to user such as scripts, rules, blobs etc. Insert user credentials into Kubernetes secret.

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name: nexus-admin-password
              data:
                password: cGFzcw==  # base64-encoded value of "pass"
                user:     dXNlcg==  # base64-encoded value of "user"
              ```

          2. Create Custom Resource `kind: Nexus` with Nexus instance URL and secret created on the previous step:

              ```yaml
              apiVersion: edp.epam.com/v1alpha1
              kind: Nexus
              metadata:
                name: nexus
              spec:
                secret: nexus-admin-password
                url: http://nexus.example.com
              ```

              Wait for the `.status` field with  `status.connected: true`

          3. Create Role using Custom Resources NexusRole:

              ```yaml
              apiVersion: edp.epam.com/v1alpha1
              kind: NexusRole
              metadata:
                name: edp-admin
              spec:
                description: Read and write access to all repos and scripts
                id: edp-admin
                name: edp-admin
                nexusRef:
                  kind: Nexus
                  name: nexus
                privileges:
                  - nx-apikey-all
                  - nx-repository-view-*-*-add
                  - nx-repository-view-*-*-browse
                  - nx-repository-view-*-*-edit
                  - nx-repository-view-*-*-read
                  - nx-script-*-add
                  - nx-script-*-delete
                  - nx-script-*-run
                  - nx-search-read
              ```
        displayName: EDP Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - artifact
        - edp
        - kuberocketci
        - nexus
        - operator
        - storage
        links:
        - name: Nexus Operator
          url: https://github.com/epam/edp-nexus-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: KubeRocketCI
          url: https://docs.kuberocketci.io
        relatedImages:
        - docker.io/epamedp/nexus-operator:3.5.0
        version: 3.5.0
      entries:
      - name: nexus-operator.v3.5.0
        version: 3.5.0
      - name: nexus-operator.v3.4.0
        version: 3.4.0
      - name: nexus-operator.v3.3.0
        version: 3.3.0
      name: stable
    defaultChannel: stable
    packageName: nexus-operator
    provider:
      name: KubeRocketCI
      url: https://docs.kuberocketci.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp, Inc.
      provider-url: https://www.netapp.com/
    name: trident-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trident-operator.v25.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "trident.netapp.io/v1",
                "kind": "TridentOrchestrator",
                "metadata": {
                  "name": "trident"
                },
                "spec": {
                  "IPv6": false,
                  "debug": true,
                  "nodePrep": [],
                  "imagePullSecrets": [],
                  "imageRegistry": "",
                  "namespace": "trident",
                  "silenceAutosupport": false
                }
              }
            ]
          categories: Storage
          containerImage: docker.io/netapp/trident-operator@sha256:c3448b6847550f45849a5cdbb9e0cedb35290d21131bc513d833d95eb42a43e0
          createdAt: "2025-09-03"
          description: Trident Operator, to manage NetApp Trident installations
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          operators.openshift.io/valid-subscription: No subscriptions are required
          repository: https://github.com/netapp/trident
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Used to deploy NetApp Trident.
            displayName: Trident Orchestrator
            kind: TridentOrchestrator
            name: tridentorchestrators.trident.netapp.io
            version: v1
          - description: Automates AWS FSxN backend configuration
            displayName: Trident Configurator
            kind: TridentConfigurator
            name: tridentconfigurators.trident.netapp.io
            version: v1
        description: |
          NetApp Trident is an open source storage provisioner and orchestrator maintained by NetApp. It enables you to create storage volumes for containerized applications managed by Docker and Kubernetes. For full release information, including patch release changes, see https://docs.netapp.com/us-en/trident/trident-rn.html.
        displayName: NetApp Trident
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - NetApp
        - ONTAP
        - Trident
        - storage
        - performance
        - open source
        - Astra
        - Cloud Native
        - AWS
        - Azure
        - GCP
        - Cloud Volumes
        links:
        - name: GitHub Repository
          url: https://github.com/NetApp/trident
        - name: Trident documentation
          url: https://docs.netapp.com/us-en/trident/index.html
        - name: Support policy
          url: https://mysupport.netapp.com/site/info/version-support
        - name: Release Notes
          url: https://docs.netapp.com/us-en/trident/trident-rn.html
        maintainers:
        - email: ng-trident-openshift-operator@netapp.com
          name: NetApp
        provider:
          name: NetApp, Inc.
          url: https://www.netapp.com/
        relatedImages:
        - docker.io/netapp/trident-operator@sha256:c3448b6847550f45849a5cdbb9e0cedb35290d21131bc513d833d95eb42a43e0
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:b88958c38883f5b1e3d307da90aac34a5ceafdb6bd3d989b062cb24ecafd1fbd
        - docker.io/netapp/trident-autosupport@sha256:21b3afbebfac3cc28c4eb319c2d385187bd8ebb728d8cfd82dbc437cc9b83b33
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:bdee8df520467b30312e7b14d7b6ef3963557ccd38269913de3fa51bb1bceec9
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:145ab76f9606e9d51dcd1e4e159c67021ecf53c24d8d80ed7bbeb2cc7fb5c42b
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6781d64bfb01e252595130eec814b9d2fe6426720a003cc61e82e9963767bfd4
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:b48bd5bc612c1bbe73d203c70e6caaae2b5360bd31eba20eb06cbd2102cc34fc
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:0ef3bf2a4dcbb5aa829ce9cc5f12efaccaf99b246b4b15abbb66b0135bec4cab
        - docker.io/netapp/trident@sha256:7925717979a84f93b62a2592682cde3443d6350ec4e88f8fd1a7d3ea812e8900
        version: 25.6.2
      entries:
      - name: trident-operator.v25.6.2
        version: 25.6.2
      - name: trident-operator.v25.6.1
        version: 25.6.1
      - name: trident-operator.v25.6.0
        version: 25.6.0
      - name: trident-operator.v25.2.1
        version: 25.2.1
      - name: trident-operator.v25.2.0
        version: 25.2.0
      name: stable
    defaultChannel: stable
    packageName: trident-operator
    provider:
      name: NetApp, Inc.
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: xspc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: xspc-operator.v1.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.hpe.com/v1",
                "kind": "XSPC",
                "metadata": {
                  "name": "xspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.connect.redhat.com/hpestorage/xspc-csi-driver-operator@sha256:8b8724998a8301057a3c4187c7b94bf4484aeb807065a12f84142cf844ea5534
          createdAt: "2024-12-17T03:53:30Z"
          description: An operator for managing HPE XP Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XSPC is the Schema for the xspcs API
            displayName: XSPC
            kind: XSPC
            name: xspcs.csi.hpe.com
            version: v1
        description: |-
          ## About
          HPE XP Storage Plug-in for Containers (XSPC) is a plugin that integrates HPE XP storage into Kubernetes based clusters.
          XSPC provides persistent volume provisioning capabilities from HPE's block storage arrays.
        displayName: HPE XP Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - XP7
        - XP8
        - HPE
        - Storage
        - Container Storage Interface
        - CSI
        links:
        - name: Knowledge Base
          url: https://www.hpe.com/us/en/storage/enterprise-xp-storage.html
        maintainers:
        - email: hpexpci@hpe.com
          name: HPE
        maturity: stable
        provider:
          name: HPE
        relatedImages:
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.connect.redhat.com/hpestorage/xspc-csi-driver@sha256:bdeac71c0b54c2779910a8d8faea03c1122000127704f17ceadba4938b685353
        - registry.connect.redhat.com/hpestorage/xspc-csi-driver-operator@sha256:8b8724998a8301057a3c4187c7b94bf4484aeb807065a12f84142cf844ea5534
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        version: 1.15.0
      entries:
      - name: xspc-operator.v1.15.0
        version: 1.15.0
      name: stable
    defaultChannel: stable
    packageName: xspc-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Entando, Inc
      provider-url: ""
    name: entando-k8s-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: entando-k8s-operator.v6.3.2-pr3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: beta
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - docker.io/entando/entando-k8s-service@sha256:775391f0005e64764288bfe007d9ea576f13618c02ea400d030a7502cfb551bd
        - docker.io/entando/entando-k8s-database-service-controller@sha256:58ca4ad2fca83a715aba9b68609f9056f633d556eb9845b3aff5c05769369539
        - docker.io/entando/app-builder@sha256:1ca90c09302b9cadf8950278d92ad4d7f3a14f02dea8704320794aa77ada8631
        - docker.io/entando/entando-k8s-app-controller@sha256:c68d3b129fc625f5d1ee10ed2a107e28c145049d4b513dcdb2c736555025e339
        - docker.io/entando/entando-k8s-plugin-controller@sha256:a20c753dda86e0f54f210e446b404733045ba426fc7186dc3936daed8224fbcb
        - docker.io/entando/entando-component-manager@sha256:668191fc86b090452baaba18ec7bbc3d07df5a99750e0c5b98759599d565cfbe
        - docker.io/entando/entando-redhat-sso@sha256:3692adb3694b3bf278bed7a678c9d0eaa0aba8e3e3ed4117ca04156525bcac72
        - registry.redhat.io/rhel8/mysql-80@sha256:570bce7dc649ac7ac85e7ecb196aabc281c77a94590aa7df7d9746869ce757ea
        - docker.io/entando/entando-k8s-dbjob@sha256:8ae47dbf93201965106663299f06daad544982fa244d4a0f365a1b717727e5ed
        - docker.io/entando/entando-k8s-composite-app-controller@sha256:4bc6797bb41b3e407b64530c3cff7a78f3fc2c22863caad9c606ccd650310b61
        - docker.io/entando/entando-k8s-app-plugin-link-controller@sha256:aa7ba0861705213434e610530c5bf5da1bdb33f1834d3c346174f5a28145c0bd
        - docker.io/entando/entando-de-app-eap@sha256:1f0e35d7f7ee8123721cf8e80ac96ac2f594f21a2b68cf9e720917d090c8095e
        - docker.io/entando/entando-k8s-keycloak-controller@sha256:f1ed42001260802353197433cea7df4c2d0e5fd9bcd926bb5ae7c654dfad2ffe
        - registry.redhat.io/rhel8/postgresql-12@sha256:f4e5c728b644bf1888ec8086424852ed74b5596a511be29e636fb10218fc9b6f
        - entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        - docker.io/entando/entando-k8s-cluster-infrastructure-controller@sha256:86dbea6696349d7b2393e12bacef4d1acec2dd3f8551e397fa26c92706a18dd3
        version: 6.3.2-pr3
      entries:
      - name: entando-k8s-operator.v6.3.2-pr3
        version: 6.3.2-pr3
      name: alpha
    - currentCSV: entando-k8s-operator.v6.3.2-pr3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: beta
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - docker.io/entando/entando-redhat-sso@sha256:3692adb3694b3bf278bed7a678c9d0eaa0aba8e3e3ed4117ca04156525bcac72
        - entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        - docker.io/entando/entando-de-app-eap@sha256:1f0e35d7f7ee8123721cf8e80ac96ac2f594f21a2b68cf9e720917d090c8095e
        - docker.io/entando/entando-k8s-composite-app-controller@sha256:4bc6797bb41b3e407b64530c3cff7a78f3fc2c22863caad9c606ccd650310b61
        - docker.io/entando/entando-k8s-app-controller@sha256:c68d3b129fc625f5d1ee10ed2a107e28c145049d4b513dcdb2c736555025e339
        - docker.io/entando/entando-k8s-plugin-controller@sha256:a20c753dda86e0f54f210e446b404733045ba426fc7186dc3936daed8224fbcb
        - docker.io/entando/entando-component-manager@sha256:668191fc86b090452baaba18ec7bbc3d07df5a99750e0c5b98759599d565cfbe
        - docker.io/entando/entando-k8s-app-plugin-link-controller@sha256:aa7ba0861705213434e610530c5bf5da1bdb33f1834d3c346174f5a28145c0bd
        - docker.io/entando/entando-k8s-database-service-controller@sha256:58ca4ad2fca83a715aba9b68609f9056f633d556eb9845b3aff5c05769369539
        - docker.io/entando/app-builder@sha256:1ca90c09302b9cadf8950278d92ad4d7f3a14f02dea8704320794aa77ada8631
        - docker.io/entando/entando-k8s-keycloak-controller@sha256:f1ed42001260802353197433cea7df4c2d0e5fd9bcd926bb5ae7c654dfad2ffe
        - registry.redhat.io/rhel8/mysql-80@sha256:570bce7dc649ac7ac85e7ecb196aabc281c77a94590aa7df7d9746869ce757ea
        - registry.redhat.io/rhel8/postgresql-12@sha256:f4e5c728b644bf1888ec8086424852ed74b5596a511be29e636fb10218fc9b6f
        - docker.io/entando/entando-k8s-cluster-infrastructure-controller@sha256:86dbea6696349d7b2393e12bacef4d1acec2dd3f8551e397fa26c92706a18dd3
        - docker.io/entando/entando-k8s-dbjob@sha256:8ae47dbf93201965106663299f06daad544982fa244d4a0f365a1b717727e5ed
        - docker.io/entando/entando-k8s-service@sha256:775391f0005e64764288bfe007d9ea576f13618c02ea400d030a7502cfb551bd
        version: 6.3.2-pr3
      entries:
      - name: entando-k8s-operator.v6.3.2-pr3
        version: 6.3.2-pr3
      name: beta
    - currentCSV: entando-k8s-operator.v6.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: registry.hub.docker.com/entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - registry.hub.docker.com/entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        version: 6.3.2
      entries:
      - name: entando-k8s-operator.v6.3.2
        version: 6.3.2
      name: stable
    defaultChannel: stable
    packageName: entando-k8s-operator-rhmp
    provider:
      name: Entando, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: submariner.io
      provider-url: ""
    name: submariner
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: submariner.v0.17.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:067ba6e58a6fe150a212881a29c90aa7cbd22f53a6877eebc413072f73f9179a
          createdAt: "2025-06-18 01:35:57"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.16.0 <0.17.6'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.17 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:d58214791a60cd9e4366c6ff534584eedbd37386da71b868867e5e212df270f9
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:a565d8c094f3755b6c9303c5b8ae03a5883796e0fea754b0202e5792f1e472c3
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:067ba6e58a6fe150a212881a29c90aa7cbd22f53a6877eebc413072f73f9179a
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:b76998e8bbef1d06aa8dbc49af4f42a3c5603d480a67e46baf11168514c7759a
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:56848aefe1d156147ab73e7353017b2663f63e2e01fe43ede57abc72cfc0c50f
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:e6dbc4bf3e8b1a3c1a7ed10eb463d134f9ed28a03c7bb523c82c5431c5c9ae90
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:5510cb1bb3e0ca5f2365f4a5fc147ca693d3c220f1a8e7835f8b13b66b957ae8
        version: 0.17.6
      entries:
      - name: submariner.v0.17.6
        version: 0.17.6
      - name: submariner.v0.17.5
        version: 0.17.5
      - name: submariner.v0.17.4
        version: 0.17.4
      - name: submariner.v0.17.2
        version: 0.17.2
      - name: submariner.v0.17.2-0.1725901206.p
        version: 0.17.2+0.1725901206.p
      - name: submariner.v0.17.1
        version: 0.17.1
      - name: submariner.v0.17.0
        version: 0.17.0
      name: stable-0.17
    - currentCSV: submariner.v0.18.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:d8d7f0147c7b65b41903e101240905703992a7bbadb33e302aa86b23000d2095
          createdAt: "2025-05-07 20:54:56"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.17.0 <0.18.5'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.18 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:e5c16f0d720e15d54781870a5e63c85540256355009584f7d1bd03ef20a2eae4
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:00072fea3fdb5e294a01da0ee44165fc52219a965a27bdf7d792276f6c3a36c3
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:e8366433d1f50e2062b86d7f749f17cd6566197560256e77f57b2c165f277733
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:a07ef1bfec7549b32ad94f2c44a168860d5f3b46fd13593c068c58bb137c77fc
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:d8d7f0147c7b65b41903e101240905703992a7bbadb33e302aa86b23000d2095
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:27f18015b8896b04457c356d03276d50c34b8c1d84142ec07342ff2c1c4e81fa
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:e913be7a440b6cc0c1dddbda93a5979fe28b99e63f12a7949966b69126fcee90
        version: 0.18.5
      entries:
      - name: submariner.v0.18.5
        version: 0.18.5
      - name: submariner.v0.18.4
        version: 0.18.4
      - name: submariner.v0.18.3
        version: 0.18.3
      - name: submariner.v0.18.2
        version: 0.18.2
      - name: submariner.v0.18.0
        version: 0.18.0
      name: stable-0.18
    - currentCSV: submariner.v0.19.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:4afaf6689e2225c5c7f6fbe002202a1cdf9444e23fd210cc2d86da71c9f78626
          createdAt: "2025-04-22 21:12:45"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.18.0 <0.19.4'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.19 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:d9d4d03013d65d11a4e8ecda411015f359e4215a3a53059a37a810acf2e3ea6c
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:4afaf6689e2225c5c7f6fbe002202a1cdf9444e23fd210cc2d86da71c9f78626
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:5b83c5889aa0d8f7838dd5afb8c901dfb0ce0fbe01175e8468ec50ddedef9223
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:70008af49a9d1ee516204715f66bd415c9caa7462f6b178dc1dc91e513605fcd
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:ef73b8964cd66ffb496191056594e0886afa9d17c03976b4159413c89ed569bf
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:5938d67a0a2653e92f0995440700512992ed82269820679f68483dcc1b2d98d9
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:1bbb624bbefbb75407071449d0144258732e219e3a7fc67ac709c3bcee560576
        version: 0.19.4
      entries:
      - name: submariner.v0.19.4
        version: 0.19.4
      - name: submariner.v0.19.3
        version: 0.19.3
      - name: submariner.v0.19.2
        version: 0.19.2
      - name: submariner.v0.19.0
        version: 0.19.0
      name: stable-0.19
    - currentCSV: submariner.v0.20.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "quay.io/submariner",
                  "version": "0.20.2"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "quay.io/submariner",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "0.20.2"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: $(SUBMARINER_OPERATOR_IMAGE)
          createdAt: "2025-09-30 14:21:54"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/submariner-io/submariner-operator
          support: Submariner
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.20 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: submariner.io
        relatedImages:
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:fd6c46b4a0570b5e7b6457c465e5a1b843ba6f9c209c1856dbebfee1068e35ea
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:70d563a6e6526b9a146a2973210b2abdae09d5acab403fa3f0b465ed5e3c9b11
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:77e0da2db6d5bb2831e6c51214e08685c9f7f3b8b1e6ac53b4fd1418644ad6a1
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:0eef6521c96b29722076c1e220770124c8ed214379abc4654d9567cff8b52d04
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:8ea02aba8f9dd901f1f0ccff97ec8d157bebe5a90a6c924affe5e01cf3454e72
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:ed521b241da60efabfe2e2b80c51639d1a82a43945945879decbfb192bbc6d23
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:4a4ac198e9d0f4a4aaed8f0ec0e1d79bd8fa712b671ee697f7ff9d5616a02d84
        version: 0.20.2
      entries:
      - name: submariner.v0.20.2
        version: 0.20.2
      - name: submariner.v0.20.1
        version: 0.20.1
      - name: submariner.v0.20.0
        version: 0.20.0
      name: stable-0.20
    - currentCSV: submariner.v0.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "quay.io/submariner",
                  "version": "v0.21.0"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "quay.io/submariner",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v0.21.0"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: $(SUBMARINER_OPERATOR_IMAGE)
          createdAt: "2025-08-13 21:57:14"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/submariner-io/submariner-operator
          support: Submariner
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.21 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: submariner.io
        relatedImages:
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:cd851eab4a2214ea7a699bd01d2abac1322d01562f27b35f58f3c0b8716f88fb
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:bcdeaa67664568c5e2a6928236ff69b175ce65ddf7a58c1664d091b926000052
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:b0ad338aed82367048cf024767b0b09e6bb09bbf7dd5128e6711e1feec6f696e
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:65765c7f08c5454b557fbda91c49efbd2cc0d4d973cf8acc7ff40ec8a381839d
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:f5909b1802986622ab169a82d5dd7877af72660903e961b3097cc29e30eb9799
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:b307f748125fd459c3474fdab645688675a06515a9924b47cedba80e038458db
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:0b74f058400f24a63cf9b57a388914cc6dda38e03d639196871f3fd76fb7f2ae
        version: 0.21.0
      entries:
      - name: submariner.v0.21.0
        version: 0.21.0
      name: stable-0.21
    defaultChannel: stable-0.21
    packageName: submariner
    provider:
      name: submariner.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Devfile
      provider-url: https://devfile.io
    name: devworkspace-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devworkspace-operator.v0.36.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/internal-objects: '["devworkspaceroutings.controller.devfile.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DevWorkspaceOperatorConfig
            name: devworkspaceoperatorconfigs.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspaceRouting
            name: devworkspaceroutings.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha2
          - kind: DevWorkspaceTemplate
            name: devworkspacetemplates.workspace.devfile.io
            version: v1alpha1
          - kind: DevWorkspaceTemplate
            name: devworkspacetemplates.workspace.devfile.io
            version: v1alpha2
        description: |
          The DevWorkspace Operator enables cluster-level support for the
          [Devfile 2.0 spec](https://docs.devfile.io), enabling static, reproducible
          configurations for launching cloud-based editors and IDEs in OpenShift and
          Kubernetes clusters.

          Leveraging the language server protocol and the Eclipse Theia web IDE, the
          DevWorkspace operator provides easy configuration of full development
          environments on the cloud with support for a wide variety of languages and
          technologies, including Go, Java, Typescript/Javascript, Python, and more.

          The DevWorkspace Operator is also used in the Web Terminal Operator to
          automatically provision Web Terminal environments.

          ## Installing the operator
          The DevWorkspace Operator can be installed directly from the OperatorHub UI and
          will be available in all namespaces on the cluster. DevWorkspace creation is
          driven by the DevWorkspace custom resource, which can be created in any
          namespace to provision a full development environment. To get started, browse
          the DevWorkspace [spec](https://devfile.io/docs/2.3.0/devfile-schema).

          Once a DevWorkspace is started, it can be accessed via the URL in its
          `.status.mainUrl` field.

          It's recommended to install the DevWorkspace Operator to the
          `openshift-operators` namespace for compatibility.

          ## Uninstalling the operator
          The DevWorkspace Operator utilizes finalizers on resources it creates and
          webhooks to restrict access to development resources. As a result, manual steps
          are required in uninstalling the operator. See the
          [documentation](https://github.com/devfile/devworkspace-operator/blob/main/docs/uninstall.md)
          for details.
        displayName: DevWorkspace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Devworkspace Operator
        - DevWorkspaces
        - Devfile
        links:
        - name: Devworkspace Operator
          url: https://github.com/devfile/devworkspace-operator
        maintainers:
        - email: dakwon@redhat.com
          name: David Kwon
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        maturity: alpha
        provider:
          name: Devfile
          url: https://devfile.io
        relatedImages:
        - registry.redhat.io/devworkspace/devworkspace-rhel9-operator@sha256:83b6abf6e57703da1125e037ea9353c224602bddc3146e5b00df47f1b1976e9a
        - registry.redhat.io/ubi9/ubi-minimal@sha256:a50731d3397a4ee28583f1699842183d4d24fadcc565c4688487af9ee4e13a44
        - registry.redhat.io/devworkspace/devworkspace-project-clone-rhel9@sha256:312a013201877cea03956e3db2ba19a3d7f842de2c3a65927c65c3e8d22a57fe
        version: 0.36.0
      entries:
      - name: devworkspace-operator.v0.36.0
        version: 0.36.0
      - name: devworkspace-operator.v0.35.1
        version: 0.35.1
      - name: devworkspace-operator.v0.34.0
        version: 0.34.0
      - name: devworkspace-operator.v0.33.0
        version: 0.33.0
      - name: devworkspace-operator.v0.32.1
        version: 0.32.1
      - name: devworkspace-operator.v0.31.2
        version: 0.31.2
      - name: devworkspace-operator.v0.31.1
        version: 0.31.1
      - name: devworkspace-operator.v0.31.0
        version: 0.31.0
      - name: devworkspace-operator.v0.30.2
        version: 0.30.2
      - name: devworkspace-operator.v0.30.1
        version: 0.30.1
      - name: devworkspace-operator.v0.30.0
        version: 0.30.0
      - name: devworkspace-operator.v0.29.0
        version: 0.29.0
      - name: devworkspace-operator.v0.28.0-0.1719314743.p
        version: 0.28.0+0.1719314743.p
      - name: devworkspace-operator.v0.28.0
        version: 0.28.0
      - name: devworkspace-operator.v0.27.0-0.1714987663.p
        version: 0.27.0+0.1714987663.p
      - name: devworkspace-operator.v0.27.0
        version: 0.27.0
      - name: devworkspace-operator.v0.26.0
        version: 0.26.0
      - name: devworkspace-operator.v0.25.0
        version: 0.25.0
      - name: devworkspace-operator.v0.24.0
        version: 0.24.0
      - name: devworkspace-operator.v0.23.0
        version: 0.23.0
      - name: devworkspace-operator.v0.22.0
        version: 0.22.0
      - name: devworkspace-operator.v0.21.2
        version: 0.21.2
      - name: devworkspace-operator.v0.21.1
        version: 0.21.1
      - name: devworkspace-operator.v0.20.0
        version: 0.20.0
      - name: devworkspace-operator.v0.19.1
        version: 0.19.1
      - name: devworkspace-operator.v0.19.1-0.1682321189.p
        version: 0.19.1+0.1682321189.p
      - name: devworkspace-operator.v0.19.1-0.1679521112.p
        version: 0.19.1+0.1679521112.p
      - name: devworkspace-operator.v0.18.1
        version: 0.18.1
      - name: devworkspace-operator.v0.18.1-0.1675929565.p
        version: 0.18.1+0.1675929565.p
      - name: devworkspace-operator.v0.17.0
        version: 0.17.0
      - name: devworkspace-operator.v0.16.0-0.1666668361.p
        version: 0.16.0+0.1666668361.p
      - name: devworkspace-operator.v0.16.0
        version: 0.16.0
      - name: devworkspace-operator.v0.15.2-0.1661828401.p
        version: 0.15.2+0.1661828401.p
      - name: devworkspace-operator.v0.15.2
        version: 0.15.2
      - name: devworkspace-operator.v0.14.1
        version: 0.14.1
      - name: devworkspace-operator.v0.13.0
        version: 0.13.0
      - name: devworkspace-operator.v0.12.0
        version: 0.12.0
      - name: devworkspace-operator.v0.11.0
        version: 0.11.0
      - name: devworkspace-operator.v0.10.0
        version: 0.10.0
      - name: devworkspace-operator.v0.9.0
        version: 0.9.0
      name: fast
    defaultChannel: fast
    packageName: devworkspace-operator
    provider:
      name: Devfile
      url: https://devfile.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microcks.io
      provider-url: ""
    name: microcks
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: microcks-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "microcks.github.io/v1alpha1",
                "kind": "MicrocksInstall",
                "metadata": {
                  "name": "my-microcksinstall"
                },
                "spec": {
                  "name": "my-microcksinstall",
                  "version": "1.10.0",
                  "microcks": {
                    "replicas": 1
                  },
                  "postman": {
                    "replicas": 1
                  },
                  "keycloak": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "1Gi"
                  },
                  "mongodb": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "2Gi"
                  }
                }
              },
              {
                "apiVersion": "microcks.github.io/v1alpha1",
                "kind": "MicrocksInstall",
                "metadata": {
                  "name": "my-microcksinstall-minikube"
                },
                "spec": {
                  "name": "my-microcksinstall-minikube",
                  "version": "1.10.0",
                  "microcks": {
                    "replicas": 2,
                    "url": "microcks.192.168.99.100.nip.io"
                  },
                  "postman": {
                    "replicas": 2
                  },
                  "keycloak": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "1Gi",
                    "url": "keycloak.192.168.99.100.nip.io"
                  },
                  "mongodb": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "2Gi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/microcks/microcks-ansible-operator:1.10.0
          createdAt: "2024-08-01T09:14:55Z"
          description: Open Source mocking and testing platform for API and microservices
          repository: https://github.com/microcks/microcks-ansible-operator
          support: microcks.github.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Microcks installation
            displayName: MicrocksInstall
            kind: MicrocksInstall
            name: microcksinstalls.microcks.github.io
            version: v1alpha1
        description: |
          Microcks is an Open source tool for API and microservices mocking and testing. It provides a platform for referencing, deploying mocks and allow contract-testing of your APIs. It can also be considered as a Service Virtualization solution because it will allow you to provide fake API or Service implementation before development being actually done. It supports Async APIs, REST APIs, gRPC, GraphQL and SOAP WebServices. It perfectly integrates into an iterative, contract-first delivery process.
          Microcks tries not reinventing the wheel and let you capitalize on already known tools of standards. It supports [AsyncAPI Specifications](https://asyncapi.com), [OpenAPI Specification](https://www.openapis.org/) `3.x`, [gRPC Protobuf](https://grpc.io/), [GraphQL Schema](https://graphql.org) out-of-the-box and can also relies on popular tools like [SoapUI](http://www.soapui.com) or [Postman](http://www.getpostman.io) to edit your API request / response samples as well as your test scripts.
          ### Operator features
          * **Create/Destroy**: Easily launch Microcks installations (`MicrocksInstall` custom resource) for your Kubernetes namespace
          ### MicrocksInstall options
          * **keycloak** component can be deployed as part of the install or you can reuse an existing instance. You can reference an existing `Secret` for providing TLS certificates for securing the `Ingress`,
          * **mongodb** component can be deployed as part of the install or you can reuse an existing instance, You can reference an existing `Secret` for providing connection credentials,
          * **features** component allow to activate optional features like the mocking of AsyncAPIs. In case you turn `features.async.enabled` to `true` and `features.async.kafka.install` to `true`, you must have the [Strimzi Operator](https://operatorhub.io/operator/strimzi-kafka-operator) already installed on your cluster/namespace.
          ### Documentation
          Documentation to the current _master_ branch as well as all releases can be found on our [GitHub](https://github.com/microcks/microcks-ansible-operator).
          ### License
          Microcks is licensed under the [Apache License, Version 2.0](https://github.com/microcks/microcks/blob/master/LICENSE).
        displayName: Microcks Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mock
        - test
        - API
        - microservice
        links:
        - name: Product Page
          url: https://microcks.io
        - name: Microcks Operator Source Code
          url: https://github.com/microcks/microcks-ansible-operator
        maintainers:
        - email: laurent.broudoux@gmail.com
          name: lbroudoux
        maturity: alpha
        provider:
          name: Microcks.io
        relatedImages:
        - quay.io/microcks/microcks-ansible-operator:1.10.0
        version: 1.10.0
      entries:
      - name: microcks-operator.v1.10.0
        version: 1.10.0
      - name: microcks-operator.v1.9.0
        version: 1.9.0
      - name: microcks-operator.v1.8.1
        version: 1.8.1
      - name: microcks-operator.v1.8.0
        version: 1.8.0
      - name: microcks-operator.v1.7.1
        version: 1.7.1
      - name: microcks-operator.v1.7.0
        version: 1.7.0
      - name: microcks-operator.v1.6.1
        version: 1.6.1
      - name: microcks-operator.v1.6.0
        version: 1.6.0
      - name: microcks-operator.v1.5.2
        version: 1.5.2
      - name: microcks-operator.v1.5.1
        version: 1.5.1
      - name: microcks-operator.v1.5.0
        version: 1.5.0
      - name: microcks-operator.v1.4.1
        version: 1.4.1
      name: stable
    defaultChannel: stable
    packageName: microcks
    provider:
      name: Microcks.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-wafv2-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-wafv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wafv2.services.k8s.aws/v1alpha1",
                "kind": "IPSet",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/wafv2-controller:1.1.1
          createdAt: "2025-09-20T21:41:21Z"
          description: AWS WAFV2 controller is a service controller for managing WAFV2
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IPSet represents the state of an AWS wafv2 IPSet resource.
            displayName: IPSet
            kind: IPSet
            name: ipsets.wafv2.services.k8s.aws
            version: v1alpha1
          - description: RuleGroup represents the state of an AWS wafv2 RuleGroup
              resource.
            displayName: RuleGroup
            kind: RuleGroup
            name: rulegroups.wafv2.services.k8s.aws
            version: v1alpha1
          - description: WebACL represents the state of an AWS wafv2 WebACL resource.
            displayName: WebACL
            kind: WebACL
            name: webacls.wafv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Web Application Firewall (Amazon WAFV2) resources in AWS from within your Kubernetes cluster.

          **About Amazon WAFV2**

          AWS WAF is a web application firewall that lets you monitor the HTTP and HTTPS requests that are forwarded to your protected web application resources. You can protect the following resource types:
          - Amazon CloudFront distribution
          - Amazon API Gateway REST API
          - Application Load Balancer
          - AWS AppSync GraphQL API
          - Amazon Cognito user pool
          - AWS App Runner service
          - AWS Verified Access instance

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon WAFV2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - wafv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon WAFV2 Developer Resources
          url: https://aws.amazon.com/waf/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: wafv2 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/wafv2-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-wafv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-wafv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-wafv2-controller.v1.0.14
        version: 1.0.14
      - name: ack-wafv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-wafv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-wafv2-controller.v1.0.10
        version: 1.0.10
      - name: ack-wafv2-controller.v1.0.9
        version: 1.0.9
      - name: ack-wafv2-controller.v1.0.8
        version: 1.0.8
      - name: ack-wafv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-wafv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-wafv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-wafv2-controller.v1.0.1
        version: 1.0.1
      - name: ack-wafv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-wafv2-controller.v0.0.4
        version: 0.0.4
      - name: ack-wafv2-controller.v0.0.3
        version: 0.0.3
      - name: ack-wafv2-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-wafv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/in-en/products/cloud-pak-for-data
    name: ibm-zen-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-zen-operator.v6.1.3
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.3 <6.1.3'
          operators.operatorframework.io/builder: operator-sdk-v1.6.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: represents zen instance
            kind: ZenService
            name: zenservices.zen.cpd.ibm.com
            version: v1
          - description: represents zen extension instance
            kind: ZenExtension
            name: zenextensions.zen.cpd.ibm.com
            version: v1
        description: |
          The Zen operator is a component in IBM® Cloud Pak foundational services. It serves a framework for user access management, defining user roles and for powering persona driven user experiences for Cloud Paks and Service offerings
        displayName: IBM Zen Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cloud Pak
        - IBM
        - Access management
        - Platform UI
        - Zen framework
        maintainers:
        - email: cp4d@ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/in-en/products/cloud-pak-for-data
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:443b51f8b10acc85bdefde7193e0f45b1c423fa7fbdcaa28342805815c43db3d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:3d5c6573faf4ae00cbb21cbfe9b2ad5cd8ffb658df0c5ba09bd98b3bf54f9979
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:fa01c7d11c23a47bb84375389e69619455c2879b5ed6bb5b9b753b945dcb02c3
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:45194b56f9fcb3af83158eb0297248364db40e8164fb4fe4f5656c58aeac3fa5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:a99b1f5a4eb0306b83562de0184fc826c34835ffbf46d929ceb9e70e0d325361
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:5f48081ffa453ec740c2562c24603d78b481b5e236d6bcc73e96203cabe153f2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:0522fc5482404fa9be179c7225a5692c8d71d755fda0b2d3928e452adcf030fb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:45ca5370b726e31c55f9d7d249374189cb48dc05ab89d675c35225d53b90e7c9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:aaee5240471872a8f1a77d4440a696d45041c5ce06849d516daccb90f5341b6d
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:abd96661823553d623eb0ea1fac8271d64391243dd0f13eb4a4fcc858dfe10a5
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:82afac013ef9b6b9d0eacb4575c979b65abd6f2821971182b90c71bd488ef366
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:b87fa8374976ccd3dad90ddbf6b65cd6ae5a028ba896562f284f089380ebe44a
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:24dff64a80b910f0b4c5442d4f611f9df7602a16c5231eff2c9d08f07b2a9f9a
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:ad027c91c719fd1d931cbac8c76735474f676b228f9ab23d1f243201246b3238
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:c6862158a2a5ae0a4e4206bfdc7f73e1c2359544966c72cf4fc3a617acb13697
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:10ede3c4e3c7e0cbc26e7b6e797cad35e39face1a1b9d4a4e2440ab0feeeb814
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:deb66a64288a370c5da5c7cc0ba4e2ba4ab816d7c5a080a0c505b05f55fa6072
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:7dbff355db7739152961cb6a97887d2e43bc960ac58837c126e03ed1a4480a3a
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:a72c2b70a9f0a93047d47247cddc1616138cbc87337289ddb208565af6e4369d
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:183ee860bf48076ea155db08032eed7004c1204a363625f54218734b00175441
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:525924934f74b28783dacf630a6401ccd4c88d253aafa2d312b7993d8b9a14c1
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:a7c050ae29cb93754f59d8f83676275ee9a91ae9a9956395b97abf8c5e20e96e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8da8b240176c00f8f6eadf32e5614c663cc1d56d546c73996cca614b715acf4a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:f06a1f50cd18d6f48762fdd0394307b44c0573d0cc91f53724c5b3c8c5e2b8e4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:c4ff5dae4560e5a943c7218df3a5aa9ca1a9ee2d1f7dfac73db05892725c34a9
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-bundle@sha256:552fc53664722903eccaa7d3782a3e84413509fd357725fe0b0f4a077c48ab2c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:719b0871f0d32da894ec80dc2a23483623b7802a5c6f935cf3672468f99faa40
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:c8baecad9c97ddbbb3f9f02c90abbf483d46974b7bc9605622a302cbeba1de74
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:b5a8fb14c3bb7c89ff970ca85f06dfcdca832e2b45ad86e8bb1b6c0a0e7e0764
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:8444703b2a7822f82141afcf4f36bf255650d755321b9e9fe2cbd1bdea41d5c7
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:e9173d81ef2a75839c0e1d677b72bb00b8e6d7ef0511afab103f08a884a03738
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:a8d2e42757c6ccbfe4372465b1827c633aad012c5194bf8b08afcd5691caa6f5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:6bbcb64cf9adb4d3b6a4c259eb910e7da5c3c58732643e60d416399b2083cc0d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:ea57c7cb1ccd5bce17f6cb35e326ab974718be8cd19b0cd38ea8112e91c51939
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ee161954766211651d7cb13f3f5a3c025ac04094ba742be453ce789c6f0e24a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:ecbc65705de76804a3c7642339ecfb4cc0f5a61891058e07c20bf2b63f1b82b8
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:1df782d0d923544767696d719bd0ac10802b8e06508a104af5e023a519f32dee
        version: 6.1.3
      entries:
      - name: ibm-zen-operator.v6.1.3
        version: 6.1.3
      name: v6.1
    defaultChannel: v6.1
    packageName: ibm-zen-operator
    provider:
      name: IBM
      url: https://www.ibm.com/in-en/products/cloud-pak-for-data
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: hpe-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-csi-operator.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.hpe.com/v1",
                "kind": "HPECSIDriver",
                "metadata": {
                  "name": "hpecsidriver-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "csp": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "disable": {
                    "alletra6000": false,
                    "alletra9000": false,
                    "alletraStorageMP": false,
                    "b10000FileService": false,
                    "nimble": false,
                    "primera": false
                  },
                  "disableHostDeletion": false,
                  "disableNodeConfiguration": false,
                  "disableNodeConformance": false,
                  "disableNodeGetVolumeStats": false,
                  "disableNodeMonitor": false,
                  "disablePreInstallHooks": false,
                  "imagePullPolicy": "IfNotPresent",
                  "images": {
                    "b10000FileServiceCSP": "quay.io/hpestorage/alletrastoragemp-b10000-nfs-csp@sha256:6a1855a2ef1a83b886bb1f45b596e8e420e5e13c84bbf446c8c202eaf27a2013",
                    "csiAttacher": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                    "csiControllerDriver": "quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676",
                    "csiExtensions": "quay.io/hpestorage/csi-extensions@sha256:27074aaa8df3b867cfa3d12318612c45f8e541d750c27db5e080734592160c89",
                    "csiNodeDriver": "quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676",
                    "csiNodeDriverRegistrar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                    "csiProvisioner": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                    "csiResizer": "registry.k8s.io/sig-storage/csi-resizer@sha256:8ddd178ba5d08973f1607f9b84619b58320948de494b31c9d7cd5375b316d6d4",
                    "csiSnapshotter": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769",
                    "csiVolumeGroupProvisioner": "quay.io/hpestorage/volume-group-provisioner@sha256:0a5f39f52113b963978a695bb8e5a5a3c4d387a02eb4d4a9647b4bc804de7399",
                    "csiVolumeGroupSnapshotter": "quay.io/hpestorage/volume-group-snapshotter@sha256:177d9abf8fbd3ac26ca4c12ca4b4adde20ebab308c79f10b04632f0171746fa8",
                    "csiVolumeMutator": "quay.io/hpestorage/volume-mutator@sha256:bf6beac1af5edcd3c9774f0dcc9579b79ef9cc8e4b74a4238082ab9aa443b5ed",
                    "nfsProvisioner": "quay.io/hpestorage/nfs-provisioner@sha256:5bf5952453ecca28e26724df85ebd5c8f0b7479ae4cb683d4a09c415da00e002",
                    "nimbleCSP": "quay.io/hpestorage/alletra-6000-and-nimble-csp@sha256:5480ffdebd74fdc547bc4a11b4434abac7bbfee3cee349f1ac893fbad84bb8d1",
                    "primera3parCSP": "quay.io/hpestorage/alletra-9000-primera-and-3par-csp@sha256:8af6e87075c47ba8d9714d55848ab2a0fef62b0768681e30d9e63ed7f396b0bf"
                  },
                  "iscsi": {
                    "chapSecretName": ""
                  },
                  "kubeletRootDir": "/var/lib/kubelet",
                  "logLevel": "info",
                  "maxVolumesPerNode": 100,
                  "node": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/hpestorage/csi-driver-operator@sha256:29e92f68cf0d7691a6c56876e7dde0d69a8cb3ff6ae6c9371c604e61cba16898
          createdAt: "2025-09-25T16:28:16Z"
          description: A Container Storage Interface (CSI) driver for HPE storage
            platforms. The CSI driver allows you to use HPE storage with your preferred
            container orchestrator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <3.0.1'
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/hpe-storage/co-deployments
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE CSI Driver for Kubernetes
            displayName: HPECSIDriver
            kind: HPECSIDriver
            name: hpecsidrivers.storage.hpe.com
            version: v1
        description: |
          The HPE CSI Operator for OpenShift deploys the HPE CSI Driver on Kubernetes and OpenShift for dynamic provisioning of persistent volumes on HPE storage systems.
          The HPE CSI Driver for Kubernetes leverages HPE storage platforms to provide scalable and persistent storage for stateful applications.
          Please refer to our CSI driver [documentation](https://scod.hpedev.io/csi_driver/index.html) for supported HPE platforms and full list of supported CSI features.
          ## Installation
          Refer to the HPE CSI Operator for OpenShift [documentation](https://scod.hpedev.io/csi_driver/deployment.html#operator).
        displayName: HPE CSI Operator for OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE CSI
        - HPE Storage
        - CSI
        links:
        - name: Documentation
          url: https://scod.hpedev.io/csi_driver/index.html
        maintainers:
        - email: dev-hi-containers@hpe.com
          name: HPE Storage Containers Team
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: HPE
        relatedImages:
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - quay.io/hpestorage/volume-mutator@sha256:bf6beac1af5edcd3c9774f0dcc9579b79ef9cc8e4b74a4238082ab9aa443b5ed
        - registry.connect.redhat.com/hpestorage/csi-driver-operator@sha256:29e92f68cf0d7691a6c56876e7dde0d69a8cb3ff6ae6c9371c604e61cba16898
        - quay.io/hpestorage/alletrastoragemp-b10000-nfs-csp@sha256:6a1855a2ef1a83b886bb1f45b596e8e420e5e13c84bbf446c8c202eaf27a2013
        - quay.io/hpestorage/alletra-9000-primera-and-3par-csp@sha256:8af6e87075c47ba8d9714d55848ab2a0fef62b0768681e30d9e63ed7f396b0bf
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - quay.io/hpestorage/nfs-provisioner@sha256:5bf5952453ecca28e26724df85ebd5c8f0b7479ae4cb683d4a09c415da00e002
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8ddd178ba5d08973f1607f9b84619b58320948de494b31c9d7cd5375b316d6d4
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769
        - quay.io/hpestorage/volume-group-provisioner@sha256:0a5f39f52113b963978a695bb8e5a5a3c4d387a02eb4d4a9647b4bc804de7399
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - quay.io/hpestorage/volume-group-snapshotter@sha256:177d9abf8fbd3ac26ca4c12ca4b4adde20ebab308c79f10b04632f0171746fa8
        - quay.io/hpestorage/csi-extensions@sha256:27074aaa8df3b867cfa3d12318612c45f8e541d750c27db5e080734592160c89
        - quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676
        - quay.io/hpestorage/alletra-6000-and-nimble-csp@sha256:5480ffdebd74fdc547bc4a11b4434abac7bbfee3cee349f1ac893fbad84bb8d1
        version: 3.0.1
      entries:
      - name: hpe-csi-operator.v3.0.1
        version: 3.0.1
      name: stable
    defaultChannel: stable
    packageName: hpe-csi-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SmartX
      provider-url: https://www.smartx.com
    name: iomesh-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: iomesh-operator.0.10.1-rc4
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
          createdAt: 2021-08-17 00:00:00+00:00
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/iomesh-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/iomesh-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/iomesh/docs
          support: SmartX, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: iomesh
            displayName: iomesh
            kind: IOMeshCluster
            name: iomeshclusters.iomesh.com
            version: v1alpha1
        description: "IOMesh (https://www.iomesh.com/) is a cloud-native storage product
          specifically designed and developed for Kubernetes. It takes ZBS as its
          core, a “production-ready” distributed storage system developed independently
          by SmartX, to provide production-level high performance and reliable persistent
          storage for business-critical applications such as MySQL, Cassandra, MongoDB,
          etc., contributing to the containerized transformation of stateful applications.
          \     \n## Pre-Requisites\n* According to https://docs.iomesh.com/deploy/prerequisites,
          make sure that your openshift cluster meets the \"Installation Requirements\"
          and set up the worker nodes correctly according to the \"Setup Worker Node\"
          chapter.\n* Run IOMesh Operator pre-installation script in an environment
          where oc or kubectl can be used to access the openshift cluster, the script
          will install the dependencies of IOMesh Operator and specific IOMesh settings
          for the openshift cluster:\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-pre-install-openshift.sh
          | sh -\n```\n## Install IOMesh Operator and IOMesh Cluster\n* Click install
          button on the current web page to install IOMesh Operator\n* On the **Installed
          Operators** -> **IOMesh Operator** -> **Create instance** -> **YAML view**,
          fill in an IOMesh Custom Resources according https://iomesh.run/iomesh.yaml,
          change the `spec.*.dataCIDR` to your own data network CIDR:\n## Post Install\n*
          Run IOMesh Operator post-installation script in an environment where oc
          or kubectl can be used to access the openshift cluster, the script will
          install IOMesh CSI Driver\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-post-install-openshift.sh
          | sh -\n```\n## Getting Started\n* Try to quickstart setup and use IOMesh:
          https://docs.iomesh.com/deploy/setup-iomesh\n"
        displayName: IOMesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iomesh-operator
        links:
        - name: IOMesh Operator
          url: https://iomesh.com
        maintainers:
        - email: na.wang@smartx.com
          name: Na Wang
        maturity: alpha
        provider:
          name: SmartX
          url: https://www.smartx.com
        relatedImages:
        - iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
        version: 0.10.1-rc4
      entries:
      - name: iomesh-operator.0.10.1-rc4
        version: 0.10.1-rc4
      name: stable
    defaultChannel: stable
    packageName: iomesh-operator-rhmp
    provider:
      name: SmartX
      url: https://www.smartx.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: openstack-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openstack-operator.v1.0.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openstack.org/v1beta1",
                "kind": "OpenStack",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openstack-operator",
                    "app.kubernetes.io/instance": "openstack",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "openstack",
                    "app.kubernetes.io/part-of": "openstack-operator"
                  },
                  "name": "openstack",
                  "namespace": "openstack-operators"
                },
                "spec": {
                  "operatorOverrides": [
                    {
                      "controllerManager": {
                        "resources": {
                          "limits": {
                            "cpu": "600m",
                            "memory": "2Gi"
                          }
                        }
                      },
                      "name": "infra",
                      "replicas": 1
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-02T13:33:09Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.14
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.openstack.org/v1beta1","kind":"OpenStack","metadata":{"name":"openstack","namespace":"openstack-operators"},"spec":{}}'
          operatorframework.io/suggested-namespace: openstack-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenStack is the Schema for the openstacks API
            displayName: Open Stack
            kind: OpenStack
            name: openstacks.operator.openstack.org
            version: v1beta1
        description: Install and configure OpenStack
        displayName: OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        links:
        - name: Openstack Operator
          url: https://github.com/openstack-k8s-operators/
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/glance-rhel9-operator@sha256:ee22cf417b7771930a442c2302277e6d37f368f5ae551fee9a50dbbe41c8fe3a
        - registry.redhat.io/rhoso-operators/mysqld-exporter-rhel9@sha256:b49e966cdd588071e0e1f2791ec6c62310078a7d39ea77f83843d3708a07c2b5
        - registry.redhat.io/rhoso/openstack-swift-proxy-server-rhel9@sha256:858551fd1c519797cae0e146e9c05412ed6e66dac01d0e68e62dd2d4bd335391
        - registry.redhat.io/rhoso/openstack-netutils-rhel9@sha256:b02ab1617922a2560bf739172bc2ce0f6b1fb3ef55054bd17e4b16bdf9512b4b
        - registry.redhat.io/rhoso/openstack-cinder-backup-rhel9@sha256:910693233ee3a1f3832ebe744f67a5139249523456de4abbd731c9d2f544b7c3
        - registry.redhat.io/rhoso/ironic-python-agent-rhel9@sha256:2ccd8e22a242059ca5c4e2bce853c36a71c36fdf40e8e36ab183a8218b0cc87b
        - registry.redhat.io/rhoso/openstack-frr-rhel9@sha256:f6516277149d1560c0bf2ac74a99cfb4ac2c84f9f090ab231b6d12bdb824de3b
        - registry.redhat.io/rhoso/openstack-octavia-housekeeping-rhel9@sha256:0a79259e398af746d0e8a6dda1aae5ee976c6bf8991b17f4a5da0d07dca1e36c
        - registry.redhat.io/rhoso/openstack-swift-object-rhel9@sha256:8fc2ab8caa9c7c35d94ef51c303763ea71e40b9c795c4555c17e6194344419b3
        - registry.redhat.io/ubi9/httpd-24@sha256:29382518403a4361c48fccc057163dc4a9f43ee159432754fa63c76485b223f5
        - registry.redhat.io/rhoso-operators/heat-rhel9-operator@sha256:583abb81a49db85949f7a4b846591ad8532afa79440ad7e2937ee00eacd972b3
        - registry.redhat.io/rhoso/openstack-barbican-keystone-listener-rhel9@sha256:cd8d4f0a94b4fe080ebf8e459f091db6637ceec6d83ac23cf04d66bd03c71e76
        - registry.redhat.io/rhoso/openstack-neutron-metadata-agent-ovn-rhel9@sha256:20323b151474703df610b6c24ccb9c748265196bbc70a7b5ece1b801ff540cd9
        - registry.redhat.io/rhoso-operators/test-rhel9-operator@sha256:df7566483abb18aa3c45d8bc4cd91531c7eb5b63345e511151a7cdca2dbcef17
        - registry.redhat.io/rhoso/openstack-iscsid-rhel9@sha256:508bd67021c271a10bb55ab4af033dc4e931708674b9c5ba94f72c07cd0c8ee2
        - registry.redhat.io/rhoso-operators/openstack-baremetal-agent-rhel9@sha256:08675b16d89d7931b1ef7b96e1e884c88e8e7fbfbe79a944f9213c52aba3ba83
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/rhoso/openstack-ironic-neutron-agent-rhel9@sha256:0abaa8f437025d2e7006c59c546ff6935556d707b57993719faef94142e1c43d
        - registry.redhat.io/rhoso/openstack-aodh-evaluator-rhel9@sha256:48c9192acf157df5b8f0827e35f0380d50cd4562676aba3c0d7791f92328af02
        - registry.redhat.io/rhoso/openstack-manila-scheduler-rhel9@sha256:8ce47f7fd8660167b3b292659d4fc58cf34ed671d5779693d24b71a9fddc1e7c
        - registry.redhat.io/rhoso/openstack-neutron-dhcp-agent-rhel9@sha256:cc9223ef705bcd5d0dcabbf71ad9f6e540ae7f003a6afd13161301f842dfa5fa
        - registry.redhat.io/rhoso/edpm-hardened-uefi-rhel9@sha256:c18ec7023fbcc4ab5183f689adc1f01a626a4923c15c61a37388f99904db7bc3
        - registry.redhat.io/rhoso/openstack-ceilometer-central-rhel9@sha256:e6d9f47406b7e3ff2a7f2256088cc64f043b8e497382b9f569eb90182598bbe9
        - registry.redhat.io/rhoso/openstack-cinder-api-rhel9@sha256:72a674caeeeecce3b60217f3607ad5b00651fa3d52f2f5d905f6e6395638dbc8
        - registry.redhat.io/rhoso/openstack-glance-api-rhel9@sha256:b1c0b09bb9aafe3bc31f43f8a99a92dde8d78cf90e781dd14300580986d8c980
        - registry.redhat.io/rhoso-operators/designate-rhel9-operator@sha256:d446af622bd32666eb4eb5f9b7f38f0c33ac81251a2230db9d5baf71081b9d6a
        - registry.redhat.io/rhoso/openstack-designate-producer-rhel9@sha256:04af4230f815a6e3fde22003cf382eebd9cbc3c3ec4056795f4474abc1e65fea
        - registry.redhat.io/rhoso-operators/mariadb-rhel9-operator@sha256:907e3a9d15941440d5289d7e448d9f0351053fc20fcb8cb3b9db0a84debd876f
        - registry.redhat.io/rhoso/openstack-manila-api-rhel9@sha256:e4dd3d5b66c5216b45a1515a975b85fd2ac4ef698a9350dd16e01417da462b32
        - registry.redhat.io/rhoso/openstack-cinder-volume-rhel9@sha256:8a09e80dac94fb4b3d35033e91812f95f0acac66532d9abc88aa1fb715d21822
        - registry.redhat.io/rhoso/openstack-designate-worker-rhel9@sha256:a6324a8a72722822df7692b8df0420388a1b4aecfff3da7e91da4afe4d9e2260
        - registry.redhat.io/rhoso/openstack-ceilometer-compute-rhel9@sha256:296f7f6c6bdcd7f5d07b8470308a4c71055f29a17c48d2ebb03b7f54a0e927ff
        - registry.redhat.io/rhoso/openstack-heat-api-cfn-rhel9@sha256:2789929ae3edd7c68579f34457d3071273e49a66f462043cd9c46e75bdda282a
        - registry.redhat.io/openshift4/ose-kube-state-metrics-rhel9@sha256:db6074b8aeedfd7603ae3adbe67071115f54e7f01a24997f1cd8fa8958aeddce
        - registry.redhat.io/rhoso-operators/cinder-rhel9-operator@sha256:12b8dd8fd3683250365eff6f07339cb976ee29da80642e591c10e9fa5ba627c9
        - registry.redhat.io/rhoso/openstack-heat-engine-rhel9@sha256:151e04b3e5a408aadef3a17954255acc3733d09961878a135e177188457447f0
        - registry.redhat.io/rhoso/openstack-rsyslog-rhel9@sha256:272e729accc2b64b8fcc0ece4ae5aa9ce6d454e34a25a8a8e2b99ac56b73361e
        - registry.redhat.io/rhoso/openstack-tempest-all-rhel9@sha256:16b1fbfd36ce9a27106bf09c414eaa93f14c9a6b11839068976c2fdb537d2ba7
        - registry.redhat.io/rhoso/openstack-octavia-api-rhel9@sha256:7b8d5927162ddab512096beb02e837ba51f6353c5187f3de72eb3db6e56ab6e7
        - registry.redhat.io/rhoso/openstack-keystone-rhel9@sha256:ac076c81247d774a9aeee1b63540ca6e18058213cf6aad10a6fd1f535a394ef4
        - registry.redhat.io/rhoso/openstack-cinder-scheduler-rhel9@sha256:2079ffe9b620e9277573ac796bf7e393b520631130e5bbb6f5ab51d872c943f3
        - registry.redhat.io/rhoso/openstack-aodh-notifier-rhel9@sha256:d2e89f27aac9afa504411ae6a6163442cd717f8c560e505a7da8e3fd9e2cff95
        - registry.redhat.io/rhoso/openstack-manila-share-rhel9@sha256:277174c885998e6f5b867d099664f6fa28f974bffb524903ff75283035cd8dc9
        - registry.redhat.io/rhoso/openstack-neutron-server-rhel9@sha256:2d3afde3201aa18b192ad0fb079c1e4f4f673565604f1af2b79aba8a7bc88bf2
        - registry.redhat.io/rhoso/openstack-ovn-northd-rhel9@sha256:668f742b2a18579bc8cfc610a862e43c5b877413cf65eabbe0ee36b3c7869446
        - registry.redhat.io/rhoso-operators/openstack-baremetal-rhel9-operator@sha256:03916e507b88943c0e00cedeb3cf20fa0787ccc762e47f38d305ff4d1b125b12
        - registry.redhat.io/rhoso/openstack-ovn-nb-db-server-rhel9@sha256:8c2ce1027779dc66d67a0e0d03e48b226d3278208cae0b2b81b35ad1b1f3ae72
        - registry.redhat.io/rhoso/openstack-neutron-ovn-agent-rhel9@sha256:bed1956c66b0e70064b51e34faac64036f77c7ecb14f46ceac4bc9ef52f0b9e0
        - registry.redhat.io/rhoso/openstack-nova-scheduler-rhel9@sha256:f0e5d3a693c9998ca3839b3ec7dfd5da0f9cf139e0b5b6913e92a10d5a13e6bc
        - registry.redhat.io/rhoso/openstack-designate-central-rhel9@sha256:d28d8d91d8bf75a1b77a5d650c7f2650a1791e2a1b0341450fa46978c955b92f
        - registry.redhat.io/rhoso/openstack-barbican-api-rhel9@sha256:bc942204843c45c6cb7ea841af8611f04efb003de7eeccc6813934aad556da73
        - registry.redhat.io/rhoso/openstack-octavia-health-manager-rhel9@sha256:8065851058019757b56f86ca4f030d44f169100eb2e37e08e0587a7b1e4d2e72
        - registry.redhat.io/rhoso/openstack-nova-api-rhel9@sha256:4116206f1fbe60fdd9458a5edee848571c94e2b0512d9c2cfcfd6890d60d4395
        - registry.redhat.io/rhoso-operators/keystone-rhel9-operator@sha256:879d4b1f9e720f8581b3a55c788a97bbe7b52f9108df3d5dce2e9bcb4200df6d
        - registry.redhat.io/rhoso/openstack-horizon-rhel9@sha256:82d0e6d24733afd305996e0f4a118df3e48621c40c9669b01f48669dbf40d868
        - registry.redhat.io/rhoso/openstack-designate-api-rhel9@sha256:b76ca7820553285bb4ee2bfb08e39aedbfe8c16b4edb478c06418a1e335a64ab
        - registry.redhat.io/rhoso/openstack-swift-container-rhel9@sha256:95fec49643fc18f402b8ba8057f2c35f214bf6138c5f0d7bbe1b96cbd78ef8d6
        - registry.redhat.io/rhoso-operators/swift-rhel9-operator@sha256:c99ef14338290d63e2bcbf129e953b434aea382e47072812bae31b1ea838e7df
        - registry.redhat.io/rhoso/openstack-ovn-base-rhel9@sha256:a3b8ec7ad1073facb8876acc787549a1a68b9bf81bebdc127f13eb6385e2d1da
        - registry.redhat.io/rhoso/openstack-ovn-bgp-agent-rhel9@sha256:a404c3416ab65e7134f438362e887529be474639e725e18f3542e86973edac87
        - registry.redhat.io/rhoso/openstack-nova-novncproxy-rhel9@sha256:04881867f9b24403bf6ef051eee3902a17e8476eee894ef579c0d5d130d2510d
        - registry.redhat.io/rhoso/openstack-ironic-inspector-rhel9@sha256:88357e83e8a4bbc3c5a37f0420ad75556821834a975cd8a1cf5fee56de2a4f99
        - registry.redhat.io/rhoso/openstack-ovn-controller-rhel9@sha256:3c42b7e0305c508d74602b9378143deb07a569443bba35850cdb7b6e3f5202e2
        - registry.redhat.io/rhoso/openstack-nova-conductor-rhel9@sha256:c67ca0e6c8126f56c6a83632b6a96bd4a35006f5f414a8779467b81864b31d12
        - registry.redhat.io/rhoso-operators/ee-openstack-ansible-ee-rhel9@sha256:61e729fb89c4872837cbe9cb1e3cea402b61f0da240751302605df487721f0aa
        - registry.redhat.io/rhoso-operators/ovn-rhel9-operator@sha256:66cd8add6078b63a36879c974db9e9f358593bb0651aa39fe85153d25404ae27
        - registry.redhat.io/rhoso/openstack-designate-backend-bind9-rhel9@sha256:82cca999357b0ab25bcc081bf802fab831587cf661db806af4ca453e0a626743
        - registry.redhat.io/rhoso/openstack-rabbitmq-rhel9@sha256:56bc2d2b201225fb98db43ea9ca68cceb4842f8f0f1a96ab3ab62ab785651d12
        - registry.redhat.io/rhoso-operators/prometheus-podman-exporter-rhel9@sha256:f33e7506db6dfdf530d3e2442baaf360d212bb1a47a7db887dba967ed0655cc2
        - registry.redhat.io/rhoso/openstack-nova-compute-rhel9@sha256:6264075310231af358803b53d5a09250dab32d63fc5966aa8fb957e5f56096a8
        - registry.redhat.io/rhoso-operators/neutron-rhel9-operator@sha256:30b6a812c47db9e6691211185c304c3e8fc6a8b9ba7ca79acf59c5ec07fb9fb0
        - registry.redhat.io/rhoso-operators/openstack-must-gather-rhel9@sha256:60b4355f4dbef3ad64103357086b10a5c0af126dd23798d6a9d725646f1e13b4
        - registry.redhat.io/rhoso/openstack-ironic-api-rhel9@sha256:50ccfe7d9ba22a50e0c1c8159cade6f4b1ac13ad8f5770f20c7be18fba58c8c3
        - registry.redhat.io/rhoso/openstack-ironic-pxe-rhel9@sha256:42e35ded74033f1f07287441917829cc934930a5df696c33cd8957be3c10fa41
        - registry.redhat.io/rhoso/openstack-openstackclient-rhel9@sha256:2952521d4b1799fd32ff773ef2d457a3abc354b639cf7f1e0f7f6615ca35ac8f
        - registry.redhat.io/rhoso-operators/manila-rhel9-operator@sha256:d9e69eb140b8ff76d5bf49deab8d700639ccdfd1ab5c8c7667bc2bbb50f3d60b
        - registry.redhat.io/rhoso-operators/horizon-rhel9-operator@sha256:9a2a13635588c6cc9949b272e1a1708c5fb53f19b05b8310ace703df5958ac63
        - registry.redhat.io/rhoso-operators/nova-rhel9-operator@sha256:3156dde249e49b40bed71092fb5213b112d43b663ba8c556e355aed524bec822
        - registry.redhat.io/rhoso-operators/ironic-rhel9-operator@sha256:dd67ffe3578ce619748735c1a5cd2a57d704d0c36c1b841d513ea36479aa4682
        - registry.redhat.io/rhoso/openstack-aodh-api-rhel9@sha256:4adc09cc1d922b387320da467de89d3f41a9c49b46d9b728458fec36505ec940
        - registry.redhat.io/rhoso/openstack-cron-rhel9@sha256:f09e724c69cec2c3b685783f9263b26c280a57b5909a723341416b342d18f9a4
        - registry.redhat.io/rhoso/openstack-ironic-conductor-rhel9@sha256:538da9fdeec546d794d192361d0a76f8b4e7682b6eb6d9cea371d8a52d603800
        - registry.redhat.io/rhoso/openstack-swift-account-rhel9@sha256:31e727f37db92fcc21319c44396576a03a118ba25d403715650162f7c0ead31e
        - registry.redhat.io/rhoso/openstack-redis-rhel9@sha256:c6dbeacbbd5aed9d13e540f1c987a574ea45747983c280446e5afadcf7c33b37
        - registry.redhat.io/rhoso/openstack-ovn-sb-db-server-rhel9@sha256:8ba4f8ed49e214f421e94133da0f17b54ce99a52bbbd367d0aef357de844ea01
        - registry.redhat.io/openshift4/ose-prometheus-node-exporter-rhel9@sha256:479c51de2fc8a3b3858d71bbe4eeafbd5b43263472346f9dde348115082894cc
        - registry.redhat.io/rhoso-operators/octavia-rhel9-operator@sha256:aa0f00f8e03dfdb547e53cf611581e9cc091eaf049558121c2bc5d702e556bcb
        - registry.redhat.io/rhoso/openstack-mariadb-rhel9@sha256:6115a7458cfa412c1659242fcec64b63f0a5ca414ecf3c5f890f0afec0bca069
        - registry.redhat.io/rhoso/openstack-multipathd-rhel9@sha256:6ff8ae176f8a633bcd78b1f8407924c74bb2f317eba16b90e7f0f9e53af74e8e
        - registry.redhat.io/rhoso/openstack-neutron-sriov-agent-rhel9@sha256:118d5f9430192ab48cf159ee2f7912f08e935ad56ef3364f18b3680210bb3e52
        - registry.redhat.io/rhoso-operators/barbican-rhel9-operator@sha256:b3b769e0dc32c12c5cc9578c2cac2622b973177ca14697bc1042d21d473e4e84
        - registry.redhat.io/rhoso/openstack-heat-api-rhel9@sha256:ccbc41b552bb34d223c9615a4e3b56bb516bc5bd8a694f048c15063ba45fca9f
        - registry.redhat.io/rhoso-operators/telemetry-rhel9-operator@sha256:0635515a5fed12f17de4cc5ac5f9567e3b3546fc2f69730c7beab25cf722f538
        - registry.redhat.io/rhoso/openstack-aodh-listener-rhel9@sha256:037845be6787f610e479345206b0fcfb2f1dc430540816bc5a54a1d26da2acbd
        - registry.redhat.io/rhoso/openstack-ceilometer-notification-rhel9@sha256:bd2b5f50a2bef4e77e3daa89aca51df6e0e55c3784886296d92c97616cd566d1
        - registry.redhat.io/rhoso-operators/sg-core-rhel9@sha256:b56923f869ff57396e6822133dfc3b21007c3df652cfe962c7ad3028214d988c
        - registry.redhat.io/rhoso/openstack-barbican-worker-rhel9@sha256:485525116fb3c22d8c60330dd87ebd3854cd71ded36bf88bcf18cee5957e22a3
        - registry.redhat.io/rhoso-operators/openstack-network-exporter-rhel9@sha256:8e8e775e37c28cb11b7aae9199405b85d5862dce195d165ca598fa1dc89bf914
        - registry.redhat.io/rhoso/openstack-placement-api-rhel9@sha256:0aa5e138947bf236f31c81bfb6f641d95af63812c06dca43141721b383c7e08e
        - registry.redhat.io/openshift-power-monitoring/kepler-rhel9@sha256:8e9e5bfeed61aaeee85bf5eefc3390556b8d4cbfc797a82cabe95bc0367126eb
        - registry.redhat.io/rhoso/openstack-ceilometer-ipmi-rhel9@sha256:2fb5723c68d0f2c3ae28b86e606ccae4b80f3bd9b16c01dd6c9adcae629db7f1
        - registry.redhat.io/rhoso-operators/openstack-rhel9-operator@sha256:64e2ad312f757101f083eedcb17e0b6c05cf6ebf6ed9e7db2e33e84c6aca9898
        - registry.redhat.io/rhoso/openstack-unbound-rhel9@sha256:7f17a719e8e49d457ff16437709c2b4eca5b9c2f58972d1e351c5c1c1ac38962
        - registry.redhat.io/rhoso-operators/infra-rhel9-operator@sha256:a2dc031a6ad2726d155e77cafb676f8e35a86e963b97ccdceed0da137e67ba0d
        - registry.redhat.io/rhoso/openstack-designate-mdns-rhel9@sha256:ce43dd1a30afd4753aa11d32df2b24ed4bf3d67847044aaae8f3e5c8901caf05
        - registry.redhat.io/rhoso/openstack-octavia-worker-rhel9@sha256:303e35926e43fe403673e18a25c2e815547c28a5f8c0f09d470cd4148a21c528
        - registry.redhat.io/rhoso-operators/placement-rhel9-operator@sha256:a16bd4fb929eb5bcf96f2aefcb5a7abf4143f4f9ed24e53de93f4a3caa89a507
        - registry.redhat.io/rhoso-operators/rabbitmq-cluster-rhel9-operator@sha256:e095e6cbda841b9c726c1ec0839a98964b894cf25ae9228052d314b1ee2f161d
        - registry.redhat.io/rhoso/openstack-memcached-rhel9@sha256:3e704b9664c65be39e83c2ed802722ffeaf969d8e571147a2be128d5c56331c1
        version: 1.0.14
      entries:
      - name: openstack-operator.v1.0.14
        version: 1.0.14
      - name: openstack-operator.v1.0.13
        version: 1.0.13
      - name: openstack-operator.v1.0.12
        version: 1.0.12
      - name: openstack-operator.v1.0.11
        version: 1.0.11
      - name: openstack-operator.v1.0.10
        version: 1.0.10
      - name: openstack-operator.v1.0.9
        version: 1.0.9
      - name: openstack-operator.v1.0.8
        version: 1.0.8
      - name: openstack-operator.v1.0.7
        version: 1.0.7
      - name: openstack-operator.v1.0.6
        version: 1.0.6
      - name: openstack-operator.v1.0.5
        version: 1.0.5
      - name: openstack-operator.v1.0.4
        version: 1.0.4
      - name: openstack-operator.v1.0.3
        version: 1.0.3
      - name: openstack-operator.v1.0.2
        version: 1.0.2
      - name: openstack-operator.v1.0.1
        version: 1.0.1
      - name: openstack-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: openstack-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: IBM
      provider-url: ""
    name: ibm-block-csi-operator-community
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12
    - currentCSV: ibm-block-csi-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
        version: 1.12.0
      entries:
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.0
    - currentCSV: ibm-block-csi-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
        version: 1.12.1
      entries:
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.1
    - currentCSV: ibm-block-csi-operator.v1.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:3fc8df547138a2009225d63c2a3111a672340bdf9f3cb24a889cbb69f2c33c2d
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:3fc8df547138a2009225d63c2a3111a672340bdf9f3cb24a889cbb69f2c33c2d
        version: 1.12.2
      entries:
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.2
    - currentCSV: ibm-block-csi-operator.v1.12.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:2bd461a1e40a1366c871585aa7a4e8039e686a66bf7192efdeb1c215209a071d
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:2bd461a1e40a1366c871585aa7a4e8039e686a66bf7192efdeb1c215209a071d
        version: 1.12.3
      entries:
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.3
    - currentCSV: ibm-block-csi-operator.v1.12.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e29db476de3778222bf049926223dee20f6bc34e55a04df5f0f3df38741ff23e
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e29db476de3778222bf049926223dee20f6bc34e55a04df5f0f3df38741ff23e
        version: 1.12.4
      entries:
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.4
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.5
    defaultChannel: stable
    packageName: ibm-block-csi-operator-community
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: ""
    name: pmem-csi-operator-os
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pmem-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pmem-csi.intel.com/v1beta1",
                "kind": "PmemCSIDeployment",
                "metadata": {
                  "name": "pmem-csi.intel.com"
                },
                "spec": {
                  "deviceMode": "lvm",
                  "nodeSelector": {
                    "feature.node.kubernetes.io/memory-nv.dax": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
          createdAt: "2021-08-26T10:45:23Z"
          description: An operator for deploying and managing the PMEM-CSI driver
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: ""
          repository: https://github.com/intel/pmem-csi/tree/release-1.0
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a PMEM-CSI driver deployment in the cluster.
            displayName: PmemCSIDeployment
            kind: PmemCSIDeployment
            name: pmemcsideployments.pmem-csi.intel.com
            version: v1beta1
        description: |2

          Intel PMEM-CSI is a [CSI](https://github.com/container-storage-interface/spec)
          storage driver for container orchestrators like Kubernetes. It makes local
          persistent memory ([PMEM](https://pmem.io/)) available as a filesystem volume to
          container applications. This is the operator to deploy and manage the
          [PMEM-CSI](https://intel.github.io/pmem-csi/1.0/README.html)
          driver on a Kubernetes cluster.

          Refer to the [PMEM-CSI documentation](https://intel.github.io/pmem-csi/1.0/README.html)
          for more details on deploying and using the PMEM-CSI driver.
        displayName: Operator for PMEM-CSI driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - PMEM-CSI
        - PMEM-CSI operator
        - Persistent Memory
        - PMEM
        links:
        - name: PMEM-CSI
          url: https://intel.github.io/pmem-csi/1.0
        maintainers:
        - email: olev.kartau@intel.com
          name: Olev Kartau
        - email: patrick.ohly@intel.com
          name: Patrick Ohly
        - email: amarnath.valluri@intel.com
          name: Amarnath Valluri
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Intel® Corporation
        relatedImages:
        - registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
        version: 1.0.0
      entries:
      - name: pmem-csi-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: pmem-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pmem-csi.intel.com/v1beta1",
                "kind": "PmemCSIDeployment",
                "metadata": {
                  "name": "pmem-csi.intel.com"
                },
                "spec": {
                  "deviceMode": "lvm",
                  "nodeSelector": {
                    "feature.node.kubernetes.io/memory-nv.dax": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
          createdAt: "2021-08-26T10:45:23Z"
          description: An operator for deploying and managing the PMEM-CSI driver
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: ""
          repository: https://github.com/intel/pmem-csi/tree/release-1.0
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a PMEM-CSI driver deployment in the cluster.
            displayName: PmemCSIDeployment
            kind: PmemCSIDeployment
            name: pmemcsideployments.pmem-csi.intel.com
            version: v1beta1
        description: |2

          Intel PMEM-CSI is a [CSI](https://github.com/container-storage-interface/spec)
          storage driver for container orchestrators like Kubernetes. It makes local
          persistent memory ([PMEM](https://pmem.io/)) available as a filesystem volume to
          container applications. This is the operator to deploy and manage the
          [PMEM-CSI](https://intel.github.io/pmem-csi/1.0/README.html)
          driver on a Kubernetes cluster.

          Refer to the [PMEM-CSI documentation](https://intel.github.io/pmem-csi/1.0/README.html)
          for more details on deploying and using the PMEM-CSI driver.
        displayName: Operator for PMEM-CSI driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - PMEM-CSI
        - PMEM-CSI operator
        - Persistent Memory
        - PMEM
        links:
        - name: PMEM-CSI
          url: https://intel.github.io/pmem-csi/1.0
        maintainers:
        - email: olev.kartau@intel.com
          name: Olev Kartau
        - email: patrick.ohly@intel.com
          name: Patrick Ohly
        - email: amarnath.valluri@intel.com
          name: Amarnath Valluri
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Intel® Corporation
        relatedImages:
        - registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
        version: 1.0.0
      entries:
      - name: pmem-csi-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pmem-csi-operator-os
    provider:
      name: Intel® Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OceanBase
      provider-url: https://oceanbase.com
    name: oceanbase-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oceanbase-operator.v2.2.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Full Lifecycle
          createdAt: "2024-08-14T16:21:06Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: OBClusterOperation
            name: obclusteroperations.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBCluster
            name: obclusters.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBParameter
            name: obparameters.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBResourceRescue
            name: obresourcerescues.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBServer
            name: observers.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantBackupPolicy
            name: obtenantbackuppolicies.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantBackup
            name: obtenantbackups.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantOperation
            name: obtenantoperations.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantRestore
            name: obtenantrestores.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenant
            name: obtenants.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBZone
            name: obzones.oceanbase.oceanbase.com
            version: v1alpha1
        description: Ob Operator description. TODO.
        displayName: Ob Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - oceanbase-operator
        - ob-operator
        links:
        - name: OceanBase Operator
          url: https://oceanbase.github.io/ob-operator
        maintainers:
        - email: hongxin.sun.vip@gmail.com
          name: Hongxin Sun
        maturity: alpha
        provider:
          name: OceanBase
          url: https://oceanbase.com
        relatedImages:
        - quay.io/oceanbasedev/oceanbase-operator@sha256:b11fe60bb8de545d3ed105acc50f2d6c5d6aa8093f3ae4ca7ebce58b1e5e7b48
        - quay.io/brancz/kube-rbac-proxy@sha256:18f12fc3a0085bbc303cfbd2ad1151127c1c959c4641e36027547b6b66eb4c92
        version: 2.2.2
      entries:
      - name: oceanbase-operator.v2.2.2
        version: 2.2.2
      name: alpha
    - currentCSV: oceanbase-operator.v2.2.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Full Lifecycle
          createdAt: "2024-08-14T16:21:06Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: OBClusterOperation
            name: obclusteroperations.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBCluster
            name: obclusters.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBParameter
            name: obparameters.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBResourceRescue
            name: obresourcerescues.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBServer
            name: observers.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantBackupPolicy
            name: obtenantbackuppolicies.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantBackup
            name: obtenantbackups.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantOperation
            name: obtenantoperations.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantRestore
            name: obtenantrestores.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenant
            name: obtenants.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBZone
            name: obzones.oceanbase.oceanbase.com
            version: v1alpha1
        description: Ob Operator description. TODO.
        displayName: Ob Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - oceanbase-operator
        - ob-operator
        links:
        - name: OceanBase Operator
          url: https://oceanbase.github.io/ob-operator
        maintainers:
        - email: hongxin.sun.vip@gmail.com
          name: Hongxin Sun
        maturity: alpha
        provider:
          name: OceanBase
          url: https://oceanbase.com
        relatedImages:
        - quay.io/oceanbasedev/oceanbase-operator@sha256:b11fe60bb8de545d3ed105acc50f2d6c5d6aa8093f3ae4ca7ebce58b1e5e7b48
        - quay.io/brancz/kube-rbac-proxy@sha256:18f12fc3a0085bbc303cfbd2ad1151127c1c959c4641e36027547b6b66eb4c92
        version: 2.2.2
      entries:
      - name: oceanbase-operator.v2.2.2
        version: 2.2.2
      name: stable
    defaultChannel: stable
    packageName: oceanbase-operator
    provider:
      name: OceanBase
      url: https://oceanbase.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.2.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"customImageNames":
            false,"externalArangoDBName": "arangodb","repository": "registry.connect.redhat.com/turbonomic","tag":"8.3.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/turbonomic/t8c-operator@sha256:d5164e6dd24a0feff10cab62530cafff1bc92d0e8ec2d16c6f4d880c6e2f50bd
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/t8c-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/t8c-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/turbonomic/t8c-operator@sha256:d5164e6dd24a0feff10cab62530cafff1bc92d0e8ec2d16c6f4d880c6e2f50bd
        version: 42.2.0
      entries:
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: stable
    defaultChannel: stable
    packageName: t8c-certified-rhmp
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ptp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ptp-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "NodePtpDevice",
                "metadata": {
                  "name": "node.example.com"
                },
                "spec": {}
              },
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "PtpConfig",
                "metadata": {
                  "name": "example-ptpconfig"
                },
                "spec": {
                  "profile": [
                    {
                      "interface": "ens787f1",
                      "name": "profile1",
                      "phc2sysOpts": "-a -r",
                      "ptp4lOpts": "-s -2",
                      "ptpClockThreshold": {}
                    }
                  ],
                  "recommend": [
                    {
                      "match": [
                        {
                          "nodeLabel": "node-role.kubernetes.io/worker=",
                          "nodeName": "node.example.com"
                        }
                      ],
                      "priority": 4,
                      "profile": "profile1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "PtpOperatorConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "daemonNodeSelector": {
                    "feature.node.kubernetes.io/ptp-capable": "yes",
                    "node-role.kubernetes.io/worker": ""
                  },
                  "ptpEventConfig": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-ptp-rhel9-operator@sha256:012334e6624d5440439824454bda1d1dc79804938174232a3b470176bf462b5e
          createdAt: "2024-03-13"
          description: This software enables configuration of Precision Time Protocol(PTP)
            on Kubernetes. It detects hardware capable PTP devices on each node, and
            configures linuxptp processes such as ptp4l, phc2sys and timemaster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.16.0-202509240123'
          operatorframework.io/suggested-namespace: openshift-ptp
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/ptp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodePtpDevice is the Schema for the nodeptpdevices API
            displayName: Node Ptp Device
            kind: NodePtpDevice
            name: nodeptpdevices.ptp.openshift.io
            version: v1
          - description: PtpConfig is the Schema for the ptpconfigs API
            displayName: Ptp Config
            kind: PtpConfig
            name: ptpconfigs.ptp.openshift.io
            version: v1
          - description: PtpOperatorConfig is the Schema for the ptpoperatorconfigs
              API
            displayName: Ptp Operator Config
            kind: PtpOperatorConfig
            name: ptpoperatorconfigs.ptp.openshift.io
            version: v1
        description: |
          # Precision Time Protocol (PTP) Operator for Openshift

          ## Introdution
          PTP Operator manages cluster wide PTP configuration.
          This operator has to run in namespace 'openshift-ptp'. An Operator Group is also required to install this operator:
          ```
          $ oc create -f - <<EOF
          apiVersion: v1
          kind: Namespace
          metadata:
            name: openshift-ptp
            labels:
              name: openshift-ptp
              openshift.io/cluster-monitoring: "true"
          EOF

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: ptp-operators
            namespace: openshift-ptp
          spec:
            targetNamespaces:
            - openshift-ptp
          EOF
          ```
        displayName: PTP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ptp
        links:
        - name: Documentation operator
          url: https://github.com/openshift/ptp-operator
        - name: Documentation operand
          url: https://github.com/openshift/linuxptp-daemon
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-cloud-event-proxy-rhel9@sha256:3fa2aad8cf026eba586645c851f07a1d752c85b30795fdcf037b138ef3bbb6aa
        - registry.redhat.io/openshift4/ose-ptp-rhel9-operator@sha256:012334e6624d5440439824454bda1d1dc79804938174232a3b470176bf462b5e
        - registry.redhat.io/openshift4/ose-ptp-rhel9@sha256:93b4717337b551603cbb6147d66d7b70a5580332953ddf42304f7092d8739c83
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        version: 4.16.0-202509240123
      entries:
      - name: ptp-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: ptp-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: ptp-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: ptp-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: ptp-operator.v4.16.0-202506201205
        version: 4.16.0-202506201205
      - name: ptp-operator.v4.16.0-202505291140
        version: 4.16.0-202505291140
      - name: ptp-operator.v4.16.0-202505161704
        version: 4.16.0-202505161704
      - name: ptp-operator.v4.16.0-202505020146
        version: 4.16.0-202505020146
      - name: ptp-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: ptp-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: ptp-operator.v4.16.0-202502260004
        version: 4.16.0-202502260004
      - name: ptp-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: ptp-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: ptp-operator.v4.16.0-202501201505
        version: 4.16.0-202501201505
      - name: ptp-operator.v4.16.0-202501072135
        version: 4.16.0-202501072135
      - name: ptp-operator.v4.16.0-202412311307
        version: 4.16.0-202412311307
      - name: ptp-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: ptp-operator.v4.16.0-202412102004
        version: 4.16.0-202412102004
      - name: ptp-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: ptp-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: ptp-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: ptp-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: ptp-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: ptp-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: ptp-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: ptp-operator.v4.16.0-202409131735
        version: 4.16.0-202409131735
      - name: ptp-operator.v4.16.0-202409062036
        version: 4.16.0-202409062036
      - name: ptp-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: ptp-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: ptp-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: ptp-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: ptp-operator.v4.16.0-202407311708
        version: 4.16.0-202407311708
      - name: ptp-operator.v4.16.0-202407301837
        version: 4.16.0-202407301837
      - name: ptp-operator.v4.16.0-202407160236
        version: 4.16.0-202407160236
      - name: ptp-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: ptp-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: ptp-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: ptp-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: ptp-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: ptp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fossul Framework
      provider-url: https://github.com/fossul/fossul
    name: fossul-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fossul-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n    \"kind\":
            \"Backup\",\n    \"metadata\": {\n      \"name\": \"mariadb-sample\"\n
            \   },\n    \"spec\": {\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\"\n    }\n  },\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n    \"kind\":
            \"Restore\",\n    \"metadata\": {\n      \"name\": \"mariadb-sample\"\n
            \   },\n    \"spec\": {\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\",\n      \"workflow_id\": \"xxxx\"\n    }\n  },        \n  {\n
            \   \"apiVersion\": \"fossul.io/v1\",\n    \"kind\": \"BackupConfig\",\n
            \   \"metadata\": {\n      \"name\": \"mariadb-sample\"\n    },\n    \"spec\":
            {\n      \"container_platform\": \"openshift\",\n      \"operator_controlled\":
            true,\n      \"auto_discovery\": false,\n      \"deployment_name\": \"mariadb\",\n
            \     \"deployment_type\": \"DeploymentConfig\",\n      \"job_retention\":
            50,\n      \"overwrite_pcv_on_restore\": true,\n      \"policies\": [\n
            \       {\n          \"policy\": \"hourly\",\n          \"retentionNumber\":
            3\n        },\n        {\n          \"policy\": \"daily\",\n          \"retentionNumber\":
            10\n        }\n      ],\n      \"pvc_deletion_timeout\": 300,\n      \"restore_to_new_pvc\":
            false,\n      \"snapshot_timeout\": 180,\n      \"storage_plugin\": \"csi.so\",\n
            \     \"app_plugin\": \"mariadb.so\"\n    }\n  },\n  {\n    \"apiVersion\":
            \"fossul.io/v1\",\n    \"kind\": \"BackupSchedule\",\n    \"metadata\":
            {\n      \"name\": \"mariadb-sample\"\n    },\n    \"spec\": {\n      \"cron_schedule\":
            \"59 23 * * *\",\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\"\n\n    }\n  },\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n
            \   \"kind\": \"Fossul\",\n    \"metadata\": {\n      \"name\": \"fossul-sample\"\n
            \   },\n    \"spec\": {\n      \"enable_routes\": false,\n      \"enable_virtualmachines\":
            false,\n      \"container_platform\": \"openshift\"\n    }\n  }\n]"
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/fossul/fossul-operator:v0.6.0
          createdAt: "2022-01-25T20:09:00Z"
          description: Fossul is a application centric container native backup and
            restore framework for databases or applications
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/fossul/fossul
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages application backup configurations
            kind: BackupConfig
            name: backupconfigs.fossul.io
            version: v1
          - description: Manages application backups
            kind: Backup
            name: backups.fossul.io
            version: v1
          - description: Manages application restores
            kind: Restore
            name: restores.fossul.io
            version: v1
          - description: Manages application backup scheduling
            kind: BackupSchedule
            name: backupschedules.fossul.io
            version: v1
          - description: Manages deployment of the Fossul Framework
            kind: Fossul
            name: fossuls.fossul.io
            version: v1
        description: "## Fossul Framework\nApplication backup and restore framework
          for container native applications on Kubernetes and OpenShift. Fossul provides
          application centric backup and recovery enabling integration with the application
          creating an application consistent backup. Since there are many applications
          and providers for storage or archive Fossul enables plugins that allow proper
          integration. Fossul provides plugins for postgresql, mariadb, mysql and
          mongo as well as plugin to integrate with CSI snapshots. Plugins can be
          written in any language. The Fossul operator deploys Fossul via the Fossul
          custom resource and additionally provides BackupConfig, Backup, BackupSchedule
          and Restore custom resources. Each user can configure and manage their own
          backups with simple custom resources that can be part of CI/CD or source
          control. Users can also opt to manage backups through Fossul CLI/API instead
          of the custom resources provided by the operator.\n\n## Installation\nIt
          is recommended to install the operator into the openshift-operators namespace
          which is the default location for operators. Once the operator is installed
          create a new project called ```fossul``` and create the Fossul custom resource
          under the fossul namespace. This will deploy the Fossul Framework. You can
          choose another namespace for the Fossul Framework but then also need to
          provide the optional 'fossul_namespace' parameter with the spec for all
          the additional custom resources. This is required so the Fossul operator
          can communicate with Fossul.\n\nOnce Fossul is deployed you can manage backups
          either through custom resources provided by the operator or the Fossul CLI/API.
          If using the customer resources, first create a BackupConfig custom resource
          for every application you want to backup. Once that is done, you can optionally
          create a backup by simply creating a Backup custom resource. You can also
          create a BackupSchedule custom resource using cron syntax which will schedule
          your backup to occur on a regular interval or do so via a Kubernetes job.
          Finally if a backup has been created you can perform a restore through a
          custom resource as well, providing the workflow id from the completed backup.
          If you are creating backups through the custom resource the workflow id
          will be appended to the spec once the backup completes. If you are using
          the Fossul CLI/API to create backups you will need to get the workflow id
          through the Fossul CLI/API directly using the backupList action.\n\nFor
          users of the cli, a cli pod is deployed with credentials and if desired
          you can create or manage backups from within that pod. To use the API you
          need the credentials stored in the fossul secret. The cli has already been
          pre-configured with these credentials. The API can also be optionally exposed
          via routes and API documentation is under ```/api/v1/index.html```.\n\n##
          Quick Setup\nWhen creating Fossul custom resources generally you will only
          need to update the ```deployment_name``` and ```deployment_type``` (DeploymentConfig,
          Deployment or VirtualMachine). All custom resources must be created in the
          namespace where the database or application exists. \n\n### Deploy Fossul
          in fossul namespace\n```$ oc new-project fossul```\n<pre>\n$ vi fossul.yaml\nkind:
          Fossul\napiVersion: fossul.io/v1\nmetadata:\n  name: fossul-sample\n  namespace:
          fossul\nspec: {}\n</pre>\n```$ oc create -f fossul.yaml```\n\n### Deploy
          Mariadb Database\nMake sure you create all Fossul custom resources in the
          same namespace as the database, application or virtual machine\n\n### Create
          MariaDB Fossul BackupConfig\nCreates a backup configuration for MariaDB
          or MySQL databases\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true\n  auto_discovery:
          false\n  deployment_name: mariadb\n  deployment_type: DeploymentConfig\n
          \ job_retention: 50\n  overwrite_pcv_on_restore: true\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc: false\n  snapshot_timeout:
          180\n  storage_plugin: csi.so\n  app_plugin: mariadb.so \n</pre>\n```$ oc
          create -f backupconfig.yaml```\n\n### Create PostgreSQL Fossul BackupConfig\nCreates
          a backup config for PostgreSQL databases, you need to ensure the user defined
          in secret has backup permissions\n\n```$ vi backupconfig.yaml```\n<pre>\nkind:
          BackupConfig\napiVersion: fossul.io/v1\nmetadata:\n  name: postgresql-sample\n
          \ namespace: databases\nspec:\n  container_platform: openshift\n  operator_controlled:
          true    \n  auto_discovery: false\n  deployment_name: postgresql \n  deployment_type:
          DeploymentConfig\n  job_retention: 50\n  overwrite_pcv_on_restore: true\n
          \ policies:\n    - policy: hourly\n      retentionNumber: 3\n    - policy:
          daily\n      retentionNumber: 10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc:
          false\n  snapshot_timeout: 180\n  storage_plugin: csi.so\n  app_plugin:
          postgres.so \n</pre>\n```$ oc create -f backupconfig.yaml```\n\n### Create
          Kubevirt Fossul BackupConfig\nCreates a backup configuration for virtual
          machines\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: rhel8-vm-sample\n  namespace: virtualmachines\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true    \n  auto_discovery:
          false\n  deployment_name: rhel8-vm\n  deployment_type: VirtualMachine\n
          \ job_retention: 50\n  overwrite_pcv_on_restore: true\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc: false\n  snapshot_timeout:
          180\n  storage_plugin: csi.so\n  app_plugin: kubevirt.so \n</pre>\n```$
          oc create -f backupconfig.yaml```\n\n### Create Tar Fossul BackupConfig\nCreates
          a backup config using tar to backup folders and files where csi snapshots
          are not desired\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: tar-sample\n  namespace: databases\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true    \n  auto_discovery:
          false\n  deployment_name: anyname\n  job_retention: 50\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  backup_source_paths: /opt/app-root/src/test\n  storage_plugin: tar.so\n</pre>\n```$
          oc create -f backupconfig.yaml```\n\n### Create BackupSchedule\nA backup
          can be scheduled per policy, defined by backup configuration using cron
          syntax\n\n```$ vi backupschedule.yaml```\n<pre>\nkind: BackupSchedule\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ cron_schedule: '59 23 * * *'\n  deployment_name: mariadb\n  policy: daily\n</pre>\n```$
          oc create -f backupschedule.yaml```\n\n### Create Backup\nA backup will
          be created according to policy and deleted either manually or as defined
          in retention policy\n\n```$ vi backup.yaml```\n<pre>\nkind: Backup\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ deployment_name: mariadb\n  policy: daily\n</pre>\n```$ oc create -f backup.yaml```\n\n###
          Perform Restore\nA restore requires a workflow_id located in the backup
          spec, upon a successful restore the custom resource will deleted automatically\n\n```$
          vi restore.yaml```\n<pre>\nkind: Restore\napiVersion: fossul.io/v1\nmetadata:\n
          \ name: mariadb-sample\n  namespace: databases\nspec:\n  deployment_name:
          mariadb\n  policy: daily\n  workflow_id: xxxx\n</pre>\n```$ oc create -f
          restore.yaml```\n"
        displayName: Fossul Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - application
        - consistent
        links:
        - name: Fossul Framework Community
          url: https://github.com/fossul/fossul
        maintainers:
        - email: ktenzer@redhat.com
          name: Keith Tenzer
        maturity: alpha
        provider:
          name: Fossul Framework
          url: https://github.com/fossul/fossul
        relatedImages:
        - quay.io/fossul/fossul-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: fossul-operator.v0.6.0
        version: 0.6.0
      - name: fossul-operator.v0.5.0
        version: 0.5.0
      - name: fossul-operator.v0.4.0
        version: 0.4.0
      - name: fossul-operator.v0.3.0
        version: 0.3.0
      - name: fossul-operator.v0.2.0
        version: 0.2.0
      - name: fossul-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: fossul-operator
    provider:
      name: Fossul Framework
      url: https://github.com/fossul/fossul
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Citrus Framework
      provider-url: ""
    name: yaks
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: yaks-operator.v0.17.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "yaks.citrusframework.org/v1alpha1",
                "kind": "Instance",
                "metadata": {
                  "labels": {
                    "app": "yaks"
                  },
                  "name": "example-instance"
                },
                "spec": {
                  "operator": {
                    "global": true
                  }
                }
              },
              {
                "apiVersion": "yaks.citrusframework.org/v1alpha1",
                "kind": "Test",
                "metadata": {
                  "labels": {
                    "app": "yaks"
                  },
                  "name": "example-test"
                },
                "spec": {
                  "source": {
                    "content": "Feature: hello world\n\n  Scenario: print slogan\n    Given YAKS does Cloud-Native BDD testing\n    Then YAKS rocks!",
                    "language": "feature",
                    "name": "example.feature"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/citrusframework/yaks:0.17.1
          createdAt: "2023-04-26T20:37:31Z"
          description: YAKS is a platform to enable Cloud Native BDD testing on Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["instances.yaks.citrusframework.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/citrusframework/yaks
          support: Citrus Framework
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance is the Schema for the yaks instance
            displayName: Instance
            kind: Instance
            name: instances.yaks.citrusframework.org
            version: v1alpha1
          - description: Test is the Schema for the tests API
            displayName: Test
            kind: Test
            name: tests.yaks.citrusframework.org
            version: v1alpha1
        description: |
          YAKS
          ====

          YAKS is a platform to enable Cloud Native BDD testing on Kubernetes.

          ## Running a YAKS test

          With the YAKS operator installed, you can run tests by creating a `Test` custom resource on the cluster.

          Tests are defined using [Gherkin](https://cucumber.io/docs/gherkin/) syntax. YAKS provides a set of predefined steps which
          help to connect with different messaging transports (Http REST, JMS, Kafka, Knative eventing) and verify responses with
          assertions on message header and body content.

          The example below defines a simple test resource.

          ```
          apiVersion: yaks.citrusframework.org/v1alpha1
          kind: Test
          metadata:
            name: example
          spec:
            source:
              name: example.feature
              content: |-
                Feature: hello world

                  Scenario: print slogan
                    Given YAKS does Cloud-Native BDD testing
                    Then YAKS rocks!
          ```

          Refer to the [YAKS repository](https://github.com/citrusframework/yaks) for more complex examples.
        displayName: YAKS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - yaks
        - testing
        - microservices
        - knative
        - kafka
        - serverless
        - camel
        - camel-k
        links:
        - name: Project page
          url: https://citrusframework.org/
        - name: YAKS source code repository
          url: https://github.com/citrusframework/yaks
        maintainers:
        - email: cdeppisch@redhat.com
          name: Christoph Deppisch
        - email: nferraro@redhat.com
          name: Nicola Ferraro
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Citrus Framework
        relatedImages:
        - docker.io/citrusframework/yaks:0.17.1
        version: 0.17.1
      entries:
      - name: yaks-operator.v0.17.1
        version: 0.17.1
      - name: yaks-operator.v0.17.0
        version: 0.17.0
      - name: yaks-operator.v0.16.0
        version: 0.16.0
      - name: yaks-operator.v0.15.1
        version: 0.15.1
      - name: yaks-operator.v0.15.0
        version: 0.15.0
      - name: yaks-operator.v0.14.3
        version: 0.14.3
      - name: yaks-operator.v0.14.2
        version: 0.14.2
      - name: yaks-operator.v0.14.1
        version: 0.14.1
      - name: yaks-operator.v0.13.0
        version: 0.13.0
      - name: yaks-operator.v0.12.2
        version: 0.12.2
      - name: yaks-operator.v0.11.0
        version: 0.11.0
      - name: yaks-operator.v0.10.0
        version: 0.10.0
      - name: yaks-operator.v0.9.0
        version: 0.9.0
      - name: yaks-operator.v0.8.0
        version: 0.8.0
      - name: yaks-operator.v0.7.0
        version: 0.7.0
      - name: yaks-operator.v0.6.0
        version: 0.6.0
      - name: yaks-operator.v0.5.0
        version: 0.5.0
      - name: yaks-operator.v0.4.0
        version: 0.4.0
      name: alpha
    defaultChannel: alpha
    packageName: yaks
    provider:
      name: Citrus Framework
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator.v1.5.0
        version: 1.5.0
      - name: dynatrace-operator.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetScaler
      provider-url: https://netscaler.com
    name: netscaler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netscaler-operator.v3.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCloudController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalercloudcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "azure": {
                    "clientsecret": "",
                    "image": "quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb",
                    "location": "",
                    "pullPolicy": "IfNotPresent",
                    "resourcegroupname": "",
                    "resources": {},
                    "subnetname": "",
                    "subscriptionid": "",
                    "tenantid": "",
                    "vnetname": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCpxWithIngressController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalercpxwithingresscontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "admSettings": {
                    "admIP": "",
                    "bandWidth": 1000,
                    "bandWidthLicense": false,
                    "cpxCores": 1,
                    "licenseEdition": "PLATINUM",
                    "licenseServerIP": "",
                    "licenseServerPort": 27000,
                    "loginSecret": "",
                    "platform": false,
                    "vCPULicense": false
                  },
                  "affinity": {},
                  "analyticsConfig": {
                    "distributedTracing": {
                      "enable": false,
                      "samplingRate": 100
                    },
                    "endpoint": {
                      "metrics": {
                        "service": ""
                      },
                      "transactions": {
                        "service": ""
                      }
                    },
                    "required": false,
                    "timeseries": {
                      "auditLogs": {
                        "enable": false
                      },
                      "events": {
                        "enable": false
                      },
                      "metrics": {
                        "enable": false,
                        "enableNativeScrape": false,
                        "exportFrequency": 30,
                        "mode": "avro",
                        "schemaFile": "schema.json"
                      },
                      "port": 5563
                    },
                    "transactions": {
                      "enable": false,
                      "port": 5557
                    }
                  },
                  "exporter": {
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4",
                    "ports": {
                      "containerPort": 8888
                    },
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {}
                  },
                  "extraVolumes": [],
                  "imagePullSecrets": [],
                  "ingressController": {
                    "defaultSSLCertSecret": "",
                    "defaultSSLSNICertSecret": "",
                    "disableAPIServerCertVerify": false,
                    "disableOpenshiftRoutes": true,
                    "enableLivenessProbe": true,
                    "enableReadinessProbe": true,
                    "entityPrefix": "k8s",
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "ingressClass": [],
                    "ipam": false,
                    "jsonLog": false,
                    "kubernetesURL": "",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "logLevel": "INFO",
                    "logProxy": "",
                    "namespaceLabels": "",
                    "openshift": true,
                    "optimizeEndpointBinding": false,
                    "profileHttpFrontend": {},
                    "profileSslFrontend": {},
                    "profileTcpFrontend": {},
                    "pullPolicy": "IfNotPresent",
                    "rbacRole": false,
                    "readinessProbe": {
                      "failureThreshold": 3,
                      "initialDelaySeconds": 50,
                      "periodSeconds": 60,
                      "successThreshold": 1
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "128m",
                        "memory": "500Mi"
                      }
                    },
                    "routeLabels": "",
                    "setAsDefaultIngressClass": false,
                    "updateIngressStatus": true
                  },
                  "netscalerCpx": {
                    "bgpSettings": {
                      "bgpConfig": [
                        {
                          "bgpRouter": {
                            "localAS": 100,
                            "neighbor": [
                              {
                                "ASOriginationInterval": 10,
                                "address": "",
                                "advertisementInterval": 10,
                                "remoteAS": 100
                              }
                            ]
                          }
                        }
                      ],
                      "bgpPort": 179
                    },
                    "cpxBgpRouter": false,
                    "cpxCommands": [],
                    "cpxLicenseAggregator": "",
                    "cpxShellCommands": [],
                    "daemonSet": false,
                    "enableLivenessProbe": true,
                    "enableStartupProbe": true,
                    "extraVolumeMounts": [],
                    "hostname": "",
                    "image": "quay.io/netscaler/netscaler-cpx@sha256:232596eaacda68c0e2e210f26b64cedc8d70ae12c5e8ae2e1a66896a479ae4f8",
                    "ingressIP": "",
                    "livenessProbe": {
                      "failureThreshold": 3,
                      "periodSeconds": 5,
                      "successThreshold": 1
                    },
                    "mgmtHttpPort": 9080,
                    "mgmtHttpsPort": 9443,
                    "nitroReadTimeout": 20,
                    "nsConfigDnsRec": false,
                    "nsCookieVersion": 0,
                    "nsDnsNameserver": "",
                    "nsEnableLabels": true,
                    "nsGateway": "192.168.1.1",
                    "nsHttp2ServerSide": "OFF",
                    "nsIP": "192.168.1.2",
                    "nsLbHashAlgo": {
                      "hashAlgorithm": "default",
                      "hashFingers": 500,
                      "required": false
                    },
                    "nsProtocol": "HTTP",
                    "nsSvcLbDnsRec": false,
                    "prometheusCredentialSecret": "",
                    "pullPolicy": "IfNotPresent",
                    "service": {
                      "annotations": {},
                      "spec": {}
                    },
                    "startupProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1
                    }
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "replicaCount": 1,
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerGslbController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalergslbcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "gslbController": {
                    "disableAPIServerCertVerify": false,
                    "entityPrefix": "",
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "kubernetesURL": "",
                    "localCluster": "",
                    "localRegion": "",
                    "logLevel": "Debug",
                    "openshift": true,
                    "pullPolicy": "IfNotPresent",
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "32m",
                        "memory": "128Mi"
                      }
                    },
                    "siteData": [
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteMask": "",
                        "siteName": "",
                        "sitePublicIp": "",
                        "siteRegion": ""
                      },
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteMask": "",
                        "siteName": "",
                        "sitePublicIp": "",
                        "siteRegion": ""
                      }
                    ]
                  },
                  "imagePullSecrets": [],
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsIP": "",
                    "nsPort": 443,
                    "nsProtocol": "HTTPS"
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerIngressController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscaleringresscontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "analyticsConfig": {
                    "distributedTracing": {
                      "enable": false,
                      "samplingRate": 100
                    },
                    "endpoint": {
                      "metrics": {
                        "service": ""
                      },
                      "transactions": {
                        "service": ""
                      }
                    },
                    "required": false,
                    "timeseries": {
                      "auditLogs": {
                        "enable": false
                      },
                      "events": {
                        "enable": false
                      },
                      "metrics": {
                        "enable": false,
                        "enableNativeScrape": false,
                        "exportFrequency": 50,
                        "mode": "avro",
                        "schemaFile": "schema.json"
                      },
                      "port": 50002
                    },
                    "transactions": {
                      "enable": false,
                      "port": 50005
                    }
                  },
                  "exporter": {
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4",
                    "ports": {
                      "containerPort": 8888
                    },
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {},
                    "serviceMonitorExtraLabels": {}
                  },
                  "extraVolumes": [],
                  "hostAliases": [],
                  "imagePullSecrets": [],
                  "ingressController": {
                    "clusterName": "",
                    "defaultSSLCertSecret": "",
                    "defaultSSLSNICertSecret": "",
                    "disableAPIServerCertVerify": false,
                    "disableOpenshiftRoutes": true,
                    "enableLivenessProbe": true,
                    "enableReadinessProbe": true,
                    "entityPrefix": "",
                    "extraVolumeMounts": [],
                    "ignoreNodeExternalIP": false,
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "ingressClass": [],
                    "ipam": false,
                    "jsonLog": false,
                    "kubernetesURL": "",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "logLevel": "INFO",
                    "logProxy": "",
                    "multiClusterPrefix": "",
                    "namespaceLabels": "",
                    "nodeLabels": "",
                    "nodeWatch": false,
                    "nsValidateCert": {},
                    "nsncPbr": false,
                    "openshift": true,
                    "optimizeEndpointBinding": false,
                    "podIPsforServiceGroupMembers": false,
                    "profileHttpFrontend": {},
                    "profileSslFrontend": {},
                    "profileTcpFrontend": {},
                    "pullPolicy": "IfNotPresent",
                    "rbacRole": false,
                    "readinessProbe": {
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 60,
                      "successThreshold": 1
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "32m",
                        "memory": "128Mi"
                      }
                    },
                    "routeLabels": "",
                    "serviceClass": [],
                    "setAsDefaultIngressClass": false,
                    "updateIngressStatus": true
                  },
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsConfigDnsRec": false,
                    "nsCookieVersion": 0,
                    "nsDnsNameserver": "",
                    "nsHttp2ServerSide": "OFF",
                    "nsIP": "",
                    "nsLbHashAlgo": {
                      "hashAlgorithm": "default",
                      "hashFingers": 256,
                      "required": false
                    },
                    "nsPort": 443,
                    "nsProtocol": "HTTPS",
                    "nsSNIPs": [],
                    "nsSvcLbDnsRec": false,
                    "nsVIP": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerIpamController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscaleripamcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "clusterName": "",
                  "image": "quay.io/netscaler/netscaler-ipam-controller@sha256:7c25b59a2e1df0e4f5615a779a9bea939c60a3eca065cb6c2295daf0d03831d8",
                  "localIPAM": {
                    "vipRange": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "reuseIngressVIP": true,
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerObservabilityExporter",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalerobservabilityexporter-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "elasticsearch": {
                    "enabled": false,
                    "server": "http://elasticsearch:9200"
                  },
                  "headless": false,
                  "image": "quay.io/netscaler/netscaler-observability-exporter@sha256:626e3a954c67bbb910477352dc740aaab21ff1c1cfb03354becf50beab17d4a4",
                  "jsonTransRateLimiting": {
                    "enabled": false,
                    "limit": 100,
                    "queueLimit": 1000,
                    "window": 5
                  },
                  "kafka": {
                    "auditLogs": "no",
                    "broker": "",
                    "dataFormat": "AVRO",
                    "enabled": false,
                    "events": "no",
                    "topic": "HTTP"
                  },
                  "nodePortRequired": false,
                  "nodeSelector": {},
                  "nsoeLogLevel": "INFO",
                  "nstracing": {
                    "enabled": false,
                    "server": "zipkin:9411/api/v1/spans"
                  },
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "replicas": 1,
                  "resources": {},
                  "splunk": {
                    "authToken": "",
                    "enabled": false,
                    "indexPrefix": "adc_nsoe",
                    "server": "http://splunk:8088"
                  },
                  "timeseries": {
                    "enabled": false
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-operator@sha256:e367c8c8ba6171e9ef5dd496592a5c62f4aa764f93d0cb78835aa29024c063d0
          createdAt: "2025-04-15T08:51:39Z"
          description: This is an operator to install Cloud Native portfolio provided
            by NetScaler
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler Cloud Controller for NetScaler VPX/MPX/SDX
              with Azure Redhat OpenShift in Azure.
            displayName: Netscaler Cloud Controller
            kind: NetscalerCloudController
            name: netscalercloudcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler CPX with Ingress Controller.
            displayName: Netscaler CPX With Ingress Controller
            kind: NetscalerCpxWithIngressController
            name: netscalercpxwithingresscontrollers.netscaler.com
            version: v1
          - description: To install NetScaler GSLB Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler GSLB Controller
            kind: NetscalerGslbController
            name: netscalergslbcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler Ingress Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler Ingress Controller
            kind: NetscalerIngressController
            name: netscaleringresscontrollers.netscaler.com
            version: v1
          - description: To install NetScaler IPAM Controller for NetScaler CPX/VPX/MPX/SDX.
            displayName: Netscaler IPAM Controller
            kind: NetscalerIpamController
            name: netscaleripamcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler Observability Exporter for NetScaler.
            displayName: Netscaler Observability Exporter
            kind: NetscalerObservabilityExporter
            name: netscalerobservabilityexporters.netscaler.com
            version: v1
        description: |
          NetScaler provides various products to empower Developer/DevOps managing your applications in OpenShift Cluster. Using this operator you can deploy below controllers for NetScaler appliance:
          1. [NetScaler Cloud Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. [NetScaler Ingress Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/unified-ingress-deployment)
          3. [NetScaler CPX with Ingress Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/two-tier-deployment)
          4. [NetScaler GSLB Controller ](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/gslb/gslb)
          5. [NetScaler Observability Exporter](https://docs.netscaler.com/en-us/netscaler-observability-exporter/deploy-coe)
          6. [NetScaler IPAM Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/ipam-installation.html)

          ### Installation
          Refer [installation instructions](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-operator-openshift).

          This operator version contains:
          1. NetScaler Ingress Controller version [2.2.10](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.2.10)
          2. NetScaler CPX version [14.1-25.111](https://docs.netscaler.com/en-us/updates?product=NetScaler-ADC-%28includes-NetScaler-Gateway%29&version=14.1&build=25.56)
          3. NetScaler GSLB Controller version [2.2.10](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.2.10)
          4. NetScaler Observability Exporter version [1.10.001](https://github.com/citrix/citrix-helm-charts/releases/tag/nsoe-v1.10.001)
          5. NetScaler Metrics Exporter [1.4.9](https://github.com/citrix/citrix-adc-metrics-exporter/releases/tag/1.4.9)
          6. NetScaler IPAM Controller [2.1.2](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.3.15)

          ### Support

          For debugging issues, please use [NetScaler Kubernetes Plugin](https://github.com/netscaler/modern-apps-toolkit/tree/main/netscaler-plugin).

          For questions and support, the following channels are available:
          - [NetScaler community portal](https://community.netscaler.com/s/)
          - [Contact us](https://www.netscaler.com/contact)
        displayName: NetScaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        - ADC
        - Kubernetes
        - OpenShift
        - Citrix
        - Controller
        - Observability
        - Metrices
        links:
        - name: Netscaler Operator
          url: https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-operator-openshift
        maintainers:
        - email: priyanka.sharma@cloud.com
          name: Priyanka Sharma
        - email: arijit.ray@cloud.com
          name: Arijit Ray
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: NetScaler
          url: https://netscaler.com
        relatedImages:
        - quay.io/netscaler/netscaler-cpx@sha256:232596eaacda68c0e2e210f26b64cedc8d70ae12c5e8ae2e1a66896a479ae4f8
        - quay.io/netscaler/netscaler-observability-exporter@sha256:626e3a954c67bbb910477352dc740aaab21ff1c1cfb03354becf50beab17d4a4
        - quay.io/netscaler/netscaler-ipam-controller@sha256:7c25b59a2e1df0e4f5615a779a9bea939c60a3eca065cb6c2295daf0d03831d8
        - quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f
        - quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:89f44c45373da73a51b0b9a02de7efe519dc2d5343b65fe382526cccc92f8f5c
        - registry.connect.redhat.com/citrix/netscaler-operator@sha256:e367c8c8ba6171e9ef5dd496592a5c62f4aa764f93d0cb78835aa29024c063d0
        - quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4
        version: 3.2.0
      entries:
      - name: netscaler-operator.v3.2.0
        version: 3.2.0
      - name: netscaler-operator.v3.1.2
        version: 3.1.2
      - name: netscaler-operator.v3.1.1
        version: 3.1.1
      - name: netscaler-operator.v3.0.1
        version: 3.0.1
      name: stable
    - currentCSV: netscaler-operator.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCloudController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netscaler-operator",
                    "app.kubernetes.io/instance": "nscc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscalercloudcontroller",
                    "app.kubernetes.io/part-of": "netscaler-operator"
                  },
                  "name": "nscc"
                },
                "spec": {
                  "azure": {
                    "clientsecret": "",
                    "image": "quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb",
                    "location": "",
                    "pullPolicy": "IfNotPresent",
                    "resourcegroupname": "",
                    "resources": {},
                    "subnetname": "",
                    "subscriptionid": "",
                    "tenantid": "",
                    "vnetname": ""
                  }
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerGslbController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netscaler-operator",
                    "app.kubernetes.io/instance": "nsgc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscalergslbcontroller",
                    "app.kubernetes.io/part-of": "netscaler-operator"
                  },
                  "name": "nsgc"
                },
                "spec": {
                  "gslbController": {
                    "acceptLicense": "Yes",
                    "cleanupGSLBSiteConfig": false,
                    "disableAPIServerCertVerify": false,
                    "entityPrefix": "",
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:88114c518ecceb54b000f38bfc330efe184bfeb5b5df0ee308950cf29e7a777f",
                    "imagePullSecrets": [],
                    "kubernetesURL": "",
                    "localCluster": "",
                    "localRegion": "",
                    "logLevel": "Debug",
                    "openshift": true,
                    "pullPolicy": "IfNotPresent",
                    "siteData": [
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteName": "",
                        "siteRegion": ""
                      },
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteName": "",
                        "siteRegion": ""
                      }
                    ]
                  },
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsIP": "",
                    "nsPort": 443,
                    "nsProtocol": "HTTPS"
                  },
                  "resources": {
                    "limits": {},
                    "requests": {
                      "cpu": "32m",
                      "memory": "128Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-operator@sha256:3f71f9a46d7b29e3da98a301965f635c03a022b5535b586d7d5c699341f68193
          createdAt: "2024-07-31T06:33:49Z"
          description: This is an operator to install Cloud Native portfolio provided
            by NetScaler
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler Cloud Controller for NetScaler VPX/MPX/SDX
              with Azure Redhat OpenShift in Azure.
            displayName: Netscaler Cloud Controller
            kind: NetscalerCloudController
            name: netscalercloudcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler GSLB Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler GSLB Controller
            kind: NetscalerGslbController
            name: netscalergslbcontrollers.netscaler.com
            version: v1
        description: |-
          NetScaler provides various products to empower Developer/DevOps managing your applications in OpenShift Cluster. Using this operator you can deploy below controllers for NetScaler appliance:
          1. [NetScaler Cloud Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. [NetScaler GSLB Controller ](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/gslb-openshift-operator)

          ### Installation
          Refer [installation instructions](https://github.com/netscaler/netscaler-k8s-ingress-controller/blob/master/docs/deploy/deploy-ns-operator.md).

          This operator version contains:
          1. NetScaler Cloud Controller version [1.2.0](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. NetScaler GSLB Controller version [1.43.7](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v1.43.7)

          ### Support

          For debugging issues, please use [NetScaler Kubernetes Plugin](https://github.com/netscaler/modern-apps-toolkit/tree/main/netscaler-plugin).

          For questions and support, the following channels are available:
          - [NetScaler community portal](https://community.netscaler.com/s/)
          - [Contact us](https://www.netscaler.com/contact)
        displayName: NetScaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        links:
        - name: Netscaler Operator
          url: https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/gslb-openshift-operator
        maintainers:
        - email: priyanka.sharma@cloud.com
          name: Priyanka Sharma
        - email: arijit.ray@cloud.com
          name: Arijit Ray
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: NetScaler
          url: https://netscaler.com
        relatedImages:
        - quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:88114c518ecceb54b000f38bfc330efe184bfeb5b5df0ee308950cf29e7a777f
        - quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:23405618050a4ceea8ea0c748cc7d9d16dbb80f10e4c39f7e82014a3ac80cec9
        - registry.connect.redhat.com/citrix/netscaler-operator@sha256:3f71f9a46d7b29e3da98a301965f635c03a022b5535b586d7d5c699341f68193
        version: 3.0.1
      entries:
      - name: netscaler-operator.v3.0.1
        version: 3.0.1
      name: tech-preview
    defaultChannel: stable
    packageName: netscaler-operator
    provider:
      name: NetScaler
      url: https://netscaler.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: open-liberty-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: open-liberty-operator.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4",
                  "expose": true,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/open-liberty-operator-controller@sha256:18f188e48725c785557c8edea46f69c464d6f81ab47860a3bb0ec978c20a26db
          createdAt: "2022-03-25T14:00:00Z"
          description: Deploy and manage applications running on Liberty
          olm.skipRange: <0.8.1
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          This advanced Operator can be used to deploy and manage Open Liberty applications with consistent, production-grade QoS. This operator is based on the [Runtime Component Operator](https://github.com/application-stacks/runtime-component-operator) and provides all of its capabilities in addition to Open Liberty specific features such as gathering traces and dumps (Day-2 operations) and easily configuring and managing the single sign-on information for your Open Liberty applications.

          Open Liberty Operator enables enterprise architects to govern the way their applications get deployed & managed in the cluster, while dramatically reducing the learning curve for developers to deploy into Kubernetes - allowing them to focus on writing the code!

          Here are some key features:

          #### Application Lifecyle
          You can deploy your Open Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the Operator will watch for any updates and will re-deploy the modified image.

          #### Custom RBAC
          This Operator is capable of using a custom ServiceAccount from the caller, allowing it to follow RBAC restrictions. By default it creates a ServiceAccount if one is not specified, which can also be bound with specific roles.

          #### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as: labels, annotations, and environment variables from a ConfigMap, a Secret or a value.

          #### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configuration, such as TLS settings, are also easily enabled.

          #### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto scaling to create (and delete) instances based on resource consumption.

          #### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you.
          This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart.
          We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          You can also easily configure and use a single storage for serviceability related Day-2 operations, such as gatherig server traces and dumps.

          #### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating kubernetes Secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to its required services without any intervention or interruption.

          #### Single Sign-On (SSO)
          Open Liberty provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows to easily configure and manage the single sign-on information for your applications.

          #### Exposing metrics to Prometheus
          Expose the Open Liberty application's metrics via the Prometheus Operator.
          You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `spec.monitoring` field to control configurations such as poll interval and security credentials.

          #### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply add the following configuration (to specify the volume size and the location to persist) to your OpenLibertyApplication CR:
          ``` storage: size: 2Gi mountPath: "/logs" ```

          #### Integration with OpenShift Serverless
          Deploy your serverless runtime component using a single toggle.  The Operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing your pod to automatically scale to 0 when it is idle.

          #### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and how they are connected.

          See our [**documentation**](https://github.com/OpenLiberty/open-liberty-operator/tree/main/doc/) for more information.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://github.com/OpenLiberty/open-liberty-operator/tree/main/doc
        - name: Open Liberty
          url: https://openliberty.io/
        - name: Open Liberty Container Images
          url: https://github.com/OpenLiberty/ci.docker
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: beta
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/open-liberty-operator-controller@sha256:18f188e48725c785557c8edea46f69c464d6f81ab47860a3bb0ec978c20a26db
        - registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4
        version: 0.8.1
      entries:
      - name: open-liberty-operator.v0.8.1
        version: 0.8.1
      - name: open-liberty-operator.v0.8.0
        version: 0.8.0
      name: beta2
    - currentCSV: open-liberty-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:d45c296d43dc4c4100f38bb736f073e04c8b4d72b68ddf1189e311aff8ad66e3
          createdAt: "2023-08-28T16:39:10Z"
          description: Deploy and manage containerized Liberty applications
          olm.skipRange: '>=0.8.0 <1.2.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:d45c296d43dc4c4100f38bb736f073e04c8b4d72b68ddf1189e311aff8ad66e3
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02
        version: 1.2.2
      entries:
      - name: open-liberty-operator.v1.2.2
        version: 1.2.2
      name: v1.2
    - currentCSV: open-liberty-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:d68bb66a8133702a198f7be655d7cff5f500bb620c460a81547971e00124ef73
          createdAt: "2024-08-01T20:19:13Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.3.3'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:d68bb66a8133702a198f7be655d7cff5f500bb620c460a81547971e00124ef73
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91
        version: 1.3.3
      entries:
      - name: open-liberty-operator.v1.3.3
        version: 1.3.3
      name: v1.3
    - currentCSV: open-liberty-operator.v1.4.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:67b5b46dfed008dab4ea7fd176d3bc31dae8940d72194a10aebb98b1dea9043e
          createdAt: "2025-07-18T20:45:06Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.4.4'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:67b5b46dfed008dab4ea7fd176d3bc31dae8940d72194a10aebb98b1dea9043e
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3
        version: 1.4.4
      entries:
      - name: open-liberty-operator.v1.4.4
        version: 1.4.4
      name: v1.4
    - currentCSV: open-liberty-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyPerformanceData",
                "metadata": {
                  "name": "openliberty-performance-data-sample"
                },
                "spec": {
                  "interval": 30,
                  "podName": "Specify_Pod_Name_Here",
                  "timespan": 240
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:c3b031fe14054f65d6612cc505b7681336873dc32ce1091451d8c7799034bc83
          createdAt: "2025-09-16T03:08:31Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.5.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server performance data
            displayName: OpenLibertyPerformanceData
            kind: OpenLibertyPerformanceData
            name: openlibertyperformancedata.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:c3b031fe14054f65d6612cc505b7681336873dc32ce1091451d8c7799034bc83
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8
        version: 1.5.0
      entries:
      - name: open-liberty-operator.v1.5.0
        version: 1.5.0
      name: v1.5
    defaultChannel: v1.5
    packageName: open-liberty-certified
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: RabbitMQ
      provider-url: https://rabbitmq.com
    name: rabbitmq-cluster-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-cluster-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rabbitmq.com/v1beta1",
                "kind": "RabbitmqCluster",
                "metadata": {
                  "name": "sample",
                  "namespace": "rabbitmq-system"
                },
                "spec": {
                  "replicas": 1,
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-07-10T12:52:23Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RabbitmqCluster is the Schema for the RabbitmqCluster API.
              Each instance of this object corresponds to a single RabbitMQ cluster.
            displayName: Rabbitmq Cluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: RabbitMQ Cluster Operator
        displayName: RabbitMQ Cluster Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        links:
        - name: Cluster Operator
          url: https://github.com/rabbitmq/cluster-operator
        maturity: alpha
        provider:
          name: RabbitMQ
          url: https://rabbitmq.com
        relatedImages:
        - registry.redhat.io/rhoso-operators/rabbitmq-cluster-rhel9-operator@sha256:e4b198236d70b7dfdd29d8ed45058dec3335b3530e0b676fcc862bfec2efce79
        version: 1.0.6
      entries:
      - name: rabbitmq-cluster-operator.v1.0.6
        version: 1.0.6
      - name: rabbitmq-cluster-operator.v1.0.5
        version: 1.0.5
      - name: rabbitmq-cluster-operator.v1.0.4
        version: 1.0.4
      - name: rabbitmq-cluster-operator.v1.0.3
        version: 1.0.3
      - name: rabbitmq-cluster-operator.v1.0.2
        version: 1.0.2
      - name: rabbitmq-cluster-operator.v1.0.1
        version: 1.0.1
      - name: rabbitmq-cluster-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: rabbitmq-cluster-operator
    provider:
      name: RabbitMQ
      url: https://rabbitmq.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-route53resolver-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-route53resolver-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "route53resolver.services.k8s.aws/v1alpha1",
                "kind": "ResolverEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "route53resolver.services.k8s.aws/v1alpha1",
                "kind": "ResolverRule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/route53resolver-controller:1.1.1
          createdAt: "2025-09-20T01:21:08Z"
          description: AWS Route53Resolver controller is a service controller for
            managing Route53Resolver resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ResolverEndpoint represents the state of an AWS route53resolver
              ResolverEndpoint resource.
            displayName: ResolverEndpoint
            kind: ResolverEndpoint
            name: resolverendpoints.route53resolver.services.k8s.aws
            version: v1alpha1
          - description: ResolverRule represents the state of an AWS route53resolver
              ResolverRule resource.
            displayName: ResolverRule
            kind: ResolverRule
            name: resolverrules.route53resolver.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Route53Resolver resources in AWS from within your Kubernetes cluster.

          **About Amazon Route53Resolver**

          Amazon Route 53 Resolver responds recursively to DNS queries from AWS resources for public records, Amazon VPC-specific DNS names, and Amazon Route 53 private hosted zones, and is available by default in all VPCs.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Route53Resolver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - route53resolver
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Route53Resolver Developer Resources
          url: https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/resolver.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: route53resolver maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/route53resolver-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-route53resolver-controller.v1.1.1
        version: 1.1.1
      - name: ack-route53resolver-controller.v1.0.15
        version: 1.0.15
      - name: ack-route53resolver-controller.v1.0.14
        version: 1.0.14
      - name: ack-route53resolver-controller.v1.0.13
        version: 1.0.13
      - name: ack-route53resolver-controller.v1.0.11
        version: 1.0.11
      - name: ack-route53resolver-controller.v1.0.10
        version: 1.0.10
      - name: ack-route53resolver-controller.v1.0.9
        version: 1.0.9
      - name: ack-route53resolver-controller.v1.0.7
        version: 1.0.7
      - name: ack-route53resolver-controller.v1.0.6
        version: 1.0.6
      - name: ack-route53resolver-controller.v1.0.5
        version: 1.0.5
      - name: ack-route53resolver-controller.v1.0.3
        version: 1.0.3
      - name: ack-route53resolver-controller.v1.0.2
        version: 1.0.2
      - name: ack-route53resolver-controller.v1.0.0
        version: 1.0.0
      - name: ack-route53resolver-controller.v0.0.11
        version: 0.0.11
      - name: ack-route53resolver-controller.v0.0.10
        version: 0.0.10
      - name: ack-route53resolver-controller.v0.0.9
        version: 0.0.9
      - name: ack-route53resolver-controller.v0.0.8
        version: 0.0.8
      - name: ack-route53resolver-controller.v0.0.7
        version: 0.0.7
      - name: ack-route53resolver-controller.v0.0.4
        version: 0.0.4
      - name: ack-route53resolver-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-route53resolver-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Weaveworks
      provider-url: ""
    name: tf-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tf-controller.v0.9.0-rc.8
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"infra.contrib.fluxcd.io/v1alpha1\",\n
            \   \"kind\": \"Terraform\",\n    \"metadata\": {\n        \"name\": \"helloworld-tf\",\n
            \       \"namespace\": \"flux-system\"\n    },\n    \"spec\": {\n        \"interval\":
            \"5m\",\n        \"approvePlan\": \"auto\",\n        \"path\": \"./\",\n
            \       \"sourceRef\": {\n          \"kind\": \"GitRepository\",\n          \"name\":
            \"helloworld\",\n          \"namespace\": \"flux-system\"\n        }      \n
            \   }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/weaveworks/tf-controller:v0.9.0-rc.8
          description: Weave TF-controller is a Flux controller that brings GitOps
            to Terraform resources.
          operatorframework.io/suggested-namespace: flux-system
          support: Weaveworks
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Terraform
            displayName: Terraform
            kind: Terraform
            name: terraforms.infra.contrib.fluxcd.io
            version: v1alpha1
        description: "**Weave GitOps Terraform Controller** (aka Weave TF-controller)
          is a Flux controller for reconciling Terraform resources in the GitOps way.
          \nWith the power of Flux together with Terraform, TF-controller allows you
          to GitOps-ify infrastructure, \nand application resources, in the Kubernetes
          and Terraform universe, together at your own pace.\n\nTF-controller offers
          many GitOps models:\n  1. **GitOps Automation Model:** GitOps your Terraform
          resources from the provision steps to the enforcement steps, like a whole
          EKS cluster.\n  2. **Hybrid GitOps Automation Model:** GitOps parts of your
          existing infrastructure resources. For example, you have an existing EKS
          cluster. You can choose to GitOps only its nodegroup, or its security group.\n
          \ 3. **State Enforcement Model:** You have a TFSTATE file, and you'd like
          to use GitOps enforce it, without changing anything else.\n  4. **Drift
          Detection Model:** You have a TFSTATE file, and you'd like to use GitOps
          just for drift detection, so you can decide to do things later when a drift
          occurs.\n\nFeatures\n  * GitOps Automation for Terraform\n  * Drift detection\n
          \ * Plan and Manual Approve"
        displayName: Weave GitOps Terraform Controller
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - weave
        - flux
        - cd
        - continuous delivery
        - terraform
        - gitops
        links:
        - name: Weave GitOps Terraform Controller
          url: https://github.com/weaveworks/tf-controller
        - name: Documentation
          url: https://weaveworks.github.io/tf-controller/
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Weaveworks
        relatedImages:
        - ghcr.io/weaveworks/tf-controller:v0.9.0-rc.8
        version: 0.9.0-rc.8
      entries:
      - name: tf-controller.v0.9.0-rc.8
        version: 0.9.0-rc.8
      - name: tf-controller.v0.9.0-rc.3
        version: 0.9.0-rc.3
      name: stable
    defaultChannel: stable
    packageName: tf-controller
    provider:
      name: Weaveworks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Randoli
      provider-url: https://www.randoli.io
    name: app-insights-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: app-insights-operator.v1.17.5-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "appinsights.randoli/v1beta1",
                "kind": "RandoliAgent",
                "metadata": {
                  "name": "randoli",
                  "namespace": "randoli-agents"
                },
                "spec": {
                  "costManagement": {
                    "opencost": {
                      "exporter": {
                        "image": {
                          "tag": "1.112.1"
                        }
                      }
                    },
                    "vpaOperator": {
                      "enabled": false
                    }
                  },
                  "logs": {
                    "vector": {
                      "agent": {
                        "containerSecurityContext": {
                          "privileged": true
                        }
                      }
                    }
                  },
                  "network": {
                    "standaloneConsole": {
                      "enable": true
                    }
                  },
                  "observability": {
                    "netobserv": {
                      "enabled": false
                    },
                    "otel": {
                      "operator": {
                        "enabled": false
                      }
                    }
                  },
                  "resources": {
                    "limits": {
                      "memory": "2Gi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "512Mi"
                    }
                  },
                  "securityContext": {
                    "fsGroup": 1001
                  },
                  "tags": {
                    "costManagement": false,
                    "observability": false
                  }
                }
              }
            ]
          categories: Developer Tools
          containerImage: docker.io/randoli/app-insights-operator:1.17.5-4@sha256:22e7150c29557f2a71186389e5d7d60bfe82a4ea022e1403033c6ad61bc95f7a
          createdAt: "2025-09-23T19:45:43Z"
          description: Randoli Agent connects OpenShift clusters to Randoli Observability
            & Cost Management.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[]'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Randoli
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: RandoliAgent
            name: randoliagents.appinsights.randoli
            version: v1beta1
        description: |-
          Randoli Agent connects OpenShift clusters to Randoli Observability & Cost Management platforms.

          This operator provides a container-based agent to be installed in your OpenShift clusters for connecting to Randoli Observability & Cost Management platforms

          **Prerequisites**

          If you are using the Randoli Observability platform, you will need to install the following Operators:
          - Red Hat build of OpenTelemetry Operator
          - Network Observability

          If you are using the Randoli Cost Management platform, you will need to install the following Operator:
          - VerticalPodAutoscaler

          For additional information please use our Openshift installation guide here: [https://docs.insights.randoli.io/agent/openshift](https://docs.insights.randoli.io/agent/openshift)

           **Randoli Observability for Kubernetes**

          [Product Page](https://www.randoli.io/product/kubernetes-observability)

          **Randoli Cost Management for Kubernetes**

          [Product Page](https://www.randoli.io/product/kubernetes-cost-management)
        displayName: Randoli Agent
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - developer tools
        - insights
        - monitoring
        links:
        - name: Randoli Observability for Kubernetes
          url: https://www.randoli.io/product/kubernetes-observability
        - name: Randoli Cost Management for Kubernetes
          url: https://www.randoli.io/product/kubernetes-cost-management
        maintainers:
        - email: contact@randoli.ca
          name: Randoli Inc.
        maturity: beta
        minKubeVersion: 1.24.0
        provider:
          name: Randoli
          url: https://www.randoli.io
        relatedImages:
        - docker.io/randoli/app-insights-operator:1.17.5-4@sha256:22e7150c29557f2a71186389e5d7d60bfe82a4ea022e1403033c6ad61bc95f7a
        version: 1.17.5-4
      entries:
      - name: app-insights-operator.v1.17.5-4
        version: 1.17.5-4
      - name: app-insights-operator.v1.17.5-3
        version: 1.17.5-3
      - name: app-insights-operator.v1.17.5-2
        version: 1.17.5-2
      - name: app-insights-operator.v1.17.5-1
        version: 1.17.5-1
      - name: app-insights-operator.v1.17.3-1
        version: 1.17.3-1
      - name: app-insights-operator.v1.17.2-1
        version: 1.17.2-1
      - name: app-insights-operator.v1.17.1-1
        version: 1.17.1-1
      - name: app-insights-operator.v1.17.0-2
        version: 1.17.0-2
      - name: app-insights-operator.v1.17.0-1
        version: 1.17.0-1
      name: stable
    defaultChannel: stable
    packageName: app-insights-operator
    provider:
      name: Randoli
      url: https://www.randoli.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ProphetStor Data Services, Inc.
      provider-url: ""
    name: federatorai-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: federatorai.v4.7.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "federatorai.containers.ai/v1alpha1",
                "kind": "AlamedaService",
                "metadata": {
                  "name": "my-alamedaservice"
                },
                "spec": {
                  "alamedaAi": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "10Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "alamedaInfluxdb": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "100Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "fedemeterInfluxdb": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "10Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "imageLocation": "registry.connect.redhat.com/prophetstor",
                  "storages": [
                    {
                      "accessMode": ["ReadWriteOnce"],
                      "size": "2Gi",
                      "type": "pvc",
                      "usage": "log"
                    }
                  ],
                  "version": "v4.7.1-b1760"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: registry.connect.redhat.com/prophetstor/federatorai-operator@sha256:bdd87f6b7d7ba9d12ed8ddb9c5151e5a89781e88d7a02b73814b16a766968047
          createdAt: "2021-10-12T01:14:00Z"
          description: Federator.ai Operator provides easy configuration and management
            of AI-based Kubernetes resource orchestrator
          operatorframework.io/suggested-namespace: federatorai
          repository: https://access.redhat.com/containers/#/registry.connect.redhat.com/prophetstor/federatorai-operator
          support: ProphetStor Data Services, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of Federator.ai.
            displayName: AlamedaService
            kind: AlamedaService
            name: alamedaservices.federatorai.containers.ai
            version: v1alpha1
        description: |-
          Federator.ai helps enterprises optimize cloud resources, maximize application performance, and save significant cost without excessive over-provisioning or under-provisioning of resources, meeting the service-level requirements of their applications.

          Enterprises often lack understanding of the resources needed to support their applications. This leads to either excessive over-provisioning or under-provisioning of resources (CPU, memory, storage). Using machine learning, Federator.ai determines the optimal cloud resources needed to support any workload on OpenShift and helps users find the best-cost instances from cloud providers for their applications.


          **Multi-layer workload prediction**

          Using machine learning and math-based algorithms, Federator.ai predicts containerized application and cluster node resource usage as the basis for resource recommendations at application level as well as at cluster node level. Federator.ai supports prediction for both physical/virtual CPUs and memories.


          **Auto-scaling via resource recommendation**

          Federator.ai utilizes the predicted resource usage to recommend the right number and size of pods for applications. Integrated with Datadog's WPA, applications are automatically scaled to meet the predicted resource usage.


          **Application-aware recommendation execution**

          Optimizing the resource usage and performance goals, Federator.ai uses application specific metrics for workload prediction and pod capacity estimation to auto-scale the right number of pods for best performance without overprovisioning.


          **Multi-cloud Cost Analysis**

          With resource usage prediction, Federator.ai analyzes potential cost of a cluster on different public cloud providers. It also recommend appropriate cluster nodes and instance types based on resource usage.


          **Custom Datadog/Sysdig Dashboards**

          Predefined custom Datadog/Sysdig Dashboards for workload prediction/recommendation visualization for cluster nodes and applications.


          **Additional resources**

          Want more product information? Explore detailed information about using this product and where to find additional help.

          * [Federator.ai Datasheet](https://prophetstor.com/wp-content/uploads/datasheets/Federator.ai.pdf)
          * [Quick Start Guide](https://prophetstor.com/wp-content/uploads/documentation/Federator.ai/Latest%20Version/ProphetStor%20Federator.ai%20Quick%20Installation%20Guide.pdf)
          * [Installation Guide](https://prophetstor.com/wp-content/uploads/2021/08/ProphetStor-Federator.ai-v4.7.0-Installation-Guide-v1.0.pdf)
          * [User Guide](https://prophetstor.com/wp-content/uploads/2021/08/Federator.ai-4.7.0-User-Guide.pdf)
          * [Release Notes](https://prophetstor.com/wp-content/uploads/2021/08/Federator.ai-4.7.0-Release-Notes.pdf)
          * [Company Information](https://www.prophetstor.com/)


          ### Prerequisite
          *  The [Openshift](https://www.openshift.com) version 4.4.x or later if using OpenShift platform.
          *  The [Kubernetes](https://kubernetes.io/) version 1.16 or later if using Kubernetes environment.
        displayName: Federator.ai
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - AI
        - Resource Orchestration
        - NoOps
        links:
        - name: Website
          url: https://prophetstor.com/federator-ai-2/
        - name: Quickstart guide
          url: https://prophetstor.com/wp-content/uploads/documentation/Federator.ai/Latest%20Version/ProphetStor%20Federator.ai%20Quick%20Installation%20Guide.pdf
        maintainers:
        - email: support@prophetstor.com
          name: ProphetStor Data Services, Inc.
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: ProphetStor Data Services, Inc.
        relatedImages:
        - registry.connect.redhat.com/prophetstor/federatorai-operator@sha256:bdd87f6b7d7ba9d12ed8ddb9c5151e5a89781e88d7a02b73814b16a766968047
        - registry.connect.redhat.com/prophetstor/alameda-recommender-ubi@sha256:13e941df9252cea3024ecb5e2247a696b64928015c1a3a4a04ee4e362388b1e4
        - registry.connect.redhat.com/prophetstor/federatorai-data-adapter@sha256:adf5d655fa030dc05133516d45642d7282fdd66e0aa8e7a7331ed614a41d3fe0
        - registry.connect.redhat.com/prophetstor/alameda-operator-ubi@sha256:9a863394cb1c89228223b26addb0d12bb91a97233fe2d5d19ecd5998e0cfe0d6
        - registry.connect.redhat.com/prophetstor/federatorai-agent-ubi@sha256:cfadc84f31f3776ded15d4d8dda92e00b7d9b1bce0ff5440277176292f9304d3
        - registry.connect.redhat.com/prophetstor/federatorai-agent-preloader@sha256:66daea6193e399f8363df5ba42e1fcd643db6df3a9075dbcd8c3464fdc91e776
        - registry.connect.redhat.com/prophetstor/federatorai-rest-ubi@sha256:3e92528615543ea900bf38387c4eb8439b60759be5be2ac5311a9dbaafc7026b
        - registry.connect.redhat.com/prophetstor/alameda-rabbitmq@sha256:b396d9aa0470e593a36f5799c04fc3c9e59b5723f703f4b019f7ef92528e6741
        - registry.connect.redhat.com/prophetstor/alameda-notifier-ubi@sha256:1f75034fb78d62924a87cb4a5435661b4afa990d6960431e784029742fbae0da
        - registry.connect.redhat.com/prophetstor/alameda-ai-dispatcher@sha256:6e0dc1136b0f8e5971d92b95ce0d6dfac475044c073bada50f23b315f4478ae4
        - registry.connect.redhat.com/prophetstor/alameda-ai@sha256:9293298bb2a72d0b0764050b6f6d8da707b46a83363a3b6aa7be60f512e2ec61
        - registry.connect.redhat.com/prophetstor/alameda-analyzer-ubi@sha256:6117ac4692793d49bc0204f13091dec119bdbc7b169e9deb86c243aca00ded27
        - registry.connect.redhat.com/prophetstor/federatorai-dashboard-backend@sha256:022dbf6696edb8eaf389fdbd1dff14c40a51c237c01f8a02a7ea65c60cd59f38
        - registry.connect.redhat.com/prophetstor/alameda-datahub-ubi@sha256:3819d6acce8d376dbe3b12469be2528d0ce9ac95c2a86507aeb76b6e3a5f5f38
        - registry.connect.redhat.com/prophetstor/alameda-executor-ubi@sha256:4512e3be7a89d0ca9f120b68f6ab9766d000bf2b01b279cd55b2dbc433aacfb6
        - registry.connect.redhat.com/prophetstor/fedemeter-api-ubi@sha256:451456207fe8d16b9824b3f64d00645766dcccbc8d685b6942e6c00d0d0c627e
        - registry.connect.redhat.com/prophetstor/federatorai-dashboard-frontend@sha256:1c1147425ca6302103363a1a3bcf27baeef13b10ec6780ef266b7f52fe27ffeb
        - registry.connect.redhat.com/prophetstor/fedemeter-influxdb@sha256:36904f003e8e29d4a9fd738655f401fbe840d60158d7d820cda79a09858dc2c2
        - registry.connect.redhat.com/prophetstor/alameda-influxdb@sha256:485c2119a6a56807e28ad03b2087129950118b55531126a66bb414063052ae76
        version: 4.7.1-1
      entries:
      - name: federatorai.v4.7.1-1
        version: 4.7.1-1
      - name: federatorai.v4.6.1-4
        version: 4.6.1-4
      name: stable
    defaultChannel: stable
    packageName: federatorai-certified
    provider:
      name: ProphetStor Data Services, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Starburst Data, Inc.
      provider-url: https://www.starburstdata.com
    name: starburst-enterprise-helm-operator-paygo-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: starburst-enterprise-helm-operator-paygo.v402.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstEnterprise",
                "metadata": {
                  "name": "starburstenterprise-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "cache": {
                    "diskUsagePercentage": 80,
                    "enabled": false,
                    "ttl": "7d",
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "catalogs": {
                    "tpch": "connector.name=tpch\ntpch.splits-per-node=4"
                  },
                  "commonLabels": {},
                  "coordinator": {
                    "additionalProperties": "",
                    "affinity": {},
                    "deploymentAnnotations": {},
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=true\nnode-scheduler.include-coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://localhost:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "resources": {
                      "cpu": 16,
                      "memory": "60Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "tolerations": []
                  },
                  "environment": null,
                  "expose": {
                    "clusterIp": {
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      }
                    },
                    "ingress": {
                      "annotations": {},
                      "host": null,
                      "ingressClassName": null,
                      "ingressName": "coordinator-ingress",
                      "path": "/",
                      "pathType": "ImplementationSpecific",
                      "serviceName": "starburst",
                      "servicePort": 8080,
                      "tls": {
                        "enabled": true,
                        "secretName": null
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "nodePort": 30080,
                          "port": 8080
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "externalSecrets": {
                    "enabled": false,
                    "goDaddy": {
                      "backendType": "secretsManager"
                    },
                    "secretPrefix": "external0",
                    "type": "goDaddy"
                  },
                  "extraArguments": [],
                  "extraSecret": {
                    "file": null,
                    "name": null
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise",
                    "tag": "402-e.1"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "initImage": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise-init",
                    "tag": "1.3.2"
                  },
                  "internal": {
                    "ports": {
                      "http": {
                        "port": 8080
                      },
                      "https": {
                        "port": 8443
                      }
                    }
                  },
                  "internalTls": false,
                  "licenseManager": {
                    "cronjob": {
                      "affinity": {},
                      "nodeSelector": {},
                      "priorityClassName": null,
                      "securityContext": {},
                      "tolerations": []
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "registry.connect.redhat.com/starburst/license-manager",
                      "tag": "0.0.2"
                    },
                    "url": "https://license-api.dogfood.eng.starburstdata.net/test/license/checkout"
                  },
                  "livenessProbe": null,
                  "prometheus": {
                    "agent": {
                      "config": "/etc/starburst/telemetry/prometheus.yaml",
                      "port": 8081,
                      "version": "0.16.1"
                    },
                    "enabled": true,
                    "rules": [
                      {
                        "attrNameSnakeCase": true,
                        "name": "$1",
                        "pattern": "trino.execution<name=QueryManager><>(running_queries|queued_queries)",
                        "type": "GAUGE"
                      },
                      {
                        "name": "failed_queries",
                        "pattern": "trino.execution<name=QueryManager><>FailedQueries\\.TotalCount",
                        "type": "COUNTER"
                      }
                    ]
                  },
                  "query": {
                    "maxConcurrentQueries": 3
                  },
                  "readinessProbe": null,
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "securityContext": {},
                  "sharedSecret": null,
                  "siac": {
                    "additionalHeapSizePercentage": 15,
                    "enabled": false
                  },
                  "spilling": {
                    "enabled": false,
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "startupProbe": null,
                  "usageMetrics": {
                    "enabled": false,
                    "usageClient": {
                      "initialDelay": "1m",
                      "interval": "1m"
                    }
                  },
                  "userDatabase": {
                    "enabled": false,
                    "users": [
                      {
                        "password": "46991b33f7a75ff79213c0dc0e610610",
                        "username": "admin"
                      }
                    ]
                  },
                  "worker": {
                    "additionalProperties": "",
                    "affinity": {},
                    "autoscaling": {
                      "enabled": false,
                      "maxReplicas": 100,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "deploymentAnnotations": {},
                    "deploymentTerminationGracePeriodSeconds": 300,
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://coordinator:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "kedaScaler": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.connect.redhat.com/starburst/keda-trino-scaler",
                        "tag": "0.1.7"
                      },
                      "port": 8021,
                      "scaledObjectSpec": {
                        "advanced": {},
                        "cooldownPeriod": 300,
                        "idleReplicaCount": 0,
                        "maxReplicaCount": 100,
                        "minReplicaCount": 1,
                        "pollingInterval": 30,
                        "scaleTargetRef": {
                          "name": "worker"
                        },
                        "triggers": [
                          {
                            "metadata": {
                              "numberOfQueriesPerWorker": "10",
                              "scaleInToIdleReplicaCountIfNoQueuedQueriesLeft": "true",
                              "scaleMethod": "query_queue"
                            },
                            "type": "external"
                          }
                        ]
                      }
                    },
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "replicas": 2,
                    "resources": {
                      "cpu": 16,
                      "memory": "100Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "starburstWorkerShutdownGracePeriodSeconds": 120,
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstHive",
                "metadata": {
                  "name": "starbursthive-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "affinity": {},
                  "commonLabels": {},
                  "database": {
                    "external": {
                      "driver": null,
                      "jdbcUrl": null,
                      "password": null,
                      "user": null
                    },
                    "internal": {
                      "databaseName": "hive",
                      "databasePassword": "HivePass1234",
                      "databaseUser": "hive",
                      "driver": "org.postgresql.Driver",
                      "env": [],
                      "envFrom": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.redhat.io/rhel8/postgresql-10",
                        "tag": "1"
                      },
                      "port": 5432,
                      "resources": {
                        "limits": {
                          "cpu": 2,
                          "memory": "1Gi"
                        },
                        "requests": {
                          "cpu": 2,
                          "memory": "1Gi"
                        }
                      },
                      "volume": {
                        "emptyDir": {}
                      }
                    },
                    "type": "internal"
                  },
                  "deploymentAnnotations": {},
                  "env": [],
                  "envFrom": [],
                  "expose": {
                    "clusterIp": {
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "nodePort": 30083,
                          "port": 9083
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "extraArguments": [],
                  "hdfs": {
                    "hadoopUserName": null
                  },
                  "heapSizePercentage": 85,
                  "hiveMetastoreStorageAuthorization": true,
                  "hiveMetastoreWarehouseDir": null,
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/hive",
                    "tag": "3.1.2-e.22"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "nodeSelector": {},
                  "objectStorage": {
                    "adl": {
                      "oauth2": {
                        "clientId": null,
                        "credential": null,
                        "refreshUrl": null
                      }
                    },
                    "awsS3": {
                      "accessKey": null,
                      "endpoint": null,
                      "pathStyleAccess": false,
                      "region": null,
                      "secretKey": null
                    },
                    "azure": {
                      "abfs": {
                        "accessKey": {
                          "accessKey": null,
                          "storageAccount": null
                        },
                        "authType": "accessKey",
                        "oauth": {
                          "clientId": null,
                          "endpoint": null,
                          "secret": null
                        }
                      },
                      "wasb": {
                        "accessKey": null,
                        "storageAccount": null
                      }
                    },
                    "gs": {
                      "cloudKeyFileSecret": null
                    }
                  },
                  "podAnnotations": {},
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "resources": {
                    "limits": {
                      "cpu": 1,
                      "memory": "1Gi"
                    },
                    "requests": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "securityContext": {},
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
          description: Deploys StarburstEnterprise Helm Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-paygo-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-paygo-rhmp/support?utm_source=openshift_console
          marketplace.redhat.com/meterDefinition: |-
            {
              "apiVersion": "marketplace.redhat.com/v1beta1",
              "kind": "MeterDefinition",
              "metadata": {
                "name": "sep-vcpu-usage"
              },
              "spec": {
                "group": "charts.starburstdata.com",
                "kind": "StarburstEnterprise",
                "resourceFilters": [
                  {
                    "namespace": {
                      "useOperatorGroup": true
                    },
                    "ownerCRD": {
                      "apiVersion": "charts.starburstdata.com/v1alpha1",
                      "kind": "StarburstEnterprise"
                    },
                    "workloadType": "Pod"
                  }
                ],
                "meters": [
                  {
                    "aggregation": "sum",
                    "period": "1h",
                    "metricId": "container_cpu_usage_core_avg",
                    "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=~\"coordinator|worker\"}[5m])*100",
                    "workloadType": "Pod"
                  }
                ]
              }
            }
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Starburst Data
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: StarburstEnterprise
            name: starburstenterprises.charts.starburstdata.com
            version: v1alpha1
          - kind: StarburstHive
            name: starbursthives.charts.starburstdata.com
            version: v1alpha1
        description: CRs for deploying starburst enterprise and hive
        displayName: Starburst Enterprise Pay as you go Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - starburst
        - trino
        - sep
        - bigdata
        - data
        links:
        - name: Starburst Enterprise Helm Operator
          url: https://www.starburstdata.com
        maintainers:
        - email: hello@starburstdata.com
          name: Starburst Data, Inc.
        maturity: alpha
        provider:
          name: Starburst Data, Inc.
          url: https://www.starburstdata.com
        relatedImages:
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
        version: 402.1.0
      entries:
      - name: starburst-enterprise-helm-operator-paygo.v402.1.0
        version: 402.1.0
      name: alpha
    defaultChannel: alpha
    packageName: starburst-enterprise-helm-operator-paygo-rhmp
    provider:
      name: Starburst Data, Inc.
      url: https://www.starburstdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: redhat-oadp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oadp-operator.v1.0.19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "restic": {
                      "enable": true
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws",
                        "kubevirt"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ResticRepository",
                "metadata": {
                  "name": "resticrepository",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/oadp/oadp-rhel8-operator@sha256:caf87a9832c153833feb34affc2b797ad0e643362448742b0405d457e5f0ba41
          createdAt: "2025-04-14T13:25:43Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <1.0.19'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/oadp/oadp-mustgather-rhel8@sha256:af4c6af48f38510c7d52f678798a9affea2211dc488ed18f83a5d005b7ef4b6b
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is a Velero resource that represents the capture of
              Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Data Protection Applications on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://github.com/openshift/oadp-operator
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/oadp/oadp-velero-plugin-for-microsoft-azure-rhel8@sha256:3e81b5e55f96504ba76d9d47bbeec50c7b1fd90b7702fc31b1dba248050ce9b4
        - registry.redhat.io/oadp/oadp-velero-plugin-for-gcp-rhel8@sha256:05a164c27da94b374ce03451c851a607a3cd436b73b9e5b70723c6dc8cacb3d8
        - registry.redhat.io/oadp/oadp-mustgather-rhel8@sha256:af4c6af48f38510c7d52f678798a9affea2211dc488ed18f83a5d005b7ef4b6b
        - registry.redhat.io/oadp/oadp-rhel8-operator@sha256:caf87a9832c153833feb34affc2b797ad0e643362448742b0405d457e5f0ba41
        - registry.redhat.io/oadp/oadp-registry-rhel8@sha256:0886b3b4ce0295080bf2c5ee95f2b7b15a4483bcd94cb739e7e7154d21150397
        - registry.redhat.io/oadp/oadp-velero-restic-restore-helper-rhel8@sha256:5d288335b5ad114b9eb61410befb4a9d1f84f0ba0c01558098bf4567fda9a8c5
        - registry.redhat.io/oadp/oadp-velero-plugin-rhel8@sha256:694603b9c1fd5eec939b28f8727fd7c9c0d380620c5081006f2655e98358464f
        - registry.redhat.io/oadp/oadp-velero-plugin-for-csi-rhel8@sha256:05ec16bdf1f0fadc4ddad244b86934f73569d2e66cc19f03b3553e257441a662
        - registry.redhat.io/oadp/oadp-velero-rhel8@sha256:39dec8b9ccdcabeaab3e56e14b79cfed4d8574cb172a886d017813a437e4ffff
        - registry.redhat.io/oadp/oadp-velero-plugin-for-aws-rhel8@sha256:28045538fff9e5051cd01e56876971cf9401a448e1dd63bff17fbb900629f22f
        - registry.redhat.io/oadp/oadp-kubevirt-velero-plugin-rhel8@sha256:d862a1b0420fcfb91403f90047b00035efba112af6bc5c50adb723014a620767
        version: 1.0.19
      entries:
      - name: oadp-operator.v1.0.19
        version: 1.0.19
      - name: oadp-operator.v1.0.18
        version: 1.0.18
      - name: oadp-operator.v1.0.17
        version: 1.0.17
      - name: oadp-operator.v1.0.16
        version: 1.0.16
      - name: oadp-operator.v1.0.15
        version: 1.0.15
      - name: oadp-operator.v1.0.14
        version: 1.0.14
      - name: oadp-operator.v1.0.13
        version: 1.0.13
      - name: oadp-operator.v1.0.12
        version: 1.0.12
      - name: oadp-operator.v1.0.11
        version: 1.0.11
      - name: oadp-operator.v1.0.10
        version: 1.0.10
      - name: oadp-operator.v1.0.9
        version: 1.0.9
      - name: oadp-operator.v1.0.8
        version: 1.0.8
      - name: oadp-operator.v1.0.7
        version: 1.0.7
      - name: oadp-operator.v1.0.6
        version: 1.0.6
      - name: oadp-operator.v1.0.5
        version: 1.0.5
      - name: oadp-operator.v1.0.4
        version: 1.0.4
      - name: oadp-operator.v1.0.3
        version: 1.0.3
      - name: oadp-operator.v1.0.2
        version: 1.0.2
      - name: oadp-operator.v1.0.1
        version: 1.0.1
      - name: oadp-operator.v1.0.0
        version: 1.0.0
      name: stable-1.0
    - currentCSV: oadp-operator.v1.4.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "nodeAgent": {
                      "enable": true,
                      "uploaderType": "restic"
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws",
                        "kubevirt"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: |
            Cloud Provider,Developer Tools,Modernization & Migration,OpenShift Optional,Storage
          certified: "false"
          containerImage: registry.redhat.io/oadp/oadp-rhel9-operator@sha256:ee010fb34448b772690a8a98808aa25aa1578a81061618f948b1d16b5d63d333
          createdAt: "2025-07-10T04:59:11Z"
          description: OADP (OpenShift API for Data Protection) operator sets up and
            installs Data Protection Applications on the OpenShift platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <1.4.5'
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/oadp/oadp-mustgather-rhel9@sha256:0cee6fecf052bae3ede6bb6f032b25c0c7cd63ced1e26328f2b02144f24fbb12
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A backup repository is an indicator of a connection from
              the restic/kopia server to the backupstoragelocation.
            displayName: BackupRepository
            kind: BackupRepository
            name: backuprepositories.velero.io
            version: v1
          - description: Backup is a Velero resource that represents the capture of
              Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - description: The CloudStorage API automates the creation of a bucket for
              object storage.
            displayName: CloudStorage
            kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataDownload represents a data download of a volume snapshot.
              There is one DataDownload created per volume to be restored.
            displayName: DataDownload
            kind: DataDownload
            name: datadownloads.velero.io
            version: v2alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DataUpload acts as the protocol between data mover plugins
              and data mover controller for the datamover backup operation
            displayName: DataUpload
            kind: DataUpload
            name: datauploads.velero.io
            version: v2alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: |
          **OpenShift API for Data Protection (OADP)** operator sets up and installs
          Velero on the OpenShift platform, allowing users to backup and restore
          applications.

          Backup and restore Kubernetes resources and internal images, at the
          granularity of a namespace, using a version of Velero appropriate for
          the installed version of OADP.

          OADP backs up Kubernetes objects and internal images by saving them as
          an archive file on object storage. OADP backs up persistent volumes
          (PVs) by creating snapshots with the native cloud snapshot API or with
          the Container Storage Interface (CSI). For cloud providers that do not
          support snapshots, OADP backs up resources and PV data with Restic or Kopia.

          * [Installing OADP for application backup and restore][install-doc]

          * [Installing OADP on a ROSA cluster and using STS, please follow the Getting Started Steps 1-3 in order to obtain the role ARN needed for using the standardized STS configuration flow via OLM][rosa-sts-doc]

          * [Frequently Asked Questions][faq-doc]

          [install-doc]: https://docs.openshift.com/container-platform/latest/backup_and_restore/application_backup_and_restore/installing/about-installing-oadp.html
          [rosa-sts-doc]: https://mobb.ninja/docs/misc/oadp/rosa-sts/
          [faq-doc]: https://access.redhat.com/articles/5456281
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - disaster backup
        - disaster recovery
        - disaster restore
        - disaster
        - backup
        - restore
        - backup automation
        - recovery
        - data protection
        - data management
        - data
        - protection
        - management
        - application resilience
        - resilience
        - velero
        - openshift
        - oadp
        - cloud-native
        - replication
        - kopia
        - restic
        - network file system
        - nfs
        - simple cloud storage
        - s3
        - minio
        - cloud storage
        - data foundation
        - ceph
        - csi
        - container storage interface
        links:
        - name: OADP Operator
          url: https://github.com/openshift/oadp-operator
        maintainers:
        - email: oadp-maintainers@redhat.com
          name: OADP Operator maintainers
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/oadp/oadp-velero-plugin-for-legacy-aws-rhel9@sha256:ab5c8d33684a361fa3bb10a006ef489be92cc955ad3d173bbfa77949dcc42aa8
        - registry.redhat.io/oadp/oadp-kubevirt-velero-plugin-rhel9@sha256:c3b32cb9f5b82021c08cd70467e68aad66746ef211d73612ecfe0af625706267
        - registry.redhat.io/oadp/oadp-velero-plugin-for-aws-rhel9@sha256:79d39186cae5696949813a0ea6c46eeededad29e6bde171d7a9b0011c993bb26
        - registry.redhat.io/oadp/oadp-velero-plugin-for-microsoft-azure-rhel9@sha256:9f05592bf0e5b76322ca582bdede387f9f4aabbbdc436549156c05d652c951d0
        - registry.redhat.io/oadp/oadp-mustgather-rhel9@sha256:0cee6fecf052bae3ede6bb6f032b25c0c7cd63ced1e26328f2b02144f24fbb12
        - registry.redhat.io/oadp/oadp-rhel9-operator@sha256:ee010fb34448b772690a8a98808aa25aa1578a81061618f948b1d16b5d63d333
        - registry.redhat.io/oadp/oadp-velero-rhel9@sha256:17160f99940aa355dc91ab3d16ba6bf91faade6423434008cf8f6e5023b1e413
        - registry.redhat.io/oadp/oadp-velero-plugin-for-gcp-rhel9@sha256:00800ead11ea308794b1536d964f8a043e13ede9657cabb23afa42cba7cf2ce6
        - registry.redhat.io/oadp/oadp-velero-plugin-rhel9@sha256:3ee9d3e799cedc28c400d364b3a8790c7e88acd5f603b09ff512a22d91e265b3
        - registry.redhat.io/oadp/oadp-velero-restic-restore-helper-rhel9@sha256:e0647defc7c428f9a04a5f6d4532dc86e57b92f31b4f907253ae14af05c5993f
        version: 1.4.5
      entries:
      - name: oadp-operator.v1.4.5
        version: 1.4.5
      - name: oadp-operator.v1.4.4
        version: 1.4.4
      - name: oadp-operator.v1.4.3
        version: 1.4.3
      - name: oadp-operator.v1.4.2
        version: 1.4.2
      - name: oadp-operator.v1.4.1
        version: 1.4.1
      - name: oadp-operator.v1.4.0
        version: 1.4.0
      name: stable-1.4
    defaultChannel: stable-1.4
    packageName: redhat-oadp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Oracle
      provider-url: https://www.oracle.com/cloud/
    name: oci-ccm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oci-ccm-operator.v0.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ccm.oraclecloud.com/v1alpha1",
                "kind": "OCICloudManager",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "oci-ccm-operator",
                    "app.kubernetes.io/instance": "ocicloudmanager",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ocicloudmanager",
                    "app.kubernetes.io/part-of": "oci-ccm-operator"
                  },
                  "name": "ocicloudmanager"
                },
                "spec": {
                  "compartment": "CompartmentId",
                  "driverNamespace": "kube-system",
                  "loadBalancer": {
                    "securityList": "SecurityListId",
                    "securityListManagementMode": "Frontend",
                    "subnet": "SubnetId"
                  },
                  "vcn": "VcnId"
                }
              },
              {
                "apiVersion": "ccm.oraclecloud.com/v1alpha1",
                "kind": "OCICsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "oci-ccm-operator",
                    "app.kubernetes.io/instance": "ocicsidriver",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ocicsidriver",
                    "app.kubernetes.io/part-of": "oci-ccm-operator"
                  },
                  "name": "ocicsidriver"
                },
                "spec": {
                  "compartment": "CompartmentId",
                  "driverNamespace": "kube-system",
                  "vcn": "VcnId"
                }
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider, Storage
          containerImage: phx.ocir.io/ax1eoqvpvbtg/oci-ccm-operator:0.0.9
          createdAt: "2023-07-28T23:42:18Z"
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Oracle
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: OCICloudManager
            name: ocicloudmanagers.ccm.oraclecloud.com
            version: v1alpha1
          - kind: OCICsiDriver
            name: ocicsidrivers.ccm.oraclecloud.com
            version: v1alpha1
        description: OCI Kubernetes Cloud Controller Manager implementation (or out-of-tree
          cloud-provider) for Oracle Cloud Infrastructure (OCI).
        displayName: oci-ccm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OCI
        - CCM
        - OpenShift
        - Kubernetes
        links:
        - name: Oracle
          url: https://www.oracle.com/cloud/
        - name: OCI Cloud Controller Manager (CCM)
          url: https://github.com/oracle/oci-cloud-controller-manager
        maintainers:
        - email: david.d.campbell@oracle.com
          name: David Campbell
        - email: alex.blinov@oracle.com
          name: Alex Blinov
        maturity: alpha
        provider:
          name: Oracle
          url: https://www.oracle.com/cloud/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - phx.ocir.io/ax1eoqvpvbtg/oci-ccm-operator:0.0.9
        version: 0.0.9
      entries:
      - name: oci-ccm-operator.v0.0.9
        version: 0.0.9
      name: alpha
    defaultChannel: alpha
    packageName: oci-ccm-operator
    provider:
      name: Oracle
      url: https://www.oracle.com/cloud/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: eventing-kogito
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eventing-kogito.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kogito.knative.dev/v1alpha1",
                "kind": "KogitoSource",
                "metadata": {
                  "name": "kogito-order-processing"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "kind": "Broker",
                      "name": "default"
                    }
                  },
                  "subject": {
                    "apiVersion": "serving.knative.dev/v1",
                    "kind": "Service",
                    "name": "serverless-workflow-order-processing"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: gcr.io/knative-releases/knative.dev/eventing-kogito/cmd/controller:v1.2.0
          createdAt: "2021-08-25T12:00:00Z"
          olm.targetNamespaces: knative-kogito
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/knative-sandbox/eventing-kogito
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoSource is the representation of a Knative Eventing
              Source for a Kogito custom Service.
            displayName: Kogito Source
            kind: KogitoSource
            name: kogitosources.kogito.knative.dev
            version: v1alpha1
        description: |-
          Knative Eventing Source for Kogito Services. This controller will enable easy configuration of any Kogito custom Service
          with the Knative Eventing platform. Please find more information about using this source in the [Official Kogito Documentation](https://docs.jboss.org/kogito/release/latest/html_single/#_kogito_knative_eventing_add_on)

          **Note:** The admission webhooks are not installed via OLM at this time. Because of that, please ensure that you create the `KogitoSource` objects with the subbject and sink information with the appropriate namespace.
        displayName: Kogito Knative Eventing Source
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless workflow
        - kogito
        - knative
        - dmn
        - bpmn
        - eventing
        - cloudevents
        - messaging
        links:
        - name: Source Repository
          url: https://github.com/knative-sandbox/eventing-kogito
        - name: Kogito Knative Add-On Documentation
          url: https://docs.jboss.org/kogito/release/latest/html_single/#_kogito_knative_eventing_add_on
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - gcr.io/knative-releases/knative.dev/eventing-kogito/cmd/controller:v1.2.0
        version: 1.2.0
      entries:
      - name: eventing-kogito.v1.2.0
        version: 1.2.0
      - name: eventing-kogito.v1.1.0
        version: 1.1.0
      name: alpha
    defaultChannel: alpha
    packageName: eventing-kogito
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KEDA Community
      provider-url: ""
    name: keda
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keda.v2.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "KedaController",
                "metadata": {
                  "name": "keda",
                  "namespace": "keda"
                },
                "spec": {
                  "watchNamespace": "",
                  "operator": {
                    "logLevel": "info",
                    "logEncoder": "console"
                  },
                  "metricsServer": {
                    "logLevel": "0"
                  },
                  "serviceAccount": {}
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledJob",
                "metadata": {
                  "name": "scaledjob-sample"
                },
                "spec": {
                  "failedJobsHistoryLimit": 5,
                  "jobTargetRef": {
                    "activeDeadlineSeconds": 600,
                    "backoffLimit": 6,
                    "completions": 1,
                    "parallelism": 1,
                    "template": null
                  },
                  "maxReplicaCount": 100,
                  "pollingInterval": 30,
                  "successfulJobsHistoryLimit": 5,
                  "triggers": [
                    {
                      "metadata": {
                        "property": "examle-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledObject",
                "metadata": {
                  "name": "example-scaledobject"
                },
                "spec": {
                  "cooldownPeriod": 300,
                  "maxReplicaCount": 100,
                  "minReplicaCount": 0,
                  "pollingInterval": 30,
                  "scaleTargetRef": {
                    "name": "example-deployment"
                  },
                  "triggers": [
                    {
                      "metadata": {
                        "property": "examle-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "TriggerAuthentication",
                "metadata": {
                  "name": "example-triggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "exmaple-secret-parameter"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ClusterTriggerAuthentication",
                "metadata": {
                  "name": "example-clustertriggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "exmaple-secret-parameter"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Cloud Provider
          certified: "false"
          containerImage: ghcr.io/kedacore/keda-olm-operator:2.7.1
          createdAt: "2022-05-11T00:00:00.000Z"
          description: '[DEPRECATED] Use Custom Metrics Autoscaler Operator instead'
          operatorframework.io/suggested-namespace: keda
          operators.operatorframework.io/builder: operator-sdk-v1.1.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/kedacore/keda-olm-operator
          support: KEDA Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              Represents an installation of a particular version of KEDA Controller.
            displayName: KedaController
            kind: KedaController
            name: kedacontrollers.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your jobs and what the triggers
            displayName: ScaledJob
            kind: ScaledJob
            name: scaledjobs.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your application and what the
              triggers are
            displayName: ScaledObject
            kind: ScaledObject
            name: scaledobjects.keda.sh
            version: v1alpha1
          - description: Describes authentication parameters
            displayName: TriggerAuthentication
            kind: TriggerAuthentication
            name: triggerauthentications.keda.sh
            version: v1alpha1
          - description: Describes cluster wide authentication parameters
            displayName: ClusterTriggerAuthentication
            kind: ClusterTriggerAuthentication
            name: clustertriggerauthentications.keda.sh
            version: v1alpha1
        description: "## Deprecated\n This KEDA community operator is deprecated,
          please use `Custom Metrics Autoscaler` Operator instead.\n## About the managed
          application\nKEDA is a Kubernetes-based event driven autoscaler.  KEDA can
          monitor event sources like Kafka, RabbitMQ, or cloud event sources and feed
          the metrics from those sources into the Kubernetes horizontal pod autoscaler.
          \ With KEDA, you can have event driven and serverless scale of deployments
          within any Kubernetes cluster.\n## About this Operator\nThe KEDA Operator
          deploys and manages installation of KEDA Controller in the cluster. Install
          this operator and follow installation instructions on how to install KEDA
          in you cluster.\n\n## Prerequisites for enabling this Operator\nPlease note
          that you can not run both KEDA 1.x and 2.x on the same Kubernetes cluster.
          You need to uninstall KEDA 1.x first, in order to install and use KEDA 2.x.\n\nDon't
          forget to uninstall KEDA v1 CRDs as well, to ensure that, please run:\n\n```
          oc delete crd scaledobjects.keda.k8s.io ```\n\n``` oc delete crd triggerauthentications.keda.k8s.io
          \ ```\n## How to install KEDA in the cluster\nThe installation of KEDA is
          triggered by the creation of `KedaController` resource, please refer to
          [KedaController Example](https://github.com/kedacore/keda-olm-operator#kedacontroller-spec)
          for more deatils on available options.\n\nOnly resource named `keda` in
          namespace `keda` will trigger the installation, reconfiguration or removal
          of the KEDA Controller resource.\n\nThere could be only one KEDA Controller
          in the cluster. \n"
        displayName: '[DEPRECATED] KEDA'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keda
        - scaling
        - event
        links:
        - name: Homepage
          url: https://keda.sh
        - name: Source Code
          url: https://github.com/kedacore/keda
        - name: Slack Channel
          url: https://kubernetes.slack.com/messages/CKZJ36A5D
        maintainers:
        - email: cncf-keda-maintainers@lists.cncf.io
          name: KEDA Community
        maturity: deprecated
        minKubeVersion: 1.17.0
        provider:
          name: KEDA Community
        relatedImages:
        - ghcr.io/kedacore/keda-olm-operator:2.7.1
        version: 2.7.1
      entries:
      - name: keda.v2.7.1
        version: 2.7.1
      - name: keda.v2.6.1
        version: 2.6.1
      - name: keda.v2.6.0
        version: 2.6.0
      - name: keda.v2.5.0
        version: 2.5.0
      - name: keda.v2.4.0
        version: 2.4.0
      - name: keda.v2.3.0
        version: 2.3.0
      - name: keda.v2.2.0
        version: 2.2.0
      - name: keda.v2.1.0
        version: 2.1.0
      - name: keda.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: keda
    provider:
      name: KEDA Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: stolostron
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stolostron.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:3443da607b276d39ba28a8d6561d296f30e9bac5ab06d5c70715bddbde0a58b2
          createdAt: "2024-06-28T04:52:57Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/multiclusterhub-operator@sha256:3443da607b276d39ba28a8d6561d296f30e9bac5ab06d5c70715bddbde0a58b2
        - quay.io/stolostron/governance-policy-addon-controller@sha256:593dc0a689091ce7c916e6be764c222c1020629ccff7eafbf17524936e4f8262
        - quay.io/stolostron/observability-e2e-test@sha256:8d9341b1814cd085a20f9feceab181a8989d6684fd44d6c8e54a0c6df4dd7216
        - quay.io/stolostron/governance-policy-framework-addon@sha256:a1c0d40561befaf5e849060beaee4b201f42d77c5090823f23ef43a75106b00b
        - quay.io/stolostron/thanos-receive-controller@sha256:b1a191714cead24f66e8de7217db3c4d4a253ecf0238a912c8b88ba8b47c972d
        - quay.io/stolostron/console@sha256:c7900403ef8dba3aac52f999532d876bbfe2f66f00a255047850594b64021830
        - quay.io/stolostron/volsync-mover-rclone@sha256:b247b23944e20104f801c172cbbb620740a5b51d1777ecdbf98b4c128f1ef171
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/prometheus@sha256:8250af047c03c6b0c07b76420e0a5c2338f053d73b2dca1ecba4bb2f211715fe
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/search-collector@sha256:136ad44ff41a29dec48e87d10af5b608f335e4242c41ac17cba2c3c4c8205bd6
        - quay.io/stolostron/multicluster-operators-channel@sha256:d791aa2c6032c1de50cf4033c426646613b182e98c406a091b88fb9ffefc0c98
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/helloworld@sha256:81a0b06c7fcf915b1bb40d33d29d0bfd4d501404463c1f46bffc503e7d1cc45b
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:02e53b8a6529e3209809dada4113054e3398de18738948cb01ed1d24dcb8e8be
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/insights-metrics@sha256:49a177357e865d5f8347840fe695531d539d17e28d2b9971821e41ebbee014bd
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/rbac-query-proxy@sha256:24aafcba20ded0892209cfcc196bff88f877ed51eb928d5d350b17041d170893
        - quay.io/stolostron/volsync-mover-restic@sha256:284c3358f0f2158d599f27487d76a7fdb21842fc0e95bc639726dfdad14b7e7d
        - quay.io/stolostron/prometheus-config-reloader@sha256:bc93ca9f1b6d95e8b963fe1af4c17f87c8c7d938a348627c29cc3cf42b70ca3e
        - quay.io/stolostron/prometheus-alertmanager@sha256:8222bbdd37c246703de54ed3dc109fa8c2f6ca17c5bb90106948683858afdc65
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/multicloud-integrations@sha256:60be5b945ead8f845a29a8be18bd26922b83f009fde1412017f1ca985eb79b55
        - quay.io/stolostron/grafana@sha256:e40d66da749108ae1eecf909c95c87bf7a289328e8b373ae4402368c474f2da5
        - quay.io/stolostron/observatorium@sha256:af4e315e4c6da3607f84e24f31e0f37f8572d161736b4748b2e57e0a2c2dd36c
        - quay.io/stolostron/volsync-addon-controller@sha256:7774767ed20263bba4857a0d4be012e12215f640fce01fbd3a085bf860d087e7
        - quay.io/stolostron/volsync-mover-syncthing@sha256:b5fad266ca8d3cd8efcbf4e8241c446004747b551b2b8e06acae10900d7a0f1b
        - quay.io/stolostron/thanos@sha256:cc78cd786643d27ed85eb3670852722e915a20b20a21479e1477fb859dec6c5e
        - quay.io/stolostron/prometheus-operator@sha256:490b0056b9a23a1dc514a3036eac6593c7c8961de27d08ff24817d280325138c
        - quay.io/stolostron/cluster-backup-controller@sha256:9c77444bd42e6e99163f4dce9f9975276aea082100f53fd938cddf74ad3b1574
        - quay.io/stolostron/multicluster-operators-application@sha256:199b696dbd7589b77e2391b71504fbcb7d3ae785972dac3d001c4c1bffa342dd
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/kube-rbac-proxy@sha256:9626cb9b2ee8f00d2abdb14cd2ce56d77435859071c10d8394bc7ee2754ce87f
        - quay.io/stolostron/volsync-mover-rsync@sha256:f7affcc16ae9bf6d12f69511b7a269f5fd065d1d87392c4c38cb5d7af5b89e23
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/search-indexer@sha256:068c38489201316dc12d475de03c8abb11575338bd75e25c1262eca8aac49cb2
        - quay.io/stolostron/search-v2-api@sha256:6b97b2c7e64de7cd154fd83da3bc690a57b67f01c83170238fe9c2e764df91a8
        - quay.io/stolostron/cert-policy-controller@sha256:27673358610dc1ca308de472dbd794dcbb38250bb299b53c736549b1e8441a0f
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/metrics-collector@sha256:4101a2f2c4230bdc1420a58b546a241dfc5414cf92b57f3b73619c1be14dbc32
        - quay.io/stolostron/observatorium-operator@sha256:5da7f478ecf3609a3b86eb563e4dd22588f3d13efdc2e33f5b672fe95583ce68
        - quay.io/stolostron/grafana-dashboard-loader@sha256:cff97fe3e32dd14ed2f3b7b02cb205715cfee6cfd6dc7e7a630101ea5147e566
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:9d1d5e2b5313b898188d2c845dcda5d604b73f16e89c71915ea308281c1f56ff
        - quay.io/stolostron/klusterlet-addon-controller@sha256:01e5fca5e5c6caca6878bf8025471b7c77164ca7b4a2d764e20101fc79290df0
        - quay.io/stolostron/search-e2e@sha256:8dfbc899ba1ffaa8fb6eb2861f6db87a13d2565cf4551fd47df7fe8253217fb2
        - quay.io/stolostron/config-policy-controller@sha256:197d8491e7c889db8f2acb726e0146d384c979acc7f634c24a9d4b21f7d444f3
        - quay.io/stolostron/must-gather@sha256:1f7c129859bdaf6bd93dda77aec01f7cd03000fbfaccd809f26d33b9e3f66e51
        - quay.io/stolostron/search-v2-operator@sha256:c41517f372fc1874895b39e70cd1c07391110697a5e49ebcf56f4a5df499d810
        - quay.io/stolostron/kube-state-metrics@sha256:ef1d671ffa098ed95186de6f15ea91fe648d103c1c7a57871ed46af07c6f7881
        - quay.io/stolostron/submariner-addon@sha256:0f1b4ae59ff5c933542f9affed20bfa6694deb804f2e97f49ebd0f2e1b925eca
        - quay.io/stolostron/insights-client@sha256:aac5b9b66800c0b862e8bfa1fea533a4aec854d8c52baddbfa03977e1ed520cb
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:bf6bec399bc2072186e9d12ed865167fe940f3f619d69ffe6c178c14a3bdadbe
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/multicluster-observability-operator@sha256:245e1815a73b5c48f790ae5a5d85fac6642cce8a5a3bf122275809e97a4d519e
        - quay.io/stolostron/grc-policy-framework-tests@sha256:f2b01192fd900860428942c7d5b900604ffcab0d42731676a1acdeebd3716557
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/cluster-permission@sha256:716d2cbcb7f69b5640d04669b15884e0adf053597f278a0bb1d8ee77e4a7af4d
        - quay.io/stolostron/governance-policy-propagator@sha256:200558b5aa2366383bd298e159d9f7966a2c9b981038c79cde536edbfbe0092c
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/volsync@sha256:ba7399b316359118b1c1b915d667fca9a4a0d12a10beae69e6dee9c7236753be
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:7f6b6e45f0be0a14d7d9d79828572b9400f4d559f084028e3254cfd40948780d
        - quay.io/stolostron/multicluster-operators-subscription@sha256:8350df0a7d4f19c8e52eb58eecf5223bb2a0ccbe4d4aa00fcf8dff832cb43e5d
        version: 0.4.0
      entries:
      - name: stolostron.v0.4.0
        version: 0.4.0
      name: community-0.4
    - currentCSV: stolostron.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:781f6712369cd4b9bdc8ba0ee2f35a12db198ca190621e199dc10cfea3f5efc9
          createdAt: "2024-10-02T08:01:20Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.5.0 <0.5.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/governance-policy-framework-addon@sha256:5f2b326a279b7c42c233ea5f11b89ecaf4cfd7c549d021ddcc4e7b0c8926fc6d
        - quay.io/stolostron/rbac-query-proxy@sha256:2f1c78176d0ca0b21c237a9f7a3f77f7d857dd848f3a42254cee1e4e469b4c3c
        - quay.io/stolostron/multiclusterhub-operator@sha256:781f6712369cd4b9bdc8ba0ee2f35a12db198ca190621e199dc10cfea3f5efc9
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:d42ddd30570ccdf388ccd65cd3f7c72de2d857d328e05aef8e855508ccf91898
        - quay.io/stolostron/observatorium@sha256:8d0f02996b5bf02be02ff3743821909b6c2396f0f0519652dde0c5d1d76f8a05
        - quay.io/stolostron/prometheus@sha256:147f3a0cee8fa7b11c82e21ca95c28278e1ab53ccb5b70f3af0425bb8a74f61b
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/volsync-mover-syncthing@sha256:b5fad266ca8d3cd8efcbf4e8241c446004747b551b2b8e06acae10900d7a0f1b
        - quay.io/stolostron/search-collector@sha256:4082f7a37aa3f10c725bb9bebf94ee8dab97d22302cd5e59b35d0e3bcc888cef
        - quay.io/stolostron/volsync-mover-rclone@sha256:b247b23944e20104f801c172cbbb620740a5b51d1777ecdbf98b4c128f1ef171
        - quay.io/stolostron/insights-metrics@sha256:49a177357e865d5f8347840fe695531d539d17e28d2b9971821e41ebbee014bd
        - quay.io/stolostron/acm-cli@sha256:85a5c55dfa4708435807d1266b5607525e33f55c5b4672fd8ba78987a1c48ef3
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/thanos-receive-controller@sha256:58c97fe5ec2131f9dcfb70563b59ed0bcae94d8bddefa51cb4784e73b1d90e4f
        - quay.io/stolostron/cluster-backup-controller@sha256:688844c722db78f383d90986150bde93bb421068e7f5b4128d19043f3e01f93e
        - quay.io/stolostron/config-policy-controller@sha256:5af6a0e17da7b7261ac71bcb1e2d991856bbb5450686d188026798e05f0617ba
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/volsync-addon-controller@sha256:d0b15da92e0ff9792d750b7b48d72b38d6abb91488c20578c7be3382c561877f
        - quay.io/stolostron/observability-e2e-test@sha256:c361f5a0f806498c6c1150c0c0d5a42222efff13a18cb28dbfd5e4bd835cb4b7
        - quay.io/stolostron/siteconfig-operator@sha256:7f47f0c61d5b53411ed78046c5a02664c6693d940ef37d6db1853670196be334
        - quay.io/stolostron/insights-client@sha256:aac5b9b66800c0b862e8bfa1fea533a4aec854d8c52baddbfa03977e1ed520cb
        - quay.io/stolostron/kube-rbac-proxy@sha256:1b708bff7fb46391f2460d34be28d30add38501cdb084306df01faae75637232
        - quay.io/stolostron/multicluster-observability-operator@sha256:fb0519952ef6436cc66496409f00e89cd6e58a41776df96859a86176a7322d05
        - quay.io/stolostron/multicluster-operators-subscription@sha256:30b1b5bf6a7a12e525450905a474babdd7e38235d9be60f3cb04d60a745a1b96
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:a8d812a527902bd4d9dfd23f90ce90f415f5a6fabb007d1c41b4c998e1ca098a
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/observatorium-operator@sha256:38f22e30db7713e9f2e239463803cdb5cdfe7108a524ffa5d7e95201bae5300b
        - quay.io/stolostron/cluster-permission@sha256:716d2cbcb7f69b5640d04669b15884e0adf053597f278a0bb1d8ee77e4a7af4d
        - quay.io/stolostron/thanos@sha256:dd4d4f8fbb9a8f51134260fc728d19f61a93b511fd7889636089944ac46163bd
        - quay.io/stolostron/search-v2-operator@sha256:6df11a36ea19a1b3da191964f851dee95b87fd993e8f0ef8906b1e63dd450d17
        - quay.io/stolostron/metrics-collector@sha256:c580106b2c7173251afa6714b710fd109811023d88a8009fb741f3d00e486e76
        - quay.io/stolostron/volsync-mover-rsync@sha256:f7affcc16ae9bf6d12f69511b7a269f5fd065d1d87392c4c38cb5d7af5b89e23
        - quay.io/stolostron/search-v2-api@sha256:338634d767383716c7a88ed2bc9c58275e9d5953932293443725883d7efcffd5
        - quay.io/stolostron/multicluster-operators-application@sha256:329c82b92fe2448fe26923fbd82d2eadacb7f96c75689d96393c573f95e6d650
        - quay.io/stolostron/must-gather@sha256:362bbfd54059975369ce6d3869d34e346a0e76159163714cfcb3597022b7c88f
        - quay.io/stolostron/prometheus-config-reloader@sha256:337f7d2036c17459a41884e38c569a3601364b02660946b49d39ce40ef97bd48
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/governance-policy-addon-controller@sha256:9f387eb8b54d1c86fc20d205dbce199738849cb0ce15a47052ddbee8781d2d9b
        - quay.io/stolostron/volsync-mover-restic@sha256:284c3358f0f2158d599f27487d76a7fdb21842fc0e95bc639726dfdad14b7e7d
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/search-e2e@sha256:ddaa50ecfafcd7501903420a47495f733dce36fcb91c8ce9b8d94f4448ed501d
        - quay.io/stolostron/submariner-addon@sha256:07287987fb0041efc16557cb9d7403dc15b5959a5d0dd0d81cd72bc2658cba10
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/klusterlet-addon-controller@sha256:cf4cad3599c4ce538dc007cda9f5b06538799e8198912d907058ecdd4bbb2451
        - quay.io/stolostron/multicloud-integrations@sha256:d4cb3a0a50bc5d22f48c96ed3a10ea9843c6ef2f83daeae411a7ddb7ea332ea6
        - quay.io/stolostron/prometheus-operator@sha256:5ff1f7a4be0216f582f5cb40de3412cbe8126433144c8201750f6c3c0fee920e
        - quay.io/stolostron/governance-policy-propagator@sha256:69ab8f492fa418596db6da578b3b579fc2f1506c5bf780e6aba3ccd05d8b02b9
        - quay.io/stolostron/console@sha256:e42d2f7e39a93f13003193142a3d1f3f9e55966cbf55b94eac93c726373a2516
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/volsync@sha256:ba7399b316359118b1c1b915d667fca9a4a0d12a10beae69e6dee9c7236753be
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:dbdbec1600d57c440173aecf19c57709d69e945dcecb348bcf4a7c37b3700b70
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/search-indexer@sha256:7471f2808686d42c2eae9f0a537f9b520648f31e87072a376243f309fd044caa
        - quay.io/stolostron/grafana@sha256:c9869c7f590bc4c54151c7e4fa9ed76dfe35c24a0ea3c19a6a1b26e61924b6e1
        - quay.io/stolostron/cert-policy-controller@sha256:809cae1aef02dd3fb68143ebdba818b20c661bb628a7922ee1d705efe4947f35
        - quay.io/stolostron/helloworld@sha256:48724b53542a3fcaf342d6ae2056bdf846cbe2ee50879613953b045bfcbe1c0a
        - quay.io/stolostron/kube-state-metrics@sha256:dc534373feda042afa385026c7da674912ca310df2a4598d83451d4474bb3f44
        - quay.io/stolostron/grc-policy-framework-tests@sha256:57794e2b15949cedf3462df2d38b95e86ca1e140a5ae81e4a9b3e7e06072847d
        - quay.io/stolostron/prometheus-alertmanager@sha256:19f2e277cbdd2bd4814b8d8034167c1a7cfe9f1d698fb16f7604fa65cb8bcb1c
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/multicluster-operators-channel@sha256:260a11aae954299ce8fbb0f9c3102afd3b5cf3f1b972b5a71e56c273ba1d532f
        - quay.io/stolostron/grafana-dashboard-loader@sha256:85031d1a56514be239c585487aec05a8cbb08c17aeb60c0c84c685eb9d9f08b5
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:358ac0b0a3afdfac06f784f2e02a305cbef4e6f648e31a00981d3b097746412a
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        version: 0.5.3
      entries:
      - name: stolostron.v0.5.3
        version: 0.5.3
      - name: stolostron.v0.5.2
        version: 0.5.2
      - name: stolostron.v0.5.1
        version: 0.5.1
      - name: stolostron.v0.5.0
        version: 0.5.0
      name: community-0.5
    - currentCSV: stolostron.v0.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:09f1f4266938fba88376a50c3e2837c5517b4f3a947cf86638a17e52eb17c1f6
          createdAt: "2025-02-24T01:42:20Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.6.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/flightctl-worker@sha256:4dddded7d2d4cce63972c8ca81f213c2cedb551cc645ec85495a9692dc36bb4c
        - quay.io/stolostron/helloworld@sha256:56f0d641d4f07b922f1c28c536eacad747c56a564e0a65f084106c2ffc9bd99d
        - quay.io/stolostron/acm-cli@sha256:8c8c059ce2f6d131c410972bf8fd0b7e399cf17cd044c7ddc89a9668c0ceac4c
        - quay.io/stolostron/search-v2-api@sha256:694523242aabb8db1f27b5afcd8ab2f63a9cdb6056c33a79badf8e68a337738d
        - quay.io/stolostron/multicluster-observability-operator@sha256:bed0dd30646d7fce894a625e6cf52e7829c272f47dde0baf1f788e9f85f66e42
        - quay.io/stolostron/cert-policy-controller@sha256:98dd74d7375ca2b546075d5c8f06a9e2c4f2ce3b77247dccf36bcaa46bf564c7
        - quay.io/stolostron/governance-policy-propagator@sha256:c21574a2421fefe3fcabeef14640d6fe699d950106c9ed056b3ba9be3cbeabe0
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/flightctl-periodic@sha256:a3ed745dc65a5886c4cc05348cf4822bfbf2c6d28b85a39fa0b8f3e1f53fdf8a
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/search-indexer@sha256:4f9be902fe976fd88a77b737cc1f56f00d48740a6a62661468328cf3bb9da31a
        - quay.io/stolostron/search-v2-operator@sha256:85e53a97240f6fd1c4251101146cc2bcc51eb91d9b171efb5b58edade9e0244e
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/multicluster-operators-application@sha256:59802fcfae5b6fd0bebe58e98412885e71f32464b8dbbf85364ddaf090aa3e00
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:9c85fb91c4be3287cac2a376465836689df885902099096f0f76ffca50a6d311
        - quay.io/stolostron/search-e2e@sha256:43e1bbfb7259612ffb408108c71b825f6e61a268c7e405b8d855f6878b6f96dd
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/flightctl-api@sha256:64fb1b655bef6aeff03cf64b6450155d3cb570ed4713f084590fe515f4fbad02
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:51a8acc47098175db148ebc19a10cdba2803d4a9f5884dd38b1c839702398789
        - quay.io/stolostron/thanos@sha256:53eee303836070a7c7621bfb2a094e3809146497d1653dc65d16055aacea332c
        - quay.io/stolostron/siteconfig-operator@sha256:97dc128666780ec9a4c955f392ab98f81f0d6038ec2159e8b1d55aab393cf118
        - quay.io/stolostron/config-policy-controller@sha256:8cf54eb6f5b49a60ed165fa4b70e4984a1362f2f31e8304c3a22e8a6a953a606
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/governance-policy-addon-controller@sha256:fa0738abc0368f9dadd8b0845d868512d26be3525a1a07b9de8a9cefafaa42b5
        - quay.io/stolostron/klusterlet-addon-controller@sha256:39e6940eb956fe9640342e11ecbcbc7d6df0169e1b340a60afc19aff7dd713a5
        - quay.io/stolostron/thanos-receive-controller@sha256:d73f98e93a258adffc1a4bd875c79a9629166599e68ef777a790a1f3146a2929
        - quay.io/stolostron/cluster-permission@sha256:443d8b30e3bb83c139c64ddfc38b0d5e6d7a25a4fae783303ab2f2c27d015b26
        - quay.io/stolostron/grafana@sha256:51092f6754106a90aff28cdd80c2f0447c85c1bbc72db15878d0b65b93e1ead6
        - quay.io/stolostron/volsync-addon-controller@sha256:d9dad837f2f735674e5711c7fff1cfe2c0c92160fceb7d1548a62348dd43b81c
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:92369b6deb21ab3cb0a12ae5c74b455cfefb8698826bca39e26648e608e48a2b
        - quay.io/stolostron/observatorium@sha256:398ecee4d3f12e84e48b12009a4b02fc8b360dfc0c1607ffecc26e6bc04d0ea2
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/prometheus-config-reloader@sha256:2b26d2fb5c69e54aecfbcaab46e652d136d726f0604844031c8771874f1fd769
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/grc-policy-framework-tests@sha256:9b84dbafef35bd568b2bf79283b6850ba1ff57d70239a27cc0cc815fb1a55515
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/metrics-collector@sha256:c7433ee4ccdb56717802acf258530e2d6d237787d760a8fbcfd642790254d5ab
        - quay.io/stolostron/cluster-backup-controller@sha256:c85f88d42b42e4366f8ff7ad6c3e6b375e98c5b6a12a9c24a6333929e48f1811
        - quay.io/stolostron/observatorium-operator@sha256:5eb4c0da92c62b664aaf3bc187924870275e687440c3b0a3ad4f1c760c8b35e9
        - quay.io/stolostron/kube-rbac-proxy@sha256:5b521e4f0e5e54c53cea714020069aa796da23d1482d14e4b7c908098ac69ae9
        - quay.io/stolostron/observability-e2e-test@sha256:03403aeb59b364f578761ae4bcb302bd3589baff500d733775e8cb4f2f961236
        - quay.io/stolostron/insights-metrics@sha256:f2f4cf14090252af32832ff6df8a9f896e5314751199156c4e5ae0755e4515e6
        - quay.io/stolostron/search-collector@sha256:54ece9741641e1bb6e30f15d97a65c92b9c6408ba2572a2f6658f537134adff1
        - quay.io/stolostron/multicluster-operators-subscription@sha256:a056ce7d438f68d33ca3dfd42ec3d9a34b347cdf05499066412784109726e7ad
        - quay.io/stolostron/submariner-addon@sha256:9c72b5177de8ddc56dd3544997e9bd222895545b7c578e43e4051596fb1fa873
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/prometheus-alertmanager@sha256:b98480f8ba6ffd19fcee3b412757ad34428e6feab7e4196bb13f2f940306e2c9
        - quay.io/stolostron/insights-client@sha256:978d8b7c3e692272024b71b8c005b752c08dc643dde8258ad75961ddcdcce67f
        - quay.io/stolostron/prometheus@sha256:3d1810567f67ae1a7bb024a1c4e494ee510cfb14c7ed21a8b5e63e41fa7bcd68
        - quay.io/stolostron/flightctl-ocp-ui@sha256:6cff6425850d2b4a8b885a684afb01244d7cd9de7035d7ad71b1e10e5bab25e9
        - quay.io/stolostron/kube-state-metrics@sha256:cd24916e55ad905e386fa3df8a648272114353a0d75283224a4dddb1955b2687
        - quay.io/stolostron/flightctl-ui@sha256:fea717f3b8b4c343ec60ce7d74a691f1124b6847d55e30a73dc0d50ea610d370
        - quay.io/stolostron/multicluster-operators-channel@sha256:501a5053bd3fd5cc9de34e7e020890d731321b090a0e5bf11b8b3350d002f573
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/stolostron/multiclusterhub-operator@sha256:09f1f4266938fba88376a50c3e2837c5517b4f3a947cf86638a17e52eb17c1f6
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/prometheus-operator@sha256:f80e79ea87cde66955e1e46d264e414c072085da9db6876ee2ae68c87d2452e5
        - quay.io/stolostron/governance-policy-framework-addon@sha256:047124271d0b30176a94454b293c82c2eb67a6f06757fc22c54c455fd0913157
        - quay.io/stolostron/console@sha256:eed52f3ab860a3a224ce178d92765ca50d72d6261f04df040a39577881738252
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/multicloud-integrations@sha256:0081065f9e59c5ed41087344668ac86012f796cdbbf3575fd8491c7dfd24fbf8
        - quay.io/stolostron/must-gather@sha256:2bf9d3c15f32a6bfe0a7a8d39b9dc62ac6cbb6fe47b7cf1f6bc5e30b17e5e849
        - quay.io/stolostron/grafana-dashboard-loader@sha256:6c6c2aa528373536d5db66fffeb87a85b79f3cf723e5d9eeaefcb74f91c1a8a1
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:05da11b343938866c5afcfe844bcd15d9c513fdb10fd7ea0d3473b455b427c2a
        - quay.io/stolostron/rbac-query-proxy@sha256:096ad23c9b9c7e90451c09b48ded67a82575e75c8cc19f22c459d63ae8b5a786
        version: 0.6.3
      entries:
      - name: stolostron.v0.6.3
        version: 0.6.3
      - name: stolostron.v0.6.2
        version: 0.6.2
      - name: stolostron.v0.6.1
        version: 0.6.1
      - name: stolostron.v0.6.0
        version: 0.6.0
      name: community-0.6
    - currentCSV: stolostron.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:ab6660283a25bdf2ea9afe8c76bfbc32f4610d69af2bb13a7bca37a1f5961bdd
          createdAt: "2025-06-09T15:10:05Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.7.1'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MulticlusterHub defines the configuration
              for an instance of a multicluster hub, a central point for managing multiple
              Kubernetes-based clusters. The deployment of multicluster hub components
              is determined based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/rbac-query-proxy@sha256:1c198b6388a72ccfe2c5b68d57a2b22893fb65d977992ec44ee8cad6ee1ca976
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/stolostron/governance-policy-addon-controller@sha256:6ecc7557168177474423cea60eadc07d8382adb3e97dee669a8590859a41b08c
        - quay.io/stolostron/flightctl-api@sha256:64fb1b655bef6aeff03cf64b6450155d3cb570ed4713f084590fe515f4fbad02
        - quay.io/stolostron/cert-policy-controller@sha256:6deb41867f8781f4e34f69a54d4aeca190d588ba2991279d39cb5a82c2a5c5be
        - quay.io/stolostron/flightctl-ui@sha256:fea717f3b8b4c343ec60ce7d74a691f1124b6847d55e30a73dc0d50ea610d370
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/cluster-backup-controller@sha256:e576c2de1f931b7d33da2dbc03d71ce323766bd9158bf4b128221ddd1ab3fb9f
        - quay.io/stolostron/observatorium@sha256:18a06117ef114a1e68474fd9832adacd30b03f60ec1aa55374285a20e5311f13
        - quay.io/stolostron/insights-metrics@sha256:861dc03fff55c1d9aff96b5ed15b9d38d5932df1c00a4423e35ead19da6eda03
        - quay.io/stolostron/siteconfig-operator@sha256:16b00f6c4d81eea26e663ca5596526d85bf7777728d1f0ca7e4dd765c64d2bb5
        - quay.io/stolostron/search-indexer@sha256:58320b54847e3bb6baa558bda65e9d0ae2198df480ba27f46f96ba1454614508
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:4b29cc08c9e35054a4a820472ea19cd51e043ab1a90495418a36922df74b9540
        - quay.io/stolostron/multiclusterhub-operator@sha256:ab6660283a25bdf2ea9afe8c76bfbc32f4610d69af2bb13a7bca37a1f5961bdd
        - quay.io/stolostron/prometheus-alertmanager@sha256:a9da6f01cfa61fef803f09ef3f6b27e33a07b97fc1f1e11b6a6c8e13b8a4b90e
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/volsync-addon-controller@sha256:224a0a03a25e5ce61f4c50f6252370876ac0cbf4f9711e997e98a00d6a124fe9
        - quay.io/stolostron/origin-cli@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/search-v2-operator@sha256:2ae9ab48615966ccf91b2022ccda12913a3db3cfbad5351dd8a20648b01137d8
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/console@sha256:ba89196efe30c812a767d58a41170b56e2b04167a72e7f13180bf40474bdc824
        - quay.io/stolostron/prometheus-config-reloader@sha256:024ee6dda10c6d22ddeab57d7d6503b7aa9ac78981cfcec8ffae94b4914500e3
        - quay.io/stolostron/submariner-addon@sha256:51359ccdc0ea29bcaf7ec2b190878e9f3333b1530cb1128e142d893fb87eba8b
        - quay.io/stolostron/prometheus-operator@sha256:8c1c12562ec00da8eaedb7d428694ebdd3a749c0b1ce03fe2b5663ae40e1c701
        - quay.io/stolostron/flightctl-periodic@sha256:a3ed745dc65a5886c4cc05348cf4822bfbf2c6d28b85a39fa0b8f3e1f53fdf8a
        - quay.io/stolostron/governance-policy-framework-addon@sha256:b1b24c40d68415525e2bae21810fc439d3e96fcc5fa75c92e7f2eee284422bfd
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:05da11b343938866c5afcfe844bcd15d9c513fdb10fd7ea0d3473b455b427c2a
        - quay.io/stolostron/helloworld@sha256:4465550664bd210cf0df8edfd706de2fc51cacf01b66a962b5607935f9666ebc
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:54c6a2eec497eafdbf861fb95d78d961fb24906b197fc69b4e96d3e8b53e6496
        - quay.io/stolostron/multicluster-observability-operator@sha256:3d95297297a301fca1915268fec71d7a6ecef98c6c2ab8494d9befacb2e169fa
        - quay.io/stolostron/thanos@sha256:9d504e2aa1001135eb6e454c58c620d102f2b0a85016b7587452dd1595794c93
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/must-gather@sha256:ad8db72d559d06f4d467f78213d4f8f2cb00f8e76e942d8e75394eb7dadd29d2
        - quay.io/stolostron/acm-cli@sha256:2f0cec3bed4ad77f573e8cd51a7e2c62d68a59e4e261ea801d5ac203fd2724fc
        - quay.io/stolostron/grafana@sha256:cebaed7e5abe0064c65ea69bd4e14794d153de9172c4c3b6de4730f23fed7d24
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/metrics-collector@sha256:c7433ee4ccdb56717802acf258530e2d6d237787d760a8fbcfd642790254d5ab
        - quay.io/stolostron/observability-e2e-test@sha256:4e60d3e3e1b5dbab3bcc1c4d37a4940a2a42a065a9df539fb243fb174dde6716
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/search-v2-api@sha256:786866ca8c71a8263ddef45ca9bd9e48cfd1b3ad6215725fbf6f79967bda20ad
        - quay.io/stolostron/multicluster-operators-channel@sha256:318d885ab33d5afe17334e831c955f1c3cdf804281b7cc8523330820ac22eee2
        - quay.io/stolostron/multicloud-integrations@sha256:665e9385531674a5849aa661eb207a0c6e39ea563fc891c57678cddbf7878d8e
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/observatorium-operator@sha256:2f11444703addec163a00c99529c228594fb52df13de47f08faac2e446bec483
        - quay.io/stolostron/governance-policy-propagator@sha256:2b77422fb03177f647c04716e7242bc08f241a1ef9c8968e492ae57816f4c4fb
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/grafana-dashboard-loader@sha256:6c6c2aa528373536d5db66fffeb87a85b79f3cf723e5d9eeaefcb74f91c1a8a1
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/cluster-permission@sha256:e45139d459951b2e751018fe7ecabe1ff22b90ac87be81fb8f17a12f7b7f6a65
        - quay.io/stolostron/search-collector@sha256:2b3dfd21c0ba7b7b84fac390002e47fae927cd5fdf13e5d1d960703aa6c5bede
        - quay.io/stolostron/insights-client@sha256:2cbd20252f10a5a64f93af50f117f05a865aa9d578e5f6e55e4249470b761604
        - quay.io/stolostron/kube-rbac-proxy@sha256:e5555e256a518ca34bea01c959bbb5497582b53dbfaaf326e1852430c175637a
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:1efd8e470b88a992e2f35ce18bdf4970f91af7ba917b8d92627b47ce8c10eb40
        - quay.io/stolostron/flightctl-worker@sha256:4dddded7d2d4cce63972c8ca81f213c2cedb551cc645ec85495a9692dc36bb4c
        - quay.io/stolostron/config-policy-controller@sha256:b1718d5facb95c5572ac658c899f9fce0a7e8c08de93804de3327cd50cdb073b
        - quay.io/stolostron/redis-7-c9s@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/search-e2e@sha256:1c7dcc9f3588620b3570a493c2d744de059f425b572b9cb5725142b89414a904
        - quay.io/stolostron/flightctl-ocp-ui@sha256:6cff6425850d2b4a8b885a684afb01244d7cd9de7035d7ad71b1e10e5bab25e9
        - quay.io/stolostron/thanos-receive-controller@sha256:f445bbfc18bbc4646e0432842022d3be5305809e52ddcb9daf0ec000844d5854
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/klusterlet-addon-controller@sha256:4cb2f4f4e14cd67a4b2f9549cba714b635d9e1c2d8b869b0e0573fd7a41bc13e
        - quay.io/stolostron/multicluster-operators-application@sha256:9f804b4f2da910e0d7f122176f72c0b0ef4b30785fde0369a66f95f4bac58a82
        - quay.io/stolostron/prometheus@sha256:b959035ff754050ca092863bb74c35ba9622d9adc9e8eb7caf3b430cfe211730
        - quay.io/stolostron/grc-policy-framework-tests@sha256:327122b13ec79ea488d6f424806549e36ebfbabd055f9b40217ba541f4a118a9
        - quay.io/stolostron/multicluster-operators-subscription@sha256:d7c137379781ce6b697f73a464b89d04fbc20c3da00ce21a195ac79ab57a6dc7
        - quay.io/stolostron/kube-state-metrics@sha256:3503df61f3d9ebf3cbf66deef865450b32c7403839fdaca42504febaac00687c
        version: 0.7.1
      entries:
      - name: stolostron.v0.7.1
        version: 0.7.1
      - name: stolostron.v0.7.0
        version: 0.7.0
      name: community-0.7
    defaultChannel: community-0.7
    packageName: stolostron
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MinIO
      provider-url: https://min.io
    name: minio-object-store-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-object-store-operator.v2025.8.19175300
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aistor.min.io/v1",
                "kind": "ObjectStore",
                "metadata": {
                  "annotations": {
                    "prometheus.io/path": "/minio/v2/metrics/cluster",
                    "prometheus.io/port": "9000",
                    "prometheus.io/scrape": "true"
                  },
                  "labels": {
                    "app": "minio"
                  },
                  "name": "my-objectstore"
                },
                "spec": {
                  "cache": {
                    "enabled": false
                  },
                  "certificates": {
                    "disableAutoCert": false
                  },
                  "configuration": {
                    "name": "storage-configuration"
                  },
                  "firewall": {
                    "enabled": false
                  },
                  "mountPath": "/export",
                  "podManagementPolicy": "Parallel",
                  "pools": [
                    {
                      "name": "pool-0",
                      "securityContext": {
                        "fsGroupChangePolicy": "OnRootMismatch"
                      },
                      "servers": 4,
                      "volumeClaimTemplate": {
                        "apiVersion": "v1",
                        "kind": "persistentvolumeclaims",
                        "spec": {
                          "accessModes": [
                            "ReadWriteOnce"
                          ],
                          "resources": {
                            "requests": {
                              "storage": "1Ti"
                            }
                          },
                          "storageClassName": "standard"
                        }
                      },
                      "volumesPerServer": 4
                    }
                  ]
                }
              },
              {
                "apiVersion": "aistor.min.io/v1alpha1",
                "kind": "AdminJob",
                "metadata": {
                  "name": "minio-job"
                },
                "spec": {
                  "commands": [
                    {
                      "args": {
                        "name": "iac-bucket"
                      },
                      "op": "make-bucket"
                    }
                  ],
                  "containerSecurityContext": {},
                  "objectStore": {
                    "name": "my-objectstore",
                    "namespace": "my-objectstore"
                  },
                  "securityContext": {},
                  "serviceAccountName": "my-objectstore-sa"
                }
              },
              {
                "apiVersion": "sts.min.io/v1beta1",
                "kind": "PolicyBinding",
                "metadata": {
                  "name": "example-binding"
                },
                "spec": {
                  "application": {
                    "namespace": "my-objectstore",
                    "serviceaccount": "my-objectstore-sa"
                  },
                  "policies": [
                    "consoleAdmin"
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,AI/Machine Learning
          createdAt: "2025-08-27T14:50:56Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AdminJob
            name: adminjobs.aistor.min.io
            version: v1alpha1
          - kind: ObjectStore
            name: objectstores.aistor.min.io
            version: v1
          - kind: PolicyBinding
            name: policybindings.sts.min.io
            version: v1alpha1
          - kind: PolicyBinding
            name: policybindings.sts.min.io
            version: v1beta1
        description: "MinIO AIStor extends the most sophisticated object store on
          the market. It is designed for the exascale data \ninfrastructure challenges
          presented by modern AI workloads. AIStor provides new features, along with
          performance and\nscalability improvements, to enable enterprises to store
          all AI data in one infrastructure.\n\nMinIO is the standard for building
          <a href=\"https://min.io/solutions/object-storage-for-ai\">large scale AI
          data infrastructure</a>.\n\nThe MinIO AIStor is a software-defined, Amazon
          S3 compatible object store that will run anywhere - from the public cloud
          to the edge.<br><br>\n\nEnterprises use MinIO to deliver against artificial
          intelligence, machine learning, analytics, application, backup and archival
          workloads - all from a single platform.\n\nThe use cases include storage
          for HDFS replacements, AI datalakes, SIEM datalakes, analytics datalakes,
          object storage as a service, streaming and database warm/hot storage.<br><br>\n\nMinIO’s
          AIStor is the fastest object storage server, with production READ/WRITE
          speeds in excess of 2.2 TiB/s on commodity hardware.\n\nBuilt for the cloud
          operating model it is native to the technologies and architectures that
          define the cloud. These include containerization, orchestration with Kubernetes,
          microservices and multi-tenancy - making it the most widely integrated object
          store in the market today.<br><br>\n\nThe AIStor offers a host of enterprise
          features including inline erasure coding, bit-rot detection, state-of-the-art
          encryption, active-active replication, object locking, lifecycle management,
          observability, key-management, caching, firewall/load balancing, global
          console, search and identity + access management.<br><br>\n\nThe AIStor
          is priced on usable capacity and includes support through the MinIO Subscription
          Network.\n\nThe Subscription Network is designed to optimize deployments
          of MinIO and reduce the risk associated with data loss and data breach while
          providing 24/7/365 direct-to-engineering support.<br><br>\n\n<a href=\"https://resources.min.io/request-a-demo-marketplaces/\">Request
          a demo today</a>"
        displayName: Minio AIStor Object Store
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Storage
        - S3
        - MinIO
        - Object Storage
        links:
        - name: Minio AIStor
          url: https://min.io
        - name: Support
          url: https://subnet.min.io
        - name: Github
          url: https://github.com/minio/aistor
        maintainers:
        - email: dev@min.io
          name: MinIO
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: MinIO
          url: https://min.io
        relatedImages:
        - quay.io/minio/aistor/operator@sha256:f80139efa062fdd83f78318b20f371a827b68e854b03dd9706b9eba4c493aeb2
        version: 2025.8.19175300
      entries:
      - name: minio-object-store-operator.v2025.8.19175300
        version: 2025.8.19175300
      - name: minio-object-store-operator.v2025.8.1184927
        version: 2025.8.1184927
      - name: minio-object-store-operator.v2025.7.1011319
        version: 2025.7.1011319
      - name: minio-object-store-operator.v2025.6.3065135
        version: 2025.6.3065135
      - name: minio-object-store-operator.v2025.5.12190907
        version: 2025.5.12190907
      name: stable
    defaultChannel: stable
    packageName: minio-object-store-operator
    provider:
      name: MinIO
      url: https://min.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: https://stakater.com
    name: template-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: template-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-07-04T00:40:36Z"
          description: Operator to distribute plain kubernetes manifests, helm charts
            and, copy secrets and configmaps to multiple namespaces
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.0.1
          operatorframework.io/suggested-namespace: template-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterTemplateInstance is the Schema for the clustertemplateinstances
              API
            displayName: Cluster Template Instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateGroupInstance is the Schema for the TemplateGroupInstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.templates.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
        description: '"Operator to distribute plain kubernetes manifests, helm charts
          and, copy secrets and configmaps to multiple namespaces"'
        displayName: Template Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - templates
        - secrets
        - configmaps
        - distribution
        - stakater
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/template-operator@sha256:ad8312417e4e9621b74d3ad03020516840feef9c4b52c4f0d894dbb6dc6f906a
        version: 0.0.1
      entries:
      - name: template-operator.v0.0.1
        version: 0.0.1
      name: release-0.0
    - currentCSV: template-operator.v0.1.5
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-08-29T06:41:42Z"
          description: Operator to distribute plain kubernetes manifests, helm charts
            and, copy secrets and configmaps to multiple namespaces
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.1.5
          operatorframework.io/suggested-namespace: template-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterTemplateInstance is the Schema for the clustertemplateinstances
              API
            displayName: Cluster Template Instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateGroupInstance is the Schema for the TemplateGroupInstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.templates.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
        description: '"Operator to distribute plain kubernetes manifests, helm charts
          and, copy secrets and configmaps to multiple namespaces"'
        displayName: Template Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - templates
        - secrets
        - configmaps
        - distribution
        - stakater
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/template-operator@sha256:88f722baf22ac07704503f098f2b0ea7d76815a1ab0552474b1c9ba6639d8142
        version: 0.1.5
      entries:
      - name: template-operator.v0.1.5
        version: 0.1.5
      name: release-0.1
    defaultChannel: release-0.1
    packageName: template-operator
    provider:
      name: Stakater
      url: https://stakater.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: loki-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-operator.v5.9.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:6728046841367066216b304ed06249b80c57c74151c6a75c9df04a6f7f76e03d
          createdAt: "2025-01-30T17:34:53Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.7.0-0 <5.9.15'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:0c3943fa009f0c649f6d592b0ac410653c8c44b4a90e41fabdfa5f3f7db1e43d
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:6728046841367066216b304ed06249b80c57c74151c6a75c9df04a6f7f76e03d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:2e164e377d4fba98e7f63f43f25b6b9dad441d2e2914c2bf02f3525d5c9aa483
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:542b28fd51194b46efc4e5af10c3241b52db4c71f0d5d1b1f184eb39764e6c53
        version: 5.9.15
      entries:
      - name: loki-operator.v5.9.15
        version: 5.9.15
      - name: loki-operator.v5.9.14
        version: 5.9.14
      - name: loki-operator.v5.9.13
        version: 5.9.13
      - name: loki-operator.v5.9.12
        version: 5.9.12
      - name: loki-operator.v5.9.11
        version: 5.9.11
      - name: loki-operator.v5.9.10
        version: 5.9.10
      - name: loki-operator.v5.9.9
        version: 5.9.9
      - name: loki-operator.v5.9.8
        version: 5.9.8
      - name: loki-operator.v5.9.7
        version: 5.9.7
      - name: loki-operator.v5.9.6
        version: 5.9.6
      - name: loki-operator.v5.9.5
        version: 5.9.5
      - name: loki-operator.v5.9.4
        version: 5.9.4
      - name: loki-operator.v5.9.3
        version: 5.9.3
      - name: loki-operator.v5.9.2
        version: 5.9.2
      - name: loki-operator.v5.9.1
        version: 5.9.1
      - name: loki-operator.v5.9.0
        version: 5.9.0
      name: stable
    - currentCSV: loki-operator.v5.9.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:6728046841367066216b304ed06249b80c57c74151c6a75c9df04a6f7f76e03d
          createdAt: "2025-01-30T17:34:53Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.7.0-0 <5.9.15'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:2e164e377d4fba98e7f63f43f25b6b9dad441d2e2914c2bf02f3525d5c9aa483
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:542b28fd51194b46efc4e5af10c3241b52db4c71f0d5d1b1f184eb39764e6c53
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:0c3943fa009f0c649f6d592b0ac410653c8c44b4a90e41fabdfa5f3f7db1e43d
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:6728046841367066216b304ed06249b80c57c74151c6a75c9df04a6f7f76e03d
        version: 5.9.15
      entries:
      - name: loki-operator.v5.9.15
        version: 5.9.15
      - name: loki-operator.v5.9.14
        version: 5.9.14
      - name: loki-operator.v5.9.13
        version: 5.9.13
      - name: loki-operator.v5.9.12
        version: 5.9.12
      - name: loki-operator.v5.9.11
        version: 5.9.11
      - name: loki-operator.v5.9.10
        version: 5.9.10
      - name: loki-operator.v5.9.9
        version: 5.9.9
      - name: loki-operator.v5.9.8
        version: 5.9.8
      - name: loki-operator.v5.9.7
        version: 5.9.7
      - name: loki-operator.v5.9.6
        version: 5.9.6
      - name: loki-operator.v5.9.5
        version: 5.9.5
      - name: loki-operator.v5.9.4
        version: 5.9.4
      - name: loki-operator.v5.9.3
        version: 5.9.3
      - name: loki-operator.v5.9.2
        version: 5.9.2
      - name: loki-operator.v5.9.1
        version: 5.9.1
      - name: loki-operator.v5.9.0
        version: 5.9.0
      name: stable-5.9
    - currentCSV: loki-operator.v6.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.14-v4.17
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:1cb1d2cad8632791529eb754938612f5e5ed373284e42f309d603059ab7ec7c5
          createdAt: "2025-03-12T12:05:37Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=5.8.0-0 <6.0.10'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.0
          operators.operatorframework.io.bundle.channels.v1: stable-6.0
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:9aa8d8e4a7fd7ddad5d842a22b09aecdbdfa31ca5eb8c4b17f0818636af5fbf8
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:69e69ab2628742774d12dd9dde50607625efdff50ea4671f2859692fbee5ecb4
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:98d2cf3437faa2400c01edb3c694816e0214c8f9dca0a0b0cdf5882a720248f3
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:1cb1d2cad8632791529eb754938612f5e5ed373284e42f309d603059ab7ec7c5
        version: 6.0.10
      entries:
      - name: loki-operator.v6.0.10
        version: 6.0.10
      - name: loki-operator.v6.0.9
        version: 6.0.9
      - name: loki-operator.v6.0.8
        version: 6.0.8
      - name: loki-operator.v6.0.7
        version: 6.0.7
      - name: loki-operator.v6.0.6
        version: 6.0.6
      - name: loki-operator.v6.0.5
        version: 6.0.5
      - name: loki-operator.v6.0.4
        version: 6.0.4
      - name: loki-operator.v6.0.3
        version: 6.0.3
      - name: loki-operator.v6.0.2
        version: 6.0.2
      - name: loki-operator.v6.0.1
        version: 6.0.1
      - name: loki-operator.v6.0.0
        version: 6.0.0
      name: stable-6.0
    - currentCSV: loki-operator.v6.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.15-v4.18
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:0bde3c1a0b667a693f443688ed3926eef1705c45ecf4cce3a3dedbddca7a2784
          createdAt: "2025-03-12T11:19:10Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=5.9.0-0 <6.1.8'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.1
          operators.operatorframework.io.bundle.channels.v1: stable-6.1
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:0bde3c1a0b667a693f443688ed3926eef1705c45ecf4cce3a3dedbddca7a2784
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:731f92c146f299bc3662b309a3068bde18fad51f65a7d29063e582b61e07042f
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:c34676932ddd14db213dbf42a1d341f51692f3f47f1b751d06187e292c99a170
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b94833f55644c289fe2c1297b68a4f9931baf977587c6a2c4c7a3e484c704f64
        version: 6.1.8
      entries:
      - name: loki-operator.v6.1.8
        version: 6.1.8
      - name: loki-operator.v6.1.7
        version: 6.1.7
      - name: loki-operator.v6.1.6
        version: 6.1.6
      - name: loki-operator.v6.1.5
        version: 6.1.5
      - name: loki-operator.v6.1.4
        version: 6.1.4
      - name: loki-operator.v6.1.3
        version: 6.1.3
      - name: loki-operator.v6.1.2
        version: 6.1.2
      - name: loki-operator.v6.1.1
        version: 6.1.1
      - name: loki-operator.v6.1.0
        version: 6.1.0
      name: stable-6.1
    - currentCSV: loki-operator.v6.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.16-v4.19
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f401be018c5ba35e8ea1e68831cc20067f0f4d1be031d05b222548c1bdfb5f7e
          createdAt: "2025-02-18T18:01:39Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=6.0.0-0 <6.2.5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.2
          operators.operatorframework.io.bundle.channels.v1: stable-6.2
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f401be018c5ba35e8ea1e68831cc20067f0f4d1be031d05b222548c1bdfb5f7e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:99d1e7c5c1f0609b33de576884782550f969b28dc622810e6661d4c60a664215
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:8955d8186ae98095623f2083b6d5389d3ac9f2612fcfcf695835219c3e1216c4
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b8c5b46a94a6c85c9d382eeeb65a1d2b4014a7dc765e9abfba706c92857462dc
        version: 6.2.5
      entries:
      - name: loki-operator.v6.2.5
        version: 6.2.5
      - name: loki-operator.v6.2.4
        version: 6.2.4
      - name: loki-operator.v6.2.3
        version: 6.2.3
      - name: loki-operator.v6.2.2
        version: 6.2.2
      - name: loki-operator.v6.2.1
        version: 6.2.1
      - name: loki-operator.v6.2.0
        version: 6.2.0
      name: stable-6.2
    defaultChannel: stable-6.2
    packageName: loki-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudtrail-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudtrail-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudtrail.services.k8s.aws/v1alpha1",
                "kind": "EventDataStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudtrail.services.k8s.aws/v1alpha1",
                "kind": "Trail",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudtrail-controller:1.1.1
          createdAt: "2025-09-19T21:16:25Z"
          description: AWS CloudTrail controller is a service controller for managing
            CloudTrail resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventDataStore represents the state of an AWS cloudtrail
              EventDataStore resource.
            displayName: EventDataStore
            kind: EventDataStore
            name: eventdatastores.cloudtrail.services.k8s.aws
            version: v1alpha1
          - description: Trail represents the state of an AWS cloudtrail Trail resource.
            displayName: Trail
            kind: Trail
            name: trails.cloudtrail.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Compute Cloud (CloudTrail) resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudTrail**

          AWS CloudTrail is an AWS service that helps you enable operational and risk auditing, governance, and compliance of your AWS account. Actions taken by a user, role, or an AWS service are recorded as events in CloudTrail. Events include actions taken in the AWS Management Console, AWS Command Line Interface, and AWS SDKs and APIs.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudTrail
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudtrail
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudTrail Developer Resources
          url: https://aws.amazon.com/cloudtrail/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudtrail maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudtrail-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-cloudtrail-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudtrail-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudtrail-controller.v1.0.28
        version: 1.0.28
      - name: ack-cloudtrail-controller.v1.0.26
        version: 1.0.26
      - name: ack-cloudtrail-controller.v1.0.25
        version: 1.0.25
      - name: ack-cloudtrail-controller.v1.0.24
        version: 1.0.24
      - name: ack-cloudtrail-controller.v1.0.23
        version: 1.0.23
      - name: ack-cloudtrail-controller.v1.0.22
        version: 1.0.22
      - name: ack-cloudtrail-controller.v1.0.21
        version: 1.0.21
      - name: ack-cloudtrail-controller.v1.0.20
        version: 1.0.20
      - name: ack-cloudtrail-controller.v1.0.19
        version: 1.0.19
      - name: ack-cloudtrail-controller.v1.0.16
        version: 1.0.16
      - name: ack-cloudtrail-controller.v1.0.15
        version: 1.0.15
      - name: ack-cloudtrail-controller.v1.0.14
        version: 1.0.14
      - name: ack-cloudtrail-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudtrail-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudtrail-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudtrail-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudtrail-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudtrail-controller.v1.0.5
        version: 1.0.5
      - name: ack-cloudtrail-controller.v1.0.4
        version: 1.0.4
      - name: ack-cloudtrail-controller.v1.0.3
        version: 1.0.3
      - name: ack-cloudtrail-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudtrail-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudtrail-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudtrail-controller.v0.1.3
        version: 0.1.3
      - name: ack-cloudtrail-controller.v0.1.2
        version: 0.1.2
      - name: ack-cloudtrail-controller.v0.1.1
        version: 0.1.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudtrail-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-route53-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-route53-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/route53-controller:1.1.1
          createdAt: "2025-09-19T19:15:31Z"
          description: AWS Route53 controller is a service controller for managing
            Route53 resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HealthCheck represents the state of an AWS route53 HealthCheck
              resource.
            displayName: HealthCheck
            kind: HealthCheck
            name: healthchecks.route53.services.k8s.aws
            version: v1alpha1
          - description: HostedZone represents the state of an AWS route53 HostedZone
              resource.
            displayName: HostedZone
            kind: HostedZone
            name: hostedzones.route53.services.k8s.aws
            version: v1alpha1
          - description: RecordSet represents the state of an AWS route53 RecordSet
              resource.
            displayName: RecordSet
            kind: RecordSet
            name: recordsets.route53.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Route53 resources in AWS from within your Kubernetes cluster.

          **About Amazon Route53**

          Amazon Route 53 is a highly available and scalable Domain Name System (DNS) web service. You can use Route 53 to perform three main functions in any combination: domain registration, DNS routing, and health checking.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Route53
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - route53
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Route53 Developer Resources
          url: https://docs.aws.amazon.com/route53/index.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/route53-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-route53-controller.v1.1.1
        version: 1.1.1
      - name: ack-route53-controller.v1.1.0
        version: 1.1.0
      - name: ack-route53-controller.v1.0.6
        version: 1.0.6
      - name: ack-route53-controller.v1.0.5
        version: 1.0.5
      - name: ack-route53-controller.v1.0.3
        version: 1.0.3
      - name: ack-route53-controller.v1.0.2
        version: 1.0.2
      - name: ack-route53-controller.v0.1.0
        version: 0.1.0
      - name: ack-route53-controller.v0.0.28
        version: 0.0.28
      - name: ack-route53-controller.v0.0.27
        version: 0.0.27
      - name: ack-route53-controller.v0.0.26
        version: 0.0.26
      - name: ack-route53-controller.v0.0.25
        version: 0.0.25
      - name: ack-route53-controller.v0.0.24
        version: 0.0.24
      - name: ack-route53-controller.v0.0.23
        version: 0.0.23
      - name: ack-route53-controller.v0.0.22
        version: 0.0.22
      - name: ack-route53-controller.v0.0.20
        version: 0.0.20
      - name: ack-route53-controller.v0.0.19
        version: 0.0.19
      - name: ack-route53-controller.v0.0.18
        version: 0.0.18
      - name: ack-route53-controller.v0.0.17
        version: 0.0.17
      - name: ack-route53-controller.v0.0.16
        version: 0.0.16
      - name: ack-route53-controller.v0.0.15
        version: 0.0.15
      - name: ack-route53-controller.v0.0.14
        version: 0.0.14
      - name: ack-route53-controller.v0.0.13
        version: 0.0.13
      - name: ack-route53-controller.v0.0.9
        version: 0.0.9
      - name: ack-route53-controller.v0.0.6
        version: 0.0.6
      - name: ack-route53-controller.v0.0.4
        version: 0.0.4
      - name: ack-route53-controller.v0.0.3
        version: 0.0.3
      - name: ack-route53-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-route53-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://www.medik8s.io/
    name: self-node-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: self-node-remediation.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/self-node-remediation-operator:v0.9.0
          createdAt: "2024-07-25 14:40:52"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.9.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Medik8s
          url: https://www.medik8s.io/
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/medik8s/self-node-remediation-operator:v0.9.0
        version: 0.9.0
      entries:
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      - name: self-node-remediation.v0.7.0
        version: 0.7.0
      - name: self-node-remediation.v0.6.1
        version: 0.6.1
      - name: self-node-remediation.v0.6.0
        version: 0.6.0
      - name: self-node-remediation.v0.5.0
        version: 0.5.0
      - name: self-node-remediation.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: self-node-remediation
    provider:
      name: Medik8s
      url: https://www.medik8s.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudfront-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudfront-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "CachePolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "Distribution",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudfront-controller:1.2.1
          createdAt: "2025-09-19T21:12:14Z"
          description: AWS CloudFront controller is a service controller for managing
            CloudFront resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CachePolicy represents the state of an AWS cloudfront CachePolicy
              resource.
            displayName: CachePolicy
            kind: CachePolicy
            name: cachepolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: Distribution represents the state of an AWS cloudfront Distribution
              resource.
            displayName: Distribution
            kind: Distribution
            name: distributions.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: Function represents the state of an AWS cloudfront Function
              resource.
            displayName: Function
            kind: Function
            name: functions.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: OriginAccessControl represents the state of an AWS cloudfront
              OriginAccessControl resource.
            displayName: OriginAccessControl
            kind: OriginAccessControl
            name: originaccesscontrols.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: OriginRequestPolicy represents the state of an AWS cloudfront
              OriginRequestPolicy resource.
            displayName: OriginRequestPolicy
            kind: OriginRequestPolicy
            name: originrequestpolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: ResponseHeadersPolicy represents the state of an AWS cloudfront
              ResponseHeadersPolicy resource.
            displayName: ResponseHeadersPolicy
            kind: ResponseHeadersPolicy
            name: responseheaderspolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: VPCOrigin represents the state of an AWS cloudfront VPCOrigin
              resource.
            displayName: VPCOrigin
            kind: VPCOrigin
            name: vpcorigins.cloudfront.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudFront resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudFront**

          Amazon CloudFront is a web service that speeds up distribution of your static and dynamic web content, such as .html, .css, .js, and image files, to your users. CloudFront delivers your content through a worldwide network of data centers called edge locations.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon CloudFront
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudfront
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudFront Developer Resources
          url: https://aws.amazon.com/cloudfront/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudfront maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudfront-controller:1.2.1
        version: 1.2.1
      entries:
      - name: ack-cloudfront-controller.v1.2.1
        version: 1.2.1
      - name: ack-cloudfront-controller.v1.2.0
        version: 1.2.0
      - name: ack-cloudfront-controller.v1.1.6
        version: 1.1.6
      - name: ack-cloudfront-controller.v1.1.4
        version: 1.1.4
      - name: ack-cloudfront-controller.v1.1.3
        version: 1.1.3
      - name: ack-cloudfront-controller.v1.1.2
        version: 1.1.2
      - name: ack-cloudfront-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudfront-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudfront-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudfront-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudfront-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudfront-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudfront-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudfront-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudfront-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudfront-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudfront-controller.v0.0.15
        version: 0.0.15
      - name: ack-cloudfront-controller.v0.0.14
        version: 0.0.14
      - name: ack-cloudfront-controller.v0.0.13
        version: 0.0.13
      - name: ack-cloudfront-controller.v0.0.12
        version: 0.0.12
      - name: ack-cloudfront-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudfront-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudfront-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudfront-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudfront-controller.v0.0.6
        version: 0.0.6
      - name: ack-cloudfront-controller.v0.0.5
        version: 0.0.5
      - name: ack-cloudfront-controller.v0.0.4
        version: 0.0.4
      - name: ack-cloudfront-controller.v0.0.3
        version: 0.0.3
      - name: ack-cloudfront-controller.v0.0.2
        version: 0.0.2
      - name: ack-cloudfront-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudfront-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/openshift/lightspeed-service
    name: lightspeed-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lightspeed-operator.v1.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ols.openshift.io/v1alpha1",
                "kind": "OLSConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightspeed-operator",
                    "app.kubernetes.io/instance": "olsconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "olsconfig",
                    "app.kubernetes.io/part-of": "lightspeed-operator"
                  },
                  "name": "cluster"
                },
                "spec": {
                  "llm": {
                    "providers": [
                      {
                        "credentialsSecretRef": {
                          "name": "credentials"
                        },
                        "models": [
                          {
                            "name": "gpt-3.5-turbo-1106"
                          }
                        ],
                        "name": "OpenAI"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-08-28T16:37:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lightspeed
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Kubernetes Engine", "OpenShift Virtualization
            Engine"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lightspeed-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Red Hat OpenShift Lightspeed instance. OLSConfig is the Schema
              for the olsconfigs API
            displayName: OLSConfig
            kind: OLSConfig
            name: olsconfigs.ols.openshift.io
            version: v1alpha1
        description: |-
          OpenShift Lightspeed Operator provides generative AI-based virtual assistant which integrates into the OpenShift web console. OpenShift Lightspeed can answer natural language questions related to OpenShift Container Platform.

          To configure the OpenShift Lightspeed Operator, you need a Large Language Model (LLM) . You can host the LLM using RHELAI or RHOAI or use one of the models from the following Software-as-a-Service providers:

          - OpenAI
          - Microsoft Azure OpenAI
          - IBM WatsonX

          For more information, see [About OpenShift Lightspeed](https://docs.redhat.com/en/documentation/red_hat_openshift_lightspeed/1.0tp1/html/about/ols-about-openshift-lightspeed) in the official product documentation.

          Questions, comments, concerns, or just want to talk to the OpenShift Lightspeed team? [Email us](mailto:openshift-lightspeed-contact-requests@redhat.com)!
        displayName: OpenShift Lightspeed Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ai
        - assistant
        - openshift
        - llm
        links:
        - name: Lightspeed Operator
          url: https://github.com/openshift/lightspeed-operator
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/lightspeed-service
        relatedImages:
        - registry.redhat.io/openshift-lightspeed/lightspeed-rhel9-operator@sha256:b12803642df4d9b3ad0cf800315021483f543c6a8827b3901e57bdb65f10dc48
        - registry.redhat.io/openshift-lightspeed/openshift-mcp-server-rhel9@sha256:f1e4521ff2cc076f361c04d8c9aeac09ffcedb89a2f1dbe71f85caa27472adab
        - registry.redhat.io/openshift-lightspeed/lightspeed-service-api-rhel9@sha256:ec359967a87bc7ee2c1d6a61e553acbdb0dc85d6f240a1865bf57d3ba5dc7193
        - registry.redhat.io/openshift-lightspeed/lightspeed-console-plugin-rhel9@sha256:dfc42c2651ac8793f61ff0e54334ecb006dc4a77487bc7c6878cdf47bd4ac080
        version: 1.0.5
      entries:
      - name: lightspeed-operator.v1.0.5
        version: 1.0.5
      name: alpha
    - currentCSV: lightspeed-operator.v1.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ols.openshift.io/v1alpha1",
                "kind": "OLSConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightspeed-operator",
                    "app.kubernetes.io/instance": "olsconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "olsconfig",
                    "app.kubernetes.io/part-of": "lightspeed-operator"
                  },
                  "name": "cluster"
                },
                "spec": {
                  "llm": {
                    "providers": [
                      {
                        "credentialsSecretRef": {
                          "name": "credentials"
                        },
                        "models": [
                          {
                            "name": "gpt-3.5-turbo-1106"
                          }
                        ],
                        "name": "OpenAI"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-08-28T16:37:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lightspeed
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Kubernetes Engine", "OpenShift Virtualization
            Engine"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lightspeed-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Red Hat OpenShift Lightspeed instance. OLSConfig is the Schema
              for the olsconfigs API
            displayName: OLSConfig
            kind: OLSConfig
            name: olsconfigs.ols.openshift.io
            version: v1alpha1
        description: |-
          OpenShift Lightspeed Operator provides generative AI-based virtual assistant which integrates into the OpenShift web console. OpenShift Lightspeed can answer natural language questions related to OpenShift Container Platform.

          To configure the OpenShift Lightspeed Operator, you need a Large Language Model (LLM) . You can host the LLM using RHELAI or RHOAI or use one of the models from the following Software-as-a-Service providers:

          - OpenAI
          - Microsoft Azure OpenAI
          - IBM WatsonX

          For more information, see [About OpenShift Lightspeed](https://docs.redhat.com/en/documentation/red_hat_openshift_lightspeed/1.0tp1/html/about/ols-about-openshift-lightspeed) in the official product documentation.

          Questions, comments, concerns, or just want to talk to the OpenShift Lightspeed team? [Email us](mailto:openshift-lightspeed-contact-requests@redhat.com)!
        displayName: OpenShift Lightspeed Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ai
        - assistant
        - openshift
        - llm
        links:
        - name: Lightspeed Operator
          url: https://github.com/openshift/lightspeed-operator
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/lightspeed-service
        relatedImages:
        - registry.redhat.io/openshift-lightspeed/lightspeed-service-api-rhel9@sha256:ec359967a87bc7ee2c1d6a61e553acbdb0dc85d6f240a1865bf57d3ba5dc7193
        - registry.redhat.io/openshift-lightspeed/lightspeed-console-plugin-rhel9@sha256:dfc42c2651ac8793f61ff0e54334ecb006dc4a77487bc7c6878cdf47bd4ac080
        - registry.redhat.io/openshift-lightspeed/openshift-mcp-server-rhel9@sha256:f1e4521ff2cc076f361c04d8c9aeac09ffcedb89a2f1dbe71f85caa27472adab
        - registry.redhat.io/openshift-lightspeed/lightspeed-rhel9-operator@sha256:b12803642df4d9b3ad0cf800315021483f543c6a8827b3901e57bdb65f10dc48
        version: 1.0.5
      entries:
      - name: lightspeed-operator.v1.0.5
        version: 1.0.5
      name: stable
    defaultChannel: stable
    packageName: lightspeed-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/openshift/lightspeed-service
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kubevirt-hyperconverged
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubevirt-hyperconverged-operator.v4.16.23
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:eaacb0d177fd64e2eefbcb2dc47633dadf4d54fcbe0f0f750b078ab3f4f189e1
          createdAt: "2025-09-29 21:23:00"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.14.1 <4.16.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:3e1c68d5695bab96c382661a33581e60597b59afd21660ec760d8fe5e20ff9e3
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.16.23-1,hyperconverged-cluster-operator-test-rhel9:v4.16.23-1,kubevirt-common-instancetypes-test-rhel9:v4.16.23-1,kubevirt-tekton-tasks-test-rhel9:v4.16.23-1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:e24191be577575fa9d462c87c6f51775a0cf0f5d1812d12c4c95a68ec34b34ff
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:6e4aa75bece3c2fe076e282b3f1ad8296299ec0551e379c7c88830a0ce8ed5bf
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:1be66eee321bc47dcc0e7ff4674a485b9a4552243080d796d5f3b0d417721dc4
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:28c0bdf60404fb435cfa0cfdc6d314f789417c44c3206ce8cd1e7b99031fb529
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:659abe1d740607b2cb3850c9ab29101fa5d5c821d2a09aab5994849697a504f3
        - registry.redhat.io/container-native-virtualization/mtq-operator-rhel9@sha256:9de5ded35edfbe98121bdd4404c84dcc72c8cba1742dc27df6382ef882db2183
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:0f0478e9357f8b6ace6d844e6e424db192398195a8cea2fa6398be8996d05be6
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:2843d199dc2748749d5f48fb9f90656962e011533d79dc7e0a4f3592b12aa69a
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:6b0c7cc17abd95fce45b6ec83ec291e843eedbc90dce0ecc449aec6e64562057
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:97d37927164f0ea986f209912e6247fdab37b680c214afc4bdbb7051caaf58b1
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:cc4ff488aa97e179c4df39ac5ae8085070017106adf1c5f2222b35d8f2234ea8
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:f8e2ea3988d4a77a4e809e1fd073ffa5f99393ca3eb69a2c1cb54dd4ab60e482
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:885670f2e9ad43224cbc96f34b23dd2642de00f688bade0f6b89dd3d3fb83dbd
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:251d6fb3197a7642d86193191aa69c85a45c5521d7fc509704edd46d5d4e90a7
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:fcbadecf01b81625d55f9219428baabc33ee7231b4ccf801c90dee3095907187
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:a0af0e5d4b142f28dc72c5636df5040c4ce92c6b12ab183d3a9bc7a9d2d2f418
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:c033ddd0ed30c0d25f4c2e9b30fca4e40959bb6d74029a1f452ff048b97de88a
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:7206ac7959c07f474f454ab48bcf3522ade95efb27fd9fbd3938d1387a1420af
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:468f117a1b55af59e72a9ff46499a8d0bb00884d450d57b692e2f7b557c9bf3d
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:e01871006b9e4cb1b6f477319de63b758eac672a70a298caed2032ec19986680
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:7587832b877f32621eb26e67a5c516728f293ecd98f74dd70d5f03474a5429d6
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:68b650d6779fc4d61f05d97e6d62b80fcd638721994153205eec68f8c1d999ce
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:19ca70dceb79b10ff458f0ea41faa6e5552a5f08828620ad7177efa34166a90e
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:84dbfa51755d60e34b194c1acf679ac083feb49038112212828fe8101dcec2a5
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:8462188ec7533d2105c0dc9154d8dcf1425b96d6945eef16ea774a5dff9bba1d
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:11030a8d8cd03efdf27fe3cfc9ebeadeb20851b6de1e9d8a93ba7f2bc6a77a96
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:8cf7993326ba220f650ae2fa5383d0d6cf3e2c7f3346f0a339a9d212b8df02f6
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:c18b6e315f5577cc18d46cd76595d48d4a49026a6b2f5763c3efb26919bde5eb
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:52b4f7c7a4f5c5c5d07a720c5eefa70709f520b1d939aec4cd971d8f3c9b12bc
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:ef8260d72239ad2126ca023250bfd976f5c050fe9c0727a0f7ef007446f5363b
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:18be4ef8b6c2cedab4a2e031939fc1f9cc23020bde7fccb7997e4222b673f57f
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:def80efdcc58d545aaf4b69b784a348cb36e97afd8e0124480521a2b980fa7c3
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:b031a12e73cb75e207203fd3785619e3a5a22d7d5a8549488c035f70374235be
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:14ea56d4149b72b98b6e7f5b33ba5bdd9115e7c643fd103c9df7a26b2a06700a
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:9e358366fd11dd167255abd930016aa4fc35b1b9dc69ce27316635488efc8ec4
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:6bf45534310768308958428692b29afa25f2740d0c62165d7182b0b3e9e944cf
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:eaacb0d177fd64e2eefbcb2dc47633dadf4d54fcbe0f0f750b078ab3f4f189e1
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:91b86228495f38d5528be1a8ffec96695ad72fafb76d2ead87f7bee25540d8b6
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:19b56bbb081db7198d570c3c691e142991aaa81857ee417d47903291ed0e7a1c
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:372652529b6a8705df581e7ab8acc2ec5c3cdf8fb273448052c59963abe4d984
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:7f89f2ba59309b26cf83f58ff3ba35569182097258fa543d5d88f3a8cd64ca00
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:bf1bb2fe09e028506fe7cf383368d76e358b8c59ecc74d3618bc250c1e97271c
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:a79579b18a7df9f23c584303593a9404c64df023458d55c715aeb71dcacc92a0
        - registry.redhat.io/container-native-virtualization/mtq-lock-server-rhel9@sha256:bb1dfc204e48a13985d6c2ec9cdf17f47330617725ee6dd854f0dae0f1901ba4
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:4005d165f1485eec98ca24ce41bef79cd8088002d43ffb709ec65f842a176bdc
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:3f3554ef043c80494db536b28eb03fa7043e53bb85ce426820c252fb999ce04e
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:531702869a34f76850d332efb84ccf571fafa281014afd1787497ca9de8aed37
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:a081e7e68def13b7de369d3f66469035d61fd1feef1bee4f95ba5f7b4041d960
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:e7672121156e050f05af771f75cfe26558699da8a074bf7051f710705a27d743
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:bd532e59edae5d2edac78b9249cd6a9f0779f44e25b019d824a1b9803f078bde
        - registry.redhat.io/container-native-virtualization/mtq-controller-rhel9@sha256:a774f0bd125c66b45e35e327c9c7e0afd7947d3c5ed222b3d5d223941366ba2c
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:52e76b4208ee4e7da8a0729b9bdb294729454d0f62580d64dd104b6921f464ed
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:3b57db7aaa3c61335044ec78cbd061f78ce99cdb4b75b3ce342e1a603da4bcef
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:7e19b42a648cee71c60cb69fbcf6ebca05c938942e16d23f1cd61a619cffdcb6
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:4046e8e862e0c55f09e9031b95e6a77af4d491593e39770c644287d44888b0bb
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:517c88a6be8842d4eb71ad30f99ce1afca954af1df2c1bf4a9584f4ec69fcf12
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:3e1c68d5695bab96c382661a33581e60597b59afd21660ec760d8fe5e20ff9e3
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:cca121312777d2727fe2c50a14583243ce69e6e356b499f133736787522095b2
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:ad24eb52bd65495224661d4894336d88fde342aeaab16b1084ceff6337f92059
        version: 4.16.23
      entries:
      - name: kubevirt-hyperconverged-operator.v4.16.23
        version: 4.16.23
      - name: kubevirt-hyperconverged-operator.v4.16.22
        version: 4.16.22
      - name: kubevirt-hyperconverged-operator.v4.16.21
        version: 4.16.21
      - name: kubevirt-hyperconverged-operator.v4.16.20
        version: 4.16.20
      - name: kubevirt-hyperconverged-operator.v4.16.19
        version: 4.16.19
      - name: kubevirt-hyperconverged-operator.v4.16.18
        version: 4.16.18
      - name: kubevirt-hyperconverged-operator.v4.16.17
        version: 4.16.17
      - name: kubevirt-hyperconverged-operator.v4.16.16
        version: 4.16.16
      - name: kubevirt-hyperconverged-operator.v4.16.15
        version: 4.16.15
      - name: kubevirt-hyperconverged-operator.v4.16.14
        version: 4.16.14
      - name: kubevirt-hyperconverged-operator.v4.16.13
        version: 4.16.13
      - name: kubevirt-hyperconverged-operator.v4.16.12
        version: 4.16.12
      - name: kubevirt-hyperconverged-operator.v4.16.11
        version: 4.16.11
      - name: kubevirt-hyperconverged-operator.v4.16.10
        version: 4.16.10
      - name: kubevirt-hyperconverged-operator.v4.16.9
        version: 4.16.9
      - name: kubevirt-hyperconverged-operator.v4.16.8
        version: 4.16.8
      name: candidate
    - currentCSV: kubevirt-hyperconverged-operator.v4.99.0-0.1723448771
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:5b2c64218342a5744182fc06f550bcdcf954e931275a9d1505ef0bbfbd0f8044
          createdAt: "2024-08-12 07:46:45"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.14.1 <4.16.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:d74c629e415d177069a0570fb8db9252d2b193d2ec7972c844686ffd14033f62
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.99.0-73,kubevirt-common-instancetypes-test-rhel9:v4.99.0-32,kubevirt-tekton-tasks-test-rhel9:v4.99.0-31
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:cc190d7c44b20f10ae61709fd2f5f7c0b248df73b768eb696d5a6cec90a34e59
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:41ee360ac43ee69ccee62d4ac3d2f07c8a8589e5b85c659b47018dd36c2b932f
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:b410508ecc71b0a3bf4825793efde4b07085e2243d8b3585da852fd1940c4be1
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:820a9e6bd0329df5c541ef0b67b3c7cdd9f81be33de87c5b5e5e43777bea197e
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:811e8e7e9872408794ef36229c42c6c437b1a4ace76c7134d91bb2331fe7815b
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:0769512df0fa1325d76a5f10a1a704cf60fc65f62f57b157ea53bc9ba682851f
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:8c64b526fd42a17c909bf45eb795121a81d502424a9cf11f8087118751333894
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:5f23fc8e9bc386622460c23b9644f58648d9f748f5670927a02888cd36c99a21
        - registry.redhat.io/container-native-virtualization/mtq-controller-rhel9@sha256:ad5938ebd5f07196ce5c35232a1c3b0a653301a83929ce84c3d1cb4dd1b91816
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:37e6b4fa5ff436572412ec0bdb1edc4a49e62c994e056c93578605f6b74adbcd
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:07435335fdbf6178bc00cc02b93cc0adb42784339151b2d9c31dc8da98b8bc6a
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:cffec516ab9ddc50cab77339d0cb3239ef4e7fd43bd2afdc73370aa487ea5704
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:dec430b74887e9b42d639f14c07d9f58a4881cac24e949dab98f96af150f1958
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:c61732be8024b4ede45eec8f786e51b198632ee5bfb0c2f5af84f37544584283
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:5ba79dd6bdf9e3778d3e301919b37e45cfc1eaf72fda77eba6588788829ba29b
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:c14cd037d14489321629908fc6b30248f7a20d866619177a165d72cfc1b6b7ef
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:83c03929cf11d7202e6c6cb9dd767bd4d7473f4fc1d562b79719ceb3d8382e68
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:6b7beed3cba9093b3aeb3109f23ec38c20f5568915dff774ad5e9e6bb5db3a58
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:e39732ebd3a4452fa7a678aeea5e64a7973fd72802eff193b16680c2f0129cb1
        - registry.redhat.io/container-native-virtualization/mtq-lock-server-rhel9@sha256:01cc6797800ccfb99453858ab36c0100c5ca343f1f71f975914b2437d1f757ba
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:9822f7a1590d830ccd64cc0729d2e30a2e04fb2b022f54498e1f1abff3748c2c
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:bf2ec130d1f3f32314d740037313f49f874a271e8b9a7f6cd2b93698a8423e8d
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:42a34caa18b5f5452b9d5603602ea74b1ee59a8bbb52bddf18901fecc27aed8a
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:18dfb38223647e11020e3ef1093a695454b90cc50922f80afb612212b7d13309
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:4c66a26b9f74d83a9bc60edb40ffd9d6ff2382d20500677d4567f8e141c3ddf0
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:758a9f6309a367d3f31cc03dd89356a8232a0fbd277570c01ebfa1fa30089a8d
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:2e5d5a2a70cda1f414f22b1c713d77d068c5fb16b1487a572019a9aa7130e6be
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:84b199e98e283c863d52a7b00044c20cb56ad563494adbec2229af265bf3f7c5
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:6baf67f3f410c071cc26f675d86edbbae5eaeb3410888cf622fad384b43b80ea
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:3a45cd2f3ea25f356167ce4fa20894a2a0706b4bcffa42c806f6017812053466
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:84679ffe8ce86c674818531897d122f5568d10c7e1e2ff919185d5771e5b4c39
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:ac33de246ee42bbfe735d7349d85c6df7e583e7599dea7dd7a2235bd0e59845f
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:e7145f016642b02bd651e21aa19edc7f93d0d1c2da3c6c7eed5d1174cf23619c
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:2564af054f52a8bb340e0270379c72da18658662a2a5bd26ac7bd979ddfccb56
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:4cff43cc145361c74c357055d51060ccc27e6bca58ccc3b5f5b983c7b8545114
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:09a30b59c906ddc4604d70f18210bcebef1f90f8bf965422b0caba363cd88989
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:63389195cf584a07ddbc2efec90cf6deaf737b91314b67db2b9f53dcc7acb57c
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:6193e569d4ab43806ac7b4870ef3634a5de3687527d6d190e59a9f735fa4778c
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:f4e7876fa8b27a8082efa99455d9bf7b0c479cff1e4c8a26136bf63092699c0b
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:d7e8e075a08b148ddfa91b44bb9cae1ef823756760279c4bb6d8489bc7e0eb2c
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:bf30b6c7173ebb6a4592246af16db87d7d391182f52286b559f8bad8f4b6cc5b
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:15179f0826831f051fdb867067be7c5528f01f442ae421b8782ae0cd1ed1b73d
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:4b3ac5ec2fafef0b7702a02847996f0903ec8d4262ccd03e8fa74885265ea4e9
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:d6dfd062df3d0d49f7efb928b835b9c8e0bcfb514a653ba2896a0152f24a0a21
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:185a83c994aced74546fe630cd71656f1e4112cb15be75f41dfcb1b9ed0966bd
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:b59e34cdb6214eacb9dc176a7b33125d7c1871f6b81696d59646294bfaf749e7
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:8e19d99232be4c37e857a4931627fb0aed721893e27b0d0e522fcf13bf67c1c9
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:06cc71822684c3ed094f327553451daf076d6bb97065c314da1a7d2adbbef3d2
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:4bdafc4404d5cfd844d05f32d88c262fba56dc931bfd5522b78c25d4ff253883
        - registry.redhat.io/container-native-virtualization/kubevirt-ipam-controller-rhel9@sha256:f33ac4d7477a730f273a462f171aed35dea38374d5ea7ef0bae2d7f74413a757
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:af5d2a2d4f0b6f0e1a13f1029073bb56980d8a70503b21628f18e8986e9b531c
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:5b2c64218342a5744182fc06f550bcdcf954e931275a9d1505ef0bbfbd0f8044
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:88f02cbebd6c3a2add6e0f8ca744e42bbd182f5a07d4fac59cf5cf07a4edfd45
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:fa0478fbf91248166f0eed8ec5f8dbb5b31d26031438376d9301836b6a069643
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:9bfdaf1cca6eac15b63a4ea6b7de28f2328516c5cf8bd5727935dd729e45657d
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:30c669770b0b992a3c84bb81c892e70a24fb14af187e8a4ee19cb3e0dd56ddbb
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:c5e8f44f647f593b00d3427172932d7aa694379d124209b9c66b4c1ffe3d4116
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:e6d9a6c41433360abd42d1118f5af3df9592d6d36af5f65082ae214cc8513c1b
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:d74c629e415d177069a0570fb8db9252d2b193d2ec7972c844686ffd14033f62
        - registry.redhat.io/container-native-virtualization/mtq-operator-rhel9@sha256:033bf7a6ef8ced1ac7275a9ddea8b16a10340a0fb0fb39619d89354d661f05b6
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:442e5cbcc63961642b662544032240bffb3863227324128a67f0bb6a05a1c364
        version: 4.99.0-0.1723448771
      entries:
      - name: kubevirt-hyperconverged-operator.v4.99.0-0.1723448771
        version: 4.99.0-0.1723448771
      name: dev-preview
    - currentCSV: kubevirt-hyperconverged-operator.v4.16.21
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:08f15657a1eee02927c5a6ef661e6e703d0064e091751bd2d8fbad5f118d9672
          createdAt: "2025-09-20 19:53:21"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.14.1 <4.16.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:88863a8dd2145bf17f7c5378c81e7c0be7a01dc47ee75bd7e5203a8fed8895b9
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.16.21-3,hyperconverged-cluster-operator-test-rhel9:v4.16.21-1,kubevirt-common-instancetypes-test-rhel9:v4.16.21-1,kubevirt-tekton-tasks-test-rhel9:v4.16.21-1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:9afe1ec9f776854caae38b89198a82e8e47a45d6c08f8dbd65a06c85f741c6a0
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:611d60b6ea4285fbc8a9f945fa5d06c09679a8f7ec1d70ec18edae7eec1a6bb3
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:299825fa33ab3f491cf954dd2e4ba23fdba2a96723b9ac3d98b3264b0b9fbb28
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:f6b075a56380143a7892e1fe5fdb212f8aa1c974f6a2d7e7e9b940789e44c010
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:d346688e02d03ddd9918f7a344d6eb3d5ddac3f1802dc82daf72bc56e7a9d496
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:406681aff7ea0a17941449593f839418edffdb073e2afffd53174f0df6292f94
        - registry.redhat.io/container-native-virtualization/mtq-controller-rhel9@sha256:253ef56b2867d33b3a475a456878b4a60f56473122eeed8c169568a9724ed30c
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:e54d87e6735151ed3c5295b23ca2f7ec666afb6421923636ef836bca1cf883d6
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:6385b346b3dd8f2bb3284871d326c851cf743a5e3fc3c49a1e51a1d43b0eafb4
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:8cd1052b9f69d0a0d8db41e90414756e83cbd07d7bd03c2faa85c41393d6539c
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:b80e68822fb0b5e78a995a97e02fe2ec448d67a5aa2224497fdf12b812182a0b
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:51a1fb0eb77e4f5c8bdfaf3b5074369b93ef63f5a7af62e22ed0789731f8dc73
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:361a48df213d52ced2a9963c2a8bde72a0981f1c497235fb918ea8c065580eeb
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:355e0946cfa963aeb5b6c3e8d476a9c3aee202baf77b95ce3f69c8e4b9e73836
        - registry.redhat.io/container-native-virtualization/mtq-lock-server-rhel9@sha256:c447251fea804340047bbaff885ad56be7a0d2612af1de26765f6cfc37da25e6
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:3daa460ed347e6eaab290b0a7fb0a08da6f210fb8be88af33f6e256511af80b6
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:dcc73f518ca7557cbd84b236367b3ecb7baab405268d54543783b4b772e14568
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:f7e5e04adb0467eb0aa4691a68b0d6485db9a52f99494097bf5b598e08f2edfb
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:2a109f94282b76ca0ff929adc35c93c0a98e34b43aba248d2975957fa149133f
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:cca121312777d2727fe2c50a14583243ce69e6e356b499f133736787522095b2
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:a0af0e5d4b142f28dc72c5636df5040c4ce92c6b12ab183d3a9bc7a9d2d2f418
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:88863a8dd2145bf17f7c5378c81e7c0be7a01dc47ee75bd7e5203a8fed8895b9
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:8cf7993326ba220f650ae2fa5383d0d6cf3e2c7f3346f0a339a9d212b8df02f6
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:edf12760e7fe4c8d37c22e172c0da270c219b7ff16d4ebb6b62729545309f6fa
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:a040696505e266e8f06a5f06b108f2db5ad3a6ed803d547d736455c1928a6a0e
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:d9560ed2758d3793139d3e13b7e5ce9d0f4ac0f8a6a8767449822c033cb2a456
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:9d997314bc7c6b30c2c3ff54b27bcbc1cdeac6dba300cab8a898b97fbce05ff7
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:dbff3d1592b192253a288fb3230a0ff125564540a30a6dd5ac881e79ed93b60a
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:9ae64f17191b98bb75f613969a9415103dad304e4fd20fb444f4fc4e57a57e0e
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:14952f38ab1d7f0abe4d54883e2c93d01ec1240ae809aefa9ca8cb4aaef080e1
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:15a971cef680f6f43e6b8ad7e0f5dab572ebacebc35cfbc6f7949a4ccd7789c2
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:d41ece8df9482ff860706edcdf27b53ac97d7c87407b3acfe8b1d9adfbf40e38
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:7581cd2330c20be2b77621e0e14da183d2b94f40fa5d3f448dede5f8c182759e
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:6fbdfdf3b9206058d5efc6653194bf6c42c07655817e3c97044da678e84b310a
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:3aa57a7ee9b342844ea4aa20bbbdf7a43b5d41f7866baa619f25473a230011f3
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:91b86228495f38d5528be1a8ffec96695ad72fafb76d2ead87f7bee25540d8b6
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:2e240514425d70039f5b48fe1908c56fae6efcc3eff93c63ec67f4aa3318a5fb
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:69388282df18b68c1d937ac98a5026b34221a54e5ce4b9cf4b89379b66623488
        - registry.redhat.io/container-native-virtualization/mtq-operator-rhel9@sha256:e29ffbe332d5da66ab4612a50ed73813ccbde92622474273b617cccd3bc24cd7
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:d253f1dbb54c70cf5874c2177bcd1cc3697381916cd4cf825bda934bfdb3d53c
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:7752f41a0f380fcd0d2652d286d4aac1ebb94a8a5fbb655b2e7058054f2eca4d
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:b20230f9a74652cf676bcf05dad81fc1ff64f9cff7232782c62cf585db48b115
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:0c5e1ee8ae0f19237a2c8b93466a7fc217a112cf1fb6dd55e2be27f95c9b66a3
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:bf7562a3bbc7655565b69fc96f2deae15729bbdef158cd90b97fc5ecc58d5bd8
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:fd2b4a337a8e45061cb6f455de3eefe27b004844c724fd821efc3b47215b74d3
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:77211068f3a8d331f1b21c64918861398f2cdbb2a3b760a5e8df51c89efea5b9
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:cddcdbf1d02b6e55cd93787991b5cf8704d0fb3a4a0bef5f58c918b57e26ee16
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:387fba10dc8f16fbfbc8c84274c5f1f5f6d01cf11f17f1b54fa0775a3300c6de
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:53ae4db672daad95d99f2955545ba335095c96630efc2a2ff19c5019a9413e87
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:63262574fa7d51d3b8c26b79e22d78f462c70239746aec6b8b3b951a02919e81
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:4046e8e862e0c55f09e9031b95e6a77af4d491593e39770c644287d44888b0bb
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:42fc0d1d112b30f65d6a48fa0442e2f41872658d2431ac8618dadf0aa7a41208
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:9e87e7d8679211a317d11e588868b40235da8c668bfb2ee10b564f6850fb91d1
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:38e767770ea0a74c4d594a382eed72bf9efa527e45ac6438acb79a1039f57fbd
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:75d76d7274f5169a414d8024125d4f0d8709888ac2b5d6121360635c7967d878
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:96092abe951f495f4f7dd0bdcb371b7efcb03a78a15e3ed8f1b37502203a6316
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:08f15657a1eee02927c5a6ef661e6e703d0064e091751bd2d8fbad5f118d9672
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:7e19b42a648cee71c60cb69fbcf6ebca05c938942e16d23f1cd61a619cffdcb6
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:878e1c8bd72cb3002457b6fbcddca99c0f4b00f6231fea52387d540b12e6329f
        version: 4.16.21
      entries:
      - name: kubevirt-hyperconverged-operator.v4.16.21
        version: 4.16.21
      - name: kubevirt-hyperconverged-operator.v4.16.17
        version: 4.16.17
      - name: kubevirt-hyperconverged-operator.v4.16.13
        version: 4.16.13
      - name: kubevirt-hyperconverged-operator.v4.16.7
        version: 4.16.7
      - name: kubevirt-hyperconverged-operator.v4.16.6
        version: 4.16.6
      - name: kubevirt-hyperconverged-operator.v4.16.5
        version: 4.16.5
      - name: kubevirt-hyperconverged-operator.v4.16.4
        version: 4.16.4
      - name: kubevirt-hyperconverged-operator.v4.16.3
        version: 4.16.3
      - name: kubevirt-hyperconverged-operator.v4.16.2
        version: 4.16.2
      - name: kubevirt-hyperconverged-operator.v4.16.1
        version: 4.16.1
      - name: kubevirt-hyperconverged-operator.v4.16.0
        version: 4.16.0
      - name: kubevirt-hyperconverged-operator.v4.15.2
        version: 4.15.2
      - name: kubevirt-hyperconverged-operator.v4.15.1
        version: 4.15.1
      - name: kubevirt-hyperconverged-operator.v4.15.0
        version: 4.15.0
      - name: kubevirt-hyperconverged-operator.v4.14.3
        version: 4.14.3
      - name: kubevirt-hyperconverged-operator.v4.14.2
        version: 4.14.2
      - name: kubevirt-hyperconverged-operator.v4.14.1
        version: 4.14.1
      - name: kubevirt-hyperconverged-operator.v4.14.0
        version: 4.14.0
      - name: kubevirt-hyperconverged-operator.v4.13.4
        version: 4.13.4
      - name: kubevirt-hyperconverged-operator.v4.13.3
        version: 4.13.3
      - name: kubevirt-hyperconverged-operator.v4.13.2
        version: 4.13.2
      - name: kubevirt-hyperconverged-operator.v4.13.1
        version: 4.13.1
      - name: kubevirt-hyperconverged-operator.v4.13.0
        version: 4.13.0
      - name: kubevirt-hyperconverged-operator.v4.12.2
        version: 4.12.2
      - name: kubevirt-hyperconverged-operator.v4.12.1
        version: 4.12.1
      - name: kubevirt-hyperconverged-operator.v4.12.0
        version: 4.12.0
      name: stable
    defaultChannel: stable
    packageName: kubevirt-hyperconverged
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: http://marketplace.redhat.com
    name: marketplace-games-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marketplace-games-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "game.marketplace.redhat.com/v1",
                "kind": "Arcade",
                "metadata": {
                  "name": "arcade-sample"
                },
                "spec": {
                  "size": 1
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Arcade is the Schema for the arcades API
            displayName: Arcade Instance
            kind: Arcade
            name: arcades.game.marketplace.redhat.com
            version: v1
        description: RHM Arcade Operator - a simple operator used to showcase how
          to build an operator around an existing web application that can can run
          on Openshift, Vanilla k8s, or on-boarded to Red Hat Marketplace.
        displayName: Arcade Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - rhm
        - red hat
        - marketplace
        - games
        - arcade
        links:
        - name: Marketplace Games Operator
          url: https://github.com/redhat-marketplace/marketplace-games-operator
        maturity: alpha
        provider:
          name: Red Hat Marketplace
          url: http://marketplace.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.connect.redhat.com/rhm-labs/arcade-operator@sha256:d6959be6693728d2dc4cd38b69251551d23c5002cb52db2684df31a0a8f7b0ab
        version: 1.0.0
      entries:
      - name: marketplace-games-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: marketplace-games-operator
    provider:
      name: Red Hat Marketplace
      url: http://marketplace.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Provider Name
      provider-url: https://your.domain
    name: smilecdr-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smilecdr-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.smilecdr.com/v1alpha1",
                "kind": "Smilecdr",
                "metadata": {
                  "name": "smilecdr"
                },
                "spec": {
                  "labels": {
                    "application": "smilecdr",
                    "client": "smilecdr",
                    "env": "dev",
                    "version": "one"
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Smilecdr
            name: smilecdrs.operator.smilecdr.com
            version: v1alpha1
        description: Smilecdr Operator description. TODO.
        displayName: Smilecdr Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - smilecdr-operator
        links:
        - name: Smilecdr Operator
          url: https://smilecdr-operator.domain
        maintainers:
        - email: your@email.com
          name: Maintainer Name
        maturity: alpha
        provider:
          name: Provider Name
          url: https://your.domain
        relatedImages:
        - quay.io/jibran_shaukat/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - quay.io/jibran_shaukat/smilecdr@sha256:cd6b2cec6f5b9ad9e0f0434532f8fb00b6919fcd840e686d96ce08874f2154db
        version: 0.0.1
      entries:
      - name: smilecdr-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: smilecdr-operator
    provider:
      name: Provider Name
      url: https://your.domain
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: model-builder-for-vision-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-modelbuilder-for-vision-rhmp.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "Modelbuilder",
                "metadata": {
                  "name": "modelbuilder-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "backup_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "metadata_storage": {
                    "storage_class": "ibmc-block-bronze",
                    "storage_size": "20G"
                  },
                  "photo_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "500G"
                  },
                  "in_memory_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              },
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "ModelbuilderAWS",
                "metadata": {
                  "name": "modelbuilder-aws-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "postgres": {
                    "storage_size": "20G",
                    "storage_class": "gp2",
                    "backup_storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
          createdAt: "2021-08-28T12:00:00Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/model-builder-for-vision-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/model-builder-for-vision-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.0.3
          operators.operatorframework.io/builder: operator-sdk-v1.11.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/ibm-modelbuilder-for-vision/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Model Builder for Vision on AWS
            displayName: Deploy-Modelbuilder on AWS
            kind: ModelbuilderAWS
            name: modelbuildersaws.modelbuilder.com
            version: v1alpha1
          - description: Deploy Model Builder for Vision on IBM
            displayName: Deploy-Modelbuilder on IBM
            kind: Modelbuilder
            name: modelbuilders.modelbuilder.com
            version: v1alpha1
        description: IBM Services Software Model Builder for Vision is an AI training
          platform that can use the capabilities of IBM Cloud and Amazon Web Services
          and their GPUs to quickly train computer vision models for compatible mobile
          apps. IBM Services Software Inspection Workbench iOS and iPadOS app will
          be used as the exclusive UI for labeling and training. Specifically, IBM
          Model Builder for Vision will train computer vision models that can then
          be deployed for use on the IBM Services Software Inspector Portable , IBM
          Services Software Inspector Wearable and the IBM Maximo Visual Inspection
          Mobile apps.
        displayName: IBM Model Builder for Vision
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Model Builder
        - AI
        - Computer Vision
        - Inspection
        - Mobile
        - Core ML
        - TensorFlow Lite
        links:
        - name: IBM Model Builder for Vision
          url: http://ibm.biz/model_builder
        maintainers:
        - email: isshelp@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.connect.redhat.com/ibm/modelbuilder-augmentor@sha256:ef5d0dde3eb334d919c2773af63007f1ef5e735ae040ef77fffe3440f854f068
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/rhel8/redis-6@sha256:2f65f46b25d84a3cb41a6a10f93786bcab47b7354ca9ac13a8e602ea833eba60
        - registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
        - registry.connect.redhat.com/ibm/modelbuilder-image-server@sha256:ecdb665af651417c50335e7259b89ef8096521abec431976bd7abe4c1edc49d6
        - registry.connect.redhat.com/ibm/modelbuilder-auth-service@sha256:f96b444fdcf7781775c9eba942bcc5e5ebdcf5f50820447f5eb26061c4c674cd
        - registry.connect.redhat.com/ibm/modelbuilder-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/modelbuilder-cv-training-service@sha256:c3c9080b08b87022c784fd98ac3e2774ec89ce29a2dd19d91dee039e3fdb2b36
        version: 1.0.3
      entries:
      - name: ibm-modelbuilder-for-vision-rhmp.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: model-builder-for-vision-certified-rhmp
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: fuse-console
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-console.v7.10.3-0.1655306919.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "replicas": 1,
                  "version": "1.10",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "100Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:37dc38375f5d553ac1d9f20f66ce005ed9495354b81d7915ae3bd7dd6af132f3
          createdAt: "2020-09-14T10:59:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.10.3'
          olm.substitutesFor: fuse-console.v7.10.3-0.1652362834.p
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.10/html/managing_fuse
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:37dc38375f5d553ac1d9f20f66ce005ed9495354b81d7915ae3bd7dd6af132f3
        version: 7.10.3+0.1655306919.p
      entries:
      - name: fuse-console.v7.10.3
        version: 7.10.3
      - name: fuse-console.v7.10.3-0.1652362834.p
        version: 7.10.3+0.1652362834.p
      - name: fuse-console.v7.10.3-0.1655306919.p
        version: 7.10.3+0.1655306919.p
      name: 7.10.x
    - currentCSV: fuse-console.v7.11.1-0.1681224695.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.11",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:518b5ace9178e2aafd7b6f247fd41d9392abd1a7bccb4cb53783531820611193
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.11.1'
          olm.substitutesFor: fuse-console.v7.11.1-0.1675972857.p
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.11/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:518b5ace9178e2aafd7b6f247fd41d9392abd1a7bccb4cb53783531820611193
        version: 7.11.1+0.1681224695.p
      entries:
      - name: fuse-console.v7.11.1
        version: 7.11.1
      - name: fuse-console.v7.11.1-0.1669922044.p
        version: 7.11.1+0.1669922044.p
      - name: fuse-console.v7.11.1-0.1675972857.p
        version: 7.11.1+0.1675972857.p
      - name: fuse-console.v7.11.1-0.1681224695.p
        version: 7.11.1+0.1681224695.p
      name: 7.11.x
    - currentCSV: fuse-console.v7.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.12",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:4e6889bf76d323956e0b594db7381e8162fea6b108f53f9e6dd0b38cf3596c8d
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.12.1'
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.12/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:4e6889bf76d323956e0b594db7381e8162fea6b108f53f9e6dd0b38cf3596c8d
        version: 7.12.1
      entries:
      - name: fuse-console.v7.12.1
        version: 7.12.1
      name: 7.12.x
    - currentCSV: fuse-console.v7.13.0-0.1757604404.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.13",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:72d98254a821bce237b3dde896f2825628cb7b6df564e23949337590586e8bea
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.7.0 <7.13.0'
          olm.substitutesFor: fuse-console.v7.13.0-0.1754559584.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.13/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:72d98254a821bce237b3dde896f2825628cb7b6df564e23949337590586e8bea
        version: 7.13.0+0.1757604404.p
      entries:
      - name: fuse-console.v7.13.0-0.1744281421.p
        version: 7.13.0+0.1744281421.p
      - name: fuse-console.v7.13.0-0.1723709432.p
        version: 7.13.0+0.1723709432.p
      - name: fuse-console.v7.13.0-0.1757604404.p
        version: 7.13.0+0.1757604404.p
      - name: fuse-console.v7.13.0-0.1721224332.p
        version: 7.13.0+0.1721224332.p
      - name: fuse-console.v7.13.0-0.1744737026.p
        version: 7.13.0+0.1744737026.p
      - name: fuse-console.v7.13.0-0.1724659726.p
        version: 7.13.0+0.1724659726.p
      - name: fuse-console.v7.13.0-0.1725026290.p
        version: 7.13.0+0.1725026290.p
      - name: fuse-console.v7.13.0-0.1748960166.p
        version: 7.13.0+0.1748960166.p
      - name: fuse-console.v7.13.0-0.1738056254.p
        version: 7.13.0+0.1738056254.p
      - name: fuse-console.v7.13.0-0.1739885034.p
        version: 7.13.0+0.1739885034.p
      - name: fuse-console.v7.13.0-0.1741875974.p
        version: 7.13.0+0.1741875974.p
      - name: fuse-console.v7.13.0-0.1741884704.p
        version: 7.13.0+0.1741884704.p
      - name: fuse-console.v7.13.0-0.1719822976.p
        version: 7.13.0+0.1719822976.p
      - name: fuse-console.v7.13.0
        version: 7.13.0
      - name: fuse-console.v7.13.0-0.1731460809.p
        version: 7.13.0+0.1731460809.p
      - name: fuse-console.v7.13.0-0.1749640277.p
        version: 7.13.0+0.1749640277.p
      - name: fuse-console.v7.13.0-0.1751291353.p
        version: 7.13.0+0.1751291353.p
      - name: fuse-console.v7.13.0-0.1753702606.p
        version: 7.13.0+0.1753702606.p
      - name: fuse-console.v7.13.0-0.1753778626.p
        version: 7.13.0+0.1753778626.p
      - name: fuse-console.v7.13.0-0.1753785080.p
        version: 7.13.0+0.1753785080.p
      - name: fuse-console.v7.13.0-0.1753860418.p
        version: 7.13.0+0.1753860418.p
      - name: fuse-console.v7.13.0-0.1754303647.p
        version: 7.13.0+0.1754303647.p
      - name: fuse-console.v7.13.0-0.1754559584.p
        version: 7.13.0+0.1754559584.p
      - name: fuse-console.v7.13.0-0.1754567932.p
        version: 7.13.0+0.1754567932.p
      - name: fuse-console.v7.12.1
        version: 7.12.1
      name: 7.13.x
    - currentCSV: fuse-console.v7.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "replicas": 1,
                  "version": "1.9",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "100Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:846d57f3fd9f307a98b0169b6733c26299209d99ed189f219c48f36ba771b198
          createdAt: "2020-09-14T10:59:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.9.0'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.9/html/managing_fuse
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:846d57f3fd9f307a98b0169b6733c26299209d99ed189f219c48f36ba771b198
        version: 7.9.0
      entries:
      - name: fuse-console.v7.9.0
        version: 7.9.0
      name: 7.9.x
    defaultChannel: 7.13.x
    packageName: fuse-console
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: opentelemetry-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opentelemetry-operator.v0.135.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "Instrumentation",
                "metadata": {
                  "name": "instrumentation"
                },
                "spec": {
                  "exporter": {
                    "endpoint": "http://otel-collector-headless:4317"
                  }
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpAMPBridge",
                "metadata": {
                  "name": "opampbridge-sample"
                },
                "spec": {
                  "capabilities": {
                    "AcceptsOpAMPConnectionSettings": true,
                    "AcceptsOtherConnectionSettings": true,
                    "AcceptsRemoteConfig": true,
                    "AcceptsRestartCommand": true,
                    "ReportsEffectiveConfig": true,
                    "ReportsHealth": true,
                    "ReportsOwnLogs": true,
                    "ReportsOwnMetrics": true,
                    "ReportsOwnTraces": true,
                    "ReportsRemoteConfig": true,
                    "ReportsStatus": true
                  },
                  "componentsAllowed": {
                    "exporters": [
                      "debug"
                    ],
                    "processors": [
                      "memory_limiter"
                    ],
                    "receivers": [
                      "otlp"
                    ]
                  },
                  "endpoint": "ws://opamp-server:4320/v1/opamp"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": "receivers:\n  otlp:\n    protocols: \n      grpc: {}\n      http: {}\n\nexporters:\n  debug: {}\n\nservice:\n  pipelines:\n    traces:\n      receivers: [otlp]\n      exporters: [debug]\n"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1beta1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": {
                    "exporters": {
                      "debug": {}
                    },
                    "receivers": {
                      "otlp": {
                        "protocols": {
                          "grpc": {},
                          "http": {}
                        }
                      }
                    },
                    "service": {
                      "pipelines": {
                        "traces": {
                          "exporters": [
                            "debug"
                          ],
                          "receivers": [
                            "otlp"
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/rhosdt/opentelemetry-rhel8-operator@sha256:45c6c52744e9a17ffd883a2864cb7722ce3f9580e522c976d4f55623118f3504
          createdAt: 24 Sep 2025, 10:19
          description: Provides the OpenTelemetry components, including the Collector
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.33.0 <0.135.0-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-opentelemetry-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-telemetry/opentelemetry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instrumentation is the spec for OpenTelemetry instrumentation.
            displayName: OpenTelemetry Instrumentation
            kind: Instrumentation
            name: instrumentations.opentelemetry.io
            version: v1alpha1
          - description: OpAMPBridge is the Schema for the opampbridges API.
            displayName: OpAMP Bridge
            kind: OpAMPBridge
            name: opampbridges.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1beta1
          - description: TargetAllocator is the Schema for the targetallocators API.
            displayName: Target Allocator
            kind: TargetAllocator
            name: targetallocators.opentelemetry.io
            version: v1alpha1
        description: |-
          Red Hat build of OpenTelemetry is a collection of tools, APIs, and SDKs. You use it to instrument, generate, collect, and export telemetry data (metrics, logs, and traces) for analysis in order to understand your software's performance and behavior.
          This operator was previously called Red Hat OpenShift distributed tracing data collection.

          ### Operator features

          * **Sidecar injection** - annotate your pods and let the operator inject a sidecar.
          * **Managed upgrades** - updating the operator will automatically update your OpenTelemetry collectors.
          * **Deployment modes** - your collector can be deployed as sidecar, daemon set, or regular deployment.
          * **Service port management** - the operator detects which ports need to be exposed based on the provided configuration.

          ## Examples

          There is a list of examples to help you deploy different configurations in [this GitHub repository](https://github.com/os-observability/redhat-rhosdt-samples)

          ### Support & Troubleshooting

          Red Hat build of OpenTelemetry is available as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/4.17/html/red_hat_build_of_opentelemetry/index) is recommended when installing, configuring, and managing the Operator and instances.
        displayName: Red Hat build of OpenTelemetry
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opentelemetry
        - tracing
        - logging
        - metrics
        - monitoring
        - troubleshooting
        links:
        - name: OpenTelemetry Operator
          url: https://github.com/open-telemetry/opentelemetry-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/opentelemetry-rhel8-operator@sha256:45c6c52744e9a17ffd883a2864cb7722ce3f9580e522c976d4f55623118f3504
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5a6edfecf1636805e6f0bcad02df10bceec764f0580f3ac8fe62aba978d50006
        - registry.redhat.io/rhosdt/opentelemetry-collector-rhel8@sha256:5f31caf1f4a12cf2832ac3e84877f4fb0c856073820498c8a5be13698348c38a
        - registry.redhat.io/rhosdt/opentelemetry-target-allocator-rhel8@sha256:f77a26edafba06a4fca1a07083e06a908664852cf228b1814455ffa8155cc979
        version: 0.135.0-1
      entries:
      - name: opentelemetry-operator.v0.135.0-1
        version: 0.135.0-1
      - name: opentelemetry-operator.v0.127.0-2
        version: 0.127.0-2
      - name: opentelemetry-operator.v0.127.0-1
        version: 0.127.0-1
      - name: opentelemetry-operator.v0.119.0-2
        version: 0.119.0-2
      - name: opentelemetry-operator.v0.119.0-1
        version: 0.119.0-1
      - name: opentelemetry-operator.v0.113.0-2
        version: 0.113.0-2
      - name: opentelemetry-operator.v0.113.0-1
        version: 0.113.0-1
      - name: opentelemetry-operator.v0.107.0-5
        version: 0.107.0-5
      - name: opentelemetry-operator.v0.107.0-4-0.1727168137.p
        version: 0.107.0-4+0.1727168137.p
      - name: opentelemetry-operator.v0.107.0-4
        version: 0.107.0-4
      - name: opentelemetry-operator.v0.102.0-3
        version: 0.102.0-3
      - name: opentelemetry-operator.v0.102.0-2
        version: 0.102.0-2
      - name: opentelemetry-operator.v0.100.1-3
        version: 0.100.1-3
      - name: opentelemetry-operator.v0.93.0-3
        version: 0.93.0-3
      - name: opentelemetry-operator.v0.93.0-2
        version: 0.93.0-2
      - name: opentelemetry-operator.v0.89.0-3
        version: 0.89.0-3
      - name: opentelemetry-operator.v0.81.1-5
        version: 0.81.1-5
      - name: opentelemetry-operator.v0.74.0-5-0.1687199949.p
        version: 0.74.0-5+0.1687199949.p
      - name: opentelemetry-operator.v0.74.0-5
        version: 0.74.0-5
      - name: opentelemetry-operator.v0.60.0-2
        version: 0.60.0-2
      - name: opentelemetry-operator.v0.56.0-1
        version: 0.56.0-1
      name: stable
    defaultChannel: stable
    packageName: opentelemetry-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://www.medik8s.io/
    name: customized-user-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: customized-user-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediation",
                "metadata": {
                  "name": "customizeduserremediation-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediationConfig",
                "metadata": {
                  "name": "customizeduserremediationconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediationTemplate",
                "metadata": {
                  "name": "customizeduserremediationtemplate-sample",
                  "namespace": "openshift-workload-availability"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/customized-user-remediation-operator:v0.1.0
          createdAt: "2024-03-25 16:17:57"
          description: Customized User Remediation Operator will remediate unhealthy
            nodes by running a pre defined user script on the unhealthy node. Since
            the user has full control over the script content, is may be quite risky
            so apply discretion.
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/customized-user-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CustomizedUserRemediationConfig is the Schema for the customizeduserremediationconfigs
              API
            displayName: Customized User Remediation Config
            kind: CustomizedUserRemediationConfig
            name: customizeduserremediationconfigs.customized-user-remediation.medik8s.io
            version: v1alpha1
          - description: CustomizedUserRemediation is the Schema for the customizeduserremediations
              API
            displayName: Customized User Remediation
            kind: CustomizedUserRemediation
            name: customizeduserremediations.customized-user-remediation.medik8s.io
            version: v1alpha1
          - description: CustomizedUserRemediationTemplate is the Schema for the customizeduserremediationtemplates
              API
            displayName: Customized User Remediation Template
            kind: CustomizedUserRemediationTemplate
            name: customizeduserremediationtemplates.customized-user-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Customized User Remediation (CUR) works with Node Health Check to remediate unhealthy nodes by running a user customized script on the unhealthy node.
          This operator does not preform any fencing on the remediated node so users should be careful with the scripts that are being used.

          ### Compatibility
          CUR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Flexibility - CUR is quite flexible sine it runs in the remediation process a user customized script.
        displayName: Customized User Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - script
        - CUR
        links:
        - name: Customized User Remediation
          url: https://customized-user-remediation.domain
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Medik8s
          url: https://www.medik8s.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/medik8s/customized-user-remediation-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: customized-user-remediation.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: customized-user-remediation
    provider:
      name: Medik8s
      url: https://www.medik8s.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-loadcore-agents
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-loadcore-agents.v0.2.25
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2024-11-20T18:10:53Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <=10.0.0'
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Agents
            name: agents.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight LoadCore Agents Operator
        displayName: Keysight LoadCore Agents Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Loadcore Agents
          url: https://keysight-loadcore-agents.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:7efeeb8b29872a6f0271f651d7ae02c91daea16d853c50e374c310f044d8c76c
        - prod.harbor.keysight.digital/loadcore/keysight-loadcore-agents@sha256:96e18b6270097c23bad5445a15613e467d06c6fbdbaf39545edb79915aacc04f
        version: 0.2.25
      entries:
      - name: keysight-loadcore-agents.v0.2.25
        version: 0.2.25
      name: alpha
    defaultChannel: alpha
    packageName: keysight-loadcore-agents
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-secret-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secret-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/tools@sha256:47204b2697dfbab5eaf9d6dcb4c160dd427a148d6b840727239b773810e74fc9
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:a6d6631d35c32b874c995b6af889636ac8dd1f39dcb68810d9f9b10a4e2c1987
        - quay.io/stackable/tools@sha256:47204b2697dfbab5eaf9d6dcb4c160dd427a148d6b840727239b773810e74fc9
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:161e754437428c68490333337303fb30b695a9ac677c3a00a1dc27565b855ab8
        - quay.io/stackable/secret-operator@sha256:1267311830649f928b528512a3cdc843efad6dd32b851d99c84639fbe70e35a3
        version: 24.11.1
      entries:
      - name: secret-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: secret-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:6eb08b99259d62514d87844e1b7470382d500696fecc9a1232f6bb536ed897fa
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:dd730457133f619d8759269abcfa79d7aeb817e01ba7af2d3aa1417df406ea56
        - quay.io/stackable/secret-operator@sha256:6eb08b99259d62514d87844e1b7470382d500696fecc9a1232f6bb536ed897fa
        version: 25.3.0
      entries:
      - name: secret-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: secret-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
          - description: Trust stores managed by the operator.
            displayName: Trust stores managed by the operator.
            kind: TrustStore
            name: truststores.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
        version: 25.7.0
      entries:
      - name: secret-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: secret-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
          - description: Trust stores managed by the operator.
            displayName: Trust stores managed by the operator.
            kind: TrustStore
            name: truststores.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        version: 25.7.0
      entries:
      - name: secret-operator.v25.7.0
        version: 25.7.0
      - name: secret-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-secret-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zettaset
      provider-url: ""
    name: xcrypt-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zts-xcrypt-operator.v0.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xcrypt.zettaset.com/v1alpha1",
                "kind": "XCrypt",
                "metadata": {
                  "name": "zts-masterset"
                },
                "spec": {
                  "status": {
                    "podNames": "zts-ca",
                    "replicas": "1,"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
          createdAt: "2021-06-16T12:29:13.97Z"
          description: XCrypt Operator deploys Zettaset XCrypt Container Encryption  -
            the only software-defined solution that offers transparent high performance
            data-at-rest encryption of critical data in container environments.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/xcrypt-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/xcrypt-operator-rhmp/support?utm_source=openshift_console
          support: Zettaset
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XCrypt is the Schema for the xcrypts API
            kind: XCrypt
            name: xcrypts.xcrypt.zettaset.com
            version: v1alpha1
        description: "Zettaset’s RedHat Certified XCrypt Operator allows users to
          deploy multiple XCrypt components on OpenShift with just a few simple commands.
          The XCrypt Operator deployment provides software-defined encryption that
          transparently protects container data throughout your OpenShift environment.\n\n##
          Core Features\n* **Automated XCrypt provisioning** - Deploying a Zettaset
          XCrypt has never been easier. Adjust cluster specific configuration and
          let the Zettaset Operator take care of provisioning XCrypt services and
          setting up the XCrypt encryption to your exact specification. \n* **CSI
          Driver deployment**: Zettaset XCrypt Operator automatically deploys Zettaset
          XCrypt CSI Driver in order to provide automated and transparent data-at-rest
          \ encryption functionality compatible with Openshift and any other container
          orchestrator. Once the Operator is provisioned, the driver runs fine tuned
          and ready to work with other XCrypt components.\n* **Multi-node topology**:
          Encrypt as many nodes and volumes as you need to. XCrypt components can
          be installed on multiple nodes across the cluster. This is defined by two
          labels: zts-master and zts-worker. Normally zts-master node is a single
          node that runs Zettaset XCrypt major services required for the product operation.
          Multiple nodes can be marked as zts-worker and those are the nodes that
          can have encrypted volumes.\n* **Seamless Encrypted Volume mount**: Volumes
          are not only being encrypted, but also are mounted/unmounted to containers
          with Persistent Volume Claims. This happens as seamlessly as regular volumes
          provision and mount/unmount.\n* **Robust failover**: A node failure doesn’t
          affect XCrypt operations on other nodes of the cluster. Encryption on a
          restarted/recovered node gets back to normal functioning if the node stays
          healthy and data hasn’t been corrupted during the restart\n\n## About Zettaset
          XCrypt Kubernetes Encryption for OpenShift\nXCrypt Kubernetes Encryption
          for OpenShift is a software-only solution that enables encryption of data
          at rest stored in Red Hat OpenShift container environments without slowing
          down the DevSecOps process. In addition to providing a transparent and high-performance
          layer of security that flexibly protects container data across on-premise,
          in hybrid, and multi-cloud environments, it also helps your organization
          efficiently transition from DevOps to DevSecOps.  \n\nMore information about
          Zettaset XCrypt Kubernetes Encryption for OpenShift can be found in the
          [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)\n\n##
          RedHat OpenShift + Zettaset: The next generation of secure enterprise infrastructure
          for DevSecOps  \nBy combining RedHat’s Container Platform with Zettaset’s
          certified software-defined encryption, you can create the next generation
          of secure enterprise infrastructure that DevSecOps requires. Together they
          create the stable, consistent, and supported base that your organization
          needs to develop applications faster, simplify updates, drive innovation,
          and reduce the risk of potential security breaches and data theft.\n###
          Core Features\n* Software-only for simple and scalable deployments \n* Negligible
          performance impact\n* Transparent to developers and administrators \n* Works
          with AWS, Google, and Azure \n* Unique encryption key per container volume
          \n* Encrypted volumes are only available when in use\n* Direct integration
          with OpenShift \n* Automated encryption policy management Secure erase of
          volumes\n* Ability to securely decommission a node\n\n## How does XCrypt
          Kubernetes Encryption fit into OpenShift? \nXCrypt XCrypt Kubernetes Encryption
          for OpenShift Encryption makes it incredibly easy to protect container data
          in your OpenShift environment. \n### Use cases include:\n* Transitioning
          from DevOps to DevSecOps \n* Ensuring data protection in OpenShift container
          environments with negligible performance impact \n* Achieving compliance
          in regulated industries like healthcare and finance\n\n## Other Information\nRole-based
          access control needs to be configured prior to the Operator installation.
          Please find details in the [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)
          \   "
        displayName: Zettaset XCrypt Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - encryption
        - storage
        links:
        - name: Zettaset
          url: https://www.zettaset.com/
        maintainers:
        - email: info@zettaset.com
          name: Zettaset
        maturity: alpha
        provider:
          name: Zettaset
        relatedImages:
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-ca-1@sha256:9f1015dfb459b69f5dbf449efa0517452095a6cf4f20826ebae278f9f800ffe7
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-kmip-server-1@sha256:9af5d21dd748dfbf4e107f27a5f10fa3c0ba928ade9b989a744b620ce34ec757
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-license-server-1@sha256:bc5486892a41e70f5dc4c4e1d82fca8db2149598fb8b5950ba5da1ee2d9d245d
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-host-manager-1@sha256:9faa4b597eb74df412761dffa057b64fb69dbf9fa016414962eb2099a343e734
        - registry.marketplace.redhat.com/rhm/zts/zts-csi-driver-1@sha256:c15fa9a7cc40891103f426206663dcebe1210c8ee3ab6037cbf64e1f86b9c7e5
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
        version: 0.0.17
      entries:
      - name: zts-xcrypt-operator.v0.0.17
        version: 0.0.17
      name: alpha
    defaultChannel: alpha
    packageName: xcrypt-operator-rhmp
    provider:
      name: Zettaset
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: openshift-secondary-scheduler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secondaryscheduleroperator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "SecondaryScheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-secondary-scheduler-operator"
                },
                "spec": {
                  "managementState": "Managed",
                  "schedulerConfig": "secondary-scheduler-config"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-rhel9-operator@sha256:210cc51760c8d3a9e1e5ba2b41f575b667e0d541b995362426e5a28f3f666600
          createdAt: 2024/10/01
          description: Runs a secondary scheduler in an OpenShift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.2.0 <1.3.1'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/secondary-scheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SecondaryScheduler is the schema for the secondaryschedulers
              API
            displayName: Secondary Scheduler
            kind: SecondaryScheduler
            name: secondaryschedulers.operator.openshift.io
            version: v1
        description: |
          The Secondary Scheduler Operator provides the ability to use a customized scheduler image that was developed using the scheduler plugin framework as a secondary scheduler in OpenShift.

          ## Secondary Scheduler Configuration

          Once the Secondary Scheduler Operator is installed, you can configure the plugins to run for your secondary scheduler.

          The configuration of the secondary scheduler is defined through a config map that wraps the KubeSchedulerConfiguration YAML file in "config.yaml".

          ## Additional Parameters

          Additionally, the following parameters can be configured:

          * `schedulerConfig` - Set the config map configuration for the secondary scheduler.
          * `schedulerImage` - Set the default operand image for the secondary scheduler.
        displayName: Secondary Scheduler Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - secondary
        - scheduling
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/secondary-scheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-rhel9-operator@sha256:210cc51760c8d3a9e1e5ba2b41f575b667e0d541b995362426e5a28f3f666600
        version: 1.3.1
      entries:
      - name: secondaryscheduleroperator.v1.3.1
        version: 1.3.1
      - name: secondaryscheduleroperator.v1.3.0
        version: 1.3.0
      - name: secondaryscheduleroperator.v1.2.2
        version: 1.2.2
      - name: secondaryscheduleroperator.v1.2.1
        version: 1.2.1
      - name: secondaryscheduleroperator.v1.2.0
        version: 1.2.0
      - name: secondaryscheduleroperator.v1.1.4
        version: 1.1.4
      - name: secondaryscheduleroperator.v1.1.2
        version: 1.1.2
      name: stable
    defaultChannel: stable
    packageName: openshift-secondary-scheduler-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: costmanagement-metrics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: costmanagement-metrics-operator.4.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "true"
          containerImage: registry.redhat.io/costmanagement/costmanagement-metrics-rhel9-operator@sha256:c5bb5291052d57390b9f4bdc660c1926827072b6dcbe09d85f5b6fc375562182
          createdAt: "2025-08-18T12:10:15Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: costmanagement-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Cost Management Metrics Operator
          ## Introduction
          The `costmanagement-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Cost Management Metrics Operator (`costmanagement-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created CostManagementMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The CostManagementMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **FIPS Compliance:** Supports deployment in high-security environments when run on an OpenShift cluster with FIPS mode enabled.

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created CostManagementMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the CostManagementMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `CostManagementMetricsConfig`, creating a new `CostManagementMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `costmanagement-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: costmanagement-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the CostManagementMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the CostManagementMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon CostManagementMetricsConfig creation** (This parameter will start to appear in CostManagementMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the CostManagementMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the costmanagement-metrics-operator
          **Note** There are separate instructions for configuring the `costmanagement-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `costmanagement-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the CostManagementMetricsConfig
          Configure the costmanagement-metrics-operator by creating a `CostManagementMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `costmanagement-metrics-operator` -> `CostManagementMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the costmanagement-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `costmanagement-metrics-operator-data` with 10Gi of storage. To configure the costmanagement-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `costmanagement-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `costmanagement-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the costmanagement-metrics-operator for a restricted network
          ##### Create the CostManagementMetricsConfig
          Configure the costmanagement-metrics-operator by creating a `CostManagementMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `costmanagement-metrics-operator` -> `CostManagementMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `costmanagement-metrics-operator-data` with 10Gi of storage. To configure the costmanagement-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `costmanagement-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: costmanagement-metrics-operator
                labels:
                  app: costmanagement-metrics-operator
              spec:
                volumes:
                - name: costmanagement-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: costmanagement-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: costmanagement-metrics-operator-reports
                    mountPath: /tmp/costmanagement-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/costmanagement-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/costmanagement-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `costmanagement-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the CostManagementMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Cost Management Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Cost Management Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/costmanagement/costmanagement-metrics-rhel9-operator@sha256:c5bb5291052d57390b9f4bdc660c1926827072b6dcbe09d85f5b6fc375562182
        version: 4.1.0
      entries:
      - name: costmanagement-metrics-operator.4.1.0
        version: 4.1.0
      - name: costmanagement-metrics-operator.4.0.0
        version: 4.0.0
      - name: costmanagement-metrics-operator.3.3.2
        version: 3.3.2
      - name: costmanagement-metrics-operator.3.3.1
        version: 3.3.1
      - name: costmanagement-metrics-operator.3.3.0
        version: 3.3.0
      - name: costmanagement-metrics-operator.3.2.1
        version: 3.2.1
      - name: costmanagement-metrics-operator.3.2.0
        version: 3.2.0
      - name: costmanagement-metrics-operator.3.1.1
        version: 3.1.1
      - name: costmanagement-metrics-operator.3.1.0
        version: 3.1.0
      - name: costmanagement-metrics-operator.3.0.1
        version: 3.0.1
      - name: costmanagement-metrics-operator.3.0.0
        version: 3.0.0
      - name: costmanagement-metrics-operator.2.0.0
        version: 2.0.0
      - name: costmanagement-metrics-operator.1.2.0
        version: 1.2.0
      - name: costmanagement-metrics-operator.1.1.9
        version: 1.1.9
      - name: costmanagement-metrics-operator.1.1.8
        version: 1.1.8
      - name: costmanagement-metrics-operator.1.1.7
        version: 1.1.7
      - name: costmanagement-metrics-operator.1.1.6
        version: 1.1.6
      - name: costmanagement-metrics-operator.1.1.5
        version: 1.1.5
      - name: costmanagement-metrics-operator.1.1.4
        version: 1.1.4
      - name: costmanagement-metrics-operator.1.1.3
        version: 1.1.3
      - name: costmanagement-metrics-operator.1.1.0
        version: 1.1.0
      - name: costmanagement-metrics-operator.1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: costmanagement-metrics-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-applicationautoscaling-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-applicationautoscaling-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "applicationautoscaling.services.k8s.aws/v1alpha1",
                "kind": "ScalableTarget",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "applicationautoscaling.services.k8s.aws/v1alpha1",
                "kind": "ScalingPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/applicationautoscaling-controller:1.1.1
          createdAt: "2025-09-20T21:43:01Z"
          description: AWS Application Auto Scaling controller is a service controller
            for managing Application Auto Scaling resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ScalableTarget represents the state of an AWS applicationautoscaling
              ScalableTarget resource.
            displayName: ScalableTarget
            kind: ScalableTarget
            name: scalabletargets.applicationautoscaling.services.k8s.aws
            version: v1alpha1
          - description: ScalingPolicy represents the state of an AWS applicationautoscaling
              ScalingPolicy resource.
            displayName: ScalingPolicy
            kind: ScalingPolicy
            name: scalingpolicies.applicationautoscaling.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Application Auto Scaling resources in AWS from within your Kubernetes cluster.
          **About Amazon Application Auto Scaling**
          Application Auto Scaling is a web service for developers and system administrators who need a solution for automatically scaling their scalable resources for individual AWS services beyond Amazon EC2. Application Auto Scaling allows you to configure automatic scaling for the following resources:
          - AppStream 2.0 fleets
          - Aurora replicas
          - Amazon Comprehend document classification and entity recognizer endpoints
          - DynamoDB tables and global secondary indexes
          - Amazon Elastic Container Service (ECS) services
          - Amazon EMR clusters
          - Amazon Keyspaces (for Apache Cassandra) tables
          - Lambda function provisioned concurrency
          - Amazon Managed Streaming for Apache Kafka (MSK) broker storage
          - SageMaker endpoint variants
          - Spot Fleet requests
          - Custom resources provided by your own applications or services. For more information, see the [GitHub repository](https://github.com/aws/aws-auto-scaling-custom-resource)

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Application Auto Scaling
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - applicationautoscaling
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Application Auto Scaling User Guide
          url: https://docs.aws.amazon.com/autoscaling/application/userguide/what-is-application-auto-scaling.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: application auto scaling maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/applicationautoscaling-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-applicationautoscaling-controller.v1.1.1
        version: 1.1.1
      - name: ack-applicationautoscaling-controller.v1.1.0
        version: 1.1.0
      - name: ack-applicationautoscaling-controller.v1.0.30
        version: 1.0.30
      - name: ack-applicationautoscaling-controller.v1.0.29
        version: 1.0.29
      - name: ack-applicationautoscaling-controller.v1.0.27
        version: 1.0.27
      - name: ack-applicationautoscaling-controller.v1.0.26
        version: 1.0.26
      - name: ack-applicationautoscaling-controller.v1.0.25
        version: 1.0.25
      - name: ack-applicationautoscaling-controller.v1.0.24
        version: 1.0.24
      - name: ack-applicationautoscaling-controller.v1.0.23
        version: 1.0.23
      - name: ack-applicationautoscaling-controller.v1.0.22
        version: 1.0.22
      - name: ack-applicationautoscaling-controller.v1.0.21
        version: 1.0.21
      - name: ack-applicationautoscaling-controller.v1.0.19
        version: 1.0.19
      - name: ack-applicationautoscaling-controller.v1.0.18
        version: 1.0.18
      - name: ack-applicationautoscaling-controller.v1.0.17
        version: 1.0.17
      - name: ack-applicationautoscaling-controller.v1.0.16
        version: 1.0.16
      - name: ack-applicationautoscaling-controller.v1.0.15
        version: 1.0.15
      - name: ack-applicationautoscaling-controller.v1.0.11
        version: 1.0.11
      - name: ack-applicationautoscaling-controller.v1.0.9
        version: 1.0.9
      - name: ack-applicationautoscaling-controller.v1.0.8
        version: 1.0.8
      - name: ack-applicationautoscaling-controller.v1.0.7
        version: 1.0.7
      - name: ack-applicationautoscaling-controller.v1.0.6
        version: 1.0.6
      - name: ack-applicationautoscaling-controller.v1.0.5
        version: 1.0.5
      - name: ack-applicationautoscaling-controller.v1.0.4
        version: 1.0.4
      - name: ack-applicationautoscaling-controller.v1.0.3
        version: 1.0.3
      - name: ack-applicationautoscaling-controller.v1.0.2
        version: 1.0.2
      - name: ack-applicationautoscaling-controller.v1.0.0
        version: 1.0.0
      - name: ack-applicationautoscaling-controller.v0.2.16
        version: 0.2.16
      - name: ack-applicationautoscaling-controller.v0.2.15
        version: 0.2.15
      - name: ack-applicationautoscaling-controller.v0.2.14
        version: 0.2.14
      - name: ack-applicationautoscaling-controller.v0.2.13
        version: 0.2.13
      - name: ack-applicationautoscaling-controller.v0.2.12
        version: 0.2.12
      - name: ack-applicationautoscaling-controller.v0.2.10
        version: 0.2.10
      - name: ack-applicationautoscaling-controller.v0.2.9
        version: 0.2.9
      - name: ack-applicationautoscaling-controller.v0.2.8
        version: 0.2.8
      - name: ack-applicationautoscaling-controller.v0.2.7
        version: 0.2.7
      - name: ack-applicationautoscaling-controller.v0.2.6
        version: 0.2.6
      - name: ack-applicationautoscaling-controller.v0.2.5
        version: 0.2.5
      - name: ack-applicationautoscaling-controller.v0.2.4
        version: 0.2.4
      - name: ack-applicationautoscaling-controller.v0.2.3
        version: 0.2.3
      - name: ack-applicationautoscaling-controller.v0.2.2
        version: 0.2.2
      - name: ack-applicationautoscaling-controller.v0.2.1
        version: 0.2.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-applicationautoscaling-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-acm-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-acm-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acm.services.k8s.aws/v1alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/acm-controller:1.1.1
          createdAt: "2025-09-20T01:19:56Z"
          description: AWS ACM controller is a service controller for managing ACM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Certificate represents the state of an AWS acm Certificate
              resource.
            displayName: Certificate
            kind: Certificate
            name: certificates.acm.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage AWS Certificate Manager (ACM) resources in AWS from within your Kubernetes cluster.

          **About Amazon ACM**

          AWS Certificate Manager (ACM) handles the complexity of creating, storing, and renewing public and private SSL/TLS X.509 certificates and keys that protect your AWS websites and applications. You can provide certificates for your [integrated AWS services](https://docs.aws.amazon.com/acm/latest/userguide/acm-services.html) either by issuing them directly with ACM or by [importing](https://docs.aws.amazon.com/acm/latest/userguide/import-certificate.html) third-party certificates into the ACM management system. ACM certificates can secure singular domain names, multiple specific domain names, wildcard domains, or combinations of these. ACM wildcard certificates can protect an unlimited number of subdomains. You can also [export](https://docs.aws.amazon.com/acm/latest/userguide/export-private.html) ACM certificates signed by AWS Private CA for use anywhere in your internal PKI.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon ACM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ACM Developer Resources
          url: https://aws.amazon.com/certificate-manager/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: acm maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/acm-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-acm-controller.v1.1.1
        version: 1.1.1
      - name: ack-acm-controller.v1.1.0
        version: 1.1.0
      - name: ack-acm-controller.v1.0.13
        version: 1.0.13
      - name: ack-acm-controller.v1.0.12
        version: 1.0.12
      - name: ack-acm-controller.v1.0.11
        version: 1.0.11
      - name: ack-acm-controller.v1.0.10
        version: 1.0.10
      - name: ack-acm-controller.v1.0.9
        version: 1.0.9
      - name: ack-acm-controller.v1.0.8
        version: 1.0.8
      - name: ack-acm-controller.v1.0.7
        version: 1.0.7
      - name: ack-acm-controller.v1.0.6
        version: 1.0.6
      - name: ack-acm-controller.v1.0.5
        version: 1.0.5
      - name: ack-acm-controller.v1.0.4
        version: 1.0.4
      - name: ack-acm-controller.v1.0.1
        version: 1.0.1
      - name: ack-acm-controller.v1.0.0
        version: 1.0.0
      - name: ack-acm-controller.v0.0.20
        version: 0.0.20
      - name: ack-acm-controller.v0.0.19
        version: 0.0.19
      - name: ack-acm-controller.v0.0.18
        version: 0.0.18
      - name: ack-acm-controller.v0.0.17
        version: 0.0.17
      - name: ack-acm-controller.v0.0.16
        version: 0.0.16
      - name: ack-acm-controller.v0.0.12
        version: 0.0.12
      - name: ack-acm-controller.v0.0.10
        version: 0.0.10
      - name: ack-acm-controller.v0.0.9
        version: 0.0.9
      - name: ack-acm-controller.v0.0.6
        version: 0.0.6
      - name: ack-acm-controller.v0.0.5
        version: 0.0.5
      - name: ack-acm-controller.v0.0.4
        version: 0.0.4
      - name: ack-acm-controller.v0.0.2
        version: 0.0.2
      - name: ack-acm-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-acm-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ricardo Maraschini
      provider-url: https://github.com/ricardomaraschini/tagger
    name: tagger
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tagger.v2.1.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tagger.dev/v1beta1",
                "kind": "Tag",
                "metadata": {
                  "name": "nginx"
                },
                "spec": {
                  "from": "docker.io/library/nginx:latest",
                  "generation": 0,
                  "mirror": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          containerImage: quay.io/tagger/operator:v2.1.17
          createdAt: "2021-08-02T00:00:00Z"
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/ricardomaraschini/tagger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tag is a Container Image representation as a Custom Resource.
            displayName: Tag
            kind: Tag
            name: tags.tagger.dev
            version: v1beta1
        description: |-
          A Container Image Orchestrator. Adds mirroring and Container Image manipulation capabilities
          to a Kubernetes cluster.

          Keeping track of all Container Images in use in a Kubernetes cluster is a complicated task.
          Container Images may come from numerous different Image Registries. In some cases, controlling
          how a stable version of a given Container Image looks escapes the user's authority. To add to
          this, Container Runtimes rely on remote registries (from the cluster's point of view) when
          obtaining Container Images, potentially making the process of pulling their blobs (manifests,
          config, and layers) slower.

          The notion of indexing Container Image versions by Tags is helpful. Still, it does not provide
          users with the right confidence to always use the intended Container Image – today's "latest"
          tag might not be tomorrow's "latest" tag. In addition to that, these Image Registries allow
          access to Container Images by their Manifest content's hash (i.e., usually sha256), which gives
          users the confidence at a cost in semantics.

          When releasing a new version of an application to Push and to Deploy are split into two
          distinct steps. Both the pusher and the puller need access to the same Image Registry, adding
          complexity.  Credentials are one example of the concerns. Other factors may pop up when
          running, for instance, in an air-gapped environment, where the cluster may not reach external
          Image Registries.

          Tagger aims to overcome these caveats by providing an image management abstraction layer. For
          instance, by providing a direct mapping between a Container Image tag (e.g., "latest") and its
          correspondent Manifest content's hash, users can then refer to the Container Image by its tag
          – and be sure to use that specific version. More than that, when allied with an Internal Image
          Registry, Tagger can also automatically mirror Container Images into the cluster.

          While using Tagger, Deployments can refer to Container Images by an arbitrarily defined name,
          such as "my-app", and Tagger will make sure that they use the right Container Image through its
          internal "tag to Manifest content's hash" mapping.

          For each new "release" of a given Container Image, Tagger creates a new Generation for it,
          making it easy to roll back to previously pushed "releases" of the same Container Image in case
          of problems.

          When integrated with an Internal Registry, Tagger allows users to push or pull Images directly
          without requiring an external Image Registry. It works as a layer between the user and the
          Internal Registry. Every time a new "release" of Container Images is pushed, all Deployments
          are updated automatically. Users don't need to know about the Internal Registry existence, if
          they are logged-in to the Kubernetes cluster, they can obtain old or provide new "Generations"
          for a Container Image.

          In summary, Tagger mirrors remote Container Images into a Kubernetes cluster indexing them in
          different Generations (allowing easy navigation through these multiple Generations), provides
          an interface allowing users to pull and push images directly to the Kubernetes cluster and
          provides full integration with Kubernetes Deployments (automatic triggers new rollouts on
          Container Image changes).

          By default Tagger won't be able to mirror until you provide it with a mirror registry config.
          After install you can configure the mirror by editing the Secret mirror-registry-config in the
          operator namespace. Follow below an example of a valid mirror-registry-config secret (you gonna
          have to provide your own address, username, password, etc).

          ```yaml
          apiVersion: v1
          kind: Secret
          metadata:
            name: mirror-registry-config
          data:
            address: cmVnaXN0cnkuaW8=
            username: YWRtaW4=
            password: d2hhdCB3ZXJlIHlvdSB0aGlua2luZz8K
            token: YW4gb3B0aW9uYWwgdG9rZW4gZm9yIHRva2VuIGJhc2VkIGF1dGg=
            insecure: dHJ1ZQ==
          ```

          To make pushing an pulling images easier you should go ahead and grab the `kubectl` plugin
          available [here](https://github.com/ricardomaraschini/tagger/releases). You can download a
          full version of the README with usage examples and configuration options
          [here](https://github.com/ricardomaraschini/tagger/releases/download/v2.1.17/README.pdf).
        displayName: Tagger
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tagger
        - mirroring
        - image
        links:
        - name: Documentation
          url: https://github.com/ricardomaraschini/tagger/releases/download/v2.1.17/README.pdf
        - name: Plugin (kubectl)
          url: https://github.com/ricardomaraschini/tagger/releases/tag/v2.1.17
        maintainers:
        - email: rmarasch@redhat.com
          name: Ricardo Maraschini
        maturity: alpha
        provider:
          name: Ricardo Maraschini
          url: https://github.com/ricardomaraschini/tagger
        relatedImages:
        - quay.io/tagger/operator:v2.1.17
        version: 2.1.17
      entries:
      - name: tagger.v2.1.17
        version: 2.1.17
      name: alpha
    defaultChannel: alpha
    packageName: tagger
    provider:
      name: Ricardo Maraschini
      url: https://github.com/ricardomaraschini/tagger
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecs-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "Service",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "TaskDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecs-controller:1.1.1
          createdAt: "2025-09-19T21:02:44Z"
          description: AWS ECS controller is a service controller for managing ECS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster represents the state of an AWS ecs Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.ecs.services.k8s.aws
            version: v1alpha1
          - description: Service represents the state of an AWS ecs Service resource.
            displayName: Service
            kind: Service
            name: services.ecs.services.k8s.aws
            version: v1alpha1
          - description: TaskDefinition represents the state of an AWS ecs TaskDefinition
              resource.
            displayName: TaskDefinition
            kind: TaskDefinition
            name: taskdefinitions.ecs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Container Service (ECS) resources in AWS from within your Kubernetes cluster.

          **About Amazon ECS**

          Amazon Elastic Container Service (Amazon ECS) is a fully managed container orchestration service that helps you easily deploy, manage, and scale containerized applications. As a fully managed service, Amazon ECS comes with AWS configuration and operational best practices built-in. It's integrated with both AWS and third-party tools, such as Amazon Elastic Container Registry and Docker. This integration makes it easier for teams to focus on building the applications, not the environment. You can run and scale your container workloads across AWS Regions in the cloud, and on-premises, without the complexity of managing a control plane.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECS Developer Resources
          url: https://aws.amazon.com/ecs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecs-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ecs-controller.v1.1.1
        version: 1.1.1
      - name: ack-ecs-controller.v1.0.15
        version: 1.0.15
      - name: ack-ecs-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecs-controller.v1.0.13
        version: 1.0.13
      - name: ack-ecs-controller.v1.0.12
        version: 1.0.12
      - name: ack-ecs-controller.v1.0.11
        version: 1.0.11
      - name: ack-ecs-controller.v1.0.10
        version: 1.0.10
      - name: ack-ecs-controller.v1.0.9
        version: 1.0.9
      - name: ack-ecs-controller.v1.0.8
        version: 1.0.8
      - name: ack-ecs-controller.v1.0.7
        version: 1.0.7
      - name: ack-ecs-controller.v1.0.6
        version: 1.0.6
      - name: ack-ecs-controller.v1.0.2
        version: 1.0.2
      - name: ack-ecs-controller.v1.0.1
        version: 1.0.1
      - name: ack-ecs-controller.v1.0.0
        version: 1.0.0
      - name: ack-ecs-controller.v0.0.8
        version: 0.0.8
      - name: ack-ecs-controller.v0.0.7
        version: 0.0.7
      - name: ack-ecs-controller.v0.0.6
        version: 0.0.6
      - name: ack-ecs-controller.v0.0.5
        version: 0.0.5
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseWeb
      provider-url: ""
    name: cnfv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cnfv-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eweb.enterpriseweb.com/v1",
                "kind": "CloudNFV",
                "metadata": {
                  "name": "cloudnfv-sample"
                },
                "spec": {
                  "size": 1
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CloudNFV is the Schema for the cloudnfvs API
            displayName: Cloud NFV
            kind: CloudNFV
            name: cloudnfvs.eweb.enterpriseweb.com
            version: v1
        description: Manages a cluster of CloudNFV Runtime Modules
        displayName: CloudNFV
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - automation
        - NFV
        - controller
        links:
        - name: Cnfv Operator
          url: https://cnfv-operator.domain
        maintainers:
        - email: bill@enterpriseweb.com
          name: Bill Malyk
        maturity: alpha
        provider:
          name: EnterpriseWeb
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61a164ad711478ac603242338f7b3c55986319479878b092be128581a823598e
        - quay.io/wmalyk28/cnfv-op@sha256:e5913ad5dd20cb1b2299d2387237ef15e645c9a8c9d4f5fe0044b86a6b3dc1df
        - quay.io/wmalyk28/cloudnfv-runtime@sha256:fcde8375576c9bff602d3a16ef1ba8ae88b780166649ba6ae355737b1d35a03f
        version: 0.0.1
      entries:
      - name: cnfv-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cnfv-operator
    provider:
      name: EnterpriseWeb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzounstructured-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzounstructured-operator.v3.1.200
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzounstructured.clusters.cambridgesemantics.com/v1",
                "kind": "AnzoUnstructured",
                "metadata": {
                  "name": "au01"
                },
                "spec": {
                  "auWorker": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-w",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-worker@sha256:964633d6b16ddc517a09ac86f3e3698f5a161a72cfa50ea9ba4e301fa631e347",
                                "name": "w",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  },
                  "msLeader": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-ms",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzo-microservices-leader@sha256:0637603c00a99533d57e16009516ec73d40703ae151e631fc281f51f1f0ca93b",
                                "name": "ms",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/unstructured-operator@sha256:633a8992e35601d99c0d690dd71af314605c223b57c2038fa7090be06977b2f9
          createdAt: "2025-07-08T09:47:55Z"
          description: Kubernetes Operator for Altair Graph Studio Unstructured
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoUnstructured is the Schema for the anzounstructureds
              API
            displayName: Altair Graph Studio Unstructured Operator
            kind: AnzoUnstructured
            name: anzounstructureds.anzounstructured.clusters.cambridgesemantics.com
            version: v1
        description: |-
          The Altair Graph Studio Unstructured Operator provides the way to install and configure an anzo unstructured setup on Red Hat K8S environment.
          Currently, this is possible only through existing Anzo installation.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v3.1.2/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Studio Unstructured Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Database
        - Big Data
        links:
        - name: Altair Graph Studio Unstructured Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v3.1.2/README.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/unstructured-operator@sha256:633a8992e35601d99c0d690dd71af314605c223b57c2038fa7090be06977b2f9
        - registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-worker@sha256:964633d6b16ddc517a09ac86f3e3698f5a161a72cfa50ea9ba4e301fa631e347
        - registry.connect.redhat.com/cambridgesemantics/anzo-microservices-leader@sha256:0637603c00a99533d57e16009516ec73d40703ae151e631fc281f51f1f0ca93b
        version: 3.1.200
      entries:
      - name: anzounstructured-operator.v3.1.200
        version: 3.1.200
      - name: anzounstructured-operator.v3.1.104
        version: 3.1.104
      - name: anzounstructured-operator.v3.1.103
        version: 3.1.103
      name: stable
    defaultChannel: stable
    packageName: anzounstructured-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: klusterlet
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: klusterlet.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "Klusterlet",
                "metadata": {
                  "name": "klusterlet"
                },
                "spec": {
                  "clusterName": "cluster1",
                  "deployOption": {
                    "mode": "Singleton"
                  },
                  "externalServerURLs": [
                    {
                      "url": "https://localhost"
                    }
                  ],
                  "imagePullSpec": "quay.io/open-cluster-management/registration-operator:v0.13.1",
                  "namespace": "open-cluster-management-agent",
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/registration-operator:v0.13.1
          createdAt: "2024-04-10T15:27:49Z"
          description: Manages the installation and upgrade of the Klusterlet.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management-io/ocm
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Klusterlet defines the configurations of agents running on
              the managed cluster
            displayName: Klusterlet
            kind: Klusterlet
            name: klusterlets.operator.open-cluster-management.io
            version: v1
        description: "Note: This operator will stop updating, please use the latest
          operator in OperatorHub.io https://operatorhub.io/operator/klusterlet. \n\nThe
          Klusterlet proivdes the registation to the Hub clusters as a managed cluster.
          This operator supports the installation and upgrade of Klusterlet.\n\n##
          Prerequisites\nYou need a Hub cluster which has installed ClusterManager.\n\n##
          How to Install\n* Install Klusterlet operator by following instructions
          in top right button `Install`.\n* Optional: Create an agent namespace used
          to deploy Klusterlet. The default namespace is `open-cluster-management-agent`.\n*
          Create a secret `bootstrap-hub-kubeconfig` with the kubeconfig file of hub
          cluster in the agent namespace.\n\n  ```\n  $ kubectl create secret generic
          bootstrap-hub-kubeconfig --from-file=kubeconfig=<kubeconfig file of hub
          clsuter> -n <agent namespace>\n  ```\n\n* Create an instance of the Klusterlet
          resource to install the Klusterlet.\n\nThe pods of Klusterlet are created
          in agent namespace.\n\n```\n$ kubectl get pods -n open-cluster-management-agent\n
          \ NAME                                READY   STATUS    RESTARTS   AGE\n
          \ klusterlet-agent-55c6574876-nfxm9   1/1     Running   0          8m38s\n```\n\nThere
          is a ManagedCluster created in the Hub cluster.\n\n```\n$ kubectl get managedclusters.cluster.open-cluster-management.io\nNAME
          \                  HUB ACCEPTED   MANAGED CLUSTER URLS     JOINED   AVAILABLE
          \  AGE\ncluster1               false                                                        27s\n```\n"
        displayName: Klusterlet
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open-cluster-management
        - Klusterlet
        links:
        - name: Registration Operator
          url: https://github.com/open-cluster-management-io/ocm
        maintainers:
        - email: zyin@redhat.com
          name: Zhiwei Yin
        - email: deads@redhat.com
          name: David Eads
        - email: jqiu@redhat.com
          name: Jian Qiu
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/open-cluster-management/registration-operator:v0.13.1
        version: 0.13.1
      entries:
      - name: klusterlet.v0.13.1
        version: 0.13.1
      - name: klusterlet.v0.13.0
        version: 0.13.0
      - name: klusterlet.v0.12.0
        version: 0.12.0
      - name: klusterlet.v0.11.0
        version: 0.11.0
      - name: klusterlet.v0.10.0
        version: 0.10.0
      - name: klusterlet.v0.9.1
        version: 0.9.1
      - name: klusterlet.v0.8.0
        version: 0.8.0
      - name: klusterlet.v0.7.0
        version: 0.7.0
      - name: klusterlet.v0.6.0
        version: 0.6.0
      - name: klusterlet.v0.5.0
        version: 0.5.0
      - name: klusterlet.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: klusterlet
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-api-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-api-controller.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "example-registry-with-studio"
                },
                "spec": {
                  "app": {
                    "host": "example-registry-with-studio-app.apps.yourcluster.example"
                  },
                  "studioUi": {
                    "enabled": true,
                    "host": "example-registry-with-studio-studio-ui.apps.yourcluster.example"
                  },
                  "ui": {
                    "host": "example-registry-with-studio-ui.apps.yourcluster.example"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-operator:3.0.5
          createdAt: "2024-12-03T11:30:52Z"
          description: Deploy and manage Apicurio API Controller on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Apicurio API
              Controller.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Apicurio API Controller

          Apicurio API Controller consists of 2 components:
           - Apicurio Registry 3
           - Apicurio Studio

          Storage options: PostgreSQL, Kafka, MySQL, SQL Server, in-memory.

          ## Apicurio API Controller Operator

          Provides a quick and easy way to deploy and manage Apicurio API Controller on Kubernetes.

          **Features**
          - Supports installation and configuration of Apicurio API Controller
          - Perform a rolling upgrade of Apicurio API Controller

          ## Prerequisites

          This operator does not deploy storage for Apicurio API Controller. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio API Controller and Apicurio API Controller Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Apicurio API Controller
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - API Management
        - Apicurio Registry
        - Apicurio Studio
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-operator:3.0.5
        version: 0.0.1
      entries:
      - name: apicurio-api-controller.v0.0.1
        version: 0.0.1
      name: 0.x
    defaultChannel: 0.x
    packageName: apicurio-api-controller
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: unsupported
      operatorframework.io/arch.s390x: unsupported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: https://www.konveyor.io
    name: konveyor-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: konveyor-operator.v0.8.0-alpha.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:c1d1942b3bb3c4e99d9b5484493e27a432455608fee4548f1d352be34a8d6ded
          createdAt: "2025-08-13T23:43:34Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0-alpha.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-hub@sha256:6be8fe8307e6686b178ace42f6db5cab8f04c5e2e29e02d2edcb5dc28ee337e9
        - quay.io/konveyor/tackle2-addon-platform@sha256:12efa5b7fbf6d829879b19e2820dcf782e648059f51544ebd25251dfc67fc7ca
        - quay.io/konveyor/postgresql-15-c9s@sha256:d75c5a1a0fa686ccf701e5e55633aa73d0cbb2cd9b5d8d4d4e80d7c46d78c121
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-operator@sha256:c1d1942b3bb3c4e99d9b5484493e27a432455608fee4548f1d352be34a8d6ded
        - quay.io/konveyor/tackle2-addon-discovery@sha256:ba865a74287ff94375ed316f348ee2c1d962d8d350c8afae5dbb0aee9b64fdea
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:7a5750fb76935e55b7b59e8363f21f5f9b48b39492d7f3e58e2e6a84d07ab027
        - quay.io/konveyor/kai-solution-server@sha256:83987d12d8290e085b6b66de56d347a8bf262815a5f99796177cb9120513ee3a
        - quay.io/konveyor/java-external-provider@sha256:42a156f1d81352f5e09fa0ffe2b51a4d93a1f476755eb0bac9a1d24bd3a89631
        - quay.io/konveyor/kantra@sha256:06ae932d9c2f51035bc5a488dd61e8a31c344e72615b013419199168701ea3fb
        - quay.io/konveyor/tackle2-ui@sha256:867bf24dd1687a79f2a911c22c1059ca7bd025a307d290290265653a696b8e98
        - quay.io/konveyor/tackle-keycloak-init@sha256:07c2ca07922d273bd2a3d9156ad8af0e5bf973aeb316496ee65d7201c381c1b1
        - quay.io/konveyor/origin-oauth-proxy@sha256:2c811da843057119b246784f21dc7eec490669a8569e8295d8e914cdce555b24
        - quay.io/konveyor/generic-external-provider@sha256:d58e867b51d5e76cd3597cfb7f2478d221b966f84509dbdd44c609ff5d9169be
        version: 0.8.0-alpha.4
      entries:
      - name: konveyor-operator.v0.8.0-alpha.4
        version: 0.8.0-alpha.4
      - name: konveyor-operator.v0.8.0-alpha.2
        version: 0.8.0-alpha.2
      - name: konveyor-operator.v0.8.0-alpha.1
        version: 0.8.0-alpha.1
      - name: konveyor-operator.v0.7.0-alpha.2
        version: 0.7.0-alpha.2
      - name: konveyor-operator.v0.7.0-alpha.1
        version: 0.7.0-alpha.1
      - name: konveyor-operator.v0.6.0-alpha.2
        version: 0.6.0-alpha.2
      - name: konveyor-operator.v0.6.0-alpha.1
        version: 0.6.0-alpha.1
      - name: konveyor-operator.v0.5.0-alpha.1
        version: 0.5.0-alpha.1
      - name: konveyor-operator.v0.4.0-alpha.1
        version: 0.4.0-alpha.1
      - name: konveyor-operator.v0.3.0-alpha.6
        version: 0.3.0-alpha.6
      - name: konveyor-operator.v0.3.0-alpha.5
        version: 0.3.0-alpha.5
      - name: konveyor-operator.v0.3.0-alpha.4
        version: 0.3.0-alpha.4
      - name: konveyor-operator.v0.3.0-alpha.3
        version: 0.3.0-alpha.3
      - name: konveyor-operator.v0.3.0-alpha.2
        version: 0.3.0-alpha.2
      - name: konveyor-operator.v0.3.0-alpha.1
        version: 0.3.0-alpha.1
      - name: konveyor-operator.v0.2.0-alpha.3
        version: 0.2.0-alpha.3
      - name: konveyor-operator.v0.2.0-alpha.2
        version: 0.2.0-alpha.2
      - name: konveyor-operator.v0.2.0-alpha.1
        version: 0.2.0-alpha.1
      name: alpha
    - currentCSV: konveyor-operator.v0.8.0-beta.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:019a79dec40445b6289948e173e5bf2ebfa9b495a7bb077abd634c0667953a54
          createdAt: "2025-09-17T19:50:24Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0-beta.5'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/kai-solution-server@sha256:1e69cf8a342090a03a9dddb1b28ae3627052b6d636ffa290a11761b94fd10648
        - quay.io/konveyor/tackle2-ui@sha256:78398832e092557639531230cf6d023bb07b589f3587afeaacf8c68640abc454
        - quay.io/konveyor/tackle2-operator@sha256:019a79dec40445b6289948e173e5bf2ebfa9b495a7bb077abd634c0667953a54
        - quay.io/konveyor/tackle-keycloak-init@sha256:0b4bfc1fea756f10a0657861c4c422303a2349d8d9b04a23238a32ceaf106ac5
        - quay.io/konveyor/origin-oauth-proxy@sha256:8988f14db7dde3edcaed0541cecb0b4fd58fab9dbaffb79a6c1c4aee13b9bd73
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/kantra@sha256:ad6f5c05e5e950e33a745accd03ad933fb7bb4907f44455f88a2b5c27e6dd15e
        - quay.io/konveyor/tackle2-addon-platform@sha256:d0280ade44ba5e15210c07342608576b38e6993af2e479740675a4706cd2c875
        - quay.io/konveyor/tackle2-addon-discovery@sha256:a29a0248e222d9ac3225e9d3df603f1383fa8844856bf68ea618d303f0b79883
        - quay.io/konveyor/java-external-provider@sha256:0cbdd91cc5057889b8daa2f7611fa0c944fab951c73bdf88a3e28795e9a5fc33
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:fed49c5ff14bb232e63d2dc180d3b530da63f11bf97055fc9d86cbb9e47104a4
        - quay.io/konveyor/generic-external-provider@sha256:a6c550bad7a09a1c51e226247f11c00d8453987698832cb566582842219f1f3c
        - quay.io/konveyor/tackle2-hub@sha256:1611913558c9f1203a844b43969d7b1539dab3578b83460e8383ee6757c3c34c
        - quay.io/konveyor/postgresql-15-c9s@sha256:157a2941ae971d8f72b2f8af1d4a29deeb5548012a978fa986c1c6e52395da4d
        version: 0.8.0-beta.5
      entries:
      - name: konveyor-operator.v0.8.0-beta.5
        version: 0.8.0-beta.5
      - name: konveyor-operator.v0.8.0-beta.4
        version: 0.8.0-beta.4
      - name: konveyor-operator.v0.8.0-beta.2
        version: 0.8.0-beta.2
      - name: konveyor-operator.v0.8.0-beta.1
        version: 0.8.0-beta.1
      - name: konveyor-operator.v0.6.0-beta.2
        version: 0.6.0-beta.2
      - name: konveyor-operator.v0.6.0-beta.1
        version: 0.6.0-beta.1
      - name: konveyor-operator.v0.5.0-beta.3
        version: 0.5.0-beta.3
      - name: konveyor-operator.v0.5.0-beta.2
        version: 0.5.0-beta.2
      - name: konveyor-operator.v0.5.0-beta.1
        version: 0.5.0-beta.1
      - name: konveyor-operator.v0.3.0-beta.3
        version: 0.3.0-beta.3
      - name: konveyor-operator.v0.3.0-beta.2
        version: 0.3.0-beta.2
      - name: konveyor-operator.v0.3.0-beta.1.2
        version: 0.3.0-beta.1.2
      - name: konveyor-operator.v0.3.0-beta.1.1
        version: 0.3.0-beta.1.1
      - name: konveyor-operator.v0.3.0-beta.1
        version: 0.3.0-beta.1
      name: beta
    - currentCSV: konveyor-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "move2kube-plan",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "image": "quay.io/konveyor/tackle-addon-move2kube-plan:latest"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:9688d0763d8c7812beee715a061c3c198006448855d7b891d45833f62cf869fb
          createdAt: "2023-05-25T15:23:45Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.1.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)

        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:9688d0763d8c7812beee715a061c3c198006448855d7b891d45833f62cf869fb
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/postgresql-12-centos7@sha256:1287d947464fbf5275feb31d106d2feb856c6117f0eaec8d947f2e03bd8618c5
        - quay.io/konveyor/tackle2-addon@sha256:35be344bbd3393f5634f52aafd21e9c3964d1e0ba5011ecad1561600d476bbe9
        - quay.io/konveyor/tackle-keycloak-init@sha256:b0347e5a608352c4b7bb796b65a67d417e98ba4896db71a487fa26243214cfcf
        - quay.io/konveyor/tackle2-hub@sha256:3d15d158ffb6f1fc4ca5aaf29a12173fe344c289773902c9161056bdf69acc92
        - quay.io/konveyor/tackle2-ui@sha256:99650804e65e2b6cb15a37c00189d7f463e4d19a0f866a79e9c09c8420a4d028
        - quay.io/konveyor/tackle2-addon-windup@sha256:81f0625cb82eb24ded51c7258f229f57778820c62e3be6ec05fb2b12d2970825
        - quay.io/konveyor/tackle-pathfinder@sha256:f9c5aa49f1d27f477cdd82c21441ae0287acdcdc9b2b3fcf197884007dc5f7dd
        version: 0.1.2
      entries:
      - name: konveyor-operator.v0.1.2
        version: 0.1.2
      - name: konveyor-operator.v0.1.1
        version: 0.1.1
      - name: konveyor-operator.v0.1.0
        version: 0.1.0
      name: konveyor-0.1
    - currentCSV: konveyor-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "move2kube-plan",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "image": "quay.io/konveyor/tackle-addon-move2kube-plan:latest"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:1860bb099ac64d9a2af27692f0671b7240e4d19746fffa9f9ae2356ff2818e10
          createdAt: "2023-07-06T06:17:40Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.2.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)

        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:1860bb099ac64d9a2af27692f0671b7240e4d19746fffa9f9ae2356ff2818e10
        - quay.io/konveyor/postgresql-12-centos7@sha256:6016530ab1fce71d29421ce9a97b3a89fbbb126b87b1b079c6aadebb7ff5664c
        - quay.io/konveyor/tackle2-hub@sha256:d9c4f21f44098a2526474968b6d88094dca8b5eaeb9042716332c9cb67df47cd
        - quay.io/konveyor/tackle2-ui@sha256:d10c1e717d160c5eb2581649a8dd7cdf4922a74d92db014f72f42a48a3134341
        - quay.io/konveyor/tackle2-addon-windup@sha256:538dfdb134d53a876d126dbd16679638584a606ef2b9e8fbe69ed1f29733c03a
        - quay.io/konveyor/tackle-keycloak-init@sha256:ca6c85bdf8417d4be116c2098b8d8ad5ae8b7ab0b636c4cbe4a392aa424e2d47
        - quay.io/konveyor/tackle2-addon@sha256:11693e16363bcd98249209d63a29f0c6cfffab3a4071649cbeea2fd276bc2440
        - quay.io/konveyor/tackle-pathfinder@sha256:190269b061b2065ca8ba9357bf9b959a2f587a547a44769ae0afd1ac86b4b197
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        version: 0.2.1
      entries:
      - name: konveyor-operator.v0.2.1
        version: 0.2.1
      - name: konveyor-operator.v0.2.0
        version: 0.2.0
      name: konveyor-0.2
    - currentCSV: konveyor-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:cd2305373f50ce927393e43f5735915e46c32d6f8a76091ee2808218885bba72
          createdAt: "2024-05-02T22:12:42Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.3.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:cd2305373f50ce927393e43f5735915e46c32d6f8a76091ee2808218885bba72
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:51bda0c6577c15684583d607a4f0cebef760814102423f4b2025016093a19fc1
        - quay.io/konveyor/origin-oauth-proxy@sha256:05f6bd1216b921b1f47a3821f0e22b84ed6b141bb2ada474c12bd9fde613108c
        - quay.io/konveyor/tackle2-hub@sha256:bf8803721d06d1a39251e22892343a013d5ef0fb4c1883c95e1eb55aa683f486
        - quay.io/konveyor/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle-keycloak-init@sha256:a63702cc0cb36830a7830ed70ca344a9aacc201ba6f3c94806ca2e93160df4dd
        - quay.io/konveyor/tackle2-ui@sha256:bbf58d2ea613489f91efae450421afafb9052416b0d7971606e2729e89fd7de6
        version: 0.3.2
      entries:
      - name: konveyor-operator.v0.3.2
        version: 0.3.2
      - name: konveyor-operator.v0.3.1
        version: 0.3.1
      - name: konveyor-operator.v0.3.0
        version: 0.3.0
      name: konveyor-0.3
    - currentCSV: konveyor-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:496db8ca859608d04008ac36cf7bdc9cafd7214294d33a9c5646083c361ffcb6
          createdAt: "2024-06-05T03:08:51Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.4.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:496db8ca859608d04008ac36cf7bdc9cafd7214294d33a9c5646083c361ffcb6
        - quay.io/konveyor/tackle-keycloak-init@sha256:4c7e07b1fce05f7dca47edecc0ba0ce373cc38e48879cc7eed72510d2865a18e
        - quay.io/konveyor/tackle2-ui@sha256:cb1255fc5fa4608fc659c3455f3164664b960cc2d9a3b2208967798e3532df78
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:e84ea36d206a8c143794e3986940377e77c7618225cad594bdf9a4047d7823dc
        - quay.io/konveyor/origin-oauth-proxy@sha256:05f6bd1216b921b1f47a3821f0e22b84ed6b141bb2ada474c12bd9fde613108c
        - quay.io/konveyor/tackle2-hub@sha256:b87acd8e2ee4a7edc4369bc375f64e2b98bc479d09556439e2ee4d74b74970a9
        - quay.io/konveyor/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        version: 0.4.0
      entries:
      - name: konveyor-operator.v0.4.0
        version: 0.4.0
      name: konveyor-0.4
    - currentCSV: konveyor-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:474d739ed470701b628f1f9e971925ac7fbe0b543f3e36305d932b963b1225fd
          createdAt: "2024-08-27T15:48:02Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.5.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-hub@sha256:faf91ca68e24979642a44fa013b1fb9edfee415eee835ccef9dce2eb418e2909
        - quay.io/konveyor/java-external-provider@sha256:33fb5d24b998f01c481a0a03e3f1bf3f7a6c93f5b662885102a0c966026e3375
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle-keycloak-init@sha256:5f652be5914ae6e23537b9fe02ccc9cd638c08de5462d9f58135a405c81f5287
        - quay.io/konveyor/tackle2-ui@sha256:36f29fe01d3078f0e100b24d1ed126850c5a773f4b0ae922fd4c4487de1ac29d
        - quay.io/konveyor/generic-external-provider@sha256:334a631538d31efc8d82adae4f52303336800fa3037b5055c72400918d81955c
        - quay.io/konveyor/postgresql-15-c9s@sha256:039ce764cadd03b4cfea954700139b8ceff415a9a0970baa7f2cf60f7f929174
        - quay.io/konveyor/tackle2-operator@sha256:474d739ed470701b628f1f9e971925ac7fbe0b543f3e36305d932b963b1225fd
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:2afbe8bea09904e084e026cd0a294fbe1c7b05d0e387f6633b09c5eaf4665e02
        - quay.io/konveyor/tackle2-addon-discovery@sha256:461f960bb1d29e719fb472fe9be9cf3deca17d984687957f54546a09047b0d3c
        version: 0.5.3
      entries:
      - name: konveyor-operator.v0.5.3
        version: 0.5.3
      - name: konveyor-operator.v0.5.1
        version: 0.5.1
      - name: konveyor-operator.v0.5.0
        version: 0.5.0
      name: konveyor-0.5
    - currentCSV: konveyor-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:8ee143c473570efd27b6f93332dc272871e11c66cc20f381c3a1e6e5688abf68
          createdAt: "2025-03-05T21:14:52Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.6.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle-keycloak-init@sha256:a3741269aa15edd2f9fb7df853a03a942d16a88eec114f3e54f82ef63b82adc4
        - quay.io/konveyor/tackle2-ui@sha256:12b0eda13a63508b9c13d4f9309bbfbe19d564627cf0de672307af1a6ec993fb
        - quay.io/konveyor/tackle2-operator@sha256:8ee143c473570efd27b6f93332dc272871e11c66cc20f381c3a1e6e5688abf68
        - quay.io/konveyor/java-external-provider@sha256:15148cde9a6fedef2207fe36f0ac3bd1ecdc6e3ac76f1cbec9eb97ffcf247231
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle2-hub@sha256:52a8b9be038fbfbc4e22172a4a8c7191d2b06fd8670e5df2ec442e3ce00d8c8a
        - quay.io/konveyor/tackle2-addon-discovery@sha256:3eb6f395c451e2c9a867871b3984a25f9f8e6ddcfe388f37fbd6d46a61a70b1e
        - quay.io/konveyor/generic-external-provider@sha256:27c1731859a4a639d2750cff8d1246281f4b0466bb5b069169c3e170c3c19b30
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:0b5cbb322fbe96335522410376c2e8505a8302088c0e454f3e5e7e28ba7c96a6
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/postgresql-15-c9s@sha256:55f8014c1edafab8f9085228a0d207c4e98820c615be34b935ce98e942d388b1
        version: 0.6.1
      entries:
      - name: konveyor-operator.v0.6.1
        version: 0.6.1
      - name: konveyor-operator.v0.6.0
        version: 0.6.0
      name: konveyor-0.6
    - currentCSV: konveyor-operator.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:f5870bc26b2a07ddacf00ed5b5f7df7d1068489010a80bcc543686c6c9576fbb
          createdAt: "2025-07-10T15:16:17Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:f5870bc26b2a07ddacf00ed5b5f7df7d1068489010a80bcc543686c6c9576fbb
        - quay.io/konveyor/tackle-keycloak-init@sha256:dab4cb1229c8a729d14b4cecb4aa3b5c4d48e94c36e84d785c922fc08d120617
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:7ed52ea9b95650241b5c2c01914775e29be23f7ed75fabb6e7c71969f3f0728f
        - quay.io/konveyor/generic-external-provider@sha256:facca8861559a658710ad3aef8b325731595bf944cce1907715d6efd33887072
        - quay.io/konveyor/tackle2-hub@sha256:52f66607f3134a4165ce691c6e9ce992d93927e9591055daafdc1bfc1dd09dd9
        - quay.io/konveyor/tackle2-ui@sha256:f87b04762fd163f40c49f3bcffd473f0c3f80a643576c6a3e2c2b4a14395c09f
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/postgresql-15-c9s@sha256:41460103a0bab189cd77acb6e741d9c5a6d3506479f7e56df40eb156d9b2cadd
        - quay.io/konveyor/tackle2-addon-discovery@sha256:60951801aa539ee59cd43f3c5f42a2ca751ca70e995cbb258b7d55ebb24d8450
        - quay.io/konveyor/java-external-provider@sha256:a0ef733e9fce3845d02b5360440add85427acad697d2157ef92242dec06bd73b
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/origin-oauth-proxy@sha256:2c811da843057119b246784f21dc7eec490669a8569e8295d8e914cdce555b24
        version: 0.7.1
      entries:
      - name: konveyor-operator.v0.7.1
        version: 0.7.1
      - name: konveyor-operator.v0.7.0
        version: 0.7.0
      name: konveyor-0.7
    - currentCSV: konveyor-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:68b6672fccdf54c65cc83f4c82db023053d6b61409bea880f58b53ccfea7b1ed
          createdAt: "2025-09-29T20:41:09Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-addon-platform@sha256:d0280ade44ba5e15210c07342608576b38e6993af2e479740675a4706cd2c875
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:8c82ad880ca2a2d5cd4b08125feca16d13d014d15f4f77677306086d1067f9aa
        - quay.io/konveyor/generic-external-provider@sha256:c7b6618bfff81d73f1b455963107ffdaf19334e6552ce076c009476da3f37f21
        - quay.io/konveyor/postgresql-15-c9s@sha256:af275f35a307367dcee9a339a6903cf79e3a8a41cabf79fb7cea7015895f0b58
        - quay.io/konveyor/origin-oauth-proxy@sha256:8988f14db7dde3edcaed0541cecb0b4fd58fab9dbaffb79a6c1c4aee13b9bd73
        - quay.io/konveyor/kai-solution-server@sha256:bab6e224c50de5b44cc37ce0297a3b91255efc08c1ad0fccf14fef9e5182775f
        - quay.io/konveyor/tackle2-operator@sha256:68b6672fccdf54c65cc83f4c82db023053d6b61409bea880f58b53ccfea7b1ed
        - quay.io/konveyor/tackle-keycloak-init@sha256:c0051b8c3a3c891fe12012623aee775080ecb47af8a4e75b8be8b0b4855d9f2c
        - quay.io/konveyor/tackle2-ui@sha256:f3f7380be65ff1fb3459c37a3d4865354bab685dc836893f78221f16ba579f70
        - quay.io/konveyor/kantra@sha256:a5b5591a51143574a6ce6678b141b9cd3e14328c6fad773412dde21b442fe364
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-hub@sha256:b3e6db1ffc195143b8e5d54b60949256e154b0f8b667736512a084e015cdf2a7
        - quay.io/konveyor/tackle2-addon-discovery@sha256:96a5b37834fbed9506014bcbd909258081fd833ef9e16daa104effd0e3a3c6ea
        - quay.io/konveyor/java-external-provider@sha256:8eab08a04e98051f89352250d63523a5b591f35a6497f5d69a5a9d556a94abf9
        version: 0.8.0
      entries:
      - name: konveyor-operator.v0.8.0
        version: 0.8.0
      name: konveyor-0.8
    - currentCSV: konveyor-operator.v0.7.0-rc.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
          createdAt: "2025-05-21T17:12:48Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.0-rc.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-ui@sha256:afe9141366a23c072286517e9af5436d14fd1df775aaa2fa13c3911764515777
        - quay.io/konveyor/postgresql-15-c9s@sha256:012cd6611aa9d3f4e0a82636d6817f381c19e3b0809199216fbe7fa3396a2ef5
        - quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/generic-external-provider@sha256:de55f78939576c110eb6f084a7ca51338fedba699fbb5135f23392783a5d7b95
        - quay.io/konveyor/tackle2-addon-discovery@sha256:89d2b0f8636897eb4dc1957a74a5b5041c0f3267a747adcf2bc6ca56b17b9937
        - quay.io/konveyor/origin-oauth-proxy@sha256:6c2847fd9667938b1d1294d988038e885e43cd9fb378862dae3419ce4f21631f
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle-keycloak-init@sha256:c3124971dc31be9201de2a6a1f6b2f886adf4c96731188dd5e750798cde3cfe7
        - quay.io/konveyor/tackle2-hub@sha256:bed834eb5488eef23b0d1e8af194b0646cf74c39c197e0d57acc5aed2299d9f5
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:46150725a2adbde8dfdc2250614d437a50e64bbf0b4562dabc73715a287de67b
        - quay.io/konveyor/java-external-provider@sha256:61ab20e37916abfeeadde685336660f0e8bbf519403d1f7100d43c1e42a72fbd
        version: 0.7.0-rc.1
      entries:
      - name: konveyor-operator.v0.7.0-rc.1
        version: 0.7.0-rc.1
      - name: konveyor-operator.v0.6.0-rc.2
        version: 0.6.0-rc.2
      - name: konveyor-operator.v0.6.0-rc.1
        version: 0.6.0-rc.1
      - name: konveyor-operator.v0.5.0-rc.1
        version: 0.5.0-rc.1
      - name: konveyor-operator.v0.3.0-rc.4
        version: 0.3.0-rc.4
      - name: konveyor-operator.v0.3.0-rc.3
        version: 0.3.0-rc.3
      - name: konveyor-operator.v0.3.0-rc.2
        version: 0.3.0-rc.2
      - name: konveyor-operator.v0.3.0-rc.1
        version: 0.3.0-rc.1
      name: rc
    - currentCSV: konveyor-operator.v0.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:33cf01f19a7043bcc5bdb12a3ed6bde7a9094981aa2383554111ebec9b4ba498
          createdAt: "2025-03-26T20:25:41Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.6.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/generic-external-provider@sha256:bc1112352f1a2c3172c5e7d5e637b5c5a69bf7d98b78f494fa593e8872e5cfd1
        - quay.io/konveyor/tackle2-ui@sha256:3c568c6047a044095080b6013b643a4041deae7b45936a443d4c4d708a81b945
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:734752be111a97e4b7e0030d563c32883f419ad031b2f350cb8e7333f2f37de3
        - quay.io/konveyor/tackle2-hub@sha256:0df49087d0c9341c2120a7c3aa2ba7907c056f94e1602ffd0496ca7113c75382
        - quay.io/konveyor/postgresql-15-c9s@sha256:7957d3b272b9f4e469bdc0c514ec9e70e8fbdc34a7250b4f4c62f9bb73cbf08d
        - quay.io/konveyor/tackle2-addon-discovery@sha256:1b89266d6bdfa497ba8137e0d33ff53785ac476dd96bf4fd9a185ef0ddf513ce
        - quay.io/konveyor/java-external-provider@sha256:fe85689b57125d678df685012b93c8a567fcb9251777cd711e9b82bdb22a734c
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/tackle2-operator@sha256:33cf01f19a7043bcc5bdb12a3ed6bde7a9094981aa2383554111ebec9b4ba498
        - quay.io/konveyor/tackle-keycloak-init@sha256:e32fa48351f410361e8bc100303deeb40d8a66d7eefb0cf7b1bc16e067a9a795
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        version: 0.6.2
      entries:
      - name: konveyor-operator.v0.6.2
        version: 0.6.2
      name: release-0.6
    - currentCSV: konveyor-operator.v0.7.0-rc.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
          createdAt: "2025-05-21T17:12:48Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.0-rc.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/generic-external-provider@sha256:de55f78939576c110eb6f084a7ca51338fedba699fbb5135f23392783a5d7b95
        - quay.io/konveyor/origin-oauth-proxy@sha256:6c2847fd9667938b1d1294d988038e885e43cd9fb378862dae3419ce4f21631f
        - quay.io/konveyor/tackle2-addon-discovery@sha256:89d2b0f8636897eb4dc1957a74a5b5041c0f3267a747adcf2bc6ca56b17b9937
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-hub@sha256:bed834eb5488eef23b0d1e8af194b0646cf74c39c197e0d57acc5aed2299d9f5
        - quay.io/konveyor/postgresql-15-c9s@sha256:012cd6611aa9d3f4e0a82636d6817f381c19e3b0809199216fbe7fa3396a2ef5
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:46150725a2adbde8dfdc2250614d437a50e64bbf0b4562dabc73715a287de67b
        - quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
        - quay.io/konveyor/tackle-keycloak-init@sha256:c3124971dc31be9201de2a6a1f6b2f886adf4c96731188dd5e750798cde3cfe7
        - quay.io/konveyor/tackle2-ui@sha256:afe9141366a23c072286517e9af5436d14fd1df775aaa2fa13c3911764515777
        - quay.io/konveyor/java-external-provider@sha256:61ab20e37916abfeeadde685336660f0e8bbf519403d1f7100d43c1e42a72fbd
        version: 0.7.0-rc.1
      entries:
      - name: konveyor-operator.v0.7.0-rc.1
        version: 0.7.0-rc.1
      name: release-0.7
    defaultChannel: konveyor-0.8
    packageName: konveyor-operator
    provider:
      name: Konveyor
      url: https://www.konveyor.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: sriov-network-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sriov-network-operator.v4.13.0-202309190544
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovIBNetwork",
                "metadata": {
                  "name": "example-sriovibnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "linkState": "auto",
                  "networkNamespace": "default",
                  "resourceName": "mellanoxnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetwork",
                "metadata": {
                  "name": "example-sriovnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "intelnics",
                  "spoofChk": "on",
                  "trust": "off",
                  "vlan": 0
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodePolicy",
                "metadata": {
                  "name": "policy-1",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "deviceType": "vfio-pci",
                  "mtu": 9000,
                  "nicSelector": {
                    "pfNames": [
                      "eth1"
                    ],
                    "rootDevices": [
                      "0000:01:00.1"
                    ],
                    "vendor": "8086"
                  },
                  "nodeSelector": {
                    "feature.node.kubernetes.io/network-sriov.capable": "true"
                  },
                  "numVfs": 6,
                  "priority": 99,
                  "resourceName": "intelnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodeState",
                "metadata": {
                  "name": "minikube",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "interfaces": []
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkPoolConfig",
                "metadata": {
                  "name": "sriovnetworkpoolconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovOperatorConfig",
                "metadata": {
                  "name": "default",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "enableInjector": true,
                  "enableOperatorWebhook": true,
                  "logLevel": 2
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: quay.io/pliurh/sriov-network-operator-manager:bd1
          createdAt: "2023-08-02T09:22:34Z"
          description: An operator for configuring SR-IOV components and initializing
            SRIOV network devices in Openshift cluster.
          olm.skipRange: '>=4.3.0-0 <4.13.0-202309190544'
          operatorframework.io/suggested-namespace: openshift-sriov-network-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "cni"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/k8snetworkplumbingwg/sriov-network-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovIBNetwork is the Schema for the sriovibnetworks API
            displayName: Sriov IBNetwork
            kind: SriovIBNetwork
            name: sriovibnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodePolicy is the Schema for the sriovnetworknodepolicies
              API
            displayName: Sriov Network Node Policy
            kind: SriovNetworkNodePolicy
            name: sriovnetworknodepolicies.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodeState is the Schema for the sriovnetworknodestates
              API
            displayName: Sriov Network Node State
            kind: SriovNetworkNodeState
            name: sriovnetworknodestates.sriovnetwork.openshift.io
            version: v1
          - kind: SriovNetworkPoolConfig
            name: sriovnetworkpoolconfigs.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetwork is the Schema for the sriovnetworks API
            displayName: Sriov Network
            kind: SriovNetwork
            name: sriovnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovOperatorConfig is the Schema for the sriovoperatorconfigs
              API
            displayName: Sriov Operator Config
            kind: SriovOperatorConfig
            name: sriovoperatorconfigs.sriovnetwork.openshift.io
            version: v1
        description: |
          # SR-IOV Network Operator for Openshift

          ## Introduction
          The sriov-network-operator is generally responsible for configuring the sriov components in a openshift cluster.

          ### Supported Features
          * Initialize the SR-IOV NICs on nodes.
          * provision SR-IOV device plugin on selected node.
          * provision SR-IOV CNI plugin on selected nodes.
          * manage configuration of SR-IOV device plugin.
          * generate net-att-def CRs for SR-IOV CNI plugin.

          This operator has to run in namespace 'openshift-sriov-network-operator'. An Operator Group is also required to install this operator:

          ```
          $ oc create namespace openshift-sriov-network-operator

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: sriov-network-operators
            namespace: openshift-sriov-network-operator
          spec:
            targetNamespaces:
            - openshift-sriov-network-operator
          EOF
          ```
        displayName: SR-IOV Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - network
        - sr-iov
        - sriov
        links:
        - name: Source Code
          url: https://github.com/k8snetworkplumbingwg/sriov-network-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-sriov-network-config-daemon@sha256:a2053c5366eb89112dfbc7a9994685fc198cce6fa87fdd6d1c0a07727e6eee03
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin@sha256:82d4846f8dfb9e0d02e7ff3f532f8f2bf44c9aba63c37d5ce2efb36969e5c453
        - registry.redhat.io/openshift4/ose-sriov-network-operator@sha256:3314ee580113b985e275583585b1f2adf381c5a0885b51100b3a01872a77a555
        - registry.redhat.io/openshift4/ose-sriov-network-webhook@sha256:bfb84d00dbb6b18d596ca2720b52d922c589249d2f988679189ea0eea985e58a
        - registry.redhat.io/openshift4/sriov-cni-rhel9@sha256:a4f0ec210cd065cb7f77dd25257aadf8345f47819cc728c5b6229acb79a09bb3
        - registry.redhat.io/openshift4/ose-sriov-dp-admission-controller@sha256:8dfaa07a5775a7156d303019879e41b17833e5643794f6e1087881d2a4dd1781
        - registry.redhat.io/openshift4/ose-sriov-infiniband-cni@sha256:716aa0b9ecfddfea0a62615f424f4abeab5362899c695e9ecc9df4df05f1add5
        version: 4.13.0-202309190544
      entries:
      - name: sriov-network-operator.v4.13.0-202309190544
        version: 4.13.0-202309190544
      name: "4.13"
    - currentCSV: sriov-network-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovIBNetwork",
                "metadata": {
                  "name": "example-sriovibnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "linkState": "auto",
                  "networkNamespace": "default",
                  "resourceName": "mellanoxnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetwork",
                "metadata": {
                  "name": "example-sriovnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "intelnics",
                  "spoofChk": "on",
                  "trust": "off",
                  "vlan": 0
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodePolicy",
                "metadata": {
                  "name": "policy-1",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "deviceType": "vfio-pci",
                  "mtu": 9000,
                  "nicSelector": {
                    "pfNames": [
                      "eth1"
                    ],
                    "rootDevices": [
                      "0000:01:00.1"
                    ],
                    "vendor": "8086"
                  },
                  "nodeSelector": {
                    "feature.node.kubernetes.io/network-sriov.capable": "true"
                  },
                  "numVfs": 6,
                  "priority": 99,
                  "resourceName": "intelnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodeState",
                "metadata": {
                  "name": "minikube",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "interfaces": []
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkPoolConfig",
                "metadata": {
                  "name": "sriovnetworkpoolconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovOperatorConfig",
                "metadata": {
                  "name": "default",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "enableInjector": true,
                  "enableOperatorWebhook": true,
                  "logLevel": 2
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-sriov-network-rhel9-operator@sha256:8dcae2a77e904713fba9c5907789f9b22ffb1f941f793adc4b604798d29a44f2
          createdAt: "2024-05-07T23:50:52Z"
          description: An operator for configuring SR-IOV components and initializing
            SRIOV network devices in Openshift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.16.0-202509240123'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "sriovnetwork.openshift.io/v1",
              "kind": "SriovOperatorConfig",
              "metadata": {
                "name": "default"
              },
              "spec": {
                "configurationMode": "daemon",
                "disableDrain": false,
                "enableInjector": true,
                "enableOperatorWebhook": true,
                "logLevel": 2
              }
            }
          operatorframework.io/suggested-namespace: openshift-sriov-network-operator
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-sriov-network-operator",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["sriovnetworknodestates.sriovnetwork.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/k8snetworkplumbingwg/sriov-network-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovIBNetwork is the Schema for the sriovibnetworks API
            displayName: Sriov IBNetwork
            kind: SriovIBNetwork
            name: sriovibnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodePolicy is the Schema for the sriovnetworknodepolicies
              API
            displayName: Sriov Network Node Policy
            kind: SriovNetworkNodePolicy
            name: sriovnetworknodepolicies.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodeState is the Schema for the sriovnetworknodestates
              API
            displayName: Sriov Network Node State
            kind: SriovNetworkNodeState
            name: sriovnetworknodestates.sriovnetwork.openshift.io
            version: v1
          - kind: SriovNetworkPoolConfig
            name: sriovnetworkpoolconfigs.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetwork is the Schema for the sriovnetworks API
            displayName: Sriov Network
            kind: SriovNetwork
            name: sriovnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovOperatorConfig is the Schema for the sriovoperatorconfigs
              API
            displayName: Sriov Operator Config
            kind: SriovOperatorConfig
            name: sriovoperatorconfigs.sriovnetwork.openshift.io
            version: v1
        description: |
          # SR-IOV Network Operator for Openshift

          ## Introduction
          The sriov-network-operator is generally responsible for configuring the sriov components in a openshift cluster.

          ### Supported Features
          * Initialize the SR-IOV NICs on nodes.
          * provision SR-IOV device plugin on selected node.
          * provision SR-IOV CNI plugin on selected nodes.
          * manage configuration of SR-IOV device plugin.
          * generate net-att-def CRs for SR-IOV CNI plugin.

          This operator has to run in namespace 'openshift-sriov-network-operator'. An Operator Group is also required to install this operator:

          ```
          $ oc create namespace openshift-sriov-network-operator

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: sriov-network-operators
            namespace: openshift-sriov-network-operator
          spec:
            targetNamespaces:
            - openshift-sriov-network-operator
          EOF
          ```
        displayName: SR-IOV Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - network
        - sr-iov
        - sriov
        links:
        - name: Source Code
          url: https://github.com/k8snetworkplumbingwg/sriov-network-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-sriov-network-config-daemon-rhel9@sha256:61afbb271a08dc3ab3b5509df535d9456f9b2c1b267554386b4bf859b6bccbe4
        - registry.redhat.io/openshift4/ose-sriov-infiniband-cni-rhel9@sha256:f828e415db73bf2df3fa7b26238a2f5c8a53cda6b396c11a0337d688108e6647
        - registry.redhat.io/openshift4/sriov-cni-rhel9@sha256:cac67ac25a0d319b5ea68cb3f63eac058339f1981811c7a3466b9daa2534a38d
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin-rhel9@sha256:5b3bc02e31eecd0d3468db729b03239d518ef73112b0fdb5ec3f05a788baa0be
        - registry.redhat.io/openshift4/ose-sriov-network-rhel9-operator@sha256:8dcae2a77e904713fba9c5907789f9b22ffb1f941f793adc4b604798d29a44f2
        - registry.redhat.io/openshift4/ose-sriov-network-webhook-rhel9@sha256:a6a5cc76d88b6b7aa86ee24b882ce43be33c8c6d4bcfae5f1b5490ead71ce8b6
        - registry.redhat.io/openshift4/ose-sriov-network-metrics-exporter-rhel9@sha256:a0580d8156de0ca1e5a119441196e9babbeb4d7edc502659c7ac66a49b1928d5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-sriov-dp-admission-controller-rhel9@sha256:20fd8a0081ffdea05796c8f6a0539a0e246da24f739f7bbf0722c90acfdeed15
        version: 4.16.0-202509240123
      entries:
      - name: sriov-network-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: sriov-network-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: sriov-network-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: sriov-network-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: sriov-network-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: sriov-network-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: sriov-network-operator.v4.16.0-202504290837
        version: 4.16.0-202504290837
      - name: sriov-network-operator.v4.16.0-202504151638
        version: 4.16.0-202504151638
      - name: sriov-network-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: sriov-network-operator.v4.16.0-202502262005
        version: 4.16.0-202502262005
      - name: sriov-network-operator.v4.16.0-202502171404
        version: 4.16.0-202502171404
      - name: sriov-network-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: sriov-network-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: sriov-network-operator.v4.16.0-202501081506
        version: 4.16.0-202501081506
      - name: sriov-network-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: sriov-network-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: sriov-network-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: sriov-network-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: sriov-network-operator.v4.16.0-202410300306
        version: 4.16.0-202410300306
      - name: sriov-network-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: sriov-network-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: sriov-network-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: sriov-network-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: sriov-network-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: sriov-network-operator.v4.16.0-202409111535
        version: 4.16.0-202409111535
      - name: sriov-network-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: sriov-network-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: sriov-network-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: sriov-network-operator.v4.16.0-202408120337
        version: 4.16.0-202408120337
      - name: sriov-network-operator.v4.16.0-202407100107
        version: 4.16.0-202407100107
      - name: sriov-network-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: sriov-network-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: sriov-network-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: sriov-network-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: sriov-network-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: The cert-manager maintainers
      provider-url: https://cert-manager.io/
    name: cert-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-manager.v1.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "example-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "example-clusterissuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example-certificate"
                },
                "spec": {
                  "dnsNames": [
                    "example.com"
                  ],
                  "issuerRef": {
                    "name": "example-issuer"
                  },
                  "secretName": "example-certificate-tls"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "name": "example-certificaterequest"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "name": "example-order"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "example-challenge"
                },
                "spec": {}
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: quay.io/jetstack/cert-manager-controller:v1.16.5
          createdAt: 2025-05-02T09:11:31
          olm.skipRange: '>=1.16.0 <1.16.5'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/internal-objects: |-
            [
              "challenges.acme.cert-manager.io",
              "orders.acme.cert-manager.io"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/cert-manager/cert-manager
          support: The cert-manager maintainers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A CertificateRequest is used to request a signed certificate from one of the
              configured issuers.

              All fields within the CertificateRequest's `spec` are immutable after creation.
              A CertificateRequest will either succeed or fail, as denoted by its `Ready` status
              condition and its `status.failureTime` field.

              A CertificateRequest is a one-shot resource, meaning it represents a single
              point in time request for a certificate and cannot be re-used.
            displayName: CertificateRequest
            kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - description: |-
              A Certificate resource should be created to ensure an up to date and signed
              X.509 certificate is stored in the Kubernetes Secret resource named in `spec.secretName`.

              The stored certificate will be renewed before it expires (as configured by `spec.renewBefore`).
            displayName: Certificate
            kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: Challenge is a type to represent a Challenge request with
              an ACME server
            displayName: (Internal) Challenge
            kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - description: |-
              A ClusterIssuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is similar to an Issuer, however it is cluster-scoped and therefore can
              be referenced by resources that exist in *any* namespace, not just the same
              namespace as the referent.
            displayName: ClusterIssuer
            kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - description: |-
              An Issuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is scoped to a single namespace and can therefore only be referenced by
              resources within the same namespace.
            displayName: Issuer
            kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - description: Order is a type to represent an Order with an ACME server
            displayName: (Internal) Order
            kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          cert-manager is a Kubernetes add-on to automate the management and issuance of TLS certificates from various issuing sources.
          It will ensure certificates are valid and up to date periodically, and attempt to renew certificates at an appropriate time before expiry.

          cert-manager has the following [built in issuers][]:

          * ACME: sign certificates using Let's Encrypt (or any other [ACME][] compliant signer).
          * Vault: sign certificates using HashiCorp Vault.
          * Venafi: sign certificates using Venafi Cloud or Venafi Trust Protection Platform.
          * SelfSigned: the private key of the certificate will be used to sign the certificate.
          * CA: a private key stored as a Kubernetes Secret will be used to sign the certificate.

          [built in issuers]: https://cert-manager.io/docs/configuration/#supported-issuer-types
          [ACME]: https://datatracker.ietf.org/doc/html/rfc8555

          cert-manager also supports [external issuers][],
          which are out-of-tree components that provide support for other certificate authorities, such as:
          [Google Cloud Certificate Authority](https://github.com/jetstack/google-cas-issuer),
          [AWS Private Certificate Authority](https://github.com/cert-manager/aws-privateca-issuer), and
          [Cloudflare Origin CA](https://github.com/cloudflare/origin-ca-issuer)

          [external issuers]: https://cert-manager.io/docs/configuration/external/

          In order to configure cert-manager to issue certificates, you must first create an Issuer or ClusterIssuer.
          These resources represent a particular signing authority and tell cert-manager how to handle your certificate requests.
          The [Issuer Documentation][] has more detail.

          [configure cert-manager]: https://cert-manager.io/docs/configuration/
          [Issuer Documentation]: https://cert-manager.io/docs/concepts/issuer/

          Once an Issuer has been configured, you can create a Certificate resource,
          which is a definition of a certificate request that is to be honored by an issuer
          and which is to be kept up-to-date.
          This is the usual way that you will interact with cert-manager to request signed certificates.
          The [Certificate Documentation][] has more detail.

          [Certificate Documentation]: https://cert-manager.io/docs/usage/certificate/
        displayName: cert-manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Documentation
          url: https://cert-manager.io/
        - name: Release Notes
          url: https://cert-manager.io/docs/release-notes
        maintainers:
        - email: cert-manager-maintainers@googlegroups.com
          name: The cert-manager maintainers
        maturity: stable
        minKubeVersion: 1.19.0-0
        provider:
          name: The cert-manager maintainers
          url: https://cert-manager.io/
        relatedImages:
        - quay.io/jetstack/cert-manager-controller:v1.16.5
        - quay.io/jetstack/cert-manager-cainjector:v1.16.5
        - quay.io/jetstack/cert-manager-webhook:v1.16.5
        version: 1.16.5
      entries:
      - name: cert-manager.v1.16.5
        version: 1.16.5
      - name: cert-manager.v1.16.1
        version: 1.16.1
      - name: cert-manager.v1.16.0-beta.0
        version: 1.16.0-beta.0
      - name: cert-manager.v1.15.2
        version: 1.15.2
      - name: cert-manager.v1.15.0
        version: 1.15.0
      - name: cert-manager.v1.14.2
        version: 1.14.2
      - name: cert-manager.v1.14.2-rc1
        version: 1.14.2-rc1
      - name: cert-manager.v1.14.1-rc1
        version: 1.14.1-rc1
      - name: cert-manager.v1.13.3
        version: 1.13.3
      - name: cert-manager.v1.13.3-rc1
        version: 1.13.3-rc1
      - name: cert-manager.v1.13.1
        version: 1.13.1
      - name: cert-manager.v1.13.1-rc1
        version: 1.13.1-rc1
      - name: cert-manager.v1.12.2
        version: 1.12.2
      - name: cert-manager.v1.12.2-rc1
        version: 1.12.2-rc1
      - name: cert-manager.v1.11.4
        version: 1.11.4
      - name: cert-manager.v1.11.4-rc1
        version: 1.11.4-rc1
      - name: cert-manager.v1.11.3-rc1
        version: 1.11.3-rc1
      - name: cert-manager.v1.11.0
        version: 1.11.0
      - name: cert-manager.v1.11.0-rc1
        version: 1.11.0-rc1
      - name: cert-manager.v1.10.2
        version: 1.10.2
      - name: cert-manager.v1.10.2-rc1
        version: 1.10.2-rc1
      - name: cert-manager.v1.10.1
        version: 1.10.1
      - name: cert-manager.v1.10.1-rc1
        version: 1.10.1-rc1
      - name: cert-manager.v1.10.0
        version: 1.10.0
      - name: cert-manager.v1.10.0-rc1
        version: 1.10.0-rc1
      - name: cert-manager.v1.9.1
        version: 1.9.1
      - name: cert-manager.v1.9.1-rc1
        version: 1.9.1-rc1
      - name: cert-manager.v1.8.2
        version: 1.8.2
      - name: cert-manager.v1.8.2-rc1
        version: 1.8.2-rc1
      - name: cert-manager.v1.8.0
        version: 1.8.0
      - name: cert-manager.v1.8.0-rc1
        version: 1.8.0-rc1
      - name: cert-manager.v1.7.2
        version: 1.7.2
      - name: cert-manager.v1.7.1
        version: 1.7.1
      - name: cert-manager.v1.7.1-rc1
        version: 1.7.1-rc1
      - name: cert-manager.v1.6.2
        version: 1.6.2
      - name: cert-manager.v1.6.2-rc1
        version: 1.6.2-rc1
      name: candidate
    - currentCSV: cert-manager.v1.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "example-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "example-clusterissuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example-certificate"
                },
                "spec": {
                  "dnsNames": [
                    "example.com"
                  ],
                  "issuerRef": {
                    "name": "example-issuer"
                  },
                  "secretName": "example-certificate-tls"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "name": "example-certificaterequest"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "name": "example-order"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "example-challenge"
                },
                "spec": {}
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: quay.io/jetstack/cert-manager-controller:v1.16.5
          createdAt: 2025-05-02T09:11:31
          olm.skipRange: '>=1.16.0 <1.16.5'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/internal-objects: |-
            [
              "challenges.acme.cert-manager.io",
              "orders.acme.cert-manager.io"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/cert-manager/cert-manager
          support: The cert-manager maintainers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A CertificateRequest is used to request a signed certificate from one of the
              configured issuers.

              All fields within the CertificateRequest's `spec` are immutable after creation.
              A CertificateRequest will either succeed or fail, as denoted by its `Ready` status
              condition and its `status.failureTime` field.

              A CertificateRequest is a one-shot resource, meaning it represents a single
              point in time request for a certificate and cannot be re-used.
            displayName: CertificateRequest
            kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - description: |-
              A Certificate resource should be created to ensure an up to date and signed
              X.509 certificate is stored in the Kubernetes Secret resource named in `spec.secretName`.

              The stored certificate will be renewed before it expires (as configured by `spec.renewBefore`).
            displayName: Certificate
            kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: Challenge is a type to represent a Challenge request with
              an ACME server
            displayName: (Internal) Challenge
            kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - description: |-
              A ClusterIssuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is similar to an Issuer, however it is cluster-scoped and therefore can
              be referenced by resources that exist in *any* namespace, not just the same
              namespace as the referent.
            displayName: ClusterIssuer
            kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - description: |-
              An Issuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is scoped to a single namespace and can therefore only be referenced by
              resources within the same namespace.
            displayName: Issuer
            kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - description: Order is a type to represent an Order with an ACME server
            displayName: (Internal) Order
            kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          cert-manager is a Kubernetes add-on to automate the management and issuance of TLS certificates from various issuing sources.
          It will ensure certificates are valid and up to date periodically, and attempt to renew certificates at an appropriate time before expiry.

          cert-manager has the following [built in issuers][]:

          * ACME: sign certificates using Let's Encrypt (or any other [ACME][] compliant signer).
          * Vault: sign certificates using HashiCorp Vault.
          * Venafi: sign certificates using Venafi Cloud or Venafi Trust Protection Platform.
          * SelfSigned: the private key of the certificate will be used to sign the certificate.
          * CA: a private key stored as a Kubernetes Secret will be used to sign the certificate.

          [built in issuers]: https://cert-manager.io/docs/configuration/#supported-issuer-types
          [ACME]: https://datatracker.ietf.org/doc/html/rfc8555

          cert-manager also supports [external issuers][],
          which are out-of-tree components that provide support for other certificate authorities, such as:
          [Google Cloud Certificate Authority](https://github.com/jetstack/google-cas-issuer),
          [AWS Private Certificate Authority](https://github.com/cert-manager/aws-privateca-issuer), and
          [Cloudflare Origin CA](https://github.com/cloudflare/origin-ca-issuer)

          [external issuers]: https://cert-manager.io/docs/configuration/external/

          In order to configure cert-manager to issue certificates, you must first create an Issuer or ClusterIssuer.
          These resources represent a particular signing authority and tell cert-manager how to handle your certificate requests.
          The [Issuer Documentation][] has more detail.

          [configure cert-manager]: https://cert-manager.io/docs/configuration/
          [Issuer Documentation]: https://cert-manager.io/docs/concepts/issuer/

          Once an Issuer has been configured, you can create a Certificate resource,
          which is a definition of a certificate request that is to be honored by an issuer
          and which is to be kept up-to-date.
          This is the usual way that you will interact with cert-manager to request signed certificates.
          The [Certificate Documentation][] has more detail.

          [Certificate Documentation]: https://cert-manager.io/docs/usage/certificate/
        displayName: cert-manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Documentation
          url: https://cert-manager.io/
        - name: Release Notes
          url: https://cert-manager.io/docs/release-notes
        maintainers:
        - email: cert-manager-maintainers@googlegroups.com
          name: The cert-manager maintainers
        maturity: stable
        minKubeVersion: 1.19.0-0
        provider:
          name: The cert-manager maintainers
          url: https://cert-manager.io/
        relatedImages:
        - quay.io/jetstack/cert-manager-controller:v1.16.5
        - quay.io/jetstack/cert-manager-cainjector:v1.16.5
        - quay.io/jetstack/cert-manager-webhook:v1.16.5
        version: 1.16.5
      entries:
      - name: cert-manager.v1.16.5
        version: 1.16.5
      - name: cert-manager.v1.16.1
        version: 1.16.1
      - name: cert-manager.v1.15.2
        version: 1.15.2
      - name: cert-manager.v1.15.0
        version: 1.15.0
      - name: cert-manager.v1.14.2
        version: 1.14.2
      - name: cert-manager.v1.13.3
        version: 1.13.3
      - name: cert-manager.v1.13.1
        version: 1.13.1
      - name: cert-manager.v1.13.1-rc1
        version: 1.13.1-rc1
      - name: cert-manager.v1.12.2
        version: 1.12.2
      - name: cert-manager.v1.11.4
        version: 1.11.4
      - name: cert-manager.v1.11.0
        version: 1.11.0
      - name: cert-manager.v1.10.2
        version: 1.10.2
      - name: cert-manager.v1.10.1
        version: 1.10.1
      - name: cert-manager.v1.10.0
        version: 1.10.0
      - name: cert-manager.v1.9.1
        version: 1.9.1
      - name: cert-manager.v1.8.2
        version: 1.8.2
      - name: cert-manager.v1.8.0
        version: 1.8.0
      - name: cert-manager.v1.7.2
        version: 1.7.2
      - name: cert-manager.v1.7.1
        version: 1.7.1
      - name: cert-manager.v1.6.2
        version: 1.6.2
      - name: cert-manager.v1.6.1
        version: 1.6.1
      - name: cert-manager.v1.6.0
        version: 1.6.0
      - name: cert-manager.v1.5.4
        version: 1.5.4
      - name: cert-manager.v1.5.3
        version: 1.5.3
      - name: cert-manager.v1.4.4
        version: 1.4.4
      - name: cert-manager.v1.4.3
        version: 1.4.3
      - name: cert-manager.v1.4.2
        version: 1.4.2
      - name: cert-manager.v1.4.1
        version: 1.4.1
      - name: cert-manager.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: cert-manager
    provider:
      name: The cert-manager maintainers
      url: https://cert-manager.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.12
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-06-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:9aed98a061b2c76f3c512c1d6453df5920f0c779b33d7168c1e7ab9d8bc928a4
        version: 0.0.12
      entries:
      - name: joget-dx8-openshift-operator.v0.0.12
        version: 0.0.12
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Dell Technologies
      provider-url: https://github.com/dell/csm-operator
    name: dell-csm-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dell-csm-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "authorization",
                  "namespace": "authorization"
                },
                "spec": {
                  "modules": [
                    {
                      "components": [
                        {
                          "enabled": true,
                          "name": "nginx"
                        },
                        {
                          "enabled": true,
                          "name": "cert-manager"
                        },
                        {
                          "authorizationController": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231",
                          "authorizationControllerReplicas": 1,
                          "certificate": "",
                          "controllerReconcileInterval": "5m",
                          "enabled": true,
                          "hostname": "csm-authorization.com",
                          "leaderElection": true,
                          "name": "proxy-server",
                          "opa": "docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095",
                          "opaKubeMgmt": "docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517",
                          "openTelemetryCollectorAddress": "",
                          "privateKey": "",
                          "proxyServerIngress": [
                            {
                              "annotations": {},
                              "hosts": [],
                              "ingressClassName": "nginx"
                            }
                          ],
                          "proxyService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682",
                          "proxyServiceReplicas": 1,
                          "roleService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c",
                          "roleServiceReplicas": 1,
                          "storageService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5",
                          "storageServiceReplicas": 1,
                          "tenantService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358",
                          "tenantServiceReplicas": 1
                        },
                        {
                          "commander": "docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a",
                          "name": "redis",
                          "redis": "docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f",
                          "redisCommander": "rediscommander",
                          "redisName": "redis-csm",
                          "redisReplicas": 5,
                          "redisSecretProviderClass": [
                            {
                              "redisPasswordKey": "",
                              "redisSecretName": "",
                              "redisUsernameKey": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "sentinel": "sentinel"
                        },
                        {
                          "configSecretProviderClass": [
                            {
                              "configSecretName": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "name": "config"
                        },
                        {
                          "name": "storage-system-credentials",
                          "secretProviderClasses": {
                            "vault": [
                              "secret-provider-class-1",
                              "secret-provider-class-2"
                            ]
                          }
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": true,
                      "forceRemoveModule": true,
                      "name": "authorization-proxy-server"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "isilon",
                  "namespace": "isilon"
                },
                "spec": {
                  "driver": {
                    "authSecret": "isilon-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VERBOSE",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_ISI_PORT",
                          "value": "8080"
                        },
                        {
                          "name": "X_CSI_ISI_PATH",
                          "value": "/ifs/data/csi"
                        },
                        {
                          "name": "X_CSI_ISI_NO_PROBE_ON_START",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_AUTOPROBE",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_AUTH_TYPE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_CUSTOM_TOPOLOGY_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOISILON_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "AZ_RECONCILE_INTERVAL",
                          "value": "1h"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_ISI_QUOTA_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_ACCESS_ZONE",
                          "value": "System"
                        },
                        {
                          "name": "X_CSI_ISI_VOLUME_PATH_PERMISSIONS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_IGNORE_UNRESOLVABLE_HOSTS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "isilon",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerscale"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSCALE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_PERFORMANCE_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSCALE_QUOTA_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "ISICLIENT_INSECURE",
                              "value": "true"
                            },
                            {
                              "name": "ISICLIENT_AUTH_TYPE",
                              "value": "1"
                            },
                            {
                              "name": "ISICLIENT_VERBOSE",
                              "value": "0"
                            },
                            {
                              "name": "POWERSCALE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSCALE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0",
                          "name": "metrics-powerscale"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-isilon/csi_sock",
                            "--mode=node",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powermax",
                  "namespace": "powermax"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powermax-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_MANAGED_ARRAYS",
                          "value": "000000000001,000000000002"
                        },
                        {
                          "name": "X_CSI_POWERMAX_PORTGROUPS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_TRANSPORT_PROTOCOL",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_K8S_CLUSTER_PREFIX",
                          "value": "CSM"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "X_CSI_VSPHERE_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VSPHERE_PORTGROUP",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VSPHERE_HOSTNAME",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VCENTER_HOST",
                          "value": ""
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "X_CSI_POWERMAX_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "X_CSI_POWERMAX_DEBUG",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_REVPROXY_USE_SECRET",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powermax",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERMAX_ISCSI_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_TOPOLOGY_CONTROL_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": [
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/memory-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/disk-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/network-unavailable",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REVPROXY_TLS_SECRET",
                              "value": "csirevproxy-tls-secret"
                            },
                            {
                              "name": "X_CSI_REVPROXY_PORT",
                              "value": "2222"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            },
                            {
                              "name": "DeployAsSidecar",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265",
                          "name": "csipowermax-reverseproxy"
                        }
                      ],
                      "configVersion": "v2.14.0",
                      "name": "csireverseproxy"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powermax"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERMAX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERMAX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414",
                          "name": "metrics-powermax"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/powermax.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powerstore",
                  "namespace": "powerstore"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powerstore-config",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_NODE_NAME_PREFIX",
                          "value": "csi-node"
                        },
                        {
                          "name": "X_CSI_FC_PORTS_FILTER_FILE_PATH",
                          "value": "/etc/fc-ports-filter"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOPOWERSTORE_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSM_AUTH_ENABLED",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_MULTI_NAS_FAILURE_THRESHOLD",
                          "value": "5"
                        },
                        {
                          "name": "X_CSI_MULTI_NAS_COOLDOWN_PERIOD",
                          "value": "5m"
                        },
                        {
                          "name": "X_CSI_NFS_ACLS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerstore",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-powerstore.dellemc.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerstore"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSTORE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_SPACE_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_ARRAY_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_FILE_SYSTEM_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSTORE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSTORE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "ZIPKIN_URI",
                              "value": ""
                            },
                            {
                              "name": "ZIPKIN_SERVICE_NAME",
                              "value": "metrics-powerstore"
                            },
                            {
                              "name": "ZIPKIN_PROBABILITY",
                              "value": "0.0"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60",
                          "name": "metrics-powerstore"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "unity",
                  "namespace": "unity"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_UNITY_ALLOW_MULTI_POD_ACCESS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_EPHEMERAL_STAGING_PATH",
                          "value": "/var/lib/kubelet/plugins/kubernetes.io/csi/pv/"
                        },
                        {
                          "name": "X_CSI_ISCSI_CHROOT",
                          "value": "/noderoot"
                        },
                        {
                          "name": "X_CSI_UNITY_SYNC_NODEINFO_INTERVAL",
                          "value": "15"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOUNITY_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOUNITY_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "TENANT_NAME",
                          "value": ""
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_UNITY_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "unity",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "vxflexos",
                  "namespace": "vxflexos"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLELISTVOLUMESNAPSHOT",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLESNAPSHOTCGDELETE",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "INFO"
                        },
                        {
                          "name": "GOSCALEIO_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOSCALEIO_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_QUOTA_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "INTERFACE_NAMES",
                          "value": null
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERFLEX_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "File",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerflex",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "initContainers": [
                      {
                        "envs": [
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          },
                          {
                            "name": "REPO_ADDRESS",
                            "value": "sftp://0.0.0.0"
                          },
                          {
                            "name": "REPO_USER",
                            "value": "sdcRepoUser"
                          },
                          {
                            "name": "MODULE_SIGCHECK",
                            "value": "0"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "imagePullPolicy": "IfNotPresent",
                        "name": "sdc"
                      }
                    ],
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_SDC_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_APPROVE_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_PREFIX",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_SDC_SFTP_REPO_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 1,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "enabled": false,
                        "envs": [
                          {
                            "name": "HOST_PID",
                            "value": "1"
                          },
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "name": "sdc-monitor"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "csi-external-health-monitor-controller"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERFLEX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_SDC_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_SDC_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERFLEX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d",
                          "name": "metrics-powerflex"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerflex"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--csisock=unix:/var/lib/kubelet/plugins/vxflexos.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
          createdAt: "2025-09-05T13:18:42Z"
          description: Easily install and manage Dell’s CSI Drivers and CSM
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/dell/csm-operator
          support: Dell Technologies
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ContainerStorageModule is the Schema for the containerstoragemodules
              API
            displayName: Container Storage Module
            kind: ContainerStorageModule
            name: containerstoragemodules.storage.dell.com
            version: v1
        description: "Dell Container Storage Modules (CSM) Operator is a Kubernetes
          Operator which can be \nused to install and manage Dell’s CSI drivers and
          CSM modules. \nBy using Dell CSM Operator, enterprises can quickly and easily
          deploy the \nCSM modules for CSI drivers making it easy for DevOps \nteams
          to build and optimize their Kubernetes workloads using enterprise grade
          storage \nwhile consistently aligning with the desired outcomes of the organization.\nThis
          operator is available as both community and certified operator for Openshift
          \nand as a community operator for upstream Kubernetes.\nOperator can be
          deployed via OperatorHub.io. using Operator Lifecycle Manager (OLM)\n\n###
          About Dell CSM\n\nDell CSM brings powerful enterprise storage features and
          functionality to Kubernetes for \neasier adoption of cloud native workloads,
          improved productivity and scalable operations. \n\nCSM delivers software
          modules for storage management that provide developers access to \nbuild
          automation for enhanced IT needs and other critical enterprise storage features.
          \n\nCurrently, the following modules are available as part of Dell CSM.
          \n*\tObservability - Delivers a high-level view of the storage capacity
          \n  and performance usage via Grafana dashboards to the Kubernetes users\n*\tReplication
          - Enables array replication capabilities for Kubernetes users\n*\tAuthorization
          - Provides both storage and Kubernetes administrators the ability \n  to
          apply RBAC for Dell CSI Drivers\n*\tResiliency - Designed to make Kubernetes
          applications more resilient to \n  node failures\n\nThese modules are available
          for various storage platforms in the Dell portfolio, \nsuch as Dell PowerStore,
          Dell PowerMax, Dell PowerFlex, Dell PowerScale, and Dell Unity XT.\n\nFor
          more information, visit [Dell’s Container Storage Module web page](https://www.dell.com/en-us/dt/solutions/kubernetes-containers/kubernetes-storage.htm)\n\nAdditional
          Materials\n*\t[Documentation](https://dell.github.io/csm-docs)\n*\t[CSM
          Solution Brief](https://www.delltechnologies.com/asset/en-us/products/storage/briefs-summaries/csm-solution-brief.pdf)\n"
        displayName: Dell Container Storage Modules
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dell Container Storage Modules
        - Dell CSI Driver
        - Dell CSM Modules
        - Powerflex
        - Powerscale
        - Powerstore
        - Unity
        - Authorization
        - Observability
        - Replication
        links:
        - name: Documentation
          url: https://dell.github.io/csm-docs
        maintainers:
        - email: container.storage.modules@dell.com
          name: Dell Container Storage Modules
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Dell Technologies
          url: https://github.com/dell/csm-operator
        relatedImages:
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0
        - registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa
        - docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265
        - registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea
        - docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682
        - registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334
        - docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f
        - quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:59335acaca515fceed5f0350c46cfcfd6037475f0cf35ae8ab41d37333039e2e
        - ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e
        - quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540
        - registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf
        - registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16
        - registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe
        version: 1.10.0
      entries:
      - name: dell-csm-operator-certified.v1.10.0
        version: 1.10.0
      - name: dell-csm-operator-certified.v1.9.1
        version: 1.9.1
      - name: dell-csm-operator-certified.v1.9.0
        version: 1.9.0
      - name: dell-csm-operator-certified.v1.8.1
        version: 1.8.1
      - name: dell-csm-operator-certified.v1.8.0
        version: 1.8.0
      - name: dell-csm-operator-certified.v1.7.0
        version: 1.7.0
      - name: dell-csm-operator-certified.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: dell-csm-operator-certified
    provider:
      name: Dell Technologies
      url: https://github.com/dell/csm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: external-dns-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: external-dns-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "provider": {
                    "type": "AWS",
                    "aws": {
                      "credentials": {
                        "name": "aws-access-key"
                      }
                    }
                  },
                  "source": {
                    "type": "Service"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:6be1403e418f9b63ec8ebcc8858e602409bc41fdee4586acc627623a78b50eb5
          createdAt: 2022-03-01T21:00:46
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## Prerequisites and Requirements
           The ExternalDNS Operator has minimal cluster level permissions. Unfortunately OLM doesn't allow the local role creation in the operand namespace yet.
           Therefore make sure the following commands are run against your cluster before the installation of the operator:
           * Create the operand namespace: `oc create ns external-dns`
           * Grant the operator access to manage operand resources: `oc apply -f https://raw.githubusercontent.com/openshift/external-dns-operator/release-0.1/config/rbac/extra-roles.yaml`
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b6fde43491c2e0733c8fe6ee8660f7f6435fe20c0b6d435cfc7f86eeab8b4748
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:6be1403e418f9b63ec8ebcc8858e602409bc41fdee4586acc627623a78b50eb5
        - registry.redhat.io/edo/external-dns-rhel8@sha256:c1134bb46172997ef7278b6cefbb0da44e72a9f808a7cd67b3c65d464754cab9
        version: 0.1.2
      entries:
      - name: external-dns-operator.v0.1.2
        version: 0.1.2
      name: alpha
    - currentCSV: external-dns-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
          createdAt: 2023-11-15T09:24:04
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e33874a51971a90917cf437fc6cbeea1da569ba07217f66f1bba96d594f58aed
        - registry.redhat.io/edo/external-dns-rhel8@sha256:6263a78b2a170e56dfca2173f14dbf14f0a5bddb6ea600bfdef248c7d5f20e4b
        version: 1.2.0
      entries:
      - name: external-dns-operator.v1.2.0
        version: 1.2.0
      name: stable-v1
    - currentCSV: external-dns-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:0877eb0a111efe0bff63173a2dfeb7a06a61afd0efbcdba053322a539e0dc667
          createdAt: 2024-01-22T09:26:37
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/edo/external-dns-rhel8@sha256:17f4127cbb9a9304769e9655c19cbea2260107c7e9371a99a9aa3e85d759c00b
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:0877eb0a111efe0bff63173a2dfeb7a06a61afd0efbcdba053322a539e0dc667
        version: 1.0.1
      entries:
      - name: external-dns-operator.v1.0.1
        version: 1.0.1
      - name: external-dns-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: external-dns-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:9ca751e84b8843cc62dc60550e5c4af741bf09a381ef62220e7eb3760ffe8524
          createdAt: 2024-01-22T09:56:14
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:9ca751e84b8843cc62dc60550e5c4af741bf09a381ef62220e7eb3760ffe8524
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/edo/external-dns-rhel8@sha256:ffcc3f2e9b9feb889e7882465f7f034aba6b9ee0699a1b6c24e83f501a63ecca
        version: 1.1.1
      entries:
      - name: external-dns-operator.v1.1.1
        version: 1.1.1
      - name: external-dns-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: external-dns-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
          createdAt: 2023-11-15T09:24:04
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e33874a51971a90917cf437fc6cbeea1da569ba07217f66f1bba96d594f58aed
        - registry.redhat.io/edo/external-dns-rhel8@sha256:6263a78b2a170e56dfca2173f14dbf14f0a5bddb6ea600bfdef248c7d5f20e4b
        version: 1.2.0
      entries:
      - name: external-dns-operator.v1.2.0
        version: 1.2.0
      name: stable-v1.2
    defaultChannel: stable-v1
    packageName: external-dns-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: rook-ceph-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rook-ceph-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephCluster",
                "metadata": {
                  "name": "my-rook-ceph",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "cephVersion": {
                    "image": "registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d"
                  },
                  "dataDirHostPath": "/var/lib/rook",
                  "mon": {
                    "count": 3
                  },
                  "dashboard": {
                    "enabled": true
                  },
                  "network": {
                    "hostNetwork": false
                  },
                  "rbdMirroring": {
                    "workers": 0
                  },
                  "storage": {
                    "useAllNodes": true,
                    "useAllDevices": true
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephBlockPool",
                "metadata": {
                  "name": "replicapool",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "failureDomain": "host",
                  "replicated": {
                    "size": 3
                  },
                  "annotations": null
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStore",
                "metadata": {
                  "name": "my-store",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "metadataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "dataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "gateway": {
                    "type": "s3",
                    "sslCertificateRef": null,
                    "port": 8080,
                    "securePort": null,
                    "instances": 1,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStoreUser",
                "metadata": {
                  "name": "my-user",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "store": "my-store",
                  "displayName": "my display name"
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephNFS",
                "metadata": {
                  "name": "my-nfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "rados": {
                    "pool": "myfs-data0",
                    "namespace": "nfs-ns"
                  },
                  "server": {
                    "active": 3,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephClient",
                "metadata": {
                  "name": "cinder",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "caps": {
                    "mon": "profile rbd",
                    "osd": "profile rbd pool=volumes, profile rbd pool=vms, profile rbd-read-only pool=images"
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephFilesystem",
                "metadata": {
                  "name": "myfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "dataPools": [
                    {
                      "compressionMode": "",
                      "crushRoot": "",
                      "deviceClass": "",
                      "erasureCoded": {
                        "algorithm": "",
                        "codingChunks": 0,
                        "dataChunks": 0
                      },
                      "failureDomain": "host",
                      "replicated": {
                        "requireSafeReplicaSize": false,
                        "size": 1,
                        "targetSizeRatio": 0.5
                      }
                    }
                  ],
                  "metadataPool": {
                    "compressionMode": "",
                    "crushRoot": "",
                    "deviceClass": "",
                    "erasureCoded": {
                      "algorithm": "",
                      "codingChunks": 0,
                      "dataChunks": 0
                    },
                    "failureDomain": "",
                    "replicated": {
                      "requireSafeReplicaSize": false,
                      "size": 1,
                      "targetSizeRatio": 0
                    }
                  },
                  "metadataServer": {
                    "activeCount": 1,
                    "activeStandby": true,
                    "placement": {},
                    "resources": {}
                  },
                  "preservePoolsOnDelete": false,
                  "preserveFilesystemOnDelete": false
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephRBDMirror",
                "metadata": {
                  "name": "my-rbd-mirror",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "annotations": null,
                  "count": 1,
                  "placement": {
                    "topologyKey": "kubernetes.io/hostname"
                  },
                  "resources": null
                }
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:cf7f72469d25fb37301a8159406613673f8968c82af1035fd45a470e0ae8acfa
          externalClusterScript: |-
            IiIiCkNvcHlyaWdodCAyMDIwIFRoZSBSb29rIEF1dGhvcnMuIEFsbCByaWdodHMgcmVzZXJ2ZWQu
            CgpMaWNlbnNlZCB1bmRlciB0aGUgQXBhY2hlIExpY2Vuc2UsIFZlcnNpb24gMi4wICh0aGUgIkxp
            Y2Vuc2UiKTsKeW91IG1heSBub3QgdXNlIHRoaXMgZmlsZSBleGNlcHQgaW4gY29tcGxpYW5jZSB3
            aXRoIHRoZSBMaWNlbnNlLgpZb3UgbWF5IG9idGFpbiBhIGNvcHkgb2YgdGhlIExpY2Vuc2UgYXQK
            CglodHRwOi8vd3d3LmFwYWNoZS5vcmcvbGljZW5zZXMvTElDRU5TRS0yLjAKClVubGVzcyByZXF1
            aXJlZCBieSBhcHBsaWNhYmxlIGxhdyBvciBhZ3JlZWQgdG8gaW4gd3JpdGluZywgc29mdHdhcmUK
            ZGlzdHJpYnV0ZWQgdW5kZXIgdGhlIExpY2Vuc2UgaXMgZGlzdHJpYnV0ZWQgb24gYW4gIkFTIElT
            IiBCQVNJUywKV0lUSE9VVCBXQVJSQU5USUVTIE9SIENPTkRJVElPTlMgT0YgQU5ZIEtJTkQsIGVp
            dGhlciBleHByZXNzIG9yIGltcGxpZWQuClNlZSB0aGUgTGljZW5zZSBmb3IgdGhlIHNwZWNpZmlj
            IGxhbmd1YWdlIGdvdmVybmluZyBwZXJtaXNzaW9ucyBhbmQKbGltaXRhdGlvbnMgdW5kZXIgdGhl
            IExpY2Vuc2UuCiIiIgoKaW1wb3J0IGVycm5vCmltcG9ydCBzeXMKaW1wb3J0IGpzb24KaW1wb3J0
            IGFyZ3BhcnNlCmltcG9ydCByZQppbXBvcnQgc3VicHJvY2VzcwppbXBvcnQgaG1hYwpmcm9tIGhh
            c2hsaWIgaW1wb3J0IHNoYTEgYXMgc2hhCmZyb20gb3MgaW1wb3J0IGxpbmVzZXAgYXMgTElORVNF
            UApmcm9tIG9zIGltcG9ydCBwYXRoCmZyb20gZW1haWwudXRpbHMgaW1wb3J0IGZvcm1hdGRhdGUK
            aW1wb3J0IHJlcXVlc3RzCmZyb20gcmVxdWVzdHMuYXV0aCBpbXBvcnQgQXV0aEJhc2UKCnB5M2sg
            PSBGYWxzZQppZiBzeXMudmVyc2lvbl9pbmZvLm1ham9yID49IDM6CiAgICBweTNrID0gVHJ1ZQog
            ICAgaW1wb3J0IHVybGxpYi5wYXJzZQogICAgZnJvbSBpcGFkZHJlc3MgaW1wb3J0IGlwX2FkZHJl
            c3MsIElQdjRBZGRyZXNzCgpNb2R1bGVOb3RGb3VuZEVycm9yID0gSW1wb3J0RXJyb3IKCnRyeToK
            ICAgIGltcG9ydCByYWRvcwpleGNlcHQgTW9kdWxlTm90Rm91bmRFcnJvciBhcyBub01vZEVycjoK
            ICAgIHByaW50KGYiRXJyb3I6IHtub01vZEVycn1cbkV4aXRpbmcgdGhlIHNjcmlwdC4uLiIpCiAg
            ICBzeXMuZXhpdCgxKQoKdHJ5OgogICAgaW1wb3J0IHJiZApleGNlcHQgTW9kdWxlTm90Rm91bmRF
            cnJvciBhcyBub01vZEVycjoKICAgIHByaW50KGYiRXJyb3I6IHtub01vZEVycn1cbkV4aXRpbmcg
            dGhlIHNjcmlwdC4uLiIpCiAgICBzeXMuZXhpdCgxKQoKdHJ5OgogICAgIyBmb3IgMi43LngKICAg
            IGZyb20gU3RyaW5nSU8gaW1wb3J0IFN0cmluZ0lPCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9y
            OgogICAgIyBmb3IgMy54CiAgICBmcm9tIGlvIGltcG9ydCBTdHJpbmdJTwoKdHJ5OgogICAgIyBm
            b3IgMi43LngKICAgIGZyb20gdXJscGFyc2UgaW1wb3J0IHVybHBhcnNlCiAgICBmcm9tIHVybGxp
            YiBpbXBvcnQgdXJsZW5jb2RlIGFzIHVybGVuY29kZQpleGNlcHQgTW9kdWxlTm90Rm91bmRFcnJv
            cjoKICAgICMgZm9yIDMueAogICAgZnJvbSB1cmxsaWIucGFyc2UgaW1wb3J0IHVybHBhcnNlCiAg
            ICBmcm9tIHVybGxpYi5wYXJzZSBpbXBvcnQgdXJsZW5jb2RlIGFzIHVybGVuY29kZQoKdHJ5Ogog
            ICAgZnJvbSBiYXNlNjQgaW1wb3J0IGVuY29kZXN0cmluZwpleGNlcHQ6CiAgICBmcm9tIGJhc2U2
            NCBpbXBvcnQgZW5jb2RlYnl0ZXMgYXMgZW5jb2Rlc3RyaW5nCgoKY2xhc3MgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbihFeGNlcHRpb24pOgogICAgcGFzcwoKCiMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMgRHVtbXlSYWRv
            cyAjIyMjIyMjIyMjIyMjIyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjCiMgdGhpcyBpcyBtYWlubHkgZm9yIHRlc3RpbmcgYW5kIGNvdWxkIGJlIHVz
            ZWQgd2hlcmUgJ3JhZG9zJyBpcyBub3QgYXZhaWxhYmxlCgoKY2xhc3MgRHVtbXlSYWRvcyhvYmpl
            Y3QpOgogICAgZGVmIF9faW5pdF9fKHNlbGYpOgogICAgICAgIHNlbGYucmV0dXJuX3ZhbCA9IDAK
            ICAgICAgICBzZWxmLmVycl9tZXNzYWdlID0gIiIKICAgICAgICBzZWxmLnN0YXRlID0gImNvbm5l
            Y3RlZCIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwID0ge30KICAgICAgICBzZWxmLmNtZF9u
            YW1lcyA9IHt9CiAgICAgICAgc2VsZi5faW5pdF9jbWRfb3V0cHV0X21hcCgpCiAgICAgICAgc2Vs
            Zi5kdW1teV9ob3N0X2lwX21hcCA9IHt9CgogICAgZGVmIF9pbml0X2NtZF9vdXRwdXRfbWFwKHNl
            bGYpOgogICAgICAgIGpzb25fZmlsZV9uYW1lID0gInRlc3QtZGF0YS9jZXBoLXN0YXR1cy1vdXQi
            CiAgICAgICAgc2NyaXB0X2RpciA9IHBhdGguYWJzcGF0aChwYXRoLmRpcm5hbWUoX19maWxlX18p
            KQogICAgICAgIGNlcGhfc3RhdHVzX3N0ciA9ICIiCiAgICAgICAgd2l0aCBvcGVuKAogICAgICAg
            ICAgICBwYXRoLmpvaW4oc2NyaXB0X2RpciwganNvbl9maWxlX25hbWUpLCBtb2RlPSJyIiwgZW5j
            b2Rpbmc9IlVURi04IgogICAgICAgICkgYXMganNvbl9maWxlOgogICAgICAgICAgICBjZXBoX3N0
            YXR1c19zdHIgPSBqc29uX2ZpbGUucmVhZCgpCiAgICAgICAgc2VsZi5jbWRfbmFtZXNbImZzIGxz
            Il0gPSAiIiJ7ImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJmcyBscyJ9IiIiCiAgICAgICAg
            c2VsZi5jbWRfbmFtZXNbInF1b3J1bV9zdGF0dXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3Jt
            YXQiOiAianNvbiIsICJwcmVmaXgiOiAicXVvcnVtX3N0YXR1cyJ9IiIiCiAgICAgICAgKQogICAg
            ICAgIHNlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXSA9ICgKICAgICAgICAgICAgIiIieyJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAibWdyIHNlcnZpY2VzIn0iIiIKICAgICAgICApCiAg
            ICAgICAgIyBhbGwgdGhlIGNvbW1hbmRzIGFuZCB0aGVpciBvdXRwdXQKICAgICAgICBzZWxmLmNt
            ZF9vdXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJmcyBscyJdXSA9ICgKICAgICAgICAgICAgIiIi
            W3sibmFtZSI6Im15ZnMiLCJtZXRhZGF0YV9wb29sIjoibXlmcy1tZXRhZGF0YSIsIm1ldGFkYXRh
            X3Bvb2xfaWQiOjIsImRhdGFfcG9vbF9pZHMiOlszXSwiZGF0YV9wb29scyI6WyJteWZzLXJlcGxp
            Y2F0ZWQiXX1dIiIiCiAgICAgICAgKQogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbc2VsZi5j
            bWRfbmFtZXNbInF1b3J1bV9zdGF0dXMiXV0gPSAoCiAgICAgICAgICAgICIiInsiZWxlY3Rpb25f
            ZXBvY2giOjMsInF1b3J1bSI6WzBdLCJxdW9ydW1fbmFtZXMiOlsiYSJdLCJxdW9ydW1fbGVhZGVy
            X25hbWUiOiJhIiwicXVvcnVtX2FnZSI6MTQzODUsImZlYXR1cmVzIjp7InF1b3J1bV9jb24iOiI0
            NTQwMTM4MjkyODM2Njk2MDYzIiwicXVvcnVtX21vbiI6WyJrcmFrZW4iLCJsdW1pbm91cyIsIm1p
            bWljIiwib3NkbWFwLXBydW5lIiwibmF1dGlsdXMiLCJvY3RvcHVzIl19LCJtb25tYXAiOnsiZXBv
            Y2giOjEsImZzaWQiOiJhZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5ZDBmMWMiLCJtb2Rp
            ZmllZCI6IjIwMjAtMDUtMDdUMDM6MzY6MzkuOTE4MDM1WiIsImNyZWF0ZWQiOiIyMDIwLTA1LTA3
            VDAzOjM2OjM5LjkxODAzNVoiLCJtaW5fbW9uX3JlbGVhc2UiOjE1LCJtaW5fbW9uX3JlbGVhc2Vf
            bmFtZSI6Im9jdG9wdXMiLCJmZWF0dXJlcyI6eyJwZXJzaXN0ZW50IjpbImtyYWtlbiIsImx1bWlu
            b3VzIiwibWltaWMiLCJvc2RtYXAtcHJ1bmUiLCJuYXV0aWx1cyIsIm9jdG9wdXMiXSwib3B0aW9u
            YWwiOltdfSwibW9ucyI6W3sicmFuayI6MCwibmFtZSI6ImEiLCJwdWJsaWNfYWRkcnMiOnsiYWRk
            cnZlYyI6W3sidHlwZSI6InYyIiwiYWRkciI6IjEwLjExMC4yMDUuMTc0OjMzMDAiLCJub25jZSI6
            MH0seyJ0eXBlIjoidjEiLCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OSIsIm5vbmNlIjowfV19
            LCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OS8wIiwicHVibGljX2FkZHIiOiIxMC4xMTAuMjA1
            LjE3NDo2Nzg5LzAiLCJwcmlvcml0eSI6MCwid2VpZ2h0IjowfV19fSIiIgogICAgICAgICkKICAg
            ICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXV0g
            PSAoCiAgICAgICAgICAgICIiInsiZGFzaGJvYXJkIjoiaHR0cHM6Ly9jZXBoLWRhc2hib2FyZDo4
            NDQzLyIsInByb21ldGhldXMiOiJodHRwOi8vY2VwaC1kYXNoYm9hcmQtZGI6OTI4My8ifSIiIgog
            ICAgICAgICkKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNh
            cHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMiLCAib3Nk
            IiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0
            YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250
            cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJdLCAiZW50aXR5Ijog
            ImNsaWVudC5oZWFsdGhjaGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRo
            IGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuaGVh
            bHRoY2hlY2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRB
            PT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cyIs
            Im9zZCI6InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5j
            b250cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJ9fV0iIiIKICAg
            ICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIs
            ICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAib3NkIiwgInBy
            b2ZpbGUgcmJkIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtbm9kZSIsICJmb3JtYXQiOiAi
            anNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIi
            W3siZW50aXR5IjoiY2xpZW50LmNzaS1yYmQtbm9kZSIsImtleSI6IkFRQk9nck5lSGJLMUF4QUF1
            YllCZVY4UzFVL0dQenE1U1ZlcTZnPT0iLCJjYXBzIjp7Im1vbiI6InByb2ZpbGUgcmJkLCBhbGxv
            dyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsIm9zZCI6InByb2ZpbGUgcmJkIn19XSIiIgogICAg
            ICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsibW9uIiwg
            InByb2ZpbGUgcmJkLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxs
            b3cgcnciLCAib3NkIiwgInByb2ZpbGUgcmJkIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQt
            cHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNy
            ZWF0ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktcmJkLXByb3Zp
            c2lvbmVyIiwia2V5IjoiQVFCTmdyTmUxZ2V5S3hBQThla1ZpUmRFK2hzczVPd2VZQmt3Tmc9PSIs
            ImNhcHMiOnsibWdyIjoiYWxsb3cgcnciLCJtb24iOiJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFu
            ZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJwcm9maWxlIHJiZCJ9fV0iIiIKICAgICAgICBzZWxm
            LmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyBy
            LCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3Nk
            IiwgImFsbG93IHJ3IHRhZyBjZXBoZnMgKj0qIiwgIm1kcyIsICJhbGxvdyBydyJdLCAiZW50aXR5
            IjogImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4Ijog
            ImF1dGggZ2V0LW9yLWNyZWF0ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVu
            dC5jc2ktY2VwaGZzLW5vZGUiLCJrZXkiOiJBUUJPZ3JOZUVOdW5LeEFBUENtZ0U3UjZHOERjWG5h
            SjFGMzJxZz09IiwiY2FwcyI6eyJtZHMiOiJhbGxvdyBydyIsIm1nciI6ImFsbG93IHJ3IiwibW9u
            IjoiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBy
            dyB0YWcgY2VwaGZzICo9KiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAg
            ICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kICdvc2Qg
            YmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3NkIiwgImFsbG93IHJ3IHRhZyBjZXBo
            ZnMgbWV0YWRhdGE9KiJdLCAiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVy
            IiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgog
            ICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIs
            ImtleSI6IkFRQk9nck5lQUZnY0dCQUF2R3FLT0FEMEQzeHhtVlkwUjkxMmRnPT0iLCJjYXBzIjp7
            Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2Nr
            bGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoifX1dIiIiCiAgICAg
            ICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAi
            YWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdyIiwgImFsbG93IHJ3
            IiwgIm9zZCIsICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiXSwgImVudGl0eSI6ICJj
            bGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVuc2hpZnQtc3RvcmFnZSIsICJmb3JtYXQi
            OiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0g
            IiIiW3siZW50aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0
            b3JhZ2UiLCJrZXkiOiJCUUJPZ3JOZUFGZ2NHQkFBdkdxS09BRDBEM3h4bVZZMFI5MTJkZz09Iiwi
            Y2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29z
            ZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIn19XSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsi
            bW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciIsICJh
            bGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT1teWZzIl0sICJl
            bnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0b3JhZ2Ut
            bXlmcyIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0i
            IiIKICAgICAgICBdID0gIiIiW3siZW50aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9u
            ZXItb3BlbnNoaWZ0LXN0b3JhZ2UtbXlmcyIsImtleSI6IkNRQk9nck5lQUZnY0dCQUF2R3FLT0FE
            MEQzeHhtVlkwUjkxMmRnPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cg
            ciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2Vw
            aGZzIG1ldGFkYXRhPW15ZnMifX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAg
            ICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCBxdW9y
            dW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIiwgImFsbG93IGNvbW1hbmQg
            Y29uZmlnIiwgIm9zZCIsICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFsbG93IHJ3eCBwb29sPWRl
            ZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9ZGVm
            YXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJndy5sb2csIGFsbG93IHgg
            cG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4Il0sICJlbnRpdHkiOiAiY2xpZW50LmhlYWx0
            aGNoZWNrZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0
            ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5oZWFsdGhjaGVja2VyIiwi
            a2V5IjoiQVFERmtiTmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsi
            bW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFu
            ZCB2ZXJzaW9uIiwgIm1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsICJvc2QiOiAicHJvZmls
            ZSByYmQtcmVhZC1vbmx5LCBhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhLCBhbGxvdyBy
            IHBvb2w9LnJndy5yb290LCBhbGxvdyBydyBwb29sPWRlZmF1bHQucmd3LmNvbnRyb2wsIGFsbG93
            IHJ4IHBvb2w9ZGVmYXVsdC5yZ3cubG9nLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0
            cy5pbmRleCJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAi
            IiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMs
            IGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cgY29tbWFuZCBjb25maWciLCAi
            b3NkIiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5j
            b250cm9sLCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1
            bHQucmd3LmJ1Y2tldHMuaW5kZXgiXSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIs
            ICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBjYXBzIn0iIiIKICAgICAgICBdID0g
            IiIiW3siZW50aXR5IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0NWJG
            UkFBVG5kTE5VU1JLcnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIjog
            ImFsbG93IGNvbW1hbmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFs
            bG93IHJ3eCBwb29sPWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFs
            bG93IHJ3IHBvb2w9ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJn
            dy5sb2csIGFsbG93IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4In19XSIiIgogICAg
            ICAgIHNlbGYuY21kX291dHB1dF9tYXBbIiIieyJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            bWdyIHNlcnZpY2VzIn0iIiJdID0gKAogICAgICAgICAgICAiIiJ7ImRhc2hib2FyZCI6ICJodHRw
            Oi8vcm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6NzAwMC8iLCAicHJvbWV0aGV1cyI6
            ICJodHRwOi8vcm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6OTI4My8ifSIiIgogICAg
            ICAgICkKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0
            eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            YXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJ7ImRhc2hib2FyZCI6ICJodHRwOi8vcm9vay1j
            ZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6NzAwMC8iLCAicHJvbWV0aGV1cyI6ICJodHRwOi8v
            cm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6OTI4My8ifSIiIgogICAgICAgIHNlbGYu
            Y21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNsaWVudC5oZWFsdGhj
            aGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldCJ9IiIiCiAgICAg
            ICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5oZWFsdGhjaGVja2VyIiwia2V5IjoiQVFERmti
            TmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjogImFsbG93
            IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFuZCB2ZXJzaW9uIiwg
            Im1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsICJvc2QiOiAicHJvZmlsZSByYmQtcmVhZC1v
            bmx5LCBhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhLCBhbGxvdyByIHBvb2w9LnJndy5y
            b290LCBhbGxvdyBydyBwb29sPWRlZmF1bHQucmd3LmNvbnRyb2wsIGFsbG93IHJ4IHBvb2w9ZGVm
            YXVsdC5yZ3cubG9nLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJ9fV0i
            IiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0eSI6
            ICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJh
            dXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0
            X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtbm9kZSIsICJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJb
            XSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5
            IjogImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRf
            b3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMt
            cHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIK
            ICAgICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAg
            ICAgICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVuc2hp
            ZnQtc3RvcmFnZSIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgog
            ICAgICAgIF0gPSAiIiJbXSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAg
            ICAgICIiInsiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlm
            dC1zdG9yYWdlLW15ZnMiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0i
            IiIKICAgICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAg
            ICAgICAgICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJb
            eyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsImtleSI6IkFRREZrYk5l
            ZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRBPT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIi
            LCAibWdyIjoiYWxsb3cgcnciLCAib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0q
            In19XSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2Fw
            cyI6IFsibW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1n
            ciIsICJhbGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIl0s
            ICJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpz
            b24iLCAicHJlZml4IjogImF1dGggY2FwcyJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmtiTmVmdDViRlJBQVRu
            ZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cgciwgIGFsbG93IGNv
            bW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciI6ImFsbG93IHJ3IiwgIm9zZCI6ImFsbG93IHJ3
            IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFw
            Wyd7ImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJzdGF0dXMifSddID0gY2VwaF9zdGF0dXNf
            c3RyCgogICAgZGVmIHNodXRkb3duKHNlbGYpOgogICAgICAgIHBhc3MKCiAgICBkZWYgZ2V0X2Zz
            aWQoc2VsZik6CiAgICAgICAgcmV0dXJuICJhZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5
            ZDBmMWMiCgogICAgZGVmIGNvbmZfcmVhZF9maWxlKHNlbGYpOgogICAgICAgIHBhc3MKCiAgICBk
            ZWYgY29ubmVjdChzZWxmKToKICAgICAgICBwYXNzCgogICAgZGVmIHBvb2xfZXhpc3RzKHNlbGYs
            IHBvb2xfbmFtZSk6CiAgICAgICAgcmV0dXJuIFRydWUKCiAgICBkZWYgbW9uX2NvbW1hbmQoc2Vs
            ZiwgY21kLCBvdXQpOgogICAgICAgIGpzb25fY21kID0ganNvbi5sb2FkcyhjbWQpCiAgICAgICAg
            anNvbl9jbWRfc3RyID0ganNvbi5kdW1wcyhqc29uX2NtZCwgc29ydF9rZXlzPVRydWUpCiAgICAg
            ICAgY21kX291dHB1dCA9IHNlbGYuY21kX291dHB1dF9tYXBbanNvbl9jbWRfc3RyXQogICAgICAg
            IHJldHVybiBzZWxmLnJldHVybl92YWwsIGNtZF9vdXRwdXQsIHN0cihzZWxmLmVycl9tZXNzYWdl
            LmVuY29kZSgidXRmLTgiKSkKCiAgICBkZWYgX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoc2VsZiwg
            aG9zdF9uYW1lKToKICAgICAgICBpcF9yZWdfeCA9IHJlLmNvbXBpbGUociJcZHsxLDN9LlxkezEs
            M30uXGR7MSwzfS5cZHsxLDN9IikKICAgICAgICAjIGlmIHByb3ZpZGVkIGhvc3QgaXMgZGlyZWN0
            bHkgYW4gSVAgYWRkcmVzcywgcmV0dXJuIHRoZSBzYW1lCiAgICAgICAgaWYgaXBfcmVnX3gubWF0
            Y2goaG9zdF9uYW1lKToKICAgICAgICAgICAgcmV0dXJuIGhvc3RfbmFtZQogICAgICAgIGltcG9y
            dCByYW5kb20KCiAgICAgICAgaG9zdF9pcCA9IHNlbGYuZHVtbXlfaG9zdF9pcF9tYXAuZ2V0KGhv
            c3RfbmFtZSwgIiIpCiAgICAgICAgaWYgbm90IGhvc3RfaXA6CiAgICAgICAgICAgIGhvc3RfaXAg
            PSBmIjE3Mi45LntyYW5kb20ucmFuZGludCgwLCAyNTQpfS57cmFuZG9tLnJhbmRpbnQoMCwgMjU0
            KX0iCiAgICAgICAgICAgIHNlbGYuZHVtbXlfaG9zdF9pcF9tYXBbaG9zdF9uYW1lXSA9IGhvc3Rf
            aXAKICAgICAgICBkZWwgcmFuZG9tCiAgICAgICAgcmV0dXJuIGhvc3RfaXAKCiAgICBAY2xhc3Nt
            ZXRob2QKICAgIGRlZiBSYWRvcyhjb25mZmlsZT1Ob25lKToKICAgICAgICByZXR1cm4gRHVtbXlS
            YWRvcygpCgoKY2xhc3MgUzNBdXRoKEF1dGhCYXNlKToKICAgICIiIkF0dGFjaGVzIEFXUyBBdXRo
            ZW50aWNhdGlvbiB0byB0aGUgZ2l2ZW4gUmVxdWVzdCBvYmplY3QuIiIiCgogICAgc2VydmljZV9i
            YXNlX3VybCA9ICJzMy5hbWF6b25hd3MuY29tIgoKICAgIGRlZiBfX2luaXRfXyhzZWxmLCBhY2Nl
            c3Nfa2V5LCBzZWNyZXRfa2V5LCBzZXJ2aWNlX3VybD1Ob25lKToKICAgICAgICBpZiBzZXJ2aWNl
            X3VybDoKICAgICAgICAgICAgc2VsZi5zZXJ2aWNlX2Jhc2VfdXJsID0gc2VydmljZV91cmwKICAg
            ICAgICBzZWxmLmFjY2Vzc19rZXkgPSBzdHIoYWNjZXNzX2tleSkKICAgICAgICBzZWxmLnNlY3Jl
            dF9rZXkgPSBzdHIoc2VjcmV0X2tleSkKCiAgICBkZWYgX19jYWxsX18oc2VsZiwgcik6CiAgICAg
            ICAgIyBDcmVhdGUgZGF0ZSBoZWFkZXIgaWYgaXQgaXMgbm90IGNyZWF0ZWQgeWV0LgogICAgICAg
            IGlmICJkYXRlIiBub3QgaW4gci5oZWFkZXJzIGFuZCAieC1hbXotZGF0ZSIgbm90IGluIHIuaGVh
            ZGVyczoKICAgICAgICAgICAgci5oZWFkZXJzWyJkYXRlIl0gPSBmb3JtYXRkYXRlKHRpbWV2YWw9
            Tm9uZSwgbG9jYWx0aW1lPUZhbHNlLCB1c2VnbXQ9VHJ1ZSkKICAgICAgICBzaWduYXR1cmUgPSBz
            ZWxmLmdldF9zaWduYXR1cmUocikKICAgICAgICBpZiBweTNrOgogICAgICAgICAgICBzaWduYXR1
            cmUgPSBzaWduYXR1cmUuZGVjb2RlKCJ1dGYtOCIpCiAgICAgICAgci5oZWFkZXJzWyJBdXRob3Jp
            emF0aW9uIl0gPSBmIkFXUyB7c2VsZi5hY2Nlc3Nfa2V5fTp7c2lnbmF0dXJlfSIKICAgICAgICBy
            ZXR1cm4gcgoKICAgIGRlZiBnZXRfc2lnbmF0dXJlKHNlbGYsIHIpOgogICAgICAgIGNhbm9uaWNh
            bF9zdHJpbmcgPSBzZWxmLmdldF9jYW5vbmljYWxfc3RyaW5nKHIudXJsLCByLmhlYWRlcnMsIHIu
            bWV0aG9kKQogICAgICAgIGlmIHB5M2s6CiAgICAgICAgICAgIGtleSA9IHNlbGYuc2VjcmV0X2tl
            eS5lbmNvZGUoInV0Zi04IikKICAgICAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZy5lbmNv
            ZGUoInV0Zi04IikKICAgICAgICBlbHNlOgogICAgICAgICAgICBrZXkgPSBzZWxmLnNlY3JldF9r
            ZXkKICAgICAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZwogICAgICAgIGggPSBobWFjLm5l
            dyhrZXksIG1zZywgZGlnZXN0bW9kPXNoYSkKICAgICAgICByZXR1cm4gZW5jb2Rlc3RyaW5nKGgu
            ZGlnZXN0KCkpLnN0cmlwKCkKCiAgICBkZWYgZ2V0X2Nhbm9uaWNhbF9zdHJpbmcoc2VsZiwgdXJs
            LCBoZWFkZXJzLCBtZXRob2QpOgogICAgICAgIHBhcnNlZHVybCA9IHVybHBhcnNlKHVybCkKICAg
            ICAgICBvYmplY3RrZXkgPSBwYXJzZWR1cmwucGF0aFsxOl0KCiAgICAgICAgYnVja2V0ID0gcGFy
            c2VkdXJsLm5ldGxvY1s6IC1sZW4oc2VsZi5zZXJ2aWNlX2Jhc2VfdXJsKV0KICAgICAgICBpZiBs
            ZW4oYnVja2V0KSA+IDE6CiAgICAgICAgICAgICMgcmVtb3ZlIGxhc3QgZG90CiAgICAgICAgICAg
            IGJ1Y2tldCA9IGJ1Y2tldFs6LTFdCgogICAgICAgIGludGVyZXN0aW5nX2hlYWRlcnMgPSB7ImNv
            bnRlbnQtbWQ1IjogIiIsICJjb250ZW50LXR5cGUiOiAiIiwgImRhdGUiOiAiIn0KICAgICAgICBm
            b3Iga2V5IGluIGhlYWRlcnM6CiAgICAgICAgICAgIGxrID0ga2V5Lmxvd2VyKCkKICAgICAgICAg
            ICAgdHJ5OgogICAgICAgICAgICAgICAgbGsgPSBsay5kZWNvZGUoInV0Zi04IikKICAgICAgICAg
            ICAgZXhjZXB0OgogICAgICAgICAgICAgICAgcGFzcwogICAgICAgICAgICBpZiBoZWFkZXJzW2tl
            eV0gYW5kICgKICAgICAgICAgICAgICAgIGxrIGluIGludGVyZXN0aW5nX2hlYWRlcnMua2V5cygp
            IG9yIGxrLnN0YXJ0c3dpdGgoIngtYW16LSIpCiAgICAgICAgICAgICk6CiAgICAgICAgICAgICAg
            ICBpbnRlcmVzdGluZ19oZWFkZXJzW2xrXSA9IGhlYWRlcnNba2V5XS5zdHJpcCgpCgogICAgICAg
            ICMgSWYgeC1hbXotZGF0ZSBpcyB1c2VkIGl0IHN1cGVyc2VkZXMgdGhlIGRhdGUgaGVhZGVyLgog
            ICAgICAgIGlmIG5vdCBweTNrOgogICAgICAgICAgICBpZiAieC1hbXotZGF0ZSIgaW4gaW50ZXJl
            c3RpbmdfaGVhZGVyczoKICAgICAgICAgICAgICAgIGludGVyZXN0aW5nX2hlYWRlcnNbImRhdGUi
            XSA9ICIiCiAgICAgICAgZWxzZToKICAgICAgICAgICAgaWYgIngtYW16LWRhdGUiIGluIGludGVy
            ZXN0aW5nX2hlYWRlcnM6CiAgICAgICAgICAgICAgICBpbnRlcmVzdGluZ19oZWFkZXJzWyJkYXRl
            Il0gPSAiIgoKICAgICAgICBidWYgPSBmInttZXRob2R9XG4iCiAgICAgICAgZm9yIGtleSBpbiBz
            b3J0ZWQoaW50ZXJlc3RpbmdfaGVhZGVycy5rZXlzKCkpOgogICAgICAgICAgICB2YWwgPSBpbnRl
            cmVzdGluZ19oZWFkZXJzW2tleV0KICAgICAgICAgICAgaWYga2V5LnN0YXJ0c3dpdGgoIngtYW16
            LSIpOgogICAgICAgICAgICAgICAgYnVmICs9IGYie2tleX06e3ZhbH1cbiIKICAgICAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgICAgIGJ1ZiArPSBmInt2YWx9XG4iCgogICAgICAgICMgYXBwZW5k
            IHRoZSBidWNrZXQgaWYgaXQgZXhpc3RzCiAgICAgICAgaWYgYnVja2V0ICE9ICIiOgogICAgICAg
            ICAgICBidWYgKz0gZiIve2J1Y2tldH0iCgogICAgICAgICMgYWRkIHRoZSBvYmplY3RrZXkuIGV2
            ZW4gaWYgaXQgZG9lc24ndCBleGlzdCwgYWRkIHRoZSBzbGFzaAogICAgICAgIGJ1ZiArPSBmIi97
            b2JqZWN0a2V5fSIKCiAgICAgICAgcmV0dXJuIGJ1ZgoKCmNsYXNzIFJhZG9zSlNPTjoKICAgIEVY
            VEVSTkFMX1VTRVJfTkFNRSA9ICJjbGllbnQuaGVhbHRoY2hlY2tlciIKICAgIEVYVEVSTkFMX1JH
            V19BRE1JTl9PUFNfVVNFUl9OQU1FID0gInJndy1hZG1pbi1vcHMtdXNlciIKICAgIEVNUFRZX09V
            VFBVVF9MSVNUID0gIkVtcHR5IG91dHB1dCBsaXN0IgogICAgREVGQVVMVF9SR1dfUE9PTF9QUkVG
            SVggPSAiZGVmYXVsdCIKICAgIERFRkFVTFRfTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUID0gIjky
            ODMiCgogICAgQGNsYXNzbWV0aG9kCiAgICBkZWYgZ2VuX2FyZ19wYXJzZXIoY2xzLCBhcmdzX3Rv
            X3BhcnNlPU5vbmUpOgogICAgICAgIGFyZ1AgPSBhcmdwYXJzZS5Bcmd1bWVudFBhcnNlcigpCgog
            ICAgICAgIGNvbW1vbl9ncm91cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJjb21tb24iKQog
            ICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoIi0tdmVyYm9zZSIsICItdiIsIGFjdGlv
            bj0ic3RvcmVfdHJ1ZSIsIGRlZmF1bHQ9RmFsc2UpCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9h
            cmd1bWVudCgKICAgICAgICAgICAgIi0tY2VwaC1jb25mIiwgIi1jIiwgaGVscD0iUHJvdmlkZSBh
            IGNlcGggY29uZiBmaWxlLiIsIHR5cGU9c3RyCiAgICAgICAgKQogICAgICAgIGNvbW1vbl9ncm91
            cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLWtleXJpbmciLCAiLWsiLCBoZWxwPSJQYXRo
            IHRvIGNlcGgga2V5cmluZyBmaWxlLiIsIHR5cGU9c3RyCiAgICAgICAgKQogICAgICAgIGNvbW1v
            bl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJ1bi1hcy11c2VyIiwKICAgICAg
            ICAgICAgIi11IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgdHlwZT1zdHIs
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIGEgdXNlciBuYW1lIHRvIGNoZWNrIHRoZSBjbHVz
            dGVyJ3MgaGVhbHRoIHN0YXR1cywgbXVzdCBiZSBwcmVmaXhlZCBieSAnY2xpZW50LiciLAogICAg
            ICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1j
            bHVzdGVyLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJL
            dWJlcm5ldGVzIGNsdXN0ZXIgbmFtZShsZWdhY3kgZmxhZyksIE5vdGU6IEVpdGhlciB1c2UgdGhp
            cyBvciAtLWs4cy1jbHVzdGVyLW5hbWUiLAogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAu
            YWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1rOHMtY2x1c3Rlci1uYW1lIiwgZGVmYXVsdD0i
            IiwgaGVscD0iS3ViZXJuZXRlcyBjbHVzdGVyIG5hbWUiCiAgICAgICAgKQogICAgICAgIGNvbW1v
            bl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLW5hbWVzcGFjZSIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIGhlbHA9Ik5hbWVzcGFjZSB3aGVyZSBDZXBoQ2x1
            c3RlciBpcyBydW5uaW5nIiwKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tcmd3LXBvb2wtcHJlZml4IiwgZGVmYXVsdD0iIiwgaGVscD0i
            UkdXIFBvb2wgcHJlZml4IgogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3Vt
            ZW50KAogICAgICAgICAgICAiLS1yZXN0cmljdGVkLWF1dGgtcGVybWlzc2lvbiIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlJlc3RyaWN0IGNlcGhDU0lLZXly
            aW5ncyBhdXRoIHBlcm1pc3Npb25zIHRvIHNwZWNpZmljIHBvb2xzLCBjbHVzdGVyLiIKICAgICAg
            ICAgICAgKyAiTWFuZGF0b3J5IGZsYWdzIHRoYXQgbmVlZCB0byBiZSBzZXQgYXJlIC0tcmJkLWRh
            dGEtcG9vbC1uYW1lLCBhbmQgLS1rOHMtY2x1c3Rlci1uYW1lLiIKICAgICAgICAgICAgKyAiLS1j
            ZXBoZnMtZmlsZXN5c3RlbS1uYW1lIGZsYWcgY2FuIGFsc28gYmUgcGFzc2VkIGluIGNhc2Ugb2Yg
            Y2VwaGZzIHVzZXIgcmVzdHJpY3Rpb24sIHNvIGl0IGNhbiByZXN0cmljdCB1c2VyIHRvIHBhcnRp
            Y3VsYXIgY2VwaGZzIGZpbGVzeXN0ZW0iCiAgICAgICAgICAgICsgInNhbXBsZSBydW46IGBweXRo
            b24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwtY2x1c3Rlci1yZXNvdXJjZXMucHkgLS1jZXBo
            ZnMtZmlsZXN5c3RlbS1uYW1lIG15ZnMgLS1yYmQtZGF0YS1wb29sLW5hbWUgcmVwbGljYXBvb2wg
            LS1rOHMtY2x1c3Rlci1uYW1lIHJvb2tzdG9yYWdlIC0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Np
            b24gdHJ1ZWAiCiAgICAgICAgICAgICsgIk5vdGU6IFJlc3RyaWN0aW5nIHRoZSBjc2ktdXNlcnMg
            cGVyIHBvb2wsIGFuZCBwZXIgY2x1c3RlciB3aWxsIHJlcXVpcmUgY3JlYXRpbmcgbmV3IGNzaS11
            c2VycyBhbmQgbmV3IHNlY3JldHMgZm9yIHRoYXQgY3NpLXVzZXJzLiIKICAgICAgICAgICAgKyAi
            U28gYXBwbHkgdGhlc2Ugc2VjcmV0cyBvbmx5IHRvIG5ldyBgQ29uc3VtZXIgY2x1c3RlcmAgZGVw
            bG95bWVudCB3aGlsZSB1c2luZyB0aGUgc2FtZSBgU291cmNlIGNsdXN0ZXJgLiIsCiAgICAgICAg
            KQogICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXYyLXBv
            cnQtZW5hYmxlIiwKICAgICAgICAgICAgYWN0aW9uPSJzdG9yZV90cnVlIiwKICAgICAgICAgICAg
            ZGVmYXVsdD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iRW5hYmxlIHYyIG1vbiBwb3J0KDMzMDAp
            IGZvciBtb25zIiwKICAgICAgICApCgogICAgICAgIG91dHB1dF9ncm91cCA9IGFyZ1AuYWRkX2Fy
            Z3VtZW50X2dyb3VwKCJvdXRwdXQiKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQo
            CiAgICAgICAgICAgICItLWZvcm1hdCIsCiAgICAgICAgICAgICItdCIsCiAgICAgICAgICAgIGNo
            b2ljZXM9WyJqc29uIiwgImJhc2giXSwKICAgICAgICAgICAgZGVmYXVsdD0ianNvbiIsCiAgICAg
            ICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBvdXRwdXQgZm9ybWF0IChqc29uIHwgYmFzaCkiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1vdXRwdXQiLAogICAgICAgICAgICAiLW8iLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAg
            ICAgICAgICBoZWxwPSJPdXRwdXQgd2lsbCBiZSBzdG9yZWQgaW50byB0aGUgcHJvdmlkZWQgZmls
            ZSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAog
            ICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgQ2VwaCBmaWxlc3lzdGVt
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tY2VwaGZzLW1ldGFkYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIs
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBjZXBoZnMgbWV0YWRh
            dGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAg
            ICAgICAgICAgICItLWNlcGhmcy1kYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9
            IiIsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBjZXBoZnMgZGF0
            YSBwb29sIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAg
            ICAgICAgICAgIi0tcmJkLWRhdGEtcG9vbC1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwK
            ICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRo
            ZSBuYW1lIG9mIHRoZSBSQkQgZGF0YXBvb2wiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3Jv
            dXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1hbGlhcy1yYmQtZGF0YS1wb29sLW5hbWUi
            LAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAg
            ICAgICAgICAgaGVscD0iUHJvdmlkZXMgYW4gYWxpYXMgZm9yIHRoZSAgUkJEIGRhdGEgcG9vbCBu
            YW1lLCBuZWNlc3NhcnkgaWYgYSBzcGVjaWFsIGNoYXJhY3RlciBpcyBwcmVzZW50IGluIHRoZSBw
            b29sIG5hbWUgc3VjaCBhcyBhIHBlcmlvZCBvciB1bmRlcnNjb3JlIiwKICAgICAgICApCiAgICAg
            ICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmd3LWVuZHBvaW50
            IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAg
            ICAgICAgICAgIGhlbHA9IlJBRE9TIEdhdGV3YXkgZW5kcG9pbnQgKGluIGA8SVB2ND46PFBPUlQ+
            YCBvciBgPFtJUHY2XT46PFBPUlQ+YCBvciBgPEZRRE4+OjxQT1JUPmAgZm9ybWF0KSIsCiAgICAg
            ICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJn
            dy10bHMtY2VydC1wYXRoIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVx
            dWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlJBRE9TIEdhdGV3YXkgZW5kcG9pbnQgVExT
            IGNlcnRpZmljYXRlIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVu
            dCgKICAgICAgICAgICAgIi0tcmd3LXNraXAtdGxzIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFs
            c2UsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9Iklnbm9yZSBU
            TFMgY2VydGlmaWNhdGlvbiB2YWxpZGF0aW9uIHdoZW4gYSBzZWxmLXNpZ25lZCBjZXJ0aWZpY2F0
            ZSBpcyBwcm92aWRlZCAoTk9UIFJFQ09NTUVOREVEIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tbW9uaXRvcmluZy1lbmRwb2ludCIs
            CiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAg
            ICAgICAgICBoZWxwPSJDZXBoIE1hbmFnZXIgcHJvbWV0aGV1cyBleHBvcnRlciBlbmRwb2ludHMg
            KGNvbW1hIHNlcGFyYXRlZCBsaXN0IG9mIChmb3JtYXQgYDxJUHY0PmAgb3IgYDxbSVB2Nl0+YCBv
            ciBgPEZRRE4+YCkgZW50cmllcyBvZiBhY3RpdmUgYW5kIHN0YW5kYnkgbWdycykiLAogICAgICAg
            ICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1tb25p
            dG9yaW5nLWVuZHBvaW50LXBvcnQiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAg
            ICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iQ2VwaCBNYW5hZ2VyIHByb21ldGhl
            dXMgZXhwb3J0ZXIgcG9ydCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJn
            dW1lbnQoCiAgICAgICAgICAgICItLXNraXAtbW9uaXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAg
            ICAgICAgIGhlbHA9IkRvIG5vdCBjaGVjayBmb3IgYSBtb25pdG9yaW5nIGVuZHBvaW50IGZvciB0
            aGUgQ2VwaCBjbHVzdGVyIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tcmJkLW1ldGFkYXRhLWVjLXBvb2wtbmFtZSIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBo
            ZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiBlcmFzdXJlIGNvZGVkIFJCRCBtZXRhZGF0YSBwb29s
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tZHJ5LXJ1biIsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFj
            dGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGhlbHA9IkRyeSBydW4gcHJpbnRzIHRoZSBl
            eGVjdXRlZCBjb21tYW5kcyB3aXRob3V0IHJ1bm5pbmcgdGhlbSIsCiAgICAgICAgKQogICAgICAg
            IG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJhZG9zLW5hbWVzcGFj
            ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAog
            ICAgICAgICAgICBoZWxwPSJEaXZpZGVzIGEgcG9vbCBpbnRvIHNlcGFyYXRlIGxvZ2ljYWwgbmFt
            ZXNwYWNlcywgdXNlZCBmb3IgY3JlYXRpbmcgUkJEIFBWQyBpbiBhIENlcGhCbG9ja1Bvb2xSYWRv
            c05hbWVzcGFjZSAoc2hvdWxkIGJlIGxvd2VyIGNhc2UpIiwKICAgICAgICApCiAgICAgICAgb3V0
            cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tc3Vidm9sdW1lLWdyb3VwIiwK
            ICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAg
            ICAgICAgIGhlbHA9InByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBzdWJ2b2x1bWUgZ3JvdXAiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1yZ3ctcmVhbG0tbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJl
            cXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUg
            cmd3LXJlYWxtIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgK
            ICAgICAgICAgICAgIi0tcmd3LXpvbmUtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAg
            ICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUg
            bmFtZSBvZiB0aGUgcmd3LXpvbmUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS1yZ3ctem9uZWdyb3VwLW5hbWUiLAogICAgICAgICAg
            ICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVs
            cD0icHJvdmlkZXMgdGhlIG5hbWUgb2YgdGhlIHJndy16b25lZ3JvdXAiLAogICAgICAgICkKICAg
            ICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS10b3BvbG9neS1w
            b29scyIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNl
            LAogICAgICAgICAgICBoZWxwPSJjb21tYS1zZXBhcmF0ZWQgbGlzdCBvZiB0b3BvbG9neS1jb25z
            dHJhaW5lZCByYmQgcG9vbHMiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2Fy
            Z3VtZW50KAogICAgICAgICAgICAiLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi1sYWJlbCIsCiAg
            ICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAg
            ICAgICBoZWxwPSJrOHMgY2x1c3RlciBmYWlsdXJlIGRvbWFpbiBsYWJlbCAoZXhhbXBsZTogem9u
            ZSwgcmFjaywgb3IgaG9zdCkgZm9yIHRoZSB0b3BvbG9neS1wb29scyB0aGF0IG1hdGNoIHRoZSBj
            ZXBoIGRvbWFpbiIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQo
            CiAgICAgICAgICAgICItLXRvcG9sb2d5LWZhaWx1cmUtZG9tYWluLXZhbHVlcyIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBo
            ZWxwPSJjb21tYS1zZXBhcmF0ZWQgbGlzdCBvZiB0aGUgazhzIGNsdXN0ZXIgZmFpbHVyZSBkb21h
            aW4gdmFsdWVzIGNvcnJlc3BvbmRpbmcgdG8gZWFjaCBvZiB0aGUgcG9vbHMgaW4gdGhlIGB0b3Bv
            bG9neS1wb29sc2AgbGlzdCIsCiAgICAgICAgKQoKICAgICAgICB1cGdyYWRlX2dyb3VwID0gYXJn
            UC5hZGRfYXJndW1lbnRfZ3JvdXAoInVwZ3JhZGUiKQogICAgICAgIHVwZ3JhZGVfZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS11cGdyYWRlIiwKICAgICAgICAgICAgYWN0aW9uPSJz
            dG9yZV90cnVlIiwKICAgICAgICAgICAgZGVmYXVsdD1GYWxzZSwKICAgICAgICAgICAgaGVscD0i
            VXBncmFkZXMgdGhlIGNlcGhDU0lLZXlyaW5ncyhGb3IgZXhhbXBsZTogY2xpZW50LmNzaS1jZXBo
            ZnMtcHJvdmlzaW9uZXIpIGFuZCBjbGllbnQuaGVhbHRoY2hlY2tlciBjZXBoIHVzZXJzIHdpdGgg
            bmV3IHBlcm1pc3Npb25zIG5lZWRlZCBmb3IgdGhlIG5ldyBjbHVzdGVyIHZlcnNpb24gYW5kIG9s
            ZGVyIHBlcm1pc3Npb24gd2lsbCBzdGlsbCBiZSBhcHBsaWVkLiIKICAgICAgICAgICAgKyAiU2Ft
            cGxlIHJ1bjogYHB5dGhvbjMgL2V0Yy9jZXBoL2NyZWF0ZS1leHRlcm5hbC1jbHVzdGVyLXJlc291
            cmNlcy5weSAtLXVwZ3JhZGVgLCB0aGlzIHdpbGwgdXBncmFkZSBhbGwgdGhlIGRlZmF1bHQgY3Np
            IHVzZXJzKG5vbi1yZXN0cmljdGVkKSIKICAgICAgICAgICAgKyAiRm9yIHJlc3RyaWN0ZWQgdXNl
            cnMoRm9yIGV4YW1wbGU6IGNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1z
            dG9yYWdlLW15ZnMpLCB1c2VycyBjcmVhdGVkIHVzaW5nIC0tcmVzdHJpY3RlZC1hdXRoLXBlcm1p
            c3Npb24gZmxhZyBuZWVkIHRvIHBhc3MgbWFuZGF0b3J5IGZsYWdzIgogICAgICAgICAgICArICJt
            YW5kYXRvcnkgZmxhZ3M6ICctLXJiZC1kYXRhLXBvb2wtbmFtZSwgLS1rOHMtY2x1c3Rlci1uYW1l
            IGFuZCAtLXJ1bi1hcy11c2VyJyBmbGFncyB3aGlsZSB1cGdyYWRpbmciCiAgICAgICAgICAgICsg
            ImluIGNhc2Ugb2YgY2VwaGZzIHVzZXJzIGlmIHlvdSBoYXZlIHBhc3NlZCAtLWNlcGhmcy1maWxl
            c3lzdGVtLW5hbWUgZmxhZyB3aGlsZSBjcmVhdGluZyB1c2VyIHRoZW4gd2hpbGUgdXBncmFkaW5n
            IGl0IHdpbGwgYmUgbWFuZGF0b3J5IHRvbyIKICAgICAgICAgICAgKyAiU2FtcGxlIHJ1bjogYHB5
            dGhvbjMgL2V0Yy9jZXBoL2NyZWF0ZS1leHRlcm5hbC1jbHVzdGVyLXJlc291cmNlcy5weSAtLXVw
            Z3JhZGUgLS1yYmQtZGF0YS1wb29sLW5hbWUgcmVwbGljYXBvb2wgLS1rOHMtY2x1c3Rlci1uYW1l
            IHJvb2tzdG9yYWdlICAtLXJ1bi1hcy11c2VyIGNsaWVudC5jc2ktcmJkLW5vZGUtcm9va3N0b3Jh
            Z2UtcmVwbGljYXBvb2xgIgogICAgICAgICAgICArICJQUzogQW4gZXhpc3Rpbmcgbm9uLXJlc3Ry
            aWN0ZWQgdXNlciBjYW5ub3QgYmUgY29udmVydGVkIHRvIGEgcmVzdHJpY3RlZCB1c2VyIGJ5IHVw
            Z3JhZGluZy4iCiAgICAgICAgICAgICsgIlVwZ3JhZGUgZmxhZyBzaG91bGQgb25seSBiZSB1c2Vk
            IHRvIGFwcGVuZCBuZXcgcGVybWlzc2lvbnMgdG8gdXNlcnMsIGl0IHNob3VsZG4ndCBiZSB1c2Vk
            IGZvciBjaGFuZ2luZyB1c2VyIGFscmVhZHkgYXBwbGllZCBwZXJtaXNzaW9uLCBmb3IgZXhhbXBs
            ZSB5b3Ugc2hvdWxkbid0IGNoYW5nZSBpbiB3aGljaCBwb29sIHVzZXIgaGFzIGFjY2VzcyIsCiAg
            ICAgICAgKQoKICAgICAgICBpZiBhcmdzX3RvX3BhcnNlOgogICAgICAgICAgICBhc3NlcnQgKAog
            ICAgICAgICAgICAgICAgdHlwZShhcmdzX3RvX3BhcnNlKSA9PSBsaXN0CiAgICAgICAgICAgICks
            ICJBcmd1bWVudCB0byAnZ2VuX2FyZ19wYXJzZXInIHNob3VsZCBiZSBhIGxpc3QiCiAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgYXJnc190b19wYXJzZSA9IHN5cy5hcmd2WzE6XQogICAgICAgIHJl
            dHVybiBhcmdQLnBhcnNlX2FyZ3MoYXJnc190b19wYXJzZSkKCiAgICBkZWYgdmFsaWRhdGVfcmJk
            X21ldGFkYXRhX2VjX3Bvb2xfbmFtZShzZWxmKToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2Vy
            LnJiZF9tZXRhZGF0YV9lY19wb29sX25hbWU6CiAgICAgICAgICAgIHJiZF9tZXRhZGF0YV9lY19w
            b29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUKICAg
            ICAgICAgICAgcmJkX3Bvb2xfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9u
            YW1lCgogICAgICAgICAgICBpZiByYmRfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiRmxh
            ZyAnLS1yYmQtZGF0YS1wb29sLW5hbWUnIHNob3VsZCBub3QgYmUgZW1wdHkiCiAgICAgICAgICAg
            ICAgICApCgogICAgICAgICAgICBpZiByYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lID09ICIiOgog
            ICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAg
            ICAgICAgICAgICAiRmxhZyAnLS1yYmQtbWV0YWRhdGEtZWMtcG9vbC1uYW1lJyBzaG91bGQgbm90
            IGJlIGVtcHR5IgogICAgICAgICAgICAgICAgKQoKICAgICAgICAgICAgY21kX2pzb24gPSB7InBy
            ZWZpeCI6ICJvc2QgZHVtcCIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgICAgIHJldF92YWws
            IGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikK
            ICAgICAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAg
            ICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAg
            ICAgICAgZiJ7Y21kX2pzb25bJ3ByZWZpeCddfSBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAg
            ICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVN
            UFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgbWV0YWRhdGFf
            cG9vbF9leGlzdCwgcG9vbF9leGlzdCA9IEZhbHNlLCBGYWxzZQoKICAgICAgICAgICAgZm9yIGtl
            eSBpbiBqc29uX291dFsicG9vbHMiXToKICAgICAgICAgICAgICAgICMgaWYgZXJhc3VyZV9jb2Rl
            X3Byb2ZpbGUgaXMgZW1wdHkgYW5kIHBvb2wgbmFtZSBleGlzdHMgdGhlbiBpdCByZXBsaWNhIHBv
            b2wKICAgICAgICAgICAgICAgIGlmICgKICAgICAgICAgICAgICAgICAgICBrZXlbImVyYXN1cmVf
            Y29kZV9wcm9maWxlIl0gPT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQga2V5WyJwb29sX25h
            bWUiXSA9PSByYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lCiAgICAgICAgICAgICAgICApOgogICAg
            ICAgICAgICAgICAgICAgIG1ldGFkYXRhX3Bvb2xfZXhpc3QgPSBUcnVlCiAgICAgICAgICAgICAg
            ICAjIGlmIGVyYXN1cmVfY29kZV9wcm9maWxlIGlzIG5vdCBlbXB0eSBhbmQgcG9vbCBuYW1lIGV4
            aXN0cyB0aGVuIGl0IGlzIGVjIHBvb2wKICAgICAgICAgICAgICAgIGlmIGtleVsiZXJhc3VyZV9j
            b2RlX3Byb2ZpbGUiXSBhbmQga2V5WyJwb29sX25hbWUiXSA9PSByYmRfcG9vbF9uYW1lOgogICAg
            ICAgICAgICAgICAgICAgIHBvb2xfZXhpc3QgPSBUcnVlCgogICAgICAgICAgICBpZiBub3QgbWV0
            YWRhdGFfcG9vbF9leGlzdDoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVF
            eGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIlByb3ZpZGVkIHJiZF9lY19tZXRhZGF0YV9w
            b29sIG5hbWUsIgogICAgICAgICAgICAgICAgICAgIGYiIHtyYmRfbWV0YWRhdGFfZWNfcG9vbF9u
            YW1lfSwgZG9lcyBub3QgZXhpc3QiCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmIG5v
            dCBwb29sX2V4aXN0OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICBmIlByb3ZpZGVkIHJiZF9kYXRhX3Bvb2wgbmFtZSwg
            e3JiZF9wb29sX25hbWV9LCBkb2VzIG5vdCBleGlzdCIKICAgICAgICAgICAgICAgICkKICAgICAg
            ICAgICAgcmV0dXJuIHJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUKCiAgICBkZWYgZHJ5X3J1bihz
            ZWxmLCBtc2cpOgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAg
            ICAgcHJpbnQoIkV4ZWN1dGU6ICIgKyAiJyIgKyBtc2cgKyAiJyIpCgogICAgZGVmIHZhbGlkYXRl
            X3Jnd19lbmRwb2ludF90bHNfY2VydChzZWxmKToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd190bHNfY2VydF9wYXRoOgogICAgICAgICAgICB3aXRoIG9wZW4oc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfdGxzX2NlcnRfcGF0aCwgZW5jb2Rpbmc9InV0ZjgiKSBhcyBmOgogICAgICAgICAgICAg
            ICAgY29udGVudHMgPSBmLnJlYWQoKQogICAgICAgICAgICAgICAgcmV0dXJuIGNvbnRlbnRzLnJz
            dHJpcCgpCgogICAgZGVmIF9jaGVja19jb25mbGljdGluZ19vcHRpb25zKHNlbGYpOgogICAgICAg
            IGlmIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnVwZ3JhZGUgYW5kIG5vdCBzZWxmLl9hcmdfcGFyc2Vy
            LnJiZF9kYXRhX3Bvb2xfbmFtZToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICJFaXRoZXIgJy0tdXBncmFkZScgb3IgJy0tcmJkLWRh
            dGEtcG9vbC1uYW1lIDxwb29sX25hbWU+JyBzaG91bGQgYmUgc3BlY2lmaWVkIgogICAgICAgICAg
            ICApCgogICAgZGVmIF9pbnZhbGlkX2VuZHBvaW50KHNlbGYsIGVuZHBvaW50X3N0cik6CiAgICAg
            ICAgIyBleHRyYWN0IHRoZSBwb3J0IGJ5IGdldHRpbmcgdGhlIGxhc3Qgc3BsaXQgb24gYDpgIGRl
            bGltaXRlcgogICAgICAgIHRyeToKICAgICAgICAgICAgZW5kcG9pbnRfc3RyX2lwLCBwb3J0ID0g
            ZW5kcG9pbnRfc3RyLnJzcGxpdCgiOiIsIDEpCiAgICAgICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAg
            ICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oZiJOb3QgYSBwcm9wZXIg
            ZW5kcG9pbnQ6IHtlbmRwb2ludF9zdHJ9IikKCiAgICAgICAgdHJ5OgogICAgICAgICAgICBpZiBl
            bmRwb2ludF9zdHJfaXBbMF0gPT0gIlsiOgogICAgICAgICAgICAgICAgZW5kcG9pbnRfc3RyX2lw
            ID0gZW5kcG9pbnRfc3RyX2lwWzEgOiBsZW4oZW5kcG9pbnRfc3RyX2lwKSAtIDFdCiAgICAgICAg
            ICAgIGlwX3R5cGUgPSAoCiAgICAgICAgICAgICAgICAiSVB2NCIgaWYgdHlwZShpcF9hZGRyZXNz
            KGVuZHBvaW50X3N0cl9pcCkpIGlzIElQdjRBZGRyZXNzIGVsc2UgIklQdjYiCiAgICAgICAgICAg
            ICkKICAgICAgICBleGNlcHQgVmFsdWVFcnJvcjoKICAgICAgICAgICAgaXBfdHlwZSA9ICJGUURO
            IgogICAgICAgIGlmIG5vdCBwb3J0LmlzZGlnaXQoKToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0
            aW9uRmFpbHVyZUV4Y2VwdGlvbihmIlBvcnQgbm90IHZhbGlkOiB7cG9ydH0iKQogICAgICAgIGlu
            dFBvcnQgPSBpbnQocG9ydCkKICAgICAgICBpZiBpbnRQb3J0IDwgMSBvciBpbnRQb3J0ID4gMioq
            MTYgLSAxOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKGYiT3V0
            IG9mIHJhbmdlIHBvcnQgbnVtYmVyOiB7cG9ydH0iKQoKICAgICAgICByZXR1cm4gaXBfdHlwZQoK
            ICAgIGRlZiBlbmRwb2ludF9kaWFsKHNlbGYsIGVuZHBvaW50X3N0ciwgaXBfdHlwZSwgdGltZW91
            dD0zLCBjZXJ0PU5vbmUpOgogICAgICAgICMgaWYgdGhlICdjbHVzdGVyJyBpbnN0YW5jZSBpcyBh
            IGR1bW15IG9uZSwKICAgICAgICAjIGRvbid0IHRyeSB0byByZWFjaCBvdXQgdG8gdGhlIGVuZHBv
            aW50CiAgICAgICAgaWYgaXNpbnN0YW5jZShzZWxmLmNsdXN0ZXIsIER1bW15UmFkb3MpOgogICAg
            ICAgICAgICByZXR1cm4gIiIsICIiLCAiIgogICAgICAgIGlmIGlwX3R5cGUgPT0gIklQdjYiOgog
            ICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBlbmRwb2ludF9zdHJfaXAsIGVuZHBvaW50
            X3N0cl9wb3J0ID0gZW5kcG9pbnRfc3RyLnJzcGxpdCgiOiIsIDEpCiAgICAgICAgICAgIGV4Y2Vw
            dCBWYWx1ZUVycm9yOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICBmIk5vdCBhIHByb3BlciBlbmRwb2ludDoge2VuZHBv
            aW50X3N0cn0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmIGVuZHBvaW50X3N0cl9p
            cFswXSAhPSAiWyI6CiAgICAgICAgICAgICAgICBlbmRwb2ludF9zdHJfaXAgPSAiWyIgKyBlbmRw
            b2ludF9zdHJfaXAgKyAiXSIKICAgICAgICAgICAgZW5kcG9pbnRfc3RyID0gIjoiLmpvaW4oW2Vu
            ZHBvaW50X3N0cl9pcCwgZW5kcG9pbnRfc3RyX3BvcnRdKQoKICAgICAgICBwcm90b2NvbHMgPSBb
            Imh0dHAiLCAiaHR0cHMiXQogICAgICAgIHJlc3BvbnNlX2Vycm9yID0gTm9uZQogICAgICAgIGZv
            ciBwcmVmaXggaW4gcHJvdG9jb2xzOgogICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBl
            cCA9IGYie3ByZWZpeH06Ly97ZW5kcG9pbnRfc3RyfSIKICAgICAgICAgICAgICAgIHZlcmlmeSA9
            IE5vbmUKICAgICAgICAgICAgICAgICMgSWYgdmVyaWZ5IGlzIHNldCB0byBhIHBhdGggdG8gYSBk
            aXJlY3RvcnksCiAgICAgICAgICAgICAgICAjIHRoZSBkaXJlY3RvcnkgbXVzdCBoYXZlIGJlZW4g
            cHJvY2Vzc2VkIHVzaW5nIHRoZSBjX3JlaGFzaCB1dGlsaXR5IHN1cHBsaWVkIHdpdGggT3BlblNT
            TC4KICAgICAgICAgICAgICAgIGlmIHByZWZpeCA9PSAiaHR0cHMiIGFuZCBzZWxmLl9hcmdfcGFy
            c2VyLnJnd19za2lwX3RsczoKICAgICAgICAgICAgICAgICAgICB2ZXJpZnkgPSBGYWxzZQogICAg
            ICAgICAgICAgICAgICAgIHIgPSByZXF1ZXN0cy5oZWFkKGVwLCB0aW1lb3V0PXRpbWVvdXQsIHZl
            cmlmeT1GYWxzZSkKICAgICAgICAgICAgICAgIGVsaWYgcHJlZml4ID09ICJodHRwcyIgYW5kIGNl
            cnQ6CiAgICAgICAgICAgICAgICAgICAgdmVyaWZ5ID0gY2VydAogICAgICAgICAgICAgICAgICAg
            IHIgPSByZXF1ZXN0cy5oZWFkKGVwLCB0aW1lb3V0PXRpbWVvdXQsIHZlcmlmeT1jZXJ0KQogICAg
            ICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgICAgICByID0gcmVxdWVzdHMuaGVhZChl
            cCwgdGltZW91dD10aW1lb3V0KQogICAgICAgICAgICAgICAgaWYgci5zdGF0dXNfY29kZSA9PSAy
            MDA6CiAgICAgICAgICAgICAgICAgICAgcmV0dXJuIHByZWZpeCwgdmVyaWZ5LCAiIgogICAgICAg
            ICAgICBleGNlcHQgRXhjZXB0aW9uIGFzIGVycjoKICAgICAgICAgICAgICAgIHJlc3BvbnNlX2Vy
            cm9yID0gZXJyCiAgICAgICAgICAgICAgICBjb250aW51ZQogICAgICAgIHN5cy5zdGRlcnIud3Jp
            dGUoCiAgICAgICAgICAgIGYidW5hYmxlIHRvIGNvbm5lY3QgdG8gZW5kcG9pbnQ6IHtlbmRwb2lu
            dF9zdHJ9LCBmYWlsZWQgZXJyb3I6IHtyZXNwb25zZV9lcnJvcn0iCiAgICAgICAgKQogICAgICAg
            IHJldHVybiAoCiAgICAgICAgICAgICIiLAogICAgICAgICAgICAiIiwKICAgICAgICAgICAgKCIt
            MSIpLAogICAgICAgICkKCiAgICBkZWYgX19pbml0X18oc2VsZiwgYXJnX2xpc3Q9Tm9uZSk6CiAg
            ICAgICAgc2VsZi5vdXRfbWFwID0ge30KICAgICAgICBzZWxmLl9leGNsdWRlZF9rZXlzID0gc2V0
            KCkKICAgICAgICBzZWxmLl9hcmdfcGFyc2VyID0gc2VsZi5nZW5fYXJnX3BhcnNlcihhcmdzX3Rv
            X3BhcnNlPWFyZ19saXN0KQogICAgICAgIHNlbGYuX2NoZWNrX2NvbmZsaWN0aW5nX29wdGlvbnMo
            KQogICAgICAgIHNlbGYucnVuX2FzX3VzZXIgPSBzZWxmLl9hcmdfcGFyc2VyLnJ1bl9hc191c2Vy
            CiAgICAgICAgc2VsZi5vdXRwdXRfZmlsZSA9IHNlbGYuX2FyZ19wYXJzZXIub3V0cHV0CiAgICAg
            ICAgc2VsZi5jZXBoX2NvbmYgPSBzZWxmLl9hcmdfcGFyc2VyLmNlcGhfY29uZgogICAgICAgIHNl
            bGYuY2VwaF9rZXlyaW5nID0gc2VsZi5fYXJnX3BhcnNlci5rZXlyaW5nCiAgICAgICAgIyBpZiB1
            c2VyIG5vdCBwcm92aWRlZCwgZ2l2ZSBhIGRlZmF1bHQgdXNlcgogICAgICAgIGlmIG5vdCBzZWxm
            LnJ1bl9hc191c2VyIGFuZCBub3Qgc2VsZi5fYXJnX3BhcnNlci51cGdyYWRlOgogICAgICAgICAg
            ICBzZWxmLnJ1bl9hc191c2VyID0gc2VsZi5FWFRFUk5BTF9VU0VSX05BTUUKICAgICAgICBpZiBu
            b3Qgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXggYW5kIG5vdCBzZWxmLl9hcmdfcGFy
            c2VyLnVwZ3JhZGU6CiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4
            ID0gc2VsZi5ERUZBVUxUX1JHV19QT09MX1BSRUZJWAogICAgICAgIGlmIHNlbGYuY2VwaF9jb25m
            OgogICAgICAgICAgICBrd2FyZ3MgPSB7fQogICAgICAgICAgICBpZiBzZWxmLmNlcGhfa2V5cmlu
            ZzoKICAgICAgICAgICAgICAgIGt3YXJnc1siY29uZiJdID0geyJrZXlyaW5nIjogc2VsZi5jZXBo
            X2tleXJpbmd9CiAgICAgICAgICAgIHNlbGYuY2x1c3RlciA9IHJhZG9zLlJhZG9zKGNvbmZmaWxl
            PXNlbGYuY2VwaF9jb25mLCAqKmt3YXJncykKICAgICAgICBlbHNlOgogICAgICAgICAgICBzZWxm
            LmNsdXN0ZXIgPSByYWRvcy5SYWRvcygpCiAgICAgICAgICAgIHNlbGYuY2x1c3Rlci5jb25mX3Jl
            YWRfZmlsZSgpCiAgICAgICAgc2VsZi5jbHVzdGVyLmNvbm5lY3QoKQoKICAgIGRlZiBzaHV0ZG93
            bihzZWxmKToKICAgICAgICBpZiBzZWxmLmNsdXN0ZXIuc3RhdGUgPT0gImNvbm5lY3RlZCI6CiAg
            ICAgICAgICAgIHNlbGYuY2x1c3Rlci5zaHV0ZG93bigpCgogICAgZGVmIGdldF9mc2lkKHNlbGYp
            OgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJu
            IHNlbGYuZHJ5X3J1bigiY2VwaCBmc2lkIikKICAgICAgICByZXR1cm4gc3RyKHNlbGYuY2x1c3Rl
            ci5nZXRfZnNpZCgpKQoKICAgIGRlZiBfY29tbW9uX2NtZF9qc29uX2dlbihzZWxmLCBjbWRfanNv
            bik6CiAgICAgICAgY21kID0ganNvbi5kdW1wcyhjbWRfanNvbiwgc29ydF9rZXlzPVRydWUpCiAg
            ICAgICAgcmV0X3ZhbCwgY21kX291dCwgZXJyX21zZyA9IHNlbGYuY2x1c3Rlci5tb25fY29tbWFu
            ZChjbWQsIGIiIikKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnZlcmJvc2U6CiAgICAgICAg
            ICAgIHByaW50KGYiQ29tbWFuZCBJbnB1dDoge2NtZH0iKQogICAgICAgICAgICBwcmludCgKICAg
            ICAgICAgICAgICAgIGYiUmV0dXJuIFZhbDoge3JldF92YWx9XG5Db21tYW5kIE91dHB1dDoge2Nt
            ZF9vdXR9XG4iCiAgICAgICAgICAgICAgICBmIkVycm9yIE1lc3NhZ2U6IHtlcnJfbXNnfVxuLS0t
            LS0tLS0tLVxuIgogICAgICAgICAgICApCiAgICAgICAganNvbl9vdXQgPSB7fQogICAgICAgICMg
            aWYgdGhlcmUgaXMgbm8gZXJyb3IgKGkuZTsgcmV0X3ZhbCBpcyBaRVJPKSBhbmQgJ2NtZF9vdXQn
            IGlzIG5vdCBlbXB0eQogICAgICAgICMgdGhlbiBjb252ZXJ0ICdjbWRfb3V0JyB0byBhIGpzb24g
            b3V0cHV0CiAgICAgICAgaWYgcmV0X3ZhbCA9PSAwIGFuZCBjbWRfb3V0OgogICAgICAgICAgICBq
            c29uX291dCA9IGpzb24ubG9hZHMoY21kX291dCkKICAgICAgICByZXR1cm4gcmV0X3ZhbCwganNv
            bl9vdXQsIGVycl9tc2cKCiAgICBkZWYgZ2V0X2NlcGhfZXh0ZXJuYWxfbW9uX2RhdGEoc2VsZik6
            CiAgICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJxdW9ydW1fc3RhdHVzIiwgImZvcm1hdCI6
            ICJqc29uIn0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAg
            IHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGggIiArIGNtZF9qc29uWyJwcmVmaXgiXSkKICAgICAg
            ICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4o
            Y21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwK
            ICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgIidxdW9y
            dW1fc3RhdHVzJyBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAgICAgIGYiRXJyb3I6IHtl
            cnJfbXNnIGlmIHJldF92YWwgIT0gMCBlbHNlIHNlbGYuRU1QVFlfT1VUUFVUX0xJU1R9IgogICAg
            ICAgICAgICApCiAgICAgICAgcV9sZWFkZXJfbmFtZSA9IGpzb25fb3V0WyJxdW9ydW1fbGVhZGVy
            X25hbWUiXQogICAgICAgIHFfbGVhZGVyX2RldGFpbHMgPSB7fQogICAgICAgIHFfbGVhZGVyX21h
            dGNoaW5nX2xpc3QgPSBbCiAgICAgICAgICAgIGwgZm9yIGwgaW4ganNvbl9vdXRbIm1vbm1hcCJd
            WyJtb25zIl0gaWYgbFsibmFtZSJdID09IHFfbGVhZGVyX25hbWUKICAgICAgICBdCiAgICAgICAg
            aWYgbGVuKHFfbGVhZGVyX21hdGNoaW5nX2xpc3QpID09IDA6CiAgICAgICAgICAgIHJhaXNlIEV4
            ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oIk5vIG1hdGNoaW5nICdtb24nIGRldGFpbHMgZm91bmQi
            KQogICAgICAgIHFfbGVhZGVyX2RldGFpbHMgPSBxX2xlYWRlcl9tYXRjaGluZ19saXN0WzBdCiAg
            ICAgICAgIyBnZXQgdGhlIGFkZHJlc3MgdmVjdG9yIG9mIHRoZSBxdW9ydW0tbGVhZGVyCiAgICAg
            ICAgcV9sZWFkZXJfYWRkcnZlYyA9IHFfbGVhZGVyX2RldGFpbHMuZ2V0KCJwdWJsaWNfYWRkcnMi
            LCB7fSkuZ2V0KCJhZGRydmVjIiwgW10pCiAgICAgICAgaXBfYWRkciA9IHN0cihxX2xlYWRlcl9k
            ZXRhaWxzWyJwdWJsaWNfYWRkciJdLnNwbGl0KCIvIilbMF0pCgogICAgICAgIGlmIHNlbGYuX2Fy
            Z19wYXJzZXIudjJfcG9ydF9lbmFibGU6CiAgICAgICAgICAgIGlmIHFfbGVhZGVyX2FkZHJ2ZWNb
            MF1bInR5cGUiXSA9PSAidjIiOgogICAgICAgICAgICAgICAgaXBfYWRkciA9IHFfbGVhZGVyX2Fk
            ZHJ2ZWNbMF1bImFkZHIiXQogICAgICAgICAgICBlbGlmIGxlbihxX2xlYWRlcl9hZGRydmVjKSA+
            IDEgYW5kIHFfbGVhZGVyX2FkZHJ2ZWNbMV1bInR5cGUiXSA9PSAidjIiOgogICAgICAgICAgICAg
            ICAgaXBfYWRkciA9IHFfbGVhZGVyX2FkZHJ2ZWNbMV1bImFkZHIiXQogICAgICAgICAgICBlbHNl
            OgogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZSgKICAgICAgICAgICAgICAgICAgICAi
            J3YyJyBhZGRyZXNzIHR5cGUgbm90IHByZXNlbnQsIGFuZCAndjItcG9ydC1lbmFibGUnIGZsYWcg
            aXMgcHJvdmlkZWQiCiAgICAgICAgICAgICAgICApCgogICAgICAgIHJldHVybiBmIntzdHIocV9s
            ZWFkZXJfbmFtZSl9PXtpcF9hZGRyfSIKCiAgICBkZWYgX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAo
            c2VsZiwgaG9zdF9uYW1lLCBwb3J0LCBpcF90eXBlKToKICAgICAgICAjIGlmICdjbHVzdGVyJyBp
            bnN0YW5jZSBpcyBhIGR1bW15IHR5cGUsCiAgICAgICAgIyBjYWxsIHRoZSBkdW1teSBpbnN0YW5j
            ZSdzICJjb252ZXJ0IiBtZXRob2QKICAgICAgICBpZiBub3QgaG9zdF9uYW1lOgogICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKCJFbXB0eSBob3N0bmFtZSBwcm92aWRl
            ZCIpCiAgICAgICAgaWYgaXNpbnN0YW5jZShzZWxmLmNsdXN0ZXIsIER1bW15UmFkb3MpOgogICAg
            ICAgICAgICByZXR1cm4gc2VsZi5jbHVzdGVyLl9jb252ZXJ0X2hvc3RuYW1lX3RvX2lwKGhvc3Rf
            bmFtZSkKCiAgICAgICAgaWYgaXBfdHlwZSA9PSAiRlFETiI6CiAgICAgICAgICAgICMgY2hlY2sg
            d2hpY2ggaXAgRlFETiBzaG91bGQgYmUgY29udmVydGVkIHRvLCBJUHY0IG9yIElQdjYKICAgICAg
            ICAgICAgIyBjaGVjayB0aGUgaG9zdCBpcCwgdGhlIGVuZHBvaW50IGlwIHR5cGUgd291bGQgYmUg
            c2ltaWxhciB0byBob3N0IGlwCiAgICAgICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAib3Jj
            aCBob3N0IGxzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICAgICAgcmV0X3ZhbCwganNvbl9v
            dXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAg
            ICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAogICAgICAgICAgICBp
            ZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICAgICAgcmFp
            c2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJ29yY2gg
            aG9zdCBscycgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAgICAgICAgICAgICAgICAgZiJFcnJvcjog
            e2Vycl9tc2cgaWYgcmV0X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAg
            ICAgICAgICAgICAgICApCiAgICAgICAgICAgIGhvc3RfYWRkciA9IGpzb25fb3V0WzBdWyJhZGRy
            Il0KICAgICAgICAgICAgIyBhZGQgOjgwIHNhbXBsZSBwb3J0IGluIGlwX3R5cGUsIGFzIF9pbnZh
            bGlkX2VuZHBvaW50IGFsc28gdmVyaWZ5IHBvcnQKICAgICAgICAgICAgaG9zdF9pcF90eXBlID0g
            c2VsZi5faW52YWxpZF9lbmRwb2ludChob3N0X2FkZHIgKyAiOjgwIikKICAgICAgICAgICAgaW1w
            b3J0IHNvY2tldAoKICAgICAgICAgICAgIyBleGFtcGxlIG91dHB1dCBbKDxBZGRyZXNzRmFtaWx5
            LkFGX0lORVQ6IDI+LCA8U29ja2V0S2luZC5TT0NLX1NUUkVBTTogMT4sIDYsICcnLCAoJzkzLjE4
            NC4yMTYuMzQnLCA4MCkpLCAuLi5dCiAgICAgICAgICAgICMgd2UgbmVlZCB0byBnZXQgOTMuMTg0
            LjIxNi4zNCBzbyBpdCB3b3VsZCBiZSBpcFswXVs0XVswXQogICAgICAgICAgICBpZiBob3N0X2lw
            X3R5cGUgPT0gIklQdjYiOgogICAgICAgICAgICAgICAgaXAgPSBzb2NrZXQuZ2V0YWRkcmluZm8o
            CiAgICAgICAgICAgICAgICAgICAgaG9zdF9uYW1lLCBwb3J0LCBmYW1pbHk9c29ja2V0LkFGX0lO
            RVQ2LCBwcm90bz1zb2NrZXQuSVBQUk9UT19UQ1AKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgZWxpZiBob3N0X2lwX3R5cGUgPT0gIklQdjQiOgogICAgICAgICAgICAgICAgaXAgPSBzb2Nr
            ZXQuZ2V0YWRkcmluZm8oCiAgICAgICAgICAgICAgICAgICAgaG9zdF9uYW1lLCBwb3J0LCBmYW1p
            bHk9c29ja2V0LkFGX0lORVQsIHByb3RvPXNvY2tldC5JUFBST1RPX1RDUAogICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICBkZWwgc29ja2V0CiAgICAgICAgICAgIHJldHVybiBpcFswXVs0XVsw
            XQogICAgICAgIHJldHVybiBob3N0X25hbWUKCiAgICBkZWYgZ2V0X2FjdGl2ZV9hbmRfc3RhbmRi
            eV9tZ3JzKHNlbGYpOgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAg
            ICAgICAgcmV0dXJuICIiLCBzZWxmLmRyeV9ydW4oImNlcGggc3RhdHVzIikKICAgICAgICBtb25p
            dG9yaW5nX2VuZHBvaW50X3BvcnQgPSBzZWxmLl9hcmdfcGFyc2VyLm1vbml0b3JpbmdfZW5kcG9p
            bnRfcG9ydAogICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdCA9IHNlbGYuX2FyZ19w
            YXJzZXIubW9uaXRvcmluZ19lbmRwb2ludAogICAgICAgIHN0YW5kYnlfbWdycyA9IFtdCiAgICAg
            ICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdDoKICAgICAgICAgICAgY21kX2pz
            b24gPSB7InByZWZpeCI6ICJzdGF0dXMiLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgICAgICBy
            ZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21k
            X2pzb24pCiAgICAgICAgICAgICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQs
            CiAgICAgICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAgICAg
            ICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAg
            ICAgICAgICInbWdyIHNlcnZpY2VzJyBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAgICAg
            ICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09V
            VFBVVF9MSVNUfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRw
            b2ludCA9ICgKICAgICAgICAgICAgICAgIGpzb25fb3V0LmdldCgibWdybWFwIiwge30pLmdldCgi
            c2VydmljZXMiLCB7fSkuZ2V0KCJwcm9tZXRoZXVzIiwgIiIpCiAgICAgICAgICAgICkKICAgICAg
            ICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnQ6CiAgICAgICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJjYW4ndCBmaW5k
            IG1vbml0b3JpbmdfZW5kcG9pbnQsIHByb21ldGhldXMgbW9kdWxlIG1pZ2h0IG5vdCBiZSBlbmFi
            bGVkLCAiCiAgICAgICAgICAgICAgICAgICAgImVuYWJsZSB0aGUgbW9kdWxlIGJ5IHJ1bm5pbmcg
            J2NlcGggbWdyIG1vZHVsZSBlbmFibGUgcHJvbWV0aGV1cyciCiAgICAgICAgICAgICAgICApCiAg
            ICAgICAgICAgICMgbm93IGNoZWNrIHRoZSBzdGFuZC1ieSBtZ3ItcwogICAgICAgICAgICBzdGFu
            ZGJ5X2FyciA9IGpzb25fb3V0LmdldCgibWdybWFwIiwge30pLmdldCgic3RhbmRieXMiLCBbXSkK
            ICAgICAgICAgICAgZm9yIGVhY2hfc3RhbmRieSBpbiBzdGFuZGJ5X2FycjoKICAgICAgICAgICAg
            ICAgIGlmICJuYW1lIiBpbiBlYWNoX3N0YW5kYnkua2V5cygpOgogICAgICAgICAgICAgICAgICAg
            IHN0YW5kYnlfbWdycy5hcHBlbmQoZWFjaF9zdGFuZGJ5WyJuYW1lIl0pCiAgICAgICAgICAgIHRy
            eToKICAgICAgICAgICAgICAgIHBhcnNlZF9lbmRwb2ludCA9IHVybHBhcnNlKG1vbml0b3Jpbmdf
            ZW5kcG9pbnQpCiAgICAgICAgICAgIGV4Y2VwdCBWYWx1ZUVycm9yOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBmImlu
            dmFsaWQgZW5kcG9pbnQ6IHttb25pdG9yaW5nX2VuZHBvaW50fSIKICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0ID0gcGFyc2VkX2VuZHBvaW50
            Lmhvc3RuYW1lCiAgICAgICAgICAgIGlmIG5vdCBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQ6CiAg
            ICAgICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQgPSBzdHIocGFyc2VkX2VuZHBv
            aW50LnBvcnQpCgogICAgICAgICMgaWYgbW9uaXRvcmluZyBlbmRwb2ludCBwb3J0IGlzIG5vdCBz
            ZXQsIHB1dCBhIGRlZmF1bHQgbW9uIHBvcnQKICAgICAgICBpZiBub3QgbW9uaXRvcmluZ19lbmRw
            b2ludF9wb3J0OgogICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQgPSBzZWxmLkRF
            RkFVTFRfTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUCgogICAgICAgICMgdXNlciBjb3VsZCBnaXZl
            IGNvbW1hIGFuZCBzcGFjZSBzZXBhcmF0ZWQgaW5wdXRzIChsaWtlIC0tbW9uaXRvcmluZy1lbmRw
            b2ludD0iPGlwMT4sIDxpcDI+IikKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Qg
            PSBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3QucmVwbGFjZSgiLCIsICIgIikKICAgICAgICBt
            b25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXQgPSBtb25pdG9yaW5nX2VuZHBvaW50X2lw
            X2xpc3Quc3BsaXQoKQogICAgICAgICMgaWYgbW9uaXRvcmluZy1lbmRwb2ludCBjb3VsZCBub3Qg
            YmUgZm91bmQsIHJhaXNlIGFuIGVycm9yCiAgICAgICAgaWYgbGVuKG1vbml0b3JpbmdfZW5kcG9p
            bnRfaXBfbGlzdF9zcGxpdCkgPT0gMDoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVy
            ZUV4Y2VwdGlvbigiTm8gJ21vbml0b3JpbmctZW5kcG9pbnQnIGZvdW5kIikKICAgICAgICAjIGZp
            cnN0IGlwIGlzIHRyZWF0ZWQgYXMgdGhlIG1haW4gbW9uaXRvcmluZy1lbmRwb2ludAogICAgICAg
            IG1vbml0b3JpbmdfZW5kcG9pbnRfaXAgPSBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3Bs
            aXRbMF0KICAgICAgICAjIHJlc3Qgb2YgdGhlIGlwLXMgYXJlIGFkZGVkIHRvIHRoZSAnc3RhbmRi
            eV9tZ3JzJyBsaXN0CiAgICAgICAgc3RhbmRieV9tZ3JzLmV4dGVuZChtb25pdG9yaW5nX2VuZHBv
            aW50X2lwX2xpc3Rfc3BsaXRbMTpdKQogICAgICAgIGZhaWxlZF9pcCA9IG1vbml0b3JpbmdfZW5k
            cG9pbnRfaXAKCiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludCA9ICI6Ii5qb2luKAogICAgICAg
            ICAgICBbbW9uaXRvcmluZ19lbmRwb2ludF9pcCwgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0XQog
            ICAgICAgICkKICAgICAgICBpcF90eXBlID0gc2VsZi5faW52YWxpZF9lbmRwb2ludChtb25pdG9y
            aW5nX2VuZHBvaW50KQogICAgICAgIHRyeToKICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2lu
            dF9pcCA9IHNlbGYuX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoCiAgICAgICAgICAgICAgICBtb25p
            dG9yaW5nX2VuZHBvaW50X2lwLCBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQsIGlwX3R5cGUKICAg
            ICAgICAgICAgKQogICAgICAgICAgICAjIGNvbGxlY3QgYWxsIHRoZSAnc3RhbmQtYnknIG1nciBp
            cHMKICAgICAgICAgICAgbWdyX2lwcyA9IFtdCiAgICAgICAgICAgIGZvciBlYWNoX3N0YW5kYnlf
            bWdyIGluIHN0YW5kYnlfbWdyczoKICAgICAgICAgICAgICAgIGZhaWxlZF9pcCA9IGVhY2hfc3Rh
            bmRieV9tZ3IKICAgICAgICAgICAgICAgIG1ncl9pcHMuYXBwZW5kKAogICAgICAgICAgICAgICAg
            ICAgIHNlbGYuX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoCiAgICAgICAgICAgICAgICAgICAgICAg
            IGVhY2hfc3RhbmRieV9tZ3IsIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydCwgaXBfdHlwZQogICAg
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgICkKICAgICAgICBleGNlcHQ6CiAgICAg
            ICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBm
            IkNvbnZlcnNpb24gb2YgaG9zdDoge2ZhaWxlZF9pcH0gdG8gSVAgZmFpbGVkLiAiCiAgICAgICAg
            ICAgICAgICAiUGxlYXNlIGVudGVyIHRoZSBJUCBhZGRyZXNzZXMgb2YgYWxsIHRoZSBjZXBoLW1n
            cnMgd2l0aCB0aGUgJy0tbW9uaXRvcmluZy1lbmRwb2ludCcgZmxhZyIKICAgICAgICAgICAgKQoK
            ICAgICAgICBfLCBfLCBlcnIgPSBzZWxmLmVuZHBvaW50X2RpYWwobW9uaXRvcmluZ19lbmRwb2lu
            dCwgaXBfdHlwZSkKICAgICAgICBpZiBlcnIgPT0gIi0xIjoKICAgICAgICAgICAgcmFpc2UgRXhl
            Y3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihlcnIpCiAgICAgICAgIyBhZGQgdGhlIHZhbGlkYXRlZCBh
            Y3RpdmUgbWdyIElQIGludG8gdGhlIGZpcnN0IGluZGV4CiAgICAgICAgbWdyX2lwcy5pbnNlcnQo
            MCwgbW9uaXRvcmluZ19lbmRwb2ludF9pcCkKICAgICAgICBhbGxfbWdyX2lwc19zdHIgPSAiLCIu
            am9pbihtZ3JfaXBzKQogICAgICAgIHJldHVybiBhbGxfbWdyX2lwc19zdHIsIG1vbml0b3Jpbmdf
            ZW5kcG9pbnRfcG9ydAoKICAgIGRlZiBjaGVja191c2VyX2V4aXN0KHNlbGYsIHVzZXIpOgogICAg
            ICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAiYXV0aCBnZXQiLCAiZW50aXR5IjogZiJ7dXNlcn0i
            LCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBfID0gc2VsZi5f
            Y29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3Ig
            bGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICByZXR1cm4gIiIKICAgICAgICByZXR1cm4g
            c3RyKGpzb25fb3V0WzBdWyJrZXkiXSkKCiAgICBkZWYgZ2V0X2NlcGhmc19wcm92aXNpb25lcl9j
            YXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gImNsaWVudC5jc2ktY2VwaGZz
            LXByb3Zpc2lvbmVyIgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAgICJtb24iOiAiYWxsb3cg
            ciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLAogICAgICAgICAgICAibWdyIjogImFs
            bG93IHJ3IiwKICAgICAgICAgICAgIm9zZCI6ICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRh
            PSoiLAogICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0
            aF9wZXJtaXNzaW9uOgogICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3Bh
            cnNlci5rOHNfY2x1c3Rlcl9uYW1lCiAgICAgICAgICAgIGlmIGs4c19jbHVzdGVyX25hbWUgPT0g
            IiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAg
            ICAgICAgICAgICAgICAgICJrOHNfY2x1c3Rlcl9uYW1lIG5vdCBmb3VuZCwgcGxlYXNlIHNldCB0
            aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgY2VwaGZzX2ZpbGVzeXN0ZW0gPSBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVt
            X25hbWUKICAgICAgICAgICAgaWYgY2VwaGZzX2ZpbGVzeXN0ZW0gPT0gIiI6CiAgICAgICAgICAg
            ICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9uYW1lfSIKICAgICAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0te2s4c19jbHVzdGVyX25h
            bWV9LXtjZXBoZnNfZmlsZXN5c3RlbX0iCiAgICAgICAgICAgICAgICBjYXBzWyJvc2QiXSA9IGYi
            YWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT17Y2VwaGZzX2ZpbGVzeXN0ZW19IgoKICAgICAg
            ICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVmIGdldF9jZXBoZnNfbm9kZV9jYXBzX2FuZF9l
            bnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiCiAg
            ICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJhbGxvdyByLCBhbGxvdyBjb21tYW5k
            ICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJtZ3IiOiAiYWxsb3cgcnciLAogICAgICAg
            ICAgICAib3NkIjogImFsbG93IHJ3IHRhZyBjZXBoZnMgKj0qIiwKICAgICAgICAgICAgIm1kcyI6
            ICJhbGxvdyBydyIsCiAgICAgICAgfQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJp
            Y3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAgIGs4c19jbHVzdGVyX25hbWUgPSBzZWxm
            Ll9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAgICAgICAgICAgaWYgazhzX2NsdXN0ZXJf
            bmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRp
            b24oCiAgICAgICAgICAgICAgICAgICAgIms4c19jbHVzdGVyX25hbWUgbm90IGZvdW5kLCBwbGVh
            c2Ugc2V0IHRoZSAnLS1rOHMtY2x1c3Rlci1uYW1lJyBmbGFnIgogICAgICAgICAgICAgICAgKQog
            ICAgICAgICAgICBjZXBoZnNfZmlsZXN5c3RlbSA9IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2Zp
            bGVzeXN0ZW1fbmFtZQogICAgICAgICAgICBpZiBjZXBoZnNfZmlsZXN5c3RlbSA9PSAiIjoKICAg
            ICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0te2s4c19jbHVzdGVyX25hbWV9IgogICAg
            ICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17azhzX2Ns
            dXN0ZXJfbmFtZX0te2NlcGhmc19maWxlc3lzdGVtfSIKICAgICAgICAgICAgICAgIGNhcHNbIm9z
            ZCJdID0gZiJhbGxvdyBydyB0YWcgY2VwaGZzICo9e2NlcGhmc19maWxlc3lzdGVtfSIKCiAgICAg
            ICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfZW50aXR5KAogICAgICAgIHNlbGYs
            CiAgICAgICAgZW50aXR5LAogICAgICAgIHJiZF9wb29sX25hbWUsCiAgICAgICAgYWxpYXNfcmJk
            X3Bvb2xfbmFtZSwKICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lLAogICAgICAgIHJhZG9zX25hbWVz
            cGFjZSwKICAgICk6CiAgICAgICAgaWYgKAogICAgICAgICAgICByYmRfcG9vbF9uYW1lLmNvdW50
            KCIuIikgIT0gMAogICAgICAgICAgICBvciByYmRfcG9vbF9uYW1lLmNvdW50KCJfIikgIT0gMAog
            ICAgICAgICAgICBvciBhbGlhc19yYmRfcG9vbF9uYW1lICE9ICIiCiAgICAgICAgICAgICMgY2hl
            Y2tpbmcgYWxpYXNfcmJkX3Bvb2xfbmFtZSBpcyBub3QgZW1wdHkgYXMgdGhlcmUgbWF5YmUgYSBz
            cGVjaWFsIGNoYXJhY3RlciB1c2VkIG90aGVyIHRoYW4gLiBvciBfCiAgICAgICAgKToKICAgICAg
            ICAgICAgaWYgYWxpYXNfcmJkX3Bvb2xfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgInBsZWFzZSBz
            ZXQgdGhlICctLWFsaWFzLXJiZC1kYXRhLXBvb2wtbmFtZScgZmxhZyBhcyB0aGUgcmJkIGRhdGEg
            cG9vbCBuYW1lIGNvbnRhaW5zICcuJyBvciAnXyciCiAgICAgICAgICAgICAgICApCiAgICAgICAg
            ICAgIGlmICgKICAgICAgICAgICAgICAgIGFsaWFzX3JiZF9wb29sX25hbWUuY291bnQoIi4iKSAh
            PSAwCiAgICAgICAgICAgICAgICBvciBhbGlhc19yYmRfcG9vbF9uYW1lLmNvdW50KCJfIikgIT0g
            MAogICAgICAgICAgICApOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJy0tYWxpYXMtcmJkLWRhdGEtcG9vbC1uYW1l
            JyBmbGFnIHZhbHVlIHNob3VsZCBub3QgY29udGFpbiAnLicgb3IgJ18nIgogICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9uYW1lfS17
            YWxpYXNfcmJkX3Bvb2xfbmFtZX0iCiAgICAgICAgZWxzZToKICAgICAgICAgICAgZW50aXR5ID0g
            ZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0te3JiZF9wb29sX25hbWV9IgoKICAgICAgICBp
            ZiByYWRvc19uYW1lc3BhY2U6CiAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0te3JhZG9z
            X25hbWVzcGFjZX0iCiAgICAgICAgcmV0dXJuIGVudGl0eQoKICAgIGRlZiBnZXRfcmJkX3Byb3Zp
            c2lvbmVyX2NhcHNfYW5kX2VudGl0eShzZWxmKToKICAgICAgICBlbnRpdHkgPSAiY2xpZW50LmNz
            aS1yYmQtcHJvdmlzaW9uZXIiCiAgICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJw
            cm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLAogICAgICAgICAgICAi
            bWdyIjogImFsbG93IHJ3IiwKICAgICAgICAgICAgIm9zZCI6ICJwcm9maWxlIHJiZCIsCiAgICAg
            ICAgfQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Np
            b246CiAgICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRh
            X3Bvb2xfbmFtZQogICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3Bh
            cnNlci5hbGlhc19yYmRfZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFt
            ZSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQogICAgICAgICAgICByYWRvc19u
            YW1lc3BhY2UgPSBzZWxmLl9hcmdfcGFyc2VyLnJhZG9zX25hbWVzcGFjZQogICAgICAgICAgICBp
            ZiByYmRfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAibWFuZGF0b3J5IGZsYWcgbm90IGZv
            dW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1yYmQtZGF0YS1wb29sLW5hbWUnIGZsYWciCiAgICAgICAg
            ICAgICAgICApCiAgICAgICAgICAgIGlmIGs4c19jbHVzdGVyX25hbWUgPT0gIiI6CiAgICAgICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLWs4cy1jbHVz
            dGVyLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVudGl0eSA9IHNl
            bGYuZ2V0X2VudGl0eSgKICAgICAgICAgICAgICAgIGVudGl0eSwKICAgICAgICAgICAgICAgIHJi
            ZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lLAogICAgICAg
            ICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAgICAgICAgICAgIHJhZG9zX25hbWVzcGFj
            ZSwKICAgICAgICAgICAgKQogICAgICAgICAgICBpZiByYWRvc19uYW1lc3BhY2UgIT0gIiI6CiAg
            ICAgICAgICAgICAgICBjYXBzWyJvc2QiXSA9ICgKICAgICAgICAgICAgICAgICAgICBmInByb2Zp
            bGUgcmJkIHBvb2w9e3JiZF9wb29sX25hbWV9IG5hbWVzcGFjZT17cmFkb3NfbmFtZXNwYWNlfSIK
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGNhcHNb
            Im9zZCJdID0gZiJwcm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9uYW1lfSIKCiAgICAgICAgcmV0
            dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfcmJkX25vZGVfY2Fwc19hbmRfZW50aXR5KHNl
            bGYpOgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3NpLXJiZC1ub2RlIgogICAgICAgIGNhcHMg
            PSB7CiAgICAgICAgICAgICJtb24iOiAicHJvZmlsZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBi
            bG9ja2xpc3QnIiwKICAgICAgICAgICAgIm9zZCI6ICJwcm9maWxlIHJiZCIsCiAgICAgICAgfQog
            ICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb246CiAg
            ICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xf
            bmFtZQogICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5h
            bGlhc19yYmRfZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNl
            bGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQogICAgICAgICAgICByYWRvc19uYW1lc3Bh
            Y2UgPSBzZWxmLl9hcmdfcGFyc2VyLnJhZG9zX25hbWVzcGFjZQogICAgICAgICAgICBpZiByYmRf
            cG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAibWFuZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBw
            bGVhc2Ugc2V0IHRoZSAnLS1yYmQtZGF0YS1wb29sLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAg
            ICApCiAgICAgICAgICAgIGlmIGs4c19jbHVzdGVyX25hbWUgPT0gIiI6CiAgICAgICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJt
            YW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLWs4cy1jbHVzdGVyLW5h
            bWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVudGl0eSA9IHNlbGYuZ2V0
            X2VudGl0eSgKICAgICAgICAgICAgICAgIGVudGl0eSwKICAgICAgICAgICAgICAgIHJiZF9wb29s
            X25hbWUsCiAgICAgICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lLAogICAgICAgICAgICAg
            ICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAgICAgICAgICAgIHJhZG9zX25hbWVzcGFjZSwKICAg
            ICAgICAgICAgKQogICAgICAgICAgICBpZiByYWRvc19uYW1lc3BhY2UgIT0gIiI6CiAgICAgICAg
            ICAgICAgICBjYXBzWyJvc2QiXSA9ICgKICAgICAgICAgICAgICAgICAgICBmInByb2ZpbGUgcmJk
            IHBvb2w9e3JiZF9wb29sX25hbWV9IG5hbWVzcGFjZT17cmFkb3NfbmFtZXNwYWNlfSIKICAgICAg
            ICAgICAgICAgICkKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJd
            ID0gZiJwcm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9uYW1lfSIKCiAgICAgICAgcmV0dXJuIGNh
            cHMsIGVudGl0eQoKICAgIGRlZiBnZXRfZGVmYXVsdFVzZXJfY2Fwc19hbmRfZW50aXR5KHNlbGYp
            OgogICAgICAgIGVudGl0eSA9IHNlbGYucnVuX2FzX3VzZXIKICAgICAgICBjYXBzID0gewogICAg
            ICAgICAgICAibW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxs
            b3cgY29tbWFuZCB2ZXJzaW9uIiwKICAgICAgICAgICAgIm1nciI6ICJhbGxvdyBjb21tYW5kIGNv
            bmZpZyIsCiAgICAgICAgICAgICJvc2QiOiBmInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cg
            cnd4IHBvb2w9e3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cubWV0YSwgYWxs
            b3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD17c2VsZi5fYXJnX3BhcnNlci5yZ3df
            cG9vbF9wcmVmaXh9LnJndy5jb250cm9sLCBhbGxvdyByeCBwb29sPXtzZWxmLl9hcmdfcGFyc2Vy
            LnJnd19wb29sX3ByZWZpeH0ucmd3LmxvZywgYWxsb3cgeCBwb29sPXtzZWxmLl9hcmdfcGFyc2Vy
            LnJnd19wb29sX3ByZWZpeH0ucmd3LmJ1Y2tldHMuaW5kZXgiLAogICAgICAgIH0KCiAgICAgICAg
            cmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfY2Fwc19hbmRfZW50aXR5KHNlbGYsIHVz
            ZXJfbmFtZSk6CiAgICAgICAgaWYgImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiBpbiB1
            c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIg
            IT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVk
            X2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZ2V0X2NlcGhm
            c19wcm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoKQogICAgICAgIGlmICJjbGllbnQuY3NpLWNl
            cGhmcy1ub2RlIiBpbiB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQuY3NpLWNlcGhm
            cy1ub2RlIiAhPSB1c2VyX25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJl
            c3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5n
            ZXRfY2VwaGZzX25vZGVfY2Fwc19hbmRfZW50aXR5KCkKICAgICAgICBpZiAiY2xpZW50LmNzaS1y
            YmQtcHJvdmlzaW9uZXIiIGluIHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVudC5jc2kt
            cmJkLXByb3Zpc2lvbmVyIiAhPSB1c2VyX25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdf
            cGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1
            cm4gc2VsZi5nZXRfcmJkX3Byb3Zpc2lvbmVyX2NhcHNfYW5kX2VudGl0eSgpCiAgICAgICAgaWYg
            ImNsaWVudC5jc2ktcmJkLW5vZGUiIGluIHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVu
            dC5jc2ktcmJkLW5vZGUiICE9IHVzZXJfbmFtZToKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19w
            YXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVlCiAgICAgICAgICAgIHJldHVy
            biBzZWxmLmdldF9yYmRfbm9kZV9jYXBzX2FuZF9lbnRpdHkoKQogICAgICAgIGlmICJjbGllbnQu
            aGVhbHRoY2hlY2tlciIgaW4gdXNlcl9uYW1lOgogICAgICAgICAgICBpZiAiY2xpZW50LmhlYWx0
            aGNoZWNrZXIiICE9IHVzZXJfbmFtZToKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIu
            cmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVlCiAgICAgICAgICAgIHJldHVybiBzZWxm
            LmdldF9kZWZhdWx0VXNlcl9jYXBzX2FuZF9lbnRpdHkoKQoKICAgICAgICByYWlzZSBFeGVjdXRp
            b25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICBmIm5vIHVzZXIgZm91bmQgd2l0aCB1c2Vy
            X25hbWU6IHt1c2VyX25hbWV9LCAiCiAgICAgICAgICAgICJnZXRfY2Fwc19hbmRfZW50aXR5IGNv
            bW1hbmQgZmFpbGVkLlxuIgogICAgICAgICkKCiAgICBkZWYgY3JlYXRlX2NlcGhDU0lLZXlyaW5n
            X3VzZXIoc2VsZiwgdXNlcik6CiAgICAgICAgIiIiCiAgICAgICAgY29tbWFuZDogY2VwaCBhdXRo
            IGdldC1vci1jcmVhdGUgY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIgbW9uICdhbGxvdyBy
            JyBtZ3IgJ2FsbG93IHJ3JyBvc2QgJ2FsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KicKICAg
            ICAgICAiIiIKICAgICAgICBjYXBzLCBlbnRpdHkgPSBzZWxmLmdldF9jYXBzX2FuZF9lbnRpdHko
            dXNlcikKICAgICAgICBjbWRfanNvbiA9IHsKICAgICAgICAgICAgInByZWZpeCI6ICJhdXRoIGdl
            dC1vci1jcmVhdGUiLAogICAgICAgICAgICAiZW50aXR5IjogZW50aXR5LAogICAgICAgICAgICAi
            Y2FwcyI6IFtjYXAgZm9yIGNhcF9saXN0IGluIGxpc3QoY2Fwcy5pdGVtcygpKSBmb3IgY2FwIGlu
            IGNhcF9saXN0XSwKICAgICAgICAgICAgImZvcm1hdCI6ICJqc29uIiwKICAgICAgICB9CgogICAg
            ICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuICgKICAg
            ICAgICAgICAgICAgIHNlbGYuZHJ5X3J1bigKICAgICAgICAgICAgICAgICAgICAiY2VwaCAiCiAg
            ICAgICAgICAgICAgICAgICAgKyBjbWRfanNvblsicHJlZml4Il0KICAgICAgICAgICAgICAgICAg
            ICArICIgIgogICAgICAgICAgICAgICAgICAgICsgY21kX2pzb25bImVudGl0eSJdCiAgICAgICAg
            ICAgICAgICAgICAgKyAiICIKICAgICAgICAgICAgICAgICAgICArICIgIi5qb2luKGNtZF9qc29u
            WyJjYXBzIl0pCiAgICAgICAgICAgICAgICApLAogICAgICAgICAgICAgICAgIiIsCiAgICAgICAg
            ICAgICkKICAgICAgICAjIGNoZWNrIGlmIHVzZXIgYWxyZWFkeSBleGlzdAogICAgICAgIHVzZXJf
            a2V5ID0gc2VsZi5jaGVja191c2VyX2V4aXN0KGVudGl0eSkKICAgICAgICBpZiB1c2VyX2tleSAh
            PSAiIjoKICAgICAgICAgICAgcmV0dXJuIHVzZXJfa2V5LCBmIntlbnRpdHkuc3BsaXQoJy4nLCAx
            KVsxXX0iCiAgICAgICAgICAgICMgZW50aXR5LnNwbGl0KCcuJywxKVsxXSB0byByZW5hbWUgZW50
            aXR5KGNsaWVudC5jc2ktcmJkLW5vZGUpIGFzIGNzaS1yYmQtbm9kZQoKICAgICAgICByZXRfdmFs
            LCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24p
            CiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwKICAgICAgICBp
            ZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiInYXV0aCBnZXQtb3It
            Y3JlYXRlIHt1c2VyfScgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAgICAgICAgICAgICBmIkVycm9y
            OiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIK
            ICAgICAgICAgICAgKQogICAgICAgIHJldHVybiBzdHIoanNvbl9vdXRbMF1bImtleSJdKSwgZiJ7
            ZW50aXR5LnNwbGl0KCcuJywgMSlbMV19IgogICAgICAgICMgZW50aXR5LnNwbGl0KCcuJywxKVsx
            XSB0byByZW5hbWUgZW50aXR5KGNsaWVudC5jc2ktcmJkLW5vZGUpIGFzIGNzaS1yYmQtbm9kZQoK
            ICAgIGRlZiBnZXRfY2VwaGZzX2RhdGFfcG9vbF9kZXRhaWxzKHNlbGYpOgogICAgICAgIGNtZF9q
            c29uID0geyJwcmVmaXgiOiAiZnMgbHMiLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIGlmIHNl
            bGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigi
            Y2VwaCAiICsgY21kX2pzb25bInByZWZpeCJdKQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBl
            cnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICAjIGlm
            IHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LCByZXBvcnQgYW4gZXJyb3IKICAgICAg
            ICBpZiByZXRfdmFsICE9IDA6CiAgICAgICAgICAgICMgaWYgZnMgYW5kIGRhdGFfcG9vbCBhcmd1
            bWVudHMgYXJlIG5vdCBzZXQsIHNpbGVudGx5IHJldHVybgogICAgICAgICAgICBpZiAoCiAgICAg
            ICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUgPT0gIiIK
            ICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFt
            ZSA9PSAiIgogICAgICAgICAgICApOgogICAgICAgICAgICAgICAgcmV0dXJuCiAgICAgICAgICAg
            ICMgaWYgdXNlciBoYXMgcHJvdmlkZWQgYW55IG9mIHRoZQogICAgICAgICAgICAjICctLWNlcGhm
            cy1maWxlc3lzdGVtLW5hbWUnIG9yICctLWNlcGhmcy1kYXRhLXBvb2wtbmFtZScgYXJndW1lbnRz
            LAogICAgICAgICAgICAjIHJhaXNlIGFuIGV4Y2VwdGlvbiBhcyB3ZSBhcmUgdW5hYmxlIHRvIHZl
            cmlmeSB0aGUgYXJncwogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KAogICAgICAgICAgICAgICAgZiInZnMgbHMnIGNlcGggY2FsbCBmYWlsZWQgd2l0aCBlcnJvcjog
            e2Vycl9tc2d9IgogICAgICAgICAgICApCgogICAgICAgIG1hdGNoaW5nX2pzb25fb3V0ID0ge30K
            ICAgICAgICAjIGlmICctLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUnIGFyZ3VtZW50IGlzIHByb3Zp
            ZGVkLAogICAgICAgICMgY2hlY2sgd2hldGhlciB0aGUgcHJvdmlkZWQgZmlsZXN5c3RlbS1uYW1l
            IGV4aXN0cyBvciBub3QKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lz
            dGVtX25hbWU6CiAgICAgICAgICAgICMgZ2V0IHRoZSBtYXRjaGluZyBsaXN0CiAgICAgICAgICAg
            IG1hdGNoaW5nX2pzb25fb3V0X2xpc3QgPSBbCiAgICAgICAgICAgICAgICBtYXRjaGVkCiAgICAg
            ICAgICAgICAgICBmb3IgbWF0Y2hlZCBpbiBqc29uX291dAogICAgICAgICAgICAgICAgaWYgc3Ry
            KG1hdGNoZWRbIm5hbWUiXSkgPT0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9u
            YW1lCiAgICAgICAgICAgIF0KICAgICAgICAgICAgIyB1bmFibGUgdG8gZmluZCBhIG1hdGNoaW5n
            IGZzLW5hbWUsIHJhaXNlIGFuIGVycm9yCiAgICAgICAgICAgIGlmIGxlbihtYXRjaGluZ19qc29u
            X291dF9saXN0KSA9PSAwOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBmIkZpbGVzeXN0ZW0gcHJvdmlkZWQsICd7c2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lfScsICIKICAgICAgICAgICAgICAg
            ICAgICBmImlzIG5vdCBmb3VuZCBpbiB0aGUgZnMtbGlzdDoge1tzdHIoeFsnbmFtZSddKSBmb3Ig
            eCBpbiBqc29uX291dF19IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBtYXRjaGluZ19q
            c29uX291dCA9IG1hdGNoaW5nX2pzb25fb3V0X2xpc3RbMF0KICAgICAgICAjIGlmIGNlcGhmcyBm
            aWxlc3lzdGVtIG5hbWUgaXMgbm90IHByb3ZpZGVkLAogICAgICAgICMgdHJ5IHRvIGdldCBhIGRl
            ZmF1bHQgZnMgbmFtZSBieSBkb2luZyB0aGUgZm9sbG93aW5nCiAgICAgICAgZWxzZToKICAgICAg
            ICAgICAgIyBhLiBjaGVjayBpZiB0aGVyZSBpcyBvbmx5IG9uZSBmaWxlc3lzdGVtIGlzIHByZXNl
            bnQKICAgICAgICAgICAgaWYgbGVuKGpzb25fb3V0KSA9PSAxOgogICAgICAgICAgICAgICAgbWF0
            Y2hpbmdfanNvbl9vdXQgPSBqc29uX291dFswXQogICAgICAgICAgICAjIGIuIG9yIGVsc2UsIGNo
            ZWNrIGlmIGRhdGFfcG9vbCBuYW1lIGlzIHByb3ZpZGVkCiAgICAgICAgICAgIGVsaWYgc2VsZi5f
            YXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWU6CiAgICAgICAgICAgICAgICAjIGFuZCBp
            ZiBwcmVzZW50LCBjaGVjayB3aGV0aGVyIHRoZXJlIGV4aXN0cyBhIGZzIHdoaWNoIGhhcyB0aGUg
            ZGF0YV9wb29sCiAgICAgICAgICAgICAgICBmb3IgZWFjaEogaW4ganNvbl9vdXQ6CiAgICAgICAg
            ICAgICAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUgaW4g
            ZWFjaEpbImRhdGFfcG9vbHMiXToKICAgICAgICAgICAgICAgICAgICAgICAgbWF0Y2hpbmdfanNv
            bl9vdXQgPSBlYWNoSgogICAgICAgICAgICAgICAgICAgICAgICBicmVhawogICAgICAgICAgICAg
            ICAgIyBpZiB0aGVyZSBpcyBubyBtYXRjaGluZyBmcyBleGlzdHMsIHRoYXQgbWVhbnMgcHJvdmlk
            ZWQgZGF0YV9wb29sIG5hbWUgaXMgaW52YWxpZAogICAgICAgICAgICAgICAgaWYgbm90IG1hdGNo
            aW5nX2pzb25fb3V0OgogICAgICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVF
            eGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgICAgIGYiUHJvdmlkZWQgZGF0YV9wb29sIG5h
            bWUsIHtzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZX0sIgogICAgICAgICAg
            ICAgICAgICAgICAgICAiIGRvZXMgbm90IGV4aXN0cyIKICAgICAgICAgICAgICAgICAgICApCiAg
            ICAgICAgICAgICMgYy4gaWYgbm90aGluZyBpcyBzZXQgYW5kIGNvdWxkbid0IGZpbmQgYSBkZWZh
            dWx0LAogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgIyBqdXN0IHJldHVybiBzaWxl
            bnRseQogICAgICAgICAgICAgICAgcmV0dXJuCgogICAgICAgIGlmIG1hdGNoaW5nX2pzb25fb3V0
            OgogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUgPSBz
            dHIobWF0Y2hpbmdfanNvbl9vdXRbIm5hbWUiXSkKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5jZXBoZnNfbWV0YWRhdGFfcG9vbF9uYW1lID0gc3RyKAogICAgICAgICAgICAgICAgbWF0Y2hp
            bmdfanNvbl9vdXRbIm1ldGFkYXRhX3Bvb2wiXQogICAgICAgICAgICApCgogICAgICAgIGlmIGlz
            aW5zdGFuY2UobWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXSwgbGlzdCk6CiAgICAgICAg
            ICAgICMgaWYgdGhlIHVzZXIgaGFzIGFscmVhZHkgcHJvdmlkZWQgZGF0YS1wb29sLW5hbWUsCiAg
            ICAgICAgICAgICMgdGhyb3VnaCAtLWNlcGhmcy1kYXRhLXBvb2wtbmFtZQogICAgICAgICAgICBp
            ZiBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZToKICAgICAgICAgICAgICAg
            ICMgaWYgdGhlIHByb3ZpZGVkIG5hbWUgaXMgbm90IG1hdGNoaW5nIHdpdGggdGhlIG9uZSBpbiB0
            aGUgbGlzdAogICAgICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2Fy
            Z19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgICAgICAgICAgbm90IGlu
            IG1hdGNoaW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0KICAgICAgICAgICAgICAgICk6CiAgICAg
            ICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAg
            ICAgICAgICAgICAgICAgZiJQcm92aWRlZCBkYXRhLXBvb2wtbmFtZTogJ3tzZWxmLl9hcmdfcGFy
            c2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZX0nLCAiCiAgICAgICAgICAgICAgICAgICAgICAgICJk
            b2Vzbid0IG1hdGNoIGZyb20gdGhlIGRhdGEtcG9vbHMgbGlzdDogIgogICAgICAgICAgICAgICAg
            ICAgICAgICBmIntbc3RyKHgpIGZvciB4IGluIG1hdGNoaW5nX2pzb25fb3V0WydkYXRhX3Bvb2xz
            J11dfSIKICAgICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICMgaWYgZGF0YV9wb29sIG5h
            bWUgaXMgbm90IHByb3ZpZGVkLAogICAgICAgICAgICAjIHRoZW4gdHJ5IHRvIGZpbmQgYSBkZWZh
            dWx0IGRhdGEgcG9vbCBuYW1lCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICAjIGlm
            IG5vIGRhdGFfcG9vbHMgZXhpc3QsIHNpbGVudGx5IHJldHVybgogICAgICAgICAgICAgICAgaWYg
            bGVuKG1hdGNoaW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0pID09IDA6CiAgICAgICAgICAgICAg
            ICAgICAgcmV0dXJuCiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRh
            X3Bvb2xfbmFtZSA9IHN0cigKICAgICAgICAgICAgICAgICAgICBtYXRjaGluZ19qc29uX291dFsi
            ZGF0YV9wb29scyJdWzBdCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICMgaWYgdGhlcmUg
            YXJlIG1vcmUgdGhhbiBvbmUgJ2RhdGFfcG9vbHMnIGV4aXN0LAogICAgICAgICAgICAjIHRoZW4g
            d2FybiB0aGUgdXNlciB0aGF0IHdlIGFyZSB1c2luZyB0aGUgc2VsZWN0ZWQgbmFtZQogICAgICAg
            ICAgICBpZiBsZW4obWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXSkgPiAxOgogICAgICAg
            ICAgICAgICAgcHJpbnQoCiAgICAgICAgICAgICAgICAgICAgIldBUk5JTkc6IE11bHRpcGxlIGRh
            dGEgcG9vbHMgZGV0ZWN0ZWQ6ICIKICAgICAgICAgICAgICAgICAgICBmIntbc3RyKHgpIGZvciB4
            IGluIG1hdGNoaW5nX2pzb25fb3V0WydkYXRhX3Bvb2xzJ11dfVxuIgogICAgICAgICAgICAgICAg
            ICAgIGYiVXNpbmcgdGhlIGRhdGEtcG9vbDogJ3tzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRh
            X3Bvb2xfbmFtZX0nXG4iCiAgICAgICAgICAgICAgICApCgogICAgZGVmIGNyZWF0ZV9jaGVja2Vy
            S2V5KHNlbGYsIHVzZXIpOgogICAgICAgIGNhcHMsIGVudGl0eSA9IHNlbGYuZ2V0X2NhcHNfYW5k
            X2VudGl0eSh1c2VyKQogICAgICAgIGNtZF9qc29uID0gewogICAgICAgICAgICAicHJlZml4Ijog
            ImF1dGggZ2V0LW9yLWNyZWF0ZSIsCiAgICAgICAgICAgICJlbnRpdHkiOiBlbnRpdHksCiAgICAg
            ICAgICAgICJjYXBzIjogW2NhcCBmb3IgY2FwX2xpc3QgaW4gbGlzdChjYXBzLml0ZW1zKCkpIGZv
            ciBjYXAgaW4gY2FwX2xpc3RdLAogICAgICAgICAgICAiZm9ybWF0IjogImpzb24iLAogICAgICAg
            IH0KCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1
            cm4gc2VsZi5kcnlfcnVuKAogICAgICAgICAgICAgICAgImNlcGggIgogICAgICAgICAgICAgICAg
            KyBjbWRfanNvblsicHJlZml4Il0KICAgICAgICAgICAgICAgICsgIiAiCiAgICAgICAgICAgICAg
            ICArIGNtZF9qc29uWyJlbnRpdHkiXQogICAgICAgICAgICAgICAgKyAiICIKICAgICAgICAgICAg
            ICAgICsgIiAiLmpvaW4oY21kX2pzb25bImNhcHMiXSkKICAgICAgICAgICAgKQogICAgICAgICMg
            Y2hlY2sgaWYgdXNlciBhbHJlYWR5IGV4aXN0CiAgICAgICAgdXNlcl9rZXkgPSBzZWxmLmNoZWNr
            X3VzZXJfZXhpc3QoZW50aXR5KQogICAgICAgIGlmIHVzZXJfa2V5ICE9ICIiOgogICAgICAgICAg
            ICByZXR1cm4gdXNlcl9rZXkKCiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBz
            ZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICMgaWYgdGhlcmUgaXMg
            YW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihq
            c29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgICAgIGYiJ2F1dGggZ2V0LW9yLWNyZWF0ZSB7c2VsZi5ydW5fYXNfdXNl
            cn0nIGNvbW1hbmQgZmFpbGVkXG4iCiAgICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBp
            ZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAg
            KQogICAgICAgIHJldHVybiBzdHIoanNvbl9vdXRbMF1bImtleSJdKQoKICAgIGRlZiBnZXRfY2Vw
            aF9kYXNoYm9hcmRfbGluayhzZWxmKToKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4IjogIm1n
            ciBzZXJ2aWNlcyIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNl
            ci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKCJjZXBoICIgKyBjbWRf
            anNvblsicHJlZml4Il0pCiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIF8gPSBzZWxmLl9jb21t
            b25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNj
            ZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkg
            PT0gMDoKICAgICAgICAgICAgcmV0dXJuIE5vbmUKICAgICAgICBpZiAiZGFzaGJvYXJkIiBub3Qg
            aW4ganNvbl9vdXQ6CiAgICAgICAgICAgIHJldHVybiBOb25lCiAgICAgICAgcmV0dXJuIGpzb25f
            b3V0WyJkYXNoYm9hcmQiXQoKICAgIGRlZiBjcmVhdGVfcmd3X2FkbWluX29wc191c2VyKHNlbGYp
            OgogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAgICAgICAg
            ICAidXNlciIsCiAgICAgICAgICAgICJjcmVhdGUiLAogICAgICAgICAgICAiLS11aWQiLAogICAg
            ICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9PUFNfVVNFUl9OQU1FLAogICAgICAgICAg
            ICAiLS1kaXNwbGF5LW5hbWUiLAogICAgICAgICAgICAiUm9vayBSR1cgQWRtaW4gT3BzIHVzZXIi
            LAogICAgICAgICAgICAiLS1jYXBzIiwKICAgICAgICAgICAgImJ1Y2tldHM9Kjt1c2Vycz0qO3Vz
            YWdlPXJlYWQ7bWV0YWRhdGE9cmVhZDt6b25lPXJlYWQiLAogICAgICAgICAgICAiLS1yZ3ctcmVh
            bG0iLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lLAogICAgICAg
            ICAgICAiLS1yZ3ctem9uZWdyb3VwIiwKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3df
            em9uZWdyb3VwX25hbWUsCiAgICAgICAgICAgICItLXJndy16b25lIiwKICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1lLAogICAgICAgIF0KICAgICAgICBpZiBzZWxmLl9h
            cmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGgg
            IiArICIgIi5qb2luKGNtZCkpCiAgICAgICAgdHJ5OgogICAgICAgICAgICBvdXRwdXQgPSBzdWJw
            cm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAg
            ZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yIGFzIGV4ZWNFcnI6CiAgICAgICAg
            ICAgICMgaWYgdGhlIHVzZXIgYWxyZWFkeSBleGlzdHMsIHdlIGp1c3QgcXVlcnkgaXQKICAgICAg
            ICAgICAgaWYgZXhlY0Vyci5yZXR1cm5jb2RlID09IGVycm5vLkVFWElTVDoKICAgICAgICAgICAg
            ICAgIGNtZCA9IFsKICAgICAgICAgICAgICAgICAgICAicmFkb3Nndy1hZG1pbiIsCiAgICAgICAg
            ICAgICAgICAgICAgInVzZXIiLAogICAgICAgICAgICAgICAgICAgICJpbmZvIiwKICAgICAgICAg
            ICAgICAgICAgICAiLS11aWQiLAogICAgICAgICAgICAgICAgICAgIHNlbGYuRVhURVJOQUxfUkdX
            X0FETUlOX09QU19VU0VSX05BTUUsCiAgICAgICAgICAgICAgICAgICAgIi0tcmd3LXJlYWxtIiwK
            ICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lLAogICAg
            ICAgICAgICAgICAgICAgICItLXJndy16b25lZ3JvdXAiLAogICAgICAgICAgICAgICAgICAgIHNl
            bGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lLAogICAgICAgICAgICAgICAgICAgICIt
            LXJndy16b25lIiwKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25l
            X25hbWUsCiAgICAgICAgICAgICAgICBdCiAgICAgICAgICAgICAgICB0cnk6CiAgICAgICAgICAg
            ICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3Vi
            cHJvY2Vzcy5QSVBFKQogICAgICAgICAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJv
            Y2Vzc0Vycm9yIGFzIGV4ZWNFcnI6CiAgICAgICAgICAgICAgICAgICAgZXJyX21zZyA9ICgKICAg
            ICAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5kIHtjbWR9LiBP
            dXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICAgICAgZiJDb2Rl
            OiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtleGVjRXJyLnN0ZGVycn0iCiAgICAgICAg
            ICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21z
            ZykKICAgICAgICAgICAgICAgICAgICByZXR1cm4gTm9uZSwgTm9uZSwgRmFsc2UsICItMSIKICAg
            ICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVycl9tc2cgPSAoCiAgICAgICAgICAgICAg
            ICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5kIHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJy
            Lm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICBmIkNvZGU6IHtleGVjRXJyLnJldHVybmNv
            ZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykKICAgICAgICAgICAgICAgIHJldHVybiBO
            b25lLCBOb25lLCBGYWxzZSwgIi0xIgoKICAgICAgICAjIGlmIGl0IGlzIHB5dGhvbjIsIGRvbid0
            IGNoZWNrIGZvciBjZXBoIHZlcnNpb24gZm9yIGFkZGluZyBgaW5mbz1yZWFkYCBjYXAocmd3X3Zh
            bGlkYXRpb24pCiAgICAgICAgaWYgc3lzLnZlcnNpb25faW5mby5tYWpvciA8IDM6CiAgICAgICAg
            ICAgIGpzb25vdXRwdXQgPSBqc29uLmxvYWRzKG91dHB1dCkKICAgICAgICAgICAgcmV0dXJuICgK
            ICAgICAgICAgICAgICAgIGpzb25vdXRwdXRbImtleXMiXVswXVsiYWNjZXNzX2tleSJdLAogICAg
            ICAgICAgICAgICAganNvbm91dHB1dFsia2V5cyJdWzBdWyJzZWNyZXRfa2V5Il0sCiAgICAgICAg
            ICAgICAgICBGYWxzZSwKICAgICAgICAgICAgICAgICIiLAogICAgICAgICAgICApCgogICAgICAg
            ICMgc2VwYXJhdGVseSBhZGQgaW5mbz1yZWFkIGNhcHMgZm9yIHJndy1lbmRwb2ludCBpcCB2YWxp
            ZGF0aW9uCiAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkID0gVHJ1ZQogICAgICAgIGNtZCA9IFsK
            ICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAgICAgICAgICAiY2FwcyIsCiAgICAgICAg
            ICAgICJhZGQiLAogICAgICAgICAgICAiLS11aWQiLAogICAgICAgICAgICBzZWxmLkVYVEVSTkFM
            X1JHV19BRE1JTl9PUFNfVVNFUl9OQU1FLAogICAgICAgICAgICAiLS1jYXBzIiwKICAgICAgICAg
            ICAgImluZm89cmVhZCIsCiAgICAgICAgICAgICItLXJndy1yZWFsbSIsCiAgICAgICAgICAgIHNl
            bGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUsCiAgICAgICAgICAgICItLXJndy16b25lZ3Jv
            dXAiLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSwKICAg
            ICAgICAgICAgIi0tcmd3LXpvbmUiLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196
            b25lX25hbWUsCiAgICAgICAgXQogICAgICAgIHRyeToKICAgICAgICAgICAgb3V0cHV0ID0gc3Vi
            cHJvY2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAg
            IGV4Y2VwdCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJvciBhcyBleGVjRXJyOgogICAgICAg
            ICAgICAjIGlmIHRoZSBjZXBoIHZlcnNpb24gbm90IHN1cHBvcnRlZCBmb3IgYWRkaW5nIGBpbmZv
            PXJlYWRgIGNhcChyZ3dfdmFsaWRhdGlvbikKICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAg
            ICAgImNvdWxkIG5vdCBhZGQgY2FwczogdW5hYmxlIHRvIGFkZCBjYXBzOiBpbmZvPXJlYWRcbiIK
            ICAgICAgICAgICAgICAgIGluIGV4ZWNFcnIuc3RkZXJyLmRlY29kZSgidXRmLTgiKQogICAgICAg
            ICAgICAgICAgYW5kIGV4ZWNFcnIucmV0dXJuY29kZSA9PSAyNDQKICAgICAgICAgICAgKToKICAg
            ICAgICAgICAgICAgIGluZm9fY2FwX3N1cHBvcnRlZCA9IEZhbHNlCiAgICAgICAgICAgIGVsc2U6
            CiAgICAgICAgICAgICAgICBlcnJfbXNnID0gKAogICAgICAgICAgICAgICAgICAgIGYiZmFpbGVk
            IHRvIGV4ZWN1dGUgY29tbWFuZCB7Y21kfS4gT3V0cHV0OiB7ZXhlY0Vyci5vdXRwdXR9LiAiCiAg
            ICAgICAgICAgICAgICAgICAgZiJDb2RlOiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtl
            eGVjRXJyLnN0ZGVycn0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICBzeXMuc3Rk
            ZXJyLndyaXRlKGVycl9tc2cpCiAgICAgICAgICAgICAgICByZXR1cm4gTm9uZSwgTm9uZSwgRmFs
            c2UsICItMSIKCiAgICAgICAganNvbm91dHB1dCA9IGpzb24ubG9hZHMob3V0cHV0KQogICAgICAg
            IHJldHVybiAoCiAgICAgICAgICAgIGpzb25vdXRwdXRbImtleXMiXVswXVsiYWNjZXNzX2tleSJd
            LAogICAgICAgICAgICBqc29ub3V0cHV0WyJrZXlzIl1bMF1bInNlY3JldF9rZXkiXSwKICAgICAg
            ICAgICAgaW5mb19jYXBfc3VwcG9ydGVkLAogICAgICAgICAgICAiIiwKICAgICAgICApCgogICAg
            ZGVmIHZhbGlkYXRlX3JiZF9wb29sKHNlbGYsIHBvb2xfbmFtZSk6CiAgICAgICAgaWYgbm90IHNl
            bGYuY2x1c3Rlci5wb29sX2V4aXN0cyhwb29sX25hbWUpOgogICAgICAgICAgICByYWlzZSBFeGVj
            dXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcG9v
            bCwgJ3twb29sX25hbWV9JywgZG9lcyBub3QgZXhpc3QiCiAgICAgICAgICAgICkKCiAgICBkZWYg
            aW5pdF9yYmRfcG9vbChzZWxmLCByYmRfcG9vbF9uYW1lKToKICAgICAgICBpZiBpc2luc3RhbmNl
            KHNlbGYuY2x1c3RlciwgRHVtbXlSYWRvcyk6CiAgICAgICAgICAgIHJldHVybgogICAgICAgIGlv
            Y3R4ID0gc2VsZi5jbHVzdGVyLm9wZW5faW9jdHgocmJkX3Bvb2xfbmFtZSkKICAgICAgICByYmRf
            aW5zdCA9IHJiZC5SQkQoKQogICAgICAgIHJiZF9pbnN0LnBvb2xfaW5pdChpb2N0eCwgVHJ1ZSkK
            CiAgICBkZWYgdmFsaWRhdGVfcmFkb3NfbmFtZXNwYWNlKHNlbGYpOgogICAgICAgIHJiZF9wb29s
            X25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgIHJhZG9z
            X25hbWVzcGFjZSA9IHNlbGYuX2FyZ19wYXJzZXIucmFkb3NfbmFtZXNwYWNlCiAgICAgICAgaWYg
            cmFkb3NfbmFtZXNwYWNlID09ICIiOgogICAgICAgICAgICByZXR1cm4KICAgICAgICBpZiByYWRv
            c19uYW1lc3BhY2UuaXNsb3dlcigpID09IEZhbHNlOgogICAgICAgICAgICByYWlzZSBFeGVjdXRp
            b25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmFkb3Mg
            TmFtZXNwYWNlLCAne3JhZG9zX25hbWVzcGFjZX0nLCAiCiAgICAgICAgICAgICAgICBmImNvbnRh
            aW5zIHVwcGVyIGNhc2UiCiAgICAgICAgICAgICkKICAgICAgICByYmRfaW5zdCA9IHJiZC5SQkQo
            KQogICAgICAgIGlvY3R4ID0gc2VsZi5jbHVzdGVyLm9wZW5faW9jdHgocmJkX3Bvb2xfbmFtZSkK
            ICAgICAgICBpZiByYmRfaW5zdC5uYW1lc3BhY2VfZXhpc3RzKGlvY3R4LCByYWRvc19uYW1lc3Bh
            Y2UpIGlzIEZhbHNlOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmFkb3MgTmFtZXNwYWNlLCAne3JhZG9z
            X25hbWVzcGFjZX0nLCAiCiAgICAgICAgICAgICAgICBmImlzIG5vdCBmb3VuZCBpbiB0aGUgcG9v
            bCAne3JiZF9wb29sX25hbWV9JyIKICAgICAgICAgICAgKQoKICAgIGRlZiBnZXRfb3JfY3JlYXRl
            X3N1YnZvbHVtZV9ncm91cChzZWxmLCBzdWJ2b2x1bWVfZ3JvdXAsIGNlcGhmc19maWxlc3lzdGVt
            X25hbWUpOgogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgImNlcGgiLAogICAgICAgICAgICAi
            ZnMiLAogICAgICAgICAgICAic3Vidm9sdW1lZ3JvdXAiLAogICAgICAgICAgICAiZ2V0cGF0aCIs
            CiAgICAgICAgICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgIHN1YnZvbHVt
            ZV9ncm91cCwKICAgICAgICBdCiAgICAgICAgdHJ5OgogICAgICAgICAgICBfID0gc3VicHJvY2Vz
            cy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2Vw
            dCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJvcjoKICAgICAgICAgICAgY21kID0gWwogICAg
            ICAgICAgICAgICAgImNlcGgiLAogICAgICAgICAgICAgICAgImZzIiwKICAgICAgICAgICAgICAg
            ICJzdWJ2b2x1bWVncm91cCIsCiAgICAgICAgICAgICAgICAiY3JlYXRlIiwKICAgICAgICAgICAg
            ICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgICAgICBzdWJ2b2x1bWVfZ3Jv
            dXAsCiAgICAgICAgICAgIF0KICAgICAgICAgICAgdHJ5OgogICAgICAgICAgICAgICAgXyA9IHN1
            YnByb2Nlc3MuY2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAgICAg
            ICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBm
            InN1YnZvbHVtZSBncm91cCB7c3Vidm9sdW1lX2dyb3VwfSBpcyBub3QgYWJsZSB0byBnZXQgY3Jl
            YXRlZCIKICAgICAgICAgICAgICAgICkKCiAgICBkZWYgcGluX3N1YnZvbHVtZSgKICAgICAgICBz
            ZWxmLCBzdWJ2b2x1bWVfZ3JvdXAsIGNlcGhmc19maWxlc3lzdGVtX25hbWUsIHBpbl90eXBlLCBw
            aW5fc2V0dGluZwogICAgKToKICAgICAgICBjbWQgPSBbCiAgICAgICAgICAgICJjZXBoIiwKICAg
            ICAgICAgICAgImZzIiwKICAgICAgICAgICAgInN1YnZvbHVtZWdyb3VwIiwKICAgICAgICAgICAg
            InBpbiIsCiAgICAgICAgICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgIHN1
            YnZvbHVtZV9ncm91cCwKICAgICAgICAgICAgcGluX3R5cGUsCiAgICAgICAgICAgIHBpbl9zZXR0
            aW5nLAogICAgICAgIF0KICAgICAgICB0cnk6CiAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNo
            ZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1
            YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJzdWJ2b2x1bWUgZ3JvdXAge3N1YnZv
            bHVtZV9ncm91cH0gaXMgbm90IGFibGUgdG8gZ2V0IHBpbm5lZCIKICAgICAgICAgICAgKQoKICAg
            IGRlZiBnZXRfcmd3X2ZzaWQoc2VsZiwgYmFzZV91cmwsIHZlcmlmeSk6CiAgICAgICAgYWNjZXNz
            X2tleSA9IHNlbGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX0FDQ0VTU19LRVkiXQogICAg
            ICAgIHNlY3JldF9rZXkgPSBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9TRUNSRVRf
            S0VZIl0KICAgICAgICByZ3dfZW5kcG9pbnQgPSBzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2lu
            dAogICAgICAgIGJhc2VfdXJsID0gYmFzZV91cmwgKyAiOi8vIiArIHJnd19lbmRwb2ludCArICIv
            YWRtaW4vaW5mbz8iCiAgICAgICAgcGFyYW1zID0geyJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAg
            cmVxdWVzdF91cmwgPSBiYXNlX3VybCArIHVybGVuY29kZShwYXJhbXMpCgogICAgICAgIHRyeToK
            ICAgICAgICAgICAgciA9IHJlcXVlc3RzLmdldCgKICAgICAgICAgICAgICAgIHJlcXVlc3RfdXJs
            LAogICAgICAgICAgICAgICAgYXV0aD1TM0F1dGgoYWNjZXNzX2tleSwgc2VjcmV0X2tleSwgcmd3
            X2VuZHBvaW50KSwKICAgICAgICAgICAgICAgIHZlcmlmeT12ZXJpZnksCiAgICAgICAgICAgICkK
            ICAgICAgICBleGNlcHQgcmVxdWVzdHMuZXhjZXB0aW9ucy5UaW1lb3V0OgogICAgICAgICAgICBz
            eXMuc3RkZXJyLndyaXRlKAogICAgICAgICAgICAgICAgZiJpbnZhbGlkIGVuZHBvaW50Oiwgbm90
            IGFibGUgdG8gY2FsbCBhZG1pbi1vcHMgYXBpe3Jnd19lbmRwb2ludH0iCiAgICAgICAgICAgICkK
            ICAgICAgICAgICAgcmV0dXJuICIiLCAiLTEiCiAgICAgICAgcjEgPSByLmpzb24oKQogICAgICAg
            IGlmIHIxIGlzIE5vbmUgb3IgcjEuZ2V0KCJpbmZvIikgaXMgTm9uZToKICAgICAgICAgICAgc3lz
            LnN0ZGVyci53cml0ZSgKICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHJndyBFbmRwb2lu
            dCwgJ3tzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludH0nLCBpcyBpbnZhbGlkLiIKICAgICAg
            ICAgICAgKQogICAgICAgICAgICByZXR1cm4gKAogICAgICAgICAgICAgICAgIiIsCiAgICAgICAg
            ICAgICAgICAiLTEiLAogICAgICAgICAgICApCgogICAgICAgIHJldHVybiByMVsiaW5mbyJdWyJz
            dG9yYWdlX2JhY2tlbmRzIl1bMF1bImNsdXN0ZXJfaWQiXSwgIiIKCiAgICBkZWYgdmFsaWRhdGVf
            cmd3X2VuZHBvaW50KHNlbGYsIGluZm9fY2FwX3N1cHBvcnRlZCk6CiAgICAgICAgIyBpZiB0aGUg
            J2NsdXN0ZXInIGluc3RhbmNlIGlzIGEgZHVtbXkgb25lLAogICAgICAgICMgZG9uJ3QgdHJ5IHRv
            IHJlYWNoIG91dCB0byB0aGUgZW5kcG9pbnQKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1
            c3RlciwgRHVtbXlSYWRvcyk6CiAgICAgICAgICAgIHJldHVybgoKICAgICAgICByZ3dfZW5kcG9p
            bnQgPSBzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludAoKICAgICAgICAjIHZhbGlkYXRlIHJn
            dyBlbmRwb2ludCBvbmx5IGlmIGlwIGFkZHJlc3MgaXMgcGFzc2VkCiAgICAgICAgaXBfdHlwZSA9
            IHNlbGYuX2ludmFsaWRfZW5kcG9pbnQocmd3X2VuZHBvaW50KQoKICAgICAgICAjIGNoZWNrIGlm
            IHRoZSByZ3cgZW5kcG9pbnQgaXMgcmVhY2hhYmxlCiAgICAgICAgY2VydCA9IE5vbmUKICAgICAg
            ICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNlci5yZ3dfc2tpcF90bHMgYW5kIHNlbGYudmFsaWRhdGVf
            cmd3X2VuZHBvaW50X3Rsc19jZXJ0KCk6CiAgICAgICAgICAgIGNlcnQgPSBzZWxmLl9hcmdfcGFy
            c2VyLnJnd190bHNfY2VydF9wYXRoCiAgICAgICAgYmFzZV91cmwsIHZlcmlmeSwgZXJyID0gc2Vs
            Zi5lbmRwb2ludF9kaWFsKHJnd19lbmRwb2ludCwgaXBfdHlwZSwgY2VydD1jZXJ0KQogICAgICAg
            IGlmIGVyciAhPSAiIjoKICAgICAgICAgICAgcmV0dXJuICItMSIKCiAgICAgICAgIyBjaGVjayBp
            ZiB0aGUgcmd3IGVuZHBvaW50IGJlbG9uZ3MgdG8gdGhlIHNhbWUgY2x1c3RlcgogICAgICAgICMg
            b25seSBjaGVjayBpZiBgaW5mb2AgY2FwIGlzIHN1cHBvcnRlZAogICAgICAgIGlmIGluZm9fY2Fw
            X3N1cHBvcnRlZDoKICAgICAgICAgICAgZnNpZCA9IHNlbGYuZ2V0X2ZzaWQoKQogICAgICAgICAg
            ICByZ3dfZnNpZCwgZXJyID0gc2VsZi5nZXRfcmd3X2ZzaWQoYmFzZV91cmwsIHZlcmlmeSkKICAg
            ICAgICAgICAgaWYgZXJyID09ICItMSI6CiAgICAgICAgICAgICAgICByZXR1cm4gIi0xIgogICAg
            ICAgICAgICBpZiBmc2lkICE9IHJnd19mc2lkOgogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53
            cml0ZSgKICAgICAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCByZ3cgRW5kcG9pbnQsICd7
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfZW5kcG9pbnR9JywgaXMgaW52YWxpZC4gV2UgYXJlIHZhbGlk
            YXRpbmcgYnkgY2FsbGluZyB0aGUgYWRtaW5vcHMgYXBpIHRocm91Z2ggcmd3LWVuZHBvaW50IGFu
            ZCB2YWxpZGF0aW5nIHRoZSBjbHVzdGVyX2lkICd7cmd3X2ZzaWR9JyBpcyBlcXVhbCB0byB0aGUg
            Y2VwaCBjbHVzdGVyIGZzaWQgJ3tmc2lkfSciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAg
            ICAgICByZXR1cm4gIi0xIgoKICAgICAgICAjIGNoZWNrIGlmIHRoZSByZ3cgZW5kcG9pbnQgcG9v
            bCBleGlzdAogICAgICAgICMgb25seSB2YWxpZGF0ZSBpZiByZ3dfcG9vbF9wcmVmaXggaXMgcGFz
            c2VkIGVsc2UgaXQgd2lsbCB0YWtlIGRlZmF1bHQgdmFsdWUgYW5kIHdlIGRvbid0IGNyZWF0ZSB0
            aGVzZSBkZWZhdWx0IHBvb2xzCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9w
            cmVmaXggIT0gImRlZmF1bHQiOgogICAgICAgICAgICByZ3dfcG9vbHNfdG9fdmFsaWRhdGUgPSBb
            CiAgICAgICAgICAgICAgICBmIntzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3
            Lm1ldGEiLAogICAgICAgICAgICAgICAgIi5yZ3cucm9vdCIsCiAgICAgICAgICAgICAgICBmIntz
            ZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmNvbnRyb2wiLAogICAgICAgICAg
            ICAgICAgZiJ7c2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5sb2ciLAogICAg
            ICAgICAgICBdCiAgICAgICAgICAgIGZvciBfcmd3X3Bvb2xfdG9fdmFsaWRhdGUgaW4gcmd3X3Bv
            b2xzX3RvX3ZhbGlkYXRlOgogICAgICAgICAgICAgICAgaWYgbm90IHNlbGYuY2x1c3Rlci5wb29s
            X2V4aXN0cyhfcmd3X3Bvb2xfdG9fdmFsaWRhdGUpOgogICAgICAgICAgICAgICAgICAgIHN5cy5z
            dGRlcnIud3JpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHBvb2ws
            ICd7X3Jnd19wb29sX3RvX3ZhbGlkYXRlfScsIGRvZXMgbm90IGV4aXN0IgogICAgICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgICAgICAgICByZXR1cm4gIi0xIgoKICAgICAgICByZXR1cm4g
            IiIKCiAgICBkZWYgdmFsaWRhdGVfcmd3X211bHRpc2l0ZShzZWxmLCByZ3dfbXVsdGlzaXRlX2Nv
            bmZpZ19uYW1lLCByZ3dfbXVsdGlzaXRlX2NvbmZpZyk6CiAgICAgICAgaWYgcmd3X211bHRpc2l0
            ZV9jb25maWcgIT0gIiI6CiAgICAgICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgICAgICJyYWRv
            c2d3LWFkbWluIiwKICAgICAgICAgICAgICAgIHJnd19tdWx0aXNpdGVfY29uZmlnLAogICAgICAg
            ICAgICAgICAgImdldCIsCiAgICAgICAgICAgICAgICAiLS1yZ3ctIiArIHJnd19tdWx0aXNpdGVf
            Y29uZmlnLAogICAgICAgICAgICAgICAgcmd3X211bHRpc2l0ZV9jb25maWdfbmFtZSwKICAgICAg
            ICAgICAgXQogICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBfID0gc3VicHJvY2Vzcy5j
            aGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAgICBleGNl
            cHQgc3VicHJvY2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3IgYXMgZXhlY0VycjoKICAgICAgICAgICAg
            ICAgIGVycl9tc2cgPSAoCiAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBj
            b21tYW5kIHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAgICAgICAgICAg
            ICAgICBmIkNvZGU6IHtleGVjRXJyLnJldHVybmNvZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJy
            fSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJy
            X21zZykKICAgICAgICAgICAgICAgIHJldHVybiAiLTEiCiAgICAgICAgcmV0dXJuICIiCgogICAg
            ZGVmIGNvbnZlcnRfY29tbWFfc2VwYXJhdGVkX3RvX2FycmF5KHNlbGYsIHZhbHVlKToKICAgICAg
            ICByZXR1cm4gdmFsdWUuc3BsaXQoIiwiKQoKICAgIGRlZiByYWlzZV9leGNlcHRpb25faWZfYW55
            X3RvcG9sb2d5X2ZsYWdfaXNfbWlzc2luZyhzZWxmKToKICAgICAgICBpZiAoCiAgICAgICAgICAg
            ICgKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMgIT0gIiIK
            ICAgICAgICAgICAgICAgIGFuZCAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCA9PSAiIgogICAgICAgICAgICAgICAgICAg
            IG9yIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfZmFpbHVyZV9kb21haW5fdmFsdWVzID09ICIi
            CiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICkKICAgICAgICAgICAgb3IgKAogICAgICAg
            ICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCAh
            PSAiIgogICAgICAgICAgICAgICAgYW5kICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdf
            cGFyc2VyLnRvcG9sb2d5X3Bvb2xzID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5f
            YXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMgPT0gIiIKICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgKQogICAgICAgICAgICBvciAoCiAgICAgICAgICAgICAgICBz
            ZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX3ZhbHVlcyAhPSAiIgogICAg
            ICAgICAgICAgICAgYW5kICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRv
            cG9sb2d5X3Bvb2xzID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5fYXJnX3BhcnNl
            ci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCA9PSAiIgogICAgICAgICAgICAgICAgKQog
            ICAgICAgICAgICApCiAgICAgICAgKToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVy
            ZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICJwcm92aWRlIGFsbCB0aGUgdG9wb2xvZ3kgZmxh
            Z3MgLS10b3BvbG9neS1wb29scywgLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi1sYWJlbCwgLS10
            b3BvbG9neS1mYWlsdXJlLWRvbWFpbi12YWx1ZXMiCiAgICAgICAgICAgICkKCiAgICBkZWYgdmFs
            aWRhdGVfdG9wb2xvZ3lfdmFsdWVzKHNlbGYsIHRvcG9sb2d5X3Bvb2xzLCB0b3BvbG9neV9mZCk6
            CiAgICAgICAgaWYgbGVuKHRvcG9sb2d5X3Bvb2xzKSAhPSBsZW4odG9wb2xvZ3lfZmQpOgogICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ZiJUaGUgcHJvdmlkZWQgdG9wb2xvZ3kgcG9vbHMsICd7dG9wb2xvZ3lfcG9vbHN9JywgYW5kICIK
            ICAgICAgICAgICAgICAgIGYidG9wb2xvZ3kgZmFpbHVyZSBkb21haW4sICd7dG9wb2xvZ3lfZmR9
            JywiCiAgICAgICAgICAgICAgICBmImFyZSBvZiBkaWZmZXJlbnQgbGVuZ3RoLCAne2xlbih0b3Bv
            bG9neV9wb29scyl9JyBhbmQgJ3tsZW4odG9wb2xvZ3lfZmQpfScgcmVzcGN0aXZlbHkiCiAgICAg
            ICAgICAgICkKICAgICAgICByZXR1cm4KCiAgICBkZWYgdmFsaWRhdGVfdG9wb2xvZ3lfcmJkX3Bv
            b2xzKHNlbGYsIHRvcG9sb2d5X3JiZF9wb29scyk6CiAgICAgICAgZm9yIHBvb2wgaW4gdG9wb2xv
            Z3lfcmJkX3Bvb2xzOgogICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3JiZF9wb29sKHBvb2wpCgog
            ICAgZGVmIGluaXRfdG9wb2xvZ3lfcmJkX3Bvb2xzKHNlbGYsIHRvcG9sb2d5X3JiZF9wb29scyk6
            CiAgICAgICAgZm9yIHBvb2wgaW4gdG9wb2xvZ3lfcmJkX3Bvb2xzOgogICAgICAgICAgICBzZWxm
            LmluaXRfcmJkX3Bvb2wocG9vbCkKCiAgICBkZWYgX2dlbl9vdXRwdXRfbWFwKHNlbGYpOgogICAg
            ICAgIGlmIHNlbGYub3V0X21hcDoKICAgICAgICAgICAgcmV0dXJuCiAgICAgICAgIyBzdXBwb3J0
            IGxlZ2FjeSBmbGFnIHdpdGggdXBncmFkZXMKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNs
            dXN0ZXJfbmFtZToKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1l
            ID0gc2VsZi5fYXJnX3BhcnNlci5jbHVzdGVyX25hbWUKICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            Lms4c19jbHVzdGVyX25hbWUgPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuazhzX2Ns
            dXN0ZXJfbmFtZS5sb3dlcigpCiAgICAgICAgKSAgIyBhbHdheXMgY29udmVydCBjbHVzdGVyIG5h
            bWUgdG8gbG93ZXJjYXNlIGNoYXJhY3RlcnMKICAgICAgICBzZWxmLnZhbGlkYXRlX3JiZF9wb29s
            KHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lKQogICAgICAgIHNlbGYuaW5pdF9y
            YmRfcG9vbChzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZSkKICAgICAgICBzZWxm
            LnZhbGlkYXRlX3JhZG9zX25hbWVzcGFjZSgpCiAgICAgICAgc2VsZi5fZXhjbHVkZWRfa2V5cy5h
            ZGQoIks4U19DTFVTVEVSX05BTUUiKQogICAgICAgIHNlbGYuZ2V0X2NlcGhmc19kYXRhX3Bvb2xf
            ZGV0YWlscygpCiAgICAgICAgc2VsZi5vdXRfbWFwWyJOQU1FU1BBQ0UiXSA9IHNlbGYuX2FyZ19w
            YXJzZXIubmFtZXNwYWNlCiAgICAgICAgc2VsZi5vdXRfbWFwWyJLOFNfQ0xVU1RFUl9OQU1FIl0g
            PSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBb
            IlJPT0tfRVhURVJOQUxfRlNJRCJdID0gc2VsZi5nZXRfZnNpZCgpCiAgICAgICAgc2VsZi5vdXRf
            bWFwWyJST09LX0VYVEVSTkFMX1VTRVJOQU1FIl0gPSBzZWxmLnJ1bl9hc191c2VyCiAgICAgICAg
            c2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0NFUEhfTU9OX0RBVEEiXSA9IHNlbGYuZ2V0X2Nl
            cGhfZXh0ZXJuYWxfbW9uX2RhdGEoKQogICAgICAgIHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5B
            TF9VU0VSX1NFQ1JFVCJdID0gc2VsZi5jcmVhdGVfY2hlY2tlcktleSgKICAgICAgICAgICAgImNs
            aWVudC5oZWFsdGhjaGVja2VyIgogICAgICAgICkKICAgICAgICBzZWxmLm91dF9tYXBbIlJPT0tf
            RVhURVJOQUxfREFTSEJPQVJEX0xJTksiXSA9IHNlbGYuZ2V0X2NlcGhfZGFzaGJvYXJkX2xpbmso
            KQogICAgICAgICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VDUkVU
            Il0sCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FIl0s
            CiAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3VzZXIoImNsaWVudC5jc2kt
            cmJkLW5vZGUiKQogICAgICAgICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BS
            T1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9SQkRfUFJP
            VklTSU9ORVJfU0VDUkVUX05BTUUiXSwKICAgICAgICApID0gc2VsZi5jcmVhdGVfY2VwaENTSUtl
            eXJpbmdfdXNlcigiY2xpZW50LmNzaS1yYmQtcHJvdmlzaW9uZXIiKQogICAgICAgIHNlbGYub3V0
            X21hcFsiQ0VQSEZTX1BPT0xfTkFNRSJdID0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9w
            b29sX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBbIkNFUEhGU19NRVRBREFUQV9QT09MX05BTUUi
            XSA9ICgKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfbWV0YWRhdGFfcG9vbF9u
            YW1lCiAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSA9IHNl
            bGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgIHNlbGYub3V0X21h
            cFsiUkVTVFJJQ1RFRF9BVVRIX1BFUk1JU1NJT04iXSA9ICgKICAgICAgICAgICAgc2VsZi5fYXJn
            X3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbgogICAgICAgICkKICAgICAgICBzZWxm
            Lm91dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdID0gc2VsZi5fYXJnX3BhcnNlci5yYWRvc19uYW1l
            c3BhY2UKICAgICAgICBzZWxmLm91dF9tYXBbIlNVQlZPTFVNRV9HUk9VUCJdID0gc2VsZi5fYXJn
            X3BhcnNlci5zdWJ2b2x1bWVfZ3JvdXAKICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNf
            Tk9ERV9TRUNSRVQiXSA9ICIiCiAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJ
            U0lPTkVSX1NFQ1JFVCJdID0gIiIKICAgICAgICAjIGNyZWF0ZSBDZXBoRlMgbm9kZSBhbmQgcHJv
            dmlzaW9uZXIga2V5cmluZyBvbmx5IHdoZW4gTURTIGV4aXN0cwogICAgICAgIGlmIHNlbGYub3V0
            X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSBhbmQgc2VsZi5vdXRfbWFwWyJDRVBIRlNfUE9PTF9OQU1F
            Il06CiAgICAgICAgICAgICgKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhG
            U19OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZT
            X05PREVfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lL
            ZXlyaW5nX3VzZXIoImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiKQogICAgICAgICAgICAoCiAgICAg
            ICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUIl0s
            CiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VD
            UkVUX05BTUUiXSwKICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3Vz
            ZXIoImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIikKICAgICAgICAgICAgIyBjcmVhdGUg
            dGhlIGRlZmF1bHQgImNzaSIgc3Vidm9sdW1lZ3JvdXAKICAgICAgICAgICAgc2VsZi5nZXRfb3Jf
            Y3JlYXRlX3N1YnZvbHVtZV9ncm91cCgKICAgICAgICAgICAgICAgICJjc2kiLCBzZWxmLl9hcmdf
            cGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICAgICAgKQogICAgICAgICAgICAj
            IHBpbiB0aGUgZGVmYXVsdCAiY3NpIiBzdWJ2b2x1bWVncm91cAogICAgICAgICAgICBzZWxmLnBp
            bl9zdWJ2b2x1bWUoCiAgICAgICAgICAgICAgICAiY3NpIiwgc2VsZi5fYXJnX3BhcnNlci5jZXBo
            ZnNfZmlsZXN5c3RlbV9uYW1lLCAiZGlzdHJpYnV0ZWQiLCAiMSIKICAgICAgICAgICAgKQogICAg
            ICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlNVQlZPTFVNRV9HUk9VUCJdOgogICAgICAgICAgICAg
            ICAgc2VsZi5nZXRfb3JfY3JlYXRlX3N1YnZvbHVtZV9ncm91cCgKICAgICAgICAgICAgICAgICAg
            ICBzZWxmLl9hcmdfcGFyc2VyLnN1YnZvbHVtZV9ncm91cCwKICAgICAgICAgICAgICAgICAgICBz
            ZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgICAgICBzZWxmLnBpbl9zdWJ2b2x1bWUoCiAgICAgICAgICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci5zdWJ2b2x1bWVfZ3JvdXAsCiAgICAgICAgICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lLAogICAgICAgICAgICAgICAgICAg
            ICJkaXN0cmlidXRlZCIsCiAgICAgICAgICAgICAgICAgICAgIjEiLAogICAgICAgICAgICAgICAg
            KQogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1RMU19DRVJUIl0gPSAiIgogICAgICAgIHNlbGYu
            b3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdID0gIiIKICAgICAgICBzZWxmLm91dF9tYXBb
            Ik1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJdID0gIiIKICAgICAgICBpZiBub3Qgc2VsZi5fYXJn
            X3BhcnNlci5za2lwX21vbml0b3JpbmdfZW5kcG9pbnQ6CiAgICAgICAgICAgICgKICAgICAgICAg
            ICAgICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdLAogICAgICAgICAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UX1BPUlQiXSwKICAgICAgICAgICAg
            KSA9IHNlbGYuZ2V0X2FjdGl2ZV9hbmRfc3RhbmRieV9tZ3JzKCkKICAgICAgICBzZWxmLm91dF9t
            YXBbIlJCRF9QT09MX05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1l
            CiAgICAgICAgc2VsZi5vdXRfbWFwWyJSQkRfTUVUQURBVEFfRUNfUE9PTF9OQU1FIl0gPSAoCiAg
            ICAgICAgICAgIHNlbGYudmFsaWRhdGVfcmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZSgpCiAgICAg
            ICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9PTFMiXSA9IHNlbGYuX2FyZ19w
            YXJzZXIudG9wb2xvZ3lfcG9vbHMKICAgICAgICBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxV
            UkVfRE9NQUlOX0xBQkVMIl0gPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xv
            Z3lfZmFpbHVyZV9kb21haW5fbGFiZWwKICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJU
            T1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiXSA9ICgKICAgICAgICAgICAgc2VsZi5fYXJn
            X3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMKICAgICAgICApCiAgICAgICAg
            aWYgKAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X3Bvb2xzICE9ICIiCiAg
            ICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX2xh
            YmVsICE9ICIiCiAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1
            cmVfZG9tYWluX3ZhbHVlcyAhPSAiIgogICAgICAgICk6CiAgICAgICAgICAgIHNlbGYudmFsaWRh
            dGVfdG9wb2xvZ3lfdmFsdWVzKAogICAgICAgICAgICAgICAgc2VsZi5jb252ZXJ0X2NvbW1hX3Nl
            cGFyYXRlZF90b19hcnJheShzZWxmLm91dF9tYXBbIlRPUE9MT0dZX1BPT0xTIl0pLAogICAgICAg
            ICAgICAgICAgc2VsZi5jb252ZXJ0X2NvbW1hX3NlcGFyYXRlZF90b19hcnJheSgKICAgICAgICAg
            ICAgICAgICAgICBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX1ZBTFVFUyJd
            CiAgICAgICAgICAgICAgICApLAogICAgICAgICAgICApCiAgICAgICAgICAgIHNlbGYudmFsaWRh
            dGVfdG9wb2xvZ3lfcmJkX3Bvb2xzKAogICAgICAgICAgICAgICAgc2VsZi5jb252ZXJ0X2NvbW1h
            X3NlcGFyYXRlZF90b19hcnJheShzZWxmLm91dF9tYXBbIlRPUE9MT0dZX1BPT0xTIl0pCiAgICAg
            ICAgICAgICkKICAgICAgICAgICAgc2VsZi5pbml0X3RvcG9sb2d5X3JiZF9wb29scygKICAgICAg
            ICAgICAgICAgIHNlbGYuY29udmVydF9jb21tYV9zZXBhcmF0ZWRfdG9fYXJyYXkoc2VsZi5vdXRf
            bWFwWyJUT1BPTE9HWV9QT09MUyJdKQogICAgICAgICAgICApCiAgICAgICAgZWxzZToKICAgICAg
            ICAgICAgc2VsZi5yYWlzZV9leGNlcHRpb25faWZfYW55X3RvcG9sb2d5X2ZsYWdfaXNfbWlzc2lu
            ZygpCgogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1BPT0xfUFJFRklYIl0gPSBzZWxmLl9hcmdf
            cGFyc2VyLnJnd19wb29sX3ByZWZpeAogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX0VORFBPSU5U
            Il0gPSAiIgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50OgogICAgICAg
            ICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgICAgICBzZWxmLmNy
            ZWF0ZV9yZ3dfYWRtaW5fb3BzX3VzZXIoKQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAg
            ICAgaWYgKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25h
            bWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9u
            ZWdyb3VwX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfem9uZV9uYW1lICE9ICIiCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAg
            ICAgIGVyciA9IHNlbGYudmFsaWRhdGVfcmd3X211bHRpc2l0ZSgKICAgICAgICAgICAgICAgICAg
            ICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwgInJlYWxtIgogICAgICAgICAg
            ICAgICAgICAgICkKICAgICAgICAgICAgICAgICAgICBlcnIgPSBzZWxmLnZhbGlkYXRlX3Jnd19t
            dWx0aXNpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pv
            bmVncm91cF9uYW1lLCAiem9uZWdyb3VwIgogICAgICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgICAgICAgICBlcnIgPSBzZWxmLnZhbGlkYXRlX3Jnd19tdWx0aXNpdGUoCiAgICAgICAgICAg
            ICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVfbmFtZSwgInpvbmUiCiAgICAg
            ICAgICAgICAgICAgICAgKQoKICAgICAgICAgICAgICAgIGlmICgKICAgICAgICAgICAgICAgICAg
            ICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lID09ICIiCiAgICAgICAgICAgICAgICAg
            ICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lID09ICIiCiAgICAgICAg
            ICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVfbmFtZSA9PSAiIgogICAg
            ICAgICAgICAgICAgKSBvciAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5y
            Z3dfcmVhbG1fbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFy
            c2VyLnJnd196b25lZ3JvdXBfbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd196b25lX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICk6CiAgICAg
            ICAgICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJH
            V19BRE1JTl9PUFNfVVNFUl9BQ0NFU1NfS0VZIl0sCiAgICAgICAgICAgICAgICAgICAgICAgIHNl
            bGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX1NFQ1JFVF9LRVkiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkLAogICAgICAgICAgICAgICAgICAgICAgICBl
            cnIsCiAgICAgICAgICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRlX3Jnd19hZG1pbl9vcHNfdXNl
            cigpCiAgICAgICAgICAgICAgICAgICAgZXJyID0gc2VsZi52YWxpZGF0ZV9yZ3dfZW5kcG9pbnQo
            aW5mb19jYXBfc3VwcG9ydGVkKQogICAgICAgICAgICAgICAgICAgIGlmIHNlbGYuX2FyZ19wYXJz
            ZXIucmd3X3Rsc19jZXJ0X3BhdGg6CiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYub3V0X21h
            cFsiUkdXX1RMU19DRVJUIl0gPSAoCiAgICAgICAgICAgICAgICAgICAgICAgICAgICBzZWxmLnZh
            bGlkYXRlX3Jnd19lbmRwb2ludF90bHNfY2VydCgpCiAgICAgICAgICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgICAgICAgICAjIGlmIHRoZXJlIGlzIG5vIGVycm9yLCBzZXQgdGhlIFJHV19F
            TkRQT0lOVAogICAgICAgICAgICAgICAgICAgIGlmIGVyciAhPSAiLTEiOgogICAgICAgICAgICAg
            ICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdID0gc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfZW5kcG9pbnQKICAgICAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICAgICAg
            ZXJyID0gIlBsZWFzZSBwcm92aWRlIGFsbCB0aGUgUkdXIG11bHRpc2l0ZSBwYXJhbWV0ZXJzIG9y
            IG5vbmUgb2YgdGhlbSIKICAgICAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKGVycikK
            CiAgICBkZWYgZ2VuX3NoZWxsX291dChzZWxmKToKICAgICAgICBzZWxmLl9nZW5fb3V0cHV0X21h
            cCgpCiAgICAgICAgc2hPdXRJTyA9IFN0cmluZ0lPKCkKICAgICAgICBmb3IgaywgdiBpbiBzZWxm
            Lm91dF9tYXAuaXRlbXMoKToKICAgICAgICAgICAgaWYgdiBhbmQgayBub3QgaW4gc2VsZi5fZXhj
            bHVkZWRfa2V5czoKICAgICAgICAgICAgICAgIHNoT3V0SU8ud3JpdGUoZiJleHBvcnQge2t9PXt2
            fXtMSU5FU0VQfSIpCiAgICAgICAgc2hPdXQgPSBzaE91dElPLmdldHZhbHVlKCkKICAgICAgICBz
            aE91dElPLmNsb3NlKCkKICAgICAgICByZXR1cm4gc2hPdXQKCiAgICBkZWYgZ2VuX2pzb25fb3V0
            KHNlbGYpOgogICAgICAgIHNlbGYuX2dlbl9vdXRwdXRfbWFwKCkKICAgICAgICBpZiBzZWxmLl9h
            cmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiAiIgogICAgICAgIGpzb25fb3V0
            ID0gWwogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFtZSI6ICJyb29rLWNlcGgtbW9u
            LWVuZHBvaW50cyIsCiAgICAgICAgICAgICAgICAia2luZCI6ICJDb25maWdNYXAiLAogICAgICAg
            ICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiBzZWxmLm91dF9t
            YXBbIlJPT0tfRVhURVJOQUxfQ0VQSF9NT05fREFUQSJdLAogICAgICAgICAgICAgICAgICAgICJt
            YXhNb25JZCI6ICIwIiwKICAgICAgICAgICAgICAgICAgICAibWFwcGluZyI6ICJ7fSIsCiAgICAg
            ICAgICAgICAgICB9LAogICAgICAgICAgICB9LAogICAgICAgICAgICB7CiAgICAgICAgICAgICAg
            ICAibmFtZSI6ICJyb29rLWNlcGgtbW9uIiwKICAgICAgICAgICAgICAgICJraW5kIjogIlNlY3Jl
            dCIsCiAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAiYWRtaW4t
            c2VjcmV0IjogImFkbWluLXNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImZzaWQiOiBzZWxm
            Lm91dF9tYXBbIlJPT0tfRVhURVJOQUxfRlNJRCJdLAogICAgICAgICAgICAgICAgICAgICJtb24t
            c2VjcmV0IjogIm1vbi1zZWNyZXQiLAogICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgfSwK
            ICAgICAgICAgICAgewogICAgICAgICAgICAgICAgIm5hbWUiOiAicm9vay1jZXBoLW9wZXJhdG9y
            LWNyZWRzIiwKICAgICAgICAgICAgICAgICJraW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAg
            ICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJS
            T09LX0VYVEVSTkFMX1VTRVJOQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgInVzZXJLZXkiOiBz
            ZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAg
            IH0sCiAgICAgICAgICAgIH0sCiAgICAgICAgXQoKICAgICAgICAjIGlmICdNT05JVE9SSU5HX0VO
            RFBPSU5UJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ21vbml0b3JpbmctZW5kcG9pbnQnIHRvIENs
            dXN0ZXIKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19F
            TkRQT0lOVCJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9J
            TlRfUE9SVCJdCiAgICAgICAgKToKICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAg
            ICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjogIm1vbml0b3JpbmctZW5kcG9p
            bnQiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIkNlcGhDbHVzdGVyIiwKICAgICAgICAg
            ICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgIk1vbml0b3JpbmdF
            bmRwb2ludCI6IHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdLAogICAgICAgICAg
            ICAgICAgICAgICAgICAiTW9uaXRvcmluZ1BvcnQiOiBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdf
            RU5EUE9JTlRfUE9SVCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9
            CiAgICAgICAgICAgICkKCiAgICAgICAgIyBpZiAnQ1NJX1JCRF9OT0RFX1NFQ1JFVCcgZXhpc3Rz
            LCB0aGVuIG9ubHkgYWRkICdyb29rLWNzaS1yYmQtcHJvdmlzaW9uZXInIFNlY3JldAogICAgICAg
            IGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VDUkVUIl0KICAg
            ICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FIl0KICAg
            ICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAg
            ICAgICAgICAgICAgICAgICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9O
            T0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQi
            LAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAi
            dXNlcklEIjogc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VDUkVUX05BTUUiXSwKICAgICAg
            ICAgICAgICAgICAgICAgICAgInVzZXJLZXkiOiBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9T
            RUNSRVQiXSwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAg
            ICAgICApCiAgICAgICAgIyBpZiAnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQnIGV4aXN0cywg
            dGhlbiBvbmx5IGFkZCAncm9vay1jc2ktcmJkLXByb3Zpc2lvbmVyJyBTZWNyZXQKICAgICAgICBp
            ZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQi
            XQogICAgICAgICAgICBhbmQgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JF
            VF9OQU1FIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAg
            ICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21h
            cFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAg
            ICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAg
            ICAgICAgICAgICAgICAgICAgInVzZXJJRCI6IHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJ
            T05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNl
            bGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAg
            ICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBpZiAnQ1NJ
            X0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAncm9vay1j
            c2ktY2VwaGZzLXByb3Zpc2lvbmVyJyBTZWNyZXQKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNl
            bGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVQiXQogICAgICAgICAgICBh
            bmQgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0KICAg
            ICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAg
            ICAgICAgICAgICAgICAgICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhG
            U19QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQi
            OiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAg
            ICAgICAgICAgImFkbWluSUQiOiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJf
            U0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgImFkbWluS2V5Ijogc2VsZi5v
            dXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAg
            ICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdDU0lf
            Q0VQSEZTX05PREVfU0VDUkVUJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ3Jvb2stY3NpLWNlcGhm
            cy1ub2RlJyBTZWNyZXQKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJ
            X0NFUEhGU19OT0RFX1NFQ1JFVCJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9D
            RVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJdCiAgICAgICAgKToKICAgICAgICAgICAganNvbl9vdXQu
            YXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjogZiJy
            b29rLXtzZWxmLm91dF9tYXBbJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAg
            ICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0
            YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImFkbWluSUQiOiBzZWxmLm91dF9tYXBbIkNT
            SV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAiYWRt
            aW5LZXkiOiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVQiXSwKICAgICAgICAg
            ICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBp
            ZiAnUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSycgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdy
            b29rLWNlcGgtZGFzaGJvYXJkLWxpbmsnIFNlY3JldAogICAgICAgIGlmIHNlbGYub3V0X21hcFsi
            Uk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSyJdOgogICAgICAgICAgICBqc29uX291dC5hcHBl
            bmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAicm9vay1j
            ZXBoLWRhc2hib2FyZC1saW5rIiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQi
            LAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAi
            dXNlcklEIjogImNlcGgtZGFzaGJvYXJkLWxpbmsiLAogICAgICAgICAgICAgICAgICAgICAgICAi
            dXNlcktleSI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSyJdLAog
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAg
            ICAgICAjIGlmICdSQURPU19OQU1FU1BBQ0UnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCB0aGUgIlJB
            RE9TX05BTUVTUEFDRSIgbmFtZXNwYWNlCiAgICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLm91
            dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIlJF
            U1RSSUNURURfQVVUSF9QRVJNSVNTSU9OIl0KICAgICAgICAgICAgYW5kIG5vdCBzZWxmLm91dF9t
            YXBbIlJCRF9NRVRBREFUQV9FQ19QT09MX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpz
            b25fb3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFt
            ZSI6ICJyYWRvcy1uYW1lc3BhY2UiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIkNlcGhC
            bG9ja1Bvb2xSYWRvc05hbWVzcGFjZSIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAg
            ICAgICAgICAgICAgICAgICAgICAgICJyYWRvc05hbWVzcGFjZU5hbWUiOiBzZWxmLm91dF9tYXBb
            IlJBRE9TX05BTUVTUEFDRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYu
            b3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAg
            ICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAg
            ICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjogImNlcGgtcmJkLXJhZG9z
            LW5hbWVzcGFjZSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwK
            ICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgInBv
            b2wiOiBzZWxmLm91dF9tYXBbIlJCRF9QT09MX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAg
            ICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNyZXQtbmFtZSI6IGYicm9vay17
            c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAg
            ICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250cm9sbGVyLWV4cGFuZC1z
            ZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NF
            Q1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5p
            by9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRf
            Tk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAg
            ICAgIH0KICAgICAgICAgICAgKQogICAgICAgIGVsc2U6CiAgICAgICAgICAgIGlmIHNlbGYub3V0
            X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdOgogICAgICAgICAgICAgICAganNvbl9v
            dXQuYXBwZW5kKAogICAgICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAgICAg
            Im5hbWUiOiAiY2VwaC1yYmQiLAogICAgICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9y
            YWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAg
            ICAgICAgICAgICAgICAgICJkYXRhUG9vbCI6IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJd
            LAogICAgICAgICAgICAgICAgICAgICAgICAgICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIlJCRF9N
            RVRBREFUQV9FQ19QT09MX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJjc2ku
            c3RvcmFnZS5rOHMuaW8vcHJvdmlzaW9uZXItc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0
            X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNlY3Jl
            dC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVU
            X05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5p
            by9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRf
            Tk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAg
            ICAgICAgICAgICAgfQogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBlbHNlOgogICAgICAg
            ICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgICAgIHsKICAgICAgICAg
            ICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaC1yYmQiLAogICAgICAgICAgICAgICAgICAgICAg
            ICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAgICAgICAiZGF0YSI6
            IHsKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJwb29sIjogc2VsZi5vdXRfbWFwWyJSQkRf
            UE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2Uuazhz
            LmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9S
            QkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAg
            ICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYi
            cm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwK
            ICAgICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vbm9kZS1zdGFn
            ZS1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX05PREVfU0VDUkVU
            X05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgICAg
            IH0KICAgICAgICAgICAgICAgICkKCiAgICAgICAgIyBpZiAnVE9QT0xPR1lfUE9PTFMnLCAnVE9Q
            T0xPR1lfRkFJTFVSRV9ET01BSU5fTEFCRUwnLCAnVE9QT0xPR1lfRkFJTFVSRV9ET01BSU5fVkFM
            VUVTJyAgZXhpc3RzLAogICAgICAgICMgdGhlbiBvbmx5IGFkZCAndG9wb2xvZ3knIFN0b3JhZ2VD
            bGFzcwogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09M
            UyJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlO
            X0xBQkVMIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfRkFJTFVSRV9E
            T01BSU5fVkFMVUVTIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAg
            ICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaC1yYmQtdG9w
            b2xvZ3kiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAg
            ICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJ0b3BvbG9n
            eUZhaWx1cmVEb21haW5MYWJlbCI6IHNlbGYub3V0X21hcFsKICAgICAgICAgICAgICAgICAgICAg
            ICAgICAgICJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9MQUJFTCIKICAgICAgICAgICAgICAgICAg
            ICAgICAgXSwKICAgICAgICAgICAgICAgICAgICAgICAgInRvcG9sb2d5RmFpbHVyZURvbWFpblZh
            bHVlcyI6IHNlbGYub3V0X21hcFsKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJUT1BPTE9H
            WV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiCiAgICAgICAgICAgICAgICAgICAgICAgIF0sCiAgICAg
            ICAgICAgICAgICAgICAgICAgICJ0b3BvbG9neVBvb2xzIjogc2VsZi5vdXRfbWFwWyJUT1BPTE9H
            WV9QT09MUyJdLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3By
            b3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJP
            VklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0
            b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxm
            Lm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUi
            OiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAg
            ICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCgogICAg
            ICAgICMgaWYgJ0NFUEhGU19GU19OQU1FJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ2NlcGhmcycg
            U3RvcmFnZUNsYXNzCiAgICAgICAgaWYgc2VsZi5vdXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdOgog
            ICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAg
            ICAgICAgICAgIm5hbWUiOiAiY2VwaGZzIiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJT
            dG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAg
            ICAgICAgICAgICAiZnNOYW1lIjogc2VsZi5vdXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdLAogICAg
            ICAgICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYub3V0X21hcFsiQ0VQSEZTX1BPT0xfTkFN
            RSJdLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lv
            bmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9DRVBIRlNfUFJPVklT
            SU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3Jh
            Z2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91
            dF9tYXBbJ0NTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUi
            OiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FJ119IiwK
            ICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAg
            ICAgICAgIyBpZiAnUkdXX0VORFBPSU5UJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ2NlcGgtcmd3
            JyBTdG9yYWdlQ2xhc3MKICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdOgog
            ICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAg
            ICAgICAgICAgIm5hbWUiOiAiY2VwaC1yZ3ciLAogICAgICAgICAgICAgICAgICAgICJraW5kIjog
            IlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAg
            ICAgICAgICAgICAgICJlbmRwb2ludCI6IHNlbGYub3V0X21hcFsiUkdXX0VORFBPSU5UIl0sCiAg
            ICAgICAgICAgICAgICAgICAgICAgICJwb29sUHJlZml4Ijogc2VsZi5vdXRfbWFwWyJSR1dfUE9P
            TF9QUkVGSVgiXSwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAg
            ICAgICAgICApCiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsK
            ICAgICAgICAgICAgICAgICAgICAibmFtZSI6ICJyZ3ctYWRtaW4tb3BzLXVzZXIiLAogICAgICAg
            ICAgICAgICAgICAgICJraW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEi
            OiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJhY2Nlc3NLZXkiOiBzZWxmLm91dF9tYXBbIlJH
            V19BRE1JTl9PUFNfVVNFUl9BQ0NFU1NfS0VZIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJz
            ZWNyZXRLZXkiOiBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9TRUNSRVRfS0VZIl0s
            CiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQog
            ICAgICAgICMgaWYgJ1JHV19UTFNfQ0VSVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkIHRoZSAiY2Vw
            aC1yZ3ctdGxzLWNlcnQiIHNlY3JldAogICAgICAgIGlmIHNlbGYub3V0X21hcFsiUkdXX1RMU19D
            RVJUIl06CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAg
            ICAgICAgICAgICAgICAgICAibmFtZSI6ICJjZXBoLXJndy10bHMtY2VydCIsCiAgICAgICAgICAg
            ICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsK
            ICAgICAgICAgICAgICAgICAgICAgICAgImNlcnQiOiBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VS
            VCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAg
            ICkKCiAgICAgICAgcmV0dXJuIGpzb24uZHVtcHMoanNvbl9vdXQpICsgTElORVNFUAoKICAgIGRl
            ZiB1cGdyYWRlX3VzZXJzX3Blcm1pc3Npb25zKHNlbGYpOgogICAgICAgIHVzZXJzID0gWwogICAg
            ICAgICAgICAiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIsCiAgICAgICAgICAgICJjbGllbnQuY3Np
            LWNlcGhmcy1wcm92aXNpb25lciIsCiAgICAgICAgICAgICJjbGllbnQuY3NpLXJiZC1ub2RlIiwK
            ICAgICAgICAgICAgImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwKICAgICAgICAgICAgImNs
            aWVudC5oZWFsdGhjaGVja2VyIiwKICAgICAgICBdCiAgICAgICAgaWYgc2VsZi5ydW5fYXNfdXNl
            ciAhPSAiIiBhbmQgc2VsZi5ydW5fYXNfdXNlciBub3QgaW4gdXNlcnM6CiAgICAgICAgICAgIHVz
            ZXJzLmFwcGVuZChzZWxmLnJ1bl9hc191c2VyKQogICAgICAgIGZvciB1c2VyIGluIHVzZXJzOgog
            ICAgICAgICAgICBzZWxmLnVwZ3JhZGVfdXNlcl9wZXJtaXNzaW9ucyh1c2VyKQoKICAgIGRlZiBn
            ZXRfcmd3X3Bvb2xfbmFtZV9kdXJpbmdfdXBncmFkZShzZWxmLCB1c2VyLCBjYXBzKToKICAgICAg
            ICBpZiB1c2VyID09ICJjbGllbnQuaGVhbHRoY2hlY2tlciI6CiAgICAgICAgICAgICMgd2hlbiBh
            ZG1pbiBoYXMgbm90IHByb3ZpZGVkIHJndyBwb29sIG5hbWUgZHVyaW5nIHVwZ3JhZGUsCiAgICAg
            ICAgICAgICMgZ2V0IHRoZSByZ3cgcG9vbCBuYW1lIGZyb20gY2xpZW50LmhlYWx0aGNoZWNrZXIg
            dXNlciB3aGljaCB3YXMgdXNlZCBkdXJpbmcgY29ubmVjdGlvbgogICAgICAgICAgICBpZiBub3Qg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXg6CiAgICAgICAgICAgICAgICAjIFRvIGdl
            dCB2YWx1ZSAnZGVmYXVsdCcgd2hpY2ggaXMgcmd3IHBvb2wgbmFtZSBmcm9tICdhbGxvdyByd3gg
            cG9vbD1kZWZhdWx0LnJndy5tZXRhJwogICAgICAgICAgICAgICAgcGF0dGVybiA9IHIicG9vbD0o
            Lio/KVwucmd3XC5tZXRhIgogICAgICAgICAgICAgICAgbWF0Y2ggPSByZS5zZWFyY2gocGF0dGVy
            biwgY2FwcykKICAgICAgICAgICAgICAgIGlmIG1hdGNoOgogICAgICAgICAgICAgICAgICAgIHNl
            bGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4ID0gbWF0Y2guZ3JvdXAoMSkKICAgICAgICAg
            ICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAgICAgImZhaWxlZCB0byBnZXQgcmd3IHBvb2wg
            bmFtZSBmb3IgdXBncmFkZSIKICAgICAgICAgICAgICAgICAgICApCgogICAgZGVmIHVwZ3JhZGVf
            dXNlcl9wZXJtaXNzaW9ucyhzZWxmLCB1c2VyKToKICAgICAgICAjIGNoZWNrIHdoZXRoZXIgdGhl
            IGdpdmVuIHVzZXIgZXhpc3RzIG9yIG5vdAogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAi
            YXV0aCBnZXQiLCAiZW50aXR5IjogZiJ7dXNlcn0iLCAiZm9ybWF0IjogImpzb24ifQogICAgICAg
            IHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihj
            bWRfanNvbikKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgog
            ICAgICAgICAgICBwcmludChmInVzZXIge3VzZXJ9IG5vdCBmb3VuZCBmb3IgdXBncmFkaW5nLiIp
            CiAgICAgICAgICAgIHJldHVybgogICAgICAgIGV4aXN0aW5nX2NhcHMgPSBqc29uX291dFswXVsi
            Y2FwcyJdCiAgICAgICAgc2VsZi5nZXRfcmd3X3Bvb2xfbmFtZV9kdXJpbmdfdXBncmFkZSh1c2Vy
            LCBzdHIoZXhpc3RpbmdfY2FwcykpCiAgICAgICAgbmV3X2NhcCwgXyA9IHNlbGYuZ2V0X2NhcHNf
            YW5kX2VudGl0eSh1c2VyKQogICAgICAgIGNhcF9rZXlzID0gWyJtb24iLCAibWdyIiwgIm9zZCIs
            ICJtZHMiXQogICAgICAgIGNhcHMgPSBbXQogICAgICAgIGZvciBlYWNoQ2FwIGluIGNhcF9rZXlz
            OgogICAgICAgICAgICBjdXJfY2FwX3ZhbHVlcyA9IGV4aXN0aW5nX2NhcHMuZ2V0KGVhY2hDYXAs
            ICIiKQogICAgICAgICAgICBuZXdfY2FwX3ZhbHVlcyA9IG5ld19jYXAuZ2V0KGVhY2hDYXAsICIi
            KQogICAgICAgICAgICBjdXJfY2FwX3Blcm1fbGlzdCA9IFsKICAgICAgICAgICAgICAgIHguc3Ry
            aXAoKSBmb3IgeCBpbiBjdXJfY2FwX3ZhbHVlcy5zcGxpdCgiLCIpIGlmIHguc3RyaXAoKQogICAg
            ICAgICAgICBdCiAgICAgICAgICAgIG5ld19jYXBfcGVybV9saXN0ID0gWwogICAgICAgICAgICAg
            ICAgeC5zdHJpcCgpIGZvciB4IGluIG5ld19jYXBfdmFsdWVzLnNwbGl0KCIsIikgaWYgeC5zdHJp
            cCgpCiAgICAgICAgICAgIF0KICAgICAgICAgICAgIyBhcHBlbmQgbmV3X2NhcF9saXN0IHRvIGN1
            cl9jYXBfbGlzdCB0byBtYWludGFpbiB0aGUgb3JkZXIgb2YgY2FwcwogICAgICAgICAgICBjdXJf
            Y2FwX3Blcm1fbGlzdC5leHRlbmQobmV3X2NhcF9wZXJtX2xpc3QpCiAgICAgICAgICAgICMgZWxp
            bWluYXRlIGR1cGxpY2F0ZXMgd2l0aG91dCB1c2luZyAnc2V0JwogICAgICAgICAgICAjIHNldCBy
            ZS1vcmRlcnMgaXRlbXMgaW4gdGhlIGxpc3QgYW5kIHdlIGhhdmUgdG8ga2VlcCB0aGUgb3JkZXIK
            ICAgICAgICAgICAgbmV3X2NhcF9saXN0ID0gW10KICAgICAgICAgICAgW25ld19jYXBfbGlzdC5h
            cHBlbmQoeCkgZm9yIHggaW4gY3VyX2NhcF9wZXJtX2xpc3QgaWYgeCBub3QgaW4gbmV3X2NhcF9s
            aXN0XQogICAgICAgICAgICBleGlzdGluZ19jYXBzW2VhY2hDYXBdID0gIiwgIi5qb2luKG5ld19j
            YXBfbGlzdCkKICAgICAgICAgICAgaWYgZXhpc3RpbmdfY2Fwc1tlYWNoQ2FwXToKICAgICAgICAg
            ICAgICAgIGNhcHMuYXBwZW5kKGVhY2hDYXApCiAgICAgICAgICAgICAgICBjYXBzLmFwcGVuZChl
            eGlzdGluZ19jYXBzW2VhY2hDYXBdKQogICAgICAgIGNtZF9qc29uID0gewogICAgICAgICAgICAi
            cHJlZml4IjogImF1dGggY2FwcyIsCiAgICAgICAgICAgICJlbnRpdHkiOiB1c2VyLAogICAgICAg
            ICAgICAiY2FwcyI6IGNhcHMsCiAgICAgICAgICAgICJmb3JtYXQiOiAianNvbiIsCiAgICAgICAg
            fQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9q
            c29uX2dlbihjbWRfanNvbikKICAgICAgICBpZiByZXRfdmFsICE9IDA6CiAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIidhdXRoIGNh
            cHMge3VzZXJ9JyBjb21tYW5kIGZhaWxlZC5cbiBFcnJvcjoge2Vycl9tc2d9IgogICAgICAgICAg
            ICApCiAgICAgICAgcHJpbnQoZiJVcGRhdGVkIHVzZXIge3VzZXJ9IHN1Y2Nlc3NmdWxseS4iKQoK
            ICAgIGRlZiBtYWluKHNlbGYpOgogICAgICAgIGdlbmVyYXRlZF9vdXRwdXQgPSAiIgogICAgICAg
            IGlmIHNlbGYuX2FyZ19wYXJzZXIudXBncmFkZToKICAgICAgICAgICAgc2VsZi51cGdyYWRlX3Vz
            ZXJzX3Blcm1pc3Npb25zKCkKICAgICAgICBlbGlmIHNlbGYuX2FyZ19wYXJzZXIuZm9ybWF0ID09
            ICJqc29uIjoKICAgICAgICAgICAgZ2VuZXJhdGVkX291dHB1dCA9IHNlbGYuZ2VuX2pzb25fb3V0
            KCkKICAgICAgICBlbGlmIHNlbGYuX2FyZ19wYXJzZXIuZm9ybWF0ID09ICJiYXNoIjoKICAgICAg
            ICAgICAgZ2VuZXJhdGVkX291dHB1dCA9IHNlbGYuZ2VuX3NoZWxsX291dCgpCiAgICAgICAgZWxz
            ZToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAg
            ICAgICAgIGYiVW5zdXBwb3J0ZWQgZm9ybWF0OiB7c2VsZi5fYXJnX3BhcnNlci5mb3JtYXR9Igog
            ICAgICAgICAgICApCiAgICAgICAgcHJpbnQoZ2VuZXJhdGVkX291dHB1dCkKICAgICAgICBpZiBz
            ZWxmLm91dHB1dF9maWxlIGFuZCBnZW5lcmF0ZWRfb3V0cHV0OgogICAgICAgICAgICBmT3V0ID0g
            b3BlbihzZWxmLm91dHB1dF9maWxlLCBtb2RlPSJ3IiwgZW5jb2Rpbmc9IlVURi04IikKICAgICAg
            ICAgICAgZk91dC53cml0ZShnZW5lcmF0ZWRfb3V0cHV0KQogICAgICAgICAgICBmT3V0LmNsb3Nl
            KCkKCgojIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMKIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIE1BSU4gIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIwppZiBfX25hbWVfXyA9PSAiX19t
            YWluX18iOgogICAgcmpPYmogPSBSYWRvc0pTT04oKQogICAgdHJ5OgogICAgICAgIHJqT2JqLm1h
            aW4oKQogICAgZXhjZXB0IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24gYXMgZXJyOgogICAgICAg
            IHByaW50KGYiRXhlY3V0aW9uIEZhaWxlZDoge2Vycn0iKQogICAgICAgIHJhaXNlIGVycgogICAg
            ZXhjZXB0IEtleUVycm9yIGFzIGtFcnI6CiAgICAgICAgcHJpbnQoZiJLZXlFcnJvcjoge2tFcnJ9
            IikKICAgIGV4Y2VwdCBPU0Vycm9yIGFzIG9zRXJyOgogICAgICAgIHByaW50KGYiRXJyb3Igd2hp
            bGUgdHJ5aW5nIHRvIG91dHB1dCB0aGUgZGF0YToge29zRXJyfSIpCiAgICBmaW5hbGx5OgogICAg
            ICAgIHJqT2JqLnNodXRkb3duKCkK
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/red-hat-storage/rook
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Ceph cluster.
            displayName: Ceph Cluster
            kind: CephCluster
            name: cephclusters.ceph.rook.io
            version: v1
          - description: Represents a Ceph Block Pool.
            displayName: Ceph Block Pool
            kind: CephBlockPool
            name: cephblockpools.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store.
            displayName: Ceph Object Store
            kind: CephObjectStore
            name: cephobjectstores.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store User.
            displayName: Ceph Object Store User
            kind: CephObjectStoreUser
            name: cephobjectstoreusers.ceph.rook.io
            version: v1
          - description: Represents a cluster of Ceph NFS ganesha gateways.
            displayName: Ceph NFS
            kind: CephNFS
            name: cephnfses.ceph.rook.io
            version: v1
          - description: Represents a Ceph User.
            displayName: Ceph Client
            kind: CephClient
            name: cephclients.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem.
            displayName: Ceph Filesystem
            kind: CephFilesystem
            name: cephfilesystems.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem Mirror.
            displayName: Ceph Filesystem Mirror
            kind: CephFilesystemMirror
            name: cephfilesystemmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph RBD Mirror.
            displayName: Ceph RBD Mirror
            kind: CephRBDMirror
            name: cephrbdmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Realm.
            displayName: Ceph Object Store Realm
            kind: CephObjectRealm
            name: cephobjectrealms.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone Group.
            displayName: Ceph Object Store Zone Group
            kind: CephObjectZoneGroup
            name: cephobjectzonegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone.
            displayName: Ceph Object Store Zone
            kind: CephObjectZone
            name: cephobjectzones.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Notification.
            displayName: Ceph Bucket Notification
            kind: CephBucketNotification
            name: cephbucketnotifications.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Topic.
            displayName: Ceph Bucket Topic
            kind: CephBucketTopic
            name: cephbuckettopics.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem SubVolumeGroup.
            displayName: Ceph Filesystem SubVolumeGroup
            kind: CephFilesystemSubVolumeGroup
            name: cephfilesystemsubvolumegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph BlockPool Rados Namespace.
            displayName: Ceph BlockPool Rados Namespace
            kind: CephBlockPoolRadosNamespace
            name: cephblockpoolradosnamespaces.ceph.rook.io
            version: v1
          - description: Represents a Ceph COSI Driver.
            displayName: Ceph COSI Driver
            kind: CephCOSIDriver
            name: cephcosidrivers.ceph.rook.io
            version: v1
        description: |2

          The Rook-Ceph storage operator packages, deploys, manages, upgrades and scales Ceph storage for providing persistent storage to infrastructure services (Logging, Metrics, Registry) as well as stateful applications in Kubernetes clusters.

          ## Rook-Ceph Storage Operator

          Rook runs as a cloud-native service in Kubernetes clusters for optimal integration with applications in need of storage, and handles the heavy-lifting behind the scenes such as provisioning and management.
          Rook orchestrates battle-tested open-source storage technology Ceph, which has years of production deployments and runs some of the worlds largest clusters.

          Ceph is a massively scalable, software-defined, cloud native storage platform that offers block, file and object storage services.
          Ceph can be used to back a wide variety of applications including relational databases, NoSQL databases, CI/CD tool-sets, messaging, AI/ML and analytics applications.
          Ceph is a proven storage platform that backs some of the world's largest storage deployments and has a large vibrant open source community backing the project.

          ## Supported features
          * **High Availability and resiliency** - Ceph has no single point of failures (SPOF) and all its components work natively in a highly available fashion
          * **Data Protection** - Ceph periodically scrub for inconsistent objects and repair them if necessary, making sure your replicas are always coherent
          * **Consistent storage platform across hybrid cloud** - Ceph can be deployed anywhere (on-premise or bare metal) and thus offers a similar experience regardless
          * **Block, File & Object storage service** - Ceph can expose your data through several storage interfaces, solving all the application use cases
          * **Scale up/down** - addition and removal of storage is fully covered by the operator.
          * **Dashboard** - The Operator deploys a dashboard for monitoring and introspecting your cluster.

          ## Before you start
          https://rook.io/docs/rook/v1.0/k8s-pre-reqs.html
        displayName: Rook-Ceph
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rook
        - ceph
        - storage
        - object storage
        - open source
        - block storage
        - shared filesystem
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/rook
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:cf7f72469d25fb37301a8159406613673f8968c82af1035fd45a470e0ae8acfa
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:80c01633a6bfca44df37ac38f764035be6998fce965c8709e4f6d54d17eeaa7b
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6fee56851da3a87b9f7f477fc59e3385e118dc0b398861a8f13459c998a5af77
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c11a599bfd09b9b22888e03eabdbe1054d4049718a591d2a45bf0ed89800111e
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:82a4b481028195cfd75471656fe40c1cbbade4c7b2d170d6bf9422d9d3b84fa7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:693a60d6a1162659213ae554b1c6ee200b0300a7da72c01415044cf507acab49
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:0afcb516eb528857887c6ee1a9cba081dc8f6b7d47d9f26a0dbb2dca0cefe07c
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:8d8326c74795c9c6bd8f33a9f9e81d2304a70b1f9be3c2f8d50e76be4b0292ff
        version: 4.16.16-rhodf
      entries:
      - name: rook-ceph-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: rook-ceph-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: rook-ceph-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: rook-ceph-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: rook-ceph-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: rook-ceph-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: rook-ceph-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: rook-ceph-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: rook-ceph-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: rook-ceph-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: rook-ceph-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: rook-ceph-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: rook-ceph-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: rook-ceph-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: rook-ceph-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: rook-ceph-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: rook-ceph-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: rook-ceph-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: logic-operator-rhel8
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: logic-operator-rhel8.v1.36.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "host/namespace",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          createdAt: "2025-05-26T14:10:15Z"
          description: OpenShift Serverless Logic Kubernetes Operator for deploying
            workflow applications based on the CNCF Serverless Workflow specification
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-serverless-logic
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kubesmarts/kie-tools
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: Sonata Flow Build
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: Sonata Flow Cluster Platform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: Sonata Flow Platform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: Sonata Flow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          OpenShift Serverless Logic Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification v0.8](https://serverlessworkflow.io/).
        displayName: OpenShift Serverless Logic Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        - workflows
        links:
        - name: Documentation
          url: https://docs.openshift.com/serverless/
        - name: Source Repository
          url: https://github.com/kiegroup/kogito-serverless-operator/
        maintainers:
        - email: serverless-logic-team@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/logic-rhel8-operator@sha256:8d3682448ebdac3aeabb2d23842b7e67a252b95f959c408af805037f9728fd3c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4564ca3dc5bac80d6faddaf94c817fbbc270698a9399d8a21ee1005d85ceda56
        version: 1.36.0
      entries:
      - name: logic-operator-rhel8.v1.36.0
        version: 1.36.0
      - name: logic-operator-rhel8.v1.35.0
        version: 1.35.0
      - name: logic-operator-rhel8.v1.34.0
        version: 1.34.0
      - name: logic-operator-rhel8.v1.33.0
        version: 1.33.0
      name: alpha
    defaultChannel: alpha
    packageName: logic-operator-rhel8
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudwatchlogs-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudwatchlogs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudwatchlogs.services.k8s.aws/v1alpha1",
                "kind": "LogGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudwatchlogs-controller:1.1.1
          createdAt: "2025-09-19T20:51:16Z"
          description: AWS CloudWatch Logs controller is a service controller for
            managing CloudWatch Logs resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LogGroup represents the state of an AWS cloudwatchlogs LogGroup
              resource.
            displayName: LogGroup
            kind: LogGroup
            name: loggroups.cloudwatchlogs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudWatch Logs resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudWatch Logs**

          You can use Amazon CloudWatch Logs to monitor, store, and access your log files from Amazon Elastic Compute Cloud (Amazon EC2) instances, AWS CloudTrail, Route 53, and other sources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudWatch Logs
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudwatchlogs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudWatch Logs Developer Resources
          url: https://aws.amazon.com/cloudwatch/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudwatch maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudwatchlogs-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-cloudwatchlogs-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudwatchlogs-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudwatchlogs-controller.v1.0.14
        version: 1.0.14
      - name: ack-cloudwatchlogs-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudwatchlogs-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudwatchlogs-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudwatchlogs-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudwatchlogs-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudwatchlogs-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudwatchlogs-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudwatchlogs-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudwatchlogs-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudwatchlogs-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudwatchlogs-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudwatchlogs-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudwatchlogs-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudwatchlogs-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudwatchlogs-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudwatchlogs-controller.v0.0.7
        version: 0.0.7
      - name: ack-cloudwatchlogs-controller.v0.0.4
        version: 0.0.4
      - name: ack-cloudwatchlogs-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudwatchlogs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: TrendMicro
      provider-url: trendmicro.com
    name: visionone-containersecurity
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: visionone-containersecurity.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "visionone-containersecurity-sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          com.redhat.openshift.versions: v4.12-v4.20
          containerImage: public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
          createdAt: "2025-09-18T22:43:41Z"
          description: Trend Vision One™ Container Security delivers full lifecycle
            protection for containers with real-time threat detection, policy enforcement,
            compliance assurance, and rapid incident response.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trendmicro-system
          operatorhub.io/ui-metadata-max-k8s-version: "1.26"
          operators.openshift.io/valid-subscription: VisionOneContainerSecurity Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trendmicro/visionone-container-security-helm
          support: TrendMicro Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trend Vision One™ Container Security delivers full lifecycle
              protection for containers with real-time threat detection, policy enforcement,
              compliance assurance, and rapid incident response.
            displayName: VisionOneContainerSecurity
            kind: VisionOneContainerSecurity
            name: visiononecontainersecurities.container-security.trendmicro.com
            version: v1alpha1
        description: "[Trend Vision One™ Container Security](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-container-security)
          delivers comprehensive, end-to-end\n protection for containerized workloads
          across OpenShift and Kubernetes environments.\n It continuously monitors
          for threats, vulnerabilities, policy violations, and\n exposed secrets ensuring
          compliance and enabling rapid, automated response. Purpose-built\n for hybrid
          cloud, it instantly isolates compromised containers to prevent lateral\n
          movement and minimize impact. Seamlessly integrating into OpenShift DevSecOps\n
          pipelines, it empowers security teams to enforce policies while enabling
          developers\n to innovate and deploy with confidence without disrupting workflows
          or compromising\n agility.\n\n **Installation**\n \n - Use the recommended
          `trendmicro-system` namespace.\n - Refer to the [Trend Vision One™ Documentation](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator)
          for registration and installation steps.\n"
        displayName: Trend Vision One™ Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trendmicro
        - visionone
        - container-security
        links:
        - name: Documentation
          url: https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator
        maintainers:
        - email: support@trendmicro.com
          name: TrendMicro VisionOne Support Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: TrendMicro
          url: trendmicro.com
        relatedImages:
        - public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
        version: 0.0.2
      entries:
      - name: visionone-containersecurity.v0.0.2
        version: 0.0.2
      - name: visionone-containersecurity.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: visionone-containersecurity.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "visionone-containersecurity-sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          com.redhat.openshift.versions: v4.12-v4.20
          containerImage: public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
          createdAt: "2025-09-18T22:43:41Z"
          description: Trend Vision One™ Container Security delivers full lifecycle
            protection for containers with real-time threat detection, policy enforcement,
            compliance assurance, and rapid incident response.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trendmicro-system
          operatorhub.io/ui-metadata-max-k8s-version: "1.26"
          operators.openshift.io/valid-subscription: VisionOneContainerSecurity Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trendmicro/visionone-container-security-helm
          support: TrendMicro Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trend Vision One™ Container Security delivers full lifecycle
              protection for containers with real-time threat detection, policy enforcement,
              compliance assurance, and rapid incident response.
            displayName: VisionOneContainerSecurity
            kind: VisionOneContainerSecurity
            name: visiononecontainersecurities.container-security.trendmicro.com
            version: v1alpha1
        description: "[Trend Vision One™ Container Security](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-container-security)
          delivers comprehensive, end-to-end\n protection for containerized workloads
          across OpenShift and Kubernetes environments.\n It continuously monitors
          for threats, vulnerabilities, policy violations, and\n exposed secrets ensuring
          compliance and enabling rapid, automated response. Purpose-built\n for hybrid
          cloud, it instantly isolates compromised containers to prevent lateral\n
          movement and minimize impact. Seamlessly integrating into OpenShift DevSecOps\n
          pipelines, it empowers security teams to enforce policies while enabling
          developers\n to innovate and deploy with confidence without disrupting workflows
          or compromising\n agility.\n\n **Installation**\n \n - Use the recommended
          `trendmicro-system` namespace.\n - Refer to the [Trend Vision One™ Documentation](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator)
          for registration and installation steps.\n"
        displayName: Trend Vision One™ Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trendmicro
        - visionone
        - container-security
        links:
        - name: Documentation
          url: https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator
        maintainers:
        - email: support@trendmicro.com
          name: TrendMicro VisionOne Support Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: TrendMicro
          url: trendmicro.com
        relatedImages:
        - public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
        version: 0.0.2
      entries:
      - name: visionone-containersecurity.v0.0.2
        version: 0.0.2
      name: stable
    defaultChannel: stable
    packageName: visionone-containersecurity
    provider:
      name: TrendMicro
      url: trendmicro.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Docling Maintainers
      provider-url: https://docling-project.github.io/docling/
    name: docling-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: docling-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2025-05-19T20:20:14Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DoclingServe is the Schema for the doclingserves API
            displayName: Docling Serve
            kind: DoclingServe
            name: doclingserves.docling.github.io
            version: v1alpha1
        description: |-
          **Overview**

          The Docling Operator is responsible for managing the full lifecycle of Docling Serve instances in your Kubernetes or OpenShift container platforms.


          **Docling Serve**

          The Docling Serve application deploys a REST API for consuming and integrating Docling in your workflows.

          The API offers endpoints for multiple types of interaction, i.e. both synchronous and asynchronous endpoints, as well as status subscriptions via websockets.

          The options for the Docling conversion can be customized at runtime by the users (e.g. which OCR engine, which enrichment models, etc). Users can also choose which of the many DoclingDocument export formats (JSON, Markdown, HTML, DocTags, etc) to receive in the result response.

          Docling Serve allows to choose between a local compute engine, i.e. running Docling in the same container as the API server, or a distributed compute engine which leverages Kubeflow Pipelines to parallelize large-scale ingestion jobs.

          **Docling**

          Docling is a powerful library which simplifies document processing, parsing diverse formats — including advanced PDF understanding — and providing seamless integrations with the gen AI ecosystem.

          Learn more about the Docling Project by visiting the GitHub organization at [https://github.com/docling-project](https://github.com/docling-project).


          **Sample DoclingServe Custom Resources**


          To deploy with the `local` compute engine with the UI enabled and Route exposed:
          ```yaml
          apiVersion: docling.github.io/v1alpha1
          kind: DoclingServe
          metadata:
            labels:
              app.kubernetes.io/name: docling-operator
              app.kubernetes.io/managed-by: kustomize
            name: doclingserve-local-sample
          spec:
            apiServer:
              image: "quay.io/docling-project/docling-serve:latest"
              enableUI: true
              instances: 1
            engine:
              local:
                numWorkers: 2
            route:
              enabled: true
          ```

          To deploy with the `KFP` compute engine with the UI enabled and Route exposed. **Note: Kubeflow Pipeline must be installed and setup in the cluster to use this option.**:
          ```yaml
          apiVersion: docling.github.io/v1alpha1
          kind: DoclingServe
          metadata:
            labels:
              app.kubernetes.io/name: docling-operator
              app.kubernetes.io/managed-by: kustomize
            name: doclingserve-kfp-sample
          spec:
            apiServer:
              image: "quay.io/docling-project/docling-serve:latest"
              enableUI: true
              instances: 1
            engine:
              kfp:
                endpoint: "https://NAME.NAMESPACE.svc.cluster.local:8888"
            route:
              enabled: true
          ```

          **Additional Configurations**


          The CRD exposes `apiServer.configMapName` to enable additional configurations of the `docling-serve` workload. For a full list of configurations can be found on docling-serve's
          [Configuration](https://github.com/docling-project/docling-serve/blob/main/docs/configuration.md) page.
        displayName: Docling Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - document
        - processing
        links:
        - name: Docling Operator
          url: https://docling-project.github.io/docling/
        maintainers:
        - email: docling-technical-discuss@lists.lfaidata.foundation
          name: Docling Maintainers
        maturity: alpha
        provider:
          name: Docling Maintainers
          url: https://docling-project.github.io/docling/
        relatedImages:
        - quay.io/docling-project/docling-operator@sha256:200d8db51874c4b1e3d20d8d1423e9a4807843fbb29f514cd5d8ba872e47f4af
        version: 0.0.1
      entries:
      - name: docling-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: docling-operator
    provider:
      name: Docling Maintainers
      url: https://docling-project.github.io/docling/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubemq.io
      provider-url: ""
    name: kubemq-operator-marketplace
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubemq-operator-marketplace.1.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
              "apiVersion": "core.k8s.kubemq.io/v1beta1",
              "kind": "KubemqCluster",
              "metadata": {
                "name": "kubemq-cluster"
              },
              "spec": {
                "key": "<Your-KubeMQ-License-Key>",
                "replicas": 3
              }
            },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqConnector",
                "metadata": {
                  "name": "kubemq-bridges"
                },
                "spec": {
                  "type": "bridges",
                  "config": "bindings: null"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.connect.redhat.com/kubemq/kubemq-operator@sha256:3a9c87e462eb1caa81925dee72407f73fca41f19a3f9bcaf4e76e569db1dfbab
          createdAt: 2020-03-11 12:00:00+00:00
          description: Kubernetes Messaging Platform
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          support: support@kubemq.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kubemq Cluster
            displayName: Kubemq Cluster
            kind: KubemqCluster
            name: kubemqclusters.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Connector
            displayName: Kubemq Connector
            kind: KubemqConnector
            name: kubemqconnectors.core.k8s.kubemq.io
            version: v1beta1
        description: |2-

          ## Kubernetes Messaging Platform
          A message-based architecture for microservices providing flexible connectivity for hybrid environments: multi-cloud, on-premises, and at the edge. Powered by enterprise-grade message broker and message queue with ready to use connectors, bridges, and control center.

          Messaging platform components:
            - **KubeMQ Cluster** - Enterprise-grade message broker and message queue, scalable, high available, and secured.
            - **KubeMQ Targets Connector** - connects a KubeMQ cluster with external systems and cloud services.
            - **KubeMQ Bridges Connector** - bridge, replicate, aggregate, and transform messages between KubeMQ clusters.
            - **KubeMQ Sources Connector** - transform external systems and cloud services data to KubeMQ cluster messages.
          \n### Kubernetes Native
          Innovative and modern message queue and message broker in a lightweight container developed to run in Kubernetes, certified in the CNCF landscape, and connect natively to the cloud-native ecosystem.

          ### Ease of use

          Simple deployment in Kubernetes in less than 1 minute. Developer friendly by simple to use SDKs and elimination of the many developers and DevOps-centered challenges to define exchanges, brokers, channels, routes, and predefined topics

          ### All messaging patterns

          KubeMQ is available with all messaging patterns such as Queue, Stream, Pub/Sub, and RPC, Saving the need to maintain multiple messaging platforms for different use cases.

          ### Hybrid Environments

          KubeMQ enables the microservices from multiple environments to communicate and build one hybrid infrastructure solution across clouds, on-prem, and at the edge.

          ### KubeMQ Cluster key features

          - All-batteries included Messaging Queue Broker for the Kubernetes environment
          - Deployed with Operator for full life cycle operation
          - Blazing fast (written in Go), small and lightweight Docker container
          - Asynchronous and Synchronous messaging with support for  `At Most Once Delivery` and `At Least Once Delivery` models
          - Supports durable FIFO based Queue, Publish-Subscribe Events, Publish-Subscribe with Persistence (Events Store), RPC Command and Query messaging patterns
          - Supports gRPC, Rest and WebSocket Transport protocols with TLS support (both RPC and Stream modes)
          - Supports Access control Authorization and Authentication
          - Supports message masticating and smart routing
          - No Message broker configuration needed (i.e., queues, exchanges)
          - .Net, Java, Python, Go, and NodeJS SDKs

          ### KubeMQ Targets Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major services types (databases, cache, messaging, serverless, HTTP, etc.)
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-targets)

          ### KubeMQ Bridges Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Build Any Topology - connects KubeMQ clusters in 1:1, 1:n , n:1, n:n
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-bridges)

          ### KubeMQ Sources Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major messaging services such as Kafka, ActiveMQ, SQS, and more.
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-sources)

          ## Getting Started

          [Start with OpenShift](https://docs.kubemq.io/getting-started/create-cluster/openshift)

          ## Get Enterprise Trial License

          [Register](https://account.kubemq.io/login/register)

          ## Request a Demo

          [See KubeMQ in Action](https://kubemq.io/request-a-demo/)
        displayName: Kubemq Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - messaging
        - queue
        - broker
        - pubsub
        - rpc
        links:
        - name: Learn more about Kubemq
          url: https://kubemq.io/
        - name: Register for free license
          url: https://account.kubemq.io/login/register
        - name: Contact Sales
          url: https://kubemq.io/contact-sales/
        - name: Documentation
          url: https://docs.kubemq.io/
        - name: Get Help
          url: https://kubemq.io/contact-support/
        maintainers:
        - email: support@kubemq.io
          name: kubemq
        maturity: alpha
        provider:
          name: Kubemq.io
        relatedImages:
        - registry.connect.redhat.com/kubemq/kubemq-operator@sha256:3a9c87e462eb1caa81925dee72407f73fca41f19a3f9bcaf4e76e569db1dfbab
        - registry.connect.redhat.com/kubemq/kubemq-uni@sha256:51154b2a112c5cdc17ee8d338524c02ccd995f9386180b1cc232b7a875a23e73
        - registry.connect.redhat.com/kubemq/kubemq-targets@sha256:2a09e1718a03729a121728e4e3683e7e1f0690789d7f9857d76bebf1cc01d436
        - registry.connect.redhat.com/kubemq/kubemq-sources@sha256:5736575b6cae39d129346f83fe2d8ab9cdbd1c1d2a99a66b4b9c01631fb5db21
        - registry.connect.redhat.com/kubemq/kubemq-sources@sha256:8ed531e58c60760888ffb18165883386bda9cc5624645353406c94750f5c3008
        version: 1.8.6
      entries:
      - name: kubemq-operator-marketplace.1.8.6
        version: 1.8.6
      - name: kubemq-operator-marketplace.1.8.5
        version: 1.8.5
      - name: kubemq-operator-marketplace.v1.8.4
        version: 1.8.4
      - name: kubemq-operator-marketplace.v1.8.3
        version: 1.8.3
      name: alpha
    defaultChannel: alpha
    packageName: kubemq-operator-marketplace
    provider:
      name: Kubemq.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: candidate
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: fast
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: stable
    defaultChannel: stable
    packageName: stackgres-certified
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzounstructured-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzounstructured-operator.v2.0.101
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzounstructured.clusters.cambridgesemantics.com/v1",
                "kind": "AnzoUnstructured",
                "metadata": {
                  "name": "au01"
                },
                "spec": {
                  "auWorker": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-w",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-unstructured-worker@sha256:4db61bdf6dc8b42824a06d2f8d8aeecddde765d123b40a63a7bbf87845719fb2",
                                "name": "w",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  },
                  "msLeader": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-ms",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-microservices-leader@sha256:0579fd97e81e436eaff3b58c5395f13aa924940a3d51f0bdea5163d327620286",
                                "name": "ms",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/unstructured-operator@sha256:a47af068a7aa55c12be84c80d263feeed5597ecc434986b4f2fff782b568bb04
          createdAt: "2021-08-02"
          description: Kubernetes Operator for Anzo Unstructured
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzounstructured-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzounstructured-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoUnstructured is the Schema for the anzounstructureds
              API
            displayName: AnzoUnstructured Operator
            kind: AnzoUnstructured
            name: anzounstructureds.anzounstructured.clusters.cambridgesemantics.com
            version: v1
        description: |-
          The Anzo Unstructured Operator provides the way to install and configure an anzo unstructured setup on Red Hat K8S environment.
          Currently, this is possible only through existing Anzo installation.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v2.0.0/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers through [ Cambridge Semantics Customer Center ]( https://customercenter.cambridgesemantics.com/ ).
        displayName: Anzo Unstructured Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Database
        - Big Data
        links:
        - name: Anzo Unstructured Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v2.0.0/README.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/unstructured-operator@sha256:a47af068a7aa55c12be84c80d263feeed5597ecc434986b4f2fff782b568bb04
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-microservices-leader@sha256:0579fd97e81e436eaff3b58c5395f13aa924940a3d51f0bdea5163d327620286
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-unstructured-worker@sha256:4db61bdf6dc8b42824a06d2f8d8aeecddde765d123b40a63a7bbf87845719fb2
        version: 2.0.0
      entries:
      - name: anzounstructured-operator.v2.0.101
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anzounstructured-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://github.com/che-incubator/kubernetes-image-puller-operator
    name: kubernetes-imagepuller-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubernetes-imagepuller-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "che.eclipse.org/v1alpha1",
                "kind": "KubernetesImagePuller",
                "metadata": {
                  "name": "image-puller"
                },
                "spec": {
                  "daemonsetName": "k8s-image-puller",
                  "images": "java11-maven=quay.io/eclipse/che-java11-maven:next;che-theia=quay.io/eclipse/che-theia:next"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/eclipse/kubernetes-image-puller-operator:1.1.0
          createdAt: "2024-01-15T16:23:04Z"
          description: Create and manage kubernetes-image-puller instances.
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/che-incubator/kubernetes-image-puller-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubernetesImagePuller is the Schema for the kubernetesimagepullers
              API
            displayName: Kubernetes Image Puller
            kind: KubernetesImagePuller
            name: kubernetesimagepullers.che.eclipse.org
            version: v1alpha1
        description: "## About the managed application\n\nCreate a `KubernetesImagePuller`
          custom resource to automatically configure and run an instance of the [kubernetes-image-puller.](https://github.com/che-incubator/kubernetes-image-puller)\n##
          About this Operator\n\nThe `KubernetesImagePuller` custom resource understands
          the following fields in the `spec`:\n\n1. `configMapName` - The name of
          the `ConfigMap` to create\t\n2. `daemonsetName` - The name of the `DaemonSet`
          to be create.\n3. `deploymentName` - The name of the `kubernetes-image-puller`
          `Deployment` to create.\n4. `images` - A list of key-value pairs separated
          by semicolons of images to pull.  For example: `java=quay.io/eclipse/che-java8-maven:latest;theia=quay.io/eclipse/che-theia:next`\n5.
          `cachingIntervalHours` - The amount of time, in hours, between `DaemonSet`
          health checks.\n6. `cachingMemoryRequest` - The memory request for each
          cached image when the puller is running.\n7. `cachingMemoryLimit` - The
          memory limit for each cached image when the puller is running.\n8. `nodeSelector`
          - Node selector applied to pods created by the `DaemonSet`.\n## Prerequisites
          for enabling this Operator\n\nThe operator requires an existing namespace
          to be installed in.\nCheck the [README](https://github.com/che-incubator/kubernetes-image-puller-operator#installing-the-operator)
          for more details.\n"
        displayName: Kubernetes Image Puller Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - che
        - kubernetes-image-puller
        links:
        - name: Repository
          url: https://github.com/che-incubator/kubernetes-image-puller-operator
        maintainers:
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        provider:
          name: Red Hat, Inc.
          url: https://github.com/che-incubator/kubernetes-image-puller-operator
        relatedImages:
        - quay.io/eclipse/kubernetes-image-puller-operator:1.1.0
        - quay.io/brancz/kube-rbac-proxy:v0.11.0
        version: 1.1.0
      entries:
      - name: kubernetes-imagepuller-operator.v1.1.0
        version: 1.1.0
      - name: kubernetes-imagepuller-operator.v1.0.6
        version: 1.0.6
      - name: kubernetes-imagepuller-operator.v1.0.5
        version: 1.0.5
      - name: kubernetes-imagepuller-operator.v1.0.4
        version: 1.0.4
      - name: kubernetes-imagepuller-operator.v1.0.3
        version: 1.0.3
      - name: kubernetes-imagepuller-operator.v1.0.2
        version: 1.0.2
      - name: kubernetes-imagepuller-operator.v1.0.1
        version: 1.0.1
      - name: kubernetes-imagepuller-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: kubernetes-imagepuller-operator
    provider:
      name: Red Hat, Inc.
      url: https://github.com/che-incubator/kubernetes-image-puller-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Palo Alto Networks
      provider-url: ""
    name: pcc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pcc-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "name": "pcc-console",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "ConsoleDefender",
                "metadata": {
                  "name": "pcc-consoledefender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "defenderConfig": {
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Defender",
                "metadata": {
                  "name": "pcc-defender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "defenderConfig": {
                    "clusterAddress": "twistlock-console.example.com",
                    "consoleAddress": "https://twistlock-console.example.com:8083",
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: quay.io/prismacloud/pcc-operator@sha256:b8fcfbd6c51286c874e00db1bd35523386cec406fa4050ef44c0a887730cf9b8
          createdAt: "2021-11-15"
          description: Deploy Prisma Cloud Compute for cloud-native security in your
            clusters
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator
          support: pcc-integrations@paloaltonetworks.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Install the Prisma Cloud Compute Console and Defenders
            displayName: Console and Defenders
            kind: ConsoleDefender
            name: consoledefenders.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install the Prisma Cloud Compute Console
            displayName: Console
            kind: Console
            name: consoles.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install Prisma Cloud Compute Defenders
            displayName: Defenders
            kind: Defender
            name: defenders.pcc.paloaltonetworks.com
            version: v1alpha1
        description: |
          ## Features
          - Deploy Console
              - Create initial local administrator user
              - Add license
          - Deploy Defenders
          - Support for offline deployments
          - Upgrade Console
          - Upgrade Defenders

          ## Installation
          See the [Prisma Cloud Compute Operator documentation](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator) for installation guides.

          ## Support
          This operator is available "as is," and should be seen as community-supported; however, Palo Alto Networks will continue contributing whenever possible.
          Palo Alto Networks does not provide technical support or help with using or troubleshooting the operator through normal support channels.
          We encourage you to open GitHub [issues](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/issues) and [pull requests](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/pulls) to track bugs and feature requests.

          ## Other links
          [Prisma Cloud Compute product documentation](https://docs.paloaltonetworks.com/prisma/prisma-cloud/prisma-cloud-admin-compute.html)

          [Prisma Cloud Compute API documentation](https://prisma.pan.dev/api/cloud/cwpp)
        displayName: Prisma Cloud Compute Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - palo
        - palo alto
        - palo alto networks
        - paloalto
        - twistlock
        - compute
        - prisma
        - prisma cloud
        - prisma cloud compute
        - security
        - image
        - images
        - container
        - containers
        - docker
        - podman
        - vulnerability
        - vulnerabilities
        - compliance
        - runtime
        - scan
        - scanning
        - monitor
        - monitoring
        - alert
        - alerting
        links:
        - name: Prisma Cloud
          url: https://www.paloaltonetworks.com/prisma/cloud
        maintainers:
        - email: pcc-integrations@paloaltonetworks.com
          name: Wyatt Gill
        maturity: alpha
        provider:
          name: Palo Alto Networks
        relatedImages:
        - quay.io/prismacloud/pcc-operator@sha256:b8fcfbd6c51286c874e00db1bd35523386cec406fa4050ef44c0a887730cf9b8
        version: 0.2.0
      entries:
      - name: pcc-operator.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: pcc-operator
    provider:
      name: Palo Alto Networks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: koku-metrics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: koku-metrics-operator.v4.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
          createdAt: "2025-09-25T19:04:35Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: koku-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Cost Management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Koku Metrics Operator
          ## Introduction
          The `koku-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Koku Metrics Operator (`koku-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created KokuMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The KokuMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.2.0:
          * Enabled gathering and reporting NVIDIA GPU metrics for resource optimization.
          * (Bugfix) Fixed custom CA certificate validation and configuration issues.
          * (Bugfix) Fixed issue where ROS queries were missing workload metadata for namespaces labeled only with the legacy `insights_cost_management_optimizations` label.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **Progress towards FIPS 140 Compliance:** The operator is using the Go Cryptographic Module v1.0.0 to progress toward FIPS 140 compliance. Although this module is not validated at the time of this release, the operator aims to meet stricter security standards when the module does successfully achieve FIPS validation.
          * **API Name Change:** The Custom Resource Definition (CRD) for configuring the upstream `Koku Metrics Operator` was renamed from `KokuMetricsConfig` to `CostManagementMetricsConfig`.

          ### NOTE:
            * The API name change impacts only users of the upstream (community) Koku Metrics Operator.
            * The downstream (Red Hat-supported) Cost Management Metrics Operator is not impacted by this change and users should continue using their existing configurations.

          **Important Upgrade Instructions:**

            If you are upgrading the upstream Koku Metrics Operator to version 4.0.0 or higher, you must manually migrate your configuration. The operator will no longer recognize existing `KokuMetricsConfig` resources.

            To successfully upgrade and retain your operator's configuration, complete the following steps. You can use the provided `oc` commands replacing the angle brackets (`< >`) with your specific values:

            1. Backup existing configuration (recommended):

               Before you upgrade, retrieve the details of your current `KokuMetricsConfig` instance to help you recreate it with the new API name.

               ```
               oc get kokumetricsconfig -n koku-metrics-operator <your-config-name> -o yaml > koku-metrics-config-backup.yaml
               ```

            2. Delete the previous `KokuMetricsConfig` instance (recommended):

                After the operator is upgraded to `version 4.0.0`, delete any existing `KokuMetricsConfig` resources.

                ```
                oc delete kokumetricsconfig -n koku-metrics-operator <your-config-name>
                ```

            3. Create the new `CostManagementMetricsConfig` instance:

                Using the information from your backup (or a new configuration), create a new `CostManagementMetricsConfig` instance. Adjust the spec section based on your specific needs. For more information about configurations, [refer to the configurable parameters](#configurable-parameters).

            4. Apply your new configuration:

                ```
                oc apply -f <your-costmanagementmetricsconfig-file>.yaml -n koku-metrics-operator
                ```

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created KokuMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the KokuMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `KokuMetricsConfig`, creating a new `KokuMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `koku-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: koku-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the KokuMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the KokuMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon KokuMetricsConfig creation** (This parameter will start to appear in KokuMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the KokuMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the koku-metrics-operator
          **Note** There are separate instructions for configuring the `koku-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `koku-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the koku-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `koku-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `koku-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the koku-metrics-operator for a restricted network
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `koku-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: koku-metrics-operator
                labels:
                  app: koku-metrics-operator
              spec:
                volumes:
                - name: koku-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: koku-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: koku-metrics-operator-reports
                    mountPath: /tmp/koku-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/koku-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/koku-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `koku-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the KokuMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Koku Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Koku Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
        version: 4.2.0
      entries:
      - name: koku-metrics-operator.v4.2.0
        version: 4.2.0
      - name: koku-metrics-operator.v4.1.0
        version: 4.1.0
      - name: koku-metrics-operator.v4.0.0
        version: 4.0.0
      - name: koku-metrics-operator.v3.3.2
        version: 3.3.2
      - name: koku-metrics-operator.v3.3.1
        version: 3.3.1
      - name: koku-metrics-operator.v3.3.0
        version: 3.3.0
      - name: koku-metrics-operator.v3.2.1
        version: 3.2.1
      - name: koku-metrics-operator.v3.2.0
        version: 3.2.0
      - name: koku-metrics-operator.v3.1.0
        version: 3.1.0
      - name: koku-metrics-operator.v3.0.1
        version: 3.0.1
      - name: koku-metrics-operator.v3.0.0
        version: 3.0.0
      - name: koku-metrics-operator.v2.0.0
        version: 2.0.0
      - name: koku-metrics-operator.v1.2.0
        version: 1.2.0
      - name: koku-metrics-operator.v1.1.9
        version: 1.1.9
      - name: koku-metrics-operator.v1.1.8
        version: 1.1.8
      - name: koku-metrics-operator.v1.1.7
        version: 1.1.7
      - name: koku-metrics-operator.v1.1.6
        version: 1.1.6
      - name: koku-metrics-operator.v1.1.5
        version: 1.1.5
      - name: koku-metrics-operator.v1.1.4
        version: 1.1.4
      - name: koku-metrics-operator.v1.1.3
        version: 1.1.3
      - name: koku-metrics-operator.v1.1.2
        version: 1.1.2
      - name: koku-metrics-operator.v1.1.1
        version: 1.1.1
      - name: koku-metrics-operator.v0.9.8
        version: 0.9.8
      - name: koku-metrics-operator.v0.9.7
        version: 0.9.7
      - name: koku-metrics-operator.v0.9.6
        version: 0.9.6
      - name: koku-metrics-operator.v0.9.5
        version: 0.9.5
      - name: koku-metrics-operator.v0.9.4
        version: 0.9.4
      - name: koku-metrics-operator.v0.9.3
        version: 0.9.3
      - name: koku-metrics-operator.v0.9.2
        version: 0.9.2
      - name: koku-metrics-operator.v0.9.1
        version: 0.9.1
      - name: koku-metrics-operator.v0.9.0
        version: 0.9.0
      name: alpha
    - currentCSV: koku-metrics-operator.v4.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
          createdAt: "2025-09-25T19:04:35Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: koku-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Cost Management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Koku Metrics Operator
          ## Introduction
          The `koku-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Koku Metrics Operator (`koku-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created KokuMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The KokuMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.2.0:
          * Enabled gathering and reporting NVIDIA GPU metrics for resource optimization.
          * (Bugfix) Fixed custom CA certificate validation and configuration issues.
          * (Bugfix) Fixed issue where ROS queries were missing workload metadata for namespaces labeled only with the legacy `insights_cost_management_optimizations` label.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **Progress towards FIPS 140 Compliance:** The operator is using the Go Cryptographic Module v1.0.0 to progress toward FIPS 140 compliance. Although this module is not validated at the time of this release, the operator aims to meet stricter security standards when the module does successfully achieve FIPS validation.
          * **API Name Change:** The Custom Resource Definition (CRD) for configuring the upstream `Koku Metrics Operator` was renamed from `KokuMetricsConfig` to `CostManagementMetricsConfig`.

          ### NOTE:
            * The API name change impacts only users of the upstream (community) Koku Metrics Operator.
            * The downstream (Red Hat-supported) Cost Management Metrics Operator is not impacted by this change and users should continue using their existing configurations.

          **Important Upgrade Instructions:**

            If you are upgrading the upstream Koku Metrics Operator to version 4.0.0 or higher, you must manually migrate your configuration. The operator will no longer recognize existing `KokuMetricsConfig` resources.

            To successfully upgrade and retain your operator's configuration, complete the following steps. You can use the provided `oc` commands replacing the angle brackets (`< >`) with your specific values:

            1. Backup existing configuration (recommended):

               Before you upgrade, retrieve the details of your current `KokuMetricsConfig` instance to help you recreate it with the new API name.

               ```
               oc get kokumetricsconfig -n koku-metrics-operator <your-config-name> -o yaml > koku-metrics-config-backup.yaml
               ```

            2. Delete the previous `KokuMetricsConfig` instance (recommended):

                After the operator is upgraded to `version 4.0.0`, delete any existing `KokuMetricsConfig` resources.

                ```
                oc delete kokumetricsconfig -n koku-metrics-operator <your-config-name>
                ```

            3. Create the new `CostManagementMetricsConfig` instance:

                Using the information from your backup (or a new configuration), create a new `CostManagementMetricsConfig` instance. Adjust the spec section based on your specific needs. For more information about configurations, [refer to the configurable parameters](#configurable-parameters).

            4. Apply your new configuration:

                ```
                oc apply -f <your-costmanagementmetricsconfig-file>.yaml -n koku-metrics-operator
                ```

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created KokuMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the KokuMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `KokuMetricsConfig`, creating a new `KokuMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `koku-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: koku-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the KokuMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the KokuMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon KokuMetricsConfig creation** (This parameter will start to appear in KokuMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the KokuMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the koku-metrics-operator
          **Note** There are separate instructions for configuring the `koku-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `koku-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the koku-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `koku-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `koku-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the koku-metrics-operator for a restricted network
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `koku-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: koku-metrics-operator
                labels:
                  app: koku-metrics-operator
              spec:
                volumes:
                - name: koku-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: koku-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: koku-metrics-operator-reports
                    mountPath: /tmp/koku-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/koku-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/koku-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `koku-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the KokuMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Koku Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Koku Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
        version: 4.2.0
      entries:
      - name: koku-metrics-operator.v4.2.0
        version: 4.2.0
      - name: koku-metrics-operator.v4.1.0
        version: 4.1.0
      - name: koku-metrics-operator.v4.0.0
        version: 4.0.0
      - name: koku-metrics-operator.v3.3.2
        version: 3.3.2
      - name: koku-metrics-operator.v3.3.1
        version: 3.3.1
      - name: koku-metrics-operator.v3.3.0
        version: 3.3.0
      - name: koku-metrics-operator.v3.2.1
        version: 3.2.1
      - name: koku-metrics-operator.v3.2.0
        version: 3.2.0
      - name: koku-metrics-operator.v3.1.0
        version: 3.1.0
      - name: koku-metrics-operator.v3.0.1
        version: 3.0.1
      - name: koku-metrics-operator.v3.0.0
        version: 3.0.0
      - name: koku-metrics-operator.v2.0.0
        version: 2.0.0
      - name: koku-metrics-operator.v1.2.0
        version: 1.2.0
      - name: koku-metrics-operator.v1.1.9
        version: 1.1.9
      - name: koku-metrics-operator.v1.1.8
        version: 1.1.8
      - name: koku-metrics-operator.v1.1.7
        version: 1.1.7
      - name: koku-metrics-operator.v1.1.6
        version: 1.1.6
      - name: koku-metrics-operator.v1.1.5
        version: 1.1.5
      - name: koku-metrics-operator.v1.1.4
        version: 1.1.4
      - name: koku-metrics-operator.v1.1.3
        version: 1.1.3
      - name: koku-metrics-operator.v1.1.2
        version: 1.1.2
      - name: koku-metrics-operator.v1.1.1
        version: 1.1.1
      - name: koku-metrics-operator.v0.9.8
        version: 0.9.8
      - name: koku-metrics-operator.v0.9.7
        version: 0.9.7
      - name: koku-metrics-operator.v0.9.6
        version: 0.9.6
      - name: koku-metrics-operator.v0.9.5
        version: 0.9.5
      - name: koku-metrics-operator.v0.9.4
        version: 0.9.4
      - name: koku-metrics-operator.v0.9.3
        version: 0.9.3
      - name: koku-metrics-operator.v0.9.2
        version: 0.9.2
      - name: koku-metrics-operator.v0.9.1
        version: 0.9.1
      - name: koku-metrics-operator.v0.9.0
        version: 0.9.0
      name: beta
    defaultChannel: beta
    packageName: koku-metrics-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Lenovo (Shanghai) Computer Technology Co., Ltd.
      provider-url: ""
    name: lenovo-locd-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lenovo-locd-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "locd.lenovo.com/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample",
                  "namespace": "locd-operator-system"
                },
                "spec": {
                  "affinity": {},
                  "argocd": {
                    "password": "xxxxxxxxxxx",
                    "server": "argocd.svc.cluster.local",
                    "user": "admin"
                  },
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "core": {
                    "image": "registry.connect.redhat.com/loc-devops/core@sha256:f771217b3f7849e1b63cc6abb93b8b0fc9636fd4a201e31807a15f773138c609",
                    "pullPolicy": "Always"
                  },
                  "fullnameOverride": "",
                  "global": {
                    "managementNamespace": "loc-devops",
                    "serviceEndpoint": "loc-devops.domain",
                    "serviceaccount": {
                      "annotations": {
                        "managedBy": "LOC DevOps"
                      },
                      "create": true,
                      "name": "loc-devops-admin"
                    }
                  },
                  "gui": {
                    "image": "registry.connect.redhat.com/loc-devops/gui@sha256:65142a36095188e879467038936b7b6b6ee39dcea6d42a437fa7066fb2bc0f44",
                    "pullPolicy": "Always"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "className": "nginx",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "loc-devops.domain",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "Prefix",
                            "servicePort": 8080
                          },
                          {
                            "path": "/api",
                            "pathType": "Prefix",
                            "servicePort": 8000
                          },
                          {
                            "path": "/swagger",
                            "pathType": "Prefix",
                            "servicePort": 8000
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "replicaCount": 1,
                  "resources": {},
                  "securityContext": {},
                  "service": {
                    "port": {
                      "core": 8000,
                      "gui": 8080
                    },
                    "type": "ClusterIP"
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Cluster
            name: clusters.locd.lenovo.com
            version: v1alpha1
        description: This is operator used to manage components of Lenovo Open Cloud
          DevOps
        displayName: lenovo-locd-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Application management
        links:
        - name: LOCD Operator
          url: https://locd-operator.domain
        maintainers:
        - email: xujt5@lenovo.com
          name: jintaoxu
        maturity: alpha
        provider:
          name: Lenovo (Shanghai) Computer Technology Co., Ltd.
        relatedImages:
        - registry.connect.redhat.com/loc-devops/gui@sha256:65142a36095188e879467038936b7b6b6ee39dcea6d42a437fa7066fb2bc0f44
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b5786bbbef725badf3dfcc2c2c7a86ead5ebb584c978c47aae8b9a62e241b80d
        - registry.connect.redhat.com/loc-devops/operator@sha256:4c17a3688ae3098e4b5c2c0530f49b1da8dedcc6fc1d3cd848513a1f189812a7
        - registry.connect.redhat.com/loc-devops/core@sha256:f771217b3f7849e1b63cc6abb93b8b0fc9636fd4a201e31807a15f773138c609
        version: 0.1.0
      entries:
      - name: lenovo-locd-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: lenovo-locd-operator
    provider:
      name: Lenovo (Shanghai) Computer Technology Co., Ltd.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtr-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtr-operator.1.2.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
             {
               "apiVersion": "windup.jboss.org/v1",
               "kind": "Windup",
               "metadata": {
                 "namespace": "mtr",
                 "name": "mtr",
                 "labels": {
                   "application": "windup"
                 }
               },
               "spec": {}
             }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mtr/mtr-rhel8-operator@sha256:4df7c4ad353339e8b958ffc651e811667767c8a0dcdd61f89352c926da52a513
          createdAt: "2024-09-03T17:22:23Z"
          description: Migration Toolkit for Runtimes is an analysis tool that supports
            the modernization and migration of Java applications.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          imagePullPolicy: Always
          operatorframework.io/suggested-namespace: mtr
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/windup/windup-operator/
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "Create an MTR Web Console application.  \n\nTo access the
              MTR Web Console application from the \"Developer\" perspective, please
              go to Topology and click on the \"Open URL\" icon on the web-console
              pod.  \n\nAlternatively from the \"Administrator\" perspective, please
              go to Networking->Routes and click on the Location hyperlink.  \n\nIn
              order to connect with default login credentials, please use \"migration\"
              as username and \"password\" as password.  \n\nKnown issue  \nIf you
              want to customize the MTR Web Console instance's parameters and you
              can not see any in the `Form view`, please switch to the `YAML view`
              and change them as needed.\n"
            displayName: Migration Toolkit for Runtimes
            kind: Windup
            name: windups.windup.jboss.org
            version: v1
        description: "The Migration Toolkit for Runtimes (MTR) is a web console application
          that supports large-scale Java application modernization and migration projects
          across a broad range of transformations and use cases.  \n\nIt analyzes
          application code, supports effort estimation, accelerates code migration,
          and enables users to move applications to containers.  \n\nFor more information
          please refer to the [https://developers.redhat.com/products/mtr/overview](https://developers.redhat.com/products/mtr/overview)
          page.\n"
        displayName: Migration Toolkit for Runtimes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - security
        - alerting
        - metric
        - troubleshooting
        - run-time
        - migration
        - modernization
        links:
        - name: Website
          url: https://developers.redhat.com/products/mtr/overview
        - name: GitHub
          url: https://github.com/windup/windup-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_runtimes
        maintainers:
        - email: migrate@redhat.com
          name: Migration Toolkit for Runtimes
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mtr/mtr-rhel8-operator@sha256:4df7c4ad353339e8b958ffc651e811667767c8a0dcdd61f89352c926da52a513
        - registry.redhat.io/mtr/mtr-web-container-rhel8@sha256:cb88a5453b506ca70563537ec6db2f44ff8cb4050c9652ec86eb5128aa313003
        - registry.redhat.io/mtr/mtr-web-executor-container-rhel8@sha256:2f0622774367f33efc321926196681322ab1535144e6a0785608c8050fc59f2e
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        version: 1.2.7
      entries:
      - name: mtr-operator.1.2.7
        version: 1.2.7
      - name: mtr-operator.1.2.6
        version: 1.2.6
      - name: mtr-operator.1.2.5
        version: 1.2.5
      - name: mtr-operator.1.2.4
        version: 1.2.4
      - name: mtr-operator.1.2.3
        version: 1.2.3
      - name: mtr-operator.1.2.2
        version: 1.2.2
      - name: mtr-operator.1.2.1
        version: 1.2.1
      - name: mtr-operator.1.2.0
        version: 1.2.0
      - name: mtr-operator.1.1.1
        version: 1.1.1
      - name: mtr-operator.1.1.0
        version: 1.1.0
      - name: mtr-operator.1.0.3
        version: 1.0.3
      - name: mtr-operator.1.0.2
        version: 1.0.2
      - name: mtr-operator.1.0.1
        version: 1.0.1
      - name: mtr-operator.1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: mtr-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: GitLab Inc
      provider-url: https://about.gitlab.com/
    name: gitlab-operator-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "certmanager": {
                        "install": false
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-26T07:16:35Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3281fe12aaae428fd7102332e7c0d7033515042d7830eba3953891ab0c3e3a29
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      - name: gitlab-operator-kubernetes.v2.1.1
        version: 2.1.1
      - name: gitlab-operator-kubernetes.v2.1.0
        version: 2.1.0
      - name: gitlab-operator-kubernetes.v2.0.2
        version: 2.0.2
      - name: gitlab-operator-kubernetes.v2.0.1
        version: 2.0.1
      - name: gitlab-operator-kubernetes.v2.0.0
        version: 2.0.0
      - name: gitlab-operator-kubernetes.v1.12.2
        version: 1.12.2
      - name: gitlab-operator-kubernetes.v1.12.1
        version: 1.12.1
      - name: gitlab-operator-kubernetes.v1.12.0
        version: 1.12.0
      - name: gitlab-operator-kubernetes.v1.11.3
        version: 1.11.3
      - name: gitlab-operator-kubernetes.v1.11.2
        version: 1.11.2
      - name: gitlab-operator-kubernetes.v1.11.1
        version: 1.11.1
      - name: gitlab-operator-kubernetes.v1.11.0
        version: 1.11.0
      - name: gitlab-operator-kubernetes.v1.10.2
        version: 1.10.2
      - name: gitlab-operator-kubernetes.v1.10.1
        version: 1.10.1
      - name: gitlab-operator-kubernetes.v1.10.0
        version: 1.10.0
      - name: gitlab-operator-kubernetes.v1.9.2
        version: 1.9.2
      - name: gitlab-operator-kubernetes.v1.9.1
        version: 1.9.1
      - name: gitlab-operator-kubernetes.v1.9.0
        version: 1.9.0
      - name: gitlab-operator-kubernetes.v1.8.3
        version: 1.8.3
      - name: gitlab-operator-kubernetes.v1.8.2
        version: 1.8.2
      - name: gitlab-operator-kubernetes.v1.8.1
        version: 1.8.1
      - name: gitlab-operator-kubernetes.v1.8.0
        version: 1.8.0
      - name: gitlab-operator-kubernetes.v1.7.2
        version: 1.7.2
      - name: gitlab-operator-kubernetes.v1.7.1
        version: 1.7.1
      - name: gitlab-operator-kubernetes.v1.7.0
        version: 1.7.0
      - name: gitlab-operator-kubernetes.v1.6.1
        version: 1.6.1
      - name: gitlab-operator-kubernetes.v1.6.0
        version: 1.6.0
      - name: gitlab-operator-kubernetes.v1.5.1
        version: 1.5.1
      - name: gitlab-operator-kubernetes.v1.4.2
        version: 1.4.2
      - name: gitlab-operator-kubernetes.v1.4.1
        version: 1.4.1
      - name: gitlab-operator-kubernetes.v1.4.0
        version: 1.4.0
      - name: gitlab-operator-kubernetes.v1.3.3
        version: 1.3.3
      - name: gitlab-operator-kubernetes.v1.3.2
        version: 1.3.2
      - name: gitlab-operator-kubernetes.v1.3.1
        version: 1.3.1
      - name: gitlab-operator-kubernetes.v1.3.0
        version: 1.3.0
      name: stable
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "certmanager": {
                        "install": false
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-26T07:16:35Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3281fe12aaae428fd7102332e7c0d7033515042d7830eba3953891ab0c3e3a29
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      - name: gitlab-operator-kubernetes.v2.1.1
        version: 2.1.1
      - name: gitlab-operator-kubernetes.v2.1.0
        version: 2.1.0
      - name: gitlab-operator-kubernetes.v2.0.2
        version: 2.0.2
      - name: gitlab-operator-kubernetes.v2.0.1
        version: 2.0.1
      - name: gitlab-operator-kubernetes.v2.0.0
        version: 2.0.0
      - name: gitlab-operator-kubernetes.v1.12.2
        version: 1.12.2
      - name: gitlab-operator-kubernetes.v1.12.1
        version: 1.12.1
      - name: gitlab-operator-kubernetes.v1.12.0
        version: 1.12.0
      - name: gitlab-operator-kubernetes.v1.11.3
        version: 1.11.3
      - name: gitlab-operator-kubernetes.v1.11.2
        version: 1.11.2
      - name: gitlab-operator-kubernetes.v1.11.1
        version: 1.11.1
      - name: gitlab-operator-kubernetes.v1.11.0
        version: 1.11.0
      - name: gitlab-operator-kubernetes.v1.10.2
        version: 1.10.2
      - name: gitlab-operator-kubernetes.v1.10.1
        version: 1.10.1
      - name: gitlab-operator-kubernetes.v1.10.0
        version: 1.10.0
      - name: gitlab-operator-kubernetes.v1.9.2
        version: 1.9.2
      - name: gitlab-operator-kubernetes.v1.9.1
        version: 1.9.1
      - name: gitlab-operator-kubernetes.v1.9.0
        version: 1.9.0
      - name: gitlab-operator-kubernetes.v1.8.3
        version: 1.8.3
      - name: gitlab-operator-kubernetes.v1.8.2
        version: 1.8.2
      - name: gitlab-operator-kubernetes.v1.8.1
        version: 1.8.1
      - name: gitlab-operator-kubernetes.v1.8.0
        version: 1.8.0
      - name: gitlab-operator-kubernetes.v1.7.2
        version: 1.7.2
      - name: gitlab-operator-kubernetes.v1.7.1
        version: 1.7.1
      - name: gitlab-operator-kubernetes.v1.7.0
        version: 1.7.0
      - name: gitlab-operator-kubernetes.v1.6.1
        version: 1.6.1
      - name: gitlab-operator-kubernetes.v1.6.0
        version: 1.6.0
      - name: gitlab-operator-kubernetes.v1.5.1
        version: 1.5.1
      - name: gitlab-operator-kubernetes.v1.4.2
        version: 1.4.2
      - name: gitlab-operator-kubernetes.v1.4.1
        version: 1.4.1
      - name: gitlab-operator-kubernetes.v1.4.0
        version: 1.4.0
      - name: gitlab-operator-kubernetes.v1.3.3
        version: 1.3.3
      - name: gitlab-operator-kubernetes.v1.3.2
        version: 1.3.2
      - name: gitlab-operator-kubernetes.v1.3.1
        version: 1.3.1
      - name: gitlab-operator-kubernetes.v1.3.0
        version: 1.3.0
      name: unstable
    defaultChannel: stable
    packageName: gitlab-operator-kubernetes
    provider:
      name: GitLab Inc
      url: https://about.gitlab.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Silicom Ltd.
      provider-url: http://www.silicom-usa.com
    name: silicom-sts-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: silicom-sts-operator.v0.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "namespace": "openshift-operators",
                  "nodeSelector": {
                    "sts.silicom.com/config": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1"
                },
                "spec": {
                  "aprLevel": 0,
                  "domainNumber": 24,
                  "esmcMode": 1,
                  "forwardable": 0,
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-TBC.8275.1",
                  "mode10MHz": 2,
                  "modePPS": 2,
                  "namespace": "openshift-operators",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  },
                  "phyLedsCtl": 0,
                  "priority2": 128,
                  "src10MHz": 1,
                  "srcPPS": 1,
                  "syncOption": 1,
                  "synceCpu": 10,
                  "synceRecClkPort": 0,
                  "tracePtpMsg": -1,
                  "twoStep": 0
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "worker2"
                },
                "spec": null,
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "images": null,
                  "sro": null
                }
              }
            ]
          capabilities: Basic Install
          categories: Drivers and plugins
          certified: "false"
          containerImage: quay.io/silicom/sts-operator:0.0.6
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicomdk/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicomDK/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/silicomDK/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - quay.io/silicom/sts-plugin@sha256:44339fd400d46d9dddb02c2da47c6f9f96ad2605a5c6de8a7dce2f8c3da3bb57
        - quay.io/silicom/phc2sys@sha256:0f3b63a3059b7d6158a916814281097f0d418b99a5fac9e8e5fd9bae58fb0b74
        - quay.io/silicom/tsyncd@sha256:78be78ec8939f025f9a8b899693ab77b3990116d1c7ffb2115c620b7c3382d9d
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/silicom/sts-operator:0.0.6
        - quay.io/silicom/grpc-tsyncd@sha256:aac7233423439e352d8eb39cef63e479c30c2f80b063c84ffc8797cc5a045117
        - quay.io/silicom/tsync_extts@sha256:be6a5950a5fdc6fdf8015767901856bf173d01bdc803be7141e4f9795ab1478a
        - quay.io/silicom/ice-driver-src@sha256:ca728f648769d31501cdedda92554bcfa890754405b0b06d41629cea47707da8
        - quay.io/silicom/gpsd@sha256:649e8a6c524920e1c564d8f797b8d64235b2527aa923f94bc9c7a0b7aba8ae7f
        version: 0.0.6
      entries:
      - name: silicom-sts-operator.v0.0.6
        version: 0.0.6
      - name: silicom-sts-operator.v0.0.5
        version: 0.0.5
      - name: silicom-sts-operator.v0.0.4
        version: 0.0.4
      - name: silicom-sts-operator.v0.0.3
        version: 0.0.3
      - name: silicom-sts-operator.v0.0.2
        version: 0.0.2
      - name: silicom-sts-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: silicom-sts-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "sts-silicom"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "namespace": "sts-silicom",
                  "nodeSelector": {
                    "sts.silicom.com/config": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1"
                },
                "spec": {
                  "aprLevel": 0,
                  "domainNumber": 24,
                  "esmcMode": 1,
                  "forwardable": 0,
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-TBC.8275.1",
                  "mode10MHz": 2,
                  "modePPS": 2,
                  "namespace": "sts-silicom",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  },
                  "phyLedsCtl": 0,
                  "priority2": 128,
                  "src10MHz": 1,
                  "srcPPS": 1,
                  "syncOption": 1,
                  "synceCpu": 10,
                  "synceRecClkPort": 0,
                  "tracePtpMsg": -1,
                  "twoStep": 0
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "worker2"
                },
                "spec": null,
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config"
                },
                "spec": {
                  "gpsSvcPort": 2947,
                  "grpcSvcPort": 50051,
                  "images": {
                    "gpsd": "quay.io/silicom/gpsd:3.23.1",
                    "grpcTsyncd": "quay.io/silicom/grpc-tsyncd:2.1.0.0",
                    "phcs2Sys": "quay.io/silicom/phcs2sys:3.1.1",
                    "stsPlugin": "quay.io/silicom/sts-plugin:0.0.2",
                    "ts2Phcs": "quay.io/silicom/ts2phcs:1.0.0",
                    "tsyncd": "quay.io/silicom/tsyncd:2.1.0.0"
                  },
                  "namespace": "sts-silicom",
                  "sro": {
                    "chart": {
                      "name": "ice-special-resource",
                      "repository": {
                        "name": "ice-special-resource",
                        "url": "http://ice-driver-src"
                      },
                      "version": "0.0.1"
                    },
                    "iceVersion": "1.7.16",
                    "namespace": "openshift-operators",
                    "srcImage": "quay.io/silicom/ice-driver-src:1.7.16",
                    "srcSvcPort": 3000
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Drivers and plugins
          certified: "false"
          containerImage: quay.io/silicom/sts-operator:0.0.2
          createdAt: 2021-11-17T10:21:24 CET
          description: This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of Silicom STS device
            plugins for Kubernetes.
          operators.operatorframework.io.bundle.channels.v1: stable,development
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicomdk/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicomDK/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/silicomDK/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/silicom/sts-operator:0.0.2
        version: 0.0.2
      entries:
      - name: silicom-sts-operator.v0.0.2
        version: 0.0.2
      - name: silicom-sts-operator.v0.0.1
        version: 0.0.1
      name: beta
    defaultChannel: alpha
    packageName: silicom-sts-operator
    provider:
      name: Silicom Ltd.
      url: http://www.silicom-usa.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-logging
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-logging.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Logging",
                "metadata": {
                  "name": "logging-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "elasticsearch": {
                    "storage": "16Gi",
                    "tag": "7.17.6-3747-35"
                  },
                  "environment": "openshift",
                  "fluentbit": {
                    "runOnMaster": false,
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "grafana": {
                    "dashboards": [
                      "wap",
                      "load-core"
                    ],
                    "dataSources": [
                      "wap",
                      "load-core"
                    ]
                  },
                  "image": {
                    "elasticInitImage": "rhel8/toolbox@sha256:a753c24cb8b03066d494f010ad654d23fb0559f14d7121542298da3e0d48340d",
                    "elasticInitRepository": "registry.access.redhat.com",
                    "fbImage": "fluent-bit:1.8.15-3747-20",
                    "fbRepository": "prod.harbor.keysight.digital/loadcore",
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "createClusterRoles": true,
                    "serviceAccountName": "load-core-service-account"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:33:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Logging
            name: loggings.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Logging Operator
        displayName: Keysight WAP Logging Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Logging
          url: https://keysight-wap-logging.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-logging@sha256:f1e3b3ad84cd0f96c92cdaf4609e29257035cb982a761e832dae2095e636c721
        version: 0.1.21
      entries:
      - name: keysight-wap-logging.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-logging
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hazelcast, Inc
      provider-url: ""
    name: hazelcast-platform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hazelcast-platform-operator.v5.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "cache-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "CronHotBackup",
                "metadata": {
                  "name": "cronhotbackup-sample"
                },
                "spec": {
                  "hotBackupTemplate": {
                    "metadata": {
                      "labels": {
                        "cron-hotbackup": "true"
                      }
                    },
                    "spec": {
                      "hazelcastResourceName": "hazelcast"
                    }
                  },
                  "schedule": "*/30 * * * *"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Flow",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hazelcast-platform-operator"
                  },
                  "name": "flow"
                },
                "spec": {
                  "database": {
                    "host": "flow-postgresql",
                    "secretName": "flow-db-secret"
                  },
                  "env": [
                    {
                      "name": "OPTIONS",
                      "value": "--flow.analytics.persistRemoteCallResponses=true --flow.analytics.persistResults=false --flow.config.custom.managementCenterUrl=http://34.140.160.158/mc\n"
                    }
                  ],
                  "externalConnectivity": {
                    "ingress": {
                      "annotations": {
                        "kubernetes.io/ingress.class": "nginx",
                        "nginx.ingress.kubernetes.io/enable-opentelemetry": "true",
                        "nginx.ingress.kubernetes.io/ssl-redirect": "false"
                      },
                      "hostname": "",
                      "ingressClassName": "nginx",
                      "path": "/"
                    }
                  },
                  "imagePullSecrets": [
                    {
                      "name": "hazelcast-cloud-hazelcast-connect-demo-pull-secret"
                    }
                  ],
                  "licenseKeySecretName": "flow-mc-license-key",
                  "repository": "quay.io/hazelcast_cloud/hazelcast-flow-internal",
                  "size": 3,
                  "version": "next"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast"
                },
                "spec": {
                  "clusterSize": 3,
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast-sample"
                },
                "spec": {
                  "clusterSize": 3,
                  "jet": {
                    "bucketConfig": {
                      "bucketURI": "gs://operator-user-code/jetJobs",
                      "secretName": "br-secret-gcp"
                    },
                    "enabled": true,
                    "resourceUploadEnabled": true
                  },
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HazelcastEndpoint",
                "metadata": {
                  "name": "hazelcastendpoint-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "port": 5701,
                  "type": "Discovery"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HotBackup",
                "metadata": {
                  "name": "hot-backup"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-longrun"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-longrun-2.0.0.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-1.0.2.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJobSnapshot",
                "metadata": {
                  "name": "jetjobsnapshot-sample"
                },
                "spec": {
                  "cancelJob": false,
                  "jetJobResourceName": "jet-job"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ManagementCenter",
                "metadata": {
                  "name": "managementcenter"
                },
                "spec": {
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "hazelcastClusters": [
                    {
                      "address": "hazelcast",
                      "name": "dev"
                    }
                  ],
                  "licenseKeySecretName": "hazelcast-license-key",
                  "persistence": {
                    "enabled": true,
                    "size": "10Gi"
                  },
                  "repository": "hazelcast/management-center"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Map",
                "metadata": {
                  "name": "map"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "MultiMap",
                "metadata": {
                  "name": "multimap-sample"
                },
                "spec": {
                  "backupCount": 3,
                  "binary": true,
                  "collectionType": "LIST",
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "queue-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ReplicatedMap",
                "metadata": {
                  "name": "replicatedmap-sample"
                },
                "spec": {
                  "asyncFillup": true,
                  "hazelcastResourceName": "hazelcast",
                  "inMemoryFormat": "OBJECT"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "topic-sample"
                },
                "spec": {
                  "globalOrderingEnabled": true,
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "UserCodeNamespace",
                "metadata": {
                  "name": "ucn-sample"
                },
                "spec": {
                  "bucketConfig": {
                    "bucketURI": "gs://operator-user-code/userCodeNamespace",
                    "secretName": "br-secret-gcp"
                  },
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "VectorCollection",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hazelcast-platform-operator"
                  },
                  "name": "vectorcollection-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "indexes": [
                    {
                      "dimension": 10,
                      "efConstruction": 256,
                      "maxDegree": 32,
                      "metric": "Dot",
                      "useDeduplication": false
                    }
                  ]
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanReplication",
                "metadata": {
                  "name": "wanreplication-sample"
                },
                "spec": {
                  "endpoints": "35.192.33.252",
                  "resources": [
                    {
                      "kind": "Map",
                      "name": "map-sample"
                    },
                    {
                      "kind": "Hazelcast",
                      "name": "hz-sample"
                    }
                  ],
                  "targetClusterName": "dev"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanSync",
                "metadata": {
                  "name": "wansync-sample"
                },
                "spec": {
                  "wanReplicationResourceName": "wan-replication-sample"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: docker.io/hazelcast/hazelcast-platform-operator@sha256:59328a4ea7a4e2110e0750364db792cf266f1e7231e3bb321e3de09d535dab23
          createdAt: "2025-03-13T11:59:59Z"
          description: Install Hazelcast clusters in Kubernetes environments.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hazelcast/hazelcast-platform-operator/
          support: Hazelcast, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.hazelcast.com
            version: v1alpha1
          - description: CronHotBackup is the Schema for the cronhotbackups API
            displayName: Cron Hot Backup
            kind: CronHotBackup
            name: cronhotbackups.hazelcast.com
            version: v1alpha1
          - description: Flow is the Schema for the flows API
            displayName: Flow
            kind: Flow
            name: flows.hazelcast.com
            version: v1alpha1
          - description: HazelcastEndpoint is the Schema for the hazelcastendpoints
              API
            displayName: Hazelcast Endpoint
            kind: HazelcastEndpoint
            name: hazelcastendpoints.hazelcast.com
            version: v1alpha1
          - description: Hazelcast is the Schema for the hazelcasts API
            displayName: Hazelcast
            kind: Hazelcast
            name: hazelcasts.hazelcast.com
            version: v1alpha1
          - description: HotBackup is the Schema for the hot backup API
            displayName: Hot Backup
            kind: HotBackup
            name: hotbackups.hazelcast.com
            version: v1alpha1
          - description: JetJob is the Schema for the jetjobs API
            displayName: Jet Job
            kind: JetJob
            name: jetjobs.hazelcast.com
            version: v1alpha1
          - description: JetJobSnapshot is the Schema for the jetjobsnapshots API
            displayName: Jet Job Snapshot
            kind: JetJobSnapshot
            name: jetjobsnapshots.hazelcast.com
            version: v1alpha1
          - description: ManagementCenter is the Schema for the managementcenters
              API
            displayName: Management Center
            kind: ManagementCenter
            name: managementcenters.hazelcast.com
            version: v1alpha1
          - description: Map is the Schema for the maps API
            displayName: Map
            kind: Map
            name: maps.hazelcast.com
            version: v1alpha1
          - description: MultiMap is the Schema for the multimaps API
            displayName: Multi Map
            kind: MultiMap
            name: multimaps.hazelcast.com
            version: v1alpha1
          - description: Queue is the Schema for the queues API
            displayName: Queue
            kind: Queue
            name: queues.hazelcast.com
            version: v1alpha1
          - description: ReplicatedMap is the Schema for the replicatedmaps API
            displayName: Replicated Map
            kind: ReplicatedMap
            name: replicatedmaps.hazelcast.com
            version: v1alpha1
          - description: Topic is the Schema for the topics API
            displayName: Topic
            kind: Topic
            name: topics.hazelcast.com
            version: v1alpha1
          - description: UserCodeNamespace is the Schema for the usercodenamespaces
              API
            displayName: User Code Namespace
            kind: UserCodeNamespace
            name: usercodenamespaces.hazelcast.com
            version: v1alpha1
          - description: VectorCollection is the Schema for the vectorcollections
              API
            displayName: Vector Collection
            kind: VectorCollection
            name: vectorcollections.hazelcast.com
            version: v1alpha1
          - description: WanReplication is the Schema for the wanreplications API
            displayName: Wan Replication
            kind: WanReplication
            name: wanreplications.hazelcast.com
            version: v1alpha1
          - description: WanSync is the Schema for the wansyncs API
            displayName: Wan Sync
            kind: WanSync
            name: wansyncs.hazelcast.com
            version: v1alpha1
        description: |
          Easily deploy Hazelcast clusters and Management Center into Kubernetes environments and manage their lifecycles.

          ## Before You Start

          If you are planning to create Hazelcast Platform Enterprise clusters, you need to [create a secret](https://docs.hazelcast.com/operator/latest/get-started#step-2-start-the-hazelcast-cluster) for the license. You can request a trial license key from [here](https://trialrequest.hazelcast.com).

          For Hazelcast Platform clusters, you can simply continue.

          ## Documentation

          1. [Get started](https://docs.hazelcast.com/operator/latest/get-started) with the Operator
          2. [Connect to the cluster from outside Kubernetes](https://guides.hazelcast.org/hazelcast-platform-operator-expose-externally)
          3. [Restore a Cluster from Cloud Storage with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-external-backup-restore)
          4. [Replicate Data between Two Hazelcast Clusters with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-wan-replication)

          ## Features

          Hazelcast Platform Operator supports the features below:

          * Custom resource for Hazelcast Platform (Enterprise) and Management Center
          * Observe status of Hazelcast clusters and Management Center instances
          * High Availability Mode configuration to create clusters that are resilient to node and zone failures
          * Support for TLS connections between members using self-signed certificates
          * Scale up and down Hazelcast clusters
          * Expose Hazelcast cluster to external clients ([Smart & Unisocket](https://docs.hazelcast.com/hazelcast/latest/clients/java#java-client-operation-modes))
          * Backup Hazelcast persistence data to cloud storage with the possibility of scheduling it and restoring the data accordingly
          * WAN Replication feature when you need to synchronize multiple Hazelcast clusters, which are connected by WANs
          * Full and Delta WAN Sync support
          * Tiered Storage support for Hazelcast and Map CRs
          * CP Subsystem configuration support for Hazelcast CR
          * User Code Deployment feature, which allows you to deploy custom and domain classes from cloud storages and URLs to Hazelcast members
          * The User Code Namespaces feature allows you to deploy custom and domain classes at runtime without a members restart
          * Support the configuration of advanced networking options
          * Support Multi-namespace configuration
          * ExecutorService and EntryProcessor support
          * Support several data structures like Map, Topic, MultiMap, ReplicatedMap, Queue and Cache which can be created dynamically via specific Custom Resources
          * MapStore, Near Cache and off-heap memory (HD memory and native memory) support for the Map CR
          * Native Memory support for the Cache CR
          * Support Jet configuration and Jet Job submission using the JetJob CR
          * Support for exporting the snapshots of JetJob CRs using JetJobSnapshot CR
          * Support for custom configurations using ConfigMap
        displayName: Hazelcast Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hazelcast
        - keyvalue
        - in-memory
        - database
        - caching
        links:
        - name: Documentation
          url: https://docs.hazelcast.com/operator/latest/
        - name: Source Code
          url: https://github.com/hazelcast/hazelcast-platform-operator
        maintainers:
        - email: cloudnative@hazelcast.com
          name: Hazelcast Cloud Native Team
        maturity: alpha
        minKubeVersion: 1.22.7
        provider:
          name: Hazelcast, Inc
        relatedImages:
        - docker.io/hazelcast/hazelcast-platform-operator@sha256:59328a4ea7a4e2110e0750364db792cf266f1e7231e3bb321e3de09d535dab23
        version: 5.15.0
      entries:
      - name: hazelcast-platform-operator.v5.15.0
        version: 5.15.0
      name: alpha
    defaultChannel: alpha
    packageName: hazelcast-platform-operator
    provider:
      name: Hazelcast, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: tang-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tang-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "daemons.redhat.com/v1alpha1",
                "kind": "TangServer",
                "metadata": {
                  "finalizers": [
                    "finalizer.daemons.tangserver.redhat.com"
                  ],
                  "name": "tangserver",
                  "namespace": "nbde"
                },
                "spec": {
                  "healthScript": "/usr/bin/tangd-health-check",
                  "image": "registry.redhat.io/rhel9/tang",
                  "keyPath": "/var/db/tang",
                  "persistentVolumeClaim": "tangserver-pvc",
                  "replicas": 1,
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:f43c96035f43040b3644e37119a3b35fc433f366d4b636459d1b9b02a05b5d3d
          createdAt: 2023-11-23T12:49:40
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TangServer is the Schema for the tangservers API
            displayName: Tang Server
            kind: TangServer
            name: tangservers.daemons.redhat.com
            version: v1alpha1
        description: NBDE Tang Server operator allows Tang Server deployment on OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nbde
        - tang
        links:
        - name: Tang Operator
          url: https://github.com/latchset/tang-operator
        maintainers:
        - email: sarroutb@redhat.com
          name: Sergio Arroutbi Braojos
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:97cade2c1ee468261aec5400728c8d44de387b459134aec7a4c3b5ec5a335d2c
        - registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:f43c96035f43040b3644e37119a3b35fc433f366d4b636459d1b9b02a05b5d3d
        version: 1.0.0
      entries:
      - name: tang-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: tang-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "daemons.redhat.com/v1alpha1",
                "kind": "TangServer",
                "metadata": {
                  "finalizers": [
                    "finalizer.daemons.tangserver.redhat.com"
                  ],
                  "name": "tangserver",
                  "namespace": "nbde"
                },
                "spec": {
                  "healthScript": "/usr/bin/tangd-health-check",
                  "image": "registry.redhat.io/rhel9/tang",
                  "keyPath": "/var/db/tang",
                  "persistentVolumeClaim": "tangserver-pvc",
                  "replicas": 1,
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:1cb2f794689817282ac570cacb106d75a2751a1926d89d3b83b619e02d869e5e
          createdAt: 2024-12-04T19:01:27
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TangServer is the Schema for the tangservers API
            displayName: Tang Server
            kind: TangServer
            name: tangservers.daemons.redhat.com
            version: v1alpha1
        description: NBDE Tang Server operator allows Tang Server deployment on OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nbde
        - tang
        links:
        - name: Tang Operator
          url: https://github.com/latchset/tang-operator
        maintainers:
        - email: sarroutb@redhat.com
          name: Sergio Arroutbi Braojos
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1172e150fff22c5eeab572f26961f3f53fbf896ee76d08c7503cfe2777c55458
        - registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:1cb2f794689817282ac570cacb106d75a2751a1926d89d3b83b619e02d869e5e
        version: 1.0.3
      entries:
      - name: tang-operator.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: alpha
    packageName: tang-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: www.redhat.com
    name: compliance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: compliance-operator.v0.1.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.17 <0.1.32'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          repository: https://github.com/openshift/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repo
          url: https://github.com/openshift/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/4.7/security/compliance_operator/compliance-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/compliance-rhel8-operator@sha256:ed68a0574ddeaf7d676403cd843f28cb586894f35dbb3ca5e06a5264bd42f030
        - registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c
        - registry.redhat.io/openshift4/compliance-openscap-rhel8@sha256:867c6d4205653d8a954d234cbaf470313009eb983d4f455fb45f49c4a7764422
        version: 0.1.32
      entries:
      - name: compliance-operator.v0.1.32
        version: 0.1.32
      name: "4.7"
    - currentCSV: compliance-operator.v0.1.61
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-node"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-platform"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "description": "Example of a tailoredProfile that extends OCP4 FedRAMP Moderate",
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.17 <0.1.61'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            displayName: Compliance Remediation
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            displayName: Compliance Scan
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            displayName: Compliance Suite
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            displayName: Profile Bundle
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            displayName: Scan Setting Binding
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            displayName: Tailored Profile
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repo
          url: https://github.com/openshift/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/compliance_operator/compliance-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b
        - registry.redhat.io/compliance/openshift-compliance-openscap-rhel8@sha256:83ad3113964ec48c5cf28315d923adb6edb95e45960e2756665d617dcd9b2e47
        - registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:5578e2225365ac1c15e4c189c5c7a8d14fd3fd26189414b36028f7f0879bb1dc
        version: 0.1.61
      entries:
      - name: compliance-operator.v0.1.61
        version: 0.1.61
      name: release-0.1
    - currentCSV: compliance-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-node"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-platform"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "description": "Example of a tailoredProfile that extends OCP4 FedRAMP Moderate",
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          must-gather-image: registry.redhat.io/compliance/openshift-compliance-must-gather-rhel8@sha256:e6468ccc55d72539b7fb980d0bce5c072e85879d4e9db39974d53eaab4d32b25
          olm.skipRange: '>=0.1.17 <1.7.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ComplianceAsCode/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            displayName: Compliance Remediation
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            displayName: Compliance Scan
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            displayName: Compliance Suite
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            displayName: Profile Bundle
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            displayName: Scan Setting Binding
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            displayName: Tailored Profile
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repository
          url: https://github.com/ComplianceAsCode/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/compliance_operator/co-overview.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4
        - registry.redhat.io/compliance/openshift-compliance-openscap-rhel8@sha256:d0f89d82845c6245a18e7db345ac3be92b2863d57f55f63e21652086128920c9
        - registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:8d3257bd6a020df53f8d7a400ae877ae029c78ffcd8ec9e35270ee3ff5c8e509
        - registry.redhat.io/compliance/openshift-compliance-must-gather-rhel8@sha256:e6468ccc55d72539b7fb980d0bce5c072e85879d4e9db39974d53eaab4d32b25
        version: 1.7.1
      entries:
      - name: compliance-operator.v1.7.1
        version: 1.7.1
      - name: compliance-operator.v1.7.0
        version: 1.7.0
      - name: compliance-operator.v1.6.2
        version: 1.6.2
      - name: compliance-operator.v1.6.1
        version: 1.6.1
      - name: compliance-operator.v1.6.0
        version: 1.6.0
      - name: compliance-operator.v1.5.1
        version: 1.5.1
      - name: compliance-operator.v1.5.0
        version: 1.5.0
      - name: compliance-operator.v1.4.1
        version: 1.4.1
      - name: compliance-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: compliance-operator
    provider:
      name: Red Hat Inc.
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: hyperfoil-bundle
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hyperfoil-operator.v0.26.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hyperfoil.io/v1alpha2",
                "kind": "Hyperfoil",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "hyperfoil-operator",
                    "app.kubernetes.io/instance": "hyperfoil-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hyperfoil",
                    "app.kubernetes.io/part-of": "hyperfoil-operator"
                  },
                  "name": "hyperfoil-sample"
                },
                "spec": {
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/hyperfoil/hyperfoil-operator:0.26.0
          createdAt: "2024-06-11T17:39:20Z"
          description: Microservice-oriented distributed benchmark framework.
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Hyperfoil/hyperfoil-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hyperfoil is the Schema for the hyperfoils API
            displayName: Hyperfoil
            kind: Hyperfoil
            name: hyperfoils.hyperfoil.io
            version: v1alpha2
        description: |
          Hyperfoil is a modern benchmark framework oriented promising
          accurate results with more flexibility and distributed architecture.

          See more on [hyperfoil.io](http://hyperfoil.io).

          Hyperfoil is licensed under Apache License 2.0
        displayName: Hyperfoil
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hyperfoil
        - benchmark
        - test
        - load
        - driver
        links:
        - name: Website
          url: http://hyperfoil.io
        - name: GitHub
          url: http://github.com/Hyperfoil
        maintainers:
        - email: rvansa@redhat.com
          name: Radim Vansa
        - email: spederse@redhat.com
          name: Ståle W. Pedersen
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/hyperfoil/hyperfoil-operator:0.26.0
        version: 0.26.0
      entries:
      - name: hyperfoil-operator.v0.26.0
        version: 0.26.0
      - name: hyperfoil-operator.v0.24.2
        version: 0.24.2
      - name: hyperfoil-operator.v0.21.0
        version: 0.21.0
      - name: hyperfoil-operator.v0.15.1
        version: 0.15.1
      - name: hyperfoil-operator.v0.15.0
        version: 0.15.0
      name: alpha
    defaultChannel: alpha
    packageName: hyperfoil-bundle
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Antline-Soft
      provider-url: ""
    name: cxl-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cxl-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-05-08T08:36:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
        description: memory extend operator
        displayName: extend-community-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - memory
        - extend
        - community
        - operator
        links:
        - name: Cxl Operator
          url: https://cxl-operator.domain
        maintainers:
        - email: tomboy@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        provider:
          name: Antline-Soft
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - rasen708/cxl-operator:latest
        version: 0.0.1
      entries:
      - name: cxl-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cxl-operator
    provider:
      name: Antline-Soft
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-msg-broker
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-msg-broker.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Messagebroker",
                "metadata": {
                  "name": "messagebroker-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "chiselServer": {
                    "enabled": false,
                    "port": 8080
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": false
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "natsCore": {
                    "reloader": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats-server-config-reloader",
                      "tag": "0.8.3-3747-30"
                    },
                    "server": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats",
                      "tag": "2.9.0-3747-21"
                    }
                  },
                  "natsHttpProxy": {
                    "enabled": false,
                    "port": 8080
                  },
                  "natsStreaming": {
                    "image": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats-streaming",
                      "tag": "0.23.2-3747-31"
                    }
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "tagSuffix": "-ocp"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:31:32Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Messagebroker
            name: messagebrokers.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Message Broker Operator
        displayName: Keysight WAP Message Broker Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Msg Broker
          url: https://keysight-wap-msg-broker.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-msg-broker@sha256:4701b6e456b099697bf1e25aeb4bc8e32b1892bb0956e36d84f9f80666b3a1fc
        version: 0.1.21
      entries:
      - name: keysight-wap-msg-broker.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-msg-broker
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC
      provider-url: https://www.nec.com
    name: sanstoragecsi-operator-bundle
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sanstoragecsi-operator-bundle.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.nec.com/v1",
                "kind": "SANStorageCSI",
                "metadata": {
                  "name": "sanstoragecsi"
                },
                "spec": {
                  "driverName": "san.csi.nec.com",
                  "images": {
                    "csiDriver": "",
                    "externalAttacher": "registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:bd8262cc4b6c405e175c39623323dd98a74bbf70e6aa90851a72612056c184ff",
                    "externalProvisioner": "registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:85f1cc8f21568e7885e492b29660f29e09796fb23ebdf5fd526aa2e99d6b2fa8",
                    "livenessProbe": "registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:1a0d15012ba70955af754b03a7488b7c6ace813646ec7c3ce845f80597050c3c",
                    "nodeRegistrar": "registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:dc7f0992be6c2853dc0e798afda368e36005a94ad76b21fa2f09790621318864"
                  },
                  "parameters": {
                    "controller": {
                      "logLevel": 5
                    },
                    "node": {
                      "livenessProbePort": 9808,
                      "logLevel": 5,
                      "maxVolumesPerNode": 1024
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          createdAt: "2022-06-15T00:00:00Z"
          description: An operator for managing the NEC Storage M Series CSI Driver
          operatorframework.io/suggested-namespace: nec-san-csi-operator-system
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: NEC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SANStorageCSI is the Schema for the sanstoragecsis API
            displayName: NEC Storage M Series CSI Driver
            kind: SANStorageCSI
            name: sanstoragecsis.csi.nec.com
            version: v1
        description: |
          This operator deploys the NEC Storage M Series CSI Driver on OpenShift Container Platform.

          NEC Storage M Series CSI Driver provides persistent storage for stateful applications using NEC Storage M Series.
          Please refer to the documentation of the NEC Storage M Series CSI Driver for detail information about supported NEC Storage Platforms and supported CSI features.
        displayName: NEC Storage M Series CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - NEC
        - M Series
        - SAN Storage
        - CSI Driver
        links:
        - name: NEC Storage M Series
          url: https://www.nec.com/en/global/prod/storage/product/san/index.html
        maturity: stable
        provider:
          name: NEC
          url: https://www.nec.com
        relatedImages:
        - registry.connect.redhat.com/nec/sanstoragecsi-operator@sha256:6656c2b5036bdc013712ecb640bbbebe0e3e90cf625c73697dbb49e3e201e2ac
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:dc7f0992be6c2853dc0e798afda368e36005a94ad76b21fa2f09790621318864
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:bd8262cc4b6c405e175c39623323dd98a74bbf70e6aa90851a72612056c184ff
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:85f1cc8f21568e7885e492b29660f29e09796fb23ebdf5fd526aa2e99d6b2fa8
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:1a0d15012ba70955af754b03a7488b7c6ace813646ec7c3ce845f80597050c3c
        version: 1.1.0
      entries:
      - name: sanstoragecsi-operator-bundle.v1.1.0
        version: 1.1.0
      - name: sanstoragecsi-operator-bundle.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: sanstoragecsi-operator-bundle
    provider:
      name: NEC
      url: https://www.nec.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nxiq-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nxiq-operator-certified.v1.193.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusIQ",
                "metadata": {
                  "name": "example-nexusiq"
                },
                "spec": {
                  "deployment": {
                    "postStart": {
                      "command": null
                    }
                  },
                  "deploymentStrategy": {
                    "type": "Recreate"
                  },
                  "fullnameOverride": "",
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [],
                    "path": "/",
                    "tls": []
                  },
                  "iq": {
                    "adminPort": 8071,
                    "applicationPort": 8070,
                    "env": [
                      {
                        "name": "JAVA_OPTS",
                        "value": "-Djava.util.prefs.userRoot=$(SONATYPE_WORK)/javaprefs -Djavax.net.ssl.trustStore=/etc/nexus-iq-server/jks/keystore.jks"
                      }
                    ],
                    "configYaml": {
                      "sonatypeWork": "/sonatype-work",
                      "createSampleData": true,
                      "server": {
                        "adminConnectors": [
                          {
                            "port": 8071,
                            "type": "http"
                          }
                        ],
                        "applicationConnectors": [
                          {
                            "port": 8070,
                            "type": "http"
                          }
                        ],
                        "requestLog": {
                          "appenders": [
                            {
                              "type": "file",
                              "currentLogFilename": "/var/log/nexus-iq-server/request.log",
                              "logFormat": "%clientHost %l %user [%date] \"%requestURL\" %statusCode %bytesSent %elapsedTime \"%header{User-Agent}\"",
                              "archivedLogFilenamePattern": "/var/log/nexus-iq-server/request-%d.log.gz",
                              "archivedFileCount": 50
                            }
                          ]
                        }
                      },
                      "logging": {
                        "level": "DEBUG",
                        "loggers": {
                          "com.sonatype.insight.scan": "INFO",
                          "eu.medsea.mimeutil.MimeUtil2": "INFO",
                          "org.apache.http": "INFO",
                          "org.apache.http.wire": "ERROR",
                          "org.eclipse.birt.report.engine.layout.pdf.font.FontConfigReader": "WARN",
                          "org.eclipse.jetty": "INFO",
                          "org.apache.shiro.web.filter.authc.BasicHttpAuthenticationFilter": "INFO",
                          "com.sonatype.insight.audit": {
                            "appenders": [
                              {
                                "type": "file",
                                "currentLogFilename": "/var/log/nexus-iq-server/audit.log",
                                "archivedLogFilenamePattern": "/var/log/nexus-iq-server/audit-%d.log.gz",
                                "archivedFileCount": 50
                              }
                            ]
                          },
                          "com.sonatype.insight.policy.violation": {
                            "appenders": [
                              {
                                "type": "console",
                                "threshold": "INFO"
                              }
                            ]
                          }
                        },
                        "appenders": [
                          {
                            "type": "console",
                            "threshold": "INFO",
                            "logFormat": "%d{'yyyy-MM-dd HH:mm:ss,SSSZ'} %level [%thread] %X{username} %logger - %msg%n"
                          },
                          {
                            "type": "file",
                            "threshold": "ALL",
                            "logFormat": "%d{'yyyy-MM-dd HH:mm:ss,SSSZ'} %level [%thread] %X{username} %logger - %msg%n",
                            "currentLogFilename": "/var/log/nexus-iq-server/clm-server.log",
                            "archivedLogFilenamePattern": "/var/log/nexus-iq-server/clm-server-%d.log.gz",
                            "archivedFileCount": 50
                          }
                        ]
                      }
                    },
                    "imageName": "registry.connect.redhat.com/sonatype/nexus-iq-server@sha256:d65df6bdff1f32bf32a87e5cb38f5e9f98b75b81e9acccbffbd967e1f71691f0",
                    "imagePullPolicy": "IfNotPresent",
                    "imagePullSecret": "",
                    "licenseSecret": "",
                    "memory": "1Gi",
                    "name": "nxiq",
                    "certificates": {}
                  },
                  "nameOverride": "",
                  "persistence": {
                    "accessMode": "ReadWriteOnce",
                    "storageSize": "1Gi"
                  },
                  "service": {
                    "annotations": {},
                    "enabled": false,
                    "labels": {},
                    "ports": [
                      {
                        "name": "nexus-service",
                        "port": 80,
                        "targetPort": 80
                      }
                    ]
                  },
                  "serviceAccount": {
                    "create": true,
                    "name": null
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nxiq-operator-certified@sha256:59c52729189da87d9b5c3de8ef85326662d94c99835ab99d6e8d7fefdcf53ef0
          createdAt: "2020-07-20"
          description: |-
            Nexus Lifecycle is an open source governance platform that continuously identifies risk,
            automatically enforces policy, and provides visibility throughout the entire SDLC.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.193.0-1
          repository: https://github.com/sonatype/operator-iq
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus IQ Server
            displayName: NexusIQ
            kind: NexusIQ
            name: nexusiqs.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Lifecycle is an open source governance platform that continuously identifies risk,
          automatically enforces policy, and provides visibility throughout the entire SDLC.
          Developers leveraging open source require a solution that helps them make safer choices
          while still delivering at DevOps speed.
          Nexus Lifecycle empowers developers and security teams to choose the best components
          and govern open source usage, ensuring organizations continue to innovate with less risk.

          ## Core Capabilities

          * **Dependency Management**:
            Generate a precise software bill of materials for all applications
            to identify open source components and gather contextual direct
            and transitive dependency information for faster action.
          * **One-Click Remediation**:
            Upgrade to the best components and versions based on real-time intelligence.
            Available in your preferred IDE or a GitHub pull request.
          * **Automated Open Source Policy Enforcement**:
            Create custom security, license, and architectural policies based
            on application type or organization and contextually enforce those policies
            across every stage of the SDLC.
          * **Continuous Monitoring**:
            Real-time notifications of newly discovered defects,
            as well as vulnerabilities based on component,
            risk level or application/container affected.

          ## Extras

          * IDE Integrations: Eclipse, IntelliJ, and Visual Studio
          * Source Control Integrations: GitHub, GitLab, and Bitbucket
          * Chrome Browser Extension
          * Red Hat Clair for Container Scanning

          ## Limitations

          This operator will be released on a quarterly basis.

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus IQ Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage

          Once the server instance is created by the operator and running,
          you'll want to expose the service as you see fit:
          1. Create a Route to the new service for iq.applicationPort (8070).
          2. Visit the URL provided by the Route, login, and set new credentials.
            The default credentials are `admin`/`admin123`.

          The Nexus IQ Server can be further configured via the NexusIQ custom resource definition:

          | Parameter            | Description                                                  | Default           |
          | -------------------- | ------------------------------------------------------------ | ----------------- |
          | `iq.applicationPort` | Port of the application connector. Must match the value in the `configYaml` property | `8070`            |
          | `iq.adminPort`       | Port of the application connector. Must match the value in the `configYaml` property | `8071`            |
          | `iq.memory`          | The amount of RAM to allocate                                | `1Gi`             |
          | `iq.env`             | Customize the server environment, including `JAVA_OPTS`      | See example YAML show when ceating a NexusIQ. |
          | `iq.licenseSecret`   | The base-64 encoded license file to be installed at startup  | `""`              |
          | `iq.configYaml`      | A YAML block which will be used as a configuration block for IQ Server. | See example YAML shown when creating a NexusIQ. |
          | `iq.certificates`   | Add your own certificates for your own trusted Certificate Authorities with names and plain armored text. | `{}` |
          | `ingress.enabled`                           | Create an ingress for Nexus         | `true`                                  |
          | `ingress.annotations`                       | Annotations to enhance ingress configuration  | `{}`                          |
          | `ingress.tls.enabled`                       | Enable TLS                          | `true`                                 |
          | `ingress.tls.secretName`                    | Name of the secret storing TLS cert, `false` to use the Ingress' default certificate | `nexus-tls`                             |
          | `ingress.path`                              | Path for ingress rules. GCP users should set to `/*` | `/`                    |
          | `deploymentStrategy.type`                   | Deployment Strategy                 |  `Recreate` |
          | `initContainers`                            | Containers to run before starting the server | `nil` |
          | `deployment.postStart.command`              | Command to run after starting the IQ Server container  | `nil`                    |
          | `deployment.terminationGracePeriodSeconds` | Time to allow for clean shutdown                        | 120                      |
          | `persistence.storageClass` | The provisioner class                        | `-` (disables dynamic provisioning)            |
          | `persistence.storageSize` | The amount of drive space to allocate                        | `1Gi`             |
          | `persistence.accessMode` | Default access mode                        | `ReadWriteOnce`             |
          | `persistence.volumeConfiguration` | A YAML block to configure the persistent volume type. Defaults to `hostPath` which should not be used in production | `hostPath`             |

          ## Configuring IQ Server

          You can define the `config.yml` for IQ Server in your CRD on startup.
          It is the `iq.configYaml` property. For more details, see the [Configuring IQ Server](https://help.sonatype.com/iqserver/configuring) help page.

          ## Installing the License

          The license file can be installed via the UI when IQ server is running, or it can be done as a part of the deploy.
          If you leave the `licenseFile` field empty/commented, IQ Server will start and prompt you to manually install the license
          when you first enter the GUI.

          ### Installing the License Automatically
          To do it automatically, first encode your `.lic` file in Base 64 with no line breaks, eg:

          ```bash
          base64 --wrap=0 mylicense.lic > lic.base64
          ```

          Then add this value to your CRD file as `iq.licenseSecret`, eg:

          ```yaml
          iq:
            licenseSecret: bXkgc2FtcGxlIGxpY2Vuc2U=
          ```

          Specify the `licenseFile` path in `iq.configYaml` as:

          ```yaml
          iq:
            configYaml:
              sonatypeWork: /sonatype-work
              createSampleData: true
              # add this line and the `licenseSecret` above to autoconfigure licensing
              licenseFile: /etc/nexus-iq-license/license_lic
              server:
                applicationConnectors:
                  - type: http
                    port: 8070
                adminConnectors:
                  - type: http
                    port: 8071
          ```

          ## Adding Certificates for Your Own Trusted Certificate Authorities

          Specify a set of certificates to be imported into the Java trust store
          of the running IQ Server. More than one certificate can be provided.

          ```yaml
          iq:
            certificates:
              my-company-cert: |
                -----BEGIN CERTIFICATE-----
                ...
                -----END CERTIFICATE-----
          ```

          ## Running Commands Before Initialization of the Server

          If you need custom initialization, you can use standard Kubernetes
          [`initContainers`](https://kubernetes.io/docs/concepts/workloads/pods/init-containers/)
          to run any series of containers and commands you need.

          ```yaml
          initContainers:
            - name: an-init-container
              image: busybox:1.28
              command: ['/bin/sh', '-c', 'echo "initializing"']
          ```
        displayName: Nexus IQ Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - sonatype
        - licensing
        - policy
        - components
        links:
        - name: Nexus Lifecycle
          url: https://www.sonatype.com/product-nexus-lifecycle
        maintainers:
        - email: support@sonaype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nxiq-operator-certified@sha256:59c52729189da87d9b5c3de8ef85326662d94c99835ab99d6e8d7fefdcf53ef0
        - registry.connect.redhat.com/sonatype/nexus-iq-server@sha256:d65df6bdff1f32bf32a87e5cb38f5e9f98b75b81e9acccbffbd967e1f71691f0
        version: 1.193.0-1
      entries:
      - name: nxiq-operator-certified.v1.193.0-1
        version: 1.193.0-1
      name: stable
    defaultChannel: stable
    packageName: nxiq-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com/
    name: openstack-baremetal-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openstack-baremetal-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "baremetal.openstack.org/v1beta1",
                "kind": "OpenStackBaremetalSet",
                "metadata": {
                  "name": "openstackbaremetalset"
                },
                "spec": {
                  "baremetalHosts": {
                    "compute-0": "172.22.0.100/24",
                    "compute-1": "172.22.0.101/24"
                  },
                  "ctlplaneGateway": "172.22.0.3",
                  "ctlplaneInterface": "enp1s0",
                  "deploymentSSHSecret": "baremetalset-deploy-secret",
                  "domainName": "osptest.openstack.org",
                  "osImage": "edpm-hardened-uefi.qcow2",
                  "passwordSecret": "baremetalset-password-secret"
                }
              },
              {
                "apiVersion": "baremetal.openstack.org/v1beta1",
                "kind": "OpenStackProvisionServer",
                "metadata": {
                  "name": "openstackprovisionserver"
                },
                "spec": {
                  "interface": "enp1s0",
                  "osImage": "edpm-hardened-uefi.qcow2",
                  "port": 8080
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:07Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenStackBaremetalSet is the Schema for the openstackbaremetalsets
              API
            displayName: OpenStack BaremetalSet
            kind: OpenStackBaremetalSet
            name: openstackbaremetalsets.baremetal.openstack.org
            version: v1beta1
          - description: OpenStackProvisionServer used to serve custom images for
              baremetal provisioning with Metal3
            displayName: OpenStackProvisionServer
            kind: OpenStackProvisionServer
            name: openstackprovisionservers.baremetal.openstack.org
            version: v1beta1
        description: An operator for handling baremetal provisioning concerns for
          OpenStack
        displayName: OpenStack Baremetal Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cn-openstack
        - openstack
        links:
        - name: Openstack Baremetal Operator
          url: https://github.com/openstack-k8s-operators/openstack-baremetal-operator
        maturity: beta
        provider:
          name: Red Hat
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/ubi9/httpd-24@sha256:dbfbe3383e12d63df7178e4cdd5b1e6e51a84c781041b9192026ea6edb72fbf3
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/openstack-baremetal-rhel9-operator@sha256:a6c69c03e59faf50973d07ccea32b80cb9c46ceb60ed00ffe7311ca42d501296
        - registry.redhat.io/rhoso/edpm-hardened-uefi-rhel9@sha256:e4bfacb9d8de056a251bee16a66b69631b96fb19fb926d8e33acd379d40a92d6
        - registry.redhat.io/rhoso-operators/openstack-baremetal-agent-rhel9@sha256:8f6f8e99042710819aa639eae6513f36c4642190ecd85d58a2ed819f1b55e38f
        version: 1.0.6
      entries:
      - name: openstack-baremetal-operator.v1.0.6
        version: 1.0.6
      - name: openstack-baremetal-operator.v1.0.5
        version: 1.0.5
      - name: openstack-baremetal-operator.v1.0.4
        version: 1.0.4
      - name: openstack-baremetal-operator.v1.0.3
        version: 1.0.3
      - name: openstack-baremetal-operator.v1.0.2
        version: 1.0.2
      - name: openstack-baremetal-operator.v1.0.1
        version: 1.0.1
      - name: openstack-baremetal-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: openstack-baremetal-operator
    provider:
      name: Red Hat
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecr-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecr-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.services.k8s.aws/v1alpha1",
                "kind": "Repository",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecr-controller:1.3.1
          createdAt: "2025-09-20T01:21:24Z"
          description: AWS ECR controller is a service controller for managing ECR
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PullThroughCacheRule represents the state of an AWS ecr PullThroughCacheRule
              resource.
            displayName: PullThroughCacheRule
            kind: PullThroughCacheRule
            name: pullthroughcacherules.ecr.services.k8s.aws
            version: v1alpha1
          - description: Repository represents the state of an AWS ecr Repository
              resource.
            displayName: Repository
            kind: Repository
            name: repositories.ecr.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Container Registry (ECR) resources in AWS from within your Kubernetes cluster.

          **About Amazon ECR**

          Amazon Elastic Container Registry (Amazon ECR) is an AWS managed container image registry service that is secure, scalable, and reliable. Amazon ECR supports private repositories with resource-based permissions using AWS IAM. This is so that specified users or Amazon EC2 instances can access your container repositories and images. You can use your preferred CLI to push, pull, and manage Docker images, Open Container Initiative (OCI) images, and OCI compatible artifacts.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECR
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECR Developer Resources
          url: https://aws.amazon.com/ecr/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecr maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecr-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-ecr-controller.v1.3.1
        version: 1.3.1
      - name: ack-ecr-controller.v1.3.0
        version: 1.3.0
      - name: ack-ecr-controller.v1.2.0
        version: 1.2.0
      - name: ack-ecr-controller.v1.1.0
        version: 1.1.0
      - name: ack-ecr-controller.v1.0.33
        version: 1.0.33
      - name: ack-ecr-controller.v1.0.32
        version: 1.0.32
      - name: ack-ecr-controller.v1.0.31
        version: 1.0.31
      - name: ack-ecr-controller.v1.0.30
        version: 1.0.30
      - name: ack-ecr-controller.v1.0.29
        version: 1.0.29
      - name: ack-ecr-controller.v1.0.28
        version: 1.0.28
      - name: ack-ecr-controller.v1.0.27
        version: 1.0.27
      - name: ack-ecr-controller.v1.0.25
        version: 1.0.25
      - name: ack-ecr-controller.v1.0.24
        version: 1.0.24
      - name: ack-ecr-controller.v1.0.23
        version: 1.0.23
      - name: ack-ecr-controller.v1.0.22
        version: 1.0.22
      - name: ack-ecr-controller.v1.0.21
        version: 1.0.21
      - name: ack-ecr-controller.v1.0.20
        version: 1.0.20
      - name: ack-ecr-controller.v1.0.19
        version: 1.0.19
      - name: ack-ecr-controller.v1.0.17
        version: 1.0.17
      - name: ack-ecr-controller.v1.0.16
        version: 1.0.16
      - name: ack-ecr-controller.v1.0.15
        version: 1.0.15
      - name: ack-ecr-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecr-controller.v1.0.12
        version: 1.0.12
      - name: ack-ecr-controller.v1.0.10
        version: 1.0.10
      - name: ack-ecr-controller.v1.0.9
        version: 1.0.9
      - name: ack-ecr-controller.v1.0.8
        version: 1.0.8
      - name: ack-ecr-controller.v1.0.7
        version: 1.0.7
      - name: ack-ecr-controller.v1.0.6
        version: 1.0.6
      - name: ack-ecr-controller.v1.0.4
        version: 1.0.4
      - name: ack-ecr-controller.v1.0.3
        version: 1.0.3
      - name: ack-ecr-controller.v1.0.2
        version: 1.0.2
      - name: ack-ecr-controller.v1.0.0
        version: 1.0.0
      - name: ack-ecr-controller.v0.1.9
        version: 0.1.9
      - name: ack-ecr-controller.v0.1.7
        version: 0.1.7
      - name: ack-ecr-controller.v0.1.6
        version: 0.1.6
      - name: ack-ecr-controller.v0.1.5
        version: 0.1.5
      - name: ack-ecr-controller.v0.1.4
        version: 0.1.4
      - name: ack-ecr-controller.v0.1.3
        version: 0.1.3
      - name: ack-ecr-controller.v0.1.2
        version: 0.1.2
      - name: ack-ecr-controller.v0.1.1
        version: 0.1.1
      - name: ack-ecr-controller.v0.1.0
        version: 0.1.0
      - name: ack-ecr-controller.v0.0.27
        version: 0.0.27
      - name: ack-ecr-controller.v0.0.26
        version: 0.0.26
      - name: ack-ecr-controller.v0.0.25
        version: 0.0.25
      - name: ack-ecr-controller.v0.0.24
        version: 0.0.24
      - name: ack-ecr-controller.v0.0.23
        version: 0.0.23
      - name: ack-ecr-controller.v0.0.22
        version: 0.0.22
      - name: ack-ecr-controller.v0.0.21
        version: 0.0.21
      - name: ack-ecr-controller.v0.0.20
        version: 0.0.20
      - name: ack-ecr-controller.v0.0.19
        version: 0.0.19
      - name: ack-ecr-controller.v0.0.18
        version: 0.0.18
      - name: ack-ecr-controller.v0.0.17
        version: 0.0.17
      - name: ack-ecr-controller.v0.0.16
        version: 0.0.16
      - name: ack-ecr-controller.v0.0.15
        version: 0.0.15
      - name: ack-ecr-controller.v0.0.14
        version: 0.0.14
      - name: ack-ecr-controller.v0.0.13
        version: 0.0.13
      - name: ack-ecr-controller.v0.0.12
        version: 0.0.12
      - name: ack-ecr-controller.v0.0.11
        version: 0.0.11
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecr-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: VictoriaMetrics
      provider-url: ""
    name: victoriametrics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: victoriametrics-operator.v0.63.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLAgent",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "persistentVolumeClaimRetentionPolicy": {
                    "whenDeleted": "Delete"
                  },
                  "remoteWrite": [
                    {
                      "maxDiskUsage": "10GB",
                      "url": "http://vlsingle-example-0.default.svc:9428/internal/insert"
                    }
                  ],
                  "remoteWriteSettings": {
                    "maxBlockSize": "30MB"
                  },
                  "replicaCount": 2,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "350Mi"
                    }
                  },
                  "storage": {
                    "volumeClaimTemplate": {
                      "spec": {
                        "resources": {
                          "requests": {
                            "storage": "50Gi"
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "victoriametrics-operator"
                  },
                  "name": "example"
                },
                "spec": {
                  "requestsLoadBalancer": {
                    "enabled": true,
                    "spec": {}
                  },
                  "vlinsert": {
                    "replicaCount": 1
                  },
                  "vlselect": {
                    "replicaCount": 1
                  },
                  "vlstorage": {
                    "replicaCount": 2,
                    "retentionPeriod": "1y",
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "50Gi"
                            }
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLSingle",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "victoriametrics-operator"
                  },
                  "name": "sample"
                },
                "spec": {
                  "retentionMaxDiskSpaceUsageBytes": "20GB",
                  "retentionPeriod": "1y",
                  "storage": {
                    "resources": {
                      "requests": {
                        "storage": "50Gi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VLogs",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "retentionPeriod": "2"
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAgent",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "extraArgs": {
                    "memory.allowedPercent": "40"
                  },
                  "remoteWrite": [
                    {
                      "maxDiskUsage": "10GB",
                      "url": "http://vmsingle-dc-1.default.svc:8429/api/v1/write"
                    },
                    {
                      "maxDiskUsage": "15GB",
                      "url": "http://vmsingle-dc-2.default.svc:8429/api/v1/write"
                    },
                    {
                      "url": "http://vminsert-main.default.svc:8480/insert/1:1/prometheus/api/v1/write"
                    }
                  ],
                  "remoteWriteSettings": {
                    "maxDiskUsagePerURL": "5GB"
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "350Mi"
                    }
                  },
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlert",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "datasource": {
                    "url": "http://vmsingle-example-pvc.default.svc:8429"
                  },
                  "evaluationInterval": "30s",
                  "notifier": {
                    "url": "http://vmalertmanager-example.default.svc:9093"
                  },
                  "replicaCount": 1,
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlertmanager",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "configSecret": "am-global-config",
                  "replicaCount": 1,
                  "selectAllByDefault": true,
                  "storage": {
                    "volumeClaimTemplate": {
                      "spec": {
                        "resources": {
                          "requests": {
                            "storage": "2Gi"
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlertmanagerConfig",
                "metadata": {
                  "name": "example-email-web"
                },
                "spec": {
                  "receivers": [
                    {
                      "email_configs": [
                        {
                          "from": "alerting@example.com",
                          "smarthost": "example.com:25",
                          "text": "ALARM",
                          "to": "some-email@example.com"
                        }
                      ],
                      "name": "email",
                      "webhook_configs": [
                        {
                          "url": "http://some-other-wh"
                        }
                      ]
                    }
                  ],
                  "route": {
                    "group_interval": "1m",
                    "receiver": "email",
                    "routes": [
                      {
                        "receiver": "email"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAuth",
                "metadata": {
                  "name": "vmauth-sample"
                },
                "spec": {
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "victoriametrics/vmauth"
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "250m",
                      "memory": "350Mi"
                    }
                  },
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMCluster",
                "metadata": {
                  "name": "example-persistent"
                },
                "spec": {
                  "replicationFactor": 2,
                  "retentionPeriod": "4",
                  "vminsert": {
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "0.5",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "vmselect": {
                    "cacheMountPath": "/select-cache",
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "0.5",
                        "memory": "500Mi"
                      }
                    },
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "2Gi"
                            }
                          }
                        }
                      }
                    }
                  },
                  "vmstorage": {
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1500Mi"
                      }
                    },
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "10Gi"
                            }
                          }
                        }
                      }
                    },
                    "storageDataPath": "/vm-data"
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMNodeScrape",
                "metadata": {
                  "name": "cadvisor-metrics"
                },
                "spec": {
                  "bearerTokenFile": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                  "interval": "10s",
                  "relabelConfigs": [
                    {
                      "action": "labelmap",
                      "regex": "__meta_kubernetes_node_label_(.+)"
                    },
                    {
                      "replacement": "kubernetes.default.svc:443",
                      "targetLabel": "__address__"
                    },
                    {
                      "regex": "(.+)",
                      "replacement": "/api/v1/nodes/$1/proxy/metrics/cadvisor",
                      "sourceLabels": [
                        "__meta_kubernetes_node_name"
                      ],
                      "targetLabel": "__metrics_path__"
                    }
                  ],
                  "scheme": "https",
                  "scrapeTimeout": "2s",
                  "selector": {},
                  "tlsConfig": {
                    "caFile": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                    "insecureSkipVerify": true
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMPodScrape",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "podMetricsEndpoints": [
                    {
                      "port": "metrics",
                      "scheme": "http"
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app.kubernetes.io/name": "example-application",
                      "monitored-by": "vm-operator"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMProbe",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "interval": "2s",
                  "jobName": "static-probe",
                  "module": "http_2xx",
                  "targets": {
                    "staticConfig": {
                      "targets": [
                        "vmagent-example.default.svc:8429/health"
                      ]
                    }
                  },
                  "vmProberSpec": {
                    "url": "prometheus-blackbox-exporter.default.svc:9115"
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMRule",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "groups": [
                    {
                      "name": "kafka",
                      "rules": [
                        {
                          "alert": "coordinator down",
                          "annotations": {
                            "description": "kafka coordinator is down",
                            "value": "{{ $value }}"
                          },
                          "expr": "ml_app_gauge{exec_context=\"consumer_group_state\"} == 0",
                          "for": "60s",
                          "labels": {
                            "job": "{{ $labels.job }}",
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMScrapeConfig",
                "metadata": {
                  "name": "mongodb"
                },
                "spec": {
                  "consulSDConfigs": [
                    {
                      "server": "https://consul-dns:8500",
                      "services": [
                        "mongodb"
                      ]
                    }
                  ],
                  "relabelConfigs": [
                    {
                      "action": "replace",
                      "sourceLabels": [
                        "__meta_consul_service"
                      ],
                      "targetLabel": "job"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMServiceScrape",
                "metadata": {
                  "name": "example-scrape"
                },
                "spec": {
                  "endpoints": [
                    {
                      "port": "http"
                    }
                  ],
                  "namespaceSelector": {},
                  "selector": {
                    "matchLabels": {
                      "app.kubernetes.io/name": "example-application",
                      "monitored-by": "vm-operator"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMSingle",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "retentionPeriod": "1"
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMSingle",
                "metadata": {
                  "name": "example-pvc"
                },
                "spec": {
                  "removePvcAfterDelete": true,
                  "retentionPeriod": "1",
                  "storage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMStaticScrape",
                "metadata": {
                  "name": "vmstaticscrape-sample"
                },
                "spec": {
                  "jobName": "static",
                  "targetEndpoints": [
                    {
                      "labels": {
                        "env": "dev",
                        "project": "operator"
                      },
                      "targets": [
                        "192.168.0.1:9100",
                        "196.168.0.50:9100"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMUser",
                "metadata": {
                  "name": "vmuser-tenant-1"
                },
                "spec": {
                  "bearerToken": "some-token",
                  "targetRefs": [
                    {
                      "crd": {
                        "kind": "VMCluster/vminsert",
                        "name": "test-persistent",
                        "namespace": "vm"
                      },
                      "target_path_suffix": "/insert/1"
                    },
                    {
                      "crd": {
                        "kind": "VMCluster/vmselect",
                        "name": "test-persistent",
                        "namespace": "vm"
                      },
                      "target_path_suffix": "/select/1"
                    },
                    {
                      "paths": [
                        "/internal/resetRollupResultCache"
                      ],
                      "static": {
                        "url": "http://vmselect-test-persistent.default.svc:8481/"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: docker.io/victoriametrics/operator:v0.63.0
          createdAt: "2025-09-12T12:58:29Z"
          description: Provides monitoring capabilities for kubernetes clusters and
            applications
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: "1.30"
          operators.openshift.io/infrastructure-features: '[fips]'
          operators.operatorframework.io.bundle.channel.default.v1: beta
          operators.operatorframework.io.bundle.channels.v1: beta
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/VictoriaMetrics/operator
          support: VictoriaMetrics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VLAgent - is a tiny but brave agent, which helps you collect
              logs from various sources and stores them in VictoriaLogs.
            displayName: VLAgent
            kind: VLAgent
            name: vlagents.operator.victoriametrics.com
            version: v1
          - description: VLCluster is fast, cost-effective and scalable logs database.
            displayName: VLCluster
            kind: VLCluster
            name: vlclusters.operator.victoriametrics.com
            version: v1
          - description: |-
              VLogs is fast, cost-effective and scalable logs database.
              VLogs is the Schema for the vlogs API
            displayName: VLogs
            kind: VLogs
            name: vlogs.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VLSingle is fast, cost-effective and scalable logs database.
              VLSingle is the Schema for the API
            displayName: VLSingle
            kind: VLSingle
            name: vlsingles.operator.victoriametrics.com
            version: v1
          - description: |-
              VMAgent - is a tiny but brave agent, which helps you collect metrics from various sources and stores them in VictoriaMetrics
              or any other Prometheus-compatible storage system that supports the remote_write protocol.
            displayName: VMAgent
            kind: VMAgent
            name: vmagents.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlertmanagerConfig is the Schema for the vmalertmanagerconfigs
              API
            displayName: VMAlertmanager Config
            kind: VMAlertmanagerConfig
            name: vmalertmanagerconfigs.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlertmanager represents Victoria-Metrics deployment for
              Alertmanager.
            displayName: VMAlertmanager
            kind: VMAlertmanager
            name: vmalertmanagers.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlert  executes a list of given alerting or recording rules
              against configured address.
            displayName: VMAlert
            kind: VMAlert
            name: vmalerts.operator.victoriametrics.com
            version: v1beta1
          - description: VMAnomaly is the Schema for the vmanomalies API.
            displayName: VMAnomaly
            kind: VMAnomaly
            name: vmanomalies.operator.victoriametrics.com
            version: v1
          - description: VMAuth is the Schema for the vmauths API
            displayName: VMAuth
            kind: VMAuth
            name: vmauths.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMCluster is fast, cost-effective and scalable time-series database.
              Cluster version with
            displayName: VMCluster
            kind: VMCluster
            name: vmclusters.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMNodeScrape defines discovery for targets placed on kubernetes nodes,
              usually its node-exporters and other host services.
              InternalIP is used as __address__ for scraping.
            displayName: VMNode Scrape
            kind: VMNodeScrape
            name: vmnodescrapes.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMPodScrape is scrape configuration for pods,
              it generates vmagent's config for scraping pod targets
              based on selectors.
            displayName: VMPod Scrape
            kind: VMPodScrape
            name: vmpodscrapes.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMProbe defines a probe for targets, that will be executed with prober,
              like blackbox exporter.
              It helps to monitor reachability of target with various checks.
            displayName: VMProbe
            kind: VMProbe
            name: vmprobes.operator.victoriametrics.com
            version: v1beta1
          - description: VMRule defines rule records for vmalert application
            displayName: VMRule
            kind: VMRule
            name: vmrules.operator.victoriametrics.com
            version: v1beta1
          - description: VMScrapeConfig specifies a set of targets and parameters
              describing how to scrape them.
            displayName: VMScrape Config
            kind: VMScrapeConfig
            name: vmscrapeconfigs.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMServiceScrape is scrape configuration for endpoints associated with
              kubernetes service,
              it generates scrape configuration for vmagent based on selectors.
              result config will scrape service endpoints
            displayName: VMService Scrape
            kind: VMServiceScrape
            name: vmservicescrapes.operator.victoriametrics.com
            version: v1beta1
          - description: VMSingle  is fast, cost-effective and scalable time-series
              database.
            displayName: VMSingle
            kind: VMSingle
            name: vmsingles.operator.victoriametrics.com
            version: v1beta1
          - description: VMStaticScrape  defines static targets configuration for
              scraping.
            displayName: VMStatic Scrape
            kind: VMStaticScrape
            name: vmstaticscrapes.operator.victoriametrics.com
            version: v1beta1
          - description: VMUser is the Schema for the vmusers API
            displayName: VMUser
            kind: VMUser
            name: vmusers.operator.victoriametrics.com
            version: v1beta1
          - description: VTCluster is fast, cost-effective and scalable traces database.
            displayName: VTCluster
            kind: VTCluster
            name: vtclusters.operator.victoriametrics.com
            version: v1
          - description: |-
              VTSingle is fast, cost-effective and scalable traces database.
              VTSingle is the Schema for the API
            displayName: VTSingle
            kind: VTSingle
            name: vtsingles.operator.victoriametrics.com
            version: v1
        description: |
          Operator manages VictoriaMetrics applications and provides monitoring features for applications running inside and outside kubernetes cluster. It has support for prometheus-operator objects and
          provides migration mechanism.

          Operator designed to use existing installation of VictoriaMetrics inside or outside kubernetes cluster or it could
          bring own resources and fully manages them.

           You can read more about operator at docs:
            - quick start [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/quick-start.md)
            - high availability [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/high-availability.md)
            - design and description of implementation [design](https://github.com/VictoriaMetrics/operator/blob/master/docs/design.md)
            - operator objects description [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/api.md)

          # VictoriaMetrics
          * VictoriaMetrics can be used as long-term storage for Prometheus or for [vmagent](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmagent/README.md).
          See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#prometheus-setup) for details.
          * Supports [Prometheus querying API](https://prometheus.io/docs/prometheus/latest/querying/api/), so it can be used as Prometheus drop-in replacement in Grafana.
          VictoriaMetrics implements [MetricsQL](https://github.com/VictoriaMetrics/VictoriaMetrics/wiki/MetricsQL) query language, which is inspired by PromQL.
          * Supports global query view. Multiple Prometheus instances may write data into VictoriaMetrics. Later this data may be used in a single query.
          * High performance and good scalability for both [inserts](https://medium.com/@valyala/high-cardinality-tsdb-benchmarks-victoriametrics-vs-timescaledb-vs-influxdb-13e6ee64dd6b)
          and [selects](https://medium.com/@valyala/when-size-matters-benchmarking-victoriametrics-vs-timescale-and-influxdb-6035811952d4).
          [Outperforms InfluxDB and TimescaleDB by up to 20x](https://medium.com/@valyala/measuring-vertical-scalability-for-time-series-databases-in-google-cloud-92550d78d8ae).
          * [Uses 10x less RAM than InfluxDB](https://medium.com/@valyala/insert-benchmarks-with-inch-influxdb-vs-victoriametrics-e31a41ae2893) when working with millions of unique time series (aka high cardinality).
          * Optimized for time series with high churn rate. Think about [prometheus-operator](https://github.com/coreos/prometheus-operator) metrics from frequent deployments in Kubernetes.
          * High data compression, so [up to 70x more data points](https://medium.com/@valyala/when-size-matters-benchmarking-victoriametrics-vs-timescale-and-influxdb-6035811952d4)
          may be crammed into limited storage comparing to TimescaleDB.
          * Optimized for storage with high-latency IO and low IOPS (HDD and network storage in AWS, Google Cloud, Microsoft Azure, etc). See [graphs from these benchmarks](https://medium.com/@valyala/high-cardinality-tsdb-benchmarks-victoriametrics-vs-timescaledb-vs-influxdb-13e6ee64dd6b).
          * A single-node VictoriaMetrics may substitute moderately sized clusters built with competing solutions such as Thanos, M3DB, Cortex, InfluxDB or TimescaleDB.
          See [vertical scalability benchmarks](https://medium.com/@valyala/measuring-vertical-scalability-for-time-series-databases-in-google-cloud-92550d78d8ae),
          [comparing Thanos to VictoriaMetrics cluster](https://medium.com/@valyala/comparing-thanos-to-victoriametrics-cluster-b193bea1683)
          and [Remote Write Storage Wars](https://promcon.io/2019-munich/talks/remote-write-storage-wars/) talk
          from [PromCon 2019](https://promcon.io/2019-munich/talks/remote-write-storage-wars/).
          * Easy operation:
              * VictoriaMetrics consists of a single [small executable](https://medium.com/@valyala/stripping-dependency-bloat-in-victoriametrics-docker-image-983fb5912b0d) without external dependencies.
              * All the configuration is done via explicit command-line flags with reasonable defaults.
              * All the data is stored in a single directory pointed by `-storageDataPath` flag.
              * Easy and fast backups from [instant snapshots](https://medium.com/@valyala/how-victoriametrics-makes-instant-snapshots-for-multi-terabyte-time-series-data-e1f3fb0e0282)
              to S3 or GCS with [vmbackup](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmbackup/README.md) / [vmrestore](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmrestore/README.md).
              See [this article](https://medium.com/@valyala/speeding-up-backups-for-big-time-series-databases-533c1a927883) for more details.
          * Storage is protected from corruption on unclean shutdown (i.e. OOM, hardware reset or `kill -9`) thanks to [the storage architecture](https://medium.com/@valyala/how-victoriametrics-makes-instant-snapshots-for-multi-terabyte-time-series-data-e1f3fb0e0282).
          * Supports metrics' scraping, ingestion and [backfilling](https://github.com/VictoriaMetrics/VictoriaMetrics#backfilling) via the following protocols:
            * [Metrics from Prometheus exporters](https://github.com/prometheus/docs/blob/master/content/docs/instrumenting/exposition_formats.md#text-based-format)
              such as [node_exporter](https://github.com/prometheus/node_exporter). See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-scrape-prometheus-exporters-such-as-node-exporter) for details.
              * [Prometheus remote write API](https://prometheus.io/docs/prometheus/latest/configuration/configuration/#remote_write)
              * [InfluxDB line protocol](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-send-data-from-influxdb-compatible-agents-such-as-telegraf) over HTTP, TCP and UDP.
              * [Graphite plaintext protocol](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-send-data-from-graphite-compatible-agents-such-as-statsd) with [tags](https://graphite.readthedocs.io/en/latest/tags.html#carbon)
              if `-graphiteListenAddr` is set.
              * [OpenTSDB put message](https://github.com/VictoriaMetrics/VictoriaMetrics#sending-data-via-telnet-put-protocol) if `-opentsdbListenAddr` is set.
              * [HTTP OpenTSDB /api/put requests](https://github.com/VictoriaMetrics/VictoriaMetrics#sending-opentsdb-data-via-http-apiput-requests) if `-opentsdbHTTPListenAddr` is set.
              * [How to import time series data](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-time-series-data).
              * [Prometheus exposition format](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-data-in-prometheus-exposition-format).
              * [Arbitrary CSV data](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-csv-data).
          * Supports metrics' relabeling. See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#relabeling) for details.
          * Ideally works with big amounts of time series data from Kubernetes, IoT sensors, connected cars, industrial telemetry, financial data and various Enterprise workloads.
        displayName: VictoriaMetrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - prometheus
        - victoriametrics
        links:
        - name: VictoriaMetrics
          url: https://victoriametrics.com
        maintainers:
        - email: nik@victoriametrics.com
          name: f41gh7
        - email: info@victoriametrics.com
          name: info
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: VictoriaMetrics
        relatedImages:
        - docker.io/victoriametrics/operator:v0.63.0
        version: 0.63.0
      entries:
      - name: victoriametrics-operator.v0.63.0
        version: 0.63.0
      - name: victoriametrics-operator.v0.62.0
        version: 0.62.0
      - name: victoriametrics-operator.v0.61.2
        version: 0.61.2
      - name: victoriametrics-operator.v0.61.1
        version: 0.61.1
      - name: victoriametrics-operator.v0.61.0
        version: 0.61.0
      - name: victoriametrics-operator.v0.60.2
        version: 0.60.2
      - name: victoriametrics-operator.v0.60.1
        version: 0.60.1
      - name: victoriametrics-operator.v0.60.0
        version: 0.60.0
      - name: victoriametrics-operator.v0.59.2
        version: 0.59.2
      - name: victoriametrics-operator.v0.59.1
        version: 0.59.1
      - name: victoriametrics-operator.v0.59.0
        version: 0.59.0
      - name: victoriametrics-operator.v0.58.1
        version: 0.58.1
      - name: victoriametrics-operator.v0.58.0
        version: 0.58.0
      - name: victoriametrics-operator.v0.57.0
        version: 0.57.0
      - name: victoriametrics-operator.v0.56.0
        version: 0.56.0
      - name: victoriametrics-operator.v0.55.0
        version: 0.55.0
      - name: victoriametrics-operator.v0.54.1
        version: 0.54.1
      - name: victoriametrics-operator.v0.53.0
        version: 0.53.0
      - name: victoriametrics-operator.v0.52.0
        version: 0.52.0
      - name: victoriametrics-operator.v0.51.1
        version: 0.51.1
      - name: victoriametrics-operator.v0.49.0
        version: 0.49.0
      - name: victoriametrics-operator.v0.48.4
        version: 0.48.4
      - name: victoriametrics-operator.v0.48.3
        version: 0.48.3
      - name: victoriametrics-operator.v0.48.1
        version: 0.48.1
      - name: victoriametrics-operator.v0.47.3
        version: 0.47.3
      - name: victoriametrics-operator.v0.47.1
        version: 0.47.1
      - name: victoriametrics-operator.v0.47.0
        version: 0.47.0
      - name: victoriametrics-operator.v0.46.4
        version: 0.46.4
      - name: victoriametrics-operator.v0.46.3
        version: 0.46.3
      - name: victoriametrics-operator.v0.39.4
        version: 0.39.4
      - name: victoriametrics-operator.v0.33.0
        version: 0.33.0
      - name: victoriametrics-operator.v0.32.1
        version: 0.32.1
      - name: victoriametrics-operator.v0.32.0
        version: 0.32.0
      - name: victoriametrics-operator.v0.29.1
        version: 0.29.1
      - name: victoriametrics-operator.v0.27.2
        version: 0.27.2
      - name: victoriametrics-operator.v0.26.2
        version: 0.26.2
      - name: victoriametrics-operator.v0.26.0
        version: 0.26.0
      - name: victoriametrics-operator.v0.25.1
        version: 0.25.1
      - name: victoriametrics-operator.v0.25.0
        version: 0.25.0
      - name: victoriametrics-operator.v0.24.0
        version: 0.24.0
      - name: victoriametrics-operator.v0.23.3
        version: 0.23.3
      - name: victoriametrics-operator.v0.22.1
        version: 0.22.1
      - name: victoriametrics-operator.v0.22.0
        version: 0.22.0
      - name: victoriametrics-operator.v0.21.0
        version: 0.21.0
      - name: victoriametrics-operator.v0.20.1
        version: 0.20.1
      - name: victoriametrics-operator.v0.19.1
        version: 0.19.1
      - name: victoriametrics-operator.v0.19.0
        version: 0.19.0
      - name: victoriametrics-operator.v0.18.2
        version: 0.18.2
      - name: victoriametrics-operator.v0.18.1
        version: 0.18.1
      - name: victoriametrics-operator.v0.18.0
        version: 0.18.0
      name: beta
    defaultChannel: beta
    packageName: victoriametrics-operator
    provider:
      name: VictoriaMetrics
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: F5 Networks Inc.
      provider-url: ""
    name: f5-bigip-ctlr-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: alpha
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: beta
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: stable
    defaultChannel: stable
    packageName: f5-bigip-ctlr-operator
    provider:
      name: F5 Networks Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dns-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dns-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSHealthCheckProbe",
                "metadata": {
                  "name": "$NAME"
                },
                "spec": {
                  "additionalHeadersRef": {
                    "name": "headers"
                  },
                  "address": "192.168.0.16",
                  "allowInsecureCertificate": true,
                  "failureThreshold": 5,
                  "hostname": "test.com",
                  "interval": "60s",
                  "path": "/healthz",
                  "port": 443,
                  "protocol": "HTTPS"
                }
              },
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSRecord",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dns-operator",
                    "app.kubernetes.io/instance": "dnsrecord-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dnsrecord",
                    "app.kubernetes.io/part-of": "dns-operator"
                  },
                  "name": "dnsrecord-sample"
                },
                "spec": {
                  "endpoints": [
                    {
                      "dnsName": "dnsrecord-simple.example.com",
                      "recordTTL": 60,
                      "recordType": "A",
                      "targets": [
                        "52.215.108.61",
                        "52.30.101.221"
                      ]
                    }
                  ],
                  "providerRef": {
                    "name": "dns-provider-creds"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/dns-rhel9-operator@sha256:8ad3876e36ba2f76355499d98a97672f54f26754fc9d219d9cb1c5b9f9874430
          createdAt: "2025-04-07T08:13:58Z"
          description: Configures how north-south traffic from outside the network
            is balanced and reaches Gateways
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kuadrant/dns-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSHealthCheckProbe is the Schema for the dnshealthcheckprobes
              API
            displayName: DNSHealthCheckProbe
            kind: DNSHealthCheckProbe
            name: dnshealthcheckprobes.kuadrant.io
            version: v1alpha1
          - description: DNSRecord is the Schema for the dnsrecords API
            displayName: DNSRecord
            kind: DNSRecord
            name: dnsrecords.kuadrant.io
            version: v1alpha1
        description: Configures how north-south traffic from outside the network is
          balanced and reaches Gateways - Part of Red Hat Connectivity Link
        displayName: Red Hat - DNS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dns
        - kuadrant
        links:
        - name: GitHub
          url: https://github.com/kuadrant/dns-operator
        - name: Documentation
          url: https://github.com/Kuadrant/dns-operator/tree/main/docs
        maintainers:
        - email: mnairn@redhat.com
          name: Michael Nairn
        - email: pbrookes@redhat.com
          name: Phil Brookes
        - email: cbrookes@redhat.com
          name: Craig Brookes
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhcl-1/dns-rhel9-operator@sha256:8ad3876e36ba2f76355499d98a97672f54f26754fc9d219d9cb1c5b9f9874430
        version: 1.1.0
      entries:
      - name: dns-operator.v1.1.0
        version: 1.1.0
      - name: dns-operator.v1.0.2
        version: 1.0.2
      - name: dns-operator.v1.0.1
        version: 1.0.1
      - name: dns-operator.v0.12.0
        version: 0.12.0
      name: stable
    defaultChannel: stable
    packageName: dns-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: patch-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: patch-operator.v0.1.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Patch",
                "metadata": {
                  "name": "test-complex-patch"
                },
                "spec": {
                  "patches": {
                    "test-complex-patch": {
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 1).metadata.name }}: {{ (index . 2).metadata.name }}    \n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "kind": "Namespace",
                          "name": "test-patch-operator"
                        },
                        {
                          "apiVersion": "v1",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "test-patch-operator"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "test-patch-operator"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/patch-operator@sha256:030ade9b9428deca3ac6390618dc4d6e6d728577d1467a22dc4d4b68c9848886
          createdAt: "2024-05-10T14:07:19Z"
          description: An operator to apply patches to Kubernetes objects in a declarative
            way.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: patch-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/patch-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Patch
            name: patches.redhatcop.redhat.io
            version: v1alpha1
        description: "The patch operator helps with defining patches in a declarative
          way. This operator has two main features:\n\n1. [ability to patch an object
          at creation time via a mutating webhook](#creation-time-patch-injection)\n2.
          [ability to enforce patches on one or more objects via a controller](#runtime-patch-enforcement)\n
          \n## Creation-time patch injection\n\nWhy apply a patch at creation time
          when you could directly create the correct object? The reason is that sometime
          the correct value depends on configuration set on the specific cluster in
          which the object is being deployed. For example, an ingress/route hostname
          might depend on the specific cluster. Consider the following example based
          on cert-manager:\n \n```yaml\napiVersion: cert-manager.io/v1\nkind: ClusterIssuer\nmetadata:\n
          \ name: letsencrypt-issuer\nspec:\n  acme:\n    server: 'https://acme-v02.api.letsencrypt.org/directory'\n
          \   email: {{ .Values.letsencrypt.email }}\n    privateKeySecretRef:\n      name:
          letsencrypt-staging\n    solvers:  \n    - dns01:\n        route53:\n          accessKeyID:
          << access_key >>\n          secretAccessKeySecretRef:\n            name:
          cert-manager-dns-credentials\n            key: aws_secret_access_key\n          region:
          << region >>\n          hostedZoneID: << hosted_zone_id >>\n ```\n\nIn this
          example the fields: `<< access_key >>`, `<< region >>` and `<< hosted_zone_id
          >>` are dependent on the specific region in which the cluster is being deployed
          and in many cases they are discoverable from other configurations already
          present in the cluster. If you want to deploy the above Cluster Issuer object
          with a gitops approach, then there is no easy way to discover those values.
          The solution so far is to manually discover those values and create a different
          gitops configuration for each cluster. But consider if you could look up
          values at deploy time based on the cluster you are deploying to. Here is
          how this object might look:\n\n```yaml\napiVersion: cert-manager.io/v1\nkind:
          ClusterIssuer\nmetadata:\n  name: letsencrypt-issuer\n  namespace: {{ .Release.Namespace
          }}\n  annotations:\n    \"redhat-cop.redhat.io/patch\": |\n      spec:\n
          \       acme:\n        - dns01:\n            route53:\n              accessKeyID:
          {{ (lookup \"v1\" \"Secret\" .metadata.namespace \"cert-manager-dns-credentials\").data.aws_access_key_id
          | b64dec }}\n              secretAccessKeySecretRef:\n                name:
          cert-manager-dns-credentials\n                key: aws_secret_access_key\n
          \             region: {{ (lookup \"config.openshift.io/v1\" \"Infrastructure\"
          \"\" \"cluster\").status.platformStatus.aws.region }}\n              hostedZoneID:
          {{ (lookup \"config.openshift.io/v1\" \"DNS\" \"\" \"cluster\").spec.publicZone.id
          }} \nspec:\n  acme:\n    server: 'https://acme-v02.api.letsencrypt.org/directory'\n
          \   email: {{ .Values.letsencrypt.email }}\n    privateKeySecretRef:\n      name:
          letsencrypt-staging\n    solvers:  \n    - dns01:\n        route53:\n          accessKeyID:
          << access_key >>\n          secretAccessKeySecretRef:\n            name:
          cert-manager-dns-credentials\n            key: aws_secret_access_key\n          region:
          << region >>\n          hostedZoneID: << hosted_zone_id >>\n ```\n\nThe
          annotation specifies a patch that will be applied by a [MutatingWebhook](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/),
          as you can see the three values are now looked up from different configurations
          in the cluster.\n\nTwo annotations influences the behavior of this MutatingWebhook\n
          \n1. \"redhat-cop.redhat.io/patch\" : this is the patch itself. The patch
          is evaluated as a template with the object itself as it's only parameter.
          The template is expressed in golang template notation and supports the same
          functions as helm template including the [lookup](https://helm.sh/docs/chart_template_guide/functions_and_pipelines/#using-the-lookup-function)
          function which plays a major role here. The patch must be expressed in yaml
          for readability. It will be converted to json by the webhook logic.\n2.
          \"redhat-cop.redhat.io/patch-type\" : this is the type of json patch. The
          possible values are: `application/json-patch+json`, `application/merge-patch+json`
          and `application/strategic-merge-patch+json`. If this annotation is omitted
          it defaults to strategic merge.\n\n### Security Considerations\n \nThe lookup
          function, if used by the template, is executed with a client which impersonates
          the user issuing the object creation/update request. This should prevent
          security permission leakage.\n\n### Installing the creation time webhook\n
          \nThe creation time webhook is not installed by the operator. This is because
          there is no way to know which specific object type should be intercepted
          and intercepting all of the types would be too inefficient. It's up to the
          administrator then to install the webhook. Here is some guidance.\n\nIf
          you installed the operator via OLM, use the following webhook template:\n\n```yaml\napiVersion:
          admissionregistration.k8s.io/v1\n kind: MutatingWebhookConfiguration\nmetadata:\n
          \ name: patch-operator-inject\n  annotations:\n    service.beta.openshift.io/inject-cabundle:
          \"true\"\nwebhooks:\n - admissionReviewVersions:\n  - v1\n  clientConfig:\n
          \   service:\n      name: patch-operator-webhook-service\n      namespace:
          patch-operator\n      path: /inject\n  failurePolicy: Fail\n  name: patch-operator-inject.redhatcop.redhat.io\n
          \ rules:\n  - << add your intercepted objects here >>\n  sideEffects: None\n```\n\nIf
          you installed the operator via the Helm chart and are using cert-manager,
          use the following webhook template:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\n
          kind: MutatingWebhookConfiguration\nmetadata:\n  name: patch-operator-inject\n
          \ annotations:\n    cert-manager.io/inject-ca-from: '{{ .Release.Namespace
          }}/webhook-server-cert'\n webhooks:\n- admissionReviewVersions:\n  - v1\n
          \ clientConfig:\n    service:\n      name: patch-operator-webhook-service\n
          \     namespace: patch-operator\n      path: /inject\n  failurePolicy: Fail\n
          \ name: patch-operator-inject.redhatcop.redhat.io\n  rules:\n  - << add
          your intercepted objects here >>\n  sideEffects: None\n```  \n\nYou should
          need to enable the webhook only for `CREATE` operations. So for example
          to enable the webhook on configmaps:\n\n```yaml\n  rules:\n  - apiGroups:\n
          \   - \"\"\n    apiVersions:\n    - v1\n    operations:\n    - CREATE\n
          \   resources:\n    - configmaps\n```\n\n## Runtime patch enforcement\n\nThere
          are situations when we need to patch pre-existing objects. Again this is
          a use case that is hard to model with gitops operators which will work only
          on object that they own. Especially with sophisticated Kubernetes distributions,
          it is not uncommon that a Kubernetes instance, at installation time, is
          configured with some default settings. Changing those configurations means
          patching those objects. For example, let's take the case of OpenShift Oauth
          configuration. This object is present by default and it is expected to be
          patched with any newly enabled authentication mechanism. This is how it
          looks like after installation:\n\n```yaml\napiVersion: config.openshift.io/v1\n
          kind: OAuth\nmetadata:\n  name: cluster\n  ownerReferences:\n    - apiVersion:
          config.openshift.io/v1\n      kind: ClusterVersion\n      name: version\n
          \     uid: 9a9d450b-3076-4e30-ac05-a889d6341fc3\n  resourceVersion: '20405124'\n
          spec: []\n```\n\nIf we need to patch it we can use the patch controller
          and the `Patch` object as follows:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: Patch\nmetadata:\n  name: gitlab-ocp-oauth-provider\n  namespace:
          openshift-config\nspec:\n  serviceAccountRef:\n    name: default\n  patches:\n
          \   gitlab-ocp-oauth-provider:\n      targetObjectRef:\n        apiVersion:
          config.openshift.io/v1\n        kind: OAuth\n        name: cluster\n      patchTemplate:
          |\n        spec:\n          identityProviders:\n          - name: my-github
          \n            mappingMethod: claim \n            type: GitHub\n            github:\n
          \             clientID: \"{{ (index . 1).data.client_id | b64dec }}\" \n
          \             clientSecret: \n                name: ocp-github-app-credentials\n
          \             organizations: \n              - my-org\n              teams:
          []            \n      patchType: application/merge-patch+json\n      sourceObjectRefs:\n
          \     - apiVersion: v1\n        kind: Secret\n        name: ocp-github-app-credentials\n
          \       namespace: openshift-config\n```\n\nThis will cause the OAuth object
          to be patched and the patch to be enforced. That means that if anything
          changes on the secret that we use a parameter (which may be rotated) or
          in Oauth object itself, the patch will be reapplied. In this case we are
          adding a gitlab authentication provider.\n\n A `patch` has the following
          fields:\n\n`targetObjectRef` this refers to the object(s) receiving the
          patch. Mutliple object can be selected based on the following rules:\n \n|
          Namespaced Type | Namespace | Name | Selection type |\n| --- | --- | ---
          | --- |\n| yes | null | null | multiple selection across namespaces |\n|
          yes | null | not null | multiple selection across namespaces where the name
          corresponds to the passed name |\n| yes | not null | null | multiple selection
          within a namespace |\n| yes | not null | not nul | single selection |\n|
          no | N/A | null | multiple selection  |\n| no | N/A | not null | single
          selection |\n\nSelection can be further narrowed down by filtering by labels
          and/or annotations using the `labelSelector` and `annotationSelector` fields.
          The patch will be applied to all of the selected instances.\n\n`sourceObjectRefs`
          these are the objects that will be watched and become part of the parameters
          of the patch template. Name and Namespace of sourceRefObjects are interpreted
          as golang templates with the current target instance and the only parameter.
          This allows to select different source object for each target object.\n
          \nSo, for example, with this patch:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: Patch\nmetadata:\n  name: multiple-namespaced-targets-patch\nspec:\n
          \ serviceAccountRef:\n    name: default\n  patches:\n    multiple-namespaced-targets-patch:\n
          \     targetObjectRef:\n        apiVersion: v1\n        kind: ServiceAccount\n
          \       name: deployer\n      patchTemplate: |\n        metadata:\n          annotations:\n
          \           {{ (index . 1).metadata.uid }}: {{ (index . 1) }}\n      patchType:
          application/strategic-merge-patch+json\n      sourceObjectRefs:\n      -
          apiVersion: v1\n        kind: ServiceAccount\n        name: default\n        namespace:
          \"{{ .metadata.namespace }}\"\n        fieldPath: $.metadata.uid\n```\n\nThe
          `deployer` service accounts from all namespaces are selected as target of
          this patch, each patch template will receive a different parameter and that
          is the `default` service account of the same namespace as the namespace
          of the `deployer` service account being processed.\n\n`sourceObjectRefs`
          also have the `fieldPath` field which can contain a jsonpath expression.
          If a value is passed the jsonpath expression will be calculate for the current
          source object and the result will be passed as parameter of the template.\n\n`patchTemplate`
          This is the the template that will be evaluated. The result must be a valid
          patch compatible with the requested type and expressed in yaml for readability.
          The parameters passed to the template are the target object and then the
          all of the source object. So if you want to refer to the target object in
          the template you can use this expression `(index . 0)`. Higher indexes refer
          to the sourceObjectRef array. The template is expressed in golang template
          notation and supports the same functions as helm template.\n\n`patchType`
          is the type of the json patch. The possible values are: `application/json-patch+json`,
          `application/merge-patch+json` and `application/strategic-merge-patch+json`.
          If this annotation is omitted it defaults to strategic merge.\n\n### Patch
          Controller Security Considerations\n\nThe patch enforcement enacted by the
          patch controller is executed with a client which uses the service account
          referenced by the `serviceAccountRef` field. So before a patch object can
          actually work an administrator must have granted the needed permissions
          to a service account in the same namespace. The `serviceAccountRef` will
          default to the `default` service account if not specified.\n\n### Patch
          Controller Performance Considerations\n\nThe patch controller will create
          a controller-manager and per `Patch` object and a reconciler for each of
          the `PatchSpec` defined in the array on patches in the `Patch` object.\nThese
          reconcilers share the same cached client. In order to be able to watch changes
          on target and source objects of a `PatchSpec`, all of the target and source
          object type instances will be cached by the client. This is a normal behavior
          of a controller-manager client, but it implies that if you create patches
          on object types that have many instances in etcd (Secrets, ServiceAccounts,
          Namespaces for example), the patch operator instance will require a significant
          amount of memory. A way to contain this issue is to try to aggregate together
          `PatchSpec` that deal with the same object types. This will cause those
          object type instances to cached only once.  \n"
        displayName: Patch Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - patch
        - enforcing
        - declarative
        links:
        - name: Patch Operator
          url: https://github.com/redhat-cop/patch-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/patch-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/patch-operator@sha256:030ade9b9428deca3ac6390618dc4d6e6d728577d1467a22dc4d4b68c9848886
        version: 0.1.11
      entries:
      - name: patch-operator.v0.1.11
        version: 0.1.11
      - name: patch-operator.v0.1.10
        version: 0.1.10
      - name: patch-operator.v0.1.9
        version: 0.1.9
      - name: patch-operator.v0.1.8
        version: 0.1.8
      - name: patch-operator.v0.1.7
        version: 0.1.7
      - name: patch-operator.v0.1.6
        version: 0.1.6
      - name: patch-operator.v0.1.5
        version: 0.1.5
      - name: patch-operator.v0.1.4
        version: 0.1.4
      - name: patch-operator.v0.1.3
        version: 0.1.3
      - name: patch-operator.v0.1.2
        version: 0.1.2
      - name: patch-operator.v0.1.1
        version: 0.1.1
      - name: patch-operator.v0.1.0
        version: 0.1.0
      - name: patch-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: patch-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-elbv2-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-elbv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "LoadBalancer",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "TargetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/elbv2-controller:1.1.1
          createdAt: "2025-09-19T22:16:38Z"
          description: AWS ELB controller is a service controller for managing ELB
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Listener represents the state of an AWS elbv2 Listener resource.
            displayName: Listener
            kind: Listener
            name: listeners.elbv2.services.k8s.aws
            version: v1alpha1
          - description: LoadBalancer represents the state of an AWS elbv2 LoadBalancer
              resource.
            displayName: LoadBalancer
            kind: LoadBalancer
            name: loadbalancers.elbv2.services.k8s.aws
            version: v1alpha1
          - description: Rule represents the state of an AWS elbv2 Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.elbv2.services.k8s.aws
            version: v1alpha1
          - description: TargetGroup represents the state of an AWS elbv2 TargetGroup
              resource.
            displayName: TargetGroup
            kind: TargetGroup
            name: targetgroups.elbv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Load Balancing (ELB) resources in AWS from within your Kubernetes cluster.

          **About Amazon ELB**

          Elastic Load Balancing automatically distributes your incoming traffic across multiple targets, such as EC2 instances, containers, and IP addresses, in one or more Availability Zones. It monitors the health of its registered targets, and routes traffic only to the healthy targets. Elastic Load Balancing scales your load balancer capacity automatically in response to changes in incoming traffic.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ELB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elbv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ELB Developer Resources
          url: https://aws.amazon.com/elasticloadbalancing
        maintainers:
        - email: ack-maintainers@amazon.com
          name: elb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/elbv2-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-elbv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-elbv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-elbv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-elbv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-elbv2-controller.v1.0.10
        version: 1.0.10
      - name: ack-elbv2-controller.v1.0.9
        version: 1.0.9
      - name: ack-elbv2-controller.v1.0.8
        version: 1.0.8
      - name: ack-elbv2-controller.v1.0.7
        version: 1.0.7
      - name: ack-elbv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-elbv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-elbv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-elbv2-controller.v1.0.3
        version: 1.0.3
      - name: ack-elbv2-controller.v1.0.2
        version: 1.0.2
      - name: ack-elbv2-controller.v1.0.1
        version: 1.0.1
      - name: ack-elbv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-elbv2-controller.v0.0.11
        version: 0.0.11
      - name: ack-elbv2-controller.v0.0.10
        version: 0.0.10
      - name: ack-elbv2-controller.v0.0.9
        version: 0.0.9
      - name: ack-elbv2-controller.v0.0.8
        version: 0.0.8
      - name: ack-elbv2-controller.v0.0.6
        version: 0.0.6
      - name: ack-elbv2-controller.v0.0.5
        version: 0.0.5
      - name: ack-elbv2-controller.v0.0.4
        version: 0.0.4
      - name: ack-elbv2-controller.v0.0.3
        version: 0.0.3
      - name: ack-elbv2-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-elbv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hazelcast, Inc
      provider-url: ""
    name: hazelcast-platform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hazelcast-platform-operator.v5.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "cache-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "CronHotBackup",
                "metadata": {
                  "name": "cronhotbackup-sample"
                },
                "spec": {
                  "hotBackupTemplate": {
                    "metadata": {
                      "labels": {
                        "cron-hotbackup": "true"
                      }
                    },
                    "spec": {
                      "hazelcastResourceName": "hazelcast"
                    }
                  },
                  "schedule": "*/30 * * * *"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast"
                },
                "spec": {
                  "clusterSize": 3,
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast-sample"
                },
                "spec": {
                  "clusterSize": 3,
                  "jet": {
                    "bucketConfig": {
                      "bucketURI": "gs://operator-user-code/jetJobs",
                      "secretName": "br-secret-gcp"
                    },
                    "enabled": true,
                    "resourceUploadEnabled": true
                  },
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HazelcastEndpoint",
                "metadata": {
                  "name": "hazelcastendpoint-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "port": 5701,
                  "type": "Discovery"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HotBackup",
                "metadata": {
                  "name": "hot-backup"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-longrun"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-longrun-2.0.0.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-1.0.2.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJobSnapshot",
                "metadata": {
                  "name": "jetjobsnapshot-sample"
                },
                "spec": {
                  "cancelJob": false,
                  "jetJobResourceName": "jet-job"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ManagementCenter",
                "metadata": {
                  "name": "managementcenter"
                },
                "spec": {
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "hazelcastClusters": [
                    {
                      "address": "hazelcast",
                      "name": "dev"
                    }
                  ],
                  "licenseKeySecretName": "hazelcast-license-key",
                  "persistence": {
                    "enabled": true,
                    "size": "10Gi"
                  },
                  "repository": "hazelcast/management-center"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Map",
                "metadata": {
                  "name": "map"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "MultiMap",
                "metadata": {
                  "name": "multimap-sample"
                },
                "spec": {
                  "backupCount": 3,
                  "binary": true,
                  "collectionType": "LIST",
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "queue-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ReplicatedMap",
                "metadata": {
                  "name": "replicatedmap-sample"
                },
                "spec": {
                  "asyncFillup": true,
                  "hazelcastResourceName": "hazelcast",
                  "inMemoryFormat": "OBJECT"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "topic-sample"
                },
                "spec": {
                  "globalOrderingEnabled": true,
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "UserCodeNamespace",
                "metadata": {
                  "name": "ucn-sample"
                },
                "spec": {
                  "bucketConfig": {
                    "bucketURI": "gs://operator-user-code/userCodeNamespace",
                    "secretName": "br-secret-gcp"
                  },
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanReplication",
                "metadata": {
                  "name": "wanreplication-sample"
                },
                "spec": {
                  "endpoints": "35.192.33.252",
                  "resources": [
                    {
                      "kind": "Map",
                      "name": "map-sample"
                    },
                    {
                      "kind": "Hazelcast",
                      "name": "hz-sample"
                    }
                  ],
                  "targetClusterName": "dev"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanSync",
                "metadata": {
                  "name": "wansync-sample"
                },
                "spec": {
                  "wanReplicationResourceName": "wan-replication-sample"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: docker.io/hazelcast/hazelcast-platform-operator@sha256:99005afda5c71ceeac200928dfed4ea35480f871053101e3aeaa44005e3fda14
          createdAt: "2024-07-29T15:35:56Z"
          description: '[DEPRECATED] Use the certified Hazelcast Operator instead'
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hazelcast/hazelcast-platform-operator/
          support: Hazelcast, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.hazelcast.com
            version: v1alpha1
          - description: CronHotBackup is the Schema for the cronhotbackups API
            displayName: Cron Hot Backup
            kind: CronHotBackup
            name: cronhotbackups.hazelcast.com
            version: v1alpha1
          - description: HazelcastEndpoint is the Schema for the hazelcastendpoints
              API
            displayName: Hazelcast Endpoint
            kind: HazelcastEndpoint
            name: hazelcastendpoints.hazelcast.com
            version: v1alpha1
          - description: Hazelcast is the Schema for the hazelcasts API
            displayName: Hazelcast
            kind: Hazelcast
            name: hazelcasts.hazelcast.com
            version: v1alpha1
          - description: HotBackup is the Schema for the hot backup API
            displayName: Hot Backup
            kind: HotBackup
            name: hotbackups.hazelcast.com
            version: v1alpha1
          - description: JetJob is the Schema for the jetjobs API
            displayName: Jet Job
            kind: JetJob
            name: jetjobs.hazelcast.com
            version: v1alpha1
          - description: JetJobSnapshot is the Schema for the jetjobsnapshots API
            displayName: Jet Job Snapshot
            kind: JetJobSnapshot
            name: jetjobsnapshots.hazelcast.com
            version: v1alpha1
          - description: ManagementCenter is the Schema for the managementcenters
              API
            displayName: Management Center
            kind: ManagementCenter
            name: managementcenters.hazelcast.com
            version: v1alpha1
          - description: Map is the Schema for the maps API
            displayName: Map
            kind: Map
            name: maps.hazelcast.com
            version: v1alpha1
          - description: MultiMap is the Schema for the multimaps API
            displayName: Multi Map
            kind: MultiMap
            name: multimaps.hazelcast.com
            version: v1alpha1
          - description: Queue is the Schema for the queues API
            displayName: Queue
            kind: Queue
            name: queues.hazelcast.com
            version: v1alpha1
          - description: ReplicatedMap is the Schema for the replicatedmaps API
            displayName: Replicated Map
            kind: ReplicatedMap
            name: replicatedmaps.hazelcast.com
            version: v1alpha1
          - description: Topic is the Schema for the topics API
            displayName: Topic
            kind: Topic
            name: topics.hazelcast.com
            version: v1alpha1
          - description: UserCodeNamespace is the Schema for the usercodenamespaces
              API
            displayName: User Code Namespace
            kind: UserCodeNamespace
            name: usercodenamespaces.hazelcast.com
            version: v1alpha1
          - description: WanReplication is the Schema for the wanreplications API
            displayName: Wan Replication
            kind: WanReplication
            name: wanreplications.hazelcast.com
            version: v1alpha1
          - description: WanSync is the Schema for the wansyncs API
            displayName: Wan Sync
            kind: WanSync
            name: wansyncs.hazelcast.com
            version: v1alpha1
        description: |
          ## Deprecated
          Use the certified Hazelcast Operator instead.
          # Hazelcast Platform Operator #

          Easily deploy Hazelcast clusters and Management Center into Kubernetes environments and manage their lifecycles.

          ## Before You Start

          If you are planning to create Hazelcast Platform Enterprise clusters, you need to [create a secret](https://docs.hazelcast.com/operator/latest/get-started#step-2-start-the-hazelcast-cluster) for the license. You can request a trial license key from [here](https://trialrequest.hazelcast.com).

          For Hazelcast Platform clusters, you can simply continue.

          ## Documentation

          1. [Get started](https://docs.hazelcast.com/operator/latest/get-started) with the Operator
          2. [Connect to the cluster from outside Kubernetes](https://guides.hazelcast.org/hazelcast-platform-operator-expose-externally)
          3. [Restore a Cluster from Cloud Storage with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-external-backup-restore)
          4. [Replicate Data between Two Hazelcast Clusters with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-wan-replication)

          ## Features

          Hazelcast Platform Operator supports the features below:

          * Custom resource for Hazelcast Platform (Enterprise) and Management Center
          * Observe status of Hazelcast clusters and Management Center instances
          * High Availability Mode configuration to create clusters that are resilient to node and zone failures
          * Support for TLS connections between members using self-signed certificates
          * Scale up and down Hazelcast clusters
          * Expose Hazelcast cluster to external clients ([Smart & Unisocket](https://docs.hazelcast.com/hazelcast/latest/clients/java#java-client-operation-modes))
          * Backup Hazelcast persistence data to cloud storage with the possibility of scheduling it and restoring the data accordingly
          * WAN Replication feature when you need to synchronize multiple Hazelcast clusters, which are connected by WANs
          * Full and Delta WAN Sync support
          * Tiered Storage support for Hazelcast and Map CRs
          * CP Subsystem configuration support for Hazelcast CR
          * User Code Deployment feature, which allows you to deploy custom and domain classes from cloud storages and URLs to Hazelcast members
          * The User Code Namespaces feature allows you to deploy custom and domain classes at runtime without a members restart
          * Support the configuration of advanced networking options
          * Support Multi-namespace configuration
          * ExecutorService and EntryProcessor support
          * Support several data structures like Map, Topic, MultiMap, ReplicatedMap, Queue and Cache which can be created dynamically via specific Custom Resources
          * MapStore, Near Cache and off-heap memory (HD memory and native memory) support for the Map CR
          * Native Memory support for the Cache CR
          * Support Jet configuration and Jet Job submission using the JetJob CR
          * Support for exporting the snapshots of JetJob CRs using JetJobSnapshot CR
          * Support for custom configurations using ConfigMap
        displayName: '[DEPRECATED] Hazelcast Platform Operator'
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hazelcast
        - keyvalue
        - in-memory
        - database
        - caching
        links:
        - name: Documentation
          url: https://docs.hazelcast.com/operator/latest/
        - name: Source Code
          url: https://github.com/hazelcast/hazelcast-platform-operator
        maintainers:
        - email: cloudnative@hazelcast.com
          name: Hazelcast Cloud Native Team
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Hazelcast, Inc
        relatedImages:
        - docker.io/hazelcast/hazelcast-platform-operator@sha256:99005afda5c71ceeac200928dfed4ea35480f871053101e3aeaa44005e3fda14
        version: 5.13.1
      entries:
      - name: hazelcast-platform-operator.v5.13.1
        version: 5.13.1
      - name: hazelcast-platform-operator.v5.13.0
        version: 5.13.0
      - name: hazelcast-platform-operator.v5.12.0
        version: 5.12.0
      - name: hazelcast-platform-operator.v5.11.0
        version: 5.11.0
      - name: hazelcast-platform-operator.v5.10.0
        version: 5.10.0
      - name: hazelcast-platform-operator.v5.9.0
        version: 5.9.0
      - name: hazelcast-platform-operator.v5.8.0
        version: 5.8.0
      - name: hazelcast-platform-operator.v5.7.0
        version: 5.7.0
      - name: hazelcast-platform-operator.v5.6.0
        version: 5.6.0
      - name: hazelcast-platform-operator.v5.5.0
        version: 5.5.0
      - name: hazelcast-platform-operator.v5.4.0
        version: 5.4.0
      - name: hazelcast-platform-operator.v5.3.0
        version: 5.3.0
      - name: hazelcast-platform-operator.v5.2.0
        version: 5.2.0
      - name: hazelcast-platform-operator.v5.1.0
        version: 5.1.0
      - name: hazelcast-platform-operator.v5.0.0
        version: 5.0.0
      name: alpha
    defaultChannel: alpha
    packageName: hazelcast-platform-operator
    provider:
      name: Hazelcast, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-storage
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-storage.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Storage",
                "metadata": {
                  "name": "storage-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore"
                  },
                  "minio": {
                    "clusterDomain": "cluster.local",
                    "helmKubectlJqImage": {
                      "repository": "docker.io/bskim45/helm-kubectl-jq"
                    },
                    "image": {
                      "repository": "prod.harbor.keysight.digital/loadcore/wap-minio",
                      "tag": "RELEASE.2021-12-29T06-49-06Z-3747-37-ocp"
                    },
                    "mcImage": {
                      "repository": "prod.harbor.keysight.digital/loadcore/wap-minio-mc",
                      "tag": "RELEASE.2021-12-29T06-52-55Z-3747-37-ocp"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "securityContext": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:32:20Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Storage
            name: storages.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Storage Operator
        displayName: Keysight WAP Storage Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Storage
          url: https://keysight-wap-storage.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-storage@sha256:9f7710cfe72658cd5456fd54f86ca6be5c9a4de2a01114c0f7531b70712bd981
        version: 0.1.21
      entries:
      - name: keysight-wap-storage.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-storage
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/cluster-nfd-operator
    name: openshift-nfd-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfd.v4.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nfd.openshift.io/v1",
                "kind": "NodeFeatureDiscovery",
                "metadata": {
                  "name": "nfd-instance",
                  "namespace": "openshift-nfd"
                },
                "spec": {
                  "customConfig": {
                    "configData": "#    - name: \"more.kernel.features\"\n#      matchOn:\n#      - loadedKMod: [\"example_kmod3\"]\n#    - name: \"more.features.by.nodename\"\n#      value: customValue\n#      matchOn:\n#      - nodename: [\"special-.*-node-.*\"]\n"
                  },
                  "operand": {
                    "image": "quay.io/openshift/origin-node-feature-discovery:4.10",
                    "servicePort": 12000
                  },
                  "workerConfig": {
                    "configData": "core:\n#  labelWhiteList:\n#  noPublish: false\n  sleepInterval: 60s\n#  sources: [all]\n#  klog:\n#    addDirHeader: false\n#    alsologtostderr: false\n#    logBacktraceAt:\n#    logtostderr: true\n#    skipHeaders: false\n#    stderrthreshold: 2\n#    v: 0\n#    vmodule:\n##   NOTE: the following options are not dynamically run-time \n##          configurable and require a nfd-worker restart to take effect\n##          after being changed\n#    logDir:\n#    logFile:\n#    logFileMaxSize: 1800\n#    skipLogHeaders: false\nsources:\n#  cpu:\n#    cpuid:\n##     NOTE: whitelist has priority over blacklist\n#      attributeBlacklist:\n#        - \"BMI1\"\n#        - \"BMI2\"\n#        - \"CLMUL\"\n#        - \"CMOV\"\n#        - \"CX16\"\n#        - \"ERMS\"\n#        - \"F16C\"\n#        - \"HTT\"\n#        - \"LZCNT\"\n#        - \"MMX\"\n#        - \"MMXEXT\"\n#        - \"NX\"\n#        - \"POPCNT\"\n#        - \"RDRAND\"\n#        - \"RDSEED\"\n#        - \"RDTSCP\"\n#        - \"SGX\"\n#        - \"SSE\"\n#        - \"SSE2\"\n#        - \"SSE3\"\n#        - \"SSE4.1\"\n#        - \"SSE4.2\"\n#        - \"SSSE3\"\n#      attributeWhitelist:\n#  kernel:\n#    kconfigFile: \"/path/to/kconfig\"\n#    configOpts:\n#      - \"NO_HZ\"\n#      - \"X86\"\n#      - \"DMI\"\n  pci:\n    deviceClassWhitelist:\n      - \"0200\"\n      - \"03\"\n      - \"12\"\n    deviceLabelFields:\n#      - \"class\"\n      - \"vendor\"\n#      - \"device\"\n#      - \"subsystem_vendor\"\n#      - \"subsystem_device\"\n#  usb:\n#    deviceClassWhitelist:\n#      - \"0e\"\n#      - \"ef\"\n#      - \"fe\"\n#      - \"ff\"\n#    deviceLabelFields:\n#      - \"class\"\n#      - \"vendor\"\n#      - \"device\"\n#  custom:\n#    - name: \"my.kernel.feature\"\n#      matchOn:\n#        - loadedKMod: [\"example_kmod1\", \"example_kmod2\"]\n#    - name: \"my.pci.feature\"\n#      matchOn:\n#        - pciId:\n#            class: [\"0200\"]\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#        - pciId :\n#            vendor: [\"8086\"]\n#            device: [\"1000\", \"1100\"]\n#    - name: \"my.usb.feature\"\n#      matchOn:\n#        - usbId:\n#          class: [\"ff\"]\n#          vendor: [\"03e7\"]\n#          device: [\"2485\"]\n#        - usbId:\n#          class: [\"fe\"]\n#          vendor: [\"1a6e\"]\n#          device: [\"089a\"]\n#    - name: \"my.combined.feature\"\n#      matchOn:\n#        - pciId:\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#          loadedKMod : [\"vendor_kmod1\", \"vendor_kmod2\"]\n"
                  }
                }
              },
              {
                "apiVersion": "nfd.openshift.io/v1alpha1",
                "kind": "NodeFeatureRule",
                "metadata": {
                  "name": "my-sample-rule-object"
                },
                "spec": {
                  "rules": [
                    {
                      "labels": {
                        "my-sample-feature": "true"
                      },
                      "matchFeatures": [
                        {
                          "feature": "kernel.loadedmodule",
                          "matchExpressions": {
                            "dummy": {
                              "op": "Exists"
                            }
                          }
                        },
                        {
                          "feature": "kernel.config",
                          "matchExpressions": {
                            "X86": {
                              "op": "In",
                              "value": [
                                "y"
                              ]
                            }
                          }
                        }
                      ],
                      "name": "my sample rule"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery,OpenShift Optional
          containerImage: quay.io/openshift/origin-cluster-nfd-operator:4.10
          description: |-
            The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

            NFD consists  of the following software components:

            The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

            ##NFD-Master
            NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

            ##NFD-Worker
            NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

            ##NFD-Topology-Updater
            NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
          olm.skipRange: '>=4.6.0 <4.11.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-nfd
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/internal-objects: '["noderesourcetopologies.topology.node.k8s.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/cluster-nfd-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              The NodeFeatureDiscovery instance is the CustomResource being watched by the NFD-Operator, and holds all the needed information to setup the behaviour of the master and worker pods
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
          - description: |
              NodeFeatureRule resource specifies a configuration for feature-based customization of node objects, such as node labeling.
            kind: NodeFeatureRule
            name: nodefeaturerules.nfd.openshift.io
            version: v1alpha1
          - description: |
              NodeResourceTopology resource describes node resources and their topology.
            kind: NodeResourceTopology
            name: noderesourcetopologies.topology.node.k8s.io
            version: v1alpha1
        description: |-
          The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

          NFD consists  of the following software components:

          The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

          ##NFD-Master
          NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

          ##NFD-Worker
          NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

          ##NFD-Topology-Updater
          NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
        displayName: Node Feature Discovery Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - feature-discovery
        - feature-detection
        - node-labels
        links:
        - name: Node Feature Discovery Operator
          url: https://docs.openshift.com/container-platform/4.9/hardware_enablement/psap-node-feature-discovery-operator.html
        - name: Node Feature Discovery Documentation
          url: https://kubernetes-sigs.github.io/node-feature-discovery/stable/get-started/index.html
        maintainers:
        - email: openshift-psap@redhat.com
          name: openshift-psap
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/cluster-nfd-operator
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy
        - quay.io/openshift/origin-cluster-nfd-operator:4.10
        version: 4.10.0
      entries:
      - name: nfd.v4.10.0
        version: 4.10.0
      - name: nfd.v4.9.0
        version: 4.9.0
      - name: nfd.v4.8.0
        version: 4.8.0
      name: stable
    defaultChannel: stable
    packageName: openshift-nfd-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/cluster-nfd-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-bedrockagent-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-bedrockagent-controller.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bedrockagent.services.k8s.aws/v1alpha1",
                "kind": "Agent",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/bedrockagent-controller:0.1.1
          createdAt: "2025-09-20T01:18:56Z"
          description: AWS Bedrock Agent controller is a service controller for managing
            Bedrock Agent resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Agent represents the state of an AWS bedrockagent Agent resource.
            displayName: Agent
            kind: Agent
            name: agents.bedrockagent.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Bedrock Agent resources in AWS from within your Kubernetes cluster.

          **About Amazon Bedrock Agent**

          Amazon Bedrock is a fully managed service that makes high-performing foundation models (FMs) from leading AI companies and Amazon available for your use through a unified API. You can choose from a wide range of foundation models to find the model that is best suited for your use case. Amazon Bedrock also offers a broad set of capabilities to build generative AI applications with security, privacy, and responsible AI. Using Amazon Bedrock, you can easily experiment with and evaluate top foundation models for your use cases, privately customize them with your data using techniques such as fine-tuning and Retrieval Augmented Generation (RAG), and build agents that execute tasks using your enterprise systems and data sources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Bedrock Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bedrockagent
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Bedrock Agent Developer Resources
          url: https://aws.amazon.com/bedrock/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: bedrock-agent maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/bedrockagent-controller:0.1.1
        version: 0.1.1
      entries:
      - name: ack-bedrockagent-controller.v0.1.1
        version: 0.1.1
      - name: ack-bedrockagent-controller.v0.1.0
        version: 0.1.0
      - name: ack-bedrockagent-controller.v0.0.6
        version: 0.0.6
      - name: ack-bedrockagent-controller.v0.0.5
        version: 0.0.5
      - name: ack-bedrockagent-controller.v0.0.4
        version: 0.0.4
      - name: ack-bedrockagent-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-bedrockagent-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: devopstales
      provider-url: devopstales.github.io
    name: community-trivy-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: community-trivy-operator.v2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "trivy-operator.devopstales.io/v1",
                "kind": "NamespaceScanner",
                "metadata": {
                  "name": "trivy-operator-main-config"
                },
                "spec": {
                  "crontab": "*/5 * * * *",
                  "namespace_selector": "trivy-scan",
                  "clusterWide": "False",
                  "policyreport": "False",
                  "registry": [
                    {
                      "name": "docker.io",
                      "user": "",
                      "password": ""
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: devopstales/trivy-operator:2.4.1
          createdAt: "2022-07-11T15:00:00Z"
          description: Trivy Operator for scheduled imagescans and an Admission Control.
          repository: https://github.com/devopstales/trivy-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Confifuration for trivy-operator.
            displayName: NamespaceScanner
            kind: NamespaceScanner
            name: namespace-scanners.trivy-operator.devopstales.io
            version: v1
          - description: |
              VulnerabilityReport summarizes vulnerabilities in application dependencies and operating system packages
              built into container images.
            displayName: VulnerabilityReport
            kind: VulnerabilityReport
            name: vulnerabilityreports.trivy-operator.devopstales.io
            version: v1
        description: |
          Trivy Operator for scheduled imagescans and an Admission Control.

          ### Add Github token to extend your Github API Rate Limit
          The Operator use Github to download Trivy database. If your Github API Rate Limit Exceeded add your Github token to the operator deployment as GITHUB_TOKEN variable.

          ```yaml
          ...
          env:
          - name: GITHUB_TOKEN
            value: "githubToken"
          ```

          ### Scheduled Image scans
          Default every 5 minutes execute a scan script. It will get image list from all namespaces with the label `trivy-scan=true`, and then scan this images with trivy, finally we will get metrics on `http://[pod-ip]:9115/metrics`

          You can chane the schedule of the image scanning by setting the `crontab` variable in the namespace-scanner object.


          ### Cache Scan results
          From the 2.4 version you can use an external redis to store scanning results and speed up the scanning process.

          ```bash
          ...
          env:
          - name: REDIS_ENABLED
            value: "True"
          - name: REDIS_BACKEND
            value: redis://redis-cache:6379
          ```

          ### Persist Trivy database cache
          To Persist Trivy cache you need to create a pvc and mount to the Deployment under `/home/trivy-operator/trivy-cache`

          ```yaml
          ...
            volumeMounts:
            - name: cache
              mountPath: "/home/trivy-operator/trivy-cache"
          volumes:
          - name: cache
            persistentVolumeClaim:
              claimName: trivy-cache
          ```

          ### Trivy Image Validator
          The admission controller function can be configured as a ValidatingWebhook in a k8s cluster. Kubernetes will send requests to the admission server when a Pod creation is initiated. The admission controller checks the image using trivy if it is in a namespace with the lable `trivy-operator-validation=true`.

          To enable the admission controller you need to set the `clusterWide` variable to `true` in the namespace-scanner object.

          Enable the scheduled scan and the admission controller for a specific namespace:
          ```bash
          kubectl label namespaces test-apps trivy-scan=true
          # or
          kubectl label namespaces test-apps trivy-operator-validation=true
          ```

          You can define policy to the Admission Controller, by adding annotation to the pod trough the deployment:

          ```bash
          spec:
            ...
            template:
              metadata:
                annotations:
                  trivy.security.devopstales.io/medium: "5"
                  trivy.security.devopstales.io/low: "10"
                  trivy.security.devopstales.io/critical: "2"
          ...
          ```

          ### Monitorng
          Trivy operator has a prometheus endpoint wher yo can get the resoults of the namespace scans and the adission controller.

          ```bash
          curl -s http://10.43.179.39:9115/metrics | grep so_vulnerabilities

          # HELP so_vulnerabilities Container vulnerabilities
          # TYPE so_vulnerabilities gauge
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="UNKNOWN"} 0
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="LOW"} 23
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="MEDIUM"} 93
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="HIGH"} 76
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="CRITICAL"} 25
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="UNKNOWN"} 0
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="LOW"} 23
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="MEDIUM"} 88
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="HIGH"} 60
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="CRITICAL"} 8
          ```

          ```bash
          curl -s http://10.43.179.39:9115/metrics | grep ac_vulnerabilities

          # HELP ac_vulnerabilities Admission Controller vulnerabilities
          # TYPE ac_vulnerabilities gauge
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="UNKNOWN"} 0.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="LOW"} 83.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="MEDIUM"} 6.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="HIGH"} 6.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="CRITICAL"} 4.0
          ```

          ### Policy Report reation
          The PolicyReport object is a protopype object probosed by the Kubernetes policy work group. The Policy Report Custom Resource Definition (CRD) can be used as a common way to provide policy results to Kubernetes cluster administrators and users, using native tools.

          Add the PolicyReport CRDs to your cluster (v1alpha2):
          ```bash
          kubectl create -f https://github.com/kubernetes-sigs/wg-policy-prototypes/raw/master/policy-report/crd/v1alpha2/wgpolicyk8s.io_policyreports.yaml
          ```

          You can enable the Policy Report creatin by setting the `policyreport` to `True` in the NamespaceScanner object.

          This objects can be visualized by the Policy Reporter UI. The Policy Reporter UI is a monitoring and Observability Tool for the PolicyReport CRD with an optional UI. It is created by Kyverno. The main goal was a tool to visualize the resoluts of the Kyverno policies, but because it uses the PolicyReports CRD it can visualize the resoults of the trivy-operator scans.
        displayName: Community Trivy Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trivy
        - security
        links:
        - name: Documentation
          url: https://devopstales.github.io/trivy-operator
        - name: Blog
          url: https://devopstales.github.io
        maintainers:
        - email: devopstales@protonmail.com
          name: devopstales
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: devopstales
          url: devopstales.github.io
        relatedImages:
        - devopstales/trivy-operator:2.4.1
        version: 2.4.0
      entries:
      - name: community-trivy-operator.v2.4.0
        version: 2.4.0
      - name: community-trivy-operator.v2.3.0
        version: 2.3.0
      - name: community-trivy-operator.v2.2.0
        version: 2.2.0
      - name: community-trivy-operator.v2.1.1
        version: 2.1.1
      name: stable
    defaultChannel: stable
    packageName: community-trivy-operator
    provider:
      name: devopstales
      url: devopstales.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: www.ibm.com
    name: ibm-elasticsearch-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-elasticsearch-operator.v1.1.2667
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
          createdAt: "2025-01-30"
          description: Elasticsearch operator for use within Cloudpak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-elasticsearch-operator
          olm.skipRange: <0.0.0 >1.1.0
          operators.openshift.io/valid-subscription: '["IBM Elasticsearch"]'
          operators.operatorframework.io/internal-objects: '["elasticsearchclusters.elasticsearch.opencontent.ibm.com"]'
          operators.operatorframework.io/operator-type: non-standalone
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ElasticsearchCluster
            name: elasticsearchclusters.elasticsearch.opencontent.ibm.com
            version: v1
        description: |-
          # Introduction
          An Elasticsearch operator for use with IBM Cloudpak services.
          # Details
          ibm-elasticsearch-operator supports installing Elasticsearch major version 6 and 7. This operator supports two CRDs Elasticsearch and Elasticsearchdetails. Currently Elasticsearch is an abstraction of Elasticsearchdetails, but there is no difference between both the CRDs. Creating either of them provisions elasticsearch instances. This abstraction layer has been created to support any changes to the operator with out breaking the CRD/API for adopting products.

          ## Prerequisites
          - Openshift 3.11+
          ## Resources Required
          In short the answer is it depends. Especially if you are running Elasticsearch 6 you need 3 or a greater odd number of replicas to make a cluster and not have split brain. Each replica will have a sidecar that needs 0.1 cpu and 2gb of ram. You should size Elasticsearch based on your use-case. The default is a full cpu and 4 cpu and 64 gb ram. To assist sizing requirements we have included a copy of [esrally](https://esrally.readthedocs.io/en/stable/) on all the docker images.
          ## Installing
          ElasticsearchCluster is the strategic Kind moving forward. There are two flavors of CRD supported that will soon be deprecated. A simple Elasticseach "All in one" crd and a more detailed CRD for teams who need greater control. Upon deploy the elasticsearch is available from the service (not the headless) created by the operator. Only https will be available on port 443. the service will also be in the ElasticsearchDetails status.

          ### Case install
          To install the component using case you must either use the native install or catalog install. Upon installing the catalog you can then use either OLM or ODLM to install but must create the OperatorGroup & Subscription for either method to avoid conflict.

          # Configuration
          Architecture:
            - x86
            - s390x
            - ppc64le
          Storage: RWO Encrypted Block Storage with 120+ IOPS/GB
          Clouds:
          - IBM CLoud (ibm-block-gold, Gen2 custom iops storage class)
          - AWS (EBS)
          - Azure (For Azure please ensure you are using a higher tier of AzureDisk storage)
          - OCS (RBD not Ceph)
          - Portworx (Non-shared replacated)
          - Please do not use with NFS

          ### Limitations
          This product is not deployable standalone and it is part of the LICCR already accepted
          ## SecurityContextConstraints Requirements
          This component requires a restricted SCC.
        displayName: IBM OpenContent Elasticsearch
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - internal
        - Elasticsearch
        maintainers:
        - email: noreply@ibm.com
          name: IBM CloudpakOpen Community
        maturity: alpha
        provider:
          name: IBM
          url: www.ibm.com
        relatedImages:
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:47a3745c78280f98f2ee352bb9192cdcd2fcd6d4c1a3627e7aa42178a0f107fd
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:30a5b2c4e3c231373fa18d8aaab16d3505cc11a8a5d4be36eaf99d7c0dee35f2
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:2668a562eb8df1117c40acbc3600d6735469d210e5aae4aa03bf9f54d0ebfe46
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:2c39ff1850b7a9d32a11e99393039ef74ab363f3fffec1e4fd542f2ae148a6b4
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:130872119947fe91cbfb8dd82b3baaf78f7adcaca5933ca394eca3f9ee84c8cc
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
        version: 1.1.2667
      entries:
      - name: ibm-elasticsearch-operator.v1.1.2667
        version: 1.1.2667
      name: v1.1
    defaultChannel: v1.1
    packageName: ibm-elasticsearch-operator
    provider:
      name: IBM
      url: www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: ovms-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openvino-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "ModelServer",
                "metadata": {
                  "name": "model-server-sample"
                },
                "spec": {
                  "image_name": "registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0",
                  "deployment_parameters": {
                    "replicas": 1
                  },
                  "service_parameters": {
                    "grpc_port": 8080,
                    "rest_port": 8081
                  },
                  "models_settings": {
                    "single_model_mode": true,
                    "config_configmap_name": "",
                    "config_path": "",
                    "model_name": "resnet",
                    "model_path": "gs://<bucket_name>/<model_dir>",
                    "nireq": 0,
                    "plugin_config": "{\"PERFORMANCE_HINT\":\"LATENCY\"}",
                    "batch_size": "",
                    "shape": "",
                    "model_version_policy": "{\"latest\": { \"num_versions\":1 }}",
                    "layout": "",
                    "target_device": "CPU",
                    "is_stateful": false,
                    "idle_sequence_cleanup": false,
                    "low_latency_transformation": true,
                    "max_sequence_number": 0
                  },
                  "server_settings": {
                    "file_system_poll_wait_seconds": 0,
                    "sequence_cleaner_poll_wait_minutes": 0,
                    "log_level": "INFO",
                    "grpc_workers": 1,
                    "rest_workers": 0
                  },
                  "models_repository": {
                    "https_proxy": "",
                    "http_proxy": "",
                    "storage_type": "google",
                    "models_host_path": "",
                    "models_volume_claim": "",
                    "aws_secret_access_key": "",
                    "aws_access_key_id": "",
                    "aws_region": "",
                    "s3_compat_api_endpoint": "",
                    "gcp_creds_secret_name": "",
                    "azure_storage_connection_string": ""
                  }
                }
              },
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "Notebook",
                "metadata": {
                  "name": "v2024.1"
                },
                "spec": {
                  "git_ref": "2024.1",
                  "auto_update_image": false
                }
              }
            ]
          alm-examples-metadata: '{ "model-server-sample":{"description":"Example
            of the model server configuration"}, "v2024":{"description":"Default notebook
            configuration"} }'
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
          createdAt: "2024-05-14T09:29:35Z"
          description: OpenVINO Toolkit Operator manages OpenVINO components in OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.34.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Provides inference serving through gRPC & REST
            displayName: Model Server
            kind: ModelServer
            name: modelservers.intel.com
            version: v1alpha1
          - description: Adds a collection of ready-to-run Jupyter notebooks for learning
              and experimenting with the OpenVINO™ Toolkit to Red Hat® OpenShift®
              Data Science
            displayName: Notebook
            kind: Notebook
            name: notebooks.intel.com
            version: v1alpha1
        description: |
          OpenVINO Toolkit Operator manages OpenVINO components in OpenShift.

          Currently there available components are ModelServer and Notebook.

          # Model Server
          [OpenVINO™ Model Server](https://github.com/openvinotoolkit/model_server) (OVMS) is a scalable, high-performance solution for serving machine learning models optimized for Intel® architectures. The server provides an inference service via gRPC or REST API for any models trained in a framework that is supported by [OpenVINO](https://software.intel.com/en-us/openvino-toolki://docs.openvino.ai/latest/index.html).
          Model Server configuration parameters is explained [here](https://github.com/openvinotoolkit/operator/blob/main/docs/modelserver_params.md).
          ## Using the cluster
          OpenVINO Model Server can be consumed as a `Service`. It is called like with `ModelServer` resource with `-ovms` suffix.
          The suffix is omitted when `ovms` phrase is included in the name.
          The service exposes gRPC and REST API interfaces to run inference requests.
          ```
          oc get pods
          NAME                                        READY   STATUS    RESTARTS   AGE
          model-server-sample-ovms-586f6f76df-dpps4   1/1     Running   0          8h

          oc get services
          NAME                       TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)             AGE
          model-server-sample-ovms   ClusterIP   172.25.199.210   <none>        8080/TCP,8081/TCP   8h
          ```

          # Notebook
          The Notebook resource integrates JupyterHub from OpenShift Data Science or Open Data Hub with a container image that includes [developer
          tools](https://github.com/openvinotoolkit/openvino/blob/master/docs/install_guides/pypi-openvino-dev.md)
          from the OpenVINO toolkit and a set of Jupyter notebook tutorials. It enables selecting a defined image `OpenVINO™ Toolkit` from
          the Jupyter Spawner choice list.

          Create the `Notebook` resource in the same project with JupyterHub and RedHat OpenShift Data Science operator as seen in [documentation](https://github.com/openvinotoolkit/operator/blob/main/docs/notebook_in_rhods.md).
          It builds the image in the cluster based on Dockerfile from [openvino_notebooks](https://github.com/openvinotoolkit/openvino_notebooks).

          ## References
          OpenVINO Model Server on [Github](https://github.com/openvinotoolkit/model_server)

          OpenVINO Model Server Operator on [Github](https://github.com/openvinotoolkit/operator)
        displayName: OpenVINO Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Inference OpenVINO
        links:
        - name: OpenVINO Toolkit Operator
          url: https://github.com/openvinotoolkit/operator
        - name: OpenVINO Model Server
          url: https://github.com/openvinotoolkit/model_server
        - name: OpenVINO Notebooks
          url: https://github.com/openvinotoolkit/openvino_notebooks
        maintainers:
        - email: openvino-operator-support@intel.com
          name: Operator Support
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5714b93ac39277affdc10c3ba142d8c032ed0ddb2ba61c04cbddac710464c4
        - registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
        - registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0
        version: 1.2.0
      entries:
      - name: openvino-operator.v1.2.0
        version: 1.2.0
      - name: openvino-operator.v1.1.0
        version: 1.1.0
      - name: openvino-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: ovms-operator
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx-openshift-operator.v0.0.33
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
          createdAt: "2019-07-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx-operator-rhmp/support?utm_source=openshift_console
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX on JBoss EAP 7
            displayName: Joget DX on JBoss EAP 7
            kind: JogetDX
            name: jogetdx.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX7/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
        - registry.connect.redhat.com/joget/joget-dx7-eap7@sha256:7245501f5fccf011d75c034b7d55542cfe3bd2c64cc7e48b3224a48b623af282
        version: 0.0.33
      entries:
      - name: joget-dx-openshift-operator.v0.0.33
        version: 0.0.33
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx-operator-rhmp
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-codeartifact-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-codeartifact-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/codeartifact-controller:1.1.1
          createdAt: "2025-09-19T22:00:24Z"
          description: AWS codeartifact controller is a service controller for managing
            codeartifact resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Domain represents the state of an AWS codeartifact Domain
              resource.
            displayName: Domain
            kind: Domain
            name: domains.codeartifact.services.k8s.aws
            version: v1alpha1
          - description: PackageGroup represents the state of an AWS codeartifact
              PackageGroup resource.
            displayName: PackageGroup
            kind: PackageGroup
            name: packagegroups.codeartifact.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon codeartifact resources in AWS from within your Kubernetes cluster.

          **About Amazon codeartifact**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon codeartifact
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - codeartifact
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon codeartifact Developer Resources
          url: https://aws.amazon.com/codeartifact/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: codeartifact maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/codeartifact-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-codeartifact-controller.v1.1.1
        version: 1.1.1
      - name: ack-codeartifact-controller.v1.1.0
        version: 1.1.0
      - name: ack-codeartifact-controller.v1.0.13
        version: 1.0.13
      - name: ack-codeartifact-controller.v1.0.11
        version: 1.0.11
      name: alpha
    defaultChannel: alpha
    packageName: ack-codeartifact-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kepler Operator Contributors
      provider-url: https://sustainable-computing.io/
    name: kepler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kepler-operator.v0.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "Kepler",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "kepler",
                    "app.kubernetes.io/name": "kepler",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "kepler"
                },
                "spec": {
                  "exporter": {
                    "deployment": {
                      "port": 9103
                    },
                    "redfish": null
                  }
                }
              },
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "PowerMonitor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "powermonitor",
                    "app.kubernetes.io/name": "powermonitor",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "power-monitor"
                },
                "spec": {
                  "kepler": {
                    "config": {
                      "logLevel": "info"
                    },
                    "deployment": {
                      "security": {
                        "mode": "none"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: quay.io/sustainable_computing_io/kepler-operator:0.20.0
          createdAt: "2025-07-18T06:13:30Z"
          description: Deploys and Manages Kepler on Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: |-
            [
              "keplerinternals.kepler.system.sustainable.computing.io",
              "powermonitorinternals.kepler.system.sustainable.computing.io"
            ]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/sustainable-computing-io/kepler-operator
          support: https://github.com/sustainable-computing-io/kepler-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KeplerInternal is the Schema for internal/unsupported API
            displayName: KeplerInternal
            kind: KeplerInternal
            name: keplerinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: Kepler is deprecated and will be removed in a future release.
              Please use PowerMonitor instead.
            displayName: Kepler
            kind: Kepler
            name: keplers.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitorInternal is the Schema for internal/unsupported
              API
            displayName: PowerMonitorInternal
            kind: PowerMonitorInternal
            name: powermonitorinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitor is the Schema for the powermonitors API
            displayName: PowerMonitor
            kind: PowerMonitor
            name: powermonitors.kepler.system.sustainable.computing.io
            version: v1alpha1
        description: |-
          Kepler (Kubernetes-based Efficient Power Level Exporter) is a Prometheus exporter that
          measures energy consumption metrics at the container, pod, and node level in Kubernetes clusters.

          ### Deprecation Notice
          **Important:** The Kepler CR (Custom Resource) is deprecated and will be removed in a future release. Please use the new PowerMonitor instead.

          ### Prerequisites
          To view the metrics exposed by Kepler, **[User Workload Monitoring](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/monitoring/configuring-user-workload-monitoring)**
          MUST be enabled in your cluster.

          ### Documentation
          * [User Documentation](https://github.com/sustainable-computing-io/kepler-operator/blob/v1alpha1/README.md)
          * [Developer Guide](https://github.com/sustainable-computing-io/kepler-operator/tree/v1alpha1/docs/developer)

          ### Contributing
          * [Report Bugs](https://github.com/sustainable-computing-io/kepler-operator/issues/new?template=bug_report.md) - Found a bug? Let us know!
          * [Browse Issues](https://github.com/sustainable-computing-io/kepler-operator/issues) - See what we're working on

          ### License
          Licensed under [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)
        displayName: Kepler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kepler
        - power
        - energy
        - monitoring
        links:
        - name: Installation Guide
          url: https://sustainable-computing.io/installation/community-operator/
        - name: Kepler Project
          url: https://sustainable-computing.io/
        maintainers:
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: vprashar@redhat.com
          name: Vibhu Prashar
        - email: vimalkum@redhat.com
          name: Vimal Kumar
        - email: kaliu@redhat.com
          name: Kaiyi Liu
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Kepler Operator Contributors
          url: https://sustainable-computing.io/
        relatedImages:
        - quay.io/sustainable_computing_io/kepler-operator:0.20.0
        - quay.io/sustainable_computing_io/kepler:v0.10.2
        - quay.io/brancz/kube-rbac-proxy:v0.19.0
        version: 0.20.0
      entries:
      - name: kepler-operator.v0.20.0
        version: 0.20.0
      - name: kepler-operator.v0.19.0
        version: 0.19.0
      - name: kepler-operator.v0.18.0
        version: 0.18.0
      - name: kepler-operator.v0.17.0
        version: 0.17.0
      - name: kepler-operator.v0.16.0
        version: 0.16.0
      - name: kepler-operator.v0.15.0
        version: 0.15.0
      name: alpha
    defaultChannel: alpha
    packageName: kepler-operator
    provider:
      name: Kepler Operator Contributors
      url: https://sustainable-computing.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MuleSoft
      provider-url: anypoint.mulesoft.com
    name: runtime-fabric-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runtime-fabric-operator.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rtf.operators.mulesoft.com/v1",
                "kind": "RuntimeFabric",
                "metadata": {
                  "name": "runtime-fabric"
                },
                "spec": {
                  "activationData": null,
                  "agent": {
                    "replicas": 1,
                    "app": {
                      "resources": {
                        "limits": {
                          "cpu": "1000m",
                          "memory": "500Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      }
                    },
                    "rtfd": {
                      "resources": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "150m",
                          "memory": "200Mi"
                        }
                      },
                      "watch": {
                        "enabled": true,
                        "components": {
                          "deployments": {
                            "enabled": true
                          },
                          "ingress": {
                            "enabled": true
                          },
                          "secrets": {
                            "enabled": true
                          },
                          "persistencegateways": {
                            "enabled": true
                          },
                          "nodes": {
                            "enabled": true
                          },
                          "jobs": {
                            "enabled": true
                          },
                          "routes": {
                            "enabled": true
                          }
                        }
                      }
                    }
                  },
                  "resource-cache": {
                    "replicas": 1,
                    "nginx": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "250Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    },
                    "fetcher": {
                      "resources": {
                        "limits": {
                          "cpu": "50m",
                          "memory": "100Mi"
                        },
                        "requests": {
                          "cpu": "50m",
                          "memory": "50Mi"
                        }
                      }
                    }
                  },
                  "mule-clusterip-service": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "global": {
                    "cluster": {
                      "flavor": "byok",
                      "type": "openshift"
                    },
                    "containerLogPaths": [
                      "/var/lib/docker/containers",
                      "/var/log/containers",
                      "/var/log/pods"
                    ],
                    "core": {
                      "affinity": {},
                      "tolerations": [],
                      "topologySpreadConstraints": [
                        {
                          "maxSkew": 1,
                          "topologyKey": "topology.kubernetes.io/zone",
                          "whenUnsatisfiable": "ScheduleAnyway"
                        }
                      ]
                    },
                    "worker": {
                      "affinity": {},
                      "tolerations": []
                    },
                    "deployment": {
                      "nodeSelector": ""
                    },
                    "image": {
                      "provider": "aws",
                      "pullSecretName": "rtf-pull-secret",
                      "rtfRegistry": "rtf-runtime-registry.kprod.msap.io",
                      "repositoryNamespace": "mulesoft"
                    },
                    "service": {
                      "clusterIp": null
                    },
                    "authorizedNamespaces": false,
                    "fipsEnabled": false,
                    "crds": {
                      "install": true,
                      "keep": true,
                      "flex": {
                        "enabled": false
                      }
                    },
                    "nodeWatcherEnabled": true,
                    "nodeReconciliation": true,
                    "deploymentRateLimitPerSecond": 1
                  },
                  "muleLicense": null,
                  "proxy": {
                    "http_no_proxy": null,
                    "http_proxy": null,
                    "monitoring_proxy": null
                  },
                  "customLog4jEnabled": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
          createdAt: "2025-07-31T21:12:41Z"
          description: Runtime Fabric Operator Application from Mulesoft
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mulesoft/rtf-ocp-operator
          support: MuleSoft
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Runtime Fabric Operator from Mulesoft
              **Activation data**
              Used to install Runtime Fabric on a Kubernetes service. Copy this value from the Runtime Fabric setup page in Runtime Manager.
              **Registry URL**
              The registry from which to retrieve the images needed to install and run Runtime Fabric. If you're using a local registry, specify the URL here.
                * Default registry URLs:
                    - US control plane: rtf-runtime-registry.kprod.msap.io
                    - EU control plane: rtf-runtime-registry.kprod-eu.msap.io
              **Registry pull secret**
              Used to retrieve images from the registry. Change this value only if you did so during the operator installation.
              **Mule license key**
              Your company’s Mule license key. The key is not required during install, but it is required to deploy applications to Runtime Fabric. Must be Base64 encoded.
            displayName: Runtime Fabric
            kind: RuntimeFabric
            name: runtimefabrics.rtf.operators.mulesoft.com
            version: v1
        description: |-
          Anypoint Runtime Fabric is a container service that you install in your existing Red Hat OpenShift cluster that enables you to run Mule applications, composite APIs, and API gateways across any environment with centralized management via Anypoint Platform.
          Capabilities of Runtime Fabric include:
          * Isolation between applications by running a separate Mule runtime server per application.
          * Ability to run multiple versions of Mule runtime server on the same set of resources.
          * Scaling applications across multiple replicas.
          * Automated application fail-over.
          * Application management with Anypoint Runtime Manager.
          Runtime Fabric supports the following Red Hat OpenShift deployment options:
          * Red Hat OpenShift Service on AWS
          * Microsoft Azure Red Hat OpenShift
          * Red Hat OpenShift Dedicated
          * Red Hat OpenShift on IBM Cloud
          * Self-managed Red Hat OpenShift editions (Performance Plus, OCP, Kubernetes engine)
          Runtime Fabric supports Red Hat OpenShift versions 4.16 to 4.19.
        displayName: Runtime Fabric Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - agent
        - rtf
        - runtime
        - fabric
        - mulesoft
        links:
        - name: Runtime Fabric Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/62901512fefb1b137ba23946
        - name: Documentation
          url: https://docs.mulesoft.com/release-notes/runtime-fabric/runtime-fabric
        maintainers:
        - email: support@mulesoft.com
          name: Mulesoft
        maturity: stable
        provider:
          name: MuleSoft
          url: anypoint.mulesoft.com
        relatedImages:
        - quay.io/rtf/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
        version: 2.11.0
      entries:
      - name: runtime-fabric-operator.v2.11.0
        version: 2.11.0
      - name: runtime-fabric-operator.v2.10.34
        version: 2.10.34
      - name: runtime-fabric-operator.v2.10.17
        version: 2.10.17
      - name: runtime-fabric-operator.v2.10.0
        version: 2.10.0
      - name: runtime-fabric-operator.v2.9.2
        version: 2.9.2
      - name: runtime-fabric-operator.v2.8.2
        version: 2.8.2
      - name: runtime-fabric-operator.v2.7.0
        version: 2.7.0
      - name: runtime-fabric-operator.v2.6.52
        version: 2.6.52
      name: alpha
    - currentCSV: runtime-fabric-operator.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rtf.operators.mulesoft.com/v1",
                "kind": "RuntimeFabric",
                "metadata": {
                  "name": "runtime-fabric"
                },
                "spec": {
                  "activationData": null,
                  "agent": {
                    "replicas": 1,
                    "app": {
                      "resources": {
                        "limits": {
                          "cpu": "1000m",
                          "memory": "500Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      }
                    },
                    "rtfd": {
                      "resources": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "150m",
                          "memory": "200Mi"
                        }
                      },
                      "watch": {
                        "enabled": true,
                        "components": {
                          "deployments": {
                            "enabled": true
                          },
                          "ingress": {
                            "enabled": true
                          },
                          "secrets": {
                            "enabled": true
                          },
                          "persistencegateways": {
                            "enabled": true
                          },
                          "nodes": {
                            "enabled": true
                          },
                          "jobs": {
                            "enabled": true
                          },
                          "routes": {
                            "enabled": true
                          }
                        }
                      }
                    }
                  },
                  "resource-cache": {
                    "replicas": 1,
                    "nginx": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "250Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    },
                    "fetcher": {
                      "resources": {
                        "limits": {
                          "cpu": "50m",
                          "memory": "100Mi"
                        },
                        "requests": {
                          "cpu": "50m",
                          "memory": "50Mi"
                        }
                      }
                    }
                  },
                  "mule-clusterip-service": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "global": {
                    "cluster": {
                      "flavor": "byok",
                      "type": "openshift"
                    },
                    "containerLogPaths": [
                      "/var/lib/docker/containers",
                      "/var/log/containers",
                      "/var/log/pods"
                    ],
                    "core": {
                      "affinity": {},
                      "tolerations": [],
                      "topologySpreadConstraints": [
                        {
                          "maxSkew": 1,
                          "topologyKey": "topology.kubernetes.io/zone",
                          "whenUnsatisfiable": "ScheduleAnyway"
                        }
                      ]
                    },
                    "worker": {
                      "affinity": {},
                      "tolerations": []
                    },
                    "deployment": {
                      "nodeSelector": ""
                    },
                    "image": {
                      "provider": "aws",
                      "pullSecretName": "rtf-pull-secret",
                      "rtfRegistry": "rtf-runtime-registry.kprod.msap.io",
                      "repositoryNamespace": "mulesoft"
                    },
                    "service": {
                      "clusterIp": null
                    },
                    "authorizedNamespaces": false,
                    "fipsEnabled": false,
                    "crds": {
                      "install": true,
                      "keep": true,
                      "flex": {
                        "enabled": false
                      }
                    },
                    "nodeWatcherEnabled": true,
                    "nodeReconciliation": true,
                    "deploymentRateLimitPerSecond": 1
                  },
                  "muleLicense": null,
                  "proxy": {
                    "http_no_proxy": null,
                    "http_proxy": null,
                    "monitoring_proxy": null
                  },
                  "customLog4jEnabled": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
          createdAt: "2025-07-31T21:12:41Z"
          description: Runtime Fabric Operator Application from Mulesoft
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mulesoft/rtf-ocp-operator
          support: MuleSoft
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Runtime Fabric Operator from Mulesoft
              **Activation data**
              Used to install Runtime Fabric on a Kubernetes service. Copy this value from the Runtime Fabric setup page in Runtime Manager.
              **Registry URL**
              The registry from which to retrieve the images needed to install and run Runtime Fabric. If you're using a local registry, specify the URL here.
                * Default registry URLs:
                    - US control plane: rtf-runtime-registry.kprod.msap.io
                    - EU control plane: rtf-runtime-registry.kprod-eu.msap.io
              **Registry pull secret**
              Used to retrieve images from the registry. Change this value only if you did so during the operator installation.
              **Mule license key**
              Your company’s Mule license key. The key is not required during install, but it is required to deploy applications to Runtime Fabric. Must be Base64 encoded.
            displayName: Runtime Fabric
            kind: RuntimeFabric
            name: runtimefabrics.rtf.operators.mulesoft.com
            version: v1
        description: |-
          Anypoint Runtime Fabric is a container service that you install in your existing Red Hat OpenShift cluster that enables you to run Mule applications, composite APIs, and API gateways across any environment with centralized management via Anypoint Platform.
          Capabilities of Runtime Fabric include:
          * Isolation between applications by running a separate Mule runtime server per application.
          * Ability to run multiple versions of Mule runtime server on the same set of resources.
          * Scaling applications across multiple replicas.
          * Automated application fail-over.
          * Application management with Anypoint Runtime Manager.
          Runtime Fabric supports the following Red Hat OpenShift deployment options:
          * Red Hat OpenShift Service on AWS
          * Microsoft Azure Red Hat OpenShift
          * Red Hat OpenShift Dedicated
          * Red Hat OpenShift on IBM Cloud
          * Self-managed Red Hat OpenShift editions (Performance Plus, OCP, Kubernetes engine)
          Runtime Fabric supports Red Hat OpenShift versions 4.16 to 4.19.
        displayName: Runtime Fabric Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - agent
        - rtf
        - runtime
        - fabric
        - mulesoft
        links:
        - name: Runtime Fabric Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/62901512fefb1b137ba23946
        - name: Documentation
          url: https://docs.mulesoft.com/release-notes/runtime-fabric/runtime-fabric
        maintainers:
        - email: support@mulesoft.com
          name: Mulesoft
        maturity: stable
        provider:
          name: MuleSoft
          url: anypoint.mulesoft.com
        relatedImages:
        - quay.io/rtf/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
        version: 2.11.0
      entries:
      - name: runtime-fabric-operator.v2.11.0
        version: 2.11.0
      - name: runtime-fabric-operator.v2.10.34
        version: 2.10.34
      - name: runtime-fabric-operator.v2.10.17
        version: 2.10.17
      - name: runtime-fabric-operator.v2.10.0
        version: 2.10.0
      - name: runtime-fabric-operator.v2.9.2
        version: 2.9.2
      - name: runtime-fabric-operator.v2.8.2
        version: 2.8.2
      - name: runtime-fabric-operator.v2.7.0
        version: 2.7.0
      - name: runtime-fabric-operator.v2.6.52
        version: 2.6.52
      name: stable
    defaultChannel: stable
    packageName: runtime-fabric-operator
    provider:
      name: MuleSoft
      url: anypoint.mulesoft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-dynamodb-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-dynamodb-controller.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "GlobalTable",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "Table",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/dynamodb-controller:1.5.2
          createdAt: "2025-09-19T19:19:47Z"
          description: AWS DynamoDB controller is a service controller for managing
            DynamoDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup represents the state of an AWS dynamodb Backup resource.
            displayName: Backup
            kind: Backup
            name: backups.dynamodb.services.k8s.aws
            version: v1alpha1
          - description: GlobalTable represents the state of an AWS dynamodb GlobalTable
              resource.
            displayName: GlobalTable
            kind: GlobalTable
            name: globaltables.dynamodb.services.k8s.aws
            version: v1alpha1
          - description: Table represents the state of an AWS dynamodb Table resource.
            displayName: Table
            kind: Table
            name: tables.dynamodb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon DynamoDB resources in AWS from within your Kubernetes cluster.

          **About Amazon DynamoDB**

          Amazon DynamoDB is a fully managed NoSQL database service that provides fast and predictable performance with seamless scalability. DynamoDB lets you offload the administrative burdens of operating and scaling a distributed database so that you don't have to worry about hardware provisioning, setup and configuration, replication, software patching, or cluster scaling. DynamoDB also offers encryption at rest, which eliminates the operational burden and complexity involved in protecting sensitive data. For more information, see [DynamoDB Encryption at Rest](https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/EncryptionAtRest.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon DynamoDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dynamodb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon DynamoDB Developer Resources
          url: https://aws.amazon.com/dynamodb/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: dynamodb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/dynamodb-controller:1.5.2
        version: 1.5.2
      entries:
      - name: ack-dynamodb-controller.v1.5.2
        version: 1.5.2
      - name: ack-dynamodb-controller.v1.5.1
        version: 1.5.1
      - name: ack-dynamodb-controller.v1.5.0
        version: 1.5.0
      - name: ack-dynamodb-controller.v1.4.10
        version: 1.4.10
      - name: ack-dynamodb-controller.v1.4.9
        version: 1.4.9
      - name: ack-dynamodb-controller.v1.4.7
        version: 1.4.7
      - name: ack-dynamodb-controller.v1.4.6
        version: 1.4.6
      - name: ack-dynamodb-controller.v1.4.5
        version: 1.4.5
      - name: ack-dynamodb-controller.v1.4.4
        version: 1.4.4
      - name: ack-dynamodb-controller.v1.4.3
        version: 1.4.3
      - name: ack-dynamodb-controller.v1.4.2
        version: 1.4.2
      - name: ack-dynamodb-controller.v1.4.1
        version: 1.4.1
      - name: ack-dynamodb-controller.v1.3.0
        version: 1.3.0
      - name: ack-dynamodb-controller.v1.2.23
        version: 1.2.23
      - name: ack-dynamodb-controller.v1.2.22
        version: 1.2.22
      - name: ack-dynamodb-controller.v1.2.21
        version: 1.2.21
      - name: ack-dynamodb-controller.v1.2.20
        version: 1.2.20
      - name: ack-dynamodb-controller.v1.2.15
        version: 1.2.15
      - name: ack-dynamodb-controller.v1.2.14
        version: 1.2.14
      - name: ack-dynamodb-controller.v1.2.13
        version: 1.2.13
      - name: ack-dynamodb-controller.v1.2.12
        version: 1.2.12
      - name: ack-dynamodb-controller.v1.2.11
        version: 1.2.11
      - name: ack-dynamodb-controller.v1.2.9
        version: 1.2.9
      - name: ack-dynamodb-controller.v1.2.8
        version: 1.2.8
      - name: ack-dynamodb-controller.v1.2.6
        version: 1.2.6
      - name: ack-dynamodb-controller.v1.2.5
        version: 1.2.5
      - name: ack-dynamodb-controller.v1.2.4
        version: 1.2.4
      - name: ack-dynamodb-controller.v1.2.2
        version: 1.2.2
      - name: ack-dynamodb-controller.v1.2.1
        version: 1.2.1
      - name: ack-dynamodb-controller.v1.2.0
        version: 1.2.0
      - name: ack-dynamodb-controller.v1.1.1
        version: 1.1.1
      - name: ack-dynamodb-controller.v1.1.0
        version: 1.1.0
      - name: ack-dynamodb-controller.v1.0.3
        version: 1.0.3
      - name: ack-dynamodb-controller.v1.0.2
        version: 1.0.2
      - name: ack-dynamodb-controller.v1.0.1
        version: 1.0.1
      - name: ack-dynamodb-controller.v1.0.0
        version: 1.0.0
      - name: ack-dynamodb-controller.v0.1.10
        version: 0.1.10
      - name: ack-dynamodb-controller.v0.1.9
        version: 0.1.9
      - name: ack-dynamodb-controller.v0.1.8
        version: 0.1.8
      - name: ack-dynamodb-controller.v0.1.7
        version: 0.1.7
      - name: ack-dynamodb-controller.v0.1.6
        version: 0.1.6
      - name: ack-dynamodb-controller.v0.1.5
        version: 0.1.5
      - name: ack-dynamodb-controller.v0.1.4
        version: 0.1.4
      - name: ack-dynamodb-controller.v0.1.3
        version: 0.1.3
      - name: ack-dynamodb-controller.v0.1.1
        version: 0.1.1
      - name: ack-dynamodb-controller.v0.1.0
        version: 0.1.0
      - name: ack-dynamodb-controller.v0.0.20
        version: 0.0.20
      - name: ack-dynamodb-controller.v0.0.19
        version: 0.0.19
      - name: ack-dynamodb-controller.v0.0.18
        version: 0.0.18
      - name: ack-dynamodb-controller.v0.0.17
        version: 0.0.17
      - name: ack-dynamodb-controller.v0.0.16
        version: 0.0.16
      - name: ack-dynamodb-controller.v0.0.15
        version: 0.0.15
      - name: ack-dynamodb-controller.v0.0.14
        version: 0.0.14
      - name: ack-dynamodb-controller.v0.0.13
        version: 0.0.13
      - name: ack-dynamodb-controller.v0.0.12
        version: 0.0.12
      - name: ack-dynamodb-controller.v0.0.11
        version: 0.0.11
      - name: ack-dynamodb-controller.v0.0.10
        version: 0.0.10
      - name: ack-dynamodb-controller.v0.0.9
        version: 0.0.9
      - name: ack-dynamodb-controller.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: ack-dynamodb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-elasticache-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-elasticache-controller.v1.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "CacheParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "CacheSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "ReplicationGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "Snapshot",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "UserGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/elasticache-controller:1.2.3
          createdAt: "2025-09-19T19:19:33Z"
          description: AWS ElastiCache controller is a service controller for managing
            ElastiCache resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CacheCluster represents the state of an AWS elasticache CacheCluster
              resource.
            displayName: CacheCluster
            kind: CacheCluster
            name: cacheclusters.elasticache.services.k8s.aws
            version: v1alpha1
          - description: CacheParameterGroup represents the state of an AWS elasticache
              CacheParameterGroup resource.
            displayName: CacheParameterGroup
            kind: CacheParameterGroup
            name: cacheparametergroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: CacheSubnetGroup represents the state of an AWS elasticache
              CacheSubnetGroup resource.
            displayName: CacheSubnetGroup
            kind: CacheSubnetGroup
            name: cachesubnetgroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ReplicationGroup represents the state of an AWS elasticache
              ReplicationGroup resource.
            displayName: ReplicationGroup
            kind: ReplicationGroup
            name: replicationgroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCache represents the state of an AWS elasticache
              ServerlessCache resource.
            displayName: ServerlessCache
            kind: ServerlessCache
            name: serverlesscaches.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCacheSnapshot represents the state of an AWS elasticache
              ServerlessCacheSnapshot resource.
            displayName: ServerlessCacheSnapshot
            kind: ServerlessCacheSnapshot
            name: serverlesscachesnapshots.elasticache.services.k8s.aws
            version: v1alpha1
          - description: Snapshot represents the state of an AWS elasticache Snapshot
              resource.
            displayName: Snapshot
            kind: Snapshot
            name: snapshots.elasticache.services.k8s.aws
            version: v1alpha1
          - description: UserGroup represents the state of an AWS elasticache UserGroup
              resource.
            displayName: UserGroup
            kind: UserGroup
            name: usergroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS elasticache User resource.
            displayName: User
            kind: User
            name: users.elasticache.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ElastiCache resources in AWS from within your Kubernetes cluster.

          **About Amazon ElastiCache**

          Amazon ElastiCache allows you to seamlessly set up, run, and scale popular open-source compatible in-memory data stores in the cloud. Build data-intensive apps or boost the performance of your existing databases by retrieving data from high throughput and low latency in-memory data stores. Amazon ElastiCache is a popular choice for real-time use cases like Caching, Session Stores, Gaming, Geospatial Services, Real-Time Analytics, and Queuing.

          Amazon ElastiCache offers fully managed Redis, voted the most loved database by developers in the Stack Overflow 2020 Developer Survey, and Memcached for your most demanding applications that require sub-millisecond response times.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ElastiCache
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticache
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ElastiCache Developer Resources
          url: https://aws.amazon.com/elasticache/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: elasticache maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/elasticache-controller:1.2.3
        version: 1.2.3
      entries:
      - name: ack-elasticache-controller.v1.2.3
        version: 1.2.3
      - name: ack-elasticache-controller.v1.2.2
        version: 1.2.2
      - name: ack-elasticache-controller.v1.2.1
        version: 1.2.1
      - name: ack-elasticache-controller.v1.2.0
        version: 1.2.0
      - name: ack-elasticache-controller.v1.1.0
        version: 1.1.0
      - name: ack-elasticache-controller.v1.0.7
        version: 1.0.7
      - name: ack-elasticache-controller.v1.0.6
        version: 1.0.6
      - name: ack-elasticache-controller.v1.0.5
        version: 1.0.5
      - name: ack-elasticache-controller.v1.0.4
        version: 1.0.4
      - name: ack-elasticache-controller.v1.0.3
        version: 1.0.3
      - name: ack-elasticache-controller.v1.0.2
        version: 1.0.2
      - name: ack-elasticache-controller.v1.0.1
        version: 1.0.1
      - name: ack-elasticache-controller.v1.0.0
        version: 1.0.0
      - name: ack-elasticache-controller.v0.2.6
        version: 0.2.6
      - name: ack-elasticache-controller.v0.2.5
        version: 0.2.5
      - name: ack-elasticache-controller.v0.2.4
        version: 0.2.4
      - name: ack-elasticache-controller.v0.2.3
        version: 0.2.3
      - name: ack-elasticache-controller.v0.2.2
        version: 0.2.2
      - name: ack-elasticache-controller.v0.2.1
        version: 0.2.1
      - name: ack-elasticache-controller.v0.2.0
        version: 0.2.0
      - name: ack-elasticache-controller.v0.1.1
        version: 0.1.1
      - name: ack-elasticache-controller.v0.1.0
        version: 0.1.0
      - name: ack-elasticache-controller.v0.0.31
        version: 0.0.31
      - name: ack-elasticache-controller.v0.0.30
        version: 0.0.30
      - name: ack-elasticache-controller.v0.0.29
        version: 0.0.29
      - name: ack-elasticache-controller.v0.0.28
        version: 0.0.28
      - name: ack-elasticache-controller.v0.0.27
        version: 0.0.27
      - name: ack-elasticache-controller.v0.0.26
        version: 0.0.26
      - name: ack-elasticache-controller.v0.0.25
        version: 0.0.25
      - name: ack-elasticache-controller.v0.0.24
        version: 0.0.24
      - name: ack-elasticache-controller.v0.0.23
        version: 0.0.23
      - name: ack-elasticache-controller.v0.0.22
        version: 0.0.22
      - name: ack-elasticache-controller.v0.0.21
        version: 0.0.21
      - name: ack-elasticache-controller.v0.0.20
        version: 0.0.20
      - name: ack-elasticache-controller.v0.0.19
        version: 0.0.19
      - name: ack-elasticache-controller.v0.0.18
        version: 0.0.18
      - name: ack-elasticache-controller.v0.0.17
        version: 0.0.17
      - name: ack-elasticache-controller.v0.0.16
        version: 0.0.16
      - name: ack-elasticache-controller.v0.0.15
        version: 0.0.15
      - name: ack-elasticache-controller.v0.0.14
        version: 0.0.14
      - name: ack-elasticache-controller.v0.0.13
        version: 0.0.13
      - name: ack-elasticache-controller.v0.0.12
        version: 0.0.12
      - name: ack-elasticache-controller.v0.0.11
        version: 0.0.11
      - name: ack-elasticache-controller.v0.0.10
        version: 0.0.10
      - name: ack-elasticache-controller.v0.0.9
        version: 0.0.9
      - name: ack-elasticache-controller.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: ack-elasticache-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ember-csi-community-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ember-csi-community-operator.v0.9.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ember-csi.io/v1alpha1",
                "kind": "EmberStorageBackend",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "config": {
                    "envVars": {
                      "X_CSI_PERSISTENCE_CONFIG": {
                        "storage": "crd"
                      },
                      "X_CSI_DEBUG_MODE": "",
                      "X_CSI_DEFAULT_MOUNT_FS": "ext4",
                      "X_CSI_EMBER_CONFIG": {
                        "grpc_workers": 30,
                        "disable_logs": false,
                        "plugin_name": "",
                        "slow_operations": true,
                        "disabled": [],
                        "enable_probe": false,
                        "debug": false,
                        "user_id": "ember_csi.io",
                        "project_id": "ember_csi.io"
                      },
                      "X_CSI_BACKEND_CONFIG": {
                        "enable_unsupported_driver": false,
                        "driver": "ACCESSIscsi",
                        "name": "",
                        "driver__ACCESSIscsi__vrts_lun_sparse": true,
                        "driver__ACCESSIscsi__vrts_target_config": "/etc/cinder/vrts_target.xml",
                        "driver__AS13000__as13000_ipsan_pools__transform_csv": "Pool0",
                        "driver__AS13000__as13000_meta_pool": "",
                        "driver__AS13000__as13000_token_available_time": 3300,
                        "driver__Acs5000ISCSI__acs5000_copy_interval": 5,
                        "driver__Acs5000ISCSI__acs5000_target": 0,
                        "driver__Acs5000ISCSI__acs5000_volpool_name__transform_csv": "pool01",
                        "driver__DPLFC__dpl_pool": "",
                        "driver__DPLFC__dpl_port": 8357,
                        "driver__DPLISCSI__dpl_pool": "",
                        "driver__DPLISCSI__dpl_port": 8357,
                        "driver__DSWARE__dsware_rest_url": "",
                        "driver__DSWARE__dsware_storage_pools": "",
                        "driver__DSWARE__manager_ips__transform_csv_kvs": "",
                        "driver__Datera__backend_availability_zone": "",
                        "driver__Datera__chap_password": "",
                        "driver__Datera__chap_username": "",
                        "driver__Datera__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__Datera__driver_client_cert": "",
                        "driver__Datera__driver_client_cert_key": "",
                        "driver__Datera__driver_data_namespace": "",
                        "driver__Datera__driver_ssl_cert_path": "",
                        "driver__Datera__driver_ssl_cert_verify": false,
                        "driver__Datera__driver_use_ssl": false,
                        "driver__Datera__filter_function": "",
                        "driver__Datera__goodness_function": "",
                        "driver__Datera__iscsi_iotype": "fileio",
                        "driver__Datera__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__Datera__iscsi_target_flags": "",
                        "driver__Datera__iscsi_write_cache": "on",
                        "driver__Datera__num_shell_tries": 3,
                        "driver__Datera__num_volume_device_scan_tries": 3,
                        "driver__Datera__report_discard_supported": false,
                        "driver__Datera__storage_protocol": "iscsi",
                        "driver__Datera__target_helper": "tgtadm",
                        "driver__Datera__target_ip_address": "$my_ip",
                        "driver__Datera__target_port": 3260,
                        "driver__Datera__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__Datera__target_protocol": "iscsi",
                        "driver__Datera__trace_flags__transform_csv": "",
                        "driver__Datera__use_chap_auth": false,
                        "driver__Datera__volume_backend_name": "",
                        "driver__Datera__volume_clear": "zero",
                        "driver__Datera__volume_clear_ionice": "",
                        "driver__Datera__volume_clear_size": 0,
                        "driver__Datera__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__Datera__volume_copy_bps_limit": 0,
                        "driver__Datera__volume_dd_blocksize": "1M",
                        "driver__Datera__volumes_dir": "$state_path/volumes",
                        "driver__FJDXFC__cinder_eternus_config_file": "/etc/cinder/cinder_fujitsu_eternus_dx.xml",
                        "driver__FJDXISCSI__cinder_eternus_config_file": "/etc/cinder/cinder_fujitsu_eternus_dx.xml",
                        "driver__FlashSystemFC__flashsystem_connection_protocol": "FC",
                        "driver__FlashSystemFC__flashsystem_multihostmap_enabled": true,
                        "driver__FlashSystemISCSI__flashsystem_connection_protocol": "FC",
                        "driver__FlashSystemISCSI__flashsystem_multihostmap_enabled": true,
                        "driver__GPFS__gpfs_images_dir": "",
                        "driver__GPFS__gpfs_images_share_mode__transform_empty_none": "",
                        "driver__GPFS__gpfs_max_clone_depth": 0,
                        "driver__GPFS__gpfs_mount_point_base": "",
                        "driver__GPFS__gpfs_sparse_volumes": true,
                        "driver__GPFS__gpfs_storage_pool": "system",
                        "driver__GPFSRemote__gpfs_hosts__transform_csv": "",
                        "driver__GPFSRemote__gpfs_hosts_key_file": "$state_path/ssh_known_hosts",
                        "driver__GPFSRemote__gpfs_images_dir": "",
                        "driver__GPFSRemote__gpfs_images_share_mode__transform_empty_none": "",
                        "driver__GPFSRemote__gpfs_max_clone_depth": 0,
                        "driver__GPFSRemote__gpfs_mount_point_base": "",
                        "driver__GPFSRemote__gpfs_private_key": "",
                        "driver__GPFSRemote__gpfs_sparse_volumes": true,
                        "driver__GPFSRemote__gpfs_ssh_port": 22,
                        "driver__GPFSRemote__gpfs_storage_pool": "system",
                        "driver__GPFSRemote__gpfs_strict_host_key_policy": false,
                        "driver__GPFSRemote__gpfs_user_login": "root",
                        "driver__GPFSRemote__gpfs_user_password": "",
                        "driver__HBSDFC__backend_availability_zone": "",
                        "driver__HBSDFC__chap_password": "",
                        "driver__HBSDFC__chap_username": "",
                        "driver__HBSDFC__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__HBSDFC__driver_client_cert": "",
                        "driver__HBSDFC__driver_client_cert_key": "",
                        "driver__HBSDFC__driver_data_namespace": "",
                        "driver__HBSDFC__driver_ssl_cert_path": "",
                        "driver__HBSDFC__driver_ssl_cert_verify": false,
                        "driver__HBSDFC__driver_use_ssl": false,
                        "driver__HBSDFC__filter_function": "",
                        "driver__HBSDFC__goodness_function": "",
                        "driver__HBSDFC__iscsi_iotype": "fileio",
                        "driver__HBSDFC__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__HBSDFC__iscsi_target_flags": "",
                        "driver__HBSDFC__iscsi_write_cache": "on",
                        "driver__HBSDFC__num_shell_tries": 3,
                        "driver__HBSDFC__num_volume_device_scan_tries": 3,
                        "driver__HBSDFC__report_discard_supported": false,
                        "driver__HBSDFC__storage_protocol": "iscsi",
                        "driver__HBSDFC__target_helper": "tgtadm",
                        "driver__HBSDFC__target_ip_address": "$my_ip",
                        "driver__HBSDFC__target_port": 3260,
                        "driver__HBSDFC__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__HBSDFC__target_protocol": "iscsi",
                        "driver__HBSDFC__trace_flags__transform_csv": "",
                        "driver__HBSDFC__use_chap_auth": false,
                        "driver__HBSDFC__volume_backend_name": "",
                        "driver__HBSDFC__volume_clear": "zero",
                        "driver__HBSDFC__volume_clear_ionice": "",
                        "driver__HBSDFC__volume_clear_size": 0,
                        "driver__HBSDFC__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__HBSDFC__volume_copy_bps_limit": 0,
                        "driver__HBSDFC__volume_dd_blocksize": "1M",
                        "driver__HBSDFC__volumes_dir": "$state_path/volumes",
                        "driver__HBSDISCSI__backend_availability_zone": "",
                        "driver__HBSDISCSI__chap_password": "",
                        "driver__HBSDISCSI__chap_username": "",
                        "driver__HBSDISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__HBSDISCSI__driver_client_cert": "",
                        "driver__HBSDISCSI__driver_client_cert_key": "",
                        "driver__HBSDISCSI__driver_data_namespace": "",
                        "driver__HBSDISCSI__driver_ssl_cert_path": "",
                        "driver__HBSDISCSI__driver_ssl_cert_verify": false,
                        "driver__HBSDISCSI__driver_use_ssl": false,
                        "driver__HBSDISCSI__filter_function": "",
                        "driver__HBSDISCSI__goodness_function": "",
                        "driver__HBSDISCSI__iscsi_iotype": "fileio",
                        "driver__HBSDISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__HBSDISCSI__iscsi_target_flags": "",
                        "driver__HBSDISCSI__iscsi_write_cache": "on",
                        "driver__HBSDISCSI__num_shell_tries": 3,
                        "driver__HBSDISCSI__num_volume_device_scan_tries": 3,
                        "driver__HBSDISCSI__report_discard_supported": false,
                        "driver__HBSDISCSI__storage_protocol": "iscsi",
                        "driver__HBSDISCSI__target_helper": "tgtadm",
                        "driver__HBSDISCSI__target_ip_address": "$my_ip",
                        "driver__HBSDISCSI__target_port": 3260,
                        "driver__HBSDISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__HBSDISCSI__target_protocol": "iscsi",
                        "driver__HBSDISCSI__trace_flags__transform_csv": "",
                        "driver__HBSDISCSI__use_chap_auth": false,
                        "driver__HBSDISCSI__volume_backend_name": "",
                        "driver__HBSDISCSI__volume_clear": "zero",
                        "driver__HBSDISCSI__volume_clear_ionice": "",
                        "driver__HBSDISCSI__volume_clear_size": 0,
                        "driver__HBSDISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__HBSDISCSI__volume_copy_bps_limit": 0,
                        "driver__HBSDISCSI__volume_dd_blocksize": "1M",
                        "driver__HBSDISCSI__volumes_dir": "$state_path/volumes",
                        "driver__HPE3PARFC__hpe3par_api_url": "",
                        "driver__HPE3PARFC__hpe3par_cpg__transform_csv": "OpenStack",
                        "driver__HPE3PARFC__hpe3par_cpg_snap": "",
                        "driver__HPE3PARFC__hpe3par_debug": false,
                        "driver__HPE3PARFC__hpe3par_iscsi_chap_enabled": false,
                        "driver__HPE3PARFC__hpe3par_iscsi_ips__transform_csv": "",
                        "driver__HPE3PARFC__hpe3par_password": "",
                        "driver__HPE3PARFC__hpe3par_snapshot_expiration": "",
                        "driver__HPE3PARFC__hpe3par_snapshot_retention": "",
                        "driver__HPE3PARFC__hpe3par_target_nsp": "",
                        "driver__HPE3PARFC__hpe3par_username": "",
                        "driver__HPE3PARFC__san_ip": "",
                        "driver__HPE3PARFC__san_login": "admin",
                        "driver__HPE3PARFC__san_password": "",
                        "driver__HPE3PARFC__san_private_key": "",
                        "driver__HPE3PARFC__san_ssh_port": 22,
                        "driver__HPE3PARFC__ssh_conn_timeout": 30,
                        "driver__HPE3PARFC__target_ip_address": "$my_ip",
                        "driver__HPE3PARFC__target_port": 3260,
                        "driver__HPE3PARFC__unique_fqdn_network": true,
                        "driver__HPE3PARISCSI__hpe3par_api_url": "",
                        "driver__HPE3PARISCSI__hpe3par_cpg__transform_csv": "OpenStack",
                        "driver__HPE3PARISCSI__hpe3par_cpg_snap": "",
                        "driver__HPE3PARISCSI__hpe3par_debug": false,
                        "driver__HPE3PARISCSI__hpe3par_iscsi_chap_enabled": false,
                        "driver__HPE3PARISCSI__hpe3par_iscsi_ips__transform_csv": "",
                        "driver__HPE3PARISCSI__hpe3par_password": "",
                        "driver__HPE3PARISCSI__hpe3par_snapshot_expiration": "",
                        "driver__HPE3PARISCSI__hpe3par_snapshot_retention": "",
                        "driver__HPE3PARISCSI__hpe3par_target_nsp": "",
                        "driver__HPE3PARISCSI__hpe3par_username": "",
                        "driver__HPE3PARISCSI__san_ip": "",
                        "driver__HPE3PARISCSI__san_login": "admin",
                        "driver__HPE3PARISCSI__san_password": "",
                        "driver__HPE3PARISCSI__san_private_key": "",
                        "driver__HPE3PARISCSI__san_ssh_port": 22,
                        "driver__HPE3PARISCSI__ssh_conn_timeout": 30,
                        "driver__HPE3PARISCSI__target_ip_address": "$my_ip",
                        "driver__HPE3PARISCSI__target_port": 3260,
                        "driver__HPE3PARISCSI__unique_fqdn_network": true,
                        "driver__HPMSAFC__driver_ssl_cert_path": "",
                        "driver__HPMSAFC__driver_ssl_cert_verify": false,
                        "driver__HPMSAFC__driver_use_ssl": false,
                        "driver__HPMSAFC__hpmsa_pool_name": "A",
                        "driver__HPMSAFC__hpmsa_pool_type": "virtual",
                        "driver__HPMSAFC__san_ip": "",
                        "driver__HPMSAFC__san_login": "admin",
                        "driver__HPMSAFC__san_password": "",
                        "driver__HPMSAISCSI__driver_ssl_cert_path": "",
                        "driver__HPMSAISCSI__driver_ssl_cert_verify": false,
                        "driver__HPMSAISCSI__driver_use_ssl": false,
                        "driver__HPMSAISCSI__hpmsa_iscsi_ips__transform_csv": "",
                        "driver__HPMSAISCSI__hpmsa_pool_name": "A",
                        "driver__HPMSAISCSI__hpmsa_pool_type": "virtual",
                        "driver__HPMSAISCSI__san_ip": "",
                        "driver__HPMSAISCSI__san_login": "admin",
                        "driver__HPMSAISCSI__san_password": "",
                        "driver__HuaweiFC__cinder_huawei_conf_file": "/etc/cinder/cinder_huawei_conf.xml",
                        "driver__HuaweiFC__hypermetro_devices": "",
                        "driver__HuaweiFC__metro_domain_name": "",
                        "driver__HuaweiFC__metro_san_address": "",
                        "driver__HuaweiFC__metro_san_password": "",
                        "driver__HuaweiFC__metro_san_user": "",
                        "driver__HuaweiFC__metro_storage_pools": "",
                        "driver__HuaweiISCSI__cinder_huawei_conf_file": "/etc/cinder/cinder_huawei_conf.xml",
                        "driver__HuaweiISCSI__hypermetro_devices": "",
                        "driver__HuaweiISCSI__metro_domain_name": "",
                        "driver__HuaweiISCSI__metro_san_address": "",
                        "driver__HuaweiISCSI__metro_san_password": "",
                        "driver__HuaweiISCSI__metro_san_user": "",
                        "driver__HuaweiISCSI__metro_storage_pools": "",
                        "driver__IBMStorage__chap": "disabled",
                        "driver__IBMStorage__connection_type": "iscsi",
                        "driver__IBMStorage__management_ips": "",
                        "driver__IBMStorage__proxy": "cinder.volume.drivers.ibm.ibm_storage.proxy.IBMStorageProxy",
                        "driver__InStorageMCSFC__instorage_mcs_allow_tenant_qos": false,
                        "driver__InStorageMCSFC__instorage_mcs_localcopy_rate": 50,
                        "driver__InStorageMCSFC__instorage_mcs_localcopy_timeout": 120,
                        "driver__InStorageMCSFC__instorage_mcs_vol_autoexpand": true,
                        "driver__InStorageMCSFC__instorage_mcs_vol_compression": false,
                        "driver__InStorageMCSFC__instorage_mcs_vol_grainsize": 256,
                        "driver__InStorageMCSFC__instorage_mcs_vol_intier": true,
                        "driver__InStorageMCSFC__instorage_mcs_vol_iogrp": "0",
                        "driver__InStorageMCSFC__instorage_mcs_vol_rsize": 2,
                        "driver__InStorageMCSFC__instorage_mcs_vol_warning": 0,
                        "driver__InStorageMCSFC__instorage_mcs_volpool_name__transform_csv": "volpool",
                        "driver__InStorageMCSFC__instorage_san_secondary_ip": "",
                        "driver__InStorageMCSISCSI__instorage_mcs_allow_tenant_qos": false,
                        "driver__InStorageMCSISCSI__instorage_mcs_localcopy_rate": 50,
                        "driver__InStorageMCSISCSI__instorage_mcs_localcopy_timeout": 120,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_autoexpand": true,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_compression": false,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_grainsize": 256,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_intier": true,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_iogrp": "0",
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_rsize": 2,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_warning": 0,
                        "driver__InStorageMCSISCSI__instorage_mcs_volpool_name__transform_csv": "volpool",
                        "driver__InStorageMCSISCSI__instorage_san_secondary_ip": "",
                        "driver__InfiniboxVolume__chap_password": "",
                        "driver__InfiniboxVolume__chap_username": "",
                        "driver__InfiniboxVolume__infinidat_iscsi_netspaces__transform_csv": "",
                        "driver__InfiniboxVolume__infinidat_pool_name": "",
                        "driver__InfiniboxVolume__infinidat_storage_protocol": "fc",
                        "driver__InfiniboxVolume__infinidat_use_compression": false,
                        "driver__InfiniboxVolume__num_volume_device_scan_tries": 3,
                        "driver__InfiniboxVolume__san_ip": "",
                        "driver__InfiniboxVolume__san_login": "admin",
                        "driver__InfiniboxVolume__san_password": "",
                        "driver__InfiniboxVolume__san_thin_provision": true,
                        "driver__InfiniboxVolume__use_chap_auth": false,
                        "driver__InfiniboxVolume__volume_dd_blocksize": "1M",
                        "driver__InfortrendCLIFC__infortrend_cli_cache": false,
                        "driver__InfortrendCLIFC__infortrend_cli_max_retries": 5,
                        "driver__InfortrendCLIFC__infortrend_cli_path": "/opt/bin/Infortrend/raidcmd_ESDS10.jar",
                        "driver__InfortrendCLIFC__infortrend_cli_timeout": 60,
                        "driver__InfortrendCLIFC__infortrend_iqn_prefix": "iqn.2002-10.com.infortrend",
                        "driver__InfortrendCLIFC__infortrend_pools_name__transform_csv": "",
                        "driver__InfortrendCLIFC__infortrend_slots_a_channels_id__transform_csv": "",
                        "driver__InfortrendCLIFC__infortrend_slots_b_channels_id__transform_csv": "",
                        "driver__InfortrendCLIFC__java_path": "/usr/bin/java",
                        "driver__InfortrendCLIISCSI__infortrend_cli_cache": false,
                        "driver__InfortrendCLIISCSI__infortrend_cli_max_retries": 5,
                        "driver__InfortrendCLIISCSI__infortrend_cli_path": "/opt/bin/Infortrend/raidcmd_ESDS10.jar",
                        "driver__InfortrendCLIISCSI__infortrend_cli_timeout": 60,
                        "driver__InfortrendCLIISCSI__infortrend_iqn_prefix": "iqn.2002-10.com.infortrend",
                        "driver__InfortrendCLIISCSI__infortrend_pools_name__transform_csv": "",
                        "driver__InfortrendCLIISCSI__infortrend_slots_a_channels_id__transform_csv": "",
                        "driver__InfortrendCLIISCSI__infortrend_slots_b_channels_id__transform_csv": "",
                        "driver__InfortrendCLIISCSI__java_path": "/usr/bin/java",
                        "driver__JovianISCSI__backend_availability_zone": "",
                        "driver__JovianISCSI__chap_password": "",
                        "driver__JovianISCSI__chap_username": "",
                        "driver__JovianISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__JovianISCSI__driver_client_cert": "",
                        "driver__JovianISCSI__driver_client_cert_key": "",
                        "driver__JovianISCSI__driver_data_namespace": "",
                        "driver__JovianISCSI__driver_ssl_cert_path": "",
                        "driver__JovianISCSI__driver_ssl_cert_verify": false,
                        "driver__JovianISCSI__driver_use_ssl": false,
                        "driver__JovianISCSI__filter_function": "",
                        "driver__JovianISCSI__goodness_function": "",
                        "driver__JovianISCSI__iscsi_iotype": "fileio",
                        "driver__JovianISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__JovianISCSI__iscsi_target_flags": "",
                        "driver__JovianISCSI__iscsi_write_cache": "on",
                        "driver__JovianISCSI__num_shell_tries": 3,
                        "driver__JovianISCSI__num_volume_device_scan_tries": 3,
                        "driver__JovianISCSI__report_discard_supported": false,
                        "driver__JovianISCSI__storage_protocol": "iscsi",
                        "driver__JovianISCSI__target_helper": "tgtadm",
                        "driver__JovianISCSI__target_ip_address": "$my_ip",
                        "driver__JovianISCSI__target_port": 3260,
                        "driver__JovianISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__JovianISCSI__target_protocol": "iscsi",
                        "driver__JovianISCSI__trace_flags__transform_csv": "",
                        "driver__JovianISCSI__use_chap_auth": false,
                        "driver__JovianISCSI__volume_backend_name": "",
                        "driver__JovianISCSI__volume_clear": "zero",
                        "driver__JovianISCSI__volume_clear_ionice": "",
                        "driver__JovianISCSI__volume_clear_size": 0,
                        "driver__JovianISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__JovianISCSI__volume_copy_bps_limit": 0,
                        "driver__JovianISCSI__volume_dd_blocksize": "1M",
                        "driver__JovianISCSI__volumes_dir": "$state_path/volumes",
                        "driver__KaminarioISCSI__disable_discovery": false,
                        "driver__KaminarioISCSI__san_ip": "",
                        "driver__KaminarioISCSI__san_login": "admin",
                        "driver__KaminarioISCSI__san_password": "",
                        "driver__KaminarioISCSI__unique_fqdn_network": true,
                        "driver__KaminarioISCSI__volume_dd_blocksize": "1M",
                        "driver__LVMVolume__iscsi_iotype": "fileio",
                        "driver__LVMVolume__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__LVMVolume__iscsi_target_flags": "",
                        "driver__LVMVolume__iscsi_write_cache": "on",
                        "driver__LVMVolume__lvm_conf_file": "/etc/cinder/lvm.conf",
                        "driver__LVMVolume__lvm_mirrors": 0,
                        "driver__LVMVolume__lvm_suppress_fd_warnings": false,
                        "driver__LVMVolume__lvm_type": "auto",
                        "driver__LVMVolume__nvmet_port_id": 1,
                        "driver__LVMVolume__scst_target_driver": "iscsi",
                        "driver__LVMVolume__scst_target_iqn_name": "",
                        "driver__LVMVolume__spdk_max_queue_depth": 64,
                        "driver__LVMVolume__spdk_rpc_ip": "",
                        "driver__LVMVolume__spdk_rpc_password": "",
                        "driver__LVMVolume__spdk_rpc_port": 8000,
                        "driver__LVMVolume__spdk_rpc_username": "",
                        "driver__LVMVolume__target_helper": "tgtadm",
                        "driver__LVMVolume__target_ip_address": "$my_ip",
                        "driver__LVMVolume__target_port": 3260,
                        "driver__LVMVolume__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__LVMVolume__target_protocol": "iscsi",
                        "driver__LVMVolume__volume_clear": "zero",
                        "driver__LVMVolume__volume_clear_size": 0,
                        "driver__LVMVolume__volume_dd_blocksize": "1M",
                        "driver__LVMVolume__volume_group": "cinder-volumes",
                        "driver__LVMVolume__volumes_dir": "$state_path/volumes",
                        "driver__LenovoFC__lenovo_pool_name": "A",
                        "driver__LenovoFC__lenovo_pool_type": "virtual",
                        "driver__LenovoISCSI__lenovo_iscsi_ips__transform_csv": "",
                        "driver__LenovoISCSI__lenovo_pool_name": "A",
                        "driver__LenovoISCSI__lenovo_pool_type": "virtual",
                        "driver__LinstorDrbd__linstor_autoplace_count": 0,
                        "driver__LinstorDrbd__linstor_controller_diskless": true,
                        "driver__LinstorDrbd__linstor_default_blocksize": 4096,
                        "driver__LinstorDrbd__linstor_default_storage_pool_name": "DfltStorPool",
                        "driver__LinstorDrbd__linstor_default_uri": "linstor://localhost",
                        "driver__LinstorDrbd__linstor_default_volume_group_name": "drbd-vg",
                        "driver__LinstorDrbd__linstor_volume_downsize_factor__transform_string_float": "4096",
                        "driver__LinstorIscsi__linstor_autoplace_count": 0,
                        "driver__LinstorIscsi__linstor_controller_diskless": true,
                        "driver__LinstorIscsi__linstor_default_blocksize": 4096,
                        "driver__LinstorIscsi__linstor_default_storage_pool_name": "DfltStorPool",
                        "driver__LinstorIscsi__linstor_default_uri": "linstor://localhost",
                        "driver__LinstorIscsi__linstor_default_volume_group_name": "drbd-vg",
                        "driver__LinstorIscsi__linstor_volume_downsize_factor__transform_string_float": "4096",
                        "driver__MStorageFC__nec_actual_free_capacity": false,
                        "driver__MStorageFC__nec_auto_accesscontrol": true,
                        "driver__MStorageFC__nec_backend_max_ld_count": 1024,
                        "driver__MStorageFC__nec_backup_ldname_format": "LX:%s",
                        "driver__MStorageFC__nec_backup_pools__transform_csv": "",
                        "driver__MStorageFC__nec_cv_ldname_format": "LX:__ControlVolume_%xh",
                        "driver__MStorageFC__nec_diskarray_name": "",
                        "driver__MStorageFC__nec_ismcli_fip": "",
                        "driver__MStorageFC__nec_ismcli_password": "",
                        "driver__MStorageFC__nec_ismcli_privkey": "",
                        "driver__MStorageFC__nec_ismcli_user": "",
                        "driver__MStorageFC__nec_ismview_alloptimize": false,
                        "driver__MStorageFC__nec_ismview_dir": "/tmp/nec/cinder",
                        "driver__MStorageFC__nec_ldname_format": "LX:%s",
                        "driver__MStorageFC__nec_ldset": "",
                        "driver__MStorageFC__nec_pools__transform_csv": "",
                        "driver__MStorageFC__nec_queryconfig_view": false,
                        "driver__MStorageFC__nec_ssh_pool_port_number": 22,
                        "driver__MStorageFC__nec_unpairthread_timeout": 3600,
                        "driver__MStorageISCSI__nec_actual_free_capacity": false,
                        "driver__MStorageISCSI__nec_auto_accesscontrol": true,
                        "driver__MStorageISCSI__nec_backend_max_ld_count": 1024,
                        "driver__MStorageISCSI__nec_backup_ldname_format": "LX:%s",
                        "driver__MStorageISCSI__nec_backup_pools__transform_csv": "",
                        "driver__MStorageISCSI__nec_cv_ldname_format": "LX:__ControlVolume_%xh",
                        "driver__MStorageISCSI__nec_diskarray_name": "",
                        "driver__MStorageISCSI__nec_ismcli_fip": "",
                        "driver__MStorageISCSI__nec_ismcli_password": "",
                        "driver__MStorageISCSI__nec_ismcli_privkey": "",
                        "driver__MStorageISCSI__nec_ismcli_user": "",
                        "driver__MStorageISCSI__nec_ismview_alloptimize": false,
                        "driver__MStorageISCSI__nec_ismview_dir": "/tmp/nec/cinder",
                        "driver__MStorageISCSI__nec_ldname_format": "LX:%s",
                        "driver__MStorageISCSI__nec_ldset": "",
                        "driver__MStorageISCSI__nec_pools__transform_csv": "",
                        "driver__MStorageISCSI__nec_queryconfig_view": false,
                        "driver__MStorageISCSI__nec_ssh_pool_port_number": 22,
                        "driver__MStorageISCSI__nec_unpairthread_timeout": 3600,
                        "driver__MacroSANFC__backend_availability_zone": "",
                        "driver__MacroSANFC__chap_password": "",
                        "driver__MacroSANFC__chap_username": "",
                        "driver__MacroSANFC__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__MacroSANFC__driver_client_cert": "",
                        "driver__MacroSANFC__driver_client_cert_key": "",
                        "driver__MacroSANFC__driver_data_namespace": "",
                        "driver__MacroSANFC__driver_ssl_cert_path": "",
                        "driver__MacroSANFC__driver_ssl_cert_verify": false,
                        "driver__MacroSANFC__driver_use_ssl": false,
                        "driver__MacroSANFC__filter_function": "",
                        "driver__MacroSANFC__goodness_function": "",
                        "driver__MacroSANFC__iscsi_iotype": "fileio",
                        "driver__MacroSANFC__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__MacroSANFC__iscsi_target_flags": "",
                        "driver__MacroSANFC__iscsi_write_cache": "on",
                        "driver__MacroSANFC__num_shell_tries": 3,
                        "driver__MacroSANFC__num_volume_device_scan_tries": 3,
                        "driver__MacroSANFC__report_discard_supported": false,
                        "driver__MacroSANFC__storage_protocol": "iscsi",
                        "driver__MacroSANFC__target_helper": "tgtadm",
                        "driver__MacroSANFC__target_ip_address": "$my_ip",
                        "driver__MacroSANFC__target_port": 3260,
                        "driver__MacroSANFC__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__MacroSANFC__target_protocol": "iscsi",
                        "driver__MacroSANFC__trace_flags__transform_csv": "",
                        "driver__MacroSANFC__use_chap_auth": false,
                        "driver__MacroSANFC__volume_backend_name": "",
                        "driver__MacroSANFC__volume_clear": "zero",
                        "driver__MacroSANFC__volume_clear_ionice": "",
                        "driver__MacroSANFC__volume_clear_size": 0,
                        "driver__MacroSANFC__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__MacroSANFC__volume_copy_bps_limit": 0,
                        "driver__MacroSANFC__volume_dd_blocksize": "1M",
                        "driver__MacroSANFC__volumes_dir": "$state_path/volumes",
                        "driver__MacroSANISCSI__backend_availability_zone": "",
                        "driver__MacroSANISCSI__chap_password": "",
                        "driver__MacroSANISCSI__chap_username": "",
                        "driver__MacroSANISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__MacroSANISCSI__driver_client_cert": "",
                        "driver__MacroSANISCSI__driver_client_cert_key": "",
                        "driver__MacroSANISCSI__driver_data_namespace": "",
                        "driver__MacroSANISCSI__driver_ssl_cert_path": "",
                        "driver__MacroSANISCSI__driver_ssl_cert_verify": false,
                        "driver__MacroSANISCSI__driver_use_ssl": false,
                        "driver__MacroSANISCSI__filter_function": "",
                        "driver__MacroSANISCSI__goodness_function": "",
                        "driver__MacroSANISCSI__iscsi_iotype": "fileio",
                        "driver__MacroSANISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__MacroSANISCSI__iscsi_target_flags": "",
                        "driver__MacroSANISCSI__iscsi_write_cache": "on",
                        "driver__MacroSANISCSI__num_shell_tries": 3,
                        "driver__MacroSANISCSI__num_volume_device_scan_tries": 3,
                        "driver__MacroSANISCSI__report_discard_supported": false,
                        "driver__MacroSANISCSI__storage_protocol": "iscsi",
                        "driver__MacroSANISCSI__target_helper": "tgtadm",
                        "driver__MacroSANISCSI__target_ip_address": "$my_ip",
                        "driver__MacroSANISCSI__target_port": 3260,
                        "driver__MacroSANISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__MacroSANISCSI__target_protocol": "iscsi",
                        "driver__MacroSANISCSI__trace_flags__transform_csv": "",
                        "driver__MacroSANISCSI__use_chap_auth": false,
                        "driver__MacroSANISCSI__volume_backend_name": "",
                        "driver__MacroSANISCSI__volume_clear": "zero",
                        "driver__MacroSANISCSI__volume_clear_ionice": "",
                        "driver__MacroSANISCSI__volume_clear_size": 0,
                        "driver__MacroSANISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__MacroSANISCSI__volume_copy_bps_limit": 0,
                        "driver__MacroSANISCSI__volume_dd_blocksize": "1M",
                        "driver__MacroSANISCSI__volumes_dir": "$state_path/volumes",
                        "driver__NetAppCmodeFibreChannel__netapp_vserver": "",
                        "driver__NetAppCmodeISCSI__netapp_vserver": "",
                        "driver__NexentaISCSI__nexenta_blocksize": 4096,
                        "driver__NexentaISCSI__nexenta_dataset_compression": "on",
                        "driver__NexentaISCSI__nexenta_dataset_dedup": "off",
                        "driver__NexentaISCSI__nexenta_dataset_description": "",
                        "driver__NexentaISCSI__nexenta_folder": "",
                        "driver__NexentaISCSI__nexenta_group_snapshot_template": "group-snapshot-%s",
                        "driver__NexentaISCSI__nexenta_host": "",
                        "driver__NexentaISCSI__nexenta_host_group_prefix": "cinder",
                        "driver__NexentaISCSI__nexenta_iscsi_target_host_group": "all",
                        "driver__NexentaISCSI__nexenta_iscsi_target_portal_groups": "",
                        "driver__NexentaISCSI__nexenta_iscsi_target_portal_port": 3260,
                        "driver__NexentaISCSI__nexenta_iscsi_target_portals": "",
                        "driver__NexentaISCSI__nexenta_lu_writebackcache_disabled": false,
                        "driver__NexentaISCSI__nexenta_luns_per_target": 100,
                        "driver__NexentaISCSI__nexenta_ns5_blocksize": 32,
                        "driver__NexentaISCSI__nexenta_origin_snapshot_template": "origin-snapshot-%s",
                        "driver__NexentaISCSI__nexenta_rest_backoff_factor__transform_string_float": "0.5",
                        "driver__NexentaISCSI__nexenta_rest_connect_timeout__transform_string_float": "30",
                        "driver__NexentaISCSI__nexenta_rest_protocol": "auto",
                        "driver__NexentaISCSI__nexenta_rest_read_timeout__transform_string_float": "300",
                        "driver__NexentaISCSI__nexenta_rest_retry_count": 3,
                        "driver__NexentaISCSI__nexenta_sparse": false,
                        "driver__NexentaISCSI__nexenta_target_group_prefix": "cinder",
                        "driver__NexentaISCSI__nexenta_target_prefix": "iqn.1986-03.com.sun:02:cinder",
                        "driver__NexentaISCSI__nexenta_use_https": true,
                        "driver__NexentaISCSI__nexenta_volume": "cinder",
                        "driver__NexentaISCSI__nexenta_volume_group": "iscsi",
                        "driver__NimbleFC__nimble_pool_name": "default",
                        "driver__NimbleFC__nimble_subnet_label": "*",
                        "driver__NimbleFC__nimble_verify_cert_path": "",
                        "driver__NimbleFC__nimble_verify_certificate": false,
                        "driver__NimbleISCSI__nimble_pool_name": "default",
                        "driver__NimbleISCSI__nimble_subnet_label": "*",
                        "driver__NimbleISCSI__nimble_verify_cert_path": "",
                        "driver__NimbleISCSI__nimble_verify_certificate": false,
                        "driver__PVMEFC__driver_ssl_cert_path": "",
                        "driver__PVMEFC__driver_ssl_cert_verify": false,
                        "driver__PVMEFC__driver_use_ssl": false,
                        "driver__PVMEFC__pvme_pool_name": "A",
                        "driver__PVMEFC__san_ip": "",
                        "driver__PVMEFC__san_login": "admin",
                        "driver__PVMEFC__san_password": "",
                        "driver__PVMEISCSI__driver_ssl_cert_path": "",
                        "driver__PVMEISCSI__driver_ssl_cert_verify": false,
                        "driver__PVMEISCSI__driver_use_ssl": false,
                        "driver__PVMEISCSI__pvme_iscsi_ips__transform_csv": "",
                        "driver__PVMEISCSI__pvme_pool_name": "A",
                        "driver__PVMEISCSI__san_ip": "",
                        "driver__PVMEISCSI__san_login": "admin",
                        "driver__PVMEISCSI__san_password": "",
                        "driver__PowerFlex__powerflex_allow_migration_during_rebuild": false,
                        "driver__PowerFlex__powerflex_allow_non_padded_volumes": false,
                        "driver__PowerFlex__powerflex_max_over_subscription_ratio__transform_string_float": "10.0",
                        "driver__PowerFlex__powerflex_rest_server_port": 443,
                        "driver__PowerFlex__powerflex_round_volume_capacity": true,
                        "driver__PowerFlex__powerflex_server_api_version": "",
                        "driver__PowerFlex__powerflex_storage_pools": "",
                        "driver__PowerFlex__powerflex_unmap_volume_before_deletion": false,
                        "driver__PowerMaxFC__driver_ssl_cert_verify": false,
                        "driver__PowerMaxFC__initiator_check": false,
                        "driver__PowerMaxFC__interval": 3,
                        "driver__PowerMaxFC__load_balance": false,
                        "driver__PowerMaxFC__load_balance_real_time": false,
                        "driver__PowerMaxFC__load_data_format": "Avg",
                        "driver__PowerMaxFC__load_look_back": 60,
                        "driver__PowerMaxFC__load_look_back_real_time": 1,
                        "driver__PowerMaxFC__port_group_load_metric": "PercentBusy",
                        "driver__PowerMaxFC__port_load_metric": "PercentBusy",
                        "driver__PowerMaxFC__powermax_array": "",
                        "driver__PowerMaxFC__powermax_array_tag_list__transform_csv": "",
                        "driver__PowerMaxFC__powermax_port_group_name_template": "portGroupName",
                        "driver__PowerMaxFC__powermax_port_groups__transform_csv": "",
                        "driver__PowerMaxFC__powermax_service_level": "",
                        "driver__PowerMaxFC__powermax_short_host_name_template": "shortHostName",
                        "driver__PowerMaxFC__powermax_snapvx_unlink_limit": 3,
                        "driver__PowerMaxFC__powermax_srp": "",
                        "driver__PowerMaxFC__retries": 200,
                        "driver__PowerMaxFC__san_ip": "",
                        "driver__PowerMaxFC__san_login": "admin",
                        "driver__PowerMaxFC__san_password": "",
                        "driver__PowerMaxFC__u4p_failover_autofailback": true,
                        "driver__PowerMaxFC__u4p_failover_backoff_factor": 1,
                        "driver__PowerMaxFC__u4p_failover_retries": 3,
                        "driver__PowerMaxFC__u4p_failover_target__transform_csv_kvs": "",
                        "driver__PowerMaxFC__u4p_failover_timeout": 20,
                        "driver__PowerMaxFC__vmax_workload": "",
                        "driver__PowerMaxISCSI__chap_password": "",
                        "driver__PowerMaxISCSI__chap_username": "",
                        "driver__PowerMaxISCSI__driver_ssl_cert_verify": false,
                        "driver__PowerMaxISCSI__initiator_check": false,
                        "driver__PowerMaxISCSI__interval": 3,
                        "driver__PowerMaxISCSI__load_balance": false,
                        "driver__PowerMaxISCSI__load_balance_real_time": false,
                        "driver__PowerMaxISCSI__load_data_format": "Avg",
                        "driver__PowerMaxISCSI__load_look_back": 60,
                        "driver__PowerMaxISCSI__load_look_back_real_time": 1,
                        "driver__PowerMaxISCSI__port_group_load_metric": "PercentBusy",
                        "driver__PowerMaxISCSI__port_load_metric": "PercentBusy",
                        "driver__PowerMaxISCSI__powermax_array": "",
                        "driver__PowerMaxISCSI__powermax_array_tag_list__transform_csv": "",
                        "driver__PowerMaxISCSI__powermax_port_group_name_template": "portGroupName",
                        "driver__PowerMaxISCSI__powermax_port_groups__transform_csv": "",
                        "driver__PowerMaxISCSI__powermax_service_level": "",
                        "driver__PowerMaxISCSI__powermax_short_host_name_template": "shortHostName",
                        "driver__PowerMaxISCSI__powermax_snapvx_unlink_limit": 3,
                        "driver__PowerMaxISCSI__powermax_srp": "",
                        "driver__PowerMaxISCSI__retries": 200,
                        "driver__PowerMaxISCSI__san_ip": "",
                        "driver__PowerMaxISCSI__san_login": "admin",
                        "driver__PowerMaxISCSI__san_password": "",
                        "driver__PowerMaxISCSI__u4p_failover_autofailback": true,
                        "driver__PowerMaxISCSI__u4p_failover_backoff_factor": 1,
                        "driver__PowerMaxISCSI__u4p_failover_retries": 3,
                        "driver__PowerMaxISCSI__u4p_failover_target__transform_csv_kvs": "",
                        "driver__PowerMaxISCSI__u4p_failover_timeout": 20,
                        "driver__PowerMaxISCSI__use_chap_auth": false,
                        "driver__PowerMaxISCSI__vmax_workload": "",
                        "driver__PowerStore__powerstore_appliances__transform_csv": "",
                        "driver__PowerStore__powerstore_ports__transform_csv": "",
                        "driver__PureFC__driver_ssl_cert_path": "",
                        "driver__PureFC__driver_ssl_cert_verify": false,
                        "driver__PureFC__pure_api_token": "",
                        "driver__PureFC__pure_automatic_max_oversubscription_ratio": true,
                        "driver__PureFC__pure_eradicate_on_delete": false,
                        "driver__PureFC__pure_host_personality__transform_empty_none": "",
                        "driver__PureFC__pure_iscsi_cidr": "0.0.0.0/0",
                        "driver__PureFC__pure_replica_interval_default": 3600,
                        "driver__PureFC__pure_replica_retention_long_term_default": 7,
                        "driver__PureFC__pure_replica_retention_long_term_per_day_default": 3,
                        "driver__PureFC__pure_replica_retention_short_term_default": 14400,
                        "driver__PureFC__pure_replication_pg_name": "cinder-group",
                        "driver__PureFC__pure_replication_pod_name": "cinder-pod",
                        "driver__PureFC__san_ip": "",
                        "driver__PureFC__use_chap_auth": false,
                        "driver__PureISCSI__driver_ssl_cert_path": "",
                        "driver__PureISCSI__driver_ssl_cert_verify": false,
                        "driver__PureISCSI__pure_api_token": "",
                        "driver__PureISCSI__pure_automatic_max_oversubscription_ratio": true,
                        "driver__PureISCSI__pure_eradicate_on_delete": false,
                        "driver__PureISCSI__pure_host_personality__transform_empty_none": "",
                        "driver__PureISCSI__pure_iscsi_cidr": "0.0.0.0/0",
                        "driver__PureISCSI__pure_replica_interval_default": 3600,
                        "driver__PureISCSI__pure_replica_retention_long_term_default": 7,
                        "driver__PureISCSI__pure_replica_retention_long_term_per_day_default": 3,
                        "driver__PureISCSI__pure_replica_retention_short_term_default": 14400,
                        "driver__PureISCSI__pure_replication_pg_name": "cinder-group",
                        "driver__PureISCSI__pure_replication_pod_name": "cinder-pod",
                        "driver__PureISCSI__san_ip": "",
                        "driver__PureISCSI__use_chap_auth": false,
                        "driver__QnapISCSI__chap_password": "",
                        "driver__QnapISCSI__chap_username": "",
                        "driver__QnapISCSI__driver_ssl_cert_verify": false,
                        "driver__QnapISCSI__qnap_management_url": "",
                        "driver__QnapISCSI__qnap_poolname": "",
                        "driver__QnapISCSI__qnap_storage_protocol": "iscsi",
                        "driver__QnapISCSI__san_login": "admin",
                        "driver__QnapISCSI__san_password": "",
                        "driver__QnapISCSI__target_ip_address": "$my_ip",
                        "driver__QnapISCSI__use_chap_auth": false,
                        "driver__Quobyte__quobyte_client_cfg": "",
                        "driver__Quobyte__quobyte_mount_point_base": "$state_path/mnt",
                        "driver__Quobyte__quobyte_overlay_volumes": false,
                        "driver__Quobyte__quobyte_qcow2_volumes": true,
                        "driver__Quobyte__quobyte_sparsed_volumes": true,
                        "driver__Quobyte__quobyte_volume_from_snapshot_cache": false,
                        "driver__Quobyte__quobyte_volume_url": "",
                        "driver__RBD__deferred_deletion_delay": 0,
                        "driver__RBD__deferred_deletion_purge_interval": 60,
                        "driver__RBD__enable_deferred_deletion": false,
                        "driver__RBD__rados_connect_timeout": -1,
                        "driver__RBD__rados_connection_interval": 5,
                        "driver__RBD__rados_connection_retries": 3,
                        "driver__RBD__rbd_ceph_conf": "",
                        "driver__RBD__rbd_cluster_name": "ceph",
                        "driver__RBD__rbd_exclusive_cinder_pool": false,
                        "driver__RBD__rbd_flatten_volume_from_snapshot": false,
                        "driver__RBD__rbd_keyring_conf": "",
                        "driver__RBD__rbd_max_clone_depth": 5,
                        "driver__RBD__rbd_pool": "rbd",
                        "driver__RBD__rbd_secret_uuid": "",
                        "driver__RBD__rbd_store_chunk_size": 4,
                        "driver__RBD__rbd_user": "",
                        "driver__RBD__report_dynamic_total_capacity": true,
                        "driver__RBD__volume_dd_blocksize": "1M",
                        "driver__RBDISCSI__deferred_deletion_delay": 0,
                        "driver__RBDISCSI__deferred_deletion_purge_interval": 60,
                        "driver__RBDISCSI__driver_ssl_cert_verify": false,
                        "driver__RBDISCSI__enable_deferred_deletion": false,
                        "driver__RBDISCSI__rados_connect_timeout": -1,
                        "driver__RBDISCSI__rados_connection_interval": 5,
                        "driver__RBDISCSI__rados_connection_retries": 3,
                        "driver__RBDISCSI__rbd_ceph_conf": "",
                        "driver__RBDISCSI__rbd_cluster_name": "ceph",
                        "driver__RBDISCSI__rbd_exclusive_cinder_pool": false,
                        "driver__RBDISCSI__rbd_flatten_volume_from_snapshot": false,
                        "driver__RBDISCSI__rbd_iscsi_api_debug": false,
                        "driver__RBDISCSI__rbd_iscsi_api_password": "",
                        "driver__RBDISCSI__rbd_iscsi_api_url": "",
                        "driver__RBDISCSI__rbd_iscsi_api_user": "",
                        "driver__RBDISCSI__rbd_iscsi_target_iqn": "",
                        "driver__RBDISCSI__rbd_max_clone_depth": 5,
                        "driver__RBDISCSI__rbd_pool": "rbd",
                        "driver__RBDISCSI__rbd_secret_uuid": "",
                        "driver__RBDISCSI__rbd_store_chunk_size": 4,
                        "driver__RBDISCSI__rbd_user": "",
                        "driver__RBDISCSI__report_dynamic_total_capacity": true,
                        "driver__RBDISCSI__suppress_requests_ssl_warnings": false,
                        "driver__RBDISCSI__volume_dd_blocksize": "1M",
                        "driver__RSD__podm_password": "",
                        "driver__RSD__podm_url": "",
                        "driver__RSD__podm_username": "",
                        "driver__SCFC__dell_api_async_rest_timeout": 15,
                        "driver__SCFC__dell_api_sync_rest_timeout": 30,
                        "driver__SCFC__dell_sc_api_port": 3033,
                        "driver__SCFC__dell_sc_server_folder": "openstack",
                        "driver__SCFC__dell_sc_ssn": 64702,
                        "driver__SCFC__dell_sc_verify_cert": false,
                        "driver__SCFC__dell_sc_volume_folder": "openstack",
                        "driver__SCFC__dell_server_os": "Red Hat Linux 6.x",
                        "driver__SCFC__excluded_domain_ips__transform_csv": "",
                        "driver__SCFC__included_domain_ips__transform_csv": "",
                        "driver__SCFC__secondary_san_ip": "",
                        "driver__SCFC__secondary_san_login": "Admin",
                        "driver__SCFC__secondary_san_password": "",
                        "driver__SCFC__secondary_sc_api_port": 3033,
                        "driver__SCISCSI__dell_api_async_rest_timeout": 15,
                        "driver__SCISCSI__dell_api_sync_rest_timeout": 30,
                        "driver__SCISCSI__dell_sc_api_port": 3033,
                        "driver__SCISCSI__dell_sc_server_folder": "openstack",
                        "driver__SCISCSI__dell_sc_ssn": 64702,
                        "driver__SCISCSI__dell_sc_verify_cert": false,
                        "driver__SCISCSI__dell_sc_volume_folder": "openstack",
                        "driver__SCISCSI__dell_server_os": "Red Hat Linux 6.x",
                        "driver__SCISCSI__excluded_domain_ips__transform_csv": "",
                        "driver__SCISCSI__included_domain_ips__transform_csv": "",
                        "driver__SCISCSI__secondary_san_ip": "",
                        "driver__SCISCSI__secondary_san_login": "Admin",
                        "driver__SCISCSI__secondary_san_password": "",
                        "driver__SCISCSI__secondary_sc_api_port": 3033,
                        "driver__SdsISCSI__backend_availability_zone": "",
                        "driver__SdsISCSI__chap_password": "",
                        "driver__SdsISCSI__chap_username": "",
                        "driver__SdsISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__SdsISCSI__driver_client_cert": "",
                        "driver__SdsISCSI__driver_client_cert_key": "",
                        "driver__SdsISCSI__driver_data_namespace": "",
                        "driver__SdsISCSI__driver_ssl_cert_path": "",
                        "driver__SdsISCSI__driver_ssl_cert_verify": false,
                        "driver__SdsISCSI__driver_use_ssl": false,
                        "driver__SdsISCSI__filter_function": "",
                        "driver__SdsISCSI__goodness_function": "",
                        "driver__SdsISCSI__iscsi_iotype": "fileio",
                        "driver__SdsISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__SdsISCSI__iscsi_target_flags": "",
                        "driver__SdsISCSI__iscsi_write_cache": "on",
                        "driver__SdsISCSI__num_shell_tries": 3,
                        "driver__SdsISCSI__num_volume_device_scan_tries": 3,
                        "driver__SdsISCSI__report_discard_supported": false,
                        "driver__SdsISCSI__storage_protocol": "iscsi",
                        "driver__SdsISCSI__target_helper": "tgtadm",
                        "driver__SdsISCSI__target_ip_address": "$my_ip",
                        "driver__SdsISCSI__target_port": 3260,
                        "driver__SdsISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__SdsISCSI__target_protocol": "iscsi",
                        "driver__SdsISCSI__trace_flags__transform_csv": "",
                        "driver__SdsISCSI__use_chap_auth": false,
                        "driver__SdsISCSI__volume_backend_name": "",
                        "driver__SdsISCSI__volume_clear": "zero",
                        "driver__SdsISCSI__volume_clear_ionice": "",
                        "driver__SdsISCSI__volume_clear_size": 0,
                        "driver__SdsISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__SdsISCSI__volume_copy_bps_limit": 0,
                        "driver__SdsISCSI__volume_dd_blocksize": "1M",
                        "driver__SdsISCSI__volumes_dir": "$state_path/volumes",
                        "driver__SolidFire__driver_ssl_cert_verify": false,
                        "driver__SolidFire__san_ip": "",
                        "driver__SolidFire__san_login": "admin",
                        "driver__SolidFire__san_password": "",
                        "driver__SolidFire__sf_account_prefix": "",
                        "driver__SolidFire__sf_allow_tenant_qos": false,
                        "driver__SolidFire__sf_api_port": 443,
                        "driver__SolidFire__sf_api_request_timeout": 30,
                        "driver__SolidFire__sf_cluster_pairing_timeout": 60,
                        "driver__SolidFire__sf_emulate_512": true,
                        "driver__SolidFire__sf_enable_vag": false,
                        "driver__SolidFire__sf_provisioning_calc": "maxProvisionedSpace",
                        "driver__SolidFire__sf_svip": "",
                        "driver__SolidFire__sf_volume_clone_timeout": 600,
                        "driver__SolidFire__sf_volume_create_timeout": 60,
                        "driver__SolidFire__sf_volume_pairing_timeout": 3600,
                        "driver__SolidFire__sf_volume_prefix": "UUID-",
                        "driver__StorPool__storpool_replication": 3,
                        "driver__StorPool__storpool_template": "",
                        "driver__StorwizeSVCFC__cycle_period_seconds": 300,
                        "driver__StorwizeSVCFC__storwize_peer_pool": "",
                        "driver__StorwizeSVCFC__storwize_preferred_host_site__transform_csv_kvs": "",
                        "driver__StorwizeSVCFC__storwize_san_secondary_ip": "",
                        "driver__StorwizeSVCFC__storwize_svc_allow_tenant_qos": false,
                        "driver__StorwizeSVCFC__storwize_svc_flashcopy_rate": 50,
                        "driver__StorwizeSVCFC__storwize_svc_flashcopy_timeout": 120,
                        "driver__StorwizeSVCFC__storwize_svc_mirror_pool": "",
                        "driver__StorwizeSVCFC__storwize_svc_multipath_enabled": false,
                        "driver__StorwizeSVCFC__storwize_svc_retain_aux_volume": false,
                        "driver__StorwizeSVCFC__storwize_svc_stretched_cluster_partner": "",
                        "driver__StorwizeSVCFC__storwize_svc_vol_autoexpand": true,
                        "driver__StorwizeSVCFC__storwize_svc_vol_compression": false,
                        "driver__StorwizeSVCFC__storwize_svc_vol_easytier": true,
                        "driver__StorwizeSVCFC__storwize_svc_vol_grainsize": 256,
                        "driver__StorwizeSVCFC__storwize_svc_vol_iogrp": "0",
                        "driver__StorwizeSVCFC__storwize_svc_vol_nofmtdisk": false,
                        "driver__StorwizeSVCFC__storwize_svc_vol_rsize": 2,
                        "driver__StorwizeSVCFC__storwize_svc_vol_warning": 0,
                        "driver__StorwizeSVCFC__storwize_svc_volpool_name__transform_csv": "volpool",
                        "driver__StorwizeSVCISCSI__cycle_period_seconds": 300,
                        "driver__StorwizeSVCISCSI__storwize_peer_pool": "",
                        "driver__StorwizeSVCISCSI__storwize_preferred_host_site__transform_csv_kvs": "",
                        "driver__StorwizeSVCISCSI__storwize_san_secondary_ip": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_allow_tenant_qos": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_flashcopy_rate": 50,
                        "driver__StorwizeSVCISCSI__storwize_svc_flashcopy_timeout": 120,
                        "driver__StorwizeSVCISCSI__storwize_svc_iscsi_chap_enabled": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_mirror_pool": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_retain_aux_volume": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_stretched_cluster_partner": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_autoexpand": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_compression": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_easytier": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_grainsize": 256,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_iogrp": "0",
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_nofmtdisk": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_rsize": 2,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_warning": 0,
                        "driver__StorwizeSVCISCSI__storwize_svc_volpool_name__transform_csv": "volpool",
                        "driver__SynoISCSI__chap_password": "",
                        "driver__SynoISCSI__chap_username": "",
                        "driver__SynoISCSI__driver_use_ssl": false,
                        "driver__SynoISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__SynoISCSI__synology_admin_port": 5000,
                        "driver__SynoISCSI__synology_device_id": "",
                        "driver__SynoISCSI__synology_one_time_pass": "",
                        "driver__SynoISCSI__synology_password": "",
                        "driver__SynoISCSI__synology_pool_name": "",
                        "driver__SynoISCSI__synology_ssl_verify": true,
                        "driver__SynoISCSI__synology_username": "admin",
                        "driver__SynoISCSI__target_ip_address": "$my_ip",
                        "driver__SynoISCSI__target_port": 3260,
                        "driver__SynoISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__SynoISCSI__target_protocol": "iscsi",
                        "driver__SynoISCSI__use_chap_auth": false,
                        "driver__Unity__remove_empty_host": false,
                        "driver__Unity__unity_io_ports__transform_csv": "",
                        "driver__Unity__unity_storage_pool_names__transform_csv": "",
                        "driver__VNX__default_timeout": 31536000,
                        "driver__VNX__destroy_empty_storage_group": false,
                        "driver__VNX__force_delete_lun_in_storagegroup": true,
                        "driver__VNX__ignore_pool_full_threshold": false,
                        "driver__VNX__initiator_auto_deregistration": false,
                        "driver__VNX__initiator_auto_registration": false,
                        "driver__VNX__io_port_list__transform_csv": "",
                        "driver__VNX__iscsi_initiators": "",
                        "driver__VNX__max_luns_per_storage_group": 255,
                        "driver__VNX__naviseccli_path": "",
                        "driver__VNX__storage_vnx_authentication_type": "global",
                        "driver__VNX__storage_vnx_pool_names__transform_csv": "",
                        "driver__VNX__storage_vnx_security_file_dir": "",
                        "driver__VNX__vnx_async_migrate": true,
                        "driver__VZStorage__vzstorage_default_volume_format": "raw",
                        "driver__VZStorage__vzstorage_mount_options__transform_csv": "",
                        "driver__VZStorage__vzstorage_mount_point_base": "$state_path/mnt",
                        "driver__VZStorage__vzstorage_shares_config": "/etc/cinder/vzstorage_shares",
                        "driver__VZStorage__vzstorage_sparsed_volumes": true,
                        "driver__VZStorage__vzstorage_used_ratio__transform_string_float": "0.95",
                        "driver__WindowsISCSI__windows_iscsi_lun_path": "C:\\iSCSIVirtualDisks",
                        "driver__WindowsSmbfs__smbfs_default_volume_format": "vhd",
                        "driver__WindowsSmbfs__smbfs_mount_point_base": "C:\\OpenStack\\_mnt",
                        "driver__WindowsSmbfs__smbfs_pool_mappings__transform_csv_kvs": "",
                        "driver__WindowsSmbfs__smbfs_shares_config": "C:\\OpenStack\\smbfs_shares.txt",
                        "driver__XtremIOFC__driver_ssl_cert_path": "",
                        "driver__XtremIOFC__driver_ssl_cert_verify": false,
                        "driver__XtremIOFC__san_ip": "",
                        "driver__XtremIOFC__san_login": "admin",
                        "driver__XtremIOFC__san_password": "",
                        "driver__XtremIOFC__xtremio_array_busy_retry_count": 5,
                        "driver__XtremIOFC__xtremio_array_busy_retry_interval": 5,
                        "driver__XtremIOFC__xtremio_clean_unused_ig": false,
                        "driver__XtremIOFC__xtremio_cluster_name": "",
                        "driver__XtremIOISCSI__driver_ssl_cert_path": "",
                        "driver__XtremIOISCSI__driver_ssl_cert_verify": false,
                        "driver__XtremIOISCSI__san_ip": "",
                        "driver__XtremIOISCSI__san_login": "admin",
                        "driver__XtremIOISCSI__san_password": "",
                        "driver__XtremIOISCSI__xtremio_array_busy_retry_count": 5,
                        "driver__XtremIOISCSI__xtremio_array_busy_retry_interval": 5,
                        "driver__XtremIOISCSI__xtremio_clean_unused_ig": false,
                        "driver__XtremIOISCSI__xtremio_cluster_name": "",
                        "driver__ZadaraVPSAISCSI__zadara_access_key": "",
                        "driver__ZadaraVPSAISCSI__zadara_default_snap_policy": false,
                        "driver__ZadaraVPSAISCSI__zadara_ssl_cert_verify": true,
                        "driver__ZadaraVPSAISCSI__zadara_use_iser": true,
                        "driver__ZadaraVPSAISCSI__zadara_vol_encrypt": false,
                        "driver__ZadaraVPSAISCSI__zadara_vol_name_template": "OS_%s",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_host": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_poolname": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_port": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_use_ssl": false
                      }
                    },
                    "sysFiles": {
                      "name": ""
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/embercsi/ember-csi-operator:latest
          createdAt: 2021-02-22:17:47:00
          description: Multi-vendor CSI plugin supporting 80+ storage drivers
          repository: https://github.com/embercsi/ember-csi-operator
          support: http://readthedocs.org/projects/ember-csi/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an Ember CSI storage backend
            displayName: Storage Backends
            kind: EmberStorageBackend
            name: embercsis.ember-csi.io
            version: v1alpha1
        description: |
          Ember-CSI is a multi-vendor CSI plugin driver supporting
          over 80 storage drivers in a single plugin to provide block and mount
          storage to container orchestration systems.

          The operator adds support for management of multiple Ember CSI instances
          within a cluster and simplifies storage backend deployments.

          ### Supported drivers
          Ember-CSI includes a good number of storage drivers, but due to limitation on hardware availability only a small number of them have been validated at one point or another. In alphabetical order they are:

          * HPE3PARFC
          * HPE3PARISCSI
          * KaminarioISCSI
          * LVMVolume
          * PowerMaxFC
          * PowerMaxISCSI
          * PureFC
          * PureISCSI
          * QnapISCSI
          * RBD
          * SolidFire
          * SynoISCSI
          * XtremIOFC
          * XtremIOISCSI

          ### Getting started
          Please have a look at our quick start documentation on how to get started:
          [github.com/embercsi/ember-csi-operator#quick-start](https://github.com/embercsi/ember-csi-operator#quick-start).
          We recommend to use the user friendly "Edit Form" when creating a new
          EmberStorageBackend. You'll find the link in the upper right corner of the
          editor.

          ### Documentation
          Documentation on Ember CSI itself can be found at [docs.ember-csi.io](http://docs.ember-csi.io/).

          ### Getting help
          Please contact us if you encounter any issues while using Ember CSI.
          Get in touch through the #ember-csi channel on Freenode, also accessible through
          [Kiwi IRC](https://kiwiirc.com/nextclient/irc.freenode.net/ember-csi).

          You can also create an issue on the [Github project](https://github.com/embercsi/ember-csi-operator)
          for bugs, enhancements, or other requests.
        displayName: Ember CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Ember-CSI
        - CSI
        links:
        - name: Learn more about the project
          url: http://ember-csi.io/
        - name: Documentation
          url: http://readthedocs.org/projects/ember-csi/
        - name: Ember-CSI Source Code
          url: https://github.com/embercsi/ember-csi
        - name: Ember-CSI Operator Source Code
          url: https://github.com/embercsi/ember-csi-operator
        maintainers:
        - email: geguileo@redhat.com
          name: Gorka Eguileor
        - email: cschwede@redhat.com
          name: Christian Schwede
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/embercsi/ember-csi-operator:latest
        version: 0.9.9
      entries:
      - name: ember-csi-community-operator.v0.9.9
        version: 0.9.9
      - name: ember-csi-community-operator.v0.9.8
        version: 0.9.8
      - name: ember-csi-community-operator.v0.9.7
        version: 0.9.7
      name: beta
    defaultChannel: beta
    packageName: ember-csi-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sqs-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sqs-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sqs.services.k8s.aws/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sqs-controller:1.2.1
          createdAt: "2025-09-19T19:14:25Z"
          description: AWS SQS controller is a service controller for managing SQS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Queue represents the state of an AWS sqs Queue resource.
            displayName: Queue
            kind: Queue
            name: queues.sqs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Simple Queue Service (SQS) resources in AWS from within your Kubernetes cluster.

          **About Amazon SQS**

          Amazon Simple Queue Service (Amazon SQS) offers a secure, durable, and available hosted queue that lets you integrate and decouple distributed software systems and components. Amazon SQS offers common constructs such as [dead-letter queues](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGuide/sqs-dead-letter-queues.html) and [cost allocation tags](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGuide/sqs-queue-tags.html). It provides a generic web services API that you can access using any programming language that the AWS SDK supports.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SQS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sqs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SQS Developer Resources
          url: https://aws.amazon.com/sqs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sqs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sqs-controller:1.2.1
        version: 1.2.1
      entries:
      - name: ack-sqs-controller.v1.2.1
        version: 1.2.1
      - name: ack-sqs-controller.v1.2.0
        version: 1.2.0
      - name: ack-sqs-controller.v1.1.16
        version: 1.1.16
      - name: ack-sqs-controller.v1.1.14
        version: 1.1.14
      - name: ack-sqs-controller.v1.1.13
        version: 1.1.13
      - name: ack-sqs-controller.v1.1.12
        version: 1.1.12
      - name: ack-sqs-controller.v1.1.11
        version: 1.1.11
      - name: ack-sqs-controller.v1.1.10
        version: 1.1.10
      - name: ack-sqs-controller.v1.1.9
        version: 1.1.9
      - name: ack-sqs-controller.v1.1.8
        version: 1.1.8
      - name: ack-sqs-controller.v1.1.7
        version: 1.1.7
      - name: ack-sqs-controller.v1.1.6
        version: 1.1.6
      - name: ack-sqs-controller.v1.1.5
        version: 1.1.5
      - name: ack-sqs-controller.v1.1.4
        version: 1.1.4
      - name: ack-sqs-controller.v1.1.3
        version: 1.1.3
      - name: ack-sqs-controller.v1.1.2
        version: 1.1.2
      - name: ack-sqs-controller.v1.1.1
        version: 1.1.1
      - name: ack-sqs-controller.v1.1.0
        version: 1.1.0
      - name: ack-sqs-controller.v1.0.17
        version: 1.0.17
      - name: ack-sqs-controller.v1.0.16
        version: 1.0.16
      - name: ack-sqs-controller.v1.0.15
        version: 1.0.15
      - name: ack-sqs-controller.v1.0.14
        version: 1.0.14
      - name: ack-sqs-controller.v1.0.13
        version: 1.0.13
      - name: ack-sqs-controller.v1.0.12
        version: 1.0.12
      - name: ack-sqs-controller.v1.0.10
        version: 1.0.10
      - name: ack-sqs-controller.v1.0.7
        version: 1.0.7
      - name: ack-sqs-controller.v1.0.5
        version: 1.0.5
      - name: ack-sqs-controller.v1.0.3
        version: 1.0.3
      - name: ack-sqs-controller.v1.0.2
        version: 1.0.2
      - name: ack-sqs-controller.v1.0.1
        version: 1.0.1
      - name: ack-sqs-controller.v1.0.0
        version: 1.0.0
      - name: ack-sqs-controller.v0.0.7
        version: 0.0.7
      - name: ack-sqs-controller.v0.0.6
        version: 0.0.6
      - name: ack-sqs-controller.v0.0.5
        version: 0.0.5
      - name: ack-sqs-controller.v0.0.4
        version: 0.0.4
      - name: ack-sqs-controller.v0.0.2
        version: 0.0.2
      - name: ack-sqs-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-sqs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/limitador-operator
    name: limitador-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: limitador-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/limitador-operator:v0.5.0
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: The Limitador operator installs and maintains limitador instances
        displayName: Limitador
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        links:
        - name: Limitador Operator
          url: https://github.com/Kuadrant/limitador-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - quay.io/kuadrant/limitador-operator:v0.5.0
        - quay.io/kuadrant/limitador:v1.2.0
        version: 0.5.0
      entries:
      - name: limitador-operator.v0.5.0
        version: 0.5.0
      - name: limitador-operator.v0.4.0
        version: 0.4.0
      - name: limitador-operator.v0.3.0
        version: 0.3.0
      name: alpha
    - currentCSV: limitador-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "name": "toy_get_route",
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/limitador-operator:v0.11.0
          createdAt: "2024-10-02T10:46:31Z"
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: The Limitador operator installs and maintains limitador instances
        displayName: Limitador
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        links:
        - name: Limitador Operator
          url: https://github.com/Kuadrant/limitador-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - quay.io/kuadrant/limitador-operator:v0.11.0
        - quay.io/kuadrant/limitador:v1.6.0
        version: 0.11.0
      entries:
      - name: limitador-operator.v0.11.0
        version: 0.11.0
      - name: limitador-operator.v0.10.0
        version: 0.10.0
      - name: limitador-operator.v0.6.0
        version: 0.6.0
      - name: limitador-operator.v0.5.0
        version: 0.5.0
      - name: limitador-operator.v0.4.0
        version: 0.4.0
      - name: limitador-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: limitador-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/limitador-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Commvault
      provider-url: ""
    name: hedvig-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hedvig-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\":\"hedvig.io/v1\",\n    \"kind\":\"HedvigDeploy\",\n
            \   \"metadata\": {\n      \"name\": \"testdeploy\",\n      \"namespace\":
            \"hedvig\"\n    },\n    \"spec\": {\n      \"cluster\": {\n        \"name\":
            \"hedvigcl\",\n        \"seeds\": [\n          \"hedvigcl1.hedvig.io\",\n
            \         \"hedvigcl2.hedvig.io\",\n          \"hedvigcl3.hedvig.io\"\n
            \       ]\n      },\n      \"k8s\": {\n        \"name\": \"k8scl\"\n      },\n
            \     \"proxy\": {\n        \"repository\": \"hedvig/hedvig-proxy\",\n
            \       \"tag\": \"v-4.3.1\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"controller\": {\n        \"repository\": \"hedvig/hedvig-csi-volume-driver\",\n
            \       \"tag\": \"v-2.0\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"node\": {\n        \"repository\": \"hedvig/hedvig-csi-volume-driver\",\n
            \       \"tag\": \"v-2.0\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"sidecars\": [\n        {\n          \"name\": \"csi-provisioner\",\n
            \         \"repository\": \"quay.io/k8scsi/csi-provisioner\",\n          \"tag\":
            \"v1.6.0\",\n          \"imagePullPolicy\": \"IfNotPresent\"\n        },\n
            \       {\n          \"name\": \"csi-snapshotter\",\n          \"repository\":
            \"quay.io/k8scsi/csi-snapshotter\",\n          \"tag\": \"v2.1.0\",\n
            \         \"imagePullPolicy\": \"IfNotPresent\"\n        },              \n
            \       {\n          \"name\": \"csi-resizer\",\n          \"repository\":
            \"quay.io/k8scsi/csi-resizer\",\n          \"tag\": \"v0.5.0\",\n          \"imagePullPolicy\":
            \"IfNotPresent\"\n        },              \n        {\n          \"name\":
            \"csi-node-driver-registrar\",\n          \"repository\": \"quay.io/k8scsi/csi-node-driver-registrar\",\n
            \         \"tag\": \"v1.0-canary\",\n          \"imagePullPolicy\": \"IfNotPresent\"\n
            \       }\n      ]\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: hedvig/hedvig-operator:2.0.0-ubi
          createdAt: ""
          description: Hedvig Operator is the official operator to deploy and manage
            Hedvig components in container orchestrators such as Kubernetes and Openshift.
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hedvig Deploy
            displayName: HedvigDeploy
            kind: HedvigDeploy
            name: hedvigdeploys.hedvig.io
            version: v1
        description: "The Hedvig Distributed Storage Platform is a software-defined,
          scale-out storage solution. It distributes data simultaneously across multiple
          locations – from on-premises data centers to the cloud – and scales capacity
          on-demand by leveraging the storage of commodity x86/ARM servers. The distributed
          write maximizes availability and protects data from hardware failures ranging
          from a single disk to an entire site, improving disaster recovery planning.\n\nHedvig
          provides a CSI Driver for accelerating your journey into the container ecosystem.
          CSI Driver enables containerized applications to maintain state by dynamically
          provisioning and consuming Hedvig virtual disks as persistent volumes. \n###
          \n\nHedvig Operator deploys the Hedvig Storage Proxy and the Hedvig CSI
          Driver to manage the life cycle of persistent storage in Kubernetes and
          Openshift.\n\n**Supported Container Orchestrators:**\n* Kubernetes 1.16+\n*
          Red Hat OpenShift 4.6+\n\n**Supported Hedvig Releases:**\n* Hedvig 4.0+\nPlease
          refer to the [documentation](https://documentation.commvault.com/commvault/hedvig/others/pdf/Hedvig_Operator_User_Guide.pdf)
          for steps to install and use the operator.\n\n\n\n\n\n"
        displayName: Hedvig Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - CSI
        - Hedvig
        - Cloud
        - Block
        - NFS
        - iSCSI
        - Snapshot
        - Volume
        links:
        - name: Documentation
          url: https://documentation.commvault.com/commvault/hedvig/article
        - name: Hedvig CSI User Guide
          url: https://documentation.commvault.com/commvault/hedvig/others/pdf/Hedvig_CSI_User_Guide.pdf
        maintainers:
        - email: ashenoy@commvault.com
          name: Abhijith Shenoy
        maturity: stable
        provider:
          name: Commvault
        relatedImages:
        - hedvig/hedvig-operator:2.0.0-ubi
        version: 1.0.2
      entries:
      - name: hedvig-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: hedvig-operator
    provider:
      name: Commvault
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: External Secrets
      provider-url: https://external-secrets.io
    name: external-secrets-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: external-secrets-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": "1234"
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "key": "provider-key",
                      "property": "provider-key-property",
                      "version": "provider-key-version"
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterExternalSecret",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "externalSecretName": "hello-world-es",
                  "externalSecretSpec": {
                    "data": [
                      {
                        "remoteRef": {
                          "key": "provider-key",
                          "property": "provider-key-property",
                          "version": "provider-key-version"
                        },
                        "secretKey": "secret-key-to-be-managed"
                      }
                    ],
                    "dataFrom": [
                      {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    ],
                    "refreshInterval": "1h",
                    "secretStoreRef": {
                      "kind": "SecretStore",
                      "name": "secret-store-name"
                    },
                    "target": {
                      "creationPolicy": "Merge",
                      "name": "my-secret",
                      "template": {
                        "data": {
                          "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                        },
                        "metadata": {
                          "annotations": {},
                          "labels": {}
                        },
                        "templateFrom": [
                          {
                            "configMap": {
                              "items": [
                                {
                                  "key": "alertmanager.yaml"
                                }
                              ],
                              "name": "alertmanager"
                            }
                          }
                        ],
                        "type": "kubernetes.io/dockerconfigjson"
                      }
                    }
                  },
                  "namespaceSelector": {
                    "matchLabels": {
                      "cool": "label"
                    }
                  },
                  "refreshTime": "1m"
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": 1234
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "extract": {
                        "conversionStrategy": "Default",
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    },
                    {
                      "find": {
                        "conversionStrategy": "Unicode",
                        "name": {
                          "regexp": ".*foobar.*"
                        },
                        "path": "path-to-filter",
                        "tags": {
                          "foo": "bar"
                        }
                      }
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "deletionPolicy": "Retain",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.external-secrets.io/v1alpha1",
                "kind": "OperatorConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "requeueInterval": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  },
                  "concurrent": 1,
                  "controllerClass": "",
                  "crds": {
                    "createClusterExternalSecret": true,
                    "createClusterSecretStore": true
                  },
                  "createOperator": true,
                  "deploymentAnnotations": {},
                  "extraArgs": {},
                  "extraEnv": [],
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/external-secrets/external-secrets",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "installCRDs": false,
                  "leaderElect": false,
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "priorityClassName": "",
                  "processClusterExternalSecret": true,
                  "processClusterStore": true,
                  "prometheus": {
                    "enabled": false,
                    "service": {
                      "port": 8080
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "scopedNamespace": "",
                  "scopedRBAC": false,
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "certCheckInterval": "5m",
                    "certDir": "/tmp/certs",
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
          createdAt: "2024-12-03T10:44:19Z"
          description: Operator to configure external-secrets helm-chart based operator
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/external-secrets/external-secrets-helm-operator
          support: external-secrets
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ACRAccessToken
            name: acraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster scoped resource that allows you to manage ExternalSecrets
              across namespaces
            displayName: ClusterExternalSecret
            kind: ClusterExternalSecret
            name: clusterexternalsecrets.external-secrets.io
            version: v1beta1
          - kind: ClusterGenerator
            name: clustergenerators.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1alpha1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1beta1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1beta1
          - kind: ECRAuthorizationToken
            name: ecrauthorizationtokens.generators.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1alpha1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1beta1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1beta1
          - kind: Fake
            name: fakes.generators.external-secrets.io
            version: v1alpha1
          - kind: GCRAccessToken
            name: gcraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - kind: GithubAccessToken
            name: githubaccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Configures external-secrets helm-chart based operator
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.external-secrets.io
            version: v1alpha1
          - kind: Password
            name: passwords.generators.external-secrets.io
            version: v1alpha1
          - kind: PushSecret
            name: pushsecrets.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1alpha1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1beta1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1beta1
          - kind: STSSessionToken
            name: stssessiontokens.generators.external-secrets.io
            version: v1alpha1
          - kind: UUID
            name: uuids.generators.external-secrets.io
            version: v1alpha1
          - kind: VaultDynamicSecret
            name: vaultdynamicsecrets.generators.external-secrets.io
            version: v1alpha1
          - kind: Webhook
            name: webhooks.generators.external-secrets.io
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK (Helm version) to configure **[official external-secrets operator helm chart](https://github.com/external-secrets/external-secrets)**, so it can be installed via OLM without having to do any change on current Helm Charts.

          Before any other resources provided by this Operator can be deployed, it is essential to create an
          OperatorConfig resource.

          The usual Helm Chart file `values.yaml`, like:
          ```yaml
          prometheus:
            enabled: true
            service:
              port: 8080
          resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          need to be encapsulated into a new custom resource called `OperatorConfig`:
          ```yaml
          apiVersion: operator.external-secrets.io/v1alpha1
          kind: OperatorConfig
          metadata:
            name: cluster
          spec:
            prometheus:
              enabled: true
              service:
                port: 8080
            resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          Once the OperatorConfig resource is deployed, the operator will create all helm chart resources, using the custom resource name as a preffix for all resources names, like for example a `Deployment` called `cluster-external-secrets`.

          ## Documentation

          * [Upstream Documentation](https://external-secrets.io)
          * [Helm Operator Install](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/install.md)
          * [Helm Operator Development](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/development.md)
          * [Helm Operator Release](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/release.md)

          ## Contributing

          You can contribute by:

          * Raising any issues you find using External Secrets Operator
          * Fixing issues by opening [Pull Requests](https://github.com/external-secrets/external-secrets-helm-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving documentation
          * Talking about External Secrets Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/external-secrets/external-secrets-helm-operator/issues).

          ## License

          External Secrets Operator is under Apache 2.0 license. See the [LICENSE](LICENSE) file for details.
        displayName: External Secrets Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secrets
        - secrets-manager
        - vault
        - aws
        - azure
        - google
        - ibm
        - akeyless
        - yandex
        - gitlab
        - oracle
        links:
        - name: External Secrets Operator
          url: https://external-secrets.io
        - name: GitHub
          url: https://github.com/external-secrets/external-secrets
        - name: Slack
          url: https://kubernetes.slack.com/messages/external-secrets
        maintainers:
        - email: contact@external-secrets.io
          name: external-secrets community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: External Secrets
          url: https://external-secrets.io
        relatedImages:
        - ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
        version: 0.11.0
      entries:
      - name: external-secrets-operator.v0.11.0
        version: 0.11.0
      - name: external-secrets-operator.v0.10.7
        version: 0.10.7
      - name: external-secrets-operator.v0.10.6
        version: 0.10.6
      - name: external-secrets-operator.v0.10.5
        version: 0.10.5
      - name: external-secrets-operator.v0.10.4
        version: 0.10.4
      - name: external-secrets-operator.v0.10.3
        version: 0.10.3
      - name: external-secrets-operator.v0.10.2
        version: 0.10.2
      - name: external-secrets-operator.v0.10.1
        version: 0.10.1
      - name: external-secrets-operator.v0.10.0
        version: 0.10.0
      - name: external-secrets-operator.v0.9.20
        version: 0.9.20
      - name: external-secrets-operator.v0.9.19
        version: 0.9.19
      - name: external-secrets-operator.v0.9.18
        version: 0.9.18
      - name: external-secrets-operator.v0.9.17
        version: 0.9.17
      - name: external-secrets-operator.v0.9.16
        version: 0.9.16
      - name: external-secrets-operator.v0.9.15-2
        version: 0.9.15-2
      - name: external-secrets-operator.v0.9.14
        version: 0.9.14
      - name: external-secrets-operator.v0.9.13
        version: 0.9.13
      - name: external-secrets-operator.v0.9.12
        version: 0.9.12
      - name: external-secrets-operator.v0.9.11
        version: 0.9.11
      - name: external-secrets-operator.v0.9.10
        version: 0.9.10
      - name: external-secrets-operator.v0.9.9
        version: 0.9.9
      - name: external-secrets-operator.v0.9.8
        version: 0.9.8
      - name: external-secrets-operator.v0.9.7
        version: 0.9.7
      - name: external-secrets-operator.v0.9.6
        version: 0.9.6
      - name: external-secrets-operator.v0.9.5
        version: 0.9.5
      - name: external-secrets-operator.v0.9.4
        version: 0.9.4
      - name: external-secrets-operator.v0.9.3
        version: 0.9.3
      - name: external-secrets-operator.v0.9.2
        version: 0.9.2
      - name: external-secrets-operator.v0.9.1
        version: 0.9.1
      - name: external-secrets-operator.v0.9.0
        version: 0.9.0
      - name: external-secrets-operator.v0.8.7
        version: 0.8.7
      - name: external-secrets-operator.v0.8.6
        version: 0.8.6
      - name: external-secrets-operator.v0.8.5
        version: 0.8.5
      - name: external-secrets-operator.v0.8.4
        version: 0.8.4
      - name: external-secrets-operator.v0.8.3
        version: 0.8.3
      - name: external-secrets-operator.v0.8.2
        version: 0.8.2
      - name: external-secrets-operator.v0.8.1
        version: 0.8.1
      - name: external-secrets-operator.v0.8.0
        version: 0.8.0
      - name: external-secrets-operator.v0.7.2
        version: 0.7.2
      - name: external-secrets-operator.v0.7.1
        version: 0.7.1
      - name: external-secrets-operator.v0.7.0
        version: 0.7.0
      - name: external-secrets-operator.v0.7.0-rc1
        version: 0.7.0-rc1
      - name: external-secrets-operator.v0.6.1
        version: 0.6.1
      - name: external-secrets-operator.v0.6.0
        version: 0.6.0
      - name: external-secrets-operator.v0.6.0-rc1
        version: 0.6.0-rc1
      - name: external-secrets-operator.v0.5.9
        version: 0.5.9
      - name: external-secrets-operator.v0.5.8
        version: 0.5.8
      - name: external-secrets-operator.v0.5.7
        version: 0.5.7
      - name: external-secrets-operator.v0.5.6
        version: 0.5.6
      - name: external-secrets-operator.v0.5.5
        version: 0.5.5
      - name: external-secrets-operator.v0.5.4
        version: 0.5.4
      - name: external-secrets-operator.v0.5.3
        version: 0.5.3
      - name: external-secrets-operator.v0.5.2
        version: 0.5.2
      - name: external-secrets-operator.v0.5.1
        version: 0.5.1
      - name: external-secrets-operator.v0.5.0
        version: 0.5.0
      - name: external-secrets-operator.v0.4.4
        version: 0.4.4
      - name: external-secrets-operator.v0.4.3
        version: 0.4.3
      - name: external-secrets-operator.v0.4.2
        version: 0.4.2
      - name: external-secrets-operator.v0.4.1
        version: 0.4.1
      - name: external-secrets-operator.v0.3.11
        version: 0.3.11
      - name: external-secrets-operator.v0.3.10
        version: 0.3.10
      - name: external-secrets-operator.v0.3.9
        version: 0.3.9
      - name: external-secrets-operator.v0.3.8
        version: 0.3.8
      name: alpha
    - currentCSV: external-secrets-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": "1234"
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "key": "provider-key",
                      "property": "provider-key-property",
                      "version": "provider-key-version"
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterExternalSecret",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "externalSecretName": "hello-world-es",
                  "externalSecretSpec": {
                    "data": [
                      {
                        "remoteRef": {
                          "key": "provider-key",
                          "property": "provider-key-property",
                          "version": "provider-key-version"
                        },
                        "secretKey": "secret-key-to-be-managed"
                      }
                    ],
                    "dataFrom": [
                      {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    ],
                    "refreshInterval": "1h",
                    "secretStoreRef": {
                      "kind": "SecretStore",
                      "name": "secret-store-name"
                    },
                    "target": {
                      "creationPolicy": "Merge",
                      "name": "my-secret",
                      "template": {
                        "data": {
                          "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                        },
                        "metadata": {
                          "annotations": {},
                          "labels": {}
                        },
                        "templateFrom": [
                          {
                            "configMap": {
                              "items": [
                                {
                                  "key": "alertmanager.yaml"
                                }
                              ],
                              "name": "alertmanager"
                            }
                          }
                        ],
                        "type": "kubernetes.io/dockerconfigjson"
                      }
                    }
                  },
                  "namespaceSelector": {
                    "matchLabels": {
                      "cool": "label"
                    }
                  },
                  "refreshTime": "1m"
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": 1234
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "extract": {
                        "conversionStrategy": "Default",
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    },
                    {
                      "find": {
                        "conversionStrategy": "Unicode",
                        "name": {
                          "regexp": ".*foobar.*"
                        },
                        "path": "path-to-filter",
                        "tags": {
                          "foo": "bar"
                        }
                      }
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "deletionPolicy": "Retain",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.external-secrets.io/v1alpha1",
                "kind": "OperatorConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "requeueInterval": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  },
                  "concurrent": 1,
                  "controllerClass": "",
                  "crds": {
                    "createClusterExternalSecret": true,
                    "createClusterSecretStore": true
                  },
                  "createOperator": true,
                  "deploymentAnnotations": {},
                  "extraArgs": {},
                  "extraEnv": [],
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/external-secrets/external-secrets",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "installCRDs": false,
                  "leaderElect": false,
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "priorityClassName": "",
                  "processClusterExternalSecret": true,
                  "processClusterStore": true,
                  "prometheus": {
                    "enabled": false,
                    "service": {
                      "port": 8080
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "scopedNamespace": "",
                  "scopedRBAC": false,
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "certCheckInterval": "5m",
                    "certDir": "/tmp/certs",
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
          createdAt: "2024-12-03T10:44:19Z"
          description: Operator to configure external-secrets helm-chart based operator
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/external-secrets/external-secrets-helm-operator
          support: external-secrets
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ACRAccessToken
            name: acraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster scoped resource that allows you to manage ExternalSecrets
              across namespaces
            displayName: ClusterExternalSecret
            kind: ClusterExternalSecret
            name: clusterexternalsecrets.external-secrets.io
            version: v1beta1
          - kind: ClusterGenerator
            name: clustergenerators.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1alpha1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1beta1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1beta1
          - kind: ECRAuthorizationToken
            name: ecrauthorizationtokens.generators.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1alpha1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1beta1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1beta1
          - kind: Fake
            name: fakes.generators.external-secrets.io
            version: v1alpha1
          - kind: GCRAccessToken
            name: gcraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - kind: GithubAccessToken
            name: githubaccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Configures external-secrets helm-chart based operator
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.external-secrets.io
            version: v1alpha1
          - kind: Password
            name: passwords.generators.external-secrets.io
            version: v1alpha1
          - kind: PushSecret
            name: pushsecrets.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1alpha1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1beta1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1beta1
          - kind: STSSessionToken
            name: stssessiontokens.generators.external-secrets.io
            version: v1alpha1
          - kind: UUID
            name: uuids.generators.external-secrets.io
            version: v1alpha1
          - kind: VaultDynamicSecret
            name: vaultdynamicsecrets.generators.external-secrets.io
            version: v1alpha1
          - kind: Webhook
            name: webhooks.generators.external-secrets.io
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK (Helm version) to configure **[official external-secrets operator helm chart](https://github.com/external-secrets/external-secrets)**, so it can be installed via OLM without having to do any change on current Helm Charts.

          Before any other resources provided by this Operator can be deployed, it is essential to create an
          OperatorConfig resource.

          The usual Helm Chart file `values.yaml`, like:
          ```yaml
          prometheus:
            enabled: true
            service:
              port: 8080
          resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          need to be encapsulated into a new custom resource called `OperatorConfig`:
          ```yaml
          apiVersion: operator.external-secrets.io/v1alpha1
          kind: OperatorConfig
          metadata:
            name: cluster
          spec:
            prometheus:
              enabled: true
              service:
                port: 8080
            resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          Once the OperatorConfig resource is deployed, the operator will create all helm chart resources, using the custom resource name as a preffix for all resources names, like for example a `Deployment` called `cluster-external-secrets`.

          ## Documentation

          * [Upstream Documentation](https://external-secrets.io)
          * [Helm Operator Install](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/install.md)
          * [Helm Operator Development](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/development.md)
          * [Helm Operator Release](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/release.md)

          ## Contributing

          You can contribute by:

          * Raising any issues you find using External Secrets Operator
          * Fixing issues by opening [Pull Requests](https://github.com/external-secrets/external-secrets-helm-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving documentation
          * Talking about External Secrets Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/external-secrets/external-secrets-helm-operator/issues).

          ## License

          External Secrets Operator is under Apache 2.0 license. See the [LICENSE](LICENSE) file for details.
        displayName: External Secrets Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secrets
        - secrets-manager
        - vault
        - aws
        - azure
        - google
        - ibm
        - akeyless
        - yandex
        - gitlab
        - oracle
        links:
        - name: External Secrets Operator
          url: https://external-secrets.io
        - name: GitHub
          url: https://github.com/external-secrets/external-secrets
        - name: Slack
          url: https://kubernetes.slack.com/messages/external-secrets
        maintainers:
        - email: contact@external-secrets.io
          name: external-secrets community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: External Secrets
          url: https://external-secrets.io
        relatedImages:
        - ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
        version: 0.11.0
      entries:
      - name: external-secrets-operator.v0.11.0
        version: 0.11.0
      - name: external-secrets-operator.v0.10.7
        version: 0.10.7
      - name: external-secrets-operator.v0.10.6
        version: 0.10.6
      - name: external-secrets-operator.v0.10.5
        version: 0.10.5
      - name: external-secrets-operator.v0.10.4
        version: 0.10.4
      - name: external-secrets-operator.v0.10.3
        version: 0.10.3
      - name: external-secrets-operator.v0.10.2
        version: 0.10.2
      - name: external-secrets-operator.v0.10.1
        version: 0.10.1
      - name: external-secrets-operator.v0.10.0
        version: 0.10.0
      - name: external-secrets-operator.v0.9.20
        version: 0.9.20
      - name: external-secrets-operator.v0.9.19
        version: 0.9.19
      - name: external-secrets-operator.v0.9.18
        version: 0.9.18
      - name: external-secrets-operator.v0.9.17
        version: 0.9.17
      - name: external-secrets-operator.v0.9.16
        version: 0.9.16
      - name: external-secrets-operator.v0.9.15-2
        version: 0.9.15-2
      - name: external-secrets-operator.v0.9.14
        version: 0.9.14
      - name: external-secrets-operator.v0.9.13
        version: 0.9.13
      - name: external-secrets-operator.v0.9.12
        version: 0.9.12
      - name: external-secrets-operator.v0.9.11
        version: 0.9.11
      - name: external-secrets-operator.v0.9.10
        version: 0.9.10
      - name: external-secrets-operator.v0.9.9
        version: 0.9.9
      - name: external-secrets-operator.v0.9.8
        version: 0.9.8
      - name: external-secrets-operator.v0.9.7
        version: 0.9.7
      - name: external-secrets-operator.v0.9.6
        version: 0.9.6
      - name: external-secrets-operator.v0.9.5
        version: 0.9.5
      - name: external-secrets-operator.v0.9.4
        version: 0.9.4
      - name: external-secrets-operator.v0.9.3
        version: 0.9.3
      - name: external-secrets-operator.v0.9.2
        version: 0.9.2
      - name: external-secrets-operator.v0.9.1
        version: 0.9.1
      - name: external-secrets-operator.v0.9.0
        version: 0.9.0
      - name: external-secrets-operator.v0.8.7
        version: 0.8.7
      - name: external-secrets-operator.v0.8.6
        version: 0.8.6
      - name: external-secrets-operator.v0.8.5
        version: 0.8.5
      - name: external-secrets-operator.v0.8.4
        version: 0.8.4
      - name: external-secrets-operator.v0.8.3
        version: 0.8.3
      - name: external-secrets-operator.v0.8.2
        version: 0.8.2
      - name: external-secrets-operator.v0.8.1
        version: 0.8.1
      - name: external-secrets-operator.v0.8.0
        version: 0.8.0
      - name: external-secrets-operator.v0.7.2
        version: 0.7.2
      - name: external-secrets-operator.v0.7.1
        version: 0.7.1
      - name: external-secrets-operator.v0.7.0
        version: 0.7.0
      - name: external-secrets-operator.v0.7.0-rc1
        version: 0.7.0-rc1
      - name: external-secrets-operator.v0.6.1
        version: 0.6.1
      - name: external-secrets-operator.v0.6.0
        version: 0.6.0
      - name: external-secrets-operator.v0.5.9
        version: 0.5.9
      - name: external-secrets-operator.v0.5.8
        version: 0.5.8
      - name: external-secrets-operator.v0.5.7
        version: 0.5.7
      - name: external-secrets-operator.v0.5.6
        version: 0.5.6
      - name: external-secrets-operator.v0.5.5
        version: 0.5.5
      - name: external-secrets-operator.v0.5.4
        version: 0.5.4
      - name: external-secrets-operator.v0.5.3
        version: 0.5.3
      - name: external-secrets-operator.v0.5.2
        version: 0.5.2
      - name: external-secrets-operator.v0.5.1
        version: 0.5.1
      - name: external-secrets-operator.v0.5.0
        version: 0.5.0
      - name: external-secrets-operator.v0.4.4
        version: 0.4.4
      - name: external-secrets-operator.v0.4.2
        version: 0.4.2
      - name: external-secrets-operator.v0.4.1
        version: 0.4.1
      - name: external-secrets-operator.v0.3.11
        version: 0.3.11
      - name: external-secrets-operator.v0.3.10
        version: 0.3.10
      - name: external-secrets-operator.v0.3.9
        version: 0.3.9
      name: stable
    defaultChannel: alpha
    packageName: external-secrets-operator
    provider:
      name: External Secrets
      url: https://external-secrets.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/kuadrant-operator
    name: rhcl-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhcl-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "authpolicy-sample"
                },
                "spec": {
                  "rules": {
                    "authentication": {
                      "apikey": {
                        "apiKey": {
                          "selector": {}
                        },
                        "credentials": {
                          "authorizationHeader": {
                            "prefix": "APIKEY"
                          }
                        }
                      }
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "DNSPolicy",
                "metadata": {
                  "name": "dnspolicy-sample"
                },
                "spec": {
                  "healthCheck": {
                    "protocol": "HTTP"
                  },
                  "providerRefs": [
                    {
                      "name": "provider-ref"
                    }
                  ],
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "ratelimitpolicy-sample"
                },
                "spec": {
                  "limits": {
                    "toys": {
                      "rates": [
                        {
                          "limit": 50,
                          "window": "1m"
                        }
                      ]
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "TLSPolicy",
                "metadata": {
                  "name": "tlspolicy-sample"
                },
                "spec": {
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "self-signed-ca"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/rhcl-rhel9-operator@sha256:92f14fd8de74bc72c407947cf5fa4739a7b455e5d351baeb64521e4c07f02a9e
          createdAt: "2025-04-28T13:10:44Z"
          description: Red Hat Connectivity Link enables you to secure, protect, and
            connect your APIs and applications in multicluster, multicloud, and hybrid
            cloud environments
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthPolicy enables authentication and authorization for service
              workloads in a Gateway API network
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1
          - description: DNSPolicy configures how North-South based traffic should
              be balanced and reach the gateways
            displayName: DNSPolicy
            kind: DNSPolicy
            name: dnspolicies.kuadrant.io
            version: v1
          - description: Kuadrant configures installations of Kuadrant Service Protection
              components
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: RateLimitPolicy enables rate limiting for service workloads
              in a Gateway API network
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1
          - description: TLSPolicy provides tls for gateway listeners by managing
              the lifecycle of tls certificates
            displayName: TLSPolicy
            kind: TLSPolicy
            name: tlspolicies.kuadrant.io
            version: v1
        description: Red Hat Connectivity Link enables you to secure, protect, and
          connect your APIs and applications in multicluster, multicloud, and hybrid
          cloud environments
        displayName: Red Hat Connectivity Link
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: GitHub
          url: https://github.com/kuadrant/kuadrant-operator
        - name: Documentation
          url: https://github.com/Kuadrant/kuadrant-operator/tree/main/docs
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - registry.redhat.io/rhcl-1/rhcl-console-plugin-rhel9@sha256:ae92d638d993d782986e761d5aa3c7c5da2b8af0aabc1decb332a4f1487dacd3
        - registry.redhat.io/rhcl-1/rhcl-rhel9-operator@sha256:92f14fd8de74bc72c407947cf5fa4739a7b455e5d351baeb64521e4c07f02a9e
        - registry.redhat.io/rhcl-1/wasm-shim-rhel9@sha256:845bb8af57f3d219aa09b9c0bb20fa945306f3b413e34e196760c8cdf624a532
        version: 1.1.0
      entries:
      - name: rhcl-operator.v1.1.0
        version: 1.1.0
      - name: rhcl-operator.v1.0.2
        version: 1.0.2
      - name: rhcl-operator.v1.0.1
        version: 1.0.1
      - name: rhcl-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: rhcl-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/kuadrant-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: akaris
      provider-url: https://github.com/andreaskaris/sosreport-operator
    name: sosreport-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sosreport-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "support.openshift.io/v1alpha1",
                "kind": "Sosreport",
                "metadata": {
                  "name": "sosreport-sample"
                },
                "spec": {
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Sosreport is the Schema for the sosreports API
            displayName: Sosreport
            kind: Sosreport
            name: sosreports.support.openshift.io
            version: v1alpha1
        description: Operator that facilitates the generation of sosreports
        displayName: sosreport-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sosreport
        - debugging
        - troubleshooting
        links:
        - name: Sosreport Operator
          url: https://github.com/andreaskaris/sosreport-operator
        maintainers:
        - email: ak.karis@gmail.com
          name: Andreas Karis
        maturity: alpha
        provider:
          name: akaris
          url: https://github.com/andreaskaris/sosreport-operator
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/akaris/sosreport-operator:0.0.2
        version: 0.0.2
      entries:
      - name: sosreport-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: sosreport-operator
    provider:
      name: akaris
      url: https://github.com/andreaskaris/sosreport-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-rds-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-rds-controller.v1.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rds.services.k8s.aws/v1alpha1",
                "kind": "DBParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "rds.services.k8s.aws/v1alpha1",
                "kind": "DBSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/rds-controller:1.6.2
          createdAt: "2025-09-22T17:27:42Z"
          description: AWS RDS controller is a service controller for managing RDS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DBClusterEndpoint represents the state of an AWS rds DBClusterEndpoint
              resource.
            displayName: DBClusterEndpoint
            kind: DBClusterEndpoint
            name: dbclusterendpoints.rds.services.k8s.aws
            version: v1alpha1
          - description: DBClusterParameterGroup represents the state of an AWS rds
              DBClusterParameterGroup resource.
            displayName: DBClusterParameterGroup
            kind: DBClusterParameterGroup
            name: dbclusterparametergroups.rds.services.k8s.aws
            version: v1alpha1
          - description: DBCluster represents the state of an AWS rds DBCluster resource.
            displayName: DBCluster
            kind: DBCluster
            name: dbclusters.rds.services.k8s.aws
            version: v1alpha1
          - description: DBClusterSnapshot represents the state of an AWS rds DBClusterSnapshot
              resource.
            displayName: DBClusterSnapshot
            kind: DBClusterSnapshot
            name: dbclustersnapshots.rds.services.k8s.aws
            version: v1alpha1
          - description: DBInstance represents the state of an AWS rds DBInstance
              resource.
            displayName: DBInstance
            kind: DBInstance
            name: dbinstances.rds.services.k8s.aws
            version: v1alpha1
          - description: DBParameterGroup represents the state of an AWS rds DBParameterGroup
              resource.
            displayName: DBParameterGroup
            kind: DBParameterGroup
            name: dbparametergroups.rds.services.k8s.aws
            version: v1alpha1
          - description: DBProxy represents the state of an AWS rds DBProxy resource.
            displayName: DBProxy
            kind: DBProxy
            name: dbproxies.rds.services.k8s.aws
            version: v1alpha1
          - description: DBSnapshot represents the state of an AWS rds DBSnapshot
              resource.
            displayName: DBSnapshot
            kind: DBSnapshot
            name: dbsnapshots.rds.services.k8s.aws
            version: v1alpha1
          - description: DBSubnetGroup represents the state of an AWS rds DBSubnetGroup
              resource.
            displayName: DBSubnetGroup
            kind: DBSubnetGroup
            name: dbsubnetgroups.rds.services.k8s.aws
            version: v1alpha1
          - description: GlobalCluster represents the state of an AWS rds GlobalCluster
              resource.
            displayName: GlobalCluster
            kind: GlobalCluster
            name: globalclusters.rds.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Relational Database Service ("RDS") resources in AWS from within your Kubernetes cluster.

          **About Amazon RDS**

          Amazon Relational Database Service (Amazon RDS) makes it easy to set up, operate, and scale a relational database in the cloud. It provides cost-efficient and resizable capacity while automating time-consuming administration tasks such as hardware provisioning, database setup, patching and backups. It frees you to focus on your applications so you can give them the fast performance, high availability, security and compatibility they need.

          Amazon RDS is available on several database instance types - optimized for memory, performance or I/O -and provides you with six familiar database engines to choose from, including Amazon Aurora, PostgreSQL, MySQL, MariaDB, Oracle Database, and SQL Server. You can use the AWS Database Migration Service to easily migrate or replicate your existing databases to Amazon RDS.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon RDS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rds
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon RDS Developer Resources
          url: https://aws.amazon.com/rds/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: rds maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/rds-controller:1.6.2
        version: 1.6.2
      entries:
      - name: ack-rds-controller.v1.6.2
        version: 1.6.2
      - name: ack-rds-controller.v1.6.1
        version: 1.6.1
      - name: ack-rds-controller.v1.6.0
        version: 1.6.0
      - name: ack-rds-controller.v1.5.0
        version: 1.5.0
      - name: ack-rds-controller.v1.4.26
        version: 1.4.26
      - name: ack-rds-controller.v1.4.25
        version: 1.4.25
      - name: ack-rds-controller.v1.4.24
        version: 1.4.24
      - name: ack-rds-controller.v1.4.22
        version: 1.4.22
      - name: ack-rds-controller.v1.4.21
        version: 1.4.21
      - name: ack-rds-controller.v1.4.20
        version: 1.4.20
      - name: ack-rds-controller.v1.4.19
        version: 1.4.19
      - name: ack-rds-controller.v1.4.18
        version: 1.4.18
      - name: ack-rds-controller.v1.4.17
        version: 1.4.17
      - name: ack-rds-controller.v1.4.16
        version: 1.4.16
      - name: ack-rds-controller.v1.4.15
        version: 1.4.15
      - name: ack-rds-controller.v1.4.14
        version: 1.4.14
      - name: ack-rds-controller.v1.4.13
        version: 1.4.13
      - name: ack-rds-controller.v1.4.12
        version: 1.4.12
      - name: ack-rds-controller.v1.4.11
        version: 1.4.11
      - name: ack-rds-controller.v1.4.10
        version: 1.4.10
      - name: ack-rds-controller.v1.4.8
        version: 1.4.8
      - name: ack-rds-controller.v1.4.7
        version: 1.4.7
      - name: ack-rds-controller.v1.4.6
        version: 1.4.6
      - name: ack-rds-controller.v1.4.4
        version: 1.4.4
      - name: ack-rds-controller.v1.4.3
        version: 1.4.3
      - name: ack-rds-controller.v1.4.2
        version: 1.4.2
      - name: ack-rds-controller.v1.4.1
        version: 1.4.1
      - name: ack-rds-controller.v1.4.0
        version: 1.4.0
      - name: ack-rds-controller.v1.3.0
        version: 1.3.0
      - name: ack-rds-controller.v1.2.5
        version: 1.2.5
      - name: ack-rds-controller.v1.2.4
        version: 1.2.4
      - name: ack-rds-controller.v1.2.3
        version: 1.2.3
      - name: ack-rds-controller.v1.2.2
        version: 1.2.2
      - name: ack-rds-controller.v1.2.1
        version: 1.2.1
      - name: ack-rds-controller.v1.2.0
        version: 1.2.0
      - name: ack-rds-controller.v1.1.11
        version: 1.1.11
      - name: ack-rds-controller.v1.1.9
        version: 1.1.9
      - name: ack-rds-controller.v1.1.8
        version: 1.1.8
      - name: ack-rds-controller.v1.1.7
        version: 1.1.7
      - name: ack-rds-controller.v1.1.6
        version: 1.1.6
      - name: ack-rds-controller.v1.1.5
        version: 1.1.5
      - name: ack-rds-controller.v1.1.4
        version: 1.1.4
      - name: ack-rds-controller.v1.1.3
        version: 1.1.3
      - name: ack-rds-controller.v1.1.2
        version: 1.1.2
      - name: ack-rds-controller.v1.1.1
        version: 1.1.1
      - name: ack-rds-controller.v1.1.0
        version: 1.1.0
      - name: ack-rds-controller.v1.0.0
        version: 1.0.0
      - name: ack-rds-controller.v0.1.3
        version: 0.1.3
      - name: ack-rds-controller.v0.1.1
        version: 0.1.1
      - name: ack-rds-controller.v0.1.0
        version: 0.1.0
      - name: ack-rds-controller.v0.0.30
        version: 0.0.30
      - name: ack-rds-controller.v0.0.29
        version: 0.0.29
      - name: ack-rds-controller.v0.0.28
        version: 0.0.28
      - name: ack-rds-controller.v0.0.27
        version: 0.0.27
      - name: ack-rds-controller.v0.0.26
        version: 0.0.26
      - name: ack-rds-controller.v0.0.25
        version: 0.0.25
      - name: ack-rds-controller.v0.0.24
        version: 0.0.24
      - name: ack-rds-controller.v0.0.23
        version: 0.0.23
      - name: ack-rds-controller.v0.0.22
        version: 0.0.22
      - name: ack-rds-controller.v0.0.21
        version: 0.0.21
      - name: ack-rds-controller.v0.0.20
        version: 0.0.20
      - name: ack-rds-controller.v0.0.19
        version: 0.0.19
      - name: ack-rds-controller.v0.0.18
        version: 0.0.18
      - name: ack-rds-controller.v0.0.17
        version: 0.0.17
      - name: ack-rds-controller.v0.0.16
        version: 0.0.16
      - name: ack-rds-controller.v0.0.15
        version: 0.0.15
      - name: ack-rds-controller.v0.0.14
        version: 0.0.14
      - name: ack-rds-controller.v0.0.13
        version: 0.0.13
      - name: ack-rds-controller.v0.0.12
        version: 0.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-rds-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: integrity-shield-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: integrity-shield-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/open-cluster-management/integrity-shield-observer",
                    "imagePullPolicy": "Always",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "provenanece": true,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  resp.status_code == 200\n  result := json.unmarshal(resp.raw_body)\n  result.allow == false\n  not is_detect_mode\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"tls_insecure_skip_verify\": true\n  })\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "requestHandlerConfig": "defaultConstraintAction:\n  audit:\n    inform: true\n  admissionControl:\n    enforce: false\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/open-cluster-management/integrity-shield-api",
                    "imagePullPolicy": "Always",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/open-cluster-management/integrity-shield-operator:0.3.0
          createdAt: "2021-10-14T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.

          [K8s Integrity Shield](https://github.com/open-cluster-management/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.

          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Detect (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster

          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).

          ## Preparations before installation

          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).


          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.

          If you want to change the settings such as default run mode (detection/enforcement) or audit interval,  please check [here](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).

          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-7b7f768bf7-fhrpg                           1/1     Running   0          20s
          integrity-shield-observer-5bc66f75f7-tn8fw                      1/1     Running   0          25s
          integrity-shield-operator-controller-manager-65b7fb58f7-j25zd   2/2     Running   0          3h5m
          ```

          After installation, you can protect Kubernetes resources by following this [document](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).

          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:

          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)

          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/open-cluster-management/integrity-shield
        - name: Operator Documentation
          url: https://github.com/open-cluster-management/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/open-cluster-management/integrity-shield-operator:0.3.0
        version: 0.3.0
      entries:
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.0
    - currentCSV: integrity-shield-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/open-cluster-management/integrity-shield-observer",
                    "imagePullPolicy": "Always",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  resp.status_code == 200\n  result := json.unmarshal(resp.raw_body)\n  result.allow == false\n  not is_detect_mode\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"tls_insecure_skip_verify\": true\n  })\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: detect\n  admissionOnly: false\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/open-cluster-management/integrity-shield-api",
                    "imagePullPolicy": "Always",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/open-cluster-management/integrity-shield-operator:0.3.1
          createdAt: "2021-10-22T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.

          [K8s Integrity Shield](https://github.com/open-cluster-management/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.

          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Detect (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster

          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).

          ## Preparations before installation

          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).


          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.

          If you want to change the settings such as default run mode (detection/enforcement) or audit interval,  please check [here](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).

          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-7b7f768bf7-fhrpg                           1/1     Running   0          20s
          integrity-shield-observer-5bc66f75f7-tn8fw                      1/1     Running   0          25s
          integrity-shield-operator-controller-manager-65b7fb58f7-j25zd   2/2     Running   0          3h5m
          ```

          After installation, you can protect Kubernetes resources by following this [document](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).

          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:

          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)

          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/open-cluster-management/integrity-shield
        - name: Operator Documentation
          url: https://github.com/open-cluster-management/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/open-cluster-management/integrity-shield-operator:0.3.1
        version: 0.3.1
      entries:
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.1
    - currentCSV: integrity-shield-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/stolostron/integrity-shield-observer",
                    "imagePullPolicy": "IfNotPresent",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  should_be_denied(resp)\n  not is_detect_mode\n  result := get_result(resp)\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"timeout\": \"2500ms\",\n    \"raise_error\": false,\n    \"tls_ca_cert\": REPLACE_WITH_CA_CERT\n  })\n}\n\n# get decision result from API response\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  not is_null(raw_body)\n  result := json.unmarshal(raw_body)\n}\n# if error, return it as result\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  is_null(raw_body)\n  error_msg := object.get(resp, \"error\", \"\")\n  result := {\"http_error\":error_msg}\n}\n\n# if API response is null, deny\nshould_be_denied(resp) {\n  is_null(resp)\n}\n\n# if API response has status_code other than 200, deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  resp.status_code != 200\n}\n\n# if decision result is \"allow == false\" or not set, then deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  result := get_result(resp)\n  allow := object.get(result, \"allow\", false)\n  allow == false\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "reporter": {
                    "image": "quay.io/stolostron/integrity-shield-reporter",
                    "imagePullPolicy": "IfNotPresent",
                    "intervalSeconds": "10",
                    "name": "integrity-shield-reporter",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: inform\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\ndecisionReporterConfig:\n  enabled: true\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/stolostron/integrity-shield-api",
                    "imagePullPolicy": "IfNotPresent",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/stolostron/integrity-shield-operator:0.3.2
          createdAt: "2022-01-25T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.15.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.
          [K8s Integrity Shield](https://github.com/stolostron/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.
          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Inform (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster
          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).
          ## Preparations before installation
          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).
          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.
          If you want to change the settings such as default run mode (inform/enforce) or audit interval,  please check [here](https://github.com/stolostron/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).
          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-5799f69896-zbcjs                           2/2     Running   0          88s
          integrity-shield-observer-757f58f7d7-7pf5h                      1/1     Running   0          55s
          integrity-shield-operator-controller-manager-6f94d7d684-dg5bt   2/2     Running   0          4m27s
          ```
          After installation, you can protect Kubernetes resources by following this [document](https://github.com/stolostron/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).
          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:
          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)
          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/stolostron/integrity-shield
        - name: Operator Documentation
          url: https://github.com/stolostron/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/stolostron/integrity-shield-operator:0.3.2
        version: 0.3.2
      entries:
      - name: integrity-shield-operator.v0.3.2
        version: 0.3.2
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.2
    - currentCSV: integrity-shield-operator.v0.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": false,
                    "image": "quay.io/stolostron/integrity-shield-observer",
                    "imagePullPolicy": "IfNotPresent",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  should_be_denied(resp)\n  not is_detect_mode\n  result := get_result(resp)\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"timeout\": \"2500ms\",\n    \"raise_error\": false,\n    \"tls_ca_cert\": REPLACE_WITH_CA_CERT\n  })\n}\n\n# get decision result from API response\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  not is_null(raw_body)\n  result := json.unmarshal(raw_body)\n}\n# if error, return it as result\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  is_null(raw_body)\n  error_msg := object.get(resp, \"error\", \"\")\n  result := {\"http_error\":error_msg}\n}\n\n# if API response is null, deny\nshould_be_denied(resp) {\n  is_null(resp)\n}\n\n# if API response has status_code other than 200, deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  resp.status_code != 200\n}\n\n# if decision result is \"allow == false\" or not set, then deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  result := get_result(resp)\n  allow := object.get(result, \"allow\", false)\n  allow == false\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "reporter": {
                    "image": "quay.io/stolostron/integrity-shield-reporter",
                    "imagePullPolicy": "IfNotPresent",
                    "intervalSeconds": "10",
                    "name": "integrity-shield-reporter",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: inform\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\ndecisionReporterConfig:\n  enabled: true\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/stolostron/integrity-shield-api",
                    "imagePullPolicy": "IfNotPresent",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/stolostron/integrity-shield-operator:0.3.3
          createdAt: "2022-03-18T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.15.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.
          [K8s Integrity Shield](https://github.com/stolostron/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.
          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Inform (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster
          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).
          ## Preparations before installation
          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).
          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.
          If you want to change the settings such as default run mode (inform/enforce) or audit interval,  please check [here](https://github.com/stolostron/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).
          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-5799f69896-zbcjs                           2/2     Running   0          88s
          integrity-shield-observer-757f58f7d7-7pf5h                      1/1     Running   0          55s
          integrity-shield-operator-controller-manager-6f94d7d684-dg5bt   2/2     Running   0          4m27s
          ```
          After installation, you can protect Kubernetes resources by following this [document](https://github.com/stolostron/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).
          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:
          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)
          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/stolostron/integrity-shield
        - name: Operator Documentation
          url: https://github.com/stolostron/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/stolostron/integrity-shield-operator:0.3.3
        version: 0.3.3
      entries:
      - name: integrity-shield-operator.v0.3.3
        version: 0.3.3
      - name: integrity-shield-operator.v0.3.2
        version: 0.3.2
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.3
    defaultChannel: alpha-0.3.3
    packageName: integrity-shield-operator
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datasance
      provider-url: ""
    name: datasance-pot-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datasance-pot-operator.v3.4.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasance.com/v3",
                "kind": "ControlPlane",
                "metadata": {
                  "name": "pot"
                },
                "spec": {
                  "replicas": {
                    "controller": 2
                  },
                  "database": {
                    "provider": "postgres",
                    "user": "user",
                    "host": "postgre.host",
                    "port": 5432,
                    "password": "password",
                    "databaseName": "dbname"
                  },
                  "auth": {
                    "url": "auth.url",
                    "realm": "realm",
                    "realmKey": "realmKey",
                    "ssl": "none",
                    "controllerClient": "controllerClient",
                    "controllerSecret": "controllerClientSecret",
                    "viewerClient": "viewerClient"
                  },
                  "images": {
                    "pullSecret": "pullsecret",
                    "controller": "ghcr.io/datasance/controller@sha256:aa896bae6c3e6eda70938a03aaa649226356d9e4a1d03ba427d4e4ee3090ca6c",
                    "portManager": "ghcr.io/datasance/port-manager@sha256:4bf3793e9e3cd10df9b46c459e983631393825ca788ffbbb3ec3659158ad61fe",
                    "proxy": "ghcr.io/datasance/proxy@sha256:276dbed0c8ef4f5b03fdf29d5c863b6f3efcae50444ce8123220c5673272860f",
                    "router": "ghcr.io/datasance/router@sha256:82657fc8f39d33d8ef0161cf27d705cbdc0a97dca07b22a5fe78fd4213ef0d31"
                  },
                  "services": {
                    "controller": {
                      "type": "LoadBalancer",
                      "annotations": {
                        "service.beta.kubernetes.io/azure-load-balancer-internal": "false"
                      }
                    },
                    "proxy": {
                      "type": "LoadBalancer",
                      "annotations": {
                        "service.beta.kubernetes.io/azure-load-balancer-internal": "false"
                      }
                    },
                    "router": {
                      "type": "LoadBalancer",
                      "annotations": {
                        "service.beta.kubernetes.io/azure-load-balancer-internal": "false"
                      }
                    }
                  },
                  "controller": {
                    "ecnViewerUrl": "pot.demo.datasance.com",
                    "https": true,
                    "secretName": "secretName"
                  },
                  "ingresses": {
                    "controller": {
                      "annotations": {
                        "cert-manager.io/cluster-issuer": "letsencrypt"
                      },
                      "ingressClassName": "nginx",
                      "host": "pot.demo.datasance.com",
                      "secretName": "secretName"
                    },
                    "router": {
                      "address": "pot.demo.datasance.com",
                      "messagePort": 5672,
                      "interiorPort": 55672,
                      "edgePort": 45672
                    },
                    "httpProxy": {
                      "address": "pot.demo.datasance.com"
                    },
                    "tcpProxy": {
                      "address": "pot.demo.datasance.com"
                    }
                  }
                },
                "status": {
                  "conditions": [
                    {
                      "lastTransitionTime": "2022-04-10T22:44:09Z",
                      "message": "",
                      "reason": "initial_status",
                      "status": "True",
                      "type": "deploying"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          containerImage: ghcr.io/datasance/operator@sha256:143b9307c2520ef16b40c4c92e6faee8b24b7fb3a97682231e53236ddd532687
          createdAt: "2024-11-04T20:45:43Z"
          description: Datasance PoT is an Enteprise version of Eclipse ioFog Platform.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Datasance PoT Enterprise Open
            Source Subscriptions based on number of Edge device that ControlPlane
            manages"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/datasance/iofog-operator
          support: Datasance Teknoloji A.S.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Datasance PoT supported ioFog ControlPlane CRD for managing
              container workloads on Edge device
            kind: ControlPlane
            name: controlplanes.datasance.com
            version: v3
        description: ioFog Control Plane on Kubernetes
        displayName: iofog-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iofog
        - datasance
        - pot
        links:
        - name: Datasance PoT // Iofog Operator
          url: https://docs.datasance.com
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Datasance
        relatedImages:
        - ghcr.io/datasance/operator@sha256:143b9307c2520ef16b40c4c92e6faee8b24b7fb3a97682231e53236ddd532687
        - ghcr.io/datasance/controller@sha256:aa896bae6c3e6eda70938a03aaa649226356d9e4a1d03ba427d4e4ee3090ca6c
        - ghcr.io/datasance/router@sha256:82657fc8f39d33d8ef0161cf27d705cbdc0a97dca07b22a5fe78fd4213ef0d31
        - ghcr.io/datasance/proxy@sha256:276dbed0c8ef4f5b03fdf29d5c863b6f3efcae50444ce8123220c5673272860f
        - ghcr.io/datasance/port-manager@sha256:4bf3793e9e3cd10df9b46c459e983631393825ca788ffbbb3ec3659158ad61fe
        version: 3.4.9
      entries:
      - name: datasance-pot-operator.v3.4.9
        version: 3.4.9
      name: stable
    defaultChannel: stable
    packageName: datasance-pot-operator
    provider:
      name: Datasance
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: gcp-filestore-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gcp-filestore-csi-driver-operator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel8-operator@sha256:1ec61d2988b95a10260c7932e9dc6a9d1d5b3eb4148b22afd438aa4ee75dc9b3
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure GCP Filestore CSI driver.
          olm.skipRange: '>=4.11.0-0 <4.13.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/gcp-filestore-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Operator that installs and configures the CSI driver for GCP Filestore. The GCP Filestore CSI Driver Operator is a Technology Preview feature. For more information on Red Hat Technology Preview features support scope, see https://access.redhat.com/support/offerings/techpreview/.
        displayName: GCP Filestore CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - filestore
        links:
        - name: Documentation
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: preview
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:700c0e50f86a70f506403aba75e020a17530ada284af7eee77d4b4ce3dbf4102
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:4f1e873080754ae9aebc361cd77f965e1ade369403346aa0cf4cf502082d18e4
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:ee45e11a6fe8cf60f3ac285cba1b23459ca95068bd9769042ae052c778c3e080
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:537fbc8fe2b212dc798361099d2c6fd981cc846c1cf5454b703ba08d488b61f8
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel8@sha256:a1a4e7b93d6609561a7a2cd732ce772bf0047448e5019298d66c4b985baecfb2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:767682dd3bd65651a8204c9fb732b9c48b99127189992b43abc6ecce027f4589
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel8-operator@sha256:1ec61d2988b95a10260c7932e9dc6a9d1d5b3eb4148b22afd438aa4ee75dc9b3
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:6f1b497772b953468f0b816face8810ace8ea44fa91dd4a3b3e583da2cac350c
        version: 4.13.0-202309181427
      entries:
      - name: gcp-filestore-csi-driver-operator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: preview
    - currentCSV: gcp-filestore-csi-driver-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9-operator@sha256:2dc13de58b391237dc48b9335d88b99f5e26d0e5edc5a664c7d39f16ca9b297d
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure GCP Filestore CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.11.0-0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/gcp-filestore-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Operator that installs and configures the CSI driver for GCP Filestore.
        displayName: GCP Filestore CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - filestore
        links:
        - name: Documentation
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:a5704d26d0b562dba9633d1c107f34e05c750c0e3a64dbcace4dab7b964fff8d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9@sha256:8e801d924071c8f2513555450e688bff60d4100b2ca56d2ba9517038a1d48d4d
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:9c3c3d2b857ce289324aee56df1ba4ab5e5ea1e76cd86ac988233463915834b4
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:78e02d937740d742eb33b08c7317f211c25020c375069d5a97344107716608d3
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:b28def976fc155dff57ff3003cc462a8e9923b325af0b319b65220c66688699e
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0627c09fe3f5a015f117fb30af1f4d17340aefa48b23994737d3901a804b63d3
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9-operator@sha256:2dc13de58b391237dc48b9335d88b99f5e26d0e5edc5a664c7d39f16ca9b297d
        version: 4.16.0-202509240123
      entries:
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202508270905
        version: 4.16.0-202508270905
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202504121405
        version: 4.16.0-202504121405
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: gcp-filestore-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-prometheusservice-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-prometheusservice-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "prometheusservice.services.k8s.aws/v1alpha1",
                "kind": "RuleGroupsNamespace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "prometheusservice.services.k8s.aws/v1alpha1",
                "kind": "Workspace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/prometheusservice-controller:1.3.1
          createdAt: "2025-09-19T21:11:38Z"
          description: AWS Prometheus controller is a service controller for managing
            Prometheus resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertManagerDefinition represents the state of an AWS prometheusservice
              AlertManagerDefinition resource.
            displayName: AlertManagerDefinition
            kind: AlertManagerDefinition
            name: alertmanagerdefinitions.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: LoggingConfiguration represents the state of an AWS prometheusservice
              LoggingConfiguration resource.
            displayName: LoggingConfiguration
            kind: LoggingConfiguration
            name: loggingconfigurations.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: RuleGroupsNamespace represents the state of an AWS prometheusservice
              RuleGroupsNamespace resource.
            displayName: RuleGroupsNamespace
            kind: RuleGroupsNamespace
            name: rulegroupsnamespaces.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: Workspace represents the state of an AWS prometheusservice
              Workspace resource.
            displayName: Workspace
            kind: Workspace
            name: workspaces.prometheusservice.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Prometheus resources in AWS from within your Kubernetes cluster.

          **About Amazon Managed Service for Prometheus**

          Amazon Managed Service for Prometheus is a serverless, Prometheus-compatible monitoring service for container metrics that makes it easier to securely monitor container environments at scale. With Amazon Managed Service for Prometheus, you can use the same open-source Prometheus data model and query language that you use today to monitor the performance of your containerized workloads, and also enjoy improved scalability, availability, and security without having to manage the underlying infrastructure.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Prometheus
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - prometheusservice
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Prometheus Developer Resources
          url: https://aws.amazon.com/prometheus/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: prometheus maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/prometheusservice-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-prometheusservice-controller.v1.3.1
        version: 1.3.1
      - name: ack-prometheusservice-controller.v1.3.0
        version: 1.3.0
      - name: ack-prometheusservice-controller.v1.2.27
        version: 1.2.27
      - name: ack-prometheusservice-controller.v1.2.26
        version: 1.2.26
      - name: ack-prometheusservice-controller.v1.2.24
        version: 1.2.24
      - name: ack-prometheusservice-controller.v1.2.23
        version: 1.2.23
      - name: ack-prometheusservice-controller.v1.2.22
        version: 1.2.22
      - name: ack-prometheusservice-controller.v1.2.20
        version: 1.2.20
      - name: ack-prometheusservice-controller.v1.2.19
        version: 1.2.19
      - name: ack-prometheusservice-controller.v1.2.17
        version: 1.2.17
      - name: ack-prometheusservice-controller.v1.2.16
        version: 1.2.16
      - name: ack-prometheusservice-controller.v1.2.15
        version: 1.2.15
      - name: ack-prometheusservice-controller.v1.2.14
        version: 1.2.14
      - name: ack-prometheusservice-controller.v1.2.13
        version: 1.2.13
      - name: ack-prometheusservice-controller.v1.2.12
        version: 1.2.12
      - name: ack-prometheusservice-controller.v1.2.11
        version: 1.2.11
      - name: ack-prometheusservice-controller.v1.2.10
        version: 1.2.10
      - name: ack-prometheusservice-controller.v1.2.9
        version: 1.2.9
      - name: ack-prometheusservice-controller.v1.2.6
        version: 1.2.6
      - name: ack-prometheusservice-controller.v1.2.5
        version: 1.2.5
      - name: ack-prometheusservice-controller.v1.2.4
        version: 1.2.4
      - name: ack-prometheusservice-controller.v1.2.3
        version: 1.2.3
      - name: ack-prometheusservice-controller.v1.2.2
        version: 1.2.2
      - name: ack-prometheusservice-controller.v1.2.1
        version: 1.2.1
      - name: ack-prometheusservice-controller.v1.2.0
        version: 1.2.0
      - name: ack-prometheusservice-controller.v1.1.0
        version: 1.1.0
      - name: ack-prometheusservice-controller.v1.0.0
        version: 1.0.0
      - name: ack-prometheusservice-controller.v0.1.3
        version: 0.1.3
      - name: ack-prometheusservice-controller.v0.1.2
        version: 0.1.2
      - name: ack-prometheusservice-controller.v0.1.1
        version: 0.1.1
      - name: ack-prometheusservice-controller.v0.1.0
        version: 0.1.0
      - name: ack-prometheusservice-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-prometheusservice-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: awss3-operator-registry
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: awss3operator.1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/screeley44/aws-s3-provisioner:v1.0.0
          createdAt: "2019-05-02 16:12:00"
          description: Manage the full lifecycle of installing, configuring and managing
            AWS S3 Provisioner.
          repository: https://github.com/yard-turkey/aws-s3-provisioner
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - description: instance of an AWS S3 Bucket
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
          - description: Request for an AWS S3 Bucket
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
        description: |
          AWS S3 Operator will deploy the AWS S3 Provisioner which will dynamically or statically
          provision AWS S3 Bucket storage and access. The operator deploys the ObjectBucket (OB) and ObjectBucketClaim (OBC)
          CustomResourceDefinitions. The OB/OBC model follows the traditional Kubernetes PV/PVC pattern, when an OBC is detected
          the operator will act on the OBC to either provision a brand new S3 Bucket in AWS or gain access to an existing
          S3 Bucket in AWS. The operator produces an OB and ConfigMap and Secret which can then be consumed by application pods.

          **Important Note**: Currently, while in preview, this operator does not
          support automatic upgrades. You must remove the old version of the operator
          manually before installing a new version.
          ## Using AWS S3 Operator

          ### Administrator Creates Secret
          This secret will contain the elevated/admin privileges needed by the provisioner
          to properly access and create S3 Buckets and IAM users and policies. The AWS Access ID
          and AWS Secret Key will be needed for this.
          1. Create the Kubernetes Secret for the Provisioner's Owner Access.
          ```yaml
          apiVersion: v1
          kind: Secret
          metadata:
            name: s3-bucket-owner [1]
            namespace: s3-provisioner [2]
          type: Opaque
          data:
            AWS_ACCESS_KEY_ID: *base64 encoded value* [3]
            AWS_SECRET_ACCESS_KEY: *base64 encoded value* [4]
          ```
          1. Name of the secret, this will be referenced in StorageClass.
          1. Namespace where the Secret will exist.
          1. Your AWS_ACCESS_KEY_ID base64 encoded.
          1. Your AWS_SECRET_ACCESS_KEY base64 encoded.
          ```
           # kubectl create -f creds.yaml
          secret/s3-bucket-owner created
          ```
          ### Administrator Creates StorageClass
          The StorageClass defines the name of the provisioner and holds other properties that are needed to provision a new bucket, including
          the Owner Secret and Namespace, and the AWS Region.
          #### Greenfield Example:
          For Greenfield, a new, dynamic bucket will be generated.
          1. Create the Kubernetes StorageClass for the Provisioner.
          ```yaml
          kind: StorageClass
          apiVersion: storage.k8s.io/v1
          metadata:
            name: s3-buckets [1]
          provisioner: aws-s3.io/bucket [2]
          parameters:
            region: us-west-1 [3]
            secretName: s3-bucket-owner [4]
            secretNamespace: s3-provisioner [5]
          reclaimPolicy: Delete [6]
          ```
          1. Name of the StorageClass, this will be referenced in the User ObjectBucketClaim.
          1. Provisioner name
          1. AWS Region that the StorageClass will serve
          1. Name of the bucket owner Secret created above
          1. Namespace where the Secret will exist
          1. reclaimPolicy (Delete or Retain) indicates if the bucket can be deleted when the OBC is deleted.
          **NOTE:** the absence of the `bucketName` Parameter key in the storage class indicates this is a new bucket and its name is based on the bucket name fields in the OBC.
          ```
           # kubectl create -f storageclass-greenfield.yaml
          storageclass.storage.k8s.io/s3-buckets created
          ```
          #### Brownfield Example:
          For brownfield, the StorageClass defines the name of the provisioner and the name of the existing bucket. It also includes other properties needed by the target
          provisioner, including: the Owner Secret and Namespace, and the AWS Region
          1. Create the Kubernetes StorageClass for the Provisioner.
          ```yaml
          kind: StorageClass
          apiVersion: storage.k8s.io/v1
          metadata:
            name: s3-existing-buckets [1]
          provisioner: aws-s3.io/bucket [2]
          parameters:
            bucketName: my-existing-bucket [3]
            region: us-west-1 [4]
            secretName: s3-bucket-owner [5]
            secretNamespace: s3-provisioner [6]
          ```
          1. Name of the StorageClass, this will be referenced in the User ObjectBucketClaim.
          1. Provisioner name
          1. Name of the existing bucket
          1. AWS Region that the StorageClass will serve
          1. Name of the bucket owner Secret created above
          1. Namespace for that bucket owner secret
          **NOTE:** the storage class's `reclaimPolicy` is ignored for existing buckets.
          ```
           # kubectl create -f storageclass-brownfield.yaml
          storageclass.storage.k8s.io/s3-buckets created
          ```
          ### User Creates ObjectBucketClaim
          An ObjectBucketClaim follows the same concept as a PVC, in that
          it is a request for Object Storage, the user doesn't need to
          concern him/herself with the underlying storage, just that
          they need access to it. The user will work with the cluster/storage
          administrator to get the proper StorageClass needed and will
          then request access via the OBC.
          #### Greenfield Request Example:
          1. Create the ObjectBucketClaim.
          ```yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucketClaim
          metadata:
            name: myobc [1]
            namespace: s3-provisioner [2]
          spec:
            generateBucketName: mybucket [3]
            bucketName: my-awesome-bucket [4]
            storageClassName: s3-buckets [5]
          ```
          1. Name of the OBC
          1. Namespace of the OBC
          1. Name prepended to a random string used to generate a bucket name. It is ignored if bucketName is defined
          1. Name of new bucket which must be unique across all AWS regions, otherwise an error occurs when creating the bucket. If present, this name overrides `generateName`
          1. StorageClass name
          **NOTE:** if both `generateBucketName` and `bucketName` are omitted, and the storage class does _not_ define a bucket name, then a new, random bucket name is generated with no prefix.
          ```
           # kubectl create -f obc-brownfield.yaml
          objectbucketclaim.objectbucket.io/myobc created
          ```
          #### Brownfield Request Example:
          1. Create the ObjectBucketClaim.
          ```yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucketClaim
          metadata:
            name: myobc [1]
            namespace: s3-provisioner [2]
          spec:
            storageClassName: s3-existing-buckets [3]
          ```
          1. Name of the OBC
          1. Namespace of the OBC
          1. StorageClass name
          **NOTE:** in the OBC here there is no reference to the bucket's name. This is defined in the storage class and is not a concern of the user creating the claim to this bucket.  An OBC does have fields for defining a bucket name for greenfield use only.
          ```
           # kubectl create -f obc-brownfield.yaml
          objectbucketclaim.objectbucket.io/myobc created
          ```
          ### Results and Recap
          Let's pause for a moment and digest what just happened.
          After creating the OBC, and assuming the S3 provisioner is running, we now have
          the following Kubernetes resources:
          .  a global ObjectBucket (OB) which contains: bucket endpoint info (including region and bucket name), a reference to the OBC, and a reference to the storage class. Unique to S3, the OB also contains the bucket Amazon Resource Name (ARN).Note: there is always a 1:1 relationship between an OBC and an OB.
          .  a ConfigMap in the same namespace as the OBC, which contains the same endpoint data found in the OB.
          .  a Secret in the same namespace as the OBC, which contains the AWS key-pairs needed to access the bucket.
          And of course, we have a *new* AWS S3 Bucket which you should be able to see via the AWS Console.
          *ObjectBucket*
          ```yaml
           # kubectl get ob obc-s3-provisioner-my-awesome-bucket -o yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucket
          metadata:
            creationTimestamp: "2019-04-03T15:42:22Z"
            generation: 1
            name: obc-s3-provisioner-my-awesome-bucket
            resourceVersion: "15057"
            selfLink: /apis/objectbucket.io/v1alpha1/objectbuckets/obc-s3-provisioner-my-awesome-bucket
            uid: 0bfe8e84-576d-4c4e-984b-f73c4460f736
          spec:
            Connection:
              additionalState:
                ARN: arn:aws:iam::<accountid>:policy/my-awesome-bucket-vSgD5 [1]
                UserName: my-awesome-bucket-vSgD5 [2]
              endpoint:
                additionalConfig: null
                bucketHost: s3-us-west-1.amazonaws.com
                bucketName: my-awesome-bucket [3]
                bucketPort: 443
                region: us-west-1
                ssl: true
                subRegion: ""
            claimRef: null [4]
            reclaimPolicy: null
            storageClassName: s3-buckets [5]
          ```
          1. The AWS Policy created for this user and bucket.
          1. The new user generated by the Provisioner to access this existing bucket.
          1. The bucket name.
          1. The reference to the OBC (not filled in yet).
          1. The reference to the StorageClass used.
          *ConfigMap*
          ```yaml
           # kubectl get cm myobc -n s3-provisioner -o yaml
          apiVersion: v1
          data:
            BUCKET_HOST: s3-us-west-1.amazonaws.com [1]
            BUCKET_NAME: my-awesome-bucket [2]
            BUCKET_PORT: "443"
            BUCKET_REGION: us-west-1
            BUCKET_SSL: "true"
            BUCKET_SUBREGION: ""
          kind: ConfigMap
          metadata:
            creationTimestamp: "2019-04-01T19:11:38Z"
            finalizers:
            - objectbucket.io/finalizer
            name: my-awesome-bucket
            namespace: s3-provisioner
            resourceVersion: "892"
            selfLink: /api/v1/namespaces/s3-provisioner/configmaps/my-awesome-bucket
            uid: 2edcc58a-aff8-4a29-814a-ffbb6439a9cd
          ```
          1. The AWS S3 host.
          1. The name of the new bucket we are gaining access to.
          *Secret*
          ```yaml
           # kubectl get secret my-awesome-bucket -n s3-provisioner -o yaml
          apiVersion: v1
          data:
            AWS_ACCESS_KEY_ID: *the_new_access_id* [1]
            AWS_SECRET_ACCESS_KEY: *the_new_access_key_value* [2]
          kind: Secret
          metadata:
            creationTimestamp: "2019-04-03T15:42:22Z"
            finalizers:
            - objectbucket.io/finalizer
            name: my-awesome-bucket
            namespace: s3-provisioner
            resourceVersion: "15058"
            selfLink: /api/v1/namespaces/s3-provisioner/secrets/screeley-provb-5
            uid: 225c71a5-9d75-4ccc-b41f-bfe91b272a13
          type: Opaque
          ```
          1. The new generated AWS Access Key ID.
          1. The new generated AWS Secret Access Key.
          What happened in AWS? The first thing we do on any OBC request is
          create a new IAM user and generate Access ID and Secret Keys.
          This allows us to also better control ACLs. We also create a policy
          in IAM which we then attach to the user and bucket. We also created a new bucket, called *my-awesome-bucket*.
          When the OBC is deleted all of its Kubernetes and AWS resources will also be deleted, which includes:
          the generated OB, Secret, ConfigMap, IAM user, and policy.
          If the _retainPolicy_ on the StorageClass for this bucket is *"Delete"*, then, in addition to the above cleanup, the physical bucket is also deleted.
          **NOTE:** The actual bucket is only deleted if the Storage Class's _reclaimPolicy_ is "Delete".
          ### User Creates Pod
          Now that we have our bucket and connection/access information, a pod
          can be used to access the bucket. This can be done in several different
          ways, but the key here is that the provisioner has provided the proper
          endpoints and keys to access the bucket. The user then simply references
          the keys.
          1. Create a Sample Pod to Access the Bucket.
          ```yaml
          apiVersion: v1
          kind: Pod
          metadata:
            name: photo1
            labels:
              name: photo1
          spec:
            containers:
            - name: photo1
              image: docker.io/screeley44/photo-gallery:latest
              imagePullPolicy: Always
              envFrom:
              - configMapRef:
                  name: my-awesome-bucket <1>
              - secretRef:
                  name: my-awesome-bucket <2>
              ports:
              - containerPort: 3000
                protocol: TCP
          ```
          1. Name of the generated configmap from the provisioning process
          1. Name of the generated secret from the provisioning process
          *[Note]* Generated ConfigMap and Secret are same name as the OBC!
          Lastly, expose the pod as a service so you can access the url from a browser. In this example,
          I exposed as a LoadBalancer
          ```
            # kubectl expose pod photo1 --type=LoadBalancer --name=photo1 -n your-namespace
          ```
          To access via a url use the EXTERNAL-IP
          ```
            # kubectl get svc photo1
            NAME                         TYPE           CLUSTER-IP       EXTERNAL-IP                                                               PORT(S)          AGE
            photo1                       LoadBalancer   100.66.124.105   a00c53ccb3c5411e9b6550a7c0e50a2a-2010797808.us-east-1.elb.amazonaws.com   3000:32344/TCP   6d
          ```
          **NOTE:** This is just one example of a Pod that can utilize the bucket information,
          there are several ways that these pod applications can be developed and therefore
          the method of getting the actual values needed from the Secrets and ConfigMaps
          will vary greatly, but the idea remains the same, that the pod consumes the generated
          ConfigMap and Secret created by the provisioner.
        displayName: AWS S3 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - awss3provisioner
        - aws
        - S3
        - provisioner
        - buckets
        links:
        - name: AWS S3 Provisioner
          url: https://github.com/yard-turkey/aws-s3-provisioner
        - name: Provisioner Examples
          url: https://github.com/yard-turkey/aws-s3-provisioner/tree/master/examples
        - name: Operator Documentation
          url: https://github.com/yard-turkey/awss3operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - quay.io/screeley44/aws-s3-provisioner:v1.0.0
        version: 1.0.1
      entries:
      - name: awss3operator.1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: awss3-operator-registry
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Managed OpenShift Black Belt
      provider-url: https://github.com/rh-mobb/ocm-operator
    name: ocm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocm-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "GitLabIdentityProvider",
                "metadata": {
                  "name": "gitlab-sample"
                },
                "spec": {
                  "clientID": "gitlab",
                  "clientSecret": {
                    "name": "gitlab"
                  },
                  "clusterName": "my-cluster",
                  "displayName": "gitlab-sample",
                  "mappingMethod": "claim",
                  "url": "https://gitlab.com"
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "LDAPIdentityProvider",
                "metadata": {
                  "name": "ldap-sample"
                },
                "spec": {
                  "bindDN": "CN=test,OU=Users,DC=example,DC=com",
                  "bindPassword": {
                    "name": "test"
                  },
                  "clusterName": "my-cluster",
                  "displayName": "ldap-sample",
                  "mappingMethod": "claim",
                  "url": "ldap://test.example.com:389"
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "MachinePool",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "aws": {
                    "spotInstances": {
                      "enabled": false
                    }
                  },
                  "clusterName": "my-cluster",
                  "instanceType": "m5.xlarge",
                  "labels": {
                    "owner": "my-user"
                  },
                  "maximumNodesPerZone": 1,
                  "minimumNodesPerZone": 1,
                  "wait": false
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "ROSACluster",
                "metadata": {
                  "name": "rosa-sample"
                },
                "spec": {
                  "accountID": "111111111111",
                  "defaultMachinePool": {
                    "instanceType": "m5.xlarge",
                    "labels": {
                      "owner": "my-owner"
                    },
                    "minimumNodesPerZone": 2
                  },
                  "displayName": "my-cluster",
                  "iam": {
                    "userRole": "arn:aws:iam::111111111111:role/ManagedOpenShift-User-my_user-Role"
                  },
                  "tags": {
                    "owner": "my-owner"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ocm-operator:v0.1.0
          createdAt: "2023-06-02T17:32:25Z"
          operatorframework.io/suggested-namespace: ocm-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ocm-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLabIdentityProvider is the Schema for the gitlabidentityproviders
              API.
            displayName: Git Lab Identity Provider
            kind: GitLabIdentityProvider
            name: gitlabidentityproviders.ocm.mobb.redhat.com
            version: v1alpha1
          - description: LDAPIdentityProvider is the Schema for the ldapidentityproviders
              API.
            displayName: LDAPIdentity Provider
            kind: LDAPIdentityProvider
            name: ldapidentityproviders.ocm.mobb.redhat.com
            version: v1alpha1
          - description: MachinePool is the Schema for the machinepools API.
            displayName: Machine Pool
            kind: MachinePool
            name: machinepools.ocm.mobb.redhat.com
            version: v1alpha1
          - description: ROSACluster is the Schema for the clusters API.
            displayName: ROSACluster
            kind: ROSACluster
            name: rosaclusters.ocm.mobb.redhat.com
            version: v1alpha1
        description: "# Quick Start\n\n## Prerequisites\n\n- [oc](https://docs.openshift.com/container-platform/4.8/cli_reference/openshift_cli/getting-started-cli.html)\n-
          [aws](https://docs.aws.amazon.com/cli/latest/userguide/getting-started-install.html)\n-
          [ROSA Cluster (4.12+)](https://mobb.ninja/docs/rosa/sts/)\n\n> **Warning:**\n>
          This is only tested on ROSA, but may work on other Kubernetes and OpenShift
          variants.  It requires\n> an AWS STS-enabled cluster.  Additionally, OpenShift
          versions 4.12+ are recommended, as they \n> enable [CRD CEL validation](https://kubernetes.io/blog/2022/09/23/crd-validation-rules-beta/).
          \ Versions\n> prior to 4.12 provide a lower version of Kubernetes that does
          not enable this feature.  They may \n> work, but provide no input validation
          when submitting custom resources to the cluster.\n\nBefore installing this
          operator, there are a couple secrets that must be created.\n\n1. OCM Token:
          used to authenticate against the OCM API.  The controller will not start
          without\nthis token as it used for all calls into OCM.\n2. AWS Credentials:
          uses an assume role to a role created in the instructions below in order\nto
          access AWS resources.  This is needed for cluster creation.  See \n[Create
          AWS IAM Policies and Roles](#create-aws-iam-policies-and-roles) for more
          details.\n\n> **Note:**\n> Certain custom resources have their own specific
          prerequisites.  Please use `oc explain` \n> or read the docs [here](https://github.com/rh-mobb/ocm-operator/tree/main/docs)
          for more details.\n\n### Create OCM Token Secret\n\n1. Create a namespace
          where you wish to install the operator:\n\n```bash\noc new-project ocm-operator\n```\n\n2.
          Create a secret containing the OCM_TOKEN.  This token can be obtained form
          \nhttps://console.redhat.com/openshift/token and is used by the operator
          to authenticate \nagainst the OCM API.  This token must exist in the same
          namespace that the operator \nis running and be named `ocm-token`.  It also
          expects the key to be called `OCM_TOKEN` \nas the operator is expecting
          this value as an environment variable.\n\n```bash\noc create secret generic
          ocm-token \\\n  --namespace=ocm-operator \\\n  --from-literal=OCM_TOKEN=${MY_OCM_TOKEN}\n```\n\n###
          Create AWS IAM Policies and Roles\n\nThe operator will need to elevate privileges
          in order to perform things like \ncreating the operator-roles for the clusters.
          \ Because of this, the operator \nmust have a specific role created to allow
          it these permissions.\n\n**NOTE:** please understand what you are doing
          if you deviate from the known good \npolicies.  If errors or more stringent
          security lockdowns are found, please submit a PR \nso that we can get this
          fixed.\n\n1. Set `ACCOUNT_ID` environment variable to define your AWS account
          ID:\n\n```bash\nexport ACCOUNT_ID=111111111111\n```\n\n2. Create the permissions
          boundary.  Because the operator needs to create policies and \nroles, the
          boundary ensures that the operator is not allowed to create additional\npermissions
          outside of the defined boundary.  The sample permission set is located \nat
          `test/aws/boundary.json` in this repository:\n\n```bash\nboundary=$(curl
          -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/boundary.json\")\naws
          iam create-policy \\\n  --policy-name \"OCMOperatorBoundary\" \\\n  --policy-document
          \"$boundary\"\n```\n\n3. Create the policy.  This policy sets what the operator
          is allowed to do.  For any \n`iam` permission, the boundary created above
          is used.  The sample permission set is \nlocated at `test/aws/iam_policy.json`
          in this repository:\n\n```bash\npolicy=$(curl -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/iam_policy.json\")\naws
          iam create-policy \\\n  --policy-name \"OCMOperator\" \\\n  --policy-document
          \"$policy\"\n```\n\n4. Create the role using a trust policy and attach the
          previously created role.  The trust \npolicy is located at `test/aws/trust_policy.json`
          in this repository.  Please note that \nthe trust policy requires an OIDC
          configuration.  The OIDC configuration refers to \n**where the operator
          is running, NOT what it is provisioning**:\n\n```bash\ntrust_policy=$(curl
          -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/trust_policy.json\")\naws
          iam create-role \\\n    --role-name OCMOperator \\\n    --assume-role-policy-document
          \"$trust_policy\"\naws iam attach-role-policy \\\n    --role-name OCMOperator
          \\\n    --policy-arn arn:aws:iam::$ACCOUNT_ID:policy/OCMOperator\n```\n\n5.
          Finally, create the secret containing the assume role credentials.  The
          previous steps allow \nthe operator to assume the role you created in the
          previous step with the permissions created \nvia the previous policies.\n\n```bash\ncat
          <<EOF > /tmp/credentials\n[default]\nrole_arn = arn:aws:iam::$ACCOUNT_ID:role/OCMOperator\nweb_identity_token_file
          = /var/run/secrets/openshift/serviceaccount/token\nEOF\n\noc create secret
          generic aws-credentials \\\n  --namespace=ocm-operator \\\n  --from-file=credentials=/tmp/credentials\n```\n\n##
          Install the Operator\n\n```bash\ncat <<EOF | oc apply -f -\napiVersion:
          operators.coreos.com/v1alpha1\nkind: Subscription\nmetadata:\n  name: ocm-operator\n
          \ namespace: ocm-operator\nspec:\n  channel: v0.1\n  installPlanApproval:
          Automatic\n  name: ocm-operator\n  source: community-operators\n  sourceNamespace:
          ocm-operator\n  startingCSV: ocm-operator.v0.1.0\nEOF\n```\n\n## Provision
          Resources\n\nOnce the operator is available and running, you can provision
          any of the \nresources that it manages.  Also note that documentation is
          always available \nonce the operator is installed as well by using the `oc
          explain` command.  For \nexample, `oc explain rosacluster.spec.clusterName`
          will give you detailed documentation \nabout what the field does.\n\nSee
          the following documentation for details:\n\n* [ROSA Clusters](https://github.com/rh-mobb/ocm-operator/blob/main/docs/clusters.md)\n*
          [Machine Pools](https://github.com/rh-mobb/ocm-operator/blob/main/docs/machinepools.md)\n*
          [Identity Providers](https://github.com/rh-mobb/ocm-operator/blob/main/docs/identityproviders.md)\n"
        displayName: OpenShift Cluster Manager Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ocm
        - rosa
        - redhat
        - machinepool
        - mobb
        links:
        - name: OCM Machine Pool Operator
          url: https://github.com/rh-mobb/ocm-operator
        maintainers:
        - email: dscott@redhat.com
          name: Dustin Scott
        maturity: alpha
        provider:
          name: Managed OpenShift Black Belt
          url: https://github.com/rh-mobb/ocm-operator
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/mobb/ocm-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: ocm-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ocm-operator
    provider:
      name: Managed OpenShift Black Belt
      url: https://github.com/rh-mobb/ocm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SAP LinuxLab - Red Hat
      provider-url: https://github.com/sap-linuxlab
    name: sap-hana-express-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sap-hana-express-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "db.sap-redhat.io/v1alpha1",
                "kind": "HanaExpress",
                "metadata": {
                  "name": "hanaexpress-sample"
                },
                "spec": {
                  "credential": {
                    "key": "hxepasswd.json",
                    "name": "hxepasswd"
                  },
                  "pvcSize": "1Gi"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/redhat-sap-cop/sap-hanaexpress-operator:0.1.4
          createdAt: "2023-07-16T20:53:02Z"
          description: Manage the lifecycle of SAP Hana Express
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HanaExpress is the Schema for the hanaexpresses API
            displayName: Hana Express
            kind: HanaExpress
            name: hanaexpresses.db.sap-redhat.io
            version: v1alpha1
        description: Operator for managing the lifecycle of SAP Hana Express on Kubernetes
          platform
        displayName: SAP Hana Express Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Database
        - SAP
        links:
        - name: Sap Hana Express Operator
          url: https://github.com/redhat-sap/sap-hana-express-operator
        maintainers:
        - email: mjiao@redhat.com
          name: RedHat@SAPLinuxLab
        maturity: alpha
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/redhat-sap-cop/sap-hanaexpress-operator:0.1.4
        version: 1.0.2
      entries:
      - name: sap-hana-express-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: sap-hana-express-operator
    provider:
      name: SAP LinuxLab - Red Hat
      url: https://github.com/sap-linuxlab
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-registry-3
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-3.v3.0.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "simple"
                },
                "spec": {
                  "app": {
                    "ingress": {
                      "host": "simple-app.apps.cluster.example"
                    }
                  },
                  "ui": {
                    "ingress": {
                      "host": "simple-ui.apps.cluster.example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-3-operator:3.0.15
          createdAt: "2025-09-22T13:04:40Z"
          description: Deploy and manage Apicurio Registry 3 on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Apicurio Registry
              version 3.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Apicurio Registry 3

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Allows you manage event schemas and API designs using the Apicurio Registry web console, REST API, Maven plug-in, or Java client
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, MySQL, SQL Server, in-memory
          - Compatible with Confluent APIs
          - Runs on a lightweight Quarkus platform

          ## Apicurio Registry 3 Operator

          Provides a quick and easy way to deploy and manage Apicurio Registry 3 on Kubernetes.

          **Features**
          - Supports installation and configuration of Apicurio Registry 3
          - Perform a rolling upgrade of Apicurio Registry 3

          ## Prerequisites

          This operator does not deploy storage for Apicurio Registry 3. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio Registry 3 and Apicurio Registry 3 Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Apicurio Registry 3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-3-operator:3.0.15
        version: 3.0.15
      entries:
      - name: apicurio-registry-3.v3.0.15
        version: 3.0.15
      - name: apicurio-registry-3.v3.0.14
        version: 3.0.14
      - name: apicurio-registry-3.v3.0.12
        version: 3.0.12
      - name: apicurio-registry-3.v3.0.9
        version: 3.0.9
      - name: apicurio-registry-3.v3.0.8
        version: 3.0.8
      - name: apicurio-registry-3.v3.0.7
        version: 3.0.7
      name: 3.x
    defaultChannel: 3.x
    packageName: apicurio-registry-3
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM Edge
      provider-url: ""
    name: infrastructure-asset-orchestrator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infrastructure-asset-orchestrator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infrastructure.asset.orchestrator.com/v1",
                "kind": "Generatekey",
                "metadata": {
                  "name": "modelbuilder-token"
                },
                "spec": {}
              },
              {
                "apiVersion": "infrastructure.asset.orchestrator.com/v1",
                "kind": "Mb-broker-service",
                "metadata": {
                  "name": "mb-broker-service-sample"
                },
                "spec": {
                  "env_type": "prod",
                  "cloud_type": "IBM",
                  "external_db": {
                    "enabled": false
                  },
                  "postgres": {
                    "enable_pg_tls": true,
                    "storage_class": "",
                    "storage_size": "20G",
                    "backup_type": "incremental",
                    "backup_frequency": "@daily"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/iao-operator@sha256:5cc395f22d86641e175339deb134af6c74de92b81470a65a5c736c8005d645fd
          createdAt: "2021-08-19T12:00:00Z"
          olm.skipRange: <1.0.2
          operators.operatorframework.io/builder: operator-sdk-v1.12.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/infrastructure-asset-orchestrator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Generate key CR
            displayName: GenerateKey
            kind: Generatekey
            name: generatekeys.infrastructure.asset.orchestrator.com
            version: v1
          - description: Deploy MB broker deployment
            displayName: MBBrokerService
            kind: Mb-broker-service
            name: mb-broker-services.infrastructure.asset.orchestrator.com
            version: v1
        description: Infrastructure Asset Orchestrator provides users access to cloud
          resources and services. Leverages the dynamic nature of the cloud platforms
          to provide significant cost savings and efficiencies. Using the Infrastructure
          Asset Orchestrator user can provision, configure, utilize and deprovision
          variety of services on the IBM Cloud Catalog.
        displayName: Infrastructure Asset Orchestrator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cloud-service
        - mbbroker-service
        links:
        - name: Infrastructure Asset Orchestrator
          url: https://infrastructure-asset-orchestrator.domain
        maintainers:
        - email: edge@us.ibm.com
          name: IBM-Edge
        maturity: alpha
        provider:
          name: IBM Edge
        relatedImages:
        - registry.connect.redhat.com/ibm/iao-operator@sha256:5cc395f22d86641e175339deb134af6c74de92b81470a65a5c736c8005d645fd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:609c19d1718947ee814b829334c37e1d720acf5fb642b0712b8ab1b14d57f53f
        - registry.connect.redhat.com/ibm/iao-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/iao-orchestrator-service@sha256:145928929fd0c2b7d9ed5fb374d7cc0e6b48269847e810c5472580a1538ad44c
        - registry.connect.redhat.com/ibm/iao-mb-broker-service@sha256:6531b4196c51246958027453c2d9c7dbcdb4a1f55c826d1da238d256cdbc8a10
        version: 1.0.2
      entries:
      - name: infrastructure-asset-orchestrator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: infrastructure-asset-orchestrator-certified
    provider:
      name: IBM Edge
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: node-maintenance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-maintenance-operator.v5.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:58846ab62346b1004e11f36bd901c96f8d315958eb39fbb272d8b328faebcf04
          createdAt: "2023-10-20 15:36:51"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=4.10.0 <5.0.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-node-maintenance-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:58846ab62346b1004e11f36bd901c96f8d315958eb39fbb272d8b328faebcf04
        version: 5.0.1
      entries:
      - name: node-maintenance-operator.v5.0.1
        version: 5.0.1
      name: 4.12-eus
    - currentCSV: node-maintenance-operator.v5.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:2efd1976d2fb110391e0e5937b90476d3eeaacb4dd29cd52375f2b8566f17b41
          createdAt: "2023-10-20 15:50:49"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=5.0.0 <5.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:2efd1976d2fb110391e0e5937b90476d3eeaacb4dd29cd52375f2b8566f17b41
        version: 5.2.1
      entries:
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.14-eus
    - currentCSV: node-maintenance-operator.v5.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:7ca5853885534198544d23067087d80047b0e1421903085a2cb328bd00def147
          createdAt: "2024-08-10 19:55:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.2.0 <5.3.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:7ca5853885534198544d23067087d80047b0e1421903085a2cb328bd00def147
        version: 5.3.1
      entries:
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.16-eus
    - currentCSV: node-maintenance-operator.v5.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-10 10:59:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.3.0 <5.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.4/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel9-operator@sha256:752bff8bebcaf3df1f9d126dc4b07b622a44b02bfcf0f82b9100a7693627079a
        version: 5.4.1
      entries:
      - name: node-maintenance-operator.v5.4.1
        version: 5.4.1
      - name: node-maintenance-operator.v5.4.0
        version: 5.4.0
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.18-eus
    - currentCSV: node-maintenance-operator.v5.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-10 10:59:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.3.0 <5.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.4/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel9-operator@sha256:752bff8bebcaf3df1f9d126dc4b07b622a44b02bfcf0f82b9100a7693627079a
        version: 5.4.1
      entries:
      - name: node-maintenance-operator.v5.4.1
        version: 5.4.1
      - name: node-maintenance-operator.v5.4.0
        version: 5.4.0
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: stable
    defaultChannel: stable
    packageName: node-maintenance-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sagemaker-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sagemaker-controller.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "App",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "DataQualityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Domain",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Endpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "EndpointConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "FeatureGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "HyperParameterTuningJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Model",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelBiasJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelExplainabilityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelPackage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelPackageGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelQualityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "MonitoringSchedule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "NotebookInstance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "NotebookInstanceLifecycleConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ProcessingJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "TrainingJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "TransformJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "UserProfile",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sagemaker-controller:1.4.1
          createdAt: "2025-09-22T19:07:36Z"
          description: ACK service controller for Amazon SageMaker manages SageMaker
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: App represents the state of an AWS sagemaker App resource.
            displayName: App
            kind: App
            name: apps.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: DataQualityJobDefinition represents the state of an AWS sagemaker
              DataQualityJobDefinition resource.
            displayName: DataQualityJobDefinition
            kind: DataQualityJobDefinition
            name: dataqualityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Domain represents the state of an AWS sagemaker Domain resource.
            displayName: Domain
            kind: Domain
            name: domains.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: EndpointConfig represents the state of an AWS sagemaker EndpointConfig
              resource.
            displayName: EndpointConfig
            kind: EndpointConfig
            name: endpointconfigs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Endpoint represents the state of an AWS sagemaker Endpoint
              resource.
            displayName: Endpoint
            kind: Endpoint
            name: endpoints.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: FeatureGroup represents the state of an AWS sagemaker FeatureGroup
              resource.
            displayName: FeatureGroup
            kind: FeatureGroup
            name: featuregroups.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: HyperParameterTuningJob represents the state of an AWS sagemaker
              HyperParameterTuningJob resource.
            displayName: HyperParameterTuningJob
            kind: HyperParameterTuningJob
            name: hyperparametertuningjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: InferenceComponent represents the state of an AWS sagemaker
              InferenceComponent resource.
            displayName: InferenceComponent
            kind: InferenceComponent
            name: inferencecomponents.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: LabelingJob represents the state of an AWS sagemaker LabelingJob
              resource.
            displayName: LabelingJob
            kind: LabelingJob
            name: labelingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelBiasJobDefinition represents the state of an AWS sagemaker
              ModelBiasJobDefinition resource.
            displayName: ModelBiasJobDefinition
            kind: ModelBiasJobDefinition
            name: modelbiasjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelExplainabilityJobDefinition represents the state of
              an AWS sagemaker ModelExplainabilityJobDefinition resource.
            displayName: ModelExplainabilityJobDefinition
            kind: ModelExplainabilityJobDefinition
            name: modelexplainabilityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelPackageGroup represents the state of an AWS sagemaker
              ModelPackageGroup resource.
            displayName: ModelPackageGroup
            kind: ModelPackageGroup
            name: modelpackagegroups.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelPackage represents the state of an AWS sagemaker ModelPackage
              resource.
            displayName: ModelPackage
            kind: ModelPackage
            name: modelpackages.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelQualityJobDefinition represents the state of an AWS
              sagemaker ModelQualityJobDefinition resource.
            displayName: ModelQualityJobDefinition
            kind: ModelQualityJobDefinition
            name: modelqualityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Model represents the state of an AWS sagemaker Model resource.
            displayName: Model
            kind: Model
            name: models.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: MonitoringSchedule represents the state of an AWS sagemaker
              MonitoringSchedule resource.
            displayName: MonitoringSchedule
            kind: MonitoringSchedule
            name: monitoringschedules.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: NotebookInstanceLifecycleConfig represents the state of an
              AWS sagemaker NotebookInstanceLifecycleConfig resource.
            displayName: NotebookInstanceLifecycleConfig
            kind: NotebookInstanceLifecycleConfig
            name: notebookinstancelifecycleconfigs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: NotebookInstance represents the state of an AWS sagemaker
              NotebookInstance resource.
            displayName: NotebookInstance
            kind: NotebookInstance
            name: notebookinstances.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: PipelineExecution represents the state of an AWS sagemaker
              PipelineExecution resource.
            displayName: PipelineExecution
            kind: PipelineExecution
            name: pipelineexecutions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Pipeline represents the state of an AWS sagemaker Pipeline
              resource.
            displayName: Pipeline
            kind: Pipeline
            name: pipelines.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ProcessingJob represents the state of an AWS sagemaker ProcessingJob
              resource.
            displayName: ProcessingJob
            kind: ProcessingJob
            name: processingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: TrainingJob represents the state of an AWS sagemaker TrainingJob
              resource.
            displayName: TrainingJob
            kind: TrainingJob
            name: trainingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: TransformJob represents the state of an AWS sagemaker TransformJob
              resource.
            displayName: TransformJob
            kind: TransformJob
            name: transformjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: UserProfile represents the state of an AWS sagemaker UserProfile
              resource.
            displayName: UserProfile
            kind: UserProfile
            name: userprofiles.sagemaker.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon SageMaker resources in AWS from within your Kubernetes cluster.

          **About Amazon SageMaker**

          Amazon SageMaker helps data scientists and developers to prepare, build, train, and deploy high-quality machine learning (ML) models quickly by bringing together a broad set of capabilities purpose-built for ML.

          For more information on Amazon SageMaker, visit the [product page](https://aws.amazon.com/sagemaker/).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SageMaker
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sagemaker
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SageMaker Developer Resources
          url: https://aws.amazon.com/sagemaker/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sagemaker maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sagemaker-controller:1.4.1
        version: 1.4.1
      entries:
      - name: ack-sagemaker-controller.v1.4.1
        version: 1.4.1
      - name: ack-sagemaker-controller.v1.4.0
        version: 1.4.0
      - name: ack-sagemaker-controller.v1.3.3
        version: 1.3.3
      - name: ack-sagemaker-controller.v1.3.2
        version: 1.3.2
      - name: ack-sagemaker-controller.v1.3.1
        version: 1.3.1
      - name: ack-sagemaker-controller.v1.3.0
        version: 1.3.0
      - name: ack-sagemaker-controller.v1.2.20
        version: 1.2.20
      - name: ack-sagemaker-controller.v1.2.19
        version: 1.2.19
      - name: ack-sagemaker-controller.v1.2.18
        version: 1.2.18
      - name: ack-sagemaker-controller.v1.2.17
        version: 1.2.17
      - name: ack-sagemaker-controller.v1.2.16
        version: 1.2.16
      - name: ack-sagemaker-controller.v1.2.13
        version: 1.2.13
      - name: ack-sagemaker-controller.v1.2.12
        version: 1.2.12
      - name: ack-sagemaker-controller.v1.2.11
        version: 1.2.11
      - name: ack-sagemaker-controller.v1.2.9
        version: 1.2.9
      - name: ack-sagemaker-controller.v1.2.8
        version: 1.2.8
      - name: ack-sagemaker-controller.v1.2.5
        version: 1.2.5
      - name: ack-sagemaker-controller.v1.2.4
        version: 1.2.4
      - name: ack-sagemaker-controller.v1.2.3
        version: 1.2.3
      - name: ack-sagemaker-controller.v1.2.2
        version: 1.2.2
      - name: ack-sagemaker-controller.v1.2.1
        version: 1.2.1
      - name: ack-sagemaker-controller.v1.2.0
        version: 1.2.0
      - name: ack-sagemaker-controller.v1.1.0
        version: 1.1.0
      - name: ack-sagemaker-controller.v1.0.0
        version: 1.0.0
      - name: ack-sagemaker-controller.v0.6.0
        version: 0.6.0
      - name: ack-sagemaker-controller.v0.5.1
        version: 0.5.1
      - name: ack-sagemaker-controller.v0.5.0
        version: 0.5.0
      - name: ack-sagemaker-controller.v0.4.5
        version: 0.4.5
      - name: ack-sagemaker-controller.v0.4.4
        version: 0.4.4
      - name: ack-sagemaker-controller.v0.4.3
        version: 0.4.3
      - name: ack-sagemaker-controller.v0.4.2
        version: 0.4.2
      - name: ack-sagemaker-controller.v0.4.0
        version: 0.4.0
      - name: ack-sagemaker-controller.v0.3.4
        version: 0.3.4
      - name: ack-sagemaker-controller.v0.3.3
        version: 0.3.3
      - name: ack-sagemaker-controller.v0.3.2
        version: 0.3.2
      - name: ack-sagemaker-controller.v0.3.1
        version: 0.3.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-sagemaker-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: networkservicemesh.io
      provider-url: networkservicemesh.io
    name: nsm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nsm-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nsm.networkservicemesh.io/v1alpha1",
                "kind": "NSM",
                "metadata": {
                  "name": "nsm-sample",
                  "namespace": "nsm"
                },
                "spec": {
                  "forwarders": [
                    {
                      "image": "ghcr.io/networkservicemesh/cmd-forwarder-vpp:v1.8.0",
                      "type": "vpp"
                    }
                  ],
                  "nsmLogLevel": "TRACE",
                  "nsmPullPolicy": "IfNotPresent",
                  "registry": {
                    "image": "ghcr.io/networkservicemesh/cmd-registry-memory:v1.8.0",
                    "type": "memory"
                  },
                  "version": "v1.8.0",
                  "webhook": {
                    "image": "ghcr.io/networkservicemesh/admission-webhook-k8s:v1.8.0"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: quay.io/acmenezes/nsm-operator:v1.8.0
          createdAt: "2023-03-02T19:36:46Z"
          operatorframework.io/suggested-namespace: nsm
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/networkservicemesh/nsm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkServiceEndpoint
            name: networkserviceendpoints.networkservicemesh.io
            version: v1
          - kind: NetworkService
            name: networkservices.networkservicemesh.io
            version: v1
          - kind: NSM
            name: nsms.nsm.networkservicemesh.io
            version: v1alpha1
        description: |-
          The Network Service Mesh Operator manages and install network service meshes.

          Check networkservicemesh.io

          Install requirements:

          Before installing the operator it's necessary to have installed [spire](https://github.com/spiffe/spire)

          We have a ready to install yaml file for spire [here](https://github.com/networkservicemesh/nsm-operator/blob/master/config/spire/spire.yaml)
        displayName: nsm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network
        - service mesh
        links:
        - name: Nsm Operator
          url: https://nsm-operator.domain
        maintainers:
        - email: adcmenezes@gmail.com
          name: Alexandre Menezes
        maturity: alpha
        provider:
          name: networkservicemesh.io
          url: networkservicemesh.io
        relatedImages:
        - quay.io/acmenezes/nsm-operator:v1.8.0
        version: 1.8.0
      entries:
      - name: nsm-operator.v1.8.0
        version: 1.8.0
      name: alpha
    defaultChannel: alpha
    packageName: nsm-operator
    provider:
      name: networkservicemesh.io
      url: networkservicemesh.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: observability-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      name: candidate
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      - name: observability-operator.v0.0.22
        version: 0.0.22
      name: development
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      name: stable
    defaultChannel: stable
    packageName: observability-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-atlas-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-atlas-kubernetes.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "id": "1",
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "WEEKLY",
                      "id": "2",
                      "retentionUnit": "DAYS",
                      "retentionValue": 6
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": true,
                  "policies": [
                    {
                      "name": "atlas-default-backuppolicy",
                      "namespace": "mongodb-atlas-system"
                    },
                    {
                      "name": "atlas-default-backuppolicy2",
                      "namespace": "mongodb-atlas-system"
                    }
                  ],
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "my-database-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "david"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-cluster"
                },
                "spec": {
                  "deploymentSpec": {
                    "name": "test-cluster",
                    "providerSettings": {
                      "instanceSizeName": "M10",
                      "providerName": "AWS",
                      "regionName": "US_EAST_1"
                    }
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          containerImage: registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator@sha256:3cfbd700e4771c2c7e75c2c41cf6de57c9ef71ecd286e96a21d2567e43424bc3
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasclusters API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB Atlas —
          the only multi-cloud document database service that gives you the versatility you need to build sophisticated and resilient applications that can adapt to changing customer demands and market trends.

          > Current Status: *Beta*. The Operator gives users the ability to provision
          > Atlas projects, clusters and database users using Kubernetes Specifications and bind connection information
          > into applications deployed to Kubernetes.

          ## Quick Start guide
          ### Step 1. Deploy Kubernetes operator by clicking Install button.

          ### Step 2. Create Atlas Cluster

          **1.** Create an Atlas API Key Secret
          In order to work with the Atlas Operator you need to provide [authentication information](https://docs.atlas.mongodb.com/configure-api-access)
           to allow the Atlas Operator to communicate with Atlas API. Once you have generated a Public and Private key in Atlas, you can create a Kuberentes Secret with:
          ```
          kubectl create secret generic mongodb-atlas-operator-api-key \
                   --from-literal="orgId=<the_atlas_organization_id>" \
                   --from-literal="publicApiKey=<the_atlas_api_public_key>" \
                   --from-literal="privateApiKey=<the_atlas_api_private_key>" \
                   -n openshift-operators
          ```
          (Note, that you should use the namespace where the Operator was installed - it's `openshift-operators` by default)

          **2.** Create an `AtlasProject` Custom Resource

          The `AtlasProject` CustomResource represents Atlas Projects in our Kubernetes cluster. You need to specify
          `projectIpAccessList` with the IP addresses or CIDR blocks of any hosts that will connect to the Atlas Cluster.
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasProject
          metadata:
            name: my-project
          spec:
            name: Test Atlas Operator Project
            projectIpAccessList:
              - ipAddress: "192.0.2.15"
                comment: "IP address for Application Server A"
              - ipAddress: "203.0.113.0/24"
                comment: "CIDR block for Application Server B - D"
          ```
          **3.** Create an `AtlasDeployment` Custom Resource.
          The example below is a minimal configuration to create an M10 Atlas cluster in the AWS US East region. For a full list of properties, check
          `atlasclusters.atlas.mongodb.com` [CRD specification](config/crd/bases/atlas.mongodb.com_atlasclusters.yaml)):
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasDeployment
          metadata:
            name: my-atlas-cluster
          spec:
            name: "Test-cluster"
            projectRef:
              name: my-project
            providerSettings:
              instanceSizeName: M10
              providerName: AWS
              regionName: US_EAST_1
          ```

          **4.** Create a database user password Kubernetes Secret
          The Secret must be created in the same namespace as the `AtlasDeployment` and `AtlasProject` were created.
          ```
          kubectl create secret generic the-user-password --from-literal="password=P@@sword%"
          ```

          **5.** Create an `AtlasDatabaseUser` Custom Resource

          In order to connect to an Atlas Cluster the database user needs to be created. `AtlasDatabaseUser` resource should reference
          the password Kubernetes Secret created in the previous step.
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasDatabaseUser
          metadata:
            name: my-database-user
          spec:
            roles:
              - roleName: "readWriteAnyDatabase"
                databaseName: "admin"
            projectRef:
              name: my-project
            username: theuser
            passwordSecretRef:
              name: the-user-password
          ```
          **6.** Wait for the `AtlasDatabaseUser` Custom Resource to be ready

          Wait until the AtlasDatabaseUser resource gets to "ready" status (it will wait until the cluster is created that may take around 10 minutes):
          ```
          kubectl get atlasdatabaseusers my-database-user -o=jsonpath='{.status.conditions[?(@.type=="Ready")].status}'
          True
          ```
          ### Step 3. Connect your application to the Atlas Cluster

          The Atlas Operator will create a Kubernetes Secret with the information necessary to connect to the Atlas Cluster created
          in the previous step. An application in the same Kubernetes Cluster can mount and use the Secret:

          ```
          ...
          containers:
                - name: test-app
                  env:
                   - name: "CONNECTION_STRING"
                     valueFrom:
                       secretKeyRef:
                         name: test-atlas-operator-project-test-cluster-theuser
                         key: connectionStringStandardSrv

          ```
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator@sha256:3cfbd700e4771c2c7e75c2c41cf6de57c9ef71ecd286e96a21d2567e43424bc3
        version: 1.0.0
      entries:
      - name: mongodb-atlas-kubernetes.v1.0.0
        version: 1.0.0
      name: beta
    - currentCSV: mongodb-atlas-kubernetes.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupCompliancePolicy",
                "metadata": {
                  "name": "my-backup-compliance-policy"
                },
                "spec": {
                  "authorizedEmail": "john.doe@example.com",
                  "authorizedUserFirstName": "John",
                  "authorizedUserLastName": "Doe",
                  "copyProtectionEnabled": false,
                  "encryptionAtRestEnabled": false,
                  "onDemandPolicy": {
                    "retentionUnit": "weeks",
                    "retentionValue": 3
                  },
                  "overwriteBackupPolicies": false,
                  "pointInTimeEnabled": true,
                  "restoreWindowDays": 42,
                  "scheduledPolicyItems": [
                    {
                      "frequencyInterval": 2,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "hourly",
                      "retentionUnit": "days",
                      "retentionValue": 2
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "weekly",
                      "retentionUnit": "weeks",
                      "retentionValue": 4
                    },
                    {
                      "frequencyInterval": 15,
                      "frequencyType": "monthly",
                      "retentionUnit": "months",
                      "retentionValue": 12
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": false,
                  "copySettings": [
                    {
                      "cloudProvider": "AWS",
                      "frequencies": [
                        "WEEKLY"
                      ],
                      "regionName": "US_WEST_1",
                      "shouldCopyOplogs": false
                    }
                  ],
                  "policy": {
                    "name": "atlasbackuppolicy-sample"
                  },
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2,
                  "useOrgAndGroupNamesInExportPrefix": true
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasCustomRole",
                "metadata": {
                  "name": "shard-operator-role"
                },
                "spec": {
                  "projectRef": {
                    "name": "my-project"
                  },
                  "role": {
                    "actions": [
                      {
                        "name": "getShardMap",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "shardingState",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "connPoolStats",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "getLog",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      }
                    ],
                    "inheritedRoles": [
                      {
                        "database": "admin",
                        "name": "operator-role-1"
                      }
                    ],
                    "name": "my-role"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDataFederation",
                "metadata": {
                  "name": "my-data-federation",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "name": "my-data-federation",
                  "privateEndpoints": [
                    {
                      "endpointId": "vpce-03f9eeaa764e32454",
                      "provider": "AWS",
                      "type": "DATA_LAKE"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project",
                    "namespace": "mongodb-atlas-system"
                  },
                  "storage": {
                    "databases": [
                      {
                        "collections": [
                          {
                            "dataSources": [
                              {
                                "storeName": "http-test",
                                "urls": [
                                  "https://data.cityofnewyork.us/api/views/vfnx-vebw/rows.csv"
                                ]
                              }
                            ],
                            "name": "test-collection-1"
                          }
                        ],
                        "name": "test-db-1"
                      }
                    ],
                    "stores": [
                      {
                        "name": "http-test",
                        "provider": "http"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "the-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "theuser"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-deployment"
                },
                "spec": {
                  "backupRef": {
                    "name": "atlasbackupschedule-sample"
                  },
                  "deploymentSpec": {
                    "backupEnabled": true,
                    "clusterType": "REPLICASET",
                    "mongoDBMajorVersion": "7.0",
                    "name": "test-deployment",
                    "replicationSpecs": [
                      {
                        "regionConfigs": [
                          {
                            "electableSpecs": {
                              "instanceSize": "M10",
                              "nodeCount": 3
                            },
                            "priority": 7,
                            "providerName": "AWS",
                            "regionName": "US_EAST_1"
                          }
                        ],
                        "zoneName": "Zone 1"
                      }
                    ],
                    "tags": [
                      {
                        "key": "team",
                        "value": "atlas-operator"
                      }
                    ]
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasFederatedAuth",
                "metadata": {
                  "name": "atlas-default-federated-auth",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "connectionSecretRef": {
                    "name": "my-org-secret",
                    "namespace": "mongodb-atlas-system"
                  },
                  "domainAllowList": [
                    "my-org-domain.com"
                  ],
                  "domainRestrictionEnabled": true,
                  "enabled": true,
                  "postAuthRoleGrants": [
                    "ORG_MEMBER"
                  ],
                  "roleMappings": [
                    {
                      "externalGroupName": "org-admin",
                      "roleAssignments": [
                        {
                          "role": "ORG_OWNER"
                        }
                      ]
                    },
                    {
                      "externalGroupName": "dev-team",
                      "roleAssignments": [
                        {
                          "role": "ORG_GROUP_CREATOR"
                        },
                        {
                          "projectName": "dev-project",
                          "role": "GROUP_OWNER"
                        }
                      ]
                    }
                  ],
                  "ssoDebugEnabled": false
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasIPAccessList",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasipaccesslist-sample"
                },
                "spec": {
                  "entries": [
                    {
                      "cidrBlock": "192.168.1.0/24"
                    },
                    {
                      "deleteAfterDate": "2025-01-20T23:59:59+02:00",
                      "ipAddress": "10.0.0.1"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkContainer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasnetworkcontainer-sample"
                },
                "spec": {
                  "cidrBlock": "10.11.0.0/16",
                  "connectionSecret": {
                    "name": "atlas-connection-secret"
                  },
                  "externalProjectRef": {
                    "projectId": "66e2f2b621571b7e69a89b66"
                  },
                  "provider": "AWS",
                  "region": "US_EAST_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkPeering",
                "metadata": {
                  "name": "atlasnetworkpeering-sample"
                },
                "spec": {
                  "atlasCidrBlock": "10.12.0.0/16",
                  "awsConfiguration": {
                    "accepterRegionName": "US_EAST_1",
                    "awsAccountId": "23214235145",
                    "routeTableCidrBlock": "10.11.0.0/16",
                    "vpcId": "vpcid-7123782134"
                  },
                  "containerId": "623412394512350",
                  "containerRegion": "US_EAST_1",
                  "projectRef": {
                    "name": "atlas-project",
                    "namespace": "namespace"
                  },
                  "provider": "AWS"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasPrivateEndpoint",
                "metadata": {
                  "name": "atlasprivateendpoint-sample"
                },
                "spec": {
                  "awsConfiguration": {
                    "id": "vpce-f4k34w51d"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "provider": "AWS",
                  "region": "EU_CENTRAL_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    },
                    {
                      "cidrBlock": "203.0.113.0/24",
                      "comment": "CIDR block for Application Server B - D"
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasSearchIndexConfig",
                "metadata": {
                  "name": "atlassearchindexconfig-sample"
                },
                "spec": {
                  "analyzer": "lucene.standard",
                  "analyzers": [
                    {
                      "name": "test-analyzer",
                      "tokenFilters": [
                        {
                          "asciiFolding": {
                            "originalTokens": "omit"
                          },
                          "type": "asciiFolding"
                        }
                      ],
                      "tokenizer": {
                        "type": "whiteSpace",
                        "whitespace": {
                          "maxTokenLenght": 200
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamConnection",
                "metadata": {
                  "name": "my-streamconnection-kafka"
                },
                "spec": {
                  "kafkaConfig": {
                    "authentication": {
                      "credentials": {
                        "name": "kafka-userpass"
                      },
                      "mechanism": "SCRAM-512"
                    },
                    "bootstrapServers": "kafka.server1:9001,kafka.server2:9002,kafka.server3:9003",
                    "security": {
                      "certificate": {
                        "name": "kafka-cert"
                      },
                      "protocol": "SSL"
                    }
                  },
                  "name": "kafka-config",
                  "type": "Kafka"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamInstance",
                "metadata": {
                  "name": "my-streaminstance-sample"
                },
                "spec": {
                  "clusterConfig": {
                    "provider": "AWS",
                    "region": "VIRGINIA_USA",
                    "tier": "SP30"
                  },
                  "connectionRegistry": [
                    {
                      "name": "my-streamconnection-sample"
                    }
                  ],
                  "name": "my-instance",
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasTeam",
                "metadata": {
                  "name": "atlasteam-sample"
                },
                "spec": {
                  "name": "my-team",
                  "usernames": [
                    "username@mongodb.com"
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasThirdPartyIntegration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasintegration-sample"
                },
                "spec": {
                  "datadog": {
                    "apiKeySecretRef": {
                      "name": "secret-ref"
                    },
                    "region": "string",
                    "sendCollectionLatencyMetrics": "string",
                    "sendDatabaseMetrics": "string"
                  },
                  "projectRef": {
                    "name": "my-atlas-project"
                  },
                  "type": "DATADOG"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:8ac3f6b1c0f31af5f32562ddeed37dd6586ac47e1e7e111947df3ed570c5bed9
          createdAt: "2025-09-25T14:34:34Z"
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/mongodb/mongodb-atlas-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The AtlasBackupCompliancePolicy is a configuration that enforces
              specific backup and retention requirements
            displayName: Atlas Backup Compliance Policy
            kind: AtlasBackupCompliancePolicy
            name: atlasbackupcompliancepolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasCustomRole is the Schema for the AtlasCustomRole API
            displayName: Atlas Custom Role
            kind: AtlasCustomRole
            name: atlascustomroles.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDataFederation is the Schema for the Atlas Data Federation
              API
            displayName: Atlas Data Federation
            kind: AtlasDataFederation
            name: atlasdatafederations.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasdeployments API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasFederatedAuth is the Schema for the Atlasfederatedauth
              API
            displayName: Atlas Federated Auth
            kind: AtlasFederatedAuth
            name: atlasfederatedauths.atlas.mongodb.com
            version: v1
          - description: AtlasIPAccessList is the Schema for the atlasipaccesslists
              API.
            displayName: Atlas IPAccess List
            kind: AtlasIPAccessList
            name: atlasipaccesslists.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkContainer is the Schema for the AtlasNetworkContainer
              API
            displayName: Atlas Network Container
            kind: AtlasNetworkContainer
            name: atlasnetworkcontainers.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkPeering is the Schema for the AtlasNetworkPeering
              API
            displayName: Atlas Network Peering
            kind: AtlasNetworkPeering
            name: atlasnetworkpeerings.atlas.mongodb.com
            version: v1
          - displayName: Atlas Org Settings
            kind: AtlasOrgSettings
            name: atlasorgsettings.atlas.mongodb.com
            version: v1
          - description: "The AtlasPrivateEndpoint custom resource definition (CRD)
              defines a desired [Private Endpoint](https://www.mongodb.com/docs/atlas/security-private-endpoint/#std-label-private-endpoint-overview)
              configuration for an Atlas project. It allows a private connection between
              your cloud provider and Atlas that doesn't send information through
              a public network. \n You can use private endpoints to create a unidirectional
              connection to Atlas clusters from your virtual network."
            displayName: Atlas Private Endpoint
            kind: AtlasPrivateEndpoint
            name: atlasprivateendpoints.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
          - description: Atlas Search Index Config is the Schema for the Atlas Search
              API
            displayName: Atlas Search Index Config
            kind: AtlasSearchIndexConfig
            name: atlassearchindexconfigs.atlas.mongodb.com
            version: v1
          - description: AtlasStreamConnection is the Schema for the atlasstreamconnections
              API
            displayName: Atlas Stream Connection
            kind: AtlasStreamConnection
            name: atlasstreamconnections.atlas.mongodb.com
            version: v1
          - description: AtlasStreamInstance is the Schema for the atlasstreaminstances
              API
            displayName: Atlas Stream Instance
            kind: AtlasStreamInstance
            name: atlasstreaminstances.atlas.mongodb.com
            version: v1
          - description: AtlasTeam is the Schema for the Atlas Teams API
            displayName: Atlas Team
            kind: AtlasTeam
            name: atlasteams.atlas.mongodb.com
            version: v1
          - description: AtlasThirdPartyIntegration is the Schema for the atlas 3rd
              party integrations API.
            displayName: Atlas Third Party Integration
            kind: AtlasThirdPartyIntegration
            name: atlasthirdpartyintegrations.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB
          Atlas — the only multi-cloud document database service that gives you the versatility you need to build sophisticated
          and resilient applications that can adapt to changing customer demands and market trends.

          The full documentation for the Operator can be found [here](https://docs.atlas.mongodb.com/atlas-operator/)

          ## Getting Started

          ### Supported features

          * Create and configure a project, or connect to an existing one.
          * Deploy, manage, scale, and tear down clusters.
          * Support for serverless instances.
          * Create and edit database users.
          * Manage IP Access Lists, network peering and private endpoints.
          * Configure and control Atlas’s fully managed cloud backup.
          * Configure federated authentication for your Atlas organization
          * Integrate Atlas monitoring with Prometheus.

          ... and more.

          To view the list of custom resources and each CR's schema, visit our [reference](https://www.mongodb.com/docs/atlas/operator/stable/custom-resources/)
          documentation. See the [Quickstart](https://www.mongodb.com/docs/atlas/operator/stable/ak8so-quick-start/) to get
          started with Atlas Kubernetes Operator.
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:8ac3f6b1c0f31af5f32562ddeed37dd6586ac47e1e7e111947df3ed570c5bed9
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:9eb6465e723ede43ac47eed6de22042d26c0efb276301a4424a6a3948a8f5b15
        version: 2.11.0
      entries:
      - name: mongodb-atlas-kubernetes.v2.11.0
        version: 2.11.0
      - name: mongodb-atlas-kubernetes.v2.10.0
        version: 2.10.0
      - name: mongodb-atlas-kubernetes.v2.9.1
        version: 2.9.1
      - name: mongodb-atlas-kubernetes.v2.9.0
        version: 2.9.0
      - name: mongodb-atlas-kubernetes.v2.8.2
        version: 2.8.2
      - name: mongodb-atlas-kubernetes.v2.8.1
        version: 2.8.1
      - name: mongodb-atlas-kubernetes.v2.8.0
        version: 2.8.0
      - name: mongodb-atlas-kubernetes.v2.7.1
        version: 2.7.1
      - name: mongodb-atlas-kubernetes.v2.7.0
        version: 2.7.0
      - name: mongodb-atlas-kubernetes.v2.6.1
        version: 2.6.1
      - name: mongodb-atlas-kubernetes.v2.6.0
        version: 2.6.0
      - name: mongodb-atlas-kubernetes.v2.5.0
        version: 2.5.0
      - name: mongodb-atlas-kubernetes.v2.4.1
        version: 2.4.1
      - name: mongodb-atlas-kubernetes.v2.4.0
        version: 2.4.0
      - name: mongodb-atlas-kubernetes.v2.3.1
        version: 2.3.1
      - name: mongodb-atlas-kubernetes.v2.3.0
        version: 2.3.0
      - name: mongodb-atlas-kubernetes.v2.2.2
        version: 2.2.2
      - name: mongodb-atlas-kubernetes.v2.2.1
        version: 2.2.1
      - name: mongodb-atlas-kubernetes.v2.2.0
        version: 2.2.0
      - name: mongodb-atlas-kubernetes.v2.1.0
        version: 2.1.0
      - name: mongodb-atlas-kubernetes.v2.0.1
        version: 2.0.1
      - name: mongodb-atlas-kubernetes.v2.0.0
        version: 2.0.0
      - name: mongodb-atlas-kubernetes.v1.9.1
        version: 1.9.1
      - name: mongodb-atlas-kubernetes.v1.9.0
        version: 1.9.0
      - name: mongodb-atlas-kubernetes.v1.8.2
        version: 1.8.2
      - name: mongodb-atlas-kubernetes.v1.8.1
        version: 1.8.1
      - name: mongodb-atlas-kubernetes.v1.8.0
        version: 1.8.0
      - name: mongodb-atlas-kubernetes.v1.7.3
        version: 1.7.3
      - name: mongodb-atlas-kubernetes.v1.7.2
        version: 1.7.2
      - name: mongodb-atlas-kubernetes.v1.7.1
        version: 1.7.1
      - name: mongodb-atlas-kubernetes.v1.7.0
        version: 1.7.0
      - name: mongodb-atlas-kubernetes.v1.6.1
        version: 1.6.1
      - name: mongodb-atlas-kubernetes.v1.6.0
        version: 1.6.0
      - name: mongodb-atlas-kubernetes.v1.5.1
        version: 1.5.1
      - name: mongodb-atlas-kubernetes.v1.5.0
        version: 1.5.0
      - name: mongodb-atlas-kubernetes.v1.4.1
        version: 1.4.1
      - name: mongodb-atlas-kubernetes.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-atlas-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco AppDynamics
      provider-url: https://www.appdynamics.com
    name: appdynamics-cloud-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appdynamics-cloud-operator.v24.7.0-133
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clustermon",
                "metadata": {
                  "name": "clustermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "clusterName": "test",
                  "image": "appdynamics/appdynamics-cloud-k8s-monitoring:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Containermon",
                "metadata": {
                  "name": "containermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "name": "ContainerMon"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "DbCollector",
                "metadata": {
                  "name": "dbcollector",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-db-collector:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "DbConfig",
                "metadata": {
                  "name": "dbconfig",
                  "namespace": "appdynamics"
                },
                "spec": null
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Inframon",
                "metadata": {
                  "name": "inframon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-k8s-monitoring:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "LogCollector",
                "metadata": {
                  "name": "logcollector-sample",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-log-collector-agent:latest",
                  "resources": {
                    "limits": {
                      "cpu": "400m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "150Mi"
                    }
                  }
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Servermon",
                "metadata": {
                  "name": "servermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "name": "ServerMon"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-08-06T08:33:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clustermon is the Schema for the clustermons API
            displayName: Clustermon
            kind: Clustermon
            name: clustermons.cluster.appdynamics.com
            version: v1alpha1
          - description: Containermon is the Schema for the containermons API
            displayName: Containermon
            kind: Containermon
            name: containermons.cluster.appdynamics.com
            version: v1alpha1
          - kind: DbCollector
            name: dbcollectors.cluster.appdynamics.com
            version: v1alpha1
          - kind: DbConfig
            name: dbconfigs.cluster.appdynamics.com
            version: v1alpha1
          - description: Inframon is the Schema for the inframons API
            displayName: Inframon
            kind: Inframon
            name: inframons.cluster.appdynamics.com
            version: v1alpha1
          - kind: LogCollector
            name: logcollectors.cluster.appdynamics.com
            version: v1alpha1
          - description: Servermon is the Schema for the servermons API
            displayName: Servermon
            kind: Servermon
            name: servermons.cluster.appdynamics.com
            version: v1alpha1
        description: The Appdynamics Cloud Operator is used to install Cluster Collector,
          Container Collector, Server Collector, Log Collector and Cluster Agent
        displayName: Cisco AppDynamics Cloud Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - appdynamics
        links:
        - name: Appdynamics Cloud Operator
          url: https://appdynamics-cloud-operator.domain
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: Cisco AppDynamics
          url: https://www.appdynamics.com
        relatedImages:
        - registry.connect.redhat.com/appdynamics/appdynamics-cloud-operator@sha256:2710d63425942f60af21f060a02eb3c40f5a5f0c952ea6593b3b9769c7ca1014
        version: 24.7.0-133
      entries:
      - name: appdynamics-cloud-operator.v24.7.0-133
        version: 24.7.0-133
      name: alpha
    defaultChannel: alpha
    packageName: appdynamics-cloud-operator
    provider:
      name: Cisco AppDynamics
      url: https://www.appdynamics.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: smb-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smb-csi-driver-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9-operator@sha256:1019812a3969f844a0f3fcd45091fccc7d0d816207e230a2f399556c41ebc705
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure CIFS/SMB CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.15.0-0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/csi-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          CIFS/SMB CSI Driver Operator provides Server Message Block (SMB) CSI Driver that enables you to create and mount CIFS/SMB PersistentVolumes.
        displayName: CIFS/SMB CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - smb
        - cifs
        links:
        - name: Documentation
          url: https://github.com/openshift/csi-operator
        - name: Source Repository
          url: https://github.com/openshift/csi-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: preview
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9-operator@sha256:1019812a3969f844a0f3fcd45091fccc7d0d816207e230a2f399556c41ebc705
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0627c09fe3f5a015f117fb30af1f4d17340aefa48b23994737d3901a804b63d3
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:a5704d26d0b562dba9633d1c107f34e05c750c0e3a64dbcace4dab7b964fff8d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9@sha256:9f516d06c279cf1f82e8948ce2290395eec5f6611abf5ea919e4d6548d81ec10
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:9c3c3d2b857ce289324aee56df1ba4ab5e5ea1e76cd86ac988233463915834b4
        version: 4.16.0-202509240123
      entries:
      - name: smb-csi-driver-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: smb-csi-driver-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: smb-csi-driver-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: smb-csi-driver-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: smb-csi-driver-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: smb-csi-driver-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: smb-csi-driver-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: smb-csi-driver-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: smb-csi-driver-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: smb-csi-driver-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: smb-csi-driver-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: smb-csi-driver-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: smb-csi-driver-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: smb-csi-driver-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: smb-csi-driver-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: smb-csi-driver-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: smb-csi-driver-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: smb-csi-driver-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: smb-csi-driver-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: smb-csi-driver-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: smb-csi-driver-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: smb-csi-driver-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: smb-csi-driver-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: smb-csi-driver-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: smb-csi-driver-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: smb-csi-driver-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: smb-csi-driver-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: smb-csi-driver-operator.v4.16.0-202406251906
        version: 4.16.0-202406251906
      - name: smb-csi-driver-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: smb-csi-driver-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: preview
    defaultChannel: preview
    packageName: smb-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: bamoe-kogito-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bamoe-kogito-operator.8.0.5-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "ruleunit-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogitoinfra-sample"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "kafka.strimzi.io/v1beta2",
                    "kind": "Kafka",
                    "name": "kogito-kafka"
                  }
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator@sha256:8a99fcd217502b7e850af1e1097de58b6cb19d1e786394811a8e075c4d871279
          description: IBM BAMOE Kogito Operator for deployment and management of
            Kogito services.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["IBM Business Automation Manager
            Open Edition"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: https://www.ibm.com/mysupport
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.rhpam.kiegroup.org
            version: v1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.rhpam.kiegroup.org
            version: v1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.rhpam.kiegroup.org
            version: v1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.rhpam.kiegroup.org
            version: v1
        description: |-
          IBM BAMOE Kogito Operator is designed for deployment and management of Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Builds a IBM BAMOE Kogito service from a remote git repository and creates a custom image for the given runtime: Red Hat build of Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Deploy IBM BAMOE Kogito service from user provided image.
          * Setup Grafana Dashboard when Grafana Operator is available in cluster.
          * Create Prometheus monitor object when Prometheus operator is available in cluster.

          **IMPORTANT!** IBM BAMOE Kogito Operator may need Grafana and Prometheus Operator to configure monitoring in cases if you do not provide these services beforehand. If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the IBM BAMOE Kogito Operator.
        displayName: IBM BAMOE Kogito Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://ibm.com
        - name: Documentation
          url: https://www.ibm.com/docs/en/ibamoe
        - name: Source
          url: https://github.com/kiegroup/kogito-operator
        maintainers:
        - email: mysphelp@us.ibm.com
          name: IBM
        maturity: beta
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9492c67eaa7ab37df6f5fa623c4aac955b7d7ffaab9ae2e57dea745e82907e40
        - registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator@sha256:8a99fcd217502b7e850af1e1097de58b6cb19d1e786394811a8e075c4d871279
        version: 8.0.5-3
      entries:
      - name: bamoe-kogito-operator.8.0.5-3
        version: 8.0.5-3
      - name: bamoe-kogito-operator.8.0.5-2
        version: 8.0.5-2
      - name: bamoe-kogito-operator.8.0.5-1
        version: 8.0.5-1
      - name: bamoe-kogito-operator.8.0.4-2
        version: 8.0.4-2
      - name: bamoe-kogito-operator.8.0.4-1
        version: 8.0.4-1
      - name: bamoe-kogito-operator.8.0.3-1
        version: 8.0.3-1
      - name: bamoe-kogito-operator.8.0.2-3
        version: 8.0.2-3
      - name: bamoe-kogito-operator.8.0.2-2
        version: 8.0.2-2
      - name: bamoe-kogito-operator.8.0.2-1
        version: 8.0.2-1
      - name: bamoe-kogito-operator.8.0.1-2
        version: 8.0.1-2
      - name: bamoe-kogito-operator.8.0.1-1
        version: 8.0.1-1
      - name: bamoe-kogito-operator.8.0.0-1
        version: 8.0.0-1
      name: 8.x
    defaultChannel: 8.x
    packageName: bamoe-kogito-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc
      provider-url: ""
    name: openshift-gitops-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-gitops-operator.v1.10.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:b0dc693854319ada341e2e3872b34d7db9823b470ed369edfde88d3ae73a40e1
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.8.19\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:a74e0a1fe3589ebe9f226a2bad20e9a5d5601a0e62c5f3290902f466f55b7160
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:2bec31b0c5407ffcecfd8b860d8afb1629eaab31a250f60191f94a824e0a52f4
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:c779ad87a7f71d851c2af818abf16e5151c43d3fd54e64eac2fc7117c00753f5
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:b0dc693854319ada341e2e3872b34d7db9823b470ed369edfde88d3ae73a40e1
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:b5c8a6c7438a45125fc7fd970946f715ab7193b7446d625fae80dde9d1f3f914
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:4c842f0ea8f0e9b53eb16d94488fc2f4298bd8e4be2f48713be9157da24549e6
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:ece2fee4227ad1becadca47028b78ceaf8717a98ec84e3f1b5188a135c4b093c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d35dff5d59e748ef0404e7e76b0237ea96b2bb19b86fa04a3a34df7a379277b2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5d5061dbc2ec5082cf14b6c600fb5400b83cf91d7ccebfa80680a238d275db
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:0905ba3f5cf87780a1e95b6a00ea7cd340d4ed1b5054e4b2c82406270baeaefa
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ec9f60018694dcc5d431ba47d5536b761b71cb3f66684978fe6bb74c157679ac
        version: 1.10.6
      entries:
      - name: openshift-gitops-operator.v1.10.6
        version: 1.10.6
      - name: openshift-gitops-operator.v1.10.5
        version: 1.10.5
      - name: openshift-gitops-operator.v1.10.4
        version: 1.10.4
      - name: openshift-gitops-operator.v1.10.3
        version: 1.10.3
      - name: openshift-gitops-operator.v1.10.2
        version: 1.10.2
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.10
    - currentCSV: openshift-gitops-operator.v1.11.7-0.1724840231.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d6ad2ee900bf99a3672049cc8ce9c94870d41176501ded1abbc1b229b9553401
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          olm.substitutesFor: openshift-gitops-operator.v1.11.7
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.9.21\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:4d22a7a6cc2f0442889512267ae3d9356f05ff0dfdbbe137c3d8ebebc24818c8
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:42f809f32f1d43186b849e3cc78063d3ebb2b54c55b2a1b77908af09a5ce044b
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:0b926fc9eae775425faa1ab9709a36fdac15afe4bfe523720e17ecf5df45bb28
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:0df37cf0a0e48efae02a0711c6da61c6165e7856db97097e98d56c8b6a137bb9
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:eade50b8321c1b2e0a8ac4dcb34596d43dea0a137b8b97b6eb68c6d469c8b200
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:20566173a8e907980e603d82fae4d96fcdfd5c43d9004c2e506cf0e09a47b91e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/rhel8/redis-6@sha256:c4a98010f7dd5faec2047bb9803061704233a8efc930be2a02386ac16d062d86
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c94c2167dc957856682a104b9a6187d3d321dec71ae794acc0dc37794d2cc373
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:7ea96000768b3cf7711150531c4c1c7edff2c45a2f29fe10bd9811ab6b281909
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d6ad2ee900bf99a3672049cc8ce9c94870d41176501ded1abbc1b229b9553401
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:959de6a95ef8978f66fdbe1033bcaa0ddeff9a376023e5d1d2698c5acb417305
        version: 1.11.7+0.1724840231.p
      entries:
      - name: openshift-gitops-operator.v1.11.7-0.1724840231.p
        version: 1.11.7+0.1724840231.p
      - name: openshift-gitops-operator.v1.11.7
        version: 1.11.7
      - name: openshift-gitops-operator.v1.11.6
        version: 1.11.6
      - name: openshift-gitops-operator.v1.11.5
        version: 1.11.5
      - name: openshift-gitops-operator.v1.11.4
        version: 1.11.4
      - name: openshift-gitops-operator.v1.11.3
        version: 1.11.3
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.11
    - currentCSV: openshift-gitops-operator.v1.12.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:48dd1254fd3a887aef2eae9239f68637eae18514a50f72ea504414d83dc727b8
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.10.18\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:dbf3038daf7c797123c511f9a609c3485cb1a5bcae53c614fad17e0a65fa7ee3
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a1d4e26ea755181818401697f18260b9e876aa441be14830f9142081ee22a24d
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:73918f3b3c507fc64006b7a500ecfa9d58151d694852f696c41a4412edffc78e
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:33bbb88343285a0510670384b87d0226dd6de8aa12c78b68921362c4d6f5a6aa
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:48dd1254fd3a887aef2eae9239f68637eae18514a50f72ea504414d83dc727b8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9e0fee30842a16d7b8a39d336123c80c5271cb2ceb50759fc7ee51a4ed09ceb3
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:06a6637c4776f4285bcb295d327ee60b1aed61b7adc5f7ab52898d79281ced70
        - registry.redhat.io/rhel8/redis-6@sha256:a848cc4edb4405846c673187dd68ce66c328dd1ad42f193f5737288c7971399d
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:b5736d1106db29fcdf42328ca20b96382d4339f90b97a07120ea73ee174ea1ad
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:7d82c719b1cef64607d30e8aaa5ab293955771c5b1dbd430e0befc988a43356d
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:1cf485a889c7586cb4944f000b53c0e97d08f97ba14e580bec57dad7cb13712e
        version: 1.12.6
      entries:
      - name: openshift-gitops-operator.v1.12.6
        version: 1.12.6
      - name: openshift-gitops-operator.v1.12.5-0.1727371523.p
        version: 1.12.5+0.1727371523.p
      - name: openshift-gitops-operator.v1.12.5
        version: 1.12.5
      - name: openshift-gitops-operator.v1.12.4
        version: 1.12.4
      - name: openshift-gitops-operator.v1.12.3
        version: 1.12.3
      - name: openshift-gitops-operator.v1.12.2
        version: 1.12.2
      - name: openshift-gitops-operator.v1.12.1
        version: 1.12.1
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.12
    - currentCSV: openshift-gitops-operator.v1.13.3-0.1741683398.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d41034bcf5c28bcb8afef08678793d0cc27a4032e2811a0c496c3bd0f10ecc17
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          olm.substitutesFor: openshift-gitops-operator.v1.13.3-0.1738142397.p
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.11.12\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_gitops/
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d41034bcf5c28bcb8afef08678793d0cc27a4032e2811a0c496c3bd0f10ecc17
        - registry.redhat.io/rhel8/redis-6@sha256:a848cc4edb4405846c673187dd68ce66c328dd1ad42f193f5737288c7971399d
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:b982df02ea7c45a7603d83fe82cac12e1faac303a1b72af6a727512bb97441df
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3f2804bc2af437ff1f06c214b48a1a780dfcfab995650d901c03270057eb9f3c
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:06a6637c4776f4285bcb295d327ee60b1aed61b7adc5f7ab52898d79281ced70
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a7c3e31ce414e6c6fecf1652343b22c52502902676bf4d7a75bf42bdabfca06
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:a366229e6ee26ebe2d7f765b2a2da6057a29e783835655b5d97545ac563b541a
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:d4d5724f67859d24c32f0f5735f5af3e325f77db97f58bc7f17ed4b831b7e534
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c02dc2c79ee24ce5450ea4e19493ac2a1997eeb5bfd359291642c529826d0cdd
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:e87515d016565bb69cb2c2ac40432e0127dd2cb20a490ac82019836032ba15e5
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:bde11fa83b7e708677b08ea995041f480b88edfb93e4a62c10380d101c0a349c
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        version: 1.13.3+0.1741683398.p
      entries:
      - name: openshift-gitops-operator.v1.13.3-0.1741683398.p
        version: 1.13.3+0.1741683398.p
      - name: openshift-gitops-operator.v1.13.3-0.1738142397.p
        version: 1.13.3+0.1738142397.p
      - name: openshift-gitops-operator.v1.13.3
        version: 1.13.3
      - name: openshift-gitops-operator.v1.13.2
        version: 1.13.2
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.13
    - currentCSV: openshift-gitops-operator.v1.14.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:7cf924af77d9c56e70710b99e080ed8aee162f05ec472f32085a76ef85cf9879
          createdAt: "2025-05-22 12:40:45"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.12.13-1\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_gitops/
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:2c45cfc01f9ec7577afc09cadc4c1a6f470836f6a0a7c24d66b791913f89bb05
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:42a79aad641e8607fc339e4ce7f9cf1d258be01f863dbc31a94d2c27edb645bd
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:97e32825f59c33cca7d8dba3580095205c20435ccd15b91cf0aff37db9841e0e
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:7cf924af77d9c56e70710b99e080ed8aee162f05ec472f32085a76ef85cf9879
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:b7d3c882fc25fbbea28951afb197d4d47ace253864889339396e20f0c6897a7f
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:bcf6b3136a5e71803cb518c8763c165bb46ee37ef7bf55503911d7fcb36b3a8f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c369df980874aae270ca7bc6455f1ca45467d75d25fe80d7bcb06a7826bdeaa3
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:cd4965d91c163160bcf43a02c54ca60b8704281c95de904f904df942bfd83bcd
        - registry.redhat.io/rhel8/redis-6@sha256:a9f30ed08abcab0122556a5485fbcf6e0fc03860f19954dcebdbf5e754876a45
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:7358787959af7e7e589edfc043a97d80ebab87536bde365d33be7a600071bf0e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1ed3b514bf9af6ac0d386e3739874d1d964e667d54e041c2d67edd769b0009fb
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:8096f824bd96e3148cf4b6936766c888d1fff0af9e9d0be440ae4feb6adc03f2
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:1112160243f227a5d393eeee165bd4fed0d3f42646a9280cbe9af60ef728aeb2
        version: 1.14.4
      entries:
      - name: openshift-gitops-operator.v1.14.4
        version: 1.14.4
      - name: openshift-gitops-operator.v1.14.3-0.1746016855.p
        version: 1.14.3+0.1746016855.p
      - name: openshift-gitops-operator.v1.14.3-0.1742836272.p
        version: 1.14.3+0.1742836272.p
      - name: openshift-gitops-operator.v1.14.3
        version: 1.14.3
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.14
    - currentCSV: openshift-gitops-operator.v1.15.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ad49a3d14f843bde0d2365093d00ffa6d3b9d59a75c81e0c56710ad37b53bd92
          createdAt: "2024-11-25T03:57:37Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v2.13.9\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:d41454824632475f0e88aa6e0031ec643bd37e1d575df36cd3ee63daf1403281
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:6cec76d7ea70d77923cf4285177e375d34e96e3aecbe77b68cd5ff769e6e540b
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:911733d78ae6f36f642b4d63f044cd102b7eee170508eaf518ac7acf0f339698
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ad49a3d14f843bde0d2365093d00ffa6d3b9d59a75c81e0c56710ad37b53bd92
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:41717c27915251bda8afeef651b609b8a9ed410e0fa09c4cf8fa7e91d6a94292
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:0c7f72546d0d918bfd21aff7157b983ff7c832b269f5efb84d2ed723a09bf612
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:5bda26e2c16beebf66dd8f3c3c55d48e2f00db870f2d9d4d44570d9d3b8bddb0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:a9e2360f00c5c9840a100a6748bd905b8f34169140a33fb0ca4af9cf6db71642
        version: 1.15.4
      entries:
      - name: openshift-gitops-operator.v1.15.4
        version: 1.15.4
      - name: openshift-gitops-operator.v1.15.3
        version: 1.15.3
      - name: openshift-gitops-operator.v1.15.2
        version: 1.15.2
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.15
    - currentCSV: openshift-gitops-operator.v1.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ead539d7cf54992cf1bfeae6a73a453059a7dffef44505991ff33319c8d40c50
          createdAt: "2025-03-07T13:13:07Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v2.14.16\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:58284a12174d19c2d04723fa1f551d96b490872e0fa2ad94af1b3e52dc6ad955
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a5e2a12ec5dea607a6651462d8e7920cc73f3272dba4041c397a14f2c73d7532
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:03d0d9093422396330aa8611adf77abdc9be759f3d2e23850e4abe778bd6002b
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:cab809fd1288d31c93f4f5c4facb7efaff9f844e21a859d1e4557831575a13c8
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c37aa141387c847eac825363dddda3eb59203d3694120dba439b00cf0c66f320
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:2598a3339bc920b91018d3f7acbe78ccbee684b5182c874e48296744d2373f9d
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ead539d7cf54992cf1bfeae6a73a453059a7dffef44505991ff33319c8d40c50
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:ceff1cde6d043daae599b578eec9759c6d49cd5cf770460b9d075fa879a16e81
        version: 1.16.3
      entries:
      - name: openshift-gitops-operator.v1.16.3
        version: 1.16.3
      - name: openshift-gitops-operator.v1.16.2
        version: 1.16.2
      - name: openshift-gitops-operator.v1.16.1
        version: 1.16.1
      - name: openshift-gitops-operator.v1.16.0-0.1746014725.p
        version: 1.16.0+0.1746014725.p
      - name: openshift-gitops-operator.v1.16.0
        version: 1.16.0
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.16
    - currentCSV: openshift-gitops-operator.v1.17.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:1b86eb4e071b1c526931b6d18c3d8f51522293699059c514488602c320271049
          createdAt: "2025-07-30T13:03:16Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.0.14\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:3fc4941fa4a6f2dbe9dd82622a0d3fb794ecd5a9dee286a38b60a7f6a947d685
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:cc45e212e7eb4e7e8daac6ea74071d5d98260de0b1f6d1cea0e7876111ccea8c
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:2f5f997bce924445de735ae0508dca1a7bba561bc4acdacf659928488233cb8a
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:5c9ea426cd60e7b8d1d8e4fe763909200612434c65596855334054e26cbfe3d0
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:70e5e3900d0c89999c69be1d58adf84c1b3664aacbffa9c5e0efc8ae1b6ef9c1
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:1b86eb4e071b1c526931b6d18c3d8f51522293699059c514488602c320271049
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:d73b4f5b355f556208751281923bb8b43edffc061a673c04d53b88f0e50ecb26
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:a2ae66772d90e734bb3385a59852e6016325a21b916a2580a33cf0da7cb0b527
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:c0a513ed54b631629efafe6e94b775d5bd0a943c44fbe96f02a0d5bb1f989099
        version: 1.17.1
      entries:
      - name: openshift-gitops-operator.v1.17.1
        version: 1.17.1
      - name: openshift-gitops-operator.v1.17.0
        version: 1.17.0
      name: gitops-1.17
    - currentCSV: openshift-gitops-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
          createdAt: "2025-08-21T01:20:45Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.18.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: NamespaceManagement
            name: namespacemanagements.argoproj.io
            version: v1beta1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.1.6\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/rhel9/redis-7@sha256:b4f58ad83ab8719a2e102f38ee526546cd1d56428901afc3add45c9a0e2a2f12
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:1eef9ecd58cf8391f29360ef74dd8aa5dcb48e1582d1500e445706d8302c951e
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a0544c14823492165550d83a6d8ba79dd632b46144d3fdcb543793726111d76
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:6284873bcf1593c210003e0292ac5583f03f03e80d937098d808d48fa0d01cf6
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:5a09fcba3d9e6fc462719e81a46af6022cf7f70a05bcb387ed959a49f6f37dbc
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:adca9c55b80d899d500bf0a7234cf61ae30ffed1b2c927ddea8d9fd07c9acb6a
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7f6ceae56c60fd432425c36f93a9e9bbbb96ef9422333ad6ea41ff3a2eaf8374
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:640e3298c7cd2faa84971b0d861446c4892ab8602108af05b8b5b049a4de624f
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c2283997d47298d49378e8200b149ea613c7735317ef12e74b7915a590f41ce2
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:08efd3e7a79c2945693c99ed31ec912223becff7238b1e28d1b1da5d63f93019
        version: 1.18.0
      entries:
      - name: openshift-gitops-operator.v1.18.0
        version: 1.18.0
      name: gitops-1.18
    - currentCSV: openshift-gitops-operator.v1.6.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:104cc7564fcc3261faa3fbf9c95720b4434e15103b876c080ed012b244ef1dae
          createdAt: "2021-09-15T22:31:21Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD v2.4.28\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces. \n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc.
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:104cc7564fcc3261faa3fbf9c95720b4434e15103b876c080ed012b244ef1dae
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:0cc4fa704e581cd5581c816668183f602fae6ef9ab3c04ce0d99d561287ef140
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:06f1150114d4c0bd236d2a5d4716c9c3e95a9733058297790389d6ba2a1bf264
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:e9e2f950cda8153cff112e43872b98e4977d3cbdf59713b8de92defef6d34b6d
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:dbcad68b70ab7918e944099aef4177c9bc3378e9ca49499dbb94a175927d3209
        - registry.redhat.io/rhel8/redis-6@sha256:53598a6effeb90e4f1b005b2521beffd2fa2b0c52d0e7f2347ee2abd2577cab3
        version: 1.6.6
      entries:
      - name: openshift-gitops-operator.v1.6.6
        version: 1.6.6
      - name: openshift-gitops-operator.v1.6.5
        version: 1.6.5
      - name: openshift-gitops-operator.v1.6.4
        version: 1.6.4
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.6
    - currentCSV: openshift-gitops-operator.v1.7.4-0.1690486082.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:e0967a4058fe5a9bfe7e998a3347529c99536cf3a44de9e40f9632035ff5a2fe
          createdAt: "2021-09-15T22:31:21Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.substitutesFor: openshift-gitops-operator.v1.7.4
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD v2.5.16\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficient privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces. \n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/rhel8/redis-6@sha256:53598a6effeb90e4f1b005b2521beffd2fa2b0c52d0e7f2347ee2abd2577cab3
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:e0967a4058fe5a9bfe7e998a3347529c99536cf3a44de9e40f9632035ff5a2fe
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c55aadf45397083bec4e42d6c0514ed79169a07fd3d9d75671904df2c0520e52
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:45789c49d9334e0ed062641313714c101bd0117dc5633316b1ec63e7a6c76c15
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:26050fb95978e90e8d00e895032045e77cb236008a0e1417459e1aeca8a48c69
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:5c11be8ec9db54331893f2f298aa27006e77b0eaded2d88d0531c77da91a421f
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:f1371d2012796e9aaca105112190ff89ec3d0be95968e05b19db101480f67975
        version: 1.7.4+0.1690486082.p
      entries:
      - name: openshift-gitops-operator.v1.7.4-0.1690486082.p
        version: 1.7.4+0.1690486082.p
      - name: openshift-gitops-operator.v1.7.4
        version: 1.7.4
      - name: openshift-gitops-operator.v1.7.3
        version: 1.7.3
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.7
    - currentCSV: openshift-gitops-operator.v1.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f32df77b3c23d6a833ff887386285ca7a5bd436641f0ec1ffc934c8247fa315f
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.skipRange: ""
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.6.153\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:8cc59901689858e06f5c2942f8c975449c17454fa8364da6153d9b5a3538a985
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:f77594bc053be144b33ff9603ee9675c7e82946ec0dbfb04d8f942c8d73155da
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f32df77b3c23d6a833ff887386285ca7a5bd436641f0ec1ffc934c8247fa315f
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:8aa68fddef88921aeeede24a849d3dabbceff01dc03185342c21e6ee58033eaf
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:16d9184345e6d5a19f326a7b52f76d8a59b96763f233c50546ff017a625e5ff5
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:0d13c03048b4435d9f5045a6fb7bd4b8808461b26aeca60c0047fa5937e1d45d
        version: 1.8.6
      entries:
      - name: openshift-gitops-operator.v1.8.6
        version: 1.8.6
      - name: openshift-gitops-operator.v1.8.5
        version: 1.8.5
      - name: openshift-gitops-operator.v1.8.4-0.1693214052.p
        version: 1.8.4+0.1693214052.p
      - name: openshift-gitops-operator.v1.8.4
        version: 1.8.4
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.8
    - currentCSV: openshift-gitops-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "argo-rollout"
                },
                "spec": null
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:11a6084e11aa32511a34bd6cfb4f4995aec690c1f5ed61861e99f6e42d62afb4
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.skipRange: ""
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.7.16\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:80f92842fe4cba008a873ad2ab6059a3a24e76e16671d8ece034dbbe666996bc
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:3d17786e2f4063154b9e3a40b96bf8ec0a14b86395cea5fe7ca18c5b43764478
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a568c8c647e2e2d6abdf12138eeef8f8bdde1fdcbf786e9ca8c8f93bc79acaf9
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:43bb961d847477902e66a92edd1539322cac8d5f74706b048a458ad037a0c7a3
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:11a6084e11aa32511a34bd6cfb4f4995aec690c1f5ed61861e99f6e42d62afb4
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ec9f60018694dcc5d431ba47d5536b761b71cb3f66684978fe6bb74c157679ac
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:3873d1f18b5643e1de2601419410270342650a21428bb26a4cbee20c0c63f98b
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:86012145d4b8798dc12f6f4f07a9316a193188aaac865e324b63f4e60e8ec584
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:136d3a936759e112d3ea2e78f7c02bd882b0fd3011f4c452f45c32c43822772d
        version: 1.9.4
      entries:
      - name: openshift-gitops-operator.v1.9.4
        version: 1.9.4
      - name: openshift-gitops-operator.v1.9.3
        version: 1.9.3
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.9
    - currentCSV: openshift-gitops-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
          createdAt: "2025-08-21T01:20:45Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.18.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: NamespaceManagement
            name: namespacemanagements.argoproj.io
            version: v1beta1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.1.6\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:adca9c55b80d899d500bf0a7234cf61ae30ffed1b2c927ddea8d9fd07c9acb6a
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7f6ceae56c60fd432425c36f93a9e9bbbb96ef9422333ad6ea41ff3a2eaf8374
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:6284873bcf1593c210003e0292ac5583f03f03e80d937098d808d48fa0d01cf6
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:08efd3e7a79c2945693c99ed31ec912223becff7238b1e28d1b1da5d63f93019
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:640e3298c7cd2faa84971b0d861446c4892ab8602108af05b8b5b049a4de624f
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c2283997d47298d49378e8200b149ea613c7735317ef12e74b7915a590f41ce2
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:5a09fcba3d9e6fc462719e81a46af6022cf7f70a05bcb387ed959a49f6f37dbc
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:1eef9ecd58cf8391f29360ef74dd8aa5dcb48e1582d1500e445706d8302c951e
        - registry.redhat.io/rhel9/redis-7@sha256:b4f58ad83ab8719a2e102f38ee526546cd1d56428901afc3add45c9a0e2a2f12
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a0544c14823492165550d83a6d8ba79dd632b46144d3fdcb543793726111d76
        version: 1.18.0
      entries:
      - name: openshift-gitops-operator.v1.18.0
        version: 1.18.0
      - name: openshift-gitops-operator.v1.17.1
        version: 1.17.1
      - name: openshift-gitops-operator.v1.17.0
        version: 1.17.0
      - name: openshift-gitops-operator.v1.16.2
        version: 1.16.2
      - name: openshift-gitops-operator.v1.16.1
        version: 1.16.1
      - name: openshift-gitops-operator.v1.16.0-0.1746014725.p
        version: 1.16.0+0.1746014725.p
      - name: openshift-gitops-operator.v1.16.0
        version: 1.16.0
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: latest
    defaultChannel: latest
    packageName: openshift-gitops-operator
    provider:
      name: Red Hat Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: node-maintenance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-maintenance-operator.v0.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-maintenance-operator:v0.17.0
          createdAt: "2024-03-13 18:56:47"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=0.12.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/medik8s/node-maintenance-operator:v0.17.0
        version: 0.17.0
      entries:
      - name: node-maintenance-operator.v0.17.0
        version: 0.17.0
      - name: node-maintenance-operator.v0.16.1
        version: 0.16.1
      - name: node-maintenance-operator.v0.16.0
        version: 0.16.0
      - name: node-maintenance-operator.v0.15.0
        version: 0.15.0
      - name: node-maintenance-operator.v0.14.0
        version: 0.14.0
      - name: node-maintenance-operator.v0.13.0
        version: 0.13.0
      - name: node-maintenance-operator.v0.12.1
        version: 0.12.1
      name: stable
    defaultChannel: stable
    packageName: node-maintenance-operator
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudwatch-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudwatch-controller.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudwatch-controller:1.3.0
          createdAt: "2025-10-02T23:05:17Z"
          description: AWS CloudWatch controller is a service controller for managing
            CloudWatch resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Dashboard represents the state of an AWS cloudwatch Dashboard
              resource.
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.cloudwatch.services.k8s.aws
            version: v1alpha1
          - description: MetricAlarm represents the state of an AWS cloudwatch MetricAlarm
              resource.
            displayName: MetricAlarm
            kind: MetricAlarm
            name: metricalarms.cloudwatch.services.k8s.aws
            version: v1alpha1
          - description: MetricStream represents the state of an AWS cloudwatch MetricStream
              resource.
            displayName: MetricStream
            kind: MetricStream
            name: metricstreams.cloudwatch.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudWatch resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudWatch**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudWatch
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudwatch
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudWatch Developer Resources
          url: https://aws.amazon.com/CloudWatch/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudwatch maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudwatch-controller:1.3.0
        version: 1.3.0
      entries:
      - name: ack-cloudwatch-controller.v1.3.0
        version: 1.3.0
      - name: ack-cloudwatch-controller.v1.2.1
        version: 1.2.1
      - name: ack-cloudwatch-controller.v1.2.0
        version: 1.2.0
      - name: ack-cloudwatch-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudwatch-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudwatch-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudwatch-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudwatch-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudwatch-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudwatch-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudwatch-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudwatch-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudwatch-controller.v1.0.5
        version: 1.0.5
      - name: ack-cloudwatch-controller.v1.0.4
        version: 1.0.4
      - name: ack-cloudwatch-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudwatch-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudwatch-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudwatch-controller.v0.0.12
        version: 0.0.12
      - name: ack-cloudwatch-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudwatch-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudwatch-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudwatch-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudwatch-controller.v0.0.3
        version: 0.0.3
      - name: ack-cloudwatch-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudwatch-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OpenShift Database Access
      provider-url: ""
    name: dbaas-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dbaas-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSConnection",
                "metadata": {
                  "name": "dbaasconnection-example"
                },
                "spec": {
                  "databaseServiceID": "instance-example",
                  "databaseServiceType": "instance",
                  "inventoryRef": {
                    "name": "inventory-example",
                    "namespace": "openshift-dbaas-operator"
                  }
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSInstance",
                "metadata": {
                  "name": "dbaasinstance-sample"
                },
                "spec": {
                  "cloudProvider": "AWS",
                  "cloudRegion": "US_EAST_1",
                  "inventoryRef": {
                    "name": "inventory-example",
                    "namespace": "openshift-dbaas-operator"
                  },
                  "name": "my-cluster-free"
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSInventory",
                "metadata": {
                  "name": "inventory-example"
                },
                "spec": {
                  "credentialsRef": {
                    "name": "dbaas-vendor-credentials"
                  },
                  "providerRef": {
                    "name": "provider-example-registration"
                  }
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSPolicy",
                "metadata": {
                  "name": "dbaaspolicy-sample"
                },
                "spec": {
                  "disableProvisions": false
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSProvider",
                "metadata": {
                  "labels": {
                    "related-to": "dbaas-operator",
                    "type": "dbaas-provider-registration"
                  },
                  "name": "provider-example-registration"
                },
                "spec": {
                  "allowsFreeTrial": true,
                  "connectionKind": "ProviderConnection",
                  "credentialFields": [
                    {
                      "displayName": "ProviderAPIKeyD",
                      "helpText": "This is the API Key for the example provider to show on UI",
                      "key": "APIKey",
                      "required": true,
                      "type": "maskedstring"
                    }
                  ],
                  "externalProvisionDescription": "Follow the guide to start a free Provider Serverless (beta) cluster",
                  "externalProvisionURL": "https://www.exmample.com/docs/provider/quickstart.html",
                  "groupVersion": "dbaas.redhat.com/v1beta1",
                  "instanceKind": "ProviderInstance",
                  "inventoryKind": "ProviderInventory",
                  "provider": {
                    "displayDescription": "This is an example for providers on how to implement their operator to integrate with DBaaS.",
                    "displayName": "DBaaS Provider Example ",
                    "icon": {
                      "base64data": "SGVsbG8sIHdvcmxkLg==",
                      "mediatype": "image/png"
                    },
                    "name": "Provider Example"
                  },
                  "provisioningParameters": {
                    "cloudProvider": {
                      "conditionalData": [
                        {
                          "defaultValue": "GCP",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "FREETRIAL"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        },
                        {
                          "defaultValue": "AWS",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "SERVERLESS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Amazon Web Services",
                              "value": "AWS"
                            },
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        },
                        {
                          "defaultValue": "AWS",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Amazon Web Services",
                              "value": "AWS"
                            },
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        }
                      ],
                      "displayName": "Cloud Provider"
                    },
                    "dedicatedLocationLabel": {
                      "displayName": "Select regions \u0026 nodes",
                      "helpText": "Select the geographical region where you want the database instance to run, and set the number of nodes you want running in this dedicated cluster."
                    },
                    "hardwareLabel": {
                      "displayName": "Hardware per node",
                      "helpText": "Select the compute and storage requirements for this database instance."
                    },
                    "machineType": {
                      "conditionalData": [
                        {
                          "defaultValue": "m5.large",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "AWS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "2 vCPU, 8 GiB RAM",
                              "value": "m5.large"
                            },
                            {
                              "displayValue": "4 vCPU, 16 GiB RAM",
                              "value": "m5.xlarge"
                            },
                            {
                              "displayValue": "8 vCPU, 32 GiB RAM",
                              "value": "m5.2xlarge"
                            },
                            {
                              "displayValue": "16 vCPU, 64 GiB RAM",
                              "value": "m5.4xlarge"
                            },
                            {
                              "displayValue": "32 vCPU, 128 GiB RAM",
                              "value": "m5.8xlarge"
                            }
                          ]
                        },
                        {
                          "defaultValue": "n1-standard-2",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "GCP"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "2 vCPU, 7.5 GiB RAM",
                              "value": "n1-standard-2"
                            },
                            {
                              "displayValue": "4 vCPU, 15 GiB RAM",
                              "value": "n1-standard-4"
                            },
                            {
                              "displayValue": "8 vCPU, 30 GiB RAM",
                              "value": "n1-standard-8"
                            },
                            {
                              "displayValue": "16 vCPU, 60 GiB RAM",
                              "value": "n1-standard-16"
                            },
                            {
                              "displayValue": "32 vCPU, 120 GiB RAM",
                              "value": "n1-standard-32"
                            }
                          ]
                        }
                      ],
                      "displayName": "Compute"
                    },
                    "name": {
                      "displayName": "Cluster name"
                    },
                    "nodes": {
                      "displayName": "nodes"
                    },
                    "plan": {
                      "conditionalData": [
                        {
                          "defaultValue": "SERVERLESS",
                          "options": [
                            {
                              "displayValue": "Free trial",
                              "value": "FREETRIAL"
                            },
                            {
                              "displayValue": "Serverless",
                              "value": "SERVERLESS"
                            },
                            {
                              "displayValue": "Dedicated",
                              "value": "DEDICATED"
                            }
                          ]
                        }
                      ],
                      "displayName": "Hosting plan"
                    },
                    "planLabel": {
                      "displayName": "Select a plan"
                    },
                    "regions": {
                      "displayName": "us-west-2"
                    },
                    "serverlessLocationLabel": {
                      "displayName": "Select regions",
                      "helpText": "Select the geographical region where you want the database instance to run."
                    },
                    "spendLimit": {
                      "conditionalData": [
                        {
                          "defaultValue": "0",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "SERVERLESS"
                            }
                          ]
                        }
                      ],
                      "displayName": "Spend limit"
                    },
                    "spendLimitLabel": {
                      "displayName": "Spend limit",
                      "helpText": "Set a spending limit on resources for this database instance."
                    },
                    "storageGib": {
                      "conditionalData": [
                        {
                          "defaultValue": "15",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "AWS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "15 GiB",
                              "value": "15"
                            },
                            {
                              "displayValue": "35 GiB",
                              "value": "35"
                            },
                            {
                              "displayValue": "75 GiB",
                              "value": "75"
                            },
                            {
                              "displayValue": "150 GiB",
                              "value": "150"
                            },
                            {
                              "displayValue": "300 GiB",
                              "value": "300"
                            },
                            {
                              "displayValue": "600 GiB",
                              "value": "600"
                            }
                          ]
                        },
                        {
                          "defaultValue": "15",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "GCP"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "15 GiB",
                              "value": "15"
                            },
                            {
                              "displayValue": "35 GiB",
                              "value": "35"
                            },
                            {
                              "displayValue": "75 GiB",
                              "value": "75"
                            },
                            {
                              "displayValue": "150 GiB",
                              "value": "150"
                            },
                            {
                              "displayValue": "300 GiB",
                              "value": "300"
                            },
                            {
                              "displayValue": "600 GiB",
                              "value": "600"
                            }
                          ]
                        }
                      ],
                      "displayName": "Storage"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: quay.io/ecosystem-appeng/dbaas-operator:v0.5.0
          description: Discover & connect with database instances hosted on 3rd-party
            ISV cloud platforms.
          operatorframework.io/suggested-namespace: openshift-dbaas-operator
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/internal-objects: '["dbaasconnections.dbaas.redhat.com","dbaasproviders.dbaas.redhat.com","dbaasplatforms.dbaas.redhat.com","dbaasinstances.dbaas.redhat.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/RHEcosystemAppEng/dbaas-operator
          support: rhoda-requests@redhat.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DBaaSConnection
            name: dbaasconnections.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSConnection defines the schema for the DBaaSConnection
              API.
            displayName: DBaaSConnection
            kind: DBaaSConnection
            name: dbaasconnections.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSInstance
            name: dbaasinstances.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSInstance defines the schema for the DBaaSInstance API.
            displayName: DBaaSInstance
            kind: DBaaSInstance
            name: dbaasinstances.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSInventory
            name: dbaasinventories.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSInventory defines the schema for the DBaaSInventory
              API. Inventory objects must be created in a valid namespace, determined
              by the existence of a DBaaSPolicy object.
            displayName: Provider Account
            kind: DBaaSInventory
            name: dbaasinventories.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSPlatform
            name: dbaasplatforms.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSPlatform defines the schema for the DBaaSPlatform API.
            displayName: DBaaSPlatform
            kind: DBaaSPlatform
            name: dbaasplatforms.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSPolicy
            name: dbaaspolicies.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSPolicy enables administrative capabilities within a
              namespace, and sets a default inventory policy. Policy defaults can
              be overridden on a per-inventory basis.
            displayName: Provider Account Policy
            kind: DBaaSPolicy
            name: dbaaspolicies.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSProvider
            name: dbaasproviders.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSProvider defines the schema for the DBaaSProvider API.
            displayName: DBaaSProvider
            kind: DBaaSProvider
            name: dbaasproviders.dbaas.redhat.com
            version: v1beta1
        description: |
          The OpenShift Database Access Operator enables OpenShift users to discover & connect with database instances
          hosted on 3rd-party ISV cloud platforms such as CrunchyData Bridge & CockroachCloud.

          ## Feedback & Support

          Please contact us at [rhoda-requests@redhat.com](mailto:rhoda-requests@redhat.com)

          For more information on how to use our service, please refer to our [Quick Starts](quickstart?keyword=database).
        displayName: OpenShift Database Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rhoda
        - oda
        - dbaas
        - database
        - db
        - data
        - access
        - crunchy
        - bridge
        - cockroach
        - cockroachdb
        - cloud
        - ack
        - amazon
        - rds
        - relational
        links:
        - name: GitHub
          url: https://github.com/RHEcosystemAppEng/dbaas-operator
        maintainers:
        - email: ecosystem-appeng@redhat.com
          name: Red Hat Ecosystem App Engineering
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: OpenShift Database Access
        relatedImages:
        - quay.io/openshift/origin-kube-rbac-proxy:4.10.0
        - quay.io/ecosystem-appeng/dbaas-dynamic-plugin:0.5.0
        - registry.developers.crunchydata.com/crunchydata/crunchy-bridge-operator-catalog:v0.0.6
        - gcr.io/cockroach-shared/ccapi-k8s-operator-catalog:v0.0.5
        - quay.io/ecosystem-appeng/rds-dbaas-operator-catalog:v0.4.0
        - quay.io/ecosystem-appeng/dbaas-operator:v0.5.0
        version: 0.5.0
      entries:
      - name: dbaas-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: dbaas-operator
    provider:
      name: OpenShift Database Access
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ControlPlane Ltd.
      provider-url: https://control-plane.io
    name: flux-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux-operator.v0.31.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "FluxInstance",
                "metadata": {
                  "name": "flux",
                  "namespace": "flux-system",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcileEvery": "1h",
                    "fluxcd.controlplane.io/reconcileTimeout": "5m"
                  }
                },
                "spec": {
                  "distribution": {
                    "version": "2.x",
                    "registry": "ghcr.io/fluxcd"
                  },
                  "components": [
                    "source-controller",
                    "kustomize-controller",
                    "helm-controller",
                    "notification-controller",
                    "image-reflector-controller",
                    "image-automation-controller"
                  ],
                  "cluster": {
                    "type": "openshift",
                    "size": "medium",
                    "domain": "cluster.local",
                    "networkPolicy": true,
                    "multitenant": true,
                    "tenantDefaultServiceAccount": "flux"
                  }
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "FluxReport",
                "metadata": {
                  "name": "flux",
                  "namespace": "flux-system"
                },
                "spec": {
                  "distribution": {
                    "entitlement": "Issued by controlplane",
                    "managedBy": "flux-operator",
                    "status": "Installed",
                    "version": "v2.3.0"
                  },
                "components": [
                    {
                      "image": "ghcr.io/fluxcd/kustomize-controller:v1.3.0@sha256:48a032574dd45c39750ba0f1488e6f1ae36756a38f40976a6b7a588d83acefc1",
                      "name": "kustomize-controller",
                      "ready": true,
                      "status": "Current Deployment is available. Replicas: 1"
                    },
                    {
                      "image": "ghcr.io/fluxcd/source-controller:v1.3.0@sha256:161da425b16b64dda4b3cec2ba0f8d7442973aba29bb446db3b340626181a0bc",
                      "name": "source-controller",
                      "ready": true,
                      "status": "Current Deployment is available. Replicas: 1"
                    }
                  ],
                "reconcilers": [
                    {
                      "apiVersion": "kustomize.toolkit.fluxcd.io/v1",
                      "kind": "Kustomization",
                      "stats": {
                        "failing": 0,
                        "running": 1,
                        "suspended": 0
                      }
                    },
                    {
                      "apiVersion": "source.toolkit.fluxcd.io/v1",
                      "kind": "GitRepository",
                      "stats": {
                        "failing": 0,
                        "running": 1,
                        "suspended": 0,
                        "totalSize": "3.7 MiB"
                      }
                    }
                  ],
                  "sync": {
                    "ready": true,
                    "id": "kustomization/flux-system",
                    "path": "clusters/production",
                    "source": "https://github.com/my-org/my-fleet.git",
                    "status": "Applied revision: refs/heads/main@sha1:a90cd1ac35de01c175f7199315d3f4cd60195911"
                  }
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "ResourceSet",
                "metadata": {
                  "name": "podinfo",
                  "namespace": "default",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcile": "enabled",
                    "fluxcd.controlplane.io/reconcileEvery": "30m",
                    "fluxcd.controlplane.io/reconcileTimeout": "5m"
                  }
                },
                "spec": {
                  "commonMetadata": {
                    "labels": {
                      "app.kubernetes.io/name": "podinfo"
                    }
                  },
                  "inputs": [
                    {
                      "tenant": "team1",
                      "version": "6.7.x",
                      "replicas": "2"
                    },
                    {
                      "tenant": "team2",
                      "version": "6.6.x",
                      "replicas": "3"
                    }
                  ],
                  "resources": [
                    {
                      "apiVersion": "source.toolkit.fluxcd.io/v1beta2",
                      "kind": "OCIRepository",
                      "metadata": {
                        "name": "podinfo-<< inputs.tenant >>",
                        "namespace": "default"
                      },
                      "spec": {
                        "interval": "10m",
                        "url": "oci://ghcr.io/stefanprodan/charts/podinfo",
                        "ref": {
                          "semver": "<< inputs.version | quote >>"
                        }
                      }
                    },
                    {
                      "apiVersion": "helm.toolkit.fluxcd.io/v2",
                      "kind": "HelmRelease",
                      "metadata": {
                        "name": "podinfo-<< inputs.tenant >>",
                        "namespace": "default"
                      },
                      "spec": {
                        "interval": "1h",
                        "releaseName": "podinfo-<< inputs.tenant >>",
                        "chartRef": {
                          "kind": "OCIRepository",
                          "name": "podinfo-<< inputs.tenant >>"
                        },
                        "values": {
                          "replicaCount": "<< inputs.replicas | int >>"
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "ResourceSetInputProvider",
                "metadata": {
                  "name": "demo",
                  "namespace": "default",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcile": "enabled",
                    "fluxcd.controlplane.io/reconcileEvery": "10m",
                    "fluxcd.controlplane.io/reconcileTimeout": "1m"
                  }
                },
                "spec": {
                  "type": "GitHubPullRequest",
                  "url": "https://github.com/fluxcd-testing/pr-testing",
                  "filter": {
                    "limit": 10,
                    "includeBranch": "^stefanprodan-patch-.*$",
                    "labels": [
                      "enhancement"
                    ]
                  },
                  "skip": {
                    "labels": [
                        "deploy/flux-preview-pause",
                        "!test-build-push/passed"
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/controlplaneio-fluxcd/flux-operator:v0.31.0-ubi
          createdAt: "2025-10-06T20:33:03Z"
          description: The Flux Operator manages the lifecycle of the CNCF Flux project.
          operatorframework.io/suggested-namespace: flux-system
          repository: https://github.com/controlplaneio-fluxcd/flux-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Flux Instance
            displayName: FluxInstance
            kind: FluxInstance
            name: fluxinstances.fluxcd.controlplane.io
            version: v1
          - description: Flux Report (Autogenerated)
            displayName: FluxReport
            kind: FluxReport
            name: fluxreports.fluxcd.controlplane.io
            version: v1
          - description: Resource Set
            displayName: ResourceSet
            kind: ResourceSet
            name: resourcesets.fluxcd.controlplane.io
            version: v1
          - description: ResourceSet Input Provider
            displayName: ResourceSetInputProvider
            kind: ResourceSetInputProvider
            name: resourcesetinputproviders.fluxcd.controlplane.io
            version: v1
        description: |-
          [Flux](https://fluxcd.io) is a leading GitOps Continuous Delivery tool used to streamline and automate application deployments on Kubernetes. Flux enables teams to achieve a reliable and auditable delivery process while promoting collaboration and traceability across different environments.
          Flux is powered by the GitOps Toolkit, a set of composable APIs and specialized tools that enable a wide range of continuous delivery use-cases, from simple Kubernetes deployment pipelines to multi-tenant and multi-cluster progressive delivery rollouts.
          For more information about the CNCF-graduated Flux project, please see the [Flux architecture overview](https://fluxcd.control-plane.io/guides/flux-architecture/).
          ### Operator Capabilities
          The [Flux Operator](https://github.com/controlplaneio-fluxcd/flux-operator) provides a declarative API for the installation, configuration and upgrade of CNCF Flux and the ControlPlane [enterprise distribution](https://fluxcd.control-plane.io/).
          The operator allows the configuration of Flux multi-tenancy lockdown, network policies, persistent storage, sharding, vertical scaling, custom patches, and the synchronization of the cluster state from Git repositories, OCI artifacts and S3-compatible storage.
          The Flux Operator supervises the Flux controllers and provides a unified view of all the Flux resources that define the GitOps workflows for the target cluster. The operator generates reports, emits events, and exports Prometheus metrics to help with monitoring and troubleshooting Flux.
          ### Operator Features
          The operator extends Flux with self-service capabilities, deployment windows, and preview environments for GitHub, GitLab and Azure DevOps pull requests testing.
          The operator ResourceSet API enables platform teams to define their own application standard as a group of Flux and Kubernetes resources that can be templated, parameterized and deployed as a single unit on self-service environments. For more information, please see the [ResourceSet documentation](https://fluxcd.control-plane.io/operator/resourcesets/introduction/).
          ### OpenShift Support
          The Flux Operator should be installed in a dedicated namespace, e.g. `flux-system`.
          To deploy Flux on OpenShift clusters, create a `FluxInstance` custom resource named `flux` in the same namespace as the operator and set the `.spec.cluster.type` field to `openshift`.
          For more information on how to configure the Flux instance, please see the [Flux Operator documentation](https://fluxcd.control-plane.io/operator/flux-config/).
        displayName: Flux Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - fluxcd
        - gitops
        - continuous delivery
        links:
        - name: Source Code
          url: https://github.com/controlplaneio-fluxcd/flux-operator
        - name: Documentation
          url: https://fluxcd.control-plane.io/operator/
        - name: Enterprise Support
          url: https://fluxcd.control-plane.io/pricing/
        maintainers:
        - email: flux-enterprise@control-plane.io
          name: ControlPlane Flux Team
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: ControlPlane Ltd.
          url: https://control-plane.io
        relatedImages:
        - ghcr.io/controlplaneio-fluxcd/flux-operator:v0.31.0-ubi
        version: 0.31.0
      entries:
      - name: flux-operator.v0.31.0
        version: 0.31.0
      - name: flux-operator.v0.30.0
        version: 0.30.0
      - name: flux-operator.v0.28.0
        version: 0.28.0
      - name: flux-operator.v0.27.0
        version: 0.27.0
      - name: flux-operator.v0.26.0
        version: 0.26.0
      - name: flux-operator.v0.24.1
        version: 0.24.1
      - name: flux-operator.v0.24.0
        version: 0.24.0
      - name: flux-operator.v0.23.0
        version: 0.23.0
      - name: flux-operator.v0.22.0
        version: 0.22.0
      - name: flux-operator.v0.21.0
        version: 0.21.0
      - name: flux-operator.v0.20.0
        version: 0.20.0
      - name: flux-operator.v0.19.0
        version: 0.19.0
      - name: flux-operator.v0.18.0
        version: 0.18.0
      name: stable
    defaultChannel: stable
    packageName: flux-operator
    provider:
      name: ControlPlane Ltd.
      url: https://control-plane.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-load-core
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-load-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Loadcore",
                "metadata": {
                  "name": "loadcore-sample"
                },
                "spec": {
                  "accessRoute": {
                    "enabled": true,
                    "host": "loadcore.apps.cluster.perf.keysight.lab"
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "reportTemplates": {
                    "enabled": false
                  },
                  "wirelessDataModel": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:42:44Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Loadcore
            name: loadcores.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight LoadCore Operator
        displayName: Keysight LoadCore Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Load Core
          url: https://keysight-load-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-load-core@sha256:245d9be60ff6e9a6e5e42a026a4ee7b8168f391a0491de23ae827d1906af88e5
        version: 0.1.21
      entries:
      - name: keysight-load-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-load-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CloudBees
      provider-url: https://www.cloudbees.com/products/continuous-integration
    name: cloudbees-ci-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudbees-ci.v3.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.cloudbees.com/v1alpha1",
                "kind": "CloudBeesCI",
                "metadata": {
                  "name": "cloudbeesci-sample"
                },
                "spec": {
                  "Agents": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-agent:2.289.3.2"
                    },
                    "ImagePullSecrets": null,
                    "SeparateNamespace": {
                      "Create": false,
                      "Enabled": false,
                      "Name": null
                    }
                  },
                  "Hibernation": {
                    "Enabled": false,
                    "Image": {
                      "dockerImage": "cloudbees/managed-master-hibernation-monitor:247.c5dfce00a179",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "NodeSelector": {},
                    "Tolerations": []
                  },
                  "Master": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-mm:2.289.3.2"
                    },
                    "JavaOpts": null,
                    "OperationsCenterNamespace": null
                  },
                  "NetworkPolicy": {
                    "Enabled": false,
                    "JMXSelectors": [],
                    "ingressControllerSelector": []
                  },
                  "OperationsCenter": {
                    "AgentListenerPort": 50000,
                    "Annotations": {},
                    "CSRF": {
                      "ProxyCompatibility": false
                    },
                    "CasC": {
                      "ConfigMapName": "oc-casc-bundle",
                      "Enabled": false
                    },
                    "ContainerPort": 8080,
                    "ContextPath": null,
                    "Enabled": true,
                    "ExtraConfigMaps": [],
                    "ExtraContainers": [],
                    "ExtraGroovyConfiguration": {},
                    "ExtraVolumeMounts": [],
                    "ExtraVolumes": [],
                    "HealthProbeLivenessFailureThreshold": 12,
                    "HealthProbes": true,
                    "HostName": null,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-cloud-core-oc:2.289.3.2",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "Ingress": {
                      "Annotations": {
                        "kubernetes.io/tls-acme": "false"
                      },
                      "Class": "nginx",
                      "tls": {
                        "Enable": false,
                        "SecretName": null
                      }
                    },
                    "JavaOpts": null,
                    "JenkinsOpts": null,
                    "LoadBalancerIP": null,
                    "LoadBalancerSourceRanges": [
                      "0.0.0.0/0"
                    ],
                    "Name": "cjoc",
                    "NodeSelector": {},
                    "Platform": "standard",
                    "Protocol": "http",
                    "Resources": {
                      "Limits": {
                        "Cpu": 1,
                        "Memory": "2G"
                      },
                      "Requests": {
                        "Cpu": 1,
                        "Memory": "2G"
                      }
                    },
                    "Route": {
                      "tls": {
                        "CACertificate": null,
                        "Certificate": null,
                        "DestinationCACertificate": null,
                        "Enable": false,
                        "InsecureEdgeTerminationPolicy": "Redirect",
                        "Key": null,
                        "Termination": "edge"
                      }
                    },
                    "ServiceAgentListenerPort": 50000,
                    "ServiceAnnotations": {},
                    "ServicePort": 80,
                    "ServiceType": "ClusterIP",
                    "Tolerations": []
                  },
                  "Persistence": {
                    "AccessMode": "ReadWriteOnce",
                    "Annotations": {},
                    "Size": "20Gi",
                    "StorageClass": null
                  },
                  "PodSecurityPolicy": {
                    "Annotations": {},
                    "Enabled": false
                  },
                  "Subdomain": false,
                  "ingress-nginx": {
                    "Enabled": false,
                    "controller": {
                      "admissionWebhooks": {
                        "patch": {
                          "nodeSelector": {
                            "kubernetes.io/os": "linux"
                          }
                        }
                      },
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "nginx-ingress": {
                    "Enabled": false,
                    "controller": {
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "rbac": {
                    "agentsServiceAccountName": "jenkins-agents",
                    "hibernationMonitorServiceAccountName": "managed-master-hibernation-monitor",
                    "install": true,
                    "masterServiceAccountName": "jenkins",
                    "serviceAccountName": "cjoc"
                  },
                  "sidecarinjector": {
                    "Enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cloudbees-ci-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cloudbees-ci-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudBeesCI
            name: cloudbeescis.charts.cloudbees.com
            version: v1alpha1
        description: CloudBees CI
        displayName: CloudBees CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudbees
        - ci
        - enterprise
        - continuous
        - integration
        - jenkins
        - automation
        - pipeline
        - devops
        links:
        - name: CloudBees CI
          url: https://www.cloudbees.com/products/continuous-integration
        maintainers:
        - email: support@cloudbees.com
          name: Support
        - email: sales@cloudbees.com
          name: Sales
        maturity: alpha
        provider:
          name: CloudBees
          url: https://www.cloudbees.com/products/continuous-integration
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.marketplace.redhat.com/rhm/cloudbees/cloudbees-ci-operator@sha256:bd0bd72e9e7807b0fa307de8a93a56967221fe2dd181bac02383f0f9be9120e4
        version: 3.34.1
      entries:
      - name: cloudbees-ci.v3.34.1
        version: 3.34.1
      name: alpha
    defaultChannel: alpha
    packageName: cloudbees-ci-rhmp
    provider:
      name: CloudBees
      url: https://www.cloudbees.com/products/continuous-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhacs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhacs-operator.v3.74.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
          createdAt: "2024-02-06T23:06:13.271846+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.73.0 < 3.74.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:2c0840f4573373a19eb8c9a26501138ef0e76efbd508fd69a6260e78194d86c8
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:0482d4464d6d39c01abab8df882604dd227c52b5420d36fe7414aaff8067cbec
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:71b9fdcaf4cc761b1499be8e6ccff6c463cc6b9856e9deb1f93baad45ea59eb0
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:74624389ac3261ed699b1f64e5717353da68895709d30a011bc8813b61a4a249
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:120b9891ee9db839ab0b2029a62e67ec4747c47c6ffe15af38e9035160f57b78
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:22bf00e1457b4a9e63901892aa4375a1d4a9d1b2c067c1f054d918a24b222b10
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:a108e9a7967001e0bdfed7124b01e9a3dbf1dec308444762973030aec57e1d5a
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:87eb62fe8aa7952d0d69dc13c4867ac2f51e821f56d149a9b588c826caec4c16
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d546f2d3f668ce2f20e594394ade0a86d808f6c8c88cadbba227083d4abac940
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:264102d4bca7d9bd384dc04e5bbff1a7b43343ea8e59cc79047cdda50a3f299f
        version: 3.74.9
      entries:
      - name: rhacs-operator.v3.74.9
        version: 3.74.9
      - name: rhacs-operator.v3.74.8
        version: 3.74.8
      - name: rhacs-operator.v3.74.7
        version: 3.74.7
      - name: rhacs-operator.v3.74.6
        version: 3.74.6
      - name: rhacs-operator.v3.74.5
        version: 3.74.5
      - name: rhacs-operator.v3.74.4
        version: 3.74.4
      - name: rhacs-operator.v3.74.3
        version: 3.74.3
      - name: rhacs-operator.v3.74.2
        version: 3.74.2
      - name: rhacs-operator.v3.74.1
        version: 3.74.1
      - name: rhacs-operator.v3.74.0
        version: 3.74.0
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: latest
    - currentCSV: rhacs-operator.v3.62.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:56905c2e7df0e29ca2b6301cd81b4f9bbe02b483374708f7f7da219a4b29428d
          createdAt: "2021-08-13T17:10:01.410173+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.61.0 < 3.62.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.5.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main page before deploying, or consult the RHACS
              documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create an cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead. If you want to deploy Secured Cluster Service in another namespace, additionally pass the `-n <namespace>` flag.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e30f7835cf52dbce9afdbaf55e5a93d8baf9c4b3a3ae488d1d42c1adba7f46af
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:56905c2e7df0e29ca2b6301cd81b4f9bbe02b483374708f7f7da219a4b29428d
        - registry.redhat.io/rh-acs/collector@sha256:98cfab15684506f1d0978eb701e8de478d91fc5bda3013ffd3b436aa95f07cc0
        - registry.redhat.io/rh-acs/main@sha256:9932304bb62b084ee1ae7e75bb21d6dfb737260c851f2d8daf05082f79230c66
        - registry.redhat.io/rh-acs/scanner@sha256:266a575c2547a3223890e8b73db0f14d69dc5a22217c7e4e3dc79644e6ab0258
        - registry.redhat.io/rh-acs/scanner-db@sha256:ff877f503e08ecca0ce7a7d8db760b2e4004e8e8ae2f27fb9b9c00a05b58c93a
        - registry.redhat.io/rh-acs/collector@sha256:558c0dabdc52efaee7bc889ef397ac2ac30f7c54562c2cc0070d0ef991a8b31b
        version: 3.62.1
      entries:
      - name: rhacs-operator.v3.62.1
        version: 3.62.1
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.62
    - currentCSV: rhacs-operator.v3.64.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fa478167f2f738f244bed6a06450679be73f8464c40f86acde4ffd390505692b
          createdAt: "2021-09-22T05:10:01.286657+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.63.0 < 3.64.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0cacd00c0a147cc8295138565ef2dc2ddbd9ad1fb25f6be817719a75ea4c6448
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fa478167f2f738f244bed6a06450679be73f8464c40f86acde4ffd390505692b
        - registry.redhat.io/rh-acs/scanner@sha256:8b7121bce8f9c7133d29245ebb265a4b8f9831a66a57f11b2d68b9b8b5cf79cc
        - registry.redhat.io/rh-acs/scanner-db@sha256:67e81bf896267b7acab8cb6755cde552173de4f0881c38eaca8e491bd926cc53
        - registry.redhat.io/rh-acs/collector@sha256:2b3a8bd3af62516a6714fe612e5fdc0433b91d873182e290d3ddc2b32ce291f0
        - registry.redhat.io/rh-acs/collector@sha256:3eb099da081cf208d46be5ccfa7b5e9ac0e2cad49f127fa4db91e49e3f26db42
        - registry.redhat.io/rh-acs/main@sha256:6aeb2223124540ac217c38b016dea99d770ab897dde4906a821d8776d91ea03d
        version: 3.64.2
      entries:
      - name: rhacs-operator.v3.64.2
        version: 3.64.2
      - name: rhacs-operator.v3.64.1
        version: 3.64.1
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.64
    - currentCSV: rhacs-operator.v3.65.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:229c7e5f967c70dfbc3cf638d521267ea6acf875b055499dfd30cb19c0b1f69c
          createdAt: "2021-09-21T13:26:28.711610+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.64.0 < 3.65.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/collector@sha256:3eb099da081cf208d46be5ccfa7b5e9ac0e2cad49f127fa4db91e49e3f26db42
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0cacd00c0a147cc8295138565ef2dc2ddbd9ad1fb25f6be817719a75ea4c6448
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:229c7e5f967c70dfbc3cf638d521267ea6acf875b055499dfd30cb19c0b1f69c
        - registry.redhat.io/rh-acs/main@sha256:e672ef9a1239065f8cf0f331757861fea082e971a9e9760d1e65795f399cbde1
        - registry.redhat.io/rh-acs/scanner@sha256:bfe9b2205bfca6aebed14827b4bda098bfad574e67ffa967ad339543062200fe
        - registry.redhat.io/rh-acs/scanner-db@sha256:20f258500ac2ba4e626df92dbcd193cf7ab8687ac251591be2442b8a46871835
        - registry.redhat.io/rh-acs/collector@sha256:2b3a8bd3af62516a6714fe612e5fdc0433b91d873182e290d3ddc2b32ce291f0
        version: 3.65.1
      entries:
      - name: rhacs-operator.v3.65.1
        version: 3.65.1
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.65
    - currentCSV: rhacs-operator.v3.66.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:10cdb0ff3e67b65737cde420536494bd84c2b7ccf65192bbed2eb43eaed66a41
          createdAt: "2021-10-20T16:38:38.091764+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.65.0 < 3.66.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bd74f0b95ff2881faf6eb4520feecd067b69a759f09bbfecd467447ce2a2a25e
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:10cdb0ff3e67b65737cde420536494bd84c2b7ccf65192bbed2eb43eaed66a41
        - registry.redhat.io/rh-acs/main@sha256:97230c01e76d90c7fe35ffeaa591be3d9175f350a8fd88599aded09a1edf930c
        - registry.redhat.io/rh-acs/scanner@sha256:323ee86d6a9cbeb32270c659c1154f3a62c36535c6ea894e4e29e93eb1f4a84f
        - registry.redhat.io/rh-acs/scanner-db@sha256:490b0bf49161e4bc6de2504726aaf360e0ca9a214aea6b886fff13153bdb263d
        - registry.redhat.io/rh-acs/collector@sha256:5673c27554a71502e06e08945a4424c6ce5da27f489854f3aa452e6d31240aee
        - registry.redhat.io/rh-acs/collector@sha256:d93e13ee1d1cc4b7f82e644a7d43e560aac42bd1e06d25b476cea9a4338bf9ce
        version: 3.66.1
      entries:
      - name: rhacs-operator.v3.66.1
        version: 3.66.1
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.66
    - currentCSV: rhacs-operator.v3.67.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:893ebcb2d0c96647e56bce3b16fe32851df59cd71f2c0fc8f6e66ca3a669bb70
          createdAt: "2021-12-16T12:54:10.117625+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.66.0 < 3.67.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/scanner@sha256:2860ed271da6064149bb53568c5ad60e129918d55a2f244f9ade4a0bea4c4933
        - registry.redhat.io/rh-acs/scanner-db@sha256:6f8d47277732eb56317d576c2d06fb1f539d46d16520e6ca34abec290392cef9
        - registry.redhat.io/rh-acs/collector@sha256:57c2dfa3dbc8368076291fbc627baa01dcb5ebc50f838e02ce7ac38a128d027b
        - registry.redhat.io/rh-acs/collector@sha256:e4648723ccbd43fa5acf005d9734bfc06d41e19e1be0bde839880e28fa361e16
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86e5fa1fa294987114be200890c2e516501e424aee0fb98ece25c95e7716295b
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:893ebcb2d0c96647e56bce3b16fe32851df59cd71f2c0fc8f6e66ca3a669bb70
        - registry.redhat.io/rh-acs/main@sha256:8db7630d094b661411ebb3f35248848d3147c735677d0922f7a04de9a41eacc9
        version: 3.67.2
      entries:
      - name: rhacs-operator.v3.67.2
        version: 3.67.2
      - name: rhacs-operator.v3.67.1
        version: 3.67.1
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.67
    - currentCSV: rhacs-operator.v3.68.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:195a694dc96096166fe82e81ecb3ceb493f96cfb5029112e576e44ab39b56abd
          createdAt: "2022-06-15T22:13:22.427644+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.67.0 < 3.68.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:195a694dc96096166fe82e81ecb3ceb493f96cfb5029112e576e44ab39b56abd
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:f0d2f5f11c2a96d34d6b4baf249cbf821fe064eade73c3f95c22f3c6981ac4d1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:105d10bd50a6062c998ed107dea388dd78f0caddb0b57713c8692b1d07042774
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:081f2198ad70bfb2ec66153b50b48b950b9a5550ba996de413c5a2aff4052783
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c9d0a5a2871c24c0b1058e419f2b090429363a319856397855412fd2c0037395
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:9b9b9146d9683d6da0ff2abbc8ea82b96d945b7a8c3740d72d8c018fb2b04fba
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:0bb76363a3a25916dd4493e177e9b2d36eac9633473b07fe0fffb853fd072da3
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:711293438bab35b17d2e1f162e7f74f667a6e16dd32e45122f917b6817fefb4e
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:b5fc3dcaa72bbc8dea005189c940e1cf2d00c94a6cf7676cf0e1790e1191d45c
        version: 3.68.2
      entries:
      - name: rhacs-operator.v3.68.2
        version: 3.68.2
      - name: rhacs-operator.v3.68.1
        version: 3.68.1
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.68
    - currentCSV: rhacs-operator.v3.69.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9b2beceb255124855ba1b352d16009b2f2cd422901a19c311f0783dab1fcc3a0
          createdAt: "2022-06-21T09:07:40.705944+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.68.0 < 3.69.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:08a7d8c892fad69c589b81ad8fbd21d36c8fa4078530dba361cfa9a8f9f05950
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:9ae03669cfdc0f4fe56b021d911edb17159c07757720ae30a1f22d7f7a897c07
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:f167e928740d1c614e26948ea58f49e569f6a4fddc438cd0abe52f60538432d0
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:867613178d491f416e57975843817251490c5fbca2fc69d8b55b169bea675279
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:8db5206532e95a7ad7cb23f389155d41757aaf5f445ddce92b008a06627d3fd5
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:77f5db63e17b140a8d6b84e3822fd77463035be9c0c832b55b4977696da13ccc
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:da6dd486a20fdba61ea12d2c77ea299d9795a9f22d9057d65ff992b1f5896e07
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:b254746ed3faa1f1cbcafdf1343a1957bf0214311b36f4e892be3000cc2aecf3
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9b2beceb255124855ba1b352d16009b2f2cd422901a19c311f0783dab1fcc3a0
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:2afb2823edc6f3434e18db60ee37381b5b79f51234c1a261b6d0f8ed74cc8b63
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:cf5629823f2dd583495bcbbcf3b83b4f29f855ace3c388f5cb7b24a6d34ed267
        version: 3.69.2
      entries:
      - name: rhacs-operator.v3.69.2
        version: 3.69.2
      - name: rhacs-operator.v3.69.1
        version: 3.69.1
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.69
    - currentCSV: rhacs-operator.v3.70.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fb903bf9e110ba03e9c63977ebaf707e6884dd8fd413bcd6b793a0ed795cdd0e
          createdAt: "2022-06-20T12:30:26.999648+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.69.0 < 3.70.1'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:dfeb239931f7a33ade5c829a102cabc01f5c348e1aa35a7ddc24757cfd549629
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:df8cfdaa40b50ff2f2670d97c2115b47ae4bea443ae2831caaa468dd62f3c75f
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fb903bf9e110ba03e9c63977ebaf707e6884dd8fd413bcd6b793a0ed795cdd0e
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:6c32b86ed8bc18e384706eb6188406e66910bf34c8869bd155a60c3525d0fea7
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:d2e827a209840b7197960c8115720972237532eb9cf9d6fee7a36107053408ce
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:3d7c8b77671740e324ee6c309a84bbaa4d925e6750979fe4e46c561bc79f1b4d
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:b47c96af1c9ffe4d8728da510a93efe21adc221225fe2b822cfcace6c6672ba2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:08a7d8c892fad69c589b81ad8fbd21d36c8fa4078530dba361cfa9a8f9f05950
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:ee55a233d2a8fd77d89ec2c69f357721c7debead6470a9f494b3cd6a001829c4
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:6b14e850cda9a54618847858292ea669abff599a40692e27fed0e337c4a60526
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:e71f2f2347dd46dec790405b1bb98d54a348a4583949f142f5703be9a790f96e
        version: 3.70.1
      entries:
      - name: rhacs-operator.v3.70.1
        version: 3.70.1
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.70
    - currentCSV: rhacs-operator.v3.71.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:469b6c1592226eb9a521131ef758a33e6a6ac7c96bc67fcc903cea30469c125e
          createdAt: "2022-12-07T08:31:36.777654+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.70.0 < 3.71.3'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4a0dc8c337299d7f6440ff75c2d5219d3ab68183f4f117f008ef065757dd5d4b
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:469b6c1592226eb9a521131ef758a33e6a6ac7c96bc67fcc903cea30469c125e
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:f16f92e074f7125b88e9e2033c7c12b44421145b93d040034fa4688626869af1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:c3ec88cd99908894d0e67153779062f90ac1b007980d66040a79b2be13798a10
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:4d8bac9b3857af6ba351c7768f70ef30ef7cb99ae4df60cb226b5a793353a052
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:7f279bb6e0f34a5208d8711ba4d7b62c78f08b683815ce33f6aa28c23c4246da
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:9bc25e81c21cda6d8b535d8ac67c1286fe57a1596001850e730f8f73bf371111
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:e282a8e5e606003ea344d44ac5ee7699563848f7160eeae183aa58ad740e0e94
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:cb2d0f470968c2b725b0d176d423cc1866fd56893344d077d28f2c16581d3614
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:df7b41999b542a477848823a1d0e7e5c29ae1b00fbac035007a5a1f33fac139f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:5fdfd3690af60bd5ecf04741cd1f73259491da501700b066120976960567f0db
        version: 3.71.3
      entries:
      - name: rhacs-operator.v3.71.3
        version: 3.71.3
      - name: rhacs-operator.v3.71.2
        version: 3.71.2
      - name: rhacs-operator.v3.71.1
        version: 3.71.1
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.71
    - currentCSV: rhacs-operator.v3.72.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:3ff61e1cc77b66a22c8a8c0ab306f0e122cd74e762005285aede23ebb6c2f6a2
          createdAt: "2023-02-28T17:02:08.443613+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.71.0 < 3.72.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:575b6c3e2914751df1668ffff83cbda187ce6e56f32214414cdf036b7a62bbd8
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d56a00b3e20e60b7169f5c35d03fa6a1bd3b9980dd3ebc9d89d759e1d6d20ad2
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:cd475a6c2356d63cfe736a11114de7fedfbd8fdcfecc56224a00b6c819103be6
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:42e5f6c4ba789f124ac3ebe7b6d1a607ede8e763a35888c97ccbd0259bec4a1c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:648b14e4158da2a85eedb46b3456fb502d680d52d375ad34bf45b4b94f8b2ef6
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:3ff61e1cc77b66a22c8a8c0ab306f0e122cd74e762005285aede23ebb6c2f6a2
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:cd0c205ee44eddcdf99b84e80fcf4403851d956efc50532b9d12013bafe44988
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:0b69595405e4e598183bbf4197ac5b7593724fa6162413f225aa01a542af18c9
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:622f30b7f591194d86a5bf95deede0225e66f7b624db197f6289a5fbbbe887a9
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:c07fae3616cbc681e4d31839abea6623b4bf4dade93273201c8db7ba0d635aad
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:cb5df858e0c5cae8f77b285e23daeb83062bd72570c90ed5ea3ce4a02d2f5f08
        version: 3.72.4
      entries:
      - name: rhacs-operator.v3.72.4
        version: 3.72.4
      - name: rhacs-operator.v3.72.3
        version: 3.72.3
      - name: rhacs-operator.v3.72.2
        version: 3.72.2
      - name: rhacs-operator.v3.72.1
        version: 3.72.1
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.72
    - currentCSV: rhacs-operator.v3.73.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a8486b66f977a639d77ef110492e5ecdde47be8acaefbaa1cbb5417085c56839
          createdAt: "2023-05-26T21:27:04.033276+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.72.0 < 3.73.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:7bf9415aa4b0746d11ca5a59d4179bc382a0fac4ded969f0c4294171fde27d58
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:e6687c12d070069f655c17832d94c619c18274a259ccc59715036f9908e162f7
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:ee2a9e4ab74f5a554d1e4ff148833a53bdaab70ac1e2668cc9ee2fceca290ba9
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:082bd6f4b40209e0591be001dea358f328cd86a08691cdceb7f71ea59d861d9d
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:4d0ead8969f120a629104a92a454842bb7103aec137deb6c8209b7db9c48a1b2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:648b14e4158da2a85eedb46b3456fb502d680d52d375ad34bf45b4b94f8b2ef6
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a8486b66f977a639d77ef110492e5ecdde47be8acaefbaa1cbb5417085c56839
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:e2c89f735969d438f75a89ff63c404a8ec4361a03fde48e880fee73006b9c669
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:e4a27d3ffb173c87728fb637ca7fde92a9e78ea9ee3edcecccc2a5e3d3748b8c
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:dce5db4f162dd40c09cad114cb802cd6d7c72d0c6986581f2e74c977493f75b0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:c020a80ff82bf4d9b5cec3eeac5de9990de76041e22c444e664a20e15c46ce3b
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:5df359156b01505ce43e5f14c3b08722856bdb86f3d0d9b4fba785a2e580f346
        version: 3.73.5
      entries:
      - name: rhacs-operator.v3.73.5
        version: 3.73.5
      - name: rhacs-operator.v3.73.4
        version: 3.73.4
      - name: rhacs-operator.v3.73.3
        version: 3.73.3
      - name: rhacs-operator.v3.73.2
        version: 3.73.2
      - name: rhacs-operator.v3.73.1
        version: 3.73.1
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.73
    - currentCSV: rhacs-operator.v3.74.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
          createdAt: "2024-02-06T23:06:13.271846+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.73.0 < 3.74.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d546f2d3f668ce2f20e594394ade0a86d808f6c8c88cadbba227083d4abac940
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:264102d4bca7d9bd384dc04e5bbff1a7b43343ea8e59cc79047cdda50a3f299f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:0482d4464d6d39c01abab8df882604dd227c52b5420d36fe7414aaff8067cbec
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:2c0840f4573373a19eb8c9a26501138ef0e76efbd508fd69a6260e78194d86c8
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:74624389ac3261ed699b1f64e5717353da68895709d30a011bc8813b61a4a249
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:87eb62fe8aa7952d0d69dc13c4867ac2f51e821f56d149a9b588c826caec4c16
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:120b9891ee9db839ab0b2029a62e67ec4747c47c6ffe15af38e9035160f57b78
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:22bf00e1457b4a9e63901892aa4375a1d4a9d1b2c067c1f054d918a24b222b10
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:a108e9a7967001e0bdfed7124b01e9a3dbf1dec308444762973030aec57e1d5a
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:71b9fdcaf4cc761b1499be8e6ccff6c463cc6b9856e9deb1f93baad45ea59eb0
        version: 3.74.9
      entries:
      - name: rhacs-operator.v3.74.9
        version: 3.74.9
      - name: rhacs-operator.v3.74.8
        version: 3.74.8
      - name: rhacs-operator.v3.74.7
        version: 3.74.7
      - name: rhacs-operator.v3.74.6
        version: 3.74.6
      - name: rhacs-operator.v3.74.5
        version: 3.74.5
      - name: rhacs-operator.v3.74.4
        version: 3.74.4
      - name: rhacs-operator.v3.74.3
        version: 3.74.3
      - name: rhacs-operator.v3.74.2
        version: 3.74.2
      - name: rhacs-operator.v3.74.1
        version: 3.74.1
      - name: rhacs-operator.v3.74.0
        version: 3.74.0
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.74
    - currentCSV: rhacs-operator.v4.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9bf3a3e0fd573f91f03794eaf5291364a90392e5e7d4b5d4ed550ef4600ab826
          createdAt: "2023-10-20T07:04:50.531740+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.74.0 < 4.0.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:08f537cba262605a0eb0a510b6faf78e3ba26465f636d03cc1f7a6ed6b40d355
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:eee1aa054d4d15b23be293f31d3ea659e0ffc1d05174ef89a6c977dd3d4377cc
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:4c848eb7465e928251e05d32ffbc93cd14e7bb77972d22b8721bd6eaa4a93d2f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:3250d89e78de852a2627593f29469f0b3ff639ea3644b4b7d6d057d0c1ad3659
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:5a492b856e0b6d9fdb020b556d13186ef5e7095b0c736a2031500c9c00dad0c1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:139d80d398e7e5cfc85839e3039a3630b48c7ae6650d755754430fccd94d6a27
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0aedd8e865f72ffa8b79fd2bab48974caec3eea3e3a6f77c89fe6dacb8947057
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9bf3a3e0fd573f91f03794eaf5291364a90392e5e7d4b5d4ed550ef4600ab826
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:cfa10bb6b461f302674f185b26c8d6efaaeba7f3f65d2cf05ae1f069aa2710d3
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:c39621837cbc8419d187a43a9575ec8cb4021c58ea1b56f215003d7264eb5d03
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:09e79b5db8977520f9951cc43c339c8e40b4659d7da513347d5a88b2500f809d
        version: 4.0.5
      entries:
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.0
    - currentCSV: rhacs-operator.v4.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:8ffb6f0ff1bf016c3919c1eefc6c3db3d4a8fbe0e79919f72b4c281afe6d4b1c
          createdAt: "2024-01-13T18:55:22.011709+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 4.0.0 < 4.1.6'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:364c107c34694ca33ac724de78db3e41ee55f9098d9ade0775a51b11f1533065
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:8ffb6f0ff1bf016c3919c1eefc6c3db3d4a8fbe0e79919f72b4c281afe6d4b1c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a8ad0e76ce4cae9f9b36aecf125c4ed56fc2d776fe3df113d41bfc1c2f8b3602
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:8547b76cf0533dfe6da2aafbbd063021bad4ff78b32b2dfed4a7607496e0a904
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:8760d6dc68c86f0c5b53d8f0bfada0df012f80bc8e0023ba3522fe6829a05a41
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:762ad216e79be0837e98873202b0072c7fd399925ee94bde57cc61c68cbd5f4e
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:01b9f9798d3ecbefe3da1ac2ac782e8e8b930e7eeceb82fe3473b5bfc1ccd40a
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:7ac504b7824d17bd7ac55b1e59c26be308506f315a0ab0d8c45d3b4934c0a16c
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:2ec62f3da93a1346ad215744257820377ccf4ead84e7471c52e348a31a431d3c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:bdbe08ef2d01b114702780fab76905e6eb1851684692ffa0a4c84604bda3ddb3
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:34694b9f4fe7fffc0c547be4e311579b5bb0c1fa92d8257bf6d1581ffd655c46
        version: 4.1.6
      entries:
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.1
    - currentCSV: rhacs-operator.v4.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4253eca84cbb0c5fc033507ddc2deef235b054c0ba944fa271ebbba21545b872
          createdAt: "2024-03-13T14:51:55.084122+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.1.0 < 4.2.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:b5a15f2985b114bf9bc24a0d9942db5d0cb818e60fd15a2f9094aa00904cce77
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4253eca84cbb0c5fc033507ddc2deef235b054c0ba944fa271ebbba21545b872
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:12db245af3b8475c9b5680bb0035c566109819e05cfe62bd1dfdfb5d6ed25bea
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:b82f7f654235688fc07fb037422a73e1106e27024d39a68ef7fac1422b69496d
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:ff27f03064633d6d08f8383649069e39c868def4aeca8f2622750397e049d0ec
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:2f326cabf03f780a3e6120995d5442641764641f62430d82e05edbe21b30c7b2
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:3858ab936d4626a11bbf895d41bad181364ee5df3312f6dd29ebeb9b7d31ec77
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:e2b07cb7d690477d78a2b8e1875545998d4bdb946f80dd0533a14187da99dbbf
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:59f1862c8ce188906499f08201ae05deeb47bd4fc5eab4be0386ca44ec065da8
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:e79a5e5681ebfe2d6349e2763ee0c1fa521b20319846958e402d23229b1f257b
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:4b35f5f39865238a8e95d3cba7ebb5f15ccfbf6a2563c34f57d91a0b5c927066
        version: 4.2.5
      entries:
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.2
    - currentCSV: rhacs-operator.v4.3.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:266cbbc270efafa26d5d15f3e333c96eaedee3b47854ff337ae1c572d34901ce
          createdAt: "2024-06-06T22:31:19.365878+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.2.0 < 4.3.8'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c36a557c9dc2ae4479c432933a9a4dbef38f2be53360a68a5b6befb614d01dac
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:b0c9b3858fc0547dcb4eb3065dedd5bbcf1e1f6707a8ab812ed06882d4f6d6ad
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:b7eb17fbcf6c4a6e4205ad66f1ffd9b10a4634f6ef5045d407c266566b1f255d
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:bdf9785bff51194636c6e8a518448c6d12a297bd31c4bb2e79d07084409c18b4
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:266cbbc270efafa26d5d15f3e333c96eaedee3b47854ff337ae1c572d34901ce
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:9fba889021a876ab627ff8c496e4b6c46c221c61142cd0c76daf33c89762e2c1
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:10d0cc82b6eddd089f3a7af37b872c33248513d0b4d3f094b6f62c01f4460825
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:6b55a25eb3a611c648d9be883828834592229c16f4166fd28ef40a25fc667061
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:2c35cb6f518393d002edd2bf5f714c1d053cd0fab55f6152a2f35676e4979202
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:2bb643651c8d51525223df3b2b675947174bbe9f1eaea8688e1fa83044cfa41f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:5092c3fbb4a38c06a2f42d125dc718ef0e1005f304ad8046afdd2a5e7ccbc580
        version: 4.3.8
      entries:
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.3
    - currentCSV: rhacs-operator.v4.4.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:2b87569d7f4ef5e06c989b273c378e007c6d87b26eec41774ac1fc67dcd9e4e4
          createdAt: "2025-02-10T19:50:13.597164+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.3.0 < 4.4.8'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:65707fdad55767ff5598a8d3679fd0779a37f59501e58971b81bebcb23f1f9fc
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:d6b4f8caf342b105c8df14c5c9180cb36123e08d34756ee8fe4f2f76c34f8064
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:664a5dc40dc1fa06c47a11e13b0ab5e0612a981d7b69bcfb29422abf3d1a13b2
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:39cfd01236e905034f6645a73063427436c14e351e0e0bc735e73e62aa5a1c3c
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:467d069ee957aea07882b83b60ae9fcc62ccca159f5d643aac13e531ccf7df88
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:2b87569d7f4ef5e06c989b273c378e007c6d87b26eec41774ac1fc67dcd9e4e4
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:0df4e95c7406aee7812d145cdd4d8c61cec32857fd2079dbc4244954a9d8288a
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:183c75920ab8e7cb947cb966f1543e39b6fdb5e1bb0caa1b0cf5f89671f0328f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:76f9dc2e2ee5595c36530d97dec216e2dbefada2fe0747e3f6a3a725b75722ab
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:763e7568704971cb4656727dc223374de13e82b4304dd3f5387ddc0f4e42b8b6
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:69c12230a785619c765e3749a7d56f76dce9ce4f530c20e8a9387b99eb65627b
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:ba6e25a413c79131b966a097f72c36478933ea52deb55a62acda7011a96b3983
        version: 4.4.8
      entries:
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.4
    - currentCSV: rhacs-operator.v4.5.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a7dfbe3590668f91fe6694335b5a35b2fca07b4735348cc5202921755d2f654d
          createdAt: "2025-04-09T22:51:15.157672+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.4.0 < 4.5.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:5914548f6ce2160e92c5d96580ffba23cdf1ae940c1f73f3f6424d15c6e97c37
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:dc7f23341c8b209f1846f29f889d00ef232fb58da9d5b4bd561f3933093b2786
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:d928600f75461465598a10de8df2495a275fd5232a92b078f609d1f5774eb778
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:1817cf8d2c08b04391792321702ef53d16a4b1ca06459c00c9c053128f534358
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:c2656b794ec4236e36f48f71bf26bdedd9de27fc4acef39deba2bdcfcddab03b
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:cc4b329cc795752ecbb178d2e4efc152bde54b04181334eb5bd70304c3354cad
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:0ff948b5517df3773cae2df02b708c348d8707409899b1de4108588e31601c0c
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a7dfbe3590668f91fe6694335b5a35b2fca07b4735348cc5202921755d2f654d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:5109b304eb9458f97af85c2ba63e08f51672fa3f9406226b7f357b86b5828dc0
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:fde78d34ecd2944c7c16b4e823c76b3667bf6c5b468e9991128ecc4272ec0e2e
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:ed33e7396357f1db03a065573c020846a5872b12a21284439770686689222434
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:f07cba0738897e0e7ae74971be0470ce2a855657f133a6e1cdd8a7092f6151b0
        version: 4.5.9
      entries:
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.5
    - currentCSV: rhacs-operator.v4.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:c4d434e1ff052508e5594d2b95dafa8726fd8adf5317754774c013a46de224d8
          createdAt: "2025-09-26T09:24:28.046251+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.5.0 < 4.6.10'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:b5cd08349931669055f03d21331499ef2ea04597f8a90fb201615d43000271bc
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:ffe41363dc00b05b0daa0c6d16efb30e534086aa4e3d7ccc599c86d33813138c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:713e61d3c8b55de1edc9d7834d3e8576718e1327665f94a6b58178fa7496c0cc
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:65793c05aec8085a53fab8bcf33a85192c17b46c91a7d86e0a457d1e3a113878
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:899da9c23000bb821bc9d544ebb56bcda4f3245ceee5b1b9c12d09f580b4b9a1
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:8fad8ab2efdcbc99d9efa736a90cca6f236f0a9d45fdb355ae0b6b4dc42d77f0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:90c47c6f95571cfcca04aa4e05a18befad75c1b44a6f3a055fcdfdab749d29c5
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:16b61ca1b9ee4e0821e7fc82fe4def75ed22950e502730f5704ec4fe63bb2e59
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:c4d434e1ff052508e5594d2b95dafa8726fd8adf5317754774c013a46de224d8
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:b630d53ad0b862151be7e14f0483ca5adac570216deae66f60e75af87a4dfd46
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:19db8c5d2425dcf18ae63b025f946fc46adaec0578e9505415cb585797069e95
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:6da59c5b82625894f6efa3fa0bc5ea0378edd4f95cf42dee772e9c8a565dba81
        version: 4.6.10
      entries:
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.6
    - currentCSV: rhacs-operator.v4.7.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:5b60d8271d6fa70df3c8f7e0ecd0d64feecfdf674dc84036500fe5f145c9415a
          createdAt: "2025-09-26T11:05:05.900682+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.6.0 < 4.7.7'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:def71e1ddea5b8c1e54c8582f39dc9ecca855b38aa9423690150aa6ef0ff189f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8882c00c9be9b579654a9ee8cd0c38537cf92d2f4bd8ebded1678aa0b6ae827
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:34ea97c6e9fc2d23427d0c6f2628075c0eaa0e151ca1c381539b039c249217f7
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:98553e6f18d3c03a29212b4f7ea1d30a42bcf8b52ca12aecbcbc47769fc35b53
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:bac79889d4717233e89bb6c8fae6732ed7bec5b9a55fd3b3de1f1311d870d2b1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:6bb454d64a9e339b1aeb81ca17b26f33db4a84f19ced4ad81858e40fd01878dc
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:5b60d8271d6fa70df3c8f7e0ecd0d64feecfdf674dc84036500fe5f145c9415a
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:7596529cbdce7118226b79d69e0667ab35ee692569e4014b74abdf0fa7415222
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:5ed9902b52830be93436d7977c10e73d346b6290532f6d7245cfe7cca9c93b31
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:0c32318ee4cbd547e2c1fd64433cd23a629d1b06523d7c0f07ddb98c2baaa708
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:d8fcfe1267495801d23a0b94101c7b4d06b48346a949c97946ebcbffe148f4b6
        version: 4.7.7
      entries:
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.7
    - currentCSV: rhacs-operator.v4.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
          createdAt: "2025-09-12T11:06:11.361461+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.7.0 < 4.8.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:adf47cb6615d344533fcca0aee41774f5e8df9b952b9b89a57c2be2b7e0fc03c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:92ff7691963e620dc5bf6da0a7694e6bcd68c2545afc1642a59a443ba2ab5611
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:db691065e46f604b980b0b202f6c2e14c8b204094d784c4a15db1e4fecb2965e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:97df59280c12727a99fdec0c9bf56eb874e6a1cacd12d849709797d4be635ce9
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:e5c200f28187c1f6971562e41cae969183ee15f3d0feb076bb202b22ff67cbcb
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8964b1441251607b23f92918e9158724174013dc4af423988f56d6098dfb2d5
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:e949830f3034c41aa5434a3654b2b1ccba7dac0373bf43d7c50a9ea43bb686fb
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:16912df2c26e5fc2e6051c0501170722856db274415063608de716cf53851038
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:d1758b76c32f03b5a1a44951d18fb2bb0c4fd0a99d24f8c42fe9e6b059574889
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0c92d5b5508032d77559606927e45686179eaafc59e16d5496e088f2f212d258
        version: 4.8.4
      entries:
      - name: rhacs-operator.v4.8.4
        version: 4.8.4
      - name: rhacs-operator.v4.8.3
        version: 4.8.3
      - name: rhacs-operator.v4.8.2
        version: 4.8.2
      - name: rhacs-operator.v4.8.1
        version: 4.8.1
      - name: rhacs-operator.v4.8.0
        version: 4.8.0
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.8
    - currentCSV: rhacs-operator.v4.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
          createdAt: "2025-09-12T11:06:11.361461+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.7.0 < 4.8.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:e5c200f28187c1f6971562e41cae969183ee15f3d0feb076bb202b22ff67cbcb
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:e949830f3034c41aa5434a3654b2b1ccba7dac0373bf43d7c50a9ea43bb686fb
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:16912df2c26e5fc2e6051c0501170722856db274415063608de716cf53851038
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:adf47cb6615d344533fcca0aee41774f5e8df9b952b9b89a57c2be2b7e0fc03c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:97df59280c12727a99fdec0c9bf56eb874e6a1cacd12d849709797d4be635ce9
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:d1758b76c32f03b5a1a44951d18fb2bb0c4fd0a99d24f8c42fe9e6b059574889
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8964b1441251607b23f92918e9158724174013dc4af423988f56d6098dfb2d5
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0c92d5b5508032d77559606927e45686179eaafc59e16d5496e088f2f212d258
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:db691065e46f604b980b0b202f6c2e14c8b204094d784c4a15db1e4fecb2965e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:92ff7691963e620dc5bf6da0a7694e6bcd68c2545afc1642a59a443ba2ab5611
        version: 4.8.4
      entries:
      - name: rhacs-operator.v4.8.4
        version: 4.8.4
      - name: rhacs-operator.v4.8.3
        version: 4.8.3
      - name: rhacs-operator.v4.8.2
        version: 4.8.2
      - name: rhacs-operator.v4.8.1
        version: 4.8.1
      - name: rhacs-operator.v4.8.0
        version: 4.8.0
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: stable
    defaultChannel: stable
    packageName: rhacs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dns-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dns-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSRecord",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dns-operator",
                    "app.kubernetes.io/instance": "dnsrecord-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dnsrecord",
                    "app.kubernetes.io/part-of": "dns-operator"
                  },
                  "name": "dnsrecord-sample"
                },
                "spec": {
                  "endpoints": [
                    {
                      "dnsName": "dnsrecord-simple.example.com",
                      "recordTTL": 60,
                      "recordType": "A",
                      "targets": [
                        "52.215.108.61",
                        "52.30.101.221"
                      ]
                    }
                  ],
                  "providerRef": {
                    "name": "dns-provider-creds"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/dns-operator:v0.6.0
          createdAt: "2024-09-13T09:55:39Z"
          description: A Kubernetes Operator to manage the lifecycle of DNS resources
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kuadrant/dns-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSRecord is the Schema for the dnsrecords API
            displayName: DNSRecord
            kind: DNSRecord
            name: dnsrecords.kuadrant.io
            version: v1alpha1
        description: A Kubernetes Operator to manage the lifecycle of DNS resources
        displayName: DNS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dns
        - kuadrant
        links:
        - name: DNS Operator
          url: https://github.com/Kuadrant/dns-operator
        maintainers:
        - email: mnairn@redhat.com
          name: Michael Nairn
        - email: pbrookes@redhat.com
          name: Phil Brookes
        - email: cbrookes@redhat.com
          name: Craig Brookes
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/dns-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: dns-operator.v0.6.0
        version: 0.6.0
      name: stable
    defaultChannel: stable
    packageName: dns-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://www.redhat.com
    name: sap-commerce-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sap-commerce-operator.v0.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hybris.hybris.org/v1alpha1",
                "kind": "HybrisApp",
                "metadata": {
                  "name": "hybrisapp-sample"
                },
                "spec": {
                  "aJPServicePort": 30000,
                  "apachejvmRouteName": "",
                  "baseImageName": "hybris-base",
                  "baseImageTag": "v0.2",
                  "hybrisANTTaskNames": "",
                  "podHealthyProbePath": "/platform/init",
                  "sourceRepoLocalPropertiesOverride": "",
                  "sourceRepoRef": "main",
                  "sourceRepoURL": "https://github.com/redhat-sap/sap-commerce-app-example.git",
                  "startupProbeFailureThreshold": 100,
                  "startupProbePeriodSecond": 60
                }
              },
              {
                "apiVersion": "hybris.hybris.org/v1alpha1",
                "kind": "HybrisBase",
                "metadata": {
                  "name": "hybrisbase-sample"
                },
                "spec": {
                  "URL": "http://nexus.redhat.com/nexus/service/local/repositories/releases/content/org/hybris/hybris/1.0/hybris-1.0.tar.gz",
                  "buildSourceRepo": "https://github.com/redhat-sap/sap-commerce-operator",
                  "buildSourceRepoBranch": "main",
                  "imageName": "hybris-base",
                  "imageTag": "v0.2",
                  "jdkURL": "https://github.com/SAP/SapMachine/releases/download/sapmachine-11.0.5/sapmachine-jdk-11.0.5-1.x86_64.rpm",
                  "password": "admin123",
                  "username": "admin"
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          containerImage: quay.io/redhat-sap-cop/hybris-operator:0.0.4
          description: Builds and runs SAP Commerce in OpenShift
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-sap/sap-commerce-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HybrisApp is the Schema for the hybrisapps API
            displayName: Hybris Application
            kind: HybrisApp
            name: hybrisapps.hybris.hybris.org
            version: v1alpha1
          - description: HybrisBase is the Schema for the Hybris Base API
            displayName: Hybris Base
            kind: HybrisBase
            name: hybrisbases.hybris.hybris.org
            version: v1alpha1
        description: |-
          Builds and runs SAP Commerce in OpenShift

          * The operator provides APIs to initialize SAP Commerce with the provided SAP Commerce Server. Though there are no restrictions from the operator on the base package to be used from SAP, the default SAP base image provides a server preconfigured with an application server based on Apache Tomcat. Both HSQLDB and Tomcat are included in the SAP Commerce package, and are designed to be used for testing, development, and demonstration purposes. The custom classes created for various business needs along with an external RDBMS configuration can be used in a HybrisApp creation to execute SAP Commerce in a production setup.

          * The operator also enables the use of SAP Commerce in a clustered model. A SAP Commerce Cluster is a number of individual SAP Commerce installations using a common set of data on one database. A cluster configuration lets you balance the load between the nodes to scale SAP Commerce in a way that is transparent to the user.

          * The clustering offers few communication protocols as options (UDP: both multicast or unicast, or JGroups). In order to support multicast for SAP Commerce clustering the corresponding OpenShift namespace has to be enabled for multicast usage.

          For more details on SAP Commerce consult the product documentation: [https://help.sap.com/viewer/product/SAP_COMMERCE](https://help.sap.com/viewer/product/SAP_COMMERCE)
        displayName: SAP Commerce Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sap-commerce-operator
        links:
        - name: SAP Commerce Operator
          url: https://github.com/redhat-sap/sap-commerce-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
          url: https://www.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/redhat-sap-cop/hybris-operator:0.0.4
        version: 0.0.4
      entries:
      - name: sap-commerce-operator.v0.0.4
        version: 0.0.4
      - name: sap-commerce-operator.v0.0.3
        version: 0.0.3
      - name: sap-commerce-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: sap-commerce-operator
    provider:
      name: Red Hat, Inc.
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cognitoidentityprovider-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cognitoidentityprovider-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cognitoidentityprovider-controller:1.1.1
          createdAt: "2025-09-19T20:58:49Z"
          description: AWS Cognito Identity Provider controller is a service controller
            for managing Cognito Identity Provider resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: UserPool represents the state of an AWS cognitoidentityprovider
              UserPool resource.
            displayName: UserPool
            kind: UserPool
            name: userpools.cognitoidentityprovider.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Cognito Identity Provider resources in AWS from within your Kubernetes cluster.

          **About Amazon Cognito Identity Provider**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Cognito Identity Provider
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cognitoidentityprovider
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Cognito Identity Provider Developer Resources
          url: https://aws.amazon.com/Cognito Identity Provider/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cognitoidentityprovider maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cognitoidentityprovider-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-cognitoidentityprovider-controller.v1.1.1
        version: 1.1.1
      - name: ack-cognitoidentityprovider-controller.v1.1.0
        version: 1.1.0
      - name: ack-cognitoidentityprovider-controller.v1.0.12
        version: 1.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-cognitoidentityprovider-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sumo Logic
      provider-url: https://www.sumologic.com/
    name: sumologic-kubernetes-collection-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "helm-operator.sumologic.com/v1alpha1",
                    "kind": "SumologicCollection",
                    "metadata": {
                        "name": "collection"
                    },
                    "spec": {
                        "sumologic": {
                            "accessId": "",
                            "accessKey": "",
                            "clusterName": "kubernetes",
                            "scc": {
                                "create": true
                            }
                        },
                        "kube-prometheus-stack": {
                            "prometheus-node-exporter": {
                                "service": {
                                    "port": 9200,
                                    "targetPort": 9200
                                }
                            }
                        },
                        "otellogs": {
                            "daemonset": {
                                "containers": {
                                    "otelcol": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                },
                                "initContainers": {
                                    "changeowner": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                }
                            }
                        }
                    }
                }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing,Monitoring
          containerImage: registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
          createdAt: "2024-07-19T10:48:00Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator.git
          support: Sumo Logic
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configuration for Sumo Logic Kubernetes Collection
            displayName: SumologicCollection
            kind: SumologicCollection
            name: sumologiccollections.helm-operator.sumologic.com
            version: v1alpha1
        description: Helm Operator for the Sumo Logic Kubernetes Collection Helm Chart
          v4.13.0-2
        displayName: Sumo Logic Kubernetes Collection Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - collection
        - logging
        - metrics
        - monitoring
        links:
        - name: Sumo Logic Kubernetes Collection Helm Operator
          url: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator
        maintainers:
        - email: collection@sumologic.com
          name: Sumo Logic
        maturity: alpha
        provider:
          name: Sumo Logic
          url: https://www.sumologic.com/
        relatedImages:
        - registry.connect.redhat.com/sumologic/autoinstrumentation-java@sha256:242958b7d8905f4ac02f6ad38897ba58b9badf2a6213a389e67dcc71fee3630c
        - registry.connect.redhat.com/sumologic/thanos@sha256:16faf7191668e86d8b576bba055dd0cbdeb009635976b9e07bd00f926032f11c
        - registry.connect.redhat.com/sumologic/tailing-sidecar@sha256:738d653a78dc415243ab3627db9135768175fb7f51d0803566f5d9e824c8744b
        - registry.connect.redhat.com/sumologic/opentelemetry-operator@sha256:5587fe5b7ebfa57ad320ad2521103f908317388d1741c40e5f7be5ed9e5f9ad7
        - registry.connect.redhat.com/sumologic/kube-state-metrics@sha256:3551ededa327a35db798ec8286cd843580e3ae395d2231e041e8c4b7621127ee
        - registry.connect.redhat.com/sumologic/tailing-sidecar-operator@sha256:c16b8b18e9b263fc3eb24059bba512bcfc3201b32105af6e5518392826ddd66c
        - registry.connect.redhat.com/sumologic/telegraf-operator@sha256:ff958222be01c0c3966474463e43bbd5bcf348790dd34bb761718cb1cea2d793
        - registry.connect.redhat.com/sumologic/kubernetes-tools-kubectl@sha256:d3155b864cc74f327adedf7d013a8f0fcb11c2e73a5655b7cafa848824e80172
        - registry.connect.redhat.com/sumologic/kubernetes-setup@sha256:f07eaf68fa11ffc708e4485a83bd3763fc5712f7d73d29edfe5b9c9e08f05df6
        - registry.connect.redhat.com/sumologic/sumologic-otel-collector@sha256:4554d43dba204e3ec8e91b36442a438f10234fa0cb4db7fb7880b257a6485112
        - registry.connect.redhat.com/sumologic/autoinstrumentation-nodejs@sha256:c206eef4fdecaa58b4f01e5e7f0dbd74e9606eab2b4debf47461d36191e248d1
        - registry.connect.redhat.com/sumologic/autoinstrumentation-dotnet@sha256:f4f385c4489c74b5c43f2fe6b6620a9eb8d9ecab470567236b3c45f636106640
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
        - registry.connect.redhat.com/sumologic/metrics-server@sha256:65d2a3ddc4b4183f438fa18b655079c93c905cb9e3b8c54977336205dc1fff07
        - registry.connect.redhat.com/sumologic/nginx-unprivileged@sha256:7260555d16c94c4fe1e016626b4cb8a5c0de6b700e2eea28744331fb298f71ae
        - registry.connect.redhat.com/sumologic/node-exporter@sha256:11e9e604432a66d9b388cb1bf70eac077c7f0372d1264342ef927d76b8ef0272
        - registry.connect.redhat.com/sumologic/prometheus-operator@sha256:58a329e6582456b32abebb470086191f8366b927af4703741613993465f116b8
        - registry.connect.redhat.com/sumologic/prometheus@sha256:0ef6eaf570a40499dd5de92a79593e950bc771b014e0ee6ff7419b4b82a36fcc
        - registry.connect.redhat.com/sumologic/autoinstrumentation-python@sha256:f0fa45fa45e64c7df88f21181ae18ee96fe83cd3988258663d0806ea6f39e6bb
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:e1f49fcf5b7e86d9eb49c16dbbf9484cb486f55f999fd51d7c738c3b1f2666cf
        - registry.connect.redhat.com/sumologic/telegraf@sha256:75efd2e509c6017ea0fc5c9103ce068cbc89087b16506a4ef48cb5ff00eab84e
        - registry.connect.redhat.com/sumologic/busybox@sha256:209bc8e0b1d958fc699f3bb75b1248544e372ad35cdc46c991bd1698f641e1f9
        - registry.connect.redhat.com/sumologic/prometheus-config-reloader@sha256:0b605d36782afd8f50fd631d0e9c7b19d39ce962af89b7a737b2f09a8ceff291
        version: 4.13.0-2
      entries:
      - name: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
        version: 4.13.0-2
      name: alpha
    defaultChannel: alpha
    packageName: sumologic-kubernetes-collection-helm-operator
    provider:
      name: Sumo Logic
      url: https://www.sumologic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: rh-service-binding-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-binding-operator.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ServiceBinding expresses intent to bind an operator-backed
              service with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding for Kubernetes. Describes the connection
              between a Provisioned Service and an Application Projection. Implements
              Service Binding Specifcation for Kubernetes (https://github.com/k8s-service-bindings/spec).
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: alpha
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:fe6629401bc8447da95ec55949b5a27e620364bb6e02f86230262699ac8a5d55
        version: 0.9.1
      entries:
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: beta
    - currentCSV: service-binding-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindablekinds-sample"
                }
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:e799bd06043028d214baeeffc683710ad2eed7b83272fc653e3b89f69711319b
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:e799bd06043028d214baeeffc683710ad2eed7b83272fc653e3b89f69711319b
        version: 0.11.0
      entries:
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: candidate
    - currentCSV: service-binding-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:3d7a3648e86848402577dd19c1f8fd130a5556ebdfbc0f663a72f9e141024812
        version: 0.10.0
      entries:
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: preview
    - currentCSV: service-binding-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindable-kinds"
                },
                "spec": {}
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  },
                  "type": "postgres",
                  "workload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: registry.redhat.io/ocp-tools-4/service-binding-rhel8-operator@sha256:9c8fb0c8f9d2f54685561af21e595aa88bf309eaebebdf23fb6997b75d026b18
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'.
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1alpha3
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1beta1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
          - description: Service Binding implementing community specification (https://github.com/k8s-service-bindings/spec)
              describing the connection between a backing service and an application
              workload.
            displayName: Service Binding (Spec API)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1beta1
        description: |
          The Service Binding Operator manages the data plane for applications and backing services. It enables developers to connect their
          applications to backing services (REST API, databases, event buses and others) with a consistent and predictable
          experience. The operator reads data made available by the control plane of backing services and projects the bindings
          data to applications either as accessible files mounted into the application’s container or environment variables.

          In Kubernetes, each service suggests a different way to access their secrets and developers will consume them in their
          application in a custom way. While this provides a good deal of flexibility, it also creates discrepancies and
          complexity dealing with each specific services, which leads large development teams to lose overall velocity.

          Service Binding Operator removes the error-prone manual configuration of binding information, provides service operators
          a low-touch administrative experience to provision and manage access to services and enriches development lifecycle.

          For additional details refer to the project [documentation](https://redhat-developer.github.io/service-binding-operator).

          Get Started with Service Binding Operator
          * Find out about SBO’s [prerequisites](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/installing-service-binding.html).
          * Refer to the [Quick Start Guide](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/quick-start.html) to see SBO in action in a simple scenario.

          Need Help
          * Raise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)

          Licence
          * Service Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE)

          Service Binding Operator is providing and supporting two different APIs.
          * `binding.operators.coreos.com/v1alpha1`: This API is compliant with the Service Binding specification. It is fully supported and considered as the mature API. It’s the API we recommend in production environments.
          * `servicebinding.io/v1beta1`: This API is the one defined in the Service Binding specification. It is available in Service Binding Operator 1.x.

          We’ll consider supporting only the API from the specification, once the Service Binding specification will be officially GA. Of course, that will be a breaking change and handle in a safe way for you.
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Documentation
          url: https://redhat-developer.github.io/service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4/service-binding-rhel8-operator@sha256:9c8fb0c8f9d2f54685561af21e595aa88bf309eaebebdf23fb6997b75d026b18
        version: 1.3.3
      entries:
      - name: service-binding-operator.v1.3.3
        version: 1.3.3
      - name: service-binding-operator.v1.3.1
        version: 1.3.1
      - name: service-binding-operator.v1.3.0
        version: 1.3.0
      - name: service-binding-operator.v1.2.0
        version: 1.2.0
      - name: service-binding-operator.v1.1.1
        version: 1.1.1
      - name: service-binding-operator.v1.1.0
        version: 1.1.0
      - name: service-binding-operator.v1.0.1
        version: 1.0.1
      - name: service-binding-operator.v1.0.0
        version: 1.0.0
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: rh-service-binding-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cluster-logging
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-logging.v5.9.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "name": "remote-elasticsearch",
                      "secret": {
                        "name": "elasticsearch"
                      },
                      "type": "elasticsearch",
                      "url": "remote.example.org:9200"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "audit",
                        "infrastructure"
                      ],
                      "name": "enable-default-log-store",
                      "outputRefs": [
                        "default"
                      ]
                    },
                    {
                      "inputRefs": [
                        "application"
                      ],
                      "name": "forward-to-remote",
                      "outputRefs": [
                        "remote-elasticsearch"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogging",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "collection": {
                    "type": "vector"
                  },
                  "logStore": {
                    "elasticsearch": {
                      "nodeCount": 3,
                      "redundancyPolicy": "SingleRedundancy",
                      "resources": {
                        "requests": {
                          "memory": "2Gi"
                        }
                      },
                      "storage": {
                        "size": "200G"
                      }
                    },
                    "retentionPolicy": {
                      "application": {
                        "maxAge": "7d"
                      }
                    },
                    "type": "elasticsearch"
                  },
                  "managementState": "Managed",
                  "visualization": {
                    "kibana": {
                      "replicas": 1
                    },
                    "type": "kibana"
                  }
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          console.openshift.io/plugins: '["logging-view-plugin"]'
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:54054faf0d2f375dd6d7fc7929a2dbb529508e1b3c004a56fc97c2a682d16be2
          createdAt: "2025-03-14T20:19:21Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing your aggregated logging stack.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.7.0-0 <5.9.15'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Logging (team-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.


              There are built-in input names for common log categories, and you can
              define custom inputs to do additional filtering.


              There is a built-in output name for the default openshift log store, but
              you can define your own outputs with a URL and other connection information
              to forward logs to other stores or processors, inside or outside the cluster.


              For more details see the documentation on the API fields.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.logging.openshift.io
            version: v1
          - description: A Red Hat OpenShift Logging instance. ClusterLogging is the
              Schema for the clusterloggings API
            displayName: Cluster Logging
            kind: ClusterLogging
            name: clusterloggings.logging.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates and manages the aggregated logging stack as a cluster-wide service.

          ##Features
          * **Create/Destroy**: Launch and create an aggregated logging stack to support the entire OCP cluster.
          * **Simplified Configuration**: Configure your aggregated logging cluster's structure like components and end points easily.
          * Configure log forwarding to an in-cluster, Red Hat managed log storage solution based on LokiStack or Elasticsearch

          ## Prerequisites
          ### Red Hat OpenShift Logging Namespace
          Cluster logging and the Red Hat OpenShift Logging Operator is only deployable to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - fluentd
        - logging
        - aggregated
        - efk
        - vector
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Fluentd
          url: https://www.fluentd.org/
        - name: Vector
          url: https://vector.dev/
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:54054faf0d2f375dd6d7fc7929a2dbb529508e1b3c004a56fc97c2a682d16be2
        - registry.redhat.io/openshift-logging/logging-view-plugin-rhel9@sha256:8023b072b45e9a21ddef83a56f4760d476053250d5a920d8eb5f03eecab8b1b3
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:54e3a5efaae5ecc2a08db03a421aa79e323b48c764115ea84d83a37c21d1d997
        - registry.redhat.io/openshift-logging/fluentd-rhel9@sha256:a676c9164ec2523b2002dc5eba234ce774926dbf7fe87de6d72b53f99cbf1db5
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:865572275b79462f14e84c69438a61afe9465cfe45ec164caf30965730b276f7
        version: 5.9.15
      entries:
      - name: cluster-logging.v5.9.15
        version: 5.9.15
      - name: cluster-logging.v5.9.14
        version: 5.9.14
      - name: cluster-logging.v5.9.13
        version: 5.9.13
      - name: cluster-logging.v5.9.12
        version: 5.9.12
      - name: cluster-logging.v5.9.11
        version: 5.9.11
      - name: cluster-logging.v5.9.10
        version: 5.9.10
      - name: cluster-logging.v5.9.9
        version: 5.9.9
      - name: cluster-logging.v5.9.8
        version: 5.9.8
      - name: cluster-logging.v5.9.7
        version: 5.9.7
      - name: cluster-logging.v5.9.6
        version: 5.9.6
      - name: cluster-logging.v5.9.5
        version: 5.9.5
      - name: cluster-logging.v5.9.4
        version: 5.9.4
      - name: cluster-logging.v5.9.3
        version: 5.9.3
      - name: cluster-logging.v5.9.2
        version: 5.9.2
      - name: cluster-logging.v5.9.1
        version: 5.9.1
      - name: cluster-logging.v5.9.0
        version: 5.9.0
      name: stable
    - currentCSV: cluster-logging.v5.8.22
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "name": "remote-elasticsearch",
                      "secret": {
                        "name": "elasticsearch"
                      },
                      "type": "elasticsearch",
                      "url": "remote.example.org:9200"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "audit",
                        "infrastructure"
                      ],
                      "name": "enable-default-log-store",
                      "outputRefs": [
                        "default"
                      ]
                    },
                    {
                      "inputRefs": [
                        "application"
                      ],
                      "name": "forward-to-remote",
                      "outputRefs": [
                        "remote-elasticsearch"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogging",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "collection": {
                    "type": "vector"
                  },
                  "logStore": {
                    "elasticsearch": {
                      "nodeCount": 3,
                      "redundancyPolicy": "SingleRedundancy",
                      "resources": {
                        "requests": {
                          "memory": "2Gi"
                        }
                      },
                      "storage": {
                        "size": "200G"
                      }
                    },
                    "retentionPolicy": {
                      "application": {
                        "maxAge": "7d"
                      }
                    },
                    "type": "elasticsearch"
                  },
                  "managementState": "Managed",
                  "visualization": {
                    "kibana": {
                      "replicas": 1
                    },
                    "type": "kibana"
                  }
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.12-v4.17
          console.openshift.io/plugins: '["logging-view-plugin"]'
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:8e13bd3e1654fe26c1d4d8d4a1bfb12e04cee7899cf16b5fa07b62b8f7494af5
          createdAt: "2024-10-17T17:45:03Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing your aggregated logging stack.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.22'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-5.8
          operators.operatorframework.io.bundle.channels.v1: stable-5.8
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Logging (team-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "ClusterLogForwarder is an API to configure forwarding logs.
              \n You configure forwarding by specifying a list of `pipelines`, which
              forward from a set of named inputs to a set of named outputs. \n There
              are built-in input names for common log categories, and you can define
              custom inputs to do additional filtering. \n There is a built-in output
              name for the default openshift log store, but you can define your own
              outputs with a URL and other connection information to forward logs
              to other stores or processors, inside or outside the cluster. \n For
              more details see the documentation on the API fields."
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.logging.openshift.io
            version: v1
          - description: A Red Hat OpenShift Logging instance. ClusterLogging is the
              Schema for the clusterloggings API
            displayName: Cluster Logging
            kind: ClusterLogging
            name: clusterloggings.logging.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          ### Notice for OCP 4.16 and 4.17 Only

          Logging 5.8 is supported on OCP 4.16 and 4.17 to help customers upgrade and migrate to Logging 6.0. **Logging 5.8 will be End of Life in November 2025.**  At that time, it will no longer be available or supported.

          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates and manages the aggregated logging stack as a cluster-wide service.

          ##Features
          * **Create/Destroy**: Launch and create an aggregated logging stack to support the entire OCP cluster.
          * **Simplified Configuration**: Configure your aggregated logging cluster's structure like components and end points easily.
          * Configure log forwarding to an in-cluster, Red Hat managed log storage solution based on LokiStack or Elasticsearch

          ## Prerequisites
          ### Red Hat OpenShift Logging Namespace
          Cluster logging and the Red Hat OpenShift Logging Operator is only deployable to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - fluentd
        - logging
        - aggregated
        - efk
        - vector
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Fluentd
          url: https://www.fluentd.org/
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/logging-view-plugin-rhel9@sha256:f00d4e7ce6b37f1a053c5dbe554359d844989c2243574000ada0f5bc678ffff8
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:8e13bd3e1654fe26c1d4d8d4a1bfb12e04cee7899cf16b5fa07b62b8f7494af5
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:f26518dd3dbf6b6a9e9e20e40741687996b1160c6ad8ad417989197800c2b3fe
        - registry.redhat.io/openshift-logging/fluentd-rhel9@sha256:9296e50fccae60dea580186bc5ae2a387172dc90ece35663d96c4553a2446a01
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:8f29d6eaeb4a11c9a61e3d763605e3c9c8655df7fbb4e39e5b8311d18449f04a
        version: 5.8.22
      entries:
      - name: cluster-logging.v5.8.22
        version: 5.8.22
      - name: cluster-logging.v5.8.21
        version: 5.8.21
      - name: cluster-logging.v5.8.20
        version: 5.8.20
      - name: cluster-logging.v5.8.19
        version: 5.8.19
      - name: cluster-logging.v5.8.18
        version: 5.8.18
      - name: cluster-logging.v5.8.17
        version: 5.8.17
      - name: cluster-logging.v5.8.16
        version: 5.8.16
      - name: cluster-logging.v5.8.15
        version: 5.8.15
      - name: cluster-logging.v5.8.14
        version: 5.8.14
      - name: cluster-logging.v5.8.13
        version: 5.8.13
      - name: cluster-logging.v5.8.12
        version: 5.8.12
      - name: cluster-logging.v5.8.11
        version: 5.8.11
      - name: cluster-logging.v5.8.10
        version: 5.8.10
      - name: cluster-logging.v5.8.9
        version: 5.8.9
      - name: cluster-logging.v5.8.8
        version: 5.8.8
      - name: cluster-logging.v5.8.7
        version: 5.8.7
      - name: cluster-logging.v5.8.6
        version: 5.8.6
      - name: cluster-logging.v5.8.5
        version: 5.8.5
      - name: cluster-logging.v5.8.4
        version: 5.8.4
      - name: cluster-logging.v5.8.3
        version: 5.8.3
      - name: cluster-logging.v5.8.2
        version: 5.8.2
      - name: cluster-logging.v5.8.1
        version: 5.8.1
      - name: cluster-logging.v5.8.0
        version: 5.8.0
      name: stable-5.8
    - currentCSV: cluster-logging.v5.9.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "name": "remote-elasticsearch",
                      "secret": {
                        "name": "elasticsearch"
                      },
                      "type": "elasticsearch",
                      "url": "remote.example.org:9200"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "audit",
                        "infrastructure"
                      ],
                      "name": "enable-default-log-store",
                      "outputRefs": [
                        "default"
                      ]
                    },
                    {
                      "inputRefs": [
                        "application"
                      ],
                      "name": "forward-to-remote",
                      "outputRefs": [
                        "remote-elasticsearch"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogging",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "collection": {
                    "type": "vector"
                  },
                  "logStore": {
                    "elasticsearch": {
                      "nodeCount": 3,
                      "redundancyPolicy": "SingleRedundancy",
                      "resources": {
                        "requests": {
                          "memory": "2Gi"
                        }
                      },
                      "storage": {
                        "size": "200G"
                      }
                    },
                    "retentionPolicy": {
                      "application": {
                        "maxAge": "7d"
                      }
                    },
                    "type": "elasticsearch"
                  },
                  "managementState": "Managed",
                  "visualization": {
                    "kibana": {
                      "replicas": 1
                    },
                    "type": "kibana"
                  }
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          console.openshift.io/plugins: '["logging-view-plugin"]'
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:54054faf0d2f375dd6d7fc7929a2dbb529508e1b3c004a56fc97c2a682d16be2
          createdAt: "2025-03-14T20:19:21Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing your aggregated logging stack.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.7.0-0 <5.9.15'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Logging (team-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.


              There are built-in input names for common log categories, and you can
              define custom inputs to do additional filtering.


              There is a built-in output name for the default openshift log store, but
              you can define your own outputs with a URL and other connection information
              to forward logs to other stores or processors, inside or outside the cluster.


              For more details see the documentation on the API fields.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.logging.openshift.io
            version: v1
          - description: A Red Hat OpenShift Logging instance. ClusterLogging is the
              Schema for the clusterloggings API
            displayName: Cluster Logging
            kind: ClusterLogging
            name: clusterloggings.logging.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates and manages the aggregated logging stack as a cluster-wide service.

          ##Features
          * **Create/Destroy**: Launch and create an aggregated logging stack to support the entire OCP cluster.
          * **Simplified Configuration**: Configure your aggregated logging cluster's structure like components and end points easily.
          * Configure log forwarding to an in-cluster, Red Hat managed log storage solution based on LokiStack or Elasticsearch

          ## Prerequisites
          ### Red Hat OpenShift Logging Namespace
          Cluster logging and the Red Hat OpenShift Logging Operator is only deployable to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - fluentd
        - logging
        - aggregated
        - efk
        - vector
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Fluentd
          url: https://www.fluentd.org/
        - name: Vector
          url: https://vector.dev/
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:54054faf0d2f375dd6d7fc7929a2dbb529508e1b3c004a56fc97c2a682d16be2
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:54e3a5efaae5ecc2a08db03a421aa79e323b48c764115ea84d83a37c21d1d997
        - registry.redhat.io/openshift-logging/fluentd-rhel9@sha256:a676c9164ec2523b2002dc5eba234ce774926dbf7fe87de6d72b53f99cbf1db5
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:865572275b79462f14e84c69438a61afe9465cfe45ec164caf30965730b276f7
        - registry.redhat.io/openshift-logging/logging-view-plugin-rhel9@sha256:8023b072b45e9a21ddef83a56f4760d476053250d5a920d8eb5f03eecab8b1b3
        version: 5.9.15
      entries:
      - name: cluster-logging.v5.9.15
        version: 5.9.15
      - name: cluster-logging.v5.9.14
        version: 5.9.14
      - name: cluster-logging.v5.9.13
        version: 5.9.13
      - name: cluster-logging.v5.9.12
        version: 5.9.12
      - name: cluster-logging.v5.9.11
        version: 5.9.11
      - name: cluster-logging.v5.9.10
        version: 5.9.10
      - name: cluster-logging.v5.9.9
        version: 5.9.9
      - name: cluster-logging.v5.9.8
        version: 5.9.8
      - name: cluster-logging.v5.9.7
        version: 5.9.7
      - name: cluster-logging.v5.9.6
        version: 5.9.6
      - name: cluster-logging.v5.9.5
        version: 5.9.5
      - name: cluster-logging.v5.9.4
        version: 5.9.4
      - name: cluster-logging.v5.9.3
        version: 5.9.3
      - name: cluster-logging.v5.9.2
        version: 5.9.2
      - name: cluster-logging.v5.9.1
        version: 5.9.1
      - name: cluster-logging.v5.9.0
        version: 5.9.0
      name: stable-5.9
    - currentCSV: cluster-logging.v6.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.14-v4.17
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:3e36ea9a808945d3817d169f736b1758875f2fbc6a7ae845266e1545a1b713c7
          createdAt: "2025-05-14T12:32:39Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.8.0-0 <6.0.10'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.0
          operators.operatorframework.io.bundle.channels.v1: stable-6.0
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:3e36ea9a808945d3817d169f736b1758875f2fbc6a7ae845266e1545a1b713c7
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:4e9936c090818d1d8f812487a9a8218fda5b11f8c06167fffe35bae1bafcadf9
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:15487d7b9d8f83d0df0b6fbf380f20df3518dfb1b831fac4c3209e811494b436
        version: 6.0.10
      entries:
      - name: cluster-logging.v6.0.10
        version: 6.0.10
      - name: cluster-logging.v6.0.9
        version: 6.0.9
      - name: cluster-logging.v6.0.8
        version: 6.0.8
      - name: cluster-logging.v6.0.7
        version: 6.0.7
      - name: cluster-logging.v6.0.6
        version: 6.0.6
      - name: cluster-logging.v6.0.5
        version: 6.0.5
      - name: cluster-logging.v6.0.4
        version: 6.0.4
      - name: cluster-logging.v6.0.3
        version: 6.0.3
      - name: cluster-logging.v6.0.2
        version: 6.0.2
      - name: cluster-logging.v6.0.1
        version: 6.0.1
      - name: cluster-logging.v6.0.0
        version: 6.0.0
      name: stable-6.0
    - currentCSV: cluster-logging.v6.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.15-v4.18
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f7101c962de4e5643bd7f760104a41cc48f020889159a97e2d6711703fbbec64
          createdAt: "2025-05-14T12:38:55Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.9.0-0 <6.1.8'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.1
          operators.operatorframework.io.bundle.channels.v1: stable-6.1
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f7101c962de4e5643bd7f760104a41cc48f020889159a97e2d6711703fbbec64
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:96e438dd5c080cfa47080410f375dc71da24f6d046cf609db41c8b873d85404f
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:76037c744b34b1c4fccf18f17888813d9670bb6c061c2c3de843aa43c0497170
        version: 6.1.8
      entries:
      - name: cluster-logging.v6.1.8
        version: 6.1.8
      - name: cluster-logging.v6.1.7
        version: 6.1.7
      - name: cluster-logging.v6.1.6
        version: 6.1.6
      - name: cluster-logging.v6.1.5
        version: 6.1.5
      - name: cluster-logging.v6.1.4
        version: 6.1.4
      - name: cluster-logging.v6.1.3
        version: 6.1.3
      - name: cluster-logging.v6.1.2
        version: 6.1.2
      - name: cluster-logging.v6.1.1
        version: 6.1.1
      - name: cluster-logging.v6.1.0
        version: 6.1.0
      name: stable-6.1
    - currentCSV: cluster-logging.v6.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.16-v4.19
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
          createdAt: "2025-07-29T15:25:15Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0-0 <6.2.5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.2
          operators.operatorframework.io.bundle.channels.v1: stable-6.2
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Installing logging](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html-single/logging/index#installing-logging-6-2)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:c76c853565fc93da88ba0116425839cfb86cab63cbde9856d7cca78278d7758a
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:502a5d2891049bbc32115fda28086841351cd9e6a3110b4c3c412f683f2b2491
        version: 6.2.5
      entries:
      - name: cluster-logging.v6.2.5
        version: 6.2.5
      - name: cluster-logging.v6.2.4
        version: 6.2.4
      - name: cluster-logging.v6.2.3
        version: 6.2.3
      - name: cluster-logging.v6.2.2
        version: 6.2.2
      - name: cluster-logging.v6.2.1
        version: 6.2.1
      - name: cluster-logging.v6.2.0
        version: 6.2.0
      name: stable-6.2
    defaultChannel: stable-6.2
    packageName: cluster-logging
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-apigateway-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-apigateway-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Resource",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "RestAPI",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Stage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "VPCLink",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/apigateway-controller:1.3.1
          createdAt: "2025-09-19T21:13:12Z"
          description: AWS API Gateway controller is a service controller for managing
            API Gateway resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIIntegrationResponse represents the state of an AWS apigateway
              APIIntegrationResponse resource.
            displayName: APIIntegrationResponse
            kind: APIIntegrationResponse
            name: apiintegrationresponses.apigateway.services.k8s.aws
            version: v1alpha1
          - description: APIKey represents the state of an AWS apigateway APIKey resource.
            displayName: APIKey
            kind: APIKey
            name: apikeys.apigateway.services.k8s.aws
            version: v1alpha1
          - description: APIMethodResponse represents the state of an AWS apigateway
              APIMethodResponse resource.
            displayName: APIMethodResponse
            kind: APIMethodResponse
            name: apimethodresponses.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Authorizer represents the state of an AWS apigateway Authorizer
              resource.
            displayName: Authorizer
            kind: Authorizer
            name: authorizers.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Deployment represents the state of an AWS apigateway Deployment
              resource.
            displayName: Deployment
            kind: Deployment
            name: deployments.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Integration represents the state of an AWS apigateway Integration
              resource.
            displayName: Integration
            kind: Integration
            name: integrations.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Method represents the state of an AWS apigateway Method resource.
            displayName: Method
            kind: Method
            name: methods.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Resource represents the state of an AWS apigateway Resource
              resource.
            displayName: Resource
            kind: Resource
            name: resources.apigateway.services.k8s.aws
            version: v1alpha1
          - description: RestAPI represents the state of an AWS apigateway RestAPI
              resource.
            displayName: RestAPI
            kind: RestAPI
            name: restapis.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Stage represents the state of an AWS apigateway Stage resource.
            displayName: Stage
            kind: Stage
            name: stages.apigateway.services.k8s.aws
            version: v1alpha1
          - description: VPCLink represents the state of an AWS apigateway VPCLink
              resource.
            displayName: VPCLink
            kind: VPCLink
            name: vpclinks.apigateway.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon API Gateway resources in AWS from within your Kubernetes cluster.

          **About Amazon API Gateway**

          Amazon API Gateway is an AWS service for creating, publishing, maintaining, monitoring, and securing REST, HTTP, and WebSocket APIs at any scale. API developers can create APIs that access AWS or other web services, as well as data stored in the AWS Cloud. As an API Gateway API developer, you can create APIs for use in your own client applications. Or you can make your APIs available to third-party app developers.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon API Gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apigateway
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon API Gateway Developer Resources
          url: https://aws.amazon.com/api-gateway/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: apigateway maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/apigateway-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-apigateway-controller.v1.3.1
        version: 1.3.1
      - name: ack-apigateway-controller.v1.3.0
        version: 1.3.0
      - name: ack-apigateway-controller.v1.2.5
        version: 1.2.5
      - name: ack-apigateway-controller.v1.2.4
        version: 1.2.4
      - name: ack-apigateway-controller.v1.2.3
        version: 1.2.3
      - name: ack-apigateway-controller.v1.2.2
        version: 1.2.2
      - name: ack-apigateway-controller.v1.2.1
        version: 1.2.1
      - name: ack-apigateway-controller.v1.1.0
        version: 1.1.0
      - name: ack-apigateway-controller.v1.0.6
        version: 1.0.6
      - name: ack-apigateway-controller.v1.0.5
        version: 1.0.5
      - name: ack-apigateway-controller.v1.0.4
        version: 1.0.4
      - name: ack-apigateway-controller.v1.0.2
        version: 1.0.2
      - name: ack-apigateway-controller.v1.0.1
        version: 1.0.1
      - name: ack-apigateway-controller.v1.0.0
        version: 1.0.0
      - name: ack-apigateway-controller.v0.0.3
        version: 0.0.3
      - name: ack-apigateway-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-apigateway-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nexus-repository-ha-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-repository-ha-operator-certified.v3.84.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusRepo",
                "metadata": {
                  "name": "nexusrepo-sample"
                },
                "spec": {
                  "ingress": {
                    "additionalRules": null,
                    "annotations": null,
                    "defaultRule": false,
                    "dockerIngress": {
                      "annotations": null,
                      "enabled": false,
                      "host": "example.com",
                      "name": "nexus-docker-ingress"
                    },
                    "dockerSubdomain": false,
                    "enabled": false,
                    "host": "example.com",
                    "name": "nexus-ingress",
                    "tls": {
                      "enabled": false,
                      "secretName": "tlsSecretName"
                    }
                  },
                  "license": {
                    "fileContentsBase64": "your_license_file_contents_in_base_64",
                    "secretName": "nexus-repo-license"
                  },
                  "secret": {
                    "nexusSecret": {
                      "enabled" : false,
                      "name": "nexus-secret.json",
                      "nexusSecretsKeyId": "super_secret_key_id",
                      "secretKeyfileContentsBase64": "secretKeyfileContentsBase64",
                      "mountPath": "/var/nexus-repo-secrets"
                    }
                  },
                  "nexusData": {
                    "pvc": {
                      "accessMode": "ReadWriteOnce",
                      "size": "2Gi"
                    },
                    "storageClass": {
                      "enabled": false,
                      "name": "nexusrepo-storage",
                      "parameters": {},
                      "provisioner": "provisioner",
                      "reclaimPolicy": "Retain",
                      "volumeBindingMode": "WaitForFirstConsumer"
                    },
                    "volumeClaimTemplate": {
                      "enabled": false
                    }
                  },
                  "service": {
                    "docker": {
                      "enabled": false,
                      "name": "nexus-repo-docker-service",
                      "port": 9090,
                      "protocol": "TCP",
                      "targetPort": 9090,
                      "type": "NodePort"
                    },
                    "nexus": {
                      "enabled": false,
                      "name": "nexus-repo-service",
                      "port": 80,
                      "protocol": "TCP",
                      "targetPort": 8081,
                      "type": "NodePort"
                    }
                  },
                  "statefulset": {
                    "container": {
                      "containerPort": 8081,
                      "env": {
                        "clustered": false,
                        "install4jAddVmParams": "-Xms2703m -Xmx2703m",
                        "jdbcUrl": null,
                        "nexusInitialPassword": "nexusabc",
                        "password": "nexus",
                        "user": "nexus",
                        "zeroDowntimeEnabled": false
                      },
                      "imageName": "registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:fbad488ff2e368845473a11b4955992fe826a94811da3abf5cc3c6a33ff3974f",
                      "pullPolicy": "IfNotPresent",
                      "resources": {
                        "limits": {
                          "cpu": 4,
                          "memory": "8Gi"
                        },
                        "requests": {
                          "cpu": 4,
                          "memory": "8Gi"
                        }
                      },
                      "terminationGracePeriod": 120
                    },
                    "imagePullSecrets": {},
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 240,
                      "path": "/",
                      "periodSeconds": 60
                    },
                    "name": "nexusrepo-statefulset",
                    "readinessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 240,
                      "path": "/",
                      "periodSeconds": 60
                    },
                    "replicaCount": 1
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-certified@sha256:99f30aa5ba8469d96de3efe69d1b21e73ff022172fc79144bba496d1119ec8a2
          createdAt: "2023-08-14"
          description: |-
            Nexus Repository is the central source of control to efficiently manage all binaries
            and build artifacts across your DevOps pipeline.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.replaces: 3.80.0-2
          olm.skipRange: <3.84.0-1
          repository: https://github.com/sonatype/operator-nexus-repository
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus Repository
            displayName: NexusRepo
            kind: NexusRepo
            name: nexusrepos.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Repository is the central source of control to efficiently manage all binaries
          and build artifacts across your DevOps pipeline.
          The flow of open source and third-party components into and through an organization
          creates a complex software supply chain.
          Nexus Repository delivers speed, efficiency, and quality to the governance
          and management of all dependencies, libraries, and applications for your DevOps teams.

          ## Core Capabilities

          * **Dependency Management**:
            Improves reliability with repeatable, fast access to secure dependencies
          * **Developer Productivity**:
            Streamline developer workflows by enabling the sharing of components and applications across teams
          * **Supply Chain Performance**:
            Improve speed-to-market and reduced build times with release advanced staging and component tagging
          * **CI/CD Integrations**:
            Increase DevOps scalability with integrations to the most popular build and deployment tools

          Version control systems and package registries do not scale when managing proprietary,
          open source, and third-party components.
          Organizations need a central binary and build artifact repository to manage dependencies
          across the entire software supply chain.

          ## Limitations

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus Repository Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage
          By default, the example Custom Resource starts up a single instance of Nexus Repository in non-Pro mode using an H2 database.

          ### Connect to a PostgreSQL database
          We strongly recommend using Nexus Repository with a PostgreSQL database. To connect a Nexus Repository instance to a PostgreSQL database, do as follows:
          * Update the `license.fileContentsBase64` parameter with a base 64 representation of your Nexus Repository Pro license.
          * Set the `statefulset.container.env.jdbcUrl` parameter to your database url.
          * Set the `statefulset.container.env.user` parameter to your database user name.
          * Set the `statefulset.container.env.password` parameter to your database password.

          ### High Availability (HA) mode
          To start Nexus Repository in HA mode, you must edit the Custom Resource YAML as follows:

          * Update the `license.fileContentsBase64` parameter with a base 64 representation of your Nexus Repository Pro license.
          * Set the `statefulset.container.env.jdbcUrl` parameter to your database url.
          * Set the `statefulset.container.env.user` parameter to your database user name.
          * Set the `statefulset.container.env.password` parameter to your database password.
          * Set the `statefulset.container.env.clustered` parameter to true.
          * Set the `statefulset.replicaCount` parameter to the number of Nexus Repository pods you require in your HA cluster
          * You must use a persistent volume as described in the 'Storage' section below.

          ### Storage
          The default configuration uses an emptyDir volume for storing Nexus Repository logs. However, we strongly recommend that
          you configure dynamic provisioning of persistent storage based on your deployment environment as explained below.

          #### Cloud deployments (AWS/Azure)
          * Ensure the appropriate Container Storage Interface (CSI) driver(s) are installed for your chosen OpenShift cloud deployment.
            Please, refer to OpenShift documentation for details on configuring CSI drivers.
          Note: if you're using Red Hat OpenShift on AWS (ROSA), CSI drivers for dynamically provisioning EBS volumes are installed
          by default; you should see associated storage classes for them in your cluster web console.

          #### On premise deployments
          1. Attach separate disks (i.e. separate from the root disk) to your worker nodes.
          2. Install the Local Storage Operator. Please refer to Open Shift persistent storage documentation:
             https://docs.openshift.com/container-platform/4.13/storage/persistent_storage/persistent_storage_local/persistent-storage-local.html
          and the Local Storage Operator documentation: https://github.com/openshift/local-storage-operator.
          3. Use the Local Storage Operator to automatically create persistent volumes for your chosen storage class name as documented here:
             https://docs.openshift.com/container-platform/4.13/storage/persistent_storage/persistent_storage_local/persistent-storage-local.html
          4. Specify the storage class name from step 3 as the value of the `nexusData.storageClass.name` property and also set `nexusData.volumeClaimTemplate.enabled` to true.

          #### Configuring the operator for dynamic persistent volume provisioning
          * Set the `nexusData.storageClass.name` parameter to a storage class name. This could be one of the default storage classes
            automatically created in your OpenShift cluster (e.g., if you're using ROSA) or one that you would like the operator to create.
          * If you would like to create a dedicated storage class (i.e., you don't want to use the default), then in addition to
            specifying a value for `nexusData.storageClass.name`, you must also set `nexusData.storageClass.enabled` parameter to `true`.
          * Set the `nexusData.volumeClaimTemplate.enabled` parameter to `true`.

          ### Additional configuration
          * Once the operator has created the server instance and it is running, you can further configure Nexus Repository via the NexusRepo custom resource:

          | Parameter                                                   | Description                         | Default                                 |
          | ----------------------------------------------------------- | ----------------------------------  | ----------------------------------------|
          | `statefulset.name`                                          |  The suffix for the statefulset's service name    | nexusrepo-statefulset     |
          | `statefulset.replicaCount`                                  |  The desired number of Nexus Repository pods    | 3                           |
          | `statefulset.container.imageName`                           |  The Nexus repository image registry url    |
          | | | registry.connect.redhat.com/sonatype/nexus-repository-manager:<app version>-ubi-1|
          | `statefulset.container.resources.requests.cpu`              |  The minimum cpu the Nexus repository pod can request    | 4 |
          | `statefulset.container.resources.requests.memory`           |  The minimum memory the Nexus repository pod can request    | 8Gi|
          | `statefulset.container.resources.limits.cpu`                |  The maximum cpu the Nexus repository pod may get.    | 4|
          | `statefulset.container.resources.limits.memory`             |  The maximum memory the Nexus repository pod may get.  | 8Gi|
          | `statefulset.container.containerPort`                       |  The Nexus Repository container's HTTP port  | 8081|
          | `statefulset.container.pullPolicy`                          |  The Nexus Repository docker image pull policy  | IfNotPresent|
          | `statefulset.container.terminationGracePeriod`              |  The time given for the pod to gracefully shut down  | 120 seconds|
          | `statefulset.container.env.clustered`                       |  Controls whether Nexus Repository should start in clustered mode  | False|
          | `statefulset.container.env.install4jAddVmParams`            |  Xmx and Xms settings for JVM  | -Xms2703m -Xmx2703m|
          | `statefulset.container.env.user`                            |  The database user name | nexus (can be overridden via Kubernetes secret. See above)|
          | `statefulset.container.env.password`                        |  The database password  | nexus (can be overridden via Kubernetes secret. See above)|
          | `statefulset.container.env.jdbcUrl`                         |  The database url |
          | | | jdbc:postgresql://your_database_server:5432/nexus (must be overridden via Kubernetes |
          | | | secret object which feeds into env variable variables defined in the operator.yaml . See above)|
          | `statefulset.container.env.nexusInitialPassword`            |  The initial admin password for Nexus Repository UI. This is for first time |
          | | installs only. Subsequent installs/restarts on the same node ignores this setting | nexusabc|
          | `statefulset.livenessProbe.initialDelaySeconds`             |  The initial delay before the Nexus Repository application is pinged for a |
          | | response i.e. to determine whether the application is still alive | 240|
          | `statefulset.livenessProbe.path`                            |  The path to ping the application | |
          | `statefulset.livenessProbe.periodSeconds`                   |  The interval for checking that the application is still alive i.e. responsive, not deadlocked etc | 60|
          | `statefulset.livenessProbe.failureThreshold`                |  The maximum number of times the application ping can fail before the |
          | | application is considered unresponsive. Kubernetes will attempt to restart the application's pod in this case | 6|
          | `statefulset.readinessProbe.initialDelaySeconds`            |  The initial delay before the Nexus Repository application is checked |
          | | whether it is ready to receive traffic | 240|
          | `statefulset.readinessProbe.path`                           |  The path where the Nexus Repository application is checked for traffic readiness | |
          | `statefulset.readinessProbe.periodSeconds`                  |  The interval for checking the Nexus Repository application for traffic readiness | 60|
          | `statefulset.readinessProbe.failureThreshold`               |  The maximum number of times to check the Nexus Repository application for traffic readiness | 6|
          | `statefulset.imagePullSecrets`                              |  The pull secret for private image registries | {}|
          | `ingress.name`                                              |  The name of an Ingress object which listens for requests to Nexus Repository on port 80 | nexus-ingress|
          | `ingress.enabled`                                           |  Whether or not to create the Ingress | false|
          | `ingress.host`                                              |  Ingress host | "example.com"|
          | `ingress.defaultRule`                                       |  Whether or not to add a default rule for the Nexus Repository |
          | | Ingress which forwards traffic to a Service object  | false|
          | `ingress.additionalRules`                                   |  Additional rules to add to the ingress | null|
          | `ingress.tls.enabled`                                       |  Whether or not TLS should be enabled for the Ingress | false|
          | `ingress.tls.secretName`                                    | The name of a Secret object in which to store the TLS secret | tlsSecretName|
          | `ingress.dockerSubdomain`                                   |  Whether or not to add rules for docker subdomains | false|
          | `ingress.annotations`                                       |  Annotations for the Ingress object | null |
          | `license.fileContentsBase64`                                |  Your Base64 encoded Nexus Repository license | your_license_file_contents_in_base_64 |
          | `license.secretName`                                        |  The name of a secret object in which to store your |
          | | Base64-encoded Nexus Repository license | nexus-repo-license|
          | `ingress.dockerIngress.name`                                |  The name of an Ingress object which listens for requests to an |
          | | existing docker repository in Nexus Repository | nexus-docker-ingress|
          | `ingress.dockerIngress.enabled`                             |  Whether or not to create an Ingress for a docker repository | false|
          | `ingress.dockerIngress.host`                                |  The docker Ingress's host | "example.com"|
          | `ingress.dockerIngress.annotations`                         |  Annotations for the docker Ingress object | null |
          | `nexusData.pvc.size`                                        |  The volume size to request for storing Nexus logs | 2Gi |
          | `nexusData.pvc.accessMode`                                  |  The persistent volume claim access mode | ReadWriteOnce |
          | `nexusData.storageClass.name`                               |  The name of the storage class to use for dynamically requesting that |
          | | storage be provisioned when using dynamic provisioning or to randomly select a pre-created persisted volume when using the Local Storage Operator. | nexusrepo-storage |
          | `nexusData.storageClass.reclaimPolicy`                      |  The reclaim policy for persistent volume's created by the storage class | Retain |
          | `nexusData.storageClass.enabled`                            |  Whether or not a storage class object should be created. | false |
          | `nexusData.storageClass.volumeBindingMode`                  |  It controls when volume binding and dynamic provisioning should occur | WaitForFirstConsumer |
          | `nexusData.storageClass.provisioner`                        |  The name of the volume provisioner | provisioner |
          | `nexusData.storageClass.parameters`                         |  The additional parameters for the storage class | {} |
          | `nexusData.volumeClaimTemplate.enabled`                           |  Whether or not volume should be requested using a claim template. |
          | | You should set this property to true when using dynamic volume provisioning (cloud deployments) |
          | | or if you're using Local Storage Operator to automatically create persistent volumes for locally attached disks (i.e. not root disk) | false |
          | `license.secretName | A name for the Kubernetes secret object which stores the base64 representation of your |
          | | Nexus Repository license specified via `license.fileContentsBase64` is stored | nexus-repo-license |
          | `license.fileContentsBase64 | A base64 representation of your Nexus Repository license | your_license_file_contents_in_base_64 |
          | `service.nexus.name`                                              |  The name of a kubernetes Service Object which allows access to |
          | | Nexus Repository's main HTTP port | nexus-repo-service |
          | `service.nexus.enabled`                                           |  Whether or not to create the Service object | false|
          | `service.nexus.type`                                              |  The type of the Kubernetes Service | "NodePort"|
          | `service.nexus.protocol`                                       |  The protocol  | TCP |
          | `service.nexus.port`                                   |  The port to listen for incoming requests | 80|
          | `service.nexus.targetPort`                                   |  The port to forward requests to | 8081 |
          | `service.docker.name`                                              |  The name of a kubernetes Service Object which allows access to a |
          | | docker repository within Nexus Repository | nexus-repo-docker-service |
          | `service.docker.enabled`                                           |  Whether or not to create a Kubernetes Service object for a given |
          | | docker repository within Nexus Repository | false |
          | `service.docker.type`                                              |  The type of the Kubernetes Service | "NodePort"|
          | `service.docker.protocol`                                       |  The protocol  | TCP |
          | `service.docker.port`                                   |  The port to listen for incoming requests | 9090 |
          | `service.docker.targetPort`                                   |  The port to forward requests to | 9090 |
        displayName: Nexus Repository HA Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - repository
        - sonatype
        links:
        - name: Nexus Repository Manager
          url: https://www.sonatype.com/product-nexus-repository
        maintainers:
        - email: support@sonatype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-certified@sha256:99f30aa5ba8469d96de3efe69d1b21e73ff022172fc79144bba496d1119ec8a2
        - registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:fbad488ff2e368845473a11b4955992fe826a94811da3abf5cc3c6a33ff3974f
        version: 3.84.0-1
      entries:
      - name: nexus-repository-ha-operator-certified.v3.84.0-1
        version: 3.84.0-1
      name: stable
    defaultChannel: stable
    packageName: nexus-repository-ha-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-kubernetes.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "ClusterMongoDBRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-enterprise"
                  },
                  "name": "clustermongodbrole-sample"
                },
                "spec": {
                  "db": "admin",
                  "role": "rootMonitor",
                  "roles": [
                    {
                      "db": "admin",
                      "role": "root"
                    },
                    {
                      "db": "admin",
                      "role": "clusterMonitor"
                    }
                  ]
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "my-replica-set"
                },
                "spec": {
                  "credentials": "my-credentials",
                  "members": 3,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "sample-sharded-cluster"
                },
                "spec": {
                  "configServerCount": 3,
                  "credentials": "my-credentials",
                  "mongodsPerShardCount": 3,
                  "mongosCount": 2,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "shardCount": 1,
                  "type": "ShardedCluster",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBMultiCluster",
                "metadata": {
                  "name": "multi-replica-set"
                },
                "spec": {
                  "clusterSpecList": [
                    {
                      "clusterName": "e2e.cluster1.mongokubernetes.com",
                      "members": 2
                    },
                    {
                      "clusterName": "e2e.cluster2.mongokubernetes.com",
                      "members": 1
                    },
                    {
                      "clusterName": "e2e.cluster3.mongokubernetes.com",
                      "members": 2
                    }
                  ],
                  "credentials": "my-credentials",
                  "duplicateServiceObjects": false,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBOpsManager",
                "metadata": {
                  "name": "ops-manager"
                },
                "spec": {
                  "adminCredentials": "ops-manager-admin",
                  "applicationDatabase": {
                    "members": 3,
                    "podSpec": {
                      "cpu": 1
                    }
                  },
                  "configuration": {
                    "mms.fromEmailAddr": "admin@thecompany.com"
                  },
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "version": "6.0.3"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBUser",
                "metadata": {
                  "name": "my-replica-set-x509-user"
                },
                "spec": {
                  "db": "$external",
                  "mongodbResourceRef": {
                    "name": "my-replica-set"
                  },
                  "roles": [
                    {
                      "db": "admin",
                      "name": "dbOwner"
                    }
                  ],
                  "username": "CN=my-replica-set-x509-user,OU=cloud,O=MongoDB,L=New York,ST=New York,C=US"
                },
                "status": {
                  "phase": "Updated"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/mongodb/mongodb-kubernetes@sha256:745cc2c6bb893c1cb1e75d79b6f3bfc7b125a6a62f7431bdc43cf6cacd373998
          createdAt: "2025-09-16T14:56:17Z"
          description: The MongoDB Controllers for Kubernetes enable easy deploys
            of MongoDB into Kubernetes clusters, using our management, monitoring
            and backup platforms, Ops Manager and Cloud Manager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mongodb/mongodb-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterMongoDBRole
            name: clustermongodbroles.mongodb.com
            version: v1
          - description: MongoDB Deployment
            displayName: MongoDB Deployment
            kind: MongoDB
            name: mongodb.mongodb.com
            version: v1
          - kind: MongoDBCommunity
            name: mongodbcommunity.mongodbcommunity.mongodb.com
            version: v1
          - description: MongoDB Multi Deployment
            displayName: MongoDB Multi Deployment
            kind: MongoDBMultiCluster
            name: mongodbmulticluster.mongodb.com
            version: v1
          - kind: MongoDBSearch
            name: mongodbsearch.mongodb.com
            version: v1
          - description: MongoDB x509 User
            displayName: MongoDB User
            kind: MongoDBUser
            name: mongodbusers.mongodb.com
            version: v1
          - description: MongoDB Ops Manager
            displayName: MongoDB Ops Manager
            kind: MongoDBOpsManager
            name: opsmanagers.mongodb.com
            version: v1
        description: |
          MongoDB Controllers for Kubernetes (MCK) is an open-source project. It represents a merge of the previous
          MongoDB Community Operator (MCO) and MongoDB Enterprise Kubernetes Operator (MEKO).

          This release brings MongoDB Community and Enterprise editions together under a single, unified operator,
          making it easier to manage, scale, and upgrade your deployments. While the first version simply brings
          the capabilities of both into a single Operator, future changes will build on this to more closely align
          how Community and Enterprise are managed in Kubernetes, to offer an even more seamless and streamlined experience.
          As an open-source project, it now allows for community contributions, helping drive quicker bug fixes and ongoing innovation.
          The MongoDB Controllers for Kubernetes enable easy deploys of MongoDB
          into Kubernetes clusters, using our management, monitoring and backup
          platforms, Ops Manager and Cloud Manager.

          ## Before You Start

          To start using the operator you''ll need an account in MongoDB Cloud Manager or
          a MongoDB Ops Manager deployment.

          * [Create a Secret with your OpsManager API key](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-operator-credentials/#procedure)

          * [Create a ConfigMap with your OpsManager project ID and URL](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-project-using-configmap/)

          By installing this integration, you will be able to deploy MongoDB instances
          with a single simple command.

          ## Required Parameters

          * `opsManager` or `cloudManager` - Enter the name of the ConfigMap containing project information
          * `credentials` - Enter the name of the Secret containing your OpsManager credentials
          * `type` - Enter MongoDB Deployment Types ("Standalone", "ReplicaSet", "ShardedCluster"

          ## Supported MongoDB Deployment Types ##

          * Standalone: An instance of mongod that is running as a single server and
          not as part of a replica set, this is, it does not do any kind of
          replication.

          * Replica Set: A replica set in MongoDB is a group of mongod processes that
          maintain the same data set. Replica sets provide redundancy and high
          availability, and are the basis for all production deployments. This section
          introduces replication in MongoDB as well as the components and architecture
          of replica sets. The section also provides tutorials for common tasks
          related to replica sets.

          * Sharded Cluster: The set of nodes comprising a sharded MongoDB deployment.
          A sharded cluster consists of config servers, shards, and one or more mongos
          routing processes. Sharding is a A database architecture that partitions
          data by key ranges and distributes the data among two or more database
          instances. Sharding enables horizontal scaling.

          ## Requirements for deploying MongoDB OpsManager

          * In order to deploy resources of type MongoDB OpsManager, you will need to
          create a secret containing the [credentials](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/plan-om-resource/#om-rsrc-prereqs)
          for the Global Owner user

          ## Security ##

          The operator can enable TLS for all traffic between servers and also between
          clients and servers. Before enabling `security.tls.enabled` to `true` you
          should create your certificates.  or you can leave the operator to create all
          the certificates for you. The operator ability to create certs is been
          deprecated due to Kubernetes API changes.

          For more information, please read the official MongoDB
          Kubernetes Operator  [docs](https://docs.mongodb.com/kubernetes-operator/stable/).
        displayName: MongoDB Controllers for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mongodb
        - databases
        - nosql
        links:
        - name: Documentation
          url: https://docs.opsmanager.mongodb.com/current/tutorial/install-k8s-operator/index.html
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: stable
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-kubernetes-init-database@sha256:3ea9106c0e35c015bc8f15657c76d974bb1cadb9220dc58be2076c9699f26fb7
        - quay.io/mongodb/mongodb-agent@sha256:6e826bbe7dd812c101464e75295370c4b10e4e9c31477db89cfceb96759853e9
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:702f529cb9d8d9042e54263f8bc37b40c062ac2695efc3e7150819779886e323
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c48631742ad1fec8a9ee745fc3cc9ca1fe7d86ef7fa68ad8e8337c7a4566dc87
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7e5361c2a1604dec3fc671d629806fe6ddc6df524b69a441b86933a4b30093
        - quay.io/mongodb/mongodb-enterprise-server@sha256:85c1d669c073caf9cbaac815e4c8ea2ca18b71deefa03bf84cd6d8b8ebd53e12
        - quay.io/mongodb/mongodb-agent@sha256:fa30370a75e8e4ac648650ceb39a7a259a754aab8a48d7e2f2801cd1b777e91b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f1e72679b2468aed2f3c3106568327911e6594c9ee15f7348e953eaa0d8fd44a
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:85b4e954c9c4bbe1019774bad1e05e9c8dc175648b01d3badf0b64143f182019
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d323c1bcd14118cc537ac282996ea5a58756d20b6b73f1eec0efa8710d246077
        - quay.io/mongodb/mongodb-enterprise-server@sha256:67238129df09d33ee4746ff1acb5515d699a29d9f42350eae3f3f69ccb7fbf54
        - quay.io/mongodb/mongodb-kubernetes-database@sha256:f8c981a6223f70d7c6bbd30c503fc8b2497f351830b2b74c30e6f81863d78d2f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:a4590af027182305913f0de1a2e37ad8612bd73b8b31f4df11068b80089021ff
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3add8b68bfbaa42596fe5dbbee4623ba5ff1a826b18bb71301525adf574dcb94
        - quay.io/mongodb/mongodb-agent@sha256:ee9f3906d893004bb2aaaa6412646f58288b0684502b9e90f33cafadcfce4349
        - quay.io/mongodb/mongodb-enterprise-server@sha256:df6081325ea2fe01d1b27d0c70005dec8232aaf51a43d91661b722b8a1761263
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:146c4f5dc443f94f54d72808b4e4df93bb904e23589b222d17700166780776b4
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:09788976bdc1327acbf18cedf8efd5fca2159e43b506faa40e03ee42ad94928a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:655896bac9e82de8e9729d59a36c3a31f66235f22b93449260a2ccb7709de084
        - quay.io/mongodb/mongodb-enterprise-server@sha256:172e338a43df01ee5408d8d1e3cb038d2c55e77761c1a2ac6ed37959d1525140
        - quay.io/mongodb/mongodb-agent@sha256:3fa5bcfda3228e0ea880434a60ced829e3872e74763b347636c18475d866d0e1
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:33bbd29b74f95baad522b380d70bcc9f933a3fee57642c37b09376c86179c40f
        - quay.io/mongodb/mongodb-kubernetes@sha256:745cc2c6bb893c1cb1e75d79b6f3bfc7b125a6a62f7431bdc43cf6cacd373998
        - quay.io/mongodb/mongodb-enterprise-server@sha256:19925e1297ab652db59c35f53e7b7c8b2b277f8ecdc4215682961aaf2b3e924e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:f71b027fccf649f18f63f8df7a59b1bac421a2cdab6712e0c985a93c08ed970a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8445d59d491a6d3890ea9c8b421214bcd0fdcfbc420ed40497a790b79f0bc89e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:895dbdfb87ddc2e9693671c7108a31ab28e16fec2e5be6d046bdd2e30319eade
        - quay.io/mongodb/mongodb-enterprise-server@sha256:dfd4277f0c16ed476ac7de51894739e8c8d13a65ea5024b9a97b3dce0f6c6975
        - quay.io/mongodb/mongodb-enterprise-server@sha256:fb8397cbcdbf7c33f6ad142492792771127f0d9a8c46388f350b6927405ee590
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7885ab6e88214f9289d8babf5189136dde0dd4c84a15a5481ccd3ef512b2e66f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:5d299820bb903f0cdfdf2626a3c44099d2eaee16188cd8234d2ff3a57259b3a7
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:439d6fefb469613432e2d05149dfc96c32b16469d743835d5f402a2ada8be8cc
        - quay.io/mongodb/mongodb-agent@sha256:da77dcb60b927bf6059c2c9bf008f8c3c0eec45cf7d1b40c34a0c263e8cc18b8
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b8aa1edbc5fb02af312a863bc7db8535a282dfd692a07b1e321e5a1acbfeb6a1
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9a3a6496be0d2d446e6d654eb56386c7726e0cd2d7232f6dcec579923edf0014
        - quay.io/mongodb/mongodb-agent@sha256:89819704d70a3b6b4d2358745a22060ef740b33ad5c6debf8e6ac4d4a6fd63ba
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b7764c316eeacffd3f9ddae62d9fd2a74eb58484d2cfd5021e460095d604764f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:e7aeedb752dda77cf0e49df3430843b5a2bba64c8404628a983f46640d316808
        - quay.io/mongodb/mongodb-search@sha256:7149a7b6f1ee84646d3972837920f80fc5db49503c598406ce92e2f794bd5f0d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa8683500e8d3632c4588bbcdddb4e3186a01f48aefe1b6d95f8860dae8b6c76
        - quay.io/mongodb/mongodb-enterprise-server@sha256:15e31285c8949a4666f7702df53416b0905d76945a28575a6028a876cc8a6f0f
        - quay.io/mongodb/mongodb-kubernetes-init-ops-manager@sha256:93c4633cbc84e9bfa3563771ac2efbf5e6515f380dec2c085494cb8326546377
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5c8553e5ba115657fb878c0fe08704af9fd5decec572a87d89d2c8c156ffc9f9
        - quay.io/mongodb/mongodb-agent@sha256:5b064c470335df840b050b714b28b75b6061444843250f1eb9215cb2b8c1a149
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a86c4d8a0b9ca3c20cc524809676eb03a944ada43378ec98c107cd72cc64700f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:5e3982493e85ac1e40536758c3145b7c2fff9a7b4be9d053ee41201363a89920
        - quay.io/mongodb/mongodb-agent@sha256:5f5031e4c2a3729f618d026b0a7f34fd4509ea73d5a4e3068b551041dc291583
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:d81eac2129b6ab8346be1749c56ea67f6da64d4ee0fb9f3ff4171ba0f61918eb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3316015ac0699d66ac73f714e5dc0784738654e0349e80b1e1f4e953b3c88964
        - quay.io/mongodb/mongodb-agent@sha256:b387b7458f5640f907c9074fcb08ae4d4785be31d4bb65f2d7c459cd6762b2d6
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:adeab0dc766a27138db7df5f37cbb1da860c14179ef631b2e001e5bb875325e4
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ef732e3eea91cc86caf2eabedc0c9115ed235a8d204f33e74bbf090a0b0259fb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a48a4d5c275fd669a27469649fb39ea38c728dbcf9f0a9da4e7cf27a2ce912ed
        - quay.io/mongodb/mongodb-enterprise-server@sha256:720056b985c1a16d48b2aecb7a88a15cba3d8186e55b3c6c24729699a85c8232
        - quay.io/mongodb/mongodb-enterprise-server@sha256:523566f44cbb6bbf0cd12b1adae939a883f2254c78dc4b8c9d0098581337b1db
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f2d9a3264a8814f33143b73c56624572fb80afeab184133d3a901713a31c1a87
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5145c97d2281aabcd8538ff515c3833aa48acb66b0b739fb1881120c1985d88a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9bef77e905dadd1e61d0439dd53b21ed1f3606e49affd9ffc4ba8156578e8f0f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:cd869ff7d661059fad7593e30141d74e8cc751295b1579577f7a853396eb5d0b
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:fa4aa97767bd19cf3711a74bf21ecb9cbbc25fed6b6b09a2eb241ebd3fbc8f3a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d49417ea85ee1ddb84429552a5aac87a82c2843904d8c0300a43b657b7a49e0e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:33a30d11969eb21952681108092fe7f916e25c8e1de43e6a4e40e126025a2bec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:778f9e29e36d6687cd17dd95ec1f8127d2a732bde1a2dd5821212eb55c1c998a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1d418c65d82ccad205b7889be4ba161d3cbfa8b1fb75b21c8fcf02de38fad9ca
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b9ab7147b59b62e40a69c6400f87d0c7faecfa5462bed6e68ca653b00e50b887
        - quay.io/mongodb/mongodb-agent@sha256:6394391f78da5e3218bb2c694e7ec9f6682fd17a6dd7e7250b0c0b5e1729796d
        - quay.io/mongodb/mongodb-agent@sha256:a8358ce465adf463d8e73916d7944765d7aaf2b0ed10bd3b8119a87fc9080667
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:78f11c9cd5e63f2b0e10c67dc1491f4ce1279fa90178c108814725414defab9a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c752a7f0726d21d55f204a359b85fa29cd727a5fe571b37289ae3ea52d6206b0
        - quay.io/mongodb/mongodb-agent@sha256:b422829aac63f068aed4a6948fd14cbd3c3bfad6d47ab76523f6839bf20b0203
        - quay.io/mongodb/mongodb-agent@sha256:6c23febaa9dd52bc29ac1e52d087fc4b2b47c578f4d6ac3a1674bb750a38cb07
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:3bbd3d7fd45d107621e7770f2b4c2a758e148cd472e0ee21719442f71bf6ade0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0fde236fe26d5b20210d6f246aadacd7845874712def516cbe2eec11d79d5181
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d2d8b59c1f39c41951e7878963f8496eb494688720366f8edc7060887cdbe6a0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7a93a0276531ff9be4c90bb8fe8d104e0a9e930c29792aafe03cc6a76a9fa89c
        - quay.io/mongodb/mongodb-agent@sha256:6cc4e953df3f09d793f0dbc6e57e55a3c5277ef63177a6d8f691a41227facf64
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3fe62f3ad52707ccf0e9993c69891e44c569a4b4e1a3244cf60e3b5fdde57bd2
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1a77851485ad3b2566e765b4bb051ff0ddfa48703492cef6c6db021dde84d87f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:bfbcfcfc5f0de22e20520b2d96f826638e424adb028a909ccc192013d11f5833
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b00ac773623e1bf422b47e6d783263d2d08737865abe9abbe106d86d1ea25165
        - quay.io/mongodb/mongodb-enterprise-server@sha256:37af0913742b6de339c7e664b30f35dcef04506529e4a09a9bbbdd57643f7149
        - quay.io/mongodb/mongodb-agent@sha256:d8a2ef79d767f03162ec57fe028f9072294b292d376af8eb97d2967eb6d7047b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7266b405ebb2dd24a9f7c12d3921024f547c3d8c01831e6bcfc1d551d63be1
        - quay.io/mongodb/mongodb-agent@sha256:af3de4f631d0b00cae7f7e80c3900378ba7e449e01830bf9dc643b873d2e6d33
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa300abf739e26350a258f643fa8e57aedacd041055c7b923bd22b21ef554cdb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d21c65780cadc9d8668d2c28935d492625d6993f13ba562722200a819281cbcc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:376ad7e5fe42651d946cba5705c0eee44aa636c1b4ff258210e4ce2fe4205b4d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:454b1585c342375dea643b10ead688d0f62347841f5f30059fcd4f92e3297ee9
        - quay.io/mongodb/mongodb-enterprise-server@sha256:11ab324698f919f84a5dab5a964d93fe63f98a48146bb3b86974cba7606d30ec
        - quay.io/mongodb/mongodb-kubernetes-init-appdb@sha256:afb0d463304fc0bfb3d8dade91459414daac6bf6fcf742347a2012cbc9fea354
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c8aec83b76dfecc05b8f5d8f40fd61b24980c0287a07294f75f321c2d42c32b4
        - quay.io/mongodb/mongodb-agent@sha256:2d5652d0e263aa94554e7a068c93e4fd94b84ec223f077361d030d172bc93f67
        version: 1.4.0
      entries:
      - name: mongodb-kubernetes.v1.4.0
        version: 1.4.0
      - name: mongodb-kubernetes.v1.3.0
        version: 1.3.0
      - name: mongodb-kubernetes.v1.2.0
        version: 1.2.0
      - name: mongodb-kubernetes.v1.1.0
        version: 1.1.0
      - name: mongodb-kubernetes.v1.0.1
        version: 1.0.1
      - name: mongodb-kubernetes.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: namespace-configuration-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: namespace-configuration-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupConfig",
                "metadata": {
                  "name": "test-groupconfig"
                },
                "spec": {
                  "labelSelector": {
                    "matchLabels": {
                      "type": "team"
                    }
                  },
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-dev\n  labels:\n    group: {{ .Name }}  \n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-qa\n  labels:\n    group: {{ .Name }}          \n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-prod\n  labels:\n    group: {{ .Name }}                           \n"
                    },
                    {
                      "objectTemplate": "apiVersion: quota.openshift.io/v1\nkind: ClusterResourceQuota\nmetadata:\n  name: {{ .Name }}-quota\nspec:\n  quota:\n    hard:\n      pods: \"4\" \n      requests.cpu: \"1\" \n      requests.memory: 1Gi \n      requests.ephemeral-storage: 2Gi \n      limits.cpu: \"2\" \n      limits.memory: 2Gi \n      limits.ephemeral-storage: 4Gi\n  selector:\n    labels:\n      matchLabels:\n        group: {{ .Name }}\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NamespaceConfig",
                "metadata": {
                  "name": "test-namespaceconfig"
                },
                "spec": {
                  "labelSelector": {
                    "matchLabels": {
                      "type": "secure"
                    }
                  },
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: ResourceQuota\nmetadata:\n  name: standard-sandbox\n  namespace: {{ .Name }}\nspec:\n  hard:\n    pods: \"4\" \n    requests.cpu: \"1\" \n    requests.memory: 1Gi \n    requests.ephemeral-storage: 2Gi \n    limits.cpu: \"2\" \n    limits.memory: 2Gi \n    limits.ephemeral-storage: 4Gi\n"
                    },
                    {
                      "objectTemplate": "kind: EgressNetworkPolicy\napiVersion: network.openshift.io/v1\nmetadata:\n  name: air-gapped\n  namespace: {{ .Name }}  \nspec:\n  egress:\n  - type: Deny\n    to:\n      cidrSelector: 0.0.0.0/0\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "UserConfig",
                "metadata": {
                  "name": "test-user-config"
                },
                "spec": {
                  "providerName": "my-provider",
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-sandbox\n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: ResourceQuota\nmetadata:\n  name: standard-sandbox\n  namespace: {{ .Name }}-sandbox\nspec:\n  hard:\n    pods: \"4\" \n    requests.cpu: \"1\" \n    requests.memory: 1Gi \n    requests.ephemeral-storage: 2Gi \n    limits.cpu: \"2\" \n    limits.memory: 2Gi \n    limits.ephemeral-storage: 4Gi\n"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/namespace-configuration-operator@sha256:49ed7d6155342adaa2b12fd80c6761c3081d8e6149d187cb7ff91a247cdf2e7a
          createdAt: "2023-11-22T15:58:50Z"
          description: This operator provides a facility to define and enforce namespace
            configurations
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: namespace-configuration-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/namespace-configuration-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GroupConfig is the Schema for the groupconfigs API
            displayName: Group Config
            kind: GroupConfig
            name: groupconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: NamespaceConfig is the Schema for the namespaceconfigs API
            displayName: Namespace Config
            kind: NamespaceConfig
            name: namespaceconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: UserConfig is the Schema for the userconfigs API
            displayName: User Config
            kind: UserConfig
            name: userconfigs.redhatcop.redhat.io
            version: v1alpha1
        description: "The `namespace-configuration-operator` helps keeping configurations
          related to Users, Groups and Namespaces aligned with one of more policies
          specified as a CRs. The purpose is to provide the foundational building
          block to create an end-to-end onboarding process. \nBy onboarding process
          we mean all the provisioning steps needed to a developer team working on
          one or more applications to OpenShift.\n This usually involves configuring
          resources such as: Groups, RoleBindings, Namespaces, ResourceQuotas, NetworkPolicies,
          EgressNetworkPolicies, etc.... . Depending on the specific environment the
          list could continue.\nNaturally such a process should be as automatic and
          scalable as possible.\n\nWith the namespace-configuration-operator one can
          create rules that will react to the creation of Users, Groups and Namespace
          and will create and enforce a set of resources.\n\nHere are some examples
          of the type of onboarding processes that one could support:\n\n1. [developer
          sandbox](https://github.com/redhat-cop/namespace-configuration-operator/blob/master/examples/user-sandbox/readme.md)\n
          2. [team onboarding](https://github.com/redhat-cop/namespace-configuration-operator//blob/master/examples/team-onboarding/readme.md)
          with support of the entire SDLC in a multitentant environment.\n\nPolicies
          can be expressed with the following CRDs:\n\n| Watched Resource | CRD |\n|--|--|\n
          | Groups | [GroupConfig](#GroupConfig) |\n| Users | [UserConfig](#UserConfig)
          |\n| Namespace | [NamespaceConfig](#NamespaceConfig) |\n\nThese CRDs all
          share some commonalities:\n\n1. Templated Resources\n2. List of ignored
          jason path\n \n### Templated Resources\n\nEach has a parameter called `templatedResources`,
          which is an array. Each element of the array has two fields `objectTemplate`
          and `excludedPaths` (see below).\n\nThe `objectTemplate` field must contains
          a [go template](https://golang.org/pkg/text/template/) that resolves to
          a single API Resource expressed in `yaml`. The template is merged with the
          object selected by the CR. For example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: UserConfig\nmetadata:\n  name: test-user-config\nspec:\n  provider:
          corp-ldap\n  templates:\n  - objectTemplate: |\n      apiVersion: v1\n      kind:
          Namespace\n      metadata:\n        name: {{ .Name }}-sandbox\n```\n\nThis
          create a rule in which every time a user from the `corp-ldap` provider is
          created, a namespace called `<username>-sandbox` is also created.\n\n###
          Excluded Paths\n\nThe logic of the `namespace-configuration-operator` is
          to enforce that the resources resolved by processing the templates \"stays
          in place\". In other words if those resources are changed and/or deleted
          they will be reset by the operator.\nBut there are situations in which at
          least part of a resource is allowed to change. Common use cases are: annotations
          and in general the metadata section of a resource can be updated by the
          various operators watching that resources. The status field is often updated
          by the main operator managing that resources. Finally, when applicable the
          `spec.replicas` field should also be allowed to change.\n\nTo handle special
          use case, one can also specify additional *jsonpaths* that should be ignored
          when comparing the desired resource and the current resource and making
          a decision on whether that resource should be reset.\n\nThe following paths
          are always included:\n \n1. `.metadata`\n2. `.status`\n3. `.spec.replicas`\n\n##
          NamespaceConfig\n\n The `NamespaceConfig` CR allows specifying one or more
          objects that will be created in the selected namespaces.\n\nNamespaces can
          be selected by labels or annotations via a label selector for example:\n\n```yaml\napiVersion:
          redhatcop.redhat.io/v1alpha1\n kind: NamespaceConfig\nmetadata:\n  name:
          small-namespace\n  selector:\n    matchLabels:\n      size: small  \n  resources:\n
          \ - apiVersion: v1\n    kind: ResourceQuota\n    metadata:\n      name:
          small-size  \n    spec:\n      hard:\n        requests.cpu: \"4\"\n        requests.memory:
          \"2Gi\"\n```\n\nHere is a `NamespaceConfig` object using a `matchExpressions`
          selector:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n kind: NamespaceConfig\nmetadata:\n
          \ name: tier-config\nspec:\n  annotationSelector:\n    matchExpressions:\n
          \     - {key: tier, operator: In, values: [gold,silver]}\n ```\n\nAlthough
          not enforced by the operator the general expectation is that the NamespaceConfig
          CR will be used to create objects inside the selected namespace.\n \nExamples
          of NamespaceConfig usages can be found [here](https://github.com/redhat-cop/namespace-configuration-operator/blob/master/examples/namespace-config/readme.md)\n
          \n## GroupConfig\n\nThe `GroupConfig` CR allows specifying one or more objects
          that will be created in the selected Group.\nGroups can be selected by labels
          or annotations via a label selector, similarly to the `NamespaceConfig`.\n\nOften
          groups are created in OpenShift by a job that synchronizes an Identity Provider
          with OCP. So the idea is that when new groups are added or deleted the configuration
          in OpenShift will adapt automatically.\n\nAlthough not enforced by the operator,
          GroupConfig are expected to create cluster-scoped resources like Namespaces,
          ClusterResourceQuotas and potentially some namespaced resources like RoleBindings.\n\n##
          UserConfig\n \nIn OpenShift an external user is defined by two entities:
          Users and Identities. There is a relationship of on to many between Users
          and Identities. Given one user, there can be one Identity per authentication
          mechanism.\n\nThe `UserConfig` CR allows specifying one or more objects
          that will be created in the selected User.\nUsers can be selected by label
          or annotation like `NamespaceConfig` and `UserConfig`.\nUSers can also be
          selected by provider name (the name of the authentication mechanism) and
          identity extra field.\n\nHere is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          UserConfig\nmetadata:\n  name: test-user-config\nspec:\n  providerName:
          okta-provider\n  identityExtraFieldSelector:\n    matchLabels:\n      sandbox_enabled:
          \"true\"\n  templates:\n  - objectTemplate: |\n      apiVersion: v1\n      kind:
          Namespace\n      metadata:\n        name: {{ .Name }}-sandbox\n ```\n\nUser
          will be selected by this `UserConfig` only if they login via the *okta-provider*
          and if the extra field was populate with the label `sandbox_enabled: \"true\"
          `. Note that not all authentication provider allow populating the extra
          fields in the Identity object.\n"
        displayName: Namespace Configuration Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - namespace
        - user
        - group
        - configuration
        - policy
        - management
        links:
        - name: Namespace Configuration Operator
          url: https://github.com/redhat-cop/namespace-configuration-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/namespace-configuration-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/namespace-configuration-operator@sha256:49ed7d6155342adaa2b12fd80c6761c3081d8e6149d187cb7ff91a247cdf2e7a
        version: 1.2.6
      entries:
      - name: namespace-configuration-operator.v1.2.6
        version: 1.2.6
      - name: namespace-configuration-operator.v1.2.5
        version: 1.2.5
      - name: namespace-configuration-operator.v1.2.4
        version: 1.2.4
      - name: namespace-configuration-operator.v1.2.3
        version: 1.2.3
      - name: namespace-configuration-operator.v1.2.2
        version: 1.2.2
      - name: namespace-configuration-operator.v1.2.1
        version: 1.2.1
      - name: namespace-configuration-operator.v1.2.0
        version: 1.2.0
      - name: namespace-configuration-operator.v1.1.0
        version: 1.1.0
      - name: namespace-configuration-operator.v1.0.7
        version: 1.0.7
      - name: namespace-configuration-operator.v1.0.6
        version: 1.0.6
      - name: namespace-configuration-operator.v1.0.5
        version: 1.0.5
      - name: namespace-configuration-operator.v1.0.4
        version: 1.0.4
      - name: namespace-configuration-operator.v1.0.3
        version: 1.0.3
      - name: namespace-configuration-operator.v1.0.2
        version: 1.0.2
      - name: namespace-configuration-operator.v1.0.1
        version: 1.0.1
      - name: namespace-configuration-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: namespace-configuration-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: geunwookim
      provider-url: ""
    name: ncn-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ncn-operator.v0.3.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-1"
                },
                "spec": null
              },
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "label-1"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
          - description: Label is the Schema for the labels API
            displayName: Label
            kind: Label
            name: labels.cxl.antline.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: jwahn@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Antline
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/rasen708/ncn-operator:v0.4.0
        version: 0.3.18
      entries:
      - name: ncn-operator.v0.3.18
        version: 0.3.18
      - name: ncn-operator.v0.3.14
        version: 0.3.14
      name: alpha
    - currentCSV: ncn-operator.v0.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-1"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-19T01:21:08Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: jwahn@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        provider:
          name: Antline
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/rasen708/ncn-operator:v0.2.2
        version: 0.2.2
      entries:
      - name: ncn-operator.v0.2.2
        version: 0.2.2
      - name: ncn-operator.v0.2.1
        version: 0.2.1
      - name: ncn-operator.v0.2.0
        version: 0.2.0
      name: beta
    - currentCSV: ncn-operator.v0.9.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cmmd.gw.com/v1",
                "kind": "CMMD",
                "metadata": {
                  "name": "my-app",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "allocateMode": "auto",
                  "enable": true,
                  "payload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "metadata": {
                      "name": "my-app"
                    },
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "name": "my-app"
                        }
                      },
                      "template": {
                        "metadata": {
                          "labels": {
                            "name": "my-app"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "quay.io/my-account/my-app:latest",
                              "name": "my-app-01"
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "cmmd.gw.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "machineconfiguration.openshift.io/role": "cmmd"
                  },
                  "name": "cmmd-node-label",
                  "namespace": "socmmd"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          console.openshift.io/plugins: '["socmmd-console-plugin"]'
          containerImage: quay.io/rasen708/ncn-operator@sha256:a09963745bd9c1419532d42d91237b7496c7a78e6a1c460a8e714ea1a0ce4d45
          createdAt: "2024-08-01T05:45:38Z"
          description: Provides the ncn extend
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "cmmd.gw.com/v1",
              "kind": "Label",
              "metadata": {
                "name": "cmmd-node-label",
                "namespace": "socmmd",
                "labels": {}
              },
              "spec": {}
            }
          operatorframework.io/suggested-namespace: socmmd-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/geunwookim/ncn-operator
          support: geunwookim
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CMMD is the Schema for the cmmds API
            displayName: CMMD
            kind: CMMD
            name: cmmds.cmmd.gw.com
            version: v1
          - description: Label is the Schema for the labels API
            displayName: Node Label
            kind: Label
            name: labels.cmmd.gw.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: geunwookim@gmail.com
          name: geunwookim
        maturity: betav1
        provider:
          name: geunwookim
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/rasen708/ncn-operator@sha256:a09963745bd9c1419532d42d91237b7496c7a78e6a1c460a8e714ea1a0ce4d45
        version: 0.9.8
      entries:
      - name: ncn-operator.v0.9.8
        version: 0.9.8
      - name: ncn-operator.v0.9.2
        version: 0.9.2
      - name: ncn-operator.v0.8.1
        version: 0.8.1
      - name: ncn-operator.v0.5.6
        version: 0.5.6
      - name: ncn-operator.v0.4.3
        version: 0.4.3
      - name: ncn-operator.v0.4.2
        version: 0.4.2
      - name: ncn-operator.v0.3.18
        version: 0.3.18
      - name: ncn-operator.v0.3.14
        version: 0.3.14
      name: betav1
    defaultChannel: betav1
    packageName: ncn-operator
    provider:
      name: geunwookim
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CNCF
      provider-url: ""
    name: etcd
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: etcdoperator.v0.6.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          description: etcd is a distributed key value store providing a reliable
            way to store data across a cluster of machines.
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
        description: |
          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader. Your applications can read and write data into etcd.
          A simple use-case is to store database connection details or feature flags within etcd as key value pairs. These values can be watched, allowing your app to reconfigure itself when they change. Advanced uses take advantage of the consistency guarantees to implement database leader elections or do distributed locking across a cluster of workers.

          _The etcd Open Cloud Service is Public Alpha. The goal before Beta is to fully implement backup features._

          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` or with the API using the automatically generated Kubernetes Service.

          [Read the complete guide to using the etcd Open Cloud Service](https://coreos.com/tectonic/docs/latest/alm/etcd-ocs.html)

          ### Supported Features
          **High availability**
          Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.
          **Automated updates**
          Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.
          **Backups included**
          Coming soon, the ability to schedule backups to happen on or off cluster.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: support@coreos.com
          name: CoreOS, Inc
        maturity: alpha
        provider:
          name: CoreOS, Inc
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:bd944a211eaf8f31da5e6d69e8541e7cada8f16a9f7a5a570b22478997819943
        version: 0.6.1
      entries:
      - name: etcdoperator.v0.6.1
        version: 0.6.1
      name: alpha
    - currentCSV: etcdoperator.v0.9.4-clusterwide
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdCluster",
                "metadata": {
                  "name": "example",
                  "annotations": {
                    "etcd.database.coreos.com/scope": "clusterwide"
                  }
                },
                "spec": {
                  "size": 3,
                  "version": "3.2.13"
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdRestore",
                "metadata": {
                  "name": "example-etcd-cluster-restore"
                },
                "spec": {
                  "etcdCluster": {
                    "name": "example-etcd-cluster"
                  },
                  "backupStorageType": "S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdBackup",
                "metadata": {
                  "name": "example-etcd-cluster-backup"
                },
                "spec": {
                  "etcdEndpoints": ["<etcd-cluster-endpoints>"],
                  "storageType":"S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
          createdAt: "2019-02-28 01:03:00"
          description: Create and maintain highly-available etcd clusters on Kubernetes
          repository: https://github.com/coreos/etcd-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to backup an etcd cluster.
            displayName: etcd Backup
            kind: EtcdBackup
            name: etcdbackups.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to restore an etcd cluster from a backup.
            displayName: etcd Restore
            kind: EtcdRestore
            name: etcdrestores.etcd.database.coreos.com
            version: v1beta2
        description: |
          The etcd Operater creates and maintains highly-available etcd clusters on Kubernetes, allowing engineers to easily deploy and manage etcd clusters for their applications.

          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader.


          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` via port forwarding:

              $ kubectl --namespace default port-forward service/example-client 2379:2379
              $ etcdctl --endpoints http://127.0.0.1:2379 get /

          Or directly to the API using the automatically generated Kubernetes Service:

              $ etcdctl --endpoints http://example-client.default.svc:2379 get /

          Be sure to secure your etcd cluster (see Common Configurations) before exposing it outside of the namespace or cluster.


          ### Supported Features

          * **High availability** - Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.

          * **Automated updates** - Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.

          * **Backups included** - Create etcd backups and restore them through the etcd Operator.

          ### Common Configurations

          * **Configure TLS** - Specify [static TLS certs](https://github.com/coreos/etcd-operator/blob/master/doc/user/cluster_tls.md) as Kubernetes secrets.

          * **Set Node Selector and Affinity** - [Spread your etcd Pods](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-node-selector-and-anti-affinity-across-nodes) across Nodes and availability zones.

          * **Set Resource Limits** - [Set the Kubernetes limit and request](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-resource-requirement) values for your etcd Pods.

          * **Customize Storage** - [Set a custom StorageClass](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#custom-persistentvolumeclaim-definition) that you would like to use.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: etcd-dev@googlegroups.com
          name: etcd Community
        maturity: alpha
        provider:
          name: CNCF
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
        version: 0.9.4-clusterwide
      entries:
      - name: etcdoperator.v0.9.4-clusterwide
        version: 0.9.4-clusterwide
      - name: etcdoperator.v0.9.2-clusterwide
        version: 0.9.2-clusterwide
      - name: etcdoperator.v0.9.0
        version: 0.9.0
      name: clusterwide-alpha
    - currentCSV: etcdoperator.v0.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "size": 3,
                  "version": "3.2.13"
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdRestore",
                "metadata": {
                  "name": "example-etcd-cluster-restore"
                },
                "spec": {
                  "etcdCluster": {
                    "name": "example-etcd-cluster"
                  },
                  "backupStorageType": "S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdBackup",
                "metadata": {
                  "name": "example-etcd-cluster-backup"
                },
                "spec": {
                  "etcdEndpoints": ["<etcd-cluster-endpoints>"],
                  "storageType":"S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
          createdAt: "2019-02-28 01:03:00"
          description: Create and maintain highly-available etcd clusters on Kubernetes
          repository: https://github.com/coreos/etcd-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to backup an etcd cluster.
            displayName: etcd Backup
            kind: EtcdBackup
            name: etcdbackups.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to restore an etcd cluster from a backup.
            displayName: etcd Restore
            kind: EtcdRestore
            name: etcdrestores.etcd.database.coreos.com
            version: v1beta2
        description: |
          The etcd Operater creates and maintains highly-available etcd clusters on Kubernetes, allowing engineers to easily deploy and manage etcd clusters for their applications.

          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader.


          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` via port forwarding:

              $ kubectl --namespace default port-forward service/example-client 2379:2379
              $ etcdctl --endpoints http://127.0.0.1:2379 get /

          Or directly to the API using the automatically generated Kubernetes Service:

              $ etcdctl --endpoints http://example-client.default.svc:2379 get /

          Be sure to secure your etcd cluster (see Common Configurations) before exposing it outside of the namespace or cluster.


          ### Supported Features

          * **High availability** - Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.

          * **Automated updates** - Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.

          * **Backups included** - Create etcd backups and restore them through the etcd Operator.

          ### Common Configurations

          * **Configure TLS** - Specify [static TLS certs](https://github.com/coreos/etcd-operator/blob/master/doc/user/cluster_tls.md) as Kubernetes secrets.

          * **Set Node Selector and Affinity** - [Spread your etcd Pods](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-node-selector-and-anti-affinity-across-nodes) across Nodes and availability zones.

          * **Set Resource Limits** - [Set the Kubernetes limit and request](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-resource-requirement) values for your etcd Pods.

          * **Customize Storage** - [Set a custom StorageClass](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#custom-persistentvolumeclaim-definition) that you would like to use.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: etcd-dev@googlegroups.com
          name: etcd Community
        maturity: alpha
        provider:
          name: CNCF
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
        version: 0.9.4
      entries:
      - name: etcdoperator.v0.9.4
        version: 0.9.4
      - name: etcdoperator.v0.9.2
        version: 0.9.2
      - name: etcdoperator.v0.9.0
        version: 0.9.0
      name: singlenamespace-alpha
    defaultChannel: singlenamespace-alpha
    packageName: etcd
    provider:
      name: CNCF
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataStax
      provider-url: ""
    name: cass-operator-community
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cass-operator.v1.23.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cassandra.datastax.com/v1beta1",
                "kind": "CassandraDatacenter",
                "metadata": {
                  "name": "dc1"
                },
                "spec": {
                  "clusterName": "development",
                  "config": {
                    "cassandra-yaml": {
                      "authenticator": "PasswordAuthenticator",
                      "authorizer": "CassandraAuthorizer",
                      "num_tokens": 16,
                      "role_manager": "CassandraRoleManager"
                    },
                    "jvm-server-options": {
                      "initial_heap_size": "1G",
                      "max_heap_size": "1G"
                    }
                  },
                  "managementApiAuth": {
                    "insecure": {}
                  },
                  "podTemplateSpec": {
                    "spec": {
                      "containers": [
                        {
                          "name": "cassandra"
                        }
                      ],
                      "securityContext": {}
                    }
                  },
                  "racks": [
                    {
                      "name": "rack1"
                    },
                    {
                      "name": "rack2"
                    },
                    {
                      "name": "rack3"
                    }
                  ],
                  "resources": {
                    "requests": {
                      "cpu": "1000m",
                      "memory": "2Gi"
                    }
                  },
                  "serverType": "cassandra",
                  "serverVersion": "4.1.2",
                  "size": 3,
                  "storageConfig": {
                    "cassandraDataVolumeClaimSpec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "server-storage"
                    }
                  }
                }
              },
              {
                "apiVersion": "control.k8ssandra.io/v1alpha1",
                "kind": "CassandraTask",
                "metadata": {
                  "name": "example-cleanup"
                },
                "spec": {
                  "datacenter": {
                    "name": "dc2",
                    "namespace": "cass-operator"
                  },
                  "jobs": [
                    {
                      "command": "cleanup",
                      "name": "cleanup-run"
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database,Big Data,OpenShift Optional
          certified: "true"
          containerImage: cr.k8ssandra.io/k8ssandra/cass-operator:v1.23.2
          createdAt: "2025-03-20"
          description: |
            Simple provisioning, turn-key operations, and automated remediation of Apache Cassandra clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/k8ssandra/cass-operator
          support: DataStax
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CassandraDatacenter is the Schema for the cassandradatacenters
              API
            displayName: Cassandra Datacenter
            kind: CassandraDatacenter
            name: cassandradatacenters.cassandra.datastax.com
            version: v1beta1
          - description: CassandraTask is the Schema for the cassandrajobs API
            displayName: Cassandra Task
            kind: CassandraTask
            name: cassandratasks.control.k8ssandra.io
            version: v1alpha1
        description: "## Apache Cassandra\n\nThe Apache Cassandra database is the
          right choice when you need scalability and\nhigh availability without compromising
          performance. Linear scalability and\nproven fault-tolerance on commodity
          hardware or cloud infrastructure make it the\nperfect platform for mission-critical
          data. Cassandra's support for replicating\nacross multiple datacenters is
          best-in-class, providing lower latency for your\nusers and the peace of
          mind of knowing that you can survive regional outages.\n\n## DataStax Enterprise\n\nThe
          most advanced distribution of Apache Cassandra™ on the market, with the\nenterprise
          functionality needed for serious production systems and backed up and\nsupported
          by the best distributed-experts in the world. It's one platform for\nall
          types of applications anywhere, any cloud, any model: key-value, graph,\ntabular,
          JSON.\n\nDataStax Enterprise is a fully integrated and optimized database,
          with graph,\nanalytics, and search included, all with a unified security
          model. Simply put,\nit's the only database capable of meeting today's demanding
          requirements\n\n## Operator Details\n\n`cass-operator` is designed as a
          modular operator for Apache Cassandra and\nderived  distributions. Apache
          Cassandra is a distributed database consisting of\nmultiple nodes working
          in concert to store data and process queries along a\nnumber of fault domains.
          `cass-operator` has the deployment of a Cassandra\ncluster around the logical
          domain of a datacenter with the `CassandraDatacenter`\ncustom resource.\n
          \   \nUpon submission of one of these resources it handles provisioning
          the underlying\nstateful sets (analogous to C\\* logical racks), services,
          and configuration.\nAdditionally through monitoring pod state via Kubernetes
          callbacks it handles day to day\noperations such as restarting failed processes,
          scaling clusters up, and deploying\nconfiguration changes in a rolling,
          non-disruptive, fashion.\n    \nThis operator is designed to be `Namespace`
          scoped. A single Kubernetes cluster may\nbe running multiple instances of
          this operator, in separate namespaces, to support\na number of C\\* clusters
          and environments. Configuration is simple with the usage of\nYAML based
          overrides in the Custom Resource paired with an `init` container.\n    \nIn
          C\\* clusters ordering and timing of certain operations are important to
          keep the system\nevenly distributed. `cass-operator` takes advantage of
          a sidecar process within the\nmain container to handle the orchestration
          of starting our main C* process."
        displayName: DataStax Kubernetes Operator for Apache Cassandra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - k8ssandra
        - cass-operator
        - DataStax Enterprise
        - Apache Cassandra
        - DSE
        - C*
        - Cassandra
        links:
        - name: DataStax
          url: https://datastax.com
        - name: Operator GitHub Repo
          url: https://github.com/k8ssandra/cass-operator
        - name: Config Builder GitHub Repo
          url: https://github.com/datastax/cass-config-builder
        - name: Documentation [1]
          url: https://github.com/k8ssandra/cass-operator/tree/master/docs/user
        - name: Documentation [2]
          url: https://docs.datastax.com/en/cass-operator/doc/cass-operator/cassOperatorTOC.html
        maintainers:
        - email: cass-operator@datastax.com
          name: DataStax Operator Team
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: DataStax
        relatedImages:
        - cr.k8ssandra.io/k8ssandra/cass-operator:v1.23.2
        version: 1.23.2
      entries:
      - name: cass-operator.v1.23.2
        version: 1.23.2
      - name: cass-operator.v1.23.1
        version: 1.23.1
      - name: cass-operator.v1.23.0
        version: 1.23.0
      - name: cass-operator.v1.21.1
        version: 1.21.1
      - name: cass-operator.v1.19.1
        version: 1.19.1
      - name: cass-operator.v1.17.1
        version: 1.17.1
      - name: cass-operator.v1.16.0
        version: 1.16.0
      - name: cass-operator.v1.14.0
        version: 1.14.0
      - name: cass-operator.v1.13.0
        version: 1.13.0
      - name: cass-operator.v1.10.4
        version: 1.10.4
      - name: cass-operator.v1.10.1
        version: 1.10.1
      name: stable
    defaultChannel: stable
    packageName: cass-operator-community
    provider:
      name: DataStax
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: openstack-ansibleee-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openstack-ansibleee-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ansibleee.openstack.org/v1beta1",
                "kind": "OpenStackAnsibleEE",
                "metadata": {
                  "name": "openstackansibleee-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-09-23T20:57:34Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.3
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenStackAnsibleEE is the Schema for the openstackansibleees
              API
            displayName: OpenStack Ansible EE
            kind: OpenStackAnsibleEE
            name: openstackansibleees.ansibleee.openstack.org
            version: v1beta1
        description: The OpenStack Ansible Execution Environment Operator
        displayName: OpenStackAnsibleEE
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - execution
        links:
        - name: OpenStack Ansible EE Operator
          url: https://github.com/openstack-k8s-operators/openstack-ansibleee-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:24b468b09b43dd6cf77f5b08a68f2c6359d115722d28982fd8cdb1923e8122de
        - registry.redhat.io/rhoso-operators/openstack-ansibleee-rhel9-operator@sha256:bf096518547e92fb6fe0ecd526813bdc4a15f928a6d89eacadb71ffaad91db69
        - registry.redhat.io/rhoso-operators/ee-openstack-ansible-ee-rhel9@sha256:640efe595fbdb993ecc1c55c14457edc2b0e3e1c6dde8ad9ea1acf4b4503339f
        version: 1.0.3
      entries:
      - name: openstack-ansibleee-operator.v1.0.3
        version: 1.0.3
      - name: openstack-ansibleee-operator.v1.0.2
        version: 1.0.2
      - name: openstack-ansibleee-operator.v1.0.1
        version: 1.0.1
      - name: openstack-ansibleee-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: openstack-ansibleee-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: ""
    name: red-hat-hawtio-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: red-hat-hawtio-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              },
              {
                "apiVersion": "hawt.io/v2",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24,
                    "internalSSL": true
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
          createdAt: "2025-08-07T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v2
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
        version: 1.2.1
      entries:
      - name: red-hat-hawtio-operator.v1.2.1
        version: 1.2.1
      - name: red-hat-hawtio-operator.v1.2.0-0.1754874242.p
        version: 1.2.0+0.1754874242.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754338551.p
        version: 1.2.0+0.1754338551.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754317392.p
        version: 1.2.0+0.1754317392.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1753863723.p
        version: 1.2.0+0.1753863723.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752679472.p
        version: 1.2.0+0.1752679472.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752452299.p
        version: 1.2.0+0.1752452299.p
      - name: red-hat-hawtio-operator.v1.2.0
        version: 1.2.0
      - name: red-hat-hawtio-operator.v1.1.1
        version: 1.1.1
      - name: red-hat-hawtio-operator.v1.1.0
        version: 1.1.0
      - name: red-hat-hawtio-operator.v1.1.0-0.1742200932.p
        version: 1.1.0+0.1742200932.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1740358952.p
        version: 1.1.0+0.1740358952.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739994212.p
        version: 1.1.0+0.1739994212.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739992853.p
        version: 1.1.0+0.1739992853.p
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: latest
    - currentCSV: red-hat-hawtio-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              },
              {
                "apiVersion": "hawt.io/v2",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24,
                    "internalSSL": true
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
          createdAt: "2025-08-07T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v2
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
        version: 1.2.1
      entries:
      - name: red-hat-hawtio-operator.v1.2.1
        version: 1.2.1
      - name: red-hat-hawtio-operator.v1.2.0-0.1754874242.p
        version: 1.2.0+0.1754874242.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754338551.p
        version: 1.2.0+0.1754338551.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754317392.p
        version: 1.2.0+0.1754317392.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1753863723.p
        version: 1.2.0+0.1753863723.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752679472.p
        version: 1.2.0+0.1752679472.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752452299.p
        version: 1.2.0+0.1752452299.p
      - name: red-hat-hawtio-operator.v1.2.0
        version: 1.2.0
      - name: red-hat-hawtio-operator.v1.1.1
        version: 1.1.1
      - name: red-hat-hawtio-operator.v1.1.0
        version: 1.1.0
      - name: red-hat-hawtio-operator.v1.1.0-0.1742200932.p
        version: 1.1.0+0.1742200932.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1740358952.p
        version: 1.1.0+0.1740358952.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739994212.p
        version: 1.1.0+0.1739994212.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739992853.p
        version: 1.1.0+0.1739992853.p
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: stable-v1
    - currentCSV: red-hat-hawtio-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel8-operator@sha256:f04d81d24c757db7b678464382ad8c5182b1580e714123a9f4f5fc9fbd9b8c57
          createdAt: "2024-01-30T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-rhel8-operator@sha256:f04d81d24c757db7b678464382ad8c5182b1580e714123a9f4f5fc9fbd9b8c57
        version: 1.0.3
      entries:
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1
    packageName: red-hat-hawtio-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MinIO
      provider-url: https://min.io
    name: minio-key-manager-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-key-manager-operator.v2025.8.19175300
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aistor.min.io/v1alpha1",
                "kind": "KeyManager",
                "metadata": {
                  "name": "my-kms",
                  "namespace": "ns-1"
                },
                "spec": {
                  "apiKeySecret": {
                    "name": "my-api-key"
                  },
                  "configuration": {
                    "name": "my-kms-server-config"
                  },
                  "hsmSecret": {
                    "name": "my-kms-hsm"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "registry-creds"
                    }
                  ],
                  "replicas": 2,
                  "volumeClaimTemplate": {
                    "metadata": {
                      "name": "key-manager-volume"
                    },
                    "spec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "100Mi"
                        }
                      },
                      "storageClassName": "standard"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security, Storage
          createdAt: "2025-08-27T14:50:59Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KeyManager
            name: keymanagers.aistor.min.io
            version: v1alpha1
        description: |-
          Highly available, powerful and operationally simple, MinIO's Enterprise Key Management Server is optimized for large storage infrastructures where billions of cryptographic keys are required.

          MinIO's AIStor KMS establishes its foundational trust using the concept of an hardware security module (but given KES is software, this is only a concept). That module assumes a pivotal role in sealing and unsealing the KMS root encryption key. The module responsibility extends to safeguarding the integrity of KMS by allowing the unsealing of its encrypted on-disk state and facilitating communication among nodes within a KMS cluster.

          It solves the challenges associated with billions of cryptographic keys and hundreds of thousands of cryptographic operations per node per second - which are commonplace in larger deployments.

          <b>High Availability and Fault Tolerance</b>
          In the dynamic landscape of large-scale systems, network or node outages are inevitable. Taking down a cluster for maintenance is rarely feasible. MinIO's AIStor KMS ensures uninterrupted availability, even when faced with such disruptions, mitigating cascading effects that can take down the entire storage infrastructure. Specifically, you could lose all but one node of a cluster and still handle any encryption, decryption or data key generation requests.

          <b>Scalability</b>
          While the amount of data usually only increases, the load on a large-scale storage system may vary significantly from time to time. MinIO's AIStor KMS supports dynamic cluster resizing and nodes can be added or removed at any point without incurring any downtime.

          <b>Multi-Tenancy</b>
          Large-scale storage infrastructures are often used by many applications and teams across the entire organization. Isolating teams and groups into their own namespaces is a core requirement. MinIO's AIStor KMS supports namespacing in the form of enclaves. Each tenant can be assigned its own enclave which is completely independent and isolated from all other enclaves on the KMS cluster.

          <b>Predictable Behavior</b>
          MinIO's AIStor KMS is designed to be easily managed, providing operators with the ability to comprehend its state intuitively. Due to its simple design, MinIO's AIStor KMS is significantly easier to operate than similar solutions that rely on more complex consensus algorithms like Raft, or Paxos.

          <b>Consistent and Performant</b>
          The responsiveness of the KMS for GET/PUT operations directly influences the overall efficiency and speed of the storage system. MinIO's AIStor KMS nodes don’t have to coordinate when handling such requests from the storage system. Therefore, the performance of a MinIO's AIStor KMS cluster increases linearly with the number of nodes. Further, MinIO's AIStor KMS supports request pipelining to handle hundreds of thousands of cryptographic operations per node and second.

          <b>Simplicity</b>
          Operating a KMS cluster does not require expertise in cryptography or distributed systems. Everything can be done from the AIStor Console.

          <a href="https://resources.min.io/request-a-demo-marketplaces/">Request a demo today</a>
        displayName: Minio AIStor Key management System (KMS)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Storage
        - MinIO
        links:
        - name: Minio AIStor
          url: https://min.io
        - name: Support
          url: https://subnet.min.io
        - name: Github
          url: https://github.com/minio/aistor
        maintainers:
        - email: dev@min.io
          name: MinIO
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: MinIO
          url: https://min.io
        relatedImages:
        - quay.io/minio/aistor/operator@sha256:f80139efa062fdd83f78318b20f371a827b68e854b03dd9706b9eba4c493aeb2
        version: 2025.8.19175300
      entries:
      - name: minio-key-manager-operator.v2025.8.19175300
        version: 2025.8.19175300
      - name: minio-key-manager-operator.v2025.8.1184927
        version: 2025.8.1184927
      - name: minio-key-manager-operator.v2025.7.1011319
        version: 2025.7.1011319
      - name: minio-key-manager-operator.v2025.6.3065135
        version: 2025.6.3065135
      - name: minio-key-manager-operator.v2025.5.12190907
        version: 2025.5.12190907
      name: stable
    defaultChannel: stable
    packageName: minio-key-manager-operator
    provider:
      name: MinIO
      url: https://min.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator-certified.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a",
                  "imagePullPolicy": "Always",
                  "initImage": "registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:3c03fb18a5d73876576b80aca2b61e7c84ca052839ccb467d14e15f6d1f9d834
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a57926b53eef862b70a53818a7ccb0e9756f878f30d4128b608090b109bdc585
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator-certified.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator-certified.v1.20.0
        version: 1.20.0
      - name: percona-server-mongodb-operator-certified.v1.19.1
        version: 1.19.1
      - name: percona-server-mongodb-operator-certified.v1.19.0
        version: 1.19.0
      - name: percona-server-mongodb-operator-certified.v1.18.0
        version: 1.18.0
      - name: percona-server-mongodb-operator-certified.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator-certified
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Cloudera
      provider-url: https://cloudera.com
    name: cloudera-streams-messaging-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster"
                },
                "spec": {
                  "kafka": {
                    "version": "3.9.0.1.4",
                    "replicas": 3,
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2,
                      "inter.broker.protocol.version": "3.9"
                    },
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "zookeeper": {
                    "replicas": 3,
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "3.9.0.1.4",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
          createdAt: "2025-07-14 10:09:41"
          description: Run Apache Kafka with Cloudera Streams Messaging - Kubernetes
            Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: A valid license and current subscription
            for Cloudera Streaming - Kubernetes Operator is required. Contact Cloudera
            sales to obtain a license. https://www.cloudera.com/contact-sales.html
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |-
          Cloudera Streams Messaging - Kubernetes Operator brings enterprise-grade Kafka deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of Kafka for enterprise deployments. To learn more about Cloudera Streaming, [read this deep dive blog post](https://blog.cloudera.com/accelerating-deployments-of-streaming-pipelines-announcing-data-in-motion-on-kubernetes/).

          Cloudera Streams Messaging - Kubernetes Operator consists of and ships multiple components including Apache Kafka, Apache ZooKeeper, Cruise Control, Strimzi, and others.

          ## Installation prerequisites

          Installing Cloudera Streams Messaging - Kubernetes Operator requires a valid license as well as Cloudera credentials (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly. The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) where installation artifacts are pulled from.

          ### Obtaining a license

          Contact [Cloudera Sales](https://www.cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets

          Before installation, create  two `Secrets`. One containing your Cloudera credentials, and one containing your license. Cloudera credentials are provided to you together with your license.

          Create a `Secret` containing your Cloudera credentials.
          ```
          oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***] --docker-server=container.repository.cloudera.com
          ```
          Create a `Secret` containing your license.
          ```
          oc create secret generic csm-op-license --from-file=license.txt=[***PATH TO LICENSE FILE***]
          ```

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces where you plan to deploy Kafka or Kafka Connect clusters. Cloudera recommends that you create the `Secret` in all required namespaces before installation if you know what namespaces you will be using to deploy Kafka or Kafka Connect.

          ## Learn more
          To learn more, visit the [Cloudera Streams Messaging - Kubernetes Operator documentation library](https://docs.cloudera.com/csm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Streams Messaging - Kubernetes Operator, visit the following resources:

          * [Strimzi](https://strimzi.io/)
          * [Apache Kafka](https://kafka.apache.org/)
          * [Apache ZooKeeper](https://zookeeper.apache.org/)
          * [Cruise Control](https://github.com/linkedin/cruise-control)
        displayName: Cloudera Streams Messaging - Kubernetes Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Documentation
          url: https://docs.cloudera.com/csm-operator
        - name: Cloudera Community Forum
          url: https://community.cloudera.com/
        - name: Cloudera Stream Processing
          url: https://www.cloudera.com/products/stream-processing.html
        - name: Contact Cloudera
          url: https://www.cloudera.com/contact-sales.html
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
        - container.repository.cloudera.com/cloudera/kafka@sha256:aa64e161b9672153a6f6559f9ae9ef88d0e40d6040a80e21bd3dc8ee39df6523
        - container.repository.cloudera.com/cloudera/kafka@sha256:b675e7426a558c4681e53d8e03f278264d7ad555f7c02422017683a29708ac2b
        - container.repository.cloudera.com/cloudera/maven-builder@sha256:9f18203b175193e5079fdc0c2b71c14887641d997f92b4a1c918546f777166df
        version: 1.4.0-b199
      entries:
      - name: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
        version: 1.4.0-b199
      name: 1.2.x
    - currentCSV: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster"
                },
                "spec": {
                  "kafka": {
                    "version": "3.9.0.1.4",
                    "replicas": 3,
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2,
                      "inter.broker.protocol.version": "3.9"
                    },
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "zookeeper": {
                    "replicas": 3,
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "3.9.0.1.4",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
          createdAt: "2025-07-14 10:09:41"
          description: Run Apache Kafka with Cloudera Streams Messaging - Kubernetes
            Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: A valid license and current subscription
            for Cloudera Streaming - Kubernetes Operator is required. Contact Cloudera
            sales to obtain a license. https://www.cloudera.com/contact-sales.html
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |-
          Cloudera Streams Messaging - Kubernetes Operator brings enterprise-grade Kafka deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of Kafka for enterprise deployments. To learn more about Cloudera Streaming, [read this deep dive blog post](https://blog.cloudera.com/accelerating-deployments-of-streaming-pipelines-announcing-data-in-motion-on-kubernetes/).

          Cloudera Streams Messaging - Kubernetes Operator consists of and ships multiple components including Apache Kafka, Apache ZooKeeper, Cruise Control, Strimzi, and others.

          ## Installation prerequisites

          Installing Cloudera Streams Messaging - Kubernetes Operator requires a valid license as well as Cloudera credentials (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly. The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) where installation artifacts are pulled from.

          ### Obtaining a license

          Contact [Cloudera Sales](https://www.cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets

          Before installation, create  two `Secrets`. One containing your Cloudera credentials, and one containing your license. Cloudera credentials are provided to you together with your license.

          Create a `Secret` containing your Cloudera credentials.
          ```
          oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***] --docker-server=container.repository.cloudera.com
          ```
          Create a `Secret` containing your license.
          ```
          oc create secret generic csm-op-license --from-file=license.txt=[***PATH TO LICENSE FILE***]
          ```

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces where you plan to deploy Kafka or Kafka Connect clusters. Cloudera recommends that you create the `Secret` in all required namespaces before installation if you know what namespaces you will be using to deploy Kafka or Kafka Connect.

          ## Learn more
          To learn more, visit the [Cloudera Streams Messaging - Kubernetes Operator documentation library](https://docs.cloudera.com/csm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Streams Messaging - Kubernetes Operator, visit the following resources:

          * [Strimzi](https://strimzi.io/)
          * [Apache Kafka](https://kafka.apache.org/)
          * [Apache ZooKeeper](https://zookeeper.apache.org/)
          * [Cruise Control](https://github.com/linkedin/cruise-control)
        displayName: Cloudera Streams Messaging - Kubernetes Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Documentation
          url: https://docs.cloudera.com/csm-operator
        - name: Cloudera Community Forum
          url: https://community.cloudera.com/
        - name: Cloudera Stream Processing
          url: https://www.cloudera.com/products/stream-processing.html
        - name: Contact Cloudera
          url: https://www.cloudera.com/contact-sales.html
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/kafka@sha256:aa64e161b9672153a6f6559f9ae9ef88d0e40d6040a80e21bd3dc8ee39df6523
        - container.repository.cloudera.com/cloudera/kafka@sha256:b675e7426a558c4681e53d8e03f278264d7ad555f7c02422017683a29708ac2b
        - container.repository.cloudera.com/cloudera/maven-builder@sha256:9f18203b175193e5079fdc0c2b71c14887641d997f92b4a1c918546f777166df
        - container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
        version: 1.4.0-b199
      entries:
      - name: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
        version: 1.4.0-b199
      name: stable
    defaultChannel: stable
    packageName: cloudera-streams-messaging-kubernetes-operator
    provider:
      name: Cloudera
      url: https://cloudera.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: vertical-pod-autoscaler
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: verticalpodautoscaler.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "autoscaling.openshift.io/v1",
                "kind": "VerticalPodAutoscalerController",
                "metadata": {
                  "name": "default",
                  "namespace": "openshift-vertical-pod-autoscaler"
                },
                "spec": {
                  "podMinCPUMillicores": 25,
                  "podMinMemoryMb": 250,
                  "recommendationOnly": false,
                  "safetyMarginFraction": 0.15,
                  "minReplicas": 2
                }
              },
              {
                 "apiVersion": "autoscaling.k8s.io/v1",
                 "kind": "VerticalPodAutoscaler",
                 "metadata": {
                    "name": "myapp-vpa"
                 },
                 "spec": {
                    "targetRef": {
                       "apiVersion": "apps/v1",
                       "kind": "Deployment",
                       "name": "myapp-deployment"
                    },
                    "resourcePolicy": {
                       "containerPolicies": [
                          {
                             "containerName": "*",
                             "minAllowed": {
                                "cpu": "100m",
                                "memory": "50Mi"
                             },
                             "maxAllowed": {
                                "cpu": 1,
                                "memory": "500Mi"
                             },
                             "controlledResources": [
                                "cpu",
                                "memory"
                             ]
                          }
                       ]
                    }
                 }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          certifiedLevel: Primed
          containerImage: registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9-operator@sha256:edf76334695c9a4aee4a9c0297af40475a16dec7e35ec7113a35b8507724536e
          createdAt: 2019-10-14 12:17:13+00:00
          description: An operator to run the OpenShift Vertical Pod Autoscaler. Vertical
            Pod Autoscaler (VPA) can be configured to monitor a workload's resource
            utilization, and then adjust its CPU and memory limits by updating the
            pod (future) or restarting the pod with updated limits.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          healthIndex: B
          olm.skipRange: '>=4.5.0 <4.16.0-202509240123'
          operatorframework.io/suggested-namespace: openshift-vertical-pod-autoscaler
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/vertical-pod-autoscaler-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of the set of VPA controllers
            displayName: VPA Controller
            kind: VerticalPodAutoscalerController
            name: verticalpodautoscalercontrollers.autoscaling.openshift.io
            version: v1
          - description: An application whose resources are managed by the VPA
            displayName: VerticalPodAutoscaler
            kind: VerticalPodAutoscaler
            name: verticalpodautoscalers.autoscaling.k8s.io
            version: v1
          - description: Represents VPA checkpoints used by the VPA controllers to
              manage workloads
            displayName: VPA Checkpoint
            kind: VerticalPodAutoscalerCheckpoint
            name: verticalpodautoscalercheckpoints.autoscaling.k8s.io
            version: v1
        description: An operator to run the OpenShift Vertical Pod Autoscaler. Vertical
          Pod Autoscaler (VPA) can be configured to monitor a workload's resource
          utilization, and then adjust its CPU and memory limits by updating the pod
          (future) or restarting the pod with updated limits.
        displayName: VerticalPodAutoscaler
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Vertical Pod Autoscaler Documentation
          url: https://docs.openshift.com/container-platform/latest/nodes/pods/nodes-pods-vertical-autoscaler.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9-operator@sha256:edf76334695c9a4aee4a9c0297af40475a16dec7e35ec7113a35b8507724536e
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9@sha256:e6c8f34895184179bb87f1659c4ff30b96b131787fa7a64e7bc6b6adbb336c16
        version: 4.16.0-202509240123
      entries:
      - name: verticalpodautoscaler.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: verticalpodautoscaler.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: verticalpodautoscaler.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: verticalpodautoscaler.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: verticalpodautoscaler.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: verticalpodautoscaler.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: verticalpodautoscaler.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: verticalpodautoscaler.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: verticalpodautoscaler.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: verticalpodautoscaler.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: verticalpodautoscaler.v4.16.0-202501260334
        version: 4.16.0-202501260334
      - name: verticalpodautoscaler.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: verticalpodautoscaler.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: verticalpodautoscaler.v4.16.0-202412021735
        version: 4.16.0-202412021735
      - name: verticalpodautoscaler.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: verticalpodautoscaler.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: verticalpodautoscaler.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: verticalpodautoscaler.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: verticalpodautoscaler.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: verticalpodautoscaler.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: verticalpodautoscaler.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: verticalpodautoscaler.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: verticalpodautoscaler.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: verticalpodautoscaler.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: verticalpodautoscaler.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: verticalpodautoscaler.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: verticalpodautoscaler.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: verticalpodautoscaler.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: verticalpodautoscaler.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: verticalpodautoscaler.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: vertical-pod-autoscaler
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-businessautomation: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: businessautomation-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: businessautomation-operator.7.13.5-16
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.kiegroup.org/v2","kind":"KieApp","metadata":{"name":"rhpam-trial"},"spec":{"environment":"rhpam-trial"}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/rhpam-7/rhpam-rhel8-operator@sha256:3b4f420def9e5d2c768d5209264bd5f9af6d6a37b22ef9a988a167ff7d9959cf
          createdAt: "2025-07-30 19:54:27"
          description: Deploys and manages Red Hat Process Automation Manager and
            Red Hat Decision Manager environments.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Process Automation
            Manager", "Red Hat Decision Manager"]'
          repository: https://github.com/kiegroup/kie-cloud-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A project prescription running an RHPAM/RHDM environment.
            displayName: KieApp
            kind: KieApp
            name: kieapps.app.kiegroup.org
            version: v2
        description: |-
          Deploys and manages Red Hat Process Automation Manager and Red Hat Decision Manager environments.

          * **Red Hat Process Automation Manager** is a platform for developing containerized microservices and applications that automate business decisions and processes. It includes business process management (BPM), business rules management (BRM), and business resource optimization and complex event processing (CEP) technologies. It also includes a user experience platform to create engaging user interfaces for process and decision services with minimal coding.

          [See more](https://www.redhat.com/en/products/process-automation).
        displayName: Business Automation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kieapp
        - pam
        - decision
        - kie
        - cloud
        - bpm
        - process
        - automation
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-process-automation-manager
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_process_automation_manager/7.13/#category-deploying-red-hat-process-automation-manager-on-openshift
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-monitoring-rhel8@sha256:8d4481a11269565f1690ad007f152e24326bae4b2099cb26ddf7b21be67031e8
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:cb37360843e9b062c59755593d9d412e1bf2c1dc1d6902e8a65a2825812c0989
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhpam-7/rhpam-kieserver-rhel8@sha256:2526d12da63b67ae9ed9294c914e9406d833c090ffa777623303ea0be22755d8
        - registry.redhat.io/rhpam-7/rhpam-process-migration-rhel8@sha256:3d7bdbe38895e50549e97573bd40da47b11691932164db7dafff0155a051aab2
        - registry.redhat.io/rhpam-7/rhpam-controller-rhel8@sha256:3a901b54a9d35d38b5aca275df2160e3190a874e139fcb606b9969b38284b3dc
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        - registry.redhat.io/rhpam-7/rhpam-kieserver-rhel8@sha256:3fbef82d1c5da8da1606d1a2d5dcfd20000da574737eb32b772d4f0123b179cb
        - registry.redhat.io/rhpam-7/rhpam-smartrouter-rhel8@sha256:d1e5461db8983a13f98259152dc3cfa2aa677c29f799104190e5cefa83cf97b5
        - registry.redhat.io/rhpam-7/rhpam-dashbuilder-rhel8@sha256:46396dc8060f2f9cc6f87ad905ef55be5b8e5216e5162bdf54df98c9d70e29ef
        - registry.redhat.io/rhscl/mysql-80-rhel7@sha256:e4f535c78bc2fdb58914c52fb805185035ff31515452b4df714f9ab623ba383a
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-rhel8@sha256:228c6ab5b83e215a98f6eb76250feb2ae3096fd21398c44e5b2701257c9c3493
        - registry.redhat.io/rhpam-7/rhpam-smartrouter-rhel8@sha256:9d7efdd97f163770d8a84b1ce5236bd4f31508154dced1f11dc6efe75a7b5089
        - registry.redhat.io/rhscl/postgresql-13-rhel7@sha256:338dd09681dcf860e6a976a3bc48a1474c2376d698815e4eb611ae9c5aa68966
        - registry.redhat.io/rhpam-7/rhpam-rhel8-operator@sha256:3b4f420def9e5d2c768d5209264bd5f9af6d6a37b22ef9a988a167ff7d9959cf
        - registry.redhat.io/rhpam-7/rhpam-dashbuilder-rhel8@sha256:4edac282119898ab1395030eecaee584990064e88fc8e65fb999f19ea2ef3f8b
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-rhel8@sha256:1abbaba9d4b2ecaec38205bcccf5ef4c7b6388ed9725b4f4c6dc8dc664588eb1
        - registry.redhat.io/rhpam-7/rhpam-controller-rhel8@sha256:989dc6dc64a2d5b6723fb5c000b0411fbdc6f78c0056666d831ffeabeab4b612
        - registry.redhat.io/openshift4/ose-cli@sha256:c3eb8807caf73e30012ae53c6d4f5114471290a5596929c27acef5a699a5b813
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/rhpam-7/rhpam-process-migration-rhel8@sha256:bd5c6586b4859dec23ad9530ee98bab33497ab251d544f863fc0af01cbd45101
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:457fc73423fa94591bc7e35729a25c58d3465583afccf342fb338b6e11188575
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-monitoring-rhel8@sha256:32912270eba0f2464da88f315bde8bf1ef5bbded814db156811d7a415a458f86
        version: 7.13.5-16
      entries:
      - name: businessautomation-operator.7.13.5-16
        version: 7.13.5-16
      - name: businessautomation-operator.7.13.5-15
        version: 7.13.5-15
      - name: businessautomation-operator.7.13.5-14
        version: 7.13.5-14
      - name: businessautomation-operator.7.13.5-13
        version: 7.13.5-13
      - name: businessautomation-operator.7.13.5-12
        version: 7.13.5-12
      - name: businessautomation-operator.7.13.5-11
        version: 7.13.5-11
      - name: businessautomation-operator.7.13.5-10
        version: 7.13.5-10
      - name: businessautomation-operator.7.13.5-9
        version: 7.13.5-9
      - name: businessautomation-operator.7.13.5-8
        version: 7.13.5-8
      - name: businessautomation-operator.7.13.5-7
        version: 7.13.5-7
      - name: businessautomation-operator.7.13.5-6
        version: 7.13.5-6
      - name: businessautomation-operator.7.13.5-5
        version: 7.13.5-5
      - name: businessautomation-operator.7.13.5-4
        version: 7.13.5-4
      - name: businessautomation-operator.7.13.5-3
        version: 7.13.5-3
      - name: businessautomation-operator.7.13.5-2
        version: 7.13.5-2
      - name: businessautomation-operator.7.13.5-1
        version: 7.13.5-1
      - name: businessautomation-operator.7.13.4-4
        version: 7.13.4-4
      - name: businessautomation-operator.7.13.4-3
        version: 7.13.4-3
      - name: businessautomation-operator.7.13.4-2
        version: 7.13.4-2
      - name: businessautomation-operator.7.13.4-1
        version: 7.13.4-1
      - name: businessautomation-operator.7.13.3-5
        version: 7.13.3-5
      - name: businessautomation-operator.7.13.3-4
        version: 7.13.3-4
      - name: businessautomation-operator.7.13.3-3
        version: 7.13.3-3
      - name: businessautomation-operator.7.13.3-2
        version: 7.13.3-2
      - name: businessautomation-operator.7.13.3-1
        version: 7.13.3-1
      - name: businessautomation-operator.7.13.2-3
        version: 7.13.2-3
      - name: businessautomation-operator.7.13.2-2
        version: 7.13.2-2
      - name: businessautomation-operator.7.13.2-1
        version: 7.13.2-1
      - name: businessautomation-operator.7.13.1-2
        version: 7.13.1-2
      - name: businessautomation-operator.7.13.1-1
        version: 7.13.1-1
      - name: businessautomation-operator.7.13.0-1
        version: 7.13.0-1
      - name: businessautomation-operator.7.12.1-4
        version: 7.12.1-4
      - name: businessautomation-operator.7.12.1-3
        version: 7.12.1-3
      - name: businessautomation-operator.7.12.1-2
        version: 7.12.1-2
      - name: businessautomation-operator.7.12.1-1
        version: 7.12.1-1
      - name: businessautomation-operator.7.12.0-2
        version: 7.12.0-2
      - name: businessautomation-operator.7.12.0-1
        version: 7.12.0-1
      - name: businessautomation-operator.7.11.1-3
        version: 7.11.1-3
      - name: businessautomation-operator.7.11.1-2
        version: 7.11.1-2
      - name: businessautomation-operator.7.11.1-1
        version: 7.11.1-1
      - name: businessautomation-operator.7.11.0-2
        version: 7.11.0-2
      - name: businessautomation-operator.7.11.0-1
        version: 7.11.0-1
      - name: businessautomation-operator.7.10.1-2
        version: 7.10.1-2
      - name: businessautomation-operator.7.10.1-1
        version: 7.10.1-1
      - name: businessautomation-operator.7.10.0-2
        version: 7.10.0-2
      - name: businessautomation-operator.7.10.0-1
        version: 7.10.0-1
      - name: businessautomation-operator.7.9.1-1
        version: 7.9.1-1
      - name: businessautomation-operator.7.9.0
        version: 7.9.0
      - name: businessautomation-operator.7.9.0-2
        version: 7.9.0-2
      - name: businessautomation-operator.7.8.1
        version: 7.8.1
      - name: businessautomation-operator.7.8.0
        version: 7.8.0
      - name: businessautomation-operator.1.4.1
        version: 1.4.1
      - name: businessautomation-operator.1.4.0
        version: 1.4.0
      - name: businessautomation-operator.1.3.0
        version: 1.3.0
      - name: businessautomation-operator.1.2.1
        version: 1.2.1
      - name: businessautomation-operator.1.2.0
        version: 1.2.0
      name: stable
    defaultChannel: stable
    packageName: businessautomation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: quay-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:d73321dfa6dcced943405fae45c2ccf841663e0ab57e8be8cc1ddba251b7ce08
          createdAt: 2021-11-02 11:11 UTC
          description: Opinionated deployment of Red Hat Quay on Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:d73321dfa6dcced943405fae45c2ccf841663e0ab57e8be8cc1ddba251b7ce08
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:0e6bef51caf479074925115fa11621c41bcada375cc1925d0511a73084b93543
        - registry.redhat.io/rhel8/postgresql-10@sha256:b2695e9af379624621bef9ec785f0cb618bfdabae98c5f17eea82ff5f8d2d5cb
        - registry.redhat.io/rhel8/redis-5@sha256:8d7e66927ce4b3050ca15ab4df4dbaeb593cb855e16fb0a6f5dde950e9db7db9
        - registry.redhat.io/quay/quay-rhel8@sha256:ed8c4249d376b4226e35bf088ffe5bf688bf68b059718c5e218aa1184fe048d0
        - registry.redhat.io/quay/clair-rhel8@sha256:af5e2fd1543365b365b1a6cce00b86e44fe5e66cde6491e9b113324c21f629c8
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:2a471300f70d3995ab7e9fcc3df35c5513e73fcc1d140fb4d651a1c3eae949eb
        version: 3.4.7
      entries:
      - name: quay-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: quay-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:1f5e4cc9ec2d9b8b89d82f1a698acfa1ac2764ca4ebc5c85b3417da7730b5867
          createdAt: 2021-09-09 20:09 UTC
          description: Opinionated deployment of Red Hat Quay on Kubernetes.
          olm.skipRange: '>=3.4.x <3.5.7'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.5.7
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7fe408a1e428b2d0660cb578c1dcd4b961f838e37166fe7b21b2bed70d288594
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:5d72577ccfd63c6d9ad712b376b604c40d2d3ecb53e391e2ee0d18caa6e0225a
        - registry.redhat.io/rhel8/postgresql-10@sha256:a723bb3a7a5046527b9a5b628ddfb211c526e3814cc4242262c4f8a016f38edd
        - registry.redhat.io/rhel8/redis-5@sha256:7916760ff03af5ca230df5e96b2bf353fbf1829386de1c063c05f623837be7f7
        - registry.redhat.io/quay/quay-rhel8@sha256:c599892160ac20c744d833603d9375923af7aec1edfd86982b7513e02fb6d463
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:1f5e4cc9ec2d9b8b89d82f1a698acfa1ac2764ca4ebc5c85b3417da7730b5867
        - registry.redhat.io/quay/clair-rhel8@sha256:d242bd9d37f95b2a5a8267b3f1eba02c671d46121712baae43466c86d1da89f9
        version: 3.5.7
      entries:
      - name: quay-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: quay-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:3c0201e3cf88e711c921870324002a02cbd42c0d2eecafb4f242aa0601e24602
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.10.15
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:fc20b8bdac88c1bfce9aec993bdce6dd2cc132f8a9d4062d174b098819c74013
        - registry.redhat.io/quay/clair-rhel8@sha256:efa6e69500ea643042b544e6205f088214d79ab09a36864a612e73d07979d1c5
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:3c0201e3cf88e711c921870324002a02cbd42c0d2eecafb4f242aa0601e24602
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7e720c6a332ec4032e8576c31421af404f61d9baaf8e88f69f8333c52e5d1683
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:774cd1f181527787ba8631c7c14437eb2250c15ddb4a9c75fdfa35c5c1a19a2e
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        version: 3.10.15
      entries:
      - name: quay-operator.v3.10.15
        version: 3.10.15
      - name: quay-operator.v3.10.14
        version: 3.10.14
      - name: quay-operator.v3.10.13
        version: 3.10.13
      - name: quay-operator.v3.10.12
        version: 3.10.12
      - name: quay-operator.v3.10.11
        version: 3.10.11
      - name: quay-operator.v3.10.10
        version: 3.10.10
      - name: quay-operator.v3.10.9
        version: 3.10.9
      - name: quay-operator.v3.10.8
        version: 3.10.8
      - name: quay-operator.v3.10.7
        version: 3.10.7
      - name: quay-operator.v3.10.6
        version: 3.10.6
      - name: quay-operator.v3.10.5
        version: 3.10.5
      - name: quay-operator.v3.10.4
        version: 3.10.4
      - name: quay-operator.v3.10.3
        version: 3.10.3
      - name: quay-operator.v3.10.2
        version: 3.10.2
      - name: quay-operator.v3.10.1
        version: 3.10.1
      - name: quay-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:cb3a41caa3a572b7981d424be2a32b8e881127cc03b557ee0a36ab13f430b897
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.11.13
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:7230fa1d4970e877cb6de8716e99a9bd43302c0ff23b4501f2aa806a8d496d9b
        - registry.redhat.io/quay/clair-rhel8@sha256:3746677ef1cb0ddd4f237ab598b14cf0bc76dcce080ffde127b5bba87ffe6693
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:1a083cd72502609427428ed22fb60846613ddb41a4613496c5cddc797b382d3a
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:11b8c9d64c42f0f0c7b2df491bba859d0410666a8ef1d2fd41eb77e4ad9d2dc8
        - registry.redhat.io/rhel8/postgresql-13@sha256:35d86936c539007a78a89869a14b87b76788fe117df442a670f63936b978ed5b
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:cb3a41caa3a572b7981d424be2a32b8e881127cc03b557ee0a36ab13f430b897
        version: 3.11.13
      entries:
      - name: quay-operator.v3.11.13
        version: 3.11.13
      - name: quay-operator.v3.11.12
        version: 3.11.12
      - name: quay-operator.v3.11.11
        version: 3.11.11
      - name: quay-operator.v3.11.10
        version: 3.11.10
      - name: quay-operator.v3.11.9
        version: 3.11.9
      - name: quay-operator.v3.11.8
        version: 3.11.8
      - name: quay-operator.v3.11.7
        version: 3.11.7
      - name: quay-operator.v3.11.6
        version: 3.11.6
      - name: quay-operator.v3.11.5
        version: 3.11.5
      - name: quay-operator.v3.11.4
        version: 3.11.4
      - name: quay-operator.v3.11.3
        version: 3.11.3
      - name: quay-operator.v3.11.2
        version: 3.11.2
      - name: quay-operator.v3.11.1
        version: 3.11.1
      - name: quay-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:accb705b6327fb2dcc1ce657a99f61e343eca752c9873d63d89fea89c10c3187
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.12.12
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:accb705b6327fb2dcc1ce657a99f61e343eca752c9873d63d89fea89c10c3187
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:ccbd83e05aea0bb96bf099f63420175b36ed413bfc8a618f4c90b6a2abc53895
        - registry.redhat.io/quay/clair-rhel8@sha256:dc3d5fb35f2bd5d00480351dca705fe260a225d9a97958cf0276801af08d8307
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:8e5dd171e6e94de0137bb6a972533b1f93b56efc1731e19b80869d5e9b8a26a6
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:8cdd677449f3c723dad79bcd9c4191c8917d6eb0beeeeeda68fe19a0988bb2a8
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        version: 3.12.12
      entries:
      - name: quay-operator.v3.12.12
        version: 3.12.12
      - name: quay-operator.v3.12.11
        version: 3.12.11
      - name: quay-operator.v3.12.10
        version: 3.12.10
      - name: quay-operator.v3.12.9
        version: 3.12.9
      - name: quay-operator.v3.12.8
        version: 3.12.8
      - name: quay-operator.v3.12.7
        version: 3.12.7
      - name: quay-operator.v3.12.6
        version: 3.12.6
      - name: quay-operator.v3.12.5
        version: 3.12.5
      - name: quay-operator.v3.12.4
        version: 3.12.4
      - name: quay-operator.v3.12.3
        version: 3.12.3
      - name: quay-operator.v3.12.2
        version: 3.12.2
      - name: quay-operator.v3.12.1
        version: 3.12.1
      - name: quay-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:5481a4b914ce4aed6ba3ceee8432eddab6e264cad2caef528404795f904100eb
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.13.8
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/clair-rhel8@sha256:069ccf36346cb14f4e4ea54de1d59afb72548912b140cfef67a6bcb7534d25e9
        - registry.redhat.io/rhel8/postgresql-13@sha256:f17fd8dd4d56d8e916cb2ad3f403d05ad259f79fd6918d67a38a6e9e4a8896cf
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:966a3dafa2dd31ac1a39e826e08c1b988fe694189fe79449bbe14a48b2631213
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:5481a4b914ce4aed6ba3ceee8432eddab6e264cad2caef528404795f904100eb
        - registry.redhat.io/quay/quay-rhel8@sha256:8887c17547c7a4be61471523eade9820e5095cf966b1885a3758081da559871c
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:31a9095d8f291ce68e8e688877692e74a17143bdda8d998fe3a07bcee32575dd
        - registry.redhat.io/rhel8/postgresql-15@sha256:a5f36e7893a519c12752c6118233c92d4e83618670bc9ea7bd3a8a34a63bd50b
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:f62cbd6632454c227b7cbb01dadc8e501e8288b2d8e0abf8e7842b806fa3195c
        version: 3.13.8
      entries:
      - name: quay-operator.v3.13.8
        version: 3.13.8
      - name: quay-operator.v3.13.7
        version: 3.13.7
      - name: quay-operator.v3.13.6
        version: 3.13.6
      - name: quay-operator.v3.13.5
        version: 3.13.5
      - name: quay-operator.v3.13.4
        version: 3.13.4
      - name: quay-operator.v3.13.3
        version: 3.13.3
      - name: quay-operator.v3.13.2
        version: 3.13.2
      - name: quay-operator.v3.13.1
        version: 3.13.1
      - name: quay-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:8ebc1f0511f7d394ceaa39a4c993e955c560dc04588072cb3f8c8fab8c85e32b
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.14.5
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/clair-rhel8@sha256:e11e8a819d17bd14c67f093b5303b4958855e731f89ce67dcca70fcb7b77f45a
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:8ebc1f0511f7d394ceaa39a4c993e955c560dc04588072cb3f8c8fab8c85e32b
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-rhel8@sha256:33e984892650d315065313fcb5f8808e145aadec5a287219c7031d85b713235e
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:43ef8adddbf9b007b6e2396f53bd6a5475150d28c2f4254f031a0fe6c7de5826
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:a12a0918f6dea114b0aae231d7119a4acfd0cd297074d895f4ecf190d2aec43a
        version: 3.14.5
      entries:
      - name: quay-operator.v3.14.5
        version: 3.14.5
      - name: quay-operator.v3.14.4
        version: 3.14.4
      - name: quay-operator.v3.14.3
        version: 3.14.3
      - name: quay-operator.v3.14.2
        version: 3.14.2
      - name: quay-operator.v3.14.1
        version: 3.14.1
      - name: quay-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:05fb981d9848281b97b3683093f752b1c1dc5ced439a2b4ba18b26ff8e11231e
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.15.2
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:873a3daf1dadf2438eb96e151072326c31ba34ed48cac041614b604413d5f8e7
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:956f3110f944d880bc27bb6df9472d0804cd75f5dcf4ed11bcf721e9d1d4542b
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:aa8c9f7b6ce57441ca1fe9dbaaefaf2bb83fd44aa367647b663bb822bf890362
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:05fb981d9848281b97b3683093f752b1c1dc5ced439a2b4ba18b26ff8e11231e
        - registry.redhat.io/quay/clair-rhel8@sha256:861be31d30142bae125cd059bb716d79b28694e9c4336333c68adf4a3823b61d
        version: 3.15.2
      entries:
      - name: quay-operator.v3.15.2
        version: 3.15.2
      - name: quay-operator.v3.15.1
        version: 3.15.1
      - name: quay-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-operator.v3.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:7a6e592d7f5f760b4451f9c974dbb96f56b1d3de92f039adcbf284f78ba13a02
          createdAt: 2021-10-12 22:10 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.6.10
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:ec36241f0a9a2983b1894cd8501898e6b70379af51a81b1e09b0ba0e9b54b33f
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:75df6b8e6bef14844a08f7d994470fa6ec4ffa66d4b15faaa0a3ba373f612fbb
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:7a6e592d7f5f760b4451f9c974dbb96f56b1d3de92f039adcbf284f78ba13a02
        - registry.redhat.io/rhel8/postgresql-10@sha256:4b912c80085b88a03309aeb7907efcc29dd3342fa3952b6ea067afb1914bfe53
        - registry.redhat.io/rhel8/redis-6@sha256:d7c7852338717308cbb59e9303e1ea35cc8e5c01ceb2818569be20c15f3f943d
        - registry.redhat.io/quay/quay-rhel8@sha256:e75e4d084eba9bb65b9cc72e38294af314d49316b12a0beb75db808dc057830b
        - registry.redhat.io/quay/clair-rhel8@sha256:cbafddd1eb50c1c0f9159ffafd86f44a8c0b976194903c96c6f144bfe62e749d
        version: 3.6.10
      entries:
      - name: quay-operator.v3.6.10
        version: 3.6.10
      - name: quay-operator.v3.6.9
        version: 3.6.9
      - name: quay-operator.v3.6.8
        version: 3.6.8
      - name: quay-operator.v3.6.7
        version: 3.6.7
      - name: quay-operator.v3.6.6
        version: 3.6.6
      - name: quay-operator.v3.6.5
        version: 3.6.5
      - name: quay-operator.v3.6.4
        version: 3.6.4
      - name: quay-operator.v3.6.2
        version: 3.6.2
      - name: quay-operator.v3.6.1
        version: 3.6.1
      - name: quay-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: quay-operator.v3.7.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:a8283c2be661e07c5f2f2e5b683c520846bcb91d580d9a3a0aa27ce994ddfff6
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.7.14
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:b15335d88b240bd6bbedeb7fe26f0a1de2f6c330af9758b0fb2edab7ca69f6df
        - registry.redhat.io/quay/quay-rhel8@sha256:7bc9e115e9bace07505ffd225242cab83d5eb6d50358188d18fd9161bdf1a5a6
        - registry.redhat.io/quay/clair-rhel8@sha256:d676af83e96740220e6f18721c98817ba071fd023a62ebdc3815424e94838083
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:c54c77e09c95dc7d80d7b31e796857c8d06f2f733426629b53d3298206d23b20
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:04ac405c152bffb43b33fb2e7f45df8b3940eb21e055ebf6aa17bd81ebdeeab4
        - registry.redhat.io/rhel8/postgresql-10@sha256:a52ad402458ec8ef3f275972c6ebed05ad64398f884404b9bb8e3010c5c95291
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:a8283c2be661e07c5f2f2e5b683c520846bcb91d580d9a3a0aa27ce994ddfff6
        version: 3.7.14
      entries:
      - name: quay-operator.v3.7.14
        version: 3.7.14
      - name: quay-operator.v3.7.13
        version: 3.7.13
      - name: quay-operator.v3.7.12
        version: 3.7.12
      - name: quay-operator.v3.7.11
        version: 3.7.11
      - name: quay-operator.v3.7.10
        version: 3.7.10
      name: stable-3.7
    - currentCSV: quay-operator.v3.8.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:abec0fa7ee641c8e424e995b5fe07b98e522ad1eb92c3b90426f69941f33eb2e
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.8.15
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7e4e57071439d671971b0de46b597cf6cff8ef3879ab711125a6ae7ecb9c6d29
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:3c047c7b22fb42557bd3a9b9c9ac333f0df7d5e4a4d333a2b02a4313f57085be
        - registry.redhat.io/rhel8/postgresql-10@sha256:b7fa644277d5e406b642685ff2ace1b1de7aea2e2a1421b3cbcda7d60405f44d
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:abec0fa7ee641c8e424e995b5fe07b98e522ad1eb92c3b90426f69941f33eb2e
        - registry.redhat.io/rhel8/redis-6@sha256:444200241cb9186cd97d564d7acf077d83fc997a8ee5ee7fdae41b604895085c
        - registry.redhat.io/quay/quay-rhel8@sha256:de2b65506f99eebb5eaafca6bb42b4ebec08e9479dd0d50aa1286f197e42d6de
        - registry.redhat.io/quay/clair-rhel8@sha256:b327fcfd7472cdb9d83a564904976b26b285fa576dc0a76e5590990ec77fd871
        version: 3.8.15
      entries:
      - name: quay-operator.v3.8.15
        version: 3.8.15
      - name: quay-operator.v3.8.14
        version: 3.8.14
      - name: quay-operator.v3.8.13
        version: 3.8.13
      - name: quay-operator.v3.8.12
        version: 3.8.12
      - name: quay-operator.v3.8.11
        version: 3.8.11
      - name: quay-operator.v3.8.10
        version: 3.8.10
      - name: quay-operator.v3.8.9
        version: 3.8.9
      - name: quay-operator.v3.8.8
        version: 3.8.8
      - name: quay-operator.v3.8.7
        version: 3.8.7
      - name: quay-operator.v3.8.6
        version: 3.8.6
      - name: quay-operator.v3.8.5
        version: 3.8.5
      - name: quay-operator.v3.8.4
        version: 3.8.4
      - name: quay-operator.v3.8.3
        version: 3.8.3
      - name: quay-operator.v3.8.2
        version: 3.8.2
      - name: quay-operator.v3.8.1
        version: 3.8.1
      - name: quay-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:8a16803946272156b7aeb00eb2202ab6cb87d875fe4a7c327554f770f0f2ecb5
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.9.16
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/clair-rhel8@sha256:6c67163c783d7d9fa4d187963379885a8b76685347ecf1894c4f8ebd6656544b
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:af8ada0c88b6f19991e61ee86d04b49a148c19dacb57c282e4ab98630e7499b4
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:5a42689198701d845ff3f05502cbbebceb7e6ad7b12762ccc038c0131cf6479e
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:f87b7e7ee6c31f6ae80b08aef0e10729b21fe1b8656fe18f6dee2310bae28ec1
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:8a16803946272156b7aeb00eb2202ab6cb87d875fe4a7c327554f770f0f2ecb5
        version: 3.9.16
      entries:
      - name: quay-operator.v3.9.16
        version: 3.9.16
      - name: quay-operator.v3.9.15
        version: 3.9.15
      - name: quay-operator.v3.9.14
        version: 3.9.14
      - name: quay-operator.v3.9.13
        version: 3.9.13
      - name: quay-operator.v3.9.12
        version: 3.9.12
      - name: quay-operator.v3.9.11
        version: 3.9.11
      - name: quay-operator.v3.9.10
        version: 3.9.10
      - name: quay-operator.v3.9.9
        version: 3.9.9
      - name: quay-operator.v3.9.8
        version: 3.9.8
      - name: quay-operator.v3.9.7
        version: 3.9.7
      - name: quay-operator.v3.9.6
        version: 3.9.6
      - name: quay-operator.v3.9.5
        version: 3.9.5
      - name: quay-operator.v3.9.4
        version: 3.9.4
      - name: quay-operator.v3.9.3
        version: 3.9.3
      - name: quay-operator.v3.9.2
        version: 3.9.2
      - name: quay-operator.v3.9.1
        version: 3.9.1
      - name: quay-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: quay-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: runtime-component-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runtime-component.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/runtime-component-operator@sha256:d37b3f62575c9ad1ffcb22c89e2c3886275c326bf27aa49fb1808acba475853d
          createdAt: "2022-03-25T14:00:00Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          olm.skipRange: <0.8.1
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          This advanced Operator is capable of deploying any runtime component image with consistent, production-grade QoS. It enables enterprise architects to govern the way their applications get deployed & managed in the cluster, while dramatically reducing the learning curve for developers to deploy into Kubernetes - allowing them to focus on writing the code!
          Here are some key features:

          #### Application Lifecyle
          You can deploy your runtime component container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the Operator will watch for any updates and will re-deploy the modified image.

          #### Custom RBAC
          This Operator is capable of using a custom ServiceAccount from the caller, allowing it to follow RBAC restrictions. By default it creates a ServiceAccount if one is not specified, which can also be bound with specific roles.

          #### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as: labels, annotations, and environment variables from a ConfigMap, a Secret or a value.

          #### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configuration, such as TLS settings, are also easily enabled.

          #### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto scaling to create (and delete) instances based on resource consumption.

          #### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you.
          This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart.
          We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.

          #### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating kubernetes Secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to its required services without any intervention or interruption.

          #### Exposing metrics to Prometheus
          The Runtime Component Operator exposes the runtime container's metrics via the Prometheus Operator.
          You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `spec.monitoring` field to control configurations such as the poll interval and security credentials.

          #### Easily mount logs and transaction directories
          If you need to mount the logs and transaction data from your runtime component to an external volume such as NFS (or any storage supported in your cluster), simply add the following (to specify the volume size and the location to persist) to your RuntimeComponent CR:
          ``` storage: size: 2Gi mountPath: "/logs" ```

          #### Integration with OpenShift Serverless
          Deploy your serverless runtime component using a single toggle.  The Operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing your pod to automatically scale to 0 when it is idle.

          #### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and how they are connected.

          See our [**documentation**](https://github.com/application-stacks/runtime-component-operator/tree/main/doc/) for more information.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://github.com/application-stacks/runtime-component-operator/tree/main/doc
        - name: Repository
          url: https://github.com/application-stacks/runtime-component-operator
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: beta
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/runtime-component-operator@sha256:d37b3f62575c9ad1ffcb22c89e2c3886275c326bf27aa49fb1808acba475853d
        - registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4
        version: 0.8.1
      entries:
      - name: runtime-component.v0.8.1
        version: 0.8.1
      - name: runtime-component.v0.8.0
        version: 0.8.0
      name: beta2
    - currentCSV: runtime-component.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:1b818f1a9c27baab6a39ae42d86f78bac085d772cae5cd97b357d0ee5799cf85
          createdAt: "2023-08-28T16:46:50Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          olm.skipRange: '>=0.8.0 <1.2.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02
        - icr.io/appcafe/runtime-component-operator@sha256:1b818f1a9c27baab6a39ae42d86f78bac085d772cae5cd97b357d0ee5799cf85
        version: 1.2.2
      entries:
      - name: runtime-component.v1.2.2
        version: 1.2.2
      name: v1.2
    - currentCSV: runtime-component.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:682c5cb3e24566bf2933b1d4a4f6bf520d9b546edcedc1ee3214ec1ac4d48629
          createdAt: "2024-08-01T20:26:42Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.3.3'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/runtime-component-operator@sha256:682c5cb3e24566bf2933b1d4a4f6bf520d9b546edcedc1ee3214ec1ac4d48629
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91
        version: 1.3.3
      entries:
      - name: runtime-component.v1.3.3
        version: 1.3.3
      name: v1.3
    - currentCSV: runtime-component.v1.4.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:6f527c5173421c28d175b42d8c690eb63c4d1c5092c31d5350fd7a5e93661eaa
          createdAt: "2025-07-18T20:38:10Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.4.4'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/runtime-component-operator@sha256:6f527c5173421c28d175b42d8c690eb63c4d1c5092c31d5350fd7a5e93661eaa
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3
        version: 1.4.4
      entries:
      - name: runtime-component.v1.4.4
        version: 1.4.4
      name: v1.4
    - currentCSV: runtime-component.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:58bfceaed2fc828bce1c0e4e21a583d25f6d72f88abdc78df2510867a4f2020f
          createdAt: "2025-09-15T18:08:27Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.5.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8
        - icr.io/appcafe/runtime-component-operator@sha256:58bfceaed2fc828bce1c0e4e21a583d25f6d72f88abdc78df2510867a4f2020f
        version: 1.5.0
      entries:
      - name: runtime-component.v1.5.0
        version: 1.5.0
      name: v1.5
    defaultChannel: v1.5
    packageName: runtime-component-operator-certified
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: glance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: glance-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "glance.openstack.org/v1beta1",
                "kind": "Glance",
                "metadata": {
                  "name": "glance"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\nenabled_backends=foo:bar,foo1:bar1\n",
                  "databaseAccount": "glance",
                  "databaseInstance": "openstack",
                  "secret": "osp-secret",
                  "serviceUser": "glance",
                  "storage": {
                    "storageClass": "local-storage",
                    "storageRequest": "10G"
                  }
                }
              },
              {
                "apiVersion": "glance.openstack.org/v1beta1",
                "kind": "GlanceAPI",
                "metadata": {
                  "annotation": {
                    "keystoneEndpoint": "true"
                  },
                  "name": "default"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\nenabled_backends=foo:bar,foo1:bar1\n[foo]\nhere-foo-config\n[foo1]\nhere-foo1-config\n",
                  "databaseAccount": "glance",
                  "databaseHostname": "glance",
                  "preserveJobs": false,
                  "replicas": 1,
                  "secret": "glance-secret",
                  "serviceAccount": "glance",
                  "serviceUser": "glance",
                  "storage": {
                    "storageRequest": "10G"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:45Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GlanceAPI is the Schema for the glanceapis API
            displayName: Glance API
            kind: GlanceAPI
            name: glanceapis.glance.openstack.org
            version: v1beta1
          - description: Glance is the Schema for the glances API
            displayName: Glance
            kind: Glance
            name: glances.glance.openstack.org
            version: v1beta1
        description: Glance Operator
        displayName: Glance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Image
        - Glance
        links:
        - name: Glance Operator
          url: https://github.com/openstack-k8s-operators/glance-operator
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/glance-rhel9-operator@sha256:74f9907b59493239fbe9c5f534399ff513c693a2bda11a037aa2cc5dc8cc3e91
        - registry.redhat.io/rhoso/openstack-glance-api-rhel9@sha256:b8f84f9b432a70f7ca7aebda60455a4890347529f6bed9a440a3b197e377e0a1
        version: 1.0.6
      entries:
      - name: glance-operator.v1.0.6
        version: 1.0.6
      - name: glance-operator.v1.0.5
        version: 1.0.5
      - name: glance-operator.v1.0.4
        version: 1.0.4
      - name: glance-operator.v1.0.3
        version: 1.0.3
      - name: glance-operator.v1.0.2
        version: 1.0.2
      - name: glance-operator.v1.0.1
        version: 1.0.1
      - name: glance-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: glance-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kafka-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kafka-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kafka-controller:1.2.1
          createdAt: "2025-09-20T21:41:07Z"
          description: AWS Kafka controller is a service controller for managing Kafka
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster represents the state of an AWS kafka Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.kafka.services.k8s.aws
            version: v1alpha1
          - description: Configuration represents the state of an AWS kafka Configuration
              resource.
            displayName: Configuration
            kind: Configuration
            name: configurations.kafka.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCluster represents the state of an AWS kafka ServerlessCluster
              resource.
            displayName: ServerlessCluster
            kind: ServerlessCluster
            name: serverlessclusters.kafka.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Kafka resources in AWS from within your Kubernetes cluster.

          **About Amazon Kafka**

          Amazon Managed Streaming for Apache Kafka (Amazon MSK) is a fully managed AWS streaming data service makes it easy for you to build and run applications that use Apache Kafka to process streaming data without needing to become an expert in operating Apache Kafka clusters. Amazon MSK manages the provisioning, configuration, and maintenance of Apache Kafka clusters and Apache ZooKeeper nodes for you. Amazon MSK also shows key Apache Kafka performance metrics in the AWS Console.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Kafka Developer Resources
          url: https://aws.amazon.com/msk/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kafka maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/kafka-controller:1.2.1
        version: 1.2.1
      entries:
      - name: ack-kafka-controller.v1.2.1
        version: 1.2.1
      - name: ack-kafka-controller.v1.2.0
        version: 1.2.0
      - name: ack-kafka-controller.v1.1.3
        version: 1.1.3
      - name: ack-kafka-controller.v1.1.2
        version: 1.1.2
      - name: ack-kafka-controller.v1.1.1
        version: 1.1.1
      - name: ack-kafka-controller.v1.1.0
        version: 1.1.0
      - name: ack-kafka-controller.v1.0.14
        version: 1.0.14
      - name: ack-kafka-controller.v1.0.13
        version: 1.0.13
      - name: ack-kafka-controller.v1.0.12
        version: 1.0.12
      - name: ack-kafka-controller.v1.0.11
        version: 1.0.11
      - name: ack-kafka-controller.v1.0.10
        version: 1.0.10
      - name: ack-kafka-controller.v1.0.8
        version: 1.0.8
      - name: ack-kafka-controller.v1.0.7
        version: 1.0.7
      - name: ack-kafka-controller.v1.0.6
        version: 1.0.6
      - name: ack-kafka-controller.v1.0.5
        version: 1.0.5
      - name: ack-kafka-controller.v1.0.4
        version: 1.0.4
      - name: ack-kafka-controller.v1.0.3
        version: 1.0.3
      - name: ack-kafka-controller.v1.0.2
        version: 1.0.2
      - name: ack-kafka-controller.v1.0.1
        version: 1.0.1
      - name: ack-kafka-controller.v1.0.0
        version: 1.0.0
      - name: ack-kafka-controller.v0.0.14
        version: 0.0.14
      - name: ack-kafka-controller.v0.0.12
        version: 0.0.12
      - name: ack-kafka-controller.v0.0.11
        version: 0.0.11
      - name: ack-kafka-controller.v0.0.10
        version: 0.0.10
      - name: ack-kafka-controller.v0.0.9
        version: 0.0.9
      - name: ack-kafka-controller.v0.0.8
        version: 0.0.8
      - name: ack-kafka-controller.v0.0.7
        version: 0.0.7
      - name: ack-kafka-controller.v0.0.6
        version: 0.0.6
      - name: ack-kafka-controller.v0.0.5
        version: 0.0.5
      - name: ack-kafka-controller.v0.0.4
        version: 0.0.4
      - name: ack-kafka-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-kafka-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-acmpca-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-acmpca-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acmpca.services.k8s.aws/v1alpha1",
                "kind": "CertificateAuthority",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/acmpca-controller:1.1.1
          createdAt: "2025-09-19T20:51:07Z"
          description: AWS ACM PCA controller is a service controller for managing
            ACM PCA resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CertificateAuthority represents the state of an AWS acmpca
              CertificateAuthority resource.
            displayName: CertificateAuthority
            kind: CertificateAuthority
            name: certificateauthorities.acmpca.services.k8s.aws
            version: v1alpha1
          - description: CertificateAuthorityActivation represents the state of an
              AWS acmpca CertificateAuthorityActivation resource.
            displayName: CertificateAuthorityActivation
            kind: CertificateAuthorityActivation
            name: certificateauthorityactivations.acmpca.services.k8s.aws
            version: v1alpha1
          - description: Certificate represents the state of an AWS acmpca Certificate
              resource.
            displayName: Certificate
            kind: Certificate
            name: certificates.acmpca.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ACM PCA resources in AWS from within your Kubernetes cluster.

          **About Amazon ACM PCA**

          AWS Private CA enables creation of private certificate authority (CA) hierarchies, including root and subordinate CAs, without the investment and maintenance costs of operating an on-premises CA. Your private CAs can issue end-entity X.509 certificates useful in scenarios including:
          - Creating encrypted TLS communication channels
          - Authenticating users, computers, API endpoints, and IoT devices
          - Cryptographically signing code
          - Implementing Online Certificate Status Protocol (OCSP) for obtaining certificate revocation status

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ACM PCA
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acmpca
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ACM PCA Developer Resources
          url: https://aws.amazon.com/private-ca/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: acmpca maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/acmpca-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-acmpca-controller.v1.1.1
        version: 1.1.1
      - name: ack-acmpca-controller.v1.1.0
        version: 1.1.0
      - name: ack-acmpca-controller.v1.0.10
        version: 1.0.10
      - name: ack-acmpca-controller.v1.0.9
        version: 1.0.9
      - name: ack-acmpca-controller.v1.0.8
        version: 1.0.8
      - name: ack-acmpca-controller.v1.0.7
        version: 1.0.7
      - name: ack-acmpca-controller.v1.0.6
        version: 1.0.6
      - name: ack-acmpca-controller.v1.0.5
        version: 1.0.5
      - name: ack-acmpca-controller.v1.0.4
        version: 1.0.4
      - name: ack-acmpca-controller.v1.0.3
        version: 1.0.3
      - name: ack-acmpca-controller.v1.0.2
        version: 1.0.2
      - name: ack-acmpca-controller.v1.0.1
        version: 1.0.1
      - name: ack-acmpca-controller.v1.0.0
        version: 1.0.0
      - name: ack-acmpca-controller.v0.0.20
        version: 0.0.20
      - name: ack-acmpca-controller.v0.0.16
        version: 0.0.16
      - name: ack-acmpca-controller.v0.0.15
        version: 0.0.15
      - name: ack-acmpca-controller.v0.0.14
        version: 0.0.14
      - name: ack-acmpca-controller.v0.0.13
        version: 0.0.13
      - name: ack-acmpca-controller.v0.0.12
        version: 0.0.12
      - name: ack-acmpca-controller.v0.0.11
        version: 0.0.11
      - name: ack-acmpca-controller.v0.0.10
        version: 0.0.10
      - name: ack-acmpca-controller.v0.0.9
        version: 0.0.9
      - name: ack-acmpca-controller.v0.0.8
        version: 0.0.8
      - name: ack-acmpca-controller.v0.0.5
        version: 0.0.5
      - name: ack-acmpca-controller.v0.0.4
        version: 0.0.4
      - name: ack-acmpca-controller.v0.0.3
        version: 0.0.3
      - name: ack-acmpca-controller.v0.0.2
        version: 0.0.2
      - name: ack-acmpca-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-acmpca-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MayaData
      provider-url: ""
    name: openebs
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openebsoperator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "openebs.io/v1alpha1",
                "kind": "OpenEBSInstallTemplate",
                "metadata": {
                  "name": "oebs",
                  "namespace": "openebs"
                },
                "spec": {
                  "rbac": {
                    "create": true,
                    "pspEnabled": false,
                    "kyvernoEnabled": false
                  },
                  "serviceAccount": {
                    "create": true,
                    "name": "openebs-maya-operator"
                  },
                  "imagePullSecrets": [],
                  "release": {
                    "version": "3.0.0"
                  },
                  "legacy": {
                    "enabled": false
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": ""
                  },
                  "apiserver": {
                    "enabled": true,
                    "image": "openebs/m-apiserver",
                    "imageTag": "2.12.2",
                    "replicas": 1,
                    "ports": {
                      "externalPort": 5656,
                      "internalPort": 5656
                    },
                    "sparse": {
                      "enabled": "false"
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "defaultStorageConfig": {
                    "enabled": "true"
                  },
                  "varDirectoryPath": {
                    "baseDir": "/var/openebs"
                  },
                  "provisioner": {
                    "enabled": true,
                    "image": "openebs/openebs-k8s-provisioner",
                    "imageTag": "2.12.2",
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "patchJivaNodeAffinity": "enabled",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "localprovisioner": {
                    "enabled": true,
                    "image": "openebs/provisioner-localpv",
                    "imageTag": "3.0.0",
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "enableDeviceClass": true,
                    "enableHostpathClass": true,
                    "basePath": "/var/openebs/local",
                    "waitForBDBindTimeoutRetryCount": "12",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "snapshotOperator": {
                    "enabled": true,
                    "controller": {
                      "image": "openebs/snapshot-controller",
                      "imageTag": "2.12.2",
                      "resources": {}
                    },
                    "provisioner": {
                      "image": "openebs/snapshot-provisioner",
                      "imageTag": "2.12.2",
                      "resources": {}
                    },
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "upgradeStrategy": "Recreate",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    }
                  },
                  "ndm": {
                    "enabled": true,
                    "image": "openebs/node-disk-manager",
                    "imageTag": "1.7.0",
                    "sparse": {
                      "path": "/var/openebs/sparse",
                      "size": "10737418240",
                      "count": "0"
                    },
                    "filters": {
                      "enableOsDiskExcludeFilter": true,
                      "osDiskExcludePaths": "/,/etc/hosts,/boot",
                      "enableVendorFilter": true,
                      "excludeVendors": "CLOUDBYT,OpenEBS",
                      "enablePathFilter": true,
                      "includePaths": "",
                      "excludePaths": "/dev/loop,/dev/fd0,/dev/sr0,/dev/ram,/dev/dm-,/dev/md,/dev/rbd,/dev/zd"
                    },
                    "probes": {
                      "enableSeachest": false
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "ndmOperator": {
                    "enabled": true,
                    "image": "openebs/node-disk-operator",
                    "imageTag": "1.7.0",
                    "replicas": 1,
                    "upgradeStrategy": "Recreate",
                    "nodeSelector": {},
                    "tolerations": [],
                    "healthCheck": {
                      "initialDelaySeconds": 15,
                      "periodSeconds": 20
                    },
                    "readinessCheck": {
                      "initialDelaySeconds": 5,
                      "periodSeconds": 10
                    },
                    "resources": {}
                  },
                  "ndmExporter": {
                    "enabled": false,
                    "image": {
                      "registry": null,
                      "repository": "openebs/node-disk-exporter",
                      "pullPolicy": "IfNotPresent",
                      "tag": "1.7.0"
                    },
                    "nodeExporter": {
                      "name": "ndm-node-exporter",
                      "podLabels": {
                        "name": "openebs-ndm-node-exporter"
                      },
                      "metricsPort": 9101
                    },
                    "clusterExporter": {
                      "name": "ndm-cluster-exporter",
                      "podLabels": {
                        "name": "openebs-ndm-cluster-exporter"
                      },
                      "metricsPort": 9100
                    }
                  },
                  "webhook": {
                    "enabled": true,
                    "image": "openebs/admission-server",
                    "imageTag": "2.12.2",
                    "failurePolicy": "Fail",
                    "replicas": 1,
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "hostNetwork": false,
                    "resources": {}
                  },
                  "helper": {
                    "image": "openebs/linux-utils",
                    "helperImageTag": "3.0.0"
                  },
                  "featureGates": {
                    "enabled": true,
                    "GPTBasedUUID": {
                      "enabled": true,
                      "featureGateFlag": "GPTBasedUUID"
                    },
                    "APIService": {
                      "enabled": false,
                      "featureGateFlag": "APIService",
                      "address": "0.0.0.0:9115"
                    },
                    "UseOSDisk": {
                      "enabled": false,
                      "featureGateFlag": "UseOSDisk"
                    },
                    "ChangeDetection": {
                      "enabled": false,
                      "featureGateFlag": "ChangeDetection"
                    }
                  },
                  "crd": {
                    "enableInstall": true
                  },
                  "policies": {
                    "monitoring": {
                      "enabled": true,
                      "image": "openebs/m-exporter",
                      "imageTag": "2.12.2"
                    }
                  },
                  "analytics": {
                    "enabled": true,
                    "pingInterval": "24h"
                  },
                  "jiva": {
                    "image": "openebs/jiva",
                    "imageTag": "2.12.2",
                    "replicas": 3,
                    "defaultStoragePath": "/var/openebs",
                    "enabled": false,
                    "openebsLocalpv": {
                      "enabled": false
                    },
                    "localpv-provisioner": {
                      "openebsNDM": {
                        "enabled": false
                      }
                    }
                  },
                  "cstor": {
                    "pool": {
                      "image": "openebs/cstor-pool",
                      "imageTag": "2.12.2"
                    },
                    "poolMgmt": {
                      "image": "openebs/cstor-pool-mgmt",
                      "imageTag": "2.12.2"
                    },
                    "target": {
                      "image": "openebs/cstor-istgt",
                      "imageTag": "2.12.2"
                    },
                    "volumeMgmt": {
                      "image": "openebs/cstor-volume-mgmt",
                      "imageTag": "2.12.2"
                    },
                    "enabled": false,
                    "openebsNDM": {
                      "enabled": false
                    }
                  },
                  "openebs-ndm": {
                    "enabled": false
                  },
                  "localpv-provisioner": {
                    "enabled": false,
                    "openebsNDM": {
                      "enabled": false
                    }
                  },
                  "zfs-localpv": {
                    "enabled": false
                  },
                  "lvm-localpv": {
                    "enabled": false
                  },
                  "nfs-provisioner": {
                    "enabled": false
                  },
                  "cleanup": {
                    "image": {
                      "registry": "",
                      "repository": "bitnami/kubectl",
                      "tag": "",
                      "imagePullSecrets": []
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: index.docker.io/openebs/helm-operator:v0.0.9
          createdAt: "2021-11-02T02:56:58Z"
          description: Creates and maintains OpenEBS Control Plane deployments
          repository: https://github.com/openebs/helm-operator
          support: https://slack.openebs.io/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a OpenEBS Install Operator
            displayName: OpenEBS Install Template
            kind: OpenEBSInstallTemplate
            name: openebsinstalltemplates.openebs.io
            version: v1alpha1
        description: |
          **OpenEBS** is a leading container attached storage solution that enables the use of containers for mission-critical, persistent workloads and for other stateful workloads such as logging or Prometheus for example.
          OpenEBS itself is deployed as just another container on your host and enables storage services that can be designated on a per pod, application, cluster or container level, including:
          * Data persistence across nodes
          * Synchronization of data across availability zones and cloud providers
          * A common layer whether you are running on the cloud, or your bare metal
          * Integration with Kubernetes, so developer and application intent flows into OpenEBS
          * Management of tiering to and from S3 and other targets.
          ## OpenEBS Operator
          OpenEBS primarily provides container attached block storage (iSCSI volumes) by leveraging/aggregating the storage on the nodes, with the storage controller itself running as a container. Different storage engines (Jiva & cStor) are supported, with tools available to dynamically provision Kubernetes Local PVs. The volumes are dynamically provisioned via PersistentVolumeClaims and are managed by a control plane component called "maya", which also runs as a deployment in the K8s cluster. In addition to maya, a typical OpenEBS installation comprises several other resources, which aid with various functionalities, ranging from snapshotting to disk management. All these components are described briefly below:
          * **Maya-API-Server** - A storage orchestrator which integrates into Kubernetes workflow to help provision
            and manage OpenEBS Jiva & cStor (storage engine) volumes

          * **Dynamic-OpenEBS-Provisioner** - A Kubernetes external storage provisioner that utilizes APIs exposed by maya-apiserver
            to perform provision & delete operations of Jiva & cStor volumes

          * **Dynamic-LocalPV-Provisioner** - A dynamic provisioner for Kubernetes Local PVs
          * **OpenEBS-Snapshot-Operator** - A Kubernetes snapshot controller that creates & restores OpenEBS volume snapshots
          * **Node-Disk-Manager** - A disk management controller which identifies available disks, maintains inventory, and
            dynamically attaches/detaches disks to backend storage pods

          The helm-based OpenEBS Operator eases the setup of all the above mentioned components, with a simple custom resource provided to define the install options, thereby enabling applications to start using the OpenEBS storageclasses in their PVCs. The OpenEBSInstallTemplate CR can be used to specify start-up parameters & also update/overwrite the definitions post install.
          ## Pre-Requisites
          Before installing OpenEBS control plane, perform the following steps to ensure successful deployment of the Node-Disk-Manager & the OpenEBS volume replicas, respectively.
          * Configure the OpenEBS service account on the openshift-operators namespace/project to use the privileged security context constraint.

            **Note**: The serviceaccount name is same as the one specified in the `spec.serviceAccount.name` field of the OpenEBSInstallTemplate CR.


            ```
            oc adm policy add-scc-to-user privileged system:serviceaccount:openshift-operators:openebs-maya-operator
            ```

          * Configure the default service account on the namespace/project in which the volume replicas are deployed to use privileged
            security context constraint.


            ```
            oc adm policy add-scc-to-user privileged system:serviceaccount:openshift-operators:default
            ```

          ## Getting Started
          * Try the quickstart [guide](https://github.com/openebs/helm-operator/blob/master/olm/README.md)
          * To learn how to contribute, please read the [contribution guide](https://github.com/openebs/helm-operator/blob/master/CONTRIBUTING.md)
          * OpenEBS welcomes your feedback and contributions in any form possible. [Join our Community](https://openebs.org/community)

          ## License
          OpenEBS is distributed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt)
        displayName: OpenEBS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenEBS
        - ContainerAttachedStorage
        links:
        - name: OpenEBS Website
          url: https://openebs.io
        - name: Operator Source Code
          url: https://github.com/openebs/helm-operator
        - name: Install Instructions
          url: https://github.com/openebs/helm-operator/blob/master/olm/README.md
        maintainers:
        - email: kiran.mova@openebs.io
          name: kmova
        maturity: alpha
        minKubeVersion: 1.12.0
        provider:
          name: MayaData
        relatedImages:
        - index.docker.io/openebs/helm-operator:v0.0.9
        version: 3.0.0
      entries:
      - name: openebsoperator.v3.0.0
        version: 3.0.0
      - name: openebsoperator.v2.12.2
        version: 2.12.2
      name: alpha
    defaultChannel: alpha
    packageName: openebs
    provider:
      name: MayaData
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-core
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Core",
                "metadata": {
                  "name": "core-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configService": {
                    "port": 8080
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "helper": {
                    "port": 8080
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "licenseService": {
                    "port": 8080
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "restStatsService": {
                    "port": 8080
                  },
                  "resultService": {
                    "port": 8080,
                    "storage": "64Gi"
                  },
                  "sessionManager": {
                    "port": 8080
                  },
                  "statsService": {
                    "port": 8080
                  },
                  "testResultsService": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:34:00Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Core
            name: cores.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Core Operator
        displayName: Keysight WAP Core Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Core
          url: https://keysight-wap-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-core@sha256:001b0d08cd84a9ebc13cd8fd35b68d431b96f84d71e933c1123ed6c0d244fd37
        version: 0.1.21
      entries:
      - name: keysight-wap-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-nats-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-nats-operator.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2023-09-19T12:44:10Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NatsCluster
            name: natsclusters.nats.io
            version: v1alpha2
          - kind: NatsServiceRole
            name: natsserviceroles.nats.io
            version: v1alpha2
        description: Keysight Nats Operator
        displayName: Keysight Nats Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Nats Operator
          url: https://keysight-nats-operator.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-nats-operator@sha256:2ee5dc8814274e83a5f82e89eabd0b643e8cde8e3d555a56535a61d05c72fd75
        version: 0.1.21
      entries:
      - name: keysight-nats-operator.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-nats-operator
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: github.com/stakater-ab/hibernation-operator
    name: hibernation-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hibernation-operator.v0.1.103
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Monitoring
          createdAt: "2025-10-02T05:44:42Z"
          description: Operator to engage and disengage hibernation on k8s resources
            using cron expressions
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.1.103
          operatorframework.io/suggested-namespace: hibernation-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Cluster Resource Supervisor
            kind: ClusterResourceSupervisor
            name: clusterresourcesupervisors.hibernation.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.hibernation.stakater.com
            version: v1beta1
        description: Schedule k8s resources using cron expressions via Hibernation
          Operator
        displayName: Hibernation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - redhat
        - k8s
        - stakater
        - hibernation
        - operator
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maintainers:
        - email: hello@stakater.com
          name: StakaterAB
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Stakater
          url: github.com/stakater-ab/hibernation-operator
        relatedImages:
        - registry.connect.redhat.com/stakater/hibernation-operator@sha256:3738f82834d53aec4f8fba5caad6258c7adb4bdc23a57bde47c7ec22030fae5c
        version: 0.1.103
      entries:
      - name: hibernation-operator.v0.1.103
        version: 0.1.103
      name: release-0.1
    defaultChannel: release-0.1
    packageName: hibernation-operator
    provider:
      name: Stakater
      url: github.com/stakater-ab/hibernation-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio Project
      provider-url: ""
    name: apicurito
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: 1.0.x
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: candidate
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: latest
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: latest
    packageName: apicurito
    provider:
      name: Apicurio Project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Apache Software Foundation
      provider-url: ""
    name: camel-k
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: camel-k-operator.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.0.1
          createdAt: "2023-09-04T12:32:38Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
          - kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.0.1
        version: 2.0.1
      entries:
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: candidate
    - currentCSV: camel-k-operator.v2.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "setBody": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationProfile",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "debug"
                },
                "spec": {
                  "traits": {
                    "logging": {
                      "level": "DEBUG"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example-source"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "setBody": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "timer-to-log-example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "log-sink"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world",
                      "period": 2000
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "timer-source"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.8.0
          createdAt: "2025-09-21T09:07:43Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: IntegrationProfile is the Schema for the integrationprofiles
              API
            displayName: Integration Profile
            kind: IntegrationProfile
            name: integrationprofiles.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Pipe is the Schema for the pipes API
            displayName: Pipe
            kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers:
          the easiest way to build and manage your Camel applications on Kubernetes.

          ## Installation

          You may need to specify the container registry where to store the container images created for the Camel applications.
          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          spec:
            build:
              registry:
                address: docker.io/my-org
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step when using Openshift cluster.

          ## Running an Integration

          Run a Camel application on the cluster by creating a sample `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            annotations:
              camel.apache.org/operator.id: camel-k
            name: test
          spec:
            flows:
            - from:
                parameters:
                  period: "1000"
                steps:
                - setBody:
                    simple: Hello Camel from ${routeId}
                - log: ${body}
                uri: timer:yaml
          ```
          Log the result of the running application:
          ```
          kubectl wait --for=condition=ready integration test --timeout=180s
          kubectl logs deployment/test -f
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.8.0
        version: 2.8.0
      entries:
      - name: camel-k-operator.v2.8.0
        version: 2.8.0
      - name: camel-k-operator.v2.7.0
        version: 2.7.0
      - name: camel-k-operator.v2.6.0
        version: 2.6.0
      - name: camel-k-operator.v2.5.1
        version: 2.5.1
      - name: camel-k-operator.v2.5.0
        version: 2.5.0
      - name: camel-k-operator.v2.4.0
        version: 2.4.0
      - name: camel-k-operator.v2.3.3
        version: 2.3.3
      - name: camel-k-operator.v2.3.2
        version: 2.3.2
      - name: camel-k-operator.v2.3.1
        version: 2.3.1
      - name: camel-k-operator.v2.3.0
        version: 2.3.0
      - name: camel-k-operator.v2.2.0
        version: 2.2.0
      - name: camel-k-operator.v2.1.0
        version: 2.1.0
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: latest
    - currentCSV: camel-k-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.12.0
          createdAt: "2023-02-27T09:01:55Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.12.0
        version: 1.12.0
      entries:
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable
    - currentCSV: camel-k-operator.v1.10.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.10.4
          createdAt: "2023-01-28T13:49:05Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.10.4
        version: 1.10.4
      entries:
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.10
    - currentCSV: camel-k-operator.v1.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.11.1
          createdAt: "2023-01-30T08:45:04Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.11.1
        version: 1.11.1
      entries:
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.11
    - currentCSV: camel-k-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.12.1
          createdAt: "2023-06-12T08:46:58Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.12.1
        version: 1.12.1
      entries:
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.12
    - currentCSV: camel-k-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.5.1
          createdAt: "2021-08-31T09:03:41Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.5.1
        version: 1.5.1
      entries:
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.5
    - currentCSV: camel-k-operator.v1.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.6.1
          createdAt: "2021-10-27T21:06:20Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.6.1
        version: 1.6.1
      entries:
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.6
    - currentCSV: camel-k-operator.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.7.0
          createdAt: "2021-11-16T10:54:12Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.8.1+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.7.0
        version: 1.7.0
      entries:
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.7
    - currentCSV: camel-k-operator.v1.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.8.2
          createdAt: "2022-03-07T07:13:19Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.8.2
        version: 1.8.2
      entries:
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.8
    - currentCSV: camel-k-operator.v1.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.9.2
          createdAt: "2022-05-23T09:05:22Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.9.2
        version: 1.9.2
      entries:
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.9
    - currentCSV: camel-k-operator.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.0.1
          createdAt: "2023-09-04T12:32:38Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
          - kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.0.1
        version: 2.0.1
      entries:
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-2.0
    - currentCSV: camel-k-operator.v2.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "setBody": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationProfile",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "debug"
                },
                "spec": {
                  "traits": {
                    "logging": {
                      "level": "DEBUG"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example-source"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "setBody": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "timer-to-log-example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "log-sink"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world",
                      "period": 2000
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "timer-source"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.8.0
          createdAt: "2025-09-21T09:07:43Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: IntegrationProfile is the Schema for the integrationprofiles
              API
            displayName: Integration Profile
            kind: IntegrationProfile
            name: integrationprofiles.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Pipe is the Schema for the pipes API
            displayName: Pipe
            kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers:
          the easiest way to build and manage your Camel applications on Kubernetes.

          ## Installation

          You may need to specify the container registry where to store the container images created for the Camel applications.
          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          spec:
            build:
              registry:
                address: docker.io/my-org
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step when using Openshift cluster.

          ## Running an Integration

          Run a Camel application on the cluster by creating a sample `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            annotations:
              camel.apache.org/operator.id: camel-k
            name: test
          spec:
            flows:
            - from:
                parameters:
                  period: "1000"
                steps:
                - setBody:
                    simple: Hello Camel from ${routeId}
                - log: ${body}
                uri: timer:yaml
          ```
          Log the result of the running application:
          ```
          kubectl wait --for=condition=ready integration test --timeout=180s
          kubectl logs deployment/test -f
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.8.0
        version: 2.8.0
      entries:
      - name: camel-k-operator.v2.8.0
        version: 2.8.0
      - name: camel-k-operator.v2.7.0
        version: 2.7.0
      - name: camel-k-operator.v2.6.0
        version: 2.6.0
      - name: camel-k-operator.v2.5.1
        version: 2.5.1
      - name: camel-k-operator.v2.5.0
        version: 2.5.0
      - name: camel-k-operator.v2.4.0
        version: 2.4.0
      - name: camel-k-operator.v2.3.3
        version: 2.3.3
      - name: camel-k-operator.v2.3.2
        version: 2.3.2
      - name: camel-k-operator.v2.3.1
        version: 2.3.1
      - name: camel-k-operator.v2.3.0
        version: 2.3.0
      - name: camel-k-operator.v2.2.0
        version: 2.2.0
      - name: camel-k-operator.v2.1.0
        version: 2.1.0
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-v2
    defaultChannel: stable-v2
    packageName: camel-k
    provider:
      name: The Apache Software Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-cpd-ccs
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-ccs.v10.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ccs.cpd.ibm.com/v1beta1",
                "kind": "CCS",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ccs-cr",
                    "app.kubernetes.io/managed-by": "ibm-cpd-ccs-operator",
                    "app.kubernetes.io/name": "ibm-cpd-ccs-operator"
                  },
                  "name": "ccs-cr"
                },
                "spec": {
                  "scaleConfig": "small",
                  "license": {
                    "accept": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
          createdAt: "2025-04-18"
          description: Common Core Services
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-ccs
          olm.skipRange: '>=7.0.0 <10.3.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM watsonx.ai"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: |-
            [
              "runtimeassemblies.runtimes.ibm.com"
            ]
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation Runtime Manager CRD. This product is not deployable
              standalone and it is part of the LICCR already accepted.  For additional
              details regarding install parameters check: http://ibm.biz/-readme.
              License By installing this product you accept the license terms https://ibm.biz/BdaaLC
              or https://ibm.biz/BdaaL9.'
            kind: RuntimeAssembly
            name: runtimeassemblies.runtimes.ibm.com
            version: v1alpha1
          - description: 'Documentation This product is not deployable standalone
              and it is part of the LICCR already accepted.  For additional details
              regarding install parameters check: http://ibm.biz/-readme. License
              By installing this product you accept the license terms https://ibm.biz/BdaaLC
              or https://ibm.biz/BdaaL9.'
            kind: CCS
            name: ccs.ccs.cpd.ibm.com
            version: v1beta1
        description: |-
          # Introduction Common Core Services Operator
          Put stuff here

          # Details
          Various Details
          ## Prerequisites
          Various Prerequisites
          ## Resources Required
          Various Resources Required
          ## Installing
          Various Installation details
          # Configuration
          Various Configuration information
          ### Limitations
          Various Limitations
          ### SecurityContextConstraints Requirements
          All pods will use the restricted SCC, I think.
        displayName: Common Core Services
        installModes:
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: OwnNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commonc-core-services
        - cpd40
        links:
        - name: Ccs
          url: https://ccs.domain
        maintainers:
        - email: test@ibm.com
          name: test
        maturity: v1
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/jobs-api@sha256:173b88f90078e1817686ac949f48387af70a706e23e6fdab4a7c6217db5ad86e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-bundle@sha256:94c7a1dd290e2abd55c5b8998cf08ad174d3fbfc659ae6fa6fb6cbab61d5fe14
        - registry.marketplace.redhat.com/cp/cpd/dap-base-job@sha256:f208686a3f56d2980e7b8640b6ebdea9057109e7520c6817a731e68d08506a36
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:bcdf9b7eea8753b3e463973a2690140ae7aa8c02c3c7726511f116bd751ec2d8
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:c268d2852580286e5c6770857fcfc3b06a698521f5acd17ed83b8e2cb9ab55cf
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:523dcc671f1c05e690d5e578a90300d28d76570f605d4de7f80645b590f4a462
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:81d8b25221d827160dd5e88efd383453f532057ad02d7a2ee3dcd9c6eb16ba65
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:0e0ae96aff8540323da8a2379af014d3af1770e4b0043b84b9b415dda44501f7
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:6f2e7cc917adb54b3112dc4909f645ac08c33e8c2c33266fbc8f5133cbcd1973
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:21c81ff7c4b8809accaf62b242fd9b3d52f235489ae427556b7f71cc21bf8926
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:351375de35d7fc9d940b905b867ce8a00b7ff1416e70d3a100918797fe3d8c56
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:4fb2127ba7c525a959bbbcf024de364ef266ebb3ca7e696f74d990d2552cda59
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:56315ce0a0a70f5899191a7eab856105489143f0e889c8cdd1b6b8c225f02eb7
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:a3a0d58f895fca4d832e602d0dee80018df4ad7bad8800ad4058eea07fda4817
        - registry.marketplace.redhat.com/cp/cpd/ngp-projects-api@sha256:df420e9f616ad8a123d9eedd117fb6cb007b30a772138db534f47ad8801fa96d
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:257bf301338f7b123099e2ffe31c3f3b793bf28d6a0570a25b889271239579ba
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:6f96b8b6afb00cb26fcf903b8e4c30b0d6344bb86d8bce2f2026fc8cb0b575ec
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:0621eb9b83c9b2b8bfe5cd09797fe08fa37a854a5506914935ff6f0df599b631
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:0ae22914dbaf09d6679bd06bc9b6a5728e744c9ce2309cbd5f01e28308c403ea
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:562281dbb2bec5191a7c6b595ab853092e6fe3ff1432e77b2471f71ccdffcbc8
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:54c86367bfe0ab658611856cc0a6c5de24c14a336bd129ac9f16293945a55238
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:f5527ddbbd0b9878555fce852996b13020fd91b1604d8c03d34a61ca2fdb2d1d
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:fbc6eab07fcfc2d0c1895f18e17f751e9481ad111875959e9ff08e14f81b9a35
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:740a1d87a43e61e32cfdea032d852af33100d959c8b57ff2f165f4baf77fc011
        - registry.marketplace.redhat.com/cp/cpd/task-credentials@sha256:e162828d5ad07ceae4db0589f930979842714e3b766552102bee373a3a570af3
        - registry.marketplace.redhat.com/cp/cpd/environments-ui@sha256:2923664475c7a8ea5e63055c90748db37825fe76c7975b845f1185d4b32ee4ea
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/cpd/runtime-manager-api@sha256:a37773df07fe757dd698e3f09f8687902e7b3592d97855e703dd92ecc832aab0
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:8a6ce0dada706056efda73ad0cc1547566003e0a7aab8ed379b50e4776861b94
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e12599ff99db058fdc8f9c1eebfba31160c636ff092138deb1f6f6705f44fd62
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:3f2f97e6b2b18526255e88ff02d1039b24312eca51ee22b7cbf76d5680f7aee1
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:d023f35b65816f244b6d699b41362e6031553f7d4c416efcee4a7f2395a61355
        - registry.marketplace.redhat.com/cp/cpd/spaces-ui-job@sha256:95dade51444e30ec034c12207ecbef675b70889b81970f49f10aa75160b2f947
        - registry.marketplace.redhat.com/cp/cpd/environment-specifications-sync@sha256:894b4624d8ead7286cf9068c6477300992c32d96339644fe764bcb29b1c8ef5c
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:ad19b65042fb26cc1244a92f81a96e5ceb4313f8c823181efe28b66e9fecf9c8
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:044b663e4b7508c62dbb28b7fb2aa267c9771fc62ab509b08d46ea1499d95431
        - registry.marketplace.redhat.com/cp/cpd/environments-api@sha256:6dd60b530f81d7a2d55a534211f7f128814abf87b8cd3575d528dba53442c44e
        - registry.marketplace.redhat.com/cp/cpd/jobs-ui@sha256:49daf607918b378593485da16ac417056ee295bca94f1a5cac37b334381ae096
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:6cffef7631ee54612fba71efc0eff4e17bc08defc3e2137aee9881d685444298
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:722944bfd0eba280255f4c198f65f562423a26e2c4ad3580591ee71cb6e9c2ef
        - registry.marketplace.redhat.com/cp/cpd/job-trigger@sha256:f2ea487072393a24e4c6864b2107c1fce072ebba07f9dbc2c139bd37fbd062da
        - registry.marketplace.redhat.com/cp/cpd/environments-init-job@sha256:6af8f900af7c11eee6894e2c3e52f740b3c700dac0ee753fadef330230bd42dd
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:6fe90f051d9261ad224dccf4d9f6bc9f19ce0e8d22fa010a46b2752374bce2b3
        version: 10.3.0
      entries:
      - name: ibm-cpd-ccs.v10.3.0
        version: 10.3.0
      name: v10.3
    defaultChannel: v10.3
    packageName: ibm-cpd-ccs
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: swift-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: swift-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "swift.openstack.org/v1beta1",
                "kind": "Swift",
                "metadata": {
                  "name": "swift"
                },
                "spec": {
                  "swiftProxy": {
                    "passwordSelectors": {
                      "service": "SwiftPassword"
                    },
                    "replicas": 1
                  },
                  "swiftRing": {
                    "ringReplicas": 1
                  },
                  "swiftStorage": {
                    "replicas": 1,
                    "storageClass": "local-storage"
                  }
                }
              },
              {
                "apiVersion": "swift.openstack.org/v1beta1",
                "kind": "SwiftProxy",
                "metadata": {
                  "name": "swift-proxy"
                },
                "spec": {
                  "passwordSelectors": {
                    "service": "SwiftPassword"
                  },
                  "replicas": 1,
                  "secret": "osp-secret",
                  "serviceUser": "swift"
                }
              },
              {
                "apiVersion": "swift.openstack.org/v1beta1",
                "kind": "SwiftRing",
                "metadata": {
                  "name": "swift-ring"
                },
                "spec": {
                  "ringReplicas": 1
                }
              },
              {
                "apiVersion": "swift.openstack.org/v1beta1",
                "kind": "SwiftStorage",
                "metadata": {
                  "name": "swift-storage"
                },
                "spec": {
                  "replicas": 1,
                  "storageClass": "local-storage",
                  "storageRequest": "10G"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:34Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SwiftProxy is the Schema for the swiftproxies API
            displayName: Swift Proxy
            kind: SwiftProxy
            name: swiftproxies.swift.openstack.org
            version: v1beta1
          - description: SwiftRing is the Schema for the swiftrings API
            displayName: Swift Ring
            kind: SwiftRing
            name: swiftrings.swift.openstack.org
            version: v1beta1
          - description: Swift is the Schema for the swifts API
            displayName: Swift
            kind: Swift
            name: swifts.swift.openstack.org
            version: v1beta1
          - description: SwiftStorage is the Schema for the swiftstorages API
            displayName: Swift Storage
            kind: SwiftStorage
            name: swiftstorages.swift.openstack.org
            version: v1beta1
        description: OpenStack Swift operator
        displayName: Swift operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Swift
        - Object Storage
        links:
        - name: Swift Operator
          url: https://github.com/openstack-k8s-operators/swift-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-swift-object-rhel9@sha256:656f23b25ca0bd80d7b77ba717a21455366dfaf2049df1bd6d6536925a18b94b
        - registry.redhat.io/rhoso/openstack-swift-proxy-server-rhel9@sha256:6e654db83d570a3f8feedb912d5d9dff35ecba4d53169481a3b4982d3a832798
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/swift-rhel9-operator@sha256:18207f02c049447836211b326f09ce660782c384cda92e592db256692e3c8e73
        - registry.redhat.io/rhoso/openstack-swift-account-rhel9@sha256:ecebf3eb430188392f7a7b8cda3d1c0c0f658bcbbaff531d99eba88a7e0c9889
        - registry.redhat.io/rhoso/openstack-swift-container-rhel9@sha256:782866f8e9661de50d6a74bc52262f4546d7e0055d0c4da83ba6b891a50c35ea
        version: 1.0.6
      entries:
      - name: swift-operator.v1.0.6
        version: 1.0.6
      - name: swift-operator.v1.0.5
        version: 1.0.5
      - name: swift-operator.v1.0.4
        version: 1.0.4
      - name: swift-operator.v1.0.3
        version: 1.0.3
      - name: swift-operator.v1.0.2
        version: 1.0.2
      - name: swift-operator.v1.0.1
        version: 1.0.1
      - name: swift-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: swift-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/component: kmm
      app.kubernetes.io/name: kmm
      app.kubernetes.io/part-of: kmm
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: kernel-module-management
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kernel-module-management.v2.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:1022205d712a2ed9d6e76968961c180d4e66a2e0fb991221b0b02ed94ed830d2
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.0.2'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel-module-management
        - out-of-tree
        - kernel
        - kmm
        - module
        - modules
        - oot
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:1022205d712a2ed9d6e76968961c180d4e66a2e0fb991221b0b02ed94ed830d2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62c44057a27abd41cf3daf60b8736da59a82dadb65fa6c32bef65d19bd49ea49
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:25f39f20f325f64676bf8ef14cd8a60cfbdc77116753a89d4bb667ea2a86172c
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:192e08b121a04ae69aea60c35199e8f5e3c42305e1ebd024a605a371de4ab472
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:94b4ddc45cb34d4ffe794a51b99c868cb709f90cc4eb8b28284d70c526a0a226
        version: 2.0.2
      entries:
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.0
    - currentCSV: kernel-module-management.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a85037778c53c3766f39efe4c4d2523d01ae7743e5c5693b05bd8b2cbb50b1d8
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.1.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a85037778c53c3766f39efe4c4d2523d01ae7743e5c5693b05bd8b2cbb50b1d8
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:611556d999ea97f9dbb78fec5963baa02de7a26626d7096ba63b177e3bcf9707
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:14e100e5853444114029116255ebfb7a54e5ed51503ba969740ad48b5ca80f70
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:1a9a19df17b464d1708b2905d03b8726e4480ca1f11aa97a15ff4c049a24ea4b
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:ea3e5fbf7eaf6bd29031e2876d7e24859779fb4d3d7ef27c26a1d310f137ad17
        version: 2.1.1
      entries:
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.1
    - currentCSV: kernel-module-management.v2.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a224708950f8defb259ef0da49011d361c5c0378404350640600c6e8fbdbd606
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.2.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:b1d98c0f14448ae8eb80821cbec236150c338674aae2a2b37f8d0ab0d5f1430c
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:53fddf0b226eb579c71c86112a8d9b6b1d90a1498d6f6985932b16b8be532f35
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:c3dc9ad4ebc68cd85a074b520903a9310e72a90a73bdf4aa7c42b7ed66b7bcba
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:ce1d68fd96f5003285f9fbc0c7abd8aa23830dbdd9b76629d38d4d2238e30032
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a224708950f8defb259ef0da49011d361c5c0378404350640600c6e8fbdbd606
        version: 2.2.1
      entries:
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.2
    - currentCSV: kernel-module-management.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:750dcc75f908cdda86932386fa1a28edaf65d7b5379ef6d0447586a064db906f
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.3.0'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:750dcc75f908cdda86932386fa1a28edaf65d7b5379ef6d0447586a064db906f
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:c5005b74402f680d11c9ea682a9a0d3f74b6ab0d01d25298d232f910f1a29c34
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:055c2d6e80bd1ec225f0cb2041572ebb28cfd1a4ed82ba47ff0d638d52ded2a8
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:f9fde334af68220943017a7491130e6a82f15e63522f5bb3931aa7d9b00ccc40
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:06ac089cc89b303e525fd71d6439b958f34ab944a72209f1cc8cb562ecfd1112
        version: 2.3.0
      entries:
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.3
    - currentCSV: kernel-module-management.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:41Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:8e0a3cfb1cb92d6d3c9ab5a3030593d8df1378c3ce755d598fb77b08ea5793d3
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        version: 2.4.1
      entries:
      - name: kernel-module-management.v2.4.1
        version: 2.4.1
      - name: kernel-module-management.v2.4.0
        version: 2.4.0
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.4
    - currentCSV: kernel-module-management.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:41Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:8e0a3cfb1cb92d6d3c9ab5a3030593d8df1378c3ce755d598fb77b08ea5793d3
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        version: 2.4.1
      entries:
      - name: kernel-module-management.v2.4.1
        version: 2.4.1
      - name: kernel-module-management.v2.4.0
        version: 2.4.0
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: kernel-module-management
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ab Initio
      provider-url: ""
    name: abinitio-runtime-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: abinitio-runtime-operator.v4.3.3-1
      currentCSVDesc:
        annotations:
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: abinitio-system
          operators.operatorframework.io/builder: operator-sdk-v1.8.0-ocp
          operators.operatorframework.io/internal-objects: '["cooperatingsystemruntimepoolmetrics.cloud.abinitio.com",
            "cooperatingsystemruntimeprofiles.cloud.abinitio.com"]'
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Defines a *runtime claim*, an interface for reserving a runtime
              object (CoOperatingSystemRuntime) in a runtime pool (CoOperatingSystemRuntimePool)
              in order to launch a job.
            displayName: Co>Operating System Runtime Claim
            kind: CoOperatingSystemRuntimeClaim
            name: cooperatingsystemruntimeclaims.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Pool Metric
            kind: CoOperatingSystemRuntimePoolMetric
            name: cooperatingsystemruntimepoolmetrics.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime pool*, a set of templated runtime objects
              (CoOperatingSystemRuntime) that are available on demand for running
              jobs. A runtime pool's capacity is configurable and can be scheduled
              in response to usage patterns.
            displayName: Co>Operating System Runtime Pool
            kind: CoOperatingSystemRuntimePool
            name: cooperatingsystemruntimepools.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Profile
            kind: CoOperatingSystemRuntimeProfile
            name: cooperatingsystemruntimeprofiles.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime object*, a set of coordinated resources
              used to run a data processing job in parallel across a launcher pod
              and a dynamically scaled set of worker pods. Such a job leverages the
              full power of the Co>Operating System for processing complex enterprise
              data.
            displayName: Co>Operating System Runtime
            kind: CoOperatingSystemRuntime
            name: cooperatingsystemruntimes.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime template*, a reusable specification (CoOperatingSystemRuntimeSpec)
              that can be referenced and overridden by runtime objects as needed,
              making it possible to specify jobs concisely. The runtime objects in
              a runtime pool (CoOperatingSystemRuntimePool) are created from a runtime
              template.
            displayName: Co>Operating System Runtime Template
            kind: CoOperatingSystemRuntimeTemplate
            name: cooperatingsystemruntimetemplates.cloud.abinitio.com
            version: v1
        description: |-
          The Co>Operating System runtime operator brings the power of Ab Initio's Co>Operating
          System to Kubernetes by defining a set of custom API resources for provisioning and
          running data processing jobs.

          The operator is installed along with the following custom resource definitions (CRDs):

          - CoOperatingSystemRuntime
          - CoOperatingSystemRuntimeClaim
          - CoOperatingSystemRuntimePool
          - CoOperatingSystemRuntimeTemplate

          In combination, these resources give you significant flexibility in scheduling, running,
          and monitoring your jobs, which can run in parallel across a dynamically scaled set of
          containers. For reference documentation for these resources, see
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-FOUR-Reference/Co-Operating-System-runtime-operator-API-reference)
          in the Ab Initio Help.

          The following topics describe the basic requirements and procedures for creating custom
          resources for running jobs after you have installed the runtime operator.

          ## Requirements for using custom resources

          The runtime operator is fully configured when you install it in your OpenShift cluster
          through OperatorHub. Importantly, the operator is installed in unrestricted mode,
          which means it can manage the CoOperatingSystemRuntime family of CRDs in any namespace
          in your cluster.

          For creating custom resources in order to run jobs using the operator, you must meet
          the requirements and configure the settings described in the following sections.

          ### Container images for job agents

          To run jobs of kind CoOperatingSystemRuntime, you must have a way to build the following
          OCI container images:

          - A *processing* image with the Co>Operating System server software installed in it.
          - A *sandbox workspace* image with the graphs and other files needed to run your jobs.
          This image is based on the Co>Operating System image and functions as the job's launcher
          agent, which coordinates with a set of worker agents (processing containers) in order to
          run a job in parallel.

          For building and managing these container images, you need the following:

          - A Co>Operating System installation package.
          - Access to a technical repository containing the projects that make up your application.
          - A container-building tool.
          - Access to a private container registry for hosting the images you build. You need pull
          and push access to this registry from your container build context and pull access from
          your OpenShift cluster (for example, via a Kubernetes Secret).

          > WARNING! Never push container images with Ab Initio software to a public registry. Push
          them to a private container registry only.

          The Co>Operating System includes reference Dockerfiles and example scripts for building
          container images. Typically, you build these images as part of your CI/CD pipeline. For
          more information, contact your Ab Initio administrator or Ab Initio Support.

          ### Ab Initio key servers

          The key servers that you use to enable jobs must be reachable from your cluster. You
          specify these key servers via the `AB_HOSTNAME_KEYSERVER_URLS` configuration variable in
          the pod specifications for launcher and worker agents in the CoOperatingSystemRuntime
          resource.

          To discuss the software activation key requirements for jobs that run in Kubernetes,
          contact your key server administrator or Ab Initio Support.

          ### Namespace and role considerations

          The runtime operator is installed with the cluster-wide RBAC roles and role bindings that
          it needs in order to manage custom resources and run jobs.

          You typically use a separate namespace or namespaces for creating and managing custom
          resources to run jobs. Each of these namespaces must be equipped with a service
          account that can list pod limits and get and patch CoOperatingSystemRuntime resources.

          For each namespace in which you intend to create custom resources, create the needed RBAC
          role and role binding by applying the following Role and RoleBinding manifests:

          ```yaml
          apiVersion: rbac.authorization.k8s.io/v1
          kind: Role
          metadata:
            name: cooperatingsystemruntime-role
          rules:
          - apiGroups:
            - ""
            resources:
            - limitranges
            verbs:
            - list
          - apiGroups:
            - cloud.abinitio.com
            resources:
            - cooperatingsystemruntimes
            verbs:
            - get
            - patch
          ---
          apiVersion: rbac.authorization.k8s.io/v1
          kind: RoleBinding
          metadata:
            name: run-cooperatingsystemruntime-rolebinding
          subjects:
          - kind: ServiceAccount
            name: default
          roleRef:
            kind: Role
            name: cooperatingsystemruntime-role
            apiGroup: rbac.authorization.k8s.io
          ```

          One way to do this is to save these manifests to a file and specify a namespace when
          using `oc` to apply them to the cluster. In the following example, the manifests are
          saved to a file named `runtime-rbac.yaml` and the role and role binding are to be created
          for the `default` service account in the `hello-world` namespace:

          ```
          oc -n hello-world apply -f runtime-rbac.yaml
          ```

          To bind the role to a service account other than `default`, simply update the value of
          `subjects[0].name` in the provided RoleBinding manifest.

          ### Dynamic volume provisioning

          The launcher and worker template sections of the CoOperatingSystemRuntime resource
          require you to specify persistent volume claims (PVCs) for job recovery data. To ensure
          that these PVCs can be bound to approriate storage volumes, dynamic volume provisioning
          must be enabled in your cluster.

          Work with your cluster administrator to identify the storage needs of your jobs and to
          ensure that dynamic volume provisioning is enabled so that PVCs can be matched to
          persistent volumes of the right storage class.

          For more information about dynamic storage provisioning, see the
          [OpenShift documentation](https://docs.openshift.com/container-platform/latest/storage/dynamic-provisioning.html).

          ## Quick start and examples

          For an overview of the key concepts associated with running jobs in Kubernetes, see
          ["About running jobs in Kubernetes"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-TWO-Running-jobs-in-Kubernetes/About-running-jobs-in-Kubernetes)
          in the Ab Initio Help.

          As described in this topic, you might begin by creating a CoOperatingSystemRuntime
          manifest that defines the configuration and compute resources needed for a standalone test
          job. You could then use the specification for this job to create a CoOperatingSystemRuntimeTemplate
          resource, which would collect common settings and make it possible to concisely specify
          CoOperatingSystemRuntime jobs.

          You could also use a runtime pool (of kind CoOperatingSystemRuntime) to extend these basic
          resources: such a pool consists of an actively running group of templated runtime objects whose
          pods can be used to run jobs on demand. Finally, for reserving resources in a pool, you could
          create a CoOperatingSystemRuntimeClaim resource.

          For more information about these custom resources, refer to the
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-FOUR-Reference/Co-Operating-System-runtime-operator-API-reference).

          For a set of integrated examples (including complete YAML manifests) that show you how to
          create and use the custom resources, see ["Example manifests for running jobs"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-TWO-Running-jobs-in-Kubernetes/Configuring-and-running-Co-Operating-System-jobs-in-Kubernetes/Example-manifests-for-CoOperatingSystemRuntime-jobs).
        displayName: Co>Operating System Runtime Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - abinitio
        - runtime
        - operator
        links:
        - name: Ab Initio
          url: https://www.abinitio.com/
        maintainers:
        - email: support@abinitio.com
          name: Ab Initio Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Ab Initio
        relatedImages:
        - registry.connect.redhat.com/abinitio/cooperating-system-runtime-operator@sha256:665efa8cc2ba97ad31f219161cdd53e15c1aef16b6c05ac60506be9bbc9ffc9a
        version: 4.3.3-1
      entries:
      - name: abinitio-runtime-operator.v4.3.3-1
        version: 4.3.3-1
      - name: abinitio-runtime-operator.v4.3.1-0
        version: 4.3.1-0
      name: release-4.3
    - currentCSV: abinitio-runtime-operator.v4.2.3-0
      currentCSVDesc:
        annotations:
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: abinitio-system
          operators.operatorframework.io/builder: operator-sdk-v1.8.0-ocp
          operators.operatorframework.io/internal-objects: '["cooperatingsystemruntimepoolmetrics.cloud.abinitio.com",
            "cooperatingsystemruntimeprofiles.cloud.abinitio.com"]'
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Defines a *runtime claim*, an interface for reserving a runtime
              object (CoOperatingSystemRuntime) in a runtime pool (CoOperatingSystemRuntimePool)
              in order to launch a job.
            displayName: Co>Operating System Runtime Claim
            kind: CoOperatingSystemRuntimeClaim
            name: cooperatingsystemruntimeclaims.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Pool Metric
            kind: CoOperatingSystemRuntimePoolMetric
            name: cooperatingsystemruntimepoolmetrics.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime pool*, a set of templated runtime objects
              (CoOperatingSystemRuntime) that are available on demand for running
              jobs. A runtime pool's capacity is configurable and can be scheduled
              in response to usage patterns.
            displayName: Co>Operating System Runtime Pool
            kind: CoOperatingSystemRuntimePool
            name: cooperatingsystemruntimepools.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Profile
            kind: CoOperatingSystemRuntimeProfile
            name: cooperatingsystemruntimeprofiles.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime object*, a set of coordinated resources
              used to run a data processing job in parallel across a launcher pod
              and a dynamically scaled set of worker pods. Such a job leverages the
              full power of the Co>Operating System for processing complex enterprise
              data.
            displayName: Co>Operating System Runtime
            kind: CoOperatingSystemRuntime
            name: cooperatingsystemruntimes.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime template*, a reusable specification (CoOperatingSystemRuntimeSpec)
              that can be referenced and overridden by runtime objects as needed,
              making it possible to specify jobs concisely. The runtime objects in
              a runtime pool (CoOperatingSystemRuntimePool) are created from a runtime
              template.
            displayName: Co>Operating System Runtime Template
            kind: CoOperatingSystemRuntimeTemplate
            name: cooperatingsystemruntimetemplates.cloud.abinitio.com
            version: v1
        description: |-
          The Co>Operating System runtime operator brings the power of Ab Initio's Co>Operating
          System to Kubernetes by defining a set of custom API resources for provisioning and
          running data processing jobs.

          The operator is installed along with the following custom resource definitions (CRDs):

          - CoOperatingSystemRuntime
          - CoOperatingSystemRuntimeClaim
          - CoOperatingSystemRuntimePool
          - CoOperatingSystemRuntimeTemplate

          In combination, these resources give you significant flexibility in scheduling, running,
          and monitoring your jobs, which can run in parallel across a dynamically scaled set of
          containers. For reference documentation for these resources, see
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Co_Operating_System_runtime_operator_API_reference)
          in the Ab Initio Help.

          The following topics describe the basic requirements and procedures for creating custom
          resources for running jobs after you have installed the runtime operator.

          ## Requirements for using custom resources

          The runtime operator is fully configured when you install it in your OpenShift cluster
          through OperatorHub. Importantly, the operator is installed in unrestricted mode,
          which means it can manage the CoOperatingSystemRuntime family of CRDs in any namespace
          in your cluster.

          For creating custom resources in order to run jobs using the operator, you must meet
          the requirements and configure the settings described in the following sections.

          ### Container images for job agents

          To run jobs of kind CoOperatingSystemRuntime, you must have a way to build the following
          OCI container images:

          - A *processing* image with the Co>Operating System server software installed in it.
          - A *sandbox workspace* image with the graphs and other files needed to run your jobs.
          This image is based on the Co>Operating System image and functions as the job's launcher
          agent, which coordinates with a set of worker agents (processing containers) in order to
          run a job in parallel.

          For building and managing these container images, you need the following:

          - A Co>Operating System installation package.
          - Access to a technical repository containing the projects that make up your application.
          - A container-building tool.
          - Access to a private container registry for hosting the images you build. You need pull
          and push access to this registry from your container build context and pull access from
          your OpenShift cluster (for example, via a Kubernetes Secret).

          > WARNING! Never push container images with Ab Initio software to a public registry. Push
          them to a private container registry only.

          The Co>Operating System includes reference Dockerfiles and example scripts for building
          container images. Typically, you build these images as part of your CI/CD pipeline. For
          more information, contact your Ab Initio administrator or Ab Initio Support.

          ### Ab Initio key servers

          The key servers that you use to enable jobs must be reachable from your cluster. You
          specify these key servers via the `AB_HOSTNAME_KEYSERVER_URLS` configuration variable in
          the pod specifications for launcher and worker agents in the CoOperatingSystemRuntime
          resource.

          To discuss the software activation key requirements for jobs that run in Kubernetes,
          contact your key server administrator or Ab Initio Support.

          ### Namespace and role considerations

          The runtime operator is installed with the cluster-wide RBAC roles and role bindings that
          it needs in order to manage custom resources and run jobs.

          You typically use a separate namespace or namespaces for creating and managing custom
          resources to run jobs. Each of these namespaces must be equipped with a service
          account that can list pod limits and get and patch CoOperatingSystemRuntime resources.

          For each namespace in which you intend to create custom resources, create the needed RBAC
          role and role binding by applying the following Role and RoleBinding manifests:

          ```yaml
          apiVersion: rbac.authorization.k8s.io/v1
          kind: Role
          metadata:
            name: cooperatingsystemruntime-role
          rules:
          - apiGroups:
            - ""
            resources:
            - limitranges
            verbs:
            - list
          - apiGroups:
            - cloud.abinitio.com
            resources:
            - cooperatingsystemruntimes
            verbs:
            - get
            - patch
          ---
          apiVersion: rbac.authorization.k8s.io/v1
          kind: RoleBinding
          metadata:
            name: run-cooperatingsystemruntime-rolebinding
          subjects:
          - kind: ServiceAccount
            name: default
          roleRef:
            kind: Role
            name: cooperatingsystemruntime-role
            apiGroup: rbac.authorization.k8s.io
          ```

          One way to do this is to save these manifests to a file and specify a namespace when
          using `oc` to apply them to the cluster. In the following example, the manifests are
          saved to a file named `runtime-rbac.yaml` and the role and role binding are to be created
          for the `default` service account in the `hello-world` namespace:

          ```
          oc -n hello-world apply -f runtime-rbac.yaml
          ```

          To bind the role to a service account other than `default`, simply update the value of
          `subjects[0].name` in the provided RoleBinding manifest.

          ### Dynamic volume provisioning

          The launcher and worker template sections of the CoOperatingSystemRuntime resource
          require you to specify persistent volume claims (PVCs) for job recovery data. To ensure
          that these PVCs can be bound to approriate storage volumes, dynamic volume provisioning
          must be enabled in your cluster.

          Work with your cluster administrator to identify the storage needs of your jobs and to
          ensure that dynamic volume provisioning is enabled so that PVCs can be matched to
          persistent volumes of the right storage class.

          For more information about dynamic storage provisioning, see the
          [OpenShift documentation](https://docs.openshift.com/container-platform/latest/storage/dynamic-provisioning.html).

          ## Quick start and examples

          For an overview of the key concepts associated with running jobs in Kubernetes, see
          ["About running jobs in Kubernetes"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Configuring_and_running_Co_Operating_System_jobs_in_Kubernetes/About_running_jobs_in_Kubernetes)
          in the Ab Initio Help.

          As described in this topic, you might begin by creating a CoOperatingSystemRuntime
          manifest that defines the configuration and compute resources needed for a standalone test
          job. You could then use the specification for this job to create a CoOperatingSystemRuntimeTemplate
          resource, which would collect common settings and make it possible to concisely specify
          CoOperatingSystemRuntime jobs.

          You could also use a runtime pool (of kind CoOperatingSystemRuntime) to extend these basic
          resources: such a pool consists of an actively running group of templated runtime objects whose
          pods can be used to run jobs on demand. Finally, for reserving resources in a pool, you could
          create a CoOperatingSystemRuntimeClaim resource.

          For more information about these custom resources, refer to the
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Co_Operating_System_runtime_operator_API_reference).

          For a set of integrated examples (including complete YAML manifests) that show you how to
          create and use the custom resources, see ["Example manifests for running jobs"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Configuring_and_running_Co_Operating_System_jobs_in_Kubernetes/Creating_custom_resources_for_running_jobs/Example_manifests_for_running_jobs).
        displayName: Co>Operating System Runtime Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - abinitio
        - runtime
        - operator
        links:
        - name: Ab Initio
          url: https://www.abinitio.com/
        maintainers:
        - email: support@abinitio.com
          name: Ab Initio Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Ab Initio
        relatedImages:
        - registry.connect.redhat.com/abinitio/cooperating-system-runtime-operator@sha256:ee30210a153e8b3ba09080737216d881cc6a07e1a888e9684be0d17fd4ff791d
        version: 4.2.3-0
      entries:
      - name: abinitio-runtime-operator.v4.2.3-0
        version: 4.2.3-0
      - name: abinitio-runtime-operator.v4.2.2
        version: 4.2.2
      name: stable
    defaultChannel: release-4.3
    packageName: abinitio-runtime-operator
    provider:
      name: Ab Initio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: recipe
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: recipe.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "Recipe",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "recipe",
                    "app.kubernetes.io/instance": "recipe-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "recipe",
                    "app.kubernetes.io/part-of": "recipe"
                  },
                  "name": "recipe-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          createdAt: "2025-09-12T06:34:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Recipe is the Schema for the recipes API
            displayName: Recipe
            kind: Recipe
            name: recipes.ramendr.openshift.io
            version: v1alpha1
        description: |
          Recipe describes a workflow used for capturing or recovering Kubernetes resources. This can be referred to by a Ramen DRPlacementControl object when kubeObjectProtection is enabled.
        displayName: Recipe
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - recipe
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/recipe
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:af262b0890300183db1766bc392017b0bda723a9ba841dc25e3d928aa413db9a
        version: 4.16.16-rhodf
      entries:
      - name: recipe.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: recipe.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: recipe.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: recipe.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: recipe.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: recipe.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: recipe.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: recipe.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: recipe.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: recipe.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: recipe.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: recipe.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: recipe.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: recipe.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: recipe.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: recipe.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: recipe.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: recipe
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: New England Research Cloud
      provider-url: https://nerc.mghpcc.org/
    name: keycloak-permissions-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keycloak-permissions-operator.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.nerc.mghpcc.org/v1",
                "kind": "KeycloakAuthorization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keycloak-permissions-operator",
                    "app.kubernetes.io/instance": "keycloakauthorization-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "keycloakauthorization",
                    "app.kubernetes.io/part-of": "keycloak-permissions-operator"
                  },
                  "name": "example"
                },
                "spec": {
                  "cleanupUndefinedAuthorizations": false,
                  "description": "",
                  "keycloak": {
                    "adminSecret": {
                      "name": "keycloak-initial-admin",
                      "passwordKey": "password",
                      "usernameKey": "username"
                    }
                  },
                  "managementState": "Managed",
                  "realm": null,
                  "validateCerts": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          com.redhat.openshift.versions: v4.11
          containerImage: quay.io/nerc-images/keycloak-permissions-operator:1.4.1
          createdAt: "2024-05-07T21:09:02Z"
          description: Managing Keycloak resources, scopes, policies, and permissions
            for fine-grained resource permissions
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/nerc-images/keycloak-permissions-operator
          support: New England Research Cloud
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Keycloak resources, scopes, policies, and permissions
              for fine-grained resource permissions.
            displayName: Keycloak Authorization
            kind: KeycloakAuthorization
            name: keycloakauthorizations.keycloak.nerc.mghpcc.org
            version: v1
        description: "An OpenShift Operator for managing Keycloak resources, scopes,
          policies,\nand permissions for fine-grained resource permissions.\nAccording
          to the [Keycloak Authorization Services documentation](https://www.keycloak.org/docs/23.0.7/authorization_services/):\n\n-
          A resource can be a web page, a RESTFul resource, a file in your file system,
          or other entity you wish to authorize access to.\n- Scopes usually represent
          the actions that can be performed on a resource, but they are not limited
          to that. You can also use scopes to represent one or more attributes within
          a resource.\n- Policies define the conditions that must be satisfied to
          access or perform operations on something (resource or scope), but they
          are not tied to what they are protecting. They are generic and can be reused
          to build permissions or even more complex policies.\n- Permissions are coupled
          with the resource they are protecting. Here you specify what you want to
          protect (resource or scope) and the policies that must be satisfied to grant
          or deny permission.\n\n## Getting started with the Keycloak Permissions
          Operator\n\nStart by defining the name of your KeycloakAuthorization and
          the same namespace where Keycloak is deployed.\n\n```yaml\napiVersion: keycloak.nerc.mghpcc.org/v1\nkind:
          KeycloakAuthorization\nmetadata:\n  name: example\n  namespace: keycloak\n```\n\n##
          KeycloakAuthorization spec\n\nNext, define the spec. \n\n- `validateCerts:
          false` If your Keycloak does not have valid certificates, if you are deploying
          to OpenShift Local for example. You don't need this line if Keycloak has
          valid certificates.\n- `cleanupUndefinedAuthorizations: true` if you would
          like to remove any undefined authorizations that may have been added manually.
          This is useful for keeping in sync with GitOps. If you do not want the operator
          removing any resources, policies, or permissions, then remove this line.\n\n```yaml\nspec:\n
          \ validateCerts: false\n  cleanupUndefinedAuthorizations: true\n```\n\n##
          Keycloak admin credentials\n\nSpecify the details to connect to keycloak,
          including the `baseUrl`, and the Secret name containing the Keycloak admin
          usernameKey and passwordKey.\n\n```yaml\n  keycloak:\n    baseUrl: https://keycloak-service:8443\n
          \   adminSecret:\n      name: keycloak-initial-admin\n      usernameKey:
          username\n      passwordKey: password\n```\n\n## Realm\n\nSpecify the realm
          id you wish to add authorization configuration to. If you want to add authorization
          configuration to multiple realms, then create a KeycloakAuthorization resource
          for each realm.\n\n```yaml\n  realm:\n    id: NERC\n```\n\n## Clients\n\nFor
          each client in the realm you wish to configure, add authorizationSettings.\n\n```yaml\n
          \   clients:\n      - id: ai4cloudops\n        authorizationSettings:\n```\n\n##
          Authorization Scopes\n\nList the scopes you wish to define for the client.\n\n```yaml\n
          \         scopes:\n            - id: prod\n              name: prod\n              displayName:
          Production\n            - id: test\n              name: test\n              displayName:
          Test\n```\n\n## Authorization Resources\n\nList the resources and corresponding
          scopes you wish to define for the client.\n\n```yaml\n          resources:\n
          \           - name: cluster\n              displayName: cluster\n              scopes:\n
          \               - id: prod\n                  name: prod\n                  displayName:
          Production\n                - id: test\n                  name: test\n                  displayName:
          Test\n```\n\n## Authorization Policies\n\nDefine the policies you wish to
          use in your permissions. You can choose from all the different types of
          permissions, including: aggregated, client, client-scope, group, regex,
          role, time, and user.\n\n```yaml\n          policies:\n            - id:
          client-example\n              name: client-example\n              logic:
          POSITIVE\n              decisionStrategy: UNANIMOUS\n              type:\n
          \               client:\n                  clients:\n                    -
          example\n            - id: group-example\n              name: group-example\n
          \             logic: POSITIVE\n              decisionStrategy: UNANIMOUS\n
          \             type:\n                group:\n                  groups:\n
          \                   - id: example\n```\n\n## Authorization Permissions\n\nDefine
          the permissions, connecting the resources, scopes, and policies together.\n\n```yaml\n
          \         permissions:\n            - name: group-example-cluster-test\n
          \             client: test\n              policy: group-example\n              resource:
          cluster\n            - name: client-example-cluster-test\n              client:
          example\n              policy: client-test\n              resource: cluster\n
          \             scope: test\n```\n\n## Other details\n\nResources, scopes,
          policies, and permissions will all be applied to Keycloak if you have configured
          the admin credentials to Keycloak successfully.\nWhen the KeycloakAuthorization
          is deleted, then the configured resources, policies, and permissions will
          be deleted as well. "
        displayName: Keycloak Permissions Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - fine-grained resource permissions
        - keycloak
        - resources
        - scopes
        - policies
        - permissions
        links:
        - name: Keycloak Permissions Operator
          url: https://keycloak-permissions-operator.domain
        maintainers:
        - email: computate@computate.org
          name: Christopher Tate
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: New England Research Cloud
          url: https://nerc.mghpcc.org/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/nerc-images/keycloak-permissions-operator:1.4.1
        version: 1.4.1
      entries:
      - name: keycloak-permissions-operator.v1.4.1
        version: 1.4.1
      - name: keycloak-permissions-operator.v1.3.0
        version: 1.3.0
      - name: keycloak-permissions-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: keycloak-permissions-operator
    provider:
      name: New England Research Cloud
      url: https://nerc.mghpcc.org/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: mongodb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"opstreelabs.in/v1alpha1","kind":"MongoDB","metadata":{"name":"mongodb"},"spec":{"kubernetesConfig":{"image":"quay.io/opstree/mongo:v5.0.6","imagePullPolicy":"IfNotPresent","securityContext":{"fsGroup":1001}},"storage":{"accessModes":["ReadWriteOnce"],"storageSize":"1Gi","storageClass":"gp2"},"mongoDBSecurity":{"mongoDBAdminUser":"admin","secretRef":{"name":"mongodb-secret","key":"password"}}}},{"apiVersion":"opstreelabs.in/v1alpha1","kind":"MongoDBCluster","metadata":{"name":"mongodb"},"spec":{"clusterSize":3,"kubernetesConfig":{"image":"quay.io/opstree/mongo:v5.0.6","imagePullPolicy":"IfNotPresent","securityContext":{"fsGroup":1001}},"storage":{"accessModes":["ReadWriteOnce"],"storageSize":"1Gi","storageClass":"gp2"},"mongoDBSecurity":{"mongoDBAdminUser":"admin","secretRef":{"name":"mongodb-secret","key":"password"}}}}]'
          capabilities: Seamless Upgrades
          categories: Database,  OpenShift Optional, Storage
          certified: "false"
          containerImage: quay.io/opstree/mongodb-operator:v0.3.0
          createdAt: 05-04-2022
          description: A golang based operator to create and manage MongoDB standalone
            setup and cluster setup
          k8sMaxVersion: ""
          k8sMinVersion: 1.17.0
          repository: https://github.com/ot-container-kit/mongodb-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Mongo DB Cluster
            displayName: MongoDBCluster
            kind: MongoDBCluster
            name: mongodbclusters.opstreelabs.in
            version: v1alpha1
          - description: Mongo DB
            displayName: MongoDB
            kind: MongoDB
            name: mongodbs.opstreelabs.in
            version: v1alpha1
        description: |+
          MongoDB Operator is an operator created in Golang to create, update, and manage MongoDB standalone, replicated, and arbiter replicated setup on Kubernetes and Openshift clusters. This operator is capable of doing the setup for MongoDB with all the required best practices.

          ### Supported Features

          * MongoDB replicated cluster setup
          * MongoDB standalone setup
          * MongoDB replicated cluster failover and recovery
          * Monitoring support with MongoDB Exporter
          * Password based authentication for MongoDB
          * Kubernetes's resources for MongoDB standalone and cluster
          * Customizable configuration changes in MongoDB
          * Insightful Grafana dashboards

          ### Documentation

          **[Documentation](https://ot-mongodb-operator.netlify.app/)**
          ### Getting Help

          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.
          ### Contributing
          You can contribute to this project by:-

          * Raising Bugs and Feature Requests
          * Fixing issue by opening a Pull Request
          * Improving Documentation

          ### License

          MongoDB  Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/mongodb-operator/blob/master/LICENSE)


        displayName: MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - MongoDB
        - Opstree
        - Operatork
        - Cloud Native
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-mongodb-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: alpha
        minKubeVersion: 1.17.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/opstree/mongodb-operator:v0.3.0
        version: 0.3.0
      entries:
      - name: mongodb-operator.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: mongodb-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: https://konveyor.io
    name: move2kube-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: move2kube-operator.v0.3.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {
                  "deployment": {
                    "api": {
                      "cpu": "300m",
                      "imageTag": "",
                      "initContainer": false,
                      "memory": "1Gi",
                      "privilegedPods": false,
                      "replicas": 1
                    },
                    "authServer": {
                      "cpu": "300m",
                      "enable": false,
                      "memory": "1Gi",
                      "replicas": 1
                    },
                    "database": {
                      "cpu": "300m",
                      "enable": false,
                      "memory": "1Gi",
                      "replicas": 1,
                      "startEmpty": true
                    }
                  },
                  "ingress": {
                    "enable": true,
                    "host": "mydomain.com",
                    "preferRoute": false,
                    "tlsSecretName": ""
                  },
                  "persistentVolumeClaim": {
                    "createNew": true,
                    "enable": false,
                    "name": "",
                    "readWriteMany": true,
                    "storageClassName": "",
                    "storageSize": "1Gi"
                  },
                  "secret": {
                    "api": {
                      "configYAML": "",
                      "createNew": true,
                      "enable": false,
                      "name": ""
                    },
                    "authServer": {
                      "adminPassword": "password",
                      "adminUsername": "admin",
                      "createNew": true,
                      "databasePassword": "password",
                      "databaseUsername": "auth-server",
                      "name": "",
                      "realmJSON": "",
                      "standaloneHAXML": ""
                    }
                  },
                  "securityContextContraints": {
                    "enable": false
                  },
                  "serviceAccount": {
                    "createNew": true,
                    "enable": false,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.10
          createdAt: "2023-09-25T02:56:15Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/konveyor/move2kube-operator:v0.3.10
        version: 0.3.10
      entries:
      - name: move2kube-operator.v0.3.10
        version: 0.3.10
      - name: move2kube-operator.v0.3.10-rc.0
        version: 0.3.10-rc.0
      - name: move2kube-operator.v0.3.9
        version: 0.3.9
      - name: move2kube-operator.v0.3.8
        version: 0.3.8
      name: alpha
    - currentCSV: move2kube-operator.v0.3.15-rc.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.15-rc.0
          createdAt: "2024-10-18T16:10:56Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/konveyor/move2kube-operator:v0.3.15-rc.0
        version: 0.3.15-rc.0
      entries:
      - name: move2kube-operator.v0.3.15-rc.0
        version: 0.3.15-rc.0
      - name: move2kube-operator.v0.3.14-rc.0
        version: 0.3.14-rc.0
      - name: move2kube-operator.v0.3.13-rc.0
        version: 0.3.13-rc.0
      - name: move2kube-operator.v0.3.11-rc.0
        version: 0.3.11-rc.0
      name: prerelease
    - currentCSV: move2kube-operator.v0.3.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.15
          createdAt: "2025-03-07T06:09:48Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/konveyor/move2kube-operator:v0.3.15
        version: 0.3.15
      entries:
      - name: move2kube-operator.v0.3.15
        version: 0.3.15
      - name: move2kube-operator.v0.3.14
        version: 0.3.14
      - name: move2kube-operator.v0.3.13
        version: 0.3.13
      - name: move2kube-operator.v0.3.11
        version: 0.3.11
      name: stable
    defaultChannel: stable
    packageName: move2kube-operator
    provider:
      name: Konveyor
      url: https://konveyor.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Dell Technologies
      provider-url: https://github.com/dell/csm-operator
    name: dell-csm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dell-csm-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "authorization",
                  "namespace": "authorization"
                },
                "spec": {
                  "modules": [
                    {
                      "components": [
                        {
                          "enabled": true,
                          "name": "nginx"
                        },
                        {
                          "enabled": true,
                          "name": "cert-manager"
                        },
                        {
                          "authorizationController": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231",
                          "authorizationControllerReplicas": 1,
                          "certificate": "",
                          "controllerReconcileInterval": "5m",
                          "enabled": true,
                          "hostname": "csm-authorization.com",
                          "leaderElection": true,
                          "name": "proxy-server",
                          "opa": "docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095",
                          "opaKubeMgmt": "docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517",
                          "openTelemetryCollectorAddress": "",
                          "privateKey": "",
                          "proxyServerIngress": [
                            {
                              "annotations": {},
                              "hosts": [],
                              "ingressClassName": "nginx"
                            }
                          ],
                          "proxyService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682",
                          "proxyServiceReplicas": 1,
                          "roleService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c",
                          "roleServiceReplicas": 1,
                          "storageService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5",
                          "storageServiceReplicas": 1,
                          "tenantService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358",
                          "tenantServiceReplicas": 1
                        },
                        {
                          "commander": "docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a",
                          "name": "redis",
                          "redis": "docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f",
                          "redisCommander": "rediscommander",
                          "redisName": "redis-csm",
                          "redisReplicas": 5,
                          "redisSecretProviderClass": [
                            {
                              "redisPasswordKey": "",
                              "redisSecretName": "",
                              "redisUsernameKey": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "sentinel": "sentinel"
                        },
                        {
                          "configSecretProviderClass": [
                            {
                              "configSecretName": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "name": "config"
                        },
                        {
                          "name": "storage-system-credentials",
                          "secretProviderClasses": {
                            "vault": [
                              "secret-provider-class-1",
                              "secret-provider-class-2"
                            ]
                          }
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": true,
                      "forceRemoveModule": true,
                      "name": "authorization-proxy-server"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "isilon",
                  "namespace": "isilon"
                },
                "spec": {
                  "driver": {
                    "authSecret": "isilon-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VERBOSE",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_ISI_PORT",
                          "value": "8080"
                        },
                        {
                          "name": "X_CSI_ISI_PATH",
                          "value": "/ifs/data/csi"
                        },
                        {
                          "name": "X_CSI_ISI_NO_PROBE_ON_START",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_AUTOPROBE",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_AUTH_TYPE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_CUSTOM_TOPOLOGY_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOISILON_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "AZ_RECONCILE_INTERVAL",
                          "value": "1h"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_ISI_QUOTA_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_ACCESS_ZONE",
                          "value": "System"
                        },
                        {
                          "name": "X_CSI_ISI_VOLUME_PATH_PERMISSIONS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_IGNORE_UNRESOLVABLE_HOSTS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "isilon",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerscale"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSCALE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_PERFORMANCE_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSCALE_QUOTA_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "ISICLIENT_INSECURE",
                              "value": "true"
                            },
                            {
                              "name": "ISICLIENT_AUTH_TYPE",
                              "value": "1"
                            },
                            {
                              "name": "ISICLIENT_VERBOSE",
                              "value": "0"
                            },
                            {
                              "name": "POWERSCALE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSCALE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0",
                          "name": "metrics-powerscale"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-isilon/csi_sock",
                            "--mode=node",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powermax",
                  "namespace": "powermax"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powermax-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_MANAGED_ARRAYS",
                          "value": "000000000001,000000000002"
                        },
                        {
                          "name": "X_CSI_POWERMAX_PORTGROUPS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_TRANSPORT_PROTOCOL",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_K8S_CLUSTER_PREFIX",
                          "value": "CSM"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "X_CSI_VSPHERE_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VSPHERE_PORTGROUP",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VSPHERE_HOSTNAME",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VCENTER_HOST",
                          "value": ""
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "X_CSI_POWERMAX_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "X_CSI_POWERMAX_DEBUG",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_REVPROXY_USE_SECRET",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powermax",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERMAX_ISCSI_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_TOPOLOGY_CONTROL_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": [
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/memory-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/disk-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/network-unavailable",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REVPROXY_TLS_SECRET",
                              "value": "csirevproxy-tls-secret"
                            },
                            {
                              "name": "X_CSI_REVPROXY_PORT",
                              "value": "2222"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            },
                            {
                              "name": "DeployAsSidecar",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265",
                          "name": "csipowermax-reverseproxy"
                        }
                      ],
                      "configVersion": "v2.14.0",
                      "name": "csireverseproxy"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powermax"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERMAX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERMAX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414",
                          "name": "metrics-powermax"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/powermax.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powerstore",
                  "namespace": "powerstore"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powerstore-config",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_NODE_NAME_PREFIX",
                          "value": "csi-node"
                        },
                        {
                          "name": "X_CSI_FC_PORTS_FILTER_FILE_PATH",
                          "value": "/etc/fc-ports-filter"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOPOWERSTORE_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSM_AUTH_ENABLED",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_MULTI_NAS_FAILURE_THRESHOLD",
                          "value": "5"
                        },
                        {
                          "name": "X_CSI_MULTI_NAS_COOLDOWN_PERIOD",
                          "value": "5m"
                        },
                        {
                          "name": "X_CSI_NFS_ACLS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerstore",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-powerstore.dellemc.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerstore"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSTORE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_SPACE_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_ARRAY_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_FILE_SYSTEM_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSTORE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSTORE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "ZIPKIN_URI",
                              "value": ""
                            },
                            {
                              "name": "ZIPKIN_SERVICE_NAME",
                              "value": "metrics-powerstore"
                            },
                            {
                              "name": "ZIPKIN_PROBABILITY",
                              "value": "0.0"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60",
                          "name": "metrics-powerstore"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "unity",
                  "namespace": "unity"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_UNITY_ALLOW_MULTI_POD_ACCESS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_EPHEMERAL_STAGING_PATH",
                          "value": "/var/lib/kubelet/plugins/kubernetes.io/csi/pv/"
                        },
                        {
                          "name": "X_CSI_ISCSI_CHROOT",
                          "value": "/noderoot"
                        },
                        {
                          "name": "X_CSI_UNITY_SYNC_NODEINFO_INTERVAL",
                          "value": "15"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOUNITY_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOUNITY_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "TENANT_NAME",
                          "value": ""
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_UNITY_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "unity",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "vxflexos",
                  "namespace": "vxflexos"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLELISTVOLUMESNAPSHOT",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLESNAPSHOTCGDELETE",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "INFO"
                        },
                        {
                          "name": "GOSCALEIO_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOSCALEIO_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_QUOTA_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "INTERFACE_NAMES",
                          "value": null
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERFLEX_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "File",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerflex",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "initContainers": [
                      {
                        "envs": [
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          },
                          {
                            "name": "REPO_ADDRESS",
                            "value": "sftp://0.0.0.0"
                          },
                          {
                            "name": "REPO_USER",
                            "value": "sdcRepoUser"
                          },
                          {
                            "name": "MODULE_SIGCHECK",
                            "value": "0"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "imagePullPolicy": "IfNotPresent",
                        "name": "sdc"
                      }
                    ],
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_SDC_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_APPROVE_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_PREFIX",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_SDC_SFTP_REPO_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 1,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "enabled": false,
                        "envs": [
                          {
                            "name": "HOST_PID",
                            "value": "1"
                          },
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "name": "sdc-monitor"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "csi-external-health-monitor-controller"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERFLEX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_SDC_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_SDC_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERFLEX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d",
                          "name": "metrics-powerflex"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerflex"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--csisock=unix:/var/lib/kubelet/plugins/vxflexos.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
          createdAt: "2025-09-05T13:18:42Z"
          description: Easily install and manage Dell’s CSI Drivers and CSM
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/dell/csm-operator
          support: Dell Technologies
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ContainerStorageModule is the Schema for the containerstoragemodules
              API
            displayName: Container Storage Module
            kind: ContainerStorageModule
            name: containerstoragemodules.storage.dell.com
            version: v1
        description: "Dell Container Storage Modules (CSM) Operator is a Kubernetes
          Operator which can be \nused to install and manage Dell’s CSI drivers and
          CSM modules. \nBy using Dell CSM Operator, enterprises can quickly and easily
          deploy the \nCSM modules for CSI drivers making it easy for DevOps \nteams
          to build and optimize their Kubernetes workloads using enterprise grade
          storage \nwhile consistently aligning with the desired outcomes of the organization.\nThis
          operator is available as both community and certified operator for Openshift
          \nand as a community operator for upstream Kubernetes.\nOperator can be
          deployed via OperatorHub.io. using Operator Lifecycle Manager (OLM)\n\n###
          About Dell CSM\n\nDell CSM brings powerful enterprise storage features and
          functionality to Kubernetes for \neasier adoption of cloud native workloads,
          improved productivity and scalable operations. \n\nCSM delivers software
          modules for storage management that provide developers access to \nbuild
          automation for enhanced IT needs and other critical enterprise storage features.
          \n\nCurrently, the following modules are available as part of Dell CSM.
          \n*\tObservability - Delivers a high-level view of the storage capacity
          \n  and performance usage via Grafana dashboards to the Kubernetes users\n*\tReplication
          - Enables array replication capabilities for Kubernetes users\n*\tAuthorization
          - Provides both storage and Kubernetes administrators the ability \n  to
          apply RBAC for Dell CSI Drivers\n*\tResiliency - Designed to make Kubernetes
          applications more resilient to \n  node failures\n\nThese modules are available
          for various storage platforms in the Dell portfolio, \nsuch as Dell PowerStore,
          Dell PowerMax, Dell PowerFlex, Dell PowerScale, and Dell Unity XT.\n\nFor
          more information, visit [Dell’s Container Storage Module web page](https://www.dell.com/en-us/dt/solutions/kubernetes-containers/kubernetes-storage.htm)\n\nAdditional
          Materials\n*\t[Documentation](https://dell.github.io/csm-docs)\n*\t[CSM
          Solution Brief](https://www.delltechnologies.com/asset/en-us/products/storage/briefs-summaries/csm-solution-brief.pdf)\n"
        displayName: Dell Container Storage Modules
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dell Container Storage Modules
        - Dell CSI Driver
        - Dell CSM Modules
        - Powerflex
        - Powerscale
        - Powerstore
        - Unity
        - Authorization
        - Observability
        - Replication
        links:
        - name: Documentation
          url: https://dell.github.io/csm-docs
        maintainers:
        - email: container.storage.modules@dell.com
          name: Dell Container Storage Modules
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Dell Technologies
          url: https://github.com/dell/csm-operator
        relatedImages:
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5
        - registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334
        - docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517
        - registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
        - registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:59335acaca515fceed5f0350c46cfcfd6037475f0cf35ae8ab41d37333039e2e
        - registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d
        - registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60
        - ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0
        - registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265
        - registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414
        - docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f
        - quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540
        - docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231
        - registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16
        - registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        version: 1.10.0
      entries:
      - name: dell-csm-operator.v1.10.0
        version: 1.10.0
      - name: dell-csm-operator.v1.9.1
        version: 1.9.1
      - name: dell-csm-operator.v1.9.0
        version: 1.9.0
      - name: dell-csm-operator.v1.8.1
        version: 1.8.1
      - name: dell-csm-operator.v1.8.0
        version: 1.8.0
      - name: dell-csm-operator.v1.7.0
        version: 1.7.0
      - name: dell-csm-operator.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: dell-csm-operator
    provider:
      name: Dell Technologies
      url: https://github.com/dell/csm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: raftt.io
      provider-url: https:///www.raftt.io
    name: raftt-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: raftt-operator.v1.25.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "raftt.io/v1beta1",
                "kind": "Raftt",
                "metadata": {
                  "name": "raftt",
                  "namespace": "raftt"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/raftt/raftt-operator@sha256:73edc6d8416ee3815d7e59bb26c33e3d35d97cf1ceadd8fa457041be13e175ca
          createdAt: "2025-04-29T11:31:02Z"
          description: This operator installs and configures a Raftt instance
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/raftt-operator/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/raftt-operator/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: raftt
          operators.openshift.io/valid-subscription: Raftt commercial license
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Raftt systems
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Raftt represents an installation of Raftt cluster controller
              & supporting services
            displayName: Raftt
            kind: Raftt
            name: raftts.raftt.io
            version: v1beta1
        description: |
          Raftt allows fast code iterations and debugging over complex Kubernetes environments.
          Raftt provides a localhost development experience for applications running on Kubernetes.
          You can use Raftt to orchestrate the environments for you, so you get the optimal development experience -
          iterations that take under a second, OOTB debugging of your code as it runs in the cluster, easy introspection, and more.
        displayName: Raftt
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dx
        - development
        - debugging
        links:
        - name: Raftt Operator
          url: https://raftt-operator.domain
        - name: Documentation
          url: https://docs.raftt.io
        maintainers:
        - email: ops@raftt.io
          name: Raftt
        maturity: alpha
        provider:
          name: raftt.io
          url: https:///www.raftt.io
        relatedImages:
        - quay.io/raftt/raftt-operator@sha256:73edc6d8416ee3815d7e59bb26c33e3d35d97cf1ceadd8fa457041be13e175ca
        version: 1.25.2
      entries:
      - name: raftt-operator.v1.25.2
        version: 1.25.2
      name: alpha
    defaultChannel: alpha
    packageName: raftt-operator
    provider:
      name: raftt.io
      url: https:///www.raftt.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HashiCorp
      provider-url: https://www.hashicorp.com/
    name: vault-secrets-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vault-secrets-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultConnection",
                "metadata": {
                  "name": "vaultconnection-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "address": "http://vault.vault.svc.cluster.local:8200"
                }
              },
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultAuth",
                "metadata": {
                  "name": "vaultauth-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "vaultConnectionRef": "vaultconnection-sample",
                  "method": "kubernetes",
                  "mount": "kubernetes",
                  "kubernetes": {
                    "role": "sample",
                    "serviceAccount": "default"
                  }
                }
              },
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultStaticSecret",
                "metadata": {
                  "name": "vaultstaticsecret-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "vaultAuthRef": "vaultauth-sample",
                  "mount": "kvv2",
                  "type": "kv-v2",
                  "path": "secret",
                  "refreshAfter": "5s",
                  "destination": {
                    "create": true,
                    "name": "secret1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.connect.redhat.com/hashicorp/vault-secrets-operator@sha256:84ac7003c4cb90c35f1e296585bb58187a86d64f972133e2661839bdb532ddb8
          createdAt: "2025-09-26T21:50:51Z"
          description: The Vault Secrets Operator (VSO) allows Pods to consume Vault
            secrets natively from Kubernetes Secrets.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/hashicorp/vault-secrets-operator
          support: HashiCorp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CSISecrets
            name: csisecrets.secrets.hashicorp.com
            version: v1beta1
          - description: HCPAuth is the Schema for the hcpauths API
            displayName: HCPAuth
            kind: HCPAuth
            name: hcpauths.secrets.hashicorp.com
            version: v1beta1
          - description: HCPVaultSecretsApp is the Schema for the hcpvaultsecretsapps
              API
            displayName: HCPVault Secrets App
            kind: HCPVaultSecretsApp
            name: hcpvaultsecretsapps.secrets.hashicorp.com
            version: v1beta1
          - description: SecretTransformation is the Schema for the secrettransformations
              API
            displayName: Secret Transformation
            kind: SecretTransformation
            name: secrettransformations.secrets.hashicorp.com
            version: v1beta1
          - description: VaultAuthGlobal is the Schema for the vaultauthglobals API
            displayName: Vault Auth Global
            kind: VaultAuthGlobal
            name: vaultauthglobals.secrets.hashicorp.com
            version: v1beta1
          - description: VaultAuth is the Schema for the vaultauths API
            displayName: Vault Auth
            kind: VaultAuth
            name: vaultauths.secrets.hashicorp.com
            version: v1beta1
          - description: VaultConnection is the Schema for the vaultconnections API
            displayName: Vault Connection
            kind: VaultConnection
            name: vaultconnections.secrets.hashicorp.com
            version: v1beta1
          - description: VaultDynamicSecret is the Schema for the vaultdynamicsecrets
              API
            displayName: Vault Dynamic Secret
            kind: VaultDynamicSecret
            name: vaultdynamicsecrets.secrets.hashicorp.com
            version: v1beta1
          - description: VaultPKISecret is the Schema for the vaultpkisecrets API
            displayName: Vault PKISecret
            kind: VaultPKISecret
            name: vaultpkisecrets.secrets.hashicorp.com
            version: v1beta1
          - description: VaultStaticSecret is the Schema for the vaultstaticsecrets
              API
            displayName: Vault Static Secret
            kind: VaultStaticSecret
            name: vaultstaticsecrets.secrets.hashicorp.com
            version: v1beta1
        description: |-
          The Vault Secrets Operator (VSO) allows Pods to consume Vault secrets
          natively from Kubernetes Secrets.

          ## Overview

          The Vault Secrets Operator operates by watching for changes to its supported set of Custom Resource Definitions (CRD).
          Each CRD provides the specification required to allow the *Operator* to synchronize a Vault Secrets to a Kubernetes Secret.
          The *Operator* writes the *source* Vault secret data directly to the *destination* Kubernetes Secret, ensuring that any
          changes made to the *source* are replicated to the *destination* over its lifetime. In this way, an application only needs
          to have access to the *destination* secret in order to make use of the secret data contained within.

          See the developer docs for more info [here](https://developer.hashicorp.com/vault/docs/platform/k8s/vso), including
          [examples](https://developer.hashicorp.com/vault/docs/platform/k8s/vso/examples) and a
          [tutorial](https://developer.hashicorp.com/vault/tutorials/kubernetes/vault-secrets-operator)
          for getting started.
        displayName: Vault Secrets Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vault
        - security
        - secrets
        links:
        - name: Vault Secrets Operator
          url: https://github.com/hashicorp/vault-secrets-operator
        - name: Documentation
          url: https://developer.hashicorp.com/vault/docs/platform/k8s/vso
        maintainers:
        - email: support@hashicorp.com
          name: HashiCorp Support
        maturity: stable
        provider:
          name: HashiCorp
          url: https://www.hashicorp.com/
        relatedImages:
        - registry.connect.redhat.com/hashicorp/vault-secrets-operator@sha256:84ac7003c4cb90c35f1e296585bb58187a86d64f972133e2661839bdb532ddb8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:503d5ee0fbcd5bd3492785f4246e19cbdb92960be0d05f3c6f123a32fe0fe84d
        version: 1.0.1
      entries:
      - name: vault-secrets-operator.v1.0.1
        version: 1.0.1
      - name: vault-secrets-operator.v1.0.0
        version: 1.0.0
      - name: vault-secrets-operator.v0.10.0
        version: 0.10.0
      - name: vault-secrets-operator.v0.9.1
        version: 0.9.1
      - name: vault-secrets-operator.v0.9.0
        version: 0.9.0
      - name: vault-secrets-operator.v0.8.1
        version: 0.8.1
      - name: vault-secrets-operator.v0.8.0
        version: 0.8.0
      - name: vault-secrets-operator.v0.7.0
        version: 0.7.0
      - name: vault-secrets-operator.v0.6.0
        version: 0.6.0
      - name: vault-secrets-operator.v0.5.2
        version: 0.5.2
      - name: vault-secrets-operator.v0.5.1
        version: 0.5.1
      - name: vault-secrets-operator.v0.5.0
        version: 0.5.0
      - name: vault-secrets-operator.v0.4.3
        version: 0.4.3
      name: stable
    defaultChannel: stable
    packageName: vault-secrets-operator
    provider:
      name: HashiCorp
      url: https://www.hashicorp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microsoft
      provider-url: ""
    name: azure-service-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: azure-service-operator.v1.0.59040
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "APIMgmtAPI",
                "metadata": {
                  "name": "aso-apimgmtapi-sample"
                },
                "spec": {
                  "apiId": "apiId0",
                  "apiService": "apimservice-sample",
                  "location": "westus",
                  "properties": {
                    "apiVersionDescription": "API version description",
                    "description": "API description",
                    "displayName": "aso-apimgmtapi-sample",
                    "isCurrent": true,
                    "isOnline": true,
                    "path": "/api/test",
                    "protocols": [
                      "http"
                    ],
                    "subscriptionRequired": false
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ApimService",
                "metadata": {
                  "name": "apimservice-sample"
                },
                "spec": {
                  "appInsightsName": "appinsights-sample-wmm",
                  "appInsightsResourceGroup": "resourcegroup-azure-operators-wmm",
                  "location": "westus",
                  "publisherEmail": {
                    "Your email": ""
                  },
                  "publisherName": {
                    "Your name": ""
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "tier": "premium",
                  "vnetName": "virtualnetwork-sample",
                  "vnetResourceGroup": "resourcegroup-azure-operators",
                  "vnetSubnetName": "test1",
                  "vnetType": "none"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AppInsights",
                "metadata": {
                  "name": "appinsights-sample"
                },
                "spec": {
                  "applicationType": "other",
                  "kind": "other",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AppInsightsApiKey",
                "metadata": {
                  "name": "key-sample"
                },
                "spec": {
                  "appInsights": "appinsights-sample",
                  "authSDKControlChannel": true,
                  "readTelemetry": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "writeAnnotations": true
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureLoadBalancer",
                "metadata": {
                  "name": "lb-sample-01"
                },
                "spec": {
                  "backendAddressPoolName": "test",
                  "backendPort": 3000,
                  "frontendPortRangeEnd": 2000,
                  "frontendPortRangeStart": 1001,
                  "inboundNatPoolName": "test",
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "lb-pip-sample-01",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureNetworkInterface",
                "metadata": {
                  "name": "nic-sample-01"
                },
                "spec": {
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "nic-pip-sample-01",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnetName": "test1",
                  "vnetName": "vnet-sample-01"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureNetworkInterface",
                "metadata": {
                  "name": "nic-sample-02"
                },
                "spec": {
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "nic-pip-sample-02",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnetName": "test1",
                  "vnetName": "vnet-sample-02"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "lb-pip-sample-01"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "nic-pip-sample-01"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "nic-pip-sample-02"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLManagedUser",
                "metadata": {
                  "name": "azuresqlmanageduser-sample"
                },
                "spec": {
                  "dbName": "azuresqldatabase-sample",
                  "managedIdentityClientId": null,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "db_owner"
                  ],
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLUser",
                "metadata": {
                  "name": "sqluser-sample"
                },
                "spec": {
                  "dbName": "azuresqldatabase-sample1",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "db_owner"
                  ],
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLVNetRule",
                "metadata": {
                  "name": "azuresqlvnetrule-sample"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-vnet",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlAction",
                "metadata": {
                  "name": "azuresqlaction-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverName": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "azuresqldatabase-sample1"
                },
                "spec": {
                  "edition": 0,
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "valueA",
                    "tag2": "valueB"
                  },
                  "name": "azuresqldatabase-sample3"
                },
                "spec": {
                  "edition": 0,
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlFailoverGroup",
                "metadata": {
                  "name": "azuresqlfailovergroup-sample"
                },
                "spec": {
                  "databaselist": [
                    "azuresqldatabase-sample1",
                    "azuresqldatabase-sample2"
                  ],
                  "failovergraceperiod": 30,
                  "failoverpolicy": "Automatic",
                  "location": "eastus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "secondaryserver": "sqlserver-samplesec",
                  "secondaryserverresourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-samplepri"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlFirewallRule",
                "metadata": {
                  "name": "sqlf-allowazuresvcaccess"
                },
                "spec": {
                  "endipaddress": "0.0.0.0",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "startipaddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample-777"
                },
                "spec": {
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample777"
                },
                "spec": {
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVMScaleSet",
                "metadata": {
                  "name": "vmss-sample-01"
                },
                "spec": {
                  "adminUserName": "azureuser7",
                  "backendAddressPoolName": "test",
                  "capacity": 3,
                  "inboundNatPoolName": "test",
                  "loadBalancerName": "lb-sample-01",
                  "location": "SouthCentralUS",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": "{ssh public key}",
                  "subnetName": "test1",
                  "virtualNetworkName": "vnet-sample-01",
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachine",
                "metadata": {
                  "name": "hpfvm1"
                },
                "spec": {
                  "adminUserName": "azureuser",
                  "location": "SouthCentralUS",
                  "networkInterfaceName": "nic-sample-02",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": {
                    "ssh key": ""
                  },
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachine",
                "metadata": {
                  "name": "hpfvm20"
                },
                "spec": {
                  "adminUserName": "azureuser",
                  "location": "SouthCentralUS",
                  "networkInterfaceName": "nic-sample-01",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": "{ssh public key}",
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachineExtension",
                "metadata": {
                  "name": "vmext20a"
                },
                "spec": {
                  "autoUpgradeMinorVersion": true,
                  "forceUpdateTag": "Test2",
                  "location": "SouthCentralUS",
                  "protectedSettings": "{}",
                  "publisher": "Microsoft.Azure.Extensions",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "settings": "{\"commandToExecute\":\"echo 'hello, world'\"}",
                  "typeHandlerVersion": "2.1",
                  "typeName": "CustomScript",
                  "vmName": "vm20"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "BlobContainer",
                "metadata": {
                  "name": "blobcontainer-sample"
                },
                "spec": {
                  "accesslevel": "Container",
                  "accountname": "storageaccountsample777",
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ConsumerGroup",
                "metadata": {
                  "name": "consumergroup-sample-1"
                },
                "spec": {
                  "eventHub": "eventhub-sample-1",
                  "namespace": "eventhubnamespace-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "CosmosDB",
                "metadata": {
                  "name": "cosmosdb-sample-1"
                },
                "spec": {
                  "kind": "GlobalDocumentDB",
                  "location": "westus",
                  "properties": {
                    "databaseAccountOfferType": "Standard",
                    "isVirtualNetworkFilterEnabled": true
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "CosmosDB",
                "metadata": {
                  "labels": {
                    "CosmosAccountType": "Production"
                  },
                  "name": "cosmosdb-sample-2"
                },
                "spec": {
                  "ipRules": [
                    "104.42.195.92,40.76.54.131,52.176.6.30,52.169.50.45,52.187.184.26"
                  ],
                  "kind": "GlobalDocumentDB",
                  "location": "eastus",
                  "locations": [
                    {
                      "failoverPriority": 0,
                      "isZoneRedundant": true,
                      "locationName": "eastus"
                    },
                    {
                      "failoverPriority": 1,
                      "locationName": "westus"
                    }
                  ],
                  "properties": {
                    "databaseAccountOfferType": "Standard",
                    "enableMultipleWriteLocations": true
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "Eventhub",
                "metadata": {
                  "name": "eventhub-sample-1"
                },
                "spec": {
                  "authorizationRule": {
                    "name": "RootManageSharedAccessKey",
                    "rights": [
                      "Listen",
                      "Manage",
                      "Send"
                    ]
                  },
                  "location": "westus",
                  "namespace": "eventhubnamespace-sample-1",
                  "properties": {
                    "messageRetentionInDays": 7,
                    "partitionCount": 2
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "Eventhub",
                "metadata": {
                  "name": "eventhub-sample-2307-10-capture"
                },
                "spec": {
                  "authorizationRule": {
                    "name": "RootManageSharedAccessKey",
                    "rights": [
                      "Listen",
                      "Manage",
                      "Send"
                    ]
                  },
                  "location": "westus",
                  "namespace": "eventhubnamespace-sample-1",
                  "properties": {
                    "captureDescription": {
                      "destination": {
                        "archiveNameFormat": "{Namespace}/{EventHub}/{PartitionId}/{Year}/{Month}/{Day}/{Hour}/{Minute}/{Second}",
                        "blobContainer": "capturecontainer",
                        "name": "EventHubArchive.AzureBlockBlob",
                        "storageAccount": {
                          "accountName": "storageaccountauv1",
                          "resourceGroup": "my-resource-group"
                        }
                      },
                      "enabled": true,
                      "intervalInSeconds": 120,
                      "sizeLimitInBytes": 524288000
                    },
                    "messageRetentionInDays": 7,
                    "partitionCount": 2
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "EventhubNamespace",
                "metadata": {
                  "name": "eventhubnamespace-sample-1"
                },
                "spec": {
                  "location": "westus",
                  "properties": {
                    "isAutoInflateEnabled": false,
                    "kafkaEnabled": false,
                    "maximumThroughputUnits": 0
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "capacity": 1,
                    "name": "Standard",
                    "tier": "Standard"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVault",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "keyvaultsample123"
                },
                "spec": {
                  "accessPolicies": [
                    {
                      "objectID": "\u003cobjectID\u003e",
                      "permissions": {
                        "certificates": [
                          "list",
                          "get"
                        ],
                        "keys": [
                          "list",
                          "get"
                        ],
                        "secrets": [
                          "list",
                          "get"
                        ],
                        "storage": [
                          "list",
                          "get"
                        ]
                      },
                      "tenantID": "\u003ctenantID\u003e"
                    }
                  ],
                  "enableSoftDelete": false,
                  "location": "westus",
                  "networkPolicies": {
                    "bypass": "AzureServices",
                    "defaultAction": "Allow",
                    "ipRules": [
                      "172.16.0.0/24",
                      "172.16.1.0/24"
                    ],
                    "virtualNetworkRules": [
                      "/subscriptions/\u003csubid\u003e/resourceGroups/rg1/providers/Microsoft.Network/virtualNetworks/test-vnet/subnets/subnet1",
                      "/subscriptions/\u003csubid\u003e/resourceGroups/rg1/providers/Microsoft.Network/virtualNetworks/test-vnet/subnets/subnet2"
                    ]
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "name": "standard"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVault",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "keyvaultsample123-simple"
                },
                "spec": {
                  "enableSoftDelete": false,
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVaultKey",
                "metadata": {
                  "name": "keyvaultkey-sample"
                },
                "spec": {
                  "keySize": 4096,
                  "keyVault": "keyvaultsample123",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "type": "RSA"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLDatabase",
                "metadata": {
                  "name": "mysqldatabase-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLDatabase",
                "metadata": {
                  "name": "mysqldatabase-sample2"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLFirewallRule",
                "metadata": {
                  "name": "mysqlfirewallrule-sample"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLFirewallRule",
                "metadata": {
                  "name": "mysqlfirewallrule-sample2"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample2",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-replica"
                },
                "spec": {
                  "createMode": "Replica",
                  "location": "eastus2",
                  "replicaProperties": {
                    "sourceServerId": "/subscriptions/{SUBID}/resourceGroups/resourcegroup-azure-operators/providers/Microsoft.DBforMySQL/servers/mysqlserver-sample"
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-sample2"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLUser",
                "metadata": {
                  "name": "mysqluser-sample"
                },
                "spec": {
                  "dbName": "mysqldatabase-sample",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "SELECT"
                  ],
                  "server": "mysqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLVNetRule",
                "metadata": {
                  "name": "mysqlvnetrule-sample"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-vnet",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLDatabase",
                "metadata": {
                  "name": "postgresqldatabase-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLDatabase",
                "metadata": {
                  "name": "postgresqldatabase-sample2"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLFirewallRule",
                "metadata": {
                  "name": "postgresqlfirewallrule-sample"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLFirewallRule",
                "metadata": {
                  "name": "postgresqlfirewallrule-sample2"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample2",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-replica3"
                },
                "spec": {
                  "createMode": "Replica",
                  "location": "eastus",
                  "replicaProperties": {
                    "sourceServerId": "/subscriptions/{subscription ID}/resourceGroups/resourcegroup-azure-operators/providers/Microsoft.DBforPostgreSQL/servers/postgresqlserver-sample"
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-sample2"
                },
                "spec": {
                  "location": "southcentralus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 2,
                    "family": "Gen5",
                    "name": "B_Gen5_2",
                    "size": "51200",
                    "tier": "Basic"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLUser",
                "metadata": {
                  "name": "psqluser-sample"
                },
                "spec": {
                  "dbName": "postgresqldatabase-sample",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "azure_pg_admin"
                  ],
                  "server": "postgresqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLVNetRule",
                "metadata": {
                  "name": "postgresqlvnetrule-sample1"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-azure-operators",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCache",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "rediscache-sample-1"
                },
                "spec": {
                  "location": "westus",
                  "properties": {
                    "enableNonSslPort": true,
                    "sku": {
                      "capacity": 1,
                      "family": "C",
                      "name": "Basic"
                    }
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCacheAction",
                "metadata": {
                  "name": "rediscacheaction-sample-1"
                },
                "spec": {
                  "actionName": "rollallkeys",
                  "cacheName": "rediscache-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCacheFirewallRule",
                "metadata": {
                  "name": "rediscachefirewallrule"
                },
                "spec": {
                  "properties": {
                    "endIP": "0.0.0.0",
                    "startIP": "0.0.0.0"
                  },
                  "redisCache": "rediscache-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ResourceGroup",
                "metadata": {
                  "name": "resourcegroup-azure-operators"
                },
                "spec": {
                  "location": "westus"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "StorageAccount",
                "metadata": {
                  "name": "storageaccountsample777"
                },
                "spec": {
                  "accessTier": "Hot",
                  "kind": "StorageV2",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "name": "Standard_RAGRS"
                  },
                  "supportsHttpsTrafficOnly": true
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "VirtualNetwork",
                "metadata": {
                  "name": "vnet-sample-01"
                },
                "spec": {
                  "addressSpace": "10.0.0.0/8",
                  "location": "SouthCentralUS",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnets": [
                    {
                      "subnetAddressPrefix": "10.1.0.0/16",
                      "subnetName": "test1"
                    },
                    {
                      "subnetAddressPrefix": "10.2.0.0/16",
                      "subnetName": "test2"
                    }
                  ]
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "VirtualNetwork",
                "metadata": {
                  "name": "vnet-sample-02"
                },
                "spec": {
                  "addressSpace": "10.0.0.0/8",
                  "location": "SouthCentralUS",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnets": [
                    {
                      "subnetAddressPrefix": "10.1.0.0/16",
                      "subnetName": "test1"
                    },
                    {
                      "subnetAddressPrefix": "10.2.0.0/16",
                      "subnetName": "test2"
                    }
                  ]
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "BlobContainer",
                "metadata": {
                  "name": "blobcontainer-sample"
                },
                "spec": {
                  "accessLevel": "Container",
                  "accountName": "storageaccountsample777",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "MySQLServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "mysqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "postgresqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "azuresqldatabase-sample"
                },
                "spec": {
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "sku": {
                    "name": "Basic",
                    "tier": "Basic"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlFailoverGroup",
                "metadata": {
                  "name": "azuresqlfailovergroup-sample"
                },
                "spec": {
                  "databaseList": [
                    "azuresqldatabase-sample",
                    "azuresqldatabase-sample2"
                  ],
                  "failoverGracePeriod": 30,
                  "failoverPolicy": "automatic",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "secondaryServer": "sqlserver-samplesec",
                  "secondaryServerResourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-samplepri"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlFirewallRule",
                "metadata": {
                  "name": "sqlf-allowazuresvcaccess"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample-777"
                },
                "spec": {
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Cloud Provider
          certified: "false"
          containerImage: mcr.microsoft.com/k8s/azureserviceoperator:1.0.59040
          createdAt: Mon 12 Jun 2023 04:24:02 UTC
          description: The Azure Service Operator helps you provision Azure resources
            and connect your applications to them from within Kubernetes.
          displayName: Azure Service Operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Azure/azure-service-operator
          support: https://github.com/Azure/azure-service-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Creates an API with the specified properties in the specified
              API Management service.
            displayName: APIMgmtAPI
            kind: APIMgmtAPI
            name: apimgmtapis.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an API Management instance into a specified Resource
              Group at the specified location. It also provides the option to link
              to an Application Insights instance for logging, and to place the API
              Management instance in a specified Virtual Network.
            displayName: ApimService
            kind: ApimService
            name: apimservices.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Application Insights instance into a specified
              Resource Group at the specified location.
            displayName: AppInsights
            kind: AppInsights
            name: appinsights.azure.microsoft.com
            version: v1alpha1
          - description: Creates an Api Key to access a given Application Insights
              instance.
            displayName: AppInsightsApiKey
            kind: AppInsightsApiKey
            name: appinsightsapikeys.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Load Balancer (LB) into a specified Resource
              Group at the specified location.
            displayName: AzureLoadBalancer
            kind: AzureLoadBalancer
            name: azureloadbalancers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Network Interface (NIC) into a specified
              Resource Group at the specified location. Users can specify underlying
              public IP address and virtual network configurations in their NIC setup.
            displayName: AzureNetworkInterface
            kind: AzureNetworkInterface
            name: azurenetworkinterfaces.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Public IP Address (PIP) into a specified
              Resource Group at the specified location. Users can specify IP allocation
              method, idle timeout, IP address version, and SKU.
            displayName: AzurePublicIPAddress
            kind: AzurePublicIPAddress
            name: azurepublicipaddresses.azure.microsoft.com
            version: v1alpha1
          - description: Allows you to roll the password for the specified SQL server.
            displayName: AzureSqlAction
            kind: AzureSqlAction
            name: azuresqlactions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an SQL database given the SQL server.
            displayName: AzureSqlDatabase
            kind: AzureSqlDatabase
            name: azuresqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an SQL database given the SQL server
            displayName: AzureSqlDatabase
            kind: AzureSqlDatabase
            name: azuresqldatabases.azure.microsoft.com
            version: v1beta1
          - description: Deploys a failover group on a specified Azure SQL server
              given the secondary server and the databases to failover.
            displayName: AzureSqlFailoverGroup
            kind: AzureSqlFailoverGroup
            name: azuresqlfailovergroups.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a failover group on a specified Azure SQL server
              given the secondary server and the databases to failover.
            displayName: AzureSqlFailoverGroup
            kind: AzureSqlFailoverGroup
            name: azuresqlfailovergroups.azure.microsoft.com
            version: v1beta1
          - description: Deploys a firewall rule to allow access to the SQL server
              from specific IPs.
            displayName: AzureSqlFirewallRule
            kind: AzureSqlFirewallRule
            name: azuresqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the SQL server
              from specific IPs.
            displayName: AzureSqlFirewallRule
            kind: AzureSqlFirewallRule
            name: azuresqlfirewallrules.azure.microsoft.com
            version: v1beta1
          - description: Binds a specified Managed Identity as a user on the SQL database
            displayName: AzureSQLManagedUser
            kind: AzureSQLManagedUser
            name: azuresqlmanagedusers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure SQL server given the location and Resource
              Group.
            displayName: AzureSqlServer
            kind: AzureSqlServer
            name: azuresqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure SQL server given the location and Resource
              Group.
            displayName: AzureSqlServer
            kind: AzureSqlServer
            name: azuresqlservers.azure.microsoft.com
            version: v1beta1
          - description: Creates an user on the specified Azure SQL database and stores
              the username/password as secrets.
            displayName: AzureSQLUser
            kind: AzureSQLUser
            name: azuresqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a VirtualNetwork rule to allow access to the SQL
              server from specific VNets.
            displayName: AzureSQLVNetRule
            kind: AzureSQLVNetRule
            name: azuresqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys Azure Virtual Machine Extensions to provide post-deployment
              configuration and automation on Azure VMs.
            displayName: AzureVirtualMachineExtension
            kind: AzureVirtualMachineExtension
            name: azurevirtualmachineextensions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Machine (VM) into a specified Resource
              Group at the specified location. Users can specify platform image, size,
              user name and public SSH key, etc. for the VM.
            displayName: AzureVirtualMachine
            kind: AzureVirtualMachine
            name: azurevirtualmachines.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Machine Scale Set (VMSS) into a
              specified Resource Group at the specified location. Users can specify
              platform image, size, user name and public SSH key, etc. for the VMSS.
            displayName: AzureVMScaleSet
            kind: AzureVMScaleSet
            name: azurevmscalesets.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a Blob Container into a specified Resource Group
              at the specified location. A Storage Account must first be created prior
              to creating a Blob Container.
            displayName: BlobContainer
            kind: BlobContainer
            name: blobcontainers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a Blob Container into a specified Resource Group
              at the specified location. A Storage Account must first be created prior
              to creating a Blob Container.
            displayName: BlobContainer
            kind: BlobContainer
            name: blobcontainers.azure.microsoft.com
            version: v1alpha2
          - description: Deploys a consumer group given the Event Hub, Event Hubs
              namespace and Resource Group.
            displayName: ConsumerGroup
            kind: ConsumerGroup
            name: consumergroups.azure.microsoft.com
            version: v1alpha1
          - description: Provision a CosmosDB instance given the Cosmos DB instance
              type, location, and Resource Group.
            displayName: CosmosDB
            kind: CosmosDB
            name: cosmosdbs.azure.microsoft.com
            version: v1alpha1
          - kind: CosmosDBSQLDatabase
            name: cosmosdbsqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Event Hubs namespace given the Resource Group
              and location. Also has the ability to configure SKU, properties, and
              network rules.
            displayName: EventhubNamespace
            kind: EventhubNamespace
            name: eventhubnamespaces.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Event Hubs instance given the Event Hubs namespace,
              Resource Group and Location.
            displayName: Eventhub
            kind: Eventhub
            name: eventhubs.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Key Vault key given the location and Resource
              Group.
            displayName: KeyVaultKey
            kind: KeyVaultKey
            name: keyvaultkeys.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Key Vault given the location and Resource
              Group.
            displayName: KeyVault
            kind: KeyVault
            name: keyvaults.azure.microsoft.com
            version: v1alpha1
          - description: Add an AAD user to an existing MySQL database.
            kind: MySQLAADUser
            name: mysqlaadusers.azure.microsoft.com
            version: v1alpha1
          - description: Add an AAD user to an existing MySQL database.
            kind: MySQLAADUser
            name: mysqlaadusers.azure.microsoft.com
            version: v1alpha2
          - description: Deploys a database under the given Azure Database for MySQL
              server
            displayName: MySQLDatabase
            kind: MySQLDatabase
            name: mysqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the Azure Database
              for MySQL server from the specified IP range.
            displayName: MySQLFirewallRule
            kind: MySQLFirewallRule
            name: mysqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Create an AAD-based admininistrative user in a MySQL server.
            kind: MySQLServerAdministrator
            name: mysqlserveradministrators.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for MySQL server given the Location,
              Resource Group and other properties. Also helps creating read replicas
              for MySQL server.
            displayName: MySQLServer
            kind: MySQLServer
            name: mysqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for MySQL server given the Location,
              Resource Group and other properties. Also helps creating read replicas
              for MySQL server.
            displayName: MySQLServer
            kind: MySQLServer
            name: mysqlservers.azure.microsoft.com
            version: v1alpha2
          - description: Add a new user to an existing MySQL database.
            displayName: MySQLUser
            kind: MySQLUser
            name: mysqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Add a new user to an existing MySQL database.
            kind: MySQLUser
            name: mysqlusers.azure.microsoft.com
            version: v1alpha2
          - description: Add virtual network rules to the MySQL server.
            displayName: MySQLVNetRule
            kind: MySQLVNetRule
            name: mysqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a database under the given Azure Database for PostgreSQL
              server.
            displayName: PostgreSQLDatabase
            kind: PostgreSQLDatabase
            name: postgresqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the Azure Database
              for PostgreSQL server from the specified IP range.
            displayName: PostgreSQLFirewallRule
            kind: PostgreSQLFirewallRule
            name: postgresqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for PostgreSQL server given the
              Location and Resource Group.
            displayName: PostgreSQLServer
            kind: PostgreSQLServer
            name: postgresqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for PostgreSQL server given the
              Location and Resource Group.
            displayName: PostgreSQLServer
            kind: PostgreSQLServer
            name: postgresqlservers.azure.microsoft.com
            version: v1alpha2
          - description: Create and Manage Users for Azure Database for PostgreSQL
              server.
            displayName: PostgreSQLUser
            kind: PostgreSQLUser
            name: postgresqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Create and manage VNet service endpoints and VNet rules in
              Azure Database for PostgreSQL.
            displayName: PostgreSQLVNetRule
            kind: PostgreSQLVNetRule
            name: postgresqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Allows you to regenerate keys and reboot the RedisCache cluster.
            displayName: RedisCacheAction
            kind: RedisCacheAction
            name: rediscacheactions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the RedisCache
              from the specified IP range.
            displayName: RedisCacheFirewallRule
            kind: RedisCacheFirewallRule
            name: rediscachefirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Cache for Redis into a specified Resource
              Group at the specified location.
            displayName: RedisCache
            kind: RedisCache
            name: rediscaches.azure.microsoft.com
            version: v1alpha1
          - description: Provisions a Resource Group given the location and a name.
            displayName: ResourceGroup
            kind: ResourceGroup
            name: resourcegroups.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Storage Account into a specified Resource
              Group at the specified location.
            displayName: StorageAccount
            kind: StorageAccount
            name: storageaccounts.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Network into a specified Resource
              Group at the specified location, in the address space provided. Users
              are also able to add subnets to their virtual network through the operator.
            displayName: VirtualNetwork
            kind: VirtualNetwork
            name: virtualnetworks.azure.microsoft.com
            version: v1alpha1
        description: |
          ## Overview

          The Azure Service Operator comprises of:

          - The Custom Resource Definitions (CRDs) for each of the Azure services a Kubernetes user can provision.
          - The Kubernetes controller that watches for requests to create Custom Resources for each of these CRDs and creates them.

          The project was built using [Kubebuilder](https://book.kubebuilder.io/).

          Curious to see how it all works? Check out our [control flow diagrams](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/controlflow.md).

          **Note**: The current version of Azure Service Operator only supports watching all namespaces in the cluster.

          ## Supported Azure Services

          - [Resource Group](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/resourcegroup/resourcegroup.md)
          - [Event Hubs](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/eventhub/eventhub.md)
          - [Azure SQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/azuresql/azuresql.md)
          - [Azure Database for PostgreSQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/postgresql/postgresql.md)
          - [Azure Database for MySQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/mysql/mysql.md)
          - [Azure Key Vault](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/keyvault/keyvault.md)
          - [Azure Cache for Redis](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/rediscache/rediscache.md)
          - [Storage Account](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/storage/storageaccount.md)
          - [Blob Storage](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/storage/blobcontainer.md)
          - [Virtual Network](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/virtualnetwork/virtualnetwork.md)
          - [Application Insights](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/appinsights/appinsights.md)
          - [API Management](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/apimgmt/apimgmt.md)
          - [Cosmos DB](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/cosmosdb/cosmosdb.md)
          - [Virtual Machine](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/virtualmachine/virtualmachine.md)
          - [Virtual Machine Scale Set](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/vmscaleset/vmscaleset.md)

          ## Instructions - Please read before installing

          Before you begin, verify you're running the Azure CLI version 2.0.53 or later. Run `az --version` to find the version. If you need to install or upgrade, see [Install Azure CLI](https://docs.microsoft.com/en-us/cli/azure/install-azure-cli).

          1. Create an Azure Service Principal. You'll need this to grant Azure Service Operator permissions to create resources in your subscription.

              First, set the following environment variables to your Azure Tenant ID and Subscription ID with your values:
              ```yaml
              AZURE_TENANT_ID=<your-tenant-id-goes-here>
              AZURE_SUBSCRIPTION_ID=<your-subscription-id-goes-here>
              ```

              You can find these values by using the [Azure CLI](https://docs.microsoft.com/en-us/cli/azure/?view=azure-cli-latest):
              ```sh
              az account show
              ```

          2. Next, we'll create a Service Principal with Contributor permissions for your subscription, so ASO can create resources in your subscription on your behalf. Note that the [ServicePrincipal](https://docs.microsoft.com/en-us/cli/azure/create-an-azure-service-principal-azure-cli) you pass to the command below needs to have access to create resources in your subscription. If you'd like to use Managed Identity for authorization instead, check out instructions [here](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/managedidentity.md).

              ```sh
              az ad sp create-for-rbac -n "azure-service-operator" --role contributor \
                  --scopes /subscriptions/$AZURE_SUBSCRIPTION_ID
              ```

              This should give you output like the following:
              ```sh
              "appId": "xxxxxxxxxx",
              "displayName": "azure-service-operator",
              "name": "http://azure-service-operator",
              "password": "xxxxxxxxxxx",
              "tenant": "xxxxxxxxxxxxx"
              ```

          3.  Once you have created a Service Principal, gather the following values:

              `AZURE_TENANT_ID` is the Tenant ID from Step 1.

              `AZURE_SUBSCRIPTION_ID` is the Subscription ID from Step 1.

              `AZURE_CLIENT_ID` is the appID from the Service Principal created in Step 2.

              `AZURE_CLIENT_SECRET` is the password from the service Principal we created in Step 2.

              `AZURE_CLOUD_ENV` is the Azure Environment you'd like to use, i.e. AzurePublicCloud, AzureUSGovernmentCloud, AzureChinaCloud, AzureGermanCloud.

          4.  Set those values in a `Secret` called `azureoperatorsettings` within the `operators` Namespace:
              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name: azureoperatorsettings
                namespace: operators
              stringData:
                AZURE_TENANT_ID: <your-tenant-id-goes-here>
                AZURE_SUBSCRIPTION_ID: <your-subscription-id-goes-here>
                AZURE_CLIENT_ID: <your-client-id>
                AZURE_CLIENT_SECRET: <your-client-secret>
                AZURE_CLOUD_ENV: <your-azure-cloud-environment>
              ```

          5. Now you can proceed to install the Azure Service Operator to the `operators` Namespace via the Install button on the top right of this page. After the operator is installed, you will then see the Azure Service Operator pod running in your cluster:

              ```console
              $ kubectl get pods -n operators
              NAME                                                READY   STATUS    RESTARTS   AGE
              azureoperator-controller-manager-7dd75bbd97-mk4s9   2/2     Running   0          35s
              ```

          ## About the project

          This project maintains [releases of the Azure Service Operator](https://github.com/Azure/azure-service-operator/releases).

          Please see the [FAQ](https://github.com/Azure/azure-service-operator/blob/main/docs/faq.md) for answers to commonly asked questions about the Azure Service Operator.

          Have more questions? Feel free to consult our documentation [here](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/contents.md).

          ## Contributing

          The [contribution guide](https://github.com/Azure/azure-service-operator/blob/main/CONTRIBUTING.md) covers everything you need to know about how you can contribute to Azure Service Operators. The [developer guide](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/contents.md#developing-azure-service-operator) will help you onboard as a developer.

          ## Support

          Azure Service Operator is an open source project that is **not** covered by the [Microsoft Azure support policy](https://support.microsoft.com/en-us/help/2941892/support-for-linux-and-open-source-technology-in-azure). Please search open issues [here](https://github.com/Azure/azure-service-operator/issues). If your issue isn't already represented, please [open a new one](https://github.com/Azure/azure-service-operator/issues/new/choose). The Azure Service Operator project maintainers will respond to the best of their abilities.

          ## Code of conduct

          This project has adopted the [Microsoft Open Source Code of Conduct](https://opensource.microsoft.com/codeofconduct/). For more information, see the [Code of Conduct FAQ](https://opensource.microsoft.com/codeofconduct/faq) or contact [opencode@microsoft.com](mailto:opencode@microsoft.com) with any additional questions or comments.
        displayName: Azure Service Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - azure, microsoft, cloud service broker
        links:
        - name: Azure Service Operator
          url: https://github.com/Azure/azure-service-operator
        maintainers:
        - email: christian.muirhead@microsoft.com
          name: Christian Muirhead
        maturity: stable
        provider:
          name: Microsoft
        relatedImages:
        - mcr.microsoft.com/k8s/azureserviceoperator:1.0.59040
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 1.0.59040
      entries:
      - name: azure-service-operator.v1.0.59040
        version: 1.0.59040
      name: stable
    defaultChannel: stable
    packageName: azure-service-operator
    provider:
      name: Microsoft
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: service-registry-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-registry-operator.v2.0.7-0.1662549961.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-kafkasql"
                },
                "spec": {
                  "configuration": {
                    "kafkasql": {
                      "bootstrapServers": "\u003cservice name\u003e.\u003cnamespace\u003e.svc:9092"
                    },
                    "persistence": "kafkasql"
                  }
                }
              },
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-sql"
                },
                "spec": {
                  "configuration": {
                    "persistence": "sql",
                    "sql": {
                      "dataSource": {
                        "password": "\u003cpassword\u003e",
                        "url": "jdbc:postgresql://\u003cservice name\u003e.\u003cnamespace\u003e.svc:5432/\u003cdatabase name\u003e",
                        "userName": "postgres"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/integration/service-registry-rhel8-operator@sha256:6dfd539951a739ec06fb6620b79c42c6159d4ee148c0d9d336547e786cd4780c
          createdAt: "2022-06-24"
          description: Deploy and manage Red Hat Integration service registry on OpenShift.
          olm.substitutesFor: service-registry-operator.v2.0.7-0.1661784420.p
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents a Red Hat Integration Service
              Registry instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Service Registry
          Service Registry stores and retrieves API designs and event schemas, and gives you control of their evolution.
          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf - Provides a REST API and web UI to manage the artifacts and artifact meta-data - Includes Serializers and Deserializers for Kafka client integration - Configurable rules to control schema validity and evolution (compatibility) - Storage options: Kafka, PostgreSQL - Compatible with Confluent and IBM APIs - Runs on a lightweight Quarkus platform - Includes Maven plugin to integrate with Maven based builds
          ## Service Registry Operator
          Provides a quick and easy way to deploy and manage service registry on OpenShift.
          **Features**
          - Supports installation and configuration of the Registry for all storage options - Easily perform a rolling upgrade of the Registry
          ## Prerequisites
          This operator does not deploy storage for the Registry. Therefore, some storage options require that the chosen persistence service is already set up.
          ## License
          Red Hat Service Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - Service Registry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - service
        - service-registry
        links:
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/integration/service-registry-rhel8-operator@sha256:6dfd539951a739ec06fb6620b79c42c6159d4ee148c0d9d336547e786cd4780c
        - registry.redhat.io/integration/service-registry-kafkasql-rhel8@sha256:fc97a2a08b2f77d620ea86ec77bb4f750bfc0109ede3ad1080c2cb694019f8b8
        - registry.redhat.io/integration/service-registry-sql-rhel8@sha256:6c97fa8ead1f6f0323811f8f35a80ae33f0e160f6c58f2bdb31713dbeaa3f014
        version: 2.0.7+0.1662549961.p
      entries:
      - name: service-registry-operator.v2.0.7-0.1662549961.p
        version: 2.0.7+0.1662549961.p
      - name: service-registry-operator.v2.0.7-0.1661784420.p
        version: 2.0.7+0.1661784420.p
      - name: service-registry-operator.v2.0.7
        version: 2.0.7
      - name: service-registry-operator.v2.0.6
        version: 2.0.6
      - name: service-registry-operator.v2.0.5-0.1654503709.p
        version: 2.0.5+0.1654503709.p
      - name: service-registry-operator.v2.0.5-0.1652887585.p
        version: 2.0.5+0.1652887585.p
      - name: service-registry-operator.v2.0.5-0.1646309885.p
        version: 2.0.5+0.1646309885.p
      - name: service-registry-operator.v2.0.5-0.1647938417.p
        version: 2.0.5+0.1647938417.p
      - name: service-registry-operator.v2.0.5-0.1648813064.p
        version: 2.0.5+0.1648813064.p
      - name: service-registry-operator.v2.0.5-0.1651047781.p
        version: 2.0.5+0.1651047781.p
      - name: service-registry-operator.v2.0.5-0.1652429755.p
        version: 2.0.5+0.1652429755.p
      - name: service-registry-operator.v2.0.5
        version: 2.0.5
      - name: service-registry-operator.v2.0.5-0.1655396591.p
        version: 2.0.5+0.1655396591.p
      - name: service-registry-operator.v2.0.4
        version: 2.0.4
      - name: service-registry-operator.v2.0.3
        version: 2.0.3
      - name: service-registry-operator.v2.0.2
        version: 2.0.2
      - name: service-registry-operator.v2.0.1
        version: 2.0.1
      - name: service-registry-operator.v2.0.0
        version: 2.0.0
      name: 2.0.x
    - currentCSV: service-registry-operator.v2.6.13-r2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-kafkasql"
                },
                "spec": {
                  "configuration": {
                    "kafkasql": {
                      "bootstrapServers": "\u003cservice name\u003e.\u003cnamespace\u003e.svc:9092"
                    },
                    "persistence": "kafkasql"
                  }
                }
              },
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-sql"
                },
                "spec": {
                  "configuration": {
                    "persistence": "sql",
                    "sql": {
                      "dataSource": {
                        "password": "\u003cpassword\u003e",
                        "url": "jdbc:postgresql://\u003cservice name\u003e.\u003cnamespace\u003e.svc:5432/\u003cdatabase name\u003e",
                        "userName": "postgres"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/apicurio/apicurio-registry-rhel8-operator@sha256:61f00395b51819312dda7abcdaa47190a11a83b1e2b523101be0ba776b1c5b7c
          createdAt: "2025-09-15"
          description: Deploy and manage Red Hat build of Apicurio Registry on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations",
            "Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents a Red Hat build of Apicurio Registry
              instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Red Hat build of Apicurio Registry

          Red Hat build of Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Provides a REST API and web UI to manage the artifacts and artifact meta-data
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL
          - Compatible with Confluent and IBM APIs
          - Runs on a lightweight Quarkus platform
          - Includes Maven plugin to integrate with Maven based builds

          ## Red Hat build of Apicurio Registry Operator

          Provides a quick and easy way to deploy and manage Red Hat build of Apicurio Registry on OpenShift.

          **Features**
          - Supports installation and configuration of the Registry for all storage options
          - Easily perform a rolling upgrade of Red Hat build of Apicurio Registry

          ## Prerequisites

          This operator does not deploy storage for Red Hat build of Apicurio Registry. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Red Hat build of Apicurio Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/main/LICENSE)
        displayName: Red Hat build of Apicurio Registry
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - service
        - service-registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/apicurio/apicurio-registry-rhel8-operator@sha256:61f00395b51819312dda7abcdaa47190a11a83b1e2b523101be0ba776b1c5b7c
        - registry.redhat.io/apicurio/apicurio-registry-kafkasql-rhel8@sha256:f2da7cd92e126187d91043e2ae40521f8574586f8b3975be27d8370fe31b6732
        - registry.redhat.io/apicurio/apicurio-registry-sql-rhel8@sha256:a55425a9af1a16959ab5e528a5740d47ea06f6935593cc925e7727cded4e7d13
        version: 2.6.13-r2
      entries:
      - name: service-registry-operator.v2.6.13-r2
        version: 2.6.13-r2
      - name: service-registry-operator.v2.6.13-r1
        version: 2.6.13-r1
      - name: service-registry-operator.v2.6.12-r1
        version: 2.6.12-r1
      - name: service-registry-operator.v2.6.11-r1
        version: 2.6.11-r1
      - name: service-registry-operator.v2.6.10
        version: 2.6.10
      - name: service-registry-operator.v2.6.9
        version: 2.6.9
      - name: service-registry-operator.v2.6.8-0.1740477005.p
        version: 2.6.8+0.1740477005.p
      - name: service-registry-operator.v2.6.8-0.1739457262.p
        version: 2.6.8+0.1739457262.p
      - name: service-registry-operator.v2.6.8
        version: 2.6.8
      - name: service-registry-operator.v2.6.6-0.1736942038.p
        version: 2.6.6+0.1736942038.p
      - name: service-registry-operator.v2.6.6
        version: 2.6.6
      - name: service-registry-operator.v2.6.5-0.1731946870.p
        version: 2.6.5+0.1731946870.p
      - name: service-registry-operator.v2.6.5
        version: 2.6.5
      - name: service-registry-operator.v2.6.5-0.1729511936.p
        version: 2.6.5+0.1729511936.p
      - name: service-registry-operator.v2.6.5-0.1730982111.p
        version: 2.6.5+0.1730982111.p
      - name: service-registry-operator.v2.6.3
        version: 2.6.3
      - name: service-registry-operator.v2.6.1-0.1724927626.p
        version: 2.6.1+0.1724927626.p
      - name: service-registry-operator.v2.6.1
        version: 2.6.1
      - name: service-registry-operator.v2.3.5
        version: 2.3.5
      - name: service-registry-operator.v2.3.4-0.1716987426.p
        version: 2.3.4+0.1716987426.p
      - name: service-registry-operator.v2.3.4
        version: 2.3.4
      - name: service-registry-operator.v2.3.3
        version: 2.3.3
      - name: service-registry-operator.v2.3.2
        version: 2.3.2
      - name: service-registry-operator.v2.3.1-0.1706181988.p
        version: 2.3.1+0.1706181988.p
      - name: service-registry-operator.v2.3.1
        version: 2.3.1
      - name: service-registry-operator.v2.3.0
        version: 2.3.0
      - name: service-registry-operator.v2.2.3
        version: 2.2.3
      - name: service-registry-operator.v2.2.3-0.1698369839.p
        version: 2.2.3+0.1698369839.p
      - name: service-registry-operator.v2.2.2
        version: 2.2.2
      - name: service-registry-operator.v2.2.2-0.1697116886.p
        version: 2.2.2+0.1697116886.p
      - name: service-registry-operator.v2.2.1
        version: 2.2.1
      - name: service-registry-operator.v2.2.0
        version: 2.2.0
      - name: service-registry-operator.v2.1.5
        version: 2.1.5
      - name: service-registry-operator.v2.1.4
        version: 2.1.4
      - name: service-registry-operator.v2.1.3
        version: 2.1.3
      - name: service-registry-operator.v2.1.2-0.1676389596.p
        version: 2.1.2+0.1676389596.p
      - name: service-registry-operator.v2.1.2
        version: 2.1.2
      - name: service-registry-operator.v2.1.1
        version: 2.1.1
      - name: service-registry-operator.v2.1.0
        version: 2.1.0
      - name: service-registry-operator.v2.0.7
        version: 2.0.7
      - name: service-registry-operator.v2.0.7-0.1662549961.p
        version: 2.0.7+0.1662549961.p
      - name: service-registry-operator.v2.0.7-0.1661784420.p
        version: 2.0.7+0.1661784420.p
      - name: service-registry-operator.v2.0.6
        version: 2.0.6
      - name: service-registry-operator.v2.0.5-0.1648813064.p
        version: 2.0.5+0.1648813064.p
      - name: service-registry-operator.v2.0.5-0.1646309885.p
        version: 2.0.5+0.1646309885.p
      - name: service-registry-operator.v2.0.5-0.1654503709.p
        version: 2.0.5+0.1654503709.p
      - name: service-registry-operator.v2.0.5-0.1652887585.p
        version: 2.0.5+0.1652887585.p
      - name: service-registry-operator.v2.0.5-0.1652429755.p
        version: 2.0.5+0.1652429755.p
      - name: service-registry-operator.v2.0.5-0.1651047781.p
        version: 2.0.5+0.1651047781.p
      - name: service-registry-operator.v2.0.5-0.1655396591.p
        version: 2.0.5+0.1655396591.p
      - name: service-registry-operator.v2.0.5-0.1647938417.p
        version: 2.0.5+0.1647938417.p
      - name: service-registry-operator.v2.0.5
        version: 2.0.5
      - name: service-registry-operator.v2.0.4
        version: 2.0.4
      - name: service-registry-operator.v2.0.3
        version: 2.0.3
      - name: service-registry-operator.v2.0.2
        version: 2.0.2
      - name: service-registry-operator.v2.0.1
        version: 2.0.1
      - name: service-registry-operator.v2.0.0
        version: 2.0.0
      name: 2.x
    defaultChannel: 2.x
    packageName: service-registry-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-keyspaces-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-keyspaces-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keyspaces.services.k8s.aws/v1alpha1",
                "kind": "Keyspace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "keyspaces.services.k8s.aws/v1alpha1",
                "kind": "Table",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/keyspaces-controller:1.1.1
          createdAt: "2025-09-20T01:23:26Z"
          description: AWS Keyspaces controller is a service controller for managing
            Keyspaces resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Keyspace represents the state of an AWS keyspaces Keyspace
              resource.
            displayName: Keyspace
            kind: Keyspace
            name: keyspaces.keyspaces.services.k8s.aws
            version: v1alpha1
          - description: Table represents the state of an AWS keyspaces Table resource.
            displayName: Table
            kind: Table
            name: tables.keyspaces.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Keyspaces resources in AWS from within your Kubernetes cluster.

          **About Amazon Keyspaces**

          Amazon Keyspaces (for Apache Cassandra) is a scalable, highly available, and managed Apache Cassandra–compatible database service. With Amazon Keyspaces, you don’t have to provision, patch, or manage servers, and you don’t have to install, maintain, or operate software.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Keyspaces
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keyspaces
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Keyspaces Developer Resources
          url: https://aws.amazon.com/keyspaces/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: keyspaces maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/keyspaces-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-keyspaces-controller.v1.1.1
        version: 1.1.1
      - name: ack-keyspaces-controller.v1.1.0
        version: 1.1.0
      - name: ack-keyspaces-controller.v1.0.14
        version: 1.0.14
      - name: ack-keyspaces-controller.v1.0.12
        version: 1.0.12
      - name: ack-keyspaces-controller.v1.0.11
        version: 1.0.11
      - name: ack-keyspaces-controller.v1.0.10
        version: 1.0.10
      - name: ack-keyspaces-controller.v1.0.9
        version: 1.0.9
      - name: ack-keyspaces-controller.v1.0.7
        version: 1.0.7
      - name: ack-keyspaces-controller.v1.0.6
        version: 1.0.6
      - name: ack-keyspaces-controller.v1.0.5
        version: 1.0.5
      - name: ack-keyspaces-controller.v1.0.3
        version: 1.0.3
      - name: ack-keyspaces-controller.v1.0.2
        version: 1.0.2
      - name: ack-keyspaces-controller.v1.0.1
        version: 1.0.1
      - name: ack-keyspaces-controller.v1.0.0
        version: 1.0.0
      - name: ack-keyspaces-controller.v0.0.16
        version: 0.0.16
      - name: ack-keyspaces-controller.v0.0.15
        version: 0.0.15
      - name: ack-keyspaces-controller.v0.0.14
        version: 0.0.14
      - name: ack-keyspaces-controller.v0.0.13
        version: 0.0.13
      - name: ack-keyspaces-controller.v0.0.12
        version: 0.0.12
      - name: ack-keyspaces-controller.v0.0.11
        version: 0.0.11
      - name: ack-keyspaces-controller.v0.0.10
        version: 0.0.10
      - name: ack-keyspaces-controller.v0.0.9
        version: 0.0.9
      - name: ack-keyspaces-controller.v0.0.8
        version: 0.0.8
      - name: ack-keyspaces-controller.v0.0.7
        version: 0.0.7
      - name: ack-keyspaces-controller.v0.0.6
        version: 0.0.6
      - name: ack-keyspaces-controller.v0.0.5
        version: 0.0.5
      - name: ack-keyspaces-controller.v0.0.4
        version: 0.0.4
      - name: ack-keyspaces-controller.v0.0.3
        version: 0.0.3
      - name: ack-keyspaces-controller.v0.0.2
        version: 0.0.2
      - name: ack-keyspaces-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-keyspaces-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/kuadrant-operator
    name: kuadrant-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kuadrant-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "authScheme": {
                    "identity": [
                      {
                        "apiKey": {
                          "allNamespaces": true,
                          "selector": {
                            "matchLabels": {
                              "app": "toystore"
                            }
                          }
                        },
                        "credentials": {
                          "in": "authorization_header",
                          "keySelector": "APIKEY"
                        },
                        "name": "friends"
                      }
                    ],
                    "response": [
                      {
                        "json": {
                          "properties": [
                            {
                              "name": "userID",
                              "valueFrom": {
                                "authJSON": "auth.identity.metadata.annotations.secret\\.kuadrant\\.io/user-id"
                              }
                            }
                          ]
                        },
                        "name": "rate-limit-apikey",
                        "wrapper": "envoyDynamicMetadata",
                        "wrapperKey": "ext_auth_data"
                      }
                    ]
                  },
                  "rules": null,
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "rateLimits": [
                    {
                      "configurations": [
                        {
                          "actions": [
                            {
                              "metadata": {
                                "default_value": "no-user",
                                "descriptor_key": "userID",
                                "metadata_key": {
                                  "key": "envoy.filters.http.ext_authz",
                                  "path": [
                                    {
                                      "segment": {
                                        "key": "ext_auth_data"
                                      }
                                    },
                                    {
                                      "segment": {
                                        "key": "userID"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          ]
                        }
                      ],
                      "limits": [
                        {
                          "conditions": [],
                          "maxValue": 2,
                          "seconds": 10,
                          "variables": [
                            "userID"
                          ]
                        }
                      ]
                    }
                  ],
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/kuadrant-operator:v0.3.1
          createdAt: "2023-06-01T12:24:17Z"
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Enable AuthN and AuthZ based access control on workloads
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1beta1
          - description: Kuadrant is the Schema for the kuadrants API
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: Enable access control on workloads based on HTTP rate limiting
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1beta1
        description: A Kubernetes Operator to manage the lifecycle of the Kuadrant
          system
        displayName: Kuadrant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: Kuadrant Operator
          url: https://github.com/Kuadrant/kuadrant-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - quay.io/kuadrant/kuadrant-operator:v0.3.1
        - oci://quay.io/kuadrant/wasm-shim:v0.2.0
        version: 0.3.1
      entries:
      - name: kuadrant-operator.v0.3.1
        version: 0.3.1
      - name: kuadrant-operator.v0.3.0
        version: 0.3.0
      - name: kuadrant-operator.v0.2.1
        version: 0.2.1
      - name: kuadrant-operator.v0.2.0
        version: 0.2.0
      name: alpha
    - currentCSV: kuadrant-operator.v0.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSPolicy",
                "metadata": {
                  "name": "dnspolicy-sample"
                },
                "spec": {
                  "healthCheck": {
                    "endpoint": "/",
                    "protocol": "HTTP"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "TLSPolicy",
                "metadata": {
                  "name": "tlspolicy-sample"
                },
                "spec": {
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "self-signed-ca"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "kuadrant.io/v1beta2",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "rules": {
                    "authentication": {
                      "apikey": {
                        "apiKey": {
                          "selector": {}
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta2",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "limits": {
                    "toys": {
                      "rates": [
                        {
                          "duration": 1,
                          "limit": 50,
                          "unit": "minute"
                        }
                      ]
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/kuadrant-operator:v0.11.0
          createdAt: "2024-10-02T20:30:45Z"
          description: '[DEPRECATED] Follow https://docs.kuadrant.io instead'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthPolicy enables authentication and authorization for service
              workloads in a Gateway API network
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1beta2
          - description: DNSPolicy configures how North-South based traffic should
              be balanced and reach the gateways
            displayName: DNSPolicy
            kind: DNSPolicy
            name: dnspolicies.kuadrant.io
            version: v1alpha1
          - description: Kuadrant configures installations of Kuadrant Service Protection
              components
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: RateLimitPolicy enables rate limiting for service workloads
              in a Gateway API network
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1beta2
          - description: TLSPolicy provides tls for gateway listeners by managing
              the lifecycle of tls certificates
            displayName: TLSPolicy
            kind: TLSPolicy
            name: tlspolicies.kuadrant.io
            version: v1alpha1
        description: |-
          ## Deprecated
          This Kuadrant community operator is deprecated, Follow https://docs.kuadrant.io instead.
        displayName: '[DEPRECATED] Kuadrant Operator'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: Kuadrant Operator
          url: https://github.com/Kuadrant/kuadrant-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - quay.io/kuadrant/kuadrant-operator:v0.11.0
        - oci://quay.io/kuadrant/wasm-shim:v0.6.0
        version: 0.11.1
      entries:
      - name: kuadrant-operator.v0.11.1
        version: 0.11.1
      - name: kuadrant-operator.v0.11.0
        version: 0.11.0
      - name: kuadrant-operator.v0.10.0
        version: 0.10.0
      - name: kuadrant-operator.v0.7.2
        version: 0.7.2
      - name: kuadrant-operator.v0.4.1
        version: 0.4.1
      - name: kuadrant-operator.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: kuadrant-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/kuadrant-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Nutanix
      provider-url: ""
    name: nutanixcsioperator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nutanixcsioperator.v2.6.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crd.nutanix.com/v1alpha1",
                "kind": "NutanixCsiStorage",
                "metadata": {
                  "name": "nutanixcsistorage"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:859b6d1b2b36a48391e347c2a7431cafa08decafb56f801e21e544c3460a4b4b
          createdAt: "2024-10-30T17:10:05Z"
          description: An Operator for deploying and managing the Nutanix CSI Driver
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2.6.11
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi","disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/nutanix/csi-operator
          support: Nutanix Support Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a new instance of Nutanix CSI Driver
            displayName: Nutanix CSI Storage
            kind: NutanixCsiStorage
            name: nutanixcsistorages.crd.nutanix.com
            version: v1alpha1
        description: "# Overview \n\nThe Nutanix CSI Operator for Kubernetes packages,
          deploys, manages, upgrades the Nutanix CSI Driver on Kubernetes and OpenShift
          for dynamic provisioning of persistent volumes on Nutanix Enterprise Cloud
          platform.\n\nThe Nutanix Container Storage Interface (CSI) Driver for Kubernetes
          leverages Nutanix Volumes and Nutanix Files to provide scalable and persistent
          storage for stateful applications.\n\nWith the Nutanix CSI Provider you
          can:\n\n*  **Provide persistent storage to your containers**\n    * Leverage
          PVC resources to consume dynamically Nutanix storage\n    * With Files storage
          classes, applications on multiple pods can access the same storage, and
          also have the benefit of multi-pod read and write access. \n\n## Warning\n\nIf
          you upgrade from a previous version take care to validate your existing
          StorageClass parameters against the corresponding [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v2_6:CSI-Volume-Driver-v2_6).\n\n##
          Supported Access mode\n\n| Storage Class mode | ReadWriteOnce | ReadOnlyMany
          | ReadWriteMany |\n|--------------------|---------------|--------------|---------------|\n|
          Nutanix Volumes    | yes           | yes          | no            |\n| Nutanix
          Files      | yes           | yes          | yes           |\n\n\n## Supported
          orchestration platforms\nThe following table details orchestration platforms
          suitable for deployment of the Nutanix CSI driver.\n\n| Orchestration platform
          | Version | Architecture |\n|------------------------|---------|--------------|\n|
          Red Hat OpenShift      | 4.12    | x86          |\n| Red Hat OpenShift      |
          4.13    | x86          |\n| Red Hat OpenShift      | 4.14    | x86          |\n|
          Red Hat OpenShift      | 4.15    | x86          |\n| Red Hat OpenShift      |
          4.16    | x86          |\n| Red Hat OpenShift      | 4.17    | x86          |\n\n##
          Configuring k8s secret and storage class\nIn order to use this driver, deploy
          your NutanixCsiStorage instance and create the relevant storage classes
          and secrets, by following the [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v2_6:CSI-Volume-Driver-v2_6)
          or the example in below section:\n\n**1. Create a NutanixCsiStorage resource
          to deploy your driver**\n\nYou can do it directly inside the Operator UI
          with the `Create instance` button or with the following ressource\n```yaml\napiVersion:
          crd.nutanix.com/v1alpha1\nkind: NutanixCsiStorage\nmetadata:\n  name: nutanixcsistorage\n
          \ namespace: openshift-cluster-csi-drivers\nspec: {}\n```\n\n**2. Create
          a Nutanix secret**\n```yaml\napiVersion: v1\nkind: Secret\nmetadata:\n  name:
          ntnx-secret\n  namespace: openshift-cluster-csi-drivers\nstringData:\n  #
          prism-ip:prism-port:username:password.\n  key: 1.2.3.4:9440:admin:password\n```\n\n**3.
          Create storage classes**\n\n- volume mode example\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n  name: nutanix-volume\nprovisioner: csi.nutanix.com\nparameters:\n
          \ csi.storage.k8s.io/provisioner-secret-name: ntnx-secret\n  csi.storage.k8s.io/provisioner-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/node-publish-secret-name:
          ntnx-secret\n  csi.storage.k8s.io/node-publish-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/controller-expand-secret-name: ntnx-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/fstype: ext4\n  storageContainer:
          default-container\n  storageType: NutanixVolumes\n  #description: \"description
          added to each storage object created by the driver\"\n  #isSegmentedIscsiNetwork:
          \"false\"\n  #whitelistIPMode: ENABLED\n  #chapAuth: ENABLED\n  #isLVMVolume:
          \"false\"\n  #numLVMDisks: 4\nallowVolumeExpansion: true\nreclaimPolicy:
          Delete\n```\n\n- dynamic files mode example\n\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n    name: nutanix-dynfiles\nprovisioner: csi.nutanix.com\nparameters:\n
          \ dynamicProv: ENABLED\n  nfsServerName: fs\n  csi.storage.k8s.io/provisioner-secret-name:
          ntnx-secret\n  csi.storage.k8s.io/provisioner-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/node-publish-secret-name: ntnx-secret\n  csi.storage.k8s.io/node-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-expand-secret-name:
          ntnx-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace: openshift-cluster-csi-drivers\n
          \ storageType: NutanixFiles\n  squashType: \"none\"\n  #description: \"description
          added to each storage object created by the driver\"\nallowVolumeExpansion:
          true\n```"
        displayName: Nutanix CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Nutanix
        - CSI
        - Volumes
        - Files
        - Storage
        links:
        - name: Documentation
          url: https://portal.nutanix.com/page/documents/list?type=software&filterKey=product&filterVal=CSI
        maintainers:
        - email: cloudnative@nutanix.com
          name: Nutanix Cloud Native
        maturity: beta
        provider:
          name: Nutanix
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:9af9bf28430b00a0cedeb2ec29acadce45e6afcecd8bdf31c793c624cfa75fa7
        - registry.k8s.io/sig-storage/livenessprobe@sha256:82adbebdf5d5a1f40f246aef8ddbee7f89dea190652aefe83336008e69f9a89f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:e998f22243869416f9860fc6a1fb07d4202eac8846defc1b85ebd015c1207605
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:76a47d100de1f2afa23803bec503d9440605db03c7179b231f963e5663a3d12a
        - registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:859b6d1b2b36a48391e347c2a7431cafa08decafb56f801e21e544c3460a4b4b
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:2cddcc716c1930775228d56b0d2d339358647629701047edfdad5fcdfaf4ebcb
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:04c55b93a032ad8075cae540d60fc740a72c31006590dbdb61ad4ab9813959d0
        - registry.connect.redhat.com/nutanix/ntnx-csi-os@sha256:92962320ad682a7113aeadc8f9de1c7849996e06819acf00925aa8aa3bf51068
        version: 2.6.11
      entries:
      - name: nutanixcsioperator.v2.6.11
        version: 2.6.11
      name: stable
    - currentCSV: nutanixcsioperator.v3.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crd.nutanix.com/v1alpha1",
                "kind": "NutanixCsiStorage",
                "metadata": {
                  "name": "nutanixcsistorage"
                },
                "spec": {
                  "ntnxInitConfigMap": {
                    "usePC": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:9cc0ffc6e9c0a82320e207dad6b6b3ddc30ab8f83fa9e099a34450d5537ca435
          createdAt: "2025-06-18T11:04:03Z"
          description: An Operator for deploying and managing the Nutanix CSI Driver
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.3.4
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi","disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/nutanix/csi-operator
          support: Nutanix Support Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a new instance of Nutanix CSI Driver
            displayName: Nutanix CSI Storage
            kind: NutanixCsiStorage
            name: nutanixcsistorages.crd.nutanix.com
            version: v1alpha1
        description: "# Important\nPlease read the following instructions carefully
          before upgrading from 2.6.x to 3.3.4, for more information please refer
          to [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)
          \n- Please do not upgrade to the CSI 3.x operator if:\n    * You are using
          LVM volumes.\n    \n- To upgrade from the CSI 2.6.x to CSI 3.3 (interacting
          with Prism Central) operator\n    * Create a Nutanix Prism Central secret
          as explained below.\n    * Delete the csidriver object from the cluster:\n\n
          \       ```\n        oc delete csidriver csi.nutanix.com\n        ```\n\n
          \   * In the installed operators, go to Nutanix CSI Operator and change
          the subscription channel from stable to stable-3.x. \n    If you have installed
          the operator with automatic update approval, the operator will be automatically
          upgraded to CSI 3.3, and then the nutanixcsistorage resource will be upgraded.\n
          \   An update plan will be generated for manual updates. Upon approval,
          the operator will be successfully upgraded.\n    \n- Direct upgrades from
          CSI 2.6.x to CSI 3.3 interacting with Prism Element are not supported. \n
          \   The only solution is to recreate the nutanixcsistorage instance by following
          the below procedure:\n    - In the installed operators, go to Nutanix CSI
          Operator and delete the nutanixcsistorage instance.\n    - Next change the
          subscription channel from stable to stable-3.x.\n    - Verify the following
          points:\n        - Ensure a Nutanix Prism Element secret is present in the
          namespace.\n        - Ensure that all the storage classes with provisioner:
          csi.nutanix.com have a controller publish secret as explained below.\n\n
          \           ```\n            csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n            csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n            ```\n\n            If this secret
          is not present in the storage class please delete and recreate the storage
          classes with the required secrets.\n    - Create a new instance of nutanixcsistorage
          from this operator by specifying `usePC: false` in YAML spec section.\n
          \   - Caution: Moving from CSI driver interacting with Prism Central to
          CSI driver interacting with Prism Element is not supported.\n    \n- Troubleshooting:\n
          \   \n    If the upgrade was unsuccessful and you want to revert to version
          CSI 2.6.x, please delete the csidriver object as explained above, uninstall
          the operator (no need to delete the nutanixcsistorage custom resource),
          and install version CSI 2.6.x from the stable channel.\n# Overview \nThe
          Nutanix CSI Operator for Kubernetes packages, deploys, manages, and upgrades
          the Nutanix CSI Driver on Kubernetes and OpenShift for dynamic provisioning
          of persistent volumes on the Nutanix Enterprise Cloud platform.\nThe Nutanix
          Container Storage Interface (CSI) Driver for Kubernetes leverages Nutanix
          Volumes and Nutanix Files to provide scalable and persistent storage for
          stateful applications.\nWith the Nutanix CSI Provider, you can:\n*  **Provide
          persistent storage to your containers**\n    * Leverage PVC resources to
          consume dynamically Nutanix storage\n    * With Files storage classes, applications
          on multiple pods can access the same storage and have the benefit of multi-pod
          read and write access. \n## Warning\nIf you upgrade from a previous version,
          make sure to validate your existing StorageClass parameters against the
          corresponding [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)\n##
          Supported Access mode\n| Storage Class mode | ReadWriteOnce | ReadOnlyMany
          | ReadWriteMany |\n|--------------------|---------------|--------------|---------------|\n|
          Nutanix Volumes    | yes           | yes          | no            |\n| Nutanix
          Files      | yes           | yes          | yes           |\n## Supported
          orchestration platforms\nThe following table details orchestration platforms
          suitable for deployment of the Nutanix CSI driver.\n| Orchestration platform
          | Version | Architecture |\n|------------------------|---------|--------------|\n|
          Red Hat OpenShift      | 4.12    | x86          |\n| Red Hat OpenShift      |
          4.13    | x86          |\n| Red Hat OpenShift      | 4.14    | x86          |\n|
          Red Hat OpenShift      | 4.15    | x86          |\n| Red Hat OpenShift      |
          4.16    | x86          |\n| Red Hat OpenShift      | 4.17    | x86          |\n|
          Red Hat OpenShift      | 4.18    | x86          |\n| Red Hat OpenShift      |
          4.19    | x86          |\n## Configuring k8s secret and storage class\nTo
          use this driver, create the relevant storage classes and secrets and deploy
          your NutanixCsiStorage instance by following the [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)
          or the example in below section:\n**1. Create a Nutanix PE secret**\n```yaml\napiVersion:
          v1\nkind: Secret\nmetadata:\n  name: ntnx-pe-secret\n  namespace: openshift-cluster-csi-drivers\nstringData:\n
          \ # prism-ip:prism-port:username:password.\n  key: 1.2.3.4:9440:admin:password\n
          \ files-key: \"fileserver01.sample.com:csi:password1\"\n```\n**2. Create
          a Nutanix PC secret**\n```yaml\napiVersion: v1\nkind: Secret\nmetadata:\n
          \ name: ntnx-pc-secret\n  namespace: openshift-cluster-csi-drivers\nstringData:\n
          \ # prism-ip:prism-port:username:password.\n  key: 1.2.3.4:9440:admin:password\n```\n**3.
          Create a NutanixCsiStorage resource to deploy your driver**\nYou can do
          it directly inside the Operator UI with the `Create instance` button or
          with the following resource\n```yaml\napiVersion: crd.nutanix.com/v1alpha1\nkind:
          NutanixCsiStorage\nmetadata:\n  name: nutanixcsistorage\n  namespace: openshift-cluster-csi-drivers\nspec:
          \n  ntnxInitConfigMap:\n    usePC : true (false for PE based installations.
          Default is true.)\n```\n**4. Create storage classes**\n- volume mode example
          for 3.x installation\n```yaml\nkind: StorageClass\napiVersion: storage.k8s.io/v1\nmetadata:\n
          \ name: nutanix-volume\nprovisioner: csi.nutanix.com\nparameters:\n  csi.storage.k8s.io/provisioner-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/node-publish-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-expand-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/fstype: ext4\n  storageContainer:
          default-container\n  storageType: NutanixVolumes\n  #description: \"description
          added to each storage object created by the driver\"\n  #hypervisorAttached:
          ENABLED\nallowVolumeExpansion: true\nreclaimPolicy: Delete\n```\n- volume
          mode example for 2.x installation\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n  name: nutanix-volume\nprovisioner: csi.nutanix.com\nparameters:\n
          \ csi.storage.k8s.io/provisioner-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/node-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/controller-expand-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/fstype: ext4\n   storageContainer:
          default-container\n   storageType: NutanixVolumes\n   #description: \"description
          added to each storage object created by the driver\"\n   #isSegmentedIscsiNetwork:
          \"false\"\n   #whitelistIPMode: ENABLED\n   #chapAuth: ENABLED\n   #isLVMVolume:
          \"false\"\n   #numLVMDisks: 4\n allowVolumeExpansion: true\n reclaimPolicy:
          Delete\n```\n- dynamic files mode example\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n    name: nutanix-dynfiles\nprovisioner: csi.nutanix.com\nparameters:\n
          \ dynamicProv: ENABLED\n  nfsServerName: fs\n  csi.storage.k8s.io/provisioner-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/node-publish-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-expand-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  storageType: NutanixFiles\n  squashType:
          \"none\"\n  #description: \"description added to each storage object created
          by the driver\"\nallowVolumeExpansion: true\n```"
        displayName: Nutanix CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Nutanix
        - CSI
        - Volumes
        - Files
        - Storage
        links:
        - name: Documentation
          url: https://portal.nutanix.com/page/documents/list?type=software&filterKey=product&filterVal=CSI
        maintainers:
        - email: cloudnative@nutanix.com
          name: Nutanix Cloud Native
        maturity: beta
        provider:
          name: Nutanix
        relatedImages:
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:9af9bf28430b00a0cedeb2ec29acadce45e6afcecd8bdf31c793c624cfa75fa7
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:04c55b93a032ad8075cae540d60fc740a72c31006590dbdb61ad4ab9813959d0
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:9e581022a4e063b9d8b4f9466a36eabe4d0ced3badc8e353f13b700074ce4553
        - registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:9cc0ffc6e9c0a82320e207dad6b6b3ddc30ab8f83fa9e099a34450d5537ca435
        - registry.connect.redhat.com/nutanix/ntnx-csi-os@sha256:f0e191e7bf98829dbf718978fc971444436a1ecb9957ee22a5d8f9ba01a931b7
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0f64602ea791246712b51df334bbd701a0f31df9950a4cb9c28c059f367baa9e
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:f1bd6ee18c4021c1c94f29edfab89b49b6a4d1b800936c19dbef2d75f8202f2d
        - registry.connect.redhat.com/nutanix/ntnx-csi-os@sha256:9f121929227acf5581a813dc2f952105f517647c11a39cd7ea0863e768082586
        - registry.k8s.io/sig-storage/livenessprobe@sha256:82adbebdf5d5a1f40f246aef8ddbee7f89dea190652aefe83336008e69f9a89f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:3c116f543f0590aeff3299c8bb0683f250817d11a77d9e9071b15a0bffdabcd9
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:10624570c0aceb03f55f1eb07147b0c537e4676869cca2e9bd4bab113f810ac4
        - registry.k8s.io/sig-storage/csi-attacher@sha256:d7325367ab72b2d469a5091d87b4fc01142d2d13d1a28b2defbbe3e6fdbc4611
        version: 3.3.4
      entries:
      - name: nutanixcsioperator.v3.3.4
        version: 3.3.4
      name: stable-3.x
    defaultChannel: stable-3.x
    packageName: nutanixcsioperator
    provider:
      name: Nutanix
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/confidential-containers
    name: trustee-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trustee-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          containerImage: registry.redhat.io/build-of-trustee/trustee-rhel9-operator@sha256:fe8e23de14ad088c79b1bd5e7aab08a1b88a86679effcbf174be81e761bd9d6d
          createdAt: "2024-06-07T10:06:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trustee-operator-system
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/trustee-operator
          support: Confidential Containers Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KbsConfig is the Schema for the kbsconfigs API
            displayName: Kbs Config
            kind: KbsConfig
            name: kbsconfigs.confidentialcontainers.org
            version: v1alpha1
        description: Remote attestation operator for confidential containers
        displayName: Red Hat build of Trustee
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trustee
        - Confidential Computing
        - Attestation
        links:
        - name: Red Hat build of Trustee
          url: https://github.com/openshift/trustee-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
          url: https://github.com/confidential-containers
        relatedImages:
        - registry.redhat.io/build-of-trustee/trustee-rhel9-operator@sha256:fe8e23de14ad088c79b1bd5e7aab08a1b88a86679effcbf174be81e761bd9d6d
        version: 0.4.2
      entries:
      - name: trustee-operator.v0.4.2
        version: 0.4.2
      - name: trustee-operator.v0.4.1
        version: 0.4.1
      - name: trustee-operator.v0.4.0
        version: 0.4.0
      - name: trustee-operator.v0.3.0
        version: 0.3.0
      - name: trustee-operator.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: trustee-operator
    provider:
      name: Red Hat
      url: https://github.com/confidential-containers
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: candidate
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: fast
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: stable
    defaultChannel: stable
    packageName: stackgres
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NooBaa
      provider-url: ""
    name: mcg-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mcg-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[{"kind":"NooBaa","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"noobaa","creationTimestamp":null},"spec":{"cleanupPolicy":{},"security":{"kms":{}},"loadBalancerSourceSubnets":{},"autoscaler":{}},"status":{"lastKeyRotateTime":null}},{"kind":"BackingStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"NamespaceStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"BucketClass","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{},"status":{}},{"kind":"NooBaaAccount","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"allow_bucket_creation":false},"status":{}}]'
          capabilities: Deep Insights
          categories: Storage,Big Data
          certified: "false"
          containerImage: registry.redhat.io/odf4/mcg-rhel9-operator@sha256:43ddea9d3c1cde5c8b38f74e8840a871dd0441ed9fd074528bacf78b439f024f
          createdAt: "2019-07-08T13:10:20.940Z"
          description: NooBaa is an object data service for hybrid and multi cloud
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operators.openshift.io/infrastructure-features: ֿ'["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/noobaa/noobaa-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A NooBaa system - Create this to start
            displayName: NooBaa
            kind: NooBaa
            name: noobaas.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos,
              PV's and more. Used in BucketClass to construct data placement policies.
            displayName: Backing Store
            kind: BackingStore
            name: backingstores.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos
              and more. Used in BucketClass to construct namespace policies.
            displayName: Namespace Store
            kind: NamespaceStore
            name: namespacestores.noobaa.io
            version: v1alpha1
          - description: Storage policy spec  tiering, mirroring, spreading, namespace
              policy. Combines BackingStores Or NamespaceStores. Referenced by ObjectBucketClaims.
            displayName: Bucket Class
            kind: BucketClass
            name: bucketclasses.noobaa.io
            version: v1alpha1
          - kind: NooBaaAccount
            name: noobaaaccounts.noobaa.io
            version: v1alpha1
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information.
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: "The noobaa operator creates and reconciles a NooBaa system in
          a Kubernetes/Openshift cluster.\n\nNooBaa provides an S3 object-store service
          abstraction and data placement policies to create hybrid and multi cloud
          data solutions.\n\nFor more information on using NooBaa refer to [Github](https://github.com/noobaa/noobaa-core)
          / [Website](https://www.noobaa.io) / [Articles](https://noobaa.desk.com).
          \n\n## How does it work?\n\n- The operator deploys the noobaa core pod and
          two services - Mgmt (UI/API) and S3 (object-store).\n- Both services require
          credentials which you will get from a secret that the operator creates -
          use describe noobaa to locate it.\n- The service addresses will also appear
          in the describe output - pick the one that is suitable for your client:\n
          \   - minikube - use the NodePort address.\n    - remote cluster - probably
          need one of the External addresses.\n    - connect an application on the
          same cluster - use Internal DNS (though any address should work)\n    \n-
          Feel free to email us or open github issues on any question.\n\n## Getting
          Started\n\n### Notes:\n- The following instructions are for **minikube**
          but it works on any Kubernetes/Openshift clusters.\n- This will setup noobaa
          in the **my-noobaa-operator** namespace.\n- You will need **jq**, **curl**,
          **kubectl** or **oc**, **aws-cli**.\n\n### 1. Install OLM (if you don't
          have it already):\n```\ncurl -sL https://github.com/operator-framework/operator-lifecycle-manager/releases/download/0.12.0/install.sh
          | bash -s 0.12.0\n```\n\n### 2. Install noobaa-operator:\n```\nkubectl create
          -f https://operatorhub.io/install/noobaa-operator.yaml\n```\nWait for it
          to be ready:\n```\nkubectl wait pod -n my-noobaa-operator -l noobaa-operator
          --for=condition=ready\n```\n\n### 3. Create noobaa system:\n```\ncurl -sL
          https://operatorhub.io/api/operator?packageName=noobaa-operator | \n    jq
          '.operator.customResourceDefinitions[0].yamlExample | .metadata.namespace=\"my-noobaa-operator\"'
          |\n    kubectl create -f -\n```\nWait for it to be ready:\n```\nkubectl
          wait pod -n my-noobaa-operator -l noobaa-core --for=condition=ready\nkubectl
          get noobaa -n my-noobaa-operator -w\n# NAME     PHASE   MGMT-ENDPOINTS                  S3-ENDPOINTS
          \                   IMAGE                    AGE\n# noobaa   **Ready**   [https://192.168.64.12:31121]
          \  [https://192.168.64.12:32557]   noobaa/noobaa-core:4.0   19m\n```\n\n###
          4. Get system information to your shell:\n```\nNOOBAA_SECRET=$(kubectl get
          noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.accounts.admin.secretRef.name'
          )\nNOOBAA_MGMT=$(kubectl get noobaa noobaa -n my-noobaa-operator -o json
          | jq -r '.status.services.serviceMgmt.nodePorts[0]' )\nNOOBAA_S3=$(kubectl
          get noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.services.serviceS3.nodePorts[0]'
          )\nNOOBAA_ACCESS_KEY=$(kubectl get secret $NOOBAA_SECRET -n my-noobaa-operator
          -o json | jq -r '.data.AWS_ACCESS_KEY_ID|@base64d')\nNOOBAA_SECRET_KEY=$(kubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq -r '.data.AWS_SECRET_ACCESS_KEY|@base64d')\n```\n\n###
          5. Connect to Mgmt UI:\n```\n# show email/password from the secret:\nkubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq '.data|map_values(@base64d)'\n\n#
          open mgmt UI login:\nopen $NOOBAA_MGMT\n```\n\n### 6. Connect to S3 with
          aws-cli:\n```\nalias s3='AWS_ACCESS_KEY_ID=$NOOBAA_ACCESS_KEY AWS_SECRET_ACCESS_KEY=$NOOBAA_SECRET_KEY
          aws --endpoint $NOOBAA_S3 --no-verify-ssl s3'\ns3 ls\ns3 sync /var/log/
          s3://first.bucket\ns3 ls s3://first.bucket\n```\n"
        displayName: NooBaa Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - noobaa
        - kubernetes
        - openshift
        - cloud
        - hybrid
        - multi
        - data
        - storage
        - s3
        - tiering
        - mirroring
        links:
        - name: Github
          url: https://github.com/noobaa/noobaa-core
        - name: Website
          url: https://www.noobaa.io
        - name: Articles
          url: https://noobaa.desk.com
        maintainers:
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: etamir@redhat.com
          name: Eran Tamir
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: NooBaa
        relatedImages:
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:f37e2c611f4cda32172d4e8938938e488076fd8a8f19c252c56a80c48044bc02
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhel8/postgresql-12@sha256:af1ab437b165fa4ee9dfd6ade637d7de5fa556d0b3851fb8edc78afe0e6fa95e
        - registry.redhat.io/odf4/mcg-rhel9-operator@sha256:43ddea9d3c1cde5c8b38f74e8840a871dd0441ed9fd074528bacf78b439f024f
        - registry.redhat.io/odf4/odf-cosi-sidecar-rhel9@sha256:bce36c7d33a85a11cd6b6b24f92887cffe6f38af70340e54d10e4f6f524aa782
        version: 4.15.20-rhodf
      entries:
      - name: mcg-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: mcg-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: mcg-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: mcg-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: mcg-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: mcg-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: mcg-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: mcg-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: mcg-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: mcg-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: mcg-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: mcg-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: mcg-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: mcg-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: mcg-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: mcg-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: mcg-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: mcg-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: mcg-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: mcg-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: mcg-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[{"kind":"NooBaa","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"noobaa","creationTimestamp":null},"spec":{"cleanupPolicy":{},"security":{"kms":{}},"loadBalancerSourceSubnets":{},"autoscaler":{}},"status":{"lastKeyRotateTime":null}},{"kind":"BackingStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"NamespaceStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"BucketClass","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{},"status":{}},{"kind":"NooBaaAccount","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"allow_bucket_creation":false},"status":{}}]'
          capabilities: Deep Insights
          categories: Storage,Big Data
          certified: "false"
          containerImage: registry.redhat.io/odf4/mcg-rhel9-operator@sha256:fdcdc8870b12d853973000877818463c3b4a5ed1d3316ff72a04a299577ed23d
          createdAt: "2019-07-08T13:10:20.940Z"
          description: NooBaa is an object data service for hybrid and multi cloud
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/infrastructure-features: ֿ'["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/noobaa/noobaa-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A NooBaa system - Create this to start
            displayName: NooBaa
            kind: NooBaa
            name: noobaas.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos,
              PV's and more. Used in BucketClass to construct data placement policies.
            displayName: Backing Store
            kind: BackingStore
            name: backingstores.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos
              and more. Used in BucketClass to construct namespace policies.
            displayName: Namespace Store
            kind: NamespaceStore
            name: namespacestores.noobaa.io
            version: v1alpha1
          - description: Storage policy spec  tiering, mirroring, spreading, namespace
              policy. Combines BackingStores Or NamespaceStores. Referenced by ObjectBucketClaims.
            displayName: Bucket Class
            kind: BucketClass
            name: bucketclasses.noobaa.io
            version: v1alpha1
          - kind: NooBaaAccount
            name: noobaaaccounts.noobaa.io
            version: v1alpha1
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information.
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: "The noobaa operator creates and reconciles a NooBaa system in
          a Kubernetes/Openshift cluster.\n\nNooBaa provides an S3 object-store service
          abstraction and data placement policies to create hybrid and multi cloud
          data solutions.\n\nFor more information on using NooBaa refer to [Github](https://github.com/noobaa/noobaa-core)
          / [Website](https://www.noobaa.io) / [Articles](https://noobaa.desk.com).
          \n\n## How does it work?\n\n- The operator deploys the noobaa core pod and
          two services - Mgmt (UI/API) and S3 (object-store).\n- Both services require
          credentials which you will get from a secret that the operator creates -
          use describe noobaa to locate it.\n- The service addresses will also appear
          in the describe output - pick the one that is suitable for your client:\n
          \   - minikube - use the NodePort address.\n    - remote cluster - probably
          need one of the External addresses.\n    - connect an application on the
          same cluster - use Internal DNS (though any address should work)\n    \n-
          Feel free to email us or open github issues on any question.\n\n## Getting
          Started\n\n### Notes:\n- The following instructions are for **minikube**
          but it works on any Kubernetes/Openshift clusters.\n- This will setup noobaa
          in the **my-noobaa-operator** namespace.\n- You will need **jq**, **curl**,
          **kubectl** or **oc**, **aws-cli**.\n\n### 1. Install OLM (if you don't
          have it already):\n```\ncurl -sL https://github.com/operator-framework/operator-lifecycle-manager/releases/download/0.12.0/install.sh
          | bash -s 0.12.0\n```\n\n### 2. Install noobaa-operator:\n```\nkubectl create
          -f https://operatorhub.io/install/noobaa-operator.yaml\n```\nWait for it
          to be ready:\n```\nkubectl wait pod -n my-noobaa-operator -l noobaa-operator
          --for=condition=ready\n```\n\n### 3. Create noobaa system:\n```\ncurl -sL
          https://operatorhub.io/api/operator?packageName=noobaa-operator | \n    jq
          '.operator.customResourceDefinitions[0].yamlExample | .metadata.namespace=\"my-noobaa-operator\"'
          |\n    kubectl create -f -\n```\nWait for it to be ready:\n```\nkubectl
          wait pod -n my-noobaa-operator -l noobaa-core --for=condition=ready\nkubectl
          get noobaa -n my-noobaa-operator -w\n# NAME     PHASE   MGMT-ENDPOINTS                  S3-ENDPOINTS
          \                   IMAGE                    AGE\n# noobaa   **Ready**   [https://192.168.64.12:31121]
          \  [https://192.168.64.12:32557]   noobaa/noobaa-core:4.0   19m\n```\n\n###
          4. Get system information to your shell:\n```\nNOOBAA_SECRET=$(kubectl get
          noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.accounts.admin.secretRef.name'
          )\nNOOBAA_MGMT=$(kubectl get noobaa noobaa -n my-noobaa-operator -o json
          | jq -r '.status.services.serviceMgmt.nodePorts[0]' )\nNOOBAA_S3=$(kubectl
          get noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.services.serviceS3.nodePorts[0]'
          )\nNOOBAA_ACCESS_KEY=$(kubectl get secret $NOOBAA_SECRET -n my-noobaa-operator
          -o json | jq -r '.data.AWS_ACCESS_KEY_ID|@base64d')\nNOOBAA_SECRET_KEY=$(kubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq -r '.data.AWS_SECRET_ACCESS_KEY|@base64d')\n```\n\n###
          5. Connect to Mgmt UI:\n```\n# show email/password from the secret:\nkubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq '.data|map_values(@base64d)'\n\n#
          open mgmt UI login:\nopen $NOOBAA_MGMT\n```\n\n### 6. Connect to S3 with
          aws-cli:\n```\nalias s3='AWS_ACCESS_KEY_ID=$NOOBAA_ACCESS_KEY AWS_SECRET_ACCESS_KEY=$NOOBAA_SECRET_KEY
          aws --endpoint $NOOBAA_S3 --no-verify-ssl s3'\ns3 ls\ns3 sync /var/log/
          s3://first.bucket\ns3 ls s3://first.bucket\n```\n"
        displayName: NooBaa Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - noobaa
        - kubernetes
        - openshift
        - cloud
        - hybrid
        - multi
        - data
        - storage
        - s3
        - tiering
        - mirroring
        links:
        - name: Github
          url: https://github.com/noobaa/noobaa-core
        - name: Website
          url: https://www.noobaa.io
        - name: Articles
          url: https://noobaa.desk.com
        maintainers:
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: etamir@redhat.com
          name: Eran Tamir
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: NooBaa
        relatedImages:
        - registry.redhat.io/odf4/mcg-rhel9-operator@sha256:fdcdc8870b12d853973000877818463c3b4a5ed1d3316ff72a04a299577ed23d
        - registry.redhat.io/odf4/odf-cosi-sidecar-rhel9@sha256:9fed589373d1586577d57f23b7a4f39ace5052fc55c99a03cb256da63a26c441
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:0084d9053ad2c74926ca02d2ede0461abd59227eee98d122c45e08a5ebaf629a
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhel8/postgresql-12@sha256:af1ab437b165fa4ee9dfd6ade637d7de5fa556d0b3851fb8edc78afe0e6fa95e
        version: 4.16.16-rhodf
      entries:
      - name: mcg-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: mcg-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: mcg-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: mcg-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: mcg-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: mcg-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: mcg-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: mcg-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: mcg-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: mcg-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: mcg-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: mcg-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: mcg-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: mcg-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: mcg-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: mcg-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: mcg-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: mcg-operator
    provider:
      name: NooBaa
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-eks-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-eks-controller.v1.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "FargateProfile",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Nodegroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/eks-controller:1.9.3
          createdAt: "2025-10-03T21:43:09Z"
          description: AWS EKS controller is a service controller for managing EKS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessEntry represents the state of an AWS eks AccessEntry
              resource.
            displayName: AccessEntry
            kind: AccessEntry
            name: accessentries.eks.services.k8s.aws
            version: v1alpha1
          - description: Addon represents the state of an AWS eks Addon resource.
            displayName: Addon
            kind: Addon
            name: addons.eks.services.k8s.aws
            version: v1alpha1
          - description: Cluster represents the state of an AWS eks Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.eks.services.k8s.aws
            version: v1alpha1
          - description: FargateProfile represents the state of an AWS eks FargateProfile
              resource.
            displayName: FargateProfile
            kind: FargateProfile
            name: fargateprofiles.eks.services.k8s.aws
            version: v1alpha1
          - description: IdentityProviderConfig represents the state of an AWS eks
              IdentityProviderConfig resource.
            displayName: IdentityProviderConfig
            kind: IdentityProviderConfig
            name: identityproviderconfigs.eks.services.k8s.aws
            version: v1alpha1
          - description: Nodegroup represents the state of an AWS eks Nodegroup resource.
            displayName: Nodegroup
            kind: Nodegroup
            name: nodegroups.eks.services.k8s.aws
            version: v1alpha1
          - description: PodIdentityAssociation represents the state of an AWS eks
              PodIdentityAssociation resource.
            displayName: PodIdentityAssociation
            kind: PodIdentityAssociation
            name: podidentityassociations.eks.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Kubernetes Service (EKS) resources in AWS from within your Kubernetes cluster.

          **About Amazon EKS**

          Amazon Elastic Kubernetes Service (Amazon EKS) is a managed service that you can use to run Kubernetes on AWS without needing to install, operate, and maintain your own Kubernetes control plane or nodes. Kubernetes is an open-source system for automating the deployment, scaling, and management of containerized applications.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EKS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eks
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EKS Developer Resources
          url: https://aws.amazon.com/eks/getting-started/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/eks-controller:1.9.3
        version: 1.9.3
      entries:
      - name: ack-eks-controller.v1.9.3
        version: 1.9.3
      - name: ack-eks-controller.v1.9.2
        version: 1.9.2
      - name: ack-eks-controller.v1.9.1
        version: 1.9.1
      - name: ack-eks-controller.v1.9.0
        version: 1.9.0
      - name: ack-eks-controller.v1.8.0
        version: 1.8.0
      - name: ack-eks-controller.v1.7.6
        version: 1.7.6
      - name: ack-eks-controller.v1.7.5
        version: 1.7.5
      - name: ack-eks-controller.v1.7.4
        version: 1.7.4
      - name: ack-eks-controller.v1.7.3
        version: 1.7.3
      - name: ack-eks-controller.v1.7.2
        version: 1.7.2
      - name: ack-eks-controller.v1.6.4
        version: 1.6.4
      - name: ack-eks-controller.v1.6.3
        version: 1.6.3
      - name: ack-eks-controller.v1.6.2
        version: 1.6.2
      - name: ack-eks-controller.v1.6.1
        version: 1.6.1
      - name: ack-eks-controller.v1.6.0
        version: 1.6.0
      - name: ack-eks-controller.v1.5.5
        version: 1.5.5
      - name: ack-eks-controller.v1.5.4
        version: 1.5.4
      - name: ack-eks-controller.v1.5.3
        version: 1.5.3
      - name: ack-eks-controller.v1.5.2
        version: 1.5.2
      - name: ack-eks-controller.v1.5.1
        version: 1.5.1
      - name: ack-eks-controller.v1.5.0
        version: 1.5.0
      - name: ack-eks-controller.v1.4.7
        version: 1.4.7
      - name: ack-eks-controller.v1.4.6
        version: 1.4.6
      - name: ack-eks-controller.v1.4.5
        version: 1.4.5
      - name: ack-eks-controller.v1.4.4
        version: 1.4.4
      - name: ack-eks-controller.v1.4.3
        version: 1.4.3
      - name: ack-eks-controller.v1.4.2
        version: 1.4.2
      - name: ack-eks-controller.v1.4.1
        version: 1.4.1
      - name: ack-eks-controller.v1.4.0
        version: 1.4.0
      - name: ack-eks-controller.v1.3.5
        version: 1.3.5
      - name: ack-eks-controller.v1.3.3
        version: 1.3.3
      - name: ack-eks-controller.v1.3.1
        version: 1.3.1
      - name: ack-eks-controller.v1.3.0
        version: 1.3.0
      - name: ack-eks-controller.v1.2.3
        version: 1.2.3
      - name: ack-eks-controller.v1.2.2
        version: 1.2.2
      - name: ack-eks-controller.v1.2.1
        version: 1.2.1
      - name: ack-eks-controller.v1.2.0
        version: 1.2.0
      - name: ack-eks-controller.v1.1.3
        version: 1.1.3
      - name: ack-eks-controller.v1.1.2
        version: 1.1.2
      - name: ack-eks-controller.v1.1.1
        version: 1.1.1
      - name: ack-eks-controller.v1.1.0
        version: 1.1.0
      - name: ack-eks-controller.v1.0.5
        version: 1.0.5
      - name: ack-eks-controller.v1.0.4
        version: 1.0.4
      - name: ack-eks-controller.v1.0.3
        version: 1.0.3
      - name: ack-eks-controller.v1.0.2
        version: 1.0.2
      - name: ack-eks-controller.v1.0.1
        version: 1.0.1
      - name: ack-eks-controller.v1.0.0
        version: 1.0.0
      - name: ack-eks-controller.v0.1.9
        version: 0.1.9
      - name: ack-eks-controller.v0.1.8
        version: 0.1.8
      - name: ack-eks-controller.v0.1.7
        version: 0.1.7
      - name: ack-eks-controller.v0.1.6
        version: 0.1.6
      - name: ack-eks-controller.v0.1.5
        version: 0.1.5
      - name: ack-eks-controller.v0.1.4
        version: 0.1.4
      - name: ack-eks-controller.v0.1.3
        version: 0.1.3
      - name: ack-eks-controller.v0.1.2
        version: 0.1.2
      - name: ack-eks-controller.v0.1.1
        version: 0.1.1
      - name: ack-eks-controller.v0.1.0
        version: 0.1.0
      - name: ack-eks-controller.v0.0.15
        version: 0.0.15
      - name: ack-eks-controller.v0.0.14
        version: 0.0.14
      - name: ack-eks-controller.v0.0.13
        version: 0.0.13
      - name: ack-eks-controller.v0.0.12
        version: 0.0.12
      - name: ack-eks-controller.v0.0.11
        version: 0.0.11
      - name: ack-eks-controller.v0.0.10
        version: 0.0.10
      - name: ack-eks-controller.v0.0.9
        version: 0.0.9
      - name: ack-eks-controller.v0.0.8
        version: 0.0.8
      - name: ack-eks-controller.v0.0.7
        version: 0.0.7
      - name: ack-eks-controller.v0.0.6
        version: 0.0.6
      - name: ack-eks-controller.v0.0.5
        version: 0.0.5
      - name: ack-eks-controller.v0.0.4
        version: 0.0.4
      - name: ack-eks-controller.v0.0.3
        version: 0.0.3
      - name: ack-eks-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-eks-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Carlos Eduardo de Paula
      provider-url: ""
    name: lbconfig-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lbconfig-operator.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-infra-sample",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "http",
                    "path": "/healthz",
                    "port": 1936
                  },
                  "ports": [
                    80,
                    443
                  ],
                  "provider": {
                    "creds": "f5-creds",
                    "host": "https://192.168.1.35",
                    "partition": "Common",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "F5_BigIP"
                  },
                  "type": "infra",
                  "vip": "192.168.1.45"
                }
              },
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-infra-sample-shard",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "http",
                    "path": "/healthz",
                    "port": 1936
                  },
                  "nodelabels": {
                    "kubernetes.io/region": "DC1",
                    "node-role.kubernetes.io/infra": ""
                  },
                  "ports": [
                    80,
                    443
                  ],
                  "provider": {
                    "creds": "f5-creds",
                    "host": "https://192.168.1.35",
                    "partition": "Common",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "F5_BigIP"
                  },
                  "type": "infra",
                  "vip": "192.168.1.46"
                }
              },
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-master-sample",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "https",
                    "path": "/healthz",
                    "port": 6443
                  },
                  "ports": [
                    6443
                  ],
                  "provider": {
                    "creds": "netscaler-creds",
                    "host": "https://192.168.1.36",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "Citrix_ADC"
                  },
                  "type": "master",
                  "vip": "192.168.1.40"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Networking
          certified: "false"
          containerImage: quay.io/carlosedp/lbconfig-operator:v0.5.1
          createdAt: "2025-01-09T15:45:28Z"
          description: Manage External Load Balancers allowing creation/update for
            VIPs and Servers dynamically via API.
          k8sMaxVersion: ""
          k8sMinVersion: ""
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/carlosedp/lbconfig-operator
          support: https://github.com/carlosedp/lbconfig-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalLoadBalancer is the Schema for the externalloadbalancers
              API
            displayName: ExternalLoadBalancer Instance
            kind: ExternalLoadBalancer
            name: externalloadbalancers.lb.lbconfig.carlosedp.com
            version: v1
          required:
          - description: ExternalLoadBalancer represents a configured instance of
              an external Load-Balancer for a specific group of nodes of the cluster.
              The Instance has a VIP and ports to be balanced to the cluster nodes
              based on a set of node labels.
            displayName: External Load-Balancer Configuration Instance
            kind: ExternalLoadBalancer
            name: externalloadbalancers.lb.lbconfig.carlosedp.com
            version: v1
        description: |
          ## About the Operator

          The LBConfig Operator, manages the configuration of External Load Balancer instances (on third-party equipment via it's API) and creates VIPs and IP Pools with Monitors for a set of OpenShift or Kubernetes nodes like Master-nodes (Control-Plane), Infra nodes (where the Routers or Ingress controllers are located) or based on it's roles and/or labels.

          The operator dynamically handles creating, updating or deleting the IPs of the pools in the Load Balancer based on the Node IPs for each role or label. On every change of the operator configuration (CRDs) or addition/change/removal or cluster Nodes, the operator updates the Load Balancer properly.

          ## Main Use Cases

          The main users for this operator is enterprise deployments or clusters composed of multiple nodes having an external load-balancer providing the balancing and high-availability to access the cluster in both API and Application levels.

          ### Create ExternalLoadBalancer Instances

          Create the instances for each Load Balancer instance you need (for example one for Master Nodes and another for the Infra Nodes). **If installing on OpenShift or Kubernetes with OLM (or in a different namespace), adjust the sample YAMLs to match the created namespace**.

          **The provider `vendor` field can be (case-insensitive):**

          * **`F5_BigIP`** - Tested on F5 BigIP version 15
          * **`Citrix_ADC`** - Tested on Citrix ADC (Netscaler) version 13
          * **`HAProxy`** - HAProxy with Dataplane API. ([Docs](./docs/haproxy/))
          * **`Dummy`** - Dummy backend used for testing to only print log messages on operations

          Create the secret holding the Load Balancer API user and password:

          ```sh
          oc create secret generic f5-creds --from-literal=username=admin --from-literal=password=admin123 --namespace lbconfig-operator-system
          ```

          #### Sample CRDs and Available Fields

          Master Nodes using a Citrix ADC LB:

          ```yaml
          apiVersion: lb.lbconfig.carlosedp.com/v1
          kind: ExternalLoadBalancer
          metadata:
            name: externalloadbalancer-master-sample
            namespace: lbconfig-operator-system
          spec:
            vip: "192.168.1.40"
            type: "master"
            ports:
              - 6443
            monitor:
              path: "/healthz"
              port: 6443
              monitortype: "https"
            provider:
              vendor: Citrix_ADC
              host: "https://192.168.1.36"
              port: 443
              creds: netscaler-creds
              validatecerts: false
          ```

          Infra Nodes using a F5 BigIP LB:

          ```yaml
          apiVersion: lb.lbconfig.carlosedp.com/v1
          kind: ExternalLoadBalancer
          metadata:
            name: externalloadbalancer-infra-sample
            namespace: lbconfig-operator-system
          spec:
            vip: "192.168.1.45"
            type: "infra"
            ports:
              - 80
              - 443
            monitor:
              path: "/healthz"
              port: 1936
              monitortype: http
            provider:
              vendor: F5_BigIP
              host: "https://192.168.1.35"
              port: 443
              creds: f5-creds
              partition: "Common"
              validatecerts: false
          ```

          To choose the nodes which will be part of the server pool, you can set either `type` or `nodelabels` fields. The yaml field `type: "master"` or `type: "infra"` selects nodes with the role label `"node-role.kubernetes.io/master"` and `"node-role.kubernetes.io/infra"` respectively. If the field `nodelabels` array is used instead, the operator will use nodes which match all labels.

          Clusters with sharded routers or using arbitrary labels to determine where the Ingress Controllers run can be configured like:

          ```yaml
          spec:
            vip: "10.0.0.6"
            ports:
              - 80
            nodelabels:
              "node.kubernetes.io/ingress-controller": "production"
              "kubernetes.io/region": "DC1"
            ...
          ```

          ## Getting Help

          This project fully open-source and is hosted at [GitHub](https://github.com/carlosedp/lbconfig-operator). Always check the project page and issues for the latest updates and/or solutions to problems encountered. If a problem is detected, please fill an issue. Feature request and improvement issues are welcome.

          Some fields inside `providers` are optional and depend on the used backend. Check the [API docs](https://pkg.go.dev/github.com/carlosedp/lbconfig-operator/api/lb.lbconfig.carlosedp.com/v1?utm_source=gopls#Provider) which fields are backend-specific.

          ## Disclaimers

          * The operator does not check if the requested configuration (names, IPs) already exists and/or conflicts with existing configuration in the Load Balancer. The user is responsible for these checks before deployment;
          * I am not responsible if the operator changes/deletes existing configuration on the Load Balancer if existing names are already configured.
          * The operator creates the entries(Pools, VIPs, Monitors) in the provided Load Balancer with the `name` of the instance configured in the CustomResource prefixed with the type. Eg. For a CR with name `externalloadbalancer-master-sample`, the operator creates a server pool named `Pool-externalloadbalancer-master-sample-6443` (suffixed with the port), a monitor named `Monitor-externalloadbalancer-master-sample` and a VIP named `VIP-externalloadbalancer-master-sample-6443` (suffixed with the port).
        displayName: External Load-Balancer Configuration Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 'load-balance '
        - infrastructure
        links:
        - name: GitHub
          url: https://github.com/carlosedp/lbconfig-operator
        - name: Documentation
          url: https://github.com/carlosedp/lbconfig-operator/docs
        - name: Issues
          url: https://github.com/carlosedp/lbconfig-operator/issues
        maintainers:
        - email: carlosedp@gmail.com
          name: Carlos Eduardo de Paula
        maturity: beta
        minKubeVersion: 1.18.0
        provider:
          name: Carlos Eduardo de Paula
        relatedImages:
        - quay.io/carlosedp/kube-rbac-proxy:v0.15.0
        - quay.io/carlosedp/lbconfig-operator:v0.5.1
        version: 0.5.1
      entries:
      - name: lbconfig-operator.v0.5.1
        version: 0.5.1
      - name: lbconfig-operator.v0.5.0
        version: 0.5.0
      - name: lbconfig-operator.v0.4.1
        version: 0.4.1
      - name: lbconfig-operator.v0.4.0
        version: 0.4.0
      - name: lbconfig-operator.v0.2.0
        version: 0.2.0
      name: beta
    defaultChannel: beta
    packageName: lbconfig-operator
    provider:
      name: Carlos Eduardo de Paula
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: group-sync-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: group-sync-operator.v0.0.36
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "azure-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "azure": {
                        "credentialsSecret": {
                          "name": "azure-group-sync",
                          "namespace": "group-sync-operator"
                        }
                      },
                      "name": "azure"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "github-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "github": {
                        "credentialsSecret": {
                          "name": "github-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "organization": "redhat-cop"
                      },
                      "name": "github"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "gitlab-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "gitlab": {
                        "credentialsSecret": {
                          "name": "gitlab-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "url": "https://gitlab.apps.openshift.com"
                      },
                      "name": "gitlab"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "keycloak-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "keycloak": {
                        "credentialsSecret": {
                          "name": "keycloak-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "insecure": true,
                        "realm": "ocp",
                        "url": "https://keycloak.apps.openshift.com"
                      },
                      "name": "keycloak"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "ldap-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "ldap": {
                        "credentialsSecret": {
                          "name": "ldap-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "insecure": true,
                        "rfc2307": {
                          "groupMembershipAttributes": [
                            "member"
                          ],
                          "groupNameAttributes": [
                            "cn"
                          ],
                          "groupUIDAttribute": "dn",
                          "groupsQuery": {
                            "baseDN": "ou=Groups,dc=example,dc=com",
                            "derefAliases": "never",
                            "filter": "(objectClass=groupofnames)",
                            "scope": "sub"
                          },
                          "tolerateMemberNotFoundErrors": true,
                          "tolerateMemberOutOfScopeErrors": true,
                          "userNameAttributes": [
                            "cn"
                          ],
                          "userUIDAttribute": "dn",
                          "usersQuery": {
                            "baseDN": "ou=Users,dc=example,dc=com",
                            "derefAliases": "never",
                            "scope": "sub"
                          }
                        },
                        "url": "ldap://localhost:389"
                      },
                      "name": "ldap"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "okta-sync"
                },
                "spec": {
                  "providers": [
                    {
                      "name": "okta",
                      "okta": {
                        "appId": "okta-sync-app-id",
                        "credentialsSecret": {
                          "name": "okta-api-token",
                          "namespace": "group-sync-operator"
                        },
                        "extractLoginUsername": true,
                        "profileKey": "login",
                        "url": "https://example.okta.com/"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/group-sync-operator@sha256:43ef79efc2a77ee5edd4e3123f9dab156a30d3e00fd1fba9f89600b96f0f68d6
          createdAt: "2025-08-12T19:19:21Z"
          description: Synchronize groups and users from external providers
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: group-sync-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/group-sync-operator
          support: Red Hat Community of Practice
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GroupSync is the Schema for the groupsyncs API
            displayName: Group Sync
            kind: GroupSync
            name: groupsyncs.redhatcop.redhat.io
            version: v1alpha1
        description: |
          Synchronizes groups from external providers into OpenShift

          ## Overview

          The OpenShift Container Platform contains functionality to synchronize groups found in external identity providers into the platform. Currently, the functionality that is included in OpenShift to limited to synchronizing LDAP only. This operator is designed to integrate with external providers in order to provide new solutions.

          Group Synchronization is facilitated by creating a `GroupSync` resource. The following describes the high level schema for this resource:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: example-groupsync
          spec:
            providers:
              - <One or more providers to synchronize>
          ```

          ## Authentication

          In most cases, authentication details must be provided in order to communicate with providers. Authentication details are provider specific with regards to the required values. In supported providers, the secret can be referenced in the `credentialsSecret` by name and namespace where it has been created as shown below:

          ```
          credentialsSecret:
            name: <secret_name>
            namespace: <secret_namespace>
          ```

          ## Providers

          Integration with external systems is made possible through a set of pluggable external providers. The following providers are currently supported:

          * [Azure](https://azure.microsoft.com/)
          * [GitHub](https://github.com)
          * [GitLab](https://gitlab.com)
          * [LDAP](https://en.wikipedia.org/wiki/Lightweight_Directory_Access_Protocol)
          * [Keycloak](https://www.keycloak.org/)/[Red Hat Single Sign On](https://access.redhat.com/products/red-hat-single-sign-on)
          * [Okta](https://www.okta.com/)
          * [IBM Security Verify](https://docs.verify.ibm.com/verify)

          The following sections describe the configuration options available for each provider


          ### Azure

          Groups contained within Azure Active Directory can be synchronized into OpenShift. The following table describes the set of configuration options for the Azure provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `authorityHost` | Azure Active Directory Endpoint | `https://login.microsoftonline.com` | No |
          | `baseGroups` | List of groups to start searching from instead of listing all groups in the directory | | No |
          | `credentialsSecret` | Name of the secret containing authentication details (See below) | | Yes |
          | `filter` | Graph API filter | | No |
          | `groups` | List of groups to filter against | | No |
          | `userNameAttributes` | Fields on a user record to use as the User Name | `userPrincipalName` | No |
          | `prune` | Prune Whether to prune groups that are no longer in Azure | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Azure provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: azure-groupsync
          spec:
            providers:
            - name: azure
              azure:
                credentialsSecret:
                  name: azure-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to Azure

          Authentication to Azure can be performed using Application Registration with access to query group information in Azure Active Directory.

          The App Registration must be granted access to the following Microsoft Graph API's:

          * Group.Read.All
          * GroupMember.Read.All
          * User.Read.All

          A secret must be created in the same namespace that contains the `GroupSync` resource:

          The following keys must be defined in the secret

          * `AZURE_TENANT_ID` - Tenant ID
          * `AZURE_CLIENT_ID` - Client ID
          * `AZURE_CLIENT_SECRET` - Client Secret

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic azure-group-sync --from-literal=AZURE_SUBSCRIPTION_ID=<AZURE_SUBSCRIPTION_ID> --from-literal=AZURE_TENANT_ID=<AZURE_TENANT_ID> --from-literal=AZURE_CLIENT_ID=<AZURE_CLIENT_ID> --from-literal=AZURE_CLIENT_SECRET=<AZURE_CLIENT_SECRET>
          ```

          ### GitHub

          Teams stored within a GitHub organization can be synchronized into OpenShift. The following table describes the set of configuration options for the GitHub provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `caSecret` | Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `organization` | Organization to synchronize against | | Yes |
          | `teams` | List of teams to filter against | | No |
          | `url` | Base URL for the GitHub or GitHub Enterprise host (Must contain a trailing slash) | | No |
          | `prune` | Prune Whether to prune groups that are no longer in GitHub | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Github provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: github-groupsync
          spec:
            providers:
            - name: github
              github:
                organization: ocp
                credentialsSecret:
                  name: github-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to GitHub

          Authentication to GitLab can be performed using a [Token](https://docs.gitlab.com/ee/security/token_overview.html) or a Username and Password (Note: 2FA not supported). A secret must be created in the same namespace that contains the `GroupSync` resource:

          When using an OAuth token, the following token types are supported:

          * Personal Access Token
          * OAuth Token
          * Job Token

          the following key is required:

          * `token` - OAuth token

          Optionally, the `tokenType` key can be specified to indicate the type of token being provided from the following values:

          * OAuth - `oauth`
          * Personal Access Token - `personal`
          * Job Token - `job`

          If no `tokenType` is provided, `oauth` is used by default

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token>
          ```

          To specify a token type, such as a Personal Access Token, the following command can be executed:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token> --from-literal=tokenType=personal
          ```

          The following keys are required for username and password:

          * `username` - Username for authenticating with GitHub
          * `password` - Password for authenticating with GitHub

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic github-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### GitLab

          Groups stored within a GitLab can be synchronized into OpenShift. The following table describes the set of configuration options for the GitLab provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `groups` | List of groups to filter against | | No |
          | `prune` | Prune Whether to prune groups that are no longer in GitLab | `false` | No |
          | `scope` | Scope for group synchronization. Options are `one` for one level or `sub` to include subgroups | `sub` | No |
          | `url` | Base URL for the GitLab instance | `https://gitlab.com` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Github provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: gitlab-groupsync
          spec:
            providers:
            - name: gitlab
              gitlab:
                credentialsSecret:
                  name: gitlab-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to GitLab

          Authentication to GitLab can be performed using an OAuth Personal Access Token or a Username and Password (Note: 2FA not supported). A secret must be created in the same namespace that contains the `GroupSync` resource:

          When using an OAuth token, the following key is required:

          * `token` - OAuth token

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token>
          ```

          The following keys are required for username and password:

          * `username` - Username for authenticating with GitLab
          * `password` - Password for authenticating with GitLab

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### LDAP

          Groups stored within an [LDAP](https://en.wikipedia.org/wiki/Lightweight_Directory_Access_Protocol) server can be synchronized into OpenShift. The LDAP provider implements the included features of the [Syncing LDAP groups](https://docs.openshift.com/container-platform/latest/authentication/ldap-syncing.html) feature and makes use of the libraries from the [OpenShift Command Line](https://docs.openshift.com/container-platform/latest/cli_reference/openshift_cli/getting-started-cli.html) tool to streamline the migration to this operator based implementation.

          The configurations of the three primary schemas (`rfc2307`, `activeDirectory` and `augmentedActiveDirectory`) can be directly migrated as is without any modification.

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | No |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `groupUIDNameMapping` | User defined name mapping | | No |
          | `rfc2307` | Configuration using the [rfc2307](https://docs.openshift.com/container-platform/latest/authentication/ldap-syncing.html#ldap-syncing-rfc2307_ldap-syncing-groups) schema | | No |
          | `activeDirectory` | Configuration using the [activeDirectory](https://docs.openshift.com/container-platform/4.5/authentication/ldap-syncing.html#ldap-syncing-activedir_ldap-syncing-groups) schema | | No |
          | `augmentedActiveDirectory` | Configuration using the [activeDirectory](https://docs.openshift.com/container-platform/4.5/authentication/ldap-syncing.html#ldap-syncing-augmented-activedir_ldap-syncing-groups) schema | | No |
          | `url` | Connection URL for the LDAP server | `ldap://ldapserver:389` | No |
          | `whitelist` | Explicit list of groups to synchronize |  | No |
          | `blacklist` | Explicit list of groups to not synchronize |  | No |
          | `prune` | Prune Whether to prune groups that are no longer in LDAP | `false` | No |

          The following is an example using the `rfc2307` schema:

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
                credentialsSecret:
                  name: ldap-group-sync
                  namespace: group-sync-operator
                insecure: true
                rfc2307:
                  groupMembershipAttributes:
                  - member
                  groupNameAttributes:
                  - cn
                  groupUIDAttribute: dn
                  groupsQuery:
                    baseDN: ou=Groups,dc=example,dc=com
                    derefAliases: never
                    filter: (objectClass=groupofnames)
                    scope: sub
                  tolerateMemberNotFoundErrors: true
                  tolerateMemberOutOfScopeErrors: true
                  userNameAttributes:
                  - cn
                  userUIDAttribute: dn
                  usersQuery:
                    baseDN: ou=Users,dc=example,dc=com
                    derefAliases: never
                    scope: sub
                url: ldap://ldapserver:389
              name: ldap
          ```

          The examples provided in the OpenShift documented referenced previously can be used to construct the schemas for the other LDAP synchronization types.

          #### Authenticating to LDAP

          If authentication is required in order to communicate with the LDAP server, a secret should be created in the same namespace that contains the `GroupSync` resource. The following keys can be defined:

          * `username` - Username (Bind DN) for authenticating with the LDAP server
          * `password` - Password for authenticating with the LDAP server

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic ldap-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          #### Whitelists and Blacklists

          Groups can be explicitly whitelisted or blacklisted in order to control the groups that are eligible to be synchronized into OpenShift. When running LDAP group synchronization using the command line, this configuration is referenced via separate files, but these are instead specified in the `blacklist` and `whitelist` properties as shown below:

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
          ...
                whitelist:
                - cn=Online Corporate Banking,ou=Groups,dc=example,dc=com
          ...
              name: ldap
          ```

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
          ...
                blacklist:
                - cn=Finance,ou=Groups,dc=example,dc=com
          ...
              name: ldap
          ```

          ### Keycloak

          Groups stored within Keycloak can be synchronized into OpenShift. The following table describes the set of configuration options for the Keycloak provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `groups` | List of groups to filter against | | No |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `loginRealm` | Realm to authenticate against | `master` | No |
          | `realm` | Realm to synchronize | | Yes |
          | `scope` | Scope for group synchronization. Options are `one` for one level or `sub` to include subgroups | `sub` | No |
          | `url` | Base URL for the Keycloak server. Older versions (<17.0.0) including Red Hat SSO should include the context path `/auth` appended to the hostname  | | Yes |
          | `prune` | Prune Whether to prune groups that are no longer in Keycloak | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Keycloak provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                url: https://keycloak-keycloak-operator.apps.openshift.com/auth
          ```

          #### Authenticating to Keycloak

          A user with rights to query for Keycloak groups must be available. The following permissions must be associated to the user:

          * Password must be set (Temporary option unselected) on the _Credentials_ tab
          * On the _Role Mappings_ tab, select _master-realm_ or _realm-management_ next to the _Client Roles_ dropdown and then select **Query Groups** and **Query Users**.

          A secret must be created in the same namespace that contains the `GroupSync` resource. It must contain the following keys for the user previously created:

          * `username` - Username for authenticating with Keycloak
          * `password` - Password for authenticating with Keycloak

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic keycloak-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### Okta

          [Okta Groups](https://help.okta.com/en/prod/Content/Topics/users-groups-profiles/usgp-main.htm) assigned to [Okta Applications](https://help.okta.com/en/prod/Content/Topics/Apps/Apps_Apps.htm) can be synchronized into OpenShift. The developer docs for the Okta API that the Okta Syncer uses can be found [here](https://developer.okta.com/docs/reference/api/apps/#list-groups-assigned-to-application).
          The following table describes the set of configuration options for the Okta provider:

            | Name | Description | Defaults | Required |
            | ----- | ---------- | -------- | ----- |
            | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
            | `groups` | List of groups to filter against | | No |
            | `url` | Okta organization URL (Must contain a trailing slash) | | Yes |
            | `appId` | Application ID of App Groups are assigned to | | Yes |
            | `extractLoginUsername` | Bool to determine if you should extract username from okta login | | No |
            | `profileKey` | Attribute field on Okta User Profile you would like to use as identity | `login` | No |
            | `groupLimit` | Integer to set the maximum number of groups to retrieve from OKTA per request. | `1000` | No |
            | `prune` | Prune Whether to prune groups that are no longer in OKTA | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with an Okta provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: okta-sync
          spec:
            providers:
              - name: okta
                okta:
                  credentialsSecret:
                    name: okta-api-token
                    namespace: group-sync-operator
                  url: "https://example.okta.com/"
                  appId: okta-sync-app-id
          ```

          #### Authenticating to Okta

          A secret must be created in the same namespace as the group-sync-operator pod. It must contain the following key:

            * `okta-api-token` - Okta API Token for interacting with Okta

          The secret can be created by executing the following command:

            ```shell
            oc create secret generic okta-api-token --from-literal=okta-api-token=<OKTA_API_TOKEN> -n group-sync-operator
            ```

          ### IBM Security Verify

          Groups defined in [IBM Security Verify](https://help.okta.com/en/prod/Content/Topics/users-groups-profiles/usgp-main.htm) (ISV) can be synchronized into OpenShift. Currently only the `userName` field from ISV will be synchronized. The developer docs for the ISV API can be found [here](https://docs.verify.ibm.com/verify/page/api-documentation).
          The following table describes the set of configuration options for the provider:

            | Name | Description | Defaults | Required |
            | ----- | ---------- | -------- | ----- |
            | `credentialsSecret` | Reference to a secret containing authentication details (see below) | `''`  | Yes |
            | `groups` | List of groups to synchronize (see below) | `nil`  | Yes |
            | `tenantUrl` | The ISV tenant URL, for example `https://my-isv.verify.ibm.com`) | `''`  | Yes |

          The following is an example of a minimal configuration that can be applied to integrate with an Okta provider:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ibmsecurityverify-sync
          spec:
              providers:
              - name: ibmsecurityverify
                ibmsecurityverify:
                  credentialsSecret:
                    name: isv-group-sync
                    namespace: group-sync-operator
                  tenantUrl: https://my-isv.verify.ibm.com
                  groups:
                    - name: 'application owners'
                      id: 645001V3V9
                    - name: developer
                      id: 645001V3VA
          ```

          #### Group Objects
          Each group object in the `groups` array must contain an `id` field. The group ID can be retrieved by pulling the group information from the ISV API. Optionally, the object may also contain a `name` which corresponds to the group's display name. When defined, the operator will confirm that the name defined in the YAML matches that received from the API when synchronization occurs; as the group IDs are not human-friendly, using the name can confirm the correct groups are configured. If the names do not match an error will be logged.

          #### Group Names
          The name of each groups created in OpenShift will match the group name in ISV. Any whitespace in the ISV group name will be replaced with a hyphen.

          #### Authenticating to IBM Security Verify

          A secret must be created in the same namespace as the group-sync-operator pod. It must contain the following keys:

          * `clientId` - The API client ID.
          * `clientSecret`- The API client secret.

          See the IBM Security Verify [API documentation](https://docs.verify.ibm.com/verify/docs/api-access) for setting up authentication.

          ### Support for Additional Metadata (Beta)

          Additional metadata based on Keycloak group are also added to the OpenShift groups as Annotations including:

          * Parent/child relationship between groups and their subgroups
          * Group attributes

          ## CA Certificates

          Several providers allow for certificates to be provided in either a _ConfigMap_ or _Secret_ to communicate securely to the target host through the use of a property called `ca`.

          The certificate can be added to a Secret called _keycloak-certs_ using the key `ca.crt` representing the certificate using the following command.

          ```
          oc create secret generic keycloak-certs --from-file=ca.crt=<file>
          ```

          An example of how the CA certificate can be added to the Keycloak provider is shown below:


          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                caSecret:
                  kind: Secret
                  name: keycloak-certs
                  namespace: group-sync-operator
                  key: ca.crt
                url: https://keycloak-keycloak-operator.apps.openshift.com
          ```

          Alteratively, a _ConfigMap_ can be used instead instead of a _Secret_. This is useful when using the [Certificate injection using Operators](https://docs.openshift.com/container-platform/latest/networking/configuring-a-custom-pki.html#certificate-injection-using-operators_configuring-a-custom-pki) feature.

          The following command can be used to create a _ConfigMap_ containing the certificate:

          ```
          oc create configmap keycloak-certs --from-file=ca.crt=<file>
          ````

          An example of how the CA certificate can be added to the Keycloak provider is shown below:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                caSecret:
                  kind: ConfigMap
                  name: keycloak-certs
                  namespace: group-sync-operator
                  key: ca.crt
                url: https://keycloak-keycloak-operator.apps.openshift.com
          ```

          ## Scheduled Execution

          A cron style expression can be specified for which a synchronization event will occur. The following specifies that a synchronization should occur nightly at 3AM

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            schedule: "0 3 * * *"
            providers:
            - ...
          ```

          If a schedule is not provided, synchronization will occur only when the object is reconciled by the platform.

          ## Metrics

          Prometheus compatible metrics are exposed by the Operator and can be integrated into OpenShift's default cluster monitoring. To enable OpenShift cluster monitoring, label the namespace the operator is deployed in with the label `openshift.io/cluster-monitoring="true"`.

          ```
          oc label namespace <namespace> openshift.io/cluster-monitoring="true"
          ```
        displayName: Group Sync Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - group
        - user
        - security
        - authorization
        links:
        - name: Group Sync Operator
          url: https://github.com/redhat-cop/group-sync-operator
        maintainers:
        - email: ablock@redhat.com
          name: Andrew Block
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:b8a174be91f3df75dff9656d4c72de28af51a731008e9e96db5eef3c7af58aac
        - quay.io/redhat-cop/group-sync-operator@sha256:43ef79efc2a77ee5edd4e3123f9dab156a30d3e00fd1fba9f89600b96f0f68d6
        version: 0.0.36
      entries:
      - name: group-sync-operator.v0.0.36
        version: 0.0.36
      - name: group-sync-operator.v0.0.35
        version: 0.0.35
      - name: group-sync-operator.v0.0.34
        version: 0.0.34
      - name: group-sync-operator.v0.0.33
        version: 0.0.33
      - name: group-sync-operator.v0.0.32
        version: 0.0.32
      - name: group-sync-operator.v0.0.31
        version: 0.0.31
      - name: group-sync-operator.v0.0.30
        version: 0.0.30
      - name: group-sync-operator.v0.0.29
        version: 0.0.29
      - name: group-sync-operator.v0.0.28
        version: 0.0.28
      - name: group-sync-operator.v0.0.27
        version: 0.0.27
      - name: group-sync-operator.v0.0.26
        version: 0.0.26
      - name: group-sync-operator.v0.0.25
        version: 0.0.25
      - name: group-sync-operator.v0.0.24
        version: 0.0.24
      - name: group-sync-operator.v0.0.23
        version: 0.0.23
      - name: group-sync-operator.v0.0.22
        version: 0.0.22
      - name: group-sync-operator.v0.0.21
        version: 0.0.21
      - name: group-sync-operator.v0.0.20
        version: 0.0.20
      - name: group-sync-operator.v0.0.19
        version: 0.0.19
      - name: group-sync-operator.v0.0.18
        version: 0.0.18
      - name: group-sync-operator.v0.0.17
        version: 0.0.17
      - name: group-sync-operator.v0.0.16
        version: 0.0.16
      - name: group-sync-operator.v0.0.15
        version: 0.0.15
      - name: group-sync-operator.v0.0.14
        version: 0.0.14
      - name: group-sync-operator.v0.0.13
        version: 0.0.13
      name: alpha
    defaultChannel: alpha
    packageName: group-sync-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubecost
      provider-url: https://www.kubecost.com
    name: kubecost-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubecost-operator.v2.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kubecost.com/v1alpha1",
                "kind": "CostAnalyzer",
                "metadata": {
                  "name": "costanalyzer-sample"
                },
                "spec": {
                  "clusterController": {
                    "image": {
                      "repository": "registry.connect.redhat.com/kubecost/kubecost-cluster-controller@sha256",
                      "tag": "1a6f11dcd7341ab3f24eba23ec06c7b75646a534a7ed960153601806425469d1"
                    }
                  },
                  "forecasting": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-modeling@sha256:8f59f8fd8696964c1b5322ca79fae78befe97323fa56c2b509bdcc67ee4c0edd"
                  },
                  "global": {
                    "platforms": {
                      "openshift": {
                        "createMonitoringClusterRoleBinding": true,
                        "createMonitoringResourceReaderRoleBinding": true,
                        "enabled": true,
                        "scc": {
                          "clusterController": true,
                          "networkCosts": true,
                          "nodeExporter": true
                        }
                      }
                    },
                    "prometheus": {
                      "enabled": false,
                      "fqdn": "https://prometheus-k8s.openshift-monitoring.svc.cluster.local:9091",
                      "kubeRBACProxy": true
                    }
                  },
                  "grafana": {
                    "image": {
                      "repository": "registry.redhat.io/rhel9/grafana@sha256",
                      "tag": "9f71de028c18034e07a8d15dac13ef68afb91cb899b9e8e7d39703547dd0eac1"
                    },
                    "sidecar": {
                      "image": {
                        "repository": "quay.io/kiwigrid/k8s-sidecar@sha256",
                        "tag": "d3749777d5bedf62549b452a7e2ccb205da53588dc7f04927efc75d0da5b6d23"
                      }
                    }
                  },
                  "kubecostAggregator": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef"
                  },
                  "kubecostFrontend": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-frontend@sha256:f9b2196dc6ac017b8074f8fe40f1a58e7a2630282c5b7a019ca9f3dbdb670dc8"
                  },
                  "kubecostModel": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef"
                  },
                  "networkCosts": {
                    "image": {
                      "repository": "registry.connect.redhat.com/kubecost/kubecost-network-costs@sha256",
                      "tag": "c444d475be4f1aa36680bd84358e1e07303dcfe5d13a95544ad9e59f80d4a5f0"
                    }
                  },
                  "prometheusRule": {
                    "enabled": true
                  },
                  "serviceMonitor": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          createdAt: "2025-08-08T19:34:34Z"
          description: Deploys Kubecost 2.8.2 via the Helm chart.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Contact Kubecost for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Kubecost.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The CostAnalyzer API describes a deployment of Kubecost and
              its dependencies.
            displayName: CostAnalyzer
            kind: CostAnalyzer
            name: costanalyzers.charts.kubecost.com
            version: v1alpha1
        description: |2-

          IMPORTANT - Before installation, you must run the following command to allow monitoring of the Kubecost namespace by the local OpenShift Promtheus. If you install in a namespace other than "Kubecost", adjust accordingly.

          `oc label namespace kubecost openshift.io/cluster-monitoring=true`

          Kubecost provides real-time cost visibility and insights for teams using Kubernetes, helping you continuously reduce your cloud costs.

          This OpenShift Operator deploys Kubecost from the official Helm chart.

          Kubecost allows you to:

          * See allocated spend across all native Kubernetes concepts and provide your teams with transparent, accurate cost data reconciled with your actual cloud bill.
          * Join in-cluster costs like CPU and memory with out-of-cluster spend from other cloud services.
          * Get complete cost visibility across OpenShift, On-Prem, AWS, GCP, Azure, Oracle, AliCloud + more.
          * Gain insights you can use to save 30-50%+ on your infrastructure spend, without exposing your private information — your data never gets shared externally, even with us.
          * Have real-time alerts and recurring reports, empowering teams to take control of their Kubernetes-enabled infrastructure and stay within budgeted limits.

          Visit the Kubecost website for more details including features, pricing, documentation, and more at www.kubecost.com.
        displayName: Kubecost
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kubernetes
        - cost
        - kubernetes costs
        - kubernetes optimization
        - k8s
        - cost monitoring
        - cloud costs
        - k8s cost
        links:
        - name: Kubecost
          url: https://www.kubecost.com
        maintainers:
        - email: support@kubecost.com
          name: Kubecost Support
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Kubecost
          url: https://www.kubecost.com
        relatedImages:
        - registry.connect.redhat.com/kubecost/kubecost-network-costs@sha256:c444d475be4f1aa36680bd84358e1e07303dcfe5d13a95544ad9e59f80d4a5f0
        - registry.redhat.io/rhel9/grafana@sha256:9f71de028c18034e07a8d15dac13ef68afb91cb899b9e8e7d39703547dd0eac1
        - quay.io/kiwigrid/k8s-sidecar@sha256:d3749777d5bedf62549b452a7e2ccb205da53588dc7f04927efc75d0da5b6d23
        - registry.connect.redhat.com/kubecost/kubecost-operator-manager@sha256:582223a66a77f35d1f7ca33460032cfca1581a8ff30f75ead289d8cd4cf62bb9
        - registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef
        - registry.connect.redhat.com/kubecost/kubecost-frontend@sha256:f9b2196dc6ac017b8074f8fe40f1a58e7a2630282c5b7a019ca9f3dbdb670dc8
        - registry.connect.redhat.com/kubecost/kubecost-modeling@sha256:8f59f8fd8696964c1b5322ca79fae78befe97323fa56c2b509bdcc67ee4c0edd
        - registry.connect.redhat.com/kubecost/kubecost-cluster-controller@sha256:1a6f11dcd7341ab3f24eba23ec06c7b75646a534a7ed960153601806425469d1
        version: 2.8.2
      entries:
      - name: kubecost-operator.v2.8.2
        version: 2.8.2
      - name: kubecost-operator.v2.8.1
        version: 2.8.1
      - name: kubecost-operator.v2.8.0
        version: 2.8.0
      - name: kubecost-operator.v2.7.2
        version: 2.7.2
      - name: kubecost-operator.v2.7.1
        version: 2.7.1
      - name: kubecost-operator.v2.5.4
        version: 2.5.4
      name: alpha
    defaultChannel: alpha
    packageName: kubecost-operator
    provider:
      name: Kubecost
      url: https://www.kubecost.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-access-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-access-operator.v25.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyAccess",
                "metadata": {
                  "name": "ivia-sample"
                },
                "spec": {
                  "image": "icr.io/ivia/ivia-wrp@sha256:c2dd5ec2b362136896e47ccbd0f3ee033af945fc2794ddaf075da603e4d4d7b2",
                  "instance": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isva/verify-access-operator@sha256:ca1d7be14b4dd313d4675f18541b2cc670ea0a0adf3ec87e347e5e20a86bc39c
          createdAt: "2025-05-01T01:27:17Z"
          description: The IBM Verify Identity Access Operator manages the lifecycle
            of IBM Verify Identity Access worker containers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM-Security/verify-access-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyAccess is the Schema for the ibmsecurityverifyaccesses
              API.
            displayName: IBMSecurity Verify Access
            kind: IBMSecurityVerifyAccess
            name: ibmsecurityverifyaccesses.ibm.com
            version: v1
        description: |+
          In a world of highly fragmented access management environments, [IBM Verify Identity Access](https://www.ibm.com/au-en/products/verify-access) helps you simplify your users' access while more securely adopting web, mobile and cloud technologies. This solution helps you strike a balance between usability and security through the use of risk-based access, single sign-on, integrated access management control, identity federation and its mobile multi-factor authentication capability, IBM Verify. Take back control of your access management with IBM Verify Identity Access.

          The IBM Verify Identity Access operator provides lifecycle management of the lightweight containers which are used to protect an environment, namely:
          * [Web Reverse Proxy](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-web-reverse-proxy)
          * [Runtime](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-runtime)
          * [Distributed Session Cache](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-distributed-session-cache)

          The operator will manage the deployment of these lightweight IBM Verify Identity Access worker containers, and also control the rolling restart of these containers when a configuration snapshot is updated.

          See the project [Readme](https://github.com/IBM-Security/verify-access-operator/blob/master/README.md) for further information and details.

        displayName: IBM Verify Identity Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Identity Access Product Information
          url: https://www.ibm.com/au-en/products/verify-access
        - name: Verify Identity Access Documentation
          url: https://www.ibm.com/docs/en/sva
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Identity Access Development Team
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - icr.io/isva/verify-access-operator@sha256:ca1d7be14b4dd313d4675f18541b2cc670ea0a0adf3ec87e347e5e20a86bc39c
        - icr.io/ivia/ivia-wrp@sha256:c2dd5ec2b362136896e47ccbd0f3ee033af945fc2794ddaf075da603e4d4d7b2
        version: 25.4.0
      entries:
      - name: ibm-security-verify-access-operator.v25.4.0
        version: 25.4.0
      - name: ibm-security-verify-access-operator.v24.12.0
        version: 24.12.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-access-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Redis
      provider-url: ""
    name: redis-enterprise-operator-cert
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-enterprise-operator.v6.2.10-45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
        - registry.connect.redhat.com/redislabs/services-manager@sha256:8b46f2aacfb3e1b243c0b64975accc55fd80b4de9c9ac4769c3c67a5676addc5
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:ffdf10a1b08115fd4de06f1cd751d36644a9bc9a8c11034cd432e9b66a5d2f73
        version: 6.2.10-45
      entries:
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: 6.2.10
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.22.0
    - currentCSV: redis-enterprise-operator.v7.22.0-11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:088e4b57c6e1ca1680b359078e5eb1bba67343cdbd9a539af26a0ea23a830593
        - registry.connect.redhat.com/redislabs/services-manager@sha256:3b96cdc65c3e42fd22eb165950be96b907725903ac74fe4f82cfa0073e28e95a
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:d56ba97e95e799210e76615095409143a5dddd5c5a1e1be1d551d7e3f9e4d3ef
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:0cfb07f774db131b5b74e5aa91291579c916c4d4cd1a693d7aaed554f0177a68
        version: 7.22.0-11.3
      entries:
      - name: redis-enterprise-operator.v7.22.0-11.3
        version: 7.22.0-11.3
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.22.0-11.3
    - currentCSV: redis-enterprise-operator.v7.4.6-7.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:473b5661049a6c8830ba113504975726154c7b172376ba3b63d891674d59c78f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5f6f32a63f4dd8be85b9d469f0fedd26d6aa76e32b7735bfa7ab7203cb272cdb
        - registry.connect.redhat.com/redislabs/services-manager@sha256:ba20ab644a473bc742af671171d5313537292d88942b16336d03a6386c061d8a
        version: 7.4.6-7.2
      entries:
      - name: redis-enterprise-operator.v7.4.6-7.2
        version: 7.4.6-7.2
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      name: 7.4.6
    - currentCSV: redis-enterprise-operator.v7.8.2-6.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:36c7a36dabb23a4c003e112085cda2873296bfb22e5d11b470d7c1aba55b7f0f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:c6e1f2109a0dac76b51c25c1cbf6030b4894b21b98147e2ad1db2dd8e7693576
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:acf817c90b5f03bc2796fbc305cf43ce5c0644dcbe1b685752090b3f600a0f7b
        version: 7.8.2-6.8
      entries:
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.2
    - currentCSV: redis-enterprise-operator.v7.8.4-9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:803be04534e59cdc7b8898b3bb336676aad1cb77cf266051e226be82679fcf6a
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:4ba0c414a74268bed9c6757e447b340db7c912afa864cfb096ddf88f11feeaf9
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:66ec124da16cc9d0abd66687e088be0b358e20cfd48f5774d34bf5874e6a4e65
        version: 7.8.4-9.0
      entries:
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.4
    - currentCSV: redis-enterprise-operator.v7.8.6-5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:1fff01daece34aeb7befcb0760d5112c8c7d261f2b86388ece416ac94076a573
        - registry.connect.redhat.com/redislabs/services-manager@sha256:88d483f5e37b5a901d1ae3d965304edc000c172d2e5e307297d5b7bcd77ae73b
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5125e83d3c93b224d1983c599994e43899a0ba0c6f2fc4eed31476f755b5fc0b
        version: 7.8.6-5.1
      entries:
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.6
    - currentCSV: redis-enterprise-operator.v6.2.10-41
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/outbound",
                      "versionTag": "6.2.10-41"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager",
                      "versionTag": "6.2.10-41"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise",
                    "versionTag": "6.2.14-9.rhel8-openshift"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/outbound@sha256:3c251754ae8650157465a56ef3a7b41d5a0ad4e9db09ef042e0014076f25e8a2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/outbound@sha256:3c251754ae8650157465a56ef3a7b41d5a0ad4e9db09ef042e0014076f25e8a2
        version: 6.2.10-41
      entries:
      - name: redis-enterprise-operator.v6.2.10-41
        version: 6.2.10-41
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: preview
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: production
    defaultChannel: production
    packageName: redis-enterprise-operator-cert
    provider:
      name: Redis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-iam-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-iam-controller.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Role",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Group",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "OpenIDConnectProvider",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Policy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/iam-controller:1.5.2
          createdAt: "2025-10-03T22:22:57Z"
          description: AWS IAM controller is a service controller for managing IAM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Group represents the state of an AWS iam Group resource.
            displayName: Group
            kind: Group
            name: groups.iam.services.k8s.aws
            version: v1alpha1
          - description: InstanceProfile represents the state of an AWS iam InstanceProfile
              resource.
            displayName: InstanceProfile
            kind: InstanceProfile
            name: instanceprofiles.iam.services.k8s.aws
            version: v1alpha1
          - description: OpenIDConnectProvider represents the state of an AWS iam
              OpenIDConnectProvider resource.
            displayName: OpenIDConnectProvider
            kind: OpenIDConnectProvider
            name: openidconnectproviders.iam.services.k8s.aws
            version: v1alpha1
          - description: Policy represents the state of an AWS iam Policy resource.
            displayName: Policy
            kind: Policy
            name: policies.iam.services.k8s.aws
            version: v1alpha1
          - description: Role represents the state of an AWS iam Role resource.
            displayName: Role
            kind: Role
            name: roles.iam.services.k8s.aws
            version: v1alpha1
          - description: ServiceLinkedRole represents the state of an AWS iam ServiceLinkedRole
              resource.
            displayName: ServiceLinkedRole
            kind: ServiceLinkedRole
            name: servicelinkedroles.iam.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS iam User resource.
            displayName: User
            kind: User
            name: users.iam.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Identity and Access Management (IAM) resources in AWS from within your Kubernetes cluster.

          **About Amazon IAM**

          AWS Identity and Access Management (IAM) is a web service that helps you securely control access to AWS resources. You use IAM to control who is authenticated (signed in) and authorized (has permissions) to use resources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon IAM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iam
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon IAM Developer Resources
          url: https://aws.amazon.com/iam/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: iam maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/iam-controller:1.5.2
        version: 1.5.2
      entries:
      - name: ack-iam-controller.v1.5.2
        version: 1.5.2
      - name: ack-iam-controller.v1.5.1
        version: 1.5.1
      - name: ack-iam-controller.v1.5.0
        version: 1.5.0
      - name: ack-iam-controller.v1.4.4
        version: 1.4.4
      - name: ack-iam-controller.v1.4.3
        version: 1.4.3
      - name: ack-iam-controller.v1.4.2
        version: 1.4.2
      - name: ack-iam-controller.v1.4.1
        version: 1.4.1
      - name: ack-iam-controller.v1.4.0
        version: 1.4.0
      - name: ack-iam-controller.v1.3.22
        version: 1.3.22
      - name: ack-iam-controller.v1.3.21
        version: 1.3.21
      - name: ack-iam-controller.v1.3.20
        version: 1.3.20
      - name: ack-iam-controller.v1.3.19
        version: 1.3.19
      - name: ack-iam-controller.v1.3.18
        version: 1.3.18
      - name: ack-iam-controller.v1.3.17
        version: 1.3.17
      - name: ack-iam-controller.v1.3.15
        version: 1.3.15
      - name: ack-iam-controller.v1.3.14
        version: 1.3.14
      - name: ack-iam-controller.v1.3.13
        version: 1.3.13
      - name: ack-iam-controller.v1.3.12
        version: 1.3.12
      - name: ack-iam-controller.v1.3.11
        version: 1.3.11
      - name: ack-iam-controller.v1.3.10
        version: 1.3.10
      - name: ack-iam-controller.v1.3.9
        version: 1.3.9
      - name: ack-iam-controller.v1.3.8
        version: 1.3.8
      - name: ack-iam-controller.v1.3.6
        version: 1.3.6
      - name: ack-iam-controller.v1.3.2
        version: 1.3.2
      - name: ack-iam-controller.v1.3.1
        version: 1.3.1
      - name: ack-iam-controller.v1.3.0
        version: 1.3.0
      - name: ack-iam-controller.v1.2.6
        version: 1.2.6
      - name: ack-iam-controller.v1.2.5
        version: 1.2.5
      - name: ack-iam-controller.v1.2.3
        version: 1.2.3
      - name: ack-iam-controller.v1.2.2
        version: 1.2.2
      - name: ack-iam-controller.v1.2.1
        version: 1.2.1
      - name: ack-iam-controller.v1.2.0
        version: 1.2.0
      - name: ack-iam-controller.v1.1.3
        version: 1.1.3
      - name: ack-iam-controller.v1.1.2
        version: 1.1.2
      - name: ack-iam-controller.v1.1.1
        version: 1.1.1
      - name: ack-iam-controller.v1.1.0
        version: 1.1.0
      - name: ack-iam-controller.v1.0.0
        version: 1.0.0
      - name: ack-iam-controller.v0.1.2
        version: 0.1.2
      - name: ack-iam-controller.v0.1.1
        version: 0.1.1
      - name: ack-iam-controller.v0.1.0
        version: 0.1.0
      - name: ack-iam-controller.v0.0.22
        version: 0.0.22
      - name: ack-iam-controller.v0.0.21
        version: 0.0.21
      - name: ack-iam-controller.v0.0.20
        version: 0.0.20
      - name: ack-iam-controller.v0.0.19
        version: 0.0.19
      - name: ack-iam-controller.v0.0.18
        version: 0.0.18
      - name: ack-iam-controller.v0.0.17
        version: 0.0.17
      - name: ack-iam-controller.v0.0.16
        version: 0.0.16
      - name: ack-iam-controller.v0.0.15
        version: 0.0.15
      - name: ack-iam-controller.v0.0.14
        version: 0.0.14
      - name: ack-iam-controller.v0.0.13
        version: 0.0.13
      - name: ack-iam-controller.v0.0.12
        version: 0.0.12
      - name: ack-iam-controller.v0.0.11
        version: 0.0.11
      - name: ack-iam-controller.v0.0.10
        version: 0.0.10
      - name: ack-iam-controller.v0.0.9
        version: 0.0.9
      - name: ack-iam-controller.v0.0.8
        version: 0.0.8
      - name: ack-iam-controller.v0.0.7
        version: 0.0.7
      - name: ack-iam-controller.v0.0.6
        version: 0.0.6
      - name: ack-iam-controller.v0.0.5
        version: 0.0.5
      - name: ack-iam-controller.v0.0.4
        version: 0.0.4
      - name: ack-iam-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-iam-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ODH
      provider-url: ""
    name: opendatahub-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opendatahub-operator.v2.35.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Managed",
                      "registriesNamespace": "odh-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "opendatahub",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "opendatahub"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v2.35.0
          createdAt: "2025-10-01T13:12:27Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",
            "workbenches.components.platform.opendatahub.io", "monitorings.services.platform.opendatahub.io",
            "servicemeshes.services.platform.opendatahub.io", "modelcontrollers.components.platform.opendatahub.io",
            "feastoperators.components.platform.opendatahub.io", "llamastackoperators.components.platform.opendatahub.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/opendatahub-io/opendatahub-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: ServiceMesh is the Schema for the servicemesh API
            displayName: Service Mesh
            kind: ServiceMesh
            name: servicemeshes.services.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack - Unified open-source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odh
        - notebooks
        - serving
        - training
        - pipelines
        - modelmesh
        - workbenches
        - dashboard
        - kserve
        - distributed workloads
        - trustyai
        - modelregistry
        - feast
        - featurestore
        - llamastack
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ODH
        relatedImages:
        - quay.io/opendatahub/opendatahub-operator:v2.35.0
        version: 2.35.0
      entries:
      - name: opendatahub-operator.v2.35.0
        version: 2.35.0
      - name: opendatahub-operator.v2.34.0
        version: 2.34.0
      - name: opendatahub-operator.v2.33.0
        version: 2.33.0
      - name: opendatahub-operator.v2.32.0
        version: 2.32.0
      - name: opendatahub-operator.v2.31.0
        version: 2.31.0
      - name: opendatahub-operator.v2.30.0
        version: 2.30.0
      - name: opendatahub-operator.v2.29.0
        version: 2.29.0
      - name: opendatahub-operator.v2.28.0
        version: 2.28.0
      - name: opendatahub-operator.v2.27.0
        version: 2.27.0
      - name: opendatahub-operator.v2.26.0
        version: 2.26.0
      - name: opendatahub-operator.v2.25.0
        version: 2.25.0
      - name: opendatahub-operator.v2.24.0
        version: 2.24.0
      - name: opendatahub-operator.v2.23.0
        version: 2.23.0
      - name: opendatahub-operator.v2.22.0
        version: 2.22.0
      - name: opendatahub-operator.v2.21.0
        version: 2.21.0
      - name: opendatahub-operator.v2.20.0
        version: 2.20.0
      - name: opendatahub-operator.v2.19.0
        version: 2.19.0
      - name: opendatahub-operator.v2.18.2
        version: 2.18.2
      - name: opendatahub-operator.v2.18.1
        version: 2.18.1
      - name: opendatahub-operator.v2.18.0
        version: 2.18.0
      - name: opendatahub-operator.v2.17.0
        version: 2.17.0
      - name: opendatahub-operator.v2.16.0
        version: 2.16.0
      - name: opendatahub-operator.v2.15.0
        version: 2.15.0
      - name: opendatahub-operator.v2.14.0
        version: 2.14.0
      - name: opendatahub-operator.v2.13.0
        version: 2.13.0
      - name: opendatahub-operator.v2.12.1
        version: 2.12.1
      - name: opendatahub-operator.v2.12.0
        version: 2.12.0
      - name: opendatahub-operator.v2.11.1
        version: 2.11.1
      - name: opendatahub-operator.v2.11.0
        version: 2.11.0
      - name: opendatahub-operator.v2.10.0
        version: 2.10.0
      name: fast
    - currentCSV: opendatahub-operator.v2.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "opendatahub",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "opendatahub"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v2.8.0
          createdAt: "2024-2-20T00:00:00Z"
          operatorframework.io/initialization-resource: |-
            {
            "apiVersion": "dscinitialization.opendatahub.io/v1",
            "kind": "DSCInitialization",
            "metadata": {
            "labels": {
             "app.kubernetes.io/created-by": "opendatahub-operator",
             "app.kubernetes.io/instance": "default",
             "app.kubernetes.io/managed-by": "kustomize",
             "app.kubernetes.io/name": "dscinitialization",
             "app.kubernetes.io/part-of": "opendatahub-operator"
            },
            "name": "default-dsci"
            },
            "spec": {
            "applicationsNamespace": "opendatahub",
            "monitoring": {
             "managementState": "Managed",
             "namespace": "opendatahub"
            },
            "serviceMesh": {
             "controlPlane": {
               "metricsCollection": "Istio",
               "name": "data-science-smcp",
               "namespace": "istio-system"
             },
             "managementState": "Managed"
            },
            "trustedCABundle": {
             "customCABundle": "",
             "managementState": "Managed"
            }
            }
            }
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: "The Open Data Hub is a machine-learning-as-a-service platform
          built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open
          Data Hub integrates multiple AI/ML open source components into one operator
          that can easily be downloaded and installed by OpenShift users.\n\nOpen
          Data Hub operator allows users to install and manage components of the Open
          Data Hub. Users can mix and match tools from each project to fulfill the
          needs of their use case. Each of the projects share some components, but
          can be mostly seen as an extension of each other to provide a complete solution
          for both novice and skilled enterprise users.\n    \n### Components\n\n*
          Open Data Hub Dashboard - A web dashboard that displays installed Open Data
          Hub components with easy access to component UIs and documentation\n* ODH
          Notebook Controller - Secure management of Jupyter Notebook in Kubernetes
          environments built on top of Kubeflow Notebook Controller with support for
          OAuth\n* Jupyter Notebooks - JupyterLab notebook that provide Python support
          for GPU workloads\n* Data Science Pipelines - Pipeline solution for end
          to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton\n*
          Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh,
          a general-purpose model serving management/routing layer\n* Distributed
          Workloads(Incubation) - Stack built to make managing distributed compute
          infrastructure in the cloud easy and intuitive for Data Scientists. This
          stack consists of three components \n                                      Codeflare
          , KubeRay and Kueue.\n* Kserve - Kserve is the Controller for for serving
          machine learning (ML) models on arbitrary frameworks"
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odh
        - notebooks
        - serving
        - training
        - pipelines
        - modelmesh
        - workbenches
        - dashboard
        - kserve
        - distributed-workloads
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: ODH
        relatedImages:
        - quay.io/opendatahub/opendatahub-operator:v2.8.0
        version: 2.8.1
      entries:
      - name: opendatahub-operator.v2.8.1
        version: 2.8.1
      name: odh-2.8.z
    - currentCSV: opendatahub-operator.v1.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kfdef.apps.kubeflow.org/v1",
                "kind": "KfDef",
                "metadata": {
                  "name": "opendatahub",
                  "namespace": "opendatahub"
                },
                "spec": {
                  "applications": [
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-common"
                        }
                      },
                      "name": "odh-common"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-dashboard"
                        }
                      },
                      "name": "odh-dashboard"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "data-science-pipelines-operator/"
                        }
                      },
                      "name": "data-science-pipelines-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-notebook-controller"
                        }
                      },
                      "name": "odh-notebook-controller"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "notebook-images"
                        }
                      },
                      "name": "notebook-images"
                    },
                    {
                      "kustomizeConfig": {
                        "parameters": [
                          {
                            "name": "monitoring-namespace",
                            "value": "opendatahub"
                          }
                        ],
                        "repoRef": {
                          "name": "manifests",
                          "path": "model-mesh"
                        }
                      },
                      "name": "model-mesh"
                    }
                  ],
                  "repos": [
                    {
                      "name": "manifests",
                      "uri": "file:///opt/manifests/odh-manifests.tar.gz"
                    }
                  ]
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhApplication",
                "metadata": {
                  "name": "example-odhapplication",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started,Model development,Model training,Notebook environments,Package management,Data Management"
                  }
                },
                "spec": {
                  "displayName": "ODHAPPLICATION DISPLAY NAME",
                  "provider": "PROVIDER_NAME",
                  "description": "Description of your ODHApplication",
                  "kfdefApplications": [
                    "<NAME OF THE ODH APPLICATION IN KFDEF>"
                  ],
                  "route": "<APPLICATION_OCP_ROUTE_OBJECT_NAME>",
                  "img": "<svg xmlns=\"http://www.w3.org/2000/svg\" role=\"img\" viewBox=\"-1.13 -1.13 58.50 48.94\"><title>Open Data Hub logo</title><g><g><g fill=\"#4d5258\" stroke-width=\".006\"><path fill=\"#bbb\" d=\"M112.30231 22.797083c-.2034 0-.35737.143517-.35737.337407 0 .192939.15397.339307.35737.339307.20339 0 .35736-.147318.35736-.339307 0-.19294-.15397-.337407-.35736-.337407zm.002.131161c.1112 0 .20149.08934.20149.206246 0 .116904-.0903.208146-.20149.208146-.1112 0-.2053-.09124-.2053-.208146 0-.116904.0931-.206246.2053-.206246zM113.04141 22.801836h-.28228v.665308h.15017v-.190088h.13211c.16918 0 .26612-.08934.26612-.244263 0-.147319-.0969-.230957-.26612-.230957zm-.008.349762h-.12451v-.225255h.12451c.0836 0 .13116.03707.13116.111202 0 .07603-.0475.114053-.13116.114053zM113.40328 22.801836v.665308h.51703v-.124508h-.36687v-.147318h.32125V23.07081h-.32125v-.144467h.35642v-.124507zM114.03814 22.801836v.665308h.14352v-.426748l.3222.426748h.13686v-.665308h-.14257v.427698L114.175 22.801836z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/><path d=\"M115.29685 22.801836h-.28228v.665308h.27658c.21004 0 .35546-.136863.35546-.332654 0-.196741-.14351-.332654-.34976-.332654zm.004.538899h-.13591v-.412491h.12926c.11405 0 .19959.08459.19959.207196 0 .121657-.0817.205295-.19294.205295zM116.21878 23.467144h.16062l-.28418-.665308h-.15492l-.28894.665308h.15493l.0523-.12926h.30794zm-.3108-.249966l.1055-.25852.10455.25852zM116.31525 22.801836v.127359h.20244v.537949h.15017v-.537949h.2034v-.127359zM117.37527 23.467144h.16063l-.28419-.665308h-.15492l-.28893.665308h.15492l.0523-.12926h.30795zm-.31079-.249966l.1055-.25852.10454.25852zM117.83433 22.801836v.665308h.15017v-.262322h.30129v.262322h.15017v-.665308h-.15017v.277528h-.30129v-.277528zM118.72025 23.188665v-.386829h-.15017v.386829c0 .175831.11785.285132.30224.285132.18343 0 .29844-.109301.29844-.285132v-.386829h-.15017v.386829c0 .09599-.0589.153971-.14827.153971-.0893 0-.15207-.05798-.15207-.153971zM119.61059 22.801836h-.30604v.665308h.31649c.15492 0 .25092-.07033.25092-.184385 0-.08744-.0532-.152071-.13876-.168228.0713-.01806.115-.07413.115-.146368 0-.102648-.0912-.166327-.23761-.166327zm-.008.268024h-.14826v-.145418h.14826c.0561 0 .0913.02566.0913.07128 0 .04657-.0352.07413-.0913.07413zm0 .274677h-.14826v-.158723h.14826c.0713 0 .11406.02946.11406.07794 0 .05132-.0428.08079-.11406.08079z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g><g fill=\"#58595b\" stroke-width=\".01\"><path d=\"M112.19192 24.15027h.0381l-.13717-.306759h-.0364l-.1376.306759h.0372l.0342-.07844h.16696zm-.18712-.110871l.0688-.157763.0697.157763zM112.27858 23.843511v.306759h.0351v-.306759zM112.64205 23.843511h-.11525v.306759h.0351v-.09729h.0802c.0758 0 .11964-.039.11964-.10649 0-.0653-.0438-.102983-.11964-.102983zm-.001.177044h-.0789V23.87594h.0789c.0552 0 .0872.02454.0872.07143 0 .0482-.032.07318-.0872.07318zM112.81775 23.825105v.325165h.0342v-.325165zM113.1137 24.15027l-.00044-.152503c-.00044-.04908-.032-.07932-.089-.07932-.0394 0-.0644.01139-.0942.03111l.0145.0241c.0241-.01709.0478-.02673.0732-.02673.0408 0 .0618.0206.0618.05434v.01446h-.0758c-.0543.000439-.0863.02586-.0863.06617 0 .03944.0307.06968.0784.06968.0368 0 .0662-.01227.0846-.03725v.03593zm-.11131-.02717c-.0324 0-.053-.01709-.053-.04251 0-.02585.0202-.03769.0587-.03769h.0719v.02805c-.009.03287-.0386.05215-.0776.05215zM113.29915 24.108201c-.0127.0074-.025.01139-.0368.01183-.0197 0-.0307-.01139-.0307-.03944v-.127086h.0732v-.02673h-.0732v-.06048h-.0337v.06048h-.0333v.02673h.0333v.132783c0 .04558.025.06573.0587.0653.0184 0 .0359-.0061.0521-.01709zM113.44228 23.854466c.0118 0 .0245.0044.0351.01183l.014-.02629c-.014-.01052-.032-.01622-.0495-.01622-.0386-.000438-.0741.02717-.0741.07713v.02586h-.0298v.02673h.0298v.196764h.0337v-.196764h.0662v-.02673h-.0662v-.03024c0-.02717.0184-.04251.0408-.04207zM113.59236 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM113.80323 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM114.22747 23.918448c-.0443.000438-.0767.01972-.0911.06004-.01-.03813-.0394-.06004-.0828-.06004-.0416.000438-.0727.01753-.0881.05346v-.05215h-.0337v.230508h.0337v-.113501c0-.04996.0285-.08458.0758-.08589.0403 0 .064.02454.064.06617v.133221h.0346v-.113501c0-.04996.0276-.08458.0754-.08589.0399 0 .064.02454.064.06617v.133221h.0342V24.00828c0-.05609-.032-.08984-.0859-.08984zM114.63474 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0487.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11088-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM114.89934 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM115.02844 23.919762l.0863.230508h.0355l.0675-.188876.0679.188876h.0355l.0855-.230508h-.0355l-.0675.195888-.0679-.195888h-.0359l-.067.195888-.0684-.195888zM115.53009 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM115.73075 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM115.95034 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11482.116569.0386 0 .071-.01402.0929-.03813l-.0188-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18843c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM116.29474 23.825105v.144615c-.0171-.03287-.0491-.05171-.0903-.05171-.0653 0-.11087.04777-.11087.116569 0 .0688.0456.117445.11131.117445.0412 0 .0727-.01884.0898-.05127v.04952h.0342v-.325165zm-.0833.297557c-.0491 0-.0837-.03637-.0837-.08721 0-.05083.0346-.08765.0837-.08765.0486 0 .0833.03681.0833.08765 0 .05083-.0346.08721-.0833.08721zM116.65984 23.918448c-.0403 0-.0719.01884-.089.05083v-.144177h-.0342v.325165h.0342v-.04952c.0171.03199.0486.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM116.89657 24.183138l.10605-.263376h-.0351l-.0732.191944-.0767-.191944h-.0359l.0955.228756-.01.02191c-.009.02191-.0241.03331-.0421.03331-.0114 0-.0219-.0035-.0324-.01095l-.0145.02717c.014.01139.0294.01665.0486.01665.0307 0 .0548-.01841.0692-.05346zM117.29947 23.842634c-.0898 0-.16083.06793-.16083.154257 0 .08633.071.154694.16083.154694.0894 0 .16039-.06836.16039-.154694 0-.08633-.071-.154257-.16039-.154257zm0 .03243c.0684 0 .12446.0539.12446.121828 0 .06792-.0561.122703-.12446.122703-.0684 0-.12533-.05478-.12533-.122703 0-.06793.057-.121828.12533-.121828zM117.64414 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0486.05083.0898.05083.0653 0 .10999-.04777.10999-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM117.9061 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM118.19616 23.918448c-.0421.000438-.0732.01709-.0894.05215v-.05084h-.0342v.230508h.0342v-.12358c.004-.0447.032-.07494.0776-.07537.0408 0 .0653.02454.0653.06573v.133221h.0342v-.141986c0-.05609-.0324-.08984-.0876-.08984zM118.65836 23.901795l.0149-.03111c-.0254-.01709-.0618-.02848-.0973-.02848-.0635 0-.10605.03068-.10605.07888 0 .109995.17266.06661.17222.148559 0 .0298-.0276.04777-.0719.04777-.0346 0-.0736-.01578-.1021-.04295l-.0158.03024c.0289.02761.0727.04689.11745.04689.0657 0 .11087-.03287.11087-.0837.00044-.11131-.17223-.06968-.17223-.149436 0-.02629.0259-.04207.0653-.04207.0245 0 .057.0075.0846.02542zM118.84168 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .11701-.04777.11701-.117007 0-.06836-.0487-.11613-.11701-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM119.18446 23.919762v.113063c0 .04996-.0281.08502-.0758.08589-.0399 0-.064-.02454-.064-.06573v-.133222h-.0342v.141986c0 .05565.0324.08984.0863.08984.0416-.000438.0723-.01753.0876-.05346v.05215h.0342v-.230508zM119.33586 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM119.62557 23.976294l.0188-.02366c-.021-.02147-.0517-.03418-.089-.03418-.0666 0-.11438.0482-.11438.116568 0 .0688.0473.116569.11438.116569.0399 0 .0719-.0149.0925-.03988l-.0197-.02016c-.0162.01928-.0412.0298-.0714.0298-.0478 0-.082-.03506-.082-.08633 0-.05083.0342-.08589.082-.08589.028 0 .0513.01008.0688.02717zM119.78997 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g></g><g><g><path fill=\"#a586f3\" d=\"m 7808.8125,2097.1875 c -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 v 329.875 c -0.8781,10.2195 7.1804,19 17.4375,19 10.2571,0 19.0227,-8.7805 18.1446,-19 v -77.7324 l 0.4179,-6.1426 c 0,-55.1597 44.2153,-99.5625 99.375,-99.5625 55.1596,0 99.5625,44.4028 99.5625,99.5625 0.01,0.1837 0.015,0.3674 0.027,0.5508 -0.026,0.4409 -0.035,0.8826 -0.027,1.3242 v 110 c 0,9.6822 7.849,17.5312 17.5312,17.5312 9.6823,0 17.5313,-7.849 17.5313,-17.5312 v -110 c 0,-0.3245 -0.012,-0.6488 -0.031,-0.9727 0.018,-0.3004 0.029,-0.6013 0.031,-0.9023 1e-4,-74.075 -60.55,-134.4374 -134.625,-134.4375 -39.9177,1e-4 -75.8465,17.5354 -100.5,45.293 V 2115 c 0.1749,-9.8356 -7.7881,-17.8833 -17.625,-17.8125 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><path fill=\"#5ed3ee\" d=\"m 7765.8027,2098.123 c -0.038,5e-4 -0.077,0 -0.1152,0 -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 V 2361 c 0,41.9333 -26.0162,79.2671 -65.4375,93.5625 -39.4213,14.2954 -83.5594,2.3741 -110.4375,-29.8125 -26.878,-32.1865 -30.6579,-77.5719 -9.5625,-113.8125 21.0955,-36.2406 62.4262,-55.2897 103.6875,-47.8125 9.5173,1.7397 18.6463,-4.5547 20.4043,-14.0686 1.758,-9.514 -4.5187,-18.6551 -14.0293,-20.4314 -55.6354,-10.082 -111.8054,15.8219 -140.25,64.6875 -28.4446,48.8656 -23.3039,110.5381 12.9375,153.9375 36.2415,43.3993 95.908,59.4005 149.0625,40.125 44.7667,-16.2337 77.0286,-54.658 86.0977,-100.1562 0,0 1.9812,-4.7686 2.5898,-9.1563 0.6086,-4.3877 0,-17.0625 0,-17.0625 v -245.0625 c 0.1769,-9.865 -7.8313,-17.9263 -17.6973,-17.8145 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><g fill=\"none\" stroke-width=\".365\" transform=\"translate(-449.265 -83.342) scale(5.91262)\"><circle cx=\"79.677\" cy=\"16.823\" r=\".645\" stroke=\"#c6ea70\"/><path stroke=\"#dff3b1\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 80.808656,18.234303 a 1.8088735,1.8088735 0 0 1 -2.309183,-0.03848 1.8088735,1.8088735 0 0 1 -0.389241,-2.276466 1.8088735,1.8088735 0 0 1 2.165211,-0.803532\"/><path stroke=\"#f2fadf\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 82.216272,17.248644 a 2.455857,2.4425189 0 0 1 -2.468226,2.016391 2.455857,2.4425189 0 0 1 -2.383743,-2.114547 2.455857,2.4425189 0 0 1 1.733108,-2.668494\"/></g></g></g></g></svg>",
                  "category": "Open Data Hub",
                  "support": "open data hub",
                  "docsLink": "http://REPLACE.THIS.URL",
                  "quickStart": "example-odhquickstart",
                  "getStartedLink": "http://REPLACE.THIS.URL",
                  "getStartedMarkDown": "# Header for Getting Started MarkDown section\nHere is where you add information about your ODHApplication in markdown format\n\n1. List item\n1. List item\n1. List item"
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhDocument",
                "metadata": {
                  "name": "example-odhdocument",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHDocument",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "type": "how-to",
                  "description": "This is an example of an ODHDocument custom resource",
                  "url": "http://REPLACE.THIS.URL",
                  "durationMinutes": 5
                }
              },
              {
                "apiVersion": "console.openshift.io/v1",
                "kind": "OdhQuickStart",
                "metadata": {
                  "name": "example-odhquickstart",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHQuickstart Name",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "durationMinutes": 5,
                  "description": "Description of an example quickstart",
                  "introduction": "### Introduction Header\nThis is the body of the introduction for an ODHQuickstart",
                  "tasks": [
                    {
                      "title": "Title of the first task",
                      "description": "### Description header for the first task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the first task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    },
                    {
                      "title": "Title of the second task",
                      "description": "### Description header for the second task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the second task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    }
                  ],
                  "conclusion": "Message for the conclusion of the ODHQuickstart",
                  "nextQuickStart": null
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v1.11.0
          createdAt: "2023-10-31T00:00:00Z"
          description: Open Data Hub Operator for deployment and management of Open
            Data Hub
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
          support: Open Data Hub
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AcceleratorProfile is the Schema for the acceleratorprofiles
            displayName: AcceleratorProfile
            kind: AcceleratorProfile
            name: acceleratorprofiles.dashboard.opendatahub.io
            version: v1alpha
          - description: KfDef is the Schema for the kfdefs API
            displayName: Kf Def
            kind: KfDef
            name: kfdefs.kfdef.apps.kubeflow.org
            version: v1
          - description: OdhApplication is the Schema for the odhapplications
            displayName: Open Data Hub Dashboard Application
            kind: OdhApplication
            name: odhapplications.dashboard.opendatahub.io
            version: v1
          - description: OdhDashboardConfig is the Schema for the odhdashboardconfigs
            displayName: Open Data Hub Dashboard Configuration
            kind: OdhDashboardConfig
            name: odhdashboardconfigs.opendatahub.io
            version: v1alpha
          - description: OdhDocument is the Schema for the odhdocuments
            displayName: Open Data Hub Dashboard Documentation
            kind: OdhDocument
            name: odhdocuments.dashboard.opendatahub.io
            version: v1
          - description: Extension for guiding user through various workflows in the
              Open Data Hub dashboard.
            displayName: Open Data Hub Dashboard QuickStart
            kind: OdhQuickStart
            name: odhquickstarts.console.openshift.io
            version: v1
        description: |
          The Open Data Hub is a machine-learning-as-a-service platform built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open Data Hub integrates multiple AI/ML open source components into one operator that can easily be downloaded and installed by OpenShift users.

          Open Data Hub operator allows users to install and manage components of the Open Data Hub. Users can mix and match tools from each project to fulfill the needs of their use case. Each of the
          projects share some components, but can be mostly seen as an extension of each other to provide a complete solution for both novice and skilled enterprise users.

          ### Components
          * Open Data Hub Dashboard - A web dashboard that displays installed Open Data Hub components with easy access to component UIs and documentation
          * ODH Notebook Controller - Secure management of Jupyter Notebook in Kubernetes environments built on top of Kubeflow Notebook Controller with support for OAuth
          * Jupyter Notebooks - JupyterLab notebook that provide Python support for GPU workloads
          * Data Science Pipelines - Pipeline solution for end to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton
          * Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh, a general-purpose model serving management/routing layer.
          * TrustyAI - TrustyAI is an open source AI Explainability (XAI) Toolkit

          To install one or multiple of these components use the default KfDef provided with the operator.
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Open Data Hub
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/opendatahub/opendatahub-operator:v1.11.0
        version: 1.11.0
      entries:
      - name: opendatahub-operator.v1.11.0
        version: 1.11.0
      - name: opendatahub-operator.v1.10.1
        version: 1.10.1
      - name: opendatahub-operator.v1.10.0
        version: 1.10.0
      - name: opendatahub-operator.v1.9.0
        version: 1.9.0
      - name: opendatahub-operator.v1.8.0
        version: 1.8.0
      - name: opendatahub-operator.v1.7.0
        version: 1.7.0
      - name: opendatahub-operator.v1.6.0
        version: 1.6.0
      - name: opendatahub-operator.v1.5.0
        version: 1.5.0
      - name: opendatahub-operator.v1.4.2
        version: 1.4.2
      - name: opendatahub-operator.v1.4.1
        version: 1.4.1
      - name: opendatahub-operator.v1.4.0
        version: 1.4.0
      - name: opendatahub-operator.v1.3.0
        version: 1.3.0
      - name: opendatahub-operator.v1.2.0
        version: 1.2.0
      - name: opendatahub-operator.v1.1.2
        version: 1.1.2
      - name: opendatahub-operator.v1.1.1
        version: 1.1.1
      name: rolling
    - currentCSV: opendatahub-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kfdef.apps.kubeflow.org/v1",
                "kind": "KfDef",
                "metadata": {
                  "name": "opendatahub",
                  "namespace": "opendatahub"
                },
                "spec": {
                  "applications": [
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-common"
                        }
                      },
                      "name": "odh-common"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-dashboard"
                        }
                      },
                      "name": "odh-dashboard"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "prometheus/cluster"
                        }
                      },
                      "name": "prometheus-cluster"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "prometheus/operator"
                        }
                      },
                      "name": "prometheus-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "grafana/cluster"
                        }
                      },
                      "name": "grafana-cluster"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "grafana/grafana"
                        }
                      },
                      "name": "grafana-instance"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-notebook-controller"
                        }
                      },
                      "name": "odh-notebook-controller"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "notebook-images"
                        }
                      },
                      "name": "notebook-images"
                    },
                    {
                      "kustomizeConfig": {
                        "overlays": [
                          "odh-model-controller"
                        ],
                        "repoRef": {
                          "name": "manifests",
                          "path": "model-mesh"
                        }
                      },
                      "name": "model-mesh"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "data-science-pipelines-operator"
                        }
                      },
                      "name": "data-science-pipelines-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "trustyai-service"
                        }
                      },
                      "name": "trustyai"
                    }
                  ],
                  "repos": [
                    {
                      "name": "manifests",
                      "uri": "https://github.com/opendatahub-io/odh-manifests/tarball/v1.5"
                    }
                  ]
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhApplication",
                "metadata": {
                  "name": "example-odhapplication",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started,Model development,Model training,Notebook environments,Package management,Data Management"
                  }
                },
                "spec": {
                  "displayName": "ODHAPPLICATION DISPLAY NAME",
                  "provider": "PROVIDER_NAME",
                  "description": "Description of your ODHApplication",
                  "kfdefApplications": [
                    "<NAME OF THE ODH APPLICATION IN KFDEF>"
                  ],
                  "route": "<APPLICATION_OCP_ROUTE_OBJECT_NAME>",
                  "img": "<svg xmlns=\"http://www.w3.org/2000/svg\" role=\"img\" viewBox=\"-1.13 -1.13 58.50 48.94\"><title>Open Data Hub logo</title><g><g><g fill=\"#4d5258\" stroke-width=\".006\"><path fill=\"#bbb\" d=\"M112.30231 22.797083c-.2034 0-.35737.143517-.35737.337407 0 .192939.15397.339307.35737.339307.20339 0 .35736-.147318.35736-.339307 0-.19294-.15397-.337407-.35736-.337407zm.002.131161c.1112 0 .20149.08934.20149.206246 0 .116904-.0903.208146-.20149.208146-.1112 0-.2053-.09124-.2053-.208146 0-.116904.0931-.206246.2053-.206246zM113.04141 22.801836h-.28228v.665308h.15017v-.190088h.13211c.16918 0 .26612-.08934.26612-.244263 0-.147319-.0969-.230957-.26612-.230957zm-.008.349762h-.12451v-.225255h.12451c.0836 0 .13116.03707.13116.111202 0 .07603-.0475.114053-.13116.114053zM113.40328 22.801836v.665308h.51703v-.124508h-.36687v-.147318h.32125V23.07081h-.32125v-.144467h.35642v-.124507zM114.03814 22.801836v.665308h.14352v-.426748l.3222.426748h.13686v-.665308h-.14257v.427698L114.175 22.801836z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/><path d=\"M115.29685 22.801836h-.28228v.665308h.27658c.21004 0 .35546-.136863.35546-.332654 0-.196741-.14351-.332654-.34976-.332654zm.004.538899h-.13591v-.412491h.12926c.11405 0 .19959.08459.19959.207196 0 .121657-.0817.205295-.19294.205295zM116.21878 23.467144h.16062l-.28418-.665308h-.15492l-.28894.665308h.15493l.0523-.12926h.30794zm-.3108-.249966l.1055-.25852.10455.25852zM116.31525 22.801836v.127359h.20244v.537949h.15017v-.537949h.2034v-.127359zM117.37527 23.467144h.16063l-.28419-.665308h-.15492l-.28893.665308h.15492l.0523-.12926h.30795zm-.31079-.249966l.1055-.25852.10454.25852zM117.83433 22.801836v.665308h.15017v-.262322h.30129v.262322h.15017v-.665308h-.15017v.277528h-.30129v-.277528zM118.72025 23.188665v-.386829h-.15017v.386829c0 .175831.11785.285132.30224.285132.18343 0 .29844-.109301.29844-.285132v-.386829h-.15017v.386829c0 .09599-.0589.153971-.14827.153971-.0893 0-.15207-.05798-.15207-.153971zM119.61059 22.801836h-.30604v.665308h.31649c.15492 0 .25092-.07033.25092-.184385 0-.08744-.0532-.152071-.13876-.168228.0713-.01806.115-.07413.115-.146368 0-.102648-.0912-.166327-.23761-.166327zm-.008.268024h-.14826v-.145418h.14826c.0561 0 .0913.02566.0913.07128 0 .04657-.0352.07413-.0913.07413zm0 .274677h-.14826v-.158723h.14826c.0713 0 .11406.02946.11406.07794 0 .05132-.0428.08079-.11406.08079z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g><g fill=\"#58595b\" stroke-width=\".01\"><path d=\"M112.19192 24.15027h.0381l-.13717-.306759h-.0364l-.1376.306759h.0372l.0342-.07844h.16696zm-.18712-.110871l.0688-.157763.0697.157763zM112.27858 23.843511v.306759h.0351v-.306759zM112.64205 23.843511h-.11525v.306759h.0351v-.09729h.0802c.0758 0 .11964-.039.11964-.10649 0-.0653-.0438-.102983-.11964-.102983zm-.001.177044h-.0789V23.87594h.0789c.0552 0 .0872.02454.0872.07143 0 .0482-.032.07318-.0872.07318zM112.81775 23.825105v.325165h.0342v-.325165zM113.1137 24.15027l-.00044-.152503c-.00044-.04908-.032-.07932-.089-.07932-.0394 0-.0644.01139-.0942.03111l.0145.0241c.0241-.01709.0478-.02673.0732-.02673.0408 0 .0618.0206.0618.05434v.01446h-.0758c-.0543.000439-.0863.02586-.0863.06617 0 .03944.0307.06968.0784.06968.0368 0 .0662-.01227.0846-.03725v.03593zm-.11131-.02717c-.0324 0-.053-.01709-.053-.04251 0-.02585.0202-.03769.0587-.03769h.0719v.02805c-.009.03287-.0386.05215-.0776.05215zM113.29915 24.108201c-.0127.0074-.025.01139-.0368.01183-.0197 0-.0307-.01139-.0307-.03944v-.127086h.0732v-.02673h-.0732v-.06048h-.0337v.06048h-.0333v.02673h.0333v.132783c0 .04558.025.06573.0587.0653.0184 0 .0359-.0061.0521-.01709zM113.44228 23.854466c.0118 0 .0245.0044.0351.01183l.014-.02629c-.014-.01052-.032-.01622-.0495-.01622-.0386-.000438-.0741.02717-.0741.07713v.02586h-.0298v.02673h.0298v.196764h.0337v-.196764h.0662v-.02673h-.0662v-.03024c0-.02717.0184-.04251.0408-.04207zM113.59236 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM113.80323 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM114.22747 23.918448c-.0443.000438-.0767.01972-.0911.06004-.01-.03813-.0394-.06004-.0828-.06004-.0416.000438-.0727.01753-.0881.05346v-.05215h-.0337v.230508h.0337v-.113501c0-.04996.0285-.08458.0758-.08589.0403 0 .064.02454.064.06617v.133221h.0346v-.113501c0-.04996.0276-.08458.0754-.08589.0399 0 .064.02454.064.06617v.133221h.0342V24.00828c0-.05609-.032-.08984-.0859-.08984zM114.63474 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0487.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11088-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM114.89934 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM115.02844 23.919762l.0863.230508h.0355l.0675-.188876.0679.188876h.0355l.0855-.230508h-.0355l-.0675.195888-.0679-.195888h-.0359l-.067.195888-.0684-.195888zM115.53009 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM115.73075 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM115.95034 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11482.116569.0386 0 .071-.01402.0929-.03813l-.0188-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18843c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM116.29474 23.825105v.144615c-.0171-.03287-.0491-.05171-.0903-.05171-.0653 0-.11087.04777-.11087.116569 0 .0688.0456.117445.11131.117445.0412 0 .0727-.01884.0898-.05127v.04952h.0342v-.325165zm-.0833.297557c-.0491 0-.0837-.03637-.0837-.08721 0-.05083.0346-.08765.0837-.08765.0486 0 .0833.03681.0833.08765 0 .05083-.0346.08721-.0833.08721zM116.65984 23.918448c-.0403 0-.0719.01884-.089.05083v-.144177h-.0342v.325165h.0342v-.04952c.0171.03199.0486.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM116.89657 24.183138l.10605-.263376h-.0351l-.0732.191944-.0767-.191944h-.0359l.0955.228756-.01.02191c-.009.02191-.0241.03331-.0421.03331-.0114 0-.0219-.0035-.0324-.01095l-.0145.02717c.014.01139.0294.01665.0486.01665.0307 0 .0548-.01841.0692-.05346zM117.29947 23.842634c-.0898 0-.16083.06793-.16083.154257 0 .08633.071.154694.16083.154694.0894 0 .16039-.06836.16039-.154694 0-.08633-.071-.154257-.16039-.154257zm0 .03243c.0684 0 .12446.0539.12446.121828 0 .06792-.0561.122703-.12446.122703-.0684 0-.12533-.05478-.12533-.122703 0-.06793.057-.121828.12533-.121828zM117.64414 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0486.05083.0898.05083.0653 0 .10999-.04777.10999-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM117.9061 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM118.19616 23.918448c-.0421.000438-.0732.01709-.0894.05215v-.05084h-.0342v.230508h.0342v-.12358c.004-.0447.032-.07494.0776-.07537.0408 0 .0653.02454.0653.06573v.133221h.0342v-.141986c0-.05609-.0324-.08984-.0876-.08984zM118.65836 23.901795l.0149-.03111c-.0254-.01709-.0618-.02848-.0973-.02848-.0635 0-.10605.03068-.10605.07888 0 .109995.17266.06661.17222.148559 0 .0298-.0276.04777-.0719.04777-.0346 0-.0736-.01578-.1021-.04295l-.0158.03024c.0289.02761.0727.04689.11745.04689.0657 0 .11087-.03287.11087-.0837.00044-.11131-.17223-.06968-.17223-.149436 0-.02629.0259-.04207.0653-.04207.0245 0 .057.0075.0846.02542zM118.84168 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .11701-.04777.11701-.117007 0-.06836-.0487-.11613-.11701-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM119.18446 23.919762v.113063c0 .04996-.0281.08502-.0758.08589-.0399 0-.064-.02454-.064-.06573v-.133222h-.0342v.141986c0 .05565.0324.08984.0863.08984.0416-.000438.0723-.01753.0876-.05346v.05215h.0342v-.230508zM119.33586 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM119.62557 23.976294l.0188-.02366c-.021-.02147-.0517-.03418-.089-.03418-.0666 0-.11438.0482-.11438.116568 0 .0688.0473.116569.11438.116569.0399 0 .0719-.0149.0925-.03988l-.0197-.02016c-.0162.01928-.0412.0298-.0714.0298-.0478 0-.082-.03506-.082-.08633 0-.05083.0342-.08589.082-.08589.028 0 .0513.01008.0688.02717zM119.78997 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g></g><g><g><path fill=\"#a586f3\" d=\"m 7808.8125,2097.1875 c -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 v 329.875 c -0.8781,10.2195 7.1804,19 17.4375,19 10.2571,0 19.0227,-8.7805 18.1446,-19 v -77.7324 l 0.4179,-6.1426 c 0,-55.1597 44.2153,-99.5625 99.375,-99.5625 55.1596,0 99.5625,44.4028 99.5625,99.5625 0.01,0.1837 0.015,0.3674 0.027,0.5508 -0.026,0.4409 -0.035,0.8826 -0.027,1.3242 v 110 c 0,9.6822 7.849,17.5312 17.5312,17.5312 9.6823,0 17.5313,-7.849 17.5313,-17.5312 v -110 c 0,-0.3245 -0.012,-0.6488 -0.031,-0.9727 0.018,-0.3004 0.029,-0.6013 0.031,-0.9023 1e-4,-74.075 -60.55,-134.4374 -134.625,-134.4375 -39.9177,1e-4 -75.8465,17.5354 -100.5,45.293 V 2115 c 0.1749,-9.8356 -7.7881,-17.8833 -17.625,-17.8125 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><path fill=\"#5ed3ee\" d=\"m 7765.8027,2098.123 c -0.038,5e-4 -0.077,0 -0.1152,0 -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 V 2361 c 0,41.9333 -26.0162,79.2671 -65.4375,93.5625 -39.4213,14.2954 -83.5594,2.3741 -110.4375,-29.8125 -26.878,-32.1865 -30.6579,-77.5719 -9.5625,-113.8125 21.0955,-36.2406 62.4262,-55.2897 103.6875,-47.8125 9.5173,1.7397 18.6463,-4.5547 20.4043,-14.0686 1.758,-9.514 -4.5187,-18.6551 -14.0293,-20.4314 -55.6354,-10.082 -111.8054,15.8219 -140.25,64.6875 -28.4446,48.8656 -23.3039,110.5381 12.9375,153.9375 36.2415,43.3993 95.908,59.4005 149.0625,40.125 44.7667,-16.2337 77.0286,-54.658 86.0977,-100.1562 0,0 1.9812,-4.7686 2.5898,-9.1563 0.6086,-4.3877 0,-17.0625 0,-17.0625 v -245.0625 c 0.1769,-9.865 -7.8313,-17.9263 -17.6973,-17.8145 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><g fill=\"none\" stroke-width=\".365\" transform=\"translate(-449.265 -83.342) scale(5.91262)\"><circle cx=\"79.677\" cy=\"16.823\" r=\".645\" stroke=\"#c6ea70\"/><path stroke=\"#dff3b1\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 80.808656,18.234303 a 1.8088735,1.8088735 0 0 1 -2.309183,-0.03848 1.8088735,1.8088735 0 0 1 -0.389241,-2.276466 1.8088735,1.8088735 0 0 1 2.165211,-0.803532\"/><path stroke=\"#f2fadf\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 82.216272,17.248644 a 2.455857,2.4425189 0 0 1 -2.468226,2.016391 2.455857,2.4425189 0 0 1 -2.383743,-2.114547 2.455857,2.4425189 0 0 1 1.733108,-2.668494\"/></g></g></g></g></svg>",
                  "category": "Open Data Hub",
                  "support": "open data hub",
                  "docsLink": "http://REPLACE.THIS.URL",
                  "quickStart": "example-odhquickstart",
                  "getStartedLink": "http://REPLACE.THIS.URL",
                  "getStartedMarkDown": "# Header for Getting Started MarkDown section\nHere is where you add information about your ODHApplication in markdown format\n\n1. List item\n1. List item\n1. List item"
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhDocument",
                "metadata": {
                  "name": "example-odhdocument",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHDocument",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "type": "how-to",
                  "description": "This is an example of an ODHDocument custom resource",
                  "url": "http://REPLACE.THIS.URL",
                  "durationMinutes": 5
                }
              },
              {
                "apiVersion": "console.openshift.io/v1",
                "kind": "OdhQuickStart",
                "metadata": {
                  "name": "example-odhquickstart",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHQuickstart Name",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "durationMinutes": 5,
                  "description": "Description of an example quickstart",
                  "introduction": "### Introduction Header\nThis is the body of the introduction for an ODHQuickstart",
                  "tasks": [
                    {
                      "title": "Title of the first task",
                      "description": "### Description header for the first task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the first task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    },
                    {
                      "title": "Title of the second task",
                      "description": "### Description header for the second task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the second task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    }
                  ],
                  "conclusion": "Message for the conclusion of the ODHQuickstart",
                  "nextQuickStart": null
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v1.5.0
          createdAt: "2022-10-18T00:00:00Z"
          description: Open Data Hub Operator for deployment and management of Open
            Data Hub
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
          support: Open Data Hub
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KfDef is the Schema for the kfdefs API
            displayName: Kf Def
            kind: KfDef
            name: kfdefs.kfdef.apps.kubeflow.org
            version: v1
          - description: OdhApplication is the Schema for the odhapplications
            displayName: Open Data Hub Dashboard Application
            kind: OdhApplication
            name: odhapplications.dashboard.opendatahub.io
            version: v1
          - description: OdhDashboardConfig is the Schema for the odhdashboardconfigs
            displayName: Open Data Hub Dashboard Configuration
            kind: OdhDashboardConfig
            name: odhdashboardconfigs.opendatahub.io
            version: v1alpha
          - description: OdhDocument is the Schema for the odhdocuments
            displayName: Open Data Hub Dashboard Documentation
            kind: OdhDocument
            name: odhdocuments.dashboard.opendatahub.io
            version: v1
          - description: Extension for guiding user through various workflows in the
              Open Data Hub dashboard.
            displayName: Open Data Hub Dashboard QuickStart
            kind: OdhQuickStart
            name: odhquickstarts.console.openshift.io
            version: v1
        description: |
          The Open Data Hub is a machine-learning-as-a-service platform built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open Data Hub integrates multiple AI/ML open source components into one operator that can easily be downloaded and installed by OpenShift users.

          Open Data Hub operator allows users to install and manage components of the Open Data Hub. Users can mix and match tools from each project to fulfill the needs of their use case. Each of the
          projects share some components, but can be mostly seen as an extension of each other to provide a complete solution for both novice and skilled enterprise users.

          ### Components
          * Open Data Hub Dashboard - A web dashboard that displays installed Open Data Hub components with easy access to component UIs and documentation
          * ODH Notebook Controller - Secure management of Jupyter Notebook in Kubernetes environments built on top of Kubeflow Notebook Controller with support for OAuth
          * Jupyter Notebooks - JupyterLab notebook that provide Python support for GPU workloads
          * Data Science Pipelines - Pipeline solution for end to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton
          * Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh, a general-purpose model serving management/routing layer.

          To install one or multiple of these components use the default KfDef provided with the operator.

          ### Available Channels

          #### Stable

          Channel `stable` offers the major releases of Open Data Hub operator and ODH manifests. It is based on Kubeflow Operator and Kustomize for deployment configuration. This version is compatible with [Open Data Hub manifests](https://github.com/opendatahub-io/odh-manifests/) as well as [Kubeflow manifests](https://github.com/opendatahub-io/manifests). This channel will offer new components and features that have been thoroughly tested by consumers of the `rolling` channel

          NOTE: The `stable` channel will be deprecated starting with the ODH 1.6 release and we will only be publishing updates on the `rolling` channel for all future releases

          #### Rolling

          Channel `rolling` will offer the latest release of Open Data Hub operator and manifests. Subscribing to this branch will provide faster updates and access to components that will be in developement and testing for the next stable release
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io
        maintainers:
        - email: maintainers@lists.opendatahub.io
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Open Data Hub
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/opendatahub/opendatahub-operator:v1.5.0
        version: 1.5.0
      entries:
      - name: opendatahub-operator.v1.5.0
        version: 1.5.0
      - name: opendatahub-operator.v1.4.2
        version: 1.4.2
      - name: opendatahub-operator.v1.4.1
        version: 1.4.1
      - name: opendatahub-operator.v1.4.0
        version: 1.4.0
      - name: opendatahub-operator.v1.3.0
        version: 1.3.0
      - name: opendatahub-operator.v1.2.0
        version: 1.2.0
      - name: opendatahub-operator.v1.1.2
        version: 1.1.2
      - name: opendatahub-operator.v1.1.1
        version: 1.1.1
      name: stable
    defaultChannel: fast
    packageName: opendatahub-operator
    provider:
      name: ODH
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Thales Group
      provider-url: https://cpl.thalesgroup.com/data-protection
    name: cte-k8s-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cte-k8s-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cte-k8s-operator.csi.cte.cpl.thalesgroup.com/v1",
                "kind": "CteK8sOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cte-k8s-operator"
                  },
                  "name": "ctek8soperator-sample"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-06-05T08:24:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CteK8sOperator is the Schema for the ctek8soperators API
            displayName: Cte K8s Operator
            kind: CteK8sOperator
            name: ctek8soperators.cte-k8s-operator.csi.cte.cpl.thalesgroup.com
            version: v1
        description: |-
          # Kubernetes Operator to deploy CTE for Kubernetes on an OCP/K8s Cluster.
          This Operator deploys CTE for Kubernetes on the OCP/K8s Cluster.

          ## About this operator

          It is a GO based operator that manages the deployment, upgrade and
          deletion of the CTE for Kubernetes on all the nodes of the Cluster on
          which it is installed.

          ## Prerequisites

          Applicable only on OCP cluster
            For an OCP cluster the operator requires a SecurityContextConstraints (SCC)
            to be created before installation. To create the SecurityContextConstraints
            you can use a definition provided by the deployment scripts at
            https://github.com/thalescpl-io/ciphertrust-transparent-encryption-kubernetes

            Look for cte-csi-scc.yaml file in deploy/kubernetes/X.Y.Z/operator-deploy/
            Where X.Y.Z is the last released version of the deploy scripts.

          For more information refer to the guide at:
            https://thalesdocs.com/ctp/cte-con/cte-k8s/latest/cte-4-k8s/index.html
        displayName: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Encryption
        - Data Protection
        - Security
        - Kubernetes
        - Openshift
        links:
        - name: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          url: https://cpl.thalesgroup.com/encryption/ciphertrust-transparent-data-encryption-kubernetes
        maintainers:
        - email: technical.support.DIS@thalesgroup.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Thales Group
          url: https://cpl.thalesgroup.com/data-protection
        relatedImages:
        - docker.io/thalesciphertrust/ciphertrust-transparent-encryption-kubernetes-operator@sha256:d58cbb65d758fd0b9d4cfba4cb48a35bdaebe2e239fb25edb9b491cd8c2c3f73
        version: 1.6.5
      entries:
      - name: cte-k8s-operator.v1.6.5
        version: 1.6.5
      name: candidate
    - currentCSV: cte-k8s-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cte-k8s-operator.csi.cte.cpl.thalesgroup.com/v1",
                "kind": "CteK8sOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cte-k8s-operator"
                  },
                  "name": "ctek8soperator-sample"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-06-05T08:24:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CteK8sOperator is the Schema for the ctek8soperators API
            displayName: Cte K8s Operator
            kind: CteK8sOperator
            name: ctek8soperators.cte-k8s-operator.csi.cte.cpl.thalesgroup.com
            version: v1
        description: |-
          # Kubernetes Operator to deploy CTE for Kubernetes on an OCP/K8s Cluster.
          This Operator deploys CTE for Kubernetes on the OCP/K8s Cluster.

          ## About this operator

          It is a GO based operator that manages the deployment, upgrade and
          deletion of the CTE for Kubernetes on all the nodes of the Cluster on
          which it is installed.

          ## Prerequisites

          Applicable only on OCP cluster
            For an OCP cluster the operator requires a SecurityContextConstraints (SCC)
            to be created before installation. To create the SecurityContextConstraints
            you can use a definition provided by the deployment scripts at
            https://github.com/thalescpl-io/ciphertrust-transparent-encryption-kubernetes

            Look for cte-csi-scc.yaml file in deploy/kubernetes/X.Y.Z/operator-deploy/
            Where X.Y.Z is the last released version of the deploy scripts.

          For more information refer to the guide at:
            https://thalesdocs.com/ctp/cte-con/cte-k8s/latest/cte-4-k8s/index.html
        displayName: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Encryption
        - Data Protection
        - Security
        - Kubernetes
        - Openshift
        links:
        - name: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          url: https://cpl.thalesgroup.com/encryption/ciphertrust-transparent-data-encryption-kubernetes
        maintainers:
        - email: technical.support.DIS@thalesgroup.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Thales Group
          url: https://cpl.thalesgroup.com/data-protection
        relatedImages:
        - docker.io/thalesciphertrust/ciphertrust-transparent-encryption-kubernetes-operator@sha256:d58cbb65d758fd0b9d4cfba4cb48a35bdaebe2e239fb25edb9b491cd8c2c3f73
        version: 1.6.5
      entries:
      - name: cte-k8s-operator.v1.6.5
        version: 1.6.5
      name: stable
    defaultChannel: stable
    packageName: cte-k8s-operator
    provider:
      name: Thales Group
      url: https://cpl.thalesgroup.com/data-protection
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-team-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-team-operator.v2024.10.2-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": null,
                        "host": null,
                        "process": null
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master",
                        "operator": "",
                        "value": ""
                      }
                    ]
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": null
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:12c1d2e9347ec447196a618b481a0970dacac7a8a72a3b38ec22efa69668d180
          createdAt: "2024-10-25T01:47:52Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2024.10.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2023.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                   operator: ""
                   value: ""

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:7c4868db021beaffa762eb0de43739bbc865a4bf5900f1d17f78caed8612c073
        - registry.connect.redhat.com/ca/uma-operator@sha256:12c1d2e9347ec447196a618b481a0970dacac7a8a72a3b38ec22efa69668d180
        version: 2024.10.2-12
      entries:
      - name: uma-team-operator.v2024.10.2-12
        version: 2024.10.2-12
      name: "2024.10"
    - currentCSV: uma-team-operator.v2025.1.3-33
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": null,
                        "host": null,
                        "process": null
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.1.1.33,990033",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master",
                        "operator": "",
                        "value": ""
                      }
                    ]
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": null
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:78b66a291f58aa94123cdd59f2aa07dce4928162c481fe9de5ce524782bc93de
          createdAt: "2025-02-18T05:35:06Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.1.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2023.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.1.1.26,990026"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                   operator: ""
                   value: ""

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:78b66a291f58aa94123cdd59f2aa07dce4928162c481fe9de5ce524782bc93de
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:ae91a28d57c8a359f81757e4bf79472f2ac189ac0ae3ee0d87e8ff9c433f1ccc
        version: 2025.1.3-33
      entries:
      - name: uma-team-operator.v2025.1.3-33
        version: 2025.1.3-33
      name: "2025.1"
    - currentCSV: uma-team-operator.v2025.2.1-25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.2.1.25,990025",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master",
                        "operator": "",
                        "value": ""
                      }
                    ]
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:d0a0090a871ff7e23d3c97b620a1b0d004a8317efe1581ba0437d073c5656825
          createdAt: "2025-03-05T05:51:44Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.2.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2023.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.1.1.26,990026"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                   operator: ""
                   value: ""

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:d0a0090a871ff7e23d3c97b620a1b0d004a8317efe1581ba0437d073c5656825
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:8bcf0dac889f38a0bf77eb9babe3f726e3b2564d5988c05bdc5c2d91ae189c61
        version: 2025.2.1-25
      entries:
      - name: uma-team-operator.v2025.2.1-25
        version: 2025.2.1-25
      name: "2025.2"
    - currentCSV: uma-team-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:60fb1f76b4d4aab569eec5f2c20f7a06d813759f5fb6ac952156b41c4d295a52
          createdAt: "2025-04-16T06:57:55Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.4.5
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.4.2.5,990005"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:60fb1f76b4d4aab569eec5f2c20f7a06d813759f5fb6ac952156b41c4d295a52
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        version: 2025.4.5-5
      entries:
      - name: uma-team-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-team-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:052055a6280f60a69caccf89f50fd8bf92da61cd59f1962f61d21467c4191970
          createdAt: "2025-07-24T14:54:58Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.5.2.12,990012"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:052055a6280f60a69caccf89f50fd8bf92da61cd59f1962f61d21467c4191970
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        version: 2025.5.4-12
      entries:
      - name: uma-team-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-team-operator.v2025.7.3-21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:e3ae2240aeb618f36f41073a5f65b03a7ad5dfdccc7f9feb830506fbd2bfd604
          createdAt: "2025-08-23T09:01:14Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.7.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.7.1.23,990023"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:e3ae2240aeb618f36f41073a5f65b03a7ad5dfdccc7f9feb830506fbd2bfd604
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.3-21
      entries:
      - name: uma-team-operator.v2025.7.3-21
        version: 2025.7.3-21
      name: "2025.7"
    - currentCSV: uma-team-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "hostmonitoring": {
                        "linux": {
                          "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:b8422d0f3a9e576948ced8244a91abc3283b26217e4c7efc7e2669ca5508930e
          createdAt: "2025-08-15T03:20:57Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 hostmonitoring:
                   linux:
                   # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                   groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.8.1.9,990009"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:b8422d0f3a9e576948ced8244a91abc3283b26217e4c7efc7e2669ca5508930e
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        version: 2025.8.2-9
      entries:
      - name: uma-team-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-team-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
          createdAt: "2025-09-24T12:24:28Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.12,990012"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-team-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-team-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "hostmonitoring": {
                        "linux": {
                          "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:8c12315745ad3564035a4db923698bf63fe32dddbc7d246f5a3867945305713e
          createdAt: "2025-07-30T08:27:50Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 hostmonitoring:
                   linux:
                   # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                   groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "99.99.0.serp_F152813,19"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:8c12315745ad3564035a4db923698bf63fe32dddbc7d246f5a3867945305713e
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        version: 2025.7.2-19
      entries:
      - name: uma-team-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-team-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:89e0505b5e526beffda9b13a1ca7b49e5ba144fd0166a7e171e9d9b084915a56
          createdAt: "2025-09-11T11:16:37Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.8,990008"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:89e0505b5e526beffda9b13a1ca7b49e5ba144fd0166a7e171e9d9b084915a56
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-team-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-team-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
          createdAt: "2025-09-24T12:24:28Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.12,990012"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-team-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-team-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ocs-client-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocs-client-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          console.openshift.io/plugins: '["odf-client-console"]'
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageClassClaim is the Schema for the storageclassclaims
              API
            displayName: Storage Class Claim
            kind: StorageClassClaim
            name: storageclassclaims.ocs.openshift.io
            version: v1alpha1
          - description: StorageClient is the Schema for the storageclients API
            displayName: Storage Client
            kind: StorageClient
            name: storageclients.ocs.openshift.io
            version: v1alpha1
        description: This operator is not intended to be directly installed
        displayName: OpenShift Data Foundation Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: OCS Client Operator
          url: https://github.com/red-hat-storage/ocs-client-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:975529d3fbdcf6f78cc1602333832f543da2918ac5c018d1ee6bea2b315106be
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:ef4276b80cacae825d4831de88d3fe6ba7cc324464087b7b28734bd2b8de3315
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel8@sha256:d83b44a1257e8015409cd261e8eed5bd8c4e581b788e8452f19114ba696a30c1
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:bc637bbbed3457c45fbb5cb1e8940b2a1dacfa64959e5bd0bfb58c5d669427fd
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:1dd7c0ef17d39ebf3311aee326202e09038cece1b7554cea23c1efe582620d0b
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:c9e5b48d84febf10ff4c9b78a8210df36c90905fbb92b6a6b5b32e0fd9a6a41d
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:9a0008f662b6afa9e2677a8a83bd662d806ad6f4d03fb49bccb78ed487ce4c7a
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:028120f2212738917f49e0330b60be5f79a208ce2974bea22b71f533aa6ce3be
        - registry.redhat.io/odf4/ocs-client-console-rhel9@sha256:ebf3f6098a9a0a06706d20fe8fc86513be34b98604bc70fd82910f0fcd17f191
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:ffcc0806ec2afb8b40d75c70ef2a6d98bb887879f5b3f573e900861316ce9e41
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel8@sha256:05cf61192ee75dc11f5777bebeacdd209232042445553329e14f542ac0084c71
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:e206dcc33863f4d8d0aa6df173a86cd50ee1d52e6bcf9efc43b39831acc20573
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:0dc0733df67bb49886ac9715b81bd4ec4a36330fe1afdde657587e1883f5ca7a
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel8@sha256:647074ada8cb88f7ed99ca14621d3b754db6a60346f12b7561cd035c05f283f7
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:390b527d10e9c7a8a9191251b0a1ba501128057cefaff35e950f0efe04ef3122
        - registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:cb93b68adb857c82f23f292cdee47d52656dc4fc5bb83fac8581110869cbe841
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:645d4fd09628f7ddcfc765e4ddacaa79fa6a7c35c03292ca2f6e317d48765d6b
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel8@sha256:1f6a17909f1fe6e09a01cad38b6a50b41d850b1c5cffd24ef9a87c8798ff15e7
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:1711a0aabd28937aac1424adebe96619a48b91f7f888f0588117f80891001da4
        version: 4.15.20-rhodf
      entries:
      - name: ocs-client-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: ocs-client-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: ocs-client-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: ocs-client-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: ocs-client-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: ocs-client-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: ocs-client-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: ocs-client-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: ocs-client-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: ocs-client-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: ocs-client-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: ocs-client-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: ocs-client-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: ocs-client-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: ocs-client-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: ocs-client-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: ocs-client-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: ocs-client-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: ocs-client-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: ocs-client-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: ocs-client-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-client-console"]'
          containerImage: registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:bd8454b5a0b235669732650ae8502831811793d93fbfbd0d0f53278fcdfed459
          createdAt: "2025-09-12T06:29:47Z"
          description: OpenShift Data Foundation client operator enables consumption
            of storage services from a remote centralized OpenShift Data Foundation
            provider cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["storageclaims.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/ocs-client-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageClaim is the Schema for the storageclaims API
            displayName: Storage Claim
            kind: StorageClaim
            name: storageclaims.ocs.openshift.io
            version: v1alpha1
          - description: StorageClient is the Schema for the storageclients API
            displayName: Storage Client
            kind: StorageClient
            name: storageclients.ocs.openshift.io
            version: v1alpha1
        description: OpenShift Data Foundation client operator enables consumption
          of storage services from a remote centralized OpenShift Data Foundation
          provider cluster.
        displayName: OpenShift Data Foundation Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: OCS Client Operator
          url: https://github.com/red-hat-storage/ocs-client-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:80c01633a6bfca44df37ac38f764035be6998fce965c8709e4f6d54d17eeaa7b
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:82a4b481028195cfd75471656fe40c1cbbade4c7b2d170d6bf9422d9d3b84fa7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel8@sha256:d83b44a1257e8015409cd261e8eed5bd8c4e581b788e8452f19114ba696a30c1
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel8@sha256:1f6a17909f1fe6e09a01cad38b6a50b41d850b1c5cffd24ef9a87c8798ff15e7
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:975529d3fbdcf6f78cc1602333832f543da2918ac5c018d1ee6bea2b315106be
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c11a599bfd09b9b22888e03eabdbe1054d4049718a591d2a45bf0ed89800111e
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6fee56851da3a87b9f7f477fc59e3385e118dc0b398861a8f13459c998a5af77
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:0afcb516eb528857887c6ee1a9cba081dc8f6b7d47d9f26a0dbb2dca0cefe07c
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:8d8326c74795c9c6bd8f33a9f9e81d2304a70b1f9be3c2f8d50e76be4b0292ff
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/odf4/ocs-client-console-rhel9@sha256:beb4f82d3923171b80aacb90dfe345a22626453bf2f3921b03a16839b204d377
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:1711a0aabd28937aac1424adebe96619a48b91f7f888f0588117f80891001da4
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:ef4276b80cacae825d4831de88d3fe6ba7cc324464087b7b28734bd2b8de3315
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:028120f2212738917f49e0330b60be5f79a208ce2974bea22b71f533aa6ce3be
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:645d4fd09628f7ddcfc765e4ddacaa79fa6a7c35c03292ca2f6e317d48765d6b
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:bc637bbbed3457c45fbb5cb1e8940b2a1dacfa64959e5bd0bfb58c5d669427fd
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:ffcc0806ec2afb8b40d75c70ef2a6d98bb887879f5b3f573e900861316ce9e41
        - registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:bd8454b5a0b235669732650ae8502831811793d93fbfbd0d0f53278fcdfed459
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:693a60d6a1162659213ae554b1c6ee200b0300a7da72c01415044cf507acab49
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:c9e5b48d84febf10ff4c9b78a8210df36c90905fbb92b6a6b5b32e0fd9a6a41d
        version: 4.16.16-rhodf
      entries:
      - name: ocs-client-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: ocs-client-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: ocs-client-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: ocs-client-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: ocs-client-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: ocs-client-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: ocs-client-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: ocs-client-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: ocs-client-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: ocs-client-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: ocs-client-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: ocs-client-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: ocs-client-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: ocs-client-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: ocs-client-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: ocs-client-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: ocs-client-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: ocs-client-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://your.domain
    name: amq-streams-proxy
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: alpha
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: amq-streams-3.0.x
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: amq-streams-3.x
    defaultChannel: alpha
    packageName: amq-streams-proxy
    provider:
      name: Red Hat
      url: https://your.domain
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-secretsmanager-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-secretsmanager-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secretsmanager.services.k8s.aws/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/secretsmanager-controller:1.1.1
          createdAt: "2025-09-19T21:14:43Z"
          description: AWS Secrets Manager controller is a service controller for
            managing Secrets Manager resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret represents the state of an AWS secretsmanager Secret
              resource.
            displayName: Secret
            kind: Secret
            name: secrets.secretsmanager.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Secrets Manager resources in AWS from within your Kubernetes cluster.

          **About Amazon Secrets Manager**

          AWS Secrets Manager helps you manage, retrieve, and rotate database credentials, application credentials, OAuth tokens, API keys, and other secrets throughout their lifecycles. [Many AWS services that use secrets](https://docs.aws.amazon.com/secretsmanager/latest/userguide/integrating.html) store them in Secrets Manager.

          Secrets Manager helps you improve your security posture, because you no longer need hard-coded credentials in application source code. Storing the credentials in Secrets Manager helps avoid possible compromise by anyone who can inspect your application or the components. You replace hard-coded credentials with a runtime call to the Secrets Manager service to retrieve credentials dynamically when you need them.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Secrets Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secretsmanager
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Secrets Manager Developer Resources
          url: https://aws.amazon.com/secrets-manager/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: secretsmanager maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/secretsmanager-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-secretsmanager-controller.v1.1.1
        version: 1.1.1
      - name: ack-secretsmanager-controller.v1.1.0
        version: 1.1.0
      - name: ack-secretsmanager-controller.v1.0.14
        version: 1.0.14
      - name: ack-secretsmanager-controller.v1.0.13
        version: 1.0.13
      - name: ack-secretsmanager-controller.v1.0.12
        version: 1.0.12
      - name: ack-secretsmanager-controller.v1.0.11
        version: 1.0.11
      - name: ack-secretsmanager-controller.v1.0.10
        version: 1.0.10
      - name: ack-secretsmanager-controller.v1.0.9
        version: 1.0.9
      - name: ack-secretsmanager-controller.v1.0.8
        version: 1.0.8
      - name: ack-secretsmanager-controller.v1.0.7
        version: 1.0.7
      - name: ack-secretsmanager-controller.v1.0.6
        version: 1.0.6
      - name: ack-secretsmanager-controller.v1.0.5
        version: 1.0.5
      - name: ack-secretsmanager-controller.v1.0.4
        version: 1.0.4
      - name: ack-secretsmanager-controller.v1.0.3
        version: 1.0.3
      - name: ack-secretsmanager-controller.v1.0.1
        version: 1.0.1
      - name: ack-secretsmanager-controller.v1.0.0
        version: 1.0.0
      - name: ack-secretsmanager-controller.v0.0.11
        version: 0.0.11
      - name: ack-secretsmanager-controller.v0.0.10
        version: 0.0.10
      - name: ack-secretsmanager-controller.v0.0.9
        version: 0.0.9
      - name: ack-secretsmanager-controller.v0.0.8
        version: 0.0.8
      - name: ack-secretsmanager-controller.v0.0.7
        version: 0.0.7
      - name: ack-secretsmanager-controller.v0.0.2
        version: 0.0.2
      - name: ack-secretsmanager-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-secretsmanager-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: crunchy-postgres-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:c559c896e174674d151f93543ac92a1b5647eb4676313f7dc6cbb8faee786f18
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:4bdd0588ffbb3f628e7df1c72670e7a5087269438343b68779012bc3a13dea31
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-exporter@sha256:3b231902b2116cd38de05c7378202716c95813b9961fb8cee811b61d36cddf86
        - registry.connect.redhat.com/crunchydata/crunchy-upgrade@sha256:68bc93cce61d9b1e44822ae3e3ac8f7a9c2b712eddfaea090eabefacd60c67f6
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:0e283616c2d32e53258d62d02702bc4e179ef8c03933592154babed1a70bf666
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:56a5e3a5758982e086c9f78201a7137e0d934316f058a31cf30df44639a569a2
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:b2c3c74fdcc0e08f555e50c527f0d322292333c5d3dce061a0dee3680a1fbc56
        - registry.connect.redhat.com/crunchydata/crunchy-pgbackrest@sha256:b70da0035b74379f2507cfabdd9d1cc1e3c3eecb9066c22f292981c9a27916e1
        - registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
        - registry.connect.redhat.com/crunchydata/crunchy-pgadmin4@sha256:0d84acb1ce6a9b046e3f8b222a583ddcc91dda0a117c9f637ef173d0ac992f57
        - registry.connect.redhat.com/crunchydata/crunchy-pgbouncer@sha256:17cb58dec78447c2f70a414a2aeff6facbf921b42a0f366d0dbf7be95aa8987a
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:a7f62f367817fc08294a0b8b82ac9826a42c0ae668efdfacaa158a4287b8a688
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.4
        version: 5.4.4
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.4.0
        version: 5.4.0
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: crunchy-postgres-operator
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hbase-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hbase-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:e56051b66b0f419874d113c0320232dba7dd2547357256cf80eeaf18bbecb298
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hbase-operator@sha256:e56051b66b0f419874d113c0320232dba7dd2547357256cf80eeaf18bbecb298
        version: 24.11.1
      entries:
      - name: hbase-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: hbase-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:232ff30ffdf4bdc33f49d87950864e15436957510ba47f352dc0b133cb43279d
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hbase-operator@sha256:232ff30ffdf4bdc33f49d87950864e15436957510ba47f352dc0b133cb43279d
        version: 25.3.0
      entries:
      - name: hbase-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: hbase-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
        version: 25.7.0
      entries:
      - name: hbase-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hbase-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
        version: 25.7.0
      entries:
      - name: hbase-operator.v25.7.0
        version: 25.7.0
      - name: hbase-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hbase-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: cinder-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cinder-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cinder.openstack.org/v1beta1",
                "kind": "Cinder",
                "metadata": {
                  "name": "cinder",
                  "namespace": "openstack"
                },
                "spec": {
                  "cinderAPI": {},
                  "cinderBackup": {},
                  "cinderScheduler": {},
                  "cinderVolumes": {
                    "volume1": {}
                  },
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "cinder",
                  "databaseInstance": "openstack",
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "cinder-secret",
                  "serviceUser": "cinder"
                }
              },
              {
                "apiVersion": "cinder.openstack.org/v1beta1",
                "kind": "CinderAPI",
                "metadata": {
                  "name": "cinderapi-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cinder.openstack.org/v1beta1",
                "kind": "CinderBackup",
                "metadata": {
                  "name": "cinderbackup-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cinder.openstack.org/v1beta1",
                "kind": "CinderScheduler",
                "metadata": {
                  "name": "cinderscheduler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cinder.openstack.org/v1beta1",
                "kind": "CinderVolume",
                "metadata": {
                  "name": "cindervolume-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:43Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CinderAPI is the Schema for the cinderapis API
            displayName: Cinder API
            kind: CinderAPI
            name: cinderapis.cinder.openstack.org
            version: v1beta1
          - description: CinderBackup is the Schema for the cinderbackups API
            displayName: Cinder Backup
            kind: CinderBackup
            name: cinderbackups.cinder.openstack.org
            version: v1beta1
          - description: Cinder is the Schema for the cinders API
            displayName: Cinder
            kind: Cinder
            name: cinders.cinder.openstack.org
            version: v1beta1
          - description: CinderScheduler is the Schema for the cinderschedulers API
            displayName: Cinder Scheduler
            kind: CinderScheduler
            name: cinderschedulers.cinder.openstack.org
            version: v1beta1
          - description: CinderVolume is the Schema for the cindervolumes API
            displayName: Cinder Volume
            kind: CinderVolume
            name: cindervolumes.cinder.openstack.org
            version: v1beta1
        description: Cinder Operator
        displayName: Cinder Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Block Storage
        - Cinder
        links:
        - name: Cinder Operator
          url: https://github.com/openstack-k8s-operators/cinder-operator
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/cinder-rhel9-operator@sha256:e0e457500af9db0831c7080b10e010d11213cd800d89e40683974aaf7966a68d
        - registry.redhat.io/rhoso/openstack-cinder-api-rhel9@sha256:ec6835efb7ffc4fa7eed10bd029440263237e8f792e7e43d4cda337d00fecdb6
        - registry.redhat.io/rhoso/openstack-cinder-backup-rhel9@sha256:936e10d80300c325d0f58b4d6ca290fcf3c5cad774eb5cb696330a20c0151561
        - registry.redhat.io/rhoso/openstack-cinder-scheduler-rhel9@sha256:753775cc575321c727b9a6371daedc8774ba3a2287c96f3d44ad8f7025f63841
        - registry.redhat.io/rhoso/openstack-cinder-volume-rhel9@sha256:47b19d993bdfc790c508861dbcc87f0ec51302db8975698af8e1476a0f5209c7
        version: 1.0.6
      entries:
      - name: cinder-operator.v1.0.6
        version: 1.0.6
      - name: cinder-operator.v1.0.5
        version: 1.0.5
      - name: cinder-operator.v1.0.4
        version: 1.0.4
      - name: cinder-operator.v1.0.3
        version: 1.0.3
      - name: cinder-operator.v1.0.2
        version: 1.0.2
      - name: cinder-operator.v1.0.1
        version: 1.0.1
      - name: cinder-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: cinder-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: node-healthcheck-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-healthcheck-operator.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:60d95ea8240ecf701ba974b6df3ee296a7fb3ef5393e0dcf43af15e782f2d153
          createdAt: "2023-11-02 11:44:50"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.3.0 <0.4.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://docs.openshift.com/container-platform/4.12/nodes/nodes/eco-node-health-check-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:60d95ea8240ecf701ba974b6df3ee296a7fb3ef5393e0dcf43af15e782f2d153
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:8d32dee5c72ec1c94e4e92f8d1da935a384e40e379f237c432086a2781f4ecb9
        version: 0.4.1
      entries:
      - name: node-healthcheck-operator.v0.4.1
        version: 0.4.1
      name: 4.12-eus
    - currentCSV: node-healthcheck-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:90cceb176fa903cc8e9aa4c0f7316560a8a354a3138dc26abbfc9884c1c576b5
          createdAt: "2023-10-30 14:21:41"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.4.0 <0.6.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/node-health-check-operator
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:90cceb176fa903cc8e9aa4c0f7316560a8a354a3138dc26abbfc9884c1c576b5
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:b8d5db068bfc81031c2eef86e8649e4fc008b7480e08c9cfab2ab4497098eecf
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel8@sha256:85054b34e119c16cc95cda9dc3f5d63f2d7bfc70634f70973410b2a1b1fa58d4
        version: 0.6.1
      entries:
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.14-eus
    - currentCSV: node-healthcheck-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:4811046e26846b1f12a3b20ddc6209980897c11b171ddd61d26b43630cc1d4b4
          createdAt: "2024-08-09T14:54:37Z"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.8.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          SNR is installed automatically when installing NHC.

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/node-health-check-operator
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:4811046e26846b1f12a3b20ddc6209980897c11b171ddd61d26b43630cc1d4b4
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:942f69efc7ae05d6c4d9c33335e4be71bfae2e6f9ef062dcd7887eebfe02120e
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel8@sha256:062cd9def39a087ae3802cd5208e2f0d5bfca29ba76ebd5139d88348e8c9e923
        version: 0.8.2
      entries:
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.16-eus
    - currentCSV: node-healthcheck-operator.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:f174ada653bb5fb7e58219510d0b10d868273a70db5c486e0b3ad05a1be50640
          createdAt: "2025-08-04 12:38:05"
          description: Detect failed Nodes and trigger remediation with a remediation
            operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <0.9.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.7/html/remediation_fencing_and_maintenance
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel9@sha256:8e9926558c3d8082f298e9e5f58bc5b473020acd44c2ce853c2916cdad71d649
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:f174ada653bb5fb7e58219510d0b10d868273a70db5c486e0b3ad05a1be50640
        - registry.redhat.io/workload-availability/node-remediation-console-rhel9@sha256:37f62407284fe574b973a1a685371d191b710ad70a6bfe8172f19a4568c11cf6
        version: 0.9.1
      entries:
      - name: node-healthcheck-operator.v0.9.1
        version: 0.9.1
      - name: node-healthcheck-operator.v0.9.0
        version: 0.9.0
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.18-eus
    - currentCSV: node-healthcheck-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:8e1dbc35126101e1025bd4c0eac4f4c614eb779ccafbb55050541a27ba7e9706
          createdAt: "2022-08-15 12:34:34"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.2.0 <0.3.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-node-health-check-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b5786bbbef725badf3dfcc2c2c7a86ead5ebb584c978c47aae8b9a62e241b80d
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:8e1dbc35126101e1025bd4c0eac4f4c614eb779ccafbb55050541a27ba7e9706
        version: 0.3.1
      entries:
      - name: node-healthcheck-operator.v0.3.1
        version: 0.3.1
      name: candidate
    - currentCSV: node-healthcheck-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:2160f2688e5b098c0bae7597cebcd380019b82ae02f4506d07c2ed85f0b3664d
          createdAt: "2025-09-01 18:07:56"
          description: Detect failed Nodes and trigger remediation with a remediation
            operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.8/html/remediation_fencing_and_maintenance
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:2160f2688e5b098c0bae7597cebcd380019b82ae02f4506d07c2ed85f0b3664d
        - registry.redhat.io/workload-availability/node-remediation-console-rhel9@sha256:cc0f671b126cde10476f8cc2061c6903932e7bf5e3dbf4e8bdda9add70d5deea
        version: 0.10.0
      entries:
      - name: node-healthcheck-operator.v0.10.0
        version: 0.10.0
      - name: node-healthcheck-operator.v0.9.1
        version: 0.9.1
      - name: node-healthcheck-operator.v0.9.0
        version: 0.9.0
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: stable
    defaultChannel: stable
    packageName: node-healthcheck-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-mq-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-mq-controller.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mq.services.k8s.aws/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/mq-controller:1.1.2
          createdAt: "2025-10-03T18:49:51Z"
          description: AWS MQ controller is a service controller for managing MQ resources
            in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker represents the state of an AWS mq Broker resource.
            displayName: Broker
            kind: Broker
            name: brokers.mq.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon MQ resources in AWS from within your Kubernetes cluster.

          **About Amazon MQ**

          Amazon MQ is a managed message broker service for Apache ActiveMQ and RabbitMQ that makes it easy to set up and operate message brokers on AWS. Amazon MQ reduces your operational responsibilities by managing the provisioning, setup, and maintenance of message brokers for you. Because Amazon MQ connects to your current applications with industry-standard APIs and protocols, you can easily migrate to AWS without having to rewrite code.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon MQ
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mq
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon MQ Developer Resources
          url: https://docs.aws.amazon.com/amazon-mq/latest/developer-guide/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: mq maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/mq-controller:1.1.2
        version: 1.1.2
      entries:
      - name: ack-mq-controller.v1.1.2
        version: 1.1.2
      - name: ack-mq-controller.v1.1.1
        version: 1.1.1
      - name: ack-mq-controller.v1.1.0
        version: 1.1.0
      - name: ack-mq-controller.v1.0.13
        version: 1.0.13
      - name: ack-mq-controller.v1.0.11
        version: 1.0.11
      - name: ack-mq-controller.v1.0.10
        version: 1.0.10
      - name: ack-mq-controller.v1.0.9
        version: 1.0.9
      - name: ack-mq-controller.v1.0.8
        version: 1.0.8
      - name: ack-mq-controller.v1.0.7
        version: 1.0.7
      - name: ack-mq-controller.v1.0.6
        version: 1.0.6
      - name: ack-mq-controller.v1.0.5
        version: 1.0.5
      - name: ack-mq-controller.v1.0.2
        version: 1.0.2
      - name: ack-mq-controller.v1.0.1
        version: 1.0.1
      - name: ack-mq-controller.v1.0.0
        version: 1.0.0
      - name: ack-mq-controller.v0.0.36
        version: 0.0.36
      - name: ack-mq-controller.v0.0.35
        version: 0.0.35
      - name: ack-mq-controller.v0.0.34
        version: 0.0.34
      - name: ack-mq-controller.v0.0.31
        version: 0.0.31
      - name: ack-mq-controller.v0.0.29
        version: 0.0.29
      - name: ack-mq-controller.v0.0.27
        version: 0.0.27
      - name: ack-mq-controller.v0.0.26
        version: 0.0.26
      - name: ack-mq-controller.v0.0.25
        version: 0.0.25
      - name: ack-mq-controller.v0.0.24
        version: 0.0.24
      - name: ack-mq-controller.v0.0.23
        version: 0.0.23
      - name: ack-mq-controller.v0.0.22
        version: 0.0.22
      - name: ack-mq-controller.v0.0.21
        version: 0.0.21
      - name: ack-mq-controller.v0.0.20
        version: 0.0.20
      - name: ack-mq-controller.v0.0.19
        version: 0.0.19
      - name: ack-mq-controller.v0.0.18
        version: 0.0.18
      - name: ack-mq-controller.v0.0.17
        version: 0.0.17
      - name: ack-mq-controller.v0.0.16
        version: 0.0.16
      - name: ack-mq-controller.v0.0.15
        version: 0.0.15
      - name: ack-mq-controller.v0.0.14
        version: 0.0.14
      - name: ack-mq-controller.v0.0.13
        version: 0.0.13
      - name: ack-mq-controller.v0.0.12
        version: 0.0.12
      - name: ack-mq-controller.v0.0.11
        version: 0.0.11
      - name: ack-mq-controller.v0.0.10
        version: 0.0.10
      - name: ack-mq-controller.v0.0.9
        version: 0.0.9
      - name: ack-mq-controller.v0.0.8
        version: 0.0.8
      - name: ack-mq-controller.v0.0.7
        version: 0.0.7
      name: alpha
    defaultChannel: alpha
    packageName: ack-mq-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Accuknox Inc.
      provider-url: https://accuknox.com
    name: accuknox-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: accuknox-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.accuknox.com/v1",
                "kind": "AccuknoxConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "accuknox-operator",
                    "app.kubernetes.io/instance": "accuknoxconfig",
                    "app.kubernetes.io/name": "accuknoxconfig-sample",
                    "app.kubernetes.io/part-of": "accuknox-operator"
                  },
                  "name": "accuknoxconfig-sample",
                  "namespace": "accuknox-agents"
                },
                "spec": {
                  "ppsHost": "PLEASE_FILL_IN",
                  "knoxGateway": "PLEASE_FILL_IN",
                  "spireHost": "PLEASE_FILL_IN"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Security
          containerImage: public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:b6fae25e229fb7fca757872fc0460642d52c4f9906f776afac50c54a69b05b68
          createdAt: "2024-02-09T17:06:30Z"
          operatorframework.io/suggested-namespace: accuknox-agents
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: public.ecr.aws/k9v9d5v2/accuknox-operator
          support: Accuknox Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccuknoxConfig is the Schema for the accuknoxconfigs API
            kind: AccuknoxConfig
            name: accuknoxconfigs.operator.accuknox.com
            version: v1
        description: |2-

          The AccuKnox Operator brings native K8s support for deploying and managing AccuKnox Components on a Kubernetes cluster.
          The AccuKnox Operator helps to easily onboard a k8s cluster to the AccuKnox SaaS.

          AccuKnox Operator deploys and manages the components:

          1. Discovery Engine

          2. Policy Enforcement Agent

          3. Shared Informer Agent

          4. Feeder Service

          5. Agents Operator

          ## Prerequisites for using this Operator

          1. Get AccuKnox licence - https://app.accuknox.com/sign-up

          2. After Successful Sign-up into AccuKnox Application.

             Please naviagate to Settings >> Manager Cluster >> Onboard Cluster >>

             After successful onboarding, Copy the join token created by AccuKnox

          3. Create a secret using jointoken by executing the below command:

          ```
          oc create secret generic accuknox-jointoken --from-literal=token=<joinToken> -n accuknox-agents
          ```

          ## Deploy AccuKnox Operator using AccuknoxConfig Custom Resource

          All the information to configure the AccuKnox Operator will be provided via Accuknox during Cluster Onboarding steps.

          ### AccuknoxConfig Spec

          | Spec | Description |
          | ---- | ----------- |
          | spireHost | "spire host url" |
          | ppsHost   |  "pps host url" |
          | knoxGateway | "knox gateway url" |
          | joinTokenSecretName | "join token secret name if it's different than accuknox-jointoken" |
          | joinTokenSecretKey | "key name used to store the jointoken with the secret" |

          *Example*

          ```
          apiVersion: operator.accuknox.com/v1
          kind: AccuknoxConfig
          metadata:
            name: accuknoxconfig-default
            namespace: accuknox-agents
          spec:
            ppsHost: <ppsHost url>
            knoxGateway: <knox gateway url>
            spireHost: <spire host url>
          ```
        displayName: Accuknox Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CNAPP
        - CSPM
        - CWPP
        - Cloud Native
        - Runtime Security
        - Zero-Trust
        maturity: alpha
        provider:
          name: Accuknox Inc.
          url: https://accuknox.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        - public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:b6fae25e229fb7fca757872fc0460642d52c4f9906f776afac50c54a69b05b68
        - public.ecr.aws/k9v9d5v2/feeder-service@sha256:f5bc340d9d3e34804dd54a03701d272e0ce5c04160d1aa41984cb3d877eb5b76
        - public.ecr.aws/k9v9d5v2/discovery-engine-discover@sha256:052efbe1064503853adab497415b9407f51578e1456931a8a0e60353aa53e0c2
        - public.ecr.aws/k9v9d5v2/discovery-engine-hardening@sha256:3eb28d44ef93e6ba331e9085e64475458e1c234017c77dc30b8bc802def434c3
        - public.ecr.aws/k9v9d5v2/agents-operator@sha256:656d6bac4371adef3e86658011ba68ae2f98514f6ddaea0667d4269eb88dd02e
        - public.ecr.aws/k9v9d5v2/shared-informer-agent@sha256:6d18d13312d4309b6e32b18c063d5b26f15fbf057fda2b9e6b7e8ebb1a5885e9
        - public.ecr.aws/k9v9d5v2/policy-enforcement-agent@sha256:f004fc3572a4d66da1bc0db9e05a279d761a42d6ed62f95de7a2bb42ce71e17b
        - public.ecr.aws/k9v9d5v2/discovery-engine-sumengine@sha256:4ca78f90b28b38badb0520f148ce702a3a6cd6b02bb62e8b1a6782a5b82e0868
        - public.ecr.aws/k9v9d5v2/discovery-engine-offloader@sha256:3c2291310e30e35a931d2af389f7ade656ce86f609d2e42ecdabf1000764695c
        version: 0.2.1
      entries:
      - name: accuknox-operator.v0.2.1
        version: 0.2.1
      name: alpha
    - currentCSV: accuknox-operator.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.accuknox.com/v1",
                "kind": "AccuknoxConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "accuknox-operator",
                    "app.kubernetes.io/instance": "accuknoxconfig",
                    "app.kubernetes.io/name": "accuknoxconfig-sample",
                    "app.kubernetes.io/part-of": "accuknox-operator"
                  },
                  "name": "accuknoxconfig-sample",
                  "namespace": "accuknox-agents"
                },
                "spec": {
                  "ppsHost": "PLEASE_FILL_IN",
                  "knoxGateway": "PLEASE_FILL_IN",
                  "spireHost": "PLEASE_FILL_IN"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Security
          containerImage: public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:a935ad806e82760bb838bada30d3ed1a54c8c61b95fe5c6109f48985c6faf4d7
          createdAt: "2024-08-23T04:25:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: accuknox-agents
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: docker.io/accuknox/accuknox-operator
          support: Accuknox Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccuknoxConfig is the Schema for the accuknoxconfigs API
            kind: AccuknoxConfig
            name: accuknoxconfigs.operator.accuknox.com
            version: v1
        description: |2-

          The AccuKnox Operator brings native K8s support for deploying and managing AccuKnox Components on a Kubernetes cluster.
          The AccuKnox Operator helps to easily onboard a k8s cluster to the AccuKnox SaaS.

          AccuKnox Operator deploys and manages the components:

          1. Discovery Engine

          2. Policy Enforcement Agent

          3. Shared Informer Agent

          4. Feeder Service

          5. Agents Operator

          ## Prerequisites for using this Operator

          1. Get AccuKnox licence - https://app.accuknox.com/sign-up

          2. After Successful Sign-up into AccuKnox Application.

             Please naviagate to Settings >> Manager Cluster >> Onboard Cluster >>

             After successful onboarding, Copy the join token created by AccuKnox

          3. Create a secret using jointoken by executing the below command:

          ```
          oc create secret generic accuknox-jointoken --from-literal=token=<joinToken> -n accuknox-agents
          ```

          ## Deploy AccuKnox Operator using AccuknoxConfig Custom Resource

          All the information to configure the AccuKnox Operator will be provided via Accuknox during Cluster Onboarding steps.

          ### AccuknoxConfig Spec

          | Spec | Description |
          | ---- | ----------- |
          | spireHost | "spire host url" |
          | ppsHost   |  "pps host url" |
          | knoxGateway | "knox gateway url" |
          | joinTokenSecretName | "join token secret name if it's different than accuknox-jointoken" |
          | joinTokenSecretKey | "key name used to store the jointoken with the secret" |

          *Example*

          ```
          apiVersion: operator.accuknox.com/v1
          kind: AccuknoxConfig
          metadata:
            name: accuknoxconfig-default
            namespace: accuknox-agents
          spec:
            ppsHost: <ppsHost url>
            knoxGateway: <knox gateway url>
            spireHost: <spire host url>
          ```
        displayName: Accuknox Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CNAPP
        - CSPM
        - CWPP
        - Cloud Native
        - Runtime Security
        - Zero-Trust
        maturity: stable
        provider:
          name: Accuknox Inc.
          url: https://accuknox.com
        relatedImages:
        - docker.io/accuknox/discovery-engine-discover@sha256:e2d37bfbd0e9d23c129c143eb7c9aab1c3ed55e9571b44888eea2b842f8a2361
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        - public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:a935ad806e82760bb838bada30d3ed1a54c8c61b95fe5c6109f48985c6faf4d7
        - docker.io/accuknox/discovery-engine-hardening@sha256:e596aaaabaec94e0abb4f92ad9266dc010c0e88a89d600e8750a360099cae3cc
        - docker.io/accuknox/feeder-service@sha256:295e897b6f27ce61c45fb0179b734ef6d4c9e91864bcc5d2884a2d1dd4db81b6
        - docker.io/accuknox/policy-enforcement-agent@sha256:a0364343afeac4537589bfc247f64fffd7aac16f8afe07ce050bc58b2af248fb
        - docker.io/accuknox/discovery-engine-offloader@sha256:59cc594009de89a574cd569e701b6765d59a6899a03ad07eee1c025331df4331
        - docker.io/accuknox/shared-informer-agent@sha256:9f53d7dba9569487c42c6cf9e88988a1adb4c0971742e8b521700dce2d22b8cc
        - docker.io/accuknox/discovery-engine-sumengine@sha256:f7927128b86fdfb635377bd599df06f0eb9b9f0d1b068a0047468e480855443e
        - docker.io/accuknox/agents-operator@sha256:b0ae2728f15ebae904efb778cd29215ccf07ffca945a59deb5484f6c19c1b21f
        version: 0.7.1
      entries:
      - name: accuknox-operator.v0.7.1
        version: 0.7.1
      name: stable
    defaultChannel: stable
    packageName: accuknox-operator-certified
    provider:
      name: Accuknox Inc.
      url: https://accuknox.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pachyderm, Inc.
      provider-url: https://pachyderm.com
    name: pachyderm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pachyderm-operator.v0.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "Pachyderm",
                "metadata": {
                  "name": "pachyderm-sample"
                },
                "spec": {
                  "console": {
                    "disable": true
                  },
                  "pachd": {
                    "metrics": {
                      "disable": false
                    },
                    "storage": {
                      "amazon": {
                        "credentialSecretName": "pachyderm-aws-secret"
                      },
                      "backend": "AMAZON"
                    }
                  },
                  "version": "v2.1.6"
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermExport",
                "metadata": {
                  "name": "pachydermexport-sample"
                },
                "spec": {
                  "backup": {
                    "target": "pachyderm-sample"
                  }
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermImport",
                "metadata": {
                  "name": "pachydermimport-sample"
                },
                "spec": {
                  "backup": "backups/pachyderm-backup-20220610.tar.gz",
                  "destination": {
                    "name": "pachyderm-restore",
                    "namespace": "pachyderm-workloads"
                  },
                  "storageSecret": "s3-credentials-secret"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
          createdAt: "2021-05-20T08:05:00Z"
          description: Pachyderm provides the data layer that allows data science
            teams to productionize and scale their machine learning lifecycle with
            data driven automation, petabyte scalability and end-to-end reproducibility.
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pachyderm/openshift-operator
          support: Pachyderm, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PachydermExport is the Schema for the pachydermexports API
            displayName: Pachyderm Export
            kind: PachydermExport
            name: pachydermexports.aiml.pachyderm.com
            version: v1beta1
          - description: PachydermImport is the Schema for the pachydermimports API
            displayName: Pachyderm Import
            kind: PachydermImport
            name: pachydermimports.aiml.pachyderm.com
            version: v1beta1
          - description: Pachyderm is the Schema for the pachyderms API
            displayName: Pachyderm
            kind: Pachyderm
            name: pachyderms.aiml.pachyderm.com
            version: v1beta1
        description: "The Pachyderm operator aims to make deploying and managing Pachyderm
          instances on Openshift container platform easier.\n\nPachyderm is the data
          foundation for machine learning.  Pachyderm provides industry leading data
          versioning, pipelines and lineage that allow data science teams to automate
          the machine learning lifecycle and optimize their machine learning operations
          (MLOps). \n\n## Getting started\n\n1. Pachyderm uses an object store to
          store your data. Set the pachd storage backend at `pachyderm.spec.pachd.storage.backend`
          to `AMAZON`, `MICROSOFT`, `GOOGLE` or `MINIO` depending on your storage
          provider. Additionally, create a user with read/write access to the object
          store.\n\n2. Setup the user credentials in the section, `pachyderm.spec.pachd.storage`.
          Required fields depend on the backend type set in step 1 above.\n\n3. Finally,
          deploy the pachyderm cluster in its own namespace.\n\nFor more information
          on getting started, reference the documentation.\n"
        displayName: Pachyderm
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pachyderm
        - versioning
        - machine learning
        - artificial intelligence
        links:
        - name: Documentation
          url: https://docs.pachyderm.com/latest/
        maintainers:
        - email: kyle@pachyderm.io
          name: Kyle White
        - email: eochieng@redhat.com
          name: Edmund Ochieng
        maturity: beta
        minKubeVersion: v1.19.0
        provider:
          name: Pachyderm, Inc.
          url: https://pachyderm.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - registry.connect.redhat.com/pachyderm/backup-handler@sha256:e2ef6bd8191e8ad58c765c0db8f5bb21ea7123caed0c0f58fe681a024201306d
        - registry.connect.redhat.com/pachyderm/etcd@sha256:7f330b51874379b1d4b71d94e37e0ecfa52b86d628296136970688cf204f53b5
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/pgbouncer@sha256:c3b6949227df24abb8934d111a34b44e87fb16194ac34a5d0fe284cff02e9ac0
        - registry.connect.redhat.com/pachyderm/init-utils@sha256:5c0258bfa0203079925648c0a7dda6d05a29b6f461854821a1eec541777ff359
        - registry.connect.redhat.com/pachyderm/worker@sha256:374f83d92899707bc87b0318fda6bffa78d2e9e523b827bb3d40693a2d1bdecd
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
        - registry.redhat.io/rhel8/postgresql-13@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/pachd@sha256:116477fefdf413ea02c6ac3d013f1f91d3f6e4b16aed2408c9890832cdb41430
        version: 0.1.4
      entries:
      - name: pachyderm-operator.v0.1.4
        version: 0.1.4
      name: stable
    defaultChannel: stable
    packageName: pachyderm-operator
    provider:
      name: Pachyderm, Inc.
      url: https://pachyderm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-listener-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: listener-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/tools@sha256:47204b2697dfbab5eaf9d6dcb4c160dd427a148d6b840727239b773810e74fc9
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/tools@sha256:47204b2697dfbab5eaf9d6dcb4c160dd427a148d6b840727239b773810e74fc9
        - quay.io/stackable/listener-operator@sha256:b9c7957e7e562889bddd08e679abff05bb677672b3f735932f68514b77eacfd5
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:a6d6631d35c32b874c995b6af889636ac8dd1f39dcb68810d9f9b10a4e2c1987
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:161e754437428c68490333337303fb30b695a9ac677c3a00a1dc27565b855ab8
        version: 24.11.1
      entries:
      - name: listener-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: listener-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/listener-operator@sha256:da53970954b889bf4d874ae3df7e411b9abf41950ef7d50d6289831826f045a0
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/listener-operator@sha256:da53970954b889bf4d874ae3df7e411b9abf41950ef7d50d6289831826f045a0
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:71f8465e35a67f4b96e4c4904e67ccac9bbb9ae347426afb8bbfaabd6709e983
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        version: 25.3.0
      entries:
      - name: listener-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: listener-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
        version: 25.7.0
      entries:
      - name: listener-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: listener-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        version: 25.7.0
      entries:
      - name: listener-operator.v25.7.0
        version: 25.7.0
      - name: listener-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: "25.7"
    packageName: stackable-listener-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      pod-security.kubernetes.io/audit: privileged
      pod-security.kubernetes.io/enforce: privileged
      pod-security.kubernetes.io/warn: privileged
      provider: NVIDIA Corporation
      provider-url: ""
    name: gpu-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gpu-operator-certified.v25.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "kernelModuleType": "auto",
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": []
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
          createdAt: Thu Sep 18 18:11:36 PDT 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <25.3.4'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:43aa99352e1dc47e9218f36340ae6328a6282ed1e7e9a61b4204475e980230ef
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:c22dc67465c33d2f17932931229e61a6e7d9e32c5ed2ecefe8d10e873ffedfeb
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:51c8f71d3b3c08ae4eb4853697e3f8e6f11e435e666e08210178e6a1faf8028f
        - nvcr.io/nvidia/driver@sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735
        - nvcr.io/nvidia/driver@sha256:44422f10e899d3419fbd611a02fc1c6127e071a2067bd36e3debb043bf251de0
        - nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:66c3d68ca2b35d47b5562300a10ff8711b953ba3a15df7dbc409d81613ebe5d0
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b689dfb7806ae51b0072620b47c26114f05477e8ad9725892ee10366e5d93d8d
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:119de9a331a47203858b99901f44d0c4a8052961b4e60327f4b100d0ab8c9df0
        - nvcr.io/nvidia/k8s-device-plugin@sha256:3c54348fe5a57e5700e7d8068e7531d2ef2d5f3ccb70c8f6bac0953432527abd
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:5c4e61f7ba83d7a64ff2523d447c209ce5bde1ddc79acaf1f32f19620b4912d6
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:6bb786c41fa7b1a6238c023566fdeaeeea3ba842a8daec015a7bd85a8037b249
        - nvcr.io/nvidia/cuda@sha256:d19fe621624c4eb6ac931b8558daa3ecc0c3f07f1e2a52e0267e083d22dceade
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3191cc0684e92d655e72fe9faf3da95fe6b8a6c2eb4ad396f75ba9b98c021aaf
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:619e7bf48168d76a7e087c3bf6190fda9b10449c9839309fa4f9ed5b8a9e8804
        version: 25.3.4
      entries:
      - name: gpu-operator-certified.v25.3.4
        version: 25.3.4
      - name: gpu-operator-certified.v25.3.3
        version: 25.3.3
      - name: gpu-operator-certified.v25.3.2
        version: 25.3.2
      - name: gpu-operator-certified.v25.3.1
        version: 25.3.1
      - name: gpu-operator-certified.v25.3.0
        version: 25.3.0
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: stable
    - currentCSV: gpu-operator-certified.v1.10.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "dcgmExporter": {
                    "config": {
                      "name": ""
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                  },
                  "devicePlugin": {
                  },
                  "driver": {
                    "enabled": true,
                    "use_ocp_driver_toolkit": true,
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "gfd": {
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "operator": {
                    "defaultRuntime": "crio",
                    "deployGFD": true,
                    "initContainer": {
                    }
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:c7f9074c1a7f58947c807f23f2eece3a8b04e11175127919156f8e864821d45a
          createdAt: Mon Apr 18 10:06:40 PDT 2022
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=v1.9.0 <v1.10.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cuda@sha256:7f0f7dc8c993a86865d4d9f9b0e36d113ea15603d5d5279d2879fa2bc21c8236
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:5b16056257acc51b517d9cdb1da3218693cefc214af93789e6e214fd2b4cacf1
        - nvcr.io/nvidia/k8s-device-plugin@sha256:aa95c16106280e36a5e32d2fe4c66e8b70f5a114860c6f4ed5b1a4085c63601b
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:e8f06ce53415e8572f633b2169289d973993cb1e49d4dcc686ac134194d88f33
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:24d804e8f005d7aeca8343aa13e5f92295d8642a4c47cb24e3ac86a22543bc37
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:67c48a8370c34f68b46eb57339e87d01c04198368a0423553fab7565c862c4df
        - nvcr.io/nvidia/driver@sha256:f99b66855e92451a59a722355dd52bf97eb985cc8a647172273de8178aeaa93c
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:c2e3d0fe41a0d227dbb70caec03c780cc76317515e5ab3875f31d50c63f41c66
        - nvcr.io/nvidia/gpu-operator@sha256:c7f9074c1a7f58947c807f23f2eece3a8b04e11175127919156f8e864821d45a
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:ac13256c295126b854a07421c45de0721e7a57974b38b3e66eb1fdc3a4e29f2e
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:9abb063e6b74085728fead5ed5bbe3bd7b3f57de60e87f6b964bba2d46f1bc67
        version: 1.10.1
      entries:
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v1.10
    - currentCSV: gpu-operator-certified.v1.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    },
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "dcgmExporter": {
                    "config": {
                      "name": ""
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                  },
                  "devicePlugin": {
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": true
                  },
                  "vgpuDeviceManager": {
                    "enabled": true,
                    "config": {
                      "name": "vgpu-devices-config",
                      "default": "default"
                      }
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:53180f81fb7e1e4bcd49a556bc6285c2262abbab8083b29d2688b02d4200dd2d
          createdAt: Thu Aug  4 11:57:56 PDT 2022
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=v1.9.0 <v1.11.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:6e2902d73c54cf57bec94edf166d206cfeb0b948b5aab70800ccbe025064b3a1
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:b90a519e9f30fbc013e1806bc081268724e9e80b4035c67c167ab3b43a37704a
        - nvcr.io/nvidia/gpu-operator@sha256:53180f81fb7e1e4bcd49a556bc6285c2262abbab8083b29d2688b02d4200dd2d
        - nvcr.io/nvidia/driver@sha256:cf8ce04907d0e8decf90a8dccfa6134af9c39509d2f8a21c355ed71223acacdb
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:39127d04c9faff3f22068dcf463ed48d02be684cd18e3f14db843bddbdd783a7
        - nvcr.io/nvidia/driver@sha256:b8799e7397fae5a27446737443091eed05b918816f53c400a48be71125ee22e8
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3804e2b03c44d25dd410e3882b7fb6d1fde41f55110d7702af47136b042a63ff
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:10d825ccbfcf87f8c74f841fa89ae345f2d381fb47a5b2b31f61b72ebbbb0915
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:f48e8b430248a9cadf716a9211bb0050d4a663d1f1386ea107e72917f82e5b5f
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:8490d555131a57a11311561667ea9ec8cf554a399d2492a6f9e6677c01ec4b9a
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7a0ccbcbe8d379bfc441e06504eb45d0d433a37e2a9e6dde83d13347296ebe81
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:e88e610d09f1a7528ab32cb54259aee340eba2b2bf66b92a28157626daadc82d
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:2d6e86db0e2b0db64c00566938e1851e25d4bb1019b1e7b696e2b199cba33397
        - nvcr.io/nvidia/driver@sha256:8c9fabc1cd773e9526eb7e8526084f065f6cca9d0fd8dc2a672a0bb717f1cc60
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:16ee5b1a8d9217f761c40a06c7ae598cbc893fbc35d99ff9b84fa886ea6b9b37
        - nvcr.io/nvidia/cuda@sha256:4400b56f81095b24fc9dde93ad7ad7d0ca6a7e9736c4c7f1d4e2bcbf0afbe45c
        version: 1.11.1
      entries:
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v1.11
    - currentCSV: gpu-operator-certified.v22.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:54d5846c0291db621931a8544c9cb6437512470a93278ca9130ba73f2e2f0055
          createdAt: Mon Jan 30 15:43:24 PST 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <22.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:c3cf59dd5d6160eba5d816ade2e81b35ebb10f4884df67971f6ace36f8e6efc1
        - nvcr.io/nvidia/driver@sha256:f250d1e4989fd4821e81bad9227eeeaa531ca8810f3a40e996f1ca2af7945f96
        - nvcr.io/nvidia/k8s-device-plugin@sha256:9c17d3a907eb77eb8f7b4f3faf52d8352e4252af92003f828083f80d629bd2c3
        - nvcr.io/nvidia/driver@sha256:b860ef04cc6cd8c481810fa19bef7221071f5a3862640b2ad6c2704728dc77c8
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:0d47dad29d2ef445b301c5c64717758eed43a606345b79f97bce2e64b40a91a8
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:64d757b4c80b910e64647a84a1d592fab2ea3313ff6dce30c25c3a08e180bd74
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:efb88937f73434994d1bbadc87b492a1df047aa9f8d6e9f5ec3b09536e6e7691
        - nvcr.io/nvidia/driver@sha256:f2ca4612f3418e0e3963be8c3771ffe463cdc1fc1b810437a3aad9d76bb6499d
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:bec9f026d9b3d9404c78d6091817a359015c6a7aa411735b34138c1518853b5d
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:00f1476548fbed9ee01961443a73bf65396c2e8bb2b84426f949dd56cb4d14cd
        - nvcr.io/nvidia/gpu-operator@sha256:54d5846c0291db621931a8544c9cb6437512470a93278ca9130ba73f2e2f0055
        - nvcr.io/nvidia/driver@sha256:c5a62bc2a6cce339798e4b7cf9cec47e289c2d30b40f451ba8c9e958a68e3e55
        - nvcr.io/nvidia/driver@sha256:268c26781c46b36fe691638e807eb5c3a2cd077029ddb095ac75264adac04925
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:9177a0ae30798b42d0387f6a20cd3ce3cd1799a91b7866bf812368764b05b1af
        - nvcr.io/nvidia/cuda@sha256:5f2a2d8977f2c59abe88394f05cc3c044194554f90994d8554b0e1117ab5900d
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:9a00cdfdddb73327ef8e8e0fa60e50926a388d380c551bba6a3d3012be40401d
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:50c2a2b998e467c82716efa79fe136aa6f7ea95fd23576cf384d251bb9628640
        version: 22.9.2
      entries:
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v22.9
    - currentCSV: gpu-operator-certified.v23.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:08ca33f491195b7ba867be1e6642b5ff8aff6a4b0b7cfc0ec87619252ccfddc7
          createdAt: Mon May 22 15:22:15 PDT 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <23.3.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:8ff0ee88fe6c4ff031ef642593f93625bfd9a48fe7be8a79a5c2736117e36941
        - nvcr.io/nvidia/k8s-device-plugin@sha256:80871275fbcf888d4683c58638c7a75b3c46091040cafc6210a8c93513adb146
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:03e43d6b1460cadb33509d5ee07a5e27603ab89d7ec35f67d8a51bb78745b463
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:cab21c93987a5c884075efe0fb4a8abaa1997e1696cbc773ba69889f42f8329b
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:84ce86490d0d313ed6517f2ac3a271e1179d7478d86c772da3846727d7feddc3
        - nvcr.io/nvidia/cuda@sha256:62d845b4bd6de4be65f0b6daaae6831c29d4529955f99db5d26f5f7179106285
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:5489125ceae5864280e4d6a9ab52ab0f650b3179349a7298c4a204feb60b661a
        - nvcr.io/nvidia/gpu-operator@sha256:08ca33f491195b7ba867be1e6642b5ff8aff6a4b0b7cfc0ec87619252ccfddc7
        - nvcr.io/nvidia/driver@sha256:73b8928a1b1fb650d2ca8d6451a7d7c551fd0509f7511c2fd5bfdcef83ae4f5d
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:2d7e32e3d30c2415b4eb0b48ff4ce5a4ccabaf69ede0486305ed51d26cab7713
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:21dfc9c56b5f8bce73e60361d6e83759c3fa14dc6afc2d5ebdf1b891a936daf6
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:0d47dad29d2ef445b301c5c64717758eed43a606345b79f97bce2e64b40a91a8
        - nvcr.io/nvidia/driver@sha256:917f92c589c3124dba54f66f89c44b93b0beaf418f01d0da6a546e1670fdf3a9
        - nvcr.io/nvidia/driver@sha256:3df118f2330e95db3c76173cca35e542b2da120ec0bb8f6aff83c4c96c0c2daf
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:8f03f74076b5aa6672a5d75a398fa7515fcde57fdb1cf8079e015b3366efd3b5
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:f10170b076badda8f669af8b0f7b47eb5470abaf35a67dda9c4e182d66138177
        version: 23.3.2
      entries:
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.3
    - currentCSV: gpu-operator-certified.v23.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:d2e0f8bd86e8e9f2ff3514e113657658584ca4a2d0b82dac2e47364cc0b30648
          createdAt: Thu Aug 31 12:22:44 PDT 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <23.6.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:878a6162a4fa0b61dae02c7cfc13ba4d28bb06f92fd62d0c521633ab507d675a
        - nvcr.io/nvidia/driver@sha256:bbdc4415fe2e71eeea1b8a6e50ee562e11704dabb46db1db4aea807102886a35
        - nvcr.io/nvidia/k8s-device-plugin@sha256:f1617c3fe54fd9f6d4a00358e5da1a9fa9e8963a4ef6b3ed06f47d4971ddff4b
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:b30f5c64791dc5406b57eaf51ac40195dc1051a18cd1f38fd0a43c65f1d6dba9
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:a229beba8540549c234a046e975bf4bf8c50d521dfe551af56f6e042b25ec017
        - nvcr.io/nvidia/gpu-operator@sha256:d2e0f8bd86e8e9f2ff3514e113657658584ca4a2d0b82dac2e47364cc0b30648
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:9484110986c80ab83bc404066ca4b7be115124ec04ca16bce775403e92bfd890
        - nvcr.io/nvidia/driver@sha256:b4517810f2884f345b4f2323e83e41d28f54cdc6e2caf0690bb7ed4f3947dbcc
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:bb845160b32fd12eb3fae3e830d2e6a7780bc7405e0d8c5b816242d48be9daa8
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:60086c33226bbe7f04b611c20d308fed5825226f8a0a5635bc58d8ed55cc08a5
        - nvcr.io/nvidia/cuda@sha256:64c0ff5fd4a27d4b1d46378634f7a7398c75f87c18c70c988dde18910d461560
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:00a91252be8e62da6f28cdf1ec3721f2e945ae296938b0b925ef9db2bf6dcc8b
        - nvcr.io/nvidia/driver@sha256:3382e254056f28831767bc6729bc2594353a5ff2a28fe9f2d94396c597bb23d8
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:0379c59b43e11861837787a0da22e70b528718373bb10de9e8b7752849bd4e7a
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:bf06c8ad380a36e5014607bd3ab05b80e8e5e7e77d23d0dd89fe53a5564152fa
        version: 23.6.1
      entries:
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.6
    - currentCSV: gpu-operator-certified.v23.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:7481a3697783dcdca9ae78e7b548a6900e86ea33ab49ec14f0ba55db2fdb1a2e",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:3988b3b570cf65b0d338cb29bd24b0b9bf1885e1563fbb398ef23cba0b59e6bb
          createdAt: Wed Mar 6 16:43:36 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <23.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:6a05d6a1923fda756aed0dddf7ed23a83c30cf1e6c519fc39dd70c0309ec8257
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:a67d8e92861a2dce5649105c07561e4422e9fe4ba81a6525dc0d70a7ef85f9c0
        - nvcr.io/nvidia/driver@sha256:56c79482582cdfbc58d3134e8672637c5bf05f328880f76898f526143d04c6af
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:8d1ec78f2b1ddb7f0c47453d0427231190747bda411733a7dd0c8f5196f09e9c
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:27c44f4720a4abf780217bd5e7903e4a008ebdbcf71238c4f106a0c22654776c
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:925f4a47710e4318ed457930f5406174c1f6d28b1bf6b1bc310687fec0fde712
        - nvcr.io/nvidia/driver@sha256:abda1ac56371d55917b96ff330109980f468e133c9d5705da0ef87429f14ccd7
        - nvcr.io/nvidia/k8s-device-plugin@sha256:bb81ae8d034d1e1c6a342091a5f1dda5884dd43ab47e96624487a133d1edd4b3
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:387021553dfb16aab633228d42f63f04fa932b4f46add07527f296dfe97e5148
        - nvcr.io/nvidia/driver@sha256:7481a3697783dcdca9ae78e7b548a6900e86ea33ab49ec14f0ba55db2fdb1a2e
        - nvcr.io/nvidia/gpu-operator@sha256:3988b3b570cf65b0d338cb29bd24b0b9bf1885e1563fbb398ef23cba0b59e6bb
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:011fb450af3fa2e8fe5d28d590e4c653631447bc23d149591ced3d89089c4f2c
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:9aefef081c3ab1123556374d2b15d0429f3990af2fbaccc3c9827801e1042703
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:59a3875e7a37eb370385e654184efa3a1b193c9ea352165818496b19cbe14aa4
        - nvcr.io/nvidia/cuda@sha256:714547d54e5fe4191019a1e5f1daffc7fab7481b619b79c378541dafc76c9e5d
        version: 23.9.2
      entries:
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.9
    - currentCSV: gpu-operator-certified.v24.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:96f25c67e5b1072d5981080e12d65ec37eb9ef2fc0494499416aa801b0a34da3",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:041e75a3df84039c2dbbd4b9d67763bd212138822dbb6dbc0008858c1c6eff8d
          createdAt: Tue Apr 30 14:18:26 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.3.0'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:ecb867c5787e44f1889b257e6c03b07748689a7b293c8f1affb1d4cb449b235c
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:f0e4e14a3081417be8e6a5d855fb0cb69e1b63bc54a74f17e6d7084abe275588
        - nvcr.io/nvidia/driver@sha256:56c79482582cdfbc58d3134e8672637c5bf05f328880f76898f526143d04c6af
        - nvcr.io/nvidia/driver@sha256:a836ccbe21da735aee1c39b81060ed5e2fdb4ffa339874baaf4634f1e9259f74
        - nvcr.io/nvidia/driver@sha256:96f25c67e5b1072d5981080e12d65ec37eb9ef2fc0494499416aa801b0a34da3
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:2edc1d4ed555830e70010c82558936198f5faa86fc29ecf5698219145102cfcc
        - nvcr.io/nvidia/k8s-device-plugin@sha256:80674b19898ecf9ed6949e39674da769d6feeeb01bea54b914ef9ff502834f49
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:d8774d6afd37110199512636172823ae8749e5ea2e13760f57c255a74f47018c
        - nvcr.io/nvidia/cuda@sha256:ae0623ec8634b6c88f815b88037763def160cbbac15013b77ddef257fc276c9a
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:1ae0b923bc57f47bab046b50c50110f6914bbaffbfef704df34b3fe332db2e31
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:0c77725de8c42d248ed825453efd2e005f4900f4be384fd23084f6b721ddd0e0
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:7798eb9b9424e09959fc808596f87bdb68ac5740174123c6646dee2166d3e7d2
        - nvcr.io/nvidia/gpu-operator@sha256:041e75a3df84039c2dbbd4b9d67763bd212138822dbb6dbc0008858c1c6eff8d
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:2cc2119d20515f8ca7a61a0f3932578d69a45bcdea49e6c320a89c56f105e7d9
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:2442fd2ec0ee29746abbced727c53741f017e8f3f615321c1ec42c1c3ffa0b12
        version: 24.3.0
      entries:
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.3
    - currentCSV: gpu-operator-certified.v24.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:858de27c152669f5a3cf4287406405b16dd5bb70c0373324eb735511997bb415",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:8e0969cffc030a89c4acd68e64d41dd54e3bce8a794106b178d4dbd636a07f1c
          createdAt: Wed Sep 25 08:25:16 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.6.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:21f4c8b88716e8e6f732f9fb4c2efaef937c227491a8631c5e55036f80f39a4d
        - nvcr.io/nvidia/driver@sha256:858de27c152669f5a3cf4287406405b16dd5bb70c0373324eb735511997bb415
        - nvcr.io/nvidia/gpu-operator@sha256:8e0969cffc030a89c4acd68e64d41dd54e3bce8a794106b178d4dbd636a07f1c
        - nvcr.io/nvidia/cuda@sha256:de5b598bca89850567c4c104411d66bb52f47c9179199e6a3be6829b7ac586a2
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:740abc3ff657545c10effd5354f09af525200ed9a1b7623f0c2e8c7bd9e4a4e2
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:96380b95396b7f29d2ed2ec8ef1ad5a8bcf9f55051db723295dc0a20db845331
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:ae63fac9a4057a7646f0cf0ee0566e8928529adde05c4c0a017cda0599e381b2
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:33de74eb590f071403c17b6c210c02963245851971168bc0c07c06c100a9f376
        - nvcr.io/nvidia/driver@sha256:07e11f85d54d49ec9648fb06e148b8d832ee1f9c3549a915eee853c9ef2949c2
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:781fb47e264d9e0fbc8da5bd046e5e678316c866bc36ddd4b56d4eb0de682d5b
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:969147c01d63be5d1fe458f32f1cc0c7408cf3062531db91408e2fc57b4d8a67
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:15dab1273345df4a5844c4c761d064dbc4b592101251dc39174e597137123027
        - nvcr.io/nvidia/driver@sha256:a6d12fb5753f267dda25dfd38910f972bc632c006a24107fa50e20bba3642d7c
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7ad2c9f71fe06f9f7745ac8635f46740fbdff4f11edd468addfab81afcdfa534
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:7bcc188703f2fac630f7a4ff8960e6733ac3a29adf6a946533b796d9a27b8acf
        version: 24.6.2
      entries:
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.6
    - currentCSV: gpu-operator-certified.v24.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:6fe74322562c726c8fade184d8c45ebae3da7b1ea0a21f0ff9dc42c66c65e692",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:92f5262bb4934e0ea29eb10ec0ad4fc293164f2f2f94d48a935922791abbfdac
          createdAt: Mon Jan 27 16:10:52 PST 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cuda@sha256:b3b340847db9abd0c39703278f5638d3e65e1369e9a526a19b3138d5a6563527
        - nvcr.io/nvidia/gpu-operator@sha256:92f5262bb4934e0ea29eb10ec0ad4fc293164f2f2f94d48a935922791abbfdac
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:34eab99d7992a57c35803ec5e0afd538d16e997341e951fef4c4019afb08793e
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:e8398b2451996e5c64fc3855e7a3c93b667a85f69c0e259049d3b0012c21545d
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:d934033f9ff89087d555850eb1da150eaaa27b11e5ad21db1b874b8bc0b8d35b
        - nvcr.io/nvidia/driver@sha256:6fe74322562c726c8fade184d8c45ebae3da7b1ea0a21f0ff9dc42c66c65e692
        - nvcr.io/nvidia/driver@sha256:38b66a8d44cab9e2c62da9e101f32cd9dbcb5e02d8e57b47671284d374ca3695
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:4ffa1cd2a6497eb647a89ed259dcfb007554737b9d80f69bc173a2c3cd72a1da
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:40830d3561c14743f484d45b498141f9e86b1308e16fae3978110783927264ab
        - nvcr.io/nvidia/driver@sha256:026f8f2d29b7058ecaaa5b98666e28e1b3646b73e7d7cda3dc8026d47a929152
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:cf39d78ced7fb5727a9668ee2cd44b14bb7a23a95b83d5464b7d755740e02121
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7089559ce6153018806857f5049085bae15b3bf6f1c8bd19d8b12f707d087dea
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b072c5793be65eee556eaff1b9cbbd115a1ef29982be95b2959adfcb4bc72382
        - nvcr.io/nvidia/driver@sha256:d55b57938866e538acc3a71ca32f8cf87e71c591abd4a34695ee428e7ec2fa73
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:7edd7a0413dcb39b6e3bcefaf06812f3293c8e480ca10783e821a561ed686200
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:857f7669ae5ce2a43a3b5691fa970085098321ef8e90a896fa8dda3dee5b1d2b
        version: 24.9.2
      entries:
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.9
    - currentCSV: gpu-operator-certified.v25.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "kernelModuleType": "auto",
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": []
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
          createdAt: Thu Sep 18 18:11:36 PDT 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <25.3.4'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cuda@sha256:d19fe621624c4eb6ac931b8558daa3ecc0c3f07f1e2a52e0267e083d22dceade
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:619e7bf48168d76a7e087c3bf6190fda9b10449c9839309fa4f9ed5b8a9e8804
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:66c3d68ca2b35d47b5562300a10ff8711b953ba3a15df7dbc409d81613ebe5d0
        - nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:5c4e61f7ba83d7a64ff2523d447c209ce5bde1ddc79acaf1f32f19620b4912d6
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:51c8f71d3b3c08ae4eb4853697e3f8e6f11e435e666e08210178e6a1faf8028f
        - nvcr.io/nvidia/driver@sha256:44422f10e899d3419fbd611a02fc1c6127e071a2067bd36e3debb043bf251de0
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3191cc0684e92d655e72fe9faf3da95fe6b8a6c2eb4ad396f75ba9b98c021aaf
        - nvcr.io/nvidia/driver@sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735
        - nvcr.io/nvidia/k8s-device-plugin@sha256:3c54348fe5a57e5700e7d8068e7531d2ef2d5f3ccb70c8f6bac0953432527abd
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:6bb786c41fa7b1a6238c023566fdeaeeea3ba842a8daec015a7bd85a8037b249
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:c22dc67465c33d2f17932931229e61a6e7d9e32c5ed2ecefe8d10e873ffedfeb
        - nvcr.io/nvidia/driver@sha256:43aa99352e1dc47e9218f36340ae6328a6282ed1e7e9a61b4204475e980230ef
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:119de9a331a47203858b99901f44d0c4a8052961b4e60327f4b100d0ab8c9df0
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b689dfb7806ae51b0072620b47c26114f05477e8ad9725892ee10366e5d93d8d
        version: 25.3.4
      entries:
      - name: gpu-operator-certified.v25.3.4
        version: 25.3.4
      - name: gpu-operator-certified.v25.3.3
        version: 25.3.3
      - name: gpu-operator-certified.v25.3.2
        version: 25.3.2
      - name: gpu-operator-certified.v25.3.1
        version: 25.3.1
      - name: gpu-operator-certified.v25.3.0
        version: 25.3.0
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v25.3
    defaultChannel: v25.3
    packageName: gpu-operator-certified
    provider:
      name: NVIDIA Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Yugabyte Inc
      provider-url: ""
    name: yugabyte-platform-operator-bundle-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: yugabyte-platform-operator.v0.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "yugabyte.com/v1alpha1",
                "kind": "YBPlatform",
                "metadata": {
                  "name": "ybplatform-sample"
                },
                "spec": {
                  "domainName": "cluster.local",
                  "helm": {
                    "package": "/opt/yugabyte/helm/yugabyte-latest.tgz",
                    "timeout": 900
                  },
                  "helm2Legacy": false,
                  "image": {
                    "commonRegistry": "",
                    "nginx": {
                      "name": "nginxinc/nginx-unprivileged",
                      "registry": "",
                      "tag": "1.17.4-amd64"
                    },
                    "postgres": {
                      "name": "postgres",
                      "registry": "",
                      "tag": 11.5
                    },
                    "prometheus": {
                      "name": "prom/prometheus",
                      "registry": "",
                      "tag": "v2.2.1"
                    },
                    "pullPolicy": "IfNotPresent",
                    "pullSecret": "yugabyte-k8s-pull-secret",
                    "repository": "registry.connect.redhat.com/yugabytedb/yugabyte-platform",
                    "tag": "latest",
                    "thirdparty-deps": {
                      "name": "yugabyte/thirdparty-deps",
                      "registry": "quay.io",
                      "tag": "latest"
                    }
                  },
                  "ip_version_support": "v4_only",
                  "ocpCompatibility": {
                    "enabled": true
                  },
                  "rbac": {
                    "create": true
                  },
                  "securityContext": {
                    "enabled": false,
                    "fsGroup": 1001,
                    "runAsUser": 1001
                  },
                  "tls": {
                    "certificate": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tDQpNSUlHTHpDQ0JSZWdBd0lCQWdJUUQ2R1ZmL0ViUDhSMFQvcWxjRGZyM3pBTkJna3Foa2lHOXcwQkFRc0ZBREJlDQpNUXN3Q1FZRFZRUUdFd0pWVXpFVk1CTUdBMVVFQ2hNTVJHbG5hVU5sY25RZ1NXNWpNUmt3RndZRFZRUUxFeEIzDQpkM2N1WkdsbmFXTmxjblF1WTI5dE1SMHdHd1lEVlFRREV4UlNZWEJwWkZOVFRDQlNVMEVnUTBFZ01qQXhPREFlDQpGdzB4T1RBNE1ERXdNREF3TURCYUZ3MHlNVEEzTXpFeE1qQXdNREJhTUI4eEhUQWJCZ05WQkFNTUZDb3VZMnh2DQpkV1F1ZVhWbllXSjVkR1V1WTI5dE1JSUJJakFOQmdrcWhraUc5dzBCQVFFRkFBT0NBUThBTUlJQkNnS0NBUUVBDQp3SmswU25ub1pRYW1hWm0rMVVJdjdRYjBLalhZeVEwckpJSGpsbE03elhFYUFoL0JqVjZ1M0h1TWxrTlZSekVFDQpGeEtGUm5nSlpXTVlPcy9sWDU1QmtBS0RocXNmaDl0ZVA4ZFVQU1RTVTN6bWRVc0dTcDJIMGpKbVhuRWNvMFpzDQpwYlVyYWhlcWlxSUUyZ1gwQlpvYldvWGRHc1EydXJiUGxyMGlVSnUxVnp4SGRoTS9nTnFQWEN5bE5YdENrbVI2DQo0eXpIa0Rob21pRTAySTMwTXBlOERxczl4d3lKd3F6T3J2ZnA0eUE0Ukhna2g1clovK1Erc0VQNkIyVzBHQXc2DQpzZDNTYytOTytYVUZoVnlFRDhqZkNsTndPVHZodi9keE5IWHZ4WHd0RWNjZTd1TVBBTFJzSFh1RW1kTkN5N2pvDQpibzJuUWFObzZBQ1c0U2s5RWJVZXB3SURBUUFCbzRJREpqQ0NBeUl3SHdZRFZSMGpCQmd3Rm9BVVU4b1hXZnhyDQp3QU1oTHhxdTVLcW9ISUpXMm5Vd0hRWURWUjBPQkJZRUZHemw3bjdOMy9GbFdDSDFvdWFCWldKMjI1UStNQjhHDQpBMVVkRVFRWU1CYUNGQ291WTJ4dmRXUXVlWFZuWVdKNWRHVXVZMjl0TUE0R0ExVWREd0VCL3dRRUF3SUZvREFkDQpCZ05WSFNVRUZqQVVCZ2dyQmdFRkJRY0RBUVlJS3dZQkJRVUhBd0l3UGdZRFZSMGZCRGN3TlRBem9ER2dMNFl0DQphSFIwY0RvdkwyTmtjQzV5WVhCcFpITnpiQzVqYjIwdlVtRndhV1JUVTB4U1UwRkRRVEl3TVRndVkzSnNNRXdHDQpBMVVkSUFSRk1FTXdOd1lKWUlaSUFZYjliQUVDTUNvd0tBWUlLd1lCQlFVSEFnRVdIR2gwZEhCek9pOHZkM2QzDQpMbVJwWjJsalpYSjBMbU52YlM5RFVGTXdDQVlHWjRFTUFRSUJNSFVHQ0NzR0FRVUZCd0VCQkdrd1p6QW1CZ2dyDQpCZ0VGQlFjd0FZWWFhSFIwY0RvdkwzTjBZWFIxY3k1eVlYQnBaSE56YkM1amIyMHdQUVlJS3dZQkJRVUhNQUtHDQpNV2gwZEhBNkx5OWpZV05sY25SekxuSmhjR2xrYzNOc0xtTnZiUzlTWVhCcFpGTlRURkpUUVVOQk1qQXhPQzVqDQpjblF3Q1FZRFZSMFRCQUl3QURDQ0FYNEdDaXNHQVFRQjFua0NCQUlFZ2dGdUJJSUJhZ0ZvQUhVQTdrdTl0M1hPDQpZTHJoUW1rZnErR2VacU1QZmwrd2N0aURBTVI3aVhxby9jc0FBQUZzVHEwL013QUFCQU1BUmpCRUFpQlg3NXJ0DQpwNkdQbUQxbVl1S1hvcXVRVytIcUtRNE8rV1RXbTNCbVBteWF1d0lnQUoya01WdmduaUx1dEh5Y3AyYzBwdDRsDQptL3M5SWN2Yi9aN09jMU82b2xJQWRnQ0hkYi9uV1h6NGpFT1pYNzN6YnY5V2pVZFdOdjlLdFdEQnRPci9YcUNEDQpEd0FBQVd4T3JUOXNBQUFFQXdCSE1FVUNJSHphY3NCcmpEN1FER1ZRSHA5YUZPMnJ4VEx4NU9GOVE2MTFYaUU2DQpZUEk3QWlFQW9xd3pTd1BVaGt4bVhYNzN0SDVEcVNwaUxuOHdrMXFWeWhZOTVmSmxZa0VBZHdCRWxHVXVzTzdPDQpyOFJBQjlpby9pakEydWFDdnRqTE1iVS8wek9XdGJhQnFBQUFBV3hPclQ1M0FBQUVBd0JJTUVZQ0lRQ1ZLaHQ5DQpWTWhlNTUvZEwrSGN6Qzk5a2xheG9DcW5zOStRNGVOTDZ2Smd4QUloQUsyZmp4a1RadGw4WFZ4WHEwc2NSTWNJDQpiSktHcGJYdEV0N3dFVzRkZDNGS01BMEdDU3FHU0liM0RRRUJDd1VBQTRJQkFRQm1GcDNCK1g2Y21Cb0FLWkhrDQpIWWxzanhIZk01bDQyZWJtSFdGdlQyTi9aVkZ3MEZ3YndBNExyNzlvd01QdU5FbVNNa2pmVVhxR011Q2VlVTdODQo0MlRwUFdPaGNzVHF6ZkxVa0hhMU9rK011L3RUZHBNcmhROEVqbkhMY3FKVE9MbmRSVy8zcE5IM2xrK2FJejNoDQpDVGFueElRVEE2cUxRNThRakszWmFMVzdqWVVNcVJFcnBGVDlBVXkxMDAxMjlOam5HVkhtYmU1bW1ZbEwvQlhMDQpmdHBJUUJJZnhCNXpHbHpTcjRHT3p5RDBES05YU040RU0yM2pIM1ljeUg4eFQrODU4ZkxSaXQ4SFlOQ3RQbkloDQoyRzR6UjV3emZFVDlnZUh6aWtuY0ZMWUpMMEpKMVIrZGVNRE91bmJWOUJqcmZJb3NNSk53TUV2ejFZczlBWXhHDQpxSFNYDQotLS0tLUVORCBDRVJUSUZJQ0FURS0tLS0tDQotLS0tLUJFR0lOIENFUlRJRklDQVRFLS0tLS0KTUlJRXNUQ0NBNW1nQXdJQkFnSVFDS1dpUnMxTFhJeUQxd0swdTZ0VFNUQU5CZ2txaGtpRzl3MEJBUXNGQURCaApNUXN3Q1FZRFZRUUdFd0pWVXpFVk1CTUdBMVVFQ2hNTVJHbG5hVU5sY25RZ1NXNWpNUmt3RndZRFZRUUxFeEIzCmQzY3VaR2xuYVdObGNuUXVZMjl0TVNBd0hnWURWUVFERXhkRWFXZHBRMlZ5ZENCSGJHOWlZV3dnVW05dmRDQkQKUVRBZUZ3MHhOekV4TURZeE1qSXpNek5hRncweU56RXhNRFl4TWpJek16TmFNRjR4Q3pBSkJnTlZCQVlUQWxWVApNUlV3RXdZRFZRUUtFd3hFYVdkcFEyVnlkQ0JKYm1NeEdUQVhCZ05WQkFzVEVIZDNkeTVrYVdkcFkyVnlkQzVqCmIyMHhIVEFiQmdOVkJBTVRGRkpoY0dsa1UxTk1JRkpUUVNCRFFTQXlNREU0TUlJQklqQU5CZ2txaGtpRzl3MEIKQVFFRkFBT0NBUThBTUlJQkNnS0NBUUVBNVMyb2loRW85bm5wZXpvemlEdHg0V1dMTENsbC9lMHQxRVllbUU1bgorTWdQNXZpYUhMeStWcEhQK25kWDVEMThJTkl1dUFWOHdGcTI2S0Y1VTBXTklaaVFwNm1MdElXalVlV0RQQTI4Ck9leWhUbGo5VExrMmJleXRidEZVNnlwYnBXVWx0bXZZNVY4bmdzcEM3bkZSTkNqcGZuREVEMmtSeUp6Tzh5b0sKTUZ6NEo0SkU4TjdOQTF1SndVRUZNVXZITHMwc2NMb1Baa0tjZXdJUm0xUlYyQXhtRlF4SmtkZjdZTjlQY2traQpmMlhnbTNiNDhCWm4wemYwcVhzU2VHdTg0dWE5Z3d6anpJN3RiVEJqYXlUcFQrL1hwV3VCVnY2ZnZhckk2YmlrCktCODU5T1NHUXV3NzNYWGdldUZ3RVBIVElSb1V0a3p1My9FUStMdHd6bmtrZFFJREFRQUJvNElCWmpDQ0FXSXcKSFFZRFZSME9CQllFRkZQS0YxbjhhOEFESVM4YXJ1U3FxQnlDVnRwMU1COEdBMVVkSXdRWU1CYUFGQVBlVURWVwowVXk3WnZDajRoc2J3NWV5UGRGVk1BNEdBMVVkRHdFQi93UUVBd0lCaGpBZEJnTlZIU1VFRmpBVUJnZ3JCZ0VGCkJRY0RBUVlJS3dZQkJRVUhBd0l3RWdZRFZSMFRBUUgvQkFnd0JnRUIvd0lCQURBMEJnZ3JCZ0VGQlFjQkFRUW8KTUNZd0pBWUlLd1lCQlFVSE1BR0dHR2gwZEhBNkx5OXZZM053TG1ScFoybGpaWEowTG1OdmJUQkNCZ05WSFI4RQpPekE1TURlZ05hQXpoakZvZEhSd09pOHZZM0pzTXk1a2FXZHBZMlZ5ZEM1amIyMHZSR2xuYVVObGNuUkhiRzlpCllXeFNiMjkwUTBFdVkzSnNNR01HQTFVZElBUmNNRm93TndZSllJWklBWWI5YkFFQ01Db3dLQVlJS3dZQkJRVUgKQWdFV0hHaDBkSEJ6T2k4dmQzZDNMbVJwWjJsalpYSjBMbU52YlM5RFVGTXdDd1lKWUlaSUFZYjliQUVCTUFnRwpCbWVCREFFQ0FUQUlCZ1puZ1F3QkFnSXdEUVlKS29aSWh2Y05BUUVMQlFBRGdnRUJBSDRqeC9MS05XNVprbEZjCllXczhFamJtMG55ektlWkMyS09WWVI3UDhnZXZLeXNsV200WG80QlN6S3IyMzVGc0o0YUZ0NnlBaXYxZVkwdFoKL1pOMThiT0dTR1N0b0VjL0pFNG9jSXpyOFA1TWcxMWtSWUhibWdZbnIxUnhla2k1bVNlYjM5REd4VHBKRDRrRwpoczVsWE5vbzRjb25VaWlKd0thcUg3dmgyYmFyeWQ4cE1JU2FnODNKVXF5VkdjMnRXUHBPMDMyOS9DV3Eya3J5CnF2NjZPU01qd3VsVXowZFhmNE9IUWFzUjdDTmZJcis0S1NjYzZBQmxRNVJERjg2UEdlRTZrZHdTUWtGaUIvY1EKeXNOeXEwakVEUVRrZmEycGptdVd0TUNOYkJuaEZYQlllamZ1YkloYVViRXYyRk9RQjNkQ2F2K0ZQZzVlRXZlWApUVnlNbkdvPQotLS0tLUVORCBDRVJUSUZJQ0FURS0tLS0tCg==",
                    "enabled": false,
                    "hostname": "localhost",
                    "key": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQpNSUlFb3dJQkFBS0NBUUVBd0prMFNubm9aUWFtYVptKzFVSXY3UWIwS2pYWXlRMHJKSUhqbGxNN3pYRWFBaC9CCmpWNnUzSHVNbGtOVlJ6RUVGeEtGUm5nSlpXTVlPcy9sWDU1QmtBS0RocXNmaDl0ZVA4ZFVQU1RTVTN6bWRVc0cKU3AySDBqSm1YbkVjbzBac3BiVXJhaGVxaXFJRTJnWDBCWm9iV29YZEdzUTJ1cmJQbHIwaVVKdTFWenhIZGhNLwpnTnFQWEN5bE5YdENrbVI2NHl6SGtEaG9taUUwMkkzME1wZThEcXM5eHd5Sndxek9ydmZwNHlBNFJIZ2toNXJaCi8rUStzRVA2QjJXMEdBdzZzZDNTYytOTytYVUZoVnlFRDhqZkNsTndPVHZodi9keE5IWHZ4WHd0RWNjZTd1TVAKQUxSc0hYdUVtZE5DeTdqb2JvMm5RYU5vNkFDVzRTazlFYlVlcHdJREFRQUJBb0lCQUJDbDdaZzJlM2tvSHdCbgowdEFhWFl3dW1ySi9YeExaK1kyNVVKUHc4KzFwUm1kQ21YZC81THVZVG1tNnZrZGVIRTFtUTNENDlVNzNSdVBtCmpwdzZVTTZDK1IyTjh6YktKcFo2dzNSWmwzaXlCalhEVzNMZHRSemRVRzMvZHd0Q0RpZ3d4ZVR3ME1RVkJ6aFYKWldlS2o2eStmNE5OekxGTkZocWdsTFBGRWVMTlk3WUgvYUdHdTZwaVpERkFMT0liMlFVUUVvVVRkUEJabDMxSwpKL0Y0cHE1RkV6ekQ0aHMwYXJzSlIxeUd4L0NCRWxqQWRSNWE4Y3FHdGhxYjFMdXJTMVcxT0dNUGkzNFdDblJKCnYwTC9zSU9ORCt5cjdjd3ZQaFFzcmI3Tk9NbGxRSDYwUGdyQ0JyOUVwalFiRm1ZN3NCTS9jYUZTWHVDTmZnTzMKT3hTQjF0RUNnWUVBNHpSMTh2aXl3Ykt5aE5xeEhkNzJ1UkxHS2k4SWZhVmhsNGxFeGdsR0pLNHQ0KzYreXhObAp2OFBkdGg3elJjL1pyQUlBUEtBRnZxa1RnVzkrc2llL3JPcE9PUzA2UnVIWVpJMjdpS3ZiL054N29YTm9TM0Y2CitXRlNJSVJqRFo0eHc2SU1iNUJFS0hOVGtKU1pBMXBkaXhYeFgzeDVTM2hCWTlmekl2NGdVVDhDZ1lFQTJRSDAKbnU5REZ0MDNmWmpvcjRMUkZCQS84NEV1dmdBL2RxMkZ4TzVvMjIrdk43Qk9LRW9IWTRuTzhYdFN2L2MzaVFZdApmTUdGZUR2eXdqRnFPU05SS2JxenMza3ZmY2gxNEhHdFltT29DVTF6Smd2VkF4SytWOFF4YWxMbmhlbjZsaTgwCnJKYmtWWUJXYnFHZHU1d2FPM1RyeGsvbHBMejFvSnliK0FQbmNKa0NnWUVBeHd3anRTeEtOc1hkL2VRMXZReWoKcTdvVjM4ajZzRkRkOGNzeGRDbVZ5UGJMZjErZXlQeUNiSUVrMExQR3ZzZDhWR1FnWHZ4MmEvVm1BRVIwZzVHdApLOGpOcE92Sm5XV29sVmRTRDBjNDRBc3RTR0Y1NG5aVm5QcTAzNFg3M1dkVkE2WUpMUWYvbmt5Q3ppeFE3UTRTCjI1SHF0NU91RS84VGttd1ROdm5aeURNQ2dZQlUxV0xmendFWEt4dktTY0NNL0VMbjZRSlhHeVUrdWJJVm52Z3cKYUZNdGxaZVYyYUowTlp4MTJTWEJXVWljS3ZUVzZFZFlBZjgvWlpBZFFsd1lRaW8wZWZ4T0tRM2J5a1hxbkFYLwo5SjE0SllMaVdMWEpJY2F2WU05QWxYQlU5WWMwdHZlTWtPQnFYR1Z0a2xpclF2Y3VUeUxzdjhjeWczd3ljWWEzCmlpOHI0UUtCZ0NaODZtcEZuUlN2aTdpTWw1MUdybDFQZER2YzZVMURRVkI5empTQlVzak1lamNTQXk4WmVJZlQKUDVoSGc3REFaUEhmUGpYLzZxUnpNQW8zWlFlZVBMdTNEYVgwbFBoekNhalF1bkpIbWZvYlA2QjFOUG45T056bQo0dE0wdEU1YTJVVmJuZW5haVZyaDRhSXlJTTFxVTdHT1ArcDNMNi9hNzVnSVQ2SVJRVlloCi0tLS0tRU5EIFJTQSBQUklWQVRFIEtFWS0tLS0tCg==",
                    "sslProtocols": ""
                  },
                  "yugaware": {
                    "health": {
                      "email": "",
                      "password": "",
                      "username": ""
                    },
                    "multiTenant": false,
                    "replicas": 1,
                    "service": {
                      "enabled": true,
                      "ip": "",
                      "type": "LoadBalancer"
                    },
                    "serviceAccount": "yugaware",
                    "storage": "100Gi",
                    "storageClass": "standard"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/yugabytedb/yugabyte-platform-operator@sha256:f227c9e2c625b04160d952349ddba552ac66dae8e7da35f56746f3447c270ce3
          description: Operator for Yugabyte Platform makes it easy to deploy and
            upgrade Yugabyte Platform on OpenShift environments.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/yugabyte-platform-operator-bundle-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/yugabyte-platform-operator-bundle-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/yugabyte/yugabyte-platform-operator
          support: YugabyteDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy and manage Yugabyte Platform as a Kubernetes native
              custom resource.
            displayName: Yugabyte Platform
            kind: YBPlatform
            name: ybplatforms.yugabyte.com
            version: v1alpha1
        description: |2

          Operator for Yugabyte Platform makes it easy to deploy and upgrade Yugabyte Platform on OpenShift environments.
          Use Yugabyte Platform's orchestration and monitoring to manage YugabyteDB universes. Yugabyte Platform can create a YugabyteDB universe with many instances (VMs, pods, machines, etc., provided by IaaS), logically grouped together to form one logical distributed database. Each universe includes one or more clusters. A universe is comprised of one primary cluster and, optionally, one or more read replica clusters. All instances belonging to a cluster run on the same type of cloud provider instance type.

          ### Yugabyte Platform Features
          * **YugabyteDB-as-a-Service on Kubernetes** - Deploy your own YugabyteDB-as-a-Service running on any Kubernetes infrastructure, Yugabyte Platform provides central management for multiple database clusters and Includes 24/7 production support.
          * **Monitoring & Alerting** - Comprehensive monitoring & alerting without any configuration.
          * **Scheduled Backups** - Use Yugabyte Platform to perform a regular scheduled backup of a full YugabyteDB universe or selected tables.

          ### Getting Started
          Here are a few resources for getting started with Yugabyte platform
          * [Quick start guide for Yugabyte Platform](https://docs.yugabyte.com/latest/yugabyte-platform/deploy/) - Install Yugabyte platform on OpenShift.
          * [Quick Start for YugabyteDB](http://docs.yugabyte.com/latest/quick-start/) - install, create a local cluster and read/write from YugabyteDB.
          * [Explore core features](https://docs.yugabyte.com/latest/explore/) - automatic sharding & re-balancing, linear scalability, fault tolerance, tunable reads etc.
          * [Ecosystem integrations](https://docs.yugabyte.com/latest/develop/ecosystem-integrations/) - integrations with Apache Kafka/KSQL, Apache Spark, JanusGraph, KairosDB, Presto and more.
          * [Real world apps](https://docs.yugabyte.com/latest/develop/realworld-apps/) - sample real-world, end-to-end applications built using YugabyteDB.

          ### Help
          Cannot find what you are looking for? Have a question? We love to hear from you - [please file a GitHub issue](https://github.com/YugaByte/yugabyte-db/issues) or talk to us live at [Slack](https://www.yugabyte.com/slack).
        displayName: Yugabyte Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - distributed sql
        links:
        - name: Yugabyte Platform Operator
          url: https://www.yugabyte.com/
        maintainers:
        - email: oss-maintainers@yugabyte.com
          name: Yugabyte
        maturity: alpha
        provider:
          name: Yugabyte Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9211b70edbb93d718428e1c9a992e5fc74a20cb213f8a0592ffe340c362890fe
        - registry.marketplace.redhat.com/rhm/yugabytedb/yugabyte-platform-operator@sha256:f227c9e2c625b04160d952349ddba552ac66dae8e7da35f56746f3447c270ce3
        version: 0.0.6
      entries:
      - name: yugabyte-platform-operator.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: yugabyte-platform-operator-bundle-rhmp
    provider:
      name: Yugabyte Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: volsync-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: volsync-product.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
          createdAt: "2025-08-27T13:12:05Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.13.1'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
        version: 0.13.1
      entries:
      - name: volsync-product.v0.13.1
        version: 0.13.1
      - name: volsync-product.v0.13.0
        version: 0.13.0
      - name: volsync-product.v0.12.2
        version: 0.12.2
      - name: volsync-product.v0.12.1
        version: 0.12.1
      - name: volsync-product.v0.12.1-0.1749677898.p
        version: 0.12.1+0.1749677898.p
      - name: volsync-product.v0.12.0
        version: 0.12.0
      name: stable
    - currentCSV: volsync-product.v0.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:d345825e953c4c170db45a5d92639182b7ee18933287aaeea035f2a2bad13c37
          createdAt: 09 Jan 2025, 00:47
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.10.2'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:d345825e953c4c170db45a5d92639182b7ee18933287aaeea035f2a2bad13c37
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 0.10.2
      entries:
      - name: volsync-product.v0.10.2
        version: 0.10.2
      - name: volsync-product.v0.10.1
        version: 0.10.1
      - name: volsync-product.v0.10.1-0.1730236570.p
        version: 0.10.1+0.1730236570.p
      - name: volsync-product.v0.10.0
        version: 0.10.0
      name: stable-0.10
    - currentCSV: volsync-product.v0.11.2-0.1749674925.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:9da8572c79b75767952cf699e5b1699b9f46d83586aa2d6a17fb23c092713b62
          createdAt: 16 Apr 2025, 06:42
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.11.2'
          olm.substitutesFor: volsync-product.v0.11.2
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:9da8572c79b75767952cf699e5b1699b9f46d83586aa2d6a17fb23c092713b62
        version: 0.11.2+0.1749674925.p
      entries:
      - name: volsync-product.v0.11.2
        version: 0.11.2
      - name: volsync-product.v0.11.2-0.1749674925.p
        version: 0.11.2+0.1749674925.p
      - name: volsync-product.v0.11.1
        version: 0.11.1
      - name: volsync-product.v0.11.0
        version: 0.11.0
      name: stable-0.11
    - currentCSV: volsync-product.v0.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:7d1ac69fe3d0fddc66acdfce9361a4fc8dd3f549f66c98450f7a0e4c5cc893a0
          createdAt: "2025-07-07T13:46:18Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.12.2'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:7d1ac69fe3d0fddc66acdfce9361a4fc8dd3f549f66c98450f7a0e4c5cc893a0
        version: 0.12.2
      entries:
      - name: volsync-product.v0.12.2
        version: 0.12.2
      - name: volsync-product.v0.12.1
        version: 0.12.1
      - name: volsync-product.v0.12.1-0.1749677898.p
        version: 0.12.1+0.1749677898.p
      - name: volsync-product.v0.12.0
        version: 0.12.0
      name: stable-0.12
    - currentCSV: volsync-product.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
          createdAt: "2025-08-27T13:12:05Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.13.1'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
        version: 0.13.1
      entries:
      - name: volsync-product.v0.13.1
        version: 0.13.1
      - name: volsync-product.v0.13.0
        version: 0.13.0
      name: stable-0.13
    - currentCSV: volsync-product.v0.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:ef44675fe70ee9ec8109fb91c6559f0c5518b45219e19175a18a5004e40b65f7
          createdAt: 22 Aug 2024, 22:06
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.9.3'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ReplicationDestination defines the destination for a replicated
              volume
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: ReplicationSource defines the source for a replicated volume
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:ef44675fe70ee9ec8109fb91c6559f0c5518b45219e19175a18a5004e40b65f7
        version: 0.9.3
      entries:
      - name: volsync-product.v0.9.3
        version: 0.9.3
      - name: volsync-product.v0.9.2
        version: 0.9.2
      - name: volsync-product.v0.9.1
        version: 0.9.1
      - name: volsync-product.v0.9.0
        version: 0.9.0
      name: stable-0.9
    defaultChannel: stable
    packageName: volsync-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: deployment-validation-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: deployment-validation-operator.v0.7.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Application Runtime, Monitoring, Security
          certified: "false"
          containerImage: quay.io/deployment-validation-operator/dv-operator:0.7.12
          createdAt: "2025-09-11T00:00:00Z"
          description: The deployment validation operator
          ignore-check.kube-linter.io/minimum-three-replicas: This deployment uses
            1 pod as currently replicating does not replicate metric data causing
            installation issues
          repository: https://github.com/app-sre/deployment-validation-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Deployment Validation Operator (DVO) checks deployments and other resources against a curated collection of best practices.
          These best practices focus mainly on ensuring that the applications are fault-tolerant. DVO reports failed validations via Prometheus metrics.
          If the best-practice check has failed, the metrics will report `1`.
        displayName: Deployment Validation Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: false
          type: MultiNamespace
        keywords:
        - dvo
        links:
        - name: repository
          url: https://github.com/app-sre/deployment-validation-operator
        - name: containerImage
          url: https://quay.io/deployment-validation-operator/dv-operator:0.7.12
        maintainers:
        - email: dvo-owners@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/deployment-validation-operator/dv-operator:0.7.12
        version: 0.7.12
      entries:
      - name: deployment-validation-operator.v0.7.12
        version: 0.7.12
      - name: deployment-validation-operator.v0.7.9
        version: 0.7.9
      - name: deployment-validation-operator.v0.7.8
        version: 0.7.8
      - name: deployment-validation-operator.v0.7.7
        version: 0.7.7
      - name: deployment-validation-operator.v0.7.6
        version: 0.7.6
      - name: deployment-validation-operator.v0.7.5
        version: 0.7.5
      - name: deployment-validation-operator.v0.7.4
        version: 0.7.4
      - name: deployment-validation-operator.v0.7.3
        version: 0.7.3
      - name: deployment-validation-operator.v0.7.2
        version: 0.7.2
      - name: deployment-validation-operator.v0.7.1
        version: 0.7.1
      - name: deployment-validation-operator.v0.7.0
        version: 0.7.0
      - name: deployment-validation-operator.v0.6.0
        version: 0.6.0
      - name: deployment-validation-operator.v0.5.0
        version: 0.5.0
      - name: deployment-validation-operator.v0.4.0
        version: 0.4.0
      - name: deployment-validation-operator.v0.3.0
        version: 0.3.0
      - name: deployment-validation-operator.v0.2.2
        version: 0.2.2
      name: alpha
    defaultChannel: alpha
    packageName: deployment-validation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: ""
    name: gitops-primer
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitops-primer.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "primer.gitops.io/v1alpha1",
                "kind": "Export",
                "metadata": {
                  "name": "primer"
                },
                "spec": {
                  "method": "download"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/migtools/gitops-primer:v0.0.11
          operatorframework.io/suggested-namespace: gitops-primer-system
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Export is the Schema for the exports API
            displayName: Export
            kind: Export
            name: exports.primer.gitops.io
            version: v1alpha1
        description: Export and normalize Kubernetes YAML files for GitOps
        displayName: gitops-primer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - operator
        links:
        - name: Gitops Primer
          url: https://github.com/migtools/gitops-primer
        maturity: alpha
        provider:
          name: Konveyor
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6d57bfd91fac9b68eb72d27226bc297472ceb136c996628b845ecc54a48b31cb
        - quay.io/migtools/gitops-primer:v0.0.11
        - quay.io/migtools/gitops-primer-webhook:v0.0.11
        - quay.io/migtools/gitops-primer-downloader:v0.0.11
        - quay.io/migtools/gitops-primer-export:v0.0.11
        - quay.io/openshift/origin-oauth-proxy:4.9
        version: 0.0.11
      entries:
      - name: gitops-primer.v0.0.11
        version: 0.0.11
      - name: gitops-primer.v0.0.10
        version: 0.0.10
      - name: gitops-primer.v0.0.9
        version: 0.0.9
      - name: gitops-primer.v0.0.8
        version: 0.0.8
      - name: gitops-primer.v0.0.7
        version: 0.0.7
      - name: gitops-primer.v0.0.6
        version: 0.0.6
      - name: gitops-primer.v0.0.5
        version: 0.0.5
      - name: gitops-primer.v0.0.4
        version: 0.0.4
      - name: gitops-primer.v0.0.3
        version: 0.0.3
      - name: gitops-primer.v0.0.2
        version: 0.0.2
      - name: gitops-primer.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: gitops-primer
    provider:
      name: Konveyor
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Craig Trought
      provider-url: ""
    name: prometheus
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometheusoperator.0.56.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "Prometheus",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "replicas": 2,
                  "serviceAccountName": "prometheus-k8s",
                  "serviceMonitorSelector": {},
                  "ruleSelector": {},
                  "podMonitorSelector": {},
                  "probeSelector": {},
                  "alerting": {
                    "alertmanagers": [
                      {
                        "namespace": "monitoring",
                        "name": "alertmanager-main",
                        "port": "web"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "ServiceMonitor",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "operated-prometheus": "true"
                    }
                  },
                  "endpoints": [
                    {
                      "port": "web",
                      "interval": "30s"
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "PodMonitor",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "app": "prometheus"
                    }
                  },
                  "podMetricsEndpoints": [
                    {
                      "port": "web",
                      "interval": "30s"
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "Alertmanager",
                "metadata": {
                  "name": "alertmanager-main"
                },
                "spec": {
                  "replicas": 3,
                  "alertmanagerConfigSelector": {}
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "PrometheusRule",
                "metadata": {
                  "name": "prometheus-example-rules"
                },
                "spec": {
                  "groups": [
                    {
                      "name": "./example.rules",
                      "rules": [
                        {
                          "alert": "ExampleAlert",
                          "expr": "vector(1)"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1alpha1",
                "kind": "AlertmanagerConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "receivers": [
                    {
                      "name": "example"
                    }
                  ],
                  "route": {
                    "receiver": "example"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/prometheus-operator/prometheus-operator:v0.56.3
          createdAt: "2022-05-24T9:40:00Z"
          description: Manage the full lifecycle of configuring and managing Prometheus
            and Alertmanager servers.
          repository: https://github.com/prometheus-operator/prometheus-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A running Prometheus instance
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.coreos.com
            version: v1
          - description: A Prometheus Rule configures groups of sequentially evaluated
              recording and alerting rules.
            displayName: Prometheus Rule
            kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - description: Configures prometheus to monitor a particular k8s service
            displayName: Service Monitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - description: Configures prometheus to monitor a particular pod
            displayName: Pod Monitor
            kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
          - description: Configures an Alertmanager for the namespace
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.coreos.com
            version: v1
          - description: Configures subsections of Alertmanager configuration specific
              to each namespace
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.coreos.com
            version: v1alpha1
          - description: Defines how groups of ingresses and static targets should
              be monitored.
            displayName: Probe
            kind: Probe
            name: probes.monitoring.coreos.com
            version: v1
          - description: Allows recording and alerting rules to be processed across
              multiple Prometheus instances.
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.coreos.com
            version: v1
        description: |
          The Prometheus Operator for Kubernetes provides easy monitoring definitions for Kubernetes services and deployment and management of Prometheus instances.

          Once installed, the Prometheus Operator provides the following features:

          * **Create/Destroy**: Easily launch a Prometheus instance for your Kubernetes namespace, a specific application or team easily using the Operator.
          * **Simple Configuration**: Configure the fundamentals of Prometheus like versions, persistence, retention policies, and replicas from a native Kubernetes resource.
          * **Target Services via Labels**: Automatically generate monitoring target configurations based on familiar Kubernetes label queries; no need to learn a Prometheus specific configuration language.

          ### Other Supported Features

          **High availability**

          Multiple instances are run across failure zones and data is replicated. This keeps your monitoring available during an outage, when you need it most.

          **Updates via automated operations**

          New Prometheus versions are deployed using a rolling update with no downtime, making it easy to stay up to date.

          **Handles the dynamic nature of containers**

          Alerting rules are attached to groups of containers instead of individual instances, which is ideal for the highly dynamic nature of container deployment.
        displayName: Prometheus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - prometheus
        - monitoring
        - tsdb
        - alerting
        - observability
        links:
        - name: Prometheus
          url: https://www.prometheus.io/
        - name: Documentation
          url: https://github.com/prometheus-operator/prometheus-operator/tree/master/Documentation
        - name: Prometheus Operator
          url: https://github.com/prometheus-operator/prometheus-operator
        maintainers:
        - email: k8s@trought.ca
          name: Craig Trought
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Craig Trought
        relatedImages:
        - quay.io/prometheus-operator/prometheus-operator:v0.56.3
        version: 0.56.3
      entries:
      - name: prometheusoperator.0.56.3
        version: 0.56.3
      - name: prometheusoperator.0.47.0
        version: 0.47.0
      name: beta
    defaultChannel: beta
    packageName: prometheus
    provider:
      name: Craig Trought
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC
      provider-url: ""
    name: nspc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nspc-operator.v1.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.nec.com/v1",
                "kind": "NSPC",
                "metadata": {
                  "name": "nspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.connect.redhat.com/nec/nspc-operator@sha256:50fe10f05d8aa7db1b97280e5842bb0b67c6f9bfb01c0d17c5be3cd11b78e853
          createdAt: "2025-07-24T00:00:00Z"
          description: An operator for managing NEC Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: NEC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NSPC is the Schema for the nspcs API
            displayName: NSPC
            kind: NSPC
            name: nspcs.csi.nec.com
            version: v1
        description: |-
          ## About
          NEC Storage Plug-in for Containers (NSPC) is a plugin that integrates NEC Storage V Series into Kubernetes based clusters.
          NSPC provides persistent volume provisioning capabilities from NEC Storage V Series arrays.

          For full documentation, refer to the reference guide.

          ## Requirements

          ### Supported Driver Version

          * NSPC v3.16.0
        displayName: NEC Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - NEC
        - Storage
        - iStorage
        - V Series
        - Container Storage Interface
        - CSI
        maturity: stable
        provider:
          name: NEC
        relatedImages:
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:81d32545fb1b2c319bfeb36a70db4b4d3aea9c51c0574b209b8f64750b32d2f0
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.connect.redhat.com/nec/nspc-operator@sha256:50fe10f05d8aa7db1b97280e5842bb0b67c6f9bfb01c0d17c5be3cd11b78e853
        - registry.connect.redhat.com/nec/nspc-csi-driver@sha256:e50d60e025b3c8ae522a8c5ae78a07f8a624a1210ed7a078c8084efd76ccc646
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        version: 1.16.0
      entries:
      - name: nspc-operator.v1.16.0
        version: 1.16.0
      name: stable
    defaultChannel: stable
    packageName: nspc-operator
    provider:
      name: NEC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: tempo-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tempo-operator.v0.18.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoMonolithic",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "jaegerui": {
                    "enabled": true,
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    }
                  },
                  "storage": {
                    "traces": {
                      "backend": "memory"
                    }
                  }
                }
              },
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoStack",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "resources": {
                    "total": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2Gi"
                      }
                    }
                  },
                  "storage": {
                    "secret": {
                      "name": "my-storage-secret",
                      "type": "s3"
                    }
                  },
                  "storageSize": "1Gi",
                  "template": {
                    "queryFrontend": {
                      "jaegerQuery": {
                        "enabled": true,
                        "ingress": {
                          "type": "route"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/rhosdt/tempo-rhel8-operator@sha256:9bfa5a2ffb09b501f001a4275379210621f06620adcf6454a9327c65538872f5
          createdAt: 22 Sep 2025, 11:15
          description: Create and manage deployments of Tempo, a high-scale distributed
            tracing backend.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.18.0-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-tempo-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/tempo-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TempoMonolithic manages a Tempo deployment in monolithic
              mode.
            displayName: TempoMonolithic
            kind: TempoMonolithic
            name: tempomonolithics.tempo.grafana.com
            version: v1alpha1
          - description: TempoStack manages a Tempo deployment in microservices mode.
            displayName: TempoStack
            kind: TempoStack
            name: tempostacks.tempo.grafana.com
            version: v1alpha1
        description: |-
          Red Hat OpenShift distributed tracing platform based on Tempo. Tempo is an open-source, easy-to-use, and highly scalable distributed tracing backend. It provides observability for microservices architectures by allowing developers to track requests as they flow through distributed systems. Tempo is optimized to handle large volumes of trace data and is designed to be highly performant even under heavy loads.
          It can ingest common open source tracing protocols including Jaeger, Zipkin, and OpenTelemetry and requires only object storage to operate.

          The Tempo Operator supports Tempo deployments in Microservices mode (`TempoStack` CR) and Monolithic mode (`TempoMonolithic` CR).

          ### Core capabilities
          Tempo is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed transaction monitoring
          * Root cause analysis
          * Performance / latency optimization

          ### Operator features
          * **Resource Limits** - Specify overall resource requests and limits in the `TempoStack` CR; the operator assigns fractions of it to each component
          * **AuthN and AuthZ** - Supports OpenID Control (OIDC) and role-based access control (RBAC)
          * **Managed upgrades** - Updating the operator will automatically update all managed Tempo clusters
          * **Multitenancy** - Multiple tenants can send traces to the same Tempo cluster
          * **mTLS** - Communication between the Tempo components can be secured via mTLS
          * **Jaeger UI** - Traces can be visualized in Jaeger UI and exposed via Ingress or OpenShift Route
          * **Observability** - The operator and `TempoStack` operands expose telemetry (metrics, traces) and integrate with Prometheus `ServiceMonitor` and `PrometheusRule`

          ### Before you start
          `TempoStack` requires object storage to store its traces. `TempoMonolithic` can store traces in-memory, in a Persistent Volume and in object storage.
          Please ensure that your system has a compatible object storage solution that is supported, such as OpenShift Data Foundation, Minio, AWS S3, Azure Storage, or Google Cloud Storage.

          ### Examples
          There is a list of examples to help you deploy different configurations in [this GitHub repository](https://github.com/os-observability/redhat-rhosdt-samples).

          ### Support & Troubleshooting
          Tempo Operator is available and supported as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/distributed_tracing) is recommended when installing, configuring, and managing the Operator and instances.
        displayName: Tempo Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tempo
        - tracing
        - observability
        - monitoring
        - database
        links:
        - name: Tempo Operator
          url: https://github.com/grafana/tempo-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6bcfa981a7a75c242a16afa6dab36bf6e3b8a1442fa863875919ad551f48eceb
        - registry.redhat.io/rhosdt/tempo-rhel8@sha256:530a0ef3aaea3f07d6210a683a771c416cea8ce5accac8f5f751dcd7232cb341
        - registry.redhat.io/rhosdt/tempo-query-rhel8@sha256:5b89188bc587a81af3ecdca3caabd6e48356e9ba4f9070860f5160544ce30eb9
        - registry.redhat.io/rhosdt/tempo-rhel8-operator@sha256:9bfa5a2ffb09b501f001a4275379210621f06620adcf6454a9327c65538872f5
        - registry.redhat.io/rhosdt/tempo-jaeger-query-rhel8@sha256:53b62fea39ce211842026e8365d61f579b8e01e8587a1ce639423d72f3b44f45
        - registry.redhat.io/rhosdt/tempo-gateway-rhel8@sha256:095ec83219660bb1955758b94a9431b3add572da4dc2cba61b42b59fa6b4bba0
        - registry.redhat.io/rhosdt/tempo-gateway-opa-rhel8@sha256:22e876837d00371d24feac3f989ead3d4f034400b292f043b18f3839bfde564a
        version: 0.18.0-1
      entries:
      - name: tempo-operator.v0.18.0-1
        version: 0.18.0-1
      - name: tempo-operator.v0.16.0-2
        version: 0.16.0-2
      - name: tempo-operator.v0.16.0-1
        version: 0.16.0-1
      - name: tempo-operator.v0.15.4-1
        version: 0.15.4-1
      - name: tempo-operator.v0.15.3-1
        version: 0.15.3-1
      - name: tempo-operator.v0.14.1-2
        version: 0.14.1-2
      - name: tempo-operator.v0.14.1-1
        version: 0.14.1-1
      - name: tempo-operator.v0.13.0-1-0.1730817419.p
        version: 0.13.0-1+0.1730817419.p
      - name: tempo-operator.v0.13.0-1
        version: 0.13.0-1
      - name: tempo-operator.v0.10.0-8
        version: 0.10.0-8
      - name: tempo-operator.v0.10.0-7
        version: 0.10.0-7
      - name: tempo-operator.v0.10.0-6
        version: 0.10.0-6
      - name: tempo-operator.v0.8.0-3
        version: 0.8.0-3
      - name: tempo-operator.v0.8.0-2
        version: 0.8.0-2
      - name: tempo-operator.v0.6.0-2
        version: 0.6.0-2
      - name: tempo-operator.v0.3.1-3
        version: 0.3.1-3
      - name: tempo-operator.v0.1.0-6
        version: 0.1.0-6
      - name: tempo-operator.v0.1.0-6-0.1687199947.p
        version: 0.1.0-6+0.1687199947.p
      name: stable
    defaultChannel: stable
    packageName: tempo-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: logging-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: logging-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Elasticsearch","metadata":{"name":"elasticsearch"},"spec":{"esClusterName":"prod","esVersion":"7.17.0","esSecurity":{"autoGeneratePassword":true,"tlsEnabled":true}}},{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Fluentd","metadata":{"name":"fluentd"},"spec":{"esCluster":{"host":"elasticsearch-master"},"esSecurity":{"tlsEnabled":true,"existingSecret":"elasticsearch-password"},"indexNameStrategy":"namespace_name"}},{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Kibana","metadata":{"name":"kibana"},"spec":{"replicas":1,"esCluster":{"host":"https://elasticsearch-master:9200","esVersion":"7.16.0","clusterName":"elasticsearch"},"esSecurity":{"tlsEnabled":true,"existingSecret":"elasticsearch-password"}}}]'
          capabilities: Seamless Upgrades
          categories: Database,  Monitoring, Big Data
          certified: "false"
          containerImage: quay.io/opstree/logging-operator:v0.4.0
          createdAt: 04-07-2022
          description: A golang-based operator to create and manage EFK (Elasticsearch,
            Fluentd, and Kibana) stack on Kubernetes.
          k8sMaxVersion: ""
          k8sMinVersion: ""
          repository: https://github.com/ot-container-kit/logging-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Elasticsearch
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Index Life Cycle
            displayName: IndexLifeCycle
            kind: IndexLifeCycle
            name: indexlifecycles.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Index Template
            displayName: IndexTemplate
            kind: IndexTemplate
            name: indextemplates.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Fluentd
            displayName: Fluentd
            kind: Fluentd
            name: fluentds.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Kibana
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.logging.opstreelabs.in
            version: v1beta1
        description: |2+

          Logging Operator is an operator created in Golang to set up and manage EFK(Elasticsearch, Fluentd, and Kibana) cluster inside Kubernetes and Openshift environment. This operator is capable of setting up each individual component of EFK cluster separately.
          ### Supported Features
          - Setup of elasticsearch cluster with different nodes type:- master, data, ingestion, and client.
          - Customizable configuration for elasticsearch, fluentd stack.
          - Setup of fluentd as light-weight log-shipper as Daemonset.
          - Kibana will be set up as visualization tool for elastic stack.
          - Seamless upgrades of elasticsearch, fluentd, and kibana.
          - Security best practices support for complete stack such as TLS, elastic security.
          - Kubernetes resources objects support like:- resources, securityContext, affinity, tolerations, etc.
          - Elasticsearch plugins and keystore seamless support with operator
          ### Documentation
          **[Documentation](https://ot-logging-operator.netlify.app/)**
          ### Getting Help
          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.
          ### Contributing
          You can contribute to this project by:-
          * Raising Bugs and Feature Requests
          * Fixing issue by opening a Pull Request
          * Improving Documentation
          ### License
          MongoDB Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/logging-operator/blob/master/LICENSE)

        displayName: Logging Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Operator
        - Elasticsearch
        - Fluentd
        - Kibana
        - Opstree
        - Logging
        - Monitoring
        - Cloud Native
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-logging-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep
        maturity: beta
        minKubeVersion: 1.17.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/opstree/logging-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: logging-operator.v0.4.0
        version: 0.4.0
      - name: logging-operator.v0.3.0
        version: 0.3.0
      name: beta
    defaultChannel: beta
    packageName: logging-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Google
      provider-url: https://cloud.google.com/alloydb/omni
    name: alloydb-omni-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: alloydb-omni-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "backupPlanRef": "backupplan1",
                  "dbclusterRef": "dbcluster-sample",
                  "manual": true,
                  "physicalBackupSpec": {
                    "backupType": "full"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "BackupPlan",
                "metadata": {
                  "name": "backupplan1"
                },
                "spec": {
                  "backupRetainDays": 14,
                  "backupSchedules": {
                    "full": "0 0 * * 0",
                    "incremental": "0 21 * * *"
                  },
                  "dbclusterRef": "dbcluster-sample",
                  "paused": false
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "clone1"
                },
                "spec": {
                  "clonedDBClusterConfig": {
                    "dbclusterName": "new-dbcluster-sample"
                  },
                  "pointInTime": "2024-02-23T19:59:43Z",
                  "sourceDBCluster": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "DBCluster",
                "metadata": {
                  "name": "dbcluster-sample"
                },
                "spec": {
                  "databaseVersion": "16.3.0",
                  "primarySpec": {
                    "adminUser": {
                      "passwordRef": {
                        "name": "db-pw-dbcluster-sample"
                      }
                    },
                    "resources": {
                      "cpu": 1,
                      "disks": [
                        {
                          "name": "DataDisk",
                          "size": "10Gi"
                        }
                      ],
                      "memory": "5Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "DBInstance",
                "metadata": {
                  "name": "dbcluster-sample-rp-1"
                },
                "spec": {
                  "dbcParent": {
                    "name": "dbcluster-sample"
                  },
                  "instanceType": "ReadPool",
                  "nodeCount": 2,
                  "resources": {
                    "cpu": 2,
                    "disks": [
                      {
                        "name": "DataDisk",
                        "size": "15Gi"
                      }
                    ],
                    "memory": "6Gi"
                  },
                  "schedulingconfig": {
                    "nodeaffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "preference": {
                            "matchExpressions": [
                              {
                                "key": "another-node-label-key",
                                "operator": "In",
                                "values": [
                                  "another-node-label-value"
                                ]
                              }
                            ]
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "podAffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "podAffinityTerm": {
                            "labelSelector": {
                              "matchExpressions": [
                                {
                                  "key": "app",
                                  "operator": "In",
                                  "values": [
                                    "store"
                                  ]
                                }
                              ]
                            },
                            "topologyKey": "kubernetes.io/hostname"
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "podAntiAffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "podAffinityTerm": {
                            "labelSelector": {
                              "matchExpressions": [
                                {
                                  "key": "security",
                                  "operator": "In",
                                  "values": [
                                    "S1"
                                  ]
                                }
                              ]
                            },
                            "topologyKey": "topology.kubernetes.io/zone"
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane",
                        "operator": "Exists"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Failover",
                "metadata": {
                  "name": "failover-sample"
                },
                "spec": {
                  "dbclusterRef": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "PgBouncer",
                "metadata": {
                  "name": "mypgbouncer"
                },
                "spec": {
                  "allowSuperUserAccess": true,
                  "dbclusterRef": "dbcluster-sample",
                  "parameters": {
                    "default_pool_size": "15",
                    "ignore_startup_parameters": "extra_float_digits",
                    "max_client_conn": "800",
                    "max_db_connections": "160",
                    "pool_mode": "transaction"
                  },
                  "podSpec": {
                    "image": "gcr.io/alloydb-omni/operator/g-pgbouncer:1.5.0",
                    "resources": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "replicaCount": 1,
                  "serviceOptions": {
                    "type": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Replication",
                "metadata": {
                  "name": "replication-downstream-sample"
                },
                "spec": {
                  "dbcluster": {
                    "name": "dbcluster-sample"
                  },
                  "downstream": {
                    "control": "setup",
                    "host": "10.10.10.10",
                    "password": {
                      "name": "ha-rep-pw-dbcluster-sample"
                    },
                    "port": 5432,
                    "replicationSlotName": "dbcluster_sample_replication_upstream_sample",
                    "username": "alloydbreplica"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Replication",
                "metadata": {
                  "name": "replication-upstream-sample"
                },
                "spec": {
                  "dbcluster": {
                    "name": "dbcluster-sample"
                  },
                  "upstream": {}
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backup": "backup1",
                  "sourceDBCluster": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Sidecar",
                "metadata": {
                  "name": "sidecar-sample"
                },
                "spec": {
                  "sidecars": [
                    {
                      "args": [
                        "-c",
                        "while [ true ]\ndo\ndate\nset -x\nls -lh /logs/diagnostic\nset +x\ndone\n"
                      ],
                      "command": [
                        "/bin/sh"
                      ],
                      "image": "busybox",
                      "name": "sidecar-sample",
                      "volumeMounts": [
                        {
                          "mountPath": "/logs",
                          "name": "obsdisk"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Switchover",
                "metadata": {
                  "name": "switchover-sample"
                },
                "spec": {
                  "dbclusterRef": "dbcluster-sample",
                  "newPrimary": "aaaa-dbcluster-sample"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: gcr.io/alloydb-omni/operator/fleet-operator:1.5.0
          createdAt: "2025-07-25T04:12:04Z"
          operatorframework.io/suggested-namespace: alloydb-omni-system
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["backuprepositories.alloydbomni.internal.dbadmin.goog",
            "failovers.alloydbomni.internal.dbadmin.goog", "instancebackupplans.alloydbomni.internal.dbadmin.goog","instancebackups.alloydbomni.internal.dbadmin.goog","instancerestores.alloydbomni.internal.dbadmin.goog","instances.alloydbomni.internal.dbadmin.goog","instanceswitchovers.alloydbomni.internal.dbadmin.goog","instanceuserdefinedauthentications.alloydbomni.internal.dbadmin.goog","lrojobs.alloydbomni.internal.dbadmin.goog","replicationconfigs.alloydbomni.internal.dbadmin.goog","sidecars.alloydbomni.internal.dbadmin.goog","createstandbyjobs.alloydbomni.internal.dbadmin.goog","deletestandbyjobs.alloydbomni.internal.dbadmin.goog"]'
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BackupPlan is the Schema for the backupplans API.
            displayName: Backup Plan
            kind: BackupPlan
            name: backupplans.alloydbomni.dbadmin.goog
            version: v1
          - description: Backup is the Schema for the backups API.
            displayName: Backup
            kind: Backup
            name: backups.alloydbomni.dbadmin.goog
            version: v1
          - description: CreateStandbyJob is an internal workflow tracking object.
              Users should not directly interact with this.
            displayName: (Internal) Create Standby Job
            kind: CreateStandbyJob
            name: createstandbyjobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: DBCluster is the Schema for the DBClusters API.
            displayName: DB Cluster
            kind: DBCluster
            name: dbclusters.alloydbomni.dbadmin.goog
            version: v1
          - description: DBInstance is the Schema for the DBInstances API.
            displayName: DB Instance
            kind: DBInstance
            name: dbinstances.alloydbomni.dbadmin.goog
            version: v1
          - description: DeleteStandbyJob is an internal workflow tracking object.
              Users should not directly interact with this.
            displayName: (Internal) Delete Standby Job
            kind: DeleteStandbyJob
            name: deletestandbyjobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Failover represents the parameters and status of a single
              failover operation.
            displayName: Failover
            kind: Failover
            name: failovers.alloydbomni.dbadmin.goog
            version: v1
          - description: PgBouncer is the Schema for the PgBouncer API.
            displayName: PgBouncer
            kind: PgBouncer
            name: pgbouncers.alloydbomni.dbadmin.goog
            version: v1
          - description: Replication is the Schema for the Replications API.
            displayName: Replication
            kind: Replication
            name: replications.alloydbomni.dbadmin.goog
            version: v1
          - description: Restore is the Schema for the Restores API.
            displayName: Restore
            kind: Restore
            name: restores.alloydbomni.dbadmin.goog
            version: v1
          - description: Sidecar is the Schema for the Sidecar API.
            displayName: Sidecar
            kind: Sidecar
            name: sidecars.alloydbomni.dbadmin.goog
            version: v1
          - description: Switchover represents the parameters and status of a single
              switchover.
            displayName: Switchover
            kind: Switchover
            name: switchovers.alloydbomni.dbadmin.goog
            version: v1
          - description: UserDefinedAuthentication is the Schema for the UserDefinedAuthentications
              API.
            displayName: User Defined Authentication
            kind: UserDefinedAuthentication
            name: userdefinedauthentications.alloydbomni.dbadmin.goog
            version: v1
          - description: BackupRepository is the Schema for the backuprepositories
              API.
            displayName: (Internal) Backup Repository
            kind: BackupRepository
            name: backuprepositories.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Failover is the Schema for the failover API.
            displayName: (Internal) Failover
            kind: Failover
            name: failovers.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceBackupPlan is the Schema for the InstanceBackupPlan
              API.
            displayName: (Internal) Instance Backup Plan
            kind: InstanceBackupPlan
            name: instancebackupplans.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceBackup is the Schema for the InstanceBackup API.
            displayName: (Internal) Instance Backup
            kind: InstanceBackup
            name: instancebackups.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceRestore is the Schema for the InstanceRestore API.
            displayName: (Internal) Instance Restore
            kind: InstanceRestore
            name: instancerestores.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Instance is the Schema for the Instance API.
            displayName: (Internal) Instance
            kind: Instance
            name: instances.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceSwitchover is the Schema for the InstanceSwitchover
              API.
            displayName: (Internal) Instance Switchover
            kind: InstanceSwitchover
            name: instanceswitchovers.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceUserDefinedAuthentication is the Schema for the InstanceUserDefinedAuthentication
              API.
            displayName: (Internal) Instance User Defined Authentication
            kind: InstanceUserDefinedAuthentication
            name: instanceuserdefinedauthentications.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: LROJob is the Schema for the LROJob API.
            displayName: (Internal) LRO Job
            kind: LROJob
            name: lrojobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: ReplicationConfig is the Schema for the ReplicationConfig
              API.
            displayName: (Internal) Replication Config
            kind: ReplicationConfig
            name: replicationconfigs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Sidecar is the Schema for the Sidecar API.
            displayName: (Internal) Sidecar
            kind: Sidecar
            name: sidecars.alloydbomni.internal.dbadmin.goog
            version: v1
        description: |-
          AlloyDB Omni is a database that lets you deploy a streamlined edition of AlloyDB for PostgreSQL in your own computing environment. While being 100% compatible with PostgreSQL applications and tools, AlloyDB Omni delivers higher performance for both OLTP and OLAP workloads and makes administration easy with a number of autopilots. AlloyDB AI provides powerful tools to build generative AI applications with the ability to create, store, and index vector embeddings right in your database.

          ## Post-installation

          After installing the operator, you need to manually create the cert-manager resources in your cluster as this is a strict dependency for our operator. For this, follow these steps:

          ```
          kubectl create ns ${NAMESPACE:?}
          ```

          ```
          kubectl apply -f - <<EOF
          apiVersion: cert-manager.io/v1
          kind: ClusterIssuer
          metadata:
            name: alloydbomni-selfsigned-cluster-issuer
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Issuer
          metadata:
            name: fleet-selfsigned-issuer
            namespace: ${NAMESPACE:?}
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Certificate
          metadata:
            name: fleet-serving-cert
            namespace: ${NAMESPACE:?}
          spec:
            dnsNames:
            - fleet-webhook-service.alloydb-omni-system.svc
            - fleet-webhook-service.alloydb-omni-system.svc.cluster.local
            issuerRef:
              kind: Issuer
              name: fleet-selfsigned-issuer
            secretName: fleet-webhook-server-cert
          ---
          apiVersion: cert-manager.io/v1
          kind: Issuer
          metadata:
            name: local-selfsigned-issuer
            namespace: ${NAMESPACE:?}
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Certificate
          metadata:
            name: local-serving-cert
            namespace: ${NAMESPACE:?}
          spec:
            dnsNames:
            - local-webhook-service.alloydb-omni-system.svc
            - local-webhook-service.alloydb-omni-system.svc.cluster.local
            issuerRef:
              kind: Issuer
              name: local-selfsigned-issuer
            secretName: local-webhook-server-cert
          EOF
          ```

          Replace the following:
          * `NAMESPACE` : the namespace where you have installed your operator. Suggested namespace: `alloydb-omni-system`.

          ## Creating a database cluster

          After you install the AlloyDB Omni operator on your Kubernetes cluster, you can create an AlloyDB Omni database cluster. For each database cluster you create, you must first create a password secret using the following format:

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: db-pw-${DB_CLUSTER_NAME:s?}
            namespace: ${DB_CLUSTER_NAMESPACE:?}
          type: Opaque
          data:
            ${DB_CLUSTER_NAME:?}: "ENCODED_PASSWORD"
          ```

          Replace the following:
          * `DB_CLUSTER_NAME`: the name of this database cluster. For example, `my-db-cluster`.
          * `DB_CLUSTER_NAMESPACE` (Optional): the namespace where you want to create the database cluster. For example, `my-db-cluster-namespace`.
          * `ENCODED_PASSWORD`: the database login password for the default postgres user role, encoded as a base64 string. For example, `Q2hhbmdlTWUxMjM=` for password `ChangeMe123`.

          Follow [these][database-cluster] steps for creating a database cluster.

          ## Connect to AlloyDB Omni running on Kubernetes
          Follow [these][connect] steps to connect to AlloyDB Omni.

          [database-cluster]: https://cloud.google.com/alloydb/omni/docs/deploy-kubernetes#create
          [connect]: https://cloud.google.com/alloydb/omni/docs/run-connect#connect-kubernetes
        displayName: AlloyDB Omni Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - sql
        - postgres
        - vector
        - ai
        - analytics
        - postgresql
        - alloydb
        links:
        - name: Documentation
          url: https://cloud.google.com/alloydb/omni/docs
        - name: Release Notes
          url: https://cloud.google.com/alloydb/docs/release-notes
        maintainers:
        - email: alloydb-omni-ext@google.com
          name: Google Cloud
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Google
          url: https://cloud.google.com/alloydb/omni
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.1
        - gcr.io/alloydb-omni/operator/local-operator:1.5.0
        - gcr.io/alloydb-omni/operator/fleet-operator:1.5.0
        version: 1.5.0
      entries:
      - name: alloydb-omni-operator.v1.5.0
        version: 1.5.0
      - name: alloydb-omni-operator.v1.4.1
        version: 1.4.1
      - name: alloydb-omni-operator.v1.4.0
        version: 1.4.0
      - name: alloydb-omni-operator.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: alloydb-omni-operator
    provider:
      name: Google
      url: https://cloud.google.com/alloydb/omni
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Trilio
      provider-url: https://trilio.io/
    name: tvo-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tvo-operator.6.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOControlPlane",
                "metadata": {
                  "name": "tvocontrolplane-sample"
                },
                "spec": {
                  "crdType": "TVOControlPlane",
                  "common": {
                    "dmapi_workers": "16",
                    "memcached_servers": "",
                    "my_ip": {
                      "host_interface": null
                    },
                    "nova_group_id": "42436",
                    "nova_user_id": "42436",
                    "oslo_enable_proxy_headers_parsing": false,
                    "trustee_role": "creator"
                  },
                  "database": {
                    "common": {
                      "host": "EDIT_DB_HOSTNAME",
                      "root_password": "EDIT_PASSWORD",
                      "root_user_name": "root"
                    },
                    "datamover_api": {
                      "database": "dmapi",
                      "password": "EDIT_PASSWORD",
                      "user": "dmapi"
                    },
                    "wlm_api": {
                      "database": "workloadmgr",
                      "password": "EDIT_PASSWORD",
                      "user": "workloadmgr"
                    }
                  },
                  "helm3_hook": true,
                  "images": {
                    "triliovault_datamover_api": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_datamover_db_sync": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_cloud_trust": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_cron": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_db_sync": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_scheduler": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_workloads": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1"
                  },
                  "keystone": {
                    "common": {
                      "admin_role_name": "admin",
                      "auth_uri": "",
                      "auth_url": "",
                      "ca_cert": "CERT MULTIPLE LINES\n",
                      "cloud_admin_domain_name": "Default",
                      "cloud_admin_project_name": "admin",
                      "cloud_admin_user_name": "admin",
                      "cloud_admin_user_password": "",
                      "is_self_signed_ssl_cert": true,
                      "keystone_auth_host": "",
                      "keystone_auth_port": "",
                      "keystone_auth_protocol": "",
                      "region_name": "regionOne",
                      "service_project_domain_name": "Default",
                      "service_project_name": "service"
                    },
                    "datamover_api": {
                      "internal_endpoint": "ENDPOINT_PROTOCOL_INTERNAL://ENDPOINT_HOST_INTERNAL:8784/v2",
                      "password": "EDIT_PASSWORD",
                      "public_auth_host": "",
                      "public_endpoint": "ENDPOINT_PROTOCOL_PUBLIC://ENDPOINT_HOST_PUBLIC:8784/v2",
                      "service_desc": "TrilioVault Datamover Api Service",
                      "service_name": "dmapi",
                      "service_type": "datamover",
                      "user": "dmapi"
                    },
                    "identity": null,
                    "keystone_interface": "internal",
                    "wlm_api": {
                      "internal_endpoint": "ENDPOINT_PROTOCOL_INTERNAL://ENDPOINT_HOST_INTERNAL:8781/v1/$(tenant_id)s",
                      "password": "EDIT_PASSWORD",
                      "public_auth_host": "",
                      "public_endpoint": "ENDPOINT_PROTOCOL_PUBLIC://ENDPOINT_HOST_PUBLIC:8781/v1/$(tenant_id)s",
                      "service_desc": "TrilioVault Workloadmgr Api Service",
                      "service_name": "TrilioVaultWLM",
                      "service_type": "workloads",
                      "user": "triliovault"
                    }
                  },
                  "labels": {
                    "datamover_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_cron": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_scheduler": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_workloads": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    }
                  },
                  "pod": {
                    "replicas": {
                      "triliovault_datamover_api": 3,
                      "triliovault_object_store": 3,
                      "triliovault_wlm_api": 3,
                      "triliovault_wlm_cron": 1,
                      "triliovault_wlm_scheduler": 3,
                      "triliovault_wlm_workloads": 3
                    },
                    "security_context": {
                      "triliovault": {
                        "container": {
                          "triliovault_datamover_api": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 0,
                            "runAsUser": 63630
                          },
                          "triliovault_datamover_api_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 63630,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_api": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_cron": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_scheduler": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_workloads": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "wlm_cloud_trust_creation": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 42436
                          }
                        }
                      }
                    }
                  },
                  "rabbitmq": {
                    "common": {
                      "admin_password": "EDIT_PASSWORD",
                      "admin_user": "admin",
                      "driver": "messagingv2",
                      "host": "EDIT_HOST",
                      "port": "5672"
                    },
                    "datamover_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "",
                      "user": "dmapi",
                      "vhost": "dmapi"
                    },
                    "wlm_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "",
                      "user": "workloadmgr",
                      "vhost": "workloadmgr"
                    }
                  },
                  "triliovault_backup_targets": [
                    {
                      "backup_target_name": "S3_BT1",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY1",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET1",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL1",
                      "s3_region_name": "REGION1",
                      "s3_secret_key": "SECRETKEY1",
                      "s3_self_signed_cert": false,
                      "s3_signature_version": "default",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "S3_BT2",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY2",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET2",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL2",
                      "s3_region_name": "REGION2",
                      "s3_secret_key": "SECRETKEY2",
                      "s3_self_signed_cert": true,
                      "s3_signature_version": "default",
                      "s3_ssl_ca_cert": "SSL_CERT_LINE_1\n",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "NFS_BT3",
                      "backup_target_type": "nfs",
                      "is_default": false,
                      "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10",
                      "nfs_shares": "11.30.1.10:/mnt/share"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:3d80cad27a1043fb61735e7a6fb156c98b269cfcc68f8b9712713a7d4f6ad6bd
          createdAt: "2025-01-14T11:20:30Z"
          description: Cloud-Native Data Protection for OpenStack
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trilio-openstack
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trilioData/triliovault-cfg-scripts
          support: Trilio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trilio for OpenStack Control Plane. Includes all control
              plane components like Datamover Api, Trilio object store and Workloadmgr
              services.
            displayName: Trilio for OpenStack Control Plane
            kind: TVOControlPlane
            name: tvocontrolplanes.tvo.trilio.io
            version: v1
          - description: Trilio for OpenStack Backup Target. This CR is responsible
              to add a new backup target.
            displayName: Trilio for OpenStack Backup Target
            kind: TVOBackupTarget
            name: tvobackuptargets.tvo.trilio.io
            version: v1
        description: |-
          Trilio for OpenStack provides scalable, native, agentless, multi-tenant backup and recovery solutions, ensuring seamless protection of workloads with flexible storage options for dynamic cloud environments.

            * OpenStack Backup and Recovery: Recover any OpenStack application quickly in the event of data loss or misconfiguration. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Such as One-Click-Restore to entire project recovery on different clusters.
            * Disaster Recovery: Recover any OpenStack application quickly in the event of a disaster. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Trilio makes OpenStack disaster recovery easy.

            * Migration: A tenant or administrator can capture an application and its data and non-disruptively migrate that exact point-in-time to another OpenStack cloud. The need to migrate OpenStack workloads can be driven by economics, security or test/dev.

            * Network Functions Virtualization backup and recovery: Business critical services like NFV can quickly be configured to any state, assuring the optimal quality of service. Trilio provides recovery from data damage and allows restoration of entire VNFs, selected VNFs, or individual data items using a point-in-time working copy. The platform also allows recovery from a disaster by restoring the OpenStack network topology and populating it with a working copy of the entire NFV.

          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Backup
        - Recovery
        - RHOSO Backup
        - Disaster Recovery
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/openstack-backup-and-recovery/
        - name: Operator Documentation
          url: https://docs.trilio.io/openstack
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        provider:
          name: Trilio
          url: https://trilio.io/
        relatedImages:
        - registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82
        - registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1
        - registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:3d80cad27a1043fb61735e7a6fb156c98b269cfcc68f8b9712713a7d4f6ad6bd
        version: 6.1.1
      entries:
      - name: tvo-operator.6.1.1
        version: 6.1.1
      name: 6.0.x
    - currentCSV: tvo-operator.6.1.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOControlPlane",
                "metadata": {
                  "name": "tvocontrolplane-sample",
                  "namespace": "trilio-openstack"
                },
                "spec": {
                  "common": {
                    "dmapi_workers": "16",
                    "memcached_servers": "memcached-0.memcached.openstack.svc:11212",
                    "nova_group_id": "42436",
                    "nova_user_id": "42436",
                    "oslo_enable_proxy_headers_parsing": false,
                    "trustee_role": "member"
                  },
                  "crdType": "TVOControlPlane",
                  "database": {
                    "common": {
                      "host": "openstack.openstack.svc",
                      "port": 3306,
                      "root_password": "openstack",
                      "root_user_name": "root"
                    },
                    "datamover_api": {
                      "database": "dmapi",
                      "password": "dsw21aSyH5gHH6XE72k11Y3t3C308jY5rP02ndwJW",
                      "user": "dmapi"
                    },
                    "wlm_api": {
                      "database": "workloadmgr",
                      "password": "dfeeYb72OwUzX5YuRKcVjxXJaELV8YT0jZczIQwys0v",
                      "user": "workloadmgr"
                    }
                  },
                  "helm3_hook": true,
                  "images": {
                    "triliovault_datamover_api": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_datamover_db_sync": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_cloud_trust": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_cron": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_db_sync": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_scheduler": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_workloads": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd"
                  },
                  "keystone": {
                    "common": {
                      "admin_role_name": "admin",
                      "auth_uri": "https://keystone-internal.openstack.svc:5000/v3",
                      "auth_url": "https://keystone-internal.openstack.svc:5000",
                      "cloud_admin_domain_name": "Default",
                      "cloud_admin_project_name": "admin",
                      "cloud_admin_user_name": "admin",
                      "cloud_admin_user_password": "openstack",
                      "is_self_signed_ssl_cert": true,
                      "keystone_auth_host": "keystone-internal.openstack.svc",
                      "keystone_auth_port": "5000",
                      "keystone_auth_protocol": "https",
                      "region_name": "regionOne",
                      "service_project_domain_name": "Default",
                      "service_project_name": "service"
                    },
                    "datamover_api": {
                      "internal_endpoint": "https://triliovault-datamover-internal.trilio-openstack.svc:8784/v2",
                      "password": "defajgh48zgt88wFzzhwEt3xHGeCOdExYbLGQN2XRm",
                      "public_auth_host": "triliovault-datamover-public-trilio-openstack.apps.trilio.trilio.bos2",
                      "public_endpoint": "https://triliovault-datamover-public-trilio-openstack.apps.trilio.trilio.bos2/v2",
                      "service_desc": "TrilioVault Datamover Api Service",
                      "service_name": "dmapi",
                      "service_type": "datamover",
                      "user": "dmapi"
                    },
                    "keystone_interface": "internal",
                    "wlm_api": {
                      "internal_endpoint": "https://triliovault-wlm-internal.trilio-openstack.svc:8781/v1/$(tenant_id)s",
                      "password": "dsdeJl8axy0gv14fURwUUxy0qUOQbugUUOfXikz6j",
                      "public_auth_host": "triliovault-wlm-public-trilio-openstack.apps.trilio.trilio.bos2",
                      "public_endpoint": "https://triliovault-wlm-public-trilio-openstack.apps.trilio.trilio.bos2/v1/$(tenant_id)s",
                      "service_desc": "TrilioVault Workloadmgr Api Service",
                      "service_name": "TrilioVaultWLM",
                      "service_type": "workloads",
                      "user": "triliovault"
                    }
                  },
                  "labels": {
                    "datamover_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_cron": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_scheduler": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_workloads": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    }
                  },
                  "pod": {
                    "replicas": {
                      "triliovault_datamover_api": 3,
                      "triliovault_object_store": 3,
                      "triliovault_wlm_api": 3,
                      "triliovault_wlm_cron": 1,
                      "triliovault_wlm_scheduler": 3,
                      "triliovault_wlm_workloads": 3
                    },
                    "security_context": {
                      "triliovault": {
                        "container": {
                          "triliovault_datamover_api": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 0,
                            "runAsUser": 63630
                          },
                          "triliovault_datamover_api_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 63630,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_api": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_cron": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_scheduler": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_workloads": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "wlm_cloud_trust_creation": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 42436
                          }
                        }
                      }
                    }
                  },
                  "rabbitmq": {
                    "common": {
                      "admin_password": "EDIT_PASSWORD",
                      "admin_user": "admin",
                      "driver": "messagingv2",
                      "host": "EDIT_HOST",
                      "port": "15671",
                      "ssl": true
                    },
                    "datamover_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "rabbit://dmapi:${PASSWORD}@${RABBITMQ_HOST}:5671/dmapi?ssl=1",
                      "user": "dmapi",
                      "vhost": "dmapi"
                    },
                    "wlm_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "rabbit://workloadmgr:${PASSWORD}@${RABBITMQ_HOST}:5671/workloadmgr?ssl=1",
                      "user": "workloadmgr",
                      "vhost": "workloadmgr"
                    }
                  },
                  "triliovault_backup_targets": [
                    {
                      "backup_target_name": "BT1_S3",
                      "backup_target_type": "s3",
                      "is_default": false,
                      "s3_access_key": "",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "",
                      "s3_region_name": "",
                      "s3_secret_key": "",
                      "s3_self_signed_cert": false,
                      "s3_signature_version": "default",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "BT2_Other_S3",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY2",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET2",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL2",
                      "s3_region_name": "REGION2",
                      "s3_secret_key": "SECRETKEY2",
                      "s3_self_signed_cert": true,
                      "s3_signature_version": "default",
                      "s3_ssl_ca_cert": "SSL_CERT_LINE_1\nSSL_CERT_LINE_2\nSSL_CERT_LINE_N\n",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "other_s3"
                    },
                    {
                      "backup_target_name": "BT3_NFS",
                      "backup_target_type": "nfs",
                      "is_default": false,
                      "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10",
                      "nfs_shares": "192.168.2.3:/nfs/share"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOBackupTarget",
                "metadata": {
                  "name": "tvobackuptarget-sample",
                  "namespace": "trilio-openstack"
                },
                "spec": {
                  "crdType": "TVOBackupTarget",
                  "images": {
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd"
                  },
                  "triliovault_backup_target": {
                    "backup_target_name": "S3_BT4",
                    "backup_target_type": "s3",
                    "is_default": false,
                    "s3_type": "other_s3",
                    "s3_access_key": "",
                    "s3_secret_key": "",
                    "s3_region_name": "",
                    "s3_bucket": "",
                    "s3_endpoint_url": "",
                    "s3_signature_version": "default",
                    "s3_auth_version": "DEFAULT",
                    "s3_bucket_object_lock_enabled": false,
                    "s3_ssl_enabled": true,
                    "s3_ssl_verify": true,
                    "s3_self_signed_cert": true,
                    "s3_ssl_ca_cert": "SSL_CERT_LINE_1\nSSL_CERT_LINE_2\nSSL_CERT_LINE_N",
                    "nfs_shares": "192.168.2.30:/nfs/share",
                    "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:e282bb056a1780e5170db3e63efdf5099a5ac5240ad44162bc7af83cac893ac7
          createdAt: "2025-01-14T11:20:30Z"
          description: Cloud-Native Data Protection for OpenStack
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: tvo-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trilioData/triliovault-cfg-scripts
          support: Trilio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trilio for OpenStack Control Plane. Includes all control
              plane components like Datamover Api, Trilio object store and Workloadmgr
              services.
            displayName: Trilio for OpenStack Control Plane
            kind: TVOControlPlane
            name: tvocontrolplanes.tvo.trilio.io
            version: v1
          - description: Trilio for OpenStack Backup Target. This CR is responsible
              to add a new backup target.
            displayName: Trilio for OpenStack Backup Target
            kind: TVOBackupTarget
            name: tvobackuptargets.tvo.trilio.io
            version: v1
        description: |-
          Trilio for OpenStack provides scalable, native, agentless, multi-tenant backup and recovery solutions, ensuring seamless protection of workloads with flexible storage options for dynamic cloud environments.

            * OpenStack Backup and Recovery: Recover any OpenStack application quickly in the event of data loss or misconfiguration. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Such as One-Click-Restore to entire project recovery on different clusters.
            * Disaster Recovery: Recover any OpenStack application quickly in the event of a disaster. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Trilio makes OpenStack disaster recovery easy.

            * Migration: A tenant or administrator can capture an application and its data and non-disruptively migrate that exact point-in-time to another OpenStack cloud. The need to migrate OpenStack workloads can be driven by economics, security or test/dev.

            * Network Functions Virtualization backup and recovery: Business critical services like NFV can quickly be configured to any state, assuring the optimal quality of service. Trilio provides recovery from data damage and allows restoration of entire VNFs, selected VNFs, or individual data items using a point-in-time working copy. The platform also allows recovery from a disaster by restoring the OpenStack network topology and populating it with a working copy of the entire NFV.

          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Backup
        - Recovery
        - RHOSO Backup
        - Disaster Recovery
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/openstack-backup-and-recovery/
        - name: Operator Documentation
          url: https://docs.trilio.io/openstack
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        provider:
          name: Trilio
          url: https://trilio.io/
        relatedImages:
        - registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd
        - registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:e282bb056a1780e5170db3e63efdf5099a5ac5240ad44162bc7af83cac893ac7
        - registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82
        version: 6.1.9
      entries:
      - name: tvo-operator.6.1.9
        version: 6.1.9
      name: 6.1.x
    defaultChannel: 6.0.x
    packageName: tvo-operator
    provider:
      name: Trilio
      url: https://trilio.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: elasticsearch-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: elasticsearch-operator.v5.8.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Elasticsearch",
                "metadata": {
                  "name": "elasticsearch"
                },
                "spec": {
                  "indexManagement": {
                    "mappings": [
                      {
                        "aliases": [
                          "infra",
                          "logs.infra"
                        ],
                        "name": "infra",
                        "policyRef": "infra-policy"
                      }
                    ],
                    "policies": [
                      {
                        "name": "infra-policy",
                        "phases": {
                          "delete": {
                            "minAge": "2d",
                            "namespaceSpec": [
                              {
                                "minAge": "5h",
                                "namespace": "openshift-monitoring"
                              }
                            ],
                            "pruneNamespacesInterval": "24h"
                          },
                          "hot": {
                            "actions": {
                              "rollover": {
                                "maxAge": "8h"
                              }
                            }
                          }
                        },
                        "pollInterval": "30m"
                      }
                    ]
                  },
                  "managementState": "Managed",
                  "nodeSpec": {
                    "resources": {
                      "limits": {
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "512Mi"
                      }
                    }
                  },
                  "nodes": [
                    {
                      "nodeCount": 1,
                      "roles": [
                        "client",
                        "data",
                        "master"
                      ],
                      "storage": {
                        "size": "20G"
                      }
                    }
                  ],
                  "redundancyPolicy": "ZeroRedundancy"
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Kibana",
                "metadata": {
                  "name": "kibana"
                },
                "spec": {
                  "managementState": "Managed",
                  "nodeSelector": {},
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "memory": "512Mi"
                    },
                    "requests": {
                      "memory": "512Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
          createdAt: "2024-09-19T15:14:33Z"
          description: |
            The Elasticsearch Operator for OCP provides a means for configuring and managing an Elasticsearch cluster for tracing and cluster logging.
            ## Prerequisites and Requirements
            ### Elasticsearch Operator Namespace
            The Elasticsearch Operator must be deployed to the global operator group namespace
            ### Memory Considerations
            Elasticsearch is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Elasticsearch cluster.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory. Each ES node can operate with a
            lower memory setting though this is not recommended for production deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.21'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Cluster Logging, Jaeger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Elasticsearch cluster instance
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.openshift.io
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.openshift.io
            version: v1
        description: "### IN OCP 4.16 AND FORWARD \n\nELASTICSEARCH IS ONLY SUPPORTED
          FOR **SERVICEMESH, TRACING, AND KIALI**, AND THIS OPERATOR **WILL BE REMOVED
          FROM THE OPENSHIFT OPERATOR CATALOG IN NOVEMBER 2025**.  ELASTICSEARCH IS
          **NO LONGER SUPPORTED FOR LOG STORAGE**. **KIBANA IS NO LONGER SUPPORTED**
          IN OCP 4.16 AND FORWARD. \n\n### About \n\nThe Elasticsearch Operator for
          OCP provides a means for configuring and managing an Elasticsearch cluster
          for use in tracing \nand cluster logging as well as a Kibana instance to
          connect to it.\nThis operator only supports OCP Cluster Logging and Jaeger.
          \ It is tightly coupled to each and is not currently capable of\nbeing used
          as a general purpose manager of Elasticsearch clusters running on OCP.\n\nPlease
          note: For a general purpose Elasticsearch operator, please use Elastic's
          Elasticsearch (ECK) Operator [here](https://catalog.redhat.com/software/containers/elastic/eck-operator/5fabf6d1ecb52450895164be?container-tabs=gti)\n\nIt
          is recommended that this operator be installed in the `openshift-operators-redhat`
          namespace to \nproperly support the Cluster Logging and Jaeger use cases.\n\nOnce
          installed, the operator provides the following features for **Elasticsearch**:\n*
          **Create/Destroy**: Deploy an Elasticsearch cluster to the same namespace
          in which the elasticsearch CR is created.\n* **Update**: Changes to the
          elasticsearch CR will be scheduled and applied to the cluster in a controlled
          manner (most often as a rolling upgrade).\n* **Cluster health**: The operator
          will periodically poll the cluster to evaluate its current health (such
          as the number of active shards and if any cluster nodes have reached their
          storage watermark usage).\n* **Redeploys**: In the case where the provided
          secrets are updated, the Elasticsearch Operator will schedule and perform
          a full cluster restart.\n* **Index management**: The Elasticsearch Operator
          will create cronjobs to perform index management such as roll over and deletion.\n\nOnce
          installed, the operator provides the following features for **Kibana**:\n*
          **Create/Destroy**: Deploy a Kibana instance to the same namespace in which
          the kibana CR is created (this should be the same namespace as the elasticsearch
          CR).\n* **Update**: Changes to the kibana CR will be scheduled and applied
          to the cluster in a controlled manner.\n* **Redeploys**: In the case where
          the provided secrets are updated, the Elasticsearch Operator will perform
          a restart.\n\n### Additionally provided features\n* Out of the box multitenancy
          that is integrated with OCP user access control.\n* Document Level Security\n*
          mTLS communication between Elasticsearch, Kibana, Index Management cronjobs,
          and CLO's Fluentd\n* OCP prometheus dashboard for Elasticsearch clusters\n*
          Prometheus Alerting rules  \n"
        displayName: OpenShift Elasticsearch Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - jaeger
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Elasticsearch Operator
          url: https://github.com/openshift/elasticsearch-operator
        maintainers:
        - email: aos-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8d603100978c655020a8f4d81c527162348f468d93989edbc00463b8297fb8a5
        - registry.redhat.io/openshift-logging/kibana6-rhel8@sha256:a2c200667738fdb079ba11ed3b0df3c5901373428d5e560bc1efea9d5e1ddd67
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
        - registry.redhat.io/openshift-logging/logging-curator5-rhel9@sha256:394e0f46fb8be44c720b29aed716443747a636d2758226c9386f930f95758470
        - registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:b26d2218aa6b1efea83a1717d62f81009c97e8082434b918937c609f5cdab912
        version: 5.8.21
      entries:
      - name: elasticsearch-operator.v5.8.21
        version: 5.8.21
      - name: elasticsearch-operator.v5.8.20
        version: 5.8.20
      - name: elasticsearch-operator.v5.8.19
        version: 5.8.19
      - name: elasticsearch-operator.v5.8.18
        version: 5.8.18
      - name: elasticsearch-operator.v5.8.17
        version: 5.8.17
      - name: elasticsearch-operator.v5.8.16
        version: 5.8.16
      - name: elasticsearch-operator.v5.8.15
        version: 5.8.15
      - name: elasticsearch-operator.v5.8.14
        version: 5.8.14
      - name: elasticsearch-operator.v5.8.13
        version: 5.8.13
      - name: elasticsearch-operator.v5.8.12
        version: 5.8.12
      - name: elasticsearch-operator.v5.8.11
        version: 5.8.11
      - name: elasticsearch-operator.v5.8.10
        version: 5.8.10
      - name: elasticsearch-operator.v5.8.9
        version: 5.8.9
      - name: elasticsearch-operator.v5.8.8
        version: 5.8.8
      - name: elasticsearch-operator.v5.8.7
        version: 5.8.7
      - name: elasticsearch-operator.v5.8.6
        version: 5.8.6
      - name: elasticsearch-operator.v5.8.5
        version: 5.8.5
      - name: elasticsearch-operator.v5.8.4
        version: 5.8.4
      - name: elasticsearch-operator.v5.8.3
        version: 5.8.3
      - name: elasticsearch-operator.v5.8.2
        version: 5.8.2
      - name: elasticsearch-operator.v5.8.1
        version: 5.8.1
      - name: elasticsearch-operator.v5.8.0
        version: 5.8.0
      name: stable
    - currentCSV: elasticsearch-operator.v5.8.22
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Elasticsearch",
                "metadata": {
                  "name": "elasticsearch"
                },
                "spec": {
                  "indexManagement": {
                    "mappings": [
                      {
                        "aliases": [
                          "infra",
                          "logs.infra"
                        ],
                        "name": "infra",
                        "policyRef": "infra-policy"
                      }
                    ],
                    "policies": [
                      {
                        "name": "infra-policy",
                        "phases": {
                          "delete": {
                            "minAge": "2d",
                            "namespaceSpec": [
                              {
                                "minAge": "5h",
                                "namespace": "openshift-monitoring"
                              }
                            ],
                            "pruneNamespacesInterval": "24h"
                          },
                          "hot": {
                            "actions": {
                              "rollover": {
                                "maxAge": "8h"
                              }
                            }
                          }
                        },
                        "pollInterval": "30m"
                      }
                    ]
                  },
                  "managementState": "Managed",
                  "nodeSpec": {
                    "resources": {
                      "limits": {
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "512Mi"
                      }
                    }
                  },
                  "nodes": [
                    {
                      "nodeCount": 1,
                      "roles": [
                        "client",
                        "data",
                        "master"
                      ],
                      "storage": {
                        "size": "20G"
                      }
                    }
                  ],
                  "redundancyPolicy": "ZeroRedundancy"
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Kibana",
                "metadata": {
                  "name": "kibana"
                },
                "spec": {
                  "managementState": "Managed",
                  "nodeSelector": {},
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "memory": "512Mi"
                    },
                    "requests": {
                      "memory": "512Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.12-v4.17
          containerImage: registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:ce1f01f7a7fd8fa3d52f6bca62f63f65aebcb4c9a8ff8f6afaca7e5d121f0406
          createdAt: "2025-09-23T09:54:59Z"
          description: |
            The Elasticsearch Operator for OCP provides a means for configuring and managing an Elasticsearch cluster for tracing and cluster logging.
            ## Prerequisites and Requirements
            ### Elasticsearch Operator Namespace
            The Elasticsearch Operator must be deployed to the global operator group namespace
            ### Memory Considerations
            Elasticsearch is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Elasticsearch cluster.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory. Each ES node can operate with a
            lower memory setting though this is not recommended for production deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.22'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-5.8
          operators.operatorframework.io.bundle.channels.v1: stable-5.8
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: elasticsearch-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Cluster Logging, Jaeger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Elasticsearch cluster instance
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.openshift.io
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.openshift.io
            version: v1
        description: |
          ### IN OCP 4.16 AND FORWARD

          IN OCP 4.16 AND FORWARD ELASTICSEARCH and KIBANA ARE **NO LONGER SUPPORTED
          FOR LOG STORAGE** AND ELASTICSEARCH IS ONLY SUPPORTED FOR **SERVICEMESH,
          TRACING, AND KIALI**

          THIS OPERATOR **WILL REACH END OF LIFE ON NOVEMBER 2025**

          ### About

          The Elasticsearch Operator for OCP provides a means for configuring and
          managing an Elasticsearch cluster for use in tracing and cluster logging as
          well as a Kibana instance to connect to it. This operator only supports OCP
          Cluster Logging and Jaeger.  It is tightly coupled to each and is not
          currently capable of being used as a general purpose manager of
          Elasticsearch clusters running on OCP.

          Please note: For a general purpose Elasticsearch operator, please use
          Elastic's Elasticsearch (ECK) Operator
          [here](https://catalog.redhat.com/software/containers/elastic/eck-operator/5fabf6d1ecb52450895164be?container-tabs=gti)

          It is recommended that this operator be installed in the
          `openshift-operators-redhat` namespace to properly support the Cluster
          Logging and Jaeger use cases.

          Once installed, the operator provides the following features for
          **Elasticsearch**: * **Create/Destroy**: Deploy an Elasticsearch cluster to
          the same namespace in which the elasticsearch CR is created.
          * **Update**: Changes to the elasticsearch CR will be scheduled and applied
          to the cluster in a controlled manner (most often as a rolling upgrade).
          * **Cluster health**: The operator will periodically poll the cluster to
          evaluate its current health (such as the number of active shards and if any
          cluster nodes have reached their storage watermark usage).
          * **Redeploys**: In the case where the provided secrets are updated, the
          Elasticsearch Operator will schedule and perform a full cluster restart.
          * **Index management**: The Elasticsearch Operator will create cronjobs to
          perform index management such as roll over and deletion.

          Once installed, the operator provides the following features for **Kibana**:

          * **Create/Destroy**: Deploy a Kibana instance to the same namespace in
          which the kibana CR is created (this should be the same namespace as the
          elasticsearch CR).
          * **Update**: Changes to the kibana CR will be scheduled and applied to the
          cluster in a controlled manner.
          * **Redeploys**: In the case where the provided secrets are updated, the
          Elasticsearch Operator will perform a restart.

          ### Additionally provided features

          * Out of the box multitenancy that is integrated with OCP user access
          control.
          * Document Level Security
          * mTLS communication between Elasticsearch, Kibana, Index Management
          cronjobs, and CLO's Fluentd
          * OCP prometheus dashboard for Elasticsearch clusters
          * Prometheus Alerting rules
        displayName: OpenShift Elasticsearch Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - jaeger
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Elasticsearch Operator
          url: https://github.com/openshift/elasticsearch-operator
        maintainers:
        - email: aos-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:6388562dc496e524a297c733720dea23a2889468fa25394812b3247f5f08c907
        - registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:69878c9e40178530baf837a28d3f78e63eb6fba54865dbf05f30dda1a661d8d0
        - registry.redhat.io/openshift-logging/kibana6-rhel8@sha256:7cfb21861cbc3c6d8c6b6f0192ac3c687e4ac9b16fbccf56841e179a08553c84
        - registry.redhat.io/openshift-logging/logging-curator5-rhel9@sha256:dd1cef6a91ca25231af46e0417c7929663098a3b4ad1230dddf4a8171871ce0e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:ce1f01f7a7fd8fa3d52f6bca62f63f65aebcb4c9a8ff8f6afaca7e5d121f0406
        version: 5.8.22
      entries:
      - name: elasticsearch-operator.v5.8.22
        version: 5.8.22
      - name: elasticsearch-operator.v5.8.21
        version: 5.8.21
      - name: elasticsearch-operator.v5.8.20
        version: 5.8.20
      - name: elasticsearch-operator.v5.8.19
        version: 5.8.19
      - name: elasticsearch-operator.v5.8.18
        version: 5.8.18
      - name: elasticsearch-operator.v5.8.17
        version: 5.8.17
      - name: elasticsearch-operator.v5.8.16
        version: 5.8.16
      - name: elasticsearch-operator.v5.8.15
        version: 5.8.15
      - name: elasticsearch-operator.v5.8.14
        version: 5.8.14
      - name: elasticsearch-operator.v5.8.13
        version: 5.8.13
      - name: elasticsearch-operator.v5.8.12
        version: 5.8.12
      - name: elasticsearch-operator.v5.8.11
        version: 5.8.11
      - name: elasticsearch-operator.v5.8.10
        version: 5.8.10
      - name: elasticsearch-operator.v5.8.9
        version: 5.8.9
      - name: elasticsearch-operator.v5.8.8
        version: 5.8.8
      - name: elasticsearch-operator.v5.8.7
        version: 5.8.7
      - name: elasticsearch-operator.v5.8.6
        version: 5.8.6
      - name: elasticsearch-operator.v5.8.5
        version: 5.8.5
      - name: elasticsearch-operator.v5.8.4
        version: 5.8.4
      - name: elasticsearch-operator.v5.8.3
        version: 5.8.3
      - name: elasticsearch-operator.v5.8.2
        version: 5.8.2
      - name: elasticsearch-operator.v5.8.1
        version: 5.8.1
      - name: elasticsearch-operator.v5.8.0
        version: 5.8.0
      name: stable-5.8
    defaultChannel: stable-5.8
    packageName: elasticsearch-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: application-services-metering-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: application-services-metering-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "applicationservices.redhat.com/v1",
                "kind": "Meter",
                "metadata": {
                  "name": "app-services-metrics-example"
                },
                "spec": {
                  "includeInfrastructure": false,
                  "meterCollectionEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Application Runtime
          certified: "false"
          containerImage: quay.io/redhat-developer/application-services-metering-operator:0.6.1
          createdAt: "2021-09-28T13:00:00Z"
          description: Collect the core usage of products from the Application Services
            portfolio into a single metric labeled by Runtimes, Integration, and Automation.
          repository: https://github.com/redhat-developer/application-services-metering-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Meter
            displayName: Meter
            kind: Meter
            name: meters.applicationservices.redhat.com
            version: v1
        description: |-
          ## About the managed application

          There is no managed application created with this operator.

          ## About this Operator

          The metering operator watches pods for predetermined labels, across a cluster or namespace(s).
          The label schema for Red Hat Application Services is in the format:

          * `com.company=Red_Hat`
          * `rht.comp=PAM`
          * `rht.comp_ver=7.10`
          * `rht.prod_name=Red_Hat_Process_Automation`
          * `rht.prod_ver=7.11`
          * `rht.subcomp=rhpam-kogito-runtime`
          * `rht.subcomp_t=application`

          It's important not to remove labels as the operator will cease monitoring those pods.

          The operator exposes a metric named: `appsvcs_cpu_usage_cores`
          The metric has a label for product category, these are Runtimes, Integration and Automation.
          There is a prometheus rule using the metric to group by product name, `appsvcs_cores_by_product:sum`.

          The operator utilises the OpenShift Monitoring Stack to provide usage observability via cloud.redhat.com

          ## Prerequisites for enabling this Operator

          None.
        displayName: Application Services Metering Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - metering
        - operator
        links:
        - name: Red Hat Application Services portfolio
          url: https://www.redhat.com/en/products/middleware
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/redhat-developer/application-services-metering-operator:0.6.1
        version: 0.6.1
      entries:
      - name: application-services-metering-operator.v0.6.1
        version: 0.6.1
      - name: application-services-metering-operator.v0.6.0
        version: 0.6.0
      name: alpha
    defaultChannel: alpha
    packageName: application-services-metering-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-licensing-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-licensing-operator.v8.0.220
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Developer"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:feeabfe4ecd36205306e3f3bf6c11517cfee735d7ae7eb4350abd01e9d76610b
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          olm.skipRange: <8.0.220
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:2f3de2bf46920cfe8c251011fc99841b43999561c1d3fe44850b205e94585a68
        version: 8.0.220
      entries:
      - name: infoscale-licensing-operator.v8.0.220
        version: 8.0.220
      - name: infoscale-licensing-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-licensing-operator.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-licensing-operator.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:b357b2203466516684c8fbc209757a98a45ec2376981f8699faf7dd5e37486a3
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:b357b2203466516684c8fbc209757a98a45ec2376981f8699faf7dd5e37486a3
        version: 8.0.330
      entries:
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-licensing-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Arctera
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Arctera
        - InfoScale
        - Licensing
        - Arctera Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
        version: 8.0.410
      entries:
      - name: infoscale-licensing-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-licensing-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-licensing-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-licensing-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Arctera
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Arctera
        - InfoScale
        - Licensing
        - Arctera Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
        version: 8.0.410
      entries:
      - name: infoscale-licensing-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-licensing-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-licensing-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-licensing-operator.v8.0.110
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Developer"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:e7b8d3c44df6151e61ded91336273e35eeaec04bd19124cf6313dc16d1224660
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:e7b8d3c44df6151e61ded91336273e35eeaec04bd19124cf6313dc16d1224660
        version: 8.0.110
      entries:
      - name: infoscale-licensing-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-licensing-operator.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-licensing-operator
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kedify
      provider-url: kedify.io
    name: kedify-agent-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kedify-agent-helm-operator.v0.2.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kedify.io/v1alpha1",
                "kind": "KedifyAgent",
                "metadata": {
                  "name": "kedifyagent-sample"
                },
                "spec": {
                  "agent": {
                    "agentId": "",
                    "apiKey": "kfy_xxxxxxxxxxxxxxxxxxxxx",
                    "createCrds": true,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v0.2.16@sha256:bfcc0b24ff75de17464ec213406f48bc211a932308e888c450ed23ee1786d37a"
                    },
                    "kedifyProxy": {
                      "globalValues": {
                        "openshift": true
                      },
                      "image": {
                        "tag": "v1.34.1@sha256:683e2f32809657c8d4a1508648545481130ac5409cd1ce63ccc795ad8c71202e"
                      }
                    },
                    "kubectlImage": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3"
                    },
                    "orgId": "00000000-0000-0000-0000-000000000000",
                    "securityContext": {
                      "readOnlyRootFilesystem": true
                    }
                  },
                  "keda": {
                    "enabled": true,
                    "image": {
                      "keda": {
                        "tag": "v2.17.2-0@sha256:802592b691a7b81a249055fa117994927c2fd376b5023a3888bcc7973c65ba52"
                      },
                      "metricsApiServer": {
                        "tag": "v2.17.2-0@sha256:df44bc1de75324158017d623a9d0adf3e3e9f1d64db6afa7133f777adc7b6389"
                      },
                      "pullPolicy": "IfNotPresent",
                      "webhooks": {
                        "tag": "v2.17.2-0@sha256:73e180bd5793dec3785fcb9ecd7e7184e645f7c37d561a5728970e5303c19a41"
                      }
                    }
                  },
                  "keda-add-ons-http": {
                    "enabled": true,
                    "interceptor": {
                      "pullPolicy": "IfNotPresent",
                      "scaledObject": {
                        "waitForCrd": true
                      }
                    },
                    "kubectlImage": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3"
                    },
                    "pinnedImages": {
                      "interceptor": {
                        "tag": "v0.10.0-17@sha256:30ccf20e79c6a62c14781872a73ad0a811063add7136b77df986bf403d6ca6bb"
                      },
                      "scaler": {
                        "tag": "v0.10.0-17@sha256:5863ee0ca09efeca225966c2eb8781f90f2dfb4aed872d84a6075b711ab042b2"
                      }
                    },
                    "scaler": {
                      "pullPolicy": "IfNotPresent"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: '{"kedify-openshift-default":{"description":"Default
            Kedify Agent CR for OpenShift. Register at https://dashboard.kedify.io
            to get started."}}'
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:d61332d50587ce2c75aff29033fdd5cb1bf0ec49d199ab72323a27d9d19ff402
          createdAt: "2025-08-07T14:04:11Z"
          description: Kedify Autoscaler, powered by proven open source technologies
            and KEDA, is an enterprise-grade service that allows teams to autoscale
            any workload using various event sources including HTTP events
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: keda
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: http://github.com/kedify/kedify-agent-helm-operator
          support: support@kedify.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KedifyAgent
            name: kedifyagents.charts.kedify.io
            version: v1alpha1
        description: Kedify Autoscaler, powered by proven open source technologies
          and KEDA, is an enterprise-grade service that allows teams to autoscale
          any workload using various event sources including HTTP events
        displayName: kedify-agent-helm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - autoscaling
        - keda
        - kedify
        links:
        - name: Kedify Homepage
          url: https://kedify.io
        maintainers:
        - email: support@kedify.io
          name: Kedify
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Kedify
          url: kedify.io
        relatedImages:
        - ghcr.io/kedify/agent:v0.2.16@sha256:9a57d5eafe368aab1d94f0eec3101e02f0ff2e356d8499984030ab89161e42ae
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:d61332d50587ce2c75aff29033fdd5cb1bf0ec49d199ab72323a27d9d19ff402
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:8e3a458912ec6540cb414446425ad7b1546528f04a5d84fcd64d8ee5424392e6
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:fc91d1f7f8b632614f31dea14c0c8d4a5f8311a83fdef82303857aa8b43b5782
        - envoyproxy/envoy:v1.34.1@sha256:007da57c2c328a90bd4e6d99b70bd899132f1b4a9426ccafe25437cf84a60c14
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:4b7277ad53dfb01bf5f2201dd31d94f24519da00d008638c3c36026b96ea78b4
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:9300a5382373b179f7f541a663cd30e50284e54f352993e8cca34460acbd890d
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:c1c093a6fda7efc9bc30f110c55f8bb075250db929dbdb2534b9071efbfc3169
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:30ccf20e79c6a62c14781872a73ad0a811063add7136b77df986bf403d6ca6bb
        - envoyproxy/envoy:v1.34.1@sha256:683e2f32809657c8d4a1508648545481130ac5409cd1ce63ccc795ad8c71202e
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:a38fdaf3a49eb4c71c4f3ab1211befa22fa5de56391f88efc11335f0bb3b1d29
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:82b041027dc900fdab717e8db6a522139067801c04cd1ebfe950572e714938ab
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:720b90a4f112a91a94b1bdafe445df1681657ddaa4fb20aa6d2904a0eb664a2a
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:5863ee0ca09efeca225966c2eb8781f90f2dfb4aed872d84a6075b711ab042b2
        - envoyproxy/envoy:v1.34.1@sha256:75545b91d050aa51b52cb0673be172f2f263c971dae3d2a82a70d128d92e87c2
        - ghcr.io/kedify/agent:v0.2.16@sha256:bfcc0b24ff75de17464ec213406f48bc211a932308e888c450ed23ee1786d37a
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:df44bc1de75324158017d623a9d0adf3e3e9f1d64db6afa7133f777adc7b6389
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:73e180bd5793dec3785fcb9ecd7e7184e645f7c37d561a5728970e5303c19a41
        - ghcr.io/kedify/agent:v0.2.16@sha256:c59d765091a31659236b8d80f4033912f5cf47d86c2e2da96bcfa133b9f6fe25
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:33cd5a8c6dd5e90614341dc1863e1f041f2c2475bd61edf5176a2a0d98ea54f9
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:850922994153104a8e382416561c251b46a95ff2b2a216aacbd7fcb4be7f6382
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:802592b691a7b81a249055fa117994927c2fd376b5023a3888bcc7973c65ba52
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:4b63409c022ba137760e7a534b6308fc1e3639f06fdb53ee1f6df32787b68400
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:ea6b22bb12c8ff7cbab4633c2f4a0e49eaf3883b07ab1f47dbebff2a3e07f029
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:9fcc76f983035eb0ceff7beb941bf32fe7fc28b341ec1aac5191ffc1471f401b
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:cb40290842976fd09f2008d8e2e64ebf90a215f0584656d4886122d322006873
        version: 0.2.16
      entries:
      - name: kedify-agent-helm-operator.v0.2.16
        version: 0.2.16
      name: stable
    defaultChannel: stable
    packageName: kedify-agent-helm-operator
    provider:
      name: Kedify
      url: kedify.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-operator.v2024.10.2-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "24.10.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "false",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": null
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": null
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:f4b9eb4f209f5428b0940056c760efcd9939553f3baac29924d1554701bc2dce
          createdAt: "2024-10-25T01:46:32Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2024.10.2
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2024.3.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "99.99.0.nextgen_uma_beta,3"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:f4b9eb4f209f5428b0940056c760efcd9939553f3baac29924d1554701bc2dce
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:7c4868db021beaffa762eb0de43739bbc865a4bf5900f1d17f78caed8612c073
        version: 2024.10.2-12
      entries:
      - name: uma-operator.v2024.10.2-12
        version: 2024.10.2-12
      name: "2024.10"
    - currentCSV: uma-operator.v2025.1.3-33
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.1.1.33,990033",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "false",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": null
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": null
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperDataQueueSize": 300,
                            "scrapperProcessingRoutines": 50,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:4a1bc124f22cf05dc6b42edbcc971d6309862e7be9c39b2ab6b1e3d50de61f3f
          createdAt: "2025-02-18T05:33:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.1.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2024.3.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       scrapperProcessingRoutines: 50
                       scrapperDataQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.1.1.26,990026"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:4a1bc124f22cf05dc6b42edbcc971d6309862e7be9c39b2ab6b1e3d50de61f3f
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:ae91a28d57c8a359f81757e4bf79472f2ac189ac0ae3ee0d87e8ff9c433f1ccc
        version: 2025.1.3-33
      entries:
      - name: uma-operator.v2025.1.3-33
        version: 2025.1.3-33
      name: "2025.1"
    - currentCSV: uma-operator.v2025.2.1-25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.2.1.25,990025",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mixedMode": "false",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperDataQueueSize": 300,
                            "scrapperProcessingRoutines": 50,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:d8f4ca750252f1570a2dd89cabf018cc4f0b2c0cd6cadfedbf48507a34d4aee9
          createdAt: "2025-03-05T05:49:42Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.2.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2024.3.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       scrapperProcessingRoutines: 50
                       scrapperDataQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.1.1.26,990026"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:d8f4ca750252f1570a2dd89cabf018cc4f0b2c0cd6cadfedbf48507a34d4aee9
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:8bcf0dac889f38a0bf77eb9babe3f726e3b2564d5988c05bdc5c2d91ae189c61
        version: 2025.2.1-25
      entries:
      - name: uma-operator.v2025.2.1-25
        version: 2025.2.1-25
      name: "2025.2"
    - currentCSV: uma-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:8ff4aab6d0043ccd2542a86b7e9c24d60670a3c376021ef2d02fc1a289130275
          createdAt: "2025-04-16T06:54:48Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.4.5
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.4.2.5,990005"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:8ff4aab6d0043ccd2542a86b7e9c24d60670a3c376021ef2d02fc1a289130275
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        version: 2025.4.5-5
      entries:
      - name: uma-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:90b4b21d93bce41f2530699cec01f03e5daa73bfbf4f30cbcc8fbd6e75e97da6
          createdAt: "2025-07-24T15:07:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.5.2.12,990012""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:90b4b21d93bce41f2530699cec01f03e5daa73bfbf4f30cbcc8fbd6e75e97da6
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        version: 2025.5.4-12
      entries:
      - name: uma-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-operator.v2025.7.4-23
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:35f74109aa1916524c1fb483972caa31cb84381c3cff7c3b26fd9b676f9d748c
          createdAt: "2025-08-25T09:00:19Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.3
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.7.1.23,990023"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:35f74109aa1916524c1fb483972caa31cb84381c3cff7c3b26fd9b676f9d748c
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.4-23
      entries:
      - name: uma-operator.v2025.7.4-23
        version: 2025.7.4-23
      name: "2025.7"
    - currentCSV: uma-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:adea186632253865406e95a0062abbb312dee23c92ff108f5914e325dc0b8833
          createdAt: "2025-08-15T03:19:46Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.8.1.9,990009"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:adea186632253865406e95a0062abbb312dee23c92ff108f5914e325dc0b8833
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        version: 2025.8.2-9
      entries:
      - name: uma-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
          createdAt: "2025-09-24T12:22:41Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.12,990012"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:c892787785c8c5e0de1692709a9e3f54eef36f1f00c53f1b9feabb613769a48c
          createdAt: "2025-07-30T08:06:11Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "99.99.0.serp_F152813,19"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:c892787785c8c5e0de1692709a9e3f54eef36f1f00c53f1b9feabb613769a48c
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        version: 2025.7.2-19
      entries:
      - name: uma-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:dc15dfdc8b82ee159b52036cc9d057e6525b08c478ed759939cf5c3fc7daeeec
          createdAt: "2025-09-11T11:15:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.8,990008"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:dc15dfdc8b82ee159b52036cc9d057e6525b08c478ed759939cf5c3fc7daeeec
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
          createdAt: "2025-09-24T12:22:41Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.12,990012"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: EDB
      provider-url: ""
    name: cloud-native-postgresql
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloud-native-postgresql.v1.27.0-rc1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:5f5bebbf4d8021dafd0fb7485d629b30aaa362e899e69eb52fa5e74757af6e36
          createdAt: "2025-07-30T09:28:05Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:5f5bebbf4d8021dafd0fb7485d629b30aaa362e899e69eb52fa5e74757af6e36
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        version: 1.27.0-rc1
      entries:
      - name: cloud-native-postgresql.v1.27.0-rc1
        version: 1.27.0-rc1
      - name: cloud-native-postgresql.v1.26.0-rc3
        version: 1.26.0-rc3
      - name: cloud-native-postgresql.v1.26.0-rc2
        version: 1.26.0-rc2
      - name: cloud-native-postgresql.v1.26.0-rc1
        version: 1.26.0-rc1
      name: candidate
    - currentCSV: cloud-native-postgresql.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
          createdAt: "2025-08-19T12:00:36Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
        version: 1.27.0
      entries:
      - name: cloud-native-postgresql.v1.27.0
        version: 1.27.0
      - name: cloud-native-postgresql.v1.26.1
        version: 1.26.1
      - name: cloud-native-postgresql.v1.26.0
        version: 1.26.0
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      - name: cloud-native-postgresql.v1.24.1
        version: 1.24.1
      - name: cloud-native-postgresql.v1.24.0
        version: 1.24.0
      - name: cloud-native-postgresql.v1.23.3
        version: 1.23.3
      - name: cloud-native-postgresql.v1.23.2
        version: 1.23.2
      - name: cloud-native-postgresql.v1.23.1
        version: 1.23.1
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      - name: cloud-native-postgresql.v1.21.1
        version: 1.21.1
      - name: cloud-native-postgresql.v1.21.0
        version: 1.21.0
      - name: cloud-native-postgresql.v1.20.2
        version: 1.20.2
      - name: cloud-native-postgresql.v1.20.1
        version: 1.20.1
      - name: cloud-native-postgresql.v1.20.0
        version: 1.20.0
      - name: cloud-native-postgresql.v1.19.1
        version: 1.19.1
      - name: cloud-native-postgresql.v1.19.0
        version: 1.19.0
      name: fast
    - currentCSV: cloud-native-postgresql.v1.25.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
          createdAt: "2025-07-26T06:26:53Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.25.3
      entries:
      - name: cloud-native-postgresql.v1.25.3
        version: 1.25.3
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      - name: cloud-native-postgresql.v1.22.9
        version: 1.22.9
      - name: cloud-native-postgresql.v1.22.8
        version: 1.22.8
      - name: cloud-native-postgresql.v1.22.7
        version: 1.22.7
      - name: cloud-native-postgresql.v1.22.6
        version: 1.22.6
      - name: cloud-native-postgresql.v1.22.5
        version: 1.22.5
      - name: cloud-native-postgresql.v1.22.4
        version: 1.22.4
      - name: cloud-native-postgresql.v1.22.3
        version: 1.22.3
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      - name: cloud-native-postgresql.v1.18.13
        version: 1.18.13
      - name: cloud-native-postgresql.v1.18.12
        version: 1.18.12
      - name: cloud-native-postgresql.v1.18.11
        version: 1.18.11
      - name: cloud-native-postgresql.v1.18.10
        version: 1.18.10
      - name: cloud-native-postgresql.v1.18.9
        version: 1.18.9
      - name: cloud-native-postgresql.v1.18.8
        version: 1.18.8
      - name: cloud-native-postgresql.v1.18.7
        version: 1.18.7
      name: stable
    - currentCSV: cloud-native-postgresql.v1.18.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:69e439a2a961881309cc122174eaca273cf6d7873e9e871efbdb2d60804f40e4
          createdAt: "2024-06-13T16:08:55Z"
          description: EDB Postgres for Kubernetes is an open source operator designed
            to manage highly available PostgreSQL databases with a primary/standby
            architecture on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 1.18.0 < 1.18.13'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloud
        - native
        - edb
        - enterprisedb
        - cnp
        - pg4k
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: john.long@enterprisedb.com
          name: John Long
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: The EDB Postgres for Kubernetes Contributors
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:69e439a2a961881309cc122174eaca273cf6d7873e9e871efbdb2d60804f40e4
        version: 1.18.13
      entries:
      - name: cloud-native-postgresql.v1.18.13
        version: 1.18.13
      - name: cloud-native-postgresql.v1.18.12
        version: 1.18.12
      - name: cloud-native-postgresql.v1.18.11
        version: 1.18.11
      - name: cloud-native-postgresql.v1.18.10
        version: 1.18.10
      - name: cloud-native-postgresql.v1.18.9
        version: 1.18.9
      - name: cloud-native-postgresql.v1.18.8
        version: 1.18.8
      - name: cloud-native-postgresql.v1.18.7
        version: 1.18.7
      name: stable-v1.18
    - currentCSV: cloud-native-postgresql.v1.19.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:2a717edd0523798bd4d0b14c557b8d6b19ae3ad1f019c0cc4f61db49f10c3dad
          createdAt: "2023-11-08T19:02:47Z"
          description: Operator to manage Postgres high availability clusters with
            a primary/standby architecture.
          olm.skipRange: '>=1.15.3 < 1.19.6'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup manager
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster manager
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PgBouncer Pooler Manager
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL backups schedule
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: "EDB Postgres for Kubernetes is an operator designed, developed,
          and supported by EDB that covers the full \nlifecycle of a highly available
          Postgres database clusters with a primary/standby architecture, using native\nstreaming
          replication. It is based on the open source CloudNativePG operator, and
          provides additional value\nsuch as compatibility with Oracle using EDB Postgres
          Advanced Server and additional supported platforms such\nas IBM Power and
          OpenShift.\n\nKey features available include:\n\n* Kubernetes API integration
          for high availability\n* Self-healing through failover and automated recreation
          of replicas\n* Capacity management with scale up/down capabilities\n* Planned
          switchovers for scheduled maintenance\n* Read-only and read-write Kubernetes
          services definitions\n* Rolling updates for Postgres minor versions and
          operator upgrades\n* Continuous backup and point-in-time recovery\n* Connection
          Pooling with PgBouncer\n* Integrated metrics exporter out of the box\n*
          PostgreSQL replication across multiple Kubernetes clusters\n* Red Hat certified
          operator for OpenShift\n\nThe operator has been renamed from Cloud Native
          PostgreSQL. Existing users of Cloud Native PostgreSQL will not\nexperience
          any change, as the underlying components and resources have not changed.\n"
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloud
        - native
        - edb
        - enterprisedb
        - cnp
        - pg4k
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: sales@enterprisedb.com
          name: Sales
        - email: support@enterprisedb.com
          name: Support
        maturity: stable
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:2a717edd0523798bd4d0b14c557b8d6b19ae3ad1f019c0cc4f61db49f10c3dad
        version: 1.19.6
      entries:
      - name: cloud-native-postgresql.v1.19.6
        version: 1.19.6
      - name: cloud-native-postgresql.v1.19.5
        version: 1.19.5
      - name: cloud-native-postgresql.v1.19.4
        version: 1.19.4
      - name: cloud-native-postgresql.v1.19.3
        version: 1.19.3
      - name: cloud-native-postgresql.v1.19.2
        version: 1.19.2
      - name: cloud-native-postgresql.v1.19.1
        version: 1.19.1
      - name: cloud-native-postgresql.v1.19.0
        version: 1.19.0
      name: stable-v1.19
    - currentCSV: cloud-native-postgresql.v1.20.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:919c6669f8bfa333b19fdd1ca03aaeceebe0c5ca74d91cb66b2369680422e916
          createdAt: "2024-02-03T17:26:23Z"
          description: Operator to manage Postgres high availability clusters with
            a primary/standby architecture.
          olm.skipRange: '>=1.15.3 < 1.20.6'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup manager
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster manager
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PgBouncer Pooler Manager
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL backups schedule
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: "EDB Postgres for Kubernetes is an operator designed, developed,
          and supported by EDB that covers the full \nlifecycle of a highly available
          Postgres database clusters with a primary/standby architecture, using native\nstreaming
          replication. It is based on the open source CloudNativePG operator, and
          provides additional value\nsuch as compatibility with Oracle using EDB Postgres
          Advanced Server and additional supported platforms such\nas IBM Power and
          OpenShift.\n\nKey features available include:\n\n* Kubernetes API integration
          for high availability\n* Self-healing through failover and automated recreation
          of replicas\n* Capacity management with scale up/down capabilities\n* Planned
          switchovers for scheduled maintenance\n* Read-only and read-write Kubernetes
          services definitions\n* Rolling updates for Postgres minor versions and
          operator upgrades\n* Continuous backup and point-in-time recovery\n* Connection
          Pooling with PgBouncer\n* Integrated metrics exporter out of the box\n*
          PostgreSQL replication across multiple Kubernetes clusters\n* Red Hat certified
          operator for OpenShift\n\nThe operator has been renamed from Cloud Native
          PostgreSQL. Existing users of Cloud Native PostgreSQL will not\nexperience
          any change, as the underlying components and resources have not changed.\n"
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloud
        - native
        - edb
        - enterprisedb
        - cnp
        - pg4k
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: sales@enterprisedb.com
          name: Sales
        - email: support@enterprisedb.com
          name: Support
        maturity: stable
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:919c6669f8bfa333b19fdd1ca03aaeceebe0c5ca74d91cb66b2369680422e916
        version: 1.20.6
      entries:
      - name: cloud-native-postgresql.v1.20.6
        version: 1.20.6
      - name: cloud-native-postgresql.v1.20.5
        version: 1.20.5
      - name: cloud-native-postgresql.v1.20.4
        version: 1.20.4
      - name: cloud-native-postgresql.v1.20.3
        version: 1.20.3
      - name: cloud-native-postgresql.v1.20.2
        version: 1.20.2
      - name: cloud-native-postgresql.v1.20.1
        version: 1.20.1
      - name: cloud-native-postgresql.v1.20.0
        version: 1.20.0
      name: stable-v1.20
    - currentCSV: cloud-native-postgresql.v1.21.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:8cc328b1fc8efa3911d3dba689bc824ff789b305f7eb3284bf21ffcdc52a9683
          createdAt: "2024-06-13T16:41:26Z"
          description: EDB Postgres for Kubernetes is an open source operator designed
            to manage highly available PostgreSQL databases with a primary/standby
            architecture on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 1.18.0 < 1.21.6'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: john.long@enterprisedb.com
          name: John Long
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: The EDB Postgres for Kubernetes Contributors
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:8cc328b1fc8efa3911d3dba689bc824ff789b305f7eb3284bf21ffcdc52a9683
        version: 1.21.6
      entries:
      - name: cloud-native-postgresql.v1.21.6
        version: 1.21.6
      - name: cloud-native-postgresql.v1.21.5
        version: 1.21.5
      - name: cloud-native-postgresql.v1.21.4
        version: 1.21.4
      - name: cloud-native-postgresql.v1.21.3
        version: 1.21.3
      - name: cloud-native-postgresql.v1.21.2
        version: 1.21.2
      - name: cloud-native-postgresql.v1.21.1
        version: 1.21.1
      - name: cloud-native-postgresql.v1.21.0
        version: 1.21.0
      name: stable-v1.21
    - currentCSV: cloud-native-postgresql.v1.22.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:4d55ad54cc7e7807bc99f5a6f96bc92e88b4060f424ffd2cb6552be9505f38f4
          createdAt: "2025-07-25T19:02:56Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images. Only
              available in 1.25 or later
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster.
              Only available in 1.25 or later
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster. Only available in 1.25 or later
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication.
              Only available in 1.25 or later
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:4d55ad54cc7e7807bc99f5a6f96bc92e88b4060f424ffd2cb6552be9505f38f4
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.22.11
      entries:
      - name: cloud-native-postgresql.v1.22.11
        version: 1.22.11
      - name: cloud-native-postgresql.v1.22.10
        version: 1.22.10
      - name: cloud-native-postgresql.v1.22.9
        version: 1.22.9
      - name: cloud-native-postgresql.v1.22.8
        version: 1.22.8
      - name: cloud-native-postgresql.v1.22.7
        version: 1.22.7
      - name: cloud-native-postgresql.v1.22.6
        version: 1.22.6
      - name: cloud-native-postgresql.v1.22.5
        version: 1.22.5
      - name: cloud-native-postgresql.v1.22.4
        version: 1.22.4
      - name: cloud-native-postgresql.v1.22.3
        version: 1.22.3
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      name: stable-v1.22
    - currentCSV: cloud-native-postgresql.v1.23.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:934e01ed5cd35cf681414b9136a337514a1b73c0110871f1c78ba98447badb0f
          createdAt: "2024-12-23T16:37:54Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 1.18.0 < 1.23.6'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:934e01ed5cd35cf681414b9136a337514a1b73c0110871f1c78ba98447badb0f
        version: 1.23.6
      entries:
      - name: cloud-native-postgresql.v1.23.6
        version: 1.23.6
      - name: cloud-native-postgresql.v1.23.5
        version: 1.23.5
      - name: cloud-native-postgresql.v1.23.4
        version: 1.23.4
      - name: cloud-native-postgresql.v1.23.3
        version: 1.23.3
      - name: cloud-native-postgresql.v1.23.2
        version: 1.23.2
      - name: cloud-native-postgresql.v1.23.1
        version: 1.23.1
      name: stable-v1.23
    - currentCSV: cloud-native-postgresql.v1.24.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:a1acaaf5a413db0f33f3824a15aefe598974d1593172bcbccd0fc3d39d7b46d0
          createdAt: "2025-05-23T15:20:37Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:a1acaaf5a413db0f33f3824a15aefe598974d1593172bcbccd0fc3d39d7b46d0
        version: 1.24.4
      entries:
      - name: cloud-native-postgresql.v1.24.4
        version: 1.24.4
      - name: cloud-native-postgresql.v1.24.3
        version: 1.24.3
      - name: cloud-native-postgresql.v1.24.2
        version: 1.24.2
      - name: cloud-native-postgresql.v1.24.1
        version: 1.24.1
      - name: cloud-native-postgresql.v1.24.0
        version: 1.24.0
      name: stable-v1.24
    - currentCSV: cloud-native-postgresql.v1.25.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
          createdAt: "2025-07-26T06:26:53Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
        version: 1.25.3
      entries:
      - name: cloud-native-postgresql.v1.25.3
        version: 1.25.3
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      name: stable-v1.25
    - currentCSV: cloud-native-postgresql.v1.26.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:47d70233908ea855ddab3ff79c7a226719c3f2be5b8a5018c7934cd44bf0f546
          createdAt: "2025-07-26T07:59:14Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:47d70233908ea855ddab3ff79c7a226719c3f2be5b8a5018c7934cd44bf0f546
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.26.1
      entries:
      - name: cloud-native-postgresql.v1.26.1
        version: 1.26.1
      - name: cloud-native-postgresql.v1.26.0
        version: 1.26.0
      name: stable-v1.26
    - currentCSV: cloud-native-postgresql.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
          createdAt: "2025-08-19T12:00:36Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.27.0
      entries:
      - name: cloud-native-postgresql.v1.27.0
        version: 1.27.0
      name: stable-v1.27
    defaultChannel: fast
    packageName: cloud-native-postgresql
    provider:
      name: EDB
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: http://marketplace.redhat.com
    name: marketplace-games-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marketplace-games-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "game.marketplace.redhat.com/v1",
                "kind": "Arcade",
                "metadata": {
                  "name": "arcade-sample"
                },
                "spec": {
                  "size": 1
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/marketplace-games-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/marketplace-games-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Arcade is the Schema for the arcades API
            displayName: Arcade Instance
            kind: Arcade
            name: arcades.game.marketplace.redhat.com
            version: v1
        description: RHM Arcade Operator - a simple operator used to showcase how
          to build an operator around an existing web application that can can run
          on Openshift, Vanilla k8s, or on-boarded to Red Hat Marketplace.
        displayName: Arcade Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - rhm
        - red hat
        - marketplace
        - games
        - arcade
        links:
        - name: Marketplace Games Operator
          url: https://github.com/redhat-marketplace/marketplace-games-operator
        maturity: alpha
        provider:
          name: Red Hat Marketplace
          url: http://marketplace.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/rhm-labs/arcade-operator@sha256:d6959be6693728d2dc4cd38b69251551d23c5002cb52db2684df31a0a8f7b0ab
        version: 1.0.0
      entries:
      - name: marketplace-games-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: marketplace-games-operator-rhmp
    provider:
      name: Red Hat Marketplace
      url: http://marketplace.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/
    name: varnish-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: varnish-operator.v0.27.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "caching.ibm.com/v1alpha1",
                "kind": "VarnishCluster",
                "metadata": {
                  "name": "varnishcluster-sample"
                },
                "spec": {
                  "backend": {
                    "port": "web",
                    "selector": {
                      "app": "nginx"
                    }
                  },
                  "replicas": 1,
                  "service": {
                    "port": 80
                  },
                  "varnish": {
                    "args": [
                      "-p",
                      "default_ttl=3600",
                      "-p",
                      "default_grace=3600"
                    ]
                  },
                  "vcl": {
                    "configMapName": "vcl-files",
                    "entrypointFileName": "entrypoint.vcl"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: ibmcom/varnish-operator:0.27.2
          createdAt: 2021-02-15T11:15:31+0000
          description: Run and manage Varnish clusters on Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/varnish-operator
          support: https://ibm.github.io/varnish-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VarnishCluster is the Schema for the varnishclusters API
            displayName: Varnish Cluster
            kind: VarnishCluster
            name: varnishclusters.caching.ibm.com
            version: v1alpha1
        description: |
          Run and manage Varnish clusters on Kubernetes

          The operator manages the whole lifecycle of the cluster: creating, deleting and keeping the cluster configuration up to date.

          The operator is responsible for building the VCL configuration using templates defined by the users and keeping the configuration up to date when relevant events occur (backend pod failure, scaling of the deployment, VCL configuration change).

          Example of a simple `VarnishCluster`:

          ```yaml
          apiVersion: caching.ibm.com/v1alpha1
          kind: VarnishCluster
          metadata:
            name: varnishcluster-sample
            namespace: varnish-ns
          spec:
            vcl:
              configMapName: vcl-files
              entrypointFileName: entrypoint.vcl
            replicas: 3
            backend:
              selector:
                app: nginx
            service:
              port: 80
          ```

          Full documentation can be found [here](https://ibm.github.io/varnish-operator/)
        displayName: Varnish Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - varnish
        - cache
        links:
        - name: Varnish Operator Docs
          url: https://ibm.github.io/varnish-operator
        - name: Varnish Operator Source Code
          url: https://github.com/IBM/varnish-operator
        maintainers:
        - email: tomash.sidei@ibm.com
          name: Tomash Sidei
        - email: cingram@us.ibm.com
          name: Craig Ingram
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/
        relatedImages:
        - ibmcom/varnish-operator:0.27.2
        version: 0.27.2
      entries:
      - name: varnish-operator.v0.27.2
        version: 0.27.2
      name: alpha
    defaultChannel: alpha
    packageName: varnish-operator
    provider:
      name: IBM
      url: https://www.ibm.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CloudBees
      provider-url: https://www.cloudbees.com/products/continuous-integration
    name: cloudbees-ci
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudbees-ci.v3.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.cloudbees.com/v1alpha1",
                "kind": "CloudBeesCI",
                "metadata": {
                  "name": "cloudbeesci-sample"
                },
                "spec": {
                  "Agents": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-agent:2.289.3.2"
                    },
                    "ImagePullSecrets": null,
                    "SeparateNamespace": {
                      "Create": false,
                      "Enabled": false,
                      "Name": null
                    }
                  },
                  "Hibernation": {
                    "Enabled": false,
                    "Image": {
                      "dockerImage": "cloudbees/managed-master-hibernation-monitor:247.c5dfce00a179",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "NodeSelector": {},
                    "Tolerations": []
                  },
                  "Master": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-mm:2.289.3.2"
                    },
                    "JavaOpts": null,
                    "OperationsCenterNamespace": null
                  },
                  "NetworkPolicy": {
                    "Enabled": false,
                    "JMXSelectors": [],
                    "ingressControllerSelector": []
                  },
                  "OperationsCenter": {
                    "AgentListenerPort": 50000,
                    "Annotations": {},
                    "CSRF": {
                      "ProxyCompatibility": false
                    },
                    "CasC": {
                      "ConfigMapName": "oc-casc-bundle",
                      "Enabled": false
                    },
                    "ContainerPort": 8080,
                    "ContextPath": null,
                    "Enabled": true,
                    "ExtraConfigMaps": [],
                    "ExtraContainers": [],
                    "ExtraGroovyConfiguration": {},
                    "ExtraVolumeMounts": [],
                    "ExtraVolumes": [],
                    "HealthProbeLivenessFailureThreshold": 12,
                    "HealthProbes": true,
                    "HostName": null,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-cloud-core-oc:2.289.3.2",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "Ingress": {
                      "Annotations": {
                        "kubernetes.io/tls-acme": "false"
                      },
                      "Class": "nginx",
                      "tls": {
                        "Enable": false,
                        "SecretName": null
                      }
                    },
                    "JavaOpts": null,
                    "JenkinsOpts": null,
                    "LoadBalancerIP": null,
                    "LoadBalancerSourceRanges": [
                      "0.0.0.0/0"
                    ],
                    "Name": "cjoc",
                    "NodeSelector": {},
                    "Platform": "standard",
                    "Protocol": "http",
                    "Resources": {
                      "Limits": {
                        "Cpu": 1,
                        "Memory": "2G"
                      },
                      "Requests": {
                        "Cpu": 1,
                        "Memory": "2G"
                      }
                    },
                    "Route": {
                      "tls": {
                        "CACertificate": null,
                        "Certificate": null,
                        "DestinationCACertificate": null,
                        "Enable": false,
                        "InsecureEdgeTerminationPolicy": "Redirect",
                        "Key": null,
                        "Termination": "edge"
                      }
                    },
                    "ServiceAgentListenerPort": 50000,
                    "ServiceAnnotations": {},
                    "ServicePort": 80,
                    "ServiceType": "ClusterIP",
                    "Tolerations": []
                  },
                  "Persistence": {
                    "AccessMode": "ReadWriteOnce",
                    "Annotations": {},
                    "Size": "20Gi",
                    "StorageClass": null
                  },
                  "PodSecurityPolicy": {
                    "Annotations": {},
                    "Enabled": false
                  },
                  "Subdomain": false,
                  "ingress-nginx": {
                    "Enabled": false,
                    "controller": {
                      "admissionWebhooks": {
                        "patch": {
                          "nodeSelector": {
                            "kubernetes.io/os": "linux"
                          }
                        }
                      },
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "nginx-ingress": {
                    "Enabled": false,
                    "controller": {
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "rbac": {
                    "agentsServiceAccountName": "jenkins-agents",
                    "hibernationMonitorServiceAccountName": "managed-master-hibernation-monitor",
                    "install": true,
                    "masterServiceAccountName": "jenkins",
                    "serviceAccountName": "cjoc"
                  },
                  "sidecarinjector": {
                    "Enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudBeesCI
            name: cloudbeescis.charts.cloudbees.com
            version: v1alpha1
        description: CloudBees CI
        displayName: CloudBees CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudbees
        - ci
        - enterprise
        - continuous
        - integration
        - jenkins
        - automation
        - pipeline
        - devops
        links:
        - name: CloudBees CI
          url: https://www.cloudbees.com/products/continuous-integration
        maintainers:
        - email: support@cloudbees.com
          name: Support
        - email: sales@cloudbees.com
          name: Sales
        maturity: alpha
        provider:
          name: CloudBees
          url: https://www.cloudbees.com/products/continuous-integration
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.connect.redhat.com/cloudbees/cloudbees-ci-operator@sha256:bd0bd72e9e7807b0fa307de8a93a56967221fe2dd181bac02383f0f9be9120e4
        version: 3.34.1
      entries:
      - name: cloudbees-ci.v3.34.1
        version: 3.34.1
      name: alpha
    defaultChannel: alpha
    packageName: cloudbees-ci
    provider:
      name: CloudBees
      url: https://www.cloudbees.com/products/continuous-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: https://github.com/Mellanox/network-operator/
    name: nvidia-network-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nvidia-network-operator.v25.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "doca3.1.0-25.07-0.9.7.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "nvcr.io/nvidia/mellanox",
                    "version": "sha256:a87096761d155eeb6f470e042d2d167bb466d57e63b4aba957f57d745e15a9b2"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:51270965284cceeda659fb1098e70a888b39664ceade7e9d455fb95ac85914b7
          createdAt: "2025-08-21T19:12:20Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/mellanox/nvidia-k8s-ipam@sha256:3f692273c1598525b9c698c9d9cd463e5ddc55b638c6c031639f38786fc67475
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:9a831bfdf85f313b1f5749b7c9b2673bb8fff18b4ff768c9242dabaa4468e449
        - nvcr.io/nvidia/mellanox/k8s-rdma-shared-dev-plugin@sha256:a87096761d155eeb6f470e042d2d167bb466d57e63b4aba957f57d745e15a9b2
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:5ad37cd5ae96153e6f87670f437ff5f6e44febe7bb6eaeb1a0f219e28284e53d
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:d41eea65b102dc5bb5792871ae212b59a65dae492ca109c52138d5626eba1ffd
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:13c2a2d055e501c03c6451413b85476ceafbc2fbffc7443952294c196f3f89f3
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:30354799264480e1aee057397eef2c52b277bbc6872184e2b7b811f4152b3a67
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:423aea08b4a0f01e2e19b6a709e14347bd08f8274f40002e0224090dfa4904fc
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:f9d4b8358c68ba28f849ca710bf769c4327a639733c0ff30282ca33a744078f6
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:51270965284cceeda659fb1098e70a888b39664ceade7e9d455fb95ac85914b7
        - nvcr.io/nvidia/mellanox/nic-feature-discovery@sha256:1330e3a7ea2491bc310bfc4766769389ceb56d050831bf9c145b7bfb6aba36c4
        - nvcr.io/nvidia/mellanox/network-operator-init-container@sha256:02ee60ac6ee163acc97706bd4229d268f608a95bcbcb6fccee90ceb1d8a6c62b
        - nvcr.io/nvidia/mellanox/ipoib-cni@sha256:7564ad1142c863e61f1dd84f5714deba9a7c6f6d85efb383ac3a2566eb4f4b71
        - nvcr.io/nvidia/mellanox/sriov-network-device-plugin@sha256:7508a160f3f124e0b78d2b4d5a52deca6b9f56e226fee957c4d156e425eb101a
        - nvcr.io/nvidia/mellanox/nic-configuration-operator-daemon@sha256:b78c15904cecd343fa48374757b3619b5cbdc22059d5ee9731da87e35b94a2e2
        - nvcr.io/nvidia/mellanox/nic-configuration-operator@sha256:03f34f7fde3f65ca7b214b66c6fdbc1250d4f3863f9c02b0fba78e8aa84efca3
        - nvcr.io/nvidia/mellanox/ib-kubernetes@sha256:1fa251979b5f5201af50bc7aa6936829103d742f42bfbf752f20852f5fe44dfb
        version: 25.7.0
      entries:
      - name: nvidia-network-operator.v25.7.0
        version: 25.7.0
      - name: nvidia-network-operator.v25.4.0
        version: 25.4.0
      name: stable
    - currentCSV: nvidia-network-operator.v24.10.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "24.10-0.7.0.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "v1.5.2"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:23aeb481d662496210d1c13083c65cfa4ffdcf67f973e5f884b1b936f693c0f5
          createdAt: "2025-01-27T08:16:45Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - ghcr.io/mellanox/ib-kubernetes@sha256:100e6addf99d5b44e81cb3aa8d2fc892fced1de3183190931961b92ee924207b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:e36c46cf2c813c081bf7cc22b47fb4cf391a3d91804e47a6ce87f31d62f42535
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:4bb13a8b850c91011013509334d5ae4f4b7aa1d1b0c12958625e161cd50309d6
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:9f468fdc4449e65e4772575f83aa85840a00f97165f9a00ba34695c91d610fbd
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:5a87d04872d5d4530cab3cbb3db528722fa24435f3375debd8ed102b04e631c2
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:23aeb481d662496210d1c13083c65cfa4ffdcf67f973e5f884b1b936f693c0f5
        - ghcr.io/mellanox/ipoib-cni@sha256:81646c07e27ce1df2603050a75dd267f264685fdc64c285eca3925ae98ed5392
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:77e3b946c20f45728dfa6dd1d115356de5ea317f91cddfd1540b7121550f66dc
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:47e1bb84ac97f9af49f97b49c2de44b4ba82b890b487880edc5c9adf86a176f7
        - ghcr.io/mellanox/network-operator-init-container@sha256:1699d23027ea30c9fa59575a914114bdfd5a87a359caf8c0a9b16d409ec0d068
        - ghcr.io/mellanox/nic-feature-discovery@sha256:92dda9434519de39be0f94c80787766e3399e8b238d14839360aff4d6232a09b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:db5797121735eeb935d8f11e305b50c80ba07aa83c291d76d05dea7e303bc740
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:ecb2cfdc510b7ed09a59daed4817ee353e83a94d2d8581c3c052f4fa4b990574
        version: 24.10.1
      entries:
      - name: nvidia-network-operator.v24.10.1
        version: 24.10.1
      name: v24.10
    - currentCSV: nvidia-network-operator.v24.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "24.10-0.7.0.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "v1.5.2"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:f258e8665d094c09159654759bc9ae41fbe0f7fffe95139ec02964759c3d8a87
          createdAt: "2024-11-26T09:58:46Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:5a87d04872d5d4530cab3cbb3db528722fa24435f3375debd8ed102b04e631c2
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:ecb2cfdc510b7ed09a59daed4817ee353e83a94d2d8581c3c052f4fa4b990574
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:9f468fdc4449e65e4772575f83aa85840a00f97165f9a00ba34695c91d610fbd
        - ghcr.io/mellanox/nic-feature-discovery@sha256:92dda9434519de39be0f94c80787766e3399e8b238d14839360aff4d6232a09b
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:f258e8665d094c09159654759bc9ae41fbe0f7fffe95139ec02964759c3d8a87
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:db5797121735eeb935d8f11e305b50c80ba07aa83c291d76d05dea7e303bc740
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:e36c46cf2c813c081bf7cc22b47fb4cf391a3d91804e47a6ce87f31d62f42535
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:4bb13a8b850c91011013509334d5ae4f4b7aa1d1b0c12958625e161cd50309d6
        - ghcr.io/mellanox/ib-kubernetes@sha256:100e6addf99d5b44e81cb3aa8d2fc892fced1de3183190931961b92ee924207b
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:77e3b946c20f45728dfa6dd1d115356de5ea317f91cddfd1540b7121550f66dc
        - ghcr.io/mellanox/ipoib-cni@sha256:81646c07e27ce1df2603050a75dd267f264685fdc64c285eca3925ae98ed5392
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:47e1bb84ac97f9af49f97b49c2de44b4ba82b890b487880edc5c9adf86a176f7
        - ghcr.io/mellanox/network-operator-init-container@sha256:1699d23027ea30c9fa59575a914114bdfd5a87a359caf8c0a9b16d409ec0d068
        version: 24.10.0
      entries:
      - name: nvidia-network-operator.v24.10.0
        version: 24.10.0
      name: v24.10.0
    - currentCSV: nvidia-network-operator.v24.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "docaTelemetryService": {
                    "image": "doca_telemetry",
                    "repository": "nvcr.io/nvidia/doca",
                    "version": "1.16.5-doca2.6.0-host"
                  },
                  "nicFeatureDiscovery": {
                    "image": "nic-feature-discovery",
                    "repository": "ghcr.io/mellanox",
                    "version": "v0.0.1"
                  },
                  "nvIpam": {
                    "enableWebhook": false,
                    "image": "nvidia-k8s-ipam",
                    "repository": "ghcr.io/mellanox",
                    "version": "v0.2.0"
                  },
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "24.07-0.6.1.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "v1.5.1"
                  },
                  "secondaryNetwork": {
                    "ipoib": {
                      "image": "ipoib-cni",
                      "repository": "ghcr.io/mellanox",
                      "version": "v1.2.0"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:fcee44995d64058cb7bf14c02da604879443e191a382402b8a6ebfe10a3a2601
          createdAt: "2024-08-29T13:23:34Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - ghcr.io/mellanox/network-operator-init-container@sha256:1699d23027ea30c9fa59575a914114bdfd5a87a359caf8c0a9b16d409ec0d068
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:93e856dde88abb0aeb8e0b991e5f6d9f41884d8beea792f4d70cda06f18e783f
        - ghcr.io/mellanox/ib-kubernetes@sha256:b56bb88963dadaa2290a17cf0b57b86c55c09e7c075b7271c184959532555c91
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:c41eaf76c06186e1dabb0d870225d946ad3ecda2a072fa4be197dd131b7695f2
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:8324c009c3d694fe7702af36305ede78b903d839190b1e4506e82362f9aa5c02
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:fcee44995d64058cb7bf14c02da604879443e191a382402b8a6ebfe10a3a2601
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:47e1bb84ac97f9af49f97b49c2de44b4ba82b890b487880edc5c9adf86a176f7
        - ghcr.io/mellanox/nic-feature-discovery@sha256:92dda9434519de39be0f94c80787766e3399e8b238d14839360aff4d6232a09b
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - ghcr.io/mellanox/ipoib-cni@sha256:81646c07e27ce1df2603050a75dd267f264685fdc64c285eca3925ae98ed5392
        version: 24.7.0
      entries:
      - name: nvidia-network-operator.v24.7.0
        version: 24.7.0
      name: v24.7.0
    - currentCSV: nvidia-network-operator.v24.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "docaTelemetryService": {
                    "image": "doca_telemetry",
                    "repository": "nvcr.io/nvidia/doca",
                    "version": "1.16.5-doca2.6.0-host"
                  },
                  "nicFeatureDiscovery": {
                    "image": "nic-feature-discovery",
                    "repository": "ghcr.io/mellanox",
                    "version": "v0.0.1"
                  },
                  "nvIpam": {
                    "enableWebhook": false,
                    "image": "nvidia-k8s-ipam",
                    "repository": "ghcr.io/mellanox",
                    "version": "v0.2.0"
                  },
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "24.07-0.6.1.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "v1.5.1"
                  },
                  "secondaryNetwork": {
                    "ipoib": {
                      "image": "ipoib-cni",
                      "repository": "ghcr.io/mellanox",
                      "version": "v1.2.0"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:b8a07696d05fb8e5292991ba731758579313d6983ed3d75dd95e7356024d987c
          createdAt: "2024-09-04T06:38:33Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:d5a163d36ce60630e14ae8620ab437510ff0821924e043b88f0d7b67cee3bd77
        - ghcr.io/mellanox/network-operator-init-container@sha256:1699d23027ea30c9fa59575a914114bdfd5a87a359caf8c0a9b16d409ec0d068
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:403be1bcf708984f90b3ccfdf24f58594554336d69f878d3843535f4fa61ceca
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:7facd05f60e2be29925ffce9d0a6b2c712c0670c8889b815425582c46c82efdc
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:8324c009c3d694fe7702af36305ede78b903d839190b1e4506e82362f9aa5c02
        - ghcr.io/mellanox/ib-kubernetes@sha256:b56bb88963dadaa2290a17cf0b57b86c55c09e7c075b7271c184959532555c91
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:10684c0c36364b1d897a03a346abb2f691a18541ae1242b596d002f34f0d7390
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:7f87324adb351513d87fa27dfd0c9d9371ae42529d76cce8ecd68ce41c4d263f
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:676c74fcbe6d2c2341a092913b1afca65ba1a2f568ee6796b371542657e7a174
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:22e6ff1e9245f0cdf67e7904235d1a6181ee2bd3f39cedf76b39d4aeaa4a6c1d
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:b3f503b1ac39ebc55840a436973f142eb7cf67e48a070284129fd1e30390c032
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:93e856dde88abb0aeb8e0b991e5f6d9f41884d8beea792f4d70cda06f18e783f
        - ghcr.io/mellanox/ipoib-cni@sha256:81646c07e27ce1df2603050a75dd267f264685fdc64c285eca3925ae98ed5392
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:47e1bb84ac97f9af49f97b49c2de44b4ba82b890b487880edc5c9adf86a176f7
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:b8a07696d05fb8e5292991ba731758579313d6983ed3d75dd95e7356024d987c
        - ghcr.io/mellanox/nic-feature-discovery@sha256:92dda9434519de39be0f94c80787766e3399e8b238d14839360aff4d6232a09b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:c41eaf76c06186e1dabb0d870225d946ad3ecda2a072fa4be197dd131b7695f2
        version: 24.7.1
      entries:
      - name: nvidia-network-operator.v24.7.1
        version: 24.7.1
      name: v24.7.1
    - currentCSV: nvidia-network-operator.v25.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "25.01-0.6.0.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "sha256:9f468fdc4449e65e4772575f83aa85840a00f97165f9a00ba34695c91d610fbd"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:12f0576f372f5adc17b60795952ae47388e50d8fad563b9595f3091eadd8eb29
          createdAt: "2025-02-26T16:55:35Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:f5c1741cc6368fcec5b250285552d9c80ff7254414f68102cc89162e12ac0400
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:cabce074d10a0f1d62135e2cc5442d65b49094b95b8297fdd024a1a5f461319f
        - ghcr.io/mellanox/nic-feature-discovery@sha256:92dda9434519de39be0f94c80787766e3399e8b238d14839360aff4d6232a09b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:5b91974c74f08fc5b1ad2edd4ca97d81b2dc6cc897045c3eac777826fe5d40d8
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:3edf0cec91eea10df17c9f182b2db430cb383887ac144e27ac44043b308284d7
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:68b3735c4edca2d889c456beba80ac2077b0cc0edba11a32769dacd3ece64892
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:a257aae9c4e02b62f9f4ddeea66c4349c517519a544dd02fe4c8d3fee5229521
        - ghcr.io/mellanox/ib-kubernetes@sha256:100e6addf99d5b44e81cb3aa8d2fc892fced1de3183190931961b92ee924207b
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:47e1bb84ac97f9af49f97b49c2de44b4ba82b890b487880edc5c9adf86a176f7
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:7604a30cbd0927eba76ddc74ea1cfbe113563ccf068e6ab3fe78123e4deec28d
        - ghcr.io/mellanox/ipoib-cni@sha256:56f19687b0b956ba45bf079cf3a655b0d1bad0314da9d5ecfe0f4d98ff31fc86
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:12f0576f372f5adc17b60795952ae47388e50d8fad563b9595f3091eadd8eb29
        - ghcr.io/mellanox/network-operator-init-container@sha256:67e93ccf3ecb61f17597567faf0f72e1b8ddcf73c5d7440baeadcc1cb6bb811b
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:9f468fdc4449e65e4772575f83aa85840a00f97165f9a00ba34695c91d610fbd
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        version: 25.1.0
      entries:
      - name: nvidia-network-operator.v25.1.0
        version: 25.1.0
      name: v25.1
    - currentCSV: nvidia-network-operator.v25.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "25.04-0.6.1.0-2"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "sha256:1eb5ab7f4b6f0cf89c954cf7ac8eb2e6fb2290a4f5f57417136a5302f20b12c8"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:93c2063f4da6a93810784be95df0083f07618486699d93245c97a29a63861dee
          createdAt: "2025-05-28T19:40:27Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:1b20b78f889339834ed74e0da621fc5da582719b2537b36d8967ddc6a04679b8
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:56a41010b4dbee0dc31e8f9c1467a739241a2d248ed8b5054d4b418f7f79f04d
        - ghcr.io/mellanox/nic-configuration-operator-daemon@sha256:0e8478898e0b30aec13030b11081b205308666c2a9c5c5bf2c00121079901b41
        - ghcr.io/mellanox/ipoib-cni@sha256:513d88d639a01ef98b65365f711cee7fbaa166137a56245b7ab32818f745418e
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:9bc09051aca763f83cde99ee8909cf989b1d6c5d6ab683fbb1ca441d47f61fdf
        - ghcr.io/mellanox/network-operator-init-container@sha256:67e93ccf3ecb61f17597567faf0f72e1b8ddcf73c5d7440baeadcc1cb6bb811b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:4a620bd39bc70928cd6e249855176ced0f4feb585c8c7624c0b9f5a3e57664c9
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:69e6ab6a0249be918f60c1e04fc53a07f280b33563f8edc953806538a9cc3c30
        - ghcr.io/mellanox/nic-configuration-operator@sha256:353ab27f5f5fba08197cbe8968222e1c115431cc1b9b8d8445f1facb3c3c4d7b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:35ff3b173b2694198e57b61720545be67e944c1da18ac004007a5b4e68562adc
        - ghcr.io/mellanox/ib-kubernetes@sha256:9774f774df6ede9ddf780b9193ca9d919618056e682d985e6b41e477b13c3432
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:cabce074d10a0f1d62135e2cc5442d65b49094b95b8297fdd024a1a5f461319f
        - ghcr.io/mellanox/nic-feature-discovery@sha256:35b1820fe94a833a43613ef5c9db1f58b293ea0dbcf96ec0742b0987f23f133c
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:76529c24ea17e68b89a2c0d4823d66504fa6f53759760a449308fda396574097
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:93c2063f4da6a93810784be95df0083f07618486699d93245c97a29a63861dee
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:1eb5ab7f4b6f0cf89c954cf7ac8eb2e6fb2290a4f5f57417136a5302f20b12c8
        version: 25.4.0
      entries:
      - name: nvidia-network-operator.v25.4.0
        version: 25.4.0
      name: v25.4
    defaultChannel: stable
    packageName: nvidia-network-operator
    provider:
      name: NVIDIA
      url: https://github.com/Mellanox/network-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: node-observability-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-observability-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservability",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "labels": {
                    "kubernetes.io/hostname": "worker-a"
                  },
                  "type": "crio-kubelet"
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservabilityMachineConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "debug": {
                    "enableCrioProfiling": true
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservabilityRun",
                "metadata": {
                  "name": "nodeobservabilityrun-sample",
                  "namespace": "node-observability-operator"
                },
                "spec": {
                  "nodeObservabilityRef": {
                    "name": "cluster"
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservability",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "nodeSelector": {
                    "kubernetes.io/hostname": "worker-a"
                  },
                  "type": "crio-kubelet"
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservabilityMachineConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "debug": {
                    "enableCrioProfiling": true
                  },
                  "nodeSelector": {
                    "kubernetes.io/hostname": "worker-a"
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservabilityRun",
                "metadata": {
                  "name": "nodeobservabilityrun-sample",
                  "namespace": "node-observability-operator"
                },
                "spec": {
                  "nodeObservabilityRef": {
                    "name": "cluster"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/noo/node-observability-rhel8-operator@sha256:53ba77f648154ac8aca1ae6c83815078dfde211dba1e42f597e69844e4954c01
          createdAt: 2022-12-05T09:25:09
          olm.skipRange: <0.2.0
          operatorframework.io/suggested-namespace: node-observability-operator
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeObservability prepares a subset of worker nodes (identified
              by a label) for running node observability queries, such as profiling
            displayName: Node Observability
            kind: NodeObservability
            name: nodeobservabilities.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservability prepares a subset of worker nodes (identified
              by a label) for running node observability queries, such as profiling
            displayName: Node Observability
            kind: NodeObservability
            name: nodeobservabilities.nodeobservability.olm.openshift.io
            version: v1alpha2
          - description: NodeObservabilityMachineConfig is the Schema for the nodeobservabilitymachineconfigs
              API
            displayName: Node Observability Machine Config
            kind: NodeObservabilityMachineConfig
            name: nodeobservabilitymachineconfigs.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservabilityMachineConfig is the Schema for the nodeobservabilitymachineconfigs
              API
            displayName: Node Observability Machine Config
            kind: NodeObservabilityMachineConfig
            name: nodeobservabilitymachineconfigs.nodeobservability.olm.openshift.io
            version: v1alpha2
          - description: NodeObservabilityRun is a request to run observability actions
              on the nodes previously selected in NodeObservability resource
            displayName: Node Observability Run
            kind: NodeObservabilityRun
            name: nodeobservabilityruns.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservabilityRun is a request to run observability actions
              on the nodes previously selected in NodeObservability resource
            displayName: Node Observability Run
            kind: NodeObservabilityRun
            name: nodeobservabilityruns.nodeobservability.olm.openshift.io
            version: v1alpha2
        description: |-
          An Operator that will be able to gather debugging/profiling data over
          a custom period of time which would be helpful to troubleshoot and resolve issues
          for OpenShift customers.
        displayName: Node Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node-observability-operator
        links:
        - name: Source Code
          url: https://github.com/openshift/node-observability-operator
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/noo/node-observability-rhel8-operator@sha256:53ba77f648154ac8aca1ae6c83815078dfde211dba1e42f597e69844e4954c01
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/noo/node-observability-agent-rhel8@sha256:a8e60d9f63e2ed0cf1e27309aac54014e3f1efc947b51ca0a99376338d220875
        version: 0.2.0
      entries:
      - name: node-observability-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: node-observability-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lib-bucket-provisioner
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lib-bucket-provisioner.v2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage,Big Data
          certified: "false"
          containerImage: quay.io/noobaa/pause
          createdAt: "2020-07-01T06:28:48Z"
          description: Library for the dynamic provisioning of object store buckets
            to be used by object store providers.
          repository: https://github.com/kube-object-storage/lib-bucket-provisioner
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: ObjectBucketClaim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information
            displayName: ObjectBucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: |
          ### CRD-only Operator

          This operator package is **CRD-only** and the operator is a no-op operator.

          Instead, bucket provisioners using this library are using these CRD's and using CSV [required-crds](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/Documentation/design/building-your-csv.md#required-crds) them so that OLM can install it as a dependency.

          ### Generic Bucket Provisioning

          Kubernetes natively supports dynamic provisioning for many types of file and block storage, but lacks support for object bucket provisioning.

          This repo is a placeholder for an object store bucket provisioning library, very similar to the Kubernetes [sig-storage-lib-external-provisioner](https://github.com/kubernetes-sigs/sig-storage-lib-external-provisioner/blob/master/controller/controller.go) library.

          ### Known Provisioners
          - https://github.com/noobaa/noobaa-operator (NooBaa)
          - https://github.com/rook/rook (Rook-Ceph)
          - https://github.com/yard-turkey/aws-s3-provisioner (AWS-S3)
        displayName: lib-bucket-provisioner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - object
        - bucket
        - storage
        - cloud
        - s3
        links:
        - name: Github
          url: https://github.com/kube-object-storage/lib-bucket-provisioner
        maintainers:
        - email: jcope@redhat.com
          name: Jon Cope
        - email: jvance@redhat.com
          name: Jeff Vance
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: dzaken@redhat.com
          name: Danny Zaken
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/noobaa/pause
        version: 2.0.0
      entries:
      - name: lib-bucket-provisioner.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: lib-bucket-provisioner
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Odigos
      provider-url: https://odigos.io
    name: odigos-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odigos-operator.v1.0.212
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.odigos.io/v1alpha1",
                "kind": "Odigos",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "odigos-operator"
                  },
                  "name": "odigos-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing
          containerImage: registry.connect.redhat.com/odigos/odigos-certified-operator-ubi9@sha256:c221eb2be22c2fcabcee728429bbca1255e0e31642c572e4dc40096d1d35e0d1
          createdAt: "2025-07-18T13:23:10Z"
          description: Odigos enables automatic distributed tracing with OpenTelemetry
            and eBPF.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Odigos Enterprise subscription
            (for enterprise features)
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Odigos
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Odigos is the Schema for the odigos API
            displayName: Odigos
            kind: Odigos
            name: odigos.operator.odigos.io
            version: v1alpha1
        description: |-
          The Odigos Operator provides options for installing and configuring Odigos.

          Upon creation of an Odigos resource, the Odigos Operator installs the following Odigos components:

          - Instrumentor
          - Odiglet
          - Autoscaler
          - Scheduler
          - Frontend UI

          With Odigos installed, follow the Odigos docs at docs.odigos.io for instructions on instrumenting applications and configuring destinations.
        displayName: Odigos Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - OpenTelemetry
        - eBPF
        - tracing
        - observability
        - distributed tracing
        - otel
        - monitoring
        links:
        - name: Odigos
          url: https://odigos.io
        - name: Odigos Documentation
          url: https://docs.odigos.io
        - name: Odigos on Github
          url: https://github.com/odigos-io/odigos
        maintainers:
        - email: mike@odigos.io
          name: Mike Dame
        maturity: alpha
        minKubeVersion: 1.20.15
        provider:
          name: Odigos
          url: https://odigos.io
        relatedImages:
        - registry.connect.redhat.com/odigos/odigos-certified-operator-ubi9@sha256:c221eb2be22c2fcabcee728429bbca1255e0e31642c572e4dc40096d1d35e0d1
        - registry.connect.redhat.com/odigos/odigos-ui-ubi9@sha256:b2a27cdfc94f1968a06ae18de1977127bb7d4e7d553e875d09a212693302f9c6
        - registry.connect.redhat.com/odigos/odigos-enterprise-instrumentor-ubi9@sha256:647e526cdbe38023bb221a802b180f0b95cb25734a0b39fa11cbbe22a07482a3
        - registry.connect.redhat.com/odigos/odigos-scheduler-ubi9@sha256:9e613c9b22d2fbfa3d21cc24f3b1e2dbc2e7287790e1f4286f1817c92565c865
        - registry.connect.redhat.com/odigos/odigos-instrumentor-ubi9@sha256:16712373c8e927b7208e43e9fe6e21180289975dbf3a81df60578e8550cf7150
        - registry.connect.redhat.com/odigos/odigos-odiglet-ubi9@sha256:770193901b4e5abcec59833f56f1530b43d4ec9514f3b5a5cb0bbb4356e06c67
        - registry.connect.redhat.com/odigos/odigos-enterprise-odiglet-ubi9@sha256:aec9b66461375994dc9aa92e5b2cc7d4151cf0332150726840cec049a7cbe7d9
        - registry.connect.redhat.com/odigos/odigos-autoscaler-ubi9@sha256:692193851b40f72ca79f7c82903136686d721c7fed77ef599037269ac6e55fdb
        - registry.connect.redhat.com/odigos/odigos-collector-ubi9@sha256:b67138d3ffd9e721602c5bf83f0c03ae1dcee230eca5e15335dcb9abced281ba
        version: 1.0.212
      entries:
      - name: odigos-operator.v1.0.212
        version: 1.0.212
      name: alpha
    defaultChannel: alpha
    packageName: odigos-operator
    provider:
      name: Odigos
      url: https://odigos.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-superset-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: superset-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:e0c9be4ac01a3d02f908f71f47f96bb1423fee14ba454c1bc627b35654d9cae3
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:e0c9be4ac01a3d02f908f71f47f96bb1423fee14ba454c1bc627b35654d9cae3
        version: 24.11.1
      entries:
      - name: superset-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: superset-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:f12cc3ed35135c14a82d49fcafed55465f0e66d46c2a36d532e5e09b9dece31e
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:f12cc3ed35135c14a82d49fcafed55465f0e66d46c2a36d532e5e09b9dece31e
        version: 25.3.0
      entries:
      - name: superset-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: superset-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
        version: 25.7.0
      entries:
      - name: superset-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: superset-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
        version: 25.7.0
      entries:
      - name: superset-operator.v25.7.0
        version: 25.7.0
      - name: superset-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-superset-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: LINBIT
      provider-url: ""
    name: linstor-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: linstor-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorController",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerImage": "",
                  "dbConnectionURL": "etcd://linstor-etcd:2379",
                  "drbdRepoCred": "",
                  "priorityClassName": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorCSIDriver",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerReplicas": 1,
                  "controllerEndpoint": "http://linstor:3370",
                  "csiControllerServiceAccountName": "csi-controller",
                  "csiNodeServiceAccountName": "csi-node",
                  "linstorPluginImage": "",
                  "imagePullSecret": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorSatelliteSet",
                "metadata": {
                  "name": "linstor-satellites"
                },
                "spec": {
                  "automaticStorageType": "None",
                  "controllerEndpoint": "http://linstor:3370",
                  "drbdRepoCred": "",
                  "kernelModuleInjectionMode": "ShippedModules",
                  "priorityClassName": "",
                  "satelliteImage": ""
                },
                "status": {
                  "errors": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          createdAt: "2021-06-22T07:57:53+00:00"
          description: LINSTOR Kubernetes Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/piraeusdatastore/piraeus-operator
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LinstorController is the Schema for the linstorcontrollersets
              API
            kind: LinstorController
            name: linstorcontrollers.linstor.linbit.com
            version: v1
          - description: LinstorCSIDriver is the Schema for the linstorcsidrivers
              API
            kind: LinstorCSIDriver
            name: linstorcsidrivers.linstor.linbit.com
            version: v1
          - description: LinstorSatelliteSet is the Schema for the linstornodesets
              API
            kind: LinstorSatelliteSet
            name: linstorsatellitesets.linstor.linbit.com
            version: v1
        description: |-
          LINSTOR is a configuration management system for storage on Linux systems.
          It manages LVM logical volumes and/or ZFS ZVOLs on a cluster of nodes.
          It leverages DRBD for replication between different nodes and to provide block
          storage devices to users and applications. It manages snapshots, encryption and
          caching of HDD backed data in SSDs via bcache.

          LINBIT provides a certified LINSTOR operator to ease deployment of LINSTOR
          on Openshift by installing DRBD, managing Satellite and Controller pods,
          and integrating with Openshift to provision persistent storage for your workloads.

          For detailed instructions and more configuration options, see our [user guide].

          [user guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift

          ## Install

          Unlike deployment via the helm chart, the certified Openshift
          operator does not deploy the needed etcd cluster. You must deploy this
          yourself ahead of time. We do this via the etcd operator available in the
          OperatorHub.

          IMPORTANT: It is advised that the etcd deployment uses persistent
          storage of some type. Either use an existing storage provisioner with
          a default `StorageClass` or simply use `hostPath` volumes.

          ### Installing the operator

          Hit "Install", select the stable update channel and a namespace for the
          operator. Use of a new namespace is recommended.

          Hit "Install" again. At this point you should have just one pod, the
          operator pod, running. Next we needs to configure the remaining provided APIs.

          #### A note on operator namespaces
          The LINSTOR operator can only watch for events and manage
          custom resources that are within the same namespace it is deployed
          within (OwnNamsespace). This means the LINSTOR Controller, LINSTOR
          Satellites, and LINSTOR CSI Driver pods all need to be deployed in the
          same namsepace as the LINSTOR Operator pod.

          ### Deploying the LINSTOR Controller

          Navigate to the left-hand control pane of the Openshift Web
          Console. Expand the "Operators" section, selecting "Installed Operators".
          Find the entry for the "Linstor Operator", then select the "LinstorController"
          from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorController". Click the "Create LinstorController" button.

          Here you will be presented with options to configure the LINSTOR
          Controller. Either via the web-form view or the YAML View. Regardless
          of which view you select, make sure that the `dbConnectionURL` matches
          the endpoint provided from your etcd deployment. Otherwise, the
          defaults are usually fine for most purposes.

          Lastly hit "Create", you should now see a linstor-controller pod
          running.

          ### Deploying the LINSTOR Satellites

          Next we need to deploy the Satellites Set. Just as before navigate
          to the left-hand control pane of the Openshift Web Console. Expand the
          "Operators" section, but this time select "Installed Operators". Find
          the entry for the "Linstor Operator", then select the
          "LinstorSatelliteSet" from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorSatelliteSet". Click the "Create LinstorSatelliteSet" button.

          Here you will be presented with the options to configure the LINSTOR
          Satellites. The defaults should be enough to get you started.
          Make sure the `controllerEndpoint` matches what is available in the
          openshift endpoints. The default is usually correct here.

          You can edit the `storagePools` section to configure LINSTOR storage pools,
          including preparing the backing devices. See our [storage guide].

          [storage guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-storage

          Below is an example manifest:

          ```yaml
          apiVersion: linstor.linbit.com/v1
          kind: LinstorSatelliteSet
          metadata:
            name: linstor-satellites
          spec:
            satelliteImage: ''
            automaticStorageType: None
            storagePools:
              lvmThinPools:
              - name: openshift-pool
                volumeGroup: ""
                thinVolume: openshift
                devicePaths:
                - /dev/vdb
            drbdRepoCred: ''
            kernelModuleInjectionMode: ShippedModules
            controllerEndpoint: 'http://linstor:3370'
            priorityClassName: ''
          ```

          Lastly hit "Create", you should now see a linstor-node pod
          running on every worker node.

          ### Deploying the LINSTOR CSI driver

          Last bit left is the CSI pods to bridge the layer between the CSI and
          LINSTOR. Just as before navigate to the left-hand control pane of the
          Openshift Web Console. Expand the "Operators" section, but this time
          select "Installed Operators". Find the entry for the "Linstor Operator",
          then select the "LinstorCSIDriver" from the "Provided APIs" column on the
          right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorCSIDriver". Click the "Create LinstorCSIDriver" button.

          Again, you will be presented with the options. Make sure that the
          `controllerEndpoint` is correct. Otherwise the defaults are fine for
          most use cases.

          Lastly hit "Create". You will now see a single "linstor-csi-controller" pod,
          as well as a "linstor-csi-node" pod on all worker nodes.

          ## Interacting with LINSTOR in Openshift.

          The Controller pod includes a LINSTOR Client, making it easy to interact directly with LINSTOR.
          For instance:

          ```
          oc exec deployment/linstor-cs-controller -- linstor storage-pool list
          ```

          This should only be necessary for investigating problems and accessing advanced functionality.
          Regular operation such as creating volumes should be achieved via the [Openshift/Kubernetes integration].

          [Openshift/Kubernetes integration]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-basic-configuration-and-deployment
        displayName: Linstor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - DRBD
        links:
        - name: LINBIT
          url: https://www.linbit.com/
        - name: User Guide
          url: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift
        maintainers:
        - email: sales@linbit.com
          name: LINBIT
        maturity: beta
        provider:
          name: LINBIT
        relatedImages:
        - registry.marketplace.redhat.com/rhm/linbit/linstor-operator@sha256:75e3ec9a9331072dc88c13979f0cacdf6232f8b3a6bfaf55c95b2ee3a03a5d5b
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:395a5a4aa4cfe3a0093d2225ce2e67acdcec0fd894e4b61e30a750f22931448d
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:bba872052e42fe0d6cbb2eea6541bf57f035575cbc7d838a3260b8582b8fe572
        - registry.marketplace.redhat.com/rhm/linbit/linstor-controller@sha256:1084ba14eaec7511682dd87c562398ce5e6690ded700c964373188b73bfe3b17
        - registry.marketplace.redhat.com/rhm/linbit/drbd-9@sha256:6337b3adfa3de87e74adcc974edafa09b618cb8e0315658c1cbdf8dd1c6c4c32
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:ba217e7e51f0748b54f477486263ff4983b0029635fd1f7ccc67160a09f81608
        - registry.marketplace.redhat.com/rhm/linbit/linstor-csi@sha256:034e3641376aa7ad47ed309fb389083c3ff82974f246f4d71b8789be8f8f618d
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:53b0f48eeecaa3ff1e211a2a413cd49266b7be85a5aa3a0431ab72d32e3a4e5c
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:8e52a6b31f9f3834d709dbca3614c2110b8ebb4af382c76b7056921221c87600
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:91cfe63acc12c91c3053ba7cc12d72cce404d33aab4afcfac285311f5ddfc7fe
        - registry.marketplace.redhat.com/rhm/linbit/linstor-satellite@sha256:4a3503092c8cb9435ccfb4837ee6a9183b680eb18a35c7d40227fa5c8f4da996
        version: 1.5.1
      entries:
      - name: linstor-operator.v1.5.1
        version: 1.5.1
      name: alpha
    - currentCSV: linstor-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorController",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerImage": "",
                  "dbConnectionURL": "etcd://linstor-etcd:2379",
                  "drbdRepoCred": "",
                  "priorityClassName": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorCSIDriver",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerReplicas": 1,
                  "controllerEndpoint": "http://linstor:3370",
                  "csiControllerServiceAccountName": "csi-controller",
                  "csiNodeServiceAccountName": "csi-node",
                  "linstorPluginImage": "",
                  "imagePullSecret": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorSatelliteSet",
                "metadata": {
                  "name": "linstor-satellites"
                },
                "spec": {
                  "automaticStorageType": "None",
                  "controllerEndpoint": "http://linstor:3370",
                  "drbdRepoCred": "",
                  "kernelModuleInjectionMode": "ShippedModules",
                  "priorityClassName": "",
                  "satelliteImage": ""
                },
                "status": {
                  "errors": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          createdAt: "2021-06-22T07:57:53+00:00"
          description: LINSTOR Kubernetes Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/piraeusdatastore/piraeus-operator
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LinstorController is the Schema for the linstorcontrollersets
              API
            kind: LinstorController
            name: linstorcontrollers.linstor.linbit.com
            version: v1
          - description: LinstorCSIDriver is the Schema for the linstorcsidrivers
              API
            kind: LinstorCSIDriver
            name: linstorcsidrivers.linstor.linbit.com
            version: v1
          - description: LinstorSatelliteSet is the Schema for the linstornodesets
              API
            kind: LinstorSatelliteSet
            name: linstorsatellitesets.linstor.linbit.com
            version: v1
        description: |-
          LINSTOR is a configuration management system for storage on Linux systems.
          It manages LVM logical volumes and/or ZFS ZVOLs on a cluster of nodes.
          It leverages DRBD for replication between different nodes and to provide block
          storage devices to users and applications. It manages snapshots, encryption and
          caching of HDD backed data in SSDs via bcache.

          LINBIT provides a certified LINSTOR operator to ease deployment of LINSTOR
          on Openshift by installing DRBD, managing Satellite and Controller pods,
          and integrating with Openshift to provision persistent storage for your workloads.

          For detailed instructions and more configuration options, see our [user guide].

          [user guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift

          ## Install

          Unlike deployment via the helm chart, the certified Openshift
          operator does not deploy the needed etcd cluster. You must deploy this
          yourself ahead of time. We do this via the etcd operator available in the
          OperatorHub.

          IMPORTANT: It is advised that the etcd deployment uses persistent
          storage of some type. Either use an existing storage provisioner with
          a default `StorageClass` or simply use `hostPath` volumes.

          ### Installing the operator

          Hit "Install", select the stable update channel and a namespace for the
          operator. Use of a new namespace is recommended.

          Hit "Install" again. At this point you should have just one pod, the
          operator pod, running. Next we needs to configure the remaining provided APIs.

          #### A note on operator namespaces
          The LINSTOR operator can only watch for events and manage
          custom resources that are within the same namespace it is deployed
          within (OwnNamsespace). This means the LINSTOR Controller, LINSTOR
          Satellites, and LINSTOR CSI Driver pods all need to be deployed in the
          same namsepace as the LINSTOR Operator pod.

          ### Deploying the LINSTOR Controller

          Navigate to the left-hand control pane of the Openshift Web
          Console. Expand the "Operators" section, selecting "Installed Operators".
          Find the entry for the "Linstor Operator", then select the "LinstorController"
          from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorController". Click the "Create LinstorController" button.

          Here you will be presented with options to configure the LINSTOR
          Controller. Either via the web-form view or the YAML View. Regardless
          of which view you select, make sure that the `dbConnectionURL` matches
          the endpoint provided from your etcd deployment. Otherwise, the
          defaults are usually fine for most purposes.

          Lastly hit "Create", you should now see a linstor-controller pod
          running.

          ### Deploying the LINSTOR Satellites

          Next we need to deploy the Satellites Set. Just as before navigate
          to the left-hand control pane of the Openshift Web Console. Expand the
          "Operators" section, but this time select "Installed Operators". Find
          the entry for the "Linstor Operator", then select the
          "LinstorSatelliteSet" from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorSatelliteSet". Click the "Create LinstorSatelliteSet" button.

          Here you will be presented with the options to configure the LINSTOR
          Satellites. The defaults should be enough to get you started.
          Make sure the `controllerEndpoint` matches what is available in the
          openshift endpoints. The default is usually correct here.

          You can edit the `storagePools` section to configure LINSTOR storage pools,
          including preparing the backing devices. See our [storage guide].

          [storage guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-storage

          Below is an example manifest:

          ```yaml
          apiVersion: linstor.linbit.com/v1
          kind: LinstorSatelliteSet
          metadata:
            name: linstor-satellites
          spec:
            satelliteImage: ''
            automaticStorageType: None
            storagePools:
              lvmThinPools:
              - name: openshift-pool
                volumeGroup: ""
                thinVolume: openshift
                devicePaths:
                - /dev/vdb
            drbdRepoCred: ''
            kernelModuleInjectionMode: ShippedModules
            controllerEndpoint: 'http://linstor:3370'
            priorityClassName: ''
          ```

          Lastly hit "Create", you should now see a linstor-node pod
          running on every worker node.

          ### Deploying the LINSTOR CSI driver

          Last bit left is the CSI pods to bridge the layer between the CSI and
          LINSTOR. Just as before navigate to the left-hand control pane of the
          Openshift Web Console. Expand the "Operators" section, but this time
          select "Installed Operators". Find the entry for the "Linstor Operator",
          then select the "LinstorCSIDriver" from the "Provided APIs" column on the
          right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorCSIDriver". Click the "Create LinstorCSIDriver" button.

          Again, you will be presented with the options. Make sure that the
          `controllerEndpoint` is correct. Otherwise the defaults are fine for
          most use cases.

          Lastly hit "Create". You will now see a single "linstor-csi-controller" pod,
          as well as a "linstor-csi-node" pod on all worker nodes.

          ## Interacting with LINSTOR in Openshift.

          The Controller pod includes a LINSTOR Client, making it easy to interact directly with LINSTOR.
          For instance:

          ```
          oc exec deployment/linstor-cs-controller -- linstor storage-pool list
          ```

          This should only be necessary for investigating problems and accessing advanced functionality.
          Regular operation such as creating volumes should be achieved via the [Openshift/Kubernetes integration].

          [Openshift/Kubernetes integration]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-basic-configuration-and-deployment
        displayName: Linstor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - DRBD
        links:
        - name: LINBIT
          url: https://www.linbit.com/
        - name: User Guide
          url: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift
        maintainers:
        - email: sales@linbit.com
          name: LINBIT
        maturity: beta
        provider:
          name: LINBIT
        relatedImages:
        - registry.marketplace.redhat.com/rhm/linbit/linstor-csi@sha256:034e3641376aa7ad47ed309fb389083c3ff82974f246f4d71b8789be8f8f618d
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:8e52a6b31f9f3834d709dbca3614c2110b8ebb4af382c76b7056921221c87600
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:395a5a4aa4cfe3a0093d2225ce2e67acdcec0fd894e4b61e30a750f22931448d
        - registry.marketplace.redhat.com/rhm/linbit/linstor-controller@sha256:1084ba14eaec7511682dd87c562398ce5e6690ded700c964373188b73bfe3b17
        - registry.marketplace.redhat.com/rhm/linbit/linstor-satellite@sha256:4a3503092c8cb9435ccfb4837ee6a9183b680eb18a35c7d40227fa5c8f4da996
        - registry.marketplace.redhat.com/rhm/linbit/linstor-operator@sha256:75e3ec9a9331072dc88c13979f0cacdf6232f8b3a6bfaf55c95b2ee3a03a5d5b
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:53b0f48eeecaa3ff1e211a2a413cd49266b7be85a5aa3a0431ab72d32e3a4e5c
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:bba872052e42fe0d6cbb2eea6541bf57f035575cbc7d838a3260b8582b8fe572
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:ba217e7e51f0748b54f477486263ff4983b0029635fd1f7ccc67160a09f81608
        - registry.marketplace.redhat.com/rhm/linbit/drbd-9@sha256:6337b3adfa3de87e74adcc974edafa09b618cb8e0315658c1cbdf8dd1c6c4c32
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:91cfe63acc12c91c3053ba7cc12d72cce404d33aab4afcfac285311f5ddfc7fe
        version: 1.5.1
      entries:
      - name: linstor-operator.v1.5.1
        version: 1.5.1
      name: stable
    defaultChannel: stable
    packageName: linstor-operator-rhmp
    provider:
      name: LINBIT
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: pg4k-pgd
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pg4k-pgd.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgd.k8s.enterprisedb.io/v1beta1",
                "kind": "PGDGroup",
                "metadata": {
                  "name": "region-a"
                },
                "spec": {
                  "backup": null,
                  "cnp": {
                    "storage": {
                      "size": "1Gi"
                    },
                    "walStorage": {
                      "size": "1Gi"
                    }
                  },
                  "connectivity": {
                    "dns": {
                      "additional": [
                        {
                          "domain": "alternate.domain"
                        },
                        {
                          "domain": "my.domain",
                          "hostSuffix": "-dc1"
                        }
                      ]
                    },
                    "tls": {
                      "clientCert": {
                        "caCertSecret": "client-ca-key-pair",
                        "certManager": {
                          "spec": {
                            "issuerRef": {
                              "group": "cert-manager.io",
                              "kind": "Issuer",
                              "name": "client-ca-issuer"
                            }
                          }
                        }
                      },
                      "mode": "verify-ca",
                      "serverCert": {
                        "caCertSecret": "server-ca-key-pair",
                        "certManager": {
                          "spec": {
                            "issuerRef": {
                              "group": "cert-manager.io",
                              "kind": "Issuer",
                              "name": "server-ca-issuer"
                            }
                          }
                        }
                      }
                    }
                  },
                  "instances": 2,
                  "pgd": {
                    "parentGroup": {
                      "create": true,
                      "name": "world"
                    }
                  },
                  "proxyInstances": 2,
                  "witness": {
                    "storage": {
                      "size": "1Gi"
                    },
                    "walStorage": {
                      "size": "1Gi"
                    }
                  },
                  "witnessInstances": 1
                }
              },
              {
                "apiVersion": "pgd.k8s.enterprisedb.io/v1beta1",
                "kind": "PGDGroupCleanup",
                "metadata": {
                  "name": "region-b-cleanup"
                },
                "spec": {
                  "executor": "region-a",
                  "target": "region-b"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: docker.enterprisedb.com/k8s_enterprise_pgd/pg4k-pgd-openshift@sha256:5127d913011619d61fd06b27be3c709df2236a4b5c0254fa58632698d58b50fb
          createdAt: "2025-09-26T01:49:29Z"
          description: EDB Postgres Distributed for Kubernetes is an operator designed
            to manage EDB Postgres Distributed (PGD) v5 workloads on Kubernetes both
            in a single Kubernetes cluster or across multiple ones.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterImageCatalog
            name: clusterimagecatalogs.pgd.k8s.enterprisedb.io
            version: v1beta1
          - kind: ImageCatalog
            name: imagecatalogs.pgd.k8s.enterprisedb.io
            version: v1beta1
          - description: Object to clean up a PGD Group
            displayName: PGD Group Cleanup
            kind: PGDGroupCleanup
            name: pgdgroupcleanups.pgd.k8s.enterprisedb.io
            version: v1beta1
          - description: A group of Postgres Clusters
            displayName: PGD Group
            kind: PGDGroup
            name: pgdgroups.pgd.k8s.enterprisedb.io
            version: v1beta1
        description: |
          EDB Postgres Distributed for Kubernetes is an operator designed to manage EDB Postgres Distributed (PGD) v5 workloads on Kubernetes both in a single Kubernetes cluster or across multiple ones.

          EDB Postgres Distributed is an extension for PostgreSQL that provides multi-master replication and data distribution with advanced conflict management, data-loss protection, and throughput up to 5X faster than native logical replication. It enables distributed PostgreSQL clusters with high availability up to five 9s.

          This operator requires access to the EDB private repositories.
        displayName: EDB Postgres Distributed for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - distributed
        - postgres
        - database
        - sql
        - cloud
        - native
        - edb
        - enterprisedb
        - cnp
        - pg4k
        - pgd
        links:
        - name: 'Documentation: EDB Postgres Distributed v5'
          url: https://www.enterprisedb.com/docs/pgd/latest/
        - name: 'Documentation: EDB Postgres Distributed for Kubernetes'
          url: https://www.enterprisedb.com/docs/postgres_distributed_for_kubernetes/latest/
        maintainers:
        - email: sales@enterprisedb.com
          name: Sales
        - email: support@enterprisedb.com
          name: Support
        maturity: alpha
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - docker.enterprisedb.com/k8s_enterprise_pgd/pg4k-pgd-openshift@sha256:5127d913011619d61fd06b27be3c709df2236a4b5c0254fa58632698d58b50fb
        version: 1.1.3
      entries:
      - name: pg4k-pgd.v1.1.3
        version: 1.1.3
      - name: pg4k-pgd.v1.1.2
        version: 1.1.2
      name: stable
    defaultChannel: stable
    packageName: pg4k-pgd
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Mantisnet
      provider-url: https://www.mantisnet.com
    name: tawon-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tawon-operator.v2.29.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Directive",
                "metadata": {
                  "name": "amf-h2"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "amf"
                    }
                  },
                  "container": {
                    "resources": {
                      "limits": {
                        "cpu": "500m"
                      }
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  },
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 8080"
                      },
                      "task": "payload"
                    },
                    {
                      "task": "h2"
                    },
                    {
                      "config": {
                        "addr": "nats://nats:4222",
                        "topic": "hello"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "TopologyAggregator",
                "metadata": {
                  "name": "topology-aggregator"
                },
                "spec": {
                  "messaging": {
                    "nats": {
                      "url": "nats://nats:4222"
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-11-13T01:16:29Z"
          olm.skipRange: '>=2.23.0 <2.29.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Directive allows to define a Tawon task pipeline that will
              be applied to a Kubernetes cluster.
            displayName: Tawon Directive
            kind: Directive
            name: directives.tawon.mantisnet.com
            version: v1alpha1
          - description: TopologyAggregator is the Schema for the topologyaggregators
              API
            displayName: Tawon Topology Aggregator
            kind: TopologyAggregator
            name: topologyaggregators.tawon.mantisnet.com
            version: v1alpha1
        description: Tawon is a cloud-native observability fabric for deeply instrumenting
          microservice-based applications.
        displayName: Tawon
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - networking
        - metrics
        - ebpf
        links:
        - name: Tawon Docs
          url: https://docs.tawon.xmantis.net
        maintainers:
        - email: elliott@mantisnet.com
          name: Elliott Starin
        - email: nicolas@mantisnet.com
          name: Nicolas Paton
        - email: thiago@mantisnet.com
          name: Thiago Navarro
        maturity: alpha
        provider:
          name: Mantisnet
          url: https://www.mantisnet.com
        relatedImages:
        - quay.io/mantisnet/kernel-headers@sha256:242e837af831b6eed98d522b6bae7416928b9fbe451fb6e02ffa678686771068
        - quay.io/mantisnet/kernel-headers@sha256:2aa3bc7871fe0fc0a79ac812d941e55cd0d25a5f0d1d9b3e2435aab6f3d70c67
        - quay.io/mantisnet/tawonctl@sha256:d436ef902bb08d830f64f34fdb86b20d2bd18d86233c8240014f0722c60e5e94
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - quay.io/mantisnet/tawon-operator@sha256:d5669dde46bb30411964baa86844d97721655886769b96809626246a4123caad
        - quay.io/mantisnet/tawon@sha256:83575079fe00ae6e53bb13005b94ef0047cc1974d95015c19e64dd53e13bcb93
        - quay.io/mantisnet/nats@sha256:6afc7ced92181a7c12d709e9693740847297dc5d60ab78cbac7b9b6c859ecdd1
        version: 2.29.1
      entries:
      - name: tawon-operator.v2.29.1
        version: 2.29.1
      name: alpha
    - currentCSV: tawon-operator.v2.39.27-rc6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "ClusterDirective",
                "metadata": {
                  "name": "capture-coredns"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "coredns"
                    }
                  },
                  "debug": {
                    "logLevel": "debug"
                  },
                  "streams": [
                    {
                      "maxage": "6h0m0s",
                      "maxmsgs": 1000,
                      "name": "coredns",
                      "retentionPolicy": "Delete"
                    }
                  ],
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 53"
                      },
                      "task": "capture"
                    },
                    {
                      "config": {
                        "name": "coredns",
                        "type": "stream"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Dashboard",
                "metadata": {
                  "name": "tawon-dashboard"
                },
                "spec": null
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Directive",
                "metadata": {
                  "name": "capture-coredns"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "coredns"
                    }
                  },
                  "debug": {
                    "logLevel": "debug"
                  },
                  "streams": [
                    {
                      "maxage": "6h0m0s",
                      "maxmsgs": 1000,
                      "name": "coredns",
                      "retentionPolicy": "Delete"
                    }
                  ],
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 53"
                      },
                      "task": "capture"
                    },
                    {
                      "config": {
                        "name": "coredns",
                        "type": "stream"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Stream",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "tawon-operator",
                    "app.kubernetes.io/instance": "stream-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "stream",
                    "app.kubernetes.io/part-of": "tawon-operator"
                  },
                  "name": "stream-sample"
                },
                "spec": {
                  "maxage": "6h",
                  "maxbytes": "1G",
                  "maxmsgs": 100000,
                  "store": {
                    "name": "mystreamstore"
                  }
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "StreamStore",
                "metadata": {
                  "name": "mystreamstore"
                },
                "spec": {
                  "defaults": {
                    "maxage": "240h",
                    "maxbytes": "1Gi",
                    "maxsize": 1000000
                  },
                  "limits": {
                    "max": 10000000,
                    "maxage": "2400h",
                    "maxbytes": "2Gi"
                  },
                  "persistentVolumeClaimRetentionPolicy": {
                    "whenDeleted": "Delete"
                  },
                  "volumeClaimTemplate": {
                    "metadata": {
                      "name": "mypvclaim"
                    },
                    "spec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "ssd-csi"
                    }
                  }
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "TopologyAggregator",
                "metadata": {
                  "name": "topology-aggregator"
                },
                "spec": {
                  "messaging": {
                    "nats": {
                      "url": "nats://nats:4222"
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing, Monitoring, Networking
          containerImage: quay.io/mantisnet/tawon-operator@sha256:d8ce5c8966d3590ce5f286d8e778463b85582b970ad7730ddf7828bc68843a90
          createdAt: "2025-03-08T18:58:04Z"
          description: Tawon is a cloud-native observability fabric for deeply instrumenting
            microservice-based applications.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Mantisnet Tawon Operator Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: MantisNet
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterDirective is the Schema for the clusterdirectives
              API.
            displayName: Cluster Tawon Directive
            kind: ClusterDirective
            name: clusterdirectives.tawon.mantisnet.com
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.tawon.mantisnet.com
            version: v1alpha1
          - description: A Directive allows to define a Tawon task pipeline that will
              be applied to a Kubernetes cluster.
            displayName: Tawon Directive
            kind: Directive
            name: directives.tawon.mantisnet.com
            version: v1alpha1
          - description: Stream is a persistent Tawon data stream.
            displayName: Stream
            kind: Stream
            name: streams.tawon.mantisnet.com
            version: v1alpha1
          - description: StreamStore is the Schema for the streamstores API
            displayName: Stream Store
            kind: StreamStore
            name: streamstores.tawon.mantisnet.com
            version: v1alpha1
          - description: TopologyAggregator is the Schema for the topologyaggregators
              API
            displayName: Tawon Topology Aggregator
            kind: TopologyAggregator
            name: topologyaggregators.tawon.mantisnet.com
            version: v1alpha1
        description: Tawon is a cloud-native observability fabric for deeply instrumenting
          microservice-based applications.
        displayName: Tawon
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - networking
        - metrics
        - ebpf
        links:
        - name: Tawon Docs
          url: https://docs.tawon.xmantis.net
        maintainers:
        - email: elliott@mantisnet.com
          name: Elliott Starin
        - email: nicolas@mantisnet.com
          name: Nicolas Paton
        - email: thiago@mantisnet.com
          name: Thiago Navarro
        maturity: alpha
        provider:
          name: Mantisnet
          url: https://www.mantisnet.com
        relatedImages:
        - quay.io/mantisnet/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/mantisnet/kernel-headers@sha256:ca3baeba7ab8ce968c1ee129112532fc8b8d6fb6741e42831b6100f0a2451629
        - quay.io/mantisnet/tawon-operator@sha256:d8ce5c8966d3590ce5f286d8e778463b85582b970ad7730ddf7828bc68843a90
        - quay.io/mantisnet/diagnose@sha256:ab15f4eaeda32f510d81a3cc865b0fd77405b55f6a0c6f285131331dcc38f2ca
        - quay.io/mantisnet/tawonctl@sha256:8bf40ca0cc9984a252363e87d6135e43ac52839113b762399cc54a085af171d8
        - quay.io/mantisnet/sengat@sha256:00c3e1194c2c3e343d16c1543b7ce2a1ee9829f6ddd1c2870785624596b46ebf
        - quay.io/mantisnet/kernel-headers@sha256:9290eb56b186ee0edf65da615c1a8aabc66f7d290168e49ad8764ab8fc2d209a
        - quay.io/mantisnet/tawon@sha256:46e654c6f275be2be604254102aaee9d927859cfd05dfa16b7cb571d36b0f1a2
        - quay.io/mantisnet/tawon-tests@sha256:cd49c24deb7cc9f4b439cb9baa885208dfe91de385cd8c5d312fd08280615c2c
        - quay.io/openshift/origin-oauth-proxy@sha256:28856aae9e481fe763776df713b4b335632645687e6fd2e31619f76d11e0f47c
        - quay.io/mantisnet/tawon@sha256:5a8bf475ff18abacb96f631e4d3ea01c94bef8a38f84c00d1dc4047073b6cc14
        - quay.io/mantisnet/nats@sha256:5f9e7f8f685fd0c1279a3465b61c7215f4600c99d81785dda86a71c966cf829a
        - quay.io/mantisnet/kernel-headers@sha256:5d00bfa9f60c95afcfa363ede59a9625fef9925cd1cb4235327a1de94a0fc632
        version: 2.39.27-rc6
      entries:
      - name: tawon-operator.v2.39.27-rc6
        version: 2.39.27-rc6
      name: candidate-v2
    - currentCSV: tawon-operator.v2.39.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "ClusterDirective",
                "metadata": {
                  "name": "capture-coredns"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "coredns"
                    }
                  },
                  "debug": {
                    "logLevel": "debug"
                  },
                  "streams": [
                    {
                      "maxage": "6h0m0s",
                      "maxmsgs": 1000,
                      "name": "coredns",
                      "retentionPolicy": "Delete"
                    }
                  ],
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 53"
                      },
                      "task": "capture"
                    },
                    {
                      "config": {
                        "name": "coredns",
                        "type": "stream"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Dashboard",
                "metadata": {
                  "name": "tawon-dashboard"
                },
                "spec": null
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Directive",
                "metadata": {
                  "name": "capture-coredns"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "coredns"
                    }
                  },
                  "debug": {
                    "logLevel": "debug"
                  },
                  "streams": [
                    {
                      "maxage": "6h0m0s",
                      "maxmsgs": 1000,
                      "name": "coredns",
                      "retentionPolicy": "Delete"
                    }
                  ],
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 53"
                      },
                      "task": "capture"
                    },
                    {
                      "config": {
                        "name": "coredns",
                        "type": "stream"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Stream",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "tawon-operator",
                    "app.kubernetes.io/instance": "stream-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "stream",
                    "app.kubernetes.io/part-of": "tawon-operator"
                  },
                  "name": "stream-sample"
                },
                "spec": {
                  "maxage": "6h",
                  "maxbytes": "1G",
                  "maxmsgs": 100000,
                  "store": {
                    "name": "mystreamstore"
                  }
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "StreamStore",
                "metadata": {
                  "name": "mystreamstore"
                },
                "spec": {
                  "defaults": {
                    "maxage": "240h",
                    "maxbytes": "1Gi",
                    "maxsize": 1000000
                  },
                  "limits": {
                    "max": 10000000,
                    "maxage": "2400h",
                    "maxbytes": "2Gi"
                  },
                  "persistentVolumeClaimRetentionPolicy": {
                    "whenDeleted": "Delete"
                  },
                  "volumeClaimTemplate": {
                    "metadata": {
                      "name": "mypvclaim"
                    },
                    "spec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "ssd-csi"
                    }
                  }
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "TopologyAggregator",
                "metadata": {
                  "name": "topology-aggregator"
                },
                "spec": {
                  "messaging": {
                    "nats": {
                      "url": "nats://nats:4222"
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing, Monitoring, Networking
          containerImage: quay.io/mantisnet/tawon-operator@sha256:f831abf0f9fb3a4e096743d4ff7356984738da25790e305325e740ca97f0be5b
          createdAt: "2025-07-18T16:24:07Z"
          description: Tawon is a cloud-native observability fabric for deeply instrumenting
            microservice-based applications.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.23.0 <2.39.32'
          operators.openshift.io/valid-subscription: Mantisnet Tawon Operator Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: MantisNet
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterDirective is the Schema for the clusterdirectives
              API.
            displayName: Cluster Tawon Directive
            kind: ClusterDirective
            name: clusterdirectives.tawon.mantisnet.com
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.tawon.mantisnet.com
            version: v1alpha1
          - description: A Directive allows to define a Tawon task pipeline that will
              be applied to a Kubernetes cluster.
            displayName: Tawon Directive
            kind: Directive
            name: directives.tawon.mantisnet.com
            version: v1alpha1
          - description: Stream is a persistent Tawon data stream.
            displayName: Stream
            kind: Stream
            name: streams.tawon.mantisnet.com
            version: v1alpha1
          - description: StreamStore is the Schema for the streamstores API
            displayName: Stream Store
            kind: StreamStore
            name: streamstores.tawon.mantisnet.com
            version: v1alpha1
          - description: TopologyAggregator is the Schema for the topologyaggregators
              API
            displayName: Tawon Topology Aggregator
            kind: TopologyAggregator
            name: topologyaggregators.tawon.mantisnet.com
            version: v1alpha1
        description: Tawon is a cloud-native observability fabric for deeply instrumenting
          microservice-based applications.
        displayName: Tawon
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - networking
        - metrics
        - ebpf
        links:
        - name: Tawon Docs
          url: https://docs.tawon.xmantis.net
        maintainers:
        - email: elliott@mantisnet.com
          name: Elliott Starin
        - email: nicolas@mantisnet.com
          name: Nicolas Paton
        - email: thiago@mantisnet.com
          name: Thiago Navarro
        maturity: alpha
        provider:
          name: Mantisnet
          url: https://www.mantisnet.com
        relatedImages:
        - quay.io/mantisnet/nats@sha256:5f9e7f8f685fd0c1279a3465b61c7215f4600c99d81785dda86a71c966cf829a
        - quay.io/mantisnet/tawon-tests@sha256:294df1580816919502dc592c1c28ffff32b13ee892686f783eff181863fa9d18
        - quay.io/mantisnet/tawon@sha256:1640758505d9b4e0f684ab502ef3344135a270192054e256e7e711209fa51adf
        - quay.io/mantisnet/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/mantisnet/nginx@sha256:c9f91949187fa1c2b4615b88d3acf7902c7e2d4a2557f33ca0cf90164269a7ae
        - quay.io/mantisnet/tawonctl@sha256:88ac81b4fe00d360c5ff127f77456f9e36990e435442a9194d3dbf50b20e7206
        - quay.io/mantisnet/diagnose@sha256:ab15f4eaeda32f510d81a3cc865b0fd77405b55f6a0c6f285131331dcc38f2ca
        - quay.io/mantisnet/tawon@sha256:846d03ada7e6f6648fc74fd0ec9e68c3c6b17e4c155df22829a3db3ca556c395
        - quay.io/mantisnet/kernel-headers@sha256:9290eb56b186ee0edf65da615c1a8aabc66f7d290168e49ad8764ab8fc2d209a
        - quay.io/mantisnet/tawon-operator@sha256:f831abf0f9fb3a4e096743d4ff7356984738da25790e305325e740ca97f0be5b
        - quay.io/mantisnet/alpine@sha256:ec1b05d1eac264d9204a57f4ad9d4dc35e9e756e9fedaea0674aefc7edb1d6a4
        - quay.io/mantisnet/kernel-headers@sha256:ca3baeba7ab8ce968c1ee129112532fc8b8d6fb6741e42831b6100f0a2451629
        - quay.io/mantisnet/kernel-headers@sha256:5d00bfa9f60c95afcfa363ede59a9625fef9925cd1cb4235327a1de94a0fc632
        - quay.io/mantisnet/kernel-headers@sha256:b0975a44cc94bc1d75679f288a9e23a65c2c0b350ce1fb2f5eb421a76a9567c6
        - quay.io/mantisnet/sengat@sha256:483fbce2bacbd95b7458773e2278b936f06a19e65acb8987adffa8403d76e40a
        - quay.io/openshift/origin-oauth-proxy@sha256:28856aae9e481fe763776df713b4b335632645687e6fd2e31619f76d11e0f47c
        version: 2.39.32
      entries:
      - name: tawon-operator.v2.39.32
        version: 2.39.32
      - name: tawon-operator.v2.39.31
        version: 2.39.31
      - name: tawon-operator.v2.39.29
        version: 2.39.29
      - name: tawon-operator.v2.39.28
        version: 2.39.28
      - name: tawon-operator.v2.39.27
        version: 2.39.27
      name: fast-v2
    defaultChannel: fast-v2
    packageName: tawon-operator
    provider:
      name: Mantisnet
      url: https://www.mantisnet.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Perficient Inc
      provider-url: http://www.perficient.com
    name: devopsinabox
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devopsinabox.v0.0.45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "DevSecOpsManager",
                "metadata": {
                  "name": "devsecopsmanager-sample"
                },
                "spec": {
                  "config": {
                    "repoPath": "",
                    "repoUrl": ""
                  },
                  "forwarding": {
                    "fluentd": {
                      "affinity": {},
                      "caFile": "files/default-openshift-logging-fluentd.crt",
                      "image": "registry.redhat.io/openshift4/ose-logging-fluentd:v4.4",
                      "keyFile": "files/default-openshift-logging-fluentd.key",
                      "loglevel": "warn",
                      "nodeSelector": {},
                      "passphrase": "",
                      "persistence": {
                        "accessMode": "ReadWriteOnce",
                        "enabled": false,
                        "size": "10Gi",
                        "storageClass": ""
                      },
                      "port": 24224,
                      "replicas": 2,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "512Mi"
                        }
                      },
                      "sharedkey": "splunkforwarding",
                      "ssl": true,
                      "tolerations": [],
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      }
                    },
                    "preserveDefault": true
                  },
                  "fullnameOverride": "",
                  "git": {
                    "password": "",
                    "server": "https://gitlab.com",
                    "token": "",
                    "username": ""
                  },
                  "nameOverride": "",
                  "prometheus": {
                    "token": ""
                  }
                }
              },
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "ManagedNamespace",
                "metadata": {
                  "name": "managednamespace-sample"
                },
                "spec": {
                  "config": {
                    "repoBranch": "master",
                    "repoPath": "",
                    "repoUrl": ""
                  },
                  "container": {
                    "password": "",
                    "registry": "https://quay.io",
                    "username": ""
                  },
                  "fullnameOverride": "",
                  "git": {
                    "password": "",
                    "server": "https://gitlab.com",
                    "token": "",
                    "username": ""
                  },
                  "manager": {
                    "namespace": ""
                  },
                  "mirrors": {
                    "maven": ""
                  },
                  "nameOverride": ""
                }
              },
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "SpringBootApp",
                "metadata": {
                  "name": "springbootapp-sample"
                },
                "spec": {
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "config": {
                    "repo": ""
                  },
                  "container": {
                    "image": "placeholder",
                    "registry": ""
                  },
                  "fullnameOverride": "",
                  "git": {
                    "branch": "master",
                    "module": "",
                    "url": ""
                  },
                  "management": {
                    "port": 8081
                  },
                  "maven": {
                    "mirror": ""
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "replicaCount": 1,
                  "resources": {},
                  "service": {
                    "port": 8080
                  },
                  "serviceAccount": {
                    "annotations": {}
                  },
                  "version": {
                    "branch": "",
                    "image": "placeholder",
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration,Developer Tools,Application Runtime
          containerImage: quay.io/mrethers/devopsinabox-operator:v0.0.45
          createdAt: 2021-03-03 10:51:09Z
          description: Perficient's DevSecOps Manager For Kubernetes Applications
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/prft-rh/devopsinabox-operator
          support: Matthieu Rethers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DevSecOpsManager
            name: devsecopsmanagers.apimc.com.perficient
            version: v1alpha1
          - kind: ManagedNamespace
            name: managednamespaces.apimc.com.perficient
            version: v1alpha1
          - kind: SpringBootApp
            name: springbootapps.apimc.com.perficient
            version: v1alpha1
        description: "The devops-in-a-box Operator manages the entire lifecycle of
          your applications in Kubernetes. It encapsulates all the best devsecops
          practices gathered by Perficient over the years working on many large projects.
          Use it to migrate existing applications with minimum changes or deploy new
          ones with little to no knowledge of Kubernetes.\n\nOnce installed, the Devops-in-a-box
          operator provides the following features:\n\n* **Create/Destroy**: Easily
          launch a new DevSecOpsManager instance to bootstrap your cluster and start
          deploying new applications with total confidence.\n\n* **Gitops Simplified**:
          Configure your cluster and individual applications through Git and watch
          your changes propagate instantly. Just point to where your code is and let
          the operator do the rest with preconfigured cloud-native CI/CD pipelines.\n\n*
          **Production-Ready**: Automatically register your applications with Openshift
          logging and monitoring systems and enjoy special additions for Spring Boot
          like custom dashboards and searchable logs.\n\n*Contact us on Github for
          installation and usage instruction*\n\n### Other Supported Features\n\n**Blue/Green,
          Canary, A/B Deployments**\n\nRun multiple versions of your applications
          simultaneously and configure traffic rules to target specific groups of
          users using Service Mesh.\n\n**Updates via automated operations**\n\nGet
          the latest enhancements automatically through the Operator Lifecycle Manager
          and upgrade all your running applications at once. "
        displayName: Devops-in-a-box
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devops
        - devsecops
        - spring
        - boot
        - java
        - gitops
        - cicd
        - argo
        links:
        - name: Devops-in-a-box Operator
          url: https://github.com/prft-rh/devopsinabox-operator
        maintainers:
        - email: matthieu.rethers@perficient.com
          name: Matthieu Rethers
        maturity: alpha
        provider:
          name: Perficient Inc
          url: http://www.perficient.com
        relatedImages:
        - quay.io/mrethers/devopsinabox-operator:v0.0.45
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        version: 0.0.45
      entries:
      - name: devopsinabox.v0.0.45
        version: 0.0.45
      name: alpha
    defaultChannel: alpha
    packageName: devopsinabox
    provider:
      name: Perficient Inc
      url: http://www.perficient.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: dapr.io
      provider-url: https://dapr.io
    name: dapr-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dapr-kubernetes-operator.v0.0.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprControlPlane",
                "metadata": {
                  "name": "dapr-control-plane",
                  "namespace": "dapr-system"
                },
                "spec": {
                  "values": {}
                }
              },
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprCruiseControl",
                "metadata": {
                  "name": "dapr-cruise-control",
                  "namespace": "dapr-system"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprInstance",
                "metadata": {
                  "name": "dapr-instance",
                  "namespace": "dapr-system"
                },
                "spec": {
                  "values": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          containerImage: docker.io/daprio/kubernetes-operator@sha256:da599a273943bf28812314a11df1052411237dd1887edf0fe6104c5857f439b0
          createdAt: "2025-10-06T18:47:59Z"
          description: Dapr Control Plane Operator
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/dapr/kubernetes-operator
          support: contact@dapr.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DaprControlPlane is the Schema for the Dapr ControlPlane
              API.
            displayName: DaprControlPlane
            kind: DaprControlPlane
            name: daprcontrolplanes.operator.dapr.io
            version: v1alpha1
          - kind: DaprCruiseControl
            name: daprcruiscontrols.operator.dapr.io
            version: v1alpha1
          - description: DaprInstance is the Schema for the Dapr Instance API.
            displayName: DaprInstance
            kind: DaprInstance
            name: daprinstances.operator.dapr.io
            version: v1alpha1
        description: Dapr Control Plane Operator
        displayName: Dapr Control Plane Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dapr
        links:
        - name: Dapr
          url: https://dapr.io
        - name: Dapr Kubernetes Operator
          url: https://github.com/dapr/kubernetes-operator
        maintainers:
        - email: lburgazzoli@gmail.com
          name: Luca Burgazzoli
        - email: salaboy@gmail.com
          name: Mauricio Salatino
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: dapr.io
          url: https://dapr.io
        relatedImages:
        - docker.io/daprio/kubernetes-operator@sha256:da599a273943bf28812314a11df1052411237dd1887edf0fe6104c5857f439b0
        version: 0.0.12
      entries:
      - name: dapr-kubernetes-operator.v0.0.12
        version: 0.0.12
      - name: dapr-kubernetes-operator.v0.0.11
        version: 0.0.11
      - name: dapr-kubernetes-operator.v0.0.10
        version: 0.0.10
      - name: dapr-kubernetes-operator.v0.0.9
        version: 0.0.9
      - name: dapr-kubernetes-operator.v0.0.8
        version: 0.0.8
      - name: dapr-kubernetes-operator.v0.0.6
        version: 0.0.6
      - name: dapr-kubernetes-operator.v0.0.5
        version: 0.0.5
      - name: dapr-kubernetes-operator.v0.0.4
        version: 0.0.4
      - name: dapr-kubernetes-operator.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: dapr-kubernetes-operator
    provider:
      name: dapr.io
      url: https://dapr.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: gitwebhook-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitwebhook-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitWebhook",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "gitwebhook-operator",
                    "app.kubernetes.io/instance": "gitwebhook-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gitwebhook",
                    "app.kubernetes.io/part-of": "gitwebhook-operator"
                  },
                  "name": "gitwebhook-sample"
                },
                "spec": null
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/redhat-cop/gitwebhook-operator@sha256:2810d75eebb24069dbbedf6a82285136c6f16b4cc58b59f1b0fbdc38560cf8a8
          createdAt: "2024-05-10T18:17:13Z"
          description: An operator to create git webhooks declaratively
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: gitwebhook-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/gitwebhook-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitWebhook is the Schema for the gitwebhooks API
            displayName: Git Webhook
            kind: GitWebhook
            name: gitwebhooks.redhatcop.redhat.io
            version: v1alpha1
        description: "This operator automates the setting up of git webhooks. Its
          main purpose is to stream line the configuration of operators such as Tekton
          or ArgoCD which rely on git webhooks for some of their functionalities.\n\n##
          The GitWebhook CRD\n \nA CRD is provided by this operator to create a git
          web hook, here is an example for github\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GitWebhook\nmetadata:\n  name: gitwebhook-github\nspec:\n  gitHub:\n    gitServerCredentials:\n
          \     name: github-pat\n  repositoryOwner: ${repo_owner}\n  ownerType: user\n
          \ repositoryName: ${repo_name}\n  webhookURL: https://hellowebhook.com\n
          \ insecureSSL: false\n  webhookSecret:\n    name: webhook-secret\n  events:\n
          \   - push  \n  contentType: json\n  active: true  \n```    \n\nhere is
          an explanation of each field:\n\n- `gihub` specifies how to connect to the
          git api server. It also requires a local reference to a secret (in the same
          namespace) containing a key `token` with a valid github token to be used
          to authenticate. A similar `gitLab` section exists when connecting to gitlab.
          Only one of `gitLab` or `gitHub` can be defined. \n- `repositoryOwner` and
          `repositoryName` identify the repository for which we want to receive events.\n-
          `ownerType` can have two values: `user` and `organization` and identifies
          the kind of owner.\n- `webhookURL` is the URL for to be called.\n - `insecureSSL`
          defines whether the target URL certificate should be validated (default
          `false`). \n- `webhookSecret` defines a local reference to a secret containing
          the `secret` key. The value is a shared secret between the webhook caller
          and the received for farther validation or identification of the caller.\n-
          `events` is the list of the repo-level events that the webhook should generate.
          The list of valid events for github can be found [here](https://docs.github.com/en/developers/webhooks-and-events/webhooks/webhook-events-and-payloads).
          The list of valid events for gitlab can be found [here](https://docs.gitlab.com/ee/user/project/integrations/webhook_events.html).\n
          - `contentType` defines the format of the webhook payload (default `json`)
          (github).\n - `active` whether the webhook should be turned on (default
          `true`) (github).\n - `pushEventBranchFilter` a regular expression to filter
          from which branches push events should be generated (gitlab only).\n\n##
          Security Considerations\n\nThis operator does not own credentials for the
          git server, but instead always allocate a new connection based on the credentials
          referenced in the CR and every reconcile cycle. As a result there is no
          risk of security escalation or credential leaking between tenants of a cluster
          using this operator. On the other hand it is the responsibility of the namespace
          owners or the platform owner to ensure that valid git credentials are always
          available in the namespace where the GitWebhook CRs need to defined.\n\n##
          Current support\n\nCurrently this operator support creating repo-level webhooks
          for github and gitlab. Potentially this operator could be extended to support
          org-level webhook or other git systems. Contributions are welcome.  "
        displayName: GitWebhook operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - webhook
        - github
        - gitlab
        links:
        - name: GitWbehook Operator
          url: https://github.com/redhat-cop/gitwebhook-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/gitwebhook-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/gitwebhook-operator@sha256:2810d75eebb24069dbbedf6a82285136c6f16b4cc58b59f1b0fbdc38560cf8a8
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        version: 0.1.2
      entries:
      - name: gitwebhook-operator.v0.1.2
        version: 0.1.2
      - name: gitwebhook-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: gitwebhook-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ec2-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ec2-controller.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "DHCPOptions",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "ElasticIPAddress",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "Instance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "InternetGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "NATGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "RouteTable",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "SecurityGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "Subnet",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "TransitGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "VPC",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "VPCEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ec2-controller:1.7.0
          createdAt: "2025-09-27T00:04:55Z"
          description: AWS EC2 controller is a service controller for managing EC2
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CapacityReservation represents the state of an AWS ec2 CapacityReservation
              resource.
            displayName: CapacityReservation
            kind: CapacityReservation
            name: capacityreservations.ec2.services.k8s.aws
            version: v1alpha1
          - description: DHCPOptions represents the state of an AWS ec2 DHCPOptions
              resource.
            displayName: DHCPOptions
            kind: DHCPOptions
            name: dhcpoptions.ec2.services.k8s.aws
            version: v1alpha1
          - description: ElasticIPAddress represents the state of an AWS ec2 ElasticIPAddress
              resource.
            displayName: ElasticIPAddress
            kind: ElasticIPAddress
            name: elasticipaddresses.ec2.services.k8s.aws
            version: v1alpha1
          - description: FlowLog represents the state of an AWS ec2 FlowLog resource.
            displayName: FlowLog
            kind: FlowLog
            name: flowlogs.ec2.services.k8s.aws
            version: v1alpha1
          - description: Instance represents the state of an AWS ec2 Instance resource.
            displayName: Instance
            kind: Instance
            name: instances.ec2.services.k8s.aws
            version: v1alpha1
          - description: InternetGateway represents the state of an AWS ec2 InternetGateway
              resource.
            displayName: InternetGateway
            kind: InternetGateway
            name: internetgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: LaunchTemplate represents the state of an AWS ec2 LaunchTemplate
              resource.
            displayName: LaunchTemplate
            kind: LaunchTemplate
            name: launchtemplates.ec2.services.k8s.aws
            version: v1alpha1
          - description: NATGateway represents the state of an AWS ec2 NATGateway
              resource.
            displayName: NATGateway
            kind: NATGateway
            name: natgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: NetworkACL represents the state of an AWS ec2 NetworkACL
              resource.
            displayName: NetworkACL
            kind: NetworkACL
            name: networkacls.ec2.services.k8s.aws
            version: v1alpha1
          - description: RouteTable represents the state of an AWS ec2 RouteTable
              resource.
            displayName: RouteTable
            kind: RouteTable
            name: routetables.ec2.services.k8s.aws
            version: v1alpha1
          - description: SecurityGroup represents the state of an AWS ec2 SecurityGroup
              resource.
            displayName: SecurityGroup
            kind: SecurityGroup
            name: securitygroups.ec2.services.k8s.aws
            version: v1alpha1
          - description: Subnet represents the state of an AWS ec2 Subnet resource.
            displayName: Subnet
            kind: Subnet
            name: subnets.ec2.services.k8s.aws
            version: v1alpha1
          - description: TransitGateway represents the state of an AWS ec2 TransitGateway
              resource.
            displayName: TransitGateway
            kind: TransitGateway
            name: transitgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: TransitGatewayVPCAttachment represents the state of an AWS
              ec2 TransitGatewayVPCAttachment resource.
            displayName: TransitGatewayVPCAttachment
            kind: TransitGatewayVPCAttachment
            name: transitgatewayvpcattachments.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCEndpoint represents the state of an AWS ec2 VPCEndpoint
              resource.
            displayName: VPCEndpoint
            kind: VPCEndpoint
            name: vpcendpoints.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCEndpointServiceConfiguration represents the state of an
              AWS ec2 VPCEndpointServiceConfiguration resource.
            displayName: VPCEndpointServiceConfiguration
            kind: VPCEndpointServiceConfiguration
            name: vpcendpointserviceconfigurations.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCPeeringConnection represents the state of an AWS ec2 VPCPeeringConnection
              resource.
            displayName: VPCPeeringConnection
            kind: VPCPeeringConnection
            name: vpcpeeringconnections.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPC represents the state of an AWS ec2 VPC resource.
            displayName: VPC
            kind: VPC
            name: vpcs.ec2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Compute Cloud (EC2) resources in AWS from within your Kubernetes cluster.

          **About Amazon EC2**

          Amazon Elastic Compute Cloud (Amazon EC2) provides scalable computing capacity in the Amazon Web Services (AWS) Cloud. Using Amazon EC2 eliminates your need to invest in hardware up front, so you can develop and deploy applications faster. You can use Amazon EC2 to launch as many or as few virtual servers as you need, configure security and networking, and manage storage. Amazon EC2 enables you to scale up or down to handle changes in requirements or spikes in popularity, reducing your need to forecast traffic.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EC2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ec2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EC2 Developer Resources
          url: https://aws.amazon.com/ec2/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ec2 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ec2-controller:1.7.0
        version: 1.7.0
      entries:
      - name: ack-ec2-controller.v1.7.0
        version: 1.7.0
      - name: ack-ec2-controller.v1.6.2
        version: 1.6.2
      - name: ack-ec2-controller.v1.6.1
        version: 1.6.1
      - name: ack-ec2-controller.v1.6.0
        version: 1.6.0
      - name: ack-ec2-controller.v1.4.11
        version: 1.4.11
      - name: ack-ec2-controller.v1.4.10
        version: 1.4.10
      - name: ack-ec2-controller.v1.4.9
        version: 1.4.9
      - name: ack-ec2-controller.v1.4.8
        version: 1.4.8
      - name: ack-ec2-controller.v1.4.7
        version: 1.4.7
      - name: ack-ec2-controller.v1.4.6
        version: 1.4.6
      - name: ack-ec2-controller.v1.4.5
        version: 1.4.5
      - name: ack-ec2-controller.v1.4.4
        version: 1.4.4
      - name: ack-ec2-controller.v1.4.2
        version: 1.4.2
      - name: ack-ec2-controller.v1.4.1
        version: 1.4.1
      - name: ack-ec2-controller.v1.4.0
        version: 1.4.0
      - name: ack-ec2-controller.v1.3.7
        version: 1.3.7
      - name: ack-ec2-controller.v1.3.6
        version: 1.3.6
      - name: ack-ec2-controller.v1.3.5
        version: 1.3.5
      - name: ack-ec2-controller.v1.3.4
        version: 1.3.4
      - name: ack-ec2-controller.v1.3.3
        version: 1.3.3
      - name: ack-ec2-controller.v1.3.1
        version: 1.3.1
      - name: ack-ec2-controller.v1.3.0
        version: 1.3.0
      - name: ack-ec2-controller.v1.2.28
        version: 1.2.28
      - name: ack-ec2-controller.v1.2.27
        version: 1.2.27
      - name: ack-ec2-controller.v1.2.26
        version: 1.2.26
      - name: ack-ec2-controller.v1.2.25
        version: 1.2.25
      - name: ack-ec2-controller.v1.2.24
        version: 1.2.24
      - name: ack-ec2-controller.v1.2.23
        version: 1.2.23
      - name: ack-ec2-controller.v1.2.22
        version: 1.2.22
      - name: ack-ec2-controller.v1.2.21
        version: 1.2.21
      - name: ack-ec2-controller.v1.2.20
        version: 1.2.20
      - name: ack-ec2-controller.v1.2.17
        version: 1.2.17
      - name: ack-ec2-controller.v1.2.16
        version: 1.2.16
      - name: ack-ec2-controller.v1.2.15
        version: 1.2.15
      - name: ack-ec2-controller.v1.2.14
        version: 1.2.14
      - name: ack-ec2-controller.v1.2.13
        version: 1.2.13
      - name: ack-ec2-controller.v1.2.12
        version: 1.2.12
      - name: ack-ec2-controller.v1.2.11
        version: 1.2.11
      - name: ack-ec2-controller.v1.2.10
        version: 1.2.10
      - name: ack-ec2-controller.v1.2.9
        version: 1.2.9
      - name: ack-ec2-controller.v1.2.8
        version: 1.2.8
      - name: ack-ec2-controller.v1.2.7
        version: 1.2.7
      - name: ack-ec2-controller.v1.2.6
        version: 1.2.6
      - name: ack-ec2-controller.v1.2.4
        version: 1.2.4
      - name: ack-ec2-controller.v1.2.1
        version: 1.2.1
      - name: ack-ec2-controller.v1.2.0
        version: 1.2.0
      - name: ack-ec2-controller.v1.1.2
        version: 1.1.2
      - name: ack-ec2-controller.v1.1.1
        version: 1.1.1
      - name: ack-ec2-controller.v1.1.0
        version: 1.1.0
      - name: ack-ec2-controller.v1.0.7
        version: 1.0.7
      - name: ack-ec2-controller.v1.0.6
        version: 1.0.6
      - name: ack-ec2-controller.v1.0.4
        version: 1.0.4
      - name: ack-ec2-controller.v1.0.3
        version: 1.0.3
      - name: ack-ec2-controller.v1.0.2
        version: 1.0.2
      - name: ack-ec2-controller.v1.0.1
        version: 1.0.1
      - name: ack-ec2-controller.v1.0.0
        version: 1.0.0
      - name: ack-ec2-controller.v0.1.1
        version: 0.1.1
      - name: ack-ec2-controller.v0.1.0
        version: 0.1.0
      - name: ack-ec2-controller.v0.0.21
        version: 0.0.21
      - name: ack-ec2-controller.v0.0.20
        version: 0.0.20
      - name: ack-ec2-controller.v0.0.19
        version: 0.0.19
      - name: ack-ec2-controller.v0.0.18
        version: 0.0.18
      - name: ack-ec2-controller.v0.0.17
        version: 0.0.17
      - name: ack-ec2-controller.v0.0.16
        version: 0.0.16
      - name: ack-ec2-controller.v0.0.15
        version: 0.0.15
      - name: ack-ec2-controller.v0.0.14
        version: 0.0.14
      - name: ack-ec2-controller.v0.0.13
        version: 0.0.13
      - name: ack-ec2-controller.v0.0.12
        version: 0.0.12
      - name: ack-ec2-controller.v0.0.11
        version: 0.0.11
      - name: ack-ec2-controller.v0.0.10
        version: 0.0.10
      - name: ack-ec2-controller.v0.0.9
        version: 0.0.9
      - name: ack-ec2-controller.v0.0.8
        version: 0.0.8
      - name: ack-ec2-controller.v0.0.7
        version: 0.0.7
      - name: ack-ec2-controller.v0.0.6
        version: 0.0.6
      - name: ack-ec2-controller.v0.0.5
        version: 0.0.5
      - name: ack-ec2-controller.v0.0.4
        version: 0.0.4
      - name: ack-ec2-controller.v0.0.3
        version: 0.0.3
      - name: ack-ec2-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-ec2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: global-load-balancer-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: global-load-balancer-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSRecord",
                "metadata": {
                  "name": "hello-global-record"
                },
                "spec": {
                  "endpoints": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    }
                  ],
                  "globalZoneRef": {
                    "name": "external-dns-zone"
                  },
                  "loadBalancingPolicy": "Multivalue",
                  "name": "hello.my.base.domain",
                  "ttl": 60
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSRecord",
                "metadata": {
                  "name": "route53-geoproximity-global-record-with-healthcheck"
                },
                "spec": {
                  "endpoints": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    }
                  ],
                  "globalZoneRef": {
                    "name": "route53-global-dns-zone"
                  },
                  "healthCheck": {
                    "failureThreshold": 3,
                    "httpGet": {
                      "host": "geoproximity-hc.${global_base_domain}",
                      "path": "/healthz",
                      "port": 80,
                      "scheme": "HTTP"
                    },
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 2
                  },
                  "loadBalancingPolicy": "Geoproximity",
                  "name": "geoproximity-hc.my.base.domain",
                  "ttl": 60
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSZone",
                "metadata": {
                  "name": "external-dns-zone"
                },
                "spec": {
                  "domain": "myglobalzone.mybasedomain.io",
                  "provider": {
                    "externalDNS": {
                      "annotations": {
                        "type": "global"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSZone",
                "metadata": {
                  "name": "route53-global-dns-zone"
                },
                "spec": {
                  "domain": "myglobalzone.mybasedomain.io",
                  "provider": {
                    "route53": {
                      "credentialsSecretRef": {
                        "name": "route53-global-zone-credentials",
                        "namespace": "mynamespace"
                      },
                      "zoneID": "Z083892239ROKNEP70BDY"
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalRouteDiscovery",
                "metadata": {
                  "name": "external-dns-globalroutediscovery"
                },
                "spec": {
                  "clusters": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3"
                    }
                  ],
                  "defaultLoadBalancingPolicy": "Multivalue",
                  "defaultTTL": 30,
                  "globalZoneRef": {
                    "name": "external-dns-zone"
                  },
                  "routeSelector": {
                    "matchLabels": {
                      "route-type": "global"
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalRouteDiscovery",
                "metadata": {
                  "name": "route53-globalroutediscovery"
                },
                "spec": {
                  "clusters": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3"
                    }
                  ],
                  "defaultLoadBalancingPolicy": "Multivalue",
                  "defaultTTL": 30,
                  "globalZoneRef": {
                    "name": "route53-global-dns-zone"
                  },
                  "routeSelector": {
                    "matchLabels": {
                      "route-type": "global"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Cloud Provider
          certified: "false"
          containerImage: quay.io/redhat-cop/global-load-balancer-operator@sha256:c0b6ac9b03571a74a56b28302d720e8aa4442233f7867d05ed365e022f7cb8f5
          createdAt: "2021-12-02T17:33:23Z"
          description: This operator creates automation around a DNS to operate as
            a global load balancer for a set of OpenShift clusters.
          operatorframework.io/suggested-namespace: global-load-balancer-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/global-load-balancer-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DNSEndpoint
            name: dnsendpoints.externaldns.k8s.io
            version: v1alpha1
          - description: GlobalDNSRecord is the Schema for the globaldnsrecords API
            displayName: Global DNSRecord
            kind: GlobalDNSRecord
            name: globaldnsrecords.redhatcop.redhat.io
            version: v1alpha1
          - description: GlobalDNSZone is the Schema for the globaldnszones API
            displayName: Global DNSZone
            kind: GlobalDNSZone
            name: globaldnszones.redhatcop.redhat.io
            version: v1alpha1
          - description: GlobalRouteDiscovery is the Schema for the globalroutediscoveries
              API
            displayName: Global Route Discovery
            kind: GlobalRouteDiscovery
            name: globalroutediscoveries.redhatcop.redhat.io
            version: v1alpha1
        description: "The global-load-balancer-operator implements automation to program
          a DNS service to act as global load balancer for applications deployed to
          multiple OpenShift clusters.\nThis operator is designed to be deployed to
          a control cluster which will watch the load balanced clusters (controlled
          clusters).\nThere are two main concepts (APIs) provided by this operator:\n\n1.
          GlobalDNSZone\n2. GlobalDNSRecord\n\n## GlobalDNSZone\n\nThe `GlobalDNSZone`
          CR allows you to configure a zone which will contain global load balanced
          records and the provider used to populate it.\nHere is an example of GlobalDNSZone:\n\n```yaml\napiVersion:
          redhatcop.redhat.io/v1alpha1\nkind: GlobalDNSZone\nmetadata:\n  name: external-dns-zone\nspec:\n
          \ # Add fields here\n  domain: global.myzone.io\n  provider:\n    externalDNS:\n
          \     annotations:\n        type: global\n```\n\nHere is a table summarizing
          the supported providers and their capabilities:\n\n| Provider  | Supports
          Health Checks  | Supports Multivalue LB | Supports Latency LB  | Supports
          GeoProximity LB  |\n|:--:|:--:|:---:|:---:|:---:|\n| External-dns  | no
          \ | yes | no  | no  |\n| Route53  | yes | yes(*)  | yes(*)  | yes(*)  |\n\n(*)
          only if all controlled clusters run on AWS.\n\n## GlobalDNSRecord\n\nThe
          `GlobalDNSRecord` CR allows you to specify the intention to create a global
          dns record. Here is an example\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GlobalDNSRecord\nmetadata:\n  name: hello-global-record\nspec:\n  name:
          hello.global.myzone.io\n  endpoints:\n  - clusterName: cluster1\n    clusterCredentialRef:\n
          \     name: kubeconfig\n      namespace: cluster1\n    loadBalancerServiceRef:\n
          \     name: router-default\n      namespace: openshift-ingress\n  - clusterName:
          cluster2\n    clusterCredentialRef:\n      name: kubeconfig\n      namespace:
          cluster2\n    loadBalancerServiceRef:\n      name: router-default\n      namespace:
          openshift-ingress\n  - clusterName: cluster3\n    clusterCredentialRef:\n
          \     name: kubeconfig\n      namespace: cluster3\n    loadBalancerServiceRef:\n
          \     name: router-default\n      namespace: openshift-ingress\n  ttl: 60\n
          \ loadBalancingPolicy: Multivalue\n  globalZoneRef:\n    name: external-dns-zone\n
          \ healthCheck:\n    failureThreshold: 3\n    httpGet:\n      host: hello.global.myzone.io\n
          \     httpHeaders:\n        - name: ciao\n          value: hello\n      path:
          /readyz\n      port: 80\n      scheme: HTTP\n    periodSeconds: 10\n    successThreshold:
          1\n    timeoutSeconds: 1     \n```\n\nThe list of endpoints identifies the
          set of `LoadBalancer` type of services to watch on the remote clusters (at
          the moment only LoadBalancer services are supported). These are the `LoadBalancer`
          services created by the [ingress controller](https://docs.openshift.com/container-platform/4.5/networking/ingress-operator.html)
          on which the routers rely. Each entry of this list requires a reference
          to the loadbalancer service on the remote cluster and a reference to a local
          secret containing the credential to connect to the remote cluster.\n\nThis
          secret must contain one entry called `kubeconfig` with a kubeconfig file
          with a default context pointing to the remote cluster. The account used
          by the kubeconfig file (presumably a service account) will require at a
          minimum cluster-level permissions as described in this [cluster role](./docs/scripts/remote-account-cluster-role.yaml).\n\nThe
          `globalZoneRef` field refers to a local (same namespace) `GlobalZone` CR.
          The DNS record represented by this `GlobalDNSRecord`, will be created in
          the referenced zone.\n\n`ttl` is the TTL of the crated DNS record.\n\n`loadBalancingPolicy`
          is the load balancing policy for this global record. It must match one of
          the policy supported by the provider of the referenced `GlobalZone`.\n\nFinally,
          `healthcheck` represent a [probe](https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.18/#probe-v1-core)
          to be used to test the health of a record. This field will be ignored if
          the provider does not support health checks.\n\n## External DNS Provider\n\nThe
          [external-dns](https://github.com/kubernetes-sigs/external-dns) provider
          delegates to external-dns the creation of the actual DNS records by creating
          a `DNSEndpoint` object.\nThe `DNSEndpoint` object will be created in the
          same namespace as the `GlobalDNSRecord` and will be owned by it.\nThe `DNSEdnpoint`
          object will have the same labels as the `GlobalDNSRecord` and the annotations
          specified in the GlobalDNSZone configuration.\nExternal-dns should be configured
          to watch for DNSEnpoints at the cluster level and to point to the desired
          provider.\nDetails on configuration can be found at the external-dns git
          repository.\nThe External-dns should be used as a fall back option when
          other options are not available as it does not support health checks and
          advanced load balancing policies.\n\n## AWS Route53 provider\n\nAWS Route53
          provider uses the Route53 service as a global loadbalancer and offers advanced
          routing capabilities via [route53 traffic policies](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/traffic-flow.html)
          (note that traffic policies will trigger an expense).\nThe following routing
          polices are currently supported:\n\n1. [Multivalue](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-multivalue)\n2.
          [Geoproximity](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-geoproximity)\n3.
          [Latency](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-latency)\n\nAWS
          Route53 provider at the moment requires that all the controlled clusters
          run in AWS.\n\nIf health checks are defined, a route53 health check originating
          from any reason (you have to ensure connectivity) will be created for each
          of the endpoint. Because the endpoint represent s a shared ELB (shared with
          other apps, that is) and the health check is app specific, we cannot use
          the ELB health check, so the route53 endpoint is created with one of the
          two IP exposed by the ELB. This is suboptimal, but it works in most situations.\n\n##
          Global Route Auto Discovery\n\nThe aboev examples showed how to create global
          DNS records. This can be good in some situations, but most of the times
          in an openshift deployment global DNS records will point to routes that
          are intended to be global.\nThe global-load-balancer operator can auto-discover
          these routes and automatically create the corresponding `GloablDNSRecord`.
          The `GlobalRouteDiscovery` CRD is used to configure the discovery process,
          here is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GlobalRouteDiscovery\nmetadata:\n  name: route53-globalroutediscovery\nspec:\n
          \ clusters:\n  - clusterName: cluster1\n    clusterCredentialRef:\n      name:
          ${cluster1_secret_name}\n      namespace: cluster1\n...\n  routeSelector:\n
          \   matchLabels:\n      route-type: global\n  defaultLoadBalancingPolicy:
          Multivalue\n  defaultTTL: 30 \n  globalZoneRef:\n    name: route53-global-dns-zone\n```\n\nThis
          global discovery route will discover routes in the provided list of cluster.
          Only the routes that match the route selector will be considered global.
          \nThe default load balancing policy and default TTL can be expressed in
          the `GlobalRouteDiscovery` CR. However with the following annotations, it's
          possible to configure route-specific values:\n\n* `global-load-balancer-operator.redhat-cop.io/load-balancing-policy`
          to set the load balancing policy\n* `global-load-balancer-operator.redhat-cop.io/ttl`
          to set the TTL\n\nThe globalZoneRef refers to the global zone to be used
          for the created `GlobalDNSRecords`.\n\nHealth checks will also be automatically
          discovered. If the pods behind the route expose a readiness check of httpGet
          kind, that configuration will be used to create the GlobalDNSRecord health
          check.\nWhen more than one container is present in the pod, by default the
          first container will be examined for health check. This behavior can be
          overridden with the this annotation on the route: `global-load-balancer-operator.redhat-cop.io/container-probe`
          where the value will container the name of the container with teh correct
          readiness probe.\n\nIf routes with the same namespace and name exists in
          multiple cluster, the following conditions must be met:\n\n* all host names
          must be the same\n* all load balancing policy must be the same\n* all TTLs
          must be the same\n* all discovered readiness checks must be the same\n\n##
          Examples\n\nThese examples are intended to help you setting up working configuration
          with each of the providers\n\n### Cluster Set up\n\nTwo approaches for cluster
          setup are provided\n\n1. [One cluster, three ingress-gateways.](https://github.com/redhat-cop/global-load-balancer-operator/docs/one-cluster-three-ingress-gateways.md)
          This approach is intended for development purposes and has the objective
          to keep resource consumption at the minimum.\n2. [Control cluster and three
          controlled clusters in different regions.](https://github.com/redhat-cop/global-load-balancer-operator/docs/three-clusters.md).
          This approach represents a more realistic set-up albeit it consumes more
          resources.\n\nYou can also set up the cluster on your own, at the end the
          following conditions must be met:\n\nThree namespace `cluster1` `cluster2`
          `cluster3` are created.\nthe following environment variables are initialized
          for each cluster:\n\n1. <cluster>_secret_name. Pointing to a secret in each
          of the cluster namespaces containing a valid kubeconfig fot that cluster\n2.
          <cluster>_service_name.  Pointing to the name of the loadbalancer service
          to be used for that cluster.\n3. <cluster>_service_namespace. Pointing to
          the namespace of the loadbalancer service to be used for that cluster.\n\nHere
          are examples for the supported provider:\n\n1. [Setting up external-dns
          as provider](https://github.com/redhat-cop/global-load-balancer-operator/docs/external-dns-provider.md)\n2.
          [Setting up route53 as a provider](https://github.com/redhat-cop/global-load-balancer-operator/docs/aws-route53-provider.md)\n"
        displayName: Global Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Networking
        - Cloud Provider
        - Global Load Balancing
        - GLB
        - GTM
        - route53
        - DNS
        links:
        - name: GigHub Repository
          url: https://github.com/redhat-cop/global-load-balancer-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/global-load-balancer-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/global-load-balancer-operator@sha256:c0b6ac9b03571a74a56b28302d720e8aa4442233f7867d05ed365e022f7cb8f5
        version: 1.4.2
      entries:
      - name: global-load-balancer-operator.v1.4.2
        version: 1.4.2
      - name: global-load-balancer-operator.v1.4.1
        version: 1.4.1
      - name: global-load-balancer-operator.v1.4.0
        version: 1.4.0
      - name: global-load-balancer-operator.v1.3.0
        version: 1.3.0
      - name: global-load-balancer-operator.v1.2.0
        version: 1.2.0
      - name: global-load-balancer-operator.v1.1.0
        version: 1.1.0
      - name: global-load-balancer-operator.v1.0.7
        version: 1.0.7
      - name: global-load-balancer-operator.v1.0.6
        version: 1.0.6
      - name: global-load-balancer-operator.v1.0.5
        version: 1.0.5
      - name: global-load-balancer-operator.v1.0.4
        version: 1.0.4
      - name: global-load-balancer-operator.v1.0.3
        version: 1.0.3
      - name: global-load-balancer-operator.v1.0.2
        version: 1.0.2
      - name: global-load-balancer-operator.v1.0.1
        version: 1.0.1
      - name: global-load-balancer-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: global-load-balancer-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Netscaler
      provider-url: https://www.netscaler.com
    name: netscaler-adm-agent-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netscaler-adm-agent-operator.v141.37.44
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "AdmAgent",
                "metadata": {
                  "name": "admagent"
                },
                "spec": {
                  "accessSecret": "",
                  "admURL": "adm.cloud.com",
                  "affinity": {},
                  "clusterPrefix": "",
                  "core": {
                    "livenessProbe": {
                      "failureThreshold": 5,
                      "initialDelaySeconds": 600,
                      "periodSeconds": 60,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "readinessProbe": {
                      "failureThreshold": 30,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 30,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "100m",
                        "memory": "2000Mi"
                      }
                    }
                  },
                  "fullnameOverride": "",
                  "helper": {
                    "image": "{{ .Values.helper.imageRegistry }}/{{ .Values.helper.imageRepository }}:{{ .Values.helper.imageTag }}",
                    "imageRegistry": "quay.io",
                    "imageRepository": "citrix/adm-agent-helper",
                    "imageTag": "2.0.1",
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "200Mi"
                      }
                    }
                  },
                  "hpa": {
                    "maxReplicas": 10,
                    "minReplicas": 1,
                    "required": false,
                    "targetCPUUtilization": "1000m",
                    "targetMemoryUtilization": "4000Mi"
                  },
                  "image": "{{ .Values.imageRegistry }}/{{ .Values.imageRepository }}:{{ .Values.imageTag }}",
                  "imagePullSecrets": [],
                  "imageRegistry": "quay.io",
                  "imageRepository": "citrix/adm-agent",
                  "imageTag": "14.1-37.44",
                  "kad": {
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "50m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "lic": {
                    "livenessProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 600,
                      "periodSeconds": 20,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "100m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "loginSecret": "",
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "redis": {
                    "livenessProbe": {
                      "failureThreshold": 5,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "readinessProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 20,
                      "periodSeconds": 10,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true
                  },
                  "sharding": {
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-adm-agent-operator@sha256:1c462c2d6161877f87d5eed4888f8d94c733501e67a94757ab0873d983fb3a4a
          createdAt: "2024-11-18T06:40:58Z"
          description: This is an operator to install ADM Agent
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler ADM Agent Operator
            displayName: NetScaler ADM Agent Operator
            kind: AdmAgent
            name: admagents.netscaler.com
            version: v1
        description: Netscaler ADM Agent Operator
        displayName: Netscaler ADM Agent Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        - ADC
        - Kubernetes
        - OpenShift
        - Citrix
        - Controller
        - Observability
        - Metrices
        - adm
        - agent
        - monitoring
        - proxy
        links:
        - name: Netscaler Adm Agent Operator
          url: https://github.com/netscaler/netscaler-helm-charts/blob/master/adm-agent/README.md
        maintainers:
        - email: pavan.belani@cloud.com
          name: Pavan Belani
        - email: swapnil.khanderaokakade@cloud.com
          name: Swapnil Khanderao
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Netscaler
          url: https://www.netscaler.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6b8c101124579dd46086ec83715618dd9a6aebe66176d676242e6c0b0900a53c
        - registry.connect.redhat.com/citrix/netscaler-adm-agent-operator@sha256:1c462c2d6161877f87d5eed4888f8d94c733501e67a94757ab0873d983fb3a4a
        - quay.io/citrix/adm-agent@sha256:45cab90524687c76a6d53f014b55063fa25c8aaeec0261f0f4f0efa6bff42968
        - quay.io/citrix/adm-agent-helper@sha256:f567a0637e8024f9d1674977f6c013741cc0fd12cf78703f8dbe01126561c3c4
        version: 141.37.44
      entries:
      - name: netscaler-adm-agent-operator.v141.37.44
        version: 141.37.44
      - name: netscaler-adm-agent-operator.v141.31.38
        version: 141.31.38
      - name: netscaler-adm-agent-operator.v141.25.33
        version: 141.25.33
      - name: netscaler-adm-agent-operator.v141.23.31
        version: 141.23.31
      - name: netscaler-adm-agent-operator.v141.21.41
        version: 141.21.41
      - name: netscaler-adm-agent-operator.v141.16.39
        version: 141.16.39
      - name: netscaler-adm-agent-operator.v141.13.23
        version: 141.13.23
      name: stable
    defaultChannel: stable
    packageName: netscaler-adm-agent-operator
    provider:
      name: Netscaler
      url: https://www.netscaler.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Redis
      provider-url: ""
    name: redis-enterprise-operator-cert-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-enterprise-operator.v6.2.10-45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:ffdf10a1b08115fd4de06f1cd751d36644a9bc9a8c11034cd432e9b66a5d2f73
        - registry.connect.redhat.com/redislabs/services-manager@sha256:8b46f2aacfb3e1b243c0b64975accc55fd80b4de9c9ac4769c3c67a5676addc5
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
        version: 6.2.10-45
      entries:
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: 6.2.10
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.22.0
    - currentCSV: redis-enterprise-operator.v7.22.0-11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:088e4b57c6e1ca1680b359078e5eb1bba67343cdbd9a539af26a0ea23a830593
        - registry.connect.redhat.com/redislabs/services-manager@sha256:3b96cdc65c3e42fd22eb165950be96b907725903ac74fe4f82cfa0073e28e95a
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:d56ba97e95e799210e76615095409143a5dddd5c5a1e1be1d551d7e3f9e4d3ef
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:0cfb07f774db131b5b74e5aa91291579c916c4d4cd1a693d7aaed554f0177a68
        version: 7.22.0-11.3
      entries:
      - name: redis-enterprise-operator.v7.22.0-11.3
        version: 7.22.0-11.3
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.22.0-11.3
    - currentCSV: redis-enterprise-operator.v7.4.6-7.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:473b5661049a6c8830ba113504975726154c7b172376ba3b63d891674d59c78f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5f6f32a63f4dd8be85b9d469f0fedd26d6aa76e32b7735bfa7ab7203cb272cdb
        - registry.connect.redhat.com/redislabs/services-manager@sha256:ba20ab644a473bc742af671171d5313537292d88942b16336d03a6386c061d8a
        version: 7.4.6-7.2
      entries:
      - name: redis-enterprise-operator.v7.4.6-7.2
        version: 7.4.6-7.2
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      name: 7.4.6
    - currentCSV: redis-enterprise-operator.v7.8.2-6.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:36c7a36dabb23a4c003e112085cda2873296bfb22e5d11b470d7c1aba55b7f0f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:c6e1f2109a0dac76b51c25c1cbf6030b4894b21b98147e2ad1db2dd8e7693576
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:acf817c90b5f03bc2796fbc305cf43ce5c0644dcbe1b685752090b3f600a0f7b
        version: 7.8.2-6.8
      entries:
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.2
    - currentCSV: redis-enterprise-operator.v7.8.4-9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:803be04534e59cdc7b8898b3bb336676aad1cb77cf266051e226be82679fcf6a
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:4ba0c414a74268bed9c6757e447b340db7c912afa864cfb096ddf88f11feeaf9
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:66ec124da16cc9d0abd66687e088be0b358e20cfd48f5774d34bf5874e6a4e65
        version: 7.8.4-9.0
      entries:
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.4
    - currentCSV: redis-enterprise-operator.v7.8.6-5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:1fff01daece34aeb7befcb0760d5112c8c7d261f2b86388ece416ac94076a573
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5125e83d3c93b224d1983c599994e43899a0ba0c6f2fc4eed31476f755b5fc0b
        - registry.connect.redhat.com/redislabs/services-manager@sha256:88d483f5e37b5a901d1ae3d965304edc000c172d2e5e307297d5b7bcd77ae73b
        version: 7.8.6-5.1
      entries:
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.6
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: production
    defaultChannel: production
    packageName: redis-enterprise-operator-cert-rhmp
    provider:
      name: Redis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc
      provider-url: ""
    name: bare-metal-event-relay
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bare-metal-event-relay.v4.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "event.redhat-cne.org/v1alpha1",
                "kind": "HardwareEvent",
                "metadata": {
                  "name": "hardwareevent-sample"
                },
                "spec": {
                  "nodeSelector": {
                    "node-role.kubernetes.io/worker": ""
                  },
                  "storageType": "emptyDir",
                  "transportHost": "http://hw-event-publisher-service.openshift-bare-metal-events.svc.cluster.local:9043"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/openshift4/baremetal-hardware-event-proxy-rhel9@sha256:aaaea040164e1fd801e13c74024b2307997cdde954c1a5e35efd90023d2d7376
          createdAt: "2023-02-23"
          description: This software manages the lifecycle of the hw-event-proxy container.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.22.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/hw-event-proxy
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HardwareEvent is the Schema for the hardwareevents API
            displayName: Hardware Event
            kind: HardwareEvent
            name: hardwareevents.event.redhat-cne.org
            version: v1alpha1
        description: |
          ## Funtionality
          Install and manage a hardware event delivery framework.
          ## Installation
          To start using Hardware event proxy, install the operator and then create the following HardwareEvent:
          ```
          apiVersion: "event.redhat-cne.org/v1alpha1"
          kind: "HardwareEvent"
          metadata:
            name: "hardware-event"
          spec:
            nodeSelector:
              node-role.kubernetes.io/worker: ""
            transportHost: "http://hw-event-publisher-service.openshift-bare-metal-events.svc.cluster.local:9043"

          ```
          This operator needs a secret to be created to access Redfish
          Message Registry.
          The secret name and spec the operand expects are under the same namespace
          the operand is deployed on. For example:
          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: redfish-basic-auth
          type: Opaque
          stringData:
            username: <bmc_username>
            password: <bmc_password>
            # BMC host DNS or IP address
            hostaddr: <bmc_host_ip_address>
          ```
        displayName: Bare Metal Event Relay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hardware
        - event
        - cloudevents
        links:
        - name: Hardware Event Proxy Operator
          url: https://github.com/redhat-cne/hw-event-proxy-operator
        maturity: alpha
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:6b7beed3cba9093b3aeb3109f23ec38c20f5568915dff774ad5e9e6bb5db3a58
        - registry.redhat.io/openshift4/bare-metal-event-relay-rhel9-operator@sha256:2219f2f623ee27ef83518d28857995287de0a70cefa193792f28acb352ed8fc0
        - registry.redhat.io/openshift4/baremetal-hardware-event-proxy-rhel9@sha256:aaaea040164e1fd801e13c74024b2307997cdde954c1a5e35efd90023d2d7376
        - registry.redhat.io/openshift4/ose-cloud-event-proxy-rhel9@sha256:c99b448820cf03a33e9237030ac74c1d1ad6a4ce17f1b2840c69b039097c5733
        version: 4.16.1
      entries:
      - name: bare-metal-event-relay.v4.16.1
        version: 4.16.1
      name: "4.16"
    - currentCSV: bare-metal-event-relay.v4.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "event.redhat-cne.org/v1alpha1",
                "kind": "HardwareEvent",
                "metadata": {
                  "name": "hardwareevent-sample"
                },
                "spec": {
                  "nodeSelector": {
                    "node-role.kubernetes.io/worker": ""
                  },
                  "storageType": "emptyDir",
                  "transportHost": "http://hw-event-publisher-service.openshift-bare-metal-events.svc.cluster.local:9043"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/openshift4/baremetal-hardware-event-proxy-rhel9@sha256:aaaea040164e1fd801e13c74024b2307997cdde954c1a5e35efd90023d2d7376
          createdAt: "2023-02-23"
          description: This software manages the lifecycle of the hw-event-proxy container.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.22.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/hw-event-proxy
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HardwareEvent is the Schema for the hardwareevents API
            displayName: Hardware Event
            kind: HardwareEvent
            name: hardwareevents.event.redhat-cne.org
            version: v1alpha1
        description: |
          ## Funtionality
          Install and manage a hardware event delivery framework.
          ## Installation
          To start using Hardware event proxy, install the operator and then create the following HardwareEvent:
          ```
          apiVersion: "event.redhat-cne.org/v1alpha1"
          kind: "HardwareEvent"
          metadata:
            name: "hardware-event"
          spec:
            nodeSelector:
              node-role.kubernetes.io/worker: ""
            transportHost: "http://hw-event-publisher-service.openshift-bare-metal-events.svc.cluster.local:9043"

          ```
          This operator needs a secret to be created to access Redfish
          Message Registry.
          The secret name and spec the operand expects are under the same namespace
          the operand is deployed on. For example:
          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: redfish-basic-auth
          type: Opaque
          stringData:
            username: <bmc_username>
            password: <bmc_password>
            # BMC host DNS or IP address
            hostaddr: <bmc_host_ip_address>
          ```
        displayName: Bare Metal Event Relay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hardware
        - event
        - cloudevents
        links:
        - name: Hardware Event Proxy Operator
          url: https://github.com/redhat-cne/hw-event-proxy-operator
        maturity: alpha
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:6b7beed3cba9093b3aeb3109f23ec38c20f5568915dff774ad5e9e6bb5db3a58
        - registry.redhat.io/openshift4/bare-metal-event-relay-rhel9-operator@sha256:2219f2f623ee27ef83518d28857995287de0a70cefa193792f28acb352ed8fc0
        - registry.redhat.io/openshift4/ose-cloud-event-proxy-rhel9@sha256:c99b448820cf03a33e9237030ac74c1d1ad6a4ce17f1b2840c69b039097c5733
        - registry.redhat.io/openshift4/baremetal-hardware-event-proxy-rhel9@sha256:aaaea040164e1fd801e13c74024b2307997cdde954c1a5e35efd90023d2d7376
        version: 4.16.1
      entries:
      - name: bare-metal-event-relay.v4.16.1
        version: 4.16.1
      name: stable
    defaultChannel: stable
    packageName: bare-metal-event-relay
    provider:
      name: Red Hat Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ocs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocs-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |2-

            [
                {
                    "apiVersion": "ocs.openshift.io/v1",
                    "kind": "StorageCluster",
                    "metadata": {
                        "name": "example-storagecluster",
                        "namespace": "openshift-storage"
                    },
                    "spec": {
                        "manageNodes": false,
                        "monPVCTemplate": {
                            "spec": {
                                "accessModes": [
                                    "ReadWriteOnce"
                                ],
                                "resources": {
                                    "requests": {
                                        "storage": "10Gi"
                                    }
                                },
                                "storageClassName": "gp2-csi"
                            }
                        },
                        "storageDeviceSets": [
                            {
                                "count": 3,
                                "dataPVCTemplate": {
                                    "spec": {
                                        "accessModes": [
                                            "ReadWriteOnce"
                                        ],
                                        "resources": {
                                            "requests": {
                                                "storage": "1Ti"
                                            }
                                        },
                                        "storageClassName": "gp2-csi",
                                        "volumeMode": "Block"
                                    }
                                },
                                "name": "example-deviceset",
                                "placement": {},
                                "portable": true,
                                "resources": {}
                            }
                        ]
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Storage
          containerImage: registry.redhat.io/odf4/ocs-rhel9-operator@sha256:b1a9492bd9b314f86b4e192cf2c83144885ff863d9932e386a41691f50be6193
          description: Red Hat OpenShift Container Storage provides hyperconverged
            storage for applications within an OpenShift cluster.
          external.features.ocs.openshift.io/export-script: |-
            IiIiCkNvcHlyaWdodCAyMDIwIFRoZSBSb29rIEF1dGhvcnMuIEFsbCByaWdodHMgcmVzZXJ2ZWQu
            CgpMaWNlbnNlZCB1bmRlciB0aGUgQXBhY2hlIExpY2Vuc2UsIFZlcnNpb24gMi4wICh0aGUgIkxp
            Y2Vuc2UiKTsKeW91IG1heSBub3QgdXNlIHRoaXMgZmlsZSBleGNlcHQgaW4gY29tcGxpYW5jZSB3
            aXRoIHRoZSBMaWNlbnNlLgpZb3UgbWF5IG9idGFpbiBhIGNvcHkgb2YgdGhlIExpY2Vuc2UgYXQK
            CglodHRwOi8vd3d3LmFwYWNoZS5vcmcvbGljZW5zZXMvTElDRU5TRS0yLjAKClVubGVzcyByZXF1
            aXJlZCBieSBhcHBsaWNhYmxlIGxhdyBvciBhZ3JlZWQgdG8gaW4gd3JpdGluZywgc29mdHdhcmUK
            ZGlzdHJpYnV0ZWQgdW5kZXIgdGhlIExpY2Vuc2UgaXMgZGlzdHJpYnV0ZWQgb24gYW4gIkFTIElT
            IiBCQVNJUywKV0lUSE9VVCBXQVJSQU5USUVTIE9SIENPTkRJVElPTlMgT0YgQU5ZIEtJTkQsIGVp
            dGhlciBleHByZXNzIG9yIGltcGxpZWQuClNlZSB0aGUgTGljZW5zZSBmb3IgdGhlIHNwZWNpZmlj
            IGxhbmd1YWdlIGdvdmVybmluZyBwZXJtaXNzaW9ucyBhbmQKbGltaXRhdGlvbnMgdW5kZXIgdGhl
            IExpY2Vuc2UuCiIiIgoKaW1wb3J0IGVycm5vCmltcG9ydCBzeXMKaW1wb3J0IGpzb24KaW1wb3J0
            IGFyZ3BhcnNlCmltcG9ydCByZQppbXBvcnQgc3VicHJvY2VzcwppbXBvcnQgaG1hYwpmcm9tIGhh
            c2hsaWIgaW1wb3J0IHNoYTEgYXMgc2hhCmZyb20gb3MgaW1wb3J0IGxpbmVzZXAgYXMgTElORVNF
            UApmcm9tIG9zIGltcG9ydCBwYXRoCmZyb20gZW1haWwudXRpbHMgaW1wb3J0IGZvcm1hdGRhdGUK
            aW1wb3J0IHJlcXVlc3RzCmZyb20gcmVxdWVzdHMuYXV0aCBpbXBvcnQgQXV0aEJhc2UKCnB5M2sg
            PSBGYWxzZQppZiBzeXMudmVyc2lvbl9pbmZvLm1ham9yID49IDM6CiAgICBweTNrID0gVHJ1ZQog
            ICAgaW1wb3J0IHVybGxpYi5wYXJzZQogICAgZnJvbSBpcGFkZHJlc3MgaW1wb3J0IGlwX2FkZHJl
            c3MsIElQdjRBZGRyZXNzCgpNb2R1bGVOb3RGb3VuZEVycm9yID0gSW1wb3J0RXJyb3IKCnRyeToK
            ICAgIGltcG9ydCByYWRvcwpleGNlcHQgTW9kdWxlTm90Rm91bmRFcnJvciBhcyBub01vZEVycjoK
            ICAgIHByaW50KGYiRXJyb3I6IHtub01vZEVycn1cbkV4aXRpbmcgdGhlIHNjcmlwdC4uLiIpCiAg
            ICBzeXMuZXhpdCgxKQoKdHJ5OgogICAgaW1wb3J0IHJiZApleGNlcHQgTW9kdWxlTm90Rm91bmRF
            cnJvciBhcyBub01vZEVycjoKICAgIHByaW50KGYiRXJyb3I6IHtub01vZEVycn1cbkV4aXRpbmcg
            dGhlIHNjcmlwdC4uLiIpCiAgICBzeXMuZXhpdCgxKQoKdHJ5OgogICAgIyBmb3IgMi43LngKICAg
            IGZyb20gU3RyaW5nSU8gaW1wb3J0IFN0cmluZ0lPCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9y
            OgogICAgIyBmb3IgMy54CiAgICBmcm9tIGlvIGltcG9ydCBTdHJpbmdJTwoKdHJ5OgogICAgIyBm
            b3IgMi43LngKICAgIGZyb20gdXJscGFyc2UgaW1wb3J0IHVybHBhcnNlCiAgICBmcm9tIHVybGxp
            YiBpbXBvcnQgdXJsZW5jb2RlIGFzIHVybGVuY29kZQpleGNlcHQgTW9kdWxlTm90Rm91bmRFcnJv
            cjoKICAgICMgZm9yIDMueAogICAgZnJvbSB1cmxsaWIucGFyc2UgaW1wb3J0IHVybHBhcnNlCiAg
            ICBmcm9tIHVybGxpYi5wYXJzZSBpbXBvcnQgdXJsZW5jb2RlIGFzIHVybGVuY29kZQoKdHJ5Ogog
            ICAgZnJvbSBiYXNlNjQgaW1wb3J0IGVuY29kZXN0cmluZwpleGNlcHQ6CiAgICBmcm9tIGJhc2U2
            NCBpbXBvcnQgZW5jb2RlYnl0ZXMgYXMgZW5jb2Rlc3RyaW5nCgoKY2xhc3MgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbihFeGNlcHRpb24pOgogICAgcGFzcwoKCiMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMgRHVtbXlSYWRv
            cyAjIyMjIyMjIyMjIyMjIyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjCiMgdGhpcyBpcyBtYWlubHkgZm9yIHRlc3RpbmcgYW5kIGNvdWxkIGJlIHVz
            ZWQgd2hlcmUgJ3JhZG9zJyBpcyBub3QgYXZhaWxhYmxlCgoKY2xhc3MgRHVtbXlSYWRvcyhvYmpl
            Y3QpOgogICAgZGVmIF9faW5pdF9fKHNlbGYpOgogICAgICAgIHNlbGYucmV0dXJuX3ZhbCA9IDAK
            ICAgICAgICBzZWxmLmVycl9tZXNzYWdlID0gIiIKICAgICAgICBzZWxmLnN0YXRlID0gImNvbm5l
            Y3RlZCIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwID0ge30KICAgICAgICBzZWxmLmNtZF9u
            YW1lcyA9IHt9CiAgICAgICAgc2VsZi5faW5pdF9jbWRfb3V0cHV0X21hcCgpCiAgICAgICAgc2Vs
            Zi5kdW1teV9ob3N0X2lwX21hcCA9IHt9CgogICAgZGVmIF9pbml0X2NtZF9vdXRwdXRfbWFwKHNl
            bGYpOgogICAgICAgIGpzb25fZmlsZV9uYW1lID0gInRlc3QtZGF0YS9jZXBoLXN0YXR1cy1vdXQi
            CiAgICAgICAgc2NyaXB0X2RpciA9IHBhdGguYWJzcGF0aChwYXRoLmRpcm5hbWUoX19maWxlX18p
            KQogICAgICAgIGNlcGhfc3RhdHVzX3N0ciA9ICIiCiAgICAgICAgd2l0aCBvcGVuKAogICAgICAg
            ICAgICBwYXRoLmpvaW4oc2NyaXB0X2RpciwganNvbl9maWxlX25hbWUpLCBtb2RlPSJyIiwgZW5j
            b2Rpbmc9IlVURi04IgogICAgICAgICkgYXMganNvbl9maWxlOgogICAgICAgICAgICBjZXBoX3N0
            YXR1c19zdHIgPSBqc29uX2ZpbGUucmVhZCgpCiAgICAgICAgc2VsZi5jbWRfbmFtZXNbImZzIGxz
            Il0gPSAiIiJ7ImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJmcyBscyJ9IiIiCiAgICAgICAg
            c2VsZi5jbWRfbmFtZXNbInF1b3J1bV9zdGF0dXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3Jt
            YXQiOiAianNvbiIsICJwcmVmaXgiOiAicXVvcnVtX3N0YXR1cyJ9IiIiCiAgICAgICAgKQogICAg
            ICAgIHNlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXSA9ICgKICAgICAgICAgICAgIiIieyJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAibWdyIHNlcnZpY2VzIn0iIiIKICAgICAgICApCiAg
            ICAgICAgIyBhbGwgdGhlIGNvbW1hbmRzIGFuZCB0aGVpciBvdXRwdXQKICAgICAgICBzZWxmLmNt
            ZF9vdXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJmcyBscyJdXSA9ICgKICAgICAgICAgICAgIiIi
            W3sibmFtZSI6Im15ZnMiLCJtZXRhZGF0YV9wb29sIjoibXlmcy1tZXRhZGF0YSIsIm1ldGFkYXRh
            X3Bvb2xfaWQiOjIsImRhdGFfcG9vbF9pZHMiOlszXSwiZGF0YV9wb29scyI6WyJteWZzLXJlcGxp
            Y2F0ZWQiXX1dIiIiCiAgICAgICAgKQogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbc2VsZi5j
            bWRfbmFtZXNbInF1b3J1bV9zdGF0dXMiXV0gPSAoCiAgICAgICAgICAgICIiInsiZWxlY3Rpb25f
            ZXBvY2giOjMsInF1b3J1bSI6WzBdLCJxdW9ydW1fbmFtZXMiOlsiYSJdLCJxdW9ydW1fbGVhZGVy
            X25hbWUiOiJhIiwicXVvcnVtX2FnZSI6MTQzODUsImZlYXR1cmVzIjp7InF1b3J1bV9jb24iOiI0
            NTQwMTM4MjkyODM2Njk2MDYzIiwicXVvcnVtX21vbiI6WyJrcmFrZW4iLCJsdW1pbm91cyIsIm1p
            bWljIiwib3NkbWFwLXBydW5lIiwibmF1dGlsdXMiLCJvY3RvcHVzIl19LCJtb25tYXAiOnsiZXBv
            Y2giOjEsImZzaWQiOiJhZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5ZDBmMWMiLCJtb2Rp
            ZmllZCI6IjIwMjAtMDUtMDdUMDM6MzY6MzkuOTE4MDM1WiIsImNyZWF0ZWQiOiIyMDIwLTA1LTA3
            VDAzOjM2OjM5LjkxODAzNVoiLCJtaW5fbW9uX3JlbGVhc2UiOjE1LCJtaW5fbW9uX3JlbGVhc2Vf
            bmFtZSI6Im9jdG9wdXMiLCJmZWF0dXJlcyI6eyJwZXJzaXN0ZW50IjpbImtyYWtlbiIsImx1bWlu
            b3VzIiwibWltaWMiLCJvc2RtYXAtcHJ1bmUiLCJuYXV0aWx1cyIsIm9jdG9wdXMiXSwib3B0aW9u
            YWwiOltdfSwibW9ucyI6W3sicmFuayI6MCwibmFtZSI6ImEiLCJwdWJsaWNfYWRkcnMiOnsiYWRk
            cnZlYyI6W3sidHlwZSI6InYyIiwiYWRkciI6IjEwLjExMC4yMDUuMTc0OjMzMDAiLCJub25jZSI6
            MH0seyJ0eXBlIjoidjEiLCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OSIsIm5vbmNlIjowfV19
            LCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OS8wIiwicHVibGljX2FkZHIiOiIxMC4xMTAuMjA1
            LjE3NDo2Nzg5LzAiLCJwcmlvcml0eSI6MCwid2VpZ2h0IjowfV19fSIiIgogICAgICAgICkKICAg
            ICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXV0g
            PSAoCiAgICAgICAgICAgICIiInsiZGFzaGJvYXJkIjoiaHR0cHM6Ly9jZXBoLWRhc2hib2FyZDo4
            NDQzLyIsInByb21ldGhldXMiOiJodHRwOi8vY2VwaC1kYXNoYm9hcmQtZGI6OTI4My8ifSIiIgog
            ICAgICAgICkKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNh
            cHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMiLCAib3Nk
            IiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0
            YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250
            cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJdLCAiZW50aXR5Ijog
            ImNsaWVudC5oZWFsdGhjaGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRo
            IGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuaGVh
            bHRoY2hlY2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRB
            PT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cyIs
            Im9zZCI6InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5j
            b250cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJ9fV0iIiIKICAg
            ICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIs
            ICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAib3NkIiwgInBy
            b2ZpbGUgcmJkIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtbm9kZSIsICJmb3JtYXQiOiAi
            anNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIi
            W3siZW50aXR5IjoiY2xpZW50LmNzaS1yYmQtbm9kZSIsImtleSI6IkFRQk9nck5lSGJLMUF4QUF1
            YllCZVY4UzFVL0dQenE1U1ZlcTZnPT0iLCJjYXBzIjp7Im1vbiI6InByb2ZpbGUgcmJkLCBhbGxv
            dyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsIm9zZCI6InByb2ZpbGUgcmJkIn19XSIiIgogICAg
            ICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsibW9uIiwg
            InByb2ZpbGUgcmJkLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxs
            b3cgcnciLCAib3NkIiwgInByb2ZpbGUgcmJkIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQt
            cHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNy
            ZWF0ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktcmJkLXByb3Zp
            c2lvbmVyIiwia2V5IjoiQVFCTmdyTmUxZ2V5S3hBQThla1ZpUmRFK2hzczVPd2VZQmt3Tmc9PSIs
            ImNhcHMiOnsibWdyIjoiYWxsb3cgcnciLCJtb24iOiJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFu
            ZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJwcm9maWxlIHJiZCJ9fV0iIiIKICAgICAgICBzZWxm
            LmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyBy
            LCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3Nk
            IiwgImFsbG93IHJ3IHRhZyBjZXBoZnMgKj0qIiwgIm1kcyIsICJhbGxvdyBydyJdLCAiZW50aXR5
            IjogImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4Ijog
            ImF1dGggZ2V0LW9yLWNyZWF0ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVu
            dC5jc2ktY2VwaGZzLW5vZGUiLCJrZXkiOiJBUUJPZ3JOZUVOdW5LeEFBUENtZ0U3UjZHOERjWG5h
            SjFGMzJxZz09IiwiY2FwcyI6eyJtZHMiOiJhbGxvdyBydyIsIm1nciI6ImFsbG93IHJ3IiwibW9u
            IjoiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBy
            dyB0YWcgY2VwaGZzICo9KiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAg
            ICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kICdvc2Qg
            YmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3NkIiwgImFsbG93IHJ3IHRhZyBjZXBo
            ZnMgbWV0YWRhdGE9KiJdLCAiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVy
            IiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgog
            ICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIs
            ImtleSI6IkFRQk9nck5lQUZnY0dCQUF2R3FLT0FEMEQzeHhtVlkwUjkxMmRnPT0iLCJjYXBzIjp7
            Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2Nr
            bGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoifX1dIiIiCiAgICAg
            ICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAi
            YWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdyIiwgImFsbG93IHJ3
            IiwgIm9zZCIsICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiXSwgImVudGl0eSI6ICJj
            bGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVuc2hpZnQtc3RvcmFnZSIsICJmb3JtYXQi
            OiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0g
            IiIiW3siZW50aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0
            b3JhZ2UiLCJrZXkiOiJCUUJPZ3JOZUFGZ2NHQkFBdkdxS09BRDBEM3h4bVZZMFI5MTJkZz09Iiwi
            Y2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29z
            ZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIn19XSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsi
            bW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciIsICJh
            bGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT1teWZzIl0sICJl
            bnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0b3JhZ2Ut
            bXlmcyIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0i
            IiIKICAgICAgICBdID0gIiIiW3siZW50aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9u
            ZXItb3BlbnNoaWZ0LXN0b3JhZ2UtbXlmcyIsImtleSI6IkNRQk9nck5lQUZnY0dCQUF2R3FLT0FE
            MEQzeHhtVlkwUjkxMmRnPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cg
            ciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2Vw
            aGZzIG1ldGFkYXRhPW15ZnMifX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAg
            ICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCBxdW9y
            dW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIiwgImFsbG93IGNvbW1hbmQg
            Y29uZmlnIiwgIm9zZCIsICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFsbG93IHJ3eCBwb29sPWRl
            ZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9ZGVm
            YXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJndy5sb2csIGFsbG93IHgg
            cG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4Il0sICJlbnRpdHkiOiAiY2xpZW50LmhlYWx0
            aGNoZWNrZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0
            ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5oZWFsdGhjaGVja2VyIiwi
            a2V5IjoiQVFERmtiTmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsi
            bW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFu
            ZCB2ZXJzaW9uIiwgIm1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsICJvc2QiOiAicHJvZmls
            ZSByYmQtcmVhZC1vbmx5LCBhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhLCBhbGxvdyBy
            IHBvb2w9LnJndy5yb290LCBhbGxvdyBydyBwb29sPWRlZmF1bHQucmd3LmNvbnRyb2wsIGFsbG93
            IHJ4IHBvb2w9ZGVmYXVsdC5yZ3cubG9nLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0
            cy5pbmRleCJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAi
            IiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMs
            IGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cgY29tbWFuZCBjb25maWciLCAi
            b3NkIiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5j
            b250cm9sLCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1
            bHQucmd3LmJ1Y2tldHMuaW5kZXgiXSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIs
            ICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBjYXBzIn0iIiIKICAgICAgICBdID0g
            IiIiW3siZW50aXR5IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0NWJG
            UkFBVG5kTE5VU1JLcnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIjog
            ImFsbG93IGNvbW1hbmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFs
            bG93IHJ3eCBwb29sPWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFs
            bG93IHJ3IHBvb2w9ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJn
            dy5sb2csIGFsbG93IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4In19XSIiIgogICAg
            ICAgIHNlbGYuY21kX291dHB1dF9tYXBbIiIieyJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            bWdyIHNlcnZpY2VzIn0iIiJdID0gKAogICAgICAgICAgICAiIiJ7ImRhc2hib2FyZCI6ICJodHRw
            Oi8vcm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6NzAwMC8iLCAicHJvbWV0aGV1cyI6
            ICJodHRwOi8vcm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6OTI4My8ifSIiIgogICAg
            ICAgICkKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0
            eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            YXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJ7ImRhc2hib2FyZCI6ICJodHRwOi8vcm9vay1j
            ZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6NzAwMC8iLCAicHJvbWV0aGV1cyI6ICJodHRwOi8v
            cm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6OTI4My8ifSIiIgogICAgICAgIHNlbGYu
            Y21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNsaWVudC5oZWFsdGhj
            aGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldCJ9IiIiCiAgICAg
            ICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5oZWFsdGhjaGVja2VyIiwia2V5IjoiQVFERmti
            TmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjogImFsbG93
            IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFuZCB2ZXJzaW9uIiwg
            Im1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsICJvc2QiOiAicHJvZmlsZSByYmQtcmVhZC1v
            bmx5LCBhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhLCBhbGxvdyByIHBvb2w9LnJndy5y
            b290LCBhbGxvdyBydyBwb29sPWRlZmF1bHQucmd3LmNvbnRyb2wsIGFsbG93IHJ4IHBvb2w9ZGVm
            YXVsdC5yZ3cubG9nLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJ9fV0i
            IiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0eSI6
            ICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJh
            dXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0
            X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtbm9kZSIsICJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJb
            XSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5
            IjogImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRf
            b3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMt
            cHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIK
            ICAgICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAg
            ICAgICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVuc2hp
            ZnQtc3RvcmFnZSIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgog
            ICAgICAgIF0gPSAiIiJbXSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAg
            ICAgICIiInsiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlm
            dC1zdG9yYWdlLW15ZnMiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0i
            IiIKICAgICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAg
            ICAgICAgICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJb
            eyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsImtleSI6IkFRREZrYk5l
            ZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRBPT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIi
            LCAibWdyIjoiYWxsb3cgcnciLCAib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0q
            In19XSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2Fw
            cyI6IFsibW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1n
            ciIsICJhbGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIl0s
            ICJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpz
            b24iLCAicHJlZml4IjogImF1dGggY2FwcyJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmtiTmVmdDViRlJBQVRu
            ZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cgciwgIGFsbG93IGNv
            bW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciI6ImFsbG93IHJ3IiwgIm9zZCI6ImFsbG93IHJ3
            IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFw
            Wyd7ImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJzdGF0dXMifSddID0gY2VwaF9zdGF0dXNf
            c3RyCgogICAgZGVmIHNodXRkb3duKHNlbGYpOgogICAgICAgIHBhc3MKCiAgICBkZWYgZ2V0X2Zz
            aWQoc2VsZik6CiAgICAgICAgcmV0dXJuICJhZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5
            ZDBmMWMiCgogICAgZGVmIGNvbmZfcmVhZF9maWxlKHNlbGYpOgogICAgICAgIHBhc3MKCiAgICBk
            ZWYgY29ubmVjdChzZWxmKToKICAgICAgICBwYXNzCgogICAgZGVmIHBvb2xfZXhpc3RzKHNlbGYs
            IHBvb2xfbmFtZSk6CiAgICAgICAgcmV0dXJuIFRydWUKCiAgICBkZWYgbW9uX2NvbW1hbmQoc2Vs
            ZiwgY21kLCBvdXQpOgogICAgICAgIGpzb25fY21kID0ganNvbi5sb2FkcyhjbWQpCiAgICAgICAg
            anNvbl9jbWRfc3RyID0ganNvbi5kdW1wcyhqc29uX2NtZCwgc29ydF9rZXlzPVRydWUpCiAgICAg
            ICAgY21kX291dHB1dCA9IHNlbGYuY21kX291dHB1dF9tYXBbanNvbl9jbWRfc3RyXQogICAgICAg
            IHJldHVybiBzZWxmLnJldHVybl92YWwsIGNtZF9vdXRwdXQsIHN0cihzZWxmLmVycl9tZXNzYWdl
            LmVuY29kZSgidXRmLTgiKSkKCiAgICBkZWYgX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoc2VsZiwg
            aG9zdF9uYW1lKToKICAgICAgICBpcF9yZWdfeCA9IHJlLmNvbXBpbGUociJcZHsxLDN9LlxkezEs
            M30uXGR7MSwzfS5cZHsxLDN9IikKICAgICAgICAjIGlmIHByb3ZpZGVkIGhvc3QgaXMgZGlyZWN0
            bHkgYW4gSVAgYWRkcmVzcywgcmV0dXJuIHRoZSBzYW1lCiAgICAgICAgaWYgaXBfcmVnX3gubWF0
            Y2goaG9zdF9uYW1lKToKICAgICAgICAgICAgcmV0dXJuIGhvc3RfbmFtZQogICAgICAgIGltcG9y
            dCByYW5kb20KCiAgICAgICAgaG9zdF9pcCA9IHNlbGYuZHVtbXlfaG9zdF9pcF9tYXAuZ2V0KGhv
            c3RfbmFtZSwgIiIpCiAgICAgICAgaWYgbm90IGhvc3RfaXA6CiAgICAgICAgICAgIGhvc3RfaXAg
            PSBmIjE3Mi45LntyYW5kb20ucmFuZGludCgwLCAyNTQpfS57cmFuZG9tLnJhbmRpbnQoMCwgMjU0
            KX0iCiAgICAgICAgICAgIHNlbGYuZHVtbXlfaG9zdF9pcF9tYXBbaG9zdF9uYW1lXSA9IGhvc3Rf
            aXAKICAgICAgICBkZWwgcmFuZG9tCiAgICAgICAgcmV0dXJuIGhvc3RfaXAKCiAgICBAY2xhc3Nt
            ZXRob2QKICAgIGRlZiBSYWRvcyhjb25mZmlsZT1Ob25lKToKICAgICAgICByZXR1cm4gRHVtbXlS
            YWRvcygpCgoKY2xhc3MgUzNBdXRoKEF1dGhCYXNlKToKICAgICIiIkF0dGFjaGVzIEFXUyBBdXRo
            ZW50aWNhdGlvbiB0byB0aGUgZ2l2ZW4gUmVxdWVzdCBvYmplY3QuIiIiCgogICAgc2VydmljZV9i
            YXNlX3VybCA9ICJzMy5hbWF6b25hd3MuY29tIgoKICAgIGRlZiBfX2luaXRfXyhzZWxmLCBhY2Nl
            c3Nfa2V5LCBzZWNyZXRfa2V5LCBzZXJ2aWNlX3VybD1Ob25lKToKICAgICAgICBpZiBzZXJ2aWNl
            X3VybDoKICAgICAgICAgICAgc2VsZi5zZXJ2aWNlX2Jhc2VfdXJsID0gc2VydmljZV91cmwKICAg
            ICAgICBzZWxmLmFjY2Vzc19rZXkgPSBzdHIoYWNjZXNzX2tleSkKICAgICAgICBzZWxmLnNlY3Jl
            dF9rZXkgPSBzdHIoc2VjcmV0X2tleSkKCiAgICBkZWYgX19jYWxsX18oc2VsZiwgcik6CiAgICAg
            ICAgIyBDcmVhdGUgZGF0ZSBoZWFkZXIgaWYgaXQgaXMgbm90IGNyZWF0ZWQgeWV0LgogICAgICAg
            IGlmICJkYXRlIiBub3QgaW4gci5oZWFkZXJzIGFuZCAieC1hbXotZGF0ZSIgbm90IGluIHIuaGVh
            ZGVyczoKICAgICAgICAgICAgci5oZWFkZXJzWyJkYXRlIl0gPSBmb3JtYXRkYXRlKHRpbWV2YWw9
            Tm9uZSwgbG9jYWx0aW1lPUZhbHNlLCB1c2VnbXQ9VHJ1ZSkKICAgICAgICBzaWduYXR1cmUgPSBz
            ZWxmLmdldF9zaWduYXR1cmUocikKICAgICAgICBpZiBweTNrOgogICAgICAgICAgICBzaWduYXR1
            cmUgPSBzaWduYXR1cmUuZGVjb2RlKCJ1dGYtOCIpCiAgICAgICAgci5oZWFkZXJzWyJBdXRob3Jp
            emF0aW9uIl0gPSBmIkFXUyB7c2VsZi5hY2Nlc3Nfa2V5fTp7c2lnbmF0dXJlfSIKICAgICAgICBy
            ZXR1cm4gcgoKICAgIGRlZiBnZXRfc2lnbmF0dXJlKHNlbGYsIHIpOgogICAgICAgIGNhbm9uaWNh
            bF9zdHJpbmcgPSBzZWxmLmdldF9jYW5vbmljYWxfc3RyaW5nKHIudXJsLCByLmhlYWRlcnMsIHIu
            bWV0aG9kKQogICAgICAgIGlmIHB5M2s6CiAgICAgICAgICAgIGtleSA9IHNlbGYuc2VjcmV0X2tl
            eS5lbmNvZGUoInV0Zi04IikKICAgICAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZy5lbmNv
            ZGUoInV0Zi04IikKICAgICAgICBlbHNlOgogICAgICAgICAgICBrZXkgPSBzZWxmLnNlY3JldF9r
            ZXkKICAgICAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZwogICAgICAgIGggPSBobWFjLm5l
            dyhrZXksIG1zZywgZGlnZXN0bW9kPXNoYSkKICAgICAgICByZXR1cm4gZW5jb2Rlc3RyaW5nKGgu
            ZGlnZXN0KCkpLnN0cmlwKCkKCiAgICBkZWYgZ2V0X2Nhbm9uaWNhbF9zdHJpbmcoc2VsZiwgdXJs
            LCBoZWFkZXJzLCBtZXRob2QpOgogICAgICAgIHBhcnNlZHVybCA9IHVybHBhcnNlKHVybCkKICAg
            ICAgICBvYmplY3RrZXkgPSBwYXJzZWR1cmwucGF0aFsxOl0KCiAgICAgICAgYnVja2V0ID0gcGFy
            c2VkdXJsLm5ldGxvY1s6IC1sZW4oc2VsZi5zZXJ2aWNlX2Jhc2VfdXJsKV0KICAgICAgICBpZiBs
            ZW4oYnVja2V0KSA+IDE6CiAgICAgICAgICAgICMgcmVtb3ZlIGxhc3QgZG90CiAgICAgICAgICAg
            IGJ1Y2tldCA9IGJ1Y2tldFs6LTFdCgogICAgICAgIGludGVyZXN0aW5nX2hlYWRlcnMgPSB7ImNv
            bnRlbnQtbWQ1IjogIiIsICJjb250ZW50LXR5cGUiOiAiIiwgImRhdGUiOiAiIn0KICAgICAgICBm
            b3Iga2V5IGluIGhlYWRlcnM6CiAgICAgICAgICAgIGxrID0ga2V5Lmxvd2VyKCkKICAgICAgICAg
            ICAgdHJ5OgogICAgICAgICAgICAgICAgbGsgPSBsay5kZWNvZGUoInV0Zi04IikKICAgICAgICAg
            ICAgZXhjZXB0OgogICAgICAgICAgICAgICAgcGFzcwogICAgICAgICAgICBpZiBoZWFkZXJzW2tl
            eV0gYW5kICgKICAgICAgICAgICAgICAgIGxrIGluIGludGVyZXN0aW5nX2hlYWRlcnMua2V5cygp
            IG9yIGxrLnN0YXJ0c3dpdGgoIngtYW16LSIpCiAgICAgICAgICAgICk6CiAgICAgICAgICAgICAg
            ICBpbnRlcmVzdGluZ19oZWFkZXJzW2xrXSA9IGhlYWRlcnNba2V5XS5zdHJpcCgpCgogICAgICAg
            ICMgSWYgeC1hbXotZGF0ZSBpcyB1c2VkIGl0IHN1cGVyc2VkZXMgdGhlIGRhdGUgaGVhZGVyLgog
            ICAgICAgIGlmIG5vdCBweTNrOgogICAgICAgICAgICBpZiAieC1hbXotZGF0ZSIgaW4gaW50ZXJl
            c3RpbmdfaGVhZGVyczoKICAgICAgICAgICAgICAgIGludGVyZXN0aW5nX2hlYWRlcnNbImRhdGUi
            XSA9ICIiCiAgICAgICAgZWxzZToKICAgICAgICAgICAgaWYgIngtYW16LWRhdGUiIGluIGludGVy
            ZXN0aW5nX2hlYWRlcnM6CiAgICAgICAgICAgICAgICBpbnRlcmVzdGluZ19oZWFkZXJzWyJkYXRl
            Il0gPSAiIgoKICAgICAgICBidWYgPSBmInttZXRob2R9XG4iCiAgICAgICAgZm9yIGtleSBpbiBz
            b3J0ZWQoaW50ZXJlc3RpbmdfaGVhZGVycy5rZXlzKCkpOgogICAgICAgICAgICB2YWwgPSBpbnRl
            cmVzdGluZ19oZWFkZXJzW2tleV0KICAgICAgICAgICAgaWYga2V5LnN0YXJ0c3dpdGgoIngtYW16
            LSIpOgogICAgICAgICAgICAgICAgYnVmICs9IGYie2tleX06e3ZhbH1cbiIKICAgICAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgICAgIGJ1ZiArPSBmInt2YWx9XG4iCgogICAgICAgICMgYXBwZW5k
            IHRoZSBidWNrZXQgaWYgaXQgZXhpc3RzCiAgICAgICAgaWYgYnVja2V0ICE9ICIiOgogICAgICAg
            ICAgICBidWYgKz0gZiIve2J1Y2tldH0iCgogICAgICAgICMgYWRkIHRoZSBvYmplY3RrZXkuIGV2
            ZW4gaWYgaXQgZG9lc24ndCBleGlzdCwgYWRkIHRoZSBzbGFzaAogICAgICAgIGJ1ZiArPSBmIi97
            b2JqZWN0a2V5fSIKCiAgICAgICAgcmV0dXJuIGJ1ZgoKCmNsYXNzIFJhZG9zSlNPTjoKICAgIEVY
            VEVSTkFMX1VTRVJfTkFNRSA9ICJjbGllbnQuaGVhbHRoY2hlY2tlciIKICAgIEVYVEVSTkFMX1JH
            V19BRE1JTl9PUFNfVVNFUl9OQU1FID0gInJndy1hZG1pbi1vcHMtdXNlciIKICAgIEVNUFRZX09V
            VFBVVF9MSVNUID0gIkVtcHR5IG91dHB1dCBsaXN0IgogICAgREVGQVVMVF9SR1dfUE9PTF9QUkVG
            SVggPSAiZGVmYXVsdCIKICAgIERFRkFVTFRfTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUID0gIjky
            ODMiCgogICAgQGNsYXNzbWV0aG9kCiAgICBkZWYgZ2VuX2FyZ19wYXJzZXIoY2xzLCBhcmdzX3Rv
            X3BhcnNlPU5vbmUpOgogICAgICAgIGFyZ1AgPSBhcmdwYXJzZS5Bcmd1bWVudFBhcnNlcigpCgog
            ICAgICAgIGNvbW1vbl9ncm91cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJjb21tb24iKQog
            ICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoIi0tdmVyYm9zZSIsICItdiIsIGFjdGlv
            bj0ic3RvcmVfdHJ1ZSIsIGRlZmF1bHQ9RmFsc2UpCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9h
            cmd1bWVudCgKICAgICAgICAgICAgIi0tY2VwaC1jb25mIiwgIi1jIiwgaGVscD0iUHJvdmlkZSBh
            IGNlcGggY29uZiBmaWxlLiIsIHR5cGU9c3RyCiAgICAgICAgKQogICAgICAgIGNvbW1vbl9ncm91
            cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLWtleXJpbmciLCAiLWsiLCBoZWxwPSJQYXRo
            IHRvIGNlcGgga2V5cmluZyBmaWxlLiIsIHR5cGU9c3RyCiAgICAgICAgKQogICAgICAgIGNvbW1v
            bl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJ1bi1hcy11c2VyIiwKICAgICAg
            ICAgICAgIi11IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgdHlwZT1zdHIs
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIGEgdXNlciBuYW1lIHRvIGNoZWNrIHRoZSBjbHVz
            dGVyJ3MgaGVhbHRoIHN0YXR1cywgbXVzdCBiZSBwcmVmaXhlZCBieSAnY2xpZW50LiciLAogICAg
            ICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1j
            bHVzdGVyLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJL
            dWJlcm5ldGVzIGNsdXN0ZXIgbmFtZShsZWdhY3kgZmxhZyksIE5vdGU6IEVpdGhlciB1c2UgdGhp
            cyBvciAtLWs4cy1jbHVzdGVyLW5hbWUiLAogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAu
            YWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1rOHMtY2x1c3Rlci1uYW1lIiwgZGVmYXVsdD0i
            IiwgaGVscD0iS3ViZXJuZXRlcyBjbHVzdGVyIG5hbWUiCiAgICAgICAgKQogICAgICAgIGNvbW1v
            bl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLW5hbWVzcGFjZSIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIGhlbHA9Ik5hbWVzcGFjZSB3aGVyZSBDZXBoQ2x1
            c3RlciBpcyBydW5uaW5nIiwKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tcmd3LXBvb2wtcHJlZml4IiwgZGVmYXVsdD0iIiwgaGVscD0i
            UkdXIFBvb2wgcHJlZml4IgogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3Vt
            ZW50KAogICAgICAgICAgICAiLS1yZXN0cmljdGVkLWF1dGgtcGVybWlzc2lvbiIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlJlc3RyaWN0IGNlcGhDU0lLZXly
            aW5ncyBhdXRoIHBlcm1pc3Npb25zIHRvIHNwZWNpZmljIHBvb2xzLCBjbHVzdGVyLiIKICAgICAg
            ICAgICAgKyAiTWFuZGF0b3J5IGZsYWdzIHRoYXQgbmVlZCB0byBiZSBzZXQgYXJlIC0tcmJkLWRh
            dGEtcG9vbC1uYW1lLCBhbmQgLS1rOHMtY2x1c3Rlci1uYW1lLiIKICAgICAgICAgICAgKyAiLS1j
            ZXBoZnMtZmlsZXN5c3RlbS1uYW1lIGZsYWcgY2FuIGFsc28gYmUgcGFzc2VkIGluIGNhc2Ugb2Yg
            Y2VwaGZzIHVzZXIgcmVzdHJpY3Rpb24sIHNvIGl0IGNhbiByZXN0cmljdCB1c2VyIHRvIHBhcnRp
            Y3VsYXIgY2VwaGZzIGZpbGVzeXN0ZW0iCiAgICAgICAgICAgICsgInNhbXBsZSBydW46IGBweXRo
            b24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwtY2x1c3Rlci1yZXNvdXJjZXMucHkgLS1jZXBo
            ZnMtZmlsZXN5c3RlbS1uYW1lIG15ZnMgLS1yYmQtZGF0YS1wb29sLW5hbWUgcmVwbGljYXBvb2wg
            LS1rOHMtY2x1c3Rlci1uYW1lIHJvb2tzdG9yYWdlIC0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Np
            b24gdHJ1ZWAiCiAgICAgICAgICAgICsgIk5vdGU6IFJlc3RyaWN0aW5nIHRoZSBjc2ktdXNlcnMg
            cGVyIHBvb2wsIGFuZCBwZXIgY2x1c3RlciB3aWxsIHJlcXVpcmUgY3JlYXRpbmcgbmV3IGNzaS11
            c2VycyBhbmQgbmV3IHNlY3JldHMgZm9yIHRoYXQgY3NpLXVzZXJzLiIKICAgICAgICAgICAgKyAi
            U28gYXBwbHkgdGhlc2Ugc2VjcmV0cyBvbmx5IHRvIG5ldyBgQ29uc3VtZXIgY2x1c3RlcmAgZGVw
            bG95bWVudCB3aGlsZSB1c2luZyB0aGUgc2FtZSBgU291cmNlIGNsdXN0ZXJgLiIsCiAgICAgICAg
            KQogICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXYyLXBv
            cnQtZW5hYmxlIiwKICAgICAgICAgICAgYWN0aW9uPSJzdG9yZV90cnVlIiwKICAgICAgICAgICAg
            ZGVmYXVsdD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iRW5hYmxlIHYyIG1vbiBwb3J0KDMzMDAp
            IGZvciBtb25zIiwKICAgICAgICApCgogICAgICAgIG91dHB1dF9ncm91cCA9IGFyZ1AuYWRkX2Fy
            Z3VtZW50X2dyb3VwKCJvdXRwdXQiKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQo
            CiAgICAgICAgICAgICItLWZvcm1hdCIsCiAgICAgICAgICAgICItdCIsCiAgICAgICAgICAgIGNo
            b2ljZXM9WyJqc29uIiwgImJhc2giXSwKICAgICAgICAgICAgZGVmYXVsdD0ianNvbiIsCiAgICAg
            ICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBvdXRwdXQgZm9ybWF0IChqc29uIHwgYmFzaCkiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1vdXRwdXQiLAogICAgICAgICAgICAiLW8iLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAg
            ICAgICAgICBoZWxwPSJPdXRwdXQgd2lsbCBiZSBzdG9yZWQgaW50byB0aGUgcHJvdmlkZWQgZmls
            ZSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAog
            ICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgQ2VwaCBmaWxlc3lzdGVt
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tY2VwaGZzLW1ldGFkYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIs
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBjZXBoZnMgbWV0YWRh
            dGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAg
            ICAgICAgICAgICItLWNlcGhmcy1kYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9
            IiIsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBjZXBoZnMgZGF0
            YSBwb29sIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAg
            ICAgICAgICAgIi0tcmJkLWRhdGEtcG9vbC1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwK
            ICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRo
            ZSBuYW1lIG9mIHRoZSBSQkQgZGF0YXBvb2wiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3Jv
            dXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1hbGlhcy1yYmQtZGF0YS1wb29sLW5hbWUi
            LAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAg
            ICAgICAgICAgaGVscD0iUHJvdmlkZXMgYW4gYWxpYXMgZm9yIHRoZSAgUkJEIGRhdGEgcG9vbCBu
            YW1lLCBuZWNlc3NhcnkgaWYgYSBzcGVjaWFsIGNoYXJhY3RlciBpcyBwcmVzZW50IGluIHRoZSBw
            b29sIG5hbWUgc3VjaCBhcyBhIHBlcmlvZCBvciB1bmRlcnNjb3JlIiwKICAgICAgICApCiAgICAg
            ICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmd3LWVuZHBvaW50
            IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAg
            ICAgICAgICAgIGhlbHA9IlJBRE9TIEdhdGV3YXkgZW5kcG9pbnQgKGluIGA8SVB2ND46PFBPUlQ+
            YCBvciBgPFtJUHY2XT46PFBPUlQ+YCBvciBgPEZRRE4+OjxQT1JUPmAgZm9ybWF0KSIsCiAgICAg
            ICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJn
            dy10bHMtY2VydC1wYXRoIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVx
            dWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlJBRE9TIEdhdGV3YXkgZW5kcG9pbnQgVExT
            IGNlcnRpZmljYXRlIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVu
            dCgKICAgICAgICAgICAgIi0tcmd3LXNraXAtdGxzIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFs
            c2UsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9Iklnbm9yZSBU
            TFMgY2VydGlmaWNhdGlvbiB2YWxpZGF0aW9uIHdoZW4gYSBzZWxmLXNpZ25lZCBjZXJ0aWZpY2F0
            ZSBpcyBwcm92aWRlZCAoTk9UIFJFQ09NTUVOREVEIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tbW9uaXRvcmluZy1lbmRwb2ludCIs
            CiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAg
            ICAgICAgICBoZWxwPSJDZXBoIE1hbmFnZXIgcHJvbWV0aGV1cyBleHBvcnRlciBlbmRwb2ludHMg
            KGNvbW1hIHNlcGFyYXRlZCBsaXN0IG9mIChmb3JtYXQgYDxJUHY0PmAgb3IgYDxbSVB2Nl0+YCBv
            ciBgPEZRRE4+YCkgZW50cmllcyBvZiBhY3RpdmUgYW5kIHN0YW5kYnkgbWdycykiLAogICAgICAg
            ICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1tb25p
            dG9yaW5nLWVuZHBvaW50LXBvcnQiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAg
            ICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iQ2VwaCBNYW5hZ2VyIHByb21ldGhl
            dXMgZXhwb3J0ZXIgcG9ydCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJn
            dW1lbnQoCiAgICAgICAgICAgICItLXNraXAtbW9uaXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAg
            ICAgICAgIGhlbHA9IkRvIG5vdCBjaGVjayBmb3IgYSBtb25pdG9yaW5nIGVuZHBvaW50IGZvciB0
            aGUgQ2VwaCBjbHVzdGVyIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tcmJkLW1ldGFkYXRhLWVjLXBvb2wtbmFtZSIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBo
            ZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiBlcmFzdXJlIGNvZGVkIFJCRCBtZXRhZGF0YSBwb29s
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tZHJ5LXJ1biIsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFj
            dGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGhlbHA9IkRyeSBydW4gcHJpbnRzIHRoZSBl
            eGVjdXRlZCBjb21tYW5kcyB3aXRob3V0IHJ1bm5pbmcgdGhlbSIsCiAgICAgICAgKQogICAgICAg
            IG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJhZG9zLW5hbWVzcGFj
            ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAog
            ICAgICAgICAgICBoZWxwPSJEaXZpZGVzIGEgcG9vbCBpbnRvIHNlcGFyYXRlIGxvZ2ljYWwgbmFt
            ZXNwYWNlcywgdXNlZCBmb3IgY3JlYXRpbmcgUkJEIFBWQyBpbiBhIENlcGhCbG9ja1Bvb2xSYWRv
            c05hbWVzcGFjZSAoc2hvdWxkIGJlIGxvd2VyIGNhc2UpIiwKICAgICAgICApCiAgICAgICAgb3V0
            cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tc3Vidm9sdW1lLWdyb3VwIiwK
            ICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAg
            ICAgICAgIGhlbHA9InByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBzdWJ2b2x1bWUgZ3JvdXAiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1yZ3ctcmVhbG0tbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJl
            cXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUg
            cmd3LXJlYWxtIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgK
            ICAgICAgICAgICAgIi0tcmd3LXpvbmUtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAg
            ICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUg
            bmFtZSBvZiB0aGUgcmd3LXpvbmUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS1yZ3ctem9uZWdyb3VwLW5hbWUiLAogICAgICAgICAg
            ICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVs
            cD0icHJvdmlkZXMgdGhlIG5hbWUgb2YgdGhlIHJndy16b25lZ3JvdXAiLAogICAgICAgICkKCiAg
            ICAgICAgdXBncmFkZV9ncm91cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJ1cGdyYWRlIikK
            ICAgICAgICB1cGdyYWRlX2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tdXBncmFk
            ZSIsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9
            RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlVwZ3JhZGVzIHRoZSBjZXBoQ1NJS2V5cmluZ3MoRm9y
            IGV4YW1wbGU6IGNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyKSBhbmQgY2xpZW50LmhlYWx0
            aGNoZWNrZXIgY2VwaCB1c2VycyB3aXRoIG5ldyBwZXJtaXNzaW9ucyBuZWVkZWQgZm9yIHRoZSBu
            ZXcgY2x1c3RlciB2ZXJzaW9uIGFuZCBvbGRlciBwZXJtaXNzaW9uIHdpbGwgc3RpbGwgYmUgYXBw
            bGllZC4iCiAgICAgICAgICAgICsgIlNhbXBsZSBydW46IGBweXRob24zIC9ldGMvY2VwaC9jcmVh
            dGUtZXh0ZXJuYWwtY2x1c3Rlci1yZXNvdXJjZXMucHkgLS11cGdyYWRlYCwgdGhpcyB3aWxsIHVw
            Z3JhZGUgYWxsIHRoZSBkZWZhdWx0IGNzaSB1c2Vycyhub24tcmVzdHJpY3RlZCkiCiAgICAgICAg
            ICAgICsgIkZvciByZXN0cmljdGVkIHVzZXJzKEZvciBleGFtcGxlOiBjbGllbnQuY3NpLWNlcGhm
            cy1wcm92aXNpb25lci1vcGVuc2hpZnQtc3RvcmFnZS1teWZzKSwgdXNlcnMgY3JlYXRlZCB1c2lu
            ZyAtLXJlc3RyaWN0ZWQtYXV0aC1wZXJtaXNzaW9uIGZsYWcgbmVlZCB0byBwYXNzIG1hbmRhdG9y
            eSBmbGFncyIKICAgICAgICAgICAgKyAibWFuZGF0b3J5IGZsYWdzOiAnLS1yYmQtZGF0YS1wb29s
            LW5hbWUsIC0tazhzLWNsdXN0ZXItbmFtZSBhbmQgLS1ydW4tYXMtdXNlcicgZmxhZ3Mgd2hpbGUg
            dXBncmFkaW5nIgogICAgICAgICAgICArICJpbiBjYXNlIG9mIGNlcGhmcyB1c2VycyBpZiB5b3Ug
            aGF2ZSBwYXNzZWQgLS1jZXBoZnMtZmlsZXN5c3RlbS1uYW1lIGZsYWcgd2hpbGUgY3JlYXRpbmcg
            dXNlciB0aGVuIHdoaWxlIHVwZ3JhZGluZyBpdCB3aWxsIGJlIG1hbmRhdG9yeSB0b28iCiAgICAg
            ICAgICAgICsgIlNhbXBsZSBydW46IGBweXRob24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwt
            Y2x1c3Rlci1yZXNvdXJjZXMucHkgLS11cGdyYWRlIC0tcmJkLWRhdGEtcG9vbC1uYW1lIHJlcGxp
            Y2Fwb29sIC0tazhzLWNsdXN0ZXItbmFtZSByb29rc3RvcmFnZSAgLS1ydW4tYXMtdXNlciBjbGll
            bnQuY3NpLXJiZC1ub2RlLXJvb2tzdG9yYWdlLXJlcGxpY2Fwb29sYCIKICAgICAgICAgICAgKyAi
            UFM6IEFuIGV4aXN0aW5nIG5vbi1yZXN0cmljdGVkIHVzZXIgY2Fubm90IGJlIGNvbnZlcnRlZCB0
            byBhIHJlc3RyaWN0ZWQgdXNlciBieSB1cGdyYWRpbmcuIgogICAgICAgICAgICArICJVcGdyYWRl
            IGZsYWcgc2hvdWxkIG9ubHkgYmUgdXNlZCB0byBhcHBlbmQgbmV3IHBlcm1pc3Npb25zIHRvIHVz
            ZXJzLCBpdCBzaG91bGRuJ3QgYmUgdXNlZCBmb3IgY2hhbmdpbmcgdXNlciBhbHJlYWR5IGFwcGxp
            ZWQgcGVybWlzc2lvbiwgZm9yIGV4YW1wbGUgeW91IHNob3VsZG4ndCBjaGFuZ2UgaW4gd2hpY2gg
            cG9vbCB1c2VyIGhhcyBhY2Nlc3MiLAogICAgICAgICkKCiAgICAgICAgaWYgYXJnc190b19wYXJz
            ZToKICAgICAgICAgICAgYXNzZXJ0ICgKICAgICAgICAgICAgICAgIHR5cGUoYXJnc190b19wYXJz
            ZSkgPT0gbGlzdAogICAgICAgICAgICApLCAiQXJndW1lbnQgdG8gJ2dlbl9hcmdfcGFyc2VyJyBz
            aG91bGQgYmUgYSBsaXN0IgogICAgICAgIGVsc2U6CiAgICAgICAgICAgIGFyZ3NfdG9fcGFyc2Ug
            PSBzeXMuYXJndlsxOl0KICAgICAgICByZXR1cm4gYXJnUC5wYXJzZV9hcmdzKGFyZ3NfdG9fcGFy
            c2UpCgogICAgZGVmIHZhbGlkYXRlX3JiZF9tZXRhZGF0YV9lY19wb29sX25hbWUoc2VsZik6CiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lOgogICAg
            ICAgICAgICByYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5yYmRf
            bWV0YWRhdGFfZWNfcG9vbF9uYW1lCiAgICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9h
            cmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQoKICAgICAgICAgICAgaWYgcmJkX3Bvb2xfbmFt
            ZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24o
            CiAgICAgICAgICAgICAgICAgICAgIkZsYWcgJy0tcmJkLWRhdGEtcG9vbC1uYW1lJyBzaG91bGQg
            bm90IGJlIGVtcHR5IgogICAgICAgICAgICAgICAgKQoKICAgICAgICAgICAgaWYgcmJkX21ldGFk
            YXRhX2VjX3Bvb2xfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIkZsYWcgJy0tcmJkLW1ldGFkYXRh
            LWVjLXBvb2wtbmFtZScgc2hvdWxkIG5vdCBiZSBlbXB0eSIKICAgICAgICAgICAgICAgICkKCiAg
            ICAgICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAib3NkIGR1bXAiLCAiZm9ybWF0IjogImpz
            b24ifQogICAgICAgICAgICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1v
            bl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBs
            ZW4oanNvbl9vdXQpID09IDA6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJl
            RXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYie2NtZF9qc29uWydwcmVmaXgnXX0gY29t
            bWFuZCBmYWlsZWQuXG4iCiAgICAgICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYg
            cmV0X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAgICAgICAg
            ICApCiAgICAgICAgICAgIG1ldGFkYXRhX3Bvb2xfZXhpc3QsIHBvb2xfZXhpc3QgPSBGYWxzZSwg
            RmFsc2UKCiAgICAgICAgICAgIGZvciBrZXkgaW4ganNvbl9vdXRbInBvb2xzIl06CiAgICAgICAg
            ICAgICAgICAjIGlmIGVyYXN1cmVfY29kZV9wcm9maWxlIGlzIGVtcHR5IGFuZCBwb29sIG5hbWUg
            ZXhpc3RzIHRoZW4gaXQgcmVwbGljYSBwb29sCiAgICAgICAgICAgICAgICBpZiAoCiAgICAgICAg
            ICAgICAgICAgICAga2V5WyJlcmFzdXJlX2NvZGVfcHJvZmlsZSJdID09ICIiCiAgICAgICAgICAg
            ICAgICAgICAgYW5kIGtleVsicG9vbF9uYW1lIl0gPT0gcmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFt
            ZQogICAgICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgICAgICBtZXRhZGF0YV9wb29sX2V4
            aXN0ID0gVHJ1ZQogICAgICAgICAgICAgICAgIyBpZiBlcmFzdXJlX2NvZGVfcHJvZmlsZSBpcyBu
            b3QgZW1wdHkgYW5kIHBvb2wgbmFtZSBleGlzdHMgdGhlbiBpdCBpcyBlYyBwb29sCiAgICAgICAg
            ICAgICAgICBpZiBrZXlbImVyYXN1cmVfY29kZV9wcm9maWxlIl0gYW5kIGtleVsicG9vbF9uYW1l
            Il0gPT0gcmJkX3Bvb2xfbmFtZToKICAgICAgICAgICAgICAgICAgICBwb29sX2V4aXN0ID0gVHJ1
            ZQoKICAgICAgICAgICAgaWYgbm90IG1ldGFkYXRhX3Bvb2xfZXhpc3Q6CiAgICAgICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJQ
            cm92aWRlZCByYmRfZWNfbWV0YWRhdGFfcG9vbCBuYW1lLCIKICAgICAgICAgICAgICAgICAgICBm
            IiB7cmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZX0sIGRvZXMgbm90IGV4aXN0IgogICAgICAgICAg
            ICAgICAgKQogICAgICAgICAgICBpZiBub3QgcG9vbF9leGlzdDoKICAgICAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgZiJQcm92
            aWRlZCByYmRfZGF0YV9wb29sIG5hbWUsIHtyYmRfcG9vbF9uYW1lfSwgZG9lcyBub3QgZXhpc3Qi
            CiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIHJldHVybiByYmRfbWV0YWRhdGFfZWNfcG9v
            bF9uYW1lCgogICAgZGVmIGRyeV9ydW4oc2VsZiwgbXNnKToKICAgICAgICBpZiBzZWxmLl9hcmdf
            cGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHByaW50KCJFeGVjdXRlOiAiICsgIiciICsgbXNn
            ICsgIiciKQoKICAgIGRlZiB2YWxpZGF0ZV9yZ3dfZW5kcG9pbnRfdGxzX2NlcnQoc2VsZik6CiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZ3dfdGxzX2NlcnRfcGF0aDoKICAgICAgICAgICAg
            d2l0aCBvcGVuKHNlbGYuX2FyZ19wYXJzZXIucmd3X3Rsc19jZXJ0X3BhdGgsIGVuY29kaW5nPSJ1
            dGY4IikgYXMgZjoKICAgICAgICAgICAgICAgIGNvbnRlbnRzID0gZi5yZWFkKCkKICAgICAgICAg
            ICAgICAgIHJldHVybiBjb250ZW50cy5yc3RyaXAoKQoKICAgIGRlZiBfY2hlY2tfY29uZmxpY3Rp
            bmdfb3B0aW9ucyhzZWxmKToKICAgICAgICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNlci51cGdyYWRl
            IGFuZCBub3Qgc2VsZi5fYXJnX3BhcnNlci5yYmRfZGF0YV9wb29sX25hbWU6CiAgICAgICAgICAg
            IHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAiRWl0aGVy
            ICctLXVwZ3JhZGUnIG9yICctLXJiZC1kYXRhLXBvb2wtbmFtZSA8cG9vbF9uYW1lPicgc2hvdWxk
            IGJlIHNwZWNpZmllZCIKICAgICAgICAgICAgKQoKICAgIGRlZiBfaW52YWxpZF9lbmRwb2ludChz
            ZWxmLCBlbmRwb2ludF9zdHIpOgogICAgICAgICMgZXh0cmFjdCB0aGUgcG9ydCBieSBnZXR0aW5n
            IHRoZSBsYXN0IHNwbGl0IG9uIGA6YCBkZWxpbWl0ZXIKICAgICAgICB0cnk6CiAgICAgICAgICAg
            IGVuZHBvaW50X3N0cl9pcCwgcG9ydCA9IGVuZHBvaW50X3N0ci5yc3BsaXQoIjoiLCAxKQogICAg
            ICAgIGV4Y2VwdCBWYWx1ZUVycm9yOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJl
            RXhjZXB0aW9uKGYiTm90IGEgcHJvcGVyIGVuZHBvaW50OiB7ZW5kcG9pbnRfc3RyfSIpCgogICAg
            ICAgIHRyeToKICAgICAgICAgICAgaWYgZW5kcG9pbnRfc3RyX2lwWzBdID09ICJbIjoKICAgICAg
            ICAgICAgICAgIGVuZHBvaW50X3N0cl9pcCA9IGVuZHBvaW50X3N0cl9pcFsxIDogbGVuKGVuZHBv
            aW50X3N0cl9pcCkgLSAxXQogICAgICAgICAgICBpcF90eXBlID0gKAogICAgICAgICAgICAgICAg
            IklQdjQiIGlmIHR5cGUoaXBfYWRkcmVzcyhlbmRwb2ludF9zdHJfaXApKSBpcyBJUHY0QWRkcmVz
            cyBlbHNlICJJUHY2IgogICAgICAgICAgICApCiAgICAgICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAg
            ICAgICAgICAgIGlwX3R5cGUgPSAiRlFETiIKICAgICAgICBpZiBub3QgcG9ydC5pc2RpZ2l0KCk6
            CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oZiJQb3J0IG5vdCB2
            YWxpZDoge3BvcnR9IikKICAgICAgICBpbnRQb3J0ID0gaW50KHBvcnQpCiAgICAgICAgaWYgaW50
            UG9ydCA8IDEgb3IgaW50UG9ydCA+IDIqKjE2IC0gMToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0
            aW9uRmFpbHVyZUV4Y2VwdGlvbihmIk91dCBvZiByYW5nZSBwb3J0IG51bWJlcjoge3BvcnR9IikK
            CiAgICAgICAgcmV0dXJuIGlwX3R5cGUKCiAgICBkZWYgZW5kcG9pbnRfZGlhbChzZWxmLCBlbmRw
            b2ludF9zdHIsIGlwX3R5cGUsIHRpbWVvdXQ9MywgY2VydD1Ob25lKToKICAgICAgICAjIGlmIHRo
            ZSAnY2x1c3RlcicgaW5zdGFuY2UgaXMgYSBkdW1teSBvbmUsCiAgICAgICAgIyBkb24ndCB0cnkg
            dG8gcmVhY2ggb3V0IHRvIHRoZSBlbmRwb2ludAogICAgICAgIGlmIGlzaW5zdGFuY2Uoc2VsZi5j
            bHVzdGVyLCBEdW1teVJhZG9zKToKICAgICAgICAgICAgcmV0dXJuICIiLCAiIiwgIiIKICAgICAg
            ICBpZiBpcF90eXBlID09ICJJUHY2IjoKICAgICAgICAgICAgdHJ5OgogICAgICAgICAgICAgICAg
            ZW5kcG9pbnRfc3RyX2lwLCBlbmRwb2ludF9zdHJfcG9ydCA9IGVuZHBvaW50X3N0ci5yc3BsaXQo
            IjoiLCAxKQogICAgICAgICAgICBleGNlcHQgVmFsdWVFcnJvcjoKICAgICAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgZiJOb3Qg
            YSBwcm9wZXIgZW5kcG9pbnQ6IHtlbmRwb2ludF9zdHJ9IgogICAgICAgICAgICAgICAgKQogICAg
            ICAgICAgICBpZiBlbmRwb2ludF9zdHJfaXBbMF0gIT0gIlsiOgogICAgICAgICAgICAgICAgZW5k
            cG9pbnRfc3RyX2lwID0gIlsiICsgZW5kcG9pbnRfc3RyX2lwICsgIl0iCiAgICAgICAgICAgIGVu
            ZHBvaW50X3N0ciA9ICI6Ii5qb2luKFtlbmRwb2ludF9zdHJfaXAsIGVuZHBvaW50X3N0cl9wb3J0
            XSkKCiAgICAgICAgcHJvdG9jb2xzID0gWyJodHRwIiwgImh0dHBzIl0KICAgICAgICByZXNwb25z
            ZV9lcnJvciA9IE5vbmUKICAgICAgICBmb3IgcHJlZml4IGluIHByb3RvY29sczoKICAgICAgICAg
            ICAgdHJ5OgogICAgICAgICAgICAgICAgZXAgPSBmIntwcmVmaXh9Oi8ve2VuZHBvaW50X3N0cn0i
            CiAgICAgICAgICAgICAgICB2ZXJpZnkgPSBOb25lCiAgICAgICAgICAgICAgICAjIElmIHZlcmlm
            eSBpcyBzZXQgdG8gYSBwYXRoIHRvIGEgZGlyZWN0b3J5LAogICAgICAgICAgICAgICAgIyB0aGUg
            ZGlyZWN0b3J5IG11c3QgaGF2ZSBiZWVuIHByb2Nlc3NlZCB1c2luZyB0aGUgY19yZWhhc2ggdXRp
            bGl0eSBzdXBwbGllZCB3aXRoIE9wZW5TU0wuCiAgICAgICAgICAgICAgICBpZiBwcmVmaXggPT0g
            Imh0dHBzIiBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfc2tpcF90bHM6CiAgICAgICAgICAgICAg
            ICAgICAgdmVyaWZ5ID0gRmFsc2UKICAgICAgICAgICAgICAgICAgICByID0gcmVxdWVzdHMuaGVh
            ZChlcCwgdGltZW91dD10aW1lb3V0LCB2ZXJpZnk9RmFsc2UpCiAgICAgICAgICAgICAgICBlbGlm
            IHByZWZpeCA9PSAiaHR0cHMiIGFuZCBjZXJ0OgogICAgICAgICAgICAgICAgICAgIHZlcmlmeSA9
            IGNlcnQKICAgICAgICAgICAgICAgICAgICByID0gcmVxdWVzdHMuaGVhZChlcCwgdGltZW91dD10
            aW1lb3V0LCB2ZXJpZnk9Y2VydCkKICAgICAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAg
            ICAgICAgciA9IHJlcXVlc3RzLmhlYWQoZXAsIHRpbWVvdXQ9dGltZW91dCkKICAgICAgICAgICAg
            ICAgIGlmIHIuc3RhdHVzX2NvZGUgPT0gMjAwOgogICAgICAgICAgICAgICAgICAgIHJldHVybiBw
            cmVmaXgsIHZlcmlmeSwgIiIKICAgICAgICAgICAgZXhjZXB0IEV4Y2VwdGlvbiBhcyBlcnI6CiAg
            ICAgICAgICAgICAgICByZXNwb25zZV9lcnJvciA9IGVycgogICAgICAgICAgICAgICAgY29udGlu
            dWUKICAgICAgICBzeXMuc3RkZXJyLndyaXRlKAogICAgICAgICAgICBmInVuYWJsZSB0byBjb25u
            ZWN0IHRvIGVuZHBvaW50OiB7ZW5kcG9pbnRfc3RyfSwgZmFpbGVkIGVycm9yOiB7cmVzcG9uc2Vf
            ZXJyb3J9IgogICAgICAgICkKICAgICAgICByZXR1cm4gKAogICAgICAgICAgICAiIiwKICAgICAg
            ICAgICAgIiIsCiAgICAgICAgICAgICgiLTEiKSwKICAgICAgICApCgogICAgZGVmIF9faW5pdF9f
            KHNlbGYsIGFyZ19saXN0PU5vbmUpOgogICAgICAgIHNlbGYub3V0X21hcCA9IHt9CiAgICAgICAg
            c2VsZi5fZXhjbHVkZWRfa2V5cyA9IHNldCgpCiAgICAgICAgc2VsZi5fYXJnX3BhcnNlciA9IHNl
            bGYuZ2VuX2FyZ19wYXJzZXIoYXJnc190b19wYXJzZT1hcmdfbGlzdCkKICAgICAgICBzZWxmLl9j
            aGVja19jb25mbGljdGluZ19vcHRpb25zKCkKICAgICAgICBzZWxmLnJ1bl9hc191c2VyID0gc2Vs
            Zi5fYXJnX3BhcnNlci5ydW5fYXNfdXNlcgogICAgICAgIHNlbGYub3V0cHV0X2ZpbGUgPSBzZWxm
            Ll9hcmdfcGFyc2VyLm91dHB1dAogICAgICAgIHNlbGYuY2VwaF9jb25mID0gc2VsZi5fYXJnX3Bh
            cnNlci5jZXBoX2NvbmYKICAgICAgICBzZWxmLmNlcGhfa2V5cmluZyA9IHNlbGYuX2FyZ19wYXJz
            ZXIua2V5cmluZwogICAgICAgICMgaWYgdXNlciBub3QgcHJvdmlkZWQsIGdpdmUgYSBkZWZhdWx0
            IHVzZXIKICAgICAgICBpZiBub3Qgc2VsZi5ydW5fYXNfdXNlciBhbmQgbm90IHNlbGYuX2FyZ19w
            YXJzZXIudXBncmFkZToKICAgICAgICAgICAgc2VsZi5ydW5fYXNfdXNlciA9IHNlbGYuRVhURVJO
            QUxfVVNFUl9OQU1FCiAgICAgICAgaWYgbm90IHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJl
            Zml4IGFuZCBub3Qgc2VsZi5fYXJnX3BhcnNlci51cGdyYWRlOgogICAgICAgICAgICBzZWxmLl9h
            cmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeCA9IHNlbGYuREVGQVVMVF9SR1dfUE9PTF9QUkVGSVgK
            ICAgICAgICBpZiBzZWxmLmNlcGhfY29uZjoKICAgICAgICAgICAga3dhcmdzID0ge30KICAgICAg
            ICAgICAgaWYgc2VsZi5jZXBoX2tleXJpbmc6CiAgICAgICAgICAgICAgICBrd2FyZ3NbImNvbmYi
            XSA9IHsia2V5cmluZyI6IHNlbGYuY2VwaF9rZXlyaW5nfQogICAgICAgICAgICBzZWxmLmNsdXN0
            ZXIgPSByYWRvcy5SYWRvcyhjb25mZmlsZT1zZWxmLmNlcGhfY29uZiwgKiprd2FyZ3MpCiAgICAg
            ICAgZWxzZToKICAgICAgICAgICAgc2VsZi5jbHVzdGVyID0gcmFkb3MuUmFkb3MoKQogICAgICAg
            ICAgICBzZWxmLmNsdXN0ZXIuY29uZl9yZWFkX2ZpbGUoKQogICAgICAgIHNlbGYuY2x1c3Rlci5j
            b25uZWN0KCkKCiAgICBkZWYgc2h1dGRvd24oc2VsZik6CiAgICAgICAgaWYgc2VsZi5jbHVzdGVy
            LnN0YXRlID09ICJjb25uZWN0ZWQiOgogICAgICAgICAgICBzZWxmLmNsdXN0ZXIuc2h1dGRvd24o
            KQoKICAgIGRlZiBnZXRfZnNpZChzZWxmKToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRy
            eV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGggZnNpZCIpCiAgICAg
            ICAgcmV0dXJuIHN0cihzZWxmLmNsdXN0ZXIuZ2V0X2ZzaWQoKSkKCiAgICBkZWYgX2NvbW1vbl9j
            bWRfanNvbl9nZW4oc2VsZiwgY21kX2pzb24pOgogICAgICAgIGNtZCA9IGpzb24uZHVtcHMoY21k
            X2pzb24sIHNvcnRfa2V5cz1UcnVlKQogICAgICAgIHJldF92YWwsIGNtZF9vdXQsIGVycl9tc2cg
            PSBzZWxmLmNsdXN0ZXIubW9uX2NvbW1hbmQoY21kLCBiIiIpCiAgICAgICAgaWYgc2VsZi5fYXJn
            X3BhcnNlci52ZXJib3NlOgogICAgICAgICAgICBwcmludChmIkNvbW1hbmQgSW5wdXQ6IHtjbWR9
            IikKICAgICAgICAgICAgcHJpbnQoCiAgICAgICAgICAgICAgICBmIlJldHVybiBWYWw6IHtyZXRf
            dmFsfVxuQ29tbWFuZCBPdXRwdXQ6IHtjbWRfb3V0fVxuIgogICAgICAgICAgICAgICAgZiJFcnJv
            ciBNZXNzYWdlOiB7ZXJyX21zZ31cbi0tLS0tLS0tLS1cbiIKICAgICAgICAgICAgKQogICAgICAg
            IGpzb25fb3V0ID0ge30KICAgICAgICAjIGlmIHRoZXJlIGlzIG5vIGVycm9yIChpLmU7IHJldF92
            YWwgaXMgWkVSTykgYW5kICdjbWRfb3V0JyBpcyBub3QgZW1wdHkKICAgICAgICAjIHRoZW4gY29u
            dmVydCAnY21kX291dCcgdG8gYSBqc29uIG91dHB1dAogICAgICAgIGlmIHJldF92YWwgPT0gMCBh
            bmQgY21kX291dDoKICAgICAgICAgICAganNvbl9vdXQgPSBqc29uLmxvYWRzKGNtZF9vdXQpCiAg
            ICAgICAgcmV0dXJuIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnCgogICAgZGVmIGdldF9jZXBo
            X2V4dGVybmFsX21vbl9kYXRhKHNlbGYpOgogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAi
            cXVvcnVtX3N0YXR1cyIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgaWYgc2VsZi5fYXJnX3Bh
            cnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKCJjZXBoICIgKyBj
            bWRfanNvblsicHJlZml4Il0pCiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBz
            ZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICMgaWYgdGhlcmUgaXMg
            YW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihq
            c29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgICAgICIncXVvcnVtX3N0YXR1cycgY29tbWFuZCBmYWlsZWQuXG4iCiAg
            ICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxm
            LkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgKQogICAgICAgIHFfbGVhZGVyX25hbWUg
            PSBqc29uX291dFsicXVvcnVtX2xlYWRlcl9uYW1lIl0KICAgICAgICBxX2xlYWRlcl9kZXRhaWxz
            ID0ge30KICAgICAgICBxX2xlYWRlcl9tYXRjaGluZ19saXN0ID0gWwogICAgICAgICAgICBsIGZv
            ciBsIGluIGpzb25fb3V0WyJtb25tYXAiXVsibW9ucyJdIGlmIGxbIm5hbWUiXSA9PSBxX2xlYWRl
            cl9uYW1lCiAgICAgICAgXQogICAgICAgIGlmIGxlbihxX2xlYWRlcl9tYXRjaGluZ19saXN0KSA9
            PSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKCJObyBtYXRj
            aGluZyAnbW9uJyBkZXRhaWxzIGZvdW5kIikKICAgICAgICBxX2xlYWRlcl9kZXRhaWxzID0gcV9s
            ZWFkZXJfbWF0Y2hpbmdfbGlzdFswXQogICAgICAgICMgZ2V0IHRoZSBhZGRyZXNzIHZlY3RvciBv
            ZiB0aGUgcXVvcnVtLWxlYWRlcgogICAgICAgIHFfbGVhZGVyX2FkZHJ2ZWMgPSBxX2xlYWRlcl9k
            ZXRhaWxzLmdldCgicHVibGljX2FkZHJzIiwge30pLmdldCgiYWRkcnZlYyIsIFtdKQogICAgICAg
            ICMgaWYgdGhlIHF1b3J1bS1sZWFkZXIgaGFzIG9ubHkgb25lIGFkZHJlc3MgaW4gdGhlIGFkZHJl
            c3MtdmVjdG9yCiAgICAgICAgIyBhbmQgaXQgaXMgb2YgdHlwZSAndjInIChpZTsgd2l0aCA8SVA+
            OjMzMDApLAogICAgICAgICMgcmFpc2UgYW4gZXhjZXB0aW9uIHRvIG1ha2UgdXNlciBhd2FyZSB0
            aGF0CiAgICAgICAgIyB0aGV5IGhhdmUgdG8gZW5hYmxlICd2MScgKGllOyB3aXRoIDxJUD46Njc4
            OSkgdHlwZSBhcyB3ZWxsCiAgICAgICAgaWYgbGVuKHFfbGVhZGVyX2FkZHJ2ZWMpID09IDEgYW5k
            IHFfbGVhZGVyX2FkZHJ2ZWNbMF1bInR5cGUiXSA9PSAidjIiOgogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgIk9ubHkgJ3YyJyBhZGRy
            ZXNzIHR5cGUgaXMgZW5hYmxlZCwgdXNlciBzaG91bGQgYWxzbyBlbmFibGUgJ3YxJyB0eXBlIGFz
            IHdlbGwiCiAgICAgICAgICAgICkKICAgICAgICBpcF9hZGRyID0gc3RyKHFfbGVhZGVyX2RldGFp
            bHNbInB1YmxpY19hZGRyIl0uc3BsaXQoIi8iKVswXSkKCiAgICAgICAgaWYgc2VsZi5fYXJnX3Bh
            cnNlci52Ml9wb3J0X2VuYWJsZToKICAgICAgICAgICAgaWYgbGVuKHFfbGVhZGVyX2FkZHJ2ZWMp
            ID4gMToKICAgICAgICAgICAgICAgIGlmIHFfbGVhZGVyX2FkZHJ2ZWNbMF1bInR5cGUiXSA9PSAi
            djIiOgogICAgICAgICAgICAgICAgICAgIGlwX2FkZHIgPSBxX2xlYWRlcl9hZGRydmVjWzBdWyJh
            ZGRyIl0KICAgICAgICAgICAgICAgIGVsaWYgcV9sZWFkZXJfYWRkcnZlY1sxXVsidHlwZSJdID09
            ICJ2MiI6CiAgICAgICAgICAgICAgICAgICAgaXBfYWRkciA9IHFfbGVhZGVyX2FkZHJ2ZWNbMV1b
            ImFkZHIiXQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0
            ZSgKICAgICAgICAgICAgICAgICAgICAiJ3YyJyBhZGRyZXNzIHR5cGUgbm90IHByZXNlbnQsIGFu
            ZCAndjItcG9ydC1lbmFibGUnIGZsYWcgaXMgcHJvdmlkZWQiCiAgICAgICAgICAgICAgICApCgog
            ICAgICAgIHJldHVybiBmIntzdHIocV9sZWFkZXJfbmFtZSl9PXtpcF9hZGRyfSIKCiAgICBkZWYg
            X2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoc2VsZiwgaG9zdF9uYW1lLCBwb3J0LCBpcF90eXBlKToK
            ICAgICAgICAjIGlmICdjbHVzdGVyJyBpbnN0YW5jZSBpcyBhIGR1bW15IHR5cGUsCiAgICAgICAg
            IyBjYWxsIHRoZSBkdW1teSBpbnN0YW5jZSdzICJjb252ZXJ0IiBtZXRob2QKICAgICAgICBpZiBu
            b3QgaG9zdF9uYW1lOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KCJFbXB0eSBob3N0bmFtZSBwcm92aWRlZCIpCiAgICAgICAgaWYgaXNpbnN0YW5jZShzZWxmLmNs
            dXN0ZXIsIER1bW15UmFkb3MpOgogICAgICAgICAgICByZXR1cm4gc2VsZi5jbHVzdGVyLl9jb252
            ZXJ0X2hvc3RuYW1lX3RvX2lwKGhvc3RfbmFtZSkKCiAgICAgICAgaWYgaXBfdHlwZSA9PSAiRlFE
            TiI6CiAgICAgICAgICAgICMgY2hlY2sgd2hpY2ggaXAgRlFETiBzaG91bGQgYmUgY29udmVydGVk
            IHRvLCBJUHY0IG9yIElQdjYKICAgICAgICAgICAgIyBjaGVjayB0aGUgaG9zdCBpcCwgdGhlIGVu
            ZHBvaW50IGlwIHR5cGUgd291bGQgYmUgc2ltaWxhciB0byBob3N0IGlwCiAgICAgICAgICAgIGNt
            ZF9qc29uID0geyJwcmVmaXgiOiAib3JjaCBob3N0IGxzIiwgImZvcm1hdCI6ICJqc29uIn0KICAg
            ICAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pz
            b25fZ2VuKGNtZF9qc29uKQogICAgICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1
            bCBhdHRlbXB0LAogICAgICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9
            PSAwOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAg
            ICAgICAgICAgICAgICAgICAiJ29yY2ggaG9zdCBscycgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAg
            ICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYgcmV0X3ZhbCAhPSAwIGVsc2Ugc2Vs
            Zi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGhvc3Rf
            YWRkciA9IGpzb25fb3V0WzBdWyJhZGRyIl0KICAgICAgICAgICAgIyBhZGQgOjgwIHNhbXBsZSBw
            b3J0IGluIGlwX3R5cGUsIGFzIF9pbnZhbGlkX2VuZHBvaW50IGFsc28gdmVyaWZ5IHBvcnQKICAg
            ICAgICAgICAgaG9zdF9pcF90eXBlID0gc2VsZi5faW52YWxpZF9lbmRwb2ludChob3N0X2FkZHIg
            KyAiOjgwIikKICAgICAgICAgICAgaW1wb3J0IHNvY2tldAoKICAgICAgICAgICAgIyBleGFtcGxl
            IG91dHB1dCBbKDxBZGRyZXNzRmFtaWx5LkFGX0lORVQ6IDI+LCA8U29ja2V0S2luZC5TT0NLX1NU
            UkVBTTogMT4sIDYsICcnLCAoJzkzLjE4NC4yMTYuMzQnLCA4MCkpLCAuLi5dCiAgICAgICAgICAg
            ICMgd2UgbmVlZCB0byBnZXQgOTMuMTg0LjIxNi4zNCBzbyBpdCB3b3VsZCBiZSBpcFswXVs0XVsw
            XQogICAgICAgICAgICBpZiBob3N0X2lwX3R5cGUgPT0gIklQdjYiOgogICAgICAgICAgICAgICAg
            aXAgPSBzb2NrZXQuZ2V0YWRkcmluZm8oCiAgICAgICAgICAgICAgICAgICAgaG9zdF9uYW1lLCBw
            b3J0LCBmYW1pbHk9c29ja2V0LkFGX0lORVQ2LCBwcm90bz1zb2NrZXQuSVBQUk9UT19UQ1AKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxpZiBob3N0X2lwX3R5cGUgPT0gIklQdjQiOgog
            ICAgICAgICAgICAgICAgaXAgPSBzb2NrZXQuZ2V0YWRkcmluZm8oCiAgICAgICAgICAgICAgICAg
            ICAgaG9zdF9uYW1lLCBwb3J0LCBmYW1pbHk9c29ja2V0LkFGX0lORVQsIHByb3RvPXNvY2tldC5J
            UFBST1RPX1RDUAogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBkZWwgc29ja2V0CiAgICAg
            ICAgICAgIHJldHVybiBpcFswXVs0XVswXQogICAgICAgIHJldHVybiBob3N0X25hbWUKCiAgICBk
            ZWYgZ2V0X2FjdGl2ZV9hbmRfc3RhbmRieV9tZ3JzKHNlbGYpOgogICAgICAgIGlmIHNlbGYuX2Fy
            Z19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuICIiLCBzZWxmLmRyeV9ydW4oImNl
            cGggc3RhdHVzIikKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQgPSBzZWxmLl9hcmdf
            cGFyc2VyLm1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydAogICAgICAgIG1vbml0b3JpbmdfZW5kcG9p
            bnRfaXBfbGlzdCA9IHNlbGYuX2FyZ19wYXJzZXIubW9uaXRvcmluZ19lbmRwb2ludAogICAgICAg
            IHN0YW5kYnlfbWdycyA9IFtdCiAgICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBf
            bGlzdDoKICAgICAgICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJzdGF0dXMiLCAiZm9ybWF0
            IjogImpzb24ifQogICAgICAgICAgICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYu
            X2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgICAgICMgaWYgdGhlcmUgaXMg
            YW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBs
            ZW4oanNvbl9vdXQpID09IDA6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJl
            RXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICInbWdyIHNlcnZpY2VzJyBjb21tYW5kIGZh
            aWxlZC5cbiIKICAgICAgICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFs
            ICE9IDAgZWxzZSBzZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgICAgICkKICAg
            ICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludCA9ICgKICAgICAgICAgICAgICAgIGpzb25fb3V0
            LmdldCgibWdybWFwIiwge30pLmdldCgic2VydmljZXMiLCB7fSkuZ2V0KCJwcm9tZXRoZXVzIiwg
            IiIpCiAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnQ6
            CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAg
            ICAgICAgICAgICAgICJjYW4ndCBmaW5kIG1vbml0b3JpbmdfZW5kcG9pbnQsIHByb21ldGhldXMg
            bW9kdWxlIG1pZ2h0IG5vdCBiZSBlbmFibGVkLCAiCiAgICAgICAgICAgICAgICAgICAgImVuYWJs
            ZSB0aGUgbW9kdWxlIGJ5IHJ1bm5pbmcgJ2NlcGggbWdyIG1vZHVsZSBlbmFibGUgcHJvbWV0aGV1
            cyciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICMgbm93IGNoZWNrIHRoZSBzdGFuZC1i
            eSBtZ3ItcwogICAgICAgICAgICBzdGFuZGJ5X2FyciA9IGpzb25fb3V0LmdldCgibWdybWFwIiwg
            e30pLmdldCgic3RhbmRieXMiLCBbXSkKICAgICAgICAgICAgZm9yIGVhY2hfc3RhbmRieSBpbiBz
            dGFuZGJ5X2FycjoKICAgICAgICAgICAgICAgIGlmICJuYW1lIiBpbiBlYWNoX3N0YW5kYnkua2V5
            cygpOgogICAgICAgICAgICAgICAgICAgIHN0YW5kYnlfbWdycy5hcHBlbmQoZWFjaF9zdGFuZGJ5
            WyJuYW1lIl0pCiAgICAgICAgICAgIHRyeToKICAgICAgICAgICAgICAgIHBhcnNlZF9lbmRwb2lu
            dCA9IHVybHBhcnNlKG1vbml0b3JpbmdfZW5kcG9pbnQpCiAgICAgICAgICAgIGV4Y2VwdCBWYWx1
            ZUVycm9yOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigK
            ICAgICAgICAgICAgICAgICAgICBmImludmFsaWQgZW5kcG9pbnQ6IHttb25pdG9yaW5nX2VuZHBv
            aW50fSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9p
            cF9saXN0ID0gcGFyc2VkX2VuZHBvaW50Lmhvc3RuYW1lCiAgICAgICAgICAgIGlmIG5vdCBtb25p
            dG9yaW5nX2VuZHBvaW50X3BvcnQ6CiAgICAgICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50
            X3BvcnQgPSBzdHIocGFyc2VkX2VuZHBvaW50LnBvcnQpCgogICAgICAgICMgaWYgbW9uaXRvcmlu
            ZyBlbmRwb2ludCBwb3J0IGlzIG5vdCBzZXQsIHB1dCBhIGRlZmF1bHQgbW9uIHBvcnQKICAgICAg
            ICBpZiBub3QgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0OgogICAgICAgICAgICBtb25pdG9yaW5n
            X2VuZHBvaW50X3BvcnQgPSBzZWxmLkRFRkFVTFRfTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUCgog
            ICAgICAgICMgdXNlciBjb3VsZCBnaXZlIGNvbW1hIGFuZCBzcGFjZSBzZXBhcmF0ZWQgaW5wdXRz
            IChsaWtlIC0tbW9uaXRvcmluZy1lbmRwb2ludD0iPGlwMT4sIDxpcDI+IikKICAgICAgICBtb25p
            dG9yaW5nX2VuZHBvaW50X2lwX2xpc3QgPSBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3QucmVw
            bGFjZSgiLCIsICIgIikKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXQg
            PSBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Quc3BsaXQoKQogICAgICAgICMgaWYgbW9uaXRv
            cmluZy1lbmRwb2ludCBjb3VsZCBub3QgYmUgZm91bmQsIHJhaXNlIGFuIGVycm9yCiAgICAgICAg
            aWYgbGVuKG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdF9zcGxpdCkgPT0gMDoKICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigiTm8gJ21vbml0b3JpbmctZW5kcG9p
            bnQnIGZvdW5kIikKICAgICAgICAjIGZpcnN0IGlwIGlzIHRyZWF0ZWQgYXMgdGhlIG1haW4gbW9u
            aXRvcmluZy1lbmRwb2ludAogICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfaXAgPSBtb25pdG9y
            aW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXRbMF0KICAgICAgICAjIHJlc3Qgb2YgdGhlIGlwLXMg
            YXJlIGFkZGVkIHRvIHRoZSAnc3RhbmRieV9tZ3JzJyBsaXN0CiAgICAgICAgc3RhbmRieV9tZ3Jz
            LmV4dGVuZChtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXRbMTpdKQogICAgICAgIGZh
            aWxlZF9pcCA9IG1vbml0b3JpbmdfZW5kcG9pbnRfaXAKCiAgICAgICAgbW9uaXRvcmluZ19lbmRw
            b2ludCA9ICI6Ii5qb2luKAogICAgICAgICAgICBbbW9uaXRvcmluZ19lbmRwb2ludF9pcCwgbW9u
            aXRvcmluZ19lbmRwb2ludF9wb3J0XQogICAgICAgICkKICAgICAgICBpcF90eXBlID0gc2VsZi5f
            aW52YWxpZF9lbmRwb2ludChtb25pdG9yaW5nX2VuZHBvaW50KQogICAgICAgIHRyeToKICAgICAg
            ICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcCA9IHNlbGYuX2NvbnZlcnRfaG9zdG5hbWVfdG9f
            aXAoCiAgICAgICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X2lwLCBtb25pdG9yaW5nX2Vu
            ZHBvaW50X3BvcnQsIGlwX3R5cGUKICAgICAgICAgICAgKQogICAgICAgICAgICAjIGNvbGxlY3Qg
            YWxsIHRoZSAnc3RhbmQtYnknIG1nciBpcHMKICAgICAgICAgICAgbWdyX2lwcyA9IFtdCiAgICAg
            ICAgICAgIGZvciBlYWNoX3N0YW5kYnlfbWdyIGluIHN0YW5kYnlfbWdyczoKICAgICAgICAgICAg
            ICAgIGZhaWxlZF9pcCA9IGVhY2hfc3RhbmRieV9tZ3IKICAgICAgICAgICAgICAgIG1ncl9pcHMu
            YXBwZW5kKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAo
            CiAgICAgICAgICAgICAgICAgICAgICAgIGVhY2hfc3RhbmRieV9tZ3IsIG1vbml0b3JpbmdfZW5k
            cG9pbnRfcG9ydCwgaXBfdHlwZQogICAgICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAg
            ICkKICAgICAgICBleGNlcHQ6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNl
            cHRpb24oCiAgICAgICAgICAgICAgICBmIkNvbnZlcnNpb24gb2YgaG9zdDoge2ZhaWxlZF9pcH0g
            dG8gSVAgZmFpbGVkLiAiCiAgICAgICAgICAgICAgICAiUGxlYXNlIGVudGVyIHRoZSBJUCBhZGRy
            ZXNzZXMgb2YgYWxsIHRoZSBjZXBoLW1ncnMgd2l0aCB0aGUgJy0tbW9uaXRvcmluZy1lbmRwb2lu
            dCcgZmxhZyIKICAgICAgICAgICAgKQoKICAgICAgICBfLCBfLCBlcnIgPSBzZWxmLmVuZHBvaW50
            X2RpYWwobW9uaXRvcmluZ19lbmRwb2ludCwgaXBfdHlwZSkKICAgICAgICBpZiBlcnIgPT0gIi0x
            IjoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihlcnIpCiAgICAg
            ICAgIyBhZGQgdGhlIHZhbGlkYXRlZCBhY3RpdmUgbWdyIElQIGludG8gdGhlIGZpcnN0IGluZGV4
            CiAgICAgICAgbWdyX2lwcy5pbnNlcnQoMCwgbW9uaXRvcmluZ19lbmRwb2ludF9pcCkKICAgICAg
            ICBhbGxfbWdyX2lwc19zdHIgPSAiLCIuam9pbihtZ3JfaXBzKQogICAgICAgIHJldHVybiBhbGxf
            bWdyX2lwc19zdHIsIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydAoKICAgIGRlZiBjaGVja191c2Vy
            X2V4aXN0KHNlbGYsIHVzZXIpOgogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAiYXV0aCBn
            ZXQiLCAiZW50aXR5IjogZiJ7dXNlcn0iLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIHJldF92
            YWwsIGpzb25fb3V0LCBfID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAg
            ICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICBy
            ZXR1cm4gIiIKICAgICAgICByZXR1cm4gc3RyKGpzb25fb3V0WzBdWyJrZXkiXSkKCiAgICBkZWYg
            Z2V0X2NlcGhmc19wcm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50
            aXR5ID0gImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIgogICAgICAgIGNhcHMgPSB7CiAg
            ICAgICAgICAgICJtb24iOiAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCci
            LAogICAgICAgICAgICAibWdyIjogImFsbG93IHJ3IiwKICAgICAgICAgICAgIm9zZCI6ICJhbGxv
            dyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiLAogICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9h
            cmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uOgogICAgICAgICAgICBrOHNfY2x1
            c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1lCiAgICAgICAgICAg
            IGlmIGs4c19jbHVzdGVyX25hbWUgPT0gIiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRp
            b25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJrOHNfY2x1c3Rlcl9uYW1l
            IG5vdCBmb3VuZCwgcGxlYXNlIHNldCB0aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW0gPSBzZWxmLl9hcmdf
            cGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICAgICAgaWYgY2VwaGZzX2ZpbGVz
            eXN0ZW0gPT0gIiI6CiAgICAgICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1
            c3Rlcl9uYW1lfSIKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYi
            e2VudGl0eX0te2s4c19jbHVzdGVyX25hbWV9LXtjZXBoZnNfZmlsZXN5c3RlbX0iCiAgICAgICAg
            ICAgICAgICBjYXBzWyJvc2QiXSA9IGYiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT17Y2Vw
            aGZzX2ZpbGVzeXN0ZW19IgoKICAgICAgICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVmIGdl
            dF9jZXBoZnNfbm9kZV9jYXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gImNs
            aWVudC5jc2ktY2VwaGZzLW5vZGUiCiAgICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6
            ICJhbGxvdyByLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJt
            Z3IiOiAiYWxsb3cgcnciLAogICAgICAgICAgICAib3NkIjogImFsbG93IHJ3IHRhZyBjZXBoZnMg
            Kj0qIiwKICAgICAgICAgICAgIm1kcyI6ICJhbGxvdyBydyIsCiAgICAgICAgfQogICAgICAgIGlm
            IHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAg
            IGs4c19jbHVzdGVyX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAg
            ICAgICAgICAgaWYgazhzX2NsdXN0ZXJfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIms4c19jbHVz
            dGVyX25hbWUgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1rOHMtY2x1c3Rlci1uYW1lJyBm
            bGFnIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBjZXBoZnNfZmlsZXN5c3RlbSA9IHNl
            bGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgICAgICBpZiBjZXBo
            ZnNfZmlsZXN5c3RlbSA9PSAiIjoKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0t
            e2s4c19jbHVzdGVyX25hbWV9IgogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgZW50
            aXR5ID0gZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0te2NlcGhmc19maWxlc3lzdGVtfSIK
            ICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJdID0gZiJhbGxvdyBydyB0YWcgY2VwaGZzICo9e2Nl
            cGhmc19maWxlc3lzdGVtfSIKCiAgICAgICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBn
            ZXRfZW50aXR5KAogICAgICAgIHNlbGYsCiAgICAgICAgZW50aXR5LAogICAgICAgIHJiZF9wb29s
            X25hbWUsCiAgICAgICAgYWxpYXNfcmJkX3Bvb2xfbmFtZSwKICAgICAgICBrOHNfY2x1c3Rlcl9u
            YW1lLAogICAgICAgIHJhZG9zX25hbWVzcGFjZSwKICAgICk6CiAgICAgICAgaWYgKAogICAgICAg
            ICAgICByYmRfcG9vbF9uYW1lLmNvdW50KCIuIikgIT0gMAogICAgICAgICAgICBvciByYmRfcG9v
            bF9uYW1lLmNvdW50KCJfIikgIT0gMAogICAgICAgICAgICBvciBhbGlhc19yYmRfcG9vbF9uYW1l
            ICE9ICIiCiAgICAgICAgICAgICMgY2hlY2tpbmcgYWxpYXNfcmJkX3Bvb2xfbmFtZSBpcyBub3Qg
            ZW1wdHkgYXMgdGhlcmUgbWF5YmUgYSBzcGVjaWFsIGNoYXJhY3RlciB1c2VkIG90aGVyIHRoYW4g
            LiBvciBfCiAgICAgICAgKToKICAgICAgICAgICAgaWYgYWxpYXNfcmJkX3Bvb2xfbmFtZSA9PSAi
            IjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAg
            ICAgICAgICAgICAgICAgInBsZWFzZSBzZXQgdGhlICctLWFsaWFzLXJiZC1kYXRhLXBvb2wtbmFt
            ZScgZmxhZyBhcyB0aGUgcmJkIGRhdGEgcG9vbCBuYW1lIGNvbnRhaW5zICcuJyBvciAnXyciCiAg
            ICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmICgKICAgICAgICAgICAgICAgIGFsaWFzX3Ji
            ZF9wb29sX25hbWUuY291bnQoIi4iKSAhPSAwCiAgICAgICAgICAgICAgICBvciBhbGlhc19yYmRf
            cG9vbF9uYW1lLmNvdW50KCJfIikgIT0gMAogICAgICAgICAgICApOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJy0t
            YWxpYXMtcmJkLWRhdGEtcG9vbC1uYW1lJyBmbGFnIHZhbHVlIHNob3VsZCBub3QgY29udGFpbiAn
            Licgb3IgJ18nIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRp
            dHl9LXtrOHNfY2x1c3Rlcl9uYW1lfS17YWxpYXNfcmJkX3Bvb2xfbmFtZX0iCiAgICAgICAgZWxz
            ZToKICAgICAgICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0te3Ji
            ZF9wb29sX25hbWV9IgoKICAgICAgICBpZiByYWRvc19uYW1lc3BhY2U6CiAgICAgICAgICAgIGVu
            dGl0eSA9IGYie2VudGl0eX0te3JhZG9zX25hbWVzcGFjZX0iCiAgICAgICAgcmV0dXJuIGVudGl0
            eQoKICAgIGRlZiBnZXRfcmJkX3Byb3Zpc2lvbmVyX2NhcHNfYW5kX2VudGl0eShzZWxmKToKICAg
            ICAgICBlbnRpdHkgPSAiY2xpZW50LmNzaS1yYmQtcHJvdmlzaW9uZXIiCiAgICAgICAgY2FwcyA9
            IHsKICAgICAgICAgICAgIm1vbiI6ICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJs
            b2NrbGlzdCciLAogICAgICAgICAgICAibWdyIjogImFsbG93IHJ3IiwKICAgICAgICAgICAgIm9z
            ZCI6ICJwcm9maWxlIHJiZCIsCiAgICAgICAgfQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIu
            cmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBz
            ZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgICAgICBhbGlhc19yYmRf
            cG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5hbGlhc19yYmRfZGF0YV9wb29sX25hbWUKICAg
            ICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJf
            bmFtZQogICAgICAgICAgICByYWRvc19uYW1lc3BhY2UgPSBzZWxmLl9hcmdfcGFyc2VyLnJhZG9z
            X25hbWVzcGFjZQogICAgICAgICAgICBpZiByYmRfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAg
            ICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAg
            ICAibWFuZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1yYmQtZGF0YS1w
            b29sLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmIGs4c19jbHVz
            dGVyX25hbWUgPT0gIiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhj
            ZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBs
            ZWFzZSBzZXQgdGhlICctLWs4cy1jbHVzdGVyLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgIGVudGl0eSA9IHNlbGYuZ2V0X2VudGl0eSgKICAgICAgICAgICAgICAgIGVu
            dGl0eSwKICAgICAgICAgICAgICAgIHJiZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBhbGlh
            c19yYmRfcG9vbF9uYW1lLAogICAgICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAg
            ICAgICAgICAgIHJhZG9zX25hbWVzcGFjZSwKICAgICAgICAgICAgKQogICAgICAgICAgICBpZiBy
            YWRvc19uYW1lc3BhY2UgIT0gIiI6CiAgICAgICAgICAgICAgICBjYXBzWyJvc2QiXSA9ICgKICAg
            ICAgICAgICAgICAgICAgICBmInByb2ZpbGUgcmJkIHBvb2w9e3JiZF9wb29sX25hbWV9IG5hbWVz
            cGFjZT17cmFkb3NfbmFtZXNwYWNlfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxz
            ZToKICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJdID0gZiJwcm9maWxlIHJiZCBwb29sPXtyYmRf
            cG9vbF9uYW1lfSIKCiAgICAgICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfcmJk
            X25vZGVfY2Fwc19hbmRfZW50aXR5KHNlbGYpOgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3Np
            LXJiZC1ub2RlIgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAgICJtb24iOiAicHJvZmlsZSBy
            YmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwKICAgICAgICAgICAgIm9zZCI6ICJw
            cm9maWxlIHJiZCIsCiAgICAgICAgfQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJp
            Y3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9h
            cmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9u
            YW1lID0gc2VsZi5fYXJnX3BhcnNlci5hbGlhc19yYmRfZGF0YV9wb29sX25hbWUKICAgICAgICAg
            ICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQog
            ICAgICAgICAgICByYWRvc19uYW1lc3BhY2UgPSBzZWxmLl9hcmdfcGFyc2VyLnJhZG9zX25hbWVz
            cGFjZQogICAgICAgICAgICBpZiByYmRfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAibWFu
            ZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1yYmQtZGF0YS1wb29sLW5h
            bWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmIGs4c19jbHVzdGVyX25h
            bWUgPT0gIiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KAogICAgICAgICAgICAgICAgICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBz
            ZXQgdGhlICctLWs4cy1jbHVzdGVyLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAg
            ICAgICAgIGVudGl0eSA9IHNlbGYuZ2V0X2VudGl0eSgKICAgICAgICAgICAgICAgIGVudGl0eSwK
            ICAgICAgICAgICAgICAgIHJiZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBhbGlhc19yYmRf
            cG9vbF9uYW1lLAogICAgICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAgICAgICAg
            ICAgIHJhZG9zX25hbWVzcGFjZSwKICAgICAgICAgICAgKQogICAgICAgICAgICBpZiByYWRvc19u
            YW1lc3BhY2UgIT0gIiI6CiAgICAgICAgICAgICAgICBjYXBzWyJvc2QiXSA9ICgKICAgICAgICAg
            ICAgICAgICAgICBmInByb2ZpbGUgcmJkIHBvb2w9e3JiZF9wb29sX25hbWV9IG5hbWVzcGFjZT17
            cmFkb3NfbmFtZXNwYWNlfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxzZToKICAg
            ICAgICAgICAgICAgIGNhcHNbIm9zZCJdID0gZiJwcm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9u
            YW1lfSIKCiAgICAgICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfZGVmYXVsdFVz
            ZXJfY2Fwc19hbmRfZW50aXR5KHNlbGYpOgogICAgICAgIGVudGl0eSA9IHNlbGYucnVuX2FzX3Vz
            ZXIKICAgICAgICBjYXBzID0gewogICAgICAgICAgICAibW9uIjogImFsbG93IHIsIGFsbG93IGNv
            bW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFuZCB2ZXJzaW9uIiwKICAgICAgICAgICAg
            Im1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsCiAgICAgICAgICAgICJvc2QiOiBmInByb2Zp
            bGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9e3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bv
            b2xfcHJlZml4fS5yZ3cubWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9v
            bD17c2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5jb250cm9sLCBhbGxvdyBy
            eCBwb29sPXtzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmxvZywgYWxsb3cg
            eCBwb29sPXtzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmJ1Y2tldHMuaW5k
            ZXgiLAogICAgICAgIH0KCiAgICAgICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRf
            Y2Fwc19hbmRfZW50aXR5KHNlbGYsIHVzZXJfbmFtZSk6CiAgICAgICAgaWYgImNsaWVudC5jc2kt
            Y2VwaGZzLXByb3Zpc2lvbmVyIiBpbiB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQu
            Y3NpLWNlcGhmcy1wcm92aXNpb25lciIgIT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAg
            ICAgcmV0dXJuIHNlbGYuZ2V0X2NlcGhmc19wcm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoKQog
            ICAgICAgIGlmICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiBpbiB1c2VyX25hbWU6CiAgICAgICAg
            ICAgIGlmICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiAhPSB1c2VyX25hbWU6CiAgICAgICAgICAg
            ICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQog
            ICAgICAgICAgICByZXR1cm4gc2VsZi5nZXRfY2VwaGZzX25vZGVfY2Fwc19hbmRfZW50aXR5KCkK
            ICAgICAgICBpZiAiY2xpZW50LmNzaS1yYmQtcHJvdmlzaW9uZXIiIGluIHVzZXJfbmFtZToKICAg
            ICAgICAgICAgaWYgImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiAhPSB1c2VyX25hbWU6CiAg
            ICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9u
            ID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5nZXRfcmJkX3Byb3Zpc2lvbmVyX2NhcHNf
            YW5kX2VudGl0eSgpCiAgICAgICAgaWYgImNsaWVudC5jc2ktcmJkLW5vZGUiIGluIHVzZXJfbmFt
            ZToKICAgICAgICAgICAgaWYgImNsaWVudC5jc2ktcmJkLW5vZGUiICE9IHVzZXJfbmFtZToKICAg
            ICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24g
            PSBUcnVlCiAgICAgICAgICAgIHJldHVybiBzZWxmLmdldF9yYmRfbm9kZV9jYXBzX2FuZF9lbnRp
            dHkoKQogICAgICAgIGlmICJjbGllbnQuaGVhbHRoY2hlY2tlciIgaW4gdXNlcl9uYW1lOgogICAg
            ICAgICAgICBpZiAiY2xpZW50LmhlYWx0aGNoZWNrZXIiICE9IHVzZXJfbmFtZToKICAgICAgICAg
            ICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVl
            CiAgICAgICAgICAgIHJldHVybiBzZWxmLmdldF9kZWZhdWx0VXNlcl9jYXBzX2FuZF9lbnRpdHko
            KQoKICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICBm
            Im5vIHVzZXIgZm91bmQgd2l0aCB1c2VyX25hbWU6IHt1c2VyX25hbWV9LCAiCiAgICAgICAgICAg
            ICJnZXRfY2Fwc19hbmRfZW50aXR5IGNvbW1hbmQgZmFpbGVkLlxuIgogICAgICAgICkKCiAgICBk
            ZWYgY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3VzZXIoc2VsZiwgdXNlcik6CiAgICAgICAgIiIiCiAg
            ICAgICAgY29tbWFuZDogY2VwaCBhdXRoIGdldC1vci1jcmVhdGUgY2xpZW50LmNzaS1jZXBoZnMt
            cHJvdmlzaW9uZXIgbW9uICdhbGxvdyByJyBtZ3IgJ2FsbG93IHJ3JyBvc2QgJ2FsbG93IHJ3IHRh
            ZyBjZXBoZnMgbWV0YWRhdGE9KicKICAgICAgICAiIiIKICAgICAgICBjYXBzLCBlbnRpdHkgPSBz
            ZWxmLmdldF9jYXBzX2FuZF9lbnRpdHkodXNlcikKICAgICAgICBjbWRfanNvbiA9IHsKICAgICAg
            ICAgICAgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUiLAogICAgICAgICAgICAiZW50aXR5
            IjogZW50aXR5LAogICAgICAgICAgICAiY2FwcyI6IFtjYXAgZm9yIGNhcF9saXN0IGluIGxpc3Qo
            Y2Fwcy5pdGVtcygpKSBmb3IgY2FwIGluIGNhcF9saXN0XSwKICAgICAgICAgICAgImZvcm1hdCI6
            ICJqc29uIiwKICAgICAgICB9CgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoK
            ICAgICAgICAgICAgcmV0dXJuICgKICAgICAgICAgICAgICAgIHNlbGYuZHJ5X3J1bigKICAgICAg
            ICAgICAgICAgICAgICAiY2VwaCAiCiAgICAgICAgICAgICAgICAgICAgKyBjbWRfanNvblsicHJl
            Zml4Il0KICAgICAgICAgICAgICAgICAgICArICIgIgogICAgICAgICAgICAgICAgICAgICsgY21k
            X2pzb25bImVudGl0eSJdCiAgICAgICAgICAgICAgICAgICAgKyAiICIKICAgICAgICAgICAgICAg
            ICAgICArICIgIi5qb2luKGNtZF9qc29uWyJjYXBzIl0pCiAgICAgICAgICAgICAgICApLAogICAg
            ICAgICAgICAgICAgIiIsCiAgICAgICAgICAgICkKICAgICAgICAjIGNoZWNrIGlmIHVzZXIgYWxy
            ZWFkeSBleGlzdAogICAgICAgIHVzZXJfa2V5ID0gc2VsZi5jaGVja191c2VyX2V4aXN0KGVudGl0
            eSkKICAgICAgICBpZiB1c2VyX2tleSAhPSAiIjoKICAgICAgICAgICAgcmV0dXJuIHVzZXJfa2V5
            LCBmIntlbnRpdHkuc3BsaXQoJy4nLCAxKVsxXX0iCiAgICAgICAgICAgICMgZW50aXR5LnNwbGl0
            KCcuJywxKVsxXSB0byByZW5hbWUgZW50aXR5KGNsaWVudC5jc2ktcmJkLW5vZGUpIGFzIGNzaS1y
            YmQtbm9kZQoKICAgICAgICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1v
            bl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nl
            c3NmdWwgYXR0ZW1wdCwKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9
            PSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAg
            ICAgICAgICAgZiInYXV0aCBnZXQtb3ItY3JlYXRlIHt1c2VyfScgY29tbWFuZCBmYWlsZWQuXG4i
            CiAgICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBz
            ZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgKQogICAgICAgIHJldHVybiBzdHIo
            anNvbl9vdXRbMF1bImtleSJdKSwgZiJ7ZW50aXR5LnNwbGl0KCcuJywgMSlbMV19IgogICAgICAg
            ICMgZW50aXR5LnNwbGl0KCcuJywxKVsxXSB0byByZW5hbWUgZW50aXR5KGNsaWVudC5jc2ktcmJk
            LW5vZGUpIGFzIGNzaS1yYmQtbm9kZQoKICAgIGRlZiBnZXRfY2VwaGZzX2RhdGFfcG9vbF9kZXRh
            aWxzKHNlbGYpOgogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAiZnMgbHMiLCAiZm9ybWF0
            IjogImpzb24ifQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAg
            ICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigiY2VwaCAiICsgY21kX2pzb25bInByZWZpeCJdKQogICAg
            ICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dl
            bihjbWRfanNvbikKICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0
            LCByZXBvcnQgYW4gZXJyb3IKICAgICAgICBpZiByZXRfdmFsICE9IDA6CiAgICAgICAgICAgICMg
            aWYgZnMgYW5kIGRhdGFfcG9vbCBhcmd1bWVudHMgYXJlIG5vdCBzZXQsIHNpbGVudGx5IHJldHVy
            bgogICAgICAgICAgICBpZiAoCiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhm
            c19maWxlc3lzdGVtX25hbWUgPT0gIiIKICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFy
            c2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZSA9PSAiIgogICAgICAgICAgICApOgogICAgICAgICAg
            ICAgICAgcmV0dXJuCiAgICAgICAgICAgICMgaWYgdXNlciBoYXMgcHJvdmlkZWQgYW55IG9mIHRo
            ZQogICAgICAgICAgICAjICctLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUnIG9yICctLWNlcGhmcy1k
            YXRhLXBvb2wtbmFtZScgYXJndW1lbnRzLAogICAgICAgICAgICAjIHJhaXNlIGFuIGV4Y2VwdGlv
            biBhcyB3ZSBhcmUgdW5hYmxlIHRvIHZlcmlmeSB0aGUgYXJncwogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiInZnMgbHMnIGNlcGgg
            Y2FsbCBmYWlsZWQgd2l0aCBlcnJvcjoge2Vycl9tc2d9IgogICAgICAgICAgICApCgogICAgICAg
            IG1hdGNoaW5nX2pzb25fb3V0ID0ge30KICAgICAgICAjIGlmICctLWNlcGhmcy1maWxlc3lzdGVt
            LW5hbWUnIGFyZ3VtZW50IGlzIHByb3ZpZGVkLAogICAgICAgICMgY2hlY2sgd2hldGhlciB0aGUg
            cHJvdmlkZWQgZmlsZXN5c3RlbS1uYW1lIGV4aXN0cyBvciBub3QKICAgICAgICBpZiBzZWxmLl9h
            cmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWU6CiAgICAgICAgICAgICMgZ2V0IHRoZSBt
            YXRjaGluZyBsaXN0CiAgICAgICAgICAgIG1hdGNoaW5nX2pzb25fb3V0X2xpc3QgPSBbCiAgICAg
            ICAgICAgICAgICBtYXRjaGVkCiAgICAgICAgICAgICAgICBmb3IgbWF0Y2hlZCBpbiBqc29uX291
            dAogICAgICAgICAgICAgICAgaWYgc3RyKG1hdGNoZWRbIm5hbWUiXSkgPT0gc2VsZi5fYXJnX3Bh
            cnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lCiAgICAgICAgICAgIF0KICAgICAgICAgICAgIyB1
            bmFibGUgdG8gZmluZCBhIG1hdGNoaW5nIGZzLW5hbWUsIHJhaXNlIGFuIGVycm9yCiAgICAgICAg
            ICAgIGlmIGxlbihtYXRjaGluZ19qc29uX291dF9saXN0KSA9PSAwOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBmIkZp
            bGVzeXN0ZW0gcHJvdmlkZWQsICd7c2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9u
            YW1lfScsICIKICAgICAgICAgICAgICAgICAgICBmImlzIG5vdCBmb3VuZCBpbiB0aGUgZnMtbGlz
            dDoge1tzdHIoeFsnbmFtZSddKSBmb3IgeCBpbiBqc29uX291dF19IgogICAgICAgICAgICAgICAg
            KQogICAgICAgICAgICBtYXRjaGluZ19qc29uX291dCA9IG1hdGNoaW5nX2pzb25fb3V0X2xpc3Rb
            MF0KICAgICAgICAjIGlmIGNlcGhmcyBmaWxlc3lzdGVtIG5hbWUgaXMgbm90IHByb3ZpZGVkLAog
            ICAgICAgICMgdHJ5IHRvIGdldCBhIGRlZmF1bHQgZnMgbmFtZSBieSBkb2luZyB0aGUgZm9sbG93
            aW5nCiAgICAgICAgZWxzZToKICAgICAgICAgICAgIyBhLiBjaGVjayBpZiB0aGVyZSBpcyBvbmx5
            IG9uZSBmaWxlc3lzdGVtIGlzIHByZXNlbnQKICAgICAgICAgICAgaWYgbGVuKGpzb25fb3V0KSA9
            PSAxOgogICAgICAgICAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXQgPSBqc29uX291dFswXQogICAg
            ICAgICAgICAjIGIuIG9yIGVsc2UsIGNoZWNrIGlmIGRhdGFfcG9vbCBuYW1lIGlzIHByb3ZpZGVk
            CiAgICAgICAgICAgIGVsaWYgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWU6
            CiAgICAgICAgICAgICAgICAjIGFuZCBpZiBwcmVzZW50LCBjaGVjayB3aGV0aGVyIHRoZXJlIGV4
            aXN0cyBhIGZzIHdoaWNoIGhhcyB0aGUgZGF0YV9wb29sCiAgICAgICAgICAgICAgICBmb3IgZWFj
            aEogaW4ganNvbl9vdXQ6CiAgICAgICAgICAgICAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5j
            ZXBoZnNfZGF0YV9wb29sX25hbWUgaW4gZWFjaEpbImRhdGFfcG9vbHMiXToKICAgICAgICAgICAg
            ICAgICAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXQgPSBlYWNoSgogICAgICAgICAgICAgICAgICAg
            ICAgICBicmVhawogICAgICAgICAgICAgICAgIyBpZiB0aGVyZSBpcyBubyBtYXRjaGluZyBmcyBl
            eGlzdHMsIHRoYXQgbWVhbnMgcHJvdmlkZWQgZGF0YV9wb29sIG5hbWUgaXMgaW52YWxpZAogICAg
            ICAgICAgICAgICAgaWYgbm90IG1hdGNoaW5nX2pzb25fb3V0OgogICAgICAgICAgICAgICAgICAg
            IHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgICAg
            IGYiUHJvdmlkZWQgZGF0YV9wb29sIG5hbWUsIHtzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRh
            X3Bvb2xfbmFtZX0sIgogICAgICAgICAgICAgICAgICAgICAgICAiIGRvZXMgbm90IGV4aXN0cyIK
            ICAgICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICMgYy4gaWYgbm90aGluZyBpcyBzZXQg
            YW5kIGNvdWxkbid0IGZpbmQgYSBkZWZhdWx0LAogICAgICAgICAgICBlbHNlOgogICAgICAgICAg
            ICAgICAgIyBqdXN0IHJldHVybiBzaWxlbnRseQogICAgICAgICAgICAgICAgcmV0dXJuCgogICAg
            ICAgIGlmIG1hdGNoaW5nX2pzb25fb3V0OgogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNl
            cGhmc19maWxlc3lzdGVtX25hbWUgPSBzdHIobWF0Y2hpbmdfanNvbl9vdXRbIm5hbWUiXSkKICAg
            ICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfbWV0YWRhdGFfcG9vbF9uYW1lID0gc3Ry
            KAogICAgICAgICAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXRbIm1ldGFkYXRhX3Bvb2wiXQogICAg
            ICAgICAgICApCgogICAgICAgIGlmIGlzaW5zdGFuY2UobWF0Y2hpbmdfanNvbl9vdXRbImRhdGFf
            cG9vbHMiXSwgbGlzdCk6CiAgICAgICAgICAgICMgaWYgdGhlIHVzZXIgaGFzIGFscmVhZHkgcHJv
            dmlkZWQgZGF0YS1wb29sLW5hbWUsCiAgICAgICAgICAgICMgdGhyb3VnaCAtLWNlcGhmcy1kYXRh
            LXBvb2wtbmFtZQogICAgICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bv
            b2xfbmFtZToKICAgICAgICAgICAgICAgICMgaWYgdGhlIHByb3ZpZGVkIG5hbWUgaXMgbm90IG1h
            dGNoaW5nIHdpdGggdGhlIG9uZSBpbiB0aGUgbGlzdAogICAgICAgICAgICAgICAgaWYgKAogICAg
            ICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lCiAg
            ICAgICAgICAgICAgICAgICAgbm90IGluIG1hdGNoaW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0K
            ICAgICAgICAgICAgICAgICk6CiAgICAgICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAgICAgZiJQcm92aWRlZCBkYXRhLXBv
            b2wtbmFtZTogJ3tzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZX0nLCAiCiAg
            ICAgICAgICAgICAgICAgICAgICAgICJkb2Vzbid0IG1hdGNoIGZyb20gdGhlIGRhdGEtcG9vbHMg
            bGlzdDogIgogICAgICAgICAgICAgICAgICAgICAgICBmIntbc3RyKHgpIGZvciB4IGluIG1hdGNo
            aW5nX2pzb25fb3V0WydkYXRhX3Bvb2xzJ11dfSIKICAgICAgICAgICAgICAgICAgICApCiAgICAg
            ICAgICAgICMgaWYgZGF0YV9wb29sIG5hbWUgaXMgbm90IHByb3ZpZGVkLAogICAgICAgICAgICAj
            IHRoZW4gdHJ5IHRvIGZpbmQgYSBkZWZhdWx0IGRhdGEgcG9vbCBuYW1lCiAgICAgICAgICAgIGVs
            c2U6CiAgICAgICAgICAgICAgICAjIGlmIG5vIGRhdGFfcG9vbHMgZXhpc3QsIHNpbGVudGx5IHJl
            dHVybgogICAgICAgICAgICAgICAgaWYgbGVuKG1hdGNoaW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xz
            Il0pID09IDA6CiAgICAgICAgICAgICAgICAgICAgcmV0dXJuCiAgICAgICAgICAgICAgICBzZWxm
            Ll9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZSA9IHN0cigKICAgICAgICAgICAgICAg
            ICAgICBtYXRjaGluZ19qc29uX291dFsiZGF0YV9wb29scyJdWzBdCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgICMgaWYgdGhlcmUgYXJlIG1vcmUgdGhhbiBvbmUgJ2RhdGFfcG9vbHMnIGV4
            aXN0LAogICAgICAgICAgICAjIHRoZW4gd2FybiB0aGUgdXNlciB0aGF0IHdlIGFyZSB1c2luZyB0
            aGUgc2VsZWN0ZWQgbmFtZQogICAgICAgICAgICBpZiBsZW4obWF0Y2hpbmdfanNvbl9vdXRbImRh
            dGFfcG9vbHMiXSkgPiAxOgogICAgICAgICAgICAgICAgcHJpbnQoCiAgICAgICAgICAgICAgICAg
            ICAgIldBUk5JTkc6IE11bHRpcGxlIGRhdGEgcG9vbHMgZGV0ZWN0ZWQ6ICIKICAgICAgICAgICAg
            ICAgICAgICBmIntbc3RyKHgpIGZvciB4IGluIG1hdGNoaW5nX2pzb25fb3V0WydkYXRhX3Bvb2xz
            J11dfVxuIgogICAgICAgICAgICAgICAgICAgIGYiVXNpbmcgdGhlIGRhdGEtcG9vbDogJ3tzZWxm
            Ll9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZX0nXG4iCiAgICAgICAgICAgICAgICAp
            CgogICAgZGVmIGNyZWF0ZV9jaGVja2VyS2V5KHNlbGYsIHVzZXIpOgogICAgICAgIGNhcHMsIGVu
            dGl0eSA9IHNlbGYuZ2V0X2NhcHNfYW5kX2VudGl0eSh1c2VyKQogICAgICAgIGNtZF9qc29uID0g
            ewogICAgICAgICAgICAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0ZSIsCiAgICAgICAgICAg
            ICJlbnRpdHkiOiBlbnRpdHksCiAgICAgICAgICAgICJjYXBzIjogW2NhcCBmb3IgY2FwX2xpc3Qg
            aW4gbGlzdChjYXBzLml0ZW1zKCkpIGZvciBjYXAgaW4gY2FwX2xpc3RdLAogICAgICAgICAgICAi
            Zm9ybWF0IjogImpzb24iLAogICAgICAgIH0KCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5k
            cnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKAogICAgICAgICAgICAgICAg
            ImNlcGggIgogICAgICAgICAgICAgICAgKyBjbWRfanNvblsicHJlZml4Il0KICAgICAgICAgICAg
            ICAgICsgIiAiCiAgICAgICAgICAgICAgICArIGNtZF9qc29uWyJlbnRpdHkiXQogICAgICAgICAg
            ICAgICAgKyAiICIKICAgICAgICAgICAgICAgICsgIiAiLmpvaW4oY21kX2pzb25bImNhcHMiXSkK
            ICAgICAgICAgICAgKQogICAgICAgICMgY2hlY2sgaWYgdXNlciBhbHJlYWR5IGV4aXN0CiAgICAg
            ICAgdXNlcl9rZXkgPSBzZWxmLmNoZWNrX3VzZXJfZXhpc3QoZW50aXR5KQogICAgICAgIGlmIHVz
            ZXJfa2V5ICE9ICIiOgogICAgICAgICAgICByZXR1cm4gdXNlcl9rZXkKCiAgICAgICAgcmV0X3Zh
            bCwganNvbl9vdXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29u
            KQogICAgICAgICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAg
            aWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmFpc2Ug
            RXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgIGYiJ2F1dGggZ2V0LW9y
            LWNyZWF0ZSB7c2VsZi5ydW5fYXNfdXNlcn0nIGNvbW1hbmQgZmFpbGVkXG4iCiAgICAgICAgICAg
            ICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09V
            VFBVVF9MSVNUfSIKICAgICAgICAgICAgKQogICAgICAgIHJldHVybiBzdHIoanNvbl9vdXRbMF1b
            ImtleSJdKQoKICAgIGRlZiBnZXRfY2VwaF9kYXNoYm9hcmRfbGluayhzZWxmKToKICAgICAgICBj
            bWRfanNvbiA9IHsicHJlZml4IjogIm1nciBzZXJ2aWNlcyIsICJmb3JtYXQiOiAianNvbiJ9CiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2Vs
            Zi5kcnlfcnVuKCJjZXBoICIgKyBjbWRfanNvblsicHJlZml4Il0pCiAgICAgICAgcmV0X3ZhbCwg
            anNvbl9vdXQsIF8gPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAg
            ICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3Zh
            bCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmV0dXJuIE5vbmUKICAg
            ICAgICBpZiAiZGFzaGJvYXJkIiBub3QgaW4ganNvbl9vdXQ6CiAgICAgICAgICAgIHJldHVybiBO
            b25lCiAgICAgICAgcmV0dXJuIGpzb25fb3V0WyJkYXNoYm9hcmQiXQoKICAgIGRlZiBjcmVhdGVf
            cmd3X2FkbWluX29wc191c2VyKHNlbGYpOgogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgInJh
            ZG9zZ3ctYWRtaW4iLAogICAgICAgICAgICAidXNlciIsCiAgICAgICAgICAgICJjcmVhdGUiLAog
            ICAgICAgICAgICAiLS11aWQiLAogICAgICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9P
            UFNfVVNFUl9OQU1FLAogICAgICAgICAgICAiLS1kaXNwbGF5LW5hbWUiLAogICAgICAgICAgICAi
            Um9vayBSR1cgQWRtaW4gT3BzIHVzZXIiLAogICAgICAgICAgICAiLS1jYXBzIiwKICAgICAgICAg
            ICAgImJ1Y2tldHM9Kjt1c2Vycz0qO3VzYWdlPXJlYWQ7bWV0YWRhdGE9cmVhZDt6b25lPXJlYWQi
            LAogICAgICAgICAgICAiLS1yZ3ctcmVhbG0iLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd19yZWFsbV9uYW1lLAogICAgICAgICAgICAiLS1yZ3ctem9uZWdyb3VwIiwKICAgICAgICAg
            ICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZWdyb3VwX25hbWUsCiAgICAgICAgICAgICItLXJn
            dy16b25lIiwKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1lLAogICAg
            ICAgIF0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJl
            dHVybiBzZWxmLmRyeV9ydW4oImNlcGggIiArICIgIi5qb2luKGNtZCkpCiAgICAgICAgdHJ5Ogog
            ICAgICAgICAgICBvdXRwdXQgPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1z
            dWJwcm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vy
            cm9yIGFzIGV4ZWNFcnI6CiAgICAgICAgICAgICMgaWYgdGhlIHVzZXIgYWxyZWFkeSBleGlzdHMs
            IHdlIGp1c3QgcXVlcnkgaXQKICAgICAgICAgICAgaWYgZXhlY0Vyci5yZXR1cm5jb2RlID09IGVy
            cm5vLkVFWElTVDoKICAgICAgICAgICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgICAgICAgICAi
            cmFkb3Nndy1hZG1pbiIsCiAgICAgICAgICAgICAgICAgICAgInVzZXIiLAogICAgICAgICAgICAg
            ICAgICAgICJpbmZvIiwKICAgICAgICAgICAgICAgICAgICAiLS11aWQiLAogICAgICAgICAgICAg
            ICAgICAgIHNlbGYuRVhURVJOQUxfUkdXX0FETUlOX09QU19VU0VSX05BTUUsCiAgICAgICAgICAg
            ICAgICAgICAgIi0tcmd3LXJlYWxtIiwKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFy
            c2VyLnJnd19yZWFsbV9uYW1lLAogICAgICAgICAgICAgICAgICAgICItLXJndy16b25lZ3JvdXAi
            LAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1l
            LAogICAgICAgICAgICAgICAgICAgICItLXJndy16b25lIiwKICAgICAgICAgICAgICAgICAgICBz
            ZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUsCiAgICAgICAgICAgICAgICBdCiAgICAgICAg
            ICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVj
            a19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAgICAgICAgZXhj
            ZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yIGFzIGV4ZWNFcnI6CiAgICAgICAgICAg
            ICAgICAgICAgZXJyX21zZyA9ICgKICAgICAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8g
            ZXhlY3V0ZSBjb21tYW5kIHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAg
            ICAgICAgICAgICAgICAgICAgZiJDb2RlOiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtl
            eGVjRXJyLnN0ZGVycn0iCiAgICAgICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAg
            IHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykKICAgICAgICAgICAgICAgICAgICByZXR1cm4gTm9u
            ZSwgTm9uZSwgRmFsc2UsICItMSIKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVy
            cl9tc2cgPSAoCiAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5k
            IHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICBm
            IkNvZGU6IHtleGVjRXJyLnJldHVybmNvZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykK
            ICAgICAgICAgICAgICAgIHJldHVybiBOb25lLCBOb25lLCBGYWxzZSwgIi0xIgoKICAgICAgICAj
            IGlmIGl0IGlzIHB5dGhvbjIsIGRvbid0IGNoZWNrIGZvciBjZXBoIHZlcnNpb24gZm9yIGFkZGlu
            ZyBgaW5mbz1yZWFkYCBjYXAocmd3X3ZhbGlkYXRpb24pCiAgICAgICAgaWYgc3lzLnZlcnNpb25f
            aW5mby5tYWpvciA8IDM6CiAgICAgICAgICAgIGpzb25vdXRwdXQgPSBqc29uLmxvYWRzKG91dHB1
            dCkKICAgICAgICAgICAgcmV0dXJuICgKICAgICAgICAgICAgICAgIGpzb25vdXRwdXRbImtleXMi
            XVswXVsiYWNjZXNzX2tleSJdLAogICAgICAgICAgICAgICAganNvbm91dHB1dFsia2V5cyJdWzBd
            WyJzZWNyZXRfa2V5Il0sCiAgICAgICAgICAgICAgICBGYWxzZSwKICAgICAgICAgICAgICAgICIi
            LAogICAgICAgICAgICApCgogICAgICAgICMgc2VwYXJhdGVseSBhZGQgaW5mbz1yZWFkIGNhcHMg
            Zm9yIHJndy1lbmRwb2ludCBpcCB2YWxpZGF0aW9uCiAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVk
            ID0gVHJ1ZQogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAg
            ICAgICAgICAiY2FwcyIsCiAgICAgICAgICAgICJhZGQiLAogICAgICAgICAgICAiLS11aWQiLAog
            ICAgICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9PUFNfVVNFUl9OQU1FLAogICAgICAg
            ICAgICAiLS1jYXBzIiwKICAgICAgICAgICAgImluZm89cmVhZCIsCiAgICAgICAgICAgICItLXJn
            dy1yZWFsbSIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUsCiAg
            ICAgICAgICAgICItLXJndy16b25lZ3JvdXAiLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd196b25lZ3JvdXBfbmFtZSwKICAgICAgICAgICAgIi0tcmd3LXpvbmUiLAogICAgICAgICAg
            ICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUsCiAgICAgICAgXQogICAgICAgIHRyeToK
            ICAgICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9
            c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2VwdCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NF
            cnJvciBhcyBleGVjRXJyOgogICAgICAgICAgICAjIGlmIHRoZSBjZXBoIHZlcnNpb24gbm90IHN1
            cHBvcnRlZCBmb3IgYWRkaW5nIGBpbmZvPXJlYWRgIGNhcChyZ3dfdmFsaWRhdGlvbikKICAgICAg
            ICAgICAgaWYgKAogICAgICAgICAgICAgICAgImNvdWxkIG5vdCBhZGQgY2FwczogdW5hYmxlIHRv
            IGFkZCBjYXBzOiBpbmZvPXJlYWRcbiIKICAgICAgICAgICAgICAgIGluIGV4ZWNFcnIuc3RkZXJy
            LmRlY29kZSgidXRmLTgiKQogICAgICAgICAgICAgICAgYW5kIGV4ZWNFcnIucmV0dXJuY29kZSA9
            PSAyNDQKICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgIGluZm9fY2FwX3N1cHBvcnRlZCA9
            IEZhbHNlCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICBlcnJfbXNnID0gKAogICAg
            ICAgICAgICAgICAgICAgIGYiZmFpbGVkIHRvIGV4ZWN1dGUgY29tbWFuZCB7Y21kfS4gT3V0cHV0
            OiB7ZXhlY0Vyci5vdXRwdXR9LiAiCiAgICAgICAgICAgICAgICAgICAgZiJDb2RlOiB7ZXhlY0Vy
            ci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtleGVjRXJyLnN0ZGVycn0iCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKGVycl9tc2cpCiAgICAgICAgICAgICAg
            ICByZXR1cm4gTm9uZSwgTm9uZSwgRmFsc2UsICItMSIKCiAgICAgICAganNvbm91dHB1dCA9IGpz
            b24ubG9hZHMob3V0cHV0KQogICAgICAgIHJldHVybiAoCiAgICAgICAgICAgIGpzb25vdXRwdXRb
            ImtleXMiXVswXVsiYWNjZXNzX2tleSJdLAogICAgICAgICAgICBqc29ub3V0cHV0WyJrZXlzIl1b
            MF1bInNlY3JldF9rZXkiXSwKICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkLAogICAgICAg
            ICAgICAiIiwKICAgICAgICApCgogICAgZGVmIHZhbGlkYXRlX3JiZF9wb29sKHNlbGYpOgogICAg
            ICAgIGlmIG5vdCBzZWxmLmNsdXN0ZXIucG9vbF9leGlzdHMoc2VsZi5fYXJnX3BhcnNlci5yYmRf
            ZGF0YV9wb29sX25hbWUpOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0
            aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcG9vbCwgJ3tzZWxmLl9hcmdfcGFy
            c2VyLnJiZF9kYXRhX3Bvb2xfbmFtZX0nLCBkb2VzIG5vdCBleGlzdCIKICAgICAgICAgICAgKQoK
            ICAgIGRlZiBpbml0X3JiZF9wb29sKHNlbGYpOgogICAgICAgIGlmIGlzaW5zdGFuY2Uoc2VsZi5j
            bHVzdGVyLCBEdW1teVJhZG9zKToKICAgICAgICAgICAgcmV0dXJuCiAgICAgICAgcmJkX3Bvb2xf
            bmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgaW9jdHgg
            PSBzZWxmLmNsdXN0ZXIub3Blbl9pb2N0eChyYmRfcG9vbF9uYW1lKQogICAgICAgIHJiZF9pbnN0
            ID0gcmJkLlJCRCgpCiAgICAgICAgcmJkX2luc3QucG9vbF9pbml0KGlvY3R4LCBUcnVlKQoKICAg
            IGRlZiB2YWxpZGF0ZV9yYWRvc19uYW1lc3BhY2Uoc2VsZik6CiAgICAgICAgcmJkX3Bvb2xfbmFt
            ZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgcmFkb3NfbmFt
            ZXNwYWNlID0gc2VsZi5fYXJnX3BhcnNlci5yYWRvc19uYW1lc3BhY2UKICAgICAgICBpZiByYWRv
            c19uYW1lc3BhY2UgPT0gIiI6CiAgICAgICAgICAgIHJldHVybgogICAgICAgIGlmIHJhZG9zX25h
            bWVzcGFjZS5pc2xvd2VyKCkgPT0gRmFsc2U6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCByYWRvcyBOYW1l
            c3BhY2UsICd7cmFkb3NfbmFtZXNwYWNlfScsICIKICAgICAgICAgICAgICAgIGYiY29udGFpbnMg
            dXBwZXIgY2FzZSIKICAgICAgICAgICAgKQogICAgICAgIHJiZF9pbnN0ID0gcmJkLlJCRCgpCiAg
            ICAgICAgaW9jdHggPSBzZWxmLmNsdXN0ZXIub3Blbl9pb2N0eChyYmRfcG9vbF9uYW1lKQogICAg
            ICAgIGlmIHJiZF9pbnN0Lm5hbWVzcGFjZV9leGlzdHMoaW9jdHgsIHJhZG9zX25hbWVzcGFjZSkg
            aXMgRmFsc2U6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAg
            ICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCByYWRvcyBOYW1lc3BhY2UsICd7cmFkb3NfbmFt
            ZXNwYWNlfScsICIKICAgICAgICAgICAgICAgIGYiaXMgbm90IGZvdW5kIGluIHRoZSBwb29sICd7
            cmJkX3Bvb2xfbmFtZX0nIgogICAgICAgICAgICApCgogICAgZGVmIGdldF9vcl9jcmVhdGVfc3Vi
            dm9sdW1lX2dyb3VwKHNlbGYsIHN1YnZvbHVtZV9ncm91cCwgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFt
            ZSk6CiAgICAgICAgY21kID0gWwogICAgICAgICAgICAiY2VwaCIsCiAgICAgICAgICAgICJmcyIs
            CiAgICAgICAgICAgICJzdWJ2b2x1bWVncm91cCIsCiAgICAgICAgICAgICJnZXRwYXRoIiwKICAg
            ICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAgICAgICAgICAgc3Vidm9sdW1lX2dy
            b3VwLAogICAgICAgIF0KICAgICAgICB0cnk6CiAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNo
            ZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1
            YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICBjbWQgPSBbCiAgICAgICAg
            ICAgICAgICAiY2VwaCIsCiAgICAgICAgICAgICAgICAiZnMiLAogICAgICAgICAgICAgICAgInN1
            YnZvbHVtZWdyb3VwIiwKICAgICAgICAgICAgICAgICJjcmVhdGUiLAogICAgICAgICAgICAgICAg
            Y2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAgICAgICAgICAgICAgIHN1YnZvbHVtZV9ncm91cCwK
            ICAgICAgICAgICAgXQogICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBfID0gc3VicHJv
            Y2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAg
            ICBleGNlcHQgc3VicHJvY2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3I6CiAgICAgICAgICAgICAgICBy
            YWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYic3Vi
            dm9sdW1lIGdyb3VwIHtzdWJ2b2x1bWVfZ3JvdXB9IGlzIG5vdCBhYmxlIHRvIGdldCBjcmVhdGVk
            IgogICAgICAgICAgICAgICAgKQoKICAgIGRlZiBwaW5fc3Vidm9sdW1lKAogICAgICAgIHNlbGYs
            IHN1YnZvbHVtZV9ncm91cCwgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwgcGluX3R5cGUsIHBpbl9z
            ZXR0aW5nCiAgICApOgogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgImNlcGgiLAogICAgICAg
            ICAgICAiZnMiLAogICAgICAgICAgICAic3Vidm9sdW1lZ3JvdXAiLAogICAgICAgICAgICAicGlu
            IiwKICAgICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAgICAgICAgICAgc3Vidm9s
            dW1lX2dyb3VwLAogICAgICAgICAgICBwaW5fdHlwZSwKICAgICAgICAgICAgcGluX3NldHRpbmcs
            CiAgICAgICAgXQogICAgICAgIHRyeToKICAgICAgICAgICAgXyA9IHN1YnByb2Nlc3MuY2hlY2tf
            b3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAgICAgICBleGNlcHQgc3VicHJv
            Y2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3I6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1
            cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmInN1YnZvbHVtZSBncm91cCB7c3Vidm9sdW1l
            X2dyb3VwfSBpcyBub3QgYWJsZSB0byBnZXQgcGlubmVkIgogICAgICAgICAgICApCgogICAgZGVm
            IGdldF9yZ3dfZnNpZChzZWxmLCBiYXNlX3VybCwgdmVyaWZ5KToKICAgICAgICBhY2Nlc3Nfa2V5
            ID0gc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJfQUNDRVNTX0tFWSJdCiAgICAgICAg
            c2VjcmV0X2tleSA9IHNlbGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX1NFQ1JFVF9LRVki
            XQogICAgICAgIHJnd19lbmRwb2ludCA9IHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50CiAg
            ICAgICAgYmFzZV91cmwgPSBiYXNlX3VybCArICI6Ly8iICsgcmd3X2VuZHBvaW50ICsgIi9hZG1p
            bi9pbmZvPyIKICAgICAgICBwYXJhbXMgPSB7ImZvcm1hdCI6ICJqc29uIn0KICAgICAgICByZXF1
            ZXN0X3VybCA9IGJhc2VfdXJsICsgdXJsZW5jb2RlKHBhcmFtcykKCiAgICAgICAgdHJ5OgogICAg
            ICAgICAgICByID0gcmVxdWVzdHMuZ2V0KAogICAgICAgICAgICAgICAgcmVxdWVzdF91cmwsCiAg
            ICAgICAgICAgICAgICBhdXRoPVMzQXV0aChhY2Nlc3Nfa2V5LCBzZWNyZXRfa2V5LCByZ3dfZW5k
            cG9pbnQpLAogICAgICAgICAgICAgICAgdmVyaWZ5PXZlcmlmeSwKICAgICAgICAgICAgKQogICAg
            ICAgIGV4Y2VwdCByZXF1ZXN0cy5leGNlcHRpb25zLlRpbWVvdXQ6CiAgICAgICAgICAgIHN5cy5z
            dGRlcnIud3JpdGUoCiAgICAgICAgICAgICAgICBmImludmFsaWQgZW5kcG9pbnQ6LCBub3QgYWJs
            ZSB0byBjYWxsIGFkbWluLW9wcyBhcGl7cmd3X2VuZHBvaW50fSIKICAgICAgICAgICAgKQogICAg
            ICAgICAgICByZXR1cm4gIiIsICItMSIKICAgICAgICByMSA9IHIuanNvbigpCiAgICAgICAgaWYg
            cjEgaXMgTm9uZSBvciByMS5nZXQoImluZm8iKSBpcyBOb25lOgogICAgICAgICAgICBzeXMuc3Rk
            ZXJyLndyaXRlKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmd3IEVuZHBvaW50LCAn
            e3NlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50fScsIGlzIGludmFsaWQuIgogICAgICAgICAg
            ICApCiAgICAgICAgICAgIHJldHVybiAoCiAgICAgICAgICAgICAgICAiIiwKICAgICAgICAgICAg
            ICAgICItMSIsCiAgICAgICAgICAgICkKCiAgICAgICAgcmV0dXJuIHIxWyJpbmZvIl1bInN0b3Jh
            Z2VfYmFja2VuZHMiXVswXVsiY2x1c3Rlcl9pZCJdLCAiIgoKICAgIGRlZiB2YWxpZGF0ZV9yZ3df
            ZW5kcG9pbnQoc2VsZiwgaW5mb19jYXBfc3VwcG9ydGVkKToKICAgICAgICAjIGlmIHRoZSAnY2x1
            c3RlcicgaW5zdGFuY2UgaXMgYSBkdW1teSBvbmUsCiAgICAgICAgIyBkb24ndCB0cnkgdG8gcmVh
            Y2ggb3V0IHRvIHRoZSBlbmRwb2ludAogICAgICAgIGlmIGlzaW5zdGFuY2Uoc2VsZi5jbHVzdGVy
            LCBEdW1teVJhZG9zKToKICAgICAgICAgICAgcmV0dXJuCgogICAgICAgIHJnd19lbmRwb2ludCA9
            IHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50CgogICAgICAgICMgdmFsaWRhdGUgcmd3IGVu
            ZHBvaW50IG9ubHkgaWYgaXAgYWRkcmVzcyBpcyBwYXNzZWQKICAgICAgICBpcF90eXBlID0gc2Vs
            Zi5faW52YWxpZF9lbmRwb2ludChyZ3dfZW5kcG9pbnQpCgogICAgICAgICMgY2hlY2sgaWYgdGhl
            IHJndyBlbmRwb2ludCBpcyByZWFjaGFibGUKICAgICAgICBjZXJ0ID0gTm9uZQogICAgICAgIGlm
            IG5vdCBzZWxmLl9hcmdfcGFyc2VyLnJnd19za2lwX3RscyBhbmQgc2VsZi52YWxpZGF0ZV9yZ3df
            ZW5kcG9pbnRfdGxzX2NlcnQoKToKICAgICAgICAgICAgY2VydCA9IHNlbGYuX2FyZ19wYXJzZXIu
            cmd3X3Rsc19jZXJ0X3BhdGgKICAgICAgICBiYXNlX3VybCwgdmVyaWZ5LCBlcnIgPSBzZWxmLmVu
            ZHBvaW50X2RpYWwocmd3X2VuZHBvaW50LCBpcF90eXBlLCBjZXJ0PWNlcnQpCiAgICAgICAgaWYg
            ZXJyICE9ICIiOgogICAgICAgICAgICByZXR1cm4gIi0xIgoKICAgICAgICAjIGNoZWNrIGlmIHRo
            ZSByZ3cgZW5kcG9pbnQgYmVsb25ncyB0byB0aGUgc2FtZSBjbHVzdGVyCiAgICAgICAgIyBvbmx5
            IGNoZWNrIGlmIGBpbmZvYCBjYXAgaXMgc3VwcG9ydGVkCiAgICAgICAgaWYgaW5mb19jYXBfc3Vw
            cG9ydGVkOgogICAgICAgICAgICBmc2lkID0gc2VsZi5nZXRfZnNpZCgpCiAgICAgICAgICAgIHJn
            d19mc2lkLCBlcnIgPSBzZWxmLmdldF9yZ3dfZnNpZChiYXNlX3VybCwgdmVyaWZ5KQogICAgICAg
            ICAgICBpZiBlcnIgPT0gIi0xIjoKICAgICAgICAgICAgICAgIHJldHVybiAiLTEiCiAgICAgICAg
            ICAgIGlmIGZzaWQgIT0gcmd3X2ZzaWQ6CiAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRl
            KAogICAgICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHJndyBFbmRwb2ludCwgJ3tzZWxm
            Ll9hcmdfcGFyc2VyLnJnd19lbmRwb2ludH0nLCBpcyBpbnZhbGlkLiBXZSBhcmUgdmFsaWRhdGlu
            ZyBieSBjYWxsaW5nIHRoZSBhZG1pbm9wcyBhcGkgdGhyb3VnaCByZ3ctZW5kcG9pbnQgYW5kIHZh
            bGlkYXRpbmcgdGhlIGNsdXN0ZXJfaWQgJ3tyZ3dfZnNpZH0nIGlzIGVxdWFsIHRvIHRoZSBjZXBo
            IGNsdXN0ZXIgZnNpZCAne2ZzaWR9JyIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAg
            IHJldHVybiAiLTEiCgogICAgICAgICMgY2hlY2sgaWYgdGhlIHJndyBlbmRwb2ludCBwb29sIGV4
            aXN0CiAgICAgICAgIyBvbmx5IHZhbGlkYXRlIGlmIHJnd19wb29sX3ByZWZpeCBpcyBwYXNzZWQg
            ZWxzZSBpdCB3aWxsIHRha2UgZGVmYXVsdCB2YWx1ZSBhbmQgd2UgZG9uJ3QgY3JlYXRlIHRoZXNl
            IGRlZmF1bHQgcG9vbHMKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZp
            eCAhPSAiZGVmYXVsdCI6CiAgICAgICAgICAgIHJnd19wb29sc190b192YWxpZGF0ZSA9IFsKICAg
            ICAgICAgICAgICAgIGYie3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cubWV0
            YSIsCiAgICAgICAgICAgICAgICAiLnJndy5yb290IiwKICAgICAgICAgICAgICAgIGYie3NlbGYu
            X2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cuY29udHJvbCIsCiAgICAgICAgICAgICAg
            ICBmIntzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmxvZyIsCiAgICAgICAg
            ICAgIF0KICAgICAgICAgICAgZm9yIF9yZ3dfcG9vbF90b192YWxpZGF0ZSBpbiByZ3dfcG9vbHNf
            dG9fdmFsaWRhdGU6CiAgICAgICAgICAgICAgICBpZiBub3Qgc2VsZi5jbHVzdGVyLnBvb2xfZXhp
            c3RzKF9yZ3dfcG9vbF90b192YWxpZGF0ZSk6CiAgICAgICAgICAgICAgICAgICAgc3lzLnN0ZGVy
            ci53cml0ZSgKICAgICAgICAgICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcG9vbCwgJ3tf
            cmd3X3Bvb2xfdG9fdmFsaWRhdGV9JywgZG9lcyBub3QgZXhpc3QiCiAgICAgICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICAgICAgICAgIHJldHVybiAiLTEiCgogICAgICAgIHJldHVybiAiIgoK
            ICAgIGRlZiB2YWxpZGF0ZV9yZ3dfbXVsdGlzaXRlKHNlbGYsIHJnd19tdWx0aXNpdGVfY29uZmln
            X25hbWUsIHJnd19tdWx0aXNpdGVfY29uZmlnKToKICAgICAgICBpZiByZ3dfbXVsdGlzaXRlX2Nv
            bmZpZyAhPSAiIjoKICAgICAgICAgICAgY21kID0gWwogICAgICAgICAgICAgICAgInJhZG9zZ3ct
            YWRtaW4iLAogICAgICAgICAgICAgICAgcmd3X211bHRpc2l0ZV9jb25maWcsCiAgICAgICAgICAg
            ICAgICAiZ2V0IiwKICAgICAgICAgICAgICAgICItLXJndy0iICsgcmd3X211bHRpc2l0ZV9jb25m
            aWcsCiAgICAgICAgICAgICAgICByZ3dfbXVsdGlzaXRlX2NvbmZpZ19uYW1lLAogICAgICAgICAg
            ICBdCiAgICAgICAgICAgIHRyeToKICAgICAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNoZWNr
            X291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgICAgIGV4Y2VwdCBz
            dWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJvciBhcyBleGVjRXJyOgogICAgICAgICAgICAgICAg
            ZXJyX21zZyA9ICgKICAgICAgICAgICAgICAgICAgICBmImZhaWxlZCB0byBleGVjdXRlIGNvbW1h
            bmQge2NtZH0uIE91dHB1dDoge2V4ZWNFcnIub3V0cHV0fS4gIgogICAgICAgICAgICAgICAgICAg
            IGYiQ29kZToge2V4ZWNFcnIucmV0dXJuY29kZX0uIEVycm9yOiB7ZXhlY0Vyci5zdGRlcnJ9Igog
            ICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZShlcnJfbXNn
            KQogICAgICAgICAgICAgICAgcmV0dXJuICItMSIKICAgICAgICByZXR1cm4gIiIKCiAgICBkZWYg
            X2dlbl9vdXRwdXRfbWFwKHNlbGYpOgogICAgICAgIGlmIHNlbGYub3V0X21hcDoKICAgICAgICAg
            ICAgcmV0dXJuCiAgICAgICAgIyBzdXBwb3J0IGxlZ2FjeSBmbGFnIHdpdGggdXBncmFkZXMKICAg
            ICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNsdXN0ZXJfbmFtZToKICAgICAgICAgICAgc2VsZi5f
            YXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5jbHVzdGVyX25h
            bWUKICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUgPSAoCiAgICAgICAg
            ICAgIHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZS5sb3dlcigpCiAgICAgICAgKSAg
            IyBhbHdheXMgY29udmVydCBjbHVzdGVyIG5hbWUgdG8gbG93ZXJjYXNlIGNoYXJhY3RlcnMKICAg
            ICAgICBzZWxmLnZhbGlkYXRlX3JiZF9wb29sKCkKICAgICAgICBzZWxmLmluaXRfcmJkX3Bvb2wo
            KQogICAgICAgIHNlbGYudmFsaWRhdGVfcmFkb3NfbmFtZXNwYWNlKCkKICAgICAgICBzZWxmLl9l
            eGNsdWRlZF9rZXlzLmFkZCgiSzhTX0NMVVNURVJfTkFNRSIpCiAgICAgICAgc2VsZi5nZXRfY2Vw
            aGZzX2RhdGFfcG9vbF9kZXRhaWxzKCkKICAgICAgICBzZWxmLm91dF9tYXBbIk5BTUVTUEFDRSJd
            ID0gc2VsZi5fYXJnX3BhcnNlci5uYW1lc3BhY2UKICAgICAgICBzZWxmLm91dF9tYXBbIks4U19D
            TFVTVEVSX05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQogICAgICAg
            IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9GU0lEIl0gPSBzZWxmLmdldF9mc2lkKCkKICAg
            ICAgICBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUk5BTUUiXSA9IHNlbGYucnVuX2Fz
            X3VzZXIKICAgICAgICBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfQ0VQSF9NT05fREFUQSJd
            ID0gc2VsZi5nZXRfY2VwaF9leHRlcm5hbF9tb25fZGF0YSgpCiAgICAgICAgc2VsZi5vdXRfbWFw
            WyJST09LX0VYVEVSTkFMX1VTRVJfU0VDUkVUIl0gPSBzZWxmLmNyZWF0ZV9jaGVja2VyS2V5KAog
            ICAgICAgICAgICAiY2xpZW50LmhlYWx0aGNoZWNrZXIiCiAgICAgICAgKQogICAgICAgIHNlbGYu
            b3V0X21hcFsiUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSyJdID0gc2VsZi5nZXRfY2VwaF9k
            YXNoYm9hcmRfbGluaygpCiAgICAgICAgKAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9S
            QkRfTk9ERV9TRUNSRVQiXSwKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVf
            U0VDUkVUX05BTUUiXSwKICAgICAgICApID0gc2VsZi5jcmVhdGVfY2VwaENTSUtleXJpbmdfdXNl
            cigiY2xpZW50LmNzaS1yYmQtbm9kZSIpCiAgICAgICAgKAogICAgICAgICAgICBzZWxmLm91dF9t
            YXBbIkNTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUIl0sCiAgICAgICAgICAgIHNlbGYub3V0X21h
            cFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICkgPSBzZWxmLmNy
            ZWF0ZV9jZXBoQ1NJS2V5cmluZ191c2VyKCJjbGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIpCiAg
            ICAgICAgc2VsZi5vdXRfbWFwWyJDRVBIRlNfUE9PTF9OQU1FIl0gPSBzZWxmLl9hcmdfcGFyc2Vy
            LmNlcGhmc19kYXRhX3Bvb2xfbmFtZQogICAgICAgIHNlbGYub3V0X21hcFsiQ0VQSEZTX01FVEFE
            QVRBX1BPT0xfTkFNRSJdID0gKAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19t
            ZXRhZGF0YV9wb29sX25hbWUKICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJDRVBIRlNf
            RlNfTkFNRSJdID0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lCiAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJSRVNUUklDVEVEX0FVVEhfUEVSTUlTU0lPTiJdID0gKAogICAgICAg
            ICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uCiAgICAgICAg
            KQogICAgICAgIHNlbGYub3V0X21hcFsiUkFET1NfTkFNRVNQQUNFIl0gPSBzZWxmLl9hcmdfcGFy
            c2VyLnJhZG9zX25hbWVzcGFjZQogICAgICAgIHNlbGYub3V0X21hcFsiU1VCVk9MVU1FX0dST1VQ
            Il0gPSBzZWxmLl9hcmdfcGFyc2VyLnN1YnZvbHVtZV9ncm91cAogICAgICAgIHNlbGYub3V0X21h
            cFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVCJdID0gIiIKICAgICAgICBzZWxmLm91dF9tYXBbIkNT
            SV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUIl0gPSAiIgogICAgICAgICMgY3JlYXRlIENlcGhG
            UyBub2RlIGFuZCBwcm92aXNpb25lciBrZXlyaW5nIG9ubHkgd2hlbiBNRFMgZXhpc3RzCiAgICAg
            ICAgaWYgc2VsZi5vdXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdIGFuZCBzZWxmLm91dF9tYXBbIkNF
            UEhGU19QT09MX05BTUUiXToKICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgc2VsZi5vdXRf
            bWFwWyJDU0lfQ0VQSEZTX05PREVfU0VDUkVUIl0sCiAgICAgICAgICAgICAgICBzZWxmLm91dF9t
            YXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICApID0gc2VsZi5j
            cmVhdGVfY2VwaENTSUtleXJpbmdfdXNlcigiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIpCiAgICAg
            ICAgICAgICgKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9WSVNJ
            T05FUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19Q
            Uk9WSVNJT05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICApID0gc2VsZi5jcmVhdGVfY2Vw
            aENTSUtleXJpbmdfdXNlcigiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiKQogICAgICAg
            ICAgICAjIGNyZWF0ZSB0aGUgZGVmYXVsdCAiY3NpIiBzdWJ2b2x1bWVncm91cAogICAgICAgICAg
            ICBzZWxmLmdldF9vcl9jcmVhdGVfc3Vidm9sdW1lX2dyb3VwKAogICAgICAgICAgICAgICAgImNz
            aSIsIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgICAgICAp
            CiAgICAgICAgICAgICMgcGluIHRoZSBkZWZhdWx0ICJjc2kiIHN1YnZvbHVtZWdyb3VwCiAgICAg
            ICAgICAgIHNlbGYucGluX3N1YnZvbHVtZSgKICAgICAgICAgICAgICAgICJjc2kiLCBzZWxmLl9h
            cmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUsICJkaXN0cmlidXRlZCIsICIxIgogICAg
            ICAgICAgICApCiAgICAgICAgICAgIGlmIHNlbGYub3V0X21hcFsiU1VCVk9MVU1FX0dST1VQIl06
            CiAgICAgICAgICAgICAgICBzZWxmLmdldF9vcl9jcmVhdGVfc3Vidm9sdW1lX2dyb3VwKAogICAg
            ICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuc3Vidm9sdW1lX2dyb3VwLAogICAgICAg
            ICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgIHNlbGYucGluX3N1YnZvbHVtZSgKICAgICAg
            ICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnN1YnZvbHVtZV9ncm91cCwKICAgICAgICAg
            ICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAg
            ICAgICAgICAgICAgICAgImRpc3RyaWJ1dGVkIiwKICAgICAgICAgICAgICAgICAgICAiMSIsCiAg
            ICAgICAgICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJSR1dfVExTX0NFUlQiXSA9ICIi
            CiAgICAgICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UIl0gPSAiIgogICAgICAg
            IHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUIl0gPSAiIgogICAgICAgIGlm
            IG5vdCBzZWxmLl9hcmdfcGFyc2VyLnNraXBfbW9uaXRvcmluZ19lbmRwb2ludDoKICAgICAgICAg
            ICAgKAogICAgICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UIl0s
            CiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJd
            LAogICAgICAgICAgICApID0gc2VsZi5nZXRfYWN0aXZlX2FuZF9zdGFuZGJ5X21ncnMoKQogICAg
            ICAgIHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdID0gc2VsZi5fYXJnX3BhcnNlci5yYmRf
            ZGF0YV9wb29sX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBbIlJCRF9NRVRBREFUQV9FQ19QT09M
            X05BTUUiXSA9ICgKICAgICAgICAgICAgc2VsZi52YWxpZGF0ZV9yYmRfbWV0YWRhdGFfZWNfcG9v
            bF9uYW1lKCkKICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJSR1dfUE9PTF9QUkVGSVgi
            XSA9IHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4CiAgICAgICAgc2VsZi5vdXRfbWFw
            WyJSR1dfRU5EUE9JTlQiXSA9ICIiCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZ3dfZW5k
            cG9pbnQ6CiAgICAgICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAg
            ICAgICAgIHNlbGYuY3JlYXRlX3Jnd19hZG1pbl9vcHNfdXNlcigpCiAgICAgICAgICAgIGVsc2U6
            CiAgICAgICAgICAgICAgICBpZiAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfcmVhbG1fbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdf
            cGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBz
            ZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICk6CiAg
            ICAgICAgICAgICAgICAgICAgZXJyID0gc2VsZi52YWxpZGF0ZV9yZ3dfbXVsdGlzaXRlKAogICAg
            ICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lLCAicmVh
            bG0iCiAgICAgICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAgIGVyciA9IHNlbGYu
            dmFsaWRhdGVfcmd3X211bHRpc2l0ZSgKICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJn
            X3BhcnNlci5yZ3dfem9uZWdyb3VwX25hbWUsICJ6b25lZ3JvdXAiCiAgICAgICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICAgICAgICAgIGVyciA9IHNlbGYudmFsaWRhdGVfcmd3X211bHRpc2l0
            ZSgKICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1l
            LCAiem9uZSIKICAgICAgICAgICAgICAgICAgICApCgogICAgICAgICAgICAgICAgaWYgKAogICAg
            ICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUgPT0gIiIKICAg
            ICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZWdyb3VwX25hbWUg
            PT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZV9u
            YW1lID09ICIiCiAgICAgICAgICAgICAgICApIG9yICgKICAgICAgICAgICAgICAgICAgICBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lICE9ICIiCiAgICAgICAgICAgICAgICAgICAgYW5k
            IHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lICE9ICIiCiAgICAgICAgICAgICAg
            ICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVfbmFtZSAhPSAiIgogICAgICAgICAg
            ICAgICAgKToKICAgICAgICAgICAgICAgICAgICAoCiAgICAgICAgICAgICAgICAgICAgICAgIHNl
            bGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX0FDQ0VTU19LRVkiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJfU0VDUkVUX0tFWSJd
            LAogICAgICAgICAgICAgICAgICAgICAgICBpbmZvX2NhcF9zdXBwb3J0ZWQsCiAgICAgICAgICAg
            ICAgICAgICAgICAgIGVyciwKICAgICAgICAgICAgICAgICAgICApID0gc2VsZi5jcmVhdGVfcmd3
            X2FkbWluX29wc191c2VyKCkKICAgICAgICAgICAgICAgICAgICBlcnIgPSBzZWxmLnZhbGlkYXRl
            X3Jnd19lbmRwb2ludChpbmZvX2NhcF9zdXBwb3J0ZWQpCiAgICAgICAgICAgICAgICAgICAgaWYg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfdGxzX2NlcnRfcGF0aDoKICAgICAgICAgICAgICAgICAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJSR1dfVExTX0NFUlQiXSA9ICgKICAgICAgICAgICAgICAgICAgICAg
            ICAgICAgIHNlbGYudmFsaWRhdGVfcmd3X2VuZHBvaW50X3Rsc19jZXJ0KCkKICAgICAgICAgICAg
            ICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAgICMgaWYgdGhlcmUgaXMgbm8gZXJyb3Is
            IHNldCB0aGUgUkdXX0VORFBPSU5UCiAgICAgICAgICAgICAgICAgICAgaWYgZXJyICE9ICItMSI6
            CiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiUkdXX0VORFBPSU5UIl0gPSBz
            ZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludAogICAgICAgICAgICAgICAgZWxzZToKICAgICAg
            ICAgICAgICAgICAgICBlcnIgPSAiUGxlYXNlIHByb3ZpZGUgYWxsIHRoZSBSR1cgbXVsdGlzaXRl
            IHBhcmFtZXRlcnMgb3Igbm9uZSBvZiB0aGVtIgogICAgICAgICAgICAgICAgICAgIHN5cy5zdGRl
            cnIud3JpdGUoZXJyKQoKICAgIGRlZiBnZW5fc2hlbGxfb3V0KHNlbGYpOgogICAgICAgIHNlbGYu
            X2dlbl9vdXRwdXRfbWFwKCkKICAgICAgICBzaE91dElPID0gU3RyaW5nSU8oKQogICAgICAgIGZv
            ciBrLCB2IGluIHNlbGYub3V0X21hcC5pdGVtcygpOgogICAgICAgICAgICBpZiB2IGFuZCBrIG5v
            dCBpbiBzZWxmLl9leGNsdWRlZF9rZXlzOgogICAgICAgICAgICAgICAgc2hPdXRJTy53cml0ZShm
            ImV4cG9ydCB7a309e3Z9e0xJTkVTRVB9IikKICAgICAgICBzaE91dCA9IHNoT3V0SU8uZ2V0dmFs
            dWUoKQogICAgICAgIHNoT3V0SU8uY2xvc2UoKQogICAgICAgIHJldHVybiBzaE91dAoKICAgIGRl
            ZiBnZW5fanNvbl9vdXQoc2VsZik6CiAgICAgICAgc2VsZi5fZ2VuX291dHB1dF9tYXAoKQogICAg
            ICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuICIiCiAg
            ICAgICAganNvbl9vdXQgPSBbCiAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICJuYW1lIjog
            InJvb2stY2VwaC1tb24tZW5kcG9pbnRzIiwKICAgICAgICAgICAgICAgICJraW5kIjogIkNvbmZp
            Z01hcCIsCiAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAiZGF0
            YSI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9DRVBIX01PTl9EQVRBIl0sCiAgICAgICAg
            ICAgICAgICAgICAgIm1heE1vbklkIjogIjAiLAogICAgICAgICAgICAgICAgICAgICJtYXBwaW5n
            IjogInt9IiwKICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgIH0sCiAgICAgICAgICAgIHsK
            ICAgICAgICAgICAgICAgICJuYW1lIjogInJvb2stY2VwaC1tb24iLAogICAgICAgICAgICAgICAg
            ImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAg
            ICAgICAgICJhZG1pbi1zZWNyZXQiOiAiYWRtaW4tc2VjcmV0IiwKICAgICAgICAgICAgICAgICAg
            ICAiZnNpZCI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9GU0lEIl0sCiAgICAgICAgICAg
            ICAgICAgICAgIm1vbi1zZWNyZXQiOiAibW9uLXNlY3JldCIsCiAgICAgICAgICAgICAgICB9LAog
            ICAgICAgICAgICB9LAogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFtZSI6ICJyb29r
            LWNlcGgtb3BlcmF0b3ItY3JlZHMiLAogICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwK
            ICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICJ1c2VySUQiOiBz
            ZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUk5BTUUiXSwKICAgICAgICAgICAgICAgICAg
            ICAidXNlcktleSI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9VU0VSX1NFQ1JFVCJdLAog
            ICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgfSwKICAgICAgICBdCgogICAgICAgICMgaWYg
            J01PTklUT1JJTkdfRU5EUE9JTlQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAnbW9uaXRvcmluZy1l
            bmRwb2ludCcgdG8gQ2x1c3RlcgogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFw
            WyJNT05JVE9SSU5HX0VORFBPSU5UIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsiTU9O
            SVRPUklOR19FTkRQT0lOVF9QT1JUIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291dC5h
            cHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAibW9u
            aXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiQ2VwaENsdXN0
            ZXIiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAg
            ICAiTW9uaXRvcmluZ0VuZHBvaW50Ijogc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5U
            Il0sCiAgICAgICAgICAgICAgICAgICAgICAgICJNb25pdG9yaW5nUG9ydCI6IHNlbGYub3V0X21h
            cFsiTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUIl0sCiAgICAgICAgICAgICAgICAgICAgfSwKICAg
            ICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQoKICAgICAgICAjIGlmICdDU0lfUkJEX05PREVf
            U0VDUkVUJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ3Jvb2stY3NpLXJiZC1wcm92aXNpb25lcicg
            U2VjcmV0CiAgICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9E
            RV9TRUNSRVQiXQogICAgICAgICAgICBhbmQgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VD
            UkVUX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAg
            ICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9vay17c2VsZi5vdXRf
            bWFwWydDU0lfUkJEX05PREVfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICJr
            aW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAg
            ICAgICAgICAgICAgICJ1c2VySUQiOiBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9TRUNSRVRf
            TkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNlbGYub3V0X21hcFsi
            Q1NJX1JCRF9OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAg
            ICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdDU0lfUkJEX1BST1ZJU0lPTkVSX1NF
            Q1JFVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdyb29rLWNzaS1yYmQtcHJvdmlzaW9uZXInIFNl
            Y3JldAogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJ
            U0lPTkVSX1NFQ1JFVCJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9SQkRfUFJP
            VklTSU9ORVJfU0VDUkVUX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFw
            cGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9v
            ay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAg
            ICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJk
            YXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJD
            U0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAg
            ICJ1c2VyS2V5Ijogc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAog
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAg
            ICAgICAjIGlmICdDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCcgZXhpc3RzLCB0aGVuIG9u
            bHkgYWRkICdyb29rLWNzaS1jZXBoZnMtcHJvdmlzaW9uZXInIFNlY3JldAogICAgICAgIGlmICgK
            ICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJd
            CiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VD
            UkVUX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAg
            ICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9vay17c2VsZi5vdXRf
            bWFwWydDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAg
            ICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewog
            ICAgICAgICAgICAgICAgICAgICAgICAiYWRtaW5JRCI6IHNlbGYub3V0X21hcFsiQ1NJX0NFUEhG
            U19QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAiYWRt
            aW5LZXkiOiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUIl0sCiAg
            ICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAg
            ICAgICMgaWYgJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAn
            cm9vay1jc2ktY2VwaGZzLW5vZGUnIFNlY3JldAogICAgICAgIGlmICgKICAgICAgICAgICAgc2Vs
            Zi5vdXRfbWFwWyJDU0lfQ0VQSEZTX05PREVfU0VDUkVUIl0KICAgICAgICAgICAgYW5kIHNlbGYu
            b3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FIl0KICAgICAgICApOgogICAgICAg
            ICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAg
            ICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9O
            QU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAg
            ICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAiYWRtaW5JRCI6IHNl
            bGYub3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FIl0sCiAgICAgICAgICAgICAg
            ICAgICAgICAgICJhZG1pbktleSI6IHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JF
            VCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAg
            ICkKICAgICAgICAjIGlmICdST09LX0VYVEVSTkFMX0RBU0hCT0FSRF9MSU5LJyBleGlzdHMsIHRo
            ZW4gb25seSBhZGQgJ3Jvb2stY2VwaC1kYXNoYm9hcmQtbGluaycgU2VjcmV0CiAgICAgICAgaWYg
            c2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0RBU0hCT0FSRF9MSU5LIl06CiAgICAgICAgICAg
            IGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAi
            bmFtZSI6ICJyb29rLWNlcGgtZGFzaGJvYXJkLWxpbmsiLAogICAgICAgICAgICAgICAgICAgICJr
            aW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAg
            ICAgICAgICAgICAgICJ1c2VySUQiOiAiY2VwaC1kYXNoYm9hcmQtbGluayIsCiAgICAgICAgICAg
            ICAgICAgICAgICAgICJ1c2VyS2V5Ijogc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0RBU0hC
            T0FSRF9MSU5LIl0sCiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAg
            ICAgICAgICAgKQogICAgICAgICMgaWYgJ1JBRE9TX05BTUVTUEFDRScgZXhpc3RzLCB0aGVuIG9u
            bHkgYWRkIHRoZSAiUkFET1NfTkFNRVNQQUNFIiBuYW1lc3BhY2UKICAgICAgICBpZiAoCiAgICAg
            ICAgICAgIHNlbGYub3V0X21hcFsiUkFET1NfTkFNRVNQQUNFIl0KICAgICAgICAgICAgYW5kIHNl
            bGYub3V0X21hcFsiUkVTVFJJQ1RFRF9BVVRIX1BFUk1JU1NJT04iXQogICAgICAgICAgICBhbmQg
            bm90IHNlbGYub3V0X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdCiAgICAgICAgKToK
            ICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAg
            ICAgICAgICAgICJuYW1lIjogInJhZG9zLW5hbWVzcGFjZSIsCiAgICAgICAgICAgICAgICAgICAg
            ImtpbmQiOiAiQ2VwaEJsb2NrUG9vbFJhZG9zTmFtZXNwYWNlIiwKICAgICAgICAgICAgICAgICAg
            ICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgInJhZG9zTmFtZXNwYWNlTmFtZSI6
            IHNlbGYub3V0X21hcFsiUkFET1NfTkFNRVNQQUNFIl0sCiAgICAgICAgICAgICAgICAgICAgICAg
            ICJwb29sIjogc2VsZi5vdXRfbWFwWyJSQkRfUE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAg
            ICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICAgICBqc29uX291
            dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAi
            Y2VwaC1yYmQtcmFkb3MtbmFtZXNwYWNlIiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJT
            dG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAg
            ICAgICAgICAgICAicG9vbCI6IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAogICAgICAg
            ICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1u
            YW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05B
            TUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRy
            b2xsZXItZXhwYW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRf
            UFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3Np
            LnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0
            X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICB9
            LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgZWxzZToKICAgICAgICAg
            ICAgaWYgc2VsZi5vdXRfbWFwWyJSQkRfTUVUQURBVEFfRUNfUE9PTF9OQU1FIl06CiAgICAgICAg
            ICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICAgICAgewogICAgICAgICAg
            ICAgICAgICAgICAgICAibmFtZSI6ICJjZXBoLXJiZCIsCiAgICAgICAgICAgICAgICAgICAgICAg
            ICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAgICAgICAgICAgICJkYXRhIjog
            ewogICAgICAgICAgICAgICAgICAgICAgICAgICAgImRhdGFQb29sIjogc2VsZi5vdXRfbWFwWyJS
            QkRfUE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYu
            b3V0X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdLAogICAgICAgICAgICAgICAgICAg
            ICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNyZXQtbmFtZSI6IGYi
            cm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwK
            ICAgICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vY29udHJvbGxl
            ci1leHBhbmQtc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9QUk9W
            SVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiY3Np
            LnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0
            X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAg
            ICAgfSwKICAgICAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICAgICApCiAgICAgICAgICAg
            IGVsc2U6CiAgICAgICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICAg
            ICAgewogICAgICAgICAgICAgICAgICAgICAgICAibmFtZSI6ICJjZXBoLXJiZCIsCiAgICAgICAg
            ICAgICAgICAgICAgICAgICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAgICAg
            ICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAgICAgInBvb2wiOiBzZWxm
            Lm91dF9tYXBbIlJCRF9QT09MX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJj
            c2kuc3RvcmFnZS5rOHMuaW8vcHJvdmlzaW9uZXItc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYu
            b3V0X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAg
            ICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNl
            Y3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VD
            UkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4
            cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9S
            QkRfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgICAgIH0sCiAgICAg
            ICAgICAgICAgICAgICAgfQogICAgICAgICAgICAgICAgKQoKICAgICAgICAjIGlmICdDRVBIRlNf
            RlNfTkFNRScgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdjZXBoZnMnIFN0b3JhZ2VDbGFzcwogICAg
            ICAgIGlmIHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXToKICAgICAgICAgICAganNvbl9v
            dXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjog
            ImNlcGhmcyIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwKICAg
            ICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImZzTmFt
            ZSI6IHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAg
            ICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIkNFUEhGU19QT09MX05BTUUiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNyZXQtbmFtZSI6
            IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1F
            J119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250cm9s
            bGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfQ0VQSEZT
            X1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNz
            aS5zdG9yYWdlLms4cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91
            dF9tYXBbJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAg
            ICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICMgaWYgJ1JHV19F
            TkRQT0lOVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdjZXBoLXJndycgU3RvcmFnZUNsYXNzCiAg
            ICAgICAgaWYgc2VsZi5vdXRfbWFwWyJSR1dfRU5EUE9JTlQiXToKICAgICAgICAgICAganNvbl9v
            dXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjog
            ImNlcGgtcmd3IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAog
            ICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAiZW5k
            cG9pbnQiOiBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdLAogICAgICAgICAgICAgICAgICAg
            ICAgICAicG9vbFByZWZpeCI6IHNlbGYub3V0X21hcFsiUkdXX1BPT0xfUFJFRklYIl0sCiAgICAg
            ICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAg
            ICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAg
            ICAgIm5hbWUiOiAicmd3LWFkbWluLW9wcy11c2VyIiwKICAgICAgICAgICAgICAgICAgICAia2lu
            ZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAg
            ICAgICAgICAgICAiYWNjZXNzS2V5Ijogc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJf
            QUNDRVNTX0tFWSJdLAogICAgICAgICAgICAgICAgICAgICAgICAic2VjcmV0S2V5Ijogc2VsZi5v
            dXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJfU0VDUkVUX0tFWSJdLAogICAgICAgICAgICAgICAg
            ICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdSR1df
            VExTX0NFUlQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCB0aGUgImNlcGgtcmd3LXRscy1jZXJ0IiBz
            ZWNyZXQKICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VSVCJdOgogICAgICAgICAg
            ICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAg
            Im5hbWUiOiAiY2VwaC1yZ3ctdGxzLWNlcnQiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjog
            IlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAg
            ICAgICAgICJjZXJ0Ijogc2VsZi5vdXRfbWFwWyJSR1dfVExTX0NFUlQiXSwKICAgICAgICAgICAg
            ICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCgogICAgICAgIHJldHVy
            biBqc29uLmR1bXBzKGpzb25fb3V0KSArIExJTkVTRVAKCiAgICBkZWYgdXBncmFkZV91c2Vyc19w
            ZXJtaXNzaW9ucyhzZWxmKToKICAgICAgICB1c2VycyA9IFsKICAgICAgICAgICAgImNsaWVudC5j
            c2ktY2VwaGZzLW5vZGUiLAogICAgICAgICAgICAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9u
            ZXIiLAogICAgICAgICAgICAiY2xpZW50LmNzaS1yYmQtbm9kZSIsCiAgICAgICAgICAgICJjbGll
            bnQuY3NpLXJiZC1wcm92aXNpb25lciIsCiAgICAgICAgICAgICJjbGllbnQuaGVhbHRoY2hlY2tl
            ciIsCiAgICAgICAgXQogICAgICAgIGlmIHNlbGYucnVuX2FzX3VzZXIgIT0gIiIgYW5kIHNlbGYu
            cnVuX2FzX3VzZXIgbm90IGluIHVzZXJzOgogICAgICAgICAgICB1c2Vycy5hcHBlbmQoc2VsZi5y
            dW5fYXNfdXNlcikKICAgICAgICBmb3IgdXNlciBpbiB1c2VyczoKICAgICAgICAgICAgc2VsZi51
            cGdyYWRlX3VzZXJfcGVybWlzc2lvbnModXNlcikKCiAgICBkZWYgZ2V0X3Jnd19wb29sX25hbWVf
            ZHVyaW5nX3VwZ3JhZGUoc2VsZiwgdXNlciwgY2Fwcyk6CiAgICAgICAgaWYgdXNlciA9PSAiY2xp
            ZW50LmhlYWx0aGNoZWNrZXIiOgogICAgICAgICAgICAjIHdoZW4gYWRtaW4gaGFzIG5vdCBwcm92
            aWRlZCByZ3cgcG9vbCBuYW1lIGR1cmluZyB1cGdyYWRlLAogICAgICAgICAgICAjIGdldCB0aGUg
            cmd3IHBvb2wgbmFtZSBmcm9tIGNsaWVudC5oZWFsdGhjaGVja2VyIHVzZXIgd2hpY2ggd2FzIHVz
            ZWQgZHVyaW5nIGNvbm5lY3Rpb24KICAgICAgICAgICAgaWYgbm90IHNlbGYuX2FyZ19wYXJzZXIu
            cmd3X3Bvb2xfcHJlZml4OgogICAgICAgICAgICAgICAgIyBUbyBnZXQgdmFsdWUgJ2RlZmF1bHQn
            IHdoaWNoIGlzIHJndyBwb29sIG5hbWUgZnJvbSAnYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YScKICAgICAgICAgICAgICAgIHBhdHRlcm4gPSByInBvb2w9KC4qPylcLnJnd1wubWV0YSIK
            ICAgICAgICAgICAgICAgIG1hdGNoID0gcmUuc2VhcmNoKHBhdHRlcm4sIGNhcHMpCiAgICAgICAg
            ICAgICAgICBpZiBtYXRjaDoKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJn
            d19wb29sX3ByZWZpeCA9IG1hdGNoLmdyb3VwKDEpCiAgICAgICAgICAgICAgICBlbHNlOgogICAg
            ICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAg
            ICAgICAgICAgICAgICAgICJmYWlsZWQgdG8gZ2V0IHJndyBwb29sIG5hbWUgZm9yIHVwZ3JhZGUi
            CiAgICAgICAgICAgICAgICAgICAgKQoKICAgIGRlZiB1cGdyYWRlX3VzZXJfcGVybWlzc2lvbnMo
            c2VsZiwgdXNlcik6CiAgICAgICAgIyBjaGVjayB3aGV0aGVyIHRoZSBnaXZlbiB1c2VyIGV4aXN0
            cyBvciBub3QKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4IjogImF1dGggZ2V0IiwgImVudGl0
            eSI6IGYie3VzZXJ9IiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICByZXRfdmFsLCBqc29uX291
            dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAg
            aWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgcHJpbnQo
            ZiJ1c2VyIHt1c2VyfSBub3QgZm91bmQgZm9yIHVwZ3JhZGluZy4iKQogICAgICAgICAgICByZXR1
            cm4KICAgICAgICBleGlzdGluZ19jYXBzID0ganNvbl9vdXRbMF1bImNhcHMiXQogICAgICAgIHNl
            bGYuZ2V0X3Jnd19wb29sX25hbWVfZHVyaW5nX3VwZ3JhZGUodXNlciwgc3RyKGV4aXN0aW5nX2Nh
            cHMpKQogICAgICAgIG5ld19jYXAsIF8gPSBzZWxmLmdldF9jYXBzX2FuZF9lbnRpdHkodXNlcikK
            ICAgICAgICBjYXBfa2V5cyA9IFsibW9uIiwgIm1nciIsICJvc2QiLCAibWRzIl0KICAgICAgICBj
            YXBzID0gW10KICAgICAgICBmb3IgZWFjaENhcCBpbiBjYXBfa2V5czoKICAgICAgICAgICAgY3Vy
            X2NhcF92YWx1ZXMgPSBleGlzdGluZ19jYXBzLmdldChlYWNoQ2FwLCAiIikKICAgICAgICAgICAg
            bmV3X2NhcF92YWx1ZXMgPSBuZXdfY2FwLmdldChlYWNoQ2FwLCAiIikKICAgICAgICAgICAgY3Vy
            X2NhcF9wZXJtX2xpc3QgPSBbCiAgICAgICAgICAgICAgICB4LnN0cmlwKCkgZm9yIHggaW4gY3Vy
            X2NhcF92YWx1ZXMuc3BsaXQoIiwiKSBpZiB4LnN0cmlwKCkKICAgICAgICAgICAgXQogICAgICAg
            ICAgICBuZXdfY2FwX3Blcm1fbGlzdCA9IFsKICAgICAgICAgICAgICAgIHguc3RyaXAoKSBmb3Ig
            eCBpbiBuZXdfY2FwX3ZhbHVlcy5zcGxpdCgiLCIpIGlmIHguc3RyaXAoKQogICAgICAgICAgICBd
            CiAgICAgICAgICAgICMgYXBwZW5kIG5ld19jYXBfbGlzdCB0byBjdXJfY2FwX2xpc3QgdG8gbWFp
            bnRhaW4gdGhlIG9yZGVyIG9mIGNhcHMKICAgICAgICAgICAgY3VyX2NhcF9wZXJtX2xpc3QuZXh0
            ZW5kKG5ld19jYXBfcGVybV9saXN0KQogICAgICAgICAgICAjIGVsaW1pbmF0ZSBkdXBsaWNhdGVz
            IHdpdGhvdXQgdXNpbmcgJ3NldCcKICAgICAgICAgICAgIyBzZXQgcmUtb3JkZXJzIGl0ZW1zIGlu
            IHRoZSBsaXN0IGFuZCB3ZSBoYXZlIHRvIGtlZXAgdGhlIG9yZGVyCiAgICAgICAgICAgIG5ld19j
            YXBfbGlzdCA9IFtdCiAgICAgICAgICAgIFtuZXdfY2FwX2xpc3QuYXBwZW5kKHgpIGZvciB4IGlu
            IGN1cl9jYXBfcGVybV9saXN0IGlmIHggbm90IGluIG5ld19jYXBfbGlzdF0KICAgICAgICAgICAg
            ZXhpc3RpbmdfY2Fwc1tlYWNoQ2FwXSA9ICIsICIuam9pbihuZXdfY2FwX2xpc3QpCiAgICAgICAg
            ICAgIGlmIGV4aXN0aW5nX2NhcHNbZWFjaENhcF06CiAgICAgICAgICAgICAgICBjYXBzLmFwcGVu
            ZChlYWNoQ2FwKQogICAgICAgICAgICAgICAgY2Fwcy5hcHBlbmQoZXhpc3RpbmdfY2Fwc1tlYWNo
            Q2FwXSkKICAgICAgICBjbWRfanNvbiA9IHsKICAgICAgICAgICAgInByZWZpeCI6ICJhdXRoIGNh
            cHMiLAogICAgICAgICAgICAiZW50aXR5IjogdXNlciwKICAgICAgICAgICAgImNhcHMiOiBjYXBz
            LAogICAgICAgICAgICAiZm9ybWF0IjogImpzb24iLAogICAgICAgIH0KICAgICAgICByZXRfdmFs
            LCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24p
            CiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiInYXV0aCBjYXBzIHt1c2VyfScgY29tbWFu
            ZCBmYWlsZWQuXG4gRXJyb3I6IHtlcnJfbXNnfSIKICAgICAgICAgICAgKQogICAgICAgIHByaW50
            KGYiVXBkYXRlZCB1c2VyIHt1c2VyfSBzdWNjZXNzZnVsbHkuIikKCiAgICBkZWYgbWFpbihzZWxm
            KToKICAgICAgICBnZW5lcmF0ZWRfb3V0cHV0ID0gIiIKICAgICAgICBpZiBzZWxmLl9hcmdfcGFy
            c2VyLnVwZ3JhZGU6CiAgICAgICAgICAgIHNlbGYudXBncmFkZV91c2Vyc19wZXJtaXNzaW9ucygp
            CiAgICAgICAgZWxpZiBzZWxmLl9hcmdfcGFyc2VyLmZvcm1hdCA9PSAianNvbiI6CiAgICAgICAg
            ICAgIGdlbmVyYXRlZF9vdXRwdXQgPSBzZWxmLmdlbl9qc29uX291dCgpCiAgICAgICAgZWxpZiBz
            ZWxmLl9hcmdfcGFyc2VyLmZvcm1hdCA9PSAiYmFzaCI6CiAgICAgICAgICAgIGdlbmVyYXRlZF9v
            dXRwdXQgPSBzZWxmLmdlbl9zaGVsbF9vdXQoKQogICAgICAgIGVsc2U6CiAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlVuc3VwcG9y
            dGVkIGZvcm1hdDoge3NlbGYuX2FyZ19wYXJzZXIuZm9ybWF0fSIKICAgICAgICAgICAgKQogICAg
            ICAgIHByaW50KGdlbmVyYXRlZF9vdXRwdXQpCiAgICAgICAgaWYgc2VsZi5vdXRwdXRfZmlsZSBh
            bmQgZ2VuZXJhdGVkX291dHB1dDoKICAgICAgICAgICAgZk91dCA9IG9wZW4oc2VsZi5vdXRwdXRf
            ZmlsZSwgbW9kZT0idyIsIGVuY29kaW5nPSJVVEYtOCIpCiAgICAgICAgICAgIGZPdXQud3JpdGUo
            Z2VuZXJhdGVkX291dHB1dCkKICAgICAgICAgICAgZk91dC5jbG9zZSgpCgoKIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyBNQUlOICMjIyMjIyMjIyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMKaWYgX19uYW1lX18gPT0gIl9fbWFpbl9fIjoKICAgIHJqT2Jq
            ID0gUmFkb3NKU09OKCkKICAgIHRyeToKICAgICAgICByak9iai5tYWluKCkKICAgIGV4Y2VwdCBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uIGFzIGVycjoKICAgICAgICBwcmludChmIkV4ZWN1dGlv
            biBGYWlsZWQ6IHtlcnJ9IikKICAgICAgICByYWlzZSBlcnIKICAgIGV4Y2VwdCBLZXlFcnJvciBh
            cyBrRXJyOgogICAgICAgIHByaW50KGYiS2V5RXJyb3I6IHtrRXJyfSIpCiAgICBleGNlcHQgT1NF
            cnJvciBhcyBvc0VycjoKICAgICAgICBwcmludChmIkVycm9yIHdoaWxlIHRyeWluZyB0byBvdXRw
            dXQgdGhlIGRhdGE6IHtvc0Vycn0iKQogICAgZmluYWxseToKICAgICAgICByak9iai5zaHV0ZG93
            bigpCg==
          external.features.ocs.openshift.io/supported-platforms: '["BareMetal", "None",
            "VSphere", "OpenStack", "oVirt"]'
          external.features.ocs.openshift.io/validation: '{"secrets":["rook-ceph-operator-creds",
            "rook-csi-rbd-node", "rook-csi-rbd-provisioner"], "configMaps": ["rook-ceph-mon-endpoints",
            "rook-ceph-mon"], "storageClasses": ["ceph-rbd"], "cephClusters": ["monitoring-endpoint"]}'
          features.ocs.openshift.io/disabled: '["ss-list", "install-wizard", "block-pool",
            "mcg-resource", "odf-dashboard", "common", "storage-provider", "storage-provisioner",
            "dashboard-resources", "csv-actions", "inventory-item", "alert-actions"]'
          features.ocs.openshift.io/enabled: '["kms", "arbiter", "flexible-scaling",
            "multus", "pool-management", "namespace-store", "mcg-standalone", "taint-nodes",
            "vault-sa-kms", "hpcs-kms"]'
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operatorframework.io/initialization-resource: "\n    {\n        \"apiVersion\":
            \"ocs.openshift.io/v1\",\n        \"kind\": \"StorageCluster\",\n        \"metadata\":
            {\n            \"name\": \"example-storagecluster\",\n            \"namespace\":
            \"openshift-storage\"\n        },\n        \"spec\": {\n            \"manageNodes\":
            false,\n            \"monPVCTemplate\": {\n                \"spec\": {\n
            \                   \"accessModes\": [\n                        \"ReadWriteOnce\"\n
            \                   ],\n                    \"resources\": {\n                        \"requests\":
            {\n                            \"storage\": \"10Gi\"\n                        }\n
            \                   },\n                    \"storageClassName\": \"gp2-csi\"\n
            \               }\n            },\n            \"storageDeviceSets\":
            [\n                {\n                    \"count\": 3,\n                    \"dataPVCTemplate\":
            {\n                        \"spec\": {\n                            \"accessModes\":
            [\n                                \"ReadWriteOnce\"\n                            ],\n
            \                           \"resources\": {\n                                \"requests\":
            {\n                                    \"storage\": \"1Ti\"\n                                }\n
            \                           },\n                            \"storageClassName\":
            \"gp2-csi\",\n                            \"volumeMode\": \"Block\"\n
            \                       }\n                    },\n                    \"name\":
            \"example-deviceset\",\n                    \"placement\": {},\n                    \"portable\":
            true,\n                    \"resources\": {}\n                }\n            ]\n
            \       }\n    }\n\t"
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:03e50284f7fe090c57eddf90feea8ff48d78aff8ce95d1b297aa91e71c3ee577
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.4
          operators.operatorframework.io/internal-objects: '["ocsinitializations.ocs.openshift.io","storageclassrequests.ocs.openshift.io","storageconsumers.ocs.openshift.io","storageprofiles.ocs.openshift.io","cephclusters.ceph.rook.io","cephobjectstores.ceph.rook.io","cephobjectstoreusers.ceph.rook.io","cephnfses.ceph.rook.io","cephclients.ceph.rook.io","cephfilesystems.ceph.rook.io","cephfilesystemmirrors.ceph.rook.io","cephrbdmirrors.ceph.rook.io","cephobjectrealms.ceph.rook.io","cephobjectzonegroups.ceph.rook.io","cephobjectzones.ceph.rook.io","cephbucketnotifications.ceph.rook.io","cephbuckettopics.ceph.rook.io","cephfilesystemsubvolumegroups.ceph.rook.io","cephblockpoolradosnamespaces.ceph.rook.io","cephcosidrivers.ceph.rook.io"]'
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/ocs-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OCSInitialization represents the initial data to be created
              when the operator is installed.
            displayName: OCS Initialization
            kind: OCSInitialization
            name: ocsinitializations.ocs.openshift.io
            version: v1
          - description: StorageClassRequest is the Schema for the StorageClassRequests
              API
            displayName: Storage Class Request
            kind: StorageClassRequest
            name: storageclassrequests.ocs.openshift.io
            version: v1alpha1
          - description: StorageCluster represents a cluster including Ceph Cluster,
              NooBaa and all the storage and compute resources required.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.ocs.openshift.io
            version: v1
          - description: StorageConsumer is the Schema for the storageconsumers API
            displayName: Storage Consumer
            kind: StorageConsumer
            name: storageconsumers.ocs.openshift.io
            version: v1alpha1
          - kind: StorageProfile
            name: storageprofiles.ocs.openshift.io
            version: v1
          - description: Represents a Ceph cluster.
            displayName: Ceph Cluster
            kind: CephCluster
            name: cephclusters.ceph.rook.io
            version: v1
          - description: Represents a Ceph Block Pool.
            displayName: Block Pools
            kind: CephBlockPool
            name: cephblockpools.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store.
            displayName: Ceph Object Store
            kind: CephObjectStore
            name: cephobjectstores.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store User.
            displayName: Ceph Object Store User
            kind: CephObjectStoreUser
            name: cephobjectstoreusers.ceph.rook.io
            version: v1
          - description: Represents a cluster of Ceph NFS ganesha gateways.
            displayName: Ceph NFS
            kind: CephNFS
            name: cephnfses.ceph.rook.io
            version: v1
          - description: Represents a Ceph User.
            displayName: Ceph Client
            kind: CephClient
            name: cephclients.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem.
            displayName: Ceph Filesystem
            kind: CephFilesystem
            name: cephfilesystems.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem Mirror.
            displayName: Ceph Filesystem Mirror
            kind: CephFilesystemMirror
            name: cephfilesystemmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph RBD Mirror.
            displayName: Ceph RBD Mirror
            kind: CephRBDMirror
            name: cephrbdmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Realm.
            displayName: Ceph Object Store Realm
            kind: CephObjectRealm
            name: cephobjectrealms.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone Group.
            displayName: Ceph Object Store Zone Group
            kind: CephObjectZoneGroup
            name: cephobjectzonegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone.
            displayName: Ceph Object Store Zone
            kind: CephObjectZone
            name: cephobjectzones.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Notification.
            displayName: Ceph Bucket Notification
            kind: CephBucketNotification
            name: cephbucketnotifications.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Topic.
            displayName: Ceph Bucket Topic
            kind: CephBucketTopic
            name: cephbuckettopics.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem SubVolumeGroup.
            displayName: Ceph Filesystem SubVolumeGroup
            kind: CephFilesystemSubVolumeGroup
            name: cephfilesystemsubvolumegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph BlockPool Rados Namespace.
            displayName: Ceph BlockPool Rados Namespace
            kind: CephBlockPoolRadosNamespace
            name: cephblockpoolradosnamespaces.ceph.rook.io
            version: v1
          - description: Represents a Ceph COSI Driver.
            displayName: Ceph COSI Driver
            kind: CephCOSIDriver
            name: cephcosidrivers.ceph.rook.io
            version: v1
        description: |
          **Red Hat OpenShift Container Storage** deploys three operators.

          ### OpenShift Container Storage operator

          The OpenShift Container Storage operator is the primary operator for OpenShift Container Storage. It serves to facilitate the other operators in OpenShift Container Storage by performing administrative tasks outside their scope as well as watching and configuring their CustomResources.

          ### Rook

          [Rook][1] deploys and manages Ceph on OpenShift, which provides block and file storage.

          # Core Capabilities

          * **Self-managing service:** No matter which supported storage technologies you choose, OpenShift Container Storage ensures that resources can be deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Container Storage you can either build dedicated storage clusters or hyper-converged clusters where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Container Storage:** OpenShift Container Storage integrates Ceph with multiple storage presentations including object storage (compatible with S3), block storage, and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Container Storage efficiently distributes and replicates your data across your cluster to minimize the risk of data loss. With snapshots, cloning, and versioning, no more losing sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data storage for your workloads, using a single namespace.

          [1]: https://rook.io
        displayName: OpenShift Container Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - rook
        - ceph
        - block storage
        - shared filesystem
        - object storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ocs-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:1711a0aabd28937aac1424adebe96619a48b91f7f888f0588117f80891001da4
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:f37e2c611f4cda32172d4e8938938e488076fd8a8f19c252c56a80c48044bc02
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:028120f2212738917f49e0330b60be5f79a208ce2974bea22b71f533aa6ce3be
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:f75a12fc0c449008028d7fc8e18ca7c1afc37ba585885ac754fed860047794a0
        - registry.redhat.io/odf4/ocs-metrics-exporter-rhel9@sha256:c58c67263c08d84a68c8e45cfd77d59a35aee2151bf79b44377016ca0fb429ee
        - registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:9a0008f662b6afa9e2677a8a83bd662d806ad6f4d03fb49bccb78ed487ce4c7a
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:645d4fd09628f7ddcfc765e4ddacaa79fa6a7c35c03292ca2f6e317d48765d6b
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:ef4276b80cacae825d4831de88d3fe6ba7cc324464087b7b28734bd2b8de3315
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:0dc0733df67bb49886ac9715b81bd4ec4a36330fe1afdde657587e1883f5ca7a
        - registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:03e50284f7fe090c57eddf90feea8ff48d78aff8ce95d1b297aa91e71c3ee577
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:ffcc0806ec2afb8b40d75c70ef2a6d98bb887879f5b3f573e900861316ce9e41
        - registry.redhat.io/odf4/ocs-rhel9-operator@sha256:b1a9492bd9b314f86b4e192cf2c83144885ff863d9932e386a41691f50be6193
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:f8bfbf64b4bd030474e239aed5e03af9dfe920eeae444c5cc8ac3dcfa640fcc1
        version: 4.15.20-rhodf
      entries:
      - name: ocs-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: ocs-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: ocs-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: ocs-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: ocs-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: ocs-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: ocs-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: ocs-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: ocs-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: ocs-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: ocs-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: ocs-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: ocs-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: ocs-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: ocs-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: ocs-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: ocs-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: ocs-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: ocs-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: ocs-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: ocs-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |2-

            [
                {
                    "apiVersion": "ocs.openshift.io/v1",
                    "kind": "StorageCluster",
                    "metadata": {
                        "name": "example-storagecluster",
                        "namespace": "openshift-storage"
                    },
                    "spec": {
                        "manageNodes": false,
                        "monPVCTemplate": {
                            "spec": {
                                "accessModes": [
                                    "ReadWriteOnce"
                                ],
                                "resources": {
                                    "requests": {
                                        "storage": "10Gi"
                                    }
                                },
                                "storageClassName": "gp2-csi"
                            }
                        },
                        "storageDeviceSets": [
                            {
                                "count": 3,
                                "dataPVCTemplate": {
                                    "spec": {
                                        "accessModes": [
                                            "ReadWriteOnce"
                                        ],
                                        "resources": {
                                            "requests": {
                                                "storage": "1Ti"
                                            }
                                        },
                                        "storageClassName": "gp2-csi",
                                        "volumeMode": "Block"
                                    }
                                },
                                "name": "example-deviceset",
                                "placement": {},
                                "portable": true,
                                "resources": {}
                            }
                        ]
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Storage
          containerImage: registry.redhat.io/odf4/ocs-rhel9-operator@sha256:3ef1c2679e9229d51ebc04bd2f2b22ce9b11e8f7fdd7bddee69630fe5b1de47c
          description: Red Hat OpenShift Container Storage provides hyperconverged
            storage for applications within an OpenShift cluster.
          external.features.ocs.openshift.io/supported-platforms: '["BareMetal", "None",
            "VSphere", "OpenStack", "oVirt"]'
          external.features.ocs.openshift.io/validation: '{"secrets":["rook-ceph-operator-creds",
            "rook-csi-rbd-node", "rook-csi-rbd-provisioner"], "configMaps": ["rook-ceph-mon-endpoints",
            "rook-ceph-mon"], "storageClasses": ["ceph-rbd"], "cephClusters": ["monitoring-endpoint"]}'
          features.ocs.openshift.io/disabled: '["ss-list", "install-wizard", "block-pool",
            "mcg-resource", "odf-dashboard", "common", "storage-provider", "storage-provisioner",
            "dashboard-resources", "csv-actions", "inventory-item", "alert-actions"]'
          features.ocs.openshift.io/enabled: '["kms", "arbiter", "flexible-scaling",
            "multus", "pool-management", "namespace-store", "mcg-standalone", "taint-nodes",
            "vault-sa-kms", "hpcs-kms"]'
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operatorframework.io/initialization-resource: "\n    {\n        \"apiVersion\":
            \"ocs.openshift.io/v1\",\n        \"kind\": \"StorageCluster\",\n        \"metadata\":
            {\n            \"name\": \"example-storagecluster\",\n            \"namespace\":
            \"openshift-storage\"\n        },\n        \"spec\": {\n            \"manageNodes\":
            false,\n            \"monPVCTemplate\": {\n                \"spec\": {\n
            \                   \"accessModes\": [\n                        \"ReadWriteOnce\"\n
            \                   ],\n                    \"resources\": {\n                        \"requests\":
            {\n                            \"storage\": \"10Gi\"\n                        }\n
            \                   },\n                    \"storageClassName\": \"gp2-csi\"\n
            \               }\n            },\n            \"storageDeviceSets\":
            [\n                {\n                    \"count\": 3,\n                    \"dataPVCTemplate\":
            {\n                        \"spec\": {\n                            \"accessModes\":
            [\n                                \"ReadWriteOnce\"\n                            ],\n
            \                           \"resources\": {\n                                \"requests\":
            {\n                                    \"storage\": \"1Ti\"\n                                }\n
            \                           },\n                            \"storageClassName\":
            \"gp2-csi\",\n                            \"volumeMode\": \"Block\"\n
            \                       }\n                    },\n                    \"name\":
            \"example-deviceset\",\n                    \"placement\": {},\n                    \"portable\":
            true,\n                    \"resources\": {}\n                }\n            ]\n
            \       }\n    }\n\t"
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:b7758b197cf423573de96bdca0cb3045852b20f34fc9c17ad1632b9c3a3cf3d0
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.4
          operators.operatorframework.io/internal-objects: '["ocsinitializations.ocs.openshift.io","storageconsumers.ocs.openshift.io","storageprofiles.ocs.openshift.io","storagerequests.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/ocs-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OCSInitialization represents the initial data to be created
              when the operator is installed.
            displayName: OCS Initialization
            kind: OCSInitialization
            name: ocsinitializations.ocs.openshift.io
            version: v1
          - description: StorageCluster represents a cluster including Ceph Cluster,
              NooBaa and all the storage and compute resources required.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.ocs.openshift.io
            version: v1
          - description: StorageConsumer is the Schema for the storageconsumers API
            displayName: Storage Consumer
            kind: StorageConsumer
            name: storageconsumers.ocs.openshift.io
            version: v1alpha1
          - description: StorageProfile is the Schema for the storageprofiles API
            displayName: Storage Profile
            kind: StorageProfile
            name: storageprofiles.ocs.openshift.io
            version: v1
          - description: StorageRequest is the Schema for the StorageRequests API
            displayName: Storage Request
            kind: StorageRequest
            name: storagerequests.ocs.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Container Storage** deploys three operators.

          ### OpenShift Container Storage operator

          The OpenShift Container Storage operator is the primary operator for OpenShift Container Storage. It serves to facilitate the other operators in OpenShift Container Storage by performing administrative tasks outside their scope as well as watching and configuring their CustomResources.

          ### Rook

          [Rook][1] deploys and manages Ceph on OpenShift, which provides block and file storage.

          # Core Capabilities

          * **Self-managing service:** No matter which supported storage technologies you choose, OpenShift Container Storage ensures that resources can be deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Container Storage you can either build dedicated storage clusters or hyper-converged clusters where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Container Storage:** OpenShift Container Storage integrates Ceph with multiple storage presentations including object storage (compatible with S3), block storage, and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Container Storage efficiently distributes and replicates your data across your cluster to minimize the risk of data loss. With snapshots, cloning, and versioning, no more losing sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data storage for your workloads, using a single namespace.

          [1]: https://rook.io
        displayName: OpenShift Container Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - rook
        - ceph
        - block storage
        - shared filesystem
        - object storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ocs-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d
        - registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:b7758b197cf423573de96bdca0cb3045852b20f34fc9c17ad1632b9c3a3cf3d0
        - registry.redhat.io/odf4/ocs-rhel9-operator@sha256:3ef1c2679e9229d51ebc04bd2f2b22ce9b11e8f7fdd7bddee69630fe5b1de47c
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:70a3c6bf4bde84c7e43fd58c60abc5612d25c792042b902f6abee94bfaa63f80
        - registry.redhat.io/odf4/ocs-metrics-exporter-rhel9@sha256:3332a93e3ea2e83af15a7edab02a9e00d70e81e44338b63a0eacdea54142ad00
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:0084d9053ad2c74926ca02d2ede0461abd59227eee98d122c45e08a5ebaf629a
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:cf7f72469d25fb37301a8159406613673f8968c82af1035fd45a470e0ae8acfa
        version: 4.16.16-rhodf
      entries:
      - name: ocs-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: ocs-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: ocs-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: ocs-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: ocs-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: ocs-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: ocs-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: ocs-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: ocs-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: ocs-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: ocs-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: ocs-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: ocs-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: ocs-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: ocs-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: ocs-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: ocs-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: ocs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: serverless-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: serverless-operator.v1.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.36.0 <1.36.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:52ba6b1252b6f3d41e2a0b3b880cb768010ab9dd925ac34596f6adaedcbbe07c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-source-rhel8@sha256:88323da16360981f27b8e4686fbbb64053e4e0e3c79c65d695edb317b334e1e0
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:3fe326ad9a00b112fc888ffd345c8d5042f0a22cf32ea575931192c39ce1fece
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:259e5dab50d36cdbfc117a42325128c4e16f34c048ba64b36d4a89132ddb4c1a
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:4344716eb2940c6a3fa9a278250f936602bb1549fece6037912837719c178b24
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-transform-jsonata-rhel8@sha256:3af4c48030263eb5769e5f36a6dd13fd390d8566efaab37532f39e1df1c95c67
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sns-sink-rhel8@sha256:51b5e18e8d4c1ca81d7c5ddfea6352a3887f757acfcc53bd2f8bd3b01337ed42
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-ddb-streams-source-rhel8@sha256:a903480159ebf1d8652eeaa702d9885fb76bb106c02d75cec5c82c31d77d6410
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:e3ae1cc6012f19631c012a07b113fb015977d2ee4f9faab3207fbc2464c25aa7
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:596a7272d6de714d7dbc6e00e9e50a2872846c269547465252d953172ef0b009
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:f4d86b196457cdeebdc380b6f1b707c10f8dee789c7779c90ee1332e511af8c4
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:f2d112ba9fdfccea988589b99b463c8a7157475f7a3f03678e8999e65b678cc3
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-sink-rhel8@sha256:afa634e029bcf79a972f49372498d44b45bcfbdb6eb28ecaf4f0b0b2a0a1824b
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:dcfe5fbc57b9f8520adbff368c8bf29eb491c9c26d1a1737a1722cffea83eb1b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-source-rhel8@sha256:27e74287774cda208f667bc2a386debe6b6255e8991abfb37c959c8525f2af53
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:9b328fff0cc8545fb9bb1f24f11d78cfac3ee76dedc11fa4def4262d79a60e8c
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:82e69a1890042f5636ac7bb645b1aaeaba0a4894851eccc1fa1c63296918e31a
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:3d780ad1ab280b5092b90e277485d589da2a18e8b08090af7d95aecd17af545d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:631eb3a8dd9d83ceafa7d8a4da4b1e2d4f2740e6017c766828dcc7575cf1322f
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-log-sink-rhel8@sha256:30b8045ef5f925a36c58941386f1899bec125d005a08aeeeba5edd49b0859149
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:c7c2e738e0d9da1d2d1993f68938e72c4709b1f28a54fa92224ef6a0e456b0aa
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:a1834b620f875681388f0ae224dd0771773a8b61b1d8e5214a942f232854b87b
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:074ebfc4494581611647ba90049655a470ca43743610a27be77db66a8633bac6
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:429d5d9568cca31fbe83c663276a300286af4cdbf3d35685afdbe5f6f2f8c6e0
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:5352e4eb4e426a41521276decadf204b6d23b4762a298e4d3dc73969e3eddc68
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:bf82b37ab8fd29c7a7cd846008a16d14b4cfdc1c09df4f731d71be3957e90123
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:6b7316cce3432d391adc6ac38c92ff0ee86dd7ae0101957cc65cd338cc69e352
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:3b04e9d8cd4b5e43a07ece838d10478723a6c6529350f010550b4a0d1d0950d1
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:6210a946f373b9c36c9b31293debe6fed090008b8234c0ba17d37fa81646172e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:c39cb88afaddc295d93a3774b8a20d001b82392df540fb502ee02316d82460a9
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:0c2cdf917bbad1f5bca8a2a2ef12a96cfcd0c77c190cc35687eb57169b54ab6f
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:187fb9d64b53c55a029ffce3c0f9e7c3304683fa8a0e832cfb7134e8a3d0c237
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:d6f338e08fa937c8a4ccfec692424de4acded565f7ca3971ea507dffeb351503
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:f3d87a7fa5b5f2818effcac3304bd56eddcdb0db97a18ac0f35a06fc7e223e7d
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:4aeb68a22f63300e54391ec6b4cda8c4e665d904b5dcdd3c40155040525742a0
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:ed60e889a677fb139f2eb4bd9ab4fda6ecc39fd050ca2b6cc1c19717e428535a
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:77e4689b2ce7a713e905553e0e42b8e128045f40cb6c6d4ee082f126cbc46b00
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:4274004b6289fa9bea3fa6e56d64077a328a6b5e63a9f0e9c41f6d561f682bd7
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:a6f565d73375e827d830dca85838f4c4cb7f08b0c078acb0c0d567b25f0d256a
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-timer-source-rhel8@sha256:8fed5ef9a1f39a55077bb5026c8854d0ee7fc8d8ff5da277ce9c1bfc2ab2474b
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:78fee4c596f2848487ca704da3fa1074e3c9c99b3f96f66833aeba63836d7b93
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:a47b6b5237a87ccd90e5a49a53d06485fd55c87eebf12d01c335b87fc3a8c494
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-sink-rhel8@sha256:d9b944e64391a996104afbe8a21f66316a948e9fb7b91734e861847285be59c7
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:ee6533b71cf280877c9d7489dfd821c9209c6802bed7fa3fdb4f8f00fbeacb09
        version: 1.36.1
      entries:
      - name: serverless-operator.v1.36.1
        version: 1.36.1
      - name: serverless-operator.v1.36.0
        version: 1.36.0
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable
    - currentCSV: serverless-operator.v1.29.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2023-07-28T05:42:31Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.29.0 <1.29.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of a particular version of Knative
              Serving
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Eventing
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Kafka components
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:a3c5d5fd8014c432830be912aedebbc58f9c60a8cabd87ae6c42c7b8d27db8f9
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:a59ba5aeb8e66db3d8ba9074b8e553603c59e6b0e35b57b82c4ce3de589b94cc
        - registry.redhat.io/ocp-tools-4-tech-preview/source-to-image-rhel8@sha256:98d8cb3a255641ca6a1bce854e5e2460c20de9fb9b28e3cc67eb459f122873dd
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:56e48afb1dcbf4e266ace944105f3c45042550d8d4a0528c982225a793d26d56
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:31d45f10a8519addec9aac115c5edc8a2e14f9621c875d95f123d1291776ef4b
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:b780634fffe2257eb845f8c1d25498d9c654d0327c1f8841b40d90c176a2bfe0
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:04866e16881f64ff2b5edd379eab0731c9d66a47c25ebfe79c8d8ad33d8116e6
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:61fb98802490323bb03ff4a7060334d8fbda9e9f95cd6366b29b1b6494e4ef8b
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:3b7449287adf98f0cf0a1dc74e03b69218be3bcf5538327bc34ef776cc1c69c4
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:10788ca67809e5a2d76d6d72df7dd64ea4407773999707a7681e0cf1b5c4695f
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:216f58dc8c8166c0c69fc0312b667daad14cf39b7aeb7bc0fe1e313376f1a5fd
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:aea5f200a845e1d4dc718d76465fdbb1eeac2b20ecae224d5d20a98214ad0357
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:e5db11a0adb889f907cad797dad9cf2de09c433e5522fa333193cb35575802d8
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:c4cff5ad5b0ac79e71a16b88640a312ab320817041317f50caf575f7e12a864e
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:c2cb80bce34a74afa76151d2b44dc25ece73fb274cdfc6fdf243b1b0cb7d989f
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:3aab995872336e6ccb590ef6488c097cff38a91348055031145427a00aa6b732
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:b520305189c7567b2f8ac867b20468259fb22cd3edc5b4576af2bfcdb0c11855
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:8273842d8e1714fa03c7e0809924b1ffc5ef282f79b9606bd8ac88c03e7a16bd
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a10e76aa15707e708f82c8d9e8b5179bf67780d99f3c3e2e99b84ae2a0aa4ecb
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:6342358e6aed2339333ebf37fc223a41a0a3e199a0e4a840f3f66e5accf8f233
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:594c3ac24bf59802e2d3fa127e4051277bc07b6444f934f535fddeb668ccc144
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:4408d1fcf34c0ffe9327d4c9c0bc772afa969d831bc5de1097164d10b4e8d900
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:03a3d6f0314acf4da4f3d3d85f1e273a3ef25828d28f5b22f394441d55a645f3
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:7dd11d385b3d5a178566d901650ecd45e4af395faec2141d57b98af8fc1e5151
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:a4943c0563c45ceccfda0a01e26be3a01c1437c89d927714c577bf6060495958
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:fd86ea2a34de43014dea3ba2b8d705a3d94347d3a24a370a8bfd032ed239dc56
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:f37f494956adf0892ffca34d2a4c9b839ce28dc9f2779ca32e8801fb4f0e9105
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:557e73127326be051aa6257e7daeb9721e2af04d2f06cc26cfc7a1073666b989
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:fe916630003194dbcd3894dd880cad8b15a041a5e534cd8f76fde6296db00797
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:397eb425f383c1273c7976bea69a603530676a00a61f461ceb5f48d67901c4cd
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:bd134e64c156dc5c81d14522c9c9efcc1f56b2e972ac8ecb1bb1b130c76adc60
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:ab90ec9b3cd9203fdcdd7c43abf95f48c9d7479a9c7af479eb1e4b6599ba9642
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:c12c471902b8e2596588d9e553ad59b501d979dd04f301d45f7d45e794737a8c
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:de0634d14962bdb8701a8c29e80b7b70684824462bc5203e5fa7f139839cb73b
        - registry.redhat.io/rhel8/buildah@sha256:a1e5cc0fb334e333e5eab69689223e8bd1f0c060810d260603b26cf8c0da2023
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:0229a0a4f026f04f8fa0a1cd8e7afea10117fe3f12e21986ec6089769c8afcaf
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ad399867cbb712111db6affaf5c9e3573699ed0eb0def8ce6057de02054ced4f
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:3d7801f7ffe6420595779673a170235921a9090333f8ca51d5b7f664e19c82b2
        version: 1.29.1
      entries:
      - name: serverless-operator.v1.29.1
        version: 1.29.1
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.29
    - currentCSV: serverless-operator.v1.30.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2023-10-30T14:03:25Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.30.0 <1.30.2'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of a particular version of Knative
              Serving
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Eventing
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Kafka components
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:9a28cb1d4c4f4dd7d494829eeeb7122db2174a27747f915244672c95dc8f9d12
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:4800c85e8de031bf77f6fae0e4c8e6c1a9912ccdfb77799cc2ea574c37d870ac
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:66149df8549380673c24e1487d0c2c8eb638b7e502455671ca64ed6f32ff82c3
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:a0c5a91348ec71697747113990652feb024143fbbd0cdd67e8ae672730383a79
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:2d3e314a6584fc6974517c1a992fb59f5133f54c2f6cfaa42f579c0d332e0300
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:ce4b36c90955a356a5774410356aafdac2b60795526a3488b0b76ac2ec1a018c
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:1f7e65009168383cdcb6b2caa22fe4db49e03c71574973a8086f6c332dc5e26e
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:32445e4a73bc7c1e8007775adeb147175d934b17e7d3d4dcf422104dbe80a52b
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:c9353a022e6f883b7c0b2d87b1bc4653172feb13806ebf0aa075d6f9f43c264e
        - registry.redhat.io/rhel8/buildah@sha256:a1e5cc0fb334e333e5eab69689223e8bd1f0c060810d260603b26cf8c0da2023
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:92a39ca90984cebb1938447a31c9170b52994ae98f83caa860acdb2c3ebb75c7
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:06e9e073f092dce67d4227e282fd8eea6dc9f38b90a8d444cf448d80d027b616
        - registry.redhat.io/openshift-serverless-1-tech-preview/eventing-istio-controller-rhel8@sha256:b6ef281fca2b80d379b1e28fe3f23119df8ff765fb2cca291d514053fb512da6
        - registry.redhat.io/ubi8/python-39@sha256:0ff58549041d32b685cc850aaa85da32d1ef73f99cd4b6e0f6882dad5f0399bf
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:21a638b481639c3c58e8b026d875917c30a5874b08fa4ceac40d447076b2bfa4
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:124087e6402caa375b24b976c3b9fd7d53890e20ac3b0b57ba3ac646e352c7ce
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:16939b802f4c16a3d9d941917d0363325f40cdef28f6e1a982bc95401117674c
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:e9505aba7c44cffc3f261a5dc7617a0e54ed98e78f73ea3c9c7a508560b792ed
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:e8c756b659f6187790bc1a5115e0f71791bff96630859a905bc5b75d14596e0e
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:a9c95fc208cc03625cf49bcebdaac73da823d1fa3bfae4239d7db20d3a0b8723
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:3fb6e7626d2b0c92c2a2fef35e9b8171382b4bdf54393f2ab1281ae7ab441761
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/ocp-tools-4-tech-preview/source-to-image-rhel8@sha256:98d8cb3a255641ca6a1bce854e5e2460c20de9fb9b28e3cc67eb459f122873dd
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:12bc62246142a605f4efed0f6bb8843aa2df7699cc1c5ab53debb94b0682e9d1
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:3c8bf480dbb05e09e8d07cab12e4736643f61bf9b5d98efa6bc0557eb31887bc
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:1455606c8edf611d996a48e29ead53e0bb4078f55384b7689058bb25fc59fb2c
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:6057a2d18406e00b0b7b1731870f707844e39193d4cce6de50137b77f17d220b
        - registry.redhat.io/ubi8/nodejs-16@sha256:8abeddc6ccfffbe313f87e5e069e18cb7047f2c7880ad63cf15bcd0a79ba71ff
        - registry.redhat.io/ubi8/openjdk-17@sha256:af305e087061cbaa4107ce0b304a1ef987f6b54bae3f75bfd529cfd084cba9b5
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:2264d2cd4ee2a720667d625ff42b6f8be4f2cd65b57a91eb6044da086a110740
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:69ea22c646c09208ca68345d4b3508453d8c49f7f7fd9a4ff4c4cb78be7bff43
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:a5154f9d13a1c373b8ba47e2bcfb9d4474fd03e08db4134c92aa4f049536f1a8
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:96d6f52711ba433ab2f273be8080840a88e626e989027480424d19c2fe1e9454
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ef116868c9c60e5ee246060688da49832571ff6a7901cfe143a42d048743736e
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:01c8fee92889a84a19136b0843e4fed5db31c1f2a6d997669f103373fd995e36
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:5c933641efc6bf919b1945be5a01dfb7774ae865975b1bb1f44653b3e4ca9f1f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d03f2bf8f5e7cba47102ba26ea0149ae6719076c645d5e451058392d6044b738
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:3508434d5ab6b38c8a7466256f1158ebcd75e8b751afcc5255da8c21566804f0
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:cd1abeb2067afa44a0f1eca301cce37307582d56c1147820ff86b9b64a40588b
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:85a369d5696771d0932b61760336e52ae3f7c66f9a3def5312aefcac03ba516e
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:96c135ced2e3bff434addb012368d96d0778ef1d294ba507ae7e624811e20d46
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:e97eb2bb7eb529cbfe67c14e823ffa1269d614e2d35723896e89efa12a499c6a
        version: 1.30.2
      entries:
      - name: serverless-operator.v1.30.2
        version: 1.30.2
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.30
    - currentCSV: serverless-operator.v1.31.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-02-01T19:15:42Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.31.0 <1.31.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:9221a833ecc66fe2354aee7aaec2a2470b21e2d89cb6ed287ece32edab2a2edb
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:22ffe08d6a0d5461158613f2b2a2ac6936025357e282a09197a750c42b5e6922
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:3dab790fa845ed7eda088e5abdc87799da1354884a53ff1ff728739b8e278418
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:1ee675125f7141e91f4ee57ebce082967834bbdabe1a870c221f31aeb8750be4
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:486bf60d17a5d265ebd65d590a6ee1a7e13bf3953ea797b2f6c2b5910789858f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:bf693eaa845373e1a06041626a9314ec00519a6fd2437b12f7f4f3f104864039
        - registry.redhat.io/ubi8/openjdk-17@sha256:a8165bc2cd5051a96d6937e25ed03155bbd4b731da6e58cebfe2ea83209c16d8
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:565becc77dbe3bf03d89b1399d6e1853253276a71e80b6b14069285d03cb14b2
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:433821e6c7f43da5ca2e38c0550722feb179a0bab021decd9a97ad020cb78b76
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:e49220eaf3143a8958f0cb1f22115b2e834fb7177755433bdc592b753a54a8ec
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:1fcecc11b55007b9c6969a24940a391968378e4102eebf7ceabbe7efda138575
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:b37dd2cf4b459fecb6d681cac36918e486c3bd22501df50b90e20b15fe164d56
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:b2480154951109a159acfb88c5472a312a515d6a6b060d832b3f7bfc16f90ac5
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:887480662b1bfe09626fbca8ec7c5da879d097c46dca11764ea98474ec78ab5d
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:cda04e3b7c2c1f37164733ad162ea0c4e52f14fdc8eb8bdef8507f0ed5454640
        - registry.redhat.io/ubi8/nodejs-16@sha256:f90b28179d1d5de68d42ef1a9166d70a9dfea85a1c6a4bf12dce2c59eeb2fbb1
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:677be039353b2cee51d5faa5d5a2abd6a7bd46992ffe9521eb033eea9925c785
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:3181ad775ead55f317f96466c2719b2b73c5b35d31933733da21273955d160c4
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:9e120ecdbea54eecbc5e0412c1774f8a2ee70d7fca682ecd0757b8627f70680b
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:eacdad8db61b23634c3f89bffa3f1c30ac5c7bbb5e938c6208003ff00f9595f1
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:df1aee67a161fa19d7f49ba56d53e0682cc1996436e985a12a2068fa419b54ea
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:3b01ebdb0704dc6a1a993ac076819489495ab3fe32e3c14609e42d1358856be2
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:d8bdd96c6a9b31f48c961ca438c254511db785fd34bd10b98084b702baf0ebab
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:63e08279d2852d494d4f8836c54e2ce942a342f452a23ef437df9640bb2f5bc9
        - registry.redhat.io/openshift-serverless-1-tech-preview/eventing-istio-controller-rhel8@sha256:5b38e5c71e422161e40f9c37dd2f46212cb80e8e37b00247767276b957080397
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:80361e052d366ffffe9a99893cd83965bd69a9ff81d8c9262976de145bbc6184
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:d2a34100f6309c3cbac7c9ccd452b7bdab2a533afc99463bfbf91e3ba2706d54
        - registry.redhat.io/rhel8/buildah@sha256:b48f410efa0ff8ab0db6ead420a5d8d866d64af846fece5efb185230d7ecf591
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:0a2989f688b966277652957b312df736ada43fff78fd89b17f60b2137e38f115
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c078071cd51c7fd5584039b54a2004331c05a897ee8b0689bfcb9204a8ebae80
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:8353ccb4f0d70597d61e4fd1f799d4bb74e5bc038f6b9916059e712aa70b97c1
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:3be4fc173b489d0beb768739f60274e07d95ffee86d70a8423e1f59678c805ea
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:3a4e7412322755b4c278540c965a403f9e178f138bd4ad7929c1c24a8036dd14
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:de62762efa0df95ed23df56f5520490d8a231e47ce13c433c512254ceff81844
        - registry.redhat.io/ubi8/python-39@sha256:4f35cbcc6f4e108c69f7d2b70cce5afb138741b0c961506376d524e001b922f4
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:99f5dcb92a33c047f0b68236a3a42ffaec78037dd0a5eb79d1144d62d6bfcf43
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:2572d98363cca2dd94984d6cefa573c5054fc206ee70048f901ad4b846d8dd52
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:dbd9c0cac86f79b8d91ab65ae7543854144781d4cb17ef16183931ec9bc234a7
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:7941b421b3e18bda8c565305241621fcdf1212cbbe0dc02f38f4ab0565d2bfcb
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:af5c30fc9f631d9ed110641b815315d8b4aef539f687277a3821d9939ca5dd46
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:ef7d28580e75d746449a90cc0e5343ef3f4ddd75bd9fb866d97836bfbc3b314d
        version: 1.31.1
      entries:
      - name: serverless-operator.v1.31.1
        version: 1.31.1
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.31
    - currentCSV: serverless-operator.v1.32.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-08-01T10:56:58Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.32.0 <1.32.2'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:16c026f09b377c44253cdb902b9a5fa0ae766fd0185cca1863798a212e0daee1
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:2f0f84767f7d76dbbf4387ca2e272766267a2c46b5bc29f5d01f05dfa39cb944
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:20566173a8e907980e603d82fae4d96fcdfd5c43d9004c2e506cf0e09a47b91e
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:04e5b3882223d45a5f5383ce6481efb7b2cbedea57ebd894caafb1454ad8c09d
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:e1e2273fdf6384d95e1128e54c0794d7d620291a272bfb1f91e0c2723be7f1b7
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:22b4d47725ed8cdd663ee26a03f31198fff9e71dc703532e70c42bdfb4840811
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:e7e6cbb46720eeca3f0f4b9829dfeeb3db373a83c89cd808b667ded4d5e139a4
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:a5a7b2be50f211d4d9d555b2a2e2302096c139ff5011d984cc386f1530007f16
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:0729de0f3f03ed220216c6cb718fc34beece0ceacd926b5e2480bdfcf44df5cd
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:ab9a1b952e9603b4ecc5f197715a1d871abc25ebb0c5e6470d4f0b365ae94276
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:2a30ae808d5c32d812406ff4ac9b472e82b6bb0555e62eee4c2c6a972f8750c2
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:b008da825ee2429f04cf39258389740c68056f82cf4bde0cbd3db850b6256885
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:5930caa25ea875d29dff0ec5e235eaa13cb6dc98980d8045335eba1a09ae761d
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:da5f813dea324ffe9679524ea7f3801536fa48f7dbbd8da56a8492b916b15fc3
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:990962ad245a6d62cd11551bd4e1f8b14dc08abbebe47beddb2f4874f78d4f47
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:b0db9f0fc85fcdcd5da980633a3a44be2bbabae068f89d0965263de38789e812
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:9d36cde68bd822016e967148eebd2ef3e7ffab9dcdd37d17f33017e9b703d9b2
        - registry.redhat.io/ubi8/openjdk-17@sha256:a166bd2fedb99cb62ebf4695ab46e19c7e451f887940cdf0b8245c81e403aea2
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:9b6524f9c0e7f78e6ffb372a8587eaf5f4f1281eaf64f93dd00b63be84a4c21e
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:42f31eb8f12dd36541c8c0b6572d942c62e2064d5418428482d676d117a6c426
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:7c8cefd5823ad953f112b905c651f12a97491aafc008b2ba0111f0c9e83e50ae
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:3be20d63d7e82f35ac88274532d6ab7839f9a7306ca272e7d532ac03215faea9
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:46b8dc1c8ea6c40401d4d1a9868ab3edbe8b9e92364a43add6722489e2eefa04
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:ab09a1710aa2331e24df9e7283f3909fdb104c1b314dea88e919d0a396f334d3
        - registry.redhat.io/ubi8/openjdk-21@sha256:bb95ea866269ac8652f666ee1587d71f4f43da33493c2be7f6a9548bcfd2407c
        - registry.redhat.io/ubi8/nodejs-16@sha256:69439fe83fa957089cb6fd3c1042ab71d719f81dd457a5d1fbc56c1e3ea600da
        - registry.redhat.io/ubi8/python-39@sha256:8c96e8648cc6651d0cefbd20496a89f208621601f9664554d3fdeef8ced25453
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:e55c70591c0498b6b85f0d68e304e88b830ea0b35219ba8881143555757be4b7
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:25cbdf78b67e7f5a360313ccfbe3b6e2c98873b4e9ddee46c115a623e77c7abb
        - registry.redhat.io/rhel8/buildah@sha256:55f5c7d88c77e1ab5a2ab8e877172b2b66cb196e21f10dc45148470d0ee8bbae
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:1a1de677297525c0c5d317fe8a7a42684fe3f1b422a7d6bea1f70f0924653b08
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:0507ea8745267fdb53329a5274d59d23c387225d2594af51dd2603b4ad99db0f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:4c9faea6fd162db3340aee8c11af11a08c9d79bbed104967c25df26c4933c8ae
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:14ca32e999767776c3e585ff5eb1aa568f7d76fd3e5ae24cae05ce7564678080
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:8033b7c1e903ba313134645c626eb3bf7ab1a0d68dc2575bc8e26884ad4f805f
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:e70dd3a9e067075eb56e5f2772642545e308348cd49b89357c02de3979b1a5be
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:3ffcf746df914f0949477b46f90cb2992bb63f8198d73ffd54208e7a777a9cb1
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:1befc9969edd2062b3df69f45aee9771b33a2659c3a14db0aaaf5c3df985af58
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:0551f1657dc158bd41eb2338dd6514d55c39fc77300cdf1744d8f5ab14e9ed1b
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:4933b76d95d411d597d4dc460fc46c92875d28e107cda661946206a9d934b8b5
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:70c26c7ac48d1eaaf1f95ca5c5b8ba8ee6b047a3c9c4a2c2db9c49a506bb2205
        version: 1.32.2
      entries:
      - name: serverless-operator.v1.32.2
        version: 1.32.2
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.32
    - currentCSV: serverless-operator.v1.33.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-11-20T08:38:34Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.33.0 <1.33.3'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:e1cb1c24333605e52364456f9bd8d86d787d1bae17ef75dd04be61257fdb2256
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:6dbfa2573a2c78e450d3c1c04c6e7cbc50d9272233e828fe9b42e575ecf386c0
        - registry.redhat.io/ubi8/nodejs-20-minimal@sha256:acee3b00ccf8ff49bcebcf10e028e34a84d45f28f855363bafb3b93c75e84f11
        - registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:e1cb1c24333605e52364456f9bd8d86d787d1bae17ef75dd04be61257fdb2256
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:eb95deb2105f14c6a9f3bb2cdd7306b446109a9b142a7bfc726b6429f270d9e5
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:ea9e4de6228d5c3ed494c2ce5317ec290543ad40bcf7c43c2a3d6565e5d94e5d
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:359f8b9fbb4c3c2f79e8c70f3b50c994693b930f6f078c34c87112ebb93037e3
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:c3dee2598088bda36ef474f4748943a29169ed9589aeb0f77bc79a7f6bef9907
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:672022b6f2a080483d9d6f6f4a5f42e52e522a5797105f3465731809c6cf4947
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:28c9d05a3ba73b76f293e8f4902b574deb44f77a2bae8cf92473b8a8b10cee92
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:a0399a993155e755409355740b260d9b0019559a62be3b6253d8693a4fb59286
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:42d16e4d9f14a427e7a5b7e3a4ea07c258e0f2c74984baeee3b4926a8c5e2640
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:a8821b59d495e24598b27c6510b32ef19de2bb04ea1e568fc5afb9135d772036
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:9b850afa750a67120f68a49f4834ba7f105563df22cbcbc563317fe1545640d1
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:09060de091d617bd4606f7a289b4528265ae0b7a04737f383e79efd5b0d3eb79
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:dd1f7c850f7cf2a4d70403deedd8a57ecc6e0ee5455d68fdfc49ff175fe42dde
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:93b2f5211e79abc38bf92f0b8fb8adf046303efa2fe62c47a2fef9dfb5662f55
        - registry.redhat.io/ubi8/openjdk-21@sha256:095175ef8573de8055e67bd5c8a21098448ac73db4e0d6ccd5269dd36b4d263c
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:b871bd08b782c9168e9a54340efffc0b0effadde8079915b848e4eff72bb3e95
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:879fca768b46676f1df89de1cdc4ded8dac310209c146c2f381ed1deb19b06be
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:993597ff83a4d62a23f642fadc92f182944b953770716b73b7d19492b5a462c3
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:eab4aa93b9d175045b1f01bff09681107d839285e0d633cd5146d2a2efab373c
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:af66cc7dbed88907d96a390fe5c8807a2a60bbe3276b8cd13b4b061cbaf7b162
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:b12272e1a9cb4f4bb62c4badc62b0bab110abb99484612dc781c46bfeac13cc0
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:b2be693c71cfc4ce934d5b01efb863dc6e391f93a6b15fe7eeb15914063a8294
        - registry.redhat.io/openshift-serverless-1-tech-preview/backstage-plugins-eventmesh-rhel8@sha256:28ae309e4cf7feabaf78ac6067e1173f9f5b33a71c06d6c27cf25e01a0b1ed64
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:05e3dd6255294b691783523f7722cfaf13aca1a1e1dab344a25eeea72aaf9f9e
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:e9b38f1df225238836aa829f3a2016372859fb8709cdd2b30496498669a90be1
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:510e98837a549f489b2d74939f3b9c7ed90a8260ba914910f681d5c0dec7655c
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:d8ae7bc888f11121fbbba65074389b879bc22453bbabc551d79d0cee68fc33a3
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:5c7f29554d04eb9348932953344f2ea0806d14e7704355619381bf395d993f58
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:76a47d100de1f2afa23803bec503d9440605db03c7179b231f963e5663a3d12a
        - registry.redhat.io/ubi8/python-39@sha256:ac12e4706e0fdc464e791a3f274474ac1bb14d5af8f217c440fbb0c76d6a5d00
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:ace72435e3e63ab21dff2345f958cf6bd171537ef2846c5d5d91cf821877140d
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:9b94875e0e5d7abdb3b9ba7b04e217a2a9b7b4410d38d1608c86d1c21f929cc1
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:56c6ac59a68976741d2ba676a53e59fc494d04dc44b513d72fdf8a9ad894a54d
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:62bf4bee0ea17ecc60366ecc71431adc79d6f0a5b9f59489913b474128c4adb0
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ccbda564735555785815f27f8f25ad9c6c27dabe4e262586fc51609331692527
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:a9c84ac7e7c3fb1e5dedaf949d54a3aee8a62397d2bd6022094671de00b651da
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:2bd241898bd54dcb248ae86470430042dc545d45e7519ed470a1a9c8c63bdc37
        version: 1.33.3
      entries:
      - name: serverless-operator.v1.33.3
        version: 1.33.3
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.33
    - currentCSV: serverless-operator.v1.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-11-20T08:53:53Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.34.0 <1.34.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:a4b3d75903de1929326041e9885213074473f07fda91b299548d45680c295c00
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi8/python-39@sha256:ac12e4706e0fdc464e791a3f274474ac1bb14d5af8f217c440fbb0c76d6a5d00
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:633ea30ea3b6d5152bbdb61acb4271e5905d50fbe3c731fb4c8d8640a57f1970
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:586fc8ed88f91298dbff31bc2fc4c63e0d6d9356e3358b2f853f49511c3ff9bf
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:d8b94bbce98325fe8079c11c0b7a6c59bad8a7edd0532a7d27ef1632ff0de708
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:ef7f34a025e40f93fbd51f3a4596998a6dc508db824f137ffa1fbaf90e88f4a8
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:2adb0124f04d6c4a552ade79dd1499c8ca997edb45855bf438ca8043b2709e19
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:f94f0e8d21a6b6341b1db00513a1a2ed5cd84fcd898a376ecde942e940a518e7
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:01ea7a06332d60ac16ae0a27adcaaede3f680a98a22ff995259a6c42ac0e358d
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:b07010595201d7cfe0ec4f18e612d323eb38464c74eb187f275617a744d9df12
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:144883b09d8a999fd4d2a109d074e44d7d8b959987512379aa71c71e4060ca28
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:a9e8583c19ca3e01593f8c0e54731cb73df9b46b2ed47a2b1c711828c2295f82
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:76a47d100de1f2afa23803bec503d9440605db03c7179b231f963e5663a3d12a
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:6d042c614431aa9a9e77cf8ba0064423e903f7c418e367ecc23bc02a24a796c0
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:86ffb8be1762782d643c147cefcfbc2b4723b0f97d017314a99bd2275d05a1cf
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:30cb41f09067959cfe3792e070fca62820fcca4035683e5c43dd46d1cce97c01
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:45bc50a993d04b775a153e985a43d587d5455883fc4c6f31b772408297cf9040
        - registry.redhat.io/openshift-serverless-1-tech-preview/backstage-plugins-eventmesh-rhel8@sha256:09dd94090eea9cac7a1935a6d671cb97fd9fd5714c715c9d2aed461e468ca54a
        - registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:a4b3d75903de1929326041e9885213074473f07fda91b299548d45680c295c00
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:946daeff59b41a6f8c4842f8c2668f97ff1b49e2a8afae5d64ba3baa916b1d95
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:75228e70fb1383bbdbcc64d0cf24bf5d977c09112e3f32316824188264c3a2bc
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:33d62d13e16f47db9951ac826671f39fdc048fe7c3fb019eeb4bf9bd0f814222
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:d5ad3c9db7ce8e4fc950261f7c388476214a03d92e0043c7bbc4f0add2b1e69e
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:b9d85323e110d8eb6600223c38f9d4cbe89b50d59e03bf1e6b8808f7cef9a3a5
        - registry.redhat.io/ubi8/nodejs-20-minimal@sha256:acee3b00ccf8ff49bcebcf10e028e34a84d45f28f855363bafb3b93c75e84f11
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:512c4d93dc13cc53d923a6990770a7527d6fb657278e2c5d8f0ece3cd5020155
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:0f126893074cd0398654c4018df6e882b7610e148c82a41ac814f4bf3830bf9d
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:a5950f81cbc64d1edb64ae7f7f120c3435d36afcd14ffc34ae522af297d87e1e
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:3d17f12cc9da578549254ba1b8934b5afc66e453b1d5d667bb90281e45b8795c
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:c0376504edae12e7af6b6cce02ac53b6b374f5058ebefcbdbd32628b551bb0c4
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:1c82d20ccf5eb5f8e7725839f1060f4efdfd1c1b4773a99bd905e690a6e36f38
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:1252960c0c0a4dc7b39090cde0fd5742a19453b1a65a08b737bc41d044a0882a
        - registry.redhat.io/ubi8/openjdk-21@sha256:095175ef8573de8055e67bd5c8a21098448ac73db4e0d6ccd5269dd36b4d263c
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:d40daf67b36387faa9f55e70e50d10d02accfd19470c0ac11b78bf969947fed1
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:84994eabeb7342a0771c7cb1bdd70e7ee3928dd45874eafa64cde5bb385fe964
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:0f8d006d0e7468783739942a1c82c7aec376c9948f23841132c8ddece8ecba06
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:b801ed7b1cb32d4a1522a959e97f280f2eff6fc92e47fea2ea022f180d37831a
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:e520d1b6b87f85dfffab5e53cfb70562ab7edf450061c07609da39e3b2498fdd
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:4b6f161c14c68536ec2c7679d30681afa9a1fa2d512548f0f591791a10a8b894
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:76480201188dd142206040af845bd55b44155af92e0c5ee9bcdb8660bea0cdb8
        version: 1.34.1
      entries:
      - name: serverless-operator.v1.34.1
        version: 1.34.1
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.34
    - currentCSV: serverless-operator.v1.35.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.35.0 <1.35.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:f23d3ab058667c6fe9ee44dc249aa7e8449c1bb9581380ddbc66624511b286d2
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:b2de6b658a1595084a202203c90704c3f9de018a7a904ca276112ad606e374f1
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:0751b9ea7ca197480c4cbca17654d0653d1c0c760ca3c818c2a1018dc952cdc8
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:da8d975527ba100313b92e87d6d2690fddae9988aa8d82d728221295955b7224
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:ab8f85bb18311b0bfee00561ef6b5d9faf810358c954ac2c6f39cf2a7739e07c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:6b842b4eddfc71a2c8f6c1fcebb8f23a05f50c95be1ba2d20a1be660d3014c8e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:e743dd93b7ae855825716403979db3f34cd300caad130e7d3011172b1aa4eaf7
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:0e58edf5ee340fdfb282faeb6c9bee8cd6323e7f29a9b282dbb91c2622131a9c
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:477b1ac0db0abf1e16ae31bb22fdb73dd98f4dbd243fa0a7e6e94852b16a813c
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:b66bb9531f05cd46360f787955c849208496e3e45541cb50b23bfb6b68570aa8
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:9b176477be4c782ce60d81a72bacb566069a499579ce6df5a987a0841f05ad5c
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:86386b54cc10743023884193b7d6c7159b90052bc5e4a0608a6cdfd47865c966
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:cb2ebe98481808e788ca70e7f0a9b3402818b622158f327b99a1e87c23226d2d
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:d0650e4d7e72241e07a3796f20931adba231fbf5e401a45d794761f2c5afae48
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:2cbe84afebac23b5397d144b4d5660af33290df451141fa3625607b6033983e0
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:020264063c50f2a4590f3d7524b271f0f77350a1f8fa3c35ca6e28f0162aa9a5
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:d00a0059f6134b2519f9b1e034b81a943cae7ce391109703f24db21e56ee923d
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:d68fae6fc07efa9da9af0532aadb49f897a15afcdf9a98dc7517ff5e17735454
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:3130a5159adb9b4eaab75a0209d2ac639156d5e9f2346a8b2e6e0e22fad4b45f
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:80a155d2637f47e249fd69760b250c9b7a1eeb6781de9308b54162bb3e1bce71
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:ffaad2214fa897e89e92e574eb5cd1c671edab7f6eab87cf2ffd4ebf162b3cf4
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:b5db9a51bbb46640b23712c61e9c5a1c7bbdbcd9a53b0b92fa36b80dcaffffc0
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:f23d3ab058667c6fe9ee44dc249aa7e8449c1bb9581380ddbc66624511b286d2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:56e7a0158a713c70765ead847732db4e5f8439aacf31e9a4a22672b583028064
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:e7aa54986c3d388c1bb00828c9ccb294ca0b6c26376c7593ed6bfddda0ef4e5d
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:ab0342b325ee67cc2a7d321aa1eda9c6797dc91fe2b6f26e3cbab53fadf564c7
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:732827652b7a7a49181e86d561831001c28b72b7e12d5e3c1d71b2e496003229
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:018459cbba2e3169bc4be6564cf3747c9cd0677445912abf237f8fe6d58e46ac
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:6f4fa254d923f3260ab0f7d6a0b459554f83b1abfcb575a5ce2c0135fba32427
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:6722aecbbbefa80dfaf93143e033c562cea0982514e2eea27654dd4638e1d072
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:2980eca437e4c1cd22a2b6c2ccae9c4d6b55921b3d575fbd4b9241f1f0a8a3a4
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:dfcd11686f161d44786a205d12f9d628d9e847d40ffdb628ec9aa4a0da0b86f0
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:acb28d84cf03b8551efa32646864005dc3bc9bb341a696544892dfd0c87fbda6
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:818dc7bd69b453ba5882b4ccb82841f2d48c7552368417bac20bef1df64f52ff
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:4ca1bfaf2d7e9477583cddfa40f0b2ce59c20ff9ab20022038774dc3cc49662b
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:f56d21089566330ecbd9aa8acf14da332d3db7360aaf6da00c9a19df83f7334c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:8a4e5e0934197c32e0dfa1efc922d750940587fb61e0df49d476a2d035c53413
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:86074674b7926ad9bfc95556e50820792fcbc5e11ab02879378625ffabdd1889
        version: 1.35.1
      entries:
      - name: serverless-operator.v1.35.1
        version: 1.35.1
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.35
    - currentCSV: serverless-operator.v1.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.36.0 <1.36.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-ddb-streams-source-rhel8@sha256:a903480159ebf1d8652eeaa702d9885fb76bb106c02d75cec5c82c31d77d6410
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-timer-source-rhel8@sha256:8fed5ef9a1f39a55077bb5026c8854d0ee7fc8d8ff5da277ce9c1bfc2ab2474b
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:77e4689b2ce7a713e905553e0e42b8e128045f40cb6c6d4ee082f126cbc46b00
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:3d780ad1ab280b5092b90e277485d589da2a18e8b08090af7d95aecd17af545d
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:429d5d9568cca31fbe83c663276a300286af4cdbf3d35685afdbe5f6f2f8c6e0
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:4344716eb2940c6a3fa9a278250f936602bb1549fece6037912837719c178b24
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:f2d112ba9fdfccea988589b99b463c8a7157475f7a3f03678e8999e65b678cc3
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:5352e4eb4e426a41521276decadf204b6d23b4762a298e4d3dc73969e3eddc68
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:187fb9d64b53c55a029ffce3c0f9e7c3304683fa8a0e832cfb7134e8a3d0c237
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:6210a946f373b9c36c9b31293debe6fed090008b8234c0ba17d37fa81646172e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-transform-jsonata-rhel8@sha256:3af4c48030263eb5769e5f36a6dd13fd390d8566efaab37532f39e1df1c95c67
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-sink-rhel8@sha256:d9b944e64391a996104afbe8a21f66316a948e9fb7b91734e861847285be59c7
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:631eb3a8dd9d83ceafa7d8a4da4b1e2d4f2740e6017c766828dcc7575cf1322f
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:a6f565d73375e827d830dca85838f4c4cb7f08b0c078acb0c0d567b25f0d256a
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:a47b6b5237a87ccd90e5a49a53d06485fd55c87eebf12d01c335b87fc3a8c494
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:ed60e889a677fb139f2eb4bd9ab4fda6ecc39fd050ca2b6cc1c19717e428535a
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:074ebfc4494581611647ba90049655a470ca43743610a27be77db66a8633bac6
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:dcfe5fbc57b9f8520adbff368c8bf29eb491c9c26d1a1737a1722cffea83eb1b
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:f3d87a7fa5b5f2818effcac3304bd56eddcdb0db97a18ac0f35a06fc7e223e7d
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:4274004b6289fa9bea3fa6e56d64077a328a6b5e63a9f0e9c41f6d561f682bd7
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-source-rhel8@sha256:27e74287774cda208f667bc2a386debe6b6255e8991abfb37c959c8525f2af53
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:3b04e9d8cd4b5e43a07ece838d10478723a6c6529350f010550b4a0d1d0950d1
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:ee6533b71cf280877c9d7489dfd821c9209c6802bed7fa3fdb4f8f00fbeacb09
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:6b7316cce3432d391adc6ac38c92ff0ee86dd7ae0101957cc65cd338cc69e352
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:c39cb88afaddc295d93a3774b8a20d001b82392df540fb502ee02316d82460a9
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:82e69a1890042f5636ac7bb645b1aaeaba0a4894851eccc1fa1c63296918e31a
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:52ba6b1252b6f3d41e2a0b3b880cb768010ab9dd925ac34596f6adaedcbbe07c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-sink-rhel8@sha256:afa634e029bcf79a972f49372498d44b45bcfbdb6eb28ecaf4f0b0b2a0a1824b
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:f4d86b196457cdeebdc380b6f1b707c10f8dee789c7779c90ee1332e511af8c4
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:259e5dab50d36cdbfc117a42325128c4e16f34c048ba64b36d4a89132ddb4c1a
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-log-sink-rhel8@sha256:30b8045ef5f925a36c58941386f1899bec125d005a08aeeeba5edd49b0859149
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:3fe326ad9a00b112fc888ffd345c8d5042f0a22cf32ea575931192c39ce1fece
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:e3ae1cc6012f19631c012a07b113fb015977d2ee4f9faab3207fbc2464c25aa7
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:bf82b37ab8fd29c7a7cd846008a16d14b4cfdc1c09df4f731d71be3957e90123
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:d6f338e08fa937c8a4ccfec692424de4acded565f7ca3971ea507dffeb351503
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:4aeb68a22f63300e54391ec6b4cda8c4e665d904b5dcdd3c40155040525742a0
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:596a7272d6de714d7dbc6e00e9e50a2872846c269547465252d953172ef0b009
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:0c2cdf917bbad1f5bca8a2a2ef12a96cfcd0c77c190cc35687eb57169b54ab6f
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:c7c2e738e0d9da1d2d1993f68938e72c4709b1f28a54fa92224ef6a0e456b0aa
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:a1834b620f875681388f0ae224dd0771773a8b61b1d8e5214a942f232854b87b
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:9b328fff0cc8545fb9bb1f24f11d78cfac3ee76dedc11fa4def4262d79a60e8c
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:78fee4c596f2848487ca704da3fa1074e3c9c99b3f96f66833aeba63836d7b93
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sns-sink-rhel8@sha256:51b5e18e8d4c1ca81d7c5ddfea6352a3887f757acfcc53bd2f8bd3b01337ed42
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-source-rhel8@sha256:88323da16360981f27b8e4686fbbb64053e4e0e3c79c65d695edb317b334e1e0
        version: 1.36.1
      entries:
      - name: serverless-operator.v1.36.1
        version: 1.36.1
      - name: serverless-operator.v1.36.0
        version: 1.36.0
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.36
    defaultChannel: stable
    packageName: serverless-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-pipes-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-pipes-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pipes.services.k8s.aws/v1alpha1",
                "kind": "Pipe",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/pipes-controller:1.1.1
          createdAt: "2025-09-19T21:15:33Z"
          description: AWS EventBridge Pipes controller is a service controller for
            managing Pipes resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pipe represents the state of an AWS pipes Pipe resource.
            displayName: Pipe
            kind: Pipe
            name: pipes.pipes.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EventBridge Pipes resources in AWS from within your Kubernetes cluster.

          **About Amazon EventBridge Pipes**

          Amazon EventBridge Pipes connects sources to targets. Pipes are intended for point-to-point integrations between supported [sources](https://docs.aws.amazon.com/eventbridge/latest/userguide/eb-pipes-event-source.html) and [targets](https://docs.aws.amazon.com/eventbridge/latest/userguide/eb-pipes-event-target.html), with support for advanced transformations and [enrichment](https://docs.aws.amazon.com/eventbridge/latest/userguide/pipes-enrichment.html). It reduces the need for specialized knowledge and integration code when developing event-driven architectures, fostering consistency across your company’s applications. To set up a pipe, you choose the source, add optional filtering, define optional enrichment, and choose the target for the event data.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EventBridge Pipes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pipes
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EventBridge Pipes Developer Resources
          url: https://aws.amazon.com/eventbridge/pipes/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: pipes maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/pipes-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-pipes-controller.v1.1.1
        version: 1.1.1
      - name: ack-pipes-controller.v1.1.0
        version: 1.1.0
      - name: ack-pipes-controller.v1.0.25
        version: 1.0.25
      - name: ack-pipes-controller.v1.0.24
        version: 1.0.24
      - name: ack-pipes-controller.v1.0.23
        version: 1.0.23
      - name: ack-pipes-controller.v1.0.22
        version: 1.0.22
      - name: ack-pipes-controller.v1.0.21
        version: 1.0.21
      - name: ack-pipes-controller.v1.0.20
        version: 1.0.20
      - name: ack-pipes-controller.v1.0.18
        version: 1.0.18
      - name: ack-pipes-controller.v1.0.17
        version: 1.0.17
      - name: ack-pipes-controller.v1.0.16
        version: 1.0.16
      - name: ack-pipes-controller.v1.0.15
        version: 1.0.15
      - name: ack-pipes-controller.v1.0.14
        version: 1.0.14
      - name: ack-pipes-controller.v1.0.13
        version: 1.0.13
      - name: ack-pipes-controller.v1.0.12
        version: 1.0.12
      - name: ack-pipes-controller.v1.0.11
        version: 1.0.11
      - name: ack-pipes-controller.v1.0.10
        version: 1.0.10
      - name: ack-pipes-controller.v1.0.9
        version: 1.0.9
      - name: ack-pipes-controller.v1.0.8
        version: 1.0.8
      - name: ack-pipes-controller.v1.0.6
        version: 1.0.6
      - name: ack-pipes-controller.v1.0.3
        version: 1.0.3
      - name: ack-pipes-controller.v1.0.2
        version: 1.0.2
      - name: ack-pipes-controller.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-pipes-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Sigstore Authors
      provider-url: https://sigstore.dev
    name: sigstore-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sigstore-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.sigstore.dev/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "name": "rekor"
                },
                "spec": {
                  "server": {
                    "ingress": {
                      "hostname": "rekor.example.com"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          containerImage: ghcr.io/sigstore/sigstore-helm-operator@sha256:d9c6eb162b97a1aa1ad3bf515db65d01a96e55b6f8bcd646864d3e64a3ee2b9e
          createdAt: "2021-09-01T21:20:28Z"
          description: Helm Charts for Sigstore Project.
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/sigstore/sigstore-helm-operator
          support: The Sigstore Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Rekor
            name: rekors.helm.sigstore.dev
            version: v1alpha1
        description: |-
          Helm Charts for the sigstore project

          ## Charts

          The following charts are available:

          * [Rekor](https://github.com/sigstore/helm-charts/tree/main/charts/rekor)
        displayName: sigstore
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        links:
        - name: Sigstore Helm Operator
          url: https://github.com/sabre1041/sigstore-helm-operator
        maintainers:
        - email: ablock@redhat.com
          name: Andrew Block
        maturity: alpha
        provider:
          name: The Sigstore Authors
          url: https://sigstore.dev
        relatedImages:
        - quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - ghcr.io/sigstore/sigstore-helm-operator@sha256:d9c6eb162b97a1aa1ad3bf515db65d01a96e55b6f8bcd646864d3e64a3ee2b9e
        version: 0.0.1
      entries:
      - name: sigstore-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: sigstore-helm-operator
    provider:
      name: The Sigstore Authors
      url: https://sigstore.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Loki SIG Operator
      provider-url: ""
    name: loki-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: docker.io/grafana/loki-operator:0.8.0
          createdAt: "2025-03-17T11:00:08Z"
          description: The Community Loki Operator provides Kubernetes native deployment
            and management of Loki and related logging components.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/loki/tree/main/operator
          support: Grafana Loki SIG Operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |-
          The Community Loki Operator provides Kubernetes native deployment and management of Loki and related logging components.
          The purpose of this project is to simplify and automate the configuration of a Loki based logging stack for Kubernetes clusters.

          ### Operator features

          The Loki Operator includes, but is not limited to, the following features:

          * Kubernetes Custom Resources: Use Kubernetes custom resources to deploy and manage Loki, Alerting rules, Recording rules, and related components.
          * Simplified Deployment Configuration: Configure the fundamentals of Loki like tenants, limits, replication factor and storage from a native Kubernetes resource.

          ### Feature Gates

          The Loki Operator Bundle provides a set of feature gates that enable/disable specific feature depending on the target Kubernetes distribution. The following feature gates are enabled by default:
          * `serviceMonitors`: Enables creating a Prometheus-Operator managed ServiceMonitor resource per LokiStack component.
          * `serviceMonitorTlsEndpoints`: Enables TLS for the ServiceMonitor endpoints.
          * `lokiStackAlerts`: Enables creating PrometheusRules for common Loki alerts.
          * `httpEncryption`: Enables TLS encryption for all HTTP LokiStack services.
          * `grpcEncryption`: Enables TLS encryption for all GRPC LokiStack services.
          * `builtInCertManagement`: Enables the built-in facility for generating and rotating TLS client and serving certificates for all LokiStack services and internal clients
          * `lokiStackGateway`: Enables reconciling the reverse-proxy lokistack-gateway component for multi-tenant authentication/authorization traffic control to Loki.
          * `runtimeSeccompProfile`: Enables the restricted seccomp profile on all Lokistack components.
          * `defaultNodeAffinity`: Enable the operator will set a default node affinity on all pods. This will limit scheduling of the pods to Nodes with Linux.
          * `lokiStackWebhook`: Enables the LokiStack CR validation and conversion webhooks.
          * `alertingRuleWebhook`: Enables the AlertingRule CR validation webhook.
          * `recordingRuleWebhook`: Enables the RecordingRule CR validation webhook.
          * `rulerConfigWebhook`: Enables the RulerConfig CR validation webhook.

          In addition it enables the following OpenShift-only related feature gates:
          * `servingCertsService`: Enables OpenShift ServiceCA annotations on the lokistack-gateway service only.
          * `ruleExtendedValidation`: Enables extended validation of AlertingRule and RecordingRule to enforce tenancy in an OpenShift context.
          * `clusterTLSPolicy`: Enables usage of TLS policies set in the API Server.
          * `clusterProxy`: Enables usage of the proxy variables set in the proxy resource.

          ### Before you start

          1. Ensure that the appropriate object storage solution, that will be used by Loki, is avaliable and configured.
        displayName: Community Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Documentation
          url: https://loki-operator.dev/
        maintainers:
        - email: loki-operator-team@googlegroups.com
          name: Grafana Loki SIG Operator
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Grafana Loki SIG Operator
        relatedImages:
        - docker.io/grafana/loki-operator:0.8.0
        - quay.io/openshift/origin-kube-rbac-proxy:latest
        - quay.io/observatorium/api:latest
        - quay.io/observatorium/opa-openshift:latest
        - docker.io/grafana/loki:3.4.2
        version: 0.8.0
      entries:
      - name: loki-operator.v0.8.0
        version: 0.8.0
      - name: loki-operator.v0.7.1
        version: 0.7.1
      - name: loki-operator.v0.7.0
        version: 0.7.0
      - name: loki-operator.v0.6.2
        version: 0.6.2
      - name: loki-operator.v0.6.1
        version: 0.6.1
      - name: loki-operator.v0.6.0
        version: 0.6.0
      - name: loki-operator.v0.5.0
        version: 0.5.0
      - name: loki-operator.v0.4.0
        version: 0.4.0
      - name: loki-operator.v0.3.0
        version: 0.3.0
      - name: loki-operator.v0.2.0
        version: 0.2.0
      - name: loki-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: loki-operator
    provider:
      name: Grafana Loki SIG Operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-license-service-reporter-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-license-service-reporter-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "IBMLicenseServiceReporter",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-license-service-reporter-operator",
                    "app.kubernetes.io/instance": "ibmlicenseservicereporter-instance",
                    "app.kubernetes.io/managed-by": "operator",
                    "app.kubernetes.io/name": "ibmlicenseservicereporter",
                    "app.kubernetes.io/part-of": "ibm-license-service-reporter-operator"
                  },
                  "name": "instance"
                },
                "spec": {
                  "authentication": {
                    "useradmin": {
                      "enabled": true
                    }
                  },
                  "license": {
                    "accept": false
                  },
                  "version": "4.2.14"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:277e2ab849907553e4992c2dccc11f03ae1e0302f20aae48c620968aa97d6afc
          createdAt: "2025-02-26T12:10:52Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <4.2.14'
          operatorframework.io/suggested-namespace: ibm-ls-reporter
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "IBM License Service Reporter is a singleton which may be
              used in a single or multi-cluster environment. It aggregates data pushed
              from IBM License Services, deployed on clusters and from ILMT. \n Documentation:
              https://ibm.biz/lsvc-rprtr. \n License: Please refer to the IBM Terms
              website (ibm.biz/lsvc-lic) to find the license terms for the particular
              IBM product for which you are deploying this component. \n IBM License
              Service Reporter is a free, optionally installed add-on – one of services
              of Cloud Pak Foundational Services. Thanks to the IBM License Service
              Reporter customer can: \n - see IBM software deployments on a single
              dashboard, verify and maintain license compliance and avoid audit infractions,
              \n - see historical data on IBM software deployments to support making
              informed decisions for future purchases, \n - see details of software
              deployments, including source (ILMT for VMs, License Service for containers),
              cluster and quantities of licenses deployed, \n - identify which software
              deployments are VMs vs. containers and can use this information to evaluate
              workloads for modernization."
            displayName: IBMLicense Service Reporter
            kind: IBMLicenseServiceReporter
            name: ibmlicenseservicereporters.operator.ibm.com
            version: v1alpha1
        description: |-
          IBM License Service Reporter is a free, optionally installed add-on – one of services of Cloud Pak Foundational Services. Thanks to the IBM License Service Reporter customer can:
           * see IBM software deployments on a single dashboard, verify and maintain license compliance and avoid audit infractions,
           * see historical data on IBM software deployments to support making informed decisions for future purchases,
           * see details of software deployments, including source (ILMT for VMs, License Service for containers), cluster and quantities of licenses deployed,
           * identify which software deployments are VMs vs. containers and can use this information to evaluate workloads for modernization.

          IBM License Service Reporter is a singleton which may be used in a single or multi-cluster environment. It aggregates data pushed from IBM License Services, deployed on clusters and from ILMT.

          **Supported platforms**

          Red Hat OpenShift Container Platform 4.10 or newer installed on Linux x86_64.

          **Prerequisites**

          Prerequisites depend on the integration of the License Service Reporter with an IBM Cloud Pak. For more information, see the applicable IBM Cloud Pak documentation.
        displayName: IBM License Service Reporter
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - licensing
        - usage
        - IBMLicenseServiceReporter
        - IBM
        - Cloud
        - License
        - License Service
        - Service
        - Reporter
        - License Service Reporter
        - License Reporter
        - Service Reporter
        maintainers:
        - email: talk2sam@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-ui@sha256:213889e3c4c1d7bb10c88926ac8276d7e78c558397993d098e36f6c5da617de1
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-oauth2-proxy@sha256:c403c24857e29ddc306ba5d9bc28c3d556b181c5265fee16879fd2b4d946f899
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator-bundle@sha256:648f92bacd52fb51655925e34da1d0bdd0da720da8489b79064e73a47dcdd54a
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:277e2ab849907553e4992c2dccc11f03ae1e0302f20aae48c620968aa97d6afc
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-postgresql@sha256:b6e11080774863f56dd16d5748d03c3f89f6f4872f662ee218d6bf600491a35f
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter@sha256:87e62695d733a60bfc2084cd751810df6801a5fbe1ab87b0a421f6814f1b887a
        version: 4.2.14
      entries:
      - name: ibm-license-service-reporter-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-license-service-reporter-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: orchestrator-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: orchestrator-operator.v1.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Orchestrator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "orchestrator-sample"
                  },
                  "name": "orchestrator-sample"
                },
                "spec": {
                  "argocd": {
                    "enabled": false,
                    "namespace": "orchestrator-gitops"
                  },
                  "platform": {
                    "eventing": {
                      "broker": {}
                    },
                    "monitoring": {
                      "enabled": false
                    },
                    "namespace": "sonataflow-infra",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "64Mi"
                      }
                    }
                  },
                  "postgres": {
                    "authSecret": {
                      "name": "sonataflow-psql-postgresql",
                      "passwordKey": "postgres-password",
                      "userKey": "postgres-username"
                    },
                    "database": "sonataflow",
                    "name": "sonataflow-psql-postgresql",
                    "namespace": "sonataflow-infra"
                  },
                  "rhdh": {
                    "devMode": true,
                    "installOperator": true,
                    "name": "my-rhdh",
                    "namespace": "rhdh",
                    "plugins": {
                      "notificationsEmail": {
                        "enabled": false,
                        "port": 587,
                        "replyTo": "",
                        "sender": ""
                      }
                    }
                  },
                  "serverless": {
                    "installOperator": true
                  },
                  "serverlessLogic": {
                    "installOperator": true
                  },
                  "tekton": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-08-15T00:51:04Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/rhdhorchestrator/orchestrator-go-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Orchestrator
            name: orchestrators.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub Orchestrator is a plugin that enables serverless asynchronous workflows to Backstage.

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://rhdhorchestrator.io)  displayName: Orchestrator Operator
        displayName: Orchestrator Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Orchestrator
        - RHDH
        links:
        - name: Orchestrator Operator
          url: https://github.com/rhdhorchestrator/orchestrator-go-operator
        maintainers:
        - email: jubah@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/rhdh-orchestrator-dev-preview-beta/controller-rhel9-operator@sha256:9f5423829ac55faf2395e2d0d3683b238339dc03285696e1b350c5c5a1c4df3a
        version: 1.6.1
      entries:
      - name: orchestrator-operator.v1.6.1
        version: 1.6.1
      - name: orchestrator-operator.v1.6.0
        version: 1.6.0
      - name: orchestrator-operator.v1.5.2
        version: 1.5.2
      - name: orchestrator-operator.v1.5.1
        version: 1.5.1
      - name: orchestrator-operator.v1.5.0
        version: 1.5.0
      - name: orchestrator-operator.v1.4.0
        version: 1.4.0
      - name: orchestrator-operator.v1.3.1
        version: 1.3.1
      - name: orchestrator-operator.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: orchestrator-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtv-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:d0158c37686bf378347dea5368c0c7838eec9674fbb0ff11f369e6fd99753652
          createdAt: "2024-09-23T16:42:11Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.7.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.5 is supported on OpenShift 4.12 and 4.13

          Migration Toolkit for Virtualization 2.6 is supported on OpenShift 4.14 and 4.15

          Migration Toolkit for Virtualization 2.7 is supported on OpenShift 4.15 and 4.16

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:d0158c37686bf378347dea5368c0c7838eec9674fbb0ff11f369e6fd99753652
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:f8b8d2ce23c1be9dd74d18c7a90324288213e91630da529ded561fbfe5d59a5d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:fffa2fd9f8c804d021264748a910ceb17cee388b4d55cc763a8598234db87f15
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:31e20712c82ae77807eb5dfba5e8616c045280ff164027443214ab8f259d964a
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:ac13419a8c9993bbe656563f478095526dedd34a93d6463979d26852d08ed217
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:19909db45872b2d665eac31f20dac2f0fa3bd9b9a343f2be839747106de6bc5b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:90742a924a1782dced8ae639e68aef93df3c14e4a7b65e2e6b1b73e22860e7a7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:3a07a1e7c74783b0de3c190717f17091d4c9d573776419b46c2ef0c55328d769
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:045cfaec032648cb1a734398fd625de7ac1cc25bfd621b7558e641da82316b58
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:8ce4bba6a69de762f221792c03f6a759afcd16dc576de4b9b1f341d660cb82b0
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:1b06a93c3458b92d40d8eb437043c9abbb9ec5290fe85b2f6ef12297bea0c838
        version: 2.7.0
      entries:
      - name: mtv-operator.v2.7.0
        version: 2.7.0
      name: candidate
    - currentCSV: mtv-operator.2.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui": "true",
                  "feature_validation": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "plan": {
                      "name": "example-plan",
                      "namespace": "openshift-mtv"
                   }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {}
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "type": "vmware",
                   "secret": {
                      "name": "",
                      "namespace": ""
                   },
                   "url": ""
                }
              },
              {
                "kind": "Host",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                     "namespace": "",
                     "name": ""
                   },
                   "id": "",
                   "ipAddress": ""
                }
              },
              {
                "kind": "NetworkMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-networkmap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "type": "pod",
                            "namespace": "",
                            "name": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "StorageMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-storagemap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "storageClass": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "Hook",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "kind": "Provisioner",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "cinder",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "name": "kubernetes.io/cinder",
                  "volumeModes": [
                    {
                      "name": "Block",
                      "priority": 0,
                      "accessModes": [
                        {
                          "name": "ReadWriteOnce",
                          "priority": 0
                        },
                        {
                          "name": "ReadWriteMany",
                          "priority": 1
                        }
                      ]
                    },
                    {
                      "name": "Filesystem",
                      "priority": 1,
                      "accessModes": [
                        {
                          "name": "ReadWriteOnce",
                          "priority": 0
                        },
                        {
                          "name": "ReadWriteMany",
                          "priority": 1
                        }
                      ]
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:74a740878707470ea79c7bf03a7f045c2ea77d924d0890042ba3554dc5613a33
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.1.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui": "true",
                "feature_validation": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          repository: https://github.com/konveyor/forklift-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: Storage Provisioner Capabilities
            displayName: Provisioner
            kind: Provisioner
            name: provisioners.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator fully manages the deployment and life cycle of Migration Toolkit for Virtualization on OpenShift.

          If you experience any issues or have feature requests, please [file an issue in our Github repository](https://github.com/konveyor/forklift-operator/issues)
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        links:
        - name: Konveyor Forklift Operator
          url: https://github.com/konveyor/forklift-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:584a696270554f4fb64755cdc5b525bc6852925e6780f776fb309b1bd8d81d0f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ui-rhel8@sha256:8089b0889f8956f2abf5a69f4ab8f4382919bc15a48663b98551fb0b865d8af7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel8@sha256:aead9b494562985e20fa466888f28119ab302d64e93d7f23bc7d8da1c60267e7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel8@sha256:8c6758c2d81f807c3875f41cec14d9e39f1fe7710584794de57a1449510110db
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:74a740878707470ea79c7bf03a7f045c2ea77d924d0890042ba3554dc5613a33
        version: 2.1.0
      entries:
      - name: mtv-operator.2.1.0
        version: 2.1.0
      name: release-v2.1.0
    - currentCSV: mtv-operator.v2.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui": "true",
                  "feature_validation": "true",
                  "feature_must_gather_api": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "plan": {
                      "name": "example-plan",
                      "namespace": "openshift-mtv"
                   }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {}
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "type": "vmware",
                   "secret": {
                      "name": "",
                      "namespace": ""
                   },
                   "url": ""
                }
              },
              {
                "kind": "Host",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                     "namespace": "",
                     "name": ""
                   },
                   "id": "",
                   "ipAddress": ""
                }
              },
              {
                "kind": "NetworkMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-networkmap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "type": "pod",
                            "namespace": "",
                            "name": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "StorageMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-storagemap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "storageClass": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "Hook",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "kind": "Provisioner",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "cinder",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "name": "kubernetes.io/cinder",
                  "volumeModes": [
                    {
                      "name": "Block",
                      "priority": 0,
                      "accessModes": [
                        {
                          "name": "ReadWriteOnce",
                          "priority": 0
                        },
                        {
                          "name": "ReadWriteMany",
                          "priority": 1
                        }
                      ]
                    },
                    {
                      "name": "Filesystem",
                      "priority": 1,
                      "accessModes": [
                        {
                          "name": "ReadWriteOnce",
                          "priority": 0
                        },
                        {
                          "name": "ReadWriteMany",
                          "priority": 1
                        }
                      ]
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:3d28138500d6b715e7d11983684db12ced28dfef9c97c619044f0b830534f4ed
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.2.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui": "true",
                "feature_validation": "true",
                "feature_must_gather_api": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          repository: https://github.com/konveyor/forklift-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: Storage Provisioner Capabilities
            displayName: Provisioner
            kind: Provisioner
            name: provisioners.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation
          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.
          * Must-gather-api, a service to generate targeted must-gather archives.

          ### Compatibility
          Migration Toolkit for Virtualization 2.0 is supported in OpenShift 4.7
          Migration Toolkit for Virtualization 2.1 is supported in OpenShift 4.8
          Migration Toolkit for Virtualization 2.2 is supported in OpenShift 4.9

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [Konveyor Community](https://github.com/konveyor/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/konveyor/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Konveyor Forklift Operator
          url: https://github.com/konveyor/forklift-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:3d28138500d6b715e7d11983684db12ced28dfef9c97c619044f0b830534f4ed
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel8@sha256:097a3c667e594060844cbe5d16adcc278d35402d1965ff3fca249c8f84a688dc
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:02ec1b0dbc12ba31fce85ee164c97560ba3a3184316324ab7c871989a319bf8d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-api-rhel8@sha256:d6bf2bb018c471f8f86f85c62ef52f4a99510b748ac8b6b7b177fc44fa53fb44
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ui-rhel8@sha256:697b20d6423ac25c0d2f41b2598f4909b6863e85265a4f52bc23b8ce77afc732
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel8@sha256:220d8f97173292dc8389bf4b1a350843757f01ea7861bbcebcf75749366b450b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel8@sha256:6f332be2991b026d5ce2b22e5bdef8d17da06c8b437fb3e5b26e8d35b75e3f95
        version: 2.2.0
      entries:
      - name: mtv-operator.v2.2.0
        version: 2.2.0
      name: release-v2.2.0
    - currentCSV: mtv-operator.v2.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui": "true",
                  "feature_validation": "true",
                  "feature_must_gather_api": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "plan": {
                      "name": "example-plan",
                      "namespace": "openshift-mtv"
                   }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {}
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "type": "vmware",
                   "secret": {
                      "name": "",
                      "namespace": ""
                   },
                   "url": ""
                }
              },
              {
                "kind": "Host",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                     "namespace": "",
                     "name": ""
                   },
                   "id": "",
                   "ipAddress": ""
                }
              },
              {
                "kind": "NetworkMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-networkmap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "type": "pod",
                            "namespace": "",
                            "name": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "StorageMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-storagemap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "storageClass": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "Hook",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:fc6a4503d00719a0d22736a78689b5828a8ebaf8c7895689a56df5cf7b43b97f
          createdAt: "2022-11-08T18:56:59Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.3.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui": "true",
                "feature_validation": "true",
                "feature_must_gather_api": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/forklift-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.
          * Must-gather-api, a service to generate targeted must-gather archives.

          ### Compatibility

          Migration Toolkit for Virtualization 2.2 is supported in OpenShift 4.9

          Migration Toolkit for Virtualization 2.3 is supported in OpenShift 4.10

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [Konveyor Community](https://github.com/konveyor/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/konveyor/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/konveyor/forklift-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ui-rhel8@sha256:e6c2337e610e0f634e57a200a92f492ebe1f111d9fccaa019f8756ba907b067b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel8@sha256:4423034aac33409721ce1b5344f6d5330c174c9b47fc6837b8de7b47018036d7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel8@sha256:872e2a803742071bf81b70d9a58aafd0ceb6ce1d80e651d876690820835854c0
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:fc6a4503d00719a0d22736a78689b5828a8ebaf8c7895689a56df5cf7b43b97f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel8@sha256:59ad5149ce8691eb7944fa291be75e5701c8e74160a0b9a74d098872be704a67
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:a0c201d3474996a08ce662b074ebb069e5082da36f3e7977cf54e3c4e1e54b1d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-api-rhel8@sha256:a7a2d6216ee87284a4f9e07d4c1973b9e001292c504c3b933515826e90073167
        version: 2.3.4
      entries:
      - name: mtv-operator.v2.3.4
        version: 2.3.4
      name: release-v2.3
    - currentCSV: mtv-operator.v2.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui": "true",
                  "feature_validation": "true",
                  "feature_must_gather_api": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "plan": {
                      "name": "example-plan",
                      "namespace": "openshift-mtv"
                   }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {}
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "type": "vmware",
                   "secret": {
                      "name": "",
                      "namespace": ""
                   },
                   "url": ""
                }
              },
              {
                "kind": "Host",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                     "namespace": "",
                     "name": ""
                   },
                   "id": "",
                   "ipAddress": ""
                }
              },
              {
                "kind": "NetworkMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-networkmap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "type": "pod",
                            "namespace": "",
                            "name": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "StorageMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-storagemap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "storageClass": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "Hook",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:fc6a4503d00719a0d22736a78689b5828a8ebaf8c7895689a56df5cf7b43b97f
          createdAt: "2022-11-08T18:56:59Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.3.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui": "true",
                "feature_validation": "true",
                "feature_must_gather_api": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/forklift-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.
          * Must-gather-api, a service to generate targeted must-gather archives.

          ### Compatibility

          Migration Toolkit for Virtualization 2.2 is supported in OpenShift 4.9

          Migration Toolkit for Virtualization 2.3 is supported in OpenShift 4.10

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [Konveyor Community](https://github.com/konveyor/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/konveyor/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/konveyor/forklift-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel8@sha256:4423034aac33409721ce1b5344f6d5330c174c9b47fc6837b8de7b47018036d7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel8@sha256:872e2a803742071bf81b70d9a58aafd0ceb6ce1d80e651d876690820835854c0
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel8@sha256:59ad5149ce8691eb7944fa291be75e5701c8e74160a0b9a74d098872be704a67
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:a0c201d3474996a08ce662b074ebb069e5082da36f3e7977cf54e3c4e1e54b1d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-api-rhel8@sha256:a7a2d6216ee87284a4f9e07d4c1973b9e001292c504c3b933515826e90073167
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ui-rhel8@sha256:e6c2337e610e0f634e57a200a92f492ebe1f111d9fccaa019f8756ba907b067b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:fc6a4503d00719a0d22736a78689b5828a8ebaf8c7895689a56df5cf7b43b97f
        version: 2.3.4
      entries:
      - name: mtv-operator.v2.3.4
        version: 2.3.4
      name: release-v2.3.0
    - currentCSV: mtv-operator.v2.4.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_must_gather_api": "true",
                  "feature_ui": "false",
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:f17fb7a2a840bda51551683e18d01ad1fae3ea73c293de1711f382ed309b8668
          createdAt: "2023-10-23T12:16:09Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.4.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui": "false",
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_must_gather_api": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.
          * Must-gather-api, a service to generate targeted must-gather archives.

          ### Compatibility

          Migration Toolkit for Virtualization 2.2 is supported in OpenShift 4.9

          Migration Toolkit for Virtualization 2.3 is supported in OpenShift 4.10 and 4.11

          Migration Toolkit for Virtualization 2.4 is supported in OpenShift 4.11 and 4.12

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-api-rhel8@sha256:cf22690c6bb420b9acc09ee5f90329ece6a35c18e5d3d6f6e40a320e2385ad85
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:7b4eb4de5dbc7e1af9c74d47fb6bba39f4973dc2d41531989c9fcf454b715b14
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:dc7ed113fae6ea3d5baaf165b08b5404a15f3a70f08089caca65b837623af4fb
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:8d3aeb8f6cfb17f5694f92f56a2361e08d220e2608183fb39ac94e92187b3fc7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-warm-rhel8@sha256:a6f91bc0ffc5e2f545a18b8888715281db7a751e1eff5f9eb89e8d319268a3ec
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:f17fb7a2a840bda51551683e18d01ad1fae3ea73c293de1711f382ed309b8668
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:8e7a7b3c2dc31b90e2102a82633f1eda95355496ff657964fedb26157f9ec229
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:ad9bacdfa4f10e020902d8e0381bdd84e48ce89ffab0f4144bd3c03fc91f7259
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:86716db780d3e43fa1e3029ab095ae6a557cf023797f4d79565a1aa5ebcb9b93
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:a7caa55c9d211a869b42e6f0101cd5099d64f5f3fd423f8a06091bad98d84aff
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ui-rhel8@sha256:54638e44269175e112a510fe2558f9a1dda4758bf3b3fb5e5b54c6617a5917bd
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:20a7b16f9f5d3dd58dbc7c061567590618a24bd595911ed089034095cc919d96
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:9aec00a487545eef76dcc4701680c5ec5bdc5e74302edbdfbc12c821b8ab4997
        version: 2.4.3
      entries:
      - name: mtv-operator.v2.4.3
        version: 2.4.3
      name: release-v2.4
    - currentCSV: mtv-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_must_gather_api": "true",
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:7dac245977fa0beeb7539b8ea5b3be68c457da132c979fd36405509fc879c7e1
          createdAt: "2023-08-31T15:03:42Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.5.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_must_gather_api": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.
          * Must-gather-api, a service to generate targeted must-gather archives.

          ### Compatibility

          Migration Toolkit for Virtualization 2.4 is supported on OpenShift 4.11 and 4.12

          Migration Toolkit for Virtualization 2.5 is supported on OpenShift 4.12 and 4.13

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:00fac23d8a514208c906ab2c911f188947368cb5e3076aacf3dc049e27b6c366
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:7dac245977fa0beeb7539b8ea5b3be68c457da132c979fd36405509fc879c7e1
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:8cc711f92258cad8f0747d3b98fef432d20661b90c4b67df8aa91650f2be3388
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-warm-rhel8@sha256:b5e0a88008c5c355844ed16b0889b0d73e63748fffafe57d0d1c52bc39c50f87
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:349cb43a620b2db9e724d874b231ef2ad896260009e3b0a93539121ac9cbe84b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:c157d6ebfb12f37a646f5f3bfbfdc90b7c8c669afe08ad342d34e474c39aa094
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:68cde15ed754189d74a528fc61cc2191861a27a4a9be112fb3ce1de3e836a77c
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-api-rhel8@sha256:c702dc948e42b9ba1e87b988f603be3df249a44dfdd6a11b0b9b49cb25f6de0c
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:55d773cad8fcf2919f5b0a5387dea62feb4ce676fb94eba310fc4d459f729d73
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:18591334cd99eb9e9f7e3d2609ff8666a9c0625e996d2f8751a14ebb9386457a
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:6097cc8e88905793bc61e48a62d77decffbbc59c1d4de83bb2eb5340ea1e8697
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:e2fca2e132bfc9a90a7c1f7e8e6b6d6c4fd8a63ec4177f5dea3f06c0c529e722
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:e665d8c5da798dc0dcd66b691db087de49564333e3c6e238d16a356c8c62d080
        version: 2.5.0
      entries:
      - name: mtv-operator.v2.5.0
        version: 2.5.0
      name: release-v2.5
    - currentCSV: mtv-operator.v2.6.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:1dfbb0993d3009608a6b272f178197f8a88edb047db4f4ff84aee8c023191d65
          createdAt: "2024-09-13T15:42:40Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.6.7'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.5 is supported on OpenShift 4.12 and 4.13

          Migration Toolkit for Virtualization 2.6 is supported on OpenShift 4.14 and 4.15

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:d34826b2a9370a1b6502d8e74a1254b2bd786684b25d90c2f0a960be0e99777d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-warm-rhel8@sha256:a46eb455027d248bcf591b612f34eb2ff4cd887cbbc3d116f78187b579ac31ab
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:3fef1c267f1f89d51ba7cd54bf39e17d7fe8ba5a4b6ecf9cdedcbe6d49d28bcc
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:1dfbb0993d3009608a6b272f178197f8a88edb047db4f4ff84aee8c023191d65
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:537f5fed16dcc46ee9cea05285a68c0e037c9b439289ff537590352865eced92
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:7dbf0f8d3d11cdb73033377552f999951331ed61d28c98cfd836903faf5601f9
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:a30fca5c5f260b94181c03a8c4c1cfe340048af87515e86571e9676e836d4b9f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:c98635ac256743873c4d74d63f4278cd26794dbaa1d56a82b15089dab510e905
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:a24acbfeca235270e75bd78046aa1dded9ff2a325a725cb46fc52ed3ff6880b8
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:403a5f8fd64926a9638342dbe568b557a7e3dc4b700cde79e64ecd9d948002a7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:6e0b0a7aca1a4afef62cdf511195e6a567feb644d3e5128c9899afc351807212
        version: 2.6.7
      entries:
      - name: mtv-operator.v2.6.7
        version: 2.6.7
      - name: mtv-operator.v2.6.6
        version: 2.6.6
      - name: mtv-operator.v2.6.5
        version: 2.6.5
      - name: mtv-operator.v2.6.4
        version: 2.6.4
      - name: mtv-operator.v2.6.3
        version: 2.6.3
      - name: mtv-operator.v2.6.2
        version: 2.6.2
      - name: mtv-operator.v2.6.1
        version: 2.6.1
      - name: mtv-operator.v2.6.0
        version: 2.6.0
      name: release-v2.6
    - currentCSV: mtv-operator.v2.7.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:db4cbc7214cdc7bf176ff16e562120111219ad8e663b7d4385dfd4c164253b7f
          createdAt: "2025-04-22T18:23:33Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.7.12'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.7 is supported on OpenShift 4.15, 4.16 and 4.17

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:a0360971c4cf448f99eb9ca79c9a153aec479e59e5985b91e019ee478382510b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:ed8cefd01f3b03df7faaf1ba0774cc18b7fc08b25f359d1e52c794b4ad60441e
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:33bae3e4c5b96df3319287d409d8d25db7db33704c7e9f8a58aec797bf20d12a
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:db4cbc7214cdc7bf176ff16e562120111219ad8e663b7d4385dfd4c164253b7f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:97ba4a4b479957e57eb296598d0c8797233f3fd0dbac20fb4686db02d3a8ed9f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:1622759bca392e2f7f66611df7988e7c63138807f917f8b34aefcc60467051ec
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:a32fba94e62f4c979e3b3596ab3a70750bd3bbf841e6004b5fd5305ac796027f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:51fdb9274911e93ae9dff9ab616a04a8a2c8fce4b6b21c05cf5f0145b1f93de0
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:a0b713409ad1d5018832cd50a2ffa3ce70ebf6cf1a4e532b3676b6f039316b52
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:ec2b67b3aff9b74f2246c310f1871198bbea8428b4e563d0e8c254a785b46a27
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:112a465da1695340286a7ff6e4dd66144209ce4389134f03813f786d30ba7508
        version: 2.7.12
      entries:
      - name: mtv-operator.v2.7.12
        version: 2.7.12
      - name: mtv-operator.v2.7.11
        version: 2.7.11
      - name: mtv-operator.v2.7.10
        version: 2.7.10
      - name: mtv-operator.v2.7.9
        version: 2.7.9
      - name: mtv-operator.v2.7.8
        version: 2.7.8
      - name: mtv-operator.v2.7.7
        version: 2.7.7
      - name: mtv-operator.v2.7.6
        version: 2.7.6
      - name: mtv-operator.v2.7.5
        version: 2.7.5
      - name: mtv-operator.v2.7.4
        version: 2.7.4
      - name: mtv-operator.v2.7.3
        version: 2.7.3
      - name: mtv-operator.v2.7.2
        version: 2.7.2
      - name: mtv-operator.v2.7.1
        version: 2.7.1
      - name: mtv-operator.v2.7.0
        version: 2.7.0
      name: release-v2.7
    - currentCSV: mtv-operator.v2.8.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "\u003cbase64-encoded-playbook\u003e"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "konveyor-forklift"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:e4145efb22dfc336d8b81f24f5b488db602d6d547479d615bf82b4fdce96c784
          createdAt: "2025-08-26T14:06:48Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: VSphere Xcopy Volume Populator
            displayName: VSphereXcopyVolumePopulator
            kind: VSphereXcopyVolumePopulator
            name: vspherexcopyvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.8 is supported on OpenShift 4.16, 4.17 and 4.18

          Migration Toolkit for Virtualization 2.9 is supported on OpenShift 4.17, 4.18 and 4.19

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/support/policy/updates/migration-toolkit-for-virtualization).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:e4145efb22dfc336d8b81f24f5b488db602d6d547479d615bf82b4fdce96c784
        version: 2.8.7
      entries:
      - name: mtv-operator.v2.8.7
        version: 2.8.7
      - name: mtv-operator.v2.8.6
        version: 2.8.6
      - name: mtv-operator.v2.8.5
        version: 2.8.5
      - name: mtv-operator.v2.8.4
        version: 2.8.4
      - name: mtv-operator.v2.8.3
        version: 2.8.3
      - name: mtv-operator.v2.8.2
        version: 2.8.2
      - name: mtv-operator.v2.8.1
        version: 2.8.1
      - name: mtv-operator.v2.8.0
        version: 2.8.0
      name: release-v2.8
    defaultChannel: release-v2.8
    packageName: mtv-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Labs
      provider-url: https://grafana.com
    name: grafana-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: grafana-operator.v5.19.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "Grafana",
                "metadata": {
                  "labels": {
                    "dashboards": "grafana-a",
                    "folders": "grafana-a"
                  },
                  "name": "grafana-a"
                },
                "spec": {
                  "config": {
                    "auth": {
                      "disable_login_form": "false"
                    },
                    "log": {
                      "mode": "console"
                    },
                    "security": {
                      "admin_password": "start",
                      "admin_user": "root"
                    }
                  }
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaAlertRuleGroup",
                "metadata": {
                  "name": "grafanaalertrulegroup-sample"
                },
                "spec": {
                  "folderRef": "test-folder-from-operator",
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "interval": "5m",
                  "rules": [
                    {
                      "condition": "B",
                      "data": [
                        {
                          "datasourceUid": "grafanacloud-demoinfra-prom",
                          "model": {
                            "datasource": {
                              "type": "prometheus",
                              "uid": "grafanacloud-demoinfra-prom"
                            },
                            "editorMode": "code",
                            "expr": "weather_temp_c{}",
                            "instant": true,
                            "intervalMs": 1000,
                            "legendFormat": "__auto",
                            "maxDataPoints": 43200,
                            "range": false,
                            "refId": "A"
                          },
                          "refId": "A",
                          "relativeTimeRange": {
                            "from": 600
                          }
                        },
                        {
                          "datasourceUid": "__expr__",
                          "model": {
                            "conditions": [
                              {
                                "evaluator": {
                                  "params": [
                                    0
                                  ],
                                  "type": "gt"
                                },
                                "operator": {
                                  "type": "and"
                                },
                                "query": {
                                  "params": [
                                    "C"
                                  ]
                                },
                                "reducer": {
                                  "params": [],
                                  "type": "last"
                                },
                                "type": "query"
                              }
                            ],
                            "datasource": {
                              "type": "__expr__",
                              "uid": "__expr__"
                            },
                            "expression": "A",
                            "intervalMs": 1000,
                            "maxDataPoints": 43200,
                            "refId": "B",
                            "type": "threshold"
                          },
                          "refId": "B",
                          "relativeTimeRange": {
                            "from": 600
                          }
                        }
                      ],
                      "execErrState": "Error",
                      "for": "5m0s",
                      "noDataState": "NoData",
                      "title": "Temperature below freezing",
                      "uid": "4843de5c-4f8a-4af0-9509-23526a04faf8"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaContactPoint",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "grafana-operator",
                    "app.kubernetes.io/instance": "grafanacontactpoint-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "grafanacontactpoint",
                    "app.kubernetes.io/part-of": "grafana-operator"
                  },
                  "name": "grafanacontactpoint-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "name": "grafanacontactpoint-sample",
                  "settings": {
                    "email": null
                  },
                  "type": "email"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaDashboard",
                "metadata": {
                  "name": "grafanadashboard-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "json": "{\n\n  \"id\": null,\n  \"title\": \"Simple Dashboard\",\n  \"tags\": [],\n  \"style\": \"dark\",\n  \"timezone\": \"browser\",\n  \"editable\": true,\n  \"hideControls\": false,\n  \"graphTooltip\": 1,\n  \"panels\": [],\n  \"time\": {\n    \"from\": \"now-6h\",\n    \"to\": \"now\"\n  },\n  \"timepicker\": {\n    \"time_options\": [],\n    \"refresh_intervals\": []\n  },\n  \"templating\": {\n    \"list\": []\n  },\n  \"annotations\": {\n    \"list\": []\n  },\n  \"refresh\": \"5s\",\n  \"schemaVersion\": 17,\n  \"version\": 0,\n  \"links\": []\n}\n"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaDatasource",
                "metadata": {
                  "name": "grafanadatasource-sample"
                },
                "spec": {
                  "datasource": {
                    "access": "proxy",
                    "isDefault": true,
                    "jsonData": {
                      "timeInterval": "5s",
                      "tlsSkipVerify": true
                    },
                    "name": "prometheus",
                    "type": "prometheus",
                    "url": "http://prometheus-service:9090"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "plugins": [
                    {
                      "name": "grafana-clock-panel",
                      "version": "1.3.0"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaFolder",
                "metadata": {
                  "name": "grafanafolder-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "title": "Example Folder"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaLibraryPanel",
                "metadata": {
                  "name": "grafana-library-panel-inline-envs"
                },
                "spec": {
                  "envs": [
                    {
                      "name": "CUSTOM_RANGE_ENV",
                      "value": "now - 12h"
                    }
                  ],
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "jsonnet": "local myRange = std.extVar('CUSTOM_RANGE_ENV'); {\n\n  model: {}\n}\n",
                  "plugins": [
                    {
                      "name": "grafana-piechart-panel",
                      "version": "1.3.9"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaMuteTiming",
                "metadata": {
                  "name": "mutetiming-sample"
                },
                "spec": {
                  "editable": false,
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "name": "mutetiming-sample",
                  "time_intervals": [
                    {
                      "days_of_month": [
                        "1",
                        "15"
                      ],
                      "location": "Asia/Shanghai",
                      "times": [
                        {
                          "end_time": "06:00",
                          "start_time": "00:00"
                        }
                      ],
                      "weekdays": [
                        "saturday"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationPolicy",
                "metadata": {
                  "name": "grafananotificationpolicy-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "route": {
                    "group_by": [
                      "grafana_folder",
                      "alertname"
                    ],
                    "receiver": "Grafana Cloud OnCall",
                    "routes": [
                      {
                        "object_matchers": [
                          [
                            "foo",
                            "=",
                            "bar"
                          ]
                        ],
                        "receiver": "grafana-default-email",
                        "routes": [
                          {
                            "object_matchers": [
                              [
                                "severity",
                                "=",
                                "critical"
                              ]
                            ],
                            "receiver": "Grafana Cloud OnCall"
                          }
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationPolicyRoute",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "grafana-operator",
                    "app.kubernetes.io/instance": "grafananotificationpolicyroute-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "grafananotificationpolicyroute",
                    "app.kubernetes.io/part-of": "grafana-operator"
                  },
                  "name": "grafananotificationpolicyroute-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationTemplate",
                "metadata": {
                  "name": "test"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "name": "test",
                  "template": "{{ define \"SlackAlert\" }}\n  [{{.Status}}] {{ .Labels.alertname }}\n  {{ .Annotations.AlertValues }}\n{{ end }}\n\n{{ define \"SlackAlertMessage\" }}\n  {{ if gt (len .Alerts.Firing) 0 }}\n    {{ len .Alerts.Firing }} firing:\n    {{ range .Alerts.Firing }} {{ template \"SlackAlert\" . }} {{ end }}\n  {{ end }}\n  {{ if gt (len .Alerts.Resolved) 0 }}\n    {{ len .Alerts.Resolved }} resolved:\n    {{ range .Alerts.Resolved }} {{ template \"SlackAlert\" . }} {{ end }}\n  {{ end }}\n{{ end }}\n\n{{ template \"SlackAlertMessage\" . }}\n"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: ghcr.io/grafana/grafana-operator@sha256:b678bc3faad11f4ef53c240e2a650ce939b9e8920dd72a47681b2efcadc7b837
          createdAt: "2025-08-21T08:38:00Z"
          description: Deploys and manages Grafana instances, dashboards and data
            sources
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/grafana/grafana-operator
          support: Grafana Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GrafanaAlertRuleGroup is the Schema for the grafanaalertrulegroups
              API
            displayName: Grafana Alert Rule Group
            kind: GrafanaAlertRuleGroup
            name: grafanaalertrulegroups.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaContactPoint is the Schema for the grafanacontactpoints
              API
            displayName: Grafana Contact Point
            kind: GrafanaContactPoint
            name: grafanacontactpoints.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaDashboard is the Schema for the grafanadashboards
              API
            displayName: Grafana Dashboard
            kind: GrafanaDashboard
            name: grafanadashboards.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaDatasource is the Schema for the grafanadatasources
              API
            displayName: Grafana Datasource
            kind: GrafanaDatasource
            name: grafanadatasources.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaFolder is the Schema for the grafanafolders API
            displayName: Grafana Folder
            kind: GrafanaFolder
            name: grafanafolders.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaLibraryPanel
            name: grafanalibrarypanels.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaMuteTiming
            name: grafanamutetimings.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaNotificationPolicy is the Schema for the GrafanaNotificationPolicy
              API
            displayName: Grafana Notification policy
            kind: GrafanaNotificationPolicy
            name: grafananotificationpolicies.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaNotificationPolicyRoute is the Schema for the grafananotificationpolicyroutes
              API
            displayName: Grafana Notification Policy Route
            kind: GrafanaNotificationPolicyRoute
            name: grafananotificationpolicyroutes.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaNotificationTemplate
            name: grafananotificationtemplates.grafana.integreatly.org
            version: v1beta1
          - description: Grafana is the Schema for the grafanas API
            displayName: Grafana
            kind: Grafana
            name: grafanas.grafana.integreatly.org
            version: v1beta1
        description: Deploys and manages Grafana instances, dashboards and data sources
        displayName: Grafana Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Grafana
        - Metrics
        - Observability
        links:
        - name: Grafana Operator
          url: https://grafana.github.io/grafana-operator
        maintainers:
        - email: grafana-operator@grafana.com
          name: Grafana-operator maintainers
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Grafana Labs
          url: https://grafana.com
        relatedImages:
        - ghcr.io/grafana/grafana-operator@sha256:b678bc3faad11f4ef53c240e2a650ce939b9e8920dd72a47681b2efcadc7b837
        - docker.io/grafana/grafana@sha256:6ac590e7cabc2fbe8d7b8fc1ce9c9f0582177b334e0df9c927ebd9670469440f
        version: 5.19.4
      entries:
      - name: grafana-operator.v5.19.4
        version: 5.19.4
      - name: grafana-operator.v5.18.0
        version: 5.18.0
      - name: grafana-operator.v5.17.1
        version: 5.17.1
      - name: grafana-operator.v5.17.0
        version: 5.17.0
      - name: grafana-operator.v5.16.0
        version: 5.16.0
      - name: grafana-operator.v5.15.1
        version: 5.15.1
      - name: grafana-operator.v5.14.0
        version: 5.14.0
      name: v5
    defaultChannel: v5
    packageName: grafana-operator
    provider:
      name: Grafana Labs
      url: https://grafana.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: PingCAP
      provider-url: ""
    name: tidb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tidb-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pingcap.com/v1alpha1",
                "kind": "TidbCluster",
                "metadata": {
                  "name": "basic"
                },
                "spec": {
                  "version": "v5.4.0",
                  "timezone": "UTC",
                  "pvReclaimPolicy": "Delete",
                  "discovery": {},
                  "pd": {
                    "baseImage": "pingcap/pd",
                    "replicas": 1,
                    "requests": {
                      "storage": "1Gi"
                    },
                    "config": {}
                  },
                  "tikv": {
                    "baseImage": "pingcap/tikv",
                    "replicas": 1,
                    "requests": {
                      "storage": "1Gi"
                    },
                    "config": {
                      "storage": {
                        "reserve-space": "0MB"
                      }
                    }
                  },
                  "tidb": {
                    "baseImage": "pingcap/tidb",
                    "replicas": 1,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "config": {}
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: pingcap/tidb-operator:v1.3.1
          createdAt: 2021-2-15T07-18-39Z
          description: TiDB Operator manages TiDB clusters on Kubernetes and automates
            tasks related to operating a TiDB cluster. It makes TiDB a truly cloud-native
            database.
          repository: https://github.com/pingcap/tidb-operator
          support: KanShiori
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup
            displayName: Backup
            kind: Backup
            name: backups.pingcap.com
            version: v1alpha1
          - description: Backup Schedule
            displayName: BackupSchedule
            kind: BackupSchedule
            name: backupschedules.pingcap.com
            version: v1alpha1
          - description: DM Cluster
            displayName: DMCluster
            kind: DMCluster
            name: dmclusters.pingcap.com
            version: v1alpha1
          - description: Restore
            displayName: Restore
            kind: Restore
            name: restores.pingcap.com
            version: v1alpha1
          - description: Tidb Cluster Auto Scaler
            displayName: TidbClusterAutoScaler
            kind: TidbClusterAutoScaler
            name: tidbclusterautoscalers.pingcap.com
            version: v1alpha1
          - description: Tidb Cluster
            displayName: TidbCluster
            kind: TidbCluster
            name: tidbclusters.pingcap.com
            version: v1alpha1
          - description: Tidb Initializer
            displayName: TidbInitializer
            kind: TidbInitializer
            name: tidbinitializers.pingcap.com
            version: v1alpha1
          - description: Tidb Monitor
            displayName: TidbMonitor
            kind: TidbMonitor
            name: tidbmonitors.pingcap.com
            version: v1alpha1
          - description: Tidb NG Monitoring
            displayName: TidbNGMonitoring
            kind: TidbNGMonitoring
            name: tidbngmonitorings.pingcap.com
            version: v1alpha1
        description: |
          ## About the managed application
          TiDB ("Ti" stands for Titanium) is an open-source, distributed, NewSQL database that supports Hybrid Transactional and Analytical Processing (HTAP) workloads. It is MySQL compatible and features horizontal scalability, strong consistency, and high availability. TiDB can be deployed on-premise or in-cloud.
          ## About this Operator

          TiDB Operator manages TiDB clusters on Kubernetes and automates tasks related to operating a TiDB cluster. It makes TiDB a truly cloud-native database.

          ## Features

          * Safely scaling the TiDB cluster

          TiDB Operator empowers TiDB with horizontal scalability on the cloud.

          * Rolling update of the TiDB cluster

          Gracefully perform rolling updates for the TiDB cluster in order, achieving zero-downtime of the TiDB cluster.

          * Multi-tenant support

          Users can deploy and manage multiple TiDB clusters on a single Kubernetes cluster easily.

          * Automatic failover

          TiDB Operator automatically performs failover for your TiDB cluster when node failures occur.

          * Kubernetes package manager support

          By embracing Kubernetes package manager Helm, users can easily deploy TiDB clusters with only one command.

          * Automatically monitoring TiDB cluster at creating

          Automatically deploy Prometheus, Grafana for TiDB cluster monitoring.

          ## Quick Start
          You can follow our [Get Started](https://docs.pingcap.com/tidb-in-kubernetes/stable/get-started) guide to quickly start a testing Kubernetes cluster and play with TiDB Operator on your own machine.

          ## Documentation

          You can see our documentation at PingCAP website for more in-depth installation and instructions for production:

          - [English](https://docs.pingcap.com/tidb-in-kubernetes/stable)
          - [简体中文](https://docs.pingcap.com/zh/tidb-in-kubernetes/stable)

          All the TiDB Operator documentation is maintained in the [docs-tidb-operator repository](https://github.com/pingcap/docs-tidb-operator).

          ## Community

          Feel free to reach out if you have any questions. The maintainers of this project are reachable via:

          - [TiDB Community Slack](https://pingcap.com/tidbslack/) in the [#sig-k8s](https://tidbcommunity.slack.com/archives/CHD0HA3LZ) channel
          - [Filing an issue](https://github.com/pingcap/tidb-operator/issue) against this repo

          Pull Requests are welcome! Check the [issue tracker](https://github.com/pingcap/tidb-operator/issue) for `status/help-wanted` issues if you're unsure where to start.

          If you're planning a new feature, please file an issue or join [#sig-k8s](https://tidbcommunity.slack.com/archives/CHD0HA3LZ) channel to discuss first.
          ## Prerequisites for enabling this Operator
        displayName: TiDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TiDB
        - SQL
        - Database
        links:
        - name: GitHub
          url: https://github.com/pingcap/tidb-operator
        - name: Documentation
          url: https://docs.pingcap.com/tidb-in-kubernetes/stable
        maintainers:
        - email: zhanghailong@pingcap.com
          name: Daniel Zhang
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: PingCAP
        relatedImages:
        - pingcap/tidb-operator:v1.3.1
        version: 1.3.1
      entries:
      - name: tidb-operator.v1.3.1
        version: 1.3.1
      name: stable
    defaultChannel: stable
    packageName: tidb-operator
    provider:
      name: PingCAP
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v0.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: gcr.io/datadoghq/operator:0.8.4
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          olm.skipRange: <0.8.4
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - gcr.io/datadoghq/operator:0.8.4
        version: 0.8.4
      entries:
      - name: datadog-operator.v0.8.4
        version: 0.8.4
      - name: datadog-operator.v0.8.2
        version: 0.8.2
      - name: datadog-operator.v0.8.1
        version: 0.8.1
      - name: datadog-operator.v0.8.0
        version: 0.8.0
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: gcr.io/datadoghq/operator:1.18.0
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - gcr.io/datadoghq/operator:1.18.0
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.2
        version: 1.4.2
      - name: datadog-operator.v1.4.1
        version: 1.4.1
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      - name: datadog-operator.v1.2.0
        version: 1.2.0
      - name: datadog-operator.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: datadog-operator
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: servicemeshoperator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: servicemeshoperator.v2.6.10
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"maistra.io/v2\",\n    \"kind\":
            \"ServiceMeshControlPlane\",\n    \"metadata\": {\n      \"name\": \"basic\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"version\": \"v2.6\",\n      \"policy\": {\n         \"type\":
            \"Istiod\"\n      },\n      \"telemetry\": {\n         \"type\": \"Istiod\"\n
            \     },\n      \"addons\": {\n        \"prometheus\": {\n          \"enabled\":
            true\n        },\n        \"kiali\": {\n          \"enabled\": true\n
            \       },\n        \"grafana\": {\n          \"enabled\": true\n        }\n
            \     }\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n    \"kind\":
            \"ServiceMeshMemberRoll\",\n    \"metadata\": {\n      \"name\": \"default\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"members\": [\n        \"your-project\",\n        \"another-of-your-projects\"
            \n      ]\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n
            \   \"kind\": \"ServiceMeshMember\",\n    \"metadata\": {\n      \"name\":
            \"default\",\n      \"namespace\": \"application-namespace\"\n    },\n
            \   \"spec\": {\n      \"controlPlaneRef\": {\n        \"name\": \"basic\",\n
            \       \"namespace\": \"control-plane-namespace\"\n      }\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
          createdAt: 2025-08-11T09:32:44CEST
          description: The OpenShift Service Mesh 2 Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh 2.6
            and earlier releases. OpenShift Service Mesh 2 is based on the open source
            Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2.6.10-0'
          operators.openshift.io/infrastructure-features: '["Disconnected","fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/maistra/istio-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Istio control plane installation
            displayName: Istio Service Mesh Control Plane
            kind: ServiceMeshControlPlane
            name: servicemeshcontrolplanes.maistra.io
            version: v2
          - description: Marks the containing namespace as a member of the referenced
              Service Mesh
            displayName: Istio Service Mesh Member
            kind: ServiceMeshMember
            name: servicemeshmembers.maistra.io
            version: v1
          - description: A list of namespaces in Service Mesh
            displayName: Istio Service Mesh Member Roll
            kind: ServiceMeshMemberRoll
            name: servicemeshmemberrolls.maistra.io
            version: v1
        description: |-
          Red Hat OpenShift Service Mesh 2 is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh 2, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh 2
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh 2 provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh 2 supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### Joining Projects Into a Mesh

          Once an instance of Red Hat OpenShift Service Mesh 2 has been installed, it will only exercise control over services within its own
          project.  Other projects may be added into the mesh using one of two methods:

          1. A **ServiceMeshMember** resource may be created in other projects to add those projects into the mesh.  The
            **ServiceMeshMember** specifies a reference to the **ServiceMeshControlPlane** object that was used to install
            the control plane.  The user creating the **ServiceMeshMember** resource must have permission to *use* the
            **ServiceMeshControlPlane** object.  The adminstrator for the project containing the control plane can grant
            individual users or groups the *use* permissions.

          2. A **ServiceMeshMemberRoll** resource may be created in the project containing the control plane.  This resource
            contains a single *members* list of all the projects that should belong in the mesh.  The resource must be named
            *default*.  The user creating the resource must have *edit* or *admin* permissions for all projects in the
            *members* list.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 2
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - maistra
        - servicemesh
        links:
        - name: Red Hat OpenShift Service Mesh 2
          url: https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html
        - name: Istio
          url: https://istio.io/
        - name: Operator Source Code
          url: https://github.com/Maistra/istio-operator
        - name: Bugs
          url: https://issues.redhat.com/projects/OSSM
        maintainers:
        - email: istio-feedback@redhat.com
          name: Red Hat, OpenShift Service Mesh
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:4ed63a5440753b1b22d44191246c82e7671e3f545c1dbdae4cb4fad163ca3a46
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6b4623c416fc85bd8e38904c16494fa3c4aeae25c90dc0da9763a0a87141bc04
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:515c5d9e6700fae981b0ea3807436b250acb1534e9ecf50043e5df8474e66dce
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:b134578e0677963197e353d8acd0ac3330df1b6d2752aa41cb072e86b5b4b91b
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6fb7d50b188dc27ffb9a396edabbfb21767fe438344fd70d4754d1ec7cf27957
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:e734ca310e09f737b4f06f3ecfeb9ac12971c63884fdcc27e3bebef9d5b18722
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:a36d5db6097749c55e946bba3b8ddd13a84b1f6f68b62d0ccbad5a803e9b46fb
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:fc500a282c6668ff987a632efbbd19ecc3aebdc021f831a7b386e360a5c8324d
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:57eb7ba03f24781501a7ceed2e160ff95317869af6c2caf8f9a83990e6a6c7b9
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:1ca80927334c6af8e386de8d44704f568b4b35d3f7afbe785a999d0fa4efd0bf
        - registry.redhat.io/openshift4/ose-prometheus@sha256:423ae2e81cfe660b7217f7662754f366a91d457d6dd1589b4fc40733896eb9d6
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:d16a60204d266bf4cc266c005fabf175b533b827500f171013884a713c88857a
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel8@sha256:3d773d16fed72278949b9e3b54dc59e346e95341dc94b715bcfc328d3c9e83b2
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:6a420228e684596ce2b900614a874f3e1284a6f2bbbba4c5a71ade158867affe
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:c77183e50a34d33b5e2fede60350f54b18ab3536c3189e331569d5df862d9979
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:a3d65983d46dcf47b94f26e7265ed32082208ce08abcf3587a9a4e15c19b274a
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:b241452cce5c9af5aa0f12d1db3a7094cc004ce355988b8a5207091810a6f84b
        - registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:3dc1bddb070fd530ce6251666989b9a849de90591718a578da4a9900f98d93ec
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:c34b4cd9a82fef40708a25f0568f6dc632878affba35d5d631a48ec31da08cd3
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:c1cadf40bdb9511d612a0257c12eb3a9738823592ebb6e32a4f51b3a8dfe08db
        - registry.redhat.io/openshift-service-mesh/prometheus-rhel8@sha256:71e193a2d403bb06a7a76686b376985315aa0c32f495b8b9dacd73c7ccce2af3
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d95762c62087072cadee5948c706619f09202a349e1310f0ed81bfb42f11203f
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:4669724b430683bf2e895c5cf941c2d99536117815999d38131e5b6f162bbef0
        version: 2.6.10-0
      entries:
      - name: servicemeshoperator.v2.6.10
        version: 2.6.10-0
      - name: servicemeshoperator.v2.6.9
        version: 2.6.9-0
      - name: servicemeshoperator.v2.6.8
        version: 2.6.8-0
      - name: servicemeshoperator.v2.6.7
        version: 2.6.7-0
      - name: servicemeshoperator.v2.6.6
        version: 2.6.6-0
      - name: servicemeshoperator.v2.6.5
        version: 2.6.5-0
      - name: servicemeshoperator.v2.6.4
        version: 2.6.4-0
      - name: servicemeshoperator.v2.6.3
        version: 2.6.3-0
      - name: servicemeshoperator.v2.6.2
        version: 2.6.2-0
      - name: servicemeshoperator.v2.6.1
        version: 2.6.1-0
      - name: servicemeshoperator.v2.6.0
        version: 2.6.0-0
      - name: servicemeshoperator.v2.5.2
        version: 2.5.2-0
      - name: servicemeshoperator.v2.5.1
        version: 2.5.1-0
      - name: servicemeshoperator.v2.5.0
        version: 2.5.0-0
      - name: servicemeshoperator.v2.4.5
        version: 2.4.5-0
      - name: servicemeshoperator.v2.4.4
        version: 2.4.4-0
      - name: servicemeshoperator.v2.4.3
        version: 2.4.3-0
      - name: servicemeshoperator.v2.4.2
        version: 2.4.2-0
      - name: servicemeshoperator.v2.4.1
        version: 2.4.1-0
      - name: servicemeshoperator.v2.4.0
        version: 2.4.0-0
      - name: servicemeshoperator.v2.3.3
        version: 2.3.3-0
      - name: servicemeshoperator.v2.3.2
        version: 2.3.2-0
      - name: servicemeshoperator.v2.3.1
        version: 2.3.1-0
      - name: servicemeshoperator.v2.3.0
        version: 2.3.0-0
      - name: servicemeshoperator.v2.2.3
        version: 2.2.3-0
      - name: servicemeshoperator.v2.2.2
        version: 2.2.2-0
      - name: servicemeshoperator.v2.2.1
        version: 2.2.1-0
      - name: servicemeshoperator.v2.2.0
        version: 2.2.0-0
      - name: servicemeshoperator.v2.1.3
        version: 2.1.3-0
      - name: servicemeshoperator.v2.1.2
        version: 2.1.2-0
      - name: servicemeshoperator.v2.1.1
        version: 2.1.1-0
      - name: servicemeshoperator.v2.1.0
        version: 2.1.0-0
      - name: servicemeshoperator.v2.0.8
        version: 2.0.8-0
      name: "1.0"
    - currentCSV: servicemeshoperator.v2.6.10
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"maistra.io/v2\",\n    \"kind\":
            \"ServiceMeshControlPlane\",\n    \"metadata\": {\n      \"name\": \"basic\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"version\": \"v2.6\",\n      \"policy\": {\n         \"type\":
            \"Istiod\"\n      },\n      \"telemetry\": {\n         \"type\": \"Istiod\"\n
            \     },\n      \"addons\": {\n        \"prometheus\": {\n          \"enabled\":
            true\n        },\n        \"kiali\": {\n          \"enabled\": true\n
            \       },\n        \"grafana\": {\n          \"enabled\": true\n        }\n
            \     }\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n    \"kind\":
            \"ServiceMeshMemberRoll\",\n    \"metadata\": {\n      \"name\": \"default\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"members\": [\n        \"your-project\",\n        \"another-of-your-projects\"
            \n      ]\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n
            \   \"kind\": \"ServiceMeshMember\",\n    \"metadata\": {\n      \"name\":
            \"default\",\n      \"namespace\": \"application-namespace\"\n    },\n
            \   \"spec\": {\n      \"controlPlaneRef\": {\n        \"name\": \"basic\",\n
            \       \"namespace\": \"control-plane-namespace\"\n      }\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
          createdAt: 2025-08-11T09:32:44CEST
          description: The OpenShift Service Mesh 2 Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh 2.6
            and earlier releases. OpenShift Service Mesh 2 is based on the open source
            Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2.6.10-0'
          operators.openshift.io/infrastructure-features: '["Disconnected","fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/maistra/istio-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Istio control plane installation
            displayName: Istio Service Mesh Control Plane
            kind: ServiceMeshControlPlane
            name: servicemeshcontrolplanes.maistra.io
            version: v2
          - description: Marks the containing namespace as a member of the referenced
              Service Mesh
            displayName: Istio Service Mesh Member
            kind: ServiceMeshMember
            name: servicemeshmembers.maistra.io
            version: v1
          - description: A list of namespaces in Service Mesh
            displayName: Istio Service Mesh Member Roll
            kind: ServiceMeshMemberRoll
            name: servicemeshmemberrolls.maistra.io
            version: v1
        description: |-
          Red Hat OpenShift Service Mesh 2 is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh 2, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh 2
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh 2 provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh 2 supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### Joining Projects Into a Mesh

          Once an instance of Red Hat OpenShift Service Mesh 2 has been installed, it will only exercise control over services within its own
          project.  Other projects may be added into the mesh using one of two methods:

          1. A **ServiceMeshMember** resource may be created in other projects to add those projects into the mesh.  The
            **ServiceMeshMember** specifies a reference to the **ServiceMeshControlPlane** object that was used to install
            the control plane.  The user creating the **ServiceMeshMember** resource must have permission to *use* the
            **ServiceMeshControlPlane** object.  The adminstrator for the project containing the control plane can grant
            individual users or groups the *use* permissions.

          2. A **ServiceMeshMemberRoll** resource may be created in the project containing the control plane.  This resource
            contains a single *members* list of all the projects that should belong in the mesh.  The resource must be named
            *default*.  The user creating the resource must have *edit* or *admin* permissions for all projects in the
            *members* list.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 2
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - maistra
        - servicemesh
        links:
        - name: Red Hat OpenShift Service Mesh 2
          url: https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html
        - name: Istio
          url: https://istio.io/
        - name: Operator Source Code
          url: https://github.com/Maistra/istio-operator
        - name: Bugs
          url: https://issues.redhat.com/projects/OSSM
        maintainers:
        - email: istio-feedback@redhat.com
          name: Red Hat, OpenShift Service Mesh
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:c1cadf40bdb9511d612a0257c12eb3a9738823592ebb6e32a4f51b3a8dfe08db
        - registry.redhat.io/openshift4/ose-prometheus@sha256:423ae2e81cfe660b7217f7662754f366a91d457d6dd1589b4fc40733896eb9d6
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:4ed63a5440753b1b22d44191246c82e7671e3f545c1dbdae4cb4fad163ca3a46
        - registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:57eb7ba03f24781501a7ceed2e160ff95317869af6c2caf8f9a83990e6a6c7b9
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:3dc1bddb070fd530ce6251666989b9a849de90591718a578da4a9900f98d93ec
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel8@sha256:3d773d16fed72278949b9e3b54dc59e346e95341dc94b715bcfc328d3c9e83b2
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:a3d65983d46dcf47b94f26e7265ed32082208ce08abcf3587a9a4e15c19b274a
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:6a420228e684596ce2b900614a874f3e1284a6f2bbbba4c5a71ade158867affe
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:c77183e50a34d33b5e2fede60350f54b18ab3536c3189e331569d5df862d9979
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:d16a60204d266bf4cc266c005fabf175b533b827500f171013884a713c88857a
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:1ca80927334c6af8e386de8d44704f568b4b35d3f7afbe785a999d0fa4efd0bf
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6b4623c416fc85bd8e38904c16494fa3c4aeae25c90dc0da9763a0a87141bc04
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:515c5d9e6700fae981b0ea3807436b250acb1534e9ecf50043e5df8474e66dce
        - registry.redhat.io/openshift-service-mesh/prometheus-rhel8@sha256:71e193a2d403bb06a7a76686b376985315aa0c32f495b8b9dacd73c7ccce2af3
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d95762c62087072cadee5948c706619f09202a349e1310f0ed81bfb42f11203f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:e734ca310e09f737b4f06f3ecfeb9ac12971c63884fdcc27e3bebef9d5b18722
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:a36d5db6097749c55e946bba3b8ddd13a84b1f6f68b62d0ccbad5a803e9b46fb
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:b134578e0677963197e353d8acd0ac3330df1b6d2752aa41cb072e86b5b4b91b
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6fb7d50b188dc27ffb9a396edabbfb21767fe438344fd70d4754d1ec7cf27957
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:4669724b430683bf2e895c5cf941c2d99536117815999d38131e5b6f162bbef0
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:fc500a282c6668ff987a632efbbd19ecc3aebdc021f831a7b386e360a5c8324d
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:c34b4cd9a82fef40708a25f0568f6dc632878affba35d5d631a48ec31da08cd3
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:b241452cce5c9af5aa0f12d1db3a7094cc004ce355988b8a5207091810a6f84b
        version: 2.6.10-0
      entries:
      - name: servicemeshoperator.v2.6.10
        version: 2.6.10-0
      - name: servicemeshoperator.v2.6.9
        version: 2.6.9-0
      - name: servicemeshoperator.v2.6.8
        version: 2.6.8-0
      - name: servicemeshoperator.v2.6.7
        version: 2.6.7-0
      - name: servicemeshoperator.v2.6.6
        version: 2.6.6-0
      - name: servicemeshoperator.v2.6.5
        version: 2.6.5-0
      - name: servicemeshoperator.v2.6.4
        version: 2.6.4-0
      - name: servicemeshoperator.v2.6.3
        version: 2.6.3-0
      - name: servicemeshoperator.v2.6.2
        version: 2.6.2-0
      - name: servicemeshoperator.v2.6.1
        version: 2.6.1-0
      - name: servicemeshoperator.v2.6.0
        version: 2.6.0-0
      - name: servicemeshoperator.v2.5.2
        version: 2.5.2-0
      - name: servicemeshoperator.v2.5.1
        version: 2.5.1-0
      - name: servicemeshoperator.v2.5.0
        version: 2.5.0-0
      - name: servicemeshoperator.v2.4.5
        version: 2.4.5-0
      - name: servicemeshoperator.v2.4.4
        version: 2.4.4-0
      - name: servicemeshoperator.v2.4.3
        version: 2.4.3-0
      - name: servicemeshoperator.v2.4.2
        version: 2.4.2-0
      - name: servicemeshoperator.v2.4.1
        version: 2.4.1-0
      - name: servicemeshoperator.v2.4.0
        version: 2.4.0-0
      - name: servicemeshoperator.v2.3.3
        version: 2.3.3-0
      - name: servicemeshoperator.v2.3.2
        version: 2.3.2-0
      - name: servicemeshoperator.v2.3.1
        version: 2.3.1-0
      - name: servicemeshoperator.v2.3.0
        version: 2.3.0-0
      - name: servicemeshoperator.v2.2.3
        version: 2.2.3-0
      - name: servicemeshoperator.v2.2.2
        version: 2.2.2-0
      - name: servicemeshoperator.v2.2.1
        version: 2.2.1-0
      - name: servicemeshoperator.v2.2.0
        version: 2.2.0-0
      - name: servicemeshoperator.v2.1.3
        version: 2.1.3-0
      - name: servicemeshoperator.v2.1.2
        version: 2.1.2-0
      - name: servicemeshoperator.v2.1.1
        version: 2.1.1-0
      - name: servicemeshoperator.v2.1.0
        version: 2.1.0-0
      - name: servicemeshoperator.v2.0.8
        version: 2.0.8-0
      name: stable
    defaultChannel: stable
    packageName: servicemeshoperator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: service-binding-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-binding-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:b355d2370926063a3de8c7a09c0ca8c2348a8d158d6a7095d0e5e8381cc95be5
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/redhat-developer/servicebinding-operator@sha256:b355d2370926063a3de8c7a09c0ca8c2348a8d158d6a7095d0e5e8381cc95be5
        version: 0.10.0
      entries:
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: beta
    - currentCSV: service-binding-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindablekinds-sample"
                }
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:c9c9e45321db2d4d3cdc2d364337af844f69c9893d0cc74d2fa56c846ffcf3da
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/redhat-developer/servicebinding-operator@sha256:c9c9e45321db2d4d3cdc2d364337af844f69c9893d0cc74d2fa56c846ffcf3da
        version: 0.11.0
      entries:
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: candidate
    - currentCSV: service-binding-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindable-kinds"
                },
                "spec": {}
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  },
                  "type": "postgres",
                  "workload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:ac47f496fb7ecdcbc371f8c809fad2687ec0c35bbc8c522a7ab63b3e5ffd90ea
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'.
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1alpha3
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1beta1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
          - description: Service Binding implementing community specification (https://github.com/k8s-service-bindings/spec)
              describing the connection between a backing service and an application
              workload.
            displayName: Service Binding (Spec API)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1beta1
        description: |
          The Service Binding Operator manages the data plane for applications and backing services. It enables developers to connect their
          applications to backing services (REST API, databases, event buses and others) with a consistent and predictable
          experience. The operator reads data made available by the control plane of backing services and projects the bindings
          data to applications either as accessible files mounted into the application’s container or environment variables.

          In Kubernetes, each service suggests a different way to access their secrets and developers will consume them in their
          application in a custom way. While this provides a good deal of flexibility, it also creates discrepancies and
          complexity dealing with each specific services, which leads large development teams to lose overall velocity.

          Service Binding Operator removes the error-prone manual configuration of binding information, provides service operators
          a low-touch administrative experience to provision and manage access to services and enriches development lifecycle.

          For additional details refer to the project [documentation](https://redhat-developer.github.io/service-binding-operator).

          Get Started with Service Binding Operator
          * Find out about SBO’s [prerequisites](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/installing-service-binding.html).
          * Refer to the [Quick Start Guide](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/quick-start.html) to see SBO in action in a simple scenario.

          Need Help
          * Raise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)

          Licence
          * Service Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE)

          Service Binding Operator is providing and supporting two different APIs.
          * `binding.operators.coreos.com/v1alpha1`: This API is compliant with the Service Binding specification. It is fully supported and considered as the mature API. It’s the API we recommend in production environments.
          * `servicebinding.io/v1beta1`: This API is the one defined in the Service Binding specification. It is available in Service Binding Operator 1.x.

          We’ll consider supporting only the API from the specification, once the Service Binding specification will be officially GA. Of course, that will be a breaking change and handle in a safe way for you.
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Documentation
          url: https://redhat-developer.github.io/service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/redhat-developer/servicebinding-operator@sha256:ac47f496fb7ecdcbc371f8c809fad2687ec0c35bbc8c522a7ab63b3e5ffd90ea
        version: 1.4.0
      entries:
      - name: service-binding-operator.v1.4.0
        version: 1.4.0
      - name: service-binding-operator.v1.3.4
        version: 1.3.4
      - name: service-binding-operator.v1.3.3
        version: 1.3.3
      - name: service-binding-operator.v1.3.2
        version: 1.3.2
      - name: service-binding-operator.v1.3.1
        version: 1.3.1
      - name: service-binding-operator.v1.3.0
        version: 1.3.0
      - name: service-binding-operator.v1.2.0
        version: 1.2.0
      - name: service-binding-operator.v1.1.1
        version: 1.1.1
      - name: service-binding-operator.v1.1.0
        version: 1.1.0
      - name: service-binding-operator.v1.0.1
        version: 1.0.1
      - name: service-binding-operator.v1.0.0
        version: 1.0.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: service-binding-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: netobserv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netobserv-operator.v1.6.3-community
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowMetric",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netobserv-operator",
                    "app.kubernetes.io/instance": "flowmetric-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "flowmetric",
                    "app.kubernetes.io/part-of": "netobserv-operator"
                  },
                  "name": "flowmetric-sample"
                },
                "spec": {
                  "charts": [
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "",
                          "promQL": "sum(rate($METRIC[2m]))"
                        }
                      ],
                      "title": "External ingress traffic",
                      "type": "SingleStat",
                      "unit": "Bps"
                    },
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "{{DstK8S_Namespace}} / {{DstK8S_OwnerName}}",
                          "promQL": "sum(rate($METRIC{DstK8S_Namespace!=\"\"}[2m])) by (DstK8S_Namespace, DstK8S_OwnerName)"
                        }
                      ],
                      "sectionName": "External",
                      "title": "Top external ingress traffic per workload",
                      "type": "StackArea",
                      "unit": "Bps"
                    }
                  ],
                  "direction": "Ingress",
                  "filters": [
                    {
                      "field": "SrcSubnetLabel",
                      "matchType": "Absence"
                    }
                  ],
                  "labels": [
                    "DstK8S_HostName",
                    "DstK8S_Namespace",
                    "DstK8S_OwnerName",
                    "DstK8S_OwnerType"
                  ],
                  "metricName": "cluster_external_ingress_bytes_total",
                  "type": "Counter",
                  "valueField": "Bytes"
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "app"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "infra"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "enable": true,
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "includeList": [
                        "node_ingress_bytes_total",
                        "workload_ingress_bytes_total",
                        "namespace_flows_total"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta2",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "metrics": {
                        "server": {
                          "port": 9400
                        }
                      },
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "eBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "Disabled"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "\"app\""
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "\"infra\""
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "\"Pod\"",
                          "src_kind": "\"Pod\""
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "\"Service\""
                        },
                        "name": "Services network"
                      }
                    ],
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    }
                  },
                  "deploymentModel": "Direct",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "enable": true,
                    "lokiStack": {
                      "name": "loki"
                    },
                    "mode": "Monolithic",
                    "monolithic": {
                      "tenantID": "netobserv",
                      "tls": {
                        "caCert": {
                          "certFile": "service-ca.crt",
                          "name": "loki-gateway-ca-bundle",
                          "type": "configmap"
                        },
                        "enable": false
                      },
                      "url": "http://loki.netobserv.svc:3100/"
                    },
                    "readTimeout": "30s",
                    "writeBatchSize": 10485760,
                    "writeBatchWait": "1s",
                    "writeTimeout": "10s"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "Flows",
                    "metrics": {
                      "disableAlerts": [],
                      "server": {
                        "port": 9401
                      }
                    },
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "prometheus": {
                    "querier": {
                      "enable": true,
                      "mode": "Auto",
                      "timeout": "30s"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:1.6.2-community
          createdAt: "2024-08-22T18:26:56Z"
          description: Network flows collector and monitoring solution
          operatorframework.io/initialization-resource: '{"apiVersion":"flows.netobserv.io/v1beta2",
            "kind":"FlowCollector","metadata":{"name":"cluster"},"spec": {}}'
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
          support: NetObserv team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta2
          - description: '`FlowMetric` is the schema for the custom metrics API, which
              allows to generate more metrics out of flow logs. You can find examples
              here: https://github.com/netobserv/network-observability-operator/tree/1.6.2-community/config/samples/flowmetrics'
            displayName: Flow Metric
            kind: FlowMetric
            name: flowmetrics.flows.netobserv.io
            version: v1alpha1
        description: |-
          DEPRECATED: this distribution channel is deprecated and doesn't receive new updates. Please read https://github.com/orgs/netobserv/discussions/996 to learn more.

          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline that consists in:
          - an eBPF agent, that generates network flows from captured packets
          - flowlogs-pipeline, a component that collects, enriches and exports these flows
          - when used in OpenShift, a Console plugin for flows visualization with powerful filtering options, a topology representation and more

          Flow data is then available in multiple ways, each optional:

          - As Prometheus metrics
          - As raw flow logs stored in Grafana Loki
          - As raw flow logs exported to a collector

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, can optionally be used as the backend to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 10GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          If you prefer to not use Loki, you must set `spec.loki.enable` to `false` in `FlowCollector`.
          In that case, you can still get the Prometheus metrics or export raw flows to a custom collector. But be aware that some of the Console plugin features will be disabled. For instance, you will not be able to view raw flows there, and the metrics / topology will have a more limited level of details, missing information such as pods or IPs.

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for example [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/config/samples/flows_v1beta2_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide on every node, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you might have to configure differently if you used another installation method. Make sure to disable it (`spec.loki.enable`) if you don't want to use Loki.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`) an optional list of exporters to which to send enriched flows. KAFKA and IPFIX exporters are supported. This allows you to define any custom storage or processing that can read from Kafka or use the IPFIX standard.

          - To enable availability zones awareness, set `spec.processor.addZone` to `true`.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - More information on [configuring metrics](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/Metrics.md).
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: '[DEPRECATED] NetObserv Operator'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/network-observability-operator:1.6.2-community
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/netobserv/netobserv-ebpf-agent:v1.6.2-community
        - quay.io/netobserv/flowlogs-pipeline:v1.6.2-community
        - quay.io/netobserv/network-observability-console-plugin:v1.6.2-community
        version: 1.6.3-community
      entries:
      - name: netobserv-operator.v1.6.3-community
        version: 1.6.3-community
      - name: netobserv-operator.v1.6.2-community
        version: 1.6.2-community
      - name: netobserv-operator.v1.6.1-community
        version: 1.6.1-community
      - name: netobserv-operator.v1.6.0-community
        version: 1.6.0-community
      - name: netobserv-operator.v1.5.0-community
        version: 1.5.0-community
      name: community
    - currentCSV: netobserv-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "ipfix": {
                      "cacheActiveTimeout": "20s",
                      "cacheMaxFlows": 400,
                      "clusterNetworkOperator": {
                        "namespace": "openshift-network-operator"
                      },
                      "ovnKubernetes": {
                        "containerName": "ovnkube-node",
                        "daemonSetName": "ovnkube-node",
                        "namespace": "ovn-kubernetes"
                      },
                      "sampling": 400
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "staticLabels": {
                      "app": "netobserv-flowcollector"
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "enableKubeProbes": true,
                    "healthPort": 8080,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:0.2.1
          createdAt: "2022-12-07T09:00:48Z"
          description: Network flows collector and monitoring solution
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
        description: |-
          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline to collect and enrich network flows. These flows can be produced by the NetObserv eBPF agent, or by any device or CNI able to export flows in IPFIX format, such as OVN-Kubernetes.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When used in OpenShift, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For a normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionnaly be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/0.2.1/config/samples/flows_v1alpha1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Agent (`spec.agent.type`) can be `EBPF` (default) or `IPFIX`. eBPF is recommended, as it should work in more situations and offers better performances. If you can't, or don't want to use eBPF, note that the IPFIX option is fully functional only when using [OVN-Kubernetes](https://github.com/ovn-org/ovn-kubernetes/) CNI. Other CNIs are not officially supported, but you may still be able to configure them manually if they allow IPFIX exports.

          - Sampling (`spec.agent.ebpf.sampling` and `spec.agent.ipfix.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50) for eBPF and 400 (1:400) for IPFIX. Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub more more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - A [F.A.Q](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: NetObserv Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: mmaciasl@redhat.com
          name: Mario Macias
        - email: ocazade@redhat.com
          name: Olivier Cazade
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/netobserv-ebpf-agent:v0.2.2
        - quay.io/netobserv/flowlogs-pipeline:v0.1.5
        - quay.io/netobserv/network-observability-console-plugin:v0.1.6
        - quay.io/netobserv/network-observability-operator:0.2.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        version: 0.2.1
      entries:
      - name: netobserv-operator.v0.2.1
        version: 0.2.1
      - name: netobserv-operator.v0.2.0
        version: 0.2.0
      - name: netobserv-operator.v0.1.4
        version: 0.1.4
      - name: netobserv-operator.v0.1.3
        version: 0.1.3
      - name: netobserv-operator.v0.1.2
        version: 0.1.2
      - name: netobserv-operator.v0.1.1
        version: 0.1.1
      - name: netobserv-operator.v0.1.0
        version: 0.1.0
      name: latest
    - currentCSV: netobserv-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "ipfix": {
                      "cacheActiveTimeout": "20s",
                      "cacheMaxFlows": 400,
                      "clusterNetworkOperator": {
                        "namespace": "openshift-network-operator"
                      },
                      "ovnKubernetes": {
                        "containerName": "ovnkube-node",
                        "daemonSetName": "ovnkube-node",
                        "namespace": "ovn-kubernetes"
                      },
                      "sampling": 400
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "staticLabels": {
                      "app": "netobserv-flowcollector"
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "enableKubeProbes": true,
                    "healthPort": 8080,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:0.2.1
          createdAt: "2022-12-07T09:00:48Z"
          description: Network flows collector and monitoring solution
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
        description: |-
          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline to collect and enrich network flows. These flows can be produced by the NetObserv eBPF agent, or by any device or CNI able to export flows in IPFIX format, such as OVN-Kubernetes.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When used in OpenShift, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For a normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionnaly be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/0.2.1/config/samples/flows_v1alpha1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Agent (`spec.agent.type`) can be `EBPF` (default) or `IPFIX`. eBPF is recommended, as it should work in more situations and offers better performances. If you can't, or don't want to use eBPF, note that the IPFIX option is fully functional only when using [OVN-Kubernetes](https://github.com/ovn-org/ovn-kubernetes/) CNI. Other CNIs are not officially supported, but you may still be able to configure them manually if they allow IPFIX exports.

          - Sampling (`spec.agent.ebpf.sampling` and `spec.agent.ipfix.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50) for eBPF and 400 (1:400) for IPFIX. Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub more more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - A [F.A.Q](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: NetObserv Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: mmaciasl@redhat.com
          name: Mario Macias
        - email: ocazade@redhat.com
          name: Olivier Cazade
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/network-observability-operator:0.2.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/netobserv/network-observability-console-plugin:v0.1.6
        - quay.io/netobserv/netobserv-ebpf-agent:v0.2.2
        - quay.io/netobserv/flowlogs-pipeline:v0.1.5
        version: 0.2.1
      entries:
      - name: netobserv-operator.v0.2.1
        version: 0.2.1
      - name: netobserv-operator.v0.2.0
        version: 0.2.0
      - name: netobserv-operator.v0.1.4
        version: 0.1.4
      - name: netobserv-operator.v0.1.3
        version: 0.1.3
      - name: netobserv-operator.v0.1.2
        version: 0.1.2
      - name: netobserv-operator.v0.1.1
        version: 0.1.1
      - name: netobserv-operator.v0.1.0
        version: 0.1.0
      name: v0.2.x
    defaultChannel: community
    packageName: netobserv-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-license-service-scanner-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-license-service-scanner-operator.v4.2.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1",
                "kind": "IBMLicenseServiceScanner",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-licensing-scanner-operator",
                    "app.kubernetes.io/instance": "ibm-licensing-scanner-instance",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ibm-licensing-scanner-instance",
                    "app.kubernetes.io/part-of": "ibm-licensing-scanner-operator"
                  },
                  "name": "ibm-licensing-scanner-instance"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "license-service-upload-secret": "ibm-licensing-scanner-upload-secret",
                  "log-level": "INFO",
                  "registry-pull-secret": "ibm-licensing-scanner-registry-pull-secret",
                  "scan": {
                    "namespaces": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:77c9db26a13dba25438a6fc5a41290606418481e08ea2ecc0a9a397062af74fb
          createdAt: "2025-05-14T12:59:00Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <4.2.16'
          operatorframework.io/suggested-namespace: ibm-licensing-scanner
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMLicenseServiceScanner is the Schema for the ibmlicenseservicescanners
              API
            displayName: IBMLicense Service Scanner
            kind: IBMLicenseServiceScanner
            name: ibmlicenseservicescanners.operator.ibm.com
            version: v1
        description: IBM License Service Scanner
        displayName: IBM License Service Scanner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - licensing
        - usage
        - IBMLicenseServiceScanner
        - IBM
        - Cloud
        - License
        - License Service
        - Service
        - Scanner
        - License Service Scanner
        - License Scanner
        - Service Scanner
        links:
        - name: Ibm License Service Scanner Operator
          url: https://ibm-license-service-scanner-operator.domain
        maintainers:
        - email: talk2sam@us.ibm.com
          name: IBM
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-bundle@sha256:b6cb9e5976b749dd7818ac19d1cfd5e24673f3c052a41d6e2fe1d2d7094694b4
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-catalog@sha256:c016904e8c5cfcf8270241ad5e28b52f528859984ed894cdd5e1283fa90cd6e8
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:77c9db26a13dba25438a6fc5a41290606418481e08ea2ecc0a9a397062af74fb
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:f3503f72a90ff8abbadaef3e78a16db6e1fbff4c90b475c33ead17bf1e69990c
        version: 4.2.16
      entries:
      - name: ibm-license-service-scanner-operator.v4.2.16
        version: 4.2.16
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-license-service-scanner-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/component: operator
      app.kubernetes.io/managed-by: kustomize
      app.kubernetes.io/name: zabbix-operator-certified
      app.kubernetes.io/part-of: zabbix-operator-certified
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zabbix LLC
      provider-url: ""
    name: zabbix-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zabbix-operator-certified.v6.0.40
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent",
                "metadata": {
                  "labels": {
                    "app": "agent",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "log_remote_commands": true,
                  "max_lines_per_second": 20,
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent2",
                "metadata": {
                  "labels": {
                    "app": "agent2",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent2"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "enable_persistent_buffer": false,
                  "enable_status_port": false,
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "persistent_buffer_claim": "",
                  "persistent_buffer_period": "1h",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixFull",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-full"
                },
                "spec": {
                  "builtin_report_service": true,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysql_volumeclaim": "mysql-full-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxyMysql",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-mysql"
                },
                "spec": {
                  "db_server_port": 3306,
                  "enable_snmp_traps": false,
                  "internal_db": true,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix_proxy",
                  "proxy": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "zabbix_mysql_volumeclaim": "mysql-proxy-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxySqlite",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-sqlite"
                },
                "spec": {
                  "db_data_volumeclaim": "zabbix-proxy-sqlite",
                  "enable_snmp_traps": false,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "proxy": {
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "node_name_as_db_name": true,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "proxy_size": 1,
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  }
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixServer",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-server"
                },
                "spec": {
                  "builtin_report_service": true,
                  "db_server_host": "mysql-server",
                  "db_server_port": 3306,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix",
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              }
            ]
          ansible.operator-sdk/reconcile-period: 5s
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          certified: "true"
          containerImage: registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
          createdAt: "2025-05-26T07:41:22Z"
          description: Zabbix operator with multiple deployment variants and different
            components
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <6.0.40
          operatorframework.io/suggested-namespace: zabbix
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: ""
          support: Zabbix
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Zabbix agent 2 is a new generation of Zabbix agent and may
              be used in place of Zabbix agent
            displayName: Zabbix agent 2
            kind: ZabbixAgent2
            name: zabbixagent2s.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix agent is deployed on a monitoring nodes to actively
              monitor local resources and applications
            displayName: Zabbix agent
            kind: ZabbixAgent
            name: zabbixagents.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database server, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Full
            kind: ZabbixFull
            name: zabbixfulls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with MySQL database server
            displayName: Zabbix proxy (MySQL)
            kind: ZabbixProxyMysql
            name: zabbixproxymysqls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with SQLite database
            displayName: Zabbix proxy (SQLite3)
            kind: ZabbixProxySqlite
            name: zabbixproxysqlites.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database support, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Server
            kind: ZabbixServer
            name: zabbixservers.kubernetes.zabbix.com
            version: v1alpha1
        description: "## About this Operator\n\nZabbix helps you to real-time monitoring
          of millions of metrics collected from tens of thousands of servers, virtual
          machines and network devices.\nThe Zabbix Operator allows users to easily
          deploy, manage, and maintain Zabbix deployments on OpenShift. By installing
          this integration you will be able to deploy Zabbix server / proxies and
          other components with a single command.\n## Supported Features\n* **Zabbix
          Server** - Simple Zabbix installation with included Zabbix server, Zabbix
          web-interface and Zabbix Java Gateway with MySQL database support. The feature
          does not provide MySQL service and requires an external MySQL database.
          \n* **Zabbix Server (Full)** - Zabbix installation with included Zabbix
          server, Zabbix web-interface, Zabbix Java Gateway and MySQL server instance.\n*
          **Zabbix proxy (SQLite3)** - Very simple way to gain power of Zabbix proxy.
          The feature has  SQLite3 support for Zabbix proxies and allow to specify
          amount of proxies. \n* **Zabbix proxy (MySQL)** - Another option of Zabbix
          proxy. The option support and deliver MySQL database.\n* **Zabbix agent**
          - Zabbix agent can be deployed on each available node for stability and
          performance monitoring on remote nodes. It allows to gather metrics with
          full automation!* \n## Prerequisites\nAll deployment options are require
          additional information during deployment. Please, check the following instructions
          and provide required configuration:\n* **Zabbix Server** - MySQL database
          host information and MySQL database credentials in specially formatted *Secret*.
          Additionally it is possible to specify SSL certificates for HTTPS support
          in *Secret*.\n* **Zabbix Server (Full)** - MySQL database credentials in
          specially formatted *Secret*. MySQL database volume name information. Additionally
          it is possible to specify SSL certificates for HTTPS support in *Secret*.\n*
          **Zabbix proxy (SQLite3)** - Zabbix server host information only.\n* **Zabbix
          proxy (MySQL)** - MySQL database credentials in specially formatted *Secrets*
          and Zabbix server host information.\n* **Zabbix agent** - Zabbix server
          host information only for outgoing and incoming connections. Zabbix agent
          uses \"privileged\" mode to monitor node resources! For example, running
          processes.* \n### MySQL credentials\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-full-secrets\ndata:\n  mysql_root_password: emFiYml4X3N1cGVyX3Jvb3Q=
          [1]\n  mysql_zabbix_password: emFiYml4X3VzZXJfcGFzc3dvcmQ= [2]\n  mysql_zabbix_username:
          emFiYml4 [3]\ntype: Opaque\n```\nUsing MySQL root password (*mysqlrootpassword*
          [1]) Zabbix server / proxy will try to create MySQL database schema with
          grant permissions to *mysqlzabbixusername* [2] and *mysqlzabbixpassword*
          [3].\n### SSL certificates for HTTPS\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-web-sslsecret\ndata:\n  ssl.crt: >-\n   < ssl.crt data>\n
          \ ssl.key: >-\n\t < ssl.key data >\n  dhparam.pem: >-\n   <  dhparam.pem
          data >\n```\nFiles *ssl.crt*, *ssl.key* and *dhparam.perm* are required
          for Zabbix web-interface for SSL support.\n"
        displayName: Zabbix Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - zabbix
        - appliance
        - agent
        - metrics
        links:
        - name: Zabbix
          url: https://www.zabbix.com
        - name: Zabbix Official Documentation
          url: https://www.zabbix.com/documentation/6.0/en/manual/quickstart
        - name: Zabbix Operator Installation
          url: https://www.zabbix.com/documentation/6.0/en/manual/installation/containers/openshift
        - name: Downloads
          url: https://www.zabbix.com/download
        maintainers:
        - email: alexey.pustovalov@zabbix.com
          name: Alexey Pustovalov
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Zabbix LLC
        relatedImages:
        - registry.connect.redhat.com/zabbix/zabbix-agent2-60@sha256:424003d52f3cb4360adf2b3619d8b9f8de9196c630ee7e4ceb5d177c1c5a36e4
        - registry.connect.redhat.com/zabbix/zabbix-snmptraps-60@sha256:e61e6b1c551dec8488cce0998f94d83d6d10b5d89a465d7119b088313032cb81
        - registry.connect.redhat.com/zabbix/zabbix-server-mysql-60@sha256:637484faec6e9a55049c13dbd1f100c03d757c3a21749388c0605040f02d05ce
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
        - quay.io/zabbix/zabbix-operator@sha256:43d26488a19967bdb2a8aafde5f43820be95011811da9cc5ec623c9029801675
        - registry.connect.redhat.com/zabbix/zabbix-web-mysql-60@sha256:611d2670e1bf95b81226976c0a1f64434b2a7766f196dadbd646a7d8319f4cb1
        - registry.connect.redhat.com/zabbix/zabbix-java-gateway-60@sha256:a4feb725c12c4ac90eb9244c6fafdb6708679f2e0e7f6aa683c765127306412e
        - registry.connect.redhat.com/zabbix/zabbix-proxy-mysql-60@sha256:c2ae74d97c083ff348dc9c55932a33da4779cd9efdd9d7c5a79acfffae5b9ed0
        - registry.connect.redhat.com/zabbix/zabbix-proxy-sqlite-60@sha256:62e62ebc8c1bc13fb610f7a83412e65d852f74b3c2c212baa75ea27fc7fa03eb
        - registry.connect.redhat.com/zabbix/zabbix-web-service-60@sha256:4a333094e110713470a897b6406c5e340e962caf32a429dddf19865041fcc448
        - registry.redhat.io/rhel9/mysql-80@sha256:c4b5246d0c0d089a766ea07768c22d9d6be8049ac1379bd345db99390ffe2a36
        - registry.connect.redhat.com/zabbix/zabbix-agent-60@sha256:2895b5e45785aaf528d4a6f78c8af0a0d494dbe9362403fdd470673d911d5004
        version: 6.0.40
      entries:
      - name: zabbix-operator-certified.v6.0.40
        version: 6.0.40
      - name: zabbix-operator-certified.v6.0.38
        version: 6.0.38
      - name: zabbix-operator-certified.v6.0.36
        version: 6.0.36
      - name: zabbix-operator-certified.v6.0.33
        version: 6.0.33
      - name: zabbix-operator-certified.v6.0.29
        version: 6.0.29
      - name: zabbix-operator-certified.v6.0.27
        version: 6.0.27
      - name: zabbix-operator-certified.v6.0.26
        version: 6.0.26
      - name: zabbix-operator-certified.v6.0.17
        version: 6.0.17
      - name: zabbix-operator-certified.v6.0.14
        version: 6.0.14
      - name: zabbix-operator-certified.v6.0.13
        version: 6.0.13
      name: lts
    defaultChannel: lts
    packageName: zabbix-operator-certified
    provider:
      name: Zabbix LLC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/multiarch-tuning-operator
    name: multiarch-tuning-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multiarch-tuning-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multiarch.openshift.io/v1beta1",
                "kind": "ClusterPodPlacementConfig",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logVerbosity": "Normal",
                  "namespaceSelector": {
                    "matchExpressions": [
                      {
                        "key": "multiarch.openshift.io/exclude-pod-placement",
                        "operator": "DoesNotExist"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Other
          console.openshift.io/disable-operand-delete: "false"
          containerImage: registry.redhat.io/multiarch-tuning/multiarch-tuning-rhel9-operator@sha256:1a1d3f5258d4b91f1a0380b0fe2c809823e323d9ff74ae4171468168d255e7e0
          createdAt: "2025-03-05T12:27:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operator.openshift.io/uninstall-message: You must remove all Operands before
            uninstalling the operator.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-multiarch-tuning-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/multiarch-tuning-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPodPlacementConfig defines the configuration for the
              architecture aware pod placement operand. Users can only deploy a single
              object named "cluster". Creating the object enables the operand.
            displayName: Cluster Pod Placement Config
            kind: ClusterPodPlacementConfig
            name: clusterpodplacementconfigs.multiarch.openshift.io
            version: v1alpha1
          - description: ClusterPodPlacementConfig defines the configuration for the
              architecture aware pod placement operand. Users can only deploy a single
              object named "cluster". Creating the object enables the operand.
            displayName: Cluster Pod Placement Config
            kind: ClusterPodPlacementConfig
            name: clusterpodplacementconfigs.multiarch.openshift.io
            version: v1beta1
        description: |
          The Multiarch Tuning Operator optimizes workload management within multi-architecture clusters and in
          single-architecture clusters transitioning to multi-architecture environments.

          By default, the scheduler does not consider the architecture of a pod's container images
          when determining the placement of new pods onto nodes.

          The Multiarch Tuning Operator introduces the ClusterPodPlacementConfig custom resource to
          enable architecture-aware scheduling, ensuring that pods are assigned to nodes with
          a corresponding architecture for their container images.

          To enable this functionality, you must create a ClusterPodPlacementConfig resource.

          When the ClusterPodPlacementConfig object is created, the Multiarch Tuning Operator deploys the necessary operands
          to support architecture-aware workload scheduling.
        displayName: Multiarch Tuning Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multiarch-tuning-operator
        - pod-placement
        links:
        - name: Multiarch Tuning Operator Documentation
          url: https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/postinstallation_configuration/configuring-multi-architecture-compute-machines-on-an-openshift-cluster#multiarch-tuning-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/multiarch-tuning-operator
        relatedImages:
        - registry.redhat.io/multiarch-tuning/multiarch-tuning-rhel9-operator@sha256:1a1d3f5258d4b91f1a0380b0fe2c809823e323d9ff74ae4171468168d255e7e0
        version: 1.1.0
      entries:
      - name: multiarch-tuning-operator.v1.1.0
        version: 1.1.0
      - name: multiarch-tuning-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: multiarch-tuning-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/multiarch-tuning-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: infra-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infra-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instanceha.openstack.org/v1beta1",
                "kind": "InstanceHa",
                "metadata": {
                  "name": "instanceha"
                },
                "spec": {
                  "caBundleSecretName": "combined-ca-bundle"
                }
              },
              {
                "apiVersion": "memcached.openstack.org/v1beta1",
                "kind": "Memcached",
                "metadata": {
                  "name": "memcached"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "network.openstack.org/v1beta1",
                "kind": "DNSData",
                "metadata": {
                  "name": "allhosts"
                },
                "spec": {
                  "hosts": [
                    {
                      "hostnames": [
                        "edpm-compute-0.ctlplane"
                      ],
                      "ip": "192.168.122.100"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-0.internalapi"
                      ],
                      "ip": "172.17.0.100"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-0.storage"
                      ],
                      "ip": "172.18.0.100"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-0.tenant"
                      ],
                      "ip": "172.19.0.100"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-1.ctlplane"
                      ],
                      "ip": "192.168.122.101"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-1.internalapi"
                      ],
                      "ip": "172.17.0.101"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-1.storage"
                      ],
                      "ip": "172.18.0.101"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-1.tenant"
                      ],
                      "ip": "172.19.0.101"
                    },
                    {
                      "hostnames": [
                        "keystone-internal.openstack.svc",
                        "placement-internal.openstack.svc",
                        "cinder-internal.openstack.svc",
                        "glance-internal.openstack.svc",
                        "nova-internal.openstack.svc",
                        "nova-metadata-internal.openstack.svc",
                        "neutron-internal.openstack.svc"
                      ],
                      "ip": "172.17.0.80"
                    },
                    {
                      "hostnames": [
                        "rabbitmq.openstack.svc"
                      ],
                      "ip": "172.17.0.85"
                    },
                    {
                      "hostnames": [
                        "rabbitmq-cell1.openstack.svc"
                      ],
                      "ip": "172.17.0.86"
                    }
                  ]
                }
              },
              {
                "apiVersion": "network.openstack.org/v1beta1",
                "kind": "DNSMasq",
                "metadata": {
                  "name": "dnsmasq"
                },
                "spec": {
                  "debug": {
                    "service": false
                  },
                  "options": [
                    {
                      "key": "server",
                      "values": [
                        "192.168.122.1"
                      ]
                    },
                    {
                      "key": "local",
                      "values": [
                        "/example.com/"
                      ]
                    }
                  ],
                  "override": {
                    "service": {
                      "metadata": {
                        "annotations": {
                          "metallb.universe.tf/address-pool": "ctlplane",
                          "metallb.universe.tf/allow-shared-ip": "ctlplane",
                          "metallb.universe.tf/loadBalancerIPs": "192.168.122.80"
                        }
                      },
                      "spec": {
                        "type": "LoadBalancer"
                      }
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "network.openstack.org/v1beta1",
                "kind": "IPSet",
                "metadata": {
                  "name": "edpm-compute-0"
                },
                "spec": {
                  "networks": [
                    {
                      "defaultRoute": true,
                      "name": "CtlPlane",
                      "subnetName": "subnet1"
                    },
                    {
                      "name": "InternalApi",
                      "subnetName": "subnet1"
                    },
                    {
                      "name": "Storage",
                      "subnetName": "subnet1"
                    },
                    {
                      "name": "Tenant",
                      "subnetName": "subnet1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "network.openstack.org/v1beta1",
                "kind": "NetConfig",
                "metadata": {
                  "name": "netconfig"
                },
                "spec": {
                  "networks": [
                    {
                      "dnsDomain": "ctlplane.example.com",
                      "name": "ctlplane",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "192.168.122.120",
                              "start": "192.168.122.100"
                            },
                            {
                              "end": "192.168.122.200",
                              "start": "192.168.122.150"
                            }
                          ],
                          "cidr": "192.168.122.0/24",
                          "gateway": "192.168.122.1",
                          "name": "subnet1"
                        }
                      ]
                    },
                    {
                      "dnsDomain": "internalapi.example.com",
                      "name": "internalapi",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "172.17.0.250",
                              "start": "172.17.0.100"
                            }
                          ],
                          "cidr": "172.17.0.0/24",
                          "name": "subnet1",
                          "vlan": 20
                        }
                      ]
                    },
                    {
                      "dnsDomain": "external.example.com",
                      "name": "external",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "10.0.0.250",
                              "start": "10.0.0.100"
                            }
                          ],
                          "cidr": "10.0.0.0/24",
                          "gateway": "10.0.0.1",
                          "name": "subnet1"
                        }
                      ]
                    },
                    {
                      "dnsDomain": "storage.example.com",
                      "name": "storage",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "172.18.0.250",
                              "start": "172.18.0.100"
                            }
                          ],
                          "cidr": "172.18.0.0/24",
                          "name": "subnet1",
                          "vlan": 21
                        }
                      ]
                    },
                    {
                      "dnsDomain": "storagemgmt.example.com",
                      "name": "storagemgmt",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "172.20.0.250",
                              "start": "172.20.0.100"
                            }
                          ],
                          "cidr": "172.20.0.0/24",
                          "name": "subnet1",
                          "vlan": 23
                        }
                      ]
                    },
                    {
                      "dnsDomain": "tenant.example.com",
                      "name": "tenant",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "172.19.0.250",
                              "start": "172.19.0.100"
                            }
                          ],
                          "cidr": "172.19.0.0/24",
                          "name": "subnet1",
                          "vlan": 22
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "network.openstack.org/v1beta1",
                "kind": "Reservation",
                "metadata": {
                  "name": "edpm-compute-0"
                },
                "spec": {
                  "ipSetRef": {
                    "name": "edpm-compute-0",
                    "namespace": "openstack",
                    "uid": "56753ffb-e486-43ab-a79a-f85f27b2ac14"
                  },
                  "reservation": {
                    "CtlPlane": {
                      "address": "192.168.122.100",
                      "network": "CtlPlane",
                      "subnet": "subnet1"
                    },
                    "InternalApi": {
                      "address": "172.17.0.100",
                      "network": "InternalApi",
                      "subnet": "subnet1"
                    },
                    "Storage": {
                      "address": "172.18.0.100",
                      "network": "Storage",
                      "subnet": "subnet1"
                    },
                    "StorageMgmt": {
                      "address": "172.20.0.100",
                      "network": "StorageMgmt",
                      "subnet": "subnet1"
                    },
                    "Tenant": {
                      "address": "172.19.0.100",
                      "network": "Tenant",
                      "subnet": "subnet1"
                    }
                  }
                }
              },
              {
                "apiVersion": "rabbitmq.openstack.org/v1beta1",
                "kind": "TransportURL",
                "metadata": {
                  "name": "transporturl"
                },
                "spec": {
                  "rabbitmqClusterName": "rabbitmq"
                }
              },
              {
                "apiVersion": "redis.openstack.org/v1beta1",
                "kind": "Redis",
                "metadata": {
                  "name": "redis"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:45Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSData is the Schema for the dnsdata API
            displayName: DNSData
            kind: DNSData
            name: dnsdata.network.openstack.org
            version: v1beta1
          - description: DNSMasq is the Schema for the dnsmasqs API
            displayName: DNSMasq
            kind: DNSMasq
            name: dnsmasqs.network.openstack.org
            version: v1beta1
          - description: InstanceHa is the Schema for the instancehas API
            displayName: Instance Ha
            kind: InstanceHa
            name: instancehas.instanceha.openstack.org
            version: v1beta1
          - description: IPSet is the Schema for the ipsets API
            displayName: IPSet
            kind: IPSet
            name: ipsets.network.openstack.org
            version: v1beta1
          - description: Memcached is the Schema for the memcacheds API
            displayName: Memcached
            kind: Memcached
            name: memcacheds.memcached.openstack.org
            version: v1beta1
          - description: NetConfig is the Schema for the netconfigs API
            displayName: Net Config
            kind: NetConfig
            name: netconfigs.network.openstack.org
            version: v1beta1
          - description: Redis is the Schema for the redises API
            displayName: Redis
            kind: Redis
            name: redises.redis.openstack.org
            version: v1beta1
          - description: Reservation is the Schema for the reservations API
            displayName: Reservation
            kind: Reservation
            name: reservations.network.openstack.org
            version: v1beta1
          - description: TransportURL is the Schema for the transporturls API
            displayName: Transport URL
            kind: TransportURL
            name: transporturls.rabbitmq.openstack.org
            version: v1beta1
        description: OpenStack Infrastructure
        displayName: OpenStack Infra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Infrastructure
        links:
        - name: Infra Operator
          url: https://github.com/openstack-k8s-operators/infra-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/infra-rhel9-operator@sha256:177077029ad8baa959e41a785cd29ed63745e8f39be618aa97509e6867fbff7c
        - registry.redhat.io/rhoso/openstack-memcached-rhel9@sha256:962a314226009a8ca5c95fb66a75b15176506e02110847d0f42aeed6dd26a751
        - registry.redhat.io/rhoso/openstack-redis-rhel9@sha256:b83d23df9f2b367e604f404d97574fd6db9e0ed530eda23e9e0bbe01292c8db8
        - registry.redhat.io/rhoso/openstack-neutron-server-rhel9@sha256:806e5e009a77a29c82699487e539669e1ec2136514a0332d222b5cb3fed6f452
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: infra-operator.v1.0.6
        version: 1.0.6
      - name: infra-operator.v1.0.5
        version: 1.0.5
      - name: infra-operator.v1.0.4
        version: 1.0.4
      - name: infra-operator.v1.0.3
        version: 1.0.3
      - name: infra-operator.v1.0.2
        version: 1.0.2
      - name: infra-operator.v1.0.1
        version: 1.0.1
      - name: infra-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: infra-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: maximilianopizarro
      provider-url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
    name: jhipster-online-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jhipster-online-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maximilianopizarro.github.io/v1alpha1",
                "kind": "JhipsterOnline",
                "metadata": {
                  "name": "jhipsteronline-sample"
                },
                "spec": {
                  "affinity": {},
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "env": {
                    "APPLICATION_GITHUB_CLIENT-ID": "CLIENT-ID",
                    "APPLICATION_GITHUB_CLIENT-SECRET": "CLIENT-SECRET",
                    "APPLICATION_GITHUB_HOST": "https://github.com",
                    "APPLICATION_JHIPSTER-CMD_CMD": "jhipster-quarkus",
                    "JAVA_APP_JAR": "jhonline-2.33.0.war",
                    "SPRING_DATASOURCE_PASSWORD": "jhipster",
                    "SPRING_DATASOURCE_URL": "jdbc:mariadb://mariadb:3306/jhipsteronline",
                    "SPRING_DATASOURCE_USERNAME": "jhipster"
                  },
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "quay.io/maximilianopizarro/jhipster-online",
                    "tag": "quarkus"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "livenessProbe": {
                    "httpGet": {
                      "path": "/jdl-studio/",
                      "port": 8080
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "readinessProbe": {
                    "httpGet": {
                      "path": "/jdl-studio/",
                      "port": 8080
                    }
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "securityContext": {},
                  "service": {
                    "port": 8080,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automount": false,
                    "create": false,
                    "name": ""
                  },
                  "tolerations": [],
                  "volumeMounts": [],
                  "volumes": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/maximilianopizarro/jhipster-online-operator:v0.1.0
          createdAt: "2025-05-30T22:42:19Z"
          description: JHipster Online provides an intuitive web interface for generating
            JHipster projects, allowing you to define your application architecture
            using a JDL (JHipster Domain Language) model and generate the code directly
            into a GitHub repository.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/maximilianoPizarro/jhipster-online-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: JhipsterOnline
            name: jhipsteronlines.maximilianopizarro.github.io
            version: v1alpha1
        description: |
          JHipster Online Operator: This repository hosts the JHipster Online Operator, designed to seamlessly deploy JHipster Online on Red Hat OpenShift and Kubernetes environments.

          JHipster Online provides an intuitive web interface for generating JHipster projects, allowing you to define your application architecture using a JDL (JHipster Domain Language) model and generate the code directly into a GitHub repository.

          Overview: The JHipster Online Operator manages the deployment and lifecycle of the JHipster Online application. It leverages the power of Kubernetes Operators to provide a native Kubernetes experience for JHipster Online users.

          This artifact includes:
          - JHipster 8.8.0. for generate Spring Boot 3.4.1 projects.
          - generator-jhipster-quarkus 3.4.0 for generate Quarkus 3.11.1 projects.
          - JDL Studio for add JDL models files by PR on your repo.

          Usage: Once the operator is running, you can create instances of JHipster Online by defining `JhipsterOnline` custom resources. An example `JhipsterOnline` custom resource is provided in the `alm-examples`. You would apply this (or a similar) YAML file to your cluster to create an instance of JHipster Online. Feel free to customize the `JhipsterOnline` Custom Resource Definition (CRD) instance to suit your needs, such as configuring GitHub integration, database connections, and resource allocations.

          Configuration: The `JhipsterOnline` custom resource supports various configuration options including `replicaCount`, `image`, `imagePullSecrets`, `serviceAccount`, `podAnnotations`, `podLabels`, `podSecurityContext`, `securityContext`, `service` type and port, `ingress` settings (enabled, `className`, `annotations`, `hosts`, `tls`), OpenShift `Route` (enabled), `autoscaling` (enabled, `minReplicas`, `maxReplicas`, `targetCPUUtilizationPercentage`), `resource` requests and limits, `nodeSelector`, `tolerations`, `affinity`, `liveness` and `readiness` probes, `env` variables, `volumes`, and `volumeMounts`. Refer to the detailed `README.md` for a full table of parameters.
        displayName: JHipster Online Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - java
        - spring-boot
        - jhipster
        - quarkus
        links:
        - name: Jhipster Online Operator
          url: https://maximilianopizarro.github.io/jhipster-online-operator/
        maintainers:
        - email: maximiliano.pizarro.5@gmail.com
          name: Maximiliano Pizarro
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: maximilianopizarro
          url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/maximilianopizarro/jhipster-online-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: jhipster-online-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: jhipster-online-operator
    provider:
      name: maximilianopizarro
      url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Skupper Project
      provider-url: https://skupper.io
    name: skupper-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: alpha
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1
    - currentCSV: skupper-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:e5f8b2ddf378cf3e40f4c7b7ec8482eb3321947cd46481c868d727932c449d78
          createdAt: "2024-02-26T18:22:07Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/config-sync@sha256:659fe28d1df7c99bba3bb854b1e9991c715ae4de99d91c88d5db5e4f9ef4c852
        - quay.io/skupper/flow-collector@sha256:3870c5bc89e623b7cebe21856bfefc22de76313705efa0ec551a5963789fff3b
        - quay.io/skupper/site-controller@sha256:e5f8b2ddf378cf3e40f4c7b7ec8482eb3321947cd46481c868d727932c449d78
        - quay.io/prometheus/prometheus@sha256:d2ab0a27783fd4ad96a8853e2847b99a0be0043687b8a5d1ebfb2dd3fa4fd1b8
        - quay.io/openshift/origin-oauth-proxy@sha256:a7dff785d821acc4b972d94964947713fade45f007dedb5deee17a2869bf5d24
        - quay.io/skupper/skupper-router@sha256:bf8d5955ce4e406275a4934278efe51c1af9a1a269a70e8fe642b7d94d80cb81
        - quay.io/skupper/service-controller@sha256:345edf2d9d72b5a97bcce0e8cea8cda82d120f7bf83756d73f9ac114c2fcced8
        version: 1.6.0
      entries:
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.6
    - currentCSV: skupper-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:4809db7263d77a95797061798e910f1bdbb5d5fe2db2bb3c743b97f5d0c7b3af
          createdAt: "2024-06-12T14:36:59Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/prometheus/prometheus@sha256:d2ab0a27783fd4ad96a8853e2847b99a0be0043687b8a5d1ebfb2dd3fa4fd1b8
        - quay.io/openshift/origin-oauth-proxy@sha256:a7dff785d821acc4b972d94964947713fade45f007dedb5deee17a2869bf5d24
        - quay.io/skupper/skupper-router@sha256:5eb258a0927e7706e7c035f1def2eb5211c8c33dd371fafa7210517fd894f68b
        - quay.io/skupper/service-controller@sha256:624e88e695297df28e485db7e802945ad1978ac6b0c9e9edca8b5e6013af8991
        - quay.io/skupper/site-controller@sha256:4809db7263d77a95797061798e910f1bdbb5d5fe2db2bb3c743b97f5d0c7b3af
        - quay.io/skupper/config-sync@sha256:a884b26a32179af5a2bfa85118bf873852eec6e2930e92da778b57506c0a0880
        - quay.io/skupper/flow-collector@sha256:bf646b0cb75e1c29bbe0261aa9943e06d33118f5e6a5ccb7503c60c974167692
        version: 1.7.1
      entries:
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.7
    - currentCSV: skupper-operator.v1.8.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:ffe3ea5067d350b64b69120bdf6d9e3ee9a30e8a2559b77170a10d66635623b8
          createdAt: "2025-03-27T20:11:53Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/service-controller@sha256:a45de211b8f8a8294a3588932a0a72766cf0d5ce64b01f4f744832c8d09aaa46
        - quay.io/skupper/config-sync@sha256:9e96d6a585f037f60ca3d89b5efee27ec92444d353e8b289ca35b94aeed1e309
        - quay.io/skupper/flow-collector@sha256:252244ab6aaf6b80118bfc06b690d4f5e8d7829b999fca60d194a84452fee7f6
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/openshift/origin-oauth-proxy@sha256:c15baf57dbd17dcc517cd2fa2299462d62124f689d6f5a48a3c9fd154fc77ccf
        - quay.io/skupper/skupper-router@sha256:7633b143f1844ee559322e4642ca80e8248a08711c8e814d489c26c39ecf9bd0
        - quay.io/skupper/site-controller@sha256:ffe3ea5067d350b64b69120bdf6d9e3ee9a30e8a2559b77170a10d66635623b8
        version: 1.8.4
      entries:
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.8
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.9
    defaultChannel: stable
    packageName: skupper-operator
    provider:
      name: Skupper Project
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Harness Inc.
      provider-url: https://harness.io
    name: harness-gitops-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: harness-gitops-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gitops.harness.io/v1alpha1",
                "kind": "GitopsAgent",
                "metadata": {
                  "annotations": {
                    "containerImage": "harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde",
                    "createdAt": "2025-07-31T12:11:18Z"
                  },
                  "name": "gitops-agent-sample"
                },
                "spec": {
                  "identity": {
                    "accountIdentifier": "accountIdentifier",
                    "agentIdentifier": "agentIdentifier",
                    "agentName": "agentName"
                  },
                  "replicas": 1,
                  "secrets": {
                    "agentSecret": "agentSecret"
                  },
                  "target": {
                    "http": "https://app.harness.io/gitops/",
                    "protocol": "HTTP1"
                  },
                  "upgrade": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde
          createdAt: "2025-07-31T06:41:18Z"
          description: Harness operator to manage and provision GitOps Agent
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Harness Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Schema for the Gitops Agents API
            displayName: Gitops Agent
            kind: GitopsAgent
            name: gitopsagents.gitops.harness.io
            version: v1alpha1
        description: |-
          ## Overview
          Harness Gitops Operator manages the full lifecycle for [Gitops Agent](https://developer.harness.io/docs/continuous-delivery/gitops/connect-and-manage/install-a-harness-git-ops-agent/)

          The operator aims to provide the following:

            * Easy configuration and installation of Harness GitOps Agent.
            * Provide seamless upgrades for Harness GitOps Agent.
            * Monitoring and autorecovery for GitOps Agent.

          ## Usage
          Deploy gitops-agent in namespace where operator is installed with basic setup.

          ### Prerequisites:
            * Go to your Harness account and create a new gitops agent
            * Copy account, org and project identifiers
            * Retrieve the agent secret key

          ```
          apiVersion: gitops.harness.io/v1alpha1
          kind: GitopsAgent
          metadata:
            name: gitops-agent-sample
          spec:
            identity:
              accountIdentifier: accountIdentifier
              agentIdentifier: agentIdentifier
              agentName: agentName
            secrets:
              agentSecret: agentSecret
            target:
              http: https://app.harness.io/gitops/
              protocol: HTTP1
            replicas: 1
            upgrade:
              enabled: true
          ```
          **Note:** Gitops Agent require an active ArgoCD instance running in the same namespace where agent is installed.

          See [Harness Documentation](https://developer.harness.io/) for more information.
        displayName: Harness GitOps Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - harness
        - agent
        - cd
        - argocd
        - kubernetes
        links:
        - name: Harness Gitops Operator
          url: https://harness-gitops-operator.domain
        maintainers:
        - email: alexandru.bosii@harness.io
          name: Alexandru Bosii
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Harness Inc.
          url: https://harness.io
        relatedImages:
        - harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde
        - docker.io/harness/gitops-agent@sha256:c9f57338be294d32a73a4cbecf90dfff50f5d36a699f3ec32d962dcb64d8d085
        version: 0.3.0
      entries:
      - name: harness-gitops-operator.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: harness-gitops-operator
    provider:
      name: Harness Inc.
      url: https://harness.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataTrucker.IO
      provider-url: https://www.datatrucker.io/
    name: datatrucker-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datatrucker-operator.v2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datatrucker.datatrucker.io/v1",
                "kind": "DatatruckerConfig",
                "metadata": {
                  "name": "datatruckerconfig-sample"
                },
                "spec": {
                  "API": {
                    "crypto": {
                      "CredentailEncryption": {
                        "Algorithm": "aes-256-cbc",
                        "IV": "eThWmZq4t7w!z$C\u0026",
                        "Key": "JaNdRgUkXp2s5u8x/A?D(G+KbPeShVmY"
                      },
                      "JWT": {
                        "authorization_keys": {
                          "private_key": "privateJWT.key",
                          "public_key": "publicJWT.key"
                        },
                        "signOptions": {
                          "algorithm": "RS256",
                          "expiresIn": "60m",
                          "issuer": "datatrucker"
                        },
                        "verify": {
                          "issuer": "datatrucker"
                        }
                      },
                      "keycloak": {
                        "client_id": "nodejs",
                        "client_secret": "7c5542ae-7264-420d-bf5f-766f4ab39ad3",
                        "jwt_publickey_verifier": "keycloak.key",
                        "realm": "TestRealm",
                        "url": "http://localhost:30015"
                      },
                      "passwordHash": {
                        "Algorithm": "sha256",
                        "SaltLength": 16
                      }
                    },
                    "cryptokeys": {
                      "keycloakkey": "LS0tLS1CRUdJTiBQVUJMSUMgS0VZLS0tLS0KTUlJQklqQU5CZ2txaGtpRzl3MEJBUUVGQUFPQ0FROEFNSUlCQ2dLQ0FRRUF4ZnhGY1gxb3RmTUN6SWlkZ09tMndZakZWZXZHN1BQT1h5dTNrTEFuTWtTVkUxUXl4TzJ4SFFvSGcydDFVcUU4SVNXeTN3ZkZTbERzdU1VK0J1VVkvaWI5L04vSlpBV3k5MGRET0YxREdIeGVwM21HMVBTUllMQWFGQ0I3NTVsbWZjL1dna3VzZFdLdmxOTmZuNkd6S1J6OXBmV3RNUnhWWkFrejhmNlVUeG0zWE02Q0tOMUludThubEcxU3ZzaFc5cmt3Lyt6UmVvOFVCV2pPcTh5RjdPV0tnS0ZZZlhENFREa090OWFxN3RaaSsxaExYOE5iV2xxRnJZYmhJcFZINXlLSko3ZTA5cUVIRkVEelg5QXJzVVNHamNjbFFFYTl1dmJxU2d4Rzc3RUZVbk82aDZxUkNnU3lNNnVYdk9qcWlIMHp4RTNmRHB4R2NhY2pvTW1EMlFJREFRQUIKLS0tLS1FTkQgUFVCTElDIEtFWS0tLS0t",
                      "privateJWTkey": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQ0KTUlJRW9RSUJBQUtDQVFCVWd1ZDN1OFlXcUJCWWtYZ3M2K25pMS84bnM5YjRGNkZudzA4RDRKdGRlalg5L1ZaMg0KWFdOamJ1ekMxUkRHMnVseFM4QjZIdTZYeEhYWU9XU2ZzOEY0UHZ2eC9TUWdmRWtWUlNDL1pQN3pzM1ZETllaMA0Kc1JScDNSZzUza29qUG80T3o2azFaeGprbFRFZEpMWUtqYlZud3U1M3A3MTFsVTJkV3B4R3VEeGYxVTJuUWgrWA0KaUhKZkhyUmxTZVA1aVhYQkpxd21VTDFkK25lTU5BOXlGakR0MnNZMG5iSDRySDVxbk1VRk4xMEpsZ3V4NFdiUw0KaGFUNklNN3h1dUVHbDRDckVENmdTejBNNGF1UVk4ekZxQWxJNlFWUTlML28zQnBCRmxYUUMwWmowb3lwQVFKLw0KQ1VLaU9FcEhLRHgram02RU9KOE1aejlMQ01yTmdLME5KSWlaQWdNQkFBRUNnZ0VBRTBMMnNLSjZob1Z0bkZJag0KZXRHV01CNmN5QmJFSXkxM215ZTNwYUdXcDVxSUhjaVU0VXR0c0tjQ2pYRjBibTVNa0liOGFwa01Oa2tQdHVWMQ0KMnJSa254eHozZmFkT2s0NWRnQ1ljeWI0dU9DcEZGVnB2TVR6YXlGNk9RNUNCWUJuYmxiRU1TaTdVclh5U0xzcQ0KUm9OLzJoTml1S0RobktBWDNZcGVYZnYxdnIyaWIrMGtiRDVvMVZLcnB6SlJ5emlsWHFwQWJwNDQ4RExFZHVGeA0KTnJKc0tWb1dXMVhmS1FvR2VDMXZqNFJudzdhWkNyU1BjMEhvMUdKb29xSXlrNjRXZnhGWjhpVElsWDJMU1gzbQ0KQm1pN1ZXZnVJK0QvUXRtWEVpN05oTE1jUmExVXB4QkdiK1ZPZXlucEVyMEpxM3VZSmlva1VIZzloOGIzSDM1Yg0KK0lJaEFRS0JnUUNtbzUrRUNpOEJUcFRHdlpqQXBtdmxGZ055NkpWd0RvVHJLSGdCcHJHVHpsVGt6Q0tLYzJhTw0KTTMvQmVIVUQ1NVFqcTdUM094eGgzQUk3cG9kT0tCWS84eXgzenlYUExQUmVCMlR0eTNVQjE5YU5CclhNUGQ5RQ0KTG5hbUQ0SmZQeTV0WHpWWEloWVB0UzBiNjdGbjZnZFl4MlNBbGFYYWFjejU4eURLZnlZL3VRS0JnUUNCMUxNYg0KTFZNTU1Pd1U4R2xsVVB1V0Q5VnBsMnlURytCN2Nnd0ZiYTlrWVlVSFQ4TmlvaUdxQzlLZWxqeklKNnlUdyt3aA0KQisvNGNJU1o5Y0JDNHB0VnRlTWJwUkg3UTQwK20xU3M1SmQzU1VFcDBMOW5VbmdVNDhrMzlpSTV5aEgxU3JTRw0KREtBamVka2Rjb3hNVWFwUHB6cVprOVNKdGVHckVsTHY5WmcvNFFLQmdHcExQTmt3MWNMenFSc01uWDlscUxVOA0KVVg3ZXRWUUJMdkFCcEdzSk5ZU2V2TEFsTm5NeUlWemJwTVo0Vzh2ZDBEVGd0bEp0dGpFRjhiYXpCdVBuMWlsOA0KcUFUVWNiMnhZejhIWUczK3RlZ3pFWHhycDFOY3RDbjlxVmxMMDFHR0w4czhTLzRFbENYNDM0bEwwcUlDT2lWNw0KS3JTS3dIeU51RjVIbzhjSFRCS2hBb0dBVjhVRVpENEs0N1g3RTlqSFY3QWN0UUQxUkpJUzI0ekFBSlJFSGpOZA0KbkIzMVFkRWhoSCtaQkkxRGtPUitLTTN4ZEYvQ05UNVpjM29WbXNPeGUwWVl2eVBjYWViNU5nVWY2cEExK1QvOA0KU2E4ak4zS2d3dWZVdUVLWjJKUXo1UjkvYm5WWGZ6dnYwcXZnSkhFNCthdEgyTnR6MlRsQU9PelZwOElXZ1pSbA0KbmVFQ2dZQXpVS0RLS1ZKaXBYdC9TRC8xd2ZzWElCazFxYWpmdENHOFVTcFlGWGp5VzVYMlZ5bTlVTldmVEdJcA0KQlJkVldXamdFbWhyOU5Da1UrVEU4WEVjTXJNRnhWbXBEeTQ4YmxsQWltWFJtOEhNMlhKODQ4TlArTG5GdEtHdA0KTWswT3JuUHdPYitualZGcXdWUllNQ0QzeFN0Z2RnZUppUHFYcWd3SDhjR0M0OGVmZ0E9PQ0KLS0tLS1FTkQgUlNBIFBSSVZBVEUgS0VZLS0tLS0=",
                      "publicJWTkey": "LS0tLS1CRUdJTiBQVUJMSUMgS0VZLS0tLS0NCk1JSUJJVEFOQmdrcWhraUc5dzBCQVFFRkFBT0NBUTRBTUlJQkNRS0NBUUJVZ3VkM3U4WVdxQkJZa1hnczYrbmkNCjEvOG5zOWI0RjZGbncwOEQ0SnRkZWpYOS9WWjJYV05qYnV6QzFSREcydWx4UzhCNkh1Nlh4SFhZT1dTZnM4RjQNClB2dngvU1FnZkVrVlJTQy9aUDd6czNWRE5ZWjBzUlJwM1JnNTNrb2pQbzRPejZrMVp4amtsVEVkSkxZS2piVm4NCnd1NTNwNzExbFUyZFdweEd1RHhmMVUyblFoK1hpSEpmSHJSbFNlUDVpWFhCSnF3bVVMMWQrbmVNTkE5eUZqRHQNCjJzWTBuYkg0ckg1cW5NVUZOMTBKbGd1eDRXYlNoYVQ2SU03eHV1RUdsNENyRUQ2Z1N6ME00YXVRWTh6RnFBbEkNCjZRVlE5TC9vM0JwQkZsWFFDMFpqMG95cEFRSi9DVUtpT0VwSEtEeCtqbTZFT0o4TVp6OUxDTXJOZ0swTkpJaVoNCkFnTUJBQUU9DQotLS0tLUVORCBQVUJMSUMgS0VZLS0tLS0=",
                      "truckerapicrt": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSUQxVENDQXIyZ0F3SUJBZ0lVUFVzZ2NYVXZRYWtuQldjeUFDeEw2dmt3b204d0RRWUpLb1pJaHZjTkFRRUwKQlFBd2RqRUxNQWtHQTFVRUJoTUNRMEV4Q3pBSkJnTlZCQWdNQWs5T01ROHdEUVlEVlFRSERBWk5hV3gwYjI0eApGekFWQmdOVkJBb01Ea1JoZEdGVWNuVmphMlZ5TGtsUE1SUXdFZ1lEVlFRTERBdEZibWRwYm1WbGNtbHVaekVhCk1CZ0dBMVVFQXd3UlJHRjBZVlJ5ZFdOclpYSlNiMjkwUTBFd0hoY05NakF4TURFeU1UWXlOalV5V2hjTk1qTXcKTVRFMU1UWXlOalV5V2pCM01Rc3dDUVlEVlFRR0V3SkRRVEVMTUFrR0ExVUVDQXdDVDA0eER6QU5CZ05WQkFjTQpCazFwYkhSdmJqRVhNQlVHQTFVRUNnd09SR0YwWVZSeWRXTnJaWEl1U1U4eEZEQVNCZ05WQkFzTUMwVnVaMmx1ClpXVnlhVzVuTVJzd0dRWURWUVFEREJKa1lYUmhkSEoxWTJ0bGNpMWhjR2t1YVc4d2dnRWlNQTBHQ1NxR1NJYjMKRFFFQkFRVUFBNElCRHdBd2dnRUtBb0lCQVFEakM4ZmVCVEg4L3NnejFQcTJQYU9QQjc5YWtyeUZvWVpGV292bQpWOXNvMmJNRHp4NVBZSFJydVFGYmJQRzlVVlc2QWZpSHdEb1p2Z3NTVXpyS3ZyR1RlYjluNUhSVlVOMTl0NExTCjdtUFppNzFRbGE4bVR3QmdvTnE4TFppcnZUdzV0S2I4K3NmcFJxQUlUQkVsRmxVUEZISzllcDl2eFdvaHhwdGgKS1dPVkwvV2ZtWldhWXlRdjlLNDBkWUJpRWpGcTFKZVB4MDlnUFdHM2pOT1VYTTZQUVlMaHJDUXFFcy9VME9GWQpwVFRHUmNXZEx4TFVFZVNkcHJKK1ZEc3lEK1lVeUFRR2phYk14VmVTVlRUN1FGbFBjLzc3OVZsRkdXMC8xWVVaCmpqakNsSkFLUWZOTHg5bG85cWJERVBZL3grZmhnRnYyWUVaSTNTdGVyZWUvVDJoWEFnTUJBQUdqV2pCWU1COEcKQTFVZEl3UVlNQmFBRkNJVy9JYkZIUUNBeEdxZC9FVFpXbVhGeGMvQ01Ba0dBMVVkRXdRQ01BQXdDd1lEVlIwUApCQVFEQWdUd01CMEdBMVVkRVFRV01CU0NFbVJoZEdGMGNuVmphMlZ5TFdGd2FTNXBiekFOQmdrcWhraUc5dzBCCkFRc0ZBQU9DQVFFQUxrMlJDbnVweDA5Wmx0SmN3M1pUaFIveUZGNS9KVlVsWkdHTnc3MlAwQmR6SEloQncrVk4KRkNucGFPMjJ0WGF1NlJVeTVNUWhwc3JrbFNKYVdJT25jOUg0Uk0vbGh3YWsyRXN0MkFhdUFwNUdYSVFJMDVBeAo3NmtDZXEyMnUyalNXRWpFS1hGRnRiZXVDZnFpWlpqNWFLcGF1YlZWRG4ydEZTWjJRMFRib2RRN2dpTFdNUTBLCk5ycVgwQWNSU2J0ZnRSNjBhdGQ1YW41aFNXY0dyZUg5ZHdFUVRvVVNFblg2RFRZSURoeElsZStjQ2V5VjVSWk8KVS9hck5RZ0lFY3ZrZUM1d1lZaFFFL1FUN3lNb1lNajFnckxmOTZSZkRhbEVFNE1seUtDWm9wYWNJWUdzSmhuKwpLMWFISEQvbGZxTC9zbnNtY3pXdHlLUncxTDF1WkxPcjJnPT0KLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=",
                      "truckerapikey": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQpNSUlFb3dJQkFBS0NBUUVBNHd2SDNnVXgvUDdJTTlUNnRqMmpqd2UvV3BLOGhhR0dSVnFMNWxmYktObXpBODhlClQyQjBhN2tCVzJ6eHZWRlZ1Z0g0aDhBNkdiNExFbE02eXI2eGszbS9aK1IwVlZEZGZiZUMwdTVqMll1OVVKV3YKSms4QVlLRGF2QzJZcTcwOE9iU20vUHJINlVhZ0NFd1JKUlpWRHhSeXZYcWZiOFZxSWNhYllTbGpsUy8xbjVtVgptbU1rTC9TdU5IV0FZaEl4YXRTWGo4ZFBZRDFodDR6VGxGek9qMEdDNGF3a0toTFAxTkRoV0tVMHhrWEZuUzhTCjFCSGtuYWF5ZmxRN01nL21GTWdFQm8ybXpNVlhrbFUwKzBCWlQzUCsrL1ZaUlJsdFA5V0ZHWTQ0d3BTUUNrSHoKUzhmWmFQYW13eEQyUDhmbjRZQmI5bUJHU04wclhxM252MDlvVndJREFRQUJBb0lCQUZSbFo2aiswMk1RcmdpZQpUWWQ2aUJhOXNZS1FjdVZWY2x6TFdKK0ozdHhqZCtpZGo4Mi9zK2JyQVRjajJ3UWRmTGxFekVaaHpQY09ybnprClZVRlIreDR2cmxlSEV6czNjTzJ2VGtEci9rKzZZdWZXakNHNFVoS1pHRlhFdVRCMWlqS0ZGV0dLTXM5aXdrL3oKUHBoRG9lQ1dJRVNRbFQxdDY1S2NpZHRuRUM0K3oxaURxR21ZZEVEa2xEZFFIVHorRGpseS80VzV2MFpVQWN2Vwo1Unkvd0JldmNURlBXQy9uZGFrN3hGUnVrU2xNZUlhTWhvTlllOU5uVCt2Nmh5VVh1MXp6cG45UEFJRTVpQWpOCnBsZkFEZitwakwrZFJDbW9sWW1LR3NQVjEzSmtCY1p2cEVoUWgrUmJnaWFwczJLc2plQ0NmTmtST0hndHc0MU0KeGxOTU1Ha0NnWUVBOHM1akpuZ2FMblIvTzZtakZEb0wvd1p2UWdnT1h3Z1RBN2tsQ0trbjBYM0xJS0hhaTZESgpyOVRGQktMdkhUazVFK1Y4NkhIZ29VMUFQQnYyMWJpaHhuQXhUYm8wSHZMdG5QOXU3YnZJejFNOWMyd2ZlWUgzCmliMG1FNkV0Y2wvNnVIUHdMVlZtRGh5bllneEZpeTRRRXp1ZHZwaFIvSnlPazQvUytQUzh2ZXNDZ1lFQTcySW8KWDlkY0xvUmYzSk0vRTF1SXhrRGRLZm1VakN6Mk1NN3pyRkpNUERobTFSUlJhQjdidHNxUzl4Mm5KeTI5amtoSQpndTR0Y2s5STdncm9WRXY0cEtkU0lOVzVNVUg0MXJPckxTbFBobCtMaTM5QXNrTERFVXVQbkFmYko3YWZibFFHCjNFWE1Ka1d2WFQxRG90aE11MUpUMTFPMjhmYlBmb3R3YVRucXFFVUNnWUE1RnArblh0dkJDbkVUNVJuTDJxNFIKWTBEcmhKdGthckNWMG9WTG1RSWN0bmY1RHpVa1dVU3lpOU5hK0xoQjViOXUxeWxjT2RJMytzTGNuektWSTVrZAo0Q2tuZmwxYzNYOVN4NWdjVzJlbW40cTJSYmpNeCtzY3ZwVmlUaVl5V3pESEI3Tno1VDFlbWd3YytJZXRac0l1CjAzaU9aSi82UmhFYnp0Z3E0dGd4eXdLQmdCanhwSDFnSm9YZVcxTFRzUlhtU3RsRnR3WUJzTDN1alhUR2k5T2YKKzd1bXNPaHhKbzV2OExzYVhyZ0xTZmltSm9ObCt0QlIyMEIzRElWcGM0NGRJSHFmeU1vTkd6QUx6TThxaDZLOApGRUpQUUc0M3cxa1A1ZDJ1ZnZkbVJpcmNONzY2V01DcE1UMm5zejlHKy81REJiUGJQMHRIdU45N3RXRmRURWhhCnRaMzFBb0dCQU5xZnVycEV4ZHozdXRNUm91RnJjTmVkV2s4VlVpVVZIV1hZMjhNNWQvRjcxMExIOTZuOW12V08Kc1cwcDFjcXBJMEsyQko2WlYwNVpybnlyRHBsVVpVdVpSQWxGR3hYR1dLU2trQzlmWm90SkFOWDBkR29KSitqTwpEYUNLNGxlcXdOMHNPRE1YZFErMlFPaUtlSUxkNU5KcDNUSW1YV1Z0RFlXUkRwOW9iSUdlCi0tLS0tRU5EIFJTQSBQUklWQVRFIEtFWS0tLS0tCg=="
                    },
                    "jobsServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "debug"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": true,
                        "FileSFTP": true,
                        "IOTKafka": true,
                        "IOTProxy": false,
                        "IOTRedis": true,
                        "Login": false,
                        "ScriptJS": true,
                        "ScriptSSH": true,
                        "ScriptShell": true,
                        "UtilEcho": true,
                        "UtilFuzzy": true,
                        "UtilSentiment": true,
                        "credentials": false,
                        "jobs": true,
                        "resources": false,
                        "statuschecks": true,
                        "ui": false,
                        "usermanagement": false
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "loginServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "info"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": false,
                        "FileSFTP": false,
                        "IOTKafka": false,
                        "IOTProxy": false,
                        "IOTRedis": false,
                        "Login": true,
                        "ScriptJS": false,
                        "ScriptSSH": false,
                        "ScriptShell": false,
                        "UtilEcho": false,
                        "UtilFuzzy": false,
                        "UtilSentiment": false,
                        "credentials": false,
                        "jobs": false,
                        "resources": false,
                        "statuschecks": true,
                        "ui": false,
                        "usermanagement": false
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "managementServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "debug"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": false,
                        "FileSFTP": false,
                        "IOTKafka": false,
                        "IOTProxy": false,
                        "IOTRedis": false,
                        "Login": true,
                        "ScriptJS": false,
                        "ScriptSSH": false,
                        "ScriptShell": false,
                        "UtilEcho": false,
                        "UtilFuzzy": false,
                        "UtilSentiment": false,
                        "credentials": true,
                        "jobs": false,
                        "resources": false,
                        "statuschecks": true,
                        "ui": true,
                        "usermanagement": true
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "name": "API",
                    "resource": {
                      "Cache": {
                        "Creds": false,
                        "CredsFlush": 10,
                        "Resource": false,
                        "ResourceFlush": 10
                      },
                      "Templates": {
                        "Scripts": "scripts",
                        "Shell": "shell",
                        "keys": "keys",
                        "resourcedefinitions": "resourcedefinitions"
                      }
                    }
                  },
                  "DB": {
                    "databasename": "userdb",
                    "hostname": "db",
                    "password": "password",
                    "port": 5432,
                    "type": "pg",
                    "user": "testuser"
                  },
                  "TempDB": {
                    "enabled": false,
                    "pvc": "samplepvc"
                  }
                }
              },
              {
                "apiVersion": "datatrucker.datatrucker.io/v1",
                "kind": "DatatruckerFlow",
                "metadata": {
                  "name": "datatruckerflow-sample"
                },
                "spec": {
                  "API": {
                    "Image": {
                      "imageName": "datatrucker-server/datatrucker-api",
                      "imageNameUI": "datatruckerio/datatrucker-ui",
                      "repository": "registry.gitlab.com/datatruckerio",
                      "tagName": "latest"
                    },
                    "name": "api"
                  },
                  "DatatruckerConfig": "datatruckerconfig-sample",
                  "JobDefinitions": [
                    {
                      "name": "echoapi",
                      "restmethod": "GET",
                      "tenant": "Admin",
                      "type": "Util-Echo",
                      "validations": {
                        "properties": {
                          "voila": {
                            "enum": [
                              "true",
                              "false"
                            ],
                            "type": "string"
                          }
                        },
                        "type": "object"
                      }
                    }
                  ],
                  "Keys": {
                    "configmap": "placeholder"
                  },
                  "Replicas": 1,
                  "Resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "256Mi"
                    },
                    "requests": {
                      "cpu": "250m",
                      "memory": "256Mi"
                    }
                  },
                  "Scripts": {
                    "configmap": "placeholder"
                  },
                  "Type": "Job"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/datatruckerio/datatrucker-operator:2.0.0
          createdAt: "2021-09-08 01:01:01"
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: hhttps://gitlab.com/datatruckerio/datatrucker-server.git
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              - creates secrets to hold server configs
              - create tempDB for protoypting
              - create configs for seperating Login, management and Jobs endpoints
              - Create crypto configs
            displayName: Datatrucker Config
            kind: DatatruckerConfig
            name: datatruckerconfigs.datatrucker.datatrucker.io
            version: v1
          - description: '- creates an API as per Job Definitions'
            displayName: Datatrucker Flow
            kind: DatatruckerFlow
            name: datatruckerflows.datatrucker.datatrucker.io
            version: v1
        description: "## DataTrucker.IO\n\nDataTrucker.IO is a simple no-code / less-code
          API Backend\n\n### What does it do ?\n\nDatatrucker.IO is a product capable
          of reading simple json/yaml configs and building th code necessary for converting
          it into an API.\nAlong with building the code, it will also host the code
          base on a nodejs server , i.e. it will immediately make it available for
          consumption.\n\n### What does it offload from a developer ?\n\nDataTrucker
          is capable of removing the most common activities which a developer needs
          to do on every new project A few of such common activities are\n\n- Creating
          an API endpoint with a specified business logic (using simple plugins)\n-
          Applying standard RBAC\n- Applying Authorization logic\n- Applying hardening
          on endpoints\n- Log management\n- Connecting to variety of systems\n- Modularizing
          business logic\n- The best of doing it with little to no code\n\n### watch
          the quick elevator pitch to understand datatrucker.io\n\n https://youtu.be/DKLFDqhjs5M\n "
        displayName: DataTrucker.IO
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - datatrucker
        - operator
        - nocode
        - apibuilder
        links:
        - name: Website
          url: https://www.datatrucker.io/
        - name: Git Repo
          url: https://gitlab.com/datatruckerio/datatrucker-server.git
        maintainers:
        - email: gauravshankar.can@gmail.com
          name: Gaurav Shankar
        maturity: alpha
        provider:
          name: DataTrucker.IO
          url: https://www.datatrucker.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - docker.io/datatruckerio/datatrucker-operator:2.0.0
        version: 2.0.0
      entries:
      - name: datatrucker-operator.v2.0.0
        version: 2.0.0
      - name: datatrucker-operator.v1.3.2
        version: 1.3.2
      - name: datatrucker-operator.v1.3.1
        version: 1.3.1
      - name: datatrucker-operator.v1.3.0
        version: 1.3.0
      name: alpha
    defaultChannel: alpha
    packageName: datatrucker-operator
    provider:
      name: DataTrucker.IO
      url: https://www.datatrucker.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: postgresql
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: registry.developers.crunchydata.com/crunchydata/postgres-operator:ubi9-5.8.3-0
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.developers.crunchydata.com/crunchydata/postgres-operator:ubi9-5.8.3-0
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: postgresql
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Vertica
      provider-url: http://www.vertica.com
    name: verticadb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: verticadb-operator.v1.11.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "EventTrigger",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "verticadb-operator",
                    "app.kubernetes.io/instance": "eventtrigger-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eventtrigger",
                    "app.kubernetes.io/part-of": "verticadb-operator"
                  },
                  "name": "eventtrigger-sample"
                },
                "spec": {
                  "matches": [
                    {
                      "condition": {
                        "status": "True",
                        "type": "DBInitialized"
                      }
                    }
                  ],
                  "references": [
                    {
                      "object": {
                        "apiVersion": "vertica.com/v1beta1",
                        "kind": "VerticaDB",
                        "name": "verticadb-sample"
                      }
                    }
                  ],
                  "template": {
                    "metadata": {
                      "generateName": "create-user-table-"
                    },
                    "spec": {
                      "template": {
                        "spec": {
                          "containers": [
                            {
                              "command": [
                                "/opt/vertica/bin/vsql",
                                "-h",
                                "verticadb-sample-defaultsubcluster",
                                "-c",
                                "CREATE TABLE T1 (C1 INT);"
                              ],
                              "image": "vertica/vertica-k8s:latest",
                              "name": "main"
                            }
                          ],
                          "restartPolicy": "OnFailure"
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "serviceName": "defaultsubcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaDB",
                "metadata": {
                  "name": "verticadb-sample"
                },
                "spec": {
                  "communal": {
                    "endpoint": "https://s3.amazonaws.com"
                  },
                  "image": "vertica/vertica-k8s:23.3.0-0-minimal",
                  "subclusters": [
                    {
                      "name": "default_subcluster",
                      "size": 3
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: docker.io/vertica/verticadb-operator:1.11.2
          createdAt: "2023-08-02T11:29:39Z"
          description: Operator that manages a Vertica Eon Mode database.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/vertica/vertica-kubernetes
          support: Vertica
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventTrigger is the Schema for the eventtriggers API
            displayName: Event Trigger
            kind: EventTrigger
            name: eventtriggers.vertica.com
            version: v1beta1
          - description: VerticaAutoscaler is a CR that allows you to autoscale one
              or more subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1beta1
          - description: VerticaDB is the CR that defines a Vertica Eon mode cluster
              that is managed by the verticadb-operator.
            displayName: Vertica DB
            kind: VerticaDB
            name: verticadbs.vertica.com
            version: v1beta1
        description: |-
          ### What is Vertica?
          Vertica is a unified analytics platform, based on a massively scalable architecture with the broadest set of analytical functions spanning event and time series, pattern matching, geospatial and end-to-end in-database machine learning. Vertica enables you to easily apply these powerful functions to the largest and most demanding analytical workloads, arming you and your customers with predictive business insights faster than any analytics data warehouse in the market. Vertica provides a unified analytics platform across major public clouds and on-premises data centers and integrates data in cloud object storage and HDFS without forcing you to move any of your data.

          ### What does this operator do?
          This operator manages a Vertica Eon Mode database, and automates the following administrative tasks:

          * Installing Vertica
          * Creating and reviving a Vertica database
          * Restarting and rescheduling DOWN pods to maintain quorum
          * Upgrading Vertica to a new version while keeping the database online
          * Subcluster scaling, both manually and automatically with the Horizontal Pod Autoscaler
          * Service management and health monitoring for pods
          * Load balancing for internal and external traffic
          * Expose a metrics endpoint to integrate with Prometheus to monitor the operator's health

          For a brief overview on how to create a cluster, see the [Vertica GitHub repository](https://github.com/vertica/vertica-kubernetes). For an in-depth look at Vertica on Kubernetes, see the [Vertica documentation](https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm).
        displayName: VerticaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - database
        - vertica
        links:
        - name: Documentation
          url: https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm
        - name: Vertica Container Images
          url: https://hub.docker.com/u/vertica
        maintainers:
        - email: matt.spilchen@vertica.com
          name: Matt Spilchen
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Vertica
          url: http://www.vertica.com
        relatedImages:
        - docker.io/vertica/verticadb-operator:1.11.2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 1.11.2
      entries:
      - name: verticadb-operator.v1.11.2
        version: 1.11.2
      - name: verticadb-operator.v1.11.1
        version: 1.11.1
      - name: verticadb-operator.v1.11.0
        version: 1.11.0
      - name: verticadb-operator.v1.10.2
        version: 1.10.2
      - name: verticadb-operator.v1.10.1
        version: 1.10.1
      - name: verticadb-operator.v1.10.0
        version: 1.10.0
      - name: verticadb-operator.v1.9.0
        version: 1.9.0
      - name: verticadb-operator.v1.8.0
        version: 1.8.0
      - name: verticadb-operator.v1.7.0
        version: 1.7.0
      - name: verticadb-operator.v1.6.0
        version: 1.6.0
      - name: verticadb-operator.v1.5.0
        version: 1.5.0
      - name: verticadb-operator.v1.4.0
        version: 1.4.0
      - name: verticadb-operator.v1.3.1
        version: 1.3.1
      - name: verticadb-operator.v1.3.0
        version: 1.3.0
      - name: verticadb-operator.v1.2.0
        version: 1.2.0
      name: stable
    - currentCSV: verticadb-operator.v25.3.1-0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vertica.com/v1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "customAutoscaler": {},
                  "scalingGranularity": "Subcluster",
                  "serviceName": "default-subcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1",
                "kind": "VerticaDB",
                "metadata": {
                  "name": "verticadb-sample"
                },
                "spec": {
                  "communal": {
                    "endpoint": "https://s3.amazonaws.com"
                  },
                  "image": "opentext/vertica-k8s:24.1.0-0-minimal",
                  "subclusters": [
                    {
                      "name": "default_subcluster",
                      "size": 3
                    }
                  ]
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "EventTrigger",
                "metadata": {
                  "name": "eventtrigger-sample"
                },
                "spec": {
                  "matches": [
                    {
                      "condition": {
                        "status": "True",
                        "type": "DBInitialized"
                      }
                    }
                  ],
                  "references": [
                    {
                      "object": {
                        "apiVersion": "vertica.com/v1beta1",
                        "kind": "VerticaDB",
                        "name": "verticadb-sample"
                      }
                    }
                  ],
                  "template": {
                    "metadata": {
                      "generateName": "create-user-table-"
                    },
                    "spec": {
                      "template": {
                        "spec": {
                          "containers": [
                            {
                              "command": [
                                "/opt/vertica/bin/vsql",
                                "-h",
                                "verticadb-sample-defaultsubcluster",
                                "-c",
                                "CREATE TABLE T1 (C1 INT);"
                              ],
                              "image": "opentext/vertica-k8s:23.4.0-0-minimal",
                              "name": "main"
                            }
                          ],
                          "restartPolicy": "OnFailure"
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "customAutoscaler": {},
                  "scalingGranularity": "Subcluster",
                  "serviceName": "default-subcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaReplicator",
                "metadata": {
                  "name": "verticareplicator-sample"
                },
                "spec": {
                  "source": {
                    "verticaDB": "sourceDBName"
                  },
                  "target": {
                    "verticaDB": "targetDBName"
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaRestorePointsQuery",
                "metadata": {
                  "name": "verticarestorepointsquery-sample"
                },
                "spec": {
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaScrutinize",
                "metadata": {
                  "name": "verticascrutinize-sample"
                },
                "spec": {
                  "verticaDBName": "verticadb-sample"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: docker.io/opentext/verticadb-operator:25.3.1-0
          createdAt: "2025-09-24T00:58:06Z"
          description: Operator that manages a Vertica Eon Mode database.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/vertica/vertica-kubernetes
          support: Vertica
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventTrigger is the Schema for the eventtriggers API
            displayName: Event Trigger
            kind: EventTrigger
            name: eventtriggers.vertica.com
            version: v1beta1
          - description: |-
              VerticaAutoscaler is a CR that allows you to autoscale one or more
              subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1
          - description: |-
              VerticaAutoscaler is a CR that allows you to autoscale one or more
              subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1beta1
          - description: VerticaDB is the CR that defines a Vertica Eon mode cluster
              that is managed by the verticadb-operator.
            displayName: Vertica DB
            kind: VerticaDB
            name: verticadbs.vertica.com
            version: v1
          - description: VerticaReplicator is the Schema for the verticareplicators
              API
            displayName: Vertica Replicator
            kind: VerticaReplicator
            name: verticareplicators.vertica.com
            version: v1beta1
          - description: VerticaRestorePointsQuery is the Schema for the verticarestorepointsqueries
              API
            displayName: Vertica Restore Points Query
            kind: VerticaRestorePointsQuery
            name: verticarestorepointsqueries.vertica.com
            version: v1beta1
          - description: VerticaScrutinize is the schema for verticascrutinize API
            displayName: Vertica Scrutinize
            kind: VerticaScrutinize
            name: verticascrutinizers.vertica.com
            version: v1beta1
        description: |-
          ### What is Vertica?
          Vertica is a unified analytics platform, based on a massively scalable architecture with the broadest set of analytical functions spanning event and time series, pattern matching, geospatial and end-to-end in-database machine learning. Vertica enables you to easily apply these powerful functions to the largest and most demanding analytical workloads, arming you and your customers with predictive business insights faster than any analytics data warehouse in the market. Vertica provides a unified analytics platform across major public clouds and on-premises data centers and integrates data in cloud object storage and HDFS without forcing you to move any of your data.

          ### What does this operator do?
          This operator manages a Vertica Eon Mode database, and automates the following administrative tasks:

          * Installing Vertica
          * Creating and reviving a Vertica database
          * Restarting and rescheduling DOWN pods to maintain quorum
          * Upgrading Vertica to a new version while keeping the database online
          * Subcluster scaling, both manually and automatically with the Horizontal Pod Autoscaler
          * Service management and health monitoring for pods
          * Load balancing for internal and external traffic
          * Expose a metrics endpoint to integrate with Prometheus to monitor the operator's health

          For a brief overview on how to create a cluster, see the [Vertica GitHub repository](https://github.com/vertica/vertica-kubernetes). For an in-depth look at Vertica on Kubernetes, see the [Vertica documentation](https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm).
        displayName: VerticaDB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - vertica
        links:
        - name: Documentation
          url: https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm
        - name: Vertica Container Images
          url: https://hub.docker.com/u/vertica
        maintainers:
        - email: cchen8@opentext.com
          name: Cai Chen
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Vertica
          url: http://www.vertica.com
        relatedImages:
        - docker.io/opentext/verticadb-operator:25.3.1-0
        version: 25.3.1-0
      entries:
      - name: verticadb-operator.v25.3.1-0
        version: 25.3.1-0
      - name: verticadb-operator.v25.3.0-0
        version: 25.3.0-0
      - name: verticadb-operator.v25.2.1-0
        version: 25.2.1-0
      - name: verticadb-operator.v25.2.0-0
        version: 25.2.0-0
      - name: verticadb-operator.v25.1.0-0
        version: 25.1.0-0
      - name: verticadb-operator.v24.4.0-0
        version: 24.4.0-0
      - name: verticadb-operator.v2.2.0
        version: 2.2.0
      - name: verticadb-operator.v2.1.3
        version: 2.1.3
      - name: verticadb-operator.v2.1.2
        version: 2.1.2
      - name: verticadb-operator.v2.1.1
        version: 2.1.1
      - name: verticadb-operator.v2.1.0
        version: 2.1.0
      - name: verticadb-operator.v2.0.2
        version: 2.0.2
      - name: verticadb-operator.v2.0.1
        version: 2.0.1
      - name: verticadb-operator.v2.0.0
        version: 2.0.0
      name: v2-stable
    defaultChannel: v2-stable
    packageName: verticadb-operator
    provider:
      name: Vertica
      url: http://www.vertica.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: F5 Networks Inc.
      provider-url: ""
    name: f5-ipam-controller-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: f5-ipam-controller-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fic.f5.com/v1",
                "kind": "F5IpamCtlr",
                "metadata": {
                  "name": "f5ipamctlr-sample"
                },
                "spec": {
                  "args": {
                    "infoblox_grid_host": "IP_ADDRESS_OF_SERVER",
                    "infoblox_labels": "CIDR_Labels",
                    "infoblox_netview": "NetView",
                    "infoblox_password": "PASSWORD",
                    "infoblox_username": "USERNAME",
                    "infoblox_wapi_version": "VERSION",
                    "insecure": true,
                    "ip_range": "IP_RANGE_LABELS",
                    "log_level": "DEBUG_OR_INFO",
                    "orchestration": "kubernetes",
                    "provider": "INFOBLOX_OR_F5-IP-PROVIDER"
                  },
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "f5-ipam-controller",
                    "user": "f5networks",
                    "version": "latest"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "serviceAccount": {
                    "create": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
          createdAt: "2025-08-08T03:48:02Z"
          description: Operator to install F5 IPAM Controller.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/f5-ipam-controller
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This Operator installs F5 IPAM controller(FIC)
            kind: F5IpamCtlr
            name: f5ipamctlrs.fic.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 IPAM Controller in your OpenShift Cluster. This enables to configure and deploy FIC using Helm Charts.
          ## F5 IPAM Controller
          The F5 IPAM Controller is a Docker container that runs in an orchestration environment and interfaces with an IPAM system. It allocates IP addresses from an IPAM system’s address pool for hostnames in an orchestration environment. The F5 IPAM Controller watches orchestration-specific resources and consumes the hostnames within each resource.
          ## Documentation
          Refer to F5 IPAM Controller documentation
          - https://github.com/F5Networks/f5-ipam-controller/blob/main/README.md
          ## Prerequisites
          ### Configure Infoblox with network and netview [Provider: Infoblox]
          Refer Infoblox documentation
          ### Create persistent volume and persistent volume claim [Provider: f5-ip-provider]
          Create persistent volume and persistent volume claim
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/f5-ipam-controller/main/docs/config_examples/f5-ip-provider/localstorage-pv-pvc-example.yaml
          ```
        displayName: F5 Ipam controller
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipam
        - ip
        - ipaddress
        - controller
        - ipv4
        - ipv6
        - F5
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/userguide/ipam/
        - name: Github Repo
          url: https://github.com/F5Networks/f5-ipam-controller
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
        - registry.connect.redhat.com/f5networks/f5-ipam-controller@sha256:09a358db300d82e00fd3b4e255917df82c02e1bc58425f8cff21cc2c77eba1b6
        version: 0.0.8
      entries:
      - name: f5-ipam-controller-operator.v0.0.8
        version: 0.0.8
      name: alpha
    - currentCSV: f5-ipam-controller-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fic.f5.com/v1",
                "kind": "F5IpamCtlr",
                "metadata": {
                  "name": "f5ipamctlr-sample"
                },
                "spec": {
                  "args": {
                    "infoblox_grid_host": "IP_ADDRESS_OF_SERVER",
                    "infoblox_labels": "CIDR_Labels",
                    "infoblox_netview": "NetView",
                    "infoblox_password": "PASSWORD",
                    "infoblox_username": "USERNAME",
                    "infoblox_wapi_version": "VERSION",
                    "insecure": true,
                    "ip_range": "IP_RANGE_LABELS",
                    "log_level": "DEBUG_OR_INFO",
                    "orchestration": "kubernetes",
                    "provider": "INFOBLOX_OR_F5-IP-PROVIDER"
                  },
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "f5-ipam-controller",
                    "user": "f5networks",
                    "version": "latest"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "serviceAccount": {
                    "create": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
          createdAt: "2025-08-08T03:48:02Z"
          description: Operator to install F5 IPAM Controller.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/f5-ipam-controller
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This Operator installs F5 IPAM controller(FIC)
            kind: F5IpamCtlr
            name: f5ipamctlrs.fic.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 IPAM Controller in your OpenShift Cluster. This enables to configure and deploy FIC using Helm Charts.
          ## F5 IPAM Controller
          The F5 IPAM Controller is a Docker container that runs in an orchestration environment and interfaces with an IPAM system. It allocates IP addresses from an IPAM system’s address pool for hostnames in an orchestration environment. The F5 IPAM Controller watches orchestration-specific resources and consumes the hostnames within each resource.
          ## Documentation
          Refer to F5 IPAM Controller documentation
          - https://github.com/F5Networks/f5-ipam-controller/blob/main/README.md
          ## Prerequisites
          ### Configure Infoblox with network and netview [Provider: Infoblox]
          Refer Infoblox documentation
          ### Create persistent volume and persistent volume claim [Provider: f5-ip-provider]
          Create persistent volume and persistent volume claim
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/f5-ipam-controller/main/docs/config_examples/f5-ip-provider/localstorage-pv-pvc-example.yaml
          ```
        displayName: F5 Ipam controller
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipam
        - ip
        - ipaddress
        - controller
        - ipv4
        - ipv6
        - F5
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/userguide/ipam/
        - name: Github Repo
          url: https://github.com/F5Networks/f5-ipam-controller
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
        - registry.connect.redhat.com/f5networks/f5-ipam-controller@sha256:09a358db300d82e00fd3b4e255917df82c02e1bc58425f8cff21cc2c77eba1b6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        version: 0.0.8
      entries:
      - name: f5-ipam-controller-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: f5-ipam-controller-operator
    provider:
      name: F5 Networks Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: smart-gateway-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smart-gateway-operator.v5.0.1757650606
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "smartgateway.infra.watch/v2",
                "kind": "SmartGateway",
                "metadata": {
                  "name": "demo"
                },
                "spec": {
                  "applications": [
                    {
                      "config": "host: localhost\nport: 8081\nwithtimestamp: false\n",
                      "name": "prometheus"
                    }
                  ],
                  "blockEventBus": false,
                  "bridge": {
                    "amqpBlock": false,
                    "amqpUrl": "amqp://amq-interconnect:5672/collectd",
                    "enabled": true,
                    "ringBufferCount": 15000,
                    "ringBufferSize": 2048,
                    "socketBlock": true,
                    "statsPeriod": 60,
                    "stopCount": 0,
                    "unixSocketPath": "/tmp",
                    "verbose": true
                  },
                  "handleErrors": false,
                  "logLevel": "info",
                  "pluginDir": "/usr/lib64/sg-core",
                  "services": [
                    {
                      "name": "prometheus",
                      "ports": [
                        {
                          "name": "prom-http",
                          "port": 8081,
                          "protocol": "TCP",
                          "targetPort": 8081
                        }
                      ]
                    }
                  ],
                  "size": 1,
                  "transports": [
                    {
                      "config": "path: /tmp/smartgateway\n",
                      "handlers": [
                        {
                          "config": "",
                          "name": "collectd-metrics"
                        }
                      ],
                      "name": "socket"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/stf/smart-gateway-rhel9-operator@sha256:3363abcb91319f8642212b9b80ca49bb9f90cae69255b1f75f03929af3d3a8c3
          createdAt: "2025-09-12T04:17:59Z"
          description: Operator for managing the Smart Gateway Custom Resources, resulting
            in deployments of the Smart Gateway.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenStack Platform", "Cloud
            Infrastructure", "Cloud Suite"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/infrawatch/smart-gateway-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of the Smart Gateway
            displayName: SmartGateway
            kind: SmartGateway
            name: smartgateways.smartgateway.infra.watch
            version: v2
        description: |-
          # Smart Gateway for Service Telemetry Framework

          The Smart Gateway is middleware for OpenShift that takes metrics and events
          data streams from an AMQP 1.x message bus, and provides a method to having
          that data stream stored within supported data stores such as Prometheus and
          ElasticSearch.

          ## Core capabilities

          The Smart Gateway provides two modes:

          * metrics: provides an HTTP scrape endpoint for Prometheus
          * events: writes events directly to an ElasticSearch endpoint

          ## Operator features

          * **Two Service Types** - Supports the `metrics` and `events` service types which defines the type of Smart Gateway to be deployed

          * **Configuration** - Configuration of the Smart Gateway to allow for it to be connected to an AMQP 1.x bus and pointed at the appropriate storage mechanism based on service type.

          ## Before getting started

          1. Ensure an AMQP 1.x bus has been setup (such as AMQ Interconnect)
          2. If the Smart Gateway is running in events mode, be sure the ElasticSearch Operator and ElasticSearch instance have been pre-deployed.
        displayName: Smart Gateway Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - servicetelemetry
        - monitoring
        - telemetry
        - notifications
        - smartgateway
        links:
        - name: Source Code
          url: https://github.com/infrawatch/smart-gateway-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/stf/smart-gateway-rhel9-operator@sha256:3363abcb91319f8642212b9b80ca49bb9f90cae69255b1f75f03929af3d3a8c3
        - registry.redhat.io/stf/sg-core-rhel9@sha256:0e12903da0b15c64421102b5b914ca4c82d107295485c9b76ed19e91a45a4420
        - registry.redhat.io/stf/sg-bridge-rhel9@sha256:7fae342cf6cecc676eaa0a8fb64ee380ad101d667d6e8f0d89c943938bb6baf4
        - registry.redhat.io/openshift4/ose-oauth-proxy:latest@sha256:dc24aa382e02c4ac058d4b824a21f0e90a6259d6d817d0341b2a41809b5daa78
        version: 5.0.1757650606
      entries:
      - name: smart-gateway-operator.v5.0.1757650606
        version: 5.0.1757650606
      - name: smart-gateway-operator.v5.0.1743671602
        version: 5.0.1743671602
      - name: smart-gateway-operator.v5.0.1740051852
        version: 5.0.1740051852
      - name: smart-gateway-operator.v5.0.1731336910
        version: 5.0.1731336910
      - name: smart-gateway-operator.v5.0.1725953602
        version: 5.0.1725953602
      name: stable-1.5
    defaultChannel: stable-1.5
    packageName: smart-gateway-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-pipelines-operator-rh
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-pipelines-operator-rh.v1.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    },
                    {
                      "name": "legacyPipelineRbac",
                      "value": "true"
                    },
                    {
                      "name": "createCABundleConfigMaps",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.19.0 <1.20.0'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.3.2
          - Tekton Triggers: v0.33.0
          - Pipelines as Code: v0.37.0
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.22.3
          - Tekton Results (tech-preview): v0.16.0
          - Manual Approval Gate (tech-preview): v0.7.0
          - Tekton Pruner (tech-preview): v0.2.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.20/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:5021158a4752a26f8b33adea7da66695b22d2ad155d346d35dde5f971bc1a55b
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:5683321a5376e63c3ff3932270a1850d5e261f76212020ce867334166f90d2f7
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:a712bb411d2433b01793dff25785c1e265fa865a52f7c8732c9fb4418ab61757
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e7998e113f75e8f29602bae210c8b66ed06634b699f1d99c5875ef879220d733
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:96db721939b22572d7d8d8b95dcb5c92fdedb98e9cd0aabb7d17b703a915b07f
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:115eea1a5ca87f3b6b72db1df05b7386bf73dc70c9c2e4dc5e1a66700d960c37
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:980ed7e81ea55522e1681dc1e9b170b68d16f07061dd1f1ffd142367d59552f1
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:1b6aecfce45ed3163227b845b212e83385cdbe999edfc2e9387a0c76f3f8d8e9
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:161df929e2bd4c5acff594be31d5458365de5ae0d2189bb2f647c3fc6657fa28
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:bc10e7d6122ede5cf4f2882c200447683255465a1080e826761e90bfb983506e
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:df15eb1f1f5f237bfe481c16ab3d40bf7fb4123f3c930bcfb121e2cc8dc42882
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:d56d6b46331b4569077ab1e7a7b609ba01f0a1dd62b695d62f8aa0f32ea1b621
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:c0f438e6095aa81f3f7d820614e0414cc43374014c1505d7641bc1dba69a3903
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:818bfa4a3a874974b5f7f8bcf2faeb6c05e1795381ee1ceacf89f17b1ec0ef7f
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:9415e889ef98df6d2b2dca1e5d798fb765eeea0d9828e48424374d5267a6b0d1
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:441e1c7124c1674e3d4c38c3e82a4fe5568e76d13c89c04bf009d6b97435de09
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:7eafffa6e4e04f10c1d4e84e02d51c00eb8649c60f8f9a13da14492f24634891
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:bc403f93ef6c5526fa30152eab384c4a25bb0bc6ab4158a1e24f637b87a57259
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:2d496752b19f765f1ea36ec94b873ebd8fa604764a56aa67d3ed94b33e5fe75c
        - registry.redhat.io/rhel9/buildah@sha256:18453ab0a62154283aaf4c45efd7543d1c8ecca1aebae46dedc7ddc0f410b232
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:036250bcdc6fb95273bf07948dfdd4dce765ab84c803ceca0d0b4abf6ca19dec
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:8f0494341244a75546a4b80b065702be04e6d200421aa4e7ec90385c9dabc6ab
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:c4d05e546c466c57379fff502816012f615be073eb8cdce54677b4ab206a3f5c
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:bbd1918660be7b3c78777dcccc87d3b26bf5adfc077844ceffbbdd53696dc0a3
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:e3e64abef53151f6b15f15bcbff85668167a571e98badbd8a2ca252ff3a5b8dc
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:dd941acd357ae2c0fd7e15fb8d32d7ee6b2847435101dd1e1f107db48af18bf3
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:823e018e0329d764ac694f177ce88af9074f364c29a1c17929f9dd0b596d03dd
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:a7163a9e1fe891de021b1003558e6e64fe17ed17ab0820ba16bba09aa112401e
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:e63298d07ccec5804dfab4e42bcc8163c29cb3ac371c91248777423a4edd31ce
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:3cf990fb2fc422f3527a7c8b43d06ccd4d32bb5487ec8cb2d4a97f0f5d13f170
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:c3013c347bca8a520443a44ac2817fbba5744a9e91069ac5f09e91447ebf6b28
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:901b5108ceb669c985e75f5fb033c496e9f028863df99ab32651d1b08433a352
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:14dab8d87323fc7d112213f7f4db3ee7cf3841747de669ce710ae6a794053457
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:458f865f1b246c5f0682f5956795771a0515b89edfcffbbc0da1a3d3f532645e
        - registry.redhat.io/ubi9/ubi-minimal@sha256:2f06ae0e6d3d9c4f610d32c480338eef474867f435d8d28625f2985e8acde6e8
        - registry.redhat.io/ubi9/openjdk-17@sha256:555c8bcd2596637befd1a6da4aa949c5a2640cf66e09c2dfb932afe37e39095e
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:c31af237afb84214bf2a4c06a7636ca071fe97203b41da90d7878109ccd64784
        - registry.redhat.io/openshift-pipelines/pipelines-opc-rhel9@sha256:c1bee4a2bb495de6d4da962b6e7b2447caa51cc36c3005e4083bd31420ef5102
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:ec74a562c18fefe36ec7ae3ba083e4eb76c56b562f916578900f8767ae98e6c2
        - registry.redhat.io/rhel9/skopeo@sha256:a65a413f8a2864389a09dc750690d97afbcdc5c70821e0f85e99e8adba7954e0
        version: 1.20.0
      entries:
      - name: openshift-pipelines-operator-rh.v1.20.0
        version: 1.20.0
      - name: openshift-pipelines-operator-rh.v1.19.3
        version: 1.19.3
      - name: openshift-pipelines-operator-rh.v1.19.2
        version: 1.19.2
      - name: openshift-pipelines-operator-rh.v1.19.1
        version: 1.19.1
      - name: openshift-pipelines-operator-rh.v1.19.0
        version: 1.19.0
      - name: openshift-pipelines-operator-rh.v1.18.1
        version: 1.18.1
      - name: openshift-pipelines-operator-rh.v1.18.0
        version: 1.18.0
      - name: openshift-pipelines-operator-rh.v1.17.1
        version: 1.17.1
      - name: openshift-pipelines-operator-rh.v1.17.0
        version: 1.17.0
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      - name: openshift-pipelines-operator-rh.v1.14.1
        version: 1.14.1
      - name: openshift-pipelines-operator-rh.v1.13.0
        version: 1.13.0
      - name: openshift-pipelines-operator-rh.v1.12.2
        version: 1.12.2
      - name: openshift-pipelines-operator-rh.v1.12.1
        version: 1.12.1
      - name: openshift-pipelines-operator-rh.v1.12.0
        version: 1.12.0
      - name: openshift-pipelines-operator-rh.v1.11.1
        version: 1.11.1
      name: latest
    - currentCSV: openshift-pipelines-operator-rh.v1.14.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "db_user": "result",
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.14.6'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.56.4
          - Tekton Triggers: v0.26.2
          - Pipelines as Code: v0.24.6
          - Tekton Chains: v0.20.1
          - Tekton Hub (tech-preview): v1.16.0
          - Tekton Results (tech-preview): v0.9.2

          ## Getting Started
          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/container-platform/OPENSHIFT_DOCS_VERSION/cicd/pipelines/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)
          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi8/openjdk-17@sha256:a8165bc2cd5051a96d6937e25ed03155bbd4b731da6e58cebfe2ea83209c16d8
        - registry.redhat.io/ubi8/ubi-minimal@sha256:6825946f10e8c81586fa7815c4f06016af05cfc231d921ec9699402a6655d201
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:059c08316152c209de13017e322d6fe5eacfde81c3b0039e55b5de8f2f27183a
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:d01ba7ead670d5ab9c1355b90010657988e46b4d1aa0bb68b194eaf4001b2d7a
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:6327385e3d3c9df1ea50e803955d980b4cd35fb15593fcfec0d477d91a45021f
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel8@sha256:6e7c7a1ca458e08ad85e1146815ec00a30016f709b45e28b464043e875c2a3e8
        - registry.redhat.io/rhel8/skopeo@sha256:e297b84ba6f1d5eb2c37e2c3b31c9a9a4c39f9590e7d7c32388165245bde4835
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:ae0260abd2725c2f4bb25c0c07d86290978892a49050e0fd23f466401aba668c
        - registry.redhat.io/rhel8/buildah@sha256:b48f410efa0ff8ab0db6ead420a5d8d866d64af846fece5efb185230d7ecf591
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:2261f4f657f70759f4d0c898e44ffafceb8dbe2144063d50227140463bdbf2c5
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:7a63e7a762b9c7b15057ed9d5bcbea08acb4c8933a300c6ec336824d49a6e937
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:8d4450d68ac03d8275e337d2c55692ef35e954400d0c9f041a943d163a8b7680
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:bf6cf2e87fb19f7aa9a490b83c16af69834c0721220a643710a1b077959e91ca
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:aa6eb4c7ade9d6add1e4ab94197c2ce72a253fea0d7461698ca0e37854f1dcbe
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:71d91c7b80b6d196025f68749bd8518b8b0affd8d5064a7c5c70d125656fbc36
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:db3ff043410d1dd0621fef4483761727f7241d788b136e446fb861953350a2cb
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:5e7f5a2813a591e7b13fea9fc4e2f599b5ec855a015cd5a4c287c4e178358756
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:9c29f6eff1f7019993446d85d226f04c5614a18351dfb4042a8bcb62276810de
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:d9972bc64dfa7e260bd9b949bbb85ef5ae532e7c6df8d874d3d3963af1097726
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:bf0c38beefa24ca3d04ebfa6140a1cc5a9a56fba81ae05799279a559093b9033
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel8@sha256:b3bedee6084e9d9da549d3e637d2bf176aa365b5818e6b6377525577edd710dc
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:82c8a91d521d7e6d81fb462fa2b804f420f52ee062df5cda39545165a2f19e0c
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:ce0ea093c522af329acdc85148c2473ea45cd44270c15f8dd32965b0d56f304f
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:abf4442b6acc714c2392b86af046d5f32d247ee0c3763bb18c57d0473edfcf02
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:e2bd929246d62038a24a0685cb3e4873c131c9b7713d6a5268ca09a8614538be
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:83913744a1bb7ff23091cfa541b807cfb82b38dde0c37e0f1dd9c17cade99775
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:63fc1351191a54d12d4e463a1bb4cc732620cba3000af86ba7f7a579514f5fe4
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:e23c901868229e4f6a4016aa7798f1aba879ff35ffe594f5b0e4d39f1bff9e52
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:f68dc18a4dc6a1886507097551f4a6fce39062d03c6eb86c067d317812b0d34a
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:f9fed20cdfebbb226e035cc62f037174f5ecdb3139a5edc7e39f019f6d2b1e98
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:ddf0cb64ea57de872515e19de63b2dec5f56738c9d7eaa283a4a489545ef4902
        - registry.redhat.io/rhel8/postgresql-13@sha256:3903d562425dd564986123280dc076548fb3d24c74138dcd857652f53052e0a7
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:2e75096ffe20d1261e8a645fefd32e37295d727aa6a280f10041a44443917469
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:57fb86ba112d79a228cbacc96cd0525c123464bcac956354722e3f2aeff42508
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:ffa8c6d8fb5bacb59da27f4ebde87385359b3f97a6a023b4a4488ced2d46c574
        version: 1.14.6
      entries:
      - name: openshift-pipelines-operator-rh.v1.14.6
        version: 1.14.6
      - name: openshift-pipelines-operator-rh.v1.14.5
        version: 1.14.5
      - name: openshift-pipelines-operator-rh.v1.14.4
        version: 1.14.4
      - name: openshift-pipelines-operator-rh.v1.14.3
        version: 1.14.3
      - name: openshift-pipelines-operator-rh.v1.14.2
        version: 1.14.2
      - name: openshift-pipelines-operator-rh.v1.14.1
        version: 1.14.1
      - name: openshift-pipelines-operator-rh.v1.13.0
        version: 1.13.0
      - name: openshift-pipelines-operator-rh.v1.12.2
        version: 1.12.2
      - name: openshift-pipelines-operator-rh.v1.12.1
        version: 1.12.1
      - name: openshift-pipelines-operator-rh.v1.12.0
        version: 1.12.0
      - name: openshift-pipelines-operator-rh.v1.11.1
        version: 1.11.1
      name: pipelines-1.14
    - currentCSV: openshift-pipelines-operator-rh.v1.15.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "db_user": "result",
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.15.3'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.59.6
          - Tekton Triggers: v0.27.0
          - Pipelines as Code: v0.27.2
          - Tekton Chains: v0.20.1
          - Tekton Hub (tech-preview): v1.17.1
          - Tekton Results (tech-preview): v0.10.0
          - Manual Approval Gate (tech-preview): v0.2.2

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.15/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:322ef2385620661a7914f47c73a6d145ec0c9ce1ccde30645bb66148545ddc96
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:49c346bf7fc69a9796690236dc8597e35a874343b02fb74c5d2f1c05d356e41c
        - registry.redhat.io/rhel8/skopeo@sha256:29465586e92ed04d2878dcdeda2d508c7e8e7539762e10296b52d631929960aa
        - registry.redhat.io/rhel8/buildah@sha256:aac6629389db17e99894c5bee0da01d4c8065d11d8c6f6e1602f9484290baa70
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:bf94c350923da706c6e0a427e2d91af3e13e751a12388a4ed75d1c34dd7b3177
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:ac1a6e057a5599fda7e048ff77030e98c92b5d51cac2472c1ef3910964fc17e6
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:61c22119f37e620a3807056cde35bc07c90867765b149294056326b1d11494f1
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:abac279a0a5ecde57952b6bef634f8ba7527cd8c8956a2ea30e9449174ad52bd
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:166789a990df9f6ab04fd3df0a8cd301bc19b9bedde94215cb27306060eb1e2c
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:83eb08bb9579cc19132af282c455bca32a0fd4f5436a35e7ad23b5efa2faa3cf
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:479e3eed93e0aa13e3489514fe00375a519eb45969d099c490b4acd056ea5adf
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:d1f86458079449c86407ac716c3cbf608cad06de29edeecbde6586f76628daa9
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:e749db81c9d7e11e2802833a2a02e93afc357c83d1a2a827e537262df281a4f7
        - registry.redhat.io/ubi8/ubi-minimal@sha256:f729a7f5685ea823e87ffd68aff988f2b8ff8d52126ade4e6de7c68088f28ebd
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:7c81ae208a16e17d0102074867add37dfb1207ee95cbbc6c732d88e61bbc59fb
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:7152c3c174d25f2a5239456ba677ac5315bfb82bc3b3ac5ee0eb8f9f74fb6632
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:aca0024cac7a8c41b1ed151eb1e1188f09bc636866f4e03f1b973f90890b37cf
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:bf6cf2e87fb19f7aa9a490b83c16af69834c0721220a643710a1b077959e91ca
        - registry.redhat.io/rhel8/postgresql-13@sha256:a92a579f1aef66ac188d24fd489c456a1a3e311d95dcce652da6b81e28fbf725
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:7f00ff1b407d9ea107d3d969988a6279b28bde86943f490a007fa922dd40a1c3
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:eed464a3b4e23f0b2623aa1b923c10232d235f4272f562955926c1748e295d2b
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:7958925cecce1b4b02ac4649e3bfd7fa3b5361c16ab3593700b8da93fdf1c79a
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:312d1f90765d05a86845856c677e1ece4c57e81a9bfe3fd1e76868d624e3d72d
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:1300d171dbf09c63071ba5586e223491a13acfc6dbab27eff04648aeb788d455
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:538c63c8e7af3cba91526f2f9ac99fb8f6b0f4781db31771c959612c209dd090
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:601fcac1abc6c8efa1248c5ca229213d6660708470a1c416d41c6aff0ba4691a
        - registry.redhat.io/ubi8/openjdk-17@sha256:e8cc2e476282b75d89c73057bfa713db22d72bdb2808d62d981a84c33beb2575
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:e7b000a2735bbf98ee79bd093f7b703ed09d7c1e11ed2a5b22c4b8ac8d440386
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:827313c2691fb8833e42418b65b4a3360a11cabc2c92a7d8e7b20183f06cbf68
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:f02af8475c33f6f38a2f90ff1be1fb1b2094860c4d5e300aefdd7eb84b4a529a
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:acf8017029f4b2e7a1c3da9bc62c65239d313d24e3012712356a191bea9a2e6f
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:75d6228eb656b8778fb3bb8dc68c686ac1c9b787e5343ae3c2467f8e724af83e
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:1db14f23077785108660dde38c98e2e30c8301f8e2e1c594dd4636e0ba29f84f
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:70bb0ebcfe5da45e5c239aa011dc5f1ef48844116d7346bb05433b503107268a
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:f04ea12675e0eb2354635cd2f9e99d6e59619d1c611e0ea95c3ce8222d615322
        version: 1.15.3
      entries:
      - name: openshift-pipelines-operator-rh.v1.15.3
        version: 1.15.3
      - name: openshift-pipelines-operator-rh.v1.15.2
        version: 1.15.2
      - name: openshift-pipelines-operator-rh.v1.15.1
        version: 1.15.1
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      - name: openshift-pipelines-operator-rh.v1.14.1
        version: 1.14.1
      - name: openshift-pipelines-operator-rh.v1.13.0
        version: 1.13.0
      - name: openshift-pipelines-operator-rh.v1.12.2
        version: 1.12.2
      - name: openshift-pipelines-operator-rh.v1.12.1
        version: 1.12.1
      - name: openshift-pipelines-operator-rh.v1.12.0
        version: 1.12.0
      - name: openshift-pipelines-operator-rh.v1.11.1
        version: 1.11.1
      name: pipelines-1.15
    - currentCSV: openshift-pipelines-operator-rh.v1.16.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.16.4'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.62.9
          - Tekton Triggers: v0.29.2
          - Pipelines as Code: v0.28.2
          - Tekton Chains: v0.22.3
          - Tekton Hub (tech-preview): v1.18.2
          - Tekton Results (tech-preview): v0.12.3
          - Manual Approval Gate (tech-preview): v0.3.1

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.16/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:81a8b903d0fc4b81e6a23fac862d18da4cc18b14cca8213665df37dfaabbc2e9
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:5672f3a375c6a274200b423eb2643c818aba75488dec0b210fec2532814ba51e
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:47f094ccd12db7a43c8a91c1a3d2a190bc2b83f1745e7eb8ce097d0c8e1e2c88
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:80d46bc5ce8c931af22a97ccbce05b59740e313167c88ad910febcc2f2dc0320
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:90e2a7ea498da756e09c7e4d19c7fb91d429cc28c57f97c6958958e3466cb58c
        - registry.redhat.io/rhel8/skopeo@sha256:1dcb4189c92da21ee58b3ca5433615e0208194c8412f7f1fb45fa491179aaf9b
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:d2bb3f9f79ef719d8fdafef3a639e3eb8bc199ee56fcfb33e85975d536d2a6ac
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:7edbd11e5a449408d673cae1fc2b9b5bab492ecff2ac6e640c8cd69a9b90ff08
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:a5a7b2be50f211d4d9d555b2a2e2302096c139ff5011d984cc386f1530007f16
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:f9122d71c3862cd2afafc8eb7f1a39c83eee306e5baf38514e6e39215c70a72b
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel8@sha256:8cb87920e6fc8185ead49aafb7799fead6a4a510eafdae0ee29e85248dae3945
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:aacdabe243dc573f3161a3d04fbf014b77d9962207305d70fcafce475f7eea93
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:92042dcacd4ab0f2a78aab03fd91a9a39d7d7366b5398fd797ecb145acf5fce4
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:2fea7d8e29d70c5893ddb6caf2bf1d4144787f5f0de8739351d2097ed09ee751
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/rhel8/buildah@sha256:5c7cd7c9a3d49e8905fc98693f6da605aeafae36bde5622dc78e12f31db3cd59
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:474658c262b539dd1c83ac22bb090b3fdac5de9cffc06587d36dbc6fa4ce208c
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:e1fe4a86211cf22e5ad2b23c989d24b69f597db4c3604e84777dbd0c585ec186
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:6acc31f179b08658976cddcb9ef42f377edb29933d1d83a2ecebfd455bb68571
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:f3e6df07fd86acb91037d55a1970e5f8c6ab12a7517752692c224c7687022812
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:8129e62906c68f04075a44c939375a7bafdeaeac673a20a4fa74ffd6285187ac
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:7d92dca9da27c066cb530172e1811493d8c916e67d8068f5ddc471fa3f3f7856
        - registry.redhat.io/ubi8/ubi-minimal@sha256:a47c89f02b39a98290f88204ed3d162845db0a0c464b319c2596cfd1e94b444e
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:b94bd566ae3d9ce5f47273d867164486826b19fca0ce11d1a7d18312413e87ba
        - registry.redhat.io/ubi8/openjdk-17@sha256:632e78989471859ca4ed1148b951a911f7e3b6e6516482be20967c4171612c55
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel8@sha256:46635d636ea2e9a6ba0760bf8c528028700ee40b0ba8466797034371185b16d0
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:856229e9e37683c926e3449658d76f93d910384eff62c2ca192768a653b71533
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:c6d654c71aba35c3c7f5c1a2ca5b8401e95adf6f649c599a9aff25177c2b3323
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:549d74a18ec52db91fa949cac9b8e3d943d478d5dec9850d2ba0cb1bb526f578
        - registry.redhat.io/rhel8/postgresql-13@sha256:a308263f5d1071be8ac5c13fd84b40609988ca30ed56c22a8bc6769e107ce09b
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:b41b888fc1a3b7215e6b71ee97bf64a1816914a8f33e96ad3930a4e2d5b5ffa7
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:4e0e4c358faa9cafc3964d6d40ee4aa73f8dfe8b6338dfb8206613d7c5241842
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:dcab4a3d519975340ade59a0ca0222104e68acd052a61f2aeb8c736be4b672cb
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:e84c2d80e92815c6bf384e1504a194dec9b796a3aceac2eed555fff241249571
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:d36dbb60c7b5caca7f3be08705f8ab91d7f01abaf8bfee4aee73bdbbb3e7e71a
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:7da7f929d2711fc55dda01fe51a474bfa7adf5aaf78be6d65a6d65850958d6e7
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:19b3dd69c1d7e17890b7163aef322c1efbc76fed7f2b1faf26aaa05fd8161184
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:aeeed543412cde0bd3a9baec965cb99d80f2b6acb77ae6356ed51b4b1262d006
        version: 1.16.4
      entries:
      - name: openshift-pipelines-operator-rh.v1.16.4
        version: 1.16.4
      - name: openshift-pipelines-operator-rh.v1.16.3
        version: 1.16.3
      - name: openshift-pipelines-operator-rh.v1.16.2
        version: 1.16.2
      - name: openshift-pipelines-operator-rh.v1.16.1
        version: 1.16.1
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      - name: openshift-pipelines-operator-rh.v1.14.1
        version: 1.14.1
      - name: openshift-pipelines-operator-rh.v1.13.0
        version: 1.13.0
      - name: openshift-pipelines-operator-rh.v1.12.2
        version: 1.12.2
      - name: openshift-pipelines-operator-rh.v1.12.1
        version: 1.12.1
      - name: openshift-pipelines-operator-rh.v1.12.0
        version: 1.12.0
      - name: openshift-pipelines-operator-rh.v1.11.1
        version: 1.11.1
      name: pipelines-1.16
    - currentCSV: openshift-pipelines-operator-rh.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.16.0 <1.17.2'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.65.7
          - Tekton Triggers: v0.30.1
          - Pipelines as Code: v0.29.1
          - Tekton Chains: v0.23.1
          - Tekton Hub (tech-preview): v1.19.2
          - Tekton Results (tech-preview): v0.13.4
          - Manual Approval Gate (tech-preview): v0.4.1

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.17/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:a14270967460883f283ec1ce17ff8b82db1e0777dda4f6d690f15ac38dcbad85
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:650391ffd3e4c7b173bf9d718405b2375c6ff2027abc0b7ce4f67e79e7ef4b81
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:37109f6de75dda750496f6668164d0dceaf9db5278e8c034708436f39de3276e
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:a392d2c85baedd7428c579e91d9f8acf8232399b90523b1093dc3447aaf56a16
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:8e52b4b31b0d9ec772abe6e6ef9c6698e89893395f364968216bebc3367da4b2
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:4b87782b35e7284a365c49f1c094aae98531ac630db1958294009c041a2d8440
        - registry.redhat.io/ubi8/openjdk-17@sha256:e5fc9e34382da41b10d4ea566639d31707493abff9e8de81f5accfd66876409e
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:4224040b00cfd76e85f80aecc5095098564cc62e9c477fe2f7a428b30df7d4d9
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:41ec6db921812fd1e80b05d51c224ddbce63bf0c57f02ccf07461d2202e97b90
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:4fabae1312c1aaf8a57bd2de63bd040956faa0c728453f2a4b4002705fba0f0c
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:7abc88bea01dd6513f6e48334f0b98501f909a1dc1ab9e83efeb3c1a834f8f92
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:eab798472b76bd51235198fd3b95189740f8d650ebc6bb01fe653c025c1defaa
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:61c9e8b1660d8935ee316ecc113e8141daf05149d172685cabd861ae8aeed3eb
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:4240501c076605a01d969328bbb8ad885c598b44ec1dc1c066312491a0eb6715
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:7c710882752f12be3c22bedc8d32096c111e42c5b9b392163f61cf8d72be3ec0
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:66309ae437c4647f48efe3d6f2b696df83859323700427d0025b94795751a859
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:41e2f2c7397a8b9e093e18977666e93943734f51780c769c99958ae1d20af53e
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:4a96f1bb7e3a5c7ca62dff2199fa15ef964b129e4e36e9619837c1221538f496
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:3696cf724ce7865626ae408d82fd87727c741e4d3eb3475c271ec6d56f0882a4
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:44da76d75f567c2ba623bae646e1d552d2c0fded1accdde3aab4f4bcdf3adc28
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:26f7b695fc1430359ab2bfd2915cf0b7e88b06d8695e26761bdc2ce8935f0fb3
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:0b6ee6cd98cb7e835cfc9544996c6f85288d7252071c100bed853eb593d6befb
        - registry.redhat.io/rhel8/buildah@sha256:6d2dcb651ba680cf4ec74331f8349dec43d071d420625a1703370acc8d984e9e
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:d851a2dd8ec95bfd7ac6ec5515c8cdadfcbfc6e6ec84b40630b4fe595f978fa5
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:429bc3d6044ca7dab780ff4b5eb91d42d38a29eb7ef428a25c53b7dda49871bf
        - registry.redhat.io/rhel8/postgresql-13@sha256:757d8dd6f243717d820846da7144a49913f28f223b74ec305b2215f54720f8a0
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel8@sha256:b5cceead6b8e9c99655c9ba810679b9d6fb978e5ff18d020a46d475c4c26fa29
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:c92ebe828b6ced6c93a3df5f4ccdfba022471729ad08b184de2f7410ecf592cc
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:df11cdbab2a1c6b9b800ef124c73f2fb5e3981402b21bc822ccc13e95836197a
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:db27cb4f0c350d99546b0f1bdf02d1ba271a8053a4552d88ff302adca60a3e29
        - registry.redhat.io/rhel8/skopeo@sha256:7fd65200b4847fc84fba0466519b87f915e28c0e9bc90d3022beecfebffd4af8
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:99f5383d8c6bc046a0bf48fd6c2064a5d055645085248327305f3ee4ae71979c
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:e673c20fff2fc684f5098282dd6ac6cb265045945249e995b81289532ce8c856
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:9668d852123b160e49511242371edc6e941a8e8b8c2ffa60080d91cc6b97d3d7
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:84eed8ba1189448e02ffd284c7d913e7b3cb7efe7a8e5b0edf890f7eb17fc439
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:922b9e86a674a6f3eb82f510569ab9cde97a1ad833c263335cd4164bf4164b66
        - registry.redhat.io/ubi8/ubi-minimal@sha256:c12e67af6a7e15113d76bc72f10bef2045c026c71ec8b7124c8a075458188a83
        version: 1.17.2
      entries:
      - name: openshift-pipelines-operator-rh.v1.17.2
        version: 1.17.2
      - name: openshift-pipelines-operator-rh.v1.17.1
        version: 1.17.1
      - name: openshift-pipelines-operator-rh.v1.17.0
        version: 1.17.0
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      - name: openshift-pipelines-operator-rh.v1.14.1
        version: 1.14.1
      - name: openshift-pipelines-operator-rh.v1.13.0
        version: 1.13.0
      - name: openshift-pipelines-operator-rh.v1.12.2
        version: 1.12.2
      - name: openshift-pipelines-operator-rh.v1.12.1
        version: 1.12.1
      - name: openshift-pipelines-operator-rh.v1.12.0
        version: 1.12.0
      - name: openshift-pipelines-operator-rh.v1.11.1
        version: 1.11.1
      name: pipelines-1.17
    - currentCSV: openshift-pipelines-operator-rh.v1.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.17.0 <1.18.1'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.68.1
          - Tekton Triggers: v0.31.0
          - Pipelines as Code: v0.33.2
          - Tekton Chains: v0.24.0
          - Tekton Hub (tech-preview): v1.20.1
          - Tekton Results: v0.14.0
          - Manual Approval Gate (tech-preview): v0.5.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.18/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi9/openjdk-17@sha256:5d619e202695acf76a5f32c07ace6709f5ee5d3999bf977e1bf95ad34197f44a
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:86ea94b12e8923035e3e4f39b6a53b3ab32c5f77d7eea170aa9b34aecc6f8577
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:0a60ce46c05256c74f770901e0bb79ad89e88ba882b643c1ff7b2a1738463acc
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:3964c3248dfb1006ea9a46295baf7f8cfa71952d6b9863b52f87e8f3d7863b1e
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:36a4c6a3273f2371ab5e679f1955221c286ea753e29e6ab8da7c983681a68f02
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:b550a8bc2a87d07a5b22a3ec3fb83efd8eb26ca6e80d404f2e453cc9b8434da7
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:cd28c2c90dd340324b0d8709640cd7636f28286114b0023649b6e21be13e03c2
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:fa5c456fc91319c55fac9a011dc7bb32ef85db293c73d74436e903b76ceb4349
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:409c14649d8e7007d4b72b6feffb71dabd26322bf2a7c77f5307f2546fdfe81b
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:96f5e1922fb9d6dc17e271576dda9bd032bd6dcc04e7496fb15a2f6b5871cd67
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:16d799503674641d4690822edc9c399c61429ccd31621d295eee4a4d1ef1c076
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:cbd89c531953a43b8055bda72f7696be80f57fd1d782567a91d82e0fbcb3d26d
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:8fec8e7a8b3e5182c67104de4e3abe32593b682c4a8986d1e9ab835dafb95ff4
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:3696cf724ce7865626ae408d82fd87727c741e4d3eb3475c271ec6d56f0882a4
        - registry.redhat.io/rhel9/skopeo@sha256:d077e35100d6d7b3795f8e3594293dbf8b65c6bc18b909aa555fe56f048b356a
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:1b57b64657eca0002145100bae262a19f02eb40068e07994df467e28a1618f90
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:20e19c02efdecf94a7b16483047e5db81c50dc2a0aa59d4895067e6c7769efcc
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e2b4c3b505c4b5aa36f8a5e661389b0de4e799c74c2c7f16a065e5522b438e89
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:80d441ec5d07c5059277a3a9173a35dfc0b8372d6e4d7823514131a94bae3fd7
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:d72842a7c6a1e50a68576c93f60047afec801cf83db64d7d3e34b25bf3c753ec
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:e6a7738cdb0e8aa9e1af381273b1c18ec1414b08a66fc851f6e11e5aa7a6706f
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:fcd5fd10de7334bd59a8da4997f91008ced95c6cde40d56d50f3b3e32c0f19e1
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:c20ab6fa719506bca942669ea84b113335b05df3ed275257b88be234cade7c8d
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:2ba6782af0d7e5c4d623de31814f43f4f7c3c38d09b57705021957d334067f87
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:d055d2b35a663aef3e1aafdbed0b12957867c0670c946ebae66e9c44a768bda2
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:72901cf4e091fe5f6a577144fdffaea8570712e9887a8bd715d5ae5bc346ff1d
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:b9675b22bc8759903472c1eb0e03f90c75460f02799d894294c161641cad9414
        - registry.redhat.io/rhel9/postgresql-13@sha256:ecd5a5583db922c517c735797db3fca587c0ef0b410e7dd494ace2267d89cf05
        - registry.redhat.io/rhel9/buildah@sha256:6886997da8c92214e99e3e8da47a0f08239172d2544b4edae058d5cb018effab
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:c90a9c3a20f088241980b76986ba4f43b0243fd6d94a826fa7b5888e01c13093
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:776d34b0ea2e35c11ad4050e6b40cf4b40c84dbe0d5f3e6ad669c494a210ea8e
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:8b4bf043e0fb237553503d63b129c432600bb4e65b2096b246b257b404b42be9
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:eaad37453a5ebef51c28d4a068f56e9f394e710c72bb86342cd2189001420836
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:8ae0ca6411e4fcd5c0605344bc6930cf122ffd80e71199509d58423a8b41de1d
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:dbc462cb50385dec3a220b5d0caed0e0211c166182026f86e3c51a1351ba0de4
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:6abc903c784bb455e1f2fde45a48afe775239668e1d0dbc1b3cc6afca3c0b477
        - registry.redhat.io/ubi9/ubi-minimal@sha256:ac61c96b93894b9169221e87718733354dd3765dd4a62b275893c7ff0d876869
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:e27a30d471dda4983a5b8192709d283a9d082558f3caf31b30db2214e418b017
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:e63d956a0534f2e726301e00e877c4fe2f34ca4559e59d2df7567c04ded311a3
        version: 1.18.1
      entries:
      - name: openshift-pipelines-operator-rh.v1.18.1
        version: 1.18.1
      - name: openshift-pipelines-operator-rh.v1.18.0
        version: 1.18.0
      name: pipelines-1.18
    - currentCSV: openshift-pipelines-operator-rh.v1.19.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.18.0 <1.19.3'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.0.0
          - Tekton Triggers: v0.32.0
          - Pipelines as Code: v0.35.3
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.21.0
          - Tekton Results (tech-preview): v0.15.3
          - Manual Approval Gate (tech-preview): v0.6.0
          - Tekton Pruner (tech-preview): v0.1.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.19/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:22cfadc42e793553182acf05a34fb6a7ae0d7055431c2e393699ecfeb171b173
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:7a0c6794e9284209dc131f3087ffc01fd5946e20a9261422808c6dd3938407de
        - registry.redhat.io/rhel9/skopeo@sha256:993f799d58c48151ecae7a84e63860f6c0039e8e18efed5ad67e7752f4b7e408
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:fd6f4cd7801d7c20cf5a6e546022136404212da5d9997750e643f6e640b00e3d
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:2c3074ecc0c74d508020bf1e11667d868150915b55d436a43c2b4fc22815bf48
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:80af5fa2d87e7c405f16575519293ca7fccc49d635b4f3d0d334dea4fbd28721
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:25bcdb30f0fe34300c4f61f62513f4853f65fe11bafde58bcf19b7b3c3fb9af7
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:251c87e2c48e7f377bcb399ed64d50dff11de3459629a73a63bac17c353eb16b
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:363e28a82ed022e507c3976d3a42ee63c053eb7871712c213b12235d8234b7ef
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:b247e9f3a2db32b575999387ef4f91d685a67c1f3950a6e265e1fe14c1fac196
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:6aaabb9af37e46a856371e29f141de6b0fac623128099bb927fc30f8f61e9db4
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:d000daec36a6cdbe41a85ee4eac24a712f7628441a412f751baf8d7778626f62
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:8bcaeaa751c61eb927d6353984f81d6a92b4bd4f9a5be9dc620a11b06ff9e6ba
        - registry.redhat.io/rhel9/buildah@sha256:a6e04c061b8d261dba85d01dcb64d3f7dba8c0b9765e03f05ec79850a248d75a
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:41d669fbe8bb48481890fb9fb52de9d4cac98cb74e6a6b284d046a8d21b06e58
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:28258903330a0715141e4685b3760ff18b0d75da0694985bba7c3345d7808b5f
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:90a5eb2c0decb72234cbea5354bb7737cebc53c467229b2c16a1425b4bf408dc
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:3f7e1d83eb1e21860cc2ae0bfe3abe8c9058e92a69698dba9d5871069016e90d
        - registry.redhat.io/ubi9/openjdk-17@sha256:cdbd64ba1b7826d09bc180fa19aa6fbc09ddb5bff5e02b7e1351353bbabebb78
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:a90a6c269b3fc4c77652d4862878211eee5eca1d8047e561d07e3ef1292e52f5
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:31a40f8095a3f8cd04750810d964572339965ae0b2d357c57c0b671a95e1cdbe
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:1f870e70669ad584db1ee31b1dcd822d8b8d0cc03028c7cb09238fb7abf19a42
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:b8cdf2a1873a3df1fb8b8f6d58d06235cef28d228abd70d0bb83cdcfbe3bc230
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:3b1c6f8acdf17a5185413147e0248baa559378ed79e08cb098c724036020576e
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:c9fbb018a70df9a542236ace556fa3343ae350f43c0629f13e525854508f73f9
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:92b8c92eb0bec102215b97f1d60d76c7a57747aeec9594ce1e0e46f7c27391d6
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:5f39f9e61ddbce92629e6502bbef737365c8a1bb2dc2ecaf9f9d8092dea844c9
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:aa1b6f374643ed4e89ba194afde8a5138e401c8cdeb45a6bb5df5b608f58358b
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:22095c436562ff92771d9349f57d007d82415dae519e3da1de7d51e3f1042a93
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:198cf9e5c9641757a0453dec2747a24b018a9b85e10eb0dfdc7a2e75ed09cf6d
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:ba596c629fe8b5b70db97923ff11392dc12115c8e7afefee0d48c12440b8dd86
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:10c620aa3a9c4bfbca185fb460e2521473a5fe6d518342b89a3dada5cf14e65a
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:13608cfe2b87535776c65728aa40cfaa84ccf63afe64d1f2e799ea0e26819239
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:68d9007c5abb2ad3b7666249a61d9428c2cf07165bd8090a6e6def0b7cf70fa3
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:37e2f1fb8a2b4c7ac3d5390ba68a9bf98a0c83a38567b6a9778764061b3d2bab
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:c3ba6b9652205407cc4d4d47b4a928fe8fa86ce4427e6b44ee1d53faf3455f21
        - registry.redhat.io/ubi9/ubi-minimal@sha256:14f14e03d68f7fd5f2b18a13478b6b127c341b346c86b6e0b886ed2b7573b8e0
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:d21cc7e094aa46ba7f6ea717a3d7927da489024a46a6c1224c0b3c5834dcb7a6
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:2d2499b7b12aace07c4b5ab4121caad445aab5bc8dd73377171683b36d417d73
        version: 1.19.3
      entries:
      - name: openshift-pipelines-operator-rh.v1.19.3
        version: 1.19.3
      - name: openshift-pipelines-operator-rh.v1.19.2
        version: 1.19.2
      - name: openshift-pipelines-operator-rh.v1.19.1
        version: 1.19.1
      - name: openshift-pipelines-operator-rh.v1.19.0
        version: 1.19.0
      name: pipelines-1.19
    - currentCSV: openshift-pipelines-operator-rh.v1.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    },
                    {
                      "name": "legacyPipelineRbac",
                      "value": "true"
                    },
                    {
                      "name": "createCABundleConfigMaps",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.19.0 <1.20.0'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.3.2
          - Tekton Triggers: v0.33.0
          - Pipelines as Code: v0.37.0
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.22.3
          - Tekton Results (tech-preview): v0.16.0
          - Manual Approval Gate (tech-preview): v0.7.0
          - Tekton Pruner (tech-preview): v0.2.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.20/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:bc403f93ef6c5526fa30152eab384c4a25bb0bc6ab4158a1e24f637b87a57259
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:980ed7e81ea55522e1681dc1e9b170b68d16f07061dd1f1ffd142367d59552f1
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:bc10e7d6122ede5cf4f2882c200447683255465a1080e826761e90bfb983506e
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:c3013c347bca8a520443a44ac2817fbba5744a9e91069ac5f09e91447ebf6b28
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:1b6aecfce45ed3163227b845b212e83385cdbe999edfc2e9387a0c76f3f8d8e9
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:5683321a5376e63c3ff3932270a1850d5e261f76212020ce867334166f90d2f7
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/ubi9/ubi-minimal@sha256:2f06ae0e6d3d9c4f610d32c480338eef474867f435d8d28625f2985e8acde6e8
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:7eafffa6e4e04f10c1d4e84e02d51c00eb8649c60f8f9a13da14492f24634891
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:d56d6b46331b4569077ab1e7a7b609ba01f0a1dd62b695d62f8aa0f32ea1b621
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:e63298d07ccec5804dfab4e42bcc8163c29cb3ac371c91248777423a4edd31ce
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:8f0494341244a75546a4b80b065702be04e6d200421aa4e7ec90385c9dabc6ab
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:c31af237afb84214bf2a4c06a7636ca071fe97203b41da90d7878109ccd64784
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:115eea1a5ca87f3b6b72db1df05b7386bf73dc70c9c2e4dc5e1a66700d960c37
        - registry.redhat.io/rhel9/skopeo@sha256:a65a413f8a2864389a09dc750690d97afbcdc5c70821e0f85e99e8adba7954e0
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:dd941acd357ae2c0fd7e15fb8d32d7ee6b2847435101dd1e1f107db48af18bf3
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:5021158a4752a26f8b33adea7da66695b22d2ad155d346d35dde5f971bc1a55b
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:823e018e0329d764ac694f177ce88af9074f364c29a1c17929f9dd0b596d03dd
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:14dab8d87323fc7d112213f7f4db3ee7cf3841747de669ce710ae6a794053457
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:bbd1918660be7b3c78777dcccc87d3b26bf5adfc077844ceffbbdd53696dc0a3
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:a7163a9e1fe891de021b1003558e6e64fe17ed17ab0820ba16bba09aa112401e
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e7998e113f75e8f29602bae210c8b66ed06634b699f1d99c5875ef879220d733
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:ec74a562c18fefe36ec7ae3ba083e4eb76c56b562f916578900f8767ae98e6c2
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:c4d05e546c466c57379fff502816012f615be073eb8cdce54677b4ab206a3f5c
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:c0f438e6095aa81f3f7d820614e0414cc43374014c1505d7641bc1dba69a3903
        - registry.redhat.io/rhel9/buildah@sha256:18453ab0a62154283aaf4c45efd7543d1c8ecca1aebae46dedc7ddc0f410b232
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:818bfa4a3a874974b5f7f8bcf2faeb6c05e1795381ee1ceacf89f17b1ec0ef7f
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:9415e889ef98df6d2b2dca1e5d798fb765eeea0d9828e48424374d5267a6b0d1
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:a712bb411d2433b01793dff25785c1e265fa865a52f7c8732c9fb4418ab61757
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:3cf990fb2fc422f3527a7c8b43d06ccd4d32bb5487ec8cb2d4a97f0f5d13f170
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:458f865f1b246c5f0682f5956795771a0515b89edfcffbbc0da1a3d3f532645e
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:96db721939b22572d7d8d8b95dcb5c92fdedb98e9cd0aabb7d17b703a915b07f
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:036250bcdc6fb95273bf07948dfdd4dce765ab84c803ceca0d0b4abf6ca19dec
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:161df929e2bd4c5acff594be31d5458365de5ae0d2189bb2f647c3fc6657fa28
        - registry.redhat.io/ubi9/openjdk-17@sha256:555c8bcd2596637befd1a6da4aa949c5a2640cf66e09c2dfb932afe37e39095e
        - registry.redhat.io/openshift-pipelines/pipelines-opc-rhel9@sha256:c1bee4a2bb495de6d4da962b6e7b2447caa51cc36c3005e4083bd31420ef5102
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:901b5108ceb669c985e75f5fb033c496e9f028863df99ab32651d1b08433a352
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:df15eb1f1f5f237bfe481c16ab3d40bf7fb4123f3c930bcfb121e2cc8dc42882
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:441e1c7124c1674e3d4c38c3e82a4fe5568e76d13c89c04bf009d6b97435de09
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:e3e64abef53151f6b15f15bcbff85668167a571e98badbd8a2ca252ff3a5b8dc
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:2d496752b19f765f1ea36ec94b873ebd8fa604764a56aa67d3ed94b33e5fe75c
        version: 1.20.0
      entries:
      - name: openshift-pipelines-operator-rh.v1.20.0
        version: 1.20.0
      name: pipelines-1.20
    defaultChannel: latest
    packageName: openshift-pipelines-operator-rh
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Coroot Inc.
      provider-url: https://coroot.com
    name: coroot-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: coroot-operator.1.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "coroot.com/v1",
                "kind": "Coroot",
                "metadata": {
                  "name": "coroot",
                  "namespace": "coroot"
                },
                "spec": {
                  "enterpriseEdition": {
                    "licenseKey": "<COROOT_LICENSE_KEY> from https://coroot.com/account"
                  },
                  "ingress": {
                    "host": "coroot.example.com"
                  },
                  "storage": {
                    "size": "50Gi"
                  },
                  "clickhouse": {
                    "replicas": 2,
                    "shards": 2,
                    "storage": {
                      "size": "100Gi"
                    }
                  },
                  "prometheus": {
                    "storage": {
                      "size": "100Gi"
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring, Logging & Tracing
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/coroot/coroot-operator
          support: support@coroot.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Coroot is the Schema for the coroots API
            displayName: Coroot
            kind: Coroot
            name: coroots.coroot.com
            version: v1
        description: An open-source observability platform using eBPF to gather telemetry
          data, providing actionable insights for quick issue detection and resolution
          in Kubernetes and VM environments.
        displayName: Coroot
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - coroot
        - observability
        links:
        - name: Documentation
          url: https://docs.coroot.com
        maintainers:
        - email: support@coroot.com
          name: Coroot Inc.
        maturity: stable
        provider:
          name: Coroot Inc.
          url: https://coroot.com
        relatedImages:
        - ghcr.io/coroot/kube-state-metrics@sha256:080a85e77b1e530df56f84c8e0ee2523c1f740adea152015367b87edbfac078d
        - ghcr.io/coroot/coroot-ee@sha256:09b9177ac07c2d26f4bd16050d07bb1e2132431055c32e047563083ee09194a8
        - ghcr.io/coroot/coroot-node-agent@sha256:3d552e98ca2ff8c6e125496857306172ec6f8e134ff13886c7f799abef29d1b8
        - ghcr.io/coroot/coroot-cluster-agent@sha256:674bcb147354873eaa0ee9c1328ddf758d43adbbe532c9fbbe41ddae7addc114
        - ghcr.io/coroot/clickhouse@sha256:afed1054626598371975c948e5b9e4dcaac54e8059b7ad9f0b905ef1c99f2260
        - ghcr.io/coroot/coroot-operator@sha256:13f1a4a07877c8fe8230a0bb7282ff47451007c9cf689a995a37d16697b9268a
        - ghcr.io/coroot/prometheus@sha256:e7117bd86e1088fc02ac12758bc94dba1029de54a5bc1ee04954569d8b79625d
        version: 1.1.5
      entries:
      - name: coroot-operator.1.1.5
        version: 1.1.5
      name: stable
    defaultChannel: stable
    packageName: coroot-operator
    provider:
      name: Coroot Inc.
      url: https://coroot.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: cpd-platform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cpd-platform-operator.v6.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cpd.ibm.com/v1",
                "kind": "Ibmcpd",
                "metadata": {
                  "name": "ibmcpd-sample"
                },
                "spec": {
                  "license": {
                    "accept": false,
                    "license": "EE"
                  },
                  "storageClass": "managed-nfs-storage"
                }
              }
            ]
          capabilities: Seamless Upgrades
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.0 <6.1.3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ibmcpd
            name: ibmcpds.cpd.ibm.com
            version: v1
          required:
          - kind: CommonService
            name: commonservices.operator.ibm.com
            version: v3
        description: "IBM Cloud Pak for Data is a cloud-native solution that enables
          you to put your data to work quickly and efficiently. \nCloud Pak for Data
          includes a suite of services that support you on your journey to AI.\n#
          Introduction\nWith Cloud Pak for Data, your users can collaborate from a
          single, unified interface that supports a large number of \nservices that
          are designed to work together. You can use the Cloud Pak for Data Operator
          to quickly provision services for \nimmediate use. These are just a few
          of the services that are available in Cloud Pak for Data:     \n- Watson
          Studio\n- Watson Machine Learning\n- Watson Knowledge Catalog\n- Watson
          OpenScale\n- Watson Machine Learning Accelerator\n- Data Virtualization\n-
          Db2 Data Management Console\n- Db2 Warehouse\n- Cognos® Dashboards\n- Analytics
          Engine Powered by Apache Spark\n- IBM Match 360 with Watson\n- Data Refinery\n-
          Db2® Big SQL\n- Decision Optimization\n- SPSS® Modeler\n- Db2 Data Gate\n\n#
          Details\n\n## Prerequisites\n- Red Hat OpenShift Container Platform 4.6
          or newer installed on x86_64\n- A user with cluster administrator role\n-
          No storage is required to install the operator, but you need storage and
          resources to install Cloud Pak for Data services. For more information,
          see [Storage considerations](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=planning-storage-considerations).\n-
          See [Pre-installation](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-pre-installation-tasks)\n-
          To install the Cloud Pak for Data control plane, review the [system\nrequirements
          for IBM Cloud Pak for Data](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=planning-system-requirements).\n-
          Before you install Cloud Pak for Data services, review the [system requirements\nfor
          services](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=requirements-system-services).\n-
          For air-gapped installations, see [Installing the Cloud Pak for Data Operator
          on air-gapped clusters](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=tasks-mirroring-images-your-container-registry).\n##
          Resources Required\n- The operator requires 100m CPU and 256Mi memory.\n-
          For the minimum requirements for services, see the [system requirements
          for services](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=requirements-system-services).\n##
          Installing\n- [Installing Cloud Pak for Data Control Plane] (https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-cloud-pak-data)\n-
          [Installing Cloud Pak for Data Service Subscriptions](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=tasks-creating-operator-subscriptions)\n-
          [Installing Cloud Pak for Data Services](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=integrations-services)\n##
          Configuration\n See [Cloud Pak for Data Post Installation](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-post-installation-tasks)\n##
          Limitations\n- `x86-64`, `PPCLE`, and `x390` platforms are supported by
          this operator.\n- See [limitations](https://www.ibm.com/docs/en/search/limitations?scope=SSQNUZ_4.0)
          in Cloud Pak for Data services\n## SecurityContextConstraints Requirements\nThe
          `restricted` security context constraint (SCC) is used for the operator.\n##
          Documentation\nSee the [Cloud Pak for Data documentation](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest)
          and [license information](https://ibm.biz/BdfVHT).\n"
        displayName: Cloud Pak for Data Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cloud Pak
        - Data
        - Analytics
        - IBM
        - AI
        - Machine Learning
        - Cloud Pak for Data
        - Auto AI
        - Db2
        - Watson
        - Data Governance
        links:
        - name: Cloud Pak for Data Operator
          url: https://www.ibm.com/products/cloud-pak-for-data/deployment-model-options
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:bbb457d4f6b7bab065e909756c2dae31af7162dce10bfb92f274aedbc98d0762
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:716b5823661ac8f1eacda11ffbe5f0ff08a0e9d2296d04acffb646b7526128a9
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:1f82599d91ab20d2024abe134d2f56caca1f9bca1f02d64352c81b5f1e0f1730
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:d6575de3156c7a9b0553887abaf8713a758d9c9916f9cdcef611923c50632867
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:113fdfb9395c96f7eb65f9a20e18ede8a2eae4aa6c91660be8fd3e517656e156
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:d0e0144eaf27e03aec1499e53cae5366e11a3da284536cd3b27ba5691b4f70d0
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:6a996123e69c7f2c3254f334e278ee0daf75fc53980183e1a72358323e41349c
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:1d6dc7a3039e6fa044053d4e89ef54ba2e83ceadd57bcc4613a6d447d027ae2f
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:7861185d55a5fd18d896dbc57501ae5cbc1de0bc02f1ae3fe558a36f0f80b4fc
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:46bbfdc1f5e418db95ba806228fa5696be22c4cf9eefc185ce27bfa68ec1d247
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:a95e42a4f9138f1f97937a74982c98d466c171d0e3e79c522ededba6c25b39a3
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:23ba1b277cf76bc28f952ce6f5bb02d6177f76eca8a8b4ff7925ef119a994a85
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:670a36f3400164cc9eec5d23422d184340e0c1edfb13a634e58f0345c325c21b
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:0dc719e58694fb501b8b36f3a375b3c0462624d27fe49afe6b835821e31393f3
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:257bf301338f7b123099e2ffe31c3f3b793bf28d6a0570a25b889271239579ba
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:d61b7b8e6a9785b59630d9c5776f72a29ffbf2e81eb753d7b11dcf2d9d77f1f0
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:30a5b2c4e3c231373fa18d8aaab16d3505cc11a8a5d4be36eaf99d7c0dee35f2
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:fbc6eab07fcfc2d0c1895f18e17f751e9481ad111875959e9ff08e14f81b9a35
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:aadebf0598c492639e7c1109ce681256731215b686fec808578bb09558a40315
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:6cd1825ba40071bca7f46c1af3e813ae7ea44a393dd3d10e3c83e62834563ffd
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:b3387bae415fe3a66db2264ee766cf43aa8f31e2b615b200baa9228216596a0a
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:508412a22fbd65074964150d6b270f38d6df898e17eb7ec5e28f279c4c2e1062
        - registry.marketplace.redhat.com/cpopen/cpd/cpd-toolbox@sha256:dcd74c128502dc301482eb302b34c21dc02173ae1cf522356d5b500ba42b4e25
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:f00158405adec00c7e992efa7d7a19a520df485cbbcf39972e579ed3c6e4d289
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-hook-utils@sha256:ec44be69b4c585e9a3d507e29bd287f5b464360d82c8e0969c5fb4ed3e95abc0
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:31470f873a0931d130484d5255692d445e3127398aae0e777a3c3cb702917743
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:f561f089c4e5037f7c04971d96fa024ba4df5e6d17834f0c44545d0e53c772d7
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:20863de342e07dddc9aa8eeb214fba5cb17478bb88613e4e1f167d77cd0af214
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:9460c98c7202ab368e732951c13bb6f5064cf18f69d36b3d26eedcfed233ec81
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:5aecbb4de6fca34f026de72b7480f35dbb2ee8d8ef1f22e524a85a71bea99a1d
        - registry.marketplace.redhat.com/cp/cpd/jobs-ui@sha256:49daf607918b378593485da16ac417056ee295bca94f1a5cac37b334381ae096
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e924abf025141f3b3437ed89830957c5915152d2610213090e33fde988bd879e
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:e6fc9d7a70877100b42e290cd091011f2e1c674a27985fce180ee245373391dc
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:8bd9fe1d35785fa0125b8fb3a60333219a50ee8be7a7ec07386e8e9d65316061
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-bundle@sha256:94c7a1dd290e2abd55c5b8998cf08ad174d3fbfc659ae6fa6fb6cbab61d5fe14
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:ad027c91c719fd1d931cbac8c76735474f676b228f9ab23d1f243201246b3238
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:2f40d8188dc8b41fb072256ce8d109167095469af416b322ac023071814cda8f
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:e4eb6a75e0d207232e479693d9c99ba863f8e62830f3191ca4d1e7ceaf3586d1
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:847751338f3b78cc06e29a4a26a18dc0f5d0bc24cbda23b765198a6a5c4503c1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:ca4da5a8a3c1e7add6ac630c3fe4713a08739f6b891b2290bb17026b94c31317
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:c268d2852580286e5c6770857fcfc3b06a698521f5acd17ed83b8e2cb9ab55cf
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:fb37fe521693bdb8b96400fd491a39c573545f35b4131fdfab501b7ba29713f6
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:4a5ed208b68613d5d88936687f6912b3aad0e93c75f0d311120a8e55200f5795
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:52ce08ac2bce7c02dc30e2c0aa74df4bde8f156debccc59960f02b4b0040ffea
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:d6cdb71edd7b7451a206d1620b35c321f88b3ba8028468329595cd538a45abf6
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:0f01a67524fa46829e455f654d3b9b00ceac44f3d870b21c022051dfcc8a00eb
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:b646ea3330f3fd9bb3174224de69f20ef58589a464d50d1c9399211141f67bd5
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:650e68c41252300836740ad495ad6776a7da94331e8070520e60a1c5e2c5f3ca
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:6f2e7cc917adb54b3112dc4909f645ac08c33e8c2c33266fbc8f5133cbcd1973
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:5e161ff695c0552d1333a1f993048bd055168a9be9e0a75adb7c4464c9c2b603
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:ff5a5be492347e75b309194661d7742d1c909c3765c9947c8b41f4c3af62b08c
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:7ac32270044a9ec08ddaa8dee62b61618d9363c8773fe44ba1dd86e1b53acbd5
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:c70d3c86d4c61bd87a427f7a7571f37fc4d0966ff4f420b73ac4a2af73ed1379
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:978cf4bf299cd3e8cde9d7ee7aa75b73fd9d9e6755567651d6c1c49036b8978b
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-usage@sha256:8438c2229b500fc9c40629280701856345e228fc645a5b468bee2135c61b2387
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:498ad7832e989e9356aeffbee96f310e3edaa6a7466320fda9b4f75034e73eec
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:7661b04bc59d64a4fafb2f660cffb99e3d326772749759ed0d2a1940ba147078
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:1f68dadaf2ab779dece300037f3abda1b0bb33603436b84336f8fb7d53d33644
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:0f46b288d060a15a48ecad8f328a8046bf232b8629c3d79da6a82ea712b363f4
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:cab91a231068a6bbef03ee8db32fadf11654317c4573c6714721a15d0e2a98dd
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:6c226a0ccde697b56c81427a970062872929eab73d99791c9dc670f4aa176ab8
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:67bfdf9115fbee782e35ac61dd93801bcb9e622d441806f1144208b4447f1430
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:2668a562eb8df1117c40acbc3600d6735469d210e5aae4aa03bf9f54d0ebfe46
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:e4888fadfadb3462e3a9ad0395a526a2631872f70caf291dc18b60269ded199f
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:f0c68715f78bd20c352c97dd48ac5a912cad58e19a00159b5e6878f6beeb00fb
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:b0735e396c09fbd08f328ce0686cd55fe9eafd5b759f36cf5677326b24cbb859
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:abd96661823553d623eb0ea1fac8271d64391243dd0f13eb4a4fcc858dfe10a5
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:48d237cd3d363da8a206ca5e092d0f91e8a5bc931ea81763498dcf6f28c0434b
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:c078bd65fb18b31ecfe03c832643bd6fd85f7330b94a1ca26f557d32d89903c1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:631a18f4b437e5092b9671fb4c86af05e99effe5540c4d6f964dc626466764a3
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:b63e2c10478735b51b6828bb11709569a2fce62a19b70157a8a5b1964820164d
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:47a3745c78280f98f2ee352bb9192cdcd2fcd6d4c1a3627e7aa42178a0f107fd
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:a28b71ef64c1967cdd36a16e01b611fff6a387b6a741dbcaf7556ed6be9b1f69
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:095aae63ced410e033d8f7f3d27bd14dc5b93a10068ef5a389f181115c5d07c2
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:00f8bcee4ba7e0c5cf5f0c52e9e7707df367041c9f4dd5508dbefe0dbdba8877
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:ec137115210dae9b69cb2899ed8a5437d94c8a658ae1ae5eb295232b34b0925e
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-velero-plugin@sha256:fe562192e86558a17d1c88a8fe21148bc000fa93c8928f7a3a897acce62e6798
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:e857323acd10f7b99c5b7fb4b0a5b89a808f68036a1d9a224eff55a7ae7744b6
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:1cf983f9b9cda18257066fc82e199bd3e22e2fcba13daf0f60c92b50a4a0e078
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:96d382b1b26a1e1e4074b7a1aa48b973cd4865bb5c3cf02795306e92c752a8b4
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:dadfa9ef372b8d165276fc5f1ee9c63c3f01367d156700810d63e1ba19f3da79
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:45194b56f9fcb3af83158eb0297248364db40e8164fb4fe4f5656c58aeac3fa5
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:0ceaaab514f65f9a1c40717e97ca652506f8b3595186222d470a99d653afcead
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:04f9ed67696fc193877f9da3d73ee3f6f4cadc7b8d09fd07fa12d0ee15f1152d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:9f7e6a1ca3faa5af1db2afe8693f6bf9182296be2a23e305ba6c990aee3fc301
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:562281dbb2bec5191a7c6b595ab853092e6fe3ff1432e77b2471f71ccdffcbc8
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-bundle@sha256:552fc53664722903eccaa7d3782a3e84413509fd357725fe0b0f4a077c48ab2c
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:cb9a33db3707c0c6045e6c823ecb21b0f229c17ff868da7d36602d0026eff351
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:d4738f64e36a8f0c88d40043eebaae1d561193895d438fb6d8bf1a042d0a1007
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:cb3620eb9aec24f7ee54c3ec6709fc223b685177500c503af8ffeb23642dcdb5
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:c330a321438943fdd264b292b68fdfb8be89c3bde93a7bd987bd92c9dd2512b0
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:9a970756cd76ec748b555d5bed97889b8c9f410db6765eedee84fad4976cd523
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:4b3ffae1ddb7a2c28f98199be6343ed826f8fd21b64fc9c584566341f44e44d9
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:48b7cfa31b9a185176c20afc119e37346a6f9380d74000edf6c729b4677ca44e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:a7c050ae29cb93754f59d8f83676275ee9a91ae9a9956395b97abf8c5e20e96e
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:65108559a5edc9a7c4f82962ab2293017e0873872730d60d8487963970761941
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:d34d561bbf8afbfba9362d2a9a58f2919faed1502abf72e11882dd9cd4fc9711
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:4f0cd4920e7cef64e9f49ae1abc5d87303102fc96819e0c3b699d86b8f8a3980
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-operator-bundle@sha256:086eaeaaaf440a1d4a254984afacd8d04583aafb9ec11ac072feb42a87b986f7
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:a2949c3868c2c91bf35e50c64b51fcc24372986744b55520e21e9ce31fde59f0
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:e21a346d15b286956f62bdd1fbb6d7b77a554ee10a475cb698a8796c42f3a267
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:6298c1561c066eae3d420f20212a3c7a8f9069a6996408685437588bf8335dec
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:47cd9776a46ae5fdac3a101d79a779baefc802ddf8e1f8fb02216f47feb55fa9
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:8444703b2a7822f82141afcf4f36bf255650d755321b9e9fe2cbd1bdea41d5c7
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:ea57c7cb1ccd5bce17f6cb35e326ab974718be8cd19b0cd38ea8112e91c51939
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:5d889be3dff96d1e888eff94bba94228aa46f4e94e5f30ba33077ef00692509b
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:ecbc65705de76804a3c7642339ecfb4cc0f5a61891058e07c20bf2b63f1b82b8
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:690e9b5218a200de93a7c8477fd2980d8d341f22be75002967201db09b5a5524
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:9c2093089ca0651bcabc1d01ee2b69d2fb0cf2a277a8b966cec9533950fdcdae
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:a1754e003811afb578dace78f9eaef91e7917e5468e590fcb41920fa71bcc22d
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:010bd5bb9943ffdffa8fd00cc8226e59d9583bb3baaeacfe70e5b00e805878f1
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:3d5c6573faf4ae00cbb21cbfe9b2ad5cd8ffb658df0c5ba09bd98b3bf54f9979
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:8cba5f256747002b62fcedb9066f7ea6dfdb1071eb359e9137fd413d20b29ff8
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:6f96b8b6afb00cb26fcf903b8e4c30b0d6344bb86d8bce2f2026fc8cb0b575ec
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:56315ce0a0a70f5899191a7eab856105489143f0e889c8cdd1b6b8c225f02eb7
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:d733c5a10a21cac2df2fff6378035151e5d0df57e3ed1f96bb3ac17c0c95b8eb
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:d06ef794046dac2c58163c201f55489b69081dd15dd91cab16387510a50e24b2
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:2c39ff1850b7a9d32a11e99393039ef74ab363f3fffec1e4fd542f2ae148a6b4
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:61ba2b05a6826f618853ba048496e2c9558e509a01648c74b50cce6d28d0d764
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:e0feb4d305c5ec9501b30ccdf8f5826cd8eb1ba5ba1d62d4d685101c785a708f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:c8786cffb4a27decb5dff3d2a9682d07d11b7096a689bda0d7cf105aa3156299
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:51e1275f6e8297418997b0ca9615642153371c29b3466f155e15d7fb9e80cda9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:f1aa6add76f59afeccdff84c9ea89a46b49c6696f62acd75ef8ff2f31a3080be
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:132670412a0ce6dd010869e569504d2e89aee0983ff221da371c9e30d5f8d8a9
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:b5a8fb14c3bb7c89ff970ca85f06dfcdca832e2b45ad86e8bb1b6c0a0e7e0764
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:ecac5e0327d939ca84d304b2fe95739ae43a06745ef775ee59e4e5378ced65c2
        - registry.marketplace.redhat.com/cp/cpd/environments-init-job@sha256:6af8f900af7c11eee6894e2c3e52f740b3c700dac0ee753fadef330230bd42dd
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:369f66883e040accb73c3deae984ed479e90eed242051b7035723f0ba21b51ac
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:7ae4dac24ea054ce527a507dea6c77a3233aca0047672ba5a367fd5742448517
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:8d7625a0ba656425f601b6ec9a398a6d30511fe1de119cf0e29f680c704ea225
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:7b2e9c373f9bfd91e16eff68ef2313171c69b94fe5a21d3d3158470c74f54a05
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:ad2215adfb8dd2dc600bf5097b7f88c257d5fbaf541409c01b36cdb67c130236
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:c41e4d3314fac0d2cf5f08e7db85801cb2098d37d0aa8104a347e9e82526e914
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:2e70036d7299168331c94af70a22e3b2063c56160f616362b1a2187c12d003a4
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter@sha256:be5fd8a8b43ae3792c58de1342cca91836287fb584bab003f7f9aeeeaca3f65f
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:54c86367bfe0ab658611856cc0a6c5de24c14a336bd129ac9f16293945a55238
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:603a6bf0e616f16b609bf157b9e21415802478aefc9c606e124008d0224a1188
        - registry.marketplace.redhat.com/cp/cpd/task-credentials@sha256:e162828d5ad07ceae4db0589f930979842714e3b766552102bee373a3a570af3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:4aa81bd0043d0bac60c9f1b0ab56e45a3035c88314614e60eaea48b2a29ffa16
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:587af68ec8b4974b29da46ede525c70054703fce8d5934f0cbe68739c6058fe4
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:fa01c7d11c23a47bb84375389e69619455c2879b5ed6bb5b9b753b945dcb02c3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:08dcf99453b4dc6dfb5620410b8f7e4eeb42fa824aedb2e480e2fb40ec8ce347
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:0164d8fbea591bbd9d67d84603162c3781b08551eddcc2a8a842aafc828af18c
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:c957787ae587a31e6a497b46649da9cd46102a2a3dff29243fd5cbac96e7e2d5
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:4e65e96603d5043e78a517a4c6e5debc5df7e3747fad4d8025ef1fdf1cd6a9da
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:e3d8d4d5285d058130345d6108ed4056ea9c871ed8c0575601a94629e86ec691
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr@sha256:43cf735b51e2ef45a911ac7032f03c74cf0125e2d2801258d4aa04831e1c6ce0
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:4a87dbcb7d2ca2d121512fbf57b645eaa9ab745439b79722f474b84efa9eb391
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:6fe90f051d9261ad224dccf4d9f6bc9f19ce0e8d22fa010a46b2752374bce2b3
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:9a537eae7e581322e2f0764ed38a797ba7cd260724a285f24d5095058b7b47a0
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:3be006596f8e1d36865a9000c77df3223c264e807bf0aae6049a23307e8079e8
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:d28e6a3c65b580051ffbe287492dfad28dbb1c669cb76aa3df23bd9428f4bbe9
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:8cb37bb0f0461efc603cab89b5420c56990aa104fe39c9b77f0393bc6b641172
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-bundle@sha256:98cf2a6c7b7af5d7cbfbc61ce140201272b30b12edabaec2b2d4287b0690d89e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:20ceaa70a3ab5ed8a2e451a1e2960d04c201f0087fba1679fb80c57c6be6117a
        - registry.marketplace.redhat.com/cp/cpd/job-trigger@sha256:f2ea487072393a24e4c6864b2107c1fce072ebba07f9dbc2c139bd37fbd062da
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:83038bb7e3f4b4637f0d5095b42dc906e7fc27bbffcbd66f6ad93dc9b1cd6bbb
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:75ca37701dea1bd079df8f8cf302cf95e8dd4ed817fe126b192ca368e92fb78c
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:b10ca594f31751faad4050b4f7717e0e98c7977fb7e26d4866e1c14125422225
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:b2034fae4f2287c8997104154064f2712b98d3c4382d1faea3df777338711f23
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:849ecfeaa36fe365c7dd8f1a76fa0d263a1196fd16af805dc66035643e8e9d9e
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-api@sha256:a76d3114418a3df1d4d3cfe03da587c37028ca1697795e2535cac9f6f190457d
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:481391c273cffdeca36f91543732ba1f082dc409ee851ae95d3c1d205a7a6a08
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:c730fd4a7baa3f46c03f6c73e2c6c7c0ff24a8900b79dfc0f2fd319909edce74
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:9d4ae40d34e38a48ad2ad6556e997a42f6308006d5f6b8638cb7bc4ce5b1e40a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:6e0a23c3434046195428792c45c95869d81fb4ac4ba75444e4343963fda6014b
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:deb66a64288a370c5da5c7cc0ba4e2ba4ab816d7c5a080a0c505b05f55fa6072
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:443b51f8b10acc85bdefde7193e0f45b1c423fa7fbdcaa28342805815c43db3d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:53d1fba68aa0227c2f4c8de26837f6c114c56ca4a397826e1f2e93bb857e5b7a
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:42ffa27ae47b577cdda214e0fbc46c10bd108211e7dc58777d8236c59cb78766
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:99589ee11b50af8ed6355a5f9c2272ce1fc7f9a13e80fda8eda5883a332782a3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:a0273c013c5943c9b9a223cf1aafbc96b3cdd2dbb147c73d2a9ed49befdbdfc2
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:9a0c76fd9ef9b2dd9af5b0f29e1813567ab7f7d35c1efdf71da1a315f2942b45
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:e3138406cfe81021370924992dca891cba22b3be2eef84546faa0e8f8aa57ad6
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:62e2b059f565f87beded59b36962c457c9662d1e401a78766fda328ea1aeb78f
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:166a3918b942a4ea7b6c371a1e30fd4c928586d7ffd5b6919c2fc728a086ccbb
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:d30397b74780cdc14930a3a02dfcf6b99264847702afe0513ce536983c32e2b1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:f662be48eeb6dc93a8a4b0c3374f59bc342b5b8f5360437978cbac8f74c3a084
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:12eece1fc0a2e14244a2f57dcf7c198a792ced9ad0d02c9606645bce1f295877
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:130872119947fe91cbfb8dd82b3baaf78f7adcaca5933ca394eca3f9ee84c8cc
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:f3f8c6c395ae860fed5761f67ce092a122e7073edbc205c9167c27a4ba85094f
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:aed24dcc7ec24707c978a17b86501b65c6f0a77bd3ea0d173c28684b9df9cd91
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:8b7941a3f518e952766a2e6f88acfa7c9914b6d2bcd4ad9ddd54df313fa30347
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:722944bfd0eba280255f4c198f65f562423a26e2c4ad3580591ee71cb6e9c2ef
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:0341c499ae144ee81dc2b92e8b99ee313ff6759374b7c1a6469156c20f522921
        - registry.marketplace.redhat.com/cp/cpd/runtime-manager-api@sha256:a37773df07fe757dd698e3f09f8687902e7b3592d97855e703dd92ecc832aab0
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bb074bdf89cbc7c703fadb17639608599ab4f66db9be5617e5f79a0550676790
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:e673f1d0c5a0e32e9abc7ac40eff905ce411f9f3d8ca0b03160e23ba5661e041
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:f2f256a2fb787419b25e99c765003d25d1edbbe60042151d91b46a5d029fc0e7
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:a16b822b41f5a745b475ceb353a459e998f9f43bb9bb867ca1640866ba51780e
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:8a6ce0dada706056efda73ad0cc1547566003e0a7aab8ed379b50e4776861b94
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:ec5e4a002cfc63c1ff1bec340111e8a1755c8624066445db36c23fdd1c3061f0
        - registry.marketplace.redhat.com/cp/cpd/environments-api@sha256:6dd60b530f81d7a2d55a534211f7f128814abf87b8cd3575d528dba53442c44e
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:b0c8bb70a01bb19c685d019e89f83d8ed327bfa670a13b78d3e803fed77499f3
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:2f2590734343c99183dba7aa9fa4debc73377ee707f9bb090680750e62acdd68
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:25420a21686a32c62e83b99fcb215df015dd13fec10f73f397f927c61226f964
        - registry.marketplace.redhat.com/cp/cpd/spaces-ui-job@sha256:95dade51444e30ec034c12207ecbef675b70889b81970f49f10aa75160b2f947
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:686700e561181ebe1431f71eb9aa11b550bc4361dab8a3668ce0c7ead217f79a
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:3d83ebdfddcbe6eca38705fc3d575f2e70580cf9f7e8735941afee0a33d3b5c5
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:99c863002ae1e93e97792fa606128d05063abbedad9e9404acf28c745ae1a5c4
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7d9b2d61a4eaac645490d1c9e6db97ba193d65015b6efa8d5a1f3ec6cccb1ad1
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:06403c4ad99655cfc10fde46b21e84f90c05ae143bb6d08464dbc8954486897a
        - registry.marketplace.redhat.com/cpopen/cpd/cpdtool@sha256:8ed09b7d65178aab93b77b9d39371bc86b1022f32c5e63a9192baef02c71016d
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:ded16bf3cc993bfa5af8907a1517cc34ecb511948e4d9ac73bab68888af5543b
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:107911ed6a60a024d151ec3724ddd6d2d1402aa70a519b9dbec1a519d88231fe
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:0157fe80f2c67786cacaaf1302c334a6e0ec0b941afd870e0c4ba163ce394216
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a24f4cb635e8b5759a5e2bda20ddefc6d5ea9b16fb1081592885c54aefb5480c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:0711e1c6001586c950e74fbdeef8f04d20ce7b93de2ad033e82d2d5421338936
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:ad892ce5e6a3f69ef484eed68e92d9c972597a86420e22ff80c247689ba02905
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:39700bf47b93c7f0db1894148c16c60eaf406196034351f54573f0a94c4c8aac
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:3112fa03dc9a530642d329666a0cc64c06a04308beb1ddca0f7472915719ef5d
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:66c1d2493e6721094e85a0c65011523004f9589f58b7f68cc6f33c5e5adb6984
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:68ae72930a346d0fa0871cff573c64285266311208f2bf64101697a540f60272
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:2712755961a1c1e23f30d49341c037858c0965dbfc065aebda88c84526291f21
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:4b6bc5640d9e68519a7f543ddbe3f16ca200f32a74ce9c07e996cf122da3b59f
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:39e6aee0bb64fe6dc18ce81496936b113de6d6746221655cea102e04c11dfa33
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:07142ce0d259412d0a9908b63dd78d845ef6b7c5f8a24442b93964b4c38372cd
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:525924934f74b28783dacf630a6401ccd4c88d253aafa2d312b7993d8b9a14c1
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:73c40c6c72b7f2399fb0ed7e462baa72c3e8b5aa246fd2fe1082904b4afc0796
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:fcb79d76f965e74f7e3459abb92e3561ba2f016b34e2ebb6f88298bb19f7bc11
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:46893c166cabdd8b3422a190e220ee2e34861e797aa068100ccf53da349e3fe7
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:a477d47eaf8f8e569601d3c23c1def66f31c098ac3f5a1ae350618c460917dd8
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d51033417b786dd2dfbc94e648db00ba68f316c23b06a0bc9e36a5c22f579124
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:86cdb5b6543e7184361cbf0b0fa556277232460737ad254e920756c6d365e47c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:e6ce3c998642b235080daa6c0b954733bd3fe973fc8d85a10ef0cdae2a73aeb0
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
        - registry.marketplace.redhat.com/cpopen/cpfs/cpdbr-hook-utils@sha256:63a5606df81013b8a7b5cb07bd1f29bf032c8998db843885184fe86a76fb93bb
        - registry.marketplace.redhat.com/cp/cpd/environment-specifications-sync@sha256:894b4624d8ead7286cf9068c6477300992c32d96339644fe764bcb29b1c8ef5c
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:3a872814d01ded278afc8dfa6b1569d9e933e8fbda28c55a991da010b4304c69
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:351375de35d7fc9d940b905b867ce8a00b7ff1416e70d3a100918797fe3d8c56
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:00c30fe372524fa82b5ed4a637509dd9d4373dc727a9abc706221f832858f9bd
        - registry.marketplace.redhat.com/cpopen/ds-operator-bundle@sha256:be2ddb821274938e5ab926627bd998279b6d19b7789d9683513e85a4b05bc2d8
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:edfb732843a53254ebf7d4a2f109af109623be8fb9bbae94366d30febd2d0494
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:ee15594f82d374da81dab59c559567af74153524a6d9bfeb28ee085bd4109b37
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:3ddc6bbed1ddae1be71043d177cbcb58830d0dad542b06567fcfa254dd11e626
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:75595bf2753fb43391e2a496fb6d59faa8785babebb4aeb1da34cd5e72c93699
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:bcfb9b7d3b379ee2bb0b449cdf5af3fbc88f99c77641441ca392ebed04403e00
        - registry.marketplace.redhat.com/cp/cpd/environments-ui@sha256:2923664475c7a8ea5e63055c90748db37825fe76c7975b845f1185d4b32ee4ea
        - registry.marketplace.redhat.com/cpopen/cpd/ibm-network-performance@sha256:30e0b58531f876a47174c14bc6f94a3c38a40f757403775a06745d4ae992060c
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:119be0bf4c60230c43bd64c0ea252108a0173adfffc9f05e5257ec260b335599
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:3c1e08c56430e9bfc0c831be2394fa28e395a3341ce5cb5201b08828298deb60
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:2698150edf65a99a1d24bf3e8336f02b57fd964fcad283e182794634b033af89
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:077107dd9cede2c49beec429decb66760372a229c476ca1089e42e4ed5ef40d9
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:9782a9de863ad10ed36341d07ea5087e40f6f213cd4453c3ac2ec100d8733ce2
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:894d67753dd0c5b8625369e53635a1ff53066d3910b73d91e5625dd8907bff43
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:6bec687b2bd5c4d04403a25f56d2245f822e914907771f26c25166462976258e
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:993f16d744a81c95e73339d3c8d5880b5cdbcf29beb914fcb2c6403af6d628a1
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:21ce62d80df2575c3aa7c8dfde706cf8af01b2fc86f72ea68f767bc1770fe733
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:2565a8ba9aa103956d6902f2bcb53ac13328b647a2d7a150bf80ea2749d5b972
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:17bac3307ad3843afea77ed213279603b5f775f24dcf4857095baf5029409c18
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:749dd365ce85a4c9ac8e6ea7f90a34b791a7bc5300319daa85d427b04f40af65
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:cff7466aa6063a2bca677815df01208822930f8c12784c1c406fa352d6bd6cb4
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:d5feb4aecf8f36531db4c52e672558c04f7e314a1a791b27b5638202f9477bf5
        - registry.marketplace.redhat.com/cpopen/cpd/ibm-service-functionality@sha256:ea4c32a19b1549698c8a68fa030d7036d3ea3cf2d000f3523b6e9fc7c6c708f1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:0e563d51d601adbd710833cbde26ab8f857092f85686213adfd6fb3fe5e2828f
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:f7e68137e1ca58a63924967a8e4c872912557fa68af8a7392e3e0d4aaaef1e78
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:009d98ff8e3cf353cf8ba71a973228e1e409e4de21373498534a6cce01c07ba5
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:fada47a5f4824d3cd5bdc262b92018ca0e697e18dd373f92fd93186fb8719a1e
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:f88a7294aa200d3c633c0a25c4e4ccda52c75de174ba508d847ffa3092d5b8d1
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:2d320dee224cc4818a84493723f2262d259891bef89dedef09e1f1e6ae5686cb
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:489d7f80d83c9890600e3c9c4cb9442a6f42b9c6c4f13dafdf0f614b72e67143
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:934bc42fb29cbe616edcd93d2a060dbffab5e5a8a2ef4caa7b2e538f26167ee3
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:96e6db93b2d16ba590252d6568b68f254548aaed73ec17e4bdd8d530ea668be6
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:5f0974565de8b045d3faf6acac7c8ebce6cd35327dd76290b39fe48b27ad5b62
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d4dff1a8968c18d984e16c4c9f1517c3a10e23a11bf166c64d734773324d4c9d
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:c411456aa41fe705b47dec60eb49c6d65d4b7bd1a64d1e0545817efc95afc67f
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:f5527ddbbd0b9878555fce852996b13020fd91b1604d8c03d34a61ca2fdb2d1d
        - registry.marketplace.redhat.com/cp/cpd/jobs-api@sha256:173b88f90078e1817686ac949f48387af70a706e23e6fdab4a7c6217db5ad86e
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:85669f2a35ecc4ebf6b46fd793ba6a6bdf996164c00662e65cd7ceece8d60cda
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ed81ba4ab2caa7ffd1cb90e11aa8915df1c5bc1b76aeb4c87532c84097bb102
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:23337786ca20ab624eb428b0d53490975e2b08a2db642ccc76d974dc0e455f55
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:b8c83fe5215b7412765181aaba84162b6cb188b5765e45972a6f29b2a9f19826
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:811d1cbda1d0287dab48cff90c7091dd1bc8aea385f562c896f2b09f075cd1f6
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:7a0e784b0722ab0a4c62754a4e817a1f25e5a2d5e960de76387971fa1ff7da41
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:36e3b8e88cbb895ca89288ed385fb800b888743844f50fb2233ebe5e6c7498c3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:de7094acd65a80daa67669d8eab809220dc715b9cd6aaa96c58e5fcbe34c460a
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90046eb6170e12439ccc5aa343f19a4d4fc99a10d98b383cf1dc0e020ebeed55
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:164f6063f3a50a36e28a240ae7c060c4f90266b7ed799bfdc075f90628bea7ff
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:78fffd8a74eb6495f104ef346bdd03a5636134e2d5e8da9ef5c2647a9ebc7624
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:19bb32e0389361e8314e627bcad2e9bd93984289485cf632466834353d9c3449
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-oadp@sha256:35594a16332ba3ac00226414be9b85243bb25dd0648e256d7ef9d42137dcf914
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:787e01e29abcbb3d1fe58e0b6051067b309085459983f38d6bd18d3811d608d9
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:2f45aabc99eb9646db4c74686569a7f2af3f27cc44362a30c92d171540bb274c
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:97b9ecef8caef156a0eb073b6eb2a34b63209b44511f5694a243f11285d88662
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:bcc579eac4be59609102836252ded82a17ee0088e5ce28c684fff41c80232a0e
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:da38999d66199df0eeabd7c9e934205fb858cf2c52be8bbaf6b34adb894dd38e
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:e29885c1ee083a73b952b0de8d1cd675b2b2830abf44d69a8577f56972a601fb
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:17f016c11039c34f90f67279b45ee488740cc3d033112a3bd589bb1a60ce6812
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:bcadf3d367f76d3454e119a9d39f6fdeeb6cb806c282f21a1bc3c38bed41b427
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:ce57bc323b95db548dcd2f90ec938e708988cf6103579e793ee39d36359d60cf
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:5f63e062489b8e1bf49b5943e7a836174fb47af8dbbc8092ace5c8b72baabb99
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:740a1d87a43e61e32cfdea032d852af33100d959c8b57ff2f165f4baf77fc011
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:d72cef9b19ea48f03f7658dd55cabe6e118be0209852025845d426b1fbf0e0ca
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:1a1bc54977b160b70cc00cc4eb1380250dcd639fa3d741363212b8081b35f72b
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:ac128758328325ae56273800ea010f9559902281cc60226c8f99de51b636b180
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:26500e941ccdddc08dea07d3a4b9aa9eae7840fd5b683191320060f6dfc88c97
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:19ee4cb8e3e60d9c1e8fa63b0bf3e218bd683b6c05b9e9553d9477c68c493a9d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:7dbff355db7739152961cb6a97887d2e43bc960ac58837c126e03ed1a4480a3a
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:075b6bc4866b812c4b5394c8902b012978449a710ce71a30f55231fead941105
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:697baab21258166a1897ce4e056e39bc4667ff837bb97cadf92dad5863242ceb
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:ef03bb6ed647d92b9c84d34810376da7049aad88589eb2e35798d82a22f5b994
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:6df671937bc55c410a9df13aba15687b2832c7c19c0eff4686c3f83648528068
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4dd54195a3283a820d260eaeaf83463722e49bee05d875333c1b99201a7e2b85
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0a773245bec514cb893f730086f39a3789d0204b82cd11f5d11a5c270b92f70f
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:fb15cbade10fb37b14afe063a93decefd9a40583895b7b6b436f36f08bd73aee
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:a6e98064effd12ffda0613db98e795e056b83ef77b5deee932a762b178ceec3f
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:be5679938413dfb7766f765fba7923d3079d2d9b6f464485a8309e8e306f19e6
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:0cc0283b768c2968849f24cbf7178213926e084da5d0c90c18d38e2260e19138
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:82f60c881048618473c07970d6463b3186518bc18e4b263e16ca61d15448ab30
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:e76b61f9f0efe16facbddd80808d80bc32d0d41b4c16aed566116f1366152987
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:b368c22956707b8e594441c0e9bac20981339a5713d4fdfbe3c64bbfdf677b35
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:9fa6eea66a3daff39b7c0c7839ef5349b483b29dc219e1db384aef6c4c31ecb4
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:dc257df716b270858ca5d07909bda9ab635a388c9887d30b015391f60cc7b5c1
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:bb45d0217b4bc4352b209e3771756cb82923b8d260846c75b33ec270b4a8b8dd
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:b797e00eb32ff9a6a56e71d355bea653ecdecfecc121ddb197a6e347b89dbf8c
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:0e0ae96aff8540323da8a2379af014d3af1770e4b0043b84b9b415dda44501f7
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:37853d2979d0ca5f88fed76fed5194b628507ef0480e668dfd6eb034fa1d2023
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:c0459e5d3acb3b2a6d9863efe6ca1a24e529b38a6ddd3cc1e13d143924e872e7
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:582ced3f60212f736f720558524225b0d4a1c79ad4890656c47dd58009ffe0de
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:351358f551dfcf817c5936c0b43c2e28db61adcde65d7d2c81cae666630426b8
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:07c13b18197ed0bdd8b97d65312a342f95d7a7fc43d50af746d788032a67f83a
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:67515618d631b0ec00ad95ecb9bcb48d7f199855af9a64e86c4b46e7ae3be69e
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:932825fc7af8b96c381d504cd0976c751661ce078257430e5c79d565f416bacb
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:5ec262d1ada13a7d5d519ad98cf82dfbbd915664bb7560408558001ea6bb8451
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:3be3ee728cbae6bb8cf29d47c37ed1a4cd5791b54e1262135089424e5bda7ea1
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-usage@sha256:fe5bc4a40472926ec0df3023777b2aa8105507128bd9ac02eac2c68ad08723c1
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:f6e8de9f11696ca22302b8576ffd5bc2c024df0c8dc3734c404cc799565f351a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:36f4c57c6594f997383c4731e1e62c4765b8642c0b5c7373ba22118220928d09
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:2d8f924a7e4ef373df9760b4f1f3ad8050facec7de0c0efccd8c27d46b756e91
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:719b0871f0d32da894ec80dc2a23483623b7802a5c6f935cf3672468f99faa40
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:b372540b630507f4cb74f5eab38d0335d7b5bd64197e3f83852b2c7c5ba9067a
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:f9c9202e106d5a159c8f6b8fd883da4d6ac0d81e75b558fa52e0f7776269404c
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:095c969026e191bc4fceb020b5760e7595e439cca91b88f0d8c3e422591d56aa
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:8d9e35c3f29f8f56d55d81a2e6cc4d1b0cc352ddc8d33d83dced83c4a111a0cb
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:88d3c1ee02acbffedbf6ddcc27b80be6727f85425c361d45cc5e5c4056cf6733
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:1279b116fbcebb1c052eb27a72c67011b44a12fe673a86c1ece28202afd911d4
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:13a9ea6ee24d9b0ac1d5fb09c3e42eff731779867e7e71ce40429b43ba2779f9
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:5e5176376a664755fa776efe623a1e11bc5d3c31a3d0af980b165e54cf06620d
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:6fc976bc622c2165dd8bf93c9244e1c29b7c973a4da7a5ce4d0a097a1c1ea4fb
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:18f44ab15289625fa2c6b72ece6e06afb51058dcba30b8f38b4fdeedf8c9e51b
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:4fb2127ba7c525a959bbbcf024de364ef266ebb3ca7e696f74d990d2552cda59
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:a99b1f5a4eb0306b83562de0184fc826c34835ffbf46d929ceb9e70e0d325361
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:2011f32eb1e68ed5b75a4e8a99c992eac546995e219701d17a511596a5a09361
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:99d0b3192e6e46167423505f48655e7d49af1a461ae00b95c9c7c9d4209e5b65
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:65d327900e3b7a874dfe4bfbd0f2c7a1ee402346aba0e8cbc628b406e5b5329b
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:bb85f2567aaf7b612343fb9819fe05e2e771ccca8383316d34282ec37fcd1383
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:b2ce782efc5349a49d9c7326e2f72b2c1d99d30eccfb6975f7cba10008f9b2d3
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:267b7da5936493b0daf8ce824e3e6b50ef3b2bcb6e3126ed18fa9e09cdb0da43
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:27195e9758afdf8521971866fd60ea3dafed91a146c3c4165876e5ea476a4622
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-bundle@sha256:cd92fd1c85a72560241fffcb3bb62cf965f96bf2a4e1f5cca84c0b52624ae962
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:a34169228887ff93161584753c990bb82a5520ad7945dfa1ff4078e70aede592
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:7dc67256c002e86ee416c60ebb2942b35d8dfe0a5f96b4db495eff7361d21466
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:7de2c2f7cd478baba2011461c8d28c1a4c70aa485404d4523d1166dcb94c0a79
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:427e431d5b8b56011a91bc4f1e1586cd5490944615011fa6ac20abc8f45f92a7
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:96451d4a9d1c4b759d82dff2eaa9aba2a6fd58518f6ac6306cc082ebb7766a64
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:eb157bfe50005f9ffd3560ce1ef7ee0fc283ddfe265c42b35a71f18409e3f561
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e1b10435f94bbb7165703720f58aa8c68acbdaa18006a2c36d1e24ffdd72cfc3
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:411a4c7164b59ab3eed86e67aa457e45bd028067c937d35d0a227de63870cf87
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:3743af183b7b801f72b279a658b0d32da0bdbfa526a00d191a9a44de6df2284f
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:2472053d912f693a5fcdd00b7d9f0a7ce98697c47eae955e31c0b602fd8a4a37
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:b2ca6bc1119571cce732f1ce2ccb9ce406b298550b55144f950bb46cf985df3a
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:7cec55c00827b747f23c2f1703d99e1ca5b30a0c445f8ad0471b25b0c9cca57c
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-ui@sha256:b9a255ea269a11c85f52fe556291b0e869efbff9918b488b65e4297b958c6059
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a4daf28771178b0af391965684337805f8de0019a677753112961710f6011991
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:b7468b824d519e46e3ea23f5a803a8ace030b74d71d4cd282819027665ff0e12
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:3f2f97e6b2b18526255e88ff02d1039b24312eca51ee22b7cbf76d5680f7aee1
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:183ee860bf48076ea155db08032eed7004c1204a363625f54218734b00175441
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:62102f79bec5a8f66c600d184478f2b42c8acfde059fc71bc262b16dd46975cb
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:3129d9c7b096f9fce9cc12002eafb5c3d731a162ddef61578455dcf8acd31fac
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:a72c2b70a9f0a93047d47247cddc1616138cbc87337289ddb208565af6e4369d
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:27210840cbff8a7c61acac2ccc328a415a405b4738422b4f47eb9a0110127d22
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:381170c9c497b1bda33eef3ddeb1bafefa07be1e6ca75e75c9eaef9dc87b8d4a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:6bbcb64cf9adb4d3b6a4c259eb910e7da5c3c58732643e60d416399b2083cc0d
        - registry.marketplace.redhat.com/cpopen/ds-operator-bundle@sha256:e30b09797e341a5fdcc3ec377657e0e65be5855fb850dfc31f35b200f43130d8
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:ef24cc79265473a8b6fea231b1930c5763293dca79eedd46c89ec63dd95812c5
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:a31e0fb7a05e969e31c10a74cc858c483d5d4da14ce3abc36f5318100e60e792
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:aba8f1d08d143a8901b4d2cd8790e067fc7b9eca9987696a1ec6f9f505a686ba
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e12599ff99db058fdc8f9c1eebfba31160c636ff092138deb1f6f6705f44fd62
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bcf193d6f30ce1c7f9337ba8efa8d9b853eb7961e0008aa82635aa383fbe9d76
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:ab31869bdab5180b816a89c45008328cb7013c26b09c531bdbe5b00382f0dd7d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8ccecc00f3a9656a3dd1b07af76ef065e301e4101bda4ae8dc5bb91097cad7ad
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:c2b9d370b6c3e3c8e30079465502ce376055261cbf0f95c48d8a2ae10ca36c76
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:a6d5bd253959be1f91b50f5f2c56ed77e1c82b06483e62094aa86ff57cd46eec
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:78206c6634f2be2bd36806eb717bb8e4db2c755263cefa6d026a425a19329324
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:673320deb0ae96c6132545032f6b3e569d40204d3a5565a568dd97c5c1a4a552
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:3f6e9eeb659857799e30011f222924cef35301c59eaa5327ab5e8d212af75ce4
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:be55d9ad6ae76bd061046af448d032e65926abaeef95c831c978e01e2b0355c6
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:f6f3758287dd31889933ed0461fcf02dfa85d872c3d193669608573ca5269524
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:14da175e95c91864142ae00e9087ac2fe0e2083fc1a3bcb20a6b29dbab99d8a3
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:0cdba17b95a959b6d04889179d105515ed3176ea3449bc104f18b86374876048
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:b19bbc36033c3e60f387a9924711aab1ed81f150ca9d8f435e5433fed7f6a5c8
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:a3a0d58f895fca4d832e602d0dee80018df4ad7bad8800ad4058eea07fda4817
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:8d744af7dfeb9892378b7cc67d0e85e29d12b6f79c8b6b2a53f7a53aeffe8231
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:fa3d8a54b1ba33465171240d66f62271286cb120c51c54267d2b4f8b731c6177
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:7a90c8e0180da12172b23f3b54536bd1afe35affe37d908ddd78996a81435751
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:d2680c5f05c886b44df644aef6cb2fdc9534bb6a238066f6ebbb97bbeaac431e
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:6ccbb20c9fc7ce00ddfbac19c8808417957481473eeb1ee62ecda165e0e275cc
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:ca175109562cd9140d98bea3be88d6b34490ccb8d79f0cbbde08c821db9c5fe9
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:4f0165926685f49cde22a93387fea401bc1799ee846041fe54bca148dfc2f62f
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:70d23c45cc636af2148fadc1c7a08802692fed14e1a31893214b8d2d82f06e1e
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:9ad526bd56829ef24834b8a790b29c50809866c1dfb60cf8e0600fb9a903d388
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:0ab7d6651cdeda78fd0416137eff72eb58a621c8cd55bbbcb16f269808dec7cd
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:545b41cc6c044d6807bfeb7453a045e074a5670cfff9f5c311f4918946901fcb
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:17f7039e49f28abb7239e0c4f3689de58c75f34c9bde9e7b33a6ffb93ba92677
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:0fb78efb9e92dd0c2e52ca2488e6f1e7431e197f87e6994e15f7d621b90319f2
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:4deb399f28d3385c50a256c6e927cf27115c311be5943021733e4b281c12a419
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:5f086cdc7a272384c8a9b680f8f90fd471799e9e96175b40c7f16126b73169e7
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:786b21db2a9ce6cb0eabd5cb7c93dc41a8b1db6c1f2d07725a5bdd75b3a1b7d5
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:cdfb097f2438b5739bb88905b8318d15d15c4c64d37cfa6fdb06f44dfe976e6d
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-velero-plugin@sha256:2dd02d8f15dca1dce7cecad098cd8791cf1cbeddd2abd8e8c6d8f066d85f9e2e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:9c635194a9c63fed440e7c0ab0be1fe6cf59b1c3f954216dfb9150fe9e0fc207
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:044b663e4b7508c62dbb28b7fb2aa267c9771fc62ab509b08d46ea1499d95431
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:9c24d02b1cbb3857d017453744bbb2e68fbe133210c35e7da5ce0341921d5b08
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:cf2b88be76a4aad436f45a79694f32d2e2175587c3630543a6d24d7725b60c5a
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:9b91aaeacb196806646a1f196d4adc83b5159de767fad1a2eab699f735938911
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:d023f35b65816f244b6d699b41362e6031553f7d4c416efcee4a7f2395a61355
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:1019e13e6f072c07da6d31277c8ac639e8ff2200599ece1666a3e34745e763d4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ee161954766211651d7cb13f3f5a3c025ac04094ba742be453ce789c6f0e24a
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:4d24fb274b5429ba61fc4bc6833e93b14b92b8d525556db64e2172b45433de31
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:772f98763527c5452c662b2ae516de59c1f6edcf339da46231f4be8aab6c580d
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:bb6492c452fa6b481551628da26bb85bad151d12afd34a89c1a272a9587ecd1c
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:00cc7d072ea6b6f5e9f3e5b95b9cb4ab8ad7bd942d8c259e70f8ba0e9c09a464
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:9c160303552a9b8a97b951311018de2ada410883468680dd61fb41519b9a9e7a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:80c0913e2c84530d9b7c6a2152de76ded5754c134f6768754d00944f46d0ed4f
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:777e090e0394ffdf39abdde285b5b2749d6835e61458f117dbdac108186c34c8
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:d7ca81372914a28c8298de5f9d3c35cdbd6d8d8582ac9b8fda5a85d22e3c0def
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:07c5e4b3dc18b921339a7a5c78b8fe9179205678eb662377e8124e46479d7f25
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:2a50cd6ee3cb352ba186a57ba73fb9df321cf98f0b79b3d81024cbe57e0c7f35
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:866ebafb730e660062cf7401023d82c5242a2b80cca0126d47a68193c84670e0
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:c2aa3566f8fce992f00f43867173e5f47593e8618357369b23044fdf3c67ba53
        - registry.marketplace.redhat.com/cpopen/cpd/cpd-toolbox@sha256:1b3469f6fcd037aacabde3e05605b908e22f012d331d7365d0450f6d933d3efd
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-api@sha256:dd9e4909317512a6fada5ecfc7cdb47c6807f266fa456a01b60d7b3486faeeb7
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:fb850a3d660729fe8281967c2cb466fbad3f01fdc302db17988c27682217ad29
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:6caf5203489492755e5cf6bffcba53be0bd32662e8cd331dada68a85b2c083ca
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:7693312b2bcee7b8502de0da8786be985607fbfe6616bc9b0da4af8a6fdedea8
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:4ca30e2017a0ff2d64100a03a84bee8d601c765f0e624e1f485cbb97748bb8a2
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:c35eaf4921bfd046985ea7a03fc325f79766bf73e3ddba69408f4467fbd7994f
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:4184f25155643f01850f286f15cc92294a92d8fdc8e5020b1b522977ba429c2c
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:61155acdb8982648ae5748be9f5944ba913db1590ec3e773206d6fbb1d8952cb
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:474187fe2837da4b86ed9482a96fbca0a7f83db149859f8ef23db84de9a6d7fb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:55d4c6a13452f76bbab485363288efcace23975623b8a4b6b83f3c549f844c58
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator-bundle@sha256:c743dd7de2edbd08c6c73e20cf15f3b147491ad77a81677339e88caffb93e84b
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:b8886b9b54e6ee4c6b2981ef092e54348a0286577d9924cf136cee881f5bf1f4
        - registry.marketplace.redhat.com/cpopen/operand-deployment-lifecycle-manager-bundle@sha256:68f71f2767a40ca9333f0a670e07493e8a8c13a5c7ca877ea5bd8a544b352d9a
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:2754384c6936b36ffd64f89df751cbe47993f73dd700a184af35f412b3408d72
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:887348b5035f5a70f1c31aede992a5c76397f5eddd2aebeb27074c3ea6248c00
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:1b29e04308bb8e3c990a8c41a5d3d1e854e197bbdf0cffb2833fea4866772ba9
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90d45b7a3c5ffcd30a1ca1447693ba4aedfb400983cf49779a8421a42dc0eecd
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:60dbe69693314c937a69bd213de438c3af0a8a290aa9b1cfaad8a57e9d04c448
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:ab686e9e4c38b7909994f9fb496943b745e1a07492fe61a61b9e8a78f756eb1c
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:abec4867d6b9294d674cc02b9f7f254e65e7de8eea0231fa94f50513f2b7c45d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:b6500764d4f6d281f1b42458eb223d21f6a413b5db55f2a2ba1a7973aad8465c
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:3c18707afb90a7dc1bd2ed1fff05c53cce89b7c85062e72261f538a12834ab37
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:245f81b08980bfd4790cc3661e609e2cf53f3dd4d4c1f988e5db4dcf316d90fb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:d76476d2ecd88ea7b4fd857e591cc9b71be9d31fe0fb0057baa1b21e6113f730
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:186e849f3cf782f0d78660eb4433853e1e61296234d230a9dc5921fa4bd41060
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:6e055b6e07cb1adaa652b8e7515c4ce91438593a56ab0ae051435f55373d96ab
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:24dff64a80b910f0b4c5442d4f611f9df7602a16c5231eff2c9d08f07b2a9f9a
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:0b640bda3afa672011d5da390af798bbd3e6dd9126029a9eef03846d41582743
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:62dda55cc870279d491679f6b43be177986991b6eb85d9dfeba3ed2e72b3f10e
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:25c98c67d09cfaa3a4d332c461fb6766700a65876a48a1aabbd73b58bc2653f3
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:bd237f42078822b038c13271f35e6710273c2a4f3fe2f3eb74429c93cd2a0e41
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:bc39ff644108a458d1b7e29fcd88bf53982ab4070ce2ea0e75b0908678b9205d
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:21dffe4500172f74345fdafb9beb476e0ccffa55f7bca59f372f312f8d28d247
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:8f9d3438f58604cf29d643691b125a4138275b2c09cca4afc8e4b29289febd77
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:e8e0ea5cb1a070443011794988f2f6c9d67758605dc2847bf0183e19bfde52eb
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:9f835bb2984329db9db7eb2ac5521371ce0ced216a7875e45ef3ec31d255b648
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:234a8dcd8a9c5a68533bce6e0f793324e3ad7a660fdedf95895c5ed73a4cee3d
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:48828f128c98e73f8e345d87bc228f0920a3ff2cdc325d4a4ffe73ad95c5c5e0
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:ceee171c5ce0a8e580dfeccfe0dbfb73474c109becd45a562411678f3ab08595
        - registry.marketplace.redhat.com/cp/cpd/dap-base-job@sha256:f208686a3f56d2980e7b8640b6ebdea9057109e7520c6817a731e68d08506a36
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:31edab3e5662489dc2fe9528d078966663c8d49836425e648b7e112f871808dc
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:bcdf9b7eea8753b3e463973a2690140ae7aa8c02c3c7726511f116bd751ec2d8
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:63d4a9fa16231cc0c01f1b8bb9363221ac1e93f1571ea132866f170b156731e8
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:8fe85b93b9dcab9c33306ae358e6910685e109f32662d02e2b6bcd8dae243968
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:da0190fee910cb8245cf7a61d6e7aa2bf599d41c679eeedc183950918ea2611f
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:ad19b65042fb26cc1244a92f81a96e5ceb4313f8c823181efe28b66e9fecf9c8
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:78e33fff4bfc2e33ea179eaed9902df9bbd11463b9152d0281d1e0924fe09563
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-bundle@sha256:842f2ac568db4db79cf0450b413b4b09b5f2180ef3541385b46fb650ac7066c0
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:c43770715b76e0aa3fd9b3c2b21998ee369f2142e3c5d04320c677c9e9b29add
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:4eab846d6e80d68ce488e2cda36329cec1a6bf2f81feadf666b1005fed294e48
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:aa8d647f007d921db565a505d52583511e25de4a6a1187fc6b4df15427944364
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:276d5d06661ed401f0b0672dbcfa3d9f07fa2cde662fbb532fcdffbe3e332227
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:c6862158a2a5ae0a4e4206bfdc7f73e1c2359544966c72cf4fc3a617acb13697
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:4794a0fdae9d060def1e0d75a98560e713ca18de2625e4d9e5b5fa4c4877fac2
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9669ed5ba26ef32ad36a6784b2658553c00ac2a89cf4e9caddadf7bf46c1592e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:8305f26c2ebc493b54abfea0757d65b679f6c96bc3a0840c89f08a034327ed9c
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:0ae22914dbaf09d6679bd06bc9b6a5728e744c9ce2309cbd5f01e28308c403ea
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:209ea4b77b3b0863c591bcdec6f53c4c39dafcb10ff13e27ae4e1cb986a59727
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:483bc8f74d02efcbfb40ceb6abb0ab0b6581523cdea576b34512709b98879233
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:f40a9375fe5722972f995a82b50c00f254e3fe37f202b0aa1bebf66b0504099f
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:e8eee5083cbe317b368c002f47ce09f98e0c5e3407e5a4712e1c5473c4090ba3
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:523dcc671f1c05e690d5e578a90300d28d76570f605d4de7f80645b590f4a462
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:7ca8d085dd9aeb568eec2c4da974c03aa72bbfe4648f5d66a62d0444a2d1c839
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:9cd2ba7ab3ecab7094c892086a5276bbde22ff96d26ce0d087df2c1791235b02
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:ab99b62d40ba05cef8e699ebc7cc65271ea6fcb5900a76fe17526e7d55b50b08
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:a64a93104dee64e0e7e3e381144340511f903155f72b7f619c82c1d440712227
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:aa0068aec14e37986620c982295c083df6fc6bbe95facba582ae3c6f769a8e3a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:a10a15f07b856a7803ce0dcd30a542741dcb591804dc74c2a5fd50c55de0a983
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:10a935e41cc2dc09f8f3939cd512d9f6bc71d68457f77793a6957c5420cf4bf5
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:b87fa8374976ccd3dad90ddbf6b65cd6ae5a028ba896562f284f089380ebe44a
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:81d8b25221d827160dd5e88efd383453f532057ad02d7a2ee3dcd9c6eb16ba65
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:973eda4612c905661ff5ebc28c7cac6d9fe16fe5edaa544548d2311496ed938e
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:b82a10ce8f75d4441e869bf62fbfb5da349a2f228b52178bb29723af62f13938
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:5436a25b6c5b419b77643a76eb073bbbf8f09517299dbf4a992ae14c1965972c
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:40064d2893ce54b358c02023f4cc3603d3d28202ba56a177b20636abd74cb559
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:ae5e6858ed698d67b4b3917c0d41d0b4f7e2564585694f834099584c21a3c012
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:268c763fec7defe810624bbd31c96407fa37fef47b904758bce4c58d95aac043
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-bundle@sha256:94662b9dba5d4bafeec53b8ad334e7a1235385b21ac0952e12975f4f0a7623a3
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:c59837981b9a2a645c9ffd84c40d4f9ddc3797911f7a785ce4e3a2bfdd0bbe5c
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:bcd879a7cc3442ef62bb9677cc5534a2bfc9e7ea248e566aa4f1f6a0242431a8
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:b87b6c32a2050f1d5ff311d163a7b45efa7d5de1d352984b2f70fa578301a952
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:b9c8a65c1742c72c6351b1c8f4a858940c2bfa00ac5ec271c57b2ee7e5e7a66b
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:78682d9b9f19eae363cec0b6d31e696b2a5fd1bd1f8429df6f65a197310b6778
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:96ebe954a21f697ebaecfe5aa2c497c60b2ddec17e9ec4b376f435e6c53b3c29
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:21c81ff7c4b8809accaf62b242fd9b3d52f235489ae427556b7f71cc21bf8926
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-oauth2-proxy@sha256:bf4052b099393b563332acd5dcba6ba37ac424aa431d2d4f21532838b5ed6d6f
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:f5248b19c4a9b605a9160ecdb5dbbe0906841b667ede6fd6cf8e1d6c4369fc8e
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator-bundle@sha256:e4340bd9481da998a6b89e94c884c46e17cfb50d9d0c1099f39a3c54e500d2b2
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-oadp@sha256:9aec2bab8c7a6941e97c18352cb9e838a99aec08b06269e8053ac764638a90b9
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:0a32e127dd305fc445808d0469147db6c9704cd15d924d57d791e035475ecb60
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:4e3ec61a34fc7c71d213999b68b1f5c3f866f97e19c12fb001d506d4166a50f1
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:99d3b5a88d20380bcb4161e3b182f305af2e4cb405460d0f2a91965dd6f99f80
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:173ef1b032594c4e070e2e013db469115ec58d970fa8b62f3b748163b46d6d3e
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:a65ed5ac55cd5be99cc43c5a8d33bf127e8694917fce5f83fb8401c04e31b1f1
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:83ca1235dc054bf9ead64932c85333ea0dbade076b4a0cd1c3fe2b6bc2e05896
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:4ccd786884e39641a74096d783742c219fb277caafe80b667a80396c3a9cf6c3
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:6cffef7631ee54612fba71efc0eff4e17bc08defc3e2137aee9881d685444298
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:a3dc5fcff6a2aeee7c7e562fe9f6e0cf0652fd7ad4ec61496021d366206e2544
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:e296fa15dc14ee827d5046e8b1434d7f6838059c41d355471e1275e09f087411
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:68a98f52b7eba7e41c78aa91602c4e88d21d4a2a2a11b9b7592f181ce30e41ea
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:dd76a8484628970b62238334e122768aeea820ab18e24d62d23a6328c83b868c
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:e9e408a13bd103fb46536f896c51e7a27cba38d9b224231b20328a28efe1e616
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:9ced60f18740063ab281b836d4caf21035242ac3f659c6bb022ddb546770bff9
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:ebcfc8941330da6991480dc18ac34d6df4fe879e562ee537e5c36df73f4bc536
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:bdbbd76c93f0fb8f838edd9f124b02cc646a6f735a68a430a00f025a06c5f407
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:c0b6ad3b88f216379fc8f14aa5f075f5adc06ad06cc46495d3a40cc765f3deba
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:bb2f19e1797a900d60fb9f05d9f115faf20d514000aa3acac55fdba701581d81
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:93f73afa23f9fe8b73fd389081281fe1b0fd74bba6aa9bf7a6df663053afd3b8
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:54d8ceaaa06196e1ff524b7be595cbe90811f3be1a57c767b1ba74f53edb2f0e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:20f05c662659be9b2b54407c02f1805b93f09a3c204011981b7f17a76caad8ee
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:758f9e49dccaec5ebd125fc0c383fb5a656ba9f2f301984f993d50b103208d12
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:7655f531c354cfcce438c61cc5cde2c7690bfee4e2eadf9f99e4b157f163cd73
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:66c11636fdf68631e37bda7467536093b9402e883999e7989dbdcc4b4d2ba7d7
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:0621eb9b83c9b2b8bfe5cd09797fe08fa37a854a5506914935ff6f0df599b631
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:57e3433f860ec2b65f5c5f86574614fa9273c09b6b37db96a9e1888f4f6932f4
        - registry.marketplace.redhat.com/cp/cpd/ngp-projects-api@sha256:df420e9f616ad8a123d9eedd117fb6cb007b30a772138db534f47ad8801fa96d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:6d903fd4bd0ef3ef361d22b209ba7b686601b49d6f8d4e29b5d030da8d949dd5
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:d904761e70d3969a4ccacd71a60ad4d77dda939c556615df8d65782ccc3273aa
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:74aadd4fe99e66b89471222f853e442fa8338a23d9c24b4e908d38536c66456a
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:c91fc0992cff4b674f5e3690398e7dfb70662c807b930b9b0e65ba507ff925d2
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:24ca69c2f80a4369e29aa382a94d64fd3546b110787df1142c45a20b8246e844
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:9021a04805bb7f21d68fc6211f844a21c1d06f700c332ec3238073065b7c224b
        - registry.marketplace.redhat.com/cpopen/cpd/cpdtool@sha256:714caed1042ee03461efc9256a3da1046bd8f43819caf3a91ec38b118d459869
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:2346f67ffdc1a802720b325e010018cdb908aa458f372fe50ac6d8a92d8fb733
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:ef38f12712c406f8206742dffae9378e344d66c774e5b37b4fcb0b657db456ca
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:7505eb74c6bf165d40f6a8f3511f8e891cc34c0201fee38240c693a08e002ab1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8da8b240176c00f8f6eadf32e5614c663cc1d56d546c73996cca614b715acf4a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:fd35c33c7361bf04a66e69112b05a078ad65fdbe8f8ca8b7f4ecc2143b1643e2
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:67a02a9264d9c6e82edc15ae36d4f99b3fa9099697c550088d27c76511189f37
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:25a4bb9cea1c3bdbeeeb205622512bcba98163f9629eba35625756346ce9d652
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:860c17d2b2c6b0e013d0f4c035a388e45d8ef1483f53aa7ad022913b025f73b0
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:7d8fb7d8e2551bf949edd33589393fd12f634871b486e26464cd089e3e4b429d
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9186c350fa54e8b792d72e47be6b54a67b610f726175c9641795010d45ad51f9
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:f289ac0d44803ff12d41c3659e6fdf04c404e0b3b233b6bc44eec671abb1b47e
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:c8d2825596b7e675c247158bdd06911e1cc0ad9af3364c9c175a5b4cf16d5d8e
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:f42395ea365edc2c1bbef914a55b091e6442d17bd87bf3730a8f9f6d401bf4f5
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:8160c688f6b66206f8428f78dcf5e21b65c3a1cadc38b765f22ba8403a87f4d6
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:aea8fcb72664a41c3881839086a47fef1b54dd672300d4aaf583fd0b575a64b4
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:b341d9af40ad6aceb4ecc792ec75295313128a1eea3a0974aac879aed24a89ba
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:838f7649df5a7c13d6155394e211ac565aa86233bd30a3a2aab9706ec4a74eec
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:8542b9e449b054e455abf6e7884dcfda708a6213015e932844568d1853b65eb2
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:3af0967b0fb0a28a83a3e007b2c78fe27ad56a5e9c153806c6a7b17d7d9b1df5
        version: 6.1.3
      entries:
      - name: cpd-platform-operator.v6.1.3
        version: 6.1.3
      name: v6.1
    defaultChannel: v6.1
    packageName: cpd-platform-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Arkmq-org
      provider-url: https://arkmq-org.io
    name: activemq-artemis-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: activemq-artemis-operator.v2.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/arkmq-org/activemq-artemis-operator:2.0.5
          createdAt: "2025-04-04 14:35:00"
          description: An operator for managing the Apache ActiveMQ Artemis message
            broker
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/arkmq-org/activemq-artemis-operator
          support: arkmq-org
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: ArkMQ Operator manages ActiveMQ Artemis messaging broker in the
          cloud
        displayName: ArkMQ Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - messaging
        - cloud
        - kubernetes
        - activemq
        - artemis
        - integration
        - operator
        links:
        - name: Documentation
          url: https://arkmq-org.io/docs
        - name: Repository
          url: https://github.com/arkmq-org/activemq-artemis-operator
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Arkmq-org
          url: https://arkmq-org.io
        relatedImages:
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:4bba8828e72fd7c9d38c22ba1c2a4f7130817a4e7eac8fb63c13e86465a8e1e7
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:b71d1a57c8b3820a8133cf7a25a6edde7b29e1c4d32d7c30325183415a8bacd1
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:49f34122ab46ab5d7ead094b6e40817b62c75b6a809f67db07d68bce750da51e
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:a4a0d0453f0761b9e114d0d4cc7c14091a2a9c30790f18ffa6bf1e365d5083fe
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:93168fd8cd4fb32f041795afb8fb0e21cf81081d6cc463a44356e153dc1ad032
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:e2438670831bad39bb08147bbebe67ef9465899fb7750f1a5a185569f99d5fb5
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:8fbdd1ba728058d5f437186b9055ca28596140068ebc69ecff4fa52717b19a82
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:81590ec33b2f00775d3c6596f610887b670de7c0b89148b09eab7584f5b0b0ec
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:ddaf00914e3a412d21ec3a814e24f3d4cc382b5c4d6c21e9b1b94dde16ddf194
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:2895b2cf6292c951a0cdde78e755514f5d7cb43ce9f7b6e7dc381d8dfeec31ab
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.7
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.0
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:3bbfb6177e4839087f615a01ee54f9e06a95acaddbb083e11851b6af9315dbf0
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.0
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.0
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:42230ce9d11a1949a5bb73b6ee218b1581f585bcd553485ee438e1c5b50fcf8b
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.10
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.7
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:8f592b2454d092c82b5f61b4e04c3aff173db246539b3c06e15a71960d3bd235
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:f10983c68aff358e285d89d827c27b3487a3ee3b8e121cdaa295bf5fa8501145
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:b293c1269166cc6181f68b8d803fdfcf20fdbd5df2787ed0453724b22e298fb2
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:f25db980c3229e66cf48690aa175f6f93e3dce3974c0e70ab23af6462e7d1893
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:9bbb915855f29ea16950a3fd50d0b9bac25f82cba7d197579e04d9c33a5b02c4
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:de82344daa0caeabc6559656da0af285436a687e8f371744c0c51bdbc309167d
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:e438bcacf099a92b143793e88c00c6ac0afb31effe4b0eed49a935c9265ad943
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:2d155bffd246f26f516a6075ec170babc1373c72bb664cba4e3c58521ffd3558
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:264b33a9aad90d32a41258904f35092a141d76dcef0c115d414475444100aace
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:56bae938a6b4cd87f9f5284931f440f9b9d116cda3b63b13f83a4b1c6c641959
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:da95bff69deb63545d9782ed20cbdd17c1a79c6b16165580bf96a1d5191e6a59
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:f2c853ff91809282c04d5b1381ceae366518bf07b16649a61aa6c60c47f10714
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.2
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:b6f61188379def8c796c738549c099bd40bcd9bb44f5c30682c5edd136347aca
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:fc9b9a4c7550d84fd90335dd4f9e0d8d39c09136c6034b28490a368368cb11f1
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.2
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:9e1fefc9a94d72a7b5a3f4e2d53ebd23e5ab221df17a5afde42ce973277838f6
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.1
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.7
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:634f3d8fd620607b56e0360a191d23a9cb8a32ba0ba1f24fe20791ec52343586
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:4be60cfa2a300b2eef9b5f10ea756bd3f55d75fbd8890232d907f8d64080449b
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:37ab7354273f4dcd19bad6936f10e92220d98b8c0266002221477a0e2d146c81
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:719178251893def955e444e706ff07042ea98d294671315d95da504c1f47de24
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.4
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:70d4c799c7d905fac27457d63e662ea72bca9e899a692e425609e4eaa45129e6
        - quay.io/arkmq-org/activemq-artemis-operator:2.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:5e768aefade6d18cd65cb1eb5ee233844f9f7e7b52631f93a0982b67d382239d
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:ffa0d7cb4068df84319776dbfb66210297da1b8466ccd5451318fc3de79736c3
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:6f2da9ecd8f5021964d0dfa85144830efcfbd0462d18f18f036971e5f17b3bb6
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:decb3932ebb5b848b90a4a23b083f0cb5237eec2844095d8c58a6798da8bb999
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:88a0098a628fe43a9b422f8ebb774086fb65ed389a195e2779482a4a63447e28
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:a4628c2591ec06caf21a3467f6e3faeaede7fa402b4f137d943b25b4dd4ad343
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.2
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:52f337cd5d7d5d199eebf4fc722f8c5279f652506f009efbc0f8f6de88e18718
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:1fd99a46758cd29fce846d0d590021758b2a7a3efedb817bb205359caa7d4f67
        version: 2.0.5
      entries:
      - name: activemq-artemis-operator.v2.0.5
        version: 2.0.5
      name: stable
    defaultChannel: stable
    packageName: activemq-artemis-operator
    provider:
      name: Arkmq-org
      url: https://arkmq-org.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Syndesis Team
      provider-url: https://github.com/syndesisio/syndesis
    name: syndesis
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: syndesis.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta2",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.12
          createdAt: "2021-08-11T12:00:00Z"
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta2
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.12
        - centos/postgresql-10-centos7
        version: 1.12.0
      entries:
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: 1.12.x
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.13
        - centos/postgresql-10-centos7
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: 1.13.x
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - centos/postgresql-10-centos7
        - quay.io/syndesis/syndesis-operator:1.13
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: candidate
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.13
        - centos/postgresql-10-centos7
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: latest
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.13
        - centos/postgresql-10-centos7
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: stable
    defaultChannel: latest
    packageName: syndesis
    provider:
      name: Syndesis Team
      url: https://github.com/syndesisio/syndesis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.22.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.5"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.22
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.22
        version: 42.22.0
      entries:
      - name: t8c-operator.v42.22.0
        version: 42.22.0
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: stable
    - currentCSV: t8c-operator.v42.10.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.10
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.10
        version: 42.10.0
      entries:
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.10.0
    - currentCSV: t8c-operator.v42.10.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.10
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.10
        version: 42.10.0
      entries:
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.12.0
    - currentCSV: t8c-operator.v42.13.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.7"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.13
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.13
        version: 42.13.0
      entries:
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.13.0
    - currentCSV: t8c-operator.v42.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.7"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.14
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.14
        version: 42.14.0
      entries:
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.14.0
    - currentCSV: t8c-operator.v42.15.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.6.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.15
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.15
        version: 42.15.0
      entries:
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.15.0
    - currentCSV: t8c-operator.v42.16.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.16
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.16
        version: 42.16.0
      entries:
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.16.0
    - currentCSV: t8c-operator.v42.17.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.6.6"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.17
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.17
        version: 42.17.0
      entries:
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.17.0
    - currentCSV: t8c-operator.v42.18.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.1"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.18
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.18
        version: 42.18.0
      entries:
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.18.0
    - currentCSV: t8c-operator.v42.19.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.2"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.19
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.19
        version: 42.19.0
      entries:
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.19.0
    - currentCSV: t8c-operator.v42.20.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.20
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.20
        version: 42.20.0
      entries:
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.20.0
    - currentCSV: t8c-operator.v42.21.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.21
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.21
        version: 42.21.0
      entries:
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.21.0
    - currentCSV: t8c-operator.v42.22.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.5"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.22
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.22
        version: 42.22.0
      entries:
      - name: t8c-operator.v42.22.0
        version: 42.22.0
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.22.0
    - currentCSV: t8c-operator.v42.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.4.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.6
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.6
        version: 42.6.0
      entries:
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.6.0
    - currentCSV: t8c-operator.v42.7.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.4.6"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.7
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.7
        version: 42.7.0
      entries:
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.7.0
    - currentCSV: t8c-operator.v42.8.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.0"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.8
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.8
        version: 42.8.0
      entries:
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.8.0
    - currentCSV: t8c-operator.v42.9.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.1"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.9
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.9
        version: 42.9.0
      entries:
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.9.0
    defaultChannel: stable
    packageName: t8c
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-db
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-db.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Db",
                "metadata": {
                  "name": "db-sample"
                },
                "spec": {
                  "clusterDomain": "cluster.local",
                  "databases": [
                    {
                      "name": "wapdb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "grafanadb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "brokerdb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "nimbusmosaic",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    }
                  ],
                  "image": {
                    "migrationJobImage": "wap-ose-cli",
                    "migrationJobRepository": "prod.harbor.keysight.digital/loadcore",
                    "migrationJobTag": "v4.13.0-3747-35-ocp",
                    "pgImage": "postgresql-ubi-image",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "pgTag": "14.2.0-3747-32"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "postgresql": {
                    "image": {
                      "pullSecrets": [
                        "load-core-secret"
                      ],
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "postgresql-ubi-image",
                      "tag": "14.2.0-3747-32"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "pgHbaConfiguration": "local all all trust\nhost all all 0.0.0.0/0 trust",
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "storageClass": "managed-loadcore-storage"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "volumePermissions": {
                        "containerSecurityContext": {
                          "runAsUser": "auto"
                        }
                      }
                    },
                    "serviceAccount": {
                      "automountServiceAccountToken": false,
                      "name": "load-core-service-account"
                    }
                  },
                  "storage": "8Gi"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:30:46Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Db
            name: dbs.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP DB Operator
        displayName: Keysight WAP DB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Db
          url: https://keysight-wap-db.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-db@sha256:5689d6a4da2fdd97223bc64aa61a8574bba810f93d09ab804cb45b4475f80ff5
        version: 0.1.21
      entries:
      - name: keysight-wap-db.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-db
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-certified-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "tag": "5.4.6",
                  "oem": "",
                  "imagePullSecrets": "",
                  "psp": false,
                  "rbac": true,
                  "serviceAccount": "default",
                  "leastPrivilege": false,
                  "global": {
                    "cattle": {
                      "url": ""
                    }
                  },
                  "bootstrapPassword": "",
                  "autoGenerateCert": true,
                  "defaultValidityPeriod": 365,
                  "internal": {
                    "certmanager": {
                      "enabled": false,
                      "secretname": "neuvector-internal"
                    },
                    "autoGenerateCert": true,
                    "autoRotateCert": true
                  },
                  "controller": {
                    "enabled": true,
                    "annotations": {},
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": "registry.connect.redhat.com/neuvector/controller@sha256:65ee3723c7c007db5ce3cd0493a785d4163d972cf957de6dec956b842807a551",
                    "imagePullPolicy": "IfNotPresent",
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "searchRegistries": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "nodePort": "",
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      }
                    },
                    "ranchersso": {
                      "enabled": false
                    },
                    "pvc": {
                      "enabled": false,
                      "existingClaim": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {
                        "userinitcfg.yaml": {
                          "users": [
                            {
                              "Fullname": "admin",
                              "Password": "",
                              "Role": "admin"
                            }
                          ]
                        }
                      }
                    },
                    "certupgrader": {
                      "env": [],
                      "schedule": "",
                      "imagePullPolicy": "IfNotPresent",
                      "timeout": 3600,
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "prime": {
                      "enabled": false,
                      "image": {
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "1.0.7"
                      }
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/enforcer@sha256:72a628a5c811c95bb267bbd0a5ffc9d74dbaca03a92aae615d43c7601b4d3b43",
                    "imagePullPolicy": "IfNotPresent",
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "env": [],
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/etcd"
                      }
                    ],
                    "resources": {},
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "manager": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/manager@sha256:89a42c9ee86f98ef56b5e695e23bdcc6334361ebd29cfb7729d7c86f04da98b0",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "",
                    "env": {
                      "ssl": true,
                      "envs": []
                    },
                    "svc": {
                      "type": "ClusterIP",
                      "nodePort": "",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": "",
                      "tls": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "topologySpreadConstraints": [],
                    "affinity": {},
                    "podLabels": {},
                    "podAnnotations": {},
                    "tolerations": [],
                    "nodeSelector": {},
                    "runAsUser": "",
                    "probes": {
                      "enabled": false,
                      "timeout": 1,
                      "periodSeconds": 10,
                      "startupFailureThreshold": 30
                    }
                  },
                  "cve": {
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "cacert": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                      "image": "registry.access.redhat.com/ubi8",
                      "imagePullPolicy": "IfNotPresent",
                      "schedule": "0 0 * * *",
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": "registry.connect.redhat.com/neuvector/scanner@sha256:b632f1d31998a4c8bad9f97216e8aa56cd7e6446bd70d2089d6c66845b964e8d",
                      "imagePullPolicy": "Always",
                      "priorityClassName": "",
                      "resources": {},
                      "topologySpreadConstraints": [],
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      },
                      "volumes": "",
                      "volumeMounts": ""
                    }
                  },
                  "resources": {},
                  "runtimePath": "",
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhooksvc": {
                    "enabled": true
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  },
                  "lease": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "true"
          containerImage: registry.connect.redhat.com/neuvector/neuvector-operator@sha256:e2f19c50ebc4127a83c943dbcfa1312ec46dec71bd36f6d99178b320d8da970b
          createdAt: "2020-12-02 14:19:21"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: Neuvector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm/tree/master/charts/core) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart.  NeuVector Operator versions are tied to NeuVector product versions. Version 1.5.2 of the NeuVector Certified Operator deploys version 5.4.6 of NeuVector.

          **Complete below steps to Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Privileged SCC is added to Service Account specified in the deploment yaml by Operator version 1.3.6 and above in new deployment. In case of upgrading NeuVector Operator from previous version to 1.3.6, please delete Privileged SCC before upgrading.

                   oc delete rolebinding -nneuvector system:openshift:scc:privileged

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s                    neuvector/default


          #Deploying the NeuVector Operator#


          Please refer to the instructions [here](https://github.com/neuvector/neuvector-operator/blob/master/README.md)


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Networking
        - Security
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - registry.connect.redhat.com/neuvector/manager@sha256:89a42c9ee86f98ef56b5e695e23bdcc6334361ebd29cfb7729d7c86f04da98b0
        - registry.connect.redhat.com/neuvector/scanner@sha256:b632f1d31998a4c8bad9f97216e8aa56cd7e6446bd70d2089d6c66845b964e8d
        - registry.access.redhat.com/ubi8@sha256:534c2c0efa4150ede18e3f9d7480d3b9ec2a52e62bc91cd54e08ee7336819619
        - registry.connect.redhat.com/neuvector/neuvector-operator@sha256:e2f19c50ebc4127a83c943dbcfa1312ec46dec71bd36f6d99178b320d8da970b
        - registry.connect.redhat.com/neuvector/controller@sha256:65ee3723c7c007db5ce3cd0493a785d4163d972cf957de6dec956b842807a551
        - registry.connect.redhat.com/neuvector/enforcer@sha256:72a628a5c811c95bb267bbd0a5ffc9d74dbaca03a92aae615d43c7601b4d3b43
        version: 1.5.2
      entries:
      - name: neuvector-operator.v1.5.2
        version: 1.5.2
      - name: neuvector-operator.v1.5.1
        version: 1.5.1
      - name: neuvector-operator.v1.5.0
        version: 1.5.0
      - name: neuvector-operator.v1.4.9
        version: 1.4.9
      - name: neuvector-operator.v1.4.8
        version: 1.4.8
      - name: neuvector-operator.v1.4.7
        version: 1.4.7
      - name: neuvector-operator.v1.4.6
        version: 1.4.6
      - name: neuvector-operator.v1.4.5
        version: 1.4.5
      - name: neuvector-operator.v1.4.4
        version: 1.4.4
      - name: neuvector-operator.v1.4.3
        version: 1.4.3
      - name: neuvector-operator.v1.4.2
        version: 1.4.2
      - name: neuvector-operator.v1.4.1
        version: 1.4.1
      - name: neuvector-operator.v1.4.0
        version: 1.4.0
      - name: neuvector-operator.v1.3.9
        version: 1.3.9
      - name: neuvector-operator.v1.3.8
        version: 1.3.8
      - name: neuvector-operator.v1.3.7
        version: 1.3.7
      - name: neuvector-operator.v1.3.6
        version: 1.3.6
      - name: neuvector-operator.v1.3.5
        version: 1.3.5
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-certified-operator
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: model-builder-for-vision-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-modelbuilder-for-vision.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "Modelbuilder",
                "metadata": {
                  "name": "modelbuilder-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "backup_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "metadata_storage": {
                    "storage_class": "ibmc-block-bronze",
                    "storage_size": "20G"
                  },
                  "photo_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "500G"
                  },
                  "in_memory_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              },
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "ModelbuilderAWS",
                "metadata": {
                  "name": "modelbuilder-aws-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "postgres": {
                    "storage_size": "20G",
                    "storage_class": "gp2",
                    "backup_storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
          createdAt: "2021-08-28T12:00:00Z"
          olm.skipRange: <1.0.3
          operators.operatorframework.io/builder: operator-sdk-v1.11.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/ibm-modelbuilder-for-vision/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Model Builder for Vision on AWS
            displayName: Deploy-Modelbuilder on AWS
            kind: ModelbuilderAWS
            name: modelbuildersaws.modelbuilder.com
            version: v1alpha1
          - description: Deploy Model Builder for Vision on IBM
            displayName: Deploy-Modelbuilder on IBM
            kind: Modelbuilder
            name: modelbuilders.modelbuilder.com
            version: v1alpha1
        description: IBM Services Software Model Builder for Vision is an AI training
          platform that can use the capabilities of IBM Cloud and Amazon Web Services
          and their GPUs to quickly train computer vision models for compatible mobile
          apps. IBM Services Software Inspection Workbench iOS and iPadOS app will
          be used as the exclusive UI for labeling and training. Specifically, IBM
          Model Builder for Vision will train computer vision models that can then
          be deployed for use on the IBM Services Software Inspector Portable , IBM
          Services Software Inspector Wearable and the IBM Maximo Visual Inspection
          Mobile apps.
        displayName: IBM Model Builder for Vision
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Model Builder
        - AI
        - Computer Vision
        - Inspection
        - Mobile
        - Core ML
        - TensorFlow Lite
        links:
        - name: IBM Model Builder for Vision
          url: http://ibm.biz/model_builder
        maintainers:
        - email: isshelp@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:2f65f46b25d84a3cb41a6a10f93786bcab47b7354ca9ac13a8e602ea833eba60
        - registry.connect.redhat.com/ibm/modelbuilder-image-server@sha256:ecdb665af651417c50335e7259b89ef8096521abec431976bd7abe4c1edc49d6
        - registry.connect.redhat.com/ibm/modelbuilder-cv-training-service@sha256:c3c9080b08b87022c784fd98ac3e2774ec89ce29a2dd19d91dee039e3fdb2b36
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d4c60531f59953f399c6a2b0f8239128cdc2640f72f2d1e9ac4d6add618f9f01
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.connect.redhat.com/ibm/modelbuilder-augmentor@sha256:ef5d0dde3eb334d919c2773af63007f1ef5e735ae040ef77fffe3440f854f068
        - registry.connect.redhat.com/ibm/modelbuilder-auth-service@sha256:f96b444fdcf7781775c9eba942bcc5e5ebdcf5f50820447f5eb26061c4c674cd
        - registry.connect.redhat.com/ibm/modelbuilder-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
        version: 1.0.3
      entries:
      - name: ibm-modelbuilder-for-vision.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: model-builder-for-vision-certified
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cluster-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-manager.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "cluster-manager"
                },
                "spec": {
                  "addOnManagerImagePullSpec": "quay.io/open-cluster-management/addon-manager:v0.13.1",
                  "deployOption": {
                    "mode": "Default"
                  },
                  "placementImagePullSpec": "quay.io/open-cluster-management/placement:v0.13.1",
                  "registrationConfiguration": {
                    "featureGates": [
                      {
                        "feature": "DefaultClusterSet",
                        "mode": "Enable"
                      }
                    ]
                  },
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "cluster-manager"
                },
                "spec": {
                  "addOnManagerImagePullSpec": "quay.io/open-cluster-management/addon-manager:v0.13.1",
                  "deployOption": {
                    "hosted": {
                      "registrationWebhookConfiguration": {
                        "address": "management-control-plane",
                        "port": 30443
                      },
                      "workWebhookConfiguration": {
                        "address": "management-control-plane",
                        "port": 31443
                      }
                    },
                    "mode": "Hosted"
                  },
                  "placementImagePullSpec": "quay.io/open-cluster-management/placement:v0.13.1",
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/registration-operator:v0.13.1
          createdAt: "2024-04-10T15:27:49Z"
          description: Manages the installation and upgrade of the ClusterManager.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management-io/ocm
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterManager defines the configurations of controllers
              running hub that govern registration and work distribution for attached
              Klusterlets
            displayName: ClusterManager
            kind: ClusterManager
            name: clustermanagers.operator.open-cluster-management.io
            version: v1
        description: "Note: This operator will stop updating, please use the latest
          operator in OperatorHub.io https://operatorhub.io/operator/cluster-manager.
          \n\nThe Cluster Manager provides the multicluster hub, which can manage
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. This operator supports the installation and upgrade of ClusterManager.\n\n##
          How to Install\nInstall ClusterManager operator by following instructions
          in top right button `Install`. After installing the operator, create an
          instance of the ClusterManager resource to install the ClusterManager.\n\nThe
          pods of ClusterManager are created in the `open-cluster-management-hub`
          namespace:\n\n```\n$ kubectl get pods -n open-cluster-management-hub\nNAME
          \                                                      READY   STATUS    RESTARTS
          \  AGE\ncluster-manager-addon-manager-controller-7fcc67bfdf-x7hbz  1/1     Running
          \  0          61m\ncluster-manager-placement-controller-dbcfd4975-26jpw
          \      1/1     Running   0          61m\ncluster-manager-registration-controller-7949d97899-wwtg4
          \  1/1     Running   0          61m\ncluster-manager-registration-webhook-7f675bc4bb-v4rhj
          \     1/1     Running   0          61m\ncluster-manager-work-webhook-6c9d65fb95-jh97b
          \             1/1     Running   0          61m\n```\n\nThere is a ManagedCluster
          created in the Hub cluster when the managed cluster is registered to the
          Hub cluster.\n\n```\n$ kubectl get managedclusters.cluster.open-cluster-management.io\nNAME
          \                  HUB ACCEPTED   MANAGED CLUSTER URLS     JOINED   AVAILABLE
          \  AGE\ncluster-dev            true                                    True
          \    True        7h20m\ncluster-test           false          https://127.0.0.1:8443
          \                       7h25m\n```\n"
        displayName: Cluster Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open-cluster-management
        - ClusterManager
        links:
        - name: Registration Operator
          url: https://github.com/open-cluster-management-io/ocm
        maintainers:
        - email: zyin@redhat.com
          name: Zhiwei Yin
        - email: deads@redhat.com
          name: David Eads
        - email: jqiu@redhat.com
          name: Jian Qiu
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/open-cluster-management/registration-operator:v0.13.1
        version: 0.13.1
      entries:
      - name: cluster-manager.v0.13.1
        version: 0.13.1
      - name: cluster-manager.v0.13.0
        version: 0.13.0
      - name: cluster-manager.v0.12.0
        version: 0.12.0
      - name: cluster-manager.v0.11.0
        version: 0.11.0
      - name: cluster-manager.v0.10.0
        version: 0.10.0
      - name: cluster-manager.v0.9.1
        version: 0.9.1
      - name: cluster-manager.v0.8.0
        version: 0.8.0
      - name: cluster-manager.v0.7.0
        version: 0.7.0
      - name: cluster-manager.v0.6.0
        version: 0.6.0
      - name: cluster-manager.v0.5.0
        version: 0.5.0
      - name: cluster-manager.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: cluster-manager
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: ScaleOps
      provider-url: https://scaleops.com
    name: scaleops-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: scaleops-operator.v1.7.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1alpha1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops-sample"
                },
                "spec": {
                  "scaleopsToken": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: ghcr.io/scaleops-sh/operator-bundle@sha256:ee627042deefa4faceb63efd56aa6f6338c90f6c13877f29d7161efe603e02f4
          createdAt: "2024-09-19T07:04:10Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: ScaleOps
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1alpha1
        description: ScaleOps Cluster Optimization
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: vlad@scaleops.com
          name: Vladislav Shub
        - email: guy@scaleops.com
          name: Guy Baron
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - ghcr.io/scaleops-sh/operator@sha256:7de0651a76d6eb63fee81e9177df93f126906d50063d7891f12c4a72c817dc86
        - ghcr.io/scaleops-sh/kube-state-metrics/ubi@sha256:d8cb65795861ce257f08726b7a3aa60a869cc44ed152f559431191bc79e98908
        - ghcr.io/scaleops-sh/operator-bundle@sha256:ee627042deefa4faceb63efd56aa6f6338c90f6c13877f29d7161efe603e02f4
        - ghcr.io/scaleops-sh/scaleops-operator-bundle@sha256:128771d92c852f7ace78103493db079a474e06d1352f06cbde7a7d4666684610
        - ghcr.io/scaleops-sh/configmap-reload/ubi@sha256:fed19f1aad43228560912abc36372a6c124d9ddf5e3e39476269dc51604d26de
        - ghcr.io/scaleops-sh/prometheus/ubi@sha256:8aeeae2a33251493d4d6e2f02a2fff1d4c2f32ab54554e6a84619ecd8b4aa69a
        version: 0.0.1
      entries:
      - name: scaleops-operator.v1.7.5
        version: 0.0.1
      name: alpha
    - currentCSV: scaleops-operator.v1.21.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops",
                  "namespace": "scaleops-system"
                },
                "spec": {
                  "scaleopsToken": "${SCALEOPS_TOKEN}"
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: registry.connect.redhat.com/scaleops-sh/operator@sha256:528f2e09d2dad4f48e01a272b51baa1753c12069515831223e3f97c7dfe03e3c
          createdAt: "2025-09-29T13:35:24Z"
          description: |-
            ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

            ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

            Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

            Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

            Contact [us](mailto:team@scaleops.com) to start your free trial today
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.21.8
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://scaleops.com
          support: ScaleOps Support support@scaleops.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This will install an instance of the ScaleOps Platform on
              your cluster
            displayName: ScaleOps Platform
            kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1
        description: |-
          ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

          ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

          Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

          Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

          Contact [us](mailto:team@scaleops.com) to start your free trial today
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        - application runtime
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: support@scaleops.com
          name: ScaleOps
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - registry.connect.redhat.com/scaleops-sh/operator@sha256:528f2e09d2dad4f48e01a272b51baa1753c12069515831223e3f97c7dfe03e3c
        - registry.connect.redhat.com/scaleops-sh/scaleops-sh@sha256:57d5d06f4f4b5470562c4f24ea5d37203633e1f4b65cd497e71d9066fd2b9e41
        - registry.connect.redhat.com/scaleops-sh/scaleops-gpu@sha256:724aef7683ea4cb245ab0a464517a6d542efa4634e5469ff35b7a7b7f00b563e
        version: 1.21.8
      entries:
      - name: scaleops-operator.v1.21.8
        version: 1.21.8
      - name: scaleops-operator.v1.21.7
        version: 1.21.7
      - name: scaleops-operator.v1.21.6
        version: 1.21.6
      - name: scaleops-operator.v1.21.5
        version: 1.21.5
      - name: scaleops-operator.v1.21.3
        version: 1.21.3
      - name: scaleops-operator.v1.21.2
        version: 1.21.2
      - name: scaleops-operator.v1.21.1
        version: 1.21.1
      - name: scaleops-operator.v1.21.0
        version: 1.21.0
      - name: scaleops-operator.v1.20.9
        version: 1.20.9
      - name: scaleops-operator.v1.20.8
        version: 1.20.8
      - name: scaleops-operator.v1.20.7
        version: 1.20.7
      - name: scaleops-operator.v1.20.6
        version: 1.20.6
      - name: scaleops-operator.v1.20.5
        version: 1.20.5
      - name: scaleops-operator.v1.20.4
        version: 1.20.4
      - name: scaleops-operator.v1.20.2
        version: 1.20.2
      - name: scaleops-operator.v1.20.1
        version: 1.20.1
      - name: scaleops-operator.v1.19.8
        version: 1.19.8
      - name: scaleops-operator.v1.19.7
        version: 1.19.7
      - name: scaleops-operator.v1.19.6
        version: 1.19.6
      - name: scaleops-operator.v1.19.5
        version: 1.19.5
      - name: scaleops-operator.v1.19.4
        version: 1.19.4
      - name: scaleops-operator.v1.19.1
        version: 1.19.1
      - name: scaleops-operator.v1.19.0
        version: 1.19.0
      - name: scaleops-operator.v1.18.8
        version: 1.18.8
      - name: scaleops-operator.v1.18.7
        version: 1.18.7
      - name: scaleops-operator.v1.18.6
        version: 1.18.6
      - name: scaleops-operator.v1.18.5
        version: 1.18.5
      - name: scaleops-operator.v1.18.4
        version: 1.18.4
      - name: scaleops-operator.v1.18.3
        version: 1.18.3
      - name: scaleops-operator.v1.18.2
        version: 1.18.2
      name: stable
    - currentCSV: scaleops-operator.v1.10.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops",
                  "namespace": "scaleops-system"
                },
                "spec": {
                  "scaleopsToken": "${SCALEOPS_TOKEN}"
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: registry.connect.redhat.com/scaleops-sh/operator@sha256:8def662a546babd6740e5417401836d647d92d52b15d559b1b3e2cebd7f77f33
          createdAt: "2025-02-12T19:06:04Z"
          description: |-
            ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

            ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

            Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

            Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

            Contact [us](mailto:team@scaleops.com) to start your free trial today
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.10.32
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://scaleops.com
          support: ScaleOps Support support@scaleops.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This will install an instance of the ScaleOps Platform on
              your cluster
            displayName: ScaleOps Platform
            kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1
        description: |-
          ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

          ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

          Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

          Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

          Contact [us](mailto:team@scaleops.com) to start your free trial today
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        - application runtime
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: support@scaleops.com
          name: ScaleOps
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - registry.connect.redhat.com/scaleops-sh/operator@sha256:8def662a546babd6740e5417401836d647d92d52b15d559b1b3e2cebd7f77f33
        - registry.connect.redhat.com/scaleops-sh/scaleops-sh@sha256:5d77e81d7959483121c208072eb61b063eb3ad9abc49e7df3c7efb8e2b77a3ae
        version: 1.10.32
      entries:
      - name: scaleops-operator.v1.10.32
        version: 1.10.32
      name: stable-1.10
    - currentCSV: scaleops-operator.v1.12.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops",
                  "namespace": "scaleops-system"
                },
                "spec": {
                  "scaleopsToken": "${SCALEOPS_TOKEN}"
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: registry.connect.redhat.com/scaleops-sh/operator@sha256:4bd74a56a8680f47aa1da7a377f74529fe7ca1568995a4e74018af0cf67763ed
          createdAt: "2025-03-11T18:10:07Z"
          description: |-
            ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

            ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

            Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

            Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

            Contact [us](mailto:team@scaleops.com) to start your free trial today
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.12.6
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://scaleops.com
          support: ScaleOps Support support@scaleops.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This will install an instance of the ScaleOps Platform on
              your cluster
            displayName: ScaleOps Platform
            kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1
        description: |-
          ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

          ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

          Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

          Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

          Contact [us](mailto:team@scaleops.com) to start your free trial today
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        - application runtime
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: support@scaleops.com
          name: ScaleOps
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - registry.connect.redhat.com/scaleops-sh/operator@sha256:4bd74a56a8680f47aa1da7a377f74529fe7ca1568995a4e74018af0cf67763ed
        - registry.connect.redhat.com/scaleops-sh/scaleops-sh@sha256:5ea4e26f2fecee3570d63adc3e2588d34f9610345d703975d4febac9c94b4bd0
        version: 1.12.6
      entries:
      - name: scaleops-operator.v1.12.6
        version: 1.12.6
      name: stable-1.12
    - currentCSV: scaleops-operator.v1.14.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops",
                  "namespace": "scaleops-system"
                },
                "spec": {
                  "scaleopsToken": "${SCALEOPS_TOKEN}"
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: registry.connect.redhat.com/scaleops-sh/operator@sha256:9267f49ce774ee4d401458531599bbd6729c6313cecc4f2bcd3bf0b5eecd92e1
          createdAt: "2025-06-08T16:15:46Z"
          description: |-
            ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

            ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

            Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

            Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

            Contact [us](mailto:team@scaleops.com) to start your free trial today
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.14.14
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://scaleops.com
          support: ScaleOps Support support@scaleops.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This will install an instance of the ScaleOps Platform on
              your cluster
            displayName: ScaleOps Platform
            kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1
        description: |-
          ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

          ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

          Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

          Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

          Contact [us](mailto:team@scaleops.com) to start your free trial today
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        - application runtime
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: support@scaleops.com
          name: ScaleOps
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - registry.connect.redhat.com/scaleops-sh/operator@sha256:9267f49ce774ee4d401458531599bbd6729c6313cecc4f2bcd3bf0b5eecd92e1
        - registry.connect.redhat.com/scaleops-sh/scaleops-sh@sha256:9c6395202ccb70e8edae3a1b540de31991ec9232b7b235dd92755588b8491720
        version: 1.14.14
      entries:
      - name: scaleops-operator.v1.14.14
        version: 1.14.14
      name: stable-1.14
    defaultChannel: stable
    packageName: scaleops-operator
    provider:
      name: ScaleOps
      url: https://scaleops.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataStax
      provider-url: ""
    name: cass-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cass-operator.v1.23.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cassandra.datastax.com/v1beta1",
                "kind": "CassandraDatacenter",
                "metadata": {
                  "name": "dc1"
                },
                "spec": {
                  "clusterName": "development",
                  "config": {
                    "cassandra-yaml": {
                      "authenticator": "PasswordAuthenticator",
                      "authorizer": "CassandraAuthorizer",
                      "num_tokens": 16,
                      "role_manager": "CassandraRoleManager"
                    },
                    "jvm-server-options": {
                      "initial_heap_size": "1G",
                      "max_heap_size": "1G"
                    }
                  },
                  "managementApiAuth": {
                    "insecure": {}
                  },
                  "podTemplateSpec": {
                    "spec": {
                      "containers": [
                        {
                          "name": "cassandra"
                        }
                      ],
                      "securityContext": {}
                    }
                  },
                  "racks": [
                    {
                      "name": "rack1"
                    },
                    {
                      "name": "rack2"
                    },
                    {
                      "name": "rack3"
                    }
                  ],
                  "resources": {
                    "requests": {
                      "cpu": "1000m",
                      "memory": "2Gi"
                    }
                  },
                  "serverType": "cassandra",
                  "serverVersion": "4.1.2",
                  "size": 3,
                  "storageConfig": {
                    "cassandraDataVolumeClaimSpec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "server-storage"
                    }
                  }
                }
              },
              {
                "apiVersion": "control.k8ssandra.io/v1alpha1",
                "kind": "CassandraTask",
                "metadata": {
                  "name": "example-cleanup"
                },
                "spec": {
                  "datacenter": {
                    "name": "dc2",
                    "namespace": "cass-operator"
                  },
                  "jobs": [
                    {
                      "command": "cleanup",
                      "name": "cleanup-run"
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database,Big Data,OpenShift Optional
          certified: "true"
          containerImage: registry.connect.redhat.com/datastax/cass-operator@sha256:6abc6253b50a0d7f6bdd4b394d2bf96320a8845952aeb8f0bbed99863192ec52
          createdAt: "2025-03-20"
          description: |
            Simple provisioning, turn-key operations, and automated remediation of Apache Cassandra clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/k8ssandra/cass-operator
          support: DataStax
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CassandraDatacenter is the Schema for the cassandradatacenters
              API
            displayName: Cassandra Datacenter
            kind: CassandraDatacenter
            name: cassandradatacenters.cassandra.datastax.com
            version: v1beta1
          - description: CassandraTask is the Schema for the cassandrajobs API
            displayName: Cassandra Task
            kind: CassandraTask
            name: cassandratasks.control.k8ssandra.io
            version: v1alpha1
        description: "## Apache Cassandra\n\nThe Apache Cassandra database is the
          right choice when you need scalability and\nhigh availability without compromising
          performance. Linear scalability and\nproven fault-tolerance on commodity
          hardware or cloud infrastructure make it the\nperfect platform for mission-critical
          data. Cassandra's support for replicating\nacross multiple datacenters is
          best-in-class, providing lower latency for your\nusers and the peace of
          mind of knowing that you can survive regional outages.\n\n## DataStax Enterprise\n\nThe
          most advanced distribution of Apache Cassandra™ on the market, with the\nenterprise
          functionality needed for serious production systems and backed up and\nsupported
          by the best distributed-experts in the world. It's one platform for\nall
          types of applications anywhere, any cloud, any model: key-value, graph,\ntabular,
          JSON.\n\nDataStax Enterprise is a fully integrated and optimized database,
          with graph,\nanalytics, and search included, all with a unified security
          model. Simply put,\nit's the only database capable of meeting today's demanding
          requirements\n\n## Operator Details\n\n`cass-operator` is designed as a
          modular operator for Apache Cassandra and\nderived  distributions. Apache
          Cassandra is a distributed database consisting of\nmultiple nodes working
          in concert to store data and process queries along a\nnumber of fault domains.
          `cass-operator` has the deployment of a Cassandra\ncluster around the logical
          domain of a datacenter with the `CassandraDatacenter`\ncustom resource.\n
          \   \nUpon submission of one of these resources it handles provisioning
          the underlying\nstateful sets (analogous to C\\* logical racks), services,
          and configuration.\nAdditionally through monitoring pod state via Kubernetes
          callbacks it handles day to day\noperations such as restarting failed processes,
          scaling clusters up, and deploying\nconfiguration changes in a rolling,
          non-disruptive, fashion.\n    \nThis operator is designed to be `Namespace`
          scoped. A single Kubernetes cluster may\nbe running multiple instances of
          this operator, in separate namespaces, to support\na number of C\\* clusters
          and environments. Configuration is simple with the usage of\nYAML based
          overrides in the Custom Resource paired with an `init` container.\n    \nIn
          C\\* clusters ordering and timing of certain operations are important to
          keep the system\nevenly distributed. `cass-operator` takes advantage of
          a sidecar process within the\nmain container to handle the orchestration
          of starting our main C* process."
        displayName: DataStax Kubernetes Operator for Apache Cassandra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - k8ssandra
        - cass-operator
        - DataStax Enterprise
        - Apache Cassandra
        - DSE
        - C*
        - Cassandra
        links:
        - name: DataStax
          url: https://datastax.com
        - name: Operator GitHub Repo
          url: https://github.com/k8ssandra/cass-operator
        - name: Config Builder GitHub Repo
          url: https://github.com/datastax/cass-config-builder
        - name: Documentation [1]
          url: https://github.com/k8ssandra/cass-operator/tree/master/docs/user
        - name: Documentation [2]
          url: https://docs.datastax.com/en/cass-operator/doc/cass-operator/cassOperatorTOC.html
        maintainers:
        - email: cass-operator@datastax.com
          name: DataStax Operator Team
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: DataStax
        relatedImages:
        - registry.connect.redhat.com/datastax/cass-operator@sha256:6abc6253b50a0d7f6bdd4b394d2bf96320a8845952aeb8f0bbed99863192ec52
        - registry.connect.redhat.com/datastax/system-logger@sha256:3408620786e6265aa835be56f85aadc48b20060b2038dc21126f614685f77c6b
        version: 1.23.2
      entries:
      - name: cass-operator.v1.23.2
        version: 1.23.2
      name: stable
    defaultChannel: stable
    packageName: cass-operator
    provider:
      name: DataStax
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
    name: aikit-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aikit-operator.v2021.2.102120
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aikit.intel/v1alpha1",
                "kind": "AIKitContainer",
                "metadata": {
                  "name": "intel-aikit-container",
                  "namespace": "redhat-ods-applications"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagestream": {
                    "name": "oneapi-aikit",
                    "namespace": "redhat-ods-applications",
                    "registry": {
                      "root": "registry.connect.redhat.com",
                      "repo": "intel",
                      "name": "oneapi-aikit",
                      "version": "latest"
                    }
                  },
                  "nameOverride": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aikit-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aikit-operator-rhmp/support?utm_source=openshift_console
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AIKitContainer
            name: aikitcontainers.aikit.intel
            version: v1alpha1
        description: |
          The Intel® oneAPI AI Analytics Toolkit gives data scientists, AI developers, and researchers familiar Python* tools and frameworks to accelerate end-to-end data science and machine learning pipelines on Intel® architectures. The components are built using oneAPI libraries for low-level compute optimizations. This toolkit maximizes performance from preprocessing through machine and deep learning phases and provides interoperability for efficient model development and deployment across single and multinodes.
        displayName: Intel® oneAPI AI Analytics Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aikit-operator
        links:
        - name: Aikit Operator
          url: https://aikit-operator.domain
        maintainers:
        - email: kam.d.kasravi@intel.com
          name: Kam Kasravi
        - email: abolfazl.shahbazi@intel.com
          name: Abolfazl Shahbazi
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Intel
          url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
        relatedImages:
        - registry.marketplace.redhat.com/rhm/intel/aikit-operator@sha256:7abb5e6164ffb3e4b87c39061f45d09848218fcfb74b48bf5e1035d56b41ee0c
        version: 2021.2.102120
      entries:
      - name: aikit-operator.v2021.2.102120
        version: 2021.2.102120
      - name: aikit-operator.v2021.2.100410
        version: 2021.2.100410
      - name: aikit-operator.v2021.2.92911
        version: 2021.2.92911
      - name: aikit-operator.v2021.2.92814
        version: 2021.2.92814
      - name: aikit-operator.v2021.2.92808-dev-a60e5ba
        version: 2021.2.92808-dev-a60e5ba
      name: alpha
    defaultChannel: alpha
    packageName: aikit-operator-rhmp
    provider:
      name: Intel
      url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Apache Software Foundation
      provider-url: ""
    name: camel-karavan-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: camel-karavan-operator.v3.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Karavan",
                "metadata": {
                  "labels": {
                    "app": "karavan"
                  },
                  "name": "karavan"
                },
                "spec": {
                  "instances": 1,
                  "auth": "public",
                  "environment": "dev",
                  "runtimes": "quarkus,spring-boot",
                  "gitPullInterval": "10s"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/apache/camel-karavan-operator:3.20.1
          repository: https://github.com/apache/camel-karavan
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Karavan
            name: karavans.camel.apache.org
            version: v1
        description: |
          Apache Camel Karavan
          ====================
          Karavan is an Integration Toolkit for Apache Camel, which makes integration easy and fun through the visualization of pipelines,
          integration with runtimes and package, image build and deploy to kubernetes out-of-the-box.

          ## Installation

          1. Install Apache Camel Karavan Operator
          2. Install Tekton operator
          3. Create namespace, ex: `karavan`
          4. Create `Secret` in the namespace, ex:
          ```
          kind: Secret
          apiVersion: v1
          metadata:
            name: karavan
            namespace: karavan
          type: Opaque
          stringData:
            master-password: karavan
            oidc-secret: secret
            oidc-server-url: https://hostname/auth/realms/karavan
            oidc-frontend-url: https://hostname/auth
            git-repository: https://github.com/repository/projects.git
            git-password: password
            git-username: username
            git-branch: main
            image-registry: image-registry.openshift-image-registry.svc:5000
          ```
          5. Create `Karavan` instance in the namespace, ex:
          ```
          apiVersion: camel.apache.org/v1
          kind: Karavan
          metadata:
            name: karavan
            namespace: karavan
          spec:
            instances: 1
            auth: public
            environment: dev
            runtimes: quarkus,spring-boot
            gitPullInterval: 10s
          ```
        displayName: Camel Karavan Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - camel
        - karavan
        - integration
        - microservices
        - low-code
        links:
        - name: Apache Camel Karavan source code repository
          url: https://github.com/apache/camel-karavan
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - ghcr.io/apache/camel-karavan-operator:3.20.1
        version: 3.20.1
      entries:
      - name: camel-karavan-operator.v3.20.1
        version: 3.20.1
      - name: camel-karavan-operator.v3.20.0
        version: 3.20.0
      - name: camel-karavan-operator.v3.18.5
        version: 3.18.5
      name: alpha
    defaultChannel: alpha
    packageName: camel-karavan-operator
    provider:
      name: The Apache Software Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: cloudnative-pg
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudnative-pg.v1.27.0-rc1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: ghcr.io/cloudnative-pg/cloudnative-pg@sha256:c2212e806e950ed5b667455f476481aef179c8e467e435352b624dc670000443
          createdAt: "2025-07-31T12:14:29Z"
          description: CloudNativePG is an open source operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/cloudnative-pg/cloudnative-pg
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.cnpg.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.cnpg.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.cnpg.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.cnpg.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.cnpg.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.cnpg.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.cnpg.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnpg` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: CloudNativePG
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnpg
        links:
        - name: CloudNativePG
          url: https://cloudnative-pg.io/
        - name: Documentation
          url: https://cloudnative-pg.io/docs/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: The CloudNativePG Contributors
        relatedImages:
        - ghcr.io/cloudnative-pg/cloudnative-pg@sha256:c2212e806e950ed5b667455f476481aef179c8e467e435352b624dc670000443
        - ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b
        - ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891
        version: 1.27.0-rc1
      entries:
      - name: cloudnative-pg.v1.27.0-rc1
        version: 1.27.0-rc1
      - name: cloudnative-pg.v1.26.0-rc3
        version: 1.26.0-rc3
      - name: cloudnative-pg.v1.26.0-rc2
        version: 1.26.0-rc2
      - name: cloudnative-pg.v1.26.0-rc1
        version: 1.26.0-rc1
      name: candidate
    - currentCSV: cloudnative-pg.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: ghcr.io/cloudnative-pg/cloudnative-pg@sha256:53995e9d6a7760c77d77e5298f78b49029427494f4dd933d5fa7a74f7af3bb44
          createdAt: "2025-08-12T14:08:35Z"
          description: CloudNativePG is an open source operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/cloudnative-pg/cloudnative-pg
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.cnpg.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.cnpg.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.cnpg.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.cnpg.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.cnpg.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.cnpg.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.cnpg.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnpg` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: CloudNativePG
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnpg
        links:
        - name: CloudNativePG
          url: https://cloudnative-pg.io/
        - name: Documentation
          url: https://cloudnative-pg.io/docs/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - ghcr.io/cloudnative-pg/cloudnative-pg@sha256:53995e9d6a7760c77d77e5298f78b49029427494f4dd933d5fa7a74f7af3bb44
        - ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b
        - ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891
        version: 1.27.0
      entries:
      - name: cloudnative-pg.v1.27.0
        version: 1.27.0
      - name: cloudnative-pg.v1.26.1
        version: 1.26.1
      - name: cloudnative-pg.v1.26.0
        version: 1.26.0
      - name: cloudnative-pg.v1.25.1
        version: 1.25.1
      - name: cloudnative-pg.v1.25.0
        version: 1.25.0
      - name: cloudnative-pg.v1.24.1
        version: 1.24.1
      - name: cloudnative-pg.v1.24.0
        version: 1.24.0
      - name: cloudnative-pg.v1.23.3
        version: 1.23.3
      - name: cloudnative-pg.v1.23.2
        version: 1.23.2
      - name: cloudnative-pg.v1.23.1
        version: 1.23.1
      name: stable-v1
    defaultChannel: stable-v1
    packageName: cloudnative-pg
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cincinnati-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: update-service-operator.v5.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "updateservice.operator.openshift.io/v1",
                "kind": "UpdateService",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          description: Creates and maintains an OpenShift Update Service instance
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-update-service
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: UpdateService is the Schema for the updateservices API.
            displayName: Update Service
            kind: UpdateService
            name: updateservices.updateservice.operator.openshift.io
            version: v1
        description: |-
          # Use Case
          Running an Update Service instance in a cluster is appealing for offline OpenShift
          clusters or for admins that want to provide their own graph data instead of
          using [https://github.com/openshift/cincinnati-graph-data](https://github.com/openshift/cincinnati-graph-data/).

          # About Update Service
          Update Service uses **Cincinnati** which is an update protocol designed to facilitate
          automatic updates. It describes a particular method for representing transitions between
          releases of a project and allowing a client to perform automatic updates between these
          releases.

          ## Components
          A **Graph Builder** iterates over the release payloads hosted by the
          storage component and builds a DAG of the releases. It is responsible for
          verifying that the graph described by the releases is acyclic and connected.

          A **Policy Engine** is in charge of altering a client's view of the graph
          by applying a set of filters which are defined within the particular Policy
          Engine instance. Both the input to and the output from Policy Engines is a
          graph, allowing multiple Policy Engines to be chained together. The first
          Policy Engine in a chain will fetch its graph from the Graph Builder and the
          last Policy Engine in a chain will serve the modified graph to the client.

          An **Update Service client** is the end consumer of the release payloads. The
          client periodically queries the Policy Engine for updates and applys them if
          available.

          # Query OpenShift's Update Service Endpoint
          $ curl --silent --header 'Accept:application/json' 'https://api.openshift.com/api/upgrades_info/v1/graph?arch=amd64&channel=stable-4.2' | jq '. as $graph | $graph.nodes | map(.version == "4.2.13") | index(true) as $orig | $graph.edges | map(select(.[0] == $orig)[1]) | map($graph.nodes[.])'
        displayName: OpenShift Update Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Update Service
        - UpdateService
        - update-service
        - OSUS
        - Cincinnati
        - version
        - upgrade
        links:
        - name: Update Service Operator
          url: https://github.com/openshift/cincinnati-operator
        - name: OpenShift Update Service
          url: https://github.com/openshift/cincinnati
        - name: OpenShift Graph Data
          url: https://github.com/openshift/cincinnati-graph-data
        maintainers:
        - email: aos-team-ota@redhat.com
          name: OpenShift Update Service maintainers
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-update-service/openshift-update-service-rhel8-operator@sha256:759f4de107a5b429933c5e44046ffb4ed75c674aea75440513a6df8fa93a83a3
        - registry.redhat.io/openshift-update-service/openshift-update-service-rhel8@sha256:a9468a662babb272917df92bc9c827535b4b37883edb440b6d2d7c32ffb2145e
        version: 5.0.3
      entries:
      - name: update-service-operator.v5.0.3
        version: 5.0.3
      - name: update-service-operator.v5.0.2
        version: 5.0.2
      - name: update-service-operator.v5.0.1
        version: 5.0.1
      - name: update-service-operator.v5.0.0
        version: 5.0.0
      - name: update-service-operator.v4.9.1
        version: 4.9.1
      - name: update-service-operator.v4.9.0
        version: 4.9.0
      name: v1
    defaultChannel: v1
    packageName: cincinnati-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: www.redhat.com
    name: dpu-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dpu-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.openshift.io/v1",
                "kind": "DpuOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dpu-operator",
                    "app.kubernetes.io/instance": "dpuoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dpuoperatorconfig",
                    "app.kubernetes.io/part-of": "dpu-operator"
                  },
                  "name": "dpuoperatorconfig-sample"
                },
                "spec": {
                  "mode": "host"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-13T13:42:44Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: dpu-operator-system
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DpuOperatorConfig is the Schema for the dpuoperatorconfigs
              API
            displayName: Dpu Operator Config
            kind: DpuOperatorConfig
            name: dpuoperatorconfigs.config.openshift.io
            version: v1
        description: This operator will manage and configure data processing unit
          (DPUs) to be used in accelerating/offloading k8s networking functions
        displayName: DPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dpu
        - dpu-operator
        - ipu
        links:
        - name: Dpu Operator
          url: https://github.com/openshift/dpu-operator
        maintainers:
        - email: wizhao@redhat.com
          name: William Zhao
        - email: bnemeth@redhat.com
          name: Balazs Nemeth
        maturity: alpha
        provider:
          name: Red Hat
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-dpu-rhel9-operator@sha256:9d8283238af545c332781b2edef6b3a2857bcf4a264c9bae26720cc48e338a4b
        - registry.redhat.io/openshift4/ose-dpu-daemon-rhel9@sha256:a9a375f70dfb4a730aca0f7e58619fb05ef226261de6980d04eeffb8a043e6a4
        version: 4.16.0-202509240123
      entries:
      - name: dpu-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: dpu-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: dpu-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: dpu-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: dpu-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: dpu-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: dpu-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: dpu-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: dpu-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: dpu-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: dpu-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: dpu-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: dpu-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: dpu-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: dpu-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: dpu-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: dpu-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: dpu-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: dpu-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: dpu-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: dpu-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: dpu-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: dpu-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: dpu-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: dpu-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: dpu-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: dpu-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: dpu-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: dpu-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      name: stable
    defaultChannel: stable
    packageName: dpu-operator
    provider:
      name: Red Hat
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: odr-cluster-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odr-cluster-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "VolumeReplicationGroup",
                "metadata": {
                  "name": "volumereplicationgroup-sample"
                },
                "spec": {
                  "async": {
                    "schedulingInterval": "10m"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ],
                        "name": "cpd"
                      },
                      {
                        "includedResources": [
                          "Deployment"
                        ],
                        "name": "deployments"
                      },
                      {
                        "excludedResources": [
                          ""
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includeClusterResources": true,
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "cpd",
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      },
                      {
                        "backupName": "deployments",
                        "includedResources": [
                          "Deployment"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "excludedResources": [
                          "ConfigMap",
                          "Secret",
                          "Deployment",
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      }
                    ]
                  },
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  },
                  "replicationState": "primary",
                  "s3Profiles": [
                    "s3-profile-of-east",
                    "s3-profile-of-west"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: MaintenanceMode
            name: maintenancemodes.ramendr.openshift.io
            version: v1alpha1
          - kind: ProtectedVolumeReplicationGroupList
            name: protectedvolumereplicationgrouplists.ramendr.openshift.io
            version: v1alpha1
          - description: VolumeReplicationGroup is the Schema for the volumereplicationgroups
              API
            displayName: Volume Replication Group
            kind: VolumeReplicationGroup
            name: volumereplicationgroups.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Cluster is a disaster-recovery orchestrator for
          stateful applications, that operates from an Advanced Cluster Management
          (ACM) managed cluster and is controlled by Openshift DR Hub operator to
          orchestrate the life-cycle of an application, and its state on the managed
          cluster.
        displayName: Openshift DR Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:769acf15d016d4992793c9c527d7c268d2a67502dd7e0fadc240e4896398195b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 4.15.20-rhodf
      entries:
      - name: odr-cluster-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odr-cluster-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odr-cluster-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odr-cluster-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odr-cluster-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odr-cluster-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odr-cluster-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odr-cluster-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odr-cluster-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odr-cluster-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odr-cluster-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odr-cluster-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odr-cluster-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odr-cluster-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odr-cluster-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odr-cluster-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odr-cluster-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odr-cluster-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odr-cluster-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odr-cluster-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odr-cluster-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "VolumeReplicationGroup",
                "metadata": {
                  "name": "volumereplicationgroup-sample"
                },
                "spec": {
                  "async": {
                    "schedulingInterval": "10m"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ],
                        "name": "cpd"
                      },
                      {
                        "includedResources": [
                          "Deployment"
                        ],
                        "name": "deployments"
                      },
                      {
                        "excludedResources": [
                          ""
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includeClusterResources": true,
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "cpd",
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      },
                      {
                        "backupName": "deployments",
                        "includedResources": [
                          "Deployment"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "excludedResources": [
                          "ConfigMap",
                          "Secret",
                          "Deployment",
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      }
                    ]
                  },
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  },
                  "replicationState": "primary",
                  "s3Profiles": [
                    "s3-profile-of-east",
                    "s3-profile-of-west"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: MaintenanceMode
            name: maintenancemodes.ramendr.openshift.io
            version: v1alpha1
          - kind: ProtectedVolumeReplicationGroupList
            name: protectedvolumereplicationgrouplists.ramendr.openshift.io
            version: v1alpha1
          - description: VolumeReplicationGroup is the Schema for the volumereplicationgroups
              API
            displayName: Volume Replication Group
            kind: VolumeReplicationGroup
            name: volumereplicationgroups.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Cluster is a disaster-recovery orchestrator for
          stateful applications, that operates from an Advanced Cluster Management
          (ACM) managed cluster and is controlled by Openshift DR Hub operator to
          orchestrate the life-cycle of an application, and its state on the managed
          cluster.
        displayName: Openshift DR Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:af262b0890300183db1766bc392017b0bda723a9ba841dc25e3d928aa413db9a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        version: 4.16.16-rhodf
      entries:
      - name: odr-cluster-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odr-cluster-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odr-cluster-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odr-cluster-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odr-cluster-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odr-cluster-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odr-cluster-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odr-cluster-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odr-cluster-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odr-cluster-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odr-cluster-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odr-cluster-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odr-cluster-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odr-cluster-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odr-cluster-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odr-cluster-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odr-cluster-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odr-cluster-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Couchbase
      provider-url: ""
    name: couchbase-enterprise-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: couchbase-operator.v2.6.4-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator@sha256:1f0a6865c360a14ba98706fd6fe8dff3bdbdede32e3622f93f19c82e377ca33f
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        version: 2.6.4-7
      entries:
      - name: couchbase-operator.v2.6.4-7
        version: 2.6.4-7
      - name: couchbase-operator.v2.6.4-6
        version: 2.6.4-6
      name: 2.6.4
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator@sha256:707571a7881bc12d5f1b0b1a94eb3c675d5c0ea0114c994a47ceb5aa79df4013
        version: 2.7.0-5
      entries:
      - name: couchbase-operator.v2.7.0-5
        version: 2.7.0-5
      name: 2.7.0
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:a79e1b6781d5f8badf664e7efea6e2889e43fae6e0d37169d766685dcd6ff394
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        version: 2.7.1-1
      entries:
      - name: couchbase-operator.v2.7.1-1
        version: 2.7.1-1
      name: 2.7.1
    - currentCSV: couchbase-operator.v2.8.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:df61928b064014d851c6e7dff6dcd63a433a27fa5623254b6e389c3eea170e36
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        version: 2.8.0-2
      entries:
      - name: couchbase-operator.v2.8.0-2
        version: 2.8.0-2
      - name: couchbase-operator.v2.8.0-1
        version: 2.8.0-1
      name: 2.8.0
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        version: 2.8.1-1
      entries:
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: 2.8.1
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        version: 2.8.1-1
      entries:
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: stable
    defaultChannel: stable
    packageName: couchbase-enterprise-certified-rhmp
    provider:
      name: Couchbase
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: topology-aware-lifecycle-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: topology-aware-lifecycle-manager.v4.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ran.openshift.io/v1alpha1",
                "kind": "ClusterGroupUpgrade",
                "metadata": {
                  "name": "ClusterGroupUpgrade-sample"
                },
                "spec": {
                  "preCaching": true
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:bf24cad2940a227f5bd2f188be5d298ff19176c634fcf609c7225d7e0a6f2fd1
          description: The topology aware lifecycle manager is an operator that facilitates
            platform upgrades of group of clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-operators",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/cluster-group-upgrades-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterGroupUpgrade is the Schema for the ClusterGroupUpgrades
              API
            displayName: Cluster Group Upgrade
            kind: ClusterGroupUpgrade
            name: clustergroupupgrades.ran.openshift.io
            version: v1alpha1
          - description: ImageBasedGroupUpgrade is the schema for upgrading a group
              of clusters using IBU
            displayName: Image-Based Group Upgrade
            kind: ImageBasedGroupUpgrade
            name: imagebasedgroupupgrades.lcm.openshift.io
            version: v1alpha1
          - description: PreCachingConfig is the Schema for the precachingconfigs
              API
            displayName: Pre-caching Config
            kind: PreCachingConfig
            name: precachingconfigs.ran.openshift.io
            version: v1alpha1
        description: |-
          Topology Aware Lifecycle Manager is an operator that facilitates
          platform and operator upgrades of group of clusters
        displayName: Topology Aware Lifecycle Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm lifecyclemanagement upgrades cluster
        links:
        - name: Cluster Group Upgrades Operator
          url: https://cluster-group-upgrades-operator.domain
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:eed8234b0fe47083ab941f513709f778bfcb80e32747b71a65fd5f62328a5ca0
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:bf24cad2940a227f5bd2f188be5d298ff19176c634fcf609c7225d7e0a6f2fd1
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-aztp-rhel9@sha256:f6a37a0de10c847a53e22406990858dbaa4cdd3be075bc047e8dcf5bc6a2f2f7
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-precache-rhel9@sha256:7673a0d7f8944b245124376f2153bf987299f5bd7350b33f0d4bd1ac18543929
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-recovery-rhel9@sha256:37e234f247c841cd12c4617f92bf8bd5a5dbcbf46fa1be042d56b966994efaad
        version: 4.16.5
      entries:
      - name: topology-aware-lifecycle-manager.v4.16.5
        version: 4.16.5
      - name: topology-aware-lifecycle-manager.v4.16.4
        version: 4.16.4
      - name: topology-aware-lifecycle-manager.v4.16.3
        version: 4.16.3
      - name: topology-aware-lifecycle-manager.v4.16.2
        version: 4.16.2
      - name: topology-aware-lifecycle-manager.v4.16.1
        version: 4.16.1
      - name: topology-aware-lifecycle-manager.v4.16.0
        version: 4.16.0
      name: "4.16"
    - currentCSV: topology-aware-lifecycle-manager.v4.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ran.openshift.io/v1alpha1",
                "kind": "ClusterGroupUpgrade",
                "metadata": {
                  "name": "ClusterGroupUpgrade-sample"
                },
                "spec": {
                  "preCaching": true
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:bf24cad2940a227f5bd2f188be5d298ff19176c634fcf609c7225d7e0a6f2fd1
          description: The topology aware lifecycle manager is an operator that facilitates
            platform upgrades of group of clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-operators",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/cluster-group-upgrades-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterGroupUpgrade is the Schema for the ClusterGroupUpgrades
              API
            displayName: Cluster Group Upgrade
            kind: ClusterGroupUpgrade
            name: clustergroupupgrades.ran.openshift.io
            version: v1alpha1
          - description: ImageBasedGroupUpgrade is the schema for upgrading a group
              of clusters using IBU
            displayName: Image-Based Group Upgrade
            kind: ImageBasedGroupUpgrade
            name: imagebasedgroupupgrades.lcm.openshift.io
            version: v1alpha1
          - description: PreCachingConfig is the Schema for the precachingconfigs
              API
            displayName: Pre-caching Config
            kind: PreCachingConfig
            name: precachingconfigs.ran.openshift.io
            version: v1alpha1
        description: |-
          Topology Aware Lifecycle Manager is an operator that facilitates
          platform and operator upgrades of group of clusters
        displayName: Topology Aware Lifecycle Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm lifecyclemanagement upgrades cluster
        links:
        - name: Cluster Group Upgrades Operator
          url: https://cluster-group-upgrades-operator.domain
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:eed8234b0fe47083ab941f513709f778bfcb80e32747b71a65fd5f62328a5ca0
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:bf24cad2940a227f5bd2f188be5d298ff19176c634fcf609c7225d7e0a6f2fd1
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-aztp-rhel9@sha256:f6a37a0de10c847a53e22406990858dbaa4cdd3be075bc047e8dcf5bc6a2f2f7
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-precache-rhel9@sha256:7673a0d7f8944b245124376f2153bf987299f5bd7350b33f0d4bd1ac18543929
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-recovery-rhel9@sha256:37e234f247c841cd12c4617f92bf8bd5a5dbcbf46fa1be042d56b966994efaad
        version: 4.16.5
      entries:
      - name: topology-aware-lifecycle-manager.v4.16.5
        version: 4.16.5
      - name: topology-aware-lifecycle-manager.v4.16.4
        version: 4.16.4
      - name: topology-aware-lifecycle-manager.v4.16.3
        version: 4.16.3
      - name: topology-aware-lifecycle-manager.v4.16.2
        version: 4.16.2
      - name: topology-aware-lifecycle-manager.v4.16.1
        version: 4.16.1
      - name: topology-aware-lifecycle-manager.v4.16.0
        version: 4.16.0
      name: stable
    defaultChannel: stable
    packageName: topology-aware-lifecycle-manager
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Korrel8r
      provider-url: https://github.com/korrel8r/korrel8r
    name: korrel8r
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: korrel8r.v0.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "korrel8r.openshift.io/v1alpha1",
                "kind": "Korrel8r",
                "metadata": {
                  "name": "korrel8r",
                  "namespace": "korrel8r"
                },
                "spec": {
                  "debug": {
                    "verbose": 1
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional,Monitoring,Logging & Tracing
          containerImage: quay.io/korrel8r/operator:0.1.8
          createdAt: "2024-09-27T17:45:00Z"
          description: Correlate cluster resources and observability data.
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "korrel8r.openshift.io/v1alpha1",
              "kind": "Korrel8r",
              "metadata": {
                "name": "korrel8r",
                "namespace": "korrel8r"
              }
            }
          operatorframework.io/suggested-namespace: korrel8r
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/korrel8r/korrel8r
          support: Korrel8r Project
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Korrel8r is a service that correlates observabililty signals
              in the cluster.
            displayName: Korrel8r
            kind: Korrel8r
            name: korrel8rs.korrel8r.openshift.io
            version: v1alpha1
        description: |
          Korrel8r helps navigate from problem symptoms to related resources and signal data that can reveal the cause.
          It can follow relationships between disjointed observability "silos" (logs, metrics, alerts and more) to bring together all the data available to solve a problem.
          This operator allows Korrel8r to be deployed as a cluster service with a REST API.
          It can be used as a back-end for data correlation by the OpenShift console and other tools.
        displayName: Korrel8r
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - signal
        - correlation
        links:
        - name: Documentation
          url: https://korrel8r.github.io/korrel8r
        - name: Source Code and Issues
          url: https://github.com/korrel8r/korrel8r
        maintainers:
        - email: korrel8r@gmail.com
          name: Korrel8r Project
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Korrel8r
          url: https://github.com/korrel8r/korrel8r
        relatedImages:
        - quay.io/korrel8r/operator:0.1.8
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 0.1.8
      entries:
      - name: korrel8r.v0.1.8
        version: 0.1.8
      - name: korrel8r.v0.1.7
        version: 0.1.7
      - name: korrel8r.v0.1.5
        version: 0.1.5
      - name: korrel8r.v0.1.3
        version: 0.1.3
      - name: korrel8r.v0.1.2
        version: 0.1.2
      - name: korrel8r.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: korrel8r
    provider:
      name: Korrel8r
      url: https://github.com/korrel8r/korrel8r
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Upbound Inc.
      provider-url: https://upbound.io
    name: universal-crossplane
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: universal-crossplane.1.5.1-up.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"pkg.crossplane.io/v1",
                "kind":"Configuration",
                "metadata":{
                  "name":"xp-getting-started-with-aws"
                },
                "spec":{
                  "package":"registry.upbound.io/xp/getting-started-with-aws:v1.2.2"
                }
              },
              {
                "apiVersion":"apiextensions.crossplane.io/v1",
                "kind":"CompositeResourceDefinition",
                "metadata":{
                  "name":"compositepostgresqlinstances.database.example.org"
                },
                "spec":{
                  "group":"database.example.org",
                  "names":{
                    "kind":"CompositePostgreSQLInstance",
                    "plural":"compositepostgresqlinstances"
                  },
                  "claimNames":{
                    "kind":"PostgreSQLInstance",
                    "plural":"postgresqlinstances"
                  },
                  "connectionSecretKeys":[
                      "username",
                      "password",
                      "endpoint",
                      "port"
                  ],
                  "versions":[
                    {
                      "name":"v1alpha1",
                      "served":true,
                      "referenceable":true,
                      "schema":{
                        "openAPIV3Schema":{
                          "type":"object",
                          "properties":{
                            "spec":{
                              "type":"object",
                              "properties":{
                                "parameters":{
                                  "type":"object",
                                  "properties":{
                                    "storageGB":{
                                      "type":"integer"
                                    }
                                  },
                                  "required":[
                                      "storageGB"
                                  ]
                                }
                              },
                              "required":[
                                  "parameters"
                              ]
                            }
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion":"apiextensions.crossplane.io/v1",
                "kind":"Composition",
                "metadata":{
                  "name":"compositepostgresqlinstances.aws.database.example.org",
                  "labels":{
                    "provider":"aws",
                    "guide":"quickstart",
                    "vpc":"default"
                  }
                },
                "spec":{
                  "writeConnectionSecretsToNamespace":"crossplane-system",
                  "compositeTypeRef":{
                    "apiVersion":"database.example.org/v1alpha1",
                    "kind":"CompositePostgreSQLInstance"
                  },
                  "resources":[
                    {
                      "name":"rdsinstance",
                      "base":{
                        "apiVersion":"database.aws.crossplane.io/v1beta1",
                        "kind":"RDSInstance",
                        "spec":{
                          "forProvider":{
                            "region":"us-east-1",
                            "dbInstanceClass":"db.t2.small",
                            "masterUsername":"masteruser",
                            "engine":"postgres",
                            "engineVersion":"9.6",
                            "skipFinalSnapshotBeforeDeletion":true,
                            "publiclyAccessible":true
                          },
                          "writeConnectionSecretToRef":{
                            "namespace":"crossplane-system"
                          }
                        }
                      },
                      "patches":[
                        {
                          "fromFieldPath":"metadata.uid",
                          "toFieldPath":"spec.writeConnectionSecretToRef.name",
                          "transforms":[
                            {
                              "type":"string",
                              "string":{
                                "fmt":"%s-postgresql"
                              }
                            }
                          ]
                        },
                        {
                          "fromFieldPath":"spec.parameters.storageGB",
                          "toFieldPath":"spec.forProvider.allocatedStorage"
                        }
                      ],
                      "connectionDetails":[
                        {
                          "fromConnectionSecretKey":"username"
                        },
                        {
                          "fromConnectionSecretKey":"password"
                        },
                        {
                          "fromConnectionSecretKey":"endpoint"
                        },
                        {
                          "fromConnectionSecretKey":"port"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion":"pkg.crossplane.io/v1",
                "kind":"Provider",
                "metadata":{
                  "name":"provider-aws"
                },
                "spec":{
                  "package":"crossplane/provider-aws:master"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime, Cloud Provider, Database, Integration &
            Delivery, Networking, Storage
          com.redhat.component: universal-crossplane
          com.redhat.delivery.backport: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.6
          containerImage: upbound/uxp-bootstrapper
          createdAt: 2021-05-18T00:00:00+0000
          description: Upbound Universal Crossplane (UXP) is Upbound's official enterprise-grade
            distribution of Crossplane.
          io.k8s.display-name: universal-crossplane
          io.openshift.tags: uxp,crossplane,upbound
          maintainer: Upbound Inc. <info@upbound.io>
          name: universal-crossplane
          ocs.tags: v4.6
          operatorframework.io/initialization-resource: |
            apiVersion: v1
            kind: Secret
            metadata:
              name: upbound-control-plane-token
              namespace: upbound-system
            type: Opaque
            data:
              token: "Output of `up cloud controlplane attach | base64` command"
          operatorframework.io/suggested-namespace: upbound-system
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io.bundle.channels.v1: stable
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: universal-crossplane
          operators.operatorframework.io/internal-objects: '["locks.pkg.crossplane.io"]'
          repository: https://github.com/upbound/universal-crossplane
          summary: Upbound Universal Crossplane (UXP) is Upbound's official enterprise-grade
            distribution of Crossplane.
          support: Upbound Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: CompositeResourceDefinition
            kind: CompositeResourceDefinition
            name: compositeresourcedefinitions.apiextensions.crossplane.io
            version: v1
          - displayName: CompositionRevision
            kind: CompositionRevision
            name: compositionrevisions.apiextensions.crossplane.io
            version: v1alpha1
          - displayName: Composition
            kind: Composition
            name: compositions.apiextensions.crossplane.io
            version: v1
          - displayName: ConfigurationRevision
            kind: ConfigurationRevision
            name: configurationrevisions.pkg.crossplane.io
            version: v1
          - displayName: Configuration
            kind: Configuration
            name: configurations.pkg.crossplane.io
            version: v1
          - displayName: ControllerConfig
            kind: ControllerConfig
            name: controllerconfigs.pkg.crossplane.io
            version: v1alpha1
          - displayName: Lock
            kind: Lock
            name: locks.pkg.crossplane.io
            version: v1alpha1
          - displayName: ProviderRevision
            kind: ProviderRevision
            name: providerrevisions.pkg.crossplane.io
            version: v1
          - displayName: Provider
            kind: Provider
            name: providers.pkg.crossplane.io
            version: v1
        description: |
          ## Overview

          Upbound Universal Crossplane (UXP) is [Upbound's][upbound] official
          enterprise-grade distribution of [Crossplane][crossplane]. It's fully compatible
          with upstream Crossplane, open source, capable of connecting to Upbound Cloud
          for real-time dashboard visibility, and maintained by Upbound. It's the easiest
          way for both individual community members and enterprises to build their
          production control planes.

          ## Quick Start

          1. Install the [Upbound CLI][upbound-cli].
              ```console
              curl -sL https://cli.upbound.io | sh
              # Install with Homebrew
              brew install upbound/tap/up
              ```

          2. [Create an Upbound account][create-account] for a free dashboard for UXP.

          3. Connect UXP to Upbound Cloud.
              ```console
              # The name of your new UXP control plane.
              UXP_NAME=mycrossplane

              up cloud login
              up cloud controlplane attach ${UXP_NAME} | up uxp connect -
              ```

          4. Manage your UXP control plane by [signing in][login] to your Upbound account.

          ## Additional Resources

          - The [UXP Documentation][uxp-documentation] provides additional information
            about UXP and resources for developers, like examples.
          - The [developer guide][developer-guide] describes how to build and run UXP
            locally from source.
          - [UXP Slack][uxp-slack] is where you can go to get all of your UXP questions
            answered.

          [upbound]: https://upbound.io
          [crossplane]: https://crossplane.io/
          [upbound-cli]: https://github.com/upbound/up
          [create-account]: https://cloud.upbound.io/register
          [login]: https://cloud.upbound.io/login
          [uxp-documentation]: https://cloud.upbound.io/docs/uxp
          [developer-guide]: https://github.com/upbound/universal-crossplane/blob/main/docs/developer-guide.md
          [uxp-slack]: https://crossplane.slack.com/archives/upbound
        displayName: Upbound Universal Crossplane (UXP)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - infrastructure
        - services
        - application
        - database
        - cache
        - bucket
        - infra
        - app
        - ops
        - oam
        - gcp
        - azure
        - aws
        - alibaba
        - cloudsql
        - rds
        - s3
        - azuredatabase
        - asparadb
        - gke
        - aks
        - eks
        links:
        - name: GitHub
          url: https://github.com/upbound/universal-crossplane
        - name: Website
          url: https://upbound.io
        - name: Twitter
          url: https://twitter.com/upbound_io
        - name: Slack
          url: https://slack.crossplane.io/
        maintainers:
        - email: info@upbound.io
          name: Upbound Inc.
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Upbound Inc.
          url: https://upbound.io
        relatedImages:
        - upbound/uxp-bootstrapper:v1.5.1-up.1
        - upbound/crossplane:v1.5.1-up.1
        - upbound/xgql:v0.1.5
        version: 1.5.1-up.1
      entries:
      - name: universal-crossplane.1.5.1-up.1
        version: 1.5.1-up.1
      - name: universal-crossplane.1.4.3-up.1
        version: 1.4.3-up.1
      - name: universal-crossplane.1.3.3-up.1
        version: 1.3.3-up.1
      - name: universal-crossplane.1.3.1-up.1
        version: 1.3.1-up.1
      - name: universal-crossplane.1.2.3-up.1
        version: 1.2.3-up.1
      - name: universal-crossplane.1.2.2-up.1
        version: 1.2.2-up.1
      - name: universal-crossplane.1.2.1-up.4
        version: 1.2.1-up.4
      - name: universal-crossplane.1.2.1-up.3
        version: 1.2.1-up.3
      name: stable
    defaultChannel: stable
    packageName: universal-crossplane
    provider:
      name: Upbound Inc.
      url: https://upbound.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Palo Alto Networks
      provider-url: ""
    name: pcc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pcc-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "name": "pcc-console",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "ConsoleDefender",
                "metadata": {
                  "name": "pcc-consoledefender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "defenderConfig": {
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Defender",
                "metadata": {
                  "name": "pcc-defender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "defenderConfig": {
                    "clusterAddress": "twistlock-console.example.com",
                    "consoleAddress": "https://twistlock-console.example.com:8083",
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: quay.io/prismacloud/pcc-operator@sha256:51a343d5ea0115e3c43ccaa546492e85ab74cf2cf2658d725c1ea7099fc89191
          description: Deploy Prisma Cloud Compute for cloud-native security in your
            clusters
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator
          support: pcc-integrations@paloaltonetworks.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Install the Prisma Cloud Compute Console and Defenders
            displayName: Console and Defenders
            kind: ConsoleDefender
            name: consoledefenders.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install the Prisma Cloud Compute Console
            displayName: Console
            kind: Console
            name: consoles.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install Prisma Cloud Compute Defenders
            displayName: Defenders
            kind: Defender
            name: defenders.pcc.paloaltonetworks.com
            version: v1alpha1
        description: |
          ## Features
          - Deploy Console
              - Create initial local administrator user
              - Add license
          - Deploy Defenders
          - Support for offline deployments
          - Upgrade Console
          - Upgrade Defenders

          ## Installation
          See the [Prisma Cloud Compute Operator documentation](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator) for installation guides.

          ## Support
          This operator is available "as is," and should be seen as community-supported; however, Palo Alto Networks will continue contributing whenever possible.
          Palo Alto Networks does not provide technical support or help with using or troubleshooting the operator through normal support channels.
          We encourage you to open GitHub [issues](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/issues) and [pull requests](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/pulls) to track bugs and feature requests.

          ## Other links
          [Prisma Cloud Compute product documentation](https://docs.paloaltonetworks.com/prisma/prisma-cloud/prisma-cloud-admin-compute.html)

          [Prisma Cloud Compute API documentation](https://prisma.pan.dev/api/cloud/cwpp)
        displayName: Prisma Cloud Compute Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - palo
        - palo alto
        - palo alto networks
        - paloalto
        - twistlock
        - compute
        - prisma
        - prisma cloud
        - prisma cloud compute
        - security
        - image
        - images
        - container
        - containers
        - docker
        - podman
        - vulnerability
        - vulnerabilities
        - compliance
        - runtime
        - scan
        - scanning
        - monitor
        - monitoring
        - alert
        - alerting
        links:
        - name: Prisma Cloud
          url: https://www.paloaltonetworks.com/prisma/cloud
        maintainers:
        - email: pcc-integrations@paloaltonetworks.com
          name: Wyatt Gill
        maturity: alpha
        provider:
          name: Palo Alto Networks
        relatedImages:
        - quay.io/prismacloud/pcc-operator@sha256:51a343d5ea0115e3c43ccaa546492e85ab74cf2cf2658d725c1ea7099fc89191
        version: 0.2.3
      entries:
      - name: pcc-operator.v0.2.3
        version: 0.2.3
      name: stable
    defaultChannel: stable
    packageName: pcc-operator
    provider:
      name: Palo Alto Networks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-iam-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-iam-operator.v4.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "Authentication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-iam-operator",
                    "app.kubernetes.io/managed-by": "ibm-iam-operator",
                    "app.kubernetes.io/name": "ibm-iam-operator"
                  },
                  "name": "example-authentication"
                },
                "spec": {
                  "auditService": {
                    "imageName": "auditService no longer used - ignore",
                    "imageRegistry": "auditService no longer used - ignore",
                    "imageTag": "auditService no longer used - ignore"
                  },
                  "authService": {
                    "imageName": "icp-platform-auth",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "ldapsCACert": "platform-auth-ldaps-ca-cert",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "650Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "400Mi",
                        "memory": "350Mi"
                      }
                    },
                    "routerCertSecret": "saml-auth-secret"
                  },
                  "clientRegistration": {
                    "imageName": "im-initcontainer",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "158Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "config": {
                    "authUniqueHosts": "internal-ip1 internal-ip2 mycluster.icp",
                    "bootstrapUserId": "kubeadmin",
                    "claimsMap": "name=\"givenName\" family_name=\"givenName\" given_name=\"givenName\" preferred_username=\"displayName\" display_name=\"displayName\"",
                    "claimsSupported": "name,family_name,display_name,given_name,preferred_username",
                    "clusterCADomain": "mycluster.icp",
                    "clusterExternalAddress": "10.0.0.1",
                    "clusterInternalAddress": "10.0.0.1",
                    "clusterName": "mycluster",
                    "defaultAdminPassword": "password",
                    "defaultAdminUser": "cpadmin",
                    "defaultLogin": "",
                    "enableImpersonation": false,
                    "fipsEnabled": true,
                    "ibmCloudSaas": false,
                    "icpPort": 8443,
                    "installType": "fresh",
                    "isOpenshiftEnv": true,
                    "nonceEnabled": true,
                    "oidcIssuerURL": "https://127.0.0.1:443/idauth/oidc/endpoint/OP",
                    "openshiftPort": 443,
                    "preferredLogin": "",
                    "providerIssuerURL": "",
                    "roksEnabled": true,
                    "roksURL": "https://roks.domain.name:443",
                    "roksUserPrefix": "changeme",
                    "saasClientRedirectUrl": "",
                    "scimAdminPassword": "scimpassword",
                    "scimAdminUser": "scimadmin",
                    "scopeClaim": "profile=\"name,family_name,display_name,given_name,preferred_username\"",
                    "wlpClientID": "4444be3a738841016ab76d71b650e836",
                    "wlpClientRegistrationSecret": "f1362ca4d20b8389af2d1ea68042c9af",
                    "wlpClientSecret": "aa73bf39752053bf723d1143fb4cf8a2",
                    "xframeDomain": "",
                    "zenFrontDoor": false
                  },
                  "identityManager": {
                    "imageName": "icp-identity-manager",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "masterNodesList": "10.0.0.1",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "550Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "ephemeral-storage": "300Mi",
                        "memory": "150Mi"
                      }
                    }
                  },
                  "identityProvider": {
                    "imageName": "icp-identity-provider",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "550Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "ephemeral-storage": "300Mi",
                        "memory": "150Mi"
                      }
                    }
                  },
                  "initMongodb": {
                    "imageName": "im-initcontainer",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "178Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "operatorVersion": "0.14.1",
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
          createdAt: "2025-02-20T00:03:25Z"
          description: The IAM operator provides a simple Kubernetes CRD-Based API
            to manage the lifecycle of IAM services. With this operator, you can simply
            deploy and upgrade the IAM services
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.11.0
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-iam-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Authentication is the Schema for the authentications API
            displayName: Authentication
            kind: Authentication
            name: authentications.operator.ibm.com
            version: v1alpha1
          - description: Client is the Schema for the clients API
            displayName: Client
            kind: Client
            name: clients.oidc.security.ibm.com
            version: v1
        description: |-
          The `ibm-iam-operator` installs the IBM Cloud Platform Common Services Identity and access management (IAM) service.

          **Important:** Do not install this operator directly. Install this operator only by using the IBM Common Service Operator. For more information about installing this operator and other Common Services operators, see [Installer documentation](http://ibm.biz/cpcs_opinstall). Additionally, you can exit this panel and navigate to the IBM Common Services tile in OperatorHub to learn more about the operator. If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For more information about IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).

          You can use the `ibm-iam-operator` to install the authentication and authorization services for the IBM Cloud Platform Common Services.

          With these services, you can configure security for IBM Cloud Platform Common Services, IBM Certified Containers (IBM products), or IBM Cloud Paks that are installed.

          For more information about the available IBM Cloud Platform Common Services, see the [IBM Knowledge Center](http://ibm.biz/cpcsdocs).

          ## Supported platforms

           - Red Hat OpenShift Container Platform 4.3 or newer installed on one of the following platforms:

              - Linux x86_64
              - Linux on Power (ppc64le)
              - Linux on IBM Z and LinuxONE

          ## Prerequisites

          Before you install this operator, you need to first install the operator dependencies and prerequisites:

          - For the list of operator dependencies, see the IBM Knowledge Center [Common Services dependencies documentation](http://ibm.biz/cpcs_opdependencies).

          - For the list of prerequisites for installing the operator, see the IBM Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).

          ## Documentation

          To install the operator by using the IBM Common Services Operator, follow the installation and configuration instructions that are in the IBM Knowledge Center.

          - If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).
          - If you are using the operator with an IBM Containerized Software, see the IBM Cloud Platform Common Services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall).
        displayName: IBM IM Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:b8886b9b54e6ee4c6b2981ef092e54348a0286577d9924cf136cee881f5bf1f4
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-bundle@sha256:98cf2a6c7b7af5d7cbfbc61ce140201272b30b12edabaec2b2d4287b0690d89e
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:f289ac0d44803ff12d41c3659e6fdf04c404e0b3b233b6bc44eec671abb1b47e
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:52ce08ac2bce7c02dc30e2c0aa74df4bde8f156debccc59960f02b4b0040ffea
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:97b9ecef8caef156a0eb073b6eb2a34b63209b44511f5694a243f11285d88662
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:0f46b288d060a15a48ecad8f328a8046bf232b8629c3d79da6a82ea712b363f4
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:7cec55c00827b747f23c2f1703d99e1ca5b30a0c445f8ad0471b25b0c9cca57c
        version: 4.11.0
      entries:
      - name: ibm-iam-operator.v4.11.0
        version: 4.11.0
      name: v4.10
    defaultChannel: v4.10
    packageName: ibm-iam-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: vfunction-operator
      app.kubernetes.io/managed-by: operator
      app.kubernetes.io/name: vfunction
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      name: vfunction-server-operator
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: vFunction
      provider-url: ""
    name: vfunction-server-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vfunction-server-operator.v4.3.1762
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vfunction.com/v1",
                "kind": "VfunctionServer",
                "metadata": {
                  "name": "vfunction-server",
                  "namespace": "vfunction",
                  "labels": {
                      "name": "vfunction",
                      "app.kubernetes.io/name": "vfunction",
                      "app.kubernetes.io/instance": "vfunction-server",
                      "app.kubernetes.io/managed-by": "operator"
                  }
                },
                "spec": {
                  "admin": {
                    "email": "admin@mycompany.com",
                    "name": "Admin",
                    "password": "Password1!"
                  },
                  "host": "http://my.domain.com",
                  "org_name": "MyCompany",
                  "upgrade": "Never",
                  "measurements": {
                    "min_number_of_services": "2",
                    "max_number_of_services": "10",
                    "max_pod_memory_capacity": "16G",
                    "max_pod_cpu_capacity": "1",
                    "pod_ephemeral_storage": "1G",
                    "S3": {
                      "bucket": "",
                      "key": "",
                      "secret": "",
                      "region": ""
                    }
                  },
                  "smtp": {
                    "identity": "",
                    "password": "",
                    "url": "",
                    "user": ""
                  },
                  "authentication": {
                    "authority": "",
                    "client_id": "",
                    "client_secret": "",
                    "jwks_url": "",
                    "issuer": "",
                    "ca_root_crt": ""
                  },
                  "tls": {
                    "use_letsencrypt": "No",
                    "crt": "",
                    "key": ""
                  },
                  "nginx": {
                    "force_http": "No",
                    "ipv6_disabled": "No",
                    "service_type": ""
                  },
                  "route": {
                    "use_route": "Yes",
                    "tls_termination": "edge",
                    "use_tls_certificate": "No"
                  },
                  "mysql": {
                    "external_mysql_ip": "",
                    "external_mysql_user": "",
                    "external_mysql_password": "",
                    "max_pod_memory_capacity": "4G",
                    "max_pod_cpu_capacity": "1",
                    "pod_ephemeral_storage": "1G"
                  },
                  "storage": {
                    "storage_class": "",
                    "mysql_persistent_volume": "",
                    "mysql_persistent_volume_claim": "",
                    "mysql_persistent_volume_claim_access_mode": "ReadWriteOnce",
                    "storage_persistent_volume": "",
                    "storage_persistent_volume_claim": "",
                    "storage_persistent_volume_claim_access_mode": "ReadWriteOnce",
                    "lets_encrypt_persistent_volume": "",
                    "lets_encrypt_persistent_volume_claim": "",
                    "lets_encrypt_persistent_volume_claim_access_mode": "ReadWriteOnce"
                  },
                  "distributed": {
                    "otlp": {
                      "min_number_of_services": "1",
                      "max_number_of_services": "10",
                      "max_pod_memory_capacity": "8G",
                      "max_pod_cpu_capacity": "1",
                      "pod_ephemeral_storage": "250M",
                      "max_folder_size_in_mb": "200"
                    },
                    "dm": {
                      "min_number_of_services": "1",
                      "max_number_of_services": "10",
                      "max_pod_memory_capacity": "8G",
                      "max_pod_cpu_capacity": "1",
                      "pod_ephemeral_storage": "1G"
                    }
                  },
                  "backup": {
                    "mysql": {
                      "restore_during_installation": "No",
                      "S3": {
                        "bucket": "",
                        "key": "",
                        "secret": "",
                        "region": ""
                      }
                    }
                  },
                  "security": {
                    "image_pull_policy": "",
                    "force_hashed_images": "No",
                    "set_pod_resources": "Yes",
                    "proxy": {
                      "http_proxy": "",
                      "https_proxy": "",
                      "additional_no_proxy": ""
                    },
                    "use_rate_limiting": "No",
                    "requests_per_second": "10"
                  },
                  "affinity": {
                    "node_affinity": [
                      {
                        "key": "",
                        "operator": "",
                        "values": "",
                        "services": ""
                      }
                    ],
                    "node_selectors": [
                      {
                        "key": "",
                        "value": "",
                        "services": ""
                      }
                    ],
                    "additional_labels": [
                      {
                        "key": "",
                        "value": "",
                        "services": ""
                      }
                    ],
                    "additional_annotations": [
                      {
                        "key": "",
                        "value": "",
                        "services": ""
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, AI/Machine Learning
          certified: Red Hat certified
          containerImage: registry.connect.redhat.com/vfunction/vfunction-server-operator@sha256:a5a0f47f7fb9219139f8d749da14950a8c39698b340644011485dfbec54ebe8b
          createdAt: "2025-08-04T16:10:32Z"
          description: vFunction is a cutting-edge code analysis, machine learning,
            and automation to boost your Java modernization projects.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.3.1762
          support: vFunction
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of vFunction running on your cluster.
            displayName: vFunction
            kind: VfunctionServer
            name: vfunctionservers.vfunction.com
            version: v1
        description: |2-

          ## Introduction

          [vFunction](https://www.vfunction.com/) is a cloud-native modernization platform that combines dynamic and static code analysis, machine learning, and automation to automatically identify and extract services from existing applications. vFunction is the only platform purpose-built for modernization of Java applications.

          The vFunction modernization process starts by learning the running monolithic application, and surfacing the interdependencies within it. Using AI, the platform analyzes and identifies services that can be separated from the application. This decomposition can present a range of micro, mini, or even macro services, depending on your application environment, each being an independently deployable and scalable application component.

          vFunction automates the extraction of these services, enabling you to modernize your monolith, quickly and easily.

          ## The vFunction platform

          The platform consists of 3 basic components; the server, the controller package, and a tools package. The server runs as an operand on an OpenShift environment. The controller package is installed on the machine that runs the monolithic application which can be either a Linux or a Windows machine, and the tools are run on a development machine, with access to the code of the monolithic application.

          The controller package consists of three elements: the vFunction agent, that collects data during the dynamic analysis phase; the vFunction Viper application, that performs static analysis on the binaries of the application; and the vFunction controller that handles all the communication between the agent, Viper, and the vFunction server.

          The vFunction agent is a mix of a Java and native agent, and needs to run on the JVM that is currently running your application. [Refer to the vFunction OOB Support Matrix document](https://drive.google.com/file/d/1ccq8LFab1FrYAimDUxwgjiCOdk4QuVzs/view) for a list of supported application servers and JVMs.

          This operator installs a vFunction server instance to be connected later with one or more vFunction controllers installed on your application machine(s).

          ## Before You Start

          The vFunction operator requires a Red Hat OpenShift Kubernetes Application Platform 4.x.

          The cluster recommended configuration (which is the same as the minimal one) should consist of:

          - *1 master node*

          - *1 worker node*

          - *Storage capacity enough for dynamically provision 2 PVs of 50G each*

          - *A default StorageClass configured*

          vFunction doesn’t provide any encryption mechanism for data stored on any attached storage. If required, encryption can be achieved by the client by using its own external means on the storage itself.

          vFunction supports backup of all critical data as part of the application. Nevertheless, for complete backup of entire data it is recommended that the client apply its own storage based backup mechanism.

          ## Install Prerequisites

          ###Cluster:
          The operator was certified and tested on OCP 4.6 - 4.17.
          If you encounter any issue with other OCP versions or cloud providers please contact info@vfunction.com.

          ###Storage:
          The operator creates two new PersistentVolumeClaims (PVCs) during installation time. For their creation, it relies on the default StorageClass to dynamically provision two PersistentVolumes (PVs). The system must have a default StorageClass in place.
          Both PersistentVolumes are defaulted to access with ReadWriteOnce mode, but you can change this by specifying a different access mode in the YAML file.
          Notes:
          - There is an option to specify a specific StorageClass for the PersistentVolumeClaims.
          - There is an option to specify the use of a specific pre-created PersistentVolume for each PersistentVolumeClaim
          - There is an option to specify a pre-created PersistentVolumeClaim for each Deployment.
          - There is an option to change the access mode for each PersistentVolumeClaim.

          ###Project:
          The operator should be installed in a new and dedicated project (namespace). If you intend to install more than one vFunction operands in the same cluster, please use different projects for each one. Installation of more than one operand on the same cluster will allow scaling by load balancing multiple applications to different vFunction servers.

          ## Installation instructions

          1. Prepare the YAML file to use in the installation:

              a. Copy the following YAML template into a text editor:

              ```yaml
                  apiVersion: vfunction.com/v1
                  kind: VfunctionServer
                  metadata:
                    name: vfunction
                    namespace: vfunction
                    labels:
                      name: vfunction
                      app.kubernetes.io/name: vfunction
                      app.kubernetes.io/instance: vfunction-server
                      app.kubernetes.io/managed-by: operator
                  spec:
                    host: "http://my.domain.com"
                    org_name: "MyCompany"
                    upgrade: "Never"
                    admin:
                      email: "admin@mycompany.com"
                      name: "Admin"
                      password: "Password1!"
                    measurements:
                      min_number_of_services: "2"
                      max_number_of_services: "10"
                      max_pod_memory_capacity: "16G"
                      max_pod_cpu_capacity: "1"
                      S3:
                        bucket: ""
                        key: ""
                        secret: ""
                    smtp:
                      password: ""
                      url: ""
                      identity: ""
                      user: ""
                    authentication:
                      authority: ""
                      client_id: ""
                      client_secret: ""
                      jwks_url: ""
                      issuer: ""
                      ca_root_crt: |
                        -----BEGIN CERTIFICATE-----
                        ...
                        -----END CERTIFICATE-----
                    tls:
                      use_letsencrypt: "No"
                      crt: |
                        -----BEGIN CERTIFICATE-----
                        ...
                        -----END CERTIFICATE-----
                      key: |
                        -----BEGIN PRIVATE KEY-----
                        ...
                        -----END PRIVATE KEY-----
                    nginx:
                      force_http: "No"
                      ipv6_disabled: "No"
                      service_type: ""
                    route:
                      use_route: "Yes"
                      tls_termination: "edge"
                      use_tls_certificate: "No"
                    mysql:
                      external_mysql_ip: ""
                      external_mysql_user: ""
                      external_mysql_password: ""
                      max_pod_memory_capacity: "4G"
                      max_pod_cpu_capacity: "1"
                    storage:
                      storage_class: ""
                      mysql_persistent_volume: ""
                      mysql_persistent_volume_claim: ""
                      mysql_persistent_volume_claim_access_mode: "ReadWriteOnce"
                      storage_persistent_volume: ""
                      storage_persistent_volume_claim: ""
                      storage_persistent_volume_claim_access_mode: "ReadWriteOnce"
                      lets_encrypt_persistent_volume: ""
                      lets_encrypt_persistent_volume_claim: ""
                      lets_encrypt_persistent_volume_claim_access_mode: "ReadWriteOnce"
                    distributed:
                        otlp:
                          min_number_of_services: "1"
                          max_number_of_services: "10"
                          max_pod_memory_capacity: "8G"
                          max_pod_cpu_capacity: "1"
                          max_folder_size_in_mb: "200"
                        dm:
                          min_number_of_services: "1"
                          max_number_of_services: "10"
                          max_pod_memory_capacity: "8G"
                          max_pod_cpu_capacity: "1"
                    backup:
                      mysql:
                        restore_during_installation: "No"
                        S3:
                          bucket: ""
                          key: ""
                          secret: ""
                    security:
                      image_pull_policy: ""
                      force_hashed_images: "No"
                      set_pod_resources: "Yes"
                      proxy:
                        http_proxy: ""
                        https_proxy: ""
                        additional_no_proxy: ""
                      use_rate_limiting: "No"
                      requests_per_second: "10"
                    affinity:
                      node_affinity:
                        - key: ""
                          operator: ""
                          values: ""
                          services: ""
                      node_selectors:
                        - key: ""
                          value: ""
                          services: ""
                      additional_labels:
                        - key: ""
                          value: ""
                          services: ""
                      additional_annotations:
                        - key: ""
                          value: ""
                          services: ""
              ```
              b. Customize the template fields by replacing with your information:

              - **host**: Enter the FQDN for accessing the vFunction dashboard.
                          Make sure that you use a domain name and not an IP address. Writing "https://" will mandate TLS while writing "http://" will expose the server through HTTP.
              - **org_name**: Enter your organization name.
              - **upgrade**: Choose auto-upgrade mode.
                  - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically.
                  - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
                  - **Never**: The automatic upgrading mechanism is off (default).
                  - **Maintenance**: Put the operator in a maintenance mode.
              - **admin.email**: Enter the email address of your administrator.
              - **admin.name**: Enter the name of your administrator.
              - **admin.password**: Enter the password you want to use for the vFunction administrator.
                                    The password should be at least 8 characters long, and consist of at least one lowercase letter, at least one uppercase letter, at least one number, and at least one special character.
              - **measurements.min\_num\_of\_services**: Minimum Number of measurement service pods.
              - **measurements.max\_num\_of\_services**: Maximum Number of measurement service pods.
              - **measurements.max\_pod\_memory\_capacity**: Maximum memory capacity limit for measurement service pods. (Default is "8G").
              - **measurements.max\_pod\_cpu\_capacity**: Maximum CPU capacity limit for measurement service pods. (Default is "1").
              - **measurements.pod\_ephemeral\_storage**: Ephemeral Storage Capacity for measurement service pods. (Default is "1G").
              - **measurements.S3**: Save all measurements data to S3, instead of local PV (optional).
                  - **measurements.S3.bucket**: Measurements S3 bucket name.
                  - **measurements.S3.key**: Measurements S3 key.
                  - **measurements.S3.secret**: Measurements S3 secret.
                  - **measurements.S3.region**: Measurements S3 region.
              - **smtp.user**: Enter the email address for a designated user for the SMTP server (optional).
              - **smtp.password**: Enter this user’s password (optional).
              - **smtp.identity**: Enter this user’s password (optional).
              - **smtp.url**: Enter the SMTP server URL (optional).
              - **authentication.authority**: Authority for authentication (Optional).
              - **authentication.client_id**: Client ID for authentication (Optional).
              - **authentication.client_secret**: Client Secret for authentication (Optional).
              - **authentication.jwks_url**: JWKS URL for authentication (Optional).
              - **authentication.issuer**: Issuer for authentication (Optional).
              - **authentication.ca\_root\_crt**: CA root certificate for authentication (Optional).
              - **tls.use_letsencrypt**: If your host FQDN starts with “https”, choose whether to use Let’s Encrypt to obtain certificates.
              - **tls.crt**: If your host FQDN starts with “https”, paste in the certifications you have for using the TLS connection (Applicable only if you don’t use Let’s Encrypt).
              - **tls.key**: If your host FQDN starts with “https”, enter the key you have for using the TLS connection (Applicable only if you don’t use Let’s Encrypt).
              - **nginx.force_http**: Force internal NGINX to listen on http even if the host FQDN starts with “https”.
                                      This is useful if you want to use a TLS termination proxy in front of vFunction, other then the route provided by the operator.
              - **nginx.ipv6_disabled**: Disable IPv6 on internal NGINX.
              - **nginx.service_type**: Set the service type for the internal NGINX service - ClusterIP, NodePort or LoadBalancer.
              - **route.use_route**: Use OpenShift route to expose vFunction. (Default is "Yes").
              - **route.tls_termination**: Set the TLS termination type for the OpenShift route - edge, passthrough or reencrypt. (Default is "edge").
              - **route.use\_tls\_certificate**: Use TLS certificate for OpenShift route (as provided in tls.crt and tls.key) - instead of the default one defined for the route controller in your cluster. (Default is "No").
              - **mysql.external\_mysql\_ip**: If you want to use an external MySQL database, enter its IP address.
              - **mysql.external\_mysql\_user**: If you want to use an external MySQL database, enter its user name.
              - **mysql.external\_mysql\_password**: If you want to use an external MySQL database, enter its password.
              - **mysql.max\_pod\_memory\_capacity**: Maximum memory capacity limit for MySQL pod. (Default is "2G").
              - **mysql.max\_pod\_cpu\_capacity**: Maximum CPU capacity limit for MySQL pod. (Default is "1").
              - **mysql.pod\_ephemeral\_storage**: Ephemeral Storage Capacity for MySQL pod. (Default is "1G").
              - **storage.storage\_class**: If you want to use a specific StorageClass for all PVCs, other than the default one, enter its name.
              - **storage.mysql\_persistent\_volume**: If you want to use a specific PV for mysql PVC, other than the dynamically provisioned one, enter its name.
              - **storage.mysql\_persistent\_volume\_claim**: If you want to use a specific pre-created PVC for mysql deployment, other than the dynamically provisioned one, enter its name.
              - **storage.mysql\_persistent\_volume\_claim\access\_mode**: If you want to use a specific access mode for mysql PVC, other than the default one, enter its name. (Default is "ReadWriteOnce").
              - **storage.storage\_persistent\_volume**: If you want to use a specific PV for storage PVC, other than the dynamically provisioned one, enter its name.
              - **storage.storage\_persistent\_volume\_claim**: If you want to use a specific pre-created PVC for storage deployment, other than the dynamically provisioned one, enter its name.
              - **storage.storage\_persistent\_volume\_claim\access\_mode**: If you want to use a specific access mode for storage PVC, other than the default one, enter its name. (Default is "ReadWriteOnce").
              - **storage.lets\_encrypt\_persistent\_volume**: If you want to use a specific PV for lets encrypt PVC, other than the dynamically provisioned one, enter its name.
              - **storage.lets\_encrypt\_persistent\_volume\_claim**: If you want to use a specific pre-created PVC for lets encrypt deployment, other than the dynamically provisioned one, enter its name.
              - **storage.lets\_encrypt\_persistent\_volume\_claim\access\_mode**: If you want to use a specific access mode for lets encrypt PVC, other than the default one, enter its name. (Default is "ReadWriteOnce").
              - **distributed.otlp.min\_number\_of\_services**: Minimum Number of OTLP pods.
              - **distributed.otlp.max\_number\_of\_services**: Maximum Number of OTLP pods.
              - **distributed.otlp.max\_pod\_memory\_capacity**: Maximum memory capacity for OTLP pods. (Default is "1G").
              - **distributed.otlp.max\_pod\_cpu\_capacity**: Maximum CPU capacity for OTLP pods. (Default is "1").
              - **distributed.otlp.pod\_ephemeral\_storage**: Ephemeral Storage Capacity for OTLP service pods. (Default is "250M").
              - **distributed.otlp.max\_folder\_size\_in\_mb**: Maximum folder size in MB for OTLP pods. (Default is "200").
              - **distributed.dm.min\_number\_of\_services**: Minimum Number of dm pods.
              - **distributed.dm.max\_number\_of\_services**: Maximum Number of dm pods.
              - **distributed.dm.max\_pod\_memory\_capacity**: Maximum memory capacity for dm pods. (Default is "8G").
              - **distributed.dm.max\_pod\_cpu\_capacity**: Maximum CPU capacity for dm pods. (Default is "1").
              - **distributed.dm.pod\_ephemeral\_storage**: Ephemeral Storage Capacity for dm service pods. (Default is "1G").
              - **backup.mysql.restore\_during\_installation**: If you want, during installation, to restore the backup from the latest kept backup with the same major version (if available on attached DB storage or S3 bucket), choose "Yes".
              - **backup.mysql.S3**: Backup MySQL to S3 (optional).
                  - **backup.mysql.S3.bucket**: MySQL backup S3 bucket name.
                  - **backup.mysql.S3.key**: MySQL backup S3 key.
                  - **backup.mysql.S3.secret**: MySQL backup S3 secret.
                  - **backup.mysql.S3.region**: MySQL backup S3 region.
              - **security.image\_pull\_policy**: If you want to use a custom image pull policy for all deployments, enter it here (optional).
              - **security.force\_hashed\_images**: If you want to use hashed images for all deployments, choose "Yes".
              - **security.set\_pod\_resources**: If you want to not set resource requests and limits for all deployments, choose "No".
              - **security.proxy.http\_proxy**: Enter the address of an HTTP proxy (optional).
              - **security.proxy.https\_proxy**: Enter the address of an HTTPS proxy (optional).
              - **security.proxy.additional\_no\_proxy**: Enter additional addresses to exclude from proxying (optional).
              - **security.use\_rate\_limiting**: If you want to use rate limiting in nginx, choose "Yes".
              - **security.requests\_per\_second**: Enter the number of requests per second for rate limiting.
              - **affinity.node\_affinity**: Enter the node affinity for the deployments (optional). There is only one nodeSelectorTerm as requiredDuringSchedulingIgnoredDuringExecution, but it can contain multiple matchExpressions (which are ANDed together).
                  - **key**: Enter the key for the node affinity.
                  - **operator**: Enter the operator for the node affinity (In, NotIn, Exists, DoesNotExist, Gt, Lt).
                  - **values**: Enter the values for the node affinity. A comma-separated list of values.
                  - **services**: Enter the services for the node affinity. A comma-separated list of services to which the node affinity applies. If not specified, the node affinity will apply to all services. List of possible services: mysql, nginx, storage, vfapi-parser, vfapi-users, vfapi-organizations, vfapi-idp, vfapi-mcp, vfapi-measurements, vfapi-otlp, vfapi-dm. (optional).
              - **affinity.node\_selectors**: Enter the node selectors for the deployments (optional).
                  - **key**: Enter the key for the node selector.
                  - **value**: Enter the value for the node selector.
                  - **services**: Enter the services for the node selector. A comma-separated list of services to which the node selector applies. If not specified, the node selector will apply to all services. List of possible services: mysql, nginx, storage, vfapi-parser, vfapi-users, vfapi-organizations, vfapi-idp, vfapi-mcp, vfapi-measurements, vfapi-otlp, vfapi-dm. (optional).
              - **affinity.additional\_labels**: Enter additional labels for the deployments (optional).
                  - **key**: Enter the key for the additional label.
                  - **value**: Enter the value for the additional label.
                  - **services**: Enter the services for the additional label. See above at the node selectors section for more details. (optional).
              - **affinity.additional\_annotations**: Enter additional annotations for the deployments (optional).
                  - **key**: Enter the key for the additional annotation.
                  - **value**: Enter the value for the additional annotation.
                  - **services**: Enter the services for the additional annotation. See above at the node selectors section for more details. (optional).
          1. In the Installed Operators window choose and click on vFunction Operator and then Create Instance.
          1. In the Create vFunctionServer window, paste the YAML file that you created in step 1.
          1. Click Create.
          1. You can now click on your new created vFunction operand in order to see its details and installation progress.

          ## Verifying the installation

          Check that the operand installed successfully:

          1. Check that the "Operand State" property shows “Working”.
          1. Check that the "Successfully Installed" property shows "Yes".

          If "Operand State" shows "Failed", it indicates the operand wasn’t installed correctly. Check the events and logs for all pods (operator and image containers) for any issues. In the event that you cannot troubleshoot, contact vFunction support.

          ## After installation

          The vFunction site is now accessible via the newly created vFunction application custom address, for example, my.domain.com. There are two ways you can access the vFunction dashboard.

          ### Access using the router canonical hostname

          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address, and to the ”vfunction” subdomain of the OpenShift canonical router hostname as the alias.
          For example, ***my.domain.com.   CNAME   vfunction.apps.ocp4.my-openshift.com.***
          2. Find your cluster Router Canonical Hostname address in the newly created vfunction-route-xxx under your vFunction OpenShift project -> **Networking** -> **Routes** -> **vfunction-route-xxx** route -> **Router Canonical Hostname** field.
          You can now access the dashboard using your defined "host" spec property (as above).

          ### Access using the nginx service location

          You may use this access method if your OpenShift is installed on a provider that supports exposing LoadBalancer-type services.
          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address and to the vfunction-nginx-xxx service location.
          For example, ***my.domain.com.   CNAME   a05951ed7cdf-1394239323.us-east-1.elb.amazonaws.com.***
          1. Point your custom domain to the vfunction-nginx-xxx service's external IP location, which you can find in the OpenShift project > **Networking** -> **Services** -> **vfunction-nginx-xxx** service > **Service Address** > **Location** field.

          ## Upgrade and Rollback

          The vFunction operator includes a built-in auto-upgrade mechanism.

          You can choose one of 4 upgrading modes:
          - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically.
          - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
          - **Never**: The automatic upgrading mechanism is off (default).
          - **Maintenance**: Put the operator in a maintenance mode.

          For an on-demand upgrade,  you can change anytime the upgrading mode from “Never” to “Always”, wait for the operand to be upgraded, and change again to “Never”.

          During the upgrade, the "Operand State" property will change to “Upgrading...” and back again to “Working” after a successful upgrade.

          In case of a failure upgrading the operand, an automatic rollback to the last working version will occur. The failed version will be marked as defective and the operator will not try to upgrade to it again.

          Choosing the "Maintenance" option will tell the operator to take down all pods, so you can safely fix storage issues, return volumes from snapshots, etc.
          After finishing the maintenance, return the upgrade mode to your original desired policy.
        displayName: vFunction Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - vfunction
        - analyzing
        - services
        links:
        - name: vFunction
          url: https://vfunction.com
        - name: Terms
          url: https://www.vfunction.com/eula/
        maintainers:
        - email: raviv@vfunction.com
          name: Raviv Halivni
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: vFunction
        relatedImages:
        - registry.connect.redhat.com/vfunction/vfunction-server-operator@sha256:a5a0f47f7fb9219139f8d749da14950a8c39698b340644011485dfbec54ebe8b
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-users@sha256:c471d450d016aa0801ffb3de4648735245f2cf6901576106aed06cacc948d6f6
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-parser@sha256:b70fe4340d0de33c03a3b3458ea29e4cbff30cdd43f0e8d94265568e9a8d1ac7
        - registry.connect.redhat.com/vfunction/vfunction-storage@sha256:8c36770f1fa5034ba089ae81c36c587cb73a5176f63969023c8ab6f90d8d7abd
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-idp@sha256:17fe2e3ef706a4bbd8fdf0e0e08bd021c49104b331674aae7b8978b31730a03d
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-organizations@sha256:8dd4d043806da1e85a4565b2f14bb9e24671fc158f8c1c2adc3804fda5213571
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-dm@sha256:2f82294b15af873c012094a8984cb0c157bf322c28cbe58b38253adc917b6662
        - registry.connect.redhat.com/vfunction/vfunction-mysql@sha256:b4b6543eb02e396e1339877f61b74cc783977dd607cd3679edce8ac6b627e9e9
        - registry.connect.redhat.com/vfunction/vfunction-nginx@sha256:8370b871d05a24da9841a93fb255bf47945428b0a192b96b4eb95b9c0efb86a0
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-mcp@sha256:d50142c1b7f8a52944f9051a13bb821c0a790f137ebef09b0a84256bf7e5d794
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-measurements@sha256:81edae1c7344c13916530f20c8be09512ac8e20542770949783b0025c37079f8
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-otlp@sha256:80ca6d64b0d74f2d4f1b267002d5666f7c602956c656fd36a5c430f77eef0672
        version: 4.3.1762
      entries:
      - name: vfunction-server-operator.v4.3.1762
        version: 4.3.1762
      - name: vfunction-server-operator.v4.0.1576
        version: 4.0.1576
      name: stable
    defaultChannel: stable
    packageName: vfunction-server-operator
    provider:
      name: vFunction
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent
      provider-url: ""
    name: cilium
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cilium.v1.10.15-x084926b
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/hubble-ui@sha256:dec6e7e541cac91ab859245dd2a7fa02de5363b75caa3509a79cf76d4f622550
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/hubble-ui-backend@sha256:a07e13df3e9d7ec0dcc3b158f243cda5a08f6eca38a25a5aa3171f6b3ec491ed
        - quay.io/cilium/operator-generic@sha256:7f2f7b59b541e4f660d65ed0a21d25ab69e49022172529771dd1134315e95c82
        - quay.io/cilium/clustermesh-apiserver@sha256:e73151c3295589de00dccae75819aabdff8fd907260817020f9e8fe3738e78b2
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - registry.connect.redhat.com/isovalent/cilium-olm:970b1d74dbe2b87c28e8dc1c68936de33123a4df-v1.10.15@sha256:afa3ac140fae24b151ab6ff5a8cdef11566a58f7234d702a6a45a398822e6d34
        - quay.io/cilium/cilium@sha256:adbb60f9d19d641a5496fcfc4769ebb55c1d4cd7699fef752f7c94754194cf83
        - quay.io/cilium/hubble-relay@sha256:67a9183fd1febd9e51f4d02bd35092b81c8612a1716015988f69fffba79d999a
        version: 1.10.15+x084926b
      entries:
      - name: cilium.v1.10.15-x084926b
        version: 1.10.15+x084926b
      - name: cilium.v1.10.13-xa8bb1e5
        version: 1.10.13+xa8bb1e5
      - name: cilium.v1.10.12-x362ea11
        version: 1.10.12+x362ea11
      - name: cilium.v1.10.10-x6adc6ff
        version: 1.10.10+x6adc6ff
      name: "1.10"
    - currentCSV: cilium.v1.11.20-xdb5ed32
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/cilium/cilium@sha256:60df3cb7155886e0b62060c7a4a31e457933c6e35af79febad5fd6e43bab2a99
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:3677840d57e1d74e3ac14db0dfceb83088e6ebd70775d775855cdd3bfed68059
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/operator-generic@sha256:1439954acf620f048ef663524ae70b4a25693c58527a2f2cee51124496e29f90
        - quay.io/cilium/hubble-ui@sha256:bbf8b041152041d3ef000d6074a8dd5d40ed910b5c0067f365b00e52e449d467
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/clustermesh-apiserver@sha256:46760182f8c98227cfac27627275616987b71509227775350573d834133a6d49
        - quay.io/cilium/hubble-relay@sha256:e2f38b901fd8bd5adc9a765a5e68836364ebd1e7dfb85c2bcd8a5488b23c3470
        - quay.io/cilium/hubble-ui-backend@sha256:46070270b8ffe7ac3f1794721be0b5a09f906dff3492f8a199a5f995db362d2a
        version: 1.11.20+xdb5ed32
      entries:
      - name: cilium.v1.11.20-xdb5ed32
        version: 1.11.20+xdb5ed32
      name: "1.11"
    - currentCSV: cilium.v1.12.19-xef55de8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.12.0 <1.12.19+xef55de8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/hubble-ui-backend@sha256:f88a73a120521eeafbcdbd908f517117f6557bf61e115847853fac371f0d774c
        - quay.io/cilium/hubble-relay@sha256:19f2dee371523888985e6e34b30d07ef3f84f2a9206ad79bdfc8dede96e25d5c
        - quay.io/cilium/clustermesh-apiserver@sha256:61f8e871760dceaa0c0459bbf44a3049dd0f4ab15e57e063d9b3284434cb3dd8
        - quay.io/cilium/hubble-ui@sha256:e15af59a2ded739e420be82de6cbdd0ce22f8d3f00e3a10b3d2d2734e074a394
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/cilium/cilium@sha256:b997994ab759443d9fc9807eb39b4784bf03e13730a8132ffe5df3edef22a312
        - quay.io/cilium/operator-generic@sha256:da4d72facecde862c88d08bf99c3f476845beb407052a0cc4c06b68cc93b7570
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:d35994cf5cf319b88ff5a7985cc972edd23df5a282360f9ba5516d610390ebf8
        version: 1.12.19+xef55de8
      entries:
      - name: cilium.v1.12.19-xef55de8
        version: 1.12.19+xef55de8
      - name: cilium.v1.12.18-x63f4f4c
        version: 1.12.18+x63f4f4c
      - name: cilium.v1.12.17-x6cc0c18
        version: 1.12.17+x6cc0c18
      - name: cilium.v1.12.16-x14b1d1d
        version: 1.12.16+x14b1d1d
      - name: cilium.v1.12.15-x7d84617
        version: 1.12.15+x7d84617
      - name: cilium.v1.12.14-x17ed60a
        version: 1.12.14+x17ed60a
      - name: cilium.v1.12.13-xb414403
        version: 1.12.13+xb414403
      - name: cilium.v1.12.12-xd280cc3
        version: 1.12.12+xd280cc3
      - name: cilium.v1.12.11-xc8fb7eb
        version: 1.12.11+xc8fb7eb
      - name: cilium.v1.12.10-xe372776
        version: 1.12.10+xe372776
      - name: cilium.v1.12.9-x2ccb2b6
        version: 1.12.9+x2ccb2b6
      - name: cilium.v1.12.8-xa53bac1
        version: 1.12.8+xa53bac1
      - name: cilium.v1.12.7-xba75145
        version: 1.12.7+xba75145
      - name: cilium.v1.12.0-xb42b669
        version: 1.12.0+xb42b669
      name: "1.12"
    - currentCSV: cilium.v1.13.12-x6e56d47
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.13.12+x6e56d47'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:b0d22d4376b4c2b2b3b988955227f1ed2ad49a363c5367e53f64edae2cec0ff5
        - quay.io/cilium/cilium@sha256:d99204aa7b3b7bd2c9ab47fd398cc9f40290799bc0c7a4386c8dc5c1780cd3d3
        - quay.io/cilium/clustermesh-apiserver@sha256:f1b1d0a85bab65e7d6adc90d000513a56ac58bdb071aa391a8580d73d20e6b6a
        - quay.io/cilium/operator-generic@sha256:f83734bbe270f961d545c7929152785507ce04a05d818ebc9776941723736d02
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/hubble-relay@sha256:01b23ea40bcd81145dde6bfcbfc4d542749d08c2a1c6348954c85123a8d2b1fc
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        version: 1.13.12+x6e56d47
      entries:
      - name: cilium.v1.13.12-x6e56d47
        version: 1.13.12+x6e56d47
      - name: cilium.v1.13.11-xba1ab0b
        version: 1.13.11+xba1ab0b
      - name: cilium.v1.13.10-x412ae64
        version: 1.13.10+x412ae64
      - name: cilium.v1.13.9-x3ae7099
        version: 1.13.9+x3ae7099
      - name: cilium.v1.13.8-xe0355c7
        version: 1.13.8+xe0355c7
      - name: cilium.v1.13.7-x37f9269
        version: 1.13.7+x37f9269
      - name: cilium.v1.13.6-xdb5afcd
        version: 1.13.6+xdb5afcd
      - name: cilium.v1.13.5-xaf5caa6
        version: 1.13.5+xaf5caa6
      - name: cilium.v1.13.4-x129f2a7
        version: 1.13.4+x129f2a7
      - name: cilium.v1.13.3-x5854d64
        version: 1.13.3+x5854d64
      - name: cilium.v1.13.2-x82c79c4
        version: 1.13.2+x82c79c4
      - name: cilium.v1.13.1-x10a5de9
        version: 1.13.1+x10a5de9
      - name: cilium.v1.13.0-x32540df
        version: 1.13.0+x32540df
      name: "1.13"
    - currentCSV: cilium.v1.14.7-x3522df7
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.14.7+x3522df7'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:291a05c58de8d7daefeb0689c23cadd92c9368aea09751de70eaa817f7fcc358
        - quay.io/cilium/operator-generic@sha256:37ef0bd85c27c765c637cd58c3ff4a559f8734ae39f9d1839a3ac7803de7b952
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/cilium@sha256:45ce2b87696082ecf7d53ba1c64ceeb4217578033e5ef28ac479ec049a48bc32
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/hubble-relay@sha256:46762393daf4a0aaef76b106614c2615942f98f10aeacd435ea3fb1a0bdf69e4
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/clustermesh-apiserver@sha256:28f3ffe53365ca79831af600f09a95c0b3e9959f5f891b416dab8cedd90c263d
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        version: 1.14.7+x3522df7
      entries:
      - name: cilium.v1.14.7-x3522df7
        version: 1.14.7+x3522df7
      - name: cilium.v1.14.6-x11464e1
        version: 1.14.6+x11464e1
      - name: cilium.v1.14.5-x3583390
        version: 1.14.5+x3583390
      - name: cilium.v1.14.4-xb89471d
        version: 1.14.4+xb89471d
      - name: cilium.v1.14.3-x041dc10
        version: 1.14.3+x041dc10
      - name: cilium.v1.14.2-xedd5ce2
        version: 1.14.2+xedd5ce2
      - name: cilium.v1.14.1-xdd859b7
        version: 1.14.1+xdd859b7
      name: "1.14"
    - currentCSV: cilium.v1.15.1-x7095b76
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.15.1+x7095b76'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:9ab6be29447125e886300e9258b9a06bedf0a9d87405832aa8b6565ed1ba4215
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        - quay.io/cilium/cilium@sha256:351d6685dc6f6ffbcd5451043167cfa8842c6decf80d8c8e426a417c73fb56d4
        - quay.io/cilium/clustermesh-apiserver@sha256:b353badd255c2ce47eaa8f394ee4cbf70666773d7294bd887693e0c33503dc37
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/operator-generic@sha256:819c7281f5a4f25ee1ce2ec4c76b6fbc69a660c68b7825e9580b1813833fa743
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/hubble-relay@sha256:3254aaf85064bc1567e8ce01ad634b6dd269e91858c83be99e47e685d4bb8012
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        version: 1.15.1+x7095b76
      entries:
      - name: cilium.v1.15.1-x7095b76
        version: 1.15.1+x7095b76
      name: "1.15"
    - currentCSV: cilium.v1.9.14-x70dc065
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Basic Install
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/hubble-relay@sha256:fd6ab1aea260abc5f64eca26c1b1e7009983e4aaa8e5d098e8d442f7659603fb
        - quay.io/cilium/clustermesh-apiserver@sha256:a0da5edf0372899647da51de1b277f0bab8e676d694aee7f939cddfdd3172010
        - quay.io/cilium/hubble-ui-backend@sha256:2cf1c7b1fde6161f575f3147c2ef885739e60b8c59a1ed444fe74edbdf22c7c1
        - quay.io/cilium/cilium@sha256:2c6ce93fa7e625979043a387eb998c17ad57df8768d89facb9b715da42a4c51c
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:3a9ca9a4c2f78109f265c4758f5f693d01865f2d0e31bc5f1ecbc065e878628b
        - docker.io/envoyproxy/envoy@sha256:55a141c0b8e39497b32074cb758b4e5dc2384bd6d9926144969a4452eaa611fd
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/hubble-ui@sha256:c267219946b84f79a5fd2849f5cc71f39453e29d93039e57b9cef2b5802e9c50
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/cilium/operator-generic@sha256:bdcfd2eade99933f2fda55ef79ea697ddfad3512f65b15bcd0ba7702518c1ba3
        version: 1.9.14+x70dc065
      entries:
      - name: cilium.v1.9.14-x70dc065
        version: 1.9.14+x70dc065
      name: "1.9"
    - currentCSV: cilium.v1.10.4-x5bfd7b3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Basic Install
          categories: Networking,Security
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:b45ff6a6552d71cbc26e358e0b25759d33818fd92301f1656453c5bcf826ff13
        version: 1.10.4+x5bfd7b3
      entries:
      - name: cilium.v1.10.4-x5bfd7b3
        version: 1.10.4+x5bfd7b3
      name: stable
    defaultChannel: "1.15"
    packageName: cilium
    provider:
      name: Isovalent
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: metallb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: metallb-operator.v4.16.0-202509230119
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "basic",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64512,
                        "neighbors": [
                          {
                            "address": "172.30.0.3",
                            "asn": 64512,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            },
                            "toReceive": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "two-neighs",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64512,
                        "neighbors": [
                          {
                            "address": "172.30.0.3",
                            "asn": 4200000000,
                            "ebgpMultiHop": true,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          },
                          {
                            "address": "172.18.0.6",
                            "asn": 4200000000,
                            "port": 179,
                            "toAdvertise": {
                              "allowed": {
                                "prefixes": [
                                  "192.168.2.0/24"
                                ]
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "two-routers-vrf",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64513,
                        "neighbors": [
                          {
                            "address": "172.32.0.3",
                            "asn": 4200000000,
                            "ebgpMultiHop": true,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ]
                      },
                      {
                        "asn": 64515,
                        "neighbors": [
                          {
                            "address": "172.31.0.3",
                            "asn": 4200000000,
                            "port": 179,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ],
                        "vrf": "red"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRNodeState",
                "metadata": {
                  "name": "frr-k8s-worker"
                },
                "status": {
                  "lastConversionResult": "success",
                  "lastReloadResult": "success",
                  "runningConfig": "Building configuration...\n\nCurrent configuration:\n!\nfrr version 9.0.2_git\nfrr defaults traditional\nhostname frr-k8s-worker\nlog file /etc/frr/frr.log informational\nlog timestamp precision 3\nno ipv6 forwarding\nservice integrated-vtysh-config\n!\nip nht resolve-via-default\n!\nipv6 nht resolve-via-default\n!\nend\n"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BFDProfile",
                "metadata": {
                  "name": "bfd-profile-sample",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "receiveInterval": 380,
                  "transmitInterval": 270
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "peers": [
                    "ebgp-single-hop0"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPoolSelectors": [
                    {
                      "matchLabels": {
                        "test": "ipv6"
                      }
                    },
                    {
                      "matchLabels": {
                        "test": "ipv4"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "communities": [
                    "65535:65282"
                  ],
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "localPref": 50
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample4",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "communities": [
                    "NO_ADVERTISE"
                  ],
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "localPref": 50
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample5",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "nodeSelectors": [
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-control-plane"
                      }
                    },
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-worker"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-beta1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "Community",
                "metadata": {
                  "name": "community1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "communities": [
                    {
                      "name": "NO_ADVERTISE",
                      "value": "65535:65282"
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "name": "ip-addresspool-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "192.168.10.0/24",
                    "192.168.9.1-192.168.9.5",
                    "fc00:f853:0ccd:e799::/124"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "labels": {
                    "test": "ipv4"
                  },
                  "name": "ip-addresspool-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "172.20.0.100/24"
                  ],
                  "autoAssign": false
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "labels": {
                    "test": "ipv6"
                  },
                  "name": "ip-addresspool-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "2002:2:2::1-2002:2:2::100"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "L2Advertisement",
                "metadata": {
                  "name": "l2-adv-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "L2Advertisement",
                "metadata": {
                  "name": "l2-adv-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "nodeSelectors": [
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-control-plane"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "MetalLB",
                "metadata": {
                  "name": "metallb",
                  "namespace": "metallb-system"
                },
                "spec": {}
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "bfdProfile": "bfd-profile-sample",
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "holdTime": "9m0s",
                  "keepaliveTime": "0s",
                  "myASN": 64512,
                  "passwordSecret": {
                    "name": "secretname",
                    "namespace": "metallb-system"
                  },
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3",
                  "peerPort": 180
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:b07c046ba2ae71b93f27d5cdca7cb93f51f3198cb84bb6e65dd6320bcda268c8
          createdAt: "2023-06-06T15:25:00Z"
          description: An operator for deploying MetalLB on a kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.8.0 <4.16.0'
          operatorframework.io/suggested-namespace: metallb-system
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/metallb-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BFDProfile is the Schema for the bfdprofiles API
            kind: BFDProfile
            name: bfdprofiles.metallb.io
            version: v1beta1
          - description: BGPAdvertisement is the Schema for the bgpadvertisements
              API
            kind: BGPAdvertisement
            name: bgpadvertisements.metallb.io
            version: v1beta1
          - description: BGPPeer is the Schema for the bgppeers API
            kind: BGPPeer
            name: bgppeers.metallb.io
            version: v1beta1
          - description: BGPPeer is the Schema for the bgppeers API
            kind: BGPPeer
            name: bgppeers.metallb.io
            version: v1beta2
          - description: Community is the Schema for the communities API
            kind: Community
            name: communities.metallb.io
            version: v1beta1
          - description: FRRConfiguration is the Schema for the frrconfigurations
              API
            kind: FRRConfiguration
            name: frrconfigurations.frrk8s.metallb.io
            version: v1beta1
          - description: FRRNodeState is the Schema for the frrnodestates API
            kind: FRRNodeState
            name: frrnodestates.frrk8s.metallb.io
            version: v1beta1
          - description: IPAddressPool is the Schema for the ipaddresspools API
            kind: IPAddressPool
            name: ipaddresspools.metallb.io
            version: v1beta1
          - description: L2Advertisement is the Schema for the l2advertisements API
            kind: L2Advertisement
            name: l2advertisements.metallb.io
            version: v1beta1
          - description: MetalLB is the Schema for the metallbs API
            displayName: MetalLB
            kind: MetalLB
            name: metallbs.metallb.io
            version: v1beta1
        description: An operator for deploying MetalLB on a kubernetes cluster.
        displayName: MetalLB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - metallb-operator
        links:
        - name: MetalLB Operator
          url: https://github.com/openshift/metallb-operator
        maintainers:
        - email: fpaoline@redhat.com
          name: Federico Paolinelli
        - email: obraunsh@redhat.com
          name: Ori Braunshtein
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/metallb-rhel9@sha256:95abda97d689078bd124ef59c92cc12e661222cac8e8b0e526579470ab54d05e
        - registry.redhat.io/openshift4/frr-rhel9@sha256:48a361644b2fa8fc632f49117172db2e9c12022774e7e0fd409561d884a6c237
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:b07c046ba2ae71b93f27d5cdca7cb93f51f3198cb84bb6e65dd6320bcda268c8
        version: 4.16.0-202509230119
      entries:
      - name: metallb-operator.v4.16.0-202509230119
        version: 4.16.0-202509230119
      - name: metallb-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: metallb-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: metallb-operator.v4.16.0-202507221936
        version: 4.16.0-202507221936
      - name: metallb-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: metallb-operator.v4.16.0-202506090335
        version: 4.16.0-202506090335
      - name: metallb-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: metallb-operator.v4.16.0-202505051339
        version: 4.16.0-202505051339
      - name: metallb-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: metallb-operator.v4.16.0-202503252033
        version: 4.16.0-202503252033
      - name: metallb-operator.v4.16.0-202502260004
        version: 4.16.0-202502260004
      - name: metallb-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: metallb-operator.v4.16.0-202501232204
        version: 4.16.0-202501232204
      - name: metallb-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: metallb-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: metallb-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: metallb-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: metallb-operator.v4.16.0-202411110236
        version: 4.16.0-202411110236
      - name: metallb-operator.v4.16.0-202410292005
        version: 4.16.0-202410292005
      - name: metallb-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: metallb-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: metallb-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: metallb-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: metallb-operator.v4.16.0-202409161407
        version: 4.16.0-202409161407
      - name: metallb-operator.v4.16.0-202409062036
        version: 4.16.0-202409062036
      - name: metallb-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: metallb-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: metallb-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: metallb-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: metallb-operator.v4.16.0-202408051837
        version: 4.16.0-202408051837
      - name: metallb-operator.v4.16.0-202407221107
        version: 4.16.0-202407221107
      - name: metallb-operator.v4.16.0-202407180936
        version: 4.16.0-202407180936
      - name: metallb-operator.v4.16.0-202407091337
        version: 4.16.0-202407091337
      - name: metallb-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: metallb-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: metallb-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: metallb-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: metallb-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app: enmasse
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq-online
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-online.1.7.4-0.1657725339.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "StandardInfraConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "broker": {
                    "resources": {
                      "memory": "1Gi",
                      "storage": "5Gi"
                    },
                    "addressFullPolicy": "FAIL"
                  },
                  "router": {
                    "linkCapacity": 50,
                    "resources": {
                      "memory": "512Mi"
                    }
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "BrokeredInfraConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "broker": {
                    "resources": {
                      "memory": "4Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta2",
                "kind": "AddressPlan",
                "metadata": {
                  "name": "standard-small-queue"
                },
                "spec": {
                  "addressType": "queue",
                  "shortDescription": "Small Queue",
                  "resources": {
                    "router": 0.01,
                    "broker": 0.1
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta2",
                "kind": "AddressSpacePlan",
                "metadata": {
                  "name": "standard-small"
                },
                "spec": {
                  "addressSpaceType": "standard",
                  "infraConfigRef": "default",
                  "shortDescription": "Small Address Space Plan",
                  "resourceLimits": {
                    "router": 1.0,
                    "broker": 2.0,
                    "aggregate": 2.0
                  },
                  "addressPlans": [
                    "standard-small-queue"
                  ]
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "AuthenticationService",
                "metadata": {
                  "name": "standard-authservice"
                },
                "spec": {
                  "type": "standard",
                  "standard": {
                    "storage": {
                      "claimName": "standard-authservice",
                      "deleteClaim": true,
                      "size": "1Gi",
                      "type": "persistent-claim"
                    }
                  }
                }
              },
              {
                "apiVersion": "enmasse.io/v1beta1",
                "kind": "AddressSpace",
                "metadata": {
                  "name": "myspace"
                },
                "spec": {
                  "plan": "standard-small",
                  "type": "standard"
                }
              },
              {
                "apiVersion": "enmasse.io/v1beta1",
                "kind": "Address",
                "metadata": {
                  "name": "myspace.myqueue"
                },
                "spec": {
                  "address": "myqueue",
                  "plan": "standard-small-queue",
                  "type": "queue"
                }
              },
              {
                "apiVersion": "user.enmasse.io/v1beta1",
                "kind": "MessagingUser",
                "metadata": {
                  "name": "myspace.user"
                },
                "spec": {
                  "authentication": {
                    "password": "ZW5tYXNzZQ==",
                    "type": "password"
                  },
                  "authorization": [
                    {
                      "addresses": [
                        "myqueue"
                      ],
                      "operations": [
                        "send",
                        "recv"
                      ]
                    }
                  ],
                  "username": "user"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator@sha256:2aaf87224a96aa56dc595e96d8e7ec8d138aaf20240afa25a1e805b405941355
          createdAt: "2019-02-19T00:00:00Z"
          description: Red Hat Integration - AMQ Online provides messaging as a managed
            service on OpenShift
          olm.substitutesFor: amq-online.1.7.4
          operators.openshift.io/capability: '["license-required"]'
          operators.openshift.io/valid-licenses: '["Red Hat AMQ", "Red Hat Integration"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: ""
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A messaging user that can connect to an Address Space
            displayName: Messaging User
            kind: MessagingUser
            name: messagingusers.user.enmasse.io
            version: v1beta1
          - description: A resource representing the available schema of plans and
              authentication services
            displayName: AddressSpaceSchema
            kind: AddressSpaceSchema
            name: addressspaceschemas.enmasse.io
            version: v1beta1
          - description: A group of messaging addresses that can be accessed via the
              same endpoint
            displayName: Address Space
            kind: AddressSpace
            name: addressspaces.enmasse.io
            version: v1beta1
          - description: A messaging address that can be used to send/receive messages
              to/from
            displayName: Address
            kind: Address
            name: addresses.enmasse.io
            version: v1beta1
          - description: Infrastructure configuration template for the standard address
              space type
            displayName: Standard Infra Config
            kind: StandardInfraConfig
            name: standardinfraconfigs.admin.enmasse.io
            version: v1beta1
          - description: Infrastructure configuration template for the brokered address
              space type
            displayName: Brokered Infra Config
            kind: BrokeredInfraConfig
            name: brokeredinfraconfigs.admin.enmasse.io
            version: v1beta1
          - description: Plan describing the resource usage of a given address type
            displayName: Address Plan
            kind: AddressPlan
            name: addressplans.admin.enmasse.io
            version: v1beta2
          - description: Plan describing the capabilities and resource limits of a
              given address space type
            displayName: Address Space Plan
            kind: AddressSpacePlan
            name: addressspaceplans.admin.enmasse.io
            version: v1beta2
          - description: Authentication service configuration available to address
              spaces.
            displayName: Authentication Service
            kind: AuthenticationService
            name: authenticationservices.admin.enmasse.io
            version: v1beta1
          - description: Console Service Singleton for deploying global console.
            displayName: Console Service
            kind: ConsoleService
            name: consoleservices.admin.enmasse.io
            version: v1beta1
        description: "**Red Hat Integration - AMQ Online** provides messaging as a
          managed service on OpenShift.\nWith Red Hat Integration - AMQ Online, administrators
          can configure a cloud-native,\nmulti-tenant messaging service either in
          the cloud or on premise.\nDevelopers can provision messaging using the Red
          Hat Integration - AMQ Online Console.\nMultiple development teams can provision
          the brokers and queues from the\nconsole, without requiring each team to
          install, configure, deploy,\nmaintain, or patch any software. \n\n**The
          core capabilities include:**\n\n  * **Built-in authentication and authorization**
          - Use the built-in authentication service or\n    plug in your own authentication
          service for authentication and\n    authorization of messaging clients.\n\n
          \ * **Self-service messaging for applications** - The service administrator
          deploys\n    and manages the messaging infrastructure, while applications
          can request\n    messaging resources regardless of the messaging infrastructure.\n\n
          \ * **Support for a wide variety of messaging patterns** - Choose between\n
          \   JMS-style messaging with strict guarantees, or messaging that supports\n
          \   a larger number of connections and higher throughput.\n\n## Post-installation
          tasks\n\nTo fully use Red Hat Integration - AMQ Online, you need to create
          a few\ninfrastructure components after the installation, including:\n\n
          \ * An authentication service\n  * Infrastructure configuration for the
          standard and brokered address space\n  * Address and address space plans\n
          \ * (Optional) Create RBAC roles to allow users to discover available plans\n
          \ * (Optional) Create RBAC roles to allow users to self-manage addresses
          and\n    address spaces.\n\nFor a complete overview of how to configure
          Red Hat Integration - AMQ Online, see\n[Configuring Red Hat Integration
          - AMQ Online](https://access.redhat.com/documentation/en-us/red_hat_amq/7.7/html/installing_and_managing_amq_online_on_openshift/configuring-messaging).\n\n###
          Quickstart infrastructure configuration\n\nIf you simply want to get started
          quickly, you can import the following\nYAML by saving the content to a file
          and apply it by running the \n`oc apply -f <file>` command. You can also
          split the content (at the `---` marker)\nand import the single YAML document
          using the Web UI: \n\n~~~yaml\n---\napiVersion: admin.enmasse.io/v1beta1\nkind:
          StandardInfraConfig\nmetadata:\n  name: default\n---\napiVersion: admin.enmasse.io/v1beta2\nkind:
          AddressPlan\nmetadata:\n  name: standard-small-queue\nspec:\n  addressType:
          queue\n  resources:\n    router: 0.01\n    broker: 0.1\n---\napiVersion:
          admin.enmasse.io/v1beta2\nkind: AddressSpacePlan\nmetadata:\n  name: standard-small\nspec:\n
          \ addressSpaceType: standard\n  infraConfigRef: default\n  addressPlans:\n
          \ - standard-small-queue\n  resourceLimits:\n    router: 2.0\n    broker:
          3.0\n    aggregate: 4.0\n---\napiVersion: admin.enmasse.io/v1beta1\nkind:
          AuthenticationService\nmetadata:\n  name: standard-authservice\nspec:\n
          \ type: standard\n  standard:\n    storage:\n      claimName: standard-authservice\n
          \     deleteClaim: true\n      size: 1Gi\n      type: persistent-claim\n~~~\n\n###
          Create RBAC roles to allow users to discover available plans\n\nFor users
          to discover the available plans, cluster-wide roles to read the available\nschema
          can be created.  Import the following YAML by saving the content to a file
          and apply it by running the \n`oc apply -f <file>` command. You can also
          split the content (at the `---` marker)\nand import the single YAML document
          using the Web UI: \n\n~~~yaml\n---\napiVersion: rbac.authorization.k8s.io/v1\nkind:
          ClusterRole\nmetadata:\n  name: enmasse.io:schema\n  labels:\n    app: enmasse\nrules:\n
          \ - apiGroups: [ \"enmasse.io\" ]\n    resources: [ \"addressspaceschemas\"
          ]\n    verbs: [ \"get\", \"list\", \"watch\" ]\n---\napiVersion: rbac.authorization.k8s.io/v1\nkind:
          ClusterRoleBinding\nmetadata:\n  name: \"enmasse.io:schema\"\n  labels:\n
          \   app: enmasse\nroleRef:\n  apiGroup: rbac.authorization.k8s.io\n  kind:
          ClusterRole\n  name: enmasse.io:schema\nsubjects:\n- kind: Group\n  name:
          system:authenticated\n~~~\n\n### Creating infrastructure using the Web UI\n\nYou
          must create a new instance of each of the following custom resources. You
          can\nuse the example data directly, which is provided when using the\nWeb
          UI:\n\n  * *Authentication Service* – Create an authentication service.\n
          \ * *Brokered Infra Config* – Create the broker infrastructure configuration.\n
          \ * *Standard Infra Config* – Create the standard infrastructure\n    configuration.\n\nYou
          must also create at least one address space plan and one address plan.\n\n*Note*:
          The name of the address space plan and address plan is required\nlater when
          creating new addresses. Some examples use specific plan\nnames, which might
          not be available in your environment. You can\ncreate those plans, or edit
          the examples to use different plan names.\n\n## Configuration for messaging
          tenants\n\nWhile service administrators perform the infrastructure configuration,
          the following\nresources are for the actual users of the system, the messaging
          tenants.\n\nYou need to create those resources to satisfy your particular
          use case.\n\n  * *Address space* – A container for addresses\n  * *Address*
          – A messaging address (address, topic, queue, and so on)\n  * *Messaging
          user* – Manages access to an address\n"
        displayName: Red Hat Integration - AMQ Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amq
        - messaging
        - amqp
        - iot
        - mqtt
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/7.7/html/using_amq_online_on_openshift/index
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-online-1-topic-forwarder@sha256:de33a701339d91a4e9275f7cf65f917ff75368a59b8ab36dd66c90bc306875aa
        - registry.redhat.io/amq7/amq-online-1-broker-plugin@sha256:2466a336c0d78b196635de3fad636aeff2d9265839f7271b8ef94f3ba97d9bc0
        - registry.redhat.io/amq7/amq-interconnect@sha256:5ebd4c4a0ed0db2725ae466550c915531b32933c420344cc2083c616da3f326f
        - registry.redhat.io/redhat-sso-7/sso73-openshift@sha256:18d15980f28f0d0c6b8582131a1f062d84463577512875b5391a612c3d7d1096
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6d489a7897bd831ec273034f1ff0e223427a40afd01b9835ebb020a8bf9ddc66
        - registry.redhat.io/amq7/amq-broker@sha256:1ec0437bbd82722148fdbb02dc55682b114c311cd6f521a8bc621a8c1cb4b32b
        - registry.redhat.io/amq7/amq-online-1-address-space-controller@sha256:9acc555c8e869aec7d72f53b60320e1a96cc82d748638e0d50970837229fdbee
        - registry.redhat.io/amq7/amq-online-1-none-auth-service@sha256:a013fb1806c002b408a7753ff943cee24fbcdf8a657eff1e9582c42b5a7b0844
        - registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator@sha256:2aaf87224a96aa56dc595e96d8e7ec8d138aaf20240afa25a1e805b405941355
        - registry.redhat.io/amq7/amq-online-1-standard-controller@sha256:070349af30cec52fa8efc185fc940077a64bb40e3e36310d7cc6e39ce0971436
        - registry.redhat.io/amq7/amq-online-1-console-init@sha256:0970804909b82fdc3d4e7d5d88280e41157d743107de8bfa2d1baa1367543176
        - registry.redhat.io/openshift3/oauth-proxy@sha256:59b20f23c03d0012ec5c95c4725892ee64277c22dedc9d0a90a7b6b0bc80fc72
        - registry.redhat.io/amq7/amq-online-1-console-server-rhel7@sha256:1dcb3d0cc7bc6e84f4f96962ff88849511b7a7686a1f5dbd9300c4c2c8c05e2f
        - registry.redhat.io/amq7/amq-online-1-agent@sha256:f98ed8ee97ac59d8fdd40ec8b2c063a5afd360a3e183057c186e7a77a32801ee
        - registry.redhat.io/amq7/amq-online-1-auth-plugin@sha256:68926f434b2f6ce4d9698938bc3c53d514c21c3f76c73b6b4cdf7fcf51f7cb0d
        version: 1.7.4+0.1657725339.p
      entries:
      - name: amq-online.1.7.4
        version: 1.7.4
      - name: amq-online.1.7.4-0.1657725339.p
        version: 1.7.4+0.1657725339.p
      - name: amq-online.1.7.3
        version: 1.7.3
      - name: amq-online.1.7.2
        version: 1.7.2
      - name: amq-online.1.7.1
        version: 1.7.1
      - name: amq-online.1.7.1-0.1628610187.p
        version: 1.7.1+0.1628610187.p
      - name: amq-online.1.7.0
        version: 1.7.0
      name: stable
    defaultChannel: stable
    packageName: amq-online
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: sandboxed-containers-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sandboxed-containers-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:051bd7f1dad8cc3251430fee32184be8d64077aba78580184cef0255d267bdcf
          createdAt: 2021-06-07-13:38:06
          olm.skipRange: <1.0.2
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          OpenShift sandboxed containers based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift’s Defense-in-Depth strategy. For more information
          [see](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5).

          # Features & benefits
          - **Isolated Developer Environments & Priviliges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as CAP_ADMIN or CAP_BPF. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select preview-1.0 from the list of available Update Channel options.
            This ensures that you install the version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              openshift-sandboxed-containers-operator namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              openshift-sandboxed-containers-operator causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a KataConfig CRD instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.8/sandboxed_containers/understanding-sandboxed-containers.html).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kernel-isolated containers
        links:
        - name: OpenShift sandboxed containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        provider:
          name: Red Hat
          url: https://www.github.com/openshift/sandboxed-containers-operator
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:051bd7f1dad8cc3251430fee32184be8d64077aba78580184cef0255d267bdcf
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:c6c589d5e47ba9564c66c84fc2bc7e5e046dae1d56a3dc99d7343f01e42e4d31
        version: 1.0.2
      entries:
      - name: sandboxed-containers-operator.v1.0.2
        version: 1.0.2
      name: preview-1.0
    - currentCSV: sandboxed-containers-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:b1e824e126c579db0f56d04c3d1796d82ed033110c6bc923de66d95b67099611
          createdAt: "2021-10-08T13:20:20.250789Z"
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift’s Defense-in-Depth strategy. For more information
          [see](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5).

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `preview-1.1` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CRD instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.8/sandboxed_containers/understanding-sandboxed-containers.html).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kernel-isolated containers
        links:
        - name: OpenShift sandboxed containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        provider:
          name: Red Hat
          url: https://www.github.com/openshift/sandboxed-containers-operator
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:b1e824e126c579db0f56d04c3d1796d82ed033110c6bc923de66d95b67099611
        version: 1.1.0
      entries:
      - name: sandboxed-containers-operator.v1.1.0
        version: 1.1.0
      - name: sandboxed-containers-operator.v1.0.2
        version: 1.0.2
      name: preview-1.1
    - currentCSV: sandboxed-containers-operator.v1.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-12T14:30:28Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.10.3'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '["peerpods.confidentialcontainers.org","peerpodconfigs.confidentialcontainers.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/sandboxed-containers-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
          - kind: PeerPod
            name: peerpods.confidentialcontainers.org
            version: v1alpha1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `candidate` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the [OpenShift sandboxed containers documentation](https://docs.redhat.com/en/documentation/openshift_sandboxed_containers/).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-cloud-api-adaptor-rhel9@sha256:7d1fa1e85af0d34c23e75868b1388eff17abe633c04a8dc55fbefcc470195c05
        - registry.redhat.io/openshift-sandboxed-containers/osc-cloud-api-adaptor-webhook-rhel9@sha256:6802f00a939f24ecd6dad150433db5ea3fd72d171f0c410cb894d3af910fc7fd
        - registry.redhat.io/openshift-sandboxed-containers/osc-podvm-builder-rhel9@sha256:c15f36a639fa07fee11f74c6563696ec8ef4da3fb1a99b7bd3f308dd04881339
        - registry.redhat.io/openshift-sandboxed-containers/osc-podvm-payload-rhel9@sha256:8d6b1262c6cedd04e33441264539842d82ed057f64eda5c4bce59566bcb72558
        - registry.redhat.io/openshift-sandboxed-containers/osc-dm-verity-image@sha256:e633bc27853f811d919e64461d0c917e659f5430458abe2304a4c85b0d871c80
        - registry.redhat.io/openshift-sandboxed-containers/osc-monitor-rhel9@sha256:037cbc221ae8fda83cdf1093127795ea806ecda53893deb49cdf981e7863774c
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel9-operator@sha256:1ffb432d9a8833854fa5e5262dcbf42530dfbaadfddf6978f7bc7804b84d2e3f
        version: 1.10.3
      entries:
      - name: sandboxed-containers-operator.v1.10.3
        version: 1.10.3
      - name: sandboxed-containers-operator.v1.10.2
        version: 1.10.2
      - name: sandboxed-containers-operator.v1.10.1
        version: 1.10.1
      - name: sandboxed-containers-operator.v1.10.0
        version: 1.10.0
      - name: sandboxed-containers-operator.v1.9.0
        version: 1.9.0
      - name: sandboxed-containers-operator.v1.8.1
        version: 1.8.1
      - name: sandboxed-containers-operator.v1.7.0
        version: 1.7.0
      - name: sandboxed-containers-operator.v1.6.0
        version: 1.6.0
      - name: sandboxed-containers-operator.v1.5.3
        version: 1.5.3
      - name: sandboxed-containers-operator.v1.5.2
        version: 1.5.2
      - name: sandboxed-containers-operator.v1.5.1
        version: 1.5.1
      - name: sandboxed-containers-operator.v1.5.0
        version: 1.5.0
      name: stable
    - currentCSV: sandboxed-containers-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:8f30a9129d817c3f4e404d2c43fb47e196d8c8da3badba4c48f65d440a4d7584
          createdAt: "2021-11-22T14:44:50.817336Z"
          olm.skipRange: '>=1.1.0 <1.2.2'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `stable-1.2` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.9/sandboxed_containers/understanding-sandboxed-containers.html)
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:a660f0b54b9139bed9a3aeef3408001c0d50ba60648364a98a09059b466fbcc1
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:8f30a9129d817c3f4e404d2c43fb47e196d8c8da3badba4c48f65d440a4d7584
        version: 1.2.2
      entries:
      - name: sandboxed-containers-operator.v1.2.2
        version: 1.2.2
      name: stable-1.2
    - currentCSV: sandboxed-containers-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          olm.skipRange: '>=1.1.0 <1.3.3'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/sandboxed-containers-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `stable-1.3` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/latest/sandboxed_containers/understanding-sandboxed-containers.html)
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:5d2b03721043e5221dfb0cf164cf59eba396ba3aae40a56c53aa3496c625eea0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:cb836456974e510eb4bccbffadbc6d99d5f57c36caec54c767a158ffd8a025d5
        version: 1.3.3
      entries:
      - name: sandboxed-containers-operator.v1.3.3
        version: 1.3.3
      name: stable-1.3
    defaultChannel: stable
    packageName: sandboxed-containers-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/stolostron/multicluster-global-hub
    name: multicluster-global-hub-operator-rh
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-global-hub-operator-rh.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:01e8dc62d5dc48af84b92f5315fab22574f648f2f705ac94f0516f4a7dab9237
          createdAt: 27 Aug 2024, 22:21
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.1.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```


          ### Documentation

          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: release-1.1
        provider:
          name: Red Hat
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:01e8dc62d5dc48af84b92f5315fab22574f648f2f705ac94f0516f4a7dab9237
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:78628c4f7a0286417023410261cc3b36797b7cdf72b38d0f4a2533a0f3db89dc
        - registry.redhat.io/rhel9/postgresql-13@sha256:c67d00643295697d196dd70dea457948ebb9e9d9d61a42019490b350c8286b50
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:5c8cea421be39712503378e32c54c332a7a0bf77a1705781d10fc123224bd60e
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:4ea8d3a5317d3b9968fa93dd68416e07e12e00e492860ff20ed90f88ca6ba225
        version: 1.1.2
      entries:
      - name: multicluster-global-hub-operator-rh.v1.1.2
        version: 1.1.2
      - name: multicluster-global-hub-operator-rh.v1.1.1
        version: 1.1.1
      - name: multicluster-global-hub-operator-rh.v1.1.0
        version: 1.1.0
      name: release-1.1
    - currentCSV: multicluster-global-hub-operator-rh.v1.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:98d470c8683c346107ab943776991554e3173c9a6431202bdacc48a75c4cf18f
          createdAt: 15 May 2025, 21:22
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          kafka-zookeeper-0                                       1/1     Running   0          21m
          kafka-zookeeper-1                                       1/1     Running   0          21m
          kafka-zookeeper-2                                       1/1     Running   0          21m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgres-0                      2/2     Running   0          22m
          ```


          ### Documentation

          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: release-1.2
        provider:
          name: Red Hat
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:c83a7b01bfe5a39d6cdf585180a011448e935d109768cb51560308b29d9ffe14
        - registry.redhat.io/rhel9/postgresql-13@sha256:b1af637cbe0e4041cd2fe6fdbcf665c5183e2991d3cc3e5b7384235f801504d9
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:16bb76ad0c68c0d6b3149b2741126a96e154150c1059621b14c2e6c30a214f16
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:98d470c8683c346107ab943776991554e3173c9a6431202bdacc48a75c4cf18f
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:92f44cff160874efe57a985609e599b13c49c0c4cef846dccc80833af2ae74f3
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:64e5ade4fbcd4cb70d62a6a62d775e051ff49838b891b682574646d4e61d9751
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:9b98fc01b08c3262e5ab99096e74fb55982f2a9c0de61e8fe84e419522e725b0
        version: 1.2.3
      entries:
      - name: multicluster-global-hub-operator-rh.v1.2.3
        version: 1.2.3
      - name: multicluster-global-hub-operator-rh.v1.2.2
        version: 1.2.2
      - name: multicluster-global-hub-operator-rh.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator-rh.v1.2.0
        version: 1.2.0
      name: release-1.2
    - currentCSV: multicluster-global-hub-operator-rh.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:0ef24ded54588aa49e366ba1cafc0971162e9d32024045bbcd7ef16e9f38a2d5
          createdAt: 01 Apr 2025, 20:25
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          kafka-zookeeper-0                                       1/1     Running   0          21m
          kafka-zookeeper-1                                       1/1     Running   0          21m
          kafka-zookeeper-2                                       1/1     Running   0          21m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgres-0                      2/2     Running   0          22m
          ```


          ### Documentation

          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: release-1.3
        provider:
          name: Red Hat
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:a139da58c681de987d847b233dd5c1e6ced5d6e86f3aab0f1b4181a77fc583d4
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:a17e90abc5e23bbc3908613c56c4208ad554eb27ed9679c4ccd06c4a1e9fcbcc
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:0ef24ded54588aa49e366ba1cafc0971162e9d32024045bbcd7ef16e9f38a2d5
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:619d8cb6b3811169af183bd8bc26ed98cdb25486ec428d88b25fee0ef9e6c552
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:e848a90d17e4f00d8021cbde47e488b82c447d8a5f6588f85f3d7cfbd6877f0f
        - registry.redhat.io/rhel9/postgresql-13@sha256:76c49b7dbd7c4ed8b430c4931b32f546db6d07ddc2146232f5d17eb17d3243b4
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-kessel-inventory-api-rhel9@sha256:dbd3924eb24db7ef28301b17851b5b5cd20e044cf259f0c556c67a0721561803
        version: 1.3.3
      entries:
      - name: multicluster-global-hub-operator-rh.v1.3.3
        version: 1.3.3
      - name: multicluster-global-hub-operator-rh.v1.3.2
        version: 1.3.2
      - name: multicluster-global-hub-operator-rh.v1.3.1
        version: 1.3.1
      - name: multicluster-global-hub-operator-rh.v1.3.0
        version: 1.3.0
      name: release-1.3
    - currentCSV: multicluster-global-hub-operator-rh.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:c5b3b3907e8135d3eedba40e27ba18a232ea4f30854d652059fcb14f82a8bf15
          createdAt: "2025-07-23T09:04:08Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: release-1.4
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:047b6c378e2b37339dd498c9025060f20812014bde4d0c98183f1a1201c21e66
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:ea906dcf8d45956251df22e269fe0b8fa8d49734e96a61b10db7b25d626b2e6a
        - registry.redhat.io/rhel9/postgresql-16@sha256:2161182c88022f38d6d9b8b467c0d59156edcededc908d383b80caa4eedf57f9
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:2be8bc79667018386235823e322ae10788e4e7e889c10e63566ae27a65dcabf1
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:0f4683d18404b4a9412e42ed417f6d3ac0024b3c57aa72c9162ff72d6ea8d394
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:c5b3b3907e8135d3eedba40e27ba18a232ea4f30854d652059fcb14f82a8bf15
        version: 1.4.2
      entries:
      - name: multicluster-global-hub-operator-rh.v1.4.2
        version: 1.4.2
      - name: multicluster-global-hub-operator-rh.v1.4.1
        version: 1.4.1
      - name: multicluster-global-hub-operator-rh.v1.4.0
        version: 1.4.0
      name: release-1.4
    - currentCSV: multicluster-global-hub-operator-rh.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:8af7e57b8c6285ad232084e7776fef32d026e080eb0bf14b91986ac28d00bb6a
          createdAt: "2025-08-18T08:30:32Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: release-1.5
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:8af7e57b8c6285ad232084e7776fef32d026e080eb0bf14b91986ac28d00bb6a
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:c95052569312d805b6750477c8bc57f7c72e1efd26143f3899ea0eddc2aed37b
        - registry.redhat.io/rhel9/postgresql-16@sha256:2161182c88022f38d6d9b8b467c0d59156edcededc908d383b80caa4eedf57f9
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:921163d139bca9e47bf95df43d7b408ff6fe2bf9b0902fb6af9f4cfc9d4598cc
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:ab3ed25a4448a7d545648ae4cea52b493918cc4d3b0c5adc56bb2510c0838c80
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:eae390ec42c3582f7896a8346df8782b287d64440794983ae93755490639e21e
        version: 1.5.1
      entries:
      - name: multicluster-global-hub-operator-rh.v1.5.1
        version: 1.5.1
      - name: multicluster-global-hub-operator-rh.v1.5.0
        version: 1.5.0
      name: release-1.5
    defaultChannel: release-1.5
    packageName: multicluster-global-hub-operator-rh
    provider:
      name: Red Hat, Inc
      url: https://github.com/stolostron/multicluster-global-hub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: https://www.turbonomic.com/
    name: t8c-tsc-client-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-tsc-client-certified.v1.4.2-candidate-20250806145822000
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "TurbonomicClient",
                "metadata": {
                  "name": "turbonomicclient-release"
                },
                "spec": {
                  "global": {
                    "version": "8.16.1"
                  },
                  "probes": {
                    "actionScript": {
                      "enabled": false
                    },
                    "actionStreamKafka": {
                      "enabled": false
                    },
                    "appDynamics": {
                      "enabled": false
                    },
                    "compellent": {
                      "enabled": false
                    },
                    "datadog": {
                      "enabled": false
                    },
                    "dynatrace": {
                      "enabled": false
                    },
                    "flexera": {
                      "enabled": false
                    },
                    "hds": {
                      "enabled": false
                    },
                    "hpe3par": {
                      "enabled": false
                    },
                    "hyperFlex": {
                      "enabled": false
                    },
                    "hyperV": {
                      "enabled": false
                    },
                    "ibmStorageFlashSystem": {
                      "enabled": false
                    },
                    "instana": {
                      "enabled": false
                    },
                    "jvm": {
                      "enabled": false
                    },
                    "mssql": {
                      "enabled": false
                    },
                    "mysql": {
                      "enabled": false
                    },
                    "netApp": {
                      "enabled": false
                    },
                    "newRelic": {
                      "enabled": false
                    },
                    "nutanix": {
                      "enabled": false
                    },
                    "oneView": {
                      "enabled": false
                    },
                    "oracle": {
                      "enabled": false
                    },
                    "powerVM": {
                      "enabled": false
                    },
                    "pure": {
                      "enabled": false
                    },
                    "scaleIO": {
                      "enabled": false
                    },
                    "serviceNow": {
                      "enabled": false
                    },
                    "snmp": {
                      "enabled": false
                    },
                    "terraform": {
                      "enabled": false
                    },
                    "tomcat": {
                      "enabled": false
                    },
                    "ucs": {
                      "enabled": false
                    },
                    "vCenter": {
                      "enabled": false
                    },
                    "vmax": {
                      "enabled": false
                    },
                    "vmm": {
                      "enabled": false
                    },
                    "vplex": {
                      "enabled": false
                    },
                    "webLogic": {
                      "enabled": false
                    },
                    "webSphere": {
                      "enabled": false
                    },
                    "wmi": {
                      "enabled": false
                    },
                    "xtremIO": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "VersionManager",
                "metadata": {
                  "name": "versionmanager-release"
                },
                "spec": {
                  "url": "http://remote-nginx-tunnel:9080/cluster-manager/clusterConfiguration"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-client-operator@sha256:8db7396c956e7a04aee92e1e90491cd3eb533cebe5f5616ca43ff16d9f3ed6e2
          createdAt: "2025-08-07T07:15:28Z"
          description: Turbonomic Secure Connect for Hybrid environment management
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.4.2-candidate-20250806145822000'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TurbonomicClient is the Schema for the turbonomicclients
              API
            displayName: Turbonomic Client
            kind: TurbonomicClient
            name: turbonomicclients.clients.turbonomic.ibm.com
            version: v1alpha1
          - description: VersionManager is the Schema for the versionmanagers API
            displayName: Version Manager
            kind: VersionManager
            name: versionmanagers.clients.turbonomic.ibm.com
            version: v1alpha1
        description: "### Turbonomic Secure Connect for Hybrid environment management\nTurbonomic
          Workload Automation for Multicloud simultaneously optimizes performance,
          compliance, and cost in real-time. Workloads are precisely resourced, automatically,
          to perform while satisfying business constraints:\n* Continuous placement
          of workload across multiple clouds both on-prem and public clouds providers.\n*
          Continuous scaling for applications and the underlying infrastructure.\n*
          Capability to connect both public cloud and on-prem targets on single saas
          instance\nIt assures application performance by giving workloads the resources
          they need when they need them.\n\n### How do I start?\nTurbonomic documentation
          can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm.\nTo
          setup Turbonomic Secure Connect client, follow the instructions below:\n1.
          Click Install on the landing page\n2. Upon moving to Install Operator page,
          do not modify the default selection for any of the configuration\n3. Select
          the Installed namespace from the drop down or add a new namespace and click
          on create project to create the entered namespace.\n4. Click Install on
          the Install Operator page.\nAt this point operator is installed.\n\nOn the
          Details tab, under Provided APIs: \n1. Click Create instance for the Turbonomic
          Client API. Enable/Disable the probes based on the requirement, and click
          Create.\n2. Click Create instance for the Version Manager API. Do not modify
          the JSON/YAML file, click on Create.\n\n### How does it work?\nTurbonomic
          uses a public APIs already exposed by application and infrastructure instrumentation
          to discover and monitor your environment.\nTurbonomic determines the right
          actions that drive continuous health, including continuous placement and
          continuous scaling for applications and the underlying cluster.\nTurbonomic
          leverages the built-on orchestration provided by the application and infrastructure
          deployment tools and automates the execution of these actions to continiously
          meet the respective service level objective of each application service."
        displayName: Turbonomic Secure Connect Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - turbonomic
        - turbo secure connect
        links:
        - name: T8c Tsc Client Certified
          url: https://t8c-tsc-certified.domain
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        relatedImages:
        - icr.io/cpopen/t8c-client-operator@sha256:8db7396c956e7a04aee92e1e90491cd3eb533cebe5f5616ca43ff16d9f3ed6e2
        version: 1.4.2-candidate-20250806145822000
      entries:
      - name: t8c-tsc-client-certified.v1.4.2-candidate-20250806145822000
        version: 1.4.2-candidate-20250806145822000
      name: candidate
    - currentCSV: t8c-tsc-client-certified.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "TurbonomicClient",
                "metadata": {
                  "name": "turbonomicclient-release"
                },
                "spec": {
                  "global": {
                    "version": "8.16.1"
                  },
                  "probes": {
                    "actionScript": {
                      "enabled": false
                    },
                    "actionStreamKafka": {
                      "enabled": false
                    },
                    "appDynamics": {
                      "enabled": false
                    },
                    "compellent": {
                      "enabled": false
                    },
                    "datadog": {
                      "enabled": false
                    },
                    "dynatrace": {
                      "enabled": false
                    },
                    "flexera": {
                      "enabled": false
                    },
                    "hds": {
                      "enabled": false
                    },
                    "hpe3par": {
                      "enabled": false
                    },
                    "hyperFlex": {
                      "enabled": false
                    },
                    "hyperV": {
                      "enabled": false
                    },
                    "ibmStorageFlashSystem": {
                      "enabled": false
                    },
                    "instana": {
                      "enabled": false
                    },
                    "jvm": {
                      "enabled": false
                    },
                    "mssql": {
                      "enabled": false
                    },
                    "mysql": {
                      "enabled": false
                    },
                    "netApp": {
                      "enabled": false
                    },
                    "newRelic": {
                      "enabled": false
                    },
                    "nutanix": {
                      "enabled": false
                    },
                    "oneView": {
                      "enabled": false
                    },
                    "oracle": {
                      "enabled": false
                    },
                    "powerVM": {
                      "enabled": false
                    },
                    "pure": {
                      "enabled": false
                    },
                    "scaleIO": {
                      "enabled": false
                    },
                    "serviceNow": {
                      "enabled": false
                    },
                    "snmp": {
                      "enabled": false
                    },
                    "terraform": {
                      "enabled": false
                    },
                    "tomcat": {
                      "enabled": false
                    },
                    "ucs": {
                      "enabled": false
                    },
                    "vCenter": {
                      "enabled": false
                    },
                    "vmax": {
                      "enabled": false
                    },
                    "vmm": {
                      "enabled": false
                    },
                    "vplex": {
                      "enabled": false
                    },
                    "webLogic": {
                      "enabled": false
                    },
                    "webSphere": {
                      "enabled": false
                    },
                    "wmi": {
                      "enabled": false
                    },
                    "xtremIO": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "VersionManager",
                "metadata": {
                  "name": "versionmanager-release"
                },
                "spec": {
                  "url": "http://remote-nginx-tunnel:9080/cluster-manager/clusterConfiguration"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-client-operator@sha256:626e024327388410f7a77bd9b8e5a5df29993fcafdb0176cd0bea08b05713c61
          createdAt: "2025-08-12T23:15:08Z"
          description: Turbonomic Secure Connect for Hybrid environment management
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.4.2'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TurbonomicClient is the Schema for the turbonomicclients
              API
            displayName: Turbonomic Client
            kind: TurbonomicClient
            name: turbonomicclients.clients.turbonomic.ibm.com
            version: v1alpha1
          - description: VersionManager is the Schema for the versionmanagers API
            displayName: Version Manager
            kind: VersionManager
            name: versionmanagers.clients.turbonomic.ibm.com
            version: v1alpha1
        description: "### Turbonomic Secure Connect for Hybrid environment management\nTurbonomic
          Workload Automation for Multicloud simultaneously optimizes performance,
          compliance, and cost in real-time. Workloads are precisely resourced, automatically,
          to perform while satisfying business constraints:\n* Continuous placement
          of workload across multiple clouds both on-prem and public clouds providers.\n*
          Continuous scaling for applications and the underlying infrastructure.\n*
          Capability to connect both public cloud and on-prem targets on single saas
          instance\nIt assures application performance by giving workloads the resources
          they need when they need them.\n\n### How do I start?\nTurbonomic documentation
          can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm.\nTo
          setup Turbonomic Secure Connect client, follow the instructions below:\n1.
          Click Install on the landing page\n2. Upon moving to Install Operator page,
          do not modify the default selection for any of the configuration\n3. Select
          the Installed namespace from the drop down or add a new namespace and click
          on create project to create the entered namespace.\n4. Click Install on
          the Install Operator page.\nAt this point operator is installed.\n\nOn the
          Details tab, under Provided APIs: \n1. Click Create instance for the Turbonomic
          Client API. Enable/Disable the probes based on the requirement, and click
          Create.\n2. Click Create instance for the Version Manager API. Do not modify
          the JSON/YAML file, click on Create.\n\n### How does it work?\nTurbonomic
          uses a public APIs already exposed by application and infrastructure instrumentation
          to discover and monitor your environment.\nTurbonomic determines the right
          actions that drive continuous health, including continuous placement and
          continuous scaling for applications and the underlying cluster.\nTurbonomic
          leverages the built-on orchestration provided by the application and infrastructure
          deployment tools and automates the execution of these actions to continiously
          meet the respective service level objective of each application service."
        displayName: Turbonomic Secure Connect Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - turbonomic
        - turbo secure connect
        links:
        - name: T8c Tsc Client Certified
          url: https://t8c-tsc-certified.domain
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        relatedImages:
        - icr.io/cpopen/t8c-client-operator@sha256:626e024327388410f7a77bd9b8e5a5df29993fcafdb0176cd0bea08b05713c61
        version: 1.4.2
      entries:
      - name: t8c-tsc-client-certified.v1.4.2
        version: 1.4.2
      name: stable
    defaultChannel: stable
    packageName: t8c-tsc-client-certified
    provider:
      name: Turbonomic, Inc.
      url: https://www.turbonomic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-recyclebin-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-recyclebin-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "recyclebin.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/recyclebin-controller:1.1.1
          createdAt: "2025-09-20T01:20:03Z"
          description: AWS Recycle Bin controller is a service controller for managing
            Recycle Bin resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Rule represents the state of an AWS recyclebin Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.recyclebin.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Recycle Bin resources in AWS from within your Kubernetes cluster.

          **About Amazon Recycle Bin**

          Recycle Bin is a data recovery feature that enables you to restore accidentally deleted Amazon EBS snapshots and EBS-backed AMIs. When using Recycle Bin, if your resources are deleted, they are retained in the Recycle Bin for a time period that you specify before being permanently deleted.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Recycle Bin
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - recyclebin
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Recycle Bin Developer Resources
          url: https://docs.aws.amazon.com/ebs/latest/userguide/recycle-bin.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: recyclebin maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/recyclebin-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-recyclebin-controller.v1.1.1
        version: 1.1.1
      - name: ack-recyclebin-controller.v1.1.0
        version: 1.1.0
      - name: ack-recyclebin-controller.v1.0.13
        version: 1.0.13
      - name: ack-recyclebin-controller.v1.0.12
        version: 1.0.12
      - name: ack-recyclebin-controller.v1.0.11
        version: 1.0.11
      - name: ack-recyclebin-controller.v1.0.10
        version: 1.0.10
      - name: ack-recyclebin-controller.v1.0.9
        version: 1.0.9
      - name: ack-recyclebin-controller.v1.0.8
        version: 1.0.8
      - name: ack-recyclebin-controller.v1.0.7
        version: 1.0.7
      - name: ack-recyclebin-controller.v1.0.6
        version: 1.0.6
      - name: ack-recyclebin-controller.v1.0.5
        version: 1.0.5
      - name: ack-recyclebin-controller.v1.0.2
        version: 1.0.2
      - name: ack-recyclebin-controller.v1.0.1
        version: 1.0.1
      - name: ack-recyclebin-controller.v1.0.0
        version: 1.0.0
      - name: ack-recyclebin-controller.v0.0.3
        version: 0.0.3
      - name: ack-recyclebin-controller.v0.0.2
        version: 0.0.2
      - name: ack-recyclebin-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-recyclebin-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ian Boden
      provider-url: ""
    name: eclipse-amlen-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eclipse-amlen-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eclipse.org/v1",
                "kind": "Amlen",
                "metadata": {
                  "name": "amlen-sample"
                },
                "spec": {
                  "_amlen_image_tag": "latest",
                  "_amlen_monitor_image_tag": "latest",
                  "_amlen_name": "amlen-sample",
                  "_amlen_namespace": "osdk-test",
                  "_amlen_persistence": "false",
                  "config": "amlen-config",
                  "device_cert_issuer": {
                    "mode": "automatic",
                    "name": "amlen-selfsigned-issuer"
                  },
                  "ha": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "size": 2
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/amlen/operator:v1.0.1a
          createdAt: "2022-11-24T15:31:00Z"
          description: An operator to run the Eclipse Amlen MQTT Broker
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/eclipse/amlen
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Amlen
            name: amlens.eclipse.org
            version: v1
        description: Operator for the eclipse Amlen MQTT Broker\n\n### Supported Feature\n\nCurrently
          supports:\n* Install - simple install process\n* Upgrade - non-disruptive
          upgrade\n*High Availability - can be enabled/disabled via the CRD\n\n###
          Documentation\nThe main Amlen website has links to product documentation,
          forum, blog and the slack channel\nhttp://www.eclipse.org/amlen\n\n### Getting
          help\nThe quickest way to get help is to use the slack channel but the forum
          is monitored as well\n\nIssues can also be added to the github repository
          https://github.com/eclipse/amlen\n\n### License\nEclipse Amlen Operator
          is licensed under the [Eclipse Public License - v 2.0](https://www.eclipse.org/legal/epl-2.0/)
        displayName: eclipse-amlen-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amlen
        - mqtt
        - broker
        links:
        - name: Eclipse Amlen Operator
          url: https://eclipse.org/amlen
        maintainers:
        - email: ianboden@uk.ibm.com
          name: Ian Boden
        maturity: stable
        provider:
          name: Ian Boden
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/amlen/operator:v1.0.1a
        version: 1.0.0
      entries:
      - name: eclipse-amlen-operator.v1.0.0
        version: 1.0.0
      - name: eclipse-amlen-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: eclipse-amlen-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eclipse.org/v1",
                "kind": "Amlen",
                "metadata": {
                  "name": "amlen-sample"
                },
                "spec": {
                  "config": "amlen-config",
                  "device_cert_issuer": {
                    "mode": "automatic",
                    "name": "amlen-selfsigned-issuer"
                  },
                  "ha": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "ldap": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "size": 2
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/amlen/operator:v1.1.3
          createdAt: "2025-05-22T10:00:00Z"
          description: An operator to run the Eclipse Amlen MQTT Broker
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/eclipse-amlen/amlen
          support: Eclipse
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Amlen
            name: amlens.eclipse.org
            version: v1
        description: |-
          Operator for the eclipse Amlen MQTT Broker

          ### Supported Feature

          Currently supports:
          * Install - simple install process
          * Upgrade - non-disruptive upgrade
          * High Availability - can be enabled/disabled via the CRD

          ### Documentation
          The main Amlen website has links to product documentation, forum, blog and the slack channel
          http://www.eclipse.org/amlen

          ### Getting help
          The quickest way to get help is to use the slack channel but the forum is monitored as well
          Issues can also be added to the github repository https://github.com/eclipse-amlen/amlen

          ### License
          Eclipse Amlen Operator is licensed under [Eclipse Public License - v 2.0](https://www.eclipse.org/legal/epl-2.0/)
        displayName: eclipse-amlen-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amlen
        - mqtt
        - broker
        links:
        - name: Eclipse Amlen Operator
          url: https://eclipse.org/amlen
        maintainers:
        - email: ianboden@uk.ibm.com
          name: Ian Boden
        maturity: stable
        provider:
          name: Eclipse Amlen
          url: https://eclipse.org/amlen
        relatedImages:
        - controller:latest
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        version: 1.1.3
      entries:
      - name: eclipse-amlen-operator.v1.1.3
        version: 1.1.3
      - name: eclipse-amlen-operator.v1.1.0
        version: 1.1.0
      - name: eclipse-amlen-operator.v1.0.2
        version: 1.0.2
      - name: eclipse-amlen-operator.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: alpha
    packageName: eclipse-amlen-operator
    provider:
      name: Ian Boden
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: vault-config-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vault-config-operator.v0.8.34
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AuthEngineMount",
                "metadata": {
                  "name": "authenginemount-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kube-authengine-mount-sample",
                  "type": "kubernetes"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azureauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azureauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azureauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureCredentials": {
                    "passwordKey": "clientsecret",
                    "secret": {
                      "name": "aad-credentials"
                    },
                    "usernameKey": "clientid"
                  },
                  "clientID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "environment": "AzurePublicCloud",
                  "maxRetries": 3,
                  "maxRetryDelay": 60,
                  "path": "azure/test",
                  "resource": "https://management.azure.com/",
                  "retryDelay": 4,
                  "tenantID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azureauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azureauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azureauthenginerole-sample"
                },
                "spec": {
                  "BoundResourceGroups": [
                    "resourcegroup1",
                    "resourcegroup2"
                  ],
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "boundGroupIDs": [
                    "group1",
                    "group2"
                  ],
                  "boundLocations": [
                    "location1",
                    "location2"
                  ],
                  "boundScaleSets": [
                    "scaleset1",
                    "scaleset1"
                  ],
                  "boundServicePrincipalIDs": [
                    "sp1",
                    "sp2"
                  ],
                  "boundSubscriptionIDs": [
                    "subscription1",
                    "subscription2"
                  ],
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "dev-role",
                  "path": "azure",
                  "policies": [
                    "policy1",
                    "policy2"
                  ],
                  "tokenBoundCIDRs": [
                    "10.20.102.13",
                    "10.20.102.15"
                  ],
                  "tokenExplicitMaxTTL": "",
                  "tokenMaxTTL": "",
                  "tokenNoDefaultPolicy": false,
                  "tokenNumUses": 0,
                  "tokenPeriod": 0,
                  "tokenPolicies": [
                    "policy1",
                    "policy2"
                  ],
                  "tokenTTL": "",
                  "tokenType": ""
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureSecretEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azuresecretengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azuresecretengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azuresecretengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureCredentials": {
                    "passwordKey": "clientsecret",
                    "secret": {
                      "name": "aad-credentials"
                    },
                    "usernameKey": "clientid"
                  },
                  "clientID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "environment": "AzurePublicCloud",
                  "passwordPolicy": "",
                  "path": "azure",
                  "rootPasswordTTL": "128d",
                  "subscriptionID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "tenantID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureSecretEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azuresecretenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azuresecretenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azuresecretenginerole-sample"
                },
                "spec": {
                  "TTL": "",
                  "applicationObjectID": "",
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureGroups": "",
                  "azureRoles": "",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "maxTTL": "",
                  "name": "azure-role",
                  "path": "azure",
                  "permanentlyDelete": "",
                  "persistApp": "",
                  "signInAudience": "",
                  "tags": ""
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "CertAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "certauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "certauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "certauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "test",
                  "path": "cert"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "CertAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "certauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "certauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "certauthenginerole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "certificate": "-----BEGIN CERTIFICATE-----\nMIIDJTCCAg2gAwIBAgIUF119kmCNlnMwK6yOOpfRs4JGhAAwDQYJKoZIhvcNAQEL\nBQAwFjEUMBIGA1UEAwwLZXhhbXBsZS5jb20wHhcNMjUwMTIxMTM0MjExWhcNMjYw\nMTIxMTM0MjExWjAWMRQwEgYDVQQDDAtleGFtcGxlLmNvbTCCASIwDQYJKoZIhvcN\nAQEBBQADggEPADCCAQoCggEBALad4Sx97VyFSfOepJoXUGYljg7hL+z20pVL2/M8\n8HXAQl4qf/i45b//Qz8VogMxdg3Y9J7O/M0n9hxbL18i1gm270VjEFyzjJCyUq0R\n72WMM4Sr3OjQPnVtJoQkWnH7wR3zsWK2N4NuIPdE3Ofb48Olyachffv7xCI+MLIa\n/adcyRggM4Z5f2nBCMcXL8D/eY03qCgd/U1aaPJ0WyJTZHLL4euL0ewKlKtB1JnU\n4MrRa5mbPQ3B+OhxcaD/7BNV6hPXAe2zftEv8LcimhbVoj3XmLJ+om58YIaEMadk\nJj61xnXbf+F/iV41onzwvTPNWCSWiLowDl8P3zkabuS583sCAwEAAaNrMGkwHQYD\nVR0OBBYEFH971CwrqkTvUz2qrUncWMS5ULwKMB8GA1UdIwQYMBaAFH971CwrqkTv\nUz2qrUncWMS5ULwKMA8GA1UdEwEB/wQFMAMBAf8wFgYDVR0RBA8wDYILZXhhbXBs\nZS5jb20wDQYJKoZIhvcNAQELBQADggEBAIQzT113TQC9BqitOCOoDFCJqcM+BHmf\nITEKlEhlqubPvA40lD4wkz67RP2SdQbF1glzZW5sDTxMjpHL+8nhdmo65CJpsQGP\nyMPwnOvx7W5JWQJa1Yt+48JNhgIymDMun+xwDsjNnlBqp0Kr7HkOxPimMPWN4R6B\nydG96puDGBYqxB1ujx5seOGfGvxPBBr3h8kNg054rtIIDrUG2iUTfpNIILJrd8Q2\nshop9//3VcIpUJO7Z1brvK/Hp4DrcKBZVo7vd8n+5uCx3vBA0tjqNAXYtFnSRvDZ\n6XV+rCZecDBFkaLuJ5YgYEK15gGe6lQAez6DTVLXMAKhZ52kceefnwM=\n-----END CERTIFICATE-----\n",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "test-role",
                  "path": "cert/test"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineConfig",
                "metadata": {
                  "name": "my-postgresql-database"
                },
                "spec": {
                  "allowedRoles": [
                    "read-only"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "connectionURL": "postgresql://{{username}}:{{password}}@my-postgresql-database.test-vault-config-operator.svc:5432",
                  "path": "test-vault-config-operator/database",
                  "pluginName": "postgresql-database-plugin",
                  "rootCredentials": {
                    "passwordKey": "postgresql-password",
                    "secret": {
                      "name": "postgresql-admin-password"
                    }
                  },
                  "username": "postgres"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineRole",
                "metadata": {
                  "name": "read-only"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "creationStatements": [
                    "CREATE ROLE \"{{name}}\" WITH LOGIN PASSWORD '{{password}}' VALID UNTIL '{{expiration}}'; GRANT SELECT ON ALL TABLES IN SCHEMA public TO \"{{name}}\";"
                  ],
                  "dBName": "my-postgresql-database",
                  "defaultTTL": "1h",
                  "maxTTL": "24h",
                  "path": "test-vault-config-operator/database"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineStaticRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "databasesecretenginestaticrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "databasesecretenginestaticrole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "databasesecretenginestaticrole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "dBName": "db-name",
                  "path": "test-vault-config-operator/database",
                  "rotationPeriod": "24h",
                  "rotationStatements": [
                    "ALTER ROLE \"{{name}}\" WITH PASSWORD '{{password}}';git add"
                  ],
                  "username": "db-user"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GCPAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "gcpauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gcpauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "gcpauthengineconfig-sample"
                },
                "spec": {
                  "GCEalias": "role_id",
                  "GCEmetadata": "default",
                  "GCPCredentials": {
                    "passwordKey": "credentials",
                    "secret": {
                      "name": "gcp-serviceaccount-credentials"
                    },
                    "usernameKey": "serviceaccount"
                  },
                  "IAMalias": "default",
                  "IAMmetadata": "default",
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "customEndpoint": {},
                  "path": "azure",
                  "serviceAccount": "vault-sa"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GCPAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "gcpauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gcpauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "gcpauthenginerole-sample"
                },
                "spec": {
                  "addGroupAliases": false,
                  "allowGCEInference": false,
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "boundInstanceGroups": [],
                  "boundLabels": [],
                  "boundProjects": {},
                  "boundRegions": [],
                  "boundServiceAccounts": {},
                  "boundZones": [],
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "maxJWTExp": "",
                  "name": "gcp-iam-role",
                  "path": "gcp",
                  "policies": [],
                  "tokenBoundCIDRs": [],
                  "tokenExplicitMaxTTL": "",
                  "tokenMaxTTL": "",
                  "tokenNoDefaultPolicy": false,
                  "tokenNumUses": 0,
                  "tokenPeriod": 0,
                  "tokenPolicies": [],
                  "tokenTTL": "",
                  "tokenType": "",
                  "type": "iam"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitHubSecretEngineConfig",
                "metadata": {
                  "name": "githubsecretengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitHubSecretEngineRole",
                "metadata": {
                  "name": "githubsecretenginerole-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Group",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "group-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "group",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "group-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "metadata": {
                    "team": "team-abc"
                  },
                  "policies": [
                    "team-abc-access"
                  ],
                  "type": "external"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupAlias",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "groupalias-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "groupalias",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "groupalias-sample"
                },
                "spec": {
                  "authEngineMountPath": "kubernetes",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "groupName": "group-sample"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "JWTOIDCAuthEngineConfig",
                "metadata": {
                  "name": "azure-oidc-config"
                },
                "spec": {
                  "OIDCClientID": "000000000-0000-0000-0000-000000000",
                  "OIDCCredentials": {
                    "passwordKey": "oidc_client_secret",
                    "secret": {
                      "name": "oidccredentials"
                    },
                    "usernameKey": "oidc_client_id"
                  },
                  "OIDCDiscoveryURL": "https://login.microsoftonline.com/000000000-0000-0000-0000-000000000/v2.0",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "oidc/azuread-oidc/"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "JWTOIDCAuthEngineRole",
                "metadata": {
                  "name": "azure-oidc-dev-role"
                },
                "spec": {
                  "OIDCScopes": [
                    "https://graph.microsoft.com/.default"
                  ],
                  "allowedRedirectURIs": [
                    "http://localhost:8250/oidc/callback",
                    "http://localhost:8200/ui/vault/auth/oidc/azuread-oidc/oidc/callback"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "groupsClaim": "groups",
                  "name": "dev-role",
                  "path": "oidc-aad/azuread-oidc",
                  "roleType": "oidc",
                  "tokenPolicies": [
                    "dev",
                    "prod"
                  ],
                  "userClaim": "email"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesAuthEngineConfig",
                "metadata": {
                  "name": "kubernetesauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kube-authengine-mount-sample"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesAuthEngineRole",
                "metadata": {
                  "name": "database-engine-admin"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kubernetes",
                  "policies": [
                    "database-engine-admin"
                  ],
                  "targetNamespaces": {
                    "targetNamespaceSelector": {
                      "matchLabels": {
                        "database-engine-admin": "true"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesSecretEngineConfig",
                "metadata": {
                  "name": "kubernetessecretengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesSecretEngineRole",
                "metadata": {
                  "name": "kubernetessecretenginerole-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "LDAPAuthEngineConfig",
                "metadata": {
                  "name": "ldapauthengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "LDAPAuthEngineGroup",
                "metadata": {
                  "name": "ldapauthenginegroup-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin",
                    "serviceAccount": {
                      "name": "default"
                    }
                  },
                  "name": "test-group",
                  "path": "ldap/test"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PKISecretEngineConfig",
                "metadata": {
                  "name": "pkisecretengineconfig-sample"
                },
                "spec": {
                  "TTL": "8760h",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "commonName": "pki-vault-demo.internal.io",
                  "crlDistributionPoints": [
                    "https://vault-internal.vault.svc:8200/v1/test-vault-config-operator/pki/crl\""
                  ],
                  "issuingCertificates": [
                    "https://vault-internal.vault.svc:8200/v1/test-vault-config-operator/pki/ca"
                  ],
                  "path": "test-vault-config-operator/pki",
                  "privateKeyType": "internal",
                  "type": "root"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PKISecretEngineRole",
                "metadata": {
                  "name": "pkisecretenginerole-sample"
                },
                "spec": {
                  "allowGlobDomains": true,
                  "allowSubdomains": true,
                  "allowedDomains": [
                    "internal.io",
                    "pki-vault-demo.svc",
                    "example.com"
                  ],
                  "allowedOtherSans": "*",
                  "allowedURISans": [
                    "*-pki-vault-demo.apps.example.com"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "maxTTL": "8760h",
                  "path": "test-vault-config-operator/pki"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PasswordPolicy",
                "metadata": {
                  "name": "simple-password-policy"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "passwordPolicy": "length = 20\nrule \"charset\" {\n  charset = \"abcdefghijklmnopqrstuvwxyz\"\n}"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Policy",
                "metadata": {
                  "name": "database-engine-admin"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "policy": "# mount database secrets engines\npath \"/sys/mounts/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n  allowed_parameters = {\n    \"type\" = [\"database\"]\n    \"*\"   = []\n  }\n}\n\n# tune database secrets engines\npath \"/sys/mounts/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/tune\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n\n# Configure database secrets engines\npath \"/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/config/+\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n\n# Configure database roles\npath \"/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/roles/+\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineConfig",
                "metadata": {
                  "name": "quaysecretengineconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineRole",
                "metadata": {
                  "name": "quaysecretenginerole-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineStaticRole",
                "metadata": {
                  "name": "quaysecretenginestaticrole-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RabbitMQSecretEngineConfig",
                "metadata": {
                  "name": "rabbitmqsecretengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "rabbitMQ-engine-admin"
                  },
                  "connectionURI": "https://test.com",
                  "leaseMaxTTL": 86400,
                  "leaseTTL": 86400,
                  "path": "vault-config-operator/rabbitmq"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RabbitMQSecretEngineRole",
                "metadata": {
                  "name": "rabbitmqsecretenginerole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "rabbitmq-engine-admin"
                  },
                  "path": "vault-config-operator/rabbitmq",
                  "tags": "administrator",
                  "vhostTopics": [
                    {
                      "topics": [
                        {
                          "permissions": {
                            "configure": ".*",
                            "read": ".*",
                            "write": ".*"
                          },
                          "topicName": "my-topic"
                        },
                        {
                          "permissions": {
                            "read": ".*"
                          },
                          "topicName": "my-read-topic"
                        }
                      ],
                      "vhostName": "/"
                    }
                  ],
                  "vhosts": [
                    {
                      "permissions": {
                        "configure": ".*",
                        "read": ".*",
                        "write": ".*"
                      },
                      "vhostName": "/"
                    },
                    {
                      "permissions": {
                        "read": "my-queue",
                        "write": "my-exchange"
                      },
                      "vhostName": "my-vhost"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RandomSecret",
                "metadata": {
                  "name": "randomsecret-password"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "secret-writer"
                  },
                  "path": "test-vault-config-operator/kv",
                  "secretFormat": {
                    "passwordPolicyName": "simple-password-policy"
                  },
                  "secretKey": "password"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "SecretEngineMount",
                "metadata": {
                  "name": "database"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "path": "test-vault-config-operator",
                  "type": "database"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "VaultSecret",
                "metadata": {
                  "name": "vaultsecret-sample"
                },
                "spec": {
                  "output": {
                    "annotations": {
                      "refresh": "every-minute"
                    },
                    "labels": {
                      "app": "test-vault-config-operator"
                    },
                    "name": "randomsecret",
                    "stringData": {
                      "anotherpassword": "{{ .anotherrandomsecret.password }}",
                      "password": "{{ .randomsecret.password }}"
                    },
                    "type": "Opaque"
                  },
                  "refreshPeriod": "3m0s",
                  "vaultSecretDefinitions": [
                    {
                      "authentication": {
                        "path": "kubernetes",
                        "role": "secret-reader",
                        "serviceAccount": {
                          "name": "default"
                        }
                      },
                      "name": "randomsecret",
                      "path": "test-vault-config-operator/kv/randomsecret-password"
                    },
                    {
                      "authentication": {
                        "path": "kubernetes",
                        "role": "secret-reader",
                        "serviceAccount": {
                          "name": "default"
                        }
                      },
                      "name": "anotherrandomsecret",
                      "path": "test-vault-config-operator/kv/another-password"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/vault-config-operator@sha256:490852417fac37f3003ab7c1ee797f9035fd2905c354cc46eca5e33d2ad780ce
          createdAt: "2025-09-18T18:56:36Z"
          description: This operator provides primitives to declaratively configure
            Hashicorp Vault.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: vault-config-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/vault-config-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthEngineMount is the Schema for the authenginemounts API
            displayName: Auth Engine Mount
            kind: AuthEngineMount
            name: authenginemounts.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureAuthEngineConfig is the Schema for the azureauthengineconfigs
              API
            displayName: Azure Auth Engine Config
            kind: AzureAuthEngineConfig
            name: azureauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureAuthEngineRole is the Schema for the azureauthengineroles
              API
            displayName: Azure Auth Engine Role
            kind: AzureAuthEngineRole
            name: azureauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - kind: AzureSecretEngineConfig
            name: azuresecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureSecretEngineRole is the Schema for the azuresecretengineroles
              API
            displayName: Azure Secret Engine Role
            kind: AzureSecretEngineRole
            name: azuresecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: CertAuthEngineConfig is the Schema for the certauthengineconfigs
              API
            displayName: Cert Auth Engine Config
            kind: CertAuthEngineConfig
            name: certauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: CertAuthEngineRole is the Schema for the certauthengineroles
              API
            displayName: Cert Auth Engine Role
            kind: CertAuthEngineRole
            name: certauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineConfig is the Schema for the databasesecretengineconfigs
              API
            displayName: Database Secret Engine Config
            kind: DatabaseSecretEngineConfig
            name: databasesecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineRole is the Schema for the databasesecretengineroles
              API
            displayName: Database Secret Engine Role
            kind: DatabaseSecretEngineRole
            name: databasesecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineStaticRole is the Schema for the databasesecretenginestaticroles
              API
            displayName: Database Secret Engine Static Role
            kind: DatabaseSecretEngineStaticRole
            name: databasesecretenginestaticroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GCPAuthEngineConfig is the Schema for the gcpauthengineconfigs
              API
            displayName: GCPAuth Engine Config
            kind: GCPAuthEngineConfig
            name: gcpauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: GCPAuthEngineRole is the Schema for the gcpauthengineroles
              API
            displayName: GCPAuth Engine Role
            kind: GCPAuthEngineRole
            name: gcpauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GitHubSecretEngineConfig is the Schema for the githubsecretengineconfigs
              API
            displayName: Git Hub Secret Engine Config
            kind: GitHubSecretEngineConfig
            name: githubsecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: GitHubSecretEngineRole is the Schema for the githubsecretengineroles
              API
            displayName: Git Hub Secret Engine Role
            kind: GitHubSecretEngineRole
            name: githubsecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GroupAlias is the Schema for the groupaliases API
            displayName: Group Alias
            kind: GroupAlias
            name: groupaliases.redhatcop.redhat.io
            version: v1alpha1
          - description: Group is the Schema for the groups API
            displayName: Group
            kind: Group
            name: groups.redhatcop.redhat.io
            version: v1alpha1
          - description: JWTOIDCAuthEngineConfig is the Schema for the jwtoidcauthengineconfigs
              API
            displayName: JWTOIDCAuth Engine Config
            kind: JWTOIDCAuthEngineConfig
            name: jwtoidcauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: JWTOIDCAuthEngineRole is the Schema for the jwtoidcauthengineroles
              API
            displayName: JWTOIDCAuth Engine Role
            kind: JWTOIDCAuthEngineRole
            name: jwtoidcauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesAuthEngineConfig is the Schema for the kubernetesauthengineconfigs
              API
            displayName: Kubernetes Auth Engine Config
            kind: KubernetesAuthEngineConfig
            name: kubernetesauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesAuthEngineRole can be used to define a KubernetesAuthEngineRole
              for the kube-auth authentication method
            displayName: Kubernetes Auth Engine Role
            kind: KubernetesAuthEngineRole
            name: kubernetesauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesSecretEngineConfig is the Schema for the kubernetessecretengineconfigs
              API
            displayName: Kubernetes Secret Engine Config
            kind: KubernetesSecretEngineConfig
            name: kubernetessecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesSecretEngineRole is the Schema for the kubernetessecretengineroles
              API
            displayName: Kubernetes Secret Engine Role
            kind: KubernetesSecretEngineRole
            name: kubernetessecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: LDAPAuthEngineConfig is the Schema for the ldapauthengineconfigs
              API
            displayName: LDAPAuth Engine Config
            kind: LDAPAuthEngineConfig
            name: ldapauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: LDAPAuthEngineGroup is the Schema for the ldapauthenginegroups
              API
            displayName: LDAPAuth Engine Group
            kind: LDAPAuthEngineGroup
            name: ldapauthenginegroups.redhatcop.redhat.io
            version: v1alpha1
          - description: PasswordPolicy is the Schema for the passwordpolicies API
            displayName: Password Policy
            kind: PasswordPolicy
            name: passwordpolicies.redhatcop.redhat.io
            version: v1alpha1
          - description: PKISecretEngineConfig is the Schema for the pkisecretengineconfigs
              API
            displayName: PKISecret Engine Config
            kind: PKISecretEngineConfig
            name: pkisecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: PKISecretEngineRole is the Schema for the pkisecretengineroles
              API
            displayName: PKISecret Engine Role
            kind: PKISecretEngineRole
            name: pkisecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: Policy is the Schema for the policies API
            displayName: Policy
            kind: Policy
            name: policies.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineConfig is the Schema for the quaysecretengineconfigs
              API
            displayName: Quay Secret Engine Config
            kind: QuaySecretEngineConfig
            name: quaysecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineRole is the Schema for the quaysecretengineroles
              API
            displayName: Quay Secret Engine Role
            kind: QuaySecretEngineRole
            name: quaysecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineStaticRole is the Schema for the quaysecretenginestaticroles
              API
            displayName: Quay Secret Engine Static Role
            kind: QuaySecretEngineStaticRole
            name: quaysecretenginestaticroles.redhatcop.redhat.io
            version: v1alpha1
          - description: RabbitMQSecretEngineConfig is the Schema for the rabbitmqsecretengineconfigs
              API
            displayName: Rabbit MQSecret Engine Config
            kind: RabbitMQSecretEngineConfig
            name: rabbitmqsecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: RabbitMQSecretEngineRole is the Schema for the rabbitmqsecretengineroles
              API
            displayName: Rabbit MQSecret Engine Role
            kind: RabbitMQSecretEngineRole
            name: rabbitmqsecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: RandomSecret is the Schema for the randomsecrets API
            displayName: Random Secret
            kind: RandomSecret
            name: randomsecrets.redhatcop.redhat.io
            version: v1alpha1
          - description: SecretEngineMount is the Schema for the secretenginemounts
              API
            displayName: Secret Engine Mount
            kind: SecretEngineMount
            name: secretenginemounts.redhatcop.redhat.io
            version: v1alpha1
          - description: VaultSecret is the Schema for the vaultsecrets API
            displayName: Vault Secret
            kind: VaultSecret
            name: vaultsecrets.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator helps set up Vault Configurations. The main intent is to do so such that subsequently pods can consume the secrets made available.
          There are two main principles through all of the capabilities of this operator:

          1. high-fidelity API. The CRD exposed by this operator reflect field by field the Vault APIs. This is because we don't want to make any assumption on the kinds of configuration workflow that user will set up. That being said the Vault API is very extensive and we are starting with enough API coverage to support, we think, some simple and very common configuration workflows.
          2. attention to security (after all we are integrating with a security tool). To prevent credential leaks we give no permissions to the operator itself against Vault. All APIs exposed by this operator contains enough information to authenticate to Vault using a local service account (local to the namespace where the API exist). In other word for a namespace user to be abel to successfully configure Vault, a service account in that namespace must have been previously given the needed Vault permissions.

          Currently this operator supports the following CRDs:

          1. [Policy](https://github.com/redhat-cop/vault-config-operator#policy) Configures Vault [Policies](https://www.vaultproject.io/docs/concepts/policies)
          2. [KubernetesAuthEngineRole](https://github.com/redhat-cop/vault-config-operator#KubernetesAuthEngineRole) Configures a Vault [Kubernetes Authentication](https://www.vaultproject.io/docs/auth/kubernetes) Role
          3. [SecretEngineMount](https://github.com/redhat-cop/vault-config-operator#SecretEngineMount) Configures a Mount point for a [SecretEngine](https://www.vaultproject.io/docs/secrets)
          4. [DatabaseSecretEngineConfig](https://github.com/redhat-cop/vault-config-operator#DatabaseSecretEngineConfig) Configures a [Database Secret Engine](https://www.vaultproject.io/docs/secrets/databases) Connection
          5. [DatabaseSecretEngineRole](https://github.com/redhat-cop/vault-config-operator#DatabaseSecretEngineRole) Configures a [Database Secret Engine](https://www.vaultproject.io/docs/secrets/databases) Role
          6. [RandomSecret](https://github.com/redhat-cop/vault-config-operator#RandomSecret) Creates a random secret in a vault [kv Secret Engine](https://www.vaultproject.io/docs/secrets/kv) with one password field generated using a [PasswordPolicy](https://www.vaultproject.io/docs/concepts/password-policies)vault-config-operator
        displayName: Vault Config Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - secrets
        - gitops
        - policy
        - credentials
        links:
        - name: Vault Config Operator
          url: https://github.com/redhat-cop/vault-config-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/vault-config-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/vault-config-operator@sha256:490852417fac37f3003ab7c1ee797f9035fd2905c354cc46eca5e33d2ad780ce
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 0.8.34
      entries:
      - name: vault-config-operator.v0.8.34
        version: 0.8.34
      - name: vault-config-operator.v0.8.33
        version: 0.8.33
      - name: vault-config-operator.v0.8.32
        version: 0.8.32
      - name: vault-config-operator.v0.8.31
        version: 0.8.31
      - name: vault-config-operator.v0.8.30
        version: 0.8.30
      - name: vault-config-operator.v0.8.29
        version: 0.8.29
      - name: vault-config-operator.v0.8.28
        version: 0.8.28
      - name: vault-config-operator.v0.8.27
        version: 0.8.27
      - name: vault-config-operator.v0.8.25
        version: 0.8.25
      - name: vault-config-operator.v0.8.24
        version: 0.8.24
      - name: vault-config-operator.v0.8.23
        version: 0.8.23
      - name: vault-config-operator.v0.8.22
        version: 0.8.22
      - name: vault-config-operator.v0.8.21
        version: 0.8.21
      - name: vault-config-operator.v0.8.20
        version: 0.8.20
      - name: vault-config-operator.v0.8.19
        version: 0.8.19
      - name: vault-config-operator.v0.8.18
        version: 0.8.18
      - name: vault-config-operator.v0.8.17
        version: 0.8.17
      - name: vault-config-operator.v0.8.16
        version: 0.8.16
      - name: vault-config-operator.v0.8.15
        version: 0.8.15
      - name: vault-config-operator.v0.8.14
        version: 0.8.14
      - name: vault-config-operator.v0.8.13
        version: 0.8.13
      - name: vault-config-operator.v0.8.12
        version: 0.8.12
      - name: vault-config-operator.v0.8.11
        version: 0.8.11
      - name: vault-config-operator.v0.8.10
        version: 0.8.10
      - name: vault-config-operator.v0.8.9
        version: 0.8.9
      - name: vault-config-operator.v0.8.7
        version: 0.8.7
      - name: vault-config-operator.v0.8.5
        version: 0.8.5
      - name: vault-config-operator.v0.8.4
        version: 0.8.4
      - name: vault-config-operator.v0.8.3
        version: 0.8.3
      - name: vault-config-operator.v0.8.2
        version: 0.8.2
      - name: vault-config-operator.v0.8.0
        version: 0.8.0
      - name: vault-config-operator.v0.7.1
        version: 0.7.1
      - name: vault-config-operator.v0.6.5
        version: 0.6.5
      - name: vault-config-operator.v0.6.4
        version: 0.6.4
      - name: vault-config-operator.v0.6.3
        version: 0.6.3
      - name: vault-config-operator.v0.6.2
        version: 0.6.2
      - name: vault-config-operator.v0.6.1
        version: 0.6.1
      - name: vault-config-operator.v0.6.0
        version: 0.6.0
      - name: vault-config-operator.v0.5.1
        version: 0.5.1
      - name: vault-config-operator.v0.5.0
        version: 0.5.0
      - name: vault-config-operator.v0.4.1
        version: 0.4.1
      - name: vault-config-operator.v0.4.0
        version: 0.4.0
      - name: vault-config-operator.v0.3.1
        version: 0.3.1
      - name: vault-config-operator.v0.3.0
        version: 0.3.0
      - name: vault-config-operator.v0.2.0
        version: 0.2.0
      - name: vault-config-operator.v0.1.2
        version: 0.1.2
      - name: vault-config-operator.v0.1.1
        version: 0.1.1
      - name: vault-config-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: vault-config-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-rest-api-browser-v1
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-rest-api-browser-v1.v0.1.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rest.loadcore.keysight.com/v1alpha1",
                "kind": "Apibrowser",
                "metadata": {
                  "name": "apibrowser-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-06-29T14:08:05Z"
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Apibrowser
            name: apibrowsers.rest.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Rest API Browser Operator v1
        displayName: Keysight Rest API Browser Operator v1
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Rest Api Browser V1
          url: https://keysight-rest-api-browser-v1.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5d5061dbc2ec5082cf14b6c600fb5400b83cf91d7ccebfa80680a238d275db
        - prod.harbor.keysight.digital/loadcore/keysight-rest-api-browser-v1@sha256:3ada6a0fa630acb0fa962a11d2c272ff679dded1106c4693d2c8f5ec9ab414a4
        version: 0.1.11
      entries:
      - name: keysight-rest-api-browser-v1.v0.1.11
        version: 0.1.11
      name: alpha
    defaultChannel: alpha
    packageName: keysight-rest-api-browser-v1
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Qualcomm Technologies, Inc.
      provider-url: https://github.com/quic/aic-operator
    name: aic-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aic-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aic.quicinc.com/v1",
                "kind": "AIC",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "aic-operator",
                    "app.kubernetes.io/instance": "aic-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "aic",
                    "app.kubernetes.io/part-of": "aic-operator"
                  },
                  "name": "aic-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "false"
          containerImage: ghcr.io/quic/cloud_ai_openshift_operator@sha256:9cf581e9b5c9cd08819fab8f8b6313836a0a20098ad4c99b80d93dbcf5ab32bf
          description: Automate the management of OpenShift Operator for Qualcomm
            AI 100 accelerators.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: Qualcomm Technologies, Inc.
          repository: https://github.com/quic/aic-operator
          support: Qualcomm Technologies, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AIC is the Schema for the aics API
            displayName: AIC
            kind: AIC
            name: aics.aic.quicinc.com
            version: v1
        description: The Openshift Operator enables the Qualcomm® Cloud AI series
          of hardware on OpenShift clusters by automating the configuration and installation
          of their Linux device drivers and its Kubernetes Device Plugin.
        displayName: AIC Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AIC
        - AIC-Operator
        links:
        - name: AIC Operator
          url: https://github.com/quic/aic-operator
        maintainers:
        - email: quic_vmellemp@quicinc.com
          name: quic_vmellemp
        - email: quic_carls@quicinc.com
          name: quic_carls
        maturity: alpha
        provider:
          name: Qualcomm Technologies, Inc.
          url: https://github.com/quic/aic-operator
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:7a943db9bcb22f817d75341621472f4bb30abc5625c53b9a781d97d431bc4d09
        - ghcr.io/quic/cloud_ai_openshift_operator@sha256:9cf581e9b5c9cd08819fab8f8b6313836a0a20098ad4c99b80d93dbcf5ab32bf
        version: 0.3.1
      entries:
      - name: aic-operator.v0.3.1
        version: 0.3.1
      name: alpha
    defaultChannel: alpha
    packageName: aic-operator
    provider:
      name: Qualcomm Technologies, Inc.
      url: https://github.com/quic/aic-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kong Inc.
      provider-url: https://konghq.com
    name: kong-gateway-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kong-gateway-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "ControlPlane",
                "metadata": {
                  "name": "controlplane-example"
                },
                "spec": {
                  "dataplane": "dataplane-example",
                  "gatewayClass": "kong"
                }
              },
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "DataPlane",
                "metadata": {
                  "name": "dataplane-example"
                },
                "spec": {
                  "env": [
                    {
                      "name": "KONG_LOG_LEVEL",
                      "value": "debug"
                    }
                  ]
                }
              },
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "GatewayConfiguration",
                "metadata": {
                  "name": "kong",
                  "namespace": "default"
                },
                "spec": {
                  "controlPlaneDeploymentOptions": {
                    "env": [
                      {
                        "name": "CONTROLLER_LOG_LEVEL",
                        "value": "debug"
                      }
                    ]
                  },
                  "dataPlaneDeploymentOptions": {
                    "containerImage": "kong",
                    "version": "2.8"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking,Integration & Delivery
          operatorframework.io/suggested-namespace: kong-system
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ControlPlane is the Schema for the controlplanes API
            displayName: Control Plane
            kind: ControlPlane
            name: controlplanes.gateway-operator.konghq.com
            version: v1alpha1
          - description: DataPlane is the Schema for the dataplanes API
            displayName: Data Plane
            kind: DataPlane
            name: dataplanes.gateway-operator.konghq.com
            version: v1alpha1
          - description: GatewayConfiguration is the Schema for the gatewayconfigurations
              API
            displayName: Gateway Configuration
            kind: GatewayConfiguration
            name: gatewayconfigurations.gateway-operator.konghq.com
            version: v1alpha1
        description: Kong Operator for Kubernetes. Technical preview.
        displayName: Kong
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - gateway
        - ingress
        - kong
        - kubernetes
        - go
        links:
        - name: Github repository for the operator
          url: https://github.com/kong/gateway-operator
        maintainers:
        - email: support@konghq.com
          name: Kong Inc.
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Kong Inc.
          url: https://konghq.com
        relatedImages:
        - ghcr.io/kong/gateway-operator@sha256:76b227715a3e1915a28f8aaa1e2ff8495349d29d0d61e7b85a72d74b03fbcc7f
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.connect.redhat.com/kong/kong@sha256:54e7c7e57bf93810a7cc36327023b8699f754c4c40c5c289548d54170caa5db2
        - registry.connect.redhat.com/kong/kong-ingress-controller@sha256:f691db070e7c604da9e163800903d0942ec6fda879eb9dcd5d8e90618ae787da
        version: 0.0.2
      entries:
      - name: kong-gateway-operator.v0.0.2
        version: 0.0.2
      - name: kong-gateway-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: kong-gateway-operator
    provider:
      name: Kong Inc.
      url: https://konghq.com
kind: List
metadata:
  resourceVersion: ""
apiVersion: v1
items:
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app: kube-green
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: kube-green Community
      provider-url: https://kube-green.dev
    name: kube-green
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kube-green.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kube-green.com/v1alpha1",
                "kind": "SleepInfo",
                "metadata": {
                  "labels": {
                    "app": "kube-green"
                  },
                  "name": "sleepinfo-sample"
                },
                "spec": {
                  "excludeRef": [
                    {
                      "apiVersion": "apps/v1",
                      "kind": "Deployment",
                      "name": "api-gateway"
                    }
                  ],
                  "sleepAt": "20:00",
                  "suspendCronJobs": true,
                  "timeZone": "Europe/Rome",
                  "wakeUpAt": "08:00",
                  "weekdays": "1-5"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          containerImage: docker.io/kubegreen/kube-green:0.7.1
          createdAt: "2025-06-29T15:45:17Z"
          description: Suspend your pods when no-one's using them to save energy and
            restore when necessary
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kube-green/kube-green
          support: kube-green Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SleepInfo is the Schema for the sleepinfos API
            displayName: SleepInfo
            kind: SleepInfo
            name: sleepinfos.kube-green.com
            version: v1alpha1
        description: |
          ## About this Operator

          [kube-green](https://kube-green.dev) shuts down (some of) your resources when you don't need them, reducing the consumption of resources of the cluster.

          Install this operator and apply the SleepInfo CRD to stop your dev namespaces from running during weekends, nights or when it is not necessary.
          See the [docs](https://kube-green.dev/docs/getting-started/) for more information.
          This operator is responsible to suspend and restart pods at specific times, defined by the SleepInfo CRD.

          An example of SleepInfo:

          ```yaml
          apiVersion: kube-green.com/v1alpha1
          kind: SleepInfo
          metadata:
            name: working-hours
          spec:
            weekdays: "1-5"
            sleepAt: "20:00"
            wakeUpAt: "08:00"
            timeZone: "Europe/Rome"
          ```

          This configuration stop pods everyday from monday to friday at 20:00 and restore them at 8:00.
          In this way, your cluster nodes could downscale, producing less CO2 and saving money.
        displayName: kube-green
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - green
        - green software
        - resources
        - downscale
        - autoscale
        - climate change
        links:
        - name: Homepage
          url: https://kube-green.dev
        - name: Source Code
          url: https://github.com/kube-green/kube-green
        maintainers:
        - email: bianchidavide12+kubegreen@gmail.com
          name: Davide
        maturity: alpha
        minKubeVersion: 1.27.0
        provider:
          name: kube-green Community
          url: https://kube-green.dev
        relatedImages:
        - docker.io/kubegreen/kube-green:0.7.1
        version: 0.7.1
      entries:
      - name: kube-green.v0.7.1
        version: 0.7.1
      - name: kube-green.v0.7.0
        version: 0.7.0
      - name: kube-green.v0.6.0
        version: 0.6.0
      - name: kube-green.v0.5.2
        version: 0.5.2
      - name: kube-green.v0.5.1
        version: 0.5.1
      - name: kube-green.v0.5.0
        version: 0.5.0
      - name: kube-green.v0.4.1
        version: 0.4.1
      - name: kube-green.v0.4.0
        version: 0.4.0
      - name: kube-green.v0.3.1
        version: 0.3.1
      - name: kube-green.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: kube-green
    provider:
      name: kube-green Community
      url: https://kube-green.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeRocketCI
      provider-url: https://docs.kuberocketci.io
    name: nexus-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-operator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexus-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexus",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexus-sample"
                },
                "spec": {
                  "secret": "nexus-secret",
                  "url": "https://nexus-example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRepository",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrepository-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrepository",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrepository-sample"
                },
                "spec": {
                  "go": {
                    "proxy": {
                      "cleanup": {
                        "policyNames": [
                          "cleanup-policy"
                        ]
                      },
                      "httpClient": {
                        "autoBlock": true,
                        "blocked": true,
                        "connection": {
                          "retries": 1440,
                          "timeout": 1440,
                          "userAgentSuffix": "user-agent-suffix"
                        }
                      },
                      "name": "go-proxy",
                      "negativeCache": {
                        "enabled": true,
                        "timeToLive": 1440
                      },
                      "online": true,
                      "proxy": {
                        "contentMaxAge": 1440,
                        "metadataMaxAge": 1440,
                        "remoteUrl": "https://test.com"
                      },
                      "routingRule": "routing-rule",
                      "storage": {
                        "blobStoreName": "blob-store-name",
                        "strictContentTypeValidation": true
                      }
                    }
                  },
                  "nexusRef": {
                    "name": "nexus-sample"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrole",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrole-sample"
                },
                "spec": {
                  "description": "test-role",
                  "id": "test-role",
                  "name": "test-role",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "privileges": [
                    "nx-blobstores-all"
                  ]
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexususer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexususer",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexususer-sample"
                },
                "spec": {
                  "email": "test@gmail.com",
                  "firstName": "John1",
                  "id": "test-user",
                  "lastName": "Doe",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "roles": [
                    "nx-admin"
                  ],
                  "secret": "$user-secret:password"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/nexus-operator:3.0.0
          createdAt: "2023-12-11T12:54:50Z"
          description: An Operator for managing Nexus entities.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-nexus-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus is the Schema for the nexus API.
            displayName: Nexus
            kind: Nexus
            name: nexuses.edp.epam.com
            version: v1alpha1
          - description: NexusRepository is the Schema for the nexusrepositories API.
            displayName: Nexus Repository
            kind: NexusRepository
            name: nexusrepositories.edp.epam.com
            version: v1alpha1
          - description: NexusRole is the Schema for the nexusroles API.
            displayName: Nexus Role
            kind: NexusRole
            name: nexusroles.edp.epam.com
            version: v1alpha1
          - description: NexusUser is the Schema for the nexususers API.
            displayName: Nexus User
            kind: NexusUser
            name: nexususers.edp.epam.com
            version: v1alpha1
        description: |
          Nexus Operator is an operator that is responsible for establishing
          a connection to provided NexusRepository Server, reconciling Nexus entities (groups,
          users, repositories, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Nexus Repository.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  nexus-secret
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Nexus` with Nexus instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1alpha1
              kind: Nexus
              metadata:
                name: nexus-sample
              spec:
                secret: nexus-secret                # Secret name
                url: https://nexus.example.com      # Nexus URL
              ```

              Wait for the `.status` field with  `status.connected: true`
        displayName: EDP Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - edp
        - nexus
        - operator
        - artifact
        - storage
        links:
        - name: Nexus Operator
          url: https://github.com/epam/edp-nexus-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/nexus-operator:3.0.0
        version: 3.0.0
      entries:
      - name: nexus-operator.v3.0.0
        version: 3.0.0
      name: alpha
    - currentCSV: nexus-operator.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexus-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexus",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexus-sample"
                },
                "spec": {
                  "secret": "nexus-secret",
                  "url": "https://nexus-example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusBlobStore",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusblobstore-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusblobstore",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusblobstore-sample"
                },
                "spec": {
                  "file": {
                    "path": "storage-blobstore/nexusblobstore-sample"
                  },
                  "name": "nexusblobstore-sample",
                  "nexusRef": {
                    "name": "nexus-sample"
                  },
                  "softQuota": {
                    "limit": 1000,
                    "type": "spaceUsedQuota"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusCleanupPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexuscleanuppolicy-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexuscleanuppolicy",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexuscleanuppolicy-sample"
                },
                "spec": {
                  "criteria": {
                    "lastBlobUpdated": 30
                  },
                  "description": "Cleanup policy for go",
                  "format": "go",
                  "name": "go-cleanup-policy",
                  "nexusRef": {
                    "name": "nexus-sample"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRepository",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrepository-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrepository",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrepository-sample"
                },
                "spec": {
                  "go": {
                    "proxy": {
                      "cleanup": {
                        "policyNames": [
                          "cleanup-policy"
                        ]
                      },
                      "httpClient": {
                        "autoBlock": true,
                        "blocked": true,
                        "connection": {
                          "retries": 1440,
                          "timeout": 1440,
                          "userAgentSuffix": "user-agent-suffix"
                        }
                      },
                      "name": "go-proxy",
                      "negativeCache": {
                        "enabled": true,
                        "timeToLive": 1440
                      },
                      "online": true,
                      "proxy": {
                        "contentMaxAge": 1440,
                        "metadataMaxAge": 1440,
                        "remoteUrl": "https://test.com"
                      },
                      "routingRule": "routing-rule",
                      "storage": {
                        "blobStoreName": "blob-store-name",
                        "strictContentTypeValidation": true
                      }
                    }
                  },
                  "nexusRef": {
                    "name": "nexus-sample"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrole",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrole-sample"
                },
                "spec": {
                  "description": "test-role",
                  "id": "test-role",
                  "name": "test-role",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "privileges": [
                    "nx-blobstores-all"
                  ]
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusScript",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusscript-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusscript",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusscript-sample"
                },
                "spec": {
                  "content": "security.setAnonymousAccess(Boolean.valueOf(args))\n",
                  "name": "anonymous",
                  "nexusRef": {
                    "name": "nexus-sample"
                  },
                  "payload": "true"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexususer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexususer",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexususer-sample"
                },
                "spec": {
                  "email": "test@gmail.com",
                  "firstName": "John1",
                  "id": "test-user",
                  "lastName": "Doe",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "roles": [
                    "nx-admin"
                  ],
                  "secret": "$user-secret:password"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/nexus-operator:3.5.0
          createdAt: "2025-04-22T11:21:24Z"
          description: An Operator for managing Nexus entities.
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/epam/edp-nexus-operator
          support: KubeRocketCI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NexusBlobStore is the Schema for the nexusblobstores API.
            displayName: Nexus Blob Store
            kind: NexusBlobStore
            name: nexusblobstores.edp.epam.com
            version: v1alpha1
          - description: NexusCleanupPolicy is the Schema for the cleanuppolicies
              API.
            displayName: Nexus Cleanup Policy
            kind: NexusCleanupPolicy
            name: nexuscleanuppolicies.edp.epam.com
            version: v1alpha1
          - description: Nexus is the Schema for the nexus API.
            displayName: Nexus
            kind: Nexus
            name: nexuses.edp.epam.com
            version: v1alpha1
          - description: NexusRepository is the Schema for the nexusrepositories API.
            displayName: Nexus Repository
            kind: NexusRepository
            name: nexusrepositories.edp.epam.com
            version: v1alpha1
          - description: NexusRole is the Schema for the nexusroles API.
            displayName: Nexus Role
            kind: NexusRole
            name: nexusroles.edp.epam.com
            version: v1alpha1
          - description: NexusScript is the Schema for the nexusscripts API.
            displayName: Nexus Script
            kind: NexusScript
            name: nexusscripts.edp.epam.com
            version: v1alpha1
          - description: NexusUser is the Schema for the nexususers API.
            displayName: Nexus User
            kind: NexusUser
            name: nexususers.edp.epam.com
            version: v1alpha1
        description: |
          Nexus Operator is an operator that is responsible for establishing
          a connection to provided NexusRepository Server, reconciling Nexus entities (groups,
          users, repositories, etc) according to the created CRs.

          ## Quick Start

          1. Login into Nexus and create user. Attach permissions to user such as scripts, rules, blobs etc. Insert user credentials into Kubernetes secret.

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name: nexus-admin-password
              data:
                password: cGFzcw==  # base64-encoded value of "pass"
                user:     dXNlcg==  # base64-encoded value of "user"
              ```

          2. Create Custom Resource `kind: Nexus` with Nexus instance URL and secret created on the previous step:

              ```yaml
              apiVersion: edp.epam.com/v1alpha1
              kind: Nexus
              metadata:
                name: nexus
              spec:
                secret: nexus-admin-password
                url: http://nexus.example.com
              ```

              Wait for the `.status` field with  `status.connected: true`

          3. Create Role using Custom Resources NexusRole:

              ```yaml
              apiVersion: edp.epam.com/v1alpha1
              kind: NexusRole
              metadata:
                name: edp-admin
              spec:
                description: Read and write access to all repos and scripts
                id: edp-admin
                name: edp-admin
                nexusRef:
                  kind: Nexus
                  name: nexus
                privileges:
                  - nx-apikey-all
                  - nx-repository-view-*-*-add
                  - nx-repository-view-*-*-browse
                  - nx-repository-view-*-*-edit
                  - nx-repository-view-*-*-read
                  - nx-script-*-add
                  - nx-script-*-delete
                  - nx-script-*-run
                  - nx-search-read
              ```
        displayName: EDP Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - artifact
        - edp
        - kuberocketci
        - nexus
        - operator
        - storage
        links:
        - name: Nexus Operator
          url: https://github.com/epam/edp-nexus-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: KubeRocketCI
          url: https://docs.kuberocketci.io
        relatedImages:
        - docker.io/epamedp/nexus-operator:3.5.0
        version: 3.5.0
      entries:
      - name: nexus-operator.v3.5.0
        version: 3.5.0
      - name: nexus-operator.v3.4.0
        version: 3.4.0
      - name: nexus-operator.v3.3.0
        version: 3.3.0
      name: stable
    defaultChannel: stable
    packageName: nexus-operator
    provider:
      name: KubeRocketCI
      url: https://docs.kuberocketci.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp, Inc.
      provider-url: https://www.netapp.com/
    name: trident-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trident-operator.v25.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "trident.netapp.io/v1",
                "kind": "TridentOrchestrator",
                "metadata": {
                  "name": "trident"
                },
                "spec": {
                  "IPv6": false,
                  "debug": true,
                  "nodePrep": [],
                  "imagePullSecrets": [],
                  "imageRegistry": "",
                  "namespace": "trident",
                  "silenceAutosupport": false
                }
              }
            ]
          categories: Storage
          containerImage: docker.io/netapp/trident-operator@sha256:c3448b6847550f45849a5cdbb9e0cedb35290d21131bc513d833d95eb42a43e0
          createdAt: "2025-09-03"
          description: Trident Operator, to manage NetApp Trident installations
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          operators.openshift.io/valid-subscription: No subscriptions are required
          repository: https://github.com/netapp/trident
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Used to deploy NetApp Trident.
            displayName: Trident Orchestrator
            kind: TridentOrchestrator
            name: tridentorchestrators.trident.netapp.io
            version: v1
          - description: Automates AWS FSxN backend configuration
            displayName: Trident Configurator
            kind: TridentConfigurator
            name: tridentconfigurators.trident.netapp.io
            version: v1
        description: |
          NetApp Trident is an open source storage provisioner and orchestrator maintained by NetApp. It enables you to create storage volumes for containerized applications managed by Docker and Kubernetes. For full release information, including patch release changes, see https://docs.netapp.com/us-en/trident/trident-rn.html.
        displayName: NetApp Trident
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - NetApp
        - ONTAP
        - Trident
        - storage
        - performance
        - open source
        - Astra
        - Cloud Native
        - AWS
        - Azure
        - GCP
        - Cloud Volumes
        links:
        - name: GitHub Repository
          url: https://github.com/NetApp/trident
        - name: Trident documentation
          url: https://docs.netapp.com/us-en/trident/index.html
        - name: Support policy
          url: https://mysupport.netapp.com/site/info/version-support
        - name: Release Notes
          url: https://docs.netapp.com/us-en/trident/trident-rn.html
        maintainers:
        - email: ng-trident-openshift-operator@netapp.com
          name: NetApp
        provider:
          name: NetApp, Inc.
          url: https://www.netapp.com/
        relatedImages:
        - docker.io/netapp/trident-operator@sha256:c3448b6847550f45849a5cdbb9e0cedb35290d21131bc513d833d95eb42a43e0
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:b88958c38883f5b1e3d307da90aac34a5ceafdb6bd3d989b062cb24ecafd1fbd
        - docker.io/netapp/trident-autosupport@sha256:21b3afbebfac3cc28c4eb319c2d385187bd8ebb728d8cfd82dbc437cc9b83b33
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:bdee8df520467b30312e7b14d7b6ef3963557ccd38269913de3fa51bb1bceec9
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:145ab76f9606e9d51dcd1e4e159c67021ecf53c24d8d80ed7bbeb2cc7fb5c42b
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6781d64bfb01e252595130eec814b9d2fe6426720a003cc61e82e9963767bfd4
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:b48bd5bc612c1bbe73d203c70e6caaae2b5360bd31eba20eb06cbd2102cc34fc
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:0ef3bf2a4dcbb5aa829ce9cc5f12efaccaf99b246b4b15abbb66b0135bec4cab
        - docker.io/netapp/trident@sha256:7925717979a84f93b62a2592682cde3443d6350ec4e88f8fd1a7d3ea812e8900
        version: 25.6.2
      entries:
      - name: trident-operator.v25.6.2
        version: 25.6.2
      - name: trident-operator.v25.6.1
        version: 25.6.1
      - name: trident-operator.v25.6.0
        version: 25.6.0
      - name: trident-operator.v25.2.1
        version: 25.2.1
      - name: trident-operator.v25.2.0
        version: 25.2.0
      name: stable
    defaultChannel: stable
    packageName: trident-operator
    provider:
      name: NetApp, Inc.
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: xspc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: xspc-operator.v1.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.hpe.com/v1",
                "kind": "XSPC",
                "metadata": {
                  "name": "xspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.connect.redhat.com/hpestorage/xspc-csi-driver-operator@sha256:8b8724998a8301057a3c4187c7b94bf4484aeb807065a12f84142cf844ea5534
          createdAt: "2024-12-17T03:53:30Z"
          description: An operator for managing HPE XP Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XSPC is the Schema for the xspcs API
            displayName: XSPC
            kind: XSPC
            name: xspcs.csi.hpe.com
            version: v1
        description: |-
          ## About
          HPE XP Storage Plug-in for Containers (XSPC) is a plugin that integrates HPE XP storage into Kubernetes based clusters.
          XSPC provides persistent volume provisioning capabilities from HPE's block storage arrays.
        displayName: HPE XP Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - XP7
        - XP8
        - HPE
        - Storage
        - Container Storage Interface
        - CSI
        links:
        - name: Knowledge Base
          url: https://www.hpe.com/us/en/storage/enterprise-xp-storage.html
        maintainers:
        - email: hpexpci@hpe.com
          name: HPE
        maturity: stable
        provider:
          name: HPE
        relatedImages:
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.connect.redhat.com/hpestorage/xspc-csi-driver@sha256:bdeac71c0b54c2779910a8d8faea03c1122000127704f17ceadba4938b685353
        - registry.connect.redhat.com/hpestorage/xspc-csi-driver-operator@sha256:8b8724998a8301057a3c4187c7b94bf4484aeb807065a12f84142cf844ea5534
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        version: 1.15.0
      entries:
      - name: xspc-operator.v1.15.0
        version: 1.15.0
      name: stable
    defaultChannel: stable
    packageName: xspc-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Entando, Inc
      provider-url: ""
    name: entando-k8s-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: entando-k8s-operator.v6.3.2-pr3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: beta
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - docker.io/entando/entando-k8s-service@sha256:775391f0005e64764288bfe007d9ea576f13618c02ea400d030a7502cfb551bd
        - docker.io/entando/entando-k8s-database-service-controller@sha256:58ca4ad2fca83a715aba9b68609f9056f633d556eb9845b3aff5c05769369539
        - docker.io/entando/app-builder@sha256:1ca90c09302b9cadf8950278d92ad4d7f3a14f02dea8704320794aa77ada8631
        - docker.io/entando/entando-k8s-app-controller@sha256:c68d3b129fc625f5d1ee10ed2a107e28c145049d4b513dcdb2c736555025e339
        - docker.io/entando/entando-k8s-plugin-controller@sha256:a20c753dda86e0f54f210e446b404733045ba426fc7186dc3936daed8224fbcb
        - docker.io/entando/entando-component-manager@sha256:668191fc86b090452baaba18ec7bbc3d07df5a99750e0c5b98759599d565cfbe
        - docker.io/entando/entando-redhat-sso@sha256:3692adb3694b3bf278bed7a678c9d0eaa0aba8e3e3ed4117ca04156525bcac72
        - registry.redhat.io/rhel8/mysql-80@sha256:570bce7dc649ac7ac85e7ecb196aabc281c77a94590aa7df7d9746869ce757ea
        - docker.io/entando/entando-k8s-dbjob@sha256:8ae47dbf93201965106663299f06daad544982fa244d4a0f365a1b717727e5ed
        - docker.io/entando/entando-k8s-composite-app-controller@sha256:4bc6797bb41b3e407b64530c3cff7a78f3fc2c22863caad9c606ccd650310b61
        - docker.io/entando/entando-k8s-app-plugin-link-controller@sha256:aa7ba0861705213434e610530c5bf5da1bdb33f1834d3c346174f5a28145c0bd
        - docker.io/entando/entando-de-app-eap@sha256:1f0e35d7f7ee8123721cf8e80ac96ac2f594f21a2b68cf9e720917d090c8095e
        - docker.io/entando/entando-k8s-keycloak-controller@sha256:f1ed42001260802353197433cea7df4c2d0e5fd9bcd926bb5ae7c654dfad2ffe
        - registry.redhat.io/rhel8/postgresql-12@sha256:f4e5c728b644bf1888ec8086424852ed74b5596a511be29e636fb10218fc9b6f
        - entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        - docker.io/entando/entando-k8s-cluster-infrastructure-controller@sha256:86dbea6696349d7b2393e12bacef4d1acec2dd3f8551e397fa26c92706a18dd3
        version: 6.3.2-pr3
      entries:
      - name: entando-k8s-operator.v6.3.2-pr3
        version: 6.3.2-pr3
      name: alpha
    - currentCSV: entando-k8s-operator.v6.3.2-pr3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: beta
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - docker.io/entando/entando-redhat-sso@sha256:3692adb3694b3bf278bed7a678c9d0eaa0aba8e3e3ed4117ca04156525bcac72
        - entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        - docker.io/entando/entando-de-app-eap@sha256:1f0e35d7f7ee8123721cf8e80ac96ac2f594f21a2b68cf9e720917d090c8095e
        - docker.io/entando/entando-k8s-composite-app-controller@sha256:4bc6797bb41b3e407b64530c3cff7a78f3fc2c22863caad9c606ccd650310b61
        - docker.io/entando/entando-k8s-app-controller@sha256:c68d3b129fc625f5d1ee10ed2a107e28c145049d4b513dcdb2c736555025e339
        - docker.io/entando/entando-k8s-plugin-controller@sha256:a20c753dda86e0f54f210e446b404733045ba426fc7186dc3936daed8224fbcb
        - docker.io/entando/entando-component-manager@sha256:668191fc86b090452baaba18ec7bbc3d07df5a99750e0c5b98759599d565cfbe
        - docker.io/entando/entando-k8s-app-plugin-link-controller@sha256:aa7ba0861705213434e610530c5bf5da1bdb33f1834d3c346174f5a28145c0bd
        - docker.io/entando/entando-k8s-database-service-controller@sha256:58ca4ad2fca83a715aba9b68609f9056f633d556eb9845b3aff5c05769369539
        - docker.io/entando/app-builder@sha256:1ca90c09302b9cadf8950278d92ad4d7f3a14f02dea8704320794aa77ada8631
        - docker.io/entando/entando-k8s-keycloak-controller@sha256:f1ed42001260802353197433cea7df4c2d0e5fd9bcd926bb5ae7c654dfad2ffe
        - registry.redhat.io/rhel8/mysql-80@sha256:570bce7dc649ac7ac85e7ecb196aabc281c77a94590aa7df7d9746869ce757ea
        - registry.redhat.io/rhel8/postgresql-12@sha256:f4e5c728b644bf1888ec8086424852ed74b5596a511be29e636fb10218fc9b6f
        - docker.io/entando/entando-k8s-cluster-infrastructure-controller@sha256:86dbea6696349d7b2393e12bacef4d1acec2dd3f8551e397fa26c92706a18dd3
        - docker.io/entando/entando-k8s-dbjob@sha256:8ae47dbf93201965106663299f06daad544982fa244d4a0f365a1b717727e5ed
        - docker.io/entando/entando-k8s-service@sha256:775391f0005e64764288bfe007d9ea576f13618c02ea400d030a7502cfb551bd
        version: 6.3.2-pr3
      entries:
      - name: entando-k8s-operator.v6.3.2-pr3
        version: 6.3.2-pr3
      name: beta
    - currentCSV: entando-k8s-operator.v6.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: registry.hub.docker.com/entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - registry.hub.docker.com/entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        version: 6.3.2
      entries:
      - name: entando-k8s-operator.v6.3.2
        version: 6.3.2
      name: stable
    defaultChannel: stable
    packageName: entando-k8s-operator-rhmp
    provider:
      name: Entando, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: submariner.io
      provider-url: ""
    name: submariner
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: submariner.v0.17.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:067ba6e58a6fe150a212881a29c90aa7cbd22f53a6877eebc413072f73f9179a
          createdAt: "2025-06-18 01:35:57"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.16.0 <0.17.6'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.17 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:d58214791a60cd9e4366c6ff534584eedbd37386da71b868867e5e212df270f9
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:a565d8c094f3755b6c9303c5b8ae03a5883796e0fea754b0202e5792f1e472c3
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:067ba6e58a6fe150a212881a29c90aa7cbd22f53a6877eebc413072f73f9179a
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:b76998e8bbef1d06aa8dbc49af4f42a3c5603d480a67e46baf11168514c7759a
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:56848aefe1d156147ab73e7353017b2663f63e2e01fe43ede57abc72cfc0c50f
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:e6dbc4bf3e8b1a3c1a7ed10eb463d134f9ed28a03c7bb523c82c5431c5c9ae90
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:5510cb1bb3e0ca5f2365f4a5fc147ca693d3c220f1a8e7835f8b13b66b957ae8
        version: 0.17.6
      entries:
      - name: submariner.v0.17.6
        version: 0.17.6
      - name: submariner.v0.17.5
        version: 0.17.5
      - name: submariner.v0.17.4
        version: 0.17.4
      - name: submariner.v0.17.2
        version: 0.17.2
      - name: submariner.v0.17.2-0.1725901206.p
        version: 0.17.2+0.1725901206.p
      - name: submariner.v0.17.1
        version: 0.17.1
      - name: submariner.v0.17.0
        version: 0.17.0
      name: stable-0.17
    - currentCSV: submariner.v0.18.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:d8d7f0147c7b65b41903e101240905703992a7bbadb33e302aa86b23000d2095
          createdAt: "2025-05-07 20:54:56"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.17.0 <0.18.5'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.18 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:e5c16f0d720e15d54781870a5e63c85540256355009584f7d1bd03ef20a2eae4
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:00072fea3fdb5e294a01da0ee44165fc52219a965a27bdf7d792276f6c3a36c3
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:e8366433d1f50e2062b86d7f749f17cd6566197560256e77f57b2c165f277733
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:a07ef1bfec7549b32ad94f2c44a168860d5f3b46fd13593c068c58bb137c77fc
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:d8d7f0147c7b65b41903e101240905703992a7bbadb33e302aa86b23000d2095
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:27f18015b8896b04457c356d03276d50c34b8c1d84142ec07342ff2c1c4e81fa
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:e913be7a440b6cc0c1dddbda93a5979fe28b99e63f12a7949966b69126fcee90
        version: 0.18.5
      entries:
      - name: submariner.v0.18.5
        version: 0.18.5
      - name: submariner.v0.18.4
        version: 0.18.4
      - name: submariner.v0.18.3
        version: 0.18.3
      - name: submariner.v0.18.2
        version: 0.18.2
      - name: submariner.v0.18.0
        version: 0.18.0
      name: stable-0.18
    - currentCSV: submariner.v0.19.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:4afaf6689e2225c5c7f6fbe002202a1cdf9444e23fd210cc2d86da71c9f78626
          createdAt: "2025-04-22 21:12:45"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.18.0 <0.19.4'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.19 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:d9d4d03013d65d11a4e8ecda411015f359e4215a3a53059a37a810acf2e3ea6c
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:4afaf6689e2225c5c7f6fbe002202a1cdf9444e23fd210cc2d86da71c9f78626
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:5b83c5889aa0d8f7838dd5afb8c901dfb0ce0fbe01175e8468ec50ddedef9223
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:70008af49a9d1ee516204715f66bd415c9caa7462f6b178dc1dc91e513605fcd
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:ef73b8964cd66ffb496191056594e0886afa9d17c03976b4159413c89ed569bf
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:5938d67a0a2653e92f0995440700512992ed82269820679f68483dcc1b2d98d9
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:1bbb624bbefbb75407071449d0144258732e219e3a7fc67ac709c3bcee560576
        version: 0.19.4
      entries:
      - name: submariner.v0.19.4
        version: 0.19.4
      - name: submariner.v0.19.3
        version: 0.19.3
      - name: submariner.v0.19.2
        version: 0.19.2
      - name: submariner.v0.19.0
        version: 0.19.0
      name: stable-0.19
    - currentCSV: submariner.v0.20.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "quay.io/submariner",
                  "version": "0.20.2"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "quay.io/submariner",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "0.20.2"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: $(SUBMARINER_OPERATOR_IMAGE)
          createdAt: "2025-09-30 14:21:54"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/submariner-io/submariner-operator
          support: Submariner
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.20 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: submariner.io
        relatedImages:
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:fd6c46b4a0570b5e7b6457c465e5a1b843ba6f9c209c1856dbebfee1068e35ea
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:70d563a6e6526b9a146a2973210b2abdae09d5acab403fa3f0b465ed5e3c9b11
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:77e0da2db6d5bb2831e6c51214e08685c9f7f3b8b1e6ac53b4fd1418644ad6a1
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:0eef6521c96b29722076c1e220770124c8ed214379abc4654d9567cff8b52d04
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:8ea02aba8f9dd901f1f0ccff97ec8d157bebe5a90a6c924affe5e01cf3454e72
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:ed521b241da60efabfe2e2b80c51639d1a82a43945945879decbfb192bbc6d23
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:4a4ac198e9d0f4a4aaed8f0ec0e1d79bd8fa712b671ee697f7ff9d5616a02d84
        version: 0.20.2
      entries:
      - name: submariner.v0.20.2
        version: 0.20.2
      - name: submariner.v0.20.1
        version: 0.20.1
      - name: submariner.v0.20.0
        version: 0.20.0
      name: stable-0.20
    - currentCSV: submariner.v0.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "quay.io/submariner",
                  "version": "v0.21.0"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "quay.io/submariner",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v0.21.0"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: $(SUBMARINER_OPERATOR_IMAGE)
          createdAt: "2025-08-13 21:57:14"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/submariner-io/submariner-operator
          support: Submariner
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.21 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: submariner.io
        relatedImages:
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:cd851eab4a2214ea7a699bd01d2abac1322d01562f27b35f58f3c0b8716f88fb
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:bcdeaa67664568c5e2a6928236ff69b175ce65ddf7a58c1664d091b926000052
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:b0ad338aed82367048cf024767b0b09e6bb09bbf7dd5128e6711e1feec6f696e
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:65765c7f08c5454b557fbda91c49efbd2cc0d4d973cf8acc7ff40ec8a381839d
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:f5909b1802986622ab169a82d5dd7877af72660903e961b3097cc29e30eb9799
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:b307f748125fd459c3474fdab645688675a06515a9924b47cedba80e038458db
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:0b74f058400f24a63cf9b57a388914cc6dda38e03d639196871f3fd76fb7f2ae
        version: 0.21.0
      entries:
      - name: submariner.v0.21.0
        version: 0.21.0
      name: stable-0.21
    defaultChannel: stable-0.21
    packageName: submariner
    provider:
      name: submariner.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Devfile
      provider-url: https://devfile.io
    name: devworkspace-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devworkspace-operator.v0.36.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/internal-objects: '["devworkspaceroutings.controller.devfile.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DevWorkspaceOperatorConfig
            name: devworkspaceoperatorconfigs.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspaceRouting
            name: devworkspaceroutings.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha2
          - kind: DevWorkspaceTemplate
            name: devworkspacetemplates.workspace.devfile.io
            version: v1alpha1
          - kind: DevWorkspaceTemplate
            name: devworkspacetemplates.workspace.devfile.io
            version: v1alpha2
        description: |
          The DevWorkspace Operator enables cluster-level support for the
          [Devfile 2.0 spec](https://docs.devfile.io), enabling static, reproducible
          configurations for launching cloud-based editors and IDEs in OpenShift and
          Kubernetes clusters.

          Leveraging the language server protocol and the Eclipse Theia web IDE, the
          DevWorkspace operator provides easy configuration of full development
          environments on the cloud with support for a wide variety of languages and
          technologies, including Go, Java, Typescript/Javascript, Python, and more.

          The DevWorkspace Operator is also used in the Web Terminal Operator to
          automatically provision Web Terminal environments.

          ## Installing the operator
          The DevWorkspace Operator can be installed directly from the OperatorHub UI and
          will be available in all namespaces on the cluster. DevWorkspace creation is
          driven by the DevWorkspace custom resource, which can be created in any
          namespace to provision a full development environment. To get started, browse
          the DevWorkspace [spec](https://devfile.io/docs/2.3.0/devfile-schema).

          Once a DevWorkspace is started, it can be accessed via the URL in its
          `.status.mainUrl` field.

          It's recommended to install the DevWorkspace Operator to the
          `openshift-operators` namespace for compatibility.

          ## Uninstalling the operator
          The DevWorkspace Operator utilizes finalizers on resources it creates and
          webhooks to restrict access to development resources. As a result, manual steps
          are required in uninstalling the operator. See the
          [documentation](https://github.com/devfile/devworkspace-operator/blob/main/docs/uninstall.md)
          for details.
        displayName: DevWorkspace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Devworkspace Operator
        - DevWorkspaces
        - Devfile
        links:
        - name: Devworkspace Operator
          url: https://github.com/devfile/devworkspace-operator
        maintainers:
        - email: dakwon@redhat.com
          name: David Kwon
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        maturity: alpha
        provider:
          name: Devfile
          url: https://devfile.io
        relatedImages:
        - registry.redhat.io/devworkspace/devworkspace-rhel9-operator@sha256:83b6abf6e57703da1125e037ea9353c224602bddc3146e5b00df47f1b1976e9a
        - registry.redhat.io/ubi9/ubi-minimal@sha256:a50731d3397a4ee28583f1699842183d4d24fadcc565c4688487af9ee4e13a44
        - registry.redhat.io/devworkspace/devworkspace-project-clone-rhel9@sha256:312a013201877cea03956e3db2ba19a3d7f842de2c3a65927c65c3e8d22a57fe
        version: 0.36.0
      entries:
      - name: devworkspace-operator.v0.36.0
        version: 0.36.0
      - name: devworkspace-operator.v0.35.1
        version: 0.35.1
      - name: devworkspace-operator.v0.34.0
        version: 0.34.0
      - name: devworkspace-operator.v0.33.0
        version: 0.33.0
      - name: devworkspace-operator.v0.32.1
        version: 0.32.1
      - name: devworkspace-operator.v0.31.2
        version: 0.31.2
      - name: devworkspace-operator.v0.31.1
        version: 0.31.1
      - name: devworkspace-operator.v0.31.0
        version: 0.31.0
      - name: devworkspace-operator.v0.30.2
        version: 0.30.2
      - name: devworkspace-operator.v0.30.1
        version: 0.30.1
      - name: devworkspace-operator.v0.30.0
        version: 0.30.0
      - name: devworkspace-operator.v0.29.0
        version: 0.29.0
      - name: devworkspace-operator.v0.28.0-0.1719314743.p
        version: 0.28.0+0.1719314743.p
      - name: devworkspace-operator.v0.28.0
        version: 0.28.0
      - name: devworkspace-operator.v0.27.0-0.1714987663.p
        version: 0.27.0+0.1714987663.p
      - name: devworkspace-operator.v0.27.0
        version: 0.27.0
      - name: devworkspace-operator.v0.26.0
        version: 0.26.0
      - name: devworkspace-operator.v0.25.0
        version: 0.25.0
      - name: devworkspace-operator.v0.24.0
        version: 0.24.0
      - name: devworkspace-operator.v0.23.0
        version: 0.23.0
      - name: devworkspace-operator.v0.22.0
        version: 0.22.0
      - name: devworkspace-operator.v0.21.2
        version: 0.21.2
      - name: devworkspace-operator.v0.21.1
        version: 0.21.1
      - name: devworkspace-operator.v0.20.0
        version: 0.20.0
      - name: devworkspace-operator.v0.19.1
        version: 0.19.1
      - name: devworkspace-operator.v0.19.1-0.1682321189.p
        version: 0.19.1+0.1682321189.p
      - name: devworkspace-operator.v0.19.1-0.1679521112.p
        version: 0.19.1+0.1679521112.p
      - name: devworkspace-operator.v0.18.1
        version: 0.18.1
      - name: devworkspace-operator.v0.18.1-0.1675929565.p
        version: 0.18.1+0.1675929565.p
      - name: devworkspace-operator.v0.17.0
        version: 0.17.0
      - name: devworkspace-operator.v0.16.0-0.1666668361.p
        version: 0.16.0+0.1666668361.p
      - name: devworkspace-operator.v0.16.0
        version: 0.16.0
      - name: devworkspace-operator.v0.15.2-0.1661828401.p
        version: 0.15.2+0.1661828401.p
      - name: devworkspace-operator.v0.15.2
        version: 0.15.2
      - name: devworkspace-operator.v0.14.1
        version: 0.14.1
      - name: devworkspace-operator.v0.13.0
        version: 0.13.0
      - name: devworkspace-operator.v0.12.0
        version: 0.12.0
      - name: devworkspace-operator.v0.11.0
        version: 0.11.0
      - name: devworkspace-operator.v0.10.0
        version: 0.10.0
      - name: devworkspace-operator.v0.9.0
        version: 0.9.0
      name: fast
    defaultChannel: fast
    packageName: devworkspace-operator
    provider:
      name: Devfile
      url: https://devfile.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microcks.io
      provider-url: ""
    name: microcks
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: microcks-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "microcks.github.io/v1alpha1",
                "kind": "MicrocksInstall",
                "metadata": {
                  "name": "my-microcksinstall"
                },
                "spec": {
                  "name": "my-microcksinstall",
                  "version": "1.10.0",
                  "microcks": {
                    "replicas": 1
                  },
                  "postman": {
                    "replicas": 1
                  },
                  "keycloak": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "1Gi"
                  },
                  "mongodb": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "2Gi"
                  }
                }
              },
              {
                "apiVersion": "microcks.github.io/v1alpha1",
                "kind": "MicrocksInstall",
                "metadata": {
                  "name": "my-microcksinstall-minikube"
                },
                "spec": {
                  "name": "my-microcksinstall-minikube",
                  "version": "1.10.0",
                  "microcks": {
                    "replicas": 2,
                    "url": "microcks.192.168.99.100.nip.io"
                  },
                  "postman": {
                    "replicas": 2
                  },
                  "keycloak": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "1Gi",
                    "url": "keycloak.192.168.99.100.nip.io"
                  },
                  "mongodb": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "2Gi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/microcks/microcks-ansible-operator:1.10.0
          createdAt: "2024-08-01T09:14:55Z"
          description: Open Source mocking and testing platform for API and microservices
          repository: https://github.com/microcks/microcks-ansible-operator
          support: microcks.github.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Microcks installation
            displayName: MicrocksInstall
            kind: MicrocksInstall
            name: microcksinstalls.microcks.github.io
            version: v1alpha1
        description: |
          Microcks is an Open source tool for API and microservices mocking and testing. It provides a platform for referencing, deploying mocks and allow contract-testing of your APIs. It can also be considered as a Service Virtualization solution because it will allow you to provide fake API or Service implementation before development being actually done. It supports Async APIs, REST APIs, gRPC, GraphQL and SOAP WebServices. It perfectly integrates into an iterative, contract-first delivery process.
          Microcks tries not reinventing the wheel and let you capitalize on already known tools of standards. It supports [AsyncAPI Specifications](https://asyncapi.com), [OpenAPI Specification](https://www.openapis.org/) `3.x`, [gRPC Protobuf](https://grpc.io/), [GraphQL Schema](https://graphql.org) out-of-the-box and can also relies on popular tools like [SoapUI](http://www.soapui.com) or [Postman](http://www.getpostman.io) to edit your API request / response samples as well as your test scripts.
          ### Operator features
          * **Create/Destroy**: Easily launch Microcks installations (`MicrocksInstall` custom resource) for your Kubernetes namespace
          ### MicrocksInstall options
          * **keycloak** component can be deployed as part of the install or you can reuse an existing instance. You can reference an existing `Secret` for providing TLS certificates for securing the `Ingress`,
          * **mongodb** component can be deployed as part of the install or you can reuse an existing instance, You can reference an existing `Secret` for providing connection credentials,
          * **features** component allow to activate optional features like the mocking of AsyncAPIs. In case you turn `features.async.enabled` to `true` and `features.async.kafka.install` to `true`, you must have the [Strimzi Operator](https://operatorhub.io/operator/strimzi-kafka-operator) already installed on your cluster/namespace.
          ### Documentation
          Documentation to the current _master_ branch as well as all releases can be found on our [GitHub](https://github.com/microcks/microcks-ansible-operator).
          ### License
          Microcks is licensed under the [Apache License, Version 2.0](https://github.com/microcks/microcks/blob/master/LICENSE).
        displayName: Microcks Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mock
        - test
        - API
        - microservice
        links:
        - name: Product Page
          url: https://microcks.io
        - name: Microcks Operator Source Code
          url: https://github.com/microcks/microcks-ansible-operator
        maintainers:
        - email: laurent.broudoux@gmail.com
          name: lbroudoux
        maturity: alpha
        provider:
          name: Microcks.io
        relatedImages:
        - quay.io/microcks/microcks-ansible-operator:1.10.0
        version: 1.10.0
      entries:
      - name: microcks-operator.v1.10.0
        version: 1.10.0
      - name: microcks-operator.v1.9.0
        version: 1.9.0
      - name: microcks-operator.v1.8.1
        version: 1.8.1
      - name: microcks-operator.v1.8.0
        version: 1.8.0
      - name: microcks-operator.v1.7.1
        version: 1.7.1
      - name: microcks-operator.v1.7.0
        version: 1.7.0
      - name: microcks-operator.v1.6.1
        version: 1.6.1
      - name: microcks-operator.v1.6.0
        version: 1.6.0
      - name: microcks-operator.v1.5.2
        version: 1.5.2
      - name: microcks-operator.v1.5.1
        version: 1.5.1
      - name: microcks-operator.v1.5.0
        version: 1.5.0
      - name: microcks-operator.v1.4.1
        version: 1.4.1
      name: stable
    defaultChannel: stable
    packageName: microcks
    provider:
      name: Microcks.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-wafv2-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-wafv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wafv2.services.k8s.aws/v1alpha1",
                "kind": "IPSet",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/wafv2-controller:1.1.1
          createdAt: "2025-09-20T21:41:21Z"
          description: AWS WAFV2 controller is a service controller for managing WAFV2
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IPSet represents the state of an AWS wafv2 IPSet resource.
            displayName: IPSet
            kind: IPSet
            name: ipsets.wafv2.services.k8s.aws
            version: v1alpha1
          - description: RuleGroup represents the state of an AWS wafv2 RuleGroup
              resource.
            displayName: RuleGroup
            kind: RuleGroup
            name: rulegroups.wafv2.services.k8s.aws
            version: v1alpha1
          - description: WebACL represents the state of an AWS wafv2 WebACL resource.
            displayName: WebACL
            kind: WebACL
            name: webacls.wafv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Web Application Firewall (Amazon WAFV2) resources in AWS from within your Kubernetes cluster.

          **About Amazon WAFV2**

          AWS WAF is a web application firewall that lets you monitor the HTTP and HTTPS requests that are forwarded to your protected web application resources. You can protect the following resource types:
          - Amazon CloudFront distribution
          - Amazon API Gateway REST API
          - Application Load Balancer
          - AWS AppSync GraphQL API
          - Amazon Cognito user pool
          - AWS App Runner service
          - AWS Verified Access instance

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon WAFV2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - wafv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon WAFV2 Developer Resources
          url: https://aws.amazon.com/waf/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: wafv2 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/wafv2-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-wafv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-wafv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-wafv2-controller.v1.0.14
        version: 1.0.14
      - name: ack-wafv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-wafv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-wafv2-controller.v1.0.10
        version: 1.0.10
      - name: ack-wafv2-controller.v1.0.9
        version: 1.0.9
      - name: ack-wafv2-controller.v1.0.8
        version: 1.0.8
      - name: ack-wafv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-wafv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-wafv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-wafv2-controller.v1.0.1
        version: 1.0.1
      - name: ack-wafv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-wafv2-controller.v0.0.4
        version: 0.0.4
      - name: ack-wafv2-controller.v0.0.3
        version: 0.0.3
      - name: ack-wafv2-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-wafv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/in-en/products/cloud-pak-for-data
    name: ibm-zen-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-zen-operator.v6.1.3
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.3 <6.1.3'
          operators.operatorframework.io/builder: operator-sdk-v1.6.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: represents zen instance
            kind: ZenService
            name: zenservices.zen.cpd.ibm.com
            version: v1
          - description: represents zen extension instance
            kind: ZenExtension
            name: zenextensions.zen.cpd.ibm.com
            version: v1
        description: |
          The Zen operator is a component in IBM® Cloud Pak foundational services. It serves a framework for user access management, defining user roles and for powering persona driven user experiences for Cloud Paks and Service offerings
        displayName: IBM Zen Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cloud Pak
        - IBM
        - Access management
        - Platform UI
        - Zen framework
        maintainers:
        - email: cp4d@ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/in-en/products/cloud-pak-for-data
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:443b51f8b10acc85bdefde7193e0f45b1c423fa7fbdcaa28342805815c43db3d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:3d5c6573faf4ae00cbb21cbfe9b2ad5cd8ffb658df0c5ba09bd98b3bf54f9979
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:fa01c7d11c23a47bb84375389e69619455c2879b5ed6bb5b9b753b945dcb02c3
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:45194b56f9fcb3af83158eb0297248364db40e8164fb4fe4f5656c58aeac3fa5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:a99b1f5a4eb0306b83562de0184fc826c34835ffbf46d929ceb9e70e0d325361
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:5f48081ffa453ec740c2562c24603d78b481b5e236d6bcc73e96203cabe153f2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:0522fc5482404fa9be179c7225a5692c8d71d755fda0b2d3928e452adcf030fb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:45ca5370b726e31c55f9d7d249374189cb48dc05ab89d675c35225d53b90e7c9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:aaee5240471872a8f1a77d4440a696d45041c5ce06849d516daccb90f5341b6d
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:abd96661823553d623eb0ea1fac8271d64391243dd0f13eb4a4fcc858dfe10a5
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:82afac013ef9b6b9d0eacb4575c979b65abd6f2821971182b90c71bd488ef366
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:b87fa8374976ccd3dad90ddbf6b65cd6ae5a028ba896562f284f089380ebe44a
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:24dff64a80b910f0b4c5442d4f611f9df7602a16c5231eff2c9d08f07b2a9f9a
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:ad027c91c719fd1d931cbac8c76735474f676b228f9ab23d1f243201246b3238
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:c6862158a2a5ae0a4e4206bfdc7f73e1c2359544966c72cf4fc3a617acb13697
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:10ede3c4e3c7e0cbc26e7b6e797cad35e39face1a1b9d4a4e2440ab0feeeb814
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:deb66a64288a370c5da5c7cc0ba4e2ba4ab816d7c5a080a0c505b05f55fa6072
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:7dbff355db7739152961cb6a97887d2e43bc960ac58837c126e03ed1a4480a3a
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:a72c2b70a9f0a93047d47247cddc1616138cbc87337289ddb208565af6e4369d
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:183ee860bf48076ea155db08032eed7004c1204a363625f54218734b00175441
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:525924934f74b28783dacf630a6401ccd4c88d253aafa2d312b7993d8b9a14c1
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:a7c050ae29cb93754f59d8f83676275ee9a91ae9a9956395b97abf8c5e20e96e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8da8b240176c00f8f6eadf32e5614c663cc1d56d546c73996cca614b715acf4a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:f06a1f50cd18d6f48762fdd0394307b44c0573d0cc91f53724c5b3c8c5e2b8e4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:c4ff5dae4560e5a943c7218df3a5aa9ca1a9ee2d1f7dfac73db05892725c34a9
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-bundle@sha256:552fc53664722903eccaa7d3782a3e84413509fd357725fe0b0f4a077c48ab2c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:719b0871f0d32da894ec80dc2a23483623b7802a5c6f935cf3672468f99faa40
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:c8baecad9c97ddbbb3f9f02c90abbf483d46974b7bc9605622a302cbeba1de74
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:b5a8fb14c3bb7c89ff970ca85f06dfcdca832e2b45ad86e8bb1b6c0a0e7e0764
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:8444703b2a7822f82141afcf4f36bf255650d755321b9e9fe2cbd1bdea41d5c7
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:e9173d81ef2a75839c0e1d677b72bb00b8e6d7ef0511afab103f08a884a03738
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:a8d2e42757c6ccbfe4372465b1827c633aad012c5194bf8b08afcd5691caa6f5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:6bbcb64cf9adb4d3b6a4c259eb910e7da5c3c58732643e60d416399b2083cc0d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:ea57c7cb1ccd5bce17f6cb35e326ab974718be8cd19b0cd38ea8112e91c51939
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ee161954766211651d7cb13f3f5a3c025ac04094ba742be453ce789c6f0e24a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:ecbc65705de76804a3c7642339ecfb4cc0f5a61891058e07c20bf2b63f1b82b8
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:1df782d0d923544767696d719bd0ac10802b8e06508a104af5e023a519f32dee
        version: 6.1.3
      entries:
      - name: ibm-zen-operator.v6.1.3
        version: 6.1.3
      name: v6.1
    defaultChannel: v6.1
    packageName: ibm-zen-operator
    provider:
      name: IBM
      url: https://www.ibm.com/in-en/products/cloud-pak-for-data
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: hpe-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-csi-operator.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.hpe.com/v1",
                "kind": "HPECSIDriver",
                "metadata": {
                  "name": "hpecsidriver-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "csp": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "disable": {
                    "alletra6000": false,
                    "alletra9000": false,
                    "alletraStorageMP": false,
                    "b10000FileService": false,
                    "nimble": false,
                    "primera": false
                  },
                  "disableHostDeletion": false,
                  "disableNodeConfiguration": false,
                  "disableNodeConformance": false,
                  "disableNodeGetVolumeStats": false,
                  "disableNodeMonitor": false,
                  "disablePreInstallHooks": false,
                  "imagePullPolicy": "IfNotPresent",
                  "images": {
                    "b10000FileServiceCSP": "quay.io/hpestorage/alletrastoragemp-b10000-nfs-csp@sha256:6a1855a2ef1a83b886bb1f45b596e8e420e5e13c84bbf446c8c202eaf27a2013",
                    "csiAttacher": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                    "csiControllerDriver": "quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676",
                    "csiExtensions": "quay.io/hpestorage/csi-extensions@sha256:27074aaa8df3b867cfa3d12318612c45f8e541d750c27db5e080734592160c89",
                    "csiNodeDriver": "quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676",
                    "csiNodeDriverRegistrar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                    "csiProvisioner": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                    "csiResizer": "registry.k8s.io/sig-storage/csi-resizer@sha256:8ddd178ba5d08973f1607f9b84619b58320948de494b31c9d7cd5375b316d6d4",
                    "csiSnapshotter": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769",
                    "csiVolumeGroupProvisioner": "quay.io/hpestorage/volume-group-provisioner@sha256:0a5f39f52113b963978a695bb8e5a5a3c4d387a02eb4d4a9647b4bc804de7399",
                    "csiVolumeGroupSnapshotter": "quay.io/hpestorage/volume-group-snapshotter@sha256:177d9abf8fbd3ac26ca4c12ca4b4adde20ebab308c79f10b04632f0171746fa8",
                    "csiVolumeMutator": "quay.io/hpestorage/volume-mutator@sha256:bf6beac1af5edcd3c9774f0dcc9579b79ef9cc8e4b74a4238082ab9aa443b5ed",
                    "nfsProvisioner": "quay.io/hpestorage/nfs-provisioner@sha256:5bf5952453ecca28e26724df85ebd5c8f0b7479ae4cb683d4a09c415da00e002",
                    "nimbleCSP": "quay.io/hpestorage/alletra-6000-and-nimble-csp@sha256:5480ffdebd74fdc547bc4a11b4434abac7bbfee3cee349f1ac893fbad84bb8d1",
                    "primera3parCSP": "quay.io/hpestorage/alletra-9000-primera-and-3par-csp@sha256:8af6e87075c47ba8d9714d55848ab2a0fef62b0768681e30d9e63ed7f396b0bf"
                  },
                  "iscsi": {
                    "chapSecretName": ""
                  },
                  "kubeletRootDir": "/var/lib/kubelet",
                  "logLevel": "info",
                  "maxVolumesPerNode": 100,
                  "node": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/hpestorage/csi-driver-operator@sha256:29e92f68cf0d7691a6c56876e7dde0d69a8cb3ff6ae6c9371c604e61cba16898
          createdAt: "2025-09-25T16:28:16Z"
          description: A Container Storage Interface (CSI) driver for HPE storage
            platforms. The CSI driver allows you to use HPE storage with your preferred
            container orchestrator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <3.0.1'
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/hpe-storage/co-deployments
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE CSI Driver for Kubernetes
            displayName: HPECSIDriver
            kind: HPECSIDriver
            name: hpecsidrivers.storage.hpe.com
            version: v1
        description: |
          The HPE CSI Operator for OpenShift deploys the HPE CSI Driver on Kubernetes and OpenShift for dynamic provisioning of persistent volumes on HPE storage systems.
          The HPE CSI Driver for Kubernetes leverages HPE storage platforms to provide scalable and persistent storage for stateful applications.
          Please refer to our CSI driver [documentation](https://scod.hpedev.io/csi_driver/index.html) for supported HPE platforms and full list of supported CSI features.
          ## Installation
          Refer to the HPE CSI Operator for OpenShift [documentation](https://scod.hpedev.io/csi_driver/deployment.html#operator).
        displayName: HPE CSI Operator for OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE CSI
        - HPE Storage
        - CSI
        links:
        - name: Documentation
          url: https://scod.hpedev.io/csi_driver/index.html
        maintainers:
        - email: dev-hi-containers@hpe.com
          name: HPE Storage Containers Team
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: HPE
        relatedImages:
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - quay.io/hpestorage/volume-mutator@sha256:bf6beac1af5edcd3c9774f0dcc9579b79ef9cc8e4b74a4238082ab9aa443b5ed
        - registry.connect.redhat.com/hpestorage/csi-driver-operator@sha256:29e92f68cf0d7691a6c56876e7dde0d69a8cb3ff6ae6c9371c604e61cba16898
        - quay.io/hpestorage/alletrastoragemp-b10000-nfs-csp@sha256:6a1855a2ef1a83b886bb1f45b596e8e420e5e13c84bbf446c8c202eaf27a2013
        - quay.io/hpestorage/alletra-9000-primera-and-3par-csp@sha256:8af6e87075c47ba8d9714d55848ab2a0fef62b0768681e30d9e63ed7f396b0bf
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - quay.io/hpestorage/nfs-provisioner@sha256:5bf5952453ecca28e26724df85ebd5c8f0b7479ae4cb683d4a09c415da00e002
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8ddd178ba5d08973f1607f9b84619b58320948de494b31c9d7cd5375b316d6d4
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769
        - quay.io/hpestorage/volume-group-provisioner@sha256:0a5f39f52113b963978a695bb8e5a5a3c4d387a02eb4d4a9647b4bc804de7399
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - quay.io/hpestorage/volume-group-snapshotter@sha256:177d9abf8fbd3ac26ca4c12ca4b4adde20ebab308c79f10b04632f0171746fa8
        - quay.io/hpestorage/csi-extensions@sha256:27074aaa8df3b867cfa3d12318612c45f8e541d750c27db5e080734592160c89
        - quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676
        - quay.io/hpestorage/alletra-6000-and-nimble-csp@sha256:5480ffdebd74fdc547bc4a11b4434abac7bbfee3cee349f1ac893fbad84bb8d1
        version: 3.0.1
      entries:
      - name: hpe-csi-operator.v3.0.1
        version: 3.0.1
      name: stable
    defaultChannel: stable
    packageName: hpe-csi-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SmartX
      provider-url: https://www.smartx.com
    name: iomesh-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: iomesh-operator.0.10.1-rc4
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
          createdAt: 2021-08-17 00:00:00+00:00
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/iomesh-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/iomesh-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/iomesh/docs
          support: SmartX, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: iomesh
            displayName: iomesh
            kind: IOMeshCluster
            name: iomeshclusters.iomesh.com
            version: v1alpha1
        description: "IOMesh (https://www.iomesh.com/) is a cloud-native storage product
          specifically designed and developed for Kubernetes. It takes ZBS as its
          core, a “production-ready” distributed storage system developed independently
          by SmartX, to provide production-level high performance and reliable persistent
          storage for business-critical applications such as MySQL, Cassandra, MongoDB,
          etc., contributing to the containerized transformation of stateful applications.
          \     \n## Pre-Requisites\n* According to https://docs.iomesh.com/deploy/prerequisites,
          make sure that your openshift cluster meets the \"Installation Requirements\"
          and set up the worker nodes correctly according to the \"Setup Worker Node\"
          chapter.\n* Run IOMesh Operator pre-installation script in an environment
          where oc or kubectl can be used to access the openshift cluster, the script
          will install the dependencies of IOMesh Operator and specific IOMesh settings
          for the openshift cluster:\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-pre-install-openshift.sh
          | sh -\n```\n## Install IOMesh Operator and IOMesh Cluster\n* Click install
          button on the current web page to install IOMesh Operator\n* On the **Installed
          Operators** -> **IOMesh Operator** -> **Create instance** -> **YAML view**,
          fill in an IOMesh Custom Resources according https://iomesh.run/iomesh.yaml,
          change the `spec.*.dataCIDR` to your own data network CIDR:\n## Post Install\n*
          Run IOMesh Operator post-installation script in an environment where oc
          or kubectl can be used to access the openshift cluster, the script will
          install IOMesh CSI Driver\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-post-install-openshift.sh
          | sh -\n```\n## Getting Started\n* Try to quickstart setup and use IOMesh:
          https://docs.iomesh.com/deploy/setup-iomesh\n"
        displayName: IOMesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iomesh-operator
        links:
        - name: IOMesh Operator
          url: https://iomesh.com
        maintainers:
        - email: na.wang@smartx.com
          name: Na Wang
        maturity: alpha
        provider:
          name: SmartX
          url: https://www.smartx.com
        relatedImages:
        - iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
        version: 0.10.1-rc4
      entries:
      - name: iomesh-operator.0.10.1-rc4
        version: 0.10.1-rc4
      name: stable
    defaultChannel: stable
    packageName: iomesh-operator-rhmp
    provider:
      name: SmartX
      url: https://www.smartx.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: openstack-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openstack-operator.v1.0.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openstack.org/v1beta1",
                "kind": "OpenStack",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openstack-operator",
                    "app.kubernetes.io/instance": "openstack",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "openstack",
                    "app.kubernetes.io/part-of": "openstack-operator"
                  },
                  "name": "openstack",
                  "namespace": "openstack-operators"
                },
                "spec": {
                  "operatorOverrides": [
                    {
                      "controllerManager": {
                        "resources": {
                          "limits": {
                            "cpu": "600m",
                            "memory": "2Gi"
                          }
                        }
                      },
                      "name": "infra",
                      "replicas": 1
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-02T13:33:09Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.14
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.openstack.org/v1beta1","kind":"OpenStack","metadata":{"name":"openstack","namespace":"openstack-operators"},"spec":{}}'
          operatorframework.io/suggested-namespace: openstack-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenStack is the Schema for the openstacks API
            displayName: Open Stack
            kind: OpenStack
            name: openstacks.operator.openstack.org
            version: v1beta1
        description: Install and configure OpenStack
        displayName: OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        links:
        - name: Openstack Operator
          url: https://github.com/openstack-k8s-operators/
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/glance-rhel9-operator@sha256:ee22cf417b7771930a442c2302277e6d37f368f5ae551fee9a50dbbe41c8fe3a
        - registry.redhat.io/rhoso-operators/mysqld-exporter-rhel9@sha256:b49e966cdd588071e0e1f2791ec6c62310078a7d39ea77f83843d3708a07c2b5
        - registry.redhat.io/rhoso/openstack-swift-proxy-server-rhel9@sha256:858551fd1c519797cae0e146e9c05412ed6e66dac01d0e68e62dd2d4bd335391
        - registry.redhat.io/rhoso/openstack-netutils-rhel9@sha256:b02ab1617922a2560bf739172bc2ce0f6b1fb3ef55054bd17e4b16bdf9512b4b
        - registry.redhat.io/rhoso/openstack-cinder-backup-rhel9@sha256:910693233ee3a1f3832ebe744f67a5139249523456de4abbd731c9d2f544b7c3
        - registry.redhat.io/rhoso/ironic-python-agent-rhel9@sha256:2ccd8e22a242059ca5c4e2bce853c36a71c36fdf40e8e36ab183a8218b0cc87b
        - registry.redhat.io/rhoso/openstack-frr-rhel9@sha256:f6516277149d1560c0bf2ac74a99cfb4ac2c84f9f090ab231b6d12bdb824de3b
        - registry.redhat.io/rhoso/openstack-octavia-housekeeping-rhel9@sha256:0a79259e398af746d0e8a6dda1aae5ee976c6bf8991b17f4a5da0d07dca1e36c
        - registry.redhat.io/rhoso/openstack-swift-object-rhel9@sha256:8fc2ab8caa9c7c35d94ef51c303763ea71e40b9c795c4555c17e6194344419b3
        - registry.redhat.io/ubi9/httpd-24@sha256:29382518403a4361c48fccc057163dc4a9f43ee159432754fa63c76485b223f5
        - registry.redhat.io/rhoso-operators/heat-rhel9-operator@sha256:583abb81a49db85949f7a4b846591ad8532afa79440ad7e2937ee00eacd972b3
        - registry.redhat.io/rhoso/openstack-barbican-keystone-listener-rhel9@sha256:cd8d4f0a94b4fe080ebf8e459f091db6637ceec6d83ac23cf04d66bd03c71e76
        - registry.redhat.io/rhoso/openstack-neutron-metadata-agent-ovn-rhel9@sha256:20323b151474703df610b6c24ccb9c748265196bbc70a7b5ece1b801ff540cd9
        - registry.redhat.io/rhoso-operators/test-rhel9-operator@sha256:df7566483abb18aa3c45d8bc4cd91531c7eb5b63345e511151a7cdca2dbcef17
        - registry.redhat.io/rhoso/openstack-iscsid-rhel9@sha256:508bd67021c271a10bb55ab4af033dc4e931708674b9c5ba94f72c07cd0c8ee2
        - registry.redhat.io/rhoso-operators/openstack-baremetal-agent-rhel9@sha256:08675b16d89d7931b1ef7b96e1e884c88e8e7fbfbe79a944f9213c52aba3ba83
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/rhoso/openstack-ironic-neutron-agent-rhel9@sha256:0abaa8f437025d2e7006c59c546ff6935556d707b57993719faef94142e1c43d
        - registry.redhat.io/rhoso/openstack-aodh-evaluator-rhel9@sha256:48c9192acf157df5b8f0827e35f0380d50cd4562676aba3c0d7791f92328af02
        - registry.redhat.io/rhoso/openstack-manila-scheduler-rhel9@sha256:8ce47f7fd8660167b3b292659d4fc58cf34ed671d5779693d24b71a9fddc1e7c
        - registry.redhat.io/rhoso/openstack-neutron-dhcp-agent-rhel9@sha256:cc9223ef705bcd5d0dcabbf71ad9f6e540ae7f003a6afd13161301f842dfa5fa
        - registry.redhat.io/rhoso/edpm-hardened-uefi-rhel9@sha256:c18ec7023fbcc4ab5183f689adc1f01a626a4923c15c61a37388f99904db7bc3
        - registry.redhat.io/rhoso/openstack-ceilometer-central-rhel9@sha256:e6d9f47406b7e3ff2a7f2256088cc64f043b8e497382b9f569eb90182598bbe9
        - registry.redhat.io/rhoso/openstack-cinder-api-rhel9@sha256:72a674caeeeecce3b60217f3607ad5b00651fa3d52f2f5d905f6e6395638dbc8
        - registry.redhat.io/rhoso/openstack-glance-api-rhel9@sha256:b1c0b09bb9aafe3bc31f43f8a99a92dde8d78cf90e781dd14300580986d8c980
        - registry.redhat.io/rhoso-operators/designate-rhel9-operator@sha256:d446af622bd32666eb4eb5f9b7f38f0c33ac81251a2230db9d5baf71081b9d6a
        - registry.redhat.io/rhoso/openstack-designate-producer-rhel9@sha256:04af4230f815a6e3fde22003cf382eebd9cbc3c3ec4056795f4474abc1e65fea
        - registry.redhat.io/rhoso-operators/mariadb-rhel9-operator@sha256:907e3a9d15941440d5289d7e448d9f0351053fc20fcb8cb3b9db0a84debd876f
        - registry.redhat.io/rhoso/openstack-manila-api-rhel9@sha256:e4dd3d5b66c5216b45a1515a975b85fd2ac4ef698a9350dd16e01417da462b32
        - registry.redhat.io/rhoso/openstack-cinder-volume-rhel9@sha256:8a09e80dac94fb4b3d35033e91812f95f0acac66532d9abc88aa1fb715d21822
        - registry.redhat.io/rhoso/openstack-designate-worker-rhel9@sha256:a6324a8a72722822df7692b8df0420388a1b4aecfff3da7e91da4afe4d9e2260
        - registry.redhat.io/rhoso/openstack-ceilometer-compute-rhel9@sha256:296f7f6c6bdcd7f5d07b8470308a4c71055f29a17c48d2ebb03b7f54a0e927ff
        - registry.redhat.io/rhoso/openstack-heat-api-cfn-rhel9@sha256:2789929ae3edd7c68579f34457d3071273e49a66f462043cd9c46e75bdda282a
        - registry.redhat.io/openshift4/ose-kube-state-metrics-rhel9@sha256:db6074b8aeedfd7603ae3adbe67071115f54e7f01a24997f1cd8fa8958aeddce
        - registry.redhat.io/rhoso-operators/cinder-rhel9-operator@sha256:12b8dd8fd3683250365eff6f07339cb976ee29da80642e591c10e9fa5ba627c9
        - registry.redhat.io/rhoso/openstack-heat-engine-rhel9@sha256:151e04b3e5a408aadef3a17954255acc3733d09961878a135e177188457447f0
        - registry.redhat.io/rhoso/openstack-rsyslog-rhel9@sha256:272e729accc2b64b8fcc0ece4ae5aa9ce6d454e34a25a8a8e2b99ac56b73361e
        - registry.redhat.io/rhoso/openstack-tempest-all-rhel9@sha256:16b1fbfd36ce9a27106bf09c414eaa93f14c9a6b11839068976c2fdb537d2ba7
        - registry.redhat.io/rhoso/openstack-octavia-api-rhel9@sha256:7b8d5927162ddab512096beb02e837ba51f6353c5187f3de72eb3db6e56ab6e7
        - registry.redhat.io/rhoso/openstack-keystone-rhel9@sha256:ac076c81247d774a9aeee1b63540ca6e18058213cf6aad10a6fd1f535a394ef4
        - registry.redhat.io/rhoso/openstack-cinder-scheduler-rhel9@sha256:2079ffe9b620e9277573ac796bf7e393b520631130e5bbb6f5ab51d872c943f3
        - registry.redhat.io/rhoso/openstack-aodh-notifier-rhel9@sha256:d2e89f27aac9afa504411ae6a6163442cd717f8c560e505a7da8e3fd9e2cff95
        - registry.redhat.io/rhoso/openstack-manila-share-rhel9@sha256:277174c885998e6f5b867d099664f6fa28f974bffb524903ff75283035cd8dc9
        - registry.redhat.io/rhoso/openstack-neutron-server-rhel9@sha256:2d3afde3201aa18b192ad0fb079c1e4f4f673565604f1af2b79aba8a7bc88bf2
        - registry.redhat.io/rhoso/openstack-ovn-northd-rhel9@sha256:668f742b2a18579bc8cfc610a862e43c5b877413cf65eabbe0ee36b3c7869446
        - registry.redhat.io/rhoso-operators/openstack-baremetal-rhel9-operator@sha256:03916e507b88943c0e00cedeb3cf20fa0787ccc762e47f38d305ff4d1b125b12
        - registry.redhat.io/rhoso/openstack-ovn-nb-db-server-rhel9@sha256:8c2ce1027779dc66d67a0e0d03e48b226d3278208cae0b2b81b35ad1b1f3ae72
        - registry.redhat.io/rhoso/openstack-neutron-ovn-agent-rhel9@sha256:bed1956c66b0e70064b51e34faac64036f77c7ecb14f46ceac4bc9ef52f0b9e0
        - registry.redhat.io/rhoso/openstack-nova-scheduler-rhel9@sha256:f0e5d3a693c9998ca3839b3ec7dfd5da0f9cf139e0b5b6913e92a10d5a13e6bc
        - registry.redhat.io/rhoso/openstack-designate-central-rhel9@sha256:d28d8d91d8bf75a1b77a5d650c7f2650a1791e2a1b0341450fa46978c955b92f
        - registry.redhat.io/rhoso/openstack-barbican-api-rhel9@sha256:bc942204843c45c6cb7ea841af8611f04efb003de7eeccc6813934aad556da73
        - registry.redhat.io/rhoso/openstack-octavia-health-manager-rhel9@sha256:8065851058019757b56f86ca4f030d44f169100eb2e37e08e0587a7b1e4d2e72
        - registry.redhat.io/rhoso/openstack-nova-api-rhel9@sha256:4116206f1fbe60fdd9458a5edee848571c94e2b0512d9c2cfcfd6890d60d4395
        - registry.redhat.io/rhoso-operators/keystone-rhel9-operator@sha256:879d4b1f9e720f8581b3a55c788a97bbe7b52f9108df3d5dce2e9bcb4200df6d
        - registry.redhat.io/rhoso/openstack-horizon-rhel9@sha256:82d0e6d24733afd305996e0f4a118df3e48621c40c9669b01f48669dbf40d868
        - registry.redhat.io/rhoso/openstack-designate-api-rhel9@sha256:b76ca7820553285bb4ee2bfb08e39aedbfe8c16b4edb478c06418a1e335a64ab
        - registry.redhat.io/rhoso/openstack-swift-container-rhel9@sha256:95fec49643fc18f402b8ba8057f2c35f214bf6138c5f0d7bbe1b96cbd78ef8d6
        - registry.redhat.io/rhoso-operators/swift-rhel9-operator@sha256:c99ef14338290d63e2bcbf129e953b434aea382e47072812bae31b1ea838e7df
        - registry.redhat.io/rhoso/openstack-ovn-base-rhel9@sha256:a3b8ec7ad1073facb8876acc787549a1a68b9bf81bebdc127f13eb6385e2d1da
        - registry.redhat.io/rhoso/openstack-ovn-bgp-agent-rhel9@sha256:a404c3416ab65e7134f438362e887529be474639e725e18f3542e86973edac87
        - registry.redhat.io/rhoso/openstack-nova-novncproxy-rhel9@sha256:04881867f9b24403bf6ef051eee3902a17e8476eee894ef579c0d5d130d2510d
        - registry.redhat.io/rhoso/openstack-ironic-inspector-rhel9@sha256:88357e83e8a4bbc3c5a37f0420ad75556821834a975cd8a1cf5fee56de2a4f99
        - registry.redhat.io/rhoso/openstack-ovn-controller-rhel9@sha256:3c42b7e0305c508d74602b9378143deb07a569443bba35850cdb7b6e3f5202e2
        - registry.redhat.io/rhoso/openstack-nova-conductor-rhel9@sha256:c67ca0e6c8126f56c6a83632b6a96bd4a35006f5f414a8779467b81864b31d12
        - registry.redhat.io/rhoso-operators/ee-openstack-ansible-ee-rhel9@sha256:61e729fb89c4872837cbe9cb1e3cea402b61f0da240751302605df487721f0aa
        - registry.redhat.io/rhoso-operators/ovn-rhel9-operator@sha256:66cd8add6078b63a36879c974db9e9f358593bb0651aa39fe85153d25404ae27
        - registry.redhat.io/rhoso/openstack-designate-backend-bind9-rhel9@sha256:82cca999357b0ab25bcc081bf802fab831587cf661db806af4ca453e0a626743
        - registry.redhat.io/rhoso/openstack-rabbitmq-rhel9@sha256:56bc2d2b201225fb98db43ea9ca68cceb4842f8f0f1a96ab3ab62ab785651d12
        - registry.redhat.io/rhoso-operators/prometheus-podman-exporter-rhel9@sha256:f33e7506db6dfdf530d3e2442baaf360d212bb1a47a7db887dba967ed0655cc2
        - registry.redhat.io/rhoso/openstack-nova-compute-rhel9@sha256:6264075310231af358803b53d5a09250dab32d63fc5966aa8fb957e5f56096a8
        - registry.redhat.io/rhoso-operators/neutron-rhel9-operator@sha256:30b6a812c47db9e6691211185c304c3e8fc6a8b9ba7ca79acf59c5ec07fb9fb0
        - registry.redhat.io/rhoso-operators/openstack-must-gather-rhel9@sha256:60b4355f4dbef3ad64103357086b10a5c0af126dd23798d6a9d725646f1e13b4
        - registry.redhat.io/rhoso/openstack-ironic-api-rhel9@sha256:50ccfe7d9ba22a50e0c1c8159cade6f4b1ac13ad8f5770f20c7be18fba58c8c3
        - registry.redhat.io/rhoso/openstack-ironic-pxe-rhel9@sha256:42e35ded74033f1f07287441917829cc934930a5df696c33cd8957be3c10fa41
        - registry.redhat.io/rhoso/openstack-openstackclient-rhel9@sha256:2952521d4b1799fd32ff773ef2d457a3abc354b639cf7f1e0f7f6615ca35ac8f
        - registry.redhat.io/rhoso-operators/manila-rhel9-operator@sha256:d9e69eb140b8ff76d5bf49deab8d700639ccdfd1ab5c8c7667bc2bbb50f3d60b
        - registry.redhat.io/rhoso-operators/horizon-rhel9-operator@sha256:9a2a13635588c6cc9949b272e1a1708c5fb53f19b05b8310ace703df5958ac63
        - registry.redhat.io/rhoso-operators/nova-rhel9-operator@sha256:3156dde249e49b40bed71092fb5213b112d43b663ba8c556e355aed524bec822
        - registry.redhat.io/rhoso-operators/ironic-rhel9-operator@sha256:dd67ffe3578ce619748735c1a5cd2a57d704d0c36c1b841d513ea36479aa4682
        - registry.redhat.io/rhoso/openstack-aodh-api-rhel9@sha256:4adc09cc1d922b387320da467de89d3f41a9c49b46d9b728458fec36505ec940
        - registry.redhat.io/rhoso/openstack-cron-rhel9@sha256:f09e724c69cec2c3b685783f9263b26c280a57b5909a723341416b342d18f9a4
        - registry.redhat.io/rhoso/openstack-ironic-conductor-rhel9@sha256:538da9fdeec546d794d192361d0a76f8b4e7682b6eb6d9cea371d8a52d603800
        - registry.redhat.io/rhoso/openstack-swift-account-rhel9@sha256:31e727f37db92fcc21319c44396576a03a118ba25d403715650162f7c0ead31e
        - registry.redhat.io/rhoso/openstack-redis-rhel9@sha256:c6dbeacbbd5aed9d13e540f1c987a574ea45747983c280446e5afadcf7c33b37
        - registry.redhat.io/rhoso/openstack-ovn-sb-db-server-rhel9@sha256:8ba4f8ed49e214f421e94133da0f17b54ce99a52bbbd367d0aef357de844ea01
        - registry.redhat.io/openshift4/ose-prometheus-node-exporter-rhel9@sha256:479c51de2fc8a3b3858d71bbe4eeafbd5b43263472346f9dde348115082894cc
        - registry.redhat.io/rhoso-operators/octavia-rhel9-operator@sha256:aa0f00f8e03dfdb547e53cf611581e9cc091eaf049558121c2bc5d702e556bcb
        - registry.redhat.io/rhoso/openstack-mariadb-rhel9@sha256:6115a7458cfa412c1659242fcec64b63f0a5ca414ecf3c5f890f0afec0bca069
        - registry.redhat.io/rhoso/openstack-multipathd-rhel9@sha256:6ff8ae176f8a633bcd78b1f8407924c74bb2f317eba16b90e7f0f9e53af74e8e
        - registry.redhat.io/rhoso/openstack-neutron-sriov-agent-rhel9@sha256:118d5f9430192ab48cf159ee2f7912f08e935ad56ef3364f18b3680210bb3e52
        - registry.redhat.io/rhoso-operators/barbican-rhel9-operator@sha256:b3b769e0dc32c12c5cc9578c2cac2622b973177ca14697bc1042d21d473e4e84
        - registry.redhat.io/rhoso/openstack-heat-api-rhel9@sha256:ccbc41b552bb34d223c9615a4e3b56bb516bc5bd8a694f048c15063ba45fca9f
        - registry.redhat.io/rhoso-operators/telemetry-rhel9-operator@sha256:0635515a5fed12f17de4cc5ac5f9567e3b3546fc2f69730c7beab25cf722f538
        - registry.redhat.io/rhoso/openstack-aodh-listener-rhel9@sha256:037845be6787f610e479345206b0fcfb2f1dc430540816bc5a54a1d26da2acbd
        - registry.redhat.io/rhoso/openstack-ceilometer-notification-rhel9@sha256:bd2b5f50a2bef4e77e3daa89aca51df6e0e55c3784886296d92c97616cd566d1
        - registry.redhat.io/rhoso-operators/sg-core-rhel9@sha256:b56923f869ff57396e6822133dfc3b21007c3df652cfe962c7ad3028214d988c
        - registry.redhat.io/rhoso/openstack-barbican-worker-rhel9@sha256:485525116fb3c22d8c60330dd87ebd3854cd71ded36bf88bcf18cee5957e22a3
        - registry.redhat.io/rhoso-operators/openstack-network-exporter-rhel9@sha256:8e8e775e37c28cb11b7aae9199405b85d5862dce195d165ca598fa1dc89bf914
        - registry.redhat.io/rhoso/openstack-placement-api-rhel9@sha256:0aa5e138947bf236f31c81bfb6f641d95af63812c06dca43141721b383c7e08e
        - registry.redhat.io/openshift-power-monitoring/kepler-rhel9@sha256:8e9e5bfeed61aaeee85bf5eefc3390556b8d4cbfc797a82cabe95bc0367126eb
        - registry.redhat.io/rhoso/openstack-ceilometer-ipmi-rhel9@sha256:2fb5723c68d0f2c3ae28b86e606ccae4b80f3bd9b16c01dd6c9adcae629db7f1
        - registry.redhat.io/rhoso-operators/openstack-rhel9-operator@sha256:64e2ad312f757101f083eedcb17e0b6c05cf6ebf6ed9e7db2e33e84c6aca9898
        - registry.redhat.io/rhoso/openstack-unbound-rhel9@sha256:7f17a719e8e49d457ff16437709c2b4eca5b9c2f58972d1e351c5c1c1ac38962
        - registry.redhat.io/rhoso-operators/infra-rhel9-operator@sha256:a2dc031a6ad2726d155e77cafb676f8e35a86e963b97ccdceed0da137e67ba0d
        - registry.redhat.io/rhoso/openstack-designate-mdns-rhel9@sha256:ce43dd1a30afd4753aa11d32df2b24ed4bf3d67847044aaae8f3e5c8901caf05
        - registry.redhat.io/rhoso/openstack-octavia-worker-rhel9@sha256:303e35926e43fe403673e18a25c2e815547c28a5f8c0f09d470cd4148a21c528
        - registry.redhat.io/rhoso-operators/placement-rhel9-operator@sha256:a16bd4fb929eb5bcf96f2aefcb5a7abf4143f4f9ed24e53de93f4a3caa89a507
        - registry.redhat.io/rhoso-operators/rabbitmq-cluster-rhel9-operator@sha256:e095e6cbda841b9c726c1ec0839a98964b894cf25ae9228052d314b1ee2f161d
        - registry.redhat.io/rhoso/openstack-memcached-rhel9@sha256:3e704b9664c65be39e83c2ed802722ffeaf969d8e571147a2be128d5c56331c1
        version: 1.0.14
      entries:
      - name: openstack-operator.v1.0.14
        version: 1.0.14
      - name: openstack-operator.v1.0.13
        version: 1.0.13
      - name: openstack-operator.v1.0.12
        version: 1.0.12
      - name: openstack-operator.v1.0.11
        version: 1.0.11
      - name: openstack-operator.v1.0.10
        version: 1.0.10
      - name: openstack-operator.v1.0.9
        version: 1.0.9
      - name: openstack-operator.v1.0.8
        version: 1.0.8
      - name: openstack-operator.v1.0.7
        version: 1.0.7
      - name: openstack-operator.v1.0.6
        version: 1.0.6
      - name: openstack-operator.v1.0.5
        version: 1.0.5
      - name: openstack-operator.v1.0.4
        version: 1.0.4
      - name: openstack-operator.v1.0.3
        version: 1.0.3
      - name: openstack-operator.v1.0.2
        version: 1.0.2
      - name: openstack-operator.v1.0.1
        version: 1.0.1
      - name: openstack-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: openstack-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: IBM
      provider-url: ""
    name: ibm-block-csi-operator-community
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12
    - currentCSV: ibm-block-csi-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
        version: 1.12.0
      entries:
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.0
    - currentCSV: ibm-block-csi-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
        version: 1.12.1
      entries:
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.1
    - currentCSV: ibm-block-csi-operator.v1.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:3fc8df547138a2009225d63c2a3111a672340bdf9f3cb24a889cbb69f2c33c2d
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:3fc8df547138a2009225d63c2a3111a672340bdf9f3cb24a889cbb69f2c33c2d
        version: 1.12.2
      entries:
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.2
    - currentCSV: ibm-block-csi-operator.v1.12.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:2bd461a1e40a1366c871585aa7a4e8039e686a66bf7192efdeb1c215209a071d
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:2bd461a1e40a1366c871585aa7a4e8039e686a66bf7192efdeb1c215209a071d
        version: 1.12.3
      entries:
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.3
    - currentCSV: ibm-block-csi-operator.v1.12.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e29db476de3778222bf049926223dee20f6bc34e55a04df5f0f3df38741ff23e
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e29db476de3778222bf049926223dee20f6bc34e55a04df5f0f3df38741ff23e
        version: 1.12.4
      entries:
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.4
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.5
    defaultChannel: stable
    packageName: ibm-block-csi-operator-community
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: ""
    name: pmem-csi-operator-os
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pmem-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pmem-csi.intel.com/v1beta1",
                "kind": "PmemCSIDeployment",
                "metadata": {
                  "name": "pmem-csi.intel.com"
                },
                "spec": {
                  "deviceMode": "lvm",
                  "nodeSelector": {
                    "feature.node.kubernetes.io/memory-nv.dax": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
          createdAt: "2021-08-26T10:45:23Z"
          description: An operator for deploying and managing the PMEM-CSI driver
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: ""
          repository: https://github.com/intel/pmem-csi/tree/release-1.0
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a PMEM-CSI driver deployment in the cluster.
            displayName: PmemCSIDeployment
            kind: PmemCSIDeployment
            name: pmemcsideployments.pmem-csi.intel.com
            version: v1beta1
        description: |2

          Intel PMEM-CSI is a [CSI](https://github.com/container-storage-interface/spec)
          storage driver for container orchestrators like Kubernetes. It makes local
          persistent memory ([PMEM](https://pmem.io/)) available as a filesystem volume to
          container applications. This is the operator to deploy and manage the
          [PMEM-CSI](https://intel.github.io/pmem-csi/1.0/README.html)
          driver on a Kubernetes cluster.

          Refer to the [PMEM-CSI documentation](https://intel.github.io/pmem-csi/1.0/README.html)
          for more details on deploying and using the PMEM-CSI driver.
        displayName: Operator for PMEM-CSI driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - PMEM-CSI
        - PMEM-CSI operator
        - Persistent Memory
        - PMEM
        links:
        - name: PMEM-CSI
          url: https://intel.github.io/pmem-csi/1.0
        maintainers:
        - email: olev.kartau@intel.com
          name: Olev Kartau
        - email: patrick.ohly@intel.com
          name: Patrick Ohly
        - email: amarnath.valluri@intel.com
          name: Amarnath Valluri
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Intel® Corporation
        relatedImages:
        - registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
        version: 1.0.0
      entries:
      - name: pmem-csi-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: pmem-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pmem-csi.intel.com/v1beta1",
                "kind": "PmemCSIDeployment",
                "metadata": {
                  "name": "pmem-csi.intel.com"
                },
                "spec": {
                  "deviceMode": "lvm",
                  "nodeSelector": {
                    "feature.node.kubernetes.io/memory-nv.dax": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
          createdAt: "2021-08-26T10:45:23Z"
          description: An operator for deploying and managing the PMEM-CSI driver
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: ""
          repository: https://github.com/intel/pmem-csi/tree/release-1.0
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a PMEM-CSI driver deployment in the cluster.
            displayName: PmemCSIDeployment
            kind: PmemCSIDeployment
            name: pmemcsideployments.pmem-csi.intel.com
            version: v1beta1
        description: |2

          Intel PMEM-CSI is a [CSI](https://github.com/container-storage-interface/spec)
          storage driver for container orchestrators like Kubernetes. It makes local
          persistent memory ([PMEM](https://pmem.io/)) available as a filesystem volume to
          container applications. This is the operator to deploy and manage the
          [PMEM-CSI](https://intel.github.io/pmem-csi/1.0/README.html)
          driver on a Kubernetes cluster.

          Refer to the [PMEM-CSI documentation](https://intel.github.io/pmem-csi/1.0/README.html)
          for more details on deploying and using the PMEM-CSI driver.
        displayName: Operator for PMEM-CSI driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - PMEM-CSI
        - PMEM-CSI operator
        - Persistent Memory
        - PMEM
        links:
        - name: PMEM-CSI
          url: https://intel.github.io/pmem-csi/1.0
        maintainers:
        - email: olev.kartau@intel.com
          name: Olev Kartau
        - email: patrick.ohly@intel.com
          name: Patrick Ohly
        - email: amarnath.valluri@intel.com
          name: Amarnath Valluri
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Intel® Corporation
        relatedImages:
        - registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
        version: 1.0.0
      entries:
      - name: pmem-csi-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pmem-csi-operator-os
    provider:
      name: Intel® Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OceanBase
      provider-url: https://oceanbase.com
    name: oceanbase-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oceanbase-operator.v2.2.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Full Lifecycle
          createdAt: "2024-08-14T16:21:06Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: OBClusterOperation
            name: obclusteroperations.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBCluster
            name: obclusters.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBParameter
            name: obparameters.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBResourceRescue
            name: obresourcerescues.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBServer
            name: observers.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantBackupPolicy
            name: obtenantbackuppolicies.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantBackup
            name: obtenantbackups.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantOperation
            name: obtenantoperations.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantRestore
            name: obtenantrestores.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenant
            name: obtenants.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBZone
            name: obzones.oceanbase.oceanbase.com
            version: v1alpha1
        description: Ob Operator description. TODO.
        displayName: Ob Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - oceanbase-operator
        - ob-operator
        links:
        - name: OceanBase Operator
          url: https://oceanbase.github.io/ob-operator
        maintainers:
        - email: hongxin.sun.vip@gmail.com
          name: Hongxin Sun
        maturity: alpha
        provider:
          name: OceanBase
          url: https://oceanbase.com
        relatedImages:
        - quay.io/oceanbasedev/oceanbase-operator@sha256:b11fe60bb8de545d3ed105acc50f2d6c5d6aa8093f3ae4ca7ebce58b1e5e7b48
        - quay.io/brancz/kube-rbac-proxy@sha256:18f12fc3a0085bbc303cfbd2ad1151127c1c959c4641e36027547b6b66eb4c92
        version: 2.2.2
      entries:
      - name: oceanbase-operator.v2.2.2
        version: 2.2.2
      name: alpha
    - currentCSV: oceanbase-operator.v2.2.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Full Lifecycle
          createdAt: "2024-08-14T16:21:06Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: OBClusterOperation
            name: obclusteroperations.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBCluster
            name: obclusters.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBParameter
            name: obparameters.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBResourceRescue
            name: obresourcerescues.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBServer
            name: observers.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantBackupPolicy
            name: obtenantbackuppolicies.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantBackup
            name: obtenantbackups.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantOperation
            name: obtenantoperations.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantRestore
            name: obtenantrestores.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenant
            name: obtenants.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBZone
            name: obzones.oceanbase.oceanbase.com
            version: v1alpha1
        description: Ob Operator description. TODO.
        displayName: Ob Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - oceanbase-operator
        - ob-operator
        links:
        - name: OceanBase Operator
          url: https://oceanbase.github.io/ob-operator
        maintainers:
        - email: hongxin.sun.vip@gmail.com
          name: Hongxin Sun
        maturity: alpha
        provider:
          name: OceanBase
          url: https://oceanbase.com
        relatedImages:
        - quay.io/oceanbasedev/oceanbase-operator@sha256:b11fe60bb8de545d3ed105acc50f2d6c5d6aa8093f3ae4ca7ebce58b1e5e7b48
        - quay.io/brancz/kube-rbac-proxy@sha256:18f12fc3a0085bbc303cfbd2ad1151127c1c959c4641e36027547b6b66eb4c92
        version: 2.2.2
      entries:
      - name: oceanbase-operator.v2.2.2
        version: 2.2.2
      name: stable
    defaultChannel: stable
    packageName: oceanbase-operator
    provider:
      name: OceanBase
      url: https://oceanbase.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.2.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"customImageNames":
            false,"externalArangoDBName": "arangodb","repository": "registry.connect.redhat.com/turbonomic","tag":"8.3.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/turbonomic/t8c-operator@sha256:d5164e6dd24a0feff10cab62530cafff1bc92d0e8ec2d16c6f4d880c6e2f50bd
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/t8c-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/t8c-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/turbonomic/t8c-operator@sha256:d5164e6dd24a0feff10cab62530cafff1bc92d0e8ec2d16c6f4d880c6e2f50bd
        version: 42.2.0
      entries:
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: stable
    defaultChannel: stable
    packageName: t8c-certified-rhmp
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ptp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ptp-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "NodePtpDevice",
                "metadata": {
                  "name": "node.example.com"
                },
                "spec": {}
              },
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "PtpConfig",
                "metadata": {
                  "name": "example-ptpconfig"
                },
                "spec": {
                  "profile": [
                    {
                      "interface": "ens787f1",
                      "name": "profile1",
                      "phc2sysOpts": "-a -r",
                      "ptp4lOpts": "-s -2",
                      "ptpClockThreshold": {}
                    }
                  ],
                  "recommend": [
                    {
                      "match": [
                        {
                          "nodeLabel": "node-role.kubernetes.io/worker=",
                          "nodeName": "node.example.com"
                        }
                      ],
                      "priority": 4,
                      "profile": "profile1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "PtpOperatorConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "daemonNodeSelector": {
                    "feature.node.kubernetes.io/ptp-capable": "yes",
                    "node-role.kubernetes.io/worker": ""
                  },
                  "ptpEventConfig": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-ptp-rhel9-operator@sha256:012334e6624d5440439824454bda1d1dc79804938174232a3b470176bf462b5e
          createdAt: "2024-03-13"
          description: This software enables configuration of Precision Time Protocol(PTP)
            on Kubernetes. It detects hardware capable PTP devices on each node, and
            configures linuxptp processes such as ptp4l, phc2sys and timemaster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.16.0-202509240123'
          operatorframework.io/suggested-namespace: openshift-ptp
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/ptp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodePtpDevice is the Schema for the nodeptpdevices API
            displayName: Node Ptp Device
            kind: NodePtpDevice
            name: nodeptpdevices.ptp.openshift.io
            version: v1
          - description: PtpConfig is the Schema for the ptpconfigs API
            displayName: Ptp Config
            kind: PtpConfig
            name: ptpconfigs.ptp.openshift.io
            version: v1
          - description: PtpOperatorConfig is the Schema for the ptpoperatorconfigs
              API
            displayName: Ptp Operator Config
            kind: PtpOperatorConfig
            name: ptpoperatorconfigs.ptp.openshift.io
            version: v1
        description: |
          # Precision Time Protocol (PTP) Operator for Openshift

          ## Introdution
          PTP Operator manages cluster wide PTP configuration.
          This operator has to run in namespace 'openshift-ptp'. An Operator Group is also required to install this operator:
          ```
          $ oc create -f - <<EOF
          apiVersion: v1
          kind: Namespace
          metadata:
            name: openshift-ptp
            labels:
              name: openshift-ptp
              openshift.io/cluster-monitoring: "true"
          EOF

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: ptp-operators
            namespace: openshift-ptp
          spec:
            targetNamespaces:
            - openshift-ptp
          EOF
          ```
        displayName: PTP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ptp
        links:
        - name: Documentation operator
          url: https://github.com/openshift/ptp-operator
        - name: Documentation operand
          url: https://github.com/openshift/linuxptp-daemon
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-cloud-event-proxy-rhel9@sha256:3fa2aad8cf026eba586645c851f07a1d752c85b30795fdcf037b138ef3bbb6aa
        - registry.redhat.io/openshift4/ose-ptp-rhel9-operator@sha256:012334e6624d5440439824454bda1d1dc79804938174232a3b470176bf462b5e
        - registry.redhat.io/openshift4/ose-ptp-rhel9@sha256:93b4717337b551603cbb6147d66d7b70a5580332953ddf42304f7092d8739c83
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        version: 4.16.0-202509240123
      entries:
      - name: ptp-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: ptp-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: ptp-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: ptp-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: ptp-operator.v4.16.0-202506201205
        version: 4.16.0-202506201205
      - name: ptp-operator.v4.16.0-202505291140
        version: 4.16.0-202505291140
      - name: ptp-operator.v4.16.0-202505161704
        version: 4.16.0-202505161704
      - name: ptp-operator.v4.16.0-202505020146
        version: 4.16.0-202505020146
      - name: ptp-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: ptp-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: ptp-operator.v4.16.0-202502260004
        version: 4.16.0-202502260004
      - name: ptp-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: ptp-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: ptp-operator.v4.16.0-202501201505
        version: 4.16.0-202501201505
      - name: ptp-operator.v4.16.0-202501072135
        version: 4.16.0-202501072135
      - name: ptp-operator.v4.16.0-202412311307
        version: 4.16.0-202412311307
      - name: ptp-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: ptp-operator.v4.16.0-202412102004
        version: 4.16.0-202412102004
      - name: ptp-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: ptp-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: ptp-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: ptp-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: ptp-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: ptp-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: ptp-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: ptp-operator.v4.16.0-202409131735
        version: 4.16.0-202409131735
      - name: ptp-operator.v4.16.0-202409062036
        version: 4.16.0-202409062036
      - name: ptp-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: ptp-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: ptp-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: ptp-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: ptp-operator.v4.16.0-202407311708
        version: 4.16.0-202407311708
      - name: ptp-operator.v4.16.0-202407301837
        version: 4.16.0-202407301837
      - name: ptp-operator.v4.16.0-202407160236
        version: 4.16.0-202407160236
      - name: ptp-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: ptp-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: ptp-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: ptp-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: ptp-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: ptp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fossul Framework
      provider-url: https://github.com/fossul/fossul
    name: fossul-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fossul-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n    \"kind\":
            \"Backup\",\n    \"metadata\": {\n      \"name\": \"mariadb-sample\"\n
            \   },\n    \"spec\": {\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\"\n    }\n  },\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n    \"kind\":
            \"Restore\",\n    \"metadata\": {\n      \"name\": \"mariadb-sample\"\n
            \   },\n    \"spec\": {\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\",\n      \"workflow_id\": \"xxxx\"\n    }\n  },        \n  {\n
            \   \"apiVersion\": \"fossul.io/v1\",\n    \"kind\": \"BackupConfig\",\n
            \   \"metadata\": {\n      \"name\": \"mariadb-sample\"\n    },\n    \"spec\":
            {\n      \"container_platform\": \"openshift\",\n      \"operator_controlled\":
            true,\n      \"auto_discovery\": false,\n      \"deployment_name\": \"mariadb\",\n
            \     \"deployment_type\": \"DeploymentConfig\",\n      \"job_retention\":
            50,\n      \"overwrite_pcv_on_restore\": true,\n      \"policies\": [\n
            \       {\n          \"policy\": \"hourly\",\n          \"retentionNumber\":
            3\n        },\n        {\n          \"policy\": \"daily\",\n          \"retentionNumber\":
            10\n        }\n      ],\n      \"pvc_deletion_timeout\": 300,\n      \"restore_to_new_pvc\":
            false,\n      \"snapshot_timeout\": 180,\n      \"storage_plugin\": \"csi.so\",\n
            \     \"app_plugin\": \"mariadb.so\"\n    }\n  },\n  {\n    \"apiVersion\":
            \"fossul.io/v1\",\n    \"kind\": \"BackupSchedule\",\n    \"metadata\":
            {\n      \"name\": \"mariadb-sample\"\n    },\n    \"spec\": {\n      \"cron_schedule\":
            \"59 23 * * *\",\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\"\n\n    }\n  },\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n
            \   \"kind\": \"Fossul\",\n    \"metadata\": {\n      \"name\": \"fossul-sample\"\n
            \   },\n    \"spec\": {\n      \"enable_routes\": false,\n      \"enable_virtualmachines\":
            false,\n      \"container_platform\": \"openshift\"\n    }\n  }\n]"
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/fossul/fossul-operator:v0.6.0
          createdAt: "2022-01-25T20:09:00Z"
          description: Fossul is a application centric container native backup and
            restore framework for databases or applications
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/fossul/fossul
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages application backup configurations
            kind: BackupConfig
            name: backupconfigs.fossul.io
            version: v1
          - description: Manages application backups
            kind: Backup
            name: backups.fossul.io
            version: v1
          - description: Manages application restores
            kind: Restore
            name: restores.fossul.io
            version: v1
          - description: Manages application backup scheduling
            kind: BackupSchedule
            name: backupschedules.fossul.io
            version: v1
          - description: Manages deployment of the Fossul Framework
            kind: Fossul
            name: fossuls.fossul.io
            version: v1
        description: "## Fossul Framework\nApplication backup and restore framework
          for container native applications on Kubernetes and OpenShift. Fossul provides
          application centric backup and recovery enabling integration with the application
          creating an application consistent backup. Since there are many applications
          and providers for storage or archive Fossul enables plugins that allow proper
          integration. Fossul provides plugins for postgresql, mariadb, mysql and
          mongo as well as plugin to integrate with CSI snapshots. Plugins can be
          written in any language. The Fossul operator deploys Fossul via the Fossul
          custom resource and additionally provides BackupConfig, Backup, BackupSchedule
          and Restore custom resources. Each user can configure and manage their own
          backups with simple custom resources that can be part of CI/CD or source
          control. Users can also opt to manage backups through Fossul CLI/API instead
          of the custom resources provided by the operator.\n\n## Installation\nIt
          is recommended to install the operator into the openshift-operators namespace
          which is the default location for operators. Once the operator is installed
          create a new project called ```fossul``` and create the Fossul custom resource
          under the fossul namespace. This will deploy the Fossul Framework. You can
          choose another namespace for the Fossul Framework but then also need to
          provide the optional 'fossul_namespace' parameter with the spec for all
          the additional custom resources. This is required so the Fossul operator
          can communicate with Fossul.\n\nOnce Fossul is deployed you can manage backups
          either through custom resources provided by the operator or the Fossul CLI/API.
          If using the customer resources, first create a BackupConfig custom resource
          for every application you want to backup. Once that is done, you can optionally
          create a backup by simply creating a Backup custom resource. You can also
          create a BackupSchedule custom resource using cron syntax which will schedule
          your backup to occur on a regular interval or do so via a Kubernetes job.
          Finally if a backup has been created you can perform a restore through a
          custom resource as well, providing the workflow id from the completed backup.
          If you are creating backups through the custom resource the workflow id
          will be appended to the spec once the backup completes. If you are using
          the Fossul CLI/API to create backups you will need to get the workflow id
          through the Fossul CLI/API directly using the backupList action.\n\nFor
          users of the cli, a cli pod is deployed with credentials and if desired
          you can create or manage backups from within that pod. To use the API you
          need the credentials stored in the fossul secret. The cli has already been
          pre-configured with these credentials. The API can also be optionally exposed
          via routes and API documentation is under ```/api/v1/index.html```.\n\n##
          Quick Setup\nWhen creating Fossul custom resources generally you will only
          need to update the ```deployment_name``` and ```deployment_type``` (DeploymentConfig,
          Deployment or VirtualMachine). All custom resources must be created in the
          namespace where the database or application exists. \n\n### Deploy Fossul
          in fossul namespace\n```$ oc new-project fossul```\n<pre>\n$ vi fossul.yaml\nkind:
          Fossul\napiVersion: fossul.io/v1\nmetadata:\n  name: fossul-sample\n  namespace:
          fossul\nspec: {}\n</pre>\n```$ oc create -f fossul.yaml```\n\n### Deploy
          Mariadb Database\nMake sure you create all Fossul custom resources in the
          same namespace as the database, application or virtual machine\n\n### Create
          MariaDB Fossul BackupConfig\nCreates a backup configuration for MariaDB
          or MySQL databases\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true\n  auto_discovery:
          false\n  deployment_name: mariadb\n  deployment_type: DeploymentConfig\n
          \ job_retention: 50\n  overwrite_pcv_on_restore: true\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc: false\n  snapshot_timeout:
          180\n  storage_plugin: csi.so\n  app_plugin: mariadb.so \n</pre>\n```$ oc
          create -f backupconfig.yaml```\n\n### Create PostgreSQL Fossul BackupConfig\nCreates
          a backup config for PostgreSQL databases, you need to ensure the user defined
          in secret has backup permissions\n\n```$ vi backupconfig.yaml```\n<pre>\nkind:
          BackupConfig\napiVersion: fossul.io/v1\nmetadata:\n  name: postgresql-sample\n
          \ namespace: databases\nspec:\n  container_platform: openshift\n  operator_controlled:
          true    \n  auto_discovery: false\n  deployment_name: postgresql \n  deployment_type:
          DeploymentConfig\n  job_retention: 50\n  overwrite_pcv_on_restore: true\n
          \ policies:\n    - policy: hourly\n      retentionNumber: 3\n    - policy:
          daily\n      retentionNumber: 10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc:
          false\n  snapshot_timeout: 180\n  storage_plugin: csi.so\n  app_plugin:
          postgres.so \n</pre>\n```$ oc create -f backupconfig.yaml```\n\n### Create
          Kubevirt Fossul BackupConfig\nCreates a backup configuration for virtual
          machines\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: rhel8-vm-sample\n  namespace: virtualmachines\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true    \n  auto_discovery:
          false\n  deployment_name: rhel8-vm\n  deployment_type: VirtualMachine\n
          \ job_retention: 50\n  overwrite_pcv_on_restore: true\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc: false\n  snapshot_timeout:
          180\n  storage_plugin: csi.so\n  app_plugin: kubevirt.so \n</pre>\n```$
          oc create -f backupconfig.yaml```\n\n### Create Tar Fossul BackupConfig\nCreates
          a backup config using tar to backup folders and files where csi snapshots
          are not desired\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: tar-sample\n  namespace: databases\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true    \n  auto_discovery:
          false\n  deployment_name: anyname\n  job_retention: 50\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  backup_source_paths: /opt/app-root/src/test\n  storage_plugin: tar.so\n</pre>\n```$
          oc create -f backupconfig.yaml```\n\n### Create BackupSchedule\nA backup
          can be scheduled per policy, defined by backup configuration using cron
          syntax\n\n```$ vi backupschedule.yaml```\n<pre>\nkind: BackupSchedule\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ cron_schedule: '59 23 * * *'\n  deployment_name: mariadb\n  policy: daily\n</pre>\n```$
          oc create -f backupschedule.yaml```\n\n### Create Backup\nA backup will
          be created according to policy and deleted either manually or as defined
          in retention policy\n\n```$ vi backup.yaml```\n<pre>\nkind: Backup\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ deployment_name: mariadb\n  policy: daily\n</pre>\n```$ oc create -f backup.yaml```\n\n###
          Perform Restore\nA restore requires a workflow_id located in the backup
          spec, upon a successful restore the custom resource will deleted automatically\n\n```$
          vi restore.yaml```\n<pre>\nkind: Restore\napiVersion: fossul.io/v1\nmetadata:\n
          \ name: mariadb-sample\n  namespace: databases\nspec:\n  deployment_name:
          mariadb\n  policy: daily\n  workflow_id: xxxx\n</pre>\n```$ oc create -f
          restore.yaml```\n"
        displayName: Fossul Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - application
        - consistent
        links:
        - name: Fossul Framework Community
          url: https://github.com/fossul/fossul
        maintainers:
        - email: ktenzer@redhat.com
          name: Keith Tenzer
        maturity: alpha
        provider:
          name: Fossul Framework
          url: https://github.com/fossul/fossul
        relatedImages:
        - quay.io/fossul/fossul-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: fossul-operator.v0.6.0
        version: 0.6.0
      - name: fossul-operator.v0.5.0
        version: 0.5.0
      - name: fossul-operator.v0.4.0
        version: 0.4.0
      - name: fossul-operator.v0.3.0
        version: 0.3.0
      - name: fossul-operator.v0.2.0
        version: 0.2.0
      - name: fossul-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: fossul-operator
    provider:
      name: Fossul Framework
      url: https://github.com/fossul/fossul
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Citrus Framework
      provider-url: ""
    name: yaks
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: yaks-operator.v0.17.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "yaks.citrusframework.org/v1alpha1",
                "kind": "Instance",
                "metadata": {
                  "labels": {
                    "app": "yaks"
                  },
                  "name": "example-instance"
                },
                "spec": {
                  "operator": {
                    "global": true
                  }
                }
              },
              {
                "apiVersion": "yaks.citrusframework.org/v1alpha1",
                "kind": "Test",
                "metadata": {
                  "labels": {
                    "app": "yaks"
                  },
                  "name": "example-test"
                },
                "spec": {
                  "source": {
                    "content": "Feature: hello world\n\n  Scenario: print slogan\n    Given YAKS does Cloud-Native BDD testing\n    Then YAKS rocks!",
                    "language": "feature",
                    "name": "example.feature"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/citrusframework/yaks:0.17.1
          createdAt: "2023-04-26T20:37:31Z"
          description: YAKS is a platform to enable Cloud Native BDD testing on Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["instances.yaks.citrusframework.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/citrusframework/yaks
          support: Citrus Framework
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance is the Schema for the yaks instance
            displayName: Instance
            kind: Instance
            name: instances.yaks.citrusframework.org
            version: v1alpha1
          - description: Test is the Schema for the tests API
            displayName: Test
            kind: Test
            name: tests.yaks.citrusframework.org
            version: v1alpha1
        description: |
          YAKS
          ====

          YAKS is a platform to enable Cloud Native BDD testing on Kubernetes.

          ## Running a YAKS test

          With the YAKS operator installed, you can run tests by creating a `Test` custom resource on the cluster.

          Tests are defined using [Gherkin](https://cucumber.io/docs/gherkin/) syntax. YAKS provides a set of predefined steps which
          help to connect with different messaging transports (Http REST, JMS, Kafka, Knative eventing) and verify responses with
          assertions on message header and body content.

          The example below defines a simple test resource.

          ```
          apiVersion: yaks.citrusframework.org/v1alpha1
          kind: Test
          metadata:
            name: example
          spec:
            source:
              name: example.feature
              content: |-
                Feature: hello world

                  Scenario: print slogan
                    Given YAKS does Cloud-Native BDD testing
                    Then YAKS rocks!
          ```

          Refer to the [YAKS repository](https://github.com/citrusframework/yaks) for more complex examples.
        displayName: YAKS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - yaks
        - testing
        - microservices
        - knative
        - kafka
        - serverless
        - camel
        - camel-k
        links:
        - name: Project page
          url: https://citrusframework.org/
        - name: YAKS source code repository
          url: https://github.com/citrusframework/yaks
        maintainers:
        - email: cdeppisch@redhat.com
          name: Christoph Deppisch
        - email: nferraro@redhat.com
          name: Nicola Ferraro
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Citrus Framework
        relatedImages:
        - docker.io/citrusframework/yaks:0.17.1
        version: 0.17.1
      entries:
      - name: yaks-operator.v0.17.1
        version: 0.17.1
      - name: yaks-operator.v0.17.0
        version: 0.17.0
      - name: yaks-operator.v0.16.0
        version: 0.16.0
      - name: yaks-operator.v0.15.1
        version: 0.15.1
      - name: yaks-operator.v0.15.0
        version: 0.15.0
      - name: yaks-operator.v0.14.3
        version: 0.14.3
      - name: yaks-operator.v0.14.2
        version: 0.14.2
      - name: yaks-operator.v0.14.1
        version: 0.14.1
      - name: yaks-operator.v0.13.0
        version: 0.13.0
      - name: yaks-operator.v0.12.2
        version: 0.12.2
      - name: yaks-operator.v0.11.0
        version: 0.11.0
      - name: yaks-operator.v0.10.0
        version: 0.10.0
      - name: yaks-operator.v0.9.0
        version: 0.9.0
      - name: yaks-operator.v0.8.0
        version: 0.8.0
      - name: yaks-operator.v0.7.0
        version: 0.7.0
      - name: yaks-operator.v0.6.0
        version: 0.6.0
      - name: yaks-operator.v0.5.0
        version: 0.5.0
      - name: yaks-operator.v0.4.0
        version: 0.4.0
      name: alpha
    defaultChannel: alpha
    packageName: yaks
    provider:
      name: Citrus Framework
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator.v1.5.0
        version: 1.5.0
      - name: dynatrace-operator.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetScaler
      provider-url: https://netscaler.com
    name: netscaler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netscaler-operator.v3.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCloudController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalercloudcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "azure": {
                    "clientsecret": "",
                    "image": "quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb",
                    "location": "",
                    "pullPolicy": "IfNotPresent",
                    "resourcegroupname": "",
                    "resources": {},
                    "subnetname": "",
                    "subscriptionid": "",
                    "tenantid": "",
                    "vnetname": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCpxWithIngressController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalercpxwithingresscontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "admSettings": {
                    "admIP": "",
                    "bandWidth": 1000,
                    "bandWidthLicense": false,
                    "cpxCores": 1,
                    "licenseEdition": "PLATINUM",
                    "licenseServerIP": "",
                    "licenseServerPort": 27000,
                    "loginSecret": "",
                    "platform": false,
                    "vCPULicense": false
                  },
                  "affinity": {},
                  "analyticsConfig": {
                    "distributedTracing": {
                      "enable": false,
                      "samplingRate": 100
                    },
                    "endpoint": {
                      "metrics": {
                        "service": ""
                      },
                      "transactions": {
                        "service": ""
                      }
                    },
                    "required": false,
                    "timeseries": {
                      "auditLogs": {
                        "enable": false
                      },
                      "events": {
                        "enable": false
                      },
                      "metrics": {
                        "enable": false,
                        "enableNativeScrape": false,
                        "exportFrequency": 30,
                        "mode": "avro",
                        "schemaFile": "schema.json"
                      },
                      "port": 5563
                    },
                    "transactions": {
                      "enable": false,
                      "port": 5557
                    }
                  },
                  "exporter": {
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4",
                    "ports": {
                      "containerPort": 8888
                    },
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {}
                  },
                  "extraVolumes": [],
                  "imagePullSecrets": [],
                  "ingressController": {
                    "defaultSSLCertSecret": "",
                    "defaultSSLSNICertSecret": "",
                    "disableAPIServerCertVerify": false,
                    "disableOpenshiftRoutes": true,
                    "enableLivenessProbe": true,
                    "enableReadinessProbe": true,
                    "entityPrefix": "k8s",
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "ingressClass": [],
                    "ipam": false,
                    "jsonLog": false,
                    "kubernetesURL": "",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "logLevel": "INFO",
                    "logProxy": "",
                    "namespaceLabels": "",
                    "openshift": true,
                    "optimizeEndpointBinding": false,
                    "profileHttpFrontend": {},
                    "profileSslFrontend": {},
                    "profileTcpFrontend": {},
                    "pullPolicy": "IfNotPresent",
                    "rbacRole": false,
                    "readinessProbe": {
                      "failureThreshold": 3,
                      "initialDelaySeconds": 50,
                      "periodSeconds": 60,
                      "successThreshold": 1
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "128m",
                        "memory": "500Mi"
                      }
                    },
                    "routeLabels": "",
                    "setAsDefaultIngressClass": false,
                    "updateIngressStatus": true
                  },
                  "netscalerCpx": {
                    "bgpSettings": {
                      "bgpConfig": [
                        {
                          "bgpRouter": {
                            "localAS": 100,
                            "neighbor": [
                              {
                                "ASOriginationInterval": 10,
                                "address": "",
                                "advertisementInterval": 10,
                                "remoteAS": 100
                              }
                            ]
                          }
                        }
                      ],
                      "bgpPort": 179
                    },
                    "cpxBgpRouter": false,
                    "cpxCommands": [],
                    "cpxLicenseAggregator": "",
                    "cpxShellCommands": [],
                    "daemonSet": false,
                    "enableLivenessProbe": true,
                    "enableStartupProbe": true,
                    "extraVolumeMounts": [],
                    "hostname": "",
                    "image": "quay.io/netscaler/netscaler-cpx@sha256:232596eaacda68c0e2e210f26b64cedc8d70ae12c5e8ae2e1a66896a479ae4f8",
                    "ingressIP": "",
                    "livenessProbe": {
                      "failureThreshold": 3,
                      "periodSeconds": 5,
                      "successThreshold": 1
                    },
                    "mgmtHttpPort": 9080,
                    "mgmtHttpsPort": 9443,
                    "nitroReadTimeout": 20,
                    "nsConfigDnsRec": false,
                    "nsCookieVersion": 0,
                    "nsDnsNameserver": "",
                    "nsEnableLabels": true,
                    "nsGateway": "192.168.1.1",
                    "nsHttp2ServerSide": "OFF",
                    "nsIP": "192.168.1.2",
                    "nsLbHashAlgo": {
                      "hashAlgorithm": "default",
                      "hashFingers": 500,
                      "required": false
                    },
                    "nsProtocol": "HTTP",
                    "nsSvcLbDnsRec": false,
                    "prometheusCredentialSecret": "",
                    "pullPolicy": "IfNotPresent",
                    "service": {
                      "annotations": {},
                      "spec": {}
                    },
                    "startupProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1
                    }
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "replicaCount": 1,
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerGslbController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalergslbcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "gslbController": {
                    "disableAPIServerCertVerify": false,
                    "entityPrefix": "",
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "kubernetesURL": "",
                    "localCluster": "",
                    "localRegion": "",
                    "logLevel": "Debug",
                    "openshift": true,
                    "pullPolicy": "IfNotPresent",
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "32m",
                        "memory": "128Mi"
                      }
                    },
                    "siteData": [
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteMask": "",
                        "siteName": "",
                        "sitePublicIp": "",
                        "siteRegion": ""
                      },
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteMask": "",
                        "siteName": "",
                        "sitePublicIp": "",
                        "siteRegion": ""
                      }
                    ]
                  },
                  "imagePullSecrets": [],
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsIP": "",
                    "nsPort": 443,
                    "nsProtocol": "HTTPS"
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerIngressController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscaleringresscontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "analyticsConfig": {
                    "distributedTracing": {
                      "enable": false,
                      "samplingRate": 100
                    },
                    "endpoint": {
                      "metrics": {
                        "service": ""
                      },
                      "transactions": {
                        "service": ""
                      }
                    },
                    "required": false,
                    "timeseries": {
                      "auditLogs": {
                        "enable": false
                      },
                      "events": {
                        "enable": false
                      },
                      "metrics": {
                        "enable": false,
                        "enableNativeScrape": false,
                        "exportFrequency": 50,
                        "mode": "avro",
                        "schemaFile": "schema.json"
                      },
                      "port": 50002
                    },
                    "transactions": {
                      "enable": false,
                      "port": 50005
                    }
                  },
                  "exporter": {
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4",
                    "ports": {
                      "containerPort": 8888
                    },
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {},
                    "serviceMonitorExtraLabels": {}
                  },
                  "extraVolumes": [],
                  "hostAliases": [],
                  "imagePullSecrets": [],
                  "ingressController": {
                    "clusterName": "",
                    "defaultSSLCertSecret": "",
                    "defaultSSLSNICertSecret": "",
                    "disableAPIServerCertVerify": false,
                    "disableOpenshiftRoutes": true,
                    "enableLivenessProbe": true,
                    "enableReadinessProbe": true,
                    "entityPrefix": "",
                    "extraVolumeMounts": [],
                    "ignoreNodeExternalIP": false,
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "ingressClass": [],
                    "ipam": false,
                    "jsonLog": false,
                    "kubernetesURL": "",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "logLevel": "INFO",
                    "logProxy": "",
                    "multiClusterPrefix": "",
                    "namespaceLabels": "",
                    "nodeLabels": "",
                    "nodeWatch": false,
                    "nsValidateCert": {},
                    "nsncPbr": false,
                    "openshift": true,
                    "optimizeEndpointBinding": false,
                    "podIPsforServiceGroupMembers": false,
                    "profileHttpFrontend": {},
                    "profileSslFrontend": {},
                    "profileTcpFrontend": {},
                    "pullPolicy": "IfNotPresent",
                    "rbacRole": false,
                    "readinessProbe": {
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 60,
                      "successThreshold": 1
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "32m",
                        "memory": "128Mi"
                      }
                    },
                    "routeLabels": "",
                    "serviceClass": [],
                    "setAsDefaultIngressClass": false,
                    "updateIngressStatus": true
                  },
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsConfigDnsRec": false,
                    "nsCookieVersion": 0,
                    "nsDnsNameserver": "",
                    "nsHttp2ServerSide": "OFF",
                    "nsIP": "",
                    "nsLbHashAlgo": {
                      "hashAlgorithm": "default",
                      "hashFingers": 256,
                      "required": false
                    },
                    "nsPort": 443,
                    "nsProtocol": "HTTPS",
                    "nsSNIPs": [],
                    "nsSvcLbDnsRec": false,
                    "nsVIP": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerIpamController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscaleripamcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "clusterName": "",
                  "image": "quay.io/netscaler/netscaler-ipam-controller@sha256:7c25b59a2e1df0e4f5615a779a9bea939c60a3eca065cb6c2295daf0d03831d8",
                  "localIPAM": {
                    "vipRange": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "reuseIngressVIP": true,
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerObservabilityExporter",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalerobservabilityexporter-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "elasticsearch": {
                    "enabled": false,
                    "server": "http://elasticsearch:9200"
                  },
                  "headless": false,
                  "image": "quay.io/netscaler/netscaler-observability-exporter@sha256:626e3a954c67bbb910477352dc740aaab21ff1c1cfb03354becf50beab17d4a4",
                  "jsonTransRateLimiting": {
                    "enabled": false,
                    "limit": 100,
                    "queueLimit": 1000,
                    "window": 5
                  },
                  "kafka": {
                    "auditLogs": "no",
                    "broker": "",
                    "dataFormat": "AVRO",
                    "enabled": false,
                    "events": "no",
                    "topic": "HTTP"
                  },
                  "nodePortRequired": false,
                  "nodeSelector": {},
                  "nsoeLogLevel": "INFO",
                  "nstracing": {
                    "enabled": false,
                    "server": "zipkin:9411/api/v1/spans"
                  },
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "replicas": 1,
                  "resources": {},
                  "splunk": {
                    "authToken": "",
                    "enabled": false,
                    "indexPrefix": "adc_nsoe",
                    "server": "http://splunk:8088"
                  },
                  "timeseries": {
                    "enabled": false
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-operator@sha256:e367c8c8ba6171e9ef5dd496592a5c62f4aa764f93d0cb78835aa29024c063d0
          createdAt: "2025-04-15T08:51:39Z"
          description: This is an operator to install Cloud Native portfolio provided
            by NetScaler
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler Cloud Controller for NetScaler VPX/MPX/SDX
              with Azure Redhat OpenShift in Azure.
            displayName: Netscaler Cloud Controller
            kind: NetscalerCloudController
            name: netscalercloudcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler CPX with Ingress Controller.
            displayName: Netscaler CPX With Ingress Controller
            kind: NetscalerCpxWithIngressController
            name: netscalercpxwithingresscontrollers.netscaler.com
            version: v1
          - description: To install NetScaler GSLB Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler GSLB Controller
            kind: NetscalerGslbController
            name: netscalergslbcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler Ingress Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler Ingress Controller
            kind: NetscalerIngressController
            name: netscaleringresscontrollers.netscaler.com
            version: v1
          - description: To install NetScaler IPAM Controller for NetScaler CPX/VPX/MPX/SDX.
            displayName: Netscaler IPAM Controller
            kind: NetscalerIpamController
            name: netscaleripamcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler Observability Exporter for NetScaler.
            displayName: Netscaler Observability Exporter
            kind: NetscalerObservabilityExporter
            name: netscalerobservabilityexporters.netscaler.com
            version: v1
        description: |
          NetScaler provides various products to empower Developer/DevOps managing your applications in OpenShift Cluster. Using this operator you can deploy below controllers for NetScaler appliance:
          1. [NetScaler Cloud Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. [NetScaler Ingress Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/unified-ingress-deployment)
          3. [NetScaler CPX with Ingress Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/two-tier-deployment)
          4. [NetScaler GSLB Controller ](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/gslb/gslb)
          5. [NetScaler Observability Exporter](https://docs.netscaler.com/en-us/netscaler-observability-exporter/deploy-coe)
          6. [NetScaler IPAM Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/ipam-installation.html)

          ### Installation
          Refer [installation instructions](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-operator-openshift).

          This operator version contains:
          1. NetScaler Ingress Controller version [2.2.10](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.2.10)
          2. NetScaler CPX version [14.1-25.111](https://docs.netscaler.com/en-us/updates?product=NetScaler-ADC-%28includes-NetScaler-Gateway%29&version=14.1&build=25.56)
          3. NetScaler GSLB Controller version [2.2.10](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.2.10)
          4. NetScaler Observability Exporter version [1.10.001](https://github.com/citrix/citrix-helm-charts/releases/tag/nsoe-v1.10.001)
          5. NetScaler Metrics Exporter [1.4.9](https://github.com/citrix/citrix-adc-metrics-exporter/releases/tag/1.4.9)
          6. NetScaler IPAM Controller [2.1.2](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.3.15)

          ### Support

          For debugging issues, please use [NetScaler Kubernetes Plugin](https://github.com/netscaler/modern-apps-toolkit/tree/main/netscaler-plugin).

          For questions and support, the following channels are available:
          - [NetScaler community portal](https://community.netscaler.com/s/)
          - [Contact us](https://www.netscaler.com/contact)
        displayName: NetScaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        - ADC
        - Kubernetes
        - OpenShift
        - Citrix
        - Controller
        - Observability
        - Metrices
        links:
        - name: Netscaler Operator
          url: https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-operator-openshift
        maintainers:
        - email: priyanka.sharma@cloud.com
          name: Priyanka Sharma
        - email: arijit.ray@cloud.com
          name: Arijit Ray
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: NetScaler
          url: https://netscaler.com
        relatedImages:
        - quay.io/netscaler/netscaler-cpx@sha256:232596eaacda68c0e2e210f26b64cedc8d70ae12c5e8ae2e1a66896a479ae4f8
        - quay.io/netscaler/netscaler-observability-exporter@sha256:626e3a954c67bbb910477352dc740aaab21ff1c1cfb03354becf50beab17d4a4
        - quay.io/netscaler/netscaler-ipam-controller@sha256:7c25b59a2e1df0e4f5615a779a9bea939c60a3eca065cb6c2295daf0d03831d8
        - quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f
        - quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:89f44c45373da73a51b0b9a02de7efe519dc2d5343b65fe382526cccc92f8f5c
        - registry.connect.redhat.com/citrix/netscaler-operator@sha256:e367c8c8ba6171e9ef5dd496592a5c62f4aa764f93d0cb78835aa29024c063d0
        - quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4
        version: 3.2.0
      entries:
      - name: netscaler-operator.v3.2.0
        version: 3.2.0
      - name: netscaler-operator.v3.1.2
        version: 3.1.2
      - name: netscaler-operator.v3.1.1
        version: 3.1.1
      - name: netscaler-operator.v3.0.1
        version: 3.0.1
      name: stable
    - currentCSV: netscaler-operator.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCloudController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netscaler-operator",
                    "app.kubernetes.io/instance": "nscc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscalercloudcontroller",
                    "app.kubernetes.io/part-of": "netscaler-operator"
                  },
                  "name": "nscc"
                },
                "spec": {
                  "azure": {
                    "clientsecret": "",
                    "image": "quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb",
                    "location": "",
                    "pullPolicy": "IfNotPresent",
                    "resourcegroupname": "",
                    "resources": {},
                    "subnetname": "",
                    "subscriptionid": "",
                    "tenantid": "",
                    "vnetname": ""
                  }
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerGslbController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netscaler-operator",
                    "app.kubernetes.io/instance": "nsgc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscalergslbcontroller",
                    "app.kubernetes.io/part-of": "netscaler-operator"
                  },
                  "name": "nsgc"
                },
                "spec": {
                  "gslbController": {
                    "acceptLicense": "Yes",
                    "cleanupGSLBSiteConfig": false,
                    "disableAPIServerCertVerify": false,
                    "entityPrefix": "",
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:88114c518ecceb54b000f38bfc330efe184bfeb5b5df0ee308950cf29e7a777f",
                    "imagePullSecrets": [],
                    "kubernetesURL": "",
                    "localCluster": "",
                    "localRegion": "",
                    "logLevel": "Debug",
                    "openshift": true,
                    "pullPolicy": "IfNotPresent",
                    "siteData": [
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteName": "",
                        "siteRegion": ""
                      },
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteName": "",
                        "siteRegion": ""
                      }
                    ]
                  },
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsIP": "",
                    "nsPort": 443,
                    "nsProtocol": "HTTPS"
                  },
                  "resources": {
                    "limits": {},
                    "requests": {
                      "cpu": "32m",
                      "memory": "128Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-operator@sha256:3f71f9a46d7b29e3da98a301965f635c03a022b5535b586d7d5c699341f68193
          createdAt: "2024-07-31T06:33:49Z"
          description: This is an operator to install Cloud Native portfolio provided
            by NetScaler
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler Cloud Controller for NetScaler VPX/MPX/SDX
              with Azure Redhat OpenShift in Azure.
            displayName: Netscaler Cloud Controller
            kind: NetscalerCloudController
            name: netscalercloudcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler GSLB Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler GSLB Controller
            kind: NetscalerGslbController
            name: netscalergslbcontrollers.netscaler.com
            version: v1
        description: |-
          NetScaler provides various products to empower Developer/DevOps managing your applications in OpenShift Cluster. Using this operator you can deploy below controllers for NetScaler appliance:
          1. [NetScaler Cloud Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. [NetScaler GSLB Controller ](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/gslb-openshift-operator)

          ### Installation
          Refer [installation instructions](https://github.com/netscaler/netscaler-k8s-ingress-controller/blob/master/docs/deploy/deploy-ns-operator.md).

          This operator version contains:
          1. NetScaler Cloud Controller version [1.2.0](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. NetScaler GSLB Controller version [1.43.7](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v1.43.7)

          ### Support

          For debugging issues, please use [NetScaler Kubernetes Plugin](https://github.com/netscaler/modern-apps-toolkit/tree/main/netscaler-plugin).

          For questions and support, the following channels are available:
          - [NetScaler community portal](https://community.netscaler.com/s/)
          - [Contact us](https://www.netscaler.com/contact)
        displayName: NetScaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        links:
        - name: Netscaler Operator
          url: https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/gslb-openshift-operator
        maintainers:
        - email: priyanka.sharma@cloud.com
          name: Priyanka Sharma
        - email: arijit.ray@cloud.com
          name: Arijit Ray
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: NetScaler
          url: https://netscaler.com
        relatedImages:
        - quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:88114c518ecceb54b000f38bfc330efe184bfeb5b5df0ee308950cf29e7a777f
        - quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:23405618050a4ceea8ea0c748cc7d9d16dbb80f10e4c39f7e82014a3ac80cec9
        - registry.connect.redhat.com/citrix/netscaler-operator@sha256:3f71f9a46d7b29e3da98a301965f635c03a022b5535b586d7d5c699341f68193
        version: 3.0.1
      entries:
      - name: netscaler-operator.v3.0.1
        version: 3.0.1
      name: tech-preview
    defaultChannel: stable
    packageName: netscaler-operator
    provider:
      name: NetScaler
      url: https://netscaler.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: open-liberty-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: open-liberty-operator.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4",
                  "expose": true,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/open-liberty-operator-controller@sha256:18f188e48725c785557c8edea46f69c464d6f81ab47860a3bb0ec978c20a26db
          createdAt: "2022-03-25T14:00:00Z"
          description: Deploy and manage applications running on Liberty
          olm.skipRange: <0.8.1
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          This advanced Operator can be used to deploy and manage Open Liberty applications with consistent, production-grade QoS. This operator is based on the [Runtime Component Operator](https://github.com/application-stacks/runtime-component-operator) and provides all of its capabilities in addition to Open Liberty specific features such as gathering traces and dumps (Day-2 operations) and easily configuring and managing the single sign-on information for your Open Liberty applications.

          Open Liberty Operator enables enterprise architects to govern the way their applications get deployed & managed in the cluster, while dramatically reducing the learning curve for developers to deploy into Kubernetes - allowing them to focus on writing the code!

          Here are some key features:

          #### Application Lifecyle
          You can deploy your Open Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the Operator will watch for any updates and will re-deploy the modified image.

          #### Custom RBAC
          This Operator is capable of using a custom ServiceAccount from the caller, allowing it to follow RBAC restrictions. By default it creates a ServiceAccount if one is not specified, which can also be bound with specific roles.

          #### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as: labels, annotations, and environment variables from a ConfigMap, a Secret or a value.

          #### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configuration, such as TLS settings, are also easily enabled.

          #### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto scaling to create (and delete) instances based on resource consumption.

          #### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you.
          This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart.
          We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          You can also easily configure and use a single storage for serviceability related Day-2 operations, such as gatherig server traces and dumps.

          #### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating kubernetes Secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to its required services without any intervention or interruption.

          #### Single Sign-On (SSO)
          Open Liberty provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows to easily configure and manage the single sign-on information for your applications.

          #### Exposing metrics to Prometheus
          Expose the Open Liberty application's metrics via the Prometheus Operator.
          You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `spec.monitoring` field to control configurations such as poll interval and security credentials.

          #### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply add the following configuration (to specify the volume size and the location to persist) to your OpenLibertyApplication CR:
          ``` storage: size: 2Gi mountPath: "/logs" ```

          #### Integration with OpenShift Serverless
          Deploy your serverless runtime component using a single toggle.  The Operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing your pod to automatically scale to 0 when it is idle.

          #### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and how they are connected.

          See our [**documentation**](https://github.com/OpenLiberty/open-liberty-operator/tree/main/doc/) for more information.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://github.com/OpenLiberty/open-liberty-operator/tree/main/doc
        - name: Open Liberty
          url: https://openliberty.io/
        - name: Open Liberty Container Images
          url: https://github.com/OpenLiberty/ci.docker
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: beta
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/open-liberty-operator-controller@sha256:18f188e48725c785557c8edea46f69c464d6f81ab47860a3bb0ec978c20a26db
        - registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4
        version: 0.8.1
      entries:
      - name: open-liberty-operator.v0.8.1
        version: 0.8.1
      - name: open-liberty-operator.v0.8.0
        version: 0.8.0
      name: beta2
    - currentCSV: open-liberty-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:d45c296d43dc4c4100f38bb736f073e04c8b4d72b68ddf1189e311aff8ad66e3
          createdAt: "2023-08-28T16:39:10Z"
          description: Deploy and manage containerized Liberty applications
          olm.skipRange: '>=0.8.0 <1.2.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:d45c296d43dc4c4100f38bb736f073e04c8b4d72b68ddf1189e311aff8ad66e3
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02
        version: 1.2.2
      entries:
      - name: open-liberty-operator.v1.2.2
        version: 1.2.2
      name: v1.2
    - currentCSV: open-liberty-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:d68bb66a8133702a198f7be655d7cff5f500bb620c460a81547971e00124ef73
          createdAt: "2024-08-01T20:19:13Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.3.3'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:d68bb66a8133702a198f7be655d7cff5f500bb620c460a81547971e00124ef73
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91
        version: 1.3.3
      entries:
      - name: open-liberty-operator.v1.3.3
        version: 1.3.3
      name: v1.3
    - currentCSV: open-liberty-operator.v1.4.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:67b5b46dfed008dab4ea7fd176d3bc31dae8940d72194a10aebb98b1dea9043e
          createdAt: "2025-07-18T20:45:06Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.4.4'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:67b5b46dfed008dab4ea7fd176d3bc31dae8940d72194a10aebb98b1dea9043e
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3
        version: 1.4.4
      entries:
      - name: open-liberty-operator.v1.4.4
        version: 1.4.4
      name: v1.4
    - currentCSV: open-liberty-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyPerformanceData",
                "metadata": {
                  "name": "openliberty-performance-data-sample"
                },
                "spec": {
                  "interval": 30,
                  "podName": "Specify_Pod_Name_Here",
                  "timespan": 240
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:c3b031fe14054f65d6612cc505b7681336873dc32ce1091451d8c7799034bc83
          createdAt: "2025-09-16T03:08:31Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.5.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server performance data
            displayName: OpenLibertyPerformanceData
            kind: OpenLibertyPerformanceData
            name: openlibertyperformancedata.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:c3b031fe14054f65d6612cc505b7681336873dc32ce1091451d8c7799034bc83
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8
        version: 1.5.0
      entries:
      - name: open-liberty-operator.v1.5.0
        version: 1.5.0
      name: v1.5
    defaultChannel: v1.5
    packageName: open-liberty-certified
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: RabbitMQ
      provider-url: https://rabbitmq.com
    name: rabbitmq-cluster-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-cluster-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rabbitmq.com/v1beta1",
                "kind": "RabbitmqCluster",
                "metadata": {
                  "name": "sample",
                  "namespace": "rabbitmq-system"
                },
                "spec": {
                  "replicas": 1,
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-07-10T12:52:23Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RabbitmqCluster is the Schema for the RabbitmqCluster API.
              Each instance of this object corresponds to a single RabbitMQ cluster.
            displayName: Rabbitmq Cluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: RabbitMQ Cluster Operator
        displayName: RabbitMQ Cluster Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        links:
        - name: Cluster Operator
          url: https://github.com/rabbitmq/cluster-operator
        maturity: alpha
        provider:
          name: RabbitMQ
          url: https://rabbitmq.com
        relatedImages:
        - registry.redhat.io/rhoso-operators/rabbitmq-cluster-rhel9-operator@sha256:e4b198236d70b7dfdd29d8ed45058dec3335b3530e0b676fcc862bfec2efce79
        version: 1.0.6
      entries:
      - name: rabbitmq-cluster-operator.v1.0.6
        version: 1.0.6
      - name: rabbitmq-cluster-operator.v1.0.5
        version: 1.0.5
      - name: rabbitmq-cluster-operator.v1.0.4
        version: 1.0.4
      - name: rabbitmq-cluster-operator.v1.0.3
        version: 1.0.3
      - name: rabbitmq-cluster-operator.v1.0.2
        version: 1.0.2
      - name: rabbitmq-cluster-operator.v1.0.1
        version: 1.0.1
      - name: rabbitmq-cluster-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: rabbitmq-cluster-operator
    provider:
      name: RabbitMQ
      url: https://rabbitmq.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-route53resolver-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-route53resolver-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "route53resolver.services.k8s.aws/v1alpha1",
                "kind": "ResolverEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "route53resolver.services.k8s.aws/v1alpha1",
                "kind": "ResolverRule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/route53resolver-controller:1.1.1
          createdAt: "2025-09-20T01:21:08Z"
          description: AWS Route53Resolver controller is a service controller for
            managing Route53Resolver resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ResolverEndpoint represents the state of an AWS route53resolver
              ResolverEndpoint resource.
            displayName: ResolverEndpoint
            kind: ResolverEndpoint
            name: resolverendpoints.route53resolver.services.k8s.aws
            version: v1alpha1
          - description: ResolverRule represents the state of an AWS route53resolver
              ResolverRule resource.
            displayName: ResolverRule
            kind: ResolverRule
            name: resolverrules.route53resolver.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Route53Resolver resources in AWS from within your Kubernetes cluster.

          **About Amazon Route53Resolver**

          Amazon Route 53 Resolver responds recursively to DNS queries from AWS resources for public records, Amazon VPC-specific DNS names, and Amazon Route 53 private hosted zones, and is available by default in all VPCs.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Route53Resolver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - route53resolver
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Route53Resolver Developer Resources
          url: https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/resolver.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: route53resolver maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/route53resolver-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-route53resolver-controller.v1.1.1
        version: 1.1.1
      - name: ack-route53resolver-controller.v1.0.15
        version: 1.0.15
      - name: ack-route53resolver-controller.v1.0.14
        version: 1.0.14
      - name: ack-route53resolver-controller.v1.0.13
        version: 1.0.13
      - name: ack-route53resolver-controller.v1.0.11
        version: 1.0.11
      - name: ack-route53resolver-controller.v1.0.10
        version: 1.0.10
      - name: ack-route53resolver-controller.v1.0.9
        version: 1.0.9
      - name: ack-route53resolver-controller.v1.0.7
        version: 1.0.7
      - name: ack-route53resolver-controller.v1.0.6
        version: 1.0.6
      - name: ack-route53resolver-controller.v1.0.5
        version: 1.0.5
      - name: ack-route53resolver-controller.v1.0.3
        version: 1.0.3
      - name: ack-route53resolver-controller.v1.0.2
        version: 1.0.2
      - name: ack-route53resolver-controller.v1.0.0
        version: 1.0.0
      - name: ack-route53resolver-controller.v0.0.11
        version: 0.0.11
      - name: ack-route53resolver-controller.v0.0.10
        version: 0.0.10
      - name: ack-route53resolver-controller.v0.0.9
        version: 0.0.9
      - name: ack-route53resolver-controller.v0.0.8
        version: 0.0.8
      - name: ack-route53resolver-controller.v0.0.7
        version: 0.0.7
      - name: ack-route53resolver-controller.v0.0.4
        version: 0.0.4
      - name: ack-route53resolver-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-route53resolver-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Weaveworks
      provider-url: ""
    name: tf-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tf-controller.v0.9.0-rc.8
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"infra.contrib.fluxcd.io/v1alpha1\",\n
            \   \"kind\": \"Terraform\",\n    \"metadata\": {\n        \"name\": \"helloworld-tf\",\n
            \       \"namespace\": \"flux-system\"\n    },\n    \"spec\": {\n        \"interval\":
            \"5m\",\n        \"approvePlan\": \"auto\",\n        \"path\": \"./\",\n
            \       \"sourceRef\": {\n          \"kind\": \"GitRepository\",\n          \"name\":
            \"helloworld\",\n          \"namespace\": \"flux-system\"\n        }      \n
            \   }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/weaveworks/tf-controller:v0.9.0-rc.8
          description: Weave TF-controller is a Flux controller that brings GitOps
            to Terraform resources.
          operatorframework.io/suggested-namespace: flux-system
          support: Weaveworks
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Terraform
            displayName: Terraform
            kind: Terraform
            name: terraforms.infra.contrib.fluxcd.io
            version: v1alpha1
        description: "**Weave GitOps Terraform Controller** (aka Weave TF-controller)
          is a Flux controller for reconciling Terraform resources in the GitOps way.
          \nWith the power of Flux together with Terraform, TF-controller allows you
          to GitOps-ify infrastructure, \nand application resources, in the Kubernetes
          and Terraform universe, together at your own pace.\n\nTF-controller offers
          many GitOps models:\n  1. **GitOps Automation Model:** GitOps your Terraform
          resources from the provision steps to the enforcement steps, like a whole
          EKS cluster.\n  2. **Hybrid GitOps Automation Model:** GitOps parts of your
          existing infrastructure resources. For example, you have an existing EKS
          cluster. You can choose to GitOps only its nodegroup, or its security group.\n
          \ 3. **State Enforcement Model:** You have a TFSTATE file, and you'd like
          to use GitOps enforce it, without changing anything else.\n  4. **Drift
          Detection Model:** You have a TFSTATE file, and you'd like to use GitOps
          just for drift detection, so you can decide to do things later when a drift
          occurs.\n\nFeatures\n  * GitOps Automation for Terraform\n  * Drift detection\n
          \ * Plan and Manual Approve"
        displayName: Weave GitOps Terraform Controller
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - weave
        - flux
        - cd
        - continuous delivery
        - terraform
        - gitops
        links:
        - name: Weave GitOps Terraform Controller
          url: https://github.com/weaveworks/tf-controller
        - name: Documentation
          url: https://weaveworks.github.io/tf-controller/
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Weaveworks
        relatedImages:
        - ghcr.io/weaveworks/tf-controller:v0.9.0-rc.8
        version: 0.9.0-rc.8
      entries:
      - name: tf-controller.v0.9.0-rc.8
        version: 0.9.0-rc.8
      - name: tf-controller.v0.9.0-rc.3
        version: 0.9.0-rc.3
      name: stable
    defaultChannel: stable
    packageName: tf-controller
    provider:
      name: Weaveworks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Randoli
      provider-url: https://www.randoli.io
    name: app-insights-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: app-insights-operator.v1.17.5-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "appinsights.randoli/v1beta1",
                "kind": "RandoliAgent",
                "metadata": {
                  "name": "randoli",
                  "namespace": "randoli-agents"
                },
                "spec": {
                  "costManagement": {
                    "opencost": {
                      "exporter": {
                        "image": {
                          "tag": "1.112.1"
                        }
                      }
                    },
                    "vpaOperator": {
                      "enabled": false
                    }
                  },
                  "logs": {
                    "vector": {
                      "agent": {
                        "containerSecurityContext": {
                          "privileged": true
                        }
                      }
                    }
                  },
                  "network": {
                    "standaloneConsole": {
                      "enable": true
                    }
                  },
                  "observability": {
                    "netobserv": {
                      "enabled": false
                    },
                    "otel": {
                      "operator": {
                        "enabled": false
                      }
                    }
                  },
                  "resources": {
                    "limits": {
                      "memory": "2Gi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "512Mi"
                    }
                  },
                  "securityContext": {
                    "fsGroup": 1001
                  },
                  "tags": {
                    "costManagement": false,
                    "observability": false
                  }
                }
              }
            ]
          categories: Developer Tools
          containerImage: docker.io/randoli/app-insights-operator:1.17.5-4@sha256:22e7150c29557f2a71186389e5d7d60bfe82a4ea022e1403033c6ad61bc95f7a
          createdAt: "2025-09-23T19:45:43Z"
          description: Randoli Agent connects OpenShift clusters to Randoli Observability
            & Cost Management.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[]'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Randoli
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: RandoliAgent
            name: randoliagents.appinsights.randoli
            version: v1beta1
        description: |-
          Randoli Agent connects OpenShift clusters to Randoli Observability & Cost Management platforms.

          This operator provides a container-based agent to be installed in your OpenShift clusters for connecting to Randoli Observability & Cost Management platforms

          **Prerequisites**

          If you are using the Randoli Observability platform, you will need to install the following Operators:
          - Red Hat build of OpenTelemetry Operator
          - Network Observability

          If you are using the Randoli Cost Management platform, you will need to install the following Operator:
          - VerticalPodAutoscaler

          For additional information please use our Openshift installation guide here: [https://docs.insights.randoli.io/agent/openshift](https://docs.insights.randoli.io/agent/openshift)

           **Randoli Observability for Kubernetes**

          [Product Page](https://www.randoli.io/product/kubernetes-observability)

          **Randoli Cost Management for Kubernetes**

          [Product Page](https://www.randoli.io/product/kubernetes-cost-management)
        displayName: Randoli Agent
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - developer tools
        - insights
        - monitoring
        links:
        - name: Randoli Observability for Kubernetes
          url: https://www.randoli.io/product/kubernetes-observability
        - name: Randoli Cost Management for Kubernetes
          url: https://www.randoli.io/product/kubernetes-cost-management
        maintainers:
        - email: contact@randoli.ca
          name: Randoli Inc.
        maturity: beta
        minKubeVersion: 1.24.0
        provider:
          name: Randoli
          url: https://www.randoli.io
        relatedImages:
        - docker.io/randoli/app-insights-operator:1.17.5-4@sha256:22e7150c29557f2a71186389e5d7d60bfe82a4ea022e1403033c6ad61bc95f7a
        version: 1.17.5-4
      entries:
      - name: app-insights-operator.v1.17.5-4
        version: 1.17.5-4
      - name: app-insights-operator.v1.17.5-3
        version: 1.17.5-3
      - name: app-insights-operator.v1.17.5-2
        version: 1.17.5-2
      - name: app-insights-operator.v1.17.5-1
        version: 1.17.5-1
      - name: app-insights-operator.v1.17.3-1
        version: 1.17.3-1
      - name: app-insights-operator.v1.17.2-1
        version: 1.17.2-1
      - name: app-insights-operator.v1.17.1-1
        version: 1.17.1-1
      - name: app-insights-operator.v1.17.0-2
        version: 1.17.0-2
      - name: app-insights-operator.v1.17.0-1
        version: 1.17.0-1
      name: stable
    defaultChannel: stable
    packageName: app-insights-operator
    provider:
      name: Randoli
      url: https://www.randoli.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ProphetStor Data Services, Inc.
      provider-url: ""
    name: federatorai-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: federatorai.v4.7.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "federatorai.containers.ai/v1alpha1",
                "kind": "AlamedaService",
                "metadata": {
                  "name": "my-alamedaservice"
                },
                "spec": {
                  "alamedaAi": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "10Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "alamedaInfluxdb": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "100Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "fedemeterInfluxdb": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "10Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "imageLocation": "registry.connect.redhat.com/prophetstor",
                  "storages": [
                    {
                      "accessMode": ["ReadWriteOnce"],
                      "size": "2Gi",
                      "type": "pvc",
                      "usage": "log"
                    }
                  ],
                  "version": "v4.7.1-b1760"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: registry.connect.redhat.com/prophetstor/federatorai-operator@sha256:bdd87f6b7d7ba9d12ed8ddb9c5151e5a89781e88d7a02b73814b16a766968047
          createdAt: "2021-10-12T01:14:00Z"
          description: Federator.ai Operator provides easy configuration and management
            of AI-based Kubernetes resource orchestrator
          operatorframework.io/suggested-namespace: federatorai
          repository: https://access.redhat.com/containers/#/registry.connect.redhat.com/prophetstor/federatorai-operator
          support: ProphetStor Data Services, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of Federator.ai.
            displayName: AlamedaService
            kind: AlamedaService
            name: alamedaservices.federatorai.containers.ai
            version: v1alpha1
        description: |-
          Federator.ai helps enterprises optimize cloud resources, maximize application performance, and save significant cost without excessive over-provisioning or under-provisioning of resources, meeting the service-level requirements of their applications.

          Enterprises often lack understanding of the resources needed to support their applications. This leads to either excessive over-provisioning or under-provisioning of resources (CPU, memory, storage). Using machine learning, Federator.ai determines the optimal cloud resources needed to support any workload on OpenShift and helps users find the best-cost instances from cloud providers for their applications.


          **Multi-layer workload prediction**

          Using machine learning and math-based algorithms, Federator.ai predicts containerized application and cluster node resource usage as the basis for resource recommendations at application level as well as at cluster node level. Federator.ai supports prediction for both physical/virtual CPUs and memories.


          **Auto-scaling via resource recommendation**

          Federator.ai utilizes the predicted resource usage to recommend the right number and size of pods for applications. Integrated with Datadog's WPA, applications are automatically scaled to meet the predicted resource usage.


          **Application-aware recommendation execution**

          Optimizing the resource usage and performance goals, Federator.ai uses application specific metrics for workload prediction and pod capacity estimation to auto-scale the right number of pods for best performance without overprovisioning.


          **Multi-cloud Cost Analysis**

          With resource usage prediction, Federator.ai analyzes potential cost of a cluster on different public cloud providers. It also recommend appropriate cluster nodes and instance types based on resource usage.


          **Custom Datadog/Sysdig Dashboards**

          Predefined custom Datadog/Sysdig Dashboards for workload prediction/recommendation visualization for cluster nodes and applications.


          **Additional resources**

          Want more product information? Explore detailed information about using this product and where to find additional help.

          * [Federator.ai Datasheet](https://prophetstor.com/wp-content/uploads/datasheets/Federator.ai.pdf)
          * [Quick Start Guide](https://prophetstor.com/wp-content/uploads/documentation/Federator.ai/Latest%20Version/ProphetStor%20Federator.ai%20Quick%20Installation%20Guide.pdf)
          * [Installation Guide](https://prophetstor.com/wp-content/uploads/2021/08/ProphetStor-Federator.ai-v4.7.0-Installation-Guide-v1.0.pdf)
          * [User Guide](https://prophetstor.com/wp-content/uploads/2021/08/Federator.ai-4.7.0-User-Guide.pdf)
          * [Release Notes](https://prophetstor.com/wp-content/uploads/2021/08/Federator.ai-4.7.0-Release-Notes.pdf)
          * [Company Information](https://www.prophetstor.com/)


          ### Prerequisite
          *  The [Openshift](https://www.openshift.com) version 4.4.x or later if using OpenShift platform.
          *  The [Kubernetes](https://kubernetes.io/) version 1.16 or later if using Kubernetes environment.
        displayName: Federator.ai
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - AI
        - Resource Orchestration
        - NoOps
        links:
        - name: Website
          url: https://prophetstor.com/federator-ai-2/
        - name: Quickstart guide
          url: https://prophetstor.com/wp-content/uploads/documentation/Federator.ai/Latest%20Version/ProphetStor%20Federator.ai%20Quick%20Installation%20Guide.pdf
        maintainers:
        - email: support@prophetstor.com
          name: ProphetStor Data Services, Inc.
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: ProphetStor Data Services, Inc.
        relatedImages:
        - registry.connect.redhat.com/prophetstor/federatorai-operator@sha256:bdd87f6b7d7ba9d12ed8ddb9c5151e5a89781e88d7a02b73814b16a766968047
        - registry.connect.redhat.com/prophetstor/alameda-recommender-ubi@sha256:13e941df9252cea3024ecb5e2247a696b64928015c1a3a4a04ee4e362388b1e4
        - registry.connect.redhat.com/prophetstor/federatorai-data-adapter@sha256:adf5d655fa030dc05133516d45642d7282fdd66e0aa8e7a7331ed614a41d3fe0
        - registry.connect.redhat.com/prophetstor/alameda-operator-ubi@sha256:9a863394cb1c89228223b26addb0d12bb91a97233fe2d5d19ecd5998e0cfe0d6
        - registry.connect.redhat.com/prophetstor/federatorai-agent-ubi@sha256:cfadc84f31f3776ded15d4d8dda92e00b7d9b1bce0ff5440277176292f9304d3
        - registry.connect.redhat.com/prophetstor/federatorai-agent-preloader@sha256:66daea6193e399f8363df5ba42e1fcd643db6df3a9075dbcd8c3464fdc91e776
        - registry.connect.redhat.com/prophetstor/federatorai-rest-ubi@sha256:3e92528615543ea900bf38387c4eb8439b60759be5be2ac5311a9dbaafc7026b
        - registry.connect.redhat.com/prophetstor/alameda-rabbitmq@sha256:b396d9aa0470e593a36f5799c04fc3c9e59b5723f703f4b019f7ef92528e6741
        - registry.connect.redhat.com/prophetstor/alameda-notifier-ubi@sha256:1f75034fb78d62924a87cb4a5435661b4afa990d6960431e784029742fbae0da
        - registry.connect.redhat.com/prophetstor/alameda-ai-dispatcher@sha256:6e0dc1136b0f8e5971d92b95ce0d6dfac475044c073bada50f23b315f4478ae4
        - registry.connect.redhat.com/prophetstor/alameda-ai@sha256:9293298bb2a72d0b0764050b6f6d8da707b46a83363a3b6aa7be60f512e2ec61
        - registry.connect.redhat.com/prophetstor/alameda-analyzer-ubi@sha256:6117ac4692793d49bc0204f13091dec119bdbc7b169e9deb86c243aca00ded27
        - registry.connect.redhat.com/prophetstor/federatorai-dashboard-backend@sha256:022dbf6696edb8eaf389fdbd1dff14c40a51c237c01f8a02a7ea65c60cd59f38
        - registry.connect.redhat.com/prophetstor/alameda-datahub-ubi@sha256:3819d6acce8d376dbe3b12469be2528d0ce9ac95c2a86507aeb76b6e3a5f5f38
        - registry.connect.redhat.com/prophetstor/alameda-executor-ubi@sha256:4512e3be7a89d0ca9f120b68f6ab9766d000bf2b01b279cd55b2dbc433aacfb6
        - registry.connect.redhat.com/prophetstor/fedemeter-api-ubi@sha256:451456207fe8d16b9824b3f64d00645766dcccbc8d685b6942e6c00d0d0c627e
        - registry.connect.redhat.com/prophetstor/federatorai-dashboard-frontend@sha256:1c1147425ca6302103363a1a3bcf27baeef13b10ec6780ef266b7f52fe27ffeb
        - registry.connect.redhat.com/prophetstor/fedemeter-influxdb@sha256:36904f003e8e29d4a9fd738655f401fbe840d60158d7d820cda79a09858dc2c2
        - registry.connect.redhat.com/prophetstor/alameda-influxdb@sha256:485c2119a6a56807e28ad03b2087129950118b55531126a66bb414063052ae76
        version: 4.7.1-1
      entries:
      - name: federatorai.v4.7.1-1
        version: 4.7.1-1
      - name: federatorai.v4.6.1-4
        version: 4.6.1-4
      name: stable
    defaultChannel: stable
    packageName: federatorai-certified
    provider:
      name: ProphetStor Data Services, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Starburst Data, Inc.
      provider-url: https://www.starburstdata.com
    name: starburst-enterprise-helm-operator-paygo-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: starburst-enterprise-helm-operator-paygo.v402.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstEnterprise",
                "metadata": {
                  "name": "starburstenterprise-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "cache": {
                    "diskUsagePercentage": 80,
                    "enabled": false,
                    "ttl": "7d",
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "catalogs": {
                    "tpch": "connector.name=tpch\ntpch.splits-per-node=4"
                  },
                  "commonLabels": {},
                  "coordinator": {
                    "additionalProperties": "",
                    "affinity": {},
                    "deploymentAnnotations": {},
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=true\nnode-scheduler.include-coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://localhost:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "resources": {
                      "cpu": 16,
                      "memory": "60Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "tolerations": []
                  },
                  "environment": null,
                  "expose": {
                    "clusterIp": {
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      }
                    },
                    "ingress": {
                      "annotations": {},
                      "host": null,
                      "ingressClassName": null,
                      "ingressName": "coordinator-ingress",
                      "path": "/",
                      "pathType": "ImplementationSpecific",
                      "serviceName": "starburst",
                      "servicePort": 8080,
                      "tls": {
                        "enabled": true,
                        "secretName": null
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "nodePort": 30080,
                          "port": 8080
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "externalSecrets": {
                    "enabled": false,
                    "goDaddy": {
                      "backendType": "secretsManager"
                    },
                    "secretPrefix": "external0",
                    "type": "goDaddy"
                  },
                  "extraArguments": [],
                  "extraSecret": {
                    "file": null,
                    "name": null
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise",
                    "tag": "402-e.1"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "initImage": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise-init",
                    "tag": "1.3.2"
                  },
                  "internal": {
                    "ports": {
                      "http": {
                        "port": 8080
                      },
                      "https": {
                        "port": 8443
                      }
                    }
                  },
                  "internalTls": false,
                  "licenseManager": {
                    "cronjob": {
                      "affinity": {},
                      "nodeSelector": {},
                      "priorityClassName": null,
                      "securityContext": {},
                      "tolerations": []
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "registry.connect.redhat.com/starburst/license-manager",
                      "tag": "0.0.2"
                    },
                    "url": "https://license-api.dogfood.eng.starburstdata.net/test/license/checkout"
                  },
                  "livenessProbe": null,
                  "prometheus": {
                    "agent": {
                      "config": "/etc/starburst/telemetry/prometheus.yaml",
                      "port": 8081,
                      "version": "0.16.1"
                    },
                    "enabled": true,
                    "rules": [
                      {
                        "attrNameSnakeCase": true,
                        "name": "$1",
                        "pattern": "trino.execution<name=QueryManager><>(running_queries|queued_queries)",
                        "type": "GAUGE"
                      },
                      {
                        "name": "failed_queries",
                        "pattern": "trino.execution<name=QueryManager><>FailedQueries\\.TotalCount",
                        "type": "COUNTER"
                      }
                    ]
                  },
                  "query": {
                    "maxConcurrentQueries": 3
                  },
                  "readinessProbe": null,
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "securityContext": {},
                  "sharedSecret": null,
                  "siac": {
                    "additionalHeapSizePercentage": 15,
                    "enabled": false
                  },
                  "spilling": {
                    "enabled": false,
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "startupProbe": null,
                  "usageMetrics": {
                    "enabled": false,
                    "usageClient": {
                      "initialDelay": "1m",
                      "interval": "1m"
                    }
                  },
                  "userDatabase": {
                    "enabled": false,
                    "users": [
                      {
                        "password": "46991b33f7a75ff79213c0dc0e610610",
                        "username": "admin"
                      }
                    ]
                  },
                  "worker": {
                    "additionalProperties": "",
                    "affinity": {},
                    "autoscaling": {
                      "enabled": false,
                      "maxReplicas": 100,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "deploymentAnnotations": {},
                    "deploymentTerminationGracePeriodSeconds": 300,
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://coordinator:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "kedaScaler": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.connect.redhat.com/starburst/keda-trino-scaler",
                        "tag": "0.1.7"
                      },
                      "port": 8021,
                      "scaledObjectSpec": {
                        "advanced": {},
                        "cooldownPeriod": 300,
                        "idleReplicaCount": 0,
                        "maxReplicaCount": 100,
                        "minReplicaCount": 1,
                        "pollingInterval": 30,
                        "scaleTargetRef": {
                          "name": "worker"
                        },
                        "triggers": [
                          {
                            "metadata": {
                              "numberOfQueriesPerWorker": "10",
                              "scaleInToIdleReplicaCountIfNoQueuedQueriesLeft": "true",
                              "scaleMethod": "query_queue"
                            },
                            "type": "external"
                          }
                        ]
                      }
                    },
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "replicas": 2,
                    "resources": {
                      "cpu": 16,
                      "memory": "100Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "starburstWorkerShutdownGracePeriodSeconds": 120,
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstHive",
                "metadata": {
                  "name": "starbursthive-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "affinity": {},
                  "commonLabels": {},
                  "database": {
                    "external": {
                      "driver": null,
                      "jdbcUrl": null,
                      "password": null,
                      "user": null
                    },
                    "internal": {
                      "databaseName": "hive",
                      "databasePassword": "HivePass1234",
                      "databaseUser": "hive",
                      "driver": "org.postgresql.Driver",
                      "env": [],
                      "envFrom": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.redhat.io/rhel8/postgresql-10",
                        "tag": "1"
                      },
                      "port": 5432,
                      "resources": {
                        "limits": {
                          "cpu": 2,
                          "memory": "1Gi"
                        },
                        "requests": {
                          "cpu": 2,
                          "memory": "1Gi"
                        }
                      },
                      "volume": {
                        "emptyDir": {}
                      }
                    },
                    "type": "internal"
                  },
                  "deploymentAnnotations": {},
                  "env": [],
                  "envFrom": [],
                  "expose": {
                    "clusterIp": {
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "nodePort": 30083,
                          "port": 9083
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "extraArguments": [],
                  "hdfs": {
                    "hadoopUserName": null
                  },
                  "heapSizePercentage": 85,
                  "hiveMetastoreStorageAuthorization": true,
                  "hiveMetastoreWarehouseDir": null,
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/hive",
                    "tag": "3.1.2-e.22"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "nodeSelector": {},
                  "objectStorage": {
                    "adl": {
                      "oauth2": {
                        "clientId": null,
                        "credential": null,
                        "refreshUrl": null
                      }
                    },
                    "awsS3": {
                      "accessKey": null,
                      "endpoint": null,
                      "pathStyleAccess": false,
                      "region": null,
                      "secretKey": null
                    },
                    "azure": {
                      "abfs": {
                        "accessKey": {
                          "accessKey": null,
                          "storageAccount": null
                        },
                        "authType": "accessKey",
                        "oauth": {
                          "clientId": null,
                          "endpoint": null,
                          "secret": null
                        }
                      },
                      "wasb": {
                        "accessKey": null,
                        "storageAccount": null
                      }
                    },
                    "gs": {
                      "cloudKeyFileSecret": null
                    }
                  },
                  "podAnnotations": {},
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "resources": {
                    "limits": {
                      "cpu": 1,
                      "memory": "1Gi"
                    },
                    "requests": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "securityContext": {},
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
          description: Deploys StarburstEnterprise Helm Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-paygo-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-paygo-rhmp/support?utm_source=openshift_console
          marketplace.redhat.com/meterDefinition: |-
            {
              "apiVersion": "marketplace.redhat.com/v1beta1",
              "kind": "MeterDefinition",
              "metadata": {
                "name": "sep-vcpu-usage"
              },
              "spec": {
                "group": "charts.starburstdata.com",
                "kind": "StarburstEnterprise",
                "resourceFilters": [
                  {
                    "namespace": {
                      "useOperatorGroup": true
                    },
                    "ownerCRD": {
                      "apiVersion": "charts.starburstdata.com/v1alpha1",
                      "kind": "StarburstEnterprise"
                    },
                    "workloadType": "Pod"
                  }
                ],
                "meters": [
                  {
                    "aggregation": "sum",
                    "period": "1h",
                    "metricId": "container_cpu_usage_core_avg",
                    "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=~\"coordinator|worker\"}[5m])*100",
                    "workloadType": "Pod"
                  }
                ]
              }
            }
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Starburst Data
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: StarburstEnterprise
            name: starburstenterprises.charts.starburstdata.com
            version: v1alpha1
          - kind: StarburstHive
            name: starbursthives.charts.starburstdata.com
            version: v1alpha1
        description: CRs for deploying starburst enterprise and hive
        displayName: Starburst Enterprise Pay as you go Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - starburst
        - trino
        - sep
        - bigdata
        - data
        links:
        - name: Starburst Enterprise Helm Operator
          url: https://www.starburstdata.com
        maintainers:
        - email: hello@starburstdata.com
          name: Starburst Data, Inc.
        maturity: alpha
        provider:
          name: Starburst Data, Inc.
          url: https://www.starburstdata.com
        relatedImages:
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
        version: 402.1.0
      entries:
      - name: starburst-enterprise-helm-operator-paygo.v402.1.0
        version: 402.1.0
      name: alpha
    defaultChannel: alpha
    packageName: starburst-enterprise-helm-operator-paygo-rhmp
    provider:
      name: Starburst Data, Inc.
      url: https://www.starburstdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: redhat-oadp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oadp-operator.v1.0.19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "restic": {
                      "enable": true
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws",
                        "kubevirt"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ResticRepository",
                "metadata": {
                  "name": "resticrepository",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/oadp/oadp-rhel8-operator@sha256:caf87a9832c153833feb34affc2b797ad0e643362448742b0405d457e5f0ba41
          createdAt: "2025-04-14T13:25:43Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <1.0.19'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/oadp/oadp-mustgather-rhel8@sha256:af4c6af48f38510c7d52f678798a9affea2211dc488ed18f83a5d005b7ef4b6b
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is a Velero resource that represents the capture of
              Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Data Protection Applications on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://github.com/openshift/oadp-operator
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/oadp/oadp-velero-plugin-for-microsoft-azure-rhel8@sha256:3e81b5e55f96504ba76d9d47bbeec50c7b1fd90b7702fc31b1dba248050ce9b4
        - registry.redhat.io/oadp/oadp-velero-plugin-for-gcp-rhel8@sha256:05a164c27da94b374ce03451c851a607a3cd436b73b9e5b70723c6dc8cacb3d8
        - registry.redhat.io/oadp/oadp-mustgather-rhel8@sha256:af4c6af48f38510c7d52f678798a9affea2211dc488ed18f83a5d005b7ef4b6b
        - registry.redhat.io/oadp/oadp-rhel8-operator@sha256:caf87a9832c153833feb34affc2b797ad0e643362448742b0405d457e5f0ba41
        - registry.redhat.io/oadp/oadp-registry-rhel8@sha256:0886b3b4ce0295080bf2c5ee95f2b7b15a4483bcd94cb739e7e7154d21150397
        - registry.redhat.io/oadp/oadp-velero-restic-restore-helper-rhel8@sha256:5d288335b5ad114b9eb61410befb4a9d1f84f0ba0c01558098bf4567fda9a8c5
        - registry.redhat.io/oadp/oadp-velero-plugin-rhel8@sha256:694603b9c1fd5eec939b28f8727fd7c9c0d380620c5081006f2655e98358464f
        - registry.redhat.io/oadp/oadp-velero-plugin-for-csi-rhel8@sha256:05ec16bdf1f0fadc4ddad244b86934f73569d2e66cc19f03b3553e257441a662
        - registry.redhat.io/oadp/oadp-velero-rhel8@sha256:39dec8b9ccdcabeaab3e56e14b79cfed4d8574cb172a886d017813a437e4ffff
        - registry.redhat.io/oadp/oadp-velero-plugin-for-aws-rhel8@sha256:28045538fff9e5051cd01e56876971cf9401a448e1dd63bff17fbb900629f22f
        - registry.redhat.io/oadp/oadp-kubevirt-velero-plugin-rhel8@sha256:d862a1b0420fcfb91403f90047b00035efba112af6bc5c50adb723014a620767
        version: 1.0.19
      entries:
      - name: oadp-operator.v1.0.19
        version: 1.0.19
      - name: oadp-operator.v1.0.18
        version: 1.0.18
      - name: oadp-operator.v1.0.17
        version: 1.0.17
      - name: oadp-operator.v1.0.16
        version: 1.0.16
      - name: oadp-operator.v1.0.15
        version: 1.0.15
      - name: oadp-operator.v1.0.14
        version: 1.0.14
      - name: oadp-operator.v1.0.13
        version: 1.0.13
      - name: oadp-operator.v1.0.12
        version: 1.0.12
      - name: oadp-operator.v1.0.11
        version: 1.0.11
      - name: oadp-operator.v1.0.10
        version: 1.0.10
      - name: oadp-operator.v1.0.9
        version: 1.0.9
      - name: oadp-operator.v1.0.8
        version: 1.0.8
      - name: oadp-operator.v1.0.7
        version: 1.0.7
      - name: oadp-operator.v1.0.6
        version: 1.0.6
      - name: oadp-operator.v1.0.5
        version: 1.0.5
      - name: oadp-operator.v1.0.4
        version: 1.0.4
      - name: oadp-operator.v1.0.3
        version: 1.0.3
      - name: oadp-operator.v1.0.2
        version: 1.0.2
      - name: oadp-operator.v1.0.1
        version: 1.0.1
      - name: oadp-operator.v1.0.0
        version: 1.0.0
      name: stable-1.0
    - currentCSV: oadp-operator.v1.4.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "nodeAgent": {
                      "enable": true,
                      "uploaderType": "restic"
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws",
                        "kubevirt"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: |
            Cloud Provider,Developer Tools,Modernization & Migration,OpenShift Optional,Storage
          certified: "false"
          containerImage: registry.redhat.io/oadp/oadp-rhel9-operator@sha256:ee010fb34448b772690a8a98808aa25aa1578a81061618f948b1d16b5d63d333
          createdAt: "2025-07-10T04:59:11Z"
          description: OADP (OpenShift API for Data Protection) operator sets up and
            installs Data Protection Applications on the OpenShift platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <1.4.5'
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/oadp/oadp-mustgather-rhel9@sha256:0cee6fecf052bae3ede6bb6f032b25c0c7cd63ced1e26328f2b02144f24fbb12
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A backup repository is an indicator of a connection from
              the restic/kopia server to the backupstoragelocation.
            displayName: BackupRepository
            kind: BackupRepository
            name: backuprepositories.velero.io
            version: v1
          - description: Backup is a Velero resource that represents the capture of
              Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - description: The CloudStorage API automates the creation of a bucket for
              object storage.
            displayName: CloudStorage
            kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataDownload represents a data download of a volume snapshot.
              There is one DataDownload created per volume to be restored.
            displayName: DataDownload
            kind: DataDownload
            name: datadownloads.velero.io
            version: v2alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DataUpload acts as the protocol between data mover plugins
              and data mover controller for the datamover backup operation
            displayName: DataUpload
            kind: DataUpload
            name: datauploads.velero.io
            version: v2alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: |
          **OpenShift API for Data Protection (OADP)** operator sets up and installs
          Velero on the OpenShift platform, allowing users to backup and restore
          applications.

          Backup and restore Kubernetes resources and internal images, at the
          granularity of a namespace, using a version of Velero appropriate for
          the installed version of OADP.

          OADP backs up Kubernetes objects and internal images by saving them as
          an archive file on object storage. OADP backs up persistent volumes
          (PVs) by creating snapshots with the native cloud snapshot API or with
          the Container Storage Interface (CSI). For cloud providers that do not
          support snapshots, OADP backs up resources and PV data with Restic or Kopia.

          * [Installing OADP for application backup and restore][install-doc]

          * [Installing OADP on a ROSA cluster and using STS, please follow the Getting Started Steps 1-3 in order to obtain the role ARN needed for using the standardized STS configuration flow via OLM][rosa-sts-doc]

          * [Frequently Asked Questions][faq-doc]

          [install-doc]: https://docs.openshift.com/container-platform/latest/backup_and_restore/application_backup_and_restore/installing/about-installing-oadp.html
          [rosa-sts-doc]: https://mobb.ninja/docs/misc/oadp/rosa-sts/
          [faq-doc]: https://access.redhat.com/articles/5456281
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - disaster backup
        - disaster recovery
        - disaster restore
        - disaster
        - backup
        - restore
        - backup automation
        - recovery
        - data protection
        - data management
        - data
        - protection
        - management
        - application resilience
        - resilience
        - velero
        - openshift
        - oadp
        - cloud-native
        - replication
        - kopia
        - restic
        - network file system
        - nfs
        - simple cloud storage
        - s3
        - minio
        - cloud storage
        - data foundation
        - ceph
        - csi
        - container storage interface
        links:
        - name: OADP Operator
          url: https://github.com/openshift/oadp-operator
        maintainers:
        - email: oadp-maintainers@redhat.com
          name: OADP Operator maintainers
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/oadp/oadp-velero-plugin-for-legacy-aws-rhel9@sha256:ab5c8d33684a361fa3bb10a006ef489be92cc955ad3d173bbfa77949dcc42aa8
        - registry.redhat.io/oadp/oadp-kubevirt-velero-plugin-rhel9@sha256:c3b32cb9f5b82021c08cd70467e68aad66746ef211d73612ecfe0af625706267
        - registry.redhat.io/oadp/oadp-velero-plugin-for-aws-rhel9@sha256:79d39186cae5696949813a0ea6c46eeededad29e6bde171d7a9b0011c993bb26
        - registry.redhat.io/oadp/oadp-velero-plugin-for-microsoft-azure-rhel9@sha256:9f05592bf0e5b76322ca582bdede387f9f4aabbbdc436549156c05d652c951d0
        - registry.redhat.io/oadp/oadp-mustgather-rhel9@sha256:0cee6fecf052bae3ede6bb6f032b25c0c7cd63ced1e26328f2b02144f24fbb12
        - registry.redhat.io/oadp/oadp-rhel9-operator@sha256:ee010fb34448b772690a8a98808aa25aa1578a81061618f948b1d16b5d63d333
        - registry.redhat.io/oadp/oadp-velero-rhel9@sha256:17160f99940aa355dc91ab3d16ba6bf91faade6423434008cf8f6e5023b1e413
        - registry.redhat.io/oadp/oadp-velero-plugin-for-gcp-rhel9@sha256:00800ead11ea308794b1536d964f8a043e13ede9657cabb23afa42cba7cf2ce6
        - registry.redhat.io/oadp/oadp-velero-plugin-rhel9@sha256:3ee9d3e799cedc28c400d364b3a8790c7e88acd5f603b09ff512a22d91e265b3
        - registry.redhat.io/oadp/oadp-velero-restic-restore-helper-rhel9@sha256:e0647defc7c428f9a04a5f6d4532dc86e57b92f31b4f907253ae14af05c5993f
        version: 1.4.5
      entries:
      - name: oadp-operator.v1.4.5
        version: 1.4.5
      - name: oadp-operator.v1.4.4
        version: 1.4.4
      - name: oadp-operator.v1.4.3
        version: 1.4.3
      - name: oadp-operator.v1.4.2
        version: 1.4.2
      - name: oadp-operator.v1.4.1
        version: 1.4.1
      - name: oadp-operator.v1.4.0
        version: 1.4.0
      name: stable-1.4
    defaultChannel: stable-1.4
    packageName: redhat-oadp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Oracle
      provider-url: https://www.oracle.com/cloud/
    name: oci-ccm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oci-ccm-operator.v0.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ccm.oraclecloud.com/v1alpha1",
                "kind": "OCICloudManager",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "oci-ccm-operator",
                    "app.kubernetes.io/instance": "ocicloudmanager",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ocicloudmanager",
                    "app.kubernetes.io/part-of": "oci-ccm-operator"
                  },
                  "name": "ocicloudmanager"
                },
                "spec": {
                  "compartment": "CompartmentId",
                  "driverNamespace": "kube-system",
                  "loadBalancer": {
                    "securityList": "SecurityListId",
                    "securityListManagementMode": "Frontend",
                    "subnet": "SubnetId"
                  },
                  "vcn": "VcnId"
                }
              },
              {
                "apiVersion": "ccm.oraclecloud.com/v1alpha1",
                "kind": "OCICsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "oci-ccm-operator",
                    "app.kubernetes.io/instance": "ocicsidriver",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ocicsidriver",
                    "app.kubernetes.io/part-of": "oci-ccm-operator"
                  },
                  "name": "ocicsidriver"
                },
                "spec": {
                  "compartment": "CompartmentId",
                  "driverNamespace": "kube-system",
                  "vcn": "VcnId"
                }
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider, Storage
          containerImage: phx.ocir.io/ax1eoqvpvbtg/oci-ccm-operator:0.0.9
          createdAt: "2023-07-28T23:42:18Z"
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Oracle
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: OCICloudManager
            name: ocicloudmanagers.ccm.oraclecloud.com
            version: v1alpha1
          - kind: OCICsiDriver
            name: ocicsidrivers.ccm.oraclecloud.com
            version: v1alpha1
        description: OCI Kubernetes Cloud Controller Manager implementation (or out-of-tree
          cloud-provider) for Oracle Cloud Infrastructure (OCI).
        displayName: oci-ccm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OCI
        - CCM
        - OpenShift
        - Kubernetes
        links:
        - name: Oracle
          url: https://www.oracle.com/cloud/
        - name: OCI Cloud Controller Manager (CCM)
          url: https://github.com/oracle/oci-cloud-controller-manager
        maintainers:
        - email: david.d.campbell@oracle.com
          name: David Campbell
        - email: alex.blinov@oracle.com
          name: Alex Blinov
        maturity: alpha
        provider:
          name: Oracle
          url: https://www.oracle.com/cloud/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - phx.ocir.io/ax1eoqvpvbtg/oci-ccm-operator:0.0.9
        version: 0.0.9
      entries:
      - name: oci-ccm-operator.v0.0.9
        version: 0.0.9
      name: alpha
    defaultChannel: alpha
    packageName: oci-ccm-operator
    provider:
      name: Oracle
      url: https://www.oracle.com/cloud/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: eventing-kogito
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eventing-kogito.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kogito.knative.dev/v1alpha1",
                "kind": "KogitoSource",
                "metadata": {
                  "name": "kogito-order-processing"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "kind": "Broker",
                      "name": "default"
                    }
                  },
                  "subject": {
                    "apiVersion": "serving.knative.dev/v1",
                    "kind": "Service",
                    "name": "serverless-workflow-order-processing"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: gcr.io/knative-releases/knative.dev/eventing-kogito/cmd/controller:v1.2.0
          createdAt: "2021-08-25T12:00:00Z"
          olm.targetNamespaces: knative-kogito
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/knative-sandbox/eventing-kogito
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoSource is the representation of a Knative Eventing
              Source for a Kogito custom Service.
            displayName: Kogito Source
            kind: KogitoSource
            name: kogitosources.kogito.knative.dev
            version: v1alpha1
        description: |-
          Knative Eventing Source for Kogito Services. This controller will enable easy configuration of any Kogito custom Service
          with the Knative Eventing platform. Please find more information about using this source in the [Official Kogito Documentation](https://docs.jboss.org/kogito/release/latest/html_single/#_kogito_knative_eventing_add_on)

          **Note:** The admission webhooks are not installed via OLM at this time. Because of that, please ensure that you create the `KogitoSource` objects with the subbject and sink information with the appropriate namespace.
        displayName: Kogito Knative Eventing Source
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless workflow
        - kogito
        - knative
        - dmn
        - bpmn
        - eventing
        - cloudevents
        - messaging
        links:
        - name: Source Repository
          url: https://github.com/knative-sandbox/eventing-kogito
        - name: Kogito Knative Add-On Documentation
          url: https://docs.jboss.org/kogito/release/latest/html_single/#_kogito_knative_eventing_add_on
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - gcr.io/knative-releases/knative.dev/eventing-kogito/cmd/controller:v1.2.0
        version: 1.2.0
      entries:
      - name: eventing-kogito.v1.2.0
        version: 1.2.0
      - name: eventing-kogito.v1.1.0
        version: 1.1.0
      name: alpha
    defaultChannel: alpha
    packageName: eventing-kogito
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KEDA Community
      provider-url: ""
    name: keda
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keda.v2.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "KedaController",
                "metadata": {
                  "name": "keda",
                  "namespace": "keda"
                },
                "spec": {
                  "watchNamespace": "",
                  "operator": {
                    "logLevel": "info",
                    "logEncoder": "console"
                  },
                  "metricsServer": {
                    "logLevel": "0"
                  },
                  "serviceAccount": {}
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledJob",
                "metadata": {
                  "name": "scaledjob-sample"
                },
                "spec": {
                  "failedJobsHistoryLimit": 5,
                  "jobTargetRef": {
                    "activeDeadlineSeconds": 600,
                    "backoffLimit": 6,
                    "completions": 1,
                    "parallelism": 1,
                    "template": null
                  },
                  "maxReplicaCount": 100,
                  "pollingInterval": 30,
                  "successfulJobsHistoryLimit": 5,
                  "triggers": [
                    {
                      "metadata": {
                        "property": "examle-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledObject",
                "metadata": {
                  "name": "example-scaledobject"
                },
                "spec": {
                  "cooldownPeriod": 300,
                  "maxReplicaCount": 100,
                  "minReplicaCount": 0,
                  "pollingInterval": 30,
                  "scaleTargetRef": {
                    "name": "example-deployment"
                  },
                  "triggers": [
                    {
                      "metadata": {
                        "property": "examle-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "TriggerAuthentication",
                "metadata": {
                  "name": "example-triggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "exmaple-secret-parameter"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ClusterTriggerAuthentication",
                "metadata": {
                  "name": "example-clustertriggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "exmaple-secret-parameter"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Cloud Provider
          certified: "false"
          containerImage: ghcr.io/kedacore/keda-olm-operator:2.7.1
          createdAt: "2022-05-11T00:00:00.000Z"
          description: '[DEPRECATED] Use Custom Metrics Autoscaler Operator instead'
          operatorframework.io/suggested-namespace: keda
          operators.operatorframework.io/builder: operator-sdk-v1.1.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/kedacore/keda-olm-operator
          support: KEDA Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              Represents an installation of a particular version of KEDA Controller.
            displayName: KedaController
            kind: KedaController
            name: kedacontrollers.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your jobs and what the triggers
            displayName: ScaledJob
            kind: ScaledJob
            name: scaledjobs.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your application and what the
              triggers are
            displayName: ScaledObject
            kind: ScaledObject
            name: scaledobjects.keda.sh
            version: v1alpha1
          - description: Describes authentication parameters
            displayName: TriggerAuthentication
            kind: TriggerAuthentication
            name: triggerauthentications.keda.sh
            version: v1alpha1
          - description: Describes cluster wide authentication parameters
            displayName: ClusterTriggerAuthentication
            kind: ClusterTriggerAuthentication
            name: clustertriggerauthentications.keda.sh
            version: v1alpha1
        description: "## Deprecated\n This KEDA community operator is deprecated,
          please use `Custom Metrics Autoscaler` Operator instead.\n## About the managed
          application\nKEDA is a Kubernetes-based event driven autoscaler.  KEDA can
          monitor event sources like Kafka, RabbitMQ, or cloud event sources and feed
          the metrics from those sources into the Kubernetes horizontal pod autoscaler.
          \ With KEDA, you can have event driven and serverless scale of deployments
          within any Kubernetes cluster.\n## About this Operator\nThe KEDA Operator
          deploys and manages installation of KEDA Controller in the cluster. Install
          this operator and follow installation instructions on how to install KEDA
          in you cluster.\n\n## Prerequisites for enabling this Operator\nPlease note
          that you can not run both KEDA 1.x and 2.x on the same Kubernetes cluster.
          You need to uninstall KEDA 1.x first, in order to install and use KEDA 2.x.\n\nDon't
          forget to uninstall KEDA v1 CRDs as well, to ensure that, please run:\n\n```
          oc delete crd scaledobjects.keda.k8s.io ```\n\n``` oc delete crd triggerauthentications.keda.k8s.io
          \ ```\n## How to install KEDA in the cluster\nThe installation of KEDA is
          triggered by the creation of `KedaController` resource, please refer to
          [KedaController Example](https://github.com/kedacore/keda-olm-operator#kedacontroller-spec)
          for more deatils on available options.\n\nOnly resource named `keda` in
          namespace `keda` will trigger the installation, reconfiguration or removal
          of the KEDA Controller resource.\n\nThere could be only one KEDA Controller
          in the cluster. \n"
        displayName: '[DEPRECATED] KEDA'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keda
        - scaling
        - event
        links:
        - name: Homepage
          url: https://keda.sh
        - name: Source Code
          url: https://github.com/kedacore/keda
        - name: Slack Channel
          url: https://kubernetes.slack.com/messages/CKZJ36A5D
        maintainers:
        - email: cncf-keda-maintainers@lists.cncf.io
          name: KEDA Community
        maturity: deprecated
        minKubeVersion: 1.17.0
        provider:
          name: KEDA Community
        relatedImages:
        - ghcr.io/kedacore/keda-olm-operator:2.7.1
        version: 2.7.1
      entries:
      - name: keda.v2.7.1
        version: 2.7.1
      - name: keda.v2.6.1
        version: 2.6.1
      - name: keda.v2.6.0
        version: 2.6.0
      - name: keda.v2.5.0
        version: 2.5.0
      - name: keda.v2.4.0
        version: 2.4.0
      - name: keda.v2.3.0
        version: 2.3.0
      - name: keda.v2.2.0
        version: 2.2.0
      - name: keda.v2.1.0
        version: 2.1.0
      - name: keda.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: keda
    provider:
      name: KEDA Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: stolostron
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stolostron.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:3443da607b276d39ba28a8d6561d296f30e9bac5ab06d5c70715bddbde0a58b2
          createdAt: "2024-06-28T04:52:57Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/multiclusterhub-operator@sha256:3443da607b276d39ba28a8d6561d296f30e9bac5ab06d5c70715bddbde0a58b2
        - quay.io/stolostron/governance-policy-addon-controller@sha256:593dc0a689091ce7c916e6be764c222c1020629ccff7eafbf17524936e4f8262
        - quay.io/stolostron/observability-e2e-test@sha256:8d9341b1814cd085a20f9feceab181a8989d6684fd44d6c8e54a0c6df4dd7216
        - quay.io/stolostron/governance-policy-framework-addon@sha256:a1c0d40561befaf5e849060beaee4b201f42d77c5090823f23ef43a75106b00b
        - quay.io/stolostron/thanos-receive-controller@sha256:b1a191714cead24f66e8de7217db3c4d4a253ecf0238a912c8b88ba8b47c972d
        - quay.io/stolostron/console@sha256:c7900403ef8dba3aac52f999532d876bbfe2f66f00a255047850594b64021830
        - quay.io/stolostron/volsync-mover-rclone@sha256:b247b23944e20104f801c172cbbb620740a5b51d1777ecdbf98b4c128f1ef171
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/prometheus@sha256:8250af047c03c6b0c07b76420e0a5c2338f053d73b2dca1ecba4bb2f211715fe
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/search-collector@sha256:136ad44ff41a29dec48e87d10af5b608f335e4242c41ac17cba2c3c4c8205bd6
        - quay.io/stolostron/multicluster-operators-channel@sha256:d791aa2c6032c1de50cf4033c426646613b182e98c406a091b88fb9ffefc0c98
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/helloworld@sha256:81a0b06c7fcf915b1bb40d33d29d0bfd4d501404463c1f46bffc503e7d1cc45b
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:02e53b8a6529e3209809dada4113054e3398de18738948cb01ed1d24dcb8e8be
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/insights-metrics@sha256:49a177357e865d5f8347840fe695531d539d17e28d2b9971821e41ebbee014bd
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/rbac-query-proxy@sha256:24aafcba20ded0892209cfcc196bff88f877ed51eb928d5d350b17041d170893
        - quay.io/stolostron/volsync-mover-restic@sha256:284c3358f0f2158d599f27487d76a7fdb21842fc0e95bc639726dfdad14b7e7d
        - quay.io/stolostron/prometheus-config-reloader@sha256:bc93ca9f1b6d95e8b963fe1af4c17f87c8c7d938a348627c29cc3cf42b70ca3e
        - quay.io/stolostron/prometheus-alertmanager@sha256:8222bbdd37c246703de54ed3dc109fa8c2f6ca17c5bb90106948683858afdc65
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/multicloud-integrations@sha256:60be5b945ead8f845a29a8be18bd26922b83f009fde1412017f1ca985eb79b55
        - quay.io/stolostron/grafana@sha256:e40d66da749108ae1eecf909c95c87bf7a289328e8b373ae4402368c474f2da5
        - quay.io/stolostron/observatorium@sha256:af4e315e4c6da3607f84e24f31e0f37f8572d161736b4748b2e57e0a2c2dd36c
        - quay.io/stolostron/volsync-addon-controller@sha256:7774767ed20263bba4857a0d4be012e12215f640fce01fbd3a085bf860d087e7
        - quay.io/stolostron/volsync-mover-syncthing@sha256:b5fad266ca8d3cd8efcbf4e8241c446004747b551b2b8e06acae10900d7a0f1b
        - quay.io/stolostron/thanos@sha256:cc78cd786643d27ed85eb3670852722e915a20b20a21479e1477fb859dec6c5e
        - quay.io/stolostron/prometheus-operator@sha256:490b0056b9a23a1dc514a3036eac6593c7c8961de27d08ff24817d280325138c
        - quay.io/stolostron/cluster-backup-controller@sha256:9c77444bd42e6e99163f4dce9f9975276aea082100f53fd938cddf74ad3b1574
        - quay.io/stolostron/multicluster-operators-application@sha256:199b696dbd7589b77e2391b71504fbcb7d3ae785972dac3d001c4c1bffa342dd
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/kube-rbac-proxy@sha256:9626cb9b2ee8f00d2abdb14cd2ce56d77435859071c10d8394bc7ee2754ce87f
        - quay.io/stolostron/volsync-mover-rsync@sha256:f7affcc16ae9bf6d12f69511b7a269f5fd065d1d87392c4c38cb5d7af5b89e23
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/search-indexer@sha256:068c38489201316dc12d475de03c8abb11575338bd75e25c1262eca8aac49cb2
        - quay.io/stolostron/search-v2-api@sha256:6b97b2c7e64de7cd154fd83da3bc690a57b67f01c83170238fe9c2e764df91a8
        - quay.io/stolostron/cert-policy-controller@sha256:27673358610dc1ca308de472dbd794dcbb38250bb299b53c736549b1e8441a0f
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/metrics-collector@sha256:4101a2f2c4230bdc1420a58b546a241dfc5414cf92b57f3b73619c1be14dbc32
        - quay.io/stolostron/observatorium-operator@sha256:5da7f478ecf3609a3b86eb563e4dd22588f3d13efdc2e33f5b672fe95583ce68
        - quay.io/stolostron/grafana-dashboard-loader@sha256:cff97fe3e32dd14ed2f3b7b02cb205715cfee6cfd6dc7e7a630101ea5147e566
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:9d1d5e2b5313b898188d2c845dcda5d604b73f16e89c71915ea308281c1f56ff
        - quay.io/stolostron/klusterlet-addon-controller@sha256:01e5fca5e5c6caca6878bf8025471b7c77164ca7b4a2d764e20101fc79290df0
        - quay.io/stolostron/search-e2e@sha256:8dfbc899ba1ffaa8fb6eb2861f6db87a13d2565cf4551fd47df7fe8253217fb2
        - quay.io/stolostron/config-policy-controller@sha256:197d8491e7c889db8f2acb726e0146d384c979acc7f634c24a9d4b21f7d444f3
        - quay.io/stolostron/must-gather@sha256:1f7c129859bdaf6bd93dda77aec01f7cd03000fbfaccd809f26d33b9e3f66e51
        - quay.io/stolostron/search-v2-operator@sha256:c41517f372fc1874895b39e70cd1c07391110697a5e49ebcf56f4a5df499d810
        - quay.io/stolostron/kube-state-metrics@sha256:ef1d671ffa098ed95186de6f15ea91fe648d103c1c7a57871ed46af07c6f7881
        - quay.io/stolostron/submariner-addon@sha256:0f1b4ae59ff5c933542f9affed20bfa6694deb804f2e97f49ebd0f2e1b925eca
        - quay.io/stolostron/insights-client@sha256:aac5b9b66800c0b862e8bfa1fea533a4aec854d8c52baddbfa03977e1ed520cb
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:bf6bec399bc2072186e9d12ed865167fe940f3f619d69ffe6c178c14a3bdadbe
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/multicluster-observability-operator@sha256:245e1815a73b5c48f790ae5a5d85fac6642cce8a5a3bf122275809e97a4d519e
        - quay.io/stolostron/grc-policy-framework-tests@sha256:f2b01192fd900860428942c7d5b900604ffcab0d42731676a1acdeebd3716557
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/cluster-permission@sha256:716d2cbcb7f69b5640d04669b15884e0adf053597f278a0bb1d8ee77e4a7af4d
        - quay.io/stolostron/governance-policy-propagator@sha256:200558b5aa2366383bd298e159d9f7966a2c9b981038c79cde536edbfbe0092c
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/volsync@sha256:ba7399b316359118b1c1b915d667fca9a4a0d12a10beae69e6dee9c7236753be
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:7f6b6e45f0be0a14d7d9d79828572b9400f4d559f084028e3254cfd40948780d
        - quay.io/stolostron/multicluster-operators-subscription@sha256:8350df0a7d4f19c8e52eb58eecf5223bb2a0ccbe4d4aa00fcf8dff832cb43e5d
        version: 0.4.0
      entries:
      - name: stolostron.v0.4.0
        version: 0.4.0
      name: community-0.4
    - currentCSV: stolostron.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:781f6712369cd4b9bdc8ba0ee2f35a12db198ca190621e199dc10cfea3f5efc9
          createdAt: "2024-10-02T08:01:20Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.5.0 <0.5.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/governance-policy-framework-addon@sha256:5f2b326a279b7c42c233ea5f11b89ecaf4cfd7c549d021ddcc4e7b0c8926fc6d
        - quay.io/stolostron/rbac-query-proxy@sha256:2f1c78176d0ca0b21c237a9f7a3f77f7d857dd848f3a42254cee1e4e469b4c3c
        - quay.io/stolostron/multiclusterhub-operator@sha256:781f6712369cd4b9bdc8ba0ee2f35a12db198ca190621e199dc10cfea3f5efc9
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:d42ddd30570ccdf388ccd65cd3f7c72de2d857d328e05aef8e855508ccf91898
        - quay.io/stolostron/observatorium@sha256:8d0f02996b5bf02be02ff3743821909b6c2396f0f0519652dde0c5d1d76f8a05
        - quay.io/stolostron/prometheus@sha256:147f3a0cee8fa7b11c82e21ca95c28278e1ab53ccb5b70f3af0425bb8a74f61b
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/volsync-mover-syncthing@sha256:b5fad266ca8d3cd8efcbf4e8241c446004747b551b2b8e06acae10900d7a0f1b
        - quay.io/stolostron/search-collector@sha256:4082f7a37aa3f10c725bb9bebf94ee8dab97d22302cd5e59b35d0e3bcc888cef
        - quay.io/stolostron/volsync-mover-rclone@sha256:b247b23944e20104f801c172cbbb620740a5b51d1777ecdbf98b4c128f1ef171
        - quay.io/stolostron/insights-metrics@sha256:49a177357e865d5f8347840fe695531d539d17e28d2b9971821e41ebbee014bd
        - quay.io/stolostron/acm-cli@sha256:85a5c55dfa4708435807d1266b5607525e33f55c5b4672fd8ba78987a1c48ef3
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/thanos-receive-controller@sha256:58c97fe5ec2131f9dcfb70563b59ed0bcae94d8bddefa51cb4784e73b1d90e4f
        - quay.io/stolostron/cluster-backup-controller@sha256:688844c722db78f383d90986150bde93bb421068e7f5b4128d19043f3e01f93e
        - quay.io/stolostron/config-policy-controller@sha256:5af6a0e17da7b7261ac71bcb1e2d991856bbb5450686d188026798e05f0617ba
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/volsync-addon-controller@sha256:d0b15da92e0ff9792d750b7b48d72b38d6abb91488c20578c7be3382c561877f
        - quay.io/stolostron/observability-e2e-test@sha256:c361f5a0f806498c6c1150c0c0d5a42222efff13a18cb28dbfd5e4bd835cb4b7
        - quay.io/stolostron/siteconfig-operator@sha256:7f47f0c61d5b53411ed78046c5a02664c6693d940ef37d6db1853670196be334
        - quay.io/stolostron/insights-client@sha256:aac5b9b66800c0b862e8bfa1fea533a4aec854d8c52baddbfa03977e1ed520cb
        - quay.io/stolostron/kube-rbac-proxy@sha256:1b708bff7fb46391f2460d34be28d30add38501cdb084306df01faae75637232
        - quay.io/stolostron/multicluster-observability-operator@sha256:fb0519952ef6436cc66496409f00e89cd6e58a41776df96859a86176a7322d05
        - quay.io/stolostron/multicluster-operators-subscription@sha256:30b1b5bf6a7a12e525450905a474babdd7e38235d9be60f3cb04d60a745a1b96
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:a8d812a527902bd4d9dfd23f90ce90f415f5a6fabb007d1c41b4c998e1ca098a
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/observatorium-operator@sha256:38f22e30db7713e9f2e239463803cdb5cdfe7108a524ffa5d7e95201bae5300b
        - quay.io/stolostron/cluster-permission@sha256:716d2cbcb7f69b5640d04669b15884e0adf053597f278a0bb1d8ee77e4a7af4d
        - quay.io/stolostron/thanos@sha256:dd4d4f8fbb9a8f51134260fc728d19f61a93b511fd7889636089944ac46163bd
        - quay.io/stolostron/search-v2-operator@sha256:6df11a36ea19a1b3da191964f851dee95b87fd993e8f0ef8906b1e63dd450d17
        - quay.io/stolostron/metrics-collector@sha256:c580106b2c7173251afa6714b710fd109811023d88a8009fb741f3d00e486e76
        - quay.io/stolostron/volsync-mover-rsync@sha256:f7affcc16ae9bf6d12f69511b7a269f5fd065d1d87392c4c38cb5d7af5b89e23
        - quay.io/stolostron/search-v2-api@sha256:338634d767383716c7a88ed2bc9c58275e9d5953932293443725883d7efcffd5
        - quay.io/stolostron/multicluster-operators-application@sha256:329c82b92fe2448fe26923fbd82d2eadacb7f96c75689d96393c573f95e6d650
        - quay.io/stolostron/must-gather@sha256:362bbfd54059975369ce6d3869d34e346a0e76159163714cfcb3597022b7c88f
        - quay.io/stolostron/prometheus-config-reloader@sha256:337f7d2036c17459a41884e38c569a3601364b02660946b49d39ce40ef97bd48
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/governance-policy-addon-controller@sha256:9f387eb8b54d1c86fc20d205dbce199738849cb0ce15a47052ddbee8781d2d9b
        - quay.io/stolostron/volsync-mover-restic@sha256:284c3358f0f2158d599f27487d76a7fdb21842fc0e95bc639726dfdad14b7e7d
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/search-e2e@sha256:ddaa50ecfafcd7501903420a47495f733dce36fcb91c8ce9b8d94f4448ed501d
        - quay.io/stolostron/submariner-addon@sha256:07287987fb0041efc16557cb9d7403dc15b5959a5d0dd0d81cd72bc2658cba10
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/klusterlet-addon-controller@sha256:cf4cad3599c4ce538dc007cda9f5b06538799e8198912d907058ecdd4bbb2451
        - quay.io/stolostron/multicloud-integrations@sha256:d4cb3a0a50bc5d22f48c96ed3a10ea9843c6ef2f83daeae411a7ddb7ea332ea6
        - quay.io/stolostron/prometheus-operator@sha256:5ff1f7a4be0216f582f5cb40de3412cbe8126433144c8201750f6c3c0fee920e
        - quay.io/stolostron/governance-policy-propagator@sha256:69ab8f492fa418596db6da578b3b579fc2f1506c5bf780e6aba3ccd05d8b02b9
        - quay.io/stolostron/console@sha256:e42d2f7e39a93f13003193142a3d1f3f9e55966cbf55b94eac93c726373a2516
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/volsync@sha256:ba7399b316359118b1c1b915d667fca9a4a0d12a10beae69e6dee9c7236753be
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:dbdbec1600d57c440173aecf19c57709d69e945dcecb348bcf4a7c37b3700b70
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/search-indexer@sha256:7471f2808686d42c2eae9f0a537f9b520648f31e87072a376243f309fd044caa
        - quay.io/stolostron/grafana@sha256:c9869c7f590bc4c54151c7e4fa9ed76dfe35c24a0ea3c19a6a1b26e61924b6e1
        - quay.io/stolostron/cert-policy-controller@sha256:809cae1aef02dd3fb68143ebdba818b20c661bb628a7922ee1d705efe4947f35
        - quay.io/stolostron/helloworld@sha256:48724b53542a3fcaf342d6ae2056bdf846cbe2ee50879613953b045bfcbe1c0a
        - quay.io/stolostron/kube-state-metrics@sha256:dc534373feda042afa385026c7da674912ca310df2a4598d83451d4474bb3f44
        - quay.io/stolostron/grc-policy-framework-tests@sha256:57794e2b15949cedf3462df2d38b95e86ca1e140a5ae81e4a9b3e7e06072847d
        - quay.io/stolostron/prometheus-alertmanager@sha256:19f2e277cbdd2bd4814b8d8034167c1a7cfe9f1d698fb16f7604fa65cb8bcb1c
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/multicluster-operators-channel@sha256:260a11aae954299ce8fbb0f9c3102afd3b5cf3f1b972b5a71e56c273ba1d532f
        - quay.io/stolostron/grafana-dashboard-loader@sha256:85031d1a56514be239c585487aec05a8cbb08c17aeb60c0c84c685eb9d9f08b5
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:358ac0b0a3afdfac06f784f2e02a305cbef4e6f648e31a00981d3b097746412a
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        version: 0.5.3
      entries:
      - name: stolostron.v0.5.3
        version: 0.5.3
      - name: stolostron.v0.5.2
        version: 0.5.2
      - name: stolostron.v0.5.1
        version: 0.5.1
      - name: stolostron.v0.5.0
        version: 0.5.0
      name: community-0.5
    - currentCSV: stolostron.v0.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:09f1f4266938fba88376a50c3e2837c5517b4f3a947cf86638a17e52eb17c1f6
          createdAt: "2025-02-24T01:42:20Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.6.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/flightctl-worker@sha256:4dddded7d2d4cce63972c8ca81f213c2cedb551cc645ec85495a9692dc36bb4c
        - quay.io/stolostron/helloworld@sha256:56f0d641d4f07b922f1c28c536eacad747c56a564e0a65f084106c2ffc9bd99d
        - quay.io/stolostron/acm-cli@sha256:8c8c059ce2f6d131c410972bf8fd0b7e399cf17cd044c7ddc89a9668c0ceac4c
        - quay.io/stolostron/search-v2-api@sha256:694523242aabb8db1f27b5afcd8ab2f63a9cdb6056c33a79badf8e68a337738d
        - quay.io/stolostron/multicluster-observability-operator@sha256:bed0dd30646d7fce894a625e6cf52e7829c272f47dde0baf1f788e9f85f66e42
        - quay.io/stolostron/cert-policy-controller@sha256:98dd74d7375ca2b546075d5c8f06a9e2c4f2ce3b77247dccf36bcaa46bf564c7
        - quay.io/stolostron/governance-policy-propagator@sha256:c21574a2421fefe3fcabeef14640d6fe699d950106c9ed056b3ba9be3cbeabe0
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/flightctl-periodic@sha256:a3ed745dc65a5886c4cc05348cf4822bfbf2c6d28b85a39fa0b8f3e1f53fdf8a
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/search-indexer@sha256:4f9be902fe976fd88a77b737cc1f56f00d48740a6a62661468328cf3bb9da31a
        - quay.io/stolostron/search-v2-operator@sha256:85e53a97240f6fd1c4251101146cc2bcc51eb91d9b171efb5b58edade9e0244e
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/multicluster-operators-application@sha256:59802fcfae5b6fd0bebe58e98412885e71f32464b8dbbf85364ddaf090aa3e00
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:9c85fb91c4be3287cac2a376465836689df885902099096f0f76ffca50a6d311
        - quay.io/stolostron/search-e2e@sha256:43e1bbfb7259612ffb408108c71b825f6e61a268c7e405b8d855f6878b6f96dd
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/flightctl-api@sha256:64fb1b655bef6aeff03cf64b6450155d3cb570ed4713f084590fe515f4fbad02
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:51a8acc47098175db148ebc19a10cdba2803d4a9f5884dd38b1c839702398789
        - quay.io/stolostron/thanos@sha256:53eee303836070a7c7621bfb2a094e3809146497d1653dc65d16055aacea332c
        - quay.io/stolostron/siteconfig-operator@sha256:97dc128666780ec9a4c955f392ab98f81f0d6038ec2159e8b1d55aab393cf118
        - quay.io/stolostron/config-policy-controller@sha256:8cf54eb6f5b49a60ed165fa4b70e4984a1362f2f31e8304c3a22e8a6a953a606
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/governance-policy-addon-controller@sha256:fa0738abc0368f9dadd8b0845d868512d26be3525a1a07b9de8a9cefafaa42b5
        - quay.io/stolostron/klusterlet-addon-controller@sha256:39e6940eb956fe9640342e11ecbcbc7d6df0169e1b340a60afc19aff7dd713a5
        - quay.io/stolostron/thanos-receive-controller@sha256:d73f98e93a258adffc1a4bd875c79a9629166599e68ef777a790a1f3146a2929
        - quay.io/stolostron/cluster-permission@sha256:443d8b30e3bb83c139c64ddfc38b0d5e6d7a25a4fae783303ab2f2c27d015b26
        - quay.io/stolostron/grafana@sha256:51092f6754106a90aff28cdd80c2f0447c85c1bbc72db15878d0b65b93e1ead6
        - quay.io/stolostron/volsync-addon-controller@sha256:d9dad837f2f735674e5711c7fff1cfe2c0c92160fceb7d1548a62348dd43b81c
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:92369b6deb21ab3cb0a12ae5c74b455cfefb8698826bca39e26648e608e48a2b
        - quay.io/stolostron/observatorium@sha256:398ecee4d3f12e84e48b12009a4b02fc8b360dfc0c1607ffecc26e6bc04d0ea2
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/prometheus-config-reloader@sha256:2b26d2fb5c69e54aecfbcaab46e652d136d726f0604844031c8771874f1fd769
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/grc-policy-framework-tests@sha256:9b84dbafef35bd568b2bf79283b6850ba1ff57d70239a27cc0cc815fb1a55515
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/metrics-collector@sha256:c7433ee4ccdb56717802acf258530e2d6d237787d760a8fbcfd642790254d5ab
        - quay.io/stolostron/cluster-backup-controller@sha256:c85f88d42b42e4366f8ff7ad6c3e6b375e98c5b6a12a9c24a6333929e48f1811
        - quay.io/stolostron/observatorium-operator@sha256:5eb4c0da92c62b664aaf3bc187924870275e687440c3b0a3ad4f1c760c8b35e9
        - quay.io/stolostron/kube-rbac-proxy@sha256:5b521e4f0e5e54c53cea714020069aa796da23d1482d14e4b7c908098ac69ae9
        - quay.io/stolostron/observability-e2e-test@sha256:03403aeb59b364f578761ae4bcb302bd3589baff500d733775e8cb4f2f961236
        - quay.io/stolostron/insights-metrics@sha256:f2f4cf14090252af32832ff6df8a9f896e5314751199156c4e5ae0755e4515e6
        - quay.io/stolostron/search-collector@sha256:54ece9741641e1bb6e30f15d97a65c92b9c6408ba2572a2f6658f537134adff1
        - quay.io/stolostron/multicluster-operators-subscription@sha256:a056ce7d438f68d33ca3dfd42ec3d9a34b347cdf05499066412784109726e7ad
        - quay.io/stolostron/submariner-addon@sha256:9c72b5177de8ddc56dd3544997e9bd222895545b7c578e43e4051596fb1fa873
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/prometheus-alertmanager@sha256:b98480f8ba6ffd19fcee3b412757ad34428e6feab7e4196bb13f2f940306e2c9
        - quay.io/stolostron/insights-client@sha256:978d8b7c3e692272024b71b8c005b752c08dc643dde8258ad75961ddcdcce67f
        - quay.io/stolostron/prometheus@sha256:3d1810567f67ae1a7bb024a1c4e494ee510cfb14c7ed21a8b5e63e41fa7bcd68
        - quay.io/stolostron/flightctl-ocp-ui@sha256:6cff6425850d2b4a8b885a684afb01244d7cd9de7035d7ad71b1e10e5bab25e9
        - quay.io/stolostron/kube-state-metrics@sha256:cd24916e55ad905e386fa3df8a648272114353a0d75283224a4dddb1955b2687
        - quay.io/stolostron/flightctl-ui@sha256:fea717f3b8b4c343ec60ce7d74a691f1124b6847d55e30a73dc0d50ea610d370
        - quay.io/stolostron/multicluster-operators-channel@sha256:501a5053bd3fd5cc9de34e7e020890d731321b090a0e5bf11b8b3350d002f573
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/stolostron/multiclusterhub-operator@sha256:09f1f4266938fba88376a50c3e2837c5517b4f3a947cf86638a17e52eb17c1f6
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/prometheus-operator@sha256:f80e79ea87cde66955e1e46d264e414c072085da9db6876ee2ae68c87d2452e5
        - quay.io/stolostron/governance-policy-framework-addon@sha256:047124271d0b30176a94454b293c82c2eb67a6f06757fc22c54c455fd0913157
        - quay.io/stolostron/console@sha256:eed52f3ab860a3a224ce178d92765ca50d72d6261f04df040a39577881738252
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/multicloud-integrations@sha256:0081065f9e59c5ed41087344668ac86012f796cdbbf3575fd8491c7dfd24fbf8
        - quay.io/stolostron/must-gather@sha256:2bf9d3c15f32a6bfe0a7a8d39b9dc62ac6cbb6fe47b7cf1f6bc5e30b17e5e849
        - quay.io/stolostron/grafana-dashboard-loader@sha256:6c6c2aa528373536d5db66fffeb87a85b79f3cf723e5d9eeaefcb74f91c1a8a1
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:05da11b343938866c5afcfe844bcd15d9c513fdb10fd7ea0d3473b455b427c2a
        - quay.io/stolostron/rbac-query-proxy@sha256:096ad23c9b9c7e90451c09b48ded67a82575e75c8cc19f22c459d63ae8b5a786
        version: 0.6.3
      entries:
      - name: stolostron.v0.6.3
        version: 0.6.3
      - name: stolostron.v0.6.2
        version: 0.6.2
      - name: stolostron.v0.6.1
        version: 0.6.1
      - name: stolostron.v0.6.0
        version: 0.6.0
      name: community-0.6
    - currentCSV: stolostron.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:ab6660283a25bdf2ea9afe8c76bfbc32f4610d69af2bb13a7bca37a1f5961bdd
          createdAt: "2025-06-09T15:10:05Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.7.1'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MulticlusterHub defines the configuration
              for an instance of a multicluster hub, a central point for managing multiple
              Kubernetes-based clusters. The deployment of multicluster hub components
              is determined based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/rbac-query-proxy@sha256:1c198b6388a72ccfe2c5b68d57a2b22893fb65d977992ec44ee8cad6ee1ca976
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/stolostron/governance-policy-addon-controller@sha256:6ecc7557168177474423cea60eadc07d8382adb3e97dee669a8590859a41b08c
        - quay.io/stolostron/flightctl-api@sha256:64fb1b655bef6aeff03cf64b6450155d3cb570ed4713f084590fe515f4fbad02
        - quay.io/stolostron/cert-policy-controller@sha256:6deb41867f8781f4e34f69a54d4aeca190d588ba2991279d39cb5a82c2a5c5be
        - quay.io/stolostron/flightctl-ui@sha256:fea717f3b8b4c343ec60ce7d74a691f1124b6847d55e30a73dc0d50ea610d370
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/cluster-backup-controller@sha256:e576c2de1f931b7d33da2dbc03d71ce323766bd9158bf4b128221ddd1ab3fb9f
        - quay.io/stolostron/observatorium@sha256:18a06117ef114a1e68474fd9832adacd30b03f60ec1aa55374285a20e5311f13
        - quay.io/stolostron/insights-metrics@sha256:861dc03fff55c1d9aff96b5ed15b9d38d5932df1c00a4423e35ead19da6eda03
        - quay.io/stolostron/siteconfig-operator@sha256:16b00f6c4d81eea26e663ca5596526d85bf7777728d1f0ca7e4dd765c64d2bb5
        - quay.io/stolostron/search-indexer@sha256:58320b54847e3bb6baa558bda65e9d0ae2198df480ba27f46f96ba1454614508
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:4b29cc08c9e35054a4a820472ea19cd51e043ab1a90495418a36922df74b9540
        - quay.io/stolostron/multiclusterhub-operator@sha256:ab6660283a25bdf2ea9afe8c76bfbc32f4610d69af2bb13a7bca37a1f5961bdd
        - quay.io/stolostron/prometheus-alertmanager@sha256:a9da6f01cfa61fef803f09ef3f6b27e33a07b97fc1f1e11b6a6c8e13b8a4b90e
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/volsync-addon-controller@sha256:224a0a03a25e5ce61f4c50f6252370876ac0cbf4f9711e997e98a00d6a124fe9
        - quay.io/stolostron/origin-cli@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/search-v2-operator@sha256:2ae9ab48615966ccf91b2022ccda12913a3db3cfbad5351dd8a20648b01137d8
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/console@sha256:ba89196efe30c812a767d58a41170b56e2b04167a72e7f13180bf40474bdc824
        - quay.io/stolostron/prometheus-config-reloader@sha256:024ee6dda10c6d22ddeab57d7d6503b7aa9ac78981cfcec8ffae94b4914500e3
        - quay.io/stolostron/submariner-addon@sha256:51359ccdc0ea29bcaf7ec2b190878e9f3333b1530cb1128e142d893fb87eba8b
        - quay.io/stolostron/prometheus-operator@sha256:8c1c12562ec00da8eaedb7d428694ebdd3a749c0b1ce03fe2b5663ae40e1c701
        - quay.io/stolostron/flightctl-periodic@sha256:a3ed745dc65a5886c4cc05348cf4822bfbf2c6d28b85a39fa0b8f3e1f53fdf8a
        - quay.io/stolostron/governance-policy-framework-addon@sha256:b1b24c40d68415525e2bae21810fc439d3e96fcc5fa75c92e7f2eee284422bfd
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:05da11b343938866c5afcfe844bcd15d9c513fdb10fd7ea0d3473b455b427c2a
        - quay.io/stolostron/helloworld@sha256:4465550664bd210cf0df8edfd706de2fc51cacf01b66a962b5607935f9666ebc
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:54c6a2eec497eafdbf861fb95d78d961fb24906b197fc69b4e96d3e8b53e6496
        - quay.io/stolostron/multicluster-observability-operator@sha256:3d95297297a301fca1915268fec71d7a6ecef98c6c2ab8494d9befacb2e169fa
        - quay.io/stolostron/thanos@sha256:9d504e2aa1001135eb6e454c58c620d102f2b0a85016b7587452dd1595794c93
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/must-gather@sha256:ad8db72d559d06f4d467f78213d4f8f2cb00f8e76e942d8e75394eb7dadd29d2
        - quay.io/stolostron/acm-cli@sha256:2f0cec3bed4ad77f573e8cd51a7e2c62d68a59e4e261ea801d5ac203fd2724fc
        - quay.io/stolostron/grafana@sha256:cebaed7e5abe0064c65ea69bd4e14794d153de9172c4c3b6de4730f23fed7d24
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/metrics-collector@sha256:c7433ee4ccdb56717802acf258530e2d6d237787d760a8fbcfd642790254d5ab
        - quay.io/stolostron/observability-e2e-test@sha256:4e60d3e3e1b5dbab3bcc1c4d37a4940a2a42a065a9df539fb243fb174dde6716
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/search-v2-api@sha256:786866ca8c71a8263ddef45ca9bd9e48cfd1b3ad6215725fbf6f79967bda20ad
        - quay.io/stolostron/multicluster-operators-channel@sha256:318d885ab33d5afe17334e831c955f1c3cdf804281b7cc8523330820ac22eee2
        - quay.io/stolostron/multicloud-integrations@sha256:665e9385531674a5849aa661eb207a0c6e39ea563fc891c57678cddbf7878d8e
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/observatorium-operator@sha256:2f11444703addec163a00c99529c228594fb52df13de47f08faac2e446bec483
        - quay.io/stolostron/governance-policy-propagator@sha256:2b77422fb03177f647c04716e7242bc08f241a1ef9c8968e492ae57816f4c4fb
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/grafana-dashboard-loader@sha256:6c6c2aa528373536d5db66fffeb87a85b79f3cf723e5d9eeaefcb74f91c1a8a1
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/cluster-permission@sha256:e45139d459951b2e751018fe7ecabe1ff22b90ac87be81fb8f17a12f7b7f6a65
        - quay.io/stolostron/search-collector@sha256:2b3dfd21c0ba7b7b84fac390002e47fae927cd5fdf13e5d1d960703aa6c5bede
        - quay.io/stolostron/insights-client@sha256:2cbd20252f10a5a64f93af50f117f05a865aa9d578e5f6e55e4249470b761604
        - quay.io/stolostron/kube-rbac-proxy@sha256:e5555e256a518ca34bea01c959bbb5497582b53dbfaaf326e1852430c175637a
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:1efd8e470b88a992e2f35ce18bdf4970f91af7ba917b8d92627b47ce8c10eb40
        - quay.io/stolostron/flightctl-worker@sha256:4dddded7d2d4cce63972c8ca81f213c2cedb551cc645ec85495a9692dc36bb4c
        - quay.io/stolostron/config-policy-controller@sha256:b1718d5facb95c5572ac658c899f9fce0a7e8c08de93804de3327cd50cdb073b
        - quay.io/stolostron/redis-7-c9s@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/search-e2e@sha256:1c7dcc9f3588620b3570a493c2d744de059f425b572b9cb5725142b89414a904
        - quay.io/stolostron/flightctl-ocp-ui@sha256:6cff6425850d2b4a8b885a684afb01244d7cd9de7035d7ad71b1e10e5bab25e9
        - quay.io/stolostron/thanos-receive-controller@sha256:f445bbfc18bbc4646e0432842022d3be5305809e52ddcb9daf0ec000844d5854
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/klusterlet-addon-controller@sha256:4cb2f4f4e14cd67a4b2f9549cba714b635d9e1c2d8b869b0e0573fd7a41bc13e
        - quay.io/stolostron/multicluster-operators-application@sha256:9f804b4f2da910e0d7f122176f72c0b0ef4b30785fde0369a66f95f4bac58a82
        - quay.io/stolostron/prometheus@sha256:b959035ff754050ca092863bb74c35ba9622d9adc9e8eb7caf3b430cfe211730
        - quay.io/stolostron/grc-policy-framework-tests@sha256:327122b13ec79ea488d6f424806549e36ebfbabd055f9b40217ba541f4a118a9
        - quay.io/stolostron/multicluster-operators-subscription@sha256:d7c137379781ce6b697f73a464b89d04fbc20c3da00ce21a195ac79ab57a6dc7
        - quay.io/stolostron/kube-state-metrics@sha256:3503df61f3d9ebf3cbf66deef865450b32c7403839fdaca42504febaac00687c
        version: 0.7.1
      entries:
      - name: stolostron.v0.7.1
        version: 0.7.1
      - name: stolostron.v0.7.0
        version: 0.7.0
      name: community-0.7
    defaultChannel: community-0.7
    packageName: stolostron
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MinIO
      provider-url: https://min.io
    name: minio-object-store-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-object-store-operator.v2025.8.19175300
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aistor.min.io/v1",
                "kind": "ObjectStore",
                "metadata": {
                  "annotations": {
                    "prometheus.io/path": "/minio/v2/metrics/cluster",
                    "prometheus.io/port": "9000",
                    "prometheus.io/scrape": "true"
                  },
                  "labels": {
                    "app": "minio"
                  },
                  "name": "my-objectstore"
                },
                "spec": {
                  "cache": {
                    "enabled": false
                  },
                  "certificates": {
                    "disableAutoCert": false
                  },
                  "configuration": {
                    "name": "storage-configuration"
                  },
                  "firewall": {
                    "enabled": false
                  },
                  "mountPath": "/export",
                  "podManagementPolicy": "Parallel",
                  "pools": [
                    {
                      "name": "pool-0",
                      "securityContext": {
                        "fsGroupChangePolicy": "OnRootMismatch"
                      },
                      "servers": 4,
                      "volumeClaimTemplate": {
                        "apiVersion": "v1",
                        "kind": "persistentvolumeclaims",
                        "spec": {
                          "accessModes": [
                            "ReadWriteOnce"
                          ],
                          "resources": {
                            "requests": {
                              "storage": "1Ti"
                            }
                          },
                          "storageClassName": "standard"
                        }
                      },
                      "volumesPerServer": 4
                    }
                  ]
                }
              },
              {
                "apiVersion": "aistor.min.io/v1alpha1",
                "kind": "AdminJob",
                "metadata": {
                  "name": "minio-job"
                },
                "spec": {
                  "commands": [
                    {
                      "args": {
                        "name": "iac-bucket"
                      },
                      "op": "make-bucket"
                    }
                  ],
                  "containerSecurityContext": {},
                  "objectStore": {
                    "name": "my-objectstore",
                    "namespace": "my-objectstore"
                  },
                  "securityContext": {},
                  "serviceAccountName": "my-objectstore-sa"
                }
              },
              {
                "apiVersion": "sts.min.io/v1beta1",
                "kind": "PolicyBinding",
                "metadata": {
                  "name": "example-binding"
                },
                "spec": {
                  "application": {
                    "namespace": "my-objectstore",
                    "serviceaccount": "my-objectstore-sa"
                  },
                  "policies": [
                    "consoleAdmin"
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,AI/Machine Learning
          createdAt: "2025-08-27T14:50:56Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AdminJob
            name: adminjobs.aistor.min.io
            version: v1alpha1
          - kind: ObjectStore
            name: objectstores.aistor.min.io
            version: v1
          - kind: PolicyBinding
            name: policybindings.sts.min.io
            version: v1alpha1
          - kind: PolicyBinding
            name: policybindings.sts.min.io
            version: v1beta1
        description: "MinIO AIStor extends the most sophisticated object store on
          the market. It is designed for the exascale data \ninfrastructure challenges
          presented by modern AI workloads. AIStor provides new features, along with
          performance and\nscalability improvements, to enable enterprises to store
          all AI data in one infrastructure.\n\nMinIO is the standard for building
          <a href=\"https://min.io/solutions/object-storage-for-ai\">large scale AI
          data infrastructure</a>.\n\nThe MinIO AIStor is a software-defined, Amazon
          S3 compatible object store that will run anywhere - from the public cloud
          to the edge.<br><br>\n\nEnterprises use MinIO to deliver against artificial
          intelligence, machine learning, analytics, application, backup and archival
          workloads - all from a single platform.\n\nThe use cases include storage
          for HDFS replacements, AI datalakes, SIEM datalakes, analytics datalakes,
          object storage as a service, streaming and database warm/hot storage.<br><br>\n\nMinIO’s
          AIStor is the fastest object storage server, with production READ/WRITE
          speeds in excess of 2.2 TiB/s on commodity hardware.\n\nBuilt for the cloud
          operating model it is native to the technologies and architectures that
          define the cloud. These include containerization, orchestration with Kubernetes,
          microservices and multi-tenancy - making it the most widely integrated object
          store in the market today.<br><br>\n\nThe AIStor offers a host of enterprise
          features including inline erasure coding, bit-rot detection, state-of-the-art
          encryption, active-active replication, object locking, lifecycle management,
          observability, key-management, caching, firewall/load balancing, global
          console, search and identity + access management.<br><br>\n\nThe AIStor
          is priced on usable capacity and includes support through the MinIO Subscription
          Network.\n\nThe Subscription Network is designed to optimize deployments
          of MinIO and reduce the risk associated with data loss and data breach while
          providing 24/7/365 direct-to-engineering support.<br><br>\n\n<a href=\"https://resources.min.io/request-a-demo-marketplaces/\">Request
          a demo today</a>"
        displayName: Minio AIStor Object Store
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Storage
        - S3
        - MinIO
        - Object Storage
        links:
        - name: Minio AIStor
          url: https://min.io
        - name: Support
          url: https://subnet.min.io
        - name: Github
          url: https://github.com/minio/aistor
        maintainers:
        - email: dev@min.io
          name: MinIO
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: MinIO
          url: https://min.io
        relatedImages:
        - quay.io/minio/aistor/operator@sha256:f80139efa062fdd83f78318b20f371a827b68e854b03dd9706b9eba4c493aeb2
        version: 2025.8.19175300
      entries:
      - name: minio-object-store-operator.v2025.8.19175300
        version: 2025.8.19175300
      - name: minio-object-store-operator.v2025.8.1184927
        version: 2025.8.1184927
      - name: minio-object-store-operator.v2025.7.1011319
        version: 2025.7.1011319
      - name: minio-object-store-operator.v2025.6.3065135
        version: 2025.6.3065135
      - name: minio-object-store-operator.v2025.5.12190907
        version: 2025.5.12190907
      name: stable
    defaultChannel: stable
    packageName: minio-object-store-operator
    provider:
      name: MinIO
      url: https://min.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: https://stakater.com
    name: template-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: template-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-07-04T00:40:36Z"
          description: Operator to distribute plain kubernetes manifests, helm charts
            and, copy secrets and configmaps to multiple namespaces
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.0.1
          operatorframework.io/suggested-namespace: template-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterTemplateInstance is the Schema for the clustertemplateinstances
              API
            displayName: Cluster Template Instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateGroupInstance is the Schema for the TemplateGroupInstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.templates.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
        description: '"Operator to distribute plain kubernetes manifests, helm charts
          and, copy secrets and configmaps to multiple namespaces"'
        displayName: Template Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - templates
        - secrets
        - configmaps
        - distribution
        - stakater
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/template-operator@sha256:ad8312417e4e9621b74d3ad03020516840feef9c4b52c4f0d894dbb6dc6f906a
        version: 0.0.1
      entries:
      - name: template-operator.v0.0.1
        version: 0.0.1
      name: release-0.0
    - currentCSV: template-operator.v0.1.5
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-08-29T06:41:42Z"
          description: Operator to distribute plain kubernetes manifests, helm charts
            and, copy secrets and configmaps to multiple namespaces
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.1.5
          operatorframework.io/suggested-namespace: template-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterTemplateInstance is the Schema for the clustertemplateinstances
              API
            displayName: Cluster Template Instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateGroupInstance is the Schema for the TemplateGroupInstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.templates.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
        description: '"Operator to distribute plain kubernetes manifests, helm charts
          and, copy secrets and configmaps to multiple namespaces"'
        displayName: Template Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - templates
        - secrets
        - configmaps
        - distribution
        - stakater
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/template-operator@sha256:88f722baf22ac07704503f098f2b0ea7d76815a1ab0552474b1c9ba6639d8142
        version: 0.1.5
      entries:
      - name: template-operator.v0.1.5
        version: 0.1.5
      name: release-0.1
    defaultChannel: release-0.1
    packageName: template-operator
    provider:
      name: Stakater
      url: https://stakater.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: loki-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-operator.v5.9.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:6728046841367066216b304ed06249b80c57c74151c6a75c9df04a6f7f76e03d
          createdAt: "2025-01-30T17:34:53Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.7.0-0 <5.9.15'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:0c3943fa009f0c649f6d592b0ac410653c8c44b4a90e41fabdfa5f3f7db1e43d
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:6728046841367066216b304ed06249b80c57c74151c6a75c9df04a6f7f76e03d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:2e164e377d4fba98e7f63f43f25b6b9dad441d2e2914c2bf02f3525d5c9aa483
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:542b28fd51194b46efc4e5af10c3241b52db4c71f0d5d1b1f184eb39764e6c53
        version: 5.9.15
      entries:
      - name: loki-operator.v5.9.15
        version: 5.9.15
      - name: loki-operator.v5.9.14
        version: 5.9.14
      - name: loki-operator.v5.9.13
        version: 5.9.13
      - name: loki-operator.v5.9.12
        version: 5.9.12
      - name: loki-operator.v5.9.11
        version: 5.9.11
      - name: loki-operator.v5.9.10
        version: 5.9.10
      - name: loki-operator.v5.9.9
        version: 5.9.9
      - name: loki-operator.v5.9.8
        version: 5.9.8
      - name: loki-operator.v5.9.7
        version: 5.9.7
      - name: loki-operator.v5.9.6
        version: 5.9.6
      - name: loki-operator.v5.9.5
        version: 5.9.5
      - name: loki-operator.v5.9.4
        version: 5.9.4
      - name: loki-operator.v5.9.3
        version: 5.9.3
      - name: loki-operator.v5.9.2
        version: 5.9.2
      - name: loki-operator.v5.9.1
        version: 5.9.1
      - name: loki-operator.v5.9.0
        version: 5.9.0
      name: stable
    - currentCSV: loki-operator.v5.9.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:6728046841367066216b304ed06249b80c57c74151c6a75c9df04a6f7f76e03d
          createdAt: "2025-01-30T17:34:53Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.7.0-0 <5.9.15'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:2e164e377d4fba98e7f63f43f25b6b9dad441d2e2914c2bf02f3525d5c9aa483
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:542b28fd51194b46efc4e5af10c3241b52db4c71f0d5d1b1f184eb39764e6c53
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:0c3943fa009f0c649f6d592b0ac410653c8c44b4a90e41fabdfa5f3f7db1e43d
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:6728046841367066216b304ed06249b80c57c74151c6a75c9df04a6f7f76e03d
        version: 5.9.15
      entries:
      - name: loki-operator.v5.9.15
        version: 5.9.15
      - name: loki-operator.v5.9.14
        version: 5.9.14
      - name: loki-operator.v5.9.13
        version: 5.9.13
      - name: loki-operator.v5.9.12
        version: 5.9.12
      - name: loki-operator.v5.9.11
        version: 5.9.11
      - name: loki-operator.v5.9.10
        version: 5.9.10
      - name: loki-operator.v5.9.9
        version: 5.9.9
      - name: loki-operator.v5.9.8
        version: 5.9.8
      - name: loki-operator.v5.9.7
        version: 5.9.7
      - name: loki-operator.v5.9.6
        version: 5.9.6
      - name: loki-operator.v5.9.5
        version: 5.9.5
      - name: loki-operator.v5.9.4
        version: 5.9.4
      - name: loki-operator.v5.9.3
        version: 5.9.3
      - name: loki-operator.v5.9.2
        version: 5.9.2
      - name: loki-operator.v5.9.1
        version: 5.9.1
      - name: loki-operator.v5.9.0
        version: 5.9.0
      name: stable-5.9
    - currentCSV: loki-operator.v6.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.14-v4.17
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:1cb1d2cad8632791529eb754938612f5e5ed373284e42f309d603059ab7ec7c5
          createdAt: "2025-03-12T12:05:37Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=5.8.0-0 <6.0.10'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.0
          operators.operatorframework.io.bundle.channels.v1: stable-6.0
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:9aa8d8e4a7fd7ddad5d842a22b09aecdbdfa31ca5eb8c4b17f0818636af5fbf8
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:69e69ab2628742774d12dd9dde50607625efdff50ea4671f2859692fbee5ecb4
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:98d2cf3437faa2400c01edb3c694816e0214c8f9dca0a0b0cdf5882a720248f3
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:1cb1d2cad8632791529eb754938612f5e5ed373284e42f309d603059ab7ec7c5
        version: 6.0.10
      entries:
      - name: loki-operator.v6.0.10
        version: 6.0.10
      - name: loki-operator.v6.0.9
        version: 6.0.9
      - name: loki-operator.v6.0.8
        version: 6.0.8
      - name: loki-operator.v6.0.7
        version: 6.0.7
      - name: loki-operator.v6.0.6
        version: 6.0.6
      - name: loki-operator.v6.0.5
        version: 6.0.5
      - name: loki-operator.v6.0.4
        version: 6.0.4
      - name: loki-operator.v6.0.3
        version: 6.0.3
      - name: loki-operator.v6.0.2
        version: 6.0.2
      - name: loki-operator.v6.0.1
        version: 6.0.1
      - name: loki-operator.v6.0.0
        version: 6.0.0
      name: stable-6.0
    - currentCSV: loki-operator.v6.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.15-v4.18
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:0bde3c1a0b667a693f443688ed3926eef1705c45ecf4cce3a3dedbddca7a2784
          createdAt: "2025-03-12T11:19:10Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=5.9.0-0 <6.1.8'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.1
          operators.operatorframework.io.bundle.channels.v1: stable-6.1
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:0bde3c1a0b667a693f443688ed3926eef1705c45ecf4cce3a3dedbddca7a2784
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:731f92c146f299bc3662b309a3068bde18fad51f65a7d29063e582b61e07042f
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:c34676932ddd14db213dbf42a1d341f51692f3f47f1b751d06187e292c99a170
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b94833f55644c289fe2c1297b68a4f9931baf977587c6a2c4c7a3e484c704f64
        version: 6.1.8
      entries:
      - name: loki-operator.v6.1.8
        version: 6.1.8
      - name: loki-operator.v6.1.7
        version: 6.1.7
      - name: loki-operator.v6.1.6
        version: 6.1.6
      - name: loki-operator.v6.1.5
        version: 6.1.5
      - name: loki-operator.v6.1.4
        version: 6.1.4
      - name: loki-operator.v6.1.3
        version: 6.1.3
      - name: loki-operator.v6.1.2
        version: 6.1.2
      - name: loki-operator.v6.1.1
        version: 6.1.1
      - name: loki-operator.v6.1.0
        version: 6.1.0
      name: stable-6.1
    - currentCSV: loki-operator.v6.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.16-v4.19
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f401be018c5ba35e8ea1e68831cc20067f0f4d1be031d05b222548c1bdfb5f7e
          createdAt: "2025-02-18T18:01:39Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=6.0.0-0 <6.2.5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.2
          operators.operatorframework.io.bundle.channels.v1: stable-6.2
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f401be018c5ba35e8ea1e68831cc20067f0f4d1be031d05b222548c1bdfb5f7e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:99d1e7c5c1f0609b33de576884782550f969b28dc622810e6661d4c60a664215
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:8955d8186ae98095623f2083b6d5389d3ac9f2612fcfcf695835219c3e1216c4
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b8c5b46a94a6c85c9d382eeeb65a1d2b4014a7dc765e9abfba706c92857462dc
        version: 6.2.5
      entries:
      - name: loki-operator.v6.2.5
        version: 6.2.5
      - name: loki-operator.v6.2.4
        version: 6.2.4
      - name: loki-operator.v6.2.3
        version: 6.2.3
      - name: loki-operator.v6.2.2
        version: 6.2.2
      - name: loki-operator.v6.2.1
        version: 6.2.1
      - name: loki-operator.v6.2.0
        version: 6.2.0
      name: stable-6.2
    defaultChannel: stable-6.2
    packageName: loki-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudtrail-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudtrail-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudtrail.services.k8s.aws/v1alpha1",
                "kind": "EventDataStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudtrail.services.k8s.aws/v1alpha1",
                "kind": "Trail",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudtrail-controller:1.1.1
          createdAt: "2025-09-19T21:16:25Z"
          description: AWS CloudTrail controller is a service controller for managing
            CloudTrail resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventDataStore represents the state of an AWS cloudtrail
              EventDataStore resource.
            displayName: EventDataStore
            kind: EventDataStore
            name: eventdatastores.cloudtrail.services.k8s.aws
            version: v1alpha1
          - description: Trail represents the state of an AWS cloudtrail Trail resource.
            displayName: Trail
            kind: Trail
            name: trails.cloudtrail.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Compute Cloud (CloudTrail) resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudTrail**

          AWS CloudTrail is an AWS service that helps you enable operational and risk auditing, governance, and compliance of your AWS account. Actions taken by a user, role, or an AWS service are recorded as events in CloudTrail. Events include actions taken in the AWS Management Console, AWS Command Line Interface, and AWS SDKs and APIs.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudTrail
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudtrail
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudTrail Developer Resources
          url: https://aws.amazon.com/cloudtrail/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudtrail maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudtrail-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-cloudtrail-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudtrail-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudtrail-controller.v1.0.28
        version: 1.0.28
      - name: ack-cloudtrail-controller.v1.0.26
        version: 1.0.26
      - name: ack-cloudtrail-controller.v1.0.25
        version: 1.0.25
      - name: ack-cloudtrail-controller.v1.0.24
        version: 1.0.24
      - name: ack-cloudtrail-controller.v1.0.23
        version: 1.0.23
      - name: ack-cloudtrail-controller.v1.0.22
        version: 1.0.22
      - name: ack-cloudtrail-controller.v1.0.21
        version: 1.0.21
      - name: ack-cloudtrail-controller.v1.0.20
        version: 1.0.20
      - name: ack-cloudtrail-controller.v1.0.19
        version: 1.0.19
      - name: ack-cloudtrail-controller.v1.0.16
        version: 1.0.16
      - name: ack-cloudtrail-controller.v1.0.15
        version: 1.0.15
      - name: ack-cloudtrail-controller.v1.0.14
        version: 1.0.14
      - name: ack-cloudtrail-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudtrail-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudtrail-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudtrail-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudtrail-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudtrail-controller.v1.0.5
        version: 1.0.5
      - name: ack-cloudtrail-controller.v1.0.4
        version: 1.0.4
      - name: ack-cloudtrail-controller.v1.0.3
        version: 1.0.3
      - name: ack-cloudtrail-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudtrail-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudtrail-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudtrail-controller.v0.1.3
        version: 0.1.3
      - name: ack-cloudtrail-controller.v0.1.2
        version: 0.1.2
      - name: ack-cloudtrail-controller.v0.1.1
        version: 0.1.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudtrail-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-route53-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-route53-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/route53-controller:1.1.1
          createdAt: "2025-09-19T19:15:31Z"
          description: AWS Route53 controller is a service controller for managing
            Route53 resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HealthCheck represents the state of an AWS route53 HealthCheck
              resource.
            displayName: HealthCheck
            kind: HealthCheck
            name: healthchecks.route53.services.k8s.aws
            version: v1alpha1
          - description: HostedZone represents the state of an AWS route53 HostedZone
              resource.
            displayName: HostedZone
            kind: HostedZone
            name: hostedzones.route53.services.k8s.aws
            version: v1alpha1
          - description: RecordSet represents the state of an AWS route53 RecordSet
              resource.
            displayName: RecordSet
            kind: RecordSet
            name: recordsets.route53.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Route53 resources in AWS from within your Kubernetes cluster.

          **About Amazon Route53**

          Amazon Route 53 is a highly available and scalable Domain Name System (DNS) web service. You can use Route 53 to perform three main functions in any combination: domain registration, DNS routing, and health checking.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Route53
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - route53
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Route53 Developer Resources
          url: https://docs.aws.amazon.com/route53/index.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/route53-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-route53-controller.v1.1.1
        version: 1.1.1
      - name: ack-route53-controller.v1.1.0
        version: 1.1.0
      - name: ack-route53-controller.v1.0.6
        version: 1.0.6
      - name: ack-route53-controller.v1.0.5
        version: 1.0.5
      - name: ack-route53-controller.v1.0.3
        version: 1.0.3
      - name: ack-route53-controller.v1.0.2
        version: 1.0.2
      - name: ack-route53-controller.v0.1.0
        version: 0.1.0
      - name: ack-route53-controller.v0.0.28
        version: 0.0.28
      - name: ack-route53-controller.v0.0.27
        version: 0.0.27
      - name: ack-route53-controller.v0.0.26
        version: 0.0.26
      - name: ack-route53-controller.v0.0.25
        version: 0.0.25
      - name: ack-route53-controller.v0.0.24
        version: 0.0.24
      - name: ack-route53-controller.v0.0.23
        version: 0.0.23
      - name: ack-route53-controller.v0.0.22
        version: 0.0.22
      - name: ack-route53-controller.v0.0.20
        version: 0.0.20
      - name: ack-route53-controller.v0.0.19
        version: 0.0.19
      - name: ack-route53-controller.v0.0.18
        version: 0.0.18
      - name: ack-route53-controller.v0.0.17
        version: 0.0.17
      - name: ack-route53-controller.v0.0.16
        version: 0.0.16
      - name: ack-route53-controller.v0.0.15
        version: 0.0.15
      - name: ack-route53-controller.v0.0.14
        version: 0.0.14
      - name: ack-route53-controller.v0.0.13
        version: 0.0.13
      - name: ack-route53-controller.v0.0.9
        version: 0.0.9
      - name: ack-route53-controller.v0.0.6
        version: 0.0.6
      - name: ack-route53-controller.v0.0.4
        version: 0.0.4
      - name: ack-route53-controller.v0.0.3
        version: 0.0.3
      - name: ack-route53-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-route53-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://www.medik8s.io/
    name: self-node-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: self-node-remediation.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/self-node-remediation-operator:v0.9.0
          createdAt: "2024-07-25 14:40:52"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.9.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Medik8s
          url: https://www.medik8s.io/
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/medik8s/self-node-remediation-operator:v0.9.0
        version: 0.9.0
      entries:
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      - name: self-node-remediation.v0.7.0
        version: 0.7.0
      - name: self-node-remediation.v0.6.1
        version: 0.6.1
      - name: self-node-remediation.v0.6.0
        version: 0.6.0
      - name: self-node-remediation.v0.5.0
        version: 0.5.0
      - name: self-node-remediation.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: self-node-remediation
    provider:
      name: Medik8s
      url: https://www.medik8s.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudfront-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudfront-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "CachePolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "Distribution",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudfront-controller:1.2.1
          createdAt: "2025-09-19T21:12:14Z"
          description: AWS CloudFront controller is a service controller for managing
            CloudFront resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CachePolicy represents the state of an AWS cloudfront CachePolicy
              resource.
            displayName: CachePolicy
            kind: CachePolicy
            name: cachepolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: Distribution represents the state of an AWS cloudfront Distribution
              resource.
            displayName: Distribution
            kind: Distribution
            name: distributions.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: Function represents the state of an AWS cloudfront Function
              resource.
            displayName: Function
            kind: Function
            name: functions.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: OriginAccessControl represents the state of an AWS cloudfront
              OriginAccessControl resource.
            displayName: OriginAccessControl
            kind: OriginAccessControl
            name: originaccesscontrols.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: OriginRequestPolicy represents the state of an AWS cloudfront
              OriginRequestPolicy resource.
            displayName: OriginRequestPolicy
            kind: OriginRequestPolicy
            name: originrequestpolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: ResponseHeadersPolicy represents the state of an AWS cloudfront
              ResponseHeadersPolicy resource.
            displayName: ResponseHeadersPolicy
            kind: ResponseHeadersPolicy
            name: responseheaderspolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: VPCOrigin represents the state of an AWS cloudfront VPCOrigin
              resource.
            displayName: VPCOrigin
            kind: VPCOrigin
            name: vpcorigins.cloudfront.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudFront resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudFront**

          Amazon CloudFront is a web service that speeds up distribution of your static and dynamic web content, such as .html, .css, .js, and image files, to your users. CloudFront delivers your content through a worldwide network of data centers called edge locations.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon CloudFront
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudfront
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudFront Developer Resources
          url: https://aws.amazon.com/cloudfront/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudfront maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudfront-controller:1.2.1
        version: 1.2.1
      entries:
      - name: ack-cloudfront-controller.v1.2.1
        version: 1.2.1
      - name: ack-cloudfront-controller.v1.2.0
        version: 1.2.0
      - name: ack-cloudfront-controller.v1.1.6
        version: 1.1.6
      - name: ack-cloudfront-controller.v1.1.4
        version: 1.1.4
      - name: ack-cloudfront-controller.v1.1.3
        version: 1.1.3
      - name: ack-cloudfront-controller.v1.1.2
        version: 1.1.2
      - name: ack-cloudfront-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudfront-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudfront-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudfront-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudfront-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudfront-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudfront-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudfront-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudfront-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudfront-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudfront-controller.v0.0.15
        version: 0.0.15
      - name: ack-cloudfront-controller.v0.0.14
        version: 0.0.14
      - name: ack-cloudfront-controller.v0.0.13
        version: 0.0.13
      - name: ack-cloudfront-controller.v0.0.12
        version: 0.0.12
      - name: ack-cloudfront-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudfront-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudfront-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudfront-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudfront-controller.v0.0.6
        version: 0.0.6
      - name: ack-cloudfront-controller.v0.0.5
        version: 0.0.5
      - name: ack-cloudfront-controller.v0.0.4
        version: 0.0.4
      - name: ack-cloudfront-controller.v0.0.3
        version: 0.0.3
      - name: ack-cloudfront-controller.v0.0.2
        version: 0.0.2
      - name: ack-cloudfront-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudfront-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/openshift/lightspeed-service
    name: lightspeed-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lightspeed-operator.v1.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ols.openshift.io/v1alpha1",
                "kind": "OLSConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightspeed-operator",
                    "app.kubernetes.io/instance": "olsconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "olsconfig",
                    "app.kubernetes.io/part-of": "lightspeed-operator"
                  },
                  "name": "cluster"
                },
                "spec": {
                  "llm": {
                    "providers": [
                      {
                        "credentialsSecretRef": {
                          "name": "credentials"
                        },
                        "models": [
                          {
                            "name": "gpt-3.5-turbo-1106"
                          }
                        ],
                        "name": "OpenAI"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-08-28T16:37:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lightspeed
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Kubernetes Engine", "OpenShift Virtualization
            Engine"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lightspeed-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Red Hat OpenShift Lightspeed instance. OLSConfig is the Schema
              for the olsconfigs API
            displayName: OLSConfig
            kind: OLSConfig
            name: olsconfigs.ols.openshift.io
            version: v1alpha1
        description: |-
          OpenShift Lightspeed Operator provides generative AI-based virtual assistant which integrates into the OpenShift web console. OpenShift Lightspeed can answer natural language questions related to OpenShift Container Platform.

          To configure the OpenShift Lightspeed Operator, you need a Large Language Model (LLM) . You can host the LLM using RHELAI or RHOAI or use one of the models from the following Software-as-a-Service providers:

          - OpenAI
          - Microsoft Azure OpenAI
          - IBM WatsonX

          For more information, see [About OpenShift Lightspeed](https://docs.redhat.com/en/documentation/red_hat_openshift_lightspeed/1.0tp1/html/about/ols-about-openshift-lightspeed) in the official product documentation.

          Questions, comments, concerns, or just want to talk to the OpenShift Lightspeed team? [Email us](mailto:openshift-lightspeed-contact-requests@redhat.com)!
        displayName: OpenShift Lightspeed Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ai
        - assistant
        - openshift
        - llm
        links:
        - name: Lightspeed Operator
          url: https://github.com/openshift/lightspeed-operator
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/lightspeed-service
        relatedImages:
        - registry.redhat.io/openshift-lightspeed/lightspeed-rhel9-operator@sha256:b12803642df4d9b3ad0cf800315021483f543c6a8827b3901e57bdb65f10dc48
        - registry.redhat.io/openshift-lightspeed/openshift-mcp-server-rhel9@sha256:f1e4521ff2cc076f361c04d8c9aeac09ffcedb89a2f1dbe71f85caa27472adab
        - registry.redhat.io/openshift-lightspeed/lightspeed-service-api-rhel9@sha256:ec359967a87bc7ee2c1d6a61e553acbdb0dc85d6f240a1865bf57d3ba5dc7193
        - registry.redhat.io/openshift-lightspeed/lightspeed-console-plugin-rhel9@sha256:dfc42c2651ac8793f61ff0e54334ecb006dc4a77487bc7c6878cdf47bd4ac080
        version: 1.0.5
      entries:
      - name: lightspeed-operator.v1.0.5
        version: 1.0.5
      name: alpha
    - currentCSV: lightspeed-operator.v1.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ols.openshift.io/v1alpha1",
                "kind": "OLSConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightspeed-operator",
                    "app.kubernetes.io/instance": "olsconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "olsconfig",
                    "app.kubernetes.io/part-of": "lightspeed-operator"
                  },
                  "name": "cluster"
                },
                "spec": {
                  "llm": {
                    "providers": [
                      {
                        "credentialsSecretRef": {
                          "name": "credentials"
                        },
                        "models": [
                          {
                            "name": "gpt-3.5-turbo-1106"
                          }
                        ],
                        "name": "OpenAI"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-08-28T16:37:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lightspeed
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Kubernetes Engine", "OpenShift Virtualization
            Engine"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lightspeed-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Red Hat OpenShift Lightspeed instance. OLSConfig is the Schema
              for the olsconfigs API
            displayName: OLSConfig
            kind: OLSConfig
            name: olsconfigs.ols.openshift.io
            version: v1alpha1
        description: |-
          OpenShift Lightspeed Operator provides generative AI-based virtual assistant which integrates into the OpenShift web console. OpenShift Lightspeed can answer natural language questions related to OpenShift Container Platform.

          To configure the OpenShift Lightspeed Operator, you need a Large Language Model (LLM) . You can host the LLM using RHELAI or RHOAI or use one of the models from the following Software-as-a-Service providers:

          - OpenAI
          - Microsoft Azure OpenAI
          - IBM WatsonX

          For more information, see [About OpenShift Lightspeed](https://docs.redhat.com/en/documentation/red_hat_openshift_lightspeed/1.0tp1/html/about/ols-about-openshift-lightspeed) in the official product documentation.

          Questions, comments, concerns, or just want to talk to the OpenShift Lightspeed team? [Email us](mailto:openshift-lightspeed-contact-requests@redhat.com)!
        displayName: OpenShift Lightspeed Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ai
        - assistant
        - openshift
        - llm
        links:
        - name: Lightspeed Operator
          url: https://github.com/openshift/lightspeed-operator
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/lightspeed-service
        relatedImages:
        - registry.redhat.io/openshift-lightspeed/lightspeed-service-api-rhel9@sha256:ec359967a87bc7ee2c1d6a61e553acbdb0dc85d6f240a1865bf57d3ba5dc7193
        - registry.redhat.io/openshift-lightspeed/lightspeed-console-plugin-rhel9@sha256:dfc42c2651ac8793f61ff0e54334ecb006dc4a77487bc7c6878cdf47bd4ac080
        - registry.redhat.io/openshift-lightspeed/openshift-mcp-server-rhel9@sha256:f1e4521ff2cc076f361c04d8c9aeac09ffcedb89a2f1dbe71f85caa27472adab
        - registry.redhat.io/openshift-lightspeed/lightspeed-rhel9-operator@sha256:b12803642df4d9b3ad0cf800315021483f543c6a8827b3901e57bdb65f10dc48
        version: 1.0.5
      entries:
      - name: lightspeed-operator.v1.0.5
        version: 1.0.5
      name: stable
    defaultChannel: stable
    packageName: lightspeed-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/openshift/lightspeed-service
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kubevirt-hyperconverged
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubevirt-hyperconverged-operator.v4.16.23
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:eaacb0d177fd64e2eefbcb2dc47633dadf4d54fcbe0f0f750b078ab3f4f189e1
          createdAt: "2025-09-29 21:23:00"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.14.1 <4.16.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:3e1c68d5695bab96c382661a33581e60597b59afd21660ec760d8fe5e20ff9e3
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.16.23-1,hyperconverged-cluster-operator-test-rhel9:v4.16.23-1,kubevirt-common-instancetypes-test-rhel9:v4.16.23-1,kubevirt-tekton-tasks-test-rhel9:v4.16.23-1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:e24191be577575fa9d462c87c6f51775a0cf0f5d1812d12c4c95a68ec34b34ff
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:6e4aa75bece3c2fe076e282b3f1ad8296299ec0551e379c7c88830a0ce8ed5bf
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:1be66eee321bc47dcc0e7ff4674a485b9a4552243080d796d5f3b0d417721dc4
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:28c0bdf60404fb435cfa0cfdc6d314f789417c44c3206ce8cd1e7b99031fb529
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:659abe1d740607b2cb3850c9ab29101fa5d5c821d2a09aab5994849697a504f3
        - registry.redhat.io/container-native-virtualization/mtq-operator-rhel9@sha256:9de5ded35edfbe98121bdd4404c84dcc72c8cba1742dc27df6382ef882db2183
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:0f0478e9357f8b6ace6d844e6e424db192398195a8cea2fa6398be8996d05be6
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:2843d199dc2748749d5f48fb9f90656962e011533d79dc7e0a4f3592b12aa69a
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:6b0c7cc17abd95fce45b6ec83ec291e843eedbc90dce0ecc449aec6e64562057
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:97d37927164f0ea986f209912e6247fdab37b680c214afc4bdbb7051caaf58b1
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:cc4ff488aa97e179c4df39ac5ae8085070017106adf1c5f2222b35d8f2234ea8
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:f8e2ea3988d4a77a4e809e1fd073ffa5f99393ca3eb69a2c1cb54dd4ab60e482
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:885670f2e9ad43224cbc96f34b23dd2642de00f688bade0f6b89dd3d3fb83dbd
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:251d6fb3197a7642d86193191aa69c85a45c5521d7fc509704edd46d5d4e90a7
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:fcbadecf01b81625d55f9219428baabc33ee7231b4ccf801c90dee3095907187
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:a0af0e5d4b142f28dc72c5636df5040c4ce92c6b12ab183d3a9bc7a9d2d2f418
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:c033ddd0ed30c0d25f4c2e9b30fca4e40959bb6d74029a1f452ff048b97de88a
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:7206ac7959c07f474f454ab48bcf3522ade95efb27fd9fbd3938d1387a1420af
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:468f117a1b55af59e72a9ff46499a8d0bb00884d450d57b692e2f7b557c9bf3d
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:e01871006b9e4cb1b6f477319de63b758eac672a70a298caed2032ec19986680
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:7587832b877f32621eb26e67a5c516728f293ecd98f74dd70d5f03474a5429d6
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:68b650d6779fc4d61f05d97e6d62b80fcd638721994153205eec68f8c1d999ce
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:19ca70dceb79b10ff458f0ea41faa6e5552a5f08828620ad7177efa34166a90e
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:84dbfa51755d60e34b194c1acf679ac083feb49038112212828fe8101dcec2a5
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:8462188ec7533d2105c0dc9154d8dcf1425b96d6945eef16ea774a5dff9bba1d
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:11030a8d8cd03efdf27fe3cfc9ebeadeb20851b6de1e9d8a93ba7f2bc6a77a96
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:8cf7993326ba220f650ae2fa5383d0d6cf3e2c7f3346f0a339a9d212b8df02f6
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:c18b6e315f5577cc18d46cd76595d48d4a49026a6b2f5763c3efb26919bde5eb
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:52b4f7c7a4f5c5c5d07a720c5eefa70709f520b1d939aec4cd971d8f3c9b12bc
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:ef8260d72239ad2126ca023250bfd976f5c050fe9c0727a0f7ef007446f5363b
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:18be4ef8b6c2cedab4a2e031939fc1f9cc23020bde7fccb7997e4222b673f57f
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:def80efdcc58d545aaf4b69b784a348cb36e97afd8e0124480521a2b980fa7c3
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:b031a12e73cb75e207203fd3785619e3a5a22d7d5a8549488c035f70374235be
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:14ea56d4149b72b98b6e7f5b33ba5bdd9115e7c643fd103c9df7a26b2a06700a
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:9e358366fd11dd167255abd930016aa4fc35b1b9dc69ce27316635488efc8ec4
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:6bf45534310768308958428692b29afa25f2740d0c62165d7182b0b3e9e944cf
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:eaacb0d177fd64e2eefbcb2dc47633dadf4d54fcbe0f0f750b078ab3f4f189e1
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:91b86228495f38d5528be1a8ffec96695ad72fafb76d2ead87f7bee25540d8b6
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:19b56bbb081db7198d570c3c691e142991aaa81857ee417d47903291ed0e7a1c
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:372652529b6a8705df581e7ab8acc2ec5c3cdf8fb273448052c59963abe4d984
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:7f89f2ba59309b26cf83f58ff3ba35569182097258fa543d5d88f3a8cd64ca00
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:bf1bb2fe09e028506fe7cf383368d76e358b8c59ecc74d3618bc250c1e97271c
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:a79579b18a7df9f23c584303593a9404c64df023458d55c715aeb71dcacc92a0
        - registry.redhat.io/container-native-virtualization/mtq-lock-server-rhel9@sha256:bb1dfc204e48a13985d6c2ec9cdf17f47330617725ee6dd854f0dae0f1901ba4
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:4005d165f1485eec98ca24ce41bef79cd8088002d43ffb709ec65f842a176bdc
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:3f3554ef043c80494db536b28eb03fa7043e53bb85ce426820c252fb999ce04e
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:531702869a34f76850d332efb84ccf571fafa281014afd1787497ca9de8aed37
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:a081e7e68def13b7de369d3f66469035d61fd1feef1bee4f95ba5f7b4041d960
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:e7672121156e050f05af771f75cfe26558699da8a074bf7051f710705a27d743
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:bd532e59edae5d2edac78b9249cd6a9f0779f44e25b019d824a1b9803f078bde
        - registry.redhat.io/container-native-virtualization/mtq-controller-rhel9@sha256:a774f0bd125c66b45e35e327c9c7e0afd7947d3c5ed222b3d5d223941366ba2c
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:52e76b4208ee4e7da8a0729b9bdb294729454d0f62580d64dd104b6921f464ed
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:3b57db7aaa3c61335044ec78cbd061f78ce99cdb4b75b3ce342e1a603da4bcef
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:7e19b42a648cee71c60cb69fbcf6ebca05c938942e16d23f1cd61a619cffdcb6
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:4046e8e862e0c55f09e9031b95e6a77af4d491593e39770c644287d44888b0bb
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:517c88a6be8842d4eb71ad30f99ce1afca954af1df2c1bf4a9584f4ec69fcf12
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:3e1c68d5695bab96c382661a33581e60597b59afd21660ec760d8fe5e20ff9e3
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:cca121312777d2727fe2c50a14583243ce69e6e356b499f133736787522095b2
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:ad24eb52bd65495224661d4894336d88fde342aeaab16b1084ceff6337f92059
        version: 4.16.23
      entries:
      - name: kubevirt-hyperconverged-operator.v4.16.23
        version: 4.16.23
      - name: kubevirt-hyperconverged-operator.v4.16.22
        version: 4.16.22
      - name: kubevirt-hyperconverged-operator.v4.16.21
        version: 4.16.21
      - name: kubevirt-hyperconverged-operator.v4.16.20
        version: 4.16.20
      - name: kubevirt-hyperconverged-operator.v4.16.19
        version: 4.16.19
      - name: kubevirt-hyperconverged-operator.v4.16.18
        version: 4.16.18
      - name: kubevirt-hyperconverged-operator.v4.16.17
        version: 4.16.17
      - name: kubevirt-hyperconverged-operator.v4.16.16
        version: 4.16.16
      - name: kubevirt-hyperconverged-operator.v4.16.15
        version: 4.16.15
      - name: kubevirt-hyperconverged-operator.v4.16.14
        version: 4.16.14
      - name: kubevirt-hyperconverged-operator.v4.16.13
        version: 4.16.13
      - name: kubevirt-hyperconverged-operator.v4.16.12
        version: 4.16.12
      - name: kubevirt-hyperconverged-operator.v4.16.11
        version: 4.16.11
      - name: kubevirt-hyperconverged-operator.v4.16.10
        version: 4.16.10
      - name: kubevirt-hyperconverged-operator.v4.16.9
        version: 4.16.9
      - name: kubevirt-hyperconverged-operator.v4.16.8
        version: 4.16.8
      name: candidate
    - currentCSV: kubevirt-hyperconverged-operator.v4.99.0-0.1723448771
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:5b2c64218342a5744182fc06f550bcdcf954e931275a9d1505ef0bbfbd0f8044
          createdAt: "2024-08-12 07:46:45"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.14.1 <4.16.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:d74c629e415d177069a0570fb8db9252d2b193d2ec7972c844686ffd14033f62
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.99.0-73,kubevirt-common-instancetypes-test-rhel9:v4.99.0-32,kubevirt-tekton-tasks-test-rhel9:v4.99.0-31
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:cc190d7c44b20f10ae61709fd2f5f7c0b248df73b768eb696d5a6cec90a34e59
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:41ee360ac43ee69ccee62d4ac3d2f07c8a8589e5b85c659b47018dd36c2b932f
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:b410508ecc71b0a3bf4825793efde4b07085e2243d8b3585da852fd1940c4be1
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:820a9e6bd0329df5c541ef0b67b3c7cdd9f81be33de87c5b5e5e43777bea197e
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:811e8e7e9872408794ef36229c42c6c437b1a4ace76c7134d91bb2331fe7815b
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:0769512df0fa1325d76a5f10a1a704cf60fc65f62f57b157ea53bc9ba682851f
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:8c64b526fd42a17c909bf45eb795121a81d502424a9cf11f8087118751333894
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:5f23fc8e9bc386622460c23b9644f58648d9f748f5670927a02888cd36c99a21
        - registry.redhat.io/container-native-virtualization/mtq-controller-rhel9@sha256:ad5938ebd5f07196ce5c35232a1c3b0a653301a83929ce84c3d1cb4dd1b91816
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:37e6b4fa5ff436572412ec0bdb1edc4a49e62c994e056c93578605f6b74adbcd
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:07435335fdbf6178bc00cc02b93cc0adb42784339151b2d9c31dc8da98b8bc6a
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:cffec516ab9ddc50cab77339d0cb3239ef4e7fd43bd2afdc73370aa487ea5704
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:dec430b74887e9b42d639f14c07d9f58a4881cac24e949dab98f96af150f1958
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:c61732be8024b4ede45eec8f786e51b198632ee5bfb0c2f5af84f37544584283
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:5ba79dd6bdf9e3778d3e301919b37e45cfc1eaf72fda77eba6588788829ba29b
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:c14cd037d14489321629908fc6b30248f7a20d866619177a165d72cfc1b6b7ef
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:83c03929cf11d7202e6c6cb9dd767bd4d7473f4fc1d562b79719ceb3d8382e68
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:6b7beed3cba9093b3aeb3109f23ec38c20f5568915dff774ad5e9e6bb5db3a58
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:e39732ebd3a4452fa7a678aeea5e64a7973fd72802eff193b16680c2f0129cb1
        - registry.redhat.io/container-native-virtualization/mtq-lock-server-rhel9@sha256:01cc6797800ccfb99453858ab36c0100c5ca343f1f71f975914b2437d1f757ba
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:9822f7a1590d830ccd64cc0729d2e30a2e04fb2b022f54498e1f1abff3748c2c
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:bf2ec130d1f3f32314d740037313f49f874a271e8b9a7f6cd2b93698a8423e8d
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:42a34caa18b5f5452b9d5603602ea74b1ee59a8bbb52bddf18901fecc27aed8a
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:18dfb38223647e11020e3ef1093a695454b90cc50922f80afb612212b7d13309
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:4c66a26b9f74d83a9bc60edb40ffd9d6ff2382d20500677d4567f8e141c3ddf0
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:758a9f6309a367d3f31cc03dd89356a8232a0fbd277570c01ebfa1fa30089a8d
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:2e5d5a2a70cda1f414f22b1c713d77d068c5fb16b1487a572019a9aa7130e6be
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:84b199e98e283c863d52a7b00044c20cb56ad563494adbec2229af265bf3f7c5
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:6baf67f3f410c071cc26f675d86edbbae5eaeb3410888cf622fad384b43b80ea
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:3a45cd2f3ea25f356167ce4fa20894a2a0706b4bcffa42c806f6017812053466
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:84679ffe8ce86c674818531897d122f5568d10c7e1e2ff919185d5771e5b4c39
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:ac33de246ee42bbfe735d7349d85c6df7e583e7599dea7dd7a2235bd0e59845f
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:e7145f016642b02bd651e21aa19edc7f93d0d1c2da3c6c7eed5d1174cf23619c
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:2564af054f52a8bb340e0270379c72da18658662a2a5bd26ac7bd979ddfccb56
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:4cff43cc145361c74c357055d51060ccc27e6bca58ccc3b5f5b983c7b8545114
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:09a30b59c906ddc4604d70f18210bcebef1f90f8bf965422b0caba363cd88989
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:63389195cf584a07ddbc2efec90cf6deaf737b91314b67db2b9f53dcc7acb57c
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:6193e569d4ab43806ac7b4870ef3634a5de3687527d6d190e59a9f735fa4778c
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:f4e7876fa8b27a8082efa99455d9bf7b0c479cff1e4c8a26136bf63092699c0b
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:d7e8e075a08b148ddfa91b44bb9cae1ef823756760279c4bb6d8489bc7e0eb2c
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:bf30b6c7173ebb6a4592246af16db87d7d391182f52286b559f8bad8f4b6cc5b
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:15179f0826831f051fdb867067be7c5528f01f442ae421b8782ae0cd1ed1b73d
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:4b3ac5ec2fafef0b7702a02847996f0903ec8d4262ccd03e8fa74885265ea4e9
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:d6dfd062df3d0d49f7efb928b835b9c8e0bcfb514a653ba2896a0152f24a0a21
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:185a83c994aced74546fe630cd71656f1e4112cb15be75f41dfcb1b9ed0966bd
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:b59e34cdb6214eacb9dc176a7b33125d7c1871f6b81696d59646294bfaf749e7
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:8e19d99232be4c37e857a4931627fb0aed721893e27b0d0e522fcf13bf67c1c9
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:06cc71822684c3ed094f327553451daf076d6bb97065c314da1a7d2adbbef3d2
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:4bdafc4404d5cfd844d05f32d88c262fba56dc931bfd5522b78c25d4ff253883
        - registry.redhat.io/container-native-virtualization/kubevirt-ipam-controller-rhel9@sha256:f33ac4d7477a730f273a462f171aed35dea38374d5ea7ef0bae2d7f74413a757
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:af5d2a2d4f0b6f0e1a13f1029073bb56980d8a70503b21628f18e8986e9b531c
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:5b2c64218342a5744182fc06f550bcdcf954e931275a9d1505ef0bbfbd0f8044
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:88f02cbebd6c3a2add6e0f8ca744e42bbd182f5a07d4fac59cf5cf07a4edfd45
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:fa0478fbf91248166f0eed8ec5f8dbb5b31d26031438376d9301836b6a069643
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:9bfdaf1cca6eac15b63a4ea6b7de28f2328516c5cf8bd5727935dd729e45657d
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:30c669770b0b992a3c84bb81c892e70a24fb14af187e8a4ee19cb3e0dd56ddbb
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:c5e8f44f647f593b00d3427172932d7aa694379d124209b9c66b4c1ffe3d4116
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:e6d9a6c41433360abd42d1118f5af3df9592d6d36af5f65082ae214cc8513c1b
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:d74c629e415d177069a0570fb8db9252d2b193d2ec7972c844686ffd14033f62
        - registry.redhat.io/container-native-virtualization/mtq-operator-rhel9@sha256:033bf7a6ef8ced1ac7275a9ddea8b16a10340a0fb0fb39619d89354d661f05b6
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:442e5cbcc63961642b662544032240bffb3863227324128a67f0bb6a05a1c364
        version: 4.99.0-0.1723448771
      entries:
      - name: kubevirt-hyperconverged-operator.v4.99.0-0.1723448771
        version: 4.99.0-0.1723448771
      name: dev-preview
    - currentCSV: kubevirt-hyperconverged-operator.v4.16.21
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:08f15657a1eee02927c5a6ef661e6e703d0064e091751bd2d8fbad5f118d9672
          createdAt: "2025-09-20 19:53:21"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.14.1 <4.16.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:88863a8dd2145bf17f7c5378c81e7c0be7a01dc47ee75bd7e5203a8fed8895b9
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.16.21-3,hyperconverged-cluster-operator-test-rhel9:v4.16.21-1,kubevirt-common-instancetypes-test-rhel9:v4.16.21-1,kubevirt-tekton-tasks-test-rhel9:v4.16.21-1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:9afe1ec9f776854caae38b89198a82e8e47a45d6c08f8dbd65a06c85f741c6a0
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:611d60b6ea4285fbc8a9f945fa5d06c09679a8f7ec1d70ec18edae7eec1a6bb3
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:299825fa33ab3f491cf954dd2e4ba23fdba2a96723b9ac3d98b3264b0b9fbb28
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:f6b075a56380143a7892e1fe5fdb212f8aa1c974f6a2d7e7e9b940789e44c010
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:d346688e02d03ddd9918f7a344d6eb3d5ddac3f1802dc82daf72bc56e7a9d496
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:406681aff7ea0a17941449593f839418edffdb073e2afffd53174f0df6292f94
        - registry.redhat.io/container-native-virtualization/mtq-controller-rhel9@sha256:253ef56b2867d33b3a475a456878b4a60f56473122eeed8c169568a9724ed30c
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:e54d87e6735151ed3c5295b23ca2f7ec666afb6421923636ef836bca1cf883d6
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:6385b346b3dd8f2bb3284871d326c851cf743a5e3fc3c49a1e51a1d43b0eafb4
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:8cd1052b9f69d0a0d8db41e90414756e83cbd07d7bd03c2faa85c41393d6539c
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:b80e68822fb0b5e78a995a97e02fe2ec448d67a5aa2224497fdf12b812182a0b
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:51a1fb0eb77e4f5c8bdfaf3b5074369b93ef63f5a7af62e22ed0789731f8dc73
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:361a48df213d52ced2a9963c2a8bde72a0981f1c497235fb918ea8c065580eeb
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:355e0946cfa963aeb5b6c3e8d476a9c3aee202baf77b95ce3f69c8e4b9e73836
        - registry.redhat.io/container-native-virtualization/mtq-lock-server-rhel9@sha256:c447251fea804340047bbaff885ad56be7a0d2612af1de26765f6cfc37da25e6
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:3daa460ed347e6eaab290b0a7fb0a08da6f210fb8be88af33f6e256511af80b6
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:dcc73f518ca7557cbd84b236367b3ecb7baab405268d54543783b4b772e14568
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:f7e5e04adb0467eb0aa4691a68b0d6485db9a52f99494097bf5b598e08f2edfb
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:2a109f94282b76ca0ff929adc35c93c0a98e34b43aba248d2975957fa149133f
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:cca121312777d2727fe2c50a14583243ce69e6e356b499f133736787522095b2
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:a0af0e5d4b142f28dc72c5636df5040c4ce92c6b12ab183d3a9bc7a9d2d2f418
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:88863a8dd2145bf17f7c5378c81e7c0be7a01dc47ee75bd7e5203a8fed8895b9
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:8cf7993326ba220f650ae2fa5383d0d6cf3e2c7f3346f0a339a9d212b8df02f6
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:edf12760e7fe4c8d37c22e172c0da270c219b7ff16d4ebb6b62729545309f6fa
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:a040696505e266e8f06a5f06b108f2db5ad3a6ed803d547d736455c1928a6a0e
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:d9560ed2758d3793139d3e13b7e5ce9d0f4ac0f8a6a8767449822c033cb2a456
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:9d997314bc7c6b30c2c3ff54b27bcbc1cdeac6dba300cab8a898b97fbce05ff7
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:dbff3d1592b192253a288fb3230a0ff125564540a30a6dd5ac881e79ed93b60a
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:9ae64f17191b98bb75f613969a9415103dad304e4fd20fb444f4fc4e57a57e0e
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:14952f38ab1d7f0abe4d54883e2c93d01ec1240ae809aefa9ca8cb4aaef080e1
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:15a971cef680f6f43e6b8ad7e0f5dab572ebacebc35cfbc6f7949a4ccd7789c2
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:d41ece8df9482ff860706edcdf27b53ac97d7c87407b3acfe8b1d9adfbf40e38
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:7581cd2330c20be2b77621e0e14da183d2b94f40fa5d3f448dede5f8c182759e
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:6fbdfdf3b9206058d5efc6653194bf6c42c07655817e3c97044da678e84b310a
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:3aa57a7ee9b342844ea4aa20bbbdf7a43b5d41f7866baa619f25473a230011f3
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:91b86228495f38d5528be1a8ffec96695ad72fafb76d2ead87f7bee25540d8b6
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:2e240514425d70039f5b48fe1908c56fae6efcc3eff93c63ec67f4aa3318a5fb
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:69388282df18b68c1d937ac98a5026b34221a54e5ce4b9cf4b89379b66623488
        - registry.redhat.io/container-native-virtualization/mtq-operator-rhel9@sha256:e29ffbe332d5da66ab4612a50ed73813ccbde92622474273b617cccd3bc24cd7
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:d253f1dbb54c70cf5874c2177bcd1cc3697381916cd4cf825bda934bfdb3d53c
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:7752f41a0f380fcd0d2652d286d4aac1ebb94a8a5fbb655b2e7058054f2eca4d
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:b20230f9a74652cf676bcf05dad81fc1ff64f9cff7232782c62cf585db48b115
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:0c5e1ee8ae0f19237a2c8b93466a7fc217a112cf1fb6dd55e2be27f95c9b66a3
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:bf7562a3bbc7655565b69fc96f2deae15729bbdef158cd90b97fc5ecc58d5bd8
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:fd2b4a337a8e45061cb6f455de3eefe27b004844c724fd821efc3b47215b74d3
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:77211068f3a8d331f1b21c64918861398f2cdbb2a3b760a5e8df51c89efea5b9
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:cddcdbf1d02b6e55cd93787991b5cf8704d0fb3a4a0bef5f58c918b57e26ee16
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:387fba10dc8f16fbfbc8c84274c5f1f5f6d01cf11f17f1b54fa0775a3300c6de
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:53ae4db672daad95d99f2955545ba335095c96630efc2a2ff19c5019a9413e87
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:63262574fa7d51d3b8c26b79e22d78f462c70239746aec6b8b3b951a02919e81
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:4046e8e862e0c55f09e9031b95e6a77af4d491593e39770c644287d44888b0bb
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:42fc0d1d112b30f65d6a48fa0442e2f41872658d2431ac8618dadf0aa7a41208
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:9e87e7d8679211a317d11e588868b40235da8c668bfb2ee10b564f6850fb91d1
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:38e767770ea0a74c4d594a382eed72bf9efa527e45ac6438acb79a1039f57fbd
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:75d76d7274f5169a414d8024125d4f0d8709888ac2b5d6121360635c7967d878
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:96092abe951f495f4f7dd0bdcb371b7efcb03a78a15e3ed8f1b37502203a6316
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:08f15657a1eee02927c5a6ef661e6e703d0064e091751bd2d8fbad5f118d9672
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:7e19b42a648cee71c60cb69fbcf6ebca05c938942e16d23f1cd61a619cffdcb6
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:878e1c8bd72cb3002457b6fbcddca99c0f4b00f6231fea52387d540b12e6329f
        version: 4.16.21
      entries:
      - name: kubevirt-hyperconverged-operator.v4.16.21
        version: 4.16.21
      - name: kubevirt-hyperconverged-operator.v4.16.17
        version: 4.16.17
      - name: kubevirt-hyperconverged-operator.v4.16.13
        version: 4.16.13
      - name: kubevirt-hyperconverged-operator.v4.16.7
        version: 4.16.7
      - name: kubevirt-hyperconverged-operator.v4.16.6
        version: 4.16.6
      - name: kubevirt-hyperconverged-operator.v4.16.5
        version: 4.16.5
      - name: kubevirt-hyperconverged-operator.v4.16.4
        version: 4.16.4
      - name: kubevirt-hyperconverged-operator.v4.16.3
        version: 4.16.3
      - name: kubevirt-hyperconverged-operator.v4.16.2
        version: 4.16.2
      - name: kubevirt-hyperconverged-operator.v4.16.1
        version: 4.16.1
      - name: kubevirt-hyperconverged-operator.v4.16.0
        version: 4.16.0
      - name: kubevirt-hyperconverged-operator.v4.15.2
        version: 4.15.2
      - name: kubevirt-hyperconverged-operator.v4.15.1
        version: 4.15.1
      - name: kubevirt-hyperconverged-operator.v4.15.0
        version: 4.15.0
      - name: kubevirt-hyperconverged-operator.v4.14.3
        version: 4.14.3
      - name: kubevirt-hyperconverged-operator.v4.14.2
        version: 4.14.2
      - name: kubevirt-hyperconverged-operator.v4.14.1
        version: 4.14.1
      - name: kubevirt-hyperconverged-operator.v4.14.0
        version: 4.14.0
      - name: kubevirt-hyperconverged-operator.v4.13.4
        version: 4.13.4
      - name: kubevirt-hyperconverged-operator.v4.13.3
        version: 4.13.3
      - name: kubevirt-hyperconverged-operator.v4.13.2
        version: 4.13.2
      - name: kubevirt-hyperconverged-operator.v4.13.1
        version: 4.13.1
      - name: kubevirt-hyperconverged-operator.v4.13.0
        version: 4.13.0
      - name: kubevirt-hyperconverged-operator.v4.12.2
        version: 4.12.2
      - name: kubevirt-hyperconverged-operator.v4.12.1
        version: 4.12.1
      - name: kubevirt-hyperconverged-operator.v4.12.0
        version: 4.12.0
      name: stable
    defaultChannel: stable
    packageName: kubevirt-hyperconverged
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: http://marketplace.redhat.com
    name: marketplace-games-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marketplace-games-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "game.marketplace.redhat.com/v1",
                "kind": "Arcade",
                "metadata": {
                  "name": "arcade-sample"
                },
                "spec": {
                  "size": 1
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Arcade is the Schema for the arcades API
            displayName: Arcade Instance
            kind: Arcade
            name: arcades.game.marketplace.redhat.com
            version: v1
        description: RHM Arcade Operator - a simple operator used to showcase how
          to build an operator around an existing web application that can can run
          on Openshift, Vanilla k8s, or on-boarded to Red Hat Marketplace.
        displayName: Arcade Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - rhm
        - red hat
        - marketplace
        - games
        - arcade
        links:
        - name: Marketplace Games Operator
          url: https://github.com/redhat-marketplace/marketplace-games-operator
        maturity: alpha
        provider:
          name: Red Hat Marketplace
          url: http://marketplace.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.connect.redhat.com/rhm-labs/arcade-operator@sha256:d6959be6693728d2dc4cd38b69251551d23c5002cb52db2684df31a0a8f7b0ab
        version: 1.0.0
      entries:
      - name: marketplace-games-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: marketplace-games-operator
    provider:
      name: Red Hat Marketplace
      url: http://marketplace.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Provider Name
      provider-url: https://your.domain
    name: smilecdr-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smilecdr-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.smilecdr.com/v1alpha1",
                "kind": "Smilecdr",
                "metadata": {
                  "name": "smilecdr"
                },
                "spec": {
                  "labels": {
                    "application": "smilecdr",
                    "client": "smilecdr",
                    "env": "dev",
                    "version": "one"
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Smilecdr
            name: smilecdrs.operator.smilecdr.com
            version: v1alpha1
        description: Smilecdr Operator description. TODO.
        displayName: Smilecdr Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - smilecdr-operator
        links:
        - name: Smilecdr Operator
          url: https://smilecdr-operator.domain
        maintainers:
        - email: your@email.com
          name: Maintainer Name
        maturity: alpha
        provider:
          name: Provider Name
          url: https://your.domain
        relatedImages:
        - quay.io/jibran_shaukat/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - quay.io/jibran_shaukat/smilecdr@sha256:cd6b2cec6f5b9ad9e0f0434532f8fb00b6919fcd840e686d96ce08874f2154db
        version: 0.0.1
      entries:
      - name: smilecdr-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: smilecdr-operator
    provider:
      name: Provider Name
      url: https://your.domain
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: model-builder-for-vision-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-modelbuilder-for-vision-rhmp.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "Modelbuilder",
                "metadata": {
                  "name": "modelbuilder-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "backup_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "metadata_storage": {
                    "storage_class": "ibmc-block-bronze",
                    "storage_size": "20G"
                  },
                  "photo_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "500G"
                  },
                  "in_memory_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              },
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "ModelbuilderAWS",
                "metadata": {
                  "name": "modelbuilder-aws-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "postgres": {
                    "storage_size": "20G",
                    "storage_class": "gp2",
                    "backup_storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
          createdAt: "2021-08-28T12:00:00Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/model-builder-for-vision-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/model-builder-for-vision-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.0.3
          operators.operatorframework.io/builder: operator-sdk-v1.11.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/ibm-modelbuilder-for-vision/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Model Builder for Vision on AWS
            displayName: Deploy-Modelbuilder on AWS
            kind: ModelbuilderAWS
            name: modelbuildersaws.modelbuilder.com
            version: v1alpha1
          - description: Deploy Model Builder for Vision on IBM
            displayName: Deploy-Modelbuilder on IBM
            kind: Modelbuilder
            name: modelbuilders.modelbuilder.com
            version: v1alpha1
        description: IBM Services Software Model Builder for Vision is an AI training
          platform that can use the capabilities of IBM Cloud and Amazon Web Services
          and their GPUs to quickly train computer vision models for compatible mobile
          apps. IBM Services Software Inspection Workbench iOS and iPadOS app will
          be used as the exclusive UI for labeling and training. Specifically, IBM
          Model Builder for Vision will train computer vision models that can then
          be deployed for use on the IBM Services Software Inspector Portable , IBM
          Services Software Inspector Wearable and the IBM Maximo Visual Inspection
          Mobile apps.
        displayName: IBM Model Builder for Vision
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Model Builder
        - AI
        - Computer Vision
        - Inspection
        - Mobile
        - Core ML
        - TensorFlow Lite
        links:
        - name: IBM Model Builder for Vision
          url: http://ibm.biz/model_builder
        maintainers:
        - email: isshelp@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.connect.redhat.com/ibm/modelbuilder-augmentor@sha256:ef5d0dde3eb334d919c2773af63007f1ef5e735ae040ef77fffe3440f854f068
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/rhel8/redis-6@sha256:2f65f46b25d84a3cb41a6a10f93786bcab47b7354ca9ac13a8e602ea833eba60
        - registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
        - registry.connect.redhat.com/ibm/modelbuilder-image-server@sha256:ecdb665af651417c50335e7259b89ef8096521abec431976bd7abe4c1edc49d6
        - registry.connect.redhat.com/ibm/modelbuilder-auth-service@sha256:f96b444fdcf7781775c9eba942bcc5e5ebdcf5f50820447f5eb26061c4c674cd
        - registry.connect.redhat.com/ibm/modelbuilder-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/modelbuilder-cv-training-service@sha256:c3c9080b08b87022c784fd98ac3e2774ec89ce29a2dd19d91dee039e3fdb2b36
        version: 1.0.3
      entries:
      - name: ibm-modelbuilder-for-vision-rhmp.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: model-builder-for-vision-certified-rhmp
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: fuse-console
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-console.v7.10.3-0.1655306919.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "replicas": 1,
                  "version": "1.10",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "100Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:37dc38375f5d553ac1d9f20f66ce005ed9495354b81d7915ae3bd7dd6af132f3
          createdAt: "2020-09-14T10:59:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.10.3'
          olm.substitutesFor: fuse-console.v7.10.3-0.1652362834.p
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.10/html/managing_fuse
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:37dc38375f5d553ac1d9f20f66ce005ed9495354b81d7915ae3bd7dd6af132f3
        version: 7.10.3+0.1655306919.p
      entries:
      - name: fuse-console.v7.10.3
        version: 7.10.3
      - name: fuse-console.v7.10.3-0.1652362834.p
        version: 7.10.3+0.1652362834.p
      - name: fuse-console.v7.10.3-0.1655306919.p
        version: 7.10.3+0.1655306919.p
      name: 7.10.x
    - currentCSV: fuse-console.v7.11.1-0.1681224695.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.11",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:518b5ace9178e2aafd7b6f247fd41d9392abd1a7bccb4cb53783531820611193
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.11.1'
          olm.substitutesFor: fuse-console.v7.11.1-0.1675972857.p
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.11/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:518b5ace9178e2aafd7b6f247fd41d9392abd1a7bccb4cb53783531820611193
        version: 7.11.1+0.1681224695.p
      entries:
      - name: fuse-console.v7.11.1
        version: 7.11.1
      - name: fuse-console.v7.11.1-0.1669922044.p
        version: 7.11.1+0.1669922044.p
      - name: fuse-console.v7.11.1-0.1675972857.p
        version: 7.11.1+0.1675972857.p
      - name: fuse-console.v7.11.1-0.1681224695.p
        version: 7.11.1+0.1681224695.p
      name: 7.11.x
    - currentCSV: fuse-console.v7.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.12",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:4e6889bf76d323956e0b594db7381e8162fea6b108f53f9e6dd0b38cf3596c8d
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.12.1'
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.12/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:4e6889bf76d323956e0b594db7381e8162fea6b108f53f9e6dd0b38cf3596c8d
        version: 7.12.1
      entries:
      - name: fuse-console.v7.12.1
        version: 7.12.1
      name: 7.12.x
    - currentCSV: fuse-console.v7.13.0-0.1757604404.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.13",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:72d98254a821bce237b3dde896f2825628cb7b6df564e23949337590586e8bea
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.7.0 <7.13.0'
          olm.substitutesFor: fuse-console.v7.13.0-0.1754559584.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.13/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:72d98254a821bce237b3dde896f2825628cb7b6df564e23949337590586e8bea
        version: 7.13.0+0.1757604404.p
      entries:
      - name: fuse-console.v7.13.0-0.1744281421.p
        version: 7.13.0+0.1744281421.p
      - name: fuse-console.v7.13.0-0.1723709432.p
        version: 7.13.0+0.1723709432.p
      - name: fuse-console.v7.13.0-0.1757604404.p
        version: 7.13.0+0.1757604404.p
      - name: fuse-console.v7.13.0-0.1721224332.p
        version: 7.13.0+0.1721224332.p
      - name: fuse-console.v7.13.0-0.1744737026.p
        version: 7.13.0+0.1744737026.p
      - name: fuse-console.v7.13.0-0.1724659726.p
        version: 7.13.0+0.1724659726.p
      - name: fuse-console.v7.13.0-0.1725026290.p
        version: 7.13.0+0.1725026290.p
      - name: fuse-console.v7.13.0-0.1748960166.p
        version: 7.13.0+0.1748960166.p
      - name: fuse-console.v7.13.0-0.1738056254.p
        version: 7.13.0+0.1738056254.p
      - name: fuse-console.v7.13.0-0.1739885034.p
        version: 7.13.0+0.1739885034.p
      - name: fuse-console.v7.13.0-0.1741875974.p
        version: 7.13.0+0.1741875974.p
      - name: fuse-console.v7.13.0-0.1741884704.p
        version: 7.13.0+0.1741884704.p
      - name: fuse-console.v7.13.0-0.1719822976.p
        version: 7.13.0+0.1719822976.p
      - name: fuse-console.v7.13.0
        version: 7.13.0
      - name: fuse-console.v7.13.0-0.1731460809.p
        version: 7.13.0+0.1731460809.p
      - name: fuse-console.v7.13.0-0.1749640277.p
        version: 7.13.0+0.1749640277.p
      - name: fuse-console.v7.13.0-0.1751291353.p
        version: 7.13.0+0.1751291353.p
      - name: fuse-console.v7.13.0-0.1753702606.p
        version: 7.13.0+0.1753702606.p
      - name: fuse-console.v7.13.0-0.1753778626.p
        version: 7.13.0+0.1753778626.p
      - name: fuse-console.v7.13.0-0.1753785080.p
        version: 7.13.0+0.1753785080.p
      - name: fuse-console.v7.13.0-0.1753860418.p
        version: 7.13.0+0.1753860418.p
      - name: fuse-console.v7.13.0-0.1754303647.p
        version: 7.13.0+0.1754303647.p
      - name: fuse-console.v7.13.0-0.1754559584.p
        version: 7.13.0+0.1754559584.p
      - name: fuse-console.v7.13.0-0.1754567932.p
        version: 7.13.0+0.1754567932.p
      - name: fuse-console.v7.12.1
        version: 7.12.1
      name: 7.13.x
    - currentCSV: fuse-console.v7.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "replicas": 1,
                  "version": "1.9",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "100Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:846d57f3fd9f307a98b0169b6733c26299209d99ed189f219c48f36ba771b198
          createdAt: "2020-09-14T10:59:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.9.0'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.9/html/managing_fuse
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:846d57f3fd9f307a98b0169b6733c26299209d99ed189f219c48f36ba771b198
        version: 7.9.0
      entries:
      - name: fuse-console.v7.9.0
        version: 7.9.0
      name: 7.9.x
    defaultChannel: 7.13.x
    packageName: fuse-console
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: opentelemetry-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opentelemetry-operator.v0.135.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "Instrumentation",
                "metadata": {
                  "name": "instrumentation"
                },
                "spec": {
                  "exporter": {
                    "endpoint": "http://otel-collector-headless:4317"
                  }
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpAMPBridge",
                "metadata": {
                  "name": "opampbridge-sample"
                },
                "spec": {
                  "capabilities": {
                    "AcceptsOpAMPConnectionSettings": true,
                    "AcceptsOtherConnectionSettings": true,
                    "AcceptsRemoteConfig": true,
                    "AcceptsRestartCommand": true,
                    "ReportsEffectiveConfig": true,
                    "ReportsHealth": true,
                    "ReportsOwnLogs": true,
                    "ReportsOwnMetrics": true,
                    "ReportsOwnTraces": true,
                    "ReportsRemoteConfig": true,
                    "ReportsStatus": true
                  },
                  "componentsAllowed": {
                    "exporters": [
                      "debug"
                    ],
                    "processors": [
                      "memory_limiter"
                    ],
                    "receivers": [
                      "otlp"
                    ]
                  },
                  "endpoint": "ws://opamp-server:4320/v1/opamp"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": "receivers:\n  otlp:\n    protocols: \n      grpc: {}\n      http: {}\n\nexporters:\n  debug: {}\n\nservice:\n  pipelines:\n    traces:\n      receivers: [otlp]\n      exporters: [debug]\n"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1beta1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": {
                    "exporters": {
                      "debug": {}
                    },
                    "receivers": {
                      "otlp": {
                        "protocols": {
                          "grpc": {},
                          "http": {}
                        }
                      }
                    },
                    "service": {
                      "pipelines": {
                        "traces": {
                          "exporters": [
                            "debug"
                          ],
                          "receivers": [
                            "otlp"
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/rhosdt/opentelemetry-rhel8-operator@sha256:45c6c52744e9a17ffd883a2864cb7722ce3f9580e522c976d4f55623118f3504
          createdAt: 24 Sep 2025, 10:19
          description: Provides the OpenTelemetry components, including the Collector
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.33.0 <0.135.0-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-opentelemetry-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-telemetry/opentelemetry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instrumentation is the spec for OpenTelemetry instrumentation.
            displayName: OpenTelemetry Instrumentation
            kind: Instrumentation
            name: instrumentations.opentelemetry.io
            version: v1alpha1
          - description: OpAMPBridge is the Schema for the opampbridges API.
            displayName: OpAMP Bridge
            kind: OpAMPBridge
            name: opampbridges.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1beta1
          - description: TargetAllocator is the Schema for the targetallocators API.
            displayName: Target Allocator
            kind: TargetAllocator
            name: targetallocators.opentelemetry.io
            version: v1alpha1
        description: |-
          Red Hat build of OpenTelemetry is a collection of tools, APIs, and SDKs. You use it to instrument, generate, collect, and export telemetry data (metrics, logs, and traces) for analysis in order to understand your software's performance and behavior.
          This operator was previously called Red Hat OpenShift distributed tracing data collection.

          ### Operator features

          * **Sidecar injection** - annotate your pods and let the operator inject a sidecar.
          * **Managed upgrades** - updating the operator will automatically update your OpenTelemetry collectors.
          * **Deployment modes** - your collector can be deployed as sidecar, daemon set, or regular deployment.
          * **Service port management** - the operator detects which ports need to be exposed based on the provided configuration.

          ## Examples

          There is a list of examples to help you deploy different configurations in [this GitHub repository](https://github.com/os-observability/redhat-rhosdt-samples)

          ### Support & Troubleshooting

          Red Hat build of OpenTelemetry is available as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/4.17/html/red_hat_build_of_opentelemetry/index) is recommended when installing, configuring, and managing the Operator and instances.
        displayName: Red Hat build of OpenTelemetry
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opentelemetry
        - tracing
        - logging
        - metrics
        - monitoring
        - troubleshooting
        links:
        - name: OpenTelemetry Operator
          url: https://github.com/open-telemetry/opentelemetry-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/opentelemetry-rhel8-operator@sha256:45c6c52744e9a17ffd883a2864cb7722ce3f9580e522c976d4f55623118f3504
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5a6edfecf1636805e6f0bcad02df10bceec764f0580f3ac8fe62aba978d50006
        - registry.redhat.io/rhosdt/opentelemetry-collector-rhel8@sha256:5f31caf1f4a12cf2832ac3e84877f4fb0c856073820498c8a5be13698348c38a
        - registry.redhat.io/rhosdt/opentelemetry-target-allocator-rhel8@sha256:f77a26edafba06a4fca1a07083e06a908664852cf228b1814455ffa8155cc979
        version: 0.135.0-1
      entries:
      - name: opentelemetry-operator.v0.135.0-1
        version: 0.135.0-1
      - name: opentelemetry-operator.v0.127.0-2
        version: 0.127.0-2
      - name: opentelemetry-operator.v0.127.0-1
        version: 0.127.0-1
      - name: opentelemetry-operator.v0.119.0-2
        version: 0.119.0-2
      - name: opentelemetry-operator.v0.119.0-1
        version: 0.119.0-1
      - name: opentelemetry-operator.v0.113.0-2
        version: 0.113.0-2
      - name: opentelemetry-operator.v0.113.0-1
        version: 0.113.0-1
      - name: opentelemetry-operator.v0.107.0-5
        version: 0.107.0-5
      - name: opentelemetry-operator.v0.107.0-4-0.1727168137.p
        version: 0.107.0-4+0.1727168137.p
      - name: opentelemetry-operator.v0.107.0-4
        version: 0.107.0-4
      - name: opentelemetry-operator.v0.102.0-3
        version: 0.102.0-3
      - name: opentelemetry-operator.v0.102.0-2
        version: 0.102.0-2
      - name: opentelemetry-operator.v0.100.1-3
        version: 0.100.1-3
      - name: opentelemetry-operator.v0.93.0-3
        version: 0.93.0-3
      - name: opentelemetry-operator.v0.93.0-2
        version: 0.93.0-2
      - name: opentelemetry-operator.v0.89.0-3
        version: 0.89.0-3
      - name: opentelemetry-operator.v0.81.1-5
        version: 0.81.1-5
      - name: opentelemetry-operator.v0.74.0-5-0.1687199949.p
        version: 0.74.0-5+0.1687199949.p
      - name: opentelemetry-operator.v0.74.0-5
        version: 0.74.0-5
      - name: opentelemetry-operator.v0.60.0-2
        version: 0.60.0-2
      - name: opentelemetry-operator.v0.56.0-1
        version: 0.56.0-1
      name: stable
    defaultChannel: stable
    packageName: opentelemetry-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://www.medik8s.io/
    name: customized-user-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: customized-user-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediation",
                "metadata": {
                  "name": "customizeduserremediation-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediationConfig",
                "metadata": {
                  "name": "customizeduserremediationconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediationTemplate",
                "metadata": {
                  "name": "customizeduserremediationtemplate-sample",
                  "namespace": "openshift-workload-availability"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/customized-user-remediation-operator:v0.1.0
          createdAt: "2024-03-25 16:17:57"
          description: Customized User Remediation Operator will remediate unhealthy
            nodes by running a pre defined user script on the unhealthy node. Since
            the user has full control over the script content, is may be quite risky
            so apply discretion.
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/customized-user-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CustomizedUserRemediationConfig is the Schema for the customizeduserremediationconfigs
              API
            displayName: Customized User Remediation Config
            kind: CustomizedUserRemediationConfig
            name: customizeduserremediationconfigs.customized-user-remediation.medik8s.io
            version: v1alpha1
          - description: CustomizedUserRemediation is the Schema for the customizeduserremediations
              API
            displayName: Customized User Remediation
            kind: CustomizedUserRemediation
            name: customizeduserremediations.customized-user-remediation.medik8s.io
            version: v1alpha1
          - description: CustomizedUserRemediationTemplate is the Schema for the customizeduserremediationtemplates
              API
            displayName: Customized User Remediation Template
            kind: CustomizedUserRemediationTemplate
            name: customizeduserremediationtemplates.customized-user-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Customized User Remediation (CUR) works with Node Health Check to remediate unhealthy nodes by running a user customized script on the unhealthy node.
          This operator does not preform any fencing on the remediated node so users should be careful with the scripts that are being used.

          ### Compatibility
          CUR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Flexibility - CUR is quite flexible sine it runs in the remediation process a user customized script.
        displayName: Customized User Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - script
        - CUR
        links:
        - name: Customized User Remediation
          url: https://customized-user-remediation.domain
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Medik8s
          url: https://www.medik8s.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/medik8s/customized-user-remediation-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: customized-user-remediation.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: customized-user-remediation
    provider:
      name: Medik8s
      url: https://www.medik8s.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-loadcore-agents
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-loadcore-agents.v0.2.25
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2024-11-20T18:10:53Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <=10.0.0'
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Agents
            name: agents.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight LoadCore Agents Operator
        displayName: Keysight LoadCore Agents Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Loadcore Agents
          url: https://keysight-loadcore-agents.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:7efeeb8b29872a6f0271f651d7ae02c91daea16d853c50e374c310f044d8c76c
        - prod.harbor.keysight.digital/loadcore/keysight-loadcore-agents@sha256:96e18b6270097c23bad5445a15613e467d06c6fbdbaf39545edb79915aacc04f
        version: 0.2.25
      entries:
      - name: keysight-loadcore-agents.v0.2.25
        version: 0.2.25
      name: alpha
    defaultChannel: alpha
    packageName: keysight-loadcore-agents
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-secret-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secret-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/tools@sha256:47204b2697dfbab5eaf9d6dcb4c160dd427a148d6b840727239b773810e74fc9
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:a6d6631d35c32b874c995b6af889636ac8dd1f39dcb68810d9f9b10a4e2c1987
        - quay.io/stackable/tools@sha256:47204b2697dfbab5eaf9d6dcb4c160dd427a148d6b840727239b773810e74fc9
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:161e754437428c68490333337303fb30b695a9ac677c3a00a1dc27565b855ab8
        - quay.io/stackable/secret-operator@sha256:1267311830649f928b528512a3cdc843efad6dd32b851d99c84639fbe70e35a3
        version: 24.11.1
      entries:
      - name: secret-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: secret-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:6eb08b99259d62514d87844e1b7470382d500696fecc9a1232f6bb536ed897fa
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:dd730457133f619d8759269abcfa79d7aeb817e01ba7af2d3aa1417df406ea56
        - quay.io/stackable/secret-operator@sha256:6eb08b99259d62514d87844e1b7470382d500696fecc9a1232f6bb536ed897fa
        version: 25.3.0
      entries:
      - name: secret-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: secret-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
          - description: Trust stores managed by the operator.
            displayName: Trust stores managed by the operator.
            kind: TrustStore
            name: truststores.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
        version: 25.7.0
      entries:
      - name: secret-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: secret-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
          - description: Trust stores managed by the operator.
            displayName: Trust stores managed by the operator.
            kind: TrustStore
            name: truststores.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        version: 25.7.0
      entries:
      - name: secret-operator.v25.7.0
        version: 25.7.0
      - name: secret-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-secret-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zettaset
      provider-url: ""
    name: xcrypt-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zts-xcrypt-operator.v0.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xcrypt.zettaset.com/v1alpha1",
                "kind": "XCrypt",
                "metadata": {
                  "name": "zts-masterset"
                },
                "spec": {
                  "status": {
                    "podNames": "zts-ca",
                    "replicas": "1,"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
          createdAt: "2021-06-16T12:29:13.97Z"
          description: XCrypt Operator deploys Zettaset XCrypt Container Encryption  -
            the only software-defined solution that offers transparent high performance
            data-at-rest encryption of critical data in container environments.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/xcrypt-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/xcrypt-operator-rhmp/support?utm_source=openshift_console
          support: Zettaset
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XCrypt is the Schema for the xcrypts API
            kind: XCrypt
            name: xcrypts.xcrypt.zettaset.com
            version: v1alpha1
        description: "Zettaset’s RedHat Certified XCrypt Operator allows users to
          deploy multiple XCrypt components on OpenShift with just a few simple commands.
          The XCrypt Operator deployment provides software-defined encryption that
          transparently protects container data throughout your OpenShift environment.\n\n##
          Core Features\n* **Automated XCrypt provisioning** - Deploying a Zettaset
          XCrypt has never been easier. Adjust cluster specific configuration and
          let the Zettaset Operator take care of provisioning XCrypt services and
          setting up the XCrypt encryption to your exact specification. \n* **CSI
          Driver deployment**: Zettaset XCrypt Operator automatically deploys Zettaset
          XCrypt CSI Driver in order to provide automated and transparent data-at-rest
          \ encryption functionality compatible with Openshift and any other container
          orchestrator. Once the Operator is provisioned, the driver runs fine tuned
          and ready to work with other XCrypt components.\n* **Multi-node topology**:
          Encrypt as many nodes and volumes as you need to. XCrypt components can
          be installed on multiple nodes across the cluster. This is defined by two
          labels: zts-master and zts-worker. Normally zts-master node is a single
          node that runs Zettaset XCrypt major services required for the product operation.
          Multiple nodes can be marked as zts-worker and those are the nodes that
          can have encrypted volumes.\n* **Seamless Encrypted Volume mount**: Volumes
          are not only being encrypted, but also are mounted/unmounted to containers
          with Persistent Volume Claims. This happens as seamlessly as regular volumes
          provision and mount/unmount.\n* **Robust failover**: A node failure doesn’t
          affect XCrypt operations on other nodes of the cluster. Encryption on a
          restarted/recovered node gets back to normal functioning if the node stays
          healthy and data hasn’t been corrupted during the restart\n\n## About Zettaset
          XCrypt Kubernetes Encryption for OpenShift\nXCrypt Kubernetes Encryption
          for OpenShift is a software-only solution that enables encryption of data
          at rest stored in Red Hat OpenShift container environments without slowing
          down the DevSecOps process. In addition to providing a transparent and high-performance
          layer of security that flexibly protects container data across on-premise,
          in hybrid, and multi-cloud environments, it also helps your organization
          efficiently transition from DevOps to DevSecOps.  \n\nMore information about
          Zettaset XCrypt Kubernetes Encryption for OpenShift can be found in the
          [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)\n\n##
          RedHat OpenShift + Zettaset: The next generation of secure enterprise infrastructure
          for DevSecOps  \nBy combining RedHat’s Container Platform with Zettaset’s
          certified software-defined encryption, you can create the next generation
          of secure enterprise infrastructure that DevSecOps requires. Together they
          create the stable, consistent, and supported base that your organization
          needs to develop applications faster, simplify updates, drive innovation,
          and reduce the risk of potential security breaches and data theft.\n###
          Core Features\n* Software-only for simple and scalable deployments \n* Negligible
          performance impact\n* Transparent to developers and administrators \n* Works
          with AWS, Google, and Azure \n* Unique encryption key per container volume
          \n* Encrypted volumes are only available when in use\n* Direct integration
          with OpenShift \n* Automated encryption policy management Secure erase of
          volumes\n* Ability to securely decommission a node\n\n## How does XCrypt
          Kubernetes Encryption fit into OpenShift? \nXCrypt XCrypt Kubernetes Encryption
          for OpenShift Encryption makes it incredibly easy to protect container data
          in your OpenShift environment. \n### Use cases include:\n* Transitioning
          from DevOps to DevSecOps \n* Ensuring data protection in OpenShift container
          environments with negligible performance impact \n* Achieving compliance
          in regulated industries like healthcare and finance\n\n## Other Information\nRole-based
          access control needs to be configured prior to the Operator installation.
          Please find details in the [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)
          \   "
        displayName: Zettaset XCrypt Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - encryption
        - storage
        links:
        - name: Zettaset
          url: https://www.zettaset.com/
        maintainers:
        - email: info@zettaset.com
          name: Zettaset
        maturity: alpha
        provider:
          name: Zettaset
        relatedImages:
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-ca-1@sha256:9f1015dfb459b69f5dbf449efa0517452095a6cf4f20826ebae278f9f800ffe7
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-kmip-server-1@sha256:9af5d21dd748dfbf4e107f27a5f10fa3c0ba928ade9b989a744b620ce34ec757
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-license-server-1@sha256:bc5486892a41e70f5dc4c4e1d82fca8db2149598fb8b5950ba5da1ee2d9d245d
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-host-manager-1@sha256:9faa4b597eb74df412761dffa057b64fb69dbf9fa016414962eb2099a343e734
        - registry.marketplace.redhat.com/rhm/zts/zts-csi-driver-1@sha256:c15fa9a7cc40891103f426206663dcebe1210c8ee3ab6037cbf64e1f86b9c7e5
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
        version: 0.0.17
      entries:
      - name: zts-xcrypt-operator.v0.0.17
        version: 0.0.17
      name: alpha
    defaultChannel: alpha
    packageName: xcrypt-operator-rhmp
    provider:
      name: Zettaset
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: openshift-secondary-scheduler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secondaryscheduleroperator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "SecondaryScheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-secondary-scheduler-operator"
                },
                "spec": {
                  "managementState": "Managed",
                  "schedulerConfig": "secondary-scheduler-config"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-rhel9-operator@sha256:210cc51760c8d3a9e1e5ba2b41f575b667e0d541b995362426e5a28f3f666600
          createdAt: 2024/10/01
          description: Runs a secondary scheduler in an OpenShift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.2.0 <1.3.1'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/secondary-scheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SecondaryScheduler is the schema for the secondaryschedulers
              API
            displayName: Secondary Scheduler
            kind: SecondaryScheduler
            name: secondaryschedulers.operator.openshift.io
            version: v1
        description: |
          The Secondary Scheduler Operator provides the ability to use a customized scheduler image that was developed using the scheduler plugin framework as a secondary scheduler in OpenShift.

          ## Secondary Scheduler Configuration

          Once the Secondary Scheduler Operator is installed, you can configure the plugins to run for your secondary scheduler.

          The configuration of the secondary scheduler is defined through a config map that wraps the KubeSchedulerConfiguration YAML file in "config.yaml".

          ## Additional Parameters

          Additionally, the following parameters can be configured:

          * `schedulerConfig` - Set the config map configuration for the secondary scheduler.
          * `schedulerImage` - Set the default operand image for the secondary scheduler.
        displayName: Secondary Scheduler Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - secondary
        - scheduling
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/secondary-scheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-rhel9-operator@sha256:210cc51760c8d3a9e1e5ba2b41f575b667e0d541b995362426e5a28f3f666600
        version: 1.3.1
      entries:
      - name: secondaryscheduleroperator.v1.3.1
        version: 1.3.1
      - name: secondaryscheduleroperator.v1.3.0
        version: 1.3.0
      - name: secondaryscheduleroperator.v1.2.2
        version: 1.2.2
      - name: secondaryscheduleroperator.v1.2.1
        version: 1.2.1
      - name: secondaryscheduleroperator.v1.2.0
        version: 1.2.0
      - name: secondaryscheduleroperator.v1.1.4
        version: 1.1.4
      - name: secondaryscheduleroperator.v1.1.2
        version: 1.1.2
      name: stable
    defaultChannel: stable
    packageName: openshift-secondary-scheduler-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: costmanagement-metrics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: costmanagement-metrics-operator.4.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "true"
          containerImage: registry.redhat.io/costmanagement/costmanagement-metrics-rhel9-operator@sha256:c5bb5291052d57390b9f4bdc660c1926827072b6dcbe09d85f5b6fc375562182
          createdAt: "2025-08-18T12:10:15Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: costmanagement-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Cost Management Metrics Operator
          ## Introduction
          The `costmanagement-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Cost Management Metrics Operator (`costmanagement-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created CostManagementMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The CostManagementMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **FIPS Compliance:** Supports deployment in high-security environments when run on an OpenShift cluster with FIPS mode enabled.

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created CostManagementMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the CostManagementMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `CostManagementMetricsConfig`, creating a new `CostManagementMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `costmanagement-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: costmanagement-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the CostManagementMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the CostManagementMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon CostManagementMetricsConfig creation** (This parameter will start to appear in CostManagementMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the CostManagementMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the costmanagement-metrics-operator
          **Note** There are separate instructions for configuring the `costmanagement-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `costmanagement-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the CostManagementMetricsConfig
          Configure the costmanagement-metrics-operator by creating a `CostManagementMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `costmanagement-metrics-operator` -> `CostManagementMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the costmanagement-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `costmanagement-metrics-operator-data` with 10Gi of storage. To configure the costmanagement-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `costmanagement-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `costmanagement-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the costmanagement-metrics-operator for a restricted network
          ##### Create the CostManagementMetricsConfig
          Configure the costmanagement-metrics-operator by creating a `CostManagementMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `costmanagement-metrics-operator` -> `CostManagementMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `costmanagement-metrics-operator-data` with 10Gi of storage. To configure the costmanagement-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `costmanagement-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: costmanagement-metrics-operator
                labels:
                  app: costmanagement-metrics-operator
              spec:
                volumes:
                - name: costmanagement-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: costmanagement-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: costmanagement-metrics-operator-reports
                    mountPath: /tmp/costmanagement-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/costmanagement-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/costmanagement-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `costmanagement-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the CostManagementMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Cost Management Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Cost Management Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/costmanagement/costmanagement-metrics-rhel9-operator@sha256:c5bb5291052d57390b9f4bdc660c1926827072b6dcbe09d85f5b6fc375562182
        version: 4.1.0
      entries:
      - name: costmanagement-metrics-operator.4.1.0
        version: 4.1.0
      - name: costmanagement-metrics-operator.4.0.0
        version: 4.0.0
      - name: costmanagement-metrics-operator.3.3.2
        version: 3.3.2
      - name: costmanagement-metrics-operator.3.3.1
        version: 3.3.1
      - name: costmanagement-metrics-operator.3.3.0
        version: 3.3.0
      - name: costmanagement-metrics-operator.3.2.1
        version: 3.2.1
      - name: costmanagement-metrics-operator.3.2.0
        version: 3.2.0
      - name: costmanagement-metrics-operator.3.1.1
        version: 3.1.1
      - name: costmanagement-metrics-operator.3.1.0
        version: 3.1.0
      - name: costmanagement-metrics-operator.3.0.1
        version: 3.0.1
      - name: costmanagement-metrics-operator.3.0.0
        version: 3.0.0
      - name: costmanagement-metrics-operator.2.0.0
        version: 2.0.0
      - name: costmanagement-metrics-operator.1.2.0
        version: 1.2.0
      - name: costmanagement-metrics-operator.1.1.9
        version: 1.1.9
      - name: costmanagement-metrics-operator.1.1.8
        version: 1.1.8
      - name: costmanagement-metrics-operator.1.1.7
        version: 1.1.7
      - name: costmanagement-metrics-operator.1.1.6
        version: 1.1.6
      - name: costmanagement-metrics-operator.1.1.5
        version: 1.1.5
      - name: costmanagement-metrics-operator.1.1.4
        version: 1.1.4
      - name: costmanagement-metrics-operator.1.1.3
        version: 1.1.3
      - name: costmanagement-metrics-operator.1.1.0
        version: 1.1.0
      - name: costmanagement-metrics-operator.1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: costmanagement-metrics-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-applicationautoscaling-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-applicationautoscaling-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "applicationautoscaling.services.k8s.aws/v1alpha1",
                "kind": "ScalableTarget",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "applicationautoscaling.services.k8s.aws/v1alpha1",
                "kind": "ScalingPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/applicationautoscaling-controller:1.1.1
          createdAt: "2025-09-20T21:43:01Z"
          description: AWS Application Auto Scaling controller is a service controller
            for managing Application Auto Scaling resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ScalableTarget represents the state of an AWS applicationautoscaling
              ScalableTarget resource.
            displayName: ScalableTarget
            kind: ScalableTarget
            name: scalabletargets.applicationautoscaling.services.k8s.aws
            version: v1alpha1
          - description: ScalingPolicy represents the state of an AWS applicationautoscaling
              ScalingPolicy resource.
            displayName: ScalingPolicy
            kind: ScalingPolicy
            name: scalingpolicies.applicationautoscaling.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Application Auto Scaling resources in AWS from within your Kubernetes cluster.
          **About Amazon Application Auto Scaling**
          Application Auto Scaling is a web service for developers and system administrators who need a solution for automatically scaling their scalable resources for individual AWS services beyond Amazon EC2. Application Auto Scaling allows you to configure automatic scaling for the following resources:
          - AppStream 2.0 fleets
          - Aurora replicas
          - Amazon Comprehend document classification and entity recognizer endpoints
          - DynamoDB tables and global secondary indexes
          - Amazon Elastic Container Service (ECS) services
          - Amazon EMR clusters
          - Amazon Keyspaces (for Apache Cassandra) tables
          - Lambda function provisioned concurrency
          - Amazon Managed Streaming for Apache Kafka (MSK) broker storage
          - SageMaker endpoint variants
          - Spot Fleet requests
          - Custom resources provided by your own applications or services. For more information, see the [GitHub repository](https://github.com/aws/aws-auto-scaling-custom-resource)

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Application Auto Scaling
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - applicationautoscaling
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Application Auto Scaling User Guide
          url: https://docs.aws.amazon.com/autoscaling/application/userguide/what-is-application-auto-scaling.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: application auto scaling maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/applicationautoscaling-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-applicationautoscaling-controller.v1.1.1
        version: 1.1.1
      - name: ack-applicationautoscaling-controller.v1.1.0
        version: 1.1.0
      - name: ack-applicationautoscaling-controller.v1.0.30
        version: 1.0.30
      - name: ack-applicationautoscaling-controller.v1.0.29
        version: 1.0.29
      - name: ack-applicationautoscaling-controller.v1.0.27
        version: 1.0.27
      - name: ack-applicationautoscaling-controller.v1.0.26
        version: 1.0.26
      - name: ack-applicationautoscaling-controller.v1.0.25
        version: 1.0.25
      - name: ack-applicationautoscaling-controller.v1.0.24
        version: 1.0.24
      - name: ack-applicationautoscaling-controller.v1.0.23
        version: 1.0.23
      - name: ack-applicationautoscaling-controller.v1.0.22
        version: 1.0.22
      - name: ack-applicationautoscaling-controller.v1.0.21
        version: 1.0.21
      - name: ack-applicationautoscaling-controller.v1.0.19
        version: 1.0.19
      - name: ack-applicationautoscaling-controller.v1.0.18
        version: 1.0.18
      - name: ack-applicationautoscaling-controller.v1.0.17
        version: 1.0.17
      - name: ack-applicationautoscaling-controller.v1.0.16
        version: 1.0.16
      - name: ack-applicationautoscaling-controller.v1.0.15
        version: 1.0.15
      - name: ack-applicationautoscaling-controller.v1.0.11
        version: 1.0.11
      - name: ack-applicationautoscaling-controller.v1.0.9
        version: 1.0.9
      - name: ack-applicationautoscaling-controller.v1.0.8
        version: 1.0.8
      - name: ack-applicationautoscaling-controller.v1.0.7
        version: 1.0.7
      - name: ack-applicationautoscaling-controller.v1.0.6
        version: 1.0.6
      - name: ack-applicationautoscaling-controller.v1.0.5
        version: 1.0.5
      - name: ack-applicationautoscaling-controller.v1.0.4
        version: 1.0.4
      - name: ack-applicationautoscaling-controller.v1.0.3
        version: 1.0.3
      - name: ack-applicationautoscaling-controller.v1.0.2
        version: 1.0.2
      - name: ack-applicationautoscaling-controller.v1.0.0
        version: 1.0.0
      - name: ack-applicationautoscaling-controller.v0.2.16
        version: 0.2.16
      - name: ack-applicationautoscaling-controller.v0.2.15
        version: 0.2.15
      - name: ack-applicationautoscaling-controller.v0.2.14
        version: 0.2.14
      - name: ack-applicationautoscaling-controller.v0.2.13
        version: 0.2.13
      - name: ack-applicationautoscaling-controller.v0.2.12
        version: 0.2.12
      - name: ack-applicationautoscaling-controller.v0.2.10
        version: 0.2.10
      - name: ack-applicationautoscaling-controller.v0.2.9
        version: 0.2.9
      - name: ack-applicationautoscaling-controller.v0.2.8
        version: 0.2.8
      - name: ack-applicationautoscaling-controller.v0.2.7
        version: 0.2.7
      - name: ack-applicationautoscaling-controller.v0.2.6
        version: 0.2.6
      - name: ack-applicationautoscaling-controller.v0.2.5
        version: 0.2.5
      - name: ack-applicationautoscaling-controller.v0.2.4
        version: 0.2.4
      - name: ack-applicationautoscaling-controller.v0.2.3
        version: 0.2.3
      - name: ack-applicationautoscaling-controller.v0.2.2
        version: 0.2.2
      - name: ack-applicationautoscaling-controller.v0.2.1
        version: 0.2.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-applicationautoscaling-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-acm-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-acm-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acm.services.k8s.aws/v1alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/acm-controller:1.1.1
          createdAt: "2025-09-20T01:19:56Z"
          description: AWS ACM controller is a service controller for managing ACM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Certificate represents the state of an AWS acm Certificate
              resource.
            displayName: Certificate
            kind: Certificate
            name: certificates.acm.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage AWS Certificate Manager (ACM) resources in AWS from within your Kubernetes cluster.

          **About Amazon ACM**

          AWS Certificate Manager (ACM) handles the complexity of creating, storing, and renewing public and private SSL/TLS X.509 certificates and keys that protect your AWS websites and applications. You can provide certificates for your [integrated AWS services](https://docs.aws.amazon.com/acm/latest/userguide/acm-services.html) either by issuing them directly with ACM or by [importing](https://docs.aws.amazon.com/acm/latest/userguide/import-certificate.html) third-party certificates into the ACM management system. ACM certificates can secure singular domain names, multiple specific domain names, wildcard domains, or combinations of these. ACM wildcard certificates can protect an unlimited number of subdomains. You can also [export](https://docs.aws.amazon.com/acm/latest/userguide/export-private.html) ACM certificates signed by AWS Private CA for use anywhere in your internal PKI.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon ACM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ACM Developer Resources
          url: https://aws.amazon.com/certificate-manager/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: acm maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/acm-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-acm-controller.v1.1.1
        version: 1.1.1
      - name: ack-acm-controller.v1.1.0
        version: 1.1.0
      - name: ack-acm-controller.v1.0.13
        version: 1.0.13
      - name: ack-acm-controller.v1.0.12
        version: 1.0.12
      - name: ack-acm-controller.v1.0.11
        version: 1.0.11
      - name: ack-acm-controller.v1.0.10
        version: 1.0.10
      - name: ack-acm-controller.v1.0.9
        version: 1.0.9
      - name: ack-acm-controller.v1.0.8
        version: 1.0.8
      - name: ack-acm-controller.v1.0.7
        version: 1.0.7
      - name: ack-acm-controller.v1.0.6
        version: 1.0.6
      - name: ack-acm-controller.v1.0.5
        version: 1.0.5
      - name: ack-acm-controller.v1.0.4
        version: 1.0.4
      - name: ack-acm-controller.v1.0.1
        version: 1.0.1
      - name: ack-acm-controller.v1.0.0
        version: 1.0.0
      - name: ack-acm-controller.v0.0.20
        version: 0.0.20
      - name: ack-acm-controller.v0.0.19
        version: 0.0.19
      - name: ack-acm-controller.v0.0.18
        version: 0.0.18
      - name: ack-acm-controller.v0.0.17
        version: 0.0.17
      - name: ack-acm-controller.v0.0.16
        version: 0.0.16
      - name: ack-acm-controller.v0.0.12
        version: 0.0.12
      - name: ack-acm-controller.v0.0.10
        version: 0.0.10
      - name: ack-acm-controller.v0.0.9
        version: 0.0.9
      - name: ack-acm-controller.v0.0.6
        version: 0.0.6
      - name: ack-acm-controller.v0.0.5
        version: 0.0.5
      - name: ack-acm-controller.v0.0.4
        version: 0.0.4
      - name: ack-acm-controller.v0.0.2
        version: 0.0.2
      - name: ack-acm-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-acm-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ricardo Maraschini
      provider-url: https://github.com/ricardomaraschini/tagger
    name: tagger
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tagger.v2.1.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tagger.dev/v1beta1",
                "kind": "Tag",
                "metadata": {
                  "name": "nginx"
                },
                "spec": {
                  "from": "docker.io/library/nginx:latest",
                  "generation": 0,
                  "mirror": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          containerImage: quay.io/tagger/operator:v2.1.17
          createdAt: "2021-08-02T00:00:00Z"
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/ricardomaraschini/tagger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tag is a Container Image representation as a Custom Resource.
            displayName: Tag
            kind: Tag
            name: tags.tagger.dev
            version: v1beta1
        description: |-
          A Container Image Orchestrator. Adds mirroring and Container Image manipulation capabilities
          to a Kubernetes cluster.

          Keeping track of all Container Images in use in a Kubernetes cluster is a complicated task.
          Container Images may come from numerous different Image Registries. In some cases, controlling
          how a stable version of a given Container Image looks escapes the user's authority. To add to
          this, Container Runtimes rely on remote registries (from the cluster's point of view) when
          obtaining Container Images, potentially making the process of pulling their blobs (manifests,
          config, and layers) slower.

          The notion of indexing Container Image versions by Tags is helpful. Still, it does not provide
          users with the right confidence to always use the intended Container Image – today's "latest"
          tag might not be tomorrow's "latest" tag. In addition to that, these Image Registries allow
          access to Container Images by their Manifest content's hash (i.e., usually sha256), which gives
          users the confidence at a cost in semantics.

          When releasing a new version of an application to Push and to Deploy are split into two
          distinct steps. Both the pusher and the puller need access to the same Image Registry, adding
          complexity.  Credentials are one example of the concerns. Other factors may pop up when
          running, for instance, in an air-gapped environment, where the cluster may not reach external
          Image Registries.

          Tagger aims to overcome these caveats by providing an image management abstraction layer. For
          instance, by providing a direct mapping between a Container Image tag (e.g., "latest") and its
          correspondent Manifest content's hash, users can then refer to the Container Image by its tag
          – and be sure to use that specific version. More than that, when allied with an Internal Image
          Registry, Tagger can also automatically mirror Container Images into the cluster.

          While using Tagger, Deployments can refer to Container Images by an arbitrarily defined name,
          such as "my-app", and Tagger will make sure that they use the right Container Image through its
          internal "tag to Manifest content's hash" mapping.

          For each new "release" of a given Container Image, Tagger creates a new Generation for it,
          making it easy to roll back to previously pushed "releases" of the same Container Image in case
          of problems.

          When integrated with an Internal Registry, Tagger allows users to push or pull Images directly
          without requiring an external Image Registry. It works as a layer between the user and the
          Internal Registry. Every time a new "release" of Container Images is pushed, all Deployments
          are updated automatically. Users don't need to know about the Internal Registry existence, if
          they are logged-in to the Kubernetes cluster, they can obtain old or provide new "Generations"
          for a Container Image.

          In summary, Tagger mirrors remote Container Images into a Kubernetes cluster indexing them in
          different Generations (allowing easy navigation through these multiple Generations), provides
          an interface allowing users to pull and push images directly to the Kubernetes cluster and
          provides full integration with Kubernetes Deployments (automatic triggers new rollouts on
          Container Image changes).

          By default Tagger won't be able to mirror until you provide it with a mirror registry config.
          After install you can configure the mirror by editing the Secret mirror-registry-config in the
          operator namespace. Follow below an example of a valid mirror-registry-config secret (you gonna
          have to provide your own address, username, password, etc).

          ```yaml
          apiVersion: v1
          kind: Secret
          metadata:
            name: mirror-registry-config
          data:
            address: cmVnaXN0cnkuaW8=
            username: YWRtaW4=
            password: d2hhdCB3ZXJlIHlvdSB0aGlua2luZz8K
            token: YW4gb3B0aW9uYWwgdG9rZW4gZm9yIHRva2VuIGJhc2VkIGF1dGg=
            insecure: dHJ1ZQ==
          ```

          To make pushing an pulling images easier you should go ahead and grab the `kubectl` plugin
          available [here](https://github.com/ricardomaraschini/tagger/releases). You can download a
          full version of the README with usage examples and configuration options
          [here](https://github.com/ricardomaraschini/tagger/releases/download/v2.1.17/README.pdf).
        displayName: Tagger
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tagger
        - mirroring
        - image
        links:
        - name: Documentation
          url: https://github.com/ricardomaraschini/tagger/releases/download/v2.1.17/README.pdf
        - name: Plugin (kubectl)
          url: https://github.com/ricardomaraschini/tagger/releases/tag/v2.1.17
        maintainers:
        - email: rmarasch@redhat.com
          name: Ricardo Maraschini
        maturity: alpha
        provider:
          name: Ricardo Maraschini
          url: https://github.com/ricardomaraschini/tagger
        relatedImages:
        - quay.io/tagger/operator:v2.1.17
        version: 2.1.17
      entries:
      - name: tagger.v2.1.17
        version: 2.1.17
      name: alpha
    defaultChannel: alpha
    packageName: tagger
    provider:
      name: Ricardo Maraschini
      url: https://github.com/ricardomaraschini/tagger
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecs-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "Service",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "TaskDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecs-controller:1.1.1
          createdAt: "2025-09-19T21:02:44Z"
          description: AWS ECS controller is a service controller for managing ECS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster represents the state of an AWS ecs Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.ecs.services.k8s.aws
            version: v1alpha1
          - description: Service represents the state of an AWS ecs Service resource.
            displayName: Service
            kind: Service
            name: services.ecs.services.k8s.aws
            version: v1alpha1
          - description: TaskDefinition represents the state of an AWS ecs TaskDefinition
              resource.
            displayName: TaskDefinition
            kind: TaskDefinition
            name: taskdefinitions.ecs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Container Service (ECS) resources in AWS from within your Kubernetes cluster.

          **About Amazon ECS**

          Amazon Elastic Container Service (Amazon ECS) is a fully managed container orchestration service that helps you easily deploy, manage, and scale containerized applications. As a fully managed service, Amazon ECS comes with AWS configuration and operational best practices built-in. It's integrated with both AWS and third-party tools, such as Amazon Elastic Container Registry and Docker. This integration makes it easier for teams to focus on building the applications, not the environment. You can run and scale your container workloads across AWS Regions in the cloud, and on-premises, without the complexity of managing a control plane.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECS Developer Resources
          url: https://aws.amazon.com/ecs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecs-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ecs-controller.v1.1.1
        version: 1.1.1
      - name: ack-ecs-controller.v1.0.15
        version: 1.0.15
      - name: ack-ecs-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecs-controller.v1.0.13
        version: 1.0.13
      - name: ack-ecs-controller.v1.0.12
        version: 1.0.12
      - name: ack-ecs-controller.v1.0.11
        version: 1.0.11
      - name: ack-ecs-controller.v1.0.10
        version: 1.0.10
      - name: ack-ecs-controller.v1.0.9
        version: 1.0.9
      - name: ack-ecs-controller.v1.0.8
        version: 1.0.8
      - name: ack-ecs-controller.v1.0.7
        version: 1.0.7
      - name: ack-ecs-controller.v1.0.6
        version: 1.0.6
      - name: ack-ecs-controller.v1.0.2
        version: 1.0.2
      - name: ack-ecs-controller.v1.0.1
        version: 1.0.1
      - name: ack-ecs-controller.v1.0.0
        version: 1.0.0
      - name: ack-ecs-controller.v0.0.8
        version: 0.0.8
      - name: ack-ecs-controller.v0.0.7
        version: 0.0.7
      - name: ack-ecs-controller.v0.0.6
        version: 0.0.6
      - name: ack-ecs-controller.v0.0.5
        version: 0.0.5
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseWeb
      provider-url: ""
    name: cnfv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cnfv-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eweb.enterpriseweb.com/v1",
                "kind": "CloudNFV",
                "metadata": {
                  "name": "cloudnfv-sample"
                },
                "spec": {
                  "size": 1
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CloudNFV is the Schema for the cloudnfvs API
            displayName: Cloud NFV
            kind: CloudNFV
            name: cloudnfvs.eweb.enterpriseweb.com
            version: v1
        description: Manages a cluster of CloudNFV Runtime Modules
        displayName: CloudNFV
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - automation
        - NFV
        - controller
        links:
        - name: Cnfv Operator
          url: https://cnfv-operator.domain
        maintainers:
        - email: bill@enterpriseweb.com
          name: Bill Malyk
        maturity: alpha
        provider:
          name: EnterpriseWeb
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61a164ad711478ac603242338f7b3c55986319479878b092be128581a823598e
        - quay.io/wmalyk28/cnfv-op@sha256:e5913ad5dd20cb1b2299d2387237ef15e645c9a8c9d4f5fe0044b86a6b3dc1df
        - quay.io/wmalyk28/cloudnfv-runtime@sha256:fcde8375576c9bff602d3a16ef1ba8ae88b780166649ba6ae355737b1d35a03f
        version: 0.0.1
      entries:
      - name: cnfv-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cnfv-operator
    provider:
      name: EnterpriseWeb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzounstructured-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzounstructured-operator.v3.1.200
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzounstructured.clusters.cambridgesemantics.com/v1",
                "kind": "AnzoUnstructured",
                "metadata": {
                  "name": "au01"
                },
                "spec": {
                  "auWorker": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-w",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-worker@sha256:964633d6b16ddc517a09ac86f3e3698f5a161a72cfa50ea9ba4e301fa631e347",
                                "name": "w",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  },
                  "msLeader": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-ms",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzo-microservices-leader@sha256:0637603c00a99533d57e16009516ec73d40703ae151e631fc281f51f1f0ca93b",
                                "name": "ms",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/unstructured-operator@sha256:633a8992e35601d99c0d690dd71af314605c223b57c2038fa7090be06977b2f9
          createdAt: "2025-07-08T09:47:55Z"
          description: Kubernetes Operator for Altair Graph Studio Unstructured
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoUnstructured is the Schema for the anzounstructureds
              API
            displayName: Altair Graph Studio Unstructured Operator
            kind: AnzoUnstructured
            name: anzounstructureds.anzounstructured.clusters.cambridgesemantics.com
            version: v1
        description: |-
          The Altair Graph Studio Unstructured Operator provides the way to install and configure an anzo unstructured setup on Red Hat K8S environment.
          Currently, this is possible only through existing Anzo installation.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v3.1.2/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Studio Unstructured Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Database
        - Big Data
        links:
        - name: Altair Graph Studio Unstructured Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v3.1.2/README.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/unstructured-operator@sha256:633a8992e35601d99c0d690dd71af314605c223b57c2038fa7090be06977b2f9
        - registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-worker@sha256:964633d6b16ddc517a09ac86f3e3698f5a161a72cfa50ea9ba4e301fa631e347
        - registry.connect.redhat.com/cambridgesemantics/anzo-microservices-leader@sha256:0637603c00a99533d57e16009516ec73d40703ae151e631fc281f51f1f0ca93b
        version: 3.1.200
      entries:
      - name: anzounstructured-operator.v3.1.200
        version: 3.1.200
      - name: anzounstructured-operator.v3.1.104
        version: 3.1.104
      - name: anzounstructured-operator.v3.1.103
        version: 3.1.103
      name: stable
    defaultChannel: stable
    packageName: anzounstructured-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: klusterlet
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: klusterlet.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "Klusterlet",
                "metadata": {
                  "name": "klusterlet"
                },
                "spec": {
                  "clusterName": "cluster1",
                  "deployOption": {
                    "mode": "Singleton"
                  },
                  "externalServerURLs": [
                    {
                      "url": "https://localhost"
                    }
                  ],
                  "imagePullSpec": "quay.io/open-cluster-management/registration-operator:v0.13.1",
                  "namespace": "open-cluster-management-agent",
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/registration-operator:v0.13.1
          createdAt: "2024-04-10T15:27:49Z"
          description: Manages the installation and upgrade of the Klusterlet.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management-io/ocm
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Klusterlet defines the configurations of agents running on
              the managed cluster
            displayName: Klusterlet
            kind: Klusterlet
            name: klusterlets.operator.open-cluster-management.io
            version: v1
        description: "Note: This operator will stop updating, please use the latest
          operator in OperatorHub.io https://operatorhub.io/operator/klusterlet. \n\nThe
          Klusterlet proivdes the registation to the Hub clusters as a managed cluster.
          This operator supports the installation and upgrade of Klusterlet.\n\n##
          Prerequisites\nYou need a Hub cluster which has installed ClusterManager.\n\n##
          How to Install\n* Install Klusterlet operator by following instructions
          in top right button `Install`.\n* Optional: Create an agent namespace used
          to deploy Klusterlet. The default namespace is `open-cluster-management-agent`.\n*
          Create a secret `bootstrap-hub-kubeconfig` with the kubeconfig file of hub
          cluster in the agent namespace.\n\n  ```\n  $ kubectl create secret generic
          bootstrap-hub-kubeconfig --from-file=kubeconfig=<kubeconfig file of hub
          clsuter> -n <agent namespace>\n  ```\n\n* Create an instance of the Klusterlet
          resource to install the Klusterlet.\n\nThe pods of Klusterlet are created
          in agent namespace.\n\n```\n$ kubectl get pods -n open-cluster-management-agent\n
          \ NAME                                READY   STATUS    RESTARTS   AGE\n
          \ klusterlet-agent-55c6574876-nfxm9   1/1     Running   0          8m38s\n```\n\nThere
          is a ManagedCluster created in the Hub cluster.\n\n```\n$ kubectl get managedclusters.cluster.open-cluster-management.io\nNAME
          \                  HUB ACCEPTED   MANAGED CLUSTER URLS     JOINED   AVAILABLE
          \  AGE\ncluster1               false                                                        27s\n```\n"
        displayName: Klusterlet
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open-cluster-management
        - Klusterlet
        links:
        - name: Registration Operator
          url: https://github.com/open-cluster-management-io/ocm
        maintainers:
        - email: zyin@redhat.com
          name: Zhiwei Yin
        - email: deads@redhat.com
          name: David Eads
        - email: jqiu@redhat.com
          name: Jian Qiu
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/open-cluster-management/registration-operator:v0.13.1
        version: 0.13.1
      entries:
      - name: klusterlet.v0.13.1
        version: 0.13.1
      - name: klusterlet.v0.13.0
        version: 0.13.0
      - name: klusterlet.v0.12.0
        version: 0.12.0
      - name: klusterlet.v0.11.0
        version: 0.11.0
      - name: klusterlet.v0.10.0
        version: 0.10.0
      - name: klusterlet.v0.9.1
        version: 0.9.1
      - name: klusterlet.v0.8.0
        version: 0.8.0
      - name: klusterlet.v0.7.0
        version: 0.7.0
      - name: klusterlet.v0.6.0
        version: 0.6.0
      - name: klusterlet.v0.5.0
        version: 0.5.0
      - name: klusterlet.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: klusterlet
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-api-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-api-controller.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "example-registry-with-studio"
                },
                "spec": {
                  "app": {
                    "host": "example-registry-with-studio-app.apps.yourcluster.example"
                  },
                  "studioUi": {
                    "enabled": true,
                    "host": "example-registry-with-studio-studio-ui.apps.yourcluster.example"
                  },
                  "ui": {
                    "host": "example-registry-with-studio-ui.apps.yourcluster.example"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-operator:3.0.5
          createdAt: "2024-12-03T11:30:52Z"
          description: Deploy and manage Apicurio API Controller on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Apicurio API
              Controller.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Apicurio API Controller

          Apicurio API Controller consists of 2 components:
           - Apicurio Registry 3
           - Apicurio Studio

          Storage options: PostgreSQL, Kafka, MySQL, SQL Server, in-memory.

          ## Apicurio API Controller Operator

          Provides a quick and easy way to deploy and manage Apicurio API Controller on Kubernetes.

          **Features**
          - Supports installation and configuration of Apicurio API Controller
          - Perform a rolling upgrade of Apicurio API Controller

          ## Prerequisites

          This operator does not deploy storage for Apicurio API Controller. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio API Controller and Apicurio API Controller Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Apicurio API Controller
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - API Management
        - Apicurio Registry
        - Apicurio Studio
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-operator:3.0.5
        version: 0.0.1
      entries:
      - name: apicurio-api-controller.v0.0.1
        version: 0.0.1
      name: 0.x
    defaultChannel: 0.x
    packageName: apicurio-api-controller
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: unsupported
      operatorframework.io/arch.s390x: unsupported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: https://www.konveyor.io
    name: konveyor-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: konveyor-operator.v0.8.0-alpha.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:c1d1942b3bb3c4e99d9b5484493e27a432455608fee4548f1d352be34a8d6ded
          createdAt: "2025-08-13T23:43:34Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0-alpha.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-hub@sha256:6be8fe8307e6686b178ace42f6db5cab8f04c5e2e29e02d2edcb5dc28ee337e9
        - quay.io/konveyor/tackle2-addon-platform@sha256:12efa5b7fbf6d829879b19e2820dcf782e648059f51544ebd25251dfc67fc7ca
        - quay.io/konveyor/postgresql-15-c9s@sha256:d75c5a1a0fa686ccf701e5e55633aa73d0cbb2cd9b5d8d4d4e80d7c46d78c121
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-operator@sha256:c1d1942b3bb3c4e99d9b5484493e27a432455608fee4548f1d352be34a8d6ded
        - quay.io/konveyor/tackle2-addon-discovery@sha256:ba865a74287ff94375ed316f348ee2c1d962d8d350c8afae5dbb0aee9b64fdea
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:7a5750fb76935e55b7b59e8363f21f5f9b48b39492d7f3e58e2e6a84d07ab027
        - quay.io/konveyor/kai-solution-server@sha256:83987d12d8290e085b6b66de56d347a8bf262815a5f99796177cb9120513ee3a
        - quay.io/konveyor/java-external-provider@sha256:42a156f1d81352f5e09fa0ffe2b51a4d93a1f476755eb0bac9a1d24bd3a89631
        - quay.io/konveyor/kantra@sha256:06ae932d9c2f51035bc5a488dd61e8a31c344e72615b013419199168701ea3fb
        - quay.io/konveyor/tackle2-ui@sha256:867bf24dd1687a79f2a911c22c1059ca7bd025a307d290290265653a696b8e98
        - quay.io/konveyor/tackle-keycloak-init@sha256:07c2ca07922d273bd2a3d9156ad8af0e5bf973aeb316496ee65d7201c381c1b1
        - quay.io/konveyor/origin-oauth-proxy@sha256:2c811da843057119b246784f21dc7eec490669a8569e8295d8e914cdce555b24
        - quay.io/konveyor/generic-external-provider@sha256:d58e867b51d5e76cd3597cfb7f2478d221b966f84509dbdd44c609ff5d9169be
        version: 0.8.0-alpha.4
      entries:
      - name: konveyor-operator.v0.8.0-alpha.4
        version: 0.8.0-alpha.4
      - name: konveyor-operator.v0.8.0-alpha.2
        version: 0.8.0-alpha.2
      - name: konveyor-operator.v0.8.0-alpha.1
        version: 0.8.0-alpha.1
      - name: konveyor-operator.v0.7.0-alpha.2
        version: 0.7.0-alpha.2
      - name: konveyor-operator.v0.7.0-alpha.1
        version: 0.7.0-alpha.1
      - name: konveyor-operator.v0.6.0-alpha.2
        version: 0.6.0-alpha.2
      - name: konveyor-operator.v0.6.0-alpha.1
        version: 0.6.0-alpha.1
      - name: konveyor-operator.v0.5.0-alpha.1
        version: 0.5.0-alpha.1
      - name: konveyor-operator.v0.4.0-alpha.1
        version: 0.4.0-alpha.1
      - name: konveyor-operator.v0.3.0-alpha.6
        version: 0.3.0-alpha.6
      - name: konveyor-operator.v0.3.0-alpha.5
        version: 0.3.0-alpha.5
      - name: konveyor-operator.v0.3.0-alpha.4
        version: 0.3.0-alpha.4
      - name: konveyor-operator.v0.3.0-alpha.3
        version: 0.3.0-alpha.3
      - name: konveyor-operator.v0.3.0-alpha.2
        version: 0.3.0-alpha.2
      - name: konveyor-operator.v0.3.0-alpha.1
        version: 0.3.0-alpha.1
      - name: konveyor-operator.v0.2.0-alpha.3
        version: 0.2.0-alpha.3
      - name: konveyor-operator.v0.2.0-alpha.2
        version: 0.2.0-alpha.2
      - name: konveyor-operator.v0.2.0-alpha.1
        version: 0.2.0-alpha.1
      name: alpha
    - currentCSV: konveyor-operator.v0.8.0-beta.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:019a79dec40445b6289948e173e5bf2ebfa9b495a7bb077abd634c0667953a54
          createdAt: "2025-09-17T19:50:24Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0-beta.5'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/kai-solution-server@sha256:1e69cf8a342090a03a9dddb1b28ae3627052b6d636ffa290a11761b94fd10648
        - quay.io/konveyor/tackle2-ui@sha256:78398832e092557639531230cf6d023bb07b589f3587afeaacf8c68640abc454
        - quay.io/konveyor/tackle2-operator@sha256:019a79dec40445b6289948e173e5bf2ebfa9b495a7bb077abd634c0667953a54
        - quay.io/konveyor/tackle-keycloak-init@sha256:0b4bfc1fea756f10a0657861c4c422303a2349d8d9b04a23238a32ceaf106ac5
        - quay.io/konveyor/origin-oauth-proxy@sha256:8988f14db7dde3edcaed0541cecb0b4fd58fab9dbaffb79a6c1c4aee13b9bd73
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/kantra@sha256:ad6f5c05e5e950e33a745accd03ad933fb7bb4907f44455f88a2b5c27e6dd15e
        - quay.io/konveyor/tackle2-addon-platform@sha256:d0280ade44ba5e15210c07342608576b38e6993af2e479740675a4706cd2c875
        - quay.io/konveyor/tackle2-addon-discovery@sha256:a29a0248e222d9ac3225e9d3df603f1383fa8844856bf68ea618d303f0b79883
        - quay.io/konveyor/java-external-provider@sha256:0cbdd91cc5057889b8daa2f7611fa0c944fab951c73bdf88a3e28795e9a5fc33
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:fed49c5ff14bb232e63d2dc180d3b530da63f11bf97055fc9d86cbb9e47104a4
        - quay.io/konveyor/generic-external-provider@sha256:a6c550bad7a09a1c51e226247f11c00d8453987698832cb566582842219f1f3c
        - quay.io/konveyor/tackle2-hub@sha256:1611913558c9f1203a844b43969d7b1539dab3578b83460e8383ee6757c3c34c
        - quay.io/konveyor/postgresql-15-c9s@sha256:157a2941ae971d8f72b2f8af1d4a29deeb5548012a978fa986c1c6e52395da4d
        version: 0.8.0-beta.5
      entries:
      - name: konveyor-operator.v0.8.0-beta.5
        version: 0.8.0-beta.5
      - name: konveyor-operator.v0.8.0-beta.4
        version: 0.8.0-beta.4
      - name: konveyor-operator.v0.8.0-beta.2
        version: 0.8.0-beta.2
      - name: konveyor-operator.v0.8.0-beta.1
        version: 0.8.0-beta.1
      - name: konveyor-operator.v0.6.0-beta.2
        version: 0.6.0-beta.2
      - name: konveyor-operator.v0.6.0-beta.1
        version: 0.6.0-beta.1
      - name: konveyor-operator.v0.5.0-beta.3
        version: 0.5.0-beta.3
      - name: konveyor-operator.v0.5.0-beta.2
        version: 0.5.0-beta.2
      - name: konveyor-operator.v0.5.0-beta.1
        version: 0.5.0-beta.1
      - name: konveyor-operator.v0.3.0-beta.3
        version: 0.3.0-beta.3
      - name: konveyor-operator.v0.3.0-beta.2
        version: 0.3.0-beta.2
      - name: konveyor-operator.v0.3.0-beta.1.2
        version: 0.3.0-beta.1.2
      - name: konveyor-operator.v0.3.0-beta.1.1
        version: 0.3.0-beta.1.1
      - name: konveyor-operator.v0.3.0-beta.1
        version: 0.3.0-beta.1
      name: beta
    - currentCSV: konveyor-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "move2kube-plan",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "image": "quay.io/konveyor/tackle-addon-move2kube-plan:latest"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:9688d0763d8c7812beee715a061c3c198006448855d7b891d45833f62cf869fb
          createdAt: "2023-05-25T15:23:45Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.1.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)

        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:9688d0763d8c7812beee715a061c3c198006448855d7b891d45833f62cf869fb
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/postgresql-12-centos7@sha256:1287d947464fbf5275feb31d106d2feb856c6117f0eaec8d947f2e03bd8618c5
        - quay.io/konveyor/tackle2-addon@sha256:35be344bbd3393f5634f52aafd21e9c3964d1e0ba5011ecad1561600d476bbe9
        - quay.io/konveyor/tackle-keycloak-init@sha256:b0347e5a608352c4b7bb796b65a67d417e98ba4896db71a487fa26243214cfcf
        - quay.io/konveyor/tackle2-hub@sha256:3d15d158ffb6f1fc4ca5aaf29a12173fe344c289773902c9161056bdf69acc92
        - quay.io/konveyor/tackle2-ui@sha256:99650804e65e2b6cb15a37c00189d7f463e4d19a0f866a79e9c09c8420a4d028
        - quay.io/konveyor/tackle2-addon-windup@sha256:81f0625cb82eb24ded51c7258f229f57778820c62e3be6ec05fb2b12d2970825
        - quay.io/konveyor/tackle-pathfinder@sha256:f9c5aa49f1d27f477cdd82c21441ae0287acdcdc9b2b3fcf197884007dc5f7dd
        version: 0.1.2
      entries:
      - name: konveyor-operator.v0.1.2
        version: 0.1.2
      - name: konveyor-operator.v0.1.1
        version: 0.1.1
      - name: konveyor-operator.v0.1.0
        version: 0.1.0
      name: konveyor-0.1
    - currentCSV: konveyor-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "move2kube-plan",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "image": "quay.io/konveyor/tackle-addon-move2kube-plan:latest"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:1860bb099ac64d9a2af27692f0671b7240e4d19746fffa9f9ae2356ff2818e10
          createdAt: "2023-07-06T06:17:40Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.2.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)

        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:1860bb099ac64d9a2af27692f0671b7240e4d19746fffa9f9ae2356ff2818e10
        - quay.io/konveyor/postgresql-12-centos7@sha256:6016530ab1fce71d29421ce9a97b3a89fbbb126b87b1b079c6aadebb7ff5664c
        - quay.io/konveyor/tackle2-hub@sha256:d9c4f21f44098a2526474968b6d88094dca8b5eaeb9042716332c9cb67df47cd
        - quay.io/konveyor/tackle2-ui@sha256:d10c1e717d160c5eb2581649a8dd7cdf4922a74d92db014f72f42a48a3134341
        - quay.io/konveyor/tackle2-addon-windup@sha256:538dfdb134d53a876d126dbd16679638584a606ef2b9e8fbe69ed1f29733c03a
        - quay.io/konveyor/tackle-keycloak-init@sha256:ca6c85bdf8417d4be116c2098b8d8ad5ae8b7ab0b636c4cbe4a392aa424e2d47
        - quay.io/konveyor/tackle2-addon@sha256:11693e16363bcd98249209d63a29f0c6cfffab3a4071649cbeea2fd276bc2440
        - quay.io/konveyor/tackle-pathfinder@sha256:190269b061b2065ca8ba9357bf9b959a2f587a547a44769ae0afd1ac86b4b197
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        version: 0.2.1
      entries:
      - name: konveyor-operator.v0.2.1
        version: 0.2.1
      - name: konveyor-operator.v0.2.0
        version: 0.2.0
      name: konveyor-0.2
    - currentCSV: konveyor-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:cd2305373f50ce927393e43f5735915e46c32d6f8a76091ee2808218885bba72
          createdAt: "2024-05-02T22:12:42Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.3.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:cd2305373f50ce927393e43f5735915e46c32d6f8a76091ee2808218885bba72
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:51bda0c6577c15684583d607a4f0cebef760814102423f4b2025016093a19fc1
        - quay.io/konveyor/origin-oauth-proxy@sha256:05f6bd1216b921b1f47a3821f0e22b84ed6b141bb2ada474c12bd9fde613108c
        - quay.io/konveyor/tackle2-hub@sha256:bf8803721d06d1a39251e22892343a013d5ef0fb4c1883c95e1eb55aa683f486
        - quay.io/konveyor/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle-keycloak-init@sha256:a63702cc0cb36830a7830ed70ca344a9aacc201ba6f3c94806ca2e93160df4dd
        - quay.io/konveyor/tackle2-ui@sha256:bbf58d2ea613489f91efae450421afafb9052416b0d7971606e2729e89fd7de6
        version: 0.3.2
      entries:
      - name: konveyor-operator.v0.3.2
        version: 0.3.2
      - name: konveyor-operator.v0.3.1
        version: 0.3.1
      - name: konveyor-operator.v0.3.0
        version: 0.3.0
      name: konveyor-0.3
    - currentCSV: konveyor-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:496db8ca859608d04008ac36cf7bdc9cafd7214294d33a9c5646083c361ffcb6
          createdAt: "2024-06-05T03:08:51Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.4.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:496db8ca859608d04008ac36cf7bdc9cafd7214294d33a9c5646083c361ffcb6
        - quay.io/konveyor/tackle-keycloak-init@sha256:4c7e07b1fce05f7dca47edecc0ba0ce373cc38e48879cc7eed72510d2865a18e
        - quay.io/konveyor/tackle2-ui@sha256:cb1255fc5fa4608fc659c3455f3164664b960cc2d9a3b2208967798e3532df78
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:e84ea36d206a8c143794e3986940377e77c7618225cad594bdf9a4047d7823dc
        - quay.io/konveyor/origin-oauth-proxy@sha256:05f6bd1216b921b1f47a3821f0e22b84ed6b141bb2ada474c12bd9fde613108c
        - quay.io/konveyor/tackle2-hub@sha256:b87acd8e2ee4a7edc4369bc375f64e2b98bc479d09556439e2ee4d74b74970a9
        - quay.io/konveyor/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        version: 0.4.0
      entries:
      - name: konveyor-operator.v0.4.0
        version: 0.4.0
      name: konveyor-0.4
    - currentCSV: konveyor-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:474d739ed470701b628f1f9e971925ac7fbe0b543f3e36305d932b963b1225fd
          createdAt: "2024-08-27T15:48:02Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.5.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-hub@sha256:faf91ca68e24979642a44fa013b1fb9edfee415eee835ccef9dce2eb418e2909
        - quay.io/konveyor/java-external-provider@sha256:33fb5d24b998f01c481a0a03e3f1bf3f7a6c93f5b662885102a0c966026e3375
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle-keycloak-init@sha256:5f652be5914ae6e23537b9fe02ccc9cd638c08de5462d9f58135a405c81f5287
        - quay.io/konveyor/tackle2-ui@sha256:36f29fe01d3078f0e100b24d1ed126850c5a773f4b0ae922fd4c4487de1ac29d
        - quay.io/konveyor/generic-external-provider@sha256:334a631538d31efc8d82adae4f52303336800fa3037b5055c72400918d81955c
        - quay.io/konveyor/postgresql-15-c9s@sha256:039ce764cadd03b4cfea954700139b8ceff415a9a0970baa7f2cf60f7f929174
        - quay.io/konveyor/tackle2-operator@sha256:474d739ed470701b628f1f9e971925ac7fbe0b543f3e36305d932b963b1225fd
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:2afbe8bea09904e084e026cd0a294fbe1c7b05d0e387f6633b09c5eaf4665e02
        - quay.io/konveyor/tackle2-addon-discovery@sha256:461f960bb1d29e719fb472fe9be9cf3deca17d984687957f54546a09047b0d3c
        version: 0.5.3
      entries:
      - name: konveyor-operator.v0.5.3
        version: 0.5.3
      - name: konveyor-operator.v0.5.1
        version: 0.5.1
      - name: konveyor-operator.v0.5.0
        version: 0.5.0
      name: konveyor-0.5
    - currentCSV: konveyor-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:8ee143c473570efd27b6f93332dc272871e11c66cc20f381c3a1e6e5688abf68
          createdAt: "2025-03-05T21:14:52Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.6.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle-keycloak-init@sha256:a3741269aa15edd2f9fb7df853a03a942d16a88eec114f3e54f82ef63b82adc4
        - quay.io/konveyor/tackle2-ui@sha256:12b0eda13a63508b9c13d4f9309bbfbe19d564627cf0de672307af1a6ec993fb
        - quay.io/konveyor/tackle2-operator@sha256:8ee143c473570efd27b6f93332dc272871e11c66cc20f381c3a1e6e5688abf68
        - quay.io/konveyor/java-external-provider@sha256:15148cde9a6fedef2207fe36f0ac3bd1ecdc6e3ac76f1cbec9eb97ffcf247231
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle2-hub@sha256:52a8b9be038fbfbc4e22172a4a8c7191d2b06fd8670e5df2ec442e3ce00d8c8a
        - quay.io/konveyor/tackle2-addon-discovery@sha256:3eb6f395c451e2c9a867871b3984a25f9f8e6ddcfe388f37fbd6d46a61a70b1e
        - quay.io/konveyor/generic-external-provider@sha256:27c1731859a4a639d2750cff8d1246281f4b0466bb5b069169c3e170c3c19b30
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:0b5cbb322fbe96335522410376c2e8505a8302088c0e454f3e5e7e28ba7c96a6
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/postgresql-15-c9s@sha256:55f8014c1edafab8f9085228a0d207c4e98820c615be34b935ce98e942d388b1
        version: 0.6.1
      entries:
      - name: konveyor-operator.v0.6.1
        version: 0.6.1
      - name: konveyor-operator.v0.6.0
        version: 0.6.0
      name: konveyor-0.6
    - currentCSV: konveyor-operator.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:f5870bc26b2a07ddacf00ed5b5f7df7d1068489010a80bcc543686c6c9576fbb
          createdAt: "2025-07-10T15:16:17Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:f5870bc26b2a07ddacf00ed5b5f7df7d1068489010a80bcc543686c6c9576fbb
        - quay.io/konveyor/tackle-keycloak-init@sha256:dab4cb1229c8a729d14b4cecb4aa3b5c4d48e94c36e84d785c922fc08d120617
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:7ed52ea9b95650241b5c2c01914775e29be23f7ed75fabb6e7c71969f3f0728f
        - quay.io/konveyor/generic-external-provider@sha256:facca8861559a658710ad3aef8b325731595bf944cce1907715d6efd33887072
        - quay.io/konveyor/tackle2-hub@sha256:52f66607f3134a4165ce691c6e9ce992d93927e9591055daafdc1bfc1dd09dd9
        - quay.io/konveyor/tackle2-ui@sha256:f87b04762fd163f40c49f3bcffd473f0c3f80a643576c6a3e2c2b4a14395c09f
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/postgresql-15-c9s@sha256:41460103a0bab189cd77acb6e741d9c5a6d3506479f7e56df40eb156d9b2cadd
        - quay.io/konveyor/tackle2-addon-discovery@sha256:60951801aa539ee59cd43f3c5f42a2ca751ca70e995cbb258b7d55ebb24d8450
        - quay.io/konveyor/java-external-provider@sha256:a0ef733e9fce3845d02b5360440add85427acad697d2157ef92242dec06bd73b
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/origin-oauth-proxy@sha256:2c811da843057119b246784f21dc7eec490669a8569e8295d8e914cdce555b24
        version: 0.7.1
      entries:
      - name: konveyor-operator.v0.7.1
        version: 0.7.1
      - name: konveyor-operator.v0.7.0
        version: 0.7.0
      name: konveyor-0.7
    - currentCSV: konveyor-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:68b6672fccdf54c65cc83f4c82db023053d6b61409bea880f58b53ccfea7b1ed
          createdAt: "2025-09-29T20:41:09Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-addon-platform@sha256:d0280ade44ba5e15210c07342608576b38e6993af2e479740675a4706cd2c875
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:8c82ad880ca2a2d5cd4b08125feca16d13d014d15f4f77677306086d1067f9aa
        - quay.io/konveyor/generic-external-provider@sha256:c7b6618bfff81d73f1b455963107ffdaf19334e6552ce076c009476da3f37f21
        - quay.io/konveyor/postgresql-15-c9s@sha256:af275f35a307367dcee9a339a6903cf79e3a8a41cabf79fb7cea7015895f0b58
        - quay.io/konveyor/origin-oauth-proxy@sha256:8988f14db7dde3edcaed0541cecb0b4fd58fab9dbaffb79a6c1c4aee13b9bd73
        - quay.io/konveyor/kai-solution-server@sha256:bab6e224c50de5b44cc37ce0297a3b91255efc08c1ad0fccf14fef9e5182775f
        - quay.io/konveyor/tackle2-operator@sha256:68b6672fccdf54c65cc83f4c82db023053d6b61409bea880f58b53ccfea7b1ed
        - quay.io/konveyor/tackle-keycloak-init@sha256:c0051b8c3a3c891fe12012623aee775080ecb47af8a4e75b8be8b0b4855d9f2c
        - quay.io/konveyor/tackle2-ui@sha256:f3f7380be65ff1fb3459c37a3d4865354bab685dc836893f78221f16ba579f70
        - quay.io/konveyor/kantra@sha256:a5b5591a51143574a6ce6678b141b9cd3e14328c6fad773412dde21b442fe364
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-hub@sha256:b3e6db1ffc195143b8e5d54b60949256e154b0f8b667736512a084e015cdf2a7
        - quay.io/konveyor/tackle2-addon-discovery@sha256:96a5b37834fbed9506014bcbd909258081fd833ef9e16daa104effd0e3a3c6ea
        - quay.io/konveyor/java-external-provider@sha256:8eab08a04e98051f89352250d63523a5b591f35a6497f5d69a5a9d556a94abf9
        version: 0.8.0
      entries:
      - name: konveyor-operator.v0.8.0
        version: 0.8.0
      name: konveyor-0.8
    - currentCSV: konveyor-operator.v0.7.0-rc.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
          createdAt: "2025-05-21T17:12:48Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.0-rc.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-ui@sha256:afe9141366a23c072286517e9af5436d14fd1df775aaa2fa13c3911764515777
        - quay.io/konveyor/postgresql-15-c9s@sha256:012cd6611aa9d3f4e0a82636d6817f381c19e3b0809199216fbe7fa3396a2ef5
        - quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/generic-external-provider@sha256:de55f78939576c110eb6f084a7ca51338fedba699fbb5135f23392783a5d7b95
        - quay.io/konveyor/tackle2-addon-discovery@sha256:89d2b0f8636897eb4dc1957a74a5b5041c0f3267a747adcf2bc6ca56b17b9937
        - quay.io/konveyor/origin-oauth-proxy@sha256:6c2847fd9667938b1d1294d988038e885e43cd9fb378862dae3419ce4f21631f
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle-keycloak-init@sha256:c3124971dc31be9201de2a6a1f6b2f886adf4c96731188dd5e750798cde3cfe7
        - quay.io/konveyor/tackle2-hub@sha256:bed834eb5488eef23b0d1e8af194b0646cf74c39c197e0d57acc5aed2299d9f5
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:46150725a2adbde8dfdc2250614d437a50e64bbf0b4562dabc73715a287de67b
        - quay.io/konveyor/java-external-provider@sha256:61ab20e37916abfeeadde685336660f0e8bbf519403d1f7100d43c1e42a72fbd
        version: 0.7.0-rc.1
      entries:
      - name: konveyor-operator.v0.7.0-rc.1
        version: 0.7.0-rc.1
      - name: konveyor-operator.v0.6.0-rc.2
        version: 0.6.0-rc.2
      - name: konveyor-operator.v0.6.0-rc.1
        version: 0.6.0-rc.1
      - name: konveyor-operator.v0.5.0-rc.1
        version: 0.5.0-rc.1
      - name: konveyor-operator.v0.3.0-rc.4
        version: 0.3.0-rc.4
      - name: konveyor-operator.v0.3.0-rc.3
        version: 0.3.0-rc.3
      - name: konveyor-operator.v0.3.0-rc.2
        version: 0.3.0-rc.2
      - name: konveyor-operator.v0.3.0-rc.1
        version: 0.3.0-rc.1
      name: rc
    - currentCSV: konveyor-operator.v0.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:33cf01f19a7043bcc5bdb12a3ed6bde7a9094981aa2383554111ebec9b4ba498
          createdAt: "2025-03-26T20:25:41Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.6.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/generic-external-provider@sha256:bc1112352f1a2c3172c5e7d5e637b5c5a69bf7d98b78f494fa593e8872e5cfd1
        - quay.io/konveyor/tackle2-ui@sha256:3c568c6047a044095080b6013b643a4041deae7b45936a443d4c4d708a81b945
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:734752be111a97e4b7e0030d563c32883f419ad031b2f350cb8e7333f2f37de3
        - quay.io/konveyor/tackle2-hub@sha256:0df49087d0c9341c2120a7c3aa2ba7907c056f94e1602ffd0496ca7113c75382
        - quay.io/konveyor/postgresql-15-c9s@sha256:7957d3b272b9f4e469bdc0c514ec9e70e8fbdc34a7250b4f4c62f9bb73cbf08d
        - quay.io/konveyor/tackle2-addon-discovery@sha256:1b89266d6bdfa497ba8137e0d33ff53785ac476dd96bf4fd9a185ef0ddf513ce
        - quay.io/konveyor/java-external-provider@sha256:fe85689b57125d678df685012b93c8a567fcb9251777cd711e9b82bdb22a734c
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/tackle2-operator@sha256:33cf01f19a7043bcc5bdb12a3ed6bde7a9094981aa2383554111ebec9b4ba498
        - quay.io/konveyor/tackle-keycloak-init@sha256:e32fa48351f410361e8bc100303deeb40d8a66d7eefb0cf7b1bc16e067a9a795
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        version: 0.6.2
      entries:
      - name: konveyor-operator.v0.6.2
        version: 0.6.2
      name: release-0.6
    - currentCSV: konveyor-operator.v0.7.0-rc.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
          createdAt: "2025-05-21T17:12:48Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.0-rc.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/generic-external-provider@sha256:de55f78939576c110eb6f084a7ca51338fedba699fbb5135f23392783a5d7b95
        - quay.io/konveyor/origin-oauth-proxy@sha256:6c2847fd9667938b1d1294d988038e885e43cd9fb378862dae3419ce4f21631f
        - quay.io/konveyor/tackle2-addon-discovery@sha256:89d2b0f8636897eb4dc1957a74a5b5041c0f3267a747adcf2bc6ca56b17b9937
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-hub@sha256:bed834eb5488eef23b0d1e8af194b0646cf74c39c197e0d57acc5aed2299d9f5
        - quay.io/konveyor/postgresql-15-c9s@sha256:012cd6611aa9d3f4e0a82636d6817f381c19e3b0809199216fbe7fa3396a2ef5
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:46150725a2adbde8dfdc2250614d437a50e64bbf0b4562dabc73715a287de67b
        - quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
        - quay.io/konveyor/tackle-keycloak-init@sha256:c3124971dc31be9201de2a6a1f6b2f886adf4c96731188dd5e750798cde3cfe7
        - quay.io/konveyor/tackle2-ui@sha256:afe9141366a23c072286517e9af5436d14fd1df775aaa2fa13c3911764515777
        - quay.io/konveyor/java-external-provider@sha256:61ab20e37916abfeeadde685336660f0e8bbf519403d1f7100d43c1e42a72fbd
        version: 0.7.0-rc.1
      entries:
      - name: konveyor-operator.v0.7.0-rc.1
        version: 0.7.0-rc.1
      name: release-0.7
    defaultChannel: konveyor-0.8
    packageName: konveyor-operator
    provider:
      name: Konveyor
      url: https://www.konveyor.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: sriov-network-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sriov-network-operator.v4.13.0-202309190544
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovIBNetwork",
                "metadata": {
                  "name": "example-sriovibnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "linkState": "auto",
                  "networkNamespace": "default",
                  "resourceName": "mellanoxnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetwork",
                "metadata": {
                  "name": "example-sriovnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "intelnics",
                  "spoofChk": "on",
                  "trust": "off",
                  "vlan": 0
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodePolicy",
                "metadata": {
                  "name": "policy-1",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "deviceType": "vfio-pci",
                  "mtu": 9000,
                  "nicSelector": {
                    "pfNames": [
                      "eth1"
                    ],
                    "rootDevices": [
                      "0000:01:00.1"
                    ],
                    "vendor": "8086"
                  },
                  "nodeSelector": {
                    "feature.node.kubernetes.io/network-sriov.capable": "true"
                  },
                  "numVfs": 6,
                  "priority": 99,
                  "resourceName": "intelnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodeState",
                "metadata": {
                  "name": "minikube",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "interfaces": []
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkPoolConfig",
                "metadata": {
                  "name": "sriovnetworkpoolconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovOperatorConfig",
                "metadata": {
                  "name": "default",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "enableInjector": true,
                  "enableOperatorWebhook": true,
                  "logLevel": 2
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: quay.io/pliurh/sriov-network-operator-manager:bd1
          createdAt: "2023-08-02T09:22:34Z"
          description: An operator for configuring SR-IOV components and initializing
            SRIOV network devices in Openshift cluster.
          olm.skipRange: '>=4.3.0-0 <4.13.0-202309190544'
          operatorframework.io/suggested-namespace: openshift-sriov-network-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "cni"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/k8snetworkplumbingwg/sriov-network-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovIBNetwork is the Schema for the sriovibnetworks API
            displayName: Sriov IBNetwork
            kind: SriovIBNetwork
            name: sriovibnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodePolicy is the Schema for the sriovnetworknodepolicies
              API
            displayName: Sriov Network Node Policy
            kind: SriovNetworkNodePolicy
            name: sriovnetworknodepolicies.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodeState is the Schema for the sriovnetworknodestates
              API
            displayName: Sriov Network Node State
            kind: SriovNetworkNodeState
            name: sriovnetworknodestates.sriovnetwork.openshift.io
            version: v1
          - kind: SriovNetworkPoolConfig
            name: sriovnetworkpoolconfigs.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetwork is the Schema for the sriovnetworks API
            displayName: Sriov Network
            kind: SriovNetwork
            name: sriovnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovOperatorConfig is the Schema for the sriovoperatorconfigs
              API
            displayName: Sriov Operator Config
            kind: SriovOperatorConfig
            name: sriovoperatorconfigs.sriovnetwork.openshift.io
            version: v1
        description: |
          # SR-IOV Network Operator for Openshift

          ## Introduction
          The sriov-network-operator is generally responsible for configuring the sriov components in a openshift cluster.

          ### Supported Features
          * Initialize the SR-IOV NICs on nodes.
          * provision SR-IOV device plugin on selected node.
          * provision SR-IOV CNI plugin on selected nodes.
          * manage configuration of SR-IOV device plugin.
          * generate net-att-def CRs for SR-IOV CNI plugin.

          This operator has to run in namespace 'openshift-sriov-network-operator'. An Operator Group is also required to install this operator:

          ```
          $ oc create namespace openshift-sriov-network-operator

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: sriov-network-operators
            namespace: openshift-sriov-network-operator
          spec:
            targetNamespaces:
            - openshift-sriov-network-operator
          EOF
          ```
        displayName: SR-IOV Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - network
        - sr-iov
        - sriov
        links:
        - name: Source Code
          url: https://github.com/k8snetworkplumbingwg/sriov-network-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-sriov-network-config-daemon@sha256:a2053c5366eb89112dfbc7a9994685fc198cce6fa87fdd6d1c0a07727e6eee03
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin@sha256:82d4846f8dfb9e0d02e7ff3f532f8f2bf44c9aba63c37d5ce2efb36969e5c453
        - registry.redhat.io/openshift4/ose-sriov-network-operator@sha256:3314ee580113b985e275583585b1f2adf381c5a0885b51100b3a01872a77a555
        - registry.redhat.io/openshift4/ose-sriov-network-webhook@sha256:bfb84d00dbb6b18d596ca2720b52d922c589249d2f988679189ea0eea985e58a
        - registry.redhat.io/openshift4/sriov-cni-rhel9@sha256:a4f0ec210cd065cb7f77dd25257aadf8345f47819cc728c5b6229acb79a09bb3
        - registry.redhat.io/openshift4/ose-sriov-dp-admission-controller@sha256:8dfaa07a5775a7156d303019879e41b17833e5643794f6e1087881d2a4dd1781
        - registry.redhat.io/openshift4/ose-sriov-infiniband-cni@sha256:716aa0b9ecfddfea0a62615f424f4abeab5362899c695e9ecc9df4df05f1add5
        version: 4.13.0-202309190544
      entries:
      - name: sriov-network-operator.v4.13.0-202309190544
        version: 4.13.0-202309190544
      name: "4.13"
    - currentCSV: sriov-network-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovIBNetwork",
                "metadata": {
                  "name": "example-sriovibnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "linkState": "auto",
                  "networkNamespace": "default",
                  "resourceName": "mellanoxnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetwork",
                "metadata": {
                  "name": "example-sriovnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "intelnics",
                  "spoofChk": "on",
                  "trust": "off",
                  "vlan": 0
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodePolicy",
                "metadata": {
                  "name": "policy-1",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "deviceType": "vfio-pci",
                  "mtu": 9000,
                  "nicSelector": {
                    "pfNames": [
                      "eth1"
                    ],
                    "rootDevices": [
                      "0000:01:00.1"
                    ],
                    "vendor": "8086"
                  },
                  "nodeSelector": {
                    "feature.node.kubernetes.io/network-sriov.capable": "true"
                  },
                  "numVfs": 6,
                  "priority": 99,
                  "resourceName": "intelnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodeState",
                "metadata": {
                  "name": "minikube",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "interfaces": []
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkPoolConfig",
                "metadata": {
                  "name": "sriovnetworkpoolconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovOperatorConfig",
                "metadata": {
                  "name": "default",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "enableInjector": true,
                  "enableOperatorWebhook": true,
                  "logLevel": 2
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-sriov-network-rhel9-operator@sha256:8dcae2a77e904713fba9c5907789f9b22ffb1f941f793adc4b604798d29a44f2
          createdAt: "2024-05-07T23:50:52Z"
          description: An operator for configuring SR-IOV components and initializing
            SRIOV network devices in Openshift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.16.0-202509240123'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "sriovnetwork.openshift.io/v1",
              "kind": "SriovOperatorConfig",
              "metadata": {
                "name": "default"
              },
              "spec": {
                "configurationMode": "daemon",
                "disableDrain": false,
                "enableInjector": true,
                "enableOperatorWebhook": true,
                "logLevel": 2
              }
            }
          operatorframework.io/suggested-namespace: openshift-sriov-network-operator
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-sriov-network-operator",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["sriovnetworknodestates.sriovnetwork.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/k8snetworkplumbingwg/sriov-network-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovIBNetwork is the Schema for the sriovibnetworks API
            displayName: Sriov IBNetwork
            kind: SriovIBNetwork
            name: sriovibnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodePolicy is the Schema for the sriovnetworknodepolicies
              API
            displayName: Sriov Network Node Policy
            kind: SriovNetworkNodePolicy
            name: sriovnetworknodepolicies.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodeState is the Schema for the sriovnetworknodestates
              API
            displayName: Sriov Network Node State
            kind: SriovNetworkNodeState
            name: sriovnetworknodestates.sriovnetwork.openshift.io
            version: v1
          - kind: SriovNetworkPoolConfig
            name: sriovnetworkpoolconfigs.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetwork is the Schema for the sriovnetworks API
            displayName: Sriov Network
            kind: SriovNetwork
            name: sriovnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovOperatorConfig is the Schema for the sriovoperatorconfigs
              API
            displayName: Sriov Operator Config
            kind: SriovOperatorConfig
            name: sriovoperatorconfigs.sriovnetwork.openshift.io
            version: v1
        description: |
          # SR-IOV Network Operator for Openshift

          ## Introduction
          The sriov-network-operator is generally responsible for configuring the sriov components in a openshift cluster.

          ### Supported Features
          * Initialize the SR-IOV NICs on nodes.
          * provision SR-IOV device plugin on selected node.
          * provision SR-IOV CNI plugin on selected nodes.
          * manage configuration of SR-IOV device plugin.
          * generate net-att-def CRs for SR-IOV CNI plugin.

          This operator has to run in namespace 'openshift-sriov-network-operator'. An Operator Group is also required to install this operator:

          ```
          $ oc create namespace openshift-sriov-network-operator

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: sriov-network-operators
            namespace: openshift-sriov-network-operator
          spec:
            targetNamespaces:
            - openshift-sriov-network-operator
          EOF
          ```
        displayName: SR-IOV Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - network
        - sr-iov
        - sriov
        links:
        - name: Source Code
          url: https://github.com/k8snetworkplumbingwg/sriov-network-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-sriov-network-config-daemon-rhel9@sha256:61afbb271a08dc3ab3b5509df535d9456f9b2c1b267554386b4bf859b6bccbe4
        - registry.redhat.io/openshift4/ose-sriov-infiniband-cni-rhel9@sha256:f828e415db73bf2df3fa7b26238a2f5c8a53cda6b396c11a0337d688108e6647
        - registry.redhat.io/openshift4/sriov-cni-rhel9@sha256:cac67ac25a0d319b5ea68cb3f63eac058339f1981811c7a3466b9daa2534a38d
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin-rhel9@sha256:5b3bc02e31eecd0d3468db729b03239d518ef73112b0fdb5ec3f05a788baa0be
        - registry.redhat.io/openshift4/ose-sriov-network-rhel9-operator@sha256:8dcae2a77e904713fba9c5907789f9b22ffb1f941f793adc4b604798d29a44f2
        - registry.redhat.io/openshift4/ose-sriov-network-webhook-rhel9@sha256:a6a5cc76d88b6b7aa86ee24b882ce43be33c8c6d4bcfae5f1b5490ead71ce8b6
        - registry.redhat.io/openshift4/ose-sriov-network-metrics-exporter-rhel9@sha256:a0580d8156de0ca1e5a119441196e9babbeb4d7edc502659c7ac66a49b1928d5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-sriov-dp-admission-controller-rhel9@sha256:20fd8a0081ffdea05796c8f6a0539a0e246da24f739f7bbf0722c90acfdeed15
        version: 4.16.0-202509240123
      entries:
      - name: sriov-network-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: sriov-network-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: sriov-network-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: sriov-network-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: sriov-network-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: sriov-network-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: sriov-network-operator.v4.16.0-202504290837
        version: 4.16.0-202504290837
      - name: sriov-network-operator.v4.16.0-202504151638
        version: 4.16.0-202504151638
      - name: sriov-network-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: sriov-network-operator.v4.16.0-202502262005
        version: 4.16.0-202502262005
      - name: sriov-network-operator.v4.16.0-202502171404
        version: 4.16.0-202502171404
      - name: sriov-network-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: sriov-network-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: sriov-network-operator.v4.16.0-202501081506
        version: 4.16.0-202501081506
      - name: sriov-network-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: sriov-network-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: sriov-network-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: sriov-network-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: sriov-network-operator.v4.16.0-202410300306
        version: 4.16.0-202410300306
      - name: sriov-network-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: sriov-network-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: sriov-network-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: sriov-network-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: sriov-network-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: sriov-network-operator.v4.16.0-202409111535
        version: 4.16.0-202409111535
      - name: sriov-network-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: sriov-network-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: sriov-network-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: sriov-network-operator.v4.16.0-202408120337
        version: 4.16.0-202408120337
      - name: sriov-network-operator.v4.16.0-202407100107
        version: 4.16.0-202407100107
      - name: sriov-network-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: sriov-network-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: sriov-network-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: sriov-network-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: sriov-network-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: The cert-manager maintainers
      provider-url: https://cert-manager.io/
    name: cert-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-manager.v1.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "example-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "example-clusterissuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example-certificate"
                },
                "spec": {
                  "dnsNames": [
                    "example.com"
                  ],
                  "issuerRef": {
                    "name": "example-issuer"
                  },
                  "secretName": "example-certificate-tls"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "name": "example-certificaterequest"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "name": "example-order"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "example-challenge"
                },
                "spec": {}
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: quay.io/jetstack/cert-manager-controller:v1.16.5
          createdAt: 2025-05-02T09:11:31
          olm.skipRange: '>=1.16.0 <1.16.5'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/internal-objects: |-
            [
              "challenges.acme.cert-manager.io",
              "orders.acme.cert-manager.io"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/cert-manager/cert-manager
          support: The cert-manager maintainers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A CertificateRequest is used to request a signed certificate from one of the
              configured issuers.

              All fields within the CertificateRequest's `spec` are immutable after creation.
              A CertificateRequest will either succeed or fail, as denoted by its `Ready` status
              condition and its `status.failureTime` field.

              A CertificateRequest is a one-shot resource, meaning it represents a single
              point in time request for a certificate and cannot be re-used.
            displayName: CertificateRequest
            kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - description: |-
              A Certificate resource should be created to ensure an up to date and signed
              X.509 certificate is stored in the Kubernetes Secret resource named in `spec.secretName`.

              The stored certificate will be renewed before it expires (as configured by `spec.renewBefore`).
            displayName: Certificate
            kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: Challenge is a type to represent a Challenge request with
              an ACME server
            displayName: (Internal) Challenge
            kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - description: |-
              A ClusterIssuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is similar to an Issuer, however it is cluster-scoped and therefore can
              be referenced by resources that exist in *any* namespace, not just the same
              namespace as the referent.
            displayName: ClusterIssuer
            kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - description: |-
              An Issuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is scoped to a single namespace and can therefore only be referenced by
              resources within the same namespace.
            displayName: Issuer
            kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - description: Order is a type to represent an Order with an ACME server
            displayName: (Internal) Order
            kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          cert-manager is a Kubernetes add-on to automate the management and issuance of TLS certificates from various issuing sources.
          It will ensure certificates are valid and up to date periodically, and attempt to renew certificates at an appropriate time before expiry.

          cert-manager has the following [built in issuers][]:

          * ACME: sign certificates using Let's Encrypt (or any other [ACME][] compliant signer).
          * Vault: sign certificates using HashiCorp Vault.
          * Venafi: sign certificates using Venafi Cloud or Venafi Trust Protection Platform.
          * SelfSigned: the private key of the certificate will be used to sign the certificate.
          * CA: a private key stored as a Kubernetes Secret will be used to sign the certificate.

          [built in issuers]: https://cert-manager.io/docs/configuration/#supported-issuer-types
          [ACME]: https://datatracker.ietf.org/doc/html/rfc8555

          cert-manager also supports [external issuers][],
          which are out-of-tree components that provide support for other certificate authorities, such as:
          [Google Cloud Certificate Authority](https://github.com/jetstack/google-cas-issuer),
          [AWS Private Certificate Authority](https://github.com/cert-manager/aws-privateca-issuer), and
          [Cloudflare Origin CA](https://github.com/cloudflare/origin-ca-issuer)

          [external issuers]: https://cert-manager.io/docs/configuration/external/

          In order to configure cert-manager to issue certificates, you must first create an Issuer or ClusterIssuer.
          These resources represent a particular signing authority and tell cert-manager how to handle your certificate requests.
          The [Issuer Documentation][] has more detail.

          [configure cert-manager]: https://cert-manager.io/docs/configuration/
          [Issuer Documentation]: https://cert-manager.io/docs/concepts/issuer/

          Once an Issuer has been configured, you can create a Certificate resource,
          which is a definition of a certificate request that is to be honored by an issuer
          and which is to be kept up-to-date.
          This is the usual way that you will interact with cert-manager to request signed certificates.
          The [Certificate Documentation][] has more detail.

          [Certificate Documentation]: https://cert-manager.io/docs/usage/certificate/
        displayName: cert-manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Documentation
          url: https://cert-manager.io/
        - name: Release Notes
          url: https://cert-manager.io/docs/release-notes
        maintainers:
        - email: cert-manager-maintainers@googlegroups.com
          name: The cert-manager maintainers
        maturity: stable
        minKubeVersion: 1.19.0-0
        provider:
          name: The cert-manager maintainers
          url: https://cert-manager.io/
        relatedImages:
        - quay.io/jetstack/cert-manager-controller:v1.16.5
        - quay.io/jetstack/cert-manager-cainjector:v1.16.5
        - quay.io/jetstack/cert-manager-webhook:v1.16.5
        version: 1.16.5
      entries:
      - name: cert-manager.v1.16.5
        version: 1.16.5
      - name: cert-manager.v1.16.1
        version: 1.16.1
      - name: cert-manager.v1.16.0-beta.0
        version: 1.16.0-beta.0
      - name: cert-manager.v1.15.2
        version: 1.15.2
      - name: cert-manager.v1.15.0
        version: 1.15.0
      - name: cert-manager.v1.14.2
        version: 1.14.2
      - name: cert-manager.v1.14.2-rc1
        version: 1.14.2-rc1
      - name: cert-manager.v1.14.1-rc1
        version: 1.14.1-rc1
      - name: cert-manager.v1.13.3
        version: 1.13.3
      - name: cert-manager.v1.13.3-rc1
        version: 1.13.3-rc1
      - name: cert-manager.v1.13.1
        version: 1.13.1
      - name: cert-manager.v1.13.1-rc1
        version: 1.13.1-rc1
      - name: cert-manager.v1.12.2
        version: 1.12.2
      - name: cert-manager.v1.12.2-rc1
        version: 1.12.2-rc1
      - name: cert-manager.v1.11.4
        version: 1.11.4
      - name: cert-manager.v1.11.4-rc1
        version: 1.11.4-rc1
      - name: cert-manager.v1.11.3-rc1
        version: 1.11.3-rc1
      - name: cert-manager.v1.11.0
        version: 1.11.0
      - name: cert-manager.v1.11.0-rc1
        version: 1.11.0-rc1
      - name: cert-manager.v1.10.2
        version: 1.10.2
      - name: cert-manager.v1.10.2-rc1
        version: 1.10.2-rc1
      - name: cert-manager.v1.10.1
        version: 1.10.1
      - name: cert-manager.v1.10.1-rc1
        version: 1.10.1-rc1
      - name: cert-manager.v1.10.0
        version: 1.10.0
      - name: cert-manager.v1.10.0-rc1
        version: 1.10.0-rc1
      - name: cert-manager.v1.9.1
        version: 1.9.1
      - name: cert-manager.v1.9.1-rc1
        version: 1.9.1-rc1
      - name: cert-manager.v1.8.2
        version: 1.8.2
      - name: cert-manager.v1.8.2-rc1
        version: 1.8.2-rc1
      - name: cert-manager.v1.8.0
        version: 1.8.0
      - name: cert-manager.v1.8.0-rc1
        version: 1.8.0-rc1
      - name: cert-manager.v1.7.2
        version: 1.7.2
      - name: cert-manager.v1.7.1
        version: 1.7.1
      - name: cert-manager.v1.7.1-rc1
        version: 1.7.1-rc1
      - name: cert-manager.v1.6.2
        version: 1.6.2
      - name: cert-manager.v1.6.2-rc1
        version: 1.6.2-rc1
      name: candidate
    - currentCSV: cert-manager.v1.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "example-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "example-clusterissuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example-certificate"
                },
                "spec": {
                  "dnsNames": [
                    "example.com"
                  ],
                  "issuerRef": {
                    "name": "example-issuer"
                  },
                  "secretName": "example-certificate-tls"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "name": "example-certificaterequest"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "name": "example-order"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "example-challenge"
                },
                "spec": {}
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: quay.io/jetstack/cert-manager-controller:v1.16.5
          createdAt: 2025-05-02T09:11:31
          olm.skipRange: '>=1.16.0 <1.16.5'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/internal-objects: |-
            [
              "challenges.acme.cert-manager.io",
              "orders.acme.cert-manager.io"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/cert-manager/cert-manager
          support: The cert-manager maintainers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A CertificateRequest is used to request a signed certificate from one of the
              configured issuers.

              All fields within the CertificateRequest's `spec` are immutable after creation.
              A CertificateRequest will either succeed or fail, as denoted by its `Ready` status
              condition and its `status.failureTime` field.

              A CertificateRequest is a one-shot resource, meaning it represents a single
              point in time request for a certificate and cannot be re-used.
            displayName: CertificateRequest
            kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - description: |-
              A Certificate resource should be created to ensure an up to date and signed
              X.509 certificate is stored in the Kubernetes Secret resource named in `spec.secretName`.

              The stored certificate will be renewed before it expires (as configured by `spec.renewBefore`).
            displayName: Certificate
            kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: Challenge is a type to represent a Challenge request with
              an ACME server
            displayName: (Internal) Challenge
            kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - description: |-
              A ClusterIssuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is similar to an Issuer, however it is cluster-scoped and therefore can
              be referenced by resources that exist in *any* namespace, not just the same
              namespace as the referent.
            displayName: ClusterIssuer
            kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - description: |-
              An Issuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is scoped to a single namespace and can therefore only be referenced by
              resources within the same namespace.
            displayName: Issuer
            kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - description: Order is a type to represent an Order with an ACME server
            displayName: (Internal) Order
            kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          cert-manager is a Kubernetes add-on to automate the management and issuance of TLS certificates from various issuing sources.
          It will ensure certificates are valid and up to date periodically, and attempt to renew certificates at an appropriate time before expiry.

          cert-manager has the following [built in issuers][]:

          * ACME: sign certificates using Let's Encrypt (or any other [ACME][] compliant signer).
          * Vault: sign certificates using HashiCorp Vault.
          * Venafi: sign certificates using Venafi Cloud or Venafi Trust Protection Platform.
          * SelfSigned: the private key of the certificate will be used to sign the certificate.
          * CA: a private key stored as a Kubernetes Secret will be used to sign the certificate.

          [built in issuers]: https://cert-manager.io/docs/configuration/#supported-issuer-types
          [ACME]: https://datatracker.ietf.org/doc/html/rfc8555

          cert-manager also supports [external issuers][],
          which are out-of-tree components that provide support for other certificate authorities, such as:
          [Google Cloud Certificate Authority](https://github.com/jetstack/google-cas-issuer),
          [AWS Private Certificate Authority](https://github.com/cert-manager/aws-privateca-issuer), and
          [Cloudflare Origin CA](https://github.com/cloudflare/origin-ca-issuer)

          [external issuers]: https://cert-manager.io/docs/configuration/external/

          In order to configure cert-manager to issue certificates, you must first create an Issuer or ClusterIssuer.
          These resources represent a particular signing authority and tell cert-manager how to handle your certificate requests.
          The [Issuer Documentation][] has more detail.

          [configure cert-manager]: https://cert-manager.io/docs/configuration/
          [Issuer Documentation]: https://cert-manager.io/docs/concepts/issuer/

          Once an Issuer has been configured, you can create a Certificate resource,
          which is a definition of a certificate request that is to be honored by an issuer
          and which is to be kept up-to-date.
          This is the usual way that you will interact with cert-manager to request signed certificates.
          The [Certificate Documentation][] has more detail.

          [Certificate Documentation]: https://cert-manager.io/docs/usage/certificate/
        displayName: cert-manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Documentation
          url: https://cert-manager.io/
        - name: Release Notes
          url: https://cert-manager.io/docs/release-notes
        maintainers:
        - email: cert-manager-maintainers@googlegroups.com
          name: The cert-manager maintainers
        maturity: stable
        minKubeVersion: 1.19.0-0
        provider:
          name: The cert-manager maintainers
          url: https://cert-manager.io/
        relatedImages:
        - quay.io/jetstack/cert-manager-controller:v1.16.5
        - quay.io/jetstack/cert-manager-cainjector:v1.16.5
        - quay.io/jetstack/cert-manager-webhook:v1.16.5
        version: 1.16.5
      entries:
      - name: cert-manager.v1.16.5
        version: 1.16.5
      - name: cert-manager.v1.16.1
        version: 1.16.1
      - name: cert-manager.v1.15.2
        version: 1.15.2
      - name: cert-manager.v1.15.0
        version: 1.15.0
      - name: cert-manager.v1.14.2
        version: 1.14.2
      - name: cert-manager.v1.13.3
        version: 1.13.3
      - name: cert-manager.v1.13.1
        version: 1.13.1
      - name: cert-manager.v1.13.1-rc1
        version: 1.13.1-rc1
      - name: cert-manager.v1.12.2
        version: 1.12.2
      - name: cert-manager.v1.11.4
        version: 1.11.4
      - name: cert-manager.v1.11.0
        version: 1.11.0
      - name: cert-manager.v1.10.2
        version: 1.10.2
      - name: cert-manager.v1.10.1
        version: 1.10.1
      - name: cert-manager.v1.10.0
        version: 1.10.0
      - name: cert-manager.v1.9.1
        version: 1.9.1
      - name: cert-manager.v1.8.2
        version: 1.8.2
      - name: cert-manager.v1.8.0
        version: 1.8.0
      - name: cert-manager.v1.7.2
        version: 1.7.2
      - name: cert-manager.v1.7.1
        version: 1.7.1
      - name: cert-manager.v1.6.2
        version: 1.6.2
      - name: cert-manager.v1.6.1
        version: 1.6.1
      - name: cert-manager.v1.6.0
        version: 1.6.0
      - name: cert-manager.v1.5.4
        version: 1.5.4
      - name: cert-manager.v1.5.3
        version: 1.5.3
      - name: cert-manager.v1.4.4
        version: 1.4.4
      - name: cert-manager.v1.4.3
        version: 1.4.3
      - name: cert-manager.v1.4.2
        version: 1.4.2
      - name: cert-manager.v1.4.1
        version: 1.4.1
      - name: cert-manager.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: cert-manager
    provider:
      name: The cert-manager maintainers
      url: https://cert-manager.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.12
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-06-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:9aed98a061b2c76f3c512c1d6453df5920f0c779b33d7168c1e7ab9d8bc928a4
        version: 0.0.12
      entries:
      - name: joget-dx8-openshift-operator.v0.0.12
        version: 0.0.12
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Dell Technologies
      provider-url: https://github.com/dell/csm-operator
    name: dell-csm-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dell-csm-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "authorization",
                  "namespace": "authorization"
                },
                "spec": {
                  "modules": [
                    {
                      "components": [
                        {
                          "enabled": true,
                          "name": "nginx"
                        },
                        {
                          "enabled": true,
                          "name": "cert-manager"
                        },
                        {
                          "authorizationController": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231",
                          "authorizationControllerReplicas": 1,
                          "certificate": "",
                          "controllerReconcileInterval": "5m",
                          "enabled": true,
                          "hostname": "csm-authorization.com",
                          "leaderElection": true,
                          "name": "proxy-server",
                          "opa": "docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095",
                          "opaKubeMgmt": "docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517",
                          "openTelemetryCollectorAddress": "",
                          "privateKey": "",
                          "proxyServerIngress": [
                            {
                              "annotations": {},
                              "hosts": [],
                              "ingressClassName": "nginx"
                            }
                          ],
                          "proxyService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682",
                          "proxyServiceReplicas": 1,
                          "roleService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c",
                          "roleServiceReplicas": 1,
                          "storageService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5",
                          "storageServiceReplicas": 1,
                          "tenantService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358",
                          "tenantServiceReplicas": 1
                        },
                        {
                          "commander": "docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a",
                          "name": "redis",
                          "redis": "docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f",
                          "redisCommander": "rediscommander",
                          "redisName": "redis-csm",
                          "redisReplicas": 5,
                          "redisSecretProviderClass": [
                            {
                              "redisPasswordKey": "",
                              "redisSecretName": "",
                              "redisUsernameKey": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "sentinel": "sentinel"
                        },
                        {
                          "configSecretProviderClass": [
                            {
                              "configSecretName": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "name": "config"
                        },
                        {
                          "name": "storage-system-credentials",
                          "secretProviderClasses": {
                            "vault": [
                              "secret-provider-class-1",
                              "secret-provider-class-2"
                            ]
                          }
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": true,
                      "forceRemoveModule": true,
                      "name": "authorization-proxy-server"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "isilon",
                  "namespace": "isilon"
                },
                "spec": {
                  "driver": {
                    "authSecret": "isilon-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VERBOSE",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_ISI_PORT",
                          "value": "8080"
                        },
                        {
                          "name": "X_CSI_ISI_PATH",
                          "value": "/ifs/data/csi"
                        },
                        {
                          "name": "X_CSI_ISI_NO_PROBE_ON_START",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_AUTOPROBE",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_AUTH_TYPE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_CUSTOM_TOPOLOGY_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOISILON_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "AZ_RECONCILE_INTERVAL",
                          "value": "1h"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_ISI_QUOTA_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_ACCESS_ZONE",
                          "value": "System"
                        },
                        {
                          "name": "X_CSI_ISI_VOLUME_PATH_PERMISSIONS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_IGNORE_UNRESOLVABLE_HOSTS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "isilon",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerscale"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSCALE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_PERFORMANCE_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSCALE_QUOTA_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "ISICLIENT_INSECURE",
                              "value": "true"
                            },
                            {
                              "name": "ISICLIENT_AUTH_TYPE",
                              "value": "1"
                            },
                            {
                              "name": "ISICLIENT_VERBOSE",
                              "value": "0"
                            },
                            {
                              "name": "POWERSCALE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSCALE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0",
                          "name": "metrics-powerscale"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-isilon/csi_sock",
                            "--mode=node",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powermax",
                  "namespace": "powermax"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powermax-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_MANAGED_ARRAYS",
                          "value": "000000000001,000000000002"
                        },
                        {
                          "name": "X_CSI_POWERMAX_PORTGROUPS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_TRANSPORT_PROTOCOL",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_K8S_CLUSTER_PREFIX",
                          "value": "CSM"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "X_CSI_VSPHERE_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VSPHERE_PORTGROUP",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VSPHERE_HOSTNAME",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VCENTER_HOST",
                          "value": ""
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "X_CSI_POWERMAX_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "X_CSI_POWERMAX_DEBUG",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_REVPROXY_USE_SECRET",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powermax",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERMAX_ISCSI_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_TOPOLOGY_CONTROL_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": [
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/memory-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/disk-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/network-unavailable",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REVPROXY_TLS_SECRET",
                              "value": "csirevproxy-tls-secret"
                            },
                            {
                              "name": "X_CSI_REVPROXY_PORT",
                              "value": "2222"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            },
                            {
                              "name": "DeployAsSidecar",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265",
                          "name": "csipowermax-reverseproxy"
                        }
                      ],
                      "configVersion": "v2.14.0",
                      "name": "csireverseproxy"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powermax"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERMAX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERMAX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414",
                          "name": "metrics-powermax"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/powermax.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powerstore",
                  "namespace": "powerstore"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powerstore-config",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_NODE_NAME_PREFIX",
                          "value": "csi-node"
                        },
                        {
                          "name": "X_CSI_FC_PORTS_FILTER_FILE_PATH",
                          "value": "/etc/fc-ports-filter"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOPOWERSTORE_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSM_AUTH_ENABLED",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_MULTI_NAS_FAILURE_THRESHOLD",
                          "value": "5"
                        },
                        {
                          "name": "X_CSI_MULTI_NAS_COOLDOWN_PERIOD",
                          "value": "5m"
                        },
                        {
                          "name": "X_CSI_NFS_ACLS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerstore",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-powerstore.dellemc.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerstore"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSTORE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_SPACE_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_ARRAY_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_FILE_SYSTEM_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSTORE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSTORE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "ZIPKIN_URI",
                              "value": ""
                            },
                            {
                              "name": "ZIPKIN_SERVICE_NAME",
                              "value": "metrics-powerstore"
                            },
                            {
                              "name": "ZIPKIN_PROBABILITY",
                              "value": "0.0"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60",
                          "name": "metrics-powerstore"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "unity",
                  "namespace": "unity"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_UNITY_ALLOW_MULTI_POD_ACCESS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_EPHEMERAL_STAGING_PATH",
                          "value": "/var/lib/kubelet/plugins/kubernetes.io/csi/pv/"
                        },
                        {
                          "name": "X_CSI_ISCSI_CHROOT",
                          "value": "/noderoot"
                        },
                        {
                          "name": "X_CSI_UNITY_SYNC_NODEINFO_INTERVAL",
                          "value": "15"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOUNITY_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOUNITY_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "TENANT_NAME",
                          "value": ""
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_UNITY_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "unity",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "vxflexos",
                  "namespace": "vxflexos"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLELISTVOLUMESNAPSHOT",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLESNAPSHOTCGDELETE",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "INFO"
                        },
                        {
                          "name": "GOSCALEIO_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOSCALEIO_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_QUOTA_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "INTERFACE_NAMES",
                          "value": null
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERFLEX_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "File",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerflex",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "initContainers": [
                      {
                        "envs": [
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          },
                          {
                            "name": "REPO_ADDRESS",
                            "value": "sftp://0.0.0.0"
                          },
                          {
                            "name": "REPO_USER",
                            "value": "sdcRepoUser"
                          },
                          {
                            "name": "MODULE_SIGCHECK",
                            "value": "0"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "imagePullPolicy": "IfNotPresent",
                        "name": "sdc"
                      }
                    ],
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_SDC_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_APPROVE_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_PREFIX",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_SDC_SFTP_REPO_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 1,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "enabled": false,
                        "envs": [
                          {
                            "name": "HOST_PID",
                            "value": "1"
                          },
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "name": "sdc-monitor"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "csi-external-health-monitor-controller"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERFLEX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_SDC_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_SDC_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERFLEX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d",
                          "name": "metrics-powerflex"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerflex"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--csisock=unix:/var/lib/kubelet/plugins/vxflexos.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
          createdAt: "2025-09-05T13:18:42Z"
          description: Easily install and manage Dell’s CSI Drivers and CSM
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/dell/csm-operator
          support: Dell Technologies
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ContainerStorageModule is the Schema for the containerstoragemodules
              API
            displayName: Container Storage Module
            kind: ContainerStorageModule
            name: containerstoragemodules.storage.dell.com
            version: v1
        description: "Dell Container Storage Modules (CSM) Operator is a Kubernetes
          Operator which can be \nused to install and manage Dell’s CSI drivers and
          CSM modules. \nBy using Dell CSM Operator, enterprises can quickly and easily
          deploy the \nCSM modules for CSI drivers making it easy for DevOps \nteams
          to build and optimize their Kubernetes workloads using enterprise grade
          storage \nwhile consistently aligning with the desired outcomes of the organization.\nThis
          operator is available as both community and certified operator for Openshift
          \nand as a community operator for upstream Kubernetes.\nOperator can be
          deployed via OperatorHub.io. using Operator Lifecycle Manager (OLM)\n\n###
          About Dell CSM\n\nDell CSM brings powerful enterprise storage features and
          functionality to Kubernetes for \neasier adoption of cloud native workloads,
          improved productivity and scalable operations. \n\nCSM delivers software
          modules for storage management that provide developers access to \nbuild
          automation for enhanced IT needs and other critical enterprise storage features.
          \n\nCurrently, the following modules are available as part of Dell CSM.
          \n*\tObservability - Delivers a high-level view of the storage capacity
          \n  and performance usage via Grafana dashboards to the Kubernetes users\n*\tReplication
          - Enables array replication capabilities for Kubernetes users\n*\tAuthorization
          - Provides both storage and Kubernetes administrators the ability \n  to
          apply RBAC for Dell CSI Drivers\n*\tResiliency - Designed to make Kubernetes
          applications more resilient to \n  node failures\n\nThese modules are available
          for various storage platforms in the Dell portfolio, \nsuch as Dell PowerStore,
          Dell PowerMax, Dell PowerFlex, Dell PowerScale, and Dell Unity XT.\n\nFor
          more information, visit [Dell’s Container Storage Module web page](https://www.dell.com/en-us/dt/solutions/kubernetes-containers/kubernetes-storage.htm)\n\nAdditional
          Materials\n*\t[Documentation](https://dell.github.io/csm-docs)\n*\t[CSM
          Solution Brief](https://www.delltechnologies.com/asset/en-us/products/storage/briefs-summaries/csm-solution-brief.pdf)\n"
        displayName: Dell Container Storage Modules
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dell Container Storage Modules
        - Dell CSI Driver
        - Dell CSM Modules
        - Powerflex
        - Powerscale
        - Powerstore
        - Unity
        - Authorization
        - Observability
        - Replication
        links:
        - name: Documentation
          url: https://dell.github.io/csm-docs
        maintainers:
        - email: container.storage.modules@dell.com
          name: Dell Container Storage Modules
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Dell Technologies
          url: https://github.com/dell/csm-operator
        relatedImages:
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0
        - registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa
        - docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265
        - registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea
        - docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682
        - registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334
        - docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f
        - quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:59335acaca515fceed5f0350c46cfcfd6037475f0cf35ae8ab41d37333039e2e
        - ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e
        - quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540
        - registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf
        - registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16
        - registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe
        version: 1.10.0
      entries:
      - name: dell-csm-operator-certified.v1.10.0
        version: 1.10.0
      - name: dell-csm-operator-certified.v1.9.1
        version: 1.9.1
      - name: dell-csm-operator-certified.v1.9.0
        version: 1.9.0
      - name: dell-csm-operator-certified.v1.8.1
        version: 1.8.1
      - name: dell-csm-operator-certified.v1.8.0
        version: 1.8.0
      - name: dell-csm-operator-certified.v1.7.0
        version: 1.7.0
      - name: dell-csm-operator-certified.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: dell-csm-operator-certified
    provider:
      name: Dell Technologies
      url: https://github.com/dell/csm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: external-dns-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: external-dns-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "provider": {
                    "type": "AWS",
                    "aws": {
                      "credentials": {
                        "name": "aws-access-key"
                      }
                    }
                  },
                  "source": {
                    "type": "Service"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:6be1403e418f9b63ec8ebcc8858e602409bc41fdee4586acc627623a78b50eb5
          createdAt: 2022-03-01T21:00:46
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## Prerequisites and Requirements
           The ExternalDNS Operator has minimal cluster level permissions. Unfortunately OLM doesn't allow the local role creation in the operand namespace yet.
           Therefore make sure the following commands are run against your cluster before the installation of the operator:
           * Create the operand namespace: `oc create ns external-dns`
           * Grant the operator access to manage operand resources: `oc apply -f https://raw.githubusercontent.com/openshift/external-dns-operator/release-0.1/config/rbac/extra-roles.yaml`
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b6fde43491c2e0733c8fe6ee8660f7f6435fe20c0b6d435cfc7f86eeab8b4748
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:6be1403e418f9b63ec8ebcc8858e602409bc41fdee4586acc627623a78b50eb5
        - registry.redhat.io/edo/external-dns-rhel8@sha256:c1134bb46172997ef7278b6cefbb0da44e72a9f808a7cd67b3c65d464754cab9
        version: 0.1.2
      entries:
      - name: external-dns-operator.v0.1.2
        version: 0.1.2
      name: alpha
    - currentCSV: external-dns-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
          createdAt: 2023-11-15T09:24:04
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e33874a51971a90917cf437fc6cbeea1da569ba07217f66f1bba96d594f58aed
        - registry.redhat.io/edo/external-dns-rhel8@sha256:6263a78b2a170e56dfca2173f14dbf14f0a5bddb6ea600bfdef248c7d5f20e4b
        version: 1.2.0
      entries:
      - name: external-dns-operator.v1.2.0
        version: 1.2.0
      name: stable-v1
    - currentCSV: external-dns-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:0877eb0a111efe0bff63173a2dfeb7a06a61afd0efbcdba053322a539e0dc667
          createdAt: 2024-01-22T09:26:37
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/edo/external-dns-rhel8@sha256:17f4127cbb9a9304769e9655c19cbea2260107c7e9371a99a9aa3e85d759c00b
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:0877eb0a111efe0bff63173a2dfeb7a06a61afd0efbcdba053322a539e0dc667
        version: 1.0.1
      entries:
      - name: external-dns-operator.v1.0.1
        version: 1.0.1
      - name: external-dns-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: external-dns-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:9ca751e84b8843cc62dc60550e5c4af741bf09a381ef62220e7eb3760ffe8524
          createdAt: 2024-01-22T09:56:14
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:9ca751e84b8843cc62dc60550e5c4af741bf09a381ef62220e7eb3760ffe8524
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/edo/external-dns-rhel8@sha256:ffcc3f2e9b9feb889e7882465f7f034aba6b9ee0699a1b6c24e83f501a63ecca
        version: 1.1.1
      entries:
      - name: external-dns-operator.v1.1.1
        version: 1.1.1
      - name: external-dns-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: external-dns-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
          createdAt: 2023-11-15T09:24:04
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e33874a51971a90917cf437fc6cbeea1da569ba07217f66f1bba96d594f58aed
        - registry.redhat.io/edo/external-dns-rhel8@sha256:6263a78b2a170e56dfca2173f14dbf14f0a5bddb6ea600bfdef248c7d5f20e4b
        version: 1.2.0
      entries:
      - name: external-dns-operator.v1.2.0
        version: 1.2.0
      name: stable-v1.2
    defaultChannel: stable-v1
    packageName: external-dns-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: rook-ceph-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rook-ceph-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephCluster",
                "metadata": {
                  "name": "my-rook-ceph",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "cephVersion": {
                    "image": "registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d"
                  },
                  "dataDirHostPath": "/var/lib/rook",
                  "mon": {
                    "count": 3
                  },
                  "dashboard": {
                    "enabled": true
                  },
                  "network": {
                    "hostNetwork": false
                  },
                  "rbdMirroring": {
                    "workers": 0
                  },
                  "storage": {
                    "useAllNodes": true,
                    "useAllDevices": true
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephBlockPool",
                "metadata": {
                  "name": "replicapool",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "failureDomain": "host",
                  "replicated": {
                    "size": 3
                  },
                  "annotations": null
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStore",
                "metadata": {
                  "name": "my-store",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "metadataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "dataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "gateway": {
                    "type": "s3",
                    "sslCertificateRef": null,
                    "port": 8080,
                    "securePort": null,
                    "instances": 1,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStoreUser",
                "metadata": {
                  "name": "my-user",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "store": "my-store",
                  "displayName": "my display name"
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephNFS",
                "metadata": {
                  "name": "my-nfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "rados": {
                    "pool": "myfs-data0",
                    "namespace": "nfs-ns"
                  },
                  "server": {
                    "active": 3,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephClient",
                "metadata": {
                  "name": "cinder",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "caps": {
                    "mon": "profile rbd",
                    "osd": "profile rbd pool=volumes, profile rbd pool=vms, profile rbd-read-only pool=images"
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephFilesystem",
                "metadata": {
                  "name": "myfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "dataPools": [
                    {
                      "compressionMode": "",
                      "crushRoot": "",
                      "deviceClass": "",
                      "erasureCoded": {
                        "algorithm": "",
                        "codingChunks": 0,
                        "dataChunks": 0
                      },
                      "failureDomain": "host",
                      "replicated": {
                        "requireSafeReplicaSize": false,
                        "size": 1,
                        "targetSizeRatio": 0.5
                      }
                    }
                  ],
                  "metadataPool": {
                    "compressionMode": "",
                    "crushRoot": "",
                    "deviceClass": "",
                    "erasureCoded": {
                      "algorithm": "",
                      "codingChunks": 0,
                      "dataChunks": 0
                    },
                    "failureDomain": "",
                    "replicated": {
                      "requireSafeReplicaSize": false,
                      "size": 1,
                      "targetSizeRatio": 0
                    }
                  },
                  "metadataServer": {
                    "activeCount": 1,
                    "activeStandby": true,
                    "placement": {},
                    "resources": {}
                  },
                  "preservePoolsOnDelete": false,
                  "preserveFilesystemOnDelete": false
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephRBDMirror",
                "metadata": {
                  "name": "my-rbd-mirror",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "annotations": null,
                  "count": 1,
                  "placement": {
                    "topologyKey": "kubernetes.io/hostname"
                  },
                  "resources": null
                }
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:cf7f72469d25fb37301a8159406613673f8968c82af1035fd45a470e0ae8acfa
          externalClusterScript: |-
            IiIiCkNvcHlyaWdodCAyMDIwIFRoZSBSb29rIEF1dGhvcnMuIEFsbCByaWdodHMgcmVzZXJ2ZWQu
            CgpMaWNlbnNlZCB1bmRlciB0aGUgQXBhY2hlIExpY2Vuc2UsIFZlcnNpb24gMi4wICh0aGUgIkxp
            Y2Vuc2UiKTsKeW91IG1heSBub3QgdXNlIHRoaXMgZmlsZSBleGNlcHQgaW4gY29tcGxpYW5jZSB3
            aXRoIHRoZSBMaWNlbnNlLgpZb3UgbWF5IG9idGFpbiBhIGNvcHkgb2YgdGhlIExpY2Vuc2UgYXQK
            CglodHRwOi8vd3d3LmFwYWNoZS5vcmcvbGljZW5zZXMvTElDRU5TRS0yLjAKClVubGVzcyByZXF1
            aXJlZCBieSBhcHBsaWNhYmxlIGxhdyBvciBhZ3JlZWQgdG8gaW4gd3JpdGluZywgc29mdHdhcmUK
            ZGlzdHJpYnV0ZWQgdW5kZXIgdGhlIExpY2Vuc2UgaXMgZGlzdHJpYnV0ZWQgb24gYW4gIkFTIElT
            IiBCQVNJUywKV0lUSE9VVCBXQVJSQU5USUVTIE9SIENPTkRJVElPTlMgT0YgQU5ZIEtJTkQsIGVp
            dGhlciBleHByZXNzIG9yIGltcGxpZWQuClNlZSB0aGUgTGljZW5zZSBmb3IgdGhlIHNwZWNpZmlj
            IGxhbmd1YWdlIGdvdmVybmluZyBwZXJtaXNzaW9ucyBhbmQKbGltaXRhdGlvbnMgdW5kZXIgdGhl
            IExpY2Vuc2UuCiIiIgoKaW1wb3J0IGVycm5vCmltcG9ydCBzeXMKaW1wb3J0IGpzb24KaW1wb3J0
            IGFyZ3BhcnNlCmltcG9ydCByZQppbXBvcnQgc3VicHJvY2VzcwppbXBvcnQgaG1hYwpmcm9tIGhh
            c2hsaWIgaW1wb3J0IHNoYTEgYXMgc2hhCmZyb20gb3MgaW1wb3J0IGxpbmVzZXAgYXMgTElORVNF
            UApmcm9tIG9zIGltcG9ydCBwYXRoCmZyb20gZW1haWwudXRpbHMgaW1wb3J0IGZvcm1hdGRhdGUK
            aW1wb3J0IHJlcXVlc3RzCmZyb20gcmVxdWVzdHMuYXV0aCBpbXBvcnQgQXV0aEJhc2UKCnB5M2sg
            PSBGYWxzZQppZiBzeXMudmVyc2lvbl9pbmZvLm1ham9yID49IDM6CiAgICBweTNrID0gVHJ1ZQog
            ICAgaW1wb3J0IHVybGxpYi5wYXJzZQogICAgZnJvbSBpcGFkZHJlc3MgaW1wb3J0IGlwX2FkZHJl
            c3MsIElQdjRBZGRyZXNzCgpNb2R1bGVOb3RGb3VuZEVycm9yID0gSW1wb3J0RXJyb3IKCnRyeToK
            ICAgIGltcG9ydCByYWRvcwpleGNlcHQgTW9kdWxlTm90Rm91bmRFcnJvciBhcyBub01vZEVycjoK
            ICAgIHByaW50KGYiRXJyb3I6IHtub01vZEVycn1cbkV4aXRpbmcgdGhlIHNjcmlwdC4uLiIpCiAg
            ICBzeXMuZXhpdCgxKQoKdHJ5OgogICAgaW1wb3J0IHJiZApleGNlcHQgTW9kdWxlTm90Rm91bmRF
            cnJvciBhcyBub01vZEVycjoKICAgIHByaW50KGYiRXJyb3I6IHtub01vZEVycn1cbkV4aXRpbmcg
            dGhlIHNjcmlwdC4uLiIpCiAgICBzeXMuZXhpdCgxKQoKdHJ5OgogICAgIyBmb3IgMi43LngKICAg
            IGZyb20gU3RyaW5nSU8gaW1wb3J0IFN0cmluZ0lPCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9y
            OgogICAgIyBmb3IgMy54CiAgICBmcm9tIGlvIGltcG9ydCBTdHJpbmdJTwoKdHJ5OgogICAgIyBm
            b3IgMi43LngKICAgIGZyb20gdXJscGFyc2UgaW1wb3J0IHVybHBhcnNlCiAgICBmcm9tIHVybGxp
            YiBpbXBvcnQgdXJsZW5jb2RlIGFzIHVybGVuY29kZQpleGNlcHQgTW9kdWxlTm90Rm91bmRFcnJv
            cjoKICAgICMgZm9yIDMueAogICAgZnJvbSB1cmxsaWIucGFyc2UgaW1wb3J0IHVybHBhcnNlCiAg
            ICBmcm9tIHVybGxpYi5wYXJzZSBpbXBvcnQgdXJsZW5jb2RlIGFzIHVybGVuY29kZQoKdHJ5Ogog
            ICAgZnJvbSBiYXNlNjQgaW1wb3J0IGVuY29kZXN0cmluZwpleGNlcHQ6CiAgICBmcm9tIGJhc2U2
            NCBpbXBvcnQgZW5jb2RlYnl0ZXMgYXMgZW5jb2Rlc3RyaW5nCgoKY2xhc3MgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbihFeGNlcHRpb24pOgogICAgcGFzcwoKCiMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMgRHVtbXlSYWRv
            cyAjIyMjIyMjIyMjIyMjIyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjCiMgdGhpcyBpcyBtYWlubHkgZm9yIHRlc3RpbmcgYW5kIGNvdWxkIGJlIHVz
            ZWQgd2hlcmUgJ3JhZG9zJyBpcyBub3QgYXZhaWxhYmxlCgoKY2xhc3MgRHVtbXlSYWRvcyhvYmpl
            Y3QpOgogICAgZGVmIF9faW5pdF9fKHNlbGYpOgogICAgICAgIHNlbGYucmV0dXJuX3ZhbCA9IDAK
            ICAgICAgICBzZWxmLmVycl9tZXNzYWdlID0gIiIKICAgICAgICBzZWxmLnN0YXRlID0gImNvbm5l
            Y3RlZCIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwID0ge30KICAgICAgICBzZWxmLmNtZF9u
            YW1lcyA9IHt9CiAgICAgICAgc2VsZi5faW5pdF9jbWRfb3V0cHV0X21hcCgpCiAgICAgICAgc2Vs
            Zi5kdW1teV9ob3N0X2lwX21hcCA9IHt9CgogICAgZGVmIF9pbml0X2NtZF9vdXRwdXRfbWFwKHNl
            bGYpOgogICAgICAgIGpzb25fZmlsZV9uYW1lID0gInRlc3QtZGF0YS9jZXBoLXN0YXR1cy1vdXQi
            CiAgICAgICAgc2NyaXB0X2RpciA9IHBhdGguYWJzcGF0aChwYXRoLmRpcm5hbWUoX19maWxlX18p
            KQogICAgICAgIGNlcGhfc3RhdHVzX3N0ciA9ICIiCiAgICAgICAgd2l0aCBvcGVuKAogICAgICAg
            ICAgICBwYXRoLmpvaW4oc2NyaXB0X2RpciwganNvbl9maWxlX25hbWUpLCBtb2RlPSJyIiwgZW5j
            b2Rpbmc9IlVURi04IgogICAgICAgICkgYXMganNvbl9maWxlOgogICAgICAgICAgICBjZXBoX3N0
            YXR1c19zdHIgPSBqc29uX2ZpbGUucmVhZCgpCiAgICAgICAgc2VsZi5jbWRfbmFtZXNbImZzIGxz
            Il0gPSAiIiJ7ImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJmcyBscyJ9IiIiCiAgICAgICAg
            c2VsZi5jbWRfbmFtZXNbInF1b3J1bV9zdGF0dXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3Jt
            YXQiOiAianNvbiIsICJwcmVmaXgiOiAicXVvcnVtX3N0YXR1cyJ9IiIiCiAgICAgICAgKQogICAg
            ICAgIHNlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXSA9ICgKICAgICAgICAgICAgIiIieyJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAibWdyIHNlcnZpY2VzIn0iIiIKICAgICAgICApCiAg
            ICAgICAgIyBhbGwgdGhlIGNvbW1hbmRzIGFuZCB0aGVpciBvdXRwdXQKICAgICAgICBzZWxmLmNt
            ZF9vdXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJmcyBscyJdXSA9ICgKICAgICAgICAgICAgIiIi
            W3sibmFtZSI6Im15ZnMiLCJtZXRhZGF0YV9wb29sIjoibXlmcy1tZXRhZGF0YSIsIm1ldGFkYXRh
            X3Bvb2xfaWQiOjIsImRhdGFfcG9vbF9pZHMiOlszXSwiZGF0YV9wb29scyI6WyJteWZzLXJlcGxp
            Y2F0ZWQiXX1dIiIiCiAgICAgICAgKQogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbc2VsZi5j
            bWRfbmFtZXNbInF1b3J1bV9zdGF0dXMiXV0gPSAoCiAgICAgICAgICAgICIiInsiZWxlY3Rpb25f
            ZXBvY2giOjMsInF1b3J1bSI6WzBdLCJxdW9ydW1fbmFtZXMiOlsiYSJdLCJxdW9ydW1fbGVhZGVy
            X25hbWUiOiJhIiwicXVvcnVtX2FnZSI6MTQzODUsImZlYXR1cmVzIjp7InF1b3J1bV9jb24iOiI0
            NTQwMTM4MjkyODM2Njk2MDYzIiwicXVvcnVtX21vbiI6WyJrcmFrZW4iLCJsdW1pbm91cyIsIm1p
            bWljIiwib3NkbWFwLXBydW5lIiwibmF1dGlsdXMiLCJvY3RvcHVzIl19LCJtb25tYXAiOnsiZXBv
            Y2giOjEsImZzaWQiOiJhZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5ZDBmMWMiLCJtb2Rp
            ZmllZCI6IjIwMjAtMDUtMDdUMDM6MzY6MzkuOTE4MDM1WiIsImNyZWF0ZWQiOiIyMDIwLTA1LTA3
            VDAzOjM2OjM5LjkxODAzNVoiLCJtaW5fbW9uX3JlbGVhc2UiOjE1LCJtaW5fbW9uX3JlbGVhc2Vf
            bmFtZSI6Im9jdG9wdXMiLCJmZWF0dXJlcyI6eyJwZXJzaXN0ZW50IjpbImtyYWtlbiIsImx1bWlu
            b3VzIiwibWltaWMiLCJvc2RtYXAtcHJ1bmUiLCJuYXV0aWx1cyIsIm9jdG9wdXMiXSwib3B0aW9u
            YWwiOltdfSwibW9ucyI6W3sicmFuayI6MCwibmFtZSI6ImEiLCJwdWJsaWNfYWRkcnMiOnsiYWRk
            cnZlYyI6W3sidHlwZSI6InYyIiwiYWRkciI6IjEwLjExMC4yMDUuMTc0OjMzMDAiLCJub25jZSI6
            MH0seyJ0eXBlIjoidjEiLCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OSIsIm5vbmNlIjowfV19
            LCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OS8wIiwicHVibGljX2FkZHIiOiIxMC4xMTAuMjA1
            LjE3NDo2Nzg5LzAiLCJwcmlvcml0eSI6MCwid2VpZ2h0IjowfV19fSIiIgogICAgICAgICkKICAg
            ICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXV0g
            PSAoCiAgICAgICAgICAgICIiInsiZGFzaGJvYXJkIjoiaHR0cHM6Ly9jZXBoLWRhc2hib2FyZDo4
            NDQzLyIsInByb21ldGhldXMiOiJodHRwOi8vY2VwaC1kYXNoYm9hcmQtZGI6OTI4My8ifSIiIgog
            ICAgICAgICkKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNh
            cHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMiLCAib3Nk
            IiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0
            YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250
            cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJdLCAiZW50aXR5Ijog
            ImNsaWVudC5oZWFsdGhjaGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRo
            IGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuaGVh
            bHRoY2hlY2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRB
            PT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cyIs
            Im9zZCI6InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5j
            b250cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJ9fV0iIiIKICAg
            ICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIs
            ICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAib3NkIiwgInBy
            b2ZpbGUgcmJkIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtbm9kZSIsICJmb3JtYXQiOiAi
            anNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIi
            W3siZW50aXR5IjoiY2xpZW50LmNzaS1yYmQtbm9kZSIsImtleSI6IkFRQk9nck5lSGJLMUF4QUF1
            YllCZVY4UzFVL0dQenE1U1ZlcTZnPT0iLCJjYXBzIjp7Im1vbiI6InByb2ZpbGUgcmJkLCBhbGxv
            dyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsIm9zZCI6InByb2ZpbGUgcmJkIn19XSIiIgogICAg
            ICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsibW9uIiwg
            InByb2ZpbGUgcmJkLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxs
            b3cgcnciLCAib3NkIiwgInByb2ZpbGUgcmJkIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQt
            cHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNy
            ZWF0ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktcmJkLXByb3Zp
            c2lvbmVyIiwia2V5IjoiQVFCTmdyTmUxZ2V5S3hBQThla1ZpUmRFK2hzczVPd2VZQmt3Tmc9PSIs
            ImNhcHMiOnsibWdyIjoiYWxsb3cgcnciLCJtb24iOiJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFu
            ZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJwcm9maWxlIHJiZCJ9fV0iIiIKICAgICAgICBzZWxm
            LmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyBy
            LCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3Nk
            IiwgImFsbG93IHJ3IHRhZyBjZXBoZnMgKj0qIiwgIm1kcyIsICJhbGxvdyBydyJdLCAiZW50aXR5
            IjogImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4Ijog
            ImF1dGggZ2V0LW9yLWNyZWF0ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVu
            dC5jc2ktY2VwaGZzLW5vZGUiLCJrZXkiOiJBUUJPZ3JOZUVOdW5LeEFBUENtZ0U3UjZHOERjWG5h
            SjFGMzJxZz09IiwiY2FwcyI6eyJtZHMiOiJhbGxvdyBydyIsIm1nciI6ImFsbG93IHJ3IiwibW9u
            IjoiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBy
            dyB0YWcgY2VwaGZzICo9KiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAg
            ICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kICdvc2Qg
            YmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3NkIiwgImFsbG93IHJ3IHRhZyBjZXBo
            ZnMgbWV0YWRhdGE9KiJdLCAiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVy
            IiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgog
            ICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIs
            ImtleSI6IkFRQk9nck5lQUZnY0dCQUF2R3FLT0FEMEQzeHhtVlkwUjkxMmRnPT0iLCJjYXBzIjp7
            Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2Nr
            bGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoifX1dIiIiCiAgICAg
            ICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAi
            YWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdyIiwgImFsbG93IHJ3
            IiwgIm9zZCIsICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiXSwgImVudGl0eSI6ICJj
            bGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVuc2hpZnQtc3RvcmFnZSIsICJmb3JtYXQi
            OiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0g
            IiIiW3siZW50aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0
            b3JhZ2UiLCJrZXkiOiJCUUJPZ3JOZUFGZ2NHQkFBdkdxS09BRDBEM3h4bVZZMFI5MTJkZz09Iiwi
            Y2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29z
            ZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIn19XSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsi
            bW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciIsICJh
            bGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT1teWZzIl0sICJl
            bnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0b3JhZ2Ut
            bXlmcyIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0i
            IiIKICAgICAgICBdID0gIiIiW3siZW50aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9u
            ZXItb3BlbnNoaWZ0LXN0b3JhZ2UtbXlmcyIsImtleSI6IkNRQk9nck5lQUZnY0dCQUF2R3FLT0FE
            MEQzeHhtVlkwUjkxMmRnPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cg
            ciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2Vw
            aGZzIG1ldGFkYXRhPW15ZnMifX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAg
            ICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCBxdW9y
            dW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIiwgImFsbG93IGNvbW1hbmQg
            Y29uZmlnIiwgIm9zZCIsICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFsbG93IHJ3eCBwb29sPWRl
            ZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9ZGVm
            YXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJndy5sb2csIGFsbG93IHgg
            cG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4Il0sICJlbnRpdHkiOiAiY2xpZW50LmhlYWx0
            aGNoZWNrZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0
            ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5oZWFsdGhjaGVja2VyIiwi
            a2V5IjoiQVFERmtiTmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsi
            bW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFu
            ZCB2ZXJzaW9uIiwgIm1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsICJvc2QiOiAicHJvZmls
            ZSByYmQtcmVhZC1vbmx5LCBhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhLCBhbGxvdyBy
            IHBvb2w9LnJndy5yb290LCBhbGxvdyBydyBwb29sPWRlZmF1bHQucmd3LmNvbnRyb2wsIGFsbG93
            IHJ4IHBvb2w9ZGVmYXVsdC5yZ3cubG9nLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0
            cy5pbmRleCJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAi
            IiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMs
            IGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cgY29tbWFuZCBjb25maWciLCAi
            b3NkIiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5j
            b250cm9sLCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1
            bHQucmd3LmJ1Y2tldHMuaW5kZXgiXSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIs
            ICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBjYXBzIn0iIiIKICAgICAgICBdID0g
            IiIiW3siZW50aXR5IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0NWJG
            UkFBVG5kTE5VU1JLcnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIjog
            ImFsbG93IGNvbW1hbmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFs
            bG93IHJ3eCBwb29sPWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFs
            bG93IHJ3IHBvb2w9ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJn
            dy5sb2csIGFsbG93IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4In19XSIiIgogICAg
            ICAgIHNlbGYuY21kX291dHB1dF9tYXBbIiIieyJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            bWdyIHNlcnZpY2VzIn0iIiJdID0gKAogICAgICAgICAgICAiIiJ7ImRhc2hib2FyZCI6ICJodHRw
            Oi8vcm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6NzAwMC8iLCAicHJvbWV0aGV1cyI6
            ICJodHRwOi8vcm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6OTI4My8ifSIiIgogICAg
            ICAgICkKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0
            eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            YXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJ7ImRhc2hib2FyZCI6ICJodHRwOi8vcm9vay1j
            ZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6NzAwMC8iLCAicHJvbWV0aGV1cyI6ICJodHRwOi8v
            cm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6OTI4My8ifSIiIgogICAgICAgIHNlbGYu
            Y21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNsaWVudC5oZWFsdGhj
            aGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldCJ9IiIiCiAgICAg
            ICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5oZWFsdGhjaGVja2VyIiwia2V5IjoiQVFERmti
            TmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjogImFsbG93
            IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFuZCB2ZXJzaW9uIiwg
            Im1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsICJvc2QiOiAicHJvZmlsZSByYmQtcmVhZC1v
            bmx5LCBhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhLCBhbGxvdyByIHBvb2w9LnJndy5y
            b290LCBhbGxvdyBydyBwb29sPWRlZmF1bHQucmd3LmNvbnRyb2wsIGFsbG93IHJ4IHBvb2w9ZGVm
            YXVsdC5yZ3cubG9nLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJ9fV0i
            IiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0eSI6
            ICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJh
            dXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0
            X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtbm9kZSIsICJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJb
            XSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5
            IjogImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRf
            b3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMt
            cHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIK
            ICAgICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAg
            ICAgICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVuc2hp
            ZnQtc3RvcmFnZSIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgog
            ICAgICAgIF0gPSAiIiJbXSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAg
            ICAgICIiInsiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlm
            dC1zdG9yYWdlLW15ZnMiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0i
            IiIKICAgICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAg
            ICAgICAgICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJb
            eyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsImtleSI6IkFRREZrYk5l
            ZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRBPT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIi
            LCAibWdyIjoiYWxsb3cgcnciLCAib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0q
            In19XSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2Fw
            cyI6IFsibW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1n
            ciIsICJhbGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIl0s
            ICJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpz
            b24iLCAicHJlZml4IjogImF1dGggY2FwcyJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmtiTmVmdDViRlJBQVRu
            ZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cgciwgIGFsbG93IGNv
            bW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciI6ImFsbG93IHJ3IiwgIm9zZCI6ImFsbG93IHJ3
            IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFw
            Wyd7ImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJzdGF0dXMifSddID0gY2VwaF9zdGF0dXNf
            c3RyCgogICAgZGVmIHNodXRkb3duKHNlbGYpOgogICAgICAgIHBhc3MKCiAgICBkZWYgZ2V0X2Zz
            aWQoc2VsZik6CiAgICAgICAgcmV0dXJuICJhZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5
            ZDBmMWMiCgogICAgZGVmIGNvbmZfcmVhZF9maWxlKHNlbGYpOgogICAgICAgIHBhc3MKCiAgICBk
            ZWYgY29ubmVjdChzZWxmKToKICAgICAgICBwYXNzCgogICAgZGVmIHBvb2xfZXhpc3RzKHNlbGYs
            IHBvb2xfbmFtZSk6CiAgICAgICAgcmV0dXJuIFRydWUKCiAgICBkZWYgbW9uX2NvbW1hbmQoc2Vs
            ZiwgY21kLCBvdXQpOgogICAgICAgIGpzb25fY21kID0ganNvbi5sb2FkcyhjbWQpCiAgICAgICAg
            anNvbl9jbWRfc3RyID0ganNvbi5kdW1wcyhqc29uX2NtZCwgc29ydF9rZXlzPVRydWUpCiAgICAg
            ICAgY21kX291dHB1dCA9IHNlbGYuY21kX291dHB1dF9tYXBbanNvbl9jbWRfc3RyXQogICAgICAg
            IHJldHVybiBzZWxmLnJldHVybl92YWwsIGNtZF9vdXRwdXQsIHN0cihzZWxmLmVycl9tZXNzYWdl
            LmVuY29kZSgidXRmLTgiKSkKCiAgICBkZWYgX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoc2VsZiwg
            aG9zdF9uYW1lKToKICAgICAgICBpcF9yZWdfeCA9IHJlLmNvbXBpbGUociJcZHsxLDN9LlxkezEs
            M30uXGR7MSwzfS5cZHsxLDN9IikKICAgICAgICAjIGlmIHByb3ZpZGVkIGhvc3QgaXMgZGlyZWN0
            bHkgYW4gSVAgYWRkcmVzcywgcmV0dXJuIHRoZSBzYW1lCiAgICAgICAgaWYgaXBfcmVnX3gubWF0
            Y2goaG9zdF9uYW1lKToKICAgICAgICAgICAgcmV0dXJuIGhvc3RfbmFtZQogICAgICAgIGltcG9y
            dCByYW5kb20KCiAgICAgICAgaG9zdF9pcCA9IHNlbGYuZHVtbXlfaG9zdF9pcF9tYXAuZ2V0KGhv
            c3RfbmFtZSwgIiIpCiAgICAgICAgaWYgbm90IGhvc3RfaXA6CiAgICAgICAgICAgIGhvc3RfaXAg
            PSBmIjE3Mi45LntyYW5kb20ucmFuZGludCgwLCAyNTQpfS57cmFuZG9tLnJhbmRpbnQoMCwgMjU0
            KX0iCiAgICAgICAgICAgIHNlbGYuZHVtbXlfaG9zdF9pcF9tYXBbaG9zdF9uYW1lXSA9IGhvc3Rf
            aXAKICAgICAgICBkZWwgcmFuZG9tCiAgICAgICAgcmV0dXJuIGhvc3RfaXAKCiAgICBAY2xhc3Nt
            ZXRob2QKICAgIGRlZiBSYWRvcyhjb25mZmlsZT1Ob25lKToKICAgICAgICByZXR1cm4gRHVtbXlS
            YWRvcygpCgoKY2xhc3MgUzNBdXRoKEF1dGhCYXNlKToKICAgICIiIkF0dGFjaGVzIEFXUyBBdXRo
            ZW50aWNhdGlvbiB0byB0aGUgZ2l2ZW4gUmVxdWVzdCBvYmplY3QuIiIiCgogICAgc2VydmljZV9i
            YXNlX3VybCA9ICJzMy5hbWF6b25hd3MuY29tIgoKICAgIGRlZiBfX2luaXRfXyhzZWxmLCBhY2Nl
            c3Nfa2V5LCBzZWNyZXRfa2V5LCBzZXJ2aWNlX3VybD1Ob25lKToKICAgICAgICBpZiBzZXJ2aWNl
            X3VybDoKICAgICAgICAgICAgc2VsZi5zZXJ2aWNlX2Jhc2VfdXJsID0gc2VydmljZV91cmwKICAg
            ICAgICBzZWxmLmFjY2Vzc19rZXkgPSBzdHIoYWNjZXNzX2tleSkKICAgICAgICBzZWxmLnNlY3Jl
            dF9rZXkgPSBzdHIoc2VjcmV0X2tleSkKCiAgICBkZWYgX19jYWxsX18oc2VsZiwgcik6CiAgICAg
            ICAgIyBDcmVhdGUgZGF0ZSBoZWFkZXIgaWYgaXQgaXMgbm90IGNyZWF0ZWQgeWV0LgogICAgICAg
            IGlmICJkYXRlIiBub3QgaW4gci5oZWFkZXJzIGFuZCAieC1hbXotZGF0ZSIgbm90IGluIHIuaGVh
            ZGVyczoKICAgICAgICAgICAgci5oZWFkZXJzWyJkYXRlIl0gPSBmb3JtYXRkYXRlKHRpbWV2YWw9
            Tm9uZSwgbG9jYWx0aW1lPUZhbHNlLCB1c2VnbXQ9VHJ1ZSkKICAgICAgICBzaWduYXR1cmUgPSBz
            ZWxmLmdldF9zaWduYXR1cmUocikKICAgICAgICBpZiBweTNrOgogICAgICAgICAgICBzaWduYXR1
            cmUgPSBzaWduYXR1cmUuZGVjb2RlKCJ1dGYtOCIpCiAgICAgICAgci5oZWFkZXJzWyJBdXRob3Jp
            emF0aW9uIl0gPSBmIkFXUyB7c2VsZi5hY2Nlc3Nfa2V5fTp7c2lnbmF0dXJlfSIKICAgICAgICBy
            ZXR1cm4gcgoKICAgIGRlZiBnZXRfc2lnbmF0dXJlKHNlbGYsIHIpOgogICAgICAgIGNhbm9uaWNh
            bF9zdHJpbmcgPSBzZWxmLmdldF9jYW5vbmljYWxfc3RyaW5nKHIudXJsLCByLmhlYWRlcnMsIHIu
            bWV0aG9kKQogICAgICAgIGlmIHB5M2s6CiAgICAgICAgICAgIGtleSA9IHNlbGYuc2VjcmV0X2tl
            eS5lbmNvZGUoInV0Zi04IikKICAgICAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZy5lbmNv
            ZGUoInV0Zi04IikKICAgICAgICBlbHNlOgogICAgICAgICAgICBrZXkgPSBzZWxmLnNlY3JldF9r
            ZXkKICAgICAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZwogICAgICAgIGggPSBobWFjLm5l
            dyhrZXksIG1zZywgZGlnZXN0bW9kPXNoYSkKICAgICAgICByZXR1cm4gZW5jb2Rlc3RyaW5nKGgu
            ZGlnZXN0KCkpLnN0cmlwKCkKCiAgICBkZWYgZ2V0X2Nhbm9uaWNhbF9zdHJpbmcoc2VsZiwgdXJs
            LCBoZWFkZXJzLCBtZXRob2QpOgogICAgICAgIHBhcnNlZHVybCA9IHVybHBhcnNlKHVybCkKICAg
            ICAgICBvYmplY3RrZXkgPSBwYXJzZWR1cmwucGF0aFsxOl0KCiAgICAgICAgYnVja2V0ID0gcGFy
            c2VkdXJsLm5ldGxvY1s6IC1sZW4oc2VsZi5zZXJ2aWNlX2Jhc2VfdXJsKV0KICAgICAgICBpZiBs
            ZW4oYnVja2V0KSA+IDE6CiAgICAgICAgICAgICMgcmVtb3ZlIGxhc3QgZG90CiAgICAgICAgICAg
            IGJ1Y2tldCA9IGJ1Y2tldFs6LTFdCgogICAgICAgIGludGVyZXN0aW5nX2hlYWRlcnMgPSB7ImNv
            bnRlbnQtbWQ1IjogIiIsICJjb250ZW50LXR5cGUiOiAiIiwgImRhdGUiOiAiIn0KICAgICAgICBm
            b3Iga2V5IGluIGhlYWRlcnM6CiAgICAgICAgICAgIGxrID0ga2V5Lmxvd2VyKCkKICAgICAgICAg
            ICAgdHJ5OgogICAgICAgICAgICAgICAgbGsgPSBsay5kZWNvZGUoInV0Zi04IikKICAgICAgICAg
            ICAgZXhjZXB0OgogICAgICAgICAgICAgICAgcGFzcwogICAgICAgICAgICBpZiBoZWFkZXJzW2tl
            eV0gYW5kICgKICAgICAgICAgICAgICAgIGxrIGluIGludGVyZXN0aW5nX2hlYWRlcnMua2V5cygp
            IG9yIGxrLnN0YXJ0c3dpdGgoIngtYW16LSIpCiAgICAgICAgICAgICk6CiAgICAgICAgICAgICAg
            ICBpbnRlcmVzdGluZ19oZWFkZXJzW2xrXSA9IGhlYWRlcnNba2V5XS5zdHJpcCgpCgogICAgICAg
            ICMgSWYgeC1hbXotZGF0ZSBpcyB1c2VkIGl0IHN1cGVyc2VkZXMgdGhlIGRhdGUgaGVhZGVyLgog
            ICAgICAgIGlmIG5vdCBweTNrOgogICAgICAgICAgICBpZiAieC1hbXotZGF0ZSIgaW4gaW50ZXJl
            c3RpbmdfaGVhZGVyczoKICAgICAgICAgICAgICAgIGludGVyZXN0aW5nX2hlYWRlcnNbImRhdGUi
            XSA9ICIiCiAgICAgICAgZWxzZToKICAgICAgICAgICAgaWYgIngtYW16LWRhdGUiIGluIGludGVy
            ZXN0aW5nX2hlYWRlcnM6CiAgICAgICAgICAgICAgICBpbnRlcmVzdGluZ19oZWFkZXJzWyJkYXRl
            Il0gPSAiIgoKICAgICAgICBidWYgPSBmInttZXRob2R9XG4iCiAgICAgICAgZm9yIGtleSBpbiBz
            b3J0ZWQoaW50ZXJlc3RpbmdfaGVhZGVycy5rZXlzKCkpOgogICAgICAgICAgICB2YWwgPSBpbnRl
            cmVzdGluZ19oZWFkZXJzW2tleV0KICAgICAgICAgICAgaWYga2V5LnN0YXJ0c3dpdGgoIngtYW16
            LSIpOgogICAgICAgICAgICAgICAgYnVmICs9IGYie2tleX06e3ZhbH1cbiIKICAgICAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgICAgIGJ1ZiArPSBmInt2YWx9XG4iCgogICAgICAgICMgYXBwZW5k
            IHRoZSBidWNrZXQgaWYgaXQgZXhpc3RzCiAgICAgICAgaWYgYnVja2V0ICE9ICIiOgogICAgICAg
            ICAgICBidWYgKz0gZiIve2J1Y2tldH0iCgogICAgICAgICMgYWRkIHRoZSBvYmplY3RrZXkuIGV2
            ZW4gaWYgaXQgZG9lc24ndCBleGlzdCwgYWRkIHRoZSBzbGFzaAogICAgICAgIGJ1ZiArPSBmIi97
            b2JqZWN0a2V5fSIKCiAgICAgICAgcmV0dXJuIGJ1ZgoKCmNsYXNzIFJhZG9zSlNPTjoKICAgIEVY
            VEVSTkFMX1VTRVJfTkFNRSA9ICJjbGllbnQuaGVhbHRoY2hlY2tlciIKICAgIEVYVEVSTkFMX1JH
            V19BRE1JTl9PUFNfVVNFUl9OQU1FID0gInJndy1hZG1pbi1vcHMtdXNlciIKICAgIEVNUFRZX09V
            VFBVVF9MSVNUID0gIkVtcHR5IG91dHB1dCBsaXN0IgogICAgREVGQVVMVF9SR1dfUE9PTF9QUkVG
            SVggPSAiZGVmYXVsdCIKICAgIERFRkFVTFRfTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUID0gIjky
            ODMiCgogICAgQGNsYXNzbWV0aG9kCiAgICBkZWYgZ2VuX2FyZ19wYXJzZXIoY2xzLCBhcmdzX3Rv
            X3BhcnNlPU5vbmUpOgogICAgICAgIGFyZ1AgPSBhcmdwYXJzZS5Bcmd1bWVudFBhcnNlcigpCgog
            ICAgICAgIGNvbW1vbl9ncm91cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJjb21tb24iKQog
            ICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoIi0tdmVyYm9zZSIsICItdiIsIGFjdGlv
            bj0ic3RvcmVfdHJ1ZSIsIGRlZmF1bHQ9RmFsc2UpCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9h
            cmd1bWVudCgKICAgICAgICAgICAgIi0tY2VwaC1jb25mIiwgIi1jIiwgaGVscD0iUHJvdmlkZSBh
            IGNlcGggY29uZiBmaWxlLiIsIHR5cGU9c3RyCiAgICAgICAgKQogICAgICAgIGNvbW1vbl9ncm91
            cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLWtleXJpbmciLCAiLWsiLCBoZWxwPSJQYXRo
            IHRvIGNlcGgga2V5cmluZyBmaWxlLiIsIHR5cGU9c3RyCiAgICAgICAgKQogICAgICAgIGNvbW1v
            bl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJ1bi1hcy11c2VyIiwKICAgICAg
            ICAgICAgIi11IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgdHlwZT1zdHIs
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIGEgdXNlciBuYW1lIHRvIGNoZWNrIHRoZSBjbHVz
            dGVyJ3MgaGVhbHRoIHN0YXR1cywgbXVzdCBiZSBwcmVmaXhlZCBieSAnY2xpZW50LiciLAogICAg
            ICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1j
            bHVzdGVyLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJL
            dWJlcm5ldGVzIGNsdXN0ZXIgbmFtZShsZWdhY3kgZmxhZyksIE5vdGU6IEVpdGhlciB1c2UgdGhp
            cyBvciAtLWs4cy1jbHVzdGVyLW5hbWUiLAogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAu
            YWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1rOHMtY2x1c3Rlci1uYW1lIiwgZGVmYXVsdD0i
            IiwgaGVscD0iS3ViZXJuZXRlcyBjbHVzdGVyIG5hbWUiCiAgICAgICAgKQogICAgICAgIGNvbW1v
            bl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLW5hbWVzcGFjZSIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIGhlbHA9Ik5hbWVzcGFjZSB3aGVyZSBDZXBoQ2x1
            c3RlciBpcyBydW5uaW5nIiwKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tcmd3LXBvb2wtcHJlZml4IiwgZGVmYXVsdD0iIiwgaGVscD0i
            UkdXIFBvb2wgcHJlZml4IgogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3Vt
            ZW50KAogICAgICAgICAgICAiLS1yZXN0cmljdGVkLWF1dGgtcGVybWlzc2lvbiIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlJlc3RyaWN0IGNlcGhDU0lLZXly
            aW5ncyBhdXRoIHBlcm1pc3Npb25zIHRvIHNwZWNpZmljIHBvb2xzLCBjbHVzdGVyLiIKICAgICAg
            ICAgICAgKyAiTWFuZGF0b3J5IGZsYWdzIHRoYXQgbmVlZCB0byBiZSBzZXQgYXJlIC0tcmJkLWRh
            dGEtcG9vbC1uYW1lLCBhbmQgLS1rOHMtY2x1c3Rlci1uYW1lLiIKICAgICAgICAgICAgKyAiLS1j
            ZXBoZnMtZmlsZXN5c3RlbS1uYW1lIGZsYWcgY2FuIGFsc28gYmUgcGFzc2VkIGluIGNhc2Ugb2Yg
            Y2VwaGZzIHVzZXIgcmVzdHJpY3Rpb24sIHNvIGl0IGNhbiByZXN0cmljdCB1c2VyIHRvIHBhcnRp
            Y3VsYXIgY2VwaGZzIGZpbGVzeXN0ZW0iCiAgICAgICAgICAgICsgInNhbXBsZSBydW46IGBweXRo
            b24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwtY2x1c3Rlci1yZXNvdXJjZXMucHkgLS1jZXBo
            ZnMtZmlsZXN5c3RlbS1uYW1lIG15ZnMgLS1yYmQtZGF0YS1wb29sLW5hbWUgcmVwbGljYXBvb2wg
            LS1rOHMtY2x1c3Rlci1uYW1lIHJvb2tzdG9yYWdlIC0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Np
            b24gdHJ1ZWAiCiAgICAgICAgICAgICsgIk5vdGU6IFJlc3RyaWN0aW5nIHRoZSBjc2ktdXNlcnMg
            cGVyIHBvb2wsIGFuZCBwZXIgY2x1c3RlciB3aWxsIHJlcXVpcmUgY3JlYXRpbmcgbmV3IGNzaS11
            c2VycyBhbmQgbmV3IHNlY3JldHMgZm9yIHRoYXQgY3NpLXVzZXJzLiIKICAgICAgICAgICAgKyAi
            U28gYXBwbHkgdGhlc2Ugc2VjcmV0cyBvbmx5IHRvIG5ldyBgQ29uc3VtZXIgY2x1c3RlcmAgZGVw
            bG95bWVudCB3aGlsZSB1c2luZyB0aGUgc2FtZSBgU291cmNlIGNsdXN0ZXJgLiIsCiAgICAgICAg
            KQogICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXYyLXBv
            cnQtZW5hYmxlIiwKICAgICAgICAgICAgYWN0aW9uPSJzdG9yZV90cnVlIiwKICAgICAgICAgICAg
            ZGVmYXVsdD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iRW5hYmxlIHYyIG1vbiBwb3J0KDMzMDAp
            IGZvciBtb25zIiwKICAgICAgICApCgogICAgICAgIG91dHB1dF9ncm91cCA9IGFyZ1AuYWRkX2Fy
            Z3VtZW50X2dyb3VwKCJvdXRwdXQiKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQo
            CiAgICAgICAgICAgICItLWZvcm1hdCIsCiAgICAgICAgICAgICItdCIsCiAgICAgICAgICAgIGNo
            b2ljZXM9WyJqc29uIiwgImJhc2giXSwKICAgICAgICAgICAgZGVmYXVsdD0ianNvbiIsCiAgICAg
            ICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBvdXRwdXQgZm9ybWF0IChqc29uIHwgYmFzaCkiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1vdXRwdXQiLAogICAgICAgICAgICAiLW8iLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAg
            ICAgICAgICBoZWxwPSJPdXRwdXQgd2lsbCBiZSBzdG9yZWQgaW50byB0aGUgcHJvdmlkZWQgZmls
            ZSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAog
            ICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgQ2VwaCBmaWxlc3lzdGVt
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tY2VwaGZzLW1ldGFkYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIs
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBjZXBoZnMgbWV0YWRh
            dGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAg
            ICAgICAgICAgICItLWNlcGhmcy1kYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9
            IiIsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBjZXBoZnMgZGF0
            YSBwb29sIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAg
            ICAgICAgICAgIi0tcmJkLWRhdGEtcG9vbC1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwK
            ICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRo
            ZSBuYW1lIG9mIHRoZSBSQkQgZGF0YXBvb2wiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3Jv
            dXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1hbGlhcy1yYmQtZGF0YS1wb29sLW5hbWUi
            LAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAg
            ICAgICAgICAgaGVscD0iUHJvdmlkZXMgYW4gYWxpYXMgZm9yIHRoZSAgUkJEIGRhdGEgcG9vbCBu
            YW1lLCBuZWNlc3NhcnkgaWYgYSBzcGVjaWFsIGNoYXJhY3RlciBpcyBwcmVzZW50IGluIHRoZSBw
            b29sIG5hbWUgc3VjaCBhcyBhIHBlcmlvZCBvciB1bmRlcnNjb3JlIiwKICAgICAgICApCiAgICAg
            ICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmd3LWVuZHBvaW50
            IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAg
            ICAgICAgICAgIGhlbHA9IlJBRE9TIEdhdGV3YXkgZW5kcG9pbnQgKGluIGA8SVB2ND46PFBPUlQ+
            YCBvciBgPFtJUHY2XT46PFBPUlQ+YCBvciBgPEZRRE4+OjxQT1JUPmAgZm9ybWF0KSIsCiAgICAg
            ICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJn
            dy10bHMtY2VydC1wYXRoIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVx
            dWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlJBRE9TIEdhdGV3YXkgZW5kcG9pbnQgVExT
            IGNlcnRpZmljYXRlIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVu
            dCgKICAgICAgICAgICAgIi0tcmd3LXNraXAtdGxzIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFs
            c2UsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9Iklnbm9yZSBU
            TFMgY2VydGlmaWNhdGlvbiB2YWxpZGF0aW9uIHdoZW4gYSBzZWxmLXNpZ25lZCBjZXJ0aWZpY2F0
            ZSBpcyBwcm92aWRlZCAoTk9UIFJFQ09NTUVOREVEIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tbW9uaXRvcmluZy1lbmRwb2ludCIs
            CiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAg
            ICAgICAgICBoZWxwPSJDZXBoIE1hbmFnZXIgcHJvbWV0aGV1cyBleHBvcnRlciBlbmRwb2ludHMg
            KGNvbW1hIHNlcGFyYXRlZCBsaXN0IG9mIChmb3JtYXQgYDxJUHY0PmAgb3IgYDxbSVB2Nl0+YCBv
            ciBgPEZRRE4+YCkgZW50cmllcyBvZiBhY3RpdmUgYW5kIHN0YW5kYnkgbWdycykiLAogICAgICAg
            ICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1tb25p
            dG9yaW5nLWVuZHBvaW50LXBvcnQiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAg
            ICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iQ2VwaCBNYW5hZ2VyIHByb21ldGhl
            dXMgZXhwb3J0ZXIgcG9ydCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJn
            dW1lbnQoCiAgICAgICAgICAgICItLXNraXAtbW9uaXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAg
            ICAgICAgIGhlbHA9IkRvIG5vdCBjaGVjayBmb3IgYSBtb25pdG9yaW5nIGVuZHBvaW50IGZvciB0
            aGUgQ2VwaCBjbHVzdGVyIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tcmJkLW1ldGFkYXRhLWVjLXBvb2wtbmFtZSIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBo
            ZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiBlcmFzdXJlIGNvZGVkIFJCRCBtZXRhZGF0YSBwb29s
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tZHJ5LXJ1biIsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFj
            dGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGhlbHA9IkRyeSBydW4gcHJpbnRzIHRoZSBl
            eGVjdXRlZCBjb21tYW5kcyB3aXRob3V0IHJ1bm5pbmcgdGhlbSIsCiAgICAgICAgKQogICAgICAg
            IG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJhZG9zLW5hbWVzcGFj
            ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAog
            ICAgICAgICAgICBoZWxwPSJEaXZpZGVzIGEgcG9vbCBpbnRvIHNlcGFyYXRlIGxvZ2ljYWwgbmFt
            ZXNwYWNlcywgdXNlZCBmb3IgY3JlYXRpbmcgUkJEIFBWQyBpbiBhIENlcGhCbG9ja1Bvb2xSYWRv
            c05hbWVzcGFjZSAoc2hvdWxkIGJlIGxvd2VyIGNhc2UpIiwKICAgICAgICApCiAgICAgICAgb3V0
            cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tc3Vidm9sdW1lLWdyb3VwIiwK
            ICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAg
            ICAgICAgIGhlbHA9InByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBzdWJ2b2x1bWUgZ3JvdXAiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1yZ3ctcmVhbG0tbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJl
            cXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUg
            cmd3LXJlYWxtIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgK
            ICAgICAgICAgICAgIi0tcmd3LXpvbmUtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAg
            ICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUg
            bmFtZSBvZiB0aGUgcmd3LXpvbmUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS1yZ3ctem9uZWdyb3VwLW5hbWUiLAogICAgICAgICAg
            ICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVs
            cD0icHJvdmlkZXMgdGhlIG5hbWUgb2YgdGhlIHJndy16b25lZ3JvdXAiLAogICAgICAgICkKICAg
            ICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS10b3BvbG9neS1w
            b29scyIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNl
            LAogICAgICAgICAgICBoZWxwPSJjb21tYS1zZXBhcmF0ZWQgbGlzdCBvZiB0b3BvbG9neS1jb25z
            dHJhaW5lZCByYmQgcG9vbHMiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2Fy
            Z3VtZW50KAogICAgICAgICAgICAiLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi1sYWJlbCIsCiAg
            ICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAg
            ICAgICBoZWxwPSJrOHMgY2x1c3RlciBmYWlsdXJlIGRvbWFpbiBsYWJlbCAoZXhhbXBsZTogem9u
            ZSwgcmFjaywgb3IgaG9zdCkgZm9yIHRoZSB0b3BvbG9neS1wb29scyB0aGF0IG1hdGNoIHRoZSBj
            ZXBoIGRvbWFpbiIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQo
            CiAgICAgICAgICAgICItLXRvcG9sb2d5LWZhaWx1cmUtZG9tYWluLXZhbHVlcyIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBo
            ZWxwPSJjb21tYS1zZXBhcmF0ZWQgbGlzdCBvZiB0aGUgazhzIGNsdXN0ZXIgZmFpbHVyZSBkb21h
            aW4gdmFsdWVzIGNvcnJlc3BvbmRpbmcgdG8gZWFjaCBvZiB0aGUgcG9vbHMgaW4gdGhlIGB0b3Bv
            bG9neS1wb29sc2AgbGlzdCIsCiAgICAgICAgKQoKICAgICAgICB1cGdyYWRlX2dyb3VwID0gYXJn
            UC5hZGRfYXJndW1lbnRfZ3JvdXAoInVwZ3JhZGUiKQogICAgICAgIHVwZ3JhZGVfZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS11cGdyYWRlIiwKICAgICAgICAgICAgYWN0aW9uPSJz
            dG9yZV90cnVlIiwKICAgICAgICAgICAgZGVmYXVsdD1GYWxzZSwKICAgICAgICAgICAgaGVscD0i
            VXBncmFkZXMgdGhlIGNlcGhDU0lLZXlyaW5ncyhGb3IgZXhhbXBsZTogY2xpZW50LmNzaS1jZXBo
            ZnMtcHJvdmlzaW9uZXIpIGFuZCBjbGllbnQuaGVhbHRoY2hlY2tlciBjZXBoIHVzZXJzIHdpdGgg
            bmV3IHBlcm1pc3Npb25zIG5lZWRlZCBmb3IgdGhlIG5ldyBjbHVzdGVyIHZlcnNpb24gYW5kIG9s
            ZGVyIHBlcm1pc3Npb24gd2lsbCBzdGlsbCBiZSBhcHBsaWVkLiIKICAgICAgICAgICAgKyAiU2Ft
            cGxlIHJ1bjogYHB5dGhvbjMgL2V0Yy9jZXBoL2NyZWF0ZS1leHRlcm5hbC1jbHVzdGVyLXJlc291
            cmNlcy5weSAtLXVwZ3JhZGVgLCB0aGlzIHdpbGwgdXBncmFkZSBhbGwgdGhlIGRlZmF1bHQgY3Np
            IHVzZXJzKG5vbi1yZXN0cmljdGVkKSIKICAgICAgICAgICAgKyAiRm9yIHJlc3RyaWN0ZWQgdXNl
            cnMoRm9yIGV4YW1wbGU6IGNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1z
            dG9yYWdlLW15ZnMpLCB1c2VycyBjcmVhdGVkIHVzaW5nIC0tcmVzdHJpY3RlZC1hdXRoLXBlcm1p
            c3Npb24gZmxhZyBuZWVkIHRvIHBhc3MgbWFuZGF0b3J5IGZsYWdzIgogICAgICAgICAgICArICJt
            YW5kYXRvcnkgZmxhZ3M6ICctLXJiZC1kYXRhLXBvb2wtbmFtZSwgLS1rOHMtY2x1c3Rlci1uYW1l
            IGFuZCAtLXJ1bi1hcy11c2VyJyBmbGFncyB3aGlsZSB1cGdyYWRpbmciCiAgICAgICAgICAgICsg
            ImluIGNhc2Ugb2YgY2VwaGZzIHVzZXJzIGlmIHlvdSBoYXZlIHBhc3NlZCAtLWNlcGhmcy1maWxl
            c3lzdGVtLW5hbWUgZmxhZyB3aGlsZSBjcmVhdGluZyB1c2VyIHRoZW4gd2hpbGUgdXBncmFkaW5n
            IGl0IHdpbGwgYmUgbWFuZGF0b3J5IHRvbyIKICAgICAgICAgICAgKyAiU2FtcGxlIHJ1bjogYHB5
            dGhvbjMgL2V0Yy9jZXBoL2NyZWF0ZS1leHRlcm5hbC1jbHVzdGVyLXJlc291cmNlcy5weSAtLXVw
            Z3JhZGUgLS1yYmQtZGF0YS1wb29sLW5hbWUgcmVwbGljYXBvb2wgLS1rOHMtY2x1c3Rlci1uYW1l
            IHJvb2tzdG9yYWdlICAtLXJ1bi1hcy11c2VyIGNsaWVudC5jc2ktcmJkLW5vZGUtcm9va3N0b3Jh
            Z2UtcmVwbGljYXBvb2xgIgogICAgICAgICAgICArICJQUzogQW4gZXhpc3Rpbmcgbm9uLXJlc3Ry
            aWN0ZWQgdXNlciBjYW5ub3QgYmUgY29udmVydGVkIHRvIGEgcmVzdHJpY3RlZCB1c2VyIGJ5IHVw
            Z3JhZGluZy4iCiAgICAgICAgICAgICsgIlVwZ3JhZGUgZmxhZyBzaG91bGQgb25seSBiZSB1c2Vk
            IHRvIGFwcGVuZCBuZXcgcGVybWlzc2lvbnMgdG8gdXNlcnMsIGl0IHNob3VsZG4ndCBiZSB1c2Vk
            IGZvciBjaGFuZ2luZyB1c2VyIGFscmVhZHkgYXBwbGllZCBwZXJtaXNzaW9uLCBmb3IgZXhhbXBs
            ZSB5b3Ugc2hvdWxkbid0IGNoYW5nZSBpbiB3aGljaCBwb29sIHVzZXIgaGFzIGFjY2VzcyIsCiAg
            ICAgICAgKQoKICAgICAgICBpZiBhcmdzX3RvX3BhcnNlOgogICAgICAgICAgICBhc3NlcnQgKAog
            ICAgICAgICAgICAgICAgdHlwZShhcmdzX3RvX3BhcnNlKSA9PSBsaXN0CiAgICAgICAgICAgICks
            ICJBcmd1bWVudCB0byAnZ2VuX2FyZ19wYXJzZXInIHNob3VsZCBiZSBhIGxpc3QiCiAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgYXJnc190b19wYXJzZSA9IHN5cy5hcmd2WzE6XQogICAgICAgIHJl
            dHVybiBhcmdQLnBhcnNlX2FyZ3MoYXJnc190b19wYXJzZSkKCiAgICBkZWYgdmFsaWRhdGVfcmJk
            X21ldGFkYXRhX2VjX3Bvb2xfbmFtZShzZWxmKToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2Vy
            LnJiZF9tZXRhZGF0YV9lY19wb29sX25hbWU6CiAgICAgICAgICAgIHJiZF9tZXRhZGF0YV9lY19w
            b29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUKICAg
            ICAgICAgICAgcmJkX3Bvb2xfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9u
            YW1lCgogICAgICAgICAgICBpZiByYmRfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiRmxh
            ZyAnLS1yYmQtZGF0YS1wb29sLW5hbWUnIHNob3VsZCBub3QgYmUgZW1wdHkiCiAgICAgICAgICAg
            ICAgICApCgogICAgICAgICAgICBpZiByYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lID09ICIiOgog
            ICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAg
            ICAgICAgICAgICAiRmxhZyAnLS1yYmQtbWV0YWRhdGEtZWMtcG9vbC1uYW1lJyBzaG91bGQgbm90
            IGJlIGVtcHR5IgogICAgICAgICAgICAgICAgKQoKICAgICAgICAgICAgY21kX2pzb24gPSB7InBy
            ZWZpeCI6ICJvc2QgZHVtcCIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgICAgIHJldF92YWws
            IGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikK
            ICAgICAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAg
            ICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAg
            ICAgICAgZiJ7Y21kX2pzb25bJ3ByZWZpeCddfSBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAg
            ICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVN
            UFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgbWV0YWRhdGFf
            cG9vbF9leGlzdCwgcG9vbF9leGlzdCA9IEZhbHNlLCBGYWxzZQoKICAgICAgICAgICAgZm9yIGtl
            eSBpbiBqc29uX291dFsicG9vbHMiXToKICAgICAgICAgICAgICAgICMgaWYgZXJhc3VyZV9jb2Rl
            X3Byb2ZpbGUgaXMgZW1wdHkgYW5kIHBvb2wgbmFtZSBleGlzdHMgdGhlbiBpdCByZXBsaWNhIHBv
            b2wKICAgICAgICAgICAgICAgIGlmICgKICAgICAgICAgICAgICAgICAgICBrZXlbImVyYXN1cmVf
            Y29kZV9wcm9maWxlIl0gPT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQga2V5WyJwb29sX25h
            bWUiXSA9PSByYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lCiAgICAgICAgICAgICAgICApOgogICAg
            ICAgICAgICAgICAgICAgIG1ldGFkYXRhX3Bvb2xfZXhpc3QgPSBUcnVlCiAgICAgICAgICAgICAg
            ICAjIGlmIGVyYXN1cmVfY29kZV9wcm9maWxlIGlzIG5vdCBlbXB0eSBhbmQgcG9vbCBuYW1lIGV4
            aXN0cyB0aGVuIGl0IGlzIGVjIHBvb2wKICAgICAgICAgICAgICAgIGlmIGtleVsiZXJhc3VyZV9j
            b2RlX3Byb2ZpbGUiXSBhbmQga2V5WyJwb29sX25hbWUiXSA9PSByYmRfcG9vbF9uYW1lOgogICAg
            ICAgICAgICAgICAgICAgIHBvb2xfZXhpc3QgPSBUcnVlCgogICAgICAgICAgICBpZiBub3QgbWV0
            YWRhdGFfcG9vbF9leGlzdDoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVF
            eGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIlByb3ZpZGVkIHJiZF9lY19tZXRhZGF0YV9w
            b29sIG5hbWUsIgogICAgICAgICAgICAgICAgICAgIGYiIHtyYmRfbWV0YWRhdGFfZWNfcG9vbF9u
            YW1lfSwgZG9lcyBub3QgZXhpc3QiCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmIG5v
            dCBwb29sX2V4aXN0OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICBmIlByb3ZpZGVkIHJiZF9kYXRhX3Bvb2wgbmFtZSwg
            e3JiZF9wb29sX25hbWV9LCBkb2VzIG5vdCBleGlzdCIKICAgICAgICAgICAgICAgICkKICAgICAg
            ICAgICAgcmV0dXJuIHJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUKCiAgICBkZWYgZHJ5X3J1bihz
            ZWxmLCBtc2cpOgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAg
            ICAgcHJpbnQoIkV4ZWN1dGU6ICIgKyAiJyIgKyBtc2cgKyAiJyIpCgogICAgZGVmIHZhbGlkYXRl
            X3Jnd19lbmRwb2ludF90bHNfY2VydChzZWxmKToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd190bHNfY2VydF9wYXRoOgogICAgICAgICAgICB3aXRoIG9wZW4oc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfdGxzX2NlcnRfcGF0aCwgZW5jb2Rpbmc9InV0ZjgiKSBhcyBmOgogICAgICAgICAgICAg
            ICAgY29udGVudHMgPSBmLnJlYWQoKQogICAgICAgICAgICAgICAgcmV0dXJuIGNvbnRlbnRzLnJz
            dHJpcCgpCgogICAgZGVmIF9jaGVja19jb25mbGljdGluZ19vcHRpb25zKHNlbGYpOgogICAgICAg
            IGlmIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnVwZ3JhZGUgYW5kIG5vdCBzZWxmLl9hcmdfcGFyc2Vy
            LnJiZF9kYXRhX3Bvb2xfbmFtZToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICJFaXRoZXIgJy0tdXBncmFkZScgb3IgJy0tcmJkLWRh
            dGEtcG9vbC1uYW1lIDxwb29sX25hbWU+JyBzaG91bGQgYmUgc3BlY2lmaWVkIgogICAgICAgICAg
            ICApCgogICAgZGVmIF9pbnZhbGlkX2VuZHBvaW50KHNlbGYsIGVuZHBvaW50X3N0cik6CiAgICAg
            ICAgIyBleHRyYWN0IHRoZSBwb3J0IGJ5IGdldHRpbmcgdGhlIGxhc3Qgc3BsaXQgb24gYDpgIGRl
            bGltaXRlcgogICAgICAgIHRyeToKICAgICAgICAgICAgZW5kcG9pbnRfc3RyX2lwLCBwb3J0ID0g
            ZW5kcG9pbnRfc3RyLnJzcGxpdCgiOiIsIDEpCiAgICAgICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAg
            ICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oZiJOb3QgYSBwcm9wZXIg
            ZW5kcG9pbnQ6IHtlbmRwb2ludF9zdHJ9IikKCiAgICAgICAgdHJ5OgogICAgICAgICAgICBpZiBl
            bmRwb2ludF9zdHJfaXBbMF0gPT0gIlsiOgogICAgICAgICAgICAgICAgZW5kcG9pbnRfc3RyX2lw
            ID0gZW5kcG9pbnRfc3RyX2lwWzEgOiBsZW4oZW5kcG9pbnRfc3RyX2lwKSAtIDFdCiAgICAgICAg
            ICAgIGlwX3R5cGUgPSAoCiAgICAgICAgICAgICAgICAiSVB2NCIgaWYgdHlwZShpcF9hZGRyZXNz
            KGVuZHBvaW50X3N0cl9pcCkpIGlzIElQdjRBZGRyZXNzIGVsc2UgIklQdjYiCiAgICAgICAgICAg
            ICkKICAgICAgICBleGNlcHQgVmFsdWVFcnJvcjoKICAgICAgICAgICAgaXBfdHlwZSA9ICJGUURO
            IgogICAgICAgIGlmIG5vdCBwb3J0LmlzZGlnaXQoKToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0
            aW9uRmFpbHVyZUV4Y2VwdGlvbihmIlBvcnQgbm90IHZhbGlkOiB7cG9ydH0iKQogICAgICAgIGlu
            dFBvcnQgPSBpbnQocG9ydCkKICAgICAgICBpZiBpbnRQb3J0IDwgMSBvciBpbnRQb3J0ID4gMioq
            MTYgLSAxOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKGYiT3V0
            IG9mIHJhbmdlIHBvcnQgbnVtYmVyOiB7cG9ydH0iKQoKICAgICAgICByZXR1cm4gaXBfdHlwZQoK
            ICAgIGRlZiBlbmRwb2ludF9kaWFsKHNlbGYsIGVuZHBvaW50X3N0ciwgaXBfdHlwZSwgdGltZW91
            dD0zLCBjZXJ0PU5vbmUpOgogICAgICAgICMgaWYgdGhlICdjbHVzdGVyJyBpbnN0YW5jZSBpcyBh
            IGR1bW15IG9uZSwKICAgICAgICAjIGRvbid0IHRyeSB0byByZWFjaCBvdXQgdG8gdGhlIGVuZHBv
            aW50CiAgICAgICAgaWYgaXNpbnN0YW5jZShzZWxmLmNsdXN0ZXIsIER1bW15UmFkb3MpOgogICAg
            ICAgICAgICByZXR1cm4gIiIsICIiLCAiIgogICAgICAgIGlmIGlwX3R5cGUgPT0gIklQdjYiOgog
            ICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBlbmRwb2ludF9zdHJfaXAsIGVuZHBvaW50
            X3N0cl9wb3J0ID0gZW5kcG9pbnRfc3RyLnJzcGxpdCgiOiIsIDEpCiAgICAgICAgICAgIGV4Y2Vw
            dCBWYWx1ZUVycm9yOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICBmIk5vdCBhIHByb3BlciBlbmRwb2ludDoge2VuZHBv
            aW50X3N0cn0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmIGVuZHBvaW50X3N0cl9p
            cFswXSAhPSAiWyI6CiAgICAgICAgICAgICAgICBlbmRwb2ludF9zdHJfaXAgPSAiWyIgKyBlbmRw
            b2ludF9zdHJfaXAgKyAiXSIKICAgICAgICAgICAgZW5kcG9pbnRfc3RyID0gIjoiLmpvaW4oW2Vu
            ZHBvaW50X3N0cl9pcCwgZW5kcG9pbnRfc3RyX3BvcnRdKQoKICAgICAgICBwcm90b2NvbHMgPSBb
            Imh0dHAiLCAiaHR0cHMiXQogICAgICAgIHJlc3BvbnNlX2Vycm9yID0gTm9uZQogICAgICAgIGZv
            ciBwcmVmaXggaW4gcHJvdG9jb2xzOgogICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBl
            cCA9IGYie3ByZWZpeH06Ly97ZW5kcG9pbnRfc3RyfSIKICAgICAgICAgICAgICAgIHZlcmlmeSA9
            IE5vbmUKICAgICAgICAgICAgICAgICMgSWYgdmVyaWZ5IGlzIHNldCB0byBhIHBhdGggdG8gYSBk
            aXJlY3RvcnksCiAgICAgICAgICAgICAgICAjIHRoZSBkaXJlY3RvcnkgbXVzdCBoYXZlIGJlZW4g
            cHJvY2Vzc2VkIHVzaW5nIHRoZSBjX3JlaGFzaCB1dGlsaXR5IHN1cHBsaWVkIHdpdGggT3BlblNT
            TC4KICAgICAgICAgICAgICAgIGlmIHByZWZpeCA9PSAiaHR0cHMiIGFuZCBzZWxmLl9hcmdfcGFy
            c2VyLnJnd19za2lwX3RsczoKICAgICAgICAgICAgICAgICAgICB2ZXJpZnkgPSBGYWxzZQogICAg
            ICAgICAgICAgICAgICAgIHIgPSByZXF1ZXN0cy5oZWFkKGVwLCB0aW1lb3V0PXRpbWVvdXQsIHZl
            cmlmeT1GYWxzZSkKICAgICAgICAgICAgICAgIGVsaWYgcHJlZml4ID09ICJodHRwcyIgYW5kIGNl
            cnQ6CiAgICAgICAgICAgICAgICAgICAgdmVyaWZ5ID0gY2VydAogICAgICAgICAgICAgICAgICAg
            IHIgPSByZXF1ZXN0cy5oZWFkKGVwLCB0aW1lb3V0PXRpbWVvdXQsIHZlcmlmeT1jZXJ0KQogICAg
            ICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgICAgICByID0gcmVxdWVzdHMuaGVhZChl
            cCwgdGltZW91dD10aW1lb3V0KQogICAgICAgICAgICAgICAgaWYgci5zdGF0dXNfY29kZSA9PSAy
            MDA6CiAgICAgICAgICAgICAgICAgICAgcmV0dXJuIHByZWZpeCwgdmVyaWZ5LCAiIgogICAgICAg
            ICAgICBleGNlcHQgRXhjZXB0aW9uIGFzIGVycjoKICAgICAgICAgICAgICAgIHJlc3BvbnNlX2Vy
            cm9yID0gZXJyCiAgICAgICAgICAgICAgICBjb250aW51ZQogICAgICAgIHN5cy5zdGRlcnIud3Jp
            dGUoCiAgICAgICAgICAgIGYidW5hYmxlIHRvIGNvbm5lY3QgdG8gZW5kcG9pbnQ6IHtlbmRwb2lu
            dF9zdHJ9LCBmYWlsZWQgZXJyb3I6IHtyZXNwb25zZV9lcnJvcn0iCiAgICAgICAgKQogICAgICAg
            IHJldHVybiAoCiAgICAgICAgICAgICIiLAogICAgICAgICAgICAiIiwKICAgICAgICAgICAgKCIt
            MSIpLAogICAgICAgICkKCiAgICBkZWYgX19pbml0X18oc2VsZiwgYXJnX2xpc3Q9Tm9uZSk6CiAg
            ICAgICAgc2VsZi5vdXRfbWFwID0ge30KICAgICAgICBzZWxmLl9leGNsdWRlZF9rZXlzID0gc2V0
            KCkKICAgICAgICBzZWxmLl9hcmdfcGFyc2VyID0gc2VsZi5nZW5fYXJnX3BhcnNlcihhcmdzX3Rv
            X3BhcnNlPWFyZ19saXN0KQogICAgICAgIHNlbGYuX2NoZWNrX2NvbmZsaWN0aW5nX29wdGlvbnMo
            KQogICAgICAgIHNlbGYucnVuX2FzX3VzZXIgPSBzZWxmLl9hcmdfcGFyc2VyLnJ1bl9hc191c2Vy
            CiAgICAgICAgc2VsZi5vdXRwdXRfZmlsZSA9IHNlbGYuX2FyZ19wYXJzZXIub3V0cHV0CiAgICAg
            ICAgc2VsZi5jZXBoX2NvbmYgPSBzZWxmLl9hcmdfcGFyc2VyLmNlcGhfY29uZgogICAgICAgIHNl
            bGYuY2VwaF9rZXlyaW5nID0gc2VsZi5fYXJnX3BhcnNlci5rZXlyaW5nCiAgICAgICAgIyBpZiB1
            c2VyIG5vdCBwcm92aWRlZCwgZ2l2ZSBhIGRlZmF1bHQgdXNlcgogICAgICAgIGlmIG5vdCBzZWxm
            LnJ1bl9hc191c2VyIGFuZCBub3Qgc2VsZi5fYXJnX3BhcnNlci51cGdyYWRlOgogICAgICAgICAg
            ICBzZWxmLnJ1bl9hc191c2VyID0gc2VsZi5FWFRFUk5BTF9VU0VSX05BTUUKICAgICAgICBpZiBu
            b3Qgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXggYW5kIG5vdCBzZWxmLl9hcmdfcGFy
            c2VyLnVwZ3JhZGU6CiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4
            ID0gc2VsZi5ERUZBVUxUX1JHV19QT09MX1BSRUZJWAogICAgICAgIGlmIHNlbGYuY2VwaF9jb25m
            OgogICAgICAgICAgICBrd2FyZ3MgPSB7fQogICAgICAgICAgICBpZiBzZWxmLmNlcGhfa2V5cmlu
            ZzoKICAgICAgICAgICAgICAgIGt3YXJnc1siY29uZiJdID0geyJrZXlyaW5nIjogc2VsZi5jZXBo
            X2tleXJpbmd9CiAgICAgICAgICAgIHNlbGYuY2x1c3RlciA9IHJhZG9zLlJhZG9zKGNvbmZmaWxl
            PXNlbGYuY2VwaF9jb25mLCAqKmt3YXJncykKICAgICAgICBlbHNlOgogICAgICAgICAgICBzZWxm
            LmNsdXN0ZXIgPSByYWRvcy5SYWRvcygpCiAgICAgICAgICAgIHNlbGYuY2x1c3Rlci5jb25mX3Jl
            YWRfZmlsZSgpCiAgICAgICAgc2VsZi5jbHVzdGVyLmNvbm5lY3QoKQoKICAgIGRlZiBzaHV0ZG93
            bihzZWxmKToKICAgICAgICBpZiBzZWxmLmNsdXN0ZXIuc3RhdGUgPT0gImNvbm5lY3RlZCI6CiAg
            ICAgICAgICAgIHNlbGYuY2x1c3Rlci5zaHV0ZG93bigpCgogICAgZGVmIGdldF9mc2lkKHNlbGYp
            OgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJu
            IHNlbGYuZHJ5X3J1bigiY2VwaCBmc2lkIikKICAgICAgICByZXR1cm4gc3RyKHNlbGYuY2x1c3Rl
            ci5nZXRfZnNpZCgpKQoKICAgIGRlZiBfY29tbW9uX2NtZF9qc29uX2dlbihzZWxmLCBjbWRfanNv
            bik6CiAgICAgICAgY21kID0ganNvbi5kdW1wcyhjbWRfanNvbiwgc29ydF9rZXlzPVRydWUpCiAg
            ICAgICAgcmV0X3ZhbCwgY21kX291dCwgZXJyX21zZyA9IHNlbGYuY2x1c3Rlci5tb25fY29tbWFu
            ZChjbWQsIGIiIikKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnZlcmJvc2U6CiAgICAgICAg
            ICAgIHByaW50KGYiQ29tbWFuZCBJbnB1dDoge2NtZH0iKQogICAgICAgICAgICBwcmludCgKICAg
            ICAgICAgICAgICAgIGYiUmV0dXJuIFZhbDoge3JldF92YWx9XG5Db21tYW5kIE91dHB1dDoge2Nt
            ZF9vdXR9XG4iCiAgICAgICAgICAgICAgICBmIkVycm9yIE1lc3NhZ2U6IHtlcnJfbXNnfVxuLS0t
            LS0tLS0tLVxuIgogICAgICAgICAgICApCiAgICAgICAganNvbl9vdXQgPSB7fQogICAgICAgICMg
            aWYgdGhlcmUgaXMgbm8gZXJyb3IgKGkuZTsgcmV0X3ZhbCBpcyBaRVJPKSBhbmQgJ2NtZF9vdXQn
            IGlzIG5vdCBlbXB0eQogICAgICAgICMgdGhlbiBjb252ZXJ0ICdjbWRfb3V0JyB0byBhIGpzb24g
            b3V0cHV0CiAgICAgICAgaWYgcmV0X3ZhbCA9PSAwIGFuZCBjbWRfb3V0OgogICAgICAgICAgICBq
            c29uX291dCA9IGpzb24ubG9hZHMoY21kX291dCkKICAgICAgICByZXR1cm4gcmV0X3ZhbCwganNv
            bl9vdXQsIGVycl9tc2cKCiAgICBkZWYgZ2V0X2NlcGhfZXh0ZXJuYWxfbW9uX2RhdGEoc2VsZik6
            CiAgICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJxdW9ydW1fc3RhdHVzIiwgImZvcm1hdCI6
            ICJqc29uIn0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAg
            IHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGggIiArIGNtZF9qc29uWyJwcmVmaXgiXSkKICAgICAg
            ICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4o
            Y21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwK
            ICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgIidxdW9y
            dW1fc3RhdHVzJyBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAgICAgIGYiRXJyb3I6IHtl
            cnJfbXNnIGlmIHJldF92YWwgIT0gMCBlbHNlIHNlbGYuRU1QVFlfT1VUUFVUX0xJU1R9IgogICAg
            ICAgICAgICApCiAgICAgICAgcV9sZWFkZXJfbmFtZSA9IGpzb25fb3V0WyJxdW9ydW1fbGVhZGVy
            X25hbWUiXQogICAgICAgIHFfbGVhZGVyX2RldGFpbHMgPSB7fQogICAgICAgIHFfbGVhZGVyX21h
            dGNoaW5nX2xpc3QgPSBbCiAgICAgICAgICAgIGwgZm9yIGwgaW4ganNvbl9vdXRbIm1vbm1hcCJd
            WyJtb25zIl0gaWYgbFsibmFtZSJdID09IHFfbGVhZGVyX25hbWUKICAgICAgICBdCiAgICAgICAg
            aWYgbGVuKHFfbGVhZGVyX21hdGNoaW5nX2xpc3QpID09IDA6CiAgICAgICAgICAgIHJhaXNlIEV4
            ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oIk5vIG1hdGNoaW5nICdtb24nIGRldGFpbHMgZm91bmQi
            KQogICAgICAgIHFfbGVhZGVyX2RldGFpbHMgPSBxX2xlYWRlcl9tYXRjaGluZ19saXN0WzBdCiAg
            ICAgICAgIyBnZXQgdGhlIGFkZHJlc3MgdmVjdG9yIG9mIHRoZSBxdW9ydW0tbGVhZGVyCiAgICAg
            ICAgcV9sZWFkZXJfYWRkcnZlYyA9IHFfbGVhZGVyX2RldGFpbHMuZ2V0KCJwdWJsaWNfYWRkcnMi
            LCB7fSkuZ2V0KCJhZGRydmVjIiwgW10pCiAgICAgICAgaXBfYWRkciA9IHN0cihxX2xlYWRlcl9k
            ZXRhaWxzWyJwdWJsaWNfYWRkciJdLnNwbGl0KCIvIilbMF0pCgogICAgICAgIGlmIHNlbGYuX2Fy
            Z19wYXJzZXIudjJfcG9ydF9lbmFibGU6CiAgICAgICAgICAgIGlmIHFfbGVhZGVyX2FkZHJ2ZWNb
            MF1bInR5cGUiXSA9PSAidjIiOgogICAgICAgICAgICAgICAgaXBfYWRkciA9IHFfbGVhZGVyX2Fk
            ZHJ2ZWNbMF1bImFkZHIiXQogICAgICAgICAgICBlbGlmIGxlbihxX2xlYWRlcl9hZGRydmVjKSA+
            IDEgYW5kIHFfbGVhZGVyX2FkZHJ2ZWNbMV1bInR5cGUiXSA9PSAidjIiOgogICAgICAgICAgICAg
            ICAgaXBfYWRkciA9IHFfbGVhZGVyX2FkZHJ2ZWNbMV1bImFkZHIiXQogICAgICAgICAgICBlbHNl
            OgogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZSgKICAgICAgICAgICAgICAgICAgICAi
            J3YyJyBhZGRyZXNzIHR5cGUgbm90IHByZXNlbnQsIGFuZCAndjItcG9ydC1lbmFibGUnIGZsYWcg
            aXMgcHJvdmlkZWQiCiAgICAgICAgICAgICAgICApCgogICAgICAgIHJldHVybiBmIntzdHIocV9s
            ZWFkZXJfbmFtZSl9PXtpcF9hZGRyfSIKCiAgICBkZWYgX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAo
            c2VsZiwgaG9zdF9uYW1lLCBwb3J0LCBpcF90eXBlKToKICAgICAgICAjIGlmICdjbHVzdGVyJyBp
            bnN0YW5jZSBpcyBhIGR1bW15IHR5cGUsCiAgICAgICAgIyBjYWxsIHRoZSBkdW1teSBpbnN0YW5j
            ZSdzICJjb252ZXJ0IiBtZXRob2QKICAgICAgICBpZiBub3QgaG9zdF9uYW1lOgogICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKCJFbXB0eSBob3N0bmFtZSBwcm92aWRl
            ZCIpCiAgICAgICAgaWYgaXNpbnN0YW5jZShzZWxmLmNsdXN0ZXIsIER1bW15UmFkb3MpOgogICAg
            ICAgICAgICByZXR1cm4gc2VsZi5jbHVzdGVyLl9jb252ZXJ0X2hvc3RuYW1lX3RvX2lwKGhvc3Rf
            bmFtZSkKCiAgICAgICAgaWYgaXBfdHlwZSA9PSAiRlFETiI6CiAgICAgICAgICAgICMgY2hlY2sg
            d2hpY2ggaXAgRlFETiBzaG91bGQgYmUgY29udmVydGVkIHRvLCBJUHY0IG9yIElQdjYKICAgICAg
            ICAgICAgIyBjaGVjayB0aGUgaG9zdCBpcCwgdGhlIGVuZHBvaW50IGlwIHR5cGUgd291bGQgYmUg
            c2ltaWxhciB0byBob3N0IGlwCiAgICAgICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAib3Jj
            aCBob3N0IGxzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICAgICAgcmV0X3ZhbCwganNvbl9v
            dXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAg
            ICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAogICAgICAgICAgICBp
            ZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICAgICAgcmFp
            c2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJ29yY2gg
            aG9zdCBscycgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAgICAgICAgICAgICAgICAgZiJFcnJvcjog
            e2Vycl9tc2cgaWYgcmV0X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAg
            ICAgICAgICAgICAgICApCiAgICAgICAgICAgIGhvc3RfYWRkciA9IGpzb25fb3V0WzBdWyJhZGRy
            Il0KICAgICAgICAgICAgIyBhZGQgOjgwIHNhbXBsZSBwb3J0IGluIGlwX3R5cGUsIGFzIF9pbnZh
            bGlkX2VuZHBvaW50IGFsc28gdmVyaWZ5IHBvcnQKICAgICAgICAgICAgaG9zdF9pcF90eXBlID0g
            c2VsZi5faW52YWxpZF9lbmRwb2ludChob3N0X2FkZHIgKyAiOjgwIikKICAgICAgICAgICAgaW1w
            b3J0IHNvY2tldAoKICAgICAgICAgICAgIyBleGFtcGxlIG91dHB1dCBbKDxBZGRyZXNzRmFtaWx5
            LkFGX0lORVQ6IDI+LCA8U29ja2V0S2luZC5TT0NLX1NUUkVBTTogMT4sIDYsICcnLCAoJzkzLjE4
            NC4yMTYuMzQnLCA4MCkpLCAuLi5dCiAgICAgICAgICAgICMgd2UgbmVlZCB0byBnZXQgOTMuMTg0
            LjIxNi4zNCBzbyBpdCB3b3VsZCBiZSBpcFswXVs0XVswXQogICAgICAgICAgICBpZiBob3N0X2lw
            X3R5cGUgPT0gIklQdjYiOgogICAgICAgICAgICAgICAgaXAgPSBzb2NrZXQuZ2V0YWRkcmluZm8o
            CiAgICAgICAgICAgICAgICAgICAgaG9zdF9uYW1lLCBwb3J0LCBmYW1pbHk9c29ja2V0LkFGX0lO
            RVQ2LCBwcm90bz1zb2NrZXQuSVBQUk9UT19UQ1AKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgZWxpZiBob3N0X2lwX3R5cGUgPT0gIklQdjQiOgogICAgICAgICAgICAgICAgaXAgPSBzb2Nr
            ZXQuZ2V0YWRkcmluZm8oCiAgICAgICAgICAgICAgICAgICAgaG9zdF9uYW1lLCBwb3J0LCBmYW1p
            bHk9c29ja2V0LkFGX0lORVQsIHByb3RvPXNvY2tldC5JUFBST1RPX1RDUAogICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICBkZWwgc29ja2V0CiAgICAgICAgICAgIHJldHVybiBpcFswXVs0XVsw
            XQogICAgICAgIHJldHVybiBob3N0X25hbWUKCiAgICBkZWYgZ2V0X2FjdGl2ZV9hbmRfc3RhbmRi
            eV9tZ3JzKHNlbGYpOgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAg
            ICAgICAgcmV0dXJuICIiLCBzZWxmLmRyeV9ydW4oImNlcGggc3RhdHVzIikKICAgICAgICBtb25p
            dG9yaW5nX2VuZHBvaW50X3BvcnQgPSBzZWxmLl9hcmdfcGFyc2VyLm1vbml0b3JpbmdfZW5kcG9p
            bnRfcG9ydAogICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdCA9IHNlbGYuX2FyZ19w
            YXJzZXIubW9uaXRvcmluZ19lbmRwb2ludAogICAgICAgIHN0YW5kYnlfbWdycyA9IFtdCiAgICAg
            ICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdDoKICAgICAgICAgICAgY21kX2pz
            b24gPSB7InByZWZpeCI6ICJzdGF0dXMiLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgICAgICBy
            ZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21k
            X2pzb24pCiAgICAgICAgICAgICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQs
            CiAgICAgICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAgICAg
            ICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAg
            ICAgICAgICInbWdyIHNlcnZpY2VzJyBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAgICAg
            ICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09V
            VFBVVF9MSVNUfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRw
            b2ludCA9ICgKICAgICAgICAgICAgICAgIGpzb25fb3V0LmdldCgibWdybWFwIiwge30pLmdldCgi
            c2VydmljZXMiLCB7fSkuZ2V0KCJwcm9tZXRoZXVzIiwgIiIpCiAgICAgICAgICAgICkKICAgICAg
            ICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnQ6CiAgICAgICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJjYW4ndCBmaW5k
            IG1vbml0b3JpbmdfZW5kcG9pbnQsIHByb21ldGhldXMgbW9kdWxlIG1pZ2h0IG5vdCBiZSBlbmFi
            bGVkLCAiCiAgICAgICAgICAgICAgICAgICAgImVuYWJsZSB0aGUgbW9kdWxlIGJ5IHJ1bm5pbmcg
            J2NlcGggbWdyIG1vZHVsZSBlbmFibGUgcHJvbWV0aGV1cyciCiAgICAgICAgICAgICAgICApCiAg
            ICAgICAgICAgICMgbm93IGNoZWNrIHRoZSBzdGFuZC1ieSBtZ3ItcwogICAgICAgICAgICBzdGFu
            ZGJ5X2FyciA9IGpzb25fb3V0LmdldCgibWdybWFwIiwge30pLmdldCgic3RhbmRieXMiLCBbXSkK
            ICAgICAgICAgICAgZm9yIGVhY2hfc3RhbmRieSBpbiBzdGFuZGJ5X2FycjoKICAgICAgICAgICAg
            ICAgIGlmICJuYW1lIiBpbiBlYWNoX3N0YW5kYnkua2V5cygpOgogICAgICAgICAgICAgICAgICAg
            IHN0YW5kYnlfbWdycy5hcHBlbmQoZWFjaF9zdGFuZGJ5WyJuYW1lIl0pCiAgICAgICAgICAgIHRy
            eToKICAgICAgICAgICAgICAgIHBhcnNlZF9lbmRwb2ludCA9IHVybHBhcnNlKG1vbml0b3Jpbmdf
            ZW5kcG9pbnQpCiAgICAgICAgICAgIGV4Y2VwdCBWYWx1ZUVycm9yOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBmImlu
            dmFsaWQgZW5kcG9pbnQ6IHttb25pdG9yaW5nX2VuZHBvaW50fSIKICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0ID0gcGFyc2VkX2VuZHBvaW50
            Lmhvc3RuYW1lCiAgICAgICAgICAgIGlmIG5vdCBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQ6CiAg
            ICAgICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQgPSBzdHIocGFyc2VkX2VuZHBv
            aW50LnBvcnQpCgogICAgICAgICMgaWYgbW9uaXRvcmluZyBlbmRwb2ludCBwb3J0IGlzIG5vdCBz
            ZXQsIHB1dCBhIGRlZmF1bHQgbW9uIHBvcnQKICAgICAgICBpZiBub3QgbW9uaXRvcmluZ19lbmRw
            b2ludF9wb3J0OgogICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQgPSBzZWxmLkRF
            RkFVTFRfTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUCgogICAgICAgICMgdXNlciBjb3VsZCBnaXZl
            IGNvbW1hIGFuZCBzcGFjZSBzZXBhcmF0ZWQgaW5wdXRzIChsaWtlIC0tbW9uaXRvcmluZy1lbmRw
            b2ludD0iPGlwMT4sIDxpcDI+IikKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Qg
            PSBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3QucmVwbGFjZSgiLCIsICIgIikKICAgICAgICBt
            b25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXQgPSBtb25pdG9yaW5nX2VuZHBvaW50X2lw
            X2xpc3Quc3BsaXQoKQogICAgICAgICMgaWYgbW9uaXRvcmluZy1lbmRwb2ludCBjb3VsZCBub3Qg
            YmUgZm91bmQsIHJhaXNlIGFuIGVycm9yCiAgICAgICAgaWYgbGVuKG1vbml0b3JpbmdfZW5kcG9p
            bnRfaXBfbGlzdF9zcGxpdCkgPT0gMDoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVy
            ZUV4Y2VwdGlvbigiTm8gJ21vbml0b3JpbmctZW5kcG9pbnQnIGZvdW5kIikKICAgICAgICAjIGZp
            cnN0IGlwIGlzIHRyZWF0ZWQgYXMgdGhlIG1haW4gbW9uaXRvcmluZy1lbmRwb2ludAogICAgICAg
            IG1vbml0b3JpbmdfZW5kcG9pbnRfaXAgPSBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3Bs
            aXRbMF0KICAgICAgICAjIHJlc3Qgb2YgdGhlIGlwLXMgYXJlIGFkZGVkIHRvIHRoZSAnc3RhbmRi
            eV9tZ3JzJyBsaXN0CiAgICAgICAgc3RhbmRieV9tZ3JzLmV4dGVuZChtb25pdG9yaW5nX2VuZHBv
            aW50X2lwX2xpc3Rfc3BsaXRbMTpdKQogICAgICAgIGZhaWxlZF9pcCA9IG1vbml0b3JpbmdfZW5k
            cG9pbnRfaXAKCiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludCA9ICI6Ii5qb2luKAogICAgICAg
            ICAgICBbbW9uaXRvcmluZ19lbmRwb2ludF9pcCwgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0XQog
            ICAgICAgICkKICAgICAgICBpcF90eXBlID0gc2VsZi5faW52YWxpZF9lbmRwb2ludChtb25pdG9y
            aW5nX2VuZHBvaW50KQogICAgICAgIHRyeToKICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2lu
            dF9pcCA9IHNlbGYuX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoCiAgICAgICAgICAgICAgICBtb25p
            dG9yaW5nX2VuZHBvaW50X2lwLCBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQsIGlwX3R5cGUKICAg
            ICAgICAgICAgKQogICAgICAgICAgICAjIGNvbGxlY3QgYWxsIHRoZSAnc3RhbmQtYnknIG1nciBp
            cHMKICAgICAgICAgICAgbWdyX2lwcyA9IFtdCiAgICAgICAgICAgIGZvciBlYWNoX3N0YW5kYnlf
            bWdyIGluIHN0YW5kYnlfbWdyczoKICAgICAgICAgICAgICAgIGZhaWxlZF9pcCA9IGVhY2hfc3Rh
            bmRieV9tZ3IKICAgICAgICAgICAgICAgIG1ncl9pcHMuYXBwZW5kKAogICAgICAgICAgICAgICAg
            ICAgIHNlbGYuX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoCiAgICAgICAgICAgICAgICAgICAgICAg
            IGVhY2hfc3RhbmRieV9tZ3IsIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydCwgaXBfdHlwZQogICAg
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgICkKICAgICAgICBleGNlcHQ6CiAgICAg
            ICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBm
            IkNvbnZlcnNpb24gb2YgaG9zdDoge2ZhaWxlZF9pcH0gdG8gSVAgZmFpbGVkLiAiCiAgICAgICAg
            ICAgICAgICAiUGxlYXNlIGVudGVyIHRoZSBJUCBhZGRyZXNzZXMgb2YgYWxsIHRoZSBjZXBoLW1n
            cnMgd2l0aCB0aGUgJy0tbW9uaXRvcmluZy1lbmRwb2ludCcgZmxhZyIKICAgICAgICAgICAgKQoK
            ICAgICAgICBfLCBfLCBlcnIgPSBzZWxmLmVuZHBvaW50X2RpYWwobW9uaXRvcmluZ19lbmRwb2lu
            dCwgaXBfdHlwZSkKICAgICAgICBpZiBlcnIgPT0gIi0xIjoKICAgICAgICAgICAgcmFpc2UgRXhl
            Y3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihlcnIpCiAgICAgICAgIyBhZGQgdGhlIHZhbGlkYXRlZCBh
            Y3RpdmUgbWdyIElQIGludG8gdGhlIGZpcnN0IGluZGV4CiAgICAgICAgbWdyX2lwcy5pbnNlcnQo
            MCwgbW9uaXRvcmluZ19lbmRwb2ludF9pcCkKICAgICAgICBhbGxfbWdyX2lwc19zdHIgPSAiLCIu
            am9pbihtZ3JfaXBzKQogICAgICAgIHJldHVybiBhbGxfbWdyX2lwc19zdHIsIG1vbml0b3Jpbmdf
            ZW5kcG9pbnRfcG9ydAoKICAgIGRlZiBjaGVja191c2VyX2V4aXN0KHNlbGYsIHVzZXIpOgogICAg
            ICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAiYXV0aCBnZXQiLCAiZW50aXR5IjogZiJ7dXNlcn0i
            LCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBfID0gc2VsZi5f
            Y29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3Ig
            bGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICByZXR1cm4gIiIKICAgICAgICByZXR1cm4g
            c3RyKGpzb25fb3V0WzBdWyJrZXkiXSkKCiAgICBkZWYgZ2V0X2NlcGhmc19wcm92aXNpb25lcl9j
            YXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gImNsaWVudC5jc2ktY2VwaGZz
            LXByb3Zpc2lvbmVyIgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAgICJtb24iOiAiYWxsb3cg
            ciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLAogICAgICAgICAgICAibWdyIjogImFs
            bG93IHJ3IiwKICAgICAgICAgICAgIm9zZCI6ICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRh
            PSoiLAogICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0
            aF9wZXJtaXNzaW9uOgogICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3Bh
            cnNlci5rOHNfY2x1c3Rlcl9uYW1lCiAgICAgICAgICAgIGlmIGs4c19jbHVzdGVyX25hbWUgPT0g
            IiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAg
            ICAgICAgICAgICAgICAgICJrOHNfY2x1c3Rlcl9uYW1lIG5vdCBmb3VuZCwgcGxlYXNlIHNldCB0
            aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgY2VwaGZzX2ZpbGVzeXN0ZW0gPSBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVt
            X25hbWUKICAgICAgICAgICAgaWYgY2VwaGZzX2ZpbGVzeXN0ZW0gPT0gIiI6CiAgICAgICAgICAg
            ICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9uYW1lfSIKICAgICAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0te2s4c19jbHVzdGVyX25h
            bWV9LXtjZXBoZnNfZmlsZXN5c3RlbX0iCiAgICAgICAgICAgICAgICBjYXBzWyJvc2QiXSA9IGYi
            YWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT17Y2VwaGZzX2ZpbGVzeXN0ZW19IgoKICAgICAg
            ICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVmIGdldF9jZXBoZnNfbm9kZV9jYXBzX2FuZF9l
            bnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiCiAg
            ICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJhbGxvdyByLCBhbGxvdyBjb21tYW5k
            ICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJtZ3IiOiAiYWxsb3cgcnciLAogICAgICAg
            ICAgICAib3NkIjogImFsbG93IHJ3IHRhZyBjZXBoZnMgKj0qIiwKICAgICAgICAgICAgIm1kcyI6
            ICJhbGxvdyBydyIsCiAgICAgICAgfQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJp
            Y3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAgIGs4c19jbHVzdGVyX25hbWUgPSBzZWxm
            Ll9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAgICAgICAgICAgaWYgazhzX2NsdXN0ZXJf
            bmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRp
            b24oCiAgICAgICAgICAgICAgICAgICAgIms4c19jbHVzdGVyX25hbWUgbm90IGZvdW5kLCBwbGVh
            c2Ugc2V0IHRoZSAnLS1rOHMtY2x1c3Rlci1uYW1lJyBmbGFnIgogICAgICAgICAgICAgICAgKQog
            ICAgICAgICAgICBjZXBoZnNfZmlsZXN5c3RlbSA9IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2Zp
            bGVzeXN0ZW1fbmFtZQogICAgICAgICAgICBpZiBjZXBoZnNfZmlsZXN5c3RlbSA9PSAiIjoKICAg
            ICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0te2s4c19jbHVzdGVyX25hbWV9IgogICAg
            ICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17azhzX2Ns
            dXN0ZXJfbmFtZX0te2NlcGhmc19maWxlc3lzdGVtfSIKICAgICAgICAgICAgICAgIGNhcHNbIm9z
            ZCJdID0gZiJhbGxvdyBydyB0YWcgY2VwaGZzICo9e2NlcGhmc19maWxlc3lzdGVtfSIKCiAgICAg
            ICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfZW50aXR5KAogICAgICAgIHNlbGYs
            CiAgICAgICAgZW50aXR5LAogICAgICAgIHJiZF9wb29sX25hbWUsCiAgICAgICAgYWxpYXNfcmJk
            X3Bvb2xfbmFtZSwKICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lLAogICAgICAgIHJhZG9zX25hbWVz
            cGFjZSwKICAgICk6CiAgICAgICAgaWYgKAogICAgICAgICAgICByYmRfcG9vbF9uYW1lLmNvdW50
            KCIuIikgIT0gMAogICAgICAgICAgICBvciByYmRfcG9vbF9uYW1lLmNvdW50KCJfIikgIT0gMAog
            ICAgICAgICAgICBvciBhbGlhc19yYmRfcG9vbF9uYW1lICE9ICIiCiAgICAgICAgICAgICMgY2hl
            Y2tpbmcgYWxpYXNfcmJkX3Bvb2xfbmFtZSBpcyBub3QgZW1wdHkgYXMgdGhlcmUgbWF5YmUgYSBz
            cGVjaWFsIGNoYXJhY3RlciB1c2VkIG90aGVyIHRoYW4gLiBvciBfCiAgICAgICAgKToKICAgICAg
            ICAgICAgaWYgYWxpYXNfcmJkX3Bvb2xfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgInBsZWFzZSBz
            ZXQgdGhlICctLWFsaWFzLXJiZC1kYXRhLXBvb2wtbmFtZScgZmxhZyBhcyB0aGUgcmJkIGRhdGEg
            cG9vbCBuYW1lIGNvbnRhaW5zICcuJyBvciAnXyciCiAgICAgICAgICAgICAgICApCiAgICAgICAg
            ICAgIGlmICgKICAgICAgICAgICAgICAgIGFsaWFzX3JiZF9wb29sX25hbWUuY291bnQoIi4iKSAh
            PSAwCiAgICAgICAgICAgICAgICBvciBhbGlhc19yYmRfcG9vbF9uYW1lLmNvdW50KCJfIikgIT0g
            MAogICAgICAgICAgICApOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJy0tYWxpYXMtcmJkLWRhdGEtcG9vbC1uYW1l
            JyBmbGFnIHZhbHVlIHNob3VsZCBub3QgY29udGFpbiAnLicgb3IgJ18nIgogICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9uYW1lfS17
            YWxpYXNfcmJkX3Bvb2xfbmFtZX0iCiAgICAgICAgZWxzZToKICAgICAgICAgICAgZW50aXR5ID0g
            ZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0te3JiZF9wb29sX25hbWV9IgoKICAgICAgICBp
            ZiByYWRvc19uYW1lc3BhY2U6CiAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0te3JhZG9z
            X25hbWVzcGFjZX0iCiAgICAgICAgcmV0dXJuIGVudGl0eQoKICAgIGRlZiBnZXRfcmJkX3Byb3Zp
            c2lvbmVyX2NhcHNfYW5kX2VudGl0eShzZWxmKToKICAgICAgICBlbnRpdHkgPSAiY2xpZW50LmNz
            aS1yYmQtcHJvdmlzaW9uZXIiCiAgICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJw
            cm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLAogICAgICAgICAgICAi
            bWdyIjogImFsbG93IHJ3IiwKICAgICAgICAgICAgIm9zZCI6ICJwcm9maWxlIHJiZCIsCiAgICAg
            ICAgfQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Np
            b246CiAgICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRh
            X3Bvb2xfbmFtZQogICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3Bh
            cnNlci5hbGlhc19yYmRfZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFt
            ZSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQogICAgICAgICAgICByYWRvc19u
            YW1lc3BhY2UgPSBzZWxmLl9hcmdfcGFyc2VyLnJhZG9zX25hbWVzcGFjZQogICAgICAgICAgICBp
            ZiByYmRfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAibWFuZGF0b3J5IGZsYWcgbm90IGZv
            dW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1yYmQtZGF0YS1wb29sLW5hbWUnIGZsYWciCiAgICAgICAg
            ICAgICAgICApCiAgICAgICAgICAgIGlmIGs4c19jbHVzdGVyX25hbWUgPT0gIiI6CiAgICAgICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLWs4cy1jbHVz
            dGVyLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVudGl0eSA9IHNl
            bGYuZ2V0X2VudGl0eSgKICAgICAgICAgICAgICAgIGVudGl0eSwKICAgICAgICAgICAgICAgIHJi
            ZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lLAogICAgICAg
            ICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAgICAgICAgICAgIHJhZG9zX25hbWVzcGFj
            ZSwKICAgICAgICAgICAgKQogICAgICAgICAgICBpZiByYWRvc19uYW1lc3BhY2UgIT0gIiI6CiAg
            ICAgICAgICAgICAgICBjYXBzWyJvc2QiXSA9ICgKICAgICAgICAgICAgICAgICAgICBmInByb2Zp
            bGUgcmJkIHBvb2w9e3JiZF9wb29sX25hbWV9IG5hbWVzcGFjZT17cmFkb3NfbmFtZXNwYWNlfSIK
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGNhcHNb
            Im9zZCJdID0gZiJwcm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9uYW1lfSIKCiAgICAgICAgcmV0
            dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfcmJkX25vZGVfY2Fwc19hbmRfZW50aXR5KHNl
            bGYpOgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3NpLXJiZC1ub2RlIgogICAgICAgIGNhcHMg
            PSB7CiAgICAgICAgICAgICJtb24iOiAicHJvZmlsZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBi
            bG9ja2xpc3QnIiwKICAgICAgICAgICAgIm9zZCI6ICJwcm9maWxlIHJiZCIsCiAgICAgICAgfQog
            ICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb246CiAg
            ICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xf
            bmFtZQogICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5h
            bGlhc19yYmRfZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNl
            bGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQogICAgICAgICAgICByYWRvc19uYW1lc3Bh
            Y2UgPSBzZWxmLl9hcmdfcGFyc2VyLnJhZG9zX25hbWVzcGFjZQogICAgICAgICAgICBpZiByYmRf
            cG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAibWFuZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBw
            bGVhc2Ugc2V0IHRoZSAnLS1yYmQtZGF0YS1wb29sLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAg
            ICApCiAgICAgICAgICAgIGlmIGs4c19jbHVzdGVyX25hbWUgPT0gIiI6CiAgICAgICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJt
            YW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLWs4cy1jbHVzdGVyLW5h
            bWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVudGl0eSA9IHNlbGYuZ2V0
            X2VudGl0eSgKICAgICAgICAgICAgICAgIGVudGl0eSwKICAgICAgICAgICAgICAgIHJiZF9wb29s
            X25hbWUsCiAgICAgICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lLAogICAgICAgICAgICAg
            ICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAgICAgICAgICAgIHJhZG9zX25hbWVzcGFjZSwKICAg
            ICAgICAgICAgKQogICAgICAgICAgICBpZiByYWRvc19uYW1lc3BhY2UgIT0gIiI6CiAgICAgICAg
            ICAgICAgICBjYXBzWyJvc2QiXSA9ICgKICAgICAgICAgICAgICAgICAgICBmInByb2ZpbGUgcmJk
            IHBvb2w9e3JiZF9wb29sX25hbWV9IG5hbWVzcGFjZT17cmFkb3NfbmFtZXNwYWNlfSIKICAgICAg
            ICAgICAgICAgICkKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJd
            ID0gZiJwcm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9uYW1lfSIKCiAgICAgICAgcmV0dXJuIGNh
            cHMsIGVudGl0eQoKICAgIGRlZiBnZXRfZGVmYXVsdFVzZXJfY2Fwc19hbmRfZW50aXR5KHNlbGYp
            OgogICAgICAgIGVudGl0eSA9IHNlbGYucnVuX2FzX3VzZXIKICAgICAgICBjYXBzID0gewogICAg
            ICAgICAgICAibW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxs
            b3cgY29tbWFuZCB2ZXJzaW9uIiwKICAgICAgICAgICAgIm1nciI6ICJhbGxvdyBjb21tYW5kIGNv
            bmZpZyIsCiAgICAgICAgICAgICJvc2QiOiBmInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cg
            cnd4IHBvb2w9e3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cubWV0YSwgYWxs
            b3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD17c2VsZi5fYXJnX3BhcnNlci5yZ3df
            cG9vbF9wcmVmaXh9LnJndy5jb250cm9sLCBhbGxvdyByeCBwb29sPXtzZWxmLl9hcmdfcGFyc2Vy
            LnJnd19wb29sX3ByZWZpeH0ucmd3LmxvZywgYWxsb3cgeCBwb29sPXtzZWxmLl9hcmdfcGFyc2Vy
            LnJnd19wb29sX3ByZWZpeH0ucmd3LmJ1Y2tldHMuaW5kZXgiLAogICAgICAgIH0KCiAgICAgICAg
            cmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfY2Fwc19hbmRfZW50aXR5KHNlbGYsIHVz
            ZXJfbmFtZSk6CiAgICAgICAgaWYgImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiBpbiB1
            c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIg
            IT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVk
            X2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZ2V0X2NlcGhm
            c19wcm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoKQogICAgICAgIGlmICJjbGllbnQuY3NpLWNl
            cGhmcy1ub2RlIiBpbiB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQuY3NpLWNlcGhm
            cy1ub2RlIiAhPSB1c2VyX25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJl
            c3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5n
            ZXRfY2VwaGZzX25vZGVfY2Fwc19hbmRfZW50aXR5KCkKICAgICAgICBpZiAiY2xpZW50LmNzaS1y
            YmQtcHJvdmlzaW9uZXIiIGluIHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVudC5jc2kt
            cmJkLXByb3Zpc2lvbmVyIiAhPSB1c2VyX25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdf
            cGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1
            cm4gc2VsZi5nZXRfcmJkX3Byb3Zpc2lvbmVyX2NhcHNfYW5kX2VudGl0eSgpCiAgICAgICAgaWYg
            ImNsaWVudC5jc2ktcmJkLW5vZGUiIGluIHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVu
            dC5jc2ktcmJkLW5vZGUiICE9IHVzZXJfbmFtZToKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19w
            YXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVlCiAgICAgICAgICAgIHJldHVy
            biBzZWxmLmdldF9yYmRfbm9kZV9jYXBzX2FuZF9lbnRpdHkoKQogICAgICAgIGlmICJjbGllbnQu
            aGVhbHRoY2hlY2tlciIgaW4gdXNlcl9uYW1lOgogICAgICAgICAgICBpZiAiY2xpZW50LmhlYWx0
            aGNoZWNrZXIiICE9IHVzZXJfbmFtZToKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIu
            cmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVlCiAgICAgICAgICAgIHJldHVybiBzZWxm
            LmdldF9kZWZhdWx0VXNlcl9jYXBzX2FuZF9lbnRpdHkoKQoKICAgICAgICByYWlzZSBFeGVjdXRp
            b25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICBmIm5vIHVzZXIgZm91bmQgd2l0aCB1c2Vy
            X25hbWU6IHt1c2VyX25hbWV9LCAiCiAgICAgICAgICAgICJnZXRfY2Fwc19hbmRfZW50aXR5IGNv
            bW1hbmQgZmFpbGVkLlxuIgogICAgICAgICkKCiAgICBkZWYgY3JlYXRlX2NlcGhDU0lLZXlyaW5n
            X3VzZXIoc2VsZiwgdXNlcik6CiAgICAgICAgIiIiCiAgICAgICAgY29tbWFuZDogY2VwaCBhdXRo
            IGdldC1vci1jcmVhdGUgY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIgbW9uICdhbGxvdyBy
            JyBtZ3IgJ2FsbG93IHJ3JyBvc2QgJ2FsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KicKICAg
            ICAgICAiIiIKICAgICAgICBjYXBzLCBlbnRpdHkgPSBzZWxmLmdldF9jYXBzX2FuZF9lbnRpdHko
            dXNlcikKICAgICAgICBjbWRfanNvbiA9IHsKICAgICAgICAgICAgInByZWZpeCI6ICJhdXRoIGdl
            dC1vci1jcmVhdGUiLAogICAgICAgICAgICAiZW50aXR5IjogZW50aXR5LAogICAgICAgICAgICAi
            Y2FwcyI6IFtjYXAgZm9yIGNhcF9saXN0IGluIGxpc3QoY2Fwcy5pdGVtcygpKSBmb3IgY2FwIGlu
            IGNhcF9saXN0XSwKICAgICAgICAgICAgImZvcm1hdCI6ICJqc29uIiwKICAgICAgICB9CgogICAg
            ICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuICgKICAg
            ICAgICAgICAgICAgIHNlbGYuZHJ5X3J1bigKICAgICAgICAgICAgICAgICAgICAiY2VwaCAiCiAg
            ICAgICAgICAgICAgICAgICAgKyBjbWRfanNvblsicHJlZml4Il0KICAgICAgICAgICAgICAgICAg
            ICArICIgIgogICAgICAgICAgICAgICAgICAgICsgY21kX2pzb25bImVudGl0eSJdCiAgICAgICAg
            ICAgICAgICAgICAgKyAiICIKICAgICAgICAgICAgICAgICAgICArICIgIi5qb2luKGNtZF9qc29u
            WyJjYXBzIl0pCiAgICAgICAgICAgICAgICApLAogICAgICAgICAgICAgICAgIiIsCiAgICAgICAg
            ICAgICkKICAgICAgICAjIGNoZWNrIGlmIHVzZXIgYWxyZWFkeSBleGlzdAogICAgICAgIHVzZXJf
            a2V5ID0gc2VsZi5jaGVja191c2VyX2V4aXN0KGVudGl0eSkKICAgICAgICBpZiB1c2VyX2tleSAh
            PSAiIjoKICAgICAgICAgICAgcmV0dXJuIHVzZXJfa2V5LCBmIntlbnRpdHkuc3BsaXQoJy4nLCAx
            KVsxXX0iCiAgICAgICAgICAgICMgZW50aXR5LnNwbGl0KCcuJywxKVsxXSB0byByZW5hbWUgZW50
            aXR5KGNsaWVudC5jc2ktcmJkLW5vZGUpIGFzIGNzaS1yYmQtbm9kZQoKICAgICAgICByZXRfdmFs
            LCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24p
            CiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwKICAgICAgICBp
            ZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiInYXV0aCBnZXQtb3It
            Y3JlYXRlIHt1c2VyfScgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAgICAgICAgICAgICBmIkVycm9y
            OiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIK
            ICAgICAgICAgICAgKQogICAgICAgIHJldHVybiBzdHIoanNvbl9vdXRbMF1bImtleSJdKSwgZiJ7
            ZW50aXR5LnNwbGl0KCcuJywgMSlbMV19IgogICAgICAgICMgZW50aXR5LnNwbGl0KCcuJywxKVsx
            XSB0byByZW5hbWUgZW50aXR5KGNsaWVudC5jc2ktcmJkLW5vZGUpIGFzIGNzaS1yYmQtbm9kZQoK
            ICAgIGRlZiBnZXRfY2VwaGZzX2RhdGFfcG9vbF9kZXRhaWxzKHNlbGYpOgogICAgICAgIGNtZF9q
            c29uID0geyJwcmVmaXgiOiAiZnMgbHMiLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIGlmIHNl
            bGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigi
            Y2VwaCAiICsgY21kX2pzb25bInByZWZpeCJdKQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBl
            cnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICAjIGlm
            IHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LCByZXBvcnQgYW4gZXJyb3IKICAgICAg
            ICBpZiByZXRfdmFsICE9IDA6CiAgICAgICAgICAgICMgaWYgZnMgYW5kIGRhdGFfcG9vbCBhcmd1
            bWVudHMgYXJlIG5vdCBzZXQsIHNpbGVudGx5IHJldHVybgogICAgICAgICAgICBpZiAoCiAgICAg
            ICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUgPT0gIiIK
            ICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFt
            ZSA9PSAiIgogICAgICAgICAgICApOgogICAgICAgICAgICAgICAgcmV0dXJuCiAgICAgICAgICAg
            ICMgaWYgdXNlciBoYXMgcHJvdmlkZWQgYW55IG9mIHRoZQogICAgICAgICAgICAjICctLWNlcGhm
            cy1maWxlc3lzdGVtLW5hbWUnIG9yICctLWNlcGhmcy1kYXRhLXBvb2wtbmFtZScgYXJndW1lbnRz
            LAogICAgICAgICAgICAjIHJhaXNlIGFuIGV4Y2VwdGlvbiBhcyB3ZSBhcmUgdW5hYmxlIHRvIHZl
            cmlmeSB0aGUgYXJncwogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KAogICAgICAgICAgICAgICAgZiInZnMgbHMnIGNlcGggY2FsbCBmYWlsZWQgd2l0aCBlcnJvcjog
            e2Vycl9tc2d9IgogICAgICAgICAgICApCgogICAgICAgIG1hdGNoaW5nX2pzb25fb3V0ID0ge30K
            ICAgICAgICAjIGlmICctLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUnIGFyZ3VtZW50IGlzIHByb3Zp
            ZGVkLAogICAgICAgICMgY2hlY2sgd2hldGhlciB0aGUgcHJvdmlkZWQgZmlsZXN5c3RlbS1uYW1l
            IGV4aXN0cyBvciBub3QKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lz
            dGVtX25hbWU6CiAgICAgICAgICAgICMgZ2V0IHRoZSBtYXRjaGluZyBsaXN0CiAgICAgICAgICAg
            IG1hdGNoaW5nX2pzb25fb3V0X2xpc3QgPSBbCiAgICAgICAgICAgICAgICBtYXRjaGVkCiAgICAg
            ICAgICAgICAgICBmb3IgbWF0Y2hlZCBpbiBqc29uX291dAogICAgICAgICAgICAgICAgaWYgc3Ry
            KG1hdGNoZWRbIm5hbWUiXSkgPT0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9u
            YW1lCiAgICAgICAgICAgIF0KICAgICAgICAgICAgIyB1bmFibGUgdG8gZmluZCBhIG1hdGNoaW5n
            IGZzLW5hbWUsIHJhaXNlIGFuIGVycm9yCiAgICAgICAgICAgIGlmIGxlbihtYXRjaGluZ19qc29u
            X291dF9saXN0KSA9PSAwOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBmIkZpbGVzeXN0ZW0gcHJvdmlkZWQsICd7c2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lfScsICIKICAgICAgICAgICAgICAg
            ICAgICBmImlzIG5vdCBmb3VuZCBpbiB0aGUgZnMtbGlzdDoge1tzdHIoeFsnbmFtZSddKSBmb3Ig
            eCBpbiBqc29uX291dF19IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBtYXRjaGluZ19q
            c29uX291dCA9IG1hdGNoaW5nX2pzb25fb3V0X2xpc3RbMF0KICAgICAgICAjIGlmIGNlcGhmcyBm
            aWxlc3lzdGVtIG5hbWUgaXMgbm90IHByb3ZpZGVkLAogICAgICAgICMgdHJ5IHRvIGdldCBhIGRl
            ZmF1bHQgZnMgbmFtZSBieSBkb2luZyB0aGUgZm9sbG93aW5nCiAgICAgICAgZWxzZToKICAgICAg
            ICAgICAgIyBhLiBjaGVjayBpZiB0aGVyZSBpcyBvbmx5IG9uZSBmaWxlc3lzdGVtIGlzIHByZXNl
            bnQKICAgICAgICAgICAgaWYgbGVuKGpzb25fb3V0KSA9PSAxOgogICAgICAgICAgICAgICAgbWF0
            Y2hpbmdfanNvbl9vdXQgPSBqc29uX291dFswXQogICAgICAgICAgICAjIGIuIG9yIGVsc2UsIGNo
            ZWNrIGlmIGRhdGFfcG9vbCBuYW1lIGlzIHByb3ZpZGVkCiAgICAgICAgICAgIGVsaWYgc2VsZi5f
            YXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWU6CiAgICAgICAgICAgICAgICAjIGFuZCBp
            ZiBwcmVzZW50LCBjaGVjayB3aGV0aGVyIHRoZXJlIGV4aXN0cyBhIGZzIHdoaWNoIGhhcyB0aGUg
            ZGF0YV9wb29sCiAgICAgICAgICAgICAgICBmb3IgZWFjaEogaW4ganNvbl9vdXQ6CiAgICAgICAg
            ICAgICAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUgaW4g
            ZWFjaEpbImRhdGFfcG9vbHMiXToKICAgICAgICAgICAgICAgICAgICAgICAgbWF0Y2hpbmdfanNv
            bl9vdXQgPSBlYWNoSgogICAgICAgICAgICAgICAgICAgICAgICBicmVhawogICAgICAgICAgICAg
            ICAgIyBpZiB0aGVyZSBpcyBubyBtYXRjaGluZyBmcyBleGlzdHMsIHRoYXQgbWVhbnMgcHJvdmlk
            ZWQgZGF0YV9wb29sIG5hbWUgaXMgaW52YWxpZAogICAgICAgICAgICAgICAgaWYgbm90IG1hdGNo
            aW5nX2pzb25fb3V0OgogICAgICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVF
            eGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgICAgIGYiUHJvdmlkZWQgZGF0YV9wb29sIG5h
            bWUsIHtzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZX0sIgogICAgICAgICAg
            ICAgICAgICAgICAgICAiIGRvZXMgbm90IGV4aXN0cyIKICAgICAgICAgICAgICAgICAgICApCiAg
            ICAgICAgICAgICMgYy4gaWYgbm90aGluZyBpcyBzZXQgYW5kIGNvdWxkbid0IGZpbmQgYSBkZWZh
            dWx0LAogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgIyBqdXN0IHJldHVybiBzaWxl
            bnRseQogICAgICAgICAgICAgICAgcmV0dXJuCgogICAgICAgIGlmIG1hdGNoaW5nX2pzb25fb3V0
            OgogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUgPSBz
            dHIobWF0Y2hpbmdfanNvbl9vdXRbIm5hbWUiXSkKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5jZXBoZnNfbWV0YWRhdGFfcG9vbF9uYW1lID0gc3RyKAogICAgICAgICAgICAgICAgbWF0Y2hp
            bmdfanNvbl9vdXRbIm1ldGFkYXRhX3Bvb2wiXQogICAgICAgICAgICApCgogICAgICAgIGlmIGlz
            aW5zdGFuY2UobWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXSwgbGlzdCk6CiAgICAgICAg
            ICAgICMgaWYgdGhlIHVzZXIgaGFzIGFscmVhZHkgcHJvdmlkZWQgZGF0YS1wb29sLW5hbWUsCiAg
            ICAgICAgICAgICMgdGhyb3VnaCAtLWNlcGhmcy1kYXRhLXBvb2wtbmFtZQogICAgICAgICAgICBp
            ZiBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZToKICAgICAgICAgICAgICAg
            ICMgaWYgdGhlIHByb3ZpZGVkIG5hbWUgaXMgbm90IG1hdGNoaW5nIHdpdGggdGhlIG9uZSBpbiB0
            aGUgbGlzdAogICAgICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2Fy
            Z19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgICAgICAgICAgbm90IGlu
            IG1hdGNoaW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0KICAgICAgICAgICAgICAgICk6CiAgICAg
            ICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAg
            ICAgICAgICAgICAgICAgZiJQcm92aWRlZCBkYXRhLXBvb2wtbmFtZTogJ3tzZWxmLl9hcmdfcGFy
            c2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZX0nLCAiCiAgICAgICAgICAgICAgICAgICAgICAgICJk
            b2Vzbid0IG1hdGNoIGZyb20gdGhlIGRhdGEtcG9vbHMgbGlzdDogIgogICAgICAgICAgICAgICAg
            ICAgICAgICBmIntbc3RyKHgpIGZvciB4IGluIG1hdGNoaW5nX2pzb25fb3V0WydkYXRhX3Bvb2xz
            J11dfSIKICAgICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICMgaWYgZGF0YV9wb29sIG5h
            bWUgaXMgbm90IHByb3ZpZGVkLAogICAgICAgICAgICAjIHRoZW4gdHJ5IHRvIGZpbmQgYSBkZWZh
            dWx0IGRhdGEgcG9vbCBuYW1lCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICAjIGlm
            IG5vIGRhdGFfcG9vbHMgZXhpc3QsIHNpbGVudGx5IHJldHVybgogICAgICAgICAgICAgICAgaWYg
            bGVuKG1hdGNoaW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0pID09IDA6CiAgICAgICAgICAgICAg
            ICAgICAgcmV0dXJuCiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRh
            X3Bvb2xfbmFtZSA9IHN0cigKICAgICAgICAgICAgICAgICAgICBtYXRjaGluZ19qc29uX291dFsi
            ZGF0YV9wb29scyJdWzBdCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICMgaWYgdGhlcmUg
            YXJlIG1vcmUgdGhhbiBvbmUgJ2RhdGFfcG9vbHMnIGV4aXN0LAogICAgICAgICAgICAjIHRoZW4g
            d2FybiB0aGUgdXNlciB0aGF0IHdlIGFyZSB1c2luZyB0aGUgc2VsZWN0ZWQgbmFtZQogICAgICAg
            ICAgICBpZiBsZW4obWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXSkgPiAxOgogICAgICAg
            ICAgICAgICAgcHJpbnQoCiAgICAgICAgICAgICAgICAgICAgIldBUk5JTkc6IE11bHRpcGxlIGRh
            dGEgcG9vbHMgZGV0ZWN0ZWQ6ICIKICAgICAgICAgICAgICAgICAgICBmIntbc3RyKHgpIGZvciB4
            IGluIG1hdGNoaW5nX2pzb25fb3V0WydkYXRhX3Bvb2xzJ11dfVxuIgogICAgICAgICAgICAgICAg
            ICAgIGYiVXNpbmcgdGhlIGRhdGEtcG9vbDogJ3tzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRh
            X3Bvb2xfbmFtZX0nXG4iCiAgICAgICAgICAgICAgICApCgogICAgZGVmIGNyZWF0ZV9jaGVja2Vy
            S2V5KHNlbGYsIHVzZXIpOgogICAgICAgIGNhcHMsIGVudGl0eSA9IHNlbGYuZ2V0X2NhcHNfYW5k
            X2VudGl0eSh1c2VyKQogICAgICAgIGNtZF9qc29uID0gewogICAgICAgICAgICAicHJlZml4Ijog
            ImF1dGggZ2V0LW9yLWNyZWF0ZSIsCiAgICAgICAgICAgICJlbnRpdHkiOiBlbnRpdHksCiAgICAg
            ICAgICAgICJjYXBzIjogW2NhcCBmb3IgY2FwX2xpc3QgaW4gbGlzdChjYXBzLml0ZW1zKCkpIGZv
            ciBjYXAgaW4gY2FwX2xpc3RdLAogICAgICAgICAgICAiZm9ybWF0IjogImpzb24iLAogICAgICAg
            IH0KCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1
            cm4gc2VsZi5kcnlfcnVuKAogICAgICAgICAgICAgICAgImNlcGggIgogICAgICAgICAgICAgICAg
            KyBjbWRfanNvblsicHJlZml4Il0KICAgICAgICAgICAgICAgICsgIiAiCiAgICAgICAgICAgICAg
            ICArIGNtZF9qc29uWyJlbnRpdHkiXQogICAgICAgICAgICAgICAgKyAiICIKICAgICAgICAgICAg
            ICAgICsgIiAiLmpvaW4oY21kX2pzb25bImNhcHMiXSkKICAgICAgICAgICAgKQogICAgICAgICMg
            Y2hlY2sgaWYgdXNlciBhbHJlYWR5IGV4aXN0CiAgICAgICAgdXNlcl9rZXkgPSBzZWxmLmNoZWNr
            X3VzZXJfZXhpc3QoZW50aXR5KQogICAgICAgIGlmIHVzZXJfa2V5ICE9ICIiOgogICAgICAgICAg
            ICByZXR1cm4gdXNlcl9rZXkKCiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBz
            ZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICMgaWYgdGhlcmUgaXMg
            YW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihq
            c29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgICAgIGYiJ2F1dGggZ2V0LW9yLWNyZWF0ZSB7c2VsZi5ydW5fYXNfdXNl
            cn0nIGNvbW1hbmQgZmFpbGVkXG4iCiAgICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBp
            ZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAg
            KQogICAgICAgIHJldHVybiBzdHIoanNvbl9vdXRbMF1bImtleSJdKQoKICAgIGRlZiBnZXRfY2Vw
            aF9kYXNoYm9hcmRfbGluayhzZWxmKToKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4IjogIm1n
            ciBzZXJ2aWNlcyIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNl
            ci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKCJjZXBoICIgKyBjbWRf
            anNvblsicHJlZml4Il0pCiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIF8gPSBzZWxmLl9jb21t
            b25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNj
            ZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkg
            PT0gMDoKICAgICAgICAgICAgcmV0dXJuIE5vbmUKICAgICAgICBpZiAiZGFzaGJvYXJkIiBub3Qg
            aW4ganNvbl9vdXQ6CiAgICAgICAgICAgIHJldHVybiBOb25lCiAgICAgICAgcmV0dXJuIGpzb25f
            b3V0WyJkYXNoYm9hcmQiXQoKICAgIGRlZiBjcmVhdGVfcmd3X2FkbWluX29wc191c2VyKHNlbGYp
            OgogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAgICAgICAg
            ICAidXNlciIsCiAgICAgICAgICAgICJjcmVhdGUiLAogICAgICAgICAgICAiLS11aWQiLAogICAg
            ICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9PUFNfVVNFUl9OQU1FLAogICAgICAgICAg
            ICAiLS1kaXNwbGF5LW5hbWUiLAogICAgICAgICAgICAiUm9vayBSR1cgQWRtaW4gT3BzIHVzZXIi
            LAogICAgICAgICAgICAiLS1jYXBzIiwKICAgICAgICAgICAgImJ1Y2tldHM9Kjt1c2Vycz0qO3Vz
            YWdlPXJlYWQ7bWV0YWRhdGE9cmVhZDt6b25lPXJlYWQiLAogICAgICAgICAgICAiLS1yZ3ctcmVh
            bG0iLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lLAogICAgICAg
            ICAgICAiLS1yZ3ctem9uZWdyb3VwIiwKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3df
            em9uZWdyb3VwX25hbWUsCiAgICAgICAgICAgICItLXJndy16b25lIiwKICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1lLAogICAgICAgIF0KICAgICAgICBpZiBzZWxmLl9h
            cmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGgg
            IiArICIgIi5qb2luKGNtZCkpCiAgICAgICAgdHJ5OgogICAgICAgICAgICBvdXRwdXQgPSBzdWJw
            cm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAg
            ZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yIGFzIGV4ZWNFcnI6CiAgICAgICAg
            ICAgICMgaWYgdGhlIHVzZXIgYWxyZWFkeSBleGlzdHMsIHdlIGp1c3QgcXVlcnkgaXQKICAgICAg
            ICAgICAgaWYgZXhlY0Vyci5yZXR1cm5jb2RlID09IGVycm5vLkVFWElTVDoKICAgICAgICAgICAg
            ICAgIGNtZCA9IFsKICAgICAgICAgICAgICAgICAgICAicmFkb3Nndy1hZG1pbiIsCiAgICAgICAg
            ICAgICAgICAgICAgInVzZXIiLAogICAgICAgICAgICAgICAgICAgICJpbmZvIiwKICAgICAgICAg
            ICAgICAgICAgICAiLS11aWQiLAogICAgICAgICAgICAgICAgICAgIHNlbGYuRVhURVJOQUxfUkdX
            X0FETUlOX09QU19VU0VSX05BTUUsCiAgICAgICAgICAgICAgICAgICAgIi0tcmd3LXJlYWxtIiwK
            ICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lLAogICAg
            ICAgICAgICAgICAgICAgICItLXJndy16b25lZ3JvdXAiLAogICAgICAgICAgICAgICAgICAgIHNl
            bGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lLAogICAgICAgICAgICAgICAgICAgICIt
            LXJndy16b25lIiwKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25l
            X25hbWUsCiAgICAgICAgICAgICAgICBdCiAgICAgICAgICAgICAgICB0cnk6CiAgICAgICAgICAg
            ICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3Vi
            cHJvY2Vzcy5QSVBFKQogICAgICAgICAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJv
            Y2Vzc0Vycm9yIGFzIGV4ZWNFcnI6CiAgICAgICAgICAgICAgICAgICAgZXJyX21zZyA9ICgKICAg
            ICAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5kIHtjbWR9LiBP
            dXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICAgICAgZiJDb2Rl
            OiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtleGVjRXJyLnN0ZGVycn0iCiAgICAgICAg
            ICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21z
            ZykKICAgICAgICAgICAgICAgICAgICByZXR1cm4gTm9uZSwgTm9uZSwgRmFsc2UsICItMSIKICAg
            ICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVycl9tc2cgPSAoCiAgICAgICAgICAgICAg
            ICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5kIHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJy
            Lm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICBmIkNvZGU6IHtleGVjRXJyLnJldHVybmNv
            ZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykKICAgICAgICAgICAgICAgIHJldHVybiBO
            b25lLCBOb25lLCBGYWxzZSwgIi0xIgoKICAgICAgICAjIGlmIGl0IGlzIHB5dGhvbjIsIGRvbid0
            IGNoZWNrIGZvciBjZXBoIHZlcnNpb24gZm9yIGFkZGluZyBgaW5mbz1yZWFkYCBjYXAocmd3X3Zh
            bGlkYXRpb24pCiAgICAgICAgaWYgc3lzLnZlcnNpb25faW5mby5tYWpvciA8IDM6CiAgICAgICAg
            ICAgIGpzb25vdXRwdXQgPSBqc29uLmxvYWRzKG91dHB1dCkKICAgICAgICAgICAgcmV0dXJuICgK
            ICAgICAgICAgICAgICAgIGpzb25vdXRwdXRbImtleXMiXVswXVsiYWNjZXNzX2tleSJdLAogICAg
            ICAgICAgICAgICAganNvbm91dHB1dFsia2V5cyJdWzBdWyJzZWNyZXRfa2V5Il0sCiAgICAgICAg
            ICAgICAgICBGYWxzZSwKICAgICAgICAgICAgICAgICIiLAogICAgICAgICAgICApCgogICAgICAg
            ICMgc2VwYXJhdGVseSBhZGQgaW5mbz1yZWFkIGNhcHMgZm9yIHJndy1lbmRwb2ludCBpcCB2YWxp
            ZGF0aW9uCiAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkID0gVHJ1ZQogICAgICAgIGNtZCA9IFsK
            ICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAgICAgICAgICAiY2FwcyIsCiAgICAgICAg
            ICAgICJhZGQiLAogICAgICAgICAgICAiLS11aWQiLAogICAgICAgICAgICBzZWxmLkVYVEVSTkFM
            X1JHV19BRE1JTl9PUFNfVVNFUl9OQU1FLAogICAgICAgICAgICAiLS1jYXBzIiwKICAgICAgICAg
            ICAgImluZm89cmVhZCIsCiAgICAgICAgICAgICItLXJndy1yZWFsbSIsCiAgICAgICAgICAgIHNl
            bGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUsCiAgICAgICAgICAgICItLXJndy16b25lZ3Jv
            dXAiLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSwKICAg
            ICAgICAgICAgIi0tcmd3LXpvbmUiLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196
            b25lX25hbWUsCiAgICAgICAgXQogICAgICAgIHRyeToKICAgICAgICAgICAgb3V0cHV0ID0gc3Vi
            cHJvY2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAg
            IGV4Y2VwdCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJvciBhcyBleGVjRXJyOgogICAgICAg
            ICAgICAjIGlmIHRoZSBjZXBoIHZlcnNpb24gbm90IHN1cHBvcnRlZCBmb3IgYWRkaW5nIGBpbmZv
            PXJlYWRgIGNhcChyZ3dfdmFsaWRhdGlvbikKICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAg
            ICAgImNvdWxkIG5vdCBhZGQgY2FwczogdW5hYmxlIHRvIGFkZCBjYXBzOiBpbmZvPXJlYWRcbiIK
            ICAgICAgICAgICAgICAgIGluIGV4ZWNFcnIuc3RkZXJyLmRlY29kZSgidXRmLTgiKQogICAgICAg
            ICAgICAgICAgYW5kIGV4ZWNFcnIucmV0dXJuY29kZSA9PSAyNDQKICAgICAgICAgICAgKToKICAg
            ICAgICAgICAgICAgIGluZm9fY2FwX3N1cHBvcnRlZCA9IEZhbHNlCiAgICAgICAgICAgIGVsc2U6
            CiAgICAgICAgICAgICAgICBlcnJfbXNnID0gKAogICAgICAgICAgICAgICAgICAgIGYiZmFpbGVk
            IHRvIGV4ZWN1dGUgY29tbWFuZCB7Y21kfS4gT3V0cHV0OiB7ZXhlY0Vyci5vdXRwdXR9LiAiCiAg
            ICAgICAgICAgICAgICAgICAgZiJDb2RlOiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtl
            eGVjRXJyLnN0ZGVycn0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICBzeXMuc3Rk
            ZXJyLndyaXRlKGVycl9tc2cpCiAgICAgICAgICAgICAgICByZXR1cm4gTm9uZSwgTm9uZSwgRmFs
            c2UsICItMSIKCiAgICAgICAganNvbm91dHB1dCA9IGpzb24ubG9hZHMob3V0cHV0KQogICAgICAg
            IHJldHVybiAoCiAgICAgICAgICAgIGpzb25vdXRwdXRbImtleXMiXVswXVsiYWNjZXNzX2tleSJd
            LAogICAgICAgICAgICBqc29ub3V0cHV0WyJrZXlzIl1bMF1bInNlY3JldF9rZXkiXSwKICAgICAg
            ICAgICAgaW5mb19jYXBfc3VwcG9ydGVkLAogICAgICAgICAgICAiIiwKICAgICAgICApCgogICAg
            ZGVmIHZhbGlkYXRlX3JiZF9wb29sKHNlbGYsIHBvb2xfbmFtZSk6CiAgICAgICAgaWYgbm90IHNl
            bGYuY2x1c3Rlci5wb29sX2V4aXN0cyhwb29sX25hbWUpOgogICAgICAgICAgICByYWlzZSBFeGVj
            dXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcG9v
            bCwgJ3twb29sX25hbWV9JywgZG9lcyBub3QgZXhpc3QiCiAgICAgICAgICAgICkKCiAgICBkZWYg
            aW5pdF9yYmRfcG9vbChzZWxmLCByYmRfcG9vbF9uYW1lKToKICAgICAgICBpZiBpc2luc3RhbmNl
            KHNlbGYuY2x1c3RlciwgRHVtbXlSYWRvcyk6CiAgICAgICAgICAgIHJldHVybgogICAgICAgIGlv
            Y3R4ID0gc2VsZi5jbHVzdGVyLm9wZW5faW9jdHgocmJkX3Bvb2xfbmFtZSkKICAgICAgICByYmRf
            aW5zdCA9IHJiZC5SQkQoKQogICAgICAgIHJiZF9pbnN0LnBvb2xfaW5pdChpb2N0eCwgVHJ1ZSkK
            CiAgICBkZWYgdmFsaWRhdGVfcmFkb3NfbmFtZXNwYWNlKHNlbGYpOgogICAgICAgIHJiZF9wb29s
            X25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgIHJhZG9z
            X25hbWVzcGFjZSA9IHNlbGYuX2FyZ19wYXJzZXIucmFkb3NfbmFtZXNwYWNlCiAgICAgICAgaWYg
            cmFkb3NfbmFtZXNwYWNlID09ICIiOgogICAgICAgICAgICByZXR1cm4KICAgICAgICBpZiByYWRv
            c19uYW1lc3BhY2UuaXNsb3dlcigpID09IEZhbHNlOgogICAgICAgICAgICByYWlzZSBFeGVjdXRp
            b25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmFkb3Mg
            TmFtZXNwYWNlLCAne3JhZG9zX25hbWVzcGFjZX0nLCAiCiAgICAgICAgICAgICAgICBmImNvbnRh
            aW5zIHVwcGVyIGNhc2UiCiAgICAgICAgICAgICkKICAgICAgICByYmRfaW5zdCA9IHJiZC5SQkQo
            KQogICAgICAgIGlvY3R4ID0gc2VsZi5jbHVzdGVyLm9wZW5faW9jdHgocmJkX3Bvb2xfbmFtZSkK
            ICAgICAgICBpZiByYmRfaW5zdC5uYW1lc3BhY2VfZXhpc3RzKGlvY3R4LCByYWRvc19uYW1lc3Bh
            Y2UpIGlzIEZhbHNlOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmFkb3MgTmFtZXNwYWNlLCAne3JhZG9z
            X25hbWVzcGFjZX0nLCAiCiAgICAgICAgICAgICAgICBmImlzIG5vdCBmb3VuZCBpbiB0aGUgcG9v
            bCAne3JiZF9wb29sX25hbWV9JyIKICAgICAgICAgICAgKQoKICAgIGRlZiBnZXRfb3JfY3JlYXRl
            X3N1YnZvbHVtZV9ncm91cChzZWxmLCBzdWJ2b2x1bWVfZ3JvdXAsIGNlcGhmc19maWxlc3lzdGVt
            X25hbWUpOgogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgImNlcGgiLAogICAgICAgICAgICAi
            ZnMiLAogICAgICAgICAgICAic3Vidm9sdW1lZ3JvdXAiLAogICAgICAgICAgICAiZ2V0cGF0aCIs
            CiAgICAgICAgICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgIHN1YnZvbHVt
            ZV9ncm91cCwKICAgICAgICBdCiAgICAgICAgdHJ5OgogICAgICAgICAgICBfID0gc3VicHJvY2Vz
            cy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2Vw
            dCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJvcjoKICAgICAgICAgICAgY21kID0gWwogICAg
            ICAgICAgICAgICAgImNlcGgiLAogICAgICAgICAgICAgICAgImZzIiwKICAgICAgICAgICAgICAg
            ICJzdWJ2b2x1bWVncm91cCIsCiAgICAgICAgICAgICAgICAiY3JlYXRlIiwKICAgICAgICAgICAg
            ICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgICAgICBzdWJ2b2x1bWVfZ3Jv
            dXAsCiAgICAgICAgICAgIF0KICAgICAgICAgICAgdHJ5OgogICAgICAgICAgICAgICAgXyA9IHN1
            YnByb2Nlc3MuY2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAgICAg
            ICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBm
            InN1YnZvbHVtZSBncm91cCB7c3Vidm9sdW1lX2dyb3VwfSBpcyBub3QgYWJsZSB0byBnZXQgY3Jl
            YXRlZCIKICAgICAgICAgICAgICAgICkKCiAgICBkZWYgcGluX3N1YnZvbHVtZSgKICAgICAgICBz
            ZWxmLCBzdWJ2b2x1bWVfZ3JvdXAsIGNlcGhmc19maWxlc3lzdGVtX25hbWUsIHBpbl90eXBlLCBw
            aW5fc2V0dGluZwogICAgKToKICAgICAgICBjbWQgPSBbCiAgICAgICAgICAgICJjZXBoIiwKICAg
            ICAgICAgICAgImZzIiwKICAgICAgICAgICAgInN1YnZvbHVtZWdyb3VwIiwKICAgICAgICAgICAg
            InBpbiIsCiAgICAgICAgICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgIHN1
            YnZvbHVtZV9ncm91cCwKICAgICAgICAgICAgcGluX3R5cGUsCiAgICAgICAgICAgIHBpbl9zZXR0
            aW5nLAogICAgICAgIF0KICAgICAgICB0cnk6CiAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNo
            ZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1
            YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJzdWJ2b2x1bWUgZ3JvdXAge3N1YnZv
            bHVtZV9ncm91cH0gaXMgbm90IGFibGUgdG8gZ2V0IHBpbm5lZCIKICAgICAgICAgICAgKQoKICAg
            IGRlZiBnZXRfcmd3X2ZzaWQoc2VsZiwgYmFzZV91cmwsIHZlcmlmeSk6CiAgICAgICAgYWNjZXNz
            X2tleSA9IHNlbGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX0FDQ0VTU19LRVkiXQogICAg
            ICAgIHNlY3JldF9rZXkgPSBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9TRUNSRVRf
            S0VZIl0KICAgICAgICByZ3dfZW5kcG9pbnQgPSBzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2lu
            dAogICAgICAgIGJhc2VfdXJsID0gYmFzZV91cmwgKyAiOi8vIiArIHJnd19lbmRwb2ludCArICIv
            YWRtaW4vaW5mbz8iCiAgICAgICAgcGFyYW1zID0geyJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAg
            cmVxdWVzdF91cmwgPSBiYXNlX3VybCArIHVybGVuY29kZShwYXJhbXMpCgogICAgICAgIHRyeToK
            ICAgICAgICAgICAgciA9IHJlcXVlc3RzLmdldCgKICAgICAgICAgICAgICAgIHJlcXVlc3RfdXJs
            LAogICAgICAgICAgICAgICAgYXV0aD1TM0F1dGgoYWNjZXNzX2tleSwgc2VjcmV0X2tleSwgcmd3
            X2VuZHBvaW50KSwKICAgICAgICAgICAgICAgIHZlcmlmeT12ZXJpZnksCiAgICAgICAgICAgICkK
            ICAgICAgICBleGNlcHQgcmVxdWVzdHMuZXhjZXB0aW9ucy5UaW1lb3V0OgogICAgICAgICAgICBz
            eXMuc3RkZXJyLndyaXRlKAogICAgICAgICAgICAgICAgZiJpbnZhbGlkIGVuZHBvaW50Oiwgbm90
            IGFibGUgdG8gY2FsbCBhZG1pbi1vcHMgYXBpe3Jnd19lbmRwb2ludH0iCiAgICAgICAgICAgICkK
            ICAgICAgICAgICAgcmV0dXJuICIiLCAiLTEiCiAgICAgICAgcjEgPSByLmpzb24oKQogICAgICAg
            IGlmIHIxIGlzIE5vbmUgb3IgcjEuZ2V0KCJpbmZvIikgaXMgTm9uZToKICAgICAgICAgICAgc3lz
            LnN0ZGVyci53cml0ZSgKICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHJndyBFbmRwb2lu
            dCwgJ3tzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludH0nLCBpcyBpbnZhbGlkLiIKICAgICAg
            ICAgICAgKQogICAgICAgICAgICByZXR1cm4gKAogICAgICAgICAgICAgICAgIiIsCiAgICAgICAg
            ICAgICAgICAiLTEiLAogICAgICAgICAgICApCgogICAgICAgIHJldHVybiByMVsiaW5mbyJdWyJz
            dG9yYWdlX2JhY2tlbmRzIl1bMF1bImNsdXN0ZXJfaWQiXSwgIiIKCiAgICBkZWYgdmFsaWRhdGVf
            cmd3X2VuZHBvaW50KHNlbGYsIGluZm9fY2FwX3N1cHBvcnRlZCk6CiAgICAgICAgIyBpZiB0aGUg
            J2NsdXN0ZXInIGluc3RhbmNlIGlzIGEgZHVtbXkgb25lLAogICAgICAgICMgZG9uJ3QgdHJ5IHRv
            IHJlYWNoIG91dCB0byB0aGUgZW5kcG9pbnQKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1
            c3RlciwgRHVtbXlSYWRvcyk6CiAgICAgICAgICAgIHJldHVybgoKICAgICAgICByZ3dfZW5kcG9p
            bnQgPSBzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludAoKICAgICAgICAjIHZhbGlkYXRlIHJn
            dyBlbmRwb2ludCBvbmx5IGlmIGlwIGFkZHJlc3MgaXMgcGFzc2VkCiAgICAgICAgaXBfdHlwZSA9
            IHNlbGYuX2ludmFsaWRfZW5kcG9pbnQocmd3X2VuZHBvaW50KQoKICAgICAgICAjIGNoZWNrIGlm
            IHRoZSByZ3cgZW5kcG9pbnQgaXMgcmVhY2hhYmxlCiAgICAgICAgY2VydCA9IE5vbmUKICAgICAg
            ICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNlci5yZ3dfc2tpcF90bHMgYW5kIHNlbGYudmFsaWRhdGVf
            cmd3X2VuZHBvaW50X3Rsc19jZXJ0KCk6CiAgICAgICAgICAgIGNlcnQgPSBzZWxmLl9hcmdfcGFy
            c2VyLnJnd190bHNfY2VydF9wYXRoCiAgICAgICAgYmFzZV91cmwsIHZlcmlmeSwgZXJyID0gc2Vs
            Zi5lbmRwb2ludF9kaWFsKHJnd19lbmRwb2ludCwgaXBfdHlwZSwgY2VydD1jZXJ0KQogICAgICAg
            IGlmIGVyciAhPSAiIjoKICAgICAgICAgICAgcmV0dXJuICItMSIKCiAgICAgICAgIyBjaGVjayBp
            ZiB0aGUgcmd3IGVuZHBvaW50IGJlbG9uZ3MgdG8gdGhlIHNhbWUgY2x1c3RlcgogICAgICAgICMg
            b25seSBjaGVjayBpZiBgaW5mb2AgY2FwIGlzIHN1cHBvcnRlZAogICAgICAgIGlmIGluZm9fY2Fw
            X3N1cHBvcnRlZDoKICAgICAgICAgICAgZnNpZCA9IHNlbGYuZ2V0X2ZzaWQoKQogICAgICAgICAg
            ICByZ3dfZnNpZCwgZXJyID0gc2VsZi5nZXRfcmd3X2ZzaWQoYmFzZV91cmwsIHZlcmlmeSkKICAg
            ICAgICAgICAgaWYgZXJyID09ICItMSI6CiAgICAgICAgICAgICAgICByZXR1cm4gIi0xIgogICAg
            ICAgICAgICBpZiBmc2lkICE9IHJnd19mc2lkOgogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53
            cml0ZSgKICAgICAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCByZ3cgRW5kcG9pbnQsICd7
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfZW5kcG9pbnR9JywgaXMgaW52YWxpZC4gV2UgYXJlIHZhbGlk
            YXRpbmcgYnkgY2FsbGluZyB0aGUgYWRtaW5vcHMgYXBpIHRocm91Z2ggcmd3LWVuZHBvaW50IGFu
            ZCB2YWxpZGF0aW5nIHRoZSBjbHVzdGVyX2lkICd7cmd3X2ZzaWR9JyBpcyBlcXVhbCB0byB0aGUg
            Y2VwaCBjbHVzdGVyIGZzaWQgJ3tmc2lkfSciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAg
            ICAgICByZXR1cm4gIi0xIgoKICAgICAgICAjIGNoZWNrIGlmIHRoZSByZ3cgZW5kcG9pbnQgcG9v
            bCBleGlzdAogICAgICAgICMgb25seSB2YWxpZGF0ZSBpZiByZ3dfcG9vbF9wcmVmaXggaXMgcGFz
            c2VkIGVsc2UgaXQgd2lsbCB0YWtlIGRlZmF1bHQgdmFsdWUgYW5kIHdlIGRvbid0IGNyZWF0ZSB0
            aGVzZSBkZWZhdWx0IHBvb2xzCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9w
            cmVmaXggIT0gImRlZmF1bHQiOgogICAgICAgICAgICByZ3dfcG9vbHNfdG9fdmFsaWRhdGUgPSBb
            CiAgICAgICAgICAgICAgICBmIntzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3
            Lm1ldGEiLAogICAgICAgICAgICAgICAgIi5yZ3cucm9vdCIsCiAgICAgICAgICAgICAgICBmIntz
            ZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmNvbnRyb2wiLAogICAgICAgICAg
            ICAgICAgZiJ7c2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5sb2ciLAogICAg
            ICAgICAgICBdCiAgICAgICAgICAgIGZvciBfcmd3X3Bvb2xfdG9fdmFsaWRhdGUgaW4gcmd3X3Bv
            b2xzX3RvX3ZhbGlkYXRlOgogICAgICAgICAgICAgICAgaWYgbm90IHNlbGYuY2x1c3Rlci5wb29s
            X2V4aXN0cyhfcmd3X3Bvb2xfdG9fdmFsaWRhdGUpOgogICAgICAgICAgICAgICAgICAgIHN5cy5z
            dGRlcnIud3JpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHBvb2ws
            ICd7X3Jnd19wb29sX3RvX3ZhbGlkYXRlfScsIGRvZXMgbm90IGV4aXN0IgogICAgICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgICAgICAgICByZXR1cm4gIi0xIgoKICAgICAgICByZXR1cm4g
            IiIKCiAgICBkZWYgdmFsaWRhdGVfcmd3X211bHRpc2l0ZShzZWxmLCByZ3dfbXVsdGlzaXRlX2Nv
            bmZpZ19uYW1lLCByZ3dfbXVsdGlzaXRlX2NvbmZpZyk6CiAgICAgICAgaWYgcmd3X211bHRpc2l0
            ZV9jb25maWcgIT0gIiI6CiAgICAgICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgICAgICJyYWRv
            c2d3LWFkbWluIiwKICAgICAgICAgICAgICAgIHJnd19tdWx0aXNpdGVfY29uZmlnLAogICAgICAg
            ICAgICAgICAgImdldCIsCiAgICAgICAgICAgICAgICAiLS1yZ3ctIiArIHJnd19tdWx0aXNpdGVf
            Y29uZmlnLAogICAgICAgICAgICAgICAgcmd3X211bHRpc2l0ZV9jb25maWdfbmFtZSwKICAgICAg
            ICAgICAgXQogICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBfID0gc3VicHJvY2Vzcy5j
            aGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAgICBleGNl
            cHQgc3VicHJvY2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3IgYXMgZXhlY0VycjoKICAgICAgICAgICAg
            ICAgIGVycl9tc2cgPSAoCiAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBj
            b21tYW5kIHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAgICAgICAgICAg
            ICAgICBmIkNvZGU6IHtleGVjRXJyLnJldHVybmNvZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJy
            fSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJy
            X21zZykKICAgICAgICAgICAgICAgIHJldHVybiAiLTEiCiAgICAgICAgcmV0dXJuICIiCgogICAg
            ZGVmIGNvbnZlcnRfY29tbWFfc2VwYXJhdGVkX3RvX2FycmF5KHNlbGYsIHZhbHVlKToKICAgICAg
            ICByZXR1cm4gdmFsdWUuc3BsaXQoIiwiKQoKICAgIGRlZiByYWlzZV9leGNlcHRpb25faWZfYW55
            X3RvcG9sb2d5X2ZsYWdfaXNfbWlzc2luZyhzZWxmKToKICAgICAgICBpZiAoCiAgICAgICAgICAg
            ICgKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMgIT0gIiIK
            ICAgICAgICAgICAgICAgIGFuZCAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCA9PSAiIgogICAgICAgICAgICAgICAgICAg
            IG9yIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfZmFpbHVyZV9kb21haW5fdmFsdWVzID09ICIi
            CiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICkKICAgICAgICAgICAgb3IgKAogICAgICAg
            ICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCAh
            PSAiIgogICAgICAgICAgICAgICAgYW5kICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdf
            cGFyc2VyLnRvcG9sb2d5X3Bvb2xzID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5f
            YXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMgPT0gIiIKICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgKQogICAgICAgICAgICBvciAoCiAgICAgICAgICAgICAgICBz
            ZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX3ZhbHVlcyAhPSAiIgogICAg
            ICAgICAgICAgICAgYW5kICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRv
            cG9sb2d5X3Bvb2xzID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5fYXJnX3BhcnNl
            ci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCA9PSAiIgogICAgICAgICAgICAgICAgKQog
            ICAgICAgICAgICApCiAgICAgICAgKToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVy
            ZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICJwcm92aWRlIGFsbCB0aGUgdG9wb2xvZ3kgZmxh
            Z3MgLS10b3BvbG9neS1wb29scywgLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi1sYWJlbCwgLS10
            b3BvbG9neS1mYWlsdXJlLWRvbWFpbi12YWx1ZXMiCiAgICAgICAgICAgICkKCiAgICBkZWYgdmFs
            aWRhdGVfdG9wb2xvZ3lfdmFsdWVzKHNlbGYsIHRvcG9sb2d5X3Bvb2xzLCB0b3BvbG9neV9mZCk6
            CiAgICAgICAgaWYgbGVuKHRvcG9sb2d5X3Bvb2xzKSAhPSBsZW4odG9wb2xvZ3lfZmQpOgogICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ZiJUaGUgcHJvdmlkZWQgdG9wb2xvZ3kgcG9vbHMsICd7dG9wb2xvZ3lfcG9vbHN9JywgYW5kICIK
            ICAgICAgICAgICAgICAgIGYidG9wb2xvZ3kgZmFpbHVyZSBkb21haW4sICd7dG9wb2xvZ3lfZmR9
            JywiCiAgICAgICAgICAgICAgICBmImFyZSBvZiBkaWZmZXJlbnQgbGVuZ3RoLCAne2xlbih0b3Bv
            bG9neV9wb29scyl9JyBhbmQgJ3tsZW4odG9wb2xvZ3lfZmQpfScgcmVzcGN0aXZlbHkiCiAgICAg
            ICAgICAgICkKICAgICAgICByZXR1cm4KCiAgICBkZWYgdmFsaWRhdGVfdG9wb2xvZ3lfcmJkX3Bv
            b2xzKHNlbGYsIHRvcG9sb2d5X3JiZF9wb29scyk6CiAgICAgICAgZm9yIHBvb2wgaW4gdG9wb2xv
            Z3lfcmJkX3Bvb2xzOgogICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3JiZF9wb29sKHBvb2wpCgog
            ICAgZGVmIGluaXRfdG9wb2xvZ3lfcmJkX3Bvb2xzKHNlbGYsIHRvcG9sb2d5X3JiZF9wb29scyk6
            CiAgICAgICAgZm9yIHBvb2wgaW4gdG9wb2xvZ3lfcmJkX3Bvb2xzOgogICAgICAgICAgICBzZWxm
            LmluaXRfcmJkX3Bvb2wocG9vbCkKCiAgICBkZWYgX2dlbl9vdXRwdXRfbWFwKHNlbGYpOgogICAg
            ICAgIGlmIHNlbGYub3V0X21hcDoKICAgICAgICAgICAgcmV0dXJuCiAgICAgICAgIyBzdXBwb3J0
            IGxlZ2FjeSBmbGFnIHdpdGggdXBncmFkZXMKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNs
            dXN0ZXJfbmFtZToKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1l
            ID0gc2VsZi5fYXJnX3BhcnNlci5jbHVzdGVyX25hbWUKICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            Lms4c19jbHVzdGVyX25hbWUgPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuazhzX2Ns
            dXN0ZXJfbmFtZS5sb3dlcigpCiAgICAgICAgKSAgIyBhbHdheXMgY29udmVydCBjbHVzdGVyIG5h
            bWUgdG8gbG93ZXJjYXNlIGNoYXJhY3RlcnMKICAgICAgICBzZWxmLnZhbGlkYXRlX3JiZF9wb29s
            KHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lKQogICAgICAgIHNlbGYuaW5pdF9y
            YmRfcG9vbChzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZSkKICAgICAgICBzZWxm
            LnZhbGlkYXRlX3JhZG9zX25hbWVzcGFjZSgpCiAgICAgICAgc2VsZi5fZXhjbHVkZWRfa2V5cy5h
            ZGQoIks4U19DTFVTVEVSX05BTUUiKQogICAgICAgIHNlbGYuZ2V0X2NlcGhmc19kYXRhX3Bvb2xf
            ZGV0YWlscygpCiAgICAgICAgc2VsZi5vdXRfbWFwWyJOQU1FU1BBQ0UiXSA9IHNlbGYuX2FyZ19w
            YXJzZXIubmFtZXNwYWNlCiAgICAgICAgc2VsZi5vdXRfbWFwWyJLOFNfQ0xVU1RFUl9OQU1FIl0g
            PSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBb
            IlJPT0tfRVhURVJOQUxfRlNJRCJdID0gc2VsZi5nZXRfZnNpZCgpCiAgICAgICAgc2VsZi5vdXRf
            bWFwWyJST09LX0VYVEVSTkFMX1VTRVJOQU1FIl0gPSBzZWxmLnJ1bl9hc191c2VyCiAgICAgICAg
            c2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0NFUEhfTU9OX0RBVEEiXSA9IHNlbGYuZ2V0X2Nl
            cGhfZXh0ZXJuYWxfbW9uX2RhdGEoKQogICAgICAgIHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5B
            TF9VU0VSX1NFQ1JFVCJdID0gc2VsZi5jcmVhdGVfY2hlY2tlcktleSgKICAgICAgICAgICAgImNs
            aWVudC5oZWFsdGhjaGVja2VyIgogICAgICAgICkKICAgICAgICBzZWxmLm91dF9tYXBbIlJPT0tf
            RVhURVJOQUxfREFTSEJPQVJEX0xJTksiXSA9IHNlbGYuZ2V0X2NlcGhfZGFzaGJvYXJkX2xpbmso
            KQogICAgICAgICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VDUkVU
            Il0sCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FIl0s
            CiAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3VzZXIoImNsaWVudC5jc2kt
            cmJkLW5vZGUiKQogICAgICAgICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BS
            T1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9SQkRfUFJP
            VklTSU9ORVJfU0VDUkVUX05BTUUiXSwKICAgICAgICApID0gc2VsZi5jcmVhdGVfY2VwaENTSUtl
            eXJpbmdfdXNlcigiY2xpZW50LmNzaS1yYmQtcHJvdmlzaW9uZXIiKQogICAgICAgIHNlbGYub3V0
            X21hcFsiQ0VQSEZTX1BPT0xfTkFNRSJdID0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9w
            b29sX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBbIkNFUEhGU19NRVRBREFUQV9QT09MX05BTUUi
            XSA9ICgKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfbWV0YWRhdGFfcG9vbF9u
            YW1lCiAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSA9IHNl
            bGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgIHNlbGYub3V0X21h
            cFsiUkVTVFJJQ1RFRF9BVVRIX1BFUk1JU1NJT04iXSA9ICgKICAgICAgICAgICAgc2VsZi5fYXJn
            X3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbgogICAgICAgICkKICAgICAgICBzZWxm
            Lm91dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdID0gc2VsZi5fYXJnX3BhcnNlci5yYWRvc19uYW1l
            c3BhY2UKICAgICAgICBzZWxmLm91dF9tYXBbIlNVQlZPTFVNRV9HUk9VUCJdID0gc2VsZi5fYXJn
            X3BhcnNlci5zdWJ2b2x1bWVfZ3JvdXAKICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNf
            Tk9ERV9TRUNSRVQiXSA9ICIiCiAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJ
            U0lPTkVSX1NFQ1JFVCJdID0gIiIKICAgICAgICAjIGNyZWF0ZSBDZXBoRlMgbm9kZSBhbmQgcHJv
            dmlzaW9uZXIga2V5cmluZyBvbmx5IHdoZW4gTURTIGV4aXN0cwogICAgICAgIGlmIHNlbGYub3V0
            X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSBhbmQgc2VsZi5vdXRfbWFwWyJDRVBIRlNfUE9PTF9OQU1F
            Il06CiAgICAgICAgICAgICgKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhG
            U19OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZT
            X05PREVfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lL
            ZXlyaW5nX3VzZXIoImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiKQogICAgICAgICAgICAoCiAgICAg
            ICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUIl0s
            CiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VD
            UkVUX05BTUUiXSwKICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3Vz
            ZXIoImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIikKICAgICAgICAgICAgIyBjcmVhdGUg
            dGhlIGRlZmF1bHQgImNzaSIgc3Vidm9sdW1lZ3JvdXAKICAgICAgICAgICAgc2VsZi5nZXRfb3Jf
            Y3JlYXRlX3N1YnZvbHVtZV9ncm91cCgKICAgICAgICAgICAgICAgICJjc2kiLCBzZWxmLl9hcmdf
            cGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICAgICAgKQogICAgICAgICAgICAj
            IHBpbiB0aGUgZGVmYXVsdCAiY3NpIiBzdWJ2b2x1bWVncm91cAogICAgICAgICAgICBzZWxmLnBp
            bl9zdWJ2b2x1bWUoCiAgICAgICAgICAgICAgICAiY3NpIiwgc2VsZi5fYXJnX3BhcnNlci5jZXBo
            ZnNfZmlsZXN5c3RlbV9uYW1lLCAiZGlzdHJpYnV0ZWQiLCAiMSIKICAgICAgICAgICAgKQogICAg
            ICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlNVQlZPTFVNRV9HUk9VUCJdOgogICAgICAgICAgICAg
            ICAgc2VsZi5nZXRfb3JfY3JlYXRlX3N1YnZvbHVtZV9ncm91cCgKICAgICAgICAgICAgICAgICAg
            ICBzZWxmLl9hcmdfcGFyc2VyLnN1YnZvbHVtZV9ncm91cCwKICAgICAgICAgICAgICAgICAgICBz
            ZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgICAgICBzZWxmLnBpbl9zdWJ2b2x1bWUoCiAgICAgICAgICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci5zdWJ2b2x1bWVfZ3JvdXAsCiAgICAgICAgICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lLAogICAgICAgICAgICAgICAgICAg
            ICJkaXN0cmlidXRlZCIsCiAgICAgICAgICAgICAgICAgICAgIjEiLAogICAgICAgICAgICAgICAg
            KQogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1RMU19DRVJUIl0gPSAiIgogICAgICAgIHNlbGYu
            b3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdID0gIiIKICAgICAgICBzZWxmLm91dF9tYXBb
            Ik1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJdID0gIiIKICAgICAgICBpZiBub3Qgc2VsZi5fYXJn
            X3BhcnNlci5za2lwX21vbml0b3JpbmdfZW5kcG9pbnQ6CiAgICAgICAgICAgICgKICAgICAgICAg
            ICAgICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdLAogICAgICAgICAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UX1BPUlQiXSwKICAgICAgICAgICAg
            KSA9IHNlbGYuZ2V0X2FjdGl2ZV9hbmRfc3RhbmRieV9tZ3JzKCkKICAgICAgICBzZWxmLm91dF9t
            YXBbIlJCRF9QT09MX05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1l
            CiAgICAgICAgc2VsZi5vdXRfbWFwWyJSQkRfTUVUQURBVEFfRUNfUE9PTF9OQU1FIl0gPSAoCiAg
            ICAgICAgICAgIHNlbGYudmFsaWRhdGVfcmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZSgpCiAgICAg
            ICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9PTFMiXSA9IHNlbGYuX2FyZ19w
            YXJzZXIudG9wb2xvZ3lfcG9vbHMKICAgICAgICBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxV
            UkVfRE9NQUlOX0xBQkVMIl0gPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xv
            Z3lfZmFpbHVyZV9kb21haW5fbGFiZWwKICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJU
            T1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiXSA9ICgKICAgICAgICAgICAgc2VsZi5fYXJn
            X3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMKICAgICAgICApCiAgICAgICAg
            aWYgKAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X3Bvb2xzICE9ICIiCiAg
            ICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX2xh
            YmVsICE9ICIiCiAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1
            cmVfZG9tYWluX3ZhbHVlcyAhPSAiIgogICAgICAgICk6CiAgICAgICAgICAgIHNlbGYudmFsaWRh
            dGVfdG9wb2xvZ3lfdmFsdWVzKAogICAgICAgICAgICAgICAgc2VsZi5jb252ZXJ0X2NvbW1hX3Nl
            cGFyYXRlZF90b19hcnJheShzZWxmLm91dF9tYXBbIlRPUE9MT0dZX1BPT0xTIl0pLAogICAgICAg
            ICAgICAgICAgc2VsZi5jb252ZXJ0X2NvbW1hX3NlcGFyYXRlZF90b19hcnJheSgKICAgICAgICAg
            ICAgICAgICAgICBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX1ZBTFVFUyJd
            CiAgICAgICAgICAgICAgICApLAogICAgICAgICAgICApCiAgICAgICAgICAgIHNlbGYudmFsaWRh
            dGVfdG9wb2xvZ3lfcmJkX3Bvb2xzKAogICAgICAgICAgICAgICAgc2VsZi5jb252ZXJ0X2NvbW1h
            X3NlcGFyYXRlZF90b19hcnJheShzZWxmLm91dF9tYXBbIlRPUE9MT0dZX1BPT0xTIl0pCiAgICAg
            ICAgICAgICkKICAgICAgICAgICAgc2VsZi5pbml0X3RvcG9sb2d5X3JiZF9wb29scygKICAgICAg
            ICAgICAgICAgIHNlbGYuY29udmVydF9jb21tYV9zZXBhcmF0ZWRfdG9fYXJyYXkoc2VsZi5vdXRf
            bWFwWyJUT1BPTE9HWV9QT09MUyJdKQogICAgICAgICAgICApCiAgICAgICAgZWxzZToKICAgICAg
            ICAgICAgc2VsZi5yYWlzZV9leGNlcHRpb25faWZfYW55X3RvcG9sb2d5X2ZsYWdfaXNfbWlzc2lu
            ZygpCgogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1BPT0xfUFJFRklYIl0gPSBzZWxmLl9hcmdf
            cGFyc2VyLnJnd19wb29sX3ByZWZpeAogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX0VORFBPSU5U
            Il0gPSAiIgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50OgogICAgICAg
            ICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgICAgICBzZWxmLmNy
            ZWF0ZV9yZ3dfYWRtaW5fb3BzX3VzZXIoKQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAg
            ICAgaWYgKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25h
            bWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9u
            ZWdyb3VwX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfem9uZV9uYW1lICE9ICIiCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAg
            ICAgIGVyciA9IHNlbGYudmFsaWRhdGVfcmd3X211bHRpc2l0ZSgKICAgICAgICAgICAgICAgICAg
            ICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwgInJlYWxtIgogICAgICAgICAg
            ICAgICAgICAgICkKICAgICAgICAgICAgICAgICAgICBlcnIgPSBzZWxmLnZhbGlkYXRlX3Jnd19t
            dWx0aXNpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pv
            bmVncm91cF9uYW1lLCAiem9uZWdyb3VwIgogICAgICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgICAgICAgICBlcnIgPSBzZWxmLnZhbGlkYXRlX3Jnd19tdWx0aXNpdGUoCiAgICAgICAgICAg
            ICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVfbmFtZSwgInpvbmUiCiAgICAg
            ICAgICAgICAgICAgICAgKQoKICAgICAgICAgICAgICAgIGlmICgKICAgICAgICAgICAgICAgICAg
            ICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lID09ICIiCiAgICAgICAgICAgICAgICAg
            ICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lID09ICIiCiAgICAgICAg
            ICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVfbmFtZSA9PSAiIgogICAg
            ICAgICAgICAgICAgKSBvciAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5y
            Z3dfcmVhbG1fbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFy
            c2VyLnJnd196b25lZ3JvdXBfbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd196b25lX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICk6CiAgICAg
            ICAgICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJH
            V19BRE1JTl9PUFNfVVNFUl9BQ0NFU1NfS0VZIl0sCiAgICAgICAgICAgICAgICAgICAgICAgIHNl
            bGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX1NFQ1JFVF9LRVkiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkLAogICAgICAgICAgICAgICAgICAgICAgICBl
            cnIsCiAgICAgICAgICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRlX3Jnd19hZG1pbl9vcHNfdXNl
            cigpCiAgICAgICAgICAgICAgICAgICAgZXJyID0gc2VsZi52YWxpZGF0ZV9yZ3dfZW5kcG9pbnQo
            aW5mb19jYXBfc3VwcG9ydGVkKQogICAgICAgICAgICAgICAgICAgIGlmIHNlbGYuX2FyZ19wYXJz
            ZXIucmd3X3Rsc19jZXJ0X3BhdGg6CiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYub3V0X21h
            cFsiUkdXX1RMU19DRVJUIl0gPSAoCiAgICAgICAgICAgICAgICAgICAgICAgICAgICBzZWxmLnZh
            bGlkYXRlX3Jnd19lbmRwb2ludF90bHNfY2VydCgpCiAgICAgICAgICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgICAgICAgICAjIGlmIHRoZXJlIGlzIG5vIGVycm9yLCBzZXQgdGhlIFJHV19F
            TkRQT0lOVAogICAgICAgICAgICAgICAgICAgIGlmIGVyciAhPSAiLTEiOgogICAgICAgICAgICAg
            ICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdID0gc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfZW5kcG9pbnQKICAgICAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICAgICAg
            ZXJyID0gIlBsZWFzZSBwcm92aWRlIGFsbCB0aGUgUkdXIG11bHRpc2l0ZSBwYXJhbWV0ZXJzIG9y
            IG5vbmUgb2YgdGhlbSIKICAgICAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKGVycikK
            CiAgICBkZWYgZ2VuX3NoZWxsX291dChzZWxmKToKICAgICAgICBzZWxmLl9nZW5fb3V0cHV0X21h
            cCgpCiAgICAgICAgc2hPdXRJTyA9IFN0cmluZ0lPKCkKICAgICAgICBmb3IgaywgdiBpbiBzZWxm
            Lm91dF9tYXAuaXRlbXMoKToKICAgICAgICAgICAgaWYgdiBhbmQgayBub3QgaW4gc2VsZi5fZXhj
            bHVkZWRfa2V5czoKICAgICAgICAgICAgICAgIHNoT3V0SU8ud3JpdGUoZiJleHBvcnQge2t9PXt2
            fXtMSU5FU0VQfSIpCiAgICAgICAgc2hPdXQgPSBzaE91dElPLmdldHZhbHVlKCkKICAgICAgICBz
            aE91dElPLmNsb3NlKCkKICAgICAgICByZXR1cm4gc2hPdXQKCiAgICBkZWYgZ2VuX2pzb25fb3V0
            KHNlbGYpOgogICAgICAgIHNlbGYuX2dlbl9vdXRwdXRfbWFwKCkKICAgICAgICBpZiBzZWxmLl9h
            cmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiAiIgogICAgICAgIGpzb25fb3V0
            ID0gWwogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFtZSI6ICJyb29rLWNlcGgtbW9u
            LWVuZHBvaW50cyIsCiAgICAgICAgICAgICAgICAia2luZCI6ICJDb25maWdNYXAiLAogICAgICAg
            ICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiBzZWxmLm91dF9t
            YXBbIlJPT0tfRVhURVJOQUxfQ0VQSF9NT05fREFUQSJdLAogICAgICAgICAgICAgICAgICAgICJt
            YXhNb25JZCI6ICIwIiwKICAgICAgICAgICAgICAgICAgICAibWFwcGluZyI6ICJ7fSIsCiAgICAg
            ICAgICAgICAgICB9LAogICAgICAgICAgICB9LAogICAgICAgICAgICB7CiAgICAgICAgICAgICAg
            ICAibmFtZSI6ICJyb29rLWNlcGgtbW9uIiwKICAgICAgICAgICAgICAgICJraW5kIjogIlNlY3Jl
            dCIsCiAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAiYWRtaW4t
            c2VjcmV0IjogImFkbWluLXNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImZzaWQiOiBzZWxm
            Lm91dF9tYXBbIlJPT0tfRVhURVJOQUxfRlNJRCJdLAogICAgICAgICAgICAgICAgICAgICJtb24t
            c2VjcmV0IjogIm1vbi1zZWNyZXQiLAogICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgfSwK
            ICAgICAgICAgICAgewogICAgICAgICAgICAgICAgIm5hbWUiOiAicm9vay1jZXBoLW9wZXJhdG9y
            LWNyZWRzIiwKICAgICAgICAgICAgICAgICJraW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAg
            ICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJS
            T09LX0VYVEVSTkFMX1VTRVJOQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgInVzZXJLZXkiOiBz
            ZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAg
            IH0sCiAgICAgICAgICAgIH0sCiAgICAgICAgXQoKICAgICAgICAjIGlmICdNT05JVE9SSU5HX0VO
            RFBPSU5UJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ21vbml0b3JpbmctZW5kcG9pbnQnIHRvIENs
            dXN0ZXIKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19F
            TkRQT0lOVCJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9J
            TlRfUE9SVCJdCiAgICAgICAgKToKICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAg
            ICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjogIm1vbml0b3JpbmctZW5kcG9p
            bnQiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIkNlcGhDbHVzdGVyIiwKICAgICAgICAg
            ICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgIk1vbml0b3JpbmdF
            bmRwb2ludCI6IHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdLAogICAgICAgICAg
            ICAgICAgICAgICAgICAiTW9uaXRvcmluZ1BvcnQiOiBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdf
            RU5EUE9JTlRfUE9SVCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9
            CiAgICAgICAgICAgICkKCiAgICAgICAgIyBpZiAnQ1NJX1JCRF9OT0RFX1NFQ1JFVCcgZXhpc3Rz
            LCB0aGVuIG9ubHkgYWRkICdyb29rLWNzaS1yYmQtcHJvdmlzaW9uZXInIFNlY3JldAogICAgICAg
            IGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VDUkVUIl0KICAg
            ICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FIl0KICAg
            ICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAg
            ICAgICAgICAgICAgICAgICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9O
            T0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQi
            LAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAi
            dXNlcklEIjogc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VDUkVUX05BTUUiXSwKICAgICAg
            ICAgICAgICAgICAgICAgICAgInVzZXJLZXkiOiBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9T
            RUNSRVQiXSwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAg
            ICAgICApCiAgICAgICAgIyBpZiAnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQnIGV4aXN0cywg
            dGhlbiBvbmx5IGFkZCAncm9vay1jc2ktcmJkLXByb3Zpc2lvbmVyJyBTZWNyZXQKICAgICAgICBp
            ZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQi
            XQogICAgICAgICAgICBhbmQgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JF
            VF9OQU1FIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAg
            ICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21h
            cFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAg
            ICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAg
            ICAgICAgICAgICAgICAgICAgInVzZXJJRCI6IHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJ
            T05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNl
            bGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAg
            ICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBpZiAnQ1NJ
            X0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAncm9vay1j
            c2ktY2VwaGZzLXByb3Zpc2lvbmVyJyBTZWNyZXQKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNl
            bGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVQiXQogICAgICAgICAgICBh
            bmQgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0KICAg
            ICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAg
            ICAgICAgICAgICAgICAgICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhG
            U19QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQi
            OiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAg
            ICAgICAgICAgImFkbWluSUQiOiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJf
            U0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgImFkbWluS2V5Ijogc2VsZi5v
            dXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAg
            ICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdDU0lf
            Q0VQSEZTX05PREVfU0VDUkVUJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ3Jvb2stY3NpLWNlcGhm
            cy1ub2RlJyBTZWNyZXQKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJ
            X0NFUEhGU19OT0RFX1NFQ1JFVCJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9D
            RVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJdCiAgICAgICAgKToKICAgICAgICAgICAganNvbl9vdXQu
            YXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjogZiJy
            b29rLXtzZWxmLm91dF9tYXBbJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAg
            ICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0
            YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImFkbWluSUQiOiBzZWxmLm91dF9tYXBbIkNT
            SV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAiYWRt
            aW5LZXkiOiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVQiXSwKICAgICAgICAg
            ICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBp
            ZiAnUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSycgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdy
            b29rLWNlcGgtZGFzaGJvYXJkLWxpbmsnIFNlY3JldAogICAgICAgIGlmIHNlbGYub3V0X21hcFsi
            Uk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSyJdOgogICAgICAgICAgICBqc29uX291dC5hcHBl
            bmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAicm9vay1j
            ZXBoLWRhc2hib2FyZC1saW5rIiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQi
            LAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAi
            dXNlcklEIjogImNlcGgtZGFzaGJvYXJkLWxpbmsiLAogICAgICAgICAgICAgICAgICAgICAgICAi
            dXNlcktleSI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSyJdLAog
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAg
            ICAgICAjIGlmICdSQURPU19OQU1FU1BBQ0UnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCB0aGUgIlJB
            RE9TX05BTUVTUEFDRSIgbmFtZXNwYWNlCiAgICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLm91
            dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIlJF
            U1RSSUNURURfQVVUSF9QRVJNSVNTSU9OIl0KICAgICAgICAgICAgYW5kIG5vdCBzZWxmLm91dF9t
            YXBbIlJCRF9NRVRBREFUQV9FQ19QT09MX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpz
            b25fb3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFt
            ZSI6ICJyYWRvcy1uYW1lc3BhY2UiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIkNlcGhC
            bG9ja1Bvb2xSYWRvc05hbWVzcGFjZSIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAg
            ICAgICAgICAgICAgICAgICAgICAgICJyYWRvc05hbWVzcGFjZU5hbWUiOiBzZWxmLm91dF9tYXBb
            IlJBRE9TX05BTUVTUEFDRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYu
            b3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAg
            ICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAg
            ICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjogImNlcGgtcmJkLXJhZG9z
            LW5hbWVzcGFjZSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwK
            ICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgInBv
            b2wiOiBzZWxmLm91dF9tYXBbIlJCRF9QT09MX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAg
            ICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNyZXQtbmFtZSI6IGYicm9vay17
            c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAg
            ICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250cm9sbGVyLWV4cGFuZC1z
            ZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NF
            Q1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5p
            by9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRf
            Tk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAg
            ICAgIH0KICAgICAgICAgICAgKQogICAgICAgIGVsc2U6CiAgICAgICAgICAgIGlmIHNlbGYub3V0
            X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdOgogICAgICAgICAgICAgICAganNvbl9v
            dXQuYXBwZW5kKAogICAgICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAgICAg
            Im5hbWUiOiAiY2VwaC1yYmQiLAogICAgICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9y
            YWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAg
            ICAgICAgICAgICAgICAgICJkYXRhUG9vbCI6IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJd
            LAogICAgICAgICAgICAgICAgICAgICAgICAgICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIlJCRF9N
            RVRBREFUQV9FQ19QT09MX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJjc2ku
            c3RvcmFnZS5rOHMuaW8vcHJvdmlzaW9uZXItc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0
            X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNlY3Jl
            dC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVU
            X05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5p
            by9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRf
            Tk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAg
            ICAgICAgICAgICAgfQogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBlbHNlOgogICAgICAg
            ICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgICAgIHsKICAgICAgICAg
            ICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaC1yYmQiLAogICAgICAgICAgICAgICAgICAgICAg
            ICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAgICAgICAiZGF0YSI6
            IHsKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJwb29sIjogc2VsZi5vdXRfbWFwWyJSQkRf
            UE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2Uuazhz
            LmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9S
            QkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAg
            ICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYi
            cm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwK
            ICAgICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vbm9kZS1zdGFn
            ZS1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX05PREVfU0VDUkVU
            X05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgICAg
            IH0KICAgICAgICAgICAgICAgICkKCiAgICAgICAgIyBpZiAnVE9QT0xPR1lfUE9PTFMnLCAnVE9Q
            T0xPR1lfRkFJTFVSRV9ET01BSU5fTEFCRUwnLCAnVE9QT0xPR1lfRkFJTFVSRV9ET01BSU5fVkFM
            VUVTJyAgZXhpc3RzLAogICAgICAgICMgdGhlbiBvbmx5IGFkZCAndG9wb2xvZ3knIFN0b3JhZ2VD
            bGFzcwogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09M
            UyJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlO
            X0xBQkVMIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfRkFJTFVSRV9E
            T01BSU5fVkFMVUVTIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAg
            ICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaC1yYmQtdG9w
            b2xvZ3kiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAg
            ICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJ0b3BvbG9n
            eUZhaWx1cmVEb21haW5MYWJlbCI6IHNlbGYub3V0X21hcFsKICAgICAgICAgICAgICAgICAgICAg
            ICAgICAgICJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9MQUJFTCIKICAgICAgICAgICAgICAgICAg
            ICAgICAgXSwKICAgICAgICAgICAgICAgICAgICAgICAgInRvcG9sb2d5RmFpbHVyZURvbWFpblZh
            bHVlcyI6IHNlbGYub3V0X21hcFsKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJUT1BPTE9H
            WV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiCiAgICAgICAgICAgICAgICAgICAgICAgIF0sCiAgICAg
            ICAgICAgICAgICAgICAgICAgICJ0b3BvbG9neVBvb2xzIjogc2VsZi5vdXRfbWFwWyJUT1BPTE9H
            WV9QT09MUyJdLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3By
            b3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJP
            VklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0
            b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxm
            Lm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUi
            OiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAg
            ICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCgogICAg
            ICAgICMgaWYgJ0NFUEhGU19GU19OQU1FJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ2NlcGhmcycg
            U3RvcmFnZUNsYXNzCiAgICAgICAgaWYgc2VsZi5vdXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdOgog
            ICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAg
            ICAgICAgICAgIm5hbWUiOiAiY2VwaGZzIiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJT
            dG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAg
            ICAgICAgICAgICAiZnNOYW1lIjogc2VsZi5vdXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdLAogICAg
            ICAgICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYub3V0X21hcFsiQ0VQSEZTX1BPT0xfTkFN
            RSJdLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lv
            bmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9DRVBIRlNfUFJPVklT
            SU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3Jh
            Z2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91
            dF9tYXBbJ0NTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUi
            OiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FJ119IiwK
            ICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAg
            ICAgICAgIyBpZiAnUkdXX0VORFBPSU5UJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ2NlcGgtcmd3
            JyBTdG9yYWdlQ2xhc3MKICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdOgog
            ICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAg
            ICAgICAgICAgIm5hbWUiOiAiY2VwaC1yZ3ciLAogICAgICAgICAgICAgICAgICAgICJraW5kIjog
            IlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAg
            ICAgICAgICAgICAgICJlbmRwb2ludCI6IHNlbGYub3V0X21hcFsiUkdXX0VORFBPSU5UIl0sCiAg
            ICAgICAgICAgICAgICAgICAgICAgICJwb29sUHJlZml4Ijogc2VsZi5vdXRfbWFwWyJSR1dfUE9P
            TF9QUkVGSVgiXSwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAg
            ICAgICAgICApCiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsK
            ICAgICAgICAgICAgICAgICAgICAibmFtZSI6ICJyZ3ctYWRtaW4tb3BzLXVzZXIiLAogICAgICAg
            ICAgICAgICAgICAgICJraW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEi
            OiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJhY2Nlc3NLZXkiOiBzZWxmLm91dF9tYXBbIlJH
            V19BRE1JTl9PUFNfVVNFUl9BQ0NFU1NfS0VZIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJz
            ZWNyZXRLZXkiOiBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9TRUNSRVRfS0VZIl0s
            CiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQog
            ICAgICAgICMgaWYgJ1JHV19UTFNfQ0VSVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkIHRoZSAiY2Vw
            aC1yZ3ctdGxzLWNlcnQiIHNlY3JldAogICAgICAgIGlmIHNlbGYub3V0X21hcFsiUkdXX1RMU19D
            RVJUIl06CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAg
            ICAgICAgICAgICAgICAgICAibmFtZSI6ICJjZXBoLXJndy10bHMtY2VydCIsCiAgICAgICAgICAg
            ICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsK
            ICAgICAgICAgICAgICAgICAgICAgICAgImNlcnQiOiBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VS
            VCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAg
            ICkKCiAgICAgICAgcmV0dXJuIGpzb24uZHVtcHMoanNvbl9vdXQpICsgTElORVNFUAoKICAgIGRl
            ZiB1cGdyYWRlX3VzZXJzX3Blcm1pc3Npb25zKHNlbGYpOgogICAgICAgIHVzZXJzID0gWwogICAg
            ICAgICAgICAiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIsCiAgICAgICAgICAgICJjbGllbnQuY3Np
            LWNlcGhmcy1wcm92aXNpb25lciIsCiAgICAgICAgICAgICJjbGllbnQuY3NpLXJiZC1ub2RlIiwK
            ICAgICAgICAgICAgImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwKICAgICAgICAgICAgImNs
            aWVudC5oZWFsdGhjaGVja2VyIiwKICAgICAgICBdCiAgICAgICAgaWYgc2VsZi5ydW5fYXNfdXNl
            ciAhPSAiIiBhbmQgc2VsZi5ydW5fYXNfdXNlciBub3QgaW4gdXNlcnM6CiAgICAgICAgICAgIHVz
            ZXJzLmFwcGVuZChzZWxmLnJ1bl9hc191c2VyKQogICAgICAgIGZvciB1c2VyIGluIHVzZXJzOgog
            ICAgICAgICAgICBzZWxmLnVwZ3JhZGVfdXNlcl9wZXJtaXNzaW9ucyh1c2VyKQoKICAgIGRlZiBn
            ZXRfcmd3X3Bvb2xfbmFtZV9kdXJpbmdfdXBncmFkZShzZWxmLCB1c2VyLCBjYXBzKToKICAgICAg
            ICBpZiB1c2VyID09ICJjbGllbnQuaGVhbHRoY2hlY2tlciI6CiAgICAgICAgICAgICMgd2hlbiBh
            ZG1pbiBoYXMgbm90IHByb3ZpZGVkIHJndyBwb29sIG5hbWUgZHVyaW5nIHVwZ3JhZGUsCiAgICAg
            ICAgICAgICMgZ2V0IHRoZSByZ3cgcG9vbCBuYW1lIGZyb20gY2xpZW50LmhlYWx0aGNoZWNrZXIg
            dXNlciB3aGljaCB3YXMgdXNlZCBkdXJpbmcgY29ubmVjdGlvbgogICAgICAgICAgICBpZiBub3Qg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXg6CiAgICAgICAgICAgICAgICAjIFRvIGdl
            dCB2YWx1ZSAnZGVmYXVsdCcgd2hpY2ggaXMgcmd3IHBvb2wgbmFtZSBmcm9tICdhbGxvdyByd3gg
            cG9vbD1kZWZhdWx0LnJndy5tZXRhJwogICAgICAgICAgICAgICAgcGF0dGVybiA9IHIicG9vbD0o
            Lio/KVwucmd3XC5tZXRhIgogICAgICAgICAgICAgICAgbWF0Y2ggPSByZS5zZWFyY2gocGF0dGVy
            biwgY2FwcykKICAgICAgICAgICAgICAgIGlmIG1hdGNoOgogICAgICAgICAgICAgICAgICAgIHNl
            bGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4ID0gbWF0Y2guZ3JvdXAoMSkKICAgICAgICAg
            ICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAgICAgImZhaWxlZCB0byBnZXQgcmd3IHBvb2wg
            bmFtZSBmb3IgdXBncmFkZSIKICAgICAgICAgICAgICAgICAgICApCgogICAgZGVmIHVwZ3JhZGVf
            dXNlcl9wZXJtaXNzaW9ucyhzZWxmLCB1c2VyKToKICAgICAgICAjIGNoZWNrIHdoZXRoZXIgdGhl
            IGdpdmVuIHVzZXIgZXhpc3RzIG9yIG5vdAogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAi
            YXV0aCBnZXQiLCAiZW50aXR5IjogZiJ7dXNlcn0iLCAiZm9ybWF0IjogImpzb24ifQogICAgICAg
            IHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihj
            bWRfanNvbikKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgog
            ICAgICAgICAgICBwcmludChmInVzZXIge3VzZXJ9IG5vdCBmb3VuZCBmb3IgdXBncmFkaW5nLiIp
            CiAgICAgICAgICAgIHJldHVybgogICAgICAgIGV4aXN0aW5nX2NhcHMgPSBqc29uX291dFswXVsi
            Y2FwcyJdCiAgICAgICAgc2VsZi5nZXRfcmd3X3Bvb2xfbmFtZV9kdXJpbmdfdXBncmFkZSh1c2Vy
            LCBzdHIoZXhpc3RpbmdfY2FwcykpCiAgICAgICAgbmV3X2NhcCwgXyA9IHNlbGYuZ2V0X2NhcHNf
            YW5kX2VudGl0eSh1c2VyKQogICAgICAgIGNhcF9rZXlzID0gWyJtb24iLCAibWdyIiwgIm9zZCIs
            ICJtZHMiXQogICAgICAgIGNhcHMgPSBbXQogICAgICAgIGZvciBlYWNoQ2FwIGluIGNhcF9rZXlz
            OgogICAgICAgICAgICBjdXJfY2FwX3ZhbHVlcyA9IGV4aXN0aW5nX2NhcHMuZ2V0KGVhY2hDYXAs
            ICIiKQogICAgICAgICAgICBuZXdfY2FwX3ZhbHVlcyA9IG5ld19jYXAuZ2V0KGVhY2hDYXAsICIi
            KQogICAgICAgICAgICBjdXJfY2FwX3Blcm1fbGlzdCA9IFsKICAgICAgICAgICAgICAgIHguc3Ry
            aXAoKSBmb3IgeCBpbiBjdXJfY2FwX3ZhbHVlcy5zcGxpdCgiLCIpIGlmIHguc3RyaXAoKQogICAg
            ICAgICAgICBdCiAgICAgICAgICAgIG5ld19jYXBfcGVybV9saXN0ID0gWwogICAgICAgICAgICAg
            ICAgeC5zdHJpcCgpIGZvciB4IGluIG5ld19jYXBfdmFsdWVzLnNwbGl0KCIsIikgaWYgeC5zdHJp
            cCgpCiAgICAgICAgICAgIF0KICAgICAgICAgICAgIyBhcHBlbmQgbmV3X2NhcF9saXN0IHRvIGN1
            cl9jYXBfbGlzdCB0byBtYWludGFpbiB0aGUgb3JkZXIgb2YgY2FwcwogICAgICAgICAgICBjdXJf
            Y2FwX3Blcm1fbGlzdC5leHRlbmQobmV3X2NhcF9wZXJtX2xpc3QpCiAgICAgICAgICAgICMgZWxp
            bWluYXRlIGR1cGxpY2F0ZXMgd2l0aG91dCB1c2luZyAnc2V0JwogICAgICAgICAgICAjIHNldCBy
            ZS1vcmRlcnMgaXRlbXMgaW4gdGhlIGxpc3QgYW5kIHdlIGhhdmUgdG8ga2VlcCB0aGUgb3JkZXIK
            ICAgICAgICAgICAgbmV3X2NhcF9saXN0ID0gW10KICAgICAgICAgICAgW25ld19jYXBfbGlzdC5h
            cHBlbmQoeCkgZm9yIHggaW4gY3VyX2NhcF9wZXJtX2xpc3QgaWYgeCBub3QgaW4gbmV3X2NhcF9s
            aXN0XQogICAgICAgICAgICBleGlzdGluZ19jYXBzW2VhY2hDYXBdID0gIiwgIi5qb2luKG5ld19j
            YXBfbGlzdCkKICAgICAgICAgICAgaWYgZXhpc3RpbmdfY2Fwc1tlYWNoQ2FwXToKICAgICAgICAg
            ICAgICAgIGNhcHMuYXBwZW5kKGVhY2hDYXApCiAgICAgICAgICAgICAgICBjYXBzLmFwcGVuZChl
            eGlzdGluZ19jYXBzW2VhY2hDYXBdKQogICAgICAgIGNtZF9qc29uID0gewogICAgICAgICAgICAi
            cHJlZml4IjogImF1dGggY2FwcyIsCiAgICAgICAgICAgICJlbnRpdHkiOiB1c2VyLAogICAgICAg
            ICAgICAiY2FwcyI6IGNhcHMsCiAgICAgICAgICAgICJmb3JtYXQiOiAianNvbiIsCiAgICAgICAg
            fQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9q
            c29uX2dlbihjbWRfanNvbikKICAgICAgICBpZiByZXRfdmFsICE9IDA6CiAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIidhdXRoIGNh
            cHMge3VzZXJ9JyBjb21tYW5kIGZhaWxlZC5cbiBFcnJvcjoge2Vycl9tc2d9IgogICAgICAgICAg
            ICApCiAgICAgICAgcHJpbnQoZiJVcGRhdGVkIHVzZXIge3VzZXJ9IHN1Y2Nlc3NmdWxseS4iKQoK
            ICAgIGRlZiBtYWluKHNlbGYpOgogICAgICAgIGdlbmVyYXRlZF9vdXRwdXQgPSAiIgogICAgICAg
            IGlmIHNlbGYuX2FyZ19wYXJzZXIudXBncmFkZToKICAgICAgICAgICAgc2VsZi51cGdyYWRlX3Vz
            ZXJzX3Blcm1pc3Npb25zKCkKICAgICAgICBlbGlmIHNlbGYuX2FyZ19wYXJzZXIuZm9ybWF0ID09
            ICJqc29uIjoKICAgICAgICAgICAgZ2VuZXJhdGVkX291dHB1dCA9IHNlbGYuZ2VuX2pzb25fb3V0
            KCkKICAgICAgICBlbGlmIHNlbGYuX2FyZ19wYXJzZXIuZm9ybWF0ID09ICJiYXNoIjoKICAgICAg
            ICAgICAgZ2VuZXJhdGVkX291dHB1dCA9IHNlbGYuZ2VuX3NoZWxsX291dCgpCiAgICAgICAgZWxz
            ZToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAg
            ICAgICAgIGYiVW5zdXBwb3J0ZWQgZm9ybWF0OiB7c2VsZi5fYXJnX3BhcnNlci5mb3JtYXR9Igog
            ICAgICAgICAgICApCiAgICAgICAgcHJpbnQoZ2VuZXJhdGVkX291dHB1dCkKICAgICAgICBpZiBz
            ZWxmLm91dHB1dF9maWxlIGFuZCBnZW5lcmF0ZWRfb3V0cHV0OgogICAgICAgICAgICBmT3V0ID0g
            b3BlbihzZWxmLm91dHB1dF9maWxlLCBtb2RlPSJ3IiwgZW5jb2Rpbmc9IlVURi04IikKICAgICAg
            ICAgICAgZk91dC53cml0ZShnZW5lcmF0ZWRfb3V0cHV0KQogICAgICAgICAgICBmT3V0LmNsb3Nl
            KCkKCgojIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMKIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIE1BSU4gIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIwppZiBfX25hbWVfXyA9PSAiX19t
            YWluX18iOgogICAgcmpPYmogPSBSYWRvc0pTT04oKQogICAgdHJ5OgogICAgICAgIHJqT2JqLm1h
            aW4oKQogICAgZXhjZXB0IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24gYXMgZXJyOgogICAgICAg
            IHByaW50KGYiRXhlY3V0aW9uIEZhaWxlZDoge2Vycn0iKQogICAgICAgIHJhaXNlIGVycgogICAg
            ZXhjZXB0IEtleUVycm9yIGFzIGtFcnI6CiAgICAgICAgcHJpbnQoZiJLZXlFcnJvcjoge2tFcnJ9
            IikKICAgIGV4Y2VwdCBPU0Vycm9yIGFzIG9zRXJyOgogICAgICAgIHByaW50KGYiRXJyb3Igd2hp
            bGUgdHJ5aW5nIHRvIG91dHB1dCB0aGUgZGF0YToge29zRXJyfSIpCiAgICBmaW5hbGx5OgogICAg
            ICAgIHJqT2JqLnNodXRkb3duKCkK
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/red-hat-storage/rook
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Ceph cluster.
            displayName: Ceph Cluster
            kind: CephCluster
            name: cephclusters.ceph.rook.io
            version: v1
          - description: Represents a Ceph Block Pool.
            displayName: Ceph Block Pool
            kind: CephBlockPool
            name: cephblockpools.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store.
            displayName: Ceph Object Store
            kind: CephObjectStore
            name: cephobjectstores.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store User.
            displayName: Ceph Object Store User
            kind: CephObjectStoreUser
            name: cephobjectstoreusers.ceph.rook.io
            version: v1
          - description: Represents a cluster of Ceph NFS ganesha gateways.
            displayName: Ceph NFS
            kind: CephNFS
            name: cephnfses.ceph.rook.io
            version: v1
          - description: Represents a Ceph User.
            displayName: Ceph Client
            kind: CephClient
            name: cephclients.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem.
            displayName: Ceph Filesystem
            kind: CephFilesystem
            name: cephfilesystems.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem Mirror.
            displayName: Ceph Filesystem Mirror
            kind: CephFilesystemMirror
            name: cephfilesystemmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph RBD Mirror.
            displayName: Ceph RBD Mirror
            kind: CephRBDMirror
            name: cephrbdmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Realm.
            displayName: Ceph Object Store Realm
            kind: CephObjectRealm
            name: cephobjectrealms.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone Group.
            displayName: Ceph Object Store Zone Group
            kind: CephObjectZoneGroup
            name: cephobjectzonegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone.
            displayName: Ceph Object Store Zone
            kind: CephObjectZone
            name: cephobjectzones.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Notification.
            displayName: Ceph Bucket Notification
            kind: CephBucketNotification
            name: cephbucketnotifications.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Topic.
            displayName: Ceph Bucket Topic
            kind: CephBucketTopic
            name: cephbuckettopics.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem SubVolumeGroup.
            displayName: Ceph Filesystem SubVolumeGroup
            kind: CephFilesystemSubVolumeGroup
            name: cephfilesystemsubvolumegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph BlockPool Rados Namespace.
            displayName: Ceph BlockPool Rados Namespace
            kind: CephBlockPoolRadosNamespace
            name: cephblockpoolradosnamespaces.ceph.rook.io
            version: v1
          - description: Represents a Ceph COSI Driver.
            displayName: Ceph COSI Driver
            kind: CephCOSIDriver
            name: cephcosidrivers.ceph.rook.io
            version: v1
        description: |2

          The Rook-Ceph storage operator packages, deploys, manages, upgrades and scales Ceph storage for providing persistent storage to infrastructure services (Logging, Metrics, Registry) as well as stateful applications in Kubernetes clusters.

          ## Rook-Ceph Storage Operator

          Rook runs as a cloud-native service in Kubernetes clusters for optimal integration with applications in need of storage, and handles the heavy-lifting behind the scenes such as provisioning and management.
          Rook orchestrates battle-tested open-source storage technology Ceph, which has years of production deployments and runs some of the worlds largest clusters.

          Ceph is a massively scalable, software-defined, cloud native storage platform that offers block, file and object storage services.
          Ceph can be used to back a wide variety of applications including relational databases, NoSQL databases, CI/CD tool-sets, messaging, AI/ML and analytics applications.
          Ceph is a proven storage platform that backs some of the world's largest storage deployments and has a large vibrant open source community backing the project.

          ## Supported features
          * **High Availability and resiliency** - Ceph has no single point of failures (SPOF) and all its components work natively in a highly available fashion
          * **Data Protection** - Ceph periodically scrub for inconsistent objects and repair them if necessary, making sure your replicas are always coherent
          * **Consistent storage platform across hybrid cloud** - Ceph can be deployed anywhere (on-premise or bare metal) and thus offers a similar experience regardless
          * **Block, File & Object storage service** - Ceph can expose your data through several storage interfaces, solving all the application use cases
          * **Scale up/down** - addition and removal of storage is fully covered by the operator.
          * **Dashboard** - The Operator deploys a dashboard for monitoring and introspecting your cluster.

          ## Before you start
          https://rook.io/docs/rook/v1.0/k8s-pre-reqs.html
        displayName: Rook-Ceph
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rook
        - ceph
        - storage
        - object storage
        - open source
        - block storage
        - shared filesystem
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/rook
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:cf7f72469d25fb37301a8159406613673f8968c82af1035fd45a470e0ae8acfa
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:80c01633a6bfca44df37ac38f764035be6998fce965c8709e4f6d54d17eeaa7b
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6fee56851da3a87b9f7f477fc59e3385e118dc0b398861a8f13459c998a5af77
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c11a599bfd09b9b22888e03eabdbe1054d4049718a591d2a45bf0ed89800111e
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:82a4b481028195cfd75471656fe40c1cbbade4c7b2d170d6bf9422d9d3b84fa7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:693a60d6a1162659213ae554b1c6ee200b0300a7da72c01415044cf507acab49
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:0afcb516eb528857887c6ee1a9cba081dc8f6b7d47d9f26a0dbb2dca0cefe07c
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:8d8326c74795c9c6bd8f33a9f9e81d2304a70b1f9be3c2f8d50e76be4b0292ff
        version: 4.16.16-rhodf
      entries:
      - name: rook-ceph-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: rook-ceph-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: rook-ceph-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: rook-ceph-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: rook-ceph-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: rook-ceph-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: rook-ceph-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: rook-ceph-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: rook-ceph-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: rook-ceph-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: rook-ceph-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: rook-ceph-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: rook-ceph-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: rook-ceph-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: rook-ceph-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: rook-ceph-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: rook-ceph-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: rook-ceph-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: logic-operator-rhel8
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: logic-operator-rhel8.v1.36.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "host/namespace",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          createdAt: "2025-05-26T14:10:15Z"
          description: OpenShift Serverless Logic Kubernetes Operator for deploying
            workflow applications based on the CNCF Serverless Workflow specification
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-serverless-logic
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kubesmarts/kie-tools
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: Sonata Flow Build
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: Sonata Flow Cluster Platform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: Sonata Flow Platform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: Sonata Flow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          OpenShift Serverless Logic Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification v0.8](https://serverlessworkflow.io/).
        displayName: OpenShift Serverless Logic Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        - workflows
        links:
        - name: Documentation
          url: https://docs.openshift.com/serverless/
        - name: Source Repository
          url: https://github.com/kiegroup/kogito-serverless-operator/
        maintainers:
        - email: serverless-logic-team@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/logic-rhel8-operator@sha256:8d3682448ebdac3aeabb2d23842b7e67a252b95f959c408af805037f9728fd3c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4564ca3dc5bac80d6faddaf94c817fbbc270698a9399d8a21ee1005d85ceda56
        version: 1.36.0
      entries:
      - name: logic-operator-rhel8.v1.36.0
        version: 1.36.0
      - name: logic-operator-rhel8.v1.35.0
        version: 1.35.0
      - name: logic-operator-rhel8.v1.34.0
        version: 1.34.0
      - name: logic-operator-rhel8.v1.33.0
        version: 1.33.0
      name: alpha
    defaultChannel: alpha
    packageName: logic-operator-rhel8
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudwatchlogs-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudwatchlogs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudwatchlogs.services.k8s.aws/v1alpha1",
                "kind": "LogGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudwatchlogs-controller:1.1.1
          createdAt: "2025-09-19T20:51:16Z"
          description: AWS CloudWatch Logs controller is a service controller for
            managing CloudWatch Logs resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LogGroup represents the state of an AWS cloudwatchlogs LogGroup
              resource.
            displayName: LogGroup
            kind: LogGroup
            name: loggroups.cloudwatchlogs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudWatch Logs resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudWatch Logs**

          You can use Amazon CloudWatch Logs to monitor, store, and access your log files from Amazon Elastic Compute Cloud (Amazon EC2) instances, AWS CloudTrail, Route 53, and other sources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudWatch Logs
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudwatchlogs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudWatch Logs Developer Resources
          url: https://aws.amazon.com/cloudwatch/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudwatch maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudwatchlogs-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-cloudwatchlogs-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudwatchlogs-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudwatchlogs-controller.v1.0.14
        version: 1.0.14
      - name: ack-cloudwatchlogs-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudwatchlogs-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudwatchlogs-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudwatchlogs-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudwatchlogs-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudwatchlogs-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudwatchlogs-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudwatchlogs-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudwatchlogs-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudwatchlogs-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudwatchlogs-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudwatchlogs-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudwatchlogs-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudwatchlogs-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudwatchlogs-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudwatchlogs-controller.v0.0.7
        version: 0.0.7
      - name: ack-cloudwatchlogs-controller.v0.0.4
        version: 0.0.4
      - name: ack-cloudwatchlogs-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudwatchlogs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: TrendMicro
      provider-url: trendmicro.com
    name: visionone-containersecurity
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: visionone-containersecurity.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "visionone-containersecurity-sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          com.redhat.openshift.versions: v4.12-v4.20
          containerImage: public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
          createdAt: "2025-09-18T22:43:41Z"
          description: Trend Vision One™ Container Security delivers full lifecycle
            protection for containers with real-time threat detection, policy enforcement,
            compliance assurance, and rapid incident response.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trendmicro-system
          operatorhub.io/ui-metadata-max-k8s-version: "1.26"
          operators.openshift.io/valid-subscription: VisionOneContainerSecurity Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trendmicro/visionone-container-security-helm
          support: TrendMicro Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trend Vision One™ Container Security delivers full lifecycle
              protection for containers with real-time threat detection, policy enforcement,
              compliance assurance, and rapid incident response.
            displayName: VisionOneContainerSecurity
            kind: VisionOneContainerSecurity
            name: visiononecontainersecurities.container-security.trendmicro.com
            version: v1alpha1
        description: "[Trend Vision One™ Container Security](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-container-security)
          delivers comprehensive, end-to-end\n protection for containerized workloads
          across OpenShift and Kubernetes environments.\n It continuously monitors
          for threats, vulnerabilities, policy violations, and\n exposed secrets ensuring
          compliance and enabling rapid, automated response. Purpose-built\n for hybrid
          cloud, it instantly isolates compromised containers to prevent lateral\n
          movement and minimize impact. Seamlessly integrating into OpenShift DevSecOps\n
          pipelines, it empowers security teams to enforce policies while enabling
          developers\n to innovate and deploy with confidence without disrupting workflows
          or compromising\n agility.\n\n **Installation**\n \n - Use the recommended
          `trendmicro-system` namespace.\n - Refer to the [Trend Vision One™ Documentation](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator)
          for registration and installation steps.\n"
        displayName: Trend Vision One™ Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trendmicro
        - visionone
        - container-security
        links:
        - name: Documentation
          url: https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator
        maintainers:
        - email: support@trendmicro.com
          name: TrendMicro VisionOne Support Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: TrendMicro
          url: trendmicro.com
        relatedImages:
        - public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
        version: 0.0.2
      entries:
      - name: visionone-containersecurity.v0.0.2
        version: 0.0.2
      - name: visionone-containersecurity.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: visionone-containersecurity.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "visionone-containersecurity-sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          com.redhat.openshift.versions: v4.12-v4.20
          containerImage: public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
          createdAt: "2025-09-18T22:43:41Z"
          description: Trend Vision One™ Container Security delivers full lifecycle
            protection for containers with real-time threat detection, policy enforcement,
            compliance assurance, and rapid incident response.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trendmicro-system
          operatorhub.io/ui-metadata-max-k8s-version: "1.26"
          operators.openshift.io/valid-subscription: VisionOneContainerSecurity Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trendmicro/visionone-container-security-helm
          support: TrendMicro Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trend Vision One™ Container Security delivers full lifecycle
              protection for containers with real-time threat detection, policy enforcement,
              compliance assurance, and rapid incident response.
            displayName: VisionOneContainerSecurity
            kind: VisionOneContainerSecurity
            name: visiononecontainersecurities.container-security.trendmicro.com
            version: v1alpha1
        description: "[Trend Vision One™ Container Security](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-container-security)
          delivers comprehensive, end-to-end\n protection for containerized workloads
          across OpenShift and Kubernetes environments.\n It continuously monitors
          for threats, vulnerabilities, policy violations, and\n exposed secrets ensuring
          compliance and enabling rapid, automated response. Purpose-built\n for hybrid
          cloud, it instantly isolates compromised containers to prevent lateral\n
          movement and minimize impact. Seamlessly integrating into OpenShift DevSecOps\n
          pipelines, it empowers security teams to enforce policies while enabling
          developers\n to innovate and deploy with confidence without disrupting workflows
          or compromising\n agility.\n\n **Installation**\n \n - Use the recommended
          `trendmicro-system` namespace.\n - Refer to the [Trend Vision One™ Documentation](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator)
          for registration and installation steps.\n"
        displayName: Trend Vision One™ Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trendmicro
        - visionone
        - container-security
        links:
        - name: Documentation
          url: https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator
        maintainers:
        - email: support@trendmicro.com
          name: TrendMicro VisionOne Support Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: TrendMicro
          url: trendmicro.com
        relatedImages:
        - public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
        version: 0.0.2
      entries:
      - name: visionone-containersecurity.v0.0.2
        version: 0.0.2
      name: stable
    defaultChannel: stable
    packageName: visionone-containersecurity
    provider:
      name: TrendMicro
      url: trendmicro.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Docling Maintainers
      provider-url: https://docling-project.github.io/docling/
    name: docling-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: docling-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2025-05-19T20:20:14Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DoclingServe is the Schema for the doclingserves API
            displayName: Docling Serve
            kind: DoclingServe
            name: doclingserves.docling.github.io
            version: v1alpha1
        description: |-
          **Overview**

          The Docling Operator is responsible for managing the full lifecycle of Docling Serve instances in your Kubernetes or OpenShift container platforms.


          **Docling Serve**

          The Docling Serve application deploys a REST API for consuming and integrating Docling in your workflows.

          The API offers endpoints for multiple types of interaction, i.e. both synchronous and asynchronous endpoints, as well as status subscriptions via websockets.

          The options for the Docling conversion can be customized at runtime by the users (e.g. which OCR engine, which enrichment models, etc). Users can also choose which of the many DoclingDocument export formats (JSON, Markdown, HTML, DocTags, etc) to receive in the result response.

          Docling Serve allows to choose between a local compute engine, i.e. running Docling in the same container as the API server, or a distributed compute engine which leverages Kubeflow Pipelines to parallelize large-scale ingestion jobs.

          **Docling**

          Docling is a powerful library which simplifies document processing, parsing diverse formats — including advanced PDF understanding — and providing seamless integrations with the gen AI ecosystem.

          Learn more about the Docling Project by visiting the GitHub organization at [https://github.com/docling-project](https://github.com/docling-project).


          **Sample DoclingServe Custom Resources**


          To deploy with the `local` compute engine with the UI enabled and Route exposed:
          ```yaml
          apiVersion: docling.github.io/v1alpha1
          kind: DoclingServe
          metadata:
            labels:
              app.kubernetes.io/name: docling-operator
              app.kubernetes.io/managed-by: kustomize
            name: doclingserve-local-sample
          spec:
            apiServer:
              image: "quay.io/docling-project/docling-serve:latest"
              enableUI: true
              instances: 1
            engine:
              local:
                numWorkers: 2
            route:
              enabled: true
          ```

          To deploy with the `KFP` compute engine with the UI enabled and Route exposed. **Note: Kubeflow Pipeline must be installed and setup in the cluster to use this option.**:
          ```yaml
          apiVersion: docling.github.io/v1alpha1
          kind: DoclingServe
          metadata:
            labels:
              app.kubernetes.io/name: docling-operator
              app.kubernetes.io/managed-by: kustomize
            name: doclingserve-kfp-sample
          spec:
            apiServer:
              image: "quay.io/docling-project/docling-serve:latest"
              enableUI: true
              instances: 1
            engine:
              kfp:
                endpoint: "https://NAME.NAMESPACE.svc.cluster.local:8888"
            route:
              enabled: true
          ```

          **Additional Configurations**


          The CRD exposes `apiServer.configMapName` to enable additional configurations of the `docling-serve` workload. For a full list of configurations can be found on docling-serve's
          [Configuration](https://github.com/docling-project/docling-serve/blob/main/docs/configuration.md) page.
        displayName: Docling Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - document
        - processing
        links:
        - name: Docling Operator
          url: https://docling-project.github.io/docling/
        maintainers:
        - email: docling-technical-discuss@lists.lfaidata.foundation
          name: Docling Maintainers
        maturity: alpha
        provider:
          name: Docling Maintainers
          url: https://docling-project.github.io/docling/
        relatedImages:
        - quay.io/docling-project/docling-operator@sha256:200d8db51874c4b1e3d20d8d1423e9a4807843fbb29f514cd5d8ba872e47f4af
        version: 0.0.1
      entries:
      - name: docling-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: docling-operator
    provider:
      name: Docling Maintainers
      url: https://docling-project.github.io/docling/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubemq.io
      provider-url: ""
    name: kubemq-operator-marketplace
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubemq-operator-marketplace.1.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
              "apiVersion": "core.k8s.kubemq.io/v1beta1",
              "kind": "KubemqCluster",
              "metadata": {
                "name": "kubemq-cluster"
              },
              "spec": {
                "key": "<Your-KubeMQ-License-Key>",
                "replicas": 3
              }
            },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqConnector",
                "metadata": {
                  "name": "kubemq-bridges"
                },
                "spec": {
                  "type": "bridges",
                  "config": "bindings: null"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.connect.redhat.com/kubemq/kubemq-operator@sha256:3a9c87e462eb1caa81925dee72407f73fca41f19a3f9bcaf4e76e569db1dfbab
          createdAt: 2020-03-11 12:00:00+00:00
          description: Kubernetes Messaging Platform
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          support: support@kubemq.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kubemq Cluster
            displayName: Kubemq Cluster
            kind: KubemqCluster
            name: kubemqclusters.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Connector
            displayName: Kubemq Connector
            kind: KubemqConnector
            name: kubemqconnectors.core.k8s.kubemq.io
            version: v1beta1
        description: |2-

          ## Kubernetes Messaging Platform
          A message-based architecture for microservices providing flexible connectivity for hybrid environments: multi-cloud, on-premises, and at the edge. Powered by enterprise-grade message broker and message queue with ready to use connectors, bridges, and control center.

          Messaging platform components:
            - **KubeMQ Cluster** - Enterprise-grade message broker and message queue, scalable, high available, and secured.
            - **KubeMQ Targets Connector** - connects a KubeMQ cluster with external systems and cloud services.
            - **KubeMQ Bridges Connector** - bridge, replicate, aggregate, and transform messages between KubeMQ clusters.
            - **KubeMQ Sources Connector** - transform external systems and cloud services data to KubeMQ cluster messages.
          \n### Kubernetes Native
          Innovative and modern message queue and message broker in a lightweight container developed to run in Kubernetes, certified in the CNCF landscape, and connect natively to the cloud-native ecosystem.

          ### Ease of use

          Simple deployment in Kubernetes in less than 1 minute. Developer friendly by simple to use SDKs and elimination of the many developers and DevOps-centered challenges to define exchanges, brokers, channels, routes, and predefined topics

          ### All messaging patterns

          KubeMQ is available with all messaging patterns such as Queue, Stream, Pub/Sub, and RPC, Saving the need to maintain multiple messaging platforms for different use cases.

          ### Hybrid Environments

          KubeMQ enables the microservices from multiple environments to communicate and build one hybrid infrastructure solution across clouds, on-prem, and at the edge.

          ### KubeMQ Cluster key features

          - All-batteries included Messaging Queue Broker for the Kubernetes environment
          - Deployed with Operator for full life cycle operation
          - Blazing fast (written in Go), small and lightweight Docker container
          - Asynchronous and Synchronous messaging with support for  `At Most Once Delivery` and `At Least Once Delivery` models
          - Supports durable FIFO based Queue, Publish-Subscribe Events, Publish-Subscribe with Persistence (Events Store), RPC Command and Query messaging patterns
          - Supports gRPC, Rest and WebSocket Transport protocols with TLS support (both RPC and Stream modes)
          - Supports Access control Authorization and Authentication
          - Supports message masticating and smart routing
          - No Message broker configuration needed (i.e., queues, exchanges)
          - .Net, Java, Python, Go, and NodeJS SDKs

          ### KubeMQ Targets Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major services types (databases, cache, messaging, serverless, HTTP, etc.)
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-targets)

          ### KubeMQ Bridges Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Build Any Topology - connects KubeMQ clusters in 1:1, 1:n , n:1, n:n
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-bridges)

          ### KubeMQ Sources Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major messaging services such as Kafka, ActiveMQ, SQS, and more.
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-sources)

          ## Getting Started

          [Start with OpenShift](https://docs.kubemq.io/getting-started/create-cluster/openshift)

          ## Get Enterprise Trial License

          [Register](https://account.kubemq.io/login/register)

          ## Request a Demo

          [See KubeMQ in Action](https://kubemq.io/request-a-demo/)
        displayName: Kubemq Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - messaging
        - queue
        - broker
        - pubsub
        - rpc
        links:
        - name: Learn more about Kubemq
          url: https://kubemq.io/
        - name: Register for free license
          url: https://account.kubemq.io/login/register
        - name: Contact Sales
          url: https://kubemq.io/contact-sales/
        - name: Documentation
          url: https://docs.kubemq.io/
        - name: Get Help
          url: https://kubemq.io/contact-support/
        maintainers:
        - email: support@kubemq.io
          name: kubemq
        maturity: alpha
        provider:
          name: Kubemq.io
        relatedImages:
        - registry.connect.redhat.com/kubemq/kubemq-operator@sha256:3a9c87e462eb1caa81925dee72407f73fca41f19a3f9bcaf4e76e569db1dfbab
        - registry.connect.redhat.com/kubemq/kubemq-uni@sha256:51154b2a112c5cdc17ee8d338524c02ccd995f9386180b1cc232b7a875a23e73
        - registry.connect.redhat.com/kubemq/kubemq-targets@sha256:2a09e1718a03729a121728e4e3683e7e1f0690789d7f9857d76bebf1cc01d436
        - registry.connect.redhat.com/kubemq/kubemq-sources@sha256:5736575b6cae39d129346f83fe2d8ab9cdbd1c1d2a99a66b4b9c01631fb5db21
        - registry.connect.redhat.com/kubemq/kubemq-sources@sha256:8ed531e58c60760888ffb18165883386bda9cc5624645353406c94750f5c3008
        version: 1.8.6
      entries:
      - name: kubemq-operator-marketplace.1.8.6
        version: 1.8.6
      - name: kubemq-operator-marketplace.1.8.5
        version: 1.8.5
      - name: kubemq-operator-marketplace.v1.8.4
        version: 1.8.4
      - name: kubemq-operator-marketplace.v1.8.3
        version: 1.8.3
      name: alpha
    defaultChannel: alpha
    packageName: kubemq-operator-marketplace
    provider:
      name: Kubemq.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: candidate
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: fast
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: stable
    defaultChannel: stable
    packageName: stackgres-certified
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzounstructured-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzounstructured-operator.v2.0.101
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzounstructured.clusters.cambridgesemantics.com/v1",
                "kind": "AnzoUnstructured",
                "metadata": {
                  "name": "au01"
                },
                "spec": {
                  "auWorker": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-w",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-unstructured-worker@sha256:4db61bdf6dc8b42824a06d2f8d8aeecddde765d123b40a63a7bbf87845719fb2",
                                "name": "w",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  },
                  "msLeader": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-ms",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-microservices-leader@sha256:0579fd97e81e436eaff3b58c5395f13aa924940a3d51f0bdea5163d327620286",
                                "name": "ms",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/unstructured-operator@sha256:a47af068a7aa55c12be84c80d263feeed5597ecc434986b4f2fff782b568bb04
          createdAt: "2021-08-02"
          description: Kubernetes Operator for Anzo Unstructured
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzounstructured-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzounstructured-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoUnstructured is the Schema for the anzounstructureds
              API
            displayName: AnzoUnstructured Operator
            kind: AnzoUnstructured
            name: anzounstructureds.anzounstructured.clusters.cambridgesemantics.com
            version: v1
        description: |-
          The Anzo Unstructured Operator provides the way to install and configure an anzo unstructured setup on Red Hat K8S environment.
          Currently, this is possible only through existing Anzo installation.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v2.0.0/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers through [ Cambridge Semantics Customer Center ]( https://customercenter.cambridgesemantics.com/ ).
        displayName: Anzo Unstructured Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Database
        - Big Data
        links:
        - name: Anzo Unstructured Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v2.0.0/README.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/unstructured-operator@sha256:a47af068a7aa55c12be84c80d263feeed5597ecc434986b4f2fff782b568bb04
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-microservices-leader@sha256:0579fd97e81e436eaff3b58c5395f13aa924940a3d51f0bdea5163d327620286
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-unstructured-worker@sha256:4db61bdf6dc8b42824a06d2f8d8aeecddde765d123b40a63a7bbf87845719fb2
        version: 2.0.0
      entries:
      - name: anzounstructured-operator.v2.0.101
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anzounstructured-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://github.com/che-incubator/kubernetes-image-puller-operator
    name: kubernetes-imagepuller-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubernetes-imagepuller-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "che.eclipse.org/v1alpha1",
                "kind": "KubernetesImagePuller",
                "metadata": {
                  "name": "image-puller"
                },
                "spec": {
                  "daemonsetName": "k8s-image-puller",
                  "images": "java11-maven=quay.io/eclipse/che-java11-maven:next;che-theia=quay.io/eclipse/che-theia:next"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/eclipse/kubernetes-image-puller-operator:1.1.0
          createdAt: "2024-01-15T16:23:04Z"
          description: Create and manage kubernetes-image-puller instances.
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/che-incubator/kubernetes-image-puller-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubernetesImagePuller is the Schema for the kubernetesimagepullers
              API
            displayName: Kubernetes Image Puller
            kind: KubernetesImagePuller
            name: kubernetesimagepullers.che.eclipse.org
            version: v1alpha1
        description: "## About the managed application\n\nCreate a `KubernetesImagePuller`
          custom resource to automatically configure and run an instance of the [kubernetes-image-puller.](https://github.com/che-incubator/kubernetes-image-puller)\n##
          About this Operator\n\nThe `KubernetesImagePuller` custom resource understands
          the following fields in the `spec`:\n\n1. `configMapName` - The name of
          the `ConfigMap` to create\t\n2. `daemonsetName` - The name of the `DaemonSet`
          to be create.\n3. `deploymentName` - The name of the `kubernetes-image-puller`
          `Deployment` to create.\n4. `images` - A list of key-value pairs separated
          by semicolons of images to pull.  For example: `java=quay.io/eclipse/che-java8-maven:latest;theia=quay.io/eclipse/che-theia:next`\n5.
          `cachingIntervalHours` - The amount of time, in hours, between `DaemonSet`
          health checks.\n6. `cachingMemoryRequest` - The memory request for each
          cached image when the puller is running.\n7. `cachingMemoryLimit` - The
          memory limit for each cached image when the puller is running.\n8. `nodeSelector`
          - Node selector applied to pods created by the `DaemonSet`.\n## Prerequisites
          for enabling this Operator\n\nThe operator requires an existing namespace
          to be installed in.\nCheck the [README](https://github.com/che-incubator/kubernetes-image-puller-operator#installing-the-operator)
          for more details.\n"
        displayName: Kubernetes Image Puller Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - che
        - kubernetes-image-puller
        links:
        - name: Repository
          url: https://github.com/che-incubator/kubernetes-image-puller-operator
        maintainers:
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        provider:
          name: Red Hat, Inc.
          url: https://github.com/che-incubator/kubernetes-image-puller-operator
        relatedImages:
        - quay.io/eclipse/kubernetes-image-puller-operator:1.1.0
        - quay.io/brancz/kube-rbac-proxy:v0.11.0
        version: 1.1.0
      entries:
      - name: kubernetes-imagepuller-operator.v1.1.0
        version: 1.1.0
      - name: kubernetes-imagepuller-operator.v1.0.6
        version: 1.0.6
      - name: kubernetes-imagepuller-operator.v1.0.5
        version: 1.0.5
      - name: kubernetes-imagepuller-operator.v1.0.4
        version: 1.0.4
      - name: kubernetes-imagepuller-operator.v1.0.3
        version: 1.0.3
      - name: kubernetes-imagepuller-operator.v1.0.2
        version: 1.0.2
      - name: kubernetes-imagepuller-operator.v1.0.1
        version: 1.0.1
      - name: kubernetes-imagepuller-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: kubernetes-imagepuller-operator
    provider:
      name: Red Hat, Inc.
      url: https://github.com/che-incubator/kubernetes-image-puller-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Palo Alto Networks
      provider-url: ""
    name: pcc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pcc-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "name": "pcc-console",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "ConsoleDefender",
                "metadata": {
                  "name": "pcc-consoledefender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "defenderConfig": {
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Defender",
                "metadata": {
                  "name": "pcc-defender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "defenderConfig": {
                    "clusterAddress": "twistlock-console.example.com",
                    "consoleAddress": "https://twistlock-console.example.com:8083",
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: quay.io/prismacloud/pcc-operator@sha256:b8fcfbd6c51286c874e00db1bd35523386cec406fa4050ef44c0a887730cf9b8
          createdAt: "2021-11-15"
          description: Deploy Prisma Cloud Compute for cloud-native security in your
            clusters
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator
          support: pcc-integrations@paloaltonetworks.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Install the Prisma Cloud Compute Console and Defenders
            displayName: Console and Defenders
            kind: ConsoleDefender
            name: consoledefenders.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install the Prisma Cloud Compute Console
            displayName: Console
            kind: Console
            name: consoles.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install Prisma Cloud Compute Defenders
            displayName: Defenders
            kind: Defender
            name: defenders.pcc.paloaltonetworks.com
            version: v1alpha1
        description: |
          ## Features
          - Deploy Console
              - Create initial local administrator user
              - Add license
          - Deploy Defenders
          - Support for offline deployments
          - Upgrade Console
          - Upgrade Defenders

          ## Installation
          See the [Prisma Cloud Compute Operator documentation](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator) for installation guides.

          ## Support
          This operator is available "as is," and should be seen as community-supported; however, Palo Alto Networks will continue contributing whenever possible.
          Palo Alto Networks does not provide technical support or help with using or troubleshooting the operator through normal support channels.
          We encourage you to open GitHub [issues](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/issues) and [pull requests](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/pulls) to track bugs and feature requests.

          ## Other links
          [Prisma Cloud Compute product documentation](https://docs.paloaltonetworks.com/prisma/prisma-cloud/prisma-cloud-admin-compute.html)

          [Prisma Cloud Compute API documentation](https://prisma.pan.dev/api/cloud/cwpp)
        displayName: Prisma Cloud Compute Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - palo
        - palo alto
        - palo alto networks
        - paloalto
        - twistlock
        - compute
        - prisma
        - prisma cloud
        - prisma cloud compute
        - security
        - image
        - images
        - container
        - containers
        - docker
        - podman
        - vulnerability
        - vulnerabilities
        - compliance
        - runtime
        - scan
        - scanning
        - monitor
        - monitoring
        - alert
        - alerting
        links:
        - name: Prisma Cloud
          url: https://www.paloaltonetworks.com/prisma/cloud
        maintainers:
        - email: pcc-integrations@paloaltonetworks.com
          name: Wyatt Gill
        maturity: alpha
        provider:
          name: Palo Alto Networks
        relatedImages:
        - quay.io/prismacloud/pcc-operator@sha256:b8fcfbd6c51286c874e00db1bd35523386cec406fa4050ef44c0a887730cf9b8
        version: 0.2.0
      entries:
      - name: pcc-operator.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: pcc-operator
    provider:
      name: Palo Alto Networks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: koku-metrics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: koku-metrics-operator.v4.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
          createdAt: "2025-09-25T19:04:35Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: koku-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Cost Management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Koku Metrics Operator
          ## Introduction
          The `koku-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Koku Metrics Operator (`koku-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created KokuMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The KokuMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.2.0:
          * Enabled gathering and reporting NVIDIA GPU metrics for resource optimization.
          * (Bugfix) Fixed custom CA certificate validation and configuration issues.
          * (Bugfix) Fixed issue where ROS queries were missing workload metadata for namespaces labeled only with the legacy `insights_cost_management_optimizations` label.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **Progress towards FIPS 140 Compliance:** The operator is using the Go Cryptographic Module v1.0.0 to progress toward FIPS 140 compliance. Although this module is not validated at the time of this release, the operator aims to meet stricter security standards when the module does successfully achieve FIPS validation.
          * **API Name Change:** The Custom Resource Definition (CRD) for configuring the upstream `Koku Metrics Operator` was renamed from `KokuMetricsConfig` to `CostManagementMetricsConfig`.

          ### NOTE:
            * The API name change impacts only users of the upstream (community) Koku Metrics Operator.
            * The downstream (Red Hat-supported) Cost Management Metrics Operator is not impacted by this change and users should continue using their existing configurations.

          **Important Upgrade Instructions:**

            If you are upgrading the upstream Koku Metrics Operator to version 4.0.0 or higher, you must manually migrate your configuration. The operator will no longer recognize existing `KokuMetricsConfig` resources.

            To successfully upgrade and retain your operator's configuration, complete the following steps. You can use the provided `oc` commands replacing the angle brackets (`< >`) with your specific values:

            1. Backup existing configuration (recommended):

               Before you upgrade, retrieve the details of your current `KokuMetricsConfig` instance to help you recreate it with the new API name.

               ```
               oc get kokumetricsconfig -n koku-metrics-operator <your-config-name> -o yaml > koku-metrics-config-backup.yaml
               ```

            2. Delete the previous `KokuMetricsConfig` instance (recommended):

                After the operator is upgraded to `version 4.0.0`, delete any existing `KokuMetricsConfig` resources.

                ```
                oc delete kokumetricsconfig -n koku-metrics-operator <your-config-name>
                ```

            3. Create the new `CostManagementMetricsConfig` instance:

                Using the information from your backup (or a new configuration), create a new `CostManagementMetricsConfig` instance. Adjust the spec section based on your specific needs. For more information about configurations, [refer to the configurable parameters](#configurable-parameters).

            4. Apply your new configuration:

                ```
                oc apply -f <your-costmanagementmetricsconfig-file>.yaml -n koku-metrics-operator
                ```

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created KokuMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the KokuMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `KokuMetricsConfig`, creating a new `KokuMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `koku-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: koku-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the KokuMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the KokuMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon KokuMetricsConfig creation** (This parameter will start to appear in KokuMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the KokuMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the koku-metrics-operator
          **Note** There are separate instructions for configuring the `koku-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `koku-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the koku-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `koku-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `koku-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the koku-metrics-operator for a restricted network
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `koku-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: koku-metrics-operator
                labels:
                  app: koku-metrics-operator
              spec:
                volumes:
                - name: koku-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: koku-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: koku-metrics-operator-reports
                    mountPath: /tmp/koku-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/koku-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/koku-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `koku-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the KokuMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Koku Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Koku Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
        version: 4.2.0
      entries:
      - name: koku-metrics-operator.v4.2.0
        version: 4.2.0
      - name: koku-metrics-operator.v4.1.0
        version: 4.1.0
      - name: koku-metrics-operator.v4.0.0
        version: 4.0.0
      - name: koku-metrics-operator.v3.3.2
        version: 3.3.2
      - name: koku-metrics-operator.v3.3.1
        version: 3.3.1
      - name: koku-metrics-operator.v3.3.0
        version: 3.3.0
      - name: koku-metrics-operator.v3.2.1
        version: 3.2.1
      - name: koku-metrics-operator.v3.2.0
        version: 3.2.0
      - name: koku-metrics-operator.v3.1.0
        version: 3.1.0
      - name: koku-metrics-operator.v3.0.1
        version: 3.0.1
      - name: koku-metrics-operator.v3.0.0
        version: 3.0.0
      - name: koku-metrics-operator.v2.0.0
        version: 2.0.0
      - name: koku-metrics-operator.v1.2.0
        version: 1.2.0
      - name: koku-metrics-operator.v1.1.9
        version: 1.1.9
      - name: koku-metrics-operator.v1.1.8
        version: 1.1.8
      - name: koku-metrics-operator.v1.1.7
        version: 1.1.7
      - name: koku-metrics-operator.v1.1.6
        version: 1.1.6
      - name: koku-metrics-operator.v1.1.5
        version: 1.1.5
      - name: koku-metrics-operator.v1.1.4
        version: 1.1.4
      - name: koku-metrics-operator.v1.1.3
        version: 1.1.3
      - name: koku-metrics-operator.v1.1.2
        version: 1.1.2
      - name: koku-metrics-operator.v1.1.1
        version: 1.1.1
      - name: koku-metrics-operator.v0.9.8
        version: 0.9.8
      - name: koku-metrics-operator.v0.9.7
        version: 0.9.7
      - name: koku-metrics-operator.v0.9.6
        version: 0.9.6
      - name: koku-metrics-operator.v0.9.5
        version: 0.9.5
      - name: koku-metrics-operator.v0.9.4
        version: 0.9.4
      - name: koku-metrics-operator.v0.9.3
        version: 0.9.3
      - name: koku-metrics-operator.v0.9.2
        version: 0.9.2
      - name: koku-metrics-operator.v0.9.1
        version: 0.9.1
      - name: koku-metrics-operator.v0.9.0
        version: 0.9.0
      name: alpha
    - currentCSV: koku-metrics-operator.v4.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
          createdAt: "2025-09-25T19:04:35Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: koku-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Cost Management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Koku Metrics Operator
          ## Introduction
          The `koku-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Koku Metrics Operator (`koku-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created KokuMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The KokuMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.2.0:
          * Enabled gathering and reporting NVIDIA GPU metrics for resource optimization.
          * (Bugfix) Fixed custom CA certificate validation and configuration issues.
          * (Bugfix) Fixed issue where ROS queries were missing workload metadata for namespaces labeled only with the legacy `insights_cost_management_optimizations` label.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **Progress towards FIPS 140 Compliance:** The operator is using the Go Cryptographic Module v1.0.0 to progress toward FIPS 140 compliance. Although this module is not validated at the time of this release, the operator aims to meet stricter security standards when the module does successfully achieve FIPS validation.
          * **API Name Change:** The Custom Resource Definition (CRD) for configuring the upstream `Koku Metrics Operator` was renamed from `KokuMetricsConfig` to `CostManagementMetricsConfig`.

          ### NOTE:
            * The API name change impacts only users of the upstream (community) Koku Metrics Operator.
            * The downstream (Red Hat-supported) Cost Management Metrics Operator is not impacted by this change and users should continue using their existing configurations.

          **Important Upgrade Instructions:**

            If you are upgrading the upstream Koku Metrics Operator to version 4.0.0 or higher, you must manually migrate your configuration. The operator will no longer recognize existing `KokuMetricsConfig` resources.

            To successfully upgrade and retain your operator's configuration, complete the following steps. You can use the provided `oc` commands replacing the angle brackets (`< >`) with your specific values:

            1. Backup existing configuration (recommended):

               Before you upgrade, retrieve the details of your current `KokuMetricsConfig` instance to help you recreate it with the new API name.

               ```
               oc get kokumetricsconfig -n koku-metrics-operator <your-config-name> -o yaml > koku-metrics-config-backup.yaml
               ```

            2. Delete the previous `KokuMetricsConfig` instance (recommended):

                After the operator is upgraded to `version 4.0.0`, delete any existing `KokuMetricsConfig` resources.

                ```
                oc delete kokumetricsconfig -n koku-metrics-operator <your-config-name>
                ```

            3. Create the new `CostManagementMetricsConfig` instance:

                Using the information from your backup (or a new configuration), create a new `CostManagementMetricsConfig` instance. Adjust the spec section based on your specific needs. For more information about configurations, [refer to the configurable parameters](#configurable-parameters).

            4. Apply your new configuration:

                ```
                oc apply -f <your-costmanagementmetricsconfig-file>.yaml -n koku-metrics-operator
                ```

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created KokuMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the KokuMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `KokuMetricsConfig`, creating a new `KokuMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `koku-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: koku-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the KokuMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the KokuMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon KokuMetricsConfig creation** (This parameter will start to appear in KokuMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the KokuMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the koku-metrics-operator
          **Note** There are separate instructions for configuring the `koku-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `koku-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the koku-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `koku-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `koku-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the koku-metrics-operator for a restricted network
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `koku-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: koku-metrics-operator
                labels:
                  app: koku-metrics-operator
              spec:
                volumes:
                - name: koku-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: koku-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: koku-metrics-operator-reports
                    mountPath: /tmp/koku-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/koku-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/koku-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `koku-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the KokuMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Koku Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Koku Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
        version: 4.2.0
      entries:
      - name: koku-metrics-operator.v4.2.0
        version: 4.2.0
      - name: koku-metrics-operator.v4.1.0
        version: 4.1.0
      - name: koku-metrics-operator.v4.0.0
        version: 4.0.0
      - name: koku-metrics-operator.v3.3.2
        version: 3.3.2
      - name: koku-metrics-operator.v3.3.1
        version: 3.3.1
      - name: koku-metrics-operator.v3.3.0
        version: 3.3.0
      - name: koku-metrics-operator.v3.2.1
        version: 3.2.1
      - name: koku-metrics-operator.v3.2.0
        version: 3.2.0
      - name: koku-metrics-operator.v3.1.0
        version: 3.1.0
      - name: koku-metrics-operator.v3.0.1
        version: 3.0.1
      - name: koku-metrics-operator.v3.0.0
        version: 3.0.0
      - name: koku-metrics-operator.v2.0.0
        version: 2.0.0
      - name: koku-metrics-operator.v1.2.0
        version: 1.2.0
      - name: koku-metrics-operator.v1.1.9
        version: 1.1.9
      - name: koku-metrics-operator.v1.1.8
        version: 1.1.8
      - name: koku-metrics-operator.v1.1.7
        version: 1.1.7
      - name: koku-metrics-operator.v1.1.6
        version: 1.1.6
      - name: koku-metrics-operator.v1.1.5
        version: 1.1.5
      - name: koku-metrics-operator.v1.1.4
        version: 1.1.4
      - name: koku-metrics-operator.v1.1.3
        version: 1.1.3
      - name: koku-metrics-operator.v1.1.2
        version: 1.1.2
      - name: koku-metrics-operator.v1.1.1
        version: 1.1.1
      - name: koku-metrics-operator.v0.9.8
        version: 0.9.8
      - name: koku-metrics-operator.v0.9.7
        version: 0.9.7
      - name: koku-metrics-operator.v0.9.6
        version: 0.9.6
      - name: koku-metrics-operator.v0.9.5
        version: 0.9.5
      - name: koku-metrics-operator.v0.9.4
        version: 0.9.4
      - name: koku-metrics-operator.v0.9.3
        version: 0.9.3
      - name: koku-metrics-operator.v0.9.2
        version: 0.9.2
      - name: koku-metrics-operator.v0.9.1
        version: 0.9.1
      - name: koku-metrics-operator.v0.9.0
        version: 0.9.0
      name: beta
    defaultChannel: beta
    packageName: koku-metrics-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Lenovo (Shanghai) Computer Technology Co., Ltd.
      provider-url: ""
    name: lenovo-locd-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lenovo-locd-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "locd.lenovo.com/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample",
                  "namespace": "locd-operator-system"
                },
                "spec": {
                  "affinity": {},
                  "argocd": {
                    "password": "xxxxxxxxxxx",
                    "server": "argocd.svc.cluster.local",
                    "user": "admin"
                  },
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "core": {
                    "image": "registry.connect.redhat.com/loc-devops/core@sha256:f771217b3f7849e1b63cc6abb93b8b0fc9636fd4a201e31807a15f773138c609",
                    "pullPolicy": "Always"
                  },
                  "fullnameOverride": "",
                  "global": {
                    "managementNamespace": "loc-devops",
                    "serviceEndpoint": "loc-devops.domain",
                    "serviceaccount": {
                      "annotations": {
                        "managedBy": "LOC DevOps"
                      },
                      "create": true,
                      "name": "loc-devops-admin"
                    }
                  },
                  "gui": {
                    "image": "registry.connect.redhat.com/loc-devops/gui@sha256:65142a36095188e879467038936b7b6b6ee39dcea6d42a437fa7066fb2bc0f44",
                    "pullPolicy": "Always"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "className": "nginx",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "loc-devops.domain",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "Prefix",
                            "servicePort": 8080
                          },
                          {
                            "path": "/api",
                            "pathType": "Prefix",
                            "servicePort": 8000
                          },
                          {
                            "path": "/swagger",
                            "pathType": "Prefix",
                            "servicePort": 8000
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "replicaCount": 1,
                  "resources": {},
                  "securityContext": {},
                  "service": {
                    "port": {
                      "core": 8000,
                      "gui": 8080
                    },
                    "type": "ClusterIP"
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Cluster
            name: clusters.locd.lenovo.com
            version: v1alpha1
        description: This is operator used to manage components of Lenovo Open Cloud
          DevOps
        displayName: lenovo-locd-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Application management
        links:
        - name: LOCD Operator
          url: https://locd-operator.domain
        maintainers:
        - email: xujt5@lenovo.com
          name: jintaoxu
        maturity: alpha
        provider:
          name: Lenovo (Shanghai) Computer Technology Co., Ltd.
        relatedImages:
        - registry.connect.redhat.com/loc-devops/gui@sha256:65142a36095188e879467038936b7b6b6ee39dcea6d42a437fa7066fb2bc0f44
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b5786bbbef725badf3dfcc2c2c7a86ead5ebb584c978c47aae8b9a62e241b80d
        - registry.connect.redhat.com/loc-devops/operator@sha256:4c17a3688ae3098e4b5c2c0530f49b1da8dedcc6fc1d3cd848513a1f189812a7
        - registry.connect.redhat.com/loc-devops/core@sha256:f771217b3f7849e1b63cc6abb93b8b0fc9636fd4a201e31807a15f773138c609
        version: 0.1.0
      entries:
      - name: lenovo-locd-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: lenovo-locd-operator
    provider:
      name: Lenovo (Shanghai) Computer Technology Co., Ltd.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtr-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtr-operator.1.2.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
             {
               "apiVersion": "windup.jboss.org/v1",
               "kind": "Windup",
               "metadata": {
                 "namespace": "mtr",
                 "name": "mtr",
                 "labels": {
                   "application": "windup"
                 }
               },
               "spec": {}
             }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mtr/mtr-rhel8-operator@sha256:4df7c4ad353339e8b958ffc651e811667767c8a0dcdd61f89352c926da52a513
          createdAt: "2024-09-03T17:22:23Z"
          description: Migration Toolkit for Runtimes is an analysis tool that supports
            the modernization and migration of Java applications.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          imagePullPolicy: Always
          operatorframework.io/suggested-namespace: mtr
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/windup/windup-operator/
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "Create an MTR Web Console application.  \n\nTo access the
              MTR Web Console application from the \"Developer\" perspective, please
              go to Topology and click on the \"Open URL\" icon on the web-console
              pod.  \n\nAlternatively from the \"Administrator\" perspective, please
              go to Networking->Routes and click on the Location hyperlink.  \n\nIn
              order to connect with default login credentials, please use \"migration\"
              as username and \"password\" as password.  \n\nKnown issue  \nIf you
              want to customize the MTR Web Console instance's parameters and you
              can not see any in the `Form view`, please switch to the `YAML view`
              and change them as needed.\n"
            displayName: Migration Toolkit for Runtimes
            kind: Windup
            name: windups.windup.jboss.org
            version: v1
        description: "The Migration Toolkit for Runtimes (MTR) is a web console application
          that supports large-scale Java application modernization and migration projects
          across a broad range of transformations and use cases.  \n\nIt analyzes
          application code, supports effort estimation, accelerates code migration,
          and enables users to move applications to containers.  \n\nFor more information
          please refer to the [https://developers.redhat.com/products/mtr/overview](https://developers.redhat.com/products/mtr/overview)
          page.\n"
        displayName: Migration Toolkit for Runtimes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - security
        - alerting
        - metric
        - troubleshooting
        - run-time
        - migration
        - modernization
        links:
        - name: Website
          url: https://developers.redhat.com/products/mtr/overview
        - name: GitHub
          url: https://github.com/windup/windup-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_runtimes
        maintainers:
        - email: migrate@redhat.com
          name: Migration Toolkit for Runtimes
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mtr/mtr-rhel8-operator@sha256:4df7c4ad353339e8b958ffc651e811667767c8a0dcdd61f89352c926da52a513
        - registry.redhat.io/mtr/mtr-web-container-rhel8@sha256:cb88a5453b506ca70563537ec6db2f44ff8cb4050c9652ec86eb5128aa313003
        - registry.redhat.io/mtr/mtr-web-executor-container-rhel8@sha256:2f0622774367f33efc321926196681322ab1535144e6a0785608c8050fc59f2e
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        version: 1.2.7
      entries:
      - name: mtr-operator.1.2.7
        version: 1.2.7
      - name: mtr-operator.1.2.6
        version: 1.2.6
      - name: mtr-operator.1.2.5
        version: 1.2.5
      - name: mtr-operator.1.2.4
        version: 1.2.4
      - name: mtr-operator.1.2.3
        version: 1.2.3
      - name: mtr-operator.1.2.2
        version: 1.2.2
      - name: mtr-operator.1.2.1
        version: 1.2.1
      - name: mtr-operator.1.2.0
        version: 1.2.0
      - name: mtr-operator.1.1.1
        version: 1.1.1
      - name: mtr-operator.1.1.0
        version: 1.1.0
      - name: mtr-operator.1.0.3
        version: 1.0.3
      - name: mtr-operator.1.0.2
        version: 1.0.2
      - name: mtr-operator.1.0.1
        version: 1.0.1
      - name: mtr-operator.1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: mtr-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: GitLab Inc
      provider-url: https://about.gitlab.com/
    name: gitlab-operator-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "certmanager": {
                        "install": false
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-26T07:16:35Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3281fe12aaae428fd7102332e7c0d7033515042d7830eba3953891ab0c3e3a29
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      - name: gitlab-operator-kubernetes.v2.1.1
        version: 2.1.1
      - name: gitlab-operator-kubernetes.v2.1.0
        version: 2.1.0
      - name: gitlab-operator-kubernetes.v2.0.2
        version: 2.0.2
      - name: gitlab-operator-kubernetes.v2.0.1
        version: 2.0.1
      - name: gitlab-operator-kubernetes.v2.0.0
        version: 2.0.0
      - name: gitlab-operator-kubernetes.v1.12.2
        version: 1.12.2
      - name: gitlab-operator-kubernetes.v1.12.1
        version: 1.12.1
      - name: gitlab-operator-kubernetes.v1.12.0
        version: 1.12.0
      - name: gitlab-operator-kubernetes.v1.11.3
        version: 1.11.3
      - name: gitlab-operator-kubernetes.v1.11.2
        version: 1.11.2
      - name: gitlab-operator-kubernetes.v1.11.1
        version: 1.11.1
      - name: gitlab-operator-kubernetes.v1.11.0
        version: 1.11.0
      - name: gitlab-operator-kubernetes.v1.10.2
        version: 1.10.2
      - name: gitlab-operator-kubernetes.v1.10.1
        version: 1.10.1
      - name: gitlab-operator-kubernetes.v1.10.0
        version: 1.10.0
      - name: gitlab-operator-kubernetes.v1.9.2
        version: 1.9.2
      - name: gitlab-operator-kubernetes.v1.9.1
        version: 1.9.1
      - name: gitlab-operator-kubernetes.v1.9.0
        version: 1.9.0
      - name: gitlab-operator-kubernetes.v1.8.3
        version: 1.8.3
      - name: gitlab-operator-kubernetes.v1.8.2
        version: 1.8.2
      - name: gitlab-operator-kubernetes.v1.8.1
        version: 1.8.1
      - name: gitlab-operator-kubernetes.v1.8.0
        version: 1.8.0
      - name: gitlab-operator-kubernetes.v1.7.2
        version: 1.7.2
      - name: gitlab-operator-kubernetes.v1.7.1
        version: 1.7.1
      - name: gitlab-operator-kubernetes.v1.7.0
        version: 1.7.0
      - name: gitlab-operator-kubernetes.v1.6.1
        version: 1.6.1
      - name: gitlab-operator-kubernetes.v1.6.0
        version: 1.6.0
      - name: gitlab-operator-kubernetes.v1.5.1
        version: 1.5.1
      - name: gitlab-operator-kubernetes.v1.4.2
        version: 1.4.2
      - name: gitlab-operator-kubernetes.v1.4.1
        version: 1.4.1
      - name: gitlab-operator-kubernetes.v1.4.0
        version: 1.4.0
      - name: gitlab-operator-kubernetes.v1.3.3
        version: 1.3.3
      - name: gitlab-operator-kubernetes.v1.3.2
        version: 1.3.2
      - name: gitlab-operator-kubernetes.v1.3.1
        version: 1.3.1
      - name: gitlab-operator-kubernetes.v1.3.0
        version: 1.3.0
      name: stable
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "certmanager": {
                        "install": false
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-26T07:16:35Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3281fe12aaae428fd7102332e7c0d7033515042d7830eba3953891ab0c3e3a29
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      - name: gitlab-operator-kubernetes.v2.1.1
        version: 2.1.1
      - name: gitlab-operator-kubernetes.v2.1.0
        version: 2.1.0
      - name: gitlab-operator-kubernetes.v2.0.2
        version: 2.0.2
      - name: gitlab-operator-kubernetes.v2.0.1
        version: 2.0.1
      - name: gitlab-operator-kubernetes.v2.0.0
        version: 2.0.0
      - name: gitlab-operator-kubernetes.v1.12.2
        version: 1.12.2
      - name: gitlab-operator-kubernetes.v1.12.1
        version: 1.12.1
      - name: gitlab-operator-kubernetes.v1.12.0
        version: 1.12.0
      - name: gitlab-operator-kubernetes.v1.11.3
        version: 1.11.3
      - name: gitlab-operator-kubernetes.v1.11.2
        version: 1.11.2
      - name: gitlab-operator-kubernetes.v1.11.1
        version: 1.11.1
      - name: gitlab-operator-kubernetes.v1.11.0
        version: 1.11.0
      - name: gitlab-operator-kubernetes.v1.10.2
        version: 1.10.2
      - name: gitlab-operator-kubernetes.v1.10.1
        version: 1.10.1
      - name: gitlab-operator-kubernetes.v1.10.0
        version: 1.10.0
      - name: gitlab-operator-kubernetes.v1.9.2
        version: 1.9.2
      - name: gitlab-operator-kubernetes.v1.9.1
        version: 1.9.1
      - name: gitlab-operator-kubernetes.v1.9.0
        version: 1.9.0
      - name: gitlab-operator-kubernetes.v1.8.3
        version: 1.8.3
      - name: gitlab-operator-kubernetes.v1.8.2
        version: 1.8.2
      - name: gitlab-operator-kubernetes.v1.8.1
        version: 1.8.1
      - name: gitlab-operator-kubernetes.v1.8.0
        version: 1.8.0
      - name: gitlab-operator-kubernetes.v1.7.2
        version: 1.7.2
      - name: gitlab-operator-kubernetes.v1.7.1
        version: 1.7.1
      - name: gitlab-operator-kubernetes.v1.7.0
        version: 1.7.0
      - name: gitlab-operator-kubernetes.v1.6.1
        version: 1.6.1
      - name: gitlab-operator-kubernetes.v1.6.0
        version: 1.6.0
      - name: gitlab-operator-kubernetes.v1.5.1
        version: 1.5.1
      - name: gitlab-operator-kubernetes.v1.4.2
        version: 1.4.2
      - name: gitlab-operator-kubernetes.v1.4.1
        version: 1.4.1
      - name: gitlab-operator-kubernetes.v1.4.0
        version: 1.4.0
      - name: gitlab-operator-kubernetes.v1.3.3
        version: 1.3.3
      - name: gitlab-operator-kubernetes.v1.3.2
        version: 1.3.2
      - name: gitlab-operator-kubernetes.v1.3.1
        version: 1.3.1
      - name: gitlab-operator-kubernetes.v1.3.0
        version: 1.3.0
      name: unstable
    defaultChannel: stable
    packageName: gitlab-operator-kubernetes
    provider:
      name: GitLab Inc
      url: https://about.gitlab.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Silicom Ltd.
      provider-url: http://www.silicom-usa.com
    name: silicom-sts-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: silicom-sts-operator.v0.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "namespace": "openshift-operators",
                  "nodeSelector": {
                    "sts.silicom.com/config": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1"
                },
                "spec": {
                  "aprLevel": 0,
                  "domainNumber": 24,
                  "esmcMode": 1,
                  "forwardable": 0,
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-TBC.8275.1",
                  "mode10MHz": 2,
                  "modePPS": 2,
                  "namespace": "openshift-operators",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  },
                  "phyLedsCtl": 0,
                  "priority2": 128,
                  "src10MHz": 1,
                  "srcPPS": 1,
                  "syncOption": 1,
                  "synceCpu": 10,
                  "synceRecClkPort": 0,
                  "tracePtpMsg": -1,
                  "twoStep": 0
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "worker2"
                },
                "spec": null,
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "images": null,
                  "sro": null
                }
              }
            ]
          capabilities: Basic Install
          categories: Drivers and plugins
          certified: "false"
          containerImage: quay.io/silicom/sts-operator:0.0.6
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicomdk/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicomDK/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/silicomDK/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - quay.io/silicom/sts-plugin@sha256:44339fd400d46d9dddb02c2da47c6f9f96ad2605a5c6de8a7dce2f8c3da3bb57
        - quay.io/silicom/phc2sys@sha256:0f3b63a3059b7d6158a916814281097f0d418b99a5fac9e8e5fd9bae58fb0b74
        - quay.io/silicom/tsyncd@sha256:78be78ec8939f025f9a8b899693ab77b3990116d1c7ffb2115c620b7c3382d9d
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/silicom/sts-operator:0.0.6
        - quay.io/silicom/grpc-tsyncd@sha256:aac7233423439e352d8eb39cef63e479c30c2f80b063c84ffc8797cc5a045117
        - quay.io/silicom/tsync_extts@sha256:be6a5950a5fdc6fdf8015767901856bf173d01bdc803be7141e4f9795ab1478a
        - quay.io/silicom/ice-driver-src@sha256:ca728f648769d31501cdedda92554bcfa890754405b0b06d41629cea47707da8
        - quay.io/silicom/gpsd@sha256:649e8a6c524920e1c564d8f797b8d64235b2527aa923f94bc9c7a0b7aba8ae7f
        version: 0.0.6
      entries:
      - name: silicom-sts-operator.v0.0.6
        version: 0.0.6
      - name: silicom-sts-operator.v0.0.5
        version: 0.0.5
      - name: silicom-sts-operator.v0.0.4
        version: 0.0.4
      - name: silicom-sts-operator.v0.0.3
        version: 0.0.3
      - name: silicom-sts-operator.v0.0.2
        version: 0.0.2
      - name: silicom-sts-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: silicom-sts-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "sts-silicom"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "namespace": "sts-silicom",
                  "nodeSelector": {
                    "sts.silicom.com/config": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1"
                },
                "spec": {
                  "aprLevel": 0,
                  "domainNumber": 24,
                  "esmcMode": 1,
                  "forwardable": 0,
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-TBC.8275.1",
                  "mode10MHz": 2,
                  "modePPS": 2,
                  "namespace": "sts-silicom",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  },
                  "phyLedsCtl": 0,
                  "priority2": 128,
                  "src10MHz": 1,
                  "srcPPS": 1,
                  "syncOption": 1,
                  "synceCpu": 10,
                  "synceRecClkPort": 0,
                  "tracePtpMsg": -1,
                  "twoStep": 0
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "worker2"
                },
                "spec": null,
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config"
                },
                "spec": {
                  "gpsSvcPort": 2947,
                  "grpcSvcPort": 50051,
                  "images": {
                    "gpsd": "quay.io/silicom/gpsd:3.23.1",
                    "grpcTsyncd": "quay.io/silicom/grpc-tsyncd:2.1.0.0",
                    "phcs2Sys": "quay.io/silicom/phcs2sys:3.1.1",
                    "stsPlugin": "quay.io/silicom/sts-plugin:0.0.2",
                    "ts2Phcs": "quay.io/silicom/ts2phcs:1.0.0",
                    "tsyncd": "quay.io/silicom/tsyncd:2.1.0.0"
                  },
                  "namespace": "sts-silicom",
                  "sro": {
                    "chart": {
                      "name": "ice-special-resource",
                      "repository": {
                        "name": "ice-special-resource",
                        "url": "http://ice-driver-src"
                      },
                      "version": "0.0.1"
                    },
                    "iceVersion": "1.7.16",
                    "namespace": "openshift-operators",
                    "srcImage": "quay.io/silicom/ice-driver-src:1.7.16",
                    "srcSvcPort": 3000
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Drivers and plugins
          certified: "false"
          containerImage: quay.io/silicom/sts-operator:0.0.2
          createdAt: 2021-11-17T10:21:24 CET
          description: This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of Silicom STS device
            plugins for Kubernetes.
          operators.operatorframework.io.bundle.channels.v1: stable,development
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicomdk/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicomDK/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/silicomDK/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/silicom/sts-operator:0.0.2
        version: 0.0.2
      entries:
      - name: silicom-sts-operator.v0.0.2
        version: 0.0.2
      - name: silicom-sts-operator.v0.0.1
        version: 0.0.1
      name: beta
    defaultChannel: alpha
    packageName: silicom-sts-operator
    provider:
      name: Silicom Ltd.
      url: http://www.silicom-usa.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-logging
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-logging.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Logging",
                "metadata": {
                  "name": "logging-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "elasticsearch": {
                    "storage": "16Gi",
                    "tag": "7.17.6-3747-35"
                  },
                  "environment": "openshift",
                  "fluentbit": {
                    "runOnMaster": false,
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "grafana": {
                    "dashboards": [
                      "wap",
                      "load-core"
                    ],
                    "dataSources": [
                      "wap",
                      "load-core"
                    ]
                  },
                  "image": {
                    "elasticInitImage": "rhel8/toolbox@sha256:a753c24cb8b03066d494f010ad654d23fb0559f14d7121542298da3e0d48340d",
                    "elasticInitRepository": "registry.access.redhat.com",
                    "fbImage": "fluent-bit:1.8.15-3747-20",
                    "fbRepository": "prod.harbor.keysight.digital/loadcore",
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "createClusterRoles": true,
                    "serviceAccountName": "load-core-service-account"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:33:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Logging
            name: loggings.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Logging Operator
        displayName: Keysight WAP Logging Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Logging
          url: https://keysight-wap-logging.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-logging@sha256:f1e3b3ad84cd0f96c92cdaf4609e29257035cb982a761e832dae2095e636c721
        version: 0.1.21
      entries:
      - name: keysight-wap-logging.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-logging
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hazelcast, Inc
      provider-url: ""
    name: hazelcast-platform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hazelcast-platform-operator.v5.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "cache-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "CronHotBackup",
                "metadata": {
                  "name": "cronhotbackup-sample"
                },
                "spec": {
                  "hotBackupTemplate": {
                    "metadata": {
                      "labels": {
                        "cron-hotbackup": "true"
                      }
                    },
                    "spec": {
                      "hazelcastResourceName": "hazelcast"
                    }
                  },
                  "schedule": "*/30 * * * *"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Flow",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hazelcast-platform-operator"
                  },
                  "name": "flow"
                },
                "spec": {
                  "database": {
                    "host": "flow-postgresql",
                    "secretName": "flow-db-secret"
                  },
                  "env": [
                    {
                      "name": "OPTIONS",
                      "value": "--flow.analytics.persistRemoteCallResponses=true --flow.analytics.persistResults=false --flow.config.custom.managementCenterUrl=http://34.140.160.158/mc\n"
                    }
                  ],
                  "externalConnectivity": {
                    "ingress": {
                      "annotations": {
                        "kubernetes.io/ingress.class": "nginx",
                        "nginx.ingress.kubernetes.io/enable-opentelemetry": "true",
                        "nginx.ingress.kubernetes.io/ssl-redirect": "false"
                      },
                      "hostname": "",
                      "ingressClassName": "nginx",
                      "path": "/"
                    }
                  },
                  "imagePullSecrets": [
                    {
                      "name": "hazelcast-cloud-hazelcast-connect-demo-pull-secret"
                    }
                  ],
                  "licenseKeySecretName": "flow-mc-license-key",
                  "repository": "quay.io/hazelcast_cloud/hazelcast-flow-internal",
                  "size": 3,
                  "version": "next"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast"
                },
                "spec": {
                  "clusterSize": 3,
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast-sample"
                },
                "spec": {
                  "clusterSize": 3,
                  "jet": {
                    "bucketConfig": {
                      "bucketURI": "gs://operator-user-code/jetJobs",
                      "secretName": "br-secret-gcp"
                    },
                    "enabled": true,
                    "resourceUploadEnabled": true
                  },
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HazelcastEndpoint",
                "metadata": {
                  "name": "hazelcastendpoint-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "port": 5701,
                  "type": "Discovery"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HotBackup",
                "metadata": {
                  "name": "hot-backup"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-longrun"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-longrun-2.0.0.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-1.0.2.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJobSnapshot",
                "metadata": {
                  "name": "jetjobsnapshot-sample"
                },
                "spec": {
                  "cancelJob": false,
                  "jetJobResourceName": "jet-job"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ManagementCenter",
                "metadata": {
                  "name": "managementcenter"
                },
                "spec": {
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "hazelcastClusters": [
                    {
                      "address": "hazelcast",
                      "name": "dev"
                    }
                  ],
                  "licenseKeySecretName": "hazelcast-license-key",
                  "persistence": {
                    "enabled": true,
                    "size": "10Gi"
                  },
                  "repository": "hazelcast/management-center"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Map",
                "metadata": {
                  "name": "map"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "MultiMap",
                "metadata": {
                  "name": "multimap-sample"
                },
                "spec": {
                  "backupCount": 3,
                  "binary": true,
                  "collectionType": "LIST",
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "queue-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ReplicatedMap",
                "metadata": {
                  "name": "replicatedmap-sample"
                },
                "spec": {
                  "asyncFillup": true,
                  "hazelcastResourceName": "hazelcast",
                  "inMemoryFormat": "OBJECT"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "topic-sample"
                },
                "spec": {
                  "globalOrderingEnabled": true,
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "UserCodeNamespace",
                "metadata": {
                  "name": "ucn-sample"
                },
                "spec": {
                  "bucketConfig": {
                    "bucketURI": "gs://operator-user-code/userCodeNamespace",
                    "secretName": "br-secret-gcp"
                  },
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "VectorCollection",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hazelcast-platform-operator"
                  },
                  "name": "vectorcollection-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "indexes": [
                    {
                      "dimension": 10,
                      "efConstruction": 256,
                      "maxDegree": 32,
                      "metric": "Dot",
                      "useDeduplication": false
                    }
                  ]
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanReplication",
                "metadata": {
                  "name": "wanreplication-sample"
                },
                "spec": {
                  "endpoints": "35.192.33.252",
                  "resources": [
                    {
                      "kind": "Map",
                      "name": "map-sample"
                    },
                    {
                      "kind": "Hazelcast",
                      "name": "hz-sample"
                    }
                  ],
                  "targetClusterName": "dev"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanSync",
                "metadata": {
                  "name": "wansync-sample"
                },
                "spec": {
                  "wanReplicationResourceName": "wan-replication-sample"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: docker.io/hazelcast/hazelcast-platform-operator@sha256:59328a4ea7a4e2110e0750364db792cf266f1e7231e3bb321e3de09d535dab23
          createdAt: "2025-03-13T11:59:59Z"
          description: Install Hazelcast clusters in Kubernetes environments.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hazelcast/hazelcast-platform-operator/
          support: Hazelcast, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.hazelcast.com
            version: v1alpha1
          - description: CronHotBackup is the Schema for the cronhotbackups API
            displayName: Cron Hot Backup
            kind: CronHotBackup
            name: cronhotbackups.hazelcast.com
            version: v1alpha1
          - description: Flow is the Schema for the flows API
            displayName: Flow
            kind: Flow
            name: flows.hazelcast.com
            version: v1alpha1
          - description: HazelcastEndpoint is the Schema for the hazelcastendpoints
              API
            displayName: Hazelcast Endpoint
            kind: HazelcastEndpoint
            name: hazelcastendpoints.hazelcast.com
            version: v1alpha1
          - description: Hazelcast is the Schema for the hazelcasts API
            displayName: Hazelcast
            kind: Hazelcast
            name: hazelcasts.hazelcast.com
            version: v1alpha1
          - description: HotBackup is the Schema for the hot backup API
            displayName: Hot Backup
            kind: HotBackup
            name: hotbackups.hazelcast.com
            version: v1alpha1
          - description: JetJob is the Schema for the jetjobs API
            displayName: Jet Job
            kind: JetJob
            name: jetjobs.hazelcast.com
            version: v1alpha1
          - description: JetJobSnapshot is the Schema for the jetjobsnapshots API
            displayName: Jet Job Snapshot
            kind: JetJobSnapshot
            name: jetjobsnapshots.hazelcast.com
            version: v1alpha1
          - description: ManagementCenter is the Schema for the managementcenters
              API
            displayName: Management Center
            kind: ManagementCenter
            name: managementcenters.hazelcast.com
            version: v1alpha1
          - description: Map is the Schema for the maps API
            displayName: Map
            kind: Map
            name: maps.hazelcast.com
            version: v1alpha1
          - description: MultiMap is the Schema for the multimaps API
            displayName: Multi Map
            kind: MultiMap
            name: multimaps.hazelcast.com
            version: v1alpha1
          - description: Queue is the Schema for the queues API
            displayName: Queue
            kind: Queue
            name: queues.hazelcast.com
            version: v1alpha1
          - description: ReplicatedMap is the Schema for the replicatedmaps API
            displayName: Replicated Map
            kind: ReplicatedMap
            name: replicatedmaps.hazelcast.com
            version: v1alpha1
          - description: Topic is the Schema for the topics API
            displayName: Topic
            kind: Topic
            name: topics.hazelcast.com
            version: v1alpha1
          - description: UserCodeNamespace is the Schema for the usercodenamespaces
              API
            displayName: User Code Namespace
            kind: UserCodeNamespace
            name: usercodenamespaces.hazelcast.com
            version: v1alpha1
          - description: VectorCollection is the Schema for the vectorcollections
              API
            displayName: Vector Collection
            kind: VectorCollection
            name: vectorcollections.hazelcast.com
            version: v1alpha1
          - description: WanReplication is the Schema for the wanreplications API
            displayName: Wan Replication
            kind: WanReplication
            name: wanreplications.hazelcast.com
            version: v1alpha1
          - description: WanSync is the Schema for the wansyncs API
            displayName: Wan Sync
            kind: WanSync
            name: wansyncs.hazelcast.com
            version: v1alpha1
        description: |
          Easily deploy Hazelcast clusters and Management Center into Kubernetes environments and manage their lifecycles.

          ## Before You Start

          If you are planning to create Hazelcast Platform Enterprise clusters, you need to [create a secret](https://docs.hazelcast.com/operator/latest/get-started#step-2-start-the-hazelcast-cluster) for the license. You can request a trial license key from [here](https://trialrequest.hazelcast.com).

          For Hazelcast Platform clusters, you can simply continue.

          ## Documentation

          1. [Get started](https://docs.hazelcast.com/operator/latest/get-started) with the Operator
          2. [Connect to the cluster from outside Kubernetes](https://guides.hazelcast.org/hazelcast-platform-operator-expose-externally)
          3. [Restore a Cluster from Cloud Storage with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-external-backup-restore)
          4. [Replicate Data between Two Hazelcast Clusters with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-wan-replication)

          ## Features

          Hazelcast Platform Operator supports the features below:

          * Custom resource for Hazelcast Platform (Enterprise) and Management Center
          * Observe status of Hazelcast clusters and Management Center instances
          * High Availability Mode configuration to create clusters that are resilient to node and zone failures
          * Support for TLS connections between members using self-signed certificates
          * Scale up and down Hazelcast clusters
          * Expose Hazelcast cluster to external clients ([Smart & Unisocket](https://docs.hazelcast.com/hazelcast/latest/clients/java#java-client-operation-modes))
          * Backup Hazelcast persistence data to cloud storage with the possibility of scheduling it and restoring the data accordingly
          * WAN Replication feature when you need to synchronize multiple Hazelcast clusters, which are connected by WANs
          * Full and Delta WAN Sync support
          * Tiered Storage support for Hazelcast and Map CRs
          * CP Subsystem configuration support for Hazelcast CR
          * User Code Deployment feature, which allows you to deploy custom and domain classes from cloud storages and URLs to Hazelcast members
          * The User Code Namespaces feature allows you to deploy custom and domain classes at runtime without a members restart
          * Support the configuration of advanced networking options
          * Support Multi-namespace configuration
          * ExecutorService and EntryProcessor support
          * Support several data structures like Map, Topic, MultiMap, ReplicatedMap, Queue and Cache which can be created dynamically via specific Custom Resources
          * MapStore, Near Cache and off-heap memory (HD memory and native memory) support for the Map CR
          * Native Memory support for the Cache CR
          * Support Jet configuration and Jet Job submission using the JetJob CR
          * Support for exporting the snapshots of JetJob CRs using JetJobSnapshot CR
          * Support for custom configurations using ConfigMap
        displayName: Hazelcast Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hazelcast
        - keyvalue
        - in-memory
        - database
        - caching
        links:
        - name: Documentation
          url: https://docs.hazelcast.com/operator/latest/
        - name: Source Code
          url: https://github.com/hazelcast/hazelcast-platform-operator
        maintainers:
        - email: cloudnative@hazelcast.com
          name: Hazelcast Cloud Native Team
        maturity: alpha
        minKubeVersion: 1.22.7
        provider:
          name: Hazelcast, Inc
        relatedImages:
        - docker.io/hazelcast/hazelcast-platform-operator@sha256:59328a4ea7a4e2110e0750364db792cf266f1e7231e3bb321e3de09d535dab23
        version: 5.15.0
      entries:
      - name: hazelcast-platform-operator.v5.15.0
        version: 5.15.0
      name: alpha
    defaultChannel: alpha
    packageName: hazelcast-platform-operator
    provider:
      name: Hazelcast, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: tang-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tang-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "daemons.redhat.com/v1alpha1",
                "kind": "TangServer",
                "metadata": {
                  "finalizers": [
                    "finalizer.daemons.tangserver.redhat.com"
                  ],
                  "name": "tangserver",
                  "namespace": "nbde"
                },
                "spec": {
                  "healthScript": "/usr/bin/tangd-health-check",
                  "image": "registry.redhat.io/rhel9/tang",
                  "keyPath": "/var/db/tang",
                  "persistentVolumeClaim": "tangserver-pvc",
                  "replicas": 1,
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:f43c96035f43040b3644e37119a3b35fc433f366d4b636459d1b9b02a05b5d3d
          createdAt: 2023-11-23T12:49:40
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TangServer is the Schema for the tangservers API
            displayName: Tang Server
            kind: TangServer
            name: tangservers.daemons.redhat.com
            version: v1alpha1
        description: NBDE Tang Server operator allows Tang Server deployment on OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nbde
        - tang
        links:
        - name: Tang Operator
          url: https://github.com/latchset/tang-operator
        maintainers:
        - email: sarroutb@redhat.com
          name: Sergio Arroutbi Braojos
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:97cade2c1ee468261aec5400728c8d44de387b459134aec7a4c3b5ec5a335d2c
        - registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:f43c96035f43040b3644e37119a3b35fc433f366d4b636459d1b9b02a05b5d3d
        version: 1.0.0
      entries:
      - name: tang-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: tang-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "daemons.redhat.com/v1alpha1",
                "kind": "TangServer",
                "metadata": {
                  "finalizers": [
                    "finalizer.daemons.tangserver.redhat.com"
                  ],
                  "name": "tangserver",
                  "namespace": "nbde"
                },
                "spec": {
                  "healthScript": "/usr/bin/tangd-health-check",
                  "image": "registry.redhat.io/rhel9/tang",
                  "keyPath": "/var/db/tang",
                  "persistentVolumeClaim": "tangserver-pvc",
                  "replicas": 1,
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:1cb2f794689817282ac570cacb106d75a2751a1926d89d3b83b619e02d869e5e
          createdAt: 2024-12-04T19:01:27
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TangServer is the Schema for the tangservers API
            displayName: Tang Server
            kind: TangServer
            name: tangservers.daemons.redhat.com
            version: v1alpha1
        description: NBDE Tang Server operator allows Tang Server deployment on OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nbde
        - tang
        links:
        - name: Tang Operator
          url: https://github.com/latchset/tang-operator
        maintainers:
        - email: sarroutb@redhat.com
          name: Sergio Arroutbi Braojos
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1172e150fff22c5eeab572f26961f3f53fbf896ee76d08c7503cfe2777c55458
        - registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:1cb2f794689817282ac570cacb106d75a2751a1926d89d3b83b619e02d869e5e
        version: 1.0.3
      entries:
      - name: tang-operator.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: alpha
    packageName: tang-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: www.redhat.com
    name: compliance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: compliance-operator.v0.1.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.17 <0.1.32'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          repository: https://github.com/openshift/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repo
          url: https://github.com/openshift/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/4.7/security/compliance_operator/compliance-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/compliance-rhel8-operator@sha256:ed68a0574ddeaf7d676403cd843f28cb586894f35dbb3ca5e06a5264bd42f030
        - registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c
        - registry.redhat.io/openshift4/compliance-openscap-rhel8@sha256:867c6d4205653d8a954d234cbaf470313009eb983d4f455fb45f49c4a7764422
        version: 0.1.32
      entries:
      - name: compliance-operator.v0.1.32
        version: 0.1.32
      name: "4.7"
    - currentCSV: compliance-operator.v0.1.61
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-node"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-platform"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "description": "Example of a tailoredProfile that extends OCP4 FedRAMP Moderate",
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.17 <0.1.61'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            displayName: Compliance Remediation
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            displayName: Compliance Scan
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            displayName: Compliance Suite
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            displayName: Profile Bundle
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            displayName: Scan Setting Binding
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            displayName: Tailored Profile
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repo
          url: https://github.com/openshift/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/compliance_operator/compliance-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b
        - registry.redhat.io/compliance/openshift-compliance-openscap-rhel8@sha256:83ad3113964ec48c5cf28315d923adb6edb95e45960e2756665d617dcd9b2e47
        - registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:5578e2225365ac1c15e4c189c5c7a8d14fd3fd26189414b36028f7f0879bb1dc
        version: 0.1.61
      entries:
      - name: compliance-operator.v0.1.61
        version: 0.1.61
      name: release-0.1
    - currentCSV: compliance-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-node"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-platform"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "description": "Example of a tailoredProfile that extends OCP4 FedRAMP Moderate",
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          must-gather-image: registry.redhat.io/compliance/openshift-compliance-must-gather-rhel8@sha256:e6468ccc55d72539b7fb980d0bce5c072e85879d4e9db39974d53eaab4d32b25
          olm.skipRange: '>=0.1.17 <1.7.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ComplianceAsCode/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            displayName: Compliance Remediation
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            displayName: Compliance Scan
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            displayName: Compliance Suite
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            displayName: Profile Bundle
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            displayName: Scan Setting Binding
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            displayName: Tailored Profile
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repository
          url: https://github.com/ComplianceAsCode/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/compliance_operator/co-overview.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4
        - registry.redhat.io/compliance/openshift-compliance-openscap-rhel8@sha256:d0f89d82845c6245a18e7db345ac3be92b2863d57f55f63e21652086128920c9
        - registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:8d3257bd6a020df53f8d7a400ae877ae029c78ffcd8ec9e35270ee3ff5c8e509
        - registry.redhat.io/compliance/openshift-compliance-must-gather-rhel8@sha256:e6468ccc55d72539b7fb980d0bce5c072e85879d4e9db39974d53eaab4d32b25
        version: 1.7.1
      entries:
      - name: compliance-operator.v1.7.1
        version: 1.7.1
      - name: compliance-operator.v1.7.0
        version: 1.7.0
      - name: compliance-operator.v1.6.2
        version: 1.6.2
      - name: compliance-operator.v1.6.1
        version: 1.6.1
      - name: compliance-operator.v1.6.0
        version: 1.6.0
      - name: compliance-operator.v1.5.1
        version: 1.5.1
      - name: compliance-operator.v1.5.0
        version: 1.5.0
      - name: compliance-operator.v1.4.1
        version: 1.4.1
      - name: compliance-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: compliance-operator
    provider:
      name: Red Hat Inc.
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: hyperfoil-bundle
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hyperfoil-operator.v0.26.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hyperfoil.io/v1alpha2",
                "kind": "Hyperfoil",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "hyperfoil-operator",
                    "app.kubernetes.io/instance": "hyperfoil-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hyperfoil",
                    "app.kubernetes.io/part-of": "hyperfoil-operator"
                  },
                  "name": "hyperfoil-sample"
                },
                "spec": {
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/hyperfoil/hyperfoil-operator:0.26.0
          createdAt: "2024-06-11T17:39:20Z"
          description: Microservice-oriented distributed benchmark framework.
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Hyperfoil/hyperfoil-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hyperfoil is the Schema for the hyperfoils API
            displayName: Hyperfoil
            kind: Hyperfoil
            name: hyperfoils.hyperfoil.io
            version: v1alpha2
        description: |
          Hyperfoil is a modern benchmark framework oriented promising
          accurate results with more flexibility and distributed architecture.

          See more on [hyperfoil.io](http://hyperfoil.io).

          Hyperfoil is licensed under Apache License 2.0
        displayName: Hyperfoil
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hyperfoil
        - benchmark
        - test
        - load
        - driver
        links:
        - name: Website
          url: http://hyperfoil.io
        - name: GitHub
          url: http://github.com/Hyperfoil
        maintainers:
        - email: rvansa@redhat.com
          name: Radim Vansa
        - email: spederse@redhat.com
          name: Ståle W. Pedersen
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/hyperfoil/hyperfoil-operator:0.26.0
        version: 0.26.0
      entries:
      - name: hyperfoil-operator.v0.26.0
        version: 0.26.0
      - name: hyperfoil-operator.v0.24.2
        version: 0.24.2
      - name: hyperfoil-operator.v0.21.0
        version: 0.21.0
      - name: hyperfoil-operator.v0.15.1
        version: 0.15.1
      - name: hyperfoil-operator.v0.15.0
        version: 0.15.0
      name: alpha
    defaultChannel: alpha
    packageName: hyperfoil-bundle
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Antline-Soft
      provider-url: ""
    name: cxl-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cxl-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-05-08T08:36:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
        description: memory extend operator
        displayName: extend-community-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - memory
        - extend
        - community
        - operator
        links:
        - name: Cxl Operator
          url: https://cxl-operator.domain
        maintainers:
        - email: tomboy@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        provider:
          name: Antline-Soft
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - rasen708/cxl-operator:latest
        version: 0.0.1
      entries:
      - name: cxl-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cxl-operator
    provider:
      name: Antline-Soft
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-msg-broker
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-msg-broker.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Messagebroker",
                "metadata": {
                  "name": "messagebroker-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "chiselServer": {
                    "enabled": false,
                    "port": 8080
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": false
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "natsCore": {
                    "reloader": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats-server-config-reloader",
                      "tag": "0.8.3-3747-30"
                    },
                    "server": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats",
                      "tag": "2.9.0-3747-21"
                    }
                  },
                  "natsHttpProxy": {
                    "enabled": false,
                    "port": 8080
                  },
                  "natsStreaming": {
                    "image": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats-streaming",
                      "tag": "0.23.2-3747-31"
                    }
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "tagSuffix": "-ocp"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:31:32Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Messagebroker
            name: messagebrokers.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Message Broker Operator
        displayName: Keysight WAP Message Broker Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Msg Broker
          url: https://keysight-wap-msg-broker.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-msg-broker@sha256:4701b6e456b099697bf1e25aeb4bc8e32b1892bb0956e36d84f9f80666b3a1fc
        version: 0.1.21
      entries:
      - name: keysight-wap-msg-broker.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-msg-broker
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC
      provider-url: https://www.nec.com
    name: sanstoragecsi-operator-bundle
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sanstoragecsi-operator-bundle.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.nec.com/v1",
                "kind": "SANStorageCSI",
                "metadata": {
                  "name": "sanstoragecsi"
                },
                "spec": {
                  "driverName": "san.csi.nec.com",
                  "images": {
                    "csiDriver": "",
                    "externalAttacher": "registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:bd8262cc4b6c405e175c39623323dd98a74bbf70e6aa90851a72612056c184ff",
                    "externalProvisioner": "registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:85f1cc8f21568e7885e492b29660f29e09796fb23ebdf5fd526aa2e99d6b2fa8",
                    "livenessProbe": "registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:1a0d15012ba70955af754b03a7488b7c6ace813646ec7c3ce845f80597050c3c",
                    "nodeRegistrar": "registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:dc7f0992be6c2853dc0e798afda368e36005a94ad76b21fa2f09790621318864"
                  },
                  "parameters": {
                    "controller": {
                      "logLevel": 5
                    },
                    "node": {
                      "livenessProbePort": 9808,
                      "logLevel": 5,
                      "maxVolumesPerNode": 1024
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          createdAt: "2022-06-15T00:00:00Z"
          description: An operator for managing the NEC Storage M Series CSI Driver
          operatorframework.io/suggested-namespace: nec-san-csi-operator-system
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: NEC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SANStorageCSI is the Schema for the sanstoragecsis API
            displayName: NEC Storage M Series CSI Driver
            kind: SANStorageCSI
            name: sanstoragecsis.csi.nec.com
            version: v1
        description: |
          This operator deploys the NEC Storage M Series CSI Driver on OpenShift Container Platform.

          NEC Storage M Series CSI Driver provides persistent storage for stateful applications using NEC Storage M Series.
          Please refer to the documentation of the NEC Storage M Series CSI Driver for detail information about supported NEC Storage Platforms and supported CSI features.
        displayName: NEC Storage M Series CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - NEC
        - M Series
        - SAN Storage
        - CSI Driver
        links:
        - name: NEC Storage M Series
          url: https://www.nec.com/en/global/prod/storage/product/san/index.html
        maturity: stable
        provider:
          name: NEC
          url: https://www.nec.com
        relatedImages:
        - registry.connect.redhat.com/nec/sanstoragecsi-operator@sha256:6656c2b5036bdc013712ecb640bbbebe0e3e90cf625c73697dbb49e3e201e2ac
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:dc7f0992be6c2853dc0e798afda368e36005a94ad76b21fa2f09790621318864
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:bd8262cc4b6c405e175c39623323dd98a74bbf70e6aa90851a72612056c184ff
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:85f1cc8f21568e7885e492b29660f29e09796fb23ebdf5fd526aa2e99d6b2fa8
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:1a0d15012ba70955af754b03a7488b7c6ace813646ec7c3ce845f80597050c3c
        version: 1.1.0
      entries:
      - name: sanstoragecsi-operator-bundle.v1.1.0
        version: 1.1.0
      - name: sanstoragecsi-operator-bundle.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: sanstoragecsi-operator-bundle
    provider:
      name: NEC
      url: https://www.nec.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nxiq-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nxiq-operator-certified.v1.193.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusIQ",
                "metadata": {
                  "name": "example-nexusiq"
                },
                "spec": {
                  "deployment": {
                    "postStart": {
                      "command": null
                    }
                  },
                  "deploymentStrategy": {
                    "type": "Recreate"
                  },
                  "fullnameOverride": "",
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [],
                    "path": "/",
                    "tls": []
                  },
                  "iq": {
                    "adminPort": 8071,
                    "applicationPort": 8070,
                    "env": [
                      {
                        "name": "JAVA_OPTS",
                        "value": "-Djava.util.prefs.userRoot=$(SONATYPE_WORK)/javaprefs -Djavax.net.ssl.trustStore=/etc/nexus-iq-server/jks/keystore.jks"
                      }
                    ],
                    "configYaml": {
                      "sonatypeWork": "/sonatype-work",
                      "createSampleData": true,
                      "server": {
                        "adminConnectors": [
                          {
                            "port": 8071,
                            "type": "http"
                          }
                        ],
                        "applicationConnectors": [
                          {
                            "port": 8070,
                            "type": "http"
                          }
                        ],
                        "requestLog": {
                          "appenders": [
                            {
                              "type": "file",
                              "currentLogFilename": "/var/log/nexus-iq-server/request.log",
                              "logFormat": "%clientHost %l %user [%date] \"%requestURL\" %statusCode %bytesSent %elapsedTime \"%header{User-Agent}\"",
                              "archivedLogFilenamePattern": "/var/log/nexus-iq-server/request-%d.log.gz",
                              "archivedFileCount": 50
                            }
                          ]
                        }
                      },
                      "logging": {
                        "level": "DEBUG",
                        "loggers": {
                          "com.sonatype.insight.scan": "INFO",
                          "eu.medsea.mimeutil.MimeUtil2": "INFO",
                          "org.apache.http": "INFO",
                          "org.apache.http.wire": "ERROR",
                          "org.eclipse.birt.report.engine.layout.pdf.font.FontConfigReader": "WARN",
                          "org.eclipse.jetty": "INFO",
                          "org.apache.shiro.web.filter.authc.BasicHttpAuthenticationFilter": "INFO",
                          "com.sonatype.insight.audit": {
                            "appenders": [
                              {
                                "type": "file",
                                "currentLogFilename": "/var/log/nexus-iq-server/audit.log",
                                "archivedLogFilenamePattern": "/var/log/nexus-iq-server/audit-%d.log.gz",
                                "archivedFileCount": 50
                              }
                            ]
                          },
                          "com.sonatype.insight.policy.violation": {
                            "appenders": [
                              {
                                "type": "console",
                                "threshold": "INFO"
                              }
                            ]
                          }
                        },
                        "appenders": [
                          {
                            "type": "console",
                            "threshold": "INFO",
                            "logFormat": "%d{'yyyy-MM-dd HH:mm:ss,SSSZ'} %level [%thread] %X{username} %logger - %msg%n"
                          },
                          {
                            "type": "file",
                            "threshold": "ALL",
                            "logFormat": "%d{'yyyy-MM-dd HH:mm:ss,SSSZ'} %level [%thread] %X{username} %logger - %msg%n",
                            "currentLogFilename": "/var/log/nexus-iq-server/clm-server.log",
                            "archivedLogFilenamePattern": "/var/log/nexus-iq-server/clm-server-%d.log.gz",
                            "archivedFileCount": 50
                          }
                        ]
                      }
                    },
                    "imageName": "registry.connect.redhat.com/sonatype/nexus-iq-server@sha256:d65df6bdff1f32bf32a87e5cb38f5e9f98b75b81e9acccbffbd967e1f71691f0",
                    "imagePullPolicy": "IfNotPresent",
                    "imagePullSecret": "",
                    "licenseSecret": "",
                    "memory": "1Gi",
                    "name": "nxiq",
                    "certificates": {}
                  },
                  "nameOverride": "",
                  "persistence": {
                    "accessMode": "ReadWriteOnce",
                    "storageSize": "1Gi"
                  },
                  "service": {
                    "annotations": {},
                    "enabled": false,
                    "labels": {},
                    "ports": [
                      {
                        "name": "nexus-service",
                        "port": 80,
                        "targetPort": 80
                      }
                    ]
                  },
                  "serviceAccount": {
                    "create": true,
                    "name": null
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nxiq-operator-certified@sha256:59c52729189da87d9b5c3de8ef85326662d94c99835ab99d6e8d7fefdcf53ef0
          createdAt: "2020-07-20"
          description: |-
            Nexus Lifecycle is an open source governance platform that continuously identifies risk,
            automatically enforces policy, and provides visibility throughout the entire SDLC.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.193.0-1
          repository: https://github.com/sonatype/operator-iq
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus IQ Server
            displayName: NexusIQ
            kind: NexusIQ
            name: nexusiqs.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Lifecycle is an open source governance platform that continuously identifies risk,
          automatically enforces policy, and provides visibility throughout the entire SDLC.
          Developers leveraging open source require a solution that helps them make safer choices
          while still delivering at DevOps speed.
          Nexus Lifecycle empowers developers and security teams to choose the best components
          and govern open source usage, ensuring organizations continue to innovate with less risk.

          ## Core Capabilities

          * **Dependency Management**:
            Generate a precise software bill of materials for all applications
            to identify open source components and gather contextual direct
            and transitive dependency information for faster action.
          * **One-Click Remediation**:
            Upgrade to the best components and versions based on real-time intelligence.
            Available in your preferred IDE or a GitHub pull request.
          * **Automated Open Source Policy Enforcement**:
            Create custom security, license, and architectural policies based
            on application type or organization and contextually enforce those policies
            across every stage of the SDLC.
          * **Continuous Monitoring**:
            Real-time notifications of newly discovered defects,
            as well as vulnerabilities based on component,
            risk level or application/container affected.

          ## Extras

          * IDE Integrations: Eclipse, IntelliJ, and Visual Studio
          * Source Control Integrations: GitHub, GitLab, and Bitbucket
          * Chrome Browser Extension
          * Red Hat Clair for Container Scanning

          ## Limitations

          This operator will be released on a quarterly basis.

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus IQ Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage

          Once the server instance is created by the operator and running,
          you'll want to expose the service as you see fit:
          1. Create a Route to the new service for iq.applicationPort (8070).
          2. Visit the URL provided by the Route, login, and set new credentials.
            The default credentials are `admin`/`admin123`.

          The Nexus IQ Server can be further configured via the NexusIQ custom resource definition:

          | Parameter            | Description                                                  | Default           |
          | -------------------- | ------------------------------------------------------------ | ----------------- |
          | `iq.applicationPort` | Port of the application connector. Must match the value in the `configYaml` property | `8070`            |
          | `iq.adminPort`       | Port of the application connector. Must match the value in the `configYaml` property | `8071`            |
          | `iq.memory`          | The amount of RAM to allocate                                | `1Gi`             |
          | `iq.env`             | Customize the server environment, including `JAVA_OPTS`      | See example YAML show when ceating a NexusIQ. |
          | `iq.licenseSecret`   | The base-64 encoded license file to be installed at startup  | `""`              |
          | `iq.configYaml`      | A YAML block which will be used as a configuration block for IQ Server. | See example YAML shown when creating a NexusIQ. |
          | `iq.certificates`   | Add your own certificates for your own trusted Certificate Authorities with names and plain armored text. | `{}` |
          | `ingress.enabled`                           | Create an ingress for Nexus         | `true`                                  |
          | `ingress.annotations`                       | Annotations to enhance ingress configuration  | `{}`                          |
          | `ingress.tls.enabled`                       | Enable TLS                          | `true`                                 |
          | `ingress.tls.secretName`                    | Name of the secret storing TLS cert, `false` to use the Ingress' default certificate | `nexus-tls`                             |
          | `ingress.path`                              | Path for ingress rules. GCP users should set to `/*` | `/`                    |
          | `deploymentStrategy.type`                   | Deployment Strategy                 |  `Recreate` |
          | `initContainers`                            | Containers to run before starting the server | `nil` |
          | `deployment.postStart.command`              | Command to run after starting the IQ Server container  | `nil`                    |
          | `deployment.terminationGracePeriodSeconds` | Time to allow for clean shutdown                        | 120                      |
          | `persistence.storageClass` | The provisioner class                        | `-` (disables dynamic provisioning)            |
          | `persistence.storageSize` | The amount of drive space to allocate                        | `1Gi`             |
          | `persistence.accessMode` | Default access mode                        | `ReadWriteOnce`             |
          | `persistence.volumeConfiguration` | A YAML block to configure the persistent volume type. Defaults to `hostPath` which should not be used in production | `hostPath`             |

          ## Configuring IQ Server

          You can define the `config.yml` for IQ Server in your CRD on startup.
          It is the `iq.configYaml` property. For more details, see the [Configuring IQ Server](https://help.sonatype.com/iqserver/configuring) help page.

          ## Installing the License

          The license file can be installed via the UI when IQ server is running, or it can be done as a part of the deploy.
          If you leave the `licenseFile` field empty/commented, IQ Server will start and prompt you to manually install the license
          when you first enter the GUI.

          ### Installing the License Automatically
          To do it automatically, first encode your `.lic` file in Base 64 with no line breaks, eg:

          ```bash
          base64 --wrap=0 mylicense.lic > lic.base64
          ```

          Then add this value to your CRD file as `iq.licenseSecret`, eg:

          ```yaml
          iq:
            licenseSecret: bXkgc2FtcGxlIGxpY2Vuc2U=
          ```

          Specify the `licenseFile` path in `iq.configYaml` as:

          ```yaml
          iq:
            configYaml:
              sonatypeWork: /sonatype-work
              createSampleData: true
              # add this line and the `licenseSecret` above to autoconfigure licensing
              licenseFile: /etc/nexus-iq-license/license_lic
              server:
                applicationConnectors:
                  - type: http
                    port: 8070
                adminConnectors:
                  - type: http
                    port: 8071
          ```

          ## Adding Certificates for Your Own Trusted Certificate Authorities

          Specify a set of certificates to be imported into the Java trust store
          of the running IQ Server. More than one certificate can be provided.

          ```yaml
          iq:
            certificates:
              my-company-cert: |
                -----BEGIN CERTIFICATE-----
                ...
                -----END CERTIFICATE-----
          ```

          ## Running Commands Before Initialization of the Server

          If you need custom initialization, you can use standard Kubernetes
          [`initContainers`](https://kubernetes.io/docs/concepts/workloads/pods/init-containers/)
          to run any series of containers and commands you need.

          ```yaml
          initContainers:
            - name: an-init-container
              image: busybox:1.28
              command: ['/bin/sh', '-c', 'echo "initializing"']
          ```
        displayName: Nexus IQ Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - sonatype
        - licensing
        - policy
        - components
        links:
        - name: Nexus Lifecycle
          url: https://www.sonatype.com/product-nexus-lifecycle
        maintainers:
        - email: support@sonaype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nxiq-operator-certified@sha256:59c52729189da87d9b5c3de8ef85326662d94c99835ab99d6e8d7fefdcf53ef0
        - registry.connect.redhat.com/sonatype/nexus-iq-server@sha256:d65df6bdff1f32bf32a87e5cb38f5e9f98b75b81e9acccbffbd967e1f71691f0
        version: 1.193.0-1
      entries:
      - name: nxiq-operator-certified.v1.193.0-1
        version: 1.193.0-1
      name: stable
    defaultChannel: stable
    packageName: nxiq-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com/
    name: openstack-baremetal-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openstack-baremetal-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "baremetal.openstack.org/v1beta1",
                "kind": "OpenStackBaremetalSet",
                "metadata": {
                  "name": "openstackbaremetalset"
                },
                "spec": {
                  "baremetalHosts": {
                    "compute-0": "172.22.0.100/24",
                    "compute-1": "172.22.0.101/24"
                  },
                  "ctlplaneGateway": "172.22.0.3",
                  "ctlplaneInterface": "enp1s0",
                  "deploymentSSHSecret": "baremetalset-deploy-secret",
                  "domainName": "osptest.openstack.org",
                  "osImage": "edpm-hardened-uefi.qcow2",
                  "passwordSecret": "baremetalset-password-secret"
                }
              },
              {
                "apiVersion": "baremetal.openstack.org/v1beta1",
                "kind": "OpenStackProvisionServer",
                "metadata": {
                  "name": "openstackprovisionserver"
                },
                "spec": {
                  "interface": "enp1s0",
                  "osImage": "edpm-hardened-uefi.qcow2",
                  "port": 8080
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:07Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenStackBaremetalSet is the Schema for the openstackbaremetalsets
              API
            displayName: OpenStack BaremetalSet
            kind: OpenStackBaremetalSet
            name: openstackbaremetalsets.baremetal.openstack.org
            version: v1beta1
          - description: OpenStackProvisionServer used to serve custom images for
              baremetal provisioning with Metal3
            displayName: OpenStackProvisionServer
            kind: OpenStackProvisionServer
            name: openstackprovisionservers.baremetal.openstack.org
            version: v1beta1
        description: An operator for handling baremetal provisioning concerns for
          OpenStack
        displayName: OpenStack Baremetal Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cn-openstack
        - openstack
        links:
        - name: Openstack Baremetal Operator
          url: https://github.com/openstack-k8s-operators/openstack-baremetal-operator
        maturity: beta
        provider:
          name: Red Hat
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/ubi9/httpd-24@sha256:dbfbe3383e12d63df7178e4cdd5b1e6e51a84c781041b9192026ea6edb72fbf3
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/openstack-baremetal-rhel9-operator@sha256:a6c69c03e59faf50973d07ccea32b80cb9c46ceb60ed00ffe7311ca42d501296
        - registry.redhat.io/rhoso/edpm-hardened-uefi-rhel9@sha256:e4bfacb9d8de056a251bee16a66b69631b96fb19fb926d8e33acd379d40a92d6
        - registry.redhat.io/rhoso-operators/openstack-baremetal-agent-rhel9@sha256:8f6f8e99042710819aa639eae6513f36c4642190ecd85d58a2ed819f1b55e38f
        version: 1.0.6
      entries:
      - name: openstack-baremetal-operator.v1.0.6
        version: 1.0.6
      - name: openstack-baremetal-operator.v1.0.5
        version: 1.0.5
      - name: openstack-baremetal-operator.v1.0.4
        version: 1.0.4
      - name: openstack-baremetal-operator.v1.0.3
        version: 1.0.3
      - name: openstack-baremetal-operator.v1.0.2
        version: 1.0.2
      - name: openstack-baremetal-operator.v1.0.1
        version: 1.0.1
      - name: openstack-baremetal-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: openstack-baremetal-operator
    provider:
      name: Red Hat
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecr-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecr-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.services.k8s.aws/v1alpha1",
                "kind": "Repository",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecr-controller:1.3.1
          createdAt: "2025-09-20T01:21:24Z"
          description: AWS ECR controller is a service controller for managing ECR
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PullThroughCacheRule represents the state of an AWS ecr PullThroughCacheRule
              resource.
            displayName: PullThroughCacheRule
            kind: PullThroughCacheRule
            name: pullthroughcacherules.ecr.services.k8s.aws
            version: v1alpha1
          - description: Repository represents the state of an AWS ecr Repository
              resource.
            displayName: Repository
            kind: Repository
            name: repositories.ecr.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Container Registry (ECR) resources in AWS from within your Kubernetes cluster.

          **About Amazon ECR**

          Amazon Elastic Container Registry (Amazon ECR) is an AWS managed container image registry service that is secure, scalable, and reliable. Amazon ECR supports private repositories with resource-based permissions using AWS IAM. This is so that specified users or Amazon EC2 instances can access your container repositories and images. You can use your preferred CLI to push, pull, and manage Docker images, Open Container Initiative (OCI) images, and OCI compatible artifacts.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECR
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECR Developer Resources
          url: https://aws.amazon.com/ecr/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecr maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecr-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-ecr-controller.v1.3.1
        version: 1.3.1
      - name: ack-ecr-controller.v1.3.0
        version: 1.3.0
      - name: ack-ecr-controller.v1.2.0
        version: 1.2.0
      - name: ack-ecr-controller.v1.1.0
        version: 1.1.0
      - name: ack-ecr-controller.v1.0.33
        version: 1.0.33
      - name: ack-ecr-controller.v1.0.32
        version: 1.0.32
      - name: ack-ecr-controller.v1.0.31
        version: 1.0.31
      - name: ack-ecr-controller.v1.0.30
        version: 1.0.30
      - name: ack-ecr-controller.v1.0.29
        version: 1.0.29
      - name: ack-ecr-controller.v1.0.28
        version: 1.0.28
      - name: ack-ecr-controller.v1.0.27
        version: 1.0.27
      - name: ack-ecr-controller.v1.0.25
        version: 1.0.25
      - name: ack-ecr-controller.v1.0.24
        version: 1.0.24
      - name: ack-ecr-controller.v1.0.23
        version: 1.0.23
      - name: ack-ecr-controller.v1.0.22
        version: 1.0.22
      - name: ack-ecr-controller.v1.0.21
        version: 1.0.21
      - name: ack-ecr-controller.v1.0.20
        version: 1.0.20
      - name: ack-ecr-controller.v1.0.19
        version: 1.0.19
      - name: ack-ecr-controller.v1.0.17
        version: 1.0.17
      - name: ack-ecr-controller.v1.0.16
        version: 1.0.16
      - name: ack-ecr-controller.v1.0.15
        version: 1.0.15
      - name: ack-ecr-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecr-controller.v1.0.12
        version: 1.0.12
      - name: ack-ecr-controller.v1.0.10
        version: 1.0.10
      - name: ack-ecr-controller.v1.0.9
        version: 1.0.9
      - name: ack-ecr-controller.v1.0.8
        version: 1.0.8
      - name: ack-ecr-controller.v1.0.7
        version: 1.0.7
      - name: ack-ecr-controller.v1.0.6
        version: 1.0.6
      - name: ack-ecr-controller.v1.0.4
        version: 1.0.4
      - name: ack-ecr-controller.v1.0.3
        version: 1.0.3
      - name: ack-ecr-controller.v1.0.2
        version: 1.0.2
      - name: ack-ecr-controller.v1.0.0
        version: 1.0.0
      - name: ack-ecr-controller.v0.1.9
        version: 0.1.9
      - name: ack-ecr-controller.v0.1.7
        version: 0.1.7
      - name: ack-ecr-controller.v0.1.6
        version: 0.1.6
      - name: ack-ecr-controller.v0.1.5
        version: 0.1.5
      - name: ack-ecr-controller.v0.1.4
        version: 0.1.4
      - name: ack-ecr-controller.v0.1.3
        version: 0.1.3
      - name: ack-ecr-controller.v0.1.2
        version: 0.1.2
      - name: ack-ecr-controller.v0.1.1
        version: 0.1.1
      - name: ack-ecr-controller.v0.1.0
        version: 0.1.0
      - name: ack-ecr-controller.v0.0.27
        version: 0.0.27
      - name: ack-ecr-controller.v0.0.26
        version: 0.0.26
      - name: ack-ecr-controller.v0.0.25
        version: 0.0.25
      - name: ack-ecr-controller.v0.0.24
        version: 0.0.24
      - name: ack-ecr-controller.v0.0.23
        version: 0.0.23
      - name: ack-ecr-controller.v0.0.22
        version: 0.0.22
      - name: ack-ecr-controller.v0.0.21
        version: 0.0.21
      - name: ack-ecr-controller.v0.0.20
        version: 0.0.20
      - name: ack-ecr-controller.v0.0.19
        version: 0.0.19
      - name: ack-ecr-controller.v0.0.18
        version: 0.0.18
      - name: ack-ecr-controller.v0.0.17
        version: 0.0.17
      - name: ack-ecr-controller.v0.0.16
        version: 0.0.16
      - name: ack-ecr-controller.v0.0.15
        version: 0.0.15
      - name: ack-ecr-controller.v0.0.14
        version: 0.0.14
      - name: ack-ecr-controller.v0.0.13
        version: 0.0.13
      - name: ack-ecr-controller.v0.0.12
        version: 0.0.12
      - name: ack-ecr-controller.v0.0.11
        version: 0.0.11
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecr-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: VictoriaMetrics
      provider-url: ""
    name: victoriametrics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: victoriametrics-operator.v0.63.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLAgent",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "persistentVolumeClaimRetentionPolicy": {
                    "whenDeleted": "Delete"
                  },
                  "remoteWrite": [
                    {
                      "maxDiskUsage": "10GB",
                      "url": "http://vlsingle-example-0.default.svc:9428/internal/insert"
                    }
                  ],
                  "remoteWriteSettings": {
                    "maxBlockSize": "30MB"
                  },
                  "replicaCount": 2,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "350Mi"
                    }
                  },
                  "storage": {
                    "volumeClaimTemplate": {
                      "spec": {
                        "resources": {
                          "requests": {
                            "storage": "50Gi"
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "victoriametrics-operator"
                  },
                  "name": "example"
                },
                "spec": {
                  "requestsLoadBalancer": {
                    "enabled": true,
                    "spec": {}
                  },
                  "vlinsert": {
                    "replicaCount": 1
                  },
                  "vlselect": {
                    "replicaCount": 1
                  },
                  "vlstorage": {
                    "replicaCount": 2,
                    "retentionPeriod": "1y",
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "50Gi"
                            }
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLSingle",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "victoriametrics-operator"
                  },
                  "name": "sample"
                },
                "spec": {
                  "retentionMaxDiskSpaceUsageBytes": "20GB",
                  "retentionPeriod": "1y",
                  "storage": {
                    "resources": {
                      "requests": {
                        "storage": "50Gi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VLogs",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "retentionPeriod": "2"
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAgent",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "extraArgs": {
                    "memory.allowedPercent": "40"
                  },
                  "remoteWrite": [
                    {
                      "maxDiskUsage": "10GB",
                      "url": "http://vmsingle-dc-1.default.svc:8429/api/v1/write"
                    },
                    {
                      "maxDiskUsage": "15GB",
                      "url": "http://vmsingle-dc-2.default.svc:8429/api/v1/write"
                    },
                    {
                      "url": "http://vminsert-main.default.svc:8480/insert/1:1/prometheus/api/v1/write"
                    }
                  ],
                  "remoteWriteSettings": {
                    "maxDiskUsagePerURL": "5GB"
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "350Mi"
                    }
                  },
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlert",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "datasource": {
                    "url": "http://vmsingle-example-pvc.default.svc:8429"
                  },
                  "evaluationInterval": "30s",
                  "notifier": {
                    "url": "http://vmalertmanager-example.default.svc:9093"
                  },
                  "replicaCount": 1,
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlertmanager",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "configSecret": "am-global-config",
                  "replicaCount": 1,
                  "selectAllByDefault": true,
                  "storage": {
                    "volumeClaimTemplate": {
                      "spec": {
                        "resources": {
                          "requests": {
                            "storage": "2Gi"
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlertmanagerConfig",
                "metadata": {
                  "name": "example-email-web"
                },
                "spec": {
                  "receivers": [
                    {
                      "email_configs": [
                        {
                          "from": "alerting@example.com",
                          "smarthost": "example.com:25",
                          "text": "ALARM",
                          "to": "some-email@example.com"
                        }
                      ],
                      "name": "email",
                      "webhook_configs": [
                        {
                          "url": "http://some-other-wh"
                        }
                      ]
                    }
                  ],
                  "route": {
                    "group_interval": "1m",
                    "receiver": "email",
                    "routes": [
                      {
                        "receiver": "email"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAuth",
                "metadata": {
                  "name": "vmauth-sample"
                },
                "spec": {
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "victoriametrics/vmauth"
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "250m",
                      "memory": "350Mi"
                    }
                  },
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMCluster",
                "metadata": {
                  "name": "example-persistent"
                },
                "spec": {
                  "replicationFactor": 2,
                  "retentionPeriod": "4",
                  "vminsert": {
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "0.5",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "vmselect": {
                    "cacheMountPath": "/select-cache",
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "0.5",
                        "memory": "500Mi"
                      }
                    },
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "2Gi"
                            }
                          }
                        }
                      }
                    }
                  },
                  "vmstorage": {
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1500Mi"
                      }
                    },
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "10Gi"
                            }
                          }
                        }
                      }
                    },
                    "storageDataPath": "/vm-data"
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMNodeScrape",
                "metadata": {
                  "name": "cadvisor-metrics"
                },
                "spec": {
                  "bearerTokenFile": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                  "interval": "10s",
                  "relabelConfigs": [
                    {
                      "action": "labelmap",
                      "regex": "__meta_kubernetes_node_label_(.+)"
                    },
                    {
                      "replacement": "kubernetes.default.svc:443",
                      "targetLabel": "__address__"
                    },
                    {
                      "regex": "(.+)",
                      "replacement": "/api/v1/nodes/$1/proxy/metrics/cadvisor",
                      "sourceLabels": [
                        "__meta_kubernetes_node_name"
                      ],
                      "targetLabel": "__metrics_path__"
                    }
                  ],
                  "scheme": "https",
                  "scrapeTimeout": "2s",
                  "selector": {},
                  "tlsConfig": {
                    "caFile": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                    "insecureSkipVerify": true
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMPodScrape",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "podMetricsEndpoints": [
                    {
                      "port": "metrics",
                      "scheme": "http"
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app.kubernetes.io/name": "example-application",
                      "monitored-by": "vm-operator"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMProbe",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "interval": "2s",
                  "jobName": "static-probe",
                  "module": "http_2xx",
                  "targets": {
                    "staticConfig": {
                      "targets": [
                        "vmagent-example.default.svc:8429/health"
                      ]
                    }
                  },
                  "vmProberSpec": {
                    "url": "prometheus-blackbox-exporter.default.svc:9115"
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMRule",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "groups": [
                    {
                      "name": "kafka",
                      "rules": [
                        {
                          "alert": "coordinator down",
                          "annotations": {
                            "description": "kafka coordinator is down",
                            "value": "{{ $value }}"
                          },
                          "expr": "ml_app_gauge{exec_context=\"consumer_group_state\"} == 0",
                          "for": "60s",
                          "labels": {
                            "job": "{{ $labels.job }}",
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMScrapeConfig",
                "metadata": {
                  "name": "mongodb"
                },
                "spec": {
                  "consulSDConfigs": [
                    {
                      "server": "https://consul-dns:8500",
                      "services": [
                        "mongodb"
                      ]
                    }
                  ],
                  "relabelConfigs": [
                    {
                      "action": "replace",
                      "sourceLabels": [
                        "__meta_consul_service"
                      ],
                      "targetLabel": "job"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMServiceScrape",
                "metadata": {
                  "name": "example-scrape"
                },
                "spec": {
                  "endpoints": [
                    {
                      "port": "http"
                    }
                  ],
                  "namespaceSelector": {},
                  "selector": {
                    "matchLabels": {
                      "app.kubernetes.io/name": "example-application",
                      "monitored-by": "vm-operator"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMSingle",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "retentionPeriod": "1"
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMSingle",
                "metadata": {
                  "name": "example-pvc"
                },
                "spec": {
                  "removePvcAfterDelete": true,
                  "retentionPeriod": "1",
                  "storage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMStaticScrape",
                "metadata": {
                  "name": "vmstaticscrape-sample"
                },
                "spec": {
                  "jobName": "static",
                  "targetEndpoints": [
                    {
                      "labels": {
                        "env": "dev",
                        "project": "operator"
                      },
                      "targets": [
                        "192.168.0.1:9100",
                        "196.168.0.50:9100"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMUser",
                "metadata": {
                  "name": "vmuser-tenant-1"
                },
                "spec": {
                  "bearerToken": "some-token",
                  "targetRefs": [
                    {
                      "crd": {
                        "kind": "VMCluster/vminsert",
                        "name": "test-persistent",
                        "namespace": "vm"
                      },
                      "target_path_suffix": "/insert/1"
                    },
                    {
                      "crd": {
                        "kind": "VMCluster/vmselect",
                        "name": "test-persistent",
                        "namespace": "vm"
                      },
                      "target_path_suffix": "/select/1"
                    },
                    {
                      "paths": [
                        "/internal/resetRollupResultCache"
                      ],
                      "static": {
                        "url": "http://vmselect-test-persistent.default.svc:8481/"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: docker.io/victoriametrics/operator:v0.63.0
          createdAt: "2025-09-12T12:58:29Z"
          description: Provides monitoring capabilities for kubernetes clusters and
            applications
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: "1.30"
          operators.openshift.io/infrastructure-features: '[fips]'
          operators.operatorframework.io.bundle.channel.default.v1: beta
          operators.operatorframework.io.bundle.channels.v1: beta
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/VictoriaMetrics/operator
          support: VictoriaMetrics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VLAgent - is a tiny but brave agent, which helps you collect
              logs from various sources and stores them in VictoriaLogs.
            displayName: VLAgent
            kind: VLAgent
            name: vlagents.operator.victoriametrics.com
            version: v1
          - description: VLCluster is fast, cost-effective and scalable logs database.
            displayName: VLCluster
            kind: VLCluster
            name: vlclusters.operator.victoriametrics.com
            version: v1
          - description: |-
              VLogs is fast, cost-effective and scalable logs database.
              VLogs is the Schema for the vlogs API
            displayName: VLogs
            kind: VLogs
            name: vlogs.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VLSingle is fast, cost-effective and scalable logs database.
              VLSingle is the Schema for the API
            displayName: VLSingle
            kind: VLSingle
            name: vlsingles.operator.victoriametrics.com
            version: v1
          - description: |-
              VMAgent - is a tiny but brave agent, which helps you collect metrics from various sources and stores them in VictoriaMetrics
              or any other Prometheus-compatible storage system that supports the remote_write protocol.
            displayName: VMAgent
            kind: VMAgent
            name: vmagents.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlertmanagerConfig is the Schema for the vmalertmanagerconfigs
              API
            displayName: VMAlertmanager Config
            kind: VMAlertmanagerConfig
            name: vmalertmanagerconfigs.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlertmanager represents Victoria-Metrics deployment for
              Alertmanager.
            displayName: VMAlertmanager
            kind: VMAlertmanager
            name: vmalertmanagers.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlert  executes a list of given alerting or recording rules
              against configured address.
            displayName: VMAlert
            kind: VMAlert
            name: vmalerts.operator.victoriametrics.com
            version: v1beta1
          - description: VMAnomaly is the Schema for the vmanomalies API.
            displayName: VMAnomaly
            kind: VMAnomaly
            name: vmanomalies.operator.victoriametrics.com
            version: v1
          - description: VMAuth is the Schema for the vmauths API
            displayName: VMAuth
            kind: VMAuth
            name: vmauths.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMCluster is fast, cost-effective and scalable time-series database.
              Cluster version with
            displayName: VMCluster
            kind: VMCluster
            name: vmclusters.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMNodeScrape defines discovery for targets placed on kubernetes nodes,
              usually its node-exporters and other host services.
              InternalIP is used as __address__ for scraping.
            displayName: VMNode Scrape
            kind: VMNodeScrape
            name: vmnodescrapes.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMPodScrape is scrape configuration for pods,
              it generates vmagent's config for scraping pod targets
              based on selectors.
            displayName: VMPod Scrape
            kind: VMPodScrape
            name: vmpodscrapes.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMProbe defines a probe for targets, that will be executed with prober,
              like blackbox exporter.
              It helps to monitor reachability of target with various checks.
            displayName: VMProbe
            kind: VMProbe
            name: vmprobes.operator.victoriametrics.com
            version: v1beta1
          - description: VMRule defines rule records for vmalert application
            displayName: VMRule
            kind: VMRule
            name: vmrules.operator.victoriametrics.com
            version: v1beta1
          - description: VMScrapeConfig specifies a set of targets and parameters
              describing how to scrape them.
            displayName: VMScrape Config
            kind: VMScrapeConfig
            name: vmscrapeconfigs.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMServiceScrape is scrape configuration for endpoints associated with
              kubernetes service,
              it generates scrape configuration for vmagent based on selectors.
              result config will scrape service endpoints
            displayName: VMService Scrape
            kind: VMServiceScrape
            name: vmservicescrapes.operator.victoriametrics.com
            version: v1beta1
          - description: VMSingle  is fast, cost-effective and scalable time-series
              database.
            displayName: VMSingle
            kind: VMSingle
            name: vmsingles.operator.victoriametrics.com
            version: v1beta1
          - description: VMStaticScrape  defines static targets configuration for
              scraping.
            displayName: VMStatic Scrape
            kind: VMStaticScrape
            name: vmstaticscrapes.operator.victoriametrics.com
            version: v1beta1
          - description: VMUser is the Schema for the vmusers API
            displayName: VMUser
            kind: VMUser
            name: vmusers.operator.victoriametrics.com
            version: v1beta1
          - description: VTCluster is fast, cost-effective and scalable traces database.
            displayName: VTCluster
            kind: VTCluster
            name: vtclusters.operator.victoriametrics.com
            version: v1
          - description: |-
              VTSingle is fast, cost-effective and scalable traces database.
              VTSingle is the Schema for the API
            displayName: VTSingle
            kind: VTSingle
            name: vtsingles.operator.victoriametrics.com
            version: v1
        description: |
          Operator manages VictoriaMetrics applications and provides monitoring features for applications running inside and outside kubernetes cluster. It has support for prometheus-operator objects and
          provides migration mechanism.

          Operator designed to use existing installation of VictoriaMetrics inside or outside kubernetes cluster or it could
          bring own resources and fully manages them.

           You can read more about operator at docs:
            - quick start [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/quick-start.md)
            - high availability [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/high-availability.md)
            - design and description of implementation [design](https://github.com/VictoriaMetrics/operator/blob/master/docs/design.md)
            - operator objects description [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/api.md)

          # VictoriaMetrics
          * VictoriaMetrics can be used as long-term storage for Prometheus or for [vmagent](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmagent/README.md).
          See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#prometheus-setup) for details.
          * Supports [Prometheus querying API](https://prometheus.io/docs/prometheus/latest/querying/api/), so it can be used as Prometheus drop-in replacement in Grafana.
          VictoriaMetrics implements [MetricsQL](https://github.com/VictoriaMetrics/VictoriaMetrics/wiki/MetricsQL) query language, which is inspired by PromQL.
          * Supports global query view. Multiple Prometheus instances may write data into VictoriaMetrics. Later this data may be used in a single query.
          * High performance and good scalability for both [inserts](https://medium.com/@valyala/high-cardinality-tsdb-benchmarks-victoriametrics-vs-timescaledb-vs-influxdb-13e6ee64dd6b)
          and [selects](https://medium.com/@valyala/when-size-matters-benchmarking-victoriametrics-vs-timescale-and-influxdb-6035811952d4).
          [Outperforms InfluxDB and TimescaleDB by up to 20x](https://medium.com/@valyala/measuring-vertical-scalability-for-time-series-databases-in-google-cloud-92550d78d8ae).
          * [Uses 10x less RAM than InfluxDB](https://medium.com/@valyala/insert-benchmarks-with-inch-influxdb-vs-victoriametrics-e31a41ae2893) when working with millions of unique time series (aka high cardinality).
          * Optimized for time series with high churn rate. Think about [prometheus-operator](https://github.com/coreos/prometheus-operator) metrics from frequent deployments in Kubernetes.
          * High data compression, so [up to 70x more data points](https://medium.com/@valyala/when-size-matters-benchmarking-victoriametrics-vs-timescale-and-influxdb-6035811952d4)
          may be crammed into limited storage comparing to TimescaleDB.
          * Optimized for storage with high-latency IO and low IOPS (HDD and network storage in AWS, Google Cloud, Microsoft Azure, etc). See [graphs from these benchmarks](https://medium.com/@valyala/high-cardinality-tsdb-benchmarks-victoriametrics-vs-timescaledb-vs-influxdb-13e6ee64dd6b).
          * A single-node VictoriaMetrics may substitute moderately sized clusters built with competing solutions such as Thanos, M3DB, Cortex, InfluxDB or TimescaleDB.
          See [vertical scalability benchmarks](https://medium.com/@valyala/measuring-vertical-scalability-for-time-series-databases-in-google-cloud-92550d78d8ae),
          [comparing Thanos to VictoriaMetrics cluster](https://medium.com/@valyala/comparing-thanos-to-victoriametrics-cluster-b193bea1683)
          and [Remote Write Storage Wars](https://promcon.io/2019-munich/talks/remote-write-storage-wars/) talk
          from [PromCon 2019](https://promcon.io/2019-munich/talks/remote-write-storage-wars/).
          * Easy operation:
              * VictoriaMetrics consists of a single [small executable](https://medium.com/@valyala/stripping-dependency-bloat-in-victoriametrics-docker-image-983fb5912b0d) without external dependencies.
              * All the configuration is done via explicit command-line flags with reasonable defaults.
              * All the data is stored in a single directory pointed by `-storageDataPath` flag.
              * Easy and fast backups from [instant snapshots](https://medium.com/@valyala/how-victoriametrics-makes-instant-snapshots-for-multi-terabyte-time-series-data-e1f3fb0e0282)
              to S3 or GCS with [vmbackup](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmbackup/README.md) / [vmrestore](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmrestore/README.md).
              See [this article](https://medium.com/@valyala/speeding-up-backups-for-big-time-series-databases-533c1a927883) for more details.
          * Storage is protected from corruption on unclean shutdown (i.e. OOM, hardware reset or `kill -9`) thanks to [the storage architecture](https://medium.com/@valyala/how-victoriametrics-makes-instant-snapshots-for-multi-terabyte-time-series-data-e1f3fb0e0282).
          * Supports metrics' scraping, ingestion and [backfilling](https://github.com/VictoriaMetrics/VictoriaMetrics#backfilling) via the following protocols:
            * [Metrics from Prometheus exporters](https://github.com/prometheus/docs/blob/master/content/docs/instrumenting/exposition_formats.md#text-based-format)
              such as [node_exporter](https://github.com/prometheus/node_exporter). See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-scrape-prometheus-exporters-such-as-node-exporter) for details.
              * [Prometheus remote write API](https://prometheus.io/docs/prometheus/latest/configuration/configuration/#remote_write)
              * [InfluxDB line protocol](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-send-data-from-influxdb-compatible-agents-such-as-telegraf) over HTTP, TCP and UDP.
              * [Graphite plaintext protocol](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-send-data-from-graphite-compatible-agents-such-as-statsd) with [tags](https://graphite.readthedocs.io/en/latest/tags.html#carbon)
              if `-graphiteListenAddr` is set.
              * [OpenTSDB put message](https://github.com/VictoriaMetrics/VictoriaMetrics#sending-data-via-telnet-put-protocol) if `-opentsdbListenAddr` is set.
              * [HTTP OpenTSDB /api/put requests](https://github.com/VictoriaMetrics/VictoriaMetrics#sending-opentsdb-data-via-http-apiput-requests) if `-opentsdbHTTPListenAddr` is set.
              * [How to import time series data](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-time-series-data).
              * [Prometheus exposition format](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-data-in-prometheus-exposition-format).
              * [Arbitrary CSV data](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-csv-data).
          * Supports metrics' relabeling. See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#relabeling) for details.
          * Ideally works with big amounts of time series data from Kubernetes, IoT sensors, connected cars, industrial telemetry, financial data and various Enterprise workloads.
        displayName: VictoriaMetrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - prometheus
        - victoriametrics
        links:
        - name: VictoriaMetrics
          url: https://victoriametrics.com
        maintainers:
        - email: nik@victoriametrics.com
          name: f41gh7
        - email: info@victoriametrics.com
          name: info
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: VictoriaMetrics
        relatedImages:
        - docker.io/victoriametrics/operator:v0.63.0
        version: 0.63.0
      entries:
      - name: victoriametrics-operator.v0.63.0
        version: 0.63.0
      - name: victoriametrics-operator.v0.62.0
        version: 0.62.0
      - name: victoriametrics-operator.v0.61.2
        version: 0.61.2
      - name: victoriametrics-operator.v0.61.1
        version: 0.61.1
      - name: victoriametrics-operator.v0.61.0
        version: 0.61.0
      - name: victoriametrics-operator.v0.60.2
        version: 0.60.2
      - name: victoriametrics-operator.v0.60.1
        version: 0.60.1
      - name: victoriametrics-operator.v0.60.0
        version: 0.60.0
      - name: victoriametrics-operator.v0.59.2
        version: 0.59.2
      - name: victoriametrics-operator.v0.59.1
        version: 0.59.1
      - name: victoriametrics-operator.v0.59.0
        version: 0.59.0
      - name: victoriametrics-operator.v0.58.1
        version: 0.58.1
      - name: victoriametrics-operator.v0.58.0
        version: 0.58.0
      - name: victoriametrics-operator.v0.57.0
        version: 0.57.0
      - name: victoriametrics-operator.v0.56.0
        version: 0.56.0
      - name: victoriametrics-operator.v0.55.0
        version: 0.55.0
      - name: victoriametrics-operator.v0.54.1
        version: 0.54.1
      - name: victoriametrics-operator.v0.53.0
        version: 0.53.0
      - name: victoriametrics-operator.v0.52.0
        version: 0.52.0
      - name: victoriametrics-operator.v0.51.1
        version: 0.51.1
      - name: victoriametrics-operator.v0.49.0
        version: 0.49.0
      - name: victoriametrics-operator.v0.48.4
        version: 0.48.4
      - name: victoriametrics-operator.v0.48.3
        version: 0.48.3
      - name: victoriametrics-operator.v0.48.1
        version: 0.48.1
      - name: victoriametrics-operator.v0.47.3
        version: 0.47.3
      - name: victoriametrics-operator.v0.47.1
        version: 0.47.1
      - name: victoriametrics-operator.v0.47.0
        version: 0.47.0
      - name: victoriametrics-operator.v0.46.4
        version: 0.46.4
      - name: victoriametrics-operator.v0.46.3
        version: 0.46.3
      - name: victoriametrics-operator.v0.39.4
        version: 0.39.4
      - name: victoriametrics-operator.v0.33.0
        version: 0.33.0
      - name: victoriametrics-operator.v0.32.1
        version: 0.32.1
      - name: victoriametrics-operator.v0.32.0
        version: 0.32.0
      - name: victoriametrics-operator.v0.29.1
        version: 0.29.1
      - name: victoriametrics-operator.v0.27.2
        version: 0.27.2
      - name: victoriametrics-operator.v0.26.2
        version: 0.26.2
      - name: victoriametrics-operator.v0.26.0
        version: 0.26.0
      - name: victoriametrics-operator.v0.25.1
        version: 0.25.1
      - name: victoriametrics-operator.v0.25.0
        version: 0.25.0
      - name: victoriametrics-operator.v0.24.0
        version: 0.24.0
      - name: victoriametrics-operator.v0.23.3
        version: 0.23.3
      - name: victoriametrics-operator.v0.22.1
        version: 0.22.1
      - name: victoriametrics-operator.v0.22.0
        version: 0.22.0
      - name: victoriametrics-operator.v0.21.0
        version: 0.21.0
      - name: victoriametrics-operator.v0.20.1
        version: 0.20.1
      - name: victoriametrics-operator.v0.19.1
        version: 0.19.1
      - name: victoriametrics-operator.v0.19.0
        version: 0.19.0
      - name: victoriametrics-operator.v0.18.2
        version: 0.18.2
      - name: victoriametrics-operator.v0.18.1
        version: 0.18.1
      - name: victoriametrics-operator.v0.18.0
        version: 0.18.0
      name: beta
    defaultChannel: beta
    packageName: victoriametrics-operator
    provider:
      name: VictoriaMetrics
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: F5 Networks Inc.
      provider-url: ""
    name: f5-bigip-ctlr-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: alpha
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: beta
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: stable
    defaultChannel: stable
    packageName: f5-bigip-ctlr-operator
    provider:
      name: F5 Networks Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dns-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dns-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSHealthCheckProbe",
                "metadata": {
                  "name": "$NAME"
                },
                "spec": {
                  "additionalHeadersRef": {
                    "name": "headers"
                  },
                  "address": "192.168.0.16",
                  "allowInsecureCertificate": true,
                  "failureThreshold": 5,
                  "hostname": "test.com",
                  "interval": "60s",
                  "path": "/healthz",
                  "port": 443,
                  "protocol": "HTTPS"
                }
              },
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSRecord",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dns-operator",
                    "app.kubernetes.io/instance": "dnsrecord-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dnsrecord",
                    "app.kubernetes.io/part-of": "dns-operator"
                  },
                  "name": "dnsrecord-sample"
                },
                "spec": {
                  "endpoints": [
                    {
                      "dnsName": "dnsrecord-simple.example.com",
                      "recordTTL": 60,
                      "recordType": "A",
                      "targets": [
                        "52.215.108.61",
                        "52.30.101.221"
                      ]
                    }
                  ],
                  "providerRef": {
                    "name": "dns-provider-creds"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/dns-rhel9-operator@sha256:8ad3876e36ba2f76355499d98a97672f54f26754fc9d219d9cb1c5b9f9874430
          createdAt: "2025-04-07T08:13:58Z"
          description: Configures how north-south traffic from outside the network
            is balanced and reaches Gateways
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kuadrant/dns-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSHealthCheckProbe is the Schema for the dnshealthcheckprobes
              API
            displayName: DNSHealthCheckProbe
            kind: DNSHealthCheckProbe
            name: dnshealthcheckprobes.kuadrant.io
            version: v1alpha1
          - description: DNSRecord is the Schema for the dnsrecords API
            displayName: DNSRecord
            kind: DNSRecord
            name: dnsrecords.kuadrant.io
            version: v1alpha1
        description: Configures how north-south traffic from outside the network is
          balanced and reaches Gateways - Part of Red Hat Connectivity Link
        displayName: Red Hat - DNS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dns
        - kuadrant
        links:
        - name: GitHub
          url: https://github.com/kuadrant/dns-operator
        - name: Documentation
          url: https://github.com/Kuadrant/dns-operator/tree/main/docs
        maintainers:
        - email: mnairn@redhat.com
          name: Michael Nairn
        - email: pbrookes@redhat.com
          name: Phil Brookes
        - email: cbrookes@redhat.com
          name: Craig Brookes
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhcl-1/dns-rhel9-operator@sha256:8ad3876e36ba2f76355499d98a97672f54f26754fc9d219d9cb1c5b9f9874430
        version: 1.1.0
      entries:
      - name: dns-operator.v1.1.0
        version: 1.1.0
      - name: dns-operator.v1.0.2
        version: 1.0.2
      - name: dns-operator.v1.0.1
        version: 1.0.1
      - name: dns-operator.v0.12.0
        version: 0.12.0
      name: stable
    defaultChannel: stable
    packageName: dns-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: patch-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: patch-operator.v0.1.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Patch",
                "metadata": {
                  "name": "test-complex-patch"
                },
                "spec": {
                  "patches": {
                    "test-complex-patch": {
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 1).metadata.name }}: {{ (index . 2).metadata.name }}    \n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "kind": "Namespace",
                          "name": "test-patch-operator"
                        },
                        {
                          "apiVersion": "v1",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "test-patch-operator"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "test-patch-operator"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/patch-operator@sha256:030ade9b9428deca3ac6390618dc4d6e6d728577d1467a22dc4d4b68c9848886
          createdAt: "2024-05-10T14:07:19Z"
          description: An operator to apply patches to Kubernetes objects in a declarative
            way.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: patch-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/patch-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Patch
            name: patches.redhatcop.redhat.io
            version: v1alpha1
        description: "The patch operator helps with defining patches in a declarative
          way. This operator has two main features:\n\n1. [ability to patch an object
          at creation time via a mutating webhook](#creation-time-patch-injection)\n2.
          [ability to enforce patches on one or more objects via a controller](#runtime-patch-enforcement)\n
          \n## Creation-time patch injection\n\nWhy apply a patch at creation time
          when you could directly create the correct object? The reason is that sometime
          the correct value depends on configuration set on the specific cluster in
          which the object is being deployed. For example, an ingress/route hostname
          might depend on the specific cluster. Consider the following example based
          on cert-manager:\n \n```yaml\napiVersion: cert-manager.io/v1\nkind: ClusterIssuer\nmetadata:\n
          \ name: letsencrypt-issuer\nspec:\n  acme:\n    server: 'https://acme-v02.api.letsencrypt.org/directory'\n
          \   email: {{ .Values.letsencrypt.email }}\n    privateKeySecretRef:\n      name:
          letsencrypt-staging\n    solvers:  \n    - dns01:\n        route53:\n          accessKeyID:
          << access_key >>\n          secretAccessKeySecretRef:\n            name:
          cert-manager-dns-credentials\n            key: aws_secret_access_key\n          region:
          << region >>\n          hostedZoneID: << hosted_zone_id >>\n ```\n\nIn this
          example the fields: `<< access_key >>`, `<< region >>` and `<< hosted_zone_id
          >>` are dependent on the specific region in which the cluster is being deployed
          and in many cases they are discoverable from other configurations already
          present in the cluster. If you want to deploy the above Cluster Issuer object
          with a gitops approach, then there is no easy way to discover those values.
          The solution so far is to manually discover those values and create a different
          gitops configuration for each cluster. But consider if you could look up
          values at deploy time based on the cluster you are deploying to. Here is
          how this object might look:\n\n```yaml\napiVersion: cert-manager.io/v1\nkind:
          ClusterIssuer\nmetadata:\n  name: letsencrypt-issuer\n  namespace: {{ .Release.Namespace
          }}\n  annotations:\n    \"redhat-cop.redhat.io/patch\": |\n      spec:\n
          \       acme:\n        - dns01:\n            route53:\n              accessKeyID:
          {{ (lookup \"v1\" \"Secret\" .metadata.namespace \"cert-manager-dns-credentials\").data.aws_access_key_id
          | b64dec }}\n              secretAccessKeySecretRef:\n                name:
          cert-manager-dns-credentials\n                key: aws_secret_access_key\n
          \             region: {{ (lookup \"config.openshift.io/v1\" \"Infrastructure\"
          \"\" \"cluster\").status.platformStatus.aws.region }}\n              hostedZoneID:
          {{ (lookup \"config.openshift.io/v1\" \"DNS\" \"\" \"cluster\").spec.publicZone.id
          }} \nspec:\n  acme:\n    server: 'https://acme-v02.api.letsencrypt.org/directory'\n
          \   email: {{ .Values.letsencrypt.email }}\n    privateKeySecretRef:\n      name:
          letsencrypt-staging\n    solvers:  \n    - dns01:\n        route53:\n          accessKeyID:
          << access_key >>\n          secretAccessKeySecretRef:\n            name:
          cert-manager-dns-credentials\n            key: aws_secret_access_key\n          region:
          << region >>\n          hostedZoneID: << hosted_zone_id >>\n ```\n\nThe
          annotation specifies a patch that will be applied by a [MutatingWebhook](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/),
          as you can see the three values are now looked up from different configurations
          in the cluster.\n\nTwo annotations influences the behavior of this MutatingWebhook\n
          \n1. \"redhat-cop.redhat.io/patch\" : this is the patch itself. The patch
          is evaluated as a template with the object itself as it's only parameter.
          The template is expressed in golang template notation and supports the same
          functions as helm template including the [lookup](https://helm.sh/docs/chart_template_guide/functions_and_pipelines/#using-the-lookup-function)
          function which plays a major role here. The patch must be expressed in yaml
          for readability. It will be converted to json by the webhook logic.\n2.
          \"redhat-cop.redhat.io/patch-type\" : this is the type of json patch. The
          possible values are: `application/json-patch+json`, `application/merge-patch+json`
          and `application/strategic-merge-patch+json`. If this annotation is omitted
          it defaults to strategic merge.\n\n### Security Considerations\n \nThe lookup
          function, if used by the template, is executed with a client which impersonates
          the user issuing the object creation/update request. This should prevent
          security permission leakage.\n\n### Installing the creation time webhook\n
          \nThe creation time webhook is not installed by the operator. This is because
          there is no way to know which specific object type should be intercepted
          and intercepting all of the types would be too inefficient. It's up to the
          administrator then to install the webhook. Here is some guidance.\n\nIf
          you installed the operator via OLM, use the following webhook template:\n\n```yaml\napiVersion:
          admissionregistration.k8s.io/v1\n kind: MutatingWebhookConfiguration\nmetadata:\n
          \ name: patch-operator-inject\n  annotations:\n    service.beta.openshift.io/inject-cabundle:
          \"true\"\nwebhooks:\n - admissionReviewVersions:\n  - v1\n  clientConfig:\n
          \   service:\n      name: patch-operator-webhook-service\n      namespace:
          patch-operator\n      path: /inject\n  failurePolicy: Fail\n  name: patch-operator-inject.redhatcop.redhat.io\n
          \ rules:\n  - << add your intercepted objects here >>\n  sideEffects: None\n```\n\nIf
          you installed the operator via the Helm chart and are using cert-manager,
          use the following webhook template:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\n
          kind: MutatingWebhookConfiguration\nmetadata:\n  name: patch-operator-inject\n
          \ annotations:\n    cert-manager.io/inject-ca-from: '{{ .Release.Namespace
          }}/webhook-server-cert'\n webhooks:\n- admissionReviewVersions:\n  - v1\n
          \ clientConfig:\n    service:\n      name: patch-operator-webhook-service\n
          \     namespace: patch-operator\n      path: /inject\n  failurePolicy: Fail\n
          \ name: patch-operator-inject.redhatcop.redhat.io\n  rules:\n  - << add
          your intercepted objects here >>\n  sideEffects: None\n```  \n\nYou should
          need to enable the webhook only for `CREATE` operations. So for example
          to enable the webhook on configmaps:\n\n```yaml\n  rules:\n  - apiGroups:\n
          \   - \"\"\n    apiVersions:\n    - v1\n    operations:\n    - CREATE\n
          \   resources:\n    - configmaps\n```\n\n## Runtime patch enforcement\n\nThere
          are situations when we need to patch pre-existing objects. Again this is
          a use case that is hard to model with gitops operators which will work only
          on object that they own. Especially with sophisticated Kubernetes distributions,
          it is not uncommon that a Kubernetes instance, at installation time, is
          configured with some default settings. Changing those configurations means
          patching those objects. For example, let's take the case of OpenShift Oauth
          configuration. This object is present by default and it is expected to be
          patched with any newly enabled authentication mechanism. This is how it
          looks like after installation:\n\n```yaml\napiVersion: config.openshift.io/v1\n
          kind: OAuth\nmetadata:\n  name: cluster\n  ownerReferences:\n    - apiVersion:
          config.openshift.io/v1\n      kind: ClusterVersion\n      name: version\n
          \     uid: 9a9d450b-3076-4e30-ac05-a889d6341fc3\n  resourceVersion: '20405124'\n
          spec: []\n```\n\nIf we need to patch it we can use the patch controller
          and the `Patch` object as follows:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: Patch\nmetadata:\n  name: gitlab-ocp-oauth-provider\n  namespace:
          openshift-config\nspec:\n  serviceAccountRef:\n    name: default\n  patches:\n
          \   gitlab-ocp-oauth-provider:\n      targetObjectRef:\n        apiVersion:
          config.openshift.io/v1\n        kind: OAuth\n        name: cluster\n      patchTemplate:
          |\n        spec:\n          identityProviders:\n          - name: my-github
          \n            mappingMethod: claim \n            type: GitHub\n            github:\n
          \             clientID: \"{{ (index . 1).data.client_id | b64dec }}\" \n
          \             clientSecret: \n                name: ocp-github-app-credentials\n
          \             organizations: \n              - my-org\n              teams:
          []            \n      patchType: application/merge-patch+json\n      sourceObjectRefs:\n
          \     - apiVersion: v1\n        kind: Secret\n        name: ocp-github-app-credentials\n
          \       namespace: openshift-config\n```\n\nThis will cause the OAuth object
          to be patched and the patch to be enforced. That means that if anything
          changes on the secret that we use a parameter (which may be rotated) or
          in Oauth object itself, the patch will be reapplied. In this case we are
          adding a gitlab authentication provider.\n\n A `patch` has the following
          fields:\n\n`targetObjectRef` this refers to the object(s) receiving the
          patch. Mutliple object can be selected based on the following rules:\n \n|
          Namespaced Type | Namespace | Name | Selection type |\n| --- | --- | ---
          | --- |\n| yes | null | null | multiple selection across namespaces |\n|
          yes | null | not null | multiple selection across namespaces where the name
          corresponds to the passed name |\n| yes | not null | null | multiple selection
          within a namespace |\n| yes | not null | not nul | single selection |\n|
          no | N/A | null | multiple selection  |\n| no | N/A | not null | single
          selection |\n\nSelection can be further narrowed down by filtering by labels
          and/or annotations using the `labelSelector` and `annotationSelector` fields.
          The patch will be applied to all of the selected instances.\n\n`sourceObjectRefs`
          these are the objects that will be watched and become part of the parameters
          of the patch template. Name and Namespace of sourceRefObjects are interpreted
          as golang templates with the current target instance and the only parameter.
          This allows to select different source object for each target object.\n
          \nSo, for example, with this patch:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: Patch\nmetadata:\n  name: multiple-namespaced-targets-patch\nspec:\n
          \ serviceAccountRef:\n    name: default\n  patches:\n    multiple-namespaced-targets-patch:\n
          \     targetObjectRef:\n        apiVersion: v1\n        kind: ServiceAccount\n
          \       name: deployer\n      patchTemplate: |\n        metadata:\n          annotations:\n
          \           {{ (index . 1).metadata.uid }}: {{ (index . 1) }}\n      patchType:
          application/strategic-merge-patch+json\n      sourceObjectRefs:\n      -
          apiVersion: v1\n        kind: ServiceAccount\n        name: default\n        namespace:
          \"{{ .metadata.namespace }}\"\n        fieldPath: $.metadata.uid\n```\n\nThe
          `deployer` service accounts from all namespaces are selected as target of
          this patch, each patch template will receive a different parameter and that
          is the `default` service account of the same namespace as the namespace
          of the `deployer` service account being processed.\n\n`sourceObjectRefs`
          also have the `fieldPath` field which can contain a jsonpath expression.
          If a value is passed the jsonpath expression will be calculate for the current
          source object and the result will be passed as parameter of the template.\n\n`patchTemplate`
          This is the the template that will be evaluated. The result must be a valid
          patch compatible with the requested type and expressed in yaml for readability.
          The parameters passed to the template are the target object and then the
          all of the source object. So if you want to refer to the target object in
          the template you can use this expression `(index . 0)`. Higher indexes refer
          to the sourceObjectRef array. The template is expressed in golang template
          notation and supports the same functions as helm template.\n\n`patchType`
          is the type of the json patch. The possible values are: `application/json-patch+json`,
          `application/merge-patch+json` and `application/strategic-merge-patch+json`.
          If this annotation is omitted it defaults to strategic merge.\n\n### Patch
          Controller Security Considerations\n\nThe patch enforcement enacted by the
          patch controller is executed with a client which uses the service account
          referenced by the `serviceAccountRef` field. So before a patch object can
          actually work an administrator must have granted the needed permissions
          to a service account in the same namespace. The `serviceAccountRef` will
          default to the `default` service account if not specified.\n\n### Patch
          Controller Performance Considerations\n\nThe patch controller will create
          a controller-manager and per `Patch` object and a reconciler for each of
          the `PatchSpec` defined in the array on patches in the `Patch` object.\nThese
          reconcilers share the same cached client. In order to be able to watch changes
          on target and source objects of a `PatchSpec`, all of the target and source
          object type instances will be cached by the client. This is a normal behavior
          of a controller-manager client, but it implies that if you create patches
          on object types that have many instances in etcd (Secrets, ServiceAccounts,
          Namespaces for example), the patch operator instance will require a significant
          amount of memory. A way to contain this issue is to try to aggregate together
          `PatchSpec` that deal with the same object types. This will cause those
          object type instances to cached only once.  \n"
        displayName: Patch Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - patch
        - enforcing
        - declarative
        links:
        - name: Patch Operator
          url: https://github.com/redhat-cop/patch-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/patch-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/patch-operator@sha256:030ade9b9428deca3ac6390618dc4d6e6d728577d1467a22dc4d4b68c9848886
        version: 0.1.11
      entries:
      - name: patch-operator.v0.1.11
        version: 0.1.11
      - name: patch-operator.v0.1.10
        version: 0.1.10
      - name: patch-operator.v0.1.9
        version: 0.1.9
      - name: patch-operator.v0.1.8
        version: 0.1.8
      - name: patch-operator.v0.1.7
        version: 0.1.7
      - name: patch-operator.v0.1.6
        version: 0.1.6
      - name: patch-operator.v0.1.5
        version: 0.1.5
      - name: patch-operator.v0.1.4
        version: 0.1.4
      - name: patch-operator.v0.1.3
        version: 0.1.3
      - name: patch-operator.v0.1.2
        version: 0.1.2
      - name: patch-operator.v0.1.1
        version: 0.1.1
      - name: patch-operator.v0.1.0
        version: 0.1.0
      - name: patch-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: patch-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-elbv2-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-elbv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "LoadBalancer",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "TargetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/elbv2-controller:1.1.1
          createdAt: "2025-09-19T22:16:38Z"
          description: AWS ELB controller is a service controller for managing ELB
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Listener represents the state of an AWS elbv2 Listener resource.
            displayName: Listener
            kind: Listener
            name: listeners.elbv2.services.k8s.aws
            version: v1alpha1
          - description: LoadBalancer represents the state of an AWS elbv2 LoadBalancer
              resource.
            displayName: LoadBalancer
            kind: LoadBalancer
            name: loadbalancers.elbv2.services.k8s.aws
            version: v1alpha1
          - description: Rule represents the state of an AWS elbv2 Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.elbv2.services.k8s.aws
            version: v1alpha1
          - description: TargetGroup represents the state of an AWS elbv2 TargetGroup
              resource.
            displayName: TargetGroup
            kind: TargetGroup
            name: targetgroups.elbv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Load Balancing (ELB) resources in AWS from within your Kubernetes cluster.

          **About Amazon ELB**

          Elastic Load Balancing automatically distributes your incoming traffic across multiple targets, such as EC2 instances, containers, and IP addresses, in one or more Availability Zones. It monitors the health of its registered targets, and routes traffic only to the healthy targets. Elastic Load Balancing scales your load balancer capacity automatically in response to changes in incoming traffic.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ELB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elbv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ELB Developer Resources
          url: https://aws.amazon.com/elasticloadbalancing
        maintainers:
        - email: ack-maintainers@amazon.com
          name: elb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/elbv2-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-elbv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-elbv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-elbv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-elbv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-elbv2-controller.v1.0.10
        version: 1.0.10
      - name: ack-elbv2-controller.v1.0.9
        version: 1.0.9
      - name: ack-elbv2-controller.v1.0.8
        version: 1.0.8
      - name: ack-elbv2-controller.v1.0.7
        version: 1.0.7
      - name: ack-elbv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-elbv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-elbv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-elbv2-controller.v1.0.3
        version: 1.0.3
      - name: ack-elbv2-controller.v1.0.2
        version: 1.0.2
      - name: ack-elbv2-controller.v1.0.1
        version: 1.0.1
      - name: ack-elbv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-elbv2-controller.v0.0.11
        version: 0.0.11
      - name: ack-elbv2-controller.v0.0.10
        version: 0.0.10
      - name: ack-elbv2-controller.v0.0.9
        version: 0.0.9
      - name: ack-elbv2-controller.v0.0.8
        version: 0.0.8
      - name: ack-elbv2-controller.v0.0.6
        version: 0.0.6
      - name: ack-elbv2-controller.v0.0.5
        version: 0.0.5
      - name: ack-elbv2-controller.v0.0.4
        version: 0.0.4
      - name: ack-elbv2-controller.v0.0.3
        version: 0.0.3
      - name: ack-elbv2-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-elbv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hazelcast, Inc
      provider-url: ""
    name: hazelcast-platform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hazelcast-platform-operator.v5.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "cache-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "CronHotBackup",
                "metadata": {
                  "name": "cronhotbackup-sample"
                },
                "spec": {
                  "hotBackupTemplate": {
                    "metadata": {
                      "labels": {
                        "cron-hotbackup": "true"
                      }
                    },
                    "spec": {
                      "hazelcastResourceName": "hazelcast"
                    }
                  },
                  "schedule": "*/30 * * * *"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast"
                },
                "spec": {
                  "clusterSize": 3,
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast-sample"
                },
                "spec": {
                  "clusterSize": 3,
                  "jet": {
                    "bucketConfig": {
                      "bucketURI": "gs://operator-user-code/jetJobs",
                      "secretName": "br-secret-gcp"
                    },
                    "enabled": true,
                    "resourceUploadEnabled": true
                  },
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HazelcastEndpoint",
                "metadata": {
                  "name": "hazelcastendpoint-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "port": 5701,
                  "type": "Discovery"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HotBackup",
                "metadata": {
                  "name": "hot-backup"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-longrun"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-longrun-2.0.0.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-1.0.2.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJobSnapshot",
                "metadata": {
                  "name": "jetjobsnapshot-sample"
                },
                "spec": {
                  "cancelJob": false,
                  "jetJobResourceName": "jet-job"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ManagementCenter",
                "metadata": {
                  "name": "managementcenter"
                },
                "spec": {
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "hazelcastClusters": [
                    {
                      "address": "hazelcast",
                      "name": "dev"
                    }
                  ],
                  "licenseKeySecretName": "hazelcast-license-key",
                  "persistence": {
                    "enabled": true,
                    "size": "10Gi"
                  },
                  "repository": "hazelcast/management-center"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Map",
                "metadata": {
                  "name": "map"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "MultiMap",
                "metadata": {
                  "name": "multimap-sample"
                },
                "spec": {
                  "backupCount": 3,
                  "binary": true,
                  "collectionType": "LIST",
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "queue-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ReplicatedMap",
                "metadata": {
                  "name": "replicatedmap-sample"
                },
                "spec": {
                  "asyncFillup": true,
                  "hazelcastResourceName": "hazelcast",
                  "inMemoryFormat": "OBJECT"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "topic-sample"
                },
                "spec": {
                  "globalOrderingEnabled": true,
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "UserCodeNamespace",
                "metadata": {
                  "name": "ucn-sample"
                },
                "spec": {
                  "bucketConfig": {
                    "bucketURI": "gs://operator-user-code/userCodeNamespace",
                    "secretName": "br-secret-gcp"
                  },
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanReplication",
                "metadata": {
                  "name": "wanreplication-sample"
                },
                "spec": {
                  "endpoints": "35.192.33.252",
                  "resources": [
                    {
                      "kind": "Map",
                      "name": "map-sample"
                    },
                    {
                      "kind": "Hazelcast",
                      "name": "hz-sample"
                    }
                  ],
                  "targetClusterName": "dev"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanSync",
                "metadata": {
                  "name": "wansync-sample"
                },
                "spec": {
                  "wanReplicationResourceName": "wan-replication-sample"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: docker.io/hazelcast/hazelcast-platform-operator@sha256:99005afda5c71ceeac200928dfed4ea35480f871053101e3aeaa44005e3fda14
          createdAt: "2024-07-29T15:35:56Z"
          description: '[DEPRECATED] Use the certified Hazelcast Operator instead'
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hazelcast/hazelcast-platform-operator/
          support: Hazelcast, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.hazelcast.com
            version: v1alpha1
          - description: CronHotBackup is the Schema for the cronhotbackups API
            displayName: Cron Hot Backup
            kind: CronHotBackup
            name: cronhotbackups.hazelcast.com
            version: v1alpha1
          - description: HazelcastEndpoint is the Schema for the hazelcastendpoints
              API
            displayName: Hazelcast Endpoint
            kind: HazelcastEndpoint
            name: hazelcastendpoints.hazelcast.com
            version: v1alpha1
          - description: Hazelcast is the Schema for the hazelcasts API
            displayName: Hazelcast
            kind: Hazelcast
            name: hazelcasts.hazelcast.com
            version: v1alpha1
          - description: HotBackup is the Schema for the hot backup API
            displayName: Hot Backup
            kind: HotBackup
            name: hotbackups.hazelcast.com
            version: v1alpha1
          - description: JetJob is the Schema for the jetjobs API
            displayName: Jet Job
            kind: JetJob
            name: jetjobs.hazelcast.com
            version: v1alpha1
          - description: JetJobSnapshot is the Schema for the jetjobsnapshots API
            displayName: Jet Job Snapshot
            kind: JetJobSnapshot
            name: jetjobsnapshots.hazelcast.com
            version: v1alpha1
          - description: ManagementCenter is the Schema for the managementcenters
              API
            displayName: Management Center
            kind: ManagementCenter
            name: managementcenters.hazelcast.com
            version: v1alpha1
          - description: Map is the Schema for the maps API
            displayName: Map
            kind: Map
            name: maps.hazelcast.com
            version: v1alpha1
          - description: MultiMap is the Schema for the multimaps API
            displayName: Multi Map
            kind: MultiMap
            name: multimaps.hazelcast.com
            version: v1alpha1
          - description: Queue is the Schema for the queues API
            displayName: Queue
            kind: Queue
            name: queues.hazelcast.com
            version: v1alpha1
          - description: ReplicatedMap is the Schema for the replicatedmaps API
            displayName: Replicated Map
            kind: ReplicatedMap
            name: replicatedmaps.hazelcast.com
            version: v1alpha1
          - description: Topic is the Schema for the topics API
            displayName: Topic
            kind: Topic
            name: topics.hazelcast.com
            version: v1alpha1
          - description: UserCodeNamespace is the Schema for the usercodenamespaces
              API
            displayName: User Code Namespace
            kind: UserCodeNamespace
            name: usercodenamespaces.hazelcast.com
            version: v1alpha1
          - description: WanReplication is the Schema for the wanreplications API
            displayName: Wan Replication
            kind: WanReplication
            name: wanreplications.hazelcast.com
            version: v1alpha1
          - description: WanSync is the Schema for the wansyncs API
            displayName: Wan Sync
            kind: WanSync
            name: wansyncs.hazelcast.com
            version: v1alpha1
        description: |
          ## Deprecated
          Use the certified Hazelcast Operator instead.
          # Hazelcast Platform Operator #

          Easily deploy Hazelcast clusters and Management Center into Kubernetes environments and manage their lifecycles.

          ## Before You Start

          If you are planning to create Hazelcast Platform Enterprise clusters, you need to [create a secret](https://docs.hazelcast.com/operator/latest/get-started#step-2-start-the-hazelcast-cluster) for the license. You can request a trial license key from [here](https://trialrequest.hazelcast.com).

          For Hazelcast Platform clusters, you can simply continue.

          ## Documentation

          1. [Get started](https://docs.hazelcast.com/operator/latest/get-started) with the Operator
          2. [Connect to the cluster from outside Kubernetes](https://guides.hazelcast.org/hazelcast-platform-operator-expose-externally)
          3. [Restore a Cluster from Cloud Storage with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-external-backup-restore)
          4. [Replicate Data between Two Hazelcast Clusters with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-wan-replication)

          ## Features

          Hazelcast Platform Operator supports the features below:

          * Custom resource for Hazelcast Platform (Enterprise) and Management Center
          * Observe status of Hazelcast clusters and Management Center instances
          * High Availability Mode configuration to create clusters that are resilient to node and zone failures
          * Support for TLS connections between members using self-signed certificates
          * Scale up and down Hazelcast clusters
          * Expose Hazelcast cluster to external clients ([Smart & Unisocket](https://docs.hazelcast.com/hazelcast/latest/clients/java#java-client-operation-modes))
          * Backup Hazelcast persistence data to cloud storage with the possibility of scheduling it and restoring the data accordingly
          * WAN Replication feature when you need to synchronize multiple Hazelcast clusters, which are connected by WANs
          * Full and Delta WAN Sync support
          * Tiered Storage support for Hazelcast and Map CRs
          * CP Subsystem configuration support for Hazelcast CR
          * User Code Deployment feature, which allows you to deploy custom and domain classes from cloud storages and URLs to Hazelcast members
          * The User Code Namespaces feature allows you to deploy custom and domain classes at runtime without a members restart
          * Support the configuration of advanced networking options
          * Support Multi-namespace configuration
          * ExecutorService and EntryProcessor support
          * Support several data structures like Map, Topic, MultiMap, ReplicatedMap, Queue and Cache which can be created dynamically via specific Custom Resources
          * MapStore, Near Cache and off-heap memory (HD memory and native memory) support for the Map CR
          * Native Memory support for the Cache CR
          * Support Jet configuration and Jet Job submission using the JetJob CR
          * Support for exporting the snapshots of JetJob CRs using JetJobSnapshot CR
          * Support for custom configurations using ConfigMap
        displayName: '[DEPRECATED] Hazelcast Platform Operator'
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hazelcast
        - keyvalue
        - in-memory
        - database
        - caching
        links:
        - name: Documentation
          url: https://docs.hazelcast.com/operator/latest/
        - name: Source Code
          url: https://github.com/hazelcast/hazelcast-platform-operator
        maintainers:
        - email: cloudnative@hazelcast.com
          name: Hazelcast Cloud Native Team
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Hazelcast, Inc
        relatedImages:
        - docker.io/hazelcast/hazelcast-platform-operator@sha256:99005afda5c71ceeac200928dfed4ea35480f871053101e3aeaa44005e3fda14
        version: 5.13.1
      entries:
      - name: hazelcast-platform-operator.v5.13.1
        version: 5.13.1
      - name: hazelcast-platform-operator.v5.13.0
        version: 5.13.0
      - name: hazelcast-platform-operator.v5.12.0
        version: 5.12.0
      - name: hazelcast-platform-operator.v5.11.0
        version: 5.11.0
      - name: hazelcast-platform-operator.v5.10.0
        version: 5.10.0
      - name: hazelcast-platform-operator.v5.9.0
        version: 5.9.0
      - name: hazelcast-platform-operator.v5.8.0
        version: 5.8.0
      - name: hazelcast-platform-operator.v5.7.0
        version: 5.7.0
      - name: hazelcast-platform-operator.v5.6.0
        version: 5.6.0
      - name: hazelcast-platform-operator.v5.5.0
        version: 5.5.0
      - name: hazelcast-platform-operator.v5.4.0
        version: 5.4.0
      - name: hazelcast-platform-operator.v5.3.0
        version: 5.3.0
      - name: hazelcast-platform-operator.v5.2.0
        version: 5.2.0
      - name: hazelcast-platform-operator.v5.1.0
        version: 5.1.0
      - name: hazelcast-platform-operator.v5.0.0
        version: 5.0.0
      name: alpha
    defaultChannel: alpha
    packageName: hazelcast-platform-operator
    provider:
      name: Hazelcast, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-storage
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-storage.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Storage",
                "metadata": {
                  "name": "storage-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore"
                  },
                  "minio": {
                    "clusterDomain": "cluster.local",
                    "helmKubectlJqImage": {
                      "repository": "docker.io/bskim45/helm-kubectl-jq"
                    },
                    "image": {
                      "repository": "prod.harbor.keysight.digital/loadcore/wap-minio",
                      "tag": "RELEASE.2021-12-29T06-49-06Z-3747-37-ocp"
                    },
                    "mcImage": {
                      "repository": "prod.harbor.keysight.digital/loadcore/wap-minio-mc",
                      "tag": "RELEASE.2021-12-29T06-52-55Z-3747-37-ocp"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "securityContext": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:32:20Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Storage
            name: storages.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Storage Operator
        displayName: Keysight WAP Storage Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Storage
          url: https://keysight-wap-storage.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-storage@sha256:9f7710cfe72658cd5456fd54f86ca6be5c9a4de2a01114c0f7531b70712bd981
        version: 0.1.21
      entries:
      - name: keysight-wap-storage.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-storage
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/cluster-nfd-operator
    name: openshift-nfd-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfd.v4.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nfd.openshift.io/v1",
                "kind": "NodeFeatureDiscovery",
                "metadata": {
                  "name": "nfd-instance",
                  "namespace": "openshift-nfd"
                },
                "spec": {
                  "customConfig": {
                    "configData": "#    - name: \"more.kernel.features\"\n#      matchOn:\n#      - loadedKMod: [\"example_kmod3\"]\n#    - name: \"more.features.by.nodename\"\n#      value: customValue\n#      matchOn:\n#      - nodename: [\"special-.*-node-.*\"]\n"
                  },
                  "operand": {
                    "image": "quay.io/openshift/origin-node-feature-discovery:4.10",
                    "servicePort": 12000
                  },
                  "workerConfig": {
                    "configData": "core:\n#  labelWhiteList:\n#  noPublish: false\n  sleepInterval: 60s\n#  sources: [all]\n#  klog:\n#    addDirHeader: false\n#    alsologtostderr: false\n#    logBacktraceAt:\n#    logtostderr: true\n#    skipHeaders: false\n#    stderrthreshold: 2\n#    v: 0\n#    vmodule:\n##   NOTE: the following options are not dynamically run-time \n##          configurable and require a nfd-worker restart to take effect\n##          after being changed\n#    logDir:\n#    logFile:\n#    logFileMaxSize: 1800\n#    skipLogHeaders: false\nsources:\n#  cpu:\n#    cpuid:\n##     NOTE: whitelist has priority over blacklist\n#      attributeBlacklist:\n#        - \"BMI1\"\n#        - \"BMI2\"\n#        - \"CLMUL\"\n#        - \"CMOV\"\n#        - \"CX16\"\n#        - \"ERMS\"\n#        - \"F16C\"\n#        - \"HTT\"\n#        - \"LZCNT\"\n#        - \"MMX\"\n#        - \"MMXEXT\"\n#        - \"NX\"\n#        - \"POPCNT\"\n#        - \"RDRAND\"\n#        - \"RDSEED\"\n#        - \"RDTSCP\"\n#        - \"SGX\"\n#        - \"SSE\"\n#        - \"SSE2\"\n#        - \"SSE3\"\n#        - \"SSE4.1\"\n#        - \"SSE4.2\"\n#        - \"SSSE3\"\n#      attributeWhitelist:\n#  kernel:\n#    kconfigFile: \"/path/to/kconfig\"\n#    configOpts:\n#      - \"NO_HZ\"\n#      - \"X86\"\n#      - \"DMI\"\n  pci:\n    deviceClassWhitelist:\n      - \"0200\"\n      - \"03\"\n      - \"12\"\n    deviceLabelFields:\n#      - \"class\"\n      - \"vendor\"\n#      - \"device\"\n#      - \"subsystem_vendor\"\n#      - \"subsystem_device\"\n#  usb:\n#    deviceClassWhitelist:\n#      - \"0e\"\n#      - \"ef\"\n#      - \"fe\"\n#      - \"ff\"\n#    deviceLabelFields:\n#      - \"class\"\n#      - \"vendor\"\n#      - \"device\"\n#  custom:\n#    - name: \"my.kernel.feature\"\n#      matchOn:\n#        - loadedKMod: [\"example_kmod1\", \"example_kmod2\"]\n#    - name: \"my.pci.feature\"\n#      matchOn:\n#        - pciId:\n#            class: [\"0200\"]\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#        - pciId :\n#            vendor: [\"8086\"]\n#            device: [\"1000\", \"1100\"]\n#    - name: \"my.usb.feature\"\n#      matchOn:\n#        - usbId:\n#          class: [\"ff\"]\n#          vendor: [\"03e7\"]\n#          device: [\"2485\"]\n#        - usbId:\n#          class: [\"fe\"]\n#          vendor: [\"1a6e\"]\n#          device: [\"089a\"]\n#    - name: \"my.combined.feature\"\n#      matchOn:\n#        - pciId:\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#          loadedKMod : [\"vendor_kmod1\", \"vendor_kmod2\"]\n"
                  }
                }
              },
              {
                "apiVersion": "nfd.openshift.io/v1alpha1",
                "kind": "NodeFeatureRule",
                "metadata": {
                  "name": "my-sample-rule-object"
                },
                "spec": {
                  "rules": [
                    {
                      "labels": {
                        "my-sample-feature": "true"
                      },
                      "matchFeatures": [
                        {
                          "feature": "kernel.loadedmodule",
                          "matchExpressions": {
                            "dummy": {
                              "op": "Exists"
                            }
                          }
                        },
                        {
                          "feature": "kernel.config",
                          "matchExpressions": {
                            "X86": {
                              "op": "In",
                              "value": [
                                "y"
                              ]
                            }
                          }
                        }
                      ],
                      "name": "my sample rule"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery,OpenShift Optional
          containerImage: quay.io/openshift/origin-cluster-nfd-operator:4.10
          description: |-
            The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

            NFD consists  of the following software components:

            The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

            ##NFD-Master
            NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

            ##NFD-Worker
            NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

            ##NFD-Topology-Updater
            NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
          olm.skipRange: '>=4.6.0 <4.11.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-nfd
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/internal-objects: '["noderesourcetopologies.topology.node.k8s.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/cluster-nfd-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              The NodeFeatureDiscovery instance is the CustomResource being watched by the NFD-Operator, and holds all the needed information to setup the behaviour of the master and worker pods
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
          - description: |
              NodeFeatureRule resource specifies a configuration for feature-based customization of node objects, such as node labeling.
            kind: NodeFeatureRule
            name: nodefeaturerules.nfd.openshift.io
            version: v1alpha1
          - description: |
              NodeResourceTopology resource describes node resources and their topology.
            kind: NodeResourceTopology
            name: noderesourcetopologies.topology.node.k8s.io
            version: v1alpha1
        description: |-
          The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

          NFD consists  of the following software components:

          The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

          ##NFD-Master
          NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

          ##NFD-Worker
          NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

          ##NFD-Topology-Updater
          NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
        displayName: Node Feature Discovery Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - feature-discovery
        - feature-detection
        - node-labels
        links:
        - name: Node Feature Discovery Operator
          url: https://docs.openshift.com/container-platform/4.9/hardware_enablement/psap-node-feature-discovery-operator.html
        - name: Node Feature Discovery Documentation
          url: https://kubernetes-sigs.github.io/node-feature-discovery/stable/get-started/index.html
        maintainers:
        - email: openshift-psap@redhat.com
          name: openshift-psap
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/cluster-nfd-operator
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy
        - quay.io/openshift/origin-cluster-nfd-operator:4.10
        version: 4.10.0
      entries:
      - name: nfd.v4.10.0
        version: 4.10.0
      - name: nfd.v4.9.0
        version: 4.9.0
      - name: nfd.v4.8.0
        version: 4.8.0
      name: stable
    defaultChannel: stable
    packageName: openshift-nfd-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/cluster-nfd-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-bedrockagent-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-bedrockagent-controller.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bedrockagent.services.k8s.aws/v1alpha1",
                "kind": "Agent",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/bedrockagent-controller:0.1.1
          createdAt: "2025-09-20T01:18:56Z"
          description: AWS Bedrock Agent controller is a service controller for managing
            Bedrock Agent resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Agent represents the state of an AWS bedrockagent Agent resource.
            displayName: Agent
            kind: Agent
            name: agents.bedrockagent.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Bedrock Agent resources in AWS from within your Kubernetes cluster.

          **About Amazon Bedrock Agent**

          Amazon Bedrock is a fully managed service that makes high-performing foundation models (FMs) from leading AI companies and Amazon available for your use through a unified API. You can choose from a wide range of foundation models to find the model that is best suited for your use case. Amazon Bedrock also offers a broad set of capabilities to build generative AI applications with security, privacy, and responsible AI. Using Amazon Bedrock, you can easily experiment with and evaluate top foundation models for your use cases, privately customize them with your data using techniques such as fine-tuning and Retrieval Augmented Generation (RAG), and build agents that execute tasks using your enterprise systems and data sources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Bedrock Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bedrockagent
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Bedrock Agent Developer Resources
          url: https://aws.amazon.com/bedrock/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: bedrock-agent maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/bedrockagent-controller:0.1.1
        version: 0.1.1
      entries:
      - name: ack-bedrockagent-controller.v0.1.1
        version: 0.1.1
      - name: ack-bedrockagent-controller.v0.1.0
        version: 0.1.0
      - name: ack-bedrockagent-controller.v0.0.6
        version: 0.0.6
      - name: ack-bedrockagent-controller.v0.0.5
        version: 0.0.5
      - name: ack-bedrockagent-controller.v0.0.4
        version: 0.0.4
      - name: ack-bedrockagent-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-bedrockagent-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: devopstales
      provider-url: devopstales.github.io
    name: community-trivy-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: community-trivy-operator.v2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "trivy-operator.devopstales.io/v1",
                "kind": "NamespaceScanner",
                "metadata": {
                  "name": "trivy-operator-main-config"
                },
                "spec": {
                  "crontab": "*/5 * * * *",
                  "namespace_selector": "trivy-scan",
                  "clusterWide": "False",
                  "policyreport": "False",
                  "registry": [
                    {
                      "name": "docker.io",
                      "user": "",
                      "password": ""
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: devopstales/trivy-operator:2.4.1
          createdAt: "2022-07-11T15:00:00Z"
          description: Trivy Operator for scheduled imagescans and an Admission Control.
          repository: https://github.com/devopstales/trivy-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Confifuration for trivy-operator.
            displayName: NamespaceScanner
            kind: NamespaceScanner
            name: namespace-scanners.trivy-operator.devopstales.io
            version: v1
          - description: |
              VulnerabilityReport summarizes vulnerabilities in application dependencies and operating system packages
              built into container images.
            displayName: VulnerabilityReport
            kind: VulnerabilityReport
            name: vulnerabilityreports.trivy-operator.devopstales.io
            version: v1
        description: |
          Trivy Operator for scheduled imagescans and an Admission Control.

          ### Add Github token to extend your Github API Rate Limit
          The Operator use Github to download Trivy database. If your Github API Rate Limit Exceeded add your Github token to the operator deployment as GITHUB_TOKEN variable.

          ```yaml
          ...
          env:
          - name: GITHUB_TOKEN
            value: "githubToken"
          ```

          ### Scheduled Image scans
          Default every 5 minutes execute a scan script. It will get image list from all namespaces with the label `trivy-scan=true`, and then scan this images with trivy, finally we will get metrics on `http://[pod-ip]:9115/metrics`

          You can chane the schedule of the image scanning by setting the `crontab` variable in the namespace-scanner object.


          ### Cache Scan results
          From the 2.4 version you can use an external redis to store scanning results and speed up the scanning process.

          ```bash
          ...
          env:
          - name: REDIS_ENABLED
            value: "True"
          - name: REDIS_BACKEND
            value: redis://redis-cache:6379
          ```

          ### Persist Trivy database cache
          To Persist Trivy cache you need to create a pvc and mount to the Deployment under `/home/trivy-operator/trivy-cache`

          ```yaml
          ...
            volumeMounts:
            - name: cache
              mountPath: "/home/trivy-operator/trivy-cache"
          volumes:
          - name: cache
            persistentVolumeClaim:
              claimName: trivy-cache
          ```

          ### Trivy Image Validator
          The admission controller function can be configured as a ValidatingWebhook in a k8s cluster. Kubernetes will send requests to the admission server when a Pod creation is initiated. The admission controller checks the image using trivy if it is in a namespace with the lable `trivy-operator-validation=true`.

          To enable the admission controller you need to set the `clusterWide` variable to `true` in the namespace-scanner object.

          Enable the scheduled scan and the admission controller for a specific namespace:
          ```bash
          kubectl label namespaces test-apps trivy-scan=true
          # or
          kubectl label namespaces test-apps trivy-operator-validation=true
          ```

          You can define policy to the Admission Controller, by adding annotation to the pod trough the deployment:

          ```bash
          spec:
            ...
            template:
              metadata:
                annotations:
                  trivy.security.devopstales.io/medium: "5"
                  trivy.security.devopstales.io/low: "10"
                  trivy.security.devopstales.io/critical: "2"
          ...
          ```

          ### Monitorng
          Trivy operator has a prometheus endpoint wher yo can get the resoults of the namespace scans and the adission controller.

          ```bash
          curl -s http://10.43.179.39:9115/metrics | grep so_vulnerabilities

          # HELP so_vulnerabilities Container vulnerabilities
          # TYPE so_vulnerabilities gauge
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="UNKNOWN"} 0
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="LOW"} 23
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="MEDIUM"} 93
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="HIGH"} 76
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="CRITICAL"} 25
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="UNKNOWN"} 0
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="LOW"} 23
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="MEDIUM"} 88
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="HIGH"} 60
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="CRITICAL"} 8
          ```

          ```bash
          curl -s http://10.43.179.39:9115/metrics | grep ac_vulnerabilities

          # HELP ac_vulnerabilities Admission Controller vulnerabilities
          # TYPE ac_vulnerabilities gauge
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="UNKNOWN"} 0.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="LOW"} 83.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="MEDIUM"} 6.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="HIGH"} 6.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="CRITICAL"} 4.0
          ```

          ### Policy Report reation
          The PolicyReport object is a protopype object probosed by the Kubernetes policy work group. The Policy Report Custom Resource Definition (CRD) can be used as a common way to provide policy results to Kubernetes cluster administrators and users, using native tools.

          Add the PolicyReport CRDs to your cluster (v1alpha2):
          ```bash
          kubectl create -f https://github.com/kubernetes-sigs/wg-policy-prototypes/raw/master/policy-report/crd/v1alpha2/wgpolicyk8s.io_policyreports.yaml
          ```

          You can enable the Policy Report creatin by setting the `policyreport` to `True` in the NamespaceScanner object.

          This objects can be visualized by the Policy Reporter UI. The Policy Reporter UI is a monitoring and Observability Tool for the PolicyReport CRD with an optional UI. It is created by Kyverno. The main goal was a tool to visualize the resoluts of the Kyverno policies, but because it uses the PolicyReports CRD it can visualize the resoults of the trivy-operator scans.
        displayName: Community Trivy Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trivy
        - security
        links:
        - name: Documentation
          url: https://devopstales.github.io/trivy-operator
        - name: Blog
          url: https://devopstales.github.io
        maintainers:
        - email: devopstales@protonmail.com
          name: devopstales
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: devopstales
          url: devopstales.github.io
        relatedImages:
        - devopstales/trivy-operator:2.4.1
        version: 2.4.0
      entries:
      - name: community-trivy-operator.v2.4.0
        version: 2.4.0
      - name: community-trivy-operator.v2.3.0
        version: 2.3.0
      - name: community-trivy-operator.v2.2.0
        version: 2.2.0
      - name: community-trivy-operator.v2.1.1
        version: 2.1.1
      name: stable
    defaultChannel: stable
    packageName: community-trivy-operator
    provider:
      name: devopstales
      url: devopstales.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: www.ibm.com
    name: ibm-elasticsearch-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-elasticsearch-operator.v1.1.2667
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
          createdAt: "2025-01-30"
          description: Elasticsearch operator for use within Cloudpak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-elasticsearch-operator
          olm.skipRange: <0.0.0 >1.1.0
          operators.openshift.io/valid-subscription: '["IBM Elasticsearch"]'
          operators.operatorframework.io/internal-objects: '["elasticsearchclusters.elasticsearch.opencontent.ibm.com"]'
          operators.operatorframework.io/operator-type: non-standalone
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ElasticsearchCluster
            name: elasticsearchclusters.elasticsearch.opencontent.ibm.com
            version: v1
        description: |-
          # Introduction
          An Elasticsearch operator for use with IBM Cloudpak services.
          # Details
          ibm-elasticsearch-operator supports installing Elasticsearch major version 6 and 7. This operator supports two CRDs Elasticsearch and Elasticsearchdetails. Currently Elasticsearch is an abstraction of Elasticsearchdetails, but there is no difference between both the CRDs. Creating either of them provisions elasticsearch instances. This abstraction layer has been created to support any changes to the operator with out breaking the CRD/API for adopting products.

          ## Prerequisites
          - Openshift 3.11+
          ## Resources Required
          In short the answer is it depends. Especially if you are running Elasticsearch 6 you need 3 or a greater odd number of replicas to make a cluster and not have split brain. Each replica will have a sidecar that needs 0.1 cpu and 2gb of ram. You should size Elasticsearch based on your use-case. The default is a full cpu and 4 cpu and 64 gb ram. To assist sizing requirements we have included a copy of [esrally](https://esrally.readthedocs.io/en/stable/) on all the docker images.
          ## Installing
          ElasticsearchCluster is the strategic Kind moving forward. There are two flavors of CRD supported that will soon be deprecated. A simple Elasticseach "All in one" crd and a more detailed CRD for teams who need greater control. Upon deploy the elasticsearch is available from the service (not the headless) created by the operator. Only https will be available on port 443. the service will also be in the ElasticsearchDetails status.

          ### Case install
          To install the component using case you must either use the native install or catalog install. Upon installing the catalog you can then use either OLM or ODLM to install but must create the OperatorGroup & Subscription for either method to avoid conflict.

          # Configuration
          Architecture:
            - x86
            - s390x
            - ppc64le
          Storage: RWO Encrypted Block Storage with 120+ IOPS/GB
          Clouds:
          - IBM CLoud (ibm-block-gold, Gen2 custom iops storage class)
          - AWS (EBS)
          - Azure (For Azure please ensure you are using a higher tier of AzureDisk storage)
          - OCS (RBD not Ceph)
          - Portworx (Non-shared replacated)
          - Please do not use with NFS

          ### Limitations
          This product is not deployable standalone and it is part of the LICCR already accepted
          ## SecurityContextConstraints Requirements
          This component requires a restricted SCC.
        displayName: IBM OpenContent Elasticsearch
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - internal
        - Elasticsearch
        maintainers:
        - email: noreply@ibm.com
          name: IBM CloudpakOpen Community
        maturity: alpha
        provider:
          name: IBM
          url: www.ibm.com
        relatedImages:
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:47a3745c78280f98f2ee352bb9192cdcd2fcd6d4c1a3627e7aa42178a0f107fd
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:30a5b2c4e3c231373fa18d8aaab16d3505cc11a8a5d4be36eaf99d7c0dee35f2
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:2668a562eb8df1117c40acbc3600d6735469d210e5aae4aa03bf9f54d0ebfe46
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:2c39ff1850b7a9d32a11e99393039ef74ab363f3fffec1e4fd542f2ae148a6b4
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:130872119947fe91cbfb8dd82b3baaf78f7adcaca5933ca394eca3f9ee84c8cc
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
        version: 1.1.2667
      entries:
      - name: ibm-elasticsearch-operator.v1.1.2667
        version: 1.1.2667
      name: v1.1
    defaultChannel: v1.1
    packageName: ibm-elasticsearch-operator
    provider:
      name: IBM
      url: www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: ovms-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openvino-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "ModelServer",
                "metadata": {
                  "name": "model-server-sample"
                },
                "spec": {
                  "image_name": "registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0",
                  "deployment_parameters": {
                    "replicas": 1
                  },
                  "service_parameters": {
                    "grpc_port": 8080,
                    "rest_port": 8081
                  },
                  "models_settings": {
                    "single_model_mode": true,
                    "config_configmap_name": "",
                    "config_path": "",
                    "model_name": "resnet",
                    "model_path": "gs://<bucket_name>/<model_dir>",
                    "nireq": 0,
                    "plugin_config": "{\"PERFORMANCE_HINT\":\"LATENCY\"}",
                    "batch_size": "",
                    "shape": "",
                    "model_version_policy": "{\"latest\": { \"num_versions\":1 }}",
                    "layout": "",
                    "target_device": "CPU",
                    "is_stateful": false,
                    "idle_sequence_cleanup": false,
                    "low_latency_transformation": true,
                    "max_sequence_number": 0
                  },
                  "server_settings": {
                    "file_system_poll_wait_seconds": 0,
                    "sequence_cleaner_poll_wait_minutes": 0,
                    "log_level": "INFO",
                    "grpc_workers": 1,
                    "rest_workers": 0
                  },
                  "models_repository": {
                    "https_proxy": "",
                    "http_proxy": "",
                    "storage_type": "google",
                    "models_host_path": "",
                    "models_volume_claim": "",
                    "aws_secret_access_key": "",
                    "aws_access_key_id": "",
                    "aws_region": "",
                    "s3_compat_api_endpoint": "",
                    "gcp_creds_secret_name": "",
                    "azure_storage_connection_string": ""
                  }
                }
              },
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "Notebook",
                "metadata": {
                  "name": "v2024.1"
                },
                "spec": {
                  "git_ref": "2024.1",
                  "auto_update_image": false
                }
              }
            ]
          alm-examples-metadata: '{ "model-server-sample":{"description":"Example
            of the model server configuration"}, "v2024":{"description":"Default notebook
            configuration"} }'
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
          createdAt: "2024-05-14T09:29:35Z"
          description: OpenVINO Toolkit Operator manages OpenVINO components in OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.34.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Provides inference serving through gRPC & REST
            displayName: Model Server
            kind: ModelServer
            name: modelservers.intel.com
            version: v1alpha1
          - description: Adds a collection of ready-to-run Jupyter notebooks for learning
              and experimenting with the OpenVINO™ Toolkit to Red Hat® OpenShift®
              Data Science
            displayName: Notebook
            kind: Notebook
            name: notebooks.intel.com
            version: v1alpha1
        description: |
          OpenVINO Toolkit Operator manages OpenVINO components in OpenShift.

          Currently there available components are ModelServer and Notebook.

          # Model Server
          [OpenVINO™ Model Server](https://github.com/openvinotoolkit/model_server) (OVMS) is a scalable, high-performance solution for serving machine learning models optimized for Intel® architectures. The server provides an inference service via gRPC or REST API for any models trained in a framework that is supported by [OpenVINO](https://software.intel.com/en-us/openvino-toolki://docs.openvino.ai/latest/index.html).
          Model Server configuration parameters is explained [here](https://github.com/openvinotoolkit/operator/blob/main/docs/modelserver_params.md).
          ## Using the cluster
          OpenVINO Model Server can be consumed as a `Service`. It is called like with `ModelServer` resource with `-ovms` suffix.
          The suffix is omitted when `ovms` phrase is included in the name.
          The service exposes gRPC and REST API interfaces to run inference requests.
          ```
          oc get pods
          NAME                                        READY   STATUS    RESTARTS   AGE
          model-server-sample-ovms-586f6f76df-dpps4   1/1     Running   0          8h

          oc get services
          NAME                       TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)             AGE
          model-server-sample-ovms   ClusterIP   172.25.199.210   <none>        8080/TCP,8081/TCP   8h
          ```

          # Notebook
          The Notebook resource integrates JupyterHub from OpenShift Data Science or Open Data Hub with a container image that includes [developer
          tools](https://github.com/openvinotoolkit/openvino/blob/master/docs/install_guides/pypi-openvino-dev.md)
          from the OpenVINO toolkit and a set of Jupyter notebook tutorials. It enables selecting a defined image `OpenVINO™ Toolkit` from
          the Jupyter Spawner choice list.

          Create the `Notebook` resource in the same project with JupyterHub and RedHat OpenShift Data Science operator as seen in [documentation](https://github.com/openvinotoolkit/operator/blob/main/docs/notebook_in_rhods.md).
          It builds the image in the cluster based on Dockerfile from [openvino_notebooks](https://github.com/openvinotoolkit/openvino_notebooks).

          ## References
          OpenVINO Model Server on [Github](https://github.com/openvinotoolkit/model_server)

          OpenVINO Model Server Operator on [Github](https://github.com/openvinotoolkit/operator)
        displayName: OpenVINO Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Inference OpenVINO
        links:
        - name: OpenVINO Toolkit Operator
          url: https://github.com/openvinotoolkit/operator
        - name: OpenVINO Model Server
          url: https://github.com/openvinotoolkit/model_server
        - name: OpenVINO Notebooks
          url: https://github.com/openvinotoolkit/openvino_notebooks
        maintainers:
        - email: openvino-operator-support@intel.com
          name: Operator Support
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5714b93ac39277affdc10c3ba142d8c032ed0ddb2ba61c04cbddac710464c4
        - registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
        - registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0
        version: 1.2.0
      entries:
      - name: openvino-operator.v1.2.0
        version: 1.2.0
      - name: openvino-operator.v1.1.0
        version: 1.1.0
      - name: openvino-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: ovms-operator
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx-openshift-operator.v0.0.33
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
          createdAt: "2019-07-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx-operator-rhmp/support?utm_source=openshift_console
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX on JBoss EAP 7
            displayName: Joget DX on JBoss EAP 7
            kind: JogetDX
            name: jogetdx.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX7/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
        - registry.connect.redhat.com/joget/joget-dx7-eap7@sha256:7245501f5fccf011d75c034b7d55542cfe3bd2c64cc7e48b3224a48b623af282
        version: 0.0.33
      entries:
      - name: joget-dx-openshift-operator.v0.0.33
        version: 0.0.33
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx-operator-rhmp
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-codeartifact-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-codeartifact-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/codeartifact-controller:1.1.1
          createdAt: "2025-09-19T22:00:24Z"
          description: AWS codeartifact controller is a service controller for managing
            codeartifact resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Domain represents the state of an AWS codeartifact Domain
              resource.
            displayName: Domain
            kind: Domain
            name: domains.codeartifact.services.k8s.aws
            version: v1alpha1
          - description: PackageGroup represents the state of an AWS codeartifact
              PackageGroup resource.
            displayName: PackageGroup
            kind: PackageGroup
            name: packagegroups.codeartifact.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon codeartifact resources in AWS from within your Kubernetes cluster.

          **About Amazon codeartifact**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon codeartifact
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - codeartifact
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon codeartifact Developer Resources
          url: https://aws.amazon.com/codeartifact/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: codeartifact maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/codeartifact-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-codeartifact-controller.v1.1.1
        version: 1.1.1
      - name: ack-codeartifact-controller.v1.1.0
        version: 1.1.0
      - name: ack-codeartifact-controller.v1.0.13
        version: 1.0.13
      - name: ack-codeartifact-controller.v1.0.11
        version: 1.0.11
      name: alpha
    defaultChannel: alpha
    packageName: ack-codeartifact-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kepler Operator Contributors
      provider-url: https://sustainable-computing.io/
    name: kepler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kepler-operator.v0.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "Kepler",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "kepler",
                    "app.kubernetes.io/name": "kepler",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "kepler"
                },
                "spec": {
                  "exporter": {
                    "deployment": {
                      "port": 9103
                    },
                    "redfish": null
                  }
                }
              },
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "PowerMonitor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "powermonitor",
                    "app.kubernetes.io/name": "powermonitor",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "power-monitor"
                },
                "spec": {
                  "kepler": {
                    "config": {
                      "logLevel": "info"
                    },
                    "deployment": {
                      "security": {
                        "mode": "none"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: quay.io/sustainable_computing_io/kepler-operator:0.20.0
          createdAt: "2025-07-18T06:13:30Z"
          description: Deploys and Manages Kepler on Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: |-
            [
              "keplerinternals.kepler.system.sustainable.computing.io",
              "powermonitorinternals.kepler.system.sustainable.computing.io"
            ]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/sustainable-computing-io/kepler-operator
          support: https://github.com/sustainable-computing-io/kepler-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KeplerInternal is the Schema for internal/unsupported API
            displayName: KeplerInternal
            kind: KeplerInternal
            name: keplerinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: Kepler is deprecated and will be removed in a future release.
              Please use PowerMonitor instead.
            displayName: Kepler
            kind: Kepler
            name: keplers.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitorInternal is the Schema for internal/unsupported
              API
            displayName: PowerMonitorInternal
            kind: PowerMonitorInternal
            name: powermonitorinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitor is the Schema for the powermonitors API
            displayName: PowerMonitor
            kind: PowerMonitor
            name: powermonitors.kepler.system.sustainable.computing.io
            version: v1alpha1
        description: |-
          Kepler (Kubernetes-based Efficient Power Level Exporter) is a Prometheus exporter that
          measures energy consumption metrics at the container, pod, and node level in Kubernetes clusters.

          ### Deprecation Notice
          **Important:** The Kepler CR (Custom Resource) is deprecated and will be removed in a future release. Please use the new PowerMonitor instead.

          ### Prerequisites
          To view the metrics exposed by Kepler, **[User Workload Monitoring](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/monitoring/configuring-user-workload-monitoring)**
          MUST be enabled in your cluster.

          ### Documentation
          * [User Documentation](https://github.com/sustainable-computing-io/kepler-operator/blob/v1alpha1/README.md)
          * [Developer Guide](https://github.com/sustainable-computing-io/kepler-operator/tree/v1alpha1/docs/developer)

          ### Contributing
          * [Report Bugs](https://github.com/sustainable-computing-io/kepler-operator/issues/new?template=bug_report.md) - Found a bug? Let us know!
          * [Browse Issues](https://github.com/sustainable-computing-io/kepler-operator/issues) - See what we're working on

          ### License
          Licensed under [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)
        displayName: Kepler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kepler
        - power
        - energy
        - monitoring
        links:
        - name: Installation Guide
          url: https://sustainable-computing.io/installation/community-operator/
        - name: Kepler Project
          url: https://sustainable-computing.io/
        maintainers:
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: vprashar@redhat.com
          name: Vibhu Prashar
        - email: vimalkum@redhat.com
          name: Vimal Kumar
        - email: kaliu@redhat.com
          name: Kaiyi Liu
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Kepler Operator Contributors
          url: https://sustainable-computing.io/
        relatedImages:
        - quay.io/sustainable_computing_io/kepler-operator:0.20.0
        - quay.io/sustainable_computing_io/kepler:v0.10.2
        - quay.io/brancz/kube-rbac-proxy:v0.19.0
        version: 0.20.0
      entries:
      - name: kepler-operator.v0.20.0
        version: 0.20.0
      - name: kepler-operator.v0.19.0
        version: 0.19.0
      - name: kepler-operator.v0.18.0
        version: 0.18.0
      - name: kepler-operator.v0.17.0
        version: 0.17.0
      - name: kepler-operator.v0.16.0
        version: 0.16.0
      - name: kepler-operator.v0.15.0
        version: 0.15.0
      name: alpha
    defaultChannel: alpha
    packageName: kepler-operator
    provider:
      name: Kepler Operator Contributors
      url: https://sustainable-computing.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MuleSoft
      provider-url: anypoint.mulesoft.com
    name: runtime-fabric-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runtime-fabric-operator.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rtf.operators.mulesoft.com/v1",
                "kind": "RuntimeFabric",
                "metadata": {
                  "name": "runtime-fabric"
                },
                "spec": {
                  "activationData": null,
                  "agent": {
                    "replicas": 1,
                    "app": {
                      "resources": {
                        "limits": {
                          "cpu": "1000m",
                          "memory": "500Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      }
                    },
                    "rtfd": {
                      "resources": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "150m",
                          "memory": "200Mi"
                        }
                      },
                      "watch": {
                        "enabled": true,
                        "components": {
                          "deployments": {
                            "enabled": true
                          },
                          "ingress": {
                            "enabled": true
                          },
                          "secrets": {
                            "enabled": true
                          },
                          "persistencegateways": {
                            "enabled": true
                          },
                          "nodes": {
                            "enabled": true
                          },
                          "jobs": {
                            "enabled": true
                          },
                          "routes": {
                            "enabled": true
                          }
                        }
                      }
                    }
                  },
                  "resource-cache": {
                    "replicas": 1,
                    "nginx": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "250Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    },
                    "fetcher": {
                      "resources": {
                        "limits": {
                          "cpu": "50m",
                          "memory": "100Mi"
                        },
                        "requests": {
                          "cpu": "50m",
                          "memory": "50Mi"
                        }
                      }
                    }
                  },
                  "mule-clusterip-service": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "global": {
                    "cluster": {
                      "flavor": "byok",
                      "type": "openshift"
                    },
                    "containerLogPaths": [
                      "/var/lib/docker/containers",
                      "/var/log/containers",
                      "/var/log/pods"
                    ],
                    "core": {
                      "affinity": {},
                      "tolerations": [],
                      "topologySpreadConstraints": [
                        {
                          "maxSkew": 1,
                          "topologyKey": "topology.kubernetes.io/zone",
                          "whenUnsatisfiable": "ScheduleAnyway"
                        }
                      ]
                    },
                    "worker": {
                      "affinity": {},
                      "tolerations": []
                    },
                    "deployment": {
                      "nodeSelector": ""
                    },
                    "image": {
                      "provider": "aws",
                      "pullSecretName": "rtf-pull-secret",
                      "rtfRegistry": "rtf-runtime-registry.kprod.msap.io",
                      "repositoryNamespace": "mulesoft"
                    },
                    "service": {
                      "clusterIp": null
                    },
                    "authorizedNamespaces": false,
                    "fipsEnabled": false,
                    "crds": {
                      "install": true,
                      "keep": true,
                      "flex": {
                        "enabled": false
                      }
                    },
                    "nodeWatcherEnabled": true,
                    "nodeReconciliation": true,
                    "deploymentRateLimitPerSecond": 1
                  },
                  "muleLicense": null,
                  "proxy": {
                    "http_no_proxy": null,
                    "http_proxy": null,
                    "monitoring_proxy": null
                  },
                  "customLog4jEnabled": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
          createdAt: "2025-07-31T21:12:41Z"
          description: Runtime Fabric Operator Application from Mulesoft
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mulesoft/rtf-ocp-operator
          support: MuleSoft
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Runtime Fabric Operator from Mulesoft
              **Activation data**
              Used to install Runtime Fabric on a Kubernetes service. Copy this value from the Runtime Fabric setup page in Runtime Manager.
              **Registry URL**
              The registry from which to retrieve the images needed to install and run Runtime Fabric. If you're using a local registry, specify the URL here.
                * Default registry URLs:
                    - US control plane: rtf-runtime-registry.kprod.msap.io
                    - EU control plane: rtf-runtime-registry.kprod-eu.msap.io
              **Registry pull secret**
              Used to retrieve images from the registry. Change this value only if you did so during the operator installation.
              **Mule license key**
              Your company’s Mule license key. The key is not required during install, but it is required to deploy applications to Runtime Fabric. Must be Base64 encoded.
            displayName: Runtime Fabric
            kind: RuntimeFabric
            name: runtimefabrics.rtf.operators.mulesoft.com
            version: v1
        description: |-
          Anypoint Runtime Fabric is a container service that you install in your existing Red Hat OpenShift cluster that enables you to run Mule applications, composite APIs, and API gateways across any environment with centralized management via Anypoint Platform.
          Capabilities of Runtime Fabric include:
          * Isolation between applications by running a separate Mule runtime server per application.
          * Ability to run multiple versions of Mule runtime server on the same set of resources.
          * Scaling applications across multiple replicas.
          * Automated application fail-over.
          * Application management with Anypoint Runtime Manager.
          Runtime Fabric supports the following Red Hat OpenShift deployment options:
          * Red Hat OpenShift Service on AWS
          * Microsoft Azure Red Hat OpenShift
          * Red Hat OpenShift Dedicated
          * Red Hat OpenShift on IBM Cloud
          * Self-managed Red Hat OpenShift editions (Performance Plus, OCP, Kubernetes engine)
          Runtime Fabric supports Red Hat OpenShift versions 4.16 to 4.19.
        displayName: Runtime Fabric Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - agent
        - rtf
        - runtime
        - fabric
        - mulesoft
        links:
        - name: Runtime Fabric Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/62901512fefb1b137ba23946
        - name: Documentation
          url: https://docs.mulesoft.com/release-notes/runtime-fabric/runtime-fabric
        maintainers:
        - email: support@mulesoft.com
          name: Mulesoft
        maturity: stable
        provider:
          name: MuleSoft
          url: anypoint.mulesoft.com
        relatedImages:
        - quay.io/rtf/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
        version: 2.11.0
      entries:
      - name: runtime-fabric-operator.v2.11.0
        version: 2.11.0
      - name: runtime-fabric-operator.v2.10.34
        version: 2.10.34
      - name: runtime-fabric-operator.v2.10.17
        version: 2.10.17
      - name: runtime-fabric-operator.v2.10.0
        version: 2.10.0
      - name: runtime-fabric-operator.v2.9.2
        version: 2.9.2
      - name: runtime-fabric-operator.v2.8.2
        version: 2.8.2
      - name: runtime-fabric-operator.v2.7.0
        version: 2.7.0
      - name: runtime-fabric-operator.v2.6.52
        version: 2.6.52
      name: alpha
    - currentCSV: runtime-fabric-operator.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rtf.operators.mulesoft.com/v1",
                "kind": "RuntimeFabric",
                "metadata": {
                  "name": "runtime-fabric"
                },
                "spec": {
                  "activationData": null,
                  "agent": {
                    "replicas": 1,
                    "app": {
                      "resources": {
                        "limits": {
                          "cpu": "1000m",
                          "memory": "500Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      }
                    },
                    "rtfd": {
                      "resources": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "150m",
                          "memory": "200Mi"
                        }
                      },
                      "watch": {
                        "enabled": true,
                        "components": {
                          "deployments": {
                            "enabled": true
                          },
                          "ingress": {
                            "enabled": true
                          },
                          "secrets": {
                            "enabled": true
                          },
                          "persistencegateways": {
                            "enabled": true
                          },
                          "nodes": {
                            "enabled": true
                          },
                          "jobs": {
                            "enabled": true
                          },
                          "routes": {
                            "enabled": true
                          }
                        }
                      }
                    }
                  },
                  "resource-cache": {
                    "replicas": 1,
                    "nginx": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "250Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    },
                    "fetcher": {
                      "resources": {
                        "limits": {
                          "cpu": "50m",
                          "memory": "100Mi"
                        },
                        "requests": {
                          "cpu": "50m",
                          "memory": "50Mi"
                        }
                      }
                    }
                  },
                  "mule-clusterip-service": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "global": {
                    "cluster": {
                      "flavor": "byok",
                      "type": "openshift"
                    },
                    "containerLogPaths": [
                      "/var/lib/docker/containers",
                      "/var/log/containers",
                      "/var/log/pods"
                    ],
                    "core": {
                      "affinity": {},
                      "tolerations": [],
                      "topologySpreadConstraints": [
                        {
                          "maxSkew": 1,
                          "topologyKey": "topology.kubernetes.io/zone",
                          "whenUnsatisfiable": "ScheduleAnyway"
                        }
                      ]
                    },
                    "worker": {
                      "affinity": {},
                      "tolerations": []
                    },
                    "deployment": {
                      "nodeSelector": ""
                    },
                    "image": {
                      "provider": "aws",
                      "pullSecretName": "rtf-pull-secret",
                      "rtfRegistry": "rtf-runtime-registry.kprod.msap.io",
                      "repositoryNamespace": "mulesoft"
                    },
                    "service": {
                      "clusterIp": null
                    },
                    "authorizedNamespaces": false,
                    "fipsEnabled": false,
                    "crds": {
                      "install": true,
                      "keep": true,
                      "flex": {
                        "enabled": false
                      }
                    },
                    "nodeWatcherEnabled": true,
                    "nodeReconciliation": true,
                    "deploymentRateLimitPerSecond": 1
                  },
                  "muleLicense": null,
                  "proxy": {
                    "http_no_proxy": null,
                    "http_proxy": null,
                    "monitoring_proxy": null
                  },
                  "customLog4jEnabled": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
          createdAt: "2025-07-31T21:12:41Z"
          description: Runtime Fabric Operator Application from Mulesoft
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mulesoft/rtf-ocp-operator
          support: MuleSoft
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Runtime Fabric Operator from Mulesoft
              **Activation data**
              Used to install Runtime Fabric on a Kubernetes service. Copy this value from the Runtime Fabric setup page in Runtime Manager.
              **Registry URL**
              The registry from which to retrieve the images needed to install and run Runtime Fabric. If you're using a local registry, specify the URL here.
                * Default registry URLs:
                    - US control plane: rtf-runtime-registry.kprod.msap.io
                    - EU control plane: rtf-runtime-registry.kprod-eu.msap.io
              **Registry pull secret**
              Used to retrieve images from the registry. Change this value only if you did so during the operator installation.
              **Mule license key**
              Your company’s Mule license key. The key is not required during install, but it is required to deploy applications to Runtime Fabric. Must be Base64 encoded.
            displayName: Runtime Fabric
            kind: RuntimeFabric
            name: runtimefabrics.rtf.operators.mulesoft.com
            version: v1
        description: |-
          Anypoint Runtime Fabric is a container service that you install in your existing Red Hat OpenShift cluster that enables you to run Mule applications, composite APIs, and API gateways across any environment with centralized management via Anypoint Platform.
          Capabilities of Runtime Fabric include:
          * Isolation between applications by running a separate Mule runtime server per application.
          * Ability to run multiple versions of Mule runtime server on the same set of resources.
          * Scaling applications across multiple replicas.
          * Automated application fail-over.
          * Application management with Anypoint Runtime Manager.
          Runtime Fabric supports the following Red Hat OpenShift deployment options:
          * Red Hat OpenShift Service on AWS
          * Microsoft Azure Red Hat OpenShift
          * Red Hat OpenShift Dedicated
          * Red Hat OpenShift on IBM Cloud
          * Self-managed Red Hat OpenShift editions (Performance Plus, OCP, Kubernetes engine)
          Runtime Fabric supports Red Hat OpenShift versions 4.16 to 4.19.
        displayName: Runtime Fabric Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - agent
        - rtf
        - runtime
        - fabric
        - mulesoft
        links:
        - name: Runtime Fabric Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/62901512fefb1b137ba23946
        - name: Documentation
          url: https://docs.mulesoft.com/release-notes/runtime-fabric/runtime-fabric
        maintainers:
        - email: support@mulesoft.com
          name: Mulesoft
        maturity: stable
        provider:
          name: MuleSoft
          url: anypoint.mulesoft.com
        relatedImages:
        - quay.io/rtf/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
        version: 2.11.0
      entries:
      - name: runtime-fabric-operator.v2.11.0
        version: 2.11.0
      - name: runtime-fabric-operator.v2.10.34
        version: 2.10.34
      - name: runtime-fabric-operator.v2.10.17
        version: 2.10.17
      - name: runtime-fabric-operator.v2.10.0
        version: 2.10.0
      - name: runtime-fabric-operator.v2.9.2
        version: 2.9.2
      - name: runtime-fabric-operator.v2.8.2
        version: 2.8.2
      - name: runtime-fabric-operator.v2.7.0
        version: 2.7.0
      - name: runtime-fabric-operator.v2.6.52
        version: 2.6.52
      name: stable
    defaultChannel: stable
    packageName: runtime-fabric-operator
    provider:
      name: MuleSoft
      url: anypoint.mulesoft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-dynamodb-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-dynamodb-controller.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "GlobalTable",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "Table",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/dynamodb-controller:1.5.2
          createdAt: "2025-09-19T19:19:47Z"
          description: AWS DynamoDB controller is a service controller for managing
            DynamoDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup represents the state of an AWS dynamodb Backup resource.
            displayName: Backup
            kind: Backup
            name: backups.dynamodb.services.k8s.aws
            version: v1alpha1
          - description: GlobalTable represents the state of an AWS dynamodb GlobalTable
              resource.
            displayName: GlobalTable
            kind: GlobalTable
            name: globaltables.dynamodb.services.k8s.aws
            version: v1alpha1
          - description: Table represents the state of an AWS dynamodb Table resource.
            displayName: Table
            kind: Table
            name: tables.dynamodb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon DynamoDB resources in AWS from within your Kubernetes cluster.

          **About Amazon DynamoDB**

          Amazon DynamoDB is a fully managed NoSQL database service that provides fast and predictable performance with seamless scalability. DynamoDB lets you offload the administrative burdens of operating and scaling a distributed database so that you don't have to worry about hardware provisioning, setup and configuration, replication, software patching, or cluster scaling. DynamoDB also offers encryption at rest, which eliminates the operational burden and complexity involved in protecting sensitive data. For more information, see [DynamoDB Encryption at Rest](https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/EncryptionAtRest.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon DynamoDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dynamodb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon DynamoDB Developer Resources
          url: https://aws.amazon.com/dynamodb/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: dynamodb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/dynamodb-controller:1.5.2
        version: 1.5.2
      entries:
      - name: ack-dynamodb-controller.v1.5.2
        version: 1.5.2
      - name: ack-dynamodb-controller.v1.5.1
        version: 1.5.1
      - name: ack-dynamodb-controller.v1.5.0
        version: 1.5.0
      - name: ack-dynamodb-controller.v1.4.10
        version: 1.4.10
      - name: ack-dynamodb-controller.v1.4.9
        version: 1.4.9
      - name: ack-dynamodb-controller.v1.4.7
        version: 1.4.7
      - name: ack-dynamodb-controller.v1.4.6
        version: 1.4.6
      - name: ack-dynamodb-controller.v1.4.5
        version: 1.4.5
      - name: ack-dynamodb-controller.v1.4.4
        version: 1.4.4
      - name: ack-dynamodb-controller.v1.4.3
        version: 1.4.3
      - name: ack-dynamodb-controller.v1.4.2
        version: 1.4.2
      - name: ack-dynamodb-controller.v1.4.1
        version: 1.4.1
      - name: ack-dynamodb-controller.v1.3.0
        version: 1.3.0
      - name: ack-dynamodb-controller.v1.2.23
        version: 1.2.23
      - name: ack-dynamodb-controller.v1.2.22
        version: 1.2.22
      - name: ack-dynamodb-controller.v1.2.21
        version: 1.2.21
      - name: ack-dynamodb-controller.v1.2.20
        version: 1.2.20
      - name: ack-dynamodb-controller.v1.2.15
        version: 1.2.15
      - name: ack-dynamodb-controller.v1.2.14
        version: 1.2.14
      - name: ack-dynamodb-controller.v1.2.13
        version: 1.2.13
      - name: ack-dynamodb-controller.v1.2.12
        version: 1.2.12
      - name: ack-dynamodb-controller.v1.2.11
        version: 1.2.11
      - name: ack-dynamodb-controller.v1.2.9
        version: 1.2.9
      - name: ack-dynamodb-controller.v1.2.8
        version: 1.2.8
      - name: ack-dynamodb-controller.v1.2.6
        version: 1.2.6
      - name: ack-dynamodb-controller.v1.2.5
        version: 1.2.5
      - name: ack-dynamodb-controller.v1.2.4
        version: 1.2.4
      - name: ack-dynamodb-controller.v1.2.2
        version: 1.2.2
      - name: ack-dynamodb-controller.v1.2.1
        version: 1.2.1
      - name: ack-dynamodb-controller.v1.2.0
        version: 1.2.0
      - name: ack-dynamodb-controller.v1.1.1
        version: 1.1.1
      - name: ack-dynamodb-controller.v1.1.0
        version: 1.1.0
      - name: ack-dynamodb-controller.v1.0.3
        version: 1.0.3
      - name: ack-dynamodb-controller.v1.0.2
        version: 1.0.2
      - name: ack-dynamodb-controller.v1.0.1
        version: 1.0.1
      - name: ack-dynamodb-controller.v1.0.0
        version: 1.0.0
      - name: ack-dynamodb-controller.v0.1.10
        version: 0.1.10
      - name: ack-dynamodb-controller.v0.1.9
        version: 0.1.9
      - name: ack-dynamodb-controller.v0.1.8
        version: 0.1.8
      - name: ack-dynamodb-controller.v0.1.7
        version: 0.1.7
      - name: ack-dynamodb-controller.v0.1.6
        version: 0.1.6
      - name: ack-dynamodb-controller.v0.1.5
        version: 0.1.5
      - name: ack-dynamodb-controller.v0.1.4
        version: 0.1.4
      - name: ack-dynamodb-controller.v0.1.3
        version: 0.1.3
      - name: ack-dynamodb-controller.v0.1.1
        version: 0.1.1
      - name: ack-dynamodb-controller.v0.1.0
        version: 0.1.0
      - name: ack-dynamodb-controller.v0.0.20
        version: 0.0.20
      - name: ack-dynamodb-controller.v0.0.19
        version: 0.0.19
      - name: ack-dynamodb-controller.v0.0.18
        version: 0.0.18
      - name: ack-dynamodb-controller.v0.0.17
        version: 0.0.17
      - name: ack-dynamodb-controller.v0.0.16
        version: 0.0.16
      - name: ack-dynamodb-controller.v0.0.15
        version: 0.0.15
      - name: ack-dynamodb-controller.v0.0.14
        version: 0.0.14
      - name: ack-dynamodb-controller.v0.0.13
        version: 0.0.13
      - name: ack-dynamodb-controller.v0.0.12
        version: 0.0.12
      - name: ack-dynamodb-controller.v0.0.11
        version: 0.0.11
      - name: ack-dynamodb-controller.v0.0.10
        version: 0.0.10
      - name: ack-dynamodb-controller.v0.0.9
        version: 0.0.9
      - name: ack-dynamodb-controller.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: ack-dynamodb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-elasticache-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-elasticache-controller.v1.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "CacheParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "CacheSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "ReplicationGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "Snapshot",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "UserGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/elasticache-controller:1.2.3
          createdAt: "2025-09-19T19:19:33Z"
          description: AWS ElastiCache controller is a service controller for managing
            ElastiCache resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CacheCluster represents the state of an AWS elasticache CacheCluster
              resource.
            displayName: CacheCluster
            kind: CacheCluster
            name: cacheclusters.elasticache.services.k8s.aws
            version: v1alpha1
          - description: CacheParameterGroup represents the state of an AWS elasticache
              CacheParameterGroup resource.
            displayName: CacheParameterGroup
            kind: CacheParameterGroup
            name: cacheparametergroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: CacheSubnetGroup represents the state of an AWS elasticache
              CacheSubnetGroup resource.
            displayName: CacheSubnetGroup
            kind: CacheSubnetGroup
            name: cachesubnetgroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ReplicationGroup represents the state of an AWS elasticache
              ReplicationGroup resource.
            displayName: ReplicationGroup
            kind: ReplicationGroup
            name: replicationgroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCache represents the state of an AWS elasticache
              ServerlessCache resource.
            displayName: ServerlessCache
            kind: ServerlessCache
            name: serverlesscaches.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCacheSnapshot represents the state of an AWS elasticache
              ServerlessCacheSnapshot resource.
            displayName: ServerlessCacheSnapshot
            kind: ServerlessCacheSnapshot
            name: serverlesscachesnapshots.elasticache.services.k8s.aws
            version: v1alpha1
          - description: Snapshot represents the state of an AWS elasticache Snapshot
              resource.
            displayName: Snapshot
            kind: Snapshot
            name: snapshots.elasticache.services.k8s.aws
            version: v1alpha1
          - description: UserGroup represents the state of an AWS elasticache UserGroup
              resource.
            displayName: UserGroup
            kind: UserGroup
            name: usergroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS elasticache User resource.
            displayName: User
            kind: User
            name: users.elasticache.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ElastiCache resources in AWS from within your Kubernetes cluster.

          **About Amazon ElastiCache**

          Amazon ElastiCache allows you to seamlessly set up, run, and scale popular open-source compatible in-memory data stores in the cloud. Build data-intensive apps or boost the performance of your existing databases by retrieving data from high throughput and low latency in-memory data stores. Amazon ElastiCache is a popular choice for real-time use cases like Caching, Session Stores, Gaming, Geospatial Services, Real-Time Analytics, and Queuing.

          Amazon ElastiCache offers fully managed Redis, voted the most loved database by developers in the Stack Overflow 2020 Developer Survey, and Memcached for your most demanding applications that require sub-millisecond response times.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ElastiCache
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticache
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ElastiCache Developer Resources
          url: https://aws.amazon.com/elasticache/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: elasticache maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/elasticache-controller:1.2.3
        version: 1.2.3
      entries:
      - name: ack-elasticache-controller.v1.2.3
        version: 1.2.3
      - name: ack-elasticache-controller.v1.2.2
        version: 1.2.2
      - name: ack-elasticache-controller.v1.2.1
        version: 1.2.1
      - name: ack-elasticache-controller.v1.2.0
        version: 1.2.0
      - name: ack-elasticache-controller.v1.1.0
        version: 1.1.0
      - name: ack-elasticache-controller.v1.0.7
        version: 1.0.7
      - name: ack-elasticache-controller.v1.0.6
        version: 1.0.6
      - name: ack-elasticache-controller.v1.0.5
        version: 1.0.5
      - name: ack-elasticache-controller.v1.0.4
        version: 1.0.4
      - name: ack-elasticache-controller.v1.0.3
        version: 1.0.3
      - name: ack-elasticache-controller.v1.0.2
        version: 1.0.2
      - name: ack-elasticache-controller.v1.0.1
        version: 1.0.1
      - name: ack-elasticache-controller.v1.0.0
        version: 1.0.0
      - name: ack-elasticache-controller.v0.2.6
        version: 0.2.6
      - name: ack-elasticache-controller.v0.2.5
        version: 0.2.5
      - name: ack-elasticache-controller.v0.2.4
        version: 0.2.4
      - name: ack-elasticache-controller.v0.2.3
        version: 0.2.3
      - name: ack-elasticache-controller.v0.2.2
        version: 0.2.2
      - name: ack-elasticache-controller.v0.2.1
        version: 0.2.1
      - name: ack-elasticache-controller.v0.2.0
        version: 0.2.0
      - name: ack-elasticache-controller.v0.1.1
        version: 0.1.1
      - name: ack-elasticache-controller.v0.1.0
        version: 0.1.0
      - name: ack-elasticache-controller.v0.0.31
        version: 0.0.31
      - name: ack-elasticache-controller.v0.0.30
        version: 0.0.30
      - name: ack-elasticache-controller.v0.0.29
        version: 0.0.29
      - name: ack-elasticache-controller.v0.0.28
        version: 0.0.28
      - name: ack-elasticache-controller.v0.0.27
        version: 0.0.27
      - name: ack-elasticache-controller.v0.0.26
        version: 0.0.26
      - name: ack-elasticache-controller.v0.0.25
        version: 0.0.25
      - name: ack-elasticache-controller.v0.0.24
        version: 0.0.24
      - name: ack-elasticache-controller.v0.0.23
        version: 0.0.23
      - name: ack-elasticache-controller.v0.0.22
        version: 0.0.22
      - name: ack-elasticache-controller.v0.0.21
        version: 0.0.21
      - name: ack-elasticache-controller.v0.0.20
        version: 0.0.20
      - name: ack-elasticache-controller.v0.0.19
        version: 0.0.19
      - name: ack-elasticache-controller.v0.0.18
        version: 0.0.18
      - name: ack-elasticache-controller.v0.0.17
        version: 0.0.17
      - name: ack-elasticache-controller.v0.0.16
        version: 0.0.16
      - name: ack-elasticache-controller.v0.0.15
        version: 0.0.15
      - name: ack-elasticache-controller.v0.0.14
        version: 0.0.14
      - name: ack-elasticache-controller.v0.0.13
        version: 0.0.13
      - name: ack-elasticache-controller.v0.0.12
        version: 0.0.12
      - name: ack-elasticache-controller.v0.0.11
        version: 0.0.11
      - name: ack-elasticache-controller.v0.0.10
        version: 0.0.10
      - name: ack-elasticache-controller.v0.0.9
        version: 0.0.9
      - name: ack-elasticache-controller.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: ack-elasticache-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ember-csi-community-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ember-csi-community-operator.v0.9.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ember-csi.io/v1alpha1",
                "kind": "EmberStorageBackend",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "config": {
                    "envVars": {
                      "X_CSI_PERSISTENCE_CONFIG": {
                        "storage": "crd"
                      },
                      "X_CSI_DEBUG_MODE": "",
                      "X_CSI_DEFAULT_MOUNT_FS": "ext4",
                      "X_CSI_EMBER_CONFIG": {
                        "grpc_workers": 30,
                        "disable_logs": false,
                        "plugin_name": "",
                        "slow_operations": true,
                        "disabled": [],
                        "enable_probe": false,
                        "debug": false,
                        "user_id": "ember_csi.io",
                        "project_id": "ember_csi.io"
                      },
                      "X_CSI_BACKEND_CONFIG": {
                        "enable_unsupported_driver": false,
                        "driver": "ACCESSIscsi",
                        "name": "",
                        "driver__ACCESSIscsi__vrts_lun_sparse": true,
                        "driver__ACCESSIscsi__vrts_target_config": "/etc/cinder/vrts_target.xml",
                        "driver__AS13000__as13000_ipsan_pools__transform_csv": "Pool0",
                        "driver__AS13000__as13000_meta_pool": "",
                        "driver__AS13000__as13000_token_available_time": 3300,
                        "driver__Acs5000ISCSI__acs5000_copy_interval": 5,
                        "driver__Acs5000ISCSI__acs5000_target": 0,
                        "driver__Acs5000ISCSI__acs5000_volpool_name__transform_csv": "pool01",
                        "driver__DPLFC__dpl_pool": "",
                        "driver__DPLFC__dpl_port": 8357,
                        "driver__DPLISCSI__dpl_pool": "",
                        "driver__DPLISCSI__dpl_port": 8357,
                        "driver__DSWARE__dsware_rest_url": "",
                        "driver__DSWARE__dsware_storage_pools": "",
                        "driver__DSWARE__manager_ips__transform_csv_kvs": "",
                        "driver__Datera__backend_availability_zone": "",
                        "driver__Datera__chap_password": "",
                        "driver__Datera__chap_username": "",
                        "driver__Datera__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__Datera__driver_client_cert": "",
                        "driver__Datera__driver_client_cert_key": "",
                        "driver__Datera__driver_data_namespace": "",
                        "driver__Datera__driver_ssl_cert_path": "",
                        "driver__Datera__driver_ssl_cert_verify": false,
                        "driver__Datera__driver_use_ssl": false,
                        "driver__Datera__filter_function": "",
                        "driver__Datera__goodness_function": "",
                        "driver__Datera__iscsi_iotype": "fileio",
                        "driver__Datera__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__Datera__iscsi_target_flags": "",
                        "driver__Datera__iscsi_write_cache": "on",
                        "driver__Datera__num_shell_tries": 3,
                        "driver__Datera__num_volume_device_scan_tries": 3,
                        "driver__Datera__report_discard_supported": false,
                        "driver__Datera__storage_protocol": "iscsi",
                        "driver__Datera__target_helper": "tgtadm",
                        "driver__Datera__target_ip_address": "$my_ip",
                        "driver__Datera__target_port": 3260,
                        "driver__Datera__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__Datera__target_protocol": "iscsi",
                        "driver__Datera__trace_flags__transform_csv": "",
                        "driver__Datera__use_chap_auth": false,
                        "driver__Datera__volume_backend_name": "",
                        "driver__Datera__volume_clear": "zero",
                        "driver__Datera__volume_clear_ionice": "",
                        "driver__Datera__volume_clear_size": 0,
                        "driver__Datera__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__Datera__volume_copy_bps_limit": 0,
                        "driver__Datera__volume_dd_blocksize": "1M",
                        "driver__Datera__volumes_dir": "$state_path/volumes",
                        "driver__FJDXFC__cinder_eternus_config_file": "/etc/cinder/cinder_fujitsu_eternus_dx.xml",
                        "driver__FJDXISCSI__cinder_eternus_config_file": "/etc/cinder/cinder_fujitsu_eternus_dx.xml",
                        "driver__FlashSystemFC__flashsystem_connection_protocol": "FC",
                        "driver__FlashSystemFC__flashsystem_multihostmap_enabled": true,
                        "driver__FlashSystemISCSI__flashsystem_connection_protocol": "FC",
                        "driver__FlashSystemISCSI__flashsystem_multihostmap_enabled": true,
                        "driver__GPFS__gpfs_images_dir": "",
                        "driver__GPFS__gpfs_images_share_mode__transform_empty_none": "",
                        "driver__GPFS__gpfs_max_clone_depth": 0,
                        "driver__GPFS__gpfs_mount_point_base": "",
                        "driver__GPFS__gpfs_sparse_volumes": true,
                        "driver__GPFS__gpfs_storage_pool": "system",
                        "driver__GPFSRemote__gpfs_hosts__transform_csv": "",
                        "driver__GPFSRemote__gpfs_hosts_key_file": "$state_path/ssh_known_hosts",
                        "driver__GPFSRemote__gpfs_images_dir": "",
                        "driver__GPFSRemote__gpfs_images_share_mode__transform_empty_none": "",
                        "driver__GPFSRemote__gpfs_max_clone_depth": 0,
                        "driver__GPFSRemote__gpfs_mount_point_base": "",
                        "driver__GPFSRemote__gpfs_private_key": "",
                        "driver__GPFSRemote__gpfs_sparse_volumes": true,
                        "driver__GPFSRemote__gpfs_ssh_port": 22,
                        "driver__GPFSRemote__gpfs_storage_pool": "system",
                        "driver__GPFSRemote__gpfs_strict_host_key_policy": false,
                        "driver__GPFSRemote__gpfs_user_login": "root",
                        "driver__GPFSRemote__gpfs_user_password": "",
                        "driver__HBSDFC__backend_availability_zone": "",
                        "driver__HBSDFC__chap_password": "",
                        "driver__HBSDFC__chap_username": "",
                        "driver__HBSDFC__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__HBSDFC__driver_client_cert": "",
                        "driver__HBSDFC__driver_client_cert_key": "",
                        "driver__HBSDFC__driver_data_namespace": "",
                        "driver__HBSDFC__driver_ssl_cert_path": "",
                        "driver__HBSDFC__driver_ssl_cert_verify": false,
                        "driver__HBSDFC__driver_use_ssl": false,
                        "driver__HBSDFC__filter_function": "",
                        "driver__HBSDFC__goodness_function": "",
                        "driver__HBSDFC__iscsi_iotype": "fileio",
                        "driver__HBSDFC__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__HBSDFC__iscsi_target_flags": "",
                        "driver__HBSDFC__iscsi_write_cache": "on",
                        "driver__HBSDFC__num_shell_tries": 3,
                        "driver__HBSDFC__num_volume_device_scan_tries": 3,
                        "driver__HBSDFC__report_discard_supported": false,
                        "driver__HBSDFC__storage_protocol": "iscsi",
                        "driver__HBSDFC__target_helper": "tgtadm",
                        "driver__HBSDFC__target_ip_address": "$my_ip",
                        "driver__HBSDFC__target_port": 3260,
                        "driver__HBSDFC__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__HBSDFC__target_protocol": "iscsi",
                        "driver__HBSDFC__trace_flags__transform_csv": "",
                        "driver__HBSDFC__use_chap_auth": false,
                        "driver__HBSDFC__volume_backend_name": "",
                        "driver__HBSDFC__volume_clear": "zero",
                        "driver__HBSDFC__volume_clear_ionice": "",
                        "driver__HBSDFC__volume_clear_size": 0,
                        "driver__HBSDFC__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__HBSDFC__volume_copy_bps_limit": 0,
                        "driver__HBSDFC__volume_dd_blocksize": "1M",
                        "driver__HBSDFC__volumes_dir": "$state_path/volumes",
                        "driver__HBSDISCSI__backend_availability_zone": "",
                        "driver__HBSDISCSI__chap_password": "",
                        "driver__HBSDISCSI__chap_username": "",
                        "driver__HBSDISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__HBSDISCSI__driver_client_cert": "",
                        "driver__HBSDISCSI__driver_client_cert_key": "",
                        "driver__HBSDISCSI__driver_data_namespace": "",
                        "driver__HBSDISCSI__driver_ssl_cert_path": "",
                        "driver__HBSDISCSI__driver_ssl_cert_verify": false,
                        "driver__HBSDISCSI__driver_use_ssl": false,
                        "driver__HBSDISCSI__filter_function": "",
                        "driver__HBSDISCSI__goodness_function": "",
                        "driver__HBSDISCSI__iscsi_iotype": "fileio",
                        "driver__HBSDISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__HBSDISCSI__iscsi_target_flags": "",
                        "driver__HBSDISCSI__iscsi_write_cache": "on",
                        "driver__HBSDISCSI__num_shell_tries": 3,
                        "driver__HBSDISCSI__num_volume_device_scan_tries": 3,
                        "driver__HBSDISCSI__report_discard_supported": false,
                        "driver__HBSDISCSI__storage_protocol": "iscsi",
                        "driver__HBSDISCSI__target_helper": "tgtadm",
                        "driver__HBSDISCSI__target_ip_address": "$my_ip",
                        "driver__HBSDISCSI__target_port": 3260,
                        "driver__HBSDISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__HBSDISCSI__target_protocol": "iscsi",
                        "driver__HBSDISCSI__trace_flags__transform_csv": "",
                        "driver__HBSDISCSI__use_chap_auth": false,
                        "driver__HBSDISCSI__volume_backend_name": "",
                        "driver__HBSDISCSI__volume_clear": "zero",
                        "driver__HBSDISCSI__volume_clear_ionice": "",
                        "driver__HBSDISCSI__volume_clear_size": 0,
                        "driver__HBSDISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__HBSDISCSI__volume_copy_bps_limit": 0,
                        "driver__HBSDISCSI__volume_dd_blocksize": "1M",
                        "driver__HBSDISCSI__volumes_dir": "$state_path/volumes",
                        "driver__HPE3PARFC__hpe3par_api_url": "",
                        "driver__HPE3PARFC__hpe3par_cpg__transform_csv": "OpenStack",
                        "driver__HPE3PARFC__hpe3par_cpg_snap": "",
                        "driver__HPE3PARFC__hpe3par_debug": false,
                        "driver__HPE3PARFC__hpe3par_iscsi_chap_enabled": false,
                        "driver__HPE3PARFC__hpe3par_iscsi_ips__transform_csv": "",
                        "driver__HPE3PARFC__hpe3par_password": "",
                        "driver__HPE3PARFC__hpe3par_snapshot_expiration": "",
                        "driver__HPE3PARFC__hpe3par_snapshot_retention": "",
                        "driver__HPE3PARFC__hpe3par_target_nsp": "",
                        "driver__HPE3PARFC__hpe3par_username": "",
                        "driver__HPE3PARFC__san_ip": "",
                        "driver__HPE3PARFC__san_login": "admin",
                        "driver__HPE3PARFC__san_password": "",
                        "driver__HPE3PARFC__san_private_key": "",
                        "driver__HPE3PARFC__san_ssh_port": 22,
                        "driver__HPE3PARFC__ssh_conn_timeout": 30,
                        "driver__HPE3PARFC__target_ip_address": "$my_ip",
                        "driver__HPE3PARFC__target_port": 3260,
                        "driver__HPE3PARFC__unique_fqdn_network": true,
                        "driver__HPE3PARISCSI__hpe3par_api_url": "",
                        "driver__HPE3PARISCSI__hpe3par_cpg__transform_csv": "OpenStack",
                        "driver__HPE3PARISCSI__hpe3par_cpg_snap": "",
                        "driver__HPE3PARISCSI__hpe3par_debug": false,
                        "driver__HPE3PARISCSI__hpe3par_iscsi_chap_enabled": false,
                        "driver__HPE3PARISCSI__hpe3par_iscsi_ips__transform_csv": "",
                        "driver__HPE3PARISCSI__hpe3par_password": "",
                        "driver__HPE3PARISCSI__hpe3par_snapshot_expiration": "",
                        "driver__HPE3PARISCSI__hpe3par_snapshot_retention": "",
                        "driver__HPE3PARISCSI__hpe3par_target_nsp": "",
                        "driver__HPE3PARISCSI__hpe3par_username": "",
                        "driver__HPE3PARISCSI__san_ip": "",
                        "driver__HPE3PARISCSI__san_login": "admin",
                        "driver__HPE3PARISCSI__san_password": "",
                        "driver__HPE3PARISCSI__san_private_key": "",
                        "driver__HPE3PARISCSI__san_ssh_port": 22,
                        "driver__HPE3PARISCSI__ssh_conn_timeout": 30,
                        "driver__HPE3PARISCSI__target_ip_address": "$my_ip",
                        "driver__HPE3PARISCSI__target_port": 3260,
                        "driver__HPE3PARISCSI__unique_fqdn_network": true,
                        "driver__HPMSAFC__driver_ssl_cert_path": "",
                        "driver__HPMSAFC__driver_ssl_cert_verify": false,
                        "driver__HPMSAFC__driver_use_ssl": false,
                        "driver__HPMSAFC__hpmsa_pool_name": "A",
                        "driver__HPMSAFC__hpmsa_pool_type": "virtual",
                        "driver__HPMSAFC__san_ip": "",
                        "driver__HPMSAFC__san_login": "admin",
                        "driver__HPMSAFC__san_password": "",
                        "driver__HPMSAISCSI__driver_ssl_cert_path": "",
                        "driver__HPMSAISCSI__driver_ssl_cert_verify": false,
                        "driver__HPMSAISCSI__driver_use_ssl": false,
                        "driver__HPMSAISCSI__hpmsa_iscsi_ips__transform_csv": "",
                        "driver__HPMSAISCSI__hpmsa_pool_name": "A",
                        "driver__HPMSAISCSI__hpmsa_pool_type": "virtual",
                        "driver__HPMSAISCSI__san_ip": "",
                        "driver__HPMSAISCSI__san_login": "admin",
                        "driver__HPMSAISCSI__san_password": "",
                        "driver__HuaweiFC__cinder_huawei_conf_file": "/etc/cinder/cinder_huawei_conf.xml",
                        "driver__HuaweiFC__hypermetro_devices": "",
                        "driver__HuaweiFC__metro_domain_name": "",
                        "driver__HuaweiFC__metro_san_address": "",
                        "driver__HuaweiFC__metro_san_password": "",
                        "driver__HuaweiFC__metro_san_user": "",
                        "driver__HuaweiFC__metro_storage_pools": "",
                        "driver__HuaweiISCSI__cinder_huawei_conf_file": "/etc/cinder/cinder_huawei_conf.xml",
                        "driver__HuaweiISCSI__hypermetro_devices": "",
                        "driver__HuaweiISCSI__metro_domain_name": "",
                        "driver__HuaweiISCSI__metro_san_address": "",
                        "driver__HuaweiISCSI__metro_san_password": "",
                        "driver__HuaweiISCSI__metro_san_user": "",
                        "driver__HuaweiISCSI__metro_storage_pools": "",
                        "driver__IBMStorage__chap": "disabled",
                        "driver__IBMStorage__connection_type": "iscsi",
                        "driver__IBMStorage__management_ips": "",
                        "driver__IBMStorage__proxy": "cinder.volume.drivers.ibm.ibm_storage.proxy.IBMStorageProxy",
                        "driver__InStorageMCSFC__instorage_mcs_allow_tenant_qos": false,
                        "driver__InStorageMCSFC__instorage_mcs_localcopy_rate": 50,
                        "driver__InStorageMCSFC__instorage_mcs_localcopy_timeout": 120,
                        "driver__InStorageMCSFC__instorage_mcs_vol_autoexpand": true,
                        "driver__InStorageMCSFC__instorage_mcs_vol_compression": false,
                        "driver__InStorageMCSFC__instorage_mcs_vol_grainsize": 256,
                        "driver__InStorageMCSFC__instorage_mcs_vol_intier": true,
                        "driver__InStorageMCSFC__instorage_mcs_vol_iogrp": "0",
                        "driver__InStorageMCSFC__instorage_mcs_vol_rsize": 2,
                        "driver__InStorageMCSFC__instorage_mcs_vol_warning": 0,
                        "driver__InStorageMCSFC__instorage_mcs_volpool_name__transform_csv": "volpool",
                        "driver__InStorageMCSFC__instorage_san_secondary_ip": "",
                        "driver__InStorageMCSISCSI__instorage_mcs_allow_tenant_qos": false,
                        "driver__InStorageMCSISCSI__instorage_mcs_localcopy_rate": 50,
                        "driver__InStorageMCSISCSI__instorage_mcs_localcopy_timeout": 120,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_autoexpand": true,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_compression": false,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_grainsize": 256,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_intier": true,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_iogrp": "0",
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_rsize": 2,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_warning": 0,
                        "driver__InStorageMCSISCSI__instorage_mcs_volpool_name__transform_csv": "volpool",
                        "driver__InStorageMCSISCSI__instorage_san_secondary_ip": "",
                        "driver__InfiniboxVolume__chap_password": "",
                        "driver__InfiniboxVolume__chap_username": "",
                        "driver__InfiniboxVolume__infinidat_iscsi_netspaces__transform_csv": "",
                        "driver__InfiniboxVolume__infinidat_pool_name": "",
                        "driver__InfiniboxVolume__infinidat_storage_protocol": "fc",
                        "driver__InfiniboxVolume__infinidat_use_compression": false,
                        "driver__InfiniboxVolume__num_volume_device_scan_tries": 3,
                        "driver__InfiniboxVolume__san_ip": "",
                        "driver__InfiniboxVolume__san_login": "admin",
                        "driver__InfiniboxVolume__san_password": "",
                        "driver__InfiniboxVolume__san_thin_provision": true,
                        "driver__InfiniboxVolume__use_chap_auth": false,
                        "driver__InfiniboxVolume__volume_dd_blocksize": "1M",
                        "driver__InfortrendCLIFC__infortrend_cli_cache": false,
                        "driver__InfortrendCLIFC__infortrend_cli_max_retries": 5,
                        "driver__InfortrendCLIFC__infortrend_cli_path": "/opt/bin/Infortrend/raidcmd_ESDS10.jar",
                        "driver__InfortrendCLIFC__infortrend_cli_timeout": 60,
                        "driver__InfortrendCLIFC__infortrend_iqn_prefix": "iqn.2002-10.com.infortrend",
                        "driver__InfortrendCLIFC__infortrend_pools_name__transform_csv": "",
                        "driver__InfortrendCLIFC__infortrend_slots_a_channels_id__transform_csv": "",
                        "driver__InfortrendCLIFC__infortrend_slots_b_channels_id__transform_csv": "",
                        "driver__InfortrendCLIFC__java_path": "/usr/bin/java",
                        "driver__InfortrendCLIISCSI__infortrend_cli_cache": false,
                        "driver__InfortrendCLIISCSI__infortrend_cli_max_retries": 5,
                        "driver__InfortrendCLIISCSI__infortrend_cli_path": "/opt/bin/Infortrend/raidcmd_ESDS10.jar",
                        "driver__InfortrendCLIISCSI__infortrend_cli_timeout": 60,
                        "driver__InfortrendCLIISCSI__infortrend_iqn_prefix": "iqn.2002-10.com.infortrend",
                        "driver__InfortrendCLIISCSI__infortrend_pools_name__transform_csv": "",
                        "driver__InfortrendCLIISCSI__infortrend_slots_a_channels_id__transform_csv": "",
                        "driver__InfortrendCLIISCSI__infortrend_slots_b_channels_id__transform_csv": "",
                        "driver__InfortrendCLIISCSI__java_path": "/usr/bin/java",
                        "driver__JovianISCSI__backend_availability_zone": "",
                        "driver__JovianISCSI__chap_password": "",
                        "driver__JovianISCSI__chap_username": "",
                        "driver__JovianISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__JovianISCSI__driver_client_cert": "",
                        "driver__JovianISCSI__driver_client_cert_key": "",
                        "driver__JovianISCSI__driver_data_namespace": "",
                        "driver__JovianISCSI__driver_ssl_cert_path": "",
                        "driver__JovianISCSI__driver_ssl_cert_verify": false,
                        "driver__JovianISCSI__driver_use_ssl": false,
                        "driver__JovianISCSI__filter_function": "",
                        "driver__JovianISCSI__goodness_function": "",
                        "driver__JovianISCSI__iscsi_iotype": "fileio",
                        "driver__JovianISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__JovianISCSI__iscsi_target_flags": "",
                        "driver__JovianISCSI__iscsi_write_cache": "on",
                        "driver__JovianISCSI__num_shell_tries": 3,
                        "driver__JovianISCSI__num_volume_device_scan_tries": 3,
                        "driver__JovianISCSI__report_discard_supported": false,
                        "driver__JovianISCSI__storage_protocol": "iscsi",
                        "driver__JovianISCSI__target_helper": "tgtadm",
                        "driver__JovianISCSI__target_ip_address": "$my_ip",
                        "driver__JovianISCSI__target_port": 3260,
                        "driver__JovianISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__JovianISCSI__target_protocol": "iscsi",
                        "driver__JovianISCSI__trace_flags__transform_csv": "",
                        "driver__JovianISCSI__use_chap_auth": false,
                        "driver__JovianISCSI__volume_backend_name": "",
                        "driver__JovianISCSI__volume_clear": "zero",
                        "driver__JovianISCSI__volume_clear_ionice": "",
                        "driver__JovianISCSI__volume_clear_size": 0,
                        "driver__JovianISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__JovianISCSI__volume_copy_bps_limit": 0,
                        "driver__JovianISCSI__volume_dd_blocksize": "1M",
                        "driver__JovianISCSI__volumes_dir": "$state_path/volumes",
                        "driver__KaminarioISCSI__disable_discovery": false,
                        "driver__KaminarioISCSI__san_ip": "",
                        "driver__KaminarioISCSI__san_login": "admin",
                        "driver__KaminarioISCSI__san_password": "",
                        "driver__KaminarioISCSI__unique_fqdn_network": true,
                        "driver__KaminarioISCSI__volume_dd_blocksize": "1M",
                        "driver__LVMVolume__iscsi_iotype": "fileio",
                        "driver__LVMVolume__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__LVMVolume__iscsi_target_flags": "",
                        "driver__LVMVolume__iscsi_write_cache": "on",
                        "driver__LVMVolume__lvm_conf_file": "/etc/cinder/lvm.conf",
                        "driver__LVMVolume__lvm_mirrors": 0,
                        "driver__LVMVolume__lvm_suppress_fd_warnings": false,
                        "driver__LVMVolume__lvm_type": "auto",
                        "driver__LVMVolume__nvmet_port_id": 1,
                        "driver__LVMVolume__scst_target_driver": "iscsi",
                        "driver__LVMVolume__scst_target_iqn_name": "",
                        "driver__LVMVolume__spdk_max_queue_depth": 64,
                        "driver__LVMVolume__spdk_rpc_ip": "",
                        "driver__LVMVolume__spdk_rpc_password": "",
                        "driver__LVMVolume__spdk_rpc_port": 8000,
                        "driver__LVMVolume__spdk_rpc_username": "",
                        "driver__LVMVolume__target_helper": "tgtadm",
                        "driver__LVMVolume__target_ip_address": "$my_ip",
                        "driver__LVMVolume__target_port": 3260,
                        "driver__LVMVolume__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__LVMVolume__target_protocol": "iscsi",
                        "driver__LVMVolume__volume_clear": "zero",
                        "driver__LVMVolume__volume_clear_size": 0,
                        "driver__LVMVolume__volume_dd_blocksize": "1M",
                        "driver__LVMVolume__volume_group": "cinder-volumes",
                        "driver__LVMVolume__volumes_dir": "$state_path/volumes",
                        "driver__LenovoFC__lenovo_pool_name": "A",
                        "driver__LenovoFC__lenovo_pool_type": "virtual",
                        "driver__LenovoISCSI__lenovo_iscsi_ips__transform_csv": "",
                        "driver__LenovoISCSI__lenovo_pool_name": "A",
                        "driver__LenovoISCSI__lenovo_pool_type": "virtual",
                        "driver__LinstorDrbd__linstor_autoplace_count": 0,
                        "driver__LinstorDrbd__linstor_controller_diskless": true,
                        "driver__LinstorDrbd__linstor_default_blocksize": 4096,
                        "driver__LinstorDrbd__linstor_default_storage_pool_name": "DfltStorPool",
                        "driver__LinstorDrbd__linstor_default_uri": "linstor://localhost",
                        "driver__LinstorDrbd__linstor_default_volume_group_name": "drbd-vg",
                        "driver__LinstorDrbd__linstor_volume_downsize_factor__transform_string_float": "4096",
                        "driver__LinstorIscsi__linstor_autoplace_count": 0,
                        "driver__LinstorIscsi__linstor_controller_diskless": true,
                        "driver__LinstorIscsi__linstor_default_blocksize": 4096,
                        "driver__LinstorIscsi__linstor_default_storage_pool_name": "DfltStorPool",
                        "driver__LinstorIscsi__linstor_default_uri": "linstor://localhost",
                        "driver__LinstorIscsi__linstor_default_volume_group_name": "drbd-vg",
                        "driver__LinstorIscsi__linstor_volume_downsize_factor__transform_string_float": "4096",
                        "driver__MStorageFC__nec_actual_free_capacity": false,
                        "driver__MStorageFC__nec_auto_accesscontrol": true,
                        "driver__MStorageFC__nec_backend_max_ld_count": 1024,
                        "driver__MStorageFC__nec_backup_ldname_format": "LX:%s",
                        "driver__MStorageFC__nec_backup_pools__transform_csv": "",
                        "driver__MStorageFC__nec_cv_ldname_format": "LX:__ControlVolume_%xh",
                        "driver__MStorageFC__nec_diskarray_name": "",
                        "driver__MStorageFC__nec_ismcli_fip": "",
                        "driver__MStorageFC__nec_ismcli_password": "",
                        "driver__MStorageFC__nec_ismcli_privkey": "",
                        "driver__MStorageFC__nec_ismcli_user": "",
                        "driver__MStorageFC__nec_ismview_alloptimize": false,
                        "driver__MStorageFC__nec_ismview_dir": "/tmp/nec/cinder",
                        "driver__MStorageFC__nec_ldname_format": "LX:%s",
                        "driver__MStorageFC__nec_ldset": "",
                        "driver__MStorageFC__nec_pools__transform_csv": "",
                        "driver__MStorageFC__nec_queryconfig_view": false,
                        "driver__MStorageFC__nec_ssh_pool_port_number": 22,
                        "driver__MStorageFC__nec_unpairthread_timeout": 3600,
                        "driver__MStorageISCSI__nec_actual_free_capacity": false,
                        "driver__MStorageISCSI__nec_auto_accesscontrol": true,
                        "driver__MStorageISCSI__nec_backend_max_ld_count": 1024,
                        "driver__MStorageISCSI__nec_backup_ldname_format": "LX:%s",
                        "driver__MStorageISCSI__nec_backup_pools__transform_csv": "",
                        "driver__MStorageISCSI__nec_cv_ldname_format": "LX:__ControlVolume_%xh",
                        "driver__MStorageISCSI__nec_diskarray_name": "",
                        "driver__MStorageISCSI__nec_ismcli_fip": "",
                        "driver__MStorageISCSI__nec_ismcli_password": "",
                        "driver__MStorageISCSI__nec_ismcli_privkey": "",
                        "driver__MStorageISCSI__nec_ismcli_user": "",
                        "driver__MStorageISCSI__nec_ismview_alloptimize": false,
                        "driver__MStorageISCSI__nec_ismview_dir": "/tmp/nec/cinder",
                        "driver__MStorageISCSI__nec_ldname_format": "LX:%s",
                        "driver__MStorageISCSI__nec_ldset": "",
                        "driver__MStorageISCSI__nec_pools__transform_csv": "",
                        "driver__MStorageISCSI__nec_queryconfig_view": false,
                        "driver__MStorageISCSI__nec_ssh_pool_port_number": 22,
                        "driver__MStorageISCSI__nec_unpairthread_timeout": 3600,
                        "driver__MacroSANFC__backend_availability_zone": "",
                        "driver__MacroSANFC__chap_password": "",
                        "driver__MacroSANFC__chap_username": "",
                        "driver__MacroSANFC__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__MacroSANFC__driver_client_cert": "",
                        "driver__MacroSANFC__driver_client_cert_key": "",
                        "driver__MacroSANFC__driver_data_namespace": "",
                        "driver__MacroSANFC__driver_ssl_cert_path": "",
                        "driver__MacroSANFC__driver_ssl_cert_verify": false,
                        "driver__MacroSANFC__driver_use_ssl": false,
                        "driver__MacroSANFC__filter_function": "",
                        "driver__MacroSANFC__goodness_function": "",
                        "driver__MacroSANFC__iscsi_iotype": "fileio",
                        "driver__MacroSANFC__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__MacroSANFC__iscsi_target_flags": "",
                        "driver__MacroSANFC__iscsi_write_cache": "on",
                        "driver__MacroSANFC__num_shell_tries": 3,
                        "driver__MacroSANFC__num_volume_device_scan_tries": 3,
                        "driver__MacroSANFC__report_discard_supported": false,
                        "driver__MacroSANFC__storage_protocol": "iscsi",
                        "driver__MacroSANFC__target_helper": "tgtadm",
                        "driver__MacroSANFC__target_ip_address": "$my_ip",
                        "driver__MacroSANFC__target_port": 3260,
                        "driver__MacroSANFC__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__MacroSANFC__target_protocol": "iscsi",
                        "driver__MacroSANFC__trace_flags__transform_csv": "",
                        "driver__MacroSANFC__use_chap_auth": false,
                        "driver__MacroSANFC__volume_backend_name": "",
                        "driver__MacroSANFC__volume_clear": "zero",
                        "driver__MacroSANFC__volume_clear_ionice": "",
                        "driver__MacroSANFC__volume_clear_size": 0,
                        "driver__MacroSANFC__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__MacroSANFC__volume_copy_bps_limit": 0,
                        "driver__MacroSANFC__volume_dd_blocksize": "1M",
                        "driver__MacroSANFC__volumes_dir": "$state_path/volumes",
                        "driver__MacroSANISCSI__backend_availability_zone": "",
                        "driver__MacroSANISCSI__chap_password": "",
                        "driver__MacroSANISCSI__chap_username": "",
                        "driver__MacroSANISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__MacroSANISCSI__driver_client_cert": "",
                        "driver__MacroSANISCSI__driver_client_cert_key": "",
                        "driver__MacroSANISCSI__driver_data_namespace": "",
                        "driver__MacroSANISCSI__driver_ssl_cert_path": "",
                        "driver__MacroSANISCSI__driver_ssl_cert_verify": false,
                        "driver__MacroSANISCSI__driver_use_ssl": false,
                        "driver__MacroSANISCSI__filter_function": "",
                        "driver__MacroSANISCSI__goodness_function": "",
                        "driver__MacroSANISCSI__iscsi_iotype": "fileio",
                        "driver__MacroSANISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__MacroSANISCSI__iscsi_target_flags": "",
                        "driver__MacroSANISCSI__iscsi_write_cache": "on",
                        "driver__MacroSANISCSI__num_shell_tries": 3,
                        "driver__MacroSANISCSI__num_volume_device_scan_tries": 3,
                        "driver__MacroSANISCSI__report_discard_supported": false,
                        "driver__MacroSANISCSI__storage_protocol": "iscsi",
                        "driver__MacroSANISCSI__target_helper": "tgtadm",
                        "driver__MacroSANISCSI__target_ip_address": "$my_ip",
                        "driver__MacroSANISCSI__target_port": 3260,
                        "driver__MacroSANISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__MacroSANISCSI__target_protocol": "iscsi",
                        "driver__MacroSANISCSI__trace_flags__transform_csv": "",
                        "driver__MacroSANISCSI__use_chap_auth": false,
                        "driver__MacroSANISCSI__volume_backend_name": "",
                        "driver__MacroSANISCSI__volume_clear": "zero",
                        "driver__MacroSANISCSI__volume_clear_ionice": "",
                        "driver__MacroSANISCSI__volume_clear_size": 0,
                        "driver__MacroSANISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__MacroSANISCSI__volume_copy_bps_limit": 0,
                        "driver__MacroSANISCSI__volume_dd_blocksize": "1M",
                        "driver__MacroSANISCSI__volumes_dir": "$state_path/volumes",
                        "driver__NetAppCmodeFibreChannel__netapp_vserver": "",
                        "driver__NetAppCmodeISCSI__netapp_vserver": "",
                        "driver__NexentaISCSI__nexenta_blocksize": 4096,
                        "driver__NexentaISCSI__nexenta_dataset_compression": "on",
                        "driver__NexentaISCSI__nexenta_dataset_dedup": "off",
                        "driver__NexentaISCSI__nexenta_dataset_description": "",
                        "driver__NexentaISCSI__nexenta_folder": "",
                        "driver__NexentaISCSI__nexenta_group_snapshot_template": "group-snapshot-%s",
                        "driver__NexentaISCSI__nexenta_host": "",
                        "driver__NexentaISCSI__nexenta_host_group_prefix": "cinder",
                        "driver__NexentaISCSI__nexenta_iscsi_target_host_group": "all",
                        "driver__NexentaISCSI__nexenta_iscsi_target_portal_groups": "",
                        "driver__NexentaISCSI__nexenta_iscsi_target_portal_port": 3260,
                        "driver__NexentaISCSI__nexenta_iscsi_target_portals": "",
                        "driver__NexentaISCSI__nexenta_lu_writebackcache_disabled": false,
                        "driver__NexentaISCSI__nexenta_luns_per_target": 100,
                        "driver__NexentaISCSI__nexenta_ns5_blocksize": 32,
                        "driver__NexentaISCSI__nexenta_origin_snapshot_template": "origin-snapshot-%s",
                        "driver__NexentaISCSI__nexenta_rest_backoff_factor__transform_string_float": "0.5",
                        "driver__NexentaISCSI__nexenta_rest_connect_timeout__transform_string_float": "30",
                        "driver__NexentaISCSI__nexenta_rest_protocol": "auto",
                        "driver__NexentaISCSI__nexenta_rest_read_timeout__transform_string_float": "300",
                        "driver__NexentaISCSI__nexenta_rest_retry_count": 3,
                        "driver__NexentaISCSI__nexenta_sparse": false,
                        "driver__NexentaISCSI__nexenta_target_group_prefix": "cinder",
                        "driver__NexentaISCSI__nexenta_target_prefix": "iqn.1986-03.com.sun:02:cinder",
                        "driver__NexentaISCSI__nexenta_use_https": true,
                        "driver__NexentaISCSI__nexenta_volume": "cinder",
                        "driver__NexentaISCSI__nexenta_volume_group": "iscsi",
                        "driver__NimbleFC__nimble_pool_name": "default",
                        "driver__NimbleFC__nimble_subnet_label": "*",
                        "driver__NimbleFC__nimble_verify_cert_path": "",
                        "driver__NimbleFC__nimble_verify_certificate": false,
                        "driver__NimbleISCSI__nimble_pool_name": "default",
                        "driver__NimbleISCSI__nimble_subnet_label": "*",
                        "driver__NimbleISCSI__nimble_verify_cert_path": "",
                        "driver__NimbleISCSI__nimble_verify_certificate": false,
                        "driver__PVMEFC__driver_ssl_cert_path": "",
                        "driver__PVMEFC__driver_ssl_cert_verify": false,
                        "driver__PVMEFC__driver_use_ssl": false,
                        "driver__PVMEFC__pvme_pool_name": "A",
                        "driver__PVMEFC__san_ip": "",
                        "driver__PVMEFC__san_login": "admin",
                        "driver__PVMEFC__san_password": "",
                        "driver__PVMEISCSI__driver_ssl_cert_path": "",
                        "driver__PVMEISCSI__driver_ssl_cert_verify": false,
                        "driver__PVMEISCSI__driver_use_ssl": false,
                        "driver__PVMEISCSI__pvme_iscsi_ips__transform_csv": "",
                        "driver__PVMEISCSI__pvme_pool_name": "A",
                        "driver__PVMEISCSI__san_ip": "",
                        "driver__PVMEISCSI__san_login": "admin",
                        "driver__PVMEISCSI__san_password": "",
                        "driver__PowerFlex__powerflex_allow_migration_during_rebuild": false,
                        "driver__PowerFlex__powerflex_allow_non_padded_volumes": false,
                        "driver__PowerFlex__powerflex_max_over_subscription_ratio__transform_string_float": "10.0",
                        "driver__PowerFlex__powerflex_rest_server_port": 443,
                        "driver__PowerFlex__powerflex_round_volume_capacity": true,
                        "driver__PowerFlex__powerflex_server_api_version": "",
                        "driver__PowerFlex__powerflex_storage_pools": "",
                        "driver__PowerFlex__powerflex_unmap_volume_before_deletion": false,
                        "driver__PowerMaxFC__driver_ssl_cert_verify": false,
                        "driver__PowerMaxFC__initiator_check": false,
                        "driver__PowerMaxFC__interval": 3,
                        "driver__PowerMaxFC__load_balance": false,
                        "driver__PowerMaxFC__load_balance_real_time": false,
                        "driver__PowerMaxFC__load_data_format": "Avg",
                        "driver__PowerMaxFC__load_look_back": 60,
                        "driver__PowerMaxFC__load_look_back_real_time": 1,
                        "driver__PowerMaxFC__port_group_load_metric": "PercentBusy",
                        "driver__PowerMaxFC__port_load_metric": "PercentBusy",
                        "driver__PowerMaxFC__powermax_array": "",
                        "driver__PowerMaxFC__powermax_array_tag_list__transform_csv": "",
                        "driver__PowerMaxFC__powermax_port_group_name_template": "portGroupName",
                        "driver__PowerMaxFC__powermax_port_groups__transform_csv": "",
                        "driver__PowerMaxFC__powermax_service_level": "",
                        "driver__PowerMaxFC__powermax_short_host_name_template": "shortHostName",
                        "driver__PowerMaxFC__powermax_snapvx_unlink_limit": 3,
                        "driver__PowerMaxFC__powermax_srp": "",
                        "driver__PowerMaxFC__retries": 200,
                        "driver__PowerMaxFC__san_ip": "",
                        "driver__PowerMaxFC__san_login": "admin",
                        "driver__PowerMaxFC__san_password": "",
                        "driver__PowerMaxFC__u4p_failover_autofailback": true,
                        "driver__PowerMaxFC__u4p_failover_backoff_factor": 1,
                        "driver__PowerMaxFC__u4p_failover_retries": 3,
                        "driver__PowerMaxFC__u4p_failover_target__transform_csv_kvs": "",
                        "driver__PowerMaxFC__u4p_failover_timeout": 20,
                        "driver__PowerMaxFC__vmax_workload": "",
                        "driver__PowerMaxISCSI__chap_password": "",
                        "driver__PowerMaxISCSI__chap_username": "",
                        "driver__PowerMaxISCSI__driver_ssl_cert_verify": false,
                        "driver__PowerMaxISCSI__initiator_check": false,
                        "driver__PowerMaxISCSI__interval": 3,
                        "driver__PowerMaxISCSI__load_balance": false,
                        "driver__PowerMaxISCSI__load_balance_real_time": false,
                        "driver__PowerMaxISCSI__load_data_format": "Avg",
                        "driver__PowerMaxISCSI__load_look_back": 60,
                        "driver__PowerMaxISCSI__load_look_back_real_time": 1,
                        "driver__PowerMaxISCSI__port_group_load_metric": "PercentBusy",
                        "driver__PowerMaxISCSI__port_load_metric": "PercentBusy",
                        "driver__PowerMaxISCSI__powermax_array": "",
                        "driver__PowerMaxISCSI__powermax_array_tag_list__transform_csv": "",
                        "driver__PowerMaxISCSI__powermax_port_group_name_template": "portGroupName",
                        "driver__PowerMaxISCSI__powermax_port_groups__transform_csv": "",
                        "driver__PowerMaxISCSI__powermax_service_level": "",
                        "driver__PowerMaxISCSI__powermax_short_host_name_template": "shortHostName",
                        "driver__PowerMaxISCSI__powermax_snapvx_unlink_limit": 3,
                        "driver__PowerMaxISCSI__powermax_srp": "",
                        "driver__PowerMaxISCSI__retries": 200,
                        "driver__PowerMaxISCSI__san_ip": "",
                        "driver__PowerMaxISCSI__san_login": "admin",
                        "driver__PowerMaxISCSI__san_password": "",
                        "driver__PowerMaxISCSI__u4p_failover_autofailback": true,
                        "driver__PowerMaxISCSI__u4p_failover_backoff_factor": 1,
                        "driver__PowerMaxISCSI__u4p_failover_retries": 3,
                        "driver__PowerMaxISCSI__u4p_failover_target__transform_csv_kvs": "",
                        "driver__PowerMaxISCSI__u4p_failover_timeout": 20,
                        "driver__PowerMaxISCSI__use_chap_auth": false,
                        "driver__PowerMaxISCSI__vmax_workload": "",
                        "driver__PowerStore__powerstore_appliances__transform_csv": "",
                        "driver__PowerStore__powerstore_ports__transform_csv": "",
                        "driver__PureFC__driver_ssl_cert_path": "",
                        "driver__PureFC__driver_ssl_cert_verify": false,
                        "driver__PureFC__pure_api_token": "",
                        "driver__PureFC__pure_automatic_max_oversubscription_ratio": true,
                        "driver__PureFC__pure_eradicate_on_delete": false,
                        "driver__PureFC__pure_host_personality__transform_empty_none": "",
                        "driver__PureFC__pure_iscsi_cidr": "0.0.0.0/0",
                        "driver__PureFC__pure_replica_interval_default": 3600,
                        "driver__PureFC__pure_replica_retention_long_term_default": 7,
                        "driver__PureFC__pure_replica_retention_long_term_per_day_default": 3,
                        "driver__PureFC__pure_replica_retention_short_term_default": 14400,
                        "driver__PureFC__pure_replication_pg_name": "cinder-group",
                        "driver__PureFC__pure_replication_pod_name": "cinder-pod",
                        "driver__PureFC__san_ip": "",
                        "driver__PureFC__use_chap_auth": false,
                        "driver__PureISCSI__driver_ssl_cert_path": "",
                        "driver__PureISCSI__driver_ssl_cert_verify": false,
                        "driver__PureISCSI__pure_api_token": "",
                        "driver__PureISCSI__pure_automatic_max_oversubscription_ratio": true,
                        "driver__PureISCSI__pure_eradicate_on_delete": false,
                        "driver__PureISCSI__pure_host_personality__transform_empty_none": "",
                        "driver__PureISCSI__pure_iscsi_cidr": "0.0.0.0/0",
                        "driver__PureISCSI__pure_replica_interval_default": 3600,
                        "driver__PureISCSI__pure_replica_retention_long_term_default": 7,
                        "driver__PureISCSI__pure_replica_retention_long_term_per_day_default": 3,
                        "driver__PureISCSI__pure_replica_retention_short_term_default": 14400,
                        "driver__PureISCSI__pure_replication_pg_name": "cinder-group",
                        "driver__PureISCSI__pure_replication_pod_name": "cinder-pod",
                        "driver__PureISCSI__san_ip": "",
                        "driver__PureISCSI__use_chap_auth": false,
                        "driver__QnapISCSI__chap_password": "",
                        "driver__QnapISCSI__chap_username": "",
                        "driver__QnapISCSI__driver_ssl_cert_verify": false,
                        "driver__QnapISCSI__qnap_management_url": "",
                        "driver__QnapISCSI__qnap_poolname": "",
                        "driver__QnapISCSI__qnap_storage_protocol": "iscsi",
                        "driver__QnapISCSI__san_login": "admin",
                        "driver__QnapISCSI__san_password": "",
                        "driver__QnapISCSI__target_ip_address": "$my_ip",
                        "driver__QnapISCSI__use_chap_auth": false,
                        "driver__Quobyte__quobyte_client_cfg": "",
                        "driver__Quobyte__quobyte_mount_point_base": "$state_path/mnt",
                        "driver__Quobyte__quobyte_overlay_volumes": false,
                        "driver__Quobyte__quobyte_qcow2_volumes": true,
                        "driver__Quobyte__quobyte_sparsed_volumes": true,
                        "driver__Quobyte__quobyte_volume_from_snapshot_cache": false,
                        "driver__Quobyte__quobyte_volume_url": "",
                        "driver__RBD__deferred_deletion_delay": 0,
                        "driver__RBD__deferred_deletion_purge_interval": 60,
                        "driver__RBD__enable_deferred_deletion": false,
                        "driver__RBD__rados_connect_timeout": -1,
                        "driver__RBD__rados_connection_interval": 5,
                        "driver__RBD__rados_connection_retries": 3,
                        "driver__RBD__rbd_ceph_conf": "",
                        "driver__RBD__rbd_cluster_name": "ceph",
                        "driver__RBD__rbd_exclusive_cinder_pool": false,
                        "driver__RBD__rbd_flatten_volume_from_snapshot": false,
                        "driver__RBD__rbd_keyring_conf": "",
                        "driver__RBD__rbd_max_clone_depth": 5,
                        "driver__RBD__rbd_pool": "rbd",
                        "driver__RBD__rbd_secret_uuid": "",
                        "driver__RBD__rbd_store_chunk_size": 4,
                        "driver__RBD__rbd_user": "",
                        "driver__RBD__report_dynamic_total_capacity": true,
                        "driver__RBD__volume_dd_blocksize": "1M",
                        "driver__RBDISCSI__deferred_deletion_delay": 0,
                        "driver__RBDISCSI__deferred_deletion_purge_interval": 60,
                        "driver__RBDISCSI__driver_ssl_cert_verify": false,
                        "driver__RBDISCSI__enable_deferred_deletion": false,
                        "driver__RBDISCSI__rados_connect_timeout": -1,
                        "driver__RBDISCSI__rados_connection_interval": 5,
                        "driver__RBDISCSI__rados_connection_retries": 3,
                        "driver__RBDISCSI__rbd_ceph_conf": "",
                        "driver__RBDISCSI__rbd_cluster_name": "ceph",
                        "driver__RBDISCSI__rbd_exclusive_cinder_pool": false,
                        "driver__RBDISCSI__rbd_flatten_volume_from_snapshot": false,
                        "driver__RBDISCSI__rbd_iscsi_api_debug": false,
                        "driver__RBDISCSI__rbd_iscsi_api_password": "",
                        "driver__RBDISCSI__rbd_iscsi_api_url": "",
                        "driver__RBDISCSI__rbd_iscsi_api_user": "",
                        "driver__RBDISCSI__rbd_iscsi_target_iqn": "",
                        "driver__RBDISCSI__rbd_max_clone_depth": 5,
                        "driver__RBDISCSI__rbd_pool": "rbd",
                        "driver__RBDISCSI__rbd_secret_uuid": "",
                        "driver__RBDISCSI__rbd_store_chunk_size": 4,
                        "driver__RBDISCSI__rbd_user": "",
                        "driver__RBDISCSI__report_dynamic_total_capacity": true,
                        "driver__RBDISCSI__suppress_requests_ssl_warnings": false,
                        "driver__RBDISCSI__volume_dd_blocksize": "1M",
                        "driver__RSD__podm_password": "",
                        "driver__RSD__podm_url": "",
                        "driver__RSD__podm_username": "",
                        "driver__SCFC__dell_api_async_rest_timeout": 15,
                        "driver__SCFC__dell_api_sync_rest_timeout": 30,
                        "driver__SCFC__dell_sc_api_port": 3033,
                        "driver__SCFC__dell_sc_server_folder": "openstack",
                        "driver__SCFC__dell_sc_ssn": 64702,
                        "driver__SCFC__dell_sc_verify_cert": false,
                        "driver__SCFC__dell_sc_volume_folder": "openstack",
                        "driver__SCFC__dell_server_os": "Red Hat Linux 6.x",
                        "driver__SCFC__excluded_domain_ips__transform_csv": "",
                        "driver__SCFC__included_domain_ips__transform_csv": "",
                        "driver__SCFC__secondary_san_ip": "",
                        "driver__SCFC__secondary_san_login": "Admin",
                        "driver__SCFC__secondary_san_password": "",
                        "driver__SCFC__secondary_sc_api_port": 3033,
                        "driver__SCISCSI__dell_api_async_rest_timeout": 15,
                        "driver__SCISCSI__dell_api_sync_rest_timeout": 30,
                        "driver__SCISCSI__dell_sc_api_port": 3033,
                        "driver__SCISCSI__dell_sc_server_folder": "openstack",
                        "driver__SCISCSI__dell_sc_ssn": 64702,
                        "driver__SCISCSI__dell_sc_verify_cert": false,
                        "driver__SCISCSI__dell_sc_volume_folder": "openstack",
                        "driver__SCISCSI__dell_server_os": "Red Hat Linux 6.x",
                        "driver__SCISCSI__excluded_domain_ips__transform_csv": "",
                        "driver__SCISCSI__included_domain_ips__transform_csv": "",
                        "driver__SCISCSI__secondary_san_ip": "",
                        "driver__SCISCSI__secondary_san_login": "Admin",
                        "driver__SCISCSI__secondary_san_password": "",
                        "driver__SCISCSI__secondary_sc_api_port": 3033,
                        "driver__SdsISCSI__backend_availability_zone": "",
                        "driver__SdsISCSI__chap_password": "",
                        "driver__SdsISCSI__chap_username": "",
                        "driver__SdsISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__SdsISCSI__driver_client_cert": "",
                        "driver__SdsISCSI__driver_client_cert_key": "",
                        "driver__SdsISCSI__driver_data_namespace": "",
                        "driver__SdsISCSI__driver_ssl_cert_path": "",
                        "driver__SdsISCSI__driver_ssl_cert_verify": false,
                        "driver__SdsISCSI__driver_use_ssl": false,
                        "driver__SdsISCSI__filter_function": "",
                        "driver__SdsISCSI__goodness_function": "",
                        "driver__SdsISCSI__iscsi_iotype": "fileio",
                        "driver__SdsISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__SdsISCSI__iscsi_target_flags": "",
                        "driver__SdsISCSI__iscsi_write_cache": "on",
                        "driver__SdsISCSI__num_shell_tries": 3,
                        "driver__SdsISCSI__num_volume_device_scan_tries": 3,
                        "driver__SdsISCSI__report_discard_supported": false,
                        "driver__SdsISCSI__storage_protocol": "iscsi",
                        "driver__SdsISCSI__target_helper": "tgtadm",
                        "driver__SdsISCSI__target_ip_address": "$my_ip",
                        "driver__SdsISCSI__target_port": 3260,
                        "driver__SdsISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__SdsISCSI__target_protocol": "iscsi",
                        "driver__SdsISCSI__trace_flags__transform_csv": "",
                        "driver__SdsISCSI__use_chap_auth": false,
                        "driver__SdsISCSI__volume_backend_name": "",
                        "driver__SdsISCSI__volume_clear": "zero",
                        "driver__SdsISCSI__volume_clear_ionice": "",
                        "driver__SdsISCSI__volume_clear_size": 0,
                        "driver__SdsISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__SdsISCSI__volume_copy_bps_limit": 0,
                        "driver__SdsISCSI__volume_dd_blocksize": "1M",
                        "driver__SdsISCSI__volumes_dir": "$state_path/volumes",
                        "driver__SolidFire__driver_ssl_cert_verify": false,
                        "driver__SolidFire__san_ip": "",
                        "driver__SolidFire__san_login": "admin",
                        "driver__SolidFire__san_password": "",
                        "driver__SolidFire__sf_account_prefix": "",
                        "driver__SolidFire__sf_allow_tenant_qos": false,
                        "driver__SolidFire__sf_api_port": 443,
                        "driver__SolidFire__sf_api_request_timeout": 30,
                        "driver__SolidFire__sf_cluster_pairing_timeout": 60,
                        "driver__SolidFire__sf_emulate_512": true,
                        "driver__SolidFire__sf_enable_vag": false,
                        "driver__SolidFire__sf_provisioning_calc": "maxProvisionedSpace",
                        "driver__SolidFire__sf_svip": "",
                        "driver__SolidFire__sf_volume_clone_timeout": 600,
                        "driver__SolidFire__sf_volume_create_timeout": 60,
                        "driver__SolidFire__sf_volume_pairing_timeout": 3600,
                        "driver__SolidFire__sf_volume_prefix": "UUID-",
                        "driver__StorPool__storpool_replication": 3,
                        "driver__StorPool__storpool_template": "",
                        "driver__StorwizeSVCFC__cycle_period_seconds": 300,
                        "driver__StorwizeSVCFC__storwize_peer_pool": "",
                        "driver__StorwizeSVCFC__storwize_preferred_host_site__transform_csv_kvs": "",
                        "driver__StorwizeSVCFC__storwize_san_secondary_ip": "",
                        "driver__StorwizeSVCFC__storwize_svc_allow_tenant_qos": false,
                        "driver__StorwizeSVCFC__storwize_svc_flashcopy_rate": 50,
                        "driver__StorwizeSVCFC__storwize_svc_flashcopy_timeout": 120,
                        "driver__StorwizeSVCFC__storwize_svc_mirror_pool": "",
                        "driver__StorwizeSVCFC__storwize_svc_multipath_enabled": false,
                        "driver__StorwizeSVCFC__storwize_svc_retain_aux_volume": false,
                        "driver__StorwizeSVCFC__storwize_svc_stretched_cluster_partner": "",
                        "driver__StorwizeSVCFC__storwize_svc_vol_autoexpand": true,
                        "driver__StorwizeSVCFC__storwize_svc_vol_compression": false,
                        "driver__StorwizeSVCFC__storwize_svc_vol_easytier": true,
                        "driver__StorwizeSVCFC__storwize_svc_vol_grainsize": 256,
                        "driver__StorwizeSVCFC__storwize_svc_vol_iogrp": "0",
                        "driver__StorwizeSVCFC__storwize_svc_vol_nofmtdisk": false,
                        "driver__StorwizeSVCFC__storwize_svc_vol_rsize": 2,
                        "driver__StorwizeSVCFC__storwize_svc_vol_warning": 0,
                        "driver__StorwizeSVCFC__storwize_svc_volpool_name__transform_csv": "volpool",
                        "driver__StorwizeSVCISCSI__cycle_period_seconds": 300,
                        "driver__StorwizeSVCISCSI__storwize_peer_pool": "",
                        "driver__StorwizeSVCISCSI__storwize_preferred_host_site__transform_csv_kvs": "",
                        "driver__StorwizeSVCISCSI__storwize_san_secondary_ip": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_allow_tenant_qos": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_flashcopy_rate": 50,
                        "driver__StorwizeSVCISCSI__storwize_svc_flashcopy_timeout": 120,
                        "driver__StorwizeSVCISCSI__storwize_svc_iscsi_chap_enabled": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_mirror_pool": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_retain_aux_volume": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_stretched_cluster_partner": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_autoexpand": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_compression": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_easytier": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_grainsize": 256,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_iogrp": "0",
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_nofmtdisk": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_rsize": 2,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_warning": 0,
                        "driver__StorwizeSVCISCSI__storwize_svc_volpool_name__transform_csv": "volpool",
                        "driver__SynoISCSI__chap_password": "",
                        "driver__SynoISCSI__chap_username": "",
                        "driver__SynoISCSI__driver_use_ssl": false,
                        "driver__SynoISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__SynoISCSI__synology_admin_port": 5000,
                        "driver__SynoISCSI__synology_device_id": "",
                        "driver__SynoISCSI__synology_one_time_pass": "",
                        "driver__SynoISCSI__synology_password": "",
                        "driver__SynoISCSI__synology_pool_name": "",
                        "driver__SynoISCSI__synology_ssl_verify": true,
                        "driver__SynoISCSI__synology_username": "admin",
                        "driver__SynoISCSI__target_ip_address": "$my_ip",
                        "driver__SynoISCSI__target_port": 3260,
                        "driver__SynoISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__SynoISCSI__target_protocol": "iscsi",
                        "driver__SynoISCSI__use_chap_auth": false,
                        "driver__Unity__remove_empty_host": false,
                        "driver__Unity__unity_io_ports__transform_csv": "",
                        "driver__Unity__unity_storage_pool_names__transform_csv": "",
                        "driver__VNX__default_timeout": 31536000,
                        "driver__VNX__destroy_empty_storage_group": false,
                        "driver__VNX__force_delete_lun_in_storagegroup": true,
                        "driver__VNX__ignore_pool_full_threshold": false,
                        "driver__VNX__initiator_auto_deregistration": false,
                        "driver__VNX__initiator_auto_registration": false,
                        "driver__VNX__io_port_list__transform_csv": "",
                        "driver__VNX__iscsi_initiators": "",
                        "driver__VNX__max_luns_per_storage_group": 255,
                        "driver__VNX__naviseccli_path": "",
                        "driver__VNX__storage_vnx_authentication_type": "global",
                        "driver__VNX__storage_vnx_pool_names__transform_csv": "",
                        "driver__VNX__storage_vnx_security_file_dir": "",
                        "driver__VNX__vnx_async_migrate": true,
                        "driver__VZStorage__vzstorage_default_volume_format": "raw",
                        "driver__VZStorage__vzstorage_mount_options__transform_csv": "",
                        "driver__VZStorage__vzstorage_mount_point_base": "$state_path/mnt",
                        "driver__VZStorage__vzstorage_shares_config": "/etc/cinder/vzstorage_shares",
                        "driver__VZStorage__vzstorage_sparsed_volumes": true,
                        "driver__VZStorage__vzstorage_used_ratio__transform_string_float": "0.95",
                        "driver__WindowsISCSI__windows_iscsi_lun_path": "C:\\iSCSIVirtualDisks",
                        "driver__WindowsSmbfs__smbfs_default_volume_format": "vhd",
                        "driver__WindowsSmbfs__smbfs_mount_point_base": "C:\\OpenStack\\_mnt",
                        "driver__WindowsSmbfs__smbfs_pool_mappings__transform_csv_kvs": "",
                        "driver__WindowsSmbfs__smbfs_shares_config": "C:\\OpenStack\\smbfs_shares.txt",
                        "driver__XtremIOFC__driver_ssl_cert_path": "",
                        "driver__XtremIOFC__driver_ssl_cert_verify": false,
                        "driver__XtremIOFC__san_ip": "",
                        "driver__XtremIOFC__san_login": "admin",
                        "driver__XtremIOFC__san_password": "",
                        "driver__XtremIOFC__xtremio_array_busy_retry_count": 5,
                        "driver__XtremIOFC__xtremio_array_busy_retry_interval": 5,
                        "driver__XtremIOFC__xtremio_clean_unused_ig": false,
                        "driver__XtremIOFC__xtremio_cluster_name": "",
                        "driver__XtremIOISCSI__driver_ssl_cert_path": "",
                        "driver__XtremIOISCSI__driver_ssl_cert_verify": false,
                        "driver__XtremIOISCSI__san_ip": "",
                        "driver__XtremIOISCSI__san_login": "admin",
                        "driver__XtremIOISCSI__san_password": "",
                        "driver__XtremIOISCSI__xtremio_array_busy_retry_count": 5,
                        "driver__XtremIOISCSI__xtremio_array_busy_retry_interval": 5,
                        "driver__XtremIOISCSI__xtremio_clean_unused_ig": false,
                        "driver__XtremIOISCSI__xtremio_cluster_name": "",
                        "driver__ZadaraVPSAISCSI__zadara_access_key": "",
                        "driver__ZadaraVPSAISCSI__zadara_default_snap_policy": false,
                        "driver__ZadaraVPSAISCSI__zadara_ssl_cert_verify": true,
                        "driver__ZadaraVPSAISCSI__zadara_use_iser": true,
                        "driver__ZadaraVPSAISCSI__zadara_vol_encrypt": false,
                        "driver__ZadaraVPSAISCSI__zadara_vol_name_template": "OS_%s",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_host": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_poolname": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_port": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_use_ssl": false
                      }
                    },
                    "sysFiles": {
                      "name": ""
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/embercsi/ember-csi-operator:latest
          createdAt: 2021-02-22:17:47:00
          description: Multi-vendor CSI plugin supporting 80+ storage drivers
          repository: https://github.com/embercsi/ember-csi-operator
          support: http://readthedocs.org/projects/ember-csi/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an Ember CSI storage backend
            displayName: Storage Backends
            kind: EmberStorageBackend
            name: embercsis.ember-csi.io
            version: v1alpha1
        description: |
          Ember-CSI is a multi-vendor CSI plugin driver supporting
          over 80 storage drivers in a single plugin to provide block and mount
          storage to container orchestration systems.

          The operator adds support for management of multiple Ember CSI instances
          within a cluster and simplifies storage backend deployments.

          ### Supported drivers
          Ember-CSI includes a good number of storage drivers, but due to limitation on hardware availability only a small number of them have been validated at one point or another. In alphabetical order they are:

          * HPE3PARFC
          * HPE3PARISCSI
          * KaminarioISCSI
          * LVMVolume
          * PowerMaxFC
          * PowerMaxISCSI
          * PureFC
          * PureISCSI
          * QnapISCSI
          * RBD
          * SolidFire
          * SynoISCSI
          * XtremIOFC
          * XtremIOISCSI

          ### Getting started
          Please have a look at our quick start documentation on how to get started:
          [github.com/embercsi/ember-csi-operator#quick-start](https://github.com/embercsi/ember-csi-operator#quick-start).
          We recommend to use the user friendly "Edit Form" when creating a new
          EmberStorageBackend. You'll find the link in the upper right corner of the
          editor.

          ### Documentation
          Documentation on Ember CSI itself can be found at [docs.ember-csi.io](http://docs.ember-csi.io/).

          ### Getting help
          Please contact us if you encounter any issues while using Ember CSI.
          Get in touch through the #ember-csi channel on Freenode, also accessible through
          [Kiwi IRC](https://kiwiirc.com/nextclient/irc.freenode.net/ember-csi).

          You can also create an issue on the [Github project](https://github.com/embercsi/ember-csi-operator)
          for bugs, enhancements, or other requests.
        displayName: Ember CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Ember-CSI
        - CSI
        links:
        - name: Learn more about the project
          url: http://ember-csi.io/
        - name: Documentation
          url: http://readthedocs.org/projects/ember-csi/
        - name: Ember-CSI Source Code
          url: https://github.com/embercsi/ember-csi
        - name: Ember-CSI Operator Source Code
          url: https://github.com/embercsi/ember-csi-operator
        maintainers:
        - email: geguileo@redhat.com
          name: Gorka Eguileor
        - email: cschwede@redhat.com
          name: Christian Schwede
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/embercsi/ember-csi-operator:latest
        version: 0.9.9
      entries:
      - name: ember-csi-community-operator.v0.9.9
        version: 0.9.9
      - name: ember-csi-community-operator.v0.9.8
        version: 0.9.8
      - name: ember-csi-community-operator.v0.9.7
        version: 0.9.7
      name: beta
    defaultChannel: beta
    packageName: ember-csi-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sqs-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sqs-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sqs.services.k8s.aws/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sqs-controller:1.2.1
          createdAt: "2025-09-19T19:14:25Z"
          description: AWS SQS controller is a service controller for managing SQS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Queue represents the state of an AWS sqs Queue resource.
            displayName: Queue
            kind: Queue
            name: queues.sqs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Simple Queue Service (SQS) resources in AWS from within your Kubernetes cluster.

          **About Amazon SQS**

          Amazon Simple Queue Service (Amazon SQS) offers a secure, durable, and available hosted queue that lets you integrate and decouple distributed software systems and components. Amazon SQS offers common constructs such as [dead-letter queues](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGuide/sqs-dead-letter-queues.html) and [cost allocation tags](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGuide/sqs-queue-tags.html). It provides a generic web services API that you can access using any programming language that the AWS SDK supports.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SQS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sqs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SQS Developer Resources
          url: https://aws.amazon.com/sqs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sqs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sqs-controller:1.2.1
        version: 1.2.1
      entries:
      - name: ack-sqs-controller.v1.2.1
        version: 1.2.1
      - name: ack-sqs-controller.v1.2.0
        version: 1.2.0
      - name: ack-sqs-controller.v1.1.16
        version: 1.1.16
      - name: ack-sqs-controller.v1.1.14
        version: 1.1.14
      - name: ack-sqs-controller.v1.1.13
        version: 1.1.13
      - name: ack-sqs-controller.v1.1.12
        version: 1.1.12
      - name: ack-sqs-controller.v1.1.11
        version: 1.1.11
      - name: ack-sqs-controller.v1.1.10
        version: 1.1.10
      - name: ack-sqs-controller.v1.1.9
        version: 1.1.9
      - name: ack-sqs-controller.v1.1.8
        version: 1.1.8
      - name: ack-sqs-controller.v1.1.7
        version: 1.1.7
      - name: ack-sqs-controller.v1.1.6
        version: 1.1.6
      - name: ack-sqs-controller.v1.1.5
        version: 1.1.5
      - name: ack-sqs-controller.v1.1.4
        version: 1.1.4
      - name: ack-sqs-controller.v1.1.3
        version: 1.1.3
      - name: ack-sqs-controller.v1.1.2
        version: 1.1.2
      - name: ack-sqs-controller.v1.1.1
        version: 1.1.1
      - name: ack-sqs-controller.v1.1.0
        version: 1.1.0
      - name: ack-sqs-controller.v1.0.17
        version: 1.0.17
      - name: ack-sqs-controller.v1.0.16
        version: 1.0.16
      - name: ack-sqs-controller.v1.0.15
        version: 1.0.15
      - name: ack-sqs-controller.v1.0.14
        version: 1.0.14
      - name: ack-sqs-controller.v1.0.13
        version: 1.0.13
      - name: ack-sqs-controller.v1.0.12
        version: 1.0.12
      - name: ack-sqs-controller.v1.0.10
        version: 1.0.10
      - name: ack-sqs-controller.v1.0.7
        version: 1.0.7
      - name: ack-sqs-controller.v1.0.5
        version: 1.0.5
      - name: ack-sqs-controller.v1.0.3
        version: 1.0.3
      - name: ack-sqs-controller.v1.0.2
        version: 1.0.2
      - name: ack-sqs-controller.v1.0.1
        version: 1.0.1
      - name: ack-sqs-controller.v1.0.0
        version: 1.0.0
      - name: ack-sqs-controller.v0.0.7
        version: 0.0.7
      - name: ack-sqs-controller.v0.0.6
        version: 0.0.6
      - name: ack-sqs-controller.v0.0.5
        version: 0.0.5
      - name: ack-sqs-controller.v0.0.4
        version: 0.0.4
      - name: ack-sqs-controller.v0.0.2
        version: 0.0.2
      - name: ack-sqs-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-sqs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/limitador-operator
    name: limitador-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: limitador-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/limitador-operator:v0.5.0
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: The Limitador operator installs and maintains limitador instances
        displayName: Limitador
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        links:
        - name: Limitador Operator
          url: https://github.com/Kuadrant/limitador-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - quay.io/kuadrant/limitador-operator:v0.5.0
        - quay.io/kuadrant/limitador:v1.2.0
        version: 0.5.0
      entries:
      - name: limitador-operator.v0.5.0
        version: 0.5.0
      - name: limitador-operator.v0.4.0
        version: 0.4.0
      - name: limitador-operator.v0.3.0
        version: 0.3.0
      name: alpha
    - currentCSV: limitador-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "name": "toy_get_route",
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/limitador-operator:v0.11.0
          createdAt: "2024-10-02T10:46:31Z"
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: The Limitador operator installs and maintains limitador instances
        displayName: Limitador
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        links:
        - name: Limitador Operator
          url: https://github.com/Kuadrant/limitador-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - quay.io/kuadrant/limitador-operator:v0.11.0
        - quay.io/kuadrant/limitador:v1.6.0
        version: 0.11.0
      entries:
      - name: limitador-operator.v0.11.0
        version: 0.11.0
      - name: limitador-operator.v0.10.0
        version: 0.10.0
      - name: limitador-operator.v0.6.0
        version: 0.6.0
      - name: limitador-operator.v0.5.0
        version: 0.5.0
      - name: limitador-operator.v0.4.0
        version: 0.4.0
      - name: limitador-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: limitador-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/limitador-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Commvault
      provider-url: ""
    name: hedvig-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hedvig-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\":\"hedvig.io/v1\",\n    \"kind\":\"HedvigDeploy\",\n
            \   \"metadata\": {\n      \"name\": \"testdeploy\",\n      \"namespace\":
            \"hedvig\"\n    },\n    \"spec\": {\n      \"cluster\": {\n        \"name\":
            \"hedvigcl\",\n        \"seeds\": [\n          \"hedvigcl1.hedvig.io\",\n
            \         \"hedvigcl2.hedvig.io\",\n          \"hedvigcl3.hedvig.io\"\n
            \       ]\n      },\n      \"k8s\": {\n        \"name\": \"k8scl\"\n      },\n
            \     \"proxy\": {\n        \"repository\": \"hedvig/hedvig-proxy\",\n
            \       \"tag\": \"v-4.3.1\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"controller\": {\n        \"repository\": \"hedvig/hedvig-csi-volume-driver\",\n
            \       \"tag\": \"v-2.0\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"node\": {\n        \"repository\": \"hedvig/hedvig-csi-volume-driver\",\n
            \       \"tag\": \"v-2.0\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"sidecars\": [\n        {\n          \"name\": \"csi-provisioner\",\n
            \         \"repository\": \"quay.io/k8scsi/csi-provisioner\",\n          \"tag\":
            \"v1.6.0\",\n          \"imagePullPolicy\": \"IfNotPresent\"\n        },\n
            \       {\n          \"name\": \"csi-snapshotter\",\n          \"repository\":
            \"quay.io/k8scsi/csi-snapshotter\",\n          \"tag\": \"v2.1.0\",\n
            \         \"imagePullPolicy\": \"IfNotPresent\"\n        },              \n
            \       {\n          \"name\": \"csi-resizer\",\n          \"repository\":
            \"quay.io/k8scsi/csi-resizer\",\n          \"tag\": \"v0.5.0\",\n          \"imagePullPolicy\":
            \"IfNotPresent\"\n        },              \n        {\n          \"name\":
            \"csi-node-driver-registrar\",\n          \"repository\": \"quay.io/k8scsi/csi-node-driver-registrar\",\n
            \         \"tag\": \"v1.0-canary\",\n          \"imagePullPolicy\": \"IfNotPresent\"\n
            \       }\n      ]\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: hedvig/hedvig-operator:2.0.0-ubi
          createdAt: ""
          description: Hedvig Operator is the official operator to deploy and manage
            Hedvig components in container orchestrators such as Kubernetes and Openshift.
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hedvig Deploy
            displayName: HedvigDeploy
            kind: HedvigDeploy
            name: hedvigdeploys.hedvig.io
            version: v1
        description: "The Hedvig Distributed Storage Platform is a software-defined,
          scale-out storage solution. It distributes data simultaneously across multiple
          locations – from on-premises data centers to the cloud – and scales capacity
          on-demand by leveraging the storage of commodity x86/ARM servers. The distributed
          write maximizes availability and protects data from hardware failures ranging
          from a single disk to an entire site, improving disaster recovery planning.\n\nHedvig
          provides a CSI Driver for accelerating your journey into the container ecosystem.
          CSI Driver enables containerized applications to maintain state by dynamically
          provisioning and consuming Hedvig virtual disks as persistent volumes. \n###
          \n\nHedvig Operator deploys the Hedvig Storage Proxy and the Hedvig CSI
          Driver to manage the life cycle of persistent storage in Kubernetes and
          Openshift.\n\n**Supported Container Orchestrators:**\n* Kubernetes 1.16+\n*
          Red Hat OpenShift 4.6+\n\n**Supported Hedvig Releases:**\n* Hedvig 4.0+\nPlease
          refer to the [documentation](https://documentation.commvault.com/commvault/hedvig/others/pdf/Hedvig_Operator_User_Guide.pdf)
          for steps to install and use the operator.\n\n\n\n\n\n"
        displayName: Hedvig Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - CSI
        - Hedvig
        - Cloud
        - Block
        - NFS
        - iSCSI
        - Snapshot
        - Volume
        links:
        - name: Documentation
          url: https://documentation.commvault.com/commvault/hedvig/article
        - name: Hedvig CSI User Guide
          url: https://documentation.commvault.com/commvault/hedvig/others/pdf/Hedvig_CSI_User_Guide.pdf
        maintainers:
        - email: ashenoy@commvault.com
          name: Abhijith Shenoy
        maturity: stable
        provider:
          name: Commvault
        relatedImages:
        - hedvig/hedvig-operator:2.0.0-ubi
        version: 1.0.2
      entries:
      - name: hedvig-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: hedvig-operator
    provider:
      name: Commvault
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: External Secrets
      provider-url: https://external-secrets.io
    name: external-secrets-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: external-secrets-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": "1234"
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "key": "provider-key",
                      "property": "provider-key-property",
                      "version": "provider-key-version"
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterExternalSecret",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "externalSecretName": "hello-world-es",
                  "externalSecretSpec": {
                    "data": [
                      {
                        "remoteRef": {
                          "key": "provider-key",
                          "property": "provider-key-property",
                          "version": "provider-key-version"
                        },
                        "secretKey": "secret-key-to-be-managed"
                      }
                    ],
                    "dataFrom": [
                      {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    ],
                    "refreshInterval": "1h",
                    "secretStoreRef": {
                      "kind": "SecretStore",
                      "name": "secret-store-name"
                    },
                    "target": {
                      "creationPolicy": "Merge",
                      "name": "my-secret",
                      "template": {
                        "data": {
                          "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                        },
                        "metadata": {
                          "annotations": {},
                          "labels": {}
                        },
                        "templateFrom": [
                          {
                            "configMap": {
                              "items": [
                                {
                                  "key": "alertmanager.yaml"
                                }
                              ],
                              "name": "alertmanager"
                            }
                          }
                        ],
                        "type": "kubernetes.io/dockerconfigjson"
                      }
                    }
                  },
                  "namespaceSelector": {
                    "matchLabels": {
                      "cool": "label"
                    }
                  },
                  "refreshTime": "1m"
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": 1234
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "extract": {
                        "conversionStrategy": "Default",
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    },
                    {
                      "find": {
                        "conversionStrategy": "Unicode",
                        "name": {
                          "regexp": ".*foobar.*"
                        },
                        "path": "path-to-filter",
                        "tags": {
                          "foo": "bar"
                        }
                      }
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "deletionPolicy": "Retain",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.external-secrets.io/v1alpha1",
                "kind": "OperatorConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "requeueInterval": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  },
                  "concurrent": 1,
                  "controllerClass": "",
                  "crds": {
                    "createClusterExternalSecret": true,
                    "createClusterSecretStore": true
                  },
                  "createOperator": true,
                  "deploymentAnnotations": {},
                  "extraArgs": {},
                  "extraEnv": [],
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/external-secrets/external-secrets",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "installCRDs": false,
                  "leaderElect": false,
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "priorityClassName": "",
                  "processClusterExternalSecret": true,
                  "processClusterStore": true,
                  "prometheus": {
                    "enabled": false,
                    "service": {
                      "port": 8080
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "scopedNamespace": "",
                  "scopedRBAC": false,
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "certCheckInterval": "5m",
                    "certDir": "/tmp/certs",
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
          createdAt: "2024-12-03T10:44:19Z"
          description: Operator to configure external-secrets helm-chart based operator
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/external-secrets/external-secrets-helm-operator
          support: external-secrets
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ACRAccessToken
            name: acraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster scoped resource that allows you to manage ExternalSecrets
              across namespaces
            displayName: ClusterExternalSecret
            kind: ClusterExternalSecret
            name: clusterexternalsecrets.external-secrets.io
            version: v1beta1
          - kind: ClusterGenerator
            name: clustergenerators.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1alpha1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1beta1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1beta1
          - kind: ECRAuthorizationToken
            name: ecrauthorizationtokens.generators.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1alpha1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1beta1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1beta1
          - kind: Fake
            name: fakes.generators.external-secrets.io
            version: v1alpha1
          - kind: GCRAccessToken
            name: gcraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - kind: GithubAccessToken
            name: githubaccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Configures external-secrets helm-chart based operator
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.external-secrets.io
            version: v1alpha1
          - kind: Password
            name: passwords.generators.external-secrets.io
            version: v1alpha1
          - kind: PushSecret
            name: pushsecrets.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1alpha1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1beta1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1beta1
          - kind: STSSessionToken
            name: stssessiontokens.generators.external-secrets.io
            version: v1alpha1
          - kind: UUID
            name: uuids.generators.external-secrets.io
            version: v1alpha1
          - kind: VaultDynamicSecret
            name: vaultdynamicsecrets.generators.external-secrets.io
            version: v1alpha1
          - kind: Webhook
            name: webhooks.generators.external-secrets.io
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK (Helm version) to configure **[official external-secrets operator helm chart](https://github.com/external-secrets/external-secrets)**, so it can be installed via OLM without having to do any change on current Helm Charts.

          Before any other resources provided by this Operator can be deployed, it is essential to create an
          OperatorConfig resource.

          The usual Helm Chart file `values.yaml`, like:
          ```yaml
          prometheus:
            enabled: true
            service:
              port: 8080
          resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          need to be encapsulated into a new custom resource called `OperatorConfig`:
          ```yaml
          apiVersion: operator.external-secrets.io/v1alpha1
          kind: OperatorConfig
          metadata:
            name: cluster
          spec:
            prometheus:
              enabled: true
              service:
                port: 8080
            resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          Once the OperatorConfig resource is deployed, the operator will create all helm chart resources, using the custom resource name as a preffix for all resources names, like for example a `Deployment` called `cluster-external-secrets`.

          ## Documentation

          * [Upstream Documentation](https://external-secrets.io)
          * [Helm Operator Install](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/install.md)
          * [Helm Operator Development](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/development.md)
          * [Helm Operator Release](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/release.md)

          ## Contributing

          You can contribute by:

          * Raising any issues you find using External Secrets Operator
          * Fixing issues by opening [Pull Requests](https://github.com/external-secrets/external-secrets-helm-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving documentation
          * Talking about External Secrets Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/external-secrets/external-secrets-helm-operator/issues).

          ## License

          External Secrets Operator is under Apache 2.0 license. See the [LICENSE](LICENSE) file for details.
        displayName: External Secrets Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secrets
        - secrets-manager
        - vault
        - aws
        - azure
        - google
        - ibm
        - akeyless
        - yandex
        - gitlab
        - oracle
        links:
        - name: External Secrets Operator
          url: https://external-secrets.io
        - name: GitHub
          url: https://github.com/external-secrets/external-secrets
        - name: Slack
          url: https://kubernetes.slack.com/messages/external-secrets
        maintainers:
        - email: contact@external-secrets.io
          name: external-secrets community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: External Secrets
          url: https://external-secrets.io
        relatedImages:
        - ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
        version: 0.11.0
      entries:
      - name: external-secrets-operator.v0.11.0
        version: 0.11.0
      - name: external-secrets-operator.v0.10.7
        version: 0.10.7
      - name: external-secrets-operator.v0.10.6
        version: 0.10.6
      - name: external-secrets-operator.v0.10.5
        version: 0.10.5
      - name: external-secrets-operator.v0.10.4
        version: 0.10.4
      - name: external-secrets-operator.v0.10.3
        version: 0.10.3
      - name: external-secrets-operator.v0.10.2
        version: 0.10.2
      - name: external-secrets-operator.v0.10.1
        version: 0.10.1
      - name: external-secrets-operator.v0.10.0
        version: 0.10.0
      - name: external-secrets-operator.v0.9.20
        version: 0.9.20
      - name: external-secrets-operator.v0.9.19
        version: 0.9.19
      - name: external-secrets-operator.v0.9.18
        version: 0.9.18
      - name: external-secrets-operator.v0.9.17
        version: 0.9.17
      - name: external-secrets-operator.v0.9.16
        version: 0.9.16
      - name: external-secrets-operator.v0.9.15-2
        version: 0.9.15-2
      - name: external-secrets-operator.v0.9.14
        version: 0.9.14
      - name: external-secrets-operator.v0.9.13
        version: 0.9.13
      - name: external-secrets-operator.v0.9.12
        version: 0.9.12
      - name: external-secrets-operator.v0.9.11
        version: 0.9.11
      - name: external-secrets-operator.v0.9.10
        version: 0.9.10
      - name: external-secrets-operator.v0.9.9
        version: 0.9.9
      - name: external-secrets-operator.v0.9.8
        version: 0.9.8
      - name: external-secrets-operator.v0.9.7
        version: 0.9.7
      - name: external-secrets-operator.v0.9.6
        version: 0.9.6
      - name: external-secrets-operator.v0.9.5
        version: 0.9.5
      - name: external-secrets-operator.v0.9.4
        version: 0.9.4
      - name: external-secrets-operator.v0.9.3
        version: 0.9.3
      - name: external-secrets-operator.v0.9.2
        version: 0.9.2
      - name: external-secrets-operator.v0.9.1
        version: 0.9.1
      - name: external-secrets-operator.v0.9.0
        version: 0.9.0
      - name: external-secrets-operator.v0.8.7
        version: 0.8.7
      - name: external-secrets-operator.v0.8.6
        version: 0.8.6
      - name: external-secrets-operator.v0.8.5
        version: 0.8.5
      - name: external-secrets-operator.v0.8.4
        version: 0.8.4
      - name: external-secrets-operator.v0.8.3
        version: 0.8.3
      - name: external-secrets-operator.v0.8.2
        version: 0.8.2
      - name: external-secrets-operator.v0.8.1
        version: 0.8.1
      - name: external-secrets-operator.v0.8.0
        version: 0.8.0
      - name: external-secrets-operator.v0.7.2
        version: 0.7.2
      - name: external-secrets-operator.v0.7.1
        version: 0.7.1
      - name: external-secrets-operator.v0.7.0
        version: 0.7.0
      - name: external-secrets-operator.v0.7.0-rc1
        version: 0.7.0-rc1
      - name: external-secrets-operator.v0.6.1
        version: 0.6.1
      - name: external-secrets-operator.v0.6.0
        version: 0.6.0
      - name: external-secrets-operator.v0.6.0-rc1
        version: 0.6.0-rc1
      - name: external-secrets-operator.v0.5.9
        version: 0.5.9
      - name: external-secrets-operator.v0.5.8
        version: 0.5.8
      - name: external-secrets-operator.v0.5.7
        version: 0.5.7
      - name: external-secrets-operator.v0.5.6
        version: 0.5.6
      - name: external-secrets-operator.v0.5.5
        version: 0.5.5
      - name: external-secrets-operator.v0.5.4
        version: 0.5.4
      - name: external-secrets-operator.v0.5.3
        version: 0.5.3
      - name: external-secrets-operator.v0.5.2
        version: 0.5.2
      - name: external-secrets-operator.v0.5.1
        version: 0.5.1
      - name: external-secrets-operator.v0.5.0
        version: 0.5.0
      - name: external-secrets-operator.v0.4.4
        version: 0.4.4
      - name: external-secrets-operator.v0.4.3
        version: 0.4.3
      - name: external-secrets-operator.v0.4.2
        version: 0.4.2
      - name: external-secrets-operator.v0.4.1
        version: 0.4.1
      - name: external-secrets-operator.v0.3.11
        version: 0.3.11
      - name: external-secrets-operator.v0.3.10
        version: 0.3.10
      - name: external-secrets-operator.v0.3.9
        version: 0.3.9
      - name: external-secrets-operator.v0.3.8
        version: 0.3.8
      name: alpha
    - currentCSV: external-secrets-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": "1234"
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "key": "provider-key",
                      "property": "provider-key-property",
                      "version": "provider-key-version"
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterExternalSecret",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "externalSecretName": "hello-world-es",
                  "externalSecretSpec": {
                    "data": [
                      {
                        "remoteRef": {
                          "key": "provider-key",
                          "property": "provider-key-property",
                          "version": "provider-key-version"
                        },
                        "secretKey": "secret-key-to-be-managed"
                      }
                    ],
                    "dataFrom": [
                      {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    ],
                    "refreshInterval": "1h",
                    "secretStoreRef": {
                      "kind": "SecretStore",
                      "name": "secret-store-name"
                    },
                    "target": {
                      "creationPolicy": "Merge",
                      "name": "my-secret",
                      "template": {
                        "data": {
                          "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                        },
                        "metadata": {
                          "annotations": {},
                          "labels": {}
                        },
                        "templateFrom": [
                          {
                            "configMap": {
                              "items": [
                                {
                                  "key": "alertmanager.yaml"
                                }
                              ],
                              "name": "alertmanager"
                            }
                          }
                        ],
                        "type": "kubernetes.io/dockerconfigjson"
                      }
                    }
                  },
                  "namespaceSelector": {
                    "matchLabels": {
                      "cool": "label"
                    }
                  },
                  "refreshTime": "1m"
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": 1234
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "extract": {
                        "conversionStrategy": "Default",
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    },
                    {
                      "find": {
                        "conversionStrategy": "Unicode",
                        "name": {
                          "regexp": ".*foobar.*"
                        },
                        "path": "path-to-filter",
                        "tags": {
                          "foo": "bar"
                        }
                      }
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "deletionPolicy": "Retain",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.external-secrets.io/v1alpha1",
                "kind": "OperatorConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "requeueInterval": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  },
                  "concurrent": 1,
                  "controllerClass": "",
                  "crds": {
                    "createClusterExternalSecret": true,
                    "createClusterSecretStore": true
                  },
                  "createOperator": true,
                  "deploymentAnnotations": {},
                  "extraArgs": {},
                  "extraEnv": [],
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/external-secrets/external-secrets",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "installCRDs": false,
                  "leaderElect": false,
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "priorityClassName": "",
                  "processClusterExternalSecret": true,
                  "processClusterStore": true,
                  "prometheus": {
                    "enabled": false,
                    "service": {
                      "port": 8080
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "scopedNamespace": "",
                  "scopedRBAC": false,
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "certCheckInterval": "5m",
                    "certDir": "/tmp/certs",
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
          createdAt: "2024-12-03T10:44:19Z"
          description: Operator to configure external-secrets helm-chart based operator
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/external-secrets/external-secrets-helm-operator
          support: external-secrets
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ACRAccessToken
            name: acraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster scoped resource that allows you to manage ExternalSecrets
              across namespaces
            displayName: ClusterExternalSecret
            kind: ClusterExternalSecret
            name: clusterexternalsecrets.external-secrets.io
            version: v1beta1
          - kind: ClusterGenerator
            name: clustergenerators.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1alpha1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1beta1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1beta1
          - kind: ECRAuthorizationToken
            name: ecrauthorizationtokens.generators.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1alpha1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1beta1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1beta1
          - kind: Fake
            name: fakes.generators.external-secrets.io
            version: v1alpha1
          - kind: GCRAccessToken
            name: gcraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - kind: GithubAccessToken
            name: githubaccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Configures external-secrets helm-chart based operator
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.external-secrets.io
            version: v1alpha1
          - kind: Password
            name: passwords.generators.external-secrets.io
            version: v1alpha1
          - kind: PushSecret
            name: pushsecrets.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1alpha1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1beta1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1beta1
          - kind: STSSessionToken
            name: stssessiontokens.generators.external-secrets.io
            version: v1alpha1
          - kind: UUID
            name: uuids.generators.external-secrets.io
            version: v1alpha1
          - kind: VaultDynamicSecret
            name: vaultdynamicsecrets.generators.external-secrets.io
            version: v1alpha1
          - kind: Webhook
            name: webhooks.generators.external-secrets.io
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK (Helm version) to configure **[official external-secrets operator helm chart](https://github.com/external-secrets/external-secrets)**, so it can be installed via OLM without having to do any change on current Helm Charts.

          Before any other resources provided by this Operator can be deployed, it is essential to create an
          OperatorConfig resource.

          The usual Helm Chart file `values.yaml`, like:
          ```yaml
          prometheus:
            enabled: true
            service:
              port: 8080
          resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          need to be encapsulated into a new custom resource called `OperatorConfig`:
          ```yaml
          apiVersion: operator.external-secrets.io/v1alpha1
          kind: OperatorConfig
          metadata:
            name: cluster
          spec:
            prometheus:
              enabled: true
              service:
                port: 8080
            resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          Once the OperatorConfig resource is deployed, the operator will create all helm chart resources, using the custom resource name as a preffix for all resources names, like for example a `Deployment` called `cluster-external-secrets`.

          ## Documentation

          * [Upstream Documentation](https://external-secrets.io)
          * [Helm Operator Install](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/install.md)
          * [Helm Operator Development](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/development.md)
          * [Helm Operator Release](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/release.md)

          ## Contributing

          You can contribute by:

          * Raising any issues you find using External Secrets Operator
          * Fixing issues by opening [Pull Requests](https://github.com/external-secrets/external-secrets-helm-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving documentation
          * Talking about External Secrets Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/external-secrets/external-secrets-helm-operator/issues).

          ## License

          External Secrets Operator is under Apache 2.0 license. See the [LICENSE](LICENSE) file for details.
        displayName: External Secrets Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secrets
        - secrets-manager
        - vault
        - aws
        - azure
        - google
        - ibm
        - akeyless
        - yandex
        - gitlab
        - oracle
        links:
        - name: External Secrets Operator
          url: https://external-secrets.io
        - name: GitHub
          url: https://github.com/external-secrets/external-secrets
        - name: Slack
          url: https://kubernetes.slack.com/messages/external-secrets
        maintainers:
        - email: contact@external-secrets.io
          name: external-secrets community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: External Secrets
          url: https://external-secrets.io
        relatedImages:
        - ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
        version: 0.11.0
      entries:
      - name: external-secrets-operator.v0.11.0
        version: 0.11.0
      - name: external-secrets-operator.v0.10.7
        version: 0.10.7
      - name: external-secrets-operator.v0.10.6
        version: 0.10.6
      - name: external-secrets-operator.v0.10.5
        version: 0.10.5
      - name: external-secrets-operator.v0.10.4
        version: 0.10.4
      - name: external-secrets-operator.v0.10.3
        version: 0.10.3
      - name: external-secrets-operator.v0.10.2
        version: 0.10.2
      - name: external-secrets-operator.v0.10.1
        version: 0.10.1
      - name: external-secrets-operator.v0.10.0
        version: 0.10.0
      - name: external-secrets-operator.v0.9.20
        version: 0.9.20
      - name: external-secrets-operator.v0.9.19
        version: 0.9.19
      - name: external-secrets-operator.v0.9.18
        version: 0.9.18
      - name: external-secrets-operator.v0.9.17
        version: 0.9.17
      - name: external-secrets-operator.v0.9.16
        version: 0.9.16
      - name: external-secrets-operator.v0.9.15-2
        version: 0.9.15-2
      - name: external-secrets-operator.v0.9.14
        version: 0.9.14
      - name: external-secrets-operator.v0.9.13
        version: 0.9.13
      - name: external-secrets-operator.v0.9.12
        version: 0.9.12
      - name: external-secrets-operator.v0.9.11
        version: 0.9.11
      - name: external-secrets-operator.v0.9.10
        version: 0.9.10
      - name: external-secrets-operator.v0.9.9
        version: 0.9.9
      - name: external-secrets-operator.v0.9.8
        version: 0.9.8
      - name: external-secrets-operator.v0.9.7
        version: 0.9.7
      - name: external-secrets-operator.v0.9.6
        version: 0.9.6
      - name: external-secrets-operator.v0.9.5
        version: 0.9.5
      - name: external-secrets-operator.v0.9.4
        version: 0.9.4
      - name: external-secrets-operator.v0.9.3
        version: 0.9.3
      - name: external-secrets-operator.v0.9.2
        version: 0.9.2
      - name: external-secrets-operator.v0.9.1
        version: 0.9.1
      - name: external-secrets-operator.v0.9.0
        version: 0.9.0
      - name: external-secrets-operator.v0.8.7
        version: 0.8.7
      - name: external-secrets-operator.v0.8.6
        version: 0.8.6
      - name: external-secrets-operator.v0.8.5
        version: 0.8.5
      - name: external-secrets-operator.v0.8.4
        version: 0.8.4
      - name: external-secrets-operator.v0.8.3
        version: 0.8.3
      - name: external-secrets-operator.v0.8.2
        version: 0.8.2
      - name: external-secrets-operator.v0.8.1
        version: 0.8.1
      - name: external-secrets-operator.v0.8.0
        version: 0.8.0
      - name: external-secrets-operator.v0.7.2
        version: 0.7.2
      - name: external-secrets-operator.v0.7.1
        version: 0.7.1
      - name: external-secrets-operator.v0.7.0
        version: 0.7.0
      - name: external-secrets-operator.v0.7.0-rc1
        version: 0.7.0-rc1
      - name: external-secrets-operator.v0.6.1
        version: 0.6.1
      - name: external-secrets-operator.v0.6.0
        version: 0.6.0
      - name: external-secrets-operator.v0.5.9
        version: 0.5.9
      - name: external-secrets-operator.v0.5.8
        version: 0.5.8
      - name: external-secrets-operator.v0.5.7
        version: 0.5.7
      - name: external-secrets-operator.v0.5.6
        version: 0.5.6
      - name: external-secrets-operator.v0.5.5
        version: 0.5.5
      - name: external-secrets-operator.v0.5.4
        version: 0.5.4
      - name: external-secrets-operator.v0.5.3
        version: 0.5.3
      - name: external-secrets-operator.v0.5.2
        version: 0.5.2
      - name: external-secrets-operator.v0.5.1
        version: 0.5.1
      - name: external-secrets-operator.v0.5.0
        version: 0.5.0
      - name: external-secrets-operator.v0.4.4
        version: 0.4.4
      - name: external-secrets-operator.v0.4.2
        version: 0.4.2
      - name: external-secrets-operator.v0.4.1
        version: 0.4.1
      - name: external-secrets-operator.v0.3.11
        version: 0.3.11
      - name: external-secrets-operator.v0.3.10
        version: 0.3.10
      - name: external-secrets-operator.v0.3.9
        version: 0.3.9
      name: stable
    defaultChannel: alpha
    packageName: external-secrets-operator
    provider:
      name: External Secrets
      url: https://external-secrets.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/kuadrant-operator
    name: rhcl-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhcl-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "authpolicy-sample"
                },
                "spec": {
                  "rules": {
                    "authentication": {
                      "apikey": {
                        "apiKey": {
                          "selector": {}
                        },
                        "credentials": {
                          "authorizationHeader": {
                            "prefix": "APIKEY"
                          }
                        }
                      }
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "DNSPolicy",
                "metadata": {
                  "name": "dnspolicy-sample"
                },
                "spec": {
                  "healthCheck": {
                    "protocol": "HTTP"
                  },
                  "providerRefs": [
                    {
                      "name": "provider-ref"
                    }
                  ],
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "ratelimitpolicy-sample"
                },
                "spec": {
                  "limits": {
                    "toys": {
                      "rates": [
                        {
                          "limit": 50,
                          "window": "1m"
                        }
                      ]
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "TLSPolicy",
                "metadata": {
                  "name": "tlspolicy-sample"
                },
                "spec": {
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "self-signed-ca"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/rhcl-rhel9-operator@sha256:92f14fd8de74bc72c407947cf5fa4739a7b455e5d351baeb64521e4c07f02a9e
          createdAt: "2025-04-28T13:10:44Z"
          description: Red Hat Connectivity Link enables you to secure, protect, and
            connect your APIs and applications in multicluster, multicloud, and hybrid
            cloud environments
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthPolicy enables authentication and authorization for service
              workloads in a Gateway API network
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1
          - description: DNSPolicy configures how North-South based traffic should
              be balanced and reach the gateways
            displayName: DNSPolicy
            kind: DNSPolicy
            name: dnspolicies.kuadrant.io
            version: v1
          - description: Kuadrant configures installations of Kuadrant Service Protection
              components
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: RateLimitPolicy enables rate limiting for service workloads
              in a Gateway API network
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1
          - description: TLSPolicy provides tls for gateway listeners by managing
              the lifecycle of tls certificates
            displayName: TLSPolicy
            kind: TLSPolicy
            name: tlspolicies.kuadrant.io
            version: v1
        description: Red Hat Connectivity Link enables you to secure, protect, and
          connect your APIs and applications in multicluster, multicloud, and hybrid
          cloud environments
        displayName: Red Hat Connectivity Link
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: GitHub
          url: https://github.com/kuadrant/kuadrant-operator
        - name: Documentation
          url: https://github.com/Kuadrant/kuadrant-operator/tree/main/docs
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - registry.redhat.io/rhcl-1/rhcl-console-plugin-rhel9@sha256:ae92d638d993d782986e761d5aa3c7c5da2b8af0aabc1decb332a4f1487dacd3
        - registry.redhat.io/rhcl-1/rhcl-rhel9-operator@sha256:92f14fd8de74bc72c407947cf5fa4739a7b455e5d351baeb64521e4c07f02a9e
        - registry.redhat.io/rhcl-1/wasm-shim-rhel9@sha256:845bb8af57f3d219aa09b9c0bb20fa945306f3b413e34e196760c8cdf624a532
        version: 1.1.0
      entries:
      - name: rhcl-operator.v1.1.0
        version: 1.1.0
      - name: rhcl-operator.v1.0.2
        version: 1.0.2
      - name: rhcl-operator.v1.0.1
        version: 1.0.1
      - name: rhcl-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: rhcl-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/kuadrant-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: akaris
      provider-url: https://github.com/andreaskaris/sosreport-operator
    name: sosreport-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sosreport-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "support.openshift.io/v1alpha1",
                "kind": "Sosreport",
                "metadata": {
                  "name": "sosreport-sample"
                },
                "spec": {
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Sosreport is the Schema for the sosreports API
            displayName: Sosreport
            kind: Sosreport
            name: sosreports.support.openshift.io
            version: v1alpha1
        description: Operator that facilitates the generation of sosreports
        displayName: sosreport-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sosreport
        - debugging
        - troubleshooting
        links:
        - name: Sosreport Operator
          url: https://github.com/andreaskaris/sosreport-operator
        maintainers:
        - email: ak.karis@gmail.com
          name: Andreas Karis
        maturity: alpha
        provider:
          name: akaris
          url: https://github.com/andreaskaris/sosreport-operator
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/akaris/sosreport-operator:0.0.2
        version: 0.0.2
      entries:
      - name: sosreport-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: sosreport-operator
    provider:
      name: akaris
      url: https://github.com/andreaskaris/sosreport-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-rds-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-rds-controller.v1.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rds.services.k8s.aws/v1alpha1",
                "kind": "DBParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "rds.services.k8s.aws/v1alpha1",
                "kind": "DBSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/rds-controller:1.6.2
          createdAt: "2025-09-22T17:27:42Z"
          description: AWS RDS controller is a service controller for managing RDS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DBClusterEndpoint represents the state of an AWS rds DBClusterEndpoint
              resource.
            displayName: DBClusterEndpoint
            kind: DBClusterEndpoint
            name: dbclusterendpoints.rds.services.k8s.aws
            version: v1alpha1
          - description: DBClusterParameterGroup represents the state of an AWS rds
              DBClusterParameterGroup resource.
            displayName: DBClusterParameterGroup
            kind: DBClusterParameterGroup
            name: dbclusterparametergroups.rds.services.k8s.aws
            version: v1alpha1
          - description: DBCluster represents the state of an AWS rds DBCluster resource.
            displayName: DBCluster
            kind: DBCluster
            name: dbclusters.rds.services.k8s.aws
            version: v1alpha1
          - description: DBClusterSnapshot represents the state of an AWS rds DBClusterSnapshot
              resource.
            displayName: DBClusterSnapshot
            kind: DBClusterSnapshot
            name: dbclustersnapshots.rds.services.k8s.aws
            version: v1alpha1
          - description: DBInstance represents the state of an AWS rds DBInstance
              resource.
            displayName: DBInstance
            kind: DBInstance
            name: dbinstances.rds.services.k8s.aws
            version: v1alpha1
          - description: DBParameterGroup represents the state of an AWS rds DBParameterGroup
              resource.
            displayName: DBParameterGroup
            kind: DBParameterGroup
            name: dbparametergroups.rds.services.k8s.aws
            version: v1alpha1
          - description: DBProxy represents the state of an AWS rds DBProxy resource.
            displayName: DBProxy
            kind: DBProxy
            name: dbproxies.rds.services.k8s.aws
            version: v1alpha1
          - description: DBSnapshot represents the state of an AWS rds DBSnapshot
              resource.
            displayName: DBSnapshot
            kind: DBSnapshot
            name: dbsnapshots.rds.services.k8s.aws
            version: v1alpha1
          - description: DBSubnetGroup represents the state of an AWS rds DBSubnetGroup
              resource.
            displayName: DBSubnetGroup
            kind: DBSubnetGroup
            name: dbsubnetgroups.rds.services.k8s.aws
            version: v1alpha1
          - description: GlobalCluster represents the state of an AWS rds GlobalCluster
              resource.
            displayName: GlobalCluster
            kind: GlobalCluster
            name: globalclusters.rds.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Relational Database Service ("RDS") resources in AWS from within your Kubernetes cluster.

          **About Amazon RDS**

          Amazon Relational Database Service (Amazon RDS) makes it easy to set up, operate, and scale a relational database in the cloud. It provides cost-efficient and resizable capacity while automating time-consuming administration tasks such as hardware provisioning, database setup, patching and backups. It frees you to focus on your applications so you can give them the fast performance, high availability, security and compatibility they need.

          Amazon RDS is available on several database instance types - optimized for memory, performance or I/O -and provides you with six familiar database engines to choose from, including Amazon Aurora, PostgreSQL, MySQL, MariaDB, Oracle Database, and SQL Server. You can use the AWS Database Migration Service to easily migrate or replicate your existing databases to Amazon RDS.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon RDS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rds
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon RDS Developer Resources
          url: https://aws.amazon.com/rds/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: rds maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/rds-controller:1.6.2
        version: 1.6.2
      entries:
      - name: ack-rds-controller.v1.6.2
        version: 1.6.2
      - name: ack-rds-controller.v1.6.1
        version: 1.6.1
      - name: ack-rds-controller.v1.6.0
        version: 1.6.0
      - name: ack-rds-controller.v1.5.0
        version: 1.5.0
      - name: ack-rds-controller.v1.4.26
        version: 1.4.26
      - name: ack-rds-controller.v1.4.25
        version: 1.4.25
      - name: ack-rds-controller.v1.4.24
        version: 1.4.24
      - name: ack-rds-controller.v1.4.22
        version: 1.4.22
      - name: ack-rds-controller.v1.4.21
        version: 1.4.21
      - name: ack-rds-controller.v1.4.20
        version: 1.4.20
      - name: ack-rds-controller.v1.4.19
        version: 1.4.19
      - name: ack-rds-controller.v1.4.18
        version: 1.4.18
      - name: ack-rds-controller.v1.4.17
        version: 1.4.17
      - name: ack-rds-controller.v1.4.16
        version: 1.4.16
      - name: ack-rds-controller.v1.4.15
        version: 1.4.15
      - name: ack-rds-controller.v1.4.14
        version: 1.4.14
      - name: ack-rds-controller.v1.4.13
        version: 1.4.13
      - name: ack-rds-controller.v1.4.12
        version: 1.4.12
      - name: ack-rds-controller.v1.4.11
        version: 1.4.11
      - name: ack-rds-controller.v1.4.10
        version: 1.4.10
      - name: ack-rds-controller.v1.4.8
        version: 1.4.8
      - name: ack-rds-controller.v1.4.7
        version: 1.4.7
      - name: ack-rds-controller.v1.4.6
        version: 1.4.6
      - name: ack-rds-controller.v1.4.4
        version: 1.4.4
      - name: ack-rds-controller.v1.4.3
        version: 1.4.3
      - name: ack-rds-controller.v1.4.2
        version: 1.4.2
      - name: ack-rds-controller.v1.4.1
        version: 1.4.1
      - name: ack-rds-controller.v1.4.0
        version: 1.4.0
      - name: ack-rds-controller.v1.3.0
        version: 1.3.0
      - name: ack-rds-controller.v1.2.5
        version: 1.2.5
      - name: ack-rds-controller.v1.2.4
        version: 1.2.4
      - name: ack-rds-controller.v1.2.3
        version: 1.2.3
      - name: ack-rds-controller.v1.2.2
        version: 1.2.2
      - name: ack-rds-controller.v1.2.1
        version: 1.2.1
      - name: ack-rds-controller.v1.2.0
        version: 1.2.0
      - name: ack-rds-controller.v1.1.11
        version: 1.1.11
      - name: ack-rds-controller.v1.1.9
        version: 1.1.9
      - name: ack-rds-controller.v1.1.8
        version: 1.1.8
      - name: ack-rds-controller.v1.1.7
        version: 1.1.7
      - name: ack-rds-controller.v1.1.6
        version: 1.1.6
      - name: ack-rds-controller.v1.1.5
        version: 1.1.5
      - name: ack-rds-controller.v1.1.4
        version: 1.1.4
      - name: ack-rds-controller.v1.1.3
        version: 1.1.3
      - name: ack-rds-controller.v1.1.2
        version: 1.1.2
      - name: ack-rds-controller.v1.1.1
        version: 1.1.1
      - name: ack-rds-controller.v1.1.0
        version: 1.1.0
      - name: ack-rds-controller.v1.0.0
        version: 1.0.0
      - name: ack-rds-controller.v0.1.3
        version: 0.1.3
      - name: ack-rds-controller.v0.1.1
        version: 0.1.1
      - name: ack-rds-controller.v0.1.0
        version: 0.1.0
      - name: ack-rds-controller.v0.0.30
        version: 0.0.30
      - name: ack-rds-controller.v0.0.29
        version: 0.0.29
      - name: ack-rds-controller.v0.0.28
        version: 0.0.28
      - name: ack-rds-controller.v0.0.27
        version: 0.0.27
      - name: ack-rds-controller.v0.0.26
        version: 0.0.26
      - name: ack-rds-controller.v0.0.25
        version: 0.0.25
      - name: ack-rds-controller.v0.0.24
        version: 0.0.24
      - name: ack-rds-controller.v0.0.23
        version: 0.0.23
      - name: ack-rds-controller.v0.0.22
        version: 0.0.22
      - name: ack-rds-controller.v0.0.21
        version: 0.0.21
      - name: ack-rds-controller.v0.0.20
        version: 0.0.20
      - name: ack-rds-controller.v0.0.19
        version: 0.0.19
      - name: ack-rds-controller.v0.0.18
        version: 0.0.18
      - name: ack-rds-controller.v0.0.17
        version: 0.0.17
      - name: ack-rds-controller.v0.0.16
        version: 0.0.16
      - name: ack-rds-controller.v0.0.15
        version: 0.0.15
      - name: ack-rds-controller.v0.0.14
        version: 0.0.14
      - name: ack-rds-controller.v0.0.13
        version: 0.0.13
      - name: ack-rds-controller.v0.0.12
        version: 0.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-rds-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: integrity-shield-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: integrity-shield-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/open-cluster-management/integrity-shield-observer",
                    "imagePullPolicy": "Always",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "provenanece": true,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  resp.status_code == 200\n  result := json.unmarshal(resp.raw_body)\n  result.allow == false\n  not is_detect_mode\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"tls_insecure_skip_verify\": true\n  })\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "requestHandlerConfig": "defaultConstraintAction:\n  audit:\n    inform: true\n  admissionControl:\n    enforce: false\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/open-cluster-management/integrity-shield-api",
                    "imagePullPolicy": "Always",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/open-cluster-management/integrity-shield-operator:0.3.0
          createdAt: "2021-10-14T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.

          [K8s Integrity Shield](https://github.com/open-cluster-management/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.

          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Detect (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster

          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).

          ## Preparations before installation

          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).


          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.

          If you want to change the settings such as default run mode (detection/enforcement) or audit interval,  please check [here](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).

          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-7b7f768bf7-fhrpg                           1/1     Running   0          20s
          integrity-shield-observer-5bc66f75f7-tn8fw                      1/1     Running   0          25s
          integrity-shield-operator-controller-manager-65b7fb58f7-j25zd   2/2     Running   0          3h5m
          ```

          After installation, you can protect Kubernetes resources by following this [document](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).

          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:

          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)

          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/open-cluster-management/integrity-shield
        - name: Operator Documentation
          url: https://github.com/open-cluster-management/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/open-cluster-management/integrity-shield-operator:0.3.0
        version: 0.3.0
      entries:
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.0
    - currentCSV: integrity-shield-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/open-cluster-management/integrity-shield-observer",
                    "imagePullPolicy": "Always",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  resp.status_code == 200\n  result := json.unmarshal(resp.raw_body)\n  result.allow == false\n  not is_detect_mode\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"tls_insecure_skip_verify\": true\n  })\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: detect\n  admissionOnly: false\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/open-cluster-management/integrity-shield-api",
                    "imagePullPolicy": "Always",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/open-cluster-management/integrity-shield-operator:0.3.1
          createdAt: "2021-10-22T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.

          [K8s Integrity Shield](https://github.com/open-cluster-management/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.

          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Detect (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster

          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).

          ## Preparations before installation

          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).


          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.

          If you want to change the settings such as default run mode (detection/enforcement) or audit interval,  please check [here](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).

          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-7b7f768bf7-fhrpg                           1/1     Running   0          20s
          integrity-shield-observer-5bc66f75f7-tn8fw                      1/1     Running   0          25s
          integrity-shield-operator-controller-manager-65b7fb58f7-j25zd   2/2     Running   0          3h5m
          ```

          After installation, you can protect Kubernetes resources by following this [document](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).

          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:

          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)

          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/open-cluster-management/integrity-shield
        - name: Operator Documentation
          url: https://github.com/open-cluster-management/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/open-cluster-management/integrity-shield-operator:0.3.1
        version: 0.3.1
      entries:
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.1
    - currentCSV: integrity-shield-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/stolostron/integrity-shield-observer",
                    "imagePullPolicy": "IfNotPresent",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  should_be_denied(resp)\n  not is_detect_mode\n  result := get_result(resp)\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"timeout\": \"2500ms\",\n    \"raise_error\": false,\n    \"tls_ca_cert\": REPLACE_WITH_CA_CERT\n  })\n}\n\n# get decision result from API response\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  not is_null(raw_body)\n  result := json.unmarshal(raw_body)\n}\n# if error, return it as result\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  is_null(raw_body)\n  error_msg := object.get(resp, \"error\", \"\")\n  result := {\"http_error\":error_msg}\n}\n\n# if API response is null, deny\nshould_be_denied(resp) {\n  is_null(resp)\n}\n\n# if API response has status_code other than 200, deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  resp.status_code != 200\n}\n\n# if decision result is \"allow == false\" or not set, then deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  result := get_result(resp)\n  allow := object.get(result, \"allow\", false)\n  allow == false\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "reporter": {
                    "image": "quay.io/stolostron/integrity-shield-reporter",
                    "imagePullPolicy": "IfNotPresent",
                    "intervalSeconds": "10",
                    "name": "integrity-shield-reporter",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: inform\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\ndecisionReporterConfig:\n  enabled: true\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/stolostron/integrity-shield-api",
                    "imagePullPolicy": "IfNotPresent",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/stolostron/integrity-shield-operator:0.3.2
          createdAt: "2022-01-25T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.15.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.
          [K8s Integrity Shield](https://github.com/stolostron/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.
          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Inform (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster
          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).
          ## Preparations before installation
          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).
          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.
          If you want to change the settings such as default run mode (inform/enforce) or audit interval,  please check [here](https://github.com/stolostron/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).
          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-5799f69896-zbcjs                           2/2     Running   0          88s
          integrity-shield-observer-757f58f7d7-7pf5h                      1/1     Running   0          55s
          integrity-shield-operator-controller-manager-6f94d7d684-dg5bt   2/2     Running   0          4m27s
          ```
          After installation, you can protect Kubernetes resources by following this [document](https://github.com/stolostron/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).
          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:
          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)
          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/stolostron/integrity-shield
        - name: Operator Documentation
          url: https://github.com/stolostron/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/stolostron/integrity-shield-operator:0.3.2
        version: 0.3.2
      entries:
      - name: integrity-shield-operator.v0.3.2
        version: 0.3.2
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.2
    - currentCSV: integrity-shield-operator.v0.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": false,
                    "image": "quay.io/stolostron/integrity-shield-observer",
                    "imagePullPolicy": "IfNotPresent",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  should_be_denied(resp)\n  not is_detect_mode\n  result := get_result(resp)\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"timeout\": \"2500ms\",\n    \"raise_error\": false,\n    \"tls_ca_cert\": REPLACE_WITH_CA_CERT\n  })\n}\n\n# get decision result from API response\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  not is_null(raw_body)\n  result := json.unmarshal(raw_body)\n}\n# if error, return it as result\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  is_null(raw_body)\n  error_msg := object.get(resp, \"error\", \"\")\n  result := {\"http_error\":error_msg}\n}\n\n# if API response is null, deny\nshould_be_denied(resp) {\n  is_null(resp)\n}\n\n# if API response has status_code other than 200, deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  resp.status_code != 200\n}\n\n# if decision result is \"allow == false\" or not set, then deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  result := get_result(resp)\n  allow := object.get(result, \"allow\", false)\n  allow == false\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "reporter": {
                    "image": "quay.io/stolostron/integrity-shield-reporter",
                    "imagePullPolicy": "IfNotPresent",
                    "intervalSeconds": "10",
                    "name": "integrity-shield-reporter",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: inform\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\ndecisionReporterConfig:\n  enabled: true\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/stolostron/integrity-shield-api",
                    "imagePullPolicy": "IfNotPresent",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/stolostron/integrity-shield-operator:0.3.3
          createdAt: "2022-03-18T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.15.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.
          [K8s Integrity Shield](https://github.com/stolostron/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.
          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Inform (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster
          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).
          ## Preparations before installation
          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).
          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.
          If you want to change the settings such as default run mode (inform/enforce) or audit interval,  please check [here](https://github.com/stolostron/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).
          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-5799f69896-zbcjs                           2/2     Running   0          88s
          integrity-shield-observer-757f58f7d7-7pf5h                      1/1     Running   0          55s
          integrity-shield-operator-controller-manager-6f94d7d684-dg5bt   2/2     Running   0          4m27s
          ```
          After installation, you can protect Kubernetes resources by following this [document](https://github.com/stolostron/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).
          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:
          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)
          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/stolostron/integrity-shield
        - name: Operator Documentation
          url: https://github.com/stolostron/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/stolostron/integrity-shield-operator:0.3.3
        version: 0.3.3
      entries:
      - name: integrity-shield-operator.v0.3.3
        version: 0.3.3
      - name: integrity-shield-operator.v0.3.2
        version: 0.3.2
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.3
    defaultChannel: alpha-0.3.3
    packageName: integrity-shield-operator
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datasance
      provider-url: ""
    name: datasance-pot-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datasance-pot-operator.v3.4.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasance.com/v3",
                "kind": "ControlPlane",
                "metadata": {
                  "name": "pot"
                },
                "spec": {
                  "replicas": {
                    "controller": 2
                  },
                  "database": {
                    "provider": "postgres",
                    "user": "user",
                    "host": "postgre.host",
                    "port": 5432,
                    "password": "password",
                    "databaseName": "dbname"
                  },
                  "auth": {
                    "url": "auth.url",
                    "realm": "realm",
                    "realmKey": "realmKey",
                    "ssl": "none",
                    "controllerClient": "controllerClient",
                    "controllerSecret": "controllerClientSecret",
                    "viewerClient": "viewerClient"
                  },
                  "images": {
                    "pullSecret": "pullsecret",
                    "controller": "ghcr.io/datasance/controller@sha256:aa896bae6c3e6eda70938a03aaa649226356d9e4a1d03ba427d4e4ee3090ca6c",
                    "portManager": "ghcr.io/datasance/port-manager@sha256:4bf3793e9e3cd10df9b46c459e983631393825ca788ffbbb3ec3659158ad61fe",
                    "proxy": "ghcr.io/datasance/proxy@sha256:276dbed0c8ef4f5b03fdf29d5c863b6f3efcae50444ce8123220c5673272860f",
                    "router": "ghcr.io/datasance/router@sha256:82657fc8f39d33d8ef0161cf27d705cbdc0a97dca07b22a5fe78fd4213ef0d31"
                  },
                  "services": {
                    "controller": {
                      "type": "LoadBalancer",
                      "annotations": {
                        "service.beta.kubernetes.io/azure-load-balancer-internal": "false"
                      }
                    },
                    "proxy": {
                      "type": "LoadBalancer",
                      "annotations": {
                        "service.beta.kubernetes.io/azure-load-balancer-internal": "false"
                      }
                    },
                    "router": {
                      "type": "LoadBalancer",
                      "annotations": {
                        "service.beta.kubernetes.io/azure-load-balancer-internal": "false"
                      }
                    }
                  },
                  "controller": {
                    "ecnViewerUrl": "pot.demo.datasance.com",
                    "https": true,
                    "secretName": "secretName"
                  },
                  "ingresses": {
                    "controller": {
                      "annotations": {
                        "cert-manager.io/cluster-issuer": "letsencrypt"
                      },
                      "ingressClassName": "nginx",
                      "host": "pot.demo.datasance.com",
                      "secretName": "secretName"
                    },
                    "router": {
                      "address": "pot.demo.datasance.com",
                      "messagePort": 5672,
                      "interiorPort": 55672,
                      "edgePort": 45672
                    },
                    "httpProxy": {
                      "address": "pot.demo.datasance.com"
                    },
                    "tcpProxy": {
                      "address": "pot.demo.datasance.com"
                    }
                  }
                },
                "status": {
                  "conditions": [
                    {
                      "lastTransitionTime": "2022-04-10T22:44:09Z",
                      "message": "",
                      "reason": "initial_status",
                      "status": "True",
                      "type": "deploying"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          containerImage: ghcr.io/datasance/operator@sha256:143b9307c2520ef16b40c4c92e6faee8b24b7fb3a97682231e53236ddd532687
          createdAt: "2024-11-04T20:45:43Z"
          description: Datasance PoT is an Enteprise version of Eclipse ioFog Platform.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Datasance PoT Enterprise Open
            Source Subscriptions based on number of Edge device that ControlPlane
            manages"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/datasance/iofog-operator
          support: Datasance Teknoloji A.S.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Datasance PoT supported ioFog ControlPlane CRD for managing
              container workloads on Edge device
            kind: ControlPlane
            name: controlplanes.datasance.com
            version: v3
        description: ioFog Control Plane on Kubernetes
        displayName: iofog-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iofog
        - datasance
        - pot
        links:
        - name: Datasance PoT // Iofog Operator
          url: https://docs.datasance.com
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Datasance
        relatedImages:
        - ghcr.io/datasance/operator@sha256:143b9307c2520ef16b40c4c92e6faee8b24b7fb3a97682231e53236ddd532687
        - ghcr.io/datasance/controller@sha256:aa896bae6c3e6eda70938a03aaa649226356d9e4a1d03ba427d4e4ee3090ca6c
        - ghcr.io/datasance/router@sha256:82657fc8f39d33d8ef0161cf27d705cbdc0a97dca07b22a5fe78fd4213ef0d31
        - ghcr.io/datasance/proxy@sha256:276dbed0c8ef4f5b03fdf29d5c863b6f3efcae50444ce8123220c5673272860f
        - ghcr.io/datasance/port-manager@sha256:4bf3793e9e3cd10df9b46c459e983631393825ca788ffbbb3ec3659158ad61fe
        version: 3.4.9
      entries:
      - name: datasance-pot-operator.v3.4.9
        version: 3.4.9
      name: stable
    defaultChannel: stable
    packageName: datasance-pot-operator
    provider:
      name: Datasance
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: gcp-filestore-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gcp-filestore-csi-driver-operator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel8-operator@sha256:1ec61d2988b95a10260c7932e9dc6a9d1d5b3eb4148b22afd438aa4ee75dc9b3
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure GCP Filestore CSI driver.
          olm.skipRange: '>=4.11.0-0 <4.13.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/gcp-filestore-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Operator that installs and configures the CSI driver for GCP Filestore. The GCP Filestore CSI Driver Operator is a Technology Preview feature. For more information on Red Hat Technology Preview features support scope, see https://access.redhat.com/support/offerings/techpreview/.
        displayName: GCP Filestore CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - filestore
        links:
        - name: Documentation
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: preview
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:700c0e50f86a70f506403aba75e020a17530ada284af7eee77d4b4ce3dbf4102
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:4f1e873080754ae9aebc361cd77f965e1ade369403346aa0cf4cf502082d18e4
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:ee45e11a6fe8cf60f3ac285cba1b23459ca95068bd9769042ae052c778c3e080
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:537fbc8fe2b212dc798361099d2c6fd981cc846c1cf5454b703ba08d488b61f8
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel8@sha256:a1a4e7b93d6609561a7a2cd732ce772bf0047448e5019298d66c4b985baecfb2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:767682dd3bd65651a8204c9fb732b9c48b99127189992b43abc6ecce027f4589
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel8-operator@sha256:1ec61d2988b95a10260c7932e9dc6a9d1d5b3eb4148b22afd438aa4ee75dc9b3
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:6f1b497772b953468f0b816face8810ace8ea44fa91dd4a3b3e583da2cac350c
        version: 4.13.0-202309181427
      entries:
      - name: gcp-filestore-csi-driver-operator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: preview
    - currentCSV: gcp-filestore-csi-driver-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9-operator@sha256:2dc13de58b391237dc48b9335d88b99f5e26d0e5edc5a664c7d39f16ca9b297d
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure GCP Filestore CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.11.0-0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/gcp-filestore-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Operator that installs and configures the CSI driver for GCP Filestore.
        displayName: GCP Filestore CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - filestore
        links:
        - name: Documentation
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:a5704d26d0b562dba9633d1c107f34e05c750c0e3a64dbcace4dab7b964fff8d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9@sha256:8e801d924071c8f2513555450e688bff60d4100b2ca56d2ba9517038a1d48d4d
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:9c3c3d2b857ce289324aee56df1ba4ab5e5ea1e76cd86ac988233463915834b4
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:78e02d937740d742eb33b08c7317f211c25020c375069d5a97344107716608d3
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:b28def976fc155dff57ff3003cc462a8e9923b325af0b319b65220c66688699e
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0627c09fe3f5a015f117fb30af1f4d17340aefa48b23994737d3901a804b63d3
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9-operator@sha256:2dc13de58b391237dc48b9335d88b99f5e26d0e5edc5a664c7d39f16ca9b297d
        version: 4.16.0-202509240123
      entries:
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202508270905
        version: 4.16.0-202508270905
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202504121405
        version: 4.16.0-202504121405
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: gcp-filestore-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-prometheusservice-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-prometheusservice-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "prometheusservice.services.k8s.aws/v1alpha1",
                "kind": "RuleGroupsNamespace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "prometheusservice.services.k8s.aws/v1alpha1",
                "kind": "Workspace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/prometheusservice-controller:1.3.1
          createdAt: "2025-09-19T21:11:38Z"
          description: AWS Prometheus controller is a service controller for managing
            Prometheus resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertManagerDefinition represents the state of an AWS prometheusservice
              AlertManagerDefinition resource.
            displayName: AlertManagerDefinition
            kind: AlertManagerDefinition
            name: alertmanagerdefinitions.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: LoggingConfiguration represents the state of an AWS prometheusservice
              LoggingConfiguration resource.
            displayName: LoggingConfiguration
            kind: LoggingConfiguration
            name: loggingconfigurations.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: RuleGroupsNamespace represents the state of an AWS prometheusservice
              RuleGroupsNamespace resource.
            displayName: RuleGroupsNamespace
            kind: RuleGroupsNamespace
            name: rulegroupsnamespaces.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: Workspace represents the state of an AWS prometheusservice
              Workspace resource.
            displayName: Workspace
            kind: Workspace
            name: workspaces.prometheusservice.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Prometheus resources in AWS from within your Kubernetes cluster.

          **About Amazon Managed Service for Prometheus**

          Amazon Managed Service for Prometheus is a serverless, Prometheus-compatible monitoring service for container metrics that makes it easier to securely monitor container environments at scale. With Amazon Managed Service for Prometheus, you can use the same open-source Prometheus data model and query language that you use today to monitor the performance of your containerized workloads, and also enjoy improved scalability, availability, and security without having to manage the underlying infrastructure.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Prometheus
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - prometheusservice
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Prometheus Developer Resources
          url: https://aws.amazon.com/prometheus/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: prometheus maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/prometheusservice-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-prometheusservice-controller.v1.3.1
        version: 1.3.1
      - name: ack-prometheusservice-controller.v1.3.0
        version: 1.3.0
      - name: ack-prometheusservice-controller.v1.2.27
        version: 1.2.27
      - name: ack-prometheusservice-controller.v1.2.26
        version: 1.2.26
      - name: ack-prometheusservice-controller.v1.2.24
        version: 1.2.24
      - name: ack-prometheusservice-controller.v1.2.23
        version: 1.2.23
      - name: ack-prometheusservice-controller.v1.2.22
        version: 1.2.22
      - name: ack-prometheusservice-controller.v1.2.20
        version: 1.2.20
      - name: ack-prometheusservice-controller.v1.2.19
        version: 1.2.19
      - name: ack-prometheusservice-controller.v1.2.17
        version: 1.2.17
      - name: ack-prometheusservice-controller.v1.2.16
        version: 1.2.16
      - name: ack-prometheusservice-controller.v1.2.15
        version: 1.2.15
      - name: ack-prometheusservice-controller.v1.2.14
        version: 1.2.14
      - name: ack-prometheusservice-controller.v1.2.13
        version: 1.2.13
      - name: ack-prometheusservice-controller.v1.2.12
        version: 1.2.12
      - name: ack-prometheusservice-controller.v1.2.11
        version: 1.2.11
      - name: ack-prometheusservice-controller.v1.2.10
        version: 1.2.10
      - name: ack-prometheusservice-controller.v1.2.9
        version: 1.2.9
      - name: ack-prometheusservice-controller.v1.2.6
        version: 1.2.6
      - name: ack-prometheusservice-controller.v1.2.5
        version: 1.2.5
      - name: ack-prometheusservice-controller.v1.2.4
        version: 1.2.4
      - name: ack-prometheusservice-controller.v1.2.3
        version: 1.2.3
      - name: ack-prometheusservice-controller.v1.2.2
        version: 1.2.2
      - name: ack-prometheusservice-controller.v1.2.1
        version: 1.2.1
      - name: ack-prometheusservice-controller.v1.2.0
        version: 1.2.0
      - name: ack-prometheusservice-controller.v1.1.0
        version: 1.1.0
      - name: ack-prometheusservice-controller.v1.0.0
        version: 1.0.0
      - name: ack-prometheusservice-controller.v0.1.3
        version: 0.1.3
      - name: ack-prometheusservice-controller.v0.1.2
        version: 0.1.2
      - name: ack-prometheusservice-controller.v0.1.1
        version: 0.1.1
      - name: ack-prometheusservice-controller.v0.1.0
        version: 0.1.0
      - name: ack-prometheusservice-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-prometheusservice-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: awss3-operator-registry
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: awss3operator.1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/screeley44/aws-s3-provisioner:v1.0.0
          createdAt: "2019-05-02 16:12:00"
          description: Manage the full lifecycle of installing, configuring and managing
            AWS S3 Provisioner.
          repository: https://github.com/yard-turkey/aws-s3-provisioner
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - description: instance of an AWS S3 Bucket
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
          - description: Request for an AWS S3 Bucket
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
        description: |
          AWS S3 Operator will deploy the AWS S3 Provisioner which will dynamically or statically
          provision AWS S3 Bucket storage and access. The operator deploys the ObjectBucket (OB) and ObjectBucketClaim (OBC)
          CustomResourceDefinitions. The OB/OBC model follows the traditional Kubernetes PV/PVC pattern, when an OBC is detected
          the operator will act on the OBC to either provision a brand new S3 Bucket in AWS or gain access to an existing
          S3 Bucket in AWS. The operator produces an OB and ConfigMap and Secret which can then be consumed by application pods.

          **Important Note**: Currently, while in preview, this operator does not
          support automatic upgrades. You must remove the old version of the operator
          manually before installing a new version.
          ## Using AWS S3 Operator

          ### Administrator Creates Secret
          This secret will contain the elevated/admin privileges needed by the provisioner
          to properly access and create S3 Buckets and IAM users and policies. The AWS Access ID
          and AWS Secret Key will be needed for this.
          1. Create the Kubernetes Secret for the Provisioner's Owner Access.
          ```yaml
          apiVersion: v1
          kind: Secret
          metadata:
            name: s3-bucket-owner [1]
            namespace: s3-provisioner [2]
          type: Opaque
          data:
            AWS_ACCESS_KEY_ID: *base64 encoded value* [3]
            AWS_SECRET_ACCESS_KEY: *base64 encoded value* [4]
          ```
          1. Name of the secret, this will be referenced in StorageClass.
          1. Namespace where the Secret will exist.
          1. Your AWS_ACCESS_KEY_ID base64 encoded.
          1. Your AWS_SECRET_ACCESS_KEY base64 encoded.
          ```
           # kubectl create -f creds.yaml
          secret/s3-bucket-owner created
          ```
          ### Administrator Creates StorageClass
          The StorageClass defines the name of the provisioner and holds other properties that are needed to provision a new bucket, including
          the Owner Secret and Namespace, and the AWS Region.
          #### Greenfield Example:
          For Greenfield, a new, dynamic bucket will be generated.
          1. Create the Kubernetes StorageClass for the Provisioner.
          ```yaml
          kind: StorageClass
          apiVersion: storage.k8s.io/v1
          metadata:
            name: s3-buckets [1]
          provisioner: aws-s3.io/bucket [2]
          parameters:
            region: us-west-1 [3]
            secretName: s3-bucket-owner [4]
            secretNamespace: s3-provisioner [5]
          reclaimPolicy: Delete [6]
          ```
          1. Name of the StorageClass, this will be referenced in the User ObjectBucketClaim.
          1. Provisioner name
          1. AWS Region that the StorageClass will serve
          1. Name of the bucket owner Secret created above
          1. Namespace where the Secret will exist
          1. reclaimPolicy (Delete or Retain) indicates if the bucket can be deleted when the OBC is deleted.
          **NOTE:** the absence of the `bucketName` Parameter key in the storage class indicates this is a new bucket and its name is based on the bucket name fields in the OBC.
          ```
           # kubectl create -f storageclass-greenfield.yaml
          storageclass.storage.k8s.io/s3-buckets created
          ```
          #### Brownfield Example:
          For brownfield, the StorageClass defines the name of the provisioner and the name of the existing bucket. It also includes other properties needed by the target
          provisioner, including: the Owner Secret and Namespace, and the AWS Region
          1. Create the Kubernetes StorageClass for the Provisioner.
          ```yaml
          kind: StorageClass
          apiVersion: storage.k8s.io/v1
          metadata:
            name: s3-existing-buckets [1]
          provisioner: aws-s3.io/bucket [2]
          parameters:
            bucketName: my-existing-bucket [3]
            region: us-west-1 [4]
            secretName: s3-bucket-owner [5]
            secretNamespace: s3-provisioner [6]
          ```
          1. Name of the StorageClass, this will be referenced in the User ObjectBucketClaim.
          1. Provisioner name
          1. Name of the existing bucket
          1. AWS Region that the StorageClass will serve
          1. Name of the bucket owner Secret created above
          1. Namespace for that bucket owner secret
          **NOTE:** the storage class's `reclaimPolicy` is ignored for existing buckets.
          ```
           # kubectl create -f storageclass-brownfield.yaml
          storageclass.storage.k8s.io/s3-buckets created
          ```
          ### User Creates ObjectBucketClaim
          An ObjectBucketClaim follows the same concept as a PVC, in that
          it is a request for Object Storage, the user doesn't need to
          concern him/herself with the underlying storage, just that
          they need access to it. The user will work with the cluster/storage
          administrator to get the proper StorageClass needed and will
          then request access via the OBC.
          #### Greenfield Request Example:
          1. Create the ObjectBucketClaim.
          ```yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucketClaim
          metadata:
            name: myobc [1]
            namespace: s3-provisioner [2]
          spec:
            generateBucketName: mybucket [3]
            bucketName: my-awesome-bucket [4]
            storageClassName: s3-buckets [5]
          ```
          1. Name of the OBC
          1. Namespace of the OBC
          1. Name prepended to a random string used to generate a bucket name. It is ignored if bucketName is defined
          1. Name of new bucket which must be unique across all AWS regions, otherwise an error occurs when creating the bucket. If present, this name overrides `generateName`
          1. StorageClass name
          **NOTE:** if both `generateBucketName` and `bucketName` are omitted, and the storage class does _not_ define a bucket name, then a new, random bucket name is generated with no prefix.
          ```
           # kubectl create -f obc-brownfield.yaml
          objectbucketclaim.objectbucket.io/myobc created
          ```
          #### Brownfield Request Example:
          1. Create the ObjectBucketClaim.
          ```yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucketClaim
          metadata:
            name: myobc [1]
            namespace: s3-provisioner [2]
          spec:
            storageClassName: s3-existing-buckets [3]
          ```
          1. Name of the OBC
          1. Namespace of the OBC
          1. StorageClass name
          **NOTE:** in the OBC here there is no reference to the bucket's name. This is defined in the storage class and is not a concern of the user creating the claim to this bucket.  An OBC does have fields for defining a bucket name for greenfield use only.
          ```
           # kubectl create -f obc-brownfield.yaml
          objectbucketclaim.objectbucket.io/myobc created
          ```
          ### Results and Recap
          Let's pause for a moment and digest what just happened.
          After creating the OBC, and assuming the S3 provisioner is running, we now have
          the following Kubernetes resources:
          .  a global ObjectBucket (OB) which contains: bucket endpoint info (including region and bucket name), a reference to the OBC, and a reference to the storage class. Unique to S3, the OB also contains the bucket Amazon Resource Name (ARN).Note: there is always a 1:1 relationship between an OBC and an OB.
          .  a ConfigMap in the same namespace as the OBC, which contains the same endpoint data found in the OB.
          .  a Secret in the same namespace as the OBC, which contains the AWS key-pairs needed to access the bucket.
          And of course, we have a *new* AWS S3 Bucket which you should be able to see via the AWS Console.
          *ObjectBucket*
          ```yaml
           # kubectl get ob obc-s3-provisioner-my-awesome-bucket -o yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucket
          metadata:
            creationTimestamp: "2019-04-03T15:42:22Z"
            generation: 1
            name: obc-s3-provisioner-my-awesome-bucket
            resourceVersion: "15057"
            selfLink: /apis/objectbucket.io/v1alpha1/objectbuckets/obc-s3-provisioner-my-awesome-bucket
            uid: 0bfe8e84-576d-4c4e-984b-f73c4460f736
          spec:
            Connection:
              additionalState:
                ARN: arn:aws:iam::<accountid>:policy/my-awesome-bucket-vSgD5 [1]
                UserName: my-awesome-bucket-vSgD5 [2]
              endpoint:
                additionalConfig: null
                bucketHost: s3-us-west-1.amazonaws.com
                bucketName: my-awesome-bucket [3]
                bucketPort: 443
                region: us-west-1
                ssl: true
                subRegion: ""
            claimRef: null [4]
            reclaimPolicy: null
            storageClassName: s3-buckets [5]
          ```
          1. The AWS Policy created for this user and bucket.
          1. The new user generated by the Provisioner to access this existing bucket.
          1. The bucket name.
          1. The reference to the OBC (not filled in yet).
          1. The reference to the StorageClass used.
          *ConfigMap*
          ```yaml
           # kubectl get cm myobc -n s3-provisioner -o yaml
          apiVersion: v1
          data:
            BUCKET_HOST: s3-us-west-1.amazonaws.com [1]
            BUCKET_NAME: my-awesome-bucket [2]
            BUCKET_PORT: "443"
            BUCKET_REGION: us-west-1
            BUCKET_SSL: "true"
            BUCKET_SUBREGION: ""
          kind: ConfigMap
          metadata:
            creationTimestamp: "2019-04-01T19:11:38Z"
            finalizers:
            - objectbucket.io/finalizer
            name: my-awesome-bucket
            namespace: s3-provisioner
            resourceVersion: "892"
            selfLink: /api/v1/namespaces/s3-provisioner/configmaps/my-awesome-bucket
            uid: 2edcc58a-aff8-4a29-814a-ffbb6439a9cd
          ```
          1. The AWS S3 host.
          1. The name of the new bucket we are gaining access to.
          *Secret*
          ```yaml
           # kubectl get secret my-awesome-bucket -n s3-provisioner -o yaml
          apiVersion: v1
          data:
            AWS_ACCESS_KEY_ID: *the_new_access_id* [1]
            AWS_SECRET_ACCESS_KEY: *the_new_access_key_value* [2]
          kind: Secret
          metadata:
            creationTimestamp: "2019-04-03T15:42:22Z"
            finalizers:
            - objectbucket.io/finalizer
            name: my-awesome-bucket
            namespace: s3-provisioner
            resourceVersion: "15058"
            selfLink: /api/v1/namespaces/s3-provisioner/secrets/screeley-provb-5
            uid: 225c71a5-9d75-4ccc-b41f-bfe91b272a13
          type: Opaque
          ```
          1. The new generated AWS Access Key ID.
          1. The new generated AWS Secret Access Key.
          What happened in AWS? The first thing we do on any OBC request is
          create a new IAM user and generate Access ID and Secret Keys.
          This allows us to also better control ACLs. We also create a policy
          in IAM which we then attach to the user and bucket. We also created a new bucket, called *my-awesome-bucket*.
          When the OBC is deleted all of its Kubernetes and AWS resources will also be deleted, which includes:
          the generated OB, Secret, ConfigMap, IAM user, and policy.
          If the _retainPolicy_ on the StorageClass for this bucket is *"Delete"*, then, in addition to the above cleanup, the physical bucket is also deleted.
          **NOTE:** The actual bucket is only deleted if the Storage Class's _reclaimPolicy_ is "Delete".
          ### User Creates Pod
          Now that we have our bucket and connection/access information, a pod
          can be used to access the bucket. This can be done in several different
          ways, but the key here is that the provisioner has provided the proper
          endpoints and keys to access the bucket. The user then simply references
          the keys.
          1. Create a Sample Pod to Access the Bucket.
          ```yaml
          apiVersion: v1
          kind: Pod
          metadata:
            name: photo1
            labels:
              name: photo1
          spec:
            containers:
            - name: photo1
              image: docker.io/screeley44/photo-gallery:latest
              imagePullPolicy: Always
              envFrom:
              - configMapRef:
                  name: my-awesome-bucket <1>
              - secretRef:
                  name: my-awesome-bucket <2>
              ports:
              - containerPort: 3000
                protocol: TCP
          ```
          1. Name of the generated configmap from the provisioning process
          1. Name of the generated secret from the provisioning process
          *[Note]* Generated ConfigMap and Secret are same name as the OBC!
          Lastly, expose the pod as a service so you can access the url from a browser. In this example,
          I exposed as a LoadBalancer
          ```
            # kubectl expose pod photo1 --type=LoadBalancer --name=photo1 -n your-namespace
          ```
          To access via a url use the EXTERNAL-IP
          ```
            # kubectl get svc photo1
            NAME                         TYPE           CLUSTER-IP       EXTERNAL-IP                                                               PORT(S)          AGE
            photo1                       LoadBalancer   100.66.124.105   a00c53ccb3c5411e9b6550a7c0e50a2a-2010797808.us-east-1.elb.amazonaws.com   3000:32344/TCP   6d
          ```
          **NOTE:** This is just one example of a Pod that can utilize the bucket information,
          there are several ways that these pod applications can be developed and therefore
          the method of getting the actual values needed from the Secrets and ConfigMaps
          will vary greatly, but the idea remains the same, that the pod consumes the generated
          ConfigMap and Secret created by the provisioner.
        displayName: AWS S3 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - awss3provisioner
        - aws
        - S3
        - provisioner
        - buckets
        links:
        - name: AWS S3 Provisioner
          url: https://github.com/yard-turkey/aws-s3-provisioner
        - name: Provisioner Examples
          url: https://github.com/yard-turkey/aws-s3-provisioner/tree/master/examples
        - name: Operator Documentation
          url: https://github.com/yard-turkey/awss3operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - quay.io/screeley44/aws-s3-provisioner:v1.0.0
        version: 1.0.1
      entries:
      - name: awss3operator.1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: awss3-operator-registry
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Managed OpenShift Black Belt
      provider-url: https://github.com/rh-mobb/ocm-operator
    name: ocm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocm-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "GitLabIdentityProvider",
                "metadata": {
                  "name": "gitlab-sample"
                },
                "spec": {
                  "clientID": "gitlab",
                  "clientSecret": {
                    "name": "gitlab"
                  },
                  "clusterName": "my-cluster",
                  "displayName": "gitlab-sample",
                  "mappingMethod": "claim",
                  "url": "https://gitlab.com"
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "LDAPIdentityProvider",
                "metadata": {
                  "name": "ldap-sample"
                },
                "spec": {
                  "bindDN": "CN=test,OU=Users,DC=example,DC=com",
                  "bindPassword": {
                    "name": "test"
                  },
                  "clusterName": "my-cluster",
                  "displayName": "ldap-sample",
                  "mappingMethod": "claim",
                  "url": "ldap://test.example.com:389"
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "MachinePool",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "aws": {
                    "spotInstances": {
                      "enabled": false
                    }
                  },
                  "clusterName": "my-cluster",
                  "instanceType": "m5.xlarge",
                  "labels": {
                    "owner": "my-user"
                  },
                  "maximumNodesPerZone": 1,
                  "minimumNodesPerZone": 1,
                  "wait": false
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "ROSACluster",
                "metadata": {
                  "name": "rosa-sample"
                },
                "spec": {
                  "accountID": "111111111111",
                  "defaultMachinePool": {
                    "instanceType": "m5.xlarge",
                    "labels": {
                      "owner": "my-owner"
                    },
                    "minimumNodesPerZone": 2
                  },
                  "displayName": "my-cluster",
                  "iam": {
                    "userRole": "arn:aws:iam::111111111111:role/ManagedOpenShift-User-my_user-Role"
                  },
                  "tags": {
                    "owner": "my-owner"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ocm-operator:v0.1.0
          createdAt: "2023-06-02T17:32:25Z"
          operatorframework.io/suggested-namespace: ocm-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ocm-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLabIdentityProvider is the Schema for the gitlabidentityproviders
              API.
            displayName: Git Lab Identity Provider
            kind: GitLabIdentityProvider
            name: gitlabidentityproviders.ocm.mobb.redhat.com
            version: v1alpha1
          - description: LDAPIdentityProvider is the Schema for the ldapidentityproviders
              API.
            displayName: LDAPIdentity Provider
            kind: LDAPIdentityProvider
            name: ldapidentityproviders.ocm.mobb.redhat.com
            version: v1alpha1
          - description: MachinePool is the Schema for the machinepools API.
            displayName: Machine Pool
            kind: MachinePool
            name: machinepools.ocm.mobb.redhat.com
            version: v1alpha1
          - description: ROSACluster is the Schema for the clusters API.
            displayName: ROSACluster
            kind: ROSACluster
            name: rosaclusters.ocm.mobb.redhat.com
            version: v1alpha1
        description: "# Quick Start\n\n## Prerequisites\n\n- [oc](https://docs.openshift.com/container-platform/4.8/cli_reference/openshift_cli/getting-started-cli.html)\n-
          [aws](https://docs.aws.amazon.com/cli/latest/userguide/getting-started-install.html)\n-
          [ROSA Cluster (4.12+)](https://mobb.ninja/docs/rosa/sts/)\n\n> **Warning:**\n>
          This is only tested on ROSA, but may work on other Kubernetes and OpenShift
          variants.  It requires\n> an AWS STS-enabled cluster.  Additionally, OpenShift
          versions 4.12+ are recommended, as they \n> enable [CRD CEL validation](https://kubernetes.io/blog/2022/09/23/crd-validation-rules-beta/).
          \ Versions\n> prior to 4.12 provide a lower version of Kubernetes that does
          not enable this feature.  They may \n> work, but provide no input validation
          when submitting custom resources to the cluster.\n\nBefore installing this
          operator, there are a couple secrets that must be created.\n\n1. OCM Token:
          used to authenticate against the OCM API.  The controller will not start
          without\nthis token as it used for all calls into OCM.\n2. AWS Credentials:
          uses an assume role to a role created in the instructions below in order\nto
          access AWS resources.  This is needed for cluster creation.  See \n[Create
          AWS IAM Policies and Roles](#create-aws-iam-policies-and-roles) for more
          details.\n\n> **Note:**\n> Certain custom resources have their own specific
          prerequisites.  Please use `oc explain` \n> or read the docs [here](https://github.com/rh-mobb/ocm-operator/tree/main/docs)
          for more details.\n\n### Create OCM Token Secret\n\n1. Create a namespace
          where you wish to install the operator:\n\n```bash\noc new-project ocm-operator\n```\n\n2.
          Create a secret containing the OCM_TOKEN.  This token can be obtained form
          \nhttps://console.redhat.com/openshift/token and is used by the operator
          to authenticate \nagainst the OCM API.  This token must exist in the same
          namespace that the operator \nis running and be named `ocm-token`.  It also
          expects the key to be called `OCM_TOKEN` \nas the operator is expecting
          this value as an environment variable.\n\n```bash\noc create secret generic
          ocm-token \\\n  --namespace=ocm-operator \\\n  --from-literal=OCM_TOKEN=${MY_OCM_TOKEN}\n```\n\n###
          Create AWS IAM Policies and Roles\n\nThe operator will need to elevate privileges
          in order to perform things like \ncreating the operator-roles for the clusters.
          \ Because of this, the operator \nmust have a specific role created to allow
          it these permissions.\n\n**NOTE:** please understand what you are doing
          if you deviate from the known good \npolicies.  If errors or more stringent
          security lockdowns are found, please submit a PR \nso that we can get this
          fixed.\n\n1. Set `ACCOUNT_ID` environment variable to define your AWS account
          ID:\n\n```bash\nexport ACCOUNT_ID=111111111111\n```\n\n2. Create the permissions
          boundary.  Because the operator needs to create policies and \nroles, the
          boundary ensures that the operator is not allowed to create additional\npermissions
          outside of the defined boundary.  The sample permission set is located \nat
          `test/aws/boundary.json` in this repository:\n\n```bash\nboundary=$(curl
          -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/boundary.json\")\naws
          iam create-policy \\\n  --policy-name \"OCMOperatorBoundary\" \\\n  --policy-document
          \"$boundary\"\n```\n\n3. Create the policy.  This policy sets what the operator
          is allowed to do.  For any \n`iam` permission, the boundary created above
          is used.  The sample permission set is \nlocated at `test/aws/iam_policy.json`
          in this repository:\n\n```bash\npolicy=$(curl -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/iam_policy.json\")\naws
          iam create-policy \\\n  --policy-name \"OCMOperator\" \\\n  --policy-document
          \"$policy\"\n```\n\n4. Create the role using a trust policy and attach the
          previously created role.  The trust \npolicy is located at `test/aws/trust_policy.json`
          in this repository.  Please note that \nthe trust policy requires an OIDC
          configuration.  The OIDC configuration refers to \n**where the operator
          is running, NOT what it is provisioning**:\n\n```bash\ntrust_policy=$(curl
          -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/trust_policy.json\")\naws
          iam create-role \\\n    --role-name OCMOperator \\\n    --assume-role-policy-document
          \"$trust_policy\"\naws iam attach-role-policy \\\n    --role-name OCMOperator
          \\\n    --policy-arn arn:aws:iam::$ACCOUNT_ID:policy/OCMOperator\n```\n\n5.
          Finally, create the secret containing the assume role credentials.  The
          previous steps allow \nthe operator to assume the role you created in the
          previous step with the permissions created \nvia the previous policies.\n\n```bash\ncat
          <<EOF > /tmp/credentials\n[default]\nrole_arn = arn:aws:iam::$ACCOUNT_ID:role/OCMOperator\nweb_identity_token_file
          = /var/run/secrets/openshift/serviceaccount/token\nEOF\n\noc create secret
          generic aws-credentials \\\n  --namespace=ocm-operator \\\n  --from-file=credentials=/tmp/credentials\n```\n\n##
          Install the Operator\n\n```bash\ncat <<EOF | oc apply -f -\napiVersion:
          operators.coreos.com/v1alpha1\nkind: Subscription\nmetadata:\n  name: ocm-operator\n
          \ namespace: ocm-operator\nspec:\n  channel: v0.1\n  installPlanApproval:
          Automatic\n  name: ocm-operator\n  source: community-operators\n  sourceNamespace:
          ocm-operator\n  startingCSV: ocm-operator.v0.1.0\nEOF\n```\n\n## Provision
          Resources\n\nOnce the operator is available and running, you can provision
          any of the \nresources that it manages.  Also note that documentation is
          always available \nonce the operator is installed as well by using the `oc
          explain` command.  For \nexample, `oc explain rosacluster.spec.clusterName`
          will give you detailed documentation \nabout what the field does.\n\nSee
          the following documentation for details:\n\n* [ROSA Clusters](https://github.com/rh-mobb/ocm-operator/blob/main/docs/clusters.md)\n*
          [Machine Pools](https://github.com/rh-mobb/ocm-operator/blob/main/docs/machinepools.md)\n*
          [Identity Providers](https://github.com/rh-mobb/ocm-operator/blob/main/docs/identityproviders.md)\n"
        displayName: OpenShift Cluster Manager Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ocm
        - rosa
        - redhat
        - machinepool
        - mobb
        links:
        - name: OCM Machine Pool Operator
          url: https://github.com/rh-mobb/ocm-operator
        maintainers:
        - email: dscott@redhat.com
          name: Dustin Scott
        maturity: alpha
        provider:
          name: Managed OpenShift Black Belt
          url: https://github.com/rh-mobb/ocm-operator
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/mobb/ocm-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: ocm-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ocm-operator
    provider:
      name: Managed OpenShift Black Belt
      url: https://github.com/rh-mobb/ocm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SAP LinuxLab - Red Hat
      provider-url: https://github.com/sap-linuxlab
    name: sap-hana-express-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sap-hana-express-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "db.sap-redhat.io/v1alpha1",
                "kind": "HanaExpress",
                "metadata": {
                  "name": "hanaexpress-sample"
                },
                "spec": {
                  "credential": {
                    "key": "hxepasswd.json",
                    "name": "hxepasswd"
                  },
                  "pvcSize": "1Gi"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/redhat-sap-cop/sap-hanaexpress-operator:0.1.4
          createdAt: "2023-07-16T20:53:02Z"
          description: Manage the lifecycle of SAP Hana Express
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HanaExpress is the Schema for the hanaexpresses API
            displayName: Hana Express
            kind: HanaExpress
            name: hanaexpresses.db.sap-redhat.io
            version: v1alpha1
        description: Operator for managing the lifecycle of SAP Hana Express on Kubernetes
          platform
        displayName: SAP Hana Express Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Database
        - SAP
        links:
        - name: Sap Hana Express Operator
          url: https://github.com/redhat-sap/sap-hana-express-operator
        maintainers:
        - email: mjiao@redhat.com
          name: RedHat@SAPLinuxLab
        maturity: alpha
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/redhat-sap-cop/sap-hanaexpress-operator:0.1.4
        version: 1.0.2
      entries:
      - name: sap-hana-express-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: sap-hana-express-operator
    provider:
      name: SAP LinuxLab - Red Hat
      url: https://github.com/sap-linuxlab
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-registry-3
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-3.v3.0.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "simple"
                },
                "spec": {
                  "app": {
                    "ingress": {
                      "host": "simple-app.apps.cluster.example"
                    }
                  },
                  "ui": {
                    "ingress": {
                      "host": "simple-ui.apps.cluster.example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-3-operator:3.0.15
          createdAt: "2025-09-22T13:04:40Z"
          description: Deploy and manage Apicurio Registry 3 on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Apicurio Registry
              version 3.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Apicurio Registry 3

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Allows you manage event schemas and API designs using the Apicurio Registry web console, REST API, Maven plug-in, or Java client
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, MySQL, SQL Server, in-memory
          - Compatible with Confluent APIs
          - Runs on a lightweight Quarkus platform

          ## Apicurio Registry 3 Operator

          Provides a quick and easy way to deploy and manage Apicurio Registry 3 on Kubernetes.

          **Features**
          - Supports installation and configuration of Apicurio Registry 3
          - Perform a rolling upgrade of Apicurio Registry 3

          ## Prerequisites

          This operator does not deploy storage for Apicurio Registry 3. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio Registry 3 and Apicurio Registry 3 Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Apicurio Registry 3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-3-operator:3.0.15
        version: 3.0.15
      entries:
      - name: apicurio-registry-3.v3.0.15
        version: 3.0.15
      - name: apicurio-registry-3.v3.0.14
        version: 3.0.14
      - name: apicurio-registry-3.v3.0.12
        version: 3.0.12
      - name: apicurio-registry-3.v3.0.9
        version: 3.0.9
      - name: apicurio-registry-3.v3.0.8
        version: 3.0.8
      - name: apicurio-registry-3.v3.0.7
        version: 3.0.7
      name: 3.x
    defaultChannel: 3.x
    packageName: apicurio-registry-3
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM Edge
      provider-url: ""
    name: infrastructure-asset-orchestrator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infrastructure-asset-orchestrator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infrastructure.asset.orchestrator.com/v1",
                "kind": "Generatekey",
                "metadata": {
                  "name": "modelbuilder-token"
                },
                "spec": {}
              },
              {
                "apiVersion": "infrastructure.asset.orchestrator.com/v1",
                "kind": "Mb-broker-service",
                "metadata": {
                  "name": "mb-broker-service-sample"
                },
                "spec": {
                  "env_type": "prod",
                  "cloud_type": "IBM",
                  "external_db": {
                    "enabled": false
                  },
                  "postgres": {
                    "enable_pg_tls": true,
                    "storage_class": "",
                    "storage_size": "20G",
                    "backup_type": "incremental",
                    "backup_frequency": "@daily"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/iao-operator@sha256:5cc395f22d86641e175339deb134af6c74de92b81470a65a5c736c8005d645fd
          createdAt: "2021-08-19T12:00:00Z"
          olm.skipRange: <1.0.2
          operators.operatorframework.io/builder: operator-sdk-v1.12.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/infrastructure-asset-orchestrator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Generate key CR
            displayName: GenerateKey
            kind: Generatekey
            name: generatekeys.infrastructure.asset.orchestrator.com
            version: v1
          - description: Deploy MB broker deployment
            displayName: MBBrokerService
            kind: Mb-broker-service
            name: mb-broker-services.infrastructure.asset.orchestrator.com
            version: v1
        description: Infrastructure Asset Orchestrator provides users access to cloud
          resources and services. Leverages the dynamic nature of the cloud platforms
          to provide significant cost savings and efficiencies. Using the Infrastructure
          Asset Orchestrator user can provision, configure, utilize and deprovision
          variety of services on the IBM Cloud Catalog.
        displayName: Infrastructure Asset Orchestrator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cloud-service
        - mbbroker-service
        links:
        - name: Infrastructure Asset Orchestrator
          url: https://infrastructure-asset-orchestrator.domain
        maintainers:
        - email: edge@us.ibm.com
          name: IBM-Edge
        maturity: alpha
        provider:
          name: IBM Edge
        relatedImages:
        - registry.connect.redhat.com/ibm/iao-operator@sha256:5cc395f22d86641e175339deb134af6c74de92b81470a65a5c736c8005d645fd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:609c19d1718947ee814b829334c37e1d720acf5fb642b0712b8ab1b14d57f53f
        - registry.connect.redhat.com/ibm/iao-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/iao-orchestrator-service@sha256:145928929fd0c2b7d9ed5fb374d7cc0e6b48269847e810c5472580a1538ad44c
        - registry.connect.redhat.com/ibm/iao-mb-broker-service@sha256:6531b4196c51246958027453c2d9c7dbcdb4a1f55c826d1da238d256cdbc8a10
        version: 1.0.2
      entries:
      - name: infrastructure-asset-orchestrator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: infrastructure-asset-orchestrator-certified
    provider:
      name: IBM Edge
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: node-maintenance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-maintenance-operator.v5.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:58846ab62346b1004e11f36bd901c96f8d315958eb39fbb272d8b328faebcf04
          createdAt: "2023-10-20 15:36:51"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=4.10.0 <5.0.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-node-maintenance-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:58846ab62346b1004e11f36bd901c96f8d315958eb39fbb272d8b328faebcf04
        version: 5.0.1
      entries:
      - name: node-maintenance-operator.v5.0.1
        version: 5.0.1
      name: 4.12-eus
    - currentCSV: node-maintenance-operator.v5.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:2efd1976d2fb110391e0e5937b90476d3eeaacb4dd29cd52375f2b8566f17b41
          createdAt: "2023-10-20 15:50:49"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=5.0.0 <5.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:2efd1976d2fb110391e0e5937b90476d3eeaacb4dd29cd52375f2b8566f17b41
        version: 5.2.1
      entries:
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.14-eus
    - currentCSV: node-maintenance-operator.v5.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:7ca5853885534198544d23067087d80047b0e1421903085a2cb328bd00def147
          createdAt: "2024-08-10 19:55:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.2.0 <5.3.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:7ca5853885534198544d23067087d80047b0e1421903085a2cb328bd00def147
        version: 5.3.1
      entries:
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.16-eus
    - currentCSV: node-maintenance-operator.v5.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-10 10:59:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.3.0 <5.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.4/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel9-operator@sha256:752bff8bebcaf3df1f9d126dc4b07b622a44b02bfcf0f82b9100a7693627079a
        version: 5.4.1
      entries:
      - name: node-maintenance-operator.v5.4.1
        version: 5.4.1
      - name: node-maintenance-operator.v5.4.0
        version: 5.4.0
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.18-eus
    - currentCSV: node-maintenance-operator.v5.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-10 10:59:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.3.0 <5.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.4/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel9-operator@sha256:752bff8bebcaf3df1f9d126dc4b07b622a44b02bfcf0f82b9100a7693627079a
        version: 5.4.1
      entries:
      - name: node-maintenance-operator.v5.4.1
        version: 5.4.1
      - name: node-maintenance-operator.v5.4.0
        version: 5.4.0
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: stable
    defaultChannel: stable
    packageName: node-maintenance-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sagemaker-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sagemaker-controller.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "App",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "DataQualityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Domain",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Endpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "EndpointConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "FeatureGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "HyperParameterTuningJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Model",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelBiasJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelExplainabilityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelPackage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelPackageGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelQualityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "MonitoringSchedule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "NotebookInstance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "NotebookInstanceLifecycleConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ProcessingJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "TrainingJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "TransformJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "UserProfile",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sagemaker-controller:1.4.1
          createdAt: "2025-09-22T19:07:36Z"
          description: ACK service controller for Amazon SageMaker manages SageMaker
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: App represents the state of an AWS sagemaker App resource.
            displayName: App
            kind: App
            name: apps.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: DataQualityJobDefinition represents the state of an AWS sagemaker
              DataQualityJobDefinition resource.
            displayName: DataQualityJobDefinition
            kind: DataQualityJobDefinition
            name: dataqualityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Domain represents the state of an AWS sagemaker Domain resource.
            displayName: Domain
            kind: Domain
            name: domains.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: EndpointConfig represents the state of an AWS sagemaker EndpointConfig
              resource.
            displayName: EndpointConfig
            kind: EndpointConfig
            name: endpointconfigs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Endpoint represents the state of an AWS sagemaker Endpoint
              resource.
            displayName: Endpoint
            kind: Endpoint
            name: endpoints.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: FeatureGroup represents the state of an AWS sagemaker FeatureGroup
              resource.
            displayName: FeatureGroup
            kind: FeatureGroup
            name: featuregroups.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: HyperParameterTuningJob represents the state of an AWS sagemaker
              HyperParameterTuningJob resource.
            displayName: HyperParameterTuningJob
            kind: HyperParameterTuningJob
            name: hyperparametertuningjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: InferenceComponent represents the state of an AWS sagemaker
              InferenceComponent resource.
            displayName: InferenceComponent
            kind: InferenceComponent
            name: inferencecomponents.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: LabelingJob represents the state of an AWS sagemaker LabelingJob
              resource.
            displayName: LabelingJob
            kind: LabelingJob
            name: labelingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelBiasJobDefinition represents the state of an AWS sagemaker
              ModelBiasJobDefinition resource.
            displayName: ModelBiasJobDefinition
            kind: ModelBiasJobDefinition
            name: modelbiasjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelExplainabilityJobDefinition represents the state of
              an AWS sagemaker ModelExplainabilityJobDefinition resource.
            displayName: ModelExplainabilityJobDefinition
            kind: ModelExplainabilityJobDefinition
            name: modelexplainabilityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelPackageGroup represents the state of an AWS sagemaker
              ModelPackageGroup resource.
            displayName: ModelPackageGroup
            kind: ModelPackageGroup
            name: modelpackagegroups.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelPackage represents the state of an AWS sagemaker ModelPackage
              resource.
            displayName: ModelPackage
            kind: ModelPackage
            name: modelpackages.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelQualityJobDefinition represents the state of an AWS
              sagemaker ModelQualityJobDefinition resource.
            displayName: ModelQualityJobDefinition
            kind: ModelQualityJobDefinition
            name: modelqualityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Model represents the state of an AWS sagemaker Model resource.
            displayName: Model
            kind: Model
            name: models.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: MonitoringSchedule represents the state of an AWS sagemaker
              MonitoringSchedule resource.
            displayName: MonitoringSchedule
            kind: MonitoringSchedule
            name: monitoringschedules.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: NotebookInstanceLifecycleConfig represents the state of an
              AWS sagemaker NotebookInstanceLifecycleConfig resource.
            displayName: NotebookInstanceLifecycleConfig
            kind: NotebookInstanceLifecycleConfig
            name: notebookinstancelifecycleconfigs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: NotebookInstance represents the state of an AWS sagemaker
              NotebookInstance resource.
            displayName: NotebookInstance
            kind: NotebookInstance
            name: notebookinstances.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: PipelineExecution represents the state of an AWS sagemaker
              PipelineExecution resource.
            displayName: PipelineExecution
            kind: PipelineExecution
            name: pipelineexecutions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Pipeline represents the state of an AWS sagemaker Pipeline
              resource.
            displayName: Pipeline
            kind: Pipeline
            name: pipelines.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ProcessingJob represents the state of an AWS sagemaker ProcessingJob
              resource.
            displayName: ProcessingJob
            kind: ProcessingJob
            name: processingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: TrainingJob represents the state of an AWS sagemaker TrainingJob
              resource.
            displayName: TrainingJob
            kind: TrainingJob
            name: trainingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: TransformJob represents the state of an AWS sagemaker TransformJob
              resource.
            displayName: TransformJob
            kind: TransformJob
            name: transformjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: UserProfile represents the state of an AWS sagemaker UserProfile
              resource.
            displayName: UserProfile
            kind: UserProfile
            name: userprofiles.sagemaker.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon SageMaker resources in AWS from within your Kubernetes cluster.

          **About Amazon SageMaker**

          Amazon SageMaker helps data scientists and developers to prepare, build, train, and deploy high-quality machine learning (ML) models quickly by bringing together a broad set of capabilities purpose-built for ML.

          For more information on Amazon SageMaker, visit the [product page](https://aws.amazon.com/sagemaker/).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SageMaker
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sagemaker
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SageMaker Developer Resources
          url: https://aws.amazon.com/sagemaker/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sagemaker maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sagemaker-controller:1.4.1
        version: 1.4.1
      entries:
      - name: ack-sagemaker-controller.v1.4.1
        version: 1.4.1
      - name: ack-sagemaker-controller.v1.4.0
        version: 1.4.0
      - name: ack-sagemaker-controller.v1.3.3
        version: 1.3.3
      - name: ack-sagemaker-controller.v1.3.2
        version: 1.3.2
      - name: ack-sagemaker-controller.v1.3.1
        version: 1.3.1
      - name: ack-sagemaker-controller.v1.3.0
        version: 1.3.0
      - name: ack-sagemaker-controller.v1.2.20
        version: 1.2.20
      - name: ack-sagemaker-controller.v1.2.19
        version: 1.2.19
      - name: ack-sagemaker-controller.v1.2.18
        version: 1.2.18
      - name: ack-sagemaker-controller.v1.2.17
        version: 1.2.17
      - name: ack-sagemaker-controller.v1.2.16
        version: 1.2.16
      - name: ack-sagemaker-controller.v1.2.13
        version: 1.2.13
      - name: ack-sagemaker-controller.v1.2.12
        version: 1.2.12
      - name: ack-sagemaker-controller.v1.2.11
        version: 1.2.11
      - name: ack-sagemaker-controller.v1.2.9
        version: 1.2.9
      - name: ack-sagemaker-controller.v1.2.8
        version: 1.2.8
      - name: ack-sagemaker-controller.v1.2.5
        version: 1.2.5
      - name: ack-sagemaker-controller.v1.2.4
        version: 1.2.4
      - name: ack-sagemaker-controller.v1.2.3
        version: 1.2.3
      - name: ack-sagemaker-controller.v1.2.2
        version: 1.2.2
      - name: ack-sagemaker-controller.v1.2.1
        version: 1.2.1
      - name: ack-sagemaker-controller.v1.2.0
        version: 1.2.0
      - name: ack-sagemaker-controller.v1.1.0
        version: 1.1.0
      - name: ack-sagemaker-controller.v1.0.0
        version: 1.0.0
      - name: ack-sagemaker-controller.v0.6.0
        version: 0.6.0
      - name: ack-sagemaker-controller.v0.5.1
        version: 0.5.1
      - name: ack-sagemaker-controller.v0.5.0
        version: 0.5.0
      - name: ack-sagemaker-controller.v0.4.5
        version: 0.4.5
      - name: ack-sagemaker-controller.v0.4.4
        version: 0.4.4
      - name: ack-sagemaker-controller.v0.4.3
        version: 0.4.3
      - name: ack-sagemaker-controller.v0.4.2
        version: 0.4.2
      - name: ack-sagemaker-controller.v0.4.0
        version: 0.4.0
      - name: ack-sagemaker-controller.v0.3.4
        version: 0.3.4
      - name: ack-sagemaker-controller.v0.3.3
        version: 0.3.3
      - name: ack-sagemaker-controller.v0.3.2
        version: 0.3.2
      - name: ack-sagemaker-controller.v0.3.1
        version: 0.3.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-sagemaker-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: networkservicemesh.io
      provider-url: networkservicemesh.io
    name: nsm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nsm-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nsm.networkservicemesh.io/v1alpha1",
                "kind": "NSM",
                "metadata": {
                  "name": "nsm-sample",
                  "namespace": "nsm"
                },
                "spec": {
                  "forwarders": [
                    {
                      "image": "ghcr.io/networkservicemesh/cmd-forwarder-vpp:v1.8.0",
                      "type": "vpp"
                    }
                  ],
                  "nsmLogLevel": "TRACE",
                  "nsmPullPolicy": "IfNotPresent",
                  "registry": {
                    "image": "ghcr.io/networkservicemesh/cmd-registry-memory:v1.8.0",
                    "type": "memory"
                  },
                  "version": "v1.8.0",
                  "webhook": {
                    "image": "ghcr.io/networkservicemesh/admission-webhook-k8s:v1.8.0"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: quay.io/acmenezes/nsm-operator:v1.8.0
          createdAt: "2023-03-02T19:36:46Z"
          operatorframework.io/suggested-namespace: nsm
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/networkservicemesh/nsm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkServiceEndpoint
            name: networkserviceendpoints.networkservicemesh.io
            version: v1
          - kind: NetworkService
            name: networkservices.networkservicemesh.io
            version: v1
          - kind: NSM
            name: nsms.nsm.networkservicemesh.io
            version: v1alpha1
        description: |-
          The Network Service Mesh Operator manages and install network service meshes.

          Check networkservicemesh.io

          Install requirements:

          Before installing the operator it's necessary to have installed [spire](https://github.com/spiffe/spire)

          We have a ready to install yaml file for spire [here](https://github.com/networkservicemesh/nsm-operator/blob/master/config/spire/spire.yaml)
        displayName: nsm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network
        - service mesh
        links:
        - name: Nsm Operator
          url: https://nsm-operator.domain
        maintainers:
        - email: adcmenezes@gmail.com
          name: Alexandre Menezes
        maturity: alpha
        provider:
          name: networkservicemesh.io
          url: networkservicemesh.io
        relatedImages:
        - quay.io/acmenezes/nsm-operator:v1.8.0
        version: 1.8.0
      entries:
      - name: nsm-operator.v1.8.0
        version: 1.8.0
      name: alpha
    defaultChannel: alpha
    packageName: nsm-operator
    provider:
      name: networkservicemesh.io
      url: networkservicemesh.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: observability-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      name: candidate
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      - name: observability-operator.v0.0.22
        version: 0.0.22
      name: development
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      name: stable
    defaultChannel: stable
    packageName: observability-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-atlas-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-atlas-kubernetes.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "id": "1",
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "WEEKLY",
                      "id": "2",
                      "retentionUnit": "DAYS",
                      "retentionValue": 6
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": true,
                  "policies": [
                    {
                      "name": "atlas-default-backuppolicy",
                      "namespace": "mongodb-atlas-system"
                    },
                    {
                      "name": "atlas-default-backuppolicy2",
                      "namespace": "mongodb-atlas-system"
                    }
                  ],
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "my-database-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "david"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-cluster"
                },
                "spec": {
                  "deploymentSpec": {
                    "name": "test-cluster",
                    "providerSettings": {
                      "instanceSizeName": "M10",
                      "providerName": "AWS",
                      "regionName": "US_EAST_1"
                    }
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          containerImage: registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator@sha256:3cfbd700e4771c2c7e75c2c41cf6de57c9ef71ecd286e96a21d2567e43424bc3
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasclusters API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB Atlas —
          the only multi-cloud document database service that gives you the versatility you need to build sophisticated and resilient applications that can adapt to changing customer demands and market trends.

          > Current Status: *Beta*. The Operator gives users the ability to provision
          > Atlas projects, clusters and database users using Kubernetes Specifications and bind connection information
          > into applications deployed to Kubernetes.

          ## Quick Start guide
          ### Step 1. Deploy Kubernetes operator by clicking Install button.

          ### Step 2. Create Atlas Cluster

          **1.** Create an Atlas API Key Secret
          In order to work with the Atlas Operator you need to provide [authentication information](https://docs.atlas.mongodb.com/configure-api-access)
           to allow the Atlas Operator to communicate with Atlas API. Once you have generated a Public and Private key in Atlas, you can create a Kuberentes Secret with:
          ```
          kubectl create secret generic mongodb-atlas-operator-api-key \
                   --from-literal="orgId=<the_atlas_organization_id>" \
                   --from-literal="publicApiKey=<the_atlas_api_public_key>" \
                   --from-literal="privateApiKey=<the_atlas_api_private_key>" \
                   -n openshift-operators
          ```
          (Note, that you should use the namespace where the Operator was installed - it's `openshift-operators` by default)

          **2.** Create an `AtlasProject` Custom Resource

          The `AtlasProject` CustomResource represents Atlas Projects in our Kubernetes cluster. You need to specify
          `projectIpAccessList` with the IP addresses or CIDR blocks of any hosts that will connect to the Atlas Cluster.
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasProject
          metadata:
            name: my-project
          spec:
            name: Test Atlas Operator Project
            projectIpAccessList:
              - ipAddress: "192.0.2.15"
                comment: "IP address for Application Server A"
              - ipAddress: "203.0.113.0/24"
                comment: "CIDR block for Application Server B - D"
          ```
          **3.** Create an `AtlasDeployment` Custom Resource.
          The example below is a minimal configuration to create an M10 Atlas cluster in the AWS US East region. For a full list of properties, check
          `atlasclusters.atlas.mongodb.com` [CRD specification](config/crd/bases/atlas.mongodb.com_atlasclusters.yaml)):
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasDeployment
          metadata:
            name: my-atlas-cluster
          spec:
            name: "Test-cluster"
            projectRef:
              name: my-project
            providerSettings:
              instanceSizeName: M10
              providerName: AWS
              regionName: US_EAST_1
          ```

          **4.** Create a database user password Kubernetes Secret
          The Secret must be created in the same namespace as the `AtlasDeployment` and `AtlasProject` were created.
          ```
          kubectl create secret generic the-user-password --from-literal="password=P@@sword%"
          ```

          **5.** Create an `AtlasDatabaseUser` Custom Resource

          In order to connect to an Atlas Cluster the database user needs to be created. `AtlasDatabaseUser` resource should reference
          the password Kubernetes Secret created in the previous step.
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasDatabaseUser
          metadata:
            name: my-database-user
          spec:
            roles:
              - roleName: "readWriteAnyDatabase"
                databaseName: "admin"
            projectRef:
              name: my-project
            username: theuser
            passwordSecretRef:
              name: the-user-password
          ```
          **6.** Wait for the `AtlasDatabaseUser` Custom Resource to be ready

          Wait until the AtlasDatabaseUser resource gets to "ready" status (it will wait until the cluster is created that may take around 10 minutes):
          ```
          kubectl get atlasdatabaseusers my-database-user -o=jsonpath='{.status.conditions[?(@.type=="Ready")].status}'
          True
          ```
          ### Step 3. Connect your application to the Atlas Cluster

          The Atlas Operator will create a Kubernetes Secret with the information necessary to connect to the Atlas Cluster created
          in the previous step. An application in the same Kubernetes Cluster can mount and use the Secret:

          ```
          ...
          containers:
                - name: test-app
                  env:
                   - name: "CONNECTION_STRING"
                     valueFrom:
                       secretKeyRef:
                         name: test-atlas-operator-project-test-cluster-theuser
                         key: connectionStringStandardSrv

          ```
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator@sha256:3cfbd700e4771c2c7e75c2c41cf6de57c9ef71ecd286e96a21d2567e43424bc3
        version: 1.0.0
      entries:
      - name: mongodb-atlas-kubernetes.v1.0.0
        version: 1.0.0
      name: beta
    - currentCSV: mongodb-atlas-kubernetes.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupCompliancePolicy",
                "metadata": {
                  "name": "my-backup-compliance-policy"
                },
                "spec": {
                  "authorizedEmail": "john.doe@example.com",
                  "authorizedUserFirstName": "John",
                  "authorizedUserLastName": "Doe",
                  "copyProtectionEnabled": false,
                  "encryptionAtRestEnabled": false,
                  "onDemandPolicy": {
                    "retentionUnit": "weeks",
                    "retentionValue": 3
                  },
                  "overwriteBackupPolicies": false,
                  "pointInTimeEnabled": true,
                  "restoreWindowDays": 42,
                  "scheduledPolicyItems": [
                    {
                      "frequencyInterval": 2,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "hourly",
                      "retentionUnit": "days",
                      "retentionValue": 2
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "weekly",
                      "retentionUnit": "weeks",
                      "retentionValue": 4
                    },
                    {
                      "frequencyInterval": 15,
                      "frequencyType": "monthly",
                      "retentionUnit": "months",
                      "retentionValue": 12
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": false,
                  "copySettings": [
                    {
                      "cloudProvider": "AWS",
                      "frequencies": [
                        "WEEKLY"
                      ],
                      "regionName": "US_WEST_1",
                      "shouldCopyOplogs": false
                    }
                  ],
                  "policy": {
                    "name": "atlasbackuppolicy-sample"
                  },
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2,
                  "useOrgAndGroupNamesInExportPrefix": true
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasCustomRole",
                "metadata": {
                  "name": "shard-operator-role"
                },
                "spec": {
                  "projectRef": {
                    "name": "my-project"
                  },
                  "role": {
                    "actions": [
                      {
                        "name": "getShardMap",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "shardingState",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "connPoolStats",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "getLog",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      }
                    ],
                    "inheritedRoles": [
                      {
                        "database": "admin",
                        "name": "operator-role-1"
                      }
                    ],
                    "name": "my-role"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDataFederation",
                "metadata": {
                  "name": "my-data-federation",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "name": "my-data-federation",
                  "privateEndpoints": [
                    {
                      "endpointId": "vpce-03f9eeaa764e32454",
                      "provider": "AWS",
                      "type": "DATA_LAKE"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project",
                    "namespace": "mongodb-atlas-system"
                  },
                  "storage": {
                    "databases": [
                      {
                        "collections": [
                          {
                            "dataSources": [
                              {
                                "storeName": "http-test",
                                "urls": [
                                  "https://data.cityofnewyork.us/api/views/vfnx-vebw/rows.csv"
                                ]
                              }
                            ],
                            "name": "test-collection-1"
                          }
                        ],
                        "name": "test-db-1"
                      }
                    ],
                    "stores": [
                      {
                        "name": "http-test",
                        "provider": "http"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "the-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "theuser"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-deployment"
                },
                "spec": {
                  "backupRef": {
                    "name": "atlasbackupschedule-sample"
                  },
                  "deploymentSpec": {
                    "backupEnabled": true,
                    "clusterType": "REPLICASET",
                    "mongoDBMajorVersion": "7.0",
                    "name": "test-deployment",
                    "replicationSpecs": [
                      {
                        "regionConfigs": [
                          {
                            "electableSpecs": {
                              "instanceSize": "M10",
                              "nodeCount": 3
                            },
                            "priority": 7,
                            "providerName": "AWS",
                            "regionName": "US_EAST_1"
                          }
                        ],
                        "zoneName": "Zone 1"
                      }
                    ],
                    "tags": [
                      {
                        "key": "team",
                        "value": "atlas-operator"
                      }
                    ]
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasFederatedAuth",
                "metadata": {
                  "name": "atlas-default-federated-auth",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "connectionSecretRef": {
                    "name": "my-org-secret",
                    "namespace": "mongodb-atlas-system"
                  },
                  "domainAllowList": [
                    "my-org-domain.com"
                  ],
                  "domainRestrictionEnabled": true,
                  "enabled": true,
                  "postAuthRoleGrants": [
                    "ORG_MEMBER"
                  ],
                  "roleMappings": [
                    {
                      "externalGroupName": "org-admin",
                      "roleAssignments": [
                        {
                          "role": "ORG_OWNER"
                        }
                      ]
                    },
                    {
                      "externalGroupName": "dev-team",
                      "roleAssignments": [
                        {
                          "role": "ORG_GROUP_CREATOR"
                        },
                        {
                          "projectName": "dev-project",
                          "role": "GROUP_OWNER"
                        }
                      ]
                    }
                  ],
                  "ssoDebugEnabled": false
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasIPAccessList",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasipaccesslist-sample"
                },
                "spec": {
                  "entries": [
                    {
                      "cidrBlock": "192.168.1.0/24"
                    },
                    {
                      "deleteAfterDate": "2025-01-20T23:59:59+02:00",
                      "ipAddress": "10.0.0.1"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkContainer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasnetworkcontainer-sample"
                },
                "spec": {
                  "cidrBlock": "10.11.0.0/16",
                  "connectionSecret": {
                    "name": "atlas-connection-secret"
                  },
                  "externalProjectRef": {
                    "projectId": "66e2f2b621571b7e69a89b66"
                  },
                  "provider": "AWS",
                  "region": "US_EAST_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkPeering",
                "metadata": {
                  "name": "atlasnetworkpeering-sample"
                },
                "spec": {
                  "atlasCidrBlock": "10.12.0.0/16",
                  "awsConfiguration": {
                    "accepterRegionName": "US_EAST_1",
                    "awsAccountId": "23214235145",
                    "routeTableCidrBlock": "10.11.0.0/16",
                    "vpcId": "vpcid-7123782134"
                  },
                  "containerId": "623412394512350",
                  "containerRegion": "US_EAST_1",
                  "projectRef": {
                    "name": "atlas-project",
                    "namespace": "namespace"
                  },
                  "provider": "AWS"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasPrivateEndpoint",
                "metadata": {
                  "name": "atlasprivateendpoint-sample"
                },
                "spec": {
                  "awsConfiguration": {
                    "id": "vpce-f4k34w51d"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "provider": "AWS",
                  "region": "EU_CENTRAL_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    },
                    {
                      "cidrBlock": "203.0.113.0/24",
                      "comment": "CIDR block for Application Server B - D"
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasSearchIndexConfig",
                "metadata": {
                  "name": "atlassearchindexconfig-sample"
                },
                "spec": {
                  "analyzer": "lucene.standard",
                  "analyzers": [
                    {
                      "name": "test-analyzer",
                      "tokenFilters": [
                        {
                          "asciiFolding": {
                            "originalTokens": "omit"
                          },
                          "type": "asciiFolding"
                        }
                      ],
                      "tokenizer": {
                        "type": "whiteSpace",
                        "whitespace": {
                          "maxTokenLenght": 200
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamConnection",
                "metadata": {
                  "name": "my-streamconnection-kafka"
                },
                "spec": {
                  "kafkaConfig": {
                    "authentication": {
                      "credentials": {
                        "name": "kafka-userpass"
                      },
                      "mechanism": "SCRAM-512"
                    },
                    "bootstrapServers": "kafka.server1:9001,kafka.server2:9002,kafka.server3:9003",
                    "security": {
                      "certificate": {
                        "name": "kafka-cert"
                      },
                      "protocol": "SSL"
                    }
                  },
                  "name": "kafka-config",
                  "type": "Kafka"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamInstance",
                "metadata": {
                  "name": "my-streaminstance-sample"
                },
                "spec": {
                  "clusterConfig": {
                    "provider": "AWS",
                    "region": "VIRGINIA_USA",
                    "tier": "SP30"
                  },
                  "connectionRegistry": [
                    {
                      "name": "my-streamconnection-sample"
                    }
                  ],
                  "name": "my-instance",
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasTeam",
                "metadata": {
                  "name": "atlasteam-sample"
                },
                "spec": {
                  "name": "my-team",
                  "usernames": [
                    "username@mongodb.com"
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasThirdPartyIntegration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasintegration-sample"
                },
                "spec": {
                  "datadog": {
                    "apiKeySecretRef": {
                      "name": "secret-ref"
                    },
                    "region": "string",
                    "sendCollectionLatencyMetrics": "string",
                    "sendDatabaseMetrics": "string"
                  },
                  "projectRef": {
                    "name": "my-atlas-project"
                  },
                  "type": "DATADOG"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:8ac3f6b1c0f31af5f32562ddeed37dd6586ac47e1e7e111947df3ed570c5bed9
          createdAt: "2025-09-25T14:34:34Z"
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/mongodb/mongodb-atlas-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The AtlasBackupCompliancePolicy is a configuration that enforces
              specific backup and retention requirements
            displayName: Atlas Backup Compliance Policy
            kind: AtlasBackupCompliancePolicy
            name: atlasbackupcompliancepolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasCustomRole is the Schema for the AtlasCustomRole API
            displayName: Atlas Custom Role
            kind: AtlasCustomRole
            name: atlascustomroles.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDataFederation is the Schema for the Atlas Data Federation
              API
            displayName: Atlas Data Federation
            kind: AtlasDataFederation
            name: atlasdatafederations.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasdeployments API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasFederatedAuth is the Schema for the Atlasfederatedauth
              API
            displayName: Atlas Federated Auth
            kind: AtlasFederatedAuth
            name: atlasfederatedauths.atlas.mongodb.com
            version: v1
          - description: AtlasIPAccessList is the Schema for the atlasipaccesslists
              API.
            displayName: Atlas IPAccess List
            kind: AtlasIPAccessList
            name: atlasipaccesslists.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkContainer is the Schema for the AtlasNetworkContainer
              API
            displayName: Atlas Network Container
            kind: AtlasNetworkContainer
            name: atlasnetworkcontainers.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkPeering is the Schema for the AtlasNetworkPeering
              API
            displayName: Atlas Network Peering
            kind: AtlasNetworkPeering
            name: atlasnetworkpeerings.atlas.mongodb.com
            version: v1
          - displayName: Atlas Org Settings
            kind: AtlasOrgSettings
            name: atlasorgsettings.atlas.mongodb.com
            version: v1
          - description: "The AtlasPrivateEndpoint custom resource definition (CRD)
              defines a desired [Private Endpoint](https://www.mongodb.com/docs/atlas/security-private-endpoint/#std-label-private-endpoint-overview)
              configuration for an Atlas project. It allows a private connection between
              your cloud provider and Atlas that doesn't send information through
              a public network. \n You can use private endpoints to create a unidirectional
              connection to Atlas clusters from your virtual network."
            displayName: Atlas Private Endpoint
            kind: AtlasPrivateEndpoint
            name: atlasprivateendpoints.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
          - description: Atlas Search Index Config is the Schema for the Atlas Search
              API
            displayName: Atlas Search Index Config
            kind: AtlasSearchIndexConfig
            name: atlassearchindexconfigs.atlas.mongodb.com
            version: v1
          - description: AtlasStreamConnection is the Schema for the atlasstreamconnections
              API
            displayName: Atlas Stream Connection
            kind: AtlasStreamConnection
            name: atlasstreamconnections.atlas.mongodb.com
            version: v1
          - description: AtlasStreamInstance is the Schema for the atlasstreaminstances
              API
            displayName: Atlas Stream Instance
            kind: AtlasStreamInstance
            name: atlasstreaminstances.atlas.mongodb.com
            version: v1
          - description: AtlasTeam is the Schema for the Atlas Teams API
            displayName: Atlas Team
            kind: AtlasTeam
            name: atlasteams.atlas.mongodb.com
            version: v1
          - description: AtlasThirdPartyIntegration is the Schema for the atlas 3rd
              party integrations API.
            displayName: Atlas Third Party Integration
            kind: AtlasThirdPartyIntegration
            name: atlasthirdpartyintegrations.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB
          Atlas — the only multi-cloud document database service that gives you the versatility you need to build sophisticated
          and resilient applications that can adapt to changing customer demands and market trends.

          The full documentation for the Operator can be found [here](https://docs.atlas.mongodb.com/atlas-operator/)

          ## Getting Started

          ### Supported features

          * Create and configure a project, or connect to an existing one.
          * Deploy, manage, scale, and tear down clusters.
          * Support for serverless instances.
          * Create and edit database users.
          * Manage IP Access Lists, network peering and private endpoints.
          * Configure and control Atlas’s fully managed cloud backup.
          * Configure federated authentication for your Atlas organization
          * Integrate Atlas monitoring with Prometheus.

          ... and more.

          To view the list of custom resources and each CR's schema, visit our [reference](https://www.mongodb.com/docs/atlas/operator/stable/custom-resources/)
          documentation. See the [Quickstart](https://www.mongodb.com/docs/atlas/operator/stable/ak8so-quick-start/) to get
          started with Atlas Kubernetes Operator.
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:8ac3f6b1c0f31af5f32562ddeed37dd6586ac47e1e7e111947df3ed570c5bed9
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:9eb6465e723ede43ac47eed6de22042d26c0efb276301a4424a6a3948a8f5b15
        version: 2.11.0
      entries:
      - name: mongodb-atlas-kubernetes.v2.11.0
        version: 2.11.0
      - name: mongodb-atlas-kubernetes.v2.10.0
        version: 2.10.0
      - name: mongodb-atlas-kubernetes.v2.9.1
        version: 2.9.1
      - name: mongodb-atlas-kubernetes.v2.9.0
        version: 2.9.0
      - name: mongodb-atlas-kubernetes.v2.8.2
        version: 2.8.2
      - name: mongodb-atlas-kubernetes.v2.8.1
        version: 2.8.1
      - name: mongodb-atlas-kubernetes.v2.8.0
        version: 2.8.0
      - name: mongodb-atlas-kubernetes.v2.7.1
        version: 2.7.1
      - name: mongodb-atlas-kubernetes.v2.7.0
        version: 2.7.0
      - name: mongodb-atlas-kubernetes.v2.6.1
        version: 2.6.1
      - name: mongodb-atlas-kubernetes.v2.6.0
        version: 2.6.0
      - name: mongodb-atlas-kubernetes.v2.5.0
        version: 2.5.0
      - name: mongodb-atlas-kubernetes.v2.4.1
        version: 2.4.1
      - name: mongodb-atlas-kubernetes.v2.4.0
        version: 2.4.0
      - name: mongodb-atlas-kubernetes.v2.3.1
        version: 2.3.1
      - name: mongodb-atlas-kubernetes.v2.3.0
        version: 2.3.0
      - name: mongodb-atlas-kubernetes.v2.2.2
        version: 2.2.2
      - name: mongodb-atlas-kubernetes.v2.2.1
        version: 2.2.1
      - name: mongodb-atlas-kubernetes.v2.2.0
        version: 2.2.0
      - name: mongodb-atlas-kubernetes.v2.1.0
        version: 2.1.0
      - name: mongodb-atlas-kubernetes.v2.0.1
        version: 2.0.1
      - name: mongodb-atlas-kubernetes.v2.0.0
        version: 2.0.0
      - name: mongodb-atlas-kubernetes.v1.9.1
        version: 1.9.1
      - name: mongodb-atlas-kubernetes.v1.9.0
        version: 1.9.0
      - name: mongodb-atlas-kubernetes.v1.8.2
        version: 1.8.2
      - name: mongodb-atlas-kubernetes.v1.8.1
        version: 1.8.1
      - name: mongodb-atlas-kubernetes.v1.8.0
        version: 1.8.0
      - name: mongodb-atlas-kubernetes.v1.7.3
        version: 1.7.3
      - name: mongodb-atlas-kubernetes.v1.7.2
        version: 1.7.2
      - name: mongodb-atlas-kubernetes.v1.7.1
        version: 1.7.1
      - name: mongodb-atlas-kubernetes.v1.7.0
        version: 1.7.0
      - name: mongodb-atlas-kubernetes.v1.6.1
        version: 1.6.1
      - name: mongodb-atlas-kubernetes.v1.6.0
        version: 1.6.0
      - name: mongodb-atlas-kubernetes.v1.5.1
        version: 1.5.1
      - name: mongodb-atlas-kubernetes.v1.5.0
        version: 1.5.0
      - name: mongodb-atlas-kubernetes.v1.4.1
        version: 1.4.1
      - name: mongodb-atlas-kubernetes.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-atlas-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco AppDynamics
      provider-url: https://www.appdynamics.com
    name: appdynamics-cloud-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appdynamics-cloud-operator.v24.7.0-133
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clustermon",
                "metadata": {
                  "name": "clustermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "clusterName": "test",
                  "image": "appdynamics/appdynamics-cloud-k8s-monitoring:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Containermon",
                "metadata": {
                  "name": "containermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "name": "ContainerMon"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "DbCollector",
                "metadata": {
                  "name": "dbcollector",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-db-collector:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "DbConfig",
                "metadata": {
                  "name": "dbconfig",
                  "namespace": "appdynamics"
                },
                "spec": null
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Inframon",
                "metadata": {
                  "name": "inframon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-k8s-monitoring:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "LogCollector",
                "metadata": {
                  "name": "logcollector-sample",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-log-collector-agent:latest",
                  "resources": {
                    "limits": {
                      "cpu": "400m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "150Mi"
                    }
                  }
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Servermon",
                "metadata": {
                  "name": "servermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "name": "ServerMon"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-08-06T08:33:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clustermon is the Schema for the clustermons API
            displayName: Clustermon
            kind: Clustermon
            name: clustermons.cluster.appdynamics.com
            version: v1alpha1
          - description: Containermon is the Schema for the containermons API
            displayName: Containermon
            kind: Containermon
            name: containermons.cluster.appdynamics.com
            version: v1alpha1
          - kind: DbCollector
            name: dbcollectors.cluster.appdynamics.com
            version: v1alpha1
          - kind: DbConfig
            name: dbconfigs.cluster.appdynamics.com
            version: v1alpha1
          - description: Inframon is the Schema for the inframons API
            displayName: Inframon
            kind: Inframon
            name: inframons.cluster.appdynamics.com
            version: v1alpha1
          - kind: LogCollector
            name: logcollectors.cluster.appdynamics.com
            version: v1alpha1
          - description: Servermon is the Schema for the servermons API
            displayName: Servermon
            kind: Servermon
            name: servermons.cluster.appdynamics.com
            version: v1alpha1
        description: The Appdynamics Cloud Operator is used to install Cluster Collector,
          Container Collector, Server Collector, Log Collector and Cluster Agent
        displayName: Cisco AppDynamics Cloud Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - appdynamics
        links:
        - name: Appdynamics Cloud Operator
          url: https://appdynamics-cloud-operator.domain
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: Cisco AppDynamics
          url: https://www.appdynamics.com
        relatedImages:
        - registry.connect.redhat.com/appdynamics/appdynamics-cloud-operator@sha256:2710d63425942f60af21f060a02eb3c40f5a5f0c952ea6593b3b9769c7ca1014
        version: 24.7.0-133
      entries:
      - name: appdynamics-cloud-operator.v24.7.0-133
        version: 24.7.0-133
      name: alpha
    defaultChannel: alpha
    packageName: appdynamics-cloud-operator
    provider:
      name: Cisco AppDynamics
      url: https://www.appdynamics.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: smb-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smb-csi-driver-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9-operator@sha256:1019812a3969f844a0f3fcd45091fccc7d0d816207e230a2f399556c41ebc705
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure CIFS/SMB CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.15.0-0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/csi-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          CIFS/SMB CSI Driver Operator provides Server Message Block (SMB) CSI Driver that enables you to create and mount CIFS/SMB PersistentVolumes.
        displayName: CIFS/SMB CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - smb
        - cifs
        links:
        - name: Documentation
          url: https://github.com/openshift/csi-operator
        - name: Source Repository
          url: https://github.com/openshift/csi-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: preview
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9-operator@sha256:1019812a3969f844a0f3fcd45091fccc7d0d816207e230a2f399556c41ebc705
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0627c09fe3f5a015f117fb30af1f4d17340aefa48b23994737d3901a804b63d3
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:a5704d26d0b562dba9633d1c107f34e05c750c0e3a64dbcace4dab7b964fff8d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9@sha256:9f516d06c279cf1f82e8948ce2290395eec5f6611abf5ea919e4d6548d81ec10
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:9c3c3d2b857ce289324aee56df1ba4ab5e5ea1e76cd86ac988233463915834b4
        version: 4.16.0-202509240123
      entries:
      - name: smb-csi-driver-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: smb-csi-driver-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: smb-csi-driver-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: smb-csi-driver-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: smb-csi-driver-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: smb-csi-driver-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: smb-csi-driver-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: smb-csi-driver-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: smb-csi-driver-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: smb-csi-driver-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: smb-csi-driver-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: smb-csi-driver-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: smb-csi-driver-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: smb-csi-driver-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: smb-csi-driver-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: smb-csi-driver-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: smb-csi-driver-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: smb-csi-driver-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: smb-csi-driver-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: smb-csi-driver-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: smb-csi-driver-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: smb-csi-driver-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: smb-csi-driver-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: smb-csi-driver-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: smb-csi-driver-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: smb-csi-driver-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: smb-csi-driver-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: smb-csi-driver-operator.v4.16.0-202406251906
        version: 4.16.0-202406251906
      - name: smb-csi-driver-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: smb-csi-driver-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: preview
    defaultChannel: preview
    packageName: smb-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: bamoe-kogito-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bamoe-kogito-operator.8.0.5-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "ruleunit-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogitoinfra-sample"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "kafka.strimzi.io/v1beta2",
                    "kind": "Kafka",
                    "name": "kogito-kafka"
                  }
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator@sha256:8a99fcd217502b7e850af1e1097de58b6cb19d1e786394811a8e075c4d871279
          description: IBM BAMOE Kogito Operator for deployment and management of
            Kogito services.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["IBM Business Automation Manager
            Open Edition"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: https://www.ibm.com/mysupport
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.rhpam.kiegroup.org
            version: v1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.rhpam.kiegroup.org
            version: v1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.rhpam.kiegroup.org
            version: v1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.rhpam.kiegroup.org
            version: v1
        description: |-
          IBM BAMOE Kogito Operator is designed for deployment and management of Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Builds a IBM BAMOE Kogito service from a remote git repository and creates a custom image for the given runtime: Red Hat build of Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Deploy IBM BAMOE Kogito service from user provided image.
          * Setup Grafana Dashboard when Grafana Operator is available in cluster.
          * Create Prometheus monitor object when Prometheus operator is available in cluster.

          **IMPORTANT!** IBM BAMOE Kogito Operator may need Grafana and Prometheus Operator to configure monitoring in cases if you do not provide these services beforehand. If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the IBM BAMOE Kogito Operator.
        displayName: IBM BAMOE Kogito Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://ibm.com
        - name: Documentation
          url: https://www.ibm.com/docs/en/ibamoe
        - name: Source
          url: https://github.com/kiegroup/kogito-operator
        maintainers:
        - email: mysphelp@us.ibm.com
          name: IBM
        maturity: beta
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9492c67eaa7ab37df6f5fa623c4aac955b7d7ffaab9ae2e57dea745e82907e40
        - registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator@sha256:8a99fcd217502b7e850af1e1097de58b6cb19d1e786394811a8e075c4d871279
        version: 8.0.5-3
      entries:
      - name: bamoe-kogito-operator.8.0.5-3
        version: 8.0.5-3
      - name: bamoe-kogito-operator.8.0.5-2
        version: 8.0.5-2
      - name: bamoe-kogito-operator.8.0.5-1
        version: 8.0.5-1
      - name: bamoe-kogito-operator.8.0.4-2
        version: 8.0.4-2
      - name: bamoe-kogito-operator.8.0.4-1
        version: 8.0.4-1
      - name: bamoe-kogito-operator.8.0.3-1
        version: 8.0.3-1
      - name: bamoe-kogito-operator.8.0.2-3
        version: 8.0.2-3
      - name: bamoe-kogito-operator.8.0.2-2
        version: 8.0.2-2
      - name: bamoe-kogito-operator.8.0.2-1
        version: 8.0.2-1
      - name: bamoe-kogito-operator.8.0.1-2
        version: 8.0.1-2
      - name: bamoe-kogito-operator.8.0.1-1
        version: 8.0.1-1
      - name: bamoe-kogito-operator.8.0.0-1
        version: 8.0.0-1
      name: 8.x
    defaultChannel: 8.x
    packageName: bamoe-kogito-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc
      provider-url: ""
    name: openshift-gitops-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-gitops-operator.v1.10.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:b0dc693854319ada341e2e3872b34d7db9823b470ed369edfde88d3ae73a40e1
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.8.19\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:a74e0a1fe3589ebe9f226a2bad20e9a5d5601a0e62c5f3290902f466f55b7160
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:2bec31b0c5407ffcecfd8b860d8afb1629eaab31a250f60191f94a824e0a52f4
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:c779ad87a7f71d851c2af818abf16e5151c43d3fd54e64eac2fc7117c00753f5
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:b0dc693854319ada341e2e3872b34d7db9823b470ed369edfde88d3ae73a40e1
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:b5c8a6c7438a45125fc7fd970946f715ab7193b7446d625fae80dde9d1f3f914
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:4c842f0ea8f0e9b53eb16d94488fc2f4298bd8e4be2f48713be9157da24549e6
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:ece2fee4227ad1becadca47028b78ceaf8717a98ec84e3f1b5188a135c4b093c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d35dff5d59e748ef0404e7e76b0237ea96b2bb19b86fa04a3a34df7a379277b2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5d5061dbc2ec5082cf14b6c600fb5400b83cf91d7ccebfa80680a238d275db
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:0905ba3f5cf87780a1e95b6a00ea7cd340d4ed1b5054e4b2c82406270baeaefa
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ec9f60018694dcc5d431ba47d5536b761b71cb3f66684978fe6bb74c157679ac
        version: 1.10.6
      entries:
      - name: openshift-gitops-operator.v1.10.6
        version: 1.10.6
      - name: openshift-gitops-operator.v1.10.5
        version: 1.10.5
      - name: openshift-gitops-operator.v1.10.4
        version: 1.10.4
      - name: openshift-gitops-operator.v1.10.3
        version: 1.10.3
      - name: openshift-gitops-operator.v1.10.2
        version: 1.10.2
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.10
    - currentCSV: openshift-gitops-operator.v1.11.7-0.1724840231.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d6ad2ee900bf99a3672049cc8ce9c94870d41176501ded1abbc1b229b9553401
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          olm.substitutesFor: openshift-gitops-operator.v1.11.7
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.9.21\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:4d22a7a6cc2f0442889512267ae3d9356f05ff0dfdbbe137c3d8ebebc24818c8
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:42f809f32f1d43186b849e3cc78063d3ebb2b54c55b2a1b77908af09a5ce044b
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:0b926fc9eae775425faa1ab9709a36fdac15afe4bfe523720e17ecf5df45bb28
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:0df37cf0a0e48efae02a0711c6da61c6165e7856db97097e98d56c8b6a137bb9
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:eade50b8321c1b2e0a8ac4dcb34596d43dea0a137b8b97b6eb68c6d469c8b200
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:20566173a8e907980e603d82fae4d96fcdfd5c43d9004c2e506cf0e09a47b91e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/rhel8/redis-6@sha256:c4a98010f7dd5faec2047bb9803061704233a8efc930be2a02386ac16d062d86
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c94c2167dc957856682a104b9a6187d3d321dec71ae794acc0dc37794d2cc373
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:7ea96000768b3cf7711150531c4c1c7edff2c45a2f29fe10bd9811ab6b281909
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d6ad2ee900bf99a3672049cc8ce9c94870d41176501ded1abbc1b229b9553401
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:959de6a95ef8978f66fdbe1033bcaa0ddeff9a376023e5d1d2698c5acb417305
        version: 1.11.7+0.1724840231.p
      entries:
      - name: openshift-gitops-operator.v1.11.7-0.1724840231.p
        version: 1.11.7+0.1724840231.p
      - name: openshift-gitops-operator.v1.11.7
        version: 1.11.7
      - name: openshift-gitops-operator.v1.11.6
        version: 1.11.6
      - name: openshift-gitops-operator.v1.11.5
        version: 1.11.5
      - name: openshift-gitops-operator.v1.11.4
        version: 1.11.4
      - name: openshift-gitops-operator.v1.11.3
        version: 1.11.3
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.11
    - currentCSV: openshift-gitops-operator.v1.12.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:48dd1254fd3a887aef2eae9239f68637eae18514a50f72ea504414d83dc727b8
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.10.18\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:dbf3038daf7c797123c511f9a609c3485cb1a5bcae53c614fad17e0a65fa7ee3
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a1d4e26ea755181818401697f18260b9e876aa441be14830f9142081ee22a24d
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:73918f3b3c507fc64006b7a500ecfa9d58151d694852f696c41a4412edffc78e
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:33bbb88343285a0510670384b87d0226dd6de8aa12c78b68921362c4d6f5a6aa
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:48dd1254fd3a887aef2eae9239f68637eae18514a50f72ea504414d83dc727b8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9e0fee30842a16d7b8a39d336123c80c5271cb2ceb50759fc7ee51a4ed09ceb3
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:06a6637c4776f4285bcb295d327ee60b1aed61b7adc5f7ab52898d79281ced70
        - registry.redhat.io/rhel8/redis-6@sha256:a848cc4edb4405846c673187dd68ce66c328dd1ad42f193f5737288c7971399d
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:b5736d1106db29fcdf42328ca20b96382d4339f90b97a07120ea73ee174ea1ad
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:7d82c719b1cef64607d30e8aaa5ab293955771c5b1dbd430e0befc988a43356d
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:1cf485a889c7586cb4944f000b53c0e97d08f97ba14e580bec57dad7cb13712e
        version: 1.12.6
      entries:
      - name: openshift-gitops-operator.v1.12.6
        version: 1.12.6
      - name: openshift-gitops-operator.v1.12.5-0.1727371523.p
        version: 1.12.5+0.1727371523.p
      - name: openshift-gitops-operator.v1.12.5
        version: 1.12.5
      - name: openshift-gitops-operator.v1.12.4
        version: 1.12.4
      - name: openshift-gitops-operator.v1.12.3
        version: 1.12.3
      - name: openshift-gitops-operator.v1.12.2
        version: 1.12.2
      - name: openshift-gitops-operator.v1.12.1
        version: 1.12.1
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.12
    - currentCSV: openshift-gitops-operator.v1.13.3-0.1741683398.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d41034bcf5c28bcb8afef08678793d0cc27a4032e2811a0c496c3bd0f10ecc17
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          olm.substitutesFor: openshift-gitops-operator.v1.13.3-0.1738142397.p
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.11.12\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_gitops/
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d41034bcf5c28bcb8afef08678793d0cc27a4032e2811a0c496c3bd0f10ecc17
        - registry.redhat.io/rhel8/redis-6@sha256:a848cc4edb4405846c673187dd68ce66c328dd1ad42f193f5737288c7971399d
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:b982df02ea7c45a7603d83fe82cac12e1faac303a1b72af6a727512bb97441df
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3f2804bc2af437ff1f06c214b48a1a780dfcfab995650d901c03270057eb9f3c
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:06a6637c4776f4285bcb295d327ee60b1aed61b7adc5f7ab52898d79281ced70
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a7c3e31ce414e6c6fecf1652343b22c52502902676bf4d7a75bf42bdabfca06
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:a366229e6ee26ebe2d7f765b2a2da6057a29e783835655b5d97545ac563b541a
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:d4d5724f67859d24c32f0f5735f5af3e325f77db97f58bc7f17ed4b831b7e534
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c02dc2c79ee24ce5450ea4e19493ac2a1997eeb5bfd359291642c529826d0cdd
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:e87515d016565bb69cb2c2ac40432e0127dd2cb20a490ac82019836032ba15e5
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:bde11fa83b7e708677b08ea995041f480b88edfb93e4a62c10380d101c0a349c
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        version: 1.13.3+0.1741683398.p
      entries:
      - name: openshift-gitops-operator.v1.13.3-0.1741683398.p
        version: 1.13.3+0.1741683398.p
      - name: openshift-gitops-operator.v1.13.3-0.1738142397.p
        version: 1.13.3+0.1738142397.p
      - name: openshift-gitops-operator.v1.13.3
        version: 1.13.3
      - name: openshift-gitops-operator.v1.13.2
        version: 1.13.2
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.13
    - currentCSV: openshift-gitops-operator.v1.14.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:7cf924af77d9c56e70710b99e080ed8aee162f05ec472f32085a76ef85cf9879
          createdAt: "2025-05-22 12:40:45"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.12.13-1\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_gitops/
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:2c45cfc01f9ec7577afc09cadc4c1a6f470836f6a0a7c24d66b791913f89bb05
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:42a79aad641e8607fc339e4ce7f9cf1d258be01f863dbc31a94d2c27edb645bd
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:97e32825f59c33cca7d8dba3580095205c20435ccd15b91cf0aff37db9841e0e
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:7cf924af77d9c56e70710b99e080ed8aee162f05ec472f32085a76ef85cf9879
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:b7d3c882fc25fbbea28951afb197d4d47ace253864889339396e20f0c6897a7f
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:bcf6b3136a5e71803cb518c8763c165bb46ee37ef7bf55503911d7fcb36b3a8f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c369df980874aae270ca7bc6455f1ca45467d75d25fe80d7bcb06a7826bdeaa3
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:cd4965d91c163160bcf43a02c54ca60b8704281c95de904f904df942bfd83bcd
        - registry.redhat.io/rhel8/redis-6@sha256:a9f30ed08abcab0122556a5485fbcf6e0fc03860f19954dcebdbf5e754876a45
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:7358787959af7e7e589edfc043a97d80ebab87536bde365d33be7a600071bf0e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1ed3b514bf9af6ac0d386e3739874d1d964e667d54e041c2d67edd769b0009fb
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:8096f824bd96e3148cf4b6936766c888d1fff0af9e9d0be440ae4feb6adc03f2
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:1112160243f227a5d393eeee165bd4fed0d3f42646a9280cbe9af60ef728aeb2
        version: 1.14.4
      entries:
      - name: openshift-gitops-operator.v1.14.4
        version: 1.14.4
      - name: openshift-gitops-operator.v1.14.3-0.1746016855.p
        version: 1.14.3+0.1746016855.p
      - name: openshift-gitops-operator.v1.14.3-0.1742836272.p
        version: 1.14.3+0.1742836272.p
      - name: openshift-gitops-operator.v1.14.3
        version: 1.14.3
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.14
    - currentCSV: openshift-gitops-operator.v1.15.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ad49a3d14f843bde0d2365093d00ffa6d3b9d59a75c81e0c56710ad37b53bd92
          createdAt: "2024-11-25T03:57:37Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v2.13.9\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:d41454824632475f0e88aa6e0031ec643bd37e1d575df36cd3ee63daf1403281
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:6cec76d7ea70d77923cf4285177e375d34e96e3aecbe77b68cd5ff769e6e540b
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:911733d78ae6f36f642b4d63f044cd102b7eee170508eaf518ac7acf0f339698
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ad49a3d14f843bde0d2365093d00ffa6d3b9d59a75c81e0c56710ad37b53bd92
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:41717c27915251bda8afeef651b609b8a9ed410e0fa09c4cf8fa7e91d6a94292
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:0c7f72546d0d918bfd21aff7157b983ff7c832b269f5efb84d2ed723a09bf612
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:5bda26e2c16beebf66dd8f3c3c55d48e2f00db870f2d9d4d44570d9d3b8bddb0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:a9e2360f00c5c9840a100a6748bd905b8f34169140a33fb0ca4af9cf6db71642
        version: 1.15.4
      entries:
      - name: openshift-gitops-operator.v1.15.4
        version: 1.15.4
      - name: openshift-gitops-operator.v1.15.3
        version: 1.15.3
      - name: openshift-gitops-operator.v1.15.2
        version: 1.15.2
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.15
    - currentCSV: openshift-gitops-operator.v1.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ead539d7cf54992cf1bfeae6a73a453059a7dffef44505991ff33319c8d40c50
          createdAt: "2025-03-07T13:13:07Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v2.14.16\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:58284a12174d19c2d04723fa1f551d96b490872e0fa2ad94af1b3e52dc6ad955
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a5e2a12ec5dea607a6651462d8e7920cc73f3272dba4041c397a14f2c73d7532
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:03d0d9093422396330aa8611adf77abdc9be759f3d2e23850e4abe778bd6002b
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:cab809fd1288d31c93f4f5c4facb7efaff9f844e21a859d1e4557831575a13c8
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c37aa141387c847eac825363dddda3eb59203d3694120dba439b00cf0c66f320
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:2598a3339bc920b91018d3f7acbe78ccbee684b5182c874e48296744d2373f9d
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ead539d7cf54992cf1bfeae6a73a453059a7dffef44505991ff33319c8d40c50
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:ceff1cde6d043daae599b578eec9759c6d49cd5cf770460b9d075fa879a16e81
        version: 1.16.3
      entries:
      - name: openshift-gitops-operator.v1.16.3
        version: 1.16.3
      - name: openshift-gitops-operator.v1.16.2
        version: 1.16.2
      - name: openshift-gitops-operator.v1.16.1
        version: 1.16.1
      - name: openshift-gitops-operator.v1.16.0-0.1746014725.p
        version: 1.16.0+0.1746014725.p
      - name: openshift-gitops-operator.v1.16.0
        version: 1.16.0
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.16
    - currentCSV: openshift-gitops-operator.v1.17.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:1b86eb4e071b1c526931b6d18c3d8f51522293699059c514488602c320271049
          createdAt: "2025-07-30T13:03:16Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.0.14\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:3fc4941fa4a6f2dbe9dd82622a0d3fb794ecd5a9dee286a38b60a7f6a947d685
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:cc45e212e7eb4e7e8daac6ea74071d5d98260de0b1f6d1cea0e7876111ccea8c
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:2f5f997bce924445de735ae0508dca1a7bba561bc4acdacf659928488233cb8a
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:5c9ea426cd60e7b8d1d8e4fe763909200612434c65596855334054e26cbfe3d0
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:70e5e3900d0c89999c69be1d58adf84c1b3664aacbffa9c5e0efc8ae1b6ef9c1
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:1b86eb4e071b1c526931b6d18c3d8f51522293699059c514488602c320271049
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:d73b4f5b355f556208751281923bb8b43edffc061a673c04d53b88f0e50ecb26
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:a2ae66772d90e734bb3385a59852e6016325a21b916a2580a33cf0da7cb0b527
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:c0a513ed54b631629efafe6e94b775d5bd0a943c44fbe96f02a0d5bb1f989099
        version: 1.17.1
      entries:
      - name: openshift-gitops-operator.v1.17.1
        version: 1.17.1
      - name: openshift-gitops-operator.v1.17.0
        version: 1.17.0
      name: gitops-1.17
    - currentCSV: openshift-gitops-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
          createdAt: "2025-08-21T01:20:45Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.18.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: NamespaceManagement
            name: namespacemanagements.argoproj.io
            version: v1beta1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.1.6\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/rhel9/redis-7@sha256:b4f58ad83ab8719a2e102f38ee526546cd1d56428901afc3add45c9a0e2a2f12
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:1eef9ecd58cf8391f29360ef74dd8aa5dcb48e1582d1500e445706d8302c951e
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a0544c14823492165550d83a6d8ba79dd632b46144d3fdcb543793726111d76
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:6284873bcf1593c210003e0292ac5583f03f03e80d937098d808d48fa0d01cf6
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:5a09fcba3d9e6fc462719e81a46af6022cf7f70a05bcb387ed959a49f6f37dbc
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:adca9c55b80d899d500bf0a7234cf61ae30ffed1b2c927ddea8d9fd07c9acb6a
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7f6ceae56c60fd432425c36f93a9e9bbbb96ef9422333ad6ea41ff3a2eaf8374
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:640e3298c7cd2faa84971b0d861446c4892ab8602108af05b8b5b049a4de624f
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c2283997d47298d49378e8200b149ea613c7735317ef12e74b7915a590f41ce2
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:08efd3e7a79c2945693c99ed31ec912223becff7238b1e28d1b1da5d63f93019
        version: 1.18.0
      entries:
      - name: openshift-gitops-operator.v1.18.0
        version: 1.18.0
      name: gitops-1.18
    - currentCSV: openshift-gitops-operator.v1.6.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:104cc7564fcc3261faa3fbf9c95720b4434e15103b876c080ed012b244ef1dae
          createdAt: "2021-09-15T22:31:21Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD v2.4.28\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces. \n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc.
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:104cc7564fcc3261faa3fbf9c95720b4434e15103b876c080ed012b244ef1dae
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:0cc4fa704e581cd5581c816668183f602fae6ef9ab3c04ce0d99d561287ef140
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:06f1150114d4c0bd236d2a5d4716c9c3e95a9733058297790389d6ba2a1bf264
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:e9e2f950cda8153cff112e43872b98e4977d3cbdf59713b8de92defef6d34b6d
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:dbcad68b70ab7918e944099aef4177c9bc3378e9ca49499dbb94a175927d3209
        - registry.redhat.io/rhel8/redis-6@sha256:53598a6effeb90e4f1b005b2521beffd2fa2b0c52d0e7f2347ee2abd2577cab3
        version: 1.6.6
      entries:
      - name: openshift-gitops-operator.v1.6.6
        version: 1.6.6
      - name: openshift-gitops-operator.v1.6.5
        version: 1.6.5
      - name: openshift-gitops-operator.v1.6.4
        version: 1.6.4
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.6
    - currentCSV: openshift-gitops-operator.v1.7.4-0.1690486082.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:e0967a4058fe5a9bfe7e998a3347529c99536cf3a44de9e40f9632035ff5a2fe
          createdAt: "2021-09-15T22:31:21Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.substitutesFor: openshift-gitops-operator.v1.7.4
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD v2.5.16\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficient privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces. \n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/rhel8/redis-6@sha256:53598a6effeb90e4f1b005b2521beffd2fa2b0c52d0e7f2347ee2abd2577cab3
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:e0967a4058fe5a9bfe7e998a3347529c99536cf3a44de9e40f9632035ff5a2fe
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c55aadf45397083bec4e42d6c0514ed79169a07fd3d9d75671904df2c0520e52
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:45789c49d9334e0ed062641313714c101bd0117dc5633316b1ec63e7a6c76c15
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:26050fb95978e90e8d00e895032045e77cb236008a0e1417459e1aeca8a48c69
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:5c11be8ec9db54331893f2f298aa27006e77b0eaded2d88d0531c77da91a421f
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:f1371d2012796e9aaca105112190ff89ec3d0be95968e05b19db101480f67975
        version: 1.7.4+0.1690486082.p
      entries:
      - name: openshift-gitops-operator.v1.7.4-0.1690486082.p
        version: 1.7.4+0.1690486082.p
      - name: openshift-gitops-operator.v1.7.4
        version: 1.7.4
      - name: openshift-gitops-operator.v1.7.3
        version: 1.7.3
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.7
    - currentCSV: openshift-gitops-operator.v1.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f32df77b3c23d6a833ff887386285ca7a5bd436641f0ec1ffc934c8247fa315f
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.skipRange: ""
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.6.153\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:8cc59901689858e06f5c2942f8c975449c17454fa8364da6153d9b5a3538a985
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:f77594bc053be144b33ff9603ee9675c7e82946ec0dbfb04d8f942c8d73155da
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f32df77b3c23d6a833ff887386285ca7a5bd436641f0ec1ffc934c8247fa315f
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:8aa68fddef88921aeeede24a849d3dabbceff01dc03185342c21e6ee58033eaf
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:16d9184345e6d5a19f326a7b52f76d8a59b96763f233c50546ff017a625e5ff5
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:0d13c03048b4435d9f5045a6fb7bd4b8808461b26aeca60c0047fa5937e1d45d
        version: 1.8.6
      entries:
      - name: openshift-gitops-operator.v1.8.6
        version: 1.8.6
      - name: openshift-gitops-operator.v1.8.5
        version: 1.8.5
      - name: openshift-gitops-operator.v1.8.4-0.1693214052.p
        version: 1.8.4+0.1693214052.p
      - name: openshift-gitops-operator.v1.8.4
        version: 1.8.4
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.8
    - currentCSV: openshift-gitops-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "argo-rollout"
                },
                "spec": null
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:11a6084e11aa32511a34bd6cfb4f4995aec690c1f5ed61861e99f6e42d62afb4
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.skipRange: ""
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.7.16\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:80f92842fe4cba008a873ad2ab6059a3a24e76e16671d8ece034dbbe666996bc
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:3d17786e2f4063154b9e3a40b96bf8ec0a14b86395cea5fe7ca18c5b43764478
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a568c8c647e2e2d6abdf12138eeef8f8bdde1fdcbf786e9ca8c8f93bc79acaf9
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:43bb961d847477902e66a92edd1539322cac8d5f74706b048a458ad037a0c7a3
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:11a6084e11aa32511a34bd6cfb4f4995aec690c1f5ed61861e99f6e42d62afb4
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ec9f60018694dcc5d431ba47d5536b761b71cb3f66684978fe6bb74c157679ac
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:3873d1f18b5643e1de2601419410270342650a21428bb26a4cbee20c0c63f98b
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:86012145d4b8798dc12f6f4f07a9316a193188aaac865e324b63f4e60e8ec584
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:136d3a936759e112d3ea2e78f7c02bd882b0fd3011f4c452f45c32c43822772d
        version: 1.9.4
      entries:
      - name: openshift-gitops-operator.v1.9.4
        version: 1.9.4
      - name: openshift-gitops-operator.v1.9.3
        version: 1.9.3
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.9
    - currentCSV: openshift-gitops-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
          createdAt: "2025-08-21T01:20:45Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.18.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: NamespaceManagement
            name: namespacemanagements.argoproj.io
            version: v1beta1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.1.6\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:adca9c55b80d899d500bf0a7234cf61ae30ffed1b2c927ddea8d9fd07c9acb6a
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7f6ceae56c60fd432425c36f93a9e9bbbb96ef9422333ad6ea41ff3a2eaf8374
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:6284873bcf1593c210003e0292ac5583f03f03e80d937098d808d48fa0d01cf6
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:08efd3e7a79c2945693c99ed31ec912223becff7238b1e28d1b1da5d63f93019
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:640e3298c7cd2faa84971b0d861446c4892ab8602108af05b8b5b049a4de624f
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c2283997d47298d49378e8200b149ea613c7735317ef12e74b7915a590f41ce2
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:5a09fcba3d9e6fc462719e81a46af6022cf7f70a05bcb387ed959a49f6f37dbc
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:1eef9ecd58cf8391f29360ef74dd8aa5dcb48e1582d1500e445706d8302c951e
        - registry.redhat.io/rhel9/redis-7@sha256:b4f58ad83ab8719a2e102f38ee526546cd1d56428901afc3add45c9a0e2a2f12
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a0544c14823492165550d83a6d8ba79dd632b46144d3fdcb543793726111d76
        version: 1.18.0
      entries:
      - name: openshift-gitops-operator.v1.18.0
        version: 1.18.0
      - name: openshift-gitops-operator.v1.17.1
        version: 1.17.1
      - name: openshift-gitops-operator.v1.17.0
        version: 1.17.0
      - name: openshift-gitops-operator.v1.16.2
        version: 1.16.2
      - name: openshift-gitops-operator.v1.16.1
        version: 1.16.1
      - name: openshift-gitops-operator.v1.16.0-0.1746014725.p
        version: 1.16.0+0.1746014725.p
      - name: openshift-gitops-operator.v1.16.0
        version: 1.16.0
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: latest
    defaultChannel: latest
    packageName: openshift-gitops-operator
    provider:
      name: Red Hat Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: node-maintenance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-maintenance-operator.v0.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-maintenance-operator:v0.17.0
          createdAt: "2024-03-13 18:56:47"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=0.12.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/medik8s/node-maintenance-operator:v0.17.0
        version: 0.17.0
      entries:
      - name: node-maintenance-operator.v0.17.0
        version: 0.17.0
      - name: node-maintenance-operator.v0.16.1
        version: 0.16.1
      - name: node-maintenance-operator.v0.16.0
        version: 0.16.0
      - name: node-maintenance-operator.v0.15.0
        version: 0.15.0
      - name: node-maintenance-operator.v0.14.0
        version: 0.14.0
      - name: node-maintenance-operator.v0.13.0
        version: 0.13.0
      - name: node-maintenance-operator.v0.12.1
        version: 0.12.1
      name: stable
    defaultChannel: stable
    packageName: node-maintenance-operator
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudwatch-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudwatch-controller.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudwatch-controller:1.3.0
          createdAt: "2025-10-02T23:05:17Z"
          description: AWS CloudWatch controller is a service controller for managing
            CloudWatch resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Dashboard represents the state of an AWS cloudwatch Dashboard
              resource.
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.cloudwatch.services.k8s.aws
            version: v1alpha1
          - description: MetricAlarm represents the state of an AWS cloudwatch MetricAlarm
              resource.
            displayName: MetricAlarm
            kind: MetricAlarm
            name: metricalarms.cloudwatch.services.k8s.aws
            version: v1alpha1
          - description: MetricStream represents the state of an AWS cloudwatch MetricStream
              resource.
            displayName: MetricStream
            kind: MetricStream
            name: metricstreams.cloudwatch.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudWatch resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudWatch**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudWatch
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudwatch
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudWatch Developer Resources
          url: https://aws.amazon.com/CloudWatch/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudwatch maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudwatch-controller:1.3.0
        version: 1.3.0
      entries:
      - name: ack-cloudwatch-controller.v1.3.0
        version: 1.3.0
      - name: ack-cloudwatch-controller.v1.2.1
        version: 1.2.1
      - name: ack-cloudwatch-controller.v1.2.0
        version: 1.2.0
      - name: ack-cloudwatch-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudwatch-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudwatch-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudwatch-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudwatch-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudwatch-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudwatch-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudwatch-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudwatch-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudwatch-controller.v1.0.5
        version: 1.0.5
      - name: ack-cloudwatch-controller.v1.0.4
        version: 1.0.4
      - name: ack-cloudwatch-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudwatch-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudwatch-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudwatch-controller.v0.0.12
        version: 0.0.12
      - name: ack-cloudwatch-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudwatch-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudwatch-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudwatch-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudwatch-controller.v0.0.3
        version: 0.0.3
      - name: ack-cloudwatch-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudwatch-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OpenShift Database Access
      provider-url: ""
    name: dbaas-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dbaas-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSConnection",
                "metadata": {
                  "name": "dbaasconnection-example"
                },
                "spec": {
                  "databaseServiceID": "instance-example",
                  "databaseServiceType": "instance",
                  "inventoryRef": {
                    "name": "inventory-example",
                    "namespace": "openshift-dbaas-operator"
                  }
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSInstance",
                "metadata": {
                  "name": "dbaasinstance-sample"
                },
                "spec": {
                  "cloudProvider": "AWS",
                  "cloudRegion": "US_EAST_1",
                  "inventoryRef": {
                    "name": "inventory-example",
                    "namespace": "openshift-dbaas-operator"
                  },
                  "name": "my-cluster-free"
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSInventory",
                "metadata": {
                  "name": "inventory-example"
                },
                "spec": {
                  "credentialsRef": {
                    "name": "dbaas-vendor-credentials"
                  },
                  "providerRef": {
                    "name": "provider-example-registration"
                  }
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSPolicy",
                "metadata": {
                  "name": "dbaaspolicy-sample"
                },
                "spec": {
                  "disableProvisions": false
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSProvider",
                "metadata": {
                  "labels": {
                    "related-to": "dbaas-operator",
                    "type": "dbaas-provider-registration"
                  },
                  "name": "provider-example-registration"
                },
                "spec": {
                  "allowsFreeTrial": true,
                  "connectionKind": "ProviderConnection",
                  "credentialFields": [
                    {
                      "displayName": "ProviderAPIKeyD",
                      "helpText": "This is the API Key for the example provider to show on UI",
                      "key": "APIKey",
                      "required": true,
                      "type": "maskedstring"
                    }
                  ],
                  "externalProvisionDescription": "Follow the guide to start a free Provider Serverless (beta) cluster",
                  "externalProvisionURL": "https://www.exmample.com/docs/provider/quickstart.html",
                  "groupVersion": "dbaas.redhat.com/v1beta1",
                  "instanceKind": "ProviderInstance",
                  "inventoryKind": "ProviderInventory",
                  "provider": {
                    "displayDescription": "This is an example for providers on how to implement their operator to integrate with DBaaS.",
                    "displayName": "DBaaS Provider Example ",
                    "icon": {
                      "base64data": "SGVsbG8sIHdvcmxkLg==",
                      "mediatype": "image/png"
                    },
                    "name": "Provider Example"
                  },
                  "provisioningParameters": {
                    "cloudProvider": {
                      "conditionalData": [
                        {
                          "defaultValue": "GCP",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "FREETRIAL"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        },
                        {
                          "defaultValue": "AWS",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "SERVERLESS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Amazon Web Services",
                              "value": "AWS"
                            },
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        },
                        {
                          "defaultValue": "AWS",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Amazon Web Services",
                              "value": "AWS"
                            },
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        }
                      ],
                      "displayName": "Cloud Provider"
                    },
                    "dedicatedLocationLabel": {
                      "displayName": "Select regions \u0026 nodes",
                      "helpText": "Select the geographical region where you want the database instance to run, and set the number of nodes you want running in this dedicated cluster."
                    },
                    "hardwareLabel": {
                      "displayName": "Hardware per node",
                      "helpText": "Select the compute and storage requirements for this database instance."
                    },
                    "machineType": {
                      "conditionalData": [
                        {
                          "defaultValue": "m5.large",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "AWS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "2 vCPU, 8 GiB RAM",
                              "value": "m5.large"
                            },
                            {
                              "displayValue": "4 vCPU, 16 GiB RAM",
                              "value": "m5.xlarge"
                            },
                            {
                              "displayValue": "8 vCPU, 32 GiB RAM",
                              "value": "m5.2xlarge"
                            },
                            {
                              "displayValue": "16 vCPU, 64 GiB RAM",
                              "value": "m5.4xlarge"
                            },
                            {
                              "displayValue": "32 vCPU, 128 GiB RAM",
                              "value": "m5.8xlarge"
                            }
                          ]
                        },
                        {
                          "defaultValue": "n1-standard-2",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "GCP"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "2 vCPU, 7.5 GiB RAM",
                              "value": "n1-standard-2"
                            },
                            {
                              "displayValue": "4 vCPU, 15 GiB RAM",
                              "value": "n1-standard-4"
                            },
                            {
                              "displayValue": "8 vCPU, 30 GiB RAM",
                              "value": "n1-standard-8"
                            },
                            {
                              "displayValue": "16 vCPU, 60 GiB RAM",
                              "value": "n1-standard-16"
                            },
                            {
                              "displayValue": "32 vCPU, 120 GiB RAM",
                              "value": "n1-standard-32"
                            }
                          ]
                        }
                      ],
                      "displayName": "Compute"
                    },
                    "name": {
                      "displayName": "Cluster name"
                    },
                    "nodes": {
                      "displayName": "nodes"
                    },
                    "plan": {
                      "conditionalData": [
                        {
                          "defaultValue": "SERVERLESS",
                          "options": [
                            {
                              "displayValue": "Free trial",
                              "value": "FREETRIAL"
                            },
                            {
                              "displayValue": "Serverless",
                              "value": "SERVERLESS"
                            },
                            {
                              "displayValue": "Dedicated",
                              "value": "DEDICATED"
                            }
                          ]
                        }
                      ],
                      "displayName": "Hosting plan"
                    },
                    "planLabel": {
                      "displayName": "Select a plan"
                    },
                    "regions": {
                      "displayName": "us-west-2"
                    },
                    "serverlessLocationLabel": {
                      "displayName": "Select regions",
                      "helpText": "Select the geographical region where you want the database instance to run."
                    },
                    "spendLimit": {
                      "conditionalData": [
                        {
                          "defaultValue": "0",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "SERVERLESS"
                            }
                          ]
                        }
                      ],
                      "displayName": "Spend limit"
                    },
                    "spendLimitLabel": {
                      "displayName": "Spend limit",
                      "helpText": "Set a spending limit on resources for this database instance."
                    },
                    "storageGib": {
                      "conditionalData": [
                        {
                          "defaultValue": "15",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "AWS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "15 GiB",
                              "value": "15"
                            },
                            {
                              "displayValue": "35 GiB",
                              "value": "35"
                            },
                            {
                              "displayValue": "75 GiB",
                              "value": "75"
                            },
                            {
                              "displayValue": "150 GiB",
                              "value": "150"
                            },
                            {
                              "displayValue": "300 GiB",
                              "value": "300"
                            },
                            {
                              "displayValue": "600 GiB",
                              "value": "600"
                            }
                          ]
                        },
                        {
                          "defaultValue": "15",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "GCP"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "15 GiB",
                              "value": "15"
                            },
                            {
                              "displayValue": "35 GiB",
                              "value": "35"
                            },
                            {
                              "displayValue": "75 GiB",
                              "value": "75"
                            },
                            {
                              "displayValue": "150 GiB",
                              "value": "150"
                            },
                            {
                              "displayValue": "300 GiB",
                              "value": "300"
                            },
                            {
                              "displayValue": "600 GiB",
                              "value": "600"
                            }
                          ]
                        }
                      ],
                      "displayName": "Storage"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: quay.io/ecosystem-appeng/dbaas-operator:v0.5.0
          description: Discover & connect with database instances hosted on 3rd-party
            ISV cloud platforms.
          operatorframework.io/suggested-namespace: openshift-dbaas-operator
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/internal-objects: '["dbaasconnections.dbaas.redhat.com","dbaasproviders.dbaas.redhat.com","dbaasplatforms.dbaas.redhat.com","dbaasinstances.dbaas.redhat.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/RHEcosystemAppEng/dbaas-operator
          support: rhoda-requests@redhat.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DBaaSConnection
            name: dbaasconnections.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSConnection defines the schema for the DBaaSConnection
              API.
            displayName: DBaaSConnection
            kind: DBaaSConnection
            name: dbaasconnections.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSInstance
            name: dbaasinstances.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSInstance defines the schema for the DBaaSInstance API.
            displayName: DBaaSInstance
            kind: DBaaSInstance
            name: dbaasinstances.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSInventory
            name: dbaasinventories.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSInventory defines the schema for the DBaaSInventory
              API. Inventory objects must be created in a valid namespace, determined
              by the existence of a DBaaSPolicy object.
            displayName: Provider Account
            kind: DBaaSInventory
            name: dbaasinventories.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSPlatform
            name: dbaasplatforms.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSPlatform defines the schema for the DBaaSPlatform API.
            displayName: DBaaSPlatform
            kind: DBaaSPlatform
            name: dbaasplatforms.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSPolicy
            name: dbaaspolicies.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSPolicy enables administrative capabilities within a
              namespace, and sets a default inventory policy. Policy defaults can
              be overridden on a per-inventory basis.
            displayName: Provider Account Policy
            kind: DBaaSPolicy
            name: dbaaspolicies.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSProvider
            name: dbaasproviders.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSProvider defines the schema for the DBaaSProvider API.
            displayName: DBaaSProvider
            kind: DBaaSProvider
            name: dbaasproviders.dbaas.redhat.com
            version: v1beta1
        description: |
          The OpenShift Database Access Operator enables OpenShift users to discover & connect with database instances
          hosted on 3rd-party ISV cloud platforms such as CrunchyData Bridge & CockroachCloud.

          ## Feedback & Support

          Please contact us at [rhoda-requests@redhat.com](mailto:rhoda-requests@redhat.com)

          For more information on how to use our service, please refer to our [Quick Starts](quickstart?keyword=database).
        displayName: OpenShift Database Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rhoda
        - oda
        - dbaas
        - database
        - db
        - data
        - access
        - crunchy
        - bridge
        - cockroach
        - cockroachdb
        - cloud
        - ack
        - amazon
        - rds
        - relational
        links:
        - name: GitHub
          url: https://github.com/RHEcosystemAppEng/dbaas-operator
        maintainers:
        - email: ecosystem-appeng@redhat.com
          name: Red Hat Ecosystem App Engineering
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: OpenShift Database Access
        relatedImages:
        - quay.io/openshift/origin-kube-rbac-proxy:4.10.0
        - quay.io/ecosystem-appeng/dbaas-dynamic-plugin:0.5.0
        - registry.developers.crunchydata.com/crunchydata/crunchy-bridge-operator-catalog:v0.0.6
        - gcr.io/cockroach-shared/ccapi-k8s-operator-catalog:v0.0.5
        - quay.io/ecosystem-appeng/rds-dbaas-operator-catalog:v0.4.0
        - quay.io/ecosystem-appeng/dbaas-operator:v0.5.0
        version: 0.5.0
      entries:
      - name: dbaas-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: dbaas-operator
    provider:
      name: OpenShift Database Access
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ControlPlane Ltd.
      provider-url: https://control-plane.io
    name: flux-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux-operator.v0.31.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "FluxInstance",
                "metadata": {
                  "name": "flux",
                  "namespace": "flux-system",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcileEvery": "1h",
                    "fluxcd.controlplane.io/reconcileTimeout": "5m"
                  }
                },
                "spec": {
                  "distribution": {
                    "version": "2.x",
                    "registry": "ghcr.io/fluxcd"
                  },
                  "components": [
                    "source-controller",
                    "kustomize-controller",
                    "helm-controller",
                    "notification-controller",
                    "image-reflector-controller",
                    "image-automation-controller"
                  ],
                  "cluster": {
                    "type": "openshift",
                    "size": "medium",
                    "domain": "cluster.local",
                    "networkPolicy": true,
                    "multitenant": true,
                    "tenantDefaultServiceAccount": "flux"
                  }
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "FluxReport",
                "metadata": {
                  "name": "flux",
                  "namespace": "flux-system"
                },
                "spec": {
                  "distribution": {
                    "entitlement": "Issued by controlplane",
                    "managedBy": "flux-operator",
                    "status": "Installed",
                    "version": "v2.3.0"
                  },
                "components": [
                    {
                      "image": "ghcr.io/fluxcd/kustomize-controller:v1.3.0@sha256:48a032574dd45c39750ba0f1488e6f1ae36756a38f40976a6b7a588d83acefc1",
                      "name": "kustomize-controller",
                      "ready": true,
                      "status": "Current Deployment is available. Replicas: 1"
                    },
                    {
                      "image": "ghcr.io/fluxcd/source-controller:v1.3.0@sha256:161da425b16b64dda4b3cec2ba0f8d7442973aba29bb446db3b340626181a0bc",
                      "name": "source-controller",
                      "ready": true,
                      "status": "Current Deployment is available. Replicas: 1"
                    }
                  ],
                "reconcilers": [
                    {
                      "apiVersion": "kustomize.toolkit.fluxcd.io/v1",
                      "kind": "Kustomization",
                      "stats": {
                        "failing": 0,
                        "running": 1,
                        "suspended": 0
                      }
                    },
                    {
                      "apiVersion": "source.toolkit.fluxcd.io/v1",
                      "kind": "GitRepository",
                      "stats": {
                        "failing": 0,
                        "running": 1,
                        "suspended": 0,
                        "totalSize": "3.7 MiB"
                      }
                    }
                  ],
                  "sync": {
                    "ready": true,
                    "id": "kustomization/flux-system",
                    "path": "clusters/production",
                    "source": "https://github.com/my-org/my-fleet.git",
                    "status": "Applied revision: refs/heads/main@sha1:a90cd1ac35de01c175f7199315d3f4cd60195911"
                  }
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "ResourceSet",
                "metadata": {
                  "name": "podinfo",
                  "namespace": "default",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcile": "enabled",
                    "fluxcd.controlplane.io/reconcileEvery": "30m",
                    "fluxcd.controlplane.io/reconcileTimeout": "5m"
                  }
                },
                "spec": {
                  "commonMetadata": {
                    "labels": {
                      "app.kubernetes.io/name": "podinfo"
                    }
                  },
                  "inputs": [
                    {
                      "tenant": "team1",
                      "version": "6.7.x",
                      "replicas": "2"
                    },
                    {
                      "tenant": "team2",
                      "version": "6.6.x",
                      "replicas": "3"
                    }
                  ],
                  "resources": [
                    {
                      "apiVersion": "source.toolkit.fluxcd.io/v1beta2",
                      "kind": "OCIRepository",
                      "metadata": {
                        "name": "podinfo-<< inputs.tenant >>",
                        "namespace": "default"
                      },
                      "spec": {
                        "interval": "10m",
                        "url": "oci://ghcr.io/stefanprodan/charts/podinfo",
                        "ref": {
                          "semver": "<< inputs.version | quote >>"
                        }
                      }
                    },
                    {
                      "apiVersion": "helm.toolkit.fluxcd.io/v2",
                      "kind": "HelmRelease",
                      "metadata": {
                        "name": "podinfo-<< inputs.tenant >>",
                        "namespace": "default"
                      },
                      "spec": {
                        "interval": "1h",
                        "releaseName": "podinfo-<< inputs.tenant >>",
                        "chartRef": {
                          "kind": "OCIRepository",
                          "name": "podinfo-<< inputs.tenant >>"
                        },
                        "values": {
                          "replicaCount": "<< inputs.replicas | int >>"
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "ResourceSetInputProvider",
                "metadata": {
                  "name": "demo",
                  "namespace": "default",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcile": "enabled",
                    "fluxcd.controlplane.io/reconcileEvery": "10m",
                    "fluxcd.controlplane.io/reconcileTimeout": "1m"
                  }
                },
                "spec": {
                  "type": "GitHubPullRequest",
                  "url": "https://github.com/fluxcd-testing/pr-testing",
                  "filter": {
                    "limit": 10,
                    "includeBranch": "^stefanprodan-patch-.*$",
                    "labels": [
                      "enhancement"
                    ]
                  },
                  "skip": {
                    "labels": [
                        "deploy/flux-preview-pause",
                        "!test-build-push/passed"
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/controlplaneio-fluxcd/flux-operator:v0.31.0-ubi
          createdAt: "2025-10-06T20:33:03Z"
          description: The Flux Operator manages the lifecycle of the CNCF Flux project.
          operatorframework.io/suggested-namespace: flux-system
          repository: https://github.com/controlplaneio-fluxcd/flux-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Flux Instance
            displayName: FluxInstance
            kind: FluxInstance
            name: fluxinstances.fluxcd.controlplane.io
            version: v1
          - description: Flux Report (Autogenerated)
            displayName: FluxReport
            kind: FluxReport
            name: fluxreports.fluxcd.controlplane.io
            version: v1
          - description: Resource Set
            displayName: ResourceSet
            kind: ResourceSet
            name: resourcesets.fluxcd.controlplane.io
            version: v1
          - description: ResourceSet Input Provider
            displayName: ResourceSetInputProvider
            kind: ResourceSetInputProvider
            name: resourcesetinputproviders.fluxcd.controlplane.io
            version: v1
        description: |-
          [Flux](https://fluxcd.io) is a leading GitOps Continuous Delivery tool used to streamline and automate application deployments on Kubernetes. Flux enables teams to achieve a reliable and auditable delivery process while promoting collaboration and traceability across different environments.
          Flux is powered by the GitOps Toolkit, a set of composable APIs and specialized tools that enable a wide range of continuous delivery use-cases, from simple Kubernetes deployment pipelines to multi-tenant and multi-cluster progressive delivery rollouts.
          For more information about the CNCF-graduated Flux project, please see the [Flux architecture overview](https://fluxcd.control-plane.io/guides/flux-architecture/).
          ### Operator Capabilities
          The [Flux Operator](https://github.com/controlplaneio-fluxcd/flux-operator) provides a declarative API for the installation, configuration and upgrade of CNCF Flux and the ControlPlane [enterprise distribution](https://fluxcd.control-plane.io/).
          The operator allows the configuration of Flux multi-tenancy lockdown, network policies, persistent storage, sharding, vertical scaling, custom patches, and the synchronization of the cluster state from Git repositories, OCI artifacts and S3-compatible storage.
          The Flux Operator supervises the Flux controllers and provides a unified view of all the Flux resources that define the GitOps workflows for the target cluster. The operator generates reports, emits events, and exports Prometheus metrics to help with monitoring and troubleshooting Flux.
          ### Operator Features
          The operator extends Flux with self-service capabilities, deployment windows, and preview environments for GitHub, GitLab and Azure DevOps pull requests testing.
          The operator ResourceSet API enables platform teams to define their own application standard as a group of Flux and Kubernetes resources that can be templated, parameterized and deployed as a single unit on self-service environments. For more information, please see the [ResourceSet documentation](https://fluxcd.control-plane.io/operator/resourcesets/introduction/).
          ### OpenShift Support
          The Flux Operator should be installed in a dedicated namespace, e.g. `flux-system`.
          To deploy Flux on OpenShift clusters, create a `FluxInstance` custom resource named `flux` in the same namespace as the operator and set the `.spec.cluster.type` field to `openshift`.
          For more information on how to configure the Flux instance, please see the [Flux Operator documentation](https://fluxcd.control-plane.io/operator/flux-config/).
        displayName: Flux Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - fluxcd
        - gitops
        - continuous delivery
        links:
        - name: Source Code
          url: https://github.com/controlplaneio-fluxcd/flux-operator
        - name: Documentation
          url: https://fluxcd.control-plane.io/operator/
        - name: Enterprise Support
          url: https://fluxcd.control-plane.io/pricing/
        maintainers:
        - email: flux-enterprise@control-plane.io
          name: ControlPlane Flux Team
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: ControlPlane Ltd.
          url: https://control-plane.io
        relatedImages:
        - ghcr.io/controlplaneio-fluxcd/flux-operator:v0.31.0-ubi
        version: 0.31.0
      entries:
      - name: flux-operator.v0.31.0
        version: 0.31.0
      - name: flux-operator.v0.30.0
        version: 0.30.0
      - name: flux-operator.v0.28.0
        version: 0.28.0
      - name: flux-operator.v0.27.0
        version: 0.27.0
      - name: flux-operator.v0.26.0
        version: 0.26.0
      - name: flux-operator.v0.24.1
        version: 0.24.1
      - name: flux-operator.v0.24.0
        version: 0.24.0
      - name: flux-operator.v0.23.0
        version: 0.23.0
      - name: flux-operator.v0.22.0
        version: 0.22.0
      - name: flux-operator.v0.21.0
        version: 0.21.0
      - name: flux-operator.v0.20.0
        version: 0.20.0
      - name: flux-operator.v0.19.0
        version: 0.19.0
      - name: flux-operator.v0.18.0
        version: 0.18.0
      name: stable
    defaultChannel: stable
    packageName: flux-operator
    provider:
      name: ControlPlane Ltd.
      url: https://control-plane.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-load-core
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-load-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Loadcore",
                "metadata": {
                  "name": "loadcore-sample"
                },
                "spec": {
                  "accessRoute": {
                    "enabled": true,
                    "host": "loadcore.apps.cluster.perf.keysight.lab"
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "reportTemplates": {
                    "enabled": false
                  },
                  "wirelessDataModel": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:42:44Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Loadcore
            name: loadcores.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight LoadCore Operator
        displayName: Keysight LoadCore Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Load Core
          url: https://keysight-load-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-load-core@sha256:245d9be60ff6e9a6e5e42a026a4ee7b8168f391a0491de23ae827d1906af88e5
        version: 0.1.21
      entries:
      - name: keysight-load-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-load-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CloudBees
      provider-url: https://www.cloudbees.com/products/continuous-integration
    name: cloudbees-ci-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudbees-ci.v3.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.cloudbees.com/v1alpha1",
                "kind": "CloudBeesCI",
                "metadata": {
                  "name": "cloudbeesci-sample"
                },
                "spec": {
                  "Agents": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-agent:2.289.3.2"
                    },
                    "ImagePullSecrets": null,
                    "SeparateNamespace": {
                      "Create": false,
                      "Enabled": false,
                      "Name": null
                    }
                  },
                  "Hibernation": {
                    "Enabled": false,
                    "Image": {
                      "dockerImage": "cloudbees/managed-master-hibernation-monitor:247.c5dfce00a179",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "NodeSelector": {},
                    "Tolerations": []
                  },
                  "Master": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-mm:2.289.3.2"
                    },
                    "JavaOpts": null,
                    "OperationsCenterNamespace": null
                  },
                  "NetworkPolicy": {
                    "Enabled": false,
                    "JMXSelectors": [],
                    "ingressControllerSelector": []
                  },
                  "OperationsCenter": {
                    "AgentListenerPort": 50000,
                    "Annotations": {},
                    "CSRF": {
                      "ProxyCompatibility": false
                    },
                    "CasC": {
                      "ConfigMapName": "oc-casc-bundle",
                      "Enabled": false
                    },
                    "ContainerPort": 8080,
                    "ContextPath": null,
                    "Enabled": true,
                    "ExtraConfigMaps": [],
                    "ExtraContainers": [],
                    "ExtraGroovyConfiguration": {},
                    "ExtraVolumeMounts": [],
                    "ExtraVolumes": [],
                    "HealthProbeLivenessFailureThreshold": 12,
                    "HealthProbes": true,
                    "HostName": null,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-cloud-core-oc:2.289.3.2",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "Ingress": {
                      "Annotations": {
                        "kubernetes.io/tls-acme": "false"
                      },
                      "Class": "nginx",
                      "tls": {
                        "Enable": false,
                        "SecretName": null
                      }
                    },
                    "JavaOpts": null,
                    "JenkinsOpts": null,
                    "LoadBalancerIP": null,
                    "LoadBalancerSourceRanges": [
                      "0.0.0.0/0"
                    ],
                    "Name": "cjoc",
                    "NodeSelector": {},
                    "Platform": "standard",
                    "Protocol": "http",
                    "Resources": {
                      "Limits": {
                        "Cpu": 1,
                        "Memory": "2G"
                      },
                      "Requests": {
                        "Cpu": 1,
                        "Memory": "2G"
                      }
                    },
                    "Route": {
                      "tls": {
                        "CACertificate": null,
                        "Certificate": null,
                        "DestinationCACertificate": null,
                        "Enable": false,
                        "InsecureEdgeTerminationPolicy": "Redirect",
                        "Key": null,
                        "Termination": "edge"
                      }
                    },
                    "ServiceAgentListenerPort": 50000,
                    "ServiceAnnotations": {},
                    "ServicePort": 80,
                    "ServiceType": "ClusterIP",
                    "Tolerations": []
                  },
                  "Persistence": {
                    "AccessMode": "ReadWriteOnce",
                    "Annotations": {},
                    "Size": "20Gi",
                    "StorageClass": null
                  },
                  "PodSecurityPolicy": {
                    "Annotations": {},
                    "Enabled": false
                  },
                  "Subdomain": false,
                  "ingress-nginx": {
                    "Enabled": false,
                    "controller": {
                      "admissionWebhooks": {
                        "patch": {
                          "nodeSelector": {
                            "kubernetes.io/os": "linux"
                          }
                        }
                      },
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "nginx-ingress": {
                    "Enabled": false,
                    "controller": {
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "rbac": {
                    "agentsServiceAccountName": "jenkins-agents",
                    "hibernationMonitorServiceAccountName": "managed-master-hibernation-monitor",
                    "install": true,
                    "masterServiceAccountName": "jenkins",
                    "serviceAccountName": "cjoc"
                  },
                  "sidecarinjector": {
                    "Enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cloudbees-ci-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cloudbees-ci-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudBeesCI
            name: cloudbeescis.charts.cloudbees.com
            version: v1alpha1
        description: CloudBees CI
        displayName: CloudBees CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudbees
        - ci
        - enterprise
        - continuous
        - integration
        - jenkins
        - automation
        - pipeline
        - devops
        links:
        - name: CloudBees CI
          url: https://www.cloudbees.com/products/continuous-integration
        maintainers:
        - email: support@cloudbees.com
          name: Support
        - email: sales@cloudbees.com
          name: Sales
        maturity: alpha
        provider:
          name: CloudBees
          url: https://www.cloudbees.com/products/continuous-integration
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.marketplace.redhat.com/rhm/cloudbees/cloudbees-ci-operator@sha256:bd0bd72e9e7807b0fa307de8a93a56967221fe2dd181bac02383f0f9be9120e4
        version: 3.34.1
      entries:
      - name: cloudbees-ci.v3.34.1
        version: 3.34.1
      name: alpha
    defaultChannel: alpha
    packageName: cloudbees-ci-rhmp
    provider:
      name: CloudBees
      url: https://www.cloudbees.com/products/continuous-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhacs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhacs-operator.v3.74.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
          createdAt: "2024-02-06T23:06:13.271846+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.73.0 < 3.74.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:2c0840f4573373a19eb8c9a26501138ef0e76efbd508fd69a6260e78194d86c8
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:0482d4464d6d39c01abab8df882604dd227c52b5420d36fe7414aaff8067cbec
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:71b9fdcaf4cc761b1499be8e6ccff6c463cc6b9856e9deb1f93baad45ea59eb0
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:74624389ac3261ed699b1f64e5717353da68895709d30a011bc8813b61a4a249
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:120b9891ee9db839ab0b2029a62e67ec4747c47c6ffe15af38e9035160f57b78
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:22bf00e1457b4a9e63901892aa4375a1d4a9d1b2c067c1f054d918a24b222b10
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:a108e9a7967001e0bdfed7124b01e9a3dbf1dec308444762973030aec57e1d5a
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:87eb62fe8aa7952d0d69dc13c4867ac2f51e821f56d149a9b588c826caec4c16
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d546f2d3f668ce2f20e594394ade0a86d808f6c8c88cadbba227083d4abac940
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:264102d4bca7d9bd384dc04e5bbff1a7b43343ea8e59cc79047cdda50a3f299f
        version: 3.74.9
      entries:
      - name: rhacs-operator.v3.74.9
        version: 3.74.9
      - name: rhacs-operator.v3.74.8
        version: 3.74.8
      - name: rhacs-operator.v3.74.7
        version: 3.74.7
      - name: rhacs-operator.v3.74.6
        version: 3.74.6
      - name: rhacs-operator.v3.74.5
        version: 3.74.5
      - name: rhacs-operator.v3.74.4
        version: 3.74.4
      - name: rhacs-operator.v3.74.3
        version: 3.74.3
      - name: rhacs-operator.v3.74.2
        version: 3.74.2
      - name: rhacs-operator.v3.74.1
        version: 3.74.1
      - name: rhacs-operator.v3.74.0
        version: 3.74.0
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: latest
    - currentCSV: rhacs-operator.v3.62.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:56905c2e7df0e29ca2b6301cd81b4f9bbe02b483374708f7f7da219a4b29428d
          createdAt: "2021-08-13T17:10:01.410173+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.61.0 < 3.62.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.5.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main page before deploying, or consult the RHACS
              documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create an cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead. If you want to deploy Secured Cluster Service in another namespace, additionally pass the `-n <namespace>` flag.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e30f7835cf52dbce9afdbaf55e5a93d8baf9c4b3a3ae488d1d42c1adba7f46af
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:56905c2e7df0e29ca2b6301cd81b4f9bbe02b483374708f7f7da219a4b29428d
        - registry.redhat.io/rh-acs/collector@sha256:98cfab15684506f1d0978eb701e8de478d91fc5bda3013ffd3b436aa95f07cc0
        - registry.redhat.io/rh-acs/main@sha256:9932304bb62b084ee1ae7e75bb21d6dfb737260c851f2d8daf05082f79230c66
        - registry.redhat.io/rh-acs/scanner@sha256:266a575c2547a3223890e8b73db0f14d69dc5a22217c7e4e3dc79644e6ab0258
        - registry.redhat.io/rh-acs/scanner-db@sha256:ff877f503e08ecca0ce7a7d8db760b2e4004e8e8ae2f27fb9b9c00a05b58c93a
        - registry.redhat.io/rh-acs/collector@sha256:558c0dabdc52efaee7bc889ef397ac2ac30f7c54562c2cc0070d0ef991a8b31b
        version: 3.62.1
      entries:
      - name: rhacs-operator.v3.62.1
        version: 3.62.1
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.62
    - currentCSV: rhacs-operator.v3.64.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fa478167f2f738f244bed6a06450679be73f8464c40f86acde4ffd390505692b
          createdAt: "2021-09-22T05:10:01.286657+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.63.0 < 3.64.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0cacd00c0a147cc8295138565ef2dc2ddbd9ad1fb25f6be817719a75ea4c6448
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fa478167f2f738f244bed6a06450679be73f8464c40f86acde4ffd390505692b
        - registry.redhat.io/rh-acs/scanner@sha256:8b7121bce8f9c7133d29245ebb265a4b8f9831a66a57f11b2d68b9b8b5cf79cc
        - registry.redhat.io/rh-acs/scanner-db@sha256:67e81bf896267b7acab8cb6755cde552173de4f0881c38eaca8e491bd926cc53
        - registry.redhat.io/rh-acs/collector@sha256:2b3a8bd3af62516a6714fe612e5fdc0433b91d873182e290d3ddc2b32ce291f0
        - registry.redhat.io/rh-acs/collector@sha256:3eb099da081cf208d46be5ccfa7b5e9ac0e2cad49f127fa4db91e49e3f26db42
        - registry.redhat.io/rh-acs/main@sha256:6aeb2223124540ac217c38b016dea99d770ab897dde4906a821d8776d91ea03d
        version: 3.64.2
      entries:
      - name: rhacs-operator.v3.64.2
        version: 3.64.2
      - name: rhacs-operator.v3.64.1
        version: 3.64.1
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.64
    - currentCSV: rhacs-operator.v3.65.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:229c7e5f967c70dfbc3cf638d521267ea6acf875b055499dfd30cb19c0b1f69c
          createdAt: "2021-09-21T13:26:28.711610+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.64.0 < 3.65.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/collector@sha256:3eb099da081cf208d46be5ccfa7b5e9ac0e2cad49f127fa4db91e49e3f26db42
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0cacd00c0a147cc8295138565ef2dc2ddbd9ad1fb25f6be817719a75ea4c6448
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:229c7e5f967c70dfbc3cf638d521267ea6acf875b055499dfd30cb19c0b1f69c
        - registry.redhat.io/rh-acs/main@sha256:e672ef9a1239065f8cf0f331757861fea082e971a9e9760d1e65795f399cbde1
        - registry.redhat.io/rh-acs/scanner@sha256:bfe9b2205bfca6aebed14827b4bda098bfad574e67ffa967ad339543062200fe
        - registry.redhat.io/rh-acs/scanner-db@sha256:20f258500ac2ba4e626df92dbcd193cf7ab8687ac251591be2442b8a46871835
        - registry.redhat.io/rh-acs/collector@sha256:2b3a8bd3af62516a6714fe612e5fdc0433b91d873182e290d3ddc2b32ce291f0
        version: 3.65.1
      entries:
      - name: rhacs-operator.v3.65.1
        version: 3.65.1
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.65
    - currentCSV: rhacs-operator.v3.66.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:10cdb0ff3e67b65737cde420536494bd84c2b7ccf65192bbed2eb43eaed66a41
          createdAt: "2021-10-20T16:38:38.091764+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.65.0 < 3.66.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bd74f0b95ff2881faf6eb4520feecd067b69a759f09bbfecd467447ce2a2a25e
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:10cdb0ff3e67b65737cde420536494bd84c2b7ccf65192bbed2eb43eaed66a41
        - registry.redhat.io/rh-acs/main@sha256:97230c01e76d90c7fe35ffeaa591be3d9175f350a8fd88599aded09a1edf930c
        - registry.redhat.io/rh-acs/scanner@sha256:323ee86d6a9cbeb32270c659c1154f3a62c36535c6ea894e4e29e93eb1f4a84f
        - registry.redhat.io/rh-acs/scanner-db@sha256:490b0bf49161e4bc6de2504726aaf360e0ca9a214aea6b886fff13153bdb263d
        - registry.redhat.io/rh-acs/collector@sha256:5673c27554a71502e06e08945a4424c6ce5da27f489854f3aa452e6d31240aee
        - registry.redhat.io/rh-acs/collector@sha256:d93e13ee1d1cc4b7f82e644a7d43e560aac42bd1e06d25b476cea9a4338bf9ce
        version: 3.66.1
      entries:
      - name: rhacs-operator.v3.66.1
        version: 3.66.1
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.66
    - currentCSV: rhacs-operator.v3.67.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:893ebcb2d0c96647e56bce3b16fe32851df59cd71f2c0fc8f6e66ca3a669bb70
          createdAt: "2021-12-16T12:54:10.117625+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.66.0 < 3.67.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/scanner@sha256:2860ed271da6064149bb53568c5ad60e129918d55a2f244f9ade4a0bea4c4933
        - registry.redhat.io/rh-acs/scanner-db@sha256:6f8d47277732eb56317d576c2d06fb1f539d46d16520e6ca34abec290392cef9
        - registry.redhat.io/rh-acs/collector@sha256:57c2dfa3dbc8368076291fbc627baa01dcb5ebc50f838e02ce7ac38a128d027b
        - registry.redhat.io/rh-acs/collector@sha256:e4648723ccbd43fa5acf005d9734bfc06d41e19e1be0bde839880e28fa361e16
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86e5fa1fa294987114be200890c2e516501e424aee0fb98ece25c95e7716295b
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:893ebcb2d0c96647e56bce3b16fe32851df59cd71f2c0fc8f6e66ca3a669bb70
        - registry.redhat.io/rh-acs/main@sha256:8db7630d094b661411ebb3f35248848d3147c735677d0922f7a04de9a41eacc9
        version: 3.67.2
      entries:
      - name: rhacs-operator.v3.67.2
        version: 3.67.2
      - name: rhacs-operator.v3.67.1
        version: 3.67.1
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.67
    - currentCSV: rhacs-operator.v3.68.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:195a694dc96096166fe82e81ecb3ceb493f96cfb5029112e576e44ab39b56abd
          createdAt: "2022-06-15T22:13:22.427644+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.67.0 < 3.68.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:195a694dc96096166fe82e81ecb3ceb493f96cfb5029112e576e44ab39b56abd
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:f0d2f5f11c2a96d34d6b4baf249cbf821fe064eade73c3f95c22f3c6981ac4d1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:105d10bd50a6062c998ed107dea388dd78f0caddb0b57713c8692b1d07042774
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:081f2198ad70bfb2ec66153b50b48b950b9a5550ba996de413c5a2aff4052783
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c9d0a5a2871c24c0b1058e419f2b090429363a319856397855412fd2c0037395
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:9b9b9146d9683d6da0ff2abbc8ea82b96d945b7a8c3740d72d8c018fb2b04fba
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:0bb76363a3a25916dd4493e177e9b2d36eac9633473b07fe0fffb853fd072da3
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:711293438bab35b17d2e1f162e7f74f667a6e16dd32e45122f917b6817fefb4e
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:b5fc3dcaa72bbc8dea005189c940e1cf2d00c94a6cf7676cf0e1790e1191d45c
        version: 3.68.2
      entries:
      - name: rhacs-operator.v3.68.2
        version: 3.68.2
      - name: rhacs-operator.v3.68.1
        version: 3.68.1
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.68
    - currentCSV: rhacs-operator.v3.69.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9b2beceb255124855ba1b352d16009b2f2cd422901a19c311f0783dab1fcc3a0
          createdAt: "2022-06-21T09:07:40.705944+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.68.0 < 3.69.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:08a7d8c892fad69c589b81ad8fbd21d36c8fa4078530dba361cfa9a8f9f05950
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:9ae03669cfdc0f4fe56b021d911edb17159c07757720ae30a1f22d7f7a897c07
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:f167e928740d1c614e26948ea58f49e569f6a4fddc438cd0abe52f60538432d0
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:867613178d491f416e57975843817251490c5fbca2fc69d8b55b169bea675279
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:8db5206532e95a7ad7cb23f389155d41757aaf5f445ddce92b008a06627d3fd5
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:77f5db63e17b140a8d6b84e3822fd77463035be9c0c832b55b4977696da13ccc
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:da6dd486a20fdba61ea12d2c77ea299d9795a9f22d9057d65ff992b1f5896e07
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:b254746ed3faa1f1cbcafdf1343a1957bf0214311b36f4e892be3000cc2aecf3
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9b2beceb255124855ba1b352d16009b2f2cd422901a19c311f0783dab1fcc3a0
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:2afb2823edc6f3434e18db60ee37381b5b79f51234c1a261b6d0f8ed74cc8b63
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:cf5629823f2dd583495bcbbcf3b83b4f29f855ace3c388f5cb7b24a6d34ed267
        version: 3.69.2
      entries:
      - name: rhacs-operator.v3.69.2
        version: 3.69.2
      - name: rhacs-operator.v3.69.1
        version: 3.69.1
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.69
    - currentCSV: rhacs-operator.v3.70.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fb903bf9e110ba03e9c63977ebaf707e6884dd8fd413bcd6b793a0ed795cdd0e
          createdAt: "2022-06-20T12:30:26.999648+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.69.0 < 3.70.1'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:dfeb239931f7a33ade5c829a102cabc01f5c348e1aa35a7ddc24757cfd549629
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:df8cfdaa40b50ff2f2670d97c2115b47ae4bea443ae2831caaa468dd62f3c75f
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fb903bf9e110ba03e9c63977ebaf707e6884dd8fd413bcd6b793a0ed795cdd0e
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:6c32b86ed8bc18e384706eb6188406e66910bf34c8869bd155a60c3525d0fea7
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:d2e827a209840b7197960c8115720972237532eb9cf9d6fee7a36107053408ce
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:3d7c8b77671740e324ee6c309a84bbaa4d925e6750979fe4e46c561bc79f1b4d
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:b47c96af1c9ffe4d8728da510a93efe21adc221225fe2b822cfcace6c6672ba2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:08a7d8c892fad69c589b81ad8fbd21d36c8fa4078530dba361cfa9a8f9f05950
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:ee55a233d2a8fd77d89ec2c69f357721c7debead6470a9f494b3cd6a001829c4
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:6b14e850cda9a54618847858292ea669abff599a40692e27fed0e337c4a60526
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:e71f2f2347dd46dec790405b1bb98d54a348a4583949f142f5703be9a790f96e
        version: 3.70.1
      entries:
      - name: rhacs-operator.v3.70.1
        version: 3.70.1
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.70
    - currentCSV: rhacs-operator.v3.71.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:469b6c1592226eb9a521131ef758a33e6a6ac7c96bc67fcc903cea30469c125e
          createdAt: "2022-12-07T08:31:36.777654+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.70.0 < 3.71.3'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4a0dc8c337299d7f6440ff75c2d5219d3ab68183f4f117f008ef065757dd5d4b
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:469b6c1592226eb9a521131ef758a33e6a6ac7c96bc67fcc903cea30469c125e
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:f16f92e074f7125b88e9e2033c7c12b44421145b93d040034fa4688626869af1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:c3ec88cd99908894d0e67153779062f90ac1b007980d66040a79b2be13798a10
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:4d8bac9b3857af6ba351c7768f70ef30ef7cb99ae4df60cb226b5a793353a052
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:7f279bb6e0f34a5208d8711ba4d7b62c78f08b683815ce33f6aa28c23c4246da
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:9bc25e81c21cda6d8b535d8ac67c1286fe57a1596001850e730f8f73bf371111
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:e282a8e5e606003ea344d44ac5ee7699563848f7160eeae183aa58ad740e0e94
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:cb2d0f470968c2b725b0d176d423cc1866fd56893344d077d28f2c16581d3614
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:df7b41999b542a477848823a1d0e7e5c29ae1b00fbac035007a5a1f33fac139f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:5fdfd3690af60bd5ecf04741cd1f73259491da501700b066120976960567f0db
        version: 3.71.3
      entries:
      - name: rhacs-operator.v3.71.3
        version: 3.71.3
      - name: rhacs-operator.v3.71.2
        version: 3.71.2
      - name: rhacs-operator.v3.71.1
        version: 3.71.1
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.71
    - currentCSV: rhacs-operator.v3.72.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:3ff61e1cc77b66a22c8a8c0ab306f0e122cd74e762005285aede23ebb6c2f6a2
          createdAt: "2023-02-28T17:02:08.443613+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.71.0 < 3.72.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:575b6c3e2914751df1668ffff83cbda187ce6e56f32214414cdf036b7a62bbd8
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d56a00b3e20e60b7169f5c35d03fa6a1bd3b9980dd3ebc9d89d759e1d6d20ad2
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:cd475a6c2356d63cfe736a11114de7fedfbd8fdcfecc56224a00b6c819103be6
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:42e5f6c4ba789f124ac3ebe7b6d1a607ede8e763a35888c97ccbd0259bec4a1c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:648b14e4158da2a85eedb46b3456fb502d680d52d375ad34bf45b4b94f8b2ef6
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:3ff61e1cc77b66a22c8a8c0ab306f0e122cd74e762005285aede23ebb6c2f6a2
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:cd0c205ee44eddcdf99b84e80fcf4403851d956efc50532b9d12013bafe44988
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:0b69595405e4e598183bbf4197ac5b7593724fa6162413f225aa01a542af18c9
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:622f30b7f591194d86a5bf95deede0225e66f7b624db197f6289a5fbbbe887a9
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:c07fae3616cbc681e4d31839abea6623b4bf4dade93273201c8db7ba0d635aad
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:cb5df858e0c5cae8f77b285e23daeb83062bd72570c90ed5ea3ce4a02d2f5f08
        version: 3.72.4
      entries:
      - name: rhacs-operator.v3.72.4
        version: 3.72.4
      - name: rhacs-operator.v3.72.3
        version: 3.72.3
      - name: rhacs-operator.v3.72.2
        version: 3.72.2
      - name: rhacs-operator.v3.72.1
        version: 3.72.1
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.72
    - currentCSV: rhacs-operator.v3.73.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a8486b66f977a639d77ef110492e5ecdde47be8acaefbaa1cbb5417085c56839
          createdAt: "2023-05-26T21:27:04.033276+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.72.0 < 3.73.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:7bf9415aa4b0746d11ca5a59d4179bc382a0fac4ded969f0c4294171fde27d58
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:e6687c12d070069f655c17832d94c619c18274a259ccc59715036f9908e162f7
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:ee2a9e4ab74f5a554d1e4ff148833a53bdaab70ac1e2668cc9ee2fceca290ba9
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:082bd6f4b40209e0591be001dea358f328cd86a08691cdceb7f71ea59d861d9d
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:4d0ead8969f120a629104a92a454842bb7103aec137deb6c8209b7db9c48a1b2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:648b14e4158da2a85eedb46b3456fb502d680d52d375ad34bf45b4b94f8b2ef6
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a8486b66f977a639d77ef110492e5ecdde47be8acaefbaa1cbb5417085c56839
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:e2c89f735969d438f75a89ff63c404a8ec4361a03fde48e880fee73006b9c669
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:e4a27d3ffb173c87728fb637ca7fde92a9e78ea9ee3edcecccc2a5e3d3748b8c
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:dce5db4f162dd40c09cad114cb802cd6d7c72d0c6986581f2e74c977493f75b0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:c020a80ff82bf4d9b5cec3eeac5de9990de76041e22c444e664a20e15c46ce3b
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:5df359156b01505ce43e5f14c3b08722856bdb86f3d0d9b4fba785a2e580f346
        version: 3.73.5
      entries:
      - name: rhacs-operator.v3.73.5
        version: 3.73.5
      - name: rhacs-operator.v3.73.4
        version: 3.73.4
      - name: rhacs-operator.v3.73.3
        version: 3.73.3
      - name: rhacs-operator.v3.73.2
        version: 3.73.2
      - name: rhacs-operator.v3.73.1
        version: 3.73.1
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.73
    - currentCSV: rhacs-operator.v3.74.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
          createdAt: "2024-02-06T23:06:13.271846+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.73.0 < 3.74.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d546f2d3f668ce2f20e594394ade0a86d808f6c8c88cadbba227083d4abac940
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:264102d4bca7d9bd384dc04e5bbff1a7b43343ea8e59cc79047cdda50a3f299f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:0482d4464d6d39c01abab8df882604dd227c52b5420d36fe7414aaff8067cbec
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:2c0840f4573373a19eb8c9a26501138ef0e76efbd508fd69a6260e78194d86c8
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:74624389ac3261ed699b1f64e5717353da68895709d30a011bc8813b61a4a249
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:87eb62fe8aa7952d0d69dc13c4867ac2f51e821f56d149a9b588c826caec4c16
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:120b9891ee9db839ab0b2029a62e67ec4747c47c6ffe15af38e9035160f57b78
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:22bf00e1457b4a9e63901892aa4375a1d4a9d1b2c067c1f054d918a24b222b10
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:a108e9a7967001e0bdfed7124b01e9a3dbf1dec308444762973030aec57e1d5a
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:71b9fdcaf4cc761b1499be8e6ccff6c463cc6b9856e9deb1f93baad45ea59eb0
        version: 3.74.9
      entries:
      - name: rhacs-operator.v3.74.9
        version: 3.74.9
      - name: rhacs-operator.v3.74.8
        version: 3.74.8
      - name: rhacs-operator.v3.74.7
        version: 3.74.7
      - name: rhacs-operator.v3.74.6
        version: 3.74.6
      - name: rhacs-operator.v3.74.5
        version: 3.74.5
      - name: rhacs-operator.v3.74.4
        version: 3.74.4
      - name: rhacs-operator.v3.74.3
        version: 3.74.3
      - name: rhacs-operator.v3.74.2
        version: 3.74.2
      - name: rhacs-operator.v3.74.1
        version: 3.74.1
      - name: rhacs-operator.v3.74.0
        version: 3.74.0
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.74
    - currentCSV: rhacs-operator.v4.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9bf3a3e0fd573f91f03794eaf5291364a90392e5e7d4b5d4ed550ef4600ab826
          createdAt: "2023-10-20T07:04:50.531740+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.74.0 < 4.0.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:08f537cba262605a0eb0a510b6faf78e3ba26465f636d03cc1f7a6ed6b40d355
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:eee1aa054d4d15b23be293f31d3ea659e0ffc1d05174ef89a6c977dd3d4377cc
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:4c848eb7465e928251e05d32ffbc93cd14e7bb77972d22b8721bd6eaa4a93d2f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:3250d89e78de852a2627593f29469f0b3ff639ea3644b4b7d6d057d0c1ad3659
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:5a492b856e0b6d9fdb020b556d13186ef5e7095b0c736a2031500c9c00dad0c1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:139d80d398e7e5cfc85839e3039a3630b48c7ae6650d755754430fccd94d6a27
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0aedd8e865f72ffa8b79fd2bab48974caec3eea3e3a6f77c89fe6dacb8947057
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9bf3a3e0fd573f91f03794eaf5291364a90392e5e7d4b5d4ed550ef4600ab826
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:cfa10bb6b461f302674f185b26c8d6efaaeba7f3f65d2cf05ae1f069aa2710d3
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:c39621837cbc8419d187a43a9575ec8cb4021c58ea1b56f215003d7264eb5d03
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:09e79b5db8977520f9951cc43c339c8e40b4659d7da513347d5a88b2500f809d
        version: 4.0.5
      entries:
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.0
    - currentCSV: rhacs-operator.v4.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:8ffb6f0ff1bf016c3919c1eefc6c3db3d4a8fbe0e79919f72b4c281afe6d4b1c
          createdAt: "2024-01-13T18:55:22.011709+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 4.0.0 < 4.1.6'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:364c107c34694ca33ac724de78db3e41ee55f9098d9ade0775a51b11f1533065
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:8ffb6f0ff1bf016c3919c1eefc6c3db3d4a8fbe0e79919f72b4c281afe6d4b1c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a8ad0e76ce4cae9f9b36aecf125c4ed56fc2d776fe3df113d41bfc1c2f8b3602
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:8547b76cf0533dfe6da2aafbbd063021bad4ff78b32b2dfed4a7607496e0a904
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:8760d6dc68c86f0c5b53d8f0bfada0df012f80bc8e0023ba3522fe6829a05a41
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:762ad216e79be0837e98873202b0072c7fd399925ee94bde57cc61c68cbd5f4e
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:01b9f9798d3ecbefe3da1ac2ac782e8e8b930e7eeceb82fe3473b5bfc1ccd40a
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:7ac504b7824d17bd7ac55b1e59c26be308506f315a0ab0d8c45d3b4934c0a16c
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:2ec62f3da93a1346ad215744257820377ccf4ead84e7471c52e348a31a431d3c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:bdbe08ef2d01b114702780fab76905e6eb1851684692ffa0a4c84604bda3ddb3
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:34694b9f4fe7fffc0c547be4e311579b5bb0c1fa92d8257bf6d1581ffd655c46
        version: 4.1.6
      entries:
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.1
    - currentCSV: rhacs-operator.v4.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4253eca84cbb0c5fc033507ddc2deef235b054c0ba944fa271ebbba21545b872
          createdAt: "2024-03-13T14:51:55.084122+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.1.0 < 4.2.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:b5a15f2985b114bf9bc24a0d9942db5d0cb818e60fd15a2f9094aa00904cce77
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4253eca84cbb0c5fc033507ddc2deef235b054c0ba944fa271ebbba21545b872
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:12db245af3b8475c9b5680bb0035c566109819e05cfe62bd1dfdfb5d6ed25bea
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:b82f7f654235688fc07fb037422a73e1106e27024d39a68ef7fac1422b69496d
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:ff27f03064633d6d08f8383649069e39c868def4aeca8f2622750397e049d0ec
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:2f326cabf03f780a3e6120995d5442641764641f62430d82e05edbe21b30c7b2
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:3858ab936d4626a11bbf895d41bad181364ee5df3312f6dd29ebeb9b7d31ec77
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:e2b07cb7d690477d78a2b8e1875545998d4bdb946f80dd0533a14187da99dbbf
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:59f1862c8ce188906499f08201ae05deeb47bd4fc5eab4be0386ca44ec065da8
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:e79a5e5681ebfe2d6349e2763ee0c1fa521b20319846958e402d23229b1f257b
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:4b35f5f39865238a8e95d3cba7ebb5f15ccfbf6a2563c34f57d91a0b5c927066
        version: 4.2.5
      entries:
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.2
    - currentCSV: rhacs-operator.v4.3.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:266cbbc270efafa26d5d15f3e333c96eaedee3b47854ff337ae1c572d34901ce
          createdAt: "2024-06-06T22:31:19.365878+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.2.0 < 4.3.8'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c36a557c9dc2ae4479c432933a9a4dbef38f2be53360a68a5b6befb614d01dac
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:b0c9b3858fc0547dcb4eb3065dedd5bbcf1e1f6707a8ab812ed06882d4f6d6ad
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:b7eb17fbcf6c4a6e4205ad66f1ffd9b10a4634f6ef5045d407c266566b1f255d
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:bdf9785bff51194636c6e8a518448c6d12a297bd31c4bb2e79d07084409c18b4
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:266cbbc270efafa26d5d15f3e333c96eaedee3b47854ff337ae1c572d34901ce
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:9fba889021a876ab627ff8c496e4b6c46c221c61142cd0c76daf33c89762e2c1
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:10d0cc82b6eddd089f3a7af37b872c33248513d0b4d3f094b6f62c01f4460825
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:6b55a25eb3a611c648d9be883828834592229c16f4166fd28ef40a25fc667061
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:2c35cb6f518393d002edd2bf5f714c1d053cd0fab55f6152a2f35676e4979202
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:2bb643651c8d51525223df3b2b675947174bbe9f1eaea8688e1fa83044cfa41f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:5092c3fbb4a38c06a2f42d125dc718ef0e1005f304ad8046afdd2a5e7ccbc580
        version: 4.3.8
      entries:
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.3
    - currentCSV: rhacs-operator.v4.4.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:2b87569d7f4ef5e06c989b273c378e007c6d87b26eec41774ac1fc67dcd9e4e4
          createdAt: "2025-02-10T19:50:13.597164+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.3.0 < 4.4.8'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:65707fdad55767ff5598a8d3679fd0779a37f59501e58971b81bebcb23f1f9fc
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:d6b4f8caf342b105c8df14c5c9180cb36123e08d34756ee8fe4f2f76c34f8064
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:664a5dc40dc1fa06c47a11e13b0ab5e0612a981d7b69bcfb29422abf3d1a13b2
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:39cfd01236e905034f6645a73063427436c14e351e0e0bc735e73e62aa5a1c3c
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:467d069ee957aea07882b83b60ae9fcc62ccca159f5d643aac13e531ccf7df88
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:2b87569d7f4ef5e06c989b273c378e007c6d87b26eec41774ac1fc67dcd9e4e4
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:0df4e95c7406aee7812d145cdd4d8c61cec32857fd2079dbc4244954a9d8288a
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:183c75920ab8e7cb947cb966f1543e39b6fdb5e1bb0caa1b0cf5f89671f0328f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:76f9dc2e2ee5595c36530d97dec216e2dbefada2fe0747e3f6a3a725b75722ab
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:763e7568704971cb4656727dc223374de13e82b4304dd3f5387ddc0f4e42b8b6
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:69c12230a785619c765e3749a7d56f76dce9ce4f530c20e8a9387b99eb65627b
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:ba6e25a413c79131b966a097f72c36478933ea52deb55a62acda7011a96b3983
        version: 4.4.8
      entries:
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.4
    - currentCSV: rhacs-operator.v4.5.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a7dfbe3590668f91fe6694335b5a35b2fca07b4735348cc5202921755d2f654d
          createdAt: "2025-04-09T22:51:15.157672+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.4.0 < 4.5.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:5914548f6ce2160e92c5d96580ffba23cdf1ae940c1f73f3f6424d15c6e97c37
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:dc7f23341c8b209f1846f29f889d00ef232fb58da9d5b4bd561f3933093b2786
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:d928600f75461465598a10de8df2495a275fd5232a92b078f609d1f5774eb778
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:1817cf8d2c08b04391792321702ef53d16a4b1ca06459c00c9c053128f534358
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:c2656b794ec4236e36f48f71bf26bdedd9de27fc4acef39deba2bdcfcddab03b
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:cc4b329cc795752ecbb178d2e4efc152bde54b04181334eb5bd70304c3354cad
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:0ff948b5517df3773cae2df02b708c348d8707409899b1de4108588e31601c0c
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a7dfbe3590668f91fe6694335b5a35b2fca07b4735348cc5202921755d2f654d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:5109b304eb9458f97af85c2ba63e08f51672fa3f9406226b7f357b86b5828dc0
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:fde78d34ecd2944c7c16b4e823c76b3667bf6c5b468e9991128ecc4272ec0e2e
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:ed33e7396357f1db03a065573c020846a5872b12a21284439770686689222434
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:f07cba0738897e0e7ae74971be0470ce2a855657f133a6e1cdd8a7092f6151b0
        version: 4.5.9
      entries:
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.5
    - currentCSV: rhacs-operator.v4.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:c4d434e1ff052508e5594d2b95dafa8726fd8adf5317754774c013a46de224d8
          createdAt: "2025-09-26T09:24:28.046251+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.5.0 < 4.6.10'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:b5cd08349931669055f03d21331499ef2ea04597f8a90fb201615d43000271bc
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:ffe41363dc00b05b0daa0c6d16efb30e534086aa4e3d7ccc599c86d33813138c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:713e61d3c8b55de1edc9d7834d3e8576718e1327665f94a6b58178fa7496c0cc
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:65793c05aec8085a53fab8bcf33a85192c17b46c91a7d86e0a457d1e3a113878
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:899da9c23000bb821bc9d544ebb56bcda4f3245ceee5b1b9c12d09f580b4b9a1
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:8fad8ab2efdcbc99d9efa736a90cca6f236f0a9d45fdb355ae0b6b4dc42d77f0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:90c47c6f95571cfcca04aa4e05a18befad75c1b44a6f3a055fcdfdab749d29c5
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:16b61ca1b9ee4e0821e7fc82fe4def75ed22950e502730f5704ec4fe63bb2e59
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:c4d434e1ff052508e5594d2b95dafa8726fd8adf5317754774c013a46de224d8
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:b630d53ad0b862151be7e14f0483ca5adac570216deae66f60e75af87a4dfd46
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:19db8c5d2425dcf18ae63b025f946fc46adaec0578e9505415cb585797069e95
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:6da59c5b82625894f6efa3fa0bc5ea0378edd4f95cf42dee772e9c8a565dba81
        version: 4.6.10
      entries:
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.6
    - currentCSV: rhacs-operator.v4.7.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:5b60d8271d6fa70df3c8f7e0ecd0d64feecfdf674dc84036500fe5f145c9415a
          createdAt: "2025-09-26T11:05:05.900682+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.6.0 < 4.7.7'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:def71e1ddea5b8c1e54c8582f39dc9ecca855b38aa9423690150aa6ef0ff189f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8882c00c9be9b579654a9ee8cd0c38537cf92d2f4bd8ebded1678aa0b6ae827
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:34ea97c6e9fc2d23427d0c6f2628075c0eaa0e151ca1c381539b039c249217f7
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:98553e6f18d3c03a29212b4f7ea1d30a42bcf8b52ca12aecbcbc47769fc35b53
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:bac79889d4717233e89bb6c8fae6732ed7bec5b9a55fd3b3de1f1311d870d2b1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:6bb454d64a9e339b1aeb81ca17b26f33db4a84f19ced4ad81858e40fd01878dc
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:5b60d8271d6fa70df3c8f7e0ecd0d64feecfdf674dc84036500fe5f145c9415a
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:7596529cbdce7118226b79d69e0667ab35ee692569e4014b74abdf0fa7415222
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:5ed9902b52830be93436d7977c10e73d346b6290532f6d7245cfe7cca9c93b31
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:0c32318ee4cbd547e2c1fd64433cd23a629d1b06523d7c0f07ddb98c2baaa708
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:d8fcfe1267495801d23a0b94101c7b4d06b48346a949c97946ebcbffe148f4b6
        version: 4.7.7
      entries:
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.7
    - currentCSV: rhacs-operator.v4.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
          createdAt: "2025-09-12T11:06:11.361461+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.7.0 < 4.8.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:adf47cb6615d344533fcca0aee41774f5e8df9b952b9b89a57c2be2b7e0fc03c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:92ff7691963e620dc5bf6da0a7694e6bcd68c2545afc1642a59a443ba2ab5611
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:db691065e46f604b980b0b202f6c2e14c8b204094d784c4a15db1e4fecb2965e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:97df59280c12727a99fdec0c9bf56eb874e6a1cacd12d849709797d4be635ce9
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:e5c200f28187c1f6971562e41cae969183ee15f3d0feb076bb202b22ff67cbcb
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8964b1441251607b23f92918e9158724174013dc4af423988f56d6098dfb2d5
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:e949830f3034c41aa5434a3654b2b1ccba7dac0373bf43d7c50a9ea43bb686fb
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:16912df2c26e5fc2e6051c0501170722856db274415063608de716cf53851038
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:d1758b76c32f03b5a1a44951d18fb2bb0c4fd0a99d24f8c42fe9e6b059574889
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0c92d5b5508032d77559606927e45686179eaafc59e16d5496e088f2f212d258
        version: 4.8.4
      entries:
      - name: rhacs-operator.v4.8.4
        version: 4.8.4
      - name: rhacs-operator.v4.8.3
        version: 4.8.3
      - name: rhacs-operator.v4.8.2
        version: 4.8.2
      - name: rhacs-operator.v4.8.1
        version: 4.8.1
      - name: rhacs-operator.v4.8.0
        version: 4.8.0
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.8
    - currentCSV: rhacs-operator.v4.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
          createdAt: "2025-09-12T11:06:11.361461+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.7.0 < 4.8.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:e5c200f28187c1f6971562e41cae969183ee15f3d0feb076bb202b22ff67cbcb
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:e949830f3034c41aa5434a3654b2b1ccba7dac0373bf43d7c50a9ea43bb686fb
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:16912df2c26e5fc2e6051c0501170722856db274415063608de716cf53851038
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:adf47cb6615d344533fcca0aee41774f5e8df9b952b9b89a57c2be2b7e0fc03c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:97df59280c12727a99fdec0c9bf56eb874e6a1cacd12d849709797d4be635ce9
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:d1758b76c32f03b5a1a44951d18fb2bb0c4fd0a99d24f8c42fe9e6b059574889
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8964b1441251607b23f92918e9158724174013dc4af423988f56d6098dfb2d5
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0c92d5b5508032d77559606927e45686179eaafc59e16d5496e088f2f212d258
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:db691065e46f604b980b0b202f6c2e14c8b204094d784c4a15db1e4fecb2965e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:92ff7691963e620dc5bf6da0a7694e6bcd68c2545afc1642a59a443ba2ab5611
        version: 4.8.4
      entries:
      - name: rhacs-operator.v4.8.4
        version: 4.8.4
      - name: rhacs-operator.v4.8.3
        version: 4.8.3
      - name: rhacs-operator.v4.8.2
        version: 4.8.2
      - name: rhacs-operator.v4.8.1
        version: 4.8.1
      - name: rhacs-operator.v4.8.0
        version: 4.8.0
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: stable
    defaultChannel: stable
    packageName: rhacs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dns-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dns-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSRecord",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dns-operator",
                    "app.kubernetes.io/instance": "dnsrecord-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dnsrecord",
                    "app.kubernetes.io/part-of": "dns-operator"
                  },
                  "name": "dnsrecord-sample"
                },
                "spec": {
                  "endpoints": [
                    {
                      "dnsName": "dnsrecord-simple.example.com",
                      "recordTTL": 60,
                      "recordType": "A",
                      "targets": [
                        "52.215.108.61",
                        "52.30.101.221"
                      ]
                    }
                  ],
                  "providerRef": {
                    "name": "dns-provider-creds"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/dns-operator:v0.6.0
          createdAt: "2024-09-13T09:55:39Z"
          description: A Kubernetes Operator to manage the lifecycle of DNS resources
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kuadrant/dns-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSRecord is the Schema for the dnsrecords API
            displayName: DNSRecord
            kind: DNSRecord
            name: dnsrecords.kuadrant.io
            version: v1alpha1
        description: A Kubernetes Operator to manage the lifecycle of DNS resources
        displayName: DNS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dns
        - kuadrant
        links:
        - name: DNS Operator
          url: https://github.com/Kuadrant/dns-operator
        maintainers:
        - email: mnairn@redhat.com
          name: Michael Nairn
        - email: pbrookes@redhat.com
          name: Phil Brookes
        - email: cbrookes@redhat.com
          name: Craig Brookes
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/dns-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: dns-operator.v0.6.0
        version: 0.6.0
      name: stable
    defaultChannel: stable
    packageName: dns-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://www.redhat.com
    name: sap-commerce-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sap-commerce-operator.v0.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hybris.hybris.org/v1alpha1",
                "kind": "HybrisApp",
                "metadata": {
                  "name": "hybrisapp-sample"
                },
                "spec": {
                  "aJPServicePort": 30000,
                  "apachejvmRouteName": "",
                  "baseImageName": "hybris-base",
                  "baseImageTag": "v0.2",
                  "hybrisANTTaskNames": "",
                  "podHealthyProbePath": "/platform/init",
                  "sourceRepoLocalPropertiesOverride": "",
                  "sourceRepoRef": "main",
                  "sourceRepoURL": "https://github.com/redhat-sap/sap-commerce-app-example.git",
                  "startupProbeFailureThreshold": 100,
                  "startupProbePeriodSecond": 60
                }
              },
              {
                "apiVersion": "hybris.hybris.org/v1alpha1",
                "kind": "HybrisBase",
                "metadata": {
                  "name": "hybrisbase-sample"
                },
                "spec": {
                  "URL": "http://nexus.redhat.com/nexus/service/local/repositories/releases/content/org/hybris/hybris/1.0/hybris-1.0.tar.gz",
                  "buildSourceRepo": "https://github.com/redhat-sap/sap-commerce-operator",
                  "buildSourceRepoBranch": "main",
                  "imageName": "hybris-base",
                  "imageTag": "v0.2",
                  "jdkURL": "https://github.com/SAP/SapMachine/releases/download/sapmachine-11.0.5/sapmachine-jdk-11.0.5-1.x86_64.rpm",
                  "password": "admin123",
                  "username": "admin"
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          containerImage: quay.io/redhat-sap-cop/hybris-operator:0.0.4
          description: Builds and runs SAP Commerce in OpenShift
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-sap/sap-commerce-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HybrisApp is the Schema for the hybrisapps API
            displayName: Hybris Application
            kind: HybrisApp
            name: hybrisapps.hybris.hybris.org
            version: v1alpha1
          - description: HybrisBase is the Schema for the Hybris Base API
            displayName: Hybris Base
            kind: HybrisBase
            name: hybrisbases.hybris.hybris.org
            version: v1alpha1
        description: |-
          Builds and runs SAP Commerce in OpenShift

          * The operator provides APIs to initialize SAP Commerce with the provided SAP Commerce Server. Though there are no restrictions from the operator on the base package to be used from SAP, the default SAP base image provides a server preconfigured with an application server based on Apache Tomcat. Both HSQLDB and Tomcat are included in the SAP Commerce package, and are designed to be used for testing, development, and demonstration purposes. The custom classes created for various business needs along with an external RDBMS configuration can be used in a HybrisApp creation to execute SAP Commerce in a production setup.

          * The operator also enables the use of SAP Commerce in a clustered model. A SAP Commerce Cluster is a number of individual SAP Commerce installations using a common set of data on one database. A cluster configuration lets you balance the load between the nodes to scale SAP Commerce in a way that is transparent to the user.

          * The clustering offers few communication protocols as options (UDP: both multicast or unicast, or JGroups). In order to support multicast for SAP Commerce clustering the corresponding OpenShift namespace has to be enabled for multicast usage.

          For more details on SAP Commerce consult the product documentation: [https://help.sap.com/viewer/product/SAP_COMMERCE](https://help.sap.com/viewer/product/SAP_COMMERCE)
        displayName: SAP Commerce Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sap-commerce-operator
        links:
        - name: SAP Commerce Operator
          url: https://github.com/redhat-sap/sap-commerce-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
          url: https://www.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/redhat-sap-cop/hybris-operator:0.0.4
        version: 0.0.4
      entries:
      - name: sap-commerce-operator.v0.0.4
        version: 0.0.4
      - name: sap-commerce-operator.v0.0.3
        version: 0.0.3
      - name: sap-commerce-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: sap-commerce-operator
    provider:
      name: Red Hat, Inc.
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cognitoidentityprovider-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cognitoidentityprovider-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cognitoidentityprovider-controller:1.1.1
          createdAt: "2025-09-19T20:58:49Z"
          description: AWS Cognito Identity Provider controller is a service controller
            for managing Cognito Identity Provider resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: UserPool represents the state of an AWS cognitoidentityprovider
              UserPool resource.
            displayName: UserPool
            kind: UserPool
            name: userpools.cognitoidentityprovider.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Cognito Identity Provider resources in AWS from within your Kubernetes cluster.

          **About Amazon Cognito Identity Provider**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Cognito Identity Provider
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cognitoidentityprovider
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Cognito Identity Provider Developer Resources
          url: https://aws.amazon.com/Cognito Identity Provider/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cognitoidentityprovider maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cognitoidentityprovider-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-cognitoidentityprovider-controller.v1.1.1
        version: 1.1.1
      - name: ack-cognitoidentityprovider-controller.v1.1.0
        version: 1.1.0
      - name: ack-cognitoidentityprovider-controller.v1.0.12
        version: 1.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-cognitoidentityprovider-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sumo Logic
      provider-url: https://www.sumologic.com/
    name: sumologic-kubernetes-collection-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "helm-operator.sumologic.com/v1alpha1",
                    "kind": "SumologicCollection",
                    "metadata": {
                        "name": "collection"
                    },
                    "spec": {
                        "sumologic": {
                            "accessId": "",
                            "accessKey": "",
                            "clusterName": "kubernetes",
                            "scc": {
                                "create": true
                            }
                        },
                        "kube-prometheus-stack": {
                            "prometheus-node-exporter": {
                                "service": {
                                    "port": 9200,
                                    "targetPort": 9200
                                }
                            }
                        },
                        "otellogs": {
                            "daemonset": {
                                "containers": {
                                    "otelcol": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                },
                                "initContainers": {
                                    "changeowner": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                }
                            }
                        }
                    }
                }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing,Monitoring
          containerImage: registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
          createdAt: "2024-07-19T10:48:00Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator.git
          support: Sumo Logic
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configuration for Sumo Logic Kubernetes Collection
            displayName: SumologicCollection
            kind: SumologicCollection
            name: sumologiccollections.helm-operator.sumologic.com
            version: v1alpha1
        description: Helm Operator for the Sumo Logic Kubernetes Collection Helm Chart
          v4.13.0-2
        displayName: Sumo Logic Kubernetes Collection Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - collection
        - logging
        - metrics
        - monitoring
        links:
        - name: Sumo Logic Kubernetes Collection Helm Operator
          url: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator
        maintainers:
        - email: collection@sumologic.com
          name: Sumo Logic
        maturity: alpha
        provider:
          name: Sumo Logic
          url: https://www.sumologic.com/
        relatedImages:
        - registry.connect.redhat.com/sumologic/autoinstrumentation-java@sha256:242958b7d8905f4ac02f6ad38897ba58b9badf2a6213a389e67dcc71fee3630c
        - registry.connect.redhat.com/sumologic/thanos@sha256:16faf7191668e86d8b576bba055dd0cbdeb009635976b9e07bd00f926032f11c
        - registry.connect.redhat.com/sumologic/tailing-sidecar@sha256:738d653a78dc415243ab3627db9135768175fb7f51d0803566f5d9e824c8744b
        - registry.connect.redhat.com/sumologic/opentelemetry-operator@sha256:5587fe5b7ebfa57ad320ad2521103f908317388d1741c40e5f7be5ed9e5f9ad7
        - registry.connect.redhat.com/sumologic/kube-state-metrics@sha256:3551ededa327a35db798ec8286cd843580e3ae395d2231e041e8c4b7621127ee
        - registry.connect.redhat.com/sumologic/tailing-sidecar-operator@sha256:c16b8b18e9b263fc3eb24059bba512bcfc3201b32105af6e5518392826ddd66c
        - registry.connect.redhat.com/sumologic/telegraf-operator@sha256:ff958222be01c0c3966474463e43bbd5bcf348790dd34bb761718cb1cea2d793
        - registry.connect.redhat.com/sumologic/kubernetes-tools-kubectl@sha256:d3155b864cc74f327adedf7d013a8f0fcb11c2e73a5655b7cafa848824e80172
        - registry.connect.redhat.com/sumologic/kubernetes-setup@sha256:f07eaf68fa11ffc708e4485a83bd3763fc5712f7d73d29edfe5b9c9e08f05df6
        - registry.connect.redhat.com/sumologic/sumologic-otel-collector@sha256:4554d43dba204e3ec8e91b36442a438f10234fa0cb4db7fb7880b257a6485112
        - registry.connect.redhat.com/sumologic/autoinstrumentation-nodejs@sha256:c206eef4fdecaa58b4f01e5e7f0dbd74e9606eab2b4debf47461d36191e248d1
        - registry.connect.redhat.com/sumologic/autoinstrumentation-dotnet@sha256:f4f385c4489c74b5c43f2fe6b6620a9eb8d9ecab470567236b3c45f636106640
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
        - registry.connect.redhat.com/sumologic/metrics-server@sha256:65d2a3ddc4b4183f438fa18b655079c93c905cb9e3b8c54977336205dc1fff07
        - registry.connect.redhat.com/sumologic/nginx-unprivileged@sha256:7260555d16c94c4fe1e016626b4cb8a5c0de6b700e2eea28744331fb298f71ae
        - registry.connect.redhat.com/sumologic/node-exporter@sha256:11e9e604432a66d9b388cb1bf70eac077c7f0372d1264342ef927d76b8ef0272
        - registry.connect.redhat.com/sumologic/prometheus-operator@sha256:58a329e6582456b32abebb470086191f8366b927af4703741613993465f116b8
        - registry.connect.redhat.com/sumologic/prometheus@sha256:0ef6eaf570a40499dd5de92a79593e950bc771b014e0ee6ff7419b4b82a36fcc
        - registry.connect.redhat.com/sumologic/autoinstrumentation-python@sha256:f0fa45fa45e64c7df88f21181ae18ee96fe83cd3988258663d0806ea6f39e6bb
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:e1f49fcf5b7e86d9eb49c16dbbf9484cb486f55f999fd51d7c738c3b1f2666cf
        - registry.connect.redhat.com/sumologic/telegraf@sha256:75efd2e509c6017ea0fc5c9103ce068cbc89087b16506a4ef48cb5ff00eab84e
        - registry.connect.redhat.com/sumologic/busybox@sha256:209bc8e0b1d958fc699f3bb75b1248544e372ad35cdc46c991bd1698f641e1f9
        - registry.connect.redhat.com/sumologic/prometheus-config-reloader@sha256:0b605d36782afd8f50fd631d0e9c7b19d39ce962af89b7a737b2f09a8ceff291
        version: 4.13.0-2
      entries:
      - name: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
        version: 4.13.0-2
      name: alpha
    defaultChannel: alpha
    packageName: sumologic-kubernetes-collection-helm-operator
    provider:
      name: Sumo Logic
      url: https://www.sumologic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: rh-service-binding-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-binding-operator.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ServiceBinding expresses intent to bind an operator-backed
              service with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding for Kubernetes. Describes the connection
              between a Provisioned Service and an Application Projection. Implements
              Service Binding Specifcation for Kubernetes (https://github.com/k8s-service-bindings/spec).
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: alpha
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:fe6629401bc8447da95ec55949b5a27e620364bb6e02f86230262699ac8a5d55
        version: 0.9.1
      entries:
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: beta
    - currentCSV: service-binding-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindablekinds-sample"
                }
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:e799bd06043028d214baeeffc683710ad2eed7b83272fc653e3b89f69711319b
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:e799bd06043028d214baeeffc683710ad2eed7b83272fc653e3b89f69711319b
        version: 0.11.0
      entries:
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: candidate
    - currentCSV: service-binding-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:3d7a3648e86848402577dd19c1f8fd130a5556ebdfbc0f663a72f9e141024812
        version: 0.10.0
      entries:
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: preview
    - currentCSV: service-binding-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindable-kinds"
                },
                "spec": {}
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  },
                  "type": "postgres",
                  "workload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: registry.redhat.io/ocp-tools-4/service-binding-rhel8-operator@sha256:9c8fb0c8f9d2f54685561af21e595aa88bf309eaebebdf23fb6997b75d026b18
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'.
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1alpha3
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1beta1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
          - description: Service Binding implementing community specification (https://github.com/k8s-service-bindings/spec)
              describing the connection between a backing service and an application
              workload.
            displayName: Service Binding (Spec API)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1beta1
        description: |
          The Service Binding Operator manages the data plane for applications and backing services. It enables developers to connect their
          applications to backing services (REST API, databases, event buses and others) with a consistent and predictable
          experience. The operator reads data made available by the control plane of backing services and projects the bindings
          data to applications either as accessible files mounted into the application’s container or environment variables.

          In Kubernetes, each service suggests a different way to access their secrets and developers will consume them in their
          application in a custom way. While this provides a good deal of flexibility, it also creates discrepancies and
          complexity dealing with each specific services, which leads large development teams to lose overall velocity.

          Service Binding Operator removes the error-prone manual configuration of binding information, provides service operators
          a low-touch administrative experience to provision and manage access to services and enriches development lifecycle.

          For additional details refer to the project [documentation](https://redhat-developer.github.io/service-binding-operator).

          Get Started with Service Binding Operator
          * Find out about SBO’s [prerequisites](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/installing-service-binding.html).
          * Refer to the [Quick Start Guide](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/quick-start.html) to see SBO in action in a simple scenario.

          Need Help
          * Raise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)

          Licence
          * Service Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE)

          Service Binding Operator is providing and supporting two different APIs.
          * `binding.operators.coreos.com/v1alpha1`: This API is compliant with the Service Binding specification. It is fully supported and considered as the mature API. It’s the API we recommend in production environments.
          * `servicebinding.io/v1beta1`: This API is the one defined in the Service Binding specification. It is available in Service Binding Operator 1.x.

          We’ll consider supporting only the API from the specification, once the Service Binding specification will be officially GA. Of course, that will be a breaking change and handle in a safe way for you.
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Documentation
          url: https://redhat-developer.github.io/service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4/service-binding-rhel8-operator@sha256:9c8fb0c8f9d2f54685561af21e595aa88bf309eaebebdf23fb6997b75d026b18
        version: 1.3.3
      entries:
      - name: service-binding-operator.v1.3.3
        version: 1.3.3
      - name: service-binding-operator.v1.3.1
        version: 1.3.1
      - name: service-binding-operator.v1.3.0
        version: 1.3.0
      - name: service-binding-operator.v1.2.0
        version: 1.2.0
      - name: service-binding-operator.v1.1.1
        version: 1.1.1
      - name: service-binding-operator.v1.1.0
        version: 1.1.0
      - name: service-binding-operator.v1.0.1
        version: 1.0.1
      - name: service-binding-operator.v1.0.0
        version: 1.0.0
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: rh-service-binding-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cluster-logging
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-logging.v5.9.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "name": "remote-elasticsearch",
                      "secret": {
                        "name": "elasticsearch"
                      },
                      "type": "elasticsearch",
                      "url": "remote.example.org:9200"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "audit",
                        "infrastructure"
                      ],
                      "name": "enable-default-log-store",
                      "outputRefs": [
                        "default"
                      ]
                    },
                    {
                      "inputRefs": [
                        "application"
                      ],
                      "name": "forward-to-remote",
                      "outputRefs": [
                        "remote-elasticsearch"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogging",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "collection": {
                    "type": "vector"
                  },
                  "logStore": {
                    "elasticsearch": {
                      "nodeCount": 3,
                      "redundancyPolicy": "SingleRedundancy",
                      "resources": {
                        "requests": {
                          "memory": "2Gi"
                        }
                      },
                      "storage": {
                        "size": "200G"
                      }
                    },
                    "retentionPolicy": {
                      "application": {
                        "maxAge": "7d"
                      }
                    },
                    "type": "elasticsearch"
                  },
                  "managementState": "Managed",
                  "visualization": {
                    "kibana": {
                      "replicas": 1
                    },
                    "type": "kibana"
                  }
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          console.openshift.io/plugins: '["logging-view-plugin"]'
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:54054faf0d2f375dd6d7fc7929a2dbb529508e1b3c004a56fc97c2a682d16be2
          createdAt: "2025-03-14T20:19:21Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing your aggregated logging stack.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.7.0-0 <5.9.15'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Logging (team-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.


              There are built-in input names for common log categories, and you can
              define custom inputs to do additional filtering.


              There is a built-in output name for the default openshift log store, but
              you can define your own outputs with a URL and other connection information
              to forward logs to other stores or processors, inside or outside the cluster.


              For more details see the documentation on the API fields.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.logging.openshift.io
            version: v1
          - description: A Red Hat OpenShift Logging instance. ClusterLogging is the
              Schema for the clusterloggings API
            displayName: Cluster Logging
            kind: ClusterLogging
            name: clusterloggings.logging.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates and manages the aggregated logging stack as a cluster-wide service.

          ##Features
          * **Create/Destroy**: Launch and create an aggregated logging stack to support the entire OCP cluster.
          * **Simplified Configuration**: Configure your aggregated logging cluster's structure like components and end points easily.
          * Configure log forwarding to an in-cluster, Red Hat managed log storage solution based on LokiStack or Elasticsearch

          ## Prerequisites
          ### Red Hat OpenShift Logging Namespace
          Cluster logging and the Red Hat OpenShift Logging Operator is only deployable to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - fluentd
        - logging
        - aggregated
        - efk
        - vector
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Fluentd
          url: https://www.fluentd.org/
        - name: Vector
          url: https://vector.dev/
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:54054faf0d2f375dd6d7fc7929a2dbb529508e1b3c004a56fc97c2a682d16be2
        - registry.redhat.io/openshift-logging/logging-view-plugin-rhel9@sha256:8023b072b45e9a21ddef83a56f4760d476053250d5a920d8eb5f03eecab8b1b3
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:54e3a5efaae5ecc2a08db03a421aa79e323b48c764115ea84d83a37c21d1d997
        - registry.redhat.io/openshift-logging/fluentd-rhel9@sha256:a676c9164ec2523b2002dc5eba234ce774926dbf7fe87de6d72b53f99cbf1db5
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:865572275b79462f14e84c69438a61afe9465cfe45ec164caf30965730b276f7
        version: 5.9.15
      entries:
      - name: cluster-logging.v5.9.15
        version: 5.9.15
      - name: cluster-logging.v5.9.14
        version: 5.9.14
      - name: cluster-logging.v5.9.13
        version: 5.9.13
      - name: cluster-logging.v5.9.12
        version: 5.9.12
      - name: cluster-logging.v5.9.11
        version: 5.9.11
      - name: cluster-logging.v5.9.10
        version: 5.9.10
      - name: cluster-logging.v5.9.9
        version: 5.9.9
      - name: cluster-logging.v5.9.8
        version: 5.9.8
      - name: cluster-logging.v5.9.7
        version: 5.9.7
      - name: cluster-logging.v5.9.6
        version: 5.9.6
      - name: cluster-logging.v5.9.5
        version: 5.9.5
      - name: cluster-logging.v5.9.4
        version: 5.9.4
      - name: cluster-logging.v5.9.3
        version: 5.9.3
      - name: cluster-logging.v5.9.2
        version: 5.9.2
      - name: cluster-logging.v5.9.1
        version: 5.9.1
      - name: cluster-logging.v5.9.0
        version: 5.9.0
      name: stable
    - currentCSV: cluster-logging.v5.8.22
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "name": "remote-elasticsearch",
                      "secret": {
                        "name": "elasticsearch"
                      },
                      "type": "elasticsearch",
                      "url": "remote.example.org:9200"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "audit",
                        "infrastructure"
                      ],
                      "name": "enable-default-log-store",
                      "outputRefs": [
                        "default"
                      ]
                    },
                    {
                      "inputRefs": [
                        "application"
                      ],
                      "name": "forward-to-remote",
                      "outputRefs": [
                        "remote-elasticsearch"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogging",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "collection": {
                    "type": "vector"
                  },
                  "logStore": {
                    "elasticsearch": {
                      "nodeCount": 3,
                      "redundancyPolicy": "SingleRedundancy",
                      "resources": {
                        "requests": {
                          "memory": "2Gi"
                        }
                      },
                      "storage": {
                        "size": "200G"
                      }
                    },
                    "retentionPolicy": {
                      "application": {
                        "maxAge": "7d"
                      }
                    },
                    "type": "elasticsearch"
                  },
                  "managementState": "Managed",
                  "visualization": {
                    "kibana": {
                      "replicas": 1
                    },
                    "type": "kibana"
                  }
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.12-v4.17
          console.openshift.io/plugins: '["logging-view-plugin"]'
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:8e13bd3e1654fe26c1d4d8d4a1bfb12e04cee7899cf16b5fa07b62b8f7494af5
          createdAt: "2024-10-17T17:45:03Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing your aggregated logging stack.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.22'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-5.8
          operators.operatorframework.io.bundle.channels.v1: stable-5.8
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Logging (team-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "ClusterLogForwarder is an API to configure forwarding logs.
              \n You configure forwarding by specifying a list of `pipelines`, which
              forward from a set of named inputs to a set of named outputs. \n There
              are built-in input names for common log categories, and you can define
              custom inputs to do additional filtering. \n There is a built-in output
              name for the default openshift log store, but you can define your own
              outputs with a URL and other connection information to forward logs
              to other stores or processors, inside or outside the cluster. \n For
              more details see the documentation on the API fields."
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.logging.openshift.io
            version: v1
          - description: A Red Hat OpenShift Logging instance. ClusterLogging is the
              Schema for the clusterloggings API
            displayName: Cluster Logging
            kind: ClusterLogging
            name: clusterloggings.logging.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          ### Notice for OCP 4.16 and 4.17 Only

          Logging 5.8 is supported on OCP 4.16 and 4.17 to help customers upgrade and migrate to Logging 6.0. **Logging 5.8 will be End of Life in November 2025.**  At that time, it will no longer be available or supported.

          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates and manages the aggregated logging stack as a cluster-wide service.

          ##Features
          * **Create/Destroy**: Launch and create an aggregated logging stack to support the entire OCP cluster.
          * **Simplified Configuration**: Configure your aggregated logging cluster's structure like components and end points easily.
          * Configure log forwarding to an in-cluster, Red Hat managed log storage solution based on LokiStack or Elasticsearch

          ## Prerequisites
          ### Red Hat OpenShift Logging Namespace
          Cluster logging and the Red Hat OpenShift Logging Operator is only deployable to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - fluentd
        - logging
        - aggregated
        - efk
        - vector
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Fluentd
          url: https://www.fluentd.org/
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/logging-view-plugin-rhel9@sha256:f00d4e7ce6b37f1a053c5dbe554359d844989c2243574000ada0f5bc678ffff8
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:8e13bd3e1654fe26c1d4d8d4a1bfb12e04cee7899cf16b5fa07b62b8f7494af5
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:f26518dd3dbf6b6a9e9e20e40741687996b1160c6ad8ad417989197800c2b3fe
        - registry.redhat.io/openshift-logging/fluentd-rhel9@sha256:9296e50fccae60dea580186bc5ae2a387172dc90ece35663d96c4553a2446a01
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:8f29d6eaeb4a11c9a61e3d763605e3c9c8655df7fbb4e39e5b8311d18449f04a
        version: 5.8.22
      entries:
      - name: cluster-logging.v5.8.22
        version: 5.8.22
      - name: cluster-logging.v5.8.21
        version: 5.8.21
      - name: cluster-logging.v5.8.20
        version: 5.8.20
      - name: cluster-logging.v5.8.19
        version: 5.8.19
      - name: cluster-logging.v5.8.18
        version: 5.8.18
      - name: cluster-logging.v5.8.17
        version: 5.8.17
      - name: cluster-logging.v5.8.16
        version: 5.8.16
      - name: cluster-logging.v5.8.15
        version: 5.8.15
      - name: cluster-logging.v5.8.14
        version: 5.8.14
      - name: cluster-logging.v5.8.13
        version: 5.8.13
      - name: cluster-logging.v5.8.12
        version: 5.8.12
      - name: cluster-logging.v5.8.11
        version: 5.8.11
      - name: cluster-logging.v5.8.10
        version: 5.8.10
      - name: cluster-logging.v5.8.9
        version: 5.8.9
      - name: cluster-logging.v5.8.8
        version: 5.8.8
      - name: cluster-logging.v5.8.7
        version: 5.8.7
      - name: cluster-logging.v5.8.6
        version: 5.8.6
      - name: cluster-logging.v5.8.5
        version: 5.8.5
      - name: cluster-logging.v5.8.4
        version: 5.8.4
      - name: cluster-logging.v5.8.3
        version: 5.8.3
      - name: cluster-logging.v5.8.2
        version: 5.8.2
      - name: cluster-logging.v5.8.1
        version: 5.8.1
      - name: cluster-logging.v5.8.0
        version: 5.8.0
      name: stable-5.8
    - currentCSV: cluster-logging.v5.9.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "name": "remote-elasticsearch",
                      "secret": {
                        "name": "elasticsearch"
                      },
                      "type": "elasticsearch",
                      "url": "remote.example.org:9200"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "audit",
                        "infrastructure"
                      ],
                      "name": "enable-default-log-store",
                      "outputRefs": [
                        "default"
                      ]
                    },
                    {
                      "inputRefs": [
                        "application"
                      ],
                      "name": "forward-to-remote",
                      "outputRefs": [
                        "remote-elasticsearch"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogging",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "collection": {
                    "type": "vector"
                  },
                  "logStore": {
                    "elasticsearch": {
                      "nodeCount": 3,
                      "redundancyPolicy": "SingleRedundancy",
                      "resources": {
                        "requests": {
                          "memory": "2Gi"
                        }
                      },
                      "storage": {
                        "size": "200G"
                      }
                    },
                    "retentionPolicy": {
                      "application": {
                        "maxAge": "7d"
                      }
                    },
                    "type": "elasticsearch"
                  },
                  "managementState": "Managed",
                  "visualization": {
                    "kibana": {
                      "replicas": 1
                    },
                    "type": "kibana"
                  }
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          console.openshift.io/plugins: '["logging-view-plugin"]'
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:54054faf0d2f375dd6d7fc7929a2dbb529508e1b3c004a56fc97c2a682d16be2
          createdAt: "2025-03-14T20:19:21Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing your aggregated logging stack.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.7.0-0 <5.9.15'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Logging (team-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.


              There are built-in input names for common log categories, and you can
              define custom inputs to do additional filtering.


              There is a built-in output name for the default openshift log store, but
              you can define your own outputs with a URL and other connection information
              to forward logs to other stores or processors, inside or outside the cluster.


              For more details see the documentation on the API fields.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.logging.openshift.io
            version: v1
          - description: A Red Hat OpenShift Logging instance. ClusterLogging is the
              Schema for the clusterloggings API
            displayName: Cluster Logging
            kind: ClusterLogging
            name: clusterloggings.logging.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates and manages the aggregated logging stack as a cluster-wide service.

          ##Features
          * **Create/Destroy**: Launch and create an aggregated logging stack to support the entire OCP cluster.
          * **Simplified Configuration**: Configure your aggregated logging cluster's structure like components and end points easily.
          * Configure log forwarding to an in-cluster, Red Hat managed log storage solution based on LokiStack or Elasticsearch

          ## Prerequisites
          ### Red Hat OpenShift Logging Namespace
          Cluster logging and the Red Hat OpenShift Logging Operator is only deployable to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - fluentd
        - logging
        - aggregated
        - efk
        - vector
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Fluentd
          url: https://www.fluentd.org/
        - name: Vector
          url: https://vector.dev/
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:54054faf0d2f375dd6d7fc7929a2dbb529508e1b3c004a56fc97c2a682d16be2
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:54e3a5efaae5ecc2a08db03a421aa79e323b48c764115ea84d83a37c21d1d997
        - registry.redhat.io/openshift-logging/fluentd-rhel9@sha256:a676c9164ec2523b2002dc5eba234ce774926dbf7fe87de6d72b53f99cbf1db5
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:865572275b79462f14e84c69438a61afe9465cfe45ec164caf30965730b276f7
        - registry.redhat.io/openshift-logging/logging-view-plugin-rhel9@sha256:8023b072b45e9a21ddef83a56f4760d476053250d5a920d8eb5f03eecab8b1b3
        version: 5.9.15
      entries:
      - name: cluster-logging.v5.9.15
        version: 5.9.15
      - name: cluster-logging.v5.9.14
        version: 5.9.14
      - name: cluster-logging.v5.9.13
        version: 5.9.13
      - name: cluster-logging.v5.9.12
        version: 5.9.12
      - name: cluster-logging.v5.9.11
        version: 5.9.11
      - name: cluster-logging.v5.9.10
        version: 5.9.10
      - name: cluster-logging.v5.9.9
        version: 5.9.9
      - name: cluster-logging.v5.9.8
        version: 5.9.8
      - name: cluster-logging.v5.9.7
        version: 5.9.7
      - name: cluster-logging.v5.9.6
        version: 5.9.6
      - name: cluster-logging.v5.9.5
        version: 5.9.5
      - name: cluster-logging.v5.9.4
        version: 5.9.4
      - name: cluster-logging.v5.9.3
        version: 5.9.3
      - name: cluster-logging.v5.9.2
        version: 5.9.2
      - name: cluster-logging.v5.9.1
        version: 5.9.1
      - name: cluster-logging.v5.9.0
        version: 5.9.0
      name: stable-5.9
    - currentCSV: cluster-logging.v6.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.14-v4.17
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:3e36ea9a808945d3817d169f736b1758875f2fbc6a7ae845266e1545a1b713c7
          createdAt: "2025-05-14T12:32:39Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.8.0-0 <6.0.10'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.0
          operators.operatorframework.io.bundle.channels.v1: stable-6.0
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:3e36ea9a808945d3817d169f736b1758875f2fbc6a7ae845266e1545a1b713c7
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:4e9936c090818d1d8f812487a9a8218fda5b11f8c06167fffe35bae1bafcadf9
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:15487d7b9d8f83d0df0b6fbf380f20df3518dfb1b831fac4c3209e811494b436
        version: 6.0.10
      entries:
      - name: cluster-logging.v6.0.10
        version: 6.0.10
      - name: cluster-logging.v6.0.9
        version: 6.0.9
      - name: cluster-logging.v6.0.8
        version: 6.0.8
      - name: cluster-logging.v6.0.7
        version: 6.0.7
      - name: cluster-logging.v6.0.6
        version: 6.0.6
      - name: cluster-logging.v6.0.5
        version: 6.0.5
      - name: cluster-logging.v6.0.4
        version: 6.0.4
      - name: cluster-logging.v6.0.3
        version: 6.0.3
      - name: cluster-logging.v6.0.2
        version: 6.0.2
      - name: cluster-logging.v6.0.1
        version: 6.0.1
      - name: cluster-logging.v6.0.0
        version: 6.0.0
      name: stable-6.0
    - currentCSV: cluster-logging.v6.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.15-v4.18
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f7101c962de4e5643bd7f760104a41cc48f020889159a97e2d6711703fbbec64
          createdAt: "2025-05-14T12:38:55Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.9.0-0 <6.1.8'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.1
          operators.operatorframework.io.bundle.channels.v1: stable-6.1
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f7101c962de4e5643bd7f760104a41cc48f020889159a97e2d6711703fbbec64
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:96e438dd5c080cfa47080410f375dc71da24f6d046cf609db41c8b873d85404f
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:76037c744b34b1c4fccf18f17888813d9670bb6c061c2c3de843aa43c0497170
        version: 6.1.8
      entries:
      - name: cluster-logging.v6.1.8
        version: 6.1.8
      - name: cluster-logging.v6.1.7
        version: 6.1.7
      - name: cluster-logging.v6.1.6
        version: 6.1.6
      - name: cluster-logging.v6.1.5
        version: 6.1.5
      - name: cluster-logging.v6.1.4
        version: 6.1.4
      - name: cluster-logging.v6.1.3
        version: 6.1.3
      - name: cluster-logging.v6.1.2
        version: 6.1.2
      - name: cluster-logging.v6.1.1
        version: 6.1.1
      - name: cluster-logging.v6.1.0
        version: 6.1.0
      name: stable-6.1
    - currentCSV: cluster-logging.v6.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.16-v4.19
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
          createdAt: "2025-07-29T15:25:15Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0-0 <6.2.5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.2
          operators.operatorframework.io.bundle.channels.v1: stable-6.2
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Installing logging](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html-single/logging/index#installing-logging-6-2)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:c76c853565fc93da88ba0116425839cfb86cab63cbde9856d7cca78278d7758a
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:502a5d2891049bbc32115fda28086841351cd9e6a3110b4c3c412f683f2b2491
        version: 6.2.5
      entries:
      - name: cluster-logging.v6.2.5
        version: 6.2.5
      - name: cluster-logging.v6.2.4
        version: 6.2.4
      - name: cluster-logging.v6.2.3
        version: 6.2.3
      - name: cluster-logging.v6.2.2
        version: 6.2.2
      - name: cluster-logging.v6.2.1
        version: 6.2.1
      - name: cluster-logging.v6.2.0
        version: 6.2.0
      name: stable-6.2
    defaultChannel: stable-6.2
    packageName: cluster-logging
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-apigateway-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-apigateway-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Resource",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "RestAPI",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Stage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "VPCLink",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/apigateway-controller:1.3.1
          createdAt: "2025-09-19T21:13:12Z"
          description: AWS API Gateway controller is a service controller for managing
            API Gateway resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIIntegrationResponse represents the state of an AWS apigateway
              APIIntegrationResponse resource.
            displayName: APIIntegrationResponse
            kind: APIIntegrationResponse
            name: apiintegrationresponses.apigateway.services.k8s.aws
            version: v1alpha1
          - description: APIKey represents the state of an AWS apigateway APIKey resource.
            displayName: APIKey
            kind: APIKey
            name: apikeys.apigateway.services.k8s.aws
            version: v1alpha1
          - description: APIMethodResponse represents the state of an AWS apigateway
              APIMethodResponse resource.
            displayName: APIMethodResponse
            kind: APIMethodResponse
            name: apimethodresponses.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Authorizer represents the state of an AWS apigateway Authorizer
              resource.
            displayName: Authorizer
            kind: Authorizer
            name: authorizers.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Deployment represents the state of an AWS apigateway Deployment
              resource.
            displayName: Deployment
            kind: Deployment
            name: deployments.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Integration represents the state of an AWS apigateway Integration
              resource.
            displayName: Integration
            kind: Integration
            name: integrations.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Method represents the state of an AWS apigateway Method resource.
            displayName: Method
            kind: Method
            name: methods.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Resource represents the state of an AWS apigateway Resource
              resource.
            displayName: Resource
            kind: Resource
            name: resources.apigateway.services.k8s.aws
            version: v1alpha1
          - description: RestAPI represents the state of an AWS apigateway RestAPI
              resource.
            displayName: RestAPI
            kind: RestAPI
            name: restapis.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Stage represents the state of an AWS apigateway Stage resource.
            displayName: Stage
            kind: Stage
            name: stages.apigateway.services.k8s.aws
            version: v1alpha1
          - description: VPCLink represents the state of an AWS apigateway VPCLink
              resource.
            displayName: VPCLink
            kind: VPCLink
            name: vpclinks.apigateway.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon API Gateway resources in AWS from within your Kubernetes cluster.

          **About Amazon API Gateway**

          Amazon API Gateway is an AWS service for creating, publishing, maintaining, monitoring, and securing REST, HTTP, and WebSocket APIs at any scale. API developers can create APIs that access AWS or other web services, as well as data stored in the AWS Cloud. As an API Gateway API developer, you can create APIs for use in your own client applications. Or you can make your APIs available to third-party app developers.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon API Gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apigateway
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon API Gateway Developer Resources
          url: https://aws.amazon.com/api-gateway/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: apigateway maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/apigateway-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-apigateway-controller.v1.3.1
        version: 1.3.1
      - name: ack-apigateway-controller.v1.3.0
        version: 1.3.0
      - name: ack-apigateway-controller.v1.2.5
        version: 1.2.5
      - name: ack-apigateway-controller.v1.2.4
        version: 1.2.4
      - name: ack-apigateway-controller.v1.2.3
        version: 1.2.3
      - name: ack-apigateway-controller.v1.2.2
        version: 1.2.2
      - name: ack-apigateway-controller.v1.2.1
        version: 1.2.1
      - name: ack-apigateway-controller.v1.1.0
        version: 1.1.0
      - name: ack-apigateway-controller.v1.0.6
        version: 1.0.6
      - name: ack-apigateway-controller.v1.0.5
        version: 1.0.5
      - name: ack-apigateway-controller.v1.0.4
        version: 1.0.4
      - name: ack-apigateway-controller.v1.0.2
        version: 1.0.2
      - name: ack-apigateway-controller.v1.0.1
        version: 1.0.1
      - name: ack-apigateway-controller.v1.0.0
        version: 1.0.0
      - name: ack-apigateway-controller.v0.0.3
        version: 0.0.3
      - name: ack-apigateway-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-apigateway-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nexus-repository-ha-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-repository-ha-operator-certified.v3.84.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusRepo",
                "metadata": {
                  "name": "nexusrepo-sample"
                },
                "spec": {
                  "ingress": {
                    "additionalRules": null,
                    "annotations": null,
                    "defaultRule": false,
                    "dockerIngress": {
                      "annotations": null,
                      "enabled": false,
                      "host": "example.com",
                      "name": "nexus-docker-ingress"
                    },
                    "dockerSubdomain": false,
                    "enabled": false,
                    "host": "example.com",
                    "name": "nexus-ingress",
                    "tls": {
                      "enabled": false,
                      "secretName": "tlsSecretName"
                    }
                  },
                  "license": {
                    "fileContentsBase64": "your_license_file_contents_in_base_64",
                    "secretName": "nexus-repo-license"
                  },
                  "secret": {
                    "nexusSecret": {
                      "enabled" : false,
                      "name": "nexus-secret.json",
                      "nexusSecretsKeyId": "super_secret_key_id",
                      "secretKeyfileContentsBase64": "secretKeyfileContentsBase64",
                      "mountPath": "/var/nexus-repo-secrets"
                    }
                  },
                  "nexusData": {
                    "pvc": {
                      "accessMode": "ReadWriteOnce",
                      "size": "2Gi"
                    },
                    "storageClass": {
                      "enabled": false,
                      "name": "nexusrepo-storage",
                      "parameters": {},
                      "provisioner": "provisioner",
                      "reclaimPolicy": "Retain",
                      "volumeBindingMode": "WaitForFirstConsumer"
                    },
                    "volumeClaimTemplate": {
                      "enabled": false
                    }
                  },
                  "service": {
                    "docker": {
                      "enabled": false,
                      "name": "nexus-repo-docker-service",
                      "port": 9090,
                      "protocol": "TCP",
                      "targetPort": 9090,
                      "type": "NodePort"
                    },
                    "nexus": {
                      "enabled": false,
                      "name": "nexus-repo-service",
                      "port": 80,
                      "protocol": "TCP",
                      "targetPort": 8081,
                      "type": "NodePort"
                    }
                  },
                  "statefulset": {
                    "container": {
                      "containerPort": 8081,
                      "env": {
                        "clustered": false,
                        "install4jAddVmParams": "-Xms2703m -Xmx2703m",
                        "jdbcUrl": null,
                        "nexusInitialPassword": "nexusabc",
                        "password": "nexus",
                        "user": "nexus",
                        "zeroDowntimeEnabled": false
                      },
                      "imageName": "registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:fbad488ff2e368845473a11b4955992fe826a94811da3abf5cc3c6a33ff3974f",
                      "pullPolicy": "IfNotPresent",
                      "resources": {
                        "limits": {
                          "cpu": 4,
                          "memory": "8Gi"
                        },
                        "requests": {
                          "cpu": 4,
                          "memory": "8Gi"
                        }
                      },
                      "terminationGracePeriod": 120
                    },
                    "imagePullSecrets": {},
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 240,
                      "path": "/",
                      "periodSeconds": 60
                    },
                    "name": "nexusrepo-statefulset",
                    "readinessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 240,
                      "path": "/",
                      "periodSeconds": 60
                    },
                    "replicaCount": 1
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-certified@sha256:99f30aa5ba8469d96de3efe69d1b21e73ff022172fc79144bba496d1119ec8a2
          createdAt: "2023-08-14"
          description: |-
            Nexus Repository is the central source of control to efficiently manage all binaries
            and build artifacts across your DevOps pipeline.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.replaces: 3.80.0-2
          olm.skipRange: <3.84.0-1
          repository: https://github.com/sonatype/operator-nexus-repository
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus Repository
            displayName: NexusRepo
            kind: NexusRepo
            name: nexusrepos.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Repository is the central source of control to efficiently manage all binaries
          and build artifacts across your DevOps pipeline.
          The flow of open source and third-party components into and through an organization
          creates a complex software supply chain.
          Nexus Repository delivers speed, efficiency, and quality to the governance
          and management of all dependencies, libraries, and applications for your DevOps teams.

          ## Core Capabilities

          * **Dependency Management**:
            Improves reliability with repeatable, fast access to secure dependencies
          * **Developer Productivity**:
            Streamline developer workflows by enabling the sharing of components and applications across teams
          * **Supply Chain Performance**:
            Improve speed-to-market and reduced build times with release advanced staging and component tagging
          * **CI/CD Integrations**:
            Increase DevOps scalability with integrations to the most popular build and deployment tools

          Version control systems and package registries do not scale when managing proprietary,
          open source, and third-party components.
          Organizations need a central binary and build artifact repository to manage dependencies
          across the entire software supply chain.

          ## Limitations

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus Repository Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage
          By default, the example Custom Resource starts up a single instance of Nexus Repository in non-Pro mode using an H2 database.

          ### Connect to a PostgreSQL database
          We strongly recommend using Nexus Repository with a PostgreSQL database. To connect a Nexus Repository instance to a PostgreSQL database, do as follows:
          * Update the `license.fileContentsBase64` parameter with a base 64 representation of your Nexus Repository Pro license.
          * Set the `statefulset.container.env.jdbcUrl` parameter to your database url.
          * Set the `statefulset.container.env.user` parameter to your database user name.
          * Set the `statefulset.container.env.password` parameter to your database password.

          ### High Availability (HA) mode
          To start Nexus Repository in HA mode, you must edit the Custom Resource YAML as follows:

          * Update the `license.fileContentsBase64` parameter with a base 64 representation of your Nexus Repository Pro license.
          * Set the `statefulset.container.env.jdbcUrl` parameter to your database url.
          * Set the `statefulset.container.env.user` parameter to your database user name.
          * Set the `statefulset.container.env.password` parameter to your database password.
          * Set the `statefulset.container.env.clustered` parameter to true.
          * Set the `statefulset.replicaCount` parameter to the number of Nexus Repository pods you require in your HA cluster
          * You must use a persistent volume as described in the 'Storage' section below.

          ### Storage
          The default configuration uses an emptyDir volume for storing Nexus Repository logs. However, we strongly recommend that
          you configure dynamic provisioning of persistent storage based on your deployment environment as explained below.

          #### Cloud deployments (AWS/Azure)
          * Ensure the appropriate Container Storage Interface (CSI) driver(s) are installed for your chosen OpenShift cloud deployment.
            Please, refer to OpenShift documentation for details on configuring CSI drivers.
          Note: if you're using Red Hat OpenShift on AWS (ROSA), CSI drivers for dynamically provisioning EBS volumes are installed
          by default; you should see associated storage classes for them in your cluster web console.

          #### On premise deployments
          1. Attach separate disks (i.e. separate from the root disk) to your worker nodes.
          2. Install the Local Storage Operator. Please refer to Open Shift persistent storage documentation:
             https://docs.openshift.com/container-platform/4.13/storage/persistent_storage/persistent_storage_local/persistent-storage-local.html
          and the Local Storage Operator documentation: https://github.com/openshift/local-storage-operator.
          3. Use the Local Storage Operator to automatically create persistent volumes for your chosen storage class name as documented here:
             https://docs.openshift.com/container-platform/4.13/storage/persistent_storage/persistent_storage_local/persistent-storage-local.html
          4. Specify the storage class name from step 3 as the value of the `nexusData.storageClass.name` property and also set `nexusData.volumeClaimTemplate.enabled` to true.

          #### Configuring the operator for dynamic persistent volume provisioning
          * Set the `nexusData.storageClass.name` parameter to a storage class name. This could be one of the default storage classes
            automatically created in your OpenShift cluster (e.g., if you're using ROSA) or one that you would like the operator to create.
          * If you would like to create a dedicated storage class (i.e., you don't want to use the default), then in addition to
            specifying a value for `nexusData.storageClass.name`, you must also set `nexusData.storageClass.enabled` parameter to `true`.
          * Set the `nexusData.volumeClaimTemplate.enabled` parameter to `true`.

          ### Additional configuration
          * Once the operator has created the server instance and it is running, you can further configure Nexus Repository via the NexusRepo custom resource:

          | Parameter                                                   | Description                         | Default                                 |
          | ----------------------------------------------------------- | ----------------------------------  | ----------------------------------------|
          | `statefulset.name`                                          |  The suffix for the statefulset's service name    | nexusrepo-statefulset     |
          | `statefulset.replicaCount`                                  |  The desired number of Nexus Repository pods    | 3                           |
          | `statefulset.container.imageName`                           |  The Nexus repository image registry url    |
          | | | registry.connect.redhat.com/sonatype/nexus-repository-manager:<app version>-ubi-1|
          | `statefulset.container.resources.requests.cpu`              |  The minimum cpu the Nexus repository pod can request    | 4 |
          | `statefulset.container.resources.requests.memory`           |  The minimum memory the Nexus repository pod can request    | 8Gi|
          | `statefulset.container.resources.limits.cpu`                |  The maximum cpu the Nexus repository pod may get.    | 4|
          | `statefulset.container.resources.limits.memory`             |  The maximum memory the Nexus repository pod may get.  | 8Gi|
          | `statefulset.container.containerPort`                       |  The Nexus Repository container's HTTP port  | 8081|
          | `statefulset.container.pullPolicy`                          |  The Nexus Repository docker image pull policy  | IfNotPresent|
          | `statefulset.container.terminationGracePeriod`              |  The time given for the pod to gracefully shut down  | 120 seconds|
          | `statefulset.container.env.clustered`                       |  Controls whether Nexus Repository should start in clustered mode  | False|
          | `statefulset.container.env.install4jAddVmParams`            |  Xmx and Xms settings for JVM  | -Xms2703m -Xmx2703m|
          | `statefulset.container.env.user`                            |  The database user name | nexus (can be overridden via Kubernetes secret. See above)|
          | `statefulset.container.env.password`                        |  The database password  | nexus (can be overridden via Kubernetes secret. See above)|
          | `statefulset.container.env.jdbcUrl`                         |  The database url |
          | | | jdbc:postgresql://your_database_server:5432/nexus (must be overridden via Kubernetes |
          | | | secret object which feeds into env variable variables defined in the operator.yaml . See above)|
          | `statefulset.container.env.nexusInitialPassword`            |  The initial admin password for Nexus Repository UI. This is for first time |
          | | installs only. Subsequent installs/restarts on the same node ignores this setting | nexusabc|
          | `statefulset.livenessProbe.initialDelaySeconds`             |  The initial delay before the Nexus Repository application is pinged for a |
          | | response i.e. to determine whether the application is still alive | 240|
          | `statefulset.livenessProbe.path`                            |  The path to ping the application | |
          | `statefulset.livenessProbe.periodSeconds`                   |  The interval for checking that the application is still alive i.e. responsive, not deadlocked etc | 60|
          | `statefulset.livenessProbe.failureThreshold`                |  The maximum number of times the application ping can fail before the |
          | | application is considered unresponsive. Kubernetes will attempt to restart the application's pod in this case | 6|
          | `statefulset.readinessProbe.initialDelaySeconds`            |  The initial delay before the Nexus Repository application is checked |
          | | whether it is ready to receive traffic | 240|
          | `statefulset.readinessProbe.path`                           |  The path where the Nexus Repository application is checked for traffic readiness | |
          | `statefulset.readinessProbe.periodSeconds`                  |  The interval for checking the Nexus Repository application for traffic readiness | 60|
          | `statefulset.readinessProbe.failureThreshold`               |  The maximum number of times to check the Nexus Repository application for traffic readiness | 6|
          | `statefulset.imagePullSecrets`                              |  The pull secret for private image registries | {}|
          | `ingress.name`                                              |  The name of an Ingress object which listens for requests to Nexus Repository on port 80 | nexus-ingress|
          | `ingress.enabled`                                           |  Whether or not to create the Ingress | false|
          | `ingress.host`                                              |  Ingress host | "example.com"|
          | `ingress.defaultRule`                                       |  Whether or not to add a default rule for the Nexus Repository |
          | | Ingress which forwards traffic to a Service object  | false|
          | `ingress.additionalRules`                                   |  Additional rules to add to the ingress | null|
          | `ingress.tls.enabled`                                       |  Whether or not TLS should be enabled for the Ingress | false|
          | `ingress.tls.secretName`                                    | The name of a Secret object in which to store the TLS secret | tlsSecretName|
          | `ingress.dockerSubdomain`                                   |  Whether or not to add rules for docker subdomains | false|
          | `ingress.annotations`                                       |  Annotations for the Ingress object | null |
          | `license.fileContentsBase64`                                |  Your Base64 encoded Nexus Repository license | your_license_file_contents_in_base_64 |
          | `license.secretName`                                        |  The name of a secret object in which to store your |
          | | Base64-encoded Nexus Repository license | nexus-repo-license|
          | `ingress.dockerIngress.name`                                |  The name of an Ingress object which listens for requests to an |
          | | existing docker repository in Nexus Repository | nexus-docker-ingress|
          | `ingress.dockerIngress.enabled`                             |  Whether or not to create an Ingress for a docker repository | false|
          | `ingress.dockerIngress.host`                                |  The docker Ingress's host | "example.com"|
          | `ingress.dockerIngress.annotations`                         |  Annotations for the docker Ingress object | null |
          | `nexusData.pvc.size`                                        |  The volume size to request for storing Nexus logs | 2Gi |
          | `nexusData.pvc.accessMode`                                  |  The persistent volume claim access mode | ReadWriteOnce |
          | `nexusData.storageClass.name`                               |  The name of the storage class to use for dynamically requesting that |
          | | storage be provisioned when using dynamic provisioning or to randomly select a pre-created persisted volume when using the Local Storage Operator. | nexusrepo-storage |
          | `nexusData.storageClass.reclaimPolicy`                      |  The reclaim policy for persistent volume's created by the storage class | Retain |
          | `nexusData.storageClass.enabled`                            |  Whether or not a storage class object should be created. | false |
          | `nexusData.storageClass.volumeBindingMode`                  |  It controls when volume binding and dynamic provisioning should occur | WaitForFirstConsumer |
          | `nexusData.storageClass.provisioner`                        |  The name of the volume provisioner | provisioner |
          | `nexusData.storageClass.parameters`                         |  The additional parameters for the storage class | {} |
          | `nexusData.volumeClaimTemplate.enabled`                           |  Whether or not volume should be requested using a claim template. |
          | | You should set this property to true when using dynamic volume provisioning (cloud deployments) |
          | | or if you're using Local Storage Operator to automatically create persistent volumes for locally attached disks (i.e. not root disk) | false |
          | `license.secretName | A name for the Kubernetes secret object which stores the base64 representation of your |
          | | Nexus Repository license specified via `license.fileContentsBase64` is stored | nexus-repo-license |
          | `license.fileContentsBase64 | A base64 representation of your Nexus Repository license | your_license_file_contents_in_base_64 |
          | `service.nexus.name`                                              |  The name of a kubernetes Service Object which allows access to |
          | | Nexus Repository's main HTTP port | nexus-repo-service |
          | `service.nexus.enabled`                                           |  Whether or not to create the Service object | false|
          | `service.nexus.type`                                              |  The type of the Kubernetes Service | "NodePort"|
          | `service.nexus.protocol`                                       |  The protocol  | TCP |
          | `service.nexus.port`                                   |  The port to listen for incoming requests | 80|
          | `service.nexus.targetPort`                                   |  The port to forward requests to | 8081 |
          | `service.docker.name`                                              |  The name of a kubernetes Service Object which allows access to a |
          | | docker repository within Nexus Repository | nexus-repo-docker-service |
          | `service.docker.enabled`                                           |  Whether or not to create a Kubernetes Service object for a given |
          | | docker repository within Nexus Repository | false |
          | `service.docker.type`                                              |  The type of the Kubernetes Service | "NodePort"|
          | `service.docker.protocol`                                       |  The protocol  | TCP |
          | `service.docker.port`                                   |  The port to listen for incoming requests | 9090 |
          | `service.docker.targetPort`                                   |  The port to forward requests to | 9090 |
        displayName: Nexus Repository HA Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - repository
        - sonatype
        links:
        - name: Nexus Repository Manager
          url: https://www.sonatype.com/product-nexus-repository
        maintainers:
        - email: support@sonatype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-certified@sha256:99f30aa5ba8469d96de3efe69d1b21e73ff022172fc79144bba496d1119ec8a2
        - registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:fbad488ff2e368845473a11b4955992fe826a94811da3abf5cc3c6a33ff3974f
        version: 3.84.0-1
      entries:
      - name: nexus-repository-ha-operator-certified.v3.84.0-1
        version: 3.84.0-1
      name: stable
    defaultChannel: stable
    packageName: nexus-repository-ha-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-kubernetes.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "ClusterMongoDBRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-enterprise"
                  },
                  "name": "clustermongodbrole-sample"
                },
                "spec": {
                  "db": "admin",
                  "role": "rootMonitor",
                  "roles": [
                    {
                      "db": "admin",
                      "role": "root"
                    },
                    {
                      "db": "admin",
                      "role": "clusterMonitor"
                    }
                  ]
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "my-replica-set"
                },
                "spec": {
                  "credentials": "my-credentials",
                  "members": 3,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "sample-sharded-cluster"
                },
                "spec": {
                  "configServerCount": 3,
                  "credentials": "my-credentials",
                  "mongodsPerShardCount": 3,
                  "mongosCount": 2,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "shardCount": 1,
                  "type": "ShardedCluster",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBMultiCluster",
                "metadata": {
                  "name": "multi-replica-set"
                },
                "spec": {
                  "clusterSpecList": [
                    {
                      "clusterName": "e2e.cluster1.mongokubernetes.com",
                      "members": 2
                    },
                    {
                      "clusterName": "e2e.cluster2.mongokubernetes.com",
                      "members": 1
                    },
                    {
                      "clusterName": "e2e.cluster3.mongokubernetes.com",
                      "members": 2
                    }
                  ],
                  "credentials": "my-credentials",
                  "duplicateServiceObjects": false,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBOpsManager",
                "metadata": {
                  "name": "ops-manager"
                },
                "spec": {
                  "adminCredentials": "ops-manager-admin",
                  "applicationDatabase": {
                    "members": 3,
                    "podSpec": {
                      "cpu": 1
                    }
                  },
                  "configuration": {
                    "mms.fromEmailAddr": "admin@thecompany.com"
                  },
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "version": "6.0.3"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBUser",
                "metadata": {
                  "name": "my-replica-set-x509-user"
                },
                "spec": {
                  "db": "$external",
                  "mongodbResourceRef": {
                    "name": "my-replica-set"
                  },
                  "roles": [
                    {
                      "db": "admin",
                      "name": "dbOwner"
                    }
                  ],
                  "username": "CN=my-replica-set-x509-user,OU=cloud,O=MongoDB,L=New York,ST=New York,C=US"
                },
                "status": {
                  "phase": "Updated"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/mongodb/mongodb-kubernetes@sha256:745cc2c6bb893c1cb1e75d79b6f3bfc7b125a6a62f7431bdc43cf6cacd373998
          createdAt: "2025-09-16T14:56:17Z"
          description: The MongoDB Controllers for Kubernetes enable easy deploys
            of MongoDB into Kubernetes clusters, using our management, monitoring
            and backup platforms, Ops Manager and Cloud Manager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mongodb/mongodb-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterMongoDBRole
            name: clustermongodbroles.mongodb.com
            version: v1
          - description: MongoDB Deployment
            displayName: MongoDB Deployment
            kind: MongoDB
            name: mongodb.mongodb.com
            version: v1
          - kind: MongoDBCommunity
            name: mongodbcommunity.mongodbcommunity.mongodb.com
            version: v1
          - description: MongoDB Multi Deployment
            displayName: MongoDB Multi Deployment
            kind: MongoDBMultiCluster
            name: mongodbmulticluster.mongodb.com
            version: v1
          - kind: MongoDBSearch
            name: mongodbsearch.mongodb.com
            version: v1
          - description: MongoDB x509 User
            displayName: MongoDB User
            kind: MongoDBUser
            name: mongodbusers.mongodb.com
            version: v1
          - description: MongoDB Ops Manager
            displayName: MongoDB Ops Manager
            kind: MongoDBOpsManager
            name: opsmanagers.mongodb.com
            version: v1
        description: |
          MongoDB Controllers for Kubernetes (MCK) is an open-source project. It represents a merge of the previous
          MongoDB Community Operator (MCO) and MongoDB Enterprise Kubernetes Operator (MEKO).

          This release brings MongoDB Community and Enterprise editions together under a single, unified operator,
          making it easier to manage, scale, and upgrade your deployments. While the first version simply brings
          the capabilities of both into a single Operator, future changes will build on this to more closely align
          how Community and Enterprise are managed in Kubernetes, to offer an even more seamless and streamlined experience.
          As an open-source project, it now allows for community contributions, helping drive quicker bug fixes and ongoing innovation.
          The MongoDB Controllers for Kubernetes enable easy deploys of MongoDB
          into Kubernetes clusters, using our management, monitoring and backup
          platforms, Ops Manager and Cloud Manager.

          ## Before You Start

          To start using the operator you''ll need an account in MongoDB Cloud Manager or
          a MongoDB Ops Manager deployment.

          * [Create a Secret with your OpsManager API key](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-operator-credentials/#procedure)

          * [Create a ConfigMap with your OpsManager project ID and URL](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-project-using-configmap/)

          By installing this integration, you will be able to deploy MongoDB instances
          with a single simple command.

          ## Required Parameters

          * `opsManager` or `cloudManager` - Enter the name of the ConfigMap containing project information
          * `credentials` - Enter the name of the Secret containing your OpsManager credentials
          * `type` - Enter MongoDB Deployment Types ("Standalone", "ReplicaSet", "ShardedCluster"

          ## Supported MongoDB Deployment Types ##

          * Standalone: An instance of mongod that is running as a single server and
          not as part of a replica set, this is, it does not do any kind of
          replication.

          * Replica Set: A replica set in MongoDB is a group of mongod processes that
          maintain the same data set. Replica sets provide redundancy and high
          availability, and are the basis for all production deployments. This section
          introduces replication in MongoDB as well as the components and architecture
          of replica sets. The section also provides tutorials for common tasks
          related to replica sets.

          * Sharded Cluster: The set of nodes comprising a sharded MongoDB deployment.
          A sharded cluster consists of config servers, shards, and one or more mongos
          routing processes. Sharding is a A database architecture that partitions
          data by key ranges and distributes the data among two or more database
          instances. Sharding enables horizontal scaling.

          ## Requirements for deploying MongoDB OpsManager

          * In order to deploy resources of type MongoDB OpsManager, you will need to
          create a secret containing the [credentials](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/plan-om-resource/#om-rsrc-prereqs)
          for the Global Owner user

          ## Security ##

          The operator can enable TLS for all traffic between servers and also between
          clients and servers. Before enabling `security.tls.enabled` to `true` you
          should create your certificates.  or you can leave the operator to create all
          the certificates for you. The operator ability to create certs is been
          deprecated due to Kubernetes API changes.

          For more information, please read the official MongoDB
          Kubernetes Operator  [docs](https://docs.mongodb.com/kubernetes-operator/stable/).
        displayName: MongoDB Controllers for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mongodb
        - databases
        - nosql
        links:
        - name: Documentation
          url: https://docs.opsmanager.mongodb.com/current/tutorial/install-k8s-operator/index.html
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: stable
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-kubernetes-init-database@sha256:3ea9106c0e35c015bc8f15657c76d974bb1cadb9220dc58be2076c9699f26fb7
        - quay.io/mongodb/mongodb-agent@sha256:6e826bbe7dd812c101464e75295370c4b10e4e9c31477db89cfceb96759853e9
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:702f529cb9d8d9042e54263f8bc37b40c062ac2695efc3e7150819779886e323
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c48631742ad1fec8a9ee745fc3cc9ca1fe7d86ef7fa68ad8e8337c7a4566dc87
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7e5361c2a1604dec3fc671d629806fe6ddc6df524b69a441b86933a4b30093
        - quay.io/mongodb/mongodb-enterprise-server@sha256:85c1d669c073caf9cbaac815e4c8ea2ca18b71deefa03bf84cd6d8b8ebd53e12
        - quay.io/mongodb/mongodb-agent@sha256:fa30370a75e8e4ac648650ceb39a7a259a754aab8a48d7e2f2801cd1b777e91b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f1e72679b2468aed2f3c3106568327911e6594c9ee15f7348e953eaa0d8fd44a
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:85b4e954c9c4bbe1019774bad1e05e9c8dc175648b01d3badf0b64143f182019
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d323c1bcd14118cc537ac282996ea5a58756d20b6b73f1eec0efa8710d246077
        - quay.io/mongodb/mongodb-enterprise-server@sha256:67238129df09d33ee4746ff1acb5515d699a29d9f42350eae3f3f69ccb7fbf54
        - quay.io/mongodb/mongodb-kubernetes-database@sha256:f8c981a6223f70d7c6bbd30c503fc8b2497f351830b2b74c30e6f81863d78d2f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:a4590af027182305913f0de1a2e37ad8612bd73b8b31f4df11068b80089021ff
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3add8b68bfbaa42596fe5dbbee4623ba5ff1a826b18bb71301525adf574dcb94
        - quay.io/mongodb/mongodb-agent@sha256:ee9f3906d893004bb2aaaa6412646f58288b0684502b9e90f33cafadcfce4349
        - quay.io/mongodb/mongodb-enterprise-server@sha256:df6081325ea2fe01d1b27d0c70005dec8232aaf51a43d91661b722b8a1761263
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:146c4f5dc443f94f54d72808b4e4df93bb904e23589b222d17700166780776b4
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:09788976bdc1327acbf18cedf8efd5fca2159e43b506faa40e03ee42ad94928a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:655896bac9e82de8e9729d59a36c3a31f66235f22b93449260a2ccb7709de084
        - quay.io/mongodb/mongodb-enterprise-server@sha256:172e338a43df01ee5408d8d1e3cb038d2c55e77761c1a2ac6ed37959d1525140
        - quay.io/mongodb/mongodb-agent@sha256:3fa5bcfda3228e0ea880434a60ced829e3872e74763b347636c18475d866d0e1
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:33bbd29b74f95baad522b380d70bcc9f933a3fee57642c37b09376c86179c40f
        - quay.io/mongodb/mongodb-kubernetes@sha256:745cc2c6bb893c1cb1e75d79b6f3bfc7b125a6a62f7431bdc43cf6cacd373998
        - quay.io/mongodb/mongodb-enterprise-server@sha256:19925e1297ab652db59c35f53e7b7c8b2b277f8ecdc4215682961aaf2b3e924e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:f71b027fccf649f18f63f8df7a59b1bac421a2cdab6712e0c985a93c08ed970a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8445d59d491a6d3890ea9c8b421214bcd0fdcfbc420ed40497a790b79f0bc89e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:895dbdfb87ddc2e9693671c7108a31ab28e16fec2e5be6d046bdd2e30319eade
        - quay.io/mongodb/mongodb-enterprise-server@sha256:dfd4277f0c16ed476ac7de51894739e8c8d13a65ea5024b9a97b3dce0f6c6975
        - quay.io/mongodb/mongodb-enterprise-server@sha256:fb8397cbcdbf7c33f6ad142492792771127f0d9a8c46388f350b6927405ee590
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7885ab6e88214f9289d8babf5189136dde0dd4c84a15a5481ccd3ef512b2e66f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:5d299820bb903f0cdfdf2626a3c44099d2eaee16188cd8234d2ff3a57259b3a7
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:439d6fefb469613432e2d05149dfc96c32b16469d743835d5f402a2ada8be8cc
        - quay.io/mongodb/mongodb-agent@sha256:da77dcb60b927bf6059c2c9bf008f8c3c0eec45cf7d1b40c34a0c263e8cc18b8
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b8aa1edbc5fb02af312a863bc7db8535a282dfd692a07b1e321e5a1acbfeb6a1
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9a3a6496be0d2d446e6d654eb56386c7726e0cd2d7232f6dcec579923edf0014
        - quay.io/mongodb/mongodb-agent@sha256:89819704d70a3b6b4d2358745a22060ef740b33ad5c6debf8e6ac4d4a6fd63ba
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b7764c316eeacffd3f9ddae62d9fd2a74eb58484d2cfd5021e460095d604764f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:e7aeedb752dda77cf0e49df3430843b5a2bba64c8404628a983f46640d316808
        - quay.io/mongodb/mongodb-search@sha256:7149a7b6f1ee84646d3972837920f80fc5db49503c598406ce92e2f794bd5f0d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa8683500e8d3632c4588bbcdddb4e3186a01f48aefe1b6d95f8860dae8b6c76
        - quay.io/mongodb/mongodb-enterprise-server@sha256:15e31285c8949a4666f7702df53416b0905d76945a28575a6028a876cc8a6f0f
        - quay.io/mongodb/mongodb-kubernetes-init-ops-manager@sha256:93c4633cbc84e9bfa3563771ac2efbf5e6515f380dec2c085494cb8326546377
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5c8553e5ba115657fb878c0fe08704af9fd5decec572a87d89d2c8c156ffc9f9
        - quay.io/mongodb/mongodb-agent@sha256:5b064c470335df840b050b714b28b75b6061444843250f1eb9215cb2b8c1a149
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a86c4d8a0b9ca3c20cc524809676eb03a944ada43378ec98c107cd72cc64700f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:5e3982493e85ac1e40536758c3145b7c2fff9a7b4be9d053ee41201363a89920
        - quay.io/mongodb/mongodb-agent@sha256:5f5031e4c2a3729f618d026b0a7f34fd4509ea73d5a4e3068b551041dc291583
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:d81eac2129b6ab8346be1749c56ea67f6da64d4ee0fb9f3ff4171ba0f61918eb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3316015ac0699d66ac73f714e5dc0784738654e0349e80b1e1f4e953b3c88964
        - quay.io/mongodb/mongodb-agent@sha256:b387b7458f5640f907c9074fcb08ae4d4785be31d4bb65f2d7c459cd6762b2d6
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:adeab0dc766a27138db7df5f37cbb1da860c14179ef631b2e001e5bb875325e4
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ef732e3eea91cc86caf2eabedc0c9115ed235a8d204f33e74bbf090a0b0259fb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a48a4d5c275fd669a27469649fb39ea38c728dbcf9f0a9da4e7cf27a2ce912ed
        - quay.io/mongodb/mongodb-enterprise-server@sha256:720056b985c1a16d48b2aecb7a88a15cba3d8186e55b3c6c24729699a85c8232
        - quay.io/mongodb/mongodb-enterprise-server@sha256:523566f44cbb6bbf0cd12b1adae939a883f2254c78dc4b8c9d0098581337b1db
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f2d9a3264a8814f33143b73c56624572fb80afeab184133d3a901713a31c1a87
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5145c97d2281aabcd8538ff515c3833aa48acb66b0b739fb1881120c1985d88a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9bef77e905dadd1e61d0439dd53b21ed1f3606e49affd9ffc4ba8156578e8f0f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:cd869ff7d661059fad7593e30141d74e8cc751295b1579577f7a853396eb5d0b
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:fa4aa97767bd19cf3711a74bf21ecb9cbbc25fed6b6b09a2eb241ebd3fbc8f3a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d49417ea85ee1ddb84429552a5aac87a82c2843904d8c0300a43b657b7a49e0e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:33a30d11969eb21952681108092fe7f916e25c8e1de43e6a4e40e126025a2bec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:778f9e29e36d6687cd17dd95ec1f8127d2a732bde1a2dd5821212eb55c1c998a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1d418c65d82ccad205b7889be4ba161d3cbfa8b1fb75b21c8fcf02de38fad9ca
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b9ab7147b59b62e40a69c6400f87d0c7faecfa5462bed6e68ca653b00e50b887
        - quay.io/mongodb/mongodb-agent@sha256:6394391f78da5e3218bb2c694e7ec9f6682fd17a6dd7e7250b0c0b5e1729796d
        - quay.io/mongodb/mongodb-agent@sha256:a8358ce465adf463d8e73916d7944765d7aaf2b0ed10bd3b8119a87fc9080667
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:78f11c9cd5e63f2b0e10c67dc1491f4ce1279fa90178c108814725414defab9a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c752a7f0726d21d55f204a359b85fa29cd727a5fe571b37289ae3ea52d6206b0
        - quay.io/mongodb/mongodb-agent@sha256:b422829aac63f068aed4a6948fd14cbd3c3bfad6d47ab76523f6839bf20b0203
        - quay.io/mongodb/mongodb-agent@sha256:6c23febaa9dd52bc29ac1e52d087fc4b2b47c578f4d6ac3a1674bb750a38cb07
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:3bbd3d7fd45d107621e7770f2b4c2a758e148cd472e0ee21719442f71bf6ade0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0fde236fe26d5b20210d6f246aadacd7845874712def516cbe2eec11d79d5181
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d2d8b59c1f39c41951e7878963f8496eb494688720366f8edc7060887cdbe6a0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7a93a0276531ff9be4c90bb8fe8d104e0a9e930c29792aafe03cc6a76a9fa89c
        - quay.io/mongodb/mongodb-agent@sha256:6cc4e953df3f09d793f0dbc6e57e55a3c5277ef63177a6d8f691a41227facf64
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3fe62f3ad52707ccf0e9993c69891e44c569a4b4e1a3244cf60e3b5fdde57bd2
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1a77851485ad3b2566e765b4bb051ff0ddfa48703492cef6c6db021dde84d87f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:bfbcfcfc5f0de22e20520b2d96f826638e424adb028a909ccc192013d11f5833
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b00ac773623e1bf422b47e6d783263d2d08737865abe9abbe106d86d1ea25165
        - quay.io/mongodb/mongodb-enterprise-server@sha256:37af0913742b6de339c7e664b30f35dcef04506529e4a09a9bbbdd57643f7149
        - quay.io/mongodb/mongodb-agent@sha256:d8a2ef79d767f03162ec57fe028f9072294b292d376af8eb97d2967eb6d7047b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7266b405ebb2dd24a9f7c12d3921024f547c3d8c01831e6bcfc1d551d63be1
        - quay.io/mongodb/mongodb-agent@sha256:af3de4f631d0b00cae7f7e80c3900378ba7e449e01830bf9dc643b873d2e6d33
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa300abf739e26350a258f643fa8e57aedacd041055c7b923bd22b21ef554cdb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d21c65780cadc9d8668d2c28935d492625d6993f13ba562722200a819281cbcc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:376ad7e5fe42651d946cba5705c0eee44aa636c1b4ff258210e4ce2fe4205b4d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:454b1585c342375dea643b10ead688d0f62347841f5f30059fcd4f92e3297ee9
        - quay.io/mongodb/mongodb-enterprise-server@sha256:11ab324698f919f84a5dab5a964d93fe63f98a48146bb3b86974cba7606d30ec
        - quay.io/mongodb/mongodb-kubernetes-init-appdb@sha256:afb0d463304fc0bfb3d8dade91459414daac6bf6fcf742347a2012cbc9fea354
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c8aec83b76dfecc05b8f5d8f40fd61b24980c0287a07294f75f321c2d42c32b4
        - quay.io/mongodb/mongodb-agent@sha256:2d5652d0e263aa94554e7a068c93e4fd94b84ec223f077361d030d172bc93f67
        version: 1.4.0
      entries:
      - name: mongodb-kubernetes.v1.4.0
        version: 1.4.0
      - name: mongodb-kubernetes.v1.3.0
        version: 1.3.0
      - name: mongodb-kubernetes.v1.2.0
        version: 1.2.0
      - name: mongodb-kubernetes.v1.1.0
        version: 1.1.0
      - name: mongodb-kubernetes.v1.0.1
        version: 1.0.1
      - name: mongodb-kubernetes.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: namespace-configuration-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: namespace-configuration-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupConfig",
                "metadata": {
                  "name": "test-groupconfig"
                },
                "spec": {
                  "labelSelector": {
                    "matchLabels": {
                      "type": "team"
                    }
                  },
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-dev\n  labels:\n    group: {{ .Name }}  \n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-qa\n  labels:\n    group: {{ .Name }}          \n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-prod\n  labels:\n    group: {{ .Name }}                           \n"
                    },
                    {
                      "objectTemplate": "apiVersion: quota.openshift.io/v1\nkind: ClusterResourceQuota\nmetadata:\n  name: {{ .Name }}-quota\nspec:\n  quota:\n    hard:\n      pods: \"4\" \n      requests.cpu: \"1\" \n      requests.memory: 1Gi \n      requests.ephemeral-storage: 2Gi \n      limits.cpu: \"2\" \n      limits.memory: 2Gi \n      limits.ephemeral-storage: 4Gi\n  selector:\n    labels:\n      matchLabels:\n        group: {{ .Name }}\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NamespaceConfig",
                "metadata": {
                  "name": "test-namespaceconfig"
                },
                "spec": {
                  "labelSelector": {
                    "matchLabels": {
                      "type": "secure"
                    }
                  },
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: ResourceQuota\nmetadata:\n  name: standard-sandbox\n  namespace: {{ .Name }}\nspec:\n  hard:\n    pods: \"4\" \n    requests.cpu: \"1\" \n    requests.memory: 1Gi \n    requests.ephemeral-storage: 2Gi \n    limits.cpu: \"2\" \n    limits.memory: 2Gi \n    limits.ephemeral-storage: 4Gi\n"
                    },
                    {
                      "objectTemplate": "kind: EgressNetworkPolicy\napiVersion: network.openshift.io/v1\nmetadata:\n  name: air-gapped\n  namespace: {{ .Name }}  \nspec:\n  egress:\n  - type: Deny\n    to:\n      cidrSelector: 0.0.0.0/0\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "UserConfig",
                "metadata": {
                  "name": "test-user-config"
                },
                "spec": {
                  "providerName": "my-provider",
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-sandbox\n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: ResourceQuota\nmetadata:\n  name: standard-sandbox\n  namespace: {{ .Name }}-sandbox\nspec:\n  hard:\n    pods: \"4\" \n    requests.cpu: \"1\" \n    requests.memory: 1Gi \n    requests.ephemeral-storage: 2Gi \n    limits.cpu: \"2\" \n    limits.memory: 2Gi \n    limits.ephemeral-storage: 4Gi\n"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/namespace-configuration-operator@sha256:49ed7d6155342adaa2b12fd80c6761c3081d8e6149d187cb7ff91a247cdf2e7a
          createdAt: "2023-11-22T15:58:50Z"
          description: This operator provides a facility to define and enforce namespace
            configurations
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: namespace-configuration-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/namespace-configuration-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GroupConfig is the Schema for the groupconfigs API
            displayName: Group Config
            kind: GroupConfig
            name: groupconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: NamespaceConfig is the Schema for the namespaceconfigs API
            displayName: Namespace Config
            kind: NamespaceConfig
            name: namespaceconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: UserConfig is the Schema for the userconfigs API
            displayName: User Config
            kind: UserConfig
            name: userconfigs.redhatcop.redhat.io
            version: v1alpha1
        description: "The `namespace-configuration-operator` helps keeping configurations
          related to Users, Groups and Namespaces aligned with one of more policies
          specified as a CRs. The purpose is to provide the foundational building
          block to create an end-to-end onboarding process. \nBy onboarding process
          we mean all the provisioning steps needed to a developer team working on
          one or more applications to OpenShift.\n This usually involves configuring
          resources such as: Groups, RoleBindings, Namespaces, ResourceQuotas, NetworkPolicies,
          EgressNetworkPolicies, etc.... . Depending on the specific environment the
          list could continue.\nNaturally such a process should be as automatic and
          scalable as possible.\n\nWith the namespace-configuration-operator one can
          create rules that will react to the creation of Users, Groups and Namespace
          and will create and enforce a set of resources.\n\nHere are some examples
          of the type of onboarding processes that one could support:\n\n1. [developer
          sandbox](https://github.com/redhat-cop/namespace-configuration-operator/blob/master/examples/user-sandbox/readme.md)\n
          2. [team onboarding](https://github.com/redhat-cop/namespace-configuration-operator//blob/master/examples/team-onboarding/readme.md)
          with support of the entire SDLC in a multitentant environment.\n\nPolicies
          can be expressed with the following CRDs:\n\n| Watched Resource | CRD |\n|--|--|\n
          | Groups | [GroupConfig](#GroupConfig) |\n| Users | [UserConfig](#UserConfig)
          |\n| Namespace | [NamespaceConfig](#NamespaceConfig) |\n\nThese CRDs all
          share some commonalities:\n\n1. Templated Resources\n2. List of ignored
          jason path\n \n### Templated Resources\n\nEach has a parameter called `templatedResources`,
          which is an array. Each element of the array has two fields `objectTemplate`
          and `excludedPaths` (see below).\n\nThe `objectTemplate` field must contains
          a [go template](https://golang.org/pkg/text/template/) that resolves to
          a single API Resource expressed in `yaml`. The template is merged with the
          object selected by the CR. For example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: UserConfig\nmetadata:\n  name: test-user-config\nspec:\n  provider:
          corp-ldap\n  templates:\n  - objectTemplate: |\n      apiVersion: v1\n      kind:
          Namespace\n      metadata:\n        name: {{ .Name }}-sandbox\n```\n\nThis
          create a rule in which every time a user from the `corp-ldap` provider is
          created, a namespace called `<username>-sandbox` is also created.\n\n###
          Excluded Paths\n\nThe logic of the `namespace-configuration-operator` is
          to enforce that the resources resolved by processing the templates \"stays
          in place\". In other words if those resources are changed and/or deleted
          they will be reset by the operator.\nBut there are situations in which at
          least part of a resource is allowed to change. Common use cases are: annotations
          and in general the metadata section of a resource can be updated by the
          various operators watching that resources. The status field is often updated
          by the main operator managing that resources. Finally, when applicable the
          `spec.replicas` field should also be allowed to change.\n\nTo handle special
          use case, one can also specify additional *jsonpaths* that should be ignored
          when comparing the desired resource and the current resource and making
          a decision on whether that resource should be reset.\n\nThe following paths
          are always included:\n \n1. `.metadata`\n2. `.status`\n3. `.spec.replicas`\n\n##
          NamespaceConfig\n\n The `NamespaceConfig` CR allows specifying one or more
          objects that will be created in the selected namespaces.\n\nNamespaces can
          be selected by labels or annotations via a label selector for example:\n\n```yaml\napiVersion:
          redhatcop.redhat.io/v1alpha1\n kind: NamespaceConfig\nmetadata:\n  name:
          small-namespace\n  selector:\n    matchLabels:\n      size: small  \n  resources:\n
          \ - apiVersion: v1\n    kind: ResourceQuota\n    metadata:\n      name:
          small-size  \n    spec:\n      hard:\n        requests.cpu: \"4\"\n        requests.memory:
          \"2Gi\"\n```\n\nHere is a `NamespaceConfig` object using a `matchExpressions`
          selector:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n kind: NamespaceConfig\nmetadata:\n
          \ name: tier-config\nspec:\n  annotationSelector:\n    matchExpressions:\n
          \     - {key: tier, operator: In, values: [gold,silver]}\n ```\n\nAlthough
          not enforced by the operator the general expectation is that the NamespaceConfig
          CR will be used to create objects inside the selected namespace.\n \nExamples
          of NamespaceConfig usages can be found [here](https://github.com/redhat-cop/namespace-configuration-operator/blob/master/examples/namespace-config/readme.md)\n
          \n## GroupConfig\n\nThe `GroupConfig` CR allows specifying one or more objects
          that will be created in the selected Group.\nGroups can be selected by labels
          or annotations via a label selector, similarly to the `NamespaceConfig`.\n\nOften
          groups are created in OpenShift by a job that synchronizes an Identity Provider
          with OCP. So the idea is that when new groups are added or deleted the configuration
          in OpenShift will adapt automatically.\n\nAlthough not enforced by the operator,
          GroupConfig are expected to create cluster-scoped resources like Namespaces,
          ClusterResourceQuotas and potentially some namespaced resources like RoleBindings.\n\n##
          UserConfig\n \nIn OpenShift an external user is defined by two entities:
          Users and Identities. There is a relationship of on to many between Users
          and Identities. Given one user, there can be one Identity per authentication
          mechanism.\n\nThe `UserConfig` CR allows specifying one or more objects
          that will be created in the selected User.\nUsers can be selected by label
          or annotation like `NamespaceConfig` and `UserConfig`.\nUSers can also be
          selected by provider name (the name of the authentication mechanism) and
          identity extra field.\n\nHere is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          UserConfig\nmetadata:\n  name: test-user-config\nspec:\n  providerName:
          okta-provider\n  identityExtraFieldSelector:\n    matchLabels:\n      sandbox_enabled:
          \"true\"\n  templates:\n  - objectTemplate: |\n      apiVersion: v1\n      kind:
          Namespace\n      metadata:\n        name: {{ .Name }}-sandbox\n ```\n\nUser
          will be selected by this `UserConfig` only if they login via the *okta-provider*
          and if the extra field was populate with the label `sandbox_enabled: \"true\"
          `. Note that not all authentication provider allow populating the extra
          fields in the Identity object.\n"
        displayName: Namespace Configuration Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - namespace
        - user
        - group
        - configuration
        - policy
        - management
        links:
        - name: Namespace Configuration Operator
          url: https://github.com/redhat-cop/namespace-configuration-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/namespace-configuration-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/namespace-configuration-operator@sha256:49ed7d6155342adaa2b12fd80c6761c3081d8e6149d187cb7ff91a247cdf2e7a
        version: 1.2.6
      entries:
      - name: namespace-configuration-operator.v1.2.6
        version: 1.2.6
      - name: namespace-configuration-operator.v1.2.5
        version: 1.2.5
      - name: namespace-configuration-operator.v1.2.4
        version: 1.2.4
      - name: namespace-configuration-operator.v1.2.3
        version: 1.2.3
      - name: namespace-configuration-operator.v1.2.2
        version: 1.2.2
      - name: namespace-configuration-operator.v1.2.1
        version: 1.2.1
      - name: namespace-configuration-operator.v1.2.0
        version: 1.2.0
      - name: namespace-configuration-operator.v1.1.0
        version: 1.1.0
      - name: namespace-configuration-operator.v1.0.7
        version: 1.0.7
      - name: namespace-configuration-operator.v1.0.6
        version: 1.0.6
      - name: namespace-configuration-operator.v1.0.5
        version: 1.0.5
      - name: namespace-configuration-operator.v1.0.4
        version: 1.0.4
      - name: namespace-configuration-operator.v1.0.3
        version: 1.0.3
      - name: namespace-configuration-operator.v1.0.2
        version: 1.0.2
      - name: namespace-configuration-operator.v1.0.1
        version: 1.0.1
      - name: namespace-configuration-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: namespace-configuration-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: geunwookim
      provider-url: ""
    name: ncn-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ncn-operator.v0.3.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-1"
                },
                "spec": null
              },
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "label-1"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
          - description: Label is the Schema for the labels API
            displayName: Label
            kind: Label
            name: labels.cxl.antline.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: jwahn@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Antline
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/rasen708/ncn-operator:v0.4.0
        version: 0.3.18
      entries:
      - name: ncn-operator.v0.3.18
        version: 0.3.18
      - name: ncn-operator.v0.3.14
        version: 0.3.14
      name: alpha
    - currentCSV: ncn-operator.v0.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-1"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-19T01:21:08Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: jwahn@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        provider:
          name: Antline
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/rasen708/ncn-operator:v0.2.2
        version: 0.2.2
      entries:
      - name: ncn-operator.v0.2.2
        version: 0.2.2
      - name: ncn-operator.v0.2.1
        version: 0.2.1
      - name: ncn-operator.v0.2.0
        version: 0.2.0
      name: beta
    - currentCSV: ncn-operator.v0.9.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cmmd.gw.com/v1",
                "kind": "CMMD",
                "metadata": {
                  "name": "my-app",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "allocateMode": "auto",
                  "enable": true,
                  "payload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "metadata": {
                      "name": "my-app"
                    },
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "name": "my-app"
                        }
                      },
                      "template": {
                        "metadata": {
                          "labels": {
                            "name": "my-app"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "quay.io/my-account/my-app:latest",
                              "name": "my-app-01"
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "cmmd.gw.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "machineconfiguration.openshift.io/role": "cmmd"
                  },
                  "name": "cmmd-node-label",
                  "namespace": "socmmd"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          console.openshift.io/plugins: '["socmmd-console-plugin"]'
          containerImage: quay.io/rasen708/ncn-operator@sha256:a09963745bd9c1419532d42d91237b7496c7a78e6a1c460a8e714ea1a0ce4d45
          createdAt: "2024-08-01T05:45:38Z"
          description: Provides the ncn extend
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "cmmd.gw.com/v1",
              "kind": "Label",
              "metadata": {
                "name": "cmmd-node-label",
                "namespace": "socmmd",
                "labels": {}
              },
              "spec": {}
            }
          operatorframework.io/suggested-namespace: socmmd-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/geunwookim/ncn-operator
          support: geunwookim
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CMMD is the Schema for the cmmds API
            displayName: CMMD
            kind: CMMD
            name: cmmds.cmmd.gw.com
            version: v1
          - description: Label is the Schema for the labels API
            displayName: Node Label
            kind: Label
            name: labels.cmmd.gw.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: geunwookim@gmail.com
          name: geunwookim
        maturity: betav1
        provider:
          name: geunwookim
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/rasen708/ncn-operator@sha256:a09963745bd9c1419532d42d91237b7496c7a78e6a1c460a8e714ea1a0ce4d45
        version: 0.9.8
      entries:
      - name: ncn-operator.v0.9.8
        version: 0.9.8
      - name: ncn-operator.v0.9.2
        version: 0.9.2
      - name: ncn-operator.v0.8.1
        version: 0.8.1
      - name: ncn-operator.v0.5.6
        version: 0.5.6
      - name: ncn-operator.v0.4.3
        version: 0.4.3
      - name: ncn-operator.v0.4.2
        version: 0.4.2
      - name: ncn-operator.v0.3.18
        version: 0.3.18
      - name: ncn-operator.v0.3.14
        version: 0.3.14
      name: betav1
    defaultChannel: betav1
    packageName: ncn-operator
    provider:
      name: geunwookim
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CNCF
      provider-url: ""
    name: etcd
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: etcdoperator.v0.6.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          description: etcd is a distributed key value store providing a reliable
            way to store data across a cluster of machines.
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
        description: |
          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader. Your applications can read and write data into etcd.
          A simple use-case is to store database connection details or feature flags within etcd as key value pairs. These values can be watched, allowing your app to reconfigure itself when they change. Advanced uses take advantage of the consistency guarantees to implement database leader elections or do distributed locking across a cluster of workers.

          _The etcd Open Cloud Service is Public Alpha. The goal before Beta is to fully implement backup features._

          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` or with the API using the automatically generated Kubernetes Service.

          [Read the complete guide to using the etcd Open Cloud Service](https://coreos.com/tectonic/docs/latest/alm/etcd-ocs.html)

          ### Supported Features
          **High availability**
          Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.
          **Automated updates**
          Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.
          **Backups included**
          Coming soon, the ability to schedule backups to happen on or off cluster.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: support@coreos.com
          name: CoreOS, Inc
        maturity: alpha
        provider:
          name: CoreOS, Inc
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:bd944a211eaf8f31da5e6d69e8541e7cada8f16a9f7a5a570b22478997819943
        version: 0.6.1
      entries:
      - name: etcdoperator.v0.6.1
        version: 0.6.1
      name: alpha
    - currentCSV: etcdoperator.v0.9.4-clusterwide
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdCluster",
                "metadata": {
                  "name": "example",
                  "annotations": {
                    "etcd.database.coreos.com/scope": "clusterwide"
                  }
                },
                "spec": {
                  "size": 3,
                  "version": "3.2.13"
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdRestore",
                "metadata": {
                  "name": "example-etcd-cluster-restore"
                },
                "spec": {
                  "etcdCluster": {
                    "name": "example-etcd-cluster"
                  },
                  "backupStorageType": "S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdBackup",
                "metadata": {
                  "name": "example-etcd-cluster-backup"
                },
                "spec": {
                  "etcdEndpoints": ["<etcd-cluster-endpoints>"],
                  "storageType":"S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
          createdAt: "2019-02-28 01:03:00"
          description: Create and maintain highly-available etcd clusters on Kubernetes
          repository: https://github.com/coreos/etcd-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to backup an etcd cluster.
            displayName: etcd Backup
            kind: EtcdBackup
            name: etcdbackups.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to restore an etcd cluster from a backup.
            displayName: etcd Restore
            kind: EtcdRestore
            name: etcdrestores.etcd.database.coreos.com
            version: v1beta2
        description: |
          The etcd Operater creates and maintains highly-available etcd clusters on Kubernetes, allowing engineers to easily deploy and manage etcd clusters for their applications.

          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader.


          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` via port forwarding:

              $ kubectl --namespace default port-forward service/example-client 2379:2379
              $ etcdctl --endpoints http://127.0.0.1:2379 get /

          Or directly to the API using the automatically generated Kubernetes Service:

              $ etcdctl --endpoints http://example-client.default.svc:2379 get /

          Be sure to secure your etcd cluster (see Common Configurations) before exposing it outside of the namespace or cluster.


          ### Supported Features

          * **High availability** - Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.

          * **Automated updates** - Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.

          * **Backups included** - Create etcd backups and restore them through the etcd Operator.

          ### Common Configurations

          * **Configure TLS** - Specify [static TLS certs](https://github.com/coreos/etcd-operator/blob/master/doc/user/cluster_tls.md) as Kubernetes secrets.

          * **Set Node Selector and Affinity** - [Spread your etcd Pods](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-node-selector-and-anti-affinity-across-nodes) across Nodes and availability zones.

          * **Set Resource Limits** - [Set the Kubernetes limit and request](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-resource-requirement) values for your etcd Pods.

          * **Customize Storage** - [Set a custom StorageClass](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#custom-persistentvolumeclaim-definition) that you would like to use.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: etcd-dev@googlegroups.com
          name: etcd Community
        maturity: alpha
        provider:
          name: CNCF
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
        version: 0.9.4-clusterwide
      entries:
      - name: etcdoperator.v0.9.4-clusterwide
        version: 0.9.4-clusterwide
      - name: etcdoperator.v0.9.2-clusterwide
        version: 0.9.2-clusterwide
      - name: etcdoperator.v0.9.0
        version: 0.9.0
      name: clusterwide-alpha
    - currentCSV: etcdoperator.v0.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "size": 3,
                  "version": "3.2.13"
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdRestore",
                "metadata": {
                  "name": "example-etcd-cluster-restore"
                },
                "spec": {
                  "etcdCluster": {
                    "name": "example-etcd-cluster"
                  },
                  "backupStorageType": "S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdBackup",
                "metadata": {
                  "name": "example-etcd-cluster-backup"
                },
                "spec": {
                  "etcdEndpoints": ["<etcd-cluster-endpoints>"],
                  "storageType":"S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
          createdAt: "2019-02-28 01:03:00"
          description: Create and maintain highly-available etcd clusters on Kubernetes
          repository: https://github.com/coreos/etcd-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to backup an etcd cluster.
            displayName: etcd Backup
            kind: EtcdBackup
            name: etcdbackups.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to restore an etcd cluster from a backup.
            displayName: etcd Restore
            kind: EtcdRestore
            name: etcdrestores.etcd.database.coreos.com
            version: v1beta2
        description: |
          The etcd Operater creates and maintains highly-available etcd clusters on Kubernetes, allowing engineers to easily deploy and manage etcd clusters for their applications.

          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader.


          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` via port forwarding:

              $ kubectl --namespace default port-forward service/example-client 2379:2379
              $ etcdctl --endpoints http://127.0.0.1:2379 get /

          Or directly to the API using the automatically generated Kubernetes Service:

              $ etcdctl --endpoints http://example-client.default.svc:2379 get /

          Be sure to secure your etcd cluster (see Common Configurations) before exposing it outside of the namespace or cluster.


          ### Supported Features

          * **High availability** - Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.

          * **Automated updates** - Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.

          * **Backups included** - Create etcd backups and restore them through the etcd Operator.

          ### Common Configurations

          * **Configure TLS** - Specify [static TLS certs](https://github.com/coreos/etcd-operator/blob/master/doc/user/cluster_tls.md) as Kubernetes secrets.

          * **Set Node Selector and Affinity** - [Spread your etcd Pods](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-node-selector-and-anti-affinity-across-nodes) across Nodes and availability zones.

          * **Set Resource Limits** - [Set the Kubernetes limit and request](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-resource-requirement) values for your etcd Pods.

          * **Customize Storage** - [Set a custom StorageClass](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#custom-persistentvolumeclaim-definition) that you would like to use.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: etcd-dev@googlegroups.com
          name: etcd Community
        maturity: alpha
        provider:
          name: CNCF
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
        version: 0.9.4
      entries:
      - name: etcdoperator.v0.9.4
        version: 0.9.4
      - name: etcdoperator.v0.9.2
        version: 0.9.2
      - name: etcdoperator.v0.9.0
        version: 0.9.0
      name: singlenamespace-alpha
    defaultChannel: singlenamespace-alpha
    packageName: etcd
    provider:
      name: CNCF
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataStax
      provider-url: ""
    name: cass-operator-community
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cass-operator.v1.23.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cassandra.datastax.com/v1beta1",
                "kind": "CassandraDatacenter",
                "metadata": {
                  "name": "dc1"
                },
                "spec": {
                  "clusterName": "development",
                  "config": {
                    "cassandra-yaml": {
                      "authenticator": "PasswordAuthenticator",
                      "authorizer": "CassandraAuthorizer",
                      "num_tokens": 16,
                      "role_manager": "CassandraRoleManager"
                    },
                    "jvm-server-options": {
                      "initial_heap_size": "1G",
                      "max_heap_size": "1G"
                    }
                  },
                  "managementApiAuth": {
                    "insecure": {}
                  },
                  "podTemplateSpec": {
                    "spec": {
                      "containers": [
                        {
                          "name": "cassandra"
                        }
                      ],
                      "securityContext": {}
                    }
                  },
                  "racks": [
                    {
                      "name": "rack1"
                    },
                    {
                      "name": "rack2"
                    },
                    {
                      "name": "rack3"
                    }
                  ],
                  "resources": {
                    "requests": {
                      "cpu": "1000m",
                      "memory": "2Gi"
                    }
                  },
                  "serverType": "cassandra",
                  "serverVersion": "4.1.2",
                  "size": 3,
                  "storageConfig": {
                    "cassandraDataVolumeClaimSpec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "server-storage"
                    }
                  }
                }
              },
              {
                "apiVersion": "control.k8ssandra.io/v1alpha1",
                "kind": "CassandraTask",
                "metadata": {
                  "name": "example-cleanup"
                },
                "spec": {
                  "datacenter": {
                    "name": "dc2",
                    "namespace": "cass-operator"
                  },
                  "jobs": [
                    {
                      "command": "cleanup",
                      "name": "cleanup-run"
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database,Big Data,OpenShift Optional
          certified: "true"
          containerImage: cr.k8ssandra.io/k8ssandra/cass-operator:v1.23.2
          createdAt: "2025-03-20"
          description: |
            Simple provisioning, turn-key operations, and automated remediation of Apache Cassandra clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/k8ssandra/cass-operator
          support: DataStax
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CassandraDatacenter is the Schema for the cassandradatacenters
              API
            displayName: Cassandra Datacenter
            kind: CassandraDatacenter
            name: cassandradatacenters.cassandra.datastax.com
            version: v1beta1
          - description: CassandraTask is the Schema for the cassandrajobs API
            displayName: Cassandra Task
            kind: CassandraTask
            name: cassandratasks.control.k8ssandra.io
            version: v1alpha1
        description: "## Apache Cassandra\n\nThe Apache Cassandra database is the
          right choice when you need scalability and\nhigh availability without compromising
          performance. Linear scalability and\nproven fault-tolerance on commodity
          hardware or cloud infrastructure make it the\nperfect platform for mission-critical
          data. Cassandra's support for replicating\nacross multiple datacenters is
          best-in-class, providing lower latency for your\nusers and the peace of
          mind of knowing that you can survive regional outages.\n\n## DataStax Enterprise\n\nThe
          most advanced distribution of Apache Cassandra™ on the market, with the\nenterprise
          functionality needed for serious production systems and backed up and\nsupported
          by the best distributed-experts in the world. It's one platform for\nall
          types of applications anywhere, any cloud, any model: key-value, graph,\ntabular,
          JSON.\n\nDataStax Enterprise is a fully integrated and optimized database,
          with graph,\nanalytics, and search included, all with a unified security
          model. Simply put,\nit's the only database capable of meeting today's demanding
          requirements\n\n## Operator Details\n\n`cass-operator` is designed as a
          modular operator for Apache Cassandra and\nderived  distributions. Apache
          Cassandra is a distributed database consisting of\nmultiple nodes working
          in concert to store data and process queries along a\nnumber of fault domains.
          `cass-operator` has the deployment of a Cassandra\ncluster around the logical
          domain of a datacenter with the `CassandraDatacenter`\ncustom resource.\n
          \   \nUpon submission of one of these resources it handles provisioning
          the underlying\nstateful sets (analogous to C\\* logical racks), services,
          and configuration.\nAdditionally through monitoring pod state via Kubernetes
          callbacks it handles day to day\noperations such as restarting failed processes,
          scaling clusters up, and deploying\nconfiguration changes in a rolling,
          non-disruptive, fashion.\n    \nThis operator is designed to be `Namespace`
          scoped. A single Kubernetes cluster may\nbe running multiple instances of
          this operator, in separate namespaces, to support\na number of C\\* clusters
          and environments. Configuration is simple with the usage of\nYAML based
          overrides in the Custom Resource paired with an `init` container.\n    \nIn
          C\\* clusters ordering and timing of certain operations are important to
          keep the system\nevenly distributed. `cass-operator` takes advantage of
          a sidecar process within the\nmain container to handle the orchestration
          of starting our main C* process."
        displayName: DataStax Kubernetes Operator for Apache Cassandra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - k8ssandra
        - cass-operator
        - DataStax Enterprise
        - Apache Cassandra
        - DSE
        - C*
        - Cassandra
        links:
        - name: DataStax
          url: https://datastax.com
        - name: Operator GitHub Repo
          url: https://github.com/k8ssandra/cass-operator
        - name: Config Builder GitHub Repo
          url: https://github.com/datastax/cass-config-builder
        - name: Documentation [1]
          url: https://github.com/k8ssandra/cass-operator/tree/master/docs/user
        - name: Documentation [2]
          url: https://docs.datastax.com/en/cass-operator/doc/cass-operator/cassOperatorTOC.html
        maintainers:
        - email: cass-operator@datastax.com
          name: DataStax Operator Team
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: DataStax
        relatedImages:
        - cr.k8ssandra.io/k8ssandra/cass-operator:v1.23.2
        version: 1.23.2
      entries:
      - name: cass-operator.v1.23.2
        version: 1.23.2
      - name: cass-operator.v1.23.1
        version: 1.23.1
      - name: cass-operator.v1.23.0
        version: 1.23.0
      - name: cass-operator.v1.21.1
        version: 1.21.1
      - name: cass-operator.v1.19.1
        version: 1.19.1
      - name: cass-operator.v1.17.1
        version: 1.17.1
      - name: cass-operator.v1.16.0
        version: 1.16.0
      - name: cass-operator.v1.14.0
        version: 1.14.0
      - name: cass-operator.v1.13.0
        version: 1.13.0
      - name: cass-operator.v1.10.4
        version: 1.10.4
      - name: cass-operator.v1.10.1
        version: 1.10.1
      name: stable
    defaultChannel: stable
    packageName: cass-operator-community
    provider:
      name: DataStax
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: openstack-ansibleee-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openstack-ansibleee-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ansibleee.openstack.org/v1beta1",
                "kind": "OpenStackAnsibleEE",
                "metadata": {
                  "name": "openstackansibleee-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-09-23T20:57:34Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.3
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenStackAnsibleEE is the Schema for the openstackansibleees
              API
            displayName: OpenStack Ansible EE
            kind: OpenStackAnsibleEE
            name: openstackansibleees.ansibleee.openstack.org
            version: v1beta1
        description: The OpenStack Ansible Execution Environment Operator
        displayName: OpenStackAnsibleEE
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - execution
        links:
        - name: OpenStack Ansible EE Operator
          url: https://github.com/openstack-k8s-operators/openstack-ansibleee-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:24b468b09b43dd6cf77f5b08a68f2c6359d115722d28982fd8cdb1923e8122de
        - registry.redhat.io/rhoso-operators/openstack-ansibleee-rhel9-operator@sha256:bf096518547e92fb6fe0ecd526813bdc4a15f928a6d89eacadb71ffaad91db69
        - registry.redhat.io/rhoso-operators/ee-openstack-ansible-ee-rhel9@sha256:640efe595fbdb993ecc1c55c14457edc2b0e3e1c6dde8ad9ea1acf4b4503339f
        version: 1.0.3
      entries:
      - name: openstack-ansibleee-operator.v1.0.3
        version: 1.0.3
      - name: openstack-ansibleee-operator.v1.0.2
        version: 1.0.2
      - name: openstack-ansibleee-operator.v1.0.1
        version: 1.0.1
      - name: openstack-ansibleee-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: openstack-ansibleee-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: ""
    name: red-hat-hawtio-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: red-hat-hawtio-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              },
              {
                "apiVersion": "hawt.io/v2",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24,
                    "internalSSL": true
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
          createdAt: "2025-08-07T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v2
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
        version: 1.2.1
      entries:
      - name: red-hat-hawtio-operator.v1.2.1
        version: 1.2.1
      - name: red-hat-hawtio-operator.v1.2.0-0.1754874242.p
        version: 1.2.0+0.1754874242.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754338551.p
        version: 1.2.0+0.1754338551.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754317392.p
        version: 1.2.0+0.1754317392.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1753863723.p
        version: 1.2.0+0.1753863723.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752679472.p
        version: 1.2.0+0.1752679472.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752452299.p
        version: 1.2.0+0.1752452299.p
      - name: red-hat-hawtio-operator.v1.2.0
        version: 1.2.0
      - name: red-hat-hawtio-operator.v1.1.1
        version: 1.1.1
      - name: red-hat-hawtio-operator.v1.1.0
        version: 1.1.0
      - name: red-hat-hawtio-operator.v1.1.0-0.1742200932.p
        version: 1.1.0+0.1742200932.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1740358952.p
        version: 1.1.0+0.1740358952.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739994212.p
        version: 1.1.0+0.1739994212.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739992853.p
        version: 1.1.0+0.1739992853.p
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: latest
    - currentCSV: red-hat-hawtio-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              },
              {
                "apiVersion": "hawt.io/v2",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24,
                    "internalSSL": true
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
          createdAt: "2025-08-07T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v2
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
        version: 1.2.1
      entries:
      - name: red-hat-hawtio-operator.v1.2.1
        version: 1.2.1
      - name: red-hat-hawtio-operator.v1.2.0-0.1754874242.p
        version: 1.2.0+0.1754874242.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754338551.p
        version: 1.2.0+0.1754338551.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754317392.p
        version: 1.2.0+0.1754317392.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1753863723.p
        version: 1.2.0+0.1753863723.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752679472.p
        version: 1.2.0+0.1752679472.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752452299.p
        version: 1.2.0+0.1752452299.p
      - name: red-hat-hawtio-operator.v1.2.0
        version: 1.2.0
      - name: red-hat-hawtio-operator.v1.1.1
        version: 1.1.1
      - name: red-hat-hawtio-operator.v1.1.0
        version: 1.1.0
      - name: red-hat-hawtio-operator.v1.1.0-0.1742200932.p
        version: 1.1.0+0.1742200932.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1740358952.p
        version: 1.1.0+0.1740358952.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739994212.p
        version: 1.1.0+0.1739994212.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739992853.p
        version: 1.1.0+0.1739992853.p
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: stable-v1
    - currentCSV: red-hat-hawtio-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel8-operator@sha256:f04d81d24c757db7b678464382ad8c5182b1580e714123a9f4f5fc9fbd9b8c57
          createdAt: "2024-01-30T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-rhel8-operator@sha256:f04d81d24c757db7b678464382ad8c5182b1580e714123a9f4f5fc9fbd9b8c57
        version: 1.0.3
      entries:
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1
    packageName: red-hat-hawtio-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MinIO
      provider-url: https://min.io
    name: minio-key-manager-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-key-manager-operator.v2025.8.19175300
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aistor.min.io/v1alpha1",
                "kind": "KeyManager",
                "metadata": {
                  "name": "my-kms",
                  "namespace": "ns-1"
                },
                "spec": {
                  "apiKeySecret": {
                    "name": "my-api-key"
                  },
                  "configuration": {
                    "name": "my-kms-server-config"
                  },
                  "hsmSecret": {
                    "name": "my-kms-hsm"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "registry-creds"
                    }
                  ],
                  "replicas": 2,
                  "volumeClaimTemplate": {
                    "metadata": {
                      "name": "key-manager-volume"
                    },
                    "spec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "100Mi"
                        }
                      },
                      "storageClassName": "standard"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security, Storage
          createdAt: "2025-08-27T14:50:59Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KeyManager
            name: keymanagers.aistor.min.io
            version: v1alpha1
        description: |-
          Highly available, powerful and operationally simple, MinIO's Enterprise Key Management Server is optimized for large storage infrastructures where billions of cryptographic keys are required.

          MinIO's AIStor KMS establishes its foundational trust using the concept of an hardware security module (but given KES is software, this is only a concept). That module assumes a pivotal role in sealing and unsealing the KMS root encryption key. The module responsibility extends to safeguarding the integrity of KMS by allowing the unsealing of its encrypted on-disk state and facilitating communication among nodes within a KMS cluster.

          It solves the challenges associated with billions of cryptographic keys and hundreds of thousands of cryptographic operations per node per second - which are commonplace in larger deployments.

          <b>High Availability and Fault Tolerance</b>
          In the dynamic landscape of large-scale systems, network or node outages are inevitable. Taking down a cluster for maintenance is rarely feasible. MinIO's AIStor KMS ensures uninterrupted availability, even when faced with such disruptions, mitigating cascading effects that can take down the entire storage infrastructure. Specifically, you could lose all but one node of a cluster and still handle any encryption, decryption or data key generation requests.

          <b>Scalability</b>
          While the amount of data usually only increases, the load on a large-scale storage system may vary significantly from time to time. MinIO's AIStor KMS supports dynamic cluster resizing and nodes can be added or removed at any point without incurring any downtime.

          <b>Multi-Tenancy</b>
          Large-scale storage infrastructures are often used by many applications and teams across the entire organization. Isolating teams and groups into their own namespaces is a core requirement. MinIO's AIStor KMS supports namespacing in the form of enclaves. Each tenant can be assigned its own enclave which is completely independent and isolated from all other enclaves on the KMS cluster.

          <b>Predictable Behavior</b>
          MinIO's AIStor KMS is designed to be easily managed, providing operators with the ability to comprehend its state intuitively. Due to its simple design, MinIO's AIStor KMS is significantly easier to operate than similar solutions that rely on more complex consensus algorithms like Raft, or Paxos.

          <b>Consistent and Performant</b>
          The responsiveness of the KMS for GET/PUT operations directly influences the overall efficiency and speed of the storage system. MinIO's AIStor KMS nodes don’t have to coordinate when handling such requests from the storage system. Therefore, the performance of a MinIO's AIStor KMS cluster increases linearly with the number of nodes. Further, MinIO's AIStor KMS supports request pipelining to handle hundreds of thousands of cryptographic operations per node and second.

          <b>Simplicity</b>
          Operating a KMS cluster does not require expertise in cryptography or distributed systems. Everything can be done from the AIStor Console.

          <a href="https://resources.min.io/request-a-demo-marketplaces/">Request a demo today</a>
        displayName: Minio AIStor Key management System (KMS)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Storage
        - MinIO
        links:
        - name: Minio AIStor
          url: https://min.io
        - name: Support
          url: https://subnet.min.io
        - name: Github
          url: https://github.com/minio/aistor
        maintainers:
        - email: dev@min.io
          name: MinIO
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: MinIO
          url: https://min.io
        relatedImages:
        - quay.io/minio/aistor/operator@sha256:f80139efa062fdd83f78318b20f371a827b68e854b03dd9706b9eba4c493aeb2
        version: 2025.8.19175300
      entries:
      - name: minio-key-manager-operator.v2025.8.19175300
        version: 2025.8.19175300
      - name: minio-key-manager-operator.v2025.8.1184927
        version: 2025.8.1184927
      - name: minio-key-manager-operator.v2025.7.1011319
        version: 2025.7.1011319
      - name: minio-key-manager-operator.v2025.6.3065135
        version: 2025.6.3065135
      - name: minio-key-manager-operator.v2025.5.12190907
        version: 2025.5.12190907
      name: stable
    defaultChannel: stable
    packageName: minio-key-manager-operator
    provider:
      name: MinIO
      url: https://min.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator-certified.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a",
                  "imagePullPolicy": "Always",
                  "initImage": "registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:3c03fb18a5d73876576b80aca2b61e7c84ca052839ccb467d14e15f6d1f9d834
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a57926b53eef862b70a53818a7ccb0e9756f878f30d4128b608090b109bdc585
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator-certified.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator-certified.v1.20.0
        version: 1.20.0
      - name: percona-server-mongodb-operator-certified.v1.19.1
        version: 1.19.1
      - name: percona-server-mongodb-operator-certified.v1.19.0
        version: 1.19.0
      - name: percona-server-mongodb-operator-certified.v1.18.0
        version: 1.18.0
      - name: percona-server-mongodb-operator-certified.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator-certified
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Cloudera
      provider-url: https://cloudera.com
    name: cloudera-streams-messaging-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster"
                },
                "spec": {
                  "kafka": {
                    "version": "3.9.0.1.4",
                    "replicas": 3,
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2,
                      "inter.broker.protocol.version": "3.9"
                    },
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "zookeeper": {
                    "replicas": 3,
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "3.9.0.1.4",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
          createdAt: "2025-07-14 10:09:41"
          description: Run Apache Kafka with Cloudera Streams Messaging - Kubernetes
            Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: A valid license and current subscription
            for Cloudera Streaming - Kubernetes Operator is required. Contact Cloudera
            sales to obtain a license. https://www.cloudera.com/contact-sales.html
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |-
          Cloudera Streams Messaging - Kubernetes Operator brings enterprise-grade Kafka deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of Kafka for enterprise deployments. To learn more about Cloudera Streaming, [read this deep dive blog post](https://blog.cloudera.com/accelerating-deployments-of-streaming-pipelines-announcing-data-in-motion-on-kubernetes/).

          Cloudera Streams Messaging - Kubernetes Operator consists of and ships multiple components including Apache Kafka, Apache ZooKeeper, Cruise Control, Strimzi, and others.

          ## Installation prerequisites

          Installing Cloudera Streams Messaging - Kubernetes Operator requires a valid license as well as Cloudera credentials (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly. The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) where installation artifacts are pulled from.

          ### Obtaining a license

          Contact [Cloudera Sales](https://www.cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets

          Before installation, create  two `Secrets`. One containing your Cloudera credentials, and one containing your license. Cloudera credentials are provided to you together with your license.

          Create a `Secret` containing your Cloudera credentials.
          ```
          oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***] --docker-server=container.repository.cloudera.com
          ```
          Create a `Secret` containing your license.
          ```
          oc create secret generic csm-op-license --from-file=license.txt=[***PATH TO LICENSE FILE***]
          ```

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces where you plan to deploy Kafka or Kafka Connect clusters. Cloudera recommends that you create the `Secret` in all required namespaces before installation if you know what namespaces you will be using to deploy Kafka or Kafka Connect.

          ## Learn more
          To learn more, visit the [Cloudera Streams Messaging - Kubernetes Operator documentation library](https://docs.cloudera.com/csm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Streams Messaging - Kubernetes Operator, visit the following resources:

          * [Strimzi](https://strimzi.io/)
          * [Apache Kafka](https://kafka.apache.org/)
          * [Apache ZooKeeper](https://zookeeper.apache.org/)
          * [Cruise Control](https://github.com/linkedin/cruise-control)
        displayName: Cloudera Streams Messaging - Kubernetes Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Documentation
          url: https://docs.cloudera.com/csm-operator
        - name: Cloudera Community Forum
          url: https://community.cloudera.com/
        - name: Cloudera Stream Processing
          url: https://www.cloudera.com/products/stream-processing.html
        - name: Contact Cloudera
          url: https://www.cloudera.com/contact-sales.html
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
        - container.repository.cloudera.com/cloudera/kafka@sha256:aa64e161b9672153a6f6559f9ae9ef88d0e40d6040a80e21bd3dc8ee39df6523
        - container.repository.cloudera.com/cloudera/kafka@sha256:b675e7426a558c4681e53d8e03f278264d7ad555f7c02422017683a29708ac2b
        - container.repository.cloudera.com/cloudera/maven-builder@sha256:9f18203b175193e5079fdc0c2b71c14887641d997f92b4a1c918546f777166df
        version: 1.4.0-b199
      entries:
      - name: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
        version: 1.4.0-b199
      name: 1.2.x
    - currentCSV: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster"
                },
                "spec": {
                  "kafka": {
                    "version": "3.9.0.1.4",
                    "replicas": 3,
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2,
                      "inter.broker.protocol.version": "3.9"
                    },
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "zookeeper": {
                    "replicas": 3,
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "3.9.0.1.4",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
          createdAt: "2025-07-14 10:09:41"
          description: Run Apache Kafka with Cloudera Streams Messaging - Kubernetes
            Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: A valid license and current subscription
            for Cloudera Streaming - Kubernetes Operator is required. Contact Cloudera
            sales to obtain a license. https://www.cloudera.com/contact-sales.html
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |-
          Cloudera Streams Messaging - Kubernetes Operator brings enterprise-grade Kafka deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of Kafka for enterprise deployments. To learn more about Cloudera Streaming, [read this deep dive blog post](https://blog.cloudera.com/accelerating-deployments-of-streaming-pipelines-announcing-data-in-motion-on-kubernetes/).

          Cloudera Streams Messaging - Kubernetes Operator consists of and ships multiple components including Apache Kafka, Apache ZooKeeper, Cruise Control, Strimzi, and others.

          ## Installation prerequisites

          Installing Cloudera Streams Messaging - Kubernetes Operator requires a valid license as well as Cloudera credentials (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly. The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) where installation artifacts are pulled from.

          ### Obtaining a license

          Contact [Cloudera Sales](https://www.cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets

          Before installation, create  two `Secrets`. One containing your Cloudera credentials, and one containing your license. Cloudera credentials are provided to you together with your license.

          Create a `Secret` containing your Cloudera credentials.
          ```
          oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***] --docker-server=container.repository.cloudera.com
          ```
          Create a `Secret` containing your license.
          ```
          oc create secret generic csm-op-license --from-file=license.txt=[***PATH TO LICENSE FILE***]
          ```

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces where you plan to deploy Kafka or Kafka Connect clusters. Cloudera recommends that you create the `Secret` in all required namespaces before installation if you know what namespaces you will be using to deploy Kafka or Kafka Connect.

          ## Learn more
          To learn more, visit the [Cloudera Streams Messaging - Kubernetes Operator documentation library](https://docs.cloudera.com/csm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Streams Messaging - Kubernetes Operator, visit the following resources:

          * [Strimzi](https://strimzi.io/)
          * [Apache Kafka](https://kafka.apache.org/)
          * [Apache ZooKeeper](https://zookeeper.apache.org/)
          * [Cruise Control](https://github.com/linkedin/cruise-control)
        displayName: Cloudera Streams Messaging - Kubernetes Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Documentation
          url: https://docs.cloudera.com/csm-operator
        - name: Cloudera Community Forum
          url: https://community.cloudera.com/
        - name: Cloudera Stream Processing
          url: https://www.cloudera.com/products/stream-processing.html
        - name: Contact Cloudera
          url: https://www.cloudera.com/contact-sales.html
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/kafka@sha256:aa64e161b9672153a6f6559f9ae9ef88d0e40d6040a80e21bd3dc8ee39df6523
        - container.repository.cloudera.com/cloudera/kafka@sha256:b675e7426a558c4681e53d8e03f278264d7ad555f7c02422017683a29708ac2b
        - container.repository.cloudera.com/cloudera/maven-builder@sha256:9f18203b175193e5079fdc0c2b71c14887641d997f92b4a1c918546f777166df
        - container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
        version: 1.4.0-b199
      entries:
      - name: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
        version: 1.4.0-b199
      name: stable
    defaultChannel: stable
    packageName: cloudera-streams-messaging-kubernetes-operator
    provider:
      name: Cloudera
      url: https://cloudera.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: vertical-pod-autoscaler
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: verticalpodautoscaler.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "autoscaling.openshift.io/v1",
                "kind": "VerticalPodAutoscalerController",
                "metadata": {
                  "name": "default",
                  "namespace": "openshift-vertical-pod-autoscaler"
                },
                "spec": {
                  "podMinCPUMillicores": 25,
                  "podMinMemoryMb": 250,
                  "recommendationOnly": false,
                  "safetyMarginFraction": 0.15,
                  "minReplicas": 2
                }
              },
              {
                 "apiVersion": "autoscaling.k8s.io/v1",
                 "kind": "VerticalPodAutoscaler",
                 "metadata": {
                    "name": "myapp-vpa"
                 },
                 "spec": {
                    "targetRef": {
                       "apiVersion": "apps/v1",
                       "kind": "Deployment",
                       "name": "myapp-deployment"
                    },
                    "resourcePolicy": {
                       "containerPolicies": [
                          {
                             "containerName": "*",
                             "minAllowed": {
                                "cpu": "100m",
                                "memory": "50Mi"
                             },
                             "maxAllowed": {
                                "cpu": 1,
                                "memory": "500Mi"
                             },
                             "controlledResources": [
                                "cpu",
                                "memory"
                             ]
                          }
                       ]
                    }
                 }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          certifiedLevel: Primed
          containerImage: registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9-operator@sha256:edf76334695c9a4aee4a9c0297af40475a16dec7e35ec7113a35b8507724536e
          createdAt: 2019-10-14 12:17:13+00:00
          description: An operator to run the OpenShift Vertical Pod Autoscaler. Vertical
            Pod Autoscaler (VPA) can be configured to monitor a workload's resource
            utilization, and then adjust its CPU and memory limits by updating the
            pod (future) or restarting the pod with updated limits.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          healthIndex: B
          olm.skipRange: '>=4.5.0 <4.16.0-202509240123'
          operatorframework.io/suggested-namespace: openshift-vertical-pod-autoscaler
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/vertical-pod-autoscaler-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of the set of VPA controllers
            displayName: VPA Controller
            kind: VerticalPodAutoscalerController
            name: verticalpodautoscalercontrollers.autoscaling.openshift.io
            version: v1
          - description: An application whose resources are managed by the VPA
            displayName: VerticalPodAutoscaler
            kind: VerticalPodAutoscaler
            name: verticalpodautoscalers.autoscaling.k8s.io
            version: v1
          - description: Represents VPA checkpoints used by the VPA controllers to
              manage workloads
            displayName: VPA Checkpoint
            kind: VerticalPodAutoscalerCheckpoint
            name: verticalpodautoscalercheckpoints.autoscaling.k8s.io
            version: v1
        description: An operator to run the OpenShift Vertical Pod Autoscaler. Vertical
          Pod Autoscaler (VPA) can be configured to monitor a workload's resource
          utilization, and then adjust its CPU and memory limits by updating the pod
          (future) or restarting the pod with updated limits.
        displayName: VerticalPodAutoscaler
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Vertical Pod Autoscaler Documentation
          url: https://docs.openshift.com/container-platform/latest/nodes/pods/nodes-pods-vertical-autoscaler.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9-operator@sha256:edf76334695c9a4aee4a9c0297af40475a16dec7e35ec7113a35b8507724536e
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9@sha256:e6c8f34895184179bb87f1659c4ff30b96b131787fa7a64e7bc6b6adbb336c16
        version: 4.16.0-202509240123
      entries:
      - name: verticalpodautoscaler.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: verticalpodautoscaler.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: verticalpodautoscaler.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: verticalpodautoscaler.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: verticalpodautoscaler.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: verticalpodautoscaler.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: verticalpodautoscaler.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: verticalpodautoscaler.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: verticalpodautoscaler.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: verticalpodautoscaler.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: verticalpodautoscaler.v4.16.0-202501260334
        version: 4.16.0-202501260334
      - name: verticalpodautoscaler.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: verticalpodautoscaler.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: verticalpodautoscaler.v4.16.0-202412021735
        version: 4.16.0-202412021735
      - name: verticalpodautoscaler.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: verticalpodautoscaler.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: verticalpodautoscaler.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: verticalpodautoscaler.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: verticalpodautoscaler.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: verticalpodautoscaler.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: verticalpodautoscaler.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: verticalpodautoscaler.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: verticalpodautoscaler.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: verticalpodautoscaler.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: verticalpodautoscaler.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: verticalpodautoscaler.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: verticalpodautoscaler.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: verticalpodautoscaler.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: verticalpodautoscaler.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: verticalpodautoscaler.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: vertical-pod-autoscaler
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-businessautomation: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: businessautomation-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: businessautomation-operator.7.13.5-16
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.kiegroup.org/v2","kind":"KieApp","metadata":{"name":"rhpam-trial"},"spec":{"environment":"rhpam-trial"}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/rhpam-7/rhpam-rhel8-operator@sha256:3b4f420def9e5d2c768d5209264bd5f9af6d6a37b22ef9a988a167ff7d9959cf
          createdAt: "2025-07-30 19:54:27"
          description: Deploys and manages Red Hat Process Automation Manager and
            Red Hat Decision Manager environments.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Process Automation
            Manager", "Red Hat Decision Manager"]'
          repository: https://github.com/kiegroup/kie-cloud-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A project prescription running an RHPAM/RHDM environment.
            displayName: KieApp
            kind: KieApp
            name: kieapps.app.kiegroup.org
            version: v2
        description: |-
          Deploys and manages Red Hat Process Automation Manager and Red Hat Decision Manager environments.

          * **Red Hat Process Automation Manager** is a platform for developing containerized microservices and applications that automate business decisions and processes. It includes business process management (BPM), business rules management (BRM), and business resource optimization and complex event processing (CEP) technologies. It also includes a user experience platform to create engaging user interfaces for process and decision services with minimal coding.

          [See more](https://www.redhat.com/en/products/process-automation).
        displayName: Business Automation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kieapp
        - pam
        - decision
        - kie
        - cloud
        - bpm
        - process
        - automation
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-process-automation-manager
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_process_automation_manager/7.13/#category-deploying-red-hat-process-automation-manager-on-openshift
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-monitoring-rhel8@sha256:8d4481a11269565f1690ad007f152e24326bae4b2099cb26ddf7b21be67031e8
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:cb37360843e9b062c59755593d9d412e1bf2c1dc1d6902e8a65a2825812c0989
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhpam-7/rhpam-kieserver-rhel8@sha256:2526d12da63b67ae9ed9294c914e9406d833c090ffa777623303ea0be22755d8
        - registry.redhat.io/rhpam-7/rhpam-process-migration-rhel8@sha256:3d7bdbe38895e50549e97573bd40da47b11691932164db7dafff0155a051aab2
        - registry.redhat.io/rhpam-7/rhpam-controller-rhel8@sha256:3a901b54a9d35d38b5aca275df2160e3190a874e139fcb606b9969b38284b3dc
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        - registry.redhat.io/rhpam-7/rhpam-kieserver-rhel8@sha256:3fbef82d1c5da8da1606d1a2d5dcfd20000da574737eb32b772d4f0123b179cb
        - registry.redhat.io/rhpam-7/rhpam-smartrouter-rhel8@sha256:d1e5461db8983a13f98259152dc3cfa2aa677c29f799104190e5cefa83cf97b5
        - registry.redhat.io/rhpam-7/rhpam-dashbuilder-rhel8@sha256:46396dc8060f2f9cc6f87ad905ef55be5b8e5216e5162bdf54df98c9d70e29ef
        - registry.redhat.io/rhscl/mysql-80-rhel7@sha256:e4f535c78bc2fdb58914c52fb805185035ff31515452b4df714f9ab623ba383a
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-rhel8@sha256:228c6ab5b83e215a98f6eb76250feb2ae3096fd21398c44e5b2701257c9c3493
        - registry.redhat.io/rhpam-7/rhpam-smartrouter-rhel8@sha256:9d7efdd97f163770d8a84b1ce5236bd4f31508154dced1f11dc6efe75a7b5089
        - registry.redhat.io/rhscl/postgresql-13-rhel7@sha256:338dd09681dcf860e6a976a3bc48a1474c2376d698815e4eb611ae9c5aa68966
        - registry.redhat.io/rhpam-7/rhpam-rhel8-operator@sha256:3b4f420def9e5d2c768d5209264bd5f9af6d6a37b22ef9a988a167ff7d9959cf
        - registry.redhat.io/rhpam-7/rhpam-dashbuilder-rhel8@sha256:4edac282119898ab1395030eecaee584990064e88fc8e65fb999f19ea2ef3f8b
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-rhel8@sha256:1abbaba9d4b2ecaec38205bcccf5ef4c7b6388ed9725b4f4c6dc8dc664588eb1
        - registry.redhat.io/rhpam-7/rhpam-controller-rhel8@sha256:989dc6dc64a2d5b6723fb5c000b0411fbdc6f78c0056666d831ffeabeab4b612
        - registry.redhat.io/openshift4/ose-cli@sha256:c3eb8807caf73e30012ae53c6d4f5114471290a5596929c27acef5a699a5b813
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/rhpam-7/rhpam-process-migration-rhel8@sha256:bd5c6586b4859dec23ad9530ee98bab33497ab251d544f863fc0af01cbd45101
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:457fc73423fa94591bc7e35729a25c58d3465583afccf342fb338b6e11188575
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-monitoring-rhel8@sha256:32912270eba0f2464da88f315bde8bf1ef5bbded814db156811d7a415a458f86
        version: 7.13.5-16
      entries:
      - name: businessautomation-operator.7.13.5-16
        version: 7.13.5-16
      - name: businessautomation-operator.7.13.5-15
        version: 7.13.5-15
      - name: businessautomation-operator.7.13.5-14
        version: 7.13.5-14
      - name: businessautomation-operator.7.13.5-13
        version: 7.13.5-13
      - name: businessautomation-operator.7.13.5-12
        version: 7.13.5-12
      - name: businessautomation-operator.7.13.5-11
        version: 7.13.5-11
      - name: businessautomation-operator.7.13.5-10
        version: 7.13.5-10
      - name: businessautomation-operator.7.13.5-9
        version: 7.13.5-9
      - name: businessautomation-operator.7.13.5-8
        version: 7.13.5-8
      - name: businessautomation-operator.7.13.5-7
        version: 7.13.5-7
      - name: businessautomation-operator.7.13.5-6
        version: 7.13.5-6
      - name: businessautomation-operator.7.13.5-5
        version: 7.13.5-5
      - name: businessautomation-operator.7.13.5-4
        version: 7.13.5-4
      - name: businessautomation-operator.7.13.5-3
        version: 7.13.5-3
      - name: businessautomation-operator.7.13.5-2
        version: 7.13.5-2
      - name: businessautomation-operator.7.13.5-1
        version: 7.13.5-1
      - name: businessautomation-operator.7.13.4-4
        version: 7.13.4-4
      - name: businessautomation-operator.7.13.4-3
        version: 7.13.4-3
      - name: businessautomation-operator.7.13.4-2
        version: 7.13.4-2
      - name: businessautomation-operator.7.13.4-1
        version: 7.13.4-1
      - name: businessautomation-operator.7.13.3-5
        version: 7.13.3-5
      - name: businessautomation-operator.7.13.3-4
        version: 7.13.3-4
      - name: businessautomation-operator.7.13.3-3
        version: 7.13.3-3
      - name: businessautomation-operator.7.13.3-2
        version: 7.13.3-2
      - name: businessautomation-operator.7.13.3-1
        version: 7.13.3-1
      - name: businessautomation-operator.7.13.2-3
        version: 7.13.2-3
      - name: businessautomation-operator.7.13.2-2
        version: 7.13.2-2
      - name: businessautomation-operator.7.13.2-1
        version: 7.13.2-1
      - name: businessautomation-operator.7.13.1-2
        version: 7.13.1-2
      - name: businessautomation-operator.7.13.1-1
        version: 7.13.1-1
      - name: businessautomation-operator.7.13.0-1
        version: 7.13.0-1
      - name: businessautomation-operator.7.12.1-4
        version: 7.12.1-4
      - name: businessautomation-operator.7.12.1-3
        version: 7.12.1-3
      - name: businessautomation-operator.7.12.1-2
        version: 7.12.1-2
      - name: businessautomation-operator.7.12.1-1
        version: 7.12.1-1
      - name: businessautomation-operator.7.12.0-2
        version: 7.12.0-2
      - name: businessautomation-operator.7.12.0-1
        version: 7.12.0-1
      - name: businessautomation-operator.7.11.1-3
        version: 7.11.1-3
      - name: businessautomation-operator.7.11.1-2
        version: 7.11.1-2
      - name: businessautomation-operator.7.11.1-1
        version: 7.11.1-1
      - name: businessautomation-operator.7.11.0-2
        version: 7.11.0-2
      - name: businessautomation-operator.7.11.0-1
        version: 7.11.0-1
      - name: businessautomation-operator.7.10.1-2
        version: 7.10.1-2
      - name: businessautomation-operator.7.10.1-1
        version: 7.10.1-1
      - name: businessautomation-operator.7.10.0-2
        version: 7.10.0-2
      - name: businessautomation-operator.7.10.0-1
        version: 7.10.0-1
      - name: businessautomation-operator.7.9.1-1
        version: 7.9.1-1
      - name: businessautomation-operator.7.9.0
        version: 7.9.0
      - name: businessautomation-operator.7.9.0-2
        version: 7.9.0-2
      - name: businessautomation-operator.7.8.1
        version: 7.8.1
      - name: businessautomation-operator.7.8.0
        version: 7.8.0
      - name: businessautomation-operator.1.4.1
        version: 1.4.1
      - name: businessautomation-operator.1.4.0
        version: 1.4.0
      - name: businessautomation-operator.1.3.0
        version: 1.3.0
      - name: businessautomation-operator.1.2.1
        version: 1.2.1
      - name: businessautomation-operator.1.2.0
        version: 1.2.0
      name: stable
    defaultChannel: stable
    packageName: businessautomation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: quay-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:d73321dfa6dcced943405fae45c2ccf841663e0ab57e8be8cc1ddba251b7ce08
          createdAt: 2021-11-02 11:11 UTC
          description: Opinionated deployment of Red Hat Quay on Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:d73321dfa6dcced943405fae45c2ccf841663e0ab57e8be8cc1ddba251b7ce08
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:0e6bef51caf479074925115fa11621c41bcada375cc1925d0511a73084b93543
        - registry.redhat.io/rhel8/postgresql-10@sha256:b2695e9af379624621bef9ec785f0cb618bfdabae98c5f17eea82ff5f8d2d5cb
        - registry.redhat.io/rhel8/redis-5@sha256:8d7e66927ce4b3050ca15ab4df4dbaeb593cb855e16fb0a6f5dde950e9db7db9
        - registry.redhat.io/quay/quay-rhel8@sha256:ed8c4249d376b4226e35bf088ffe5bf688bf68b059718c5e218aa1184fe048d0
        - registry.redhat.io/quay/clair-rhel8@sha256:af5e2fd1543365b365b1a6cce00b86e44fe5e66cde6491e9b113324c21f629c8
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:2a471300f70d3995ab7e9fcc3df35c5513e73fcc1d140fb4d651a1c3eae949eb
        version: 3.4.7
      entries:
      - name: quay-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: quay-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:1f5e4cc9ec2d9b8b89d82f1a698acfa1ac2764ca4ebc5c85b3417da7730b5867
          createdAt: 2021-09-09 20:09 UTC
          description: Opinionated deployment of Red Hat Quay on Kubernetes.
          olm.skipRange: '>=3.4.x <3.5.7'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.5.7
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7fe408a1e428b2d0660cb578c1dcd4b961f838e37166fe7b21b2bed70d288594
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:5d72577ccfd63c6d9ad712b376b604c40d2d3ecb53e391e2ee0d18caa6e0225a
        - registry.redhat.io/rhel8/postgresql-10@sha256:a723bb3a7a5046527b9a5b628ddfb211c526e3814cc4242262c4f8a016f38edd
        - registry.redhat.io/rhel8/redis-5@sha256:7916760ff03af5ca230df5e96b2bf353fbf1829386de1c063c05f623837be7f7
        - registry.redhat.io/quay/quay-rhel8@sha256:c599892160ac20c744d833603d9375923af7aec1edfd86982b7513e02fb6d463
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:1f5e4cc9ec2d9b8b89d82f1a698acfa1ac2764ca4ebc5c85b3417da7730b5867
        - registry.redhat.io/quay/clair-rhel8@sha256:d242bd9d37f95b2a5a8267b3f1eba02c671d46121712baae43466c86d1da89f9
        version: 3.5.7
      entries:
      - name: quay-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: quay-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:3c0201e3cf88e711c921870324002a02cbd42c0d2eecafb4f242aa0601e24602
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.10.15
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:fc20b8bdac88c1bfce9aec993bdce6dd2cc132f8a9d4062d174b098819c74013
        - registry.redhat.io/quay/clair-rhel8@sha256:efa6e69500ea643042b544e6205f088214d79ab09a36864a612e73d07979d1c5
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:3c0201e3cf88e711c921870324002a02cbd42c0d2eecafb4f242aa0601e24602
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7e720c6a332ec4032e8576c31421af404f61d9baaf8e88f69f8333c52e5d1683
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:774cd1f181527787ba8631c7c14437eb2250c15ddb4a9c75fdfa35c5c1a19a2e
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        version: 3.10.15
      entries:
      - name: quay-operator.v3.10.15
        version: 3.10.15
      - name: quay-operator.v3.10.14
        version: 3.10.14
      - name: quay-operator.v3.10.13
        version: 3.10.13
      - name: quay-operator.v3.10.12
        version: 3.10.12
      - name: quay-operator.v3.10.11
        version: 3.10.11
      - name: quay-operator.v3.10.10
        version: 3.10.10
      - name: quay-operator.v3.10.9
        version: 3.10.9
      - name: quay-operator.v3.10.8
        version: 3.10.8
      - name: quay-operator.v3.10.7
        version: 3.10.7
      - name: quay-operator.v3.10.6
        version: 3.10.6
      - name: quay-operator.v3.10.5
        version: 3.10.5
      - name: quay-operator.v3.10.4
        version: 3.10.4
      - name: quay-operator.v3.10.3
        version: 3.10.3
      - name: quay-operator.v3.10.2
        version: 3.10.2
      - name: quay-operator.v3.10.1
        version: 3.10.1
      - name: quay-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:cb3a41caa3a572b7981d424be2a32b8e881127cc03b557ee0a36ab13f430b897
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.11.13
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:7230fa1d4970e877cb6de8716e99a9bd43302c0ff23b4501f2aa806a8d496d9b
        - registry.redhat.io/quay/clair-rhel8@sha256:3746677ef1cb0ddd4f237ab598b14cf0bc76dcce080ffde127b5bba87ffe6693
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:1a083cd72502609427428ed22fb60846613ddb41a4613496c5cddc797b382d3a
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:11b8c9d64c42f0f0c7b2df491bba859d0410666a8ef1d2fd41eb77e4ad9d2dc8
        - registry.redhat.io/rhel8/postgresql-13@sha256:35d86936c539007a78a89869a14b87b76788fe117df442a670f63936b978ed5b
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:cb3a41caa3a572b7981d424be2a32b8e881127cc03b557ee0a36ab13f430b897
        version: 3.11.13
      entries:
      - name: quay-operator.v3.11.13
        version: 3.11.13
      - name: quay-operator.v3.11.12
        version: 3.11.12
      - name: quay-operator.v3.11.11
        version: 3.11.11
      - name: quay-operator.v3.11.10
        version: 3.11.10
      - name: quay-operator.v3.11.9
        version: 3.11.9
      - name: quay-operator.v3.11.8
        version: 3.11.8
      - name: quay-operator.v3.11.7
        version: 3.11.7
      - name: quay-operator.v3.11.6
        version: 3.11.6
      - name: quay-operator.v3.11.5
        version: 3.11.5
      - name: quay-operator.v3.11.4
        version: 3.11.4
      - name: quay-operator.v3.11.3
        version: 3.11.3
      - name: quay-operator.v3.11.2
        version: 3.11.2
      - name: quay-operator.v3.11.1
        version: 3.11.1
      - name: quay-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:accb705b6327fb2dcc1ce657a99f61e343eca752c9873d63d89fea89c10c3187
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.12.12
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:accb705b6327fb2dcc1ce657a99f61e343eca752c9873d63d89fea89c10c3187
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:ccbd83e05aea0bb96bf099f63420175b36ed413bfc8a618f4c90b6a2abc53895
        - registry.redhat.io/quay/clair-rhel8@sha256:dc3d5fb35f2bd5d00480351dca705fe260a225d9a97958cf0276801af08d8307
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:8e5dd171e6e94de0137bb6a972533b1f93b56efc1731e19b80869d5e9b8a26a6
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:8cdd677449f3c723dad79bcd9c4191c8917d6eb0beeeeeda68fe19a0988bb2a8
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        version: 3.12.12
      entries:
      - name: quay-operator.v3.12.12
        version: 3.12.12
      - name: quay-operator.v3.12.11
        version: 3.12.11
      - name: quay-operator.v3.12.10
        version: 3.12.10
      - name: quay-operator.v3.12.9
        version: 3.12.9
      - name: quay-operator.v3.12.8
        version: 3.12.8
      - name: quay-operator.v3.12.7
        version: 3.12.7
      - name: quay-operator.v3.12.6
        version: 3.12.6
      - name: quay-operator.v3.12.5
        version: 3.12.5
      - name: quay-operator.v3.12.4
        version: 3.12.4
      - name: quay-operator.v3.12.3
        version: 3.12.3
      - name: quay-operator.v3.12.2
        version: 3.12.2
      - name: quay-operator.v3.12.1
        version: 3.12.1
      - name: quay-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:5481a4b914ce4aed6ba3ceee8432eddab6e264cad2caef528404795f904100eb
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.13.8
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/clair-rhel8@sha256:069ccf36346cb14f4e4ea54de1d59afb72548912b140cfef67a6bcb7534d25e9
        - registry.redhat.io/rhel8/postgresql-13@sha256:f17fd8dd4d56d8e916cb2ad3f403d05ad259f79fd6918d67a38a6e9e4a8896cf
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:966a3dafa2dd31ac1a39e826e08c1b988fe694189fe79449bbe14a48b2631213
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:5481a4b914ce4aed6ba3ceee8432eddab6e264cad2caef528404795f904100eb
        - registry.redhat.io/quay/quay-rhel8@sha256:8887c17547c7a4be61471523eade9820e5095cf966b1885a3758081da559871c
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:31a9095d8f291ce68e8e688877692e74a17143bdda8d998fe3a07bcee32575dd
        - registry.redhat.io/rhel8/postgresql-15@sha256:a5f36e7893a519c12752c6118233c92d4e83618670bc9ea7bd3a8a34a63bd50b
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:f62cbd6632454c227b7cbb01dadc8e501e8288b2d8e0abf8e7842b806fa3195c
        version: 3.13.8
      entries:
      - name: quay-operator.v3.13.8
        version: 3.13.8
      - name: quay-operator.v3.13.7
        version: 3.13.7
      - name: quay-operator.v3.13.6
        version: 3.13.6
      - name: quay-operator.v3.13.5
        version: 3.13.5
      - name: quay-operator.v3.13.4
        version: 3.13.4
      - name: quay-operator.v3.13.3
        version: 3.13.3
      - name: quay-operator.v3.13.2
        version: 3.13.2
      - name: quay-operator.v3.13.1
        version: 3.13.1
      - name: quay-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:8ebc1f0511f7d394ceaa39a4c993e955c560dc04588072cb3f8c8fab8c85e32b
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.14.5
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/clair-rhel8@sha256:e11e8a819d17bd14c67f093b5303b4958855e731f89ce67dcca70fcb7b77f45a
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:8ebc1f0511f7d394ceaa39a4c993e955c560dc04588072cb3f8c8fab8c85e32b
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-rhel8@sha256:33e984892650d315065313fcb5f8808e145aadec5a287219c7031d85b713235e
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:43ef8adddbf9b007b6e2396f53bd6a5475150d28c2f4254f031a0fe6c7de5826
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:a12a0918f6dea114b0aae231d7119a4acfd0cd297074d895f4ecf190d2aec43a
        version: 3.14.5
      entries:
      - name: quay-operator.v3.14.5
        version: 3.14.5
      - name: quay-operator.v3.14.4
        version: 3.14.4
      - name: quay-operator.v3.14.3
        version: 3.14.3
      - name: quay-operator.v3.14.2
        version: 3.14.2
      - name: quay-operator.v3.14.1
        version: 3.14.1
      - name: quay-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:05fb981d9848281b97b3683093f752b1c1dc5ced439a2b4ba18b26ff8e11231e
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.15.2
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:873a3daf1dadf2438eb96e151072326c31ba34ed48cac041614b604413d5f8e7
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:956f3110f944d880bc27bb6df9472d0804cd75f5dcf4ed11bcf721e9d1d4542b
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:aa8c9f7b6ce57441ca1fe9dbaaefaf2bb83fd44aa367647b663bb822bf890362
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:05fb981d9848281b97b3683093f752b1c1dc5ced439a2b4ba18b26ff8e11231e
        - registry.redhat.io/quay/clair-rhel8@sha256:861be31d30142bae125cd059bb716d79b28694e9c4336333c68adf4a3823b61d
        version: 3.15.2
      entries:
      - name: quay-operator.v3.15.2
        version: 3.15.2
      - name: quay-operator.v3.15.1
        version: 3.15.1
      - name: quay-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-operator.v3.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:7a6e592d7f5f760b4451f9c974dbb96f56b1d3de92f039adcbf284f78ba13a02
          createdAt: 2021-10-12 22:10 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.6.10
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:ec36241f0a9a2983b1894cd8501898e6b70379af51a81b1e09b0ba0e9b54b33f
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:75df6b8e6bef14844a08f7d994470fa6ec4ffa66d4b15faaa0a3ba373f612fbb
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:7a6e592d7f5f760b4451f9c974dbb96f56b1d3de92f039adcbf284f78ba13a02
        - registry.redhat.io/rhel8/postgresql-10@sha256:4b912c80085b88a03309aeb7907efcc29dd3342fa3952b6ea067afb1914bfe53
        - registry.redhat.io/rhel8/redis-6@sha256:d7c7852338717308cbb59e9303e1ea35cc8e5c01ceb2818569be20c15f3f943d
        - registry.redhat.io/quay/quay-rhel8@sha256:e75e4d084eba9bb65b9cc72e38294af314d49316b12a0beb75db808dc057830b
        - registry.redhat.io/quay/clair-rhel8@sha256:cbafddd1eb50c1c0f9159ffafd86f44a8c0b976194903c96c6f144bfe62e749d
        version: 3.6.10
      entries:
      - name: quay-operator.v3.6.10
        version: 3.6.10
      - name: quay-operator.v3.6.9
        version: 3.6.9
      - name: quay-operator.v3.6.8
        version: 3.6.8
      - name: quay-operator.v3.6.7
        version: 3.6.7
      - name: quay-operator.v3.6.6
        version: 3.6.6
      - name: quay-operator.v3.6.5
        version: 3.6.5
      - name: quay-operator.v3.6.4
        version: 3.6.4
      - name: quay-operator.v3.6.2
        version: 3.6.2
      - name: quay-operator.v3.6.1
        version: 3.6.1
      - name: quay-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: quay-operator.v3.7.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:a8283c2be661e07c5f2f2e5b683c520846bcb91d580d9a3a0aa27ce994ddfff6
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.7.14
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:b15335d88b240bd6bbedeb7fe26f0a1de2f6c330af9758b0fb2edab7ca69f6df
        - registry.redhat.io/quay/quay-rhel8@sha256:7bc9e115e9bace07505ffd225242cab83d5eb6d50358188d18fd9161bdf1a5a6
        - registry.redhat.io/quay/clair-rhel8@sha256:d676af83e96740220e6f18721c98817ba071fd023a62ebdc3815424e94838083
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:c54c77e09c95dc7d80d7b31e796857c8d06f2f733426629b53d3298206d23b20
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:04ac405c152bffb43b33fb2e7f45df8b3940eb21e055ebf6aa17bd81ebdeeab4
        - registry.redhat.io/rhel8/postgresql-10@sha256:a52ad402458ec8ef3f275972c6ebed05ad64398f884404b9bb8e3010c5c95291
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:a8283c2be661e07c5f2f2e5b683c520846bcb91d580d9a3a0aa27ce994ddfff6
        version: 3.7.14
      entries:
      - name: quay-operator.v3.7.14
        version: 3.7.14
      - name: quay-operator.v3.7.13
        version: 3.7.13
      - name: quay-operator.v3.7.12
        version: 3.7.12
      - name: quay-operator.v3.7.11
        version: 3.7.11
      - name: quay-operator.v3.7.10
        version: 3.7.10
      name: stable-3.7
    - currentCSV: quay-operator.v3.8.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:abec0fa7ee641c8e424e995b5fe07b98e522ad1eb92c3b90426f69941f33eb2e
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.8.15
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7e4e57071439d671971b0de46b597cf6cff8ef3879ab711125a6ae7ecb9c6d29
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:3c047c7b22fb42557bd3a9b9c9ac333f0df7d5e4a4d333a2b02a4313f57085be
        - registry.redhat.io/rhel8/postgresql-10@sha256:b7fa644277d5e406b642685ff2ace1b1de7aea2e2a1421b3cbcda7d60405f44d
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:abec0fa7ee641c8e424e995b5fe07b98e522ad1eb92c3b90426f69941f33eb2e
        - registry.redhat.io/rhel8/redis-6@sha256:444200241cb9186cd97d564d7acf077d83fc997a8ee5ee7fdae41b604895085c
        - registry.redhat.io/quay/quay-rhel8@sha256:de2b65506f99eebb5eaafca6bb42b4ebec08e9479dd0d50aa1286f197e42d6de
        - registry.redhat.io/quay/clair-rhel8@sha256:b327fcfd7472cdb9d83a564904976b26b285fa576dc0a76e5590990ec77fd871
        version: 3.8.15
      entries:
      - name: quay-operator.v3.8.15
        version: 3.8.15
      - name: quay-operator.v3.8.14
        version: 3.8.14
      - name: quay-operator.v3.8.13
        version: 3.8.13
      - name: quay-operator.v3.8.12
        version: 3.8.12
      - name: quay-operator.v3.8.11
        version: 3.8.11
      - name: quay-operator.v3.8.10
        version: 3.8.10
      - name: quay-operator.v3.8.9
        version: 3.8.9
      - name: quay-operator.v3.8.8
        version: 3.8.8
      - name: quay-operator.v3.8.7
        version: 3.8.7
      - name: quay-operator.v3.8.6
        version: 3.8.6
      - name: quay-operator.v3.8.5
        version: 3.8.5
      - name: quay-operator.v3.8.4
        version: 3.8.4
      - name: quay-operator.v3.8.3
        version: 3.8.3
      - name: quay-operator.v3.8.2
        version: 3.8.2
      - name: quay-operator.v3.8.1
        version: 3.8.1
      - name: quay-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:8a16803946272156b7aeb00eb2202ab6cb87d875fe4a7c327554f770f0f2ecb5
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.9.16
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/clair-rhel8@sha256:6c67163c783d7d9fa4d187963379885a8b76685347ecf1894c4f8ebd6656544b
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:af8ada0c88b6f19991e61ee86d04b49a148c19dacb57c282e4ab98630e7499b4
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:5a42689198701d845ff3f05502cbbebceb7e6ad7b12762ccc038c0131cf6479e
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:f87b7e7ee6c31f6ae80b08aef0e10729b21fe1b8656fe18f6dee2310bae28ec1
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:8a16803946272156b7aeb00eb2202ab6cb87d875fe4a7c327554f770f0f2ecb5
        version: 3.9.16
      entries:
      - name: quay-operator.v3.9.16
        version: 3.9.16
      - name: quay-operator.v3.9.15
        version: 3.9.15
      - name: quay-operator.v3.9.14
        version: 3.9.14
      - name: quay-operator.v3.9.13
        version: 3.9.13
      - name: quay-operator.v3.9.12
        version: 3.9.12
      - name: quay-operator.v3.9.11
        version: 3.9.11
      - name: quay-operator.v3.9.10
        version: 3.9.10
      - name: quay-operator.v3.9.9
        version: 3.9.9
      - name: quay-operator.v3.9.8
        version: 3.9.8
      - name: quay-operator.v3.9.7
        version: 3.9.7
      - name: quay-operator.v3.9.6
        version: 3.9.6
      - name: quay-operator.v3.9.5
        version: 3.9.5
      - name: quay-operator.v3.9.4
        version: 3.9.4
      - name: quay-operator.v3.9.3
        version: 3.9.3
      - name: quay-operator.v3.9.2
        version: 3.9.2
      - name: quay-operator.v3.9.1
        version: 3.9.1
      - name: quay-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: quay-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: runtime-component-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runtime-component.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/runtime-component-operator@sha256:d37b3f62575c9ad1ffcb22c89e2c3886275c326bf27aa49fb1808acba475853d
          createdAt: "2022-03-25T14:00:00Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          olm.skipRange: <0.8.1
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          This advanced Operator is capable of deploying any runtime component image with consistent, production-grade QoS. It enables enterprise architects to govern the way their applications get deployed & managed in the cluster, while dramatically reducing the learning curve for developers to deploy into Kubernetes - allowing them to focus on writing the code!
          Here are some key features:

          #### Application Lifecyle
          You can deploy your runtime component container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the Operator will watch for any updates and will re-deploy the modified image.

          #### Custom RBAC
          This Operator is capable of using a custom ServiceAccount from the caller, allowing it to follow RBAC restrictions. By default it creates a ServiceAccount if one is not specified, which can also be bound with specific roles.

          #### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as: labels, annotations, and environment variables from a ConfigMap, a Secret or a value.

          #### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configuration, such as TLS settings, are also easily enabled.

          #### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto scaling to create (and delete) instances based on resource consumption.

          #### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you.
          This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart.
          We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.

          #### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating kubernetes Secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to its required services without any intervention or interruption.

          #### Exposing metrics to Prometheus
          The Runtime Component Operator exposes the runtime container's metrics via the Prometheus Operator.
          You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `spec.monitoring` field to control configurations such as the poll interval and security credentials.

          #### Easily mount logs and transaction directories
          If you need to mount the logs and transaction data from your runtime component to an external volume such as NFS (or any storage supported in your cluster), simply add the following (to specify the volume size and the location to persist) to your RuntimeComponent CR:
          ``` storage: size: 2Gi mountPath: "/logs" ```

          #### Integration with OpenShift Serverless
          Deploy your serverless runtime component using a single toggle.  The Operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing your pod to automatically scale to 0 when it is idle.

          #### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and how they are connected.

          See our [**documentation**](https://github.com/application-stacks/runtime-component-operator/tree/main/doc/) for more information.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://github.com/application-stacks/runtime-component-operator/tree/main/doc
        - name: Repository
          url: https://github.com/application-stacks/runtime-component-operator
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: beta
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/runtime-component-operator@sha256:d37b3f62575c9ad1ffcb22c89e2c3886275c326bf27aa49fb1808acba475853d
        - registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4
        version: 0.8.1
      entries:
      - name: runtime-component.v0.8.1
        version: 0.8.1
      - name: runtime-component.v0.8.0
        version: 0.8.0
      name: beta2
    - currentCSV: runtime-component.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:1b818f1a9c27baab6a39ae42d86f78bac085d772cae5cd97b357d0ee5799cf85
          createdAt: "2023-08-28T16:46:50Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          olm.skipRange: '>=0.8.0 <1.2.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02
        - icr.io/appcafe/runtime-component-operator@sha256:1b818f1a9c27baab6a39ae42d86f78bac085d772cae5cd97b357d0ee5799cf85
        version: 1.2.2
      entries:
      - name: runtime-component.v1.2.2
        version: 1.2.2
      name: v1.2
    - currentCSV: runtime-component.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:682c5cb3e24566bf2933b1d4a4f6bf520d9b546edcedc1ee3214ec1ac4d48629
          createdAt: "2024-08-01T20:26:42Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.3.3'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/runtime-component-operator@sha256:682c5cb3e24566bf2933b1d4a4f6bf520d9b546edcedc1ee3214ec1ac4d48629
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91
        version: 1.3.3
      entries:
      - name: runtime-component.v1.3.3
        version: 1.3.3
      name: v1.3
    - currentCSV: runtime-component.v1.4.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:6f527c5173421c28d175b42d8c690eb63c4d1c5092c31d5350fd7a5e93661eaa
          createdAt: "2025-07-18T20:38:10Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.4.4'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/runtime-component-operator@sha256:6f527c5173421c28d175b42d8c690eb63c4d1c5092c31d5350fd7a5e93661eaa
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3
        version: 1.4.4
      entries:
      - name: runtime-component.v1.4.4
        version: 1.4.4
      name: v1.4
    - currentCSV: runtime-component.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:58bfceaed2fc828bce1c0e4e21a583d25f6d72f88abdc78df2510867a4f2020f
          createdAt: "2025-09-15T18:08:27Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.5.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8
        - icr.io/appcafe/runtime-component-operator@sha256:58bfceaed2fc828bce1c0e4e21a583d25f6d72f88abdc78df2510867a4f2020f
        version: 1.5.0
      entries:
      - name: runtime-component.v1.5.0
        version: 1.5.0
      name: v1.5
    defaultChannel: v1.5
    packageName: runtime-component-operator-certified
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: glance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: glance-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "glance.openstack.org/v1beta1",
                "kind": "Glance",
                "metadata": {
                  "name": "glance"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\nenabled_backends=foo:bar,foo1:bar1\n",
                  "databaseAccount": "glance",
                  "databaseInstance": "openstack",
                  "secret": "osp-secret",
                  "serviceUser": "glance",
                  "storage": {
                    "storageClass": "local-storage",
                    "storageRequest": "10G"
                  }
                }
              },
              {
                "apiVersion": "glance.openstack.org/v1beta1",
                "kind": "GlanceAPI",
                "metadata": {
                  "annotation": {
                    "keystoneEndpoint": "true"
                  },
                  "name": "default"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\nenabled_backends=foo:bar,foo1:bar1\n[foo]\nhere-foo-config\n[foo1]\nhere-foo1-config\n",
                  "databaseAccount": "glance",
                  "databaseHostname": "glance",
                  "preserveJobs": false,
                  "replicas": 1,
                  "secret": "glance-secret",
                  "serviceAccount": "glance",
                  "serviceUser": "glance",
                  "storage": {
                    "storageRequest": "10G"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:45Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GlanceAPI is the Schema for the glanceapis API
            displayName: Glance API
            kind: GlanceAPI
            name: glanceapis.glance.openstack.org
            version: v1beta1
          - description: Glance is the Schema for the glances API
            displayName: Glance
            kind: Glance
            name: glances.glance.openstack.org
            version: v1beta1
        description: Glance Operator
        displayName: Glance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Image
        - Glance
        links:
        - name: Glance Operator
          url: https://github.com/openstack-k8s-operators/glance-operator
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/glance-rhel9-operator@sha256:74f9907b59493239fbe9c5f534399ff513c693a2bda11a037aa2cc5dc8cc3e91
        - registry.redhat.io/rhoso/openstack-glance-api-rhel9@sha256:b8f84f9b432a70f7ca7aebda60455a4890347529f6bed9a440a3b197e377e0a1
        version: 1.0.6
      entries:
      - name: glance-operator.v1.0.6
        version: 1.0.6
      - name: glance-operator.v1.0.5
        version: 1.0.5
      - name: glance-operator.v1.0.4
        version: 1.0.4
      - name: glance-operator.v1.0.3
        version: 1.0.3
      - name: glance-operator.v1.0.2
        version: 1.0.2
      - name: glance-operator.v1.0.1
        version: 1.0.1
      - name: glance-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: glance-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kafka-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kafka-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kafka-controller:1.2.1
          createdAt: "2025-09-20T21:41:07Z"
          description: AWS Kafka controller is a service controller for managing Kafka
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster represents the state of an AWS kafka Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.kafka.services.k8s.aws
            version: v1alpha1
          - description: Configuration represents the state of an AWS kafka Configuration
              resource.
            displayName: Configuration
            kind: Configuration
            name: configurations.kafka.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCluster represents the state of an AWS kafka ServerlessCluster
              resource.
            displayName: ServerlessCluster
            kind: ServerlessCluster
            name: serverlessclusters.kafka.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Kafka resources in AWS from within your Kubernetes cluster.

          **About Amazon Kafka**

          Amazon Managed Streaming for Apache Kafka (Amazon MSK) is a fully managed AWS streaming data service makes it easy for you to build and run applications that use Apache Kafka to process streaming data without needing to become an expert in operating Apache Kafka clusters. Amazon MSK manages the provisioning, configuration, and maintenance of Apache Kafka clusters and Apache ZooKeeper nodes for you. Amazon MSK also shows key Apache Kafka performance metrics in the AWS Console.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Kafka Developer Resources
          url: https://aws.amazon.com/msk/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kafka maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/kafka-controller:1.2.1
        version: 1.2.1
      entries:
      - name: ack-kafka-controller.v1.2.1
        version: 1.2.1
      - name: ack-kafka-controller.v1.2.0
        version: 1.2.0
      - name: ack-kafka-controller.v1.1.3
        version: 1.1.3
      - name: ack-kafka-controller.v1.1.2
        version: 1.1.2
      - name: ack-kafka-controller.v1.1.1
        version: 1.1.1
      - name: ack-kafka-controller.v1.1.0
        version: 1.1.0
      - name: ack-kafka-controller.v1.0.14
        version: 1.0.14
      - name: ack-kafka-controller.v1.0.13
        version: 1.0.13
      - name: ack-kafka-controller.v1.0.12
        version: 1.0.12
      - name: ack-kafka-controller.v1.0.11
        version: 1.0.11
      - name: ack-kafka-controller.v1.0.10
        version: 1.0.10
      - name: ack-kafka-controller.v1.0.8
        version: 1.0.8
      - name: ack-kafka-controller.v1.0.7
        version: 1.0.7
      - name: ack-kafka-controller.v1.0.6
        version: 1.0.6
      - name: ack-kafka-controller.v1.0.5
        version: 1.0.5
      - name: ack-kafka-controller.v1.0.4
        version: 1.0.4
      - name: ack-kafka-controller.v1.0.3
        version: 1.0.3
      - name: ack-kafka-controller.v1.0.2
        version: 1.0.2
      - name: ack-kafka-controller.v1.0.1
        version: 1.0.1
      - name: ack-kafka-controller.v1.0.0
        version: 1.0.0
      - name: ack-kafka-controller.v0.0.14
        version: 0.0.14
      - name: ack-kafka-controller.v0.0.12
        version: 0.0.12
      - name: ack-kafka-controller.v0.0.11
        version: 0.0.11
      - name: ack-kafka-controller.v0.0.10
        version: 0.0.10
      - name: ack-kafka-controller.v0.0.9
        version: 0.0.9
      - name: ack-kafka-controller.v0.0.8
        version: 0.0.8
      - name: ack-kafka-controller.v0.0.7
        version: 0.0.7
      - name: ack-kafka-controller.v0.0.6
        version: 0.0.6
      - name: ack-kafka-controller.v0.0.5
        version: 0.0.5
      - name: ack-kafka-controller.v0.0.4
        version: 0.0.4
      - name: ack-kafka-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-kafka-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-acmpca-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-acmpca-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acmpca.services.k8s.aws/v1alpha1",
                "kind": "CertificateAuthority",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/acmpca-controller:1.1.1
          createdAt: "2025-09-19T20:51:07Z"
          description: AWS ACM PCA controller is a service controller for managing
            ACM PCA resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CertificateAuthority represents the state of an AWS acmpca
              CertificateAuthority resource.
            displayName: CertificateAuthority
            kind: CertificateAuthority
            name: certificateauthorities.acmpca.services.k8s.aws
            version: v1alpha1
          - description: CertificateAuthorityActivation represents the state of an
              AWS acmpca CertificateAuthorityActivation resource.
            displayName: CertificateAuthorityActivation
            kind: CertificateAuthorityActivation
            name: certificateauthorityactivations.acmpca.services.k8s.aws
            version: v1alpha1
          - description: Certificate represents the state of an AWS acmpca Certificate
              resource.
            displayName: Certificate
            kind: Certificate
            name: certificates.acmpca.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ACM PCA resources in AWS from within your Kubernetes cluster.

          **About Amazon ACM PCA**

          AWS Private CA enables creation of private certificate authority (CA) hierarchies, including root and subordinate CAs, without the investment and maintenance costs of operating an on-premises CA. Your private CAs can issue end-entity X.509 certificates useful in scenarios including:
          - Creating encrypted TLS communication channels
          - Authenticating users, computers, API endpoints, and IoT devices
          - Cryptographically signing code
          - Implementing Online Certificate Status Protocol (OCSP) for obtaining certificate revocation status

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ACM PCA
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acmpca
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ACM PCA Developer Resources
          url: https://aws.amazon.com/private-ca/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: acmpca maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/acmpca-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-acmpca-controller.v1.1.1
        version: 1.1.1
      - name: ack-acmpca-controller.v1.1.0
        version: 1.1.0
      - name: ack-acmpca-controller.v1.0.10
        version: 1.0.10
      - name: ack-acmpca-controller.v1.0.9
        version: 1.0.9
      - name: ack-acmpca-controller.v1.0.8
        version: 1.0.8
      - name: ack-acmpca-controller.v1.0.7
        version: 1.0.7
      - name: ack-acmpca-controller.v1.0.6
        version: 1.0.6
      - name: ack-acmpca-controller.v1.0.5
        version: 1.0.5
      - name: ack-acmpca-controller.v1.0.4
        version: 1.0.4
      - name: ack-acmpca-controller.v1.0.3
        version: 1.0.3
      - name: ack-acmpca-controller.v1.0.2
        version: 1.0.2
      - name: ack-acmpca-controller.v1.0.1
        version: 1.0.1
      - name: ack-acmpca-controller.v1.0.0
        version: 1.0.0
      - name: ack-acmpca-controller.v0.0.20
        version: 0.0.20
      - name: ack-acmpca-controller.v0.0.16
        version: 0.0.16
      - name: ack-acmpca-controller.v0.0.15
        version: 0.0.15
      - name: ack-acmpca-controller.v0.0.14
        version: 0.0.14
      - name: ack-acmpca-controller.v0.0.13
        version: 0.0.13
      - name: ack-acmpca-controller.v0.0.12
        version: 0.0.12
      - name: ack-acmpca-controller.v0.0.11
        version: 0.0.11
      - name: ack-acmpca-controller.v0.0.10
        version: 0.0.10
      - name: ack-acmpca-controller.v0.0.9
        version: 0.0.9
      - name: ack-acmpca-controller.v0.0.8
        version: 0.0.8
      - name: ack-acmpca-controller.v0.0.5
        version: 0.0.5
      - name: ack-acmpca-controller.v0.0.4
        version: 0.0.4
      - name: ack-acmpca-controller.v0.0.3
        version: 0.0.3
      - name: ack-acmpca-controller.v0.0.2
        version: 0.0.2
      - name: ack-acmpca-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-acmpca-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MayaData
      provider-url: ""
    name: openebs
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openebsoperator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "openebs.io/v1alpha1",
                "kind": "OpenEBSInstallTemplate",
                "metadata": {
                  "name": "oebs",
                  "namespace": "openebs"
                },
                "spec": {
                  "rbac": {
                    "create": true,
                    "pspEnabled": false,
                    "kyvernoEnabled": false
                  },
                  "serviceAccount": {
                    "create": true,
                    "name": "openebs-maya-operator"
                  },
                  "imagePullSecrets": [],
                  "release": {
                    "version": "3.0.0"
                  },
                  "legacy": {
                    "enabled": false
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": ""
                  },
                  "apiserver": {
                    "enabled": true,
                    "image": "openebs/m-apiserver",
                    "imageTag": "2.12.2",
                    "replicas": 1,
                    "ports": {
                      "externalPort": 5656,
                      "internalPort": 5656
                    },
                    "sparse": {
                      "enabled": "false"
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "defaultStorageConfig": {
                    "enabled": "true"
                  },
                  "varDirectoryPath": {
                    "baseDir": "/var/openebs"
                  },
                  "provisioner": {
                    "enabled": true,
                    "image": "openebs/openebs-k8s-provisioner",
                    "imageTag": "2.12.2",
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "patchJivaNodeAffinity": "enabled",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "localprovisioner": {
                    "enabled": true,
                    "image": "openebs/provisioner-localpv",
                    "imageTag": "3.0.0",
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "enableDeviceClass": true,
                    "enableHostpathClass": true,
                    "basePath": "/var/openebs/local",
                    "waitForBDBindTimeoutRetryCount": "12",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "snapshotOperator": {
                    "enabled": true,
                    "controller": {
                      "image": "openebs/snapshot-controller",
                      "imageTag": "2.12.2",
                      "resources": {}
                    },
                    "provisioner": {
                      "image": "openebs/snapshot-provisioner",
                      "imageTag": "2.12.2",
                      "resources": {}
                    },
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "upgradeStrategy": "Recreate",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    }
                  },
                  "ndm": {
                    "enabled": true,
                    "image": "openebs/node-disk-manager",
                    "imageTag": "1.7.0",
                    "sparse": {
                      "path": "/var/openebs/sparse",
                      "size": "10737418240",
                      "count": "0"
                    },
                    "filters": {
                      "enableOsDiskExcludeFilter": true,
                      "osDiskExcludePaths": "/,/etc/hosts,/boot",
                      "enableVendorFilter": true,
                      "excludeVendors": "CLOUDBYT,OpenEBS",
                      "enablePathFilter": true,
                      "includePaths": "",
                      "excludePaths": "/dev/loop,/dev/fd0,/dev/sr0,/dev/ram,/dev/dm-,/dev/md,/dev/rbd,/dev/zd"
                    },
                    "probes": {
                      "enableSeachest": false
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "ndmOperator": {
                    "enabled": true,
                    "image": "openebs/node-disk-operator",
                    "imageTag": "1.7.0",
                    "replicas": 1,
                    "upgradeStrategy": "Recreate",
                    "nodeSelector": {},
                    "tolerations": [],
                    "healthCheck": {
                      "initialDelaySeconds": 15,
                      "periodSeconds": 20
                    },
                    "readinessCheck": {
                      "initialDelaySeconds": 5,
                      "periodSeconds": 10
                    },
                    "resources": {}
                  },
                  "ndmExporter": {
                    "enabled": false,
                    "image": {
                      "registry": null,
                      "repository": "openebs/node-disk-exporter",
                      "pullPolicy": "IfNotPresent",
                      "tag": "1.7.0"
                    },
                    "nodeExporter": {
                      "name": "ndm-node-exporter",
                      "podLabels": {
                        "name": "openebs-ndm-node-exporter"
                      },
                      "metricsPort": 9101
                    },
                    "clusterExporter": {
                      "name": "ndm-cluster-exporter",
                      "podLabels": {
                        "name": "openebs-ndm-cluster-exporter"
                      },
                      "metricsPort": 9100
                    }
                  },
                  "webhook": {
                    "enabled": true,
                    "image": "openebs/admission-server",
                    "imageTag": "2.12.2",
                    "failurePolicy": "Fail",
                    "replicas": 1,
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "hostNetwork": false,
                    "resources": {}
                  },
                  "helper": {
                    "image": "openebs/linux-utils",
                    "helperImageTag": "3.0.0"
                  },
                  "featureGates": {
                    "enabled": true,
                    "GPTBasedUUID": {
                      "enabled": true,
                      "featureGateFlag": "GPTBasedUUID"
                    },
                    "APIService": {
                      "enabled": false,
                      "featureGateFlag": "APIService",
                      "address": "0.0.0.0:9115"
                    },
                    "UseOSDisk": {
                      "enabled": false,
                      "featureGateFlag": "UseOSDisk"
                    },
                    "ChangeDetection": {
                      "enabled": false,
                      "featureGateFlag": "ChangeDetection"
                    }
                  },
                  "crd": {
                    "enableInstall": true
                  },
                  "policies": {
                    "monitoring": {
                      "enabled": true,
                      "image": "openebs/m-exporter",
                      "imageTag": "2.12.2"
                    }
                  },
                  "analytics": {
                    "enabled": true,
                    "pingInterval": "24h"
                  },
                  "jiva": {
                    "image": "openebs/jiva",
                    "imageTag": "2.12.2",
                    "replicas": 3,
                    "defaultStoragePath": "/var/openebs",
                    "enabled": false,
                    "openebsLocalpv": {
                      "enabled": false
                    },
                    "localpv-provisioner": {
                      "openebsNDM": {
                        "enabled": false
                      }
                    }
                  },
                  "cstor": {
                    "pool": {
                      "image": "openebs/cstor-pool",
                      "imageTag": "2.12.2"
                    },
                    "poolMgmt": {
                      "image": "openebs/cstor-pool-mgmt",
                      "imageTag": "2.12.2"
                    },
                    "target": {
                      "image": "openebs/cstor-istgt",
                      "imageTag": "2.12.2"
                    },
                    "volumeMgmt": {
                      "image": "openebs/cstor-volume-mgmt",
                      "imageTag": "2.12.2"
                    },
                    "enabled": false,
                    "openebsNDM": {
                      "enabled": false
                    }
                  },
                  "openebs-ndm": {
                    "enabled": false
                  },
                  "localpv-provisioner": {
                    "enabled": false,
                    "openebsNDM": {
                      "enabled": false
                    }
                  },
                  "zfs-localpv": {
                    "enabled": false
                  },
                  "lvm-localpv": {
                    "enabled": false
                  },
                  "nfs-provisioner": {
                    "enabled": false
                  },
                  "cleanup": {
                    "image": {
                      "registry": "",
                      "repository": "bitnami/kubectl",
                      "tag": "",
                      "imagePullSecrets": []
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: index.docker.io/openebs/helm-operator:v0.0.9
          createdAt: "2021-11-02T02:56:58Z"
          description: Creates and maintains OpenEBS Control Plane deployments
          repository: https://github.com/openebs/helm-operator
          support: https://slack.openebs.io/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a OpenEBS Install Operator
            displayName: OpenEBS Install Template
            kind: OpenEBSInstallTemplate
            name: openebsinstalltemplates.openebs.io
            version: v1alpha1
        description: |
          **OpenEBS** is a leading container attached storage solution that enables the use of containers for mission-critical, persistent workloads and for other stateful workloads such as logging or Prometheus for example.
          OpenEBS itself is deployed as just another container on your host and enables storage services that can be designated on a per pod, application, cluster or container level, including:
          * Data persistence across nodes
          * Synchronization of data across availability zones and cloud providers
          * A common layer whether you are running on the cloud, or your bare metal
          * Integration with Kubernetes, so developer and application intent flows into OpenEBS
          * Management of tiering to and from S3 and other targets.
          ## OpenEBS Operator
          OpenEBS primarily provides container attached block storage (iSCSI volumes) by leveraging/aggregating the storage on the nodes, with the storage controller itself running as a container. Different storage engines (Jiva & cStor) are supported, with tools available to dynamically provision Kubernetes Local PVs. The volumes are dynamically provisioned via PersistentVolumeClaims and are managed by a control plane component called "maya", which also runs as a deployment in the K8s cluster. In addition to maya, a typical OpenEBS installation comprises several other resources, which aid with various functionalities, ranging from snapshotting to disk management. All these components are described briefly below:
          * **Maya-API-Server** - A storage orchestrator which integrates into Kubernetes workflow to help provision
            and manage OpenEBS Jiva & cStor (storage engine) volumes

          * **Dynamic-OpenEBS-Provisioner** - A Kubernetes external storage provisioner that utilizes APIs exposed by maya-apiserver
            to perform provision & delete operations of Jiva & cStor volumes

          * **Dynamic-LocalPV-Provisioner** - A dynamic provisioner for Kubernetes Local PVs
          * **OpenEBS-Snapshot-Operator** - A Kubernetes snapshot controller that creates & restores OpenEBS volume snapshots
          * **Node-Disk-Manager** - A disk management controller which identifies available disks, maintains inventory, and
            dynamically attaches/detaches disks to backend storage pods

          The helm-based OpenEBS Operator eases the setup of all the above mentioned components, with a simple custom resource provided to define the install options, thereby enabling applications to start using the OpenEBS storageclasses in their PVCs. The OpenEBSInstallTemplate CR can be used to specify start-up parameters & also update/overwrite the definitions post install.
          ## Pre-Requisites
          Before installing OpenEBS control plane, perform the following steps to ensure successful deployment of the Node-Disk-Manager & the OpenEBS volume replicas, respectively.
          * Configure the OpenEBS service account on the openshift-operators namespace/project to use the privileged security context constraint.

            **Note**: The serviceaccount name is same as the one specified in the `spec.serviceAccount.name` field of the OpenEBSInstallTemplate CR.


            ```
            oc adm policy add-scc-to-user privileged system:serviceaccount:openshift-operators:openebs-maya-operator
            ```

          * Configure the default service account on the namespace/project in which the volume replicas are deployed to use privileged
            security context constraint.


            ```
            oc adm policy add-scc-to-user privileged system:serviceaccount:openshift-operators:default
            ```

          ## Getting Started
          * Try the quickstart [guide](https://github.com/openebs/helm-operator/blob/master/olm/README.md)
          * To learn how to contribute, please read the [contribution guide](https://github.com/openebs/helm-operator/blob/master/CONTRIBUTING.md)
          * OpenEBS welcomes your feedback and contributions in any form possible. [Join our Community](https://openebs.org/community)

          ## License
          OpenEBS is distributed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt)
        displayName: OpenEBS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenEBS
        - ContainerAttachedStorage
        links:
        - name: OpenEBS Website
          url: https://openebs.io
        - name: Operator Source Code
          url: https://github.com/openebs/helm-operator
        - name: Install Instructions
          url: https://github.com/openebs/helm-operator/blob/master/olm/README.md
        maintainers:
        - email: kiran.mova@openebs.io
          name: kmova
        maturity: alpha
        minKubeVersion: 1.12.0
        provider:
          name: MayaData
        relatedImages:
        - index.docker.io/openebs/helm-operator:v0.0.9
        version: 3.0.0
      entries:
      - name: openebsoperator.v3.0.0
        version: 3.0.0
      - name: openebsoperator.v2.12.2
        version: 2.12.2
      name: alpha
    defaultChannel: alpha
    packageName: openebs
    provider:
      name: MayaData
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-core
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Core",
                "metadata": {
                  "name": "core-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configService": {
                    "port": 8080
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "helper": {
                    "port": 8080
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "licenseService": {
                    "port": 8080
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "restStatsService": {
                    "port": 8080
                  },
                  "resultService": {
                    "port": 8080,
                    "storage": "64Gi"
                  },
                  "sessionManager": {
                    "port": 8080
                  },
                  "statsService": {
                    "port": 8080
                  },
                  "testResultsService": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:34:00Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Core
            name: cores.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Core Operator
        displayName: Keysight WAP Core Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Core
          url: https://keysight-wap-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-core@sha256:001b0d08cd84a9ebc13cd8fd35b68d431b96f84d71e933c1123ed6c0d244fd37
        version: 0.1.21
      entries:
      - name: keysight-wap-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-nats-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-nats-operator.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2023-09-19T12:44:10Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NatsCluster
            name: natsclusters.nats.io
            version: v1alpha2
          - kind: NatsServiceRole
            name: natsserviceroles.nats.io
            version: v1alpha2
        description: Keysight Nats Operator
        displayName: Keysight Nats Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Nats Operator
          url: https://keysight-nats-operator.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-nats-operator@sha256:2ee5dc8814274e83a5f82e89eabd0b643e8cde8e3d555a56535a61d05c72fd75
        version: 0.1.21
      entries:
      - name: keysight-nats-operator.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-nats-operator
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: github.com/stakater-ab/hibernation-operator
    name: hibernation-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hibernation-operator.v0.1.103
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Monitoring
          createdAt: "2025-10-02T05:44:42Z"
          description: Operator to engage and disengage hibernation on k8s resources
            using cron expressions
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.1.103
          operatorframework.io/suggested-namespace: hibernation-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Cluster Resource Supervisor
            kind: ClusterResourceSupervisor
            name: clusterresourcesupervisors.hibernation.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.hibernation.stakater.com
            version: v1beta1
        description: Schedule k8s resources using cron expressions via Hibernation
          Operator
        displayName: Hibernation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - redhat
        - k8s
        - stakater
        - hibernation
        - operator
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maintainers:
        - email: hello@stakater.com
          name: StakaterAB
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Stakater
          url: github.com/stakater-ab/hibernation-operator
        relatedImages:
        - registry.connect.redhat.com/stakater/hibernation-operator@sha256:3738f82834d53aec4f8fba5caad6258c7adb4bdc23a57bde47c7ec22030fae5c
        version: 0.1.103
      entries:
      - name: hibernation-operator.v0.1.103
        version: 0.1.103
      name: release-0.1
    defaultChannel: release-0.1
    packageName: hibernation-operator
    provider:
      name: Stakater
      url: github.com/stakater-ab/hibernation-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio Project
      provider-url: ""
    name: apicurito
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: 1.0.x
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: candidate
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: latest
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: latest
    packageName: apicurito
    provider:
      name: Apicurio Project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Apache Software Foundation
      provider-url: ""
    name: camel-k
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: camel-k-operator.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.0.1
          createdAt: "2023-09-04T12:32:38Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
          - kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.0.1
        version: 2.0.1
      entries:
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: candidate
    - currentCSV: camel-k-operator.v2.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "setBody": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationProfile",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "debug"
                },
                "spec": {
                  "traits": {
                    "logging": {
                      "level": "DEBUG"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example-source"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "setBody": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "timer-to-log-example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "log-sink"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world",
                      "period": 2000
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "timer-source"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.8.0
          createdAt: "2025-09-21T09:07:43Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: IntegrationProfile is the Schema for the integrationprofiles
              API
            displayName: Integration Profile
            kind: IntegrationProfile
            name: integrationprofiles.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Pipe is the Schema for the pipes API
            displayName: Pipe
            kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers:
          the easiest way to build and manage your Camel applications on Kubernetes.

          ## Installation

          You may need to specify the container registry where to store the container images created for the Camel applications.
          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          spec:
            build:
              registry:
                address: docker.io/my-org
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step when using Openshift cluster.

          ## Running an Integration

          Run a Camel application on the cluster by creating a sample `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            annotations:
              camel.apache.org/operator.id: camel-k
            name: test
          spec:
            flows:
            - from:
                parameters:
                  period: "1000"
                steps:
                - setBody:
                    simple: Hello Camel from ${routeId}
                - log: ${body}
                uri: timer:yaml
          ```
          Log the result of the running application:
          ```
          kubectl wait --for=condition=ready integration test --timeout=180s
          kubectl logs deployment/test -f
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.8.0
        version: 2.8.0
      entries:
      - name: camel-k-operator.v2.8.0
        version: 2.8.0
      - name: camel-k-operator.v2.7.0
        version: 2.7.0
      - name: camel-k-operator.v2.6.0
        version: 2.6.0
      - name: camel-k-operator.v2.5.1
        version: 2.5.1
      - name: camel-k-operator.v2.5.0
        version: 2.5.0
      - name: camel-k-operator.v2.4.0
        version: 2.4.0
      - name: camel-k-operator.v2.3.3
        version: 2.3.3
      - name: camel-k-operator.v2.3.2
        version: 2.3.2
      - name: camel-k-operator.v2.3.1
        version: 2.3.1
      - name: camel-k-operator.v2.3.0
        version: 2.3.0
      - name: camel-k-operator.v2.2.0
        version: 2.2.0
      - name: camel-k-operator.v2.1.0
        version: 2.1.0
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: latest
    - currentCSV: camel-k-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.12.0
          createdAt: "2023-02-27T09:01:55Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.12.0
        version: 1.12.0
      entries:
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable
    - currentCSV: camel-k-operator.v1.10.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.10.4
          createdAt: "2023-01-28T13:49:05Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.10.4
        version: 1.10.4
      entries:
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.10
    - currentCSV: camel-k-operator.v1.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.11.1
          createdAt: "2023-01-30T08:45:04Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.11.1
        version: 1.11.1
      entries:
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.11
    - currentCSV: camel-k-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.12.1
          createdAt: "2023-06-12T08:46:58Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.12.1
        version: 1.12.1
      entries:
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.12
    - currentCSV: camel-k-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.5.1
          createdAt: "2021-08-31T09:03:41Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.5.1
        version: 1.5.1
      entries:
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.5
    - currentCSV: camel-k-operator.v1.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.6.1
          createdAt: "2021-10-27T21:06:20Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.6.1
        version: 1.6.1
      entries:
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.6
    - currentCSV: camel-k-operator.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.7.0
          createdAt: "2021-11-16T10:54:12Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.8.1+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.7.0
        version: 1.7.0
      entries:
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.7
    - currentCSV: camel-k-operator.v1.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.8.2
          createdAt: "2022-03-07T07:13:19Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.8.2
        version: 1.8.2
      entries:
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.8
    - currentCSV: camel-k-operator.v1.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.9.2
          createdAt: "2022-05-23T09:05:22Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.9.2
        version: 1.9.2
      entries:
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.9
    - currentCSV: camel-k-operator.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.0.1
          createdAt: "2023-09-04T12:32:38Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
          - kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.0.1
        version: 2.0.1
      entries:
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-2.0
    - currentCSV: camel-k-operator.v2.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "setBody": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationProfile",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "debug"
                },
                "spec": {
                  "traits": {
                    "logging": {
                      "level": "DEBUG"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example-source"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "setBody": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "timer-to-log-example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "log-sink"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world",
                      "period": 2000
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "timer-source"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.8.0
          createdAt: "2025-09-21T09:07:43Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: IntegrationProfile is the Schema for the integrationprofiles
              API
            displayName: Integration Profile
            kind: IntegrationProfile
            name: integrationprofiles.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Pipe is the Schema for the pipes API
            displayName: Pipe
            kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers:
          the easiest way to build and manage your Camel applications on Kubernetes.

          ## Installation

          You may need to specify the container registry where to store the container images created for the Camel applications.
          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          spec:
            build:
              registry:
                address: docker.io/my-org
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step when using Openshift cluster.

          ## Running an Integration

          Run a Camel application on the cluster by creating a sample `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            annotations:
              camel.apache.org/operator.id: camel-k
            name: test
          spec:
            flows:
            - from:
                parameters:
                  period: "1000"
                steps:
                - setBody:
                    simple: Hello Camel from ${routeId}
                - log: ${body}
                uri: timer:yaml
          ```
          Log the result of the running application:
          ```
          kubectl wait --for=condition=ready integration test --timeout=180s
          kubectl logs deployment/test -f
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.8.0
        version: 2.8.0
      entries:
      - name: camel-k-operator.v2.8.0
        version: 2.8.0
      - name: camel-k-operator.v2.7.0
        version: 2.7.0
      - name: camel-k-operator.v2.6.0
        version: 2.6.0
      - name: camel-k-operator.v2.5.1
        version: 2.5.1
      - name: camel-k-operator.v2.5.0
        version: 2.5.0
      - name: camel-k-operator.v2.4.0
        version: 2.4.0
      - name: camel-k-operator.v2.3.3
        version: 2.3.3
      - name: camel-k-operator.v2.3.2
        version: 2.3.2
      - name: camel-k-operator.v2.3.1
        version: 2.3.1
      - name: camel-k-operator.v2.3.0
        version: 2.3.0
      - name: camel-k-operator.v2.2.0
        version: 2.2.0
      - name: camel-k-operator.v2.1.0
        version: 2.1.0
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-v2
    defaultChannel: stable-v2
    packageName: camel-k
    provider:
      name: The Apache Software Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-cpd-ccs
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-ccs.v10.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ccs.cpd.ibm.com/v1beta1",
                "kind": "CCS",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ccs-cr",
                    "app.kubernetes.io/managed-by": "ibm-cpd-ccs-operator",
                    "app.kubernetes.io/name": "ibm-cpd-ccs-operator"
                  },
                  "name": "ccs-cr"
                },
                "spec": {
                  "scaleConfig": "small",
                  "license": {
                    "accept": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
          createdAt: "2025-04-18"
          description: Common Core Services
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-ccs
          olm.skipRange: '>=7.0.0 <10.3.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM watsonx.ai"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: |-
            [
              "runtimeassemblies.runtimes.ibm.com"
            ]
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation Runtime Manager CRD. This product is not deployable
              standalone and it is part of the LICCR already accepted.  For additional
              details regarding install parameters check: http://ibm.biz/-readme.
              License By installing this product you accept the license terms https://ibm.biz/BdaaLC
              or https://ibm.biz/BdaaL9.'
            kind: RuntimeAssembly
            name: runtimeassemblies.runtimes.ibm.com
            version: v1alpha1
          - description: 'Documentation This product is not deployable standalone
              and it is part of the LICCR already accepted.  For additional details
              regarding install parameters check: http://ibm.biz/-readme. License
              By installing this product you accept the license terms https://ibm.biz/BdaaLC
              or https://ibm.biz/BdaaL9.'
            kind: CCS
            name: ccs.ccs.cpd.ibm.com
            version: v1beta1
        description: |-
          # Introduction Common Core Services Operator
          Put stuff here

          # Details
          Various Details
          ## Prerequisites
          Various Prerequisites
          ## Resources Required
          Various Resources Required
          ## Installing
          Various Installation details
          # Configuration
          Various Configuration information
          ### Limitations
          Various Limitations
          ### SecurityContextConstraints Requirements
          All pods will use the restricted SCC, I think.
        displayName: Common Core Services
        installModes:
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: OwnNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commonc-core-services
        - cpd40
        links:
        - name: Ccs
          url: https://ccs.domain
        maintainers:
        - email: test@ibm.com
          name: test
        maturity: v1
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/jobs-api@sha256:173b88f90078e1817686ac949f48387af70a706e23e6fdab4a7c6217db5ad86e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-bundle@sha256:94c7a1dd290e2abd55c5b8998cf08ad174d3fbfc659ae6fa6fb6cbab61d5fe14
        - registry.marketplace.redhat.com/cp/cpd/dap-base-job@sha256:f208686a3f56d2980e7b8640b6ebdea9057109e7520c6817a731e68d08506a36
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:bcdf9b7eea8753b3e463973a2690140ae7aa8c02c3c7726511f116bd751ec2d8
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:c268d2852580286e5c6770857fcfc3b06a698521f5acd17ed83b8e2cb9ab55cf
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:523dcc671f1c05e690d5e578a90300d28d76570f605d4de7f80645b590f4a462
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:81d8b25221d827160dd5e88efd383453f532057ad02d7a2ee3dcd9c6eb16ba65
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:0e0ae96aff8540323da8a2379af014d3af1770e4b0043b84b9b415dda44501f7
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:6f2e7cc917adb54b3112dc4909f645ac08c33e8c2c33266fbc8f5133cbcd1973
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:21c81ff7c4b8809accaf62b242fd9b3d52f235489ae427556b7f71cc21bf8926
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:351375de35d7fc9d940b905b867ce8a00b7ff1416e70d3a100918797fe3d8c56
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:4fb2127ba7c525a959bbbcf024de364ef266ebb3ca7e696f74d990d2552cda59
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:56315ce0a0a70f5899191a7eab856105489143f0e889c8cdd1b6b8c225f02eb7
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:a3a0d58f895fca4d832e602d0dee80018df4ad7bad8800ad4058eea07fda4817
        - registry.marketplace.redhat.com/cp/cpd/ngp-projects-api@sha256:df420e9f616ad8a123d9eedd117fb6cb007b30a772138db534f47ad8801fa96d
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:257bf301338f7b123099e2ffe31c3f3b793bf28d6a0570a25b889271239579ba
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:6f96b8b6afb00cb26fcf903b8e4c30b0d6344bb86d8bce2f2026fc8cb0b575ec
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:0621eb9b83c9b2b8bfe5cd09797fe08fa37a854a5506914935ff6f0df599b631
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:0ae22914dbaf09d6679bd06bc9b6a5728e744c9ce2309cbd5f01e28308c403ea
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:562281dbb2bec5191a7c6b595ab853092e6fe3ff1432e77b2471f71ccdffcbc8
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:54c86367bfe0ab658611856cc0a6c5de24c14a336bd129ac9f16293945a55238
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:f5527ddbbd0b9878555fce852996b13020fd91b1604d8c03d34a61ca2fdb2d1d
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:fbc6eab07fcfc2d0c1895f18e17f751e9481ad111875959e9ff08e14f81b9a35
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:740a1d87a43e61e32cfdea032d852af33100d959c8b57ff2f165f4baf77fc011
        - registry.marketplace.redhat.com/cp/cpd/task-credentials@sha256:e162828d5ad07ceae4db0589f930979842714e3b766552102bee373a3a570af3
        - registry.marketplace.redhat.com/cp/cpd/environments-ui@sha256:2923664475c7a8ea5e63055c90748db37825fe76c7975b845f1185d4b32ee4ea
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/cpd/runtime-manager-api@sha256:a37773df07fe757dd698e3f09f8687902e7b3592d97855e703dd92ecc832aab0
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:8a6ce0dada706056efda73ad0cc1547566003e0a7aab8ed379b50e4776861b94
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e12599ff99db058fdc8f9c1eebfba31160c636ff092138deb1f6f6705f44fd62
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:3f2f97e6b2b18526255e88ff02d1039b24312eca51ee22b7cbf76d5680f7aee1
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:d023f35b65816f244b6d699b41362e6031553f7d4c416efcee4a7f2395a61355
        - registry.marketplace.redhat.com/cp/cpd/spaces-ui-job@sha256:95dade51444e30ec034c12207ecbef675b70889b81970f49f10aa75160b2f947
        - registry.marketplace.redhat.com/cp/cpd/environment-specifications-sync@sha256:894b4624d8ead7286cf9068c6477300992c32d96339644fe764bcb29b1c8ef5c
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:ad19b65042fb26cc1244a92f81a96e5ceb4313f8c823181efe28b66e9fecf9c8
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:044b663e4b7508c62dbb28b7fb2aa267c9771fc62ab509b08d46ea1499d95431
        - registry.marketplace.redhat.com/cp/cpd/environments-api@sha256:6dd60b530f81d7a2d55a534211f7f128814abf87b8cd3575d528dba53442c44e
        - registry.marketplace.redhat.com/cp/cpd/jobs-ui@sha256:49daf607918b378593485da16ac417056ee295bca94f1a5cac37b334381ae096
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:6cffef7631ee54612fba71efc0eff4e17bc08defc3e2137aee9881d685444298
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:722944bfd0eba280255f4c198f65f562423a26e2c4ad3580591ee71cb6e9c2ef
        - registry.marketplace.redhat.com/cp/cpd/job-trigger@sha256:f2ea487072393a24e4c6864b2107c1fce072ebba07f9dbc2c139bd37fbd062da
        - registry.marketplace.redhat.com/cp/cpd/environments-init-job@sha256:6af8f900af7c11eee6894e2c3e52f740b3c700dac0ee753fadef330230bd42dd
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:6fe90f051d9261ad224dccf4d9f6bc9f19ce0e8d22fa010a46b2752374bce2b3
        version: 10.3.0
      entries:
      - name: ibm-cpd-ccs.v10.3.0
        version: 10.3.0
      name: v10.3
    defaultChannel: v10.3
    packageName: ibm-cpd-ccs
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: swift-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: swift-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "swift.openstack.org/v1beta1",
                "kind": "Swift",
                "metadata": {
                  "name": "swift"
                },
                "spec": {
                  "swiftProxy": {
                    "passwordSelectors": {
                      "service": "SwiftPassword"
                    },
                    "replicas": 1
                  },
                  "swiftRing": {
                    "ringReplicas": 1
                  },
                  "swiftStorage": {
                    "replicas": 1,
                    "storageClass": "local-storage"
                  }
                }
              },
              {
                "apiVersion": "swift.openstack.org/v1beta1",
                "kind": "SwiftProxy",
                "metadata": {
                  "name": "swift-proxy"
                },
                "spec": {
                  "passwordSelectors": {
                    "service": "SwiftPassword"
                  },
                  "replicas": 1,
                  "secret": "osp-secret",
                  "serviceUser": "swift"
                }
              },
              {
                "apiVersion": "swift.openstack.org/v1beta1",
                "kind": "SwiftRing",
                "metadata": {
                  "name": "swift-ring"
                },
                "spec": {
                  "ringReplicas": 1
                }
              },
              {
                "apiVersion": "swift.openstack.org/v1beta1",
                "kind": "SwiftStorage",
                "metadata": {
                  "name": "swift-storage"
                },
                "spec": {
                  "replicas": 1,
                  "storageClass": "local-storage",
                  "storageRequest": "10G"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:34Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SwiftProxy is the Schema for the swiftproxies API
            displayName: Swift Proxy
            kind: SwiftProxy
            name: swiftproxies.swift.openstack.org
            version: v1beta1
          - description: SwiftRing is the Schema for the swiftrings API
            displayName: Swift Ring
            kind: SwiftRing
            name: swiftrings.swift.openstack.org
            version: v1beta1
          - description: Swift is the Schema for the swifts API
            displayName: Swift
            kind: Swift
            name: swifts.swift.openstack.org
            version: v1beta1
          - description: SwiftStorage is the Schema for the swiftstorages API
            displayName: Swift Storage
            kind: SwiftStorage
            name: swiftstorages.swift.openstack.org
            version: v1beta1
        description: OpenStack Swift operator
        displayName: Swift operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Swift
        - Object Storage
        links:
        - name: Swift Operator
          url: https://github.com/openstack-k8s-operators/swift-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-swift-object-rhel9@sha256:656f23b25ca0bd80d7b77ba717a21455366dfaf2049df1bd6d6536925a18b94b
        - registry.redhat.io/rhoso/openstack-swift-proxy-server-rhel9@sha256:6e654db83d570a3f8feedb912d5d9dff35ecba4d53169481a3b4982d3a832798
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/swift-rhel9-operator@sha256:18207f02c049447836211b326f09ce660782c384cda92e592db256692e3c8e73
        - registry.redhat.io/rhoso/openstack-swift-account-rhel9@sha256:ecebf3eb430188392f7a7b8cda3d1c0c0f658bcbbaff531d99eba88a7e0c9889
        - registry.redhat.io/rhoso/openstack-swift-container-rhel9@sha256:782866f8e9661de50d6a74bc52262f4546d7e0055d0c4da83ba6b891a50c35ea
        version: 1.0.6
      entries:
      - name: swift-operator.v1.0.6
        version: 1.0.6
      - name: swift-operator.v1.0.5
        version: 1.0.5
      - name: swift-operator.v1.0.4
        version: 1.0.4
      - name: swift-operator.v1.0.3
        version: 1.0.3
      - name: swift-operator.v1.0.2
        version: 1.0.2
      - name: swift-operator.v1.0.1
        version: 1.0.1
      - name: swift-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: swift-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/component: kmm
      app.kubernetes.io/name: kmm
      app.kubernetes.io/part-of: kmm
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: kernel-module-management
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kernel-module-management.v2.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:1022205d712a2ed9d6e76968961c180d4e66a2e0fb991221b0b02ed94ed830d2
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.0.2'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel-module-management
        - out-of-tree
        - kernel
        - kmm
        - module
        - modules
        - oot
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:1022205d712a2ed9d6e76968961c180d4e66a2e0fb991221b0b02ed94ed830d2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62c44057a27abd41cf3daf60b8736da59a82dadb65fa6c32bef65d19bd49ea49
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:25f39f20f325f64676bf8ef14cd8a60cfbdc77116753a89d4bb667ea2a86172c
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:192e08b121a04ae69aea60c35199e8f5e3c42305e1ebd024a605a371de4ab472
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:94b4ddc45cb34d4ffe794a51b99c868cb709f90cc4eb8b28284d70c526a0a226
        version: 2.0.2
      entries:
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.0
    - currentCSV: kernel-module-management.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a85037778c53c3766f39efe4c4d2523d01ae7743e5c5693b05bd8b2cbb50b1d8
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.1.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a85037778c53c3766f39efe4c4d2523d01ae7743e5c5693b05bd8b2cbb50b1d8
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:611556d999ea97f9dbb78fec5963baa02de7a26626d7096ba63b177e3bcf9707
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:14e100e5853444114029116255ebfb7a54e5ed51503ba969740ad48b5ca80f70
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:1a9a19df17b464d1708b2905d03b8726e4480ca1f11aa97a15ff4c049a24ea4b
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:ea3e5fbf7eaf6bd29031e2876d7e24859779fb4d3d7ef27c26a1d310f137ad17
        version: 2.1.1
      entries:
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.1
    - currentCSV: kernel-module-management.v2.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a224708950f8defb259ef0da49011d361c5c0378404350640600c6e8fbdbd606
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.2.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:b1d98c0f14448ae8eb80821cbec236150c338674aae2a2b37f8d0ab0d5f1430c
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:53fddf0b226eb579c71c86112a8d9b6b1d90a1498d6f6985932b16b8be532f35
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:c3dc9ad4ebc68cd85a074b520903a9310e72a90a73bdf4aa7c42b7ed66b7bcba
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:ce1d68fd96f5003285f9fbc0c7abd8aa23830dbdd9b76629d38d4d2238e30032
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a224708950f8defb259ef0da49011d361c5c0378404350640600c6e8fbdbd606
        version: 2.2.1
      entries:
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.2
    - currentCSV: kernel-module-management.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:750dcc75f908cdda86932386fa1a28edaf65d7b5379ef6d0447586a064db906f
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.3.0'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:750dcc75f908cdda86932386fa1a28edaf65d7b5379ef6d0447586a064db906f
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:c5005b74402f680d11c9ea682a9a0d3f74b6ab0d01d25298d232f910f1a29c34
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:055c2d6e80bd1ec225f0cb2041572ebb28cfd1a4ed82ba47ff0d638d52ded2a8
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:f9fde334af68220943017a7491130e6a82f15e63522f5bb3931aa7d9b00ccc40
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:06ac089cc89b303e525fd71d6439b958f34ab944a72209f1cc8cb562ecfd1112
        version: 2.3.0
      entries:
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.3
    - currentCSV: kernel-module-management.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:41Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:8e0a3cfb1cb92d6d3c9ab5a3030593d8df1378c3ce755d598fb77b08ea5793d3
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        version: 2.4.1
      entries:
      - name: kernel-module-management.v2.4.1
        version: 2.4.1
      - name: kernel-module-management.v2.4.0
        version: 2.4.0
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.4
    - currentCSV: kernel-module-management.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:41Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:8e0a3cfb1cb92d6d3c9ab5a3030593d8df1378c3ce755d598fb77b08ea5793d3
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        version: 2.4.1
      entries:
      - name: kernel-module-management.v2.4.1
        version: 2.4.1
      - name: kernel-module-management.v2.4.0
        version: 2.4.0
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: kernel-module-management
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ab Initio
      provider-url: ""
    name: abinitio-runtime-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: abinitio-runtime-operator.v4.3.3-1
      currentCSVDesc:
        annotations:
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: abinitio-system
          operators.operatorframework.io/builder: operator-sdk-v1.8.0-ocp
          operators.operatorframework.io/internal-objects: '["cooperatingsystemruntimepoolmetrics.cloud.abinitio.com",
            "cooperatingsystemruntimeprofiles.cloud.abinitio.com"]'
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Defines a *runtime claim*, an interface for reserving a runtime
              object (CoOperatingSystemRuntime) in a runtime pool (CoOperatingSystemRuntimePool)
              in order to launch a job.
            displayName: Co>Operating System Runtime Claim
            kind: CoOperatingSystemRuntimeClaim
            name: cooperatingsystemruntimeclaims.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Pool Metric
            kind: CoOperatingSystemRuntimePoolMetric
            name: cooperatingsystemruntimepoolmetrics.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime pool*, a set of templated runtime objects
              (CoOperatingSystemRuntime) that are available on demand for running
              jobs. A runtime pool's capacity is configurable and can be scheduled
              in response to usage patterns.
            displayName: Co>Operating System Runtime Pool
            kind: CoOperatingSystemRuntimePool
            name: cooperatingsystemruntimepools.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Profile
            kind: CoOperatingSystemRuntimeProfile
            name: cooperatingsystemruntimeprofiles.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime object*, a set of coordinated resources
              used to run a data processing job in parallel across a launcher pod
              and a dynamically scaled set of worker pods. Such a job leverages the
              full power of the Co>Operating System for processing complex enterprise
              data.
            displayName: Co>Operating System Runtime
            kind: CoOperatingSystemRuntime
            name: cooperatingsystemruntimes.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime template*, a reusable specification (CoOperatingSystemRuntimeSpec)
              that can be referenced and overridden by runtime objects as needed,
              making it possible to specify jobs concisely. The runtime objects in
              a runtime pool (CoOperatingSystemRuntimePool) are created from a runtime
              template.
            displayName: Co>Operating System Runtime Template
            kind: CoOperatingSystemRuntimeTemplate
            name: cooperatingsystemruntimetemplates.cloud.abinitio.com
            version: v1
        description: |-
          The Co>Operating System runtime operator brings the power of Ab Initio's Co>Operating
          System to Kubernetes by defining a set of custom API resources for provisioning and
          running data processing jobs.

          The operator is installed along with the following custom resource definitions (CRDs):

          - CoOperatingSystemRuntime
          - CoOperatingSystemRuntimeClaim
          - CoOperatingSystemRuntimePool
          - CoOperatingSystemRuntimeTemplate

          In combination, these resources give you significant flexibility in scheduling, running,
          and monitoring your jobs, which can run in parallel across a dynamically scaled set of
          containers. For reference documentation for these resources, see
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-FOUR-Reference/Co-Operating-System-runtime-operator-API-reference)
          in the Ab Initio Help.

          The following topics describe the basic requirements and procedures for creating custom
          resources for running jobs after you have installed the runtime operator.

          ## Requirements for using custom resources

          The runtime operator is fully configured when you install it in your OpenShift cluster
          through OperatorHub. Importantly, the operator is installed in unrestricted mode,
          which means it can manage the CoOperatingSystemRuntime family of CRDs in any namespace
          in your cluster.

          For creating custom resources in order to run jobs using the operator, you must meet
          the requirements and configure the settings described in the following sections.

          ### Container images for job agents

          To run jobs of kind CoOperatingSystemRuntime, you must have a way to build the following
          OCI container images:

          - A *processing* image with the Co>Operating System server software installed in it.
          - A *sandbox workspace* image with the graphs and other files needed to run your jobs.
          This image is based on the Co>Operating System image and functions as the job's launcher
          agent, which coordinates with a set of worker agents (processing containers) in order to
          run a job in parallel.

          For building and managing these container images, you need the following:

          - A Co>Operating System installation package.
          - Access to a technical repository containing the projects that make up your application.
          - A container-building tool.
          - Access to a private container registry for hosting the images you build. You need pull
          and push access to this registry from your container build context and pull access from
          your OpenShift cluster (for example, via a Kubernetes Secret).

          > WARNING! Never push container images with Ab Initio software to a public registry. Push
          them to a private container registry only.

          The Co>Operating System includes reference Dockerfiles and example scripts for building
          container images. Typically, you build these images as part of your CI/CD pipeline. For
          more information, contact your Ab Initio administrator or Ab Initio Support.

          ### Ab Initio key servers

          The key servers that you use to enable jobs must be reachable from your cluster. You
          specify these key servers via the `AB_HOSTNAME_KEYSERVER_URLS` configuration variable in
          the pod specifications for launcher and worker agents in the CoOperatingSystemRuntime
          resource.

          To discuss the software activation key requirements for jobs that run in Kubernetes,
          contact your key server administrator or Ab Initio Support.

          ### Namespace and role considerations

          The runtime operator is installed with the cluster-wide RBAC roles and role bindings that
          it needs in order to manage custom resources and run jobs.

          You typically use a separate namespace or namespaces for creating and managing custom
          resources to run jobs. Each of these namespaces must be equipped with a service
          account that can list pod limits and get and patch CoOperatingSystemRuntime resources.

          For each namespace in which you intend to create custom resources, create the needed RBAC
          role and role binding by applying the following Role and RoleBinding manifests:

          ```yaml
          apiVersion: rbac.authorization.k8s.io/v1
          kind: Role
          metadata:
            name: cooperatingsystemruntime-role
          rules:
          - apiGroups:
            - ""
            resources:
            - limitranges
            verbs:
            - list
          - apiGroups:
            - cloud.abinitio.com
            resources:
            - cooperatingsystemruntimes
            verbs:
            - get
            - patch
          ---
          apiVersion: rbac.authorization.k8s.io/v1
          kind: RoleBinding
          metadata:
            name: run-cooperatingsystemruntime-rolebinding
          subjects:
          - kind: ServiceAccount
            name: default
          roleRef:
            kind: Role
            name: cooperatingsystemruntime-role
            apiGroup: rbac.authorization.k8s.io
          ```

          One way to do this is to save these manifests to a file and specify a namespace when
          using `oc` to apply them to the cluster. In the following example, the manifests are
          saved to a file named `runtime-rbac.yaml` and the role and role binding are to be created
          for the `default` service account in the `hello-world` namespace:

          ```
          oc -n hello-world apply -f runtime-rbac.yaml
          ```

          To bind the role to a service account other than `default`, simply update the value of
          `subjects[0].name` in the provided RoleBinding manifest.

          ### Dynamic volume provisioning

          The launcher and worker template sections of the CoOperatingSystemRuntime resource
          require you to specify persistent volume claims (PVCs) for job recovery data. To ensure
          that these PVCs can be bound to approriate storage volumes, dynamic volume provisioning
          must be enabled in your cluster.

          Work with your cluster administrator to identify the storage needs of your jobs and to
          ensure that dynamic volume provisioning is enabled so that PVCs can be matched to
          persistent volumes of the right storage class.

          For more information about dynamic storage provisioning, see the
          [OpenShift documentation](https://docs.openshift.com/container-platform/latest/storage/dynamic-provisioning.html).

          ## Quick start and examples

          For an overview of the key concepts associated with running jobs in Kubernetes, see
          ["About running jobs in Kubernetes"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-TWO-Running-jobs-in-Kubernetes/About-running-jobs-in-Kubernetes)
          in the Ab Initio Help.

          As described in this topic, you might begin by creating a CoOperatingSystemRuntime
          manifest that defines the configuration and compute resources needed for a standalone test
          job. You could then use the specification for this job to create a CoOperatingSystemRuntimeTemplate
          resource, which would collect common settings and make it possible to concisely specify
          CoOperatingSystemRuntime jobs.

          You could also use a runtime pool (of kind CoOperatingSystemRuntime) to extend these basic
          resources: such a pool consists of an actively running group of templated runtime objects whose
          pods can be used to run jobs on demand. Finally, for reserving resources in a pool, you could
          create a CoOperatingSystemRuntimeClaim resource.

          For more information about these custom resources, refer to the
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-FOUR-Reference/Co-Operating-System-runtime-operator-API-reference).

          For a set of integrated examples (including complete YAML manifests) that show you how to
          create and use the custom resources, see ["Example manifests for running jobs"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-TWO-Running-jobs-in-Kubernetes/Configuring-and-running-Co-Operating-System-jobs-in-Kubernetes/Example-manifests-for-CoOperatingSystemRuntime-jobs).
        displayName: Co>Operating System Runtime Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - abinitio
        - runtime
        - operator
        links:
        - name: Ab Initio
          url: https://www.abinitio.com/
        maintainers:
        - email: support@abinitio.com
          name: Ab Initio Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Ab Initio
        relatedImages:
        - registry.connect.redhat.com/abinitio/cooperating-system-runtime-operator@sha256:665efa8cc2ba97ad31f219161cdd53e15c1aef16b6c05ac60506be9bbc9ffc9a
        version: 4.3.3-1
      entries:
      - name: abinitio-runtime-operator.v4.3.3-1
        version: 4.3.3-1
      - name: abinitio-runtime-operator.v4.3.1-0
        version: 4.3.1-0
      name: release-4.3
    - currentCSV: abinitio-runtime-operator.v4.2.3-0
      currentCSVDesc:
        annotations:
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: abinitio-system
          operators.operatorframework.io/builder: operator-sdk-v1.8.0-ocp
          operators.operatorframework.io/internal-objects: '["cooperatingsystemruntimepoolmetrics.cloud.abinitio.com",
            "cooperatingsystemruntimeprofiles.cloud.abinitio.com"]'
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Defines a *runtime claim*, an interface for reserving a runtime
              object (CoOperatingSystemRuntime) in a runtime pool (CoOperatingSystemRuntimePool)
              in order to launch a job.
            displayName: Co>Operating System Runtime Claim
            kind: CoOperatingSystemRuntimeClaim
            name: cooperatingsystemruntimeclaims.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Pool Metric
            kind: CoOperatingSystemRuntimePoolMetric
            name: cooperatingsystemruntimepoolmetrics.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime pool*, a set of templated runtime objects
              (CoOperatingSystemRuntime) that are available on demand for running
              jobs. A runtime pool's capacity is configurable and can be scheduled
              in response to usage patterns.
            displayName: Co>Operating System Runtime Pool
            kind: CoOperatingSystemRuntimePool
            name: cooperatingsystemruntimepools.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Profile
            kind: CoOperatingSystemRuntimeProfile
            name: cooperatingsystemruntimeprofiles.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime object*, a set of coordinated resources
              used to run a data processing job in parallel across a launcher pod
              and a dynamically scaled set of worker pods. Such a job leverages the
              full power of the Co>Operating System for processing complex enterprise
              data.
            displayName: Co>Operating System Runtime
            kind: CoOperatingSystemRuntime
            name: cooperatingsystemruntimes.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime template*, a reusable specification (CoOperatingSystemRuntimeSpec)
              that can be referenced and overridden by runtime objects as needed,
              making it possible to specify jobs concisely. The runtime objects in
              a runtime pool (CoOperatingSystemRuntimePool) are created from a runtime
              template.
            displayName: Co>Operating System Runtime Template
            kind: CoOperatingSystemRuntimeTemplate
            name: cooperatingsystemruntimetemplates.cloud.abinitio.com
            version: v1
        description: |-
          The Co>Operating System runtime operator brings the power of Ab Initio's Co>Operating
          System to Kubernetes by defining a set of custom API resources for provisioning and
          running data processing jobs.

          The operator is installed along with the following custom resource definitions (CRDs):

          - CoOperatingSystemRuntime
          - CoOperatingSystemRuntimeClaim
          - CoOperatingSystemRuntimePool
          - CoOperatingSystemRuntimeTemplate

          In combination, these resources give you significant flexibility in scheduling, running,
          and monitoring your jobs, which can run in parallel across a dynamically scaled set of
          containers. For reference documentation for these resources, see
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Co_Operating_System_runtime_operator_API_reference)
          in the Ab Initio Help.

          The following topics describe the basic requirements and procedures for creating custom
          resources for running jobs after you have installed the runtime operator.

          ## Requirements for using custom resources

          The runtime operator is fully configured when you install it in your OpenShift cluster
          through OperatorHub. Importantly, the operator is installed in unrestricted mode,
          which means it can manage the CoOperatingSystemRuntime family of CRDs in any namespace
          in your cluster.

          For creating custom resources in order to run jobs using the operator, you must meet
          the requirements and configure the settings described in the following sections.

          ### Container images for job agents

          To run jobs of kind CoOperatingSystemRuntime, you must have a way to build the following
          OCI container images:

          - A *processing* image with the Co>Operating System server software installed in it.
          - A *sandbox workspace* image with the graphs and other files needed to run your jobs.
          This image is based on the Co>Operating System image and functions as the job's launcher
          agent, which coordinates with a set of worker agents (processing containers) in order to
          run a job in parallel.

          For building and managing these container images, you need the following:

          - A Co>Operating System installation package.
          - Access to a technical repository containing the projects that make up your application.
          - A container-building tool.
          - Access to a private container registry for hosting the images you build. You need pull
          and push access to this registry from your container build context and pull access from
          your OpenShift cluster (for example, via a Kubernetes Secret).

          > WARNING! Never push container images with Ab Initio software to a public registry. Push
          them to a private container registry only.

          The Co>Operating System includes reference Dockerfiles and example scripts for building
          container images. Typically, you build these images as part of your CI/CD pipeline. For
          more information, contact your Ab Initio administrator or Ab Initio Support.

          ### Ab Initio key servers

          The key servers that you use to enable jobs must be reachable from your cluster. You
          specify these key servers via the `AB_HOSTNAME_KEYSERVER_URLS` configuration variable in
          the pod specifications for launcher and worker agents in the CoOperatingSystemRuntime
          resource.

          To discuss the software activation key requirements for jobs that run in Kubernetes,
          contact your key server administrator or Ab Initio Support.

          ### Namespace and role considerations

          The runtime operator is installed with the cluster-wide RBAC roles and role bindings that
          it needs in order to manage custom resources and run jobs.

          You typically use a separate namespace or namespaces for creating and managing custom
          resources to run jobs. Each of these namespaces must be equipped with a service
          account that can list pod limits and get and patch CoOperatingSystemRuntime resources.

          For each namespace in which you intend to create custom resources, create the needed RBAC
          role and role binding by applying the following Role and RoleBinding manifests:

          ```yaml
          apiVersion: rbac.authorization.k8s.io/v1
          kind: Role
          metadata:
            name: cooperatingsystemruntime-role
          rules:
          - apiGroups:
            - ""
            resources:
            - limitranges
            verbs:
            - list
          - apiGroups:
            - cloud.abinitio.com
            resources:
            - cooperatingsystemruntimes
            verbs:
            - get
            - patch
          ---
          apiVersion: rbac.authorization.k8s.io/v1
          kind: RoleBinding
          metadata:
            name: run-cooperatingsystemruntime-rolebinding
          subjects:
          - kind: ServiceAccount
            name: default
          roleRef:
            kind: Role
            name: cooperatingsystemruntime-role
            apiGroup: rbac.authorization.k8s.io
          ```

          One way to do this is to save these manifests to a file and specify a namespace when
          using `oc` to apply them to the cluster. In the following example, the manifests are
          saved to a file named `runtime-rbac.yaml` and the role and role binding are to be created
          for the `default` service account in the `hello-world` namespace:

          ```
          oc -n hello-world apply -f runtime-rbac.yaml
          ```

          To bind the role to a service account other than `default`, simply update the value of
          `subjects[0].name` in the provided RoleBinding manifest.

          ### Dynamic volume provisioning

          The launcher and worker template sections of the CoOperatingSystemRuntime resource
          require you to specify persistent volume claims (PVCs) for job recovery data. To ensure
          that these PVCs can be bound to approriate storage volumes, dynamic volume provisioning
          must be enabled in your cluster.

          Work with your cluster administrator to identify the storage needs of your jobs and to
          ensure that dynamic volume provisioning is enabled so that PVCs can be matched to
          persistent volumes of the right storage class.

          For more information about dynamic storage provisioning, see the
          [OpenShift documentation](https://docs.openshift.com/container-platform/latest/storage/dynamic-provisioning.html).

          ## Quick start and examples

          For an overview of the key concepts associated with running jobs in Kubernetes, see
          ["About running jobs in Kubernetes"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Configuring_and_running_Co_Operating_System_jobs_in_Kubernetes/About_running_jobs_in_Kubernetes)
          in the Ab Initio Help.

          As described in this topic, you might begin by creating a CoOperatingSystemRuntime
          manifest that defines the configuration and compute resources needed for a standalone test
          job. You could then use the specification for this job to create a CoOperatingSystemRuntimeTemplate
          resource, which would collect common settings and make it possible to concisely specify
          CoOperatingSystemRuntime jobs.

          You could also use a runtime pool (of kind CoOperatingSystemRuntime) to extend these basic
          resources: such a pool consists of an actively running group of templated runtime objects whose
          pods can be used to run jobs on demand. Finally, for reserving resources in a pool, you could
          create a CoOperatingSystemRuntimeClaim resource.

          For more information about these custom resources, refer to the
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Co_Operating_System_runtime_operator_API_reference).

          For a set of integrated examples (including complete YAML manifests) that show you how to
          create and use the custom resources, see ["Example manifests for running jobs"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Configuring_and_running_Co_Operating_System_jobs_in_Kubernetes/Creating_custom_resources_for_running_jobs/Example_manifests_for_running_jobs).
        displayName: Co>Operating System Runtime Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - abinitio
        - runtime
        - operator
        links:
        - name: Ab Initio
          url: https://www.abinitio.com/
        maintainers:
        - email: support@abinitio.com
          name: Ab Initio Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Ab Initio
        relatedImages:
        - registry.connect.redhat.com/abinitio/cooperating-system-runtime-operator@sha256:ee30210a153e8b3ba09080737216d881cc6a07e1a888e9684be0d17fd4ff791d
        version: 4.2.3-0
      entries:
      - name: abinitio-runtime-operator.v4.2.3-0
        version: 4.2.3-0
      - name: abinitio-runtime-operator.v4.2.2
        version: 4.2.2
      name: stable
    defaultChannel: release-4.3
    packageName: abinitio-runtime-operator
    provider:
      name: Ab Initio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: recipe
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: recipe.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "Recipe",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "recipe",
                    "app.kubernetes.io/instance": "recipe-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "recipe",
                    "app.kubernetes.io/part-of": "recipe"
                  },
                  "name": "recipe-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          createdAt: "2025-09-12T06:34:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Recipe is the Schema for the recipes API
            displayName: Recipe
            kind: Recipe
            name: recipes.ramendr.openshift.io
            version: v1alpha1
        description: |
          Recipe describes a workflow used for capturing or recovering Kubernetes resources. This can be referred to by a Ramen DRPlacementControl object when kubeObjectProtection is enabled.
        displayName: Recipe
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - recipe
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/recipe
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:af262b0890300183db1766bc392017b0bda723a9ba841dc25e3d928aa413db9a
        version: 4.16.16-rhodf
      entries:
      - name: recipe.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: recipe.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: recipe.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: recipe.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: recipe.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: recipe.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: recipe.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: recipe.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: recipe.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: recipe.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: recipe.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: recipe.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: recipe.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: recipe.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: recipe.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: recipe.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: recipe.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: recipe
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: New England Research Cloud
      provider-url: https://nerc.mghpcc.org/
    name: keycloak-permissions-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keycloak-permissions-operator.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.nerc.mghpcc.org/v1",
                "kind": "KeycloakAuthorization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keycloak-permissions-operator",
                    "app.kubernetes.io/instance": "keycloakauthorization-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "keycloakauthorization",
                    "app.kubernetes.io/part-of": "keycloak-permissions-operator"
                  },
                  "name": "example"
                },
                "spec": {
                  "cleanupUndefinedAuthorizations": false,
                  "description": "",
                  "keycloak": {
                    "adminSecret": {
                      "name": "keycloak-initial-admin",
                      "passwordKey": "password",
                      "usernameKey": "username"
                    }
                  },
                  "managementState": "Managed",
                  "realm": null,
                  "validateCerts": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          com.redhat.openshift.versions: v4.11
          containerImage: quay.io/nerc-images/keycloak-permissions-operator:1.4.1
          createdAt: "2024-05-07T21:09:02Z"
          description: Managing Keycloak resources, scopes, policies, and permissions
            for fine-grained resource permissions
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/nerc-images/keycloak-permissions-operator
          support: New England Research Cloud
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Keycloak resources, scopes, policies, and permissions
              for fine-grained resource permissions.
            displayName: Keycloak Authorization
            kind: KeycloakAuthorization
            name: keycloakauthorizations.keycloak.nerc.mghpcc.org
            version: v1
        description: "An OpenShift Operator for managing Keycloak resources, scopes,
          policies,\nand permissions for fine-grained resource permissions.\nAccording
          to the [Keycloak Authorization Services documentation](https://www.keycloak.org/docs/23.0.7/authorization_services/):\n\n-
          A resource can be a web page, a RESTFul resource, a file in your file system,
          or other entity you wish to authorize access to.\n- Scopes usually represent
          the actions that can be performed on a resource, but they are not limited
          to that. You can also use scopes to represent one or more attributes within
          a resource.\n- Policies define the conditions that must be satisfied to
          access or perform operations on something (resource or scope), but they
          are not tied to what they are protecting. They are generic and can be reused
          to build permissions or even more complex policies.\n- Permissions are coupled
          with the resource they are protecting. Here you specify what you want to
          protect (resource or scope) and the policies that must be satisfied to grant
          or deny permission.\n\n## Getting started with the Keycloak Permissions
          Operator\n\nStart by defining the name of your KeycloakAuthorization and
          the same namespace where Keycloak is deployed.\n\n```yaml\napiVersion: keycloak.nerc.mghpcc.org/v1\nkind:
          KeycloakAuthorization\nmetadata:\n  name: example\n  namespace: keycloak\n```\n\n##
          KeycloakAuthorization spec\n\nNext, define the spec. \n\n- `validateCerts:
          false` If your Keycloak does not have valid certificates, if you are deploying
          to OpenShift Local for example. You don't need this line if Keycloak has
          valid certificates.\n- `cleanupUndefinedAuthorizations: true` if you would
          like to remove any undefined authorizations that may have been added manually.
          This is useful for keeping in sync with GitOps. If you do not want the operator
          removing any resources, policies, or permissions, then remove this line.\n\n```yaml\nspec:\n
          \ validateCerts: false\n  cleanupUndefinedAuthorizations: true\n```\n\n##
          Keycloak admin credentials\n\nSpecify the details to connect to keycloak,
          including the `baseUrl`, and the Secret name containing the Keycloak admin
          usernameKey and passwordKey.\n\n```yaml\n  keycloak:\n    baseUrl: https://keycloak-service:8443\n
          \   adminSecret:\n      name: keycloak-initial-admin\n      usernameKey:
          username\n      passwordKey: password\n```\n\n## Realm\n\nSpecify the realm
          id you wish to add authorization configuration to. If you want to add authorization
          configuration to multiple realms, then create a KeycloakAuthorization resource
          for each realm.\n\n```yaml\n  realm:\n    id: NERC\n```\n\n## Clients\n\nFor
          each client in the realm you wish to configure, add authorizationSettings.\n\n```yaml\n
          \   clients:\n      - id: ai4cloudops\n        authorizationSettings:\n```\n\n##
          Authorization Scopes\n\nList the scopes you wish to define for the client.\n\n```yaml\n
          \         scopes:\n            - id: prod\n              name: prod\n              displayName:
          Production\n            - id: test\n              name: test\n              displayName:
          Test\n```\n\n## Authorization Resources\n\nList the resources and corresponding
          scopes you wish to define for the client.\n\n```yaml\n          resources:\n
          \           - name: cluster\n              displayName: cluster\n              scopes:\n
          \               - id: prod\n                  name: prod\n                  displayName:
          Production\n                - id: test\n                  name: test\n                  displayName:
          Test\n```\n\n## Authorization Policies\n\nDefine the policies you wish to
          use in your permissions. You can choose from all the different types of
          permissions, including: aggregated, client, client-scope, group, regex,
          role, time, and user.\n\n```yaml\n          policies:\n            - id:
          client-example\n              name: client-example\n              logic:
          POSITIVE\n              decisionStrategy: UNANIMOUS\n              type:\n
          \               client:\n                  clients:\n                    -
          example\n            - id: group-example\n              name: group-example\n
          \             logic: POSITIVE\n              decisionStrategy: UNANIMOUS\n
          \             type:\n                group:\n                  groups:\n
          \                   - id: example\n```\n\n## Authorization Permissions\n\nDefine
          the permissions, connecting the resources, scopes, and policies together.\n\n```yaml\n
          \         permissions:\n            - name: group-example-cluster-test\n
          \             client: test\n              policy: group-example\n              resource:
          cluster\n            - name: client-example-cluster-test\n              client:
          example\n              policy: client-test\n              resource: cluster\n
          \             scope: test\n```\n\n## Other details\n\nResources, scopes,
          policies, and permissions will all be applied to Keycloak if you have configured
          the admin credentials to Keycloak successfully.\nWhen the KeycloakAuthorization
          is deleted, then the configured resources, policies, and permissions will
          be deleted as well. "
        displayName: Keycloak Permissions Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - fine-grained resource permissions
        - keycloak
        - resources
        - scopes
        - policies
        - permissions
        links:
        - name: Keycloak Permissions Operator
          url: https://keycloak-permissions-operator.domain
        maintainers:
        - email: computate@computate.org
          name: Christopher Tate
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: New England Research Cloud
          url: https://nerc.mghpcc.org/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/nerc-images/keycloak-permissions-operator:1.4.1
        version: 1.4.1
      entries:
      - name: keycloak-permissions-operator.v1.4.1
        version: 1.4.1
      - name: keycloak-permissions-operator.v1.3.0
        version: 1.3.0
      - name: keycloak-permissions-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: keycloak-permissions-operator
    provider:
      name: New England Research Cloud
      url: https://nerc.mghpcc.org/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: mongodb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"opstreelabs.in/v1alpha1","kind":"MongoDB","metadata":{"name":"mongodb"},"spec":{"kubernetesConfig":{"image":"quay.io/opstree/mongo:v5.0.6","imagePullPolicy":"IfNotPresent","securityContext":{"fsGroup":1001}},"storage":{"accessModes":["ReadWriteOnce"],"storageSize":"1Gi","storageClass":"gp2"},"mongoDBSecurity":{"mongoDBAdminUser":"admin","secretRef":{"name":"mongodb-secret","key":"password"}}}},{"apiVersion":"opstreelabs.in/v1alpha1","kind":"MongoDBCluster","metadata":{"name":"mongodb"},"spec":{"clusterSize":3,"kubernetesConfig":{"image":"quay.io/opstree/mongo:v5.0.6","imagePullPolicy":"IfNotPresent","securityContext":{"fsGroup":1001}},"storage":{"accessModes":["ReadWriteOnce"],"storageSize":"1Gi","storageClass":"gp2"},"mongoDBSecurity":{"mongoDBAdminUser":"admin","secretRef":{"name":"mongodb-secret","key":"password"}}}}]'
          capabilities: Seamless Upgrades
          categories: Database,  OpenShift Optional, Storage
          certified: "false"
          containerImage: quay.io/opstree/mongodb-operator:v0.3.0
          createdAt: 05-04-2022
          description: A golang based operator to create and manage MongoDB standalone
            setup and cluster setup
          k8sMaxVersion: ""
          k8sMinVersion: 1.17.0
          repository: https://github.com/ot-container-kit/mongodb-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Mongo DB Cluster
            displayName: MongoDBCluster
            kind: MongoDBCluster
            name: mongodbclusters.opstreelabs.in
            version: v1alpha1
          - description: Mongo DB
            displayName: MongoDB
            kind: MongoDB
            name: mongodbs.opstreelabs.in
            version: v1alpha1
        description: |+
          MongoDB Operator is an operator created in Golang to create, update, and manage MongoDB standalone, replicated, and arbiter replicated setup on Kubernetes and Openshift clusters. This operator is capable of doing the setup for MongoDB with all the required best practices.

          ### Supported Features

          * MongoDB replicated cluster setup
          * MongoDB standalone setup
          * MongoDB replicated cluster failover and recovery
          * Monitoring support with MongoDB Exporter
          * Password based authentication for MongoDB
          * Kubernetes's resources for MongoDB standalone and cluster
          * Customizable configuration changes in MongoDB
          * Insightful Grafana dashboards

          ### Documentation

          **[Documentation](https://ot-mongodb-operator.netlify.app/)**
          ### Getting Help

          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.
          ### Contributing
          You can contribute to this project by:-

          * Raising Bugs and Feature Requests
          * Fixing issue by opening a Pull Request
          * Improving Documentation

          ### License

          MongoDB  Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/mongodb-operator/blob/master/LICENSE)


        displayName: MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - MongoDB
        - Opstree
        - Operatork
        - Cloud Native
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-mongodb-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: alpha
        minKubeVersion: 1.17.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/opstree/mongodb-operator:v0.3.0
        version: 0.3.0
      entries:
      - name: mongodb-operator.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: mongodb-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: https://konveyor.io
    name: move2kube-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: move2kube-operator.v0.3.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {
                  "deployment": {
                    "api": {
                      "cpu": "300m",
                      "imageTag": "",
                      "initContainer": false,
                      "memory": "1Gi",
                      "privilegedPods": false,
                      "replicas": 1
                    },
                    "authServer": {
                      "cpu": "300m",
                      "enable": false,
                      "memory": "1Gi",
                      "replicas": 1
                    },
                    "database": {
                      "cpu": "300m",
                      "enable": false,
                      "memory": "1Gi",
                      "replicas": 1,
                      "startEmpty": true
                    }
                  },
                  "ingress": {
                    "enable": true,
                    "host": "mydomain.com",
                    "preferRoute": false,
                    "tlsSecretName": ""
                  },
                  "persistentVolumeClaim": {
                    "createNew": true,
                    "enable": false,
                    "name": "",
                    "readWriteMany": true,
                    "storageClassName": "",
                    "storageSize": "1Gi"
                  },
                  "secret": {
                    "api": {
                      "configYAML": "",
                      "createNew": true,
                      "enable": false,
                      "name": ""
                    },
                    "authServer": {
                      "adminPassword": "password",
                      "adminUsername": "admin",
                      "createNew": true,
                      "databasePassword": "password",
                      "databaseUsername": "auth-server",
                      "name": "",
                      "realmJSON": "",
                      "standaloneHAXML": ""
                    }
                  },
                  "securityContextContraints": {
                    "enable": false
                  },
                  "serviceAccount": {
                    "createNew": true,
                    "enable": false,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.10
          createdAt: "2023-09-25T02:56:15Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/konveyor/move2kube-operator:v0.3.10
        version: 0.3.10
      entries:
      - name: move2kube-operator.v0.3.10
        version: 0.3.10
      - name: move2kube-operator.v0.3.10-rc.0
        version: 0.3.10-rc.0
      - name: move2kube-operator.v0.3.9
        version: 0.3.9
      - name: move2kube-operator.v0.3.8
        version: 0.3.8
      name: alpha
    - currentCSV: move2kube-operator.v0.3.15-rc.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.15-rc.0
          createdAt: "2024-10-18T16:10:56Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/konveyor/move2kube-operator:v0.3.15-rc.0
        version: 0.3.15-rc.0
      entries:
      - name: move2kube-operator.v0.3.15-rc.0
        version: 0.3.15-rc.0
      - name: move2kube-operator.v0.3.14-rc.0
        version: 0.3.14-rc.0
      - name: move2kube-operator.v0.3.13-rc.0
        version: 0.3.13-rc.0
      - name: move2kube-operator.v0.3.11-rc.0
        version: 0.3.11-rc.0
      name: prerelease
    - currentCSV: move2kube-operator.v0.3.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.15
          createdAt: "2025-03-07T06:09:48Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/konveyor/move2kube-operator:v0.3.15
        version: 0.3.15
      entries:
      - name: move2kube-operator.v0.3.15
        version: 0.3.15
      - name: move2kube-operator.v0.3.14
        version: 0.3.14
      - name: move2kube-operator.v0.3.13
        version: 0.3.13
      - name: move2kube-operator.v0.3.11
        version: 0.3.11
      name: stable
    defaultChannel: stable
    packageName: move2kube-operator
    provider:
      name: Konveyor
      url: https://konveyor.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Dell Technologies
      provider-url: https://github.com/dell/csm-operator
    name: dell-csm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dell-csm-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "authorization",
                  "namespace": "authorization"
                },
                "spec": {
                  "modules": [
                    {
                      "components": [
                        {
                          "enabled": true,
                          "name": "nginx"
                        },
                        {
                          "enabled": true,
                          "name": "cert-manager"
                        },
                        {
                          "authorizationController": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231",
                          "authorizationControllerReplicas": 1,
                          "certificate": "",
                          "controllerReconcileInterval": "5m",
                          "enabled": true,
                          "hostname": "csm-authorization.com",
                          "leaderElection": true,
                          "name": "proxy-server",
                          "opa": "docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095",
                          "opaKubeMgmt": "docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517",
                          "openTelemetryCollectorAddress": "",
                          "privateKey": "",
                          "proxyServerIngress": [
                            {
                              "annotations": {},
                              "hosts": [],
                              "ingressClassName": "nginx"
                            }
                          ],
                          "proxyService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682",
                          "proxyServiceReplicas": 1,
                          "roleService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c",
                          "roleServiceReplicas": 1,
                          "storageService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5",
                          "storageServiceReplicas": 1,
                          "tenantService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358",
                          "tenantServiceReplicas": 1
                        },
                        {
                          "commander": "docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a",
                          "name": "redis",
                          "redis": "docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f",
                          "redisCommander": "rediscommander",
                          "redisName": "redis-csm",
                          "redisReplicas": 5,
                          "redisSecretProviderClass": [
                            {
                              "redisPasswordKey": "",
                              "redisSecretName": "",
                              "redisUsernameKey": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "sentinel": "sentinel"
                        },
                        {
                          "configSecretProviderClass": [
                            {
                              "configSecretName": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "name": "config"
                        },
                        {
                          "name": "storage-system-credentials",
                          "secretProviderClasses": {
                            "vault": [
                              "secret-provider-class-1",
                              "secret-provider-class-2"
                            ]
                          }
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": true,
                      "forceRemoveModule": true,
                      "name": "authorization-proxy-server"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "isilon",
                  "namespace": "isilon"
                },
                "spec": {
                  "driver": {
                    "authSecret": "isilon-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VERBOSE",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_ISI_PORT",
                          "value": "8080"
                        },
                        {
                          "name": "X_CSI_ISI_PATH",
                          "value": "/ifs/data/csi"
                        },
                        {
                          "name": "X_CSI_ISI_NO_PROBE_ON_START",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_AUTOPROBE",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_AUTH_TYPE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_CUSTOM_TOPOLOGY_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOISILON_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "AZ_RECONCILE_INTERVAL",
                          "value": "1h"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_ISI_QUOTA_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_ACCESS_ZONE",
                          "value": "System"
                        },
                        {
                          "name": "X_CSI_ISI_VOLUME_PATH_PERMISSIONS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_IGNORE_UNRESOLVABLE_HOSTS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "isilon",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerscale"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSCALE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_PERFORMANCE_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSCALE_QUOTA_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "ISICLIENT_INSECURE",
                              "value": "true"
                            },
                            {
                              "name": "ISICLIENT_AUTH_TYPE",
                              "value": "1"
                            },
                            {
                              "name": "ISICLIENT_VERBOSE",
                              "value": "0"
                            },
                            {
                              "name": "POWERSCALE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSCALE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0",
                          "name": "metrics-powerscale"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-isilon/csi_sock",
                            "--mode=node",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powermax",
                  "namespace": "powermax"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powermax-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_MANAGED_ARRAYS",
                          "value": "000000000001,000000000002"
                        },
                        {
                          "name": "X_CSI_POWERMAX_PORTGROUPS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_TRANSPORT_PROTOCOL",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_K8S_CLUSTER_PREFIX",
                          "value": "CSM"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "X_CSI_VSPHERE_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VSPHERE_PORTGROUP",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VSPHERE_HOSTNAME",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VCENTER_HOST",
                          "value": ""
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "X_CSI_POWERMAX_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "X_CSI_POWERMAX_DEBUG",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_REVPROXY_USE_SECRET",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powermax",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERMAX_ISCSI_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_TOPOLOGY_CONTROL_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": [
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/memory-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/disk-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/network-unavailable",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REVPROXY_TLS_SECRET",
                              "value": "csirevproxy-tls-secret"
                            },
                            {
                              "name": "X_CSI_REVPROXY_PORT",
                              "value": "2222"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            },
                            {
                              "name": "DeployAsSidecar",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265",
                          "name": "csipowermax-reverseproxy"
                        }
                      ],
                      "configVersion": "v2.14.0",
                      "name": "csireverseproxy"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powermax"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERMAX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERMAX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414",
                          "name": "metrics-powermax"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/powermax.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powerstore",
                  "namespace": "powerstore"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powerstore-config",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_NODE_NAME_PREFIX",
                          "value": "csi-node"
                        },
                        {
                          "name": "X_CSI_FC_PORTS_FILTER_FILE_PATH",
                          "value": "/etc/fc-ports-filter"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOPOWERSTORE_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSM_AUTH_ENABLED",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_MULTI_NAS_FAILURE_THRESHOLD",
                          "value": "5"
                        },
                        {
                          "name": "X_CSI_MULTI_NAS_COOLDOWN_PERIOD",
                          "value": "5m"
                        },
                        {
                          "name": "X_CSI_NFS_ACLS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerstore",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-powerstore.dellemc.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerstore"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSTORE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_SPACE_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_ARRAY_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_FILE_SYSTEM_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSTORE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSTORE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "ZIPKIN_URI",
                              "value": ""
                            },
                            {
                              "name": "ZIPKIN_SERVICE_NAME",
                              "value": "metrics-powerstore"
                            },
                            {
                              "name": "ZIPKIN_PROBABILITY",
                              "value": "0.0"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60",
                          "name": "metrics-powerstore"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "unity",
                  "namespace": "unity"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_UNITY_ALLOW_MULTI_POD_ACCESS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_EPHEMERAL_STAGING_PATH",
                          "value": "/var/lib/kubelet/plugins/kubernetes.io/csi/pv/"
                        },
                        {
                          "name": "X_CSI_ISCSI_CHROOT",
                          "value": "/noderoot"
                        },
                        {
                          "name": "X_CSI_UNITY_SYNC_NODEINFO_INTERVAL",
                          "value": "15"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOUNITY_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOUNITY_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "TENANT_NAME",
                          "value": ""
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_UNITY_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "unity",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "vxflexos",
                  "namespace": "vxflexos"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLELISTVOLUMESNAPSHOT",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLESNAPSHOTCGDELETE",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "INFO"
                        },
                        {
                          "name": "GOSCALEIO_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOSCALEIO_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_QUOTA_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "INTERFACE_NAMES",
                          "value": null
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERFLEX_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "File",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerflex",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "initContainers": [
                      {
                        "envs": [
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          },
                          {
                            "name": "REPO_ADDRESS",
                            "value": "sftp://0.0.0.0"
                          },
                          {
                            "name": "REPO_USER",
                            "value": "sdcRepoUser"
                          },
                          {
                            "name": "MODULE_SIGCHECK",
                            "value": "0"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "imagePullPolicy": "IfNotPresent",
                        "name": "sdc"
                      }
                    ],
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_SDC_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_APPROVE_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_PREFIX",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_SDC_SFTP_REPO_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 1,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "enabled": false,
                        "envs": [
                          {
                            "name": "HOST_PID",
                            "value": "1"
                          },
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "name": "sdc-monitor"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "csi-external-health-monitor-controller"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERFLEX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_SDC_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_SDC_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERFLEX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d",
                          "name": "metrics-powerflex"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerflex"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--csisock=unix:/var/lib/kubelet/plugins/vxflexos.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
          createdAt: "2025-09-05T13:18:42Z"
          description: Easily install and manage Dell’s CSI Drivers and CSM
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/dell/csm-operator
          support: Dell Technologies
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ContainerStorageModule is the Schema for the containerstoragemodules
              API
            displayName: Container Storage Module
            kind: ContainerStorageModule
            name: containerstoragemodules.storage.dell.com
            version: v1
        description: "Dell Container Storage Modules (CSM) Operator is a Kubernetes
          Operator which can be \nused to install and manage Dell’s CSI drivers and
          CSM modules. \nBy using Dell CSM Operator, enterprises can quickly and easily
          deploy the \nCSM modules for CSI drivers making it easy for DevOps \nteams
          to build and optimize their Kubernetes workloads using enterprise grade
          storage \nwhile consistently aligning with the desired outcomes of the organization.\nThis
          operator is available as both community and certified operator for Openshift
          \nand as a community operator for upstream Kubernetes.\nOperator can be
          deployed via OperatorHub.io. using Operator Lifecycle Manager (OLM)\n\n###
          About Dell CSM\n\nDell CSM brings powerful enterprise storage features and
          functionality to Kubernetes for \neasier adoption of cloud native workloads,
          improved productivity and scalable operations. \n\nCSM delivers software
          modules for storage management that provide developers access to \nbuild
          automation for enhanced IT needs and other critical enterprise storage features.
          \n\nCurrently, the following modules are available as part of Dell CSM.
          \n*\tObservability - Delivers a high-level view of the storage capacity
          \n  and performance usage via Grafana dashboards to the Kubernetes users\n*\tReplication
          - Enables array replication capabilities for Kubernetes users\n*\tAuthorization
          - Provides both storage and Kubernetes administrators the ability \n  to
          apply RBAC for Dell CSI Drivers\n*\tResiliency - Designed to make Kubernetes
          applications more resilient to \n  node failures\n\nThese modules are available
          for various storage platforms in the Dell portfolio, \nsuch as Dell PowerStore,
          Dell PowerMax, Dell PowerFlex, Dell PowerScale, and Dell Unity XT.\n\nFor
          more information, visit [Dell’s Container Storage Module web page](https://www.dell.com/en-us/dt/solutions/kubernetes-containers/kubernetes-storage.htm)\n\nAdditional
          Materials\n*\t[Documentation](https://dell.github.io/csm-docs)\n*\t[CSM
          Solution Brief](https://www.delltechnologies.com/asset/en-us/products/storage/briefs-summaries/csm-solution-brief.pdf)\n"
        displayName: Dell Container Storage Modules
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dell Container Storage Modules
        - Dell CSI Driver
        - Dell CSM Modules
        - Powerflex
        - Powerscale
        - Powerstore
        - Unity
        - Authorization
        - Observability
        - Replication
        links:
        - name: Documentation
          url: https://dell.github.io/csm-docs
        maintainers:
        - email: container.storage.modules@dell.com
          name: Dell Container Storage Modules
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Dell Technologies
          url: https://github.com/dell/csm-operator
        relatedImages:
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5
        - registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334
        - docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517
        - registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
        - registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:59335acaca515fceed5f0350c46cfcfd6037475f0cf35ae8ab41d37333039e2e
        - registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d
        - registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60
        - ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0
        - registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265
        - registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414
        - docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f
        - quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540
        - docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231
        - registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16
        - registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        version: 1.10.0
      entries:
      - name: dell-csm-operator.v1.10.0
        version: 1.10.0
      - name: dell-csm-operator.v1.9.1
        version: 1.9.1
      - name: dell-csm-operator.v1.9.0
        version: 1.9.0
      - name: dell-csm-operator.v1.8.1
        version: 1.8.1
      - name: dell-csm-operator.v1.8.0
        version: 1.8.0
      - name: dell-csm-operator.v1.7.0
        version: 1.7.0
      - name: dell-csm-operator.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: dell-csm-operator
    provider:
      name: Dell Technologies
      url: https://github.com/dell/csm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: raftt.io
      provider-url: https:///www.raftt.io
    name: raftt-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: raftt-operator.v1.25.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "raftt.io/v1beta1",
                "kind": "Raftt",
                "metadata": {
                  "name": "raftt",
                  "namespace": "raftt"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/raftt/raftt-operator@sha256:73edc6d8416ee3815d7e59bb26c33e3d35d97cf1ceadd8fa457041be13e175ca
          createdAt: "2025-04-29T11:31:02Z"
          description: This operator installs and configures a Raftt instance
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/raftt-operator/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/raftt-operator/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: raftt
          operators.openshift.io/valid-subscription: Raftt commercial license
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Raftt systems
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Raftt represents an installation of Raftt cluster controller
              & supporting services
            displayName: Raftt
            kind: Raftt
            name: raftts.raftt.io
            version: v1beta1
        description: |
          Raftt allows fast code iterations and debugging over complex Kubernetes environments.
          Raftt provides a localhost development experience for applications running on Kubernetes.
          You can use Raftt to orchestrate the environments for you, so you get the optimal development experience -
          iterations that take under a second, OOTB debugging of your code as it runs in the cluster, easy introspection, and more.
        displayName: Raftt
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dx
        - development
        - debugging
        links:
        - name: Raftt Operator
          url: https://raftt-operator.domain
        - name: Documentation
          url: https://docs.raftt.io
        maintainers:
        - email: ops@raftt.io
          name: Raftt
        maturity: alpha
        provider:
          name: raftt.io
          url: https:///www.raftt.io
        relatedImages:
        - quay.io/raftt/raftt-operator@sha256:73edc6d8416ee3815d7e59bb26c33e3d35d97cf1ceadd8fa457041be13e175ca
        version: 1.25.2
      entries:
      - name: raftt-operator.v1.25.2
        version: 1.25.2
      name: alpha
    defaultChannel: alpha
    packageName: raftt-operator
    provider:
      name: raftt.io
      url: https:///www.raftt.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HashiCorp
      provider-url: https://www.hashicorp.com/
    name: vault-secrets-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vault-secrets-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultConnection",
                "metadata": {
                  "name": "vaultconnection-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "address": "http://vault.vault.svc.cluster.local:8200"
                }
              },
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultAuth",
                "metadata": {
                  "name": "vaultauth-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "vaultConnectionRef": "vaultconnection-sample",
                  "method": "kubernetes",
                  "mount": "kubernetes",
                  "kubernetes": {
                    "role": "sample",
                    "serviceAccount": "default"
                  }
                }
              },
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultStaticSecret",
                "metadata": {
                  "name": "vaultstaticsecret-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "vaultAuthRef": "vaultauth-sample",
                  "mount": "kvv2",
                  "type": "kv-v2",
                  "path": "secret",
                  "refreshAfter": "5s",
                  "destination": {
                    "create": true,
                    "name": "secret1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.connect.redhat.com/hashicorp/vault-secrets-operator@sha256:84ac7003c4cb90c35f1e296585bb58187a86d64f972133e2661839bdb532ddb8
          createdAt: "2025-09-26T21:50:51Z"
          description: The Vault Secrets Operator (VSO) allows Pods to consume Vault
            secrets natively from Kubernetes Secrets.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/hashicorp/vault-secrets-operator
          support: HashiCorp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CSISecrets
            name: csisecrets.secrets.hashicorp.com
            version: v1beta1
          - description: HCPAuth is the Schema for the hcpauths API
            displayName: HCPAuth
            kind: HCPAuth
            name: hcpauths.secrets.hashicorp.com
            version: v1beta1
          - description: HCPVaultSecretsApp is the Schema for the hcpvaultsecretsapps
              API
            displayName: HCPVault Secrets App
            kind: HCPVaultSecretsApp
            name: hcpvaultsecretsapps.secrets.hashicorp.com
            version: v1beta1
          - description: SecretTransformation is the Schema for the secrettransformations
              API
            displayName: Secret Transformation
            kind: SecretTransformation
            name: secrettransformations.secrets.hashicorp.com
            version: v1beta1
          - description: VaultAuthGlobal is the Schema for the vaultauthglobals API
            displayName: Vault Auth Global
            kind: VaultAuthGlobal
            name: vaultauthglobals.secrets.hashicorp.com
            version: v1beta1
          - description: VaultAuth is the Schema for the vaultauths API
            displayName: Vault Auth
            kind: VaultAuth
            name: vaultauths.secrets.hashicorp.com
            version: v1beta1
          - description: VaultConnection is the Schema for the vaultconnections API
            displayName: Vault Connection
            kind: VaultConnection
            name: vaultconnections.secrets.hashicorp.com
            version: v1beta1
          - description: VaultDynamicSecret is the Schema for the vaultdynamicsecrets
              API
            displayName: Vault Dynamic Secret
            kind: VaultDynamicSecret
            name: vaultdynamicsecrets.secrets.hashicorp.com
            version: v1beta1
          - description: VaultPKISecret is the Schema for the vaultpkisecrets API
            displayName: Vault PKISecret
            kind: VaultPKISecret
            name: vaultpkisecrets.secrets.hashicorp.com
            version: v1beta1
          - description: VaultStaticSecret is the Schema for the vaultstaticsecrets
              API
            displayName: Vault Static Secret
            kind: VaultStaticSecret
            name: vaultstaticsecrets.secrets.hashicorp.com
            version: v1beta1
        description: |-
          The Vault Secrets Operator (VSO) allows Pods to consume Vault secrets
          natively from Kubernetes Secrets.

          ## Overview

          The Vault Secrets Operator operates by watching for changes to its supported set of Custom Resource Definitions (CRD).
          Each CRD provides the specification required to allow the *Operator* to synchronize a Vault Secrets to a Kubernetes Secret.
          The *Operator* writes the *source* Vault secret data directly to the *destination* Kubernetes Secret, ensuring that any
          changes made to the *source* are replicated to the *destination* over its lifetime. In this way, an application only needs
          to have access to the *destination* secret in order to make use of the secret data contained within.

          See the developer docs for more info [here](https://developer.hashicorp.com/vault/docs/platform/k8s/vso), including
          [examples](https://developer.hashicorp.com/vault/docs/platform/k8s/vso/examples) and a
          [tutorial](https://developer.hashicorp.com/vault/tutorials/kubernetes/vault-secrets-operator)
          for getting started.
        displayName: Vault Secrets Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vault
        - security
        - secrets
        links:
        - name: Vault Secrets Operator
          url: https://github.com/hashicorp/vault-secrets-operator
        - name: Documentation
          url: https://developer.hashicorp.com/vault/docs/platform/k8s/vso
        maintainers:
        - email: support@hashicorp.com
          name: HashiCorp Support
        maturity: stable
        provider:
          name: HashiCorp
          url: https://www.hashicorp.com/
        relatedImages:
        - registry.connect.redhat.com/hashicorp/vault-secrets-operator@sha256:84ac7003c4cb90c35f1e296585bb58187a86d64f972133e2661839bdb532ddb8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:503d5ee0fbcd5bd3492785f4246e19cbdb92960be0d05f3c6f123a32fe0fe84d
        version: 1.0.1
      entries:
      - name: vault-secrets-operator.v1.0.1
        version: 1.0.1
      - name: vault-secrets-operator.v1.0.0
        version: 1.0.0
      - name: vault-secrets-operator.v0.10.0
        version: 0.10.0
      - name: vault-secrets-operator.v0.9.1
        version: 0.9.1
      - name: vault-secrets-operator.v0.9.0
        version: 0.9.0
      - name: vault-secrets-operator.v0.8.1
        version: 0.8.1
      - name: vault-secrets-operator.v0.8.0
        version: 0.8.0
      - name: vault-secrets-operator.v0.7.0
        version: 0.7.0
      - name: vault-secrets-operator.v0.6.0
        version: 0.6.0
      - name: vault-secrets-operator.v0.5.2
        version: 0.5.2
      - name: vault-secrets-operator.v0.5.1
        version: 0.5.1
      - name: vault-secrets-operator.v0.5.0
        version: 0.5.0
      - name: vault-secrets-operator.v0.4.3
        version: 0.4.3
      name: stable
    defaultChannel: stable
    packageName: vault-secrets-operator
    provider:
      name: HashiCorp
      url: https://www.hashicorp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microsoft
      provider-url: ""
    name: azure-service-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: azure-service-operator.v1.0.59040
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "APIMgmtAPI",
                "metadata": {
                  "name": "aso-apimgmtapi-sample"
                },
                "spec": {
                  "apiId": "apiId0",
                  "apiService": "apimservice-sample",
                  "location": "westus",
                  "properties": {
                    "apiVersionDescription": "API version description",
                    "description": "API description",
                    "displayName": "aso-apimgmtapi-sample",
                    "isCurrent": true,
                    "isOnline": true,
                    "path": "/api/test",
                    "protocols": [
                      "http"
                    ],
                    "subscriptionRequired": false
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ApimService",
                "metadata": {
                  "name": "apimservice-sample"
                },
                "spec": {
                  "appInsightsName": "appinsights-sample-wmm",
                  "appInsightsResourceGroup": "resourcegroup-azure-operators-wmm",
                  "location": "westus",
                  "publisherEmail": {
                    "Your email": ""
                  },
                  "publisherName": {
                    "Your name": ""
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "tier": "premium",
                  "vnetName": "virtualnetwork-sample",
                  "vnetResourceGroup": "resourcegroup-azure-operators",
                  "vnetSubnetName": "test1",
                  "vnetType": "none"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AppInsights",
                "metadata": {
                  "name": "appinsights-sample"
                },
                "spec": {
                  "applicationType": "other",
                  "kind": "other",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AppInsightsApiKey",
                "metadata": {
                  "name": "key-sample"
                },
                "spec": {
                  "appInsights": "appinsights-sample",
                  "authSDKControlChannel": true,
                  "readTelemetry": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "writeAnnotations": true
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureLoadBalancer",
                "metadata": {
                  "name": "lb-sample-01"
                },
                "spec": {
                  "backendAddressPoolName": "test",
                  "backendPort": 3000,
                  "frontendPortRangeEnd": 2000,
                  "frontendPortRangeStart": 1001,
                  "inboundNatPoolName": "test",
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "lb-pip-sample-01",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureNetworkInterface",
                "metadata": {
                  "name": "nic-sample-01"
                },
                "spec": {
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "nic-pip-sample-01",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnetName": "test1",
                  "vnetName": "vnet-sample-01"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureNetworkInterface",
                "metadata": {
                  "name": "nic-sample-02"
                },
                "spec": {
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "nic-pip-sample-02",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnetName": "test1",
                  "vnetName": "vnet-sample-02"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "lb-pip-sample-01"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "nic-pip-sample-01"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "nic-pip-sample-02"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLManagedUser",
                "metadata": {
                  "name": "azuresqlmanageduser-sample"
                },
                "spec": {
                  "dbName": "azuresqldatabase-sample",
                  "managedIdentityClientId": null,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "db_owner"
                  ],
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLUser",
                "metadata": {
                  "name": "sqluser-sample"
                },
                "spec": {
                  "dbName": "azuresqldatabase-sample1",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "db_owner"
                  ],
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLVNetRule",
                "metadata": {
                  "name": "azuresqlvnetrule-sample"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-vnet",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlAction",
                "metadata": {
                  "name": "azuresqlaction-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverName": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "azuresqldatabase-sample1"
                },
                "spec": {
                  "edition": 0,
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "valueA",
                    "tag2": "valueB"
                  },
                  "name": "azuresqldatabase-sample3"
                },
                "spec": {
                  "edition": 0,
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlFailoverGroup",
                "metadata": {
                  "name": "azuresqlfailovergroup-sample"
                },
                "spec": {
                  "databaselist": [
                    "azuresqldatabase-sample1",
                    "azuresqldatabase-sample2"
                  ],
                  "failovergraceperiod": 30,
                  "failoverpolicy": "Automatic",
                  "location": "eastus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "secondaryserver": "sqlserver-samplesec",
                  "secondaryserverresourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-samplepri"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlFirewallRule",
                "metadata": {
                  "name": "sqlf-allowazuresvcaccess"
                },
                "spec": {
                  "endipaddress": "0.0.0.0",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "startipaddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample-777"
                },
                "spec": {
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample777"
                },
                "spec": {
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVMScaleSet",
                "metadata": {
                  "name": "vmss-sample-01"
                },
                "spec": {
                  "adminUserName": "azureuser7",
                  "backendAddressPoolName": "test",
                  "capacity": 3,
                  "inboundNatPoolName": "test",
                  "loadBalancerName": "lb-sample-01",
                  "location": "SouthCentralUS",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": "{ssh public key}",
                  "subnetName": "test1",
                  "virtualNetworkName": "vnet-sample-01",
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachine",
                "metadata": {
                  "name": "hpfvm1"
                },
                "spec": {
                  "adminUserName": "azureuser",
                  "location": "SouthCentralUS",
                  "networkInterfaceName": "nic-sample-02",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": {
                    "ssh key": ""
                  },
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachine",
                "metadata": {
                  "name": "hpfvm20"
                },
                "spec": {
                  "adminUserName": "azureuser",
                  "location": "SouthCentralUS",
                  "networkInterfaceName": "nic-sample-01",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": "{ssh public key}",
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachineExtension",
                "metadata": {
                  "name": "vmext20a"
                },
                "spec": {
                  "autoUpgradeMinorVersion": true,
                  "forceUpdateTag": "Test2",
                  "location": "SouthCentralUS",
                  "protectedSettings": "{}",
                  "publisher": "Microsoft.Azure.Extensions",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "settings": "{\"commandToExecute\":\"echo 'hello, world'\"}",
                  "typeHandlerVersion": "2.1",
                  "typeName": "CustomScript",
                  "vmName": "vm20"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "BlobContainer",
                "metadata": {
                  "name": "blobcontainer-sample"
                },
                "spec": {
                  "accesslevel": "Container",
                  "accountname": "storageaccountsample777",
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ConsumerGroup",
                "metadata": {
                  "name": "consumergroup-sample-1"
                },
                "spec": {
                  "eventHub": "eventhub-sample-1",
                  "namespace": "eventhubnamespace-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "CosmosDB",
                "metadata": {
                  "name": "cosmosdb-sample-1"
                },
                "spec": {
                  "kind": "GlobalDocumentDB",
                  "location": "westus",
                  "properties": {
                    "databaseAccountOfferType": "Standard",
                    "isVirtualNetworkFilterEnabled": true
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "CosmosDB",
                "metadata": {
                  "labels": {
                    "CosmosAccountType": "Production"
                  },
                  "name": "cosmosdb-sample-2"
                },
                "spec": {
                  "ipRules": [
                    "104.42.195.92,40.76.54.131,52.176.6.30,52.169.50.45,52.187.184.26"
                  ],
                  "kind": "GlobalDocumentDB",
                  "location": "eastus",
                  "locations": [
                    {
                      "failoverPriority": 0,
                      "isZoneRedundant": true,
                      "locationName": "eastus"
                    },
                    {
                      "failoverPriority": 1,
                      "locationName": "westus"
                    }
                  ],
                  "properties": {
                    "databaseAccountOfferType": "Standard",
                    "enableMultipleWriteLocations": true
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "Eventhub",
                "metadata": {
                  "name": "eventhub-sample-1"
                },
                "spec": {
                  "authorizationRule": {
                    "name": "RootManageSharedAccessKey",
                    "rights": [
                      "Listen",
                      "Manage",
                      "Send"
                    ]
                  },
                  "location": "westus",
                  "namespace": "eventhubnamespace-sample-1",
                  "properties": {
                    "messageRetentionInDays": 7,
                    "partitionCount": 2
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "Eventhub",
                "metadata": {
                  "name": "eventhub-sample-2307-10-capture"
                },
                "spec": {
                  "authorizationRule": {
                    "name": "RootManageSharedAccessKey",
                    "rights": [
                      "Listen",
                      "Manage",
                      "Send"
                    ]
                  },
                  "location": "westus",
                  "namespace": "eventhubnamespace-sample-1",
                  "properties": {
                    "captureDescription": {
                      "destination": {
                        "archiveNameFormat": "{Namespace}/{EventHub}/{PartitionId}/{Year}/{Month}/{Day}/{Hour}/{Minute}/{Second}",
                        "blobContainer": "capturecontainer",
                        "name": "EventHubArchive.AzureBlockBlob",
                        "storageAccount": {
                          "accountName": "storageaccountauv1",
                          "resourceGroup": "my-resource-group"
                        }
                      },
                      "enabled": true,
                      "intervalInSeconds": 120,
                      "sizeLimitInBytes": 524288000
                    },
                    "messageRetentionInDays": 7,
                    "partitionCount": 2
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "EventhubNamespace",
                "metadata": {
                  "name": "eventhubnamespace-sample-1"
                },
                "spec": {
                  "location": "westus",
                  "properties": {
                    "isAutoInflateEnabled": false,
                    "kafkaEnabled": false,
                    "maximumThroughputUnits": 0
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "capacity": 1,
                    "name": "Standard",
                    "tier": "Standard"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVault",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "keyvaultsample123"
                },
                "spec": {
                  "accessPolicies": [
                    {
                      "objectID": "\u003cobjectID\u003e",
                      "permissions": {
                        "certificates": [
                          "list",
                          "get"
                        ],
                        "keys": [
                          "list",
                          "get"
                        ],
                        "secrets": [
                          "list",
                          "get"
                        ],
                        "storage": [
                          "list",
                          "get"
                        ]
                      },
                      "tenantID": "\u003ctenantID\u003e"
                    }
                  ],
                  "enableSoftDelete": false,
                  "location": "westus",
                  "networkPolicies": {
                    "bypass": "AzureServices",
                    "defaultAction": "Allow",
                    "ipRules": [
                      "172.16.0.0/24",
                      "172.16.1.0/24"
                    ],
                    "virtualNetworkRules": [
                      "/subscriptions/\u003csubid\u003e/resourceGroups/rg1/providers/Microsoft.Network/virtualNetworks/test-vnet/subnets/subnet1",
                      "/subscriptions/\u003csubid\u003e/resourceGroups/rg1/providers/Microsoft.Network/virtualNetworks/test-vnet/subnets/subnet2"
                    ]
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "name": "standard"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVault",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "keyvaultsample123-simple"
                },
                "spec": {
                  "enableSoftDelete": false,
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVaultKey",
                "metadata": {
                  "name": "keyvaultkey-sample"
                },
                "spec": {
                  "keySize": 4096,
                  "keyVault": "keyvaultsample123",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "type": "RSA"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLDatabase",
                "metadata": {
                  "name": "mysqldatabase-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLDatabase",
                "metadata": {
                  "name": "mysqldatabase-sample2"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLFirewallRule",
                "metadata": {
                  "name": "mysqlfirewallrule-sample"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLFirewallRule",
                "metadata": {
                  "name": "mysqlfirewallrule-sample2"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample2",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-replica"
                },
                "spec": {
                  "createMode": "Replica",
                  "location": "eastus2",
                  "replicaProperties": {
                    "sourceServerId": "/subscriptions/{SUBID}/resourceGroups/resourcegroup-azure-operators/providers/Microsoft.DBforMySQL/servers/mysqlserver-sample"
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-sample2"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLUser",
                "metadata": {
                  "name": "mysqluser-sample"
                },
                "spec": {
                  "dbName": "mysqldatabase-sample",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "SELECT"
                  ],
                  "server": "mysqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLVNetRule",
                "metadata": {
                  "name": "mysqlvnetrule-sample"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-vnet",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLDatabase",
                "metadata": {
                  "name": "postgresqldatabase-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLDatabase",
                "metadata": {
                  "name": "postgresqldatabase-sample2"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLFirewallRule",
                "metadata": {
                  "name": "postgresqlfirewallrule-sample"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLFirewallRule",
                "metadata": {
                  "name": "postgresqlfirewallrule-sample2"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample2",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-replica3"
                },
                "spec": {
                  "createMode": "Replica",
                  "location": "eastus",
                  "replicaProperties": {
                    "sourceServerId": "/subscriptions/{subscription ID}/resourceGroups/resourcegroup-azure-operators/providers/Microsoft.DBforPostgreSQL/servers/postgresqlserver-sample"
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-sample2"
                },
                "spec": {
                  "location": "southcentralus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 2,
                    "family": "Gen5",
                    "name": "B_Gen5_2",
                    "size": "51200",
                    "tier": "Basic"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLUser",
                "metadata": {
                  "name": "psqluser-sample"
                },
                "spec": {
                  "dbName": "postgresqldatabase-sample",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "azure_pg_admin"
                  ],
                  "server": "postgresqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLVNetRule",
                "metadata": {
                  "name": "postgresqlvnetrule-sample1"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-azure-operators",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCache",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "rediscache-sample-1"
                },
                "spec": {
                  "location": "westus",
                  "properties": {
                    "enableNonSslPort": true,
                    "sku": {
                      "capacity": 1,
                      "family": "C",
                      "name": "Basic"
                    }
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCacheAction",
                "metadata": {
                  "name": "rediscacheaction-sample-1"
                },
                "spec": {
                  "actionName": "rollallkeys",
                  "cacheName": "rediscache-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCacheFirewallRule",
                "metadata": {
                  "name": "rediscachefirewallrule"
                },
                "spec": {
                  "properties": {
                    "endIP": "0.0.0.0",
                    "startIP": "0.0.0.0"
                  },
                  "redisCache": "rediscache-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ResourceGroup",
                "metadata": {
                  "name": "resourcegroup-azure-operators"
                },
                "spec": {
                  "location": "westus"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "StorageAccount",
                "metadata": {
                  "name": "storageaccountsample777"
                },
                "spec": {
                  "accessTier": "Hot",
                  "kind": "StorageV2",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "name": "Standard_RAGRS"
                  },
                  "supportsHttpsTrafficOnly": true
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "VirtualNetwork",
                "metadata": {
                  "name": "vnet-sample-01"
                },
                "spec": {
                  "addressSpace": "10.0.0.0/8",
                  "location": "SouthCentralUS",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnets": [
                    {
                      "subnetAddressPrefix": "10.1.0.0/16",
                      "subnetName": "test1"
                    },
                    {
                      "subnetAddressPrefix": "10.2.0.0/16",
                      "subnetName": "test2"
                    }
                  ]
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "VirtualNetwork",
                "metadata": {
                  "name": "vnet-sample-02"
                },
                "spec": {
                  "addressSpace": "10.0.0.0/8",
                  "location": "SouthCentralUS",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnets": [
                    {
                      "subnetAddressPrefix": "10.1.0.0/16",
                      "subnetName": "test1"
                    },
                    {
                      "subnetAddressPrefix": "10.2.0.0/16",
                      "subnetName": "test2"
                    }
                  ]
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "BlobContainer",
                "metadata": {
                  "name": "blobcontainer-sample"
                },
                "spec": {
                  "accessLevel": "Container",
                  "accountName": "storageaccountsample777",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "MySQLServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "mysqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "postgresqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "azuresqldatabase-sample"
                },
                "spec": {
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "sku": {
                    "name": "Basic",
                    "tier": "Basic"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlFailoverGroup",
                "metadata": {
                  "name": "azuresqlfailovergroup-sample"
                },
                "spec": {
                  "databaseList": [
                    "azuresqldatabase-sample",
                    "azuresqldatabase-sample2"
                  ],
                  "failoverGracePeriod": 30,
                  "failoverPolicy": "automatic",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "secondaryServer": "sqlserver-samplesec",
                  "secondaryServerResourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-samplepri"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlFirewallRule",
                "metadata": {
                  "name": "sqlf-allowazuresvcaccess"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample-777"
                },
                "spec": {
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Cloud Provider
          certified: "false"
          containerImage: mcr.microsoft.com/k8s/azureserviceoperator:1.0.59040
          createdAt: Mon 12 Jun 2023 04:24:02 UTC
          description: The Azure Service Operator helps you provision Azure resources
            and connect your applications to them from within Kubernetes.
          displayName: Azure Service Operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Azure/azure-service-operator
          support: https://github.com/Azure/azure-service-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Creates an API with the specified properties in the specified
              API Management service.
            displayName: APIMgmtAPI
            kind: APIMgmtAPI
            name: apimgmtapis.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an API Management instance into a specified Resource
              Group at the specified location. It also provides the option to link
              to an Application Insights instance for logging, and to place the API
              Management instance in a specified Virtual Network.
            displayName: ApimService
            kind: ApimService
            name: apimservices.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Application Insights instance into a specified
              Resource Group at the specified location.
            displayName: AppInsights
            kind: AppInsights
            name: appinsights.azure.microsoft.com
            version: v1alpha1
          - description: Creates an Api Key to access a given Application Insights
              instance.
            displayName: AppInsightsApiKey
            kind: AppInsightsApiKey
            name: appinsightsapikeys.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Load Balancer (LB) into a specified Resource
              Group at the specified location.
            displayName: AzureLoadBalancer
            kind: AzureLoadBalancer
            name: azureloadbalancers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Network Interface (NIC) into a specified
              Resource Group at the specified location. Users can specify underlying
              public IP address and virtual network configurations in their NIC setup.
            displayName: AzureNetworkInterface
            kind: AzureNetworkInterface
            name: azurenetworkinterfaces.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Public IP Address (PIP) into a specified
              Resource Group at the specified location. Users can specify IP allocation
              method, idle timeout, IP address version, and SKU.
            displayName: AzurePublicIPAddress
            kind: AzurePublicIPAddress
            name: azurepublicipaddresses.azure.microsoft.com
            version: v1alpha1
          - description: Allows you to roll the password for the specified SQL server.
            displayName: AzureSqlAction
            kind: AzureSqlAction
            name: azuresqlactions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an SQL database given the SQL server.
            displayName: AzureSqlDatabase
            kind: AzureSqlDatabase
            name: azuresqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an SQL database given the SQL server
            displayName: AzureSqlDatabase
            kind: AzureSqlDatabase
            name: azuresqldatabases.azure.microsoft.com
            version: v1beta1
          - description: Deploys a failover group on a specified Azure SQL server
              given the secondary server and the databases to failover.
            displayName: AzureSqlFailoverGroup
            kind: AzureSqlFailoverGroup
            name: azuresqlfailovergroups.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a failover group on a specified Azure SQL server
              given the secondary server and the databases to failover.
            displayName: AzureSqlFailoverGroup
            kind: AzureSqlFailoverGroup
            name: azuresqlfailovergroups.azure.microsoft.com
            version: v1beta1
          - description: Deploys a firewall rule to allow access to the SQL server
              from specific IPs.
            displayName: AzureSqlFirewallRule
            kind: AzureSqlFirewallRule
            name: azuresqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the SQL server
              from specific IPs.
            displayName: AzureSqlFirewallRule
            kind: AzureSqlFirewallRule
            name: azuresqlfirewallrules.azure.microsoft.com
            version: v1beta1
          - description: Binds a specified Managed Identity as a user on the SQL database
            displayName: AzureSQLManagedUser
            kind: AzureSQLManagedUser
            name: azuresqlmanagedusers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure SQL server given the location and Resource
              Group.
            displayName: AzureSqlServer
            kind: AzureSqlServer
            name: azuresqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure SQL server given the location and Resource
              Group.
            displayName: AzureSqlServer
            kind: AzureSqlServer
            name: azuresqlservers.azure.microsoft.com
            version: v1beta1
          - description: Creates an user on the specified Azure SQL database and stores
              the username/password as secrets.
            displayName: AzureSQLUser
            kind: AzureSQLUser
            name: azuresqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a VirtualNetwork rule to allow access to the SQL
              server from specific VNets.
            displayName: AzureSQLVNetRule
            kind: AzureSQLVNetRule
            name: azuresqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys Azure Virtual Machine Extensions to provide post-deployment
              configuration and automation on Azure VMs.
            displayName: AzureVirtualMachineExtension
            kind: AzureVirtualMachineExtension
            name: azurevirtualmachineextensions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Machine (VM) into a specified Resource
              Group at the specified location. Users can specify platform image, size,
              user name and public SSH key, etc. for the VM.
            displayName: AzureVirtualMachine
            kind: AzureVirtualMachine
            name: azurevirtualmachines.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Machine Scale Set (VMSS) into a
              specified Resource Group at the specified location. Users can specify
              platform image, size, user name and public SSH key, etc. for the VMSS.
            displayName: AzureVMScaleSet
            kind: AzureVMScaleSet
            name: azurevmscalesets.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a Blob Container into a specified Resource Group
              at the specified location. A Storage Account must first be created prior
              to creating a Blob Container.
            displayName: BlobContainer
            kind: BlobContainer
            name: blobcontainers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a Blob Container into a specified Resource Group
              at the specified location. A Storage Account must first be created prior
              to creating a Blob Container.
            displayName: BlobContainer
            kind: BlobContainer
            name: blobcontainers.azure.microsoft.com
            version: v1alpha2
          - description: Deploys a consumer group given the Event Hub, Event Hubs
              namespace and Resource Group.
            displayName: ConsumerGroup
            kind: ConsumerGroup
            name: consumergroups.azure.microsoft.com
            version: v1alpha1
          - description: Provision a CosmosDB instance given the Cosmos DB instance
              type, location, and Resource Group.
            displayName: CosmosDB
            kind: CosmosDB
            name: cosmosdbs.azure.microsoft.com
            version: v1alpha1
          - kind: CosmosDBSQLDatabase
            name: cosmosdbsqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Event Hubs namespace given the Resource Group
              and location. Also has the ability to configure SKU, properties, and
              network rules.
            displayName: EventhubNamespace
            kind: EventhubNamespace
            name: eventhubnamespaces.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Event Hubs instance given the Event Hubs namespace,
              Resource Group and Location.
            displayName: Eventhub
            kind: Eventhub
            name: eventhubs.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Key Vault key given the location and Resource
              Group.
            displayName: KeyVaultKey
            kind: KeyVaultKey
            name: keyvaultkeys.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Key Vault given the location and Resource
              Group.
            displayName: KeyVault
            kind: KeyVault
            name: keyvaults.azure.microsoft.com
            version: v1alpha1
          - description: Add an AAD user to an existing MySQL database.
            kind: MySQLAADUser
            name: mysqlaadusers.azure.microsoft.com
            version: v1alpha1
          - description: Add an AAD user to an existing MySQL database.
            kind: MySQLAADUser
            name: mysqlaadusers.azure.microsoft.com
            version: v1alpha2
          - description: Deploys a database under the given Azure Database for MySQL
              server
            displayName: MySQLDatabase
            kind: MySQLDatabase
            name: mysqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the Azure Database
              for MySQL server from the specified IP range.
            displayName: MySQLFirewallRule
            kind: MySQLFirewallRule
            name: mysqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Create an AAD-based admininistrative user in a MySQL server.
            kind: MySQLServerAdministrator
            name: mysqlserveradministrators.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for MySQL server given the Location,
              Resource Group and other properties. Also helps creating read replicas
              for MySQL server.
            displayName: MySQLServer
            kind: MySQLServer
            name: mysqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for MySQL server given the Location,
              Resource Group and other properties. Also helps creating read replicas
              for MySQL server.
            displayName: MySQLServer
            kind: MySQLServer
            name: mysqlservers.azure.microsoft.com
            version: v1alpha2
          - description: Add a new user to an existing MySQL database.
            displayName: MySQLUser
            kind: MySQLUser
            name: mysqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Add a new user to an existing MySQL database.
            kind: MySQLUser
            name: mysqlusers.azure.microsoft.com
            version: v1alpha2
          - description: Add virtual network rules to the MySQL server.
            displayName: MySQLVNetRule
            kind: MySQLVNetRule
            name: mysqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a database under the given Azure Database for PostgreSQL
              server.
            displayName: PostgreSQLDatabase
            kind: PostgreSQLDatabase
            name: postgresqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the Azure Database
              for PostgreSQL server from the specified IP range.
            displayName: PostgreSQLFirewallRule
            kind: PostgreSQLFirewallRule
            name: postgresqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for PostgreSQL server given the
              Location and Resource Group.
            displayName: PostgreSQLServer
            kind: PostgreSQLServer
            name: postgresqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for PostgreSQL server given the
              Location and Resource Group.
            displayName: PostgreSQLServer
            kind: PostgreSQLServer
            name: postgresqlservers.azure.microsoft.com
            version: v1alpha2
          - description: Create and Manage Users for Azure Database for PostgreSQL
              server.
            displayName: PostgreSQLUser
            kind: PostgreSQLUser
            name: postgresqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Create and manage VNet service endpoints and VNet rules in
              Azure Database for PostgreSQL.
            displayName: PostgreSQLVNetRule
            kind: PostgreSQLVNetRule
            name: postgresqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Allows you to regenerate keys and reboot the RedisCache cluster.
            displayName: RedisCacheAction
            kind: RedisCacheAction
            name: rediscacheactions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the RedisCache
              from the specified IP range.
            displayName: RedisCacheFirewallRule
            kind: RedisCacheFirewallRule
            name: rediscachefirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Cache for Redis into a specified Resource
              Group at the specified location.
            displayName: RedisCache
            kind: RedisCache
            name: rediscaches.azure.microsoft.com
            version: v1alpha1
          - description: Provisions a Resource Group given the location and a name.
            displayName: ResourceGroup
            kind: ResourceGroup
            name: resourcegroups.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Storage Account into a specified Resource
              Group at the specified location.
            displayName: StorageAccount
            kind: StorageAccount
            name: storageaccounts.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Network into a specified Resource
              Group at the specified location, in the address space provided. Users
              are also able to add subnets to their virtual network through the operator.
            displayName: VirtualNetwork
            kind: VirtualNetwork
            name: virtualnetworks.azure.microsoft.com
            version: v1alpha1
        description: |
          ## Overview

          The Azure Service Operator comprises of:

          - The Custom Resource Definitions (CRDs) for each of the Azure services a Kubernetes user can provision.
          - The Kubernetes controller that watches for requests to create Custom Resources for each of these CRDs and creates them.

          The project was built using [Kubebuilder](https://book.kubebuilder.io/).

          Curious to see how it all works? Check out our [control flow diagrams](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/controlflow.md).

          **Note**: The current version of Azure Service Operator only supports watching all namespaces in the cluster.

          ## Supported Azure Services

          - [Resource Group](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/resourcegroup/resourcegroup.md)
          - [Event Hubs](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/eventhub/eventhub.md)
          - [Azure SQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/azuresql/azuresql.md)
          - [Azure Database for PostgreSQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/postgresql/postgresql.md)
          - [Azure Database for MySQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/mysql/mysql.md)
          - [Azure Key Vault](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/keyvault/keyvault.md)
          - [Azure Cache for Redis](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/rediscache/rediscache.md)
          - [Storage Account](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/storage/storageaccount.md)
          - [Blob Storage](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/storage/blobcontainer.md)
          - [Virtual Network](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/virtualnetwork/virtualnetwork.md)
          - [Application Insights](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/appinsights/appinsights.md)
          - [API Management](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/apimgmt/apimgmt.md)
          - [Cosmos DB](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/cosmosdb/cosmosdb.md)
          - [Virtual Machine](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/virtualmachine/virtualmachine.md)
          - [Virtual Machine Scale Set](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/vmscaleset/vmscaleset.md)

          ## Instructions - Please read before installing

          Before you begin, verify you're running the Azure CLI version 2.0.53 or later. Run `az --version` to find the version. If you need to install or upgrade, see [Install Azure CLI](https://docs.microsoft.com/en-us/cli/azure/install-azure-cli).

          1. Create an Azure Service Principal. You'll need this to grant Azure Service Operator permissions to create resources in your subscription.

              First, set the following environment variables to your Azure Tenant ID and Subscription ID with your values:
              ```yaml
              AZURE_TENANT_ID=<your-tenant-id-goes-here>
              AZURE_SUBSCRIPTION_ID=<your-subscription-id-goes-here>
              ```

              You can find these values by using the [Azure CLI](https://docs.microsoft.com/en-us/cli/azure/?view=azure-cli-latest):
              ```sh
              az account show
              ```

          2. Next, we'll create a Service Principal with Contributor permissions for your subscription, so ASO can create resources in your subscription on your behalf. Note that the [ServicePrincipal](https://docs.microsoft.com/en-us/cli/azure/create-an-azure-service-principal-azure-cli) you pass to the command below needs to have access to create resources in your subscription. If you'd like to use Managed Identity for authorization instead, check out instructions [here](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/managedidentity.md).

              ```sh
              az ad sp create-for-rbac -n "azure-service-operator" --role contributor \
                  --scopes /subscriptions/$AZURE_SUBSCRIPTION_ID
              ```

              This should give you output like the following:
              ```sh
              "appId": "xxxxxxxxxx",
              "displayName": "azure-service-operator",
              "name": "http://azure-service-operator",
              "password": "xxxxxxxxxxx",
              "tenant": "xxxxxxxxxxxxx"
              ```

          3.  Once you have created a Service Principal, gather the following values:

              `AZURE_TENANT_ID` is the Tenant ID from Step 1.

              `AZURE_SUBSCRIPTION_ID` is the Subscription ID from Step 1.

              `AZURE_CLIENT_ID` is the appID from the Service Principal created in Step 2.

              `AZURE_CLIENT_SECRET` is the password from the service Principal we created in Step 2.

              `AZURE_CLOUD_ENV` is the Azure Environment you'd like to use, i.e. AzurePublicCloud, AzureUSGovernmentCloud, AzureChinaCloud, AzureGermanCloud.

          4.  Set those values in a `Secret` called `azureoperatorsettings` within the `operators` Namespace:
              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name: azureoperatorsettings
                namespace: operators
              stringData:
                AZURE_TENANT_ID: <your-tenant-id-goes-here>
                AZURE_SUBSCRIPTION_ID: <your-subscription-id-goes-here>
                AZURE_CLIENT_ID: <your-client-id>
                AZURE_CLIENT_SECRET: <your-client-secret>
                AZURE_CLOUD_ENV: <your-azure-cloud-environment>
              ```

          5. Now you can proceed to install the Azure Service Operator to the `operators` Namespace via the Install button on the top right of this page. After the operator is installed, you will then see the Azure Service Operator pod running in your cluster:

              ```console
              $ kubectl get pods -n operators
              NAME                                                READY   STATUS    RESTARTS   AGE
              azureoperator-controller-manager-7dd75bbd97-mk4s9   2/2     Running   0          35s
              ```

          ## About the project

          This project maintains [releases of the Azure Service Operator](https://github.com/Azure/azure-service-operator/releases).

          Please see the [FAQ](https://github.com/Azure/azure-service-operator/blob/main/docs/faq.md) for answers to commonly asked questions about the Azure Service Operator.

          Have more questions? Feel free to consult our documentation [here](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/contents.md).

          ## Contributing

          The [contribution guide](https://github.com/Azure/azure-service-operator/blob/main/CONTRIBUTING.md) covers everything you need to know about how you can contribute to Azure Service Operators. The [developer guide](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/contents.md#developing-azure-service-operator) will help you onboard as a developer.

          ## Support

          Azure Service Operator is an open source project that is **not** covered by the [Microsoft Azure support policy](https://support.microsoft.com/en-us/help/2941892/support-for-linux-and-open-source-technology-in-azure). Please search open issues [here](https://github.com/Azure/azure-service-operator/issues). If your issue isn't already represented, please [open a new one](https://github.com/Azure/azure-service-operator/issues/new/choose). The Azure Service Operator project maintainers will respond to the best of their abilities.

          ## Code of conduct

          This project has adopted the [Microsoft Open Source Code of Conduct](https://opensource.microsoft.com/codeofconduct/). For more information, see the [Code of Conduct FAQ](https://opensource.microsoft.com/codeofconduct/faq) or contact [opencode@microsoft.com](mailto:opencode@microsoft.com) with any additional questions or comments.
        displayName: Azure Service Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - azure, microsoft, cloud service broker
        links:
        - name: Azure Service Operator
          url: https://github.com/Azure/azure-service-operator
        maintainers:
        - email: christian.muirhead@microsoft.com
          name: Christian Muirhead
        maturity: stable
        provider:
          name: Microsoft
        relatedImages:
        - mcr.microsoft.com/k8s/azureserviceoperator:1.0.59040
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 1.0.59040
      entries:
      - name: azure-service-operator.v1.0.59040
        version: 1.0.59040
      name: stable
    defaultChannel: stable
    packageName: azure-service-operator
    provider:
      name: Microsoft
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: service-registry-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-registry-operator.v2.0.7-0.1662549961.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-kafkasql"
                },
                "spec": {
                  "configuration": {
                    "kafkasql": {
                      "bootstrapServers": "\u003cservice name\u003e.\u003cnamespace\u003e.svc:9092"
                    },
                    "persistence": "kafkasql"
                  }
                }
              },
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-sql"
                },
                "spec": {
                  "configuration": {
                    "persistence": "sql",
                    "sql": {
                      "dataSource": {
                        "password": "\u003cpassword\u003e",
                        "url": "jdbc:postgresql://\u003cservice name\u003e.\u003cnamespace\u003e.svc:5432/\u003cdatabase name\u003e",
                        "userName": "postgres"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/integration/service-registry-rhel8-operator@sha256:6dfd539951a739ec06fb6620b79c42c6159d4ee148c0d9d336547e786cd4780c
          createdAt: "2022-06-24"
          description: Deploy and manage Red Hat Integration service registry on OpenShift.
          olm.substitutesFor: service-registry-operator.v2.0.7-0.1661784420.p
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents a Red Hat Integration Service
              Registry instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Service Registry
          Service Registry stores and retrieves API designs and event schemas, and gives you control of their evolution.
          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf - Provides a REST API and web UI to manage the artifacts and artifact meta-data - Includes Serializers and Deserializers for Kafka client integration - Configurable rules to control schema validity and evolution (compatibility) - Storage options: Kafka, PostgreSQL - Compatible with Confluent and IBM APIs - Runs on a lightweight Quarkus platform - Includes Maven plugin to integrate with Maven based builds
          ## Service Registry Operator
          Provides a quick and easy way to deploy and manage service registry on OpenShift.
          **Features**
          - Supports installation and configuration of the Registry for all storage options - Easily perform a rolling upgrade of the Registry
          ## Prerequisites
          This operator does not deploy storage for the Registry. Therefore, some storage options require that the chosen persistence service is already set up.
          ## License
          Red Hat Service Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - Service Registry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - service
        - service-registry
        links:
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/integration/service-registry-rhel8-operator@sha256:6dfd539951a739ec06fb6620b79c42c6159d4ee148c0d9d336547e786cd4780c
        - registry.redhat.io/integration/service-registry-kafkasql-rhel8@sha256:fc97a2a08b2f77d620ea86ec77bb4f750bfc0109ede3ad1080c2cb694019f8b8
        - registry.redhat.io/integration/service-registry-sql-rhel8@sha256:6c97fa8ead1f6f0323811f8f35a80ae33f0e160f6c58f2bdb31713dbeaa3f014
        version: 2.0.7+0.1662549961.p
      entries:
      - name: service-registry-operator.v2.0.7-0.1662549961.p
        version: 2.0.7+0.1662549961.p
      - name: service-registry-operator.v2.0.7-0.1661784420.p
        version: 2.0.7+0.1661784420.p
      - name: service-registry-operator.v2.0.7
        version: 2.0.7
      - name: service-registry-operator.v2.0.6
        version: 2.0.6
      - name: service-registry-operator.v2.0.5-0.1654503709.p
        version: 2.0.5+0.1654503709.p
      - name: service-registry-operator.v2.0.5-0.1652887585.p
        version: 2.0.5+0.1652887585.p
      - name: service-registry-operator.v2.0.5-0.1646309885.p
        version: 2.0.5+0.1646309885.p
      - name: service-registry-operator.v2.0.5-0.1647938417.p
        version: 2.0.5+0.1647938417.p
      - name: service-registry-operator.v2.0.5-0.1648813064.p
        version: 2.0.5+0.1648813064.p
      - name: service-registry-operator.v2.0.5-0.1651047781.p
        version: 2.0.5+0.1651047781.p
      - name: service-registry-operator.v2.0.5-0.1652429755.p
        version: 2.0.5+0.1652429755.p
      - name: service-registry-operator.v2.0.5
        version: 2.0.5
      - name: service-registry-operator.v2.0.5-0.1655396591.p
        version: 2.0.5+0.1655396591.p
      - name: service-registry-operator.v2.0.4
        version: 2.0.4
      - name: service-registry-operator.v2.0.3
        version: 2.0.3
      - name: service-registry-operator.v2.0.2
        version: 2.0.2
      - name: service-registry-operator.v2.0.1
        version: 2.0.1
      - name: service-registry-operator.v2.0.0
        version: 2.0.0
      name: 2.0.x
    - currentCSV: service-registry-operator.v2.6.13-r2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-kafkasql"
                },
                "spec": {
                  "configuration": {
                    "kafkasql": {
                      "bootstrapServers": "\u003cservice name\u003e.\u003cnamespace\u003e.svc:9092"
                    },
                    "persistence": "kafkasql"
                  }
                }
              },
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-sql"
                },
                "spec": {
                  "configuration": {
                    "persistence": "sql",
                    "sql": {
                      "dataSource": {
                        "password": "\u003cpassword\u003e",
                        "url": "jdbc:postgresql://\u003cservice name\u003e.\u003cnamespace\u003e.svc:5432/\u003cdatabase name\u003e",
                        "userName": "postgres"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/apicurio/apicurio-registry-rhel8-operator@sha256:61f00395b51819312dda7abcdaa47190a11a83b1e2b523101be0ba776b1c5b7c
          createdAt: "2025-09-15"
          description: Deploy and manage Red Hat build of Apicurio Registry on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations",
            "Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents a Red Hat build of Apicurio Registry
              instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Red Hat build of Apicurio Registry

          Red Hat build of Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Provides a REST API and web UI to manage the artifacts and artifact meta-data
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL
          - Compatible with Confluent and IBM APIs
          - Runs on a lightweight Quarkus platform
          - Includes Maven plugin to integrate with Maven based builds

          ## Red Hat build of Apicurio Registry Operator

          Provides a quick and easy way to deploy and manage Red Hat build of Apicurio Registry on OpenShift.

          **Features**
          - Supports installation and configuration of the Registry for all storage options
          - Easily perform a rolling upgrade of Red Hat build of Apicurio Registry

          ## Prerequisites

          This operator does not deploy storage for Red Hat build of Apicurio Registry. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Red Hat build of Apicurio Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/main/LICENSE)
        displayName: Red Hat build of Apicurio Registry
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - service
        - service-registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/apicurio/apicurio-registry-rhel8-operator@sha256:61f00395b51819312dda7abcdaa47190a11a83b1e2b523101be0ba776b1c5b7c
        - registry.redhat.io/apicurio/apicurio-registry-kafkasql-rhel8@sha256:f2da7cd92e126187d91043e2ae40521f8574586f8b3975be27d8370fe31b6732
        - registry.redhat.io/apicurio/apicurio-registry-sql-rhel8@sha256:a55425a9af1a16959ab5e528a5740d47ea06f6935593cc925e7727cded4e7d13
        version: 2.6.13-r2
      entries:
      - name: service-registry-operator.v2.6.13-r2
        version: 2.6.13-r2
      - name: service-registry-operator.v2.6.13-r1
        version: 2.6.13-r1
      - name: service-registry-operator.v2.6.12-r1
        version: 2.6.12-r1
      - name: service-registry-operator.v2.6.11-r1
        version: 2.6.11-r1
      - name: service-registry-operator.v2.6.10
        version: 2.6.10
      - name: service-registry-operator.v2.6.9
        version: 2.6.9
      - name: service-registry-operator.v2.6.8-0.1740477005.p
        version: 2.6.8+0.1740477005.p
      - name: service-registry-operator.v2.6.8-0.1739457262.p
        version: 2.6.8+0.1739457262.p
      - name: service-registry-operator.v2.6.8
        version: 2.6.8
      - name: service-registry-operator.v2.6.6-0.1736942038.p
        version: 2.6.6+0.1736942038.p
      - name: service-registry-operator.v2.6.6
        version: 2.6.6
      - name: service-registry-operator.v2.6.5-0.1731946870.p
        version: 2.6.5+0.1731946870.p
      - name: service-registry-operator.v2.6.5
        version: 2.6.5
      - name: service-registry-operator.v2.6.5-0.1729511936.p
        version: 2.6.5+0.1729511936.p
      - name: service-registry-operator.v2.6.5-0.1730982111.p
        version: 2.6.5+0.1730982111.p
      - name: service-registry-operator.v2.6.3
        version: 2.6.3
      - name: service-registry-operator.v2.6.1-0.1724927626.p
        version: 2.6.1+0.1724927626.p
      - name: service-registry-operator.v2.6.1
        version: 2.6.1
      - name: service-registry-operator.v2.3.5
        version: 2.3.5
      - name: service-registry-operator.v2.3.4-0.1716987426.p
        version: 2.3.4+0.1716987426.p
      - name: service-registry-operator.v2.3.4
        version: 2.3.4
      - name: service-registry-operator.v2.3.3
        version: 2.3.3
      - name: service-registry-operator.v2.3.2
        version: 2.3.2
      - name: service-registry-operator.v2.3.1-0.1706181988.p
        version: 2.3.1+0.1706181988.p
      - name: service-registry-operator.v2.3.1
        version: 2.3.1
      - name: service-registry-operator.v2.3.0
        version: 2.3.0
      - name: service-registry-operator.v2.2.3
        version: 2.2.3
      - name: service-registry-operator.v2.2.3-0.1698369839.p
        version: 2.2.3+0.1698369839.p
      - name: service-registry-operator.v2.2.2
        version: 2.2.2
      - name: service-registry-operator.v2.2.2-0.1697116886.p
        version: 2.2.2+0.1697116886.p
      - name: service-registry-operator.v2.2.1
        version: 2.2.1
      - name: service-registry-operator.v2.2.0
        version: 2.2.0
      - name: service-registry-operator.v2.1.5
        version: 2.1.5
      - name: service-registry-operator.v2.1.4
        version: 2.1.4
      - name: service-registry-operator.v2.1.3
        version: 2.1.3
      - name: service-registry-operator.v2.1.2-0.1676389596.p
        version: 2.1.2+0.1676389596.p
      - name: service-registry-operator.v2.1.2
        version: 2.1.2
      - name: service-registry-operator.v2.1.1
        version: 2.1.1
      - name: service-registry-operator.v2.1.0
        version: 2.1.0
      - name: service-registry-operator.v2.0.7
        version: 2.0.7
      - name: service-registry-operator.v2.0.7-0.1662549961.p
        version: 2.0.7+0.1662549961.p
      - name: service-registry-operator.v2.0.7-0.1661784420.p
        version: 2.0.7+0.1661784420.p
      - name: service-registry-operator.v2.0.6
        version: 2.0.6
      - name: service-registry-operator.v2.0.5-0.1648813064.p
        version: 2.0.5+0.1648813064.p
      - name: service-registry-operator.v2.0.5-0.1646309885.p
        version: 2.0.5+0.1646309885.p
      - name: service-registry-operator.v2.0.5-0.1654503709.p
        version: 2.0.5+0.1654503709.p
      - name: service-registry-operator.v2.0.5-0.1652887585.p
        version: 2.0.5+0.1652887585.p
      - name: service-registry-operator.v2.0.5-0.1652429755.p
        version: 2.0.5+0.1652429755.p
      - name: service-registry-operator.v2.0.5-0.1651047781.p
        version: 2.0.5+0.1651047781.p
      - name: service-registry-operator.v2.0.5-0.1655396591.p
        version: 2.0.5+0.1655396591.p
      - name: service-registry-operator.v2.0.5-0.1647938417.p
        version: 2.0.5+0.1647938417.p
      - name: service-registry-operator.v2.0.5
        version: 2.0.5
      - name: service-registry-operator.v2.0.4
        version: 2.0.4
      - name: service-registry-operator.v2.0.3
        version: 2.0.3
      - name: service-registry-operator.v2.0.2
        version: 2.0.2
      - name: service-registry-operator.v2.0.1
        version: 2.0.1
      - name: service-registry-operator.v2.0.0
        version: 2.0.0
      name: 2.x
    defaultChannel: 2.x
    packageName: service-registry-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-keyspaces-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-keyspaces-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keyspaces.services.k8s.aws/v1alpha1",
                "kind": "Keyspace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "keyspaces.services.k8s.aws/v1alpha1",
                "kind": "Table",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/keyspaces-controller:1.1.1
          createdAt: "2025-09-20T01:23:26Z"
          description: AWS Keyspaces controller is a service controller for managing
            Keyspaces resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Keyspace represents the state of an AWS keyspaces Keyspace
              resource.
            displayName: Keyspace
            kind: Keyspace
            name: keyspaces.keyspaces.services.k8s.aws
            version: v1alpha1
          - description: Table represents the state of an AWS keyspaces Table resource.
            displayName: Table
            kind: Table
            name: tables.keyspaces.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Keyspaces resources in AWS from within your Kubernetes cluster.

          **About Amazon Keyspaces**

          Amazon Keyspaces (for Apache Cassandra) is a scalable, highly available, and managed Apache Cassandra–compatible database service. With Amazon Keyspaces, you don’t have to provision, patch, or manage servers, and you don’t have to install, maintain, or operate software.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Keyspaces
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keyspaces
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Keyspaces Developer Resources
          url: https://aws.amazon.com/keyspaces/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: keyspaces maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/keyspaces-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-keyspaces-controller.v1.1.1
        version: 1.1.1
      - name: ack-keyspaces-controller.v1.1.0
        version: 1.1.0
      - name: ack-keyspaces-controller.v1.0.14
        version: 1.0.14
      - name: ack-keyspaces-controller.v1.0.12
        version: 1.0.12
      - name: ack-keyspaces-controller.v1.0.11
        version: 1.0.11
      - name: ack-keyspaces-controller.v1.0.10
        version: 1.0.10
      - name: ack-keyspaces-controller.v1.0.9
        version: 1.0.9
      - name: ack-keyspaces-controller.v1.0.7
        version: 1.0.7
      - name: ack-keyspaces-controller.v1.0.6
        version: 1.0.6
      - name: ack-keyspaces-controller.v1.0.5
        version: 1.0.5
      - name: ack-keyspaces-controller.v1.0.3
        version: 1.0.3
      - name: ack-keyspaces-controller.v1.0.2
        version: 1.0.2
      - name: ack-keyspaces-controller.v1.0.1
        version: 1.0.1
      - name: ack-keyspaces-controller.v1.0.0
        version: 1.0.0
      - name: ack-keyspaces-controller.v0.0.16
        version: 0.0.16
      - name: ack-keyspaces-controller.v0.0.15
        version: 0.0.15
      - name: ack-keyspaces-controller.v0.0.14
        version: 0.0.14
      - name: ack-keyspaces-controller.v0.0.13
        version: 0.0.13
      - name: ack-keyspaces-controller.v0.0.12
        version: 0.0.12
      - name: ack-keyspaces-controller.v0.0.11
        version: 0.0.11
      - name: ack-keyspaces-controller.v0.0.10
        version: 0.0.10
      - name: ack-keyspaces-controller.v0.0.9
        version: 0.0.9
      - name: ack-keyspaces-controller.v0.0.8
        version: 0.0.8
      - name: ack-keyspaces-controller.v0.0.7
        version: 0.0.7
      - name: ack-keyspaces-controller.v0.0.6
        version: 0.0.6
      - name: ack-keyspaces-controller.v0.0.5
        version: 0.0.5
      - name: ack-keyspaces-controller.v0.0.4
        version: 0.0.4
      - name: ack-keyspaces-controller.v0.0.3
        version: 0.0.3
      - name: ack-keyspaces-controller.v0.0.2
        version: 0.0.2
      - name: ack-keyspaces-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-keyspaces-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/kuadrant-operator
    name: kuadrant-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kuadrant-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "authScheme": {
                    "identity": [
                      {
                        "apiKey": {
                          "allNamespaces": true,
                          "selector": {
                            "matchLabels": {
                              "app": "toystore"
                            }
                          }
                        },
                        "credentials": {
                          "in": "authorization_header",
                          "keySelector": "APIKEY"
                        },
                        "name": "friends"
                      }
                    ],
                    "response": [
                      {
                        "json": {
                          "properties": [
                            {
                              "name": "userID",
                              "valueFrom": {
                                "authJSON": "auth.identity.metadata.annotations.secret\\.kuadrant\\.io/user-id"
                              }
                            }
                          ]
                        },
                        "name": "rate-limit-apikey",
                        "wrapper": "envoyDynamicMetadata",
                        "wrapperKey": "ext_auth_data"
                      }
                    ]
                  },
                  "rules": null,
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "rateLimits": [
                    {
                      "configurations": [
                        {
                          "actions": [
                            {
                              "metadata": {
                                "default_value": "no-user",
                                "descriptor_key": "userID",
                                "metadata_key": {
                                  "key": "envoy.filters.http.ext_authz",
                                  "path": [
                                    {
                                      "segment": {
                                        "key": "ext_auth_data"
                                      }
                                    },
                                    {
                                      "segment": {
                                        "key": "userID"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          ]
                        }
                      ],
                      "limits": [
                        {
                          "conditions": [],
                          "maxValue": 2,
                          "seconds": 10,
                          "variables": [
                            "userID"
                          ]
                        }
                      ]
                    }
                  ],
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/kuadrant-operator:v0.3.1
          createdAt: "2023-06-01T12:24:17Z"
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Enable AuthN and AuthZ based access control on workloads
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1beta1
          - description: Kuadrant is the Schema for the kuadrants API
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: Enable access control on workloads based on HTTP rate limiting
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1beta1
        description: A Kubernetes Operator to manage the lifecycle of the Kuadrant
          system
        displayName: Kuadrant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: Kuadrant Operator
          url: https://github.com/Kuadrant/kuadrant-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - quay.io/kuadrant/kuadrant-operator:v0.3.1
        - oci://quay.io/kuadrant/wasm-shim:v0.2.0
        version: 0.3.1
      entries:
      - name: kuadrant-operator.v0.3.1
        version: 0.3.1
      - name: kuadrant-operator.v0.3.0
        version: 0.3.0
      - name: kuadrant-operator.v0.2.1
        version: 0.2.1
      - name: kuadrant-operator.v0.2.0
        version: 0.2.0
      name: alpha
    - currentCSV: kuadrant-operator.v0.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSPolicy",
                "metadata": {
                  "name": "dnspolicy-sample"
                },
                "spec": {
                  "healthCheck": {
                    "endpoint": "/",
                    "protocol": "HTTP"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "TLSPolicy",
                "metadata": {
                  "name": "tlspolicy-sample"
                },
                "spec": {
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "self-signed-ca"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "kuadrant.io/v1beta2",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "rules": {
                    "authentication": {
                      "apikey": {
                        "apiKey": {
                          "selector": {}
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta2",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "limits": {
                    "toys": {
                      "rates": [
                        {
                          "duration": 1,
                          "limit": 50,
                          "unit": "minute"
                        }
                      ]
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/kuadrant-operator:v0.11.0
          createdAt: "2024-10-02T20:30:45Z"
          description: '[DEPRECATED] Follow https://docs.kuadrant.io instead'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthPolicy enables authentication and authorization for service
              workloads in a Gateway API network
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1beta2
          - description: DNSPolicy configures how North-South based traffic should
              be balanced and reach the gateways
            displayName: DNSPolicy
            kind: DNSPolicy
            name: dnspolicies.kuadrant.io
            version: v1alpha1
          - description: Kuadrant configures installations of Kuadrant Service Protection
              components
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: RateLimitPolicy enables rate limiting for service workloads
              in a Gateway API network
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1beta2
          - description: TLSPolicy provides tls for gateway listeners by managing
              the lifecycle of tls certificates
            displayName: TLSPolicy
            kind: TLSPolicy
            name: tlspolicies.kuadrant.io
            version: v1alpha1
        description: |-
          ## Deprecated
          This Kuadrant community operator is deprecated, Follow https://docs.kuadrant.io instead.
        displayName: '[DEPRECATED] Kuadrant Operator'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: Kuadrant Operator
          url: https://github.com/Kuadrant/kuadrant-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - quay.io/kuadrant/kuadrant-operator:v0.11.0
        - oci://quay.io/kuadrant/wasm-shim:v0.6.0
        version: 0.11.1
      entries:
      - name: kuadrant-operator.v0.11.1
        version: 0.11.1
      - name: kuadrant-operator.v0.11.0
        version: 0.11.0
      - name: kuadrant-operator.v0.10.0
        version: 0.10.0
      - name: kuadrant-operator.v0.7.2
        version: 0.7.2
      - name: kuadrant-operator.v0.4.1
        version: 0.4.1
      - name: kuadrant-operator.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: kuadrant-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/kuadrant-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Nutanix
      provider-url: ""
    name: nutanixcsioperator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nutanixcsioperator.v2.6.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crd.nutanix.com/v1alpha1",
                "kind": "NutanixCsiStorage",
                "metadata": {
                  "name": "nutanixcsistorage"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:859b6d1b2b36a48391e347c2a7431cafa08decafb56f801e21e544c3460a4b4b
          createdAt: "2024-10-30T17:10:05Z"
          description: An Operator for deploying and managing the Nutanix CSI Driver
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2.6.11
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi","disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/nutanix/csi-operator
          support: Nutanix Support Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a new instance of Nutanix CSI Driver
            displayName: Nutanix CSI Storage
            kind: NutanixCsiStorage
            name: nutanixcsistorages.crd.nutanix.com
            version: v1alpha1
        description: "# Overview \n\nThe Nutanix CSI Operator for Kubernetes packages,
          deploys, manages, upgrades the Nutanix CSI Driver on Kubernetes and OpenShift
          for dynamic provisioning of persistent volumes on Nutanix Enterprise Cloud
          platform.\n\nThe Nutanix Container Storage Interface (CSI) Driver for Kubernetes
          leverages Nutanix Volumes and Nutanix Files to provide scalable and persistent
          storage for stateful applications.\n\nWith the Nutanix CSI Provider you
          can:\n\n*  **Provide persistent storage to your containers**\n    * Leverage
          PVC resources to consume dynamically Nutanix storage\n    * With Files storage
          classes, applications on multiple pods can access the same storage, and
          also have the benefit of multi-pod read and write access. \n\n## Warning\n\nIf
          you upgrade from a previous version take care to validate your existing
          StorageClass parameters against the corresponding [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v2_6:CSI-Volume-Driver-v2_6).\n\n##
          Supported Access mode\n\n| Storage Class mode | ReadWriteOnce | ReadOnlyMany
          | ReadWriteMany |\n|--------------------|---------------|--------------|---------------|\n|
          Nutanix Volumes    | yes           | yes          | no            |\n| Nutanix
          Files      | yes           | yes          | yes           |\n\n\n## Supported
          orchestration platforms\nThe following table details orchestration platforms
          suitable for deployment of the Nutanix CSI driver.\n\n| Orchestration platform
          | Version | Architecture |\n|------------------------|---------|--------------|\n|
          Red Hat OpenShift      | 4.12    | x86          |\n| Red Hat OpenShift      |
          4.13    | x86          |\n| Red Hat OpenShift      | 4.14    | x86          |\n|
          Red Hat OpenShift      | 4.15    | x86          |\n| Red Hat OpenShift      |
          4.16    | x86          |\n| Red Hat OpenShift      | 4.17    | x86          |\n\n##
          Configuring k8s secret and storage class\nIn order to use this driver, deploy
          your NutanixCsiStorage instance and create the relevant storage classes
          and secrets, by following the [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v2_6:CSI-Volume-Driver-v2_6)
          or the example in below section:\n\n**1. Create a NutanixCsiStorage resource
          to deploy your driver**\n\nYou can do it directly inside the Operator UI
          with the `Create instance` button or with the following ressource\n```yaml\napiVersion:
          crd.nutanix.com/v1alpha1\nkind: NutanixCsiStorage\nmetadata:\n  name: nutanixcsistorage\n
          \ namespace: openshift-cluster-csi-drivers\nspec: {}\n```\n\n**2. Create
          a Nutanix secret**\n```yaml\napiVersion: v1\nkind: Secret\nmetadata:\n  name:
          ntnx-secret\n  namespace: openshift-cluster-csi-drivers\nstringData:\n  #
          prism-ip:prism-port:username:password.\n  key: 1.2.3.4:9440:admin:password\n```\n\n**3.
          Create storage classes**\n\n- volume mode example\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n  name: nutanix-volume\nprovisioner: csi.nutanix.com\nparameters:\n
          \ csi.storage.k8s.io/provisioner-secret-name: ntnx-secret\n  csi.storage.k8s.io/provisioner-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/node-publish-secret-name:
          ntnx-secret\n  csi.storage.k8s.io/node-publish-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/controller-expand-secret-name: ntnx-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/fstype: ext4\n  storageContainer:
          default-container\n  storageType: NutanixVolumes\n  #description: \"description
          added to each storage object created by the driver\"\n  #isSegmentedIscsiNetwork:
          \"false\"\n  #whitelistIPMode: ENABLED\n  #chapAuth: ENABLED\n  #isLVMVolume:
          \"false\"\n  #numLVMDisks: 4\nallowVolumeExpansion: true\nreclaimPolicy:
          Delete\n```\n\n- dynamic files mode example\n\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n    name: nutanix-dynfiles\nprovisioner: csi.nutanix.com\nparameters:\n
          \ dynamicProv: ENABLED\n  nfsServerName: fs\n  csi.storage.k8s.io/provisioner-secret-name:
          ntnx-secret\n  csi.storage.k8s.io/provisioner-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/node-publish-secret-name: ntnx-secret\n  csi.storage.k8s.io/node-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-expand-secret-name:
          ntnx-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace: openshift-cluster-csi-drivers\n
          \ storageType: NutanixFiles\n  squashType: \"none\"\n  #description: \"description
          added to each storage object created by the driver\"\nallowVolumeExpansion:
          true\n```"
        displayName: Nutanix CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Nutanix
        - CSI
        - Volumes
        - Files
        - Storage
        links:
        - name: Documentation
          url: https://portal.nutanix.com/page/documents/list?type=software&filterKey=product&filterVal=CSI
        maintainers:
        - email: cloudnative@nutanix.com
          name: Nutanix Cloud Native
        maturity: beta
        provider:
          name: Nutanix
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:9af9bf28430b00a0cedeb2ec29acadce45e6afcecd8bdf31c793c624cfa75fa7
        - registry.k8s.io/sig-storage/livenessprobe@sha256:82adbebdf5d5a1f40f246aef8ddbee7f89dea190652aefe83336008e69f9a89f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:e998f22243869416f9860fc6a1fb07d4202eac8846defc1b85ebd015c1207605
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:76a47d100de1f2afa23803bec503d9440605db03c7179b231f963e5663a3d12a
        - registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:859b6d1b2b36a48391e347c2a7431cafa08decafb56f801e21e544c3460a4b4b
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:2cddcc716c1930775228d56b0d2d339358647629701047edfdad5fcdfaf4ebcb
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:04c55b93a032ad8075cae540d60fc740a72c31006590dbdb61ad4ab9813959d0
        - registry.connect.redhat.com/nutanix/ntnx-csi-os@sha256:92962320ad682a7113aeadc8f9de1c7849996e06819acf00925aa8aa3bf51068
        version: 2.6.11
      entries:
      - name: nutanixcsioperator.v2.6.11
        version: 2.6.11
      name: stable
    - currentCSV: nutanixcsioperator.v3.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crd.nutanix.com/v1alpha1",
                "kind": "NutanixCsiStorage",
                "metadata": {
                  "name": "nutanixcsistorage"
                },
                "spec": {
                  "ntnxInitConfigMap": {
                    "usePC": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:9cc0ffc6e9c0a82320e207dad6b6b3ddc30ab8f83fa9e099a34450d5537ca435
          createdAt: "2025-06-18T11:04:03Z"
          description: An Operator for deploying and managing the Nutanix CSI Driver
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.3.4
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi","disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/nutanix/csi-operator
          support: Nutanix Support Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a new instance of Nutanix CSI Driver
            displayName: Nutanix CSI Storage
            kind: NutanixCsiStorage
            name: nutanixcsistorages.crd.nutanix.com
            version: v1alpha1
        description: "# Important\nPlease read the following instructions carefully
          before upgrading from 2.6.x to 3.3.4, for more information please refer
          to [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)
          \n- Please do not upgrade to the CSI 3.x operator if:\n    * You are using
          LVM volumes.\n    \n- To upgrade from the CSI 2.6.x to CSI 3.3 (interacting
          with Prism Central) operator\n    * Create a Nutanix Prism Central secret
          as explained below.\n    * Delete the csidriver object from the cluster:\n\n
          \       ```\n        oc delete csidriver csi.nutanix.com\n        ```\n\n
          \   * In the installed operators, go to Nutanix CSI Operator and change
          the subscription channel from stable to stable-3.x. \n    If you have installed
          the operator with automatic update approval, the operator will be automatically
          upgraded to CSI 3.3, and then the nutanixcsistorage resource will be upgraded.\n
          \   An update plan will be generated for manual updates. Upon approval,
          the operator will be successfully upgraded.\n    \n- Direct upgrades from
          CSI 2.6.x to CSI 3.3 interacting with Prism Element are not supported. \n
          \   The only solution is to recreate the nutanixcsistorage instance by following
          the below procedure:\n    - In the installed operators, go to Nutanix CSI
          Operator and delete the nutanixcsistorage instance.\n    - Next change the
          subscription channel from stable to stable-3.x.\n    - Verify the following
          points:\n        - Ensure a Nutanix Prism Element secret is present in the
          namespace.\n        - Ensure that all the storage classes with provisioner:
          csi.nutanix.com have a controller publish secret as explained below.\n\n
          \           ```\n            csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n            csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n            ```\n\n            If this secret
          is not present in the storage class please delete and recreate the storage
          classes with the required secrets.\n    - Create a new instance of nutanixcsistorage
          from this operator by specifying `usePC: false` in YAML spec section.\n
          \   - Caution: Moving from CSI driver interacting with Prism Central to
          CSI driver interacting with Prism Element is not supported.\n    \n- Troubleshooting:\n
          \   \n    If the upgrade was unsuccessful and you want to revert to version
          CSI 2.6.x, please delete the csidriver object as explained above, uninstall
          the operator (no need to delete the nutanixcsistorage custom resource),
          and install version CSI 2.6.x from the stable channel.\n# Overview \nThe
          Nutanix CSI Operator for Kubernetes packages, deploys, manages, and upgrades
          the Nutanix CSI Driver on Kubernetes and OpenShift for dynamic provisioning
          of persistent volumes on the Nutanix Enterprise Cloud platform.\nThe Nutanix
          Container Storage Interface (CSI) Driver for Kubernetes leverages Nutanix
          Volumes and Nutanix Files to provide scalable and persistent storage for
          stateful applications.\nWith the Nutanix CSI Provider, you can:\n*  **Provide
          persistent storage to your containers**\n    * Leverage PVC resources to
          consume dynamically Nutanix storage\n    * With Files storage classes, applications
          on multiple pods can access the same storage and have the benefit of multi-pod
          read and write access. \n## Warning\nIf you upgrade from a previous version,
          make sure to validate your existing StorageClass parameters against the
          corresponding [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)\n##
          Supported Access mode\n| Storage Class mode | ReadWriteOnce | ReadOnlyMany
          | ReadWriteMany |\n|--------------------|---------------|--------------|---------------|\n|
          Nutanix Volumes    | yes           | yes          | no            |\n| Nutanix
          Files      | yes           | yes          | yes           |\n## Supported
          orchestration platforms\nThe following table details orchestration platforms
          suitable for deployment of the Nutanix CSI driver.\n| Orchestration platform
          | Version | Architecture |\n|------------------------|---------|--------------|\n|
          Red Hat OpenShift      | 4.12    | x86          |\n| Red Hat OpenShift      |
          4.13    | x86          |\n| Red Hat OpenShift      | 4.14    | x86          |\n|
          Red Hat OpenShift      | 4.15    | x86          |\n| Red Hat OpenShift      |
          4.16    | x86          |\n| Red Hat OpenShift      | 4.17    | x86          |\n|
          Red Hat OpenShift      | 4.18    | x86          |\n| Red Hat OpenShift      |
          4.19    | x86          |\n## Configuring k8s secret and storage class\nTo
          use this driver, create the relevant storage classes and secrets and deploy
          your NutanixCsiStorage instance by following the [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)
          or the example in below section:\n**1. Create a Nutanix PE secret**\n```yaml\napiVersion:
          v1\nkind: Secret\nmetadata:\n  name: ntnx-pe-secret\n  namespace: openshift-cluster-csi-drivers\nstringData:\n
          \ # prism-ip:prism-port:username:password.\n  key: 1.2.3.4:9440:admin:password\n
          \ files-key: \"fileserver01.sample.com:csi:password1\"\n```\n**2. Create
          a Nutanix PC secret**\n```yaml\napiVersion: v1\nkind: Secret\nmetadata:\n
          \ name: ntnx-pc-secret\n  namespace: openshift-cluster-csi-drivers\nstringData:\n
          \ # prism-ip:prism-port:username:password.\n  key: 1.2.3.4:9440:admin:password\n```\n**3.
          Create a NutanixCsiStorage resource to deploy your driver**\nYou can do
          it directly inside the Operator UI with the `Create instance` button or
          with the following resource\n```yaml\napiVersion: crd.nutanix.com/v1alpha1\nkind:
          NutanixCsiStorage\nmetadata:\n  name: nutanixcsistorage\n  namespace: openshift-cluster-csi-drivers\nspec:
          \n  ntnxInitConfigMap:\n    usePC : true (false for PE based installations.
          Default is true.)\n```\n**4. Create storage classes**\n- volume mode example
          for 3.x installation\n```yaml\nkind: StorageClass\napiVersion: storage.k8s.io/v1\nmetadata:\n
          \ name: nutanix-volume\nprovisioner: csi.nutanix.com\nparameters:\n  csi.storage.k8s.io/provisioner-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/node-publish-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-expand-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/fstype: ext4\n  storageContainer:
          default-container\n  storageType: NutanixVolumes\n  #description: \"description
          added to each storage object created by the driver\"\n  #hypervisorAttached:
          ENABLED\nallowVolumeExpansion: true\nreclaimPolicy: Delete\n```\n- volume
          mode example for 2.x installation\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n  name: nutanix-volume\nprovisioner: csi.nutanix.com\nparameters:\n
          \ csi.storage.k8s.io/provisioner-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/node-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/controller-expand-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/fstype: ext4\n   storageContainer:
          default-container\n   storageType: NutanixVolumes\n   #description: \"description
          added to each storage object created by the driver\"\n   #isSegmentedIscsiNetwork:
          \"false\"\n   #whitelistIPMode: ENABLED\n   #chapAuth: ENABLED\n   #isLVMVolume:
          \"false\"\n   #numLVMDisks: 4\n allowVolumeExpansion: true\n reclaimPolicy:
          Delete\n```\n- dynamic files mode example\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n    name: nutanix-dynfiles\nprovisioner: csi.nutanix.com\nparameters:\n
          \ dynamicProv: ENABLED\n  nfsServerName: fs\n  csi.storage.k8s.io/provisioner-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/node-publish-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-expand-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  storageType: NutanixFiles\n  squashType:
          \"none\"\n  #description: \"description added to each storage object created
          by the driver\"\nallowVolumeExpansion: true\n```"
        displayName: Nutanix CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Nutanix
        - CSI
        - Volumes
        - Files
        - Storage
        links:
        - name: Documentation
          url: https://portal.nutanix.com/page/documents/list?type=software&filterKey=product&filterVal=CSI
        maintainers:
        - email: cloudnative@nutanix.com
          name: Nutanix Cloud Native
        maturity: beta
        provider:
          name: Nutanix
        relatedImages:
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:9af9bf28430b00a0cedeb2ec29acadce45e6afcecd8bdf31c793c624cfa75fa7
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:04c55b93a032ad8075cae540d60fc740a72c31006590dbdb61ad4ab9813959d0
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:9e581022a4e063b9d8b4f9466a36eabe4d0ced3badc8e353f13b700074ce4553
        - registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:9cc0ffc6e9c0a82320e207dad6b6b3ddc30ab8f83fa9e099a34450d5537ca435
        - registry.connect.redhat.com/nutanix/ntnx-csi-os@sha256:f0e191e7bf98829dbf718978fc971444436a1ecb9957ee22a5d8f9ba01a931b7
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0f64602ea791246712b51df334bbd701a0f31df9950a4cb9c28c059f367baa9e
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:f1bd6ee18c4021c1c94f29edfab89b49b6a4d1b800936c19dbef2d75f8202f2d
        - registry.connect.redhat.com/nutanix/ntnx-csi-os@sha256:9f121929227acf5581a813dc2f952105f517647c11a39cd7ea0863e768082586
        - registry.k8s.io/sig-storage/livenessprobe@sha256:82adbebdf5d5a1f40f246aef8ddbee7f89dea190652aefe83336008e69f9a89f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:3c116f543f0590aeff3299c8bb0683f250817d11a77d9e9071b15a0bffdabcd9
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:10624570c0aceb03f55f1eb07147b0c537e4676869cca2e9bd4bab113f810ac4
        - registry.k8s.io/sig-storage/csi-attacher@sha256:d7325367ab72b2d469a5091d87b4fc01142d2d13d1a28b2defbbe3e6fdbc4611
        version: 3.3.4
      entries:
      - name: nutanixcsioperator.v3.3.4
        version: 3.3.4
      name: stable-3.x
    defaultChannel: stable-3.x
    packageName: nutanixcsioperator
    provider:
      name: Nutanix
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/confidential-containers
    name: trustee-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trustee-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          containerImage: registry.redhat.io/build-of-trustee/trustee-rhel9-operator@sha256:fe8e23de14ad088c79b1bd5e7aab08a1b88a86679effcbf174be81e761bd9d6d
          createdAt: "2024-06-07T10:06:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trustee-operator-system
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/trustee-operator
          support: Confidential Containers Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KbsConfig is the Schema for the kbsconfigs API
            displayName: Kbs Config
            kind: KbsConfig
            name: kbsconfigs.confidentialcontainers.org
            version: v1alpha1
        description: Remote attestation operator for confidential containers
        displayName: Red Hat build of Trustee
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trustee
        - Confidential Computing
        - Attestation
        links:
        - name: Red Hat build of Trustee
          url: https://github.com/openshift/trustee-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
          url: https://github.com/confidential-containers
        relatedImages:
        - registry.redhat.io/build-of-trustee/trustee-rhel9-operator@sha256:fe8e23de14ad088c79b1bd5e7aab08a1b88a86679effcbf174be81e761bd9d6d
        version: 0.4.2
      entries:
      - name: trustee-operator.v0.4.2
        version: 0.4.2
      - name: trustee-operator.v0.4.1
        version: 0.4.1
      - name: trustee-operator.v0.4.0
        version: 0.4.0
      - name: trustee-operator.v0.3.0
        version: 0.3.0
      - name: trustee-operator.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: trustee-operator
    provider:
      name: Red Hat
      url: https://github.com/confidential-containers
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: candidate
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: fast
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: stable
    defaultChannel: stable
    packageName: stackgres
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NooBaa
      provider-url: ""
    name: mcg-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mcg-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[{"kind":"NooBaa","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"noobaa","creationTimestamp":null},"spec":{"cleanupPolicy":{},"security":{"kms":{}},"loadBalancerSourceSubnets":{},"autoscaler":{}},"status":{"lastKeyRotateTime":null}},{"kind":"BackingStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"NamespaceStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"BucketClass","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{},"status":{}},{"kind":"NooBaaAccount","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"allow_bucket_creation":false},"status":{}}]'
          capabilities: Deep Insights
          categories: Storage,Big Data
          certified: "false"
          containerImage: registry.redhat.io/odf4/mcg-rhel9-operator@sha256:43ddea9d3c1cde5c8b38f74e8840a871dd0441ed9fd074528bacf78b439f024f
          createdAt: "2019-07-08T13:10:20.940Z"
          description: NooBaa is an object data service for hybrid and multi cloud
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operators.openshift.io/infrastructure-features: ֿ'["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/noobaa/noobaa-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A NooBaa system - Create this to start
            displayName: NooBaa
            kind: NooBaa
            name: noobaas.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos,
              PV's and more. Used in BucketClass to construct data placement policies.
            displayName: Backing Store
            kind: BackingStore
            name: backingstores.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos
              and more. Used in BucketClass to construct namespace policies.
            displayName: Namespace Store
            kind: NamespaceStore
            name: namespacestores.noobaa.io
            version: v1alpha1
          - description: Storage policy spec  tiering, mirroring, spreading, namespace
              policy. Combines BackingStores Or NamespaceStores. Referenced by ObjectBucketClaims.
            displayName: Bucket Class
            kind: BucketClass
            name: bucketclasses.noobaa.io
            version: v1alpha1
          - kind: NooBaaAccount
            name: noobaaaccounts.noobaa.io
            version: v1alpha1
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information.
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: "The noobaa operator creates and reconciles a NooBaa system in
          a Kubernetes/Openshift cluster.\n\nNooBaa provides an S3 object-store service
          abstraction and data placement policies to create hybrid and multi cloud
          data solutions.\n\nFor more information on using NooBaa refer to [Github](https://github.com/noobaa/noobaa-core)
          / [Website](https://www.noobaa.io) / [Articles](https://noobaa.desk.com).
          \n\n## How does it work?\n\n- The operator deploys the noobaa core pod and
          two services - Mgmt (UI/API) and S3 (object-store).\n- Both services require
          credentials which you will get from a secret that the operator creates -
          use describe noobaa to locate it.\n- The service addresses will also appear
          in the describe output - pick the one that is suitable for your client:\n
          \   - minikube - use the NodePort address.\n    - remote cluster - probably
          need one of the External addresses.\n    - connect an application on the
          same cluster - use Internal DNS (though any address should work)\n    \n-
          Feel free to email us or open github issues on any question.\n\n## Getting
          Started\n\n### Notes:\n- The following instructions are for **minikube**
          but it works on any Kubernetes/Openshift clusters.\n- This will setup noobaa
          in the **my-noobaa-operator** namespace.\n- You will need **jq**, **curl**,
          **kubectl** or **oc**, **aws-cli**.\n\n### 1. Install OLM (if you don't
          have it already):\n```\ncurl -sL https://github.com/operator-framework/operator-lifecycle-manager/releases/download/0.12.0/install.sh
          | bash -s 0.12.0\n```\n\n### 2. Install noobaa-operator:\n```\nkubectl create
          -f https://operatorhub.io/install/noobaa-operator.yaml\n```\nWait for it
          to be ready:\n```\nkubectl wait pod -n my-noobaa-operator -l noobaa-operator
          --for=condition=ready\n```\n\n### 3. Create noobaa system:\n```\ncurl -sL
          https://operatorhub.io/api/operator?packageName=noobaa-operator | \n    jq
          '.operator.customResourceDefinitions[0].yamlExample | .metadata.namespace=\"my-noobaa-operator\"'
          |\n    kubectl create -f -\n```\nWait for it to be ready:\n```\nkubectl
          wait pod -n my-noobaa-operator -l noobaa-core --for=condition=ready\nkubectl
          get noobaa -n my-noobaa-operator -w\n# NAME     PHASE   MGMT-ENDPOINTS                  S3-ENDPOINTS
          \                   IMAGE                    AGE\n# noobaa   **Ready**   [https://192.168.64.12:31121]
          \  [https://192.168.64.12:32557]   noobaa/noobaa-core:4.0   19m\n```\n\n###
          4. Get system information to your shell:\n```\nNOOBAA_SECRET=$(kubectl get
          noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.accounts.admin.secretRef.name'
          )\nNOOBAA_MGMT=$(kubectl get noobaa noobaa -n my-noobaa-operator -o json
          | jq -r '.status.services.serviceMgmt.nodePorts[0]' )\nNOOBAA_S3=$(kubectl
          get noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.services.serviceS3.nodePorts[0]'
          )\nNOOBAA_ACCESS_KEY=$(kubectl get secret $NOOBAA_SECRET -n my-noobaa-operator
          -o json | jq -r '.data.AWS_ACCESS_KEY_ID|@base64d')\nNOOBAA_SECRET_KEY=$(kubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq -r '.data.AWS_SECRET_ACCESS_KEY|@base64d')\n```\n\n###
          5. Connect to Mgmt UI:\n```\n# show email/password from the secret:\nkubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq '.data|map_values(@base64d)'\n\n#
          open mgmt UI login:\nopen $NOOBAA_MGMT\n```\n\n### 6. Connect to S3 with
          aws-cli:\n```\nalias s3='AWS_ACCESS_KEY_ID=$NOOBAA_ACCESS_KEY AWS_SECRET_ACCESS_KEY=$NOOBAA_SECRET_KEY
          aws --endpoint $NOOBAA_S3 --no-verify-ssl s3'\ns3 ls\ns3 sync /var/log/
          s3://first.bucket\ns3 ls s3://first.bucket\n```\n"
        displayName: NooBaa Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - noobaa
        - kubernetes
        - openshift
        - cloud
        - hybrid
        - multi
        - data
        - storage
        - s3
        - tiering
        - mirroring
        links:
        - name: Github
          url: https://github.com/noobaa/noobaa-core
        - name: Website
          url: https://www.noobaa.io
        - name: Articles
          url: https://noobaa.desk.com
        maintainers:
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: etamir@redhat.com
          name: Eran Tamir
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: NooBaa
        relatedImages:
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:f37e2c611f4cda32172d4e8938938e488076fd8a8f19c252c56a80c48044bc02
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhel8/postgresql-12@sha256:af1ab437b165fa4ee9dfd6ade637d7de5fa556d0b3851fb8edc78afe0e6fa95e
        - registry.redhat.io/odf4/mcg-rhel9-operator@sha256:43ddea9d3c1cde5c8b38f74e8840a871dd0441ed9fd074528bacf78b439f024f
        - registry.redhat.io/odf4/odf-cosi-sidecar-rhel9@sha256:bce36c7d33a85a11cd6b6b24f92887cffe6f38af70340e54d10e4f6f524aa782
        version: 4.15.20-rhodf
      entries:
      - name: mcg-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: mcg-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: mcg-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: mcg-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: mcg-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: mcg-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: mcg-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: mcg-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: mcg-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: mcg-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: mcg-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: mcg-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: mcg-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: mcg-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: mcg-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: mcg-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: mcg-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: mcg-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: mcg-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: mcg-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: mcg-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[{"kind":"NooBaa","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"noobaa","creationTimestamp":null},"spec":{"cleanupPolicy":{},"security":{"kms":{}},"loadBalancerSourceSubnets":{},"autoscaler":{}},"status":{"lastKeyRotateTime":null}},{"kind":"BackingStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"NamespaceStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"BucketClass","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{},"status":{}},{"kind":"NooBaaAccount","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"allow_bucket_creation":false},"status":{}}]'
          capabilities: Deep Insights
          categories: Storage,Big Data
          certified: "false"
          containerImage: registry.redhat.io/odf4/mcg-rhel9-operator@sha256:fdcdc8870b12d853973000877818463c3b4a5ed1d3316ff72a04a299577ed23d
          createdAt: "2019-07-08T13:10:20.940Z"
          description: NooBaa is an object data service for hybrid and multi cloud
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/infrastructure-features: ֿ'["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/noobaa/noobaa-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A NooBaa system - Create this to start
            displayName: NooBaa
            kind: NooBaa
            name: noobaas.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos,
              PV's and more. Used in BucketClass to construct data placement policies.
            displayName: Backing Store
            kind: BackingStore
            name: backingstores.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos
              and more. Used in BucketClass to construct namespace policies.
            displayName: Namespace Store
            kind: NamespaceStore
            name: namespacestores.noobaa.io
            version: v1alpha1
          - description: Storage policy spec  tiering, mirroring, spreading, namespace
              policy. Combines BackingStores Or NamespaceStores. Referenced by ObjectBucketClaims.
            displayName: Bucket Class
            kind: BucketClass
            name: bucketclasses.noobaa.io
            version: v1alpha1
          - kind: NooBaaAccount
            name: noobaaaccounts.noobaa.io
            version: v1alpha1
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information.
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: "The noobaa operator creates and reconciles a NooBaa system in
          a Kubernetes/Openshift cluster.\n\nNooBaa provides an S3 object-store service
          abstraction and data placement policies to create hybrid and multi cloud
          data solutions.\n\nFor more information on using NooBaa refer to [Github](https://github.com/noobaa/noobaa-core)
          / [Website](https://www.noobaa.io) / [Articles](https://noobaa.desk.com).
          \n\n## How does it work?\n\n- The operator deploys the noobaa core pod and
          two services - Mgmt (UI/API) and S3 (object-store).\n- Both services require
          credentials which you will get from a secret that the operator creates -
          use describe noobaa to locate it.\n- The service addresses will also appear
          in the describe output - pick the one that is suitable for your client:\n
          \   - minikube - use the NodePort address.\n    - remote cluster - probably
          need one of the External addresses.\n    - connect an application on the
          same cluster - use Internal DNS (though any address should work)\n    \n-
          Feel free to email us or open github issues on any question.\n\n## Getting
          Started\n\n### Notes:\n- The following instructions are for **minikube**
          but it works on any Kubernetes/Openshift clusters.\n- This will setup noobaa
          in the **my-noobaa-operator** namespace.\n- You will need **jq**, **curl**,
          **kubectl** or **oc**, **aws-cli**.\n\n### 1. Install OLM (if you don't
          have it already):\n```\ncurl -sL https://github.com/operator-framework/operator-lifecycle-manager/releases/download/0.12.0/install.sh
          | bash -s 0.12.0\n```\n\n### 2. Install noobaa-operator:\n```\nkubectl create
          -f https://operatorhub.io/install/noobaa-operator.yaml\n```\nWait for it
          to be ready:\n```\nkubectl wait pod -n my-noobaa-operator -l noobaa-operator
          --for=condition=ready\n```\n\n### 3. Create noobaa system:\n```\ncurl -sL
          https://operatorhub.io/api/operator?packageName=noobaa-operator | \n    jq
          '.operator.customResourceDefinitions[0].yamlExample | .metadata.namespace=\"my-noobaa-operator\"'
          |\n    kubectl create -f -\n```\nWait for it to be ready:\n```\nkubectl
          wait pod -n my-noobaa-operator -l noobaa-core --for=condition=ready\nkubectl
          get noobaa -n my-noobaa-operator -w\n# NAME     PHASE   MGMT-ENDPOINTS                  S3-ENDPOINTS
          \                   IMAGE                    AGE\n# noobaa   **Ready**   [https://192.168.64.12:31121]
          \  [https://192.168.64.12:32557]   noobaa/noobaa-core:4.0   19m\n```\n\n###
          4. Get system information to your shell:\n```\nNOOBAA_SECRET=$(kubectl get
          noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.accounts.admin.secretRef.name'
          )\nNOOBAA_MGMT=$(kubectl get noobaa noobaa -n my-noobaa-operator -o json
          | jq -r '.status.services.serviceMgmt.nodePorts[0]' )\nNOOBAA_S3=$(kubectl
          get noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.services.serviceS3.nodePorts[0]'
          )\nNOOBAA_ACCESS_KEY=$(kubectl get secret $NOOBAA_SECRET -n my-noobaa-operator
          -o json | jq -r '.data.AWS_ACCESS_KEY_ID|@base64d')\nNOOBAA_SECRET_KEY=$(kubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq -r '.data.AWS_SECRET_ACCESS_KEY|@base64d')\n```\n\n###
          5. Connect to Mgmt UI:\n```\n# show email/password from the secret:\nkubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq '.data|map_values(@base64d)'\n\n#
          open mgmt UI login:\nopen $NOOBAA_MGMT\n```\n\n### 6. Connect to S3 with
          aws-cli:\n```\nalias s3='AWS_ACCESS_KEY_ID=$NOOBAA_ACCESS_KEY AWS_SECRET_ACCESS_KEY=$NOOBAA_SECRET_KEY
          aws --endpoint $NOOBAA_S3 --no-verify-ssl s3'\ns3 ls\ns3 sync /var/log/
          s3://first.bucket\ns3 ls s3://first.bucket\n```\n"
        displayName: NooBaa Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - noobaa
        - kubernetes
        - openshift
        - cloud
        - hybrid
        - multi
        - data
        - storage
        - s3
        - tiering
        - mirroring
        links:
        - name: Github
          url: https://github.com/noobaa/noobaa-core
        - name: Website
          url: https://www.noobaa.io
        - name: Articles
          url: https://noobaa.desk.com
        maintainers:
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: etamir@redhat.com
          name: Eran Tamir
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: NooBaa
        relatedImages:
        - registry.redhat.io/odf4/mcg-rhel9-operator@sha256:fdcdc8870b12d853973000877818463c3b4a5ed1d3316ff72a04a299577ed23d
        - registry.redhat.io/odf4/odf-cosi-sidecar-rhel9@sha256:9fed589373d1586577d57f23b7a4f39ace5052fc55c99a03cb256da63a26c441
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:0084d9053ad2c74926ca02d2ede0461abd59227eee98d122c45e08a5ebaf629a
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhel8/postgresql-12@sha256:af1ab437b165fa4ee9dfd6ade637d7de5fa556d0b3851fb8edc78afe0e6fa95e
        version: 4.16.16-rhodf
      entries:
      - name: mcg-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: mcg-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: mcg-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: mcg-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: mcg-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: mcg-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: mcg-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: mcg-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: mcg-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: mcg-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: mcg-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: mcg-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: mcg-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: mcg-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: mcg-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: mcg-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: mcg-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: mcg-operator
    provider:
      name: NooBaa
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-eks-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-eks-controller.v1.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "FargateProfile",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Nodegroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/eks-controller:1.9.3
          createdAt: "2025-10-03T21:43:09Z"
          description: AWS EKS controller is a service controller for managing EKS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessEntry represents the state of an AWS eks AccessEntry
              resource.
            displayName: AccessEntry
            kind: AccessEntry
            name: accessentries.eks.services.k8s.aws
            version: v1alpha1
          - description: Addon represents the state of an AWS eks Addon resource.
            displayName: Addon
            kind: Addon
            name: addons.eks.services.k8s.aws
            version: v1alpha1
          - description: Cluster represents the state of an AWS eks Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.eks.services.k8s.aws
            version: v1alpha1
          - description: FargateProfile represents the state of an AWS eks FargateProfile
              resource.
            displayName: FargateProfile
            kind: FargateProfile
            name: fargateprofiles.eks.services.k8s.aws
            version: v1alpha1
          - description: IdentityProviderConfig represents the state of an AWS eks
              IdentityProviderConfig resource.
            displayName: IdentityProviderConfig
            kind: IdentityProviderConfig
            name: identityproviderconfigs.eks.services.k8s.aws
            version: v1alpha1
          - description: Nodegroup represents the state of an AWS eks Nodegroup resource.
            displayName: Nodegroup
            kind: Nodegroup
            name: nodegroups.eks.services.k8s.aws
            version: v1alpha1
          - description: PodIdentityAssociation represents the state of an AWS eks
              PodIdentityAssociation resource.
            displayName: PodIdentityAssociation
            kind: PodIdentityAssociation
            name: podidentityassociations.eks.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Kubernetes Service (EKS) resources in AWS from within your Kubernetes cluster.

          **About Amazon EKS**

          Amazon Elastic Kubernetes Service (Amazon EKS) is a managed service that you can use to run Kubernetes on AWS without needing to install, operate, and maintain your own Kubernetes control plane or nodes. Kubernetes is an open-source system for automating the deployment, scaling, and management of containerized applications.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EKS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eks
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EKS Developer Resources
          url: https://aws.amazon.com/eks/getting-started/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/eks-controller:1.9.3
        version: 1.9.3
      entries:
      - name: ack-eks-controller.v1.9.3
        version: 1.9.3
      - name: ack-eks-controller.v1.9.2
        version: 1.9.2
      - name: ack-eks-controller.v1.9.1
        version: 1.9.1
      - name: ack-eks-controller.v1.9.0
        version: 1.9.0
      - name: ack-eks-controller.v1.8.0
        version: 1.8.0
      - name: ack-eks-controller.v1.7.6
        version: 1.7.6
      - name: ack-eks-controller.v1.7.5
        version: 1.7.5
      - name: ack-eks-controller.v1.7.4
        version: 1.7.4
      - name: ack-eks-controller.v1.7.3
        version: 1.7.3
      - name: ack-eks-controller.v1.7.2
        version: 1.7.2
      - name: ack-eks-controller.v1.6.4
        version: 1.6.4
      - name: ack-eks-controller.v1.6.3
        version: 1.6.3
      - name: ack-eks-controller.v1.6.2
        version: 1.6.2
      - name: ack-eks-controller.v1.6.1
        version: 1.6.1
      - name: ack-eks-controller.v1.6.0
        version: 1.6.0
      - name: ack-eks-controller.v1.5.5
        version: 1.5.5
      - name: ack-eks-controller.v1.5.4
        version: 1.5.4
      - name: ack-eks-controller.v1.5.3
        version: 1.5.3
      - name: ack-eks-controller.v1.5.2
        version: 1.5.2
      - name: ack-eks-controller.v1.5.1
        version: 1.5.1
      - name: ack-eks-controller.v1.5.0
        version: 1.5.0
      - name: ack-eks-controller.v1.4.7
        version: 1.4.7
      - name: ack-eks-controller.v1.4.6
        version: 1.4.6
      - name: ack-eks-controller.v1.4.5
        version: 1.4.5
      - name: ack-eks-controller.v1.4.4
        version: 1.4.4
      - name: ack-eks-controller.v1.4.3
        version: 1.4.3
      - name: ack-eks-controller.v1.4.2
        version: 1.4.2
      - name: ack-eks-controller.v1.4.1
        version: 1.4.1
      - name: ack-eks-controller.v1.4.0
        version: 1.4.0
      - name: ack-eks-controller.v1.3.5
        version: 1.3.5
      - name: ack-eks-controller.v1.3.3
        version: 1.3.3
      - name: ack-eks-controller.v1.3.1
        version: 1.3.1
      - name: ack-eks-controller.v1.3.0
        version: 1.3.0
      - name: ack-eks-controller.v1.2.3
        version: 1.2.3
      - name: ack-eks-controller.v1.2.2
        version: 1.2.2
      - name: ack-eks-controller.v1.2.1
        version: 1.2.1
      - name: ack-eks-controller.v1.2.0
        version: 1.2.0
      - name: ack-eks-controller.v1.1.3
        version: 1.1.3
      - name: ack-eks-controller.v1.1.2
        version: 1.1.2
      - name: ack-eks-controller.v1.1.1
        version: 1.1.1
      - name: ack-eks-controller.v1.1.0
        version: 1.1.0
      - name: ack-eks-controller.v1.0.5
        version: 1.0.5
      - name: ack-eks-controller.v1.0.4
        version: 1.0.4
      - name: ack-eks-controller.v1.0.3
        version: 1.0.3
      - name: ack-eks-controller.v1.0.2
        version: 1.0.2
      - name: ack-eks-controller.v1.0.1
        version: 1.0.1
      - name: ack-eks-controller.v1.0.0
        version: 1.0.0
      - name: ack-eks-controller.v0.1.9
        version: 0.1.9
      - name: ack-eks-controller.v0.1.8
        version: 0.1.8
      - name: ack-eks-controller.v0.1.7
        version: 0.1.7
      - name: ack-eks-controller.v0.1.6
        version: 0.1.6
      - name: ack-eks-controller.v0.1.5
        version: 0.1.5
      - name: ack-eks-controller.v0.1.4
        version: 0.1.4
      - name: ack-eks-controller.v0.1.3
        version: 0.1.3
      - name: ack-eks-controller.v0.1.2
        version: 0.1.2
      - name: ack-eks-controller.v0.1.1
        version: 0.1.1
      - name: ack-eks-controller.v0.1.0
        version: 0.1.0
      - name: ack-eks-controller.v0.0.15
        version: 0.0.15
      - name: ack-eks-controller.v0.0.14
        version: 0.0.14
      - name: ack-eks-controller.v0.0.13
        version: 0.0.13
      - name: ack-eks-controller.v0.0.12
        version: 0.0.12
      - name: ack-eks-controller.v0.0.11
        version: 0.0.11
      - name: ack-eks-controller.v0.0.10
        version: 0.0.10
      - name: ack-eks-controller.v0.0.9
        version: 0.0.9
      - name: ack-eks-controller.v0.0.8
        version: 0.0.8
      - name: ack-eks-controller.v0.0.7
        version: 0.0.7
      - name: ack-eks-controller.v0.0.6
        version: 0.0.6
      - name: ack-eks-controller.v0.0.5
        version: 0.0.5
      - name: ack-eks-controller.v0.0.4
        version: 0.0.4
      - name: ack-eks-controller.v0.0.3
        version: 0.0.3
      - name: ack-eks-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-eks-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Carlos Eduardo de Paula
      provider-url: ""
    name: lbconfig-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lbconfig-operator.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-infra-sample",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "http",
                    "path": "/healthz",
                    "port": 1936
                  },
                  "ports": [
                    80,
                    443
                  ],
                  "provider": {
                    "creds": "f5-creds",
                    "host": "https://192.168.1.35",
                    "partition": "Common",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "F5_BigIP"
                  },
                  "type": "infra",
                  "vip": "192.168.1.45"
                }
              },
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-infra-sample-shard",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "http",
                    "path": "/healthz",
                    "port": 1936
                  },
                  "nodelabels": {
                    "kubernetes.io/region": "DC1",
                    "node-role.kubernetes.io/infra": ""
                  },
                  "ports": [
                    80,
                    443
                  ],
                  "provider": {
                    "creds": "f5-creds",
                    "host": "https://192.168.1.35",
                    "partition": "Common",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "F5_BigIP"
                  },
                  "type": "infra",
                  "vip": "192.168.1.46"
                }
              },
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-master-sample",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "https",
                    "path": "/healthz",
                    "port": 6443
                  },
                  "ports": [
                    6443
                  ],
                  "provider": {
                    "creds": "netscaler-creds",
                    "host": "https://192.168.1.36",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "Citrix_ADC"
                  },
                  "type": "master",
                  "vip": "192.168.1.40"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Networking
          certified: "false"
          containerImage: quay.io/carlosedp/lbconfig-operator:v0.5.1
          createdAt: "2025-01-09T15:45:28Z"
          description: Manage External Load Balancers allowing creation/update for
            VIPs and Servers dynamically via API.
          k8sMaxVersion: ""
          k8sMinVersion: ""
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/carlosedp/lbconfig-operator
          support: https://github.com/carlosedp/lbconfig-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalLoadBalancer is the Schema for the externalloadbalancers
              API
            displayName: ExternalLoadBalancer Instance
            kind: ExternalLoadBalancer
            name: externalloadbalancers.lb.lbconfig.carlosedp.com
            version: v1
          required:
          - description: ExternalLoadBalancer represents a configured instance of
              an external Load-Balancer for a specific group of nodes of the cluster.
              The Instance has a VIP and ports to be balanced to the cluster nodes
              based on a set of node labels.
            displayName: External Load-Balancer Configuration Instance
            kind: ExternalLoadBalancer
            name: externalloadbalancers.lb.lbconfig.carlosedp.com
            version: v1
        description: |
          ## About the Operator

          The LBConfig Operator, manages the configuration of External Load Balancer instances (on third-party equipment via it's API) and creates VIPs and IP Pools with Monitors for a set of OpenShift or Kubernetes nodes like Master-nodes (Control-Plane), Infra nodes (where the Routers or Ingress controllers are located) or based on it's roles and/or labels.

          The operator dynamically handles creating, updating or deleting the IPs of the pools in the Load Balancer based on the Node IPs for each role or label. On every change of the operator configuration (CRDs) or addition/change/removal or cluster Nodes, the operator updates the Load Balancer properly.

          ## Main Use Cases

          The main users for this operator is enterprise deployments or clusters composed of multiple nodes having an external load-balancer providing the balancing and high-availability to access the cluster in both API and Application levels.

          ### Create ExternalLoadBalancer Instances

          Create the instances for each Load Balancer instance you need (for example one for Master Nodes and another for the Infra Nodes). **If installing on OpenShift or Kubernetes with OLM (or in a different namespace), adjust the sample YAMLs to match the created namespace**.

          **The provider `vendor` field can be (case-insensitive):**

          * **`F5_BigIP`** - Tested on F5 BigIP version 15
          * **`Citrix_ADC`** - Tested on Citrix ADC (Netscaler) version 13
          * **`HAProxy`** - HAProxy with Dataplane API. ([Docs](./docs/haproxy/))
          * **`Dummy`** - Dummy backend used for testing to only print log messages on operations

          Create the secret holding the Load Balancer API user and password:

          ```sh
          oc create secret generic f5-creds --from-literal=username=admin --from-literal=password=admin123 --namespace lbconfig-operator-system
          ```

          #### Sample CRDs and Available Fields

          Master Nodes using a Citrix ADC LB:

          ```yaml
          apiVersion: lb.lbconfig.carlosedp.com/v1
          kind: ExternalLoadBalancer
          metadata:
            name: externalloadbalancer-master-sample
            namespace: lbconfig-operator-system
          spec:
            vip: "192.168.1.40"
            type: "master"
            ports:
              - 6443
            monitor:
              path: "/healthz"
              port: 6443
              monitortype: "https"
            provider:
              vendor: Citrix_ADC
              host: "https://192.168.1.36"
              port: 443
              creds: netscaler-creds
              validatecerts: false
          ```

          Infra Nodes using a F5 BigIP LB:

          ```yaml
          apiVersion: lb.lbconfig.carlosedp.com/v1
          kind: ExternalLoadBalancer
          metadata:
            name: externalloadbalancer-infra-sample
            namespace: lbconfig-operator-system
          spec:
            vip: "192.168.1.45"
            type: "infra"
            ports:
              - 80
              - 443
            monitor:
              path: "/healthz"
              port: 1936
              monitortype: http
            provider:
              vendor: F5_BigIP
              host: "https://192.168.1.35"
              port: 443
              creds: f5-creds
              partition: "Common"
              validatecerts: false
          ```

          To choose the nodes which will be part of the server pool, you can set either `type` or `nodelabels` fields. The yaml field `type: "master"` or `type: "infra"` selects nodes with the role label `"node-role.kubernetes.io/master"` and `"node-role.kubernetes.io/infra"` respectively. If the field `nodelabels` array is used instead, the operator will use nodes which match all labels.

          Clusters with sharded routers or using arbitrary labels to determine where the Ingress Controllers run can be configured like:

          ```yaml
          spec:
            vip: "10.0.0.6"
            ports:
              - 80
            nodelabels:
              "node.kubernetes.io/ingress-controller": "production"
              "kubernetes.io/region": "DC1"
            ...
          ```

          ## Getting Help

          This project fully open-source and is hosted at [GitHub](https://github.com/carlosedp/lbconfig-operator). Always check the project page and issues for the latest updates and/or solutions to problems encountered. If a problem is detected, please fill an issue. Feature request and improvement issues are welcome.

          Some fields inside `providers` are optional and depend on the used backend. Check the [API docs](https://pkg.go.dev/github.com/carlosedp/lbconfig-operator/api/lb.lbconfig.carlosedp.com/v1?utm_source=gopls#Provider) which fields are backend-specific.

          ## Disclaimers

          * The operator does not check if the requested configuration (names, IPs) already exists and/or conflicts with existing configuration in the Load Balancer. The user is responsible for these checks before deployment;
          * I am not responsible if the operator changes/deletes existing configuration on the Load Balancer if existing names are already configured.
          * The operator creates the entries(Pools, VIPs, Monitors) in the provided Load Balancer with the `name` of the instance configured in the CustomResource prefixed with the type. Eg. For a CR with name `externalloadbalancer-master-sample`, the operator creates a server pool named `Pool-externalloadbalancer-master-sample-6443` (suffixed with the port), a monitor named `Monitor-externalloadbalancer-master-sample` and a VIP named `VIP-externalloadbalancer-master-sample-6443` (suffixed with the port).
        displayName: External Load-Balancer Configuration Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 'load-balance '
        - infrastructure
        links:
        - name: GitHub
          url: https://github.com/carlosedp/lbconfig-operator
        - name: Documentation
          url: https://github.com/carlosedp/lbconfig-operator/docs
        - name: Issues
          url: https://github.com/carlosedp/lbconfig-operator/issues
        maintainers:
        - email: carlosedp@gmail.com
          name: Carlos Eduardo de Paula
        maturity: beta
        minKubeVersion: 1.18.0
        provider:
          name: Carlos Eduardo de Paula
        relatedImages:
        - quay.io/carlosedp/kube-rbac-proxy:v0.15.0
        - quay.io/carlosedp/lbconfig-operator:v0.5.1
        version: 0.5.1
      entries:
      - name: lbconfig-operator.v0.5.1
        version: 0.5.1
      - name: lbconfig-operator.v0.5.0
        version: 0.5.0
      - name: lbconfig-operator.v0.4.1
        version: 0.4.1
      - name: lbconfig-operator.v0.4.0
        version: 0.4.0
      - name: lbconfig-operator.v0.2.0
        version: 0.2.0
      name: beta
    defaultChannel: beta
    packageName: lbconfig-operator
    provider:
      name: Carlos Eduardo de Paula
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: group-sync-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: group-sync-operator.v0.0.36
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "azure-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "azure": {
                        "credentialsSecret": {
                          "name": "azure-group-sync",
                          "namespace": "group-sync-operator"
                        }
                      },
                      "name": "azure"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "github-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "github": {
                        "credentialsSecret": {
                          "name": "github-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "organization": "redhat-cop"
                      },
                      "name": "github"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "gitlab-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "gitlab": {
                        "credentialsSecret": {
                          "name": "gitlab-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "url": "https://gitlab.apps.openshift.com"
                      },
                      "name": "gitlab"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "keycloak-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "keycloak": {
                        "credentialsSecret": {
                          "name": "keycloak-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "insecure": true,
                        "realm": "ocp",
                        "url": "https://keycloak.apps.openshift.com"
                      },
                      "name": "keycloak"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "ldap-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "ldap": {
                        "credentialsSecret": {
                          "name": "ldap-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "insecure": true,
                        "rfc2307": {
                          "groupMembershipAttributes": [
                            "member"
                          ],
                          "groupNameAttributes": [
                            "cn"
                          ],
                          "groupUIDAttribute": "dn",
                          "groupsQuery": {
                            "baseDN": "ou=Groups,dc=example,dc=com",
                            "derefAliases": "never",
                            "filter": "(objectClass=groupofnames)",
                            "scope": "sub"
                          },
                          "tolerateMemberNotFoundErrors": true,
                          "tolerateMemberOutOfScopeErrors": true,
                          "userNameAttributes": [
                            "cn"
                          ],
                          "userUIDAttribute": "dn",
                          "usersQuery": {
                            "baseDN": "ou=Users,dc=example,dc=com",
                            "derefAliases": "never",
                            "scope": "sub"
                          }
                        },
                        "url": "ldap://localhost:389"
                      },
                      "name": "ldap"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "okta-sync"
                },
                "spec": {
                  "providers": [
                    {
                      "name": "okta",
                      "okta": {
                        "appId": "okta-sync-app-id",
                        "credentialsSecret": {
                          "name": "okta-api-token",
                          "namespace": "group-sync-operator"
                        },
                        "extractLoginUsername": true,
                        "profileKey": "login",
                        "url": "https://example.okta.com/"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/group-sync-operator@sha256:43ef79efc2a77ee5edd4e3123f9dab156a30d3e00fd1fba9f89600b96f0f68d6
          createdAt: "2025-08-12T19:19:21Z"
          description: Synchronize groups and users from external providers
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: group-sync-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/group-sync-operator
          support: Red Hat Community of Practice
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GroupSync is the Schema for the groupsyncs API
            displayName: Group Sync
            kind: GroupSync
            name: groupsyncs.redhatcop.redhat.io
            version: v1alpha1
        description: |
          Synchronizes groups from external providers into OpenShift

          ## Overview

          The OpenShift Container Platform contains functionality to synchronize groups found in external identity providers into the platform. Currently, the functionality that is included in OpenShift to limited to synchronizing LDAP only. This operator is designed to integrate with external providers in order to provide new solutions.

          Group Synchronization is facilitated by creating a `GroupSync` resource. The following describes the high level schema for this resource:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: example-groupsync
          spec:
            providers:
              - <One or more providers to synchronize>
          ```

          ## Authentication

          In most cases, authentication details must be provided in order to communicate with providers. Authentication details are provider specific with regards to the required values. In supported providers, the secret can be referenced in the `credentialsSecret` by name and namespace where it has been created as shown below:

          ```
          credentialsSecret:
            name: <secret_name>
            namespace: <secret_namespace>
          ```

          ## Providers

          Integration with external systems is made possible through a set of pluggable external providers. The following providers are currently supported:

          * [Azure](https://azure.microsoft.com/)
          * [GitHub](https://github.com)
          * [GitLab](https://gitlab.com)
          * [LDAP](https://en.wikipedia.org/wiki/Lightweight_Directory_Access_Protocol)
          * [Keycloak](https://www.keycloak.org/)/[Red Hat Single Sign On](https://access.redhat.com/products/red-hat-single-sign-on)
          * [Okta](https://www.okta.com/)
          * [IBM Security Verify](https://docs.verify.ibm.com/verify)

          The following sections describe the configuration options available for each provider


          ### Azure

          Groups contained within Azure Active Directory can be synchronized into OpenShift. The following table describes the set of configuration options for the Azure provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `authorityHost` | Azure Active Directory Endpoint | `https://login.microsoftonline.com` | No |
          | `baseGroups` | List of groups to start searching from instead of listing all groups in the directory | | No |
          | `credentialsSecret` | Name of the secret containing authentication details (See below) | | Yes |
          | `filter` | Graph API filter | | No |
          | `groups` | List of groups to filter against | | No |
          | `userNameAttributes` | Fields on a user record to use as the User Name | `userPrincipalName` | No |
          | `prune` | Prune Whether to prune groups that are no longer in Azure | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Azure provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: azure-groupsync
          spec:
            providers:
            - name: azure
              azure:
                credentialsSecret:
                  name: azure-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to Azure

          Authentication to Azure can be performed using Application Registration with access to query group information in Azure Active Directory.

          The App Registration must be granted access to the following Microsoft Graph API's:

          * Group.Read.All
          * GroupMember.Read.All
          * User.Read.All

          A secret must be created in the same namespace that contains the `GroupSync` resource:

          The following keys must be defined in the secret

          * `AZURE_TENANT_ID` - Tenant ID
          * `AZURE_CLIENT_ID` - Client ID
          * `AZURE_CLIENT_SECRET` - Client Secret

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic azure-group-sync --from-literal=AZURE_SUBSCRIPTION_ID=<AZURE_SUBSCRIPTION_ID> --from-literal=AZURE_TENANT_ID=<AZURE_TENANT_ID> --from-literal=AZURE_CLIENT_ID=<AZURE_CLIENT_ID> --from-literal=AZURE_CLIENT_SECRET=<AZURE_CLIENT_SECRET>
          ```

          ### GitHub

          Teams stored within a GitHub organization can be synchronized into OpenShift. The following table describes the set of configuration options for the GitHub provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `caSecret` | Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `organization` | Organization to synchronize against | | Yes |
          | `teams` | List of teams to filter against | | No |
          | `url` | Base URL for the GitHub or GitHub Enterprise host (Must contain a trailing slash) | | No |
          | `prune` | Prune Whether to prune groups that are no longer in GitHub | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Github provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: github-groupsync
          spec:
            providers:
            - name: github
              github:
                organization: ocp
                credentialsSecret:
                  name: github-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to GitHub

          Authentication to GitLab can be performed using a [Token](https://docs.gitlab.com/ee/security/token_overview.html) or a Username and Password (Note: 2FA not supported). A secret must be created in the same namespace that contains the `GroupSync` resource:

          When using an OAuth token, the following token types are supported:

          * Personal Access Token
          * OAuth Token
          * Job Token

          the following key is required:

          * `token` - OAuth token

          Optionally, the `tokenType` key can be specified to indicate the type of token being provided from the following values:

          * OAuth - `oauth`
          * Personal Access Token - `personal`
          * Job Token - `job`

          If no `tokenType` is provided, `oauth` is used by default

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token>
          ```

          To specify a token type, such as a Personal Access Token, the following command can be executed:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token> --from-literal=tokenType=personal
          ```

          The following keys are required for username and password:

          * `username` - Username for authenticating with GitHub
          * `password` - Password for authenticating with GitHub

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic github-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### GitLab

          Groups stored within a GitLab can be synchronized into OpenShift. The following table describes the set of configuration options for the GitLab provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `groups` | List of groups to filter against | | No |
          | `prune` | Prune Whether to prune groups that are no longer in GitLab | `false` | No |
          | `scope` | Scope for group synchronization. Options are `one` for one level or `sub` to include subgroups | `sub` | No |
          | `url` | Base URL for the GitLab instance | `https://gitlab.com` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Github provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: gitlab-groupsync
          spec:
            providers:
            - name: gitlab
              gitlab:
                credentialsSecret:
                  name: gitlab-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to GitLab

          Authentication to GitLab can be performed using an OAuth Personal Access Token or a Username and Password (Note: 2FA not supported). A secret must be created in the same namespace that contains the `GroupSync` resource:

          When using an OAuth token, the following key is required:

          * `token` - OAuth token

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token>
          ```

          The following keys are required for username and password:

          * `username` - Username for authenticating with GitLab
          * `password` - Password for authenticating with GitLab

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### LDAP

          Groups stored within an [LDAP](https://en.wikipedia.org/wiki/Lightweight_Directory_Access_Protocol) server can be synchronized into OpenShift. The LDAP provider implements the included features of the [Syncing LDAP groups](https://docs.openshift.com/container-platform/latest/authentication/ldap-syncing.html) feature and makes use of the libraries from the [OpenShift Command Line](https://docs.openshift.com/container-platform/latest/cli_reference/openshift_cli/getting-started-cli.html) tool to streamline the migration to this operator based implementation.

          The configurations of the three primary schemas (`rfc2307`, `activeDirectory` and `augmentedActiveDirectory`) can be directly migrated as is without any modification.

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | No |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `groupUIDNameMapping` | User defined name mapping | | No |
          | `rfc2307` | Configuration using the [rfc2307](https://docs.openshift.com/container-platform/latest/authentication/ldap-syncing.html#ldap-syncing-rfc2307_ldap-syncing-groups) schema | | No |
          | `activeDirectory` | Configuration using the [activeDirectory](https://docs.openshift.com/container-platform/4.5/authentication/ldap-syncing.html#ldap-syncing-activedir_ldap-syncing-groups) schema | | No |
          | `augmentedActiveDirectory` | Configuration using the [activeDirectory](https://docs.openshift.com/container-platform/4.5/authentication/ldap-syncing.html#ldap-syncing-augmented-activedir_ldap-syncing-groups) schema | | No |
          | `url` | Connection URL for the LDAP server | `ldap://ldapserver:389` | No |
          | `whitelist` | Explicit list of groups to synchronize |  | No |
          | `blacklist` | Explicit list of groups to not synchronize |  | No |
          | `prune` | Prune Whether to prune groups that are no longer in LDAP | `false` | No |

          The following is an example using the `rfc2307` schema:

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
                credentialsSecret:
                  name: ldap-group-sync
                  namespace: group-sync-operator
                insecure: true
                rfc2307:
                  groupMembershipAttributes:
                  - member
                  groupNameAttributes:
                  - cn
                  groupUIDAttribute: dn
                  groupsQuery:
                    baseDN: ou=Groups,dc=example,dc=com
                    derefAliases: never
                    filter: (objectClass=groupofnames)
                    scope: sub
                  tolerateMemberNotFoundErrors: true
                  tolerateMemberOutOfScopeErrors: true
                  userNameAttributes:
                  - cn
                  userUIDAttribute: dn
                  usersQuery:
                    baseDN: ou=Users,dc=example,dc=com
                    derefAliases: never
                    scope: sub
                url: ldap://ldapserver:389
              name: ldap
          ```

          The examples provided in the OpenShift documented referenced previously can be used to construct the schemas for the other LDAP synchronization types.

          #### Authenticating to LDAP

          If authentication is required in order to communicate with the LDAP server, a secret should be created in the same namespace that contains the `GroupSync` resource. The following keys can be defined:

          * `username` - Username (Bind DN) for authenticating with the LDAP server
          * `password` - Password for authenticating with the LDAP server

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic ldap-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          #### Whitelists and Blacklists

          Groups can be explicitly whitelisted or blacklisted in order to control the groups that are eligible to be synchronized into OpenShift. When running LDAP group synchronization using the command line, this configuration is referenced via separate files, but these are instead specified in the `blacklist` and `whitelist` properties as shown below:

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
          ...
                whitelist:
                - cn=Online Corporate Banking,ou=Groups,dc=example,dc=com
          ...
              name: ldap
          ```

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
          ...
                blacklist:
                - cn=Finance,ou=Groups,dc=example,dc=com
          ...
              name: ldap
          ```

          ### Keycloak

          Groups stored within Keycloak can be synchronized into OpenShift. The following table describes the set of configuration options for the Keycloak provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `groups` | List of groups to filter against | | No |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `loginRealm` | Realm to authenticate against | `master` | No |
          | `realm` | Realm to synchronize | | Yes |
          | `scope` | Scope for group synchronization. Options are `one` for one level or `sub` to include subgroups | `sub` | No |
          | `url` | Base URL for the Keycloak server. Older versions (<17.0.0) including Red Hat SSO should include the context path `/auth` appended to the hostname  | | Yes |
          | `prune` | Prune Whether to prune groups that are no longer in Keycloak | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Keycloak provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                url: https://keycloak-keycloak-operator.apps.openshift.com/auth
          ```

          #### Authenticating to Keycloak

          A user with rights to query for Keycloak groups must be available. The following permissions must be associated to the user:

          * Password must be set (Temporary option unselected) on the _Credentials_ tab
          * On the _Role Mappings_ tab, select _master-realm_ or _realm-management_ next to the _Client Roles_ dropdown and then select **Query Groups** and **Query Users**.

          A secret must be created in the same namespace that contains the `GroupSync` resource. It must contain the following keys for the user previously created:

          * `username` - Username for authenticating with Keycloak
          * `password` - Password for authenticating with Keycloak

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic keycloak-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### Okta

          [Okta Groups](https://help.okta.com/en/prod/Content/Topics/users-groups-profiles/usgp-main.htm) assigned to [Okta Applications](https://help.okta.com/en/prod/Content/Topics/Apps/Apps_Apps.htm) can be synchronized into OpenShift. The developer docs for the Okta API that the Okta Syncer uses can be found [here](https://developer.okta.com/docs/reference/api/apps/#list-groups-assigned-to-application).
          The following table describes the set of configuration options for the Okta provider:

            | Name | Description | Defaults | Required |
            | ----- | ---------- | -------- | ----- |
            | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
            | `groups` | List of groups to filter against | | No |
            | `url` | Okta organization URL (Must contain a trailing slash) | | Yes |
            | `appId` | Application ID of App Groups are assigned to | | Yes |
            | `extractLoginUsername` | Bool to determine if you should extract username from okta login | | No |
            | `profileKey` | Attribute field on Okta User Profile you would like to use as identity | `login` | No |
            | `groupLimit` | Integer to set the maximum number of groups to retrieve from OKTA per request. | `1000` | No |
            | `prune` | Prune Whether to prune groups that are no longer in OKTA | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with an Okta provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: okta-sync
          spec:
            providers:
              - name: okta
                okta:
                  credentialsSecret:
                    name: okta-api-token
                    namespace: group-sync-operator
                  url: "https://example.okta.com/"
                  appId: okta-sync-app-id
          ```

          #### Authenticating to Okta

          A secret must be created in the same namespace as the group-sync-operator pod. It must contain the following key:

            * `okta-api-token` - Okta API Token for interacting with Okta

          The secret can be created by executing the following command:

            ```shell
            oc create secret generic okta-api-token --from-literal=okta-api-token=<OKTA_API_TOKEN> -n group-sync-operator
            ```

          ### IBM Security Verify

          Groups defined in [IBM Security Verify](https://help.okta.com/en/prod/Content/Topics/users-groups-profiles/usgp-main.htm) (ISV) can be synchronized into OpenShift. Currently only the `userName` field from ISV will be synchronized. The developer docs for the ISV API can be found [here](https://docs.verify.ibm.com/verify/page/api-documentation).
          The following table describes the set of configuration options for the provider:

            | Name | Description | Defaults | Required |
            | ----- | ---------- | -------- | ----- |
            | `credentialsSecret` | Reference to a secret containing authentication details (see below) | `''`  | Yes |
            | `groups` | List of groups to synchronize (see below) | `nil`  | Yes |
            | `tenantUrl` | The ISV tenant URL, for example `https://my-isv.verify.ibm.com`) | `''`  | Yes |

          The following is an example of a minimal configuration that can be applied to integrate with an Okta provider:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ibmsecurityverify-sync
          spec:
              providers:
              - name: ibmsecurityverify
                ibmsecurityverify:
                  credentialsSecret:
                    name: isv-group-sync
                    namespace: group-sync-operator
                  tenantUrl: https://my-isv.verify.ibm.com
                  groups:
                    - name: 'application owners'
                      id: 645001V3V9
                    - name: developer
                      id: 645001V3VA
          ```

          #### Group Objects
          Each group object in the `groups` array must contain an `id` field. The group ID can be retrieved by pulling the group information from the ISV API. Optionally, the object may also contain a `name` which corresponds to the group's display name. When defined, the operator will confirm that the name defined in the YAML matches that received from the API when synchronization occurs; as the group IDs are not human-friendly, using the name can confirm the correct groups are configured. If the names do not match an error will be logged.

          #### Group Names
          The name of each groups created in OpenShift will match the group name in ISV. Any whitespace in the ISV group name will be replaced with a hyphen.

          #### Authenticating to IBM Security Verify

          A secret must be created in the same namespace as the group-sync-operator pod. It must contain the following keys:

          * `clientId` - The API client ID.
          * `clientSecret`- The API client secret.

          See the IBM Security Verify [API documentation](https://docs.verify.ibm.com/verify/docs/api-access) for setting up authentication.

          ### Support for Additional Metadata (Beta)

          Additional metadata based on Keycloak group are also added to the OpenShift groups as Annotations including:

          * Parent/child relationship between groups and their subgroups
          * Group attributes

          ## CA Certificates

          Several providers allow for certificates to be provided in either a _ConfigMap_ or _Secret_ to communicate securely to the target host through the use of a property called `ca`.

          The certificate can be added to a Secret called _keycloak-certs_ using the key `ca.crt` representing the certificate using the following command.

          ```
          oc create secret generic keycloak-certs --from-file=ca.crt=<file>
          ```

          An example of how the CA certificate can be added to the Keycloak provider is shown below:


          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                caSecret:
                  kind: Secret
                  name: keycloak-certs
                  namespace: group-sync-operator
                  key: ca.crt
                url: https://keycloak-keycloak-operator.apps.openshift.com
          ```

          Alteratively, a _ConfigMap_ can be used instead instead of a _Secret_. This is useful when using the [Certificate injection using Operators](https://docs.openshift.com/container-platform/latest/networking/configuring-a-custom-pki.html#certificate-injection-using-operators_configuring-a-custom-pki) feature.

          The following command can be used to create a _ConfigMap_ containing the certificate:

          ```
          oc create configmap keycloak-certs --from-file=ca.crt=<file>
          ````

          An example of how the CA certificate can be added to the Keycloak provider is shown below:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                caSecret:
                  kind: ConfigMap
                  name: keycloak-certs
                  namespace: group-sync-operator
                  key: ca.crt
                url: https://keycloak-keycloak-operator.apps.openshift.com
          ```

          ## Scheduled Execution

          A cron style expression can be specified for which a synchronization event will occur. The following specifies that a synchronization should occur nightly at 3AM

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            schedule: "0 3 * * *"
            providers:
            - ...
          ```

          If a schedule is not provided, synchronization will occur only when the object is reconciled by the platform.

          ## Metrics

          Prometheus compatible metrics are exposed by the Operator and can be integrated into OpenShift's default cluster monitoring. To enable OpenShift cluster monitoring, label the namespace the operator is deployed in with the label `openshift.io/cluster-monitoring="true"`.

          ```
          oc label namespace <namespace> openshift.io/cluster-monitoring="true"
          ```
        displayName: Group Sync Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - group
        - user
        - security
        - authorization
        links:
        - name: Group Sync Operator
          url: https://github.com/redhat-cop/group-sync-operator
        maintainers:
        - email: ablock@redhat.com
          name: Andrew Block
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:b8a174be91f3df75dff9656d4c72de28af51a731008e9e96db5eef3c7af58aac
        - quay.io/redhat-cop/group-sync-operator@sha256:43ef79efc2a77ee5edd4e3123f9dab156a30d3e00fd1fba9f89600b96f0f68d6
        version: 0.0.36
      entries:
      - name: group-sync-operator.v0.0.36
        version: 0.0.36
      - name: group-sync-operator.v0.0.35
        version: 0.0.35
      - name: group-sync-operator.v0.0.34
        version: 0.0.34
      - name: group-sync-operator.v0.0.33
        version: 0.0.33
      - name: group-sync-operator.v0.0.32
        version: 0.0.32
      - name: group-sync-operator.v0.0.31
        version: 0.0.31
      - name: group-sync-operator.v0.0.30
        version: 0.0.30
      - name: group-sync-operator.v0.0.29
        version: 0.0.29
      - name: group-sync-operator.v0.0.28
        version: 0.0.28
      - name: group-sync-operator.v0.0.27
        version: 0.0.27
      - name: group-sync-operator.v0.0.26
        version: 0.0.26
      - name: group-sync-operator.v0.0.25
        version: 0.0.25
      - name: group-sync-operator.v0.0.24
        version: 0.0.24
      - name: group-sync-operator.v0.0.23
        version: 0.0.23
      - name: group-sync-operator.v0.0.22
        version: 0.0.22
      - name: group-sync-operator.v0.0.21
        version: 0.0.21
      - name: group-sync-operator.v0.0.20
        version: 0.0.20
      - name: group-sync-operator.v0.0.19
        version: 0.0.19
      - name: group-sync-operator.v0.0.18
        version: 0.0.18
      - name: group-sync-operator.v0.0.17
        version: 0.0.17
      - name: group-sync-operator.v0.0.16
        version: 0.0.16
      - name: group-sync-operator.v0.0.15
        version: 0.0.15
      - name: group-sync-operator.v0.0.14
        version: 0.0.14
      - name: group-sync-operator.v0.0.13
        version: 0.0.13
      name: alpha
    defaultChannel: alpha
    packageName: group-sync-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubecost
      provider-url: https://www.kubecost.com
    name: kubecost-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubecost-operator.v2.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kubecost.com/v1alpha1",
                "kind": "CostAnalyzer",
                "metadata": {
                  "name": "costanalyzer-sample"
                },
                "spec": {
                  "clusterController": {
                    "image": {
                      "repository": "registry.connect.redhat.com/kubecost/kubecost-cluster-controller@sha256",
                      "tag": "1a6f11dcd7341ab3f24eba23ec06c7b75646a534a7ed960153601806425469d1"
                    }
                  },
                  "forecasting": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-modeling@sha256:8f59f8fd8696964c1b5322ca79fae78befe97323fa56c2b509bdcc67ee4c0edd"
                  },
                  "global": {
                    "platforms": {
                      "openshift": {
                        "createMonitoringClusterRoleBinding": true,
                        "createMonitoringResourceReaderRoleBinding": true,
                        "enabled": true,
                        "scc": {
                          "clusterController": true,
                          "networkCosts": true,
                          "nodeExporter": true
                        }
                      }
                    },
                    "prometheus": {
                      "enabled": false,
                      "fqdn": "https://prometheus-k8s.openshift-monitoring.svc.cluster.local:9091",
                      "kubeRBACProxy": true
                    }
                  },
                  "grafana": {
                    "image": {
                      "repository": "registry.redhat.io/rhel9/grafana@sha256",
                      "tag": "9f71de028c18034e07a8d15dac13ef68afb91cb899b9e8e7d39703547dd0eac1"
                    },
                    "sidecar": {
                      "image": {
                        "repository": "quay.io/kiwigrid/k8s-sidecar@sha256",
                        "tag": "d3749777d5bedf62549b452a7e2ccb205da53588dc7f04927efc75d0da5b6d23"
                      }
                    }
                  },
                  "kubecostAggregator": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef"
                  },
                  "kubecostFrontend": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-frontend@sha256:f9b2196dc6ac017b8074f8fe40f1a58e7a2630282c5b7a019ca9f3dbdb670dc8"
                  },
                  "kubecostModel": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef"
                  },
                  "networkCosts": {
                    "image": {
                      "repository": "registry.connect.redhat.com/kubecost/kubecost-network-costs@sha256",
                      "tag": "c444d475be4f1aa36680bd84358e1e07303dcfe5d13a95544ad9e59f80d4a5f0"
                    }
                  },
                  "prometheusRule": {
                    "enabled": true
                  },
                  "serviceMonitor": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          createdAt: "2025-08-08T19:34:34Z"
          description: Deploys Kubecost 2.8.2 via the Helm chart.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Contact Kubecost for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Kubecost.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The CostAnalyzer API describes a deployment of Kubecost and
              its dependencies.
            displayName: CostAnalyzer
            kind: CostAnalyzer
            name: costanalyzers.charts.kubecost.com
            version: v1alpha1
        description: |2-

          IMPORTANT - Before installation, you must run the following command to allow monitoring of the Kubecost namespace by the local OpenShift Promtheus. If you install in a namespace other than "Kubecost", adjust accordingly.

          `oc label namespace kubecost openshift.io/cluster-monitoring=true`

          Kubecost provides real-time cost visibility and insights for teams using Kubernetes, helping you continuously reduce your cloud costs.

          This OpenShift Operator deploys Kubecost from the official Helm chart.

          Kubecost allows you to:

          * See allocated spend across all native Kubernetes concepts and provide your teams with transparent, accurate cost data reconciled with your actual cloud bill.
          * Join in-cluster costs like CPU and memory with out-of-cluster spend from other cloud services.
          * Get complete cost visibility across OpenShift, On-Prem, AWS, GCP, Azure, Oracle, AliCloud + more.
          * Gain insights you can use to save 30-50%+ on your infrastructure spend, without exposing your private information — your data never gets shared externally, even with us.
          * Have real-time alerts and recurring reports, empowering teams to take control of their Kubernetes-enabled infrastructure and stay within budgeted limits.

          Visit the Kubecost website for more details including features, pricing, documentation, and more at www.kubecost.com.
        displayName: Kubecost
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kubernetes
        - cost
        - kubernetes costs
        - kubernetes optimization
        - k8s
        - cost monitoring
        - cloud costs
        - k8s cost
        links:
        - name: Kubecost
          url: https://www.kubecost.com
        maintainers:
        - email: support@kubecost.com
          name: Kubecost Support
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Kubecost
          url: https://www.kubecost.com
        relatedImages:
        - registry.connect.redhat.com/kubecost/kubecost-network-costs@sha256:c444d475be4f1aa36680bd84358e1e07303dcfe5d13a95544ad9e59f80d4a5f0
        - registry.redhat.io/rhel9/grafana@sha256:9f71de028c18034e07a8d15dac13ef68afb91cb899b9e8e7d39703547dd0eac1
        - quay.io/kiwigrid/k8s-sidecar@sha256:d3749777d5bedf62549b452a7e2ccb205da53588dc7f04927efc75d0da5b6d23
        - registry.connect.redhat.com/kubecost/kubecost-operator-manager@sha256:582223a66a77f35d1f7ca33460032cfca1581a8ff30f75ead289d8cd4cf62bb9
        - registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef
        - registry.connect.redhat.com/kubecost/kubecost-frontend@sha256:f9b2196dc6ac017b8074f8fe40f1a58e7a2630282c5b7a019ca9f3dbdb670dc8
        - registry.connect.redhat.com/kubecost/kubecost-modeling@sha256:8f59f8fd8696964c1b5322ca79fae78befe97323fa56c2b509bdcc67ee4c0edd
        - registry.connect.redhat.com/kubecost/kubecost-cluster-controller@sha256:1a6f11dcd7341ab3f24eba23ec06c7b75646a534a7ed960153601806425469d1
        version: 2.8.2
      entries:
      - name: kubecost-operator.v2.8.2
        version: 2.8.2
      - name: kubecost-operator.v2.8.1
        version: 2.8.1
      - name: kubecost-operator.v2.8.0
        version: 2.8.0
      - name: kubecost-operator.v2.7.2
        version: 2.7.2
      - name: kubecost-operator.v2.7.1
        version: 2.7.1
      - name: kubecost-operator.v2.5.4
        version: 2.5.4
      name: alpha
    defaultChannel: alpha
    packageName: kubecost-operator
    provider:
      name: Kubecost
      url: https://www.kubecost.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-access-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-access-operator.v25.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyAccess",
                "metadata": {
                  "name": "ivia-sample"
                },
                "spec": {
                  "image": "icr.io/ivia/ivia-wrp@sha256:c2dd5ec2b362136896e47ccbd0f3ee033af945fc2794ddaf075da603e4d4d7b2",
                  "instance": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isva/verify-access-operator@sha256:ca1d7be14b4dd313d4675f18541b2cc670ea0a0adf3ec87e347e5e20a86bc39c
          createdAt: "2025-05-01T01:27:17Z"
          description: The IBM Verify Identity Access Operator manages the lifecycle
            of IBM Verify Identity Access worker containers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM-Security/verify-access-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyAccess is the Schema for the ibmsecurityverifyaccesses
              API.
            displayName: IBMSecurity Verify Access
            kind: IBMSecurityVerifyAccess
            name: ibmsecurityverifyaccesses.ibm.com
            version: v1
        description: |+
          In a world of highly fragmented access management environments, [IBM Verify Identity Access](https://www.ibm.com/au-en/products/verify-access) helps you simplify your users' access while more securely adopting web, mobile and cloud technologies. This solution helps you strike a balance between usability and security through the use of risk-based access, single sign-on, integrated access management control, identity federation and its mobile multi-factor authentication capability, IBM Verify. Take back control of your access management with IBM Verify Identity Access.

          The IBM Verify Identity Access operator provides lifecycle management of the lightweight containers which are used to protect an environment, namely:
          * [Web Reverse Proxy](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-web-reverse-proxy)
          * [Runtime](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-runtime)
          * [Distributed Session Cache](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-distributed-session-cache)

          The operator will manage the deployment of these lightweight IBM Verify Identity Access worker containers, and also control the rolling restart of these containers when a configuration snapshot is updated.

          See the project [Readme](https://github.com/IBM-Security/verify-access-operator/blob/master/README.md) for further information and details.

        displayName: IBM Verify Identity Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Identity Access Product Information
          url: https://www.ibm.com/au-en/products/verify-access
        - name: Verify Identity Access Documentation
          url: https://www.ibm.com/docs/en/sva
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Identity Access Development Team
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - icr.io/isva/verify-access-operator@sha256:ca1d7be14b4dd313d4675f18541b2cc670ea0a0adf3ec87e347e5e20a86bc39c
        - icr.io/ivia/ivia-wrp@sha256:c2dd5ec2b362136896e47ccbd0f3ee033af945fc2794ddaf075da603e4d4d7b2
        version: 25.4.0
      entries:
      - name: ibm-security-verify-access-operator.v25.4.0
        version: 25.4.0
      - name: ibm-security-verify-access-operator.v24.12.0
        version: 24.12.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-access-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Redis
      provider-url: ""
    name: redis-enterprise-operator-cert
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-enterprise-operator.v6.2.10-45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
        - registry.connect.redhat.com/redislabs/services-manager@sha256:8b46f2aacfb3e1b243c0b64975accc55fd80b4de9c9ac4769c3c67a5676addc5
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:ffdf10a1b08115fd4de06f1cd751d36644a9bc9a8c11034cd432e9b66a5d2f73
        version: 6.2.10-45
      entries:
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: 6.2.10
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.22.0
    - currentCSV: redis-enterprise-operator.v7.22.0-11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:088e4b57c6e1ca1680b359078e5eb1bba67343cdbd9a539af26a0ea23a830593
        - registry.connect.redhat.com/redislabs/services-manager@sha256:3b96cdc65c3e42fd22eb165950be96b907725903ac74fe4f82cfa0073e28e95a
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:d56ba97e95e799210e76615095409143a5dddd5c5a1e1be1d551d7e3f9e4d3ef
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:0cfb07f774db131b5b74e5aa91291579c916c4d4cd1a693d7aaed554f0177a68
        version: 7.22.0-11.3
      entries:
      - name: redis-enterprise-operator.v7.22.0-11.3
        version: 7.22.0-11.3
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.22.0-11.3
    - currentCSV: redis-enterprise-operator.v7.4.6-7.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:473b5661049a6c8830ba113504975726154c7b172376ba3b63d891674d59c78f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5f6f32a63f4dd8be85b9d469f0fedd26d6aa76e32b7735bfa7ab7203cb272cdb
        - registry.connect.redhat.com/redislabs/services-manager@sha256:ba20ab644a473bc742af671171d5313537292d88942b16336d03a6386c061d8a
        version: 7.4.6-7.2
      entries:
      - name: redis-enterprise-operator.v7.4.6-7.2
        version: 7.4.6-7.2
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      name: 7.4.6
    - currentCSV: redis-enterprise-operator.v7.8.2-6.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:36c7a36dabb23a4c003e112085cda2873296bfb22e5d11b470d7c1aba55b7f0f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:c6e1f2109a0dac76b51c25c1cbf6030b4894b21b98147e2ad1db2dd8e7693576
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:acf817c90b5f03bc2796fbc305cf43ce5c0644dcbe1b685752090b3f600a0f7b
        version: 7.8.2-6.8
      entries:
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.2
    - currentCSV: redis-enterprise-operator.v7.8.4-9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:803be04534e59cdc7b8898b3bb336676aad1cb77cf266051e226be82679fcf6a
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:4ba0c414a74268bed9c6757e447b340db7c912afa864cfb096ddf88f11feeaf9
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:66ec124da16cc9d0abd66687e088be0b358e20cfd48f5774d34bf5874e6a4e65
        version: 7.8.4-9.0
      entries:
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.4
    - currentCSV: redis-enterprise-operator.v7.8.6-5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:1fff01daece34aeb7befcb0760d5112c8c7d261f2b86388ece416ac94076a573
        - registry.connect.redhat.com/redislabs/services-manager@sha256:88d483f5e37b5a901d1ae3d965304edc000c172d2e5e307297d5b7bcd77ae73b
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5125e83d3c93b224d1983c599994e43899a0ba0c6f2fc4eed31476f755b5fc0b
        version: 7.8.6-5.1
      entries:
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.6
    - currentCSV: redis-enterprise-operator.v6.2.10-41
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/outbound",
                      "versionTag": "6.2.10-41"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager",
                      "versionTag": "6.2.10-41"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise",
                    "versionTag": "6.2.14-9.rhel8-openshift"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/outbound@sha256:3c251754ae8650157465a56ef3a7b41d5a0ad4e9db09ef042e0014076f25e8a2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/outbound@sha256:3c251754ae8650157465a56ef3a7b41d5a0ad4e9db09ef042e0014076f25e8a2
        version: 6.2.10-41
      entries:
      - name: redis-enterprise-operator.v6.2.10-41
        version: 6.2.10-41
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: preview
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: production
    defaultChannel: production
    packageName: redis-enterprise-operator-cert
    provider:
      name: Redis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-iam-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-iam-controller.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Role",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Group",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "OpenIDConnectProvider",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Policy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/iam-controller:1.5.2
          createdAt: "2025-10-03T22:22:57Z"
          description: AWS IAM controller is a service controller for managing IAM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Group represents the state of an AWS iam Group resource.
            displayName: Group
            kind: Group
            name: groups.iam.services.k8s.aws
            version: v1alpha1
          - description: InstanceProfile represents the state of an AWS iam InstanceProfile
              resource.
            displayName: InstanceProfile
            kind: InstanceProfile
            name: instanceprofiles.iam.services.k8s.aws
            version: v1alpha1
          - description: OpenIDConnectProvider represents the state of an AWS iam
              OpenIDConnectProvider resource.
            displayName: OpenIDConnectProvider
            kind: OpenIDConnectProvider
            name: openidconnectproviders.iam.services.k8s.aws
            version: v1alpha1
          - description: Policy represents the state of an AWS iam Policy resource.
            displayName: Policy
            kind: Policy
            name: policies.iam.services.k8s.aws
            version: v1alpha1
          - description: Role represents the state of an AWS iam Role resource.
            displayName: Role
            kind: Role
            name: roles.iam.services.k8s.aws
            version: v1alpha1
          - description: ServiceLinkedRole represents the state of an AWS iam ServiceLinkedRole
              resource.
            displayName: ServiceLinkedRole
            kind: ServiceLinkedRole
            name: servicelinkedroles.iam.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS iam User resource.
            displayName: User
            kind: User
            name: users.iam.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Identity and Access Management (IAM) resources in AWS from within your Kubernetes cluster.

          **About Amazon IAM**

          AWS Identity and Access Management (IAM) is a web service that helps you securely control access to AWS resources. You use IAM to control who is authenticated (signed in) and authorized (has permissions) to use resources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon IAM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iam
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon IAM Developer Resources
          url: https://aws.amazon.com/iam/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: iam maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/iam-controller:1.5.2
        version: 1.5.2
      entries:
      - name: ack-iam-controller.v1.5.2
        version: 1.5.2
      - name: ack-iam-controller.v1.5.1
        version: 1.5.1
      - name: ack-iam-controller.v1.5.0
        version: 1.5.0
      - name: ack-iam-controller.v1.4.4
        version: 1.4.4
      - name: ack-iam-controller.v1.4.3
        version: 1.4.3
      - name: ack-iam-controller.v1.4.2
        version: 1.4.2
      - name: ack-iam-controller.v1.4.1
        version: 1.4.1
      - name: ack-iam-controller.v1.4.0
        version: 1.4.0
      - name: ack-iam-controller.v1.3.22
        version: 1.3.22
      - name: ack-iam-controller.v1.3.21
        version: 1.3.21
      - name: ack-iam-controller.v1.3.20
        version: 1.3.20
      - name: ack-iam-controller.v1.3.19
        version: 1.3.19
      - name: ack-iam-controller.v1.3.18
        version: 1.3.18
      - name: ack-iam-controller.v1.3.17
        version: 1.3.17
      - name: ack-iam-controller.v1.3.15
        version: 1.3.15
      - name: ack-iam-controller.v1.3.14
        version: 1.3.14
      - name: ack-iam-controller.v1.3.13
        version: 1.3.13
      - name: ack-iam-controller.v1.3.12
        version: 1.3.12
      - name: ack-iam-controller.v1.3.11
        version: 1.3.11
      - name: ack-iam-controller.v1.3.10
        version: 1.3.10
      - name: ack-iam-controller.v1.3.9
        version: 1.3.9
      - name: ack-iam-controller.v1.3.8
        version: 1.3.8
      - name: ack-iam-controller.v1.3.6
        version: 1.3.6
      - name: ack-iam-controller.v1.3.2
        version: 1.3.2
      - name: ack-iam-controller.v1.3.1
        version: 1.3.1
      - name: ack-iam-controller.v1.3.0
        version: 1.3.0
      - name: ack-iam-controller.v1.2.6
        version: 1.2.6
      - name: ack-iam-controller.v1.2.5
        version: 1.2.5
      - name: ack-iam-controller.v1.2.3
        version: 1.2.3
      - name: ack-iam-controller.v1.2.2
        version: 1.2.2
      - name: ack-iam-controller.v1.2.1
        version: 1.2.1
      - name: ack-iam-controller.v1.2.0
        version: 1.2.0
      - name: ack-iam-controller.v1.1.3
        version: 1.1.3
      - name: ack-iam-controller.v1.1.2
        version: 1.1.2
      - name: ack-iam-controller.v1.1.1
        version: 1.1.1
      - name: ack-iam-controller.v1.1.0
        version: 1.1.0
      - name: ack-iam-controller.v1.0.0
        version: 1.0.0
      - name: ack-iam-controller.v0.1.2
        version: 0.1.2
      - name: ack-iam-controller.v0.1.1
        version: 0.1.1
      - name: ack-iam-controller.v0.1.0
        version: 0.1.0
      - name: ack-iam-controller.v0.0.22
        version: 0.0.22
      - name: ack-iam-controller.v0.0.21
        version: 0.0.21
      - name: ack-iam-controller.v0.0.20
        version: 0.0.20
      - name: ack-iam-controller.v0.0.19
        version: 0.0.19
      - name: ack-iam-controller.v0.0.18
        version: 0.0.18
      - name: ack-iam-controller.v0.0.17
        version: 0.0.17
      - name: ack-iam-controller.v0.0.16
        version: 0.0.16
      - name: ack-iam-controller.v0.0.15
        version: 0.0.15
      - name: ack-iam-controller.v0.0.14
        version: 0.0.14
      - name: ack-iam-controller.v0.0.13
        version: 0.0.13
      - name: ack-iam-controller.v0.0.12
        version: 0.0.12
      - name: ack-iam-controller.v0.0.11
        version: 0.0.11
      - name: ack-iam-controller.v0.0.10
        version: 0.0.10
      - name: ack-iam-controller.v0.0.9
        version: 0.0.9
      - name: ack-iam-controller.v0.0.8
        version: 0.0.8
      - name: ack-iam-controller.v0.0.7
        version: 0.0.7
      - name: ack-iam-controller.v0.0.6
        version: 0.0.6
      - name: ack-iam-controller.v0.0.5
        version: 0.0.5
      - name: ack-iam-controller.v0.0.4
        version: 0.0.4
      - name: ack-iam-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-iam-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ODH
      provider-url: ""
    name: opendatahub-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opendatahub-operator.v2.35.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Managed",
                      "registriesNamespace": "odh-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "opendatahub",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "opendatahub"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v2.35.0
          createdAt: "2025-10-01T13:12:27Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",
            "workbenches.components.platform.opendatahub.io", "monitorings.services.platform.opendatahub.io",
            "servicemeshes.services.platform.opendatahub.io", "modelcontrollers.components.platform.opendatahub.io",
            "feastoperators.components.platform.opendatahub.io", "llamastackoperators.components.platform.opendatahub.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/opendatahub-io/opendatahub-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: ServiceMesh is the Schema for the servicemesh API
            displayName: Service Mesh
            kind: ServiceMesh
            name: servicemeshes.services.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack - Unified open-source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odh
        - notebooks
        - serving
        - training
        - pipelines
        - modelmesh
        - workbenches
        - dashboard
        - kserve
        - distributed workloads
        - trustyai
        - modelregistry
        - feast
        - featurestore
        - llamastack
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ODH
        relatedImages:
        - quay.io/opendatahub/opendatahub-operator:v2.35.0
        version: 2.35.0
      entries:
      - name: opendatahub-operator.v2.35.0
        version: 2.35.0
      - name: opendatahub-operator.v2.34.0
        version: 2.34.0
      - name: opendatahub-operator.v2.33.0
        version: 2.33.0
      - name: opendatahub-operator.v2.32.0
        version: 2.32.0
      - name: opendatahub-operator.v2.31.0
        version: 2.31.0
      - name: opendatahub-operator.v2.30.0
        version: 2.30.0
      - name: opendatahub-operator.v2.29.0
        version: 2.29.0
      - name: opendatahub-operator.v2.28.0
        version: 2.28.0
      - name: opendatahub-operator.v2.27.0
        version: 2.27.0
      - name: opendatahub-operator.v2.26.0
        version: 2.26.0
      - name: opendatahub-operator.v2.25.0
        version: 2.25.0
      - name: opendatahub-operator.v2.24.0
        version: 2.24.0
      - name: opendatahub-operator.v2.23.0
        version: 2.23.0
      - name: opendatahub-operator.v2.22.0
        version: 2.22.0
      - name: opendatahub-operator.v2.21.0
        version: 2.21.0
      - name: opendatahub-operator.v2.20.0
        version: 2.20.0
      - name: opendatahub-operator.v2.19.0
        version: 2.19.0
      - name: opendatahub-operator.v2.18.2
        version: 2.18.2
      - name: opendatahub-operator.v2.18.1
        version: 2.18.1
      - name: opendatahub-operator.v2.18.0
        version: 2.18.0
      - name: opendatahub-operator.v2.17.0
        version: 2.17.0
      - name: opendatahub-operator.v2.16.0
        version: 2.16.0
      - name: opendatahub-operator.v2.15.0
        version: 2.15.0
      - name: opendatahub-operator.v2.14.0
        version: 2.14.0
      - name: opendatahub-operator.v2.13.0
        version: 2.13.0
      - name: opendatahub-operator.v2.12.1
        version: 2.12.1
      - name: opendatahub-operator.v2.12.0
        version: 2.12.0
      - name: opendatahub-operator.v2.11.1
        version: 2.11.1
      - name: opendatahub-operator.v2.11.0
        version: 2.11.0
      - name: opendatahub-operator.v2.10.0
        version: 2.10.0
      name: fast
    - currentCSV: opendatahub-operator.v2.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "opendatahub",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "opendatahub"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v2.8.0
          createdAt: "2024-2-20T00:00:00Z"
          operatorframework.io/initialization-resource: |-
            {
            "apiVersion": "dscinitialization.opendatahub.io/v1",
            "kind": "DSCInitialization",
            "metadata": {
            "labels": {
             "app.kubernetes.io/created-by": "opendatahub-operator",
             "app.kubernetes.io/instance": "default",
             "app.kubernetes.io/managed-by": "kustomize",
             "app.kubernetes.io/name": "dscinitialization",
             "app.kubernetes.io/part-of": "opendatahub-operator"
            },
            "name": "default-dsci"
            },
            "spec": {
            "applicationsNamespace": "opendatahub",
            "monitoring": {
             "managementState": "Managed",
             "namespace": "opendatahub"
            },
            "serviceMesh": {
             "controlPlane": {
               "metricsCollection": "Istio",
               "name": "data-science-smcp",
               "namespace": "istio-system"
             },
             "managementState": "Managed"
            },
            "trustedCABundle": {
             "customCABundle": "",
             "managementState": "Managed"
            }
            }
            }
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: "The Open Data Hub is a machine-learning-as-a-service platform
          built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open
          Data Hub integrates multiple AI/ML open source components into one operator
          that can easily be downloaded and installed by OpenShift users.\n\nOpen
          Data Hub operator allows users to install and manage components of the Open
          Data Hub. Users can mix and match tools from each project to fulfill the
          needs of their use case. Each of the projects share some components, but
          can be mostly seen as an extension of each other to provide a complete solution
          for both novice and skilled enterprise users.\n    \n### Components\n\n*
          Open Data Hub Dashboard - A web dashboard that displays installed Open Data
          Hub components with easy access to component UIs and documentation\n* ODH
          Notebook Controller - Secure management of Jupyter Notebook in Kubernetes
          environments built on top of Kubeflow Notebook Controller with support for
          OAuth\n* Jupyter Notebooks - JupyterLab notebook that provide Python support
          for GPU workloads\n* Data Science Pipelines - Pipeline solution for end
          to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton\n*
          Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh,
          a general-purpose model serving management/routing layer\n* Distributed
          Workloads(Incubation) - Stack built to make managing distributed compute
          infrastructure in the cloud easy and intuitive for Data Scientists. This
          stack consists of three components \n                                      Codeflare
          , KubeRay and Kueue.\n* Kserve - Kserve is the Controller for for serving
          machine learning (ML) models on arbitrary frameworks"
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odh
        - notebooks
        - serving
        - training
        - pipelines
        - modelmesh
        - workbenches
        - dashboard
        - kserve
        - distributed-workloads
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: ODH
        relatedImages:
        - quay.io/opendatahub/opendatahub-operator:v2.8.0
        version: 2.8.1
      entries:
      - name: opendatahub-operator.v2.8.1
        version: 2.8.1
      name: odh-2.8.z
    - currentCSV: opendatahub-operator.v1.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kfdef.apps.kubeflow.org/v1",
                "kind": "KfDef",
                "metadata": {
                  "name": "opendatahub",
                  "namespace": "opendatahub"
                },
                "spec": {
                  "applications": [
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-common"
                        }
                      },
                      "name": "odh-common"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-dashboard"
                        }
                      },
                      "name": "odh-dashboard"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "data-science-pipelines-operator/"
                        }
                      },
                      "name": "data-science-pipelines-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-notebook-controller"
                        }
                      },
                      "name": "odh-notebook-controller"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "notebook-images"
                        }
                      },
                      "name": "notebook-images"
                    },
                    {
                      "kustomizeConfig": {
                        "parameters": [
                          {
                            "name": "monitoring-namespace",
                            "value": "opendatahub"
                          }
                        ],
                        "repoRef": {
                          "name": "manifests",
                          "path": "model-mesh"
                        }
                      },
                      "name": "model-mesh"
                    }
                  ],
                  "repos": [
                    {
                      "name": "manifests",
                      "uri": "file:///opt/manifests/odh-manifests.tar.gz"
                    }
                  ]
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhApplication",
                "metadata": {
                  "name": "example-odhapplication",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started,Model development,Model training,Notebook environments,Package management,Data Management"
                  }
                },
                "spec": {
                  "displayName": "ODHAPPLICATION DISPLAY NAME",
                  "provider": "PROVIDER_NAME",
                  "description": "Description of your ODHApplication",
                  "kfdefApplications": [
                    "<NAME OF THE ODH APPLICATION IN KFDEF>"
                  ],
                  "route": "<APPLICATION_OCP_ROUTE_OBJECT_NAME>",
                  "img": "<svg xmlns=\"http://www.w3.org/2000/svg\" role=\"img\" viewBox=\"-1.13 -1.13 58.50 48.94\"><title>Open Data Hub logo</title><g><g><g fill=\"#4d5258\" stroke-width=\".006\"><path fill=\"#bbb\" d=\"M112.30231 22.797083c-.2034 0-.35737.143517-.35737.337407 0 .192939.15397.339307.35737.339307.20339 0 .35736-.147318.35736-.339307 0-.19294-.15397-.337407-.35736-.337407zm.002.131161c.1112 0 .20149.08934.20149.206246 0 .116904-.0903.208146-.20149.208146-.1112 0-.2053-.09124-.2053-.208146 0-.116904.0931-.206246.2053-.206246zM113.04141 22.801836h-.28228v.665308h.15017v-.190088h.13211c.16918 0 .26612-.08934.26612-.244263 0-.147319-.0969-.230957-.26612-.230957zm-.008.349762h-.12451v-.225255h.12451c.0836 0 .13116.03707.13116.111202 0 .07603-.0475.114053-.13116.114053zM113.40328 22.801836v.665308h.51703v-.124508h-.36687v-.147318h.32125V23.07081h-.32125v-.144467h.35642v-.124507zM114.03814 22.801836v.665308h.14352v-.426748l.3222.426748h.13686v-.665308h-.14257v.427698L114.175 22.801836z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/><path d=\"M115.29685 22.801836h-.28228v.665308h.27658c.21004 0 .35546-.136863.35546-.332654 0-.196741-.14351-.332654-.34976-.332654zm.004.538899h-.13591v-.412491h.12926c.11405 0 .19959.08459.19959.207196 0 .121657-.0817.205295-.19294.205295zM116.21878 23.467144h.16062l-.28418-.665308h-.15492l-.28894.665308h.15493l.0523-.12926h.30794zm-.3108-.249966l.1055-.25852.10455.25852zM116.31525 22.801836v.127359h.20244v.537949h.15017v-.537949h.2034v-.127359zM117.37527 23.467144h.16063l-.28419-.665308h-.15492l-.28893.665308h.15492l.0523-.12926h.30795zm-.31079-.249966l.1055-.25852.10454.25852zM117.83433 22.801836v.665308h.15017v-.262322h.30129v.262322h.15017v-.665308h-.15017v.277528h-.30129v-.277528zM118.72025 23.188665v-.386829h-.15017v.386829c0 .175831.11785.285132.30224.285132.18343 0 .29844-.109301.29844-.285132v-.386829h-.15017v.386829c0 .09599-.0589.153971-.14827.153971-.0893 0-.15207-.05798-.15207-.153971zM119.61059 22.801836h-.30604v.665308h.31649c.15492 0 .25092-.07033.25092-.184385 0-.08744-.0532-.152071-.13876-.168228.0713-.01806.115-.07413.115-.146368 0-.102648-.0912-.166327-.23761-.166327zm-.008.268024h-.14826v-.145418h.14826c.0561 0 .0913.02566.0913.07128 0 .04657-.0352.07413-.0913.07413zm0 .274677h-.14826v-.158723h.14826c.0713 0 .11406.02946.11406.07794 0 .05132-.0428.08079-.11406.08079z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g><g fill=\"#58595b\" stroke-width=\".01\"><path d=\"M112.19192 24.15027h.0381l-.13717-.306759h-.0364l-.1376.306759h.0372l.0342-.07844h.16696zm-.18712-.110871l.0688-.157763.0697.157763zM112.27858 23.843511v.306759h.0351v-.306759zM112.64205 23.843511h-.11525v.306759h.0351v-.09729h.0802c.0758 0 .11964-.039.11964-.10649 0-.0653-.0438-.102983-.11964-.102983zm-.001.177044h-.0789V23.87594h.0789c.0552 0 .0872.02454.0872.07143 0 .0482-.032.07318-.0872.07318zM112.81775 23.825105v.325165h.0342v-.325165zM113.1137 24.15027l-.00044-.152503c-.00044-.04908-.032-.07932-.089-.07932-.0394 0-.0644.01139-.0942.03111l.0145.0241c.0241-.01709.0478-.02673.0732-.02673.0408 0 .0618.0206.0618.05434v.01446h-.0758c-.0543.000439-.0863.02586-.0863.06617 0 .03944.0307.06968.0784.06968.0368 0 .0662-.01227.0846-.03725v.03593zm-.11131-.02717c-.0324 0-.053-.01709-.053-.04251 0-.02585.0202-.03769.0587-.03769h.0719v.02805c-.009.03287-.0386.05215-.0776.05215zM113.29915 24.108201c-.0127.0074-.025.01139-.0368.01183-.0197 0-.0307-.01139-.0307-.03944v-.127086h.0732v-.02673h-.0732v-.06048h-.0337v.06048h-.0333v.02673h.0333v.132783c0 .04558.025.06573.0587.0653.0184 0 .0359-.0061.0521-.01709zM113.44228 23.854466c.0118 0 .0245.0044.0351.01183l.014-.02629c-.014-.01052-.032-.01622-.0495-.01622-.0386-.000438-.0741.02717-.0741.07713v.02586h-.0298v.02673h.0298v.196764h.0337v-.196764h.0662v-.02673h-.0662v-.03024c0-.02717.0184-.04251.0408-.04207zM113.59236 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM113.80323 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM114.22747 23.918448c-.0443.000438-.0767.01972-.0911.06004-.01-.03813-.0394-.06004-.0828-.06004-.0416.000438-.0727.01753-.0881.05346v-.05215h-.0337v.230508h.0337v-.113501c0-.04996.0285-.08458.0758-.08589.0403 0 .064.02454.064.06617v.133221h.0346v-.113501c0-.04996.0276-.08458.0754-.08589.0399 0 .064.02454.064.06617v.133221h.0342V24.00828c0-.05609-.032-.08984-.0859-.08984zM114.63474 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0487.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11088-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM114.89934 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM115.02844 23.919762l.0863.230508h.0355l.0675-.188876.0679.188876h.0355l.0855-.230508h-.0355l-.0675.195888-.0679-.195888h-.0359l-.067.195888-.0684-.195888zM115.53009 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM115.73075 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM115.95034 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11482.116569.0386 0 .071-.01402.0929-.03813l-.0188-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18843c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM116.29474 23.825105v.144615c-.0171-.03287-.0491-.05171-.0903-.05171-.0653 0-.11087.04777-.11087.116569 0 .0688.0456.117445.11131.117445.0412 0 .0727-.01884.0898-.05127v.04952h.0342v-.325165zm-.0833.297557c-.0491 0-.0837-.03637-.0837-.08721 0-.05083.0346-.08765.0837-.08765.0486 0 .0833.03681.0833.08765 0 .05083-.0346.08721-.0833.08721zM116.65984 23.918448c-.0403 0-.0719.01884-.089.05083v-.144177h-.0342v.325165h.0342v-.04952c.0171.03199.0486.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM116.89657 24.183138l.10605-.263376h-.0351l-.0732.191944-.0767-.191944h-.0359l.0955.228756-.01.02191c-.009.02191-.0241.03331-.0421.03331-.0114 0-.0219-.0035-.0324-.01095l-.0145.02717c.014.01139.0294.01665.0486.01665.0307 0 .0548-.01841.0692-.05346zM117.29947 23.842634c-.0898 0-.16083.06793-.16083.154257 0 .08633.071.154694.16083.154694.0894 0 .16039-.06836.16039-.154694 0-.08633-.071-.154257-.16039-.154257zm0 .03243c.0684 0 .12446.0539.12446.121828 0 .06792-.0561.122703-.12446.122703-.0684 0-.12533-.05478-.12533-.122703 0-.06793.057-.121828.12533-.121828zM117.64414 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0486.05083.0898.05083.0653 0 .10999-.04777.10999-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM117.9061 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM118.19616 23.918448c-.0421.000438-.0732.01709-.0894.05215v-.05084h-.0342v.230508h.0342v-.12358c.004-.0447.032-.07494.0776-.07537.0408 0 .0653.02454.0653.06573v.133221h.0342v-.141986c0-.05609-.0324-.08984-.0876-.08984zM118.65836 23.901795l.0149-.03111c-.0254-.01709-.0618-.02848-.0973-.02848-.0635 0-.10605.03068-.10605.07888 0 .109995.17266.06661.17222.148559 0 .0298-.0276.04777-.0719.04777-.0346 0-.0736-.01578-.1021-.04295l-.0158.03024c.0289.02761.0727.04689.11745.04689.0657 0 .11087-.03287.11087-.0837.00044-.11131-.17223-.06968-.17223-.149436 0-.02629.0259-.04207.0653-.04207.0245 0 .057.0075.0846.02542zM118.84168 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .11701-.04777.11701-.117007 0-.06836-.0487-.11613-.11701-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM119.18446 23.919762v.113063c0 .04996-.0281.08502-.0758.08589-.0399 0-.064-.02454-.064-.06573v-.133222h-.0342v.141986c0 .05565.0324.08984.0863.08984.0416-.000438.0723-.01753.0876-.05346v.05215h.0342v-.230508zM119.33586 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM119.62557 23.976294l.0188-.02366c-.021-.02147-.0517-.03418-.089-.03418-.0666 0-.11438.0482-.11438.116568 0 .0688.0473.116569.11438.116569.0399 0 .0719-.0149.0925-.03988l-.0197-.02016c-.0162.01928-.0412.0298-.0714.0298-.0478 0-.082-.03506-.082-.08633 0-.05083.0342-.08589.082-.08589.028 0 .0513.01008.0688.02717zM119.78997 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g></g><g><g><path fill=\"#a586f3\" d=\"m 7808.8125,2097.1875 c -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 v 329.875 c -0.8781,10.2195 7.1804,19 17.4375,19 10.2571,0 19.0227,-8.7805 18.1446,-19 v -77.7324 l 0.4179,-6.1426 c 0,-55.1597 44.2153,-99.5625 99.375,-99.5625 55.1596,0 99.5625,44.4028 99.5625,99.5625 0.01,0.1837 0.015,0.3674 0.027,0.5508 -0.026,0.4409 -0.035,0.8826 -0.027,1.3242 v 110 c 0,9.6822 7.849,17.5312 17.5312,17.5312 9.6823,0 17.5313,-7.849 17.5313,-17.5312 v -110 c 0,-0.3245 -0.012,-0.6488 -0.031,-0.9727 0.018,-0.3004 0.029,-0.6013 0.031,-0.9023 1e-4,-74.075 -60.55,-134.4374 -134.625,-134.4375 -39.9177,1e-4 -75.8465,17.5354 -100.5,45.293 V 2115 c 0.1749,-9.8356 -7.7881,-17.8833 -17.625,-17.8125 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><path fill=\"#5ed3ee\" d=\"m 7765.8027,2098.123 c -0.038,5e-4 -0.077,0 -0.1152,0 -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 V 2361 c 0,41.9333 -26.0162,79.2671 -65.4375,93.5625 -39.4213,14.2954 -83.5594,2.3741 -110.4375,-29.8125 -26.878,-32.1865 -30.6579,-77.5719 -9.5625,-113.8125 21.0955,-36.2406 62.4262,-55.2897 103.6875,-47.8125 9.5173,1.7397 18.6463,-4.5547 20.4043,-14.0686 1.758,-9.514 -4.5187,-18.6551 -14.0293,-20.4314 -55.6354,-10.082 -111.8054,15.8219 -140.25,64.6875 -28.4446,48.8656 -23.3039,110.5381 12.9375,153.9375 36.2415,43.3993 95.908,59.4005 149.0625,40.125 44.7667,-16.2337 77.0286,-54.658 86.0977,-100.1562 0,0 1.9812,-4.7686 2.5898,-9.1563 0.6086,-4.3877 0,-17.0625 0,-17.0625 v -245.0625 c 0.1769,-9.865 -7.8313,-17.9263 -17.6973,-17.8145 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><g fill=\"none\" stroke-width=\".365\" transform=\"translate(-449.265 -83.342) scale(5.91262)\"><circle cx=\"79.677\" cy=\"16.823\" r=\".645\" stroke=\"#c6ea70\"/><path stroke=\"#dff3b1\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 80.808656,18.234303 a 1.8088735,1.8088735 0 0 1 -2.309183,-0.03848 1.8088735,1.8088735 0 0 1 -0.389241,-2.276466 1.8088735,1.8088735 0 0 1 2.165211,-0.803532\"/><path stroke=\"#f2fadf\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 82.216272,17.248644 a 2.455857,2.4425189 0 0 1 -2.468226,2.016391 2.455857,2.4425189 0 0 1 -2.383743,-2.114547 2.455857,2.4425189 0 0 1 1.733108,-2.668494\"/></g></g></g></g></svg>",
                  "category": "Open Data Hub",
                  "support": "open data hub",
                  "docsLink": "http://REPLACE.THIS.URL",
                  "quickStart": "example-odhquickstart",
                  "getStartedLink": "http://REPLACE.THIS.URL",
                  "getStartedMarkDown": "# Header for Getting Started MarkDown section\nHere is where you add information about your ODHApplication in markdown format\n\n1. List item\n1. List item\n1. List item"
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhDocument",
                "metadata": {
                  "name": "example-odhdocument",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHDocument",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "type": "how-to",
                  "description": "This is an example of an ODHDocument custom resource",
                  "url": "http://REPLACE.THIS.URL",
                  "durationMinutes": 5
                }
              },
              {
                "apiVersion": "console.openshift.io/v1",
                "kind": "OdhQuickStart",
                "metadata": {
                  "name": "example-odhquickstart",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHQuickstart Name",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "durationMinutes": 5,
                  "description": "Description of an example quickstart",
                  "introduction": "### Introduction Header\nThis is the body of the introduction for an ODHQuickstart",
                  "tasks": [
                    {
                      "title": "Title of the first task",
                      "description": "### Description header for the first task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the first task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    },
                    {
                      "title": "Title of the second task",
                      "description": "### Description header for the second task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the second task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    }
                  ],
                  "conclusion": "Message for the conclusion of the ODHQuickstart",
                  "nextQuickStart": null
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v1.11.0
          createdAt: "2023-10-31T00:00:00Z"
          description: Open Data Hub Operator for deployment and management of Open
            Data Hub
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
          support: Open Data Hub
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AcceleratorProfile is the Schema for the acceleratorprofiles
            displayName: AcceleratorProfile
            kind: AcceleratorProfile
            name: acceleratorprofiles.dashboard.opendatahub.io
            version: v1alpha
          - description: KfDef is the Schema for the kfdefs API
            displayName: Kf Def
            kind: KfDef
            name: kfdefs.kfdef.apps.kubeflow.org
            version: v1
          - description: OdhApplication is the Schema for the odhapplications
            displayName: Open Data Hub Dashboard Application
            kind: OdhApplication
            name: odhapplications.dashboard.opendatahub.io
            version: v1
          - description: OdhDashboardConfig is the Schema for the odhdashboardconfigs
            displayName: Open Data Hub Dashboard Configuration
            kind: OdhDashboardConfig
            name: odhdashboardconfigs.opendatahub.io
            version: v1alpha
          - description: OdhDocument is the Schema for the odhdocuments
            displayName: Open Data Hub Dashboard Documentation
            kind: OdhDocument
            name: odhdocuments.dashboard.opendatahub.io
            version: v1
          - description: Extension for guiding user through various workflows in the
              Open Data Hub dashboard.
            displayName: Open Data Hub Dashboard QuickStart
            kind: OdhQuickStart
            name: odhquickstarts.console.openshift.io
            version: v1
        description: |
          The Open Data Hub is a machine-learning-as-a-service platform built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open Data Hub integrates multiple AI/ML open source components into one operator that can easily be downloaded and installed by OpenShift users.

          Open Data Hub operator allows users to install and manage components of the Open Data Hub. Users can mix and match tools from each project to fulfill the needs of their use case. Each of the
          projects share some components, but can be mostly seen as an extension of each other to provide a complete solution for both novice and skilled enterprise users.

          ### Components
          * Open Data Hub Dashboard - A web dashboard that displays installed Open Data Hub components with easy access to component UIs and documentation
          * ODH Notebook Controller - Secure management of Jupyter Notebook in Kubernetes environments built on top of Kubeflow Notebook Controller with support for OAuth
          * Jupyter Notebooks - JupyterLab notebook that provide Python support for GPU workloads
          * Data Science Pipelines - Pipeline solution for end to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton
          * Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh, a general-purpose model serving management/routing layer.
          * TrustyAI - TrustyAI is an open source AI Explainability (XAI) Toolkit

          To install one or multiple of these components use the default KfDef provided with the operator.
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Open Data Hub
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/opendatahub/opendatahub-operator:v1.11.0
        version: 1.11.0
      entries:
      - name: opendatahub-operator.v1.11.0
        version: 1.11.0
      - name: opendatahub-operator.v1.10.1
        version: 1.10.1
      - name: opendatahub-operator.v1.10.0
        version: 1.10.0
      - name: opendatahub-operator.v1.9.0
        version: 1.9.0
      - name: opendatahub-operator.v1.8.0
        version: 1.8.0
      - name: opendatahub-operator.v1.7.0
        version: 1.7.0
      - name: opendatahub-operator.v1.6.0
        version: 1.6.0
      - name: opendatahub-operator.v1.5.0
        version: 1.5.0
      - name: opendatahub-operator.v1.4.2
        version: 1.4.2
      - name: opendatahub-operator.v1.4.1
        version: 1.4.1
      - name: opendatahub-operator.v1.4.0
        version: 1.4.0
      - name: opendatahub-operator.v1.3.0
        version: 1.3.0
      - name: opendatahub-operator.v1.2.0
        version: 1.2.0
      - name: opendatahub-operator.v1.1.2
        version: 1.1.2
      - name: opendatahub-operator.v1.1.1
        version: 1.1.1
      name: rolling
    - currentCSV: opendatahub-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kfdef.apps.kubeflow.org/v1",
                "kind": "KfDef",
                "metadata": {
                  "name": "opendatahub",
                  "namespace": "opendatahub"
                },
                "spec": {
                  "applications": [
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-common"
                        }
                      },
                      "name": "odh-common"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-dashboard"
                        }
                      },
                      "name": "odh-dashboard"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "prometheus/cluster"
                        }
                      },
                      "name": "prometheus-cluster"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "prometheus/operator"
                        }
                      },
                      "name": "prometheus-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "grafana/cluster"
                        }
                      },
                      "name": "grafana-cluster"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "grafana/grafana"
                        }
                      },
                      "name": "grafana-instance"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-notebook-controller"
                        }
                      },
                      "name": "odh-notebook-controller"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "notebook-images"
                        }
                      },
                      "name": "notebook-images"
                    },
                    {
                      "kustomizeConfig": {
                        "overlays": [
                          "odh-model-controller"
                        ],
                        "repoRef": {
                          "name": "manifests",
                          "path": "model-mesh"
                        }
                      },
                      "name": "model-mesh"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "data-science-pipelines-operator"
                        }
                      },
                      "name": "data-science-pipelines-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "trustyai-service"
                        }
                      },
                      "name": "trustyai"
                    }
                  ],
                  "repos": [
                    {
                      "name": "manifests",
                      "uri": "https://github.com/opendatahub-io/odh-manifests/tarball/v1.5"
                    }
                  ]
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhApplication",
                "metadata": {
                  "name": "example-odhapplication",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started,Model development,Model training,Notebook environments,Package management,Data Management"
                  }
                },
                "spec": {
                  "displayName": "ODHAPPLICATION DISPLAY NAME",
                  "provider": "PROVIDER_NAME",
                  "description": "Description of your ODHApplication",
                  "kfdefApplications": [
                    "<NAME OF THE ODH APPLICATION IN KFDEF>"
                  ],
                  "route": "<APPLICATION_OCP_ROUTE_OBJECT_NAME>",
                  "img": "<svg xmlns=\"http://www.w3.org/2000/svg\" role=\"img\" viewBox=\"-1.13 -1.13 58.50 48.94\"><title>Open Data Hub logo</title><g><g><g fill=\"#4d5258\" stroke-width=\".006\"><path fill=\"#bbb\" d=\"M112.30231 22.797083c-.2034 0-.35737.143517-.35737.337407 0 .192939.15397.339307.35737.339307.20339 0 .35736-.147318.35736-.339307 0-.19294-.15397-.337407-.35736-.337407zm.002.131161c.1112 0 .20149.08934.20149.206246 0 .116904-.0903.208146-.20149.208146-.1112 0-.2053-.09124-.2053-.208146 0-.116904.0931-.206246.2053-.206246zM113.04141 22.801836h-.28228v.665308h.15017v-.190088h.13211c.16918 0 .26612-.08934.26612-.244263 0-.147319-.0969-.230957-.26612-.230957zm-.008.349762h-.12451v-.225255h.12451c.0836 0 .13116.03707.13116.111202 0 .07603-.0475.114053-.13116.114053zM113.40328 22.801836v.665308h.51703v-.124508h-.36687v-.147318h.32125V23.07081h-.32125v-.144467h.35642v-.124507zM114.03814 22.801836v.665308h.14352v-.426748l.3222.426748h.13686v-.665308h-.14257v.427698L114.175 22.801836z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/><path d=\"M115.29685 22.801836h-.28228v.665308h.27658c.21004 0 .35546-.136863.35546-.332654 0-.196741-.14351-.332654-.34976-.332654zm.004.538899h-.13591v-.412491h.12926c.11405 0 .19959.08459.19959.207196 0 .121657-.0817.205295-.19294.205295zM116.21878 23.467144h.16062l-.28418-.665308h-.15492l-.28894.665308h.15493l.0523-.12926h.30794zm-.3108-.249966l.1055-.25852.10455.25852zM116.31525 22.801836v.127359h.20244v.537949h.15017v-.537949h.2034v-.127359zM117.37527 23.467144h.16063l-.28419-.665308h-.15492l-.28893.665308h.15492l.0523-.12926h.30795zm-.31079-.249966l.1055-.25852.10454.25852zM117.83433 22.801836v.665308h.15017v-.262322h.30129v.262322h.15017v-.665308h-.15017v.277528h-.30129v-.277528zM118.72025 23.188665v-.386829h-.15017v.386829c0 .175831.11785.285132.30224.285132.18343 0 .29844-.109301.29844-.285132v-.386829h-.15017v.386829c0 .09599-.0589.153971-.14827.153971-.0893 0-.15207-.05798-.15207-.153971zM119.61059 22.801836h-.30604v.665308h.31649c.15492 0 .25092-.07033.25092-.184385 0-.08744-.0532-.152071-.13876-.168228.0713-.01806.115-.07413.115-.146368 0-.102648-.0912-.166327-.23761-.166327zm-.008.268024h-.14826v-.145418h.14826c.0561 0 .0913.02566.0913.07128 0 .04657-.0352.07413-.0913.07413zm0 .274677h-.14826v-.158723h.14826c.0713 0 .11406.02946.11406.07794 0 .05132-.0428.08079-.11406.08079z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g><g fill=\"#58595b\" stroke-width=\".01\"><path d=\"M112.19192 24.15027h.0381l-.13717-.306759h-.0364l-.1376.306759h.0372l.0342-.07844h.16696zm-.18712-.110871l.0688-.157763.0697.157763zM112.27858 23.843511v.306759h.0351v-.306759zM112.64205 23.843511h-.11525v.306759h.0351v-.09729h.0802c.0758 0 .11964-.039.11964-.10649 0-.0653-.0438-.102983-.11964-.102983zm-.001.177044h-.0789V23.87594h.0789c.0552 0 .0872.02454.0872.07143 0 .0482-.032.07318-.0872.07318zM112.81775 23.825105v.325165h.0342v-.325165zM113.1137 24.15027l-.00044-.152503c-.00044-.04908-.032-.07932-.089-.07932-.0394 0-.0644.01139-.0942.03111l.0145.0241c.0241-.01709.0478-.02673.0732-.02673.0408 0 .0618.0206.0618.05434v.01446h-.0758c-.0543.000439-.0863.02586-.0863.06617 0 .03944.0307.06968.0784.06968.0368 0 .0662-.01227.0846-.03725v.03593zm-.11131-.02717c-.0324 0-.053-.01709-.053-.04251 0-.02585.0202-.03769.0587-.03769h.0719v.02805c-.009.03287-.0386.05215-.0776.05215zM113.29915 24.108201c-.0127.0074-.025.01139-.0368.01183-.0197 0-.0307-.01139-.0307-.03944v-.127086h.0732v-.02673h-.0732v-.06048h-.0337v.06048h-.0333v.02673h.0333v.132783c0 .04558.025.06573.0587.0653.0184 0 .0359-.0061.0521-.01709zM113.44228 23.854466c.0118 0 .0245.0044.0351.01183l.014-.02629c-.014-.01052-.032-.01622-.0495-.01622-.0386-.000438-.0741.02717-.0741.07713v.02586h-.0298v.02673h.0298v.196764h.0337v-.196764h.0662v-.02673h-.0662v-.03024c0-.02717.0184-.04251.0408-.04207zM113.59236 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM113.80323 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM114.22747 23.918448c-.0443.000438-.0767.01972-.0911.06004-.01-.03813-.0394-.06004-.0828-.06004-.0416.000438-.0727.01753-.0881.05346v-.05215h-.0337v.230508h.0337v-.113501c0-.04996.0285-.08458.0758-.08589.0403 0 .064.02454.064.06617v.133221h.0346v-.113501c0-.04996.0276-.08458.0754-.08589.0399 0 .064.02454.064.06617v.133221h.0342V24.00828c0-.05609-.032-.08984-.0859-.08984zM114.63474 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0487.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11088-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM114.89934 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM115.02844 23.919762l.0863.230508h.0355l.0675-.188876.0679.188876h.0355l.0855-.230508h-.0355l-.0675.195888-.0679-.195888h-.0359l-.067.195888-.0684-.195888zM115.53009 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM115.73075 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM115.95034 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11482.116569.0386 0 .071-.01402.0929-.03813l-.0188-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18843c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM116.29474 23.825105v.144615c-.0171-.03287-.0491-.05171-.0903-.05171-.0653 0-.11087.04777-.11087.116569 0 .0688.0456.117445.11131.117445.0412 0 .0727-.01884.0898-.05127v.04952h.0342v-.325165zm-.0833.297557c-.0491 0-.0837-.03637-.0837-.08721 0-.05083.0346-.08765.0837-.08765.0486 0 .0833.03681.0833.08765 0 .05083-.0346.08721-.0833.08721zM116.65984 23.918448c-.0403 0-.0719.01884-.089.05083v-.144177h-.0342v.325165h.0342v-.04952c.0171.03199.0486.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM116.89657 24.183138l.10605-.263376h-.0351l-.0732.191944-.0767-.191944h-.0359l.0955.228756-.01.02191c-.009.02191-.0241.03331-.0421.03331-.0114 0-.0219-.0035-.0324-.01095l-.0145.02717c.014.01139.0294.01665.0486.01665.0307 0 .0548-.01841.0692-.05346zM117.29947 23.842634c-.0898 0-.16083.06793-.16083.154257 0 .08633.071.154694.16083.154694.0894 0 .16039-.06836.16039-.154694 0-.08633-.071-.154257-.16039-.154257zm0 .03243c.0684 0 .12446.0539.12446.121828 0 .06792-.0561.122703-.12446.122703-.0684 0-.12533-.05478-.12533-.122703 0-.06793.057-.121828.12533-.121828zM117.64414 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0486.05083.0898.05083.0653 0 .10999-.04777.10999-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM117.9061 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM118.19616 23.918448c-.0421.000438-.0732.01709-.0894.05215v-.05084h-.0342v.230508h.0342v-.12358c.004-.0447.032-.07494.0776-.07537.0408 0 .0653.02454.0653.06573v.133221h.0342v-.141986c0-.05609-.0324-.08984-.0876-.08984zM118.65836 23.901795l.0149-.03111c-.0254-.01709-.0618-.02848-.0973-.02848-.0635 0-.10605.03068-.10605.07888 0 .109995.17266.06661.17222.148559 0 .0298-.0276.04777-.0719.04777-.0346 0-.0736-.01578-.1021-.04295l-.0158.03024c.0289.02761.0727.04689.11745.04689.0657 0 .11087-.03287.11087-.0837.00044-.11131-.17223-.06968-.17223-.149436 0-.02629.0259-.04207.0653-.04207.0245 0 .057.0075.0846.02542zM118.84168 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .11701-.04777.11701-.117007 0-.06836-.0487-.11613-.11701-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM119.18446 23.919762v.113063c0 .04996-.0281.08502-.0758.08589-.0399 0-.064-.02454-.064-.06573v-.133222h-.0342v.141986c0 .05565.0324.08984.0863.08984.0416-.000438.0723-.01753.0876-.05346v.05215h.0342v-.230508zM119.33586 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM119.62557 23.976294l.0188-.02366c-.021-.02147-.0517-.03418-.089-.03418-.0666 0-.11438.0482-.11438.116568 0 .0688.0473.116569.11438.116569.0399 0 .0719-.0149.0925-.03988l-.0197-.02016c-.0162.01928-.0412.0298-.0714.0298-.0478 0-.082-.03506-.082-.08633 0-.05083.0342-.08589.082-.08589.028 0 .0513.01008.0688.02717zM119.78997 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g></g><g><g><path fill=\"#a586f3\" d=\"m 7808.8125,2097.1875 c -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 v 329.875 c -0.8781,10.2195 7.1804,19 17.4375,19 10.2571,0 19.0227,-8.7805 18.1446,-19 v -77.7324 l 0.4179,-6.1426 c 0,-55.1597 44.2153,-99.5625 99.375,-99.5625 55.1596,0 99.5625,44.4028 99.5625,99.5625 0.01,0.1837 0.015,0.3674 0.027,0.5508 -0.026,0.4409 -0.035,0.8826 -0.027,1.3242 v 110 c 0,9.6822 7.849,17.5312 17.5312,17.5312 9.6823,0 17.5313,-7.849 17.5313,-17.5312 v -110 c 0,-0.3245 -0.012,-0.6488 -0.031,-0.9727 0.018,-0.3004 0.029,-0.6013 0.031,-0.9023 1e-4,-74.075 -60.55,-134.4374 -134.625,-134.4375 -39.9177,1e-4 -75.8465,17.5354 -100.5,45.293 V 2115 c 0.1749,-9.8356 -7.7881,-17.8833 -17.625,-17.8125 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><path fill=\"#5ed3ee\" d=\"m 7765.8027,2098.123 c -0.038,5e-4 -0.077,0 -0.1152,0 -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 V 2361 c 0,41.9333 -26.0162,79.2671 -65.4375,93.5625 -39.4213,14.2954 -83.5594,2.3741 -110.4375,-29.8125 -26.878,-32.1865 -30.6579,-77.5719 -9.5625,-113.8125 21.0955,-36.2406 62.4262,-55.2897 103.6875,-47.8125 9.5173,1.7397 18.6463,-4.5547 20.4043,-14.0686 1.758,-9.514 -4.5187,-18.6551 -14.0293,-20.4314 -55.6354,-10.082 -111.8054,15.8219 -140.25,64.6875 -28.4446,48.8656 -23.3039,110.5381 12.9375,153.9375 36.2415,43.3993 95.908,59.4005 149.0625,40.125 44.7667,-16.2337 77.0286,-54.658 86.0977,-100.1562 0,0 1.9812,-4.7686 2.5898,-9.1563 0.6086,-4.3877 0,-17.0625 0,-17.0625 v -245.0625 c 0.1769,-9.865 -7.8313,-17.9263 -17.6973,-17.8145 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><g fill=\"none\" stroke-width=\".365\" transform=\"translate(-449.265 -83.342) scale(5.91262)\"><circle cx=\"79.677\" cy=\"16.823\" r=\".645\" stroke=\"#c6ea70\"/><path stroke=\"#dff3b1\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 80.808656,18.234303 a 1.8088735,1.8088735 0 0 1 -2.309183,-0.03848 1.8088735,1.8088735 0 0 1 -0.389241,-2.276466 1.8088735,1.8088735 0 0 1 2.165211,-0.803532\"/><path stroke=\"#f2fadf\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 82.216272,17.248644 a 2.455857,2.4425189 0 0 1 -2.468226,2.016391 2.455857,2.4425189 0 0 1 -2.383743,-2.114547 2.455857,2.4425189 0 0 1 1.733108,-2.668494\"/></g></g></g></g></svg>",
                  "category": "Open Data Hub",
                  "support": "open data hub",
                  "docsLink": "http://REPLACE.THIS.URL",
                  "quickStart": "example-odhquickstart",
                  "getStartedLink": "http://REPLACE.THIS.URL",
                  "getStartedMarkDown": "# Header for Getting Started MarkDown section\nHere is where you add information about your ODHApplication in markdown format\n\n1. List item\n1. List item\n1. List item"
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhDocument",
                "metadata": {
                  "name": "example-odhdocument",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHDocument",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "type": "how-to",
                  "description": "This is an example of an ODHDocument custom resource",
                  "url": "http://REPLACE.THIS.URL",
                  "durationMinutes": 5
                }
              },
              {
                "apiVersion": "console.openshift.io/v1",
                "kind": "OdhQuickStart",
                "metadata": {
                  "name": "example-odhquickstart",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHQuickstart Name",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "durationMinutes": 5,
                  "description": "Description of an example quickstart",
                  "introduction": "### Introduction Header\nThis is the body of the introduction for an ODHQuickstart",
                  "tasks": [
                    {
                      "title": "Title of the first task",
                      "description": "### Description header for the first task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the first task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    },
                    {
                      "title": "Title of the second task",
                      "description": "### Description header for the second task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the second task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    }
                  ],
                  "conclusion": "Message for the conclusion of the ODHQuickstart",
                  "nextQuickStart": null
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v1.5.0
          createdAt: "2022-10-18T00:00:00Z"
          description: Open Data Hub Operator for deployment and management of Open
            Data Hub
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
          support: Open Data Hub
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KfDef is the Schema for the kfdefs API
            displayName: Kf Def
            kind: KfDef
            name: kfdefs.kfdef.apps.kubeflow.org
            version: v1
          - description: OdhApplication is the Schema for the odhapplications
            displayName: Open Data Hub Dashboard Application
            kind: OdhApplication
            name: odhapplications.dashboard.opendatahub.io
            version: v1
          - description: OdhDashboardConfig is the Schema for the odhdashboardconfigs
            displayName: Open Data Hub Dashboard Configuration
            kind: OdhDashboardConfig
            name: odhdashboardconfigs.opendatahub.io
            version: v1alpha
          - description: OdhDocument is the Schema for the odhdocuments
            displayName: Open Data Hub Dashboard Documentation
            kind: OdhDocument
            name: odhdocuments.dashboard.opendatahub.io
            version: v1
          - description: Extension for guiding user through various workflows in the
              Open Data Hub dashboard.
            displayName: Open Data Hub Dashboard QuickStart
            kind: OdhQuickStart
            name: odhquickstarts.console.openshift.io
            version: v1
        description: |
          The Open Data Hub is a machine-learning-as-a-service platform built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open Data Hub integrates multiple AI/ML open source components into one operator that can easily be downloaded and installed by OpenShift users.

          Open Data Hub operator allows users to install and manage components of the Open Data Hub. Users can mix and match tools from each project to fulfill the needs of their use case. Each of the
          projects share some components, but can be mostly seen as an extension of each other to provide a complete solution for both novice and skilled enterprise users.

          ### Components
          * Open Data Hub Dashboard - A web dashboard that displays installed Open Data Hub components with easy access to component UIs and documentation
          * ODH Notebook Controller - Secure management of Jupyter Notebook in Kubernetes environments built on top of Kubeflow Notebook Controller with support for OAuth
          * Jupyter Notebooks - JupyterLab notebook that provide Python support for GPU workloads
          * Data Science Pipelines - Pipeline solution for end to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton
          * Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh, a general-purpose model serving management/routing layer.

          To install one or multiple of these components use the default KfDef provided with the operator.

          ### Available Channels

          #### Stable

          Channel `stable` offers the major releases of Open Data Hub operator and ODH manifests. It is based on Kubeflow Operator and Kustomize for deployment configuration. This version is compatible with [Open Data Hub manifests](https://github.com/opendatahub-io/odh-manifests/) as well as [Kubeflow manifests](https://github.com/opendatahub-io/manifests). This channel will offer new components and features that have been thoroughly tested by consumers of the `rolling` channel

          NOTE: The `stable` channel will be deprecated starting with the ODH 1.6 release and we will only be publishing updates on the `rolling` channel for all future releases

          #### Rolling

          Channel `rolling` will offer the latest release of Open Data Hub operator and manifests. Subscribing to this branch will provide faster updates and access to components that will be in developement and testing for the next stable release
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io
        maintainers:
        - email: maintainers@lists.opendatahub.io
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Open Data Hub
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/opendatahub/opendatahub-operator:v1.5.0
        version: 1.5.0
      entries:
      - name: opendatahub-operator.v1.5.0
        version: 1.5.0
      - name: opendatahub-operator.v1.4.2
        version: 1.4.2
      - name: opendatahub-operator.v1.4.1
        version: 1.4.1
      - name: opendatahub-operator.v1.4.0
        version: 1.4.0
      - name: opendatahub-operator.v1.3.0
        version: 1.3.0
      - name: opendatahub-operator.v1.2.0
        version: 1.2.0
      - name: opendatahub-operator.v1.1.2
        version: 1.1.2
      - name: opendatahub-operator.v1.1.1
        version: 1.1.1
      name: stable
    defaultChannel: fast
    packageName: opendatahub-operator
    provider:
      name: ODH
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Thales Group
      provider-url: https://cpl.thalesgroup.com/data-protection
    name: cte-k8s-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cte-k8s-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cte-k8s-operator.csi.cte.cpl.thalesgroup.com/v1",
                "kind": "CteK8sOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cte-k8s-operator"
                  },
                  "name": "ctek8soperator-sample"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-06-05T08:24:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CteK8sOperator is the Schema for the ctek8soperators API
            displayName: Cte K8s Operator
            kind: CteK8sOperator
            name: ctek8soperators.cte-k8s-operator.csi.cte.cpl.thalesgroup.com
            version: v1
        description: |-
          # Kubernetes Operator to deploy CTE for Kubernetes on an OCP/K8s Cluster.
          This Operator deploys CTE for Kubernetes on the OCP/K8s Cluster.

          ## About this operator

          It is a GO based operator that manages the deployment, upgrade and
          deletion of the CTE for Kubernetes on all the nodes of the Cluster on
          which it is installed.

          ## Prerequisites

          Applicable only on OCP cluster
            For an OCP cluster the operator requires a SecurityContextConstraints (SCC)
            to be created before installation. To create the SecurityContextConstraints
            you can use a definition provided by the deployment scripts at
            https://github.com/thalescpl-io/ciphertrust-transparent-encryption-kubernetes

            Look for cte-csi-scc.yaml file in deploy/kubernetes/X.Y.Z/operator-deploy/
            Where X.Y.Z is the last released version of the deploy scripts.

          For more information refer to the guide at:
            https://thalesdocs.com/ctp/cte-con/cte-k8s/latest/cte-4-k8s/index.html
        displayName: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Encryption
        - Data Protection
        - Security
        - Kubernetes
        - Openshift
        links:
        - name: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          url: https://cpl.thalesgroup.com/encryption/ciphertrust-transparent-data-encryption-kubernetes
        maintainers:
        - email: technical.support.DIS@thalesgroup.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Thales Group
          url: https://cpl.thalesgroup.com/data-protection
        relatedImages:
        - docker.io/thalesciphertrust/ciphertrust-transparent-encryption-kubernetes-operator@sha256:d58cbb65d758fd0b9d4cfba4cb48a35bdaebe2e239fb25edb9b491cd8c2c3f73
        version: 1.6.5
      entries:
      - name: cte-k8s-operator.v1.6.5
        version: 1.6.5
      name: candidate
    - currentCSV: cte-k8s-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cte-k8s-operator.csi.cte.cpl.thalesgroup.com/v1",
                "kind": "CteK8sOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cte-k8s-operator"
                  },
                  "name": "ctek8soperator-sample"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-06-05T08:24:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CteK8sOperator is the Schema for the ctek8soperators API
            displayName: Cte K8s Operator
            kind: CteK8sOperator
            name: ctek8soperators.cte-k8s-operator.csi.cte.cpl.thalesgroup.com
            version: v1
        description: |-
          # Kubernetes Operator to deploy CTE for Kubernetes on an OCP/K8s Cluster.
          This Operator deploys CTE for Kubernetes on the OCP/K8s Cluster.

          ## About this operator

          It is a GO based operator that manages the deployment, upgrade and
          deletion of the CTE for Kubernetes on all the nodes of the Cluster on
          which it is installed.

          ## Prerequisites

          Applicable only on OCP cluster
            For an OCP cluster the operator requires a SecurityContextConstraints (SCC)
            to be created before installation. To create the SecurityContextConstraints
            you can use a definition provided by the deployment scripts at
            https://github.com/thalescpl-io/ciphertrust-transparent-encryption-kubernetes

            Look for cte-csi-scc.yaml file in deploy/kubernetes/X.Y.Z/operator-deploy/
            Where X.Y.Z is the last released version of the deploy scripts.

          For more information refer to the guide at:
            https://thalesdocs.com/ctp/cte-con/cte-k8s/latest/cte-4-k8s/index.html
        displayName: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Encryption
        - Data Protection
        - Security
        - Kubernetes
        - Openshift
        links:
        - name: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          url: https://cpl.thalesgroup.com/encryption/ciphertrust-transparent-data-encryption-kubernetes
        maintainers:
        - email: technical.support.DIS@thalesgroup.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Thales Group
          url: https://cpl.thalesgroup.com/data-protection
        relatedImages:
        - docker.io/thalesciphertrust/ciphertrust-transparent-encryption-kubernetes-operator@sha256:d58cbb65d758fd0b9d4cfba4cb48a35bdaebe2e239fb25edb9b491cd8c2c3f73
        version: 1.6.5
      entries:
      - name: cte-k8s-operator.v1.6.5
        version: 1.6.5
      name: stable
    defaultChannel: stable
    packageName: cte-k8s-operator
    provider:
      name: Thales Group
      url: https://cpl.thalesgroup.com/data-protection
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-team-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-team-operator.v2024.10.2-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": null,
                        "host": null,
                        "process": null
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master",
                        "operator": "",
                        "value": ""
                      }
                    ]
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": null
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:12c1d2e9347ec447196a618b481a0970dacac7a8a72a3b38ec22efa69668d180
          createdAt: "2024-10-25T01:47:52Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2024.10.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2023.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                   operator: ""
                   value: ""

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:7c4868db021beaffa762eb0de43739bbc865a4bf5900f1d17f78caed8612c073
        - registry.connect.redhat.com/ca/uma-operator@sha256:12c1d2e9347ec447196a618b481a0970dacac7a8a72a3b38ec22efa69668d180
        version: 2024.10.2-12
      entries:
      - name: uma-team-operator.v2024.10.2-12
        version: 2024.10.2-12
      name: "2024.10"
    - currentCSV: uma-team-operator.v2025.1.3-33
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": null,
                        "host": null,
                        "process": null
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.1.1.33,990033",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master",
                        "operator": "",
                        "value": ""
                      }
                    ]
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": null
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:78b66a291f58aa94123cdd59f2aa07dce4928162c481fe9de5ce524782bc93de
          createdAt: "2025-02-18T05:35:06Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.1.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2023.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.1.1.26,990026"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                   operator: ""
                   value: ""

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:78b66a291f58aa94123cdd59f2aa07dce4928162c481fe9de5ce524782bc93de
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:ae91a28d57c8a359f81757e4bf79472f2ac189ac0ae3ee0d87e8ff9c433f1ccc
        version: 2025.1.3-33
      entries:
      - name: uma-team-operator.v2025.1.3-33
        version: 2025.1.3-33
      name: "2025.1"
    - currentCSV: uma-team-operator.v2025.2.1-25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.2.1.25,990025",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master",
                        "operator": "",
                        "value": ""
                      }
                    ]
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:d0a0090a871ff7e23d3c97b620a1b0d004a8317efe1581ba0437d073c5656825
          createdAt: "2025-03-05T05:51:44Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.2.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2023.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.1.1.26,990026"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                   operator: ""
                   value: ""

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:d0a0090a871ff7e23d3c97b620a1b0d004a8317efe1581ba0437d073c5656825
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:8bcf0dac889f38a0bf77eb9babe3f726e3b2564d5988c05bdc5c2d91ae189c61
        version: 2025.2.1-25
      entries:
      - name: uma-team-operator.v2025.2.1-25
        version: 2025.2.1-25
      name: "2025.2"
    - currentCSV: uma-team-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:60fb1f76b4d4aab569eec5f2c20f7a06d813759f5fb6ac952156b41c4d295a52
          createdAt: "2025-04-16T06:57:55Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.4.5
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.4.2.5,990005"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:60fb1f76b4d4aab569eec5f2c20f7a06d813759f5fb6ac952156b41c4d295a52
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        version: 2025.4.5-5
      entries:
      - name: uma-team-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-team-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:052055a6280f60a69caccf89f50fd8bf92da61cd59f1962f61d21467c4191970
          createdAt: "2025-07-24T14:54:58Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.5.2.12,990012"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:052055a6280f60a69caccf89f50fd8bf92da61cd59f1962f61d21467c4191970
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        version: 2025.5.4-12
      entries:
      - name: uma-team-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-team-operator.v2025.7.3-21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:e3ae2240aeb618f36f41073a5f65b03a7ad5dfdccc7f9feb830506fbd2bfd604
          createdAt: "2025-08-23T09:01:14Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.7.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.7.1.23,990023"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:e3ae2240aeb618f36f41073a5f65b03a7ad5dfdccc7f9feb830506fbd2bfd604
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.3-21
      entries:
      - name: uma-team-operator.v2025.7.3-21
        version: 2025.7.3-21
      name: "2025.7"
    - currentCSV: uma-team-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "hostmonitoring": {
                        "linux": {
                          "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:b8422d0f3a9e576948ced8244a91abc3283b26217e4c7efc7e2669ca5508930e
          createdAt: "2025-08-15T03:20:57Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 hostmonitoring:
                   linux:
                   # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                   groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.8.1.9,990009"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:b8422d0f3a9e576948ced8244a91abc3283b26217e4c7efc7e2669ca5508930e
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        version: 2025.8.2-9
      entries:
      - name: uma-team-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-team-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
          createdAt: "2025-09-24T12:24:28Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.12,990012"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-team-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-team-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "hostmonitoring": {
                        "linux": {
                          "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:8c12315745ad3564035a4db923698bf63fe32dddbc7d246f5a3867945305713e
          createdAt: "2025-07-30T08:27:50Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 hostmonitoring:
                   linux:
                   # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                   groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "99.99.0.serp_F152813,19"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:8c12315745ad3564035a4db923698bf63fe32dddbc7d246f5a3867945305713e
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        version: 2025.7.2-19
      entries:
      - name: uma-team-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-team-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:89e0505b5e526beffda9b13a1ca7b49e5ba144fd0166a7e171e9d9b084915a56
          createdAt: "2025-09-11T11:16:37Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.8,990008"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:89e0505b5e526beffda9b13a1ca7b49e5ba144fd0166a7e171e9d9b084915a56
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-team-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-team-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
          createdAt: "2025-09-24T12:24:28Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.12,990012"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-team-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-team-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ocs-client-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocs-client-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          console.openshift.io/plugins: '["odf-client-console"]'
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageClassClaim is the Schema for the storageclassclaims
              API
            displayName: Storage Class Claim
            kind: StorageClassClaim
            name: storageclassclaims.ocs.openshift.io
            version: v1alpha1
          - description: StorageClient is the Schema for the storageclients API
            displayName: Storage Client
            kind: StorageClient
            name: storageclients.ocs.openshift.io
            version: v1alpha1
        description: This operator is not intended to be directly installed
        displayName: OpenShift Data Foundation Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: OCS Client Operator
          url: https://github.com/red-hat-storage/ocs-client-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:975529d3fbdcf6f78cc1602333832f543da2918ac5c018d1ee6bea2b315106be
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:ef4276b80cacae825d4831de88d3fe6ba7cc324464087b7b28734bd2b8de3315
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel8@sha256:d83b44a1257e8015409cd261e8eed5bd8c4e581b788e8452f19114ba696a30c1
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:bc637bbbed3457c45fbb5cb1e8940b2a1dacfa64959e5bd0bfb58c5d669427fd
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:1dd7c0ef17d39ebf3311aee326202e09038cece1b7554cea23c1efe582620d0b
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:c9e5b48d84febf10ff4c9b78a8210df36c90905fbb92b6a6b5b32e0fd9a6a41d
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:9a0008f662b6afa9e2677a8a83bd662d806ad6f4d03fb49bccb78ed487ce4c7a
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:028120f2212738917f49e0330b60be5f79a208ce2974bea22b71f533aa6ce3be
        - registry.redhat.io/odf4/ocs-client-console-rhel9@sha256:ebf3f6098a9a0a06706d20fe8fc86513be34b98604bc70fd82910f0fcd17f191
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:ffcc0806ec2afb8b40d75c70ef2a6d98bb887879f5b3f573e900861316ce9e41
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel8@sha256:05cf61192ee75dc11f5777bebeacdd209232042445553329e14f542ac0084c71
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:e206dcc33863f4d8d0aa6df173a86cd50ee1d52e6bcf9efc43b39831acc20573
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:0dc0733df67bb49886ac9715b81bd4ec4a36330fe1afdde657587e1883f5ca7a
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel8@sha256:647074ada8cb88f7ed99ca14621d3b754db6a60346f12b7561cd035c05f283f7
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:390b527d10e9c7a8a9191251b0a1ba501128057cefaff35e950f0efe04ef3122
        - registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:cb93b68adb857c82f23f292cdee47d52656dc4fc5bb83fac8581110869cbe841
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:645d4fd09628f7ddcfc765e4ddacaa79fa6a7c35c03292ca2f6e317d48765d6b
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel8@sha256:1f6a17909f1fe6e09a01cad38b6a50b41d850b1c5cffd24ef9a87c8798ff15e7
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:1711a0aabd28937aac1424adebe96619a48b91f7f888f0588117f80891001da4
        version: 4.15.20-rhodf
      entries:
      - name: ocs-client-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: ocs-client-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: ocs-client-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: ocs-client-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: ocs-client-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: ocs-client-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: ocs-client-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: ocs-client-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: ocs-client-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: ocs-client-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: ocs-client-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: ocs-client-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: ocs-client-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: ocs-client-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: ocs-client-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: ocs-client-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: ocs-client-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: ocs-client-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: ocs-client-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: ocs-client-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: ocs-client-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-client-console"]'
          containerImage: registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:bd8454b5a0b235669732650ae8502831811793d93fbfbd0d0f53278fcdfed459
          createdAt: "2025-09-12T06:29:47Z"
          description: OpenShift Data Foundation client operator enables consumption
            of storage services from a remote centralized OpenShift Data Foundation
            provider cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["storageclaims.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/ocs-client-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageClaim is the Schema for the storageclaims API
            displayName: Storage Claim
            kind: StorageClaim
            name: storageclaims.ocs.openshift.io
            version: v1alpha1
          - description: StorageClient is the Schema for the storageclients API
            displayName: Storage Client
            kind: StorageClient
            name: storageclients.ocs.openshift.io
            version: v1alpha1
        description: OpenShift Data Foundation client operator enables consumption
          of storage services from a remote centralized OpenShift Data Foundation
          provider cluster.
        displayName: OpenShift Data Foundation Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: OCS Client Operator
          url: https://github.com/red-hat-storage/ocs-client-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:80c01633a6bfca44df37ac38f764035be6998fce965c8709e4f6d54d17eeaa7b
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:82a4b481028195cfd75471656fe40c1cbbade4c7b2d170d6bf9422d9d3b84fa7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel8@sha256:d83b44a1257e8015409cd261e8eed5bd8c4e581b788e8452f19114ba696a30c1
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel8@sha256:1f6a17909f1fe6e09a01cad38b6a50b41d850b1c5cffd24ef9a87c8798ff15e7
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:975529d3fbdcf6f78cc1602333832f543da2918ac5c018d1ee6bea2b315106be
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c11a599bfd09b9b22888e03eabdbe1054d4049718a591d2a45bf0ed89800111e
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6fee56851da3a87b9f7f477fc59e3385e118dc0b398861a8f13459c998a5af77
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:0afcb516eb528857887c6ee1a9cba081dc8f6b7d47d9f26a0dbb2dca0cefe07c
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:8d8326c74795c9c6bd8f33a9f9e81d2304a70b1f9be3c2f8d50e76be4b0292ff
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/odf4/ocs-client-console-rhel9@sha256:beb4f82d3923171b80aacb90dfe345a22626453bf2f3921b03a16839b204d377
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:1711a0aabd28937aac1424adebe96619a48b91f7f888f0588117f80891001da4
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:ef4276b80cacae825d4831de88d3fe6ba7cc324464087b7b28734bd2b8de3315
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:028120f2212738917f49e0330b60be5f79a208ce2974bea22b71f533aa6ce3be
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:645d4fd09628f7ddcfc765e4ddacaa79fa6a7c35c03292ca2f6e317d48765d6b
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:bc637bbbed3457c45fbb5cb1e8940b2a1dacfa64959e5bd0bfb58c5d669427fd
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:ffcc0806ec2afb8b40d75c70ef2a6d98bb887879f5b3f573e900861316ce9e41
        - registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:bd8454b5a0b235669732650ae8502831811793d93fbfbd0d0f53278fcdfed459
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:693a60d6a1162659213ae554b1c6ee200b0300a7da72c01415044cf507acab49
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:c9e5b48d84febf10ff4c9b78a8210df36c90905fbb92b6a6b5b32e0fd9a6a41d
        version: 4.16.16-rhodf
      entries:
      - name: ocs-client-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: ocs-client-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: ocs-client-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: ocs-client-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: ocs-client-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: ocs-client-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: ocs-client-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: ocs-client-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: ocs-client-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: ocs-client-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: ocs-client-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: ocs-client-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: ocs-client-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: ocs-client-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: ocs-client-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: ocs-client-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: ocs-client-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: ocs-client-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://your.domain
    name: amq-streams-proxy
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: alpha
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: amq-streams-3.0.x
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: amq-streams-3.x
    defaultChannel: alpha
    packageName: amq-streams-proxy
    provider:
      name: Red Hat
      url: https://your.domain
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-secretsmanager-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-secretsmanager-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secretsmanager.services.k8s.aws/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/secretsmanager-controller:1.1.1
          createdAt: "2025-09-19T21:14:43Z"
          description: AWS Secrets Manager controller is a service controller for
            managing Secrets Manager resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret represents the state of an AWS secretsmanager Secret
              resource.
            displayName: Secret
            kind: Secret
            name: secrets.secretsmanager.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Secrets Manager resources in AWS from within your Kubernetes cluster.

          **About Amazon Secrets Manager**

          AWS Secrets Manager helps you manage, retrieve, and rotate database credentials, application credentials, OAuth tokens, API keys, and other secrets throughout their lifecycles. [Many AWS services that use secrets](https://docs.aws.amazon.com/secretsmanager/latest/userguide/integrating.html) store them in Secrets Manager.

          Secrets Manager helps you improve your security posture, because you no longer need hard-coded credentials in application source code. Storing the credentials in Secrets Manager helps avoid possible compromise by anyone who can inspect your application or the components. You replace hard-coded credentials with a runtime call to the Secrets Manager service to retrieve credentials dynamically when you need them.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Secrets Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secretsmanager
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Secrets Manager Developer Resources
          url: https://aws.amazon.com/secrets-manager/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: secretsmanager maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/secretsmanager-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-secretsmanager-controller.v1.1.1
        version: 1.1.1
      - name: ack-secretsmanager-controller.v1.1.0
        version: 1.1.0
      - name: ack-secretsmanager-controller.v1.0.14
        version: 1.0.14
      - name: ack-secretsmanager-controller.v1.0.13
        version: 1.0.13
      - name: ack-secretsmanager-controller.v1.0.12
        version: 1.0.12
      - name: ack-secretsmanager-controller.v1.0.11
        version: 1.0.11
      - name: ack-secretsmanager-controller.v1.0.10
        version: 1.0.10
      - name: ack-secretsmanager-controller.v1.0.9
        version: 1.0.9
      - name: ack-secretsmanager-controller.v1.0.8
        version: 1.0.8
      - name: ack-secretsmanager-controller.v1.0.7
        version: 1.0.7
      - name: ack-secretsmanager-controller.v1.0.6
        version: 1.0.6
      - name: ack-secretsmanager-controller.v1.0.5
        version: 1.0.5
      - name: ack-secretsmanager-controller.v1.0.4
        version: 1.0.4
      - name: ack-secretsmanager-controller.v1.0.3
        version: 1.0.3
      - name: ack-secretsmanager-controller.v1.0.1
        version: 1.0.1
      - name: ack-secretsmanager-controller.v1.0.0
        version: 1.0.0
      - name: ack-secretsmanager-controller.v0.0.11
        version: 0.0.11
      - name: ack-secretsmanager-controller.v0.0.10
        version: 0.0.10
      - name: ack-secretsmanager-controller.v0.0.9
        version: 0.0.9
      - name: ack-secretsmanager-controller.v0.0.8
        version: 0.0.8
      - name: ack-secretsmanager-controller.v0.0.7
        version: 0.0.7
      - name: ack-secretsmanager-controller.v0.0.2
        version: 0.0.2
      - name: ack-secretsmanager-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-secretsmanager-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: crunchy-postgres-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:c559c896e174674d151f93543ac92a1b5647eb4676313f7dc6cbb8faee786f18
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:4bdd0588ffbb3f628e7df1c72670e7a5087269438343b68779012bc3a13dea31
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-exporter@sha256:3b231902b2116cd38de05c7378202716c95813b9961fb8cee811b61d36cddf86
        - registry.connect.redhat.com/crunchydata/crunchy-upgrade@sha256:68bc93cce61d9b1e44822ae3e3ac8f7a9c2b712eddfaea090eabefacd60c67f6
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:0e283616c2d32e53258d62d02702bc4e179ef8c03933592154babed1a70bf666
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:56a5e3a5758982e086c9f78201a7137e0d934316f058a31cf30df44639a569a2
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:b2c3c74fdcc0e08f555e50c527f0d322292333c5d3dce061a0dee3680a1fbc56
        - registry.connect.redhat.com/crunchydata/crunchy-pgbackrest@sha256:b70da0035b74379f2507cfabdd9d1cc1e3c3eecb9066c22f292981c9a27916e1
        - registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
        - registry.connect.redhat.com/crunchydata/crunchy-pgadmin4@sha256:0d84acb1ce6a9b046e3f8b222a583ddcc91dda0a117c9f637ef173d0ac992f57
        - registry.connect.redhat.com/crunchydata/crunchy-pgbouncer@sha256:17cb58dec78447c2f70a414a2aeff6facbf921b42a0f366d0dbf7be95aa8987a
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:a7f62f367817fc08294a0b8b82ac9826a42c0ae668efdfacaa158a4287b8a688
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.4
        version: 5.4.4
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.4.0
        version: 5.4.0
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: crunchy-postgres-operator
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hbase-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hbase-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:e56051b66b0f419874d113c0320232dba7dd2547357256cf80eeaf18bbecb298
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hbase-operator@sha256:e56051b66b0f419874d113c0320232dba7dd2547357256cf80eeaf18bbecb298
        version: 24.11.1
      entries:
      - name: hbase-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: hbase-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:232ff30ffdf4bdc33f49d87950864e15436957510ba47f352dc0b133cb43279d
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hbase-operator@sha256:232ff30ffdf4bdc33f49d87950864e15436957510ba47f352dc0b133cb43279d
        version: 25.3.0
      entries:
      - name: hbase-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: hbase-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
        version: 25.7.0
      entries:
      - name: hbase-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hbase-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
        version: 25.7.0
      entries:
      - name: hbase-operator.v25.7.0
        version: 25.7.0
      - name: hbase-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hbase-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: cinder-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cinder-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cinder.openstack.org/v1beta1",
                "kind": "Cinder",
                "metadata": {
                  "name": "cinder",
                  "namespace": "openstack"
                },
                "spec": {
                  "cinderAPI": {},
                  "cinderBackup": {},
                  "cinderScheduler": {},
                  "cinderVolumes": {
                    "volume1": {}
                  },
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "cinder",
                  "databaseInstance": "openstack",
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "cinder-secret",
                  "serviceUser": "cinder"
                }
              },
              {
                "apiVersion": "cinder.openstack.org/v1beta1",
                "kind": "CinderAPI",
                "metadata": {
                  "name": "cinderapi-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cinder.openstack.org/v1beta1",
                "kind": "CinderBackup",
                "metadata": {
                  "name": "cinderbackup-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cinder.openstack.org/v1beta1",
                "kind": "CinderScheduler",
                "metadata": {
                  "name": "cinderscheduler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cinder.openstack.org/v1beta1",
                "kind": "CinderVolume",
                "metadata": {
                  "name": "cindervolume-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:43Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CinderAPI is the Schema for the cinderapis API
            displayName: Cinder API
            kind: CinderAPI
            name: cinderapis.cinder.openstack.org
            version: v1beta1
          - description: CinderBackup is the Schema for the cinderbackups API
            displayName: Cinder Backup
            kind: CinderBackup
            name: cinderbackups.cinder.openstack.org
            version: v1beta1
          - description: Cinder is the Schema for the cinders API
            displayName: Cinder
            kind: Cinder
            name: cinders.cinder.openstack.org
            version: v1beta1
          - description: CinderScheduler is the Schema for the cinderschedulers API
            displayName: Cinder Scheduler
            kind: CinderScheduler
            name: cinderschedulers.cinder.openstack.org
            version: v1beta1
          - description: CinderVolume is the Schema for the cindervolumes API
            displayName: Cinder Volume
            kind: CinderVolume
            name: cindervolumes.cinder.openstack.org
            version: v1beta1
        description: Cinder Operator
        displayName: Cinder Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Block Storage
        - Cinder
        links:
        - name: Cinder Operator
          url: https://github.com/openstack-k8s-operators/cinder-operator
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/cinder-rhel9-operator@sha256:e0e457500af9db0831c7080b10e010d11213cd800d89e40683974aaf7966a68d
        - registry.redhat.io/rhoso/openstack-cinder-api-rhel9@sha256:ec6835efb7ffc4fa7eed10bd029440263237e8f792e7e43d4cda337d00fecdb6
        - registry.redhat.io/rhoso/openstack-cinder-backup-rhel9@sha256:936e10d80300c325d0f58b4d6ca290fcf3c5cad774eb5cb696330a20c0151561
        - registry.redhat.io/rhoso/openstack-cinder-scheduler-rhel9@sha256:753775cc575321c727b9a6371daedc8774ba3a2287c96f3d44ad8f7025f63841
        - registry.redhat.io/rhoso/openstack-cinder-volume-rhel9@sha256:47b19d993bdfc790c508861dbcc87f0ec51302db8975698af8e1476a0f5209c7
        version: 1.0.6
      entries:
      - name: cinder-operator.v1.0.6
        version: 1.0.6
      - name: cinder-operator.v1.0.5
        version: 1.0.5
      - name: cinder-operator.v1.0.4
        version: 1.0.4
      - name: cinder-operator.v1.0.3
        version: 1.0.3
      - name: cinder-operator.v1.0.2
        version: 1.0.2
      - name: cinder-operator.v1.0.1
        version: 1.0.1
      - name: cinder-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: cinder-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: node-healthcheck-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-healthcheck-operator.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:60d95ea8240ecf701ba974b6df3ee296a7fb3ef5393e0dcf43af15e782f2d153
          createdAt: "2023-11-02 11:44:50"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.3.0 <0.4.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://docs.openshift.com/container-platform/4.12/nodes/nodes/eco-node-health-check-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:60d95ea8240ecf701ba974b6df3ee296a7fb3ef5393e0dcf43af15e782f2d153
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:8d32dee5c72ec1c94e4e92f8d1da935a384e40e379f237c432086a2781f4ecb9
        version: 0.4.1
      entries:
      - name: node-healthcheck-operator.v0.4.1
        version: 0.4.1
      name: 4.12-eus
    - currentCSV: node-healthcheck-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:90cceb176fa903cc8e9aa4c0f7316560a8a354a3138dc26abbfc9884c1c576b5
          createdAt: "2023-10-30 14:21:41"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.4.0 <0.6.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/node-health-check-operator
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:90cceb176fa903cc8e9aa4c0f7316560a8a354a3138dc26abbfc9884c1c576b5
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:b8d5db068bfc81031c2eef86e8649e4fc008b7480e08c9cfab2ab4497098eecf
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel8@sha256:85054b34e119c16cc95cda9dc3f5d63f2d7bfc70634f70973410b2a1b1fa58d4
        version: 0.6.1
      entries:
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.14-eus
    - currentCSV: node-healthcheck-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:4811046e26846b1f12a3b20ddc6209980897c11b171ddd61d26b43630cc1d4b4
          createdAt: "2024-08-09T14:54:37Z"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.8.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          SNR is installed automatically when installing NHC.

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/node-health-check-operator
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:4811046e26846b1f12a3b20ddc6209980897c11b171ddd61d26b43630cc1d4b4
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:942f69efc7ae05d6c4d9c33335e4be71bfae2e6f9ef062dcd7887eebfe02120e
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel8@sha256:062cd9def39a087ae3802cd5208e2f0d5bfca29ba76ebd5139d88348e8c9e923
        version: 0.8.2
      entries:
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.16-eus
    - currentCSV: node-healthcheck-operator.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:f174ada653bb5fb7e58219510d0b10d868273a70db5c486e0b3ad05a1be50640
          createdAt: "2025-08-04 12:38:05"
          description: Detect failed Nodes and trigger remediation with a remediation
            operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <0.9.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.7/html/remediation_fencing_and_maintenance
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel9@sha256:8e9926558c3d8082f298e9e5f58bc5b473020acd44c2ce853c2916cdad71d649
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:f174ada653bb5fb7e58219510d0b10d868273a70db5c486e0b3ad05a1be50640
        - registry.redhat.io/workload-availability/node-remediation-console-rhel9@sha256:37f62407284fe574b973a1a685371d191b710ad70a6bfe8172f19a4568c11cf6
        version: 0.9.1
      entries:
      - name: node-healthcheck-operator.v0.9.1
        version: 0.9.1
      - name: node-healthcheck-operator.v0.9.0
        version: 0.9.0
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.18-eus
    - currentCSV: node-healthcheck-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:8e1dbc35126101e1025bd4c0eac4f4c614eb779ccafbb55050541a27ba7e9706
          createdAt: "2022-08-15 12:34:34"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.2.0 <0.3.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-node-health-check-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b5786bbbef725badf3dfcc2c2c7a86ead5ebb584c978c47aae8b9a62e241b80d
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:8e1dbc35126101e1025bd4c0eac4f4c614eb779ccafbb55050541a27ba7e9706
        version: 0.3.1
      entries:
      - name: node-healthcheck-operator.v0.3.1
        version: 0.3.1
      name: candidate
    - currentCSV: node-healthcheck-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:2160f2688e5b098c0bae7597cebcd380019b82ae02f4506d07c2ed85f0b3664d
          createdAt: "2025-09-01 18:07:56"
          description: Detect failed Nodes and trigger remediation with a remediation
            operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.8/html/remediation_fencing_and_maintenance
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:2160f2688e5b098c0bae7597cebcd380019b82ae02f4506d07c2ed85f0b3664d
        - registry.redhat.io/workload-availability/node-remediation-console-rhel9@sha256:cc0f671b126cde10476f8cc2061c6903932e7bf5e3dbf4e8bdda9add70d5deea
        version: 0.10.0
      entries:
      - name: node-healthcheck-operator.v0.10.0
        version: 0.10.0
      - name: node-healthcheck-operator.v0.9.1
        version: 0.9.1
      - name: node-healthcheck-operator.v0.9.0
        version: 0.9.0
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: stable
    defaultChannel: stable
    packageName: node-healthcheck-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-mq-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-mq-controller.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mq.services.k8s.aws/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/mq-controller:1.1.2
          createdAt: "2025-10-03T18:49:51Z"
          description: AWS MQ controller is a service controller for managing MQ resources
            in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker represents the state of an AWS mq Broker resource.
            displayName: Broker
            kind: Broker
            name: brokers.mq.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon MQ resources in AWS from within your Kubernetes cluster.

          **About Amazon MQ**

          Amazon MQ is a managed message broker service for Apache ActiveMQ and RabbitMQ that makes it easy to set up and operate message brokers on AWS. Amazon MQ reduces your operational responsibilities by managing the provisioning, setup, and maintenance of message brokers for you. Because Amazon MQ connects to your current applications with industry-standard APIs and protocols, you can easily migrate to AWS without having to rewrite code.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon MQ
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mq
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon MQ Developer Resources
          url: https://docs.aws.amazon.com/amazon-mq/latest/developer-guide/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: mq maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/mq-controller:1.1.2
        version: 1.1.2
      entries:
      - name: ack-mq-controller.v1.1.2
        version: 1.1.2
      - name: ack-mq-controller.v1.1.1
        version: 1.1.1
      - name: ack-mq-controller.v1.1.0
        version: 1.1.0
      - name: ack-mq-controller.v1.0.13
        version: 1.0.13
      - name: ack-mq-controller.v1.0.11
        version: 1.0.11
      - name: ack-mq-controller.v1.0.10
        version: 1.0.10
      - name: ack-mq-controller.v1.0.9
        version: 1.0.9
      - name: ack-mq-controller.v1.0.8
        version: 1.0.8
      - name: ack-mq-controller.v1.0.7
        version: 1.0.7
      - name: ack-mq-controller.v1.0.6
        version: 1.0.6
      - name: ack-mq-controller.v1.0.5
        version: 1.0.5
      - name: ack-mq-controller.v1.0.2
        version: 1.0.2
      - name: ack-mq-controller.v1.0.1
        version: 1.0.1
      - name: ack-mq-controller.v1.0.0
        version: 1.0.0
      - name: ack-mq-controller.v0.0.36
        version: 0.0.36
      - name: ack-mq-controller.v0.0.35
        version: 0.0.35
      - name: ack-mq-controller.v0.0.34
        version: 0.0.34
      - name: ack-mq-controller.v0.0.31
        version: 0.0.31
      - name: ack-mq-controller.v0.0.29
        version: 0.0.29
      - name: ack-mq-controller.v0.0.27
        version: 0.0.27
      - name: ack-mq-controller.v0.0.26
        version: 0.0.26
      - name: ack-mq-controller.v0.0.25
        version: 0.0.25
      - name: ack-mq-controller.v0.0.24
        version: 0.0.24
      - name: ack-mq-controller.v0.0.23
        version: 0.0.23
      - name: ack-mq-controller.v0.0.22
        version: 0.0.22
      - name: ack-mq-controller.v0.0.21
        version: 0.0.21
      - name: ack-mq-controller.v0.0.20
        version: 0.0.20
      - name: ack-mq-controller.v0.0.19
        version: 0.0.19
      - name: ack-mq-controller.v0.0.18
        version: 0.0.18
      - name: ack-mq-controller.v0.0.17
        version: 0.0.17
      - name: ack-mq-controller.v0.0.16
        version: 0.0.16
      - name: ack-mq-controller.v0.0.15
        version: 0.0.15
      - name: ack-mq-controller.v0.0.14
        version: 0.0.14
      - name: ack-mq-controller.v0.0.13
        version: 0.0.13
      - name: ack-mq-controller.v0.0.12
        version: 0.0.12
      - name: ack-mq-controller.v0.0.11
        version: 0.0.11
      - name: ack-mq-controller.v0.0.10
        version: 0.0.10
      - name: ack-mq-controller.v0.0.9
        version: 0.0.9
      - name: ack-mq-controller.v0.0.8
        version: 0.0.8
      - name: ack-mq-controller.v0.0.7
        version: 0.0.7
      name: alpha
    defaultChannel: alpha
    packageName: ack-mq-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Accuknox Inc.
      provider-url: https://accuknox.com
    name: accuknox-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: accuknox-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.accuknox.com/v1",
                "kind": "AccuknoxConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "accuknox-operator",
                    "app.kubernetes.io/instance": "accuknoxconfig",
                    "app.kubernetes.io/name": "accuknoxconfig-sample",
                    "app.kubernetes.io/part-of": "accuknox-operator"
                  },
                  "name": "accuknoxconfig-sample",
                  "namespace": "accuknox-agents"
                },
                "spec": {
                  "ppsHost": "PLEASE_FILL_IN",
                  "knoxGateway": "PLEASE_FILL_IN",
                  "spireHost": "PLEASE_FILL_IN"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Security
          containerImage: public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:b6fae25e229fb7fca757872fc0460642d52c4f9906f776afac50c54a69b05b68
          createdAt: "2024-02-09T17:06:30Z"
          operatorframework.io/suggested-namespace: accuknox-agents
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: public.ecr.aws/k9v9d5v2/accuknox-operator
          support: Accuknox Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccuknoxConfig is the Schema for the accuknoxconfigs API
            kind: AccuknoxConfig
            name: accuknoxconfigs.operator.accuknox.com
            version: v1
        description: |2-

          The AccuKnox Operator brings native K8s support for deploying and managing AccuKnox Components on a Kubernetes cluster.
          The AccuKnox Operator helps to easily onboard a k8s cluster to the AccuKnox SaaS.

          AccuKnox Operator deploys and manages the components:

          1. Discovery Engine

          2. Policy Enforcement Agent

          3. Shared Informer Agent

          4. Feeder Service

          5. Agents Operator

          ## Prerequisites for using this Operator

          1. Get AccuKnox licence - https://app.accuknox.com/sign-up

          2. After Successful Sign-up into AccuKnox Application.

             Please naviagate to Settings >> Manager Cluster >> Onboard Cluster >>

             After successful onboarding, Copy the join token created by AccuKnox

          3. Create a secret using jointoken by executing the below command:

          ```
          oc create secret generic accuknox-jointoken --from-literal=token=<joinToken> -n accuknox-agents
          ```

          ## Deploy AccuKnox Operator using AccuknoxConfig Custom Resource

          All the information to configure the AccuKnox Operator will be provided via Accuknox during Cluster Onboarding steps.

          ### AccuknoxConfig Spec

          | Spec | Description |
          | ---- | ----------- |
          | spireHost | "spire host url" |
          | ppsHost   |  "pps host url" |
          | knoxGateway | "knox gateway url" |
          | joinTokenSecretName | "join token secret name if it's different than accuknox-jointoken" |
          | joinTokenSecretKey | "key name used to store the jointoken with the secret" |

          *Example*

          ```
          apiVersion: operator.accuknox.com/v1
          kind: AccuknoxConfig
          metadata:
            name: accuknoxconfig-default
            namespace: accuknox-agents
          spec:
            ppsHost: <ppsHost url>
            knoxGateway: <knox gateway url>
            spireHost: <spire host url>
          ```
        displayName: Accuknox Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CNAPP
        - CSPM
        - CWPP
        - Cloud Native
        - Runtime Security
        - Zero-Trust
        maturity: alpha
        provider:
          name: Accuknox Inc.
          url: https://accuknox.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        - public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:b6fae25e229fb7fca757872fc0460642d52c4f9906f776afac50c54a69b05b68
        - public.ecr.aws/k9v9d5v2/feeder-service@sha256:f5bc340d9d3e34804dd54a03701d272e0ce5c04160d1aa41984cb3d877eb5b76
        - public.ecr.aws/k9v9d5v2/discovery-engine-discover@sha256:052efbe1064503853adab497415b9407f51578e1456931a8a0e60353aa53e0c2
        - public.ecr.aws/k9v9d5v2/discovery-engine-hardening@sha256:3eb28d44ef93e6ba331e9085e64475458e1c234017c77dc30b8bc802def434c3
        - public.ecr.aws/k9v9d5v2/agents-operator@sha256:656d6bac4371adef3e86658011ba68ae2f98514f6ddaea0667d4269eb88dd02e
        - public.ecr.aws/k9v9d5v2/shared-informer-agent@sha256:6d18d13312d4309b6e32b18c063d5b26f15fbf057fda2b9e6b7e8ebb1a5885e9
        - public.ecr.aws/k9v9d5v2/policy-enforcement-agent@sha256:f004fc3572a4d66da1bc0db9e05a279d761a42d6ed62f95de7a2bb42ce71e17b
        - public.ecr.aws/k9v9d5v2/discovery-engine-sumengine@sha256:4ca78f90b28b38badb0520f148ce702a3a6cd6b02bb62e8b1a6782a5b82e0868
        - public.ecr.aws/k9v9d5v2/discovery-engine-offloader@sha256:3c2291310e30e35a931d2af389f7ade656ce86f609d2e42ecdabf1000764695c
        version: 0.2.1
      entries:
      - name: accuknox-operator.v0.2.1
        version: 0.2.1
      name: alpha
    - currentCSV: accuknox-operator.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.accuknox.com/v1",
                "kind": "AccuknoxConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "accuknox-operator",
                    "app.kubernetes.io/instance": "accuknoxconfig",
                    "app.kubernetes.io/name": "accuknoxconfig-sample",
                    "app.kubernetes.io/part-of": "accuknox-operator"
                  },
                  "name": "accuknoxconfig-sample",
                  "namespace": "accuknox-agents"
                },
                "spec": {
                  "ppsHost": "PLEASE_FILL_IN",
                  "knoxGateway": "PLEASE_FILL_IN",
                  "spireHost": "PLEASE_FILL_IN"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Security
          containerImage: public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:a935ad806e82760bb838bada30d3ed1a54c8c61b95fe5c6109f48985c6faf4d7
          createdAt: "2024-08-23T04:25:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: accuknox-agents
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: docker.io/accuknox/accuknox-operator
          support: Accuknox Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccuknoxConfig is the Schema for the accuknoxconfigs API
            kind: AccuknoxConfig
            name: accuknoxconfigs.operator.accuknox.com
            version: v1
        description: |2-

          The AccuKnox Operator brings native K8s support for deploying and managing AccuKnox Components on a Kubernetes cluster.
          The AccuKnox Operator helps to easily onboard a k8s cluster to the AccuKnox SaaS.

          AccuKnox Operator deploys and manages the components:

          1. Discovery Engine

          2. Policy Enforcement Agent

          3. Shared Informer Agent

          4. Feeder Service

          5. Agents Operator

          ## Prerequisites for using this Operator

          1. Get AccuKnox licence - https://app.accuknox.com/sign-up

          2. After Successful Sign-up into AccuKnox Application.

             Please naviagate to Settings >> Manager Cluster >> Onboard Cluster >>

             After successful onboarding, Copy the join token created by AccuKnox

          3. Create a secret using jointoken by executing the below command:

          ```
          oc create secret generic accuknox-jointoken --from-literal=token=<joinToken> -n accuknox-agents
          ```

          ## Deploy AccuKnox Operator using AccuknoxConfig Custom Resource

          All the information to configure the AccuKnox Operator will be provided via Accuknox during Cluster Onboarding steps.

          ### AccuknoxConfig Spec

          | Spec | Description |
          | ---- | ----------- |
          | spireHost | "spire host url" |
          | ppsHost   |  "pps host url" |
          | knoxGateway | "knox gateway url" |
          | joinTokenSecretName | "join token secret name if it's different than accuknox-jointoken" |
          | joinTokenSecretKey | "key name used to store the jointoken with the secret" |

          *Example*

          ```
          apiVersion: operator.accuknox.com/v1
          kind: AccuknoxConfig
          metadata:
            name: accuknoxconfig-default
            namespace: accuknox-agents
          spec:
            ppsHost: <ppsHost url>
            knoxGateway: <knox gateway url>
            spireHost: <spire host url>
          ```
        displayName: Accuknox Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CNAPP
        - CSPM
        - CWPP
        - Cloud Native
        - Runtime Security
        - Zero-Trust
        maturity: stable
        provider:
          name: Accuknox Inc.
          url: https://accuknox.com
        relatedImages:
        - docker.io/accuknox/discovery-engine-discover@sha256:e2d37bfbd0e9d23c129c143eb7c9aab1c3ed55e9571b44888eea2b842f8a2361
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        - public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:a935ad806e82760bb838bada30d3ed1a54c8c61b95fe5c6109f48985c6faf4d7
        - docker.io/accuknox/discovery-engine-hardening@sha256:e596aaaabaec94e0abb4f92ad9266dc010c0e88a89d600e8750a360099cae3cc
        - docker.io/accuknox/feeder-service@sha256:295e897b6f27ce61c45fb0179b734ef6d4c9e91864bcc5d2884a2d1dd4db81b6
        - docker.io/accuknox/policy-enforcement-agent@sha256:a0364343afeac4537589bfc247f64fffd7aac16f8afe07ce050bc58b2af248fb
        - docker.io/accuknox/discovery-engine-offloader@sha256:59cc594009de89a574cd569e701b6765d59a6899a03ad07eee1c025331df4331
        - docker.io/accuknox/shared-informer-agent@sha256:9f53d7dba9569487c42c6cf9e88988a1adb4c0971742e8b521700dce2d22b8cc
        - docker.io/accuknox/discovery-engine-sumengine@sha256:f7927128b86fdfb635377bd599df06f0eb9b9f0d1b068a0047468e480855443e
        - docker.io/accuknox/agents-operator@sha256:b0ae2728f15ebae904efb778cd29215ccf07ffca945a59deb5484f6c19c1b21f
        version: 0.7.1
      entries:
      - name: accuknox-operator.v0.7.1
        version: 0.7.1
      name: stable
    defaultChannel: stable
    packageName: accuknox-operator-certified
    provider:
      name: Accuknox Inc.
      url: https://accuknox.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pachyderm, Inc.
      provider-url: https://pachyderm.com
    name: pachyderm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pachyderm-operator.v0.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "Pachyderm",
                "metadata": {
                  "name": "pachyderm-sample"
                },
                "spec": {
                  "console": {
                    "disable": true
                  },
                  "pachd": {
                    "metrics": {
                      "disable": false
                    },
                    "storage": {
                      "amazon": {
                        "credentialSecretName": "pachyderm-aws-secret"
                      },
                      "backend": "AMAZON"
                    }
                  },
                  "version": "v2.1.6"
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermExport",
                "metadata": {
                  "name": "pachydermexport-sample"
                },
                "spec": {
                  "backup": {
                    "target": "pachyderm-sample"
                  }
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermImport",
                "metadata": {
                  "name": "pachydermimport-sample"
                },
                "spec": {
                  "backup": "backups/pachyderm-backup-20220610.tar.gz",
                  "destination": {
                    "name": "pachyderm-restore",
                    "namespace": "pachyderm-workloads"
                  },
                  "storageSecret": "s3-credentials-secret"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
          createdAt: "2021-05-20T08:05:00Z"
          description: Pachyderm provides the data layer that allows data science
            teams to productionize and scale their machine learning lifecycle with
            data driven automation, petabyte scalability and end-to-end reproducibility.
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pachyderm/openshift-operator
          support: Pachyderm, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PachydermExport is the Schema for the pachydermexports API
            displayName: Pachyderm Export
            kind: PachydermExport
            name: pachydermexports.aiml.pachyderm.com
            version: v1beta1
          - description: PachydermImport is the Schema for the pachydermimports API
            displayName: Pachyderm Import
            kind: PachydermImport
            name: pachydermimports.aiml.pachyderm.com
            version: v1beta1
          - description: Pachyderm is the Schema for the pachyderms API
            displayName: Pachyderm
            kind: Pachyderm
            name: pachyderms.aiml.pachyderm.com
            version: v1beta1
        description: "The Pachyderm operator aims to make deploying and managing Pachyderm
          instances on Openshift container platform easier.\n\nPachyderm is the data
          foundation for machine learning.  Pachyderm provides industry leading data
          versioning, pipelines and lineage that allow data science teams to automate
          the machine learning lifecycle and optimize their machine learning operations
          (MLOps). \n\n## Getting started\n\n1. Pachyderm uses an object store to
          store your data. Set the pachd storage backend at `pachyderm.spec.pachd.storage.backend`
          to `AMAZON`, `MICROSOFT`, `GOOGLE` or `MINIO` depending on your storage
          provider. Additionally, create a user with read/write access to the object
          store.\n\n2. Setup the user credentials in the section, `pachyderm.spec.pachd.storage`.
          Required fields depend on the backend type set in step 1 above.\n\n3. Finally,
          deploy the pachyderm cluster in its own namespace.\n\nFor more information
          on getting started, reference the documentation.\n"
        displayName: Pachyderm
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pachyderm
        - versioning
        - machine learning
        - artificial intelligence
        links:
        - name: Documentation
          url: https://docs.pachyderm.com/latest/
        maintainers:
        - email: kyle@pachyderm.io
          name: Kyle White
        - email: eochieng@redhat.com
          name: Edmund Ochieng
        maturity: beta
        minKubeVersion: v1.19.0
        provider:
          name: Pachyderm, Inc.
          url: https://pachyderm.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - registry.connect.redhat.com/pachyderm/backup-handler@sha256:e2ef6bd8191e8ad58c765c0db8f5bb21ea7123caed0c0f58fe681a024201306d
        - registry.connect.redhat.com/pachyderm/etcd@sha256:7f330b51874379b1d4b71d94e37e0ecfa52b86d628296136970688cf204f53b5
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/pgbouncer@sha256:c3b6949227df24abb8934d111a34b44e87fb16194ac34a5d0fe284cff02e9ac0
        - registry.connect.redhat.com/pachyderm/init-utils@sha256:5c0258bfa0203079925648c0a7dda6d05a29b6f461854821a1eec541777ff359
        - registry.connect.redhat.com/pachyderm/worker@sha256:374f83d92899707bc87b0318fda6bffa78d2e9e523b827bb3d40693a2d1bdecd
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
        - registry.redhat.io/rhel8/postgresql-13@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/pachd@sha256:116477fefdf413ea02c6ac3d013f1f91d3f6e4b16aed2408c9890832cdb41430
        version: 0.1.4
      entries:
      - name: pachyderm-operator.v0.1.4
        version: 0.1.4
      name: stable
    defaultChannel: stable
    packageName: pachyderm-operator
    provider:
      name: Pachyderm, Inc.
      url: https://pachyderm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-listener-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: listener-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/tools@sha256:47204b2697dfbab5eaf9d6dcb4c160dd427a148d6b840727239b773810e74fc9
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/tools@sha256:47204b2697dfbab5eaf9d6dcb4c160dd427a148d6b840727239b773810e74fc9
        - quay.io/stackable/listener-operator@sha256:b9c7957e7e562889bddd08e679abff05bb677672b3f735932f68514b77eacfd5
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:a6d6631d35c32b874c995b6af889636ac8dd1f39dcb68810d9f9b10a4e2c1987
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:161e754437428c68490333337303fb30b695a9ac677c3a00a1dc27565b855ab8
        version: 24.11.1
      entries:
      - name: listener-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: listener-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/listener-operator@sha256:da53970954b889bf4d874ae3df7e411b9abf41950ef7d50d6289831826f045a0
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/listener-operator@sha256:da53970954b889bf4d874ae3df7e411b9abf41950ef7d50d6289831826f045a0
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:71f8465e35a67f4b96e4c4904e67ccac9bbb9ae347426afb8bbfaabd6709e983
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        version: 25.3.0
      entries:
      - name: listener-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: listener-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
        version: 25.7.0
      entries:
      - name: listener-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: listener-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        version: 25.7.0
      entries:
      - name: listener-operator.v25.7.0
        version: 25.7.0
      - name: listener-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: "25.7"
    packageName: stackable-listener-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      pod-security.kubernetes.io/audit: privileged
      pod-security.kubernetes.io/enforce: privileged
      pod-security.kubernetes.io/warn: privileged
      provider: NVIDIA Corporation
      provider-url: ""
    name: gpu-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gpu-operator-certified.v25.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "kernelModuleType": "auto",
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": []
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
          createdAt: Thu Sep 18 18:11:36 PDT 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <25.3.4'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:43aa99352e1dc47e9218f36340ae6328a6282ed1e7e9a61b4204475e980230ef
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:c22dc67465c33d2f17932931229e61a6e7d9e32c5ed2ecefe8d10e873ffedfeb
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:51c8f71d3b3c08ae4eb4853697e3f8e6f11e435e666e08210178e6a1faf8028f
        - nvcr.io/nvidia/driver@sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735
        - nvcr.io/nvidia/driver@sha256:44422f10e899d3419fbd611a02fc1c6127e071a2067bd36e3debb043bf251de0
        - nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:66c3d68ca2b35d47b5562300a10ff8711b953ba3a15df7dbc409d81613ebe5d0
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b689dfb7806ae51b0072620b47c26114f05477e8ad9725892ee10366e5d93d8d
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:119de9a331a47203858b99901f44d0c4a8052961b4e60327f4b100d0ab8c9df0
        - nvcr.io/nvidia/k8s-device-plugin@sha256:3c54348fe5a57e5700e7d8068e7531d2ef2d5f3ccb70c8f6bac0953432527abd
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:5c4e61f7ba83d7a64ff2523d447c209ce5bde1ddc79acaf1f32f19620b4912d6
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:6bb786c41fa7b1a6238c023566fdeaeeea3ba842a8daec015a7bd85a8037b249
        - nvcr.io/nvidia/cuda@sha256:d19fe621624c4eb6ac931b8558daa3ecc0c3f07f1e2a52e0267e083d22dceade
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3191cc0684e92d655e72fe9faf3da95fe6b8a6c2eb4ad396f75ba9b98c021aaf
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:619e7bf48168d76a7e087c3bf6190fda9b10449c9839309fa4f9ed5b8a9e8804
        version: 25.3.4
      entries:
      - name: gpu-operator-certified.v25.3.4
        version: 25.3.4
      - name: gpu-operator-certified.v25.3.3
        version: 25.3.3
      - name: gpu-operator-certified.v25.3.2
        version: 25.3.2
      - name: gpu-operator-certified.v25.3.1
        version: 25.3.1
      - name: gpu-operator-certified.v25.3.0
        version: 25.3.0
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: stable
    - currentCSV: gpu-operator-certified.v1.10.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "dcgmExporter": {
                    "config": {
                      "name": ""
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                  },
                  "devicePlugin": {
                  },
                  "driver": {
                    "enabled": true,
                    "use_ocp_driver_toolkit": true,
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "gfd": {
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "operator": {
                    "defaultRuntime": "crio",
                    "deployGFD": true,
                    "initContainer": {
                    }
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:c7f9074c1a7f58947c807f23f2eece3a8b04e11175127919156f8e864821d45a
          createdAt: Mon Apr 18 10:06:40 PDT 2022
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=v1.9.0 <v1.10.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cuda@sha256:7f0f7dc8c993a86865d4d9f9b0e36d113ea15603d5d5279d2879fa2bc21c8236
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:5b16056257acc51b517d9cdb1da3218693cefc214af93789e6e214fd2b4cacf1
        - nvcr.io/nvidia/k8s-device-plugin@sha256:aa95c16106280e36a5e32d2fe4c66e8b70f5a114860c6f4ed5b1a4085c63601b
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:e8f06ce53415e8572f633b2169289d973993cb1e49d4dcc686ac134194d88f33
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:24d804e8f005d7aeca8343aa13e5f92295d8642a4c47cb24e3ac86a22543bc37
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:67c48a8370c34f68b46eb57339e87d01c04198368a0423553fab7565c862c4df
        - nvcr.io/nvidia/driver@sha256:f99b66855e92451a59a722355dd52bf97eb985cc8a647172273de8178aeaa93c
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:c2e3d0fe41a0d227dbb70caec03c780cc76317515e5ab3875f31d50c63f41c66
        - nvcr.io/nvidia/gpu-operator@sha256:c7f9074c1a7f58947c807f23f2eece3a8b04e11175127919156f8e864821d45a
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:ac13256c295126b854a07421c45de0721e7a57974b38b3e66eb1fdc3a4e29f2e
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:9abb063e6b74085728fead5ed5bbe3bd7b3f57de60e87f6b964bba2d46f1bc67
        version: 1.10.1
      entries:
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v1.10
    - currentCSV: gpu-operator-certified.v1.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    },
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "dcgmExporter": {
                    "config": {
                      "name": ""
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                  },
                  "devicePlugin": {
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": true
                  },
                  "vgpuDeviceManager": {
                    "enabled": true,
                    "config": {
                      "name": "vgpu-devices-config",
                      "default": "default"
                      }
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:53180f81fb7e1e4bcd49a556bc6285c2262abbab8083b29d2688b02d4200dd2d
          createdAt: Thu Aug  4 11:57:56 PDT 2022
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=v1.9.0 <v1.11.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:6e2902d73c54cf57bec94edf166d206cfeb0b948b5aab70800ccbe025064b3a1
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:b90a519e9f30fbc013e1806bc081268724e9e80b4035c67c167ab3b43a37704a
        - nvcr.io/nvidia/gpu-operator@sha256:53180f81fb7e1e4bcd49a556bc6285c2262abbab8083b29d2688b02d4200dd2d
        - nvcr.io/nvidia/driver@sha256:cf8ce04907d0e8decf90a8dccfa6134af9c39509d2f8a21c355ed71223acacdb
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:39127d04c9faff3f22068dcf463ed48d02be684cd18e3f14db843bddbdd783a7
        - nvcr.io/nvidia/driver@sha256:b8799e7397fae5a27446737443091eed05b918816f53c400a48be71125ee22e8
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3804e2b03c44d25dd410e3882b7fb6d1fde41f55110d7702af47136b042a63ff
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:10d825ccbfcf87f8c74f841fa89ae345f2d381fb47a5b2b31f61b72ebbbb0915
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:f48e8b430248a9cadf716a9211bb0050d4a663d1f1386ea107e72917f82e5b5f
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:8490d555131a57a11311561667ea9ec8cf554a399d2492a6f9e6677c01ec4b9a
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7a0ccbcbe8d379bfc441e06504eb45d0d433a37e2a9e6dde83d13347296ebe81
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:e88e610d09f1a7528ab32cb54259aee340eba2b2bf66b92a28157626daadc82d
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:2d6e86db0e2b0db64c00566938e1851e25d4bb1019b1e7b696e2b199cba33397
        - nvcr.io/nvidia/driver@sha256:8c9fabc1cd773e9526eb7e8526084f065f6cca9d0fd8dc2a672a0bb717f1cc60
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:16ee5b1a8d9217f761c40a06c7ae598cbc893fbc35d99ff9b84fa886ea6b9b37
        - nvcr.io/nvidia/cuda@sha256:4400b56f81095b24fc9dde93ad7ad7d0ca6a7e9736c4c7f1d4e2bcbf0afbe45c
        version: 1.11.1
      entries:
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v1.11
    - currentCSV: gpu-operator-certified.v22.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:54d5846c0291db621931a8544c9cb6437512470a93278ca9130ba73f2e2f0055
          createdAt: Mon Jan 30 15:43:24 PST 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <22.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:c3cf59dd5d6160eba5d816ade2e81b35ebb10f4884df67971f6ace36f8e6efc1
        - nvcr.io/nvidia/driver@sha256:f250d1e4989fd4821e81bad9227eeeaa531ca8810f3a40e996f1ca2af7945f96
        - nvcr.io/nvidia/k8s-device-plugin@sha256:9c17d3a907eb77eb8f7b4f3faf52d8352e4252af92003f828083f80d629bd2c3
        - nvcr.io/nvidia/driver@sha256:b860ef04cc6cd8c481810fa19bef7221071f5a3862640b2ad6c2704728dc77c8
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:0d47dad29d2ef445b301c5c64717758eed43a606345b79f97bce2e64b40a91a8
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:64d757b4c80b910e64647a84a1d592fab2ea3313ff6dce30c25c3a08e180bd74
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:efb88937f73434994d1bbadc87b492a1df047aa9f8d6e9f5ec3b09536e6e7691
        - nvcr.io/nvidia/driver@sha256:f2ca4612f3418e0e3963be8c3771ffe463cdc1fc1b810437a3aad9d76bb6499d
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:bec9f026d9b3d9404c78d6091817a359015c6a7aa411735b34138c1518853b5d
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:00f1476548fbed9ee01961443a73bf65396c2e8bb2b84426f949dd56cb4d14cd
        - nvcr.io/nvidia/gpu-operator@sha256:54d5846c0291db621931a8544c9cb6437512470a93278ca9130ba73f2e2f0055
        - nvcr.io/nvidia/driver@sha256:c5a62bc2a6cce339798e4b7cf9cec47e289c2d30b40f451ba8c9e958a68e3e55
        - nvcr.io/nvidia/driver@sha256:268c26781c46b36fe691638e807eb5c3a2cd077029ddb095ac75264adac04925
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:9177a0ae30798b42d0387f6a20cd3ce3cd1799a91b7866bf812368764b05b1af
        - nvcr.io/nvidia/cuda@sha256:5f2a2d8977f2c59abe88394f05cc3c044194554f90994d8554b0e1117ab5900d
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:9a00cdfdddb73327ef8e8e0fa60e50926a388d380c551bba6a3d3012be40401d
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:50c2a2b998e467c82716efa79fe136aa6f7ea95fd23576cf384d251bb9628640
        version: 22.9.2
      entries:
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v22.9
    - currentCSV: gpu-operator-certified.v23.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:08ca33f491195b7ba867be1e6642b5ff8aff6a4b0b7cfc0ec87619252ccfddc7
          createdAt: Mon May 22 15:22:15 PDT 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <23.3.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:8ff0ee88fe6c4ff031ef642593f93625bfd9a48fe7be8a79a5c2736117e36941
        - nvcr.io/nvidia/k8s-device-plugin@sha256:80871275fbcf888d4683c58638c7a75b3c46091040cafc6210a8c93513adb146
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:03e43d6b1460cadb33509d5ee07a5e27603ab89d7ec35f67d8a51bb78745b463
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:cab21c93987a5c884075efe0fb4a8abaa1997e1696cbc773ba69889f42f8329b
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:84ce86490d0d313ed6517f2ac3a271e1179d7478d86c772da3846727d7feddc3
        - nvcr.io/nvidia/cuda@sha256:62d845b4bd6de4be65f0b6daaae6831c29d4529955f99db5d26f5f7179106285
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:5489125ceae5864280e4d6a9ab52ab0f650b3179349a7298c4a204feb60b661a
        - nvcr.io/nvidia/gpu-operator@sha256:08ca33f491195b7ba867be1e6642b5ff8aff6a4b0b7cfc0ec87619252ccfddc7
        - nvcr.io/nvidia/driver@sha256:73b8928a1b1fb650d2ca8d6451a7d7c551fd0509f7511c2fd5bfdcef83ae4f5d
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:2d7e32e3d30c2415b4eb0b48ff4ce5a4ccabaf69ede0486305ed51d26cab7713
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:21dfc9c56b5f8bce73e60361d6e83759c3fa14dc6afc2d5ebdf1b891a936daf6
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:0d47dad29d2ef445b301c5c64717758eed43a606345b79f97bce2e64b40a91a8
        - nvcr.io/nvidia/driver@sha256:917f92c589c3124dba54f66f89c44b93b0beaf418f01d0da6a546e1670fdf3a9
        - nvcr.io/nvidia/driver@sha256:3df118f2330e95db3c76173cca35e542b2da120ec0bb8f6aff83c4c96c0c2daf
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:8f03f74076b5aa6672a5d75a398fa7515fcde57fdb1cf8079e015b3366efd3b5
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:f10170b076badda8f669af8b0f7b47eb5470abaf35a67dda9c4e182d66138177
        version: 23.3.2
      entries:
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.3
    - currentCSV: gpu-operator-certified.v23.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:d2e0f8bd86e8e9f2ff3514e113657658584ca4a2d0b82dac2e47364cc0b30648
          createdAt: Thu Aug 31 12:22:44 PDT 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <23.6.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:878a6162a4fa0b61dae02c7cfc13ba4d28bb06f92fd62d0c521633ab507d675a
        - nvcr.io/nvidia/driver@sha256:bbdc4415fe2e71eeea1b8a6e50ee562e11704dabb46db1db4aea807102886a35
        - nvcr.io/nvidia/k8s-device-plugin@sha256:f1617c3fe54fd9f6d4a00358e5da1a9fa9e8963a4ef6b3ed06f47d4971ddff4b
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:b30f5c64791dc5406b57eaf51ac40195dc1051a18cd1f38fd0a43c65f1d6dba9
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:a229beba8540549c234a046e975bf4bf8c50d521dfe551af56f6e042b25ec017
        - nvcr.io/nvidia/gpu-operator@sha256:d2e0f8bd86e8e9f2ff3514e113657658584ca4a2d0b82dac2e47364cc0b30648
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:9484110986c80ab83bc404066ca4b7be115124ec04ca16bce775403e92bfd890
        - nvcr.io/nvidia/driver@sha256:b4517810f2884f345b4f2323e83e41d28f54cdc6e2caf0690bb7ed4f3947dbcc
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:bb845160b32fd12eb3fae3e830d2e6a7780bc7405e0d8c5b816242d48be9daa8
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:60086c33226bbe7f04b611c20d308fed5825226f8a0a5635bc58d8ed55cc08a5
        - nvcr.io/nvidia/cuda@sha256:64c0ff5fd4a27d4b1d46378634f7a7398c75f87c18c70c988dde18910d461560
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:00a91252be8e62da6f28cdf1ec3721f2e945ae296938b0b925ef9db2bf6dcc8b
        - nvcr.io/nvidia/driver@sha256:3382e254056f28831767bc6729bc2594353a5ff2a28fe9f2d94396c597bb23d8
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:0379c59b43e11861837787a0da22e70b528718373bb10de9e8b7752849bd4e7a
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:bf06c8ad380a36e5014607bd3ab05b80e8e5e7e77d23d0dd89fe53a5564152fa
        version: 23.6.1
      entries:
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.6
    - currentCSV: gpu-operator-certified.v23.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:7481a3697783dcdca9ae78e7b548a6900e86ea33ab49ec14f0ba55db2fdb1a2e",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:3988b3b570cf65b0d338cb29bd24b0b9bf1885e1563fbb398ef23cba0b59e6bb
          createdAt: Wed Mar 6 16:43:36 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <23.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:6a05d6a1923fda756aed0dddf7ed23a83c30cf1e6c519fc39dd70c0309ec8257
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:a67d8e92861a2dce5649105c07561e4422e9fe4ba81a6525dc0d70a7ef85f9c0
        - nvcr.io/nvidia/driver@sha256:56c79482582cdfbc58d3134e8672637c5bf05f328880f76898f526143d04c6af
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:8d1ec78f2b1ddb7f0c47453d0427231190747bda411733a7dd0c8f5196f09e9c
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:27c44f4720a4abf780217bd5e7903e4a008ebdbcf71238c4f106a0c22654776c
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:925f4a47710e4318ed457930f5406174c1f6d28b1bf6b1bc310687fec0fde712
        - nvcr.io/nvidia/driver@sha256:abda1ac56371d55917b96ff330109980f468e133c9d5705da0ef87429f14ccd7
        - nvcr.io/nvidia/k8s-device-plugin@sha256:bb81ae8d034d1e1c6a342091a5f1dda5884dd43ab47e96624487a133d1edd4b3
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:387021553dfb16aab633228d42f63f04fa932b4f46add07527f296dfe97e5148
        - nvcr.io/nvidia/driver@sha256:7481a3697783dcdca9ae78e7b548a6900e86ea33ab49ec14f0ba55db2fdb1a2e
        - nvcr.io/nvidia/gpu-operator@sha256:3988b3b570cf65b0d338cb29bd24b0b9bf1885e1563fbb398ef23cba0b59e6bb
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:011fb450af3fa2e8fe5d28d590e4c653631447bc23d149591ced3d89089c4f2c
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:9aefef081c3ab1123556374d2b15d0429f3990af2fbaccc3c9827801e1042703
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:59a3875e7a37eb370385e654184efa3a1b193c9ea352165818496b19cbe14aa4
        - nvcr.io/nvidia/cuda@sha256:714547d54e5fe4191019a1e5f1daffc7fab7481b619b79c378541dafc76c9e5d
        version: 23.9.2
      entries:
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.9
    - currentCSV: gpu-operator-certified.v24.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:96f25c67e5b1072d5981080e12d65ec37eb9ef2fc0494499416aa801b0a34da3",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:041e75a3df84039c2dbbd4b9d67763bd212138822dbb6dbc0008858c1c6eff8d
          createdAt: Tue Apr 30 14:18:26 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.3.0'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:ecb867c5787e44f1889b257e6c03b07748689a7b293c8f1affb1d4cb449b235c
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:f0e4e14a3081417be8e6a5d855fb0cb69e1b63bc54a74f17e6d7084abe275588
        - nvcr.io/nvidia/driver@sha256:56c79482582cdfbc58d3134e8672637c5bf05f328880f76898f526143d04c6af
        - nvcr.io/nvidia/driver@sha256:a836ccbe21da735aee1c39b81060ed5e2fdb4ffa339874baaf4634f1e9259f74
        - nvcr.io/nvidia/driver@sha256:96f25c67e5b1072d5981080e12d65ec37eb9ef2fc0494499416aa801b0a34da3
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:2edc1d4ed555830e70010c82558936198f5faa86fc29ecf5698219145102cfcc
        - nvcr.io/nvidia/k8s-device-plugin@sha256:80674b19898ecf9ed6949e39674da769d6feeeb01bea54b914ef9ff502834f49
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:d8774d6afd37110199512636172823ae8749e5ea2e13760f57c255a74f47018c
        - nvcr.io/nvidia/cuda@sha256:ae0623ec8634b6c88f815b88037763def160cbbac15013b77ddef257fc276c9a
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:1ae0b923bc57f47bab046b50c50110f6914bbaffbfef704df34b3fe332db2e31
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:0c77725de8c42d248ed825453efd2e005f4900f4be384fd23084f6b721ddd0e0
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:7798eb9b9424e09959fc808596f87bdb68ac5740174123c6646dee2166d3e7d2
        - nvcr.io/nvidia/gpu-operator@sha256:041e75a3df84039c2dbbd4b9d67763bd212138822dbb6dbc0008858c1c6eff8d
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:2cc2119d20515f8ca7a61a0f3932578d69a45bcdea49e6c320a89c56f105e7d9
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:2442fd2ec0ee29746abbced727c53741f017e8f3f615321c1ec42c1c3ffa0b12
        version: 24.3.0
      entries:
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.3
    - currentCSV: gpu-operator-certified.v24.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:858de27c152669f5a3cf4287406405b16dd5bb70c0373324eb735511997bb415",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:8e0969cffc030a89c4acd68e64d41dd54e3bce8a794106b178d4dbd636a07f1c
          createdAt: Wed Sep 25 08:25:16 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.6.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:21f4c8b88716e8e6f732f9fb4c2efaef937c227491a8631c5e55036f80f39a4d
        - nvcr.io/nvidia/driver@sha256:858de27c152669f5a3cf4287406405b16dd5bb70c0373324eb735511997bb415
        - nvcr.io/nvidia/gpu-operator@sha256:8e0969cffc030a89c4acd68e64d41dd54e3bce8a794106b178d4dbd636a07f1c
        - nvcr.io/nvidia/cuda@sha256:de5b598bca89850567c4c104411d66bb52f47c9179199e6a3be6829b7ac586a2
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:740abc3ff657545c10effd5354f09af525200ed9a1b7623f0c2e8c7bd9e4a4e2
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:96380b95396b7f29d2ed2ec8ef1ad5a8bcf9f55051db723295dc0a20db845331
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:ae63fac9a4057a7646f0cf0ee0566e8928529adde05c4c0a017cda0599e381b2
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:33de74eb590f071403c17b6c210c02963245851971168bc0c07c06c100a9f376
        - nvcr.io/nvidia/driver@sha256:07e11f85d54d49ec9648fb06e148b8d832ee1f9c3549a915eee853c9ef2949c2
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:781fb47e264d9e0fbc8da5bd046e5e678316c866bc36ddd4b56d4eb0de682d5b
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:969147c01d63be5d1fe458f32f1cc0c7408cf3062531db91408e2fc57b4d8a67
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:15dab1273345df4a5844c4c761d064dbc4b592101251dc39174e597137123027
        - nvcr.io/nvidia/driver@sha256:a6d12fb5753f267dda25dfd38910f972bc632c006a24107fa50e20bba3642d7c
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7ad2c9f71fe06f9f7745ac8635f46740fbdff4f11edd468addfab81afcdfa534
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:7bcc188703f2fac630f7a4ff8960e6733ac3a29adf6a946533b796d9a27b8acf
        version: 24.6.2
      entries:
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.6
    - currentCSV: gpu-operator-certified.v24.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:6fe74322562c726c8fade184d8c45ebae3da7b1ea0a21f0ff9dc42c66c65e692",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:92f5262bb4934e0ea29eb10ec0ad4fc293164f2f2f94d48a935922791abbfdac
          createdAt: Mon Jan 27 16:10:52 PST 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cuda@sha256:b3b340847db9abd0c39703278f5638d3e65e1369e9a526a19b3138d5a6563527
        - nvcr.io/nvidia/gpu-operator@sha256:92f5262bb4934e0ea29eb10ec0ad4fc293164f2f2f94d48a935922791abbfdac
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:34eab99d7992a57c35803ec5e0afd538d16e997341e951fef4c4019afb08793e
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:e8398b2451996e5c64fc3855e7a3c93b667a85f69c0e259049d3b0012c21545d
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:d934033f9ff89087d555850eb1da150eaaa27b11e5ad21db1b874b8bc0b8d35b
        - nvcr.io/nvidia/driver@sha256:6fe74322562c726c8fade184d8c45ebae3da7b1ea0a21f0ff9dc42c66c65e692
        - nvcr.io/nvidia/driver@sha256:38b66a8d44cab9e2c62da9e101f32cd9dbcb5e02d8e57b47671284d374ca3695
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:4ffa1cd2a6497eb647a89ed259dcfb007554737b9d80f69bc173a2c3cd72a1da
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:40830d3561c14743f484d45b498141f9e86b1308e16fae3978110783927264ab
        - nvcr.io/nvidia/driver@sha256:026f8f2d29b7058ecaaa5b98666e28e1b3646b73e7d7cda3dc8026d47a929152
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:cf39d78ced7fb5727a9668ee2cd44b14bb7a23a95b83d5464b7d755740e02121
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7089559ce6153018806857f5049085bae15b3bf6f1c8bd19d8b12f707d087dea
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b072c5793be65eee556eaff1b9cbbd115a1ef29982be95b2959adfcb4bc72382
        - nvcr.io/nvidia/driver@sha256:d55b57938866e538acc3a71ca32f8cf87e71c591abd4a34695ee428e7ec2fa73
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:7edd7a0413dcb39b6e3bcefaf06812f3293c8e480ca10783e821a561ed686200
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:857f7669ae5ce2a43a3b5691fa970085098321ef8e90a896fa8dda3dee5b1d2b
        version: 24.9.2
      entries:
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.9
    - currentCSV: gpu-operator-certified.v25.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "kernelModuleType": "auto",
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": []
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
          createdAt: Thu Sep 18 18:11:36 PDT 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <25.3.4'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cuda@sha256:d19fe621624c4eb6ac931b8558daa3ecc0c3f07f1e2a52e0267e083d22dceade
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:619e7bf48168d76a7e087c3bf6190fda9b10449c9839309fa4f9ed5b8a9e8804
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:66c3d68ca2b35d47b5562300a10ff8711b953ba3a15df7dbc409d81613ebe5d0
        - nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:5c4e61f7ba83d7a64ff2523d447c209ce5bde1ddc79acaf1f32f19620b4912d6
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:51c8f71d3b3c08ae4eb4853697e3f8e6f11e435e666e08210178e6a1faf8028f
        - nvcr.io/nvidia/driver@sha256:44422f10e899d3419fbd611a02fc1c6127e071a2067bd36e3debb043bf251de0
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3191cc0684e92d655e72fe9faf3da95fe6b8a6c2eb4ad396f75ba9b98c021aaf
        - nvcr.io/nvidia/driver@sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735
        - nvcr.io/nvidia/k8s-device-plugin@sha256:3c54348fe5a57e5700e7d8068e7531d2ef2d5f3ccb70c8f6bac0953432527abd
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:6bb786c41fa7b1a6238c023566fdeaeeea3ba842a8daec015a7bd85a8037b249
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:c22dc67465c33d2f17932931229e61a6e7d9e32c5ed2ecefe8d10e873ffedfeb
        - nvcr.io/nvidia/driver@sha256:43aa99352e1dc47e9218f36340ae6328a6282ed1e7e9a61b4204475e980230ef
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:119de9a331a47203858b99901f44d0c4a8052961b4e60327f4b100d0ab8c9df0
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b689dfb7806ae51b0072620b47c26114f05477e8ad9725892ee10366e5d93d8d
        version: 25.3.4
      entries:
      - name: gpu-operator-certified.v25.3.4
        version: 25.3.4
      - name: gpu-operator-certified.v25.3.3
        version: 25.3.3
      - name: gpu-operator-certified.v25.3.2
        version: 25.3.2
      - name: gpu-operator-certified.v25.3.1
        version: 25.3.1
      - name: gpu-operator-certified.v25.3.0
        version: 25.3.0
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v25.3
    defaultChannel: v25.3
    packageName: gpu-operator-certified
    provider:
      name: NVIDIA Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Yugabyte Inc
      provider-url: ""
    name: yugabyte-platform-operator-bundle-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: yugabyte-platform-operator.v0.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "yugabyte.com/v1alpha1",
                "kind": "YBPlatform",
                "metadata": {
                  "name": "ybplatform-sample"
                },
                "spec": {
                  "domainName": "cluster.local",
                  "helm": {
                    "package": "/opt/yugabyte/helm/yugabyte-latest.tgz",
                    "timeout": 900
                  },
                  "helm2Legacy": false,
                  "image": {
                    "commonRegistry": "",
                    "nginx": {
                      "name": "nginxinc/nginx-unprivileged",
                      "registry": "",
                      "tag": "1.17.4-amd64"
                    },
                    "postgres": {
                      "name": "postgres",
                      "registry": "",
                      "tag": 11.5
                    },
                    "prometheus": {
                      "name": "prom/prometheus",
                      "registry": "",
                      "tag": "v2.2.1"
                    },
                    "pullPolicy": "IfNotPresent",
                    "pullSecret": "yugabyte-k8s-pull-secret",
                    "repository": "registry.connect.redhat.com/yugabytedb/yugabyte-platform",
                    "tag": "latest",
                    "thirdparty-deps": {
                      "name": "yugabyte/thirdparty-deps",
                      "registry": "quay.io",
                      "tag": "latest"
                    }
                  },
                  "ip_version_support": "v4_only",
                  "ocpCompatibility": {
                    "enabled": true
                  },
                  "rbac": {
                    "create": true
                  },
                  "securityContext": {
                    "enabled": false,
                    "fsGroup": 1001,
                    "runAsUser": 1001
                  },
                  "tls": {
                    "certificate": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tDQpNSUlHTHpDQ0JSZWdBd0lCQWdJUUQ2R1ZmL0ViUDhSMFQvcWxjRGZyM3pBTkJna3Foa2lHOXcwQkFRc0ZBREJlDQpNUXN3Q1FZRFZRUUdFd0pWVXpFVk1CTUdBMVVFQ2hNTVJHbG5hVU5sY25RZ1NXNWpNUmt3RndZRFZRUUxFeEIzDQpkM2N1WkdsbmFXTmxjblF1WTI5dE1SMHdHd1lEVlFRREV4UlNZWEJwWkZOVFRDQlNVMEVnUTBFZ01qQXhPREFlDQpGdzB4T1RBNE1ERXdNREF3TURCYUZ3MHlNVEEzTXpFeE1qQXdNREJhTUI4eEhUQWJCZ05WQkFNTUZDb3VZMnh2DQpkV1F1ZVhWbllXSjVkR1V1WTI5dE1JSUJJakFOQmdrcWhraUc5dzBCQVFFRkFBT0NBUThBTUlJQkNnS0NBUUVBDQp3SmswU25ub1pRYW1hWm0rMVVJdjdRYjBLalhZeVEwckpJSGpsbE03elhFYUFoL0JqVjZ1M0h1TWxrTlZSekVFDQpGeEtGUm5nSlpXTVlPcy9sWDU1QmtBS0RocXNmaDl0ZVA4ZFVQU1RTVTN6bWRVc0dTcDJIMGpKbVhuRWNvMFpzDQpwYlVyYWhlcWlxSUUyZ1gwQlpvYldvWGRHc1EydXJiUGxyMGlVSnUxVnp4SGRoTS9nTnFQWEN5bE5YdENrbVI2DQo0eXpIa0Rob21pRTAySTMwTXBlOERxczl4d3lKd3F6T3J2ZnA0eUE0Ukhna2g1clovK1Erc0VQNkIyVzBHQXc2DQpzZDNTYytOTytYVUZoVnlFRDhqZkNsTndPVHZodi9keE5IWHZ4WHd0RWNjZTd1TVBBTFJzSFh1RW1kTkN5N2pvDQpibzJuUWFObzZBQ1c0U2s5RWJVZXB3SURBUUFCbzRJREpqQ0NBeUl3SHdZRFZSMGpCQmd3Rm9BVVU4b1hXZnhyDQp3QU1oTHhxdTVLcW9ISUpXMm5Vd0hRWURWUjBPQkJZRUZHemw3bjdOMy9GbFdDSDFvdWFCWldKMjI1UStNQjhHDQpBMVVkRVFRWU1CYUNGQ291WTJ4dmRXUXVlWFZuWVdKNWRHVXVZMjl0TUE0R0ExVWREd0VCL3dRRUF3SUZvREFkDQpCZ05WSFNVRUZqQVVCZ2dyQmdFRkJRY0RBUVlJS3dZQkJRVUhBd0l3UGdZRFZSMGZCRGN3TlRBem9ER2dMNFl0DQphSFIwY0RvdkwyTmtjQzV5WVhCcFpITnpiQzVqYjIwdlVtRndhV1JUVTB4U1UwRkRRVEl3TVRndVkzSnNNRXdHDQpBMVVkSUFSRk1FTXdOd1lKWUlaSUFZYjliQUVDTUNvd0tBWUlLd1lCQlFVSEFnRVdIR2gwZEhCek9pOHZkM2QzDQpMbVJwWjJsalpYSjBMbU52YlM5RFVGTXdDQVlHWjRFTUFRSUJNSFVHQ0NzR0FRVUZCd0VCQkdrd1p6QW1CZ2dyDQpCZ0VGQlFjd0FZWWFhSFIwY0RvdkwzTjBZWFIxY3k1eVlYQnBaSE56YkM1amIyMHdQUVlJS3dZQkJRVUhNQUtHDQpNV2gwZEhBNkx5OWpZV05sY25SekxuSmhjR2xrYzNOc0xtTnZiUzlTWVhCcFpGTlRURkpUUVVOQk1qQXhPQzVqDQpjblF3Q1FZRFZSMFRCQUl3QURDQ0FYNEdDaXNHQVFRQjFua0NCQUlFZ2dGdUJJSUJhZ0ZvQUhVQTdrdTl0M1hPDQpZTHJoUW1rZnErR2VacU1QZmwrd2N0aURBTVI3aVhxby9jc0FBQUZzVHEwL013QUFCQU1BUmpCRUFpQlg3NXJ0DQpwNkdQbUQxbVl1S1hvcXVRVytIcUtRNE8rV1RXbTNCbVBteWF1d0lnQUoya01WdmduaUx1dEh5Y3AyYzBwdDRsDQptL3M5SWN2Yi9aN09jMU82b2xJQWRnQ0hkYi9uV1h6NGpFT1pYNzN6YnY5V2pVZFdOdjlLdFdEQnRPci9YcUNEDQpEd0FBQVd4T3JUOXNBQUFFQXdCSE1FVUNJSHphY3NCcmpEN1FER1ZRSHA5YUZPMnJ4VEx4NU9GOVE2MTFYaUU2DQpZUEk3QWlFQW9xd3pTd1BVaGt4bVhYNzN0SDVEcVNwaUxuOHdrMXFWeWhZOTVmSmxZa0VBZHdCRWxHVXVzTzdPDQpyOFJBQjlpby9pakEydWFDdnRqTE1iVS8wek9XdGJhQnFBQUFBV3hPclQ1M0FBQUVBd0JJTUVZQ0lRQ1ZLaHQ5DQpWTWhlNTUvZEwrSGN6Qzk5a2xheG9DcW5zOStRNGVOTDZ2Smd4QUloQUsyZmp4a1RadGw4WFZ4WHEwc2NSTWNJDQpiSktHcGJYdEV0N3dFVzRkZDNGS01BMEdDU3FHU0liM0RRRUJDd1VBQTRJQkFRQm1GcDNCK1g2Y21Cb0FLWkhrDQpIWWxzanhIZk01bDQyZWJtSFdGdlQyTi9aVkZ3MEZ3YndBNExyNzlvd01QdU5FbVNNa2pmVVhxR011Q2VlVTdODQo0MlRwUFdPaGNzVHF6ZkxVa0hhMU9rK011L3RUZHBNcmhROEVqbkhMY3FKVE9MbmRSVy8zcE5IM2xrK2FJejNoDQpDVGFueElRVEE2cUxRNThRakszWmFMVzdqWVVNcVJFcnBGVDlBVXkxMDAxMjlOam5HVkhtYmU1bW1ZbEwvQlhMDQpmdHBJUUJJZnhCNXpHbHpTcjRHT3p5RDBES05YU040RU0yM2pIM1ljeUg4eFQrODU4ZkxSaXQ4SFlOQ3RQbkloDQoyRzR6UjV3emZFVDlnZUh6aWtuY0ZMWUpMMEpKMVIrZGVNRE91bmJWOUJqcmZJb3NNSk53TUV2ejFZczlBWXhHDQpxSFNYDQotLS0tLUVORCBDRVJUSUZJQ0FURS0tLS0tDQotLS0tLUJFR0lOIENFUlRJRklDQVRFLS0tLS0KTUlJRXNUQ0NBNW1nQXdJQkFnSVFDS1dpUnMxTFhJeUQxd0swdTZ0VFNUQU5CZ2txaGtpRzl3MEJBUXNGQURCaApNUXN3Q1FZRFZRUUdFd0pWVXpFVk1CTUdBMVVFQ2hNTVJHbG5hVU5sY25RZ1NXNWpNUmt3RndZRFZRUUxFeEIzCmQzY3VaR2xuYVdObGNuUXVZMjl0TVNBd0hnWURWUVFERXhkRWFXZHBRMlZ5ZENCSGJHOWlZV3dnVW05dmRDQkQKUVRBZUZ3MHhOekV4TURZeE1qSXpNek5hRncweU56RXhNRFl4TWpJek16TmFNRjR4Q3pBSkJnTlZCQVlUQWxWVApNUlV3RXdZRFZRUUtFd3hFYVdkcFEyVnlkQ0JKYm1NeEdUQVhCZ05WQkFzVEVIZDNkeTVrYVdkcFkyVnlkQzVqCmIyMHhIVEFiQmdOVkJBTVRGRkpoY0dsa1UxTk1JRkpUUVNCRFFTQXlNREU0TUlJQklqQU5CZ2txaGtpRzl3MEIKQVFFRkFBT0NBUThBTUlJQkNnS0NBUUVBNVMyb2loRW85bm5wZXpvemlEdHg0V1dMTENsbC9lMHQxRVllbUU1bgorTWdQNXZpYUhMeStWcEhQK25kWDVEMThJTkl1dUFWOHdGcTI2S0Y1VTBXTklaaVFwNm1MdElXalVlV0RQQTI4Ck9leWhUbGo5VExrMmJleXRidEZVNnlwYnBXVWx0bXZZNVY4bmdzcEM3bkZSTkNqcGZuREVEMmtSeUp6Tzh5b0sKTUZ6NEo0SkU4TjdOQTF1SndVRUZNVXZITHMwc2NMb1Baa0tjZXdJUm0xUlYyQXhtRlF4SmtkZjdZTjlQY2traQpmMlhnbTNiNDhCWm4wemYwcVhzU2VHdTg0dWE5Z3d6anpJN3RiVEJqYXlUcFQrL1hwV3VCVnY2ZnZhckk2YmlrCktCODU5T1NHUXV3NzNYWGdldUZ3RVBIVElSb1V0a3p1My9FUStMdHd6bmtrZFFJREFRQUJvNElCWmpDQ0FXSXcKSFFZRFZSME9CQllFRkZQS0YxbjhhOEFESVM4YXJ1U3FxQnlDVnRwMU1COEdBMVVkSXdRWU1CYUFGQVBlVURWVwowVXk3WnZDajRoc2J3NWV5UGRGVk1BNEdBMVVkRHdFQi93UUVBd0lCaGpBZEJnTlZIU1VFRmpBVUJnZ3JCZ0VGCkJRY0RBUVlJS3dZQkJRVUhBd0l3RWdZRFZSMFRBUUgvQkFnd0JnRUIvd0lCQURBMEJnZ3JCZ0VGQlFjQkFRUW8KTUNZd0pBWUlLd1lCQlFVSE1BR0dHR2gwZEhBNkx5OXZZM053TG1ScFoybGpaWEowTG1OdmJUQkNCZ05WSFI4RQpPekE1TURlZ05hQXpoakZvZEhSd09pOHZZM0pzTXk1a2FXZHBZMlZ5ZEM1amIyMHZSR2xuYVVObGNuUkhiRzlpCllXeFNiMjkwUTBFdVkzSnNNR01HQTFVZElBUmNNRm93TndZSllJWklBWWI5YkFFQ01Db3dLQVlJS3dZQkJRVUgKQWdFV0hHaDBkSEJ6T2k4dmQzZDNMbVJwWjJsalpYSjBMbU52YlM5RFVGTXdDd1lKWUlaSUFZYjliQUVCTUFnRwpCbWVCREFFQ0FUQUlCZ1puZ1F3QkFnSXdEUVlKS29aSWh2Y05BUUVMQlFBRGdnRUJBSDRqeC9MS05XNVprbEZjCllXczhFamJtMG55ektlWkMyS09WWVI3UDhnZXZLeXNsV200WG80QlN6S3IyMzVGc0o0YUZ0NnlBaXYxZVkwdFoKL1pOMThiT0dTR1N0b0VjL0pFNG9jSXpyOFA1TWcxMWtSWUhibWdZbnIxUnhla2k1bVNlYjM5REd4VHBKRDRrRwpoczVsWE5vbzRjb25VaWlKd0thcUg3dmgyYmFyeWQ4cE1JU2FnODNKVXF5VkdjMnRXUHBPMDMyOS9DV3Eya3J5CnF2NjZPU01qd3VsVXowZFhmNE9IUWFzUjdDTmZJcis0S1NjYzZBQmxRNVJERjg2UEdlRTZrZHdTUWtGaUIvY1EKeXNOeXEwakVEUVRrZmEycGptdVd0TUNOYkJuaEZYQlllamZ1YkloYVViRXYyRk9RQjNkQ2F2K0ZQZzVlRXZlWApUVnlNbkdvPQotLS0tLUVORCBDRVJUSUZJQ0FURS0tLS0tCg==",
                    "enabled": false,
                    "hostname": "localhost",
                    "key": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQpNSUlFb3dJQkFBS0NBUUVBd0prMFNubm9aUWFtYVptKzFVSXY3UWIwS2pYWXlRMHJKSUhqbGxNN3pYRWFBaC9CCmpWNnUzSHVNbGtOVlJ6RUVGeEtGUm5nSlpXTVlPcy9sWDU1QmtBS0RocXNmaDl0ZVA4ZFVQU1RTVTN6bWRVc0cKU3AySDBqSm1YbkVjbzBac3BiVXJhaGVxaXFJRTJnWDBCWm9iV29YZEdzUTJ1cmJQbHIwaVVKdTFWenhIZGhNLwpnTnFQWEN5bE5YdENrbVI2NHl6SGtEaG9taUUwMkkzME1wZThEcXM5eHd5Sndxek9ydmZwNHlBNFJIZ2toNXJaCi8rUStzRVA2QjJXMEdBdzZzZDNTYytOTytYVUZoVnlFRDhqZkNsTndPVHZodi9keE5IWHZ4WHd0RWNjZTd1TVAKQUxSc0hYdUVtZE5DeTdqb2JvMm5RYU5vNkFDVzRTazlFYlVlcHdJREFRQUJBb0lCQUJDbDdaZzJlM2tvSHdCbgowdEFhWFl3dW1ySi9YeExaK1kyNVVKUHc4KzFwUm1kQ21YZC81THVZVG1tNnZrZGVIRTFtUTNENDlVNzNSdVBtCmpwdzZVTTZDK1IyTjh6YktKcFo2dzNSWmwzaXlCalhEVzNMZHRSemRVRzMvZHd0Q0RpZ3d4ZVR3ME1RVkJ6aFYKWldlS2o2eStmNE5OekxGTkZocWdsTFBGRWVMTlk3WUgvYUdHdTZwaVpERkFMT0liMlFVUUVvVVRkUEJabDMxSwpKL0Y0cHE1RkV6ekQ0aHMwYXJzSlIxeUd4L0NCRWxqQWRSNWE4Y3FHdGhxYjFMdXJTMVcxT0dNUGkzNFdDblJKCnYwTC9zSU9ORCt5cjdjd3ZQaFFzcmI3Tk9NbGxRSDYwUGdyQ0JyOUVwalFiRm1ZN3NCTS9jYUZTWHVDTmZnTzMKT3hTQjF0RUNnWUVBNHpSMTh2aXl3Ykt5aE5xeEhkNzJ1UkxHS2k4SWZhVmhsNGxFeGdsR0pLNHQ0KzYreXhObAp2OFBkdGg3elJjL1pyQUlBUEtBRnZxa1RnVzkrc2llL3JPcE9PUzA2UnVIWVpJMjdpS3ZiL054N29YTm9TM0Y2CitXRlNJSVJqRFo0eHc2SU1iNUJFS0hOVGtKU1pBMXBkaXhYeFgzeDVTM2hCWTlmekl2NGdVVDhDZ1lFQTJRSDAKbnU5REZ0MDNmWmpvcjRMUkZCQS84NEV1dmdBL2RxMkZ4TzVvMjIrdk43Qk9LRW9IWTRuTzhYdFN2L2MzaVFZdApmTUdGZUR2eXdqRnFPU05SS2JxenMza3ZmY2gxNEhHdFltT29DVTF6Smd2VkF4SytWOFF4YWxMbmhlbjZsaTgwCnJKYmtWWUJXYnFHZHU1d2FPM1RyeGsvbHBMejFvSnliK0FQbmNKa0NnWUVBeHd3anRTeEtOc1hkL2VRMXZReWoKcTdvVjM4ajZzRkRkOGNzeGRDbVZ5UGJMZjErZXlQeUNiSUVrMExQR3ZzZDhWR1FnWHZ4MmEvVm1BRVIwZzVHdApLOGpOcE92Sm5XV29sVmRTRDBjNDRBc3RTR0Y1NG5aVm5QcTAzNFg3M1dkVkE2WUpMUWYvbmt5Q3ppeFE3UTRTCjI1SHF0NU91RS84VGttd1ROdm5aeURNQ2dZQlUxV0xmendFWEt4dktTY0NNL0VMbjZRSlhHeVUrdWJJVm52Z3cKYUZNdGxaZVYyYUowTlp4MTJTWEJXVWljS3ZUVzZFZFlBZjgvWlpBZFFsd1lRaW8wZWZ4T0tRM2J5a1hxbkFYLwo5SjE0SllMaVdMWEpJY2F2WU05QWxYQlU5WWMwdHZlTWtPQnFYR1Z0a2xpclF2Y3VUeUxzdjhjeWczd3ljWWEzCmlpOHI0UUtCZ0NaODZtcEZuUlN2aTdpTWw1MUdybDFQZER2YzZVMURRVkI5empTQlVzak1lamNTQXk4WmVJZlQKUDVoSGc3REFaUEhmUGpYLzZxUnpNQW8zWlFlZVBMdTNEYVgwbFBoekNhalF1bkpIbWZvYlA2QjFOUG45T056bQo0dE0wdEU1YTJVVmJuZW5haVZyaDRhSXlJTTFxVTdHT1ArcDNMNi9hNzVnSVQ2SVJRVlloCi0tLS0tRU5EIFJTQSBQUklWQVRFIEtFWS0tLS0tCg==",
                    "sslProtocols": ""
                  },
                  "yugaware": {
                    "health": {
                      "email": "",
                      "password": "",
                      "username": ""
                    },
                    "multiTenant": false,
                    "replicas": 1,
                    "service": {
                      "enabled": true,
                      "ip": "",
                      "type": "LoadBalancer"
                    },
                    "serviceAccount": "yugaware",
                    "storage": "100Gi",
                    "storageClass": "standard"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/yugabytedb/yugabyte-platform-operator@sha256:f227c9e2c625b04160d952349ddba552ac66dae8e7da35f56746f3447c270ce3
          description: Operator for Yugabyte Platform makes it easy to deploy and
            upgrade Yugabyte Platform on OpenShift environments.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/yugabyte-platform-operator-bundle-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/yugabyte-platform-operator-bundle-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/yugabyte/yugabyte-platform-operator
          support: YugabyteDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy and manage Yugabyte Platform as a Kubernetes native
              custom resource.
            displayName: Yugabyte Platform
            kind: YBPlatform
            name: ybplatforms.yugabyte.com
            version: v1alpha1
        description: |2

          Operator for Yugabyte Platform makes it easy to deploy and upgrade Yugabyte Platform on OpenShift environments.
          Use Yugabyte Platform's orchestration and monitoring to manage YugabyteDB universes. Yugabyte Platform can create a YugabyteDB universe with many instances (VMs, pods, machines, etc., provided by IaaS), logically grouped together to form one logical distributed database. Each universe includes one or more clusters. A universe is comprised of one primary cluster and, optionally, one or more read replica clusters. All instances belonging to a cluster run on the same type of cloud provider instance type.

          ### Yugabyte Platform Features
          * **YugabyteDB-as-a-Service on Kubernetes** - Deploy your own YugabyteDB-as-a-Service running on any Kubernetes infrastructure, Yugabyte Platform provides central management for multiple database clusters and Includes 24/7 production support.
          * **Monitoring & Alerting** - Comprehensive monitoring & alerting without any configuration.
          * **Scheduled Backups** - Use Yugabyte Platform to perform a regular scheduled backup of a full YugabyteDB universe or selected tables.

          ### Getting Started
          Here are a few resources for getting started with Yugabyte platform
          * [Quick start guide for Yugabyte Platform](https://docs.yugabyte.com/latest/yugabyte-platform/deploy/) - Install Yugabyte platform on OpenShift.
          * [Quick Start for YugabyteDB](http://docs.yugabyte.com/latest/quick-start/) - install, create a local cluster and read/write from YugabyteDB.
          * [Explore core features](https://docs.yugabyte.com/latest/explore/) - automatic sharding & re-balancing, linear scalability, fault tolerance, tunable reads etc.
          * [Ecosystem integrations](https://docs.yugabyte.com/latest/develop/ecosystem-integrations/) - integrations with Apache Kafka/KSQL, Apache Spark, JanusGraph, KairosDB, Presto and more.
          * [Real world apps](https://docs.yugabyte.com/latest/develop/realworld-apps/) - sample real-world, end-to-end applications built using YugabyteDB.

          ### Help
          Cannot find what you are looking for? Have a question? We love to hear from you - [please file a GitHub issue](https://github.com/YugaByte/yugabyte-db/issues) or talk to us live at [Slack](https://www.yugabyte.com/slack).
        displayName: Yugabyte Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - distributed sql
        links:
        - name: Yugabyte Platform Operator
          url: https://www.yugabyte.com/
        maintainers:
        - email: oss-maintainers@yugabyte.com
          name: Yugabyte
        maturity: alpha
        provider:
          name: Yugabyte Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9211b70edbb93d718428e1c9a992e5fc74a20cb213f8a0592ffe340c362890fe
        - registry.marketplace.redhat.com/rhm/yugabytedb/yugabyte-platform-operator@sha256:f227c9e2c625b04160d952349ddba552ac66dae8e7da35f56746f3447c270ce3
        version: 0.0.6
      entries:
      - name: yugabyte-platform-operator.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: yugabyte-platform-operator-bundle-rhmp
    provider:
      name: Yugabyte Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: volsync-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: volsync-product.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
          createdAt: "2025-08-27T13:12:05Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.13.1'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
        version: 0.13.1
      entries:
      - name: volsync-product.v0.13.1
        version: 0.13.1
      - name: volsync-product.v0.13.0
        version: 0.13.0
      - name: volsync-product.v0.12.2
        version: 0.12.2
      - name: volsync-product.v0.12.1
        version: 0.12.1
      - name: volsync-product.v0.12.1-0.1749677898.p
        version: 0.12.1+0.1749677898.p
      - name: volsync-product.v0.12.0
        version: 0.12.0
      name: stable
    - currentCSV: volsync-product.v0.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:d345825e953c4c170db45a5d92639182b7ee18933287aaeea035f2a2bad13c37
          createdAt: 09 Jan 2025, 00:47
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.10.2'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:d345825e953c4c170db45a5d92639182b7ee18933287aaeea035f2a2bad13c37
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 0.10.2
      entries:
      - name: volsync-product.v0.10.2
        version: 0.10.2
      - name: volsync-product.v0.10.1
        version: 0.10.1
      - name: volsync-product.v0.10.1-0.1730236570.p
        version: 0.10.1+0.1730236570.p
      - name: volsync-product.v0.10.0
        version: 0.10.0
      name: stable-0.10
    - currentCSV: volsync-product.v0.11.2-0.1749674925.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:9da8572c79b75767952cf699e5b1699b9f46d83586aa2d6a17fb23c092713b62
          createdAt: 16 Apr 2025, 06:42
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.11.2'
          olm.substitutesFor: volsync-product.v0.11.2
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:9da8572c79b75767952cf699e5b1699b9f46d83586aa2d6a17fb23c092713b62
        version: 0.11.2+0.1749674925.p
      entries:
      - name: volsync-product.v0.11.2
        version: 0.11.2
      - name: volsync-product.v0.11.2-0.1749674925.p
        version: 0.11.2+0.1749674925.p
      - name: volsync-product.v0.11.1
        version: 0.11.1
      - name: volsync-product.v0.11.0
        version: 0.11.0
      name: stable-0.11
    - currentCSV: volsync-product.v0.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:7d1ac69fe3d0fddc66acdfce9361a4fc8dd3f549f66c98450f7a0e4c5cc893a0
          createdAt: "2025-07-07T13:46:18Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.12.2'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:7d1ac69fe3d0fddc66acdfce9361a4fc8dd3f549f66c98450f7a0e4c5cc893a0
        version: 0.12.2
      entries:
      - name: volsync-product.v0.12.2
        version: 0.12.2
      - name: volsync-product.v0.12.1
        version: 0.12.1
      - name: volsync-product.v0.12.1-0.1749677898.p
        version: 0.12.1+0.1749677898.p
      - name: volsync-product.v0.12.0
        version: 0.12.0
      name: stable-0.12
    - currentCSV: volsync-product.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
          createdAt: "2025-08-27T13:12:05Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.13.1'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
        version: 0.13.1
      entries:
      - name: volsync-product.v0.13.1
        version: 0.13.1
      - name: volsync-product.v0.13.0
        version: 0.13.0
      name: stable-0.13
    - currentCSV: volsync-product.v0.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:ef44675fe70ee9ec8109fb91c6559f0c5518b45219e19175a18a5004e40b65f7
          createdAt: 22 Aug 2024, 22:06
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.9.3'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ReplicationDestination defines the destination for a replicated
              volume
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: ReplicationSource defines the source for a replicated volume
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:ef44675fe70ee9ec8109fb91c6559f0c5518b45219e19175a18a5004e40b65f7
        version: 0.9.3
      entries:
      - name: volsync-product.v0.9.3
        version: 0.9.3
      - name: volsync-product.v0.9.2
        version: 0.9.2
      - name: volsync-product.v0.9.1
        version: 0.9.1
      - name: volsync-product.v0.9.0
        version: 0.9.0
      name: stable-0.9
    defaultChannel: stable
    packageName: volsync-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: deployment-validation-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: deployment-validation-operator.v0.7.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Application Runtime, Monitoring, Security
          certified: "false"
          containerImage: quay.io/deployment-validation-operator/dv-operator:0.7.12
          createdAt: "2025-09-11T00:00:00Z"
          description: The deployment validation operator
          ignore-check.kube-linter.io/minimum-three-replicas: This deployment uses
            1 pod as currently replicating does not replicate metric data causing
            installation issues
          repository: https://github.com/app-sre/deployment-validation-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Deployment Validation Operator (DVO) checks deployments and other resources against a curated collection of best practices.
          These best practices focus mainly on ensuring that the applications are fault-tolerant. DVO reports failed validations via Prometheus metrics.
          If the best-practice check has failed, the metrics will report `1`.
        displayName: Deployment Validation Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: false
          type: MultiNamespace
        keywords:
        - dvo
        links:
        - name: repository
          url: https://github.com/app-sre/deployment-validation-operator
        - name: containerImage
          url: https://quay.io/deployment-validation-operator/dv-operator:0.7.12
        maintainers:
        - email: dvo-owners@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/deployment-validation-operator/dv-operator:0.7.12
        version: 0.7.12
      entries:
      - name: deployment-validation-operator.v0.7.12
        version: 0.7.12
      - name: deployment-validation-operator.v0.7.9
        version: 0.7.9
      - name: deployment-validation-operator.v0.7.8
        version: 0.7.8
      - name: deployment-validation-operator.v0.7.7
        version: 0.7.7
      - name: deployment-validation-operator.v0.7.6
        version: 0.7.6
      - name: deployment-validation-operator.v0.7.5
        version: 0.7.5
      - name: deployment-validation-operator.v0.7.4
        version: 0.7.4
      - name: deployment-validation-operator.v0.7.3
        version: 0.7.3
      - name: deployment-validation-operator.v0.7.2
        version: 0.7.2
      - name: deployment-validation-operator.v0.7.1
        version: 0.7.1
      - name: deployment-validation-operator.v0.7.0
        version: 0.7.0
      - name: deployment-validation-operator.v0.6.0
        version: 0.6.0
      - name: deployment-validation-operator.v0.5.0
        version: 0.5.0
      - name: deployment-validation-operator.v0.4.0
        version: 0.4.0
      - name: deployment-validation-operator.v0.3.0
        version: 0.3.0
      - name: deployment-validation-operator.v0.2.2
        version: 0.2.2
      name: alpha
    defaultChannel: alpha
    packageName: deployment-validation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: ""
    name: gitops-primer
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitops-primer.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "primer.gitops.io/v1alpha1",
                "kind": "Export",
                "metadata": {
                  "name": "primer"
                },
                "spec": {
                  "method": "download"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/migtools/gitops-primer:v0.0.11
          operatorframework.io/suggested-namespace: gitops-primer-system
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Export is the Schema for the exports API
            displayName: Export
            kind: Export
            name: exports.primer.gitops.io
            version: v1alpha1
        description: Export and normalize Kubernetes YAML files for GitOps
        displayName: gitops-primer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - operator
        links:
        - name: Gitops Primer
          url: https://github.com/migtools/gitops-primer
        maturity: alpha
        provider:
          name: Konveyor
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6d57bfd91fac9b68eb72d27226bc297472ceb136c996628b845ecc54a48b31cb
        - quay.io/migtools/gitops-primer:v0.0.11
        - quay.io/migtools/gitops-primer-webhook:v0.0.11
        - quay.io/migtools/gitops-primer-downloader:v0.0.11
        - quay.io/migtools/gitops-primer-export:v0.0.11
        - quay.io/openshift/origin-oauth-proxy:4.9
        version: 0.0.11
      entries:
      - name: gitops-primer.v0.0.11
        version: 0.0.11
      - name: gitops-primer.v0.0.10
        version: 0.0.10
      - name: gitops-primer.v0.0.9
        version: 0.0.9
      - name: gitops-primer.v0.0.8
        version: 0.0.8
      - name: gitops-primer.v0.0.7
        version: 0.0.7
      - name: gitops-primer.v0.0.6
        version: 0.0.6
      - name: gitops-primer.v0.0.5
        version: 0.0.5
      - name: gitops-primer.v0.0.4
        version: 0.0.4
      - name: gitops-primer.v0.0.3
        version: 0.0.3
      - name: gitops-primer.v0.0.2
        version: 0.0.2
      - name: gitops-primer.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: gitops-primer
    provider:
      name: Konveyor
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Craig Trought
      provider-url: ""
    name: prometheus
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometheusoperator.0.56.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "Prometheus",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "replicas": 2,
                  "serviceAccountName": "prometheus-k8s",
                  "serviceMonitorSelector": {},
                  "ruleSelector": {},
                  "podMonitorSelector": {},
                  "probeSelector": {},
                  "alerting": {
                    "alertmanagers": [
                      {
                        "namespace": "monitoring",
                        "name": "alertmanager-main",
                        "port": "web"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "ServiceMonitor",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "operated-prometheus": "true"
                    }
                  },
                  "endpoints": [
                    {
                      "port": "web",
                      "interval": "30s"
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "PodMonitor",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "app": "prometheus"
                    }
                  },
                  "podMetricsEndpoints": [
                    {
                      "port": "web",
                      "interval": "30s"
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "Alertmanager",
                "metadata": {
                  "name": "alertmanager-main"
                },
                "spec": {
                  "replicas": 3,
                  "alertmanagerConfigSelector": {}
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "PrometheusRule",
                "metadata": {
                  "name": "prometheus-example-rules"
                },
                "spec": {
                  "groups": [
                    {
                      "name": "./example.rules",
                      "rules": [
                        {
                          "alert": "ExampleAlert",
                          "expr": "vector(1)"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1alpha1",
                "kind": "AlertmanagerConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "receivers": [
                    {
                      "name": "example"
                    }
                  ],
                  "route": {
                    "receiver": "example"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/prometheus-operator/prometheus-operator:v0.56.3
          createdAt: "2022-05-24T9:40:00Z"
          description: Manage the full lifecycle of configuring and managing Prometheus
            and Alertmanager servers.
          repository: https://github.com/prometheus-operator/prometheus-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A running Prometheus instance
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.coreos.com
            version: v1
          - description: A Prometheus Rule configures groups of sequentially evaluated
              recording and alerting rules.
            displayName: Prometheus Rule
            kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - description: Configures prometheus to monitor a particular k8s service
            displayName: Service Monitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - description: Configures prometheus to monitor a particular pod
            displayName: Pod Monitor
            kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
          - description: Configures an Alertmanager for the namespace
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.coreos.com
            version: v1
          - description: Configures subsections of Alertmanager configuration specific
              to each namespace
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.coreos.com
            version: v1alpha1
          - description: Defines how groups of ingresses and static targets should
              be monitored.
            displayName: Probe
            kind: Probe
            name: probes.monitoring.coreos.com
            version: v1
          - description: Allows recording and alerting rules to be processed across
              multiple Prometheus instances.
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.coreos.com
            version: v1
        description: |
          The Prometheus Operator for Kubernetes provides easy monitoring definitions for Kubernetes services and deployment and management of Prometheus instances.

          Once installed, the Prometheus Operator provides the following features:

          * **Create/Destroy**: Easily launch a Prometheus instance for your Kubernetes namespace, a specific application or team easily using the Operator.
          * **Simple Configuration**: Configure the fundamentals of Prometheus like versions, persistence, retention policies, and replicas from a native Kubernetes resource.
          * **Target Services via Labels**: Automatically generate monitoring target configurations based on familiar Kubernetes label queries; no need to learn a Prometheus specific configuration language.

          ### Other Supported Features

          **High availability**

          Multiple instances are run across failure zones and data is replicated. This keeps your monitoring available during an outage, when you need it most.

          **Updates via automated operations**

          New Prometheus versions are deployed using a rolling update with no downtime, making it easy to stay up to date.

          **Handles the dynamic nature of containers**

          Alerting rules are attached to groups of containers instead of individual instances, which is ideal for the highly dynamic nature of container deployment.
        displayName: Prometheus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - prometheus
        - monitoring
        - tsdb
        - alerting
        - observability
        links:
        - name: Prometheus
          url: https://www.prometheus.io/
        - name: Documentation
          url: https://github.com/prometheus-operator/prometheus-operator/tree/master/Documentation
        - name: Prometheus Operator
          url: https://github.com/prometheus-operator/prometheus-operator
        maintainers:
        - email: k8s@trought.ca
          name: Craig Trought
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Craig Trought
        relatedImages:
        - quay.io/prometheus-operator/prometheus-operator:v0.56.3
        version: 0.56.3
      entries:
      - name: prometheusoperator.0.56.3
        version: 0.56.3
      - name: prometheusoperator.0.47.0
        version: 0.47.0
      name: beta
    defaultChannel: beta
    packageName: prometheus
    provider:
      name: Craig Trought
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC
      provider-url: ""
    name: nspc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nspc-operator.v1.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.nec.com/v1",
                "kind": "NSPC",
                "metadata": {
                  "name": "nspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.connect.redhat.com/nec/nspc-operator@sha256:50fe10f05d8aa7db1b97280e5842bb0b67c6f9bfb01c0d17c5be3cd11b78e853
          createdAt: "2025-07-24T00:00:00Z"
          description: An operator for managing NEC Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: NEC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NSPC is the Schema for the nspcs API
            displayName: NSPC
            kind: NSPC
            name: nspcs.csi.nec.com
            version: v1
        description: |-
          ## About
          NEC Storage Plug-in for Containers (NSPC) is a plugin that integrates NEC Storage V Series into Kubernetes based clusters.
          NSPC provides persistent volume provisioning capabilities from NEC Storage V Series arrays.

          For full documentation, refer to the reference guide.

          ## Requirements

          ### Supported Driver Version

          * NSPC v3.16.0
        displayName: NEC Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - NEC
        - Storage
        - iStorage
        - V Series
        - Container Storage Interface
        - CSI
        maturity: stable
        provider:
          name: NEC
        relatedImages:
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:81d32545fb1b2c319bfeb36a70db4b4d3aea9c51c0574b209b8f64750b32d2f0
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.connect.redhat.com/nec/nspc-operator@sha256:50fe10f05d8aa7db1b97280e5842bb0b67c6f9bfb01c0d17c5be3cd11b78e853
        - registry.connect.redhat.com/nec/nspc-csi-driver@sha256:e50d60e025b3c8ae522a8c5ae78a07f8a624a1210ed7a078c8084efd76ccc646
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        version: 1.16.0
      entries:
      - name: nspc-operator.v1.16.0
        version: 1.16.0
      name: stable
    defaultChannel: stable
    packageName: nspc-operator
    provider:
      name: NEC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: tempo-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tempo-operator.v0.18.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoMonolithic",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "jaegerui": {
                    "enabled": true,
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    }
                  },
                  "storage": {
                    "traces": {
                      "backend": "memory"
                    }
                  }
                }
              },
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoStack",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "resources": {
                    "total": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2Gi"
                      }
                    }
                  },
                  "storage": {
                    "secret": {
                      "name": "my-storage-secret",
                      "type": "s3"
                    }
                  },
                  "storageSize": "1Gi",
                  "template": {
                    "queryFrontend": {
                      "jaegerQuery": {
                        "enabled": true,
                        "ingress": {
                          "type": "route"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/rhosdt/tempo-rhel8-operator@sha256:9bfa5a2ffb09b501f001a4275379210621f06620adcf6454a9327c65538872f5
          createdAt: 22 Sep 2025, 11:15
          description: Create and manage deployments of Tempo, a high-scale distributed
            tracing backend.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.18.0-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-tempo-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/tempo-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TempoMonolithic manages a Tempo deployment in monolithic
              mode.
            displayName: TempoMonolithic
            kind: TempoMonolithic
            name: tempomonolithics.tempo.grafana.com
            version: v1alpha1
          - description: TempoStack manages a Tempo deployment in microservices mode.
            displayName: TempoStack
            kind: TempoStack
            name: tempostacks.tempo.grafana.com
            version: v1alpha1
        description: |-
          Red Hat OpenShift distributed tracing platform based on Tempo. Tempo is an open-source, easy-to-use, and highly scalable distributed tracing backend. It provides observability for microservices architectures by allowing developers to track requests as they flow through distributed systems. Tempo is optimized to handle large volumes of trace data and is designed to be highly performant even under heavy loads.
          It can ingest common open source tracing protocols including Jaeger, Zipkin, and OpenTelemetry and requires only object storage to operate.

          The Tempo Operator supports Tempo deployments in Microservices mode (`TempoStack` CR) and Monolithic mode (`TempoMonolithic` CR).

          ### Core capabilities
          Tempo is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed transaction monitoring
          * Root cause analysis
          * Performance / latency optimization

          ### Operator features
          * **Resource Limits** - Specify overall resource requests and limits in the `TempoStack` CR; the operator assigns fractions of it to each component
          * **AuthN and AuthZ** - Supports OpenID Control (OIDC) and role-based access control (RBAC)
          * **Managed upgrades** - Updating the operator will automatically update all managed Tempo clusters
          * **Multitenancy** - Multiple tenants can send traces to the same Tempo cluster
          * **mTLS** - Communication between the Tempo components can be secured via mTLS
          * **Jaeger UI** - Traces can be visualized in Jaeger UI and exposed via Ingress or OpenShift Route
          * **Observability** - The operator and `TempoStack` operands expose telemetry (metrics, traces) and integrate with Prometheus `ServiceMonitor` and `PrometheusRule`

          ### Before you start
          `TempoStack` requires object storage to store its traces. `TempoMonolithic` can store traces in-memory, in a Persistent Volume and in object storage.
          Please ensure that your system has a compatible object storage solution that is supported, such as OpenShift Data Foundation, Minio, AWS S3, Azure Storage, or Google Cloud Storage.

          ### Examples
          There is a list of examples to help you deploy different configurations in [this GitHub repository](https://github.com/os-observability/redhat-rhosdt-samples).

          ### Support & Troubleshooting
          Tempo Operator is available and supported as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/distributed_tracing) is recommended when installing, configuring, and managing the Operator and instances.
        displayName: Tempo Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tempo
        - tracing
        - observability
        - monitoring
        - database
        links:
        - name: Tempo Operator
          url: https://github.com/grafana/tempo-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6bcfa981a7a75c242a16afa6dab36bf6e3b8a1442fa863875919ad551f48eceb
        - registry.redhat.io/rhosdt/tempo-rhel8@sha256:530a0ef3aaea3f07d6210a683a771c416cea8ce5accac8f5f751dcd7232cb341
        - registry.redhat.io/rhosdt/tempo-query-rhel8@sha256:5b89188bc587a81af3ecdca3caabd6e48356e9ba4f9070860f5160544ce30eb9
        - registry.redhat.io/rhosdt/tempo-rhel8-operator@sha256:9bfa5a2ffb09b501f001a4275379210621f06620adcf6454a9327c65538872f5
        - registry.redhat.io/rhosdt/tempo-jaeger-query-rhel8@sha256:53b62fea39ce211842026e8365d61f579b8e01e8587a1ce639423d72f3b44f45
        - registry.redhat.io/rhosdt/tempo-gateway-rhel8@sha256:095ec83219660bb1955758b94a9431b3add572da4dc2cba61b42b59fa6b4bba0
        - registry.redhat.io/rhosdt/tempo-gateway-opa-rhel8@sha256:22e876837d00371d24feac3f989ead3d4f034400b292f043b18f3839bfde564a
        version: 0.18.0-1
      entries:
      - name: tempo-operator.v0.18.0-1
        version: 0.18.0-1
      - name: tempo-operator.v0.16.0-2
        version: 0.16.0-2
      - name: tempo-operator.v0.16.0-1
        version: 0.16.0-1
      - name: tempo-operator.v0.15.4-1
        version: 0.15.4-1
      - name: tempo-operator.v0.15.3-1
        version: 0.15.3-1
      - name: tempo-operator.v0.14.1-2
        version: 0.14.1-2
      - name: tempo-operator.v0.14.1-1
        version: 0.14.1-1
      - name: tempo-operator.v0.13.0-1-0.1730817419.p
        version: 0.13.0-1+0.1730817419.p
      - name: tempo-operator.v0.13.0-1
        version: 0.13.0-1
      - name: tempo-operator.v0.10.0-8
        version: 0.10.0-8
      - name: tempo-operator.v0.10.0-7
        version: 0.10.0-7
      - name: tempo-operator.v0.10.0-6
        version: 0.10.0-6
      - name: tempo-operator.v0.8.0-3
        version: 0.8.0-3
      - name: tempo-operator.v0.8.0-2
        version: 0.8.0-2
      - name: tempo-operator.v0.6.0-2
        version: 0.6.0-2
      - name: tempo-operator.v0.3.1-3
        version: 0.3.1-3
      - name: tempo-operator.v0.1.0-6
        version: 0.1.0-6
      - name: tempo-operator.v0.1.0-6-0.1687199947.p
        version: 0.1.0-6+0.1687199947.p
      name: stable
    defaultChannel: stable
    packageName: tempo-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: logging-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: logging-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Elasticsearch","metadata":{"name":"elasticsearch"},"spec":{"esClusterName":"prod","esVersion":"7.17.0","esSecurity":{"autoGeneratePassword":true,"tlsEnabled":true}}},{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Fluentd","metadata":{"name":"fluentd"},"spec":{"esCluster":{"host":"elasticsearch-master"},"esSecurity":{"tlsEnabled":true,"existingSecret":"elasticsearch-password"},"indexNameStrategy":"namespace_name"}},{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Kibana","metadata":{"name":"kibana"},"spec":{"replicas":1,"esCluster":{"host":"https://elasticsearch-master:9200","esVersion":"7.16.0","clusterName":"elasticsearch"},"esSecurity":{"tlsEnabled":true,"existingSecret":"elasticsearch-password"}}}]'
          capabilities: Seamless Upgrades
          categories: Database,  Monitoring, Big Data
          certified: "false"
          containerImage: quay.io/opstree/logging-operator:v0.4.0
          createdAt: 04-07-2022
          description: A golang-based operator to create and manage EFK (Elasticsearch,
            Fluentd, and Kibana) stack on Kubernetes.
          k8sMaxVersion: ""
          k8sMinVersion: ""
          repository: https://github.com/ot-container-kit/logging-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Elasticsearch
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Index Life Cycle
            displayName: IndexLifeCycle
            kind: IndexLifeCycle
            name: indexlifecycles.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Index Template
            displayName: IndexTemplate
            kind: IndexTemplate
            name: indextemplates.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Fluentd
            displayName: Fluentd
            kind: Fluentd
            name: fluentds.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Kibana
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.logging.opstreelabs.in
            version: v1beta1
        description: |2+

          Logging Operator is an operator created in Golang to set up and manage EFK(Elasticsearch, Fluentd, and Kibana) cluster inside Kubernetes and Openshift environment. This operator is capable of setting up each individual component of EFK cluster separately.
          ### Supported Features
          - Setup of elasticsearch cluster with different nodes type:- master, data, ingestion, and client.
          - Customizable configuration for elasticsearch, fluentd stack.
          - Setup of fluentd as light-weight log-shipper as Daemonset.
          - Kibana will be set up as visualization tool for elastic stack.
          - Seamless upgrades of elasticsearch, fluentd, and kibana.
          - Security best practices support for complete stack such as TLS, elastic security.
          - Kubernetes resources objects support like:- resources, securityContext, affinity, tolerations, etc.
          - Elasticsearch plugins and keystore seamless support with operator
          ### Documentation
          **[Documentation](https://ot-logging-operator.netlify.app/)**
          ### Getting Help
          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.
          ### Contributing
          You can contribute to this project by:-
          * Raising Bugs and Feature Requests
          * Fixing issue by opening a Pull Request
          * Improving Documentation
          ### License
          MongoDB Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/logging-operator/blob/master/LICENSE)

        displayName: Logging Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Operator
        - Elasticsearch
        - Fluentd
        - Kibana
        - Opstree
        - Logging
        - Monitoring
        - Cloud Native
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-logging-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep
        maturity: beta
        minKubeVersion: 1.17.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/opstree/logging-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: logging-operator.v0.4.0
        version: 0.4.0
      - name: logging-operator.v0.3.0
        version: 0.3.0
      name: beta
    defaultChannel: beta
    packageName: logging-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Google
      provider-url: https://cloud.google.com/alloydb/omni
    name: alloydb-omni-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: alloydb-omni-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "backupPlanRef": "backupplan1",
                  "dbclusterRef": "dbcluster-sample",
                  "manual": true,
                  "physicalBackupSpec": {
                    "backupType": "full"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "BackupPlan",
                "metadata": {
                  "name": "backupplan1"
                },
                "spec": {
                  "backupRetainDays": 14,
                  "backupSchedules": {
                    "full": "0 0 * * 0",
                    "incremental": "0 21 * * *"
                  },
                  "dbclusterRef": "dbcluster-sample",
                  "paused": false
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "clone1"
                },
                "spec": {
                  "clonedDBClusterConfig": {
                    "dbclusterName": "new-dbcluster-sample"
                  },
                  "pointInTime": "2024-02-23T19:59:43Z",
                  "sourceDBCluster": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "DBCluster",
                "metadata": {
                  "name": "dbcluster-sample"
                },
                "spec": {
                  "databaseVersion": "16.3.0",
                  "primarySpec": {
                    "adminUser": {
                      "passwordRef": {
                        "name": "db-pw-dbcluster-sample"
                      }
                    },
                    "resources": {
                      "cpu": 1,
                      "disks": [
                        {
                          "name": "DataDisk",
                          "size": "10Gi"
                        }
                      ],
                      "memory": "5Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "DBInstance",
                "metadata": {
                  "name": "dbcluster-sample-rp-1"
                },
                "spec": {
                  "dbcParent": {
                    "name": "dbcluster-sample"
                  },
                  "instanceType": "ReadPool",
                  "nodeCount": 2,
                  "resources": {
                    "cpu": 2,
                    "disks": [
                      {
                        "name": "DataDisk",
                        "size": "15Gi"
                      }
                    ],
                    "memory": "6Gi"
                  },
                  "schedulingconfig": {
                    "nodeaffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "preference": {
                            "matchExpressions": [
                              {
                                "key": "another-node-label-key",
                                "operator": "In",
                                "values": [
                                  "another-node-label-value"
                                ]
                              }
                            ]
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "podAffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "podAffinityTerm": {
                            "labelSelector": {
                              "matchExpressions": [
                                {
                                  "key": "app",
                                  "operator": "In",
                                  "values": [
                                    "store"
                                  ]
                                }
                              ]
                            },
                            "topologyKey": "kubernetes.io/hostname"
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "podAntiAffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "podAffinityTerm": {
                            "labelSelector": {
                              "matchExpressions": [
                                {
                                  "key": "security",
                                  "operator": "In",
                                  "values": [
                                    "S1"
                                  ]
                                }
                              ]
                            },
                            "topologyKey": "topology.kubernetes.io/zone"
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane",
                        "operator": "Exists"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Failover",
                "metadata": {
                  "name": "failover-sample"
                },
                "spec": {
                  "dbclusterRef": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "PgBouncer",
                "metadata": {
                  "name": "mypgbouncer"
                },
                "spec": {
                  "allowSuperUserAccess": true,
                  "dbclusterRef": "dbcluster-sample",
                  "parameters": {
                    "default_pool_size": "15",
                    "ignore_startup_parameters": "extra_float_digits",
                    "max_client_conn": "800",
                    "max_db_connections": "160",
                    "pool_mode": "transaction"
                  },
                  "podSpec": {
                    "image": "gcr.io/alloydb-omni/operator/g-pgbouncer:1.5.0",
                    "resources": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "replicaCount": 1,
                  "serviceOptions": {
                    "type": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Replication",
                "metadata": {
                  "name": "replication-downstream-sample"
                },
                "spec": {
                  "dbcluster": {
                    "name": "dbcluster-sample"
                  },
                  "downstream": {
                    "control": "setup",
                    "host": "10.10.10.10",
                    "password": {
                      "name": "ha-rep-pw-dbcluster-sample"
                    },
                    "port": 5432,
                    "replicationSlotName": "dbcluster_sample_replication_upstream_sample",
                    "username": "alloydbreplica"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Replication",
                "metadata": {
                  "name": "replication-upstream-sample"
                },
                "spec": {
                  "dbcluster": {
                    "name": "dbcluster-sample"
                  },
                  "upstream": {}
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backup": "backup1",
                  "sourceDBCluster": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Sidecar",
                "metadata": {
                  "name": "sidecar-sample"
                },
                "spec": {
                  "sidecars": [
                    {
                      "args": [
                        "-c",
                        "while [ true ]\ndo\ndate\nset -x\nls -lh /logs/diagnostic\nset +x\ndone\n"
                      ],
                      "command": [
                        "/bin/sh"
                      ],
                      "image": "busybox",
                      "name": "sidecar-sample",
                      "volumeMounts": [
                        {
                          "mountPath": "/logs",
                          "name": "obsdisk"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Switchover",
                "metadata": {
                  "name": "switchover-sample"
                },
                "spec": {
                  "dbclusterRef": "dbcluster-sample",
                  "newPrimary": "aaaa-dbcluster-sample"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: gcr.io/alloydb-omni/operator/fleet-operator:1.5.0
          createdAt: "2025-07-25T04:12:04Z"
          operatorframework.io/suggested-namespace: alloydb-omni-system
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["backuprepositories.alloydbomni.internal.dbadmin.goog",
            "failovers.alloydbomni.internal.dbadmin.goog", "instancebackupplans.alloydbomni.internal.dbadmin.goog","instancebackups.alloydbomni.internal.dbadmin.goog","instancerestores.alloydbomni.internal.dbadmin.goog","instances.alloydbomni.internal.dbadmin.goog","instanceswitchovers.alloydbomni.internal.dbadmin.goog","instanceuserdefinedauthentications.alloydbomni.internal.dbadmin.goog","lrojobs.alloydbomni.internal.dbadmin.goog","replicationconfigs.alloydbomni.internal.dbadmin.goog","sidecars.alloydbomni.internal.dbadmin.goog","createstandbyjobs.alloydbomni.internal.dbadmin.goog","deletestandbyjobs.alloydbomni.internal.dbadmin.goog"]'
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BackupPlan is the Schema for the backupplans API.
            displayName: Backup Plan
            kind: BackupPlan
            name: backupplans.alloydbomni.dbadmin.goog
            version: v1
          - description: Backup is the Schema for the backups API.
            displayName: Backup
            kind: Backup
            name: backups.alloydbomni.dbadmin.goog
            version: v1
          - description: CreateStandbyJob is an internal workflow tracking object.
              Users should not directly interact with this.
            displayName: (Internal) Create Standby Job
            kind: CreateStandbyJob
            name: createstandbyjobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: DBCluster is the Schema for the DBClusters API.
            displayName: DB Cluster
            kind: DBCluster
            name: dbclusters.alloydbomni.dbadmin.goog
            version: v1
          - description: DBInstance is the Schema for the DBInstances API.
            displayName: DB Instance
            kind: DBInstance
            name: dbinstances.alloydbomni.dbadmin.goog
            version: v1
          - description: DeleteStandbyJob is an internal workflow tracking object.
              Users should not directly interact with this.
            displayName: (Internal) Delete Standby Job
            kind: DeleteStandbyJob
            name: deletestandbyjobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Failover represents the parameters and status of a single
              failover operation.
            displayName: Failover
            kind: Failover
            name: failovers.alloydbomni.dbadmin.goog
            version: v1
          - description: PgBouncer is the Schema for the PgBouncer API.
            displayName: PgBouncer
            kind: PgBouncer
            name: pgbouncers.alloydbomni.dbadmin.goog
            version: v1
          - description: Replication is the Schema for the Replications API.
            displayName: Replication
            kind: Replication
            name: replications.alloydbomni.dbadmin.goog
            version: v1
          - description: Restore is the Schema for the Restores API.
            displayName: Restore
            kind: Restore
            name: restores.alloydbomni.dbadmin.goog
            version: v1
          - description: Sidecar is the Schema for the Sidecar API.
            displayName: Sidecar
            kind: Sidecar
            name: sidecars.alloydbomni.dbadmin.goog
            version: v1
          - description: Switchover represents the parameters and status of a single
              switchover.
            displayName: Switchover
            kind: Switchover
            name: switchovers.alloydbomni.dbadmin.goog
            version: v1
          - description: UserDefinedAuthentication is the Schema for the UserDefinedAuthentications
              API.
            displayName: User Defined Authentication
            kind: UserDefinedAuthentication
            name: userdefinedauthentications.alloydbomni.dbadmin.goog
            version: v1
          - description: BackupRepository is the Schema for the backuprepositories
              API.
            displayName: (Internal) Backup Repository
            kind: BackupRepository
            name: backuprepositories.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Failover is the Schema for the failover API.
            displayName: (Internal) Failover
            kind: Failover
            name: failovers.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceBackupPlan is the Schema for the InstanceBackupPlan
              API.
            displayName: (Internal) Instance Backup Plan
            kind: InstanceBackupPlan
            name: instancebackupplans.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceBackup is the Schema for the InstanceBackup API.
            displayName: (Internal) Instance Backup
            kind: InstanceBackup
            name: instancebackups.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceRestore is the Schema for the InstanceRestore API.
            displayName: (Internal) Instance Restore
            kind: InstanceRestore
            name: instancerestores.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Instance is the Schema for the Instance API.
            displayName: (Internal) Instance
            kind: Instance
            name: instances.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceSwitchover is the Schema for the InstanceSwitchover
              API.
            displayName: (Internal) Instance Switchover
            kind: InstanceSwitchover
            name: instanceswitchovers.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceUserDefinedAuthentication is the Schema for the InstanceUserDefinedAuthentication
              API.
            displayName: (Internal) Instance User Defined Authentication
            kind: InstanceUserDefinedAuthentication
            name: instanceuserdefinedauthentications.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: LROJob is the Schema for the LROJob API.
            displayName: (Internal) LRO Job
            kind: LROJob
            name: lrojobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: ReplicationConfig is the Schema for the ReplicationConfig
              API.
            displayName: (Internal) Replication Config
            kind: ReplicationConfig
            name: replicationconfigs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Sidecar is the Schema for the Sidecar API.
            displayName: (Internal) Sidecar
            kind: Sidecar
            name: sidecars.alloydbomni.internal.dbadmin.goog
            version: v1
        description: |-
          AlloyDB Omni is a database that lets you deploy a streamlined edition of AlloyDB for PostgreSQL in your own computing environment. While being 100% compatible with PostgreSQL applications and tools, AlloyDB Omni delivers higher performance for both OLTP and OLAP workloads and makes administration easy with a number of autopilots. AlloyDB AI provides powerful tools to build generative AI applications with the ability to create, store, and index vector embeddings right in your database.

          ## Post-installation

          After installing the operator, you need to manually create the cert-manager resources in your cluster as this is a strict dependency for our operator. For this, follow these steps:

          ```
          kubectl create ns ${NAMESPACE:?}
          ```

          ```
          kubectl apply -f - <<EOF
          apiVersion: cert-manager.io/v1
          kind: ClusterIssuer
          metadata:
            name: alloydbomni-selfsigned-cluster-issuer
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Issuer
          metadata:
            name: fleet-selfsigned-issuer
            namespace: ${NAMESPACE:?}
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Certificate
          metadata:
            name: fleet-serving-cert
            namespace: ${NAMESPACE:?}
          spec:
            dnsNames:
            - fleet-webhook-service.alloydb-omni-system.svc
            - fleet-webhook-service.alloydb-omni-system.svc.cluster.local
            issuerRef:
              kind: Issuer
              name: fleet-selfsigned-issuer
            secretName: fleet-webhook-server-cert
          ---
          apiVersion: cert-manager.io/v1
          kind: Issuer
          metadata:
            name: local-selfsigned-issuer
            namespace: ${NAMESPACE:?}
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Certificate
          metadata:
            name: local-serving-cert
            namespace: ${NAMESPACE:?}
          spec:
            dnsNames:
            - local-webhook-service.alloydb-omni-system.svc
            - local-webhook-service.alloydb-omni-system.svc.cluster.local
            issuerRef:
              kind: Issuer
              name: local-selfsigned-issuer
            secretName: local-webhook-server-cert
          EOF
          ```

          Replace the following:
          * `NAMESPACE` : the namespace where you have installed your operator. Suggested namespace: `alloydb-omni-system`.

          ## Creating a database cluster

          After you install the AlloyDB Omni operator on your Kubernetes cluster, you can create an AlloyDB Omni database cluster. For each database cluster you create, you must first create a password secret using the following format:

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: db-pw-${DB_CLUSTER_NAME:s?}
            namespace: ${DB_CLUSTER_NAMESPACE:?}
          type: Opaque
          data:
            ${DB_CLUSTER_NAME:?}: "ENCODED_PASSWORD"
          ```

          Replace the following:
          * `DB_CLUSTER_NAME`: the name of this database cluster. For example, `my-db-cluster`.
          * `DB_CLUSTER_NAMESPACE` (Optional): the namespace where you want to create the database cluster. For example, `my-db-cluster-namespace`.
          * `ENCODED_PASSWORD`: the database login password for the default postgres user role, encoded as a base64 string. For example, `Q2hhbmdlTWUxMjM=` for password `ChangeMe123`.

          Follow [these][database-cluster] steps for creating a database cluster.

          ## Connect to AlloyDB Omni running on Kubernetes
          Follow [these][connect] steps to connect to AlloyDB Omni.

          [database-cluster]: https://cloud.google.com/alloydb/omni/docs/deploy-kubernetes#create
          [connect]: https://cloud.google.com/alloydb/omni/docs/run-connect#connect-kubernetes
        displayName: AlloyDB Omni Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - sql
        - postgres
        - vector
        - ai
        - analytics
        - postgresql
        - alloydb
        links:
        - name: Documentation
          url: https://cloud.google.com/alloydb/omni/docs
        - name: Release Notes
          url: https://cloud.google.com/alloydb/docs/release-notes
        maintainers:
        - email: alloydb-omni-ext@google.com
          name: Google Cloud
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Google
          url: https://cloud.google.com/alloydb/omni
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.1
        - gcr.io/alloydb-omni/operator/local-operator:1.5.0
        - gcr.io/alloydb-omni/operator/fleet-operator:1.5.0
        version: 1.5.0
      entries:
      - name: alloydb-omni-operator.v1.5.0
        version: 1.5.0
      - name: alloydb-omni-operator.v1.4.1
        version: 1.4.1
      - name: alloydb-omni-operator.v1.4.0
        version: 1.4.0
      - name: alloydb-omni-operator.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: alloydb-omni-operator
    provider:
      name: Google
      url: https://cloud.google.com/alloydb/omni
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Trilio
      provider-url: https://trilio.io/
    name: tvo-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tvo-operator.6.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOControlPlane",
                "metadata": {
                  "name": "tvocontrolplane-sample"
                },
                "spec": {
                  "crdType": "TVOControlPlane",
                  "common": {
                    "dmapi_workers": "16",
                    "memcached_servers": "",
                    "my_ip": {
                      "host_interface": null
                    },
                    "nova_group_id": "42436",
                    "nova_user_id": "42436",
                    "oslo_enable_proxy_headers_parsing": false,
                    "trustee_role": "creator"
                  },
                  "database": {
                    "common": {
                      "host": "EDIT_DB_HOSTNAME",
                      "root_password": "EDIT_PASSWORD",
                      "root_user_name": "root"
                    },
                    "datamover_api": {
                      "database": "dmapi",
                      "password": "EDIT_PASSWORD",
                      "user": "dmapi"
                    },
                    "wlm_api": {
                      "database": "workloadmgr",
                      "password": "EDIT_PASSWORD",
                      "user": "workloadmgr"
                    }
                  },
                  "helm3_hook": true,
                  "images": {
                    "triliovault_datamover_api": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_datamover_db_sync": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_cloud_trust": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_cron": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_db_sync": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_scheduler": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_workloads": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1"
                  },
                  "keystone": {
                    "common": {
                      "admin_role_name": "admin",
                      "auth_uri": "",
                      "auth_url": "",
                      "ca_cert": "CERT MULTIPLE LINES\n",
                      "cloud_admin_domain_name": "Default",
                      "cloud_admin_project_name": "admin",
                      "cloud_admin_user_name": "admin",
                      "cloud_admin_user_password": "",
                      "is_self_signed_ssl_cert": true,
                      "keystone_auth_host": "",
                      "keystone_auth_port": "",
                      "keystone_auth_protocol": "",
                      "region_name": "regionOne",
                      "service_project_domain_name": "Default",
                      "service_project_name": "service"
                    },
                    "datamover_api": {
                      "internal_endpoint": "ENDPOINT_PROTOCOL_INTERNAL://ENDPOINT_HOST_INTERNAL:8784/v2",
                      "password": "EDIT_PASSWORD",
                      "public_auth_host": "",
                      "public_endpoint": "ENDPOINT_PROTOCOL_PUBLIC://ENDPOINT_HOST_PUBLIC:8784/v2",
                      "service_desc": "TrilioVault Datamover Api Service",
                      "service_name": "dmapi",
                      "service_type": "datamover",
                      "user": "dmapi"
                    },
                    "identity": null,
                    "keystone_interface": "internal",
                    "wlm_api": {
                      "internal_endpoint": "ENDPOINT_PROTOCOL_INTERNAL://ENDPOINT_HOST_INTERNAL:8781/v1/$(tenant_id)s",
                      "password": "EDIT_PASSWORD",
                      "public_auth_host": "",
                      "public_endpoint": "ENDPOINT_PROTOCOL_PUBLIC://ENDPOINT_HOST_PUBLIC:8781/v1/$(tenant_id)s",
                      "service_desc": "TrilioVault Workloadmgr Api Service",
                      "service_name": "TrilioVaultWLM",
                      "service_type": "workloads",
                      "user": "triliovault"
                    }
                  },
                  "labels": {
                    "datamover_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_cron": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_scheduler": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_workloads": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    }
                  },
                  "pod": {
                    "replicas": {
                      "triliovault_datamover_api": 3,
                      "triliovault_object_store": 3,
                      "triliovault_wlm_api": 3,
                      "triliovault_wlm_cron": 1,
                      "triliovault_wlm_scheduler": 3,
                      "triliovault_wlm_workloads": 3
                    },
                    "security_context": {
                      "triliovault": {
                        "container": {
                          "triliovault_datamover_api": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 0,
                            "runAsUser": 63630
                          },
                          "triliovault_datamover_api_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 63630,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_api": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_cron": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_scheduler": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_workloads": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "wlm_cloud_trust_creation": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 42436
                          }
                        }
                      }
                    }
                  },
                  "rabbitmq": {
                    "common": {
                      "admin_password": "EDIT_PASSWORD",
                      "admin_user": "admin",
                      "driver": "messagingv2",
                      "host": "EDIT_HOST",
                      "port": "5672"
                    },
                    "datamover_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "",
                      "user": "dmapi",
                      "vhost": "dmapi"
                    },
                    "wlm_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "",
                      "user": "workloadmgr",
                      "vhost": "workloadmgr"
                    }
                  },
                  "triliovault_backup_targets": [
                    {
                      "backup_target_name": "S3_BT1",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY1",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET1",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL1",
                      "s3_region_name": "REGION1",
                      "s3_secret_key": "SECRETKEY1",
                      "s3_self_signed_cert": false,
                      "s3_signature_version": "default",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "S3_BT2",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY2",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET2",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL2",
                      "s3_region_name": "REGION2",
                      "s3_secret_key": "SECRETKEY2",
                      "s3_self_signed_cert": true,
                      "s3_signature_version": "default",
                      "s3_ssl_ca_cert": "SSL_CERT_LINE_1\n",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "NFS_BT3",
                      "backup_target_type": "nfs",
                      "is_default": false,
                      "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10",
                      "nfs_shares": "11.30.1.10:/mnt/share"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:3d80cad27a1043fb61735e7a6fb156c98b269cfcc68f8b9712713a7d4f6ad6bd
          createdAt: "2025-01-14T11:20:30Z"
          description: Cloud-Native Data Protection for OpenStack
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trilio-openstack
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trilioData/triliovault-cfg-scripts
          support: Trilio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trilio for OpenStack Control Plane. Includes all control
              plane components like Datamover Api, Trilio object store and Workloadmgr
              services.
            displayName: Trilio for OpenStack Control Plane
            kind: TVOControlPlane
            name: tvocontrolplanes.tvo.trilio.io
            version: v1
          - description: Trilio for OpenStack Backup Target. This CR is responsible
              to add a new backup target.
            displayName: Trilio for OpenStack Backup Target
            kind: TVOBackupTarget
            name: tvobackuptargets.tvo.trilio.io
            version: v1
        description: |-
          Trilio for OpenStack provides scalable, native, agentless, multi-tenant backup and recovery solutions, ensuring seamless protection of workloads with flexible storage options for dynamic cloud environments.

            * OpenStack Backup and Recovery: Recover any OpenStack application quickly in the event of data loss or misconfiguration. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Such as One-Click-Restore to entire project recovery on different clusters.
            * Disaster Recovery: Recover any OpenStack application quickly in the event of a disaster. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Trilio makes OpenStack disaster recovery easy.

            * Migration: A tenant or administrator can capture an application and its data and non-disruptively migrate that exact point-in-time to another OpenStack cloud. The need to migrate OpenStack workloads can be driven by economics, security or test/dev.

            * Network Functions Virtualization backup and recovery: Business critical services like NFV can quickly be configured to any state, assuring the optimal quality of service. Trilio provides recovery from data damage and allows restoration of entire VNFs, selected VNFs, or individual data items using a point-in-time working copy. The platform also allows recovery from a disaster by restoring the OpenStack network topology and populating it with a working copy of the entire NFV.

          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Backup
        - Recovery
        - RHOSO Backup
        - Disaster Recovery
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/openstack-backup-and-recovery/
        - name: Operator Documentation
          url: https://docs.trilio.io/openstack
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        provider:
          name: Trilio
          url: https://trilio.io/
        relatedImages:
        - registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82
        - registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1
        - registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:3d80cad27a1043fb61735e7a6fb156c98b269cfcc68f8b9712713a7d4f6ad6bd
        version: 6.1.1
      entries:
      - name: tvo-operator.6.1.1
        version: 6.1.1
      name: 6.0.x
    - currentCSV: tvo-operator.6.1.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOControlPlane",
                "metadata": {
                  "name": "tvocontrolplane-sample",
                  "namespace": "trilio-openstack"
                },
                "spec": {
                  "common": {
                    "dmapi_workers": "16",
                    "memcached_servers": "memcached-0.memcached.openstack.svc:11212",
                    "nova_group_id": "42436",
                    "nova_user_id": "42436",
                    "oslo_enable_proxy_headers_parsing": false,
                    "trustee_role": "member"
                  },
                  "crdType": "TVOControlPlane",
                  "database": {
                    "common": {
                      "host": "openstack.openstack.svc",
                      "port": 3306,
                      "root_password": "openstack",
                      "root_user_name": "root"
                    },
                    "datamover_api": {
                      "database": "dmapi",
                      "password": "dsw21aSyH5gHH6XE72k11Y3t3C308jY5rP02ndwJW",
                      "user": "dmapi"
                    },
                    "wlm_api": {
                      "database": "workloadmgr",
                      "password": "dfeeYb72OwUzX5YuRKcVjxXJaELV8YT0jZczIQwys0v",
                      "user": "workloadmgr"
                    }
                  },
                  "helm3_hook": true,
                  "images": {
                    "triliovault_datamover_api": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_datamover_db_sync": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_cloud_trust": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_cron": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_db_sync": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_scheduler": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_workloads": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd"
                  },
                  "keystone": {
                    "common": {
                      "admin_role_name": "admin",
                      "auth_uri": "https://keystone-internal.openstack.svc:5000/v3",
                      "auth_url": "https://keystone-internal.openstack.svc:5000",
                      "cloud_admin_domain_name": "Default",
                      "cloud_admin_project_name": "admin",
                      "cloud_admin_user_name": "admin",
                      "cloud_admin_user_password": "openstack",
                      "is_self_signed_ssl_cert": true,
                      "keystone_auth_host": "keystone-internal.openstack.svc",
                      "keystone_auth_port": "5000",
                      "keystone_auth_protocol": "https",
                      "region_name": "regionOne",
                      "service_project_domain_name": "Default",
                      "service_project_name": "service"
                    },
                    "datamover_api": {
                      "internal_endpoint": "https://triliovault-datamover-internal.trilio-openstack.svc:8784/v2",
                      "password": "defajgh48zgt88wFzzhwEt3xHGeCOdExYbLGQN2XRm",
                      "public_auth_host": "triliovault-datamover-public-trilio-openstack.apps.trilio.trilio.bos2",
                      "public_endpoint": "https://triliovault-datamover-public-trilio-openstack.apps.trilio.trilio.bos2/v2",
                      "service_desc": "TrilioVault Datamover Api Service",
                      "service_name": "dmapi",
                      "service_type": "datamover",
                      "user": "dmapi"
                    },
                    "keystone_interface": "internal",
                    "wlm_api": {
                      "internal_endpoint": "https://triliovault-wlm-internal.trilio-openstack.svc:8781/v1/$(tenant_id)s",
                      "password": "dsdeJl8axy0gv14fURwUUxy0qUOQbugUUOfXikz6j",
                      "public_auth_host": "triliovault-wlm-public-trilio-openstack.apps.trilio.trilio.bos2",
                      "public_endpoint": "https://triliovault-wlm-public-trilio-openstack.apps.trilio.trilio.bos2/v1/$(tenant_id)s",
                      "service_desc": "TrilioVault Workloadmgr Api Service",
                      "service_name": "TrilioVaultWLM",
                      "service_type": "workloads",
                      "user": "triliovault"
                    }
                  },
                  "labels": {
                    "datamover_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_cron": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_scheduler": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_workloads": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    }
                  },
                  "pod": {
                    "replicas": {
                      "triliovault_datamover_api": 3,
                      "triliovault_object_store": 3,
                      "triliovault_wlm_api": 3,
                      "triliovault_wlm_cron": 1,
                      "triliovault_wlm_scheduler": 3,
                      "triliovault_wlm_workloads": 3
                    },
                    "security_context": {
                      "triliovault": {
                        "container": {
                          "triliovault_datamover_api": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 0,
                            "runAsUser": 63630
                          },
                          "triliovault_datamover_api_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 63630,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_api": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_cron": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_scheduler": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_workloads": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "wlm_cloud_trust_creation": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 42436
                          }
                        }
                      }
                    }
                  },
                  "rabbitmq": {
                    "common": {
                      "admin_password": "EDIT_PASSWORD",
                      "admin_user": "admin",
                      "driver": "messagingv2",
                      "host": "EDIT_HOST",
                      "port": "15671",
                      "ssl": true
                    },
                    "datamover_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "rabbit://dmapi:${PASSWORD}@${RABBITMQ_HOST}:5671/dmapi?ssl=1",
                      "user": "dmapi",
                      "vhost": "dmapi"
                    },
                    "wlm_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "rabbit://workloadmgr:${PASSWORD}@${RABBITMQ_HOST}:5671/workloadmgr?ssl=1",
                      "user": "workloadmgr",
                      "vhost": "workloadmgr"
                    }
                  },
                  "triliovault_backup_targets": [
                    {
                      "backup_target_name": "BT1_S3",
                      "backup_target_type": "s3",
                      "is_default": false,
                      "s3_access_key": "",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "",
                      "s3_region_name": "",
                      "s3_secret_key": "",
                      "s3_self_signed_cert": false,
                      "s3_signature_version": "default",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "BT2_Other_S3",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY2",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET2",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL2",
                      "s3_region_name": "REGION2",
                      "s3_secret_key": "SECRETKEY2",
                      "s3_self_signed_cert": true,
                      "s3_signature_version": "default",
                      "s3_ssl_ca_cert": "SSL_CERT_LINE_1\nSSL_CERT_LINE_2\nSSL_CERT_LINE_N\n",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "other_s3"
                    },
                    {
                      "backup_target_name": "BT3_NFS",
                      "backup_target_type": "nfs",
                      "is_default": false,
                      "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10",
                      "nfs_shares": "192.168.2.3:/nfs/share"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOBackupTarget",
                "metadata": {
                  "name": "tvobackuptarget-sample",
                  "namespace": "trilio-openstack"
                },
                "spec": {
                  "crdType": "TVOBackupTarget",
                  "images": {
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd"
                  },
                  "triliovault_backup_target": {
                    "backup_target_name": "S3_BT4",
                    "backup_target_type": "s3",
                    "is_default": false,
                    "s3_type": "other_s3",
                    "s3_access_key": "",
                    "s3_secret_key": "",
                    "s3_region_name": "",
                    "s3_bucket": "",
                    "s3_endpoint_url": "",
                    "s3_signature_version": "default",
                    "s3_auth_version": "DEFAULT",
                    "s3_bucket_object_lock_enabled": false,
                    "s3_ssl_enabled": true,
                    "s3_ssl_verify": true,
                    "s3_self_signed_cert": true,
                    "s3_ssl_ca_cert": "SSL_CERT_LINE_1\nSSL_CERT_LINE_2\nSSL_CERT_LINE_N",
                    "nfs_shares": "192.168.2.30:/nfs/share",
                    "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:e282bb056a1780e5170db3e63efdf5099a5ac5240ad44162bc7af83cac893ac7
          createdAt: "2025-01-14T11:20:30Z"
          description: Cloud-Native Data Protection for OpenStack
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: tvo-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trilioData/triliovault-cfg-scripts
          support: Trilio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trilio for OpenStack Control Plane. Includes all control
              plane components like Datamover Api, Trilio object store and Workloadmgr
              services.
            displayName: Trilio for OpenStack Control Plane
            kind: TVOControlPlane
            name: tvocontrolplanes.tvo.trilio.io
            version: v1
          - description: Trilio for OpenStack Backup Target. This CR is responsible
              to add a new backup target.
            displayName: Trilio for OpenStack Backup Target
            kind: TVOBackupTarget
            name: tvobackuptargets.tvo.trilio.io
            version: v1
        description: |-
          Trilio for OpenStack provides scalable, native, agentless, multi-tenant backup and recovery solutions, ensuring seamless protection of workloads with flexible storage options for dynamic cloud environments.

            * OpenStack Backup and Recovery: Recover any OpenStack application quickly in the event of data loss or misconfiguration. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Such as One-Click-Restore to entire project recovery on different clusters.
            * Disaster Recovery: Recover any OpenStack application quickly in the event of a disaster. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Trilio makes OpenStack disaster recovery easy.

            * Migration: A tenant or administrator can capture an application and its data and non-disruptively migrate that exact point-in-time to another OpenStack cloud. The need to migrate OpenStack workloads can be driven by economics, security or test/dev.

            * Network Functions Virtualization backup and recovery: Business critical services like NFV can quickly be configured to any state, assuring the optimal quality of service. Trilio provides recovery from data damage and allows restoration of entire VNFs, selected VNFs, or individual data items using a point-in-time working copy. The platform also allows recovery from a disaster by restoring the OpenStack network topology and populating it with a working copy of the entire NFV.

          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Backup
        - Recovery
        - RHOSO Backup
        - Disaster Recovery
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/openstack-backup-and-recovery/
        - name: Operator Documentation
          url: https://docs.trilio.io/openstack
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        provider:
          name: Trilio
          url: https://trilio.io/
        relatedImages:
        - registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd
        - registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:e282bb056a1780e5170db3e63efdf5099a5ac5240ad44162bc7af83cac893ac7
        - registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82
        version: 6.1.9
      entries:
      - name: tvo-operator.6.1.9
        version: 6.1.9
      name: 6.1.x
    defaultChannel: 6.0.x
    packageName: tvo-operator
    provider:
      name: Trilio
      url: https://trilio.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: elasticsearch-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: elasticsearch-operator.v5.8.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Elasticsearch",
                "metadata": {
                  "name": "elasticsearch"
                },
                "spec": {
                  "indexManagement": {
                    "mappings": [
                      {
                        "aliases": [
                          "infra",
                          "logs.infra"
                        ],
                        "name": "infra",
                        "policyRef": "infra-policy"
                      }
                    ],
                    "policies": [
                      {
                        "name": "infra-policy",
                        "phases": {
                          "delete": {
                            "minAge": "2d",
                            "namespaceSpec": [
                              {
                                "minAge": "5h",
                                "namespace": "openshift-monitoring"
                              }
                            ],
                            "pruneNamespacesInterval": "24h"
                          },
                          "hot": {
                            "actions": {
                              "rollover": {
                                "maxAge": "8h"
                              }
                            }
                          }
                        },
                        "pollInterval": "30m"
                      }
                    ]
                  },
                  "managementState": "Managed",
                  "nodeSpec": {
                    "resources": {
                      "limits": {
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "512Mi"
                      }
                    }
                  },
                  "nodes": [
                    {
                      "nodeCount": 1,
                      "roles": [
                        "client",
                        "data",
                        "master"
                      ],
                      "storage": {
                        "size": "20G"
                      }
                    }
                  ],
                  "redundancyPolicy": "ZeroRedundancy"
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Kibana",
                "metadata": {
                  "name": "kibana"
                },
                "spec": {
                  "managementState": "Managed",
                  "nodeSelector": {},
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "memory": "512Mi"
                    },
                    "requests": {
                      "memory": "512Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
          createdAt: "2024-09-19T15:14:33Z"
          description: |
            The Elasticsearch Operator for OCP provides a means for configuring and managing an Elasticsearch cluster for tracing and cluster logging.
            ## Prerequisites and Requirements
            ### Elasticsearch Operator Namespace
            The Elasticsearch Operator must be deployed to the global operator group namespace
            ### Memory Considerations
            Elasticsearch is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Elasticsearch cluster.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory. Each ES node can operate with a
            lower memory setting though this is not recommended for production deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.21'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Cluster Logging, Jaeger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Elasticsearch cluster instance
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.openshift.io
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.openshift.io
            version: v1
        description: "### IN OCP 4.16 AND FORWARD \n\nELASTICSEARCH IS ONLY SUPPORTED
          FOR **SERVICEMESH, TRACING, AND KIALI**, AND THIS OPERATOR **WILL BE REMOVED
          FROM THE OPENSHIFT OPERATOR CATALOG IN NOVEMBER 2025**.  ELASTICSEARCH IS
          **NO LONGER SUPPORTED FOR LOG STORAGE**. **KIBANA IS NO LONGER SUPPORTED**
          IN OCP 4.16 AND FORWARD. \n\n### About \n\nThe Elasticsearch Operator for
          OCP provides a means for configuring and managing an Elasticsearch cluster
          for use in tracing \nand cluster logging as well as a Kibana instance to
          connect to it.\nThis operator only supports OCP Cluster Logging and Jaeger.
          \ It is tightly coupled to each and is not currently capable of\nbeing used
          as a general purpose manager of Elasticsearch clusters running on OCP.\n\nPlease
          note: For a general purpose Elasticsearch operator, please use Elastic's
          Elasticsearch (ECK) Operator [here](https://catalog.redhat.com/software/containers/elastic/eck-operator/5fabf6d1ecb52450895164be?container-tabs=gti)\n\nIt
          is recommended that this operator be installed in the `openshift-operators-redhat`
          namespace to \nproperly support the Cluster Logging and Jaeger use cases.\n\nOnce
          installed, the operator provides the following features for **Elasticsearch**:\n*
          **Create/Destroy**: Deploy an Elasticsearch cluster to the same namespace
          in which the elasticsearch CR is created.\n* **Update**: Changes to the
          elasticsearch CR will be scheduled and applied to the cluster in a controlled
          manner (most often as a rolling upgrade).\n* **Cluster health**: The operator
          will periodically poll the cluster to evaluate its current health (such
          as the number of active shards and if any cluster nodes have reached their
          storage watermark usage).\n* **Redeploys**: In the case where the provided
          secrets are updated, the Elasticsearch Operator will schedule and perform
          a full cluster restart.\n* **Index management**: The Elasticsearch Operator
          will create cronjobs to perform index management such as roll over and deletion.\n\nOnce
          installed, the operator provides the following features for **Kibana**:\n*
          **Create/Destroy**: Deploy a Kibana instance to the same namespace in which
          the kibana CR is created (this should be the same namespace as the elasticsearch
          CR).\n* **Update**: Changes to the kibana CR will be scheduled and applied
          to the cluster in a controlled manner.\n* **Redeploys**: In the case where
          the provided secrets are updated, the Elasticsearch Operator will perform
          a restart.\n\n### Additionally provided features\n* Out of the box multitenancy
          that is integrated with OCP user access control.\n* Document Level Security\n*
          mTLS communication between Elasticsearch, Kibana, Index Management cronjobs,
          and CLO's Fluentd\n* OCP prometheus dashboard for Elasticsearch clusters\n*
          Prometheus Alerting rules  \n"
        displayName: OpenShift Elasticsearch Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - jaeger
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Elasticsearch Operator
          url: https://github.com/openshift/elasticsearch-operator
        maintainers:
        - email: aos-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8d603100978c655020a8f4d81c527162348f468d93989edbc00463b8297fb8a5
        - registry.redhat.io/openshift-logging/kibana6-rhel8@sha256:a2c200667738fdb079ba11ed3b0df3c5901373428d5e560bc1efea9d5e1ddd67
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
        - registry.redhat.io/openshift-logging/logging-curator5-rhel9@sha256:394e0f46fb8be44c720b29aed716443747a636d2758226c9386f930f95758470
        - registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:b26d2218aa6b1efea83a1717d62f81009c97e8082434b918937c609f5cdab912
        version: 5.8.21
      entries:
      - name: elasticsearch-operator.v5.8.21
        version: 5.8.21
      - name: elasticsearch-operator.v5.8.20
        version: 5.8.20
      - name: elasticsearch-operator.v5.8.19
        version: 5.8.19
      - name: elasticsearch-operator.v5.8.18
        version: 5.8.18
      - name: elasticsearch-operator.v5.8.17
        version: 5.8.17
      - name: elasticsearch-operator.v5.8.16
        version: 5.8.16
      - name: elasticsearch-operator.v5.8.15
        version: 5.8.15
      - name: elasticsearch-operator.v5.8.14
        version: 5.8.14
      - name: elasticsearch-operator.v5.8.13
        version: 5.8.13
      - name: elasticsearch-operator.v5.8.12
        version: 5.8.12
      - name: elasticsearch-operator.v5.8.11
        version: 5.8.11
      - name: elasticsearch-operator.v5.8.10
        version: 5.8.10
      - name: elasticsearch-operator.v5.8.9
        version: 5.8.9
      - name: elasticsearch-operator.v5.8.8
        version: 5.8.8
      - name: elasticsearch-operator.v5.8.7
        version: 5.8.7
      - name: elasticsearch-operator.v5.8.6
        version: 5.8.6
      - name: elasticsearch-operator.v5.8.5
        version: 5.8.5
      - name: elasticsearch-operator.v5.8.4
        version: 5.8.4
      - name: elasticsearch-operator.v5.8.3
        version: 5.8.3
      - name: elasticsearch-operator.v5.8.2
        version: 5.8.2
      - name: elasticsearch-operator.v5.8.1
        version: 5.8.1
      - name: elasticsearch-operator.v5.8.0
        version: 5.8.0
      name: stable
    - currentCSV: elasticsearch-operator.v5.8.22
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Elasticsearch",
                "metadata": {
                  "name": "elasticsearch"
                },
                "spec": {
                  "indexManagement": {
                    "mappings": [
                      {
                        "aliases": [
                          "infra",
                          "logs.infra"
                        ],
                        "name": "infra",
                        "policyRef": "infra-policy"
                      }
                    ],
                    "policies": [
                      {
                        "name": "infra-policy",
                        "phases": {
                          "delete": {
                            "minAge": "2d",
                            "namespaceSpec": [
                              {
                                "minAge": "5h",
                                "namespace": "openshift-monitoring"
                              }
                            ],
                            "pruneNamespacesInterval": "24h"
                          },
                          "hot": {
                            "actions": {
                              "rollover": {
                                "maxAge": "8h"
                              }
                            }
                          }
                        },
                        "pollInterval": "30m"
                      }
                    ]
                  },
                  "managementState": "Managed",
                  "nodeSpec": {
                    "resources": {
                      "limits": {
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "512Mi"
                      }
                    }
                  },
                  "nodes": [
                    {
                      "nodeCount": 1,
                      "roles": [
                        "client",
                        "data",
                        "master"
                      ],
                      "storage": {
                        "size": "20G"
                      }
                    }
                  ],
                  "redundancyPolicy": "ZeroRedundancy"
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Kibana",
                "metadata": {
                  "name": "kibana"
                },
                "spec": {
                  "managementState": "Managed",
                  "nodeSelector": {},
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "memory": "512Mi"
                    },
                    "requests": {
                      "memory": "512Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.12-v4.17
          containerImage: registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:ce1f01f7a7fd8fa3d52f6bca62f63f65aebcb4c9a8ff8f6afaca7e5d121f0406
          createdAt: "2025-09-23T09:54:59Z"
          description: |
            The Elasticsearch Operator for OCP provides a means for configuring and managing an Elasticsearch cluster for tracing and cluster logging.
            ## Prerequisites and Requirements
            ### Elasticsearch Operator Namespace
            The Elasticsearch Operator must be deployed to the global operator group namespace
            ### Memory Considerations
            Elasticsearch is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Elasticsearch cluster.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory. Each ES node can operate with a
            lower memory setting though this is not recommended for production deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.22'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-5.8
          operators.operatorframework.io.bundle.channels.v1: stable-5.8
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: elasticsearch-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Cluster Logging, Jaeger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Elasticsearch cluster instance
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.openshift.io
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.openshift.io
            version: v1
        description: |
          ### IN OCP 4.16 AND FORWARD

          IN OCP 4.16 AND FORWARD ELASTICSEARCH and KIBANA ARE **NO LONGER SUPPORTED
          FOR LOG STORAGE** AND ELASTICSEARCH IS ONLY SUPPORTED FOR **SERVICEMESH,
          TRACING, AND KIALI**

          THIS OPERATOR **WILL REACH END OF LIFE ON NOVEMBER 2025**

          ### About

          The Elasticsearch Operator for OCP provides a means for configuring and
          managing an Elasticsearch cluster for use in tracing and cluster logging as
          well as a Kibana instance to connect to it. This operator only supports OCP
          Cluster Logging and Jaeger.  It is tightly coupled to each and is not
          currently capable of being used as a general purpose manager of
          Elasticsearch clusters running on OCP.

          Please note: For a general purpose Elasticsearch operator, please use
          Elastic's Elasticsearch (ECK) Operator
          [here](https://catalog.redhat.com/software/containers/elastic/eck-operator/5fabf6d1ecb52450895164be?container-tabs=gti)

          It is recommended that this operator be installed in the
          `openshift-operators-redhat` namespace to properly support the Cluster
          Logging and Jaeger use cases.

          Once installed, the operator provides the following features for
          **Elasticsearch**: * **Create/Destroy**: Deploy an Elasticsearch cluster to
          the same namespace in which the elasticsearch CR is created.
          * **Update**: Changes to the elasticsearch CR will be scheduled and applied
          to the cluster in a controlled manner (most often as a rolling upgrade).
          * **Cluster health**: The operator will periodically poll the cluster to
          evaluate its current health (such as the number of active shards and if any
          cluster nodes have reached their storage watermark usage).
          * **Redeploys**: In the case where the provided secrets are updated, the
          Elasticsearch Operator will schedule and perform a full cluster restart.
          * **Index management**: The Elasticsearch Operator will create cronjobs to
          perform index management such as roll over and deletion.

          Once installed, the operator provides the following features for **Kibana**:

          * **Create/Destroy**: Deploy a Kibana instance to the same namespace in
          which the kibana CR is created (this should be the same namespace as the
          elasticsearch CR).
          * **Update**: Changes to the kibana CR will be scheduled and applied to the
          cluster in a controlled manner.
          * **Redeploys**: In the case where the provided secrets are updated, the
          Elasticsearch Operator will perform a restart.

          ### Additionally provided features

          * Out of the box multitenancy that is integrated with OCP user access
          control.
          * Document Level Security
          * mTLS communication between Elasticsearch, Kibana, Index Management
          cronjobs, and CLO's Fluentd
          * OCP prometheus dashboard for Elasticsearch clusters
          * Prometheus Alerting rules
        displayName: OpenShift Elasticsearch Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - jaeger
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Elasticsearch Operator
          url: https://github.com/openshift/elasticsearch-operator
        maintainers:
        - email: aos-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:6388562dc496e524a297c733720dea23a2889468fa25394812b3247f5f08c907
        - registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:69878c9e40178530baf837a28d3f78e63eb6fba54865dbf05f30dda1a661d8d0
        - registry.redhat.io/openshift-logging/kibana6-rhel8@sha256:7cfb21861cbc3c6d8c6b6f0192ac3c687e4ac9b16fbccf56841e179a08553c84
        - registry.redhat.io/openshift-logging/logging-curator5-rhel9@sha256:dd1cef6a91ca25231af46e0417c7929663098a3b4ad1230dddf4a8171871ce0e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:ce1f01f7a7fd8fa3d52f6bca62f63f65aebcb4c9a8ff8f6afaca7e5d121f0406
        version: 5.8.22
      entries:
      - name: elasticsearch-operator.v5.8.22
        version: 5.8.22
      - name: elasticsearch-operator.v5.8.21
        version: 5.8.21
      - name: elasticsearch-operator.v5.8.20
        version: 5.8.20
      - name: elasticsearch-operator.v5.8.19
        version: 5.8.19
      - name: elasticsearch-operator.v5.8.18
        version: 5.8.18
      - name: elasticsearch-operator.v5.8.17
        version: 5.8.17
      - name: elasticsearch-operator.v5.8.16
        version: 5.8.16
      - name: elasticsearch-operator.v5.8.15
        version: 5.8.15
      - name: elasticsearch-operator.v5.8.14
        version: 5.8.14
      - name: elasticsearch-operator.v5.8.13
        version: 5.8.13
      - name: elasticsearch-operator.v5.8.12
        version: 5.8.12
      - name: elasticsearch-operator.v5.8.11
        version: 5.8.11
      - name: elasticsearch-operator.v5.8.10
        version: 5.8.10
      - name: elasticsearch-operator.v5.8.9
        version: 5.8.9
      - name: elasticsearch-operator.v5.8.8
        version: 5.8.8
      - name: elasticsearch-operator.v5.8.7
        version: 5.8.7
      - name: elasticsearch-operator.v5.8.6
        version: 5.8.6
      - name: elasticsearch-operator.v5.8.5
        version: 5.8.5
      - name: elasticsearch-operator.v5.8.4
        version: 5.8.4
      - name: elasticsearch-operator.v5.8.3
        version: 5.8.3
      - name: elasticsearch-operator.v5.8.2
        version: 5.8.2
      - name: elasticsearch-operator.v5.8.1
        version: 5.8.1
      - name: elasticsearch-operator.v5.8.0
        version: 5.8.0
      name: stable-5.8
    defaultChannel: stable-5.8
    packageName: elasticsearch-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: application-services-metering-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: application-services-metering-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "applicationservices.redhat.com/v1",
                "kind": "Meter",
                "metadata": {
                  "name": "app-services-metrics-example"
                },
                "spec": {
                  "includeInfrastructure": false,
                  "meterCollectionEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Application Runtime
          certified: "false"
          containerImage: quay.io/redhat-developer/application-services-metering-operator:0.6.1
          createdAt: "2021-09-28T13:00:00Z"
          description: Collect the core usage of products from the Application Services
            portfolio into a single metric labeled by Runtimes, Integration, and Automation.
          repository: https://github.com/redhat-developer/application-services-metering-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Meter
            displayName: Meter
            kind: Meter
            name: meters.applicationservices.redhat.com
            version: v1
        description: |-
          ## About the managed application

          There is no managed application created with this operator.

          ## About this Operator

          The metering operator watches pods for predetermined labels, across a cluster or namespace(s).
          The label schema for Red Hat Application Services is in the format:

          * `com.company=Red_Hat`
          * `rht.comp=PAM`
          * `rht.comp_ver=7.10`
          * `rht.prod_name=Red_Hat_Process_Automation`
          * `rht.prod_ver=7.11`
          * `rht.subcomp=rhpam-kogito-runtime`
          * `rht.subcomp_t=application`

          It's important not to remove labels as the operator will cease monitoring those pods.

          The operator exposes a metric named: `appsvcs_cpu_usage_cores`
          The metric has a label for product category, these are Runtimes, Integration and Automation.
          There is a prometheus rule using the metric to group by product name, `appsvcs_cores_by_product:sum`.

          The operator utilises the OpenShift Monitoring Stack to provide usage observability via cloud.redhat.com

          ## Prerequisites for enabling this Operator

          None.
        displayName: Application Services Metering Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - metering
        - operator
        links:
        - name: Red Hat Application Services portfolio
          url: https://www.redhat.com/en/products/middleware
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/redhat-developer/application-services-metering-operator:0.6.1
        version: 0.6.1
      entries:
      - name: application-services-metering-operator.v0.6.1
        version: 0.6.1
      - name: application-services-metering-operator.v0.6.0
        version: 0.6.0
      name: alpha
    defaultChannel: alpha
    packageName: application-services-metering-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-licensing-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-licensing-operator.v8.0.220
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Developer"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:feeabfe4ecd36205306e3f3bf6c11517cfee735d7ae7eb4350abd01e9d76610b
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          olm.skipRange: <8.0.220
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:2f3de2bf46920cfe8c251011fc99841b43999561c1d3fe44850b205e94585a68
        version: 8.0.220
      entries:
      - name: infoscale-licensing-operator.v8.0.220
        version: 8.0.220
      - name: infoscale-licensing-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-licensing-operator.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-licensing-operator.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:b357b2203466516684c8fbc209757a98a45ec2376981f8699faf7dd5e37486a3
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:b357b2203466516684c8fbc209757a98a45ec2376981f8699faf7dd5e37486a3
        version: 8.0.330
      entries:
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-licensing-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Arctera
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Arctera
        - InfoScale
        - Licensing
        - Arctera Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
        version: 8.0.410
      entries:
      - name: infoscale-licensing-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-licensing-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-licensing-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-licensing-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Arctera
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Arctera
        - InfoScale
        - Licensing
        - Arctera Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
        version: 8.0.410
      entries:
      - name: infoscale-licensing-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-licensing-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-licensing-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-licensing-operator.v8.0.110
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Developer"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:e7b8d3c44df6151e61ded91336273e35eeaec04bd19124cf6313dc16d1224660
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:e7b8d3c44df6151e61ded91336273e35eeaec04bd19124cf6313dc16d1224660
        version: 8.0.110
      entries:
      - name: infoscale-licensing-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-licensing-operator.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-licensing-operator
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kedify
      provider-url: kedify.io
    name: kedify-agent-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kedify-agent-helm-operator.v0.2.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kedify.io/v1alpha1",
                "kind": "KedifyAgent",
                "metadata": {
                  "name": "kedifyagent-sample"
                },
                "spec": {
                  "agent": {
                    "agentId": "",
                    "apiKey": "kfy_xxxxxxxxxxxxxxxxxxxxx",
                    "createCrds": true,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v0.2.16@sha256:bfcc0b24ff75de17464ec213406f48bc211a932308e888c450ed23ee1786d37a"
                    },
                    "kedifyProxy": {
                      "globalValues": {
                        "openshift": true
                      },
                      "image": {
                        "tag": "v1.34.1@sha256:683e2f32809657c8d4a1508648545481130ac5409cd1ce63ccc795ad8c71202e"
                      }
                    },
                    "kubectlImage": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3"
                    },
                    "orgId": "00000000-0000-0000-0000-000000000000",
                    "securityContext": {
                      "readOnlyRootFilesystem": true
                    }
                  },
                  "keda": {
                    "enabled": true,
                    "image": {
                      "keda": {
                        "tag": "v2.17.2-0@sha256:802592b691a7b81a249055fa117994927c2fd376b5023a3888bcc7973c65ba52"
                      },
                      "metricsApiServer": {
                        "tag": "v2.17.2-0@sha256:df44bc1de75324158017d623a9d0adf3e3e9f1d64db6afa7133f777adc7b6389"
                      },
                      "pullPolicy": "IfNotPresent",
                      "webhooks": {
                        "tag": "v2.17.2-0@sha256:73e180bd5793dec3785fcb9ecd7e7184e645f7c37d561a5728970e5303c19a41"
                      }
                    }
                  },
                  "keda-add-ons-http": {
                    "enabled": true,
                    "interceptor": {
                      "pullPolicy": "IfNotPresent",
                      "scaledObject": {
                        "waitForCrd": true
                      }
                    },
                    "kubectlImage": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3"
                    },
                    "pinnedImages": {
                      "interceptor": {
                        "tag": "v0.10.0-17@sha256:30ccf20e79c6a62c14781872a73ad0a811063add7136b77df986bf403d6ca6bb"
                      },
                      "scaler": {
                        "tag": "v0.10.0-17@sha256:5863ee0ca09efeca225966c2eb8781f90f2dfb4aed872d84a6075b711ab042b2"
                      }
                    },
                    "scaler": {
                      "pullPolicy": "IfNotPresent"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: '{"kedify-openshift-default":{"description":"Default
            Kedify Agent CR for OpenShift. Register at https://dashboard.kedify.io
            to get started."}}'
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:d61332d50587ce2c75aff29033fdd5cb1bf0ec49d199ab72323a27d9d19ff402
          createdAt: "2025-08-07T14:04:11Z"
          description: Kedify Autoscaler, powered by proven open source technologies
            and KEDA, is an enterprise-grade service that allows teams to autoscale
            any workload using various event sources including HTTP events
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: keda
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: http://github.com/kedify/kedify-agent-helm-operator
          support: support@kedify.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KedifyAgent
            name: kedifyagents.charts.kedify.io
            version: v1alpha1
        description: Kedify Autoscaler, powered by proven open source technologies
          and KEDA, is an enterprise-grade service that allows teams to autoscale
          any workload using various event sources including HTTP events
        displayName: kedify-agent-helm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - autoscaling
        - keda
        - kedify
        links:
        - name: Kedify Homepage
          url: https://kedify.io
        maintainers:
        - email: support@kedify.io
          name: Kedify
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Kedify
          url: kedify.io
        relatedImages:
        - ghcr.io/kedify/agent:v0.2.16@sha256:9a57d5eafe368aab1d94f0eec3101e02f0ff2e356d8499984030ab89161e42ae
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:d61332d50587ce2c75aff29033fdd5cb1bf0ec49d199ab72323a27d9d19ff402
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:8e3a458912ec6540cb414446425ad7b1546528f04a5d84fcd64d8ee5424392e6
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:fc91d1f7f8b632614f31dea14c0c8d4a5f8311a83fdef82303857aa8b43b5782
        - envoyproxy/envoy:v1.34.1@sha256:007da57c2c328a90bd4e6d99b70bd899132f1b4a9426ccafe25437cf84a60c14
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:4b7277ad53dfb01bf5f2201dd31d94f24519da00d008638c3c36026b96ea78b4
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:9300a5382373b179f7f541a663cd30e50284e54f352993e8cca34460acbd890d
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:c1c093a6fda7efc9bc30f110c55f8bb075250db929dbdb2534b9071efbfc3169
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:30ccf20e79c6a62c14781872a73ad0a811063add7136b77df986bf403d6ca6bb
        - envoyproxy/envoy:v1.34.1@sha256:683e2f32809657c8d4a1508648545481130ac5409cd1ce63ccc795ad8c71202e
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:a38fdaf3a49eb4c71c4f3ab1211befa22fa5de56391f88efc11335f0bb3b1d29
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:82b041027dc900fdab717e8db6a522139067801c04cd1ebfe950572e714938ab
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:720b90a4f112a91a94b1bdafe445df1681657ddaa4fb20aa6d2904a0eb664a2a
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:5863ee0ca09efeca225966c2eb8781f90f2dfb4aed872d84a6075b711ab042b2
        - envoyproxy/envoy:v1.34.1@sha256:75545b91d050aa51b52cb0673be172f2f263c971dae3d2a82a70d128d92e87c2
        - ghcr.io/kedify/agent:v0.2.16@sha256:bfcc0b24ff75de17464ec213406f48bc211a932308e888c450ed23ee1786d37a
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:df44bc1de75324158017d623a9d0adf3e3e9f1d64db6afa7133f777adc7b6389
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:73e180bd5793dec3785fcb9ecd7e7184e645f7c37d561a5728970e5303c19a41
        - ghcr.io/kedify/agent:v0.2.16@sha256:c59d765091a31659236b8d80f4033912f5cf47d86c2e2da96bcfa133b9f6fe25
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:33cd5a8c6dd5e90614341dc1863e1f041f2c2475bd61edf5176a2a0d98ea54f9
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:850922994153104a8e382416561c251b46a95ff2b2a216aacbd7fcb4be7f6382
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:802592b691a7b81a249055fa117994927c2fd376b5023a3888bcc7973c65ba52
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:4b63409c022ba137760e7a534b6308fc1e3639f06fdb53ee1f6df32787b68400
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:ea6b22bb12c8ff7cbab4633c2f4a0e49eaf3883b07ab1f47dbebff2a3e07f029
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:9fcc76f983035eb0ceff7beb941bf32fe7fc28b341ec1aac5191ffc1471f401b
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:cb40290842976fd09f2008d8e2e64ebf90a215f0584656d4886122d322006873
        version: 0.2.16
      entries:
      - name: kedify-agent-helm-operator.v0.2.16
        version: 0.2.16
      name: stable
    defaultChannel: stable
    packageName: kedify-agent-helm-operator
    provider:
      name: Kedify
      url: kedify.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-operator.v2024.10.2-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "24.10.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "false",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": null
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": null
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:f4b9eb4f209f5428b0940056c760efcd9939553f3baac29924d1554701bc2dce
          createdAt: "2024-10-25T01:46:32Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2024.10.2
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2024.3.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "99.99.0.nextgen_uma_beta,3"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:f4b9eb4f209f5428b0940056c760efcd9939553f3baac29924d1554701bc2dce
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:7c4868db021beaffa762eb0de43739bbc865a4bf5900f1d17f78caed8612c073
        version: 2024.10.2-12
      entries:
      - name: uma-operator.v2024.10.2-12
        version: 2024.10.2-12
      name: "2024.10"
    - currentCSV: uma-operator.v2025.1.3-33
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.1.1.33,990033",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "false",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": null
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": null
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperDataQueueSize": 300,
                            "scrapperProcessingRoutines": 50,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:4a1bc124f22cf05dc6b42edbcc971d6309862e7be9c39b2ab6b1e3d50de61f3f
          createdAt: "2025-02-18T05:33:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.1.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2024.3.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       scrapperProcessingRoutines: 50
                       scrapperDataQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.1.1.26,990026"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:4a1bc124f22cf05dc6b42edbcc971d6309862e7be9c39b2ab6b1e3d50de61f3f
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:ae91a28d57c8a359f81757e4bf79472f2ac189ac0ae3ee0d87e8ff9c433f1ccc
        version: 2025.1.3-33
      entries:
      - name: uma-operator.v2025.1.3-33
        version: 2025.1.3-33
      name: "2025.1"
    - currentCSV: uma-operator.v2025.2.1-25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.2.1.25,990025",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mixedMode": "false",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperDataQueueSize": 300,
                            "scrapperProcessingRoutines": 50,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:d8f4ca750252f1570a2dd89cabf018cc4f0b2c0cd6cadfedbf48507a34d4aee9
          createdAt: "2025-03-05T05:49:42Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.2.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2024.3.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       scrapperProcessingRoutines: 50
                       scrapperDataQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.1.1.26,990026"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:d8f4ca750252f1570a2dd89cabf018cc4f0b2c0cd6cadfedbf48507a34d4aee9
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:8bcf0dac889f38a0bf77eb9babe3f726e3b2564d5988c05bdc5c2d91ae189c61
        version: 2025.2.1-25
      entries:
      - name: uma-operator.v2025.2.1-25
        version: 2025.2.1-25
      name: "2025.2"
    - currentCSV: uma-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:8ff4aab6d0043ccd2542a86b7e9c24d60670a3c376021ef2d02fc1a289130275
          createdAt: "2025-04-16T06:54:48Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.4.5
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.4.2.5,990005"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:8ff4aab6d0043ccd2542a86b7e9c24d60670a3c376021ef2d02fc1a289130275
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        version: 2025.4.5-5
      entries:
      - name: uma-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:90b4b21d93bce41f2530699cec01f03e5daa73bfbf4f30cbcc8fbd6e75e97da6
          createdAt: "2025-07-24T15:07:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.5.2.12,990012""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:90b4b21d93bce41f2530699cec01f03e5daa73bfbf4f30cbcc8fbd6e75e97da6
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        version: 2025.5.4-12
      entries:
      - name: uma-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-operator.v2025.7.4-23
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:35f74109aa1916524c1fb483972caa31cb84381c3cff7c3b26fd9b676f9d748c
          createdAt: "2025-08-25T09:00:19Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.3
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.7.1.23,990023"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:35f74109aa1916524c1fb483972caa31cb84381c3cff7c3b26fd9b676f9d748c
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.4-23
      entries:
      - name: uma-operator.v2025.7.4-23
        version: 2025.7.4-23
      name: "2025.7"
    - currentCSV: uma-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:adea186632253865406e95a0062abbb312dee23c92ff108f5914e325dc0b8833
          createdAt: "2025-08-15T03:19:46Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.8.1.9,990009"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:adea186632253865406e95a0062abbb312dee23c92ff108f5914e325dc0b8833
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        version: 2025.8.2-9
      entries:
      - name: uma-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
          createdAt: "2025-09-24T12:22:41Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.12,990012"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:c892787785c8c5e0de1692709a9e3f54eef36f1f00c53f1b9feabb613769a48c
          createdAt: "2025-07-30T08:06:11Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "99.99.0.serp_F152813,19"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:c892787785c8c5e0de1692709a9e3f54eef36f1f00c53f1b9feabb613769a48c
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        version: 2025.7.2-19
      entries:
      - name: uma-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:dc15dfdc8b82ee159b52036cc9d057e6525b08c478ed759939cf5c3fc7daeeec
          createdAt: "2025-09-11T11:15:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.8,990008"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:dc15dfdc8b82ee159b52036cc9d057e6525b08c478ed759939cf5c3fc7daeeec
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
          createdAt: "2025-09-24T12:22:41Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.12,990012"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: EDB
      provider-url: ""
    name: cloud-native-postgresql
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloud-native-postgresql.v1.27.0-rc1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:5f5bebbf4d8021dafd0fb7485d629b30aaa362e899e69eb52fa5e74757af6e36
          createdAt: "2025-07-30T09:28:05Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:5f5bebbf4d8021dafd0fb7485d629b30aaa362e899e69eb52fa5e74757af6e36
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        version: 1.27.0-rc1
      entries:
      - name: cloud-native-postgresql.v1.27.0-rc1
        version: 1.27.0-rc1
      - name: cloud-native-postgresql.v1.26.0-rc3
        version: 1.26.0-rc3
      - name: cloud-native-postgresql.v1.26.0-rc2
        version: 1.26.0-rc2
      - name: cloud-native-postgresql.v1.26.0-rc1
        version: 1.26.0-rc1
      name: candidate
    - currentCSV: cloud-native-postgresql.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
          createdAt: "2025-08-19T12:00:36Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
        version: 1.27.0
      entries:
      - name: cloud-native-postgresql.v1.27.0
        version: 1.27.0
      - name: cloud-native-postgresql.v1.26.1
        version: 1.26.1
      - name: cloud-native-postgresql.v1.26.0
        version: 1.26.0
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      - name: cloud-native-postgresql.v1.24.1
        version: 1.24.1
      - name: cloud-native-postgresql.v1.24.0
        version: 1.24.0
      - name: cloud-native-postgresql.v1.23.3
        version: 1.23.3
      - name: cloud-native-postgresql.v1.23.2
        version: 1.23.2
      - name: cloud-native-postgresql.v1.23.1
        version: 1.23.1
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      - name: cloud-native-postgresql.v1.21.1
        version: 1.21.1
      - name: cloud-native-postgresql.v1.21.0
        version: 1.21.0
      - name: cloud-native-postgresql.v1.20.2
        version: 1.20.2
      - name: cloud-native-postgresql.v1.20.1
        version: 1.20.1
      - name: cloud-native-postgresql.v1.20.0
        version: 1.20.0
      - name: cloud-native-postgresql.v1.19.1
        version: 1.19.1
      - name: cloud-native-postgresql.v1.19.0
        version: 1.19.0
      name: fast
    - currentCSV: cloud-native-postgresql.v1.25.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
          createdAt: "2025-07-26T06:26:53Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.25.3
      entries:
      - name: cloud-native-postgresql.v1.25.3
        version: 1.25.3
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      - name: cloud-native-postgresql.v1.22.9
        version: 1.22.9
      - name: cloud-native-postgresql.v1.22.8
        version: 1.22.8
      - name: cloud-native-postgresql.v1.22.7
        version: 1.22.7
      - name: cloud-native-postgresql.v1.22.6
        version: 1.22.6
      - name: cloud-native-postgresql.v1.22.5
        version: 1.22.5
      - name: cloud-native-postgresql.v1.22.4
        version: 1.22.4
      - name: cloud-native-postgresql.v1.22.3
        version: 1.22.3
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      - name: cloud-native-postgresql.v1.18.13
        version: 1.18.13
      - name: cloud-native-postgresql.v1.18.12
        version: 1.18.12
      - name: cloud-native-postgresql.v1.18.11
        version: 1.18.11
      - name: cloud-native-postgresql.v1.18.10
        version: 1.18.10
      - name: cloud-native-postgresql.v1.18.9
        version: 1.18.9
      - name: cloud-native-postgresql.v1.18.8
        version: 1.18.8
      - name: cloud-native-postgresql.v1.18.7
        version: 1.18.7
      name: stable
    - currentCSV: cloud-native-postgresql.v1.18.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:69e439a2a961881309cc122174eaca273cf6d7873e9e871efbdb2d60804f40e4
          createdAt: "2024-06-13T16:08:55Z"
          description: EDB Postgres for Kubernetes is an open source operator designed
            to manage highly available PostgreSQL databases with a primary/standby
            architecture on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 1.18.0 < 1.18.13'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloud
        - native
        - edb
        - enterprisedb
        - cnp
        - pg4k
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: john.long@enterprisedb.com
          name: John Long
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: The EDB Postgres for Kubernetes Contributors
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:69e439a2a961881309cc122174eaca273cf6d7873e9e871efbdb2d60804f40e4
        version: 1.18.13
      entries:
      - name: cloud-native-postgresql.v1.18.13
        version: 1.18.13
      - name: cloud-native-postgresql.v1.18.12
        version: 1.18.12
      - name: cloud-native-postgresql.v1.18.11
        version: 1.18.11
      - name: cloud-native-postgresql.v1.18.10
        version: 1.18.10
      - name: cloud-native-postgresql.v1.18.9
        version: 1.18.9
      - name: cloud-native-postgresql.v1.18.8
        version: 1.18.8
      - name: cloud-native-postgresql.v1.18.7
        version: 1.18.7
      name: stable-v1.18
    - currentCSV: cloud-native-postgresql.v1.19.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:2a717edd0523798bd4d0b14c557b8d6b19ae3ad1f019c0cc4f61db49f10c3dad
          createdAt: "2023-11-08T19:02:47Z"
          description: Operator to manage Postgres high availability clusters with
            a primary/standby architecture.
          olm.skipRange: '>=1.15.3 < 1.19.6'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup manager
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster manager
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PgBouncer Pooler Manager
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL backups schedule
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: "EDB Postgres for Kubernetes is an operator designed, developed,
          and supported by EDB that covers the full \nlifecycle of a highly available
          Postgres database clusters with a primary/standby architecture, using native\nstreaming
          replication. It is based on the open source CloudNativePG operator, and
          provides additional value\nsuch as compatibility with Oracle using EDB Postgres
          Advanced Server and additional supported platforms such\nas IBM Power and
          OpenShift.\n\nKey features available include:\n\n* Kubernetes API integration
          for high availability\n* Self-healing through failover and automated recreation
          of replicas\n* Capacity management with scale up/down capabilities\n* Planned
          switchovers for scheduled maintenance\n* Read-only and read-write Kubernetes
          services definitions\n* Rolling updates for Postgres minor versions and
          operator upgrades\n* Continuous backup and point-in-time recovery\n* Connection
          Pooling with PgBouncer\n* Integrated metrics exporter out of the box\n*
          PostgreSQL replication across multiple Kubernetes clusters\n* Red Hat certified
          operator for OpenShift\n\nThe operator has been renamed from Cloud Native
          PostgreSQL. Existing users of Cloud Native PostgreSQL will not\nexperience
          any change, as the underlying components and resources have not changed.\n"
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloud
        - native
        - edb
        - enterprisedb
        - cnp
        - pg4k
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: sales@enterprisedb.com
          name: Sales
        - email: support@enterprisedb.com
          name: Support
        maturity: stable
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:2a717edd0523798bd4d0b14c557b8d6b19ae3ad1f019c0cc4f61db49f10c3dad
        version: 1.19.6
      entries:
      - name: cloud-native-postgresql.v1.19.6
        version: 1.19.6
      - name: cloud-native-postgresql.v1.19.5
        version: 1.19.5
      - name: cloud-native-postgresql.v1.19.4
        version: 1.19.4
      - name: cloud-native-postgresql.v1.19.3
        version: 1.19.3
      - name: cloud-native-postgresql.v1.19.2
        version: 1.19.2
      - name: cloud-native-postgresql.v1.19.1
        version: 1.19.1
      - name: cloud-native-postgresql.v1.19.0
        version: 1.19.0
      name: stable-v1.19
    - currentCSV: cloud-native-postgresql.v1.20.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:919c6669f8bfa333b19fdd1ca03aaeceebe0c5ca74d91cb66b2369680422e916
          createdAt: "2024-02-03T17:26:23Z"
          description: Operator to manage Postgres high availability clusters with
            a primary/standby architecture.
          olm.skipRange: '>=1.15.3 < 1.20.6'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup manager
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster manager
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PgBouncer Pooler Manager
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL backups schedule
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: "EDB Postgres for Kubernetes is an operator designed, developed,
          and supported by EDB that covers the full \nlifecycle of a highly available
          Postgres database clusters with a primary/standby architecture, using native\nstreaming
          replication. It is based on the open source CloudNativePG operator, and
          provides additional value\nsuch as compatibility with Oracle using EDB Postgres
          Advanced Server and additional supported platforms such\nas IBM Power and
          OpenShift.\n\nKey features available include:\n\n* Kubernetes API integration
          for high availability\n* Self-healing through failover and automated recreation
          of replicas\n* Capacity management with scale up/down capabilities\n* Planned
          switchovers for scheduled maintenance\n* Read-only and read-write Kubernetes
          services definitions\n* Rolling updates for Postgres minor versions and
          operator upgrades\n* Continuous backup and point-in-time recovery\n* Connection
          Pooling with PgBouncer\n* Integrated metrics exporter out of the box\n*
          PostgreSQL replication across multiple Kubernetes clusters\n* Red Hat certified
          operator for OpenShift\n\nThe operator has been renamed from Cloud Native
          PostgreSQL. Existing users of Cloud Native PostgreSQL will not\nexperience
          any change, as the underlying components and resources have not changed.\n"
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloud
        - native
        - edb
        - enterprisedb
        - cnp
        - pg4k
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: sales@enterprisedb.com
          name: Sales
        - email: support@enterprisedb.com
          name: Support
        maturity: stable
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:919c6669f8bfa333b19fdd1ca03aaeceebe0c5ca74d91cb66b2369680422e916
        version: 1.20.6
      entries:
      - name: cloud-native-postgresql.v1.20.6
        version: 1.20.6
      - name: cloud-native-postgresql.v1.20.5
        version: 1.20.5
      - name: cloud-native-postgresql.v1.20.4
        version: 1.20.4
      - name: cloud-native-postgresql.v1.20.3
        version: 1.20.3
      - name: cloud-native-postgresql.v1.20.2
        version: 1.20.2
      - name: cloud-native-postgresql.v1.20.1
        version: 1.20.1
      - name: cloud-native-postgresql.v1.20.0
        version: 1.20.0
      name: stable-v1.20
    - currentCSV: cloud-native-postgresql.v1.21.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:8cc328b1fc8efa3911d3dba689bc824ff789b305f7eb3284bf21ffcdc52a9683
          createdAt: "2024-06-13T16:41:26Z"
          description: EDB Postgres for Kubernetes is an open source operator designed
            to manage highly available PostgreSQL databases with a primary/standby
            architecture on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 1.18.0 < 1.21.6'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: john.long@enterprisedb.com
          name: John Long
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: The EDB Postgres for Kubernetes Contributors
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:8cc328b1fc8efa3911d3dba689bc824ff789b305f7eb3284bf21ffcdc52a9683
        version: 1.21.6
      entries:
      - name: cloud-native-postgresql.v1.21.6
        version: 1.21.6
      - name: cloud-native-postgresql.v1.21.5
        version: 1.21.5
      - name: cloud-native-postgresql.v1.21.4
        version: 1.21.4
      - name: cloud-native-postgresql.v1.21.3
        version: 1.21.3
      - name: cloud-native-postgresql.v1.21.2
        version: 1.21.2
      - name: cloud-native-postgresql.v1.21.1
        version: 1.21.1
      - name: cloud-native-postgresql.v1.21.0
        version: 1.21.0
      name: stable-v1.21
    - currentCSV: cloud-native-postgresql.v1.22.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:4d55ad54cc7e7807bc99f5a6f96bc92e88b4060f424ffd2cb6552be9505f38f4
          createdAt: "2025-07-25T19:02:56Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images. Only
              available in 1.25 or later
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster.
              Only available in 1.25 or later
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster. Only available in 1.25 or later
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication.
              Only available in 1.25 or later
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:4d55ad54cc7e7807bc99f5a6f96bc92e88b4060f424ffd2cb6552be9505f38f4
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.22.11
      entries:
      - name: cloud-native-postgresql.v1.22.11
        version: 1.22.11
      - name: cloud-native-postgresql.v1.22.10
        version: 1.22.10
      - name: cloud-native-postgresql.v1.22.9
        version: 1.22.9
      - name: cloud-native-postgresql.v1.22.8
        version: 1.22.8
      - name: cloud-native-postgresql.v1.22.7
        version: 1.22.7
      - name: cloud-native-postgresql.v1.22.6
        version: 1.22.6
      - name: cloud-native-postgresql.v1.22.5
        version: 1.22.5
      - name: cloud-native-postgresql.v1.22.4
        version: 1.22.4
      - name: cloud-native-postgresql.v1.22.3
        version: 1.22.3
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      name: stable-v1.22
    - currentCSV: cloud-native-postgresql.v1.23.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:934e01ed5cd35cf681414b9136a337514a1b73c0110871f1c78ba98447badb0f
          createdAt: "2024-12-23T16:37:54Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 1.18.0 < 1.23.6'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:934e01ed5cd35cf681414b9136a337514a1b73c0110871f1c78ba98447badb0f
        version: 1.23.6
      entries:
      - name: cloud-native-postgresql.v1.23.6
        version: 1.23.6
      - name: cloud-native-postgresql.v1.23.5
        version: 1.23.5
      - name: cloud-native-postgresql.v1.23.4
        version: 1.23.4
      - name: cloud-native-postgresql.v1.23.3
        version: 1.23.3
      - name: cloud-native-postgresql.v1.23.2
        version: 1.23.2
      - name: cloud-native-postgresql.v1.23.1
        version: 1.23.1
      name: stable-v1.23
    - currentCSV: cloud-native-postgresql.v1.24.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:a1acaaf5a413db0f33f3824a15aefe598974d1593172bcbccd0fc3d39d7b46d0
          createdAt: "2025-05-23T15:20:37Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:a1acaaf5a413db0f33f3824a15aefe598974d1593172bcbccd0fc3d39d7b46d0
        version: 1.24.4
      entries:
      - name: cloud-native-postgresql.v1.24.4
        version: 1.24.4
      - name: cloud-native-postgresql.v1.24.3
        version: 1.24.3
      - name: cloud-native-postgresql.v1.24.2
        version: 1.24.2
      - name: cloud-native-postgresql.v1.24.1
        version: 1.24.1
      - name: cloud-native-postgresql.v1.24.0
        version: 1.24.0
      name: stable-v1.24
    - currentCSV: cloud-native-postgresql.v1.25.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
          createdAt: "2025-07-26T06:26:53Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
        version: 1.25.3
      entries:
      - name: cloud-native-postgresql.v1.25.3
        version: 1.25.3
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      name: stable-v1.25
    - currentCSV: cloud-native-postgresql.v1.26.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:47d70233908ea855ddab3ff79c7a226719c3f2be5b8a5018c7934cd44bf0f546
          createdAt: "2025-07-26T07:59:14Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:47d70233908ea855ddab3ff79c7a226719c3f2be5b8a5018c7934cd44bf0f546
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.26.1
      entries:
      - name: cloud-native-postgresql.v1.26.1
        version: 1.26.1
      - name: cloud-native-postgresql.v1.26.0
        version: 1.26.0
      name: stable-v1.26
    - currentCSV: cloud-native-postgresql.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
          createdAt: "2025-08-19T12:00:36Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.27.0
      entries:
      - name: cloud-native-postgresql.v1.27.0
        version: 1.27.0
      name: stable-v1.27
    defaultChannel: fast
    packageName: cloud-native-postgresql
    provider:
      name: EDB
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: http://marketplace.redhat.com
    name: marketplace-games-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marketplace-games-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "game.marketplace.redhat.com/v1",
                "kind": "Arcade",
                "metadata": {
                  "name": "arcade-sample"
                },
                "spec": {
                  "size": 1
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/marketplace-games-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/marketplace-games-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Arcade is the Schema for the arcades API
            displayName: Arcade Instance
            kind: Arcade
            name: arcades.game.marketplace.redhat.com
            version: v1
        description: RHM Arcade Operator - a simple operator used to showcase how
          to build an operator around an existing web application that can can run
          on Openshift, Vanilla k8s, or on-boarded to Red Hat Marketplace.
        displayName: Arcade Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - rhm
        - red hat
        - marketplace
        - games
        - arcade
        links:
        - name: Marketplace Games Operator
          url: https://github.com/redhat-marketplace/marketplace-games-operator
        maturity: alpha
        provider:
          name: Red Hat Marketplace
          url: http://marketplace.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/rhm-labs/arcade-operator@sha256:d6959be6693728d2dc4cd38b69251551d23c5002cb52db2684df31a0a8f7b0ab
        version: 1.0.0
      entries:
      - name: marketplace-games-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: marketplace-games-operator-rhmp
    provider:
      name: Red Hat Marketplace
      url: http://marketplace.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/
    name: varnish-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: varnish-operator.v0.27.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "caching.ibm.com/v1alpha1",
                "kind": "VarnishCluster",
                "metadata": {
                  "name": "varnishcluster-sample"
                },
                "spec": {
                  "backend": {
                    "port": "web",
                    "selector": {
                      "app": "nginx"
                    }
                  },
                  "replicas": 1,
                  "service": {
                    "port": 80
                  },
                  "varnish": {
                    "args": [
                      "-p",
                      "default_ttl=3600",
                      "-p",
                      "default_grace=3600"
                    ]
                  },
                  "vcl": {
                    "configMapName": "vcl-files",
                    "entrypointFileName": "entrypoint.vcl"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: ibmcom/varnish-operator:0.27.2
          createdAt: 2021-02-15T11:15:31+0000
          description: Run and manage Varnish clusters on Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/varnish-operator
          support: https://ibm.github.io/varnish-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VarnishCluster is the Schema for the varnishclusters API
            displayName: Varnish Cluster
            kind: VarnishCluster
            name: varnishclusters.caching.ibm.com
            version: v1alpha1
        description: |
          Run and manage Varnish clusters on Kubernetes

          The operator manages the whole lifecycle of the cluster: creating, deleting and keeping the cluster configuration up to date.

          The operator is responsible for building the VCL configuration using templates defined by the users and keeping the configuration up to date when relevant events occur (backend pod failure, scaling of the deployment, VCL configuration change).

          Example of a simple `VarnishCluster`:

          ```yaml
          apiVersion: caching.ibm.com/v1alpha1
          kind: VarnishCluster
          metadata:
            name: varnishcluster-sample
            namespace: varnish-ns
          spec:
            vcl:
              configMapName: vcl-files
              entrypointFileName: entrypoint.vcl
            replicas: 3
            backend:
              selector:
                app: nginx
            service:
              port: 80
          ```

          Full documentation can be found [here](https://ibm.github.io/varnish-operator/)
        displayName: Varnish Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - varnish
        - cache
        links:
        - name: Varnish Operator Docs
          url: https://ibm.github.io/varnish-operator
        - name: Varnish Operator Source Code
          url: https://github.com/IBM/varnish-operator
        maintainers:
        - email: tomash.sidei@ibm.com
          name: Tomash Sidei
        - email: cingram@us.ibm.com
          name: Craig Ingram
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/
        relatedImages:
        - ibmcom/varnish-operator:0.27.2
        version: 0.27.2
      entries:
      - name: varnish-operator.v0.27.2
        version: 0.27.2
      name: alpha
    defaultChannel: alpha
    packageName: varnish-operator
    provider:
      name: IBM
      url: https://www.ibm.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CloudBees
      provider-url: https://www.cloudbees.com/products/continuous-integration
    name: cloudbees-ci
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudbees-ci.v3.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.cloudbees.com/v1alpha1",
                "kind": "CloudBeesCI",
                "metadata": {
                  "name": "cloudbeesci-sample"
                },
                "spec": {
                  "Agents": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-agent:2.289.3.2"
                    },
                    "ImagePullSecrets": null,
                    "SeparateNamespace": {
                      "Create": false,
                      "Enabled": false,
                      "Name": null
                    }
                  },
                  "Hibernation": {
                    "Enabled": false,
                    "Image": {
                      "dockerImage": "cloudbees/managed-master-hibernation-monitor:247.c5dfce00a179",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "NodeSelector": {},
                    "Tolerations": []
                  },
                  "Master": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-mm:2.289.3.2"
                    },
                    "JavaOpts": null,
                    "OperationsCenterNamespace": null
                  },
                  "NetworkPolicy": {
                    "Enabled": false,
                    "JMXSelectors": [],
                    "ingressControllerSelector": []
                  },
                  "OperationsCenter": {
                    "AgentListenerPort": 50000,
                    "Annotations": {},
                    "CSRF": {
                      "ProxyCompatibility": false
                    },
                    "CasC": {
                      "ConfigMapName": "oc-casc-bundle",
                      "Enabled": false
                    },
                    "ContainerPort": 8080,
                    "ContextPath": null,
                    "Enabled": true,
                    "ExtraConfigMaps": [],
                    "ExtraContainers": [],
                    "ExtraGroovyConfiguration": {},
                    "ExtraVolumeMounts": [],
                    "ExtraVolumes": [],
                    "HealthProbeLivenessFailureThreshold": 12,
                    "HealthProbes": true,
                    "HostName": null,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-cloud-core-oc:2.289.3.2",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "Ingress": {
                      "Annotations": {
                        "kubernetes.io/tls-acme": "false"
                      },
                      "Class": "nginx",
                      "tls": {
                        "Enable": false,
                        "SecretName": null
                      }
                    },
                    "JavaOpts": null,
                    "JenkinsOpts": null,
                    "LoadBalancerIP": null,
                    "LoadBalancerSourceRanges": [
                      "0.0.0.0/0"
                    ],
                    "Name": "cjoc",
                    "NodeSelector": {},
                    "Platform": "standard",
                    "Protocol": "http",
                    "Resources": {
                      "Limits": {
                        "Cpu": 1,
                        "Memory": "2G"
                      },
                      "Requests": {
                        "Cpu": 1,
                        "Memory": "2G"
                      }
                    },
                    "Route": {
                      "tls": {
                        "CACertificate": null,
                        "Certificate": null,
                        "DestinationCACertificate": null,
                        "Enable": false,
                        "InsecureEdgeTerminationPolicy": "Redirect",
                        "Key": null,
                        "Termination": "edge"
                      }
                    },
                    "ServiceAgentListenerPort": 50000,
                    "ServiceAnnotations": {},
                    "ServicePort": 80,
                    "ServiceType": "ClusterIP",
                    "Tolerations": []
                  },
                  "Persistence": {
                    "AccessMode": "ReadWriteOnce",
                    "Annotations": {},
                    "Size": "20Gi",
                    "StorageClass": null
                  },
                  "PodSecurityPolicy": {
                    "Annotations": {},
                    "Enabled": false
                  },
                  "Subdomain": false,
                  "ingress-nginx": {
                    "Enabled": false,
                    "controller": {
                      "admissionWebhooks": {
                        "patch": {
                          "nodeSelector": {
                            "kubernetes.io/os": "linux"
                          }
                        }
                      },
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "nginx-ingress": {
                    "Enabled": false,
                    "controller": {
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "rbac": {
                    "agentsServiceAccountName": "jenkins-agents",
                    "hibernationMonitorServiceAccountName": "managed-master-hibernation-monitor",
                    "install": true,
                    "masterServiceAccountName": "jenkins",
                    "serviceAccountName": "cjoc"
                  },
                  "sidecarinjector": {
                    "Enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudBeesCI
            name: cloudbeescis.charts.cloudbees.com
            version: v1alpha1
        description: CloudBees CI
        displayName: CloudBees CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudbees
        - ci
        - enterprise
        - continuous
        - integration
        - jenkins
        - automation
        - pipeline
        - devops
        links:
        - name: CloudBees CI
          url: https://www.cloudbees.com/products/continuous-integration
        maintainers:
        - email: support@cloudbees.com
          name: Support
        - email: sales@cloudbees.com
          name: Sales
        maturity: alpha
        provider:
          name: CloudBees
          url: https://www.cloudbees.com/products/continuous-integration
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.connect.redhat.com/cloudbees/cloudbees-ci-operator@sha256:bd0bd72e9e7807b0fa307de8a93a56967221fe2dd181bac02383f0f9be9120e4
        version: 3.34.1
      entries:
      - name: cloudbees-ci.v3.34.1
        version: 3.34.1
      name: alpha
    defaultChannel: alpha
    packageName: cloudbees-ci
    provider:
      name: CloudBees
      url: https://www.cloudbees.com/products/continuous-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: https://github.com/Mellanox/network-operator/
    name: nvidia-network-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nvidia-network-operator.v25.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "doca3.1.0-25.07-0.9.7.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "nvcr.io/nvidia/mellanox",
                    "version": "sha256:a87096761d155eeb6f470e042d2d167bb466d57e63b4aba957f57d745e15a9b2"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:51270965284cceeda659fb1098e70a888b39664ceade7e9d455fb95ac85914b7
          createdAt: "2025-08-21T19:12:20Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/mellanox/nvidia-k8s-ipam@sha256:3f692273c1598525b9c698c9d9cd463e5ddc55b638c6c031639f38786fc67475
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:9a831bfdf85f313b1f5749b7c9b2673bb8fff18b4ff768c9242dabaa4468e449
        - nvcr.io/nvidia/mellanox/k8s-rdma-shared-dev-plugin@sha256:a87096761d155eeb6f470e042d2d167bb466d57e63b4aba957f57d745e15a9b2
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:5ad37cd5ae96153e6f87670f437ff5f6e44febe7bb6eaeb1a0f219e28284e53d
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:d41eea65b102dc5bb5792871ae212b59a65dae492ca109c52138d5626eba1ffd
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:13c2a2d055e501c03c6451413b85476ceafbc2fbffc7443952294c196f3f89f3
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:30354799264480e1aee057397eef2c52b277bbc6872184e2b7b811f4152b3a67
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:423aea08b4a0f01e2e19b6a709e14347bd08f8274f40002e0224090dfa4904fc
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:f9d4b8358c68ba28f849ca710bf769c4327a639733c0ff30282ca33a744078f6
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:51270965284cceeda659fb1098e70a888b39664ceade7e9d455fb95ac85914b7
        - nvcr.io/nvidia/mellanox/nic-feature-discovery@sha256:1330e3a7ea2491bc310bfc4766769389ceb56d050831bf9c145b7bfb6aba36c4
        - nvcr.io/nvidia/mellanox/network-operator-init-container@sha256:02ee60ac6ee163acc97706bd4229d268f608a95bcbcb6fccee90ceb1d8a6c62b
        - nvcr.io/nvidia/mellanox/ipoib-cni@sha256:7564ad1142c863e61f1dd84f5714deba9a7c6f6d85efb383ac3a2566eb4f4b71
        - nvcr.io/nvidia/mellanox/sriov-network-device-plugin@sha256:7508a160f3f124e0b78d2b4d5a52deca6b9f56e226fee957c4d156e425eb101a
        - nvcr.io/nvidia/mellanox/nic-configuration-operator-daemon@sha256:b78c15904cecd343fa48374757b3619b5cbdc22059d5ee9731da87e35b94a2e2
        - nvcr.io/nvidia/mellanox/nic-configuration-operator@sha256:03f34f7fde3f65ca7b214b66c6fdbc1250d4f3863f9c02b0fba78e8aa84efca3
        - nvcr.io/nvidia/mellanox/ib-kubernetes@sha256:1fa251979b5f5201af50bc7aa6936829103d742f42bfbf752f20852f5fe44dfb
        version: 25.7.0
      entries:
      - name: nvidia-network-operator.v25.7.0
        version: 25.7.0
      - name: nvidia-network-operator.v25.4.0
        version: 25.4.0
      name: stable
    - currentCSV: nvidia-network-operator.v24.10.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "24.10-0.7.0.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "v1.5.2"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:23aeb481d662496210d1c13083c65cfa4ffdcf67f973e5f884b1b936f693c0f5
          createdAt: "2025-01-27T08:16:45Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - ghcr.io/mellanox/ib-kubernetes@sha256:100e6addf99d5b44e81cb3aa8d2fc892fced1de3183190931961b92ee924207b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:e36c46cf2c813c081bf7cc22b47fb4cf391a3d91804e47a6ce87f31d62f42535
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:4bb13a8b850c91011013509334d5ae4f4b7aa1d1b0c12958625e161cd50309d6
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:9f468fdc4449e65e4772575f83aa85840a00f97165f9a00ba34695c91d610fbd
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:5a87d04872d5d4530cab3cbb3db528722fa24435f3375debd8ed102b04e631c2
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:23aeb481d662496210d1c13083c65cfa4ffdcf67f973e5f884b1b936f693c0f5
        - ghcr.io/mellanox/ipoib-cni@sha256:81646c07e27ce1df2603050a75dd267f264685fdc64c285eca3925ae98ed5392
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:77e3b946c20f45728dfa6dd1d115356de5ea317f91cddfd1540b7121550f66dc
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:47e1bb84ac97f9af49f97b49c2de44b4ba82b890b487880edc5c9adf86a176f7
        - ghcr.io/mellanox/network-operator-init-container@sha256:1699d23027ea30c9fa59575a914114bdfd5a87a359caf8c0a9b16d409ec0d068
        - ghcr.io/mellanox/nic-feature-discovery@sha256:92dda9434519de39be0f94c80787766e3399e8b238d14839360aff4d6232a09b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:db5797121735eeb935d8f11e305b50c80ba07aa83c291d76d05dea7e303bc740
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:ecb2cfdc510b7ed09a59daed4817ee353e83a94d2d8581c3c052f4fa4b990574
        version: 24.10.1
      entries:
      - name: nvidia-network-operator.v24.10.1
        version: 24.10.1
      name: v24.10
    - currentCSV: nvidia-network-operator.v24.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "24.10-0.7.0.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "v1.5.2"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:f258e8665d094c09159654759bc9ae41fbe0f7fffe95139ec02964759c3d8a87
          createdAt: "2024-11-26T09:58:46Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:5a87d04872d5d4530cab3cbb3db528722fa24435f3375debd8ed102b04e631c2
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:ecb2cfdc510b7ed09a59daed4817ee353e83a94d2d8581c3c052f4fa4b990574
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:9f468fdc4449e65e4772575f83aa85840a00f97165f9a00ba34695c91d610fbd
        - ghcr.io/mellanox/nic-feature-discovery@sha256:92dda9434519de39be0f94c80787766e3399e8b238d14839360aff4d6232a09b
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:f258e8665d094c09159654759bc9ae41fbe0f7fffe95139ec02964759c3d8a87
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:db5797121735eeb935d8f11e305b50c80ba07aa83c291d76d05dea7e303bc740
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:e36c46cf2c813c081bf7cc22b47fb4cf391a3d91804e47a6ce87f31d62f42535
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:4bb13a8b850c91011013509334d5ae4f4b7aa1d1b0c12958625e161cd50309d6
        - ghcr.io/mellanox/ib-kubernetes@sha256:100e6addf99d5b44e81cb3aa8d2fc892fced1de3183190931961b92ee924207b
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:77e3b946c20f45728dfa6dd1d115356de5ea317f91cddfd1540b7121550f66dc
        - ghcr.io/mellanox/ipoib-cni@sha256:81646c07e27ce1df2603050a75dd267f264685fdc64c285eca3925ae98ed5392
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:47e1bb84ac97f9af49f97b49c2de44b4ba82b890b487880edc5c9adf86a176f7
        - ghcr.io/mellanox/network-operator-init-container@sha256:1699d23027ea30c9fa59575a914114bdfd5a87a359caf8c0a9b16d409ec0d068
        version: 24.10.0
      entries:
      - name: nvidia-network-operator.v24.10.0
        version: 24.10.0
      name: v24.10.0
    - currentCSV: nvidia-network-operator.v24.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "docaTelemetryService": {
                    "image": "doca_telemetry",
                    "repository": "nvcr.io/nvidia/doca",
                    "version": "1.16.5-doca2.6.0-host"
                  },
                  "nicFeatureDiscovery": {
                    "image": "nic-feature-discovery",
                    "repository": "ghcr.io/mellanox",
                    "version": "v0.0.1"
                  },
                  "nvIpam": {
                    "enableWebhook": false,
                    "image": "nvidia-k8s-ipam",
                    "repository": "ghcr.io/mellanox",
                    "version": "v0.2.0"
                  },
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "24.07-0.6.1.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "v1.5.1"
                  },
                  "secondaryNetwork": {
                    "ipoib": {
                      "image": "ipoib-cni",
                      "repository": "ghcr.io/mellanox",
                      "version": "v1.2.0"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:fcee44995d64058cb7bf14c02da604879443e191a382402b8a6ebfe10a3a2601
          createdAt: "2024-08-29T13:23:34Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - ghcr.io/mellanox/network-operator-init-container@sha256:1699d23027ea30c9fa59575a914114bdfd5a87a359caf8c0a9b16d409ec0d068
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:93e856dde88abb0aeb8e0b991e5f6d9f41884d8beea792f4d70cda06f18e783f
        - ghcr.io/mellanox/ib-kubernetes@sha256:b56bb88963dadaa2290a17cf0b57b86c55c09e7c075b7271c184959532555c91
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:c41eaf76c06186e1dabb0d870225d946ad3ecda2a072fa4be197dd131b7695f2
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:8324c009c3d694fe7702af36305ede78b903d839190b1e4506e82362f9aa5c02
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:fcee44995d64058cb7bf14c02da604879443e191a382402b8a6ebfe10a3a2601
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:47e1bb84ac97f9af49f97b49c2de44b4ba82b890b487880edc5c9adf86a176f7
        - ghcr.io/mellanox/nic-feature-discovery@sha256:92dda9434519de39be0f94c80787766e3399e8b238d14839360aff4d6232a09b
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - ghcr.io/mellanox/ipoib-cni@sha256:81646c07e27ce1df2603050a75dd267f264685fdc64c285eca3925ae98ed5392
        version: 24.7.0
      entries:
      - name: nvidia-network-operator.v24.7.0
        version: 24.7.0
      name: v24.7.0
    - currentCSV: nvidia-network-operator.v24.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "docaTelemetryService": {
                    "image": "doca_telemetry",
                    "repository": "nvcr.io/nvidia/doca",
                    "version": "1.16.5-doca2.6.0-host"
                  },
                  "nicFeatureDiscovery": {
                    "image": "nic-feature-discovery",
                    "repository": "ghcr.io/mellanox",
                    "version": "v0.0.1"
                  },
                  "nvIpam": {
                    "enableWebhook": false,
                    "image": "nvidia-k8s-ipam",
                    "repository": "ghcr.io/mellanox",
                    "version": "v0.2.0"
                  },
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "24.07-0.6.1.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "v1.5.1"
                  },
                  "secondaryNetwork": {
                    "ipoib": {
                      "image": "ipoib-cni",
                      "repository": "ghcr.io/mellanox",
                      "version": "v1.2.0"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:b8a07696d05fb8e5292991ba731758579313d6983ed3d75dd95e7356024d987c
          createdAt: "2024-09-04T06:38:33Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:d5a163d36ce60630e14ae8620ab437510ff0821924e043b88f0d7b67cee3bd77
        - ghcr.io/mellanox/network-operator-init-container@sha256:1699d23027ea30c9fa59575a914114bdfd5a87a359caf8c0a9b16d409ec0d068
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:403be1bcf708984f90b3ccfdf24f58594554336d69f878d3843535f4fa61ceca
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:7facd05f60e2be29925ffce9d0a6b2c712c0670c8889b815425582c46c82efdc
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:8324c009c3d694fe7702af36305ede78b903d839190b1e4506e82362f9aa5c02
        - ghcr.io/mellanox/ib-kubernetes@sha256:b56bb88963dadaa2290a17cf0b57b86c55c09e7c075b7271c184959532555c91
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:10684c0c36364b1d897a03a346abb2f691a18541ae1242b596d002f34f0d7390
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:7f87324adb351513d87fa27dfd0c9d9371ae42529d76cce8ecd68ce41c4d263f
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:676c74fcbe6d2c2341a092913b1afca65ba1a2f568ee6796b371542657e7a174
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:22e6ff1e9245f0cdf67e7904235d1a6181ee2bd3f39cedf76b39d4aeaa4a6c1d
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:b3f503b1ac39ebc55840a436973f142eb7cf67e48a070284129fd1e30390c032
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:93e856dde88abb0aeb8e0b991e5f6d9f41884d8beea792f4d70cda06f18e783f
        - ghcr.io/mellanox/ipoib-cni@sha256:81646c07e27ce1df2603050a75dd267f264685fdc64c285eca3925ae98ed5392
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:47e1bb84ac97f9af49f97b49c2de44b4ba82b890b487880edc5c9adf86a176f7
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:b8a07696d05fb8e5292991ba731758579313d6983ed3d75dd95e7356024d987c
        - ghcr.io/mellanox/nic-feature-discovery@sha256:92dda9434519de39be0f94c80787766e3399e8b238d14839360aff4d6232a09b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:c41eaf76c06186e1dabb0d870225d946ad3ecda2a072fa4be197dd131b7695f2
        version: 24.7.1
      entries:
      - name: nvidia-network-operator.v24.7.1
        version: 24.7.1
      name: v24.7.1
    - currentCSV: nvidia-network-operator.v25.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "25.01-0.6.0.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "sha256:9f468fdc4449e65e4772575f83aa85840a00f97165f9a00ba34695c91d610fbd"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:12f0576f372f5adc17b60795952ae47388e50d8fad563b9595f3091eadd8eb29
          createdAt: "2025-02-26T16:55:35Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:f5c1741cc6368fcec5b250285552d9c80ff7254414f68102cc89162e12ac0400
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:cabce074d10a0f1d62135e2cc5442d65b49094b95b8297fdd024a1a5f461319f
        - ghcr.io/mellanox/nic-feature-discovery@sha256:92dda9434519de39be0f94c80787766e3399e8b238d14839360aff4d6232a09b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:5b91974c74f08fc5b1ad2edd4ca97d81b2dc6cc897045c3eac777826fe5d40d8
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:3edf0cec91eea10df17c9f182b2db430cb383887ac144e27ac44043b308284d7
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:68b3735c4edca2d889c456beba80ac2077b0cc0edba11a32769dacd3ece64892
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:a257aae9c4e02b62f9f4ddeea66c4349c517519a544dd02fe4c8d3fee5229521
        - ghcr.io/mellanox/ib-kubernetes@sha256:100e6addf99d5b44e81cb3aa8d2fc892fced1de3183190931961b92ee924207b
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:47e1bb84ac97f9af49f97b49c2de44b4ba82b890b487880edc5c9adf86a176f7
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:7604a30cbd0927eba76ddc74ea1cfbe113563ccf068e6ab3fe78123e4deec28d
        - ghcr.io/mellanox/ipoib-cni@sha256:56f19687b0b956ba45bf079cf3a655b0d1bad0314da9d5ecfe0f4d98ff31fc86
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:12f0576f372f5adc17b60795952ae47388e50d8fad563b9595f3091eadd8eb29
        - ghcr.io/mellanox/network-operator-init-container@sha256:67e93ccf3ecb61f17597567faf0f72e1b8ddcf73c5d7440baeadcc1cb6bb811b
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:9f468fdc4449e65e4772575f83aa85840a00f97165f9a00ba34695c91d610fbd
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        version: 25.1.0
      entries:
      - name: nvidia-network-operator.v25.1.0
        version: 25.1.0
      name: v25.1
    - currentCSV: nvidia-network-operator.v25.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "25.04-0.6.1.0-2"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "sha256:1eb5ab7f4b6f0cf89c954cf7ac8eb2e6fb2290a4f5f57417136a5302f20b12c8"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:93c2063f4da6a93810784be95df0083f07618486699d93245c97a29a63861dee
          createdAt: "2025-05-28T19:40:27Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:1b20b78f889339834ed74e0da621fc5da582719b2537b36d8967ddc6a04679b8
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:56a41010b4dbee0dc31e8f9c1467a739241a2d248ed8b5054d4b418f7f79f04d
        - ghcr.io/mellanox/nic-configuration-operator-daemon@sha256:0e8478898e0b30aec13030b11081b205308666c2a9c5c5bf2c00121079901b41
        - ghcr.io/mellanox/ipoib-cni@sha256:513d88d639a01ef98b65365f711cee7fbaa166137a56245b7ab32818f745418e
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:9bc09051aca763f83cde99ee8909cf989b1d6c5d6ab683fbb1ca441d47f61fdf
        - ghcr.io/mellanox/network-operator-init-container@sha256:67e93ccf3ecb61f17597567faf0f72e1b8ddcf73c5d7440baeadcc1cb6bb811b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:4a620bd39bc70928cd6e249855176ced0f4feb585c8c7624c0b9f5a3e57664c9
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:69e6ab6a0249be918f60c1e04fc53a07f280b33563f8edc953806538a9cc3c30
        - ghcr.io/mellanox/nic-configuration-operator@sha256:353ab27f5f5fba08197cbe8968222e1c115431cc1b9b8d8445f1facb3c3c4d7b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:35ff3b173b2694198e57b61720545be67e944c1da18ac004007a5b4e68562adc
        - ghcr.io/mellanox/ib-kubernetes@sha256:9774f774df6ede9ddf780b9193ca9d919618056e682d985e6b41e477b13c3432
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:cabce074d10a0f1d62135e2cc5442d65b49094b95b8297fdd024a1a5f461319f
        - ghcr.io/mellanox/nic-feature-discovery@sha256:35b1820fe94a833a43613ef5c9db1f58b293ea0dbcf96ec0742b0987f23f133c
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:76529c24ea17e68b89a2c0d4823d66504fa6f53759760a449308fda396574097
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:93c2063f4da6a93810784be95df0083f07618486699d93245c97a29a63861dee
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:1eb5ab7f4b6f0cf89c954cf7ac8eb2e6fb2290a4f5f57417136a5302f20b12c8
        version: 25.4.0
      entries:
      - name: nvidia-network-operator.v25.4.0
        version: 25.4.0
      name: v25.4
    defaultChannel: stable
    packageName: nvidia-network-operator
    provider:
      name: NVIDIA
      url: https://github.com/Mellanox/network-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: node-observability-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-observability-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservability",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "labels": {
                    "kubernetes.io/hostname": "worker-a"
                  },
                  "type": "crio-kubelet"
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservabilityMachineConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "debug": {
                    "enableCrioProfiling": true
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservabilityRun",
                "metadata": {
                  "name": "nodeobservabilityrun-sample",
                  "namespace": "node-observability-operator"
                },
                "spec": {
                  "nodeObservabilityRef": {
                    "name": "cluster"
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservability",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "nodeSelector": {
                    "kubernetes.io/hostname": "worker-a"
                  },
                  "type": "crio-kubelet"
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservabilityMachineConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "debug": {
                    "enableCrioProfiling": true
                  },
                  "nodeSelector": {
                    "kubernetes.io/hostname": "worker-a"
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservabilityRun",
                "metadata": {
                  "name": "nodeobservabilityrun-sample",
                  "namespace": "node-observability-operator"
                },
                "spec": {
                  "nodeObservabilityRef": {
                    "name": "cluster"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/noo/node-observability-rhel8-operator@sha256:53ba77f648154ac8aca1ae6c83815078dfde211dba1e42f597e69844e4954c01
          createdAt: 2022-12-05T09:25:09
          olm.skipRange: <0.2.0
          operatorframework.io/suggested-namespace: node-observability-operator
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeObservability prepares a subset of worker nodes (identified
              by a label) for running node observability queries, such as profiling
            displayName: Node Observability
            kind: NodeObservability
            name: nodeobservabilities.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservability prepares a subset of worker nodes (identified
              by a label) for running node observability queries, such as profiling
            displayName: Node Observability
            kind: NodeObservability
            name: nodeobservabilities.nodeobservability.olm.openshift.io
            version: v1alpha2
          - description: NodeObservabilityMachineConfig is the Schema for the nodeobservabilitymachineconfigs
              API
            displayName: Node Observability Machine Config
            kind: NodeObservabilityMachineConfig
            name: nodeobservabilitymachineconfigs.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservabilityMachineConfig is the Schema for the nodeobservabilitymachineconfigs
              API
            displayName: Node Observability Machine Config
            kind: NodeObservabilityMachineConfig
            name: nodeobservabilitymachineconfigs.nodeobservability.olm.openshift.io
            version: v1alpha2
          - description: NodeObservabilityRun is a request to run observability actions
              on the nodes previously selected in NodeObservability resource
            displayName: Node Observability Run
            kind: NodeObservabilityRun
            name: nodeobservabilityruns.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservabilityRun is a request to run observability actions
              on the nodes previously selected in NodeObservability resource
            displayName: Node Observability Run
            kind: NodeObservabilityRun
            name: nodeobservabilityruns.nodeobservability.olm.openshift.io
            version: v1alpha2
        description: |-
          An Operator that will be able to gather debugging/profiling data over
          a custom period of time which would be helpful to troubleshoot and resolve issues
          for OpenShift customers.
        displayName: Node Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node-observability-operator
        links:
        - name: Source Code
          url: https://github.com/openshift/node-observability-operator
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/noo/node-observability-rhel8-operator@sha256:53ba77f648154ac8aca1ae6c83815078dfde211dba1e42f597e69844e4954c01
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/noo/node-observability-agent-rhel8@sha256:a8e60d9f63e2ed0cf1e27309aac54014e3f1efc947b51ca0a99376338d220875
        version: 0.2.0
      entries:
      - name: node-observability-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: node-observability-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lib-bucket-provisioner
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lib-bucket-provisioner.v2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage,Big Data
          certified: "false"
          containerImage: quay.io/noobaa/pause
          createdAt: "2020-07-01T06:28:48Z"
          description: Library for the dynamic provisioning of object store buckets
            to be used by object store providers.
          repository: https://github.com/kube-object-storage/lib-bucket-provisioner
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: ObjectBucketClaim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information
            displayName: ObjectBucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: |
          ### CRD-only Operator

          This operator package is **CRD-only** and the operator is a no-op operator.

          Instead, bucket provisioners using this library are using these CRD's and using CSV [required-crds](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/Documentation/design/building-your-csv.md#required-crds) them so that OLM can install it as a dependency.

          ### Generic Bucket Provisioning

          Kubernetes natively supports dynamic provisioning for many types of file and block storage, but lacks support for object bucket provisioning.

          This repo is a placeholder for an object store bucket provisioning library, very similar to the Kubernetes [sig-storage-lib-external-provisioner](https://github.com/kubernetes-sigs/sig-storage-lib-external-provisioner/blob/master/controller/controller.go) library.

          ### Known Provisioners
          - https://github.com/noobaa/noobaa-operator (NooBaa)
          - https://github.com/rook/rook (Rook-Ceph)
          - https://github.com/yard-turkey/aws-s3-provisioner (AWS-S3)
        displayName: lib-bucket-provisioner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - object
        - bucket
        - storage
        - cloud
        - s3
        links:
        - name: Github
          url: https://github.com/kube-object-storage/lib-bucket-provisioner
        maintainers:
        - email: jcope@redhat.com
          name: Jon Cope
        - email: jvance@redhat.com
          name: Jeff Vance
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: dzaken@redhat.com
          name: Danny Zaken
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/noobaa/pause
        version: 2.0.0
      entries:
      - name: lib-bucket-provisioner.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: lib-bucket-provisioner
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Odigos
      provider-url: https://odigos.io
    name: odigos-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odigos-operator.v1.0.212
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.odigos.io/v1alpha1",
                "kind": "Odigos",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "odigos-operator"
                  },
                  "name": "odigos-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing
          containerImage: registry.connect.redhat.com/odigos/odigos-certified-operator-ubi9@sha256:c221eb2be22c2fcabcee728429bbca1255e0e31642c572e4dc40096d1d35e0d1
          createdAt: "2025-07-18T13:23:10Z"
          description: Odigos enables automatic distributed tracing with OpenTelemetry
            and eBPF.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Odigos Enterprise subscription
            (for enterprise features)
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Odigos
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Odigos is the Schema for the odigos API
            displayName: Odigos
            kind: Odigos
            name: odigos.operator.odigos.io
            version: v1alpha1
        description: |-
          The Odigos Operator provides options for installing and configuring Odigos.

          Upon creation of an Odigos resource, the Odigos Operator installs the following Odigos components:

          - Instrumentor
          - Odiglet
          - Autoscaler
          - Scheduler
          - Frontend UI

          With Odigos installed, follow the Odigos docs at docs.odigos.io for instructions on instrumenting applications and configuring destinations.
        displayName: Odigos Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - OpenTelemetry
        - eBPF
        - tracing
        - observability
        - distributed tracing
        - otel
        - monitoring
        links:
        - name: Odigos
          url: https://odigos.io
        - name: Odigos Documentation
          url: https://docs.odigos.io
        - name: Odigos on Github
          url: https://github.com/odigos-io/odigos
        maintainers:
        - email: mike@odigos.io
          name: Mike Dame
        maturity: alpha
        minKubeVersion: 1.20.15
        provider:
          name: Odigos
          url: https://odigos.io
        relatedImages:
        - registry.connect.redhat.com/odigos/odigos-certified-operator-ubi9@sha256:c221eb2be22c2fcabcee728429bbca1255e0e31642c572e4dc40096d1d35e0d1
        - registry.connect.redhat.com/odigos/odigos-ui-ubi9@sha256:b2a27cdfc94f1968a06ae18de1977127bb7d4e7d553e875d09a212693302f9c6
        - registry.connect.redhat.com/odigos/odigos-enterprise-instrumentor-ubi9@sha256:647e526cdbe38023bb221a802b180f0b95cb25734a0b39fa11cbbe22a07482a3
        - registry.connect.redhat.com/odigos/odigos-scheduler-ubi9@sha256:9e613c9b22d2fbfa3d21cc24f3b1e2dbc2e7287790e1f4286f1817c92565c865
        - registry.connect.redhat.com/odigos/odigos-instrumentor-ubi9@sha256:16712373c8e927b7208e43e9fe6e21180289975dbf3a81df60578e8550cf7150
        - registry.connect.redhat.com/odigos/odigos-odiglet-ubi9@sha256:770193901b4e5abcec59833f56f1530b43d4ec9514f3b5a5cb0bbb4356e06c67
        - registry.connect.redhat.com/odigos/odigos-enterprise-odiglet-ubi9@sha256:aec9b66461375994dc9aa92e5b2cc7d4151cf0332150726840cec049a7cbe7d9
        - registry.connect.redhat.com/odigos/odigos-autoscaler-ubi9@sha256:692193851b40f72ca79f7c82903136686d721c7fed77ef599037269ac6e55fdb
        - registry.connect.redhat.com/odigos/odigos-collector-ubi9@sha256:b67138d3ffd9e721602c5bf83f0c03ae1dcee230eca5e15335dcb9abced281ba
        version: 1.0.212
      entries:
      - name: odigos-operator.v1.0.212
        version: 1.0.212
      name: alpha
    defaultChannel: alpha
    packageName: odigos-operator
    provider:
      name: Odigos
      url: https://odigos.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-superset-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: superset-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:e0c9be4ac01a3d02f908f71f47f96bb1423fee14ba454c1bc627b35654d9cae3
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:e0c9be4ac01a3d02f908f71f47f96bb1423fee14ba454c1bc627b35654d9cae3
        version: 24.11.1
      entries:
      - name: superset-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: superset-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:f12cc3ed35135c14a82d49fcafed55465f0e66d46c2a36d532e5e09b9dece31e
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:f12cc3ed35135c14a82d49fcafed55465f0e66d46c2a36d532e5e09b9dece31e
        version: 25.3.0
      entries:
      - name: superset-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: superset-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
        version: 25.7.0
      entries:
      - name: superset-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: superset-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
        version: 25.7.0
      entries:
      - name: superset-operator.v25.7.0
        version: 25.7.0
      - name: superset-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-superset-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: LINBIT
      provider-url: ""
    name: linstor-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: linstor-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorController",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerImage": "",
                  "dbConnectionURL": "etcd://linstor-etcd:2379",
                  "drbdRepoCred": "",
                  "priorityClassName": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorCSIDriver",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerReplicas": 1,
                  "controllerEndpoint": "http://linstor:3370",
                  "csiControllerServiceAccountName": "csi-controller",
                  "csiNodeServiceAccountName": "csi-node",
                  "linstorPluginImage": "",
                  "imagePullSecret": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorSatelliteSet",
                "metadata": {
                  "name": "linstor-satellites"
                },
                "spec": {
                  "automaticStorageType": "None",
                  "controllerEndpoint": "http://linstor:3370",
                  "drbdRepoCred": "",
                  "kernelModuleInjectionMode": "ShippedModules",
                  "priorityClassName": "",
                  "satelliteImage": ""
                },
                "status": {
                  "errors": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          createdAt: "2021-06-22T07:57:53+00:00"
          description: LINSTOR Kubernetes Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/piraeusdatastore/piraeus-operator
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LinstorController is the Schema for the linstorcontrollersets
              API
            kind: LinstorController
            name: linstorcontrollers.linstor.linbit.com
            version: v1
          - description: LinstorCSIDriver is the Schema for the linstorcsidrivers
              API
            kind: LinstorCSIDriver
            name: linstorcsidrivers.linstor.linbit.com
            version: v1
          - description: LinstorSatelliteSet is the Schema for the linstornodesets
              API
            kind: LinstorSatelliteSet
            name: linstorsatellitesets.linstor.linbit.com
            version: v1
        description: |-
          LINSTOR is a configuration management system for storage on Linux systems.
          It manages LVM logical volumes and/or ZFS ZVOLs on a cluster of nodes.
          It leverages DRBD for replication between different nodes and to provide block
          storage devices to users and applications. It manages snapshots, encryption and
          caching of HDD backed data in SSDs via bcache.

          LINBIT provides a certified LINSTOR operator to ease deployment of LINSTOR
          on Openshift by installing DRBD, managing Satellite and Controller pods,
          and integrating with Openshift to provision persistent storage for your workloads.

          For detailed instructions and more configuration options, see our [user guide].

          [user guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift

          ## Install

          Unlike deployment via the helm chart, the certified Openshift
          operator does not deploy the needed etcd cluster. You must deploy this
          yourself ahead of time. We do this via the etcd operator available in the
          OperatorHub.

          IMPORTANT: It is advised that the etcd deployment uses persistent
          storage of some type. Either use an existing storage provisioner with
          a default `StorageClass` or simply use `hostPath` volumes.

          ### Installing the operator

          Hit "Install", select the stable update channel and a namespace for the
          operator. Use of a new namespace is recommended.

          Hit "Install" again. At this point you should have just one pod, the
          operator pod, running. Next we needs to configure the remaining provided APIs.

          #### A note on operator namespaces
          The LINSTOR operator can only watch for events and manage
          custom resources that are within the same namespace it is deployed
          within (OwnNamsespace). This means the LINSTOR Controller, LINSTOR
          Satellites, and LINSTOR CSI Driver pods all need to be deployed in the
          same namsepace as the LINSTOR Operator pod.

          ### Deploying the LINSTOR Controller

          Navigate to the left-hand control pane of the Openshift Web
          Console. Expand the "Operators" section, selecting "Installed Operators".
          Find the entry for the "Linstor Operator", then select the "LinstorController"
          from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorController". Click the "Create LinstorController" button.

          Here you will be presented with options to configure the LINSTOR
          Controller. Either via the web-form view or the YAML View. Regardless
          of which view you select, make sure that the `dbConnectionURL` matches
          the endpoint provided from your etcd deployment. Otherwise, the
          defaults are usually fine for most purposes.

          Lastly hit "Create", you should now see a linstor-controller pod
          running.

          ### Deploying the LINSTOR Satellites

          Next we need to deploy the Satellites Set. Just as before navigate
          to the left-hand control pane of the Openshift Web Console. Expand the
          "Operators" section, but this time select "Installed Operators". Find
          the entry for the "Linstor Operator", then select the
          "LinstorSatelliteSet" from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorSatelliteSet". Click the "Create LinstorSatelliteSet" button.

          Here you will be presented with the options to configure the LINSTOR
          Satellites. The defaults should be enough to get you started.
          Make sure the `controllerEndpoint` matches what is available in the
          openshift endpoints. The default is usually correct here.

          You can edit the `storagePools` section to configure LINSTOR storage pools,
          including preparing the backing devices. See our [storage guide].

          [storage guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-storage

          Below is an example manifest:

          ```yaml
          apiVersion: linstor.linbit.com/v1
          kind: LinstorSatelliteSet
          metadata:
            name: linstor-satellites
          spec:
            satelliteImage: ''
            automaticStorageType: None
            storagePools:
              lvmThinPools:
              - name: openshift-pool
                volumeGroup: ""
                thinVolume: openshift
                devicePaths:
                - /dev/vdb
            drbdRepoCred: ''
            kernelModuleInjectionMode: ShippedModules
            controllerEndpoint: 'http://linstor:3370'
            priorityClassName: ''
          ```

          Lastly hit "Create", you should now see a linstor-node pod
          running on every worker node.

          ### Deploying the LINSTOR CSI driver

          Last bit left is the CSI pods to bridge the layer between the CSI and
          LINSTOR. Just as before navigate to the left-hand control pane of the
          Openshift Web Console. Expand the "Operators" section, but this time
          select "Installed Operators". Find the entry for the "Linstor Operator",
          then select the "LinstorCSIDriver" from the "Provided APIs" column on the
          right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorCSIDriver". Click the "Create LinstorCSIDriver" button.

          Again, you will be presented with the options. Make sure that the
          `controllerEndpoint` is correct. Otherwise the defaults are fine for
          most use cases.

          Lastly hit "Create". You will now see a single "linstor-csi-controller" pod,
          as well as a "linstor-csi-node" pod on all worker nodes.

          ## Interacting with LINSTOR in Openshift.

          The Controller pod includes a LINSTOR Client, making it easy to interact directly with LINSTOR.
          For instance:

          ```
          oc exec deployment/linstor-cs-controller -- linstor storage-pool list
          ```

          This should only be necessary for investigating problems and accessing advanced functionality.
          Regular operation such as creating volumes should be achieved via the [Openshift/Kubernetes integration].

          [Openshift/Kubernetes integration]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-basic-configuration-and-deployment
        displayName: Linstor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - DRBD
        links:
        - name: LINBIT
          url: https://www.linbit.com/
        - name: User Guide
          url: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift
        maintainers:
        - email: sales@linbit.com
          name: LINBIT
        maturity: beta
        provider:
          name: LINBIT
        relatedImages:
        - registry.marketplace.redhat.com/rhm/linbit/linstor-operator@sha256:75e3ec9a9331072dc88c13979f0cacdf6232f8b3a6bfaf55c95b2ee3a03a5d5b
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:395a5a4aa4cfe3a0093d2225ce2e67acdcec0fd894e4b61e30a750f22931448d
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:bba872052e42fe0d6cbb2eea6541bf57f035575cbc7d838a3260b8582b8fe572
        - registry.marketplace.redhat.com/rhm/linbit/linstor-controller@sha256:1084ba14eaec7511682dd87c562398ce5e6690ded700c964373188b73bfe3b17
        - registry.marketplace.redhat.com/rhm/linbit/drbd-9@sha256:6337b3adfa3de87e74adcc974edafa09b618cb8e0315658c1cbdf8dd1c6c4c32
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:ba217e7e51f0748b54f477486263ff4983b0029635fd1f7ccc67160a09f81608
        - registry.marketplace.redhat.com/rhm/linbit/linstor-csi@sha256:034e3641376aa7ad47ed309fb389083c3ff82974f246f4d71b8789be8f8f618d
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:53b0f48eeecaa3ff1e211a2a413cd49266b7be85a5aa3a0431ab72d32e3a4e5c
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:8e52a6b31f9f3834d709dbca3614c2110b8ebb4af382c76b7056921221c87600
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:91cfe63acc12c91c3053ba7cc12d72cce404d33aab4afcfac285311f5ddfc7fe
        - registry.marketplace.redhat.com/rhm/linbit/linstor-satellite@sha256:4a3503092c8cb9435ccfb4837ee6a9183b680eb18a35c7d40227fa5c8f4da996
        version: 1.5.1
      entries:
      - name: linstor-operator.v1.5.1
        version: 1.5.1
      name: alpha
    - currentCSV: linstor-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorController",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerImage": "",
                  "dbConnectionURL": "etcd://linstor-etcd:2379",
                  "drbdRepoCred": "",
                  "priorityClassName": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorCSIDriver",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerReplicas": 1,
                  "controllerEndpoint": "http://linstor:3370",
                  "csiControllerServiceAccountName": "csi-controller",
                  "csiNodeServiceAccountName": "csi-node",
                  "linstorPluginImage": "",
                  "imagePullSecret": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorSatelliteSet",
                "metadata": {
                  "name": "linstor-satellites"
                },
                "spec": {
                  "automaticStorageType": "None",
                  "controllerEndpoint": "http://linstor:3370",
                  "drbdRepoCred": "",
                  "kernelModuleInjectionMode": "ShippedModules",
                  "priorityClassName": "",
                  "satelliteImage": ""
                },
                "status": {
                  "errors": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          createdAt: "2021-06-22T07:57:53+00:00"
          description: LINSTOR Kubernetes Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/piraeusdatastore/piraeus-operator
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LinstorController is the Schema for the linstorcontrollersets
              API
            kind: LinstorController
            name: linstorcontrollers.linstor.linbit.com
            version: v1
          - description: LinstorCSIDriver is the Schema for the linstorcsidrivers
              API
            kind: LinstorCSIDriver
            name: linstorcsidrivers.linstor.linbit.com
            version: v1
          - description: LinstorSatelliteSet is the Schema for the linstornodesets
              API
            kind: LinstorSatelliteSet
            name: linstorsatellitesets.linstor.linbit.com
            version: v1
        description: |-
          LINSTOR is a configuration management system for storage on Linux systems.
          It manages LVM logical volumes and/or ZFS ZVOLs on a cluster of nodes.
          It leverages DRBD for replication between different nodes and to provide block
          storage devices to users and applications. It manages snapshots, encryption and
          caching of HDD backed data in SSDs via bcache.

          LINBIT provides a certified LINSTOR operator to ease deployment of LINSTOR
          on Openshift by installing DRBD, managing Satellite and Controller pods,
          and integrating with Openshift to provision persistent storage for your workloads.

          For detailed instructions and more configuration options, see our [user guide].

          [user guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift

          ## Install

          Unlike deployment via the helm chart, the certified Openshift
          operator does not deploy the needed etcd cluster. You must deploy this
          yourself ahead of time. We do this via the etcd operator available in the
          OperatorHub.

          IMPORTANT: It is advised that the etcd deployment uses persistent
          storage of some type. Either use an existing storage provisioner with
          a default `StorageClass` or simply use `hostPath` volumes.

          ### Installing the operator

          Hit "Install", select the stable update channel and a namespace for the
          operator. Use of a new namespace is recommended.

          Hit "Install" again. At this point you should have just one pod, the
          operator pod, running. Next we needs to configure the remaining provided APIs.

          #### A note on operator namespaces
          The LINSTOR operator can only watch for events and manage
          custom resources that are within the same namespace it is deployed
          within (OwnNamsespace). This means the LINSTOR Controller, LINSTOR
          Satellites, and LINSTOR CSI Driver pods all need to be deployed in the
          same namsepace as the LINSTOR Operator pod.

          ### Deploying the LINSTOR Controller

          Navigate to the left-hand control pane of the Openshift Web
          Console. Expand the "Operators" section, selecting "Installed Operators".
          Find the entry for the "Linstor Operator", then select the "LinstorController"
          from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorController". Click the "Create LinstorController" button.

          Here you will be presented with options to configure the LINSTOR
          Controller. Either via the web-form view or the YAML View. Regardless
          of which view you select, make sure that the `dbConnectionURL` matches
          the endpoint provided from your etcd deployment. Otherwise, the
          defaults are usually fine for most purposes.

          Lastly hit "Create", you should now see a linstor-controller pod
          running.

          ### Deploying the LINSTOR Satellites

          Next we need to deploy the Satellites Set. Just as before navigate
          to the left-hand control pane of the Openshift Web Console. Expand the
          "Operators" section, but this time select "Installed Operators". Find
          the entry for the "Linstor Operator", then select the
          "LinstorSatelliteSet" from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorSatelliteSet". Click the "Create LinstorSatelliteSet" button.

          Here you will be presented with the options to configure the LINSTOR
          Satellites. The defaults should be enough to get you started.
          Make sure the `controllerEndpoint` matches what is available in the
          openshift endpoints. The default is usually correct here.

          You can edit the `storagePools` section to configure LINSTOR storage pools,
          including preparing the backing devices. See our [storage guide].

          [storage guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-storage

          Below is an example manifest:

          ```yaml
          apiVersion: linstor.linbit.com/v1
          kind: LinstorSatelliteSet
          metadata:
            name: linstor-satellites
          spec:
            satelliteImage: ''
            automaticStorageType: None
            storagePools:
              lvmThinPools:
              - name: openshift-pool
                volumeGroup: ""
                thinVolume: openshift
                devicePaths:
                - /dev/vdb
            drbdRepoCred: ''
            kernelModuleInjectionMode: ShippedModules
            controllerEndpoint: 'http://linstor:3370'
            priorityClassName: ''
          ```

          Lastly hit "Create", you should now see a linstor-node pod
          running on every worker node.

          ### Deploying the LINSTOR CSI driver

          Last bit left is the CSI pods to bridge the layer between the CSI and
          LINSTOR. Just as before navigate to the left-hand control pane of the
          Openshift Web Console. Expand the "Operators" section, but this time
          select "Installed Operators". Find the entry for the "Linstor Operator",
          then select the "LinstorCSIDriver" from the "Provided APIs" column on the
          right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorCSIDriver". Click the "Create LinstorCSIDriver" button.

          Again, you will be presented with the options. Make sure that the
          `controllerEndpoint` is correct. Otherwise the defaults are fine for
          most use cases.

          Lastly hit "Create". You will now see a single "linstor-csi-controller" pod,
          as well as a "linstor-csi-node" pod on all worker nodes.

          ## Interacting with LINSTOR in Openshift.

          The Controller pod includes a LINSTOR Client, making it easy to interact directly with LINSTOR.
          For instance:

          ```
          oc exec deployment/linstor-cs-controller -- linstor storage-pool list
          ```

          This should only be necessary for investigating problems and accessing advanced functionality.
          Regular operation such as creating volumes should be achieved via the [Openshift/Kubernetes integration].

          [Openshift/Kubernetes integration]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-basic-configuration-and-deployment
        displayName: Linstor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - DRBD
        links:
        - name: LINBIT
          url: https://www.linbit.com/
        - name: User Guide
          url: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift
        maintainers:
        - email: sales@linbit.com
          name: LINBIT
        maturity: beta
        provider:
          name: LINBIT
        relatedImages:
        - registry.marketplace.redhat.com/rhm/linbit/linstor-csi@sha256:034e3641376aa7ad47ed309fb389083c3ff82974f246f4d71b8789be8f8f618d
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:8e52a6b31f9f3834d709dbca3614c2110b8ebb4af382c76b7056921221c87600
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:395a5a4aa4cfe3a0093d2225ce2e67acdcec0fd894e4b61e30a750f22931448d
        - registry.marketplace.redhat.com/rhm/linbit/linstor-controller@sha256:1084ba14eaec7511682dd87c562398ce5e6690ded700c964373188b73bfe3b17
        - registry.marketplace.redhat.com/rhm/linbit/linstor-satellite@sha256:4a3503092c8cb9435ccfb4837ee6a9183b680eb18a35c7d40227fa5c8f4da996
        - registry.marketplace.redhat.com/rhm/linbit/linstor-operator@sha256:75e3ec9a9331072dc88c13979f0cacdf6232f8b3a6bfaf55c95b2ee3a03a5d5b
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:53b0f48eeecaa3ff1e211a2a413cd49266b7be85a5aa3a0431ab72d32e3a4e5c
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:bba872052e42fe0d6cbb2eea6541bf57f035575cbc7d838a3260b8582b8fe572
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:ba217e7e51f0748b54f477486263ff4983b0029635fd1f7ccc67160a09f81608
        - registry.marketplace.redhat.com/rhm/linbit/drbd-9@sha256:6337b3adfa3de87e74adcc974edafa09b618cb8e0315658c1cbdf8dd1c6c4c32
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:91cfe63acc12c91c3053ba7cc12d72cce404d33aab4afcfac285311f5ddfc7fe
        version: 1.5.1
      entries:
      - name: linstor-operator.v1.5.1
        version: 1.5.1
      name: stable
    defaultChannel: stable
    packageName: linstor-operator-rhmp
    provider:
      name: LINBIT
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: pg4k-pgd
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pg4k-pgd.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgd.k8s.enterprisedb.io/v1beta1",
                "kind": "PGDGroup",
                "metadata": {
                  "name": "region-a"
                },
                "spec": {
                  "backup": null,
                  "cnp": {
                    "storage": {
                      "size": "1Gi"
                    },
                    "walStorage": {
                      "size": "1Gi"
                    }
                  },
                  "connectivity": {
                    "dns": {
                      "additional": [
                        {
                          "domain": "alternate.domain"
                        },
                        {
                          "domain": "my.domain",
                          "hostSuffix": "-dc1"
                        }
                      ]
                    },
                    "tls": {
                      "clientCert": {
                        "caCertSecret": "client-ca-key-pair",
                        "certManager": {
                          "spec": {
                            "issuerRef": {
                              "group": "cert-manager.io",
                              "kind": "Issuer",
                              "name": "client-ca-issuer"
                            }
                          }
                        }
                      },
                      "mode": "verify-ca",
                      "serverCert": {
                        "caCertSecret": "server-ca-key-pair",
                        "certManager": {
                          "spec": {
                            "issuerRef": {
                              "group": "cert-manager.io",
                              "kind": "Issuer",
                              "name": "server-ca-issuer"
                            }
                          }
                        }
                      }
                    }
                  },
                  "instances": 2,
                  "pgd": {
                    "parentGroup": {
                      "create": true,
                      "name": "world"
                    }
                  },
                  "proxyInstances": 2,
                  "witness": {
                    "storage": {
                      "size": "1Gi"
                    },
                    "walStorage": {
                      "size": "1Gi"
                    }
                  },
                  "witnessInstances": 1
                }
              },
              {
                "apiVersion": "pgd.k8s.enterprisedb.io/v1beta1",
                "kind": "PGDGroupCleanup",
                "metadata": {
                  "name": "region-b-cleanup"
                },
                "spec": {
                  "executor": "region-a",
                  "target": "region-b"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: docker.enterprisedb.com/k8s_enterprise_pgd/pg4k-pgd-openshift@sha256:5127d913011619d61fd06b27be3c709df2236a4b5c0254fa58632698d58b50fb
          createdAt: "2025-09-26T01:49:29Z"
          description: EDB Postgres Distributed for Kubernetes is an operator designed
            to manage EDB Postgres Distributed (PGD) v5 workloads on Kubernetes both
            in a single Kubernetes cluster or across multiple ones.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterImageCatalog
            name: clusterimagecatalogs.pgd.k8s.enterprisedb.io
            version: v1beta1
          - kind: ImageCatalog
            name: imagecatalogs.pgd.k8s.enterprisedb.io
            version: v1beta1
          - description: Object to clean up a PGD Group
            displayName: PGD Group Cleanup
            kind: PGDGroupCleanup
            name: pgdgroupcleanups.pgd.k8s.enterprisedb.io
            version: v1beta1
          - description: A group of Postgres Clusters
            displayName: PGD Group
            kind: PGDGroup
            name: pgdgroups.pgd.k8s.enterprisedb.io
            version: v1beta1
        description: |
          EDB Postgres Distributed for Kubernetes is an operator designed to manage EDB Postgres Distributed (PGD) v5 workloads on Kubernetes both in a single Kubernetes cluster or across multiple ones.

          EDB Postgres Distributed is an extension for PostgreSQL that provides multi-master replication and data distribution with advanced conflict management, data-loss protection, and throughput up to 5X faster than native logical replication. It enables distributed PostgreSQL clusters with high availability up to five 9s.

          This operator requires access to the EDB private repositories.
        displayName: EDB Postgres Distributed for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - distributed
        - postgres
        - database
        - sql
        - cloud
        - native
        - edb
        - enterprisedb
        - cnp
        - pg4k
        - pgd
        links:
        - name: 'Documentation: EDB Postgres Distributed v5'
          url: https://www.enterprisedb.com/docs/pgd/latest/
        - name: 'Documentation: EDB Postgres Distributed for Kubernetes'
          url: https://www.enterprisedb.com/docs/postgres_distributed_for_kubernetes/latest/
        maintainers:
        - email: sales@enterprisedb.com
          name: Sales
        - email: support@enterprisedb.com
          name: Support
        maturity: alpha
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - docker.enterprisedb.com/k8s_enterprise_pgd/pg4k-pgd-openshift@sha256:5127d913011619d61fd06b27be3c709df2236a4b5c0254fa58632698d58b50fb
        version: 1.1.3
      entries:
      - name: pg4k-pgd.v1.1.3
        version: 1.1.3
      - name: pg4k-pgd.v1.1.2
        version: 1.1.2
      name: stable
    defaultChannel: stable
    packageName: pg4k-pgd
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Mantisnet
      provider-url: https://www.mantisnet.com
    name: tawon-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tawon-operator.v2.29.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Directive",
                "metadata": {
                  "name": "amf-h2"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "amf"
                    }
                  },
                  "container": {
                    "resources": {
                      "limits": {
                        "cpu": "500m"
                      }
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  },
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 8080"
                      },
                      "task": "payload"
                    },
                    {
                      "task": "h2"
                    },
                    {
                      "config": {
                        "addr": "nats://nats:4222",
                        "topic": "hello"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "TopologyAggregator",
                "metadata": {
                  "name": "topology-aggregator"
                },
                "spec": {
                  "messaging": {
                    "nats": {
                      "url": "nats://nats:4222"
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-11-13T01:16:29Z"
          olm.skipRange: '>=2.23.0 <2.29.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Directive allows to define a Tawon task pipeline that will
              be applied to a Kubernetes cluster.
            displayName: Tawon Directive
            kind: Directive
            name: directives.tawon.mantisnet.com
            version: v1alpha1
          - description: TopologyAggregator is the Schema for the topologyaggregators
              API
            displayName: Tawon Topology Aggregator
            kind: TopologyAggregator
            name: topologyaggregators.tawon.mantisnet.com
            version: v1alpha1
        description: Tawon is a cloud-native observability fabric for deeply instrumenting
          microservice-based applications.
        displayName: Tawon
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - networking
        - metrics
        - ebpf
        links:
        - name: Tawon Docs
          url: https://docs.tawon.xmantis.net
        maintainers:
        - email: elliott@mantisnet.com
          name: Elliott Starin
        - email: nicolas@mantisnet.com
          name: Nicolas Paton
        - email: thiago@mantisnet.com
          name: Thiago Navarro
        maturity: alpha
        provider:
          name: Mantisnet
          url: https://www.mantisnet.com
        relatedImages:
        - quay.io/mantisnet/kernel-headers@sha256:242e837af831b6eed98d522b6bae7416928b9fbe451fb6e02ffa678686771068
        - quay.io/mantisnet/kernel-headers@sha256:2aa3bc7871fe0fc0a79ac812d941e55cd0d25a5f0d1d9b3e2435aab6f3d70c67
        - quay.io/mantisnet/tawonctl@sha256:d436ef902bb08d830f64f34fdb86b20d2bd18d86233c8240014f0722c60e5e94
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - quay.io/mantisnet/tawon-operator@sha256:d5669dde46bb30411964baa86844d97721655886769b96809626246a4123caad
        - quay.io/mantisnet/tawon@sha256:83575079fe00ae6e53bb13005b94ef0047cc1974d95015c19e64dd53e13bcb93
        - quay.io/mantisnet/nats@sha256:6afc7ced92181a7c12d709e9693740847297dc5d60ab78cbac7b9b6c859ecdd1
        version: 2.29.1
      entries:
      - name: tawon-operator.v2.29.1
        version: 2.29.1
      name: alpha
    - currentCSV: tawon-operator.v2.39.27-rc6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "ClusterDirective",
                "metadata": {
                  "name": "capture-coredns"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "coredns"
                    }
                  },
                  "debug": {
                    "logLevel": "debug"
                  },
                  "streams": [
                    {
                      "maxage": "6h0m0s",
                      "maxmsgs": 1000,
                      "name": "coredns",
                      "retentionPolicy": "Delete"
                    }
                  ],
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 53"
                      },
                      "task": "capture"
                    },
                    {
                      "config": {
                        "name": "coredns",
                        "type": "stream"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Dashboard",
                "metadata": {
                  "name": "tawon-dashboard"
                },
                "spec": null
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Directive",
                "metadata": {
                  "name": "capture-coredns"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "coredns"
                    }
                  },
                  "debug": {
                    "logLevel": "debug"
                  },
                  "streams": [
                    {
                      "maxage": "6h0m0s",
                      "maxmsgs": 1000,
                      "name": "coredns",
                      "retentionPolicy": "Delete"
                    }
                  ],
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 53"
                      },
                      "task": "capture"
                    },
                    {
                      "config": {
                        "name": "coredns",
                        "type": "stream"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Stream",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "tawon-operator",
                    "app.kubernetes.io/instance": "stream-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "stream",
                    "app.kubernetes.io/part-of": "tawon-operator"
                  },
                  "name": "stream-sample"
                },
                "spec": {
                  "maxage": "6h",
                  "maxbytes": "1G",
                  "maxmsgs": 100000,
                  "store": {
                    "name": "mystreamstore"
                  }
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "StreamStore",
                "metadata": {
                  "name": "mystreamstore"
                },
                "spec": {
                  "defaults": {
                    "maxage": "240h",
                    "maxbytes": "1Gi",
                    "maxsize": 1000000
                  },
                  "limits": {
                    "max": 10000000,
                    "maxage": "2400h",
                    "maxbytes": "2Gi"
                  },
                  "persistentVolumeClaimRetentionPolicy": {
                    "whenDeleted": "Delete"
                  },
                  "volumeClaimTemplate": {
                    "metadata": {
                      "name": "mypvclaim"
                    },
                    "spec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "ssd-csi"
                    }
                  }
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "TopologyAggregator",
                "metadata": {
                  "name": "topology-aggregator"
                },
                "spec": {
                  "messaging": {
                    "nats": {
                      "url": "nats://nats:4222"
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing, Monitoring, Networking
          containerImage: quay.io/mantisnet/tawon-operator@sha256:d8ce5c8966d3590ce5f286d8e778463b85582b970ad7730ddf7828bc68843a90
          createdAt: "2025-03-08T18:58:04Z"
          description: Tawon is a cloud-native observability fabric for deeply instrumenting
            microservice-based applications.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Mantisnet Tawon Operator Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: MantisNet
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterDirective is the Schema for the clusterdirectives
              API.
            displayName: Cluster Tawon Directive
            kind: ClusterDirective
            name: clusterdirectives.tawon.mantisnet.com
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.tawon.mantisnet.com
            version: v1alpha1
          - description: A Directive allows to define a Tawon task pipeline that will
              be applied to a Kubernetes cluster.
            displayName: Tawon Directive
            kind: Directive
            name: directives.tawon.mantisnet.com
            version: v1alpha1
          - description: Stream is a persistent Tawon data stream.
            displayName: Stream
            kind: Stream
            name: streams.tawon.mantisnet.com
            version: v1alpha1
          - description: StreamStore is the Schema for the streamstores API
            displayName: Stream Store
            kind: StreamStore
            name: streamstores.tawon.mantisnet.com
            version: v1alpha1
          - description: TopologyAggregator is the Schema for the topologyaggregators
              API
            displayName: Tawon Topology Aggregator
            kind: TopologyAggregator
            name: topologyaggregators.tawon.mantisnet.com
            version: v1alpha1
        description: Tawon is a cloud-native observability fabric for deeply instrumenting
          microservice-based applications.
        displayName: Tawon
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - networking
        - metrics
        - ebpf
        links:
        - name: Tawon Docs
          url: https://docs.tawon.xmantis.net
        maintainers:
        - email: elliott@mantisnet.com
          name: Elliott Starin
        - email: nicolas@mantisnet.com
          name: Nicolas Paton
        - email: thiago@mantisnet.com
          name: Thiago Navarro
        maturity: alpha
        provider:
          name: Mantisnet
          url: https://www.mantisnet.com
        relatedImages:
        - quay.io/mantisnet/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/mantisnet/kernel-headers@sha256:ca3baeba7ab8ce968c1ee129112532fc8b8d6fb6741e42831b6100f0a2451629
        - quay.io/mantisnet/tawon-operator@sha256:d8ce5c8966d3590ce5f286d8e778463b85582b970ad7730ddf7828bc68843a90
        - quay.io/mantisnet/diagnose@sha256:ab15f4eaeda32f510d81a3cc865b0fd77405b55f6a0c6f285131331dcc38f2ca
        - quay.io/mantisnet/tawonctl@sha256:8bf40ca0cc9984a252363e87d6135e43ac52839113b762399cc54a085af171d8
        - quay.io/mantisnet/sengat@sha256:00c3e1194c2c3e343d16c1543b7ce2a1ee9829f6ddd1c2870785624596b46ebf
        - quay.io/mantisnet/kernel-headers@sha256:9290eb56b186ee0edf65da615c1a8aabc66f7d290168e49ad8764ab8fc2d209a
        - quay.io/mantisnet/tawon@sha256:46e654c6f275be2be604254102aaee9d927859cfd05dfa16b7cb571d36b0f1a2
        - quay.io/mantisnet/tawon-tests@sha256:cd49c24deb7cc9f4b439cb9baa885208dfe91de385cd8c5d312fd08280615c2c
        - quay.io/openshift/origin-oauth-proxy@sha256:28856aae9e481fe763776df713b4b335632645687e6fd2e31619f76d11e0f47c
        - quay.io/mantisnet/tawon@sha256:5a8bf475ff18abacb96f631e4d3ea01c94bef8a38f84c00d1dc4047073b6cc14
        - quay.io/mantisnet/nats@sha256:5f9e7f8f685fd0c1279a3465b61c7215f4600c99d81785dda86a71c966cf829a
        - quay.io/mantisnet/kernel-headers@sha256:5d00bfa9f60c95afcfa363ede59a9625fef9925cd1cb4235327a1de94a0fc632
        version: 2.39.27-rc6
      entries:
      - name: tawon-operator.v2.39.27-rc6
        version: 2.39.27-rc6
      name: candidate-v2
    - currentCSV: tawon-operator.v2.39.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "ClusterDirective",
                "metadata": {
                  "name": "capture-coredns"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "coredns"
                    }
                  },
                  "debug": {
                    "logLevel": "debug"
                  },
                  "streams": [
                    {
                      "maxage": "6h0m0s",
                      "maxmsgs": 1000,
                      "name": "coredns",
                      "retentionPolicy": "Delete"
                    }
                  ],
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 53"
                      },
                      "task": "capture"
                    },
                    {
                      "config": {
                        "name": "coredns",
                        "type": "stream"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Dashboard",
                "metadata": {
                  "name": "tawon-dashboard"
                },
                "spec": null
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Directive",
                "metadata": {
                  "name": "capture-coredns"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "coredns"
                    }
                  },
                  "debug": {
                    "logLevel": "debug"
                  },
                  "streams": [
                    {
                      "maxage": "6h0m0s",
                      "maxmsgs": 1000,
                      "name": "coredns",
                      "retentionPolicy": "Delete"
                    }
                  ],
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 53"
                      },
                      "task": "capture"
                    },
                    {
                      "config": {
                        "name": "coredns",
                        "type": "stream"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Stream",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "tawon-operator",
                    "app.kubernetes.io/instance": "stream-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "stream",
                    "app.kubernetes.io/part-of": "tawon-operator"
                  },
                  "name": "stream-sample"
                },
                "spec": {
                  "maxage": "6h",
                  "maxbytes": "1G",
                  "maxmsgs": 100000,
                  "store": {
                    "name": "mystreamstore"
                  }
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "StreamStore",
                "metadata": {
                  "name": "mystreamstore"
                },
                "spec": {
                  "defaults": {
                    "maxage": "240h",
                    "maxbytes": "1Gi",
                    "maxsize": 1000000
                  },
                  "limits": {
                    "max": 10000000,
                    "maxage": "2400h",
                    "maxbytes": "2Gi"
                  },
                  "persistentVolumeClaimRetentionPolicy": {
                    "whenDeleted": "Delete"
                  },
                  "volumeClaimTemplate": {
                    "metadata": {
                      "name": "mypvclaim"
                    },
                    "spec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "ssd-csi"
                    }
                  }
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "TopologyAggregator",
                "metadata": {
                  "name": "topology-aggregator"
                },
                "spec": {
                  "messaging": {
                    "nats": {
                      "url": "nats://nats:4222"
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing, Monitoring, Networking
          containerImage: quay.io/mantisnet/tawon-operator@sha256:f831abf0f9fb3a4e096743d4ff7356984738da25790e305325e740ca97f0be5b
          createdAt: "2025-07-18T16:24:07Z"
          description: Tawon is a cloud-native observability fabric for deeply instrumenting
            microservice-based applications.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.23.0 <2.39.32'
          operators.openshift.io/valid-subscription: Mantisnet Tawon Operator Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: MantisNet
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterDirective is the Schema for the clusterdirectives
              API.
            displayName: Cluster Tawon Directive
            kind: ClusterDirective
            name: clusterdirectives.tawon.mantisnet.com
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.tawon.mantisnet.com
            version: v1alpha1
          - description: A Directive allows to define a Tawon task pipeline that will
              be applied to a Kubernetes cluster.
            displayName: Tawon Directive
            kind: Directive
            name: directives.tawon.mantisnet.com
            version: v1alpha1
          - description: Stream is a persistent Tawon data stream.
            displayName: Stream
            kind: Stream
            name: streams.tawon.mantisnet.com
            version: v1alpha1
          - description: StreamStore is the Schema for the streamstores API
            displayName: Stream Store
            kind: StreamStore
            name: streamstores.tawon.mantisnet.com
            version: v1alpha1
          - description: TopologyAggregator is the Schema for the topologyaggregators
              API
            displayName: Tawon Topology Aggregator
            kind: TopologyAggregator
            name: topologyaggregators.tawon.mantisnet.com
            version: v1alpha1
        description: Tawon is a cloud-native observability fabric for deeply instrumenting
          microservice-based applications.
        displayName: Tawon
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - networking
        - metrics
        - ebpf
        links:
        - name: Tawon Docs
          url: https://docs.tawon.xmantis.net
        maintainers:
        - email: elliott@mantisnet.com
          name: Elliott Starin
        - email: nicolas@mantisnet.com
          name: Nicolas Paton
        - email: thiago@mantisnet.com
          name: Thiago Navarro
        maturity: alpha
        provider:
          name: Mantisnet
          url: https://www.mantisnet.com
        relatedImages:
        - quay.io/mantisnet/nats@sha256:5f9e7f8f685fd0c1279a3465b61c7215f4600c99d81785dda86a71c966cf829a
        - quay.io/mantisnet/tawon-tests@sha256:294df1580816919502dc592c1c28ffff32b13ee892686f783eff181863fa9d18
        - quay.io/mantisnet/tawon@sha256:1640758505d9b4e0f684ab502ef3344135a270192054e256e7e711209fa51adf
        - quay.io/mantisnet/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/mantisnet/nginx@sha256:c9f91949187fa1c2b4615b88d3acf7902c7e2d4a2557f33ca0cf90164269a7ae
        - quay.io/mantisnet/tawonctl@sha256:88ac81b4fe00d360c5ff127f77456f9e36990e435442a9194d3dbf50b20e7206
        - quay.io/mantisnet/diagnose@sha256:ab15f4eaeda32f510d81a3cc865b0fd77405b55f6a0c6f285131331dcc38f2ca
        - quay.io/mantisnet/tawon@sha256:846d03ada7e6f6648fc74fd0ec9e68c3c6b17e4c155df22829a3db3ca556c395
        - quay.io/mantisnet/kernel-headers@sha256:9290eb56b186ee0edf65da615c1a8aabc66f7d290168e49ad8764ab8fc2d209a
        - quay.io/mantisnet/tawon-operator@sha256:f831abf0f9fb3a4e096743d4ff7356984738da25790e305325e740ca97f0be5b
        - quay.io/mantisnet/alpine@sha256:ec1b05d1eac264d9204a57f4ad9d4dc35e9e756e9fedaea0674aefc7edb1d6a4
        - quay.io/mantisnet/kernel-headers@sha256:ca3baeba7ab8ce968c1ee129112532fc8b8d6fb6741e42831b6100f0a2451629
        - quay.io/mantisnet/kernel-headers@sha256:5d00bfa9f60c95afcfa363ede59a9625fef9925cd1cb4235327a1de94a0fc632
        - quay.io/mantisnet/kernel-headers@sha256:b0975a44cc94bc1d75679f288a9e23a65c2c0b350ce1fb2f5eb421a76a9567c6
        - quay.io/mantisnet/sengat@sha256:483fbce2bacbd95b7458773e2278b936f06a19e65acb8987adffa8403d76e40a
        - quay.io/openshift/origin-oauth-proxy@sha256:28856aae9e481fe763776df713b4b335632645687e6fd2e31619f76d11e0f47c
        version: 2.39.32
      entries:
      - name: tawon-operator.v2.39.32
        version: 2.39.32
      - name: tawon-operator.v2.39.31
        version: 2.39.31
      - name: tawon-operator.v2.39.29
        version: 2.39.29
      - name: tawon-operator.v2.39.28
        version: 2.39.28
      - name: tawon-operator.v2.39.27
        version: 2.39.27
      name: fast-v2
    defaultChannel: fast-v2
    packageName: tawon-operator
    provider:
      name: Mantisnet
      url: https://www.mantisnet.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Perficient Inc
      provider-url: http://www.perficient.com
    name: devopsinabox
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devopsinabox.v0.0.45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "DevSecOpsManager",
                "metadata": {
                  "name": "devsecopsmanager-sample"
                },
                "spec": {
                  "config": {
                    "repoPath": "",
                    "repoUrl": ""
                  },
                  "forwarding": {
                    "fluentd": {
                      "affinity": {},
                      "caFile": "files/default-openshift-logging-fluentd.crt",
                      "image": "registry.redhat.io/openshift4/ose-logging-fluentd:v4.4",
                      "keyFile": "files/default-openshift-logging-fluentd.key",
                      "loglevel": "warn",
                      "nodeSelector": {},
                      "passphrase": "",
                      "persistence": {
                        "accessMode": "ReadWriteOnce",
                        "enabled": false,
                        "size": "10Gi",
                        "storageClass": ""
                      },
                      "port": 24224,
                      "replicas": 2,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "512Mi"
                        }
                      },
                      "sharedkey": "splunkforwarding",
                      "ssl": true,
                      "tolerations": [],
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      }
                    },
                    "preserveDefault": true
                  },
                  "fullnameOverride": "",
                  "git": {
                    "password": "",
                    "server": "https://gitlab.com",
                    "token": "",
                    "username": ""
                  },
                  "nameOverride": "",
                  "prometheus": {
                    "token": ""
                  }
                }
              },
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "ManagedNamespace",
                "metadata": {
                  "name": "managednamespace-sample"
                },
                "spec": {
                  "config": {
                    "repoBranch": "master",
                    "repoPath": "",
                    "repoUrl": ""
                  },
                  "container": {
                    "password": "",
                    "registry": "https://quay.io",
                    "username": ""
                  },
                  "fullnameOverride": "",
                  "git": {
                    "password": "",
                    "server": "https://gitlab.com",
                    "token": "",
                    "username": ""
                  },
                  "manager": {
                    "namespace": ""
                  },
                  "mirrors": {
                    "maven": ""
                  },
                  "nameOverride": ""
                }
              },
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "SpringBootApp",
                "metadata": {
                  "name": "springbootapp-sample"
                },
                "spec": {
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "config": {
                    "repo": ""
                  },
                  "container": {
                    "image": "placeholder",
                    "registry": ""
                  },
                  "fullnameOverride": "",
                  "git": {
                    "branch": "master",
                    "module": "",
                    "url": ""
                  },
                  "management": {
                    "port": 8081
                  },
                  "maven": {
                    "mirror": ""
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "replicaCount": 1,
                  "resources": {},
                  "service": {
                    "port": 8080
                  },
                  "serviceAccount": {
                    "annotations": {}
                  },
                  "version": {
                    "branch": "",
                    "image": "placeholder",
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration,Developer Tools,Application Runtime
          containerImage: quay.io/mrethers/devopsinabox-operator:v0.0.45
          createdAt: 2021-03-03 10:51:09Z
          description: Perficient's DevSecOps Manager For Kubernetes Applications
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/prft-rh/devopsinabox-operator
          support: Matthieu Rethers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DevSecOpsManager
            name: devsecopsmanagers.apimc.com.perficient
            version: v1alpha1
          - kind: ManagedNamespace
            name: managednamespaces.apimc.com.perficient
            version: v1alpha1
          - kind: SpringBootApp
            name: springbootapps.apimc.com.perficient
            version: v1alpha1
        description: "The devops-in-a-box Operator manages the entire lifecycle of
          your applications in Kubernetes. It encapsulates all the best devsecops
          practices gathered by Perficient over the years working on many large projects.
          Use it to migrate existing applications with minimum changes or deploy new
          ones with little to no knowledge of Kubernetes.\n\nOnce installed, the Devops-in-a-box
          operator provides the following features:\n\n* **Create/Destroy**: Easily
          launch a new DevSecOpsManager instance to bootstrap your cluster and start
          deploying new applications with total confidence.\n\n* **Gitops Simplified**:
          Configure your cluster and individual applications through Git and watch
          your changes propagate instantly. Just point to where your code is and let
          the operator do the rest with preconfigured cloud-native CI/CD pipelines.\n\n*
          **Production-Ready**: Automatically register your applications with Openshift
          logging and monitoring systems and enjoy special additions for Spring Boot
          like custom dashboards and searchable logs.\n\n*Contact us on Github for
          installation and usage instruction*\n\n### Other Supported Features\n\n**Blue/Green,
          Canary, A/B Deployments**\n\nRun multiple versions of your applications
          simultaneously and configure traffic rules to target specific groups of
          users using Service Mesh.\n\n**Updates via automated operations**\n\nGet
          the latest enhancements automatically through the Operator Lifecycle Manager
          and upgrade all your running applications at once. "
        displayName: Devops-in-a-box
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devops
        - devsecops
        - spring
        - boot
        - java
        - gitops
        - cicd
        - argo
        links:
        - name: Devops-in-a-box Operator
          url: https://github.com/prft-rh/devopsinabox-operator
        maintainers:
        - email: matthieu.rethers@perficient.com
          name: Matthieu Rethers
        maturity: alpha
        provider:
          name: Perficient Inc
          url: http://www.perficient.com
        relatedImages:
        - quay.io/mrethers/devopsinabox-operator:v0.0.45
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        version: 0.0.45
      entries:
      - name: devopsinabox.v0.0.45
        version: 0.0.45
      name: alpha
    defaultChannel: alpha
    packageName: devopsinabox
    provider:
      name: Perficient Inc
      url: http://www.perficient.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: dapr.io
      provider-url: https://dapr.io
    name: dapr-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dapr-kubernetes-operator.v0.0.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprControlPlane",
                "metadata": {
                  "name": "dapr-control-plane",
                  "namespace": "dapr-system"
                },
                "spec": {
                  "values": {}
                }
              },
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprCruiseControl",
                "metadata": {
                  "name": "dapr-cruise-control",
                  "namespace": "dapr-system"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprInstance",
                "metadata": {
                  "name": "dapr-instance",
                  "namespace": "dapr-system"
                },
                "spec": {
                  "values": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          containerImage: docker.io/daprio/kubernetes-operator@sha256:da599a273943bf28812314a11df1052411237dd1887edf0fe6104c5857f439b0
          createdAt: "2025-10-06T18:47:59Z"
          description: Dapr Control Plane Operator
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/dapr/kubernetes-operator
          support: contact@dapr.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DaprControlPlane is the Schema for the Dapr ControlPlane
              API.
            displayName: DaprControlPlane
            kind: DaprControlPlane
            name: daprcontrolplanes.operator.dapr.io
            version: v1alpha1
          - kind: DaprCruiseControl
            name: daprcruiscontrols.operator.dapr.io
            version: v1alpha1
          - description: DaprInstance is the Schema for the Dapr Instance API.
            displayName: DaprInstance
            kind: DaprInstance
            name: daprinstances.operator.dapr.io
            version: v1alpha1
        description: Dapr Control Plane Operator
        displayName: Dapr Control Plane Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dapr
        links:
        - name: Dapr
          url: https://dapr.io
        - name: Dapr Kubernetes Operator
          url: https://github.com/dapr/kubernetes-operator
        maintainers:
        - email: lburgazzoli@gmail.com
          name: Luca Burgazzoli
        - email: salaboy@gmail.com
          name: Mauricio Salatino
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: dapr.io
          url: https://dapr.io
        relatedImages:
        - docker.io/daprio/kubernetes-operator@sha256:da599a273943bf28812314a11df1052411237dd1887edf0fe6104c5857f439b0
        version: 0.0.12
      entries:
      - name: dapr-kubernetes-operator.v0.0.12
        version: 0.0.12
      - name: dapr-kubernetes-operator.v0.0.11
        version: 0.0.11
      - name: dapr-kubernetes-operator.v0.0.10
        version: 0.0.10
      - name: dapr-kubernetes-operator.v0.0.9
        version: 0.0.9
      - name: dapr-kubernetes-operator.v0.0.8
        version: 0.0.8
      - name: dapr-kubernetes-operator.v0.0.6
        version: 0.0.6
      - name: dapr-kubernetes-operator.v0.0.5
        version: 0.0.5
      - name: dapr-kubernetes-operator.v0.0.4
        version: 0.0.4
      - name: dapr-kubernetes-operator.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: dapr-kubernetes-operator
    provider:
      name: dapr.io
      url: https://dapr.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: gitwebhook-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitwebhook-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitWebhook",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "gitwebhook-operator",
                    "app.kubernetes.io/instance": "gitwebhook-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gitwebhook",
                    "app.kubernetes.io/part-of": "gitwebhook-operator"
                  },
                  "name": "gitwebhook-sample"
                },
                "spec": null
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/redhat-cop/gitwebhook-operator@sha256:2810d75eebb24069dbbedf6a82285136c6f16b4cc58b59f1b0fbdc38560cf8a8
          createdAt: "2024-05-10T18:17:13Z"
          description: An operator to create git webhooks declaratively
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: gitwebhook-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/gitwebhook-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitWebhook is the Schema for the gitwebhooks API
            displayName: Git Webhook
            kind: GitWebhook
            name: gitwebhooks.redhatcop.redhat.io
            version: v1alpha1
        description: "This operator automates the setting up of git webhooks. Its
          main purpose is to stream line the configuration of operators such as Tekton
          or ArgoCD which rely on git webhooks for some of their functionalities.\n\n##
          The GitWebhook CRD\n \nA CRD is provided by this operator to create a git
          web hook, here is an example for github\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GitWebhook\nmetadata:\n  name: gitwebhook-github\nspec:\n  gitHub:\n    gitServerCredentials:\n
          \     name: github-pat\n  repositoryOwner: ${repo_owner}\n  ownerType: user\n
          \ repositoryName: ${repo_name}\n  webhookURL: https://hellowebhook.com\n
          \ insecureSSL: false\n  webhookSecret:\n    name: webhook-secret\n  events:\n
          \   - push  \n  contentType: json\n  active: true  \n```    \n\nhere is
          an explanation of each field:\n\n- `gihub` specifies how to connect to the
          git api server. It also requires a local reference to a secret (in the same
          namespace) containing a key `token` with a valid github token to be used
          to authenticate. A similar `gitLab` section exists when connecting to gitlab.
          Only one of `gitLab` or `gitHub` can be defined. \n- `repositoryOwner` and
          `repositoryName` identify the repository for which we want to receive events.\n-
          `ownerType` can have two values: `user` and `organization` and identifies
          the kind of owner.\n- `webhookURL` is the URL for to be called.\n - `insecureSSL`
          defines whether the target URL certificate should be validated (default
          `false`). \n- `webhookSecret` defines a local reference to a secret containing
          the `secret` key. The value is a shared secret between the webhook caller
          and the received for farther validation or identification of the caller.\n-
          `events` is the list of the repo-level events that the webhook should generate.
          The list of valid events for github can be found [here](https://docs.github.com/en/developers/webhooks-and-events/webhooks/webhook-events-and-payloads).
          The list of valid events for gitlab can be found [here](https://docs.gitlab.com/ee/user/project/integrations/webhook_events.html).\n
          - `contentType` defines the format of the webhook payload (default `json`)
          (github).\n - `active` whether the webhook should be turned on (default
          `true`) (github).\n - `pushEventBranchFilter` a regular expression to filter
          from which branches push events should be generated (gitlab only).\n\n##
          Security Considerations\n\nThis operator does not own credentials for the
          git server, but instead always allocate a new connection based on the credentials
          referenced in the CR and every reconcile cycle. As a result there is no
          risk of security escalation or credential leaking between tenants of a cluster
          using this operator. On the other hand it is the responsibility of the namespace
          owners or the platform owner to ensure that valid git credentials are always
          available in the namespace where the GitWebhook CRs need to defined.\n\n##
          Current support\n\nCurrently this operator support creating repo-level webhooks
          for github and gitlab. Potentially this operator could be extended to support
          org-level webhook or other git systems. Contributions are welcome.  "
        displayName: GitWebhook operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - webhook
        - github
        - gitlab
        links:
        - name: GitWbehook Operator
          url: https://github.com/redhat-cop/gitwebhook-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/gitwebhook-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/gitwebhook-operator@sha256:2810d75eebb24069dbbedf6a82285136c6f16b4cc58b59f1b0fbdc38560cf8a8
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        version: 0.1.2
      entries:
      - name: gitwebhook-operator.v0.1.2
        version: 0.1.2
      - name: gitwebhook-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: gitwebhook-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ec2-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ec2-controller.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "DHCPOptions",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "ElasticIPAddress",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "Instance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "InternetGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "NATGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "RouteTable",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "SecurityGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "Subnet",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "TransitGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "VPC",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "VPCEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ec2-controller:1.7.0
          createdAt: "2025-09-27T00:04:55Z"
          description: AWS EC2 controller is a service controller for managing EC2
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CapacityReservation represents the state of an AWS ec2 CapacityReservation
              resource.
            displayName: CapacityReservation
            kind: CapacityReservation
            name: capacityreservations.ec2.services.k8s.aws
            version: v1alpha1
          - description: DHCPOptions represents the state of an AWS ec2 DHCPOptions
              resource.
            displayName: DHCPOptions
            kind: DHCPOptions
            name: dhcpoptions.ec2.services.k8s.aws
            version: v1alpha1
          - description: ElasticIPAddress represents the state of an AWS ec2 ElasticIPAddress
              resource.
            displayName: ElasticIPAddress
            kind: ElasticIPAddress
            name: elasticipaddresses.ec2.services.k8s.aws
            version: v1alpha1
          - description: FlowLog represents the state of an AWS ec2 FlowLog resource.
            displayName: FlowLog
            kind: FlowLog
            name: flowlogs.ec2.services.k8s.aws
            version: v1alpha1
          - description: Instance represents the state of an AWS ec2 Instance resource.
            displayName: Instance
            kind: Instance
            name: instances.ec2.services.k8s.aws
            version: v1alpha1
          - description: InternetGateway represents the state of an AWS ec2 InternetGateway
              resource.
            displayName: InternetGateway
            kind: InternetGateway
            name: internetgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: LaunchTemplate represents the state of an AWS ec2 LaunchTemplate
              resource.
            displayName: LaunchTemplate
            kind: LaunchTemplate
            name: launchtemplates.ec2.services.k8s.aws
            version: v1alpha1
          - description: NATGateway represents the state of an AWS ec2 NATGateway
              resource.
            displayName: NATGateway
            kind: NATGateway
            name: natgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: NetworkACL represents the state of an AWS ec2 NetworkACL
              resource.
            displayName: NetworkACL
            kind: NetworkACL
            name: networkacls.ec2.services.k8s.aws
            version: v1alpha1
          - description: RouteTable represents the state of an AWS ec2 RouteTable
              resource.
            displayName: RouteTable
            kind: RouteTable
            name: routetables.ec2.services.k8s.aws
            version: v1alpha1
          - description: SecurityGroup represents the state of an AWS ec2 SecurityGroup
              resource.
            displayName: SecurityGroup
            kind: SecurityGroup
            name: securitygroups.ec2.services.k8s.aws
            version: v1alpha1
          - description: Subnet represents the state of an AWS ec2 Subnet resource.
            displayName: Subnet
            kind: Subnet
            name: subnets.ec2.services.k8s.aws
            version: v1alpha1
          - description: TransitGateway represents the state of an AWS ec2 TransitGateway
              resource.
            displayName: TransitGateway
            kind: TransitGateway
            name: transitgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: TransitGatewayVPCAttachment represents the state of an AWS
              ec2 TransitGatewayVPCAttachment resource.
            displayName: TransitGatewayVPCAttachment
            kind: TransitGatewayVPCAttachment
            name: transitgatewayvpcattachments.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCEndpoint represents the state of an AWS ec2 VPCEndpoint
              resource.
            displayName: VPCEndpoint
            kind: VPCEndpoint
            name: vpcendpoints.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCEndpointServiceConfiguration represents the state of an
              AWS ec2 VPCEndpointServiceConfiguration resource.
            displayName: VPCEndpointServiceConfiguration
            kind: VPCEndpointServiceConfiguration
            name: vpcendpointserviceconfigurations.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCPeeringConnection represents the state of an AWS ec2 VPCPeeringConnection
              resource.
            displayName: VPCPeeringConnection
            kind: VPCPeeringConnection
            name: vpcpeeringconnections.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPC represents the state of an AWS ec2 VPC resource.
            displayName: VPC
            kind: VPC
            name: vpcs.ec2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Compute Cloud (EC2) resources in AWS from within your Kubernetes cluster.

          **About Amazon EC2**

          Amazon Elastic Compute Cloud (Amazon EC2) provides scalable computing capacity in the Amazon Web Services (AWS) Cloud. Using Amazon EC2 eliminates your need to invest in hardware up front, so you can develop and deploy applications faster. You can use Amazon EC2 to launch as many or as few virtual servers as you need, configure security and networking, and manage storage. Amazon EC2 enables you to scale up or down to handle changes in requirements or spikes in popularity, reducing your need to forecast traffic.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EC2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ec2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EC2 Developer Resources
          url: https://aws.amazon.com/ec2/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ec2 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ec2-controller:1.7.0
        version: 1.7.0
      entries:
      - name: ack-ec2-controller.v1.7.0
        version: 1.7.0
      - name: ack-ec2-controller.v1.6.2
        version: 1.6.2
      - name: ack-ec2-controller.v1.6.1
        version: 1.6.1
      - name: ack-ec2-controller.v1.6.0
        version: 1.6.0
      - name: ack-ec2-controller.v1.4.11
        version: 1.4.11
      - name: ack-ec2-controller.v1.4.10
        version: 1.4.10
      - name: ack-ec2-controller.v1.4.9
        version: 1.4.9
      - name: ack-ec2-controller.v1.4.8
        version: 1.4.8
      - name: ack-ec2-controller.v1.4.7
        version: 1.4.7
      - name: ack-ec2-controller.v1.4.6
        version: 1.4.6
      - name: ack-ec2-controller.v1.4.5
        version: 1.4.5
      - name: ack-ec2-controller.v1.4.4
        version: 1.4.4
      - name: ack-ec2-controller.v1.4.2
        version: 1.4.2
      - name: ack-ec2-controller.v1.4.1
        version: 1.4.1
      - name: ack-ec2-controller.v1.4.0
        version: 1.4.0
      - name: ack-ec2-controller.v1.3.7
        version: 1.3.7
      - name: ack-ec2-controller.v1.3.6
        version: 1.3.6
      - name: ack-ec2-controller.v1.3.5
        version: 1.3.5
      - name: ack-ec2-controller.v1.3.4
        version: 1.3.4
      - name: ack-ec2-controller.v1.3.3
        version: 1.3.3
      - name: ack-ec2-controller.v1.3.1
        version: 1.3.1
      - name: ack-ec2-controller.v1.3.0
        version: 1.3.0
      - name: ack-ec2-controller.v1.2.28
        version: 1.2.28
      - name: ack-ec2-controller.v1.2.27
        version: 1.2.27
      - name: ack-ec2-controller.v1.2.26
        version: 1.2.26
      - name: ack-ec2-controller.v1.2.25
        version: 1.2.25
      - name: ack-ec2-controller.v1.2.24
        version: 1.2.24
      - name: ack-ec2-controller.v1.2.23
        version: 1.2.23
      - name: ack-ec2-controller.v1.2.22
        version: 1.2.22
      - name: ack-ec2-controller.v1.2.21
        version: 1.2.21
      - name: ack-ec2-controller.v1.2.20
        version: 1.2.20
      - name: ack-ec2-controller.v1.2.17
        version: 1.2.17
      - name: ack-ec2-controller.v1.2.16
        version: 1.2.16
      - name: ack-ec2-controller.v1.2.15
        version: 1.2.15
      - name: ack-ec2-controller.v1.2.14
        version: 1.2.14
      - name: ack-ec2-controller.v1.2.13
        version: 1.2.13
      - name: ack-ec2-controller.v1.2.12
        version: 1.2.12
      - name: ack-ec2-controller.v1.2.11
        version: 1.2.11
      - name: ack-ec2-controller.v1.2.10
        version: 1.2.10
      - name: ack-ec2-controller.v1.2.9
        version: 1.2.9
      - name: ack-ec2-controller.v1.2.8
        version: 1.2.8
      - name: ack-ec2-controller.v1.2.7
        version: 1.2.7
      - name: ack-ec2-controller.v1.2.6
        version: 1.2.6
      - name: ack-ec2-controller.v1.2.4
        version: 1.2.4
      - name: ack-ec2-controller.v1.2.1
        version: 1.2.1
      - name: ack-ec2-controller.v1.2.0
        version: 1.2.0
      - name: ack-ec2-controller.v1.1.2
        version: 1.1.2
      - name: ack-ec2-controller.v1.1.1
        version: 1.1.1
      - name: ack-ec2-controller.v1.1.0
        version: 1.1.0
      - name: ack-ec2-controller.v1.0.7
        version: 1.0.7
      - name: ack-ec2-controller.v1.0.6
        version: 1.0.6
      - name: ack-ec2-controller.v1.0.4
        version: 1.0.4
      - name: ack-ec2-controller.v1.0.3
        version: 1.0.3
      - name: ack-ec2-controller.v1.0.2
        version: 1.0.2
      - name: ack-ec2-controller.v1.0.1
        version: 1.0.1
      - name: ack-ec2-controller.v1.0.0
        version: 1.0.0
      - name: ack-ec2-controller.v0.1.1
        version: 0.1.1
      - name: ack-ec2-controller.v0.1.0
        version: 0.1.0
      - name: ack-ec2-controller.v0.0.21
        version: 0.0.21
      - name: ack-ec2-controller.v0.0.20
        version: 0.0.20
      - name: ack-ec2-controller.v0.0.19
        version: 0.0.19
      - name: ack-ec2-controller.v0.0.18
        version: 0.0.18
      - name: ack-ec2-controller.v0.0.17
        version: 0.0.17
      - name: ack-ec2-controller.v0.0.16
        version: 0.0.16
      - name: ack-ec2-controller.v0.0.15
        version: 0.0.15
      - name: ack-ec2-controller.v0.0.14
        version: 0.0.14
      - name: ack-ec2-controller.v0.0.13
        version: 0.0.13
      - name: ack-ec2-controller.v0.0.12
        version: 0.0.12
      - name: ack-ec2-controller.v0.0.11
        version: 0.0.11
      - name: ack-ec2-controller.v0.0.10
        version: 0.0.10
      - name: ack-ec2-controller.v0.0.9
        version: 0.0.9
      - name: ack-ec2-controller.v0.0.8
        version: 0.0.8
      - name: ack-ec2-controller.v0.0.7
        version: 0.0.7
      - name: ack-ec2-controller.v0.0.6
        version: 0.0.6
      - name: ack-ec2-controller.v0.0.5
        version: 0.0.5
      - name: ack-ec2-controller.v0.0.4
        version: 0.0.4
      - name: ack-ec2-controller.v0.0.3
        version: 0.0.3
      - name: ack-ec2-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-ec2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: global-load-balancer-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: global-load-balancer-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSRecord",
                "metadata": {
                  "name": "hello-global-record"
                },
                "spec": {
                  "endpoints": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    }
                  ],
                  "globalZoneRef": {
                    "name": "external-dns-zone"
                  },
                  "loadBalancingPolicy": "Multivalue",
                  "name": "hello.my.base.domain",
                  "ttl": 60
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSRecord",
                "metadata": {
                  "name": "route53-geoproximity-global-record-with-healthcheck"
                },
                "spec": {
                  "endpoints": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    }
                  ],
                  "globalZoneRef": {
                    "name": "route53-global-dns-zone"
                  },
                  "healthCheck": {
                    "failureThreshold": 3,
                    "httpGet": {
                      "host": "geoproximity-hc.${global_base_domain}",
                      "path": "/healthz",
                      "port": 80,
                      "scheme": "HTTP"
                    },
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 2
                  },
                  "loadBalancingPolicy": "Geoproximity",
                  "name": "geoproximity-hc.my.base.domain",
                  "ttl": 60
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSZone",
                "metadata": {
                  "name": "external-dns-zone"
                },
                "spec": {
                  "domain": "myglobalzone.mybasedomain.io",
                  "provider": {
                    "externalDNS": {
                      "annotations": {
                        "type": "global"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSZone",
                "metadata": {
                  "name": "route53-global-dns-zone"
                },
                "spec": {
                  "domain": "myglobalzone.mybasedomain.io",
                  "provider": {
                    "route53": {
                      "credentialsSecretRef": {
                        "name": "route53-global-zone-credentials",
                        "namespace": "mynamespace"
                      },
                      "zoneID": "Z083892239ROKNEP70BDY"
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalRouteDiscovery",
                "metadata": {
                  "name": "external-dns-globalroutediscovery"
                },
                "spec": {
                  "clusters": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3"
                    }
                  ],
                  "defaultLoadBalancingPolicy": "Multivalue",
                  "defaultTTL": 30,
                  "globalZoneRef": {
                    "name": "external-dns-zone"
                  },
                  "routeSelector": {
                    "matchLabels": {
                      "route-type": "global"
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalRouteDiscovery",
                "metadata": {
                  "name": "route53-globalroutediscovery"
                },
                "spec": {
                  "clusters": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3"
                    }
                  ],
                  "defaultLoadBalancingPolicy": "Multivalue",
                  "defaultTTL": 30,
                  "globalZoneRef": {
                    "name": "route53-global-dns-zone"
                  },
                  "routeSelector": {
                    "matchLabels": {
                      "route-type": "global"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Cloud Provider
          certified: "false"
          containerImage: quay.io/redhat-cop/global-load-balancer-operator@sha256:c0b6ac9b03571a74a56b28302d720e8aa4442233f7867d05ed365e022f7cb8f5
          createdAt: "2021-12-02T17:33:23Z"
          description: This operator creates automation around a DNS to operate as
            a global load balancer for a set of OpenShift clusters.
          operatorframework.io/suggested-namespace: global-load-balancer-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/global-load-balancer-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DNSEndpoint
            name: dnsendpoints.externaldns.k8s.io
            version: v1alpha1
          - description: GlobalDNSRecord is the Schema for the globaldnsrecords API
            displayName: Global DNSRecord
            kind: GlobalDNSRecord
            name: globaldnsrecords.redhatcop.redhat.io
            version: v1alpha1
          - description: GlobalDNSZone is the Schema for the globaldnszones API
            displayName: Global DNSZone
            kind: GlobalDNSZone
            name: globaldnszones.redhatcop.redhat.io
            version: v1alpha1
          - description: GlobalRouteDiscovery is the Schema for the globalroutediscoveries
              API
            displayName: Global Route Discovery
            kind: GlobalRouteDiscovery
            name: globalroutediscoveries.redhatcop.redhat.io
            version: v1alpha1
        description: "The global-load-balancer-operator implements automation to program
          a DNS service to act as global load balancer for applications deployed to
          multiple OpenShift clusters.\nThis operator is designed to be deployed to
          a control cluster which will watch the load balanced clusters (controlled
          clusters).\nThere are two main concepts (APIs) provided by this operator:\n\n1.
          GlobalDNSZone\n2. GlobalDNSRecord\n\n## GlobalDNSZone\n\nThe `GlobalDNSZone`
          CR allows you to configure a zone which will contain global load balanced
          records and the provider used to populate it.\nHere is an example of GlobalDNSZone:\n\n```yaml\napiVersion:
          redhatcop.redhat.io/v1alpha1\nkind: GlobalDNSZone\nmetadata:\n  name: external-dns-zone\nspec:\n
          \ # Add fields here\n  domain: global.myzone.io\n  provider:\n    externalDNS:\n
          \     annotations:\n        type: global\n```\n\nHere is a table summarizing
          the supported providers and their capabilities:\n\n| Provider  | Supports
          Health Checks  | Supports Multivalue LB | Supports Latency LB  | Supports
          GeoProximity LB  |\n|:--:|:--:|:---:|:---:|:---:|\n| External-dns  | no
          \ | yes | no  | no  |\n| Route53  | yes | yes(*)  | yes(*)  | yes(*)  |\n\n(*)
          only if all controlled clusters run on AWS.\n\n## GlobalDNSRecord\n\nThe
          `GlobalDNSRecord` CR allows you to specify the intention to create a global
          dns record. Here is an example\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GlobalDNSRecord\nmetadata:\n  name: hello-global-record\nspec:\n  name:
          hello.global.myzone.io\n  endpoints:\n  - clusterName: cluster1\n    clusterCredentialRef:\n
          \     name: kubeconfig\n      namespace: cluster1\n    loadBalancerServiceRef:\n
          \     name: router-default\n      namespace: openshift-ingress\n  - clusterName:
          cluster2\n    clusterCredentialRef:\n      name: kubeconfig\n      namespace:
          cluster2\n    loadBalancerServiceRef:\n      name: router-default\n      namespace:
          openshift-ingress\n  - clusterName: cluster3\n    clusterCredentialRef:\n
          \     name: kubeconfig\n      namespace: cluster3\n    loadBalancerServiceRef:\n
          \     name: router-default\n      namespace: openshift-ingress\n  ttl: 60\n
          \ loadBalancingPolicy: Multivalue\n  globalZoneRef:\n    name: external-dns-zone\n
          \ healthCheck:\n    failureThreshold: 3\n    httpGet:\n      host: hello.global.myzone.io\n
          \     httpHeaders:\n        - name: ciao\n          value: hello\n      path:
          /readyz\n      port: 80\n      scheme: HTTP\n    periodSeconds: 10\n    successThreshold:
          1\n    timeoutSeconds: 1     \n```\n\nThe list of endpoints identifies the
          set of `LoadBalancer` type of services to watch on the remote clusters (at
          the moment only LoadBalancer services are supported). These are the `LoadBalancer`
          services created by the [ingress controller](https://docs.openshift.com/container-platform/4.5/networking/ingress-operator.html)
          on which the routers rely. Each entry of this list requires a reference
          to the loadbalancer service on the remote cluster and a reference to a local
          secret containing the credential to connect to the remote cluster.\n\nThis
          secret must contain one entry called `kubeconfig` with a kubeconfig file
          with a default context pointing to the remote cluster. The account used
          by the kubeconfig file (presumably a service account) will require at a
          minimum cluster-level permissions as described in this [cluster role](./docs/scripts/remote-account-cluster-role.yaml).\n\nThe
          `globalZoneRef` field refers to a local (same namespace) `GlobalZone` CR.
          The DNS record represented by this `GlobalDNSRecord`, will be created in
          the referenced zone.\n\n`ttl` is the TTL of the crated DNS record.\n\n`loadBalancingPolicy`
          is the load balancing policy for this global record. It must match one of
          the policy supported by the provider of the referenced `GlobalZone`.\n\nFinally,
          `healthcheck` represent a [probe](https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.18/#probe-v1-core)
          to be used to test the health of a record. This field will be ignored if
          the provider does not support health checks.\n\n## External DNS Provider\n\nThe
          [external-dns](https://github.com/kubernetes-sigs/external-dns) provider
          delegates to external-dns the creation of the actual DNS records by creating
          a `DNSEndpoint` object.\nThe `DNSEndpoint` object will be created in the
          same namespace as the `GlobalDNSRecord` and will be owned by it.\nThe `DNSEdnpoint`
          object will have the same labels as the `GlobalDNSRecord` and the annotations
          specified in the GlobalDNSZone configuration.\nExternal-dns should be configured
          to watch for DNSEnpoints at the cluster level and to point to the desired
          provider.\nDetails on configuration can be found at the external-dns git
          repository.\nThe External-dns should be used as a fall back option when
          other options are not available as it does not support health checks and
          advanced load balancing policies.\n\n## AWS Route53 provider\n\nAWS Route53
          provider uses the Route53 service as a global loadbalancer and offers advanced
          routing capabilities via [route53 traffic policies](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/traffic-flow.html)
          (note that traffic policies will trigger an expense).\nThe following routing
          polices are currently supported:\n\n1. [Multivalue](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-multivalue)\n2.
          [Geoproximity](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-geoproximity)\n3.
          [Latency](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-latency)\n\nAWS
          Route53 provider at the moment requires that all the controlled clusters
          run in AWS.\n\nIf health checks are defined, a route53 health check originating
          from any reason (you have to ensure connectivity) will be created for each
          of the endpoint. Because the endpoint represent s a shared ELB (shared with
          other apps, that is) and the health check is app specific, we cannot use
          the ELB health check, so the route53 endpoint is created with one of the
          two IP exposed by the ELB. This is suboptimal, but it works in most situations.\n\n##
          Global Route Auto Discovery\n\nThe aboev examples showed how to create global
          DNS records. This can be good in some situations, but most of the times
          in an openshift deployment global DNS records will point to routes that
          are intended to be global.\nThe global-load-balancer operator can auto-discover
          these routes and automatically create the corresponding `GloablDNSRecord`.
          The `GlobalRouteDiscovery` CRD is used to configure the discovery process,
          here is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GlobalRouteDiscovery\nmetadata:\n  name: route53-globalroutediscovery\nspec:\n
          \ clusters:\n  - clusterName: cluster1\n    clusterCredentialRef:\n      name:
          ${cluster1_secret_name}\n      namespace: cluster1\n...\n  routeSelector:\n
          \   matchLabels:\n      route-type: global\n  defaultLoadBalancingPolicy:
          Multivalue\n  defaultTTL: 30 \n  globalZoneRef:\n    name: route53-global-dns-zone\n```\n\nThis
          global discovery route will discover routes in the provided list of cluster.
          Only the routes that match the route selector will be considered global.
          \nThe default load balancing policy and default TTL can be expressed in
          the `GlobalRouteDiscovery` CR. However with the following annotations, it's
          possible to configure route-specific values:\n\n* `global-load-balancer-operator.redhat-cop.io/load-balancing-policy`
          to set the load balancing policy\n* `global-load-balancer-operator.redhat-cop.io/ttl`
          to set the TTL\n\nThe globalZoneRef refers to the global zone to be used
          for the created `GlobalDNSRecords`.\n\nHealth checks will also be automatically
          discovered. If the pods behind the route expose a readiness check of httpGet
          kind, that configuration will be used to create the GlobalDNSRecord health
          check.\nWhen more than one container is present in the pod, by default the
          first container will be examined for health check. This behavior can be
          overridden with the this annotation on the route: `global-load-balancer-operator.redhat-cop.io/container-probe`
          where the value will container the name of the container with teh correct
          readiness probe.\n\nIf routes with the same namespace and name exists in
          multiple cluster, the following conditions must be met:\n\n* all host names
          must be the same\n* all load balancing policy must be the same\n* all TTLs
          must be the same\n* all discovered readiness checks must be the same\n\n##
          Examples\n\nThese examples are intended to help you setting up working configuration
          with each of the providers\n\n### Cluster Set up\n\nTwo approaches for cluster
          setup are provided\n\n1. [One cluster, three ingress-gateways.](https://github.com/redhat-cop/global-load-balancer-operator/docs/one-cluster-three-ingress-gateways.md)
          This approach is intended for development purposes and has the objective
          to keep resource consumption at the minimum.\n2. [Control cluster and three
          controlled clusters in different regions.](https://github.com/redhat-cop/global-load-balancer-operator/docs/three-clusters.md).
          This approach represents a more realistic set-up albeit it consumes more
          resources.\n\nYou can also set up the cluster on your own, at the end the
          following conditions must be met:\n\nThree namespace `cluster1` `cluster2`
          `cluster3` are created.\nthe following environment variables are initialized
          for each cluster:\n\n1. <cluster>_secret_name. Pointing to a secret in each
          of the cluster namespaces containing a valid kubeconfig fot that cluster\n2.
          <cluster>_service_name.  Pointing to the name of the loadbalancer service
          to be used for that cluster.\n3. <cluster>_service_namespace. Pointing to
          the namespace of the loadbalancer service to be used for that cluster.\n\nHere
          are examples for the supported provider:\n\n1. [Setting up external-dns
          as provider](https://github.com/redhat-cop/global-load-balancer-operator/docs/external-dns-provider.md)\n2.
          [Setting up route53 as a provider](https://github.com/redhat-cop/global-load-balancer-operator/docs/aws-route53-provider.md)\n"
        displayName: Global Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Networking
        - Cloud Provider
        - Global Load Balancing
        - GLB
        - GTM
        - route53
        - DNS
        links:
        - name: GigHub Repository
          url: https://github.com/redhat-cop/global-load-balancer-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/global-load-balancer-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/global-load-balancer-operator@sha256:c0b6ac9b03571a74a56b28302d720e8aa4442233f7867d05ed365e022f7cb8f5
        version: 1.4.2
      entries:
      - name: global-load-balancer-operator.v1.4.2
        version: 1.4.2
      - name: global-load-balancer-operator.v1.4.1
        version: 1.4.1
      - name: global-load-balancer-operator.v1.4.0
        version: 1.4.0
      - name: global-load-balancer-operator.v1.3.0
        version: 1.3.0
      - name: global-load-balancer-operator.v1.2.0
        version: 1.2.0
      - name: global-load-balancer-operator.v1.1.0
        version: 1.1.0
      - name: global-load-balancer-operator.v1.0.7
        version: 1.0.7
      - name: global-load-balancer-operator.v1.0.6
        version: 1.0.6
      - name: global-load-balancer-operator.v1.0.5
        version: 1.0.5
      - name: global-load-balancer-operator.v1.0.4
        version: 1.0.4
      - name: global-load-balancer-operator.v1.0.3
        version: 1.0.3
      - name: global-load-balancer-operator.v1.0.2
        version: 1.0.2
      - name: global-load-balancer-operator.v1.0.1
        version: 1.0.1
      - name: global-load-balancer-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: global-load-balancer-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Netscaler
      provider-url: https://www.netscaler.com
    name: netscaler-adm-agent-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netscaler-adm-agent-operator.v141.37.44
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "AdmAgent",
                "metadata": {
                  "name": "admagent"
                },
                "spec": {
                  "accessSecret": "",
                  "admURL": "adm.cloud.com",
                  "affinity": {},
                  "clusterPrefix": "",
                  "core": {
                    "livenessProbe": {
                      "failureThreshold": 5,
                      "initialDelaySeconds": 600,
                      "periodSeconds": 60,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "readinessProbe": {
                      "failureThreshold": 30,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 30,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "100m",
                        "memory": "2000Mi"
                      }
                    }
                  },
                  "fullnameOverride": "",
                  "helper": {
                    "image": "{{ .Values.helper.imageRegistry }}/{{ .Values.helper.imageRepository }}:{{ .Values.helper.imageTag }}",
                    "imageRegistry": "quay.io",
                    "imageRepository": "citrix/adm-agent-helper",
                    "imageTag": "2.0.1",
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "200Mi"
                      }
                    }
                  },
                  "hpa": {
                    "maxReplicas": 10,
                    "minReplicas": 1,
                    "required": false,
                    "targetCPUUtilization": "1000m",
                    "targetMemoryUtilization": "4000Mi"
                  },
                  "image": "{{ .Values.imageRegistry }}/{{ .Values.imageRepository }}:{{ .Values.imageTag }}",
                  "imagePullSecrets": [],
                  "imageRegistry": "quay.io",
                  "imageRepository": "citrix/adm-agent",
                  "imageTag": "14.1-37.44",
                  "kad": {
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "50m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "lic": {
                    "livenessProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 600,
                      "periodSeconds": 20,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "100m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "loginSecret": "",
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "redis": {
                    "livenessProbe": {
                      "failureThreshold": 5,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "readinessProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 20,
                      "periodSeconds": 10,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true
                  },
                  "sharding": {
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-adm-agent-operator@sha256:1c462c2d6161877f87d5eed4888f8d94c733501e67a94757ab0873d983fb3a4a
          createdAt: "2024-11-18T06:40:58Z"
          description: This is an operator to install ADM Agent
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler ADM Agent Operator
            displayName: NetScaler ADM Agent Operator
            kind: AdmAgent
            name: admagents.netscaler.com
            version: v1
        description: Netscaler ADM Agent Operator
        displayName: Netscaler ADM Agent Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        - ADC
        - Kubernetes
        - OpenShift
        - Citrix
        - Controller
        - Observability
        - Metrices
        - adm
        - agent
        - monitoring
        - proxy
        links:
        - name: Netscaler Adm Agent Operator
          url: https://github.com/netscaler/netscaler-helm-charts/blob/master/adm-agent/README.md
        maintainers:
        - email: pavan.belani@cloud.com
          name: Pavan Belani
        - email: swapnil.khanderaokakade@cloud.com
          name: Swapnil Khanderao
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Netscaler
          url: https://www.netscaler.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6b8c101124579dd46086ec83715618dd9a6aebe66176d676242e6c0b0900a53c
        - registry.connect.redhat.com/citrix/netscaler-adm-agent-operator@sha256:1c462c2d6161877f87d5eed4888f8d94c733501e67a94757ab0873d983fb3a4a
        - quay.io/citrix/adm-agent@sha256:45cab90524687c76a6d53f014b55063fa25c8aaeec0261f0f4f0efa6bff42968
        - quay.io/citrix/adm-agent-helper@sha256:f567a0637e8024f9d1674977f6c013741cc0fd12cf78703f8dbe01126561c3c4
        version: 141.37.44
      entries:
      - name: netscaler-adm-agent-operator.v141.37.44
        version: 141.37.44
      - name: netscaler-adm-agent-operator.v141.31.38
        version: 141.31.38
      - name: netscaler-adm-agent-operator.v141.25.33
        version: 141.25.33
      - name: netscaler-adm-agent-operator.v141.23.31
        version: 141.23.31
      - name: netscaler-adm-agent-operator.v141.21.41
        version: 141.21.41
      - name: netscaler-adm-agent-operator.v141.16.39
        version: 141.16.39
      - name: netscaler-adm-agent-operator.v141.13.23
        version: 141.13.23
      name: stable
    defaultChannel: stable
    packageName: netscaler-adm-agent-operator
    provider:
      name: Netscaler
      url: https://www.netscaler.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Redis
      provider-url: ""
    name: redis-enterprise-operator-cert-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-enterprise-operator.v6.2.10-45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:ffdf10a1b08115fd4de06f1cd751d36644a9bc9a8c11034cd432e9b66a5d2f73
        - registry.connect.redhat.com/redislabs/services-manager@sha256:8b46f2aacfb3e1b243c0b64975accc55fd80b4de9c9ac4769c3c67a5676addc5
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
        version: 6.2.10-45
      entries:
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: 6.2.10
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.22.0
    - currentCSV: redis-enterprise-operator.v7.22.0-11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:088e4b57c6e1ca1680b359078e5eb1bba67343cdbd9a539af26a0ea23a830593
        - registry.connect.redhat.com/redislabs/services-manager@sha256:3b96cdc65c3e42fd22eb165950be96b907725903ac74fe4f82cfa0073e28e95a
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:d56ba97e95e799210e76615095409143a5dddd5c5a1e1be1d551d7e3f9e4d3ef
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:0cfb07f774db131b5b74e5aa91291579c916c4d4cd1a693d7aaed554f0177a68
        version: 7.22.0-11.3
      entries:
      - name: redis-enterprise-operator.v7.22.0-11.3
        version: 7.22.0-11.3
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.22.0-11.3
    - currentCSV: redis-enterprise-operator.v7.4.6-7.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:473b5661049a6c8830ba113504975726154c7b172376ba3b63d891674d59c78f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5f6f32a63f4dd8be85b9d469f0fedd26d6aa76e32b7735bfa7ab7203cb272cdb
        - registry.connect.redhat.com/redislabs/services-manager@sha256:ba20ab644a473bc742af671171d5313537292d88942b16336d03a6386c061d8a
        version: 7.4.6-7.2
      entries:
      - name: redis-enterprise-operator.v7.4.6-7.2
        version: 7.4.6-7.2
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      name: 7.4.6
    - currentCSV: redis-enterprise-operator.v7.8.2-6.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:36c7a36dabb23a4c003e112085cda2873296bfb22e5d11b470d7c1aba55b7f0f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:c6e1f2109a0dac76b51c25c1cbf6030b4894b21b98147e2ad1db2dd8e7693576
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:acf817c90b5f03bc2796fbc305cf43ce5c0644dcbe1b685752090b3f600a0f7b
        version: 7.8.2-6.8
      entries:
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.2
    - currentCSV: redis-enterprise-operator.v7.8.4-9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:803be04534e59cdc7b8898b3bb336676aad1cb77cf266051e226be82679fcf6a
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:4ba0c414a74268bed9c6757e447b340db7c912afa864cfb096ddf88f11feeaf9
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:66ec124da16cc9d0abd66687e088be0b358e20cfd48f5774d34bf5874e6a4e65
        version: 7.8.4-9.0
      entries:
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.4
    - currentCSV: redis-enterprise-operator.v7.8.6-5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:1fff01daece34aeb7befcb0760d5112c8c7d261f2b86388ece416ac94076a573
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5125e83d3c93b224d1983c599994e43899a0ba0c6f2fc4eed31476f755b5fc0b
        - registry.connect.redhat.com/redislabs/services-manager@sha256:88d483f5e37b5a901d1ae3d965304edc000c172d2e5e307297d5b7bcd77ae73b
        version: 7.8.6-5.1
      entries:
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.6
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: production
    defaultChannel: production
    packageName: redis-enterprise-operator-cert-rhmp
    provider:
      name: Redis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc
      provider-url: ""
    name: bare-metal-event-relay
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bare-metal-event-relay.v4.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "event.redhat-cne.org/v1alpha1",
                "kind": "HardwareEvent",
                "metadata": {
                  "name": "hardwareevent-sample"
                },
                "spec": {
                  "nodeSelector": {
                    "node-role.kubernetes.io/worker": ""
                  },
                  "storageType": "emptyDir",
                  "transportHost": "http://hw-event-publisher-service.openshift-bare-metal-events.svc.cluster.local:9043"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/openshift4/baremetal-hardware-event-proxy-rhel9@sha256:aaaea040164e1fd801e13c74024b2307997cdde954c1a5e35efd90023d2d7376
          createdAt: "2023-02-23"
          description: This software manages the lifecycle of the hw-event-proxy container.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.22.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/hw-event-proxy
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HardwareEvent is the Schema for the hardwareevents API
            displayName: Hardware Event
            kind: HardwareEvent
            name: hardwareevents.event.redhat-cne.org
            version: v1alpha1
        description: |
          ## Funtionality
          Install and manage a hardware event delivery framework.
          ## Installation
          To start using Hardware event proxy, install the operator and then create the following HardwareEvent:
          ```
          apiVersion: "event.redhat-cne.org/v1alpha1"
          kind: "HardwareEvent"
          metadata:
            name: "hardware-event"
          spec:
            nodeSelector:
              node-role.kubernetes.io/worker: ""
            transportHost: "http://hw-event-publisher-service.openshift-bare-metal-events.svc.cluster.local:9043"

          ```
          This operator needs a secret to be created to access Redfish
          Message Registry.
          The secret name and spec the operand expects are under the same namespace
          the operand is deployed on. For example:
          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: redfish-basic-auth
          type: Opaque
          stringData:
            username: <bmc_username>
            password: <bmc_password>
            # BMC host DNS or IP address
            hostaddr: <bmc_host_ip_address>
          ```
        displayName: Bare Metal Event Relay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hardware
        - event
        - cloudevents
        links:
        - name: Hardware Event Proxy Operator
          url: https://github.com/redhat-cne/hw-event-proxy-operator
        maturity: alpha
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:6b7beed3cba9093b3aeb3109f23ec38c20f5568915dff774ad5e9e6bb5db3a58
        - registry.redhat.io/openshift4/bare-metal-event-relay-rhel9-operator@sha256:2219f2f623ee27ef83518d28857995287de0a70cefa193792f28acb352ed8fc0
        - registry.redhat.io/openshift4/baremetal-hardware-event-proxy-rhel9@sha256:aaaea040164e1fd801e13c74024b2307997cdde954c1a5e35efd90023d2d7376
        - registry.redhat.io/openshift4/ose-cloud-event-proxy-rhel9@sha256:c99b448820cf03a33e9237030ac74c1d1ad6a4ce17f1b2840c69b039097c5733
        version: 4.16.1
      entries:
      - name: bare-metal-event-relay.v4.16.1
        version: 4.16.1
      name: "4.16"
    - currentCSV: bare-metal-event-relay.v4.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "event.redhat-cne.org/v1alpha1",
                "kind": "HardwareEvent",
                "metadata": {
                  "name": "hardwareevent-sample"
                },
                "spec": {
                  "nodeSelector": {
                    "node-role.kubernetes.io/worker": ""
                  },
                  "storageType": "emptyDir",
                  "transportHost": "http://hw-event-publisher-service.openshift-bare-metal-events.svc.cluster.local:9043"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/openshift4/baremetal-hardware-event-proxy-rhel9@sha256:aaaea040164e1fd801e13c74024b2307997cdde954c1a5e35efd90023d2d7376
          createdAt: "2023-02-23"
          description: This software manages the lifecycle of the hw-event-proxy container.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.22.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/hw-event-proxy
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HardwareEvent is the Schema for the hardwareevents API
            displayName: Hardware Event
            kind: HardwareEvent
            name: hardwareevents.event.redhat-cne.org
            version: v1alpha1
        description: |
          ## Funtionality
          Install and manage a hardware event delivery framework.
          ## Installation
          To start using Hardware event proxy, install the operator and then create the following HardwareEvent:
          ```
          apiVersion: "event.redhat-cne.org/v1alpha1"
          kind: "HardwareEvent"
          metadata:
            name: "hardware-event"
          spec:
            nodeSelector:
              node-role.kubernetes.io/worker: ""
            transportHost: "http://hw-event-publisher-service.openshift-bare-metal-events.svc.cluster.local:9043"

          ```
          This operator needs a secret to be created to access Redfish
          Message Registry.
          The secret name and spec the operand expects are under the same namespace
          the operand is deployed on. For example:
          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: redfish-basic-auth
          type: Opaque
          stringData:
            username: <bmc_username>
            password: <bmc_password>
            # BMC host DNS or IP address
            hostaddr: <bmc_host_ip_address>
          ```
        displayName: Bare Metal Event Relay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hardware
        - event
        - cloudevents
        links:
        - name: Hardware Event Proxy Operator
          url: https://github.com/redhat-cne/hw-event-proxy-operator
        maturity: alpha
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:6b7beed3cba9093b3aeb3109f23ec38c20f5568915dff774ad5e9e6bb5db3a58
        - registry.redhat.io/openshift4/bare-metal-event-relay-rhel9-operator@sha256:2219f2f623ee27ef83518d28857995287de0a70cefa193792f28acb352ed8fc0
        - registry.redhat.io/openshift4/ose-cloud-event-proxy-rhel9@sha256:c99b448820cf03a33e9237030ac74c1d1ad6a4ce17f1b2840c69b039097c5733
        - registry.redhat.io/openshift4/baremetal-hardware-event-proxy-rhel9@sha256:aaaea040164e1fd801e13c74024b2307997cdde954c1a5e35efd90023d2d7376
        version: 4.16.1
      entries:
      - name: bare-metal-event-relay.v4.16.1
        version: 4.16.1
      name: stable
    defaultChannel: stable
    packageName: bare-metal-event-relay
    provider:
      name: Red Hat Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ocs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocs-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |2-

            [
                {
                    "apiVersion": "ocs.openshift.io/v1",
                    "kind": "StorageCluster",
                    "metadata": {
                        "name": "example-storagecluster",
                        "namespace": "openshift-storage"
                    },
                    "spec": {
                        "manageNodes": false,
                        "monPVCTemplate": {
                            "spec": {
                                "accessModes": [
                                    "ReadWriteOnce"
                                ],
                                "resources": {
                                    "requests": {
                                        "storage": "10Gi"
                                    }
                                },
                                "storageClassName": "gp2-csi"
                            }
                        },
                        "storageDeviceSets": [
                            {
                                "count": 3,
                                "dataPVCTemplate": {
                                    "spec": {
                                        "accessModes": [
                                            "ReadWriteOnce"
                                        ],
                                        "resources": {
                                            "requests": {
                                                "storage": "1Ti"
                                            }
                                        },
                                        "storageClassName": "gp2-csi",
                                        "volumeMode": "Block"
                                    }
                                },
                                "name": "example-deviceset",
                                "placement": {},
                                "portable": true,
                                "resources": {}
                            }
                        ]
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Storage
          containerImage: registry.redhat.io/odf4/ocs-rhel9-operator@sha256:b1a9492bd9b314f86b4e192cf2c83144885ff863d9932e386a41691f50be6193
          description: Red Hat OpenShift Container Storage provides hyperconverged
            storage for applications within an OpenShift cluster.
          external.features.ocs.openshift.io/export-script: |-
            IiIiCkNvcHlyaWdodCAyMDIwIFRoZSBSb29rIEF1dGhvcnMuIEFsbCByaWdodHMgcmVzZXJ2ZWQu
            CgpMaWNlbnNlZCB1bmRlciB0aGUgQXBhY2hlIExpY2Vuc2UsIFZlcnNpb24gMi4wICh0aGUgIkxp
            Y2Vuc2UiKTsKeW91IG1heSBub3QgdXNlIHRoaXMgZmlsZSBleGNlcHQgaW4gY29tcGxpYW5jZSB3
            aXRoIHRoZSBMaWNlbnNlLgpZb3UgbWF5IG9idGFpbiBhIGNvcHkgb2YgdGhlIExpY2Vuc2UgYXQK
            CglodHRwOi8vd3d3LmFwYWNoZS5vcmcvbGljZW5zZXMvTElDRU5TRS0yLjAKClVubGVzcyByZXF1
            aXJlZCBieSBhcHBsaWNhYmxlIGxhdyBvciBhZ3JlZWQgdG8gaW4gd3JpdGluZywgc29mdHdhcmUK
            ZGlzdHJpYnV0ZWQgdW5kZXIgdGhlIExpY2Vuc2UgaXMgZGlzdHJpYnV0ZWQgb24gYW4gIkFTIElT
            IiBCQVNJUywKV0lUSE9VVCBXQVJSQU5USUVTIE9SIENPTkRJVElPTlMgT0YgQU5ZIEtJTkQsIGVp
            dGhlciBleHByZXNzIG9yIGltcGxpZWQuClNlZSB0aGUgTGljZW5zZSBmb3IgdGhlIHNwZWNpZmlj
            IGxhbmd1YWdlIGdvdmVybmluZyBwZXJtaXNzaW9ucyBhbmQKbGltaXRhdGlvbnMgdW5kZXIgdGhl
            IExpY2Vuc2UuCiIiIgoKaW1wb3J0IGVycm5vCmltcG9ydCBzeXMKaW1wb3J0IGpzb24KaW1wb3J0
            IGFyZ3BhcnNlCmltcG9ydCByZQppbXBvcnQgc3VicHJvY2VzcwppbXBvcnQgaG1hYwpmcm9tIGhh
            c2hsaWIgaW1wb3J0IHNoYTEgYXMgc2hhCmZyb20gb3MgaW1wb3J0IGxpbmVzZXAgYXMgTElORVNF
            UApmcm9tIG9zIGltcG9ydCBwYXRoCmZyb20gZW1haWwudXRpbHMgaW1wb3J0IGZvcm1hdGRhdGUK
            aW1wb3J0IHJlcXVlc3RzCmZyb20gcmVxdWVzdHMuYXV0aCBpbXBvcnQgQXV0aEJhc2UKCnB5M2sg
            PSBGYWxzZQppZiBzeXMudmVyc2lvbl9pbmZvLm1ham9yID49IDM6CiAgICBweTNrID0gVHJ1ZQog
            ICAgaW1wb3J0IHVybGxpYi5wYXJzZQogICAgZnJvbSBpcGFkZHJlc3MgaW1wb3J0IGlwX2FkZHJl
            c3MsIElQdjRBZGRyZXNzCgpNb2R1bGVOb3RGb3VuZEVycm9yID0gSW1wb3J0RXJyb3IKCnRyeToK
            ICAgIGltcG9ydCByYWRvcwpleGNlcHQgTW9kdWxlTm90Rm91bmRFcnJvciBhcyBub01vZEVycjoK
            ICAgIHByaW50KGYiRXJyb3I6IHtub01vZEVycn1cbkV4aXRpbmcgdGhlIHNjcmlwdC4uLiIpCiAg
            ICBzeXMuZXhpdCgxKQoKdHJ5OgogICAgaW1wb3J0IHJiZApleGNlcHQgTW9kdWxlTm90Rm91bmRF
            cnJvciBhcyBub01vZEVycjoKICAgIHByaW50KGYiRXJyb3I6IHtub01vZEVycn1cbkV4aXRpbmcg
            dGhlIHNjcmlwdC4uLiIpCiAgICBzeXMuZXhpdCgxKQoKdHJ5OgogICAgIyBmb3IgMi43LngKICAg
            IGZyb20gU3RyaW5nSU8gaW1wb3J0IFN0cmluZ0lPCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9y
            OgogICAgIyBmb3IgMy54CiAgICBmcm9tIGlvIGltcG9ydCBTdHJpbmdJTwoKdHJ5OgogICAgIyBm
            b3IgMi43LngKICAgIGZyb20gdXJscGFyc2UgaW1wb3J0IHVybHBhcnNlCiAgICBmcm9tIHVybGxp
            YiBpbXBvcnQgdXJsZW5jb2RlIGFzIHVybGVuY29kZQpleGNlcHQgTW9kdWxlTm90Rm91bmRFcnJv
            cjoKICAgICMgZm9yIDMueAogICAgZnJvbSB1cmxsaWIucGFyc2UgaW1wb3J0IHVybHBhcnNlCiAg
            ICBmcm9tIHVybGxpYi5wYXJzZSBpbXBvcnQgdXJsZW5jb2RlIGFzIHVybGVuY29kZQoKdHJ5Ogog
            ICAgZnJvbSBiYXNlNjQgaW1wb3J0IGVuY29kZXN0cmluZwpleGNlcHQ6CiAgICBmcm9tIGJhc2U2
            NCBpbXBvcnQgZW5jb2RlYnl0ZXMgYXMgZW5jb2Rlc3RyaW5nCgoKY2xhc3MgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbihFeGNlcHRpb24pOgogICAgcGFzcwoKCiMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMgRHVtbXlSYWRv
            cyAjIyMjIyMjIyMjIyMjIyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjCiMgdGhpcyBpcyBtYWlubHkgZm9yIHRlc3RpbmcgYW5kIGNvdWxkIGJlIHVz
            ZWQgd2hlcmUgJ3JhZG9zJyBpcyBub3QgYXZhaWxhYmxlCgoKY2xhc3MgRHVtbXlSYWRvcyhvYmpl
            Y3QpOgogICAgZGVmIF9faW5pdF9fKHNlbGYpOgogICAgICAgIHNlbGYucmV0dXJuX3ZhbCA9IDAK
            ICAgICAgICBzZWxmLmVycl9tZXNzYWdlID0gIiIKICAgICAgICBzZWxmLnN0YXRlID0gImNvbm5l
            Y3RlZCIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwID0ge30KICAgICAgICBzZWxmLmNtZF9u
            YW1lcyA9IHt9CiAgICAgICAgc2VsZi5faW5pdF9jbWRfb3V0cHV0X21hcCgpCiAgICAgICAgc2Vs
            Zi5kdW1teV9ob3N0X2lwX21hcCA9IHt9CgogICAgZGVmIF9pbml0X2NtZF9vdXRwdXRfbWFwKHNl
            bGYpOgogICAgICAgIGpzb25fZmlsZV9uYW1lID0gInRlc3QtZGF0YS9jZXBoLXN0YXR1cy1vdXQi
            CiAgICAgICAgc2NyaXB0X2RpciA9IHBhdGguYWJzcGF0aChwYXRoLmRpcm5hbWUoX19maWxlX18p
            KQogICAgICAgIGNlcGhfc3RhdHVzX3N0ciA9ICIiCiAgICAgICAgd2l0aCBvcGVuKAogICAgICAg
            ICAgICBwYXRoLmpvaW4oc2NyaXB0X2RpciwganNvbl9maWxlX25hbWUpLCBtb2RlPSJyIiwgZW5j
            b2Rpbmc9IlVURi04IgogICAgICAgICkgYXMganNvbl9maWxlOgogICAgICAgICAgICBjZXBoX3N0
            YXR1c19zdHIgPSBqc29uX2ZpbGUucmVhZCgpCiAgICAgICAgc2VsZi5jbWRfbmFtZXNbImZzIGxz
            Il0gPSAiIiJ7ImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJmcyBscyJ9IiIiCiAgICAgICAg
            c2VsZi5jbWRfbmFtZXNbInF1b3J1bV9zdGF0dXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3Jt
            YXQiOiAianNvbiIsICJwcmVmaXgiOiAicXVvcnVtX3N0YXR1cyJ9IiIiCiAgICAgICAgKQogICAg
            ICAgIHNlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXSA9ICgKICAgICAgICAgICAgIiIieyJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAibWdyIHNlcnZpY2VzIn0iIiIKICAgICAgICApCiAg
            ICAgICAgIyBhbGwgdGhlIGNvbW1hbmRzIGFuZCB0aGVpciBvdXRwdXQKICAgICAgICBzZWxmLmNt
            ZF9vdXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJmcyBscyJdXSA9ICgKICAgICAgICAgICAgIiIi
            W3sibmFtZSI6Im15ZnMiLCJtZXRhZGF0YV9wb29sIjoibXlmcy1tZXRhZGF0YSIsIm1ldGFkYXRh
            X3Bvb2xfaWQiOjIsImRhdGFfcG9vbF9pZHMiOlszXSwiZGF0YV9wb29scyI6WyJteWZzLXJlcGxp
            Y2F0ZWQiXX1dIiIiCiAgICAgICAgKQogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbc2VsZi5j
            bWRfbmFtZXNbInF1b3J1bV9zdGF0dXMiXV0gPSAoCiAgICAgICAgICAgICIiInsiZWxlY3Rpb25f
            ZXBvY2giOjMsInF1b3J1bSI6WzBdLCJxdW9ydW1fbmFtZXMiOlsiYSJdLCJxdW9ydW1fbGVhZGVy
            X25hbWUiOiJhIiwicXVvcnVtX2FnZSI6MTQzODUsImZlYXR1cmVzIjp7InF1b3J1bV9jb24iOiI0
            NTQwMTM4MjkyODM2Njk2MDYzIiwicXVvcnVtX21vbiI6WyJrcmFrZW4iLCJsdW1pbm91cyIsIm1p
            bWljIiwib3NkbWFwLXBydW5lIiwibmF1dGlsdXMiLCJvY3RvcHVzIl19LCJtb25tYXAiOnsiZXBv
            Y2giOjEsImZzaWQiOiJhZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5ZDBmMWMiLCJtb2Rp
            ZmllZCI6IjIwMjAtMDUtMDdUMDM6MzY6MzkuOTE4MDM1WiIsImNyZWF0ZWQiOiIyMDIwLTA1LTA3
            VDAzOjM2OjM5LjkxODAzNVoiLCJtaW5fbW9uX3JlbGVhc2UiOjE1LCJtaW5fbW9uX3JlbGVhc2Vf
            bmFtZSI6Im9jdG9wdXMiLCJmZWF0dXJlcyI6eyJwZXJzaXN0ZW50IjpbImtyYWtlbiIsImx1bWlu
            b3VzIiwibWltaWMiLCJvc2RtYXAtcHJ1bmUiLCJuYXV0aWx1cyIsIm9jdG9wdXMiXSwib3B0aW9u
            YWwiOltdfSwibW9ucyI6W3sicmFuayI6MCwibmFtZSI6ImEiLCJwdWJsaWNfYWRkcnMiOnsiYWRk
            cnZlYyI6W3sidHlwZSI6InYyIiwiYWRkciI6IjEwLjExMC4yMDUuMTc0OjMzMDAiLCJub25jZSI6
            MH0seyJ0eXBlIjoidjEiLCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OSIsIm5vbmNlIjowfV19
            LCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OS8wIiwicHVibGljX2FkZHIiOiIxMC4xMTAuMjA1
            LjE3NDo2Nzg5LzAiLCJwcmlvcml0eSI6MCwid2VpZ2h0IjowfV19fSIiIgogICAgICAgICkKICAg
            ICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXV0g
            PSAoCiAgICAgICAgICAgICIiInsiZGFzaGJvYXJkIjoiaHR0cHM6Ly9jZXBoLWRhc2hib2FyZDo4
            NDQzLyIsInByb21ldGhldXMiOiJodHRwOi8vY2VwaC1kYXNoYm9hcmQtZGI6OTI4My8ifSIiIgog
            ICAgICAgICkKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNh
            cHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMiLCAib3Nk
            IiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0
            YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250
            cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJdLCAiZW50aXR5Ijog
            ImNsaWVudC5oZWFsdGhjaGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRo
            IGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuaGVh
            bHRoY2hlY2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRB
            PT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cyIs
            Im9zZCI6InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5j
            b250cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJ9fV0iIiIKICAg
            ICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIs
            ICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAib3NkIiwgInBy
            b2ZpbGUgcmJkIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtbm9kZSIsICJmb3JtYXQiOiAi
            anNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIi
            W3siZW50aXR5IjoiY2xpZW50LmNzaS1yYmQtbm9kZSIsImtleSI6IkFRQk9nck5lSGJLMUF4QUF1
            YllCZVY4UzFVL0dQenE1U1ZlcTZnPT0iLCJjYXBzIjp7Im1vbiI6InByb2ZpbGUgcmJkLCBhbGxv
            dyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsIm9zZCI6InByb2ZpbGUgcmJkIn19XSIiIgogICAg
            ICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsibW9uIiwg
            InByb2ZpbGUgcmJkLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxs
            b3cgcnciLCAib3NkIiwgInByb2ZpbGUgcmJkIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQt
            cHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNy
            ZWF0ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktcmJkLXByb3Zp
            c2lvbmVyIiwia2V5IjoiQVFCTmdyTmUxZ2V5S3hBQThla1ZpUmRFK2hzczVPd2VZQmt3Tmc9PSIs
            ImNhcHMiOnsibWdyIjoiYWxsb3cgcnciLCJtb24iOiJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFu
            ZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJwcm9maWxlIHJiZCJ9fV0iIiIKICAgICAgICBzZWxm
            LmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyBy
            LCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3Nk
            IiwgImFsbG93IHJ3IHRhZyBjZXBoZnMgKj0qIiwgIm1kcyIsICJhbGxvdyBydyJdLCAiZW50aXR5
            IjogImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4Ijog
            ImF1dGggZ2V0LW9yLWNyZWF0ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVu
            dC5jc2ktY2VwaGZzLW5vZGUiLCJrZXkiOiJBUUJPZ3JOZUVOdW5LeEFBUENtZ0U3UjZHOERjWG5h
            SjFGMzJxZz09IiwiY2FwcyI6eyJtZHMiOiJhbGxvdyBydyIsIm1nciI6ImFsbG93IHJ3IiwibW9u
            IjoiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBy
            dyB0YWcgY2VwaGZzICo9KiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAg
            ICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kICdvc2Qg
            YmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3NkIiwgImFsbG93IHJ3IHRhZyBjZXBo
            ZnMgbWV0YWRhdGE9KiJdLCAiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVy
            IiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgog
            ICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIs
            ImtleSI6IkFRQk9nck5lQUZnY0dCQUF2R3FLT0FEMEQzeHhtVlkwUjkxMmRnPT0iLCJjYXBzIjp7
            Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2Nr
            bGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoifX1dIiIiCiAgICAg
            ICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAi
            YWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdyIiwgImFsbG93IHJ3
            IiwgIm9zZCIsICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiXSwgImVudGl0eSI6ICJj
            bGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVuc2hpZnQtc3RvcmFnZSIsICJmb3JtYXQi
            OiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0g
            IiIiW3siZW50aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0
            b3JhZ2UiLCJrZXkiOiJCUUJPZ3JOZUFGZ2NHQkFBdkdxS09BRDBEM3h4bVZZMFI5MTJkZz09Iiwi
            Y2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29z
            ZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIn19XSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsi
            bW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciIsICJh
            bGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT1teWZzIl0sICJl
            bnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0b3JhZ2Ut
            bXlmcyIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0i
            IiIKICAgICAgICBdID0gIiIiW3siZW50aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9u
            ZXItb3BlbnNoaWZ0LXN0b3JhZ2UtbXlmcyIsImtleSI6IkNRQk9nck5lQUZnY0dCQUF2R3FLT0FE
            MEQzeHhtVlkwUjkxMmRnPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cg
            ciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2Vw
            aGZzIG1ldGFkYXRhPW15ZnMifX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAg
            ICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCBxdW9y
            dW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIiwgImFsbG93IGNvbW1hbmQg
            Y29uZmlnIiwgIm9zZCIsICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFsbG93IHJ3eCBwb29sPWRl
            ZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9ZGVm
            YXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJndy5sb2csIGFsbG93IHgg
            cG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4Il0sICJlbnRpdHkiOiAiY2xpZW50LmhlYWx0
            aGNoZWNrZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0
            ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5oZWFsdGhjaGVja2VyIiwi
            a2V5IjoiQVFERmtiTmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsi
            bW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFu
            ZCB2ZXJzaW9uIiwgIm1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsICJvc2QiOiAicHJvZmls
            ZSByYmQtcmVhZC1vbmx5LCBhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhLCBhbGxvdyBy
            IHBvb2w9LnJndy5yb290LCBhbGxvdyBydyBwb29sPWRlZmF1bHQucmd3LmNvbnRyb2wsIGFsbG93
            IHJ4IHBvb2w9ZGVmYXVsdC5yZ3cubG9nLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0
            cy5pbmRleCJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAi
            IiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMs
            IGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cgY29tbWFuZCBjb25maWciLCAi
            b3NkIiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5j
            b250cm9sLCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1
            bHQucmd3LmJ1Y2tldHMuaW5kZXgiXSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIs
            ICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBjYXBzIn0iIiIKICAgICAgICBdID0g
            IiIiW3siZW50aXR5IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0NWJG
            UkFBVG5kTE5VU1JLcnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIjog
            ImFsbG93IGNvbW1hbmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFs
            bG93IHJ3eCBwb29sPWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFs
            bG93IHJ3IHBvb2w9ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJn
            dy5sb2csIGFsbG93IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4In19XSIiIgogICAg
            ICAgIHNlbGYuY21kX291dHB1dF9tYXBbIiIieyJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            bWdyIHNlcnZpY2VzIn0iIiJdID0gKAogICAgICAgICAgICAiIiJ7ImRhc2hib2FyZCI6ICJodHRw
            Oi8vcm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6NzAwMC8iLCAicHJvbWV0aGV1cyI6
            ICJodHRwOi8vcm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6OTI4My8ifSIiIgogICAg
            ICAgICkKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0
            eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            YXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJ7ImRhc2hib2FyZCI6ICJodHRwOi8vcm9vay1j
            ZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6NzAwMC8iLCAicHJvbWV0aGV1cyI6ICJodHRwOi8v
            cm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6OTI4My8ifSIiIgogICAgICAgIHNlbGYu
            Y21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNsaWVudC5oZWFsdGhj
            aGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldCJ9IiIiCiAgICAg
            ICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5oZWFsdGhjaGVja2VyIiwia2V5IjoiQVFERmti
            TmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjogImFsbG93
            IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFuZCB2ZXJzaW9uIiwg
            Im1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsICJvc2QiOiAicHJvZmlsZSByYmQtcmVhZC1v
            bmx5LCBhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhLCBhbGxvdyByIHBvb2w9LnJndy5y
            b290LCBhbGxvdyBydyBwb29sPWRlZmF1bHQucmd3LmNvbnRyb2wsIGFsbG93IHJ4IHBvb2w9ZGVm
            YXVsdC5yZ3cubG9nLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJ9fV0i
            IiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0eSI6
            ICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJh
            dXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0
            X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtbm9kZSIsICJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJb
            XSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5
            IjogImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRf
            b3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMt
            cHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIK
            ICAgICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAg
            ICAgICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVuc2hp
            ZnQtc3RvcmFnZSIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgog
            ICAgICAgIF0gPSAiIiJbXSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAg
            ICAgICIiInsiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlm
            dC1zdG9yYWdlLW15ZnMiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0i
            IiIKICAgICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAg
            ICAgICAgICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJb
            eyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsImtleSI6IkFRREZrYk5l
            ZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRBPT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIi
            LCAibWdyIjoiYWxsb3cgcnciLCAib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0q
            In19XSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2Fw
            cyI6IFsibW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1n
            ciIsICJhbGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIl0s
            ICJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpz
            b24iLCAicHJlZml4IjogImF1dGggY2FwcyJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmtiTmVmdDViRlJBQVRu
            ZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cgciwgIGFsbG93IGNv
            bW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciI6ImFsbG93IHJ3IiwgIm9zZCI6ImFsbG93IHJ3
            IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFw
            Wyd7ImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJzdGF0dXMifSddID0gY2VwaF9zdGF0dXNf
            c3RyCgogICAgZGVmIHNodXRkb3duKHNlbGYpOgogICAgICAgIHBhc3MKCiAgICBkZWYgZ2V0X2Zz
            aWQoc2VsZik6CiAgICAgICAgcmV0dXJuICJhZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5
            ZDBmMWMiCgogICAgZGVmIGNvbmZfcmVhZF9maWxlKHNlbGYpOgogICAgICAgIHBhc3MKCiAgICBk
            ZWYgY29ubmVjdChzZWxmKToKICAgICAgICBwYXNzCgogICAgZGVmIHBvb2xfZXhpc3RzKHNlbGYs
            IHBvb2xfbmFtZSk6CiAgICAgICAgcmV0dXJuIFRydWUKCiAgICBkZWYgbW9uX2NvbW1hbmQoc2Vs
            ZiwgY21kLCBvdXQpOgogICAgICAgIGpzb25fY21kID0ganNvbi5sb2FkcyhjbWQpCiAgICAgICAg
            anNvbl9jbWRfc3RyID0ganNvbi5kdW1wcyhqc29uX2NtZCwgc29ydF9rZXlzPVRydWUpCiAgICAg
            ICAgY21kX291dHB1dCA9IHNlbGYuY21kX291dHB1dF9tYXBbanNvbl9jbWRfc3RyXQogICAgICAg
            IHJldHVybiBzZWxmLnJldHVybl92YWwsIGNtZF9vdXRwdXQsIHN0cihzZWxmLmVycl9tZXNzYWdl
            LmVuY29kZSgidXRmLTgiKSkKCiAgICBkZWYgX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoc2VsZiwg
            aG9zdF9uYW1lKToKICAgICAgICBpcF9yZWdfeCA9IHJlLmNvbXBpbGUociJcZHsxLDN9LlxkezEs
            M30uXGR7MSwzfS5cZHsxLDN9IikKICAgICAgICAjIGlmIHByb3ZpZGVkIGhvc3QgaXMgZGlyZWN0
            bHkgYW4gSVAgYWRkcmVzcywgcmV0dXJuIHRoZSBzYW1lCiAgICAgICAgaWYgaXBfcmVnX3gubWF0
            Y2goaG9zdF9uYW1lKToKICAgICAgICAgICAgcmV0dXJuIGhvc3RfbmFtZQogICAgICAgIGltcG9y
            dCByYW5kb20KCiAgICAgICAgaG9zdF9pcCA9IHNlbGYuZHVtbXlfaG9zdF9pcF9tYXAuZ2V0KGhv
            c3RfbmFtZSwgIiIpCiAgICAgICAgaWYgbm90IGhvc3RfaXA6CiAgICAgICAgICAgIGhvc3RfaXAg
            PSBmIjE3Mi45LntyYW5kb20ucmFuZGludCgwLCAyNTQpfS57cmFuZG9tLnJhbmRpbnQoMCwgMjU0
            KX0iCiAgICAgICAgICAgIHNlbGYuZHVtbXlfaG9zdF9pcF9tYXBbaG9zdF9uYW1lXSA9IGhvc3Rf
            aXAKICAgICAgICBkZWwgcmFuZG9tCiAgICAgICAgcmV0dXJuIGhvc3RfaXAKCiAgICBAY2xhc3Nt
            ZXRob2QKICAgIGRlZiBSYWRvcyhjb25mZmlsZT1Ob25lKToKICAgICAgICByZXR1cm4gRHVtbXlS
            YWRvcygpCgoKY2xhc3MgUzNBdXRoKEF1dGhCYXNlKToKICAgICIiIkF0dGFjaGVzIEFXUyBBdXRo
            ZW50aWNhdGlvbiB0byB0aGUgZ2l2ZW4gUmVxdWVzdCBvYmplY3QuIiIiCgogICAgc2VydmljZV9i
            YXNlX3VybCA9ICJzMy5hbWF6b25hd3MuY29tIgoKICAgIGRlZiBfX2luaXRfXyhzZWxmLCBhY2Nl
            c3Nfa2V5LCBzZWNyZXRfa2V5LCBzZXJ2aWNlX3VybD1Ob25lKToKICAgICAgICBpZiBzZXJ2aWNl
            X3VybDoKICAgICAgICAgICAgc2VsZi5zZXJ2aWNlX2Jhc2VfdXJsID0gc2VydmljZV91cmwKICAg
            ICAgICBzZWxmLmFjY2Vzc19rZXkgPSBzdHIoYWNjZXNzX2tleSkKICAgICAgICBzZWxmLnNlY3Jl
            dF9rZXkgPSBzdHIoc2VjcmV0X2tleSkKCiAgICBkZWYgX19jYWxsX18oc2VsZiwgcik6CiAgICAg
            ICAgIyBDcmVhdGUgZGF0ZSBoZWFkZXIgaWYgaXQgaXMgbm90IGNyZWF0ZWQgeWV0LgogICAgICAg
            IGlmICJkYXRlIiBub3QgaW4gci5oZWFkZXJzIGFuZCAieC1hbXotZGF0ZSIgbm90IGluIHIuaGVh
            ZGVyczoKICAgICAgICAgICAgci5oZWFkZXJzWyJkYXRlIl0gPSBmb3JtYXRkYXRlKHRpbWV2YWw9
            Tm9uZSwgbG9jYWx0aW1lPUZhbHNlLCB1c2VnbXQ9VHJ1ZSkKICAgICAgICBzaWduYXR1cmUgPSBz
            ZWxmLmdldF9zaWduYXR1cmUocikKICAgICAgICBpZiBweTNrOgogICAgICAgICAgICBzaWduYXR1
            cmUgPSBzaWduYXR1cmUuZGVjb2RlKCJ1dGYtOCIpCiAgICAgICAgci5oZWFkZXJzWyJBdXRob3Jp
            emF0aW9uIl0gPSBmIkFXUyB7c2VsZi5hY2Nlc3Nfa2V5fTp7c2lnbmF0dXJlfSIKICAgICAgICBy
            ZXR1cm4gcgoKICAgIGRlZiBnZXRfc2lnbmF0dXJlKHNlbGYsIHIpOgogICAgICAgIGNhbm9uaWNh
            bF9zdHJpbmcgPSBzZWxmLmdldF9jYW5vbmljYWxfc3RyaW5nKHIudXJsLCByLmhlYWRlcnMsIHIu
            bWV0aG9kKQogICAgICAgIGlmIHB5M2s6CiAgICAgICAgICAgIGtleSA9IHNlbGYuc2VjcmV0X2tl
            eS5lbmNvZGUoInV0Zi04IikKICAgICAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZy5lbmNv
            ZGUoInV0Zi04IikKICAgICAgICBlbHNlOgogICAgICAgICAgICBrZXkgPSBzZWxmLnNlY3JldF9r
            ZXkKICAgICAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZwogICAgICAgIGggPSBobWFjLm5l
            dyhrZXksIG1zZywgZGlnZXN0bW9kPXNoYSkKICAgICAgICByZXR1cm4gZW5jb2Rlc3RyaW5nKGgu
            ZGlnZXN0KCkpLnN0cmlwKCkKCiAgICBkZWYgZ2V0X2Nhbm9uaWNhbF9zdHJpbmcoc2VsZiwgdXJs
            LCBoZWFkZXJzLCBtZXRob2QpOgogICAgICAgIHBhcnNlZHVybCA9IHVybHBhcnNlKHVybCkKICAg
            ICAgICBvYmplY3RrZXkgPSBwYXJzZWR1cmwucGF0aFsxOl0KCiAgICAgICAgYnVja2V0ID0gcGFy
            c2VkdXJsLm5ldGxvY1s6IC1sZW4oc2VsZi5zZXJ2aWNlX2Jhc2VfdXJsKV0KICAgICAgICBpZiBs
            ZW4oYnVja2V0KSA+IDE6CiAgICAgICAgICAgICMgcmVtb3ZlIGxhc3QgZG90CiAgICAgICAgICAg
            IGJ1Y2tldCA9IGJ1Y2tldFs6LTFdCgogICAgICAgIGludGVyZXN0aW5nX2hlYWRlcnMgPSB7ImNv
            bnRlbnQtbWQ1IjogIiIsICJjb250ZW50LXR5cGUiOiAiIiwgImRhdGUiOiAiIn0KICAgICAgICBm
            b3Iga2V5IGluIGhlYWRlcnM6CiAgICAgICAgICAgIGxrID0ga2V5Lmxvd2VyKCkKICAgICAgICAg
            ICAgdHJ5OgogICAgICAgICAgICAgICAgbGsgPSBsay5kZWNvZGUoInV0Zi04IikKICAgICAgICAg
            ICAgZXhjZXB0OgogICAgICAgICAgICAgICAgcGFzcwogICAgICAgICAgICBpZiBoZWFkZXJzW2tl
            eV0gYW5kICgKICAgICAgICAgICAgICAgIGxrIGluIGludGVyZXN0aW5nX2hlYWRlcnMua2V5cygp
            IG9yIGxrLnN0YXJ0c3dpdGgoIngtYW16LSIpCiAgICAgICAgICAgICk6CiAgICAgICAgICAgICAg
            ICBpbnRlcmVzdGluZ19oZWFkZXJzW2xrXSA9IGhlYWRlcnNba2V5XS5zdHJpcCgpCgogICAgICAg
            ICMgSWYgeC1hbXotZGF0ZSBpcyB1c2VkIGl0IHN1cGVyc2VkZXMgdGhlIGRhdGUgaGVhZGVyLgog
            ICAgICAgIGlmIG5vdCBweTNrOgogICAgICAgICAgICBpZiAieC1hbXotZGF0ZSIgaW4gaW50ZXJl
            c3RpbmdfaGVhZGVyczoKICAgICAgICAgICAgICAgIGludGVyZXN0aW5nX2hlYWRlcnNbImRhdGUi
            XSA9ICIiCiAgICAgICAgZWxzZToKICAgICAgICAgICAgaWYgIngtYW16LWRhdGUiIGluIGludGVy
            ZXN0aW5nX2hlYWRlcnM6CiAgICAgICAgICAgICAgICBpbnRlcmVzdGluZ19oZWFkZXJzWyJkYXRl
            Il0gPSAiIgoKICAgICAgICBidWYgPSBmInttZXRob2R9XG4iCiAgICAgICAgZm9yIGtleSBpbiBz
            b3J0ZWQoaW50ZXJlc3RpbmdfaGVhZGVycy5rZXlzKCkpOgogICAgICAgICAgICB2YWwgPSBpbnRl
            cmVzdGluZ19oZWFkZXJzW2tleV0KICAgICAgICAgICAgaWYga2V5LnN0YXJ0c3dpdGgoIngtYW16
            LSIpOgogICAgICAgICAgICAgICAgYnVmICs9IGYie2tleX06e3ZhbH1cbiIKICAgICAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgICAgIGJ1ZiArPSBmInt2YWx9XG4iCgogICAgICAgICMgYXBwZW5k
            IHRoZSBidWNrZXQgaWYgaXQgZXhpc3RzCiAgICAgICAgaWYgYnVja2V0ICE9ICIiOgogICAgICAg
            ICAgICBidWYgKz0gZiIve2J1Y2tldH0iCgogICAgICAgICMgYWRkIHRoZSBvYmplY3RrZXkuIGV2
            ZW4gaWYgaXQgZG9lc24ndCBleGlzdCwgYWRkIHRoZSBzbGFzaAogICAgICAgIGJ1ZiArPSBmIi97
            b2JqZWN0a2V5fSIKCiAgICAgICAgcmV0dXJuIGJ1ZgoKCmNsYXNzIFJhZG9zSlNPTjoKICAgIEVY
            VEVSTkFMX1VTRVJfTkFNRSA9ICJjbGllbnQuaGVhbHRoY2hlY2tlciIKICAgIEVYVEVSTkFMX1JH
            V19BRE1JTl9PUFNfVVNFUl9OQU1FID0gInJndy1hZG1pbi1vcHMtdXNlciIKICAgIEVNUFRZX09V
            VFBVVF9MSVNUID0gIkVtcHR5IG91dHB1dCBsaXN0IgogICAgREVGQVVMVF9SR1dfUE9PTF9QUkVG
            SVggPSAiZGVmYXVsdCIKICAgIERFRkFVTFRfTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUID0gIjky
            ODMiCgogICAgQGNsYXNzbWV0aG9kCiAgICBkZWYgZ2VuX2FyZ19wYXJzZXIoY2xzLCBhcmdzX3Rv
            X3BhcnNlPU5vbmUpOgogICAgICAgIGFyZ1AgPSBhcmdwYXJzZS5Bcmd1bWVudFBhcnNlcigpCgog
            ICAgICAgIGNvbW1vbl9ncm91cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJjb21tb24iKQog
            ICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoIi0tdmVyYm9zZSIsICItdiIsIGFjdGlv
            bj0ic3RvcmVfdHJ1ZSIsIGRlZmF1bHQ9RmFsc2UpCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9h
            cmd1bWVudCgKICAgICAgICAgICAgIi0tY2VwaC1jb25mIiwgIi1jIiwgaGVscD0iUHJvdmlkZSBh
            IGNlcGggY29uZiBmaWxlLiIsIHR5cGU9c3RyCiAgICAgICAgKQogICAgICAgIGNvbW1vbl9ncm91
            cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLWtleXJpbmciLCAiLWsiLCBoZWxwPSJQYXRo
            IHRvIGNlcGgga2V5cmluZyBmaWxlLiIsIHR5cGU9c3RyCiAgICAgICAgKQogICAgICAgIGNvbW1v
            bl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJ1bi1hcy11c2VyIiwKICAgICAg
            ICAgICAgIi11IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgdHlwZT1zdHIs
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIGEgdXNlciBuYW1lIHRvIGNoZWNrIHRoZSBjbHVz
            dGVyJ3MgaGVhbHRoIHN0YXR1cywgbXVzdCBiZSBwcmVmaXhlZCBieSAnY2xpZW50LiciLAogICAg
            ICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1j
            bHVzdGVyLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJL
            dWJlcm5ldGVzIGNsdXN0ZXIgbmFtZShsZWdhY3kgZmxhZyksIE5vdGU6IEVpdGhlciB1c2UgdGhp
            cyBvciAtLWs4cy1jbHVzdGVyLW5hbWUiLAogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAu
            YWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1rOHMtY2x1c3Rlci1uYW1lIiwgZGVmYXVsdD0i
            IiwgaGVscD0iS3ViZXJuZXRlcyBjbHVzdGVyIG5hbWUiCiAgICAgICAgKQogICAgICAgIGNvbW1v
            bl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLW5hbWVzcGFjZSIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIGhlbHA9Ik5hbWVzcGFjZSB3aGVyZSBDZXBoQ2x1
            c3RlciBpcyBydW5uaW5nIiwKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tcmd3LXBvb2wtcHJlZml4IiwgZGVmYXVsdD0iIiwgaGVscD0i
            UkdXIFBvb2wgcHJlZml4IgogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3Vt
            ZW50KAogICAgICAgICAgICAiLS1yZXN0cmljdGVkLWF1dGgtcGVybWlzc2lvbiIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlJlc3RyaWN0IGNlcGhDU0lLZXly
            aW5ncyBhdXRoIHBlcm1pc3Npb25zIHRvIHNwZWNpZmljIHBvb2xzLCBjbHVzdGVyLiIKICAgICAg
            ICAgICAgKyAiTWFuZGF0b3J5IGZsYWdzIHRoYXQgbmVlZCB0byBiZSBzZXQgYXJlIC0tcmJkLWRh
            dGEtcG9vbC1uYW1lLCBhbmQgLS1rOHMtY2x1c3Rlci1uYW1lLiIKICAgICAgICAgICAgKyAiLS1j
            ZXBoZnMtZmlsZXN5c3RlbS1uYW1lIGZsYWcgY2FuIGFsc28gYmUgcGFzc2VkIGluIGNhc2Ugb2Yg
            Y2VwaGZzIHVzZXIgcmVzdHJpY3Rpb24sIHNvIGl0IGNhbiByZXN0cmljdCB1c2VyIHRvIHBhcnRp
            Y3VsYXIgY2VwaGZzIGZpbGVzeXN0ZW0iCiAgICAgICAgICAgICsgInNhbXBsZSBydW46IGBweXRo
            b24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwtY2x1c3Rlci1yZXNvdXJjZXMucHkgLS1jZXBo
            ZnMtZmlsZXN5c3RlbS1uYW1lIG15ZnMgLS1yYmQtZGF0YS1wb29sLW5hbWUgcmVwbGljYXBvb2wg
            LS1rOHMtY2x1c3Rlci1uYW1lIHJvb2tzdG9yYWdlIC0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Np
            b24gdHJ1ZWAiCiAgICAgICAgICAgICsgIk5vdGU6IFJlc3RyaWN0aW5nIHRoZSBjc2ktdXNlcnMg
            cGVyIHBvb2wsIGFuZCBwZXIgY2x1c3RlciB3aWxsIHJlcXVpcmUgY3JlYXRpbmcgbmV3IGNzaS11
            c2VycyBhbmQgbmV3IHNlY3JldHMgZm9yIHRoYXQgY3NpLXVzZXJzLiIKICAgICAgICAgICAgKyAi
            U28gYXBwbHkgdGhlc2Ugc2VjcmV0cyBvbmx5IHRvIG5ldyBgQ29uc3VtZXIgY2x1c3RlcmAgZGVw
            bG95bWVudCB3aGlsZSB1c2luZyB0aGUgc2FtZSBgU291cmNlIGNsdXN0ZXJgLiIsCiAgICAgICAg
            KQogICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXYyLXBv
            cnQtZW5hYmxlIiwKICAgICAgICAgICAgYWN0aW9uPSJzdG9yZV90cnVlIiwKICAgICAgICAgICAg
            ZGVmYXVsdD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iRW5hYmxlIHYyIG1vbiBwb3J0KDMzMDAp
            IGZvciBtb25zIiwKICAgICAgICApCgogICAgICAgIG91dHB1dF9ncm91cCA9IGFyZ1AuYWRkX2Fy
            Z3VtZW50X2dyb3VwKCJvdXRwdXQiKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQo
            CiAgICAgICAgICAgICItLWZvcm1hdCIsCiAgICAgICAgICAgICItdCIsCiAgICAgICAgICAgIGNo
            b2ljZXM9WyJqc29uIiwgImJhc2giXSwKICAgICAgICAgICAgZGVmYXVsdD0ianNvbiIsCiAgICAg
            ICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBvdXRwdXQgZm9ybWF0IChqc29uIHwgYmFzaCkiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1vdXRwdXQiLAogICAgICAgICAgICAiLW8iLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAg
            ICAgICAgICBoZWxwPSJPdXRwdXQgd2lsbCBiZSBzdG9yZWQgaW50byB0aGUgcHJvdmlkZWQgZmls
            ZSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAog
            ICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgQ2VwaCBmaWxlc3lzdGVt
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tY2VwaGZzLW1ldGFkYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIs
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBjZXBoZnMgbWV0YWRh
            dGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAg
            ICAgICAgICAgICItLWNlcGhmcy1kYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9
            IiIsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBjZXBoZnMgZGF0
            YSBwb29sIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAg
            ICAgICAgICAgIi0tcmJkLWRhdGEtcG9vbC1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwK
            ICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRo
            ZSBuYW1lIG9mIHRoZSBSQkQgZGF0YXBvb2wiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3Jv
            dXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1hbGlhcy1yYmQtZGF0YS1wb29sLW5hbWUi
            LAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAg
            ICAgICAgICAgaGVscD0iUHJvdmlkZXMgYW4gYWxpYXMgZm9yIHRoZSAgUkJEIGRhdGEgcG9vbCBu
            YW1lLCBuZWNlc3NhcnkgaWYgYSBzcGVjaWFsIGNoYXJhY3RlciBpcyBwcmVzZW50IGluIHRoZSBw
            b29sIG5hbWUgc3VjaCBhcyBhIHBlcmlvZCBvciB1bmRlcnNjb3JlIiwKICAgICAgICApCiAgICAg
            ICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmd3LWVuZHBvaW50
            IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAg
            ICAgICAgICAgIGhlbHA9IlJBRE9TIEdhdGV3YXkgZW5kcG9pbnQgKGluIGA8SVB2ND46PFBPUlQ+
            YCBvciBgPFtJUHY2XT46PFBPUlQ+YCBvciBgPEZRRE4+OjxQT1JUPmAgZm9ybWF0KSIsCiAgICAg
            ICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJn
            dy10bHMtY2VydC1wYXRoIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVx
            dWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlJBRE9TIEdhdGV3YXkgZW5kcG9pbnQgVExT
            IGNlcnRpZmljYXRlIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVu
            dCgKICAgICAgICAgICAgIi0tcmd3LXNraXAtdGxzIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFs
            c2UsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9Iklnbm9yZSBU
            TFMgY2VydGlmaWNhdGlvbiB2YWxpZGF0aW9uIHdoZW4gYSBzZWxmLXNpZ25lZCBjZXJ0aWZpY2F0
            ZSBpcyBwcm92aWRlZCAoTk9UIFJFQ09NTUVOREVEIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tbW9uaXRvcmluZy1lbmRwb2ludCIs
            CiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAg
            ICAgICAgICBoZWxwPSJDZXBoIE1hbmFnZXIgcHJvbWV0aGV1cyBleHBvcnRlciBlbmRwb2ludHMg
            KGNvbW1hIHNlcGFyYXRlZCBsaXN0IG9mIChmb3JtYXQgYDxJUHY0PmAgb3IgYDxbSVB2Nl0+YCBv
            ciBgPEZRRE4+YCkgZW50cmllcyBvZiBhY3RpdmUgYW5kIHN0YW5kYnkgbWdycykiLAogICAgICAg
            ICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1tb25p
            dG9yaW5nLWVuZHBvaW50LXBvcnQiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAg
            ICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iQ2VwaCBNYW5hZ2VyIHByb21ldGhl
            dXMgZXhwb3J0ZXIgcG9ydCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJn
            dW1lbnQoCiAgICAgICAgICAgICItLXNraXAtbW9uaXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAg
            ICAgICAgIGhlbHA9IkRvIG5vdCBjaGVjayBmb3IgYSBtb25pdG9yaW5nIGVuZHBvaW50IGZvciB0
            aGUgQ2VwaCBjbHVzdGVyIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tcmJkLW1ldGFkYXRhLWVjLXBvb2wtbmFtZSIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBo
            ZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiBlcmFzdXJlIGNvZGVkIFJCRCBtZXRhZGF0YSBwb29s
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tZHJ5LXJ1biIsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFj
            dGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGhlbHA9IkRyeSBydW4gcHJpbnRzIHRoZSBl
            eGVjdXRlZCBjb21tYW5kcyB3aXRob3V0IHJ1bm5pbmcgdGhlbSIsCiAgICAgICAgKQogICAgICAg
            IG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJhZG9zLW5hbWVzcGFj
            ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAog
            ICAgICAgICAgICBoZWxwPSJEaXZpZGVzIGEgcG9vbCBpbnRvIHNlcGFyYXRlIGxvZ2ljYWwgbmFt
            ZXNwYWNlcywgdXNlZCBmb3IgY3JlYXRpbmcgUkJEIFBWQyBpbiBhIENlcGhCbG9ja1Bvb2xSYWRv
            c05hbWVzcGFjZSAoc2hvdWxkIGJlIGxvd2VyIGNhc2UpIiwKICAgICAgICApCiAgICAgICAgb3V0
            cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tc3Vidm9sdW1lLWdyb3VwIiwK
            ICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAg
            ICAgICAgIGhlbHA9InByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBzdWJ2b2x1bWUgZ3JvdXAiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1yZ3ctcmVhbG0tbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJl
            cXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUg
            cmd3LXJlYWxtIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgK
            ICAgICAgICAgICAgIi0tcmd3LXpvbmUtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAg
            ICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUg
            bmFtZSBvZiB0aGUgcmd3LXpvbmUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS1yZ3ctem9uZWdyb3VwLW5hbWUiLAogICAgICAgICAg
            ICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVs
            cD0icHJvdmlkZXMgdGhlIG5hbWUgb2YgdGhlIHJndy16b25lZ3JvdXAiLAogICAgICAgICkKCiAg
            ICAgICAgdXBncmFkZV9ncm91cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJ1cGdyYWRlIikK
            ICAgICAgICB1cGdyYWRlX2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tdXBncmFk
            ZSIsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9
            RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlVwZ3JhZGVzIHRoZSBjZXBoQ1NJS2V5cmluZ3MoRm9y
            IGV4YW1wbGU6IGNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyKSBhbmQgY2xpZW50LmhlYWx0
            aGNoZWNrZXIgY2VwaCB1c2VycyB3aXRoIG5ldyBwZXJtaXNzaW9ucyBuZWVkZWQgZm9yIHRoZSBu
            ZXcgY2x1c3RlciB2ZXJzaW9uIGFuZCBvbGRlciBwZXJtaXNzaW9uIHdpbGwgc3RpbGwgYmUgYXBw
            bGllZC4iCiAgICAgICAgICAgICsgIlNhbXBsZSBydW46IGBweXRob24zIC9ldGMvY2VwaC9jcmVh
            dGUtZXh0ZXJuYWwtY2x1c3Rlci1yZXNvdXJjZXMucHkgLS11cGdyYWRlYCwgdGhpcyB3aWxsIHVw
            Z3JhZGUgYWxsIHRoZSBkZWZhdWx0IGNzaSB1c2Vycyhub24tcmVzdHJpY3RlZCkiCiAgICAgICAg
            ICAgICsgIkZvciByZXN0cmljdGVkIHVzZXJzKEZvciBleGFtcGxlOiBjbGllbnQuY3NpLWNlcGhm
            cy1wcm92aXNpb25lci1vcGVuc2hpZnQtc3RvcmFnZS1teWZzKSwgdXNlcnMgY3JlYXRlZCB1c2lu
            ZyAtLXJlc3RyaWN0ZWQtYXV0aC1wZXJtaXNzaW9uIGZsYWcgbmVlZCB0byBwYXNzIG1hbmRhdG9y
            eSBmbGFncyIKICAgICAgICAgICAgKyAibWFuZGF0b3J5IGZsYWdzOiAnLS1yYmQtZGF0YS1wb29s
            LW5hbWUsIC0tazhzLWNsdXN0ZXItbmFtZSBhbmQgLS1ydW4tYXMtdXNlcicgZmxhZ3Mgd2hpbGUg
            dXBncmFkaW5nIgogICAgICAgICAgICArICJpbiBjYXNlIG9mIGNlcGhmcyB1c2VycyBpZiB5b3Ug
            aGF2ZSBwYXNzZWQgLS1jZXBoZnMtZmlsZXN5c3RlbS1uYW1lIGZsYWcgd2hpbGUgY3JlYXRpbmcg
            dXNlciB0aGVuIHdoaWxlIHVwZ3JhZGluZyBpdCB3aWxsIGJlIG1hbmRhdG9yeSB0b28iCiAgICAg
            ICAgICAgICsgIlNhbXBsZSBydW46IGBweXRob24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwt
            Y2x1c3Rlci1yZXNvdXJjZXMucHkgLS11cGdyYWRlIC0tcmJkLWRhdGEtcG9vbC1uYW1lIHJlcGxp
            Y2Fwb29sIC0tazhzLWNsdXN0ZXItbmFtZSByb29rc3RvcmFnZSAgLS1ydW4tYXMtdXNlciBjbGll
            bnQuY3NpLXJiZC1ub2RlLXJvb2tzdG9yYWdlLXJlcGxpY2Fwb29sYCIKICAgICAgICAgICAgKyAi
            UFM6IEFuIGV4aXN0aW5nIG5vbi1yZXN0cmljdGVkIHVzZXIgY2Fubm90IGJlIGNvbnZlcnRlZCB0
            byBhIHJlc3RyaWN0ZWQgdXNlciBieSB1cGdyYWRpbmcuIgogICAgICAgICAgICArICJVcGdyYWRl
            IGZsYWcgc2hvdWxkIG9ubHkgYmUgdXNlZCB0byBhcHBlbmQgbmV3IHBlcm1pc3Npb25zIHRvIHVz
            ZXJzLCBpdCBzaG91bGRuJ3QgYmUgdXNlZCBmb3IgY2hhbmdpbmcgdXNlciBhbHJlYWR5IGFwcGxp
            ZWQgcGVybWlzc2lvbiwgZm9yIGV4YW1wbGUgeW91IHNob3VsZG4ndCBjaGFuZ2UgaW4gd2hpY2gg
            cG9vbCB1c2VyIGhhcyBhY2Nlc3MiLAogICAgICAgICkKCiAgICAgICAgaWYgYXJnc190b19wYXJz
            ZToKICAgICAgICAgICAgYXNzZXJ0ICgKICAgICAgICAgICAgICAgIHR5cGUoYXJnc190b19wYXJz
            ZSkgPT0gbGlzdAogICAgICAgICAgICApLCAiQXJndW1lbnQgdG8gJ2dlbl9hcmdfcGFyc2VyJyBz
            aG91bGQgYmUgYSBsaXN0IgogICAgICAgIGVsc2U6CiAgICAgICAgICAgIGFyZ3NfdG9fcGFyc2Ug
            PSBzeXMuYXJndlsxOl0KICAgICAgICByZXR1cm4gYXJnUC5wYXJzZV9hcmdzKGFyZ3NfdG9fcGFy
            c2UpCgogICAgZGVmIHZhbGlkYXRlX3JiZF9tZXRhZGF0YV9lY19wb29sX25hbWUoc2VsZik6CiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lOgogICAg
            ICAgICAgICByYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5yYmRf
            bWV0YWRhdGFfZWNfcG9vbF9uYW1lCiAgICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9h
            cmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQoKICAgICAgICAgICAgaWYgcmJkX3Bvb2xfbmFt
            ZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24o
            CiAgICAgICAgICAgICAgICAgICAgIkZsYWcgJy0tcmJkLWRhdGEtcG9vbC1uYW1lJyBzaG91bGQg
            bm90IGJlIGVtcHR5IgogICAgICAgICAgICAgICAgKQoKICAgICAgICAgICAgaWYgcmJkX21ldGFk
            YXRhX2VjX3Bvb2xfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIkZsYWcgJy0tcmJkLW1ldGFkYXRh
            LWVjLXBvb2wtbmFtZScgc2hvdWxkIG5vdCBiZSBlbXB0eSIKICAgICAgICAgICAgICAgICkKCiAg
            ICAgICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAib3NkIGR1bXAiLCAiZm9ybWF0IjogImpz
            b24ifQogICAgICAgICAgICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1v
            bl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBs
            ZW4oanNvbl9vdXQpID09IDA6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJl
            RXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYie2NtZF9qc29uWydwcmVmaXgnXX0gY29t
            bWFuZCBmYWlsZWQuXG4iCiAgICAgICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYg
            cmV0X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAgICAgICAg
            ICApCiAgICAgICAgICAgIG1ldGFkYXRhX3Bvb2xfZXhpc3QsIHBvb2xfZXhpc3QgPSBGYWxzZSwg
            RmFsc2UKCiAgICAgICAgICAgIGZvciBrZXkgaW4ganNvbl9vdXRbInBvb2xzIl06CiAgICAgICAg
            ICAgICAgICAjIGlmIGVyYXN1cmVfY29kZV9wcm9maWxlIGlzIGVtcHR5IGFuZCBwb29sIG5hbWUg
            ZXhpc3RzIHRoZW4gaXQgcmVwbGljYSBwb29sCiAgICAgICAgICAgICAgICBpZiAoCiAgICAgICAg
            ICAgICAgICAgICAga2V5WyJlcmFzdXJlX2NvZGVfcHJvZmlsZSJdID09ICIiCiAgICAgICAgICAg
            ICAgICAgICAgYW5kIGtleVsicG9vbF9uYW1lIl0gPT0gcmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFt
            ZQogICAgICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgICAgICBtZXRhZGF0YV9wb29sX2V4
            aXN0ID0gVHJ1ZQogICAgICAgICAgICAgICAgIyBpZiBlcmFzdXJlX2NvZGVfcHJvZmlsZSBpcyBu
            b3QgZW1wdHkgYW5kIHBvb2wgbmFtZSBleGlzdHMgdGhlbiBpdCBpcyBlYyBwb29sCiAgICAgICAg
            ICAgICAgICBpZiBrZXlbImVyYXN1cmVfY29kZV9wcm9maWxlIl0gYW5kIGtleVsicG9vbF9uYW1l
            Il0gPT0gcmJkX3Bvb2xfbmFtZToKICAgICAgICAgICAgICAgICAgICBwb29sX2V4aXN0ID0gVHJ1
            ZQoKICAgICAgICAgICAgaWYgbm90IG1ldGFkYXRhX3Bvb2xfZXhpc3Q6CiAgICAgICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJQ
            cm92aWRlZCByYmRfZWNfbWV0YWRhdGFfcG9vbCBuYW1lLCIKICAgICAgICAgICAgICAgICAgICBm
            IiB7cmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZX0sIGRvZXMgbm90IGV4aXN0IgogICAgICAgICAg
            ICAgICAgKQogICAgICAgICAgICBpZiBub3QgcG9vbF9leGlzdDoKICAgICAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgZiJQcm92
            aWRlZCByYmRfZGF0YV9wb29sIG5hbWUsIHtyYmRfcG9vbF9uYW1lfSwgZG9lcyBub3QgZXhpc3Qi
            CiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIHJldHVybiByYmRfbWV0YWRhdGFfZWNfcG9v
            bF9uYW1lCgogICAgZGVmIGRyeV9ydW4oc2VsZiwgbXNnKToKICAgICAgICBpZiBzZWxmLl9hcmdf
            cGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHByaW50KCJFeGVjdXRlOiAiICsgIiciICsgbXNn
            ICsgIiciKQoKICAgIGRlZiB2YWxpZGF0ZV9yZ3dfZW5kcG9pbnRfdGxzX2NlcnQoc2VsZik6CiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZ3dfdGxzX2NlcnRfcGF0aDoKICAgICAgICAgICAg
            d2l0aCBvcGVuKHNlbGYuX2FyZ19wYXJzZXIucmd3X3Rsc19jZXJ0X3BhdGgsIGVuY29kaW5nPSJ1
            dGY4IikgYXMgZjoKICAgICAgICAgICAgICAgIGNvbnRlbnRzID0gZi5yZWFkKCkKICAgICAgICAg
            ICAgICAgIHJldHVybiBjb250ZW50cy5yc3RyaXAoKQoKICAgIGRlZiBfY2hlY2tfY29uZmxpY3Rp
            bmdfb3B0aW9ucyhzZWxmKToKICAgICAgICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNlci51cGdyYWRl
            IGFuZCBub3Qgc2VsZi5fYXJnX3BhcnNlci5yYmRfZGF0YV9wb29sX25hbWU6CiAgICAgICAgICAg
            IHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAiRWl0aGVy
            ICctLXVwZ3JhZGUnIG9yICctLXJiZC1kYXRhLXBvb2wtbmFtZSA8cG9vbF9uYW1lPicgc2hvdWxk
            IGJlIHNwZWNpZmllZCIKICAgICAgICAgICAgKQoKICAgIGRlZiBfaW52YWxpZF9lbmRwb2ludChz
            ZWxmLCBlbmRwb2ludF9zdHIpOgogICAgICAgICMgZXh0cmFjdCB0aGUgcG9ydCBieSBnZXR0aW5n
            IHRoZSBsYXN0IHNwbGl0IG9uIGA6YCBkZWxpbWl0ZXIKICAgICAgICB0cnk6CiAgICAgICAgICAg
            IGVuZHBvaW50X3N0cl9pcCwgcG9ydCA9IGVuZHBvaW50X3N0ci5yc3BsaXQoIjoiLCAxKQogICAg
            ICAgIGV4Y2VwdCBWYWx1ZUVycm9yOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJl
            RXhjZXB0aW9uKGYiTm90IGEgcHJvcGVyIGVuZHBvaW50OiB7ZW5kcG9pbnRfc3RyfSIpCgogICAg
            ICAgIHRyeToKICAgICAgICAgICAgaWYgZW5kcG9pbnRfc3RyX2lwWzBdID09ICJbIjoKICAgICAg
            ICAgICAgICAgIGVuZHBvaW50X3N0cl9pcCA9IGVuZHBvaW50X3N0cl9pcFsxIDogbGVuKGVuZHBv
            aW50X3N0cl9pcCkgLSAxXQogICAgICAgICAgICBpcF90eXBlID0gKAogICAgICAgICAgICAgICAg
            IklQdjQiIGlmIHR5cGUoaXBfYWRkcmVzcyhlbmRwb2ludF9zdHJfaXApKSBpcyBJUHY0QWRkcmVz
            cyBlbHNlICJJUHY2IgogICAgICAgICAgICApCiAgICAgICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAg
            ICAgICAgICAgIGlwX3R5cGUgPSAiRlFETiIKICAgICAgICBpZiBub3QgcG9ydC5pc2RpZ2l0KCk6
            CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oZiJQb3J0IG5vdCB2
            YWxpZDoge3BvcnR9IikKICAgICAgICBpbnRQb3J0ID0gaW50KHBvcnQpCiAgICAgICAgaWYgaW50
            UG9ydCA8IDEgb3IgaW50UG9ydCA+IDIqKjE2IC0gMToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0
            aW9uRmFpbHVyZUV4Y2VwdGlvbihmIk91dCBvZiByYW5nZSBwb3J0IG51bWJlcjoge3BvcnR9IikK
            CiAgICAgICAgcmV0dXJuIGlwX3R5cGUKCiAgICBkZWYgZW5kcG9pbnRfZGlhbChzZWxmLCBlbmRw
            b2ludF9zdHIsIGlwX3R5cGUsIHRpbWVvdXQ9MywgY2VydD1Ob25lKToKICAgICAgICAjIGlmIHRo
            ZSAnY2x1c3RlcicgaW5zdGFuY2UgaXMgYSBkdW1teSBvbmUsCiAgICAgICAgIyBkb24ndCB0cnkg
            dG8gcmVhY2ggb3V0IHRvIHRoZSBlbmRwb2ludAogICAgICAgIGlmIGlzaW5zdGFuY2Uoc2VsZi5j
            bHVzdGVyLCBEdW1teVJhZG9zKToKICAgICAgICAgICAgcmV0dXJuICIiLCAiIiwgIiIKICAgICAg
            ICBpZiBpcF90eXBlID09ICJJUHY2IjoKICAgICAgICAgICAgdHJ5OgogICAgICAgICAgICAgICAg
            ZW5kcG9pbnRfc3RyX2lwLCBlbmRwb2ludF9zdHJfcG9ydCA9IGVuZHBvaW50X3N0ci5yc3BsaXQo
            IjoiLCAxKQogICAgICAgICAgICBleGNlcHQgVmFsdWVFcnJvcjoKICAgICAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgZiJOb3Qg
            YSBwcm9wZXIgZW5kcG9pbnQ6IHtlbmRwb2ludF9zdHJ9IgogICAgICAgICAgICAgICAgKQogICAg
            ICAgICAgICBpZiBlbmRwb2ludF9zdHJfaXBbMF0gIT0gIlsiOgogICAgICAgICAgICAgICAgZW5k
            cG9pbnRfc3RyX2lwID0gIlsiICsgZW5kcG9pbnRfc3RyX2lwICsgIl0iCiAgICAgICAgICAgIGVu
            ZHBvaW50X3N0ciA9ICI6Ii5qb2luKFtlbmRwb2ludF9zdHJfaXAsIGVuZHBvaW50X3N0cl9wb3J0
            XSkKCiAgICAgICAgcHJvdG9jb2xzID0gWyJodHRwIiwgImh0dHBzIl0KICAgICAgICByZXNwb25z
            ZV9lcnJvciA9IE5vbmUKICAgICAgICBmb3IgcHJlZml4IGluIHByb3RvY29sczoKICAgICAgICAg
            ICAgdHJ5OgogICAgICAgICAgICAgICAgZXAgPSBmIntwcmVmaXh9Oi8ve2VuZHBvaW50X3N0cn0i
            CiAgICAgICAgICAgICAgICB2ZXJpZnkgPSBOb25lCiAgICAgICAgICAgICAgICAjIElmIHZlcmlm
            eSBpcyBzZXQgdG8gYSBwYXRoIHRvIGEgZGlyZWN0b3J5LAogICAgICAgICAgICAgICAgIyB0aGUg
            ZGlyZWN0b3J5IG11c3QgaGF2ZSBiZWVuIHByb2Nlc3NlZCB1c2luZyB0aGUgY19yZWhhc2ggdXRp
            bGl0eSBzdXBwbGllZCB3aXRoIE9wZW5TU0wuCiAgICAgICAgICAgICAgICBpZiBwcmVmaXggPT0g
            Imh0dHBzIiBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfc2tpcF90bHM6CiAgICAgICAgICAgICAg
            ICAgICAgdmVyaWZ5ID0gRmFsc2UKICAgICAgICAgICAgICAgICAgICByID0gcmVxdWVzdHMuaGVh
            ZChlcCwgdGltZW91dD10aW1lb3V0LCB2ZXJpZnk9RmFsc2UpCiAgICAgICAgICAgICAgICBlbGlm
            IHByZWZpeCA9PSAiaHR0cHMiIGFuZCBjZXJ0OgogICAgICAgICAgICAgICAgICAgIHZlcmlmeSA9
            IGNlcnQKICAgICAgICAgICAgICAgICAgICByID0gcmVxdWVzdHMuaGVhZChlcCwgdGltZW91dD10
            aW1lb3V0LCB2ZXJpZnk9Y2VydCkKICAgICAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAg
            ICAgICAgciA9IHJlcXVlc3RzLmhlYWQoZXAsIHRpbWVvdXQ9dGltZW91dCkKICAgICAgICAgICAg
            ICAgIGlmIHIuc3RhdHVzX2NvZGUgPT0gMjAwOgogICAgICAgICAgICAgICAgICAgIHJldHVybiBw
            cmVmaXgsIHZlcmlmeSwgIiIKICAgICAgICAgICAgZXhjZXB0IEV4Y2VwdGlvbiBhcyBlcnI6CiAg
            ICAgICAgICAgICAgICByZXNwb25zZV9lcnJvciA9IGVycgogICAgICAgICAgICAgICAgY29udGlu
            dWUKICAgICAgICBzeXMuc3RkZXJyLndyaXRlKAogICAgICAgICAgICBmInVuYWJsZSB0byBjb25u
            ZWN0IHRvIGVuZHBvaW50OiB7ZW5kcG9pbnRfc3RyfSwgZmFpbGVkIGVycm9yOiB7cmVzcG9uc2Vf
            ZXJyb3J9IgogICAgICAgICkKICAgICAgICByZXR1cm4gKAogICAgICAgICAgICAiIiwKICAgICAg
            ICAgICAgIiIsCiAgICAgICAgICAgICgiLTEiKSwKICAgICAgICApCgogICAgZGVmIF9faW5pdF9f
            KHNlbGYsIGFyZ19saXN0PU5vbmUpOgogICAgICAgIHNlbGYub3V0X21hcCA9IHt9CiAgICAgICAg
            c2VsZi5fZXhjbHVkZWRfa2V5cyA9IHNldCgpCiAgICAgICAgc2VsZi5fYXJnX3BhcnNlciA9IHNl
            bGYuZ2VuX2FyZ19wYXJzZXIoYXJnc190b19wYXJzZT1hcmdfbGlzdCkKICAgICAgICBzZWxmLl9j
            aGVja19jb25mbGljdGluZ19vcHRpb25zKCkKICAgICAgICBzZWxmLnJ1bl9hc191c2VyID0gc2Vs
            Zi5fYXJnX3BhcnNlci5ydW5fYXNfdXNlcgogICAgICAgIHNlbGYub3V0cHV0X2ZpbGUgPSBzZWxm
            Ll9hcmdfcGFyc2VyLm91dHB1dAogICAgICAgIHNlbGYuY2VwaF9jb25mID0gc2VsZi5fYXJnX3Bh
            cnNlci5jZXBoX2NvbmYKICAgICAgICBzZWxmLmNlcGhfa2V5cmluZyA9IHNlbGYuX2FyZ19wYXJz
            ZXIua2V5cmluZwogICAgICAgICMgaWYgdXNlciBub3QgcHJvdmlkZWQsIGdpdmUgYSBkZWZhdWx0
            IHVzZXIKICAgICAgICBpZiBub3Qgc2VsZi5ydW5fYXNfdXNlciBhbmQgbm90IHNlbGYuX2FyZ19w
            YXJzZXIudXBncmFkZToKICAgICAgICAgICAgc2VsZi5ydW5fYXNfdXNlciA9IHNlbGYuRVhURVJO
            QUxfVVNFUl9OQU1FCiAgICAgICAgaWYgbm90IHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJl
            Zml4IGFuZCBub3Qgc2VsZi5fYXJnX3BhcnNlci51cGdyYWRlOgogICAgICAgICAgICBzZWxmLl9h
            cmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeCA9IHNlbGYuREVGQVVMVF9SR1dfUE9PTF9QUkVGSVgK
            ICAgICAgICBpZiBzZWxmLmNlcGhfY29uZjoKICAgICAgICAgICAga3dhcmdzID0ge30KICAgICAg
            ICAgICAgaWYgc2VsZi5jZXBoX2tleXJpbmc6CiAgICAgICAgICAgICAgICBrd2FyZ3NbImNvbmYi
            XSA9IHsia2V5cmluZyI6IHNlbGYuY2VwaF9rZXlyaW5nfQogICAgICAgICAgICBzZWxmLmNsdXN0
            ZXIgPSByYWRvcy5SYWRvcyhjb25mZmlsZT1zZWxmLmNlcGhfY29uZiwgKiprd2FyZ3MpCiAgICAg
            ICAgZWxzZToKICAgICAgICAgICAgc2VsZi5jbHVzdGVyID0gcmFkb3MuUmFkb3MoKQogICAgICAg
            ICAgICBzZWxmLmNsdXN0ZXIuY29uZl9yZWFkX2ZpbGUoKQogICAgICAgIHNlbGYuY2x1c3Rlci5j
            b25uZWN0KCkKCiAgICBkZWYgc2h1dGRvd24oc2VsZik6CiAgICAgICAgaWYgc2VsZi5jbHVzdGVy
            LnN0YXRlID09ICJjb25uZWN0ZWQiOgogICAgICAgICAgICBzZWxmLmNsdXN0ZXIuc2h1dGRvd24o
            KQoKICAgIGRlZiBnZXRfZnNpZChzZWxmKToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRy
            eV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGggZnNpZCIpCiAgICAg
            ICAgcmV0dXJuIHN0cihzZWxmLmNsdXN0ZXIuZ2V0X2ZzaWQoKSkKCiAgICBkZWYgX2NvbW1vbl9j
            bWRfanNvbl9nZW4oc2VsZiwgY21kX2pzb24pOgogICAgICAgIGNtZCA9IGpzb24uZHVtcHMoY21k
            X2pzb24sIHNvcnRfa2V5cz1UcnVlKQogICAgICAgIHJldF92YWwsIGNtZF9vdXQsIGVycl9tc2cg
            PSBzZWxmLmNsdXN0ZXIubW9uX2NvbW1hbmQoY21kLCBiIiIpCiAgICAgICAgaWYgc2VsZi5fYXJn
            X3BhcnNlci52ZXJib3NlOgogICAgICAgICAgICBwcmludChmIkNvbW1hbmQgSW5wdXQ6IHtjbWR9
            IikKICAgICAgICAgICAgcHJpbnQoCiAgICAgICAgICAgICAgICBmIlJldHVybiBWYWw6IHtyZXRf
            dmFsfVxuQ29tbWFuZCBPdXRwdXQ6IHtjbWRfb3V0fVxuIgogICAgICAgICAgICAgICAgZiJFcnJv
            ciBNZXNzYWdlOiB7ZXJyX21zZ31cbi0tLS0tLS0tLS1cbiIKICAgICAgICAgICAgKQogICAgICAg
            IGpzb25fb3V0ID0ge30KICAgICAgICAjIGlmIHRoZXJlIGlzIG5vIGVycm9yIChpLmU7IHJldF92
            YWwgaXMgWkVSTykgYW5kICdjbWRfb3V0JyBpcyBub3QgZW1wdHkKICAgICAgICAjIHRoZW4gY29u
            dmVydCAnY21kX291dCcgdG8gYSBqc29uIG91dHB1dAogICAgICAgIGlmIHJldF92YWwgPT0gMCBh
            bmQgY21kX291dDoKICAgICAgICAgICAganNvbl9vdXQgPSBqc29uLmxvYWRzKGNtZF9vdXQpCiAg
            ICAgICAgcmV0dXJuIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnCgogICAgZGVmIGdldF9jZXBo
            X2V4dGVybmFsX21vbl9kYXRhKHNlbGYpOgogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAi
            cXVvcnVtX3N0YXR1cyIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgaWYgc2VsZi5fYXJnX3Bh
            cnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKCJjZXBoICIgKyBj
            bWRfanNvblsicHJlZml4Il0pCiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBz
            ZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICMgaWYgdGhlcmUgaXMg
            YW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihq
            c29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgICAgICIncXVvcnVtX3N0YXR1cycgY29tbWFuZCBmYWlsZWQuXG4iCiAg
            ICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxm
            LkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgKQogICAgICAgIHFfbGVhZGVyX25hbWUg
            PSBqc29uX291dFsicXVvcnVtX2xlYWRlcl9uYW1lIl0KICAgICAgICBxX2xlYWRlcl9kZXRhaWxz
            ID0ge30KICAgICAgICBxX2xlYWRlcl9tYXRjaGluZ19saXN0ID0gWwogICAgICAgICAgICBsIGZv
            ciBsIGluIGpzb25fb3V0WyJtb25tYXAiXVsibW9ucyJdIGlmIGxbIm5hbWUiXSA9PSBxX2xlYWRl
            cl9uYW1lCiAgICAgICAgXQogICAgICAgIGlmIGxlbihxX2xlYWRlcl9tYXRjaGluZ19saXN0KSA9
            PSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKCJObyBtYXRj
            aGluZyAnbW9uJyBkZXRhaWxzIGZvdW5kIikKICAgICAgICBxX2xlYWRlcl9kZXRhaWxzID0gcV9s
            ZWFkZXJfbWF0Y2hpbmdfbGlzdFswXQogICAgICAgICMgZ2V0IHRoZSBhZGRyZXNzIHZlY3RvciBv
            ZiB0aGUgcXVvcnVtLWxlYWRlcgogICAgICAgIHFfbGVhZGVyX2FkZHJ2ZWMgPSBxX2xlYWRlcl9k
            ZXRhaWxzLmdldCgicHVibGljX2FkZHJzIiwge30pLmdldCgiYWRkcnZlYyIsIFtdKQogICAgICAg
            ICMgaWYgdGhlIHF1b3J1bS1sZWFkZXIgaGFzIG9ubHkgb25lIGFkZHJlc3MgaW4gdGhlIGFkZHJl
            c3MtdmVjdG9yCiAgICAgICAgIyBhbmQgaXQgaXMgb2YgdHlwZSAndjInIChpZTsgd2l0aCA8SVA+
            OjMzMDApLAogICAgICAgICMgcmFpc2UgYW4gZXhjZXB0aW9uIHRvIG1ha2UgdXNlciBhd2FyZSB0
            aGF0CiAgICAgICAgIyB0aGV5IGhhdmUgdG8gZW5hYmxlICd2MScgKGllOyB3aXRoIDxJUD46Njc4
            OSkgdHlwZSBhcyB3ZWxsCiAgICAgICAgaWYgbGVuKHFfbGVhZGVyX2FkZHJ2ZWMpID09IDEgYW5k
            IHFfbGVhZGVyX2FkZHJ2ZWNbMF1bInR5cGUiXSA9PSAidjIiOgogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgIk9ubHkgJ3YyJyBhZGRy
            ZXNzIHR5cGUgaXMgZW5hYmxlZCwgdXNlciBzaG91bGQgYWxzbyBlbmFibGUgJ3YxJyB0eXBlIGFz
            IHdlbGwiCiAgICAgICAgICAgICkKICAgICAgICBpcF9hZGRyID0gc3RyKHFfbGVhZGVyX2RldGFp
            bHNbInB1YmxpY19hZGRyIl0uc3BsaXQoIi8iKVswXSkKCiAgICAgICAgaWYgc2VsZi5fYXJnX3Bh
            cnNlci52Ml9wb3J0X2VuYWJsZToKICAgICAgICAgICAgaWYgbGVuKHFfbGVhZGVyX2FkZHJ2ZWMp
            ID4gMToKICAgICAgICAgICAgICAgIGlmIHFfbGVhZGVyX2FkZHJ2ZWNbMF1bInR5cGUiXSA9PSAi
            djIiOgogICAgICAgICAgICAgICAgICAgIGlwX2FkZHIgPSBxX2xlYWRlcl9hZGRydmVjWzBdWyJh
            ZGRyIl0KICAgICAgICAgICAgICAgIGVsaWYgcV9sZWFkZXJfYWRkcnZlY1sxXVsidHlwZSJdID09
            ICJ2MiI6CiAgICAgICAgICAgICAgICAgICAgaXBfYWRkciA9IHFfbGVhZGVyX2FkZHJ2ZWNbMV1b
            ImFkZHIiXQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0
            ZSgKICAgICAgICAgICAgICAgICAgICAiJ3YyJyBhZGRyZXNzIHR5cGUgbm90IHByZXNlbnQsIGFu
            ZCAndjItcG9ydC1lbmFibGUnIGZsYWcgaXMgcHJvdmlkZWQiCiAgICAgICAgICAgICAgICApCgog
            ICAgICAgIHJldHVybiBmIntzdHIocV9sZWFkZXJfbmFtZSl9PXtpcF9hZGRyfSIKCiAgICBkZWYg
            X2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoc2VsZiwgaG9zdF9uYW1lLCBwb3J0LCBpcF90eXBlKToK
            ICAgICAgICAjIGlmICdjbHVzdGVyJyBpbnN0YW5jZSBpcyBhIGR1bW15IHR5cGUsCiAgICAgICAg
            IyBjYWxsIHRoZSBkdW1teSBpbnN0YW5jZSdzICJjb252ZXJ0IiBtZXRob2QKICAgICAgICBpZiBu
            b3QgaG9zdF9uYW1lOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KCJFbXB0eSBob3N0bmFtZSBwcm92aWRlZCIpCiAgICAgICAgaWYgaXNpbnN0YW5jZShzZWxmLmNs
            dXN0ZXIsIER1bW15UmFkb3MpOgogICAgICAgICAgICByZXR1cm4gc2VsZi5jbHVzdGVyLl9jb252
            ZXJ0X2hvc3RuYW1lX3RvX2lwKGhvc3RfbmFtZSkKCiAgICAgICAgaWYgaXBfdHlwZSA9PSAiRlFE
            TiI6CiAgICAgICAgICAgICMgY2hlY2sgd2hpY2ggaXAgRlFETiBzaG91bGQgYmUgY29udmVydGVk
            IHRvLCBJUHY0IG9yIElQdjYKICAgICAgICAgICAgIyBjaGVjayB0aGUgaG9zdCBpcCwgdGhlIGVu
            ZHBvaW50IGlwIHR5cGUgd291bGQgYmUgc2ltaWxhciB0byBob3N0IGlwCiAgICAgICAgICAgIGNt
            ZF9qc29uID0geyJwcmVmaXgiOiAib3JjaCBob3N0IGxzIiwgImZvcm1hdCI6ICJqc29uIn0KICAg
            ICAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pz
            b25fZ2VuKGNtZF9qc29uKQogICAgICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1
            bCBhdHRlbXB0LAogICAgICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9
            PSAwOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAg
            ICAgICAgICAgICAgICAgICAiJ29yY2ggaG9zdCBscycgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAg
            ICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYgcmV0X3ZhbCAhPSAwIGVsc2Ugc2Vs
            Zi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGhvc3Rf
            YWRkciA9IGpzb25fb3V0WzBdWyJhZGRyIl0KICAgICAgICAgICAgIyBhZGQgOjgwIHNhbXBsZSBw
            b3J0IGluIGlwX3R5cGUsIGFzIF9pbnZhbGlkX2VuZHBvaW50IGFsc28gdmVyaWZ5IHBvcnQKICAg
            ICAgICAgICAgaG9zdF9pcF90eXBlID0gc2VsZi5faW52YWxpZF9lbmRwb2ludChob3N0X2FkZHIg
            KyAiOjgwIikKICAgICAgICAgICAgaW1wb3J0IHNvY2tldAoKICAgICAgICAgICAgIyBleGFtcGxl
            IG91dHB1dCBbKDxBZGRyZXNzRmFtaWx5LkFGX0lORVQ6IDI+LCA8U29ja2V0S2luZC5TT0NLX1NU
            UkVBTTogMT4sIDYsICcnLCAoJzkzLjE4NC4yMTYuMzQnLCA4MCkpLCAuLi5dCiAgICAgICAgICAg
            ICMgd2UgbmVlZCB0byBnZXQgOTMuMTg0LjIxNi4zNCBzbyBpdCB3b3VsZCBiZSBpcFswXVs0XVsw
            XQogICAgICAgICAgICBpZiBob3N0X2lwX3R5cGUgPT0gIklQdjYiOgogICAgICAgICAgICAgICAg
            aXAgPSBzb2NrZXQuZ2V0YWRkcmluZm8oCiAgICAgICAgICAgICAgICAgICAgaG9zdF9uYW1lLCBw
            b3J0LCBmYW1pbHk9c29ja2V0LkFGX0lORVQ2LCBwcm90bz1zb2NrZXQuSVBQUk9UT19UQ1AKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxpZiBob3N0X2lwX3R5cGUgPT0gIklQdjQiOgog
            ICAgICAgICAgICAgICAgaXAgPSBzb2NrZXQuZ2V0YWRkcmluZm8oCiAgICAgICAgICAgICAgICAg
            ICAgaG9zdF9uYW1lLCBwb3J0LCBmYW1pbHk9c29ja2V0LkFGX0lORVQsIHByb3RvPXNvY2tldC5J
            UFBST1RPX1RDUAogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBkZWwgc29ja2V0CiAgICAg
            ICAgICAgIHJldHVybiBpcFswXVs0XVswXQogICAgICAgIHJldHVybiBob3N0X25hbWUKCiAgICBk
            ZWYgZ2V0X2FjdGl2ZV9hbmRfc3RhbmRieV9tZ3JzKHNlbGYpOgogICAgICAgIGlmIHNlbGYuX2Fy
            Z19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuICIiLCBzZWxmLmRyeV9ydW4oImNl
            cGggc3RhdHVzIikKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQgPSBzZWxmLl9hcmdf
            cGFyc2VyLm1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydAogICAgICAgIG1vbml0b3JpbmdfZW5kcG9p
            bnRfaXBfbGlzdCA9IHNlbGYuX2FyZ19wYXJzZXIubW9uaXRvcmluZ19lbmRwb2ludAogICAgICAg
            IHN0YW5kYnlfbWdycyA9IFtdCiAgICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBf
            bGlzdDoKICAgICAgICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJzdGF0dXMiLCAiZm9ybWF0
            IjogImpzb24ifQogICAgICAgICAgICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYu
            X2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgICAgICMgaWYgdGhlcmUgaXMg
            YW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBs
            ZW4oanNvbl9vdXQpID09IDA6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJl
            RXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICInbWdyIHNlcnZpY2VzJyBjb21tYW5kIGZh
            aWxlZC5cbiIKICAgICAgICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFs
            ICE9IDAgZWxzZSBzZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgICAgICkKICAg
            ICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludCA9ICgKICAgICAgICAgICAgICAgIGpzb25fb3V0
            LmdldCgibWdybWFwIiwge30pLmdldCgic2VydmljZXMiLCB7fSkuZ2V0KCJwcm9tZXRoZXVzIiwg
            IiIpCiAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnQ6
            CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAg
            ICAgICAgICAgICAgICJjYW4ndCBmaW5kIG1vbml0b3JpbmdfZW5kcG9pbnQsIHByb21ldGhldXMg
            bW9kdWxlIG1pZ2h0IG5vdCBiZSBlbmFibGVkLCAiCiAgICAgICAgICAgICAgICAgICAgImVuYWJs
            ZSB0aGUgbW9kdWxlIGJ5IHJ1bm5pbmcgJ2NlcGggbWdyIG1vZHVsZSBlbmFibGUgcHJvbWV0aGV1
            cyciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICMgbm93IGNoZWNrIHRoZSBzdGFuZC1i
            eSBtZ3ItcwogICAgICAgICAgICBzdGFuZGJ5X2FyciA9IGpzb25fb3V0LmdldCgibWdybWFwIiwg
            e30pLmdldCgic3RhbmRieXMiLCBbXSkKICAgICAgICAgICAgZm9yIGVhY2hfc3RhbmRieSBpbiBz
            dGFuZGJ5X2FycjoKICAgICAgICAgICAgICAgIGlmICJuYW1lIiBpbiBlYWNoX3N0YW5kYnkua2V5
            cygpOgogICAgICAgICAgICAgICAgICAgIHN0YW5kYnlfbWdycy5hcHBlbmQoZWFjaF9zdGFuZGJ5
            WyJuYW1lIl0pCiAgICAgICAgICAgIHRyeToKICAgICAgICAgICAgICAgIHBhcnNlZF9lbmRwb2lu
            dCA9IHVybHBhcnNlKG1vbml0b3JpbmdfZW5kcG9pbnQpCiAgICAgICAgICAgIGV4Y2VwdCBWYWx1
            ZUVycm9yOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigK
            ICAgICAgICAgICAgICAgICAgICBmImludmFsaWQgZW5kcG9pbnQ6IHttb25pdG9yaW5nX2VuZHBv
            aW50fSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9p
            cF9saXN0ID0gcGFyc2VkX2VuZHBvaW50Lmhvc3RuYW1lCiAgICAgICAgICAgIGlmIG5vdCBtb25p
            dG9yaW5nX2VuZHBvaW50X3BvcnQ6CiAgICAgICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50
            X3BvcnQgPSBzdHIocGFyc2VkX2VuZHBvaW50LnBvcnQpCgogICAgICAgICMgaWYgbW9uaXRvcmlu
            ZyBlbmRwb2ludCBwb3J0IGlzIG5vdCBzZXQsIHB1dCBhIGRlZmF1bHQgbW9uIHBvcnQKICAgICAg
            ICBpZiBub3QgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0OgogICAgICAgICAgICBtb25pdG9yaW5n
            X2VuZHBvaW50X3BvcnQgPSBzZWxmLkRFRkFVTFRfTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUCgog
            ICAgICAgICMgdXNlciBjb3VsZCBnaXZlIGNvbW1hIGFuZCBzcGFjZSBzZXBhcmF0ZWQgaW5wdXRz
            IChsaWtlIC0tbW9uaXRvcmluZy1lbmRwb2ludD0iPGlwMT4sIDxpcDI+IikKICAgICAgICBtb25p
            dG9yaW5nX2VuZHBvaW50X2lwX2xpc3QgPSBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3QucmVw
            bGFjZSgiLCIsICIgIikKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXQg
            PSBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Quc3BsaXQoKQogICAgICAgICMgaWYgbW9uaXRv
            cmluZy1lbmRwb2ludCBjb3VsZCBub3QgYmUgZm91bmQsIHJhaXNlIGFuIGVycm9yCiAgICAgICAg
            aWYgbGVuKG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdF9zcGxpdCkgPT0gMDoKICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigiTm8gJ21vbml0b3JpbmctZW5kcG9p
            bnQnIGZvdW5kIikKICAgICAgICAjIGZpcnN0IGlwIGlzIHRyZWF0ZWQgYXMgdGhlIG1haW4gbW9u
            aXRvcmluZy1lbmRwb2ludAogICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfaXAgPSBtb25pdG9y
            aW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXRbMF0KICAgICAgICAjIHJlc3Qgb2YgdGhlIGlwLXMg
            YXJlIGFkZGVkIHRvIHRoZSAnc3RhbmRieV9tZ3JzJyBsaXN0CiAgICAgICAgc3RhbmRieV9tZ3Jz
            LmV4dGVuZChtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXRbMTpdKQogICAgICAgIGZh
            aWxlZF9pcCA9IG1vbml0b3JpbmdfZW5kcG9pbnRfaXAKCiAgICAgICAgbW9uaXRvcmluZ19lbmRw
            b2ludCA9ICI6Ii5qb2luKAogICAgICAgICAgICBbbW9uaXRvcmluZ19lbmRwb2ludF9pcCwgbW9u
            aXRvcmluZ19lbmRwb2ludF9wb3J0XQogICAgICAgICkKICAgICAgICBpcF90eXBlID0gc2VsZi5f
            aW52YWxpZF9lbmRwb2ludChtb25pdG9yaW5nX2VuZHBvaW50KQogICAgICAgIHRyeToKICAgICAg
            ICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcCA9IHNlbGYuX2NvbnZlcnRfaG9zdG5hbWVfdG9f
            aXAoCiAgICAgICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X2lwLCBtb25pdG9yaW5nX2Vu
            ZHBvaW50X3BvcnQsIGlwX3R5cGUKICAgICAgICAgICAgKQogICAgICAgICAgICAjIGNvbGxlY3Qg
            YWxsIHRoZSAnc3RhbmQtYnknIG1nciBpcHMKICAgICAgICAgICAgbWdyX2lwcyA9IFtdCiAgICAg
            ICAgICAgIGZvciBlYWNoX3N0YW5kYnlfbWdyIGluIHN0YW5kYnlfbWdyczoKICAgICAgICAgICAg
            ICAgIGZhaWxlZF9pcCA9IGVhY2hfc3RhbmRieV9tZ3IKICAgICAgICAgICAgICAgIG1ncl9pcHMu
            YXBwZW5kKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAo
            CiAgICAgICAgICAgICAgICAgICAgICAgIGVhY2hfc3RhbmRieV9tZ3IsIG1vbml0b3JpbmdfZW5k
            cG9pbnRfcG9ydCwgaXBfdHlwZQogICAgICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAg
            ICkKICAgICAgICBleGNlcHQ6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNl
            cHRpb24oCiAgICAgICAgICAgICAgICBmIkNvbnZlcnNpb24gb2YgaG9zdDoge2ZhaWxlZF9pcH0g
            dG8gSVAgZmFpbGVkLiAiCiAgICAgICAgICAgICAgICAiUGxlYXNlIGVudGVyIHRoZSBJUCBhZGRy
            ZXNzZXMgb2YgYWxsIHRoZSBjZXBoLW1ncnMgd2l0aCB0aGUgJy0tbW9uaXRvcmluZy1lbmRwb2lu
            dCcgZmxhZyIKICAgICAgICAgICAgKQoKICAgICAgICBfLCBfLCBlcnIgPSBzZWxmLmVuZHBvaW50
            X2RpYWwobW9uaXRvcmluZ19lbmRwb2ludCwgaXBfdHlwZSkKICAgICAgICBpZiBlcnIgPT0gIi0x
            IjoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihlcnIpCiAgICAg
            ICAgIyBhZGQgdGhlIHZhbGlkYXRlZCBhY3RpdmUgbWdyIElQIGludG8gdGhlIGZpcnN0IGluZGV4
            CiAgICAgICAgbWdyX2lwcy5pbnNlcnQoMCwgbW9uaXRvcmluZ19lbmRwb2ludF9pcCkKICAgICAg
            ICBhbGxfbWdyX2lwc19zdHIgPSAiLCIuam9pbihtZ3JfaXBzKQogICAgICAgIHJldHVybiBhbGxf
            bWdyX2lwc19zdHIsIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydAoKICAgIGRlZiBjaGVja191c2Vy
            X2V4aXN0KHNlbGYsIHVzZXIpOgogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAiYXV0aCBn
            ZXQiLCAiZW50aXR5IjogZiJ7dXNlcn0iLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIHJldF92
            YWwsIGpzb25fb3V0LCBfID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAg
            ICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICBy
            ZXR1cm4gIiIKICAgICAgICByZXR1cm4gc3RyKGpzb25fb3V0WzBdWyJrZXkiXSkKCiAgICBkZWYg
            Z2V0X2NlcGhmc19wcm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50
            aXR5ID0gImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIgogICAgICAgIGNhcHMgPSB7CiAg
            ICAgICAgICAgICJtb24iOiAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCci
            LAogICAgICAgICAgICAibWdyIjogImFsbG93IHJ3IiwKICAgICAgICAgICAgIm9zZCI6ICJhbGxv
            dyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiLAogICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9h
            cmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uOgogICAgICAgICAgICBrOHNfY2x1
            c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1lCiAgICAgICAgICAg
            IGlmIGs4c19jbHVzdGVyX25hbWUgPT0gIiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRp
            b25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJrOHNfY2x1c3Rlcl9uYW1l
            IG5vdCBmb3VuZCwgcGxlYXNlIHNldCB0aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW0gPSBzZWxmLl9hcmdf
            cGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICAgICAgaWYgY2VwaGZzX2ZpbGVz
            eXN0ZW0gPT0gIiI6CiAgICAgICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1
            c3Rlcl9uYW1lfSIKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYi
            e2VudGl0eX0te2s4c19jbHVzdGVyX25hbWV9LXtjZXBoZnNfZmlsZXN5c3RlbX0iCiAgICAgICAg
            ICAgICAgICBjYXBzWyJvc2QiXSA9IGYiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT17Y2Vw
            aGZzX2ZpbGVzeXN0ZW19IgoKICAgICAgICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVmIGdl
            dF9jZXBoZnNfbm9kZV9jYXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gImNs
            aWVudC5jc2ktY2VwaGZzLW5vZGUiCiAgICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6
            ICJhbGxvdyByLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJt
            Z3IiOiAiYWxsb3cgcnciLAogICAgICAgICAgICAib3NkIjogImFsbG93IHJ3IHRhZyBjZXBoZnMg
            Kj0qIiwKICAgICAgICAgICAgIm1kcyI6ICJhbGxvdyBydyIsCiAgICAgICAgfQogICAgICAgIGlm
            IHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAg
            IGs4c19jbHVzdGVyX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAg
            ICAgICAgICAgaWYgazhzX2NsdXN0ZXJfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIms4c19jbHVz
            dGVyX25hbWUgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1rOHMtY2x1c3Rlci1uYW1lJyBm
            bGFnIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBjZXBoZnNfZmlsZXN5c3RlbSA9IHNl
            bGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgICAgICBpZiBjZXBo
            ZnNfZmlsZXN5c3RlbSA9PSAiIjoKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0t
            e2s4c19jbHVzdGVyX25hbWV9IgogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgZW50
            aXR5ID0gZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0te2NlcGhmc19maWxlc3lzdGVtfSIK
            ICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJdID0gZiJhbGxvdyBydyB0YWcgY2VwaGZzICo9e2Nl
            cGhmc19maWxlc3lzdGVtfSIKCiAgICAgICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBn
            ZXRfZW50aXR5KAogICAgICAgIHNlbGYsCiAgICAgICAgZW50aXR5LAogICAgICAgIHJiZF9wb29s
            X25hbWUsCiAgICAgICAgYWxpYXNfcmJkX3Bvb2xfbmFtZSwKICAgICAgICBrOHNfY2x1c3Rlcl9u
            YW1lLAogICAgICAgIHJhZG9zX25hbWVzcGFjZSwKICAgICk6CiAgICAgICAgaWYgKAogICAgICAg
            ICAgICByYmRfcG9vbF9uYW1lLmNvdW50KCIuIikgIT0gMAogICAgICAgICAgICBvciByYmRfcG9v
            bF9uYW1lLmNvdW50KCJfIikgIT0gMAogICAgICAgICAgICBvciBhbGlhc19yYmRfcG9vbF9uYW1l
            ICE9ICIiCiAgICAgICAgICAgICMgY2hlY2tpbmcgYWxpYXNfcmJkX3Bvb2xfbmFtZSBpcyBub3Qg
            ZW1wdHkgYXMgdGhlcmUgbWF5YmUgYSBzcGVjaWFsIGNoYXJhY3RlciB1c2VkIG90aGVyIHRoYW4g
            LiBvciBfCiAgICAgICAgKToKICAgICAgICAgICAgaWYgYWxpYXNfcmJkX3Bvb2xfbmFtZSA9PSAi
            IjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAg
            ICAgICAgICAgICAgICAgInBsZWFzZSBzZXQgdGhlICctLWFsaWFzLXJiZC1kYXRhLXBvb2wtbmFt
            ZScgZmxhZyBhcyB0aGUgcmJkIGRhdGEgcG9vbCBuYW1lIGNvbnRhaW5zICcuJyBvciAnXyciCiAg
            ICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmICgKICAgICAgICAgICAgICAgIGFsaWFzX3Ji
            ZF9wb29sX25hbWUuY291bnQoIi4iKSAhPSAwCiAgICAgICAgICAgICAgICBvciBhbGlhc19yYmRf
            cG9vbF9uYW1lLmNvdW50KCJfIikgIT0gMAogICAgICAgICAgICApOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJy0t
            YWxpYXMtcmJkLWRhdGEtcG9vbC1uYW1lJyBmbGFnIHZhbHVlIHNob3VsZCBub3QgY29udGFpbiAn
            Licgb3IgJ18nIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRp
            dHl9LXtrOHNfY2x1c3Rlcl9uYW1lfS17YWxpYXNfcmJkX3Bvb2xfbmFtZX0iCiAgICAgICAgZWxz
            ZToKICAgICAgICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0te3Ji
            ZF9wb29sX25hbWV9IgoKICAgICAgICBpZiByYWRvc19uYW1lc3BhY2U6CiAgICAgICAgICAgIGVu
            dGl0eSA9IGYie2VudGl0eX0te3JhZG9zX25hbWVzcGFjZX0iCiAgICAgICAgcmV0dXJuIGVudGl0
            eQoKICAgIGRlZiBnZXRfcmJkX3Byb3Zpc2lvbmVyX2NhcHNfYW5kX2VudGl0eShzZWxmKToKICAg
            ICAgICBlbnRpdHkgPSAiY2xpZW50LmNzaS1yYmQtcHJvdmlzaW9uZXIiCiAgICAgICAgY2FwcyA9
            IHsKICAgICAgICAgICAgIm1vbiI6ICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJs
            b2NrbGlzdCciLAogICAgICAgICAgICAibWdyIjogImFsbG93IHJ3IiwKICAgICAgICAgICAgIm9z
            ZCI6ICJwcm9maWxlIHJiZCIsCiAgICAgICAgfQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIu
            cmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBz
            ZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgICAgICBhbGlhc19yYmRf
            cG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5hbGlhc19yYmRfZGF0YV9wb29sX25hbWUKICAg
            ICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJf
            bmFtZQogICAgICAgICAgICByYWRvc19uYW1lc3BhY2UgPSBzZWxmLl9hcmdfcGFyc2VyLnJhZG9z
            X25hbWVzcGFjZQogICAgICAgICAgICBpZiByYmRfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAg
            ICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAg
            ICAibWFuZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1yYmQtZGF0YS1w
            b29sLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmIGs4c19jbHVz
            dGVyX25hbWUgPT0gIiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhj
            ZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBs
            ZWFzZSBzZXQgdGhlICctLWs4cy1jbHVzdGVyLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgIGVudGl0eSA9IHNlbGYuZ2V0X2VudGl0eSgKICAgICAgICAgICAgICAgIGVu
            dGl0eSwKICAgICAgICAgICAgICAgIHJiZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBhbGlh
            c19yYmRfcG9vbF9uYW1lLAogICAgICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAg
            ICAgICAgICAgIHJhZG9zX25hbWVzcGFjZSwKICAgICAgICAgICAgKQogICAgICAgICAgICBpZiBy
            YWRvc19uYW1lc3BhY2UgIT0gIiI6CiAgICAgICAgICAgICAgICBjYXBzWyJvc2QiXSA9ICgKICAg
            ICAgICAgICAgICAgICAgICBmInByb2ZpbGUgcmJkIHBvb2w9e3JiZF9wb29sX25hbWV9IG5hbWVz
            cGFjZT17cmFkb3NfbmFtZXNwYWNlfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxz
            ZToKICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJdID0gZiJwcm9maWxlIHJiZCBwb29sPXtyYmRf
            cG9vbF9uYW1lfSIKCiAgICAgICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfcmJk
            X25vZGVfY2Fwc19hbmRfZW50aXR5KHNlbGYpOgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3Np
            LXJiZC1ub2RlIgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAgICJtb24iOiAicHJvZmlsZSBy
            YmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwKICAgICAgICAgICAgIm9zZCI6ICJw
            cm9maWxlIHJiZCIsCiAgICAgICAgfQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJp
            Y3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9h
            cmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9u
            YW1lID0gc2VsZi5fYXJnX3BhcnNlci5hbGlhc19yYmRfZGF0YV9wb29sX25hbWUKICAgICAgICAg
            ICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQog
            ICAgICAgICAgICByYWRvc19uYW1lc3BhY2UgPSBzZWxmLl9hcmdfcGFyc2VyLnJhZG9zX25hbWVz
            cGFjZQogICAgICAgICAgICBpZiByYmRfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAibWFu
            ZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1yYmQtZGF0YS1wb29sLW5h
            bWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmIGs4c19jbHVzdGVyX25h
            bWUgPT0gIiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KAogICAgICAgICAgICAgICAgICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBz
            ZXQgdGhlICctLWs4cy1jbHVzdGVyLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAg
            ICAgICAgIGVudGl0eSA9IHNlbGYuZ2V0X2VudGl0eSgKICAgICAgICAgICAgICAgIGVudGl0eSwK
            ICAgICAgICAgICAgICAgIHJiZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBhbGlhc19yYmRf
            cG9vbF9uYW1lLAogICAgICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAgICAgICAg
            ICAgIHJhZG9zX25hbWVzcGFjZSwKICAgICAgICAgICAgKQogICAgICAgICAgICBpZiByYWRvc19u
            YW1lc3BhY2UgIT0gIiI6CiAgICAgICAgICAgICAgICBjYXBzWyJvc2QiXSA9ICgKICAgICAgICAg
            ICAgICAgICAgICBmInByb2ZpbGUgcmJkIHBvb2w9e3JiZF9wb29sX25hbWV9IG5hbWVzcGFjZT17
            cmFkb3NfbmFtZXNwYWNlfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxzZToKICAg
            ICAgICAgICAgICAgIGNhcHNbIm9zZCJdID0gZiJwcm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9u
            YW1lfSIKCiAgICAgICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfZGVmYXVsdFVz
            ZXJfY2Fwc19hbmRfZW50aXR5KHNlbGYpOgogICAgICAgIGVudGl0eSA9IHNlbGYucnVuX2FzX3Vz
            ZXIKICAgICAgICBjYXBzID0gewogICAgICAgICAgICAibW9uIjogImFsbG93IHIsIGFsbG93IGNv
            bW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFuZCB2ZXJzaW9uIiwKICAgICAgICAgICAg
            Im1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsCiAgICAgICAgICAgICJvc2QiOiBmInByb2Zp
            bGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9e3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bv
            b2xfcHJlZml4fS5yZ3cubWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9v
            bD17c2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5jb250cm9sLCBhbGxvdyBy
            eCBwb29sPXtzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmxvZywgYWxsb3cg
            eCBwb29sPXtzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmJ1Y2tldHMuaW5k
            ZXgiLAogICAgICAgIH0KCiAgICAgICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRf
            Y2Fwc19hbmRfZW50aXR5KHNlbGYsIHVzZXJfbmFtZSk6CiAgICAgICAgaWYgImNsaWVudC5jc2kt
            Y2VwaGZzLXByb3Zpc2lvbmVyIiBpbiB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQu
            Y3NpLWNlcGhmcy1wcm92aXNpb25lciIgIT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAg
            ICAgcmV0dXJuIHNlbGYuZ2V0X2NlcGhmc19wcm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoKQog
            ICAgICAgIGlmICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiBpbiB1c2VyX25hbWU6CiAgICAgICAg
            ICAgIGlmICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiAhPSB1c2VyX25hbWU6CiAgICAgICAgICAg
            ICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQog
            ICAgICAgICAgICByZXR1cm4gc2VsZi5nZXRfY2VwaGZzX25vZGVfY2Fwc19hbmRfZW50aXR5KCkK
            ICAgICAgICBpZiAiY2xpZW50LmNzaS1yYmQtcHJvdmlzaW9uZXIiIGluIHVzZXJfbmFtZToKICAg
            ICAgICAgICAgaWYgImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiAhPSB1c2VyX25hbWU6CiAg
            ICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9u
            ID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5nZXRfcmJkX3Byb3Zpc2lvbmVyX2NhcHNf
            YW5kX2VudGl0eSgpCiAgICAgICAgaWYgImNsaWVudC5jc2ktcmJkLW5vZGUiIGluIHVzZXJfbmFt
            ZToKICAgICAgICAgICAgaWYgImNsaWVudC5jc2ktcmJkLW5vZGUiICE9IHVzZXJfbmFtZToKICAg
            ICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24g
            PSBUcnVlCiAgICAgICAgICAgIHJldHVybiBzZWxmLmdldF9yYmRfbm9kZV9jYXBzX2FuZF9lbnRp
            dHkoKQogICAgICAgIGlmICJjbGllbnQuaGVhbHRoY2hlY2tlciIgaW4gdXNlcl9uYW1lOgogICAg
            ICAgICAgICBpZiAiY2xpZW50LmhlYWx0aGNoZWNrZXIiICE9IHVzZXJfbmFtZToKICAgICAgICAg
            ICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVl
            CiAgICAgICAgICAgIHJldHVybiBzZWxmLmdldF9kZWZhdWx0VXNlcl9jYXBzX2FuZF9lbnRpdHko
            KQoKICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICBm
            Im5vIHVzZXIgZm91bmQgd2l0aCB1c2VyX25hbWU6IHt1c2VyX25hbWV9LCAiCiAgICAgICAgICAg
            ICJnZXRfY2Fwc19hbmRfZW50aXR5IGNvbW1hbmQgZmFpbGVkLlxuIgogICAgICAgICkKCiAgICBk
            ZWYgY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3VzZXIoc2VsZiwgdXNlcik6CiAgICAgICAgIiIiCiAg
            ICAgICAgY29tbWFuZDogY2VwaCBhdXRoIGdldC1vci1jcmVhdGUgY2xpZW50LmNzaS1jZXBoZnMt
            cHJvdmlzaW9uZXIgbW9uICdhbGxvdyByJyBtZ3IgJ2FsbG93IHJ3JyBvc2QgJ2FsbG93IHJ3IHRh
            ZyBjZXBoZnMgbWV0YWRhdGE9KicKICAgICAgICAiIiIKICAgICAgICBjYXBzLCBlbnRpdHkgPSBz
            ZWxmLmdldF9jYXBzX2FuZF9lbnRpdHkodXNlcikKICAgICAgICBjbWRfanNvbiA9IHsKICAgICAg
            ICAgICAgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUiLAogICAgICAgICAgICAiZW50aXR5
            IjogZW50aXR5LAogICAgICAgICAgICAiY2FwcyI6IFtjYXAgZm9yIGNhcF9saXN0IGluIGxpc3Qo
            Y2Fwcy5pdGVtcygpKSBmb3IgY2FwIGluIGNhcF9saXN0XSwKICAgICAgICAgICAgImZvcm1hdCI6
            ICJqc29uIiwKICAgICAgICB9CgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoK
            ICAgICAgICAgICAgcmV0dXJuICgKICAgICAgICAgICAgICAgIHNlbGYuZHJ5X3J1bigKICAgICAg
            ICAgICAgICAgICAgICAiY2VwaCAiCiAgICAgICAgICAgICAgICAgICAgKyBjbWRfanNvblsicHJl
            Zml4Il0KICAgICAgICAgICAgICAgICAgICArICIgIgogICAgICAgICAgICAgICAgICAgICsgY21k
            X2pzb25bImVudGl0eSJdCiAgICAgICAgICAgICAgICAgICAgKyAiICIKICAgICAgICAgICAgICAg
            ICAgICArICIgIi5qb2luKGNtZF9qc29uWyJjYXBzIl0pCiAgICAgICAgICAgICAgICApLAogICAg
            ICAgICAgICAgICAgIiIsCiAgICAgICAgICAgICkKICAgICAgICAjIGNoZWNrIGlmIHVzZXIgYWxy
            ZWFkeSBleGlzdAogICAgICAgIHVzZXJfa2V5ID0gc2VsZi5jaGVja191c2VyX2V4aXN0KGVudGl0
            eSkKICAgICAgICBpZiB1c2VyX2tleSAhPSAiIjoKICAgICAgICAgICAgcmV0dXJuIHVzZXJfa2V5
            LCBmIntlbnRpdHkuc3BsaXQoJy4nLCAxKVsxXX0iCiAgICAgICAgICAgICMgZW50aXR5LnNwbGl0
            KCcuJywxKVsxXSB0byByZW5hbWUgZW50aXR5KGNsaWVudC5jc2ktcmJkLW5vZGUpIGFzIGNzaS1y
            YmQtbm9kZQoKICAgICAgICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1v
            bl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nl
            c3NmdWwgYXR0ZW1wdCwKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9
            PSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAg
            ICAgICAgICAgZiInYXV0aCBnZXQtb3ItY3JlYXRlIHt1c2VyfScgY29tbWFuZCBmYWlsZWQuXG4i
            CiAgICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBz
            ZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgKQogICAgICAgIHJldHVybiBzdHIo
            anNvbl9vdXRbMF1bImtleSJdKSwgZiJ7ZW50aXR5LnNwbGl0KCcuJywgMSlbMV19IgogICAgICAg
            ICMgZW50aXR5LnNwbGl0KCcuJywxKVsxXSB0byByZW5hbWUgZW50aXR5KGNsaWVudC5jc2ktcmJk
            LW5vZGUpIGFzIGNzaS1yYmQtbm9kZQoKICAgIGRlZiBnZXRfY2VwaGZzX2RhdGFfcG9vbF9kZXRh
            aWxzKHNlbGYpOgogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAiZnMgbHMiLCAiZm9ybWF0
            IjogImpzb24ifQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAg
            ICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigiY2VwaCAiICsgY21kX2pzb25bInByZWZpeCJdKQogICAg
            ICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dl
            bihjbWRfanNvbikKICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0
            LCByZXBvcnQgYW4gZXJyb3IKICAgICAgICBpZiByZXRfdmFsICE9IDA6CiAgICAgICAgICAgICMg
            aWYgZnMgYW5kIGRhdGFfcG9vbCBhcmd1bWVudHMgYXJlIG5vdCBzZXQsIHNpbGVudGx5IHJldHVy
            bgogICAgICAgICAgICBpZiAoCiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhm
            c19maWxlc3lzdGVtX25hbWUgPT0gIiIKICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFy
            c2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZSA9PSAiIgogICAgICAgICAgICApOgogICAgICAgICAg
            ICAgICAgcmV0dXJuCiAgICAgICAgICAgICMgaWYgdXNlciBoYXMgcHJvdmlkZWQgYW55IG9mIHRo
            ZQogICAgICAgICAgICAjICctLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUnIG9yICctLWNlcGhmcy1k
            YXRhLXBvb2wtbmFtZScgYXJndW1lbnRzLAogICAgICAgICAgICAjIHJhaXNlIGFuIGV4Y2VwdGlv
            biBhcyB3ZSBhcmUgdW5hYmxlIHRvIHZlcmlmeSB0aGUgYXJncwogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiInZnMgbHMnIGNlcGgg
            Y2FsbCBmYWlsZWQgd2l0aCBlcnJvcjoge2Vycl9tc2d9IgogICAgICAgICAgICApCgogICAgICAg
            IG1hdGNoaW5nX2pzb25fb3V0ID0ge30KICAgICAgICAjIGlmICctLWNlcGhmcy1maWxlc3lzdGVt
            LW5hbWUnIGFyZ3VtZW50IGlzIHByb3ZpZGVkLAogICAgICAgICMgY2hlY2sgd2hldGhlciB0aGUg
            cHJvdmlkZWQgZmlsZXN5c3RlbS1uYW1lIGV4aXN0cyBvciBub3QKICAgICAgICBpZiBzZWxmLl9h
            cmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWU6CiAgICAgICAgICAgICMgZ2V0IHRoZSBt
            YXRjaGluZyBsaXN0CiAgICAgICAgICAgIG1hdGNoaW5nX2pzb25fb3V0X2xpc3QgPSBbCiAgICAg
            ICAgICAgICAgICBtYXRjaGVkCiAgICAgICAgICAgICAgICBmb3IgbWF0Y2hlZCBpbiBqc29uX291
            dAogICAgICAgICAgICAgICAgaWYgc3RyKG1hdGNoZWRbIm5hbWUiXSkgPT0gc2VsZi5fYXJnX3Bh
            cnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lCiAgICAgICAgICAgIF0KICAgICAgICAgICAgIyB1
            bmFibGUgdG8gZmluZCBhIG1hdGNoaW5nIGZzLW5hbWUsIHJhaXNlIGFuIGVycm9yCiAgICAgICAg
            ICAgIGlmIGxlbihtYXRjaGluZ19qc29uX291dF9saXN0KSA9PSAwOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBmIkZp
            bGVzeXN0ZW0gcHJvdmlkZWQsICd7c2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9u
            YW1lfScsICIKICAgICAgICAgICAgICAgICAgICBmImlzIG5vdCBmb3VuZCBpbiB0aGUgZnMtbGlz
            dDoge1tzdHIoeFsnbmFtZSddKSBmb3IgeCBpbiBqc29uX291dF19IgogICAgICAgICAgICAgICAg
            KQogICAgICAgICAgICBtYXRjaGluZ19qc29uX291dCA9IG1hdGNoaW5nX2pzb25fb3V0X2xpc3Rb
            MF0KICAgICAgICAjIGlmIGNlcGhmcyBmaWxlc3lzdGVtIG5hbWUgaXMgbm90IHByb3ZpZGVkLAog
            ICAgICAgICMgdHJ5IHRvIGdldCBhIGRlZmF1bHQgZnMgbmFtZSBieSBkb2luZyB0aGUgZm9sbG93
            aW5nCiAgICAgICAgZWxzZToKICAgICAgICAgICAgIyBhLiBjaGVjayBpZiB0aGVyZSBpcyBvbmx5
            IG9uZSBmaWxlc3lzdGVtIGlzIHByZXNlbnQKICAgICAgICAgICAgaWYgbGVuKGpzb25fb3V0KSA9
            PSAxOgogICAgICAgICAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXQgPSBqc29uX291dFswXQogICAg
            ICAgICAgICAjIGIuIG9yIGVsc2UsIGNoZWNrIGlmIGRhdGFfcG9vbCBuYW1lIGlzIHByb3ZpZGVk
            CiAgICAgICAgICAgIGVsaWYgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWU6
            CiAgICAgICAgICAgICAgICAjIGFuZCBpZiBwcmVzZW50LCBjaGVjayB3aGV0aGVyIHRoZXJlIGV4
            aXN0cyBhIGZzIHdoaWNoIGhhcyB0aGUgZGF0YV9wb29sCiAgICAgICAgICAgICAgICBmb3IgZWFj
            aEogaW4ganNvbl9vdXQ6CiAgICAgICAgICAgICAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5j
            ZXBoZnNfZGF0YV9wb29sX25hbWUgaW4gZWFjaEpbImRhdGFfcG9vbHMiXToKICAgICAgICAgICAg
            ICAgICAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXQgPSBlYWNoSgogICAgICAgICAgICAgICAgICAg
            ICAgICBicmVhawogICAgICAgICAgICAgICAgIyBpZiB0aGVyZSBpcyBubyBtYXRjaGluZyBmcyBl
            eGlzdHMsIHRoYXQgbWVhbnMgcHJvdmlkZWQgZGF0YV9wb29sIG5hbWUgaXMgaW52YWxpZAogICAg
            ICAgICAgICAgICAgaWYgbm90IG1hdGNoaW5nX2pzb25fb3V0OgogICAgICAgICAgICAgICAgICAg
            IHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgICAg
            IGYiUHJvdmlkZWQgZGF0YV9wb29sIG5hbWUsIHtzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRh
            X3Bvb2xfbmFtZX0sIgogICAgICAgICAgICAgICAgICAgICAgICAiIGRvZXMgbm90IGV4aXN0cyIK
            ICAgICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICMgYy4gaWYgbm90aGluZyBpcyBzZXQg
            YW5kIGNvdWxkbid0IGZpbmQgYSBkZWZhdWx0LAogICAgICAgICAgICBlbHNlOgogICAgICAgICAg
            ICAgICAgIyBqdXN0IHJldHVybiBzaWxlbnRseQogICAgICAgICAgICAgICAgcmV0dXJuCgogICAg
            ICAgIGlmIG1hdGNoaW5nX2pzb25fb3V0OgogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNl
            cGhmc19maWxlc3lzdGVtX25hbWUgPSBzdHIobWF0Y2hpbmdfanNvbl9vdXRbIm5hbWUiXSkKICAg
            ICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfbWV0YWRhdGFfcG9vbF9uYW1lID0gc3Ry
            KAogICAgICAgICAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXRbIm1ldGFkYXRhX3Bvb2wiXQogICAg
            ICAgICAgICApCgogICAgICAgIGlmIGlzaW5zdGFuY2UobWF0Y2hpbmdfanNvbl9vdXRbImRhdGFf
            cG9vbHMiXSwgbGlzdCk6CiAgICAgICAgICAgICMgaWYgdGhlIHVzZXIgaGFzIGFscmVhZHkgcHJv
            dmlkZWQgZGF0YS1wb29sLW5hbWUsCiAgICAgICAgICAgICMgdGhyb3VnaCAtLWNlcGhmcy1kYXRh
            LXBvb2wtbmFtZQogICAgICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bv
            b2xfbmFtZToKICAgICAgICAgICAgICAgICMgaWYgdGhlIHByb3ZpZGVkIG5hbWUgaXMgbm90IG1h
            dGNoaW5nIHdpdGggdGhlIG9uZSBpbiB0aGUgbGlzdAogICAgICAgICAgICAgICAgaWYgKAogICAg
            ICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lCiAg
            ICAgICAgICAgICAgICAgICAgbm90IGluIG1hdGNoaW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0K
            ICAgICAgICAgICAgICAgICk6CiAgICAgICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAgICAgZiJQcm92aWRlZCBkYXRhLXBv
            b2wtbmFtZTogJ3tzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZX0nLCAiCiAg
            ICAgICAgICAgICAgICAgICAgICAgICJkb2Vzbid0IG1hdGNoIGZyb20gdGhlIGRhdGEtcG9vbHMg
            bGlzdDogIgogICAgICAgICAgICAgICAgICAgICAgICBmIntbc3RyKHgpIGZvciB4IGluIG1hdGNo
            aW5nX2pzb25fb3V0WydkYXRhX3Bvb2xzJ11dfSIKICAgICAgICAgICAgICAgICAgICApCiAgICAg
            ICAgICAgICMgaWYgZGF0YV9wb29sIG5hbWUgaXMgbm90IHByb3ZpZGVkLAogICAgICAgICAgICAj
            IHRoZW4gdHJ5IHRvIGZpbmQgYSBkZWZhdWx0IGRhdGEgcG9vbCBuYW1lCiAgICAgICAgICAgIGVs
            c2U6CiAgICAgICAgICAgICAgICAjIGlmIG5vIGRhdGFfcG9vbHMgZXhpc3QsIHNpbGVudGx5IHJl
            dHVybgogICAgICAgICAgICAgICAgaWYgbGVuKG1hdGNoaW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xz
            Il0pID09IDA6CiAgICAgICAgICAgICAgICAgICAgcmV0dXJuCiAgICAgICAgICAgICAgICBzZWxm
            Ll9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZSA9IHN0cigKICAgICAgICAgICAgICAg
            ICAgICBtYXRjaGluZ19qc29uX291dFsiZGF0YV9wb29scyJdWzBdCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgICMgaWYgdGhlcmUgYXJlIG1vcmUgdGhhbiBvbmUgJ2RhdGFfcG9vbHMnIGV4
            aXN0LAogICAgICAgICAgICAjIHRoZW4gd2FybiB0aGUgdXNlciB0aGF0IHdlIGFyZSB1c2luZyB0
            aGUgc2VsZWN0ZWQgbmFtZQogICAgICAgICAgICBpZiBsZW4obWF0Y2hpbmdfanNvbl9vdXRbImRh
            dGFfcG9vbHMiXSkgPiAxOgogICAgICAgICAgICAgICAgcHJpbnQoCiAgICAgICAgICAgICAgICAg
            ICAgIldBUk5JTkc6IE11bHRpcGxlIGRhdGEgcG9vbHMgZGV0ZWN0ZWQ6ICIKICAgICAgICAgICAg
            ICAgICAgICBmIntbc3RyKHgpIGZvciB4IGluIG1hdGNoaW5nX2pzb25fb3V0WydkYXRhX3Bvb2xz
            J11dfVxuIgogICAgICAgICAgICAgICAgICAgIGYiVXNpbmcgdGhlIGRhdGEtcG9vbDogJ3tzZWxm
            Ll9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZX0nXG4iCiAgICAgICAgICAgICAgICAp
            CgogICAgZGVmIGNyZWF0ZV9jaGVja2VyS2V5KHNlbGYsIHVzZXIpOgogICAgICAgIGNhcHMsIGVu
            dGl0eSA9IHNlbGYuZ2V0X2NhcHNfYW5kX2VudGl0eSh1c2VyKQogICAgICAgIGNtZF9qc29uID0g
            ewogICAgICAgICAgICAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0ZSIsCiAgICAgICAgICAg
            ICJlbnRpdHkiOiBlbnRpdHksCiAgICAgICAgICAgICJjYXBzIjogW2NhcCBmb3IgY2FwX2xpc3Qg
            aW4gbGlzdChjYXBzLml0ZW1zKCkpIGZvciBjYXAgaW4gY2FwX2xpc3RdLAogICAgICAgICAgICAi
            Zm9ybWF0IjogImpzb24iLAogICAgICAgIH0KCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5k
            cnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKAogICAgICAgICAgICAgICAg
            ImNlcGggIgogICAgICAgICAgICAgICAgKyBjbWRfanNvblsicHJlZml4Il0KICAgICAgICAgICAg
            ICAgICsgIiAiCiAgICAgICAgICAgICAgICArIGNtZF9qc29uWyJlbnRpdHkiXQogICAgICAgICAg
            ICAgICAgKyAiICIKICAgICAgICAgICAgICAgICsgIiAiLmpvaW4oY21kX2pzb25bImNhcHMiXSkK
            ICAgICAgICAgICAgKQogICAgICAgICMgY2hlY2sgaWYgdXNlciBhbHJlYWR5IGV4aXN0CiAgICAg
            ICAgdXNlcl9rZXkgPSBzZWxmLmNoZWNrX3VzZXJfZXhpc3QoZW50aXR5KQogICAgICAgIGlmIHVz
            ZXJfa2V5ICE9ICIiOgogICAgICAgICAgICByZXR1cm4gdXNlcl9rZXkKCiAgICAgICAgcmV0X3Zh
            bCwganNvbl9vdXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29u
            KQogICAgICAgICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAg
            aWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmFpc2Ug
            RXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgIGYiJ2F1dGggZ2V0LW9y
            LWNyZWF0ZSB7c2VsZi5ydW5fYXNfdXNlcn0nIGNvbW1hbmQgZmFpbGVkXG4iCiAgICAgICAgICAg
            ICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09V
            VFBVVF9MSVNUfSIKICAgICAgICAgICAgKQogICAgICAgIHJldHVybiBzdHIoanNvbl9vdXRbMF1b
            ImtleSJdKQoKICAgIGRlZiBnZXRfY2VwaF9kYXNoYm9hcmRfbGluayhzZWxmKToKICAgICAgICBj
            bWRfanNvbiA9IHsicHJlZml4IjogIm1nciBzZXJ2aWNlcyIsICJmb3JtYXQiOiAianNvbiJ9CiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2Vs
            Zi5kcnlfcnVuKCJjZXBoICIgKyBjbWRfanNvblsicHJlZml4Il0pCiAgICAgICAgcmV0X3ZhbCwg
            anNvbl9vdXQsIF8gPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAg
            ICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3Zh
            bCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmV0dXJuIE5vbmUKICAg
            ICAgICBpZiAiZGFzaGJvYXJkIiBub3QgaW4ganNvbl9vdXQ6CiAgICAgICAgICAgIHJldHVybiBO
            b25lCiAgICAgICAgcmV0dXJuIGpzb25fb3V0WyJkYXNoYm9hcmQiXQoKICAgIGRlZiBjcmVhdGVf
            cmd3X2FkbWluX29wc191c2VyKHNlbGYpOgogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgInJh
            ZG9zZ3ctYWRtaW4iLAogICAgICAgICAgICAidXNlciIsCiAgICAgICAgICAgICJjcmVhdGUiLAog
            ICAgICAgICAgICAiLS11aWQiLAogICAgICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9P
            UFNfVVNFUl9OQU1FLAogICAgICAgICAgICAiLS1kaXNwbGF5LW5hbWUiLAogICAgICAgICAgICAi
            Um9vayBSR1cgQWRtaW4gT3BzIHVzZXIiLAogICAgICAgICAgICAiLS1jYXBzIiwKICAgICAgICAg
            ICAgImJ1Y2tldHM9Kjt1c2Vycz0qO3VzYWdlPXJlYWQ7bWV0YWRhdGE9cmVhZDt6b25lPXJlYWQi
            LAogICAgICAgICAgICAiLS1yZ3ctcmVhbG0iLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd19yZWFsbV9uYW1lLAogICAgICAgICAgICAiLS1yZ3ctem9uZWdyb3VwIiwKICAgICAgICAg
            ICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZWdyb3VwX25hbWUsCiAgICAgICAgICAgICItLXJn
            dy16b25lIiwKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1lLAogICAg
            ICAgIF0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJl
            dHVybiBzZWxmLmRyeV9ydW4oImNlcGggIiArICIgIi5qb2luKGNtZCkpCiAgICAgICAgdHJ5Ogog
            ICAgICAgICAgICBvdXRwdXQgPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1z
            dWJwcm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vy
            cm9yIGFzIGV4ZWNFcnI6CiAgICAgICAgICAgICMgaWYgdGhlIHVzZXIgYWxyZWFkeSBleGlzdHMs
            IHdlIGp1c3QgcXVlcnkgaXQKICAgICAgICAgICAgaWYgZXhlY0Vyci5yZXR1cm5jb2RlID09IGVy
            cm5vLkVFWElTVDoKICAgICAgICAgICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgICAgICAgICAi
            cmFkb3Nndy1hZG1pbiIsCiAgICAgICAgICAgICAgICAgICAgInVzZXIiLAogICAgICAgICAgICAg
            ICAgICAgICJpbmZvIiwKICAgICAgICAgICAgICAgICAgICAiLS11aWQiLAogICAgICAgICAgICAg
            ICAgICAgIHNlbGYuRVhURVJOQUxfUkdXX0FETUlOX09QU19VU0VSX05BTUUsCiAgICAgICAgICAg
            ICAgICAgICAgIi0tcmd3LXJlYWxtIiwKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFy
            c2VyLnJnd19yZWFsbV9uYW1lLAogICAgICAgICAgICAgICAgICAgICItLXJndy16b25lZ3JvdXAi
            LAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1l
            LAogICAgICAgICAgICAgICAgICAgICItLXJndy16b25lIiwKICAgICAgICAgICAgICAgICAgICBz
            ZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUsCiAgICAgICAgICAgICAgICBdCiAgICAgICAg
            ICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVj
            a19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAgICAgICAgZXhj
            ZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yIGFzIGV4ZWNFcnI6CiAgICAgICAgICAg
            ICAgICAgICAgZXJyX21zZyA9ICgKICAgICAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8g
            ZXhlY3V0ZSBjb21tYW5kIHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAg
            ICAgICAgICAgICAgICAgICAgZiJDb2RlOiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtl
            eGVjRXJyLnN0ZGVycn0iCiAgICAgICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAg
            IHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykKICAgICAgICAgICAgICAgICAgICByZXR1cm4gTm9u
            ZSwgTm9uZSwgRmFsc2UsICItMSIKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVy
            cl9tc2cgPSAoCiAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5k
            IHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICBm
            IkNvZGU6IHtleGVjRXJyLnJldHVybmNvZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykK
            ICAgICAgICAgICAgICAgIHJldHVybiBOb25lLCBOb25lLCBGYWxzZSwgIi0xIgoKICAgICAgICAj
            IGlmIGl0IGlzIHB5dGhvbjIsIGRvbid0IGNoZWNrIGZvciBjZXBoIHZlcnNpb24gZm9yIGFkZGlu
            ZyBgaW5mbz1yZWFkYCBjYXAocmd3X3ZhbGlkYXRpb24pCiAgICAgICAgaWYgc3lzLnZlcnNpb25f
            aW5mby5tYWpvciA8IDM6CiAgICAgICAgICAgIGpzb25vdXRwdXQgPSBqc29uLmxvYWRzKG91dHB1
            dCkKICAgICAgICAgICAgcmV0dXJuICgKICAgICAgICAgICAgICAgIGpzb25vdXRwdXRbImtleXMi
            XVswXVsiYWNjZXNzX2tleSJdLAogICAgICAgICAgICAgICAganNvbm91dHB1dFsia2V5cyJdWzBd
            WyJzZWNyZXRfa2V5Il0sCiAgICAgICAgICAgICAgICBGYWxzZSwKICAgICAgICAgICAgICAgICIi
            LAogICAgICAgICAgICApCgogICAgICAgICMgc2VwYXJhdGVseSBhZGQgaW5mbz1yZWFkIGNhcHMg
            Zm9yIHJndy1lbmRwb2ludCBpcCB2YWxpZGF0aW9uCiAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVk
            ID0gVHJ1ZQogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAg
            ICAgICAgICAiY2FwcyIsCiAgICAgICAgICAgICJhZGQiLAogICAgICAgICAgICAiLS11aWQiLAog
            ICAgICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9PUFNfVVNFUl9OQU1FLAogICAgICAg
            ICAgICAiLS1jYXBzIiwKICAgICAgICAgICAgImluZm89cmVhZCIsCiAgICAgICAgICAgICItLXJn
            dy1yZWFsbSIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUsCiAg
            ICAgICAgICAgICItLXJndy16b25lZ3JvdXAiLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd196b25lZ3JvdXBfbmFtZSwKICAgICAgICAgICAgIi0tcmd3LXpvbmUiLAogICAgICAgICAg
            ICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUsCiAgICAgICAgXQogICAgICAgIHRyeToK
            ICAgICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9
            c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2VwdCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NF
            cnJvciBhcyBleGVjRXJyOgogICAgICAgICAgICAjIGlmIHRoZSBjZXBoIHZlcnNpb24gbm90IHN1
            cHBvcnRlZCBmb3IgYWRkaW5nIGBpbmZvPXJlYWRgIGNhcChyZ3dfdmFsaWRhdGlvbikKICAgICAg
            ICAgICAgaWYgKAogICAgICAgICAgICAgICAgImNvdWxkIG5vdCBhZGQgY2FwczogdW5hYmxlIHRv
            IGFkZCBjYXBzOiBpbmZvPXJlYWRcbiIKICAgICAgICAgICAgICAgIGluIGV4ZWNFcnIuc3RkZXJy
            LmRlY29kZSgidXRmLTgiKQogICAgICAgICAgICAgICAgYW5kIGV4ZWNFcnIucmV0dXJuY29kZSA9
            PSAyNDQKICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgIGluZm9fY2FwX3N1cHBvcnRlZCA9
            IEZhbHNlCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICBlcnJfbXNnID0gKAogICAg
            ICAgICAgICAgICAgICAgIGYiZmFpbGVkIHRvIGV4ZWN1dGUgY29tbWFuZCB7Y21kfS4gT3V0cHV0
            OiB7ZXhlY0Vyci5vdXRwdXR9LiAiCiAgICAgICAgICAgICAgICAgICAgZiJDb2RlOiB7ZXhlY0Vy
            ci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtleGVjRXJyLnN0ZGVycn0iCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKGVycl9tc2cpCiAgICAgICAgICAgICAg
            ICByZXR1cm4gTm9uZSwgTm9uZSwgRmFsc2UsICItMSIKCiAgICAgICAganNvbm91dHB1dCA9IGpz
            b24ubG9hZHMob3V0cHV0KQogICAgICAgIHJldHVybiAoCiAgICAgICAgICAgIGpzb25vdXRwdXRb
            ImtleXMiXVswXVsiYWNjZXNzX2tleSJdLAogICAgICAgICAgICBqc29ub3V0cHV0WyJrZXlzIl1b
            MF1bInNlY3JldF9rZXkiXSwKICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkLAogICAgICAg
            ICAgICAiIiwKICAgICAgICApCgogICAgZGVmIHZhbGlkYXRlX3JiZF9wb29sKHNlbGYpOgogICAg
            ICAgIGlmIG5vdCBzZWxmLmNsdXN0ZXIucG9vbF9leGlzdHMoc2VsZi5fYXJnX3BhcnNlci5yYmRf
            ZGF0YV9wb29sX25hbWUpOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0
            aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcG9vbCwgJ3tzZWxmLl9hcmdfcGFy
            c2VyLnJiZF9kYXRhX3Bvb2xfbmFtZX0nLCBkb2VzIG5vdCBleGlzdCIKICAgICAgICAgICAgKQoK
            ICAgIGRlZiBpbml0X3JiZF9wb29sKHNlbGYpOgogICAgICAgIGlmIGlzaW5zdGFuY2Uoc2VsZi5j
            bHVzdGVyLCBEdW1teVJhZG9zKToKICAgICAgICAgICAgcmV0dXJuCiAgICAgICAgcmJkX3Bvb2xf
            bmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgaW9jdHgg
            PSBzZWxmLmNsdXN0ZXIub3Blbl9pb2N0eChyYmRfcG9vbF9uYW1lKQogICAgICAgIHJiZF9pbnN0
            ID0gcmJkLlJCRCgpCiAgICAgICAgcmJkX2luc3QucG9vbF9pbml0KGlvY3R4LCBUcnVlKQoKICAg
            IGRlZiB2YWxpZGF0ZV9yYWRvc19uYW1lc3BhY2Uoc2VsZik6CiAgICAgICAgcmJkX3Bvb2xfbmFt
            ZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgcmFkb3NfbmFt
            ZXNwYWNlID0gc2VsZi5fYXJnX3BhcnNlci5yYWRvc19uYW1lc3BhY2UKICAgICAgICBpZiByYWRv
            c19uYW1lc3BhY2UgPT0gIiI6CiAgICAgICAgICAgIHJldHVybgogICAgICAgIGlmIHJhZG9zX25h
            bWVzcGFjZS5pc2xvd2VyKCkgPT0gRmFsc2U6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCByYWRvcyBOYW1l
            c3BhY2UsICd7cmFkb3NfbmFtZXNwYWNlfScsICIKICAgICAgICAgICAgICAgIGYiY29udGFpbnMg
            dXBwZXIgY2FzZSIKICAgICAgICAgICAgKQogICAgICAgIHJiZF9pbnN0ID0gcmJkLlJCRCgpCiAg
            ICAgICAgaW9jdHggPSBzZWxmLmNsdXN0ZXIub3Blbl9pb2N0eChyYmRfcG9vbF9uYW1lKQogICAg
            ICAgIGlmIHJiZF9pbnN0Lm5hbWVzcGFjZV9leGlzdHMoaW9jdHgsIHJhZG9zX25hbWVzcGFjZSkg
            aXMgRmFsc2U6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAg
            ICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCByYWRvcyBOYW1lc3BhY2UsICd7cmFkb3NfbmFt
            ZXNwYWNlfScsICIKICAgICAgICAgICAgICAgIGYiaXMgbm90IGZvdW5kIGluIHRoZSBwb29sICd7
            cmJkX3Bvb2xfbmFtZX0nIgogICAgICAgICAgICApCgogICAgZGVmIGdldF9vcl9jcmVhdGVfc3Vi
            dm9sdW1lX2dyb3VwKHNlbGYsIHN1YnZvbHVtZV9ncm91cCwgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFt
            ZSk6CiAgICAgICAgY21kID0gWwogICAgICAgICAgICAiY2VwaCIsCiAgICAgICAgICAgICJmcyIs
            CiAgICAgICAgICAgICJzdWJ2b2x1bWVncm91cCIsCiAgICAgICAgICAgICJnZXRwYXRoIiwKICAg
            ICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAgICAgICAgICAgc3Vidm9sdW1lX2dy
            b3VwLAogICAgICAgIF0KICAgICAgICB0cnk6CiAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNo
            ZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1
            YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICBjbWQgPSBbCiAgICAgICAg
            ICAgICAgICAiY2VwaCIsCiAgICAgICAgICAgICAgICAiZnMiLAogICAgICAgICAgICAgICAgInN1
            YnZvbHVtZWdyb3VwIiwKICAgICAgICAgICAgICAgICJjcmVhdGUiLAogICAgICAgICAgICAgICAg
            Y2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAgICAgICAgICAgICAgIHN1YnZvbHVtZV9ncm91cCwK
            ICAgICAgICAgICAgXQogICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBfID0gc3VicHJv
            Y2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAg
            ICBleGNlcHQgc3VicHJvY2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3I6CiAgICAgICAgICAgICAgICBy
            YWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYic3Vi
            dm9sdW1lIGdyb3VwIHtzdWJ2b2x1bWVfZ3JvdXB9IGlzIG5vdCBhYmxlIHRvIGdldCBjcmVhdGVk
            IgogICAgICAgICAgICAgICAgKQoKICAgIGRlZiBwaW5fc3Vidm9sdW1lKAogICAgICAgIHNlbGYs
            IHN1YnZvbHVtZV9ncm91cCwgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwgcGluX3R5cGUsIHBpbl9z
            ZXR0aW5nCiAgICApOgogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgImNlcGgiLAogICAgICAg
            ICAgICAiZnMiLAogICAgICAgICAgICAic3Vidm9sdW1lZ3JvdXAiLAogICAgICAgICAgICAicGlu
            IiwKICAgICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAgICAgICAgICAgc3Vidm9s
            dW1lX2dyb3VwLAogICAgICAgICAgICBwaW5fdHlwZSwKICAgICAgICAgICAgcGluX3NldHRpbmcs
            CiAgICAgICAgXQogICAgICAgIHRyeToKICAgICAgICAgICAgXyA9IHN1YnByb2Nlc3MuY2hlY2tf
            b3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAgICAgICBleGNlcHQgc3VicHJv
            Y2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3I6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1
            cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmInN1YnZvbHVtZSBncm91cCB7c3Vidm9sdW1l
            X2dyb3VwfSBpcyBub3QgYWJsZSB0byBnZXQgcGlubmVkIgogICAgICAgICAgICApCgogICAgZGVm
            IGdldF9yZ3dfZnNpZChzZWxmLCBiYXNlX3VybCwgdmVyaWZ5KToKICAgICAgICBhY2Nlc3Nfa2V5
            ID0gc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJfQUNDRVNTX0tFWSJdCiAgICAgICAg
            c2VjcmV0X2tleSA9IHNlbGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX1NFQ1JFVF9LRVki
            XQogICAgICAgIHJnd19lbmRwb2ludCA9IHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50CiAg
            ICAgICAgYmFzZV91cmwgPSBiYXNlX3VybCArICI6Ly8iICsgcmd3X2VuZHBvaW50ICsgIi9hZG1p
            bi9pbmZvPyIKICAgICAgICBwYXJhbXMgPSB7ImZvcm1hdCI6ICJqc29uIn0KICAgICAgICByZXF1
            ZXN0X3VybCA9IGJhc2VfdXJsICsgdXJsZW5jb2RlKHBhcmFtcykKCiAgICAgICAgdHJ5OgogICAg
            ICAgICAgICByID0gcmVxdWVzdHMuZ2V0KAogICAgICAgICAgICAgICAgcmVxdWVzdF91cmwsCiAg
            ICAgICAgICAgICAgICBhdXRoPVMzQXV0aChhY2Nlc3Nfa2V5LCBzZWNyZXRfa2V5LCByZ3dfZW5k
            cG9pbnQpLAogICAgICAgICAgICAgICAgdmVyaWZ5PXZlcmlmeSwKICAgICAgICAgICAgKQogICAg
            ICAgIGV4Y2VwdCByZXF1ZXN0cy5leGNlcHRpb25zLlRpbWVvdXQ6CiAgICAgICAgICAgIHN5cy5z
            dGRlcnIud3JpdGUoCiAgICAgICAgICAgICAgICBmImludmFsaWQgZW5kcG9pbnQ6LCBub3QgYWJs
            ZSB0byBjYWxsIGFkbWluLW9wcyBhcGl7cmd3X2VuZHBvaW50fSIKICAgICAgICAgICAgKQogICAg
            ICAgICAgICByZXR1cm4gIiIsICItMSIKICAgICAgICByMSA9IHIuanNvbigpCiAgICAgICAgaWYg
            cjEgaXMgTm9uZSBvciByMS5nZXQoImluZm8iKSBpcyBOb25lOgogICAgICAgICAgICBzeXMuc3Rk
            ZXJyLndyaXRlKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmd3IEVuZHBvaW50LCAn
            e3NlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50fScsIGlzIGludmFsaWQuIgogICAgICAgICAg
            ICApCiAgICAgICAgICAgIHJldHVybiAoCiAgICAgICAgICAgICAgICAiIiwKICAgICAgICAgICAg
            ICAgICItMSIsCiAgICAgICAgICAgICkKCiAgICAgICAgcmV0dXJuIHIxWyJpbmZvIl1bInN0b3Jh
            Z2VfYmFja2VuZHMiXVswXVsiY2x1c3Rlcl9pZCJdLCAiIgoKICAgIGRlZiB2YWxpZGF0ZV9yZ3df
            ZW5kcG9pbnQoc2VsZiwgaW5mb19jYXBfc3VwcG9ydGVkKToKICAgICAgICAjIGlmIHRoZSAnY2x1
            c3RlcicgaW5zdGFuY2UgaXMgYSBkdW1teSBvbmUsCiAgICAgICAgIyBkb24ndCB0cnkgdG8gcmVh
            Y2ggb3V0IHRvIHRoZSBlbmRwb2ludAogICAgICAgIGlmIGlzaW5zdGFuY2Uoc2VsZi5jbHVzdGVy
            LCBEdW1teVJhZG9zKToKICAgICAgICAgICAgcmV0dXJuCgogICAgICAgIHJnd19lbmRwb2ludCA9
            IHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50CgogICAgICAgICMgdmFsaWRhdGUgcmd3IGVu
            ZHBvaW50IG9ubHkgaWYgaXAgYWRkcmVzcyBpcyBwYXNzZWQKICAgICAgICBpcF90eXBlID0gc2Vs
            Zi5faW52YWxpZF9lbmRwb2ludChyZ3dfZW5kcG9pbnQpCgogICAgICAgICMgY2hlY2sgaWYgdGhl
            IHJndyBlbmRwb2ludCBpcyByZWFjaGFibGUKICAgICAgICBjZXJ0ID0gTm9uZQogICAgICAgIGlm
            IG5vdCBzZWxmLl9hcmdfcGFyc2VyLnJnd19za2lwX3RscyBhbmQgc2VsZi52YWxpZGF0ZV9yZ3df
            ZW5kcG9pbnRfdGxzX2NlcnQoKToKICAgICAgICAgICAgY2VydCA9IHNlbGYuX2FyZ19wYXJzZXIu
            cmd3X3Rsc19jZXJ0X3BhdGgKICAgICAgICBiYXNlX3VybCwgdmVyaWZ5LCBlcnIgPSBzZWxmLmVu
            ZHBvaW50X2RpYWwocmd3X2VuZHBvaW50LCBpcF90eXBlLCBjZXJ0PWNlcnQpCiAgICAgICAgaWYg
            ZXJyICE9ICIiOgogICAgICAgICAgICByZXR1cm4gIi0xIgoKICAgICAgICAjIGNoZWNrIGlmIHRo
            ZSByZ3cgZW5kcG9pbnQgYmVsb25ncyB0byB0aGUgc2FtZSBjbHVzdGVyCiAgICAgICAgIyBvbmx5
            IGNoZWNrIGlmIGBpbmZvYCBjYXAgaXMgc3VwcG9ydGVkCiAgICAgICAgaWYgaW5mb19jYXBfc3Vw
            cG9ydGVkOgogICAgICAgICAgICBmc2lkID0gc2VsZi5nZXRfZnNpZCgpCiAgICAgICAgICAgIHJn
            d19mc2lkLCBlcnIgPSBzZWxmLmdldF9yZ3dfZnNpZChiYXNlX3VybCwgdmVyaWZ5KQogICAgICAg
            ICAgICBpZiBlcnIgPT0gIi0xIjoKICAgICAgICAgICAgICAgIHJldHVybiAiLTEiCiAgICAgICAg
            ICAgIGlmIGZzaWQgIT0gcmd3X2ZzaWQ6CiAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRl
            KAogICAgICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHJndyBFbmRwb2ludCwgJ3tzZWxm
            Ll9hcmdfcGFyc2VyLnJnd19lbmRwb2ludH0nLCBpcyBpbnZhbGlkLiBXZSBhcmUgdmFsaWRhdGlu
            ZyBieSBjYWxsaW5nIHRoZSBhZG1pbm9wcyBhcGkgdGhyb3VnaCByZ3ctZW5kcG9pbnQgYW5kIHZh
            bGlkYXRpbmcgdGhlIGNsdXN0ZXJfaWQgJ3tyZ3dfZnNpZH0nIGlzIGVxdWFsIHRvIHRoZSBjZXBo
            IGNsdXN0ZXIgZnNpZCAne2ZzaWR9JyIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAg
            IHJldHVybiAiLTEiCgogICAgICAgICMgY2hlY2sgaWYgdGhlIHJndyBlbmRwb2ludCBwb29sIGV4
            aXN0CiAgICAgICAgIyBvbmx5IHZhbGlkYXRlIGlmIHJnd19wb29sX3ByZWZpeCBpcyBwYXNzZWQg
            ZWxzZSBpdCB3aWxsIHRha2UgZGVmYXVsdCB2YWx1ZSBhbmQgd2UgZG9uJ3QgY3JlYXRlIHRoZXNl
            IGRlZmF1bHQgcG9vbHMKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZp
            eCAhPSAiZGVmYXVsdCI6CiAgICAgICAgICAgIHJnd19wb29sc190b192YWxpZGF0ZSA9IFsKICAg
            ICAgICAgICAgICAgIGYie3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cubWV0
            YSIsCiAgICAgICAgICAgICAgICAiLnJndy5yb290IiwKICAgICAgICAgICAgICAgIGYie3NlbGYu
            X2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cuY29udHJvbCIsCiAgICAgICAgICAgICAg
            ICBmIntzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmxvZyIsCiAgICAgICAg
            ICAgIF0KICAgICAgICAgICAgZm9yIF9yZ3dfcG9vbF90b192YWxpZGF0ZSBpbiByZ3dfcG9vbHNf
            dG9fdmFsaWRhdGU6CiAgICAgICAgICAgICAgICBpZiBub3Qgc2VsZi5jbHVzdGVyLnBvb2xfZXhp
            c3RzKF9yZ3dfcG9vbF90b192YWxpZGF0ZSk6CiAgICAgICAgICAgICAgICAgICAgc3lzLnN0ZGVy
            ci53cml0ZSgKICAgICAgICAgICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcG9vbCwgJ3tf
            cmd3X3Bvb2xfdG9fdmFsaWRhdGV9JywgZG9lcyBub3QgZXhpc3QiCiAgICAgICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICAgICAgICAgIHJldHVybiAiLTEiCgogICAgICAgIHJldHVybiAiIgoK
            ICAgIGRlZiB2YWxpZGF0ZV9yZ3dfbXVsdGlzaXRlKHNlbGYsIHJnd19tdWx0aXNpdGVfY29uZmln
            X25hbWUsIHJnd19tdWx0aXNpdGVfY29uZmlnKToKICAgICAgICBpZiByZ3dfbXVsdGlzaXRlX2Nv
            bmZpZyAhPSAiIjoKICAgICAgICAgICAgY21kID0gWwogICAgICAgICAgICAgICAgInJhZG9zZ3ct
            YWRtaW4iLAogICAgICAgICAgICAgICAgcmd3X211bHRpc2l0ZV9jb25maWcsCiAgICAgICAgICAg
            ICAgICAiZ2V0IiwKICAgICAgICAgICAgICAgICItLXJndy0iICsgcmd3X211bHRpc2l0ZV9jb25m
            aWcsCiAgICAgICAgICAgICAgICByZ3dfbXVsdGlzaXRlX2NvbmZpZ19uYW1lLAogICAgICAgICAg
            ICBdCiAgICAgICAgICAgIHRyeToKICAgICAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNoZWNr
            X291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgICAgIGV4Y2VwdCBz
            dWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJvciBhcyBleGVjRXJyOgogICAgICAgICAgICAgICAg
            ZXJyX21zZyA9ICgKICAgICAgICAgICAgICAgICAgICBmImZhaWxlZCB0byBleGVjdXRlIGNvbW1h
            bmQge2NtZH0uIE91dHB1dDoge2V4ZWNFcnIub3V0cHV0fS4gIgogICAgICAgICAgICAgICAgICAg
            IGYiQ29kZToge2V4ZWNFcnIucmV0dXJuY29kZX0uIEVycm9yOiB7ZXhlY0Vyci5zdGRlcnJ9Igog
            ICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZShlcnJfbXNn
            KQogICAgICAgICAgICAgICAgcmV0dXJuICItMSIKICAgICAgICByZXR1cm4gIiIKCiAgICBkZWYg
            X2dlbl9vdXRwdXRfbWFwKHNlbGYpOgogICAgICAgIGlmIHNlbGYub3V0X21hcDoKICAgICAgICAg
            ICAgcmV0dXJuCiAgICAgICAgIyBzdXBwb3J0IGxlZ2FjeSBmbGFnIHdpdGggdXBncmFkZXMKICAg
            ICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNsdXN0ZXJfbmFtZToKICAgICAgICAgICAgc2VsZi5f
            YXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5jbHVzdGVyX25h
            bWUKICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUgPSAoCiAgICAgICAg
            ICAgIHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZS5sb3dlcigpCiAgICAgICAgKSAg
            IyBhbHdheXMgY29udmVydCBjbHVzdGVyIG5hbWUgdG8gbG93ZXJjYXNlIGNoYXJhY3RlcnMKICAg
            ICAgICBzZWxmLnZhbGlkYXRlX3JiZF9wb29sKCkKICAgICAgICBzZWxmLmluaXRfcmJkX3Bvb2wo
            KQogICAgICAgIHNlbGYudmFsaWRhdGVfcmFkb3NfbmFtZXNwYWNlKCkKICAgICAgICBzZWxmLl9l
            eGNsdWRlZF9rZXlzLmFkZCgiSzhTX0NMVVNURVJfTkFNRSIpCiAgICAgICAgc2VsZi5nZXRfY2Vw
            aGZzX2RhdGFfcG9vbF9kZXRhaWxzKCkKICAgICAgICBzZWxmLm91dF9tYXBbIk5BTUVTUEFDRSJd
            ID0gc2VsZi5fYXJnX3BhcnNlci5uYW1lc3BhY2UKICAgICAgICBzZWxmLm91dF9tYXBbIks4U19D
            TFVTVEVSX05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQogICAgICAg
            IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9GU0lEIl0gPSBzZWxmLmdldF9mc2lkKCkKICAg
            ICAgICBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUk5BTUUiXSA9IHNlbGYucnVuX2Fz
            X3VzZXIKICAgICAgICBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfQ0VQSF9NT05fREFUQSJd
            ID0gc2VsZi5nZXRfY2VwaF9leHRlcm5hbF9tb25fZGF0YSgpCiAgICAgICAgc2VsZi5vdXRfbWFw
            WyJST09LX0VYVEVSTkFMX1VTRVJfU0VDUkVUIl0gPSBzZWxmLmNyZWF0ZV9jaGVja2VyS2V5KAog
            ICAgICAgICAgICAiY2xpZW50LmhlYWx0aGNoZWNrZXIiCiAgICAgICAgKQogICAgICAgIHNlbGYu
            b3V0X21hcFsiUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSyJdID0gc2VsZi5nZXRfY2VwaF9k
            YXNoYm9hcmRfbGluaygpCiAgICAgICAgKAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9S
            QkRfTk9ERV9TRUNSRVQiXSwKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVf
            U0VDUkVUX05BTUUiXSwKICAgICAgICApID0gc2VsZi5jcmVhdGVfY2VwaENTSUtleXJpbmdfdXNl
            cigiY2xpZW50LmNzaS1yYmQtbm9kZSIpCiAgICAgICAgKAogICAgICAgICAgICBzZWxmLm91dF9t
            YXBbIkNTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUIl0sCiAgICAgICAgICAgIHNlbGYub3V0X21h
            cFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICkgPSBzZWxmLmNy
            ZWF0ZV9jZXBoQ1NJS2V5cmluZ191c2VyKCJjbGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIpCiAg
            ICAgICAgc2VsZi5vdXRfbWFwWyJDRVBIRlNfUE9PTF9OQU1FIl0gPSBzZWxmLl9hcmdfcGFyc2Vy
            LmNlcGhmc19kYXRhX3Bvb2xfbmFtZQogICAgICAgIHNlbGYub3V0X21hcFsiQ0VQSEZTX01FVEFE
            QVRBX1BPT0xfTkFNRSJdID0gKAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19t
            ZXRhZGF0YV9wb29sX25hbWUKICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJDRVBIRlNf
            RlNfTkFNRSJdID0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lCiAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJSRVNUUklDVEVEX0FVVEhfUEVSTUlTU0lPTiJdID0gKAogICAgICAg
            ICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uCiAgICAgICAg
            KQogICAgICAgIHNlbGYub3V0X21hcFsiUkFET1NfTkFNRVNQQUNFIl0gPSBzZWxmLl9hcmdfcGFy
            c2VyLnJhZG9zX25hbWVzcGFjZQogICAgICAgIHNlbGYub3V0X21hcFsiU1VCVk9MVU1FX0dST1VQ
            Il0gPSBzZWxmLl9hcmdfcGFyc2VyLnN1YnZvbHVtZV9ncm91cAogICAgICAgIHNlbGYub3V0X21h
            cFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVCJdID0gIiIKICAgICAgICBzZWxmLm91dF9tYXBbIkNT
            SV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUIl0gPSAiIgogICAgICAgICMgY3JlYXRlIENlcGhG
            UyBub2RlIGFuZCBwcm92aXNpb25lciBrZXlyaW5nIG9ubHkgd2hlbiBNRFMgZXhpc3RzCiAgICAg
            ICAgaWYgc2VsZi5vdXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdIGFuZCBzZWxmLm91dF9tYXBbIkNF
            UEhGU19QT09MX05BTUUiXToKICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgc2VsZi5vdXRf
            bWFwWyJDU0lfQ0VQSEZTX05PREVfU0VDUkVUIl0sCiAgICAgICAgICAgICAgICBzZWxmLm91dF9t
            YXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICApID0gc2VsZi5j
            cmVhdGVfY2VwaENTSUtleXJpbmdfdXNlcigiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIpCiAgICAg
            ICAgICAgICgKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9WSVNJ
            T05FUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19Q
            Uk9WSVNJT05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICApID0gc2VsZi5jcmVhdGVfY2Vw
            aENTSUtleXJpbmdfdXNlcigiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiKQogICAgICAg
            ICAgICAjIGNyZWF0ZSB0aGUgZGVmYXVsdCAiY3NpIiBzdWJ2b2x1bWVncm91cAogICAgICAgICAg
            ICBzZWxmLmdldF9vcl9jcmVhdGVfc3Vidm9sdW1lX2dyb3VwKAogICAgICAgICAgICAgICAgImNz
            aSIsIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgICAgICAp
            CiAgICAgICAgICAgICMgcGluIHRoZSBkZWZhdWx0ICJjc2kiIHN1YnZvbHVtZWdyb3VwCiAgICAg
            ICAgICAgIHNlbGYucGluX3N1YnZvbHVtZSgKICAgICAgICAgICAgICAgICJjc2kiLCBzZWxmLl9h
            cmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUsICJkaXN0cmlidXRlZCIsICIxIgogICAg
            ICAgICAgICApCiAgICAgICAgICAgIGlmIHNlbGYub3V0X21hcFsiU1VCVk9MVU1FX0dST1VQIl06
            CiAgICAgICAgICAgICAgICBzZWxmLmdldF9vcl9jcmVhdGVfc3Vidm9sdW1lX2dyb3VwKAogICAg
            ICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuc3Vidm9sdW1lX2dyb3VwLAogICAgICAg
            ICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgIHNlbGYucGluX3N1YnZvbHVtZSgKICAgICAg
            ICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnN1YnZvbHVtZV9ncm91cCwKICAgICAgICAg
            ICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAg
            ICAgICAgICAgICAgICAgImRpc3RyaWJ1dGVkIiwKICAgICAgICAgICAgICAgICAgICAiMSIsCiAg
            ICAgICAgICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJSR1dfVExTX0NFUlQiXSA9ICIi
            CiAgICAgICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UIl0gPSAiIgogICAgICAg
            IHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUIl0gPSAiIgogICAgICAgIGlm
            IG5vdCBzZWxmLl9hcmdfcGFyc2VyLnNraXBfbW9uaXRvcmluZ19lbmRwb2ludDoKICAgICAgICAg
            ICAgKAogICAgICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UIl0s
            CiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJd
            LAogICAgICAgICAgICApID0gc2VsZi5nZXRfYWN0aXZlX2FuZF9zdGFuZGJ5X21ncnMoKQogICAg
            ICAgIHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdID0gc2VsZi5fYXJnX3BhcnNlci5yYmRf
            ZGF0YV9wb29sX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBbIlJCRF9NRVRBREFUQV9FQ19QT09M
            X05BTUUiXSA9ICgKICAgICAgICAgICAgc2VsZi52YWxpZGF0ZV9yYmRfbWV0YWRhdGFfZWNfcG9v
            bF9uYW1lKCkKICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJSR1dfUE9PTF9QUkVGSVgi
            XSA9IHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4CiAgICAgICAgc2VsZi5vdXRfbWFw
            WyJSR1dfRU5EUE9JTlQiXSA9ICIiCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZ3dfZW5k
            cG9pbnQ6CiAgICAgICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAg
            ICAgICAgIHNlbGYuY3JlYXRlX3Jnd19hZG1pbl9vcHNfdXNlcigpCiAgICAgICAgICAgIGVsc2U6
            CiAgICAgICAgICAgICAgICBpZiAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfcmVhbG1fbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdf
            cGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBz
            ZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICk6CiAg
            ICAgICAgICAgICAgICAgICAgZXJyID0gc2VsZi52YWxpZGF0ZV9yZ3dfbXVsdGlzaXRlKAogICAg
            ICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lLCAicmVh
            bG0iCiAgICAgICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAgIGVyciA9IHNlbGYu
            dmFsaWRhdGVfcmd3X211bHRpc2l0ZSgKICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJn
            X3BhcnNlci5yZ3dfem9uZWdyb3VwX25hbWUsICJ6b25lZ3JvdXAiCiAgICAgICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICAgICAgICAgIGVyciA9IHNlbGYudmFsaWRhdGVfcmd3X211bHRpc2l0
            ZSgKICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1l
            LCAiem9uZSIKICAgICAgICAgICAgICAgICAgICApCgogICAgICAgICAgICAgICAgaWYgKAogICAg
            ICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUgPT0gIiIKICAg
            ICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZWdyb3VwX25hbWUg
            PT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZV9u
            YW1lID09ICIiCiAgICAgICAgICAgICAgICApIG9yICgKICAgICAgICAgICAgICAgICAgICBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lICE9ICIiCiAgICAgICAgICAgICAgICAgICAgYW5k
            IHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lICE9ICIiCiAgICAgICAgICAgICAg
            ICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVfbmFtZSAhPSAiIgogICAgICAgICAg
            ICAgICAgKToKICAgICAgICAgICAgICAgICAgICAoCiAgICAgICAgICAgICAgICAgICAgICAgIHNl
            bGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX0FDQ0VTU19LRVkiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJfU0VDUkVUX0tFWSJd
            LAogICAgICAgICAgICAgICAgICAgICAgICBpbmZvX2NhcF9zdXBwb3J0ZWQsCiAgICAgICAgICAg
            ICAgICAgICAgICAgIGVyciwKICAgICAgICAgICAgICAgICAgICApID0gc2VsZi5jcmVhdGVfcmd3
            X2FkbWluX29wc191c2VyKCkKICAgICAgICAgICAgICAgICAgICBlcnIgPSBzZWxmLnZhbGlkYXRl
            X3Jnd19lbmRwb2ludChpbmZvX2NhcF9zdXBwb3J0ZWQpCiAgICAgICAgICAgICAgICAgICAgaWYg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfdGxzX2NlcnRfcGF0aDoKICAgICAgICAgICAgICAgICAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJSR1dfVExTX0NFUlQiXSA9ICgKICAgICAgICAgICAgICAgICAgICAg
            ICAgICAgIHNlbGYudmFsaWRhdGVfcmd3X2VuZHBvaW50X3Rsc19jZXJ0KCkKICAgICAgICAgICAg
            ICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAgICMgaWYgdGhlcmUgaXMgbm8gZXJyb3Is
            IHNldCB0aGUgUkdXX0VORFBPSU5UCiAgICAgICAgICAgICAgICAgICAgaWYgZXJyICE9ICItMSI6
            CiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiUkdXX0VORFBPSU5UIl0gPSBz
            ZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludAogICAgICAgICAgICAgICAgZWxzZToKICAgICAg
            ICAgICAgICAgICAgICBlcnIgPSAiUGxlYXNlIHByb3ZpZGUgYWxsIHRoZSBSR1cgbXVsdGlzaXRl
            IHBhcmFtZXRlcnMgb3Igbm9uZSBvZiB0aGVtIgogICAgICAgICAgICAgICAgICAgIHN5cy5zdGRl
            cnIud3JpdGUoZXJyKQoKICAgIGRlZiBnZW5fc2hlbGxfb3V0KHNlbGYpOgogICAgICAgIHNlbGYu
            X2dlbl9vdXRwdXRfbWFwKCkKICAgICAgICBzaE91dElPID0gU3RyaW5nSU8oKQogICAgICAgIGZv
            ciBrLCB2IGluIHNlbGYub3V0X21hcC5pdGVtcygpOgogICAgICAgICAgICBpZiB2IGFuZCBrIG5v
            dCBpbiBzZWxmLl9leGNsdWRlZF9rZXlzOgogICAgICAgICAgICAgICAgc2hPdXRJTy53cml0ZShm
            ImV4cG9ydCB7a309e3Z9e0xJTkVTRVB9IikKICAgICAgICBzaE91dCA9IHNoT3V0SU8uZ2V0dmFs
            dWUoKQogICAgICAgIHNoT3V0SU8uY2xvc2UoKQogICAgICAgIHJldHVybiBzaE91dAoKICAgIGRl
            ZiBnZW5fanNvbl9vdXQoc2VsZik6CiAgICAgICAgc2VsZi5fZ2VuX291dHB1dF9tYXAoKQogICAg
            ICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuICIiCiAg
            ICAgICAganNvbl9vdXQgPSBbCiAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICJuYW1lIjog
            InJvb2stY2VwaC1tb24tZW5kcG9pbnRzIiwKICAgICAgICAgICAgICAgICJraW5kIjogIkNvbmZp
            Z01hcCIsCiAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAiZGF0
            YSI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9DRVBIX01PTl9EQVRBIl0sCiAgICAgICAg
            ICAgICAgICAgICAgIm1heE1vbklkIjogIjAiLAogICAgICAgICAgICAgICAgICAgICJtYXBwaW5n
            IjogInt9IiwKICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgIH0sCiAgICAgICAgICAgIHsK
            ICAgICAgICAgICAgICAgICJuYW1lIjogInJvb2stY2VwaC1tb24iLAogICAgICAgICAgICAgICAg
            ImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAg
            ICAgICAgICJhZG1pbi1zZWNyZXQiOiAiYWRtaW4tc2VjcmV0IiwKICAgICAgICAgICAgICAgICAg
            ICAiZnNpZCI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9GU0lEIl0sCiAgICAgICAgICAg
            ICAgICAgICAgIm1vbi1zZWNyZXQiOiAibW9uLXNlY3JldCIsCiAgICAgICAgICAgICAgICB9LAog
            ICAgICAgICAgICB9LAogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFtZSI6ICJyb29r
            LWNlcGgtb3BlcmF0b3ItY3JlZHMiLAogICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwK
            ICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICJ1c2VySUQiOiBz
            ZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUk5BTUUiXSwKICAgICAgICAgICAgICAgICAg
            ICAidXNlcktleSI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9VU0VSX1NFQ1JFVCJdLAog
            ICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgfSwKICAgICAgICBdCgogICAgICAgICMgaWYg
            J01PTklUT1JJTkdfRU5EUE9JTlQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAnbW9uaXRvcmluZy1l
            bmRwb2ludCcgdG8gQ2x1c3RlcgogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFw
            WyJNT05JVE9SSU5HX0VORFBPSU5UIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsiTU9O
            SVRPUklOR19FTkRQT0lOVF9QT1JUIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291dC5h
            cHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAibW9u
            aXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiQ2VwaENsdXN0
            ZXIiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAg
            ICAiTW9uaXRvcmluZ0VuZHBvaW50Ijogc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5U
            Il0sCiAgICAgICAgICAgICAgICAgICAgICAgICJNb25pdG9yaW5nUG9ydCI6IHNlbGYub3V0X21h
            cFsiTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUIl0sCiAgICAgICAgICAgICAgICAgICAgfSwKICAg
            ICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQoKICAgICAgICAjIGlmICdDU0lfUkJEX05PREVf
            U0VDUkVUJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ3Jvb2stY3NpLXJiZC1wcm92aXNpb25lcicg
            U2VjcmV0CiAgICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9E
            RV9TRUNSRVQiXQogICAgICAgICAgICBhbmQgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VD
            UkVUX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAg
            ICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9vay17c2VsZi5vdXRf
            bWFwWydDU0lfUkJEX05PREVfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICJr
            aW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAg
            ICAgICAgICAgICAgICJ1c2VySUQiOiBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9TRUNSRVRf
            TkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNlbGYub3V0X21hcFsi
            Q1NJX1JCRF9OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAg
            ICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdDU0lfUkJEX1BST1ZJU0lPTkVSX1NF
            Q1JFVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdyb29rLWNzaS1yYmQtcHJvdmlzaW9uZXInIFNl
            Y3JldAogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJ
            U0lPTkVSX1NFQ1JFVCJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9SQkRfUFJP
            VklTSU9ORVJfU0VDUkVUX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFw
            cGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9v
            ay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAg
            ICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJk
            YXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJD
            U0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAg
            ICJ1c2VyS2V5Ijogc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAog
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAg
            ICAgICAjIGlmICdDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCcgZXhpc3RzLCB0aGVuIG9u
            bHkgYWRkICdyb29rLWNzaS1jZXBoZnMtcHJvdmlzaW9uZXInIFNlY3JldAogICAgICAgIGlmICgK
            ICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJd
            CiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VD
            UkVUX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAg
            ICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9vay17c2VsZi5vdXRf
            bWFwWydDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAg
            ICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewog
            ICAgICAgICAgICAgICAgICAgICAgICAiYWRtaW5JRCI6IHNlbGYub3V0X21hcFsiQ1NJX0NFUEhG
            U19QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAiYWRt
            aW5LZXkiOiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUIl0sCiAg
            ICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAg
            ICAgICMgaWYgJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAn
            cm9vay1jc2ktY2VwaGZzLW5vZGUnIFNlY3JldAogICAgICAgIGlmICgKICAgICAgICAgICAgc2Vs
            Zi5vdXRfbWFwWyJDU0lfQ0VQSEZTX05PREVfU0VDUkVUIl0KICAgICAgICAgICAgYW5kIHNlbGYu
            b3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FIl0KICAgICAgICApOgogICAgICAg
            ICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAg
            ICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9O
            QU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAg
            ICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAiYWRtaW5JRCI6IHNl
            bGYub3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FIl0sCiAgICAgICAgICAgICAg
            ICAgICAgICAgICJhZG1pbktleSI6IHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JF
            VCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAg
            ICkKICAgICAgICAjIGlmICdST09LX0VYVEVSTkFMX0RBU0hCT0FSRF9MSU5LJyBleGlzdHMsIHRo
            ZW4gb25seSBhZGQgJ3Jvb2stY2VwaC1kYXNoYm9hcmQtbGluaycgU2VjcmV0CiAgICAgICAgaWYg
            c2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0RBU0hCT0FSRF9MSU5LIl06CiAgICAgICAgICAg
            IGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAi
            bmFtZSI6ICJyb29rLWNlcGgtZGFzaGJvYXJkLWxpbmsiLAogICAgICAgICAgICAgICAgICAgICJr
            aW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAg
            ICAgICAgICAgICAgICJ1c2VySUQiOiAiY2VwaC1kYXNoYm9hcmQtbGluayIsCiAgICAgICAgICAg
            ICAgICAgICAgICAgICJ1c2VyS2V5Ijogc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0RBU0hC
            T0FSRF9MSU5LIl0sCiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAg
            ICAgICAgICAgKQogICAgICAgICMgaWYgJ1JBRE9TX05BTUVTUEFDRScgZXhpc3RzLCB0aGVuIG9u
            bHkgYWRkIHRoZSAiUkFET1NfTkFNRVNQQUNFIiBuYW1lc3BhY2UKICAgICAgICBpZiAoCiAgICAg
            ICAgICAgIHNlbGYub3V0X21hcFsiUkFET1NfTkFNRVNQQUNFIl0KICAgICAgICAgICAgYW5kIHNl
            bGYub3V0X21hcFsiUkVTVFJJQ1RFRF9BVVRIX1BFUk1JU1NJT04iXQogICAgICAgICAgICBhbmQg
            bm90IHNlbGYub3V0X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdCiAgICAgICAgKToK
            ICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAg
            ICAgICAgICAgICJuYW1lIjogInJhZG9zLW5hbWVzcGFjZSIsCiAgICAgICAgICAgICAgICAgICAg
            ImtpbmQiOiAiQ2VwaEJsb2NrUG9vbFJhZG9zTmFtZXNwYWNlIiwKICAgICAgICAgICAgICAgICAg
            ICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgInJhZG9zTmFtZXNwYWNlTmFtZSI6
            IHNlbGYub3V0X21hcFsiUkFET1NfTkFNRVNQQUNFIl0sCiAgICAgICAgICAgICAgICAgICAgICAg
            ICJwb29sIjogc2VsZi5vdXRfbWFwWyJSQkRfUE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAg
            ICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICAgICBqc29uX291
            dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAi
            Y2VwaC1yYmQtcmFkb3MtbmFtZXNwYWNlIiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJT
            dG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAg
            ICAgICAgICAgICAicG9vbCI6IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAogICAgICAg
            ICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1u
            YW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05B
            TUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRy
            b2xsZXItZXhwYW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRf
            UFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3Np
            LnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0
            X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICB9
            LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgZWxzZToKICAgICAgICAg
            ICAgaWYgc2VsZi5vdXRfbWFwWyJSQkRfTUVUQURBVEFfRUNfUE9PTF9OQU1FIl06CiAgICAgICAg
            ICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICAgICAgewogICAgICAgICAg
            ICAgICAgICAgICAgICAibmFtZSI6ICJjZXBoLXJiZCIsCiAgICAgICAgICAgICAgICAgICAgICAg
            ICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAgICAgICAgICAgICJkYXRhIjog
            ewogICAgICAgICAgICAgICAgICAgICAgICAgICAgImRhdGFQb29sIjogc2VsZi5vdXRfbWFwWyJS
            QkRfUE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYu
            b3V0X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdLAogICAgICAgICAgICAgICAgICAg
            ICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNyZXQtbmFtZSI6IGYi
            cm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwK
            ICAgICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vY29udHJvbGxl
            ci1leHBhbmQtc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9QUk9W
            SVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiY3Np
            LnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0
            X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAg
            ICAgfSwKICAgICAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICAgICApCiAgICAgICAgICAg
            IGVsc2U6CiAgICAgICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICAg
            ICAgewogICAgICAgICAgICAgICAgICAgICAgICAibmFtZSI6ICJjZXBoLXJiZCIsCiAgICAgICAg
            ICAgICAgICAgICAgICAgICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAgICAg
            ICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAgICAgInBvb2wiOiBzZWxm
            Lm91dF9tYXBbIlJCRF9QT09MX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJj
            c2kuc3RvcmFnZS5rOHMuaW8vcHJvdmlzaW9uZXItc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYu
            b3V0X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAg
            ICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNl
            Y3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VD
            UkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4
            cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9S
            QkRfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgICAgIH0sCiAgICAg
            ICAgICAgICAgICAgICAgfQogICAgICAgICAgICAgICAgKQoKICAgICAgICAjIGlmICdDRVBIRlNf
            RlNfTkFNRScgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdjZXBoZnMnIFN0b3JhZ2VDbGFzcwogICAg
            ICAgIGlmIHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXToKICAgICAgICAgICAganNvbl9v
            dXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjog
            ImNlcGhmcyIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwKICAg
            ICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImZzTmFt
            ZSI6IHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAg
            ICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIkNFUEhGU19QT09MX05BTUUiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNyZXQtbmFtZSI6
            IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1F
            J119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250cm9s
            bGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfQ0VQSEZT
            X1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNz
            aS5zdG9yYWdlLms4cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91
            dF9tYXBbJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAg
            ICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICMgaWYgJ1JHV19F
            TkRQT0lOVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdjZXBoLXJndycgU3RvcmFnZUNsYXNzCiAg
            ICAgICAgaWYgc2VsZi5vdXRfbWFwWyJSR1dfRU5EUE9JTlQiXToKICAgICAgICAgICAganNvbl9v
            dXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjog
            ImNlcGgtcmd3IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAog
            ICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAiZW5k
            cG9pbnQiOiBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdLAogICAgICAgICAgICAgICAgICAg
            ICAgICAicG9vbFByZWZpeCI6IHNlbGYub3V0X21hcFsiUkdXX1BPT0xfUFJFRklYIl0sCiAgICAg
            ICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAg
            ICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAg
            ICAgIm5hbWUiOiAicmd3LWFkbWluLW9wcy11c2VyIiwKICAgICAgICAgICAgICAgICAgICAia2lu
            ZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAg
            ICAgICAgICAgICAiYWNjZXNzS2V5Ijogc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJf
            QUNDRVNTX0tFWSJdLAogICAgICAgICAgICAgICAgICAgICAgICAic2VjcmV0S2V5Ijogc2VsZi5v
            dXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJfU0VDUkVUX0tFWSJdLAogICAgICAgICAgICAgICAg
            ICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdSR1df
            VExTX0NFUlQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCB0aGUgImNlcGgtcmd3LXRscy1jZXJ0IiBz
            ZWNyZXQKICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VSVCJdOgogICAgICAgICAg
            ICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAg
            Im5hbWUiOiAiY2VwaC1yZ3ctdGxzLWNlcnQiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjog
            IlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAg
            ICAgICAgICJjZXJ0Ijogc2VsZi5vdXRfbWFwWyJSR1dfVExTX0NFUlQiXSwKICAgICAgICAgICAg
            ICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCgogICAgICAgIHJldHVy
            biBqc29uLmR1bXBzKGpzb25fb3V0KSArIExJTkVTRVAKCiAgICBkZWYgdXBncmFkZV91c2Vyc19w
            ZXJtaXNzaW9ucyhzZWxmKToKICAgICAgICB1c2VycyA9IFsKICAgICAgICAgICAgImNsaWVudC5j
            c2ktY2VwaGZzLW5vZGUiLAogICAgICAgICAgICAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9u
            ZXIiLAogICAgICAgICAgICAiY2xpZW50LmNzaS1yYmQtbm9kZSIsCiAgICAgICAgICAgICJjbGll
            bnQuY3NpLXJiZC1wcm92aXNpb25lciIsCiAgICAgICAgICAgICJjbGllbnQuaGVhbHRoY2hlY2tl
            ciIsCiAgICAgICAgXQogICAgICAgIGlmIHNlbGYucnVuX2FzX3VzZXIgIT0gIiIgYW5kIHNlbGYu
            cnVuX2FzX3VzZXIgbm90IGluIHVzZXJzOgogICAgICAgICAgICB1c2Vycy5hcHBlbmQoc2VsZi5y
            dW5fYXNfdXNlcikKICAgICAgICBmb3IgdXNlciBpbiB1c2VyczoKICAgICAgICAgICAgc2VsZi51
            cGdyYWRlX3VzZXJfcGVybWlzc2lvbnModXNlcikKCiAgICBkZWYgZ2V0X3Jnd19wb29sX25hbWVf
            ZHVyaW5nX3VwZ3JhZGUoc2VsZiwgdXNlciwgY2Fwcyk6CiAgICAgICAgaWYgdXNlciA9PSAiY2xp
            ZW50LmhlYWx0aGNoZWNrZXIiOgogICAgICAgICAgICAjIHdoZW4gYWRtaW4gaGFzIG5vdCBwcm92
            aWRlZCByZ3cgcG9vbCBuYW1lIGR1cmluZyB1cGdyYWRlLAogICAgICAgICAgICAjIGdldCB0aGUg
            cmd3IHBvb2wgbmFtZSBmcm9tIGNsaWVudC5oZWFsdGhjaGVja2VyIHVzZXIgd2hpY2ggd2FzIHVz
            ZWQgZHVyaW5nIGNvbm5lY3Rpb24KICAgICAgICAgICAgaWYgbm90IHNlbGYuX2FyZ19wYXJzZXIu
            cmd3X3Bvb2xfcHJlZml4OgogICAgICAgICAgICAgICAgIyBUbyBnZXQgdmFsdWUgJ2RlZmF1bHQn
            IHdoaWNoIGlzIHJndyBwb29sIG5hbWUgZnJvbSAnYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YScKICAgICAgICAgICAgICAgIHBhdHRlcm4gPSByInBvb2w9KC4qPylcLnJnd1wubWV0YSIK
            ICAgICAgICAgICAgICAgIG1hdGNoID0gcmUuc2VhcmNoKHBhdHRlcm4sIGNhcHMpCiAgICAgICAg
            ICAgICAgICBpZiBtYXRjaDoKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJn
            d19wb29sX3ByZWZpeCA9IG1hdGNoLmdyb3VwKDEpCiAgICAgICAgICAgICAgICBlbHNlOgogICAg
            ICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAg
            ICAgICAgICAgICAgICAgICJmYWlsZWQgdG8gZ2V0IHJndyBwb29sIG5hbWUgZm9yIHVwZ3JhZGUi
            CiAgICAgICAgICAgICAgICAgICAgKQoKICAgIGRlZiB1cGdyYWRlX3VzZXJfcGVybWlzc2lvbnMo
            c2VsZiwgdXNlcik6CiAgICAgICAgIyBjaGVjayB3aGV0aGVyIHRoZSBnaXZlbiB1c2VyIGV4aXN0
            cyBvciBub3QKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4IjogImF1dGggZ2V0IiwgImVudGl0
            eSI6IGYie3VzZXJ9IiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICByZXRfdmFsLCBqc29uX291
            dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAg
            aWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgcHJpbnQo
            ZiJ1c2VyIHt1c2VyfSBub3QgZm91bmQgZm9yIHVwZ3JhZGluZy4iKQogICAgICAgICAgICByZXR1
            cm4KICAgICAgICBleGlzdGluZ19jYXBzID0ganNvbl9vdXRbMF1bImNhcHMiXQogICAgICAgIHNl
            bGYuZ2V0X3Jnd19wb29sX25hbWVfZHVyaW5nX3VwZ3JhZGUodXNlciwgc3RyKGV4aXN0aW5nX2Nh
            cHMpKQogICAgICAgIG5ld19jYXAsIF8gPSBzZWxmLmdldF9jYXBzX2FuZF9lbnRpdHkodXNlcikK
            ICAgICAgICBjYXBfa2V5cyA9IFsibW9uIiwgIm1nciIsICJvc2QiLCAibWRzIl0KICAgICAgICBj
            YXBzID0gW10KICAgICAgICBmb3IgZWFjaENhcCBpbiBjYXBfa2V5czoKICAgICAgICAgICAgY3Vy
            X2NhcF92YWx1ZXMgPSBleGlzdGluZ19jYXBzLmdldChlYWNoQ2FwLCAiIikKICAgICAgICAgICAg
            bmV3X2NhcF92YWx1ZXMgPSBuZXdfY2FwLmdldChlYWNoQ2FwLCAiIikKICAgICAgICAgICAgY3Vy
            X2NhcF9wZXJtX2xpc3QgPSBbCiAgICAgICAgICAgICAgICB4LnN0cmlwKCkgZm9yIHggaW4gY3Vy
            X2NhcF92YWx1ZXMuc3BsaXQoIiwiKSBpZiB4LnN0cmlwKCkKICAgICAgICAgICAgXQogICAgICAg
            ICAgICBuZXdfY2FwX3Blcm1fbGlzdCA9IFsKICAgICAgICAgICAgICAgIHguc3RyaXAoKSBmb3Ig
            eCBpbiBuZXdfY2FwX3ZhbHVlcy5zcGxpdCgiLCIpIGlmIHguc3RyaXAoKQogICAgICAgICAgICBd
            CiAgICAgICAgICAgICMgYXBwZW5kIG5ld19jYXBfbGlzdCB0byBjdXJfY2FwX2xpc3QgdG8gbWFp
            bnRhaW4gdGhlIG9yZGVyIG9mIGNhcHMKICAgICAgICAgICAgY3VyX2NhcF9wZXJtX2xpc3QuZXh0
            ZW5kKG5ld19jYXBfcGVybV9saXN0KQogICAgICAgICAgICAjIGVsaW1pbmF0ZSBkdXBsaWNhdGVz
            IHdpdGhvdXQgdXNpbmcgJ3NldCcKICAgICAgICAgICAgIyBzZXQgcmUtb3JkZXJzIGl0ZW1zIGlu
            IHRoZSBsaXN0IGFuZCB3ZSBoYXZlIHRvIGtlZXAgdGhlIG9yZGVyCiAgICAgICAgICAgIG5ld19j
            YXBfbGlzdCA9IFtdCiAgICAgICAgICAgIFtuZXdfY2FwX2xpc3QuYXBwZW5kKHgpIGZvciB4IGlu
            IGN1cl9jYXBfcGVybV9saXN0IGlmIHggbm90IGluIG5ld19jYXBfbGlzdF0KICAgICAgICAgICAg
            ZXhpc3RpbmdfY2Fwc1tlYWNoQ2FwXSA9ICIsICIuam9pbihuZXdfY2FwX2xpc3QpCiAgICAgICAg
            ICAgIGlmIGV4aXN0aW5nX2NhcHNbZWFjaENhcF06CiAgICAgICAgICAgICAgICBjYXBzLmFwcGVu
            ZChlYWNoQ2FwKQogICAgICAgICAgICAgICAgY2Fwcy5hcHBlbmQoZXhpc3RpbmdfY2Fwc1tlYWNo
            Q2FwXSkKICAgICAgICBjbWRfanNvbiA9IHsKICAgICAgICAgICAgInByZWZpeCI6ICJhdXRoIGNh
            cHMiLAogICAgICAgICAgICAiZW50aXR5IjogdXNlciwKICAgICAgICAgICAgImNhcHMiOiBjYXBz
            LAogICAgICAgICAgICAiZm9ybWF0IjogImpzb24iLAogICAgICAgIH0KICAgICAgICByZXRfdmFs
            LCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24p
            CiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiInYXV0aCBjYXBzIHt1c2VyfScgY29tbWFu
            ZCBmYWlsZWQuXG4gRXJyb3I6IHtlcnJfbXNnfSIKICAgICAgICAgICAgKQogICAgICAgIHByaW50
            KGYiVXBkYXRlZCB1c2VyIHt1c2VyfSBzdWNjZXNzZnVsbHkuIikKCiAgICBkZWYgbWFpbihzZWxm
            KToKICAgICAgICBnZW5lcmF0ZWRfb3V0cHV0ID0gIiIKICAgICAgICBpZiBzZWxmLl9hcmdfcGFy
            c2VyLnVwZ3JhZGU6CiAgICAgICAgICAgIHNlbGYudXBncmFkZV91c2Vyc19wZXJtaXNzaW9ucygp
            CiAgICAgICAgZWxpZiBzZWxmLl9hcmdfcGFyc2VyLmZvcm1hdCA9PSAianNvbiI6CiAgICAgICAg
            ICAgIGdlbmVyYXRlZF9vdXRwdXQgPSBzZWxmLmdlbl9qc29uX291dCgpCiAgICAgICAgZWxpZiBz
            ZWxmLl9hcmdfcGFyc2VyLmZvcm1hdCA9PSAiYmFzaCI6CiAgICAgICAgICAgIGdlbmVyYXRlZF9v
            dXRwdXQgPSBzZWxmLmdlbl9zaGVsbF9vdXQoKQogICAgICAgIGVsc2U6CiAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlVuc3VwcG9y
            dGVkIGZvcm1hdDoge3NlbGYuX2FyZ19wYXJzZXIuZm9ybWF0fSIKICAgICAgICAgICAgKQogICAg
            ICAgIHByaW50KGdlbmVyYXRlZF9vdXRwdXQpCiAgICAgICAgaWYgc2VsZi5vdXRwdXRfZmlsZSBh
            bmQgZ2VuZXJhdGVkX291dHB1dDoKICAgICAgICAgICAgZk91dCA9IG9wZW4oc2VsZi5vdXRwdXRf
            ZmlsZSwgbW9kZT0idyIsIGVuY29kaW5nPSJVVEYtOCIpCiAgICAgICAgICAgIGZPdXQud3JpdGUo
            Z2VuZXJhdGVkX291dHB1dCkKICAgICAgICAgICAgZk91dC5jbG9zZSgpCgoKIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyBNQUlOICMjIyMjIyMjIyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMKaWYgX19uYW1lX18gPT0gIl9fbWFpbl9fIjoKICAgIHJqT2Jq
            ID0gUmFkb3NKU09OKCkKICAgIHRyeToKICAgICAgICByak9iai5tYWluKCkKICAgIGV4Y2VwdCBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uIGFzIGVycjoKICAgICAgICBwcmludChmIkV4ZWN1dGlv
            biBGYWlsZWQ6IHtlcnJ9IikKICAgICAgICByYWlzZSBlcnIKICAgIGV4Y2VwdCBLZXlFcnJvciBh
            cyBrRXJyOgogICAgICAgIHByaW50KGYiS2V5RXJyb3I6IHtrRXJyfSIpCiAgICBleGNlcHQgT1NF
            cnJvciBhcyBvc0VycjoKICAgICAgICBwcmludChmIkVycm9yIHdoaWxlIHRyeWluZyB0byBvdXRw
            dXQgdGhlIGRhdGE6IHtvc0Vycn0iKQogICAgZmluYWxseToKICAgICAgICByak9iai5zaHV0ZG93
            bigpCg==
          external.features.ocs.openshift.io/supported-platforms: '["BareMetal", "None",
            "VSphere", "OpenStack", "oVirt"]'
          external.features.ocs.openshift.io/validation: '{"secrets":["rook-ceph-operator-creds",
            "rook-csi-rbd-node", "rook-csi-rbd-provisioner"], "configMaps": ["rook-ceph-mon-endpoints",
            "rook-ceph-mon"], "storageClasses": ["ceph-rbd"], "cephClusters": ["monitoring-endpoint"]}'
          features.ocs.openshift.io/disabled: '["ss-list", "install-wizard", "block-pool",
            "mcg-resource", "odf-dashboard", "common", "storage-provider", "storage-provisioner",
            "dashboard-resources", "csv-actions", "inventory-item", "alert-actions"]'
          features.ocs.openshift.io/enabled: '["kms", "arbiter", "flexible-scaling",
            "multus", "pool-management", "namespace-store", "mcg-standalone", "taint-nodes",
            "vault-sa-kms", "hpcs-kms"]'
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operatorframework.io/initialization-resource: "\n    {\n        \"apiVersion\":
            \"ocs.openshift.io/v1\",\n        \"kind\": \"StorageCluster\",\n        \"metadata\":
            {\n            \"name\": \"example-storagecluster\",\n            \"namespace\":
            \"openshift-storage\"\n        },\n        \"spec\": {\n            \"manageNodes\":
            false,\n            \"monPVCTemplate\": {\n                \"spec\": {\n
            \                   \"accessModes\": [\n                        \"ReadWriteOnce\"\n
            \                   ],\n                    \"resources\": {\n                        \"requests\":
            {\n                            \"storage\": \"10Gi\"\n                        }\n
            \                   },\n                    \"storageClassName\": \"gp2-csi\"\n
            \               }\n            },\n            \"storageDeviceSets\":
            [\n                {\n                    \"count\": 3,\n                    \"dataPVCTemplate\":
            {\n                        \"spec\": {\n                            \"accessModes\":
            [\n                                \"ReadWriteOnce\"\n                            ],\n
            \                           \"resources\": {\n                                \"requests\":
            {\n                                    \"storage\": \"1Ti\"\n                                }\n
            \                           },\n                            \"storageClassName\":
            \"gp2-csi\",\n                            \"volumeMode\": \"Block\"\n
            \                       }\n                    },\n                    \"name\":
            \"example-deviceset\",\n                    \"placement\": {},\n                    \"portable\":
            true,\n                    \"resources\": {}\n                }\n            ]\n
            \       }\n    }\n\t"
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:03e50284f7fe090c57eddf90feea8ff48d78aff8ce95d1b297aa91e71c3ee577
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.4
          operators.operatorframework.io/internal-objects: '["ocsinitializations.ocs.openshift.io","storageclassrequests.ocs.openshift.io","storageconsumers.ocs.openshift.io","storageprofiles.ocs.openshift.io","cephclusters.ceph.rook.io","cephobjectstores.ceph.rook.io","cephobjectstoreusers.ceph.rook.io","cephnfses.ceph.rook.io","cephclients.ceph.rook.io","cephfilesystems.ceph.rook.io","cephfilesystemmirrors.ceph.rook.io","cephrbdmirrors.ceph.rook.io","cephobjectrealms.ceph.rook.io","cephobjectzonegroups.ceph.rook.io","cephobjectzones.ceph.rook.io","cephbucketnotifications.ceph.rook.io","cephbuckettopics.ceph.rook.io","cephfilesystemsubvolumegroups.ceph.rook.io","cephblockpoolradosnamespaces.ceph.rook.io","cephcosidrivers.ceph.rook.io"]'
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/ocs-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OCSInitialization represents the initial data to be created
              when the operator is installed.
            displayName: OCS Initialization
            kind: OCSInitialization
            name: ocsinitializations.ocs.openshift.io
            version: v1
          - description: StorageClassRequest is the Schema for the StorageClassRequests
              API
            displayName: Storage Class Request
            kind: StorageClassRequest
            name: storageclassrequests.ocs.openshift.io
            version: v1alpha1
          - description: StorageCluster represents a cluster including Ceph Cluster,
              NooBaa and all the storage and compute resources required.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.ocs.openshift.io
            version: v1
          - description: StorageConsumer is the Schema for the storageconsumers API
            displayName: Storage Consumer
            kind: StorageConsumer
            name: storageconsumers.ocs.openshift.io
            version: v1alpha1
          - kind: StorageProfile
            name: storageprofiles.ocs.openshift.io
            version: v1
          - description: Represents a Ceph cluster.
            displayName: Ceph Cluster
            kind: CephCluster
            name: cephclusters.ceph.rook.io
            version: v1
          - description: Represents a Ceph Block Pool.
            displayName: Block Pools
            kind: CephBlockPool
            name: cephblockpools.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store.
            displayName: Ceph Object Store
            kind: CephObjectStore
            name: cephobjectstores.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store User.
            displayName: Ceph Object Store User
            kind: CephObjectStoreUser
            name: cephobjectstoreusers.ceph.rook.io
            version: v1
          - description: Represents a cluster of Ceph NFS ganesha gateways.
            displayName: Ceph NFS
            kind: CephNFS
            name: cephnfses.ceph.rook.io
            version: v1
          - description: Represents a Ceph User.
            displayName: Ceph Client
            kind: CephClient
            name: cephclients.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem.
            displayName: Ceph Filesystem
            kind: CephFilesystem
            name: cephfilesystems.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem Mirror.
            displayName: Ceph Filesystem Mirror
            kind: CephFilesystemMirror
            name: cephfilesystemmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph RBD Mirror.
            displayName: Ceph RBD Mirror
            kind: CephRBDMirror
            name: cephrbdmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Realm.
            displayName: Ceph Object Store Realm
            kind: CephObjectRealm
            name: cephobjectrealms.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone Group.
            displayName: Ceph Object Store Zone Group
            kind: CephObjectZoneGroup
            name: cephobjectzonegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone.
            displayName: Ceph Object Store Zone
            kind: CephObjectZone
            name: cephobjectzones.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Notification.
            displayName: Ceph Bucket Notification
            kind: CephBucketNotification
            name: cephbucketnotifications.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Topic.
            displayName: Ceph Bucket Topic
            kind: CephBucketTopic
            name: cephbuckettopics.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem SubVolumeGroup.
            displayName: Ceph Filesystem SubVolumeGroup
            kind: CephFilesystemSubVolumeGroup
            name: cephfilesystemsubvolumegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph BlockPool Rados Namespace.
            displayName: Ceph BlockPool Rados Namespace
            kind: CephBlockPoolRadosNamespace
            name: cephblockpoolradosnamespaces.ceph.rook.io
            version: v1
          - description: Represents a Ceph COSI Driver.
            displayName: Ceph COSI Driver
            kind: CephCOSIDriver
            name: cephcosidrivers.ceph.rook.io
            version: v1
        description: |
          **Red Hat OpenShift Container Storage** deploys three operators.

          ### OpenShift Container Storage operator

          The OpenShift Container Storage operator is the primary operator for OpenShift Container Storage. It serves to facilitate the other operators in OpenShift Container Storage by performing administrative tasks outside their scope as well as watching and configuring their CustomResources.

          ### Rook

          [Rook][1] deploys and manages Ceph on OpenShift, which provides block and file storage.

          # Core Capabilities

          * **Self-managing service:** No matter which supported storage technologies you choose, OpenShift Container Storage ensures that resources can be deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Container Storage you can either build dedicated storage clusters or hyper-converged clusters where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Container Storage:** OpenShift Container Storage integrates Ceph with multiple storage presentations including object storage (compatible with S3), block storage, and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Container Storage efficiently distributes and replicates your data across your cluster to minimize the risk of data loss. With snapshots, cloning, and versioning, no more losing sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data storage for your workloads, using a single namespace.

          [1]: https://rook.io
        displayName: OpenShift Container Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - rook
        - ceph
        - block storage
        - shared filesystem
        - object storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ocs-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:1711a0aabd28937aac1424adebe96619a48b91f7f888f0588117f80891001da4
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:f37e2c611f4cda32172d4e8938938e488076fd8a8f19c252c56a80c48044bc02
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:028120f2212738917f49e0330b60be5f79a208ce2974bea22b71f533aa6ce3be
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:f75a12fc0c449008028d7fc8e18ca7c1afc37ba585885ac754fed860047794a0
        - registry.redhat.io/odf4/ocs-metrics-exporter-rhel9@sha256:c58c67263c08d84a68c8e45cfd77d59a35aee2151bf79b44377016ca0fb429ee
        - registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:9a0008f662b6afa9e2677a8a83bd662d806ad6f4d03fb49bccb78ed487ce4c7a
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:645d4fd09628f7ddcfc765e4ddacaa79fa6a7c35c03292ca2f6e317d48765d6b
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:ef4276b80cacae825d4831de88d3fe6ba7cc324464087b7b28734bd2b8de3315
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:0dc0733df67bb49886ac9715b81bd4ec4a36330fe1afdde657587e1883f5ca7a
        - registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:03e50284f7fe090c57eddf90feea8ff48d78aff8ce95d1b297aa91e71c3ee577
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:ffcc0806ec2afb8b40d75c70ef2a6d98bb887879f5b3f573e900861316ce9e41
        - registry.redhat.io/odf4/ocs-rhel9-operator@sha256:b1a9492bd9b314f86b4e192cf2c83144885ff863d9932e386a41691f50be6193
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:f8bfbf64b4bd030474e239aed5e03af9dfe920eeae444c5cc8ac3dcfa640fcc1
        version: 4.15.20-rhodf
      entries:
      - name: ocs-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: ocs-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: ocs-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: ocs-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: ocs-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: ocs-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: ocs-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: ocs-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: ocs-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: ocs-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: ocs-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: ocs-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: ocs-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: ocs-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: ocs-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: ocs-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: ocs-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: ocs-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: ocs-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: ocs-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: ocs-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |2-

            [
                {
                    "apiVersion": "ocs.openshift.io/v1",
                    "kind": "StorageCluster",
                    "metadata": {
                        "name": "example-storagecluster",
                        "namespace": "openshift-storage"
                    },
                    "spec": {
                        "manageNodes": false,
                        "monPVCTemplate": {
                            "spec": {
                                "accessModes": [
                                    "ReadWriteOnce"
                                ],
                                "resources": {
                                    "requests": {
                                        "storage": "10Gi"
                                    }
                                },
                                "storageClassName": "gp2-csi"
                            }
                        },
                        "storageDeviceSets": [
                            {
                                "count": 3,
                                "dataPVCTemplate": {
                                    "spec": {
                                        "accessModes": [
                                            "ReadWriteOnce"
                                        ],
                                        "resources": {
                                            "requests": {
                                                "storage": "1Ti"
                                            }
                                        },
                                        "storageClassName": "gp2-csi",
                                        "volumeMode": "Block"
                                    }
                                },
                                "name": "example-deviceset",
                                "placement": {},
                                "portable": true,
                                "resources": {}
                            }
                        ]
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Storage
          containerImage: registry.redhat.io/odf4/ocs-rhel9-operator@sha256:3ef1c2679e9229d51ebc04bd2f2b22ce9b11e8f7fdd7bddee69630fe5b1de47c
          description: Red Hat OpenShift Container Storage provides hyperconverged
            storage for applications within an OpenShift cluster.
          external.features.ocs.openshift.io/supported-platforms: '["BareMetal", "None",
            "VSphere", "OpenStack", "oVirt"]'
          external.features.ocs.openshift.io/validation: '{"secrets":["rook-ceph-operator-creds",
            "rook-csi-rbd-node", "rook-csi-rbd-provisioner"], "configMaps": ["rook-ceph-mon-endpoints",
            "rook-ceph-mon"], "storageClasses": ["ceph-rbd"], "cephClusters": ["monitoring-endpoint"]}'
          features.ocs.openshift.io/disabled: '["ss-list", "install-wizard", "block-pool",
            "mcg-resource", "odf-dashboard", "common", "storage-provider", "storage-provisioner",
            "dashboard-resources", "csv-actions", "inventory-item", "alert-actions"]'
          features.ocs.openshift.io/enabled: '["kms", "arbiter", "flexible-scaling",
            "multus", "pool-management", "namespace-store", "mcg-standalone", "taint-nodes",
            "vault-sa-kms", "hpcs-kms"]'
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operatorframework.io/initialization-resource: "\n    {\n        \"apiVersion\":
            \"ocs.openshift.io/v1\",\n        \"kind\": \"StorageCluster\",\n        \"metadata\":
            {\n            \"name\": \"example-storagecluster\",\n            \"namespace\":
            \"openshift-storage\"\n        },\n        \"spec\": {\n            \"manageNodes\":
            false,\n            \"monPVCTemplate\": {\n                \"spec\": {\n
            \                   \"accessModes\": [\n                        \"ReadWriteOnce\"\n
            \                   ],\n                    \"resources\": {\n                        \"requests\":
            {\n                            \"storage\": \"10Gi\"\n                        }\n
            \                   },\n                    \"storageClassName\": \"gp2-csi\"\n
            \               }\n            },\n            \"storageDeviceSets\":
            [\n                {\n                    \"count\": 3,\n                    \"dataPVCTemplate\":
            {\n                        \"spec\": {\n                            \"accessModes\":
            [\n                                \"ReadWriteOnce\"\n                            ],\n
            \                           \"resources\": {\n                                \"requests\":
            {\n                                    \"storage\": \"1Ti\"\n                                }\n
            \                           },\n                            \"storageClassName\":
            \"gp2-csi\",\n                            \"volumeMode\": \"Block\"\n
            \                       }\n                    },\n                    \"name\":
            \"example-deviceset\",\n                    \"placement\": {},\n                    \"portable\":
            true,\n                    \"resources\": {}\n                }\n            ]\n
            \       }\n    }\n\t"
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:b7758b197cf423573de96bdca0cb3045852b20f34fc9c17ad1632b9c3a3cf3d0
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.4
          operators.operatorframework.io/internal-objects: '["ocsinitializations.ocs.openshift.io","storageconsumers.ocs.openshift.io","storageprofiles.ocs.openshift.io","storagerequests.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/ocs-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OCSInitialization represents the initial data to be created
              when the operator is installed.
            displayName: OCS Initialization
            kind: OCSInitialization
            name: ocsinitializations.ocs.openshift.io
            version: v1
          - description: StorageCluster represents a cluster including Ceph Cluster,
              NooBaa and all the storage and compute resources required.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.ocs.openshift.io
            version: v1
          - description: StorageConsumer is the Schema for the storageconsumers API
            displayName: Storage Consumer
            kind: StorageConsumer
            name: storageconsumers.ocs.openshift.io
            version: v1alpha1
          - description: StorageProfile is the Schema for the storageprofiles API
            displayName: Storage Profile
            kind: StorageProfile
            name: storageprofiles.ocs.openshift.io
            version: v1
          - description: StorageRequest is the Schema for the StorageRequests API
            displayName: Storage Request
            kind: StorageRequest
            name: storagerequests.ocs.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Container Storage** deploys three operators.

          ### OpenShift Container Storage operator

          The OpenShift Container Storage operator is the primary operator for OpenShift Container Storage. It serves to facilitate the other operators in OpenShift Container Storage by performing administrative tasks outside their scope as well as watching and configuring their CustomResources.

          ### Rook

          [Rook][1] deploys and manages Ceph on OpenShift, which provides block and file storage.

          # Core Capabilities

          * **Self-managing service:** No matter which supported storage technologies you choose, OpenShift Container Storage ensures that resources can be deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Container Storage you can either build dedicated storage clusters or hyper-converged clusters where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Container Storage:** OpenShift Container Storage integrates Ceph with multiple storage presentations including object storage (compatible with S3), block storage, and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Container Storage efficiently distributes and replicates your data across your cluster to minimize the risk of data loss. With snapshots, cloning, and versioning, no more losing sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data storage for your workloads, using a single namespace.

          [1]: https://rook.io
        displayName: OpenShift Container Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - rook
        - ceph
        - block storage
        - shared filesystem
        - object storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ocs-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d
        - registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:b7758b197cf423573de96bdca0cb3045852b20f34fc9c17ad1632b9c3a3cf3d0
        - registry.redhat.io/odf4/ocs-rhel9-operator@sha256:3ef1c2679e9229d51ebc04bd2f2b22ce9b11e8f7fdd7bddee69630fe5b1de47c
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:70a3c6bf4bde84c7e43fd58c60abc5612d25c792042b902f6abee94bfaa63f80
        - registry.redhat.io/odf4/ocs-metrics-exporter-rhel9@sha256:3332a93e3ea2e83af15a7edab02a9e00d70e81e44338b63a0eacdea54142ad00
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:0084d9053ad2c74926ca02d2ede0461abd59227eee98d122c45e08a5ebaf629a
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:cf7f72469d25fb37301a8159406613673f8968c82af1035fd45a470e0ae8acfa
        version: 4.16.16-rhodf
      entries:
      - name: ocs-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: ocs-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: ocs-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: ocs-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: ocs-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: ocs-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: ocs-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: ocs-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: ocs-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: ocs-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: ocs-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: ocs-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: ocs-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: ocs-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: ocs-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: ocs-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: ocs-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: ocs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: serverless-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: serverless-operator.v1.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.36.0 <1.36.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:52ba6b1252b6f3d41e2a0b3b880cb768010ab9dd925ac34596f6adaedcbbe07c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-source-rhel8@sha256:88323da16360981f27b8e4686fbbb64053e4e0e3c79c65d695edb317b334e1e0
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:3fe326ad9a00b112fc888ffd345c8d5042f0a22cf32ea575931192c39ce1fece
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:259e5dab50d36cdbfc117a42325128c4e16f34c048ba64b36d4a89132ddb4c1a
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:4344716eb2940c6a3fa9a278250f936602bb1549fece6037912837719c178b24
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-transform-jsonata-rhel8@sha256:3af4c48030263eb5769e5f36a6dd13fd390d8566efaab37532f39e1df1c95c67
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sns-sink-rhel8@sha256:51b5e18e8d4c1ca81d7c5ddfea6352a3887f757acfcc53bd2f8bd3b01337ed42
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-ddb-streams-source-rhel8@sha256:a903480159ebf1d8652eeaa702d9885fb76bb106c02d75cec5c82c31d77d6410
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:e3ae1cc6012f19631c012a07b113fb015977d2ee4f9faab3207fbc2464c25aa7
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:596a7272d6de714d7dbc6e00e9e50a2872846c269547465252d953172ef0b009
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:f4d86b196457cdeebdc380b6f1b707c10f8dee789c7779c90ee1332e511af8c4
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:f2d112ba9fdfccea988589b99b463c8a7157475f7a3f03678e8999e65b678cc3
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-sink-rhel8@sha256:afa634e029bcf79a972f49372498d44b45bcfbdb6eb28ecaf4f0b0b2a0a1824b
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:dcfe5fbc57b9f8520adbff368c8bf29eb491c9c26d1a1737a1722cffea83eb1b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-source-rhel8@sha256:27e74287774cda208f667bc2a386debe6b6255e8991abfb37c959c8525f2af53
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:9b328fff0cc8545fb9bb1f24f11d78cfac3ee76dedc11fa4def4262d79a60e8c
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:82e69a1890042f5636ac7bb645b1aaeaba0a4894851eccc1fa1c63296918e31a
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:3d780ad1ab280b5092b90e277485d589da2a18e8b08090af7d95aecd17af545d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:631eb3a8dd9d83ceafa7d8a4da4b1e2d4f2740e6017c766828dcc7575cf1322f
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-log-sink-rhel8@sha256:30b8045ef5f925a36c58941386f1899bec125d005a08aeeeba5edd49b0859149
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:c7c2e738e0d9da1d2d1993f68938e72c4709b1f28a54fa92224ef6a0e456b0aa
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:a1834b620f875681388f0ae224dd0771773a8b61b1d8e5214a942f232854b87b
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:074ebfc4494581611647ba90049655a470ca43743610a27be77db66a8633bac6
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:429d5d9568cca31fbe83c663276a300286af4cdbf3d35685afdbe5f6f2f8c6e0
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:5352e4eb4e426a41521276decadf204b6d23b4762a298e4d3dc73969e3eddc68
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:bf82b37ab8fd29c7a7cd846008a16d14b4cfdc1c09df4f731d71be3957e90123
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:6b7316cce3432d391adc6ac38c92ff0ee86dd7ae0101957cc65cd338cc69e352
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:3b04e9d8cd4b5e43a07ece838d10478723a6c6529350f010550b4a0d1d0950d1
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:6210a946f373b9c36c9b31293debe6fed090008b8234c0ba17d37fa81646172e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:c39cb88afaddc295d93a3774b8a20d001b82392df540fb502ee02316d82460a9
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:0c2cdf917bbad1f5bca8a2a2ef12a96cfcd0c77c190cc35687eb57169b54ab6f
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:187fb9d64b53c55a029ffce3c0f9e7c3304683fa8a0e832cfb7134e8a3d0c237
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:d6f338e08fa937c8a4ccfec692424de4acded565f7ca3971ea507dffeb351503
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:f3d87a7fa5b5f2818effcac3304bd56eddcdb0db97a18ac0f35a06fc7e223e7d
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:4aeb68a22f63300e54391ec6b4cda8c4e665d904b5dcdd3c40155040525742a0
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:ed60e889a677fb139f2eb4bd9ab4fda6ecc39fd050ca2b6cc1c19717e428535a
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:77e4689b2ce7a713e905553e0e42b8e128045f40cb6c6d4ee082f126cbc46b00
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:4274004b6289fa9bea3fa6e56d64077a328a6b5e63a9f0e9c41f6d561f682bd7
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:a6f565d73375e827d830dca85838f4c4cb7f08b0c078acb0c0d567b25f0d256a
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-timer-source-rhel8@sha256:8fed5ef9a1f39a55077bb5026c8854d0ee7fc8d8ff5da277ce9c1bfc2ab2474b
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:78fee4c596f2848487ca704da3fa1074e3c9c99b3f96f66833aeba63836d7b93
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:a47b6b5237a87ccd90e5a49a53d06485fd55c87eebf12d01c335b87fc3a8c494
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-sink-rhel8@sha256:d9b944e64391a996104afbe8a21f66316a948e9fb7b91734e861847285be59c7
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:ee6533b71cf280877c9d7489dfd821c9209c6802bed7fa3fdb4f8f00fbeacb09
        version: 1.36.1
      entries:
      - name: serverless-operator.v1.36.1
        version: 1.36.1
      - name: serverless-operator.v1.36.0
        version: 1.36.0
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable
    - currentCSV: serverless-operator.v1.29.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2023-07-28T05:42:31Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.29.0 <1.29.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of a particular version of Knative
              Serving
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Eventing
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Kafka components
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:a3c5d5fd8014c432830be912aedebbc58f9c60a8cabd87ae6c42c7b8d27db8f9
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:a59ba5aeb8e66db3d8ba9074b8e553603c59e6b0e35b57b82c4ce3de589b94cc
        - registry.redhat.io/ocp-tools-4-tech-preview/source-to-image-rhel8@sha256:98d8cb3a255641ca6a1bce854e5e2460c20de9fb9b28e3cc67eb459f122873dd
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:56e48afb1dcbf4e266ace944105f3c45042550d8d4a0528c982225a793d26d56
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:31d45f10a8519addec9aac115c5edc8a2e14f9621c875d95f123d1291776ef4b
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:b780634fffe2257eb845f8c1d25498d9c654d0327c1f8841b40d90c176a2bfe0
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:04866e16881f64ff2b5edd379eab0731c9d66a47c25ebfe79c8d8ad33d8116e6
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:61fb98802490323bb03ff4a7060334d8fbda9e9f95cd6366b29b1b6494e4ef8b
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:3b7449287adf98f0cf0a1dc74e03b69218be3bcf5538327bc34ef776cc1c69c4
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:10788ca67809e5a2d76d6d72df7dd64ea4407773999707a7681e0cf1b5c4695f
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:216f58dc8c8166c0c69fc0312b667daad14cf39b7aeb7bc0fe1e313376f1a5fd
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:aea5f200a845e1d4dc718d76465fdbb1eeac2b20ecae224d5d20a98214ad0357
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:e5db11a0adb889f907cad797dad9cf2de09c433e5522fa333193cb35575802d8
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:c4cff5ad5b0ac79e71a16b88640a312ab320817041317f50caf575f7e12a864e
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:c2cb80bce34a74afa76151d2b44dc25ece73fb274cdfc6fdf243b1b0cb7d989f
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:3aab995872336e6ccb590ef6488c097cff38a91348055031145427a00aa6b732
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:b520305189c7567b2f8ac867b20468259fb22cd3edc5b4576af2bfcdb0c11855
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:8273842d8e1714fa03c7e0809924b1ffc5ef282f79b9606bd8ac88c03e7a16bd
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a10e76aa15707e708f82c8d9e8b5179bf67780d99f3c3e2e99b84ae2a0aa4ecb
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:6342358e6aed2339333ebf37fc223a41a0a3e199a0e4a840f3f66e5accf8f233
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:594c3ac24bf59802e2d3fa127e4051277bc07b6444f934f535fddeb668ccc144
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:4408d1fcf34c0ffe9327d4c9c0bc772afa969d831bc5de1097164d10b4e8d900
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:03a3d6f0314acf4da4f3d3d85f1e273a3ef25828d28f5b22f394441d55a645f3
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:7dd11d385b3d5a178566d901650ecd45e4af395faec2141d57b98af8fc1e5151
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:a4943c0563c45ceccfda0a01e26be3a01c1437c89d927714c577bf6060495958
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:fd86ea2a34de43014dea3ba2b8d705a3d94347d3a24a370a8bfd032ed239dc56
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:f37f494956adf0892ffca34d2a4c9b839ce28dc9f2779ca32e8801fb4f0e9105
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:557e73127326be051aa6257e7daeb9721e2af04d2f06cc26cfc7a1073666b989
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:fe916630003194dbcd3894dd880cad8b15a041a5e534cd8f76fde6296db00797
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:397eb425f383c1273c7976bea69a603530676a00a61f461ceb5f48d67901c4cd
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:bd134e64c156dc5c81d14522c9c9efcc1f56b2e972ac8ecb1bb1b130c76adc60
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:ab90ec9b3cd9203fdcdd7c43abf95f48c9d7479a9c7af479eb1e4b6599ba9642
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:c12c471902b8e2596588d9e553ad59b501d979dd04f301d45f7d45e794737a8c
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:de0634d14962bdb8701a8c29e80b7b70684824462bc5203e5fa7f139839cb73b
        - registry.redhat.io/rhel8/buildah@sha256:a1e5cc0fb334e333e5eab69689223e8bd1f0c060810d260603b26cf8c0da2023
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:0229a0a4f026f04f8fa0a1cd8e7afea10117fe3f12e21986ec6089769c8afcaf
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ad399867cbb712111db6affaf5c9e3573699ed0eb0def8ce6057de02054ced4f
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:3d7801f7ffe6420595779673a170235921a9090333f8ca51d5b7f664e19c82b2
        version: 1.29.1
      entries:
      - name: serverless-operator.v1.29.1
        version: 1.29.1
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.29
    - currentCSV: serverless-operator.v1.30.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2023-10-30T14:03:25Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.30.0 <1.30.2'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of a particular version of Knative
              Serving
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Eventing
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Kafka components
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:9a28cb1d4c4f4dd7d494829eeeb7122db2174a27747f915244672c95dc8f9d12
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:4800c85e8de031bf77f6fae0e4c8e6c1a9912ccdfb77799cc2ea574c37d870ac
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:66149df8549380673c24e1487d0c2c8eb638b7e502455671ca64ed6f32ff82c3
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:a0c5a91348ec71697747113990652feb024143fbbd0cdd67e8ae672730383a79
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:2d3e314a6584fc6974517c1a992fb59f5133f54c2f6cfaa42f579c0d332e0300
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:ce4b36c90955a356a5774410356aafdac2b60795526a3488b0b76ac2ec1a018c
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:1f7e65009168383cdcb6b2caa22fe4db49e03c71574973a8086f6c332dc5e26e
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:32445e4a73bc7c1e8007775adeb147175d934b17e7d3d4dcf422104dbe80a52b
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:c9353a022e6f883b7c0b2d87b1bc4653172feb13806ebf0aa075d6f9f43c264e
        - registry.redhat.io/rhel8/buildah@sha256:a1e5cc0fb334e333e5eab69689223e8bd1f0c060810d260603b26cf8c0da2023
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:92a39ca90984cebb1938447a31c9170b52994ae98f83caa860acdb2c3ebb75c7
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:06e9e073f092dce67d4227e282fd8eea6dc9f38b90a8d444cf448d80d027b616
        - registry.redhat.io/openshift-serverless-1-tech-preview/eventing-istio-controller-rhel8@sha256:b6ef281fca2b80d379b1e28fe3f23119df8ff765fb2cca291d514053fb512da6
        - registry.redhat.io/ubi8/python-39@sha256:0ff58549041d32b685cc850aaa85da32d1ef73f99cd4b6e0f6882dad5f0399bf
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:21a638b481639c3c58e8b026d875917c30a5874b08fa4ceac40d447076b2bfa4
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:124087e6402caa375b24b976c3b9fd7d53890e20ac3b0b57ba3ac646e352c7ce
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:16939b802f4c16a3d9d941917d0363325f40cdef28f6e1a982bc95401117674c
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:e9505aba7c44cffc3f261a5dc7617a0e54ed98e78f73ea3c9c7a508560b792ed
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:e8c756b659f6187790bc1a5115e0f71791bff96630859a905bc5b75d14596e0e
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:a9c95fc208cc03625cf49bcebdaac73da823d1fa3bfae4239d7db20d3a0b8723
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:3fb6e7626d2b0c92c2a2fef35e9b8171382b4bdf54393f2ab1281ae7ab441761
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/ocp-tools-4-tech-preview/source-to-image-rhel8@sha256:98d8cb3a255641ca6a1bce854e5e2460c20de9fb9b28e3cc67eb459f122873dd
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:12bc62246142a605f4efed0f6bb8843aa2df7699cc1c5ab53debb94b0682e9d1
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:3c8bf480dbb05e09e8d07cab12e4736643f61bf9b5d98efa6bc0557eb31887bc
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:1455606c8edf611d996a48e29ead53e0bb4078f55384b7689058bb25fc59fb2c
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:6057a2d18406e00b0b7b1731870f707844e39193d4cce6de50137b77f17d220b
        - registry.redhat.io/ubi8/nodejs-16@sha256:8abeddc6ccfffbe313f87e5e069e18cb7047f2c7880ad63cf15bcd0a79ba71ff
        - registry.redhat.io/ubi8/openjdk-17@sha256:af305e087061cbaa4107ce0b304a1ef987f6b54bae3f75bfd529cfd084cba9b5
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:2264d2cd4ee2a720667d625ff42b6f8be4f2cd65b57a91eb6044da086a110740
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:69ea22c646c09208ca68345d4b3508453d8c49f7f7fd9a4ff4c4cb78be7bff43
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:a5154f9d13a1c373b8ba47e2bcfb9d4474fd03e08db4134c92aa4f049536f1a8
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:96d6f52711ba433ab2f273be8080840a88e626e989027480424d19c2fe1e9454
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ef116868c9c60e5ee246060688da49832571ff6a7901cfe143a42d048743736e
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:01c8fee92889a84a19136b0843e4fed5db31c1f2a6d997669f103373fd995e36
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:5c933641efc6bf919b1945be5a01dfb7774ae865975b1bb1f44653b3e4ca9f1f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d03f2bf8f5e7cba47102ba26ea0149ae6719076c645d5e451058392d6044b738
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:3508434d5ab6b38c8a7466256f1158ebcd75e8b751afcc5255da8c21566804f0
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:cd1abeb2067afa44a0f1eca301cce37307582d56c1147820ff86b9b64a40588b
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:85a369d5696771d0932b61760336e52ae3f7c66f9a3def5312aefcac03ba516e
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:96c135ced2e3bff434addb012368d96d0778ef1d294ba507ae7e624811e20d46
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:e97eb2bb7eb529cbfe67c14e823ffa1269d614e2d35723896e89efa12a499c6a
        version: 1.30.2
      entries:
      - name: serverless-operator.v1.30.2
        version: 1.30.2
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.30
    - currentCSV: serverless-operator.v1.31.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-02-01T19:15:42Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.31.0 <1.31.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:9221a833ecc66fe2354aee7aaec2a2470b21e2d89cb6ed287ece32edab2a2edb
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:22ffe08d6a0d5461158613f2b2a2ac6936025357e282a09197a750c42b5e6922
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:3dab790fa845ed7eda088e5abdc87799da1354884a53ff1ff728739b8e278418
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:1ee675125f7141e91f4ee57ebce082967834bbdabe1a870c221f31aeb8750be4
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:486bf60d17a5d265ebd65d590a6ee1a7e13bf3953ea797b2f6c2b5910789858f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:bf693eaa845373e1a06041626a9314ec00519a6fd2437b12f7f4f3f104864039
        - registry.redhat.io/ubi8/openjdk-17@sha256:a8165bc2cd5051a96d6937e25ed03155bbd4b731da6e58cebfe2ea83209c16d8
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:565becc77dbe3bf03d89b1399d6e1853253276a71e80b6b14069285d03cb14b2
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:433821e6c7f43da5ca2e38c0550722feb179a0bab021decd9a97ad020cb78b76
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:e49220eaf3143a8958f0cb1f22115b2e834fb7177755433bdc592b753a54a8ec
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:1fcecc11b55007b9c6969a24940a391968378e4102eebf7ceabbe7efda138575
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:b37dd2cf4b459fecb6d681cac36918e486c3bd22501df50b90e20b15fe164d56
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:b2480154951109a159acfb88c5472a312a515d6a6b060d832b3f7bfc16f90ac5
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:887480662b1bfe09626fbca8ec7c5da879d097c46dca11764ea98474ec78ab5d
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:cda04e3b7c2c1f37164733ad162ea0c4e52f14fdc8eb8bdef8507f0ed5454640
        - registry.redhat.io/ubi8/nodejs-16@sha256:f90b28179d1d5de68d42ef1a9166d70a9dfea85a1c6a4bf12dce2c59eeb2fbb1
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:677be039353b2cee51d5faa5d5a2abd6a7bd46992ffe9521eb033eea9925c785
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:3181ad775ead55f317f96466c2719b2b73c5b35d31933733da21273955d160c4
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:9e120ecdbea54eecbc5e0412c1774f8a2ee70d7fca682ecd0757b8627f70680b
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:eacdad8db61b23634c3f89bffa3f1c30ac5c7bbb5e938c6208003ff00f9595f1
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:df1aee67a161fa19d7f49ba56d53e0682cc1996436e985a12a2068fa419b54ea
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:3b01ebdb0704dc6a1a993ac076819489495ab3fe32e3c14609e42d1358856be2
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:d8bdd96c6a9b31f48c961ca438c254511db785fd34bd10b98084b702baf0ebab
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:63e08279d2852d494d4f8836c54e2ce942a342f452a23ef437df9640bb2f5bc9
        - registry.redhat.io/openshift-serverless-1-tech-preview/eventing-istio-controller-rhel8@sha256:5b38e5c71e422161e40f9c37dd2f46212cb80e8e37b00247767276b957080397
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:80361e052d366ffffe9a99893cd83965bd69a9ff81d8c9262976de145bbc6184
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:d2a34100f6309c3cbac7c9ccd452b7bdab2a533afc99463bfbf91e3ba2706d54
        - registry.redhat.io/rhel8/buildah@sha256:b48f410efa0ff8ab0db6ead420a5d8d866d64af846fece5efb185230d7ecf591
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:0a2989f688b966277652957b312df736ada43fff78fd89b17f60b2137e38f115
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c078071cd51c7fd5584039b54a2004331c05a897ee8b0689bfcb9204a8ebae80
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:8353ccb4f0d70597d61e4fd1f799d4bb74e5bc038f6b9916059e712aa70b97c1
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:3be4fc173b489d0beb768739f60274e07d95ffee86d70a8423e1f59678c805ea
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:3a4e7412322755b4c278540c965a403f9e178f138bd4ad7929c1c24a8036dd14
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:de62762efa0df95ed23df56f5520490d8a231e47ce13c433c512254ceff81844
        - registry.redhat.io/ubi8/python-39@sha256:4f35cbcc6f4e108c69f7d2b70cce5afb138741b0c961506376d524e001b922f4
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:99f5dcb92a33c047f0b68236a3a42ffaec78037dd0a5eb79d1144d62d6bfcf43
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:2572d98363cca2dd94984d6cefa573c5054fc206ee70048f901ad4b846d8dd52
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:dbd9c0cac86f79b8d91ab65ae7543854144781d4cb17ef16183931ec9bc234a7
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:7941b421b3e18bda8c565305241621fcdf1212cbbe0dc02f38f4ab0565d2bfcb
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:af5c30fc9f631d9ed110641b815315d8b4aef539f687277a3821d9939ca5dd46
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:ef7d28580e75d746449a90cc0e5343ef3f4ddd75bd9fb866d97836bfbc3b314d
        version: 1.31.1
      entries:
      - name: serverless-operator.v1.31.1
        version: 1.31.1
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.31
    - currentCSV: serverless-operator.v1.32.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-08-01T10:56:58Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.32.0 <1.32.2'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:16c026f09b377c44253cdb902b9a5fa0ae766fd0185cca1863798a212e0daee1
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:2f0f84767f7d76dbbf4387ca2e272766267a2c46b5bc29f5d01f05dfa39cb944
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:20566173a8e907980e603d82fae4d96fcdfd5c43d9004c2e506cf0e09a47b91e
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:04e5b3882223d45a5f5383ce6481efb7b2cbedea57ebd894caafb1454ad8c09d
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:e1e2273fdf6384d95e1128e54c0794d7d620291a272bfb1f91e0c2723be7f1b7
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:22b4d47725ed8cdd663ee26a03f31198fff9e71dc703532e70c42bdfb4840811
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:e7e6cbb46720eeca3f0f4b9829dfeeb3db373a83c89cd808b667ded4d5e139a4
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:a5a7b2be50f211d4d9d555b2a2e2302096c139ff5011d984cc386f1530007f16
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:0729de0f3f03ed220216c6cb718fc34beece0ceacd926b5e2480bdfcf44df5cd
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:ab9a1b952e9603b4ecc5f197715a1d871abc25ebb0c5e6470d4f0b365ae94276
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:2a30ae808d5c32d812406ff4ac9b472e82b6bb0555e62eee4c2c6a972f8750c2
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:b008da825ee2429f04cf39258389740c68056f82cf4bde0cbd3db850b6256885
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:5930caa25ea875d29dff0ec5e235eaa13cb6dc98980d8045335eba1a09ae761d
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:da5f813dea324ffe9679524ea7f3801536fa48f7dbbd8da56a8492b916b15fc3
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:990962ad245a6d62cd11551bd4e1f8b14dc08abbebe47beddb2f4874f78d4f47
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:b0db9f0fc85fcdcd5da980633a3a44be2bbabae068f89d0965263de38789e812
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:9d36cde68bd822016e967148eebd2ef3e7ffab9dcdd37d17f33017e9b703d9b2
        - registry.redhat.io/ubi8/openjdk-17@sha256:a166bd2fedb99cb62ebf4695ab46e19c7e451f887940cdf0b8245c81e403aea2
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:9b6524f9c0e7f78e6ffb372a8587eaf5f4f1281eaf64f93dd00b63be84a4c21e
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:42f31eb8f12dd36541c8c0b6572d942c62e2064d5418428482d676d117a6c426
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:7c8cefd5823ad953f112b905c651f12a97491aafc008b2ba0111f0c9e83e50ae
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:3be20d63d7e82f35ac88274532d6ab7839f9a7306ca272e7d532ac03215faea9
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:46b8dc1c8ea6c40401d4d1a9868ab3edbe8b9e92364a43add6722489e2eefa04
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:ab09a1710aa2331e24df9e7283f3909fdb104c1b314dea88e919d0a396f334d3
        - registry.redhat.io/ubi8/openjdk-21@sha256:bb95ea866269ac8652f666ee1587d71f4f43da33493c2be7f6a9548bcfd2407c
        - registry.redhat.io/ubi8/nodejs-16@sha256:69439fe83fa957089cb6fd3c1042ab71d719f81dd457a5d1fbc56c1e3ea600da
        - registry.redhat.io/ubi8/python-39@sha256:8c96e8648cc6651d0cefbd20496a89f208621601f9664554d3fdeef8ced25453
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:e55c70591c0498b6b85f0d68e304e88b830ea0b35219ba8881143555757be4b7
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:25cbdf78b67e7f5a360313ccfbe3b6e2c98873b4e9ddee46c115a623e77c7abb
        - registry.redhat.io/rhel8/buildah@sha256:55f5c7d88c77e1ab5a2ab8e877172b2b66cb196e21f10dc45148470d0ee8bbae
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:1a1de677297525c0c5d317fe8a7a42684fe3f1b422a7d6bea1f70f0924653b08
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:0507ea8745267fdb53329a5274d59d23c387225d2594af51dd2603b4ad99db0f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:4c9faea6fd162db3340aee8c11af11a08c9d79bbed104967c25df26c4933c8ae
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:14ca32e999767776c3e585ff5eb1aa568f7d76fd3e5ae24cae05ce7564678080
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:8033b7c1e903ba313134645c626eb3bf7ab1a0d68dc2575bc8e26884ad4f805f
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:e70dd3a9e067075eb56e5f2772642545e308348cd49b89357c02de3979b1a5be
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:3ffcf746df914f0949477b46f90cb2992bb63f8198d73ffd54208e7a777a9cb1
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:1befc9969edd2062b3df69f45aee9771b33a2659c3a14db0aaaf5c3df985af58
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:0551f1657dc158bd41eb2338dd6514d55c39fc77300cdf1744d8f5ab14e9ed1b
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:4933b76d95d411d597d4dc460fc46c92875d28e107cda661946206a9d934b8b5
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:70c26c7ac48d1eaaf1f95ca5c5b8ba8ee6b047a3c9c4a2c2db9c49a506bb2205
        version: 1.32.2
      entries:
      - name: serverless-operator.v1.32.2
        version: 1.32.2
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.32
    - currentCSV: serverless-operator.v1.33.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-11-20T08:38:34Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.33.0 <1.33.3'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:e1cb1c24333605e52364456f9bd8d86d787d1bae17ef75dd04be61257fdb2256
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:6dbfa2573a2c78e450d3c1c04c6e7cbc50d9272233e828fe9b42e575ecf386c0
        - registry.redhat.io/ubi8/nodejs-20-minimal@sha256:acee3b00ccf8ff49bcebcf10e028e34a84d45f28f855363bafb3b93c75e84f11
        - registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:e1cb1c24333605e52364456f9bd8d86d787d1bae17ef75dd04be61257fdb2256
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:eb95deb2105f14c6a9f3bb2cdd7306b446109a9b142a7bfc726b6429f270d9e5
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:ea9e4de6228d5c3ed494c2ce5317ec290543ad40bcf7c43c2a3d6565e5d94e5d
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:359f8b9fbb4c3c2f79e8c70f3b50c994693b930f6f078c34c87112ebb93037e3
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:c3dee2598088bda36ef474f4748943a29169ed9589aeb0f77bc79a7f6bef9907
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:672022b6f2a080483d9d6f6f4a5f42e52e522a5797105f3465731809c6cf4947
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:28c9d05a3ba73b76f293e8f4902b574deb44f77a2bae8cf92473b8a8b10cee92
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:a0399a993155e755409355740b260d9b0019559a62be3b6253d8693a4fb59286
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:42d16e4d9f14a427e7a5b7e3a4ea07c258e0f2c74984baeee3b4926a8c5e2640
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:a8821b59d495e24598b27c6510b32ef19de2bb04ea1e568fc5afb9135d772036
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:9b850afa750a67120f68a49f4834ba7f105563df22cbcbc563317fe1545640d1
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:09060de091d617bd4606f7a289b4528265ae0b7a04737f383e79efd5b0d3eb79
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:dd1f7c850f7cf2a4d70403deedd8a57ecc6e0ee5455d68fdfc49ff175fe42dde
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:93b2f5211e79abc38bf92f0b8fb8adf046303efa2fe62c47a2fef9dfb5662f55
        - registry.redhat.io/ubi8/openjdk-21@sha256:095175ef8573de8055e67bd5c8a21098448ac73db4e0d6ccd5269dd36b4d263c
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:b871bd08b782c9168e9a54340efffc0b0effadde8079915b848e4eff72bb3e95
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:879fca768b46676f1df89de1cdc4ded8dac310209c146c2f381ed1deb19b06be
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:993597ff83a4d62a23f642fadc92f182944b953770716b73b7d19492b5a462c3
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:eab4aa93b9d175045b1f01bff09681107d839285e0d633cd5146d2a2efab373c
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:af66cc7dbed88907d96a390fe5c8807a2a60bbe3276b8cd13b4b061cbaf7b162
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:b12272e1a9cb4f4bb62c4badc62b0bab110abb99484612dc781c46bfeac13cc0
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:b2be693c71cfc4ce934d5b01efb863dc6e391f93a6b15fe7eeb15914063a8294
        - registry.redhat.io/openshift-serverless-1-tech-preview/backstage-plugins-eventmesh-rhel8@sha256:28ae309e4cf7feabaf78ac6067e1173f9f5b33a71c06d6c27cf25e01a0b1ed64
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:05e3dd6255294b691783523f7722cfaf13aca1a1e1dab344a25eeea72aaf9f9e
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:e9b38f1df225238836aa829f3a2016372859fb8709cdd2b30496498669a90be1
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:510e98837a549f489b2d74939f3b9c7ed90a8260ba914910f681d5c0dec7655c
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:d8ae7bc888f11121fbbba65074389b879bc22453bbabc551d79d0cee68fc33a3
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:5c7f29554d04eb9348932953344f2ea0806d14e7704355619381bf395d993f58
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:76a47d100de1f2afa23803bec503d9440605db03c7179b231f963e5663a3d12a
        - registry.redhat.io/ubi8/python-39@sha256:ac12e4706e0fdc464e791a3f274474ac1bb14d5af8f217c440fbb0c76d6a5d00
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:ace72435e3e63ab21dff2345f958cf6bd171537ef2846c5d5d91cf821877140d
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:9b94875e0e5d7abdb3b9ba7b04e217a2a9b7b4410d38d1608c86d1c21f929cc1
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:56c6ac59a68976741d2ba676a53e59fc494d04dc44b513d72fdf8a9ad894a54d
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:62bf4bee0ea17ecc60366ecc71431adc79d6f0a5b9f59489913b474128c4adb0
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ccbda564735555785815f27f8f25ad9c6c27dabe4e262586fc51609331692527
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:a9c84ac7e7c3fb1e5dedaf949d54a3aee8a62397d2bd6022094671de00b651da
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:2bd241898bd54dcb248ae86470430042dc545d45e7519ed470a1a9c8c63bdc37
        version: 1.33.3
      entries:
      - name: serverless-operator.v1.33.3
        version: 1.33.3
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.33
    - currentCSV: serverless-operator.v1.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-11-20T08:53:53Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.34.0 <1.34.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:a4b3d75903de1929326041e9885213074473f07fda91b299548d45680c295c00
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi8/python-39@sha256:ac12e4706e0fdc464e791a3f274474ac1bb14d5af8f217c440fbb0c76d6a5d00
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:633ea30ea3b6d5152bbdb61acb4271e5905d50fbe3c731fb4c8d8640a57f1970
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:586fc8ed88f91298dbff31bc2fc4c63e0d6d9356e3358b2f853f49511c3ff9bf
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:d8b94bbce98325fe8079c11c0b7a6c59bad8a7edd0532a7d27ef1632ff0de708
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:ef7f34a025e40f93fbd51f3a4596998a6dc508db824f137ffa1fbaf90e88f4a8
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:2adb0124f04d6c4a552ade79dd1499c8ca997edb45855bf438ca8043b2709e19
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:f94f0e8d21a6b6341b1db00513a1a2ed5cd84fcd898a376ecde942e940a518e7
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:01ea7a06332d60ac16ae0a27adcaaede3f680a98a22ff995259a6c42ac0e358d
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:b07010595201d7cfe0ec4f18e612d323eb38464c74eb187f275617a744d9df12
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:144883b09d8a999fd4d2a109d074e44d7d8b959987512379aa71c71e4060ca28
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:a9e8583c19ca3e01593f8c0e54731cb73df9b46b2ed47a2b1c711828c2295f82
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:76a47d100de1f2afa23803bec503d9440605db03c7179b231f963e5663a3d12a
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:6d042c614431aa9a9e77cf8ba0064423e903f7c418e367ecc23bc02a24a796c0
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:86ffb8be1762782d643c147cefcfbc2b4723b0f97d017314a99bd2275d05a1cf
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:30cb41f09067959cfe3792e070fca62820fcca4035683e5c43dd46d1cce97c01
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:45bc50a993d04b775a153e985a43d587d5455883fc4c6f31b772408297cf9040
        - registry.redhat.io/openshift-serverless-1-tech-preview/backstage-plugins-eventmesh-rhel8@sha256:09dd94090eea9cac7a1935a6d671cb97fd9fd5714c715c9d2aed461e468ca54a
        - registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:a4b3d75903de1929326041e9885213074473f07fda91b299548d45680c295c00
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:946daeff59b41a6f8c4842f8c2668f97ff1b49e2a8afae5d64ba3baa916b1d95
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:75228e70fb1383bbdbcc64d0cf24bf5d977c09112e3f32316824188264c3a2bc
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:33d62d13e16f47db9951ac826671f39fdc048fe7c3fb019eeb4bf9bd0f814222
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:d5ad3c9db7ce8e4fc950261f7c388476214a03d92e0043c7bbc4f0add2b1e69e
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:b9d85323e110d8eb6600223c38f9d4cbe89b50d59e03bf1e6b8808f7cef9a3a5
        - registry.redhat.io/ubi8/nodejs-20-minimal@sha256:acee3b00ccf8ff49bcebcf10e028e34a84d45f28f855363bafb3b93c75e84f11
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:512c4d93dc13cc53d923a6990770a7527d6fb657278e2c5d8f0ece3cd5020155
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:0f126893074cd0398654c4018df6e882b7610e148c82a41ac814f4bf3830bf9d
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:a5950f81cbc64d1edb64ae7f7f120c3435d36afcd14ffc34ae522af297d87e1e
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:3d17f12cc9da578549254ba1b8934b5afc66e453b1d5d667bb90281e45b8795c
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:c0376504edae12e7af6b6cce02ac53b6b374f5058ebefcbdbd32628b551bb0c4
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:1c82d20ccf5eb5f8e7725839f1060f4efdfd1c1b4773a99bd905e690a6e36f38
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:1252960c0c0a4dc7b39090cde0fd5742a19453b1a65a08b737bc41d044a0882a
        - registry.redhat.io/ubi8/openjdk-21@sha256:095175ef8573de8055e67bd5c8a21098448ac73db4e0d6ccd5269dd36b4d263c
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:d40daf67b36387faa9f55e70e50d10d02accfd19470c0ac11b78bf969947fed1
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:84994eabeb7342a0771c7cb1bdd70e7ee3928dd45874eafa64cde5bb385fe964
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:0f8d006d0e7468783739942a1c82c7aec376c9948f23841132c8ddece8ecba06
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:b801ed7b1cb32d4a1522a959e97f280f2eff6fc92e47fea2ea022f180d37831a
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:e520d1b6b87f85dfffab5e53cfb70562ab7edf450061c07609da39e3b2498fdd
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:4b6f161c14c68536ec2c7679d30681afa9a1fa2d512548f0f591791a10a8b894
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:76480201188dd142206040af845bd55b44155af92e0c5ee9bcdb8660bea0cdb8
        version: 1.34.1
      entries:
      - name: serverless-operator.v1.34.1
        version: 1.34.1
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.34
    - currentCSV: serverless-operator.v1.35.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.35.0 <1.35.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:f23d3ab058667c6fe9ee44dc249aa7e8449c1bb9581380ddbc66624511b286d2
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:b2de6b658a1595084a202203c90704c3f9de018a7a904ca276112ad606e374f1
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:0751b9ea7ca197480c4cbca17654d0653d1c0c760ca3c818c2a1018dc952cdc8
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:da8d975527ba100313b92e87d6d2690fddae9988aa8d82d728221295955b7224
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:ab8f85bb18311b0bfee00561ef6b5d9faf810358c954ac2c6f39cf2a7739e07c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:6b842b4eddfc71a2c8f6c1fcebb8f23a05f50c95be1ba2d20a1be660d3014c8e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:e743dd93b7ae855825716403979db3f34cd300caad130e7d3011172b1aa4eaf7
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:0e58edf5ee340fdfb282faeb6c9bee8cd6323e7f29a9b282dbb91c2622131a9c
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:477b1ac0db0abf1e16ae31bb22fdb73dd98f4dbd243fa0a7e6e94852b16a813c
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:b66bb9531f05cd46360f787955c849208496e3e45541cb50b23bfb6b68570aa8
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:9b176477be4c782ce60d81a72bacb566069a499579ce6df5a987a0841f05ad5c
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:86386b54cc10743023884193b7d6c7159b90052bc5e4a0608a6cdfd47865c966
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:cb2ebe98481808e788ca70e7f0a9b3402818b622158f327b99a1e87c23226d2d
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:d0650e4d7e72241e07a3796f20931adba231fbf5e401a45d794761f2c5afae48
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:2cbe84afebac23b5397d144b4d5660af33290df451141fa3625607b6033983e0
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:020264063c50f2a4590f3d7524b271f0f77350a1f8fa3c35ca6e28f0162aa9a5
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:d00a0059f6134b2519f9b1e034b81a943cae7ce391109703f24db21e56ee923d
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:d68fae6fc07efa9da9af0532aadb49f897a15afcdf9a98dc7517ff5e17735454
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:3130a5159adb9b4eaab75a0209d2ac639156d5e9f2346a8b2e6e0e22fad4b45f
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:80a155d2637f47e249fd69760b250c9b7a1eeb6781de9308b54162bb3e1bce71
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:ffaad2214fa897e89e92e574eb5cd1c671edab7f6eab87cf2ffd4ebf162b3cf4
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:b5db9a51bbb46640b23712c61e9c5a1c7bbdbcd9a53b0b92fa36b80dcaffffc0
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:f23d3ab058667c6fe9ee44dc249aa7e8449c1bb9581380ddbc66624511b286d2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:56e7a0158a713c70765ead847732db4e5f8439aacf31e9a4a22672b583028064
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:e7aa54986c3d388c1bb00828c9ccb294ca0b6c26376c7593ed6bfddda0ef4e5d
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:ab0342b325ee67cc2a7d321aa1eda9c6797dc91fe2b6f26e3cbab53fadf564c7
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:732827652b7a7a49181e86d561831001c28b72b7e12d5e3c1d71b2e496003229
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:018459cbba2e3169bc4be6564cf3747c9cd0677445912abf237f8fe6d58e46ac
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:6f4fa254d923f3260ab0f7d6a0b459554f83b1abfcb575a5ce2c0135fba32427
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:6722aecbbbefa80dfaf93143e033c562cea0982514e2eea27654dd4638e1d072
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:2980eca437e4c1cd22a2b6c2ccae9c4d6b55921b3d575fbd4b9241f1f0a8a3a4
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:dfcd11686f161d44786a205d12f9d628d9e847d40ffdb628ec9aa4a0da0b86f0
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:acb28d84cf03b8551efa32646864005dc3bc9bb341a696544892dfd0c87fbda6
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:818dc7bd69b453ba5882b4ccb82841f2d48c7552368417bac20bef1df64f52ff
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:4ca1bfaf2d7e9477583cddfa40f0b2ce59c20ff9ab20022038774dc3cc49662b
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:f56d21089566330ecbd9aa8acf14da332d3db7360aaf6da00c9a19df83f7334c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:8a4e5e0934197c32e0dfa1efc922d750940587fb61e0df49d476a2d035c53413
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:86074674b7926ad9bfc95556e50820792fcbc5e11ab02879378625ffabdd1889
        version: 1.35.1
      entries:
      - name: serverless-operator.v1.35.1
        version: 1.35.1
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.35
    - currentCSV: serverless-operator.v1.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.36.0 <1.36.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-ddb-streams-source-rhel8@sha256:a903480159ebf1d8652eeaa702d9885fb76bb106c02d75cec5c82c31d77d6410
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-timer-source-rhel8@sha256:8fed5ef9a1f39a55077bb5026c8854d0ee7fc8d8ff5da277ce9c1bfc2ab2474b
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:77e4689b2ce7a713e905553e0e42b8e128045f40cb6c6d4ee082f126cbc46b00
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:3d780ad1ab280b5092b90e277485d589da2a18e8b08090af7d95aecd17af545d
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:429d5d9568cca31fbe83c663276a300286af4cdbf3d35685afdbe5f6f2f8c6e0
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:4344716eb2940c6a3fa9a278250f936602bb1549fece6037912837719c178b24
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:f2d112ba9fdfccea988589b99b463c8a7157475f7a3f03678e8999e65b678cc3
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:5352e4eb4e426a41521276decadf204b6d23b4762a298e4d3dc73969e3eddc68
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:187fb9d64b53c55a029ffce3c0f9e7c3304683fa8a0e832cfb7134e8a3d0c237
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:6210a946f373b9c36c9b31293debe6fed090008b8234c0ba17d37fa81646172e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-transform-jsonata-rhel8@sha256:3af4c48030263eb5769e5f36a6dd13fd390d8566efaab37532f39e1df1c95c67
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-sink-rhel8@sha256:d9b944e64391a996104afbe8a21f66316a948e9fb7b91734e861847285be59c7
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:631eb3a8dd9d83ceafa7d8a4da4b1e2d4f2740e6017c766828dcc7575cf1322f
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:a6f565d73375e827d830dca85838f4c4cb7f08b0c078acb0c0d567b25f0d256a
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:a47b6b5237a87ccd90e5a49a53d06485fd55c87eebf12d01c335b87fc3a8c494
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:ed60e889a677fb139f2eb4bd9ab4fda6ecc39fd050ca2b6cc1c19717e428535a
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:074ebfc4494581611647ba90049655a470ca43743610a27be77db66a8633bac6
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:dcfe5fbc57b9f8520adbff368c8bf29eb491c9c26d1a1737a1722cffea83eb1b
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:f3d87a7fa5b5f2818effcac3304bd56eddcdb0db97a18ac0f35a06fc7e223e7d
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:4274004b6289fa9bea3fa6e56d64077a328a6b5e63a9f0e9c41f6d561f682bd7
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-source-rhel8@sha256:27e74287774cda208f667bc2a386debe6b6255e8991abfb37c959c8525f2af53
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:3b04e9d8cd4b5e43a07ece838d10478723a6c6529350f010550b4a0d1d0950d1
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:ee6533b71cf280877c9d7489dfd821c9209c6802bed7fa3fdb4f8f00fbeacb09
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:6b7316cce3432d391adc6ac38c92ff0ee86dd7ae0101957cc65cd338cc69e352
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:c39cb88afaddc295d93a3774b8a20d001b82392df540fb502ee02316d82460a9
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:82e69a1890042f5636ac7bb645b1aaeaba0a4894851eccc1fa1c63296918e31a
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:52ba6b1252b6f3d41e2a0b3b880cb768010ab9dd925ac34596f6adaedcbbe07c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-sink-rhel8@sha256:afa634e029bcf79a972f49372498d44b45bcfbdb6eb28ecaf4f0b0b2a0a1824b
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:f4d86b196457cdeebdc380b6f1b707c10f8dee789c7779c90ee1332e511af8c4
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:259e5dab50d36cdbfc117a42325128c4e16f34c048ba64b36d4a89132ddb4c1a
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-log-sink-rhel8@sha256:30b8045ef5f925a36c58941386f1899bec125d005a08aeeeba5edd49b0859149
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:3fe326ad9a00b112fc888ffd345c8d5042f0a22cf32ea575931192c39ce1fece
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:e3ae1cc6012f19631c012a07b113fb015977d2ee4f9faab3207fbc2464c25aa7
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:bf82b37ab8fd29c7a7cd846008a16d14b4cfdc1c09df4f731d71be3957e90123
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:d6f338e08fa937c8a4ccfec692424de4acded565f7ca3971ea507dffeb351503
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:4aeb68a22f63300e54391ec6b4cda8c4e665d904b5dcdd3c40155040525742a0
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:596a7272d6de714d7dbc6e00e9e50a2872846c269547465252d953172ef0b009
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:0c2cdf917bbad1f5bca8a2a2ef12a96cfcd0c77c190cc35687eb57169b54ab6f
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:c7c2e738e0d9da1d2d1993f68938e72c4709b1f28a54fa92224ef6a0e456b0aa
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:a1834b620f875681388f0ae224dd0771773a8b61b1d8e5214a942f232854b87b
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:9b328fff0cc8545fb9bb1f24f11d78cfac3ee76dedc11fa4def4262d79a60e8c
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:78fee4c596f2848487ca704da3fa1074e3c9c99b3f96f66833aeba63836d7b93
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sns-sink-rhel8@sha256:51b5e18e8d4c1ca81d7c5ddfea6352a3887f757acfcc53bd2f8bd3b01337ed42
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-source-rhel8@sha256:88323da16360981f27b8e4686fbbb64053e4e0e3c79c65d695edb317b334e1e0
        version: 1.36.1
      entries:
      - name: serverless-operator.v1.36.1
        version: 1.36.1
      - name: serverless-operator.v1.36.0
        version: 1.36.0
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.36
    defaultChannel: stable
    packageName: serverless-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-pipes-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-pipes-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pipes.services.k8s.aws/v1alpha1",
                "kind": "Pipe",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/pipes-controller:1.1.1
          createdAt: "2025-09-19T21:15:33Z"
          description: AWS EventBridge Pipes controller is a service controller for
            managing Pipes resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pipe represents the state of an AWS pipes Pipe resource.
            displayName: Pipe
            kind: Pipe
            name: pipes.pipes.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EventBridge Pipes resources in AWS from within your Kubernetes cluster.

          **About Amazon EventBridge Pipes**

          Amazon EventBridge Pipes connects sources to targets. Pipes are intended for point-to-point integrations between supported [sources](https://docs.aws.amazon.com/eventbridge/latest/userguide/eb-pipes-event-source.html) and [targets](https://docs.aws.amazon.com/eventbridge/latest/userguide/eb-pipes-event-target.html), with support for advanced transformations and [enrichment](https://docs.aws.amazon.com/eventbridge/latest/userguide/pipes-enrichment.html). It reduces the need for specialized knowledge and integration code when developing event-driven architectures, fostering consistency across your company’s applications. To set up a pipe, you choose the source, add optional filtering, define optional enrichment, and choose the target for the event data.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EventBridge Pipes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pipes
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EventBridge Pipes Developer Resources
          url: https://aws.amazon.com/eventbridge/pipes/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: pipes maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/pipes-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-pipes-controller.v1.1.1
        version: 1.1.1
      - name: ack-pipes-controller.v1.1.0
        version: 1.1.0
      - name: ack-pipes-controller.v1.0.25
        version: 1.0.25
      - name: ack-pipes-controller.v1.0.24
        version: 1.0.24
      - name: ack-pipes-controller.v1.0.23
        version: 1.0.23
      - name: ack-pipes-controller.v1.0.22
        version: 1.0.22
      - name: ack-pipes-controller.v1.0.21
        version: 1.0.21
      - name: ack-pipes-controller.v1.0.20
        version: 1.0.20
      - name: ack-pipes-controller.v1.0.18
        version: 1.0.18
      - name: ack-pipes-controller.v1.0.17
        version: 1.0.17
      - name: ack-pipes-controller.v1.0.16
        version: 1.0.16
      - name: ack-pipes-controller.v1.0.15
        version: 1.0.15
      - name: ack-pipes-controller.v1.0.14
        version: 1.0.14
      - name: ack-pipes-controller.v1.0.13
        version: 1.0.13
      - name: ack-pipes-controller.v1.0.12
        version: 1.0.12
      - name: ack-pipes-controller.v1.0.11
        version: 1.0.11
      - name: ack-pipes-controller.v1.0.10
        version: 1.0.10
      - name: ack-pipes-controller.v1.0.9
        version: 1.0.9
      - name: ack-pipes-controller.v1.0.8
        version: 1.0.8
      - name: ack-pipes-controller.v1.0.6
        version: 1.0.6
      - name: ack-pipes-controller.v1.0.3
        version: 1.0.3
      - name: ack-pipes-controller.v1.0.2
        version: 1.0.2
      - name: ack-pipes-controller.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-pipes-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Sigstore Authors
      provider-url: https://sigstore.dev
    name: sigstore-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sigstore-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.sigstore.dev/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "name": "rekor"
                },
                "spec": {
                  "server": {
                    "ingress": {
                      "hostname": "rekor.example.com"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          containerImage: ghcr.io/sigstore/sigstore-helm-operator@sha256:d9c6eb162b97a1aa1ad3bf515db65d01a96e55b6f8bcd646864d3e64a3ee2b9e
          createdAt: "2021-09-01T21:20:28Z"
          description: Helm Charts for Sigstore Project.
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/sigstore/sigstore-helm-operator
          support: The Sigstore Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Rekor
            name: rekors.helm.sigstore.dev
            version: v1alpha1
        description: |-
          Helm Charts for the sigstore project

          ## Charts

          The following charts are available:

          * [Rekor](https://github.com/sigstore/helm-charts/tree/main/charts/rekor)
        displayName: sigstore
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        links:
        - name: Sigstore Helm Operator
          url: https://github.com/sabre1041/sigstore-helm-operator
        maintainers:
        - email: ablock@redhat.com
          name: Andrew Block
        maturity: alpha
        provider:
          name: The Sigstore Authors
          url: https://sigstore.dev
        relatedImages:
        - quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - ghcr.io/sigstore/sigstore-helm-operator@sha256:d9c6eb162b97a1aa1ad3bf515db65d01a96e55b6f8bcd646864d3e64a3ee2b9e
        version: 0.0.1
      entries:
      - name: sigstore-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: sigstore-helm-operator
    provider:
      name: The Sigstore Authors
      url: https://sigstore.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Loki SIG Operator
      provider-url: ""
    name: loki-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: docker.io/grafana/loki-operator:0.8.0
          createdAt: "2025-03-17T11:00:08Z"
          description: The Community Loki Operator provides Kubernetes native deployment
            and management of Loki and related logging components.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/loki/tree/main/operator
          support: Grafana Loki SIG Operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |-
          The Community Loki Operator provides Kubernetes native deployment and management of Loki and related logging components.
          The purpose of this project is to simplify and automate the configuration of a Loki based logging stack for Kubernetes clusters.

          ### Operator features

          The Loki Operator includes, but is not limited to, the following features:

          * Kubernetes Custom Resources: Use Kubernetes custom resources to deploy and manage Loki, Alerting rules, Recording rules, and related components.
          * Simplified Deployment Configuration: Configure the fundamentals of Loki like tenants, limits, replication factor and storage from a native Kubernetes resource.

          ### Feature Gates

          The Loki Operator Bundle provides a set of feature gates that enable/disable specific feature depending on the target Kubernetes distribution. The following feature gates are enabled by default:
          * `serviceMonitors`: Enables creating a Prometheus-Operator managed ServiceMonitor resource per LokiStack component.
          * `serviceMonitorTlsEndpoints`: Enables TLS for the ServiceMonitor endpoints.
          * `lokiStackAlerts`: Enables creating PrometheusRules for common Loki alerts.
          * `httpEncryption`: Enables TLS encryption for all HTTP LokiStack services.
          * `grpcEncryption`: Enables TLS encryption for all GRPC LokiStack services.
          * `builtInCertManagement`: Enables the built-in facility for generating and rotating TLS client and serving certificates for all LokiStack services and internal clients
          * `lokiStackGateway`: Enables reconciling the reverse-proxy lokistack-gateway component for multi-tenant authentication/authorization traffic control to Loki.
          * `runtimeSeccompProfile`: Enables the restricted seccomp profile on all Lokistack components.
          * `defaultNodeAffinity`: Enable the operator will set a default node affinity on all pods. This will limit scheduling of the pods to Nodes with Linux.
          * `lokiStackWebhook`: Enables the LokiStack CR validation and conversion webhooks.
          * `alertingRuleWebhook`: Enables the AlertingRule CR validation webhook.
          * `recordingRuleWebhook`: Enables the RecordingRule CR validation webhook.
          * `rulerConfigWebhook`: Enables the RulerConfig CR validation webhook.

          In addition it enables the following OpenShift-only related feature gates:
          * `servingCertsService`: Enables OpenShift ServiceCA annotations on the lokistack-gateway service only.
          * `ruleExtendedValidation`: Enables extended validation of AlertingRule and RecordingRule to enforce tenancy in an OpenShift context.
          * `clusterTLSPolicy`: Enables usage of TLS policies set in the API Server.
          * `clusterProxy`: Enables usage of the proxy variables set in the proxy resource.

          ### Before you start

          1. Ensure that the appropriate object storage solution, that will be used by Loki, is avaliable and configured.
        displayName: Community Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Documentation
          url: https://loki-operator.dev/
        maintainers:
        - email: loki-operator-team@googlegroups.com
          name: Grafana Loki SIG Operator
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Grafana Loki SIG Operator
        relatedImages:
        - docker.io/grafana/loki-operator:0.8.0
        - quay.io/openshift/origin-kube-rbac-proxy:latest
        - quay.io/observatorium/api:latest
        - quay.io/observatorium/opa-openshift:latest
        - docker.io/grafana/loki:3.4.2
        version: 0.8.0
      entries:
      - name: loki-operator.v0.8.0
        version: 0.8.0
      - name: loki-operator.v0.7.1
        version: 0.7.1
      - name: loki-operator.v0.7.0
        version: 0.7.0
      - name: loki-operator.v0.6.2
        version: 0.6.2
      - name: loki-operator.v0.6.1
        version: 0.6.1
      - name: loki-operator.v0.6.0
        version: 0.6.0
      - name: loki-operator.v0.5.0
        version: 0.5.0
      - name: loki-operator.v0.4.0
        version: 0.4.0
      - name: loki-operator.v0.3.0
        version: 0.3.0
      - name: loki-operator.v0.2.0
        version: 0.2.0
      - name: loki-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: loki-operator
    provider:
      name: Grafana Loki SIG Operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-license-service-reporter-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-license-service-reporter-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "IBMLicenseServiceReporter",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-license-service-reporter-operator",
                    "app.kubernetes.io/instance": "ibmlicenseservicereporter-instance",
                    "app.kubernetes.io/managed-by": "operator",
                    "app.kubernetes.io/name": "ibmlicenseservicereporter",
                    "app.kubernetes.io/part-of": "ibm-license-service-reporter-operator"
                  },
                  "name": "instance"
                },
                "spec": {
                  "authentication": {
                    "useradmin": {
                      "enabled": true
                    }
                  },
                  "license": {
                    "accept": false
                  },
                  "version": "4.2.14"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:277e2ab849907553e4992c2dccc11f03ae1e0302f20aae48c620968aa97d6afc
          createdAt: "2025-02-26T12:10:52Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <4.2.14'
          operatorframework.io/suggested-namespace: ibm-ls-reporter
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "IBM License Service Reporter is a singleton which may be
              used in a single or multi-cluster environment. It aggregates data pushed
              from IBM License Services, deployed on clusters and from ILMT. \n Documentation:
              https://ibm.biz/lsvc-rprtr. \n License: Please refer to the IBM Terms
              website (ibm.biz/lsvc-lic) to find the license terms for the particular
              IBM product for which you are deploying this component. \n IBM License
              Service Reporter is a free, optionally installed add-on – one of services
              of Cloud Pak Foundational Services. Thanks to the IBM License Service
              Reporter customer can: \n - see IBM software deployments on a single
              dashboard, verify and maintain license compliance and avoid audit infractions,
              \n - see historical data on IBM software deployments to support making
              informed decisions for future purchases, \n - see details of software
              deployments, including source (ILMT for VMs, License Service for containers),
              cluster and quantities of licenses deployed, \n - identify which software
              deployments are VMs vs. containers and can use this information to evaluate
              workloads for modernization."
            displayName: IBMLicense Service Reporter
            kind: IBMLicenseServiceReporter
            name: ibmlicenseservicereporters.operator.ibm.com
            version: v1alpha1
        description: |-
          IBM License Service Reporter is a free, optionally installed add-on – one of services of Cloud Pak Foundational Services. Thanks to the IBM License Service Reporter customer can:
           * see IBM software deployments on a single dashboard, verify and maintain license compliance and avoid audit infractions,
           * see historical data on IBM software deployments to support making informed decisions for future purchases,
           * see details of software deployments, including source (ILMT for VMs, License Service for containers), cluster and quantities of licenses deployed,
           * identify which software deployments are VMs vs. containers and can use this information to evaluate workloads for modernization.

          IBM License Service Reporter is a singleton which may be used in a single or multi-cluster environment. It aggregates data pushed from IBM License Services, deployed on clusters and from ILMT.

          **Supported platforms**

          Red Hat OpenShift Container Platform 4.10 or newer installed on Linux x86_64.

          **Prerequisites**

          Prerequisites depend on the integration of the License Service Reporter with an IBM Cloud Pak. For more information, see the applicable IBM Cloud Pak documentation.
        displayName: IBM License Service Reporter
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - licensing
        - usage
        - IBMLicenseServiceReporter
        - IBM
        - Cloud
        - License
        - License Service
        - Service
        - Reporter
        - License Service Reporter
        - License Reporter
        - Service Reporter
        maintainers:
        - email: talk2sam@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-ui@sha256:213889e3c4c1d7bb10c88926ac8276d7e78c558397993d098e36f6c5da617de1
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-oauth2-proxy@sha256:c403c24857e29ddc306ba5d9bc28c3d556b181c5265fee16879fd2b4d946f899
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator-bundle@sha256:648f92bacd52fb51655925e34da1d0bdd0da720da8489b79064e73a47dcdd54a
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:277e2ab849907553e4992c2dccc11f03ae1e0302f20aae48c620968aa97d6afc
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-postgresql@sha256:b6e11080774863f56dd16d5748d03c3f89f6f4872f662ee218d6bf600491a35f
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter@sha256:87e62695d733a60bfc2084cd751810df6801a5fbe1ab87b0a421f6814f1b887a
        version: 4.2.14
      entries:
      - name: ibm-license-service-reporter-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-license-service-reporter-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: orchestrator-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: orchestrator-operator.v1.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Orchestrator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "orchestrator-sample"
                  },
                  "name": "orchestrator-sample"
                },
                "spec": {
                  "argocd": {
                    "enabled": false,
                    "namespace": "orchestrator-gitops"
                  },
                  "platform": {
                    "eventing": {
                      "broker": {}
                    },
                    "monitoring": {
                      "enabled": false
                    },
                    "namespace": "sonataflow-infra",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "64Mi"
                      }
                    }
                  },
                  "postgres": {
                    "authSecret": {
                      "name": "sonataflow-psql-postgresql",
                      "passwordKey": "postgres-password",
                      "userKey": "postgres-username"
                    },
                    "database": "sonataflow",
                    "name": "sonataflow-psql-postgresql",
                    "namespace": "sonataflow-infra"
                  },
                  "rhdh": {
                    "devMode": true,
                    "installOperator": true,
                    "name": "my-rhdh",
                    "namespace": "rhdh",
                    "plugins": {
                      "notificationsEmail": {
                        "enabled": false,
                        "port": 587,
                        "replyTo": "",
                        "sender": ""
                      }
                    }
                  },
                  "serverless": {
                    "installOperator": true
                  },
                  "serverlessLogic": {
                    "installOperator": true
                  },
                  "tekton": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-08-15T00:51:04Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/rhdhorchestrator/orchestrator-go-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Orchestrator
            name: orchestrators.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub Orchestrator is a plugin that enables serverless asynchronous workflows to Backstage.

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://rhdhorchestrator.io)  displayName: Orchestrator Operator
        displayName: Orchestrator Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Orchestrator
        - RHDH
        links:
        - name: Orchestrator Operator
          url: https://github.com/rhdhorchestrator/orchestrator-go-operator
        maintainers:
        - email: jubah@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/rhdh-orchestrator-dev-preview-beta/controller-rhel9-operator@sha256:9f5423829ac55faf2395e2d0d3683b238339dc03285696e1b350c5c5a1c4df3a
        version: 1.6.1
      entries:
      - name: orchestrator-operator.v1.6.1
        version: 1.6.1
      - name: orchestrator-operator.v1.6.0
        version: 1.6.0
      - name: orchestrator-operator.v1.5.2
        version: 1.5.2
      - name: orchestrator-operator.v1.5.1
        version: 1.5.1
      - name: orchestrator-operator.v1.5.0
        version: 1.5.0
      - name: orchestrator-operator.v1.4.0
        version: 1.4.0
      - name: orchestrator-operator.v1.3.1
        version: 1.3.1
      - name: orchestrator-operator.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: orchestrator-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtv-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:d0158c37686bf378347dea5368c0c7838eec9674fbb0ff11f369e6fd99753652
          createdAt: "2024-09-23T16:42:11Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.7.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.5 is supported on OpenShift 4.12 and 4.13

          Migration Toolkit for Virtualization 2.6 is supported on OpenShift 4.14 and 4.15

          Migration Toolkit for Virtualization 2.7 is supported on OpenShift 4.15 and 4.16

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:d0158c37686bf378347dea5368c0c7838eec9674fbb0ff11f369e6fd99753652
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:f8b8d2ce23c1be9dd74d18c7a90324288213e91630da529ded561fbfe5d59a5d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:fffa2fd9f8c804d021264748a910ceb17cee388b4d55cc763a8598234db87f15
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:31e20712c82ae77807eb5dfba5e8616c045280ff164027443214ab8f259d964a
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:ac13419a8c9993bbe656563f478095526dedd34a93d6463979d26852d08ed217
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:19909db45872b2d665eac31f20dac2f0fa3bd9b9a343f2be839747106de6bc5b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:90742a924a1782dced8ae639e68aef93df3c14e4a7b65e2e6b1b73e22860e7a7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:3a07a1e7c74783b0de3c190717f17091d4c9d573776419b46c2ef0c55328d769
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:045cfaec032648cb1a734398fd625de7ac1cc25bfd621b7558e641da82316b58
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:8ce4bba6a69de762f221792c03f6a759afcd16dc576de4b9b1f341d660cb82b0
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:1b06a93c3458b92d40d8eb437043c9abbb9ec5290fe85b2f6ef12297bea0c838
        version: 2.7.0
      entries:
      - name: mtv-operator.v2.7.0
        version: 2.7.0
      name: candidate
    - currentCSV: mtv-operator.2.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui": "true",
                  "feature_validation": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "plan": {
                      "name": "example-plan",
                      "namespace": "openshift-mtv"
                   }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {}
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "type": "vmware",
                   "secret": {
                      "name": "",
                      "namespace": ""
                   },
                   "url": ""
                }
              },
              {
                "kind": "Host",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                     "namespace": "",
                     "name": ""
                   },
                   "id": "",
                   "ipAddress": ""
                }
              },
              {
                "kind": "NetworkMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-networkmap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "type": "pod",
                            "namespace": "",
                            "name": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "StorageMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-storagemap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "storageClass": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "Hook",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "kind": "Provisioner",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "cinder",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "name": "kubernetes.io/cinder",
                  "volumeModes": [
                    {
                      "name": "Block",
                      "priority": 0,
                      "accessModes": [
                        {
                          "name": "ReadWriteOnce",
                          "priority": 0
                        },
                        {
                          "name": "ReadWriteMany",
                          "priority": 1
                        }
                      ]
                    },
                    {
                      "name": "Filesystem",
                      "priority": 1,
                      "accessModes": [
                        {
                          "name": "ReadWriteOnce",
                          "priority": 0
                        },
                        {
                          "name": "ReadWriteMany",
                          "priority": 1
                        }
                      ]
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:74a740878707470ea79c7bf03a7f045c2ea77d924d0890042ba3554dc5613a33
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.1.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui": "true",
                "feature_validation": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          repository: https://github.com/konveyor/forklift-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: Storage Provisioner Capabilities
            displayName: Provisioner
            kind: Provisioner
            name: provisioners.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator fully manages the deployment and life cycle of Migration Toolkit for Virtualization on OpenShift.

          If you experience any issues or have feature requests, please [file an issue in our Github repository](https://github.com/konveyor/forklift-operator/issues)
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        links:
        - name: Konveyor Forklift Operator
          url: https://github.com/konveyor/forklift-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:584a696270554f4fb64755cdc5b525bc6852925e6780f776fb309b1bd8d81d0f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ui-rhel8@sha256:8089b0889f8956f2abf5a69f4ab8f4382919bc15a48663b98551fb0b865d8af7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel8@sha256:aead9b494562985e20fa466888f28119ab302d64e93d7f23bc7d8da1c60267e7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel8@sha256:8c6758c2d81f807c3875f41cec14d9e39f1fe7710584794de57a1449510110db
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:74a740878707470ea79c7bf03a7f045c2ea77d924d0890042ba3554dc5613a33
        version: 2.1.0
      entries:
      - name: mtv-operator.2.1.0
        version: 2.1.0
      name: release-v2.1.0
    - currentCSV: mtv-operator.v2.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui": "true",
                  "feature_validation": "true",
                  "feature_must_gather_api": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "plan": {
                      "name": "example-plan",
                      "namespace": "openshift-mtv"
                   }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {}
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "type": "vmware",
                   "secret": {
                      "name": "",
                      "namespace": ""
                   },
                   "url": ""
                }
              },
              {
                "kind": "Host",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                     "namespace": "",
                     "name": ""
                   },
                   "id": "",
                   "ipAddress": ""
                }
              },
              {
                "kind": "NetworkMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-networkmap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "type": "pod",
                            "namespace": "",
                            "name": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "StorageMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-storagemap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "storageClass": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "Hook",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "kind": "Provisioner",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "cinder",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "name": "kubernetes.io/cinder",
                  "volumeModes": [
                    {
                      "name": "Block",
                      "priority": 0,
                      "accessModes": [
                        {
                          "name": "ReadWriteOnce",
                          "priority": 0
                        },
                        {
                          "name": "ReadWriteMany",
                          "priority": 1
                        }
                      ]
                    },
                    {
                      "name": "Filesystem",
                      "priority": 1,
                      "accessModes": [
                        {
                          "name": "ReadWriteOnce",
                          "priority": 0
                        },
                        {
                          "name": "ReadWriteMany",
                          "priority": 1
                        }
                      ]
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:3d28138500d6b715e7d11983684db12ced28dfef9c97c619044f0b830534f4ed
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.2.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui": "true",
                "feature_validation": "true",
                "feature_must_gather_api": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          repository: https://github.com/konveyor/forklift-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: Storage Provisioner Capabilities
            displayName: Provisioner
            kind: Provisioner
            name: provisioners.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation
          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.
          * Must-gather-api, a service to generate targeted must-gather archives.

          ### Compatibility
          Migration Toolkit for Virtualization 2.0 is supported in OpenShift 4.7
          Migration Toolkit for Virtualization 2.1 is supported in OpenShift 4.8
          Migration Toolkit for Virtualization 2.2 is supported in OpenShift 4.9

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [Konveyor Community](https://github.com/konveyor/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/konveyor/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Konveyor Forklift Operator
          url: https://github.com/konveyor/forklift-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:3d28138500d6b715e7d11983684db12ced28dfef9c97c619044f0b830534f4ed
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel8@sha256:097a3c667e594060844cbe5d16adcc278d35402d1965ff3fca249c8f84a688dc
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:02ec1b0dbc12ba31fce85ee164c97560ba3a3184316324ab7c871989a319bf8d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-api-rhel8@sha256:d6bf2bb018c471f8f86f85c62ef52f4a99510b748ac8b6b7b177fc44fa53fb44
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ui-rhel8@sha256:697b20d6423ac25c0d2f41b2598f4909b6863e85265a4f52bc23b8ce77afc732
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel8@sha256:220d8f97173292dc8389bf4b1a350843757f01ea7861bbcebcf75749366b450b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel8@sha256:6f332be2991b026d5ce2b22e5bdef8d17da06c8b437fb3e5b26e8d35b75e3f95
        version: 2.2.0
      entries:
      - name: mtv-operator.v2.2.0
        version: 2.2.0
      name: release-v2.2.0
    - currentCSV: mtv-operator.v2.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui": "true",
                  "feature_validation": "true",
                  "feature_must_gather_api": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "plan": {
                      "name": "example-plan",
                      "namespace": "openshift-mtv"
                   }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {}
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "type": "vmware",
                   "secret": {
                      "name": "",
                      "namespace": ""
                   },
                   "url": ""
                }
              },
              {
                "kind": "Host",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                     "namespace": "",
                     "name": ""
                   },
                   "id": "",
                   "ipAddress": ""
                }
              },
              {
                "kind": "NetworkMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-networkmap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "type": "pod",
                            "namespace": "",
                            "name": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "StorageMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-storagemap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "storageClass": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "Hook",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:fc6a4503d00719a0d22736a78689b5828a8ebaf8c7895689a56df5cf7b43b97f
          createdAt: "2022-11-08T18:56:59Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.3.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui": "true",
                "feature_validation": "true",
                "feature_must_gather_api": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/forklift-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.
          * Must-gather-api, a service to generate targeted must-gather archives.

          ### Compatibility

          Migration Toolkit for Virtualization 2.2 is supported in OpenShift 4.9

          Migration Toolkit for Virtualization 2.3 is supported in OpenShift 4.10

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [Konveyor Community](https://github.com/konveyor/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/konveyor/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/konveyor/forklift-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ui-rhel8@sha256:e6c2337e610e0f634e57a200a92f492ebe1f111d9fccaa019f8756ba907b067b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel8@sha256:4423034aac33409721ce1b5344f6d5330c174c9b47fc6837b8de7b47018036d7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel8@sha256:872e2a803742071bf81b70d9a58aafd0ceb6ce1d80e651d876690820835854c0
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:fc6a4503d00719a0d22736a78689b5828a8ebaf8c7895689a56df5cf7b43b97f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel8@sha256:59ad5149ce8691eb7944fa291be75e5701c8e74160a0b9a74d098872be704a67
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:a0c201d3474996a08ce662b074ebb069e5082da36f3e7977cf54e3c4e1e54b1d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-api-rhel8@sha256:a7a2d6216ee87284a4f9e07d4c1973b9e001292c504c3b933515826e90073167
        version: 2.3.4
      entries:
      - name: mtv-operator.v2.3.4
        version: 2.3.4
      name: release-v2.3
    - currentCSV: mtv-operator.v2.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui": "true",
                  "feature_validation": "true",
                  "feature_must_gather_api": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "plan": {
                      "name": "example-plan",
                      "namespace": "openshift-mtv"
                   }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {}
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "type": "vmware",
                   "secret": {
                      "name": "",
                      "namespace": ""
                   },
                   "url": ""
                }
              },
              {
                "kind": "Host",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                     "namespace": "",
                     "name": ""
                   },
                   "id": "",
                   "ipAddress": ""
                }
              },
              {
                "kind": "NetworkMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-networkmap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "type": "pod",
                            "namespace": "",
                            "name": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "StorageMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-storagemap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "storageClass": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "Hook",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:fc6a4503d00719a0d22736a78689b5828a8ebaf8c7895689a56df5cf7b43b97f
          createdAt: "2022-11-08T18:56:59Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.3.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui": "true",
                "feature_validation": "true",
                "feature_must_gather_api": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/forklift-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.
          * Must-gather-api, a service to generate targeted must-gather archives.

          ### Compatibility

          Migration Toolkit for Virtualization 2.2 is supported in OpenShift 4.9

          Migration Toolkit for Virtualization 2.3 is supported in OpenShift 4.10

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [Konveyor Community](https://github.com/konveyor/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/konveyor/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/konveyor/forklift-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel8@sha256:4423034aac33409721ce1b5344f6d5330c174c9b47fc6837b8de7b47018036d7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel8@sha256:872e2a803742071bf81b70d9a58aafd0ceb6ce1d80e651d876690820835854c0
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel8@sha256:59ad5149ce8691eb7944fa291be75e5701c8e74160a0b9a74d098872be704a67
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:a0c201d3474996a08ce662b074ebb069e5082da36f3e7977cf54e3c4e1e54b1d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-api-rhel8@sha256:a7a2d6216ee87284a4f9e07d4c1973b9e001292c504c3b933515826e90073167
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ui-rhel8@sha256:e6c2337e610e0f634e57a200a92f492ebe1f111d9fccaa019f8756ba907b067b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:fc6a4503d00719a0d22736a78689b5828a8ebaf8c7895689a56df5cf7b43b97f
        version: 2.3.4
      entries:
      - name: mtv-operator.v2.3.4
        version: 2.3.4
      name: release-v2.3.0
    - currentCSV: mtv-operator.v2.4.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_must_gather_api": "true",
                  "feature_ui": "false",
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:f17fb7a2a840bda51551683e18d01ad1fae3ea73c293de1711f382ed309b8668
          createdAt: "2023-10-23T12:16:09Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.4.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui": "false",
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_must_gather_api": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.
          * Must-gather-api, a service to generate targeted must-gather archives.

          ### Compatibility

          Migration Toolkit for Virtualization 2.2 is supported in OpenShift 4.9

          Migration Toolkit for Virtualization 2.3 is supported in OpenShift 4.10 and 4.11

          Migration Toolkit for Virtualization 2.4 is supported in OpenShift 4.11 and 4.12

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-api-rhel8@sha256:cf22690c6bb420b9acc09ee5f90329ece6a35c18e5d3d6f6e40a320e2385ad85
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:7b4eb4de5dbc7e1af9c74d47fb6bba39f4973dc2d41531989c9fcf454b715b14
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:dc7ed113fae6ea3d5baaf165b08b5404a15f3a70f08089caca65b837623af4fb
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:8d3aeb8f6cfb17f5694f92f56a2361e08d220e2608183fb39ac94e92187b3fc7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-warm-rhel8@sha256:a6f91bc0ffc5e2f545a18b8888715281db7a751e1eff5f9eb89e8d319268a3ec
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:f17fb7a2a840bda51551683e18d01ad1fae3ea73c293de1711f382ed309b8668
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:8e7a7b3c2dc31b90e2102a82633f1eda95355496ff657964fedb26157f9ec229
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:ad9bacdfa4f10e020902d8e0381bdd84e48ce89ffab0f4144bd3c03fc91f7259
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:86716db780d3e43fa1e3029ab095ae6a557cf023797f4d79565a1aa5ebcb9b93
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:a7caa55c9d211a869b42e6f0101cd5099d64f5f3fd423f8a06091bad98d84aff
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ui-rhel8@sha256:54638e44269175e112a510fe2558f9a1dda4758bf3b3fb5e5b54c6617a5917bd
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:20a7b16f9f5d3dd58dbc7c061567590618a24bd595911ed089034095cc919d96
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:9aec00a487545eef76dcc4701680c5ec5bdc5e74302edbdfbc12c821b8ab4997
        version: 2.4.3
      entries:
      - name: mtv-operator.v2.4.3
        version: 2.4.3
      name: release-v2.4
    - currentCSV: mtv-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_must_gather_api": "true",
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:7dac245977fa0beeb7539b8ea5b3be68c457da132c979fd36405509fc879c7e1
          createdAt: "2023-08-31T15:03:42Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.5.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_must_gather_api": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.
          * Must-gather-api, a service to generate targeted must-gather archives.

          ### Compatibility

          Migration Toolkit for Virtualization 2.4 is supported on OpenShift 4.11 and 4.12

          Migration Toolkit for Virtualization 2.5 is supported on OpenShift 4.12 and 4.13

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:00fac23d8a514208c906ab2c911f188947368cb5e3076aacf3dc049e27b6c366
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:7dac245977fa0beeb7539b8ea5b3be68c457da132c979fd36405509fc879c7e1
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:8cc711f92258cad8f0747d3b98fef432d20661b90c4b67df8aa91650f2be3388
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-warm-rhel8@sha256:b5e0a88008c5c355844ed16b0889b0d73e63748fffafe57d0d1c52bc39c50f87
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:349cb43a620b2db9e724d874b231ef2ad896260009e3b0a93539121ac9cbe84b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:c157d6ebfb12f37a646f5f3bfbfdc90b7c8c669afe08ad342d34e474c39aa094
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:68cde15ed754189d74a528fc61cc2191861a27a4a9be112fb3ce1de3e836a77c
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-api-rhel8@sha256:c702dc948e42b9ba1e87b988f603be3df249a44dfdd6a11b0b9b49cb25f6de0c
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:55d773cad8fcf2919f5b0a5387dea62feb4ce676fb94eba310fc4d459f729d73
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:18591334cd99eb9e9f7e3d2609ff8666a9c0625e996d2f8751a14ebb9386457a
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:6097cc8e88905793bc61e48a62d77decffbbc59c1d4de83bb2eb5340ea1e8697
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:e2fca2e132bfc9a90a7c1f7e8e6b6d6c4fd8a63ec4177f5dea3f06c0c529e722
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:e665d8c5da798dc0dcd66b691db087de49564333e3c6e238d16a356c8c62d080
        version: 2.5.0
      entries:
      - name: mtv-operator.v2.5.0
        version: 2.5.0
      name: release-v2.5
    - currentCSV: mtv-operator.v2.6.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:1dfbb0993d3009608a6b272f178197f8a88edb047db4f4ff84aee8c023191d65
          createdAt: "2024-09-13T15:42:40Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.6.7'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.5 is supported on OpenShift 4.12 and 4.13

          Migration Toolkit for Virtualization 2.6 is supported on OpenShift 4.14 and 4.15

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:d34826b2a9370a1b6502d8e74a1254b2bd786684b25d90c2f0a960be0e99777d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-warm-rhel8@sha256:a46eb455027d248bcf591b612f34eb2ff4cd887cbbc3d116f78187b579ac31ab
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:3fef1c267f1f89d51ba7cd54bf39e17d7fe8ba5a4b6ecf9cdedcbe6d49d28bcc
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:1dfbb0993d3009608a6b272f178197f8a88edb047db4f4ff84aee8c023191d65
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:537f5fed16dcc46ee9cea05285a68c0e037c9b439289ff537590352865eced92
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:7dbf0f8d3d11cdb73033377552f999951331ed61d28c98cfd836903faf5601f9
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:a30fca5c5f260b94181c03a8c4c1cfe340048af87515e86571e9676e836d4b9f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:c98635ac256743873c4d74d63f4278cd26794dbaa1d56a82b15089dab510e905
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:a24acbfeca235270e75bd78046aa1dded9ff2a325a725cb46fc52ed3ff6880b8
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:403a5f8fd64926a9638342dbe568b557a7e3dc4b700cde79e64ecd9d948002a7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:6e0b0a7aca1a4afef62cdf511195e6a567feb644d3e5128c9899afc351807212
        version: 2.6.7
      entries:
      - name: mtv-operator.v2.6.7
        version: 2.6.7
      - name: mtv-operator.v2.6.6
        version: 2.6.6
      - name: mtv-operator.v2.6.5
        version: 2.6.5
      - name: mtv-operator.v2.6.4
        version: 2.6.4
      - name: mtv-operator.v2.6.3
        version: 2.6.3
      - name: mtv-operator.v2.6.2
        version: 2.6.2
      - name: mtv-operator.v2.6.1
        version: 2.6.1
      - name: mtv-operator.v2.6.0
        version: 2.6.0
      name: release-v2.6
    - currentCSV: mtv-operator.v2.7.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:db4cbc7214cdc7bf176ff16e562120111219ad8e663b7d4385dfd4c164253b7f
          createdAt: "2025-04-22T18:23:33Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.7.12'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.7 is supported on OpenShift 4.15, 4.16 and 4.17

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:a0360971c4cf448f99eb9ca79c9a153aec479e59e5985b91e019ee478382510b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:ed8cefd01f3b03df7faaf1ba0774cc18b7fc08b25f359d1e52c794b4ad60441e
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:33bae3e4c5b96df3319287d409d8d25db7db33704c7e9f8a58aec797bf20d12a
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:db4cbc7214cdc7bf176ff16e562120111219ad8e663b7d4385dfd4c164253b7f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:97ba4a4b479957e57eb296598d0c8797233f3fd0dbac20fb4686db02d3a8ed9f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:1622759bca392e2f7f66611df7988e7c63138807f917f8b34aefcc60467051ec
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:a32fba94e62f4c979e3b3596ab3a70750bd3bbf841e6004b5fd5305ac796027f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:51fdb9274911e93ae9dff9ab616a04a8a2c8fce4b6b21c05cf5f0145b1f93de0
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:a0b713409ad1d5018832cd50a2ffa3ce70ebf6cf1a4e532b3676b6f039316b52
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:ec2b67b3aff9b74f2246c310f1871198bbea8428b4e563d0e8c254a785b46a27
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:112a465da1695340286a7ff6e4dd66144209ce4389134f03813f786d30ba7508
        version: 2.7.12
      entries:
      - name: mtv-operator.v2.7.12
        version: 2.7.12
      - name: mtv-operator.v2.7.11
        version: 2.7.11
      - name: mtv-operator.v2.7.10
        version: 2.7.10
      - name: mtv-operator.v2.7.9
        version: 2.7.9
      - name: mtv-operator.v2.7.8
        version: 2.7.8
      - name: mtv-operator.v2.7.7
        version: 2.7.7
      - name: mtv-operator.v2.7.6
        version: 2.7.6
      - name: mtv-operator.v2.7.5
        version: 2.7.5
      - name: mtv-operator.v2.7.4
        version: 2.7.4
      - name: mtv-operator.v2.7.3
        version: 2.7.3
      - name: mtv-operator.v2.7.2
        version: 2.7.2
      - name: mtv-operator.v2.7.1
        version: 2.7.1
      - name: mtv-operator.v2.7.0
        version: 2.7.0
      name: release-v2.7
    - currentCSV: mtv-operator.v2.8.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "\u003cbase64-encoded-playbook\u003e"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "konveyor-forklift"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:e4145efb22dfc336d8b81f24f5b488db602d6d547479d615bf82b4fdce96c784
          createdAt: "2025-08-26T14:06:48Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: VSphere Xcopy Volume Populator
            displayName: VSphereXcopyVolumePopulator
            kind: VSphereXcopyVolumePopulator
            name: vspherexcopyvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.8 is supported on OpenShift 4.16, 4.17 and 4.18

          Migration Toolkit for Virtualization 2.9 is supported on OpenShift 4.17, 4.18 and 4.19

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/support/policy/updates/migration-toolkit-for-virtualization).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:e4145efb22dfc336d8b81f24f5b488db602d6d547479d615bf82b4fdce96c784
        version: 2.8.7
      entries:
      - name: mtv-operator.v2.8.7
        version: 2.8.7
      - name: mtv-operator.v2.8.6
        version: 2.8.6
      - name: mtv-operator.v2.8.5
        version: 2.8.5
      - name: mtv-operator.v2.8.4
        version: 2.8.4
      - name: mtv-operator.v2.8.3
        version: 2.8.3
      - name: mtv-operator.v2.8.2
        version: 2.8.2
      - name: mtv-operator.v2.8.1
        version: 2.8.1
      - name: mtv-operator.v2.8.0
        version: 2.8.0
      name: release-v2.8
    defaultChannel: release-v2.8
    packageName: mtv-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Labs
      provider-url: https://grafana.com
    name: grafana-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: grafana-operator.v5.19.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "Grafana",
                "metadata": {
                  "labels": {
                    "dashboards": "grafana-a",
                    "folders": "grafana-a"
                  },
                  "name": "grafana-a"
                },
                "spec": {
                  "config": {
                    "auth": {
                      "disable_login_form": "false"
                    },
                    "log": {
                      "mode": "console"
                    },
                    "security": {
                      "admin_password": "start",
                      "admin_user": "root"
                    }
                  }
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaAlertRuleGroup",
                "metadata": {
                  "name": "grafanaalertrulegroup-sample"
                },
                "spec": {
                  "folderRef": "test-folder-from-operator",
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "interval": "5m",
                  "rules": [
                    {
                      "condition": "B",
                      "data": [
                        {
                          "datasourceUid": "grafanacloud-demoinfra-prom",
                          "model": {
                            "datasource": {
                              "type": "prometheus",
                              "uid": "grafanacloud-demoinfra-prom"
                            },
                            "editorMode": "code",
                            "expr": "weather_temp_c{}",
                            "instant": true,
                            "intervalMs": 1000,
                            "legendFormat": "__auto",
                            "maxDataPoints": 43200,
                            "range": false,
                            "refId": "A"
                          },
                          "refId": "A",
                          "relativeTimeRange": {
                            "from": 600
                          }
                        },
                        {
                          "datasourceUid": "__expr__",
                          "model": {
                            "conditions": [
                              {
                                "evaluator": {
                                  "params": [
                                    0
                                  ],
                                  "type": "gt"
                                },
                                "operator": {
                                  "type": "and"
                                },
                                "query": {
                                  "params": [
                                    "C"
                                  ]
                                },
                                "reducer": {
                                  "params": [],
                                  "type": "last"
                                },
                                "type": "query"
                              }
                            ],
                            "datasource": {
                              "type": "__expr__",
                              "uid": "__expr__"
                            },
                            "expression": "A",
                            "intervalMs": 1000,
                            "maxDataPoints": 43200,
                            "refId": "B",
                            "type": "threshold"
                          },
                          "refId": "B",
                          "relativeTimeRange": {
                            "from": 600
                          }
                        }
                      ],
                      "execErrState": "Error",
                      "for": "5m0s",
                      "noDataState": "NoData",
                      "title": "Temperature below freezing",
                      "uid": "4843de5c-4f8a-4af0-9509-23526a04faf8"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaContactPoint",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "grafana-operator",
                    "app.kubernetes.io/instance": "grafanacontactpoint-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "grafanacontactpoint",
                    "app.kubernetes.io/part-of": "grafana-operator"
                  },
                  "name": "grafanacontactpoint-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "name": "grafanacontactpoint-sample",
                  "settings": {
                    "email": null
                  },
                  "type": "email"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaDashboard",
                "metadata": {
                  "name": "grafanadashboard-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "json": "{\n\n  \"id\": null,\n  \"title\": \"Simple Dashboard\",\n  \"tags\": [],\n  \"style\": \"dark\",\n  \"timezone\": \"browser\",\n  \"editable\": true,\n  \"hideControls\": false,\n  \"graphTooltip\": 1,\n  \"panels\": [],\n  \"time\": {\n    \"from\": \"now-6h\",\n    \"to\": \"now\"\n  },\n  \"timepicker\": {\n    \"time_options\": [],\n    \"refresh_intervals\": []\n  },\n  \"templating\": {\n    \"list\": []\n  },\n  \"annotations\": {\n    \"list\": []\n  },\n  \"refresh\": \"5s\",\n  \"schemaVersion\": 17,\n  \"version\": 0,\n  \"links\": []\n}\n"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaDatasource",
                "metadata": {
                  "name": "grafanadatasource-sample"
                },
                "spec": {
                  "datasource": {
                    "access": "proxy",
                    "isDefault": true,
                    "jsonData": {
                      "timeInterval": "5s",
                      "tlsSkipVerify": true
                    },
                    "name": "prometheus",
                    "type": "prometheus",
                    "url": "http://prometheus-service:9090"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "plugins": [
                    {
                      "name": "grafana-clock-panel",
                      "version": "1.3.0"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaFolder",
                "metadata": {
                  "name": "grafanafolder-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "title": "Example Folder"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaLibraryPanel",
                "metadata": {
                  "name": "grafana-library-panel-inline-envs"
                },
                "spec": {
                  "envs": [
                    {
                      "name": "CUSTOM_RANGE_ENV",
                      "value": "now - 12h"
                    }
                  ],
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "jsonnet": "local myRange = std.extVar('CUSTOM_RANGE_ENV'); {\n\n  model: {}\n}\n",
                  "plugins": [
                    {
                      "name": "grafana-piechart-panel",
                      "version": "1.3.9"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaMuteTiming",
                "metadata": {
                  "name": "mutetiming-sample"
                },
                "spec": {
                  "editable": false,
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "name": "mutetiming-sample",
                  "time_intervals": [
                    {
                      "days_of_month": [
                        "1",
                        "15"
                      ],
                      "location": "Asia/Shanghai",
                      "times": [
                        {
                          "end_time": "06:00",
                          "start_time": "00:00"
                        }
                      ],
                      "weekdays": [
                        "saturday"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationPolicy",
                "metadata": {
                  "name": "grafananotificationpolicy-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "route": {
                    "group_by": [
                      "grafana_folder",
                      "alertname"
                    ],
                    "receiver": "Grafana Cloud OnCall",
                    "routes": [
                      {
                        "object_matchers": [
                          [
                            "foo",
                            "=",
                            "bar"
                          ]
                        ],
                        "receiver": "grafana-default-email",
                        "routes": [
                          {
                            "object_matchers": [
                              [
                                "severity",
                                "=",
                                "critical"
                              ]
                            ],
                            "receiver": "Grafana Cloud OnCall"
                          }
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationPolicyRoute",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "grafana-operator",
                    "app.kubernetes.io/instance": "grafananotificationpolicyroute-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "grafananotificationpolicyroute",
                    "app.kubernetes.io/part-of": "grafana-operator"
                  },
                  "name": "grafananotificationpolicyroute-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationTemplate",
                "metadata": {
                  "name": "test"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "name": "test",
                  "template": "{{ define \"SlackAlert\" }}\n  [{{.Status}}] {{ .Labels.alertname }}\n  {{ .Annotations.AlertValues }}\n{{ end }}\n\n{{ define \"SlackAlertMessage\" }}\n  {{ if gt (len .Alerts.Firing) 0 }}\n    {{ len .Alerts.Firing }} firing:\n    {{ range .Alerts.Firing }} {{ template \"SlackAlert\" . }} {{ end }}\n  {{ end }}\n  {{ if gt (len .Alerts.Resolved) 0 }}\n    {{ len .Alerts.Resolved }} resolved:\n    {{ range .Alerts.Resolved }} {{ template \"SlackAlert\" . }} {{ end }}\n  {{ end }}\n{{ end }}\n\n{{ template \"SlackAlertMessage\" . }}\n"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: ghcr.io/grafana/grafana-operator@sha256:b678bc3faad11f4ef53c240e2a650ce939b9e8920dd72a47681b2efcadc7b837
          createdAt: "2025-08-21T08:38:00Z"
          description: Deploys and manages Grafana instances, dashboards and data
            sources
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/grafana/grafana-operator
          support: Grafana Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GrafanaAlertRuleGroup is the Schema for the grafanaalertrulegroups
              API
            displayName: Grafana Alert Rule Group
            kind: GrafanaAlertRuleGroup
            name: grafanaalertrulegroups.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaContactPoint is the Schema for the grafanacontactpoints
              API
            displayName: Grafana Contact Point
            kind: GrafanaContactPoint
            name: grafanacontactpoints.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaDashboard is the Schema for the grafanadashboards
              API
            displayName: Grafana Dashboard
            kind: GrafanaDashboard
            name: grafanadashboards.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaDatasource is the Schema for the grafanadatasources
              API
            displayName: Grafana Datasource
            kind: GrafanaDatasource
            name: grafanadatasources.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaFolder is the Schema for the grafanafolders API
            displayName: Grafana Folder
            kind: GrafanaFolder
            name: grafanafolders.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaLibraryPanel
            name: grafanalibrarypanels.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaMuteTiming
            name: grafanamutetimings.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaNotificationPolicy is the Schema for the GrafanaNotificationPolicy
              API
            displayName: Grafana Notification policy
            kind: GrafanaNotificationPolicy
            name: grafananotificationpolicies.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaNotificationPolicyRoute is the Schema for the grafananotificationpolicyroutes
              API
            displayName: Grafana Notification Policy Route
            kind: GrafanaNotificationPolicyRoute
            name: grafananotificationpolicyroutes.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaNotificationTemplate
            name: grafananotificationtemplates.grafana.integreatly.org
            version: v1beta1
          - description: Grafana is the Schema for the grafanas API
            displayName: Grafana
            kind: Grafana
            name: grafanas.grafana.integreatly.org
            version: v1beta1
        description: Deploys and manages Grafana instances, dashboards and data sources
        displayName: Grafana Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Grafana
        - Metrics
        - Observability
        links:
        - name: Grafana Operator
          url: https://grafana.github.io/grafana-operator
        maintainers:
        - email: grafana-operator@grafana.com
          name: Grafana-operator maintainers
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Grafana Labs
          url: https://grafana.com
        relatedImages:
        - ghcr.io/grafana/grafana-operator@sha256:b678bc3faad11f4ef53c240e2a650ce939b9e8920dd72a47681b2efcadc7b837
        - docker.io/grafana/grafana@sha256:6ac590e7cabc2fbe8d7b8fc1ce9c9f0582177b334e0df9c927ebd9670469440f
        version: 5.19.4
      entries:
      - name: grafana-operator.v5.19.4
        version: 5.19.4
      - name: grafana-operator.v5.18.0
        version: 5.18.0
      - name: grafana-operator.v5.17.1
        version: 5.17.1
      - name: grafana-operator.v5.17.0
        version: 5.17.0
      - name: grafana-operator.v5.16.0
        version: 5.16.0
      - name: grafana-operator.v5.15.1
        version: 5.15.1
      - name: grafana-operator.v5.14.0
        version: 5.14.0
      name: v5
    defaultChannel: v5
    packageName: grafana-operator
    provider:
      name: Grafana Labs
      url: https://grafana.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: PingCAP
      provider-url: ""
    name: tidb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tidb-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pingcap.com/v1alpha1",
                "kind": "TidbCluster",
                "metadata": {
                  "name": "basic"
                },
                "spec": {
                  "version": "v5.4.0",
                  "timezone": "UTC",
                  "pvReclaimPolicy": "Delete",
                  "discovery": {},
                  "pd": {
                    "baseImage": "pingcap/pd",
                    "replicas": 1,
                    "requests": {
                      "storage": "1Gi"
                    },
                    "config": {}
                  },
                  "tikv": {
                    "baseImage": "pingcap/tikv",
                    "replicas": 1,
                    "requests": {
                      "storage": "1Gi"
                    },
                    "config": {
                      "storage": {
                        "reserve-space": "0MB"
                      }
                    }
                  },
                  "tidb": {
                    "baseImage": "pingcap/tidb",
                    "replicas": 1,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "config": {}
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: pingcap/tidb-operator:v1.3.1
          createdAt: 2021-2-15T07-18-39Z
          description: TiDB Operator manages TiDB clusters on Kubernetes and automates
            tasks related to operating a TiDB cluster. It makes TiDB a truly cloud-native
            database.
          repository: https://github.com/pingcap/tidb-operator
          support: KanShiori
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup
            displayName: Backup
            kind: Backup
            name: backups.pingcap.com
            version: v1alpha1
          - description: Backup Schedule
            displayName: BackupSchedule
            kind: BackupSchedule
            name: backupschedules.pingcap.com
            version: v1alpha1
          - description: DM Cluster
            displayName: DMCluster
            kind: DMCluster
            name: dmclusters.pingcap.com
            version: v1alpha1
          - description: Restore
            displayName: Restore
            kind: Restore
            name: restores.pingcap.com
            version: v1alpha1
          - description: Tidb Cluster Auto Scaler
            displayName: TidbClusterAutoScaler
            kind: TidbClusterAutoScaler
            name: tidbclusterautoscalers.pingcap.com
            version: v1alpha1
          - description: Tidb Cluster
            displayName: TidbCluster
            kind: TidbCluster
            name: tidbclusters.pingcap.com
            version: v1alpha1
          - description: Tidb Initializer
            displayName: TidbInitializer
            kind: TidbInitializer
            name: tidbinitializers.pingcap.com
            version: v1alpha1
          - description: Tidb Monitor
            displayName: TidbMonitor
            kind: TidbMonitor
            name: tidbmonitors.pingcap.com
            version: v1alpha1
          - description: Tidb NG Monitoring
            displayName: TidbNGMonitoring
            kind: TidbNGMonitoring
            name: tidbngmonitorings.pingcap.com
            version: v1alpha1
        description: |
          ## About the managed application
          TiDB ("Ti" stands for Titanium) is an open-source, distributed, NewSQL database that supports Hybrid Transactional and Analytical Processing (HTAP) workloads. It is MySQL compatible and features horizontal scalability, strong consistency, and high availability. TiDB can be deployed on-premise or in-cloud.
          ## About this Operator

          TiDB Operator manages TiDB clusters on Kubernetes and automates tasks related to operating a TiDB cluster. It makes TiDB a truly cloud-native database.

          ## Features

          * Safely scaling the TiDB cluster

          TiDB Operator empowers TiDB with horizontal scalability on the cloud.

          * Rolling update of the TiDB cluster

          Gracefully perform rolling updates for the TiDB cluster in order, achieving zero-downtime of the TiDB cluster.

          * Multi-tenant support

          Users can deploy and manage multiple TiDB clusters on a single Kubernetes cluster easily.

          * Automatic failover

          TiDB Operator automatically performs failover for your TiDB cluster when node failures occur.

          * Kubernetes package manager support

          By embracing Kubernetes package manager Helm, users can easily deploy TiDB clusters with only one command.

          * Automatically monitoring TiDB cluster at creating

          Automatically deploy Prometheus, Grafana for TiDB cluster monitoring.

          ## Quick Start
          You can follow our [Get Started](https://docs.pingcap.com/tidb-in-kubernetes/stable/get-started) guide to quickly start a testing Kubernetes cluster and play with TiDB Operator on your own machine.

          ## Documentation

          You can see our documentation at PingCAP website for more in-depth installation and instructions for production:

          - [English](https://docs.pingcap.com/tidb-in-kubernetes/stable)
          - [简体中文](https://docs.pingcap.com/zh/tidb-in-kubernetes/stable)

          All the TiDB Operator documentation is maintained in the [docs-tidb-operator repository](https://github.com/pingcap/docs-tidb-operator).

          ## Community

          Feel free to reach out if you have any questions. The maintainers of this project are reachable via:

          - [TiDB Community Slack](https://pingcap.com/tidbslack/) in the [#sig-k8s](https://tidbcommunity.slack.com/archives/CHD0HA3LZ) channel
          - [Filing an issue](https://github.com/pingcap/tidb-operator/issue) against this repo

          Pull Requests are welcome! Check the [issue tracker](https://github.com/pingcap/tidb-operator/issue) for `status/help-wanted` issues if you're unsure where to start.

          If you're planning a new feature, please file an issue or join [#sig-k8s](https://tidbcommunity.slack.com/archives/CHD0HA3LZ) channel to discuss first.
          ## Prerequisites for enabling this Operator
        displayName: TiDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TiDB
        - SQL
        - Database
        links:
        - name: GitHub
          url: https://github.com/pingcap/tidb-operator
        - name: Documentation
          url: https://docs.pingcap.com/tidb-in-kubernetes/stable
        maintainers:
        - email: zhanghailong@pingcap.com
          name: Daniel Zhang
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: PingCAP
        relatedImages:
        - pingcap/tidb-operator:v1.3.1
        version: 1.3.1
      entries:
      - name: tidb-operator.v1.3.1
        version: 1.3.1
      name: stable
    defaultChannel: stable
    packageName: tidb-operator
    provider:
      name: PingCAP
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v0.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: gcr.io/datadoghq/operator:0.8.4
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          olm.skipRange: <0.8.4
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - gcr.io/datadoghq/operator:0.8.4
        version: 0.8.4
      entries:
      - name: datadog-operator.v0.8.4
        version: 0.8.4
      - name: datadog-operator.v0.8.2
        version: 0.8.2
      - name: datadog-operator.v0.8.1
        version: 0.8.1
      - name: datadog-operator.v0.8.0
        version: 0.8.0
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: gcr.io/datadoghq/operator:1.18.0
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - gcr.io/datadoghq/operator:1.18.0
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.2
        version: 1.4.2
      - name: datadog-operator.v1.4.1
        version: 1.4.1
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      - name: datadog-operator.v1.2.0
        version: 1.2.0
      - name: datadog-operator.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: datadog-operator
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: servicemeshoperator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: servicemeshoperator.v2.6.10
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"maistra.io/v2\",\n    \"kind\":
            \"ServiceMeshControlPlane\",\n    \"metadata\": {\n      \"name\": \"basic\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"version\": \"v2.6\",\n      \"policy\": {\n         \"type\":
            \"Istiod\"\n      },\n      \"telemetry\": {\n         \"type\": \"Istiod\"\n
            \     },\n      \"addons\": {\n        \"prometheus\": {\n          \"enabled\":
            true\n        },\n        \"kiali\": {\n          \"enabled\": true\n
            \       },\n        \"grafana\": {\n          \"enabled\": true\n        }\n
            \     }\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n    \"kind\":
            \"ServiceMeshMemberRoll\",\n    \"metadata\": {\n      \"name\": \"default\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"members\": [\n        \"your-project\",\n        \"another-of-your-projects\"
            \n      ]\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n
            \   \"kind\": \"ServiceMeshMember\",\n    \"metadata\": {\n      \"name\":
            \"default\",\n      \"namespace\": \"application-namespace\"\n    },\n
            \   \"spec\": {\n      \"controlPlaneRef\": {\n        \"name\": \"basic\",\n
            \       \"namespace\": \"control-plane-namespace\"\n      }\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
          createdAt: 2025-08-11T09:32:44CEST
          description: The OpenShift Service Mesh 2 Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh 2.6
            and earlier releases. OpenShift Service Mesh 2 is based on the open source
            Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2.6.10-0'
          operators.openshift.io/infrastructure-features: '["Disconnected","fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/maistra/istio-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Istio control plane installation
            displayName: Istio Service Mesh Control Plane
            kind: ServiceMeshControlPlane
            name: servicemeshcontrolplanes.maistra.io
            version: v2
          - description: Marks the containing namespace as a member of the referenced
              Service Mesh
            displayName: Istio Service Mesh Member
            kind: ServiceMeshMember
            name: servicemeshmembers.maistra.io
            version: v1
          - description: A list of namespaces in Service Mesh
            displayName: Istio Service Mesh Member Roll
            kind: ServiceMeshMemberRoll
            name: servicemeshmemberrolls.maistra.io
            version: v1
        description: |-
          Red Hat OpenShift Service Mesh 2 is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh 2, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh 2
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh 2 provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh 2 supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### Joining Projects Into a Mesh

          Once an instance of Red Hat OpenShift Service Mesh 2 has been installed, it will only exercise control over services within its own
          project.  Other projects may be added into the mesh using one of two methods:

          1. A **ServiceMeshMember** resource may be created in other projects to add those projects into the mesh.  The
            **ServiceMeshMember** specifies a reference to the **ServiceMeshControlPlane** object that was used to install
            the control plane.  The user creating the **ServiceMeshMember** resource must have permission to *use* the
            **ServiceMeshControlPlane** object.  The adminstrator for the project containing the control plane can grant
            individual users or groups the *use* permissions.

          2. A **ServiceMeshMemberRoll** resource may be created in the project containing the control plane.  This resource
            contains a single *members* list of all the projects that should belong in the mesh.  The resource must be named
            *default*.  The user creating the resource must have *edit* or *admin* permissions for all projects in the
            *members* list.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 2
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - maistra
        - servicemesh
        links:
        - name: Red Hat OpenShift Service Mesh 2
          url: https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html
        - name: Istio
          url: https://istio.io/
        - name: Operator Source Code
          url: https://github.com/Maistra/istio-operator
        - name: Bugs
          url: https://issues.redhat.com/projects/OSSM
        maintainers:
        - email: istio-feedback@redhat.com
          name: Red Hat, OpenShift Service Mesh
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:4ed63a5440753b1b22d44191246c82e7671e3f545c1dbdae4cb4fad163ca3a46
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6b4623c416fc85bd8e38904c16494fa3c4aeae25c90dc0da9763a0a87141bc04
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:515c5d9e6700fae981b0ea3807436b250acb1534e9ecf50043e5df8474e66dce
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:b134578e0677963197e353d8acd0ac3330df1b6d2752aa41cb072e86b5b4b91b
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6fb7d50b188dc27ffb9a396edabbfb21767fe438344fd70d4754d1ec7cf27957
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:e734ca310e09f737b4f06f3ecfeb9ac12971c63884fdcc27e3bebef9d5b18722
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:a36d5db6097749c55e946bba3b8ddd13a84b1f6f68b62d0ccbad5a803e9b46fb
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:fc500a282c6668ff987a632efbbd19ecc3aebdc021f831a7b386e360a5c8324d
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:57eb7ba03f24781501a7ceed2e160ff95317869af6c2caf8f9a83990e6a6c7b9
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:1ca80927334c6af8e386de8d44704f568b4b35d3f7afbe785a999d0fa4efd0bf
        - registry.redhat.io/openshift4/ose-prometheus@sha256:423ae2e81cfe660b7217f7662754f366a91d457d6dd1589b4fc40733896eb9d6
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:d16a60204d266bf4cc266c005fabf175b533b827500f171013884a713c88857a
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel8@sha256:3d773d16fed72278949b9e3b54dc59e346e95341dc94b715bcfc328d3c9e83b2
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:6a420228e684596ce2b900614a874f3e1284a6f2bbbba4c5a71ade158867affe
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:c77183e50a34d33b5e2fede60350f54b18ab3536c3189e331569d5df862d9979
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:a3d65983d46dcf47b94f26e7265ed32082208ce08abcf3587a9a4e15c19b274a
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:b241452cce5c9af5aa0f12d1db3a7094cc004ce355988b8a5207091810a6f84b
        - registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:3dc1bddb070fd530ce6251666989b9a849de90591718a578da4a9900f98d93ec
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:c34b4cd9a82fef40708a25f0568f6dc632878affba35d5d631a48ec31da08cd3
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:c1cadf40bdb9511d612a0257c12eb3a9738823592ebb6e32a4f51b3a8dfe08db
        - registry.redhat.io/openshift-service-mesh/prometheus-rhel8@sha256:71e193a2d403bb06a7a76686b376985315aa0c32f495b8b9dacd73c7ccce2af3
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d95762c62087072cadee5948c706619f09202a349e1310f0ed81bfb42f11203f
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:4669724b430683bf2e895c5cf941c2d99536117815999d38131e5b6f162bbef0
        version: 2.6.10-0
      entries:
      - name: servicemeshoperator.v2.6.10
        version: 2.6.10-0
      - name: servicemeshoperator.v2.6.9
        version: 2.6.9-0
      - name: servicemeshoperator.v2.6.8
        version: 2.6.8-0
      - name: servicemeshoperator.v2.6.7
        version: 2.6.7-0
      - name: servicemeshoperator.v2.6.6
        version: 2.6.6-0
      - name: servicemeshoperator.v2.6.5
        version: 2.6.5-0
      - name: servicemeshoperator.v2.6.4
        version: 2.6.4-0
      - name: servicemeshoperator.v2.6.3
        version: 2.6.3-0
      - name: servicemeshoperator.v2.6.2
        version: 2.6.2-0
      - name: servicemeshoperator.v2.6.1
        version: 2.6.1-0
      - name: servicemeshoperator.v2.6.0
        version: 2.6.0-0
      - name: servicemeshoperator.v2.5.2
        version: 2.5.2-0
      - name: servicemeshoperator.v2.5.1
        version: 2.5.1-0
      - name: servicemeshoperator.v2.5.0
        version: 2.5.0-0
      - name: servicemeshoperator.v2.4.5
        version: 2.4.5-0
      - name: servicemeshoperator.v2.4.4
        version: 2.4.4-0
      - name: servicemeshoperator.v2.4.3
        version: 2.4.3-0
      - name: servicemeshoperator.v2.4.2
        version: 2.4.2-0
      - name: servicemeshoperator.v2.4.1
        version: 2.4.1-0
      - name: servicemeshoperator.v2.4.0
        version: 2.4.0-0
      - name: servicemeshoperator.v2.3.3
        version: 2.3.3-0
      - name: servicemeshoperator.v2.3.2
        version: 2.3.2-0
      - name: servicemeshoperator.v2.3.1
        version: 2.3.1-0
      - name: servicemeshoperator.v2.3.0
        version: 2.3.0-0
      - name: servicemeshoperator.v2.2.3
        version: 2.2.3-0
      - name: servicemeshoperator.v2.2.2
        version: 2.2.2-0
      - name: servicemeshoperator.v2.2.1
        version: 2.2.1-0
      - name: servicemeshoperator.v2.2.0
        version: 2.2.0-0
      - name: servicemeshoperator.v2.1.3
        version: 2.1.3-0
      - name: servicemeshoperator.v2.1.2
        version: 2.1.2-0
      - name: servicemeshoperator.v2.1.1
        version: 2.1.1-0
      - name: servicemeshoperator.v2.1.0
        version: 2.1.0-0
      - name: servicemeshoperator.v2.0.8
        version: 2.0.8-0
      name: "1.0"
    - currentCSV: servicemeshoperator.v2.6.10
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"maistra.io/v2\",\n    \"kind\":
            \"ServiceMeshControlPlane\",\n    \"metadata\": {\n      \"name\": \"basic\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"version\": \"v2.6\",\n      \"policy\": {\n         \"type\":
            \"Istiod\"\n      },\n      \"telemetry\": {\n         \"type\": \"Istiod\"\n
            \     },\n      \"addons\": {\n        \"prometheus\": {\n          \"enabled\":
            true\n        },\n        \"kiali\": {\n          \"enabled\": true\n
            \       },\n        \"grafana\": {\n          \"enabled\": true\n        }\n
            \     }\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n    \"kind\":
            \"ServiceMeshMemberRoll\",\n    \"metadata\": {\n      \"name\": \"default\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"members\": [\n        \"your-project\",\n        \"another-of-your-projects\"
            \n      ]\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n
            \   \"kind\": \"ServiceMeshMember\",\n    \"metadata\": {\n      \"name\":
            \"default\",\n      \"namespace\": \"application-namespace\"\n    },\n
            \   \"spec\": {\n      \"controlPlaneRef\": {\n        \"name\": \"basic\",\n
            \       \"namespace\": \"control-plane-namespace\"\n      }\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
          createdAt: 2025-08-11T09:32:44CEST
          description: The OpenShift Service Mesh 2 Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh 2.6
            and earlier releases. OpenShift Service Mesh 2 is based on the open source
            Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2.6.10-0'
          operators.openshift.io/infrastructure-features: '["Disconnected","fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/maistra/istio-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Istio control plane installation
            displayName: Istio Service Mesh Control Plane
            kind: ServiceMeshControlPlane
            name: servicemeshcontrolplanes.maistra.io
            version: v2
          - description: Marks the containing namespace as a member of the referenced
              Service Mesh
            displayName: Istio Service Mesh Member
            kind: ServiceMeshMember
            name: servicemeshmembers.maistra.io
            version: v1
          - description: A list of namespaces in Service Mesh
            displayName: Istio Service Mesh Member Roll
            kind: ServiceMeshMemberRoll
            name: servicemeshmemberrolls.maistra.io
            version: v1
        description: |-
          Red Hat OpenShift Service Mesh 2 is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh 2, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh 2
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh 2 provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh 2 supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### Joining Projects Into a Mesh

          Once an instance of Red Hat OpenShift Service Mesh 2 has been installed, it will only exercise control over services within its own
          project.  Other projects may be added into the mesh using one of two methods:

          1. A **ServiceMeshMember** resource may be created in other projects to add those projects into the mesh.  The
            **ServiceMeshMember** specifies a reference to the **ServiceMeshControlPlane** object that was used to install
            the control plane.  The user creating the **ServiceMeshMember** resource must have permission to *use* the
            **ServiceMeshControlPlane** object.  The adminstrator for the project containing the control plane can grant
            individual users or groups the *use* permissions.

          2. A **ServiceMeshMemberRoll** resource may be created in the project containing the control plane.  This resource
            contains a single *members* list of all the projects that should belong in the mesh.  The resource must be named
            *default*.  The user creating the resource must have *edit* or *admin* permissions for all projects in the
            *members* list.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 2
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - maistra
        - servicemesh
        links:
        - name: Red Hat OpenShift Service Mesh 2
          url: https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html
        - name: Istio
          url: https://istio.io/
        - name: Operator Source Code
          url: https://github.com/Maistra/istio-operator
        - name: Bugs
          url: https://issues.redhat.com/projects/OSSM
        maintainers:
        - email: istio-feedback@redhat.com
          name: Red Hat, OpenShift Service Mesh
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:c1cadf40bdb9511d612a0257c12eb3a9738823592ebb6e32a4f51b3a8dfe08db
        - registry.redhat.io/openshift4/ose-prometheus@sha256:423ae2e81cfe660b7217f7662754f366a91d457d6dd1589b4fc40733896eb9d6
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:4ed63a5440753b1b22d44191246c82e7671e3f545c1dbdae4cb4fad163ca3a46
        - registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:57eb7ba03f24781501a7ceed2e160ff95317869af6c2caf8f9a83990e6a6c7b9
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:3dc1bddb070fd530ce6251666989b9a849de90591718a578da4a9900f98d93ec
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel8@sha256:3d773d16fed72278949b9e3b54dc59e346e95341dc94b715bcfc328d3c9e83b2
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:a3d65983d46dcf47b94f26e7265ed32082208ce08abcf3587a9a4e15c19b274a
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:6a420228e684596ce2b900614a874f3e1284a6f2bbbba4c5a71ade158867affe
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:c77183e50a34d33b5e2fede60350f54b18ab3536c3189e331569d5df862d9979
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:d16a60204d266bf4cc266c005fabf175b533b827500f171013884a713c88857a
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:1ca80927334c6af8e386de8d44704f568b4b35d3f7afbe785a999d0fa4efd0bf
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6b4623c416fc85bd8e38904c16494fa3c4aeae25c90dc0da9763a0a87141bc04
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:515c5d9e6700fae981b0ea3807436b250acb1534e9ecf50043e5df8474e66dce
        - registry.redhat.io/openshift-service-mesh/prometheus-rhel8@sha256:71e193a2d403bb06a7a76686b376985315aa0c32f495b8b9dacd73c7ccce2af3
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d95762c62087072cadee5948c706619f09202a349e1310f0ed81bfb42f11203f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:e734ca310e09f737b4f06f3ecfeb9ac12971c63884fdcc27e3bebef9d5b18722
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:a36d5db6097749c55e946bba3b8ddd13a84b1f6f68b62d0ccbad5a803e9b46fb
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:b134578e0677963197e353d8acd0ac3330df1b6d2752aa41cb072e86b5b4b91b
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6fb7d50b188dc27ffb9a396edabbfb21767fe438344fd70d4754d1ec7cf27957
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:4669724b430683bf2e895c5cf941c2d99536117815999d38131e5b6f162bbef0
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:fc500a282c6668ff987a632efbbd19ecc3aebdc021f831a7b386e360a5c8324d
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:c34b4cd9a82fef40708a25f0568f6dc632878affba35d5d631a48ec31da08cd3
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:b241452cce5c9af5aa0f12d1db3a7094cc004ce355988b8a5207091810a6f84b
        version: 2.6.10-0
      entries:
      - name: servicemeshoperator.v2.6.10
        version: 2.6.10-0
      - name: servicemeshoperator.v2.6.9
        version: 2.6.9-0
      - name: servicemeshoperator.v2.6.8
        version: 2.6.8-0
      - name: servicemeshoperator.v2.6.7
        version: 2.6.7-0
      - name: servicemeshoperator.v2.6.6
        version: 2.6.6-0
      - name: servicemeshoperator.v2.6.5
        version: 2.6.5-0
      - name: servicemeshoperator.v2.6.4
        version: 2.6.4-0
      - name: servicemeshoperator.v2.6.3
        version: 2.6.3-0
      - name: servicemeshoperator.v2.6.2
        version: 2.6.2-0
      - name: servicemeshoperator.v2.6.1
        version: 2.6.1-0
      - name: servicemeshoperator.v2.6.0
        version: 2.6.0-0
      - name: servicemeshoperator.v2.5.2
        version: 2.5.2-0
      - name: servicemeshoperator.v2.5.1
        version: 2.5.1-0
      - name: servicemeshoperator.v2.5.0
        version: 2.5.0-0
      - name: servicemeshoperator.v2.4.5
        version: 2.4.5-0
      - name: servicemeshoperator.v2.4.4
        version: 2.4.4-0
      - name: servicemeshoperator.v2.4.3
        version: 2.4.3-0
      - name: servicemeshoperator.v2.4.2
        version: 2.4.2-0
      - name: servicemeshoperator.v2.4.1
        version: 2.4.1-0
      - name: servicemeshoperator.v2.4.0
        version: 2.4.0-0
      - name: servicemeshoperator.v2.3.3
        version: 2.3.3-0
      - name: servicemeshoperator.v2.3.2
        version: 2.3.2-0
      - name: servicemeshoperator.v2.3.1
        version: 2.3.1-0
      - name: servicemeshoperator.v2.3.0
        version: 2.3.0-0
      - name: servicemeshoperator.v2.2.3
        version: 2.2.3-0
      - name: servicemeshoperator.v2.2.2
        version: 2.2.2-0
      - name: servicemeshoperator.v2.2.1
        version: 2.2.1-0
      - name: servicemeshoperator.v2.2.0
        version: 2.2.0-0
      - name: servicemeshoperator.v2.1.3
        version: 2.1.3-0
      - name: servicemeshoperator.v2.1.2
        version: 2.1.2-0
      - name: servicemeshoperator.v2.1.1
        version: 2.1.1-0
      - name: servicemeshoperator.v2.1.0
        version: 2.1.0-0
      - name: servicemeshoperator.v2.0.8
        version: 2.0.8-0
      name: stable
    defaultChannel: stable
    packageName: servicemeshoperator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: service-binding-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-binding-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:b355d2370926063a3de8c7a09c0ca8c2348a8d158d6a7095d0e5e8381cc95be5
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/redhat-developer/servicebinding-operator@sha256:b355d2370926063a3de8c7a09c0ca8c2348a8d158d6a7095d0e5e8381cc95be5
        version: 0.10.0
      entries:
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: beta
    - currentCSV: service-binding-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindablekinds-sample"
                }
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:c9c9e45321db2d4d3cdc2d364337af844f69c9893d0cc74d2fa56c846ffcf3da
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/redhat-developer/servicebinding-operator@sha256:c9c9e45321db2d4d3cdc2d364337af844f69c9893d0cc74d2fa56c846ffcf3da
        version: 0.11.0
      entries:
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: candidate
    - currentCSV: service-binding-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindable-kinds"
                },
                "spec": {}
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  },
                  "type": "postgres",
                  "workload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:ac47f496fb7ecdcbc371f8c809fad2687ec0c35bbc8c522a7ab63b3e5ffd90ea
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'.
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1alpha3
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1beta1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
          - description: Service Binding implementing community specification (https://github.com/k8s-service-bindings/spec)
              describing the connection between a backing service and an application
              workload.
            displayName: Service Binding (Spec API)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1beta1
        description: |
          The Service Binding Operator manages the data plane for applications and backing services. It enables developers to connect their
          applications to backing services (REST API, databases, event buses and others) with a consistent and predictable
          experience. The operator reads data made available by the control plane of backing services and projects the bindings
          data to applications either as accessible files mounted into the application’s container or environment variables.

          In Kubernetes, each service suggests a different way to access their secrets and developers will consume them in their
          application in a custom way. While this provides a good deal of flexibility, it also creates discrepancies and
          complexity dealing with each specific services, which leads large development teams to lose overall velocity.

          Service Binding Operator removes the error-prone manual configuration of binding information, provides service operators
          a low-touch administrative experience to provision and manage access to services and enriches development lifecycle.

          For additional details refer to the project [documentation](https://redhat-developer.github.io/service-binding-operator).

          Get Started with Service Binding Operator
          * Find out about SBO’s [prerequisites](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/installing-service-binding.html).
          * Refer to the [Quick Start Guide](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/quick-start.html) to see SBO in action in a simple scenario.

          Need Help
          * Raise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)

          Licence
          * Service Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE)

          Service Binding Operator is providing and supporting two different APIs.
          * `binding.operators.coreos.com/v1alpha1`: This API is compliant with the Service Binding specification. It is fully supported and considered as the mature API. It’s the API we recommend in production environments.
          * `servicebinding.io/v1beta1`: This API is the one defined in the Service Binding specification. It is available in Service Binding Operator 1.x.

          We’ll consider supporting only the API from the specification, once the Service Binding specification will be officially GA. Of course, that will be a breaking change and handle in a safe way for you.
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Documentation
          url: https://redhat-developer.github.io/service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/redhat-developer/servicebinding-operator@sha256:ac47f496fb7ecdcbc371f8c809fad2687ec0c35bbc8c522a7ab63b3e5ffd90ea
        version: 1.4.0
      entries:
      - name: service-binding-operator.v1.4.0
        version: 1.4.0
      - name: service-binding-operator.v1.3.4
        version: 1.3.4
      - name: service-binding-operator.v1.3.3
        version: 1.3.3
      - name: service-binding-operator.v1.3.2
        version: 1.3.2
      - name: service-binding-operator.v1.3.1
        version: 1.3.1
      - name: service-binding-operator.v1.3.0
        version: 1.3.0
      - name: service-binding-operator.v1.2.0
        version: 1.2.0
      - name: service-binding-operator.v1.1.1
        version: 1.1.1
      - name: service-binding-operator.v1.1.0
        version: 1.1.0
      - name: service-binding-operator.v1.0.1
        version: 1.0.1
      - name: service-binding-operator.v1.0.0
        version: 1.0.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: service-binding-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: netobserv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netobserv-operator.v1.6.3-community
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowMetric",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netobserv-operator",
                    "app.kubernetes.io/instance": "flowmetric-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "flowmetric",
                    "app.kubernetes.io/part-of": "netobserv-operator"
                  },
                  "name": "flowmetric-sample"
                },
                "spec": {
                  "charts": [
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "",
                          "promQL": "sum(rate($METRIC[2m]))"
                        }
                      ],
                      "title": "External ingress traffic",
                      "type": "SingleStat",
                      "unit": "Bps"
                    },
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "{{DstK8S_Namespace}} / {{DstK8S_OwnerName}}",
                          "promQL": "sum(rate($METRIC{DstK8S_Namespace!=\"\"}[2m])) by (DstK8S_Namespace, DstK8S_OwnerName)"
                        }
                      ],
                      "sectionName": "External",
                      "title": "Top external ingress traffic per workload",
                      "type": "StackArea",
                      "unit": "Bps"
                    }
                  ],
                  "direction": "Ingress",
                  "filters": [
                    {
                      "field": "SrcSubnetLabel",
                      "matchType": "Absence"
                    }
                  ],
                  "labels": [
                    "DstK8S_HostName",
                    "DstK8S_Namespace",
                    "DstK8S_OwnerName",
                    "DstK8S_OwnerType"
                  ],
                  "metricName": "cluster_external_ingress_bytes_total",
                  "type": "Counter",
                  "valueField": "Bytes"
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "app"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "infra"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "enable": true,
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "includeList": [
                        "node_ingress_bytes_total",
                        "workload_ingress_bytes_total",
                        "namespace_flows_total"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta2",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "metrics": {
                        "server": {
                          "port": 9400
                        }
                      },
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "eBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "Disabled"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "\"app\""
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "\"infra\""
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "\"Pod\"",
                          "src_kind": "\"Pod\""
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "\"Service\""
                        },
                        "name": "Services network"
                      }
                    ],
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    }
                  },
                  "deploymentModel": "Direct",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "enable": true,
                    "lokiStack": {
                      "name": "loki"
                    },
                    "mode": "Monolithic",
                    "monolithic": {
                      "tenantID": "netobserv",
                      "tls": {
                        "caCert": {
                          "certFile": "service-ca.crt",
                          "name": "loki-gateway-ca-bundle",
                          "type": "configmap"
                        },
                        "enable": false
                      },
                      "url": "http://loki.netobserv.svc:3100/"
                    },
                    "readTimeout": "30s",
                    "writeBatchSize": 10485760,
                    "writeBatchWait": "1s",
                    "writeTimeout": "10s"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "Flows",
                    "metrics": {
                      "disableAlerts": [],
                      "server": {
                        "port": 9401
                      }
                    },
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "prometheus": {
                    "querier": {
                      "enable": true,
                      "mode": "Auto",
                      "timeout": "30s"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:1.6.2-community
          createdAt: "2024-08-22T18:26:56Z"
          description: Network flows collector and monitoring solution
          operatorframework.io/initialization-resource: '{"apiVersion":"flows.netobserv.io/v1beta2",
            "kind":"FlowCollector","metadata":{"name":"cluster"},"spec": {}}'
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
          support: NetObserv team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta2
          - description: '`FlowMetric` is the schema for the custom metrics API, which
              allows to generate more metrics out of flow logs. You can find examples
              here: https://github.com/netobserv/network-observability-operator/tree/1.6.2-community/config/samples/flowmetrics'
            displayName: Flow Metric
            kind: FlowMetric
            name: flowmetrics.flows.netobserv.io
            version: v1alpha1
        description: |-
          DEPRECATED: this distribution channel is deprecated and doesn't receive new updates. Please read https://github.com/orgs/netobserv/discussions/996 to learn more.

          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline that consists in:
          - an eBPF agent, that generates network flows from captured packets
          - flowlogs-pipeline, a component that collects, enriches and exports these flows
          - when used in OpenShift, a Console plugin for flows visualization with powerful filtering options, a topology representation and more

          Flow data is then available in multiple ways, each optional:

          - As Prometheus metrics
          - As raw flow logs stored in Grafana Loki
          - As raw flow logs exported to a collector

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, can optionally be used as the backend to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 10GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          If you prefer to not use Loki, you must set `spec.loki.enable` to `false` in `FlowCollector`.
          In that case, you can still get the Prometheus metrics or export raw flows to a custom collector. But be aware that some of the Console plugin features will be disabled. For instance, you will not be able to view raw flows there, and the metrics / topology will have a more limited level of details, missing information such as pods or IPs.

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for example [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/config/samples/flows_v1beta2_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide on every node, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you might have to configure differently if you used another installation method. Make sure to disable it (`spec.loki.enable`) if you don't want to use Loki.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`) an optional list of exporters to which to send enriched flows. KAFKA and IPFIX exporters are supported. This allows you to define any custom storage or processing that can read from Kafka or use the IPFIX standard.

          - To enable availability zones awareness, set `spec.processor.addZone` to `true`.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - More information on [configuring metrics](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/Metrics.md).
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: '[DEPRECATED] NetObserv Operator'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/network-observability-operator:1.6.2-community
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/netobserv/netobserv-ebpf-agent:v1.6.2-community
        - quay.io/netobserv/flowlogs-pipeline:v1.6.2-community
        - quay.io/netobserv/network-observability-console-plugin:v1.6.2-community
        version: 1.6.3-community
      entries:
      - name: netobserv-operator.v1.6.3-community
        version: 1.6.3-community
      - name: netobserv-operator.v1.6.2-community
        version: 1.6.2-community
      - name: netobserv-operator.v1.6.1-community
        version: 1.6.1-community
      - name: netobserv-operator.v1.6.0-community
        version: 1.6.0-community
      - name: netobserv-operator.v1.5.0-community
        version: 1.5.0-community
      name: community
    - currentCSV: netobserv-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "ipfix": {
                      "cacheActiveTimeout": "20s",
                      "cacheMaxFlows": 400,
                      "clusterNetworkOperator": {
                        "namespace": "openshift-network-operator"
                      },
                      "ovnKubernetes": {
                        "containerName": "ovnkube-node",
                        "daemonSetName": "ovnkube-node",
                        "namespace": "ovn-kubernetes"
                      },
                      "sampling": 400
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "staticLabels": {
                      "app": "netobserv-flowcollector"
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "enableKubeProbes": true,
                    "healthPort": 8080,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:0.2.1
          createdAt: "2022-12-07T09:00:48Z"
          description: Network flows collector and monitoring solution
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
        description: |-
          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline to collect and enrich network flows. These flows can be produced by the NetObserv eBPF agent, or by any device or CNI able to export flows in IPFIX format, such as OVN-Kubernetes.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When used in OpenShift, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For a normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionnaly be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/0.2.1/config/samples/flows_v1alpha1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Agent (`spec.agent.type`) can be `EBPF` (default) or `IPFIX`. eBPF is recommended, as it should work in more situations and offers better performances. If you can't, or don't want to use eBPF, note that the IPFIX option is fully functional only when using [OVN-Kubernetes](https://github.com/ovn-org/ovn-kubernetes/) CNI. Other CNIs are not officially supported, but you may still be able to configure them manually if they allow IPFIX exports.

          - Sampling (`spec.agent.ebpf.sampling` and `spec.agent.ipfix.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50) for eBPF and 400 (1:400) for IPFIX. Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub more more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - A [F.A.Q](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: NetObserv Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: mmaciasl@redhat.com
          name: Mario Macias
        - email: ocazade@redhat.com
          name: Olivier Cazade
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/netobserv-ebpf-agent:v0.2.2
        - quay.io/netobserv/flowlogs-pipeline:v0.1.5
        - quay.io/netobserv/network-observability-console-plugin:v0.1.6
        - quay.io/netobserv/network-observability-operator:0.2.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        version: 0.2.1
      entries:
      - name: netobserv-operator.v0.2.1
        version: 0.2.1
      - name: netobserv-operator.v0.2.0
        version: 0.2.0
      - name: netobserv-operator.v0.1.4
        version: 0.1.4
      - name: netobserv-operator.v0.1.3
        version: 0.1.3
      - name: netobserv-operator.v0.1.2
        version: 0.1.2
      - name: netobserv-operator.v0.1.1
        version: 0.1.1
      - name: netobserv-operator.v0.1.0
        version: 0.1.0
      name: latest
    - currentCSV: netobserv-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "ipfix": {
                      "cacheActiveTimeout": "20s",
                      "cacheMaxFlows": 400,
                      "clusterNetworkOperator": {
                        "namespace": "openshift-network-operator"
                      },
                      "ovnKubernetes": {
                        "containerName": "ovnkube-node",
                        "daemonSetName": "ovnkube-node",
                        "namespace": "ovn-kubernetes"
                      },
                      "sampling": 400
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "staticLabels": {
                      "app": "netobserv-flowcollector"
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "enableKubeProbes": true,
                    "healthPort": 8080,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:0.2.1
          createdAt: "2022-12-07T09:00:48Z"
          description: Network flows collector and monitoring solution
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
        description: |-
          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline to collect and enrich network flows. These flows can be produced by the NetObserv eBPF agent, or by any device or CNI able to export flows in IPFIX format, such as OVN-Kubernetes.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When used in OpenShift, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For a normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionnaly be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/0.2.1/config/samples/flows_v1alpha1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Agent (`spec.agent.type`) can be `EBPF` (default) or `IPFIX`. eBPF is recommended, as it should work in more situations and offers better performances. If you can't, or don't want to use eBPF, note that the IPFIX option is fully functional only when using [OVN-Kubernetes](https://github.com/ovn-org/ovn-kubernetes/) CNI. Other CNIs are not officially supported, but you may still be able to configure them manually if they allow IPFIX exports.

          - Sampling (`spec.agent.ebpf.sampling` and `spec.agent.ipfix.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50) for eBPF and 400 (1:400) for IPFIX. Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub more more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - A [F.A.Q](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: NetObserv Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: mmaciasl@redhat.com
          name: Mario Macias
        - email: ocazade@redhat.com
          name: Olivier Cazade
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/network-observability-operator:0.2.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/netobserv/network-observability-console-plugin:v0.1.6
        - quay.io/netobserv/netobserv-ebpf-agent:v0.2.2
        - quay.io/netobserv/flowlogs-pipeline:v0.1.5
        version: 0.2.1
      entries:
      - name: netobserv-operator.v0.2.1
        version: 0.2.1
      - name: netobserv-operator.v0.2.0
        version: 0.2.0
      - name: netobserv-operator.v0.1.4
        version: 0.1.4
      - name: netobserv-operator.v0.1.3
        version: 0.1.3
      - name: netobserv-operator.v0.1.2
        version: 0.1.2
      - name: netobserv-operator.v0.1.1
        version: 0.1.1
      - name: netobserv-operator.v0.1.0
        version: 0.1.0
      name: v0.2.x
    defaultChannel: community
    packageName: netobserv-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-license-service-scanner-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-license-service-scanner-operator.v4.2.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1",
                "kind": "IBMLicenseServiceScanner",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-licensing-scanner-operator",
                    "app.kubernetes.io/instance": "ibm-licensing-scanner-instance",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ibm-licensing-scanner-instance",
                    "app.kubernetes.io/part-of": "ibm-licensing-scanner-operator"
                  },
                  "name": "ibm-licensing-scanner-instance"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "license-service-upload-secret": "ibm-licensing-scanner-upload-secret",
                  "log-level": "INFO",
                  "registry-pull-secret": "ibm-licensing-scanner-registry-pull-secret",
                  "scan": {
                    "namespaces": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:77c9db26a13dba25438a6fc5a41290606418481e08ea2ecc0a9a397062af74fb
          createdAt: "2025-05-14T12:59:00Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <4.2.16'
          operatorframework.io/suggested-namespace: ibm-licensing-scanner
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMLicenseServiceScanner is the Schema for the ibmlicenseservicescanners
              API
            displayName: IBMLicense Service Scanner
            kind: IBMLicenseServiceScanner
            name: ibmlicenseservicescanners.operator.ibm.com
            version: v1
        description: IBM License Service Scanner
        displayName: IBM License Service Scanner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - licensing
        - usage
        - IBMLicenseServiceScanner
        - IBM
        - Cloud
        - License
        - License Service
        - Service
        - Scanner
        - License Service Scanner
        - License Scanner
        - Service Scanner
        links:
        - name: Ibm License Service Scanner Operator
          url: https://ibm-license-service-scanner-operator.domain
        maintainers:
        - email: talk2sam@us.ibm.com
          name: IBM
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-bundle@sha256:b6cb9e5976b749dd7818ac19d1cfd5e24673f3c052a41d6e2fe1d2d7094694b4
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-catalog@sha256:c016904e8c5cfcf8270241ad5e28b52f528859984ed894cdd5e1283fa90cd6e8
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:77c9db26a13dba25438a6fc5a41290606418481e08ea2ecc0a9a397062af74fb
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:f3503f72a90ff8abbadaef3e78a16db6e1fbff4c90b475c33ead17bf1e69990c
        version: 4.2.16
      entries:
      - name: ibm-license-service-scanner-operator.v4.2.16
        version: 4.2.16
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-license-service-scanner-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/component: operator
      app.kubernetes.io/managed-by: kustomize
      app.kubernetes.io/name: zabbix-operator-certified
      app.kubernetes.io/part-of: zabbix-operator-certified
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zabbix LLC
      provider-url: ""
    name: zabbix-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zabbix-operator-certified.v6.0.40
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent",
                "metadata": {
                  "labels": {
                    "app": "agent",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "log_remote_commands": true,
                  "max_lines_per_second": 20,
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent2",
                "metadata": {
                  "labels": {
                    "app": "agent2",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent2"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "enable_persistent_buffer": false,
                  "enable_status_port": false,
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "persistent_buffer_claim": "",
                  "persistent_buffer_period": "1h",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixFull",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-full"
                },
                "spec": {
                  "builtin_report_service": true,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysql_volumeclaim": "mysql-full-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxyMysql",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-mysql"
                },
                "spec": {
                  "db_server_port": 3306,
                  "enable_snmp_traps": false,
                  "internal_db": true,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix_proxy",
                  "proxy": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "zabbix_mysql_volumeclaim": "mysql-proxy-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxySqlite",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-sqlite"
                },
                "spec": {
                  "db_data_volumeclaim": "zabbix-proxy-sqlite",
                  "enable_snmp_traps": false,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "proxy": {
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "node_name_as_db_name": true,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "proxy_size": 1,
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  }
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixServer",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-server"
                },
                "spec": {
                  "builtin_report_service": true,
                  "db_server_host": "mysql-server",
                  "db_server_port": 3306,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix",
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              }
            ]
          ansible.operator-sdk/reconcile-period: 5s
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          certified: "true"
          containerImage: registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
          createdAt: "2025-05-26T07:41:22Z"
          description: Zabbix operator with multiple deployment variants and different
            components
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <6.0.40
          operatorframework.io/suggested-namespace: zabbix
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: ""
          support: Zabbix
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Zabbix agent 2 is a new generation of Zabbix agent and may
              be used in place of Zabbix agent
            displayName: Zabbix agent 2
            kind: ZabbixAgent2
            name: zabbixagent2s.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix agent is deployed on a monitoring nodes to actively
              monitor local resources and applications
            displayName: Zabbix agent
            kind: ZabbixAgent
            name: zabbixagents.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database server, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Full
            kind: ZabbixFull
            name: zabbixfulls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with MySQL database server
            displayName: Zabbix proxy (MySQL)
            kind: ZabbixProxyMysql
            name: zabbixproxymysqls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with SQLite database
            displayName: Zabbix proxy (SQLite3)
            kind: ZabbixProxySqlite
            name: zabbixproxysqlites.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database support, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Server
            kind: ZabbixServer
            name: zabbixservers.kubernetes.zabbix.com
            version: v1alpha1
        description: "## About this Operator\n\nZabbix helps you to real-time monitoring
          of millions of metrics collected from tens of thousands of servers, virtual
          machines and network devices.\nThe Zabbix Operator allows users to easily
          deploy, manage, and maintain Zabbix deployments on OpenShift. By installing
          this integration you will be able to deploy Zabbix server / proxies and
          other components with a single command.\n## Supported Features\n* **Zabbix
          Server** - Simple Zabbix installation with included Zabbix server, Zabbix
          web-interface and Zabbix Java Gateway with MySQL database support. The feature
          does not provide MySQL service and requires an external MySQL database.
          \n* **Zabbix Server (Full)** - Zabbix installation with included Zabbix
          server, Zabbix web-interface, Zabbix Java Gateway and MySQL server instance.\n*
          **Zabbix proxy (SQLite3)** - Very simple way to gain power of Zabbix proxy.
          The feature has  SQLite3 support for Zabbix proxies and allow to specify
          amount of proxies. \n* **Zabbix proxy (MySQL)** - Another option of Zabbix
          proxy. The option support and deliver MySQL database.\n* **Zabbix agent**
          - Zabbix agent can be deployed on each available node for stability and
          performance monitoring on remote nodes. It allows to gather metrics with
          full automation!* \n## Prerequisites\nAll deployment options are require
          additional information during deployment. Please, check the following instructions
          and provide required configuration:\n* **Zabbix Server** - MySQL database
          host information and MySQL database credentials in specially formatted *Secret*.
          Additionally it is possible to specify SSL certificates for HTTPS support
          in *Secret*.\n* **Zabbix Server (Full)** - MySQL database credentials in
          specially formatted *Secret*. MySQL database volume name information. Additionally
          it is possible to specify SSL certificates for HTTPS support in *Secret*.\n*
          **Zabbix proxy (SQLite3)** - Zabbix server host information only.\n* **Zabbix
          proxy (MySQL)** - MySQL database credentials in specially formatted *Secrets*
          and Zabbix server host information.\n* **Zabbix agent** - Zabbix server
          host information only for outgoing and incoming connections. Zabbix agent
          uses \"privileged\" mode to monitor node resources! For example, running
          processes.* \n### MySQL credentials\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-full-secrets\ndata:\n  mysql_root_password: emFiYml4X3N1cGVyX3Jvb3Q=
          [1]\n  mysql_zabbix_password: emFiYml4X3VzZXJfcGFzc3dvcmQ= [2]\n  mysql_zabbix_username:
          emFiYml4 [3]\ntype: Opaque\n```\nUsing MySQL root password (*mysqlrootpassword*
          [1]) Zabbix server / proxy will try to create MySQL database schema with
          grant permissions to *mysqlzabbixusername* [2] and *mysqlzabbixpassword*
          [3].\n### SSL certificates for HTTPS\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-web-sslsecret\ndata:\n  ssl.crt: >-\n   < ssl.crt data>\n
          \ ssl.key: >-\n\t < ssl.key data >\n  dhparam.pem: >-\n   <  dhparam.pem
          data >\n```\nFiles *ssl.crt*, *ssl.key* and *dhparam.perm* are required
          for Zabbix web-interface for SSL support.\n"
        displayName: Zabbix Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - zabbix
        - appliance
        - agent
        - metrics
        links:
        - name: Zabbix
          url: https://www.zabbix.com
        - name: Zabbix Official Documentation
          url: https://www.zabbix.com/documentation/6.0/en/manual/quickstart
        - name: Zabbix Operator Installation
          url: https://www.zabbix.com/documentation/6.0/en/manual/installation/containers/openshift
        - name: Downloads
          url: https://www.zabbix.com/download
        maintainers:
        - email: alexey.pustovalov@zabbix.com
          name: Alexey Pustovalov
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Zabbix LLC
        relatedImages:
        - registry.connect.redhat.com/zabbix/zabbix-agent2-60@sha256:424003d52f3cb4360adf2b3619d8b9f8de9196c630ee7e4ceb5d177c1c5a36e4
        - registry.connect.redhat.com/zabbix/zabbix-snmptraps-60@sha256:e61e6b1c551dec8488cce0998f94d83d6d10b5d89a465d7119b088313032cb81
        - registry.connect.redhat.com/zabbix/zabbix-server-mysql-60@sha256:637484faec6e9a55049c13dbd1f100c03d757c3a21749388c0605040f02d05ce
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
        - quay.io/zabbix/zabbix-operator@sha256:43d26488a19967bdb2a8aafde5f43820be95011811da9cc5ec623c9029801675
        - registry.connect.redhat.com/zabbix/zabbix-web-mysql-60@sha256:611d2670e1bf95b81226976c0a1f64434b2a7766f196dadbd646a7d8319f4cb1
        - registry.connect.redhat.com/zabbix/zabbix-java-gateway-60@sha256:a4feb725c12c4ac90eb9244c6fafdb6708679f2e0e7f6aa683c765127306412e
        - registry.connect.redhat.com/zabbix/zabbix-proxy-mysql-60@sha256:c2ae74d97c083ff348dc9c55932a33da4779cd9efdd9d7c5a79acfffae5b9ed0
        - registry.connect.redhat.com/zabbix/zabbix-proxy-sqlite-60@sha256:62e62ebc8c1bc13fb610f7a83412e65d852f74b3c2c212baa75ea27fc7fa03eb
        - registry.connect.redhat.com/zabbix/zabbix-web-service-60@sha256:4a333094e110713470a897b6406c5e340e962caf32a429dddf19865041fcc448
        - registry.redhat.io/rhel9/mysql-80@sha256:c4b5246d0c0d089a766ea07768c22d9d6be8049ac1379bd345db99390ffe2a36
        - registry.connect.redhat.com/zabbix/zabbix-agent-60@sha256:2895b5e45785aaf528d4a6f78c8af0a0d494dbe9362403fdd470673d911d5004
        version: 6.0.40
      entries:
      - name: zabbix-operator-certified.v6.0.40
        version: 6.0.40
      - name: zabbix-operator-certified.v6.0.38
        version: 6.0.38
      - name: zabbix-operator-certified.v6.0.36
        version: 6.0.36
      - name: zabbix-operator-certified.v6.0.33
        version: 6.0.33
      - name: zabbix-operator-certified.v6.0.29
        version: 6.0.29
      - name: zabbix-operator-certified.v6.0.27
        version: 6.0.27
      - name: zabbix-operator-certified.v6.0.26
        version: 6.0.26
      - name: zabbix-operator-certified.v6.0.17
        version: 6.0.17
      - name: zabbix-operator-certified.v6.0.14
        version: 6.0.14
      - name: zabbix-operator-certified.v6.0.13
        version: 6.0.13
      name: lts
    defaultChannel: lts
    packageName: zabbix-operator-certified
    provider:
      name: Zabbix LLC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/multiarch-tuning-operator
    name: multiarch-tuning-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multiarch-tuning-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multiarch.openshift.io/v1beta1",
                "kind": "ClusterPodPlacementConfig",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logVerbosity": "Normal",
                  "namespaceSelector": {
                    "matchExpressions": [
                      {
                        "key": "multiarch.openshift.io/exclude-pod-placement",
                        "operator": "DoesNotExist"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Other
          console.openshift.io/disable-operand-delete: "false"
          containerImage: registry.redhat.io/multiarch-tuning/multiarch-tuning-rhel9-operator@sha256:1a1d3f5258d4b91f1a0380b0fe2c809823e323d9ff74ae4171468168d255e7e0
          createdAt: "2025-03-05T12:27:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operator.openshift.io/uninstall-message: You must remove all Operands before
            uninstalling the operator.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-multiarch-tuning-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/multiarch-tuning-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPodPlacementConfig defines the configuration for the
              architecture aware pod placement operand. Users can only deploy a single
              object named "cluster". Creating the object enables the operand.
            displayName: Cluster Pod Placement Config
            kind: ClusterPodPlacementConfig
            name: clusterpodplacementconfigs.multiarch.openshift.io
            version: v1alpha1
          - description: ClusterPodPlacementConfig defines the configuration for the
              architecture aware pod placement operand. Users can only deploy a single
              object named "cluster". Creating the object enables the operand.
            displayName: Cluster Pod Placement Config
            kind: ClusterPodPlacementConfig
            name: clusterpodplacementconfigs.multiarch.openshift.io
            version: v1beta1
        description: |
          The Multiarch Tuning Operator optimizes workload management within multi-architecture clusters and in
          single-architecture clusters transitioning to multi-architecture environments.

          By default, the scheduler does not consider the architecture of a pod's container images
          when determining the placement of new pods onto nodes.

          The Multiarch Tuning Operator introduces the ClusterPodPlacementConfig custom resource to
          enable architecture-aware scheduling, ensuring that pods are assigned to nodes with
          a corresponding architecture for their container images.

          To enable this functionality, you must create a ClusterPodPlacementConfig resource.

          When the ClusterPodPlacementConfig object is created, the Multiarch Tuning Operator deploys the necessary operands
          to support architecture-aware workload scheduling.
        displayName: Multiarch Tuning Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multiarch-tuning-operator
        - pod-placement
        links:
        - name: Multiarch Tuning Operator Documentation
          url: https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/postinstallation_configuration/configuring-multi-architecture-compute-machines-on-an-openshift-cluster#multiarch-tuning-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/multiarch-tuning-operator
        relatedImages:
        - registry.redhat.io/multiarch-tuning/multiarch-tuning-rhel9-operator@sha256:1a1d3f5258d4b91f1a0380b0fe2c809823e323d9ff74ae4171468168d255e7e0
        version: 1.1.0
      entries:
      - name: multiarch-tuning-operator.v1.1.0
        version: 1.1.0
      - name: multiarch-tuning-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: multiarch-tuning-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/multiarch-tuning-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: infra-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infra-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instanceha.openstack.org/v1beta1",
                "kind": "InstanceHa",
                "metadata": {
                  "name": "instanceha"
                },
                "spec": {
                  "caBundleSecretName": "combined-ca-bundle"
                }
              },
              {
                "apiVersion": "memcached.openstack.org/v1beta1",
                "kind": "Memcached",
                "metadata": {
                  "name": "memcached"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "network.openstack.org/v1beta1",
                "kind": "DNSData",
                "metadata": {
                  "name": "allhosts"
                },
                "spec": {
                  "hosts": [
                    {
                      "hostnames": [
                        "edpm-compute-0.ctlplane"
                      ],
                      "ip": "192.168.122.100"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-0.internalapi"
                      ],
                      "ip": "172.17.0.100"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-0.storage"
                      ],
                      "ip": "172.18.0.100"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-0.tenant"
                      ],
                      "ip": "172.19.0.100"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-1.ctlplane"
                      ],
                      "ip": "192.168.122.101"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-1.internalapi"
                      ],
                      "ip": "172.17.0.101"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-1.storage"
                      ],
                      "ip": "172.18.0.101"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-1.tenant"
                      ],
                      "ip": "172.19.0.101"
                    },
                    {
                      "hostnames": [
                        "keystone-internal.openstack.svc",
                        "placement-internal.openstack.svc",
                        "cinder-internal.openstack.svc",
                        "glance-internal.openstack.svc",
                        "nova-internal.openstack.svc",
                        "nova-metadata-internal.openstack.svc",
                        "neutron-internal.openstack.svc"
                      ],
                      "ip": "172.17.0.80"
                    },
                    {
                      "hostnames": [
                        "rabbitmq.openstack.svc"
                      ],
                      "ip": "172.17.0.85"
                    },
                    {
                      "hostnames": [
                        "rabbitmq-cell1.openstack.svc"
                      ],
                      "ip": "172.17.0.86"
                    }
                  ]
                }
              },
              {
                "apiVersion": "network.openstack.org/v1beta1",
                "kind": "DNSMasq",
                "metadata": {
                  "name": "dnsmasq"
                },
                "spec": {
                  "debug": {
                    "service": false
                  },
                  "options": [
                    {
                      "key": "server",
                      "values": [
                        "192.168.122.1"
                      ]
                    },
                    {
                      "key": "local",
                      "values": [
                        "/example.com/"
                      ]
                    }
                  ],
                  "override": {
                    "service": {
                      "metadata": {
                        "annotations": {
                          "metallb.universe.tf/address-pool": "ctlplane",
                          "metallb.universe.tf/allow-shared-ip": "ctlplane",
                          "metallb.universe.tf/loadBalancerIPs": "192.168.122.80"
                        }
                      },
                      "spec": {
                        "type": "LoadBalancer"
                      }
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "network.openstack.org/v1beta1",
                "kind": "IPSet",
                "metadata": {
                  "name": "edpm-compute-0"
                },
                "spec": {
                  "networks": [
                    {
                      "defaultRoute": true,
                      "name": "CtlPlane",
                      "subnetName": "subnet1"
                    },
                    {
                      "name": "InternalApi",
                      "subnetName": "subnet1"
                    },
                    {
                      "name": "Storage",
                      "subnetName": "subnet1"
                    },
                    {
                      "name": "Tenant",
                      "subnetName": "subnet1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "network.openstack.org/v1beta1",
                "kind": "NetConfig",
                "metadata": {
                  "name": "netconfig"
                },
                "spec": {
                  "networks": [
                    {
                      "dnsDomain": "ctlplane.example.com",
                      "name": "ctlplane",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "192.168.122.120",
                              "start": "192.168.122.100"
                            },
                            {
                              "end": "192.168.122.200",
                              "start": "192.168.122.150"
                            }
                          ],
                          "cidr": "192.168.122.0/24",
                          "gateway": "192.168.122.1",
                          "name": "subnet1"
                        }
                      ]
                    },
                    {
                      "dnsDomain": "internalapi.example.com",
                      "name": "internalapi",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "172.17.0.250",
                              "start": "172.17.0.100"
                            }
                          ],
                          "cidr": "172.17.0.0/24",
                          "name": "subnet1",
                          "vlan": 20
                        }
                      ]
                    },
                    {
                      "dnsDomain": "external.example.com",
                      "name": "external",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "10.0.0.250",
                              "start": "10.0.0.100"
                            }
                          ],
                          "cidr": "10.0.0.0/24",
                          "gateway": "10.0.0.1",
                          "name": "subnet1"
                        }
                      ]
                    },
                    {
                      "dnsDomain": "storage.example.com",
                      "name": "storage",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "172.18.0.250",
                              "start": "172.18.0.100"
                            }
                          ],
                          "cidr": "172.18.0.0/24",
                          "name": "subnet1",
                          "vlan": 21
                        }
                      ]
                    },
                    {
                      "dnsDomain": "storagemgmt.example.com",
                      "name": "storagemgmt",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "172.20.0.250",
                              "start": "172.20.0.100"
                            }
                          ],
                          "cidr": "172.20.0.0/24",
                          "name": "subnet1",
                          "vlan": 23
                        }
                      ]
                    },
                    {
                      "dnsDomain": "tenant.example.com",
                      "name": "tenant",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "172.19.0.250",
                              "start": "172.19.0.100"
                            }
                          ],
                          "cidr": "172.19.0.0/24",
                          "name": "subnet1",
                          "vlan": 22
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "network.openstack.org/v1beta1",
                "kind": "Reservation",
                "metadata": {
                  "name": "edpm-compute-0"
                },
                "spec": {
                  "ipSetRef": {
                    "name": "edpm-compute-0",
                    "namespace": "openstack",
                    "uid": "56753ffb-e486-43ab-a79a-f85f27b2ac14"
                  },
                  "reservation": {
                    "CtlPlane": {
                      "address": "192.168.122.100",
                      "network": "CtlPlane",
                      "subnet": "subnet1"
                    },
                    "InternalApi": {
                      "address": "172.17.0.100",
                      "network": "InternalApi",
                      "subnet": "subnet1"
                    },
                    "Storage": {
                      "address": "172.18.0.100",
                      "network": "Storage",
                      "subnet": "subnet1"
                    },
                    "StorageMgmt": {
                      "address": "172.20.0.100",
                      "network": "StorageMgmt",
                      "subnet": "subnet1"
                    },
                    "Tenant": {
                      "address": "172.19.0.100",
                      "network": "Tenant",
                      "subnet": "subnet1"
                    }
                  }
                }
              },
              {
                "apiVersion": "rabbitmq.openstack.org/v1beta1",
                "kind": "TransportURL",
                "metadata": {
                  "name": "transporturl"
                },
                "spec": {
                  "rabbitmqClusterName": "rabbitmq"
                }
              },
              {
                "apiVersion": "redis.openstack.org/v1beta1",
                "kind": "Redis",
                "metadata": {
                  "name": "redis"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:45Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSData is the Schema for the dnsdata API
            displayName: DNSData
            kind: DNSData
            name: dnsdata.network.openstack.org
            version: v1beta1
          - description: DNSMasq is the Schema for the dnsmasqs API
            displayName: DNSMasq
            kind: DNSMasq
            name: dnsmasqs.network.openstack.org
            version: v1beta1
          - description: InstanceHa is the Schema for the instancehas API
            displayName: Instance Ha
            kind: InstanceHa
            name: instancehas.instanceha.openstack.org
            version: v1beta1
          - description: IPSet is the Schema for the ipsets API
            displayName: IPSet
            kind: IPSet
            name: ipsets.network.openstack.org
            version: v1beta1
          - description: Memcached is the Schema for the memcacheds API
            displayName: Memcached
            kind: Memcached
            name: memcacheds.memcached.openstack.org
            version: v1beta1
          - description: NetConfig is the Schema for the netconfigs API
            displayName: Net Config
            kind: NetConfig
            name: netconfigs.network.openstack.org
            version: v1beta1
          - description: Redis is the Schema for the redises API
            displayName: Redis
            kind: Redis
            name: redises.redis.openstack.org
            version: v1beta1
          - description: Reservation is the Schema for the reservations API
            displayName: Reservation
            kind: Reservation
            name: reservations.network.openstack.org
            version: v1beta1
          - description: TransportURL is the Schema for the transporturls API
            displayName: Transport URL
            kind: TransportURL
            name: transporturls.rabbitmq.openstack.org
            version: v1beta1
        description: OpenStack Infrastructure
        displayName: OpenStack Infra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Infrastructure
        links:
        - name: Infra Operator
          url: https://github.com/openstack-k8s-operators/infra-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/infra-rhel9-operator@sha256:177077029ad8baa959e41a785cd29ed63745e8f39be618aa97509e6867fbff7c
        - registry.redhat.io/rhoso/openstack-memcached-rhel9@sha256:962a314226009a8ca5c95fb66a75b15176506e02110847d0f42aeed6dd26a751
        - registry.redhat.io/rhoso/openstack-redis-rhel9@sha256:b83d23df9f2b367e604f404d97574fd6db9e0ed530eda23e9e0bbe01292c8db8
        - registry.redhat.io/rhoso/openstack-neutron-server-rhel9@sha256:806e5e009a77a29c82699487e539669e1ec2136514a0332d222b5cb3fed6f452
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: infra-operator.v1.0.6
        version: 1.0.6
      - name: infra-operator.v1.0.5
        version: 1.0.5
      - name: infra-operator.v1.0.4
        version: 1.0.4
      - name: infra-operator.v1.0.3
        version: 1.0.3
      - name: infra-operator.v1.0.2
        version: 1.0.2
      - name: infra-operator.v1.0.1
        version: 1.0.1
      - name: infra-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: infra-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: maximilianopizarro
      provider-url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
    name: jhipster-online-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jhipster-online-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maximilianopizarro.github.io/v1alpha1",
                "kind": "JhipsterOnline",
                "metadata": {
                  "name": "jhipsteronline-sample"
                },
                "spec": {
                  "affinity": {},
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "env": {
                    "APPLICATION_GITHUB_CLIENT-ID": "CLIENT-ID",
                    "APPLICATION_GITHUB_CLIENT-SECRET": "CLIENT-SECRET",
                    "APPLICATION_GITHUB_HOST": "https://github.com",
                    "APPLICATION_JHIPSTER-CMD_CMD": "jhipster-quarkus",
                    "JAVA_APP_JAR": "jhonline-2.33.0.war",
                    "SPRING_DATASOURCE_PASSWORD": "jhipster",
                    "SPRING_DATASOURCE_URL": "jdbc:mariadb://mariadb:3306/jhipsteronline",
                    "SPRING_DATASOURCE_USERNAME": "jhipster"
                  },
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "quay.io/maximilianopizarro/jhipster-online",
                    "tag": "quarkus"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "livenessProbe": {
                    "httpGet": {
                      "path": "/jdl-studio/",
                      "port": 8080
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "readinessProbe": {
                    "httpGet": {
                      "path": "/jdl-studio/",
                      "port": 8080
                    }
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "securityContext": {},
                  "service": {
                    "port": 8080,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automount": false,
                    "create": false,
                    "name": ""
                  },
                  "tolerations": [],
                  "volumeMounts": [],
                  "volumes": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/maximilianopizarro/jhipster-online-operator:v0.1.0
          createdAt: "2025-05-30T22:42:19Z"
          description: JHipster Online provides an intuitive web interface for generating
            JHipster projects, allowing you to define your application architecture
            using a JDL (JHipster Domain Language) model and generate the code directly
            into a GitHub repository.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/maximilianoPizarro/jhipster-online-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: JhipsterOnline
            name: jhipsteronlines.maximilianopizarro.github.io
            version: v1alpha1
        description: |
          JHipster Online Operator: This repository hosts the JHipster Online Operator, designed to seamlessly deploy JHipster Online on Red Hat OpenShift and Kubernetes environments.

          JHipster Online provides an intuitive web interface for generating JHipster projects, allowing you to define your application architecture using a JDL (JHipster Domain Language) model and generate the code directly into a GitHub repository.

          Overview: The JHipster Online Operator manages the deployment and lifecycle of the JHipster Online application. It leverages the power of Kubernetes Operators to provide a native Kubernetes experience for JHipster Online users.

          This artifact includes:
          - JHipster 8.8.0. for generate Spring Boot 3.4.1 projects.
          - generator-jhipster-quarkus 3.4.0 for generate Quarkus 3.11.1 projects.
          - JDL Studio for add JDL models files by PR on your repo.

          Usage: Once the operator is running, you can create instances of JHipster Online by defining `JhipsterOnline` custom resources. An example `JhipsterOnline` custom resource is provided in the `alm-examples`. You would apply this (or a similar) YAML file to your cluster to create an instance of JHipster Online. Feel free to customize the `JhipsterOnline` Custom Resource Definition (CRD) instance to suit your needs, such as configuring GitHub integration, database connections, and resource allocations.

          Configuration: The `JhipsterOnline` custom resource supports various configuration options including `replicaCount`, `image`, `imagePullSecrets`, `serviceAccount`, `podAnnotations`, `podLabels`, `podSecurityContext`, `securityContext`, `service` type and port, `ingress` settings (enabled, `className`, `annotations`, `hosts`, `tls`), OpenShift `Route` (enabled), `autoscaling` (enabled, `minReplicas`, `maxReplicas`, `targetCPUUtilizationPercentage`), `resource` requests and limits, `nodeSelector`, `tolerations`, `affinity`, `liveness` and `readiness` probes, `env` variables, `volumes`, and `volumeMounts`. Refer to the detailed `README.md` for a full table of parameters.
        displayName: JHipster Online Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - java
        - spring-boot
        - jhipster
        - quarkus
        links:
        - name: Jhipster Online Operator
          url: https://maximilianopizarro.github.io/jhipster-online-operator/
        maintainers:
        - email: maximiliano.pizarro.5@gmail.com
          name: Maximiliano Pizarro
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: maximilianopizarro
          url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/maximilianopizarro/jhipster-online-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: jhipster-online-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: jhipster-online-operator
    provider:
      name: maximilianopizarro
      url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Skupper Project
      provider-url: https://skupper.io
    name: skupper-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: alpha
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1
    - currentCSV: skupper-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:e5f8b2ddf378cf3e40f4c7b7ec8482eb3321947cd46481c868d727932c449d78
          createdAt: "2024-02-26T18:22:07Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/config-sync@sha256:659fe28d1df7c99bba3bb854b1e9991c715ae4de99d91c88d5db5e4f9ef4c852
        - quay.io/skupper/flow-collector@sha256:3870c5bc89e623b7cebe21856bfefc22de76313705efa0ec551a5963789fff3b
        - quay.io/skupper/site-controller@sha256:e5f8b2ddf378cf3e40f4c7b7ec8482eb3321947cd46481c868d727932c449d78
        - quay.io/prometheus/prometheus@sha256:d2ab0a27783fd4ad96a8853e2847b99a0be0043687b8a5d1ebfb2dd3fa4fd1b8
        - quay.io/openshift/origin-oauth-proxy@sha256:a7dff785d821acc4b972d94964947713fade45f007dedb5deee17a2869bf5d24
        - quay.io/skupper/skupper-router@sha256:bf8d5955ce4e406275a4934278efe51c1af9a1a269a70e8fe642b7d94d80cb81
        - quay.io/skupper/service-controller@sha256:345edf2d9d72b5a97bcce0e8cea8cda82d120f7bf83756d73f9ac114c2fcced8
        version: 1.6.0
      entries:
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.6
    - currentCSV: skupper-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:4809db7263d77a95797061798e910f1bdbb5d5fe2db2bb3c743b97f5d0c7b3af
          createdAt: "2024-06-12T14:36:59Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/prometheus/prometheus@sha256:d2ab0a27783fd4ad96a8853e2847b99a0be0043687b8a5d1ebfb2dd3fa4fd1b8
        - quay.io/openshift/origin-oauth-proxy@sha256:a7dff785d821acc4b972d94964947713fade45f007dedb5deee17a2869bf5d24
        - quay.io/skupper/skupper-router@sha256:5eb258a0927e7706e7c035f1def2eb5211c8c33dd371fafa7210517fd894f68b
        - quay.io/skupper/service-controller@sha256:624e88e695297df28e485db7e802945ad1978ac6b0c9e9edca8b5e6013af8991
        - quay.io/skupper/site-controller@sha256:4809db7263d77a95797061798e910f1bdbb5d5fe2db2bb3c743b97f5d0c7b3af
        - quay.io/skupper/config-sync@sha256:a884b26a32179af5a2bfa85118bf873852eec6e2930e92da778b57506c0a0880
        - quay.io/skupper/flow-collector@sha256:bf646b0cb75e1c29bbe0261aa9943e06d33118f5e6a5ccb7503c60c974167692
        version: 1.7.1
      entries:
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.7
    - currentCSV: skupper-operator.v1.8.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:ffe3ea5067d350b64b69120bdf6d9e3ee9a30e8a2559b77170a10d66635623b8
          createdAt: "2025-03-27T20:11:53Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/service-controller@sha256:a45de211b8f8a8294a3588932a0a72766cf0d5ce64b01f4f744832c8d09aaa46
        - quay.io/skupper/config-sync@sha256:9e96d6a585f037f60ca3d89b5efee27ec92444d353e8b289ca35b94aeed1e309
        - quay.io/skupper/flow-collector@sha256:252244ab6aaf6b80118bfc06b690d4f5e8d7829b999fca60d194a84452fee7f6
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/openshift/origin-oauth-proxy@sha256:c15baf57dbd17dcc517cd2fa2299462d62124f689d6f5a48a3c9fd154fc77ccf
        - quay.io/skupper/skupper-router@sha256:7633b143f1844ee559322e4642ca80e8248a08711c8e814d489c26c39ecf9bd0
        - quay.io/skupper/site-controller@sha256:ffe3ea5067d350b64b69120bdf6d9e3ee9a30e8a2559b77170a10d66635623b8
        version: 1.8.4
      entries:
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.8
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.9
    defaultChannel: stable
    packageName: skupper-operator
    provider:
      name: Skupper Project
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Harness Inc.
      provider-url: https://harness.io
    name: harness-gitops-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: harness-gitops-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gitops.harness.io/v1alpha1",
                "kind": "GitopsAgent",
                "metadata": {
                  "annotations": {
                    "containerImage": "harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde",
                    "createdAt": "2025-07-31T12:11:18Z"
                  },
                  "name": "gitops-agent-sample"
                },
                "spec": {
                  "identity": {
                    "accountIdentifier": "accountIdentifier",
                    "agentIdentifier": "agentIdentifier",
                    "agentName": "agentName"
                  },
                  "replicas": 1,
                  "secrets": {
                    "agentSecret": "agentSecret"
                  },
                  "target": {
                    "http": "https://app.harness.io/gitops/",
                    "protocol": "HTTP1"
                  },
                  "upgrade": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde
          createdAt: "2025-07-31T06:41:18Z"
          description: Harness operator to manage and provision GitOps Agent
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Harness Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Schema for the Gitops Agents API
            displayName: Gitops Agent
            kind: GitopsAgent
            name: gitopsagents.gitops.harness.io
            version: v1alpha1
        description: |-
          ## Overview
          Harness Gitops Operator manages the full lifecycle for [Gitops Agent](https://developer.harness.io/docs/continuous-delivery/gitops/connect-and-manage/install-a-harness-git-ops-agent/)

          The operator aims to provide the following:

            * Easy configuration and installation of Harness GitOps Agent.
            * Provide seamless upgrades for Harness GitOps Agent.
            * Monitoring and autorecovery for GitOps Agent.

          ## Usage
          Deploy gitops-agent in namespace where operator is installed with basic setup.

          ### Prerequisites:
            * Go to your Harness account and create a new gitops agent
            * Copy account, org and project identifiers
            * Retrieve the agent secret key

          ```
          apiVersion: gitops.harness.io/v1alpha1
          kind: GitopsAgent
          metadata:
            name: gitops-agent-sample
          spec:
            identity:
              accountIdentifier: accountIdentifier
              agentIdentifier: agentIdentifier
              agentName: agentName
            secrets:
              agentSecret: agentSecret
            target:
              http: https://app.harness.io/gitops/
              protocol: HTTP1
            replicas: 1
            upgrade:
              enabled: true
          ```
          **Note:** Gitops Agent require an active ArgoCD instance running in the same namespace where agent is installed.

          See [Harness Documentation](https://developer.harness.io/) for more information.
        displayName: Harness GitOps Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - harness
        - agent
        - cd
        - argocd
        - kubernetes
        links:
        - name: Harness Gitops Operator
          url: https://harness-gitops-operator.domain
        maintainers:
        - email: alexandru.bosii@harness.io
          name: Alexandru Bosii
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Harness Inc.
          url: https://harness.io
        relatedImages:
        - harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde
        - docker.io/harness/gitops-agent@sha256:c9f57338be294d32a73a4cbecf90dfff50f5d36a699f3ec32d962dcb64d8d085
        version: 0.3.0
      entries:
      - name: harness-gitops-operator.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: harness-gitops-operator
    provider:
      name: Harness Inc.
      url: https://harness.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataTrucker.IO
      provider-url: https://www.datatrucker.io/
    name: datatrucker-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datatrucker-operator.v2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datatrucker.datatrucker.io/v1",
                "kind": "DatatruckerConfig",
                "metadata": {
                  "name": "datatruckerconfig-sample"
                },
                "spec": {
                  "API": {
                    "crypto": {
                      "CredentailEncryption": {
                        "Algorithm": "aes-256-cbc",
                        "IV": "eThWmZq4t7w!z$C\u0026",
                        "Key": "JaNdRgUkXp2s5u8x/A?D(G+KbPeShVmY"
                      },
                      "JWT": {
                        "authorization_keys": {
                          "private_key": "privateJWT.key",
                          "public_key": "publicJWT.key"
                        },
                        "signOptions": {
                          "algorithm": "RS256",
                          "expiresIn": "60m",
                          "issuer": "datatrucker"
                        },
                        "verify": {
                          "issuer": "datatrucker"
                        }
                      },
                      "keycloak": {
                        "client_id": "nodejs",
                        "client_secret": "7c5542ae-7264-420d-bf5f-766f4ab39ad3",
                        "jwt_publickey_verifier": "keycloak.key",
                        "realm": "TestRealm",
                        "url": "http://localhost:30015"
                      },
                      "passwordHash": {
                        "Algorithm": "sha256",
                        "SaltLength": 16
                      }
                    },
                    "cryptokeys": {
                      "keycloakkey": "LS0tLS1CRUdJTiBQVUJMSUMgS0VZLS0tLS0KTUlJQklqQU5CZ2txaGtpRzl3MEJBUUVGQUFPQ0FROEFNSUlCQ2dLQ0FRRUF4ZnhGY1gxb3RmTUN6SWlkZ09tMndZakZWZXZHN1BQT1h5dTNrTEFuTWtTVkUxUXl4TzJ4SFFvSGcydDFVcUU4SVNXeTN3ZkZTbERzdU1VK0J1VVkvaWI5L04vSlpBV3k5MGRET0YxREdIeGVwM21HMVBTUllMQWFGQ0I3NTVsbWZjL1dna3VzZFdLdmxOTmZuNkd6S1J6OXBmV3RNUnhWWkFrejhmNlVUeG0zWE02Q0tOMUludThubEcxU3ZzaFc5cmt3Lyt6UmVvOFVCV2pPcTh5RjdPV0tnS0ZZZlhENFREa090OWFxN3RaaSsxaExYOE5iV2xxRnJZYmhJcFZINXlLSko3ZTA5cUVIRkVEelg5QXJzVVNHamNjbFFFYTl1dmJxU2d4Rzc3RUZVbk82aDZxUkNnU3lNNnVYdk9qcWlIMHp4RTNmRHB4R2NhY2pvTW1EMlFJREFRQUIKLS0tLS1FTkQgUFVCTElDIEtFWS0tLS0t",
                      "privateJWTkey": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQ0KTUlJRW9RSUJBQUtDQVFCVWd1ZDN1OFlXcUJCWWtYZ3M2K25pMS84bnM5YjRGNkZudzA4RDRKdGRlalg5L1ZaMg0KWFdOamJ1ekMxUkRHMnVseFM4QjZIdTZYeEhYWU9XU2ZzOEY0UHZ2eC9TUWdmRWtWUlNDL1pQN3pzM1ZETllaMA0Kc1JScDNSZzUza29qUG80T3o2azFaeGprbFRFZEpMWUtqYlZud3U1M3A3MTFsVTJkV3B4R3VEeGYxVTJuUWgrWA0KaUhKZkhyUmxTZVA1aVhYQkpxd21VTDFkK25lTU5BOXlGakR0MnNZMG5iSDRySDVxbk1VRk4xMEpsZ3V4NFdiUw0KaGFUNklNN3h1dUVHbDRDckVENmdTejBNNGF1UVk4ekZxQWxJNlFWUTlML28zQnBCRmxYUUMwWmowb3lwQVFKLw0KQ1VLaU9FcEhLRHgram02RU9KOE1aejlMQ01yTmdLME5KSWlaQWdNQkFBRUNnZ0VBRTBMMnNLSjZob1Z0bkZJag0KZXRHV01CNmN5QmJFSXkxM215ZTNwYUdXcDVxSUhjaVU0VXR0c0tjQ2pYRjBibTVNa0liOGFwa01Oa2tQdHVWMQ0KMnJSa254eHozZmFkT2s0NWRnQ1ljeWI0dU9DcEZGVnB2TVR6YXlGNk9RNUNCWUJuYmxiRU1TaTdVclh5U0xzcQ0KUm9OLzJoTml1S0RobktBWDNZcGVYZnYxdnIyaWIrMGtiRDVvMVZLcnB6SlJ5emlsWHFwQWJwNDQ4RExFZHVGeA0KTnJKc0tWb1dXMVhmS1FvR2VDMXZqNFJudzdhWkNyU1BjMEhvMUdKb29xSXlrNjRXZnhGWjhpVElsWDJMU1gzbQ0KQm1pN1ZXZnVJK0QvUXRtWEVpN05oTE1jUmExVXB4QkdiK1ZPZXlucEVyMEpxM3VZSmlva1VIZzloOGIzSDM1Yg0KK0lJaEFRS0JnUUNtbzUrRUNpOEJUcFRHdlpqQXBtdmxGZ055NkpWd0RvVHJLSGdCcHJHVHpsVGt6Q0tLYzJhTw0KTTMvQmVIVUQ1NVFqcTdUM094eGgzQUk3cG9kT0tCWS84eXgzenlYUExQUmVCMlR0eTNVQjE5YU5CclhNUGQ5RQ0KTG5hbUQ0SmZQeTV0WHpWWEloWVB0UzBiNjdGbjZnZFl4MlNBbGFYYWFjejU4eURLZnlZL3VRS0JnUUNCMUxNYg0KTFZNTU1Pd1U4R2xsVVB1V0Q5VnBsMnlURytCN2Nnd0ZiYTlrWVlVSFQ4TmlvaUdxQzlLZWxqeklKNnlUdyt3aA0KQisvNGNJU1o5Y0JDNHB0VnRlTWJwUkg3UTQwK20xU3M1SmQzU1VFcDBMOW5VbmdVNDhrMzlpSTV5aEgxU3JTRw0KREtBamVka2Rjb3hNVWFwUHB6cVprOVNKdGVHckVsTHY5WmcvNFFLQmdHcExQTmt3MWNMenFSc01uWDlscUxVOA0KVVg3ZXRWUUJMdkFCcEdzSk5ZU2V2TEFsTm5NeUlWemJwTVo0Vzh2ZDBEVGd0bEp0dGpFRjhiYXpCdVBuMWlsOA0KcUFUVWNiMnhZejhIWUczK3RlZ3pFWHhycDFOY3RDbjlxVmxMMDFHR0w4czhTLzRFbENYNDM0bEwwcUlDT2lWNw0KS3JTS3dIeU51RjVIbzhjSFRCS2hBb0dBVjhVRVpENEs0N1g3RTlqSFY3QWN0UUQxUkpJUzI0ekFBSlJFSGpOZA0KbkIzMVFkRWhoSCtaQkkxRGtPUitLTTN4ZEYvQ05UNVpjM29WbXNPeGUwWVl2eVBjYWViNU5nVWY2cEExK1QvOA0KU2E4ak4zS2d3dWZVdUVLWjJKUXo1UjkvYm5WWGZ6dnYwcXZnSkhFNCthdEgyTnR6MlRsQU9PelZwOElXZ1pSbA0KbmVFQ2dZQXpVS0RLS1ZKaXBYdC9TRC8xd2ZzWElCazFxYWpmdENHOFVTcFlGWGp5VzVYMlZ5bTlVTldmVEdJcA0KQlJkVldXamdFbWhyOU5Da1UrVEU4WEVjTXJNRnhWbXBEeTQ4YmxsQWltWFJtOEhNMlhKODQ4TlArTG5GdEtHdA0KTWswT3JuUHdPYitualZGcXdWUllNQ0QzeFN0Z2RnZUppUHFYcWd3SDhjR0M0OGVmZ0E9PQ0KLS0tLS1FTkQgUlNBIFBSSVZBVEUgS0VZLS0tLS0=",
                      "publicJWTkey": "LS0tLS1CRUdJTiBQVUJMSUMgS0VZLS0tLS0NCk1JSUJJVEFOQmdrcWhraUc5dzBCQVFFRkFBT0NBUTRBTUlJQkNRS0NBUUJVZ3VkM3U4WVdxQkJZa1hnczYrbmkNCjEvOG5zOWI0RjZGbncwOEQ0SnRkZWpYOS9WWjJYV05qYnV6QzFSREcydWx4UzhCNkh1Nlh4SFhZT1dTZnM4RjQNClB2dngvU1FnZkVrVlJTQy9aUDd6czNWRE5ZWjBzUlJwM1JnNTNrb2pQbzRPejZrMVp4amtsVEVkSkxZS2piVm4NCnd1NTNwNzExbFUyZFdweEd1RHhmMVUyblFoK1hpSEpmSHJSbFNlUDVpWFhCSnF3bVVMMWQrbmVNTkE5eUZqRHQNCjJzWTBuYkg0ckg1cW5NVUZOMTBKbGd1eDRXYlNoYVQ2SU03eHV1RUdsNENyRUQ2Z1N6ME00YXVRWTh6RnFBbEkNCjZRVlE5TC9vM0JwQkZsWFFDMFpqMG95cEFRSi9DVUtpT0VwSEtEeCtqbTZFT0o4TVp6OUxDTXJOZ0swTkpJaVoNCkFnTUJBQUU9DQotLS0tLUVORCBQVUJMSUMgS0VZLS0tLS0=",
                      "truckerapicrt": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSUQxVENDQXIyZ0F3SUJBZ0lVUFVzZ2NYVXZRYWtuQldjeUFDeEw2dmt3b204d0RRWUpLb1pJaHZjTkFRRUwKQlFBd2RqRUxNQWtHQTFVRUJoTUNRMEV4Q3pBSkJnTlZCQWdNQWs5T01ROHdEUVlEVlFRSERBWk5hV3gwYjI0eApGekFWQmdOVkJBb01Ea1JoZEdGVWNuVmphMlZ5TGtsUE1SUXdFZ1lEVlFRTERBdEZibWRwYm1WbGNtbHVaekVhCk1CZ0dBMVVFQXd3UlJHRjBZVlJ5ZFdOclpYSlNiMjkwUTBFd0hoY05NakF4TURFeU1UWXlOalV5V2hjTk1qTXcKTVRFMU1UWXlOalV5V2pCM01Rc3dDUVlEVlFRR0V3SkRRVEVMTUFrR0ExVUVDQXdDVDA0eER6QU5CZ05WQkFjTQpCazFwYkhSdmJqRVhNQlVHQTFVRUNnd09SR0YwWVZSeWRXTnJaWEl1U1U4eEZEQVNCZ05WQkFzTUMwVnVaMmx1ClpXVnlhVzVuTVJzd0dRWURWUVFEREJKa1lYUmhkSEoxWTJ0bGNpMWhjR2t1YVc4d2dnRWlNQTBHQ1NxR1NJYjMKRFFFQkFRVUFBNElCRHdBd2dnRUtBb0lCQVFEakM4ZmVCVEg4L3NnejFQcTJQYU9QQjc5YWtyeUZvWVpGV292bQpWOXNvMmJNRHp4NVBZSFJydVFGYmJQRzlVVlc2QWZpSHdEb1p2Z3NTVXpyS3ZyR1RlYjluNUhSVlVOMTl0NExTCjdtUFppNzFRbGE4bVR3QmdvTnE4TFppcnZUdzV0S2I4K3NmcFJxQUlUQkVsRmxVUEZISzllcDl2eFdvaHhwdGgKS1dPVkwvV2ZtWldhWXlRdjlLNDBkWUJpRWpGcTFKZVB4MDlnUFdHM2pOT1VYTTZQUVlMaHJDUXFFcy9VME9GWQpwVFRHUmNXZEx4TFVFZVNkcHJKK1ZEc3lEK1lVeUFRR2phYk14VmVTVlRUN1FGbFBjLzc3OVZsRkdXMC8xWVVaCmpqakNsSkFLUWZOTHg5bG85cWJERVBZL3grZmhnRnYyWUVaSTNTdGVyZWUvVDJoWEFnTUJBQUdqV2pCWU1COEcKQTFVZEl3UVlNQmFBRkNJVy9JYkZIUUNBeEdxZC9FVFpXbVhGeGMvQ01Ba0dBMVVkRXdRQ01BQXdDd1lEVlIwUApCQVFEQWdUd01CMEdBMVVkRVFRV01CU0NFbVJoZEdGMGNuVmphMlZ5TFdGd2FTNXBiekFOQmdrcWhraUc5dzBCCkFRc0ZBQU9DQVFFQUxrMlJDbnVweDA5Wmx0SmN3M1pUaFIveUZGNS9KVlVsWkdHTnc3MlAwQmR6SEloQncrVk4KRkNucGFPMjJ0WGF1NlJVeTVNUWhwc3JrbFNKYVdJT25jOUg0Uk0vbGh3YWsyRXN0MkFhdUFwNUdYSVFJMDVBeAo3NmtDZXEyMnUyalNXRWpFS1hGRnRiZXVDZnFpWlpqNWFLcGF1YlZWRG4ydEZTWjJRMFRib2RRN2dpTFdNUTBLCk5ycVgwQWNSU2J0ZnRSNjBhdGQ1YW41aFNXY0dyZUg5ZHdFUVRvVVNFblg2RFRZSURoeElsZStjQ2V5VjVSWk8KVS9hck5RZ0lFY3ZrZUM1d1lZaFFFL1FUN3lNb1lNajFnckxmOTZSZkRhbEVFNE1seUtDWm9wYWNJWUdzSmhuKwpLMWFISEQvbGZxTC9zbnNtY3pXdHlLUncxTDF1WkxPcjJnPT0KLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=",
                      "truckerapikey": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQpNSUlFb3dJQkFBS0NBUUVBNHd2SDNnVXgvUDdJTTlUNnRqMmpqd2UvV3BLOGhhR0dSVnFMNWxmYktObXpBODhlClQyQjBhN2tCVzJ6eHZWRlZ1Z0g0aDhBNkdiNExFbE02eXI2eGszbS9aK1IwVlZEZGZiZUMwdTVqMll1OVVKV3YKSms4QVlLRGF2QzJZcTcwOE9iU20vUHJINlVhZ0NFd1JKUlpWRHhSeXZYcWZiOFZxSWNhYllTbGpsUy8xbjVtVgptbU1rTC9TdU5IV0FZaEl4YXRTWGo4ZFBZRDFodDR6VGxGek9qMEdDNGF3a0toTFAxTkRoV0tVMHhrWEZuUzhTCjFCSGtuYWF5ZmxRN01nL21GTWdFQm8ybXpNVlhrbFUwKzBCWlQzUCsrL1ZaUlJsdFA5V0ZHWTQ0d3BTUUNrSHoKUzhmWmFQYW13eEQyUDhmbjRZQmI5bUJHU04wclhxM252MDlvVndJREFRQUJBb0lCQUZSbFo2aiswMk1RcmdpZQpUWWQ2aUJhOXNZS1FjdVZWY2x6TFdKK0ozdHhqZCtpZGo4Mi9zK2JyQVRjajJ3UWRmTGxFekVaaHpQY09ybnprClZVRlIreDR2cmxlSEV6czNjTzJ2VGtEci9rKzZZdWZXakNHNFVoS1pHRlhFdVRCMWlqS0ZGV0dLTXM5aXdrL3oKUHBoRG9lQ1dJRVNRbFQxdDY1S2NpZHRuRUM0K3oxaURxR21ZZEVEa2xEZFFIVHorRGpseS80VzV2MFpVQWN2Vwo1Unkvd0JldmNURlBXQy9uZGFrN3hGUnVrU2xNZUlhTWhvTlllOU5uVCt2Nmh5VVh1MXp6cG45UEFJRTVpQWpOCnBsZkFEZitwakwrZFJDbW9sWW1LR3NQVjEzSmtCY1p2cEVoUWgrUmJnaWFwczJLc2plQ0NmTmtST0hndHc0MU0KeGxOTU1Ha0NnWUVBOHM1akpuZ2FMblIvTzZtakZEb0wvd1p2UWdnT1h3Z1RBN2tsQ0trbjBYM0xJS0hhaTZESgpyOVRGQktMdkhUazVFK1Y4NkhIZ29VMUFQQnYyMWJpaHhuQXhUYm8wSHZMdG5QOXU3YnZJejFNOWMyd2ZlWUgzCmliMG1FNkV0Y2wvNnVIUHdMVlZtRGh5bllneEZpeTRRRXp1ZHZwaFIvSnlPazQvUytQUzh2ZXNDZ1lFQTcySW8KWDlkY0xvUmYzSk0vRTF1SXhrRGRLZm1VakN6Mk1NN3pyRkpNUERobTFSUlJhQjdidHNxUzl4Mm5KeTI5amtoSQpndTR0Y2s5STdncm9WRXY0cEtkU0lOVzVNVUg0MXJPckxTbFBobCtMaTM5QXNrTERFVXVQbkFmYko3YWZibFFHCjNFWE1Ka1d2WFQxRG90aE11MUpUMTFPMjhmYlBmb3R3YVRucXFFVUNnWUE1RnArblh0dkJDbkVUNVJuTDJxNFIKWTBEcmhKdGthckNWMG9WTG1RSWN0bmY1RHpVa1dVU3lpOU5hK0xoQjViOXUxeWxjT2RJMytzTGNuektWSTVrZAo0Q2tuZmwxYzNYOVN4NWdjVzJlbW40cTJSYmpNeCtzY3ZwVmlUaVl5V3pESEI3Tno1VDFlbWd3YytJZXRac0l1CjAzaU9aSi82UmhFYnp0Z3E0dGd4eXdLQmdCanhwSDFnSm9YZVcxTFRzUlhtU3RsRnR3WUJzTDN1alhUR2k5T2YKKzd1bXNPaHhKbzV2OExzYVhyZ0xTZmltSm9ObCt0QlIyMEIzRElWcGM0NGRJSHFmeU1vTkd6QUx6TThxaDZLOApGRUpQUUc0M3cxa1A1ZDJ1ZnZkbVJpcmNONzY2V01DcE1UMm5zejlHKy81REJiUGJQMHRIdU45N3RXRmRURWhhCnRaMzFBb0dCQU5xZnVycEV4ZHozdXRNUm91RnJjTmVkV2s4VlVpVVZIV1hZMjhNNWQvRjcxMExIOTZuOW12V08Kc1cwcDFjcXBJMEsyQko2WlYwNVpybnlyRHBsVVpVdVpSQWxGR3hYR1dLU2trQzlmWm90SkFOWDBkR29KSitqTwpEYUNLNGxlcXdOMHNPRE1YZFErMlFPaUtlSUxkNU5KcDNUSW1YV1Z0RFlXUkRwOW9iSUdlCi0tLS0tRU5EIFJTQSBQUklWQVRFIEtFWS0tLS0tCg=="
                    },
                    "jobsServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "debug"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": true,
                        "FileSFTP": true,
                        "IOTKafka": true,
                        "IOTProxy": false,
                        "IOTRedis": true,
                        "Login": false,
                        "ScriptJS": true,
                        "ScriptSSH": true,
                        "ScriptShell": true,
                        "UtilEcho": true,
                        "UtilFuzzy": true,
                        "UtilSentiment": true,
                        "credentials": false,
                        "jobs": true,
                        "resources": false,
                        "statuschecks": true,
                        "ui": false,
                        "usermanagement": false
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "loginServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "info"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": false,
                        "FileSFTP": false,
                        "IOTKafka": false,
                        "IOTProxy": false,
                        "IOTRedis": false,
                        "Login": true,
                        "ScriptJS": false,
                        "ScriptSSH": false,
                        "ScriptShell": false,
                        "UtilEcho": false,
                        "UtilFuzzy": false,
                        "UtilSentiment": false,
                        "credentials": false,
                        "jobs": false,
                        "resources": false,
                        "statuschecks": true,
                        "ui": false,
                        "usermanagement": false
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "managementServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "debug"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": false,
                        "FileSFTP": false,
                        "IOTKafka": false,
                        "IOTProxy": false,
                        "IOTRedis": false,
                        "Login": true,
                        "ScriptJS": false,
                        "ScriptSSH": false,
                        "ScriptShell": false,
                        "UtilEcho": false,
                        "UtilFuzzy": false,
                        "UtilSentiment": false,
                        "credentials": true,
                        "jobs": false,
                        "resources": false,
                        "statuschecks": true,
                        "ui": true,
                        "usermanagement": true
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "name": "API",
                    "resource": {
                      "Cache": {
                        "Creds": false,
                        "CredsFlush": 10,
                        "Resource": false,
                        "ResourceFlush": 10
                      },
                      "Templates": {
                        "Scripts": "scripts",
                        "Shell": "shell",
                        "keys": "keys",
                        "resourcedefinitions": "resourcedefinitions"
                      }
                    }
                  },
                  "DB": {
                    "databasename": "userdb",
                    "hostname": "db",
                    "password": "password",
                    "port": 5432,
                    "type": "pg",
                    "user": "testuser"
                  },
                  "TempDB": {
                    "enabled": false,
                    "pvc": "samplepvc"
                  }
                }
              },
              {
                "apiVersion": "datatrucker.datatrucker.io/v1",
                "kind": "DatatruckerFlow",
                "metadata": {
                  "name": "datatruckerflow-sample"
                },
                "spec": {
                  "API": {
                    "Image": {
                      "imageName": "datatrucker-server/datatrucker-api",
                      "imageNameUI": "datatruckerio/datatrucker-ui",
                      "repository": "registry.gitlab.com/datatruckerio",
                      "tagName": "latest"
                    },
                    "name": "api"
                  },
                  "DatatruckerConfig": "datatruckerconfig-sample",
                  "JobDefinitions": [
                    {
                      "name": "echoapi",
                      "restmethod": "GET",
                      "tenant": "Admin",
                      "type": "Util-Echo",
                      "validations": {
                        "properties": {
                          "voila": {
                            "enum": [
                              "true",
                              "false"
                            ],
                            "type": "string"
                          }
                        },
                        "type": "object"
                      }
                    }
                  ],
                  "Keys": {
                    "configmap": "placeholder"
                  },
                  "Replicas": 1,
                  "Resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "256Mi"
                    },
                    "requests": {
                      "cpu": "250m",
                      "memory": "256Mi"
                    }
                  },
                  "Scripts": {
                    "configmap": "placeholder"
                  },
                  "Type": "Job"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/datatruckerio/datatrucker-operator:2.0.0
          createdAt: "2021-09-08 01:01:01"
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: hhttps://gitlab.com/datatruckerio/datatrucker-server.git
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              - creates secrets to hold server configs
              - create tempDB for protoypting
              - create configs for seperating Login, management and Jobs endpoints
              - Create crypto configs
            displayName: Datatrucker Config
            kind: DatatruckerConfig
            name: datatruckerconfigs.datatrucker.datatrucker.io
            version: v1
          - description: '- creates an API as per Job Definitions'
            displayName: Datatrucker Flow
            kind: DatatruckerFlow
            name: datatruckerflows.datatrucker.datatrucker.io
            version: v1
        description: "## DataTrucker.IO\n\nDataTrucker.IO is a simple no-code / less-code
          API Backend\n\n### What does it do ?\n\nDatatrucker.IO is a product capable
          of reading simple json/yaml configs and building th code necessary for converting
          it into an API.\nAlong with building the code, it will also host the code
          base on a nodejs server , i.e. it will immediately make it available for
          consumption.\n\n### What does it offload from a developer ?\n\nDataTrucker
          is capable of removing the most common activities which a developer needs
          to do on every new project A few of such common activities are\n\n- Creating
          an API endpoint with a specified business logic (using simple plugins)\n-
          Applying standard RBAC\n- Applying Authorization logic\n- Applying hardening
          on endpoints\n- Log management\n- Connecting to variety of systems\n- Modularizing
          business logic\n- The best of doing it with little to no code\n\n### watch
          the quick elevator pitch to understand datatrucker.io\n\n https://youtu.be/DKLFDqhjs5M\n "
        displayName: DataTrucker.IO
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - datatrucker
        - operator
        - nocode
        - apibuilder
        links:
        - name: Website
          url: https://www.datatrucker.io/
        - name: Git Repo
          url: https://gitlab.com/datatruckerio/datatrucker-server.git
        maintainers:
        - email: gauravshankar.can@gmail.com
          name: Gaurav Shankar
        maturity: alpha
        provider:
          name: DataTrucker.IO
          url: https://www.datatrucker.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - docker.io/datatruckerio/datatrucker-operator:2.0.0
        version: 2.0.0
      entries:
      - name: datatrucker-operator.v2.0.0
        version: 2.0.0
      - name: datatrucker-operator.v1.3.2
        version: 1.3.2
      - name: datatrucker-operator.v1.3.1
        version: 1.3.1
      - name: datatrucker-operator.v1.3.0
        version: 1.3.0
      name: alpha
    defaultChannel: alpha
    packageName: datatrucker-operator
    provider:
      name: DataTrucker.IO
      url: https://www.datatrucker.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: postgresql
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: registry.developers.crunchydata.com/crunchydata/postgres-operator:ubi9-5.8.3-0
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.developers.crunchydata.com/crunchydata/postgres-operator:ubi9-5.8.3-0
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: postgresql
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Vertica
      provider-url: http://www.vertica.com
    name: verticadb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: verticadb-operator.v1.11.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "EventTrigger",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "verticadb-operator",
                    "app.kubernetes.io/instance": "eventtrigger-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eventtrigger",
                    "app.kubernetes.io/part-of": "verticadb-operator"
                  },
                  "name": "eventtrigger-sample"
                },
                "spec": {
                  "matches": [
                    {
                      "condition": {
                        "status": "True",
                        "type": "DBInitialized"
                      }
                    }
                  ],
                  "references": [
                    {
                      "object": {
                        "apiVersion": "vertica.com/v1beta1",
                        "kind": "VerticaDB",
                        "name": "verticadb-sample"
                      }
                    }
                  ],
                  "template": {
                    "metadata": {
                      "generateName": "create-user-table-"
                    },
                    "spec": {
                      "template": {
                        "spec": {
                          "containers": [
                            {
                              "command": [
                                "/opt/vertica/bin/vsql",
                                "-h",
                                "verticadb-sample-defaultsubcluster",
                                "-c",
                                "CREATE TABLE T1 (C1 INT);"
                              ],
                              "image": "vertica/vertica-k8s:latest",
                              "name": "main"
                            }
                          ],
                          "restartPolicy": "OnFailure"
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "serviceName": "defaultsubcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaDB",
                "metadata": {
                  "name": "verticadb-sample"
                },
                "spec": {
                  "communal": {
                    "endpoint": "https://s3.amazonaws.com"
                  },
                  "image": "vertica/vertica-k8s:23.3.0-0-minimal",
                  "subclusters": [
                    {
                      "name": "default_subcluster",
                      "size": 3
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: docker.io/vertica/verticadb-operator:1.11.2
          createdAt: "2023-08-02T11:29:39Z"
          description: Operator that manages a Vertica Eon Mode database.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/vertica/vertica-kubernetes
          support: Vertica
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventTrigger is the Schema for the eventtriggers API
            displayName: Event Trigger
            kind: EventTrigger
            name: eventtriggers.vertica.com
            version: v1beta1
          - description: VerticaAutoscaler is a CR that allows you to autoscale one
              or more subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1beta1
          - description: VerticaDB is the CR that defines a Vertica Eon mode cluster
              that is managed by the verticadb-operator.
            displayName: Vertica DB
            kind: VerticaDB
            name: verticadbs.vertica.com
            version: v1beta1
        description: |-
          ### What is Vertica?
          Vertica is a unified analytics platform, based on a massively scalable architecture with the broadest set of analytical functions spanning event and time series, pattern matching, geospatial and end-to-end in-database machine learning. Vertica enables you to easily apply these powerful functions to the largest and most demanding analytical workloads, arming you and your customers with predictive business insights faster than any analytics data warehouse in the market. Vertica provides a unified analytics platform across major public clouds and on-premises data centers and integrates data in cloud object storage and HDFS without forcing you to move any of your data.

          ### What does this operator do?
          This operator manages a Vertica Eon Mode database, and automates the following administrative tasks:

          * Installing Vertica
          * Creating and reviving a Vertica database
          * Restarting and rescheduling DOWN pods to maintain quorum
          * Upgrading Vertica to a new version while keeping the database online
          * Subcluster scaling, both manually and automatically with the Horizontal Pod Autoscaler
          * Service management and health monitoring for pods
          * Load balancing for internal and external traffic
          * Expose a metrics endpoint to integrate with Prometheus to monitor the operator's health

          For a brief overview on how to create a cluster, see the [Vertica GitHub repository](https://github.com/vertica/vertica-kubernetes). For an in-depth look at Vertica on Kubernetes, see the [Vertica documentation](https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm).
        displayName: VerticaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - database
        - vertica
        links:
        - name: Documentation
          url: https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm
        - name: Vertica Container Images
          url: https://hub.docker.com/u/vertica
        maintainers:
        - email: matt.spilchen@vertica.com
          name: Matt Spilchen
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Vertica
          url: http://www.vertica.com
        relatedImages:
        - docker.io/vertica/verticadb-operator:1.11.2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 1.11.2
      entries:
      - name: verticadb-operator.v1.11.2
        version: 1.11.2
      - name: verticadb-operator.v1.11.1
        version: 1.11.1
      - name: verticadb-operator.v1.11.0
        version: 1.11.0
      - name: verticadb-operator.v1.10.2
        version: 1.10.2
      - name: verticadb-operator.v1.10.1
        version: 1.10.1
      - name: verticadb-operator.v1.10.0
        version: 1.10.0
      - name: verticadb-operator.v1.9.0
        version: 1.9.0
      - name: verticadb-operator.v1.8.0
        version: 1.8.0
      - name: verticadb-operator.v1.7.0
        version: 1.7.0
      - name: verticadb-operator.v1.6.0
        version: 1.6.0
      - name: verticadb-operator.v1.5.0
        version: 1.5.0
      - name: verticadb-operator.v1.4.0
        version: 1.4.0
      - name: verticadb-operator.v1.3.1
        version: 1.3.1
      - name: verticadb-operator.v1.3.0
        version: 1.3.0
      - name: verticadb-operator.v1.2.0
        version: 1.2.0
      name: stable
    - currentCSV: verticadb-operator.v25.3.1-0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vertica.com/v1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "customAutoscaler": {},
                  "scalingGranularity": "Subcluster",
                  "serviceName": "default-subcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1",
                "kind": "VerticaDB",
                "metadata": {
                  "name": "verticadb-sample"
                },
                "spec": {
                  "communal": {
                    "endpoint": "https://s3.amazonaws.com"
                  },
                  "image": "opentext/vertica-k8s:24.1.0-0-minimal",
                  "subclusters": [
                    {
                      "name": "default_subcluster",
                      "size": 3
                    }
                  ]
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "EventTrigger",
                "metadata": {
                  "name": "eventtrigger-sample"
                },
                "spec": {
                  "matches": [
                    {
                      "condition": {
                        "status": "True",
                        "type": "DBInitialized"
                      }
                    }
                  ],
                  "references": [
                    {
                      "object": {
                        "apiVersion": "vertica.com/v1beta1",
                        "kind": "VerticaDB",
                        "name": "verticadb-sample"
                      }
                    }
                  ],
                  "template": {
                    "metadata": {
                      "generateName": "create-user-table-"
                    },
                    "spec": {
                      "template": {
                        "spec": {
                          "containers": [
                            {
                              "command": [
                                "/opt/vertica/bin/vsql",
                                "-h",
                                "verticadb-sample-defaultsubcluster",
                                "-c",
                                "CREATE TABLE T1 (C1 INT);"
                              ],
                              "image": "opentext/vertica-k8s:23.4.0-0-minimal",
                              "name": "main"
                            }
                          ],
                          "restartPolicy": "OnFailure"
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "customAutoscaler": {},
                  "scalingGranularity": "Subcluster",
                  "serviceName": "default-subcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaReplicator",
                "metadata": {
                  "name": "verticareplicator-sample"
                },
                "spec": {
                  "source": {
                    "verticaDB": "sourceDBName"
                  },
                  "target": {
                    "verticaDB": "targetDBName"
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaRestorePointsQuery",
                "metadata": {
                  "name": "verticarestorepointsquery-sample"
                },
                "spec": {
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaScrutinize",
                "metadata": {
                  "name": "verticascrutinize-sample"
                },
                "spec": {
                  "verticaDBName": "verticadb-sample"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: docker.io/opentext/verticadb-operator:25.3.1-0
          createdAt: "2025-09-24T00:58:06Z"
          description: Operator that manages a Vertica Eon Mode database.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/vertica/vertica-kubernetes
          support: Vertica
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventTrigger is the Schema for the eventtriggers API
            displayName: Event Trigger
            kind: EventTrigger
            name: eventtriggers.vertica.com
            version: v1beta1
          - description: |-
              VerticaAutoscaler is a CR that allows you to autoscale one or more
              subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1
          - description: |-
              VerticaAutoscaler is a CR that allows you to autoscale one or more
              subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1beta1
          - description: VerticaDB is the CR that defines a Vertica Eon mode cluster
              that is managed by the verticadb-operator.
            displayName: Vertica DB
            kind: VerticaDB
            name: verticadbs.vertica.com
            version: v1
          - description: VerticaReplicator is the Schema for the verticareplicators
              API
            displayName: Vertica Replicator
            kind: VerticaReplicator
            name: verticareplicators.vertica.com
            version: v1beta1
          - description: VerticaRestorePointsQuery is the Schema for the verticarestorepointsqueries
              API
            displayName: Vertica Restore Points Query
            kind: VerticaRestorePointsQuery
            name: verticarestorepointsqueries.vertica.com
            version: v1beta1
          - description: VerticaScrutinize is the schema for verticascrutinize API
            displayName: Vertica Scrutinize
            kind: VerticaScrutinize
            name: verticascrutinizers.vertica.com
            version: v1beta1
        description: |-
          ### What is Vertica?
          Vertica is a unified analytics platform, based on a massively scalable architecture with the broadest set of analytical functions spanning event and time series, pattern matching, geospatial and end-to-end in-database machine learning. Vertica enables you to easily apply these powerful functions to the largest and most demanding analytical workloads, arming you and your customers with predictive business insights faster than any analytics data warehouse in the market. Vertica provides a unified analytics platform across major public clouds and on-premises data centers and integrates data in cloud object storage and HDFS without forcing you to move any of your data.

          ### What does this operator do?
          This operator manages a Vertica Eon Mode database, and automates the following administrative tasks:

          * Installing Vertica
          * Creating and reviving a Vertica database
          * Restarting and rescheduling DOWN pods to maintain quorum
          * Upgrading Vertica to a new version while keeping the database online
          * Subcluster scaling, both manually and automatically with the Horizontal Pod Autoscaler
          * Service management and health monitoring for pods
          * Load balancing for internal and external traffic
          * Expose a metrics endpoint to integrate with Prometheus to monitor the operator's health

          For a brief overview on how to create a cluster, see the [Vertica GitHub repository](https://github.com/vertica/vertica-kubernetes). For an in-depth look at Vertica on Kubernetes, see the [Vertica documentation](https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm).
        displayName: VerticaDB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - vertica
        links:
        - name: Documentation
          url: https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm
        - name: Vertica Container Images
          url: https://hub.docker.com/u/vertica
        maintainers:
        - email: cchen8@opentext.com
          name: Cai Chen
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Vertica
          url: http://www.vertica.com
        relatedImages:
        - docker.io/opentext/verticadb-operator:25.3.1-0
        version: 25.3.1-0
      entries:
      - name: verticadb-operator.v25.3.1-0
        version: 25.3.1-0
      - name: verticadb-operator.v25.3.0-0
        version: 25.3.0-0
      - name: verticadb-operator.v25.2.1-0
        version: 25.2.1-0
      - name: verticadb-operator.v25.2.0-0
        version: 25.2.0-0
      - name: verticadb-operator.v25.1.0-0
        version: 25.1.0-0
      - name: verticadb-operator.v24.4.0-0
        version: 24.4.0-0
      - name: verticadb-operator.v2.2.0
        version: 2.2.0
      - name: verticadb-operator.v2.1.3
        version: 2.1.3
      - name: verticadb-operator.v2.1.2
        version: 2.1.2
      - name: verticadb-operator.v2.1.1
        version: 2.1.1
      - name: verticadb-operator.v2.1.0
        version: 2.1.0
      - name: verticadb-operator.v2.0.2
        version: 2.0.2
      - name: verticadb-operator.v2.0.1
        version: 2.0.1
      - name: verticadb-operator.v2.0.0
        version: 2.0.0
      name: v2-stable
    defaultChannel: v2-stable
    packageName: verticadb-operator
    provider:
      name: Vertica
      url: http://www.vertica.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: F5 Networks Inc.
      provider-url: ""
    name: f5-ipam-controller-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: f5-ipam-controller-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fic.f5.com/v1",
                "kind": "F5IpamCtlr",
                "metadata": {
                  "name": "f5ipamctlr-sample"
                },
                "spec": {
                  "args": {
                    "infoblox_grid_host": "IP_ADDRESS_OF_SERVER",
                    "infoblox_labels": "CIDR_Labels",
                    "infoblox_netview": "NetView",
                    "infoblox_password": "PASSWORD",
                    "infoblox_username": "USERNAME",
                    "infoblox_wapi_version": "VERSION",
                    "insecure": true,
                    "ip_range": "IP_RANGE_LABELS",
                    "log_level": "DEBUG_OR_INFO",
                    "orchestration": "kubernetes",
                    "provider": "INFOBLOX_OR_F5-IP-PROVIDER"
                  },
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "f5-ipam-controller",
                    "user": "f5networks",
                    "version": "latest"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "serviceAccount": {
                    "create": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
          createdAt: "2025-08-08T03:48:02Z"
          description: Operator to install F5 IPAM Controller.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/f5-ipam-controller
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This Operator installs F5 IPAM controller(FIC)
            kind: F5IpamCtlr
            name: f5ipamctlrs.fic.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 IPAM Controller in your OpenShift Cluster. This enables to configure and deploy FIC using Helm Charts.
          ## F5 IPAM Controller
          The F5 IPAM Controller is a Docker container that runs in an orchestration environment and interfaces with an IPAM system. It allocates IP addresses from an IPAM system’s address pool for hostnames in an orchestration environment. The F5 IPAM Controller watches orchestration-specific resources and consumes the hostnames within each resource.
          ## Documentation
          Refer to F5 IPAM Controller documentation
          - https://github.com/F5Networks/f5-ipam-controller/blob/main/README.md
          ## Prerequisites
          ### Configure Infoblox with network and netview [Provider: Infoblox]
          Refer Infoblox documentation
          ### Create persistent volume and persistent volume claim [Provider: f5-ip-provider]
          Create persistent volume and persistent volume claim
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/f5-ipam-controller/main/docs/config_examples/f5-ip-provider/localstorage-pv-pvc-example.yaml
          ```
        displayName: F5 Ipam controller
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipam
        - ip
        - ipaddress
        - controller
        - ipv4
        - ipv6
        - F5
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/userguide/ipam/
        - name: Github Repo
          url: https://github.com/F5Networks/f5-ipam-controller
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
        - registry.connect.redhat.com/f5networks/f5-ipam-controller@sha256:09a358db300d82e00fd3b4e255917df82c02e1bc58425f8cff21cc2c77eba1b6
        version: 0.0.8
      entries:
      - name: f5-ipam-controller-operator.v0.0.8
        version: 0.0.8
      name: alpha
    - currentCSV: f5-ipam-controller-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fic.f5.com/v1",
                "kind": "F5IpamCtlr",
                "metadata": {
                  "name": "f5ipamctlr-sample"
                },
                "spec": {
                  "args": {
                    "infoblox_grid_host": "IP_ADDRESS_OF_SERVER",
                    "infoblox_labels": "CIDR_Labels",
                    "infoblox_netview": "NetView",
                    "infoblox_password": "PASSWORD",
                    "infoblox_username": "USERNAME",
                    "infoblox_wapi_version": "VERSION",
                    "insecure": true,
                    "ip_range": "IP_RANGE_LABELS",
                    "log_level": "DEBUG_OR_INFO",
                    "orchestration": "kubernetes",
                    "provider": "INFOBLOX_OR_F5-IP-PROVIDER"
                  },
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "f5-ipam-controller",
                    "user": "f5networks",
                    "version": "latest"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "serviceAccount": {
                    "create": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
          createdAt: "2025-08-08T03:48:02Z"
          description: Operator to install F5 IPAM Controller.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/f5-ipam-controller
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This Operator installs F5 IPAM controller(FIC)
            kind: F5IpamCtlr
            name: f5ipamctlrs.fic.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 IPAM Controller in your OpenShift Cluster. This enables to configure and deploy FIC using Helm Charts.
          ## F5 IPAM Controller
          The F5 IPAM Controller is a Docker container that runs in an orchestration environment and interfaces with an IPAM system. It allocates IP addresses from an IPAM system’s address pool for hostnames in an orchestration environment. The F5 IPAM Controller watches orchestration-specific resources and consumes the hostnames within each resource.
          ## Documentation
          Refer to F5 IPAM Controller documentation
          - https://github.com/F5Networks/f5-ipam-controller/blob/main/README.md
          ## Prerequisites
          ### Configure Infoblox with network and netview [Provider: Infoblox]
          Refer Infoblox documentation
          ### Create persistent volume and persistent volume claim [Provider: f5-ip-provider]
          Create persistent volume and persistent volume claim
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/f5-ipam-controller/main/docs/config_examples/f5-ip-provider/localstorage-pv-pvc-example.yaml
          ```
        displayName: F5 Ipam controller
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipam
        - ip
        - ipaddress
        - controller
        - ipv4
        - ipv6
        - F5
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/userguide/ipam/
        - name: Github Repo
          url: https://github.com/F5Networks/f5-ipam-controller
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
        - registry.connect.redhat.com/f5networks/f5-ipam-controller@sha256:09a358db300d82e00fd3b4e255917df82c02e1bc58425f8cff21cc2c77eba1b6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        version: 0.0.8
      entries:
      - name: f5-ipam-controller-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: f5-ipam-controller-operator
    provider:
      name: F5 Networks Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: smart-gateway-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smart-gateway-operator.v5.0.1757650606
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "smartgateway.infra.watch/v2",
                "kind": "SmartGateway",
                "metadata": {
                  "name": "demo"
                },
                "spec": {
                  "applications": [
                    {
                      "config": "host: localhost\nport: 8081\nwithtimestamp: false\n",
                      "name": "prometheus"
                    }
                  ],
                  "blockEventBus": false,
                  "bridge": {
                    "amqpBlock": false,
                    "amqpUrl": "amqp://amq-interconnect:5672/collectd",
                    "enabled": true,
                    "ringBufferCount": 15000,
                    "ringBufferSize": 2048,
                    "socketBlock": true,
                    "statsPeriod": 60,
                    "stopCount": 0,
                    "unixSocketPath": "/tmp",
                    "verbose": true
                  },
                  "handleErrors": false,
                  "logLevel": "info",
                  "pluginDir": "/usr/lib64/sg-core",
                  "services": [
                    {
                      "name": "prometheus",
                      "ports": [
                        {
                          "name": "prom-http",
                          "port": 8081,
                          "protocol": "TCP",
                          "targetPort": 8081
                        }
                      ]
                    }
                  ],
                  "size": 1,
                  "transports": [
                    {
                      "config": "path: /tmp/smartgateway\n",
                      "handlers": [
                        {
                          "config": "",
                          "name": "collectd-metrics"
                        }
                      ],
                      "name": "socket"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/stf/smart-gateway-rhel9-operator@sha256:3363abcb91319f8642212b9b80ca49bb9f90cae69255b1f75f03929af3d3a8c3
          createdAt: "2025-09-12T04:17:59Z"
          description: Operator for managing the Smart Gateway Custom Resources, resulting
            in deployments of the Smart Gateway.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenStack Platform", "Cloud
            Infrastructure", "Cloud Suite"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/infrawatch/smart-gateway-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of the Smart Gateway
            displayName: SmartGateway
            kind: SmartGateway
            name: smartgateways.smartgateway.infra.watch
            version: v2
        description: |-
          # Smart Gateway for Service Telemetry Framework

          The Smart Gateway is middleware for OpenShift that takes metrics and events
          data streams from an AMQP 1.x message bus, and provides a method to having
          that data stream stored within supported data stores such as Prometheus and
          ElasticSearch.

          ## Core capabilities

          The Smart Gateway provides two modes:

          * metrics: provides an HTTP scrape endpoint for Prometheus
          * events: writes events directly to an ElasticSearch endpoint

          ## Operator features

          * **Two Service Types** - Supports the `metrics` and `events` service types which defines the type of Smart Gateway to be deployed

          * **Configuration** - Configuration of the Smart Gateway to allow for it to be connected to an AMQP 1.x bus and pointed at the appropriate storage mechanism based on service type.

          ## Before getting started

          1. Ensure an AMQP 1.x bus has been setup (such as AMQ Interconnect)
          2. If the Smart Gateway is running in events mode, be sure the ElasticSearch Operator and ElasticSearch instance have been pre-deployed.
        displayName: Smart Gateway Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - servicetelemetry
        - monitoring
        - telemetry
        - notifications
        - smartgateway
        links:
        - name: Source Code
          url: https://github.com/infrawatch/smart-gateway-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/stf/smart-gateway-rhel9-operator@sha256:3363abcb91319f8642212b9b80ca49bb9f90cae69255b1f75f03929af3d3a8c3
        - registry.redhat.io/stf/sg-core-rhel9@sha256:0e12903da0b15c64421102b5b914ca4c82d107295485c9b76ed19e91a45a4420
        - registry.redhat.io/stf/sg-bridge-rhel9@sha256:7fae342cf6cecc676eaa0a8fb64ee380ad101d667d6e8f0d89c943938bb6baf4
        - registry.redhat.io/openshift4/ose-oauth-proxy:latest@sha256:dc24aa382e02c4ac058d4b824a21f0e90a6259d6d817d0341b2a41809b5daa78
        version: 5.0.1757650606
      entries:
      - name: smart-gateway-operator.v5.0.1757650606
        version: 5.0.1757650606
      - name: smart-gateway-operator.v5.0.1743671602
        version: 5.0.1743671602
      - name: smart-gateway-operator.v5.0.1740051852
        version: 5.0.1740051852
      - name: smart-gateway-operator.v5.0.1731336910
        version: 5.0.1731336910
      - name: smart-gateway-operator.v5.0.1725953602
        version: 5.0.1725953602
      name: stable-1.5
    defaultChannel: stable-1.5
    packageName: smart-gateway-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-pipelines-operator-rh
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-pipelines-operator-rh.v1.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    },
                    {
                      "name": "legacyPipelineRbac",
                      "value": "true"
                    },
                    {
                      "name": "createCABundleConfigMaps",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.19.0 <1.20.0'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.3.2
          - Tekton Triggers: v0.33.0
          - Pipelines as Code: v0.37.0
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.22.3
          - Tekton Results (tech-preview): v0.16.0
          - Manual Approval Gate (tech-preview): v0.7.0
          - Tekton Pruner (tech-preview): v0.2.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.20/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:5021158a4752a26f8b33adea7da66695b22d2ad155d346d35dde5f971bc1a55b
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:5683321a5376e63c3ff3932270a1850d5e261f76212020ce867334166f90d2f7
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:a712bb411d2433b01793dff25785c1e265fa865a52f7c8732c9fb4418ab61757
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e7998e113f75e8f29602bae210c8b66ed06634b699f1d99c5875ef879220d733
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:96db721939b22572d7d8d8b95dcb5c92fdedb98e9cd0aabb7d17b703a915b07f
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:115eea1a5ca87f3b6b72db1df05b7386bf73dc70c9c2e4dc5e1a66700d960c37
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:980ed7e81ea55522e1681dc1e9b170b68d16f07061dd1f1ffd142367d59552f1
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:1b6aecfce45ed3163227b845b212e83385cdbe999edfc2e9387a0c76f3f8d8e9
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:161df929e2bd4c5acff594be31d5458365de5ae0d2189bb2f647c3fc6657fa28
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:bc10e7d6122ede5cf4f2882c200447683255465a1080e826761e90bfb983506e
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:df15eb1f1f5f237bfe481c16ab3d40bf7fb4123f3c930bcfb121e2cc8dc42882
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:d56d6b46331b4569077ab1e7a7b609ba01f0a1dd62b695d62f8aa0f32ea1b621
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:c0f438e6095aa81f3f7d820614e0414cc43374014c1505d7641bc1dba69a3903
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:818bfa4a3a874974b5f7f8bcf2faeb6c05e1795381ee1ceacf89f17b1ec0ef7f
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:9415e889ef98df6d2b2dca1e5d798fb765eeea0d9828e48424374d5267a6b0d1
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:441e1c7124c1674e3d4c38c3e82a4fe5568e76d13c89c04bf009d6b97435de09
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:7eafffa6e4e04f10c1d4e84e02d51c00eb8649c60f8f9a13da14492f24634891
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:bc403f93ef6c5526fa30152eab384c4a25bb0bc6ab4158a1e24f637b87a57259
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:2d496752b19f765f1ea36ec94b873ebd8fa604764a56aa67d3ed94b33e5fe75c
        - registry.redhat.io/rhel9/buildah@sha256:18453ab0a62154283aaf4c45efd7543d1c8ecca1aebae46dedc7ddc0f410b232
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:036250bcdc6fb95273bf07948dfdd4dce765ab84c803ceca0d0b4abf6ca19dec
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:8f0494341244a75546a4b80b065702be04e6d200421aa4e7ec90385c9dabc6ab
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:c4d05e546c466c57379fff502816012f615be073eb8cdce54677b4ab206a3f5c
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:bbd1918660be7b3c78777dcccc87d3b26bf5adfc077844ceffbbdd53696dc0a3
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:e3e64abef53151f6b15f15bcbff85668167a571e98badbd8a2ca252ff3a5b8dc
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:dd941acd357ae2c0fd7e15fb8d32d7ee6b2847435101dd1e1f107db48af18bf3
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:823e018e0329d764ac694f177ce88af9074f364c29a1c17929f9dd0b596d03dd
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:a7163a9e1fe891de021b1003558e6e64fe17ed17ab0820ba16bba09aa112401e
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:e63298d07ccec5804dfab4e42bcc8163c29cb3ac371c91248777423a4edd31ce
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:3cf990fb2fc422f3527a7c8b43d06ccd4d32bb5487ec8cb2d4a97f0f5d13f170
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:c3013c347bca8a520443a44ac2817fbba5744a9e91069ac5f09e91447ebf6b28
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:901b5108ceb669c985e75f5fb033c496e9f028863df99ab32651d1b08433a352
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:14dab8d87323fc7d112213f7f4db3ee7cf3841747de669ce710ae6a794053457
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:458f865f1b246c5f0682f5956795771a0515b89edfcffbbc0da1a3d3f532645e
        - registry.redhat.io/ubi9/ubi-minimal@sha256:2f06ae0e6d3d9c4f610d32c480338eef474867f435d8d28625f2985e8acde6e8
        - registry.redhat.io/ubi9/openjdk-17@sha256:555c8bcd2596637befd1a6da4aa949c5a2640cf66e09c2dfb932afe37e39095e
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:c31af237afb84214bf2a4c06a7636ca071fe97203b41da90d7878109ccd64784
        - registry.redhat.io/openshift-pipelines/pipelines-opc-rhel9@sha256:c1bee4a2bb495de6d4da962b6e7b2447caa51cc36c3005e4083bd31420ef5102
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:ec74a562c18fefe36ec7ae3ba083e4eb76c56b562f916578900f8767ae98e6c2
        - registry.redhat.io/rhel9/skopeo@sha256:a65a413f8a2864389a09dc750690d97afbcdc5c70821e0f85e99e8adba7954e0
        version: 1.20.0
      entries:
      - name: openshift-pipelines-operator-rh.v1.20.0
        version: 1.20.0
      - name: openshift-pipelines-operator-rh.v1.19.3
        version: 1.19.3
      - name: openshift-pipelines-operator-rh.v1.19.2
        version: 1.19.2
      - name: openshift-pipelines-operator-rh.v1.19.1
        version: 1.19.1
      - name: openshift-pipelines-operator-rh.v1.19.0
        version: 1.19.0
      - name: openshift-pipelines-operator-rh.v1.18.1
        version: 1.18.1
      - name: openshift-pipelines-operator-rh.v1.18.0
        version: 1.18.0
      - name: openshift-pipelines-operator-rh.v1.17.1
        version: 1.17.1
      - name: openshift-pipelines-operator-rh.v1.17.0
        version: 1.17.0
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      - name: openshift-pipelines-operator-rh.v1.14.1
        version: 1.14.1
      - name: openshift-pipelines-operator-rh.v1.13.0
        version: 1.13.0
      - name: openshift-pipelines-operator-rh.v1.12.2
        version: 1.12.2
      - name: openshift-pipelines-operator-rh.v1.12.1
        version: 1.12.1
      - name: openshift-pipelines-operator-rh.v1.12.0
        version: 1.12.0
      - name: openshift-pipelines-operator-rh.v1.11.1
        version: 1.11.1
      name: latest
    - currentCSV: openshift-pipelines-operator-rh.v1.14.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "db_user": "result",
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.14.6'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.56.4
          - Tekton Triggers: v0.26.2
          - Pipelines as Code: v0.24.6
          - Tekton Chains: v0.20.1
          - Tekton Hub (tech-preview): v1.16.0
          - Tekton Results (tech-preview): v0.9.2

          ## Getting Started
          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/container-platform/OPENSHIFT_DOCS_VERSION/cicd/pipelines/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)
          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi8/openjdk-17@sha256:a8165bc2cd5051a96d6937e25ed03155bbd4b731da6e58cebfe2ea83209c16d8
        - registry.redhat.io/ubi8/ubi-minimal@sha256:6825946f10e8c81586fa7815c4f06016af05cfc231d921ec9699402a6655d201
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:059c08316152c209de13017e322d6fe5eacfde81c3b0039e55b5de8f2f27183a
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:d01ba7ead670d5ab9c1355b90010657988e46b4d1aa0bb68b194eaf4001b2d7a
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:6327385e3d3c9df1ea50e803955d980b4cd35fb15593fcfec0d477d91a45021f
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel8@sha256:6e7c7a1ca458e08ad85e1146815ec00a30016f709b45e28b464043e875c2a3e8
        - registry.redhat.io/rhel8/skopeo@sha256:e297b84ba6f1d5eb2c37e2c3b31c9a9a4c39f9590e7d7c32388165245bde4835
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:ae0260abd2725c2f4bb25c0c07d86290978892a49050e0fd23f466401aba668c
        - registry.redhat.io/rhel8/buildah@sha256:b48f410efa0ff8ab0db6ead420a5d8d866d64af846fece5efb185230d7ecf591
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:2261f4f657f70759f4d0c898e44ffafceb8dbe2144063d50227140463bdbf2c5
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:7a63e7a762b9c7b15057ed9d5bcbea08acb4c8933a300c6ec336824d49a6e937
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:8d4450d68ac03d8275e337d2c55692ef35e954400d0c9f041a943d163a8b7680
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:bf6cf2e87fb19f7aa9a490b83c16af69834c0721220a643710a1b077959e91ca
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:aa6eb4c7ade9d6add1e4ab94197c2ce72a253fea0d7461698ca0e37854f1dcbe
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:71d91c7b80b6d196025f68749bd8518b8b0affd8d5064a7c5c70d125656fbc36
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:db3ff043410d1dd0621fef4483761727f7241d788b136e446fb861953350a2cb
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:5e7f5a2813a591e7b13fea9fc4e2f599b5ec855a015cd5a4c287c4e178358756
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:9c29f6eff1f7019993446d85d226f04c5614a18351dfb4042a8bcb62276810de
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:d9972bc64dfa7e260bd9b949bbb85ef5ae532e7c6df8d874d3d3963af1097726
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:bf0c38beefa24ca3d04ebfa6140a1cc5a9a56fba81ae05799279a559093b9033
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel8@sha256:b3bedee6084e9d9da549d3e637d2bf176aa365b5818e6b6377525577edd710dc
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:82c8a91d521d7e6d81fb462fa2b804f420f52ee062df5cda39545165a2f19e0c
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:ce0ea093c522af329acdc85148c2473ea45cd44270c15f8dd32965b0d56f304f
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:abf4442b6acc714c2392b86af046d5f32d247ee0c3763bb18c57d0473edfcf02
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:e2bd929246d62038a24a0685cb3e4873c131c9b7713d6a5268ca09a8614538be
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:83913744a1bb7ff23091cfa541b807cfb82b38dde0c37e0f1dd9c17cade99775
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:63fc1351191a54d12d4e463a1bb4cc732620cba3000af86ba7f7a579514f5fe4
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:e23c901868229e4f6a4016aa7798f1aba879ff35ffe594f5b0e4d39f1bff9e52
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:f68dc18a4dc6a1886507097551f4a6fce39062d03c6eb86c067d317812b0d34a
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:f9fed20cdfebbb226e035cc62f037174f5ecdb3139a5edc7e39f019f6d2b1e98
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:ddf0cb64ea57de872515e19de63b2dec5f56738c9d7eaa283a4a489545ef4902
        - registry.redhat.io/rhel8/postgresql-13@sha256:3903d562425dd564986123280dc076548fb3d24c74138dcd857652f53052e0a7
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:2e75096ffe20d1261e8a645fefd32e37295d727aa6a280f10041a44443917469
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:57fb86ba112d79a228cbacc96cd0525c123464bcac956354722e3f2aeff42508
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:ffa8c6d8fb5bacb59da27f4ebde87385359b3f97a6a023b4a4488ced2d46c574
        version: 1.14.6
      entries:
      - name: openshift-pipelines-operator-rh.v1.14.6
        version: 1.14.6
      - name: openshift-pipelines-operator-rh.v1.14.5
        version: 1.14.5
      - name: openshift-pipelines-operator-rh.v1.14.4
        version: 1.14.4
      - name: openshift-pipelines-operator-rh.v1.14.3
        version: 1.14.3
      - name: openshift-pipelines-operator-rh.v1.14.2
        version: 1.14.2
      - name: openshift-pipelines-operator-rh.v1.14.1
        version: 1.14.1
      - name: openshift-pipelines-operator-rh.v1.13.0
        version: 1.13.0
      - name: openshift-pipelines-operator-rh.v1.12.2
        version: 1.12.2
      - name: openshift-pipelines-operator-rh.v1.12.1
        version: 1.12.1
      - name: openshift-pipelines-operator-rh.v1.12.0
        version: 1.12.0
      - name: openshift-pipelines-operator-rh.v1.11.1
        version: 1.11.1
      name: pipelines-1.14
    - currentCSV: openshift-pipelines-operator-rh.v1.15.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "db_user": "result",
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.15.3'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.59.6
          - Tekton Triggers: v0.27.0
          - Pipelines as Code: v0.27.2
          - Tekton Chains: v0.20.1
          - Tekton Hub (tech-preview): v1.17.1
          - Tekton Results (tech-preview): v0.10.0
          - Manual Approval Gate (tech-preview): v0.2.2

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.15/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:322ef2385620661a7914f47c73a6d145ec0c9ce1ccde30645bb66148545ddc96
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:49c346bf7fc69a9796690236dc8597e35a874343b02fb74c5d2f1c05d356e41c
        - registry.redhat.io/rhel8/skopeo@sha256:29465586e92ed04d2878dcdeda2d508c7e8e7539762e10296b52d631929960aa
        - registry.redhat.io/rhel8/buildah@sha256:aac6629389db17e99894c5bee0da01d4c8065d11d8c6f6e1602f9484290baa70
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:bf94c350923da706c6e0a427e2d91af3e13e751a12388a4ed75d1c34dd7b3177
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:ac1a6e057a5599fda7e048ff77030e98c92b5d51cac2472c1ef3910964fc17e6
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:61c22119f37e620a3807056cde35bc07c90867765b149294056326b1d11494f1
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:abac279a0a5ecde57952b6bef634f8ba7527cd8c8956a2ea30e9449174ad52bd
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:166789a990df9f6ab04fd3df0a8cd301bc19b9bedde94215cb27306060eb1e2c
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:83eb08bb9579cc19132af282c455bca32a0fd4f5436a35e7ad23b5efa2faa3cf
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:479e3eed93e0aa13e3489514fe00375a519eb45969d099c490b4acd056ea5adf
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:d1f86458079449c86407ac716c3cbf608cad06de29edeecbde6586f76628daa9
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:e749db81c9d7e11e2802833a2a02e93afc357c83d1a2a827e537262df281a4f7
        - registry.redhat.io/ubi8/ubi-minimal@sha256:f729a7f5685ea823e87ffd68aff988f2b8ff8d52126ade4e6de7c68088f28ebd
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:7c81ae208a16e17d0102074867add37dfb1207ee95cbbc6c732d88e61bbc59fb
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:7152c3c174d25f2a5239456ba677ac5315bfb82bc3b3ac5ee0eb8f9f74fb6632
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:aca0024cac7a8c41b1ed151eb1e1188f09bc636866f4e03f1b973f90890b37cf
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:bf6cf2e87fb19f7aa9a490b83c16af69834c0721220a643710a1b077959e91ca
        - registry.redhat.io/rhel8/postgresql-13@sha256:a92a579f1aef66ac188d24fd489c456a1a3e311d95dcce652da6b81e28fbf725
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:7f00ff1b407d9ea107d3d969988a6279b28bde86943f490a007fa922dd40a1c3
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:eed464a3b4e23f0b2623aa1b923c10232d235f4272f562955926c1748e295d2b
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:7958925cecce1b4b02ac4649e3bfd7fa3b5361c16ab3593700b8da93fdf1c79a
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:312d1f90765d05a86845856c677e1ece4c57e81a9bfe3fd1e76868d624e3d72d
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:1300d171dbf09c63071ba5586e223491a13acfc6dbab27eff04648aeb788d455
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:538c63c8e7af3cba91526f2f9ac99fb8f6b0f4781db31771c959612c209dd090
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:601fcac1abc6c8efa1248c5ca229213d6660708470a1c416d41c6aff0ba4691a
        - registry.redhat.io/ubi8/openjdk-17@sha256:e8cc2e476282b75d89c73057bfa713db22d72bdb2808d62d981a84c33beb2575
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:e7b000a2735bbf98ee79bd093f7b703ed09d7c1e11ed2a5b22c4b8ac8d440386
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:827313c2691fb8833e42418b65b4a3360a11cabc2c92a7d8e7b20183f06cbf68
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:f02af8475c33f6f38a2f90ff1be1fb1b2094860c4d5e300aefdd7eb84b4a529a
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:acf8017029f4b2e7a1c3da9bc62c65239d313d24e3012712356a191bea9a2e6f
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:75d6228eb656b8778fb3bb8dc68c686ac1c9b787e5343ae3c2467f8e724af83e
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:1db14f23077785108660dde38c98e2e30c8301f8e2e1c594dd4636e0ba29f84f
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:70bb0ebcfe5da45e5c239aa011dc5f1ef48844116d7346bb05433b503107268a
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:f04ea12675e0eb2354635cd2f9e99d6e59619d1c611e0ea95c3ce8222d615322
        version: 1.15.3
      entries:
      - name: openshift-pipelines-operator-rh.v1.15.3
        version: 1.15.3
      - name: openshift-pipelines-operator-rh.v1.15.2
        version: 1.15.2
      - name: openshift-pipelines-operator-rh.v1.15.1
        version: 1.15.1
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      - name: openshift-pipelines-operator-rh.v1.14.1
        version: 1.14.1
      - name: openshift-pipelines-operator-rh.v1.13.0
        version: 1.13.0
      - name: openshift-pipelines-operator-rh.v1.12.2
        version: 1.12.2
      - name: openshift-pipelines-operator-rh.v1.12.1
        version: 1.12.1
      - name: openshift-pipelines-operator-rh.v1.12.0
        version: 1.12.0
      - name: openshift-pipelines-operator-rh.v1.11.1
        version: 1.11.1
      name: pipelines-1.15
    - currentCSV: openshift-pipelines-operator-rh.v1.16.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.16.4'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.62.9
          - Tekton Triggers: v0.29.2
          - Pipelines as Code: v0.28.2
          - Tekton Chains: v0.22.3
          - Tekton Hub (tech-preview): v1.18.2
          - Tekton Results (tech-preview): v0.12.3
          - Manual Approval Gate (tech-preview): v0.3.1

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.16/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:81a8b903d0fc4b81e6a23fac862d18da4cc18b14cca8213665df37dfaabbc2e9
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:5672f3a375c6a274200b423eb2643c818aba75488dec0b210fec2532814ba51e
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:47f094ccd12db7a43c8a91c1a3d2a190bc2b83f1745e7eb8ce097d0c8e1e2c88
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:80d46bc5ce8c931af22a97ccbce05b59740e313167c88ad910febcc2f2dc0320
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:90e2a7ea498da756e09c7e4d19c7fb91d429cc28c57f97c6958958e3466cb58c
        - registry.redhat.io/rhel8/skopeo@sha256:1dcb4189c92da21ee58b3ca5433615e0208194c8412f7f1fb45fa491179aaf9b
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:d2bb3f9f79ef719d8fdafef3a639e3eb8bc199ee56fcfb33e85975d536d2a6ac
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:7edbd11e5a449408d673cae1fc2b9b5bab492ecff2ac6e640c8cd69a9b90ff08
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:a5a7b2be50f211d4d9d555b2a2e2302096c139ff5011d984cc386f1530007f16
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:f9122d71c3862cd2afafc8eb7f1a39c83eee306e5baf38514e6e39215c70a72b
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel8@sha256:8cb87920e6fc8185ead49aafb7799fead6a4a510eafdae0ee29e85248dae3945
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:aacdabe243dc573f3161a3d04fbf014b77d9962207305d70fcafce475f7eea93
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:92042dcacd4ab0f2a78aab03fd91a9a39d7d7366b5398fd797ecb145acf5fce4
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:2fea7d8e29d70c5893ddb6caf2bf1d4144787f5f0de8739351d2097ed09ee751
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/rhel8/buildah@sha256:5c7cd7c9a3d49e8905fc98693f6da605aeafae36bde5622dc78e12f31db3cd59
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:474658c262b539dd1c83ac22bb090b3fdac5de9cffc06587d36dbc6fa4ce208c
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:e1fe4a86211cf22e5ad2b23c989d24b69f597db4c3604e84777dbd0c585ec186
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:6acc31f179b08658976cddcb9ef42f377edb29933d1d83a2ecebfd455bb68571
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:f3e6df07fd86acb91037d55a1970e5f8c6ab12a7517752692c224c7687022812
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:8129e62906c68f04075a44c939375a7bafdeaeac673a20a4fa74ffd6285187ac
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:7d92dca9da27c066cb530172e1811493d8c916e67d8068f5ddc471fa3f3f7856
        - registry.redhat.io/ubi8/ubi-minimal@sha256:a47c89f02b39a98290f88204ed3d162845db0a0c464b319c2596cfd1e94b444e
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:b94bd566ae3d9ce5f47273d867164486826b19fca0ce11d1a7d18312413e87ba
        - registry.redhat.io/ubi8/openjdk-17@sha256:632e78989471859ca4ed1148b951a911f7e3b6e6516482be20967c4171612c55
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel8@sha256:46635d636ea2e9a6ba0760bf8c528028700ee40b0ba8466797034371185b16d0
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:856229e9e37683c926e3449658d76f93d910384eff62c2ca192768a653b71533
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:c6d654c71aba35c3c7f5c1a2ca5b8401e95adf6f649c599a9aff25177c2b3323
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:549d74a18ec52db91fa949cac9b8e3d943d478d5dec9850d2ba0cb1bb526f578
        - registry.redhat.io/rhel8/postgresql-13@sha256:a308263f5d1071be8ac5c13fd84b40609988ca30ed56c22a8bc6769e107ce09b
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:b41b888fc1a3b7215e6b71ee97bf64a1816914a8f33e96ad3930a4e2d5b5ffa7
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:4e0e4c358faa9cafc3964d6d40ee4aa73f8dfe8b6338dfb8206613d7c5241842
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:dcab4a3d519975340ade59a0ca0222104e68acd052a61f2aeb8c736be4b672cb
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:e84c2d80e92815c6bf384e1504a194dec9b796a3aceac2eed555fff241249571
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:d36dbb60c7b5caca7f3be08705f8ab91d7f01abaf8bfee4aee73bdbbb3e7e71a
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:7da7f929d2711fc55dda01fe51a474bfa7adf5aaf78be6d65a6d65850958d6e7
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:19b3dd69c1d7e17890b7163aef322c1efbc76fed7f2b1faf26aaa05fd8161184
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:aeeed543412cde0bd3a9baec965cb99d80f2b6acb77ae6356ed51b4b1262d006
        version: 1.16.4
      entries:
      - name: openshift-pipelines-operator-rh.v1.16.4
        version: 1.16.4
      - name: openshift-pipelines-operator-rh.v1.16.3
        version: 1.16.3
      - name: openshift-pipelines-operator-rh.v1.16.2
        version: 1.16.2
      - name: openshift-pipelines-operator-rh.v1.16.1
        version: 1.16.1
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      - name: openshift-pipelines-operator-rh.v1.14.1
        version: 1.14.1
      - name: openshift-pipelines-operator-rh.v1.13.0
        version: 1.13.0
      - name: openshift-pipelines-operator-rh.v1.12.2
        version: 1.12.2
      - name: openshift-pipelines-operator-rh.v1.12.1
        version: 1.12.1
      - name: openshift-pipelines-operator-rh.v1.12.0
        version: 1.12.0
      - name: openshift-pipelines-operator-rh.v1.11.1
        version: 1.11.1
      name: pipelines-1.16
    - currentCSV: openshift-pipelines-operator-rh.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.16.0 <1.17.2'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.65.7
          - Tekton Triggers: v0.30.1
          - Pipelines as Code: v0.29.1
          - Tekton Chains: v0.23.1
          - Tekton Hub (tech-preview): v1.19.2
          - Tekton Results (tech-preview): v0.13.4
          - Manual Approval Gate (tech-preview): v0.4.1

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.17/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:a14270967460883f283ec1ce17ff8b82db1e0777dda4f6d690f15ac38dcbad85
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:650391ffd3e4c7b173bf9d718405b2375c6ff2027abc0b7ce4f67e79e7ef4b81
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:37109f6de75dda750496f6668164d0dceaf9db5278e8c034708436f39de3276e
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:a392d2c85baedd7428c579e91d9f8acf8232399b90523b1093dc3447aaf56a16
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:8e52b4b31b0d9ec772abe6e6ef9c6698e89893395f364968216bebc3367da4b2
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:4b87782b35e7284a365c49f1c094aae98531ac630db1958294009c041a2d8440
        - registry.redhat.io/ubi8/openjdk-17@sha256:e5fc9e34382da41b10d4ea566639d31707493abff9e8de81f5accfd66876409e
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:4224040b00cfd76e85f80aecc5095098564cc62e9c477fe2f7a428b30df7d4d9
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:41ec6db921812fd1e80b05d51c224ddbce63bf0c57f02ccf07461d2202e97b90
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:4fabae1312c1aaf8a57bd2de63bd040956faa0c728453f2a4b4002705fba0f0c
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:7abc88bea01dd6513f6e48334f0b98501f909a1dc1ab9e83efeb3c1a834f8f92
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:eab798472b76bd51235198fd3b95189740f8d650ebc6bb01fe653c025c1defaa
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:61c9e8b1660d8935ee316ecc113e8141daf05149d172685cabd861ae8aeed3eb
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:4240501c076605a01d969328bbb8ad885c598b44ec1dc1c066312491a0eb6715
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:7c710882752f12be3c22bedc8d32096c111e42c5b9b392163f61cf8d72be3ec0
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:66309ae437c4647f48efe3d6f2b696df83859323700427d0025b94795751a859
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:41e2f2c7397a8b9e093e18977666e93943734f51780c769c99958ae1d20af53e
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:4a96f1bb7e3a5c7ca62dff2199fa15ef964b129e4e36e9619837c1221538f496
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:3696cf724ce7865626ae408d82fd87727c741e4d3eb3475c271ec6d56f0882a4
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:44da76d75f567c2ba623bae646e1d552d2c0fded1accdde3aab4f4bcdf3adc28
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:26f7b695fc1430359ab2bfd2915cf0b7e88b06d8695e26761bdc2ce8935f0fb3
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:0b6ee6cd98cb7e835cfc9544996c6f85288d7252071c100bed853eb593d6befb
        - registry.redhat.io/rhel8/buildah@sha256:6d2dcb651ba680cf4ec74331f8349dec43d071d420625a1703370acc8d984e9e
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:d851a2dd8ec95bfd7ac6ec5515c8cdadfcbfc6e6ec84b40630b4fe595f978fa5
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:429bc3d6044ca7dab780ff4b5eb91d42d38a29eb7ef428a25c53b7dda49871bf
        - registry.redhat.io/rhel8/postgresql-13@sha256:757d8dd6f243717d820846da7144a49913f28f223b74ec305b2215f54720f8a0
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel8@sha256:b5cceead6b8e9c99655c9ba810679b9d6fb978e5ff18d020a46d475c4c26fa29
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:c92ebe828b6ced6c93a3df5f4ccdfba022471729ad08b184de2f7410ecf592cc
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:df11cdbab2a1c6b9b800ef124c73f2fb5e3981402b21bc822ccc13e95836197a
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:db27cb4f0c350d99546b0f1bdf02d1ba271a8053a4552d88ff302adca60a3e29
        - registry.redhat.io/rhel8/skopeo@sha256:7fd65200b4847fc84fba0466519b87f915e28c0e9bc90d3022beecfebffd4af8
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:99f5383d8c6bc046a0bf48fd6c2064a5d055645085248327305f3ee4ae71979c
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:e673c20fff2fc684f5098282dd6ac6cb265045945249e995b81289532ce8c856
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:9668d852123b160e49511242371edc6e941a8e8b8c2ffa60080d91cc6b97d3d7
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:84eed8ba1189448e02ffd284c7d913e7b3cb7efe7a8e5b0edf890f7eb17fc439
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:922b9e86a674a6f3eb82f510569ab9cde97a1ad833c263335cd4164bf4164b66
        - registry.redhat.io/ubi8/ubi-minimal@sha256:c12e67af6a7e15113d76bc72f10bef2045c026c71ec8b7124c8a075458188a83
        version: 1.17.2
      entries:
      - name: openshift-pipelines-operator-rh.v1.17.2
        version: 1.17.2
      - name: openshift-pipelines-operator-rh.v1.17.1
        version: 1.17.1
      - name: openshift-pipelines-operator-rh.v1.17.0
        version: 1.17.0
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      - name: openshift-pipelines-operator-rh.v1.14.1
        version: 1.14.1
      - name: openshift-pipelines-operator-rh.v1.13.0
        version: 1.13.0
      - name: openshift-pipelines-operator-rh.v1.12.2
        version: 1.12.2
      - name: openshift-pipelines-operator-rh.v1.12.1
        version: 1.12.1
      - name: openshift-pipelines-operator-rh.v1.12.0
        version: 1.12.0
      - name: openshift-pipelines-operator-rh.v1.11.1
        version: 1.11.1
      name: pipelines-1.17
    - currentCSV: openshift-pipelines-operator-rh.v1.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.17.0 <1.18.1'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.68.1
          - Tekton Triggers: v0.31.0
          - Pipelines as Code: v0.33.2
          - Tekton Chains: v0.24.0
          - Tekton Hub (tech-preview): v1.20.1
          - Tekton Results: v0.14.0
          - Manual Approval Gate (tech-preview): v0.5.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.18/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi9/openjdk-17@sha256:5d619e202695acf76a5f32c07ace6709f5ee5d3999bf977e1bf95ad34197f44a
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:86ea94b12e8923035e3e4f39b6a53b3ab32c5f77d7eea170aa9b34aecc6f8577
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:0a60ce46c05256c74f770901e0bb79ad89e88ba882b643c1ff7b2a1738463acc
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:3964c3248dfb1006ea9a46295baf7f8cfa71952d6b9863b52f87e8f3d7863b1e
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:36a4c6a3273f2371ab5e679f1955221c286ea753e29e6ab8da7c983681a68f02
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:b550a8bc2a87d07a5b22a3ec3fb83efd8eb26ca6e80d404f2e453cc9b8434da7
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:cd28c2c90dd340324b0d8709640cd7636f28286114b0023649b6e21be13e03c2
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:fa5c456fc91319c55fac9a011dc7bb32ef85db293c73d74436e903b76ceb4349
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:409c14649d8e7007d4b72b6feffb71dabd26322bf2a7c77f5307f2546fdfe81b
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:96f5e1922fb9d6dc17e271576dda9bd032bd6dcc04e7496fb15a2f6b5871cd67
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:16d799503674641d4690822edc9c399c61429ccd31621d295eee4a4d1ef1c076
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:cbd89c531953a43b8055bda72f7696be80f57fd1d782567a91d82e0fbcb3d26d
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:8fec8e7a8b3e5182c67104de4e3abe32593b682c4a8986d1e9ab835dafb95ff4
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:3696cf724ce7865626ae408d82fd87727c741e4d3eb3475c271ec6d56f0882a4
        - registry.redhat.io/rhel9/skopeo@sha256:d077e35100d6d7b3795f8e3594293dbf8b65c6bc18b909aa555fe56f048b356a
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:1b57b64657eca0002145100bae262a19f02eb40068e07994df467e28a1618f90
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:20e19c02efdecf94a7b16483047e5db81c50dc2a0aa59d4895067e6c7769efcc
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e2b4c3b505c4b5aa36f8a5e661389b0de4e799c74c2c7f16a065e5522b438e89
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:80d441ec5d07c5059277a3a9173a35dfc0b8372d6e4d7823514131a94bae3fd7
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:d72842a7c6a1e50a68576c93f60047afec801cf83db64d7d3e34b25bf3c753ec
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:e6a7738cdb0e8aa9e1af381273b1c18ec1414b08a66fc851f6e11e5aa7a6706f
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:fcd5fd10de7334bd59a8da4997f91008ced95c6cde40d56d50f3b3e32c0f19e1
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:c20ab6fa719506bca942669ea84b113335b05df3ed275257b88be234cade7c8d
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:2ba6782af0d7e5c4d623de31814f43f4f7c3c38d09b57705021957d334067f87
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:d055d2b35a663aef3e1aafdbed0b12957867c0670c946ebae66e9c44a768bda2
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:72901cf4e091fe5f6a577144fdffaea8570712e9887a8bd715d5ae5bc346ff1d
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:b9675b22bc8759903472c1eb0e03f90c75460f02799d894294c161641cad9414
        - registry.redhat.io/rhel9/postgresql-13@sha256:ecd5a5583db922c517c735797db3fca587c0ef0b410e7dd494ace2267d89cf05
        - registry.redhat.io/rhel9/buildah@sha256:6886997da8c92214e99e3e8da47a0f08239172d2544b4edae058d5cb018effab
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:c90a9c3a20f088241980b76986ba4f43b0243fd6d94a826fa7b5888e01c13093
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:776d34b0ea2e35c11ad4050e6b40cf4b40c84dbe0d5f3e6ad669c494a210ea8e
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:8b4bf043e0fb237553503d63b129c432600bb4e65b2096b246b257b404b42be9
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:eaad37453a5ebef51c28d4a068f56e9f394e710c72bb86342cd2189001420836
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:8ae0ca6411e4fcd5c0605344bc6930cf122ffd80e71199509d58423a8b41de1d
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:dbc462cb50385dec3a220b5d0caed0e0211c166182026f86e3c51a1351ba0de4
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:6abc903c784bb455e1f2fde45a48afe775239668e1d0dbc1b3cc6afca3c0b477
        - registry.redhat.io/ubi9/ubi-minimal@sha256:ac61c96b93894b9169221e87718733354dd3765dd4a62b275893c7ff0d876869
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:e27a30d471dda4983a5b8192709d283a9d082558f3caf31b30db2214e418b017
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:e63d956a0534f2e726301e00e877c4fe2f34ca4559e59d2df7567c04ded311a3
        version: 1.18.1
      entries:
      - name: openshift-pipelines-operator-rh.v1.18.1
        version: 1.18.1
      - name: openshift-pipelines-operator-rh.v1.18.0
        version: 1.18.0
      name: pipelines-1.18
    - currentCSV: openshift-pipelines-operator-rh.v1.19.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.18.0 <1.19.3'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.0.0
          - Tekton Triggers: v0.32.0
          - Pipelines as Code: v0.35.3
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.21.0
          - Tekton Results (tech-preview): v0.15.3
          - Manual Approval Gate (tech-preview): v0.6.0
          - Tekton Pruner (tech-preview): v0.1.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.19/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:22cfadc42e793553182acf05a34fb6a7ae0d7055431c2e393699ecfeb171b173
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:7a0c6794e9284209dc131f3087ffc01fd5946e20a9261422808c6dd3938407de
        - registry.redhat.io/rhel9/skopeo@sha256:993f799d58c48151ecae7a84e63860f6c0039e8e18efed5ad67e7752f4b7e408
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:fd6f4cd7801d7c20cf5a6e546022136404212da5d9997750e643f6e640b00e3d
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:2c3074ecc0c74d508020bf1e11667d868150915b55d436a43c2b4fc22815bf48
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:80af5fa2d87e7c405f16575519293ca7fccc49d635b4f3d0d334dea4fbd28721
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:25bcdb30f0fe34300c4f61f62513f4853f65fe11bafde58bcf19b7b3c3fb9af7
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:251c87e2c48e7f377bcb399ed64d50dff11de3459629a73a63bac17c353eb16b
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:363e28a82ed022e507c3976d3a42ee63c053eb7871712c213b12235d8234b7ef
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:b247e9f3a2db32b575999387ef4f91d685a67c1f3950a6e265e1fe14c1fac196
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:6aaabb9af37e46a856371e29f141de6b0fac623128099bb927fc30f8f61e9db4
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:d000daec36a6cdbe41a85ee4eac24a712f7628441a412f751baf8d7778626f62
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:8bcaeaa751c61eb927d6353984f81d6a92b4bd4f9a5be9dc620a11b06ff9e6ba
        - registry.redhat.io/rhel9/buildah@sha256:a6e04c061b8d261dba85d01dcb64d3f7dba8c0b9765e03f05ec79850a248d75a
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:41d669fbe8bb48481890fb9fb52de9d4cac98cb74e6a6b284d046a8d21b06e58
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:28258903330a0715141e4685b3760ff18b0d75da0694985bba7c3345d7808b5f
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:90a5eb2c0decb72234cbea5354bb7737cebc53c467229b2c16a1425b4bf408dc
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:3f7e1d83eb1e21860cc2ae0bfe3abe8c9058e92a69698dba9d5871069016e90d
        - registry.redhat.io/ubi9/openjdk-17@sha256:cdbd64ba1b7826d09bc180fa19aa6fbc09ddb5bff5e02b7e1351353bbabebb78
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:a90a6c269b3fc4c77652d4862878211eee5eca1d8047e561d07e3ef1292e52f5
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:31a40f8095a3f8cd04750810d964572339965ae0b2d357c57c0b671a95e1cdbe
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:1f870e70669ad584db1ee31b1dcd822d8b8d0cc03028c7cb09238fb7abf19a42
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:b8cdf2a1873a3df1fb8b8f6d58d06235cef28d228abd70d0bb83cdcfbe3bc230
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:3b1c6f8acdf17a5185413147e0248baa559378ed79e08cb098c724036020576e
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:c9fbb018a70df9a542236ace556fa3343ae350f43c0629f13e525854508f73f9
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:92b8c92eb0bec102215b97f1d60d76c7a57747aeec9594ce1e0e46f7c27391d6
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:5f39f9e61ddbce92629e6502bbef737365c8a1bb2dc2ecaf9f9d8092dea844c9
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:aa1b6f374643ed4e89ba194afde8a5138e401c8cdeb45a6bb5df5b608f58358b
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:22095c436562ff92771d9349f57d007d82415dae519e3da1de7d51e3f1042a93
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:198cf9e5c9641757a0453dec2747a24b018a9b85e10eb0dfdc7a2e75ed09cf6d
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:ba596c629fe8b5b70db97923ff11392dc12115c8e7afefee0d48c12440b8dd86
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:10c620aa3a9c4bfbca185fb460e2521473a5fe6d518342b89a3dada5cf14e65a
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:13608cfe2b87535776c65728aa40cfaa84ccf63afe64d1f2e799ea0e26819239
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:68d9007c5abb2ad3b7666249a61d9428c2cf07165bd8090a6e6def0b7cf70fa3
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:37e2f1fb8a2b4c7ac3d5390ba68a9bf98a0c83a38567b6a9778764061b3d2bab
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:c3ba6b9652205407cc4d4d47b4a928fe8fa86ce4427e6b44ee1d53faf3455f21
        - registry.redhat.io/ubi9/ubi-minimal@sha256:14f14e03d68f7fd5f2b18a13478b6b127c341b346c86b6e0b886ed2b7573b8e0
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:d21cc7e094aa46ba7f6ea717a3d7927da489024a46a6c1224c0b3c5834dcb7a6
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:2d2499b7b12aace07c4b5ab4121caad445aab5bc8dd73377171683b36d417d73
        version: 1.19.3
      entries:
      - name: openshift-pipelines-operator-rh.v1.19.3
        version: 1.19.3
      - name: openshift-pipelines-operator-rh.v1.19.2
        version: 1.19.2
      - name: openshift-pipelines-operator-rh.v1.19.1
        version: 1.19.1
      - name: openshift-pipelines-operator-rh.v1.19.0
        version: 1.19.0
      name: pipelines-1.19
    - currentCSV: openshift-pipelines-operator-rh.v1.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    },
                    {
                      "name": "legacyPipelineRbac",
                      "value": "true"
                    },
                    {
                      "name": "createCABundleConfigMaps",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.19.0 <1.20.0'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.3.2
          - Tekton Triggers: v0.33.0
          - Pipelines as Code: v0.37.0
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.22.3
          - Tekton Results (tech-preview): v0.16.0
          - Manual Approval Gate (tech-preview): v0.7.0
          - Tekton Pruner (tech-preview): v0.2.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.20/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:bc403f93ef6c5526fa30152eab384c4a25bb0bc6ab4158a1e24f637b87a57259
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:980ed7e81ea55522e1681dc1e9b170b68d16f07061dd1f1ffd142367d59552f1
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:bc10e7d6122ede5cf4f2882c200447683255465a1080e826761e90bfb983506e
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:c3013c347bca8a520443a44ac2817fbba5744a9e91069ac5f09e91447ebf6b28
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:1b6aecfce45ed3163227b845b212e83385cdbe999edfc2e9387a0c76f3f8d8e9
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:5683321a5376e63c3ff3932270a1850d5e261f76212020ce867334166f90d2f7
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/ubi9/ubi-minimal@sha256:2f06ae0e6d3d9c4f610d32c480338eef474867f435d8d28625f2985e8acde6e8
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:7eafffa6e4e04f10c1d4e84e02d51c00eb8649c60f8f9a13da14492f24634891
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:d56d6b46331b4569077ab1e7a7b609ba01f0a1dd62b695d62f8aa0f32ea1b621
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:e63298d07ccec5804dfab4e42bcc8163c29cb3ac371c91248777423a4edd31ce
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:8f0494341244a75546a4b80b065702be04e6d200421aa4e7ec90385c9dabc6ab
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:c31af237afb84214bf2a4c06a7636ca071fe97203b41da90d7878109ccd64784
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:115eea1a5ca87f3b6b72db1df05b7386bf73dc70c9c2e4dc5e1a66700d960c37
        - registry.redhat.io/rhel9/skopeo@sha256:a65a413f8a2864389a09dc750690d97afbcdc5c70821e0f85e99e8adba7954e0
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:dd941acd357ae2c0fd7e15fb8d32d7ee6b2847435101dd1e1f107db48af18bf3
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:5021158a4752a26f8b33adea7da66695b22d2ad155d346d35dde5f971bc1a55b
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:823e018e0329d764ac694f177ce88af9074f364c29a1c17929f9dd0b596d03dd
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:14dab8d87323fc7d112213f7f4db3ee7cf3841747de669ce710ae6a794053457
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:bbd1918660be7b3c78777dcccc87d3b26bf5adfc077844ceffbbdd53696dc0a3
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:a7163a9e1fe891de021b1003558e6e64fe17ed17ab0820ba16bba09aa112401e
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e7998e113f75e8f29602bae210c8b66ed06634b699f1d99c5875ef879220d733
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:ec74a562c18fefe36ec7ae3ba083e4eb76c56b562f916578900f8767ae98e6c2
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:c4d05e546c466c57379fff502816012f615be073eb8cdce54677b4ab206a3f5c
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:c0f438e6095aa81f3f7d820614e0414cc43374014c1505d7641bc1dba69a3903
        - registry.redhat.io/rhel9/buildah@sha256:18453ab0a62154283aaf4c45efd7543d1c8ecca1aebae46dedc7ddc0f410b232
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:818bfa4a3a874974b5f7f8bcf2faeb6c05e1795381ee1ceacf89f17b1ec0ef7f
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:9415e889ef98df6d2b2dca1e5d798fb765eeea0d9828e48424374d5267a6b0d1
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:a712bb411d2433b01793dff25785c1e265fa865a52f7c8732c9fb4418ab61757
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:3cf990fb2fc422f3527a7c8b43d06ccd4d32bb5487ec8cb2d4a97f0f5d13f170
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:458f865f1b246c5f0682f5956795771a0515b89edfcffbbc0da1a3d3f532645e
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:96db721939b22572d7d8d8b95dcb5c92fdedb98e9cd0aabb7d17b703a915b07f
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:036250bcdc6fb95273bf07948dfdd4dce765ab84c803ceca0d0b4abf6ca19dec
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:161df929e2bd4c5acff594be31d5458365de5ae0d2189bb2f647c3fc6657fa28
        - registry.redhat.io/ubi9/openjdk-17@sha256:555c8bcd2596637befd1a6da4aa949c5a2640cf66e09c2dfb932afe37e39095e
        - registry.redhat.io/openshift-pipelines/pipelines-opc-rhel9@sha256:c1bee4a2bb495de6d4da962b6e7b2447caa51cc36c3005e4083bd31420ef5102
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:901b5108ceb669c985e75f5fb033c496e9f028863df99ab32651d1b08433a352
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:df15eb1f1f5f237bfe481c16ab3d40bf7fb4123f3c930bcfb121e2cc8dc42882
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:441e1c7124c1674e3d4c38c3e82a4fe5568e76d13c89c04bf009d6b97435de09
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:e3e64abef53151f6b15f15bcbff85668167a571e98badbd8a2ca252ff3a5b8dc
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:2d496752b19f765f1ea36ec94b873ebd8fa604764a56aa67d3ed94b33e5fe75c
        version: 1.20.0
      entries:
      - name: openshift-pipelines-operator-rh.v1.20.0
        version: 1.20.0
      name: pipelines-1.20
    defaultChannel: latest
    packageName: openshift-pipelines-operator-rh
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Coroot Inc.
      provider-url: https://coroot.com
    name: coroot-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: coroot-operator.1.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "coroot.com/v1",
                "kind": "Coroot",
                "metadata": {
                  "name": "coroot",
                  "namespace": "coroot"
                },
                "spec": {
                  "enterpriseEdition": {
                    "licenseKey": "<COROOT_LICENSE_KEY> from https://coroot.com/account"
                  },
                  "ingress": {
                    "host": "coroot.example.com"
                  },
                  "storage": {
                    "size": "50Gi"
                  },
                  "clickhouse": {
                    "replicas": 2,
                    "shards": 2,
                    "storage": {
                      "size": "100Gi"
                    }
                  },
                  "prometheus": {
                    "storage": {
                      "size": "100Gi"
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring, Logging & Tracing
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/coroot/coroot-operator
          support: support@coroot.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Coroot is the Schema for the coroots API
            displayName: Coroot
            kind: Coroot
            name: coroots.coroot.com
            version: v1
        description: An open-source observability platform using eBPF to gather telemetry
          data, providing actionable insights for quick issue detection and resolution
          in Kubernetes and VM environments.
        displayName: Coroot
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - coroot
        - observability
        links:
        - name: Documentation
          url: https://docs.coroot.com
        maintainers:
        - email: support@coroot.com
          name: Coroot Inc.
        maturity: stable
        provider:
          name: Coroot Inc.
          url: https://coroot.com
        relatedImages:
        - ghcr.io/coroot/kube-state-metrics@sha256:080a85e77b1e530df56f84c8e0ee2523c1f740adea152015367b87edbfac078d
        - ghcr.io/coroot/coroot-ee@sha256:09b9177ac07c2d26f4bd16050d07bb1e2132431055c32e047563083ee09194a8
        - ghcr.io/coroot/coroot-node-agent@sha256:3d552e98ca2ff8c6e125496857306172ec6f8e134ff13886c7f799abef29d1b8
        - ghcr.io/coroot/coroot-cluster-agent@sha256:674bcb147354873eaa0ee9c1328ddf758d43adbbe532c9fbbe41ddae7addc114
        - ghcr.io/coroot/clickhouse@sha256:afed1054626598371975c948e5b9e4dcaac54e8059b7ad9f0b905ef1c99f2260
        - ghcr.io/coroot/coroot-operator@sha256:13f1a4a07877c8fe8230a0bb7282ff47451007c9cf689a995a37d16697b9268a
        - ghcr.io/coroot/prometheus@sha256:e7117bd86e1088fc02ac12758bc94dba1029de54a5bc1ee04954569d8b79625d
        version: 1.1.5
      entries:
      - name: coroot-operator.1.1.5
        version: 1.1.5
      name: stable
    defaultChannel: stable
    packageName: coroot-operator
    provider:
      name: Coroot Inc.
      url: https://coroot.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: cpd-platform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cpd-platform-operator.v6.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cpd.ibm.com/v1",
                "kind": "Ibmcpd",
                "metadata": {
                  "name": "ibmcpd-sample"
                },
                "spec": {
                  "license": {
                    "accept": false,
                    "license": "EE"
                  },
                  "storageClass": "managed-nfs-storage"
                }
              }
            ]
          capabilities: Seamless Upgrades
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.0 <6.1.3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ibmcpd
            name: ibmcpds.cpd.ibm.com
            version: v1
          required:
          - kind: CommonService
            name: commonservices.operator.ibm.com
            version: v3
        description: "IBM Cloud Pak for Data is a cloud-native solution that enables
          you to put your data to work quickly and efficiently. \nCloud Pak for Data
          includes a suite of services that support you on your journey to AI.\n#
          Introduction\nWith Cloud Pak for Data, your users can collaborate from a
          single, unified interface that supports a large number of \nservices that
          are designed to work together. You can use the Cloud Pak for Data Operator
          to quickly provision services for \nimmediate use. These are just a few
          of the services that are available in Cloud Pak for Data:     \n- Watson
          Studio\n- Watson Machine Learning\n- Watson Knowledge Catalog\n- Watson
          OpenScale\n- Watson Machine Learning Accelerator\n- Data Virtualization\n-
          Db2 Data Management Console\n- Db2 Warehouse\n- Cognos® Dashboards\n- Analytics
          Engine Powered by Apache Spark\n- IBM Match 360 with Watson\n- Data Refinery\n-
          Db2® Big SQL\n- Decision Optimization\n- SPSS® Modeler\n- Db2 Data Gate\n\n#
          Details\n\n## Prerequisites\n- Red Hat OpenShift Container Platform 4.6
          or newer installed on x86_64\n- A user with cluster administrator role\n-
          No storage is required to install the operator, but you need storage and
          resources to install Cloud Pak for Data services. For more information,
          see [Storage considerations](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=planning-storage-considerations).\n-
          See [Pre-installation](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-pre-installation-tasks)\n-
          To install the Cloud Pak for Data control plane, review the [system\nrequirements
          for IBM Cloud Pak for Data](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=planning-system-requirements).\n-
          Before you install Cloud Pak for Data services, review the [system requirements\nfor
          services](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=requirements-system-services).\n-
          For air-gapped installations, see [Installing the Cloud Pak for Data Operator
          on air-gapped clusters](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=tasks-mirroring-images-your-container-registry).\n##
          Resources Required\n- The operator requires 100m CPU and 256Mi memory.\n-
          For the minimum requirements for services, see the [system requirements
          for services](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=requirements-system-services).\n##
          Installing\n- [Installing Cloud Pak for Data Control Plane] (https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-cloud-pak-data)\n-
          [Installing Cloud Pak for Data Service Subscriptions](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=tasks-creating-operator-subscriptions)\n-
          [Installing Cloud Pak for Data Services](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=integrations-services)\n##
          Configuration\n See [Cloud Pak for Data Post Installation](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-post-installation-tasks)\n##
          Limitations\n- `x86-64`, `PPCLE`, and `x390` platforms are supported by
          this operator.\n- See [limitations](https://www.ibm.com/docs/en/search/limitations?scope=SSQNUZ_4.0)
          in Cloud Pak for Data services\n## SecurityContextConstraints Requirements\nThe
          `restricted` security context constraint (SCC) is used for the operator.\n##
          Documentation\nSee the [Cloud Pak for Data documentation](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest)
          and [license information](https://ibm.biz/BdfVHT).\n"
        displayName: Cloud Pak for Data Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cloud Pak
        - Data
        - Analytics
        - IBM
        - AI
        - Machine Learning
        - Cloud Pak for Data
        - Auto AI
        - Db2
        - Watson
        - Data Governance
        links:
        - name: Cloud Pak for Data Operator
          url: https://www.ibm.com/products/cloud-pak-for-data/deployment-model-options
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:bbb457d4f6b7bab065e909756c2dae31af7162dce10bfb92f274aedbc98d0762
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:716b5823661ac8f1eacda11ffbe5f0ff08a0e9d2296d04acffb646b7526128a9
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:1f82599d91ab20d2024abe134d2f56caca1f9bca1f02d64352c81b5f1e0f1730
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:d6575de3156c7a9b0553887abaf8713a758d9c9916f9cdcef611923c50632867
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:113fdfb9395c96f7eb65f9a20e18ede8a2eae4aa6c91660be8fd3e517656e156
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:d0e0144eaf27e03aec1499e53cae5366e11a3da284536cd3b27ba5691b4f70d0
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:6a996123e69c7f2c3254f334e278ee0daf75fc53980183e1a72358323e41349c
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:1d6dc7a3039e6fa044053d4e89ef54ba2e83ceadd57bcc4613a6d447d027ae2f
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:7861185d55a5fd18d896dbc57501ae5cbc1de0bc02f1ae3fe558a36f0f80b4fc
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:46bbfdc1f5e418db95ba806228fa5696be22c4cf9eefc185ce27bfa68ec1d247
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:a95e42a4f9138f1f97937a74982c98d466c171d0e3e79c522ededba6c25b39a3
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:23ba1b277cf76bc28f952ce6f5bb02d6177f76eca8a8b4ff7925ef119a994a85
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:670a36f3400164cc9eec5d23422d184340e0c1edfb13a634e58f0345c325c21b
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:0dc719e58694fb501b8b36f3a375b3c0462624d27fe49afe6b835821e31393f3
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:257bf301338f7b123099e2ffe31c3f3b793bf28d6a0570a25b889271239579ba
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:d61b7b8e6a9785b59630d9c5776f72a29ffbf2e81eb753d7b11dcf2d9d77f1f0
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:30a5b2c4e3c231373fa18d8aaab16d3505cc11a8a5d4be36eaf99d7c0dee35f2
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:fbc6eab07fcfc2d0c1895f18e17f751e9481ad111875959e9ff08e14f81b9a35
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:aadebf0598c492639e7c1109ce681256731215b686fec808578bb09558a40315
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:6cd1825ba40071bca7f46c1af3e813ae7ea44a393dd3d10e3c83e62834563ffd
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:b3387bae415fe3a66db2264ee766cf43aa8f31e2b615b200baa9228216596a0a
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:508412a22fbd65074964150d6b270f38d6df898e17eb7ec5e28f279c4c2e1062
        - registry.marketplace.redhat.com/cpopen/cpd/cpd-toolbox@sha256:dcd74c128502dc301482eb302b34c21dc02173ae1cf522356d5b500ba42b4e25
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:f00158405adec00c7e992efa7d7a19a520df485cbbcf39972e579ed3c6e4d289
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-hook-utils@sha256:ec44be69b4c585e9a3d507e29bd287f5b464360d82c8e0969c5fb4ed3e95abc0
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:31470f873a0931d130484d5255692d445e3127398aae0e777a3c3cb702917743
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:f561f089c4e5037f7c04971d96fa024ba4df5e6d17834f0c44545d0e53c772d7
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:20863de342e07dddc9aa8eeb214fba5cb17478bb88613e4e1f167d77cd0af214
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:9460c98c7202ab368e732951c13bb6f5064cf18f69d36b3d26eedcfed233ec81
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:5aecbb4de6fca34f026de72b7480f35dbb2ee8d8ef1f22e524a85a71bea99a1d
        - registry.marketplace.redhat.com/cp/cpd/jobs-ui@sha256:49daf607918b378593485da16ac417056ee295bca94f1a5cac37b334381ae096
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e924abf025141f3b3437ed89830957c5915152d2610213090e33fde988bd879e
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:e6fc9d7a70877100b42e290cd091011f2e1c674a27985fce180ee245373391dc
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:8bd9fe1d35785fa0125b8fb3a60333219a50ee8be7a7ec07386e8e9d65316061
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-bundle@sha256:94c7a1dd290e2abd55c5b8998cf08ad174d3fbfc659ae6fa6fb6cbab61d5fe14
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:ad027c91c719fd1d931cbac8c76735474f676b228f9ab23d1f243201246b3238
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:2f40d8188dc8b41fb072256ce8d109167095469af416b322ac023071814cda8f
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:e4eb6a75e0d207232e479693d9c99ba863f8e62830f3191ca4d1e7ceaf3586d1
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:847751338f3b78cc06e29a4a26a18dc0f5d0bc24cbda23b765198a6a5c4503c1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:ca4da5a8a3c1e7add6ac630c3fe4713a08739f6b891b2290bb17026b94c31317
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:c268d2852580286e5c6770857fcfc3b06a698521f5acd17ed83b8e2cb9ab55cf
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:fb37fe521693bdb8b96400fd491a39c573545f35b4131fdfab501b7ba29713f6
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:4a5ed208b68613d5d88936687f6912b3aad0e93c75f0d311120a8e55200f5795
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:52ce08ac2bce7c02dc30e2c0aa74df4bde8f156debccc59960f02b4b0040ffea
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:d6cdb71edd7b7451a206d1620b35c321f88b3ba8028468329595cd538a45abf6
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:0f01a67524fa46829e455f654d3b9b00ceac44f3d870b21c022051dfcc8a00eb
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:b646ea3330f3fd9bb3174224de69f20ef58589a464d50d1c9399211141f67bd5
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:650e68c41252300836740ad495ad6776a7da94331e8070520e60a1c5e2c5f3ca
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:6f2e7cc917adb54b3112dc4909f645ac08c33e8c2c33266fbc8f5133cbcd1973
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:5e161ff695c0552d1333a1f993048bd055168a9be9e0a75adb7c4464c9c2b603
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:ff5a5be492347e75b309194661d7742d1c909c3765c9947c8b41f4c3af62b08c
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:7ac32270044a9ec08ddaa8dee62b61618d9363c8773fe44ba1dd86e1b53acbd5
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:c70d3c86d4c61bd87a427f7a7571f37fc4d0966ff4f420b73ac4a2af73ed1379
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:978cf4bf299cd3e8cde9d7ee7aa75b73fd9d9e6755567651d6c1c49036b8978b
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-usage@sha256:8438c2229b500fc9c40629280701856345e228fc645a5b468bee2135c61b2387
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:498ad7832e989e9356aeffbee96f310e3edaa6a7466320fda9b4f75034e73eec
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:7661b04bc59d64a4fafb2f660cffb99e3d326772749759ed0d2a1940ba147078
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:1f68dadaf2ab779dece300037f3abda1b0bb33603436b84336f8fb7d53d33644
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:0f46b288d060a15a48ecad8f328a8046bf232b8629c3d79da6a82ea712b363f4
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:cab91a231068a6bbef03ee8db32fadf11654317c4573c6714721a15d0e2a98dd
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:6c226a0ccde697b56c81427a970062872929eab73d99791c9dc670f4aa176ab8
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:67bfdf9115fbee782e35ac61dd93801bcb9e622d441806f1144208b4447f1430
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:2668a562eb8df1117c40acbc3600d6735469d210e5aae4aa03bf9f54d0ebfe46
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:e4888fadfadb3462e3a9ad0395a526a2631872f70caf291dc18b60269ded199f
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:f0c68715f78bd20c352c97dd48ac5a912cad58e19a00159b5e6878f6beeb00fb
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:b0735e396c09fbd08f328ce0686cd55fe9eafd5b759f36cf5677326b24cbb859
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:abd96661823553d623eb0ea1fac8271d64391243dd0f13eb4a4fcc858dfe10a5
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:48d237cd3d363da8a206ca5e092d0f91e8a5bc931ea81763498dcf6f28c0434b
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:c078bd65fb18b31ecfe03c832643bd6fd85f7330b94a1ca26f557d32d89903c1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:631a18f4b437e5092b9671fb4c86af05e99effe5540c4d6f964dc626466764a3
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:b63e2c10478735b51b6828bb11709569a2fce62a19b70157a8a5b1964820164d
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:47a3745c78280f98f2ee352bb9192cdcd2fcd6d4c1a3627e7aa42178a0f107fd
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:a28b71ef64c1967cdd36a16e01b611fff6a387b6a741dbcaf7556ed6be9b1f69
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:095aae63ced410e033d8f7f3d27bd14dc5b93a10068ef5a389f181115c5d07c2
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:00f8bcee4ba7e0c5cf5f0c52e9e7707df367041c9f4dd5508dbefe0dbdba8877
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:ec137115210dae9b69cb2899ed8a5437d94c8a658ae1ae5eb295232b34b0925e
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-velero-plugin@sha256:fe562192e86558a17d1c88a8fe21148bc000fa93c8928f7a3a897acce62e6798
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:e857323acd10f7b99c5b7fb4b0a5b89a808f68036a1d9a224eff55a7ae7744b6
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:1cf983f9b9cda18257066fc82e199bd3e22e2fcba13daf0f60c92b50a4a0e078
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:96d382b1b26a1e1e4074b7a1aa48b973cd4865bb5c3cf02795306e92c752a8b4
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:dadfa9ef372b8d165276fc5f1ee9c63c3f01367d156700810d63e1ba19f3da79
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:45194b56f9fcb3af83158eb0297248364db40e8164fb4fe4f5656c58aeac3fa5
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:0ceaaab514f65f9a1c40717e97ca652506f8b3595186222d470a99d653afcead
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:04f9ed67696fc193877f9da3d73ee3f6f4cadc7b8d09fd07fa12d0ee15f1152d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:9f7e6a1ca3faa5af1db2afe8693f6bf9182296be2a23e305ba6c990aee3fc301
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:562281dbb2bec5191a7c6b595ab853092e6fe3ff1432e77b2471f71ccdffcbc8
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-bundle@sha256:552fc53664722903eccaa7d3782a3e84413509fd357725fe0b0f4a077c48ab2c
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:cb9a33db3707c0c6045e6c823ecb21b0f229c17ff868da7d36602d0026eff351
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:d4738f64e36a8f0c88d40043eebaae1d561193895d438fb6d8bf1a042d0a1007
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:cb3620eb9aec24f7ee54c3ec6709fc223b685177500c503af8ffeb23642dcdb5
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:c330a321438943fdd264b292b68fdfb8be89c3bde93a7bd987bd92c9dd2512b0
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:9a970756cd76ec748b555d5bed97889b8c9f410db6765eedee84fad4976cd523
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:4b3ffae1ddb7a2c28f98199be6343ed826f8fd21b64fc9c584566341f44e44d9
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:48b7cfa31b9a185176c20afc119e37346a6f9380d74000edf6c729b4677ca44e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:a7c050ae29cb93754f59d8f83676275ee9a91ae9a9956395b97abf8c5e20e96e
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:65108559a5edc9a7c4f82962ab2293017e0873872730d60d8487963970761941
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:d34d561bbf8afbfba9362d2a9a58f2919faed1502abf72e11882dd9cd4fc9711
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:4f0cd4920e7cef64e9f49ae1abc5d87303102fc96819e0c3b699d86b8f8a3980
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-operator-bundle@sha256:086eaeaaaf440a1d4a254984afacd8d04583aafb9ec11ac072feb42a87b986f7
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:a2949c3868c2c91bf35e50c64b51fcc24372986744b55520e21e9ce31fde59f0
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:e21a346d15b286956f62bdd1fbb6d7b77a554ee10a475cb698a8796c42f3a267
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:6298c1561c066eae3d420f20212a3c7a8f9069a6996408685437588bf8335dec
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:47cd9776a46ae5fdac3a101d79a779baefc802ddf8e1f8fb02216f47feb55fa9
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:8444703b2a7822f82141afcf4f36bf255650d755321b9e9fe2cbd1bdea41d5c7
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:ea57c7cb1ccd5bce17f6cb35e326ab974718be8cd19b0cd38ea8112e91c51939
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:5d889be3dff96d1e888eff94bba94228aa46f4e94e5f30ba33077ef00692509b
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:ecbc65705de76804a3c7642339ecfb4cc0f5a61891058e07c20bf2b63f1b82b8
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:690e9b5218a200de93a7c8477fd2980d8d341f22be75002967201db09b5a5524
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:9c2093089ca0651bcabc1d01ee2b69d2fb0cf2a277a8b966cec9533950fdcdae
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:a1754e003811afb578dace78f9eaef91e7917e5468e590fcb41920fa71bcc22d
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:010bd5bb9943ffdffa8fd00cc8226e59d9583bb3baaeacfe70e5b00e805878f1
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:3d5c6573faf4ae00cbb21cbfe9b2ad5cd8ffb658df0c5ba09bd98b3bf54f9979
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:8cba5f256747002b62fcedb9066f7ea6dfdb1071eb359e9137fd413d20b29ff8
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:6f96b8b6afb00cb26fcf903b8e4c30b0d6344bb86d8bce2f2026fc8cb0b575ec
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:56315ce0a0a70f5899191a7eab856105489143f0e889c8cdd1b6b8c225f02eb7
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:d733c5a10a21cac2df2fff6378035151e5d0df57e3ed1f96bb3ac17c0c95b8eb
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:d06ef794046dac2c58163c201f55489b69081dd15dd91cab16387510a50e24b2
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:2c39ff1850b7a9d32a11e99393039ef74ab363f3fffec1e4fd542f2ae148a6b4
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:61ba2b05a6826f618853ba048496e2c9558e509a01648c74b50cce6d28d0d764
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:e0feb4d305c5ec9501b30ccdf8f5826cd8eb1ba5ba1d62d4d685101c785a708f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:c8786cffb4a27decb5dff3d2a9682d07d11b7096a689bda0d7cf105aa3156299
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:51e1275f6e8297418997b0ca9615642153371c29b3466f155e15d7fb9e80cda9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:f1aa6add76f59afeccdff84c9ea89a46b49c6696f62acd75ef8ff2f31a3080be
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:132670412a0ce6dd010869e569504d2e89aee0983ff221da371c9e30d5f8d8a9
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:b5a8fb14c3bb7c89ff970ca85f06dfcdca832e2b45ad86e8bb1b6c0a0e7e0764
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:ecac5e0327d939ca84d304b2fe95739ae43a06745ef775ee59e4e5378ced65c2
        - registry.marketplace.redhat.com/cp/cpd/environments-init-job@sha256:6af8f900af7c11eee6894e2c3e52f740b3c700dac0ee753fadef330230bd42dd
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:369f66883e040accb73c3deae984ed479e90eed242051b7035723f0ba21b51ac
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:7ae4dac24ea054ce527a507dea6c77a3233aca0047672ba5a367fd5742448517
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:8d7625a0ba656425f601b6ec9a398a6d30511fe1de119cf0e29f680c704ea225
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:7b2e9c373f9bfd91e16eff68ef2313171c69b94fe5a21d3d3158470c74f54a05
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:ad2215adfb8dd2dc600bf5097b7f88c257d5fbaf541409c01b36cdb67c130236
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:c41e4d3314fac0d2cf5f08e7db85801cb2098d37d0aa8104a347e9e82526e914
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:2e70036d7299168331c94af70a22e3b2063c56160f616362b1a2187c12d003a4
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter@sha256:be5fd8a8b43ae3792c58de1342cca91836287fb584bab003f7f9aeeeaca3f65f
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:54c86367bfe0ab658611856cc0a6c5de24c14a336bd129ac9f16293945a55238
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:603a6bf0e616f16b609bf157b9e21415802478aefc9c606e124008d0224a1188
        - registry.marketplace.redhat.com/cp/cpd/task-credentials@sha256:e162828d5ad07ceae4db0589f930979842714e3b766552102bee373a3a570af3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:4aa81bd0043d0bac60c9f1b0ab56e45a3035c88314614e60eaea48b2a29ffa16
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:587af68ec8b4974b29da46ede525c70054703fce8d5934f0cbe68739c6058fe4
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:fa01c7d11c23a47bb84375389e69619455c2879b5ed6bb5b9b753b945dcb02c3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:08dcf99453b4dc6dfb5620410b8f7e4eeb42fa824aedb2e480e2fb40ec8ce347
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:0164d8fbea591bbd9d67d84603162c3781b08551eddcc2a8a842aafc828af18c
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:c957787ae587a31e6a497b46649da9cd46102a2a3dff29243fd5cbac96e7e2d5
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:4e65e96603d5043e78a517a4c6e5debc5df7e3747fad4d8025ef1fdf1cd6a9da
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:e3d8d4d5285d058130345d6108ed4056ea9c871ed8c0575601a94629e86ec691
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr@sha256:43cf735b51e2ef45a911ac7032f03c74cf0125e2d2801258d4aa04831e1c6ce0
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:4a87dbcb7d2ca2d121512fbf57b645eaa9ab745439b79722f474b84efa9eb391
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:6fe90f051d9261ad224dccf4d9f6bc9f19ce0e8d22fa010a46b2752374bce2b3
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:9a537eae7e581322e2f0764ed38a797ba7cd260724a285f24d5095058b7b47a0
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:3be006596f8e1d36865a9000c77df3223c264e807bf0aae6049a23307e8079e8
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:d28e6a3c65b580051ffbe287492dfad28dbb1c669cb76aa3df23bd9428f4bbe9
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:8cb37bb0f0461efc603cab89b5420c56990aa104fe39c9b77f0393bc6b641172
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-bundle@sha256:98cf2a6c7b7af5d7cbfbc61ce140201272b30b12edabaec2b2d4287b0690d89e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:20ceaa70a3ab5ed8a2e451a1e2960d04c201f0087fba1679fb80c57c6be6117a
        - registry.marketplace.redhat.com/cp/cpd/job-trigger@sha256:f2ea487072393a24e4c6864b2107c1fce072ebba07f9dbc2c139bd37fbd062da
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:83038bb7e3f4b4637f0d5095b42dc906e7fc27bbffcbd66f6ad93dc9b1cd6bbb
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:75ca37701dea1bd079df8f8cf302cf95e8dd4ed817fe126b192ca368e92fb78c
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:b10ca594f31751faad4050b4f7717e0e98c7977fb7e26d4866e1c14125422225
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:b2034fae4f2287c8997104154064f2712b98d3c4382d1faea3df777338711f23
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:849ecfeaa36fe365c7dd8f1a76fa0d263a1196fd16af805dc66035643e8e9d9e
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-api@sha256:a76d3114418a3df1d4d3cfe03da587c37028ca1697795e2535cac9f6f190457d
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:481391c273cffdeca36f91543732ba1f082dc409ee851ae95d3c1d205a7a6a08
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:c730fd4a7baa3f46c03f6c73e2c6c7c0ff24a8900b79dfc0f2fd319909edce74
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:9d4ae40d34e38a48ad2ad6556e997a42f6308006d5f6b8638cb7bc4ce5b1e40a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:6e0a23c3434046195428792c45c95869d81fb4ac4ba75444e4343963fda6014b
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:deb66a64288a370c5da5c7cc0ba4e2ba4ab816d7c5a080a0c505b05f55fa6072
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:443b51f8b10acc85bdefde7193e0f45b1c423fa7fbdcaa28342805815c43db3d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:53d1fba68aa0227c2f4c8de26837f6c114c56ca4a397826e1f2e93bb857e5b7a
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:42ffa27ae47b577cdda214e0fbc46c10bd108211e7dc58777d8236c59cb78766
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:99589ee11b50af8ed6355a5f9c2272ce1fc7f9a13e80fda8eda5883a332782a3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:a0273c013c5943c9b9a223cf1aafbc96b3cdd2dbb147c73d2a9ed49befdbdfc2
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:9a0c76fd9ef9b2dd9af5b0f29e1813567ab7f7d35c1efdf71da1a315f2942b45
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:e3138406cfe81021370924992dca891cba22b3be2eef84546faa0e8f8aa57ad6
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:62e2b059f565f87beded59b36962c457c9662d1e401a78766fda328ea1aeb78f
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:166a3918b942a4ea7b6c371a1e30fd4c928586d7ffd5b6919c2fc728a086ccbb
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:d30397b74780cdc14930a3a02dfcf6b99264847702afe0513ce536983c32e2b1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:f662be48eeb6dc93a8a4b0c3374f59bc342b5b8f5360437978cbac8f74c3a084
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:12eece1fc0a2e14244a2f57dcf7c198a792ced9ad0d02c9606645bce1f295877
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:130872119947fe91cbfb8dd82b3baaf78f7adcaca5933ca394eca3f9ee84c8cc
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:f3f8c6c395ae860fed5761f67ce092a122e7073edbc205c9167c27a4ba85094f
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:aed24dcc7ec24707c978a17b86501b65c6f0a77bd3ea0d173c28684b9df9cd91
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:8b7941a3f518e952766a2e6f88acfa7c9914b6d2bcd4ad9ddd54df313fa30347
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:722944bfd0eba280255f4c198f65f562423a26e2c4ad3580591ee71cb6e9c2ef
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:0341c499ae144ee81dc2b92e8b99ee313ff6759374b7c1a6469156c20f522921
        - registry.marketplace.redhat.com/cp/cpd/runtime-manager-api@sha256:a37773df07fe757dd698e3f09f8687902e7b3592d97855e703dd92ecc832aab0
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bb074bdf89cbc7c703fadb17639608599ab4f66db9be5617e5f79a0550676790
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:e673f1d0c5a0e32e9abc7ac40eff905ce411f9f3d8ca0b03160e23ba5661e041
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:f2f256a2fb787419b25e99c765003d25d1edbbe60042151d91b46a5d029fc0e7
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:a16b822b41f5a745b475ceb353a459e998f9f43bb9bb867ca1640866ba51780e
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:8a6ce0dada706056efda73ad0cc1547566003e0a7aab8ed379b50e4776861b94
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:ec5e4a002cfc63c1ff1bec340111e8a1755c8624066445db36c23fdd1c3061f0
        - registry.marketplace.redhat.com/cp/cpd/environments-api@sha256:6dd60b530f81d7a2d55a534211f7f128814abf87b8cd3575d528dba53442c44e
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:b0c8bb70a01bb19c685d019e89f83d8ed327bfa670a13b78d3e803fed77499f3
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:2f2590734343c99183dba7aa9fa4debc73377ee707f9bb090680750e62acdd68
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:25420a21686a32c62e83b99fcb215df015dd13fec10f73f397f927c61226f964
        - registry.marketplace.redhat.com/cp/cpd/spaces-ui-job@sha256:95dade51444e30ec034c12207ecbef675b70889b81970f49f10aa75160b2f947
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:686700e561181ebe1431f71eb9aa11b550bc4361dab8a3668ce0c7ead217f79a
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:3d83ebdfddcbe6eca38705fc3d575f2e70580cf9f7e8735941afee0a33d3b5c5
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:99c863002ae1e93e97792fa606128d05063abbedad9e9404acf28c745ae1a5c4
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7d9b2d61a4eaac645490d1c9e6db97ba193d65015b6efa8d5a1f3ec6cccb1ad1
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:06403c4ad99655cfc10fde46b21e84f90c05ae143bb6d08464dbc8954486897a
        - registry.marketplace.redhat.com/cpopen/cpd/cpdtool@sha256:8ed09b7d65178aab93b77b9d39371bc86b1022f32c5e63a9192baef02c71016d
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:ded16bf3cc993bfa5af8907a1517cc34ecb511948e4d9ac73bab68888af5543b
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:107911ed6a60a024d151ec3724ddd6d2d1402aa70a519b9dbec1a519d88231fe
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:0157fe80f2c67786cacaaf1302c334a6e0ec0b941afd870e0c4ba163ce394216
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a24f4cb635e8b5759a5e2bda20ddefc6d5ea9b16fb1081592885c54aefb5480c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:0711e1c6001586c950e74fbdeef8f04d20ce7b93de2ad033e82d2d5421338936
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:ad892ce5e6a3f69ef484eed68e92d9c972597a86420e22ff80c247689ba02905
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:39700bf47b93c7f0db1894148c16c60eaf406196034351f54573f0a94c4c8aac
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:3112fa03dc9a530642d329666a0cc64c06a04308beb1ddca0f7472915719ef5d
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:66c1d2493e6721094e85a0c65011523004f9589f58b7f68cc6f33c5e5adb6984
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:68ae72930a346d0fa0871cff573c64285266311208f2bf64101697a540f60272
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:2712755961a1c1e23f30d49341c037858c0965dbfc065aebda88c84526291f21
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:4b6bc5640d9e68519a7f543ddbe3f16ca200f32a74ce9c07e996cf122da3b59f
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:39e6aee0bb64fe6dc18ce81496936b113de6d6746221655cea102e04c11dfa33
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:07142ce0d259412d0a9908b63dd78d845ef6b7c5f8a24442b93964b4c38372cd
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:525924934f74b28783dacf630a6401ccd4c88d253aafa2d312b7993d8b9a14c1
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:73c40c6c72b7f2399fb0ed7e462baa72c3e8b5aa246fd2fe1082904b4afc0796
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:fcb79d76f965e74f7e3459abb92e3561ba2f016b34e2ebb6f88298bb19f7bc11
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:46893c166cabdd8b3422a190e220ee2e34861e797aa068100ccf53da349e3fe7
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:a477d47eaf8f8e569601d3c23c1def66f31c098ac3f5a1ae350618c460917dd8
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d51033417b786dd2dfbc94e648db00ba68f316c23b06a0bc9e36a5c22f579124
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:86cdb5b6543e7184361cbf0b0fa556277232460737ad254e920756c6d365e47c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:e6ce3c998642b235080daa6c0b954733bd3fe973fc8d85a10ef0cdae2a73aeb0
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
        - registry.marketplace.redhat.com/cpopen/cpfs/cpdbr-hook-utils@sha256:63a5606df81013b8a7b5cb07bd1f29bf032c8998db843885184fe86a76fb93bb
        - registry.marketplace.redhat.com/cp/cpd/environment-specifications-sync@sha256:894b4624d8ead7286cf9068c6477300992c32d96339644fe764bcb29b1c8ef5c
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:3a872814d01ded278afc8dfa6b1569d9e933e8fbda28c55a991da010b4304c69
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:351375de35d7fc9d940b905b867ce8a00b7ff1416e70d3a100918797fe3d8c56
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:00c30fe372524fa82b5ed4a637509dd9d4373dc727a9abc706221f832858f9bd
        - registry.marketplace.redhat.com/cpopen/ds-operator-bundle@sha256:be2ddb821274938e5ab926627bd998279b6d19b7789d9683513e85a4b05bc2d8
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:edfb732843a53254ebf7d4a2f109af109623be8fb9bbae94366d30febd2d0494
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:ee15594f82d374da81dab59c559567af74153524a6d9bfeb28ee085bd4109b37
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:3ddc6bbed1ddae1be71043d177cbcb58830d0dad542b06567fcfa254dd11e626
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:75595bf2753fb43391e2a496fb6d59faa8785babebb4aeb1da34cd5e72c93699
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:bcfb9b7d3b379ee2bb0b449cdf5af3fbc88f99c77641441ca392ebed04403e00
        - registry.marketplace.redhat.com/cp/cpd/environments-ui@sha256:2923664475c7a8ea5e63055c90748db37825fe76c7975b845f1185d4b32ee4ea
        - registry.marketplace.redhat.com/cpopen/cpd/ibm-network-performance@sha256:30e0b58531f876a47174c14bc6f94a3c38a40f757403775a06745d4ae992060c
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:119be0bf4c60230c43bd64c0ea252108a0173adfffc9f05e5257ec260b335599
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:3c1e08c56430e9bfc0c831be2394fa28e395a3341ce5cb5201b08828298deb60
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:2698150edf65a99a1d24bf3e8336f02b57fd964fcad283e182794634b033af89
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:077107dd9cede2c49beec429decb66760372a229c476ca1089e42e4ed5ef40d9
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:9782a9de863ad10ed36341d07ea5087e40f6f213cd4453c3ac2ec100d8733ce2
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:894d67753dd0c5b8625369e53635a1ff53066d3910b73d91e5625dd8907bff43
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:6bec687b2bd5c4d04403a25f56d2245f822e914907771f26c25166462976258e
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:993f16d744a81c95e73339d3c8d5880b5cdbcf29beb914fcb2c6403af6d628a1
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:21ce62d80df2575c3aa7c8dfde706cf8af01b2fc86f72ea68f767bc1770fe733
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:2565a8ba9aa103956d6902f2bcb53ac13328b647a2d7a150bf80ea2749d5b972
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:17bac3307ad3843afea77ed213279603b5f775f24dcf4857095baf5029409c18
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:749dd365ce85a4c9ac8e6ea7f90a34b791a7bc5300319daa85d427b04f40af65
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:cff7466aa6063a2bca677815df01208822930f8c12784c1c406fa352d6bd6cb4
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:d5feb4aecf8f36531db4c52e672558c04f7e314a1a791b27b5638202f9477bf5
        - registry.marketplace.redhat.com/cpopen/cpd/ibm-service-functionality@sha256:ea4c32a19b1549698c8a68fa030d7036d3ea3cf2d000f3523b6e9fc7c6c708f1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:0e563d51d601adbd710833cbde26ab8f857092f85686213adfd6fb3fe5e2828f
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:f7e68137e1ca58a63924967a8e4c872912557fa68af8a7392e3e0d4aaaef1e78
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:009d98ff8e3cf353cf8ba71a973228e1e409e4de21373498534a6cce01c07ba5
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:fada47a5f4824d3cd5bdc262b92018ca0e697e18dd373f92fd93186fb8719a1e
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:f88a7294aa200d3c633c0a25c4e4ccda52c75de174ba508d847ffa3092d5b8d1
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:2d320dee224cc4818a84493723f2262d259891bef89dedef09e1f1e6ae5686cb
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:489d7f80d83c9890600e3c9c4cb9442a6f42b9c6c4f13dafdf0f614b72e67143
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:934bc42fb29cbe616edcd93d2a060dbffab5e5a8a2ef4caa7b2e538f26167ee3
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:96e6db93b2d16ba590252d6568b68f254548aaed73ec17e4bdd8d530ea668be6
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:5f0974565de8b045d3faf6acac7c8ebce6cd35327dd76290b39fe48b27ad5b62
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d4dff1a8968c18d984e16c4c9f1517c3a10e23a11bf166c64d734773324d4c9d
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:c411456aa41fe705b47dec60eb49c6d65d4b7bd1a64d1e0545817efc95afc67f
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:f5527ddbbd0b9878555fce852996b13020fd91b1604d8c03d34a61ca2fdb2d1d
        - registry.marketplace.redhat.com/cp/cpd/jobs-api@sha256:173b88f90078e1817686ac949f48387af70a706e23e6fdab4a7c6217db5ad86e
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:85669f2a35ecc4ebf6b46fd793ba6a6bdf996164c00662e65cd7ceece8d60cda
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ed81ba4ab2caa7ffd1cb90e11aa8915df1c5bc1b76aeb4c87532c84097bb102
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:23337786ca20ab624eb428b0d53490975e2b08a2db642ccc76d974dc0e455f55
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:b8c83fe5215b7412765181aaba84162b6cb188b5765e45972a6f29b2a9f19826
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:811d1cbda1d0287dab48cff90c7091dd1bc8aea385f562c896f2b09f075cd1f6
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:7a0e784b0722ab0a4c62754a4e817a1f25e5a2d5e960de76387971fa1ff7da41
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:36e3b8e88cbb895ca89288ed385fb800b888743844f50fb2233ebe5e6c7498c3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:de7094acd65a80daa67669d8eab809220dc715b9cd6aaa96c58e5fcbe34c460a
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90046eb6170e12439ccc5aa343f19a4d4fc99a10d98b383cf1dc0e020ebeed55
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:164f6063f3a50a36e28a240ae7c060c4f90266b7ed799bfdc075f90628bea7ff
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:78fffd8a74eb6495f104ef346bdd03a5636134e2d5e8da9ef5c2647a9ebc7624
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:19bb32e0389361e8314e627bcad2e9bd93984289485cf632466834353d9c3449
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-oadp@sha256:35594a16332ba3ac00226414be9b85243bb25dd0648e256d7ef9d42137dcf914
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:787e01e29abcbb3d1fe58e0b6051067b309085459983f38d6bd18d3811d608d9
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:2f45aabc99eb9646db4c74686569a7f2af3f27cc44362a30c92d171540bb274c
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:97b9ecef8caef156a0eb073b6eb2a34b63209b44511f5694a243f11285d88662
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:bcc579eac4be59609102836252ded82a17ee0088e5ce28c684fff41c80232a0e
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:da38999d66199df0eeabd7c9e934205fb858cf2c52be8bbaf6b34adb894dd38e
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:e29885c1ee083a73b952b0de8d1cd675b2b2830abf44d69a8577f56972a601fb
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:17f016c11039c34f90f67279b45ee488740cc3d033112a3bd589bb1a60ce6812
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:bcadf3d367f76d3454e119a9d39f6fdeeb6cb806c282f21a1bc3c38bed41b427
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:ce57bc323b95db548dcd2f90ec938e708988cf6103579e793ee39d36359d60cf
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:5f63e062489b8e1bf49b5943e7a836174fb47af8dbbc8092ace5c8b72baabb99
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:740a1d87a43e61e32cfdea032d852af33100d959c8b57ff2f165f4baf77fc011
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:d72cef9b19ea48f03f7658dd55cabe6e118be0209852025845d426b1fbf0e0ca
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:1a1bc54977b160b70cc00cc4eb1380250dcd639fa3d741363212b8081b35f72b
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:ac128758328325ae56273800ea010f9559902281cc60226c8f99de51b636b180
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:26500e941ccdddc08dea07d3a4b9aa9eae7840fd5b683191320060f6dfc88c97
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:19ee4cb8e3e60d9c1e8fa63b0bf3e218bd683b6c05b9e9553d9477c68c493a9d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:7dbff355db7739152961cb6a97887d2e43bc960ac58837c126e03ed1a4480a3a
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:075b6bc4866b812c4b5394c8902b012978449a710ce71a30f55231fead941105
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:697baab21258166a1897ce4e056e39bc4667ff837bb97cadf92dad5863242ceb
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:ef03bb6ed647d92b9c84d34810376da7049aad88589eb2e35798d82a22f5b994
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:6df671937bc55c410a9df13aba15687b2832c7c19c0eff4686c3f83648528068
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4dd54195a3283a820d260eaeaf83463722e49bee05d875333c1b99201a7e2b85
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0a773245bec514cb893f730086f39a3789d0204b82cd11f5d11a5c270b92f70f
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:fb15cbade10fb37b14afe063a93decefd9a40583895b7b6b436f36f08bd73aee
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:a6e98064effd12ffda0613db98e795e056b83ef77b5deee932a762b178ceec3f
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:be5679938413dfb7766f765fba7923d3079d2d9b6f464485a8309e8e306f19e6
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:0cc0283b768c2968849f24cbf7178213926e084da5d0c90c18d38e2260e19138
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:82f60c881048618473c07970d6463b3186518bc18e4b263e16ca61d15448ab30
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:e76b61f9f0efe16facbddd80808d80bc32d0d41b4c16aed566116f1366152987
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:b368c22956707b8e594441c0e9bac20981339a5713d4fdfbe3c64bbfdf677b35
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:9fa6eea66a3daff39b7c0c7839ef5349b483b29dc219e1db384aef6c4c31ecb4
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:dc257df716b270858ca5d07909bda9ab635a388c9887d30b015391f60cc7b5c1
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:bb45d0217b4bc4352b209e3771756cb82923b8d260846c75b33ec270b4a8b8dd
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:b797e00eb32ff9a6a56e71d355bea653ecdecfecc121ddb197a6e347b89dbf8c
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:0e0ae96aff8540323da8a2379af014d3af1770e4b0043b84b9b415dda44501f7
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:37853d2979d0ca5f88fed76fed5194b628507ef0480e668dfd6eb034fa1d2023
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:c0459e5d3acb3b2a6d9863efe6ca1a24e529b38a6ddd3cc1e13d143924e872e7
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:582ced3f60212f736f720558524225b0d4a1c79ad4890656c47dd58009ffe0de
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:351358f551dfcf817c5936c0b43c2e28db61adcde65d7d2c81cae666630426b8
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:07c13b18197ed0bdd8b97d65312a342f95d7a7fc43d50af746d788032a67f83a
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:67515618d631b0ec00ad95ecb9bcb48d7f199855af9a64e86c4b46e7ae3be69e
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:932825fc7af8b96c381d504cd0976c751661ce078257430e5c79d565f416bacb
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:5ec262d1ada13a7d5d519ad98cf82dfbbd915664bb7560408558001ea6bb8451
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:3be3ee728cbae6bb8cf29d47c37ed1a4cd5791b54e1262135089424e5bda7ea1
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-usage@sha256:fe5bc4a40472926ec0df3023777b2aa8105507128bd9ac02eac2c68ad08723c1
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:f6e8de9f11696ca22302b8576ffd5bc2c024df0c8dc3734c404cc799565f351a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:36f4c57c6594f997383c4731e1e62c4765b8642c0b5c7373ba22118220928d09
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:2d8f924a7e4ef373df9760b4f1f3ad8050facec7de0c0efccd8c27d46b756e91
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:719b0871f0d32da894ec80dc2a23483623b7802a5c6f935cf3672468f99faa40
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:b372540b630507f4cb74f5eab38d0335d7b5bd64197e3f83852b2c7c5ba9067a
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:f9c9202e106d5a159c8f6b8fd883da4d6ac0d81e75b558fa52e0f7776269404c
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:095c969026e191bc4fceb020b5760e7595e439cca91b88f0d8c3e422591d56aa
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:8d9e35c3f29f8f56d55d81a2e6cc4d1b0cc352ddc8d33d83dced83c4a111a0cb
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:88d3c1ee02acbffedbf6ddcc27b80be6727f85425c361d45cc5e5c4056cf6733
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:1279b116fbcebb1c052eb27a72c67011b44a12fe673a86c1ece28202afd911d4
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:13a9ea6ee24d9b0ac1d5fb09c3e42eff731779867e7e71ce40429b43ba2779f9
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:5e5176376a664755fa776efe623a1e11bc5d3c31a3d0af980b165e54cf06620d
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:6fc976bc622c2165dd8bf93c9244e1c29b7c973a4da7a5ce4d0a097a1c1ea4fb
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:18f44ab15289625fa2c6b72ece6e06afb51058dcba30b8f38b4fdeedf8c9e51b
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:4fb2127ba7c525a959bbbcf024de364ef266ebb3ca7e696f74d990d2552cda59
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:a99b1f5a4eb0306b83562de0184fc826c34835ffbf46d929ceb9e70e0d325361
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:2011f32eb1e68ed5b75a4e8a99c992eac546995e219701d17a511596a5a09361
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:99d0b3192e6e46167423505f48655e7d49af1a461ae00b95c9c7c9d4209e5b65
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:65d327900e3b7a874dfe4bfbd0f2c7a1ee402346aba0e8cbc628b406e5b5329b
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:bb85f2567aaf7b612343fb9819fe05e2e771ccca8383316d34282ec37fcd1383
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:b2ce782efc5349a49d9c7326e2f72b2c1d99d30eccfb6975f7cba10008f9b2d3
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:267b7da5936493b0daf8ce824e3e6b50ef3b2bcb6e3126ed18fa9e09cdb0da43
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:27195e9758afdf8521971866fd60ea3dafed91a146c3c4165876e5ea476a4622
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-bundle@sha256:cd92fd1c85a72560241fffcb3bb62cf965f96bf2a4e1f5cca84c0b52624ae962
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:a34169228887ff93161584753c990bb82a5520ad7945dfa1ff4078e70aede592
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:7dc67256c002e86ee416c60ebb2942b35d8dfe0a5f96b4db495eff7361d21466
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:7de2c2f7cd478baba2011461c8d28c1a4c70aa485404d4523d1166dcb94c0a79
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:427e431d5b8b56011a91bc4f1e1586cd5490944615011fa6ac20abc8f45f92a7
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:96451d4a9d1c4b759d82dff2eaa9aba2a6fd58518f6ac6306cc082ebb7766a64
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:eb157bfe50005f9ffd3560ce1ef7ee0fc283ddfe265c42b35a71f18409e3f561
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e1b10435f94bbb7165703720f58aa8c68acbdaa18006a2c36d1e24ffdd72cfc3
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:411a4c7164b59ab3eed86e67aa457e45bd028067c937d35d0a227de63870cf87
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:3743af183b7b801f72b279a658b0d32da0bdbfa526a00d191a9a44de6df2284f
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:2472053d912f693a5fcdd00b7d9f0a7ce98697c47eae955e31c0b602fd8a4a37
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:b2ca6bc1119571cce732f1ce2ccb9ce406b298550b55144f950bb46cf985df3a
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:7cec55c00827b747f23c2f1703d99e1ca5b30a0c445f8ad0471b25b0c9cca57c
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-ui@sha256:b9a255ea269a11c85f52fe556291b0e869efbff9918b488b65e4297b958c6059
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a4daf28771178b0af391965684337805f8de0019a677753112961710f6011991
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:b7468b824d519e46e3ea23f5a803a8ace030b74d71d4cd282819027665ff0e12
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:3f2f97e6b2b18526255e88ff02d1039b24312eca51ee22b7cbf76d5680f7aee1
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:183ee860bf48076ea155db08032eed7004c1204a363625f54218734b00175441
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:62102f79bec5a8f66c600d184478f2b42c8acfde059fc71bc262b16dd46975cb
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:3129d9c7b096f9fce9cc12002eafb5c3d731a162ddef61578455dcf8acd31fac
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:a72c2b70a9f0a93047d47247cddc1616138cbc87337289ddb208565af6e4369d
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:27210840cbff8a7c61acac2ccc328a415a405b4738422b4f47eb9a0110127d22
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:381170c9c497b1bda33eef3ddeb1bafefa07be1e6ca75e75c9eaef9dc87b8d4a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:6bbcb64cf9adb4d3b6a4c259eb910e7da5c3c58732643e60d416399b2083cc0d
        - registry.marketplace.redhat.com/cpopen/ds-operator-bundle@sha256:e30b09797e341a5fdcc3ec377657e0e65be5855fb850dfc31f35b200f43130d8
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:ef24cc79265473a8b6fea231b1930c5763293dca79eedd46c89ec63dd95812c5
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:a31e0fb7a05e969e31c10a74cc858c483d5d4da14ce3abc36f5318100e60e792
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:aba8f1d08d143a8901b4d2cd8790e067fc7b9eca9987696a1ec6f9f505a686ba
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e12599ff99db058fdc8f9c1eebfba31160c636ff092138deb1f6f6705f44fd62
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bcf193d6f30ce1c7f9337ba8efa8d9b853eb7961e0008aa82635aa383fbe9d76
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:ab31869bdab5180b816a89c45008328cb7013c26b09c531bdbe5b00382f0dd7d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8ccecc00f3a9656a3dd1b07af76ef065e301e4101bda4ae8dc5bb91097cad7ad
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:c2b9d370b6c3e3c8e30079465502ce376055261cbf0f95c48d8a2ae10ca36c76
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:a6d5bd253959be1f91b50f5f2c56ed77e1c82b06483e62094aa86ff57cd46eec
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:78206c6634f2be2bd36806eb717bb8e4db2c755263cefa6d026a425a19329324
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:673320deb0ae96c6132545032f6b3e569d40204d3a5565a568dd97c5c1a4a552
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:3f6e9eeb659857799e30011f222924cef35301c59eaa5327ab5e8d212af75ce4
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:be55d9ad6ae76bd061046af448d032e65926abaeef95c831c978e01e2b0355c6
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:f6f3758287dd31889933ed0461fcf02dfa85d872c3d193669608573ca5269524
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:14da175e95c91864142ae00e9087ac2fe0e2083fc1a3bcb20a6b29dbab99d8a3
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:0cdba17b95a959b6d04889179d105515ed3176ea3449bc104f18b86374876048
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:b19bbc36033c3e60f387a9924711aab1ed81f150ca9d8f435e5433fed7f6a5c8
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:a3a0d58f895fca4d832e602d0dee80018df4ad7bad8800ad4058eea07fda4817
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:8d744af7dfeb9892378b7cc67d0e85e29d12b6f79c8b6b2a53f7a53aeffe8231
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:fa3d8a54b1ba33465171240d66f62271286cb120c51c54267d2b4f8b731c6177
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:7a90c8e0180da12172b23f3b54536bd1afe35affe37d908ddd78996a81435751
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:d2680c5f05c886b44df644aef6cb2fdc9534bb6a238066f6ebbb97bbeaac431e
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:6ccbb20c9fc7ce00ddfbac19c8808417957481473eeb1ee62ecda165e0e275cc
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:ca175109562cd9140d98bea3be88d6b34490ccb8d79f0cbbde08c821db9c5fe9
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:4f0165926685f49cde22a93387fea401bc1799ee846041fe54bca148dfc2f62f
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:70d23c45cc636af2148fadc1c7a08802692fed14e1a31893214b8d2d82f06e1e
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:9ad526bd56829ef24834b8a790b29c50809866c1dfb60cf8e0600fb9a903d388
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:0ab7d6651cdeda78fd0416137eff72eb58a621c8cd55bbbcb16f269808dec7cd
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:545b41cc6c044d6807bfeb7453a045e074a5670cfff9f5c311f4918946901fcb
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:17f7039e49f28abb7239e0c4f3689de58c75f34c9bde9e7b33a6ffb93ba92677
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:0fb78efb9e92dd0c2e52ca2488e6f1e7431e197f87e6994e15f7d621b90319f2
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:4deb399f28d3385c50a256c6e927cf27115c311be5943021733e4b281c12a419
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:5f086cdc7a272384c8a9b680f8f90fd471799e9e96175b40c7f16126b73169e7
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:786b21db2a9ce6cb0eabd5cb7c93dc41a8b1db6c1f2d07725a5bdd75b3a1b7d5
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:cdfb097f2438b5739bb88905b8318d15d15c4c64d37cfa6fdb06f44dfe976e6d
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-velero-plugin@sha256:2dd02d8f15dca1dce7cecad098cd8791cf1cbeddd2abd8e8c6d8f066d85f9e2e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:9c635194a9c63fed440e7c0ab0be1fe6cf59b1c3f954216dfb9150fe9e0fc207
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:044b663e4b7508c62dbb28b7fb2aa267c9771fc62ab509b08d46ea1499d95431
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:9c24d02b1cbb3857d017453744bbb2e68fbe133210c35e7da5ce0341921d5b08
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:cf2b88be76a4aad436f45a79694f32d2e2175587c3630543a6d24d7725b60c5a
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:9b91aaeacb196806646a1f196d4adc83b5159de767fad1a2eab699f735938911
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:d023f35b65816f244b6d699b41362e6031553f7d4c416efcee4a7f2395a61355
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:1019e13e6f072c07da6d31277c8ac639e8ff2200599ece1666a3e34745e763d4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ee161954766211651d7cb13f3f5a3c025ac04094ba742be453ce789c6f0e24a
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:4d24fb274b5429ba61fc4bc6833e93b14b92b8d525556db64e2172b45433de31
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:772f98763527c5452c662b2ae516de59c1f6edcf339da46231f4be8aab6c580d
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:bb6492c452fa6b481551628da26bb85bad151d12afd34a89c1a272a9587ecd1c
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:00cc7d072ea6b6f5e9f3e5b95b9cb4ab8ad7bd942d8c259e70f8ba0e9c09a464
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:9c160303552a9b8a97b951311018de2ada410883468680dd61fb41519b9a9e7a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:80c0913e2c84530d9b7c6a2152de76ded5754c134f6768754d00944f46d0ed4f
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:777e090e0394ffdf39abdde285b5b2749d6835e61458f117dbdac108186c34c8
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:d7ca81372914a28c8298de5f9d3c35cdbd6d8d8582ac9b8fda5a85d22e3c0def
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:07c5e4b3dc18b921339a7a5c78b8fe9179205678eb662377e8124e46479d7f25
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:2a50cd6ee3cb352ba186a57ba73fb9df321cf98f0b79b3d81024cbe57e0c7f35
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:866ebafb730e660062cf7401023d82c5242a2b80cca0126d47a68193c84670e0
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:c2aa3566f8fce992f00f43867173e5f47593e8618357369b23044fdf3c67ba53
        - registry.marketplace.redhat.com/cpopen/cpd/cpd-toolbox@sha256:1b3469f6fcd037aacabde3e05605b908e22f012d331d7365d0450f6d933d3efd
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-api@sha256:dd9e4909317512a6fada5ecfc7cdb47c6807f266fa456a01b60d7b3486faeeb7
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:fb850a3d660729fe8281967c2cb466fbad3f01fdc302db17988c27682217ad29
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:6caf5203489492755e5cf6bffcba53be0bd32662e8cd331dada68a85b2c083ca
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:7693312b2bcee7b8502de0da8786be985607fbfe6616bc9b0da4af8a6fdedea8
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:4ca30e2017a0ff2d64100a03a84bee8d601c765f0e624e1f485cbb97748bb8a2
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:c35eaf4921bfd046985ea7a03fc325f79766bf73e3ddba69408f4467fbd7994f
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:4184f25155643f01850f286f15cc92294a92d8fdc8e5020b1b522977ba429c2c
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:61155acdb8982648ae5748be9f5944ba913db1590ec3e773206d6fbb1d8952cb
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:474187fe2837da4b86ed9482a96fbca0a7f83db149859f8ef23db84de9a6d7fb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:55d4c6a13452f76bbab485363288efcace23975623b8a4b6b83f3c549f844c58
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator-bundle@sha256:c743dd7de2edbd08c6c73e20cf15f3b147491ad77a81677339e88caffb93e84b
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:b8886b9b54e6ee4c6b2981ef092e54348a0286577d9924cf136cee881f5bf1f4
        - registry.marketplace.redhat.com/cpopen/operand-deployment-lifecycle-manager-bundle@sha256:68f71f2767a40ca9333f0a670e07493e8a8c13a5c7ca877ea5bd8a544b352d9a
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:2754384c6936b36ffd64f89df751cbe47993f73dd700a184af35f412b3408d72
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:887348b5035f5a70f1c31aede992a5c76397f5eddd2aebeb27074c3ea6248c00
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:1b29e04308bb8e3c990a8c41a5d3d1e854e197bbdf0cffb2833fea4866772ba9
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90d45b7a3c5ffcd30a1ca1447693ba4aedfb400983cf49779a8421a42dc0eecd
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:60dbe69693314c937a69bd213de438c3af0a8a290aa9b1cfaad8a57e9d04c448
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:ab686e9e4c38b7909994f9fb496943b745e1a07492fe61a61b9e8a78f756eb1c
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:abec4867d6b9294d674cc02b9f7f254e65e7de8eea0231fa94f50513f2b7c45d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:b6500764d4f6d281f1b42458eb223d21f6a413b5db55f2a2ba1a7973aad8465c
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:3c18707afb90a7dc1bd2ed1fff05c53cce89b7c85062e72261f538a12834ab37
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:245f81b08980bfd4790cc3661e609e2cf53f3dd4d4c1f988e5db4dcf316d90fb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:d76476d2ecd88ea7b4fd857e591cc9b71be9d31fe0fb0057baa1b21e6113f730
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:186e849f3cf782f0d78660eb4433853e1e61296234d230a9dc5921fa4bd41060
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:6e055b6e07cb1adaa652b8e7515c4ce91438593a56ab0ae051435f55373d96ab
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:24dff64a80b910f0b4c5442d4f611f9df7602a16c5231eff2c9d08f07b2a9f9a
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:0b640bda3afa672011d5da390af798bbd3e6dd9126029a9eef03846d41582743
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:62dda55cc870279d491679f6b43be177986991b6eb85d9dfeba3ed2e72b3f10e
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:25c98c67d09cfaa3a4d332c461fb6766700a65876a48a1aabbd73b58bc2653f3
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:bd237f42078822b038c13271f35e6710273c2a4f3fe2f3eb74429c93cd2a0e41
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:bc39ff644108a458d1b7e29fcd88bf53982ab4070ce2ea0e75b0908678b9205d
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:21dffe4500172f74345fdafb9beb476e0ccffa55f7bca59f372f312f8d28d247
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:8f9d3438f58604cf29d643691b125a4138275b2c09cca4afc8e4b29289febd77
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:e8e0ea5cb1a070443011794988f2f6c9d67758605dc2847bf0183e19bfde52eb
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:9f835bb2984329db9db7eb2ac5521371ce0ced216a7875e45ef3ec31d255b648
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:234a8dcd8a9c5a68533bce6e0f793324e3ad7a660fdedf95895c5ed73a4cee3d
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:48828f128c98e73f8e345d87bc228f0920a3ff2cdc325d4a4ffe73ad95c5c5e0
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:ceee171c5ce0a8e580dfeccfe0dbfb73474c109becd45a562411678f3ab08595
        - registry.marketplace.redhat.com/cp/cpd/dap-base-job@sha256:f208686a3f56d2980e7b8640b6ebdea9057109e7520c6817a731e68d08506a36
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:31edab3e5662489dc2fe9528d078966663c8d49836425e648b7e112f871808dc
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:bcdf9b7eea8753b3e463973a2690140ae7aa8c02c3c7726511f116bd751ec2d8
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:63d4a9fa16231cc0c01f1b8bb9363221ac1e93f1571ea132866f170b156731e8
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:8fe85b93b9dcab9c33306ae358e6910685e109f32662d02e2b6bcd8dae243968
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:da0190fee910cb8245cf7a61d6e7aa2bf599d41c679eeedc183950918ea2611f
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:ad19b65042fb26cc1244a92f81a96e5ceb4313f8c823181efe28b66e9fecf9c8
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:78e33fff4bfc2e33ea179eaed9902df9bbd11463b9152d0281d1e0924fe09563
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-bundle@sha256:842f2ac568db4db79cf0450b413b4b09b5f2180ef3541385b46fb650ac7066c0
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:c43770715b76e0aa3fd9b3c2b21998ee369f2142e3c5d04320c677c9e9b29add
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:4eab846d6e80d68ce488e2cda36329cec1a6bf2f81feadf666b1005fed294e48
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:aa8d647f007d921db565a505d52583511e25de4a6a1187fc6b4df15427944364
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:276d5d06661ed401f0b0672dbcfa3d9f07fa2cde662fbb532fcdffbe3e332227
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:c6862158a2a5ae0a4e4206bfdc7f73e1c2359544966c72cf4fc3a617acb13697
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:4794a0fdae9d060def1e0d75a98560e713ca18de2625e4d9e5b5fa4c4877fac2
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9669ed5ba26ef32ad36a6784b2658553c00ac2a89cf4e9caddadf7bf46c1592e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:8305f26c2ebc493b54abfea0757d65b679f6c96bc3a0840c89f08a034327ed9c
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:0ae22914dbaf09d6679bd06bc9b6a5728e744c9ce2309cbd5f01e28308c403ea
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:209ea4b77b3b0863c591bcdec6f53c4c39dafcb10ff13e27ae4e1cb986a59727
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:483bc8f74d02efcbfb40ceb6abb0ab0b6581523cdea576b34512709b98879233
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:f40a9375fe5722972f995a82b50c00f254e3fe37f202b0aa1bebf66b0504099f
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:e8eee5083cbe317b368c002f47ce09f98e0c5e3407e5a4712e1c5473c4090ba3
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:523dcc671f1c05e690d5e578a90300d28d76570f605d4de7f80645b590f4a462
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:7ca8d085dd9aeb568eec2c4da974c03aa72bbfe4648f5d66a62d0444a2d1c839
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:9cd2ba7ab3ecab7094c892086a5276bbde22ff96d26ce0d087df2c1791235b02
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:ab99b62d40ba05cef8e699ebc7cc65271ea6fcb5900a76fe17526e7d55b50b08
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:a64a93104dee64e0e7e3e381144340511f903155f72b7f619c82c1d440712227
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:aa0068aec14e37986620c982295c083df6fc6bbe95facba582ae3c6f769a8e3a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:a10a15f07b856a7803ce0dcd30a542741dcb591804dc74c2a5fd50c55de0a983
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:10a935e41cc2dc09f8f3939cd512d9f6bc71d68457f77793a6957c5420cf4bf5
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:b87fa8374976ccd3dad90ddbf6b65cd6ae5a028ba896562f284f089380ebe44a
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:81d8b25221d827160dd5e88efd383453f532057ad02d7a2ee3dcd9c6eb16ba65
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:973eda4612c905661ff5ebc28c7cac6d9fe16fe5edaa544548d2311496ed938e
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:b82a10ce8f75d4441e869bf62fbfb5da349a2f228b52178bb29723af62f13938
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:5436a25b6c5b419b77643a76eb073bbbf8f09517299dbf4a992ae14c1965972c
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:40064d2893ce54b358c02023f4cc3603d3d28202ba56a177b20636abd74cb559
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:ae5e6858ed698d67b4b3917c0d41d0b4f7e2564585694f834099584c21a3c012
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:268c763fec7defe810624bbd31c96407fa37fef47b904758bce4c58d95aac043
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-bundle@sha256:94662b9dba5d4bafeec53b8ad334e7a1235385b21ac0952e12975f4f0a7623a3
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:c59837981b9a2a645c9ffd84c40d4f9ddc3797911f7a785ce4e3a2bfdd0bbe5c
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:bcd879a7cc3442ef62bb9677cc5534a2bfc9e7ea248e566aa4f1f6a0242431a8
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:b87b6c32a2050f1d5ff311d163a7b45efa7d5de1d352984b2f70fa578301a952
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:b9c8a65c1742c72c6351b1c8f4a858940c2bfa00ac5ec271c57b2ee7e5e7a66b
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:78682d9b9f19eae363cec0b6d31e696b2a5fd1bd1f8429df6f65a197310b6778
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:96ebe954a21f697ebaecfe5aa2c497c60b2ddec17e9ec4b376f435e6c53b3c29
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:21c81ff7c4b8809accaf62b242fd9b3d52f235489ae427556b7f71cc21bf8926
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-oauth2-proxy@sha256:bf4052b099393b563332acd5dcba6ba37ac424aa431d2d4f21532838b5ed6d6f
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:f5248b19c4a9b605a9160ecdb5dbbe0906841b667ede6fd6cf8e1d6c4369fc8e
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator-bundle@sha256:e4340bd9481da998a6b89e94c884c46e17cfb50d9d0c1099f39a3c54e500d2b2
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-oadp@sha256:9aec2bab8c7a6941e97c18352cb9e838a99aec08b06269e8053ac764638a90b9
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:0a32e127dd305fc445808d0469147db6c9704cd15d924d57d791e035475ecb60
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:4e3ec61a34fc7c71d213999b68b1f5c3f866f97e19c12fb001d506d4166a50f1
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:99d3b5a88d20380bcb4161e3b182f305af2e4cb405460d0f2a91965dd6f99f80
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:173ef1b032594c4e070e2e013db469115ec58d970fa8b62f3b748163b46d6d3e
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:a65ed5ac55cd5be99cc43c5a8d33bf127e8694917fce5f83fb8401c04e31b1f1
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:83ca1235dc054bf9ead64932c85333ea0dbade076b4a0cd1c3fe2b6bc2e05896
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:4ccd786884e39641a74096d783742c219fb277caafe80b667a80396c3a9cf6c3
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:6cffef7631ee54612fba71efc0eff4e17bc08defc3e2137aee9881d685444298
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:a3dc5fcff6a2aeee7c7e562fe9f6e0cf0652fd7ad4ec61496021d366206e2544
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:e296fa15dc14ee827d5046e8b1434d7f6838059c41d355471e1275e09f087411
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:68a98f52b7eba7e41c78aa91602c4e88d21d4a2a2a11b9b7592f181ce30e41ea
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:dd76a8484628970b62238334e122768aeea820ab18e24d62d23a6328c83b868c
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:e9e408a13bd103fb46536f896c51e7a27cba38d9b224231b20328a28efe1e616
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:9ced60f18740063ab281b836d4caf21035242ac3f659c6bb022ddb546770bff9
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:ebcfc8941330da6991480dc18ac34d6df4fe879e562ee537e5c36df73f4bc536
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:bdbbd76c93f0fb8f838edd9f124b02cc646a6f735a68a430a00f025a06c5f407
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:c0b6ad3b88f216379fc8f14aa5f075f5adc06ad06cc46495d3a40cc765f3deba
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:bb2f19e1797a900d60fb9f05d9f115faf20d514000aa3acac55fdba701581d81
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:93f73afa23f9fe8b73fd389081281fe1b0fd74bba6aa9bf7a6df663053afd3b8
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:54d8ceaaa06196e1ff524b7be595cbe90811f3be1a57c767b1ba74f53edb2f0e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:20f05c662659be9b2b54407c02f1805b93f09a3c204011981b7f17a76caad8ee
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:758f9e49dccaec5ebd125fc0c383fb5a656ba9f2f301984f993d50b103208d12
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:7655f531c354cfcce438c61cc5cde2c7690bfee4e2eadf9f99e4b157f163cd73
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:66c11636fdf68631e37bda7467536093b9402e883999e7989dbdcc4b4d2ba7d7
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:0621eb9b83c9b2b8bfe5cd09797fe08fa37a854a5506914935ff6f0df599b631
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:57e3433f860ec2b65f5c5f86574614fa9273c09b6b37db96a9e1888f4f6932f4
        - registry.marketplace.redhat.com/cp/cpd/ngp-projects-api@sha256:df420e9f616ad8a123d9eedd117fb6cb007b30a772138db534f47ad8801fa96d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:6d903fd4bd0ef3ef361d22b209ba7b686601b49d6f8d4e29b5d030da8d949dd5
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:d904761e70d3969a4ccacd71a60ad4d77dda939c556615df8d65782ccc3273aa
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:74aadd4fe99e66b89471222f853e442fa8338a23d9c24b4e908d38536c66456a
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:c91fc0992cff4b674f5e3690398e7dfb70662c807b930b9b0e65ba507ff925d2
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:24ca69c2f80a4369e29aa382a94d64fd3546b110787df1142c45a20b8246e844
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:9021a04805bb7f21d68fc6211f844a21c1d06f700c332ec3238073065b7c224b
        - registry.marketplace.redhat.com/cpopen/cpd/cpdtool@sha256:714caed1042ee03461efc9256a3da1046bd8f43819caf3a91ec38b118d459869
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:2346f67ffdc1a802720b325e010018cdb908aa458f372fe50ac6d8a92d8fb733
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:ef38f12712c406f8206742dffae9378e344d66c774e5b37b4fcb0b657db456ca
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:7505eb74c6bf165d40f6a8f3511f8e891cc34c0201fee38240c693a08e002ab1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8da8b240176c00f8f6eadf32e5614c663cc1d56d546c73996cca614b715acf4a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:fd35c33c7361bf04a66e69112b05a078ad65fdbe8f8ca8b7f4ecc2143b1643e2
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:67a02a9264d9c6e82edc15ae36d4f99b3fa9099697c550088d27c76511189f37
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:25a4bb9cea1c3bdbeeeb205622512bcba98163f9629eba35625756346ce9d652
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:860c17d2b2c6b0e013d0f4c035a388e45d8ef1483f53aa7ad022913b025f73b0
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:7d8fb7d8e2551bf949edd33589393fd12f634871b486e26464cd089e3e4b429d
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9186c350fa54e8b792d72e47be6b54a67b610f726175c9641795010d45ad51f9
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:f289ac0d44803ff12d41c3659e6fdf04c404e0b3b233b6bc44eec671abb1b47e
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:c8d2825596b7e675c247158bdd06911e1cc0ad9af3364c9c175a5b4cf16d5d8e
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:f42395ea365edc2c1bbef914a55b091e6442d17bd87bf3730a8f9f6d401bf4f5
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:8160c688f6b66206f8428f78dcf5e21b65c3a1cadc38b765f22ba8403a87f4d6
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:aea8fcb72664a41c3881839086a47fef1b54dd672300d4aaf583fd0b575a64b4
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:b341d9af40ad6aceb4ecc792ec75295313128a1eea3a0974aac879aed24a89ba
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:838f7649df5a7c13d6155394e211ac565aa86233bd30a3a2aab9706ec4a74eec
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:8542b9e449b054e455abf6e7884dcfda708a6213015e932844568d1853b65eb2
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:3af0967b0fb0a28a83a3e007b2c78fe27ad56a5e9c153806c6a7b17d7d9b1df5
        version: 6.1.3
      entries:
      - name: cpd-platform-operator.v6.1.3
        version: 6.1.3
      name: v6.1
    defaultChannel: v6.1
    packageName: cpd-platform-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Arkmq-org
      provider-url: https://arkmq-org.io
    name: activemq-artemis-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: activemq-artemis-operator.v2.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/arkmq-org/activemq-artemis-operator:2.0.5
          createdAt: "2025-04-04 14:35:00"
          description: An operator for managing the Apache ActiveMQ Artemis message
            broker
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/arkmq-org/activemq-artemis-operator
          support: arkmq-org
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: ArkMQ Operator manages ActiveMQ Artemis messaging broker in the
          cloud
        displayName: ArkMQ Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - messaging
        - cloud
        - kubernetes
        - activemq
        - artemis
        - integration
        - operator
        links:
        - name: Documentation
          url: https://arkmq-org.io/docs
        - name: Repository
          url: https://github.com/arkmq-org/activemq-artemis-operator
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Arkmq-org
          url: https://arkmq-org.io
        relatedImages:
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:4bba8828e72fd7c9d38c22ba1c2a4f7130817a4e7eac8fb63c13e86465a8e1e7
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:b71d1a57c8b3820a8133cf7a25a6edde7b29e1c4d32d7c30325183415a8bacd1
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:49f34122ab46ab5d7ead094b6e40817b62c75b6a809f67db07d68bce750da51e
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:a4a0d0453f0761b9e114d0d4cc7c14091a2a9c30790f18ffa6bf1e365d5083fe
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:93168fd8cd4fb32f041795afb8fb0e21cf81081d6cc463a44356e153dc1ad032
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:e2438670831bad39bb08147bbebe67ef9465899fb7750f1a5a185569f99d5fb5
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:8fbdd1ba728058d5f437186b9055ca28596140068ebc69ecff4fa52717b19a82
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:81590ec33b2f00775d3c6596f610887b670de7c0b89148b09eab7584f5b0b0ec
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:ddaf00914e3a412d21ec3a814e24f3d4cc382b5c4d6c21e9b1b94dde16ddf194
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:2895b2cf6292c951a0cdde78e755514f5d7cb43ce9f7b6e7dc381d8dfeec31ab
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.7
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.0
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:3bbfb6177e4839087f615a01ee54f9e06a95acaddbb083e11851b6af9315dbf0
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.0
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.0
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:42230ce9d11a1949a5bb73b6ee218b1581f585bcd553485ee438e1c5b50fcf8b
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.10
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.7
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:8f592b2454d092c82b5f61b4e04c3aff173db246539b3c06e15a71960d3bd235
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:f10983c68aff358e285d89d827c27b3487a3ee3b8e121cdaa295bf5fa8501145
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:b293c1269166cc6181f68b8d803fdfcf20fdbd5df2787ed0453724b22e298fb2
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:f25db980c3229e66cf48690aa175f6f93e3dce3974c0e70ab23af6462e7d1893
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:9bbb915855f29ea16950a3fd50d0b9bac25f82cba7d197579e04d9c33a5b02c4
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:de82344daa0caeabc6559656da0af285436a687e8f371744c0c51bdbc309167d
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:e438bcacf099a92b143793e88c00c6ac0afb31effe4b0eed49a935c9265ad943
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:2d155bffd246f26f516a6075ec170babc1373c72bb664cba4e3c58521ffd3558
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:264b33a9aad90d32a41258904f35092a141d76dcef0c115d414475444100aace
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:56bae938a6b4cd87f9f5284931f440f9b9d116cda3b63b13f83a4b1c6c641959
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:da95bff69deb63545d9782ed20cbdd17c1a79c6b16165580bf96a1d5191e6a59
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:f2c853ff91809282c04d5b1381ceae366518bf07b16649a61aa6c60c47f10714
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.2
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:b6f61188379def8c796c738549c099bd40bcd9bb44f5c30682c5edd136347aca
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:fc9b9a4c7550d84fd90335dd4f9e0d8d39c09136c6034b28490a368368cb11f1
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.2
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:9e1fefc9a94d72a7b5a3f4e2d53ebd23e5ab221df17a5afde42ce973277838f6
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.1
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.7
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:634f3d8fd620607b56e0360a191d23a9cb8a32ba0ba1f24fe20791ec52343586
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:4be60cfa2a300b2eef9b5f10ea756bd3f55d75fbd8890232d907f8d64080449b
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:37ab7354273f4dcd19bad6936f10e92220d98b8c0266002221477a0e2d146c81
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:719178251893def955e444e706ff07042ea98d294671315d95da504c1f47de24
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.4
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:70d4c799c7d905fac27457d63e662ea72bca9e899a692e425609e4eaa45129e6
        - quay.io/arkmq-org/activemq-artemis-operator:2.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:5e768aefade6d18cd65cb1eb5ee233844f9f7e7b52631f93a0982b67d382239d
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:ffa0d7cb4068df84319776dbfb66210297da1b8466ccd5451318fc3de79736c3
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:6f2da9ecd8f5021964d0dfa85144830efcfbd0462d18f18f036971e5f17b3bb6
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:decb3932ebb5b848b90a4a23b083f0cb5237eec2844095d8c58a6798da8bb999
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:88a0098a628fe43a9b422f8ebb774086fb65ed389a195e2779482a4a63447e28
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:a4628c2591ec06caf21a3467f6e3faeaede7fa402b4f137d943b25b4dd4ad343
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.2
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:52f337cd5d7d5d199eebf4fc722f8c5279f652506f009efbc0f8f6de88e18718
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:1fd99a46758cd29fce846d0d590021758b2a7a3efedb817bb205359caa7d4f67
        version: 2.0.5
      entries:
      - name: activemq-artemis-operator.v2.0.5
        version: 2.0.5
      name: stable
    defaultChannel: stable
    packageName: activemq-artemis-operator
    provider:
      name: Arkmq-org
      url: https://arkmq-org.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Syndesis Team
      provider-url: https://github.com/syndesisio/syndesis
    name: syndesis
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: syndesis.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta2",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.12
          createdAt: "2021-08-11T12:00:00Z"
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta2
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.12
        - centos/postgresql-10-centos7
        version: 1.12.0
      entries:
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: 1.12.x
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.13
        - centos/postgresql-10-centos7
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: 1.13.x
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - centos/postgresql-10-centos7
        - quay.io/syndesis/syndesis-operator:1.13
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: candidate
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.13
        - centos/postgresql-10-centos7
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: latest
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.13
        - centos/postgresql-10-centos7
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: stable
    defaultChannel: latest
    packageName: syndesis
    provider:
      name: Syndesis Team
      url: https://github.com/syndesisio/syndesis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.22.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.5"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.22
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.22
        version: 42.22.0
      entries:
      - name: t8c-operator.v42.22.0
        version: 42.22.0
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: stable
    - currentCSV: t8c-operator.v42.10.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.10
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.10
        version: 42.10.0
      entries:
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.10.0
    - currentCSV: t8c-operator.v42.10.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.10
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.10
        version: 42.10.0
      entries:
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.12.0
    - currentCSV: t8c-operator.v42.13.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.7"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.13
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.13
        version: 42.13.0
      entries:
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.13.0
    - currentCSV: t8c-operator.v42.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.7"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.14
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.14
        version: 42.14.0
      entries:
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.14.0
    - currentCSV: t8c-operator.v42.15.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.6.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.15
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.15
        version: 42.15.0
      entries:
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.15.0
    - currentCSV: t8c-operator.v42.16.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.16
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.16
        version: 42.16.0
      entries:
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.16.0
    - currentCSV: t8c-operator.v42.17.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.6.6"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.17
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.17
        version: 42.17.0
      entries:
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.17.0
    - currentCSV: t8c-operator.v42.18.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.1"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.18
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.18
        version: 42.18.0
      entries:
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.18.0
    - currentCSV: t8c-operator.v42.19.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.2"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.19
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.19
        version: 42.19.0
      entries:
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.19.0
    - currentCSV: t8c-operator.v42.20.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.20
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.20
        version: 42.20.0
      entries:
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.20.0
    - currentCSV: t8c-operator.v42.21.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.21
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.21
        version: 42.21.0
      entries:
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.21.0
    - currentCSV: t8c-operator.v42.22.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.5"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.22
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.22
        version: 42.22.0
      entries:
      - name: t8c-operator.v42.22.0
        version: 42.22.0
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.22.0
    - currentCSV: t8c-operator.v42.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.4.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.6
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.6
        version: 42.6.0
      entries:
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.6.0
    - currentCSV: t8c-operator.v42.7.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.4.6"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.7
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.7
        version: 42.7.0
      entries:
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.7.0
    - currentCSV: t8c-operator.v42.8.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.0"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.8
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.8
        version: 42.8.0
      entries:
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.8.0
    - currentCSV: t8c-operator.v42.9.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.1"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.9
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.9
        version: 42.9.0
      entries:
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.9.0
    defaultChannel: stable
    packageName: t8c
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-db
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-db.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Db",
                "metadata": {
                  "name": "db-sample"
                },
                "spec": {
                  "clusterDomain": "cluster.local",
                  "databases": [
                    {
                      "name": "wapdb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "grafanadb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "brokerdb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "nimbusmosaic",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    }
                  ],
                  "image": {
                    "migrationJobImage": "wap-ose-cli",
                    "migrationJobRepository": "prod.harbor.keysight.digital/loadcore",
                    "migrationJobTag": "v4.13.0-3747-35-ocp",
                    "pgImage": "postgresql-ubi-image",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "pgTag": "14.2.0-3747-32"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "postgresql": {
                    "image": {
                      "pullSecrets": [
                        "load-core-secret"
                      ],
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "postgresql-ubi-image",
                      "tag": "14.2.0-3747-32"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "pgHbaConfiguration": "local all all trust\nhost all all 0.0.0.0/0 trust",
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "storageClass": "managed-loadcore-storage"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "volumePermissions": {
                        "containerSecurityContext": {
                          "runAsUser": "auto"
                        }
                      }
                    },
                    "serviceAccount": {
                      "automountServiceAccountToken": false,
                      "name": "load-core-service-account"
                    }
                  },
                  "storage": "8Gi"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:30:46Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Db
            name: dbs.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP DB Operator
        displayName: Keysight WAP DB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Db
          url: https://keysight-wap-db.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-db@sha256:5689d6a4da2fdd97223bc64aa61a8574bba810f93d09ab804cb45b4475f80ff5
        version: 0.1.21
      entries:
      - name: keysight-wap-db.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-db
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-certified-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "tag": "5.4.6",
                  "oem": "",
                  "imagePullSecrets": "",
                  "psp": false,
                  "rbac": true,
                  "serviceAccount": "default",
                  "leastPrivilege": false,
                  "global": {
                    "cattle": {
                      "url": ""
                    }
                  },
                  "bootstrapPassword": "",
                  "autoGenerateCert": true,
                  "defaultValidityPeriod": 365,
                  "internal": {
                    "certmanager": {
                      "enabled": false,
                      "secretname": "neuvector-internal"
                    },
                    "autoGenerateCert": true,
                    "autoRotateCert": true
                  },
                  "controller": {
                    "enabled": true,
                    "annotations": {},
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": "registry.connect.redhat.com/neuvector/controller@sha256:65ee3723c7c007db5ce3cd0493a785d4163d972cf957de6dec956b842807a551",
                    "imagePullPolicy": "IfNotPresent",
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "searchRegistries": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "nodePort": "",
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      }
                    },
                    "ranchersso": {
                      "enabled": false
                    },
                    "pvc": {
                      "enabled": false,
                      "existingClaim": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {
                        "userinitcfg.yaml": {
                          "users": [
                            {
                              "Fullname": "admin",
                              "Password": "",
                              "Role": "admin"
                            }
                          ]
                        }
                      }
                    },
                    "certupgrader": {
                      "env": [],
                      "schedule": "",
                      "imagePullPolicy": "IfNotPresent",
                      "timeout": 3600,
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "prime": {
                      "enabled": false,
                      "image": {
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "1.0.7"
                      }
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/enforcer@sha256:72a628a5c811c95bb267bbd0a5ffc9d74dbaca03a92aae615d43c7601b4d3b43",
                    "imagePullPolicy": "IfNotPresent",
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "env": [],
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/etcd"
                      }
                    ],
                    "resources": {},
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "manager": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/manager@sha256:89a42c9ee86f98ef56b5e695e23bdcc6334361ebd29cfb7729d7c86f04da98b0",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "",
                    "env": {
                      "ssl": true,
                      "envs": []
                    },
                    "svc": {
                      "type": "ClusterIP",
                      "nodePort": "",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": "",
                      "tls": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "topologySpreadConstraints": [],
                    "affinity": {},
                    "podLabels": {},
                    "podAnnotations": {},
                    "tolerations": [],
                    "nodeSelector": {},
                    "runAsUser": "",
                    "probes": {
                      "enabled": false,
                      "timeout": 1,
                      "periodSeconds": 10,
                      "startupFailureThreshold": 30
                    }
                  },
                  "cve": {
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "cacert": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                      "image": "registry.access.redhat.com/ubi8",
                      "imagePullPolicy": "IfNotPresent",
                      "schedule": "0 0 * * *",
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": "registry.connect.redhat.com/neuvector/scanner@sha256:b632f1d31998a4c8bad9f97216e8aa56cd7e6446bd70d2089d6c66845b964e8d",
                      "imagePullPolicy": "Always",
                      "priorityClassName": "",
                      "resources": {},
                      "topologySpreadConstraints": [],
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      },
                      "volumes": "",
                      "volumeMounts": ""
                    }
                  },
                  "resources": {},
                  "runtimePath": "",
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhooksvc": {
                    "enabled": true
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  },
                  "lease": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "true"
          containerImage: registry.connect.redhat.com/neuvector/neuvector-operator@sha256:e2f19c50ebc4127a83c943dbcfa1312ec46dec71bd36f6d99178b320d8da970b
          createdAt: "2020-12-02 14:19:21"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: Neuvector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm/tree/master/charts/core) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart.  NeuVector Operator versions are tied to NeuVector product versions. Version 1.5.2 of the NeuVector Certified Operator deploys version 5.4.6 of NeuVector.

          **Complete below steps to Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Privileged SCC is added to Service Account specified in the deploment yaml by Operator version 1.3.6 and above in new deployment. In case of upgrading NeuVector Operator from previous version to 1.3.6, please delete Privileged SCC before upgrading.

                   oc delete rolebinding -nneuvector system:openshift:scc:privileged

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s                    neuvector/default


          #Deploying the NeuVector Operator#


          Please refer to the instructions [here](https://github.com/neuvector/neuvector-operator/blob/master/README.md)


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Networking
        - Security
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - registry.connect.redhat.com/neuvector/manager@sha256:89a42c9ee86f98ef56b5e695e23bdcc6334361ebd29cfb7729d7c86f04da98b0
        - registry.connect.redhat.com/neuvector/scanner@sha256:b632f1d31998a4c8bad9f97216e8aa56cd7e6446bd70d2089d6c66845b964e8d
        - registry.access.redhat.com/ubi8@sha256:534c2c0efa4150ede18e3f9d7480d3b9ec2a52e62bc91cd54e08ee7336819619
        - registry.connect.redhat.com/neuvector/neuvector-operator@sha256:e2f19c50ebc4127a83c943dbcfa1312ec46dec71bd36f6d99178b320d8da970b
        - registry.connect.redhat.com/neuvector/controller@sha256:65ee3723c7c007db5ce3cd0493a785d4163d972cf957de6dec956b842807a551
        - registry.connect.redhat.com/neuvector/enforcer@sha256:72a628a5c811c95bb267bbd0a5ffc9d74dbaca03a92aae615d43c7601b4d3b43
        version: 1.5.2
      entries:
      - name: neuvector-operator.v1.5.2
        version: 1.5.2
      - name: neuvector-operator.v1.5.1
        version: 1.5.1
      - name: neuvector-operator.v1.5.0
        version: 1.5.0
      - name: neuvector-operator.v1.4.9
        version: 1.4.9
      - name: neuvector-operator.v1.4.8
        version: 1.4.8
      - name: neuvector-operator.v1.4.7
        version: 1.4.7
      - name: neuvector-operator.v1.4.6
        version: 1.4.6
      - name: neuvector-operator.v1.4.5
        version: 1.4.5
      - name: neuvector-operator.v1.4.4
        version: 1.4.4
      - name: neuvector-operator.v1.4.3
        version: 1.4.3
      - name: neuvector-operator.v1.4.2
        version: 1.4.2
      - name: neuvector-operator.v1.4.1
        version: 1.4.1
      - name: neuvector-operator.v1.4.0
        version: 1.4.0
      - name: neuvector-operator.v1.3.9
        version: 1.3.9
      - name: neuvector-operator.v1.3.8
        version: 1.3.8
      - name: neuvector-operator.v1.3.7
        version: 1.3.7
      - name: neuvector-operator.v1.3.6
        version: 1.3.6
      - name: neuvector-operator.v1.3.5
        version: 1.3.5
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-certified-operator
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: model-builder-for-vision-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-modelbuilder-for-vision.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "Modelbuilder",
                "metadata": {
                  "name": "modelbuilder-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "backup_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "metadata_storage": {
                    "storage_class": "ibmc-block-bronze",
                    "storage_size": "20G"
                  },
                  "photo_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "500G"
                  },
                  "in_memory_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              },
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "ModelbuilderAWS",
                "metadata": {
                  "name": "modelbuilder-aws-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "postgres": {
                    "storage_size": "20G",
                    "storage_class": "gp2",
                    "backup_storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
          createdAt: "2021-08-28T12:00:00Z"
          olm.skipRange: <1.0.3
          operators.operatorframework.io/builder: operator-sdk-v1.11.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/ibm-modelbuilder-for-vision/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Model Builder for Vision on AWS
            displayName: Deploy-Modelbuilder on AWS
            kind: ModelbuilderAWS
            name: modelbuildersaws.modelbuilder.com
            version: v1alpha1
          - description: Deploy Model Builder for Vision on IBM
            displayName: Deploy-Modelbuilder on IBM
            kind: Modelbuilder
            name: modelbuilders.modelbuilder.com
            version: v1alpha1
        description: IBM Services Software Model Builder for Vision is an AI training
          platform that can use the capabilities of IBM Cloud and Amazon Web Services
          and their GPUs to quickly train computer vision models for compatible mobile
          apps. IBM Services Software Inspection Workbench iOS and iPadOS app will
          be used as the exclusive UI for labeling and training. Specifically, IBM
          Model Builder for Vision will train computer vision models that can then
          be deployed for use on the IBM Services Software Inspector Portable , IBM
          Services Software Inspector Wearable and the IBM Maximo Visual Inspection
          Mobile apps.
        displayName: IBM Model Builder for Vision
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Model Builder
        - AI
        - Computer Vision
        - Inspection
        - Mobile
        - Core ML
        - TensorFlow Lite
        links:
        - name: IBM Model Builder for Vision
          url: http://ibm.biz/model_builder
        maintainers:
        - email: isshelp@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:2f65f46b25d84a3cb41a6a10f93786bcab47b7354ca9ac13a8e602ea833eba60
        - registry.connect.redhat.com/ibm/modelbuilder-image-server@sha256:ecdb665af651417c50335e7259b89ef8096521abec431976bd7abe4c1edc49d6
        - registry.connect.redhat.com/ibm/modelbuilder-cv-training-service@sha256:c3c9080b08b87022c784fd98ac3e2774ec89ce29a2dd19d91dee039e3fdb2b36
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d4c60531f59953f399c6a2b0f8239128cdc2640f72f2d1e9ac4d6add618f9f01
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.connect.redhat.com/ibm/modelbuilder-augmentor@sha256:ef5d0dde3eb334d919c2773af63007f1ef5e735ae040ef77fffe3440f854f068
        - registry.connect.redhat.com/ibm/modelbuilder-auth-service@sha256:f96b444fdcf7781775c9eba942bcc5e5ebdcf5f50820447f5eb26061c4c674cd
        - registry.connect.redhat.com/ibm/modelbuilder-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
        version: 1.0.3
      entries:
      - name: ibm-modelbuilder-for-vision.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: model-builder-for-vision-certified
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cluster-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-manager.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "cluster-manager"
                },
                "spec": {
                  "addOnManagerImagePullSpec": "quay.io/open-cluster-management/addon-manager:v0.13.1",
                  "deployOption": {
                    "mode": "Default"
                  },
                  "placementImagePullSpec": "quay.io/open-cluster-management/placement:v0.13.1",
                  "registrationConfiguration": {
                    "featureGates": [
                      {
                        "feature": "DefaultClusterSet",
                        "mode": "Enable"
                      }
                    ]
                  },
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "cluster-manager"
                },
                "spec": {
                  "addOnManagerImagePullSpec": "quay.io/open-cluster-management/addon-manager:v0.13.1",
                  "deployOption": {
                    "hosted": {
                      "registrationWebhookConfiguration": {
                        "address": "management-control-plane",
                        "port": 30443
                      },
                      "workWebhookConfiguration": {
                        "address": "management-control-plane",
                        "port": 31443
                      }
                    },
                    "mode": "Hosted"
                  },
                  "placementImagePullSpec": "quay.io/open-cluster-management/placement:v0.13.1",
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/registration-operator:v0.13.1
          createdAt: "2024-04-10T15:27:49Z"
          description: Manages the installation and upgrade of the ClusterManager.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management-io/ocm
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterManager defines the configurations of controllers
              running hub that govern registration and work distribution for attached
              Klusterlets
            displayName: ClusterManager
            kind: ClusterManager
            name: clustermanagers.operator.open-cluster-management.io
            version: v1
        description: "Note: This operator will stop updating, please use the latest
          operator in OperatorHub.io https://operatorhub.io/operator/cluster-manager.
          \n\nThe Cluster Manager provides the multicluster hub, which can manage
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. This operator supports the installation and upgrade of ClusterManager.\n\n##
          How to Install\nInstall ClusterManager operator by following instructions
          in top right button `Install`. After installing the operator, create an
          instance of the ClusterManager resource to install the ClusterManager.\n\nThe
          pods of ClusterManager are created in the `open-cluster-management-hub`
          namespace:\n\n```\n$ kubectl get pods -n open-cluster-management-hub\nNAME
          \                                                      READY   STATUS    RESTARTS
          \  AGE\ncluster-manager-addon-manager-controller-7fcc67bfdf-x7hbz  1/1     Running
          \  0          61m\ncluster-manager-placement-controller-dbcfd4975-26jpw
          \      1/1     Running   0          61m\ncluster-manager-registration-controller-7949d97899-wwtg4
          \  1/1     Running   0          61m\ncluster-manager-registration-webhook-7f675bc4bb-v4rhj
          \     1/1     Running   0          61m\ncluster-manager-work-webhook-6c9d65fb95-jh97b
          \             1/1     Running   0          61m\n```\n\nThere is a ManagedCluster
          created in the Hub cluster when the managed cluster is registered to the
          Hub cluster.\n\n```\n$ kubectl get managedclusters.cluster.open-cluster-management.io\nNAME
          \                  HUB ACCEPTED   MANAGED CLUSTER URLS     JOINED   AVAILABLE
          \  AGE\ncluster-dev            true                                    True
          \    True        7h20m\ncluster-test           false          https://127.0.0.1:8443
          \                       7h25m\n```\n"
        displayName: Cluster Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open-cluster-management
        - ClusterManager
        links:
        - name: Registration Operator
          url: https://github.com/open-cluster-management-io/ocm
        maintainers:
        - email: zyin@redhat.com
          name: Zhiwei Yin
        - email: deads@redhat.com
          name: David Eads
        - email: jqiu@redhat.com
          name: Jian Qiu
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/open-cluster-management/registration-operator:v0.13.1
        version: 0.13.1
      entries:
      - name: cluster-manager.v0.13.1
        version: 0.13.1
      - name: cluster-manager.v0.13.0
        version: 0.13.0
      - name: cluster-manager.v0.12.0
        version: 0.12.0
      - name: cluster-manager.v0.11.0
        version: 0.11.0
      - name: cluster-manager.v0.10.0
        version: 0.10.0
      - name: cluster-manager.v0.9.1
        version: 0.9.1
      - name: cluster-manager.v0.8.0
        version: 0.8.0
      - name: cluster-manager.v0.7.0
        version: 0.7.0
      - name: cluster-manager.v0.6.0
        version: 0.6.0
      - name: cluster-manager.v0.5.0
        version: 0.5.0
      - name: cluster-manager.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: cluster-manager
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: ScaleOps
      provider-url: https://scaleops.com
    name: scaleops-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: scaleops-operator.v1.7.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1alpha1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops-sample"
                },
                "spec": {
                  "scaleopsToken": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: ghcr.io/scaleops-sh/operator-bundle@sha256:ee627042deefa4faceb63efd56aa6f6338c90f6c13877f29d7161efe603e02f4
          createdAt: "2024-09-19T07:04:10Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: ScaleOps
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1alpha1
        description: ScaleOps Cluster Optimization
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: vlad@scaleops.com
          name: Vladislav Shub
        - email: guy@scaleops.com
          name: Guy Baron
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - ghcr.io/scaleops-sh/operator@sha256:7de0651a76d6eb63fee81e9177df93f126906d50063d7891f12c4a72c817dc86
        - ghcr.io/scaleops-sh/kube-state-metrics/ubi@sha256:d8cb65795861ce257f08726b7a3aa60a869cc44ed152f559431191bc79e98908
        - ghcr.io/scaleops-sh/operator-bundle@sha256:ee627042deefa4faceb63efd56aa6f6338c90f6c13877f29d7161efe603e02f4
        - ghcr.io/scaleops-sh/scaleops-operator-bundle@sha256:128771d92c852f7ace78103493db079a474e06d1352f06cbde7a7d4666684610
        - ghcr.io/scaleops-sh/configmap-reload/ubi@sha256:fed19f1aad43228560912abc36372a6c124d9ddf5e3e39476269dc51604d26de
        - ghcr.io/scaleops-sh/prometheus/ubi@sha256:8aeeae2a33251493d4d6e2f02a2fff1d4c2f32ab54554e6a84619ecd8b4aa69a
        version: 0.0.1
      entries:
      - name: scaleops-operator.v1.7.5
        version: 0.0.1
      name: alpha
    - currentCSV: scaleops-operator.v1.21.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops",
                  "namespace": "scaleops-system"
                },
                "spec": {
                  "scaleopsToken": "${SCALEOPS_TOKEN}"
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: registry.connect.redhat.com/scaleops-sh/operator@sha256:528f2e09d2dad4f48e01a272b51baa1753c12069515831223e3f97c7dfe03e3c
          createdAt: "2025-09-29T13:35:24Z"
          description: |-
            ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

            ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

            Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

            Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

            Contact [us](mailto:team@scaleops.com) to start your free trial today
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.21.8
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://scaleops.com
          support: ScaleOps Support support@scaleops.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This will install an instance of the ScaleOps Platform on
              your cluster
            displayName: ScaleOps Platform
            kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1
        description: |-
          ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

          ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

          Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

          Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

          Contact [us](mailto:team@scaleops.com) to start your free trial today
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        - application runtime
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: support@scaleops.com
          name: ScaleOps
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - registry.connect.redhat.com/scaleops-sh/operator@sha256:528f2e09d2dad4f48e01a272b51baa1753c12069515831223e3f97c7dfe03e3c
        - registry.connect.redhat.com/scaleops-sh/scaleops-sh@sha256:57d5d06f4f4b5470562c4f24ea5d37203633e1f4b65cd497e71d9066fd2b9e41
        - registry.connect.redhat.com/scaleops-sh/scaleops-gpu@sha256:724aef7683ea4cb245ab0a464517a6d542efa4634e5469ff35b7a7b7f00b563e
        version: 1.21.8
      entries:
      - name: scaleops-operator.v1.21.8
        version: 1.21.8
      - name: scaleops-operator.v1.21.7
        version: 1.21.7
      - name: scaleops-operator.v1.21.6
        version: 1.21.6
      - name: scaleops-operator.v1.21.5
        version: 1.21.5
      - name: scaleops-operator.v1.21.3
        version: 1.21.3
      - name: scaleops-operator.v1.21.2
        version: 1.21.2
      - name: scaleops-operator.v1.21.1
        version: 1.21.1
      - name: scaleops-operator.v1.21.0
        version: 1.21.0
      - name: scaleops-operator.v1.20.9
        version: 1.20.9
      - name: scaleops-operator.v1.20.8
        version: 1.20.8
      - name: scaleops-operator.v1.20.7
        version: 1.20.7
      - name: scaleops-operator.v1.20.6
        version: 1.20.6
      - name: scaleops-operator.v1.20.5
        version: 1.20.5
      - name: scaleops-operator.v1.20.4
        version: 1.20.4
      - name: scaleops-operator.v1.20.2
        version: 1.20.2
      - name: scaleops-operator.v1.20.1
        version: 1.20.1
      - name: scaleops-operator.v1.19.8
        version: 1.19.8
      - name: scaleops-operator.v1.19.7
        version: 1.19.7
      - name: scaleops-operator.v1.19.6
        version: 1.19.6
      - name: scaleops-operator.v1.19.5
        version: 1.19.5
      - name: scaleops-operator.v1.19.4
        version: 1.19.4
      - name: scaleops-operator.v1.19.1
        version: 1.19.1
      - name: scaleops-operator.v1.19.0
        version: 1.19.0
      - name: scaleops-operator.v1.18.8
        version: 1.18.8
      - name: scaleops-operator.v1.18.7
        version: 1.18.7
      - name: scaleops-operator.v1.18.6
        version: 1.18.6
      - name: scaleops-operator.v1.18.5
        version: 1.18.5
      - name: scaleops-operator.v1.18.4
        version: 1.18.4
      - name: scaleops-operator.v1.18.3
        version: 1.18.3
      - name: scaleops-operator.v1.18.2
        version: 1.18.2
      name: stable
    - currentCSV: scaleops-operator.v1.10.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops",
                  "namespace": "scaleops-system"
                },
                "spec": {
                  "scaleopsToken": "${SCALEOPS_TOKEN}"
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: registry.connect.redhat.com/scaleops-sh/operator@sha256:8def662a546babd6740e5417401836d647d92d52b15d559b1b3e2cebd7f77f33
          createdAt: "2025-02-12T19:06:04Z"
          description: |-
            ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

            ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

            Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

            Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

            Contact [us](mailto:team@scaleops.com) to start your free trial today
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.10.32
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://scaleops.com
          support: ScaleOps Support support@scaleops.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This will install an instance of the ScaleOps Platform on
              your cluster
            displayName: ScaleOps Platform
            kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1
        description: |-
          ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

          ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

          Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

          Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

          Contact [us](mailto:team@scaleops.com) to start your free trial today
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        - application runtime
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: support@scaleops.com
          name: ScaleOps
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - registry.connect.redhat.com/scaleops-sh/operator@sha256:8def662a546babd6740e5417401836d647d92d52b15d559b1b3e2cebd7f77f33
        - registry.connect.redhat.com/scaleops-sh/scaleops-sh@sha256:5d77e81d7959483121c208072eb61b063eb3ad9abc49e7df3c7efb8e2b77a3ae
        version: 1.10.32
      entries:
      - name: scaleops-operator.v1.10.32
        version: 1.10.32
      name: stable-1.10
    - currentCSV: scaleops-operator.v1.12.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops",
                  "namespace": "scaleops-system"
                },
                "spec": {
                  "scaleopsToken": "${SCALEOPS_TOKEN}"
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: registry.connect.redhat.com/scaleops-sh/operator@sha256:4bd74a56a8680f47aa1da7a377f74529fe7ca1568995a4e74018af0cf67763ed
          createdAt: "2025-03-11T18:10:07Z"
          description: |-
            ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

            ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

            Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

            Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

            Contact [us](mailto:team@scaleops.com) to start your free trial today
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.12.6
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://scaleops.com
          support: ScaleOps Support support@scaleops.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This will install an instance of the ScaleOps Platform on
              your cluster
            displayName: ScaleOps Platform
            kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1
        description: |-
          ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

          ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

          Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

          Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

          Contact [us](mailto:team@scaleops.com) to start your free trial today
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        - application runtime
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: support@scaleops.com
          name: ScaleOps
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - registry.connect.redhat.com/scaleops-sh/operator@sha256:4bd74a56a8680f47aa1da7a377f74529fe7ca1568995a4e74018af0cf67763ed
        - registry.connect.redhat.com/scaleops-sh/scaleops-sh@sha256:5ea4e26f2fecee3570d63adc3e2588d34f9610345d703975d4febac9c94b4bd0
        version: 1.12.6
      entries:
      - name: scaleops-operator.v1.12.6
        version: 1.12.6
      name: stable-1.12
    - currentCSV: scaleops-operator.v1.14.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops",
                  "namespace": "scaleops-system"
                },
                "spec": {
                  "scaleopsToken": "${SCALEOPS_TOKEN}"
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: registry.connect.redhat.com/scaleops-sh/operator@sha256:9267f49ce774ee4d401458531599bbd6729c6313cecc4f2bcd3bf0b5eecd92e1
          createdAt: "2025-06-08T16:15:46Z"
          description: |-
            ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

            ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

            Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

            Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

            Contact [us](mailto:team@scaleops.com) to start your free trial today
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.14.14
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://scaleops.com
          support: ScaleOps Support support@scaleops.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This will install an instance of the ScaleOps Platform on
              your cluster
            displayName: ScaleOps Platform
            kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1
        description: |-
          ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

          ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

          Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

          Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

          Contact [us](mailto:team@scaleops.com) to start your free trial today
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        - application runtime
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: support@scaleops.com
          name: ScaleOps
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - registry.connect.redhat.com/scaleops-sh/operator@sha256:9267f49ce774ee4d401458531599bbd6729c6313cecc4f2bcd3bf0b5eecd92e1
        - registry.connect.redhat.com/scaleops-sh/scaleops-sh@sha256:9c6395202ccb70e8edae3a1b540de31991ec9232b7b235dd92755588b8491720
        version: 1.14.14
      entries:
      - name: scaleops-operator.v1.14.14
        version: 1.14.14
      name: stable-1.14
    defaultChannel: stable
    packageName: scaleops-operator
    provider:
      name: ScaleOps
      url: https://scaleops.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataStax
      provider-url: ""
    name: cass-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cass-operator.v1.23.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cassandra.datastax.com/v1beta1",
                "kind": "CassandraDatacenter",
                "metadata": {
                  "name": "dc1"
                },
                "spec": {
                  "clusterName": "development",
                  "config": {
                    "cassandra-yaml": {
                      "authenticator": "PasswordAuthenticator",
                      "authorizer": "CassandraAuthorizer",
                      "num_tokens": 16,
                      "role_manager": "CassandraRoleManager"
                    },
                    "jvm-server-options": {
                      "initial_heap_size": "1G",
                      "max_heap_size": "1G"
                    }
                  },
                  "managementApiAuth": {
                    "insecure": {}
                  },
                  "podTemplateSpec": {
                    "spec": {
                      "containers": [
                        {
                          "name": "cassandra"
                        }
                      ],
                      "securityContext": {}
                    }
                  },
                  "racks": [
                    {
                      "name": "rack1"
                    },
                    {
                      "name": "rack2"
                    },
                    {
                      "name": "rack3"
                    }
                  ],
                  "resources": {
                    "requests": {
                      "cpu": "1000m",
                      "memory": "2Gi"
                    }
                  },
                  "serverType": "cassandra",
                  "serverVersion": "4.1.2",
                  "size": 3,
                  "storageConfig": {
                    "cassandraDataVolumeClaimSpec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "server-storage"
                    }
                  }
                }
              },
              {
                "apiVersion": "control.k8ssandra.io/v1alpha1",
                "kind": "CassandraTask",
                "metadata": {
                  "name": "example-cleanup"
                },
                "spec": {
                  "datacenter": {
                    "name": "dc2",
                    "namespace": "cass-operator"
                  },
                  "jobs": [
                    {
                      "command": "cleanup",
                      "name": "cleanup-run"
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database,Big Data,OpenShift Optional
          certified: "true"
          containerImage: registry.connect.redhat.com/datastax/cass-operator@sha256:6abc6253b50a0d7f6bdd4b394d2bf96320a8845952aeb8f0bbed99863192ec52
          createdAt: "2025-03-20"
          description: |
            Simple provisioning, turn-key operations, and automated remediation of Apache Cassandra clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/k8ssandra/cass-operator
          support: DataStax
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CassandraDatacenter is the Schema for the cassandradatacenters
              API
            displayName: Cassandra Datacenter
            kind: CassandraDatacenter
            name: cassandradatacenters.cassandra.datastax.com
            version: v1beta1
          - description: CassandraTask is the Schema for the cassandrajobs API
            displayName: Cassandra Task
            kind: CassandraTask
            name: cassandratasks.control.k8ssandra.io
            version: v1alpha1
        description: "## Apache Cassandra\n\nThe Apache Cassandra database is the
          right choice when you need scalability and\nhigh availability without compromising
          performance. Linear scalability and\nproven fault-tolerance on commodity
          hardware or cloud infrastructure make it the\nperfect platform for mission-critical
          data. Cassandra's support for replicating\nacross multiple datacenters is
          best-in-class, providing lower latency for your\nusers and the peace of
          mind of knowing that you can survive regional outages.\n\n## DataStax Enterprise\n\nThe
          most advanced distribution of Apache Cassandra™ on the market, with the\nenterprise
          functionality needed for serious production systems and backed up and\nsupported
          by the best distributed-experts in the world. It's one platform for\nall
          types of applications anywhere, any cloud, any model: key-value, graph,\ntabular,
          JSON.\n\nDataStax Enterprise is a fully integrated and optimized database,
          with graph,\nanalytics, and search included, all with a unified security
          model. Simply put,\nit's the only database capable of meeting today's demanding
          requirements\n\n## Operator Details\n\n`cass-operator` is designed as a
          modular operator for Apache Cassandra and\nderived  distributions. Apache
          Cassandra is a distributed database consisting of\nmultiple nodes working
          in concert to store data and process queries along a\nnumber of fault domains.
          `cass-operator` has the deployment of a Cassandra\ncluster around the logical
          domain of a datacenter with the `CassandraDatacenter`\ncustom resource.\n
          \   \nUpon submission of one of these resources it handles provisioning
          the underlying\nstateful sets (analogous to C\\* logical racks), services,
          and configuration.\nAdditionally through monitoring pod state via Kubernetes
          callbacks it handles day to day\noperations such as restarting failed processes,
          scaling clusters up, and deploying\nconfiguration changes in a rolling,
          non-disruptive, fashion.\n    \nThis operator is designed to be `Namespace`
          scoped. A single Kubernetes cluster may\nbe running multiple instances of
          this operator, in separate namespaces, to support\na number of C\\* clusters
          and environments. Configuration is simple with the usage of\nYAML based
          overrides in the Custom Resource paired with an `init` container.\n    \nIn
          C\\* clusters ordering and timing of certain operations are important to
          keep the system\nevenly distributed. `cass-operator` takes advantage of
          a sidecar process within the\nmain container to handle the orchestration
          of starting our main C* process."
        displayName: DataStax Kubernetes Operator for Apache Cassandra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - k8ssandra
        - cass-operator
        - DataStax Enterprise
        - Apache Cassandra
        - DSE
        - C*
        - Cassandra
        links:
        - name: DataStax
          url: https://datastax.com
        - name: Operator GitHub Repo
          url: https://github.com/k8ssandra/cass-operator
        - name: Config Builder GitHub Repo
          url: https://github.com/datastax/cass-config-builder
        - name: Documentation [1]
          url: https://github.com/k8ssandra/cass-operator/tree/master/docs/user
        - name: Documentation [2]
          url: https://docs.datastax.com/en/cass-operator/doc/cass-operator/cassOperatorTOC.html
        maintainers:
        - email: cass-operator@datastax.com
          name: DataStax Operator Team
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: DataStax
        relatedImages:
        - registry.connect.redhat.com/datastax/cass-operator@sha256:6abc6253b50a0d7f6bdd4b394d2bf96320a8845952aeb8f0bbed99863192ec52
        - registry.connect.redhat.com/datastax/system-logger@sha256:3408620786e6265aa835be56f85aadc48b20060b2038dc21126f614685f77c6b
        version: 1.23.2
      entries:
      - name: cass-operator.v1.23.2
        version: 1.23.2
      name: stable
    defaultChannel: stable
    packageName: cass-operator
    provider:
      name: DataStax
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
    name: aikit-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aikit-operator.v2021.2.102120
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aikit.intel/v1alpha1",
                "kind": "AIKitContainer",
                "metadata": {
                  "name": "intel-aikit-container",
                  "namespace": "redhat-ods-applications"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagestream": {
                    "name": "oneapi-aikit",
                    "namespace": "redhat-ods-applications",
                    "registry": {
                      "root": "registry.connect.redhat.com",
                      "repo": "intel",
                      "name": "oneapi-aikit",
                      "version": "latest"
                    }
                  },
                  "nameOverride": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aikit-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aikit-operator-rhmp/support?utm_source=openshift_console
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AIKitContainer
            name: aikitcontainers.aikit.intel
            version: v1alpha1
        description: |
          The Intel® oneAPI AI Analytics Toolkit gives data scientists, AI developers, and researchers familiar Python* tools and frameworks to accelerate end-to-end data science and machine learning pipelines on Intel® architectures. The components are built using oneAPI libraries for low-level compute optimizations. This toolkit maximizes performance from preprocessing through machine and deep learning phases and provides interoperability for efficient model development and deployment across single and multinodes.
        displayName: Intel® oneAPI AI Analytics Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aikit-operator
        links:
        - name: Aikit Operator
          url: https://aikit-operator.domain
        maintainers:
        - email: kam.d.kasravi@intel.com
          name: Kam Kasravi
        - email: abolfazl.shahbazi@intel.com
          name: Abolfazl Shahbazi
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Intel
          url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
        relatedImages:
        - registry.marketplace.redhat.com/rhm/intel/aikit-operator@sha256:7abb5e6164ffb3e4b87c39061f45d09848218fcfb74b48bf5e1035d56b41ee0c
        version: 2021.2.102120
      entries:
      - name: aikit-operator.v2021.2.102120
        version: 2021.2.102120
      - name: aikit-operator.v2021.2.100410
        version: 2021.2.100410
      - name: aikit-operator.v2021.2.92911
        version: 2021.2.92911
      - name: aikit-operator.v2021.2.92814
        version: 2021.2.92814
      - name: aikit-operator.v2021.2.92808-dev-a60e5ba
        version: 2021.2.92808-dev-a60e5ba
      name: alpha
    defaultChannel: alpha
    packageName: aikit-operator-rhmp
    provider:
      name: Intel
      url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Apache Software Foundation
      provider-url: ""
    name: camel-karavan-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: camel-karavan-operator.v3.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Karavan",
                "metadata": {
                  "labels": {
                    "app": "karavan"
                  },
                  "name": "karavan"
                },
                "spec": {
                  "instances": 1,
                  "auth": "public",
                  "environment": "dev",
                  "runtimes": "quarkus,spring-boot",
                  "gitPullInterval": "10s"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/apache/camel-karavan-operator:3.20.1
          repository: https://github.com/apache/camel-karavan
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Karavan
            name: karavans.camel.apache.org
            version: v1
        description: |
          Apache Camel Karavan
          ====================
          Karavan is an Integration Toolkit for Apache Camel, which makes integration easy and fun through the visualization of pipelines,
          integration with runtimes and package, image build and deploy to kubernetes out-of-the-box.

          ## Installation

          1. Install Apache Camel Karavan Operator
          2. Install Tekton operator
          3. Create namespace, ex: `karavan`
          4. Create `Secret` in the namespace, ex:
          ```
          kind: Secret
          apiVersion: v1
          metadata:
            name: karavan
            namespace: karavan
          type: Opaque
          stringData:
            master-password: karavan
            oidc-secret: secret
            oidc-server-url: https://hostname/auth/realms/karavan
            oidc-frontend-url: https://hostname/auth
            git-repository: https://github.com/repository/projects.git
            git-password: password
            git-username: username
            git-branch: main
            image-registry: image-registry.openshift-image-registry.svc:5000
          ```
          5. Create `Karavan` instance in the namespace, ex:
          ```
          apiVersion: camel.apache.org/v1
          kind: Karavan
          metadata:
            name: karavan
            namespace: karavan
          spec:
            instances: 1
            auth: public
            environment: dev
            runtimes: quarkus,spring-boot
            gitPullInterval: 10s
          ```
        displayName: Camel Karavan Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - camel
        - karavan
        - integration
        - microservices
        - low-code
        links:
        - name: Apache Camel Karavan source code repository
          url: https://github.com/apache/camel-karavan
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - ghcr.io/apache/camel-karavan-operator:3.20.1
        version: 3.20.1
      entries:
      - name: camel-karavan-operator.v3.20.1
        version: 3.20.1
      - name: camel-karavan-operator.v3.20.0
        version: 3.20.0
      - name: camel-karavan-operator.v3.18.5
        version: 3.18.5
      name: alpha
    defaultChannel: alpha
    packageName: camel-karavan-operator
    provider:
      name: The Apache Software Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: cloudnative-pg
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudnative-pg.v1.27.0-rc1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: ghcr.io/cloudnative-pg/cloudnative-pg@sha256:c2212e806e950ed5b667455f476481aef179c8e467e435352b624dc670000443
          createdAt: "2025-07-31T12:14:29Z"
          description: CloudNativePG is an open source operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/cloudnative-pg/cloudnative-pg
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.cnpg.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.cnpg.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.cnpg.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.cnpg.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.cnpg.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.cnpg.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.cnpg.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnpg` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: CloudNativePG
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnpg
        links:
        - name: CloudNativePG
          url: https://cloudnative-pg.io/
        - name: Documentation
          url: https://cloudnative-pg.io/docs/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: The CloudNativePG Contributors
        relatedImages:
        - ghcr.io/cloudnative-pg/cloudnative-pg@sha256:c2212e806e950ed5b667455f476481aef179c8e467e435352b624dc670000443
        - ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b
        - ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891
        version: 1.27.0-rc1
      entries:
      - name: cloudnative-pg.v1.27.0-rc1
        version: 1.27.0-rc1
      - name: cloudnative-pg.v1.26.0-rc3
        version: 1.26.0-rc3
      - name: cloudnative-pg.v1.26.0-rc2
        version: 1.26.0-rc2
      - name: cloudnative-pg.v1.26.0-rc1
        version: 1.26.0-rc1
      name: candidate
    - currentCSV: cloudnative-pg.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: ghcr.io/cloudnative-pg/cloudnative-pg@sha256:53995e9d6a7760c77d77e5298f78b49029427494f4dd933d5fa7a74f7af3bb44
          createdAt: "2025-08-12T14:08:35Z"
          description: CloudNativePG is an open source operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/cloudnative-pg/cloudnative-pg
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.cnpg.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.cnpg.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.cnpg.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.cnpg.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.cnpg.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.cnpg.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.cnpg.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnpg` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: CloudNativePG
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnpg
        links:
        - name: CloudNativePG
          url: https://cloudnative-pg.io/
        - name: Documentation
          url: https://cloudnative-pg.io/docs/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - ghcr.io/cloudnative-pg/cloudnative-pg@sha256:53995e9d6a7760c77d77e5298f78b49029427494f4dd933d5fa7a74f7af3bb44
        - ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b
        - ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891
        version: 1.27.0
      entries:
      - name: cloudnative-pg.v1.27.0
        version: 1.27.0
      - name: cloudnative-pg.v1.26.1
        version: 1.26.1
      - name: cloudnative-pg.v1.26.0
        version: 1.26.0
      - name: cloudnative-pg.v1.25.1
        version: 1.25.1
      - name: cloudnative-pg.v1.25.0
        version: 1.25.0
      - name: cloudnative-pg.v1.24.1
        version: 1.24.1
      - name: cloudnative-pg.v1.24.0
        version: 1.24.0
      - name: cloudnative-pg.v1.23.3
        version: 1.23.3
      - name: cloudnative-pg.v1.23.2
        version: 1.23.2
      - name: cloudnative-pg.v1.23.1
        version: 1.23.1
      name: stable-v1
    defaultChannel: stable-v1
    packageName: cloudnative-pg
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cincinnati-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: update-service-operator.v5.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "updateservice.operator.openshift.io/v1",
                "kind": "UpdateService",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          description: Creates and maintains an OpenShift Update Service instance
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-update-service
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: UpdateService is the Schema for the updateservices API.
            displayName: Update Service
            kind: UpdateService
            name: updateservices.updateservice.operator.openshift.io
            version: v1
        description: |-
          # Use Case
          Running an Update Service instance in a cluster is appealing for offline OpenShift
          clusters or for admins that want to provide their own graph data instead of
          using [https://github.com/openshift/cincinnati-graph-data](https://github.com/openshift/cincinnati-graph-data/).

          # About Update Service
          Update Service uses **Cincinnati** which is an update protocol designed to facilitate
          automatic updates. It describes a particular method for representing transitions between
          releases of a project and allowing a client to perform automatic updates between these
          releases.

          ## Components
          A **Graph Builder** iterates over the release payloads hosted by the
          storage component and builds a DAG of the releases. It is responsible for
          verifying that the graph described by the releases is acyclic and connected.

          A **Policy Engine** is in charge of altering a client's view of the graph
          by applying a set of filters which are defined within the particular Policy
          Engine instance. Both the input to and the output from Policy Engines is a
          graph, allowing multiple Policy Engines to be chained together. The first
          Policy Engine in a chain will fetch its graph from the Graph Builder and the
          last Policy Engine in a chain will serve the modified graph to the client.

          An **Update Service client** is the end consumer of the release payloads. The
          client periodically queries the Policy Engine for updates and applys them if
          available.

          # Query OpenShift's Update Service Endpoint
          $ curl --silent --header 'Accept:application/json' 'https://api.openshift.com/api/upgrades_info/v1/graph?arch=amd64&channel=stable-4.2' | jq '. as $graph | $graph.nodes | map(.version == "4.2.13") | index(true) as $orig | $graph.edges | map(select(.[0] == $orig)[1]) | map($graph.nodes[.])'
        displayName: OpenShift Update Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Update Service
        - UpdateService
        - update-service
        - OSUS
        - Cincinnati
        - version
        - upgrade
        links:
        - name: Update Service Operator
          url: https://github.com/openshift/cincinnati-operator
        - name: OpenShift Update Service
          url: https://github.com/openshift/cincinnati
        - name: OpenShift Graph Data
          url: https://github.com/openshift/cincinnati-graph-data
        maintainers:
        - email: aos-team-ota@redhat.com
          name: OpenShift Update Service maintainers
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-update-service/openshift-update-service-rhel8-operator@sha256:759f4de107a5b429933c5e44046ffb4ed75c674aea75440513a6df8fa93a83a3
        - registry.redhat.io/openshift-update-service/openshift-update-service-rhel8@sha256:a9468a662babb272917df92bc9c827535b4b37883edb440b6d2d7c32ffb2145e
        version: 5.0.3
      entries:
      - name: update-service-operator.v5.0.3
        version: 5.0.3
      - name: update-service-operator.v5.0.2
        version: 5.0.2
      - name: update-service-operator.v5.0.1
        version: 5.0.1
      - name: update-service-operator.v5.0.0
        version: 5.0.0
      - name: update-service-operator.v4.9.1
        version: 4.9.1
      - name: update-service-operator.v4.9.0
        version: 4.9.0
      name: v1
    defaultChannel: v1
    packageName: cincinnati-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: www.redhat.com
    name: dpu-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dpu-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.openshift.io/v1",
                "kind": "DpuOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dpu-operator",
                    "app.kubernetes.io/instance": "dpuoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dpuoperatorconfig",
                    "app.kubernetes.io/part-of": "dpu-operator"
                  },
                  "name": "dpuoperatorconfig-sample"
                },
                "spec": {
                  "mode": "host"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-13T13:42:44Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: dpu-operator-system
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DpuOperatorConfig is the Schema for the dpuoperatorconfigs
              API
            displayName: Dpu Operator Config
            kind: DpuOperatorConfig
            name: dpuoperatorconfigs.config.openshift.io
            version: v1
        description: This operator will manage and configure data processing unit
          (DPUs) to be used in accelerating/offloading k8s networking functions
        displayName: DPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dpu
        - dpu-operator
        - ipu
        links:
        - name: Dpu Operator
          url: https://github.com/openshift/dpu-operator
        maintainers:
        - email: wizhao@redhat.com
          name: William Zhao
        - email: bnemeth@redhat.com
          name: Balazs Nemeth
        maturity: alpha
        provider:
          name: Red Hat
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-dpu-rhel9-operator@sha256:9d8283238af545c332781b2edef6b3a2857bcf4a264c9bae26720cc48e338a4b
        - registry.redhat.io/openshift4/ose-dpu-daemon-rhel9@sha256:a9a375f70dfb4a730aca0f7e58619fb05ef226261de6980d04eeffb8a043e6a4
        version: 4.16.0-202509240123
      entries:
      - name: dpu-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: dpu-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: dpu-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: dpu-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: dpu-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: dpu-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: dpu-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: dpu-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: dpu-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: dpu-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: dpu-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: dpu-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: dpu-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: dpu-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: dpu-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: dpu-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: dpu-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: dpu-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: dpu-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: dpu-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: dpu-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: dpu-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: dpu-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: dpu-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: dpu-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: dpu-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: dpu-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: dpu-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: dpu-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      name: stable
    defaultChannel: stable
    packageName: dpu-operator
    provider:
      name: Red Hat
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: odr-cluster-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odr-cluster-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "VolumeReplicationGroup",
                "metadata": {
                  "name": "volumereplicationgroup-sample"
                },
                "spec": {
                  "async": {
                    "schedulingInterval": "10m"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ],
                        "name": "cpd"
                      },
                      {
                        "includedResources": [
                          "Deployment"
                        ],
                        "name": "deployments"
                      },
                      {
                        "excludedResources": [
                          ""
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includeClusterResources": true,
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "cpd",
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      },
                      {
                        "backupName": "deployments",
                        "includedResources": [
                          "Deployment"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "excludedResources": [
                          "ConfigMap",
                          "Secret",
                          "Deployment",
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      }
                    ]
                  },
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  },
                  "replicationState": "primary",
                  "s3Profiles": [
                    "s3-profile-of-east",
                    "s3-profile-of-west"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: MaintenanceMode
            name: maintenancemodes.ramendr.openshift.io
            version: v1alpha1
          - kind: ProtectedVolumeReplicationGroupList
            name: protectedvolumereplicationgrouplists.ramendr.openshift.io
            version: v1alpha1
          - description: VolumeReplicationGroup is the Schema for the volumereplicationgroups
              API
            displayName: Volume Replication Group
            kind: VolumeReplicationGroup
            name: volumereplicationgroups.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Cluster is a disaster-recovery orchestrator for
          stateful applications, that operates from an Advanced Cluster Management
          (ACM) managed cluster and is controlled by Openshift DR Hub operator to
          orchestrate the life-cycle of an application, and its state on the managed
          cluster.
        displayName: Openshift DR Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:769acf15d016d4992793c9c527d7c268d2a67502dd7e0fadc240e4896398195b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 4.15.20-rhodf
      entries:
      - name: odr-cluster-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odr-cluster-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odr-cluster-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odr-cluster-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odr-cluster-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odr-cluster-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odr-cluster-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odr-cluster-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odr-cluster-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odr-cluster-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odr-cluster-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odr-cluster-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odr-cluster-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odr-cluster-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odr-cluster-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odr-cluster-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odr-cluster-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odr-cluster-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odr-cluster-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odr-cluster-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odr-cluster-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "VolumeReplicationGroup",
                "metadata": {
                  "name": "volumereplicationgroup-sample"
                },
                "spec": {
                  "async": {
                    "schedulingInterval": "10m"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ],
                        "name": "cpd"
                      },
                      {
                        "includedResources": [
                          "Deployment"
                        ],
                        "name": "deployments"
                      },
                      {
                        "excludedResources": [
                          ""
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includeClusterResources": true,
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "cpd",
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      },
                      {
                        "backupName": "deployments",
                        "includedResources": [
                          "Deployment"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "excludedResources": [
                          "ConfigMap",
                          "Secret",
                          "Deployment",
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      }
                    ]
                  },
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  },
                  "replicationState": "primary",
                  "s3Profiles": [
                    "s3-profile-of-east",
                    "s3-profile-of-west"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: MaintenanceMode
            name: maintenancemodes.ramendr.openshift.io
            version: v1alpha1
          - kind: ProtectedVolumeReplicationGroupList
            name: protectedvolumereplicationgrouplists.ramendr.openshift.io
            version: v1alpha1
          - description: VolumeReplicationGroup is the Schema for the volumereplicationgroups
              API
            displayName: Volume Replication Group
            kind: VolumeReplicationGroup
            name: volumereplicationgroups.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Cluster is a disaster-recovery orchestrator for
          stateful applications, that operates from an Advanced Cluster Management
          (ACM) managed cluster and is controlled by Openshift DR Hub operator to
          orchestrate the life-cycle of an application, and its state on the managed
          cluster.
        displayName: Openshift DR Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:af262b0890300183db1766bc392017b0bda723a9ba841dc25e3d928aa413db9a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        version: 4.16.16-rhodf
      entries:
      - name: odr-cluster-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odr-cluster-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odr-cluster-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odr-cluster-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odr-cluster-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odr-cluster-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odr-cluster-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odr-cluster-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odr-cluster-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odr-cluster-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odr-cluster-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odr-cluster-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odr-cluster-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odr-cluster-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odr-cluster-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odr-cluster-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odr-cluster-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odr-cluster-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Couchbase
      provider-url: ""
    name: couchbase-enterprise-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: couchbase-operator.v2.6.4-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator@sha256:1f0a6865c360a14ba98706fd6fe8dff3bdbdede32e3622f93f19c82e377ca33f
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        version: 2.6.4-7
      entries:
      - name: couchbase-operator.v2.6.4-7
        version: 2.6.4-7
      - name: couchbase-operator.v2.6.4-6
        version: 2.6.4-6
      name: 2.6.4
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator@sha256:707571a7881bc12d5f1b0b1a94eb3c675d5c0ea0114c994a47ceb5aa79df4013
        version: 2.7.0-5
      entries:
      - name: couchbase-operator.v2.7.0-5
        version: 2.7.0-5
      name: 2.7.0
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:a79e1b6781d5f8badf664e7efea6e2889e43fae6e0d37169d766685dcd6ff394
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        version: 2.7.1-1
      entries:
      - name: couchbase-operator.v2.7.1-1
        version: 2.7.1-1
      name: 2.7.1
    - currentCSV: couchbase-operator.v2.8.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:df61928b064014d851c6e7dff6dcd63a433a27fa5623254b6e389c3eea170e36
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        version: 2.8.0-2
      entries:
      - name: couchbase-operator.v2.8.0-2
        version: 2.8.0-2
      - name: couchbase-operator.v2.8.0-1
        version: 2.8.0-1
      name: 2.8.0
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        version: 2.8.1-1
      entries:
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: 2.8.1
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        version: 2.8.1-1
      entries:
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: stable
    defaultChannel: stable
    packageName: couchbase-enterprise-certified-rhmp
    provider:
      name: Couchbase
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: topology-aware-lifecycle-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: topology-aware-lifecycle-manager.v4.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ran.openshift.io/v1alpha1",
                "kind": "ClusterGroupUpgrade",
                "metadata": {
                  "name": "ClusterGroupUpgrade-sample"
                },
                "spec": {
                  "preCaching": true
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:bf24cad2940a227f5bd2f188be5d298ff19176c634fcf609c7225d7e0a6f2fd1
          description: The topology aware lifecycle manager is an operator that facilitates
            platform upgrades of group of clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-operators",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/cluster-group-upgrades-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterGroupUpgrade is the Schema for the ClusterGroupUpgrades
              API
            displayName: Cluster Group Upgrade
            kind: ClusterGroupUpgrade
            name: clustergroupupgrades.ran.openshift.io
            version: v1alpha1
          - description: ImageBasedGroupUpgrade is the schema for upgrading a group
              of clusters using IBU
            displayName: Image-Based Group Upgrade
            kind: ImageBasedGroupUpgrade
            name: imagebasedgroupupgrades.lcm.openshift.io
            version: v1alpha1
          - description: PreCachingConfig is the Schema for the precachingconfigs
              API
            displayName: Pre-caching Config
            kind: PreCachingConfig
            name: precachingconfigs.ran.openshift.io
            version: v1alpha1
        description: |-
          Topology Aware Lifecycle Manager is an operator that facilitates
          platform and operator upgrades of group of clusters
        displayName: Topology Aware Lifecycle Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm lifecyclemanagement upgrades cluster
        links:
        - name: Cluster Group Upgrades Operator
          url: https://cluster-group-upgrades-operator.domain
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:eed8234b0fe47083ab941f513709f778bfcb80e32747b71a65fd5f62328a5ca0
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:bf24cad2940a227f5bd2f188be5d298ff19176c634fcf609c7225d7e0a6f2fd1
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-aztp-rhel9@sha256:f6a37a0de10c847a53e22406990858dbaa4cdd3be075bc047e8dcf5bc6a2f2f7
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-precache-rhel9@sha256:7673a0d7f8944b245124376f2153bf987299f5bd7350b33f0d4bd1ac18543929
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-recovery-rhel9@sha256:37e234f247c841cd12c4617f92bf8bd5a5dbcbf46fa1be042d56b966994efaad
        version: 4.16.5
      entries:
      - name: topology-aware-lifecycle-manager.v4.16.5
        version: 4.16.5
      - name: topology-aware-lifecycle-manager.v4.16.4
        version: 4.16.4
      - name: topology-aware-lifecycle-manager.v4.16.3
        version: 4.16.3
      - name: topology-aware-lifecycle-manager.v4.16.2
        version: 4.16.2
      - name: topology-aware-lifecycle-manager.v4.16.1
        version: 4.16.1
      - name: topology-aware-lifecycle-manager.v4.16.0
        version: 4.16.0
      name: "4.16"
    - currentCSV: topology-aware-lifecycle-manager.v4.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ran.openshift.io/v1alpha1",
                "kind": "ClusterGroupUpgrade",
                "metadata": {
                  "name": "ClusterGroupUpgrade-sample"
                },
                "spec": {
                  "preCaching": true
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:bf24cad2940a227f5bd2f188be5d298ff19176c634fcf609c7225d7e0a6f2fd1
          description: The topology aware lifecycle manager is an operator that facilitates
            platform upgrades of group of clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-operators",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/cluster-group-upgrades-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterGroupUpgrade is the Schema for the ClusterGroupUpgrades
              API
            displayName: Cluster Group Upgrade
            kind: ClusterGroupUpgrade
            name: clustergroupupgrades.ran.openshift.io
            version: v1alpha1
          - description: ImageBasedGroupUpgrade is the schema for upgrading a group
              of clusters using IBU
            displayName: Image-Based Group Upgrade
            kind: ImageBasedGroupUpgrade
            name: imagebasedgroupupgrades.lcm.openshift.io
            version: v1alpha1
          - description: PreCachingConfig is the Schema for the precachingconfigs
              API
            displayName: Pre-caching Config
            kind: PreCachingConfig
            name: precachingconfigs.ran.openshift.io
            version: v1alpha1
        description: |-
          Topology Aware Lifecycle Manager is an operator that facilitates
          platform and operator upgrades of group of clusters
        displayName: Topology Aware Lifecycle Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm lifecyclemanagement upgrades cluster
        links:
        - name: Cluster Group Upgrades Operator
          url: https://cluster-group-upgrades-operator.domain
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:eed8234b0fe47083ab941f513709f778bfcb80e32747b71a65fd5f62328a5ca0
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:bf24cad2940a227f5bd2f188be5d298ff19176c634fcf609c7225d7e0a6f2fd1
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-aztp-rhel9@sha256:f6a37a0de10c847a53e22406990858dbaa4cdd3be075bc047e8dcf5bc6a2f2f7
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-precache-rhel9@sha256:7673a0d7f8944b245124376f2153bf987299f5bd7350b33f0d4bd1ac18543929
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-recovery-rhel9@sha256:37e234f247c841cd12c4617f92bf8bd5a5dbcbf46fa1be042d56b966994efaad
        version: 4.16.5
      entries:
      - name: topology-aware-lifecycle-manager.v4.16.5
        version: 4.16.5
      - name: topology-aware-lifecycle-manager.v4.16.4
        version: 4.16.4
      - name: topology-aware-lifecycle-manager.v4.16.3
        version: 4.16.3
      - name: topology-aware-lifecycle-manager.v4.16.2
        version: 4.16.2
      - name: topology-aware-lifecycle-manager.v4.16.1
        version: 4.16.1
      - name: topology-aware-lifecycle-manager.v4.16.0
        version: 4.16.0
      name: stable
    defaultChannel: stable
    packageName: topology-aware-lifecycle-manager
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Korrel8r
      provider-url: https://github.com/korrel8r/korrel8r
    name: korrel8r
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: korrel8r.v0.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "korrel8r.openshift.io/v1alpha1",
                "kind": "Korrel8r",
                "metadata": {
                  "name": "korrel8r",
                  "namespace": "korrel8r"
                },
                "spec": {
                  "debug": {
                    "verbose": 1
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional,Monitoring,Logging & Tracing
          containerImage: quay.io/korrel8r/operator:0.1.8
          createdAt: "2024-09-27T17:45:00Z"
          description: Correlate cluster resources and observability data.
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "korrel8r.openshift.io/v1alpha1",
              "kind": "Korrel8r",
              "metadata": {
                "name": "korrel8r",
                "namespace": "korrel8r"
              }
            }
          operatorframework.io/suggested-namespace: korrel8r
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/korrel8r/korrel8r
          support: Korrel8r Project
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Korrel8r is a service that correlates observabililty signals
              in the cluster.
            displayName: Korrel8r
            kind: Korrel8r
            name: korrel8rs.korrel8r.openshift.io
            version: v1alpha1
        description: |
          Korrel8r helps navigate from problem symptoms to related resources and signal data that can reveal the cause.
          It can follow relationships between disjointed observability "silos" (logs, metrics, alerts and more) to bring together all the data available to solve a problem.
          This operator allows Korrel8r to be deployed as a cluster service with a REST API.
          It can be used as a back-end for data correlation by the OpenShift console and other tools.
        displayName: Korrel8r
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - signal
        - correlation
        links:
        - name: Documentation
          url: https://korrel8r.github.io/korrel8r
        - name: Source Code and Issues
          url: https://github.com/korrel8r/korrel8r
        maintainers:
        - email: korrel8r@gmail.com
          name: Korrel8r Project
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Korrel8r
          url: https://github.com/korrel8r/korrel8r
        relatedImages:
        - quay.io/korrel8r/operator:0.1.8
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 0.1.8
      entries:
      - name: korrel8r.v0.1.8
        version: 0.1.8
      - name: korrel8r.v0.1.7
        version: 0.1.7
      - name: korrel8r.v0.1.5
        version: 0.1.5
      - name: korrel8r.v0.1.3
        version: 0.1.3
      - name: korrel8r.v0.1.2
        version: 0.1.2
      - name: korrel8r.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: korrel8r
    provider:
      name: Korrel8r
      url: https://github.com/korrel8r/korrel8r
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Upbound Inc.
      provider-url: https://upbound.io
    name: universal-crossplane
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: universal-crossplane.1.5.1-up.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"pkg.crossplane.io/v1",
                "kind":"Configuration",
                "metadata":{
                  "name":"xp-getting-started-with-aws"
                },
                "spec":{
                  "package":"registry.upbound.io/xp/getting-started-with-aws:v1.2.2"
                }
              },
              {
                "apiVersion":"apiextensions.crossplane.io/v1",
                "kind":"CompositeResourceDefinition",
                "metadata":{
                  "name":"compositepostgresqlinstances.database.example.org"
                },
                "spec":{
                  "group":"database.example.org",
                  "names":{
                    "kind":"CompositePostgreSQLInstance",
                    "plural":"compositepostgresqlinstances"
                  },
                  "claimNames":{
                    "kind":"PostgreSQLInstance",
                    "plural":"postgresqlinstances"
                  },
                  "connectionSecretKeys":[
                      "username",
                      "password",
                      "endpoint",
                      "port"
                  ],
                  "versions":[
                    {
                      "name":"v1alpha1",
                      "served":true,
                      "referenceable":true,
                      "schema":{
                        "openAPIV3Schema":{
                          "type":"object",
                          "properties":{
                            "spec":{
                              "type":"object",
                              "properties":{
                                "parameters":{
                                  "type":"object",
                                  "properties":{
                                    "storageGB":{
                                      "type":"integer"
                                    }
                                  },
                                  "required":[
                                      "storageGB"
                                  ]
                                }
                              },
                              "required":[
                                  "parameters"
                              ]
                            }
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion":"apiextensions.crossplane.io/v1",
                "kind":"Composition",
                "metadata":{
                  "name":"compositepostgresqlinstances.aws.database.example.org",
                  "labels":{
                    "provider":"aws",
                    "guide":"quickstart",
                    "vpc":"default"
                  }
                },
                "spec":{
                  "writeConnectionSecretsToNamespace":"crossplane-system",
                  "compositeTypeRef":{
                    "apiVersion":"database.example.org/v1alpha1",
                    "kind":"CompositePostgreSQLInstance"
                  },
                  "resources":[
                    {
                      "name":"rdsinstance",
                      "base":{
                        "apiVersion":"database.aws.crossplane.io/v1beta1",
                        "kind":"RDSInstance",
                        "spec":{
                          "forProvider":{
                            "region":"us-east-1",
                            "dbInstanceClass":"db.t2.small",
                            "masterUsername":"masteruser",
                            "engine":"postgres",
                            "engineVersion":"9.6",
                            "skipFinalSnapshotBeforeDeletion":true,
                            "publiclyAccessible":true
                          },
                          "writeConnectionSecretToRef":{
                            "namespace":"crossplane-system"
                          }
                        }
                      },
                      "patches":[
                        {
                          "fromFieldPath":"metadata.uid",
                          "toFieldPath":"spec.writeConnectionSecretToRef.name",
                          "transforms":[
                            {
                              "type":"string",
                              "string":{
                                "fmt":"%s-postgresql"
                              }
                            }
                          ]
                        },
                        {
                          "fromFieldPath":"spec.parameters.storageGB",
                          "toFieldPath":"spec.forProvider.allocatedStorage"
                        }
                      ],
                      "connectionDetails":[
                        {
                          "fromConnectionSecretKey":"username"
                        },
                        {
                          "fromConnectionSecretKey":"password"
                        },
                        {
                          "fromConnectionSecretKey":"endpoint"
                        },
                        {
                          "fromConnectionSecretKey":"port"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion":"pkg.crossplane.io/v1",
                "kind":"Provider",
                "metadata":{
                  "name":"provider-aws"
                },
                "spec":{
                  "package":"crossplane/provider-aws:master"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime, Cloud Provider, Database, Integration &
            Delivery, Networking, Storage
          com.redhat.component: universal-crossplane
          com.redhat.delivery.backport: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.6
          containerImage: upbound/uxp-bootstrapper
          createdAt: 2021-05-18T00:00:00+0000
          description: Upbound Universal Crossplane (UXP) is Upbound's official enterprise-grade
            distribution of Crossplane.
          io.k8s.display-name: universal-crossplane
          io.openshift.tags: uxp,crossplane,upbound
          maintainer: Upbound Inc. <info@upbound.io>
          name: universal-crossplane
          ocs.tags: v4.6
          operatorframework.io/initialization-resource: |
            apiVersion: v1
            kind: Secret
            metadata:
              name: upbound-control-plane-token
              namespace: upbound-system
            type: Opaque
            data:
              token: "Output of `up cloud controlplane attach | base64` command"
          operatorframework.io/suggested-namespace: upbound-system
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io.bundle.channels.v1: stable
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: universal-crossplane
          operators.operatorframework.io/internal-objects: '["locks.pkg.crossplane.io"]'
          repository: https://github.com/upbound/universal-crossplane
          summary: Upbound Universal Crossplane (UXP) is Upbound's official enterprise-grade
            distribution of Crossplane.
          support: Upbound Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: CompositeResourceDefinition
            kind: CompositeResourceDefinition
            name: compositeresourcedefinitions.apiextensions.crossplane.io
            version: v1
          - displayName: CompositionRevision
            kind: CompositionRevision
            name: compositionrevisions.apiextensions.crossplane.io
            version: v1alpha1
          - displayName: Composition
            kind: Composition
            name: compositions.apiextensions.crossplane.io
            version: v1
          - displayName: ConfigurationRevision
            kind: ConfigurationRevision
            name: configurationrevisions.pkg.crossplane.io
            version: v1
          - displayName: Configuration
            kind: Configuration
            name: configurations.pkg.crossplane.io
            version: v1
          - displayName: ControllerConfig
            kind: ControllerConfig
            name: controllerconfigs.pkg.crossplane.io
            version: v1alpha1
          - displayName: Lock
            kind: Lock
            name: locks.pkg.crossplane.io
            version: v1alpha1
          - displayName: ProviderRevision
            kind: ProviderRevision
            name: providerrevisions.pkg.crossplane.io
            version: v1
          - displayName: Provider
            kind: Provider
            name: providers.pkg.crossplane.io
            version: v1
        description: |
          ## Overview

          Upbound Universal Crossplane (UXP) is [Upbound's][upbound] official
          enterprise-grade distribution of [Crossplane][crossplane]. It's fully compatible
          with upstream Crossplane, open source, capable of connecting to Upbound Cloud
          for real-time dashboard visibility, and maintained by Upbound. It's the easiest
          way for both individual community members and enterprises to build their
          production control planes.

          ## Quick Start

          1. Install the [Upbound CLI][upbound-cli].
              ```console
              curl -sL https://cli.upbound.io | sh
              # Install with Homebrew
              brew install upbound/tap/up
              ```

          2. [Create an Upbound account][create-account] for a free dashboard for UXP.

          3. Connect UXP to Upbound Cloud.
              ```console
              # The name of your new UXP control plane.
              UXP_NAME=mycrossplane

              up cloud login
              up cloud controlplane attach ${UXP_NAME} | up uxp connect -
              ```

          4. Manage your UXP control plane by [signing in][login] to your Upbound account.

          ## Additional Resources

          - The [UXP Documentation][uxp-documentation] provides additional information
            about UXP and resources for developers, like examples.
          - The [developer guide][developer-guide] describes how to build and run UXP
            locally from source.
          - [UXP Slack][uxp-slack] is where you can go to get all of your UXP questions
            answered.

          [upbound]: https://upbound.io
          [crossplane]: https://crossplane.io/
          [upbound-cli]: https://github.com/upbound/up
          [create-account]: https://cloud.upbound.io/register
          [login]: https://cloud.upbound.io/login
          [uxp-documentation]: https://cloud.upbound.io/docs/uxp
          [developer-guide]: https://github.com/upbound/universal-crossplane/blob/main/docs/developer-guide.md
          [uxp-slack]: https://crossplane.slack.com/archives/upbound
        displayName: Upbound Universal Crossplane (UXP)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - infrastructure
        - services
        - application
        - database
        - cache
        - bucket
        - infra
        - app
        - ops
        - oam
        - gcp
        - azure
        - aws
        - alibaba
        - cloudsql
        - rds
        - s3
        - azuredatabase
        - asparadb
        - gke
        - aks
        - eks
        links:
        - name: GitHub
          url: https://github.com/upbound/universal-crossplane
        - name: Website
          url: https://upbound.io
        - name: Twitter
          url: https://twitter.com/upbound_io
        - name: Slack
          url: https://slack.crossplane.io/
        maintainers:
        - email: info@upbound.io
          name: Upbound Inc.
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Upbound Inc.
          url: https://upbound.io
        relatedImages:
        - upbound/uxp-bootstrapper:v1.5.1-up.1
        - upbound/crossplane:v1.5.1-up.1
        - upbound/xgql:v0.1.5
        version: 1.5.1-up.1
      entries:
      - name: universal-crossplane.1.5.1-up.1
        version: 1.5.1-up.1
      - name: universal-crossplane.1.4.3-up.1
        version: 1.4.3-up.1
      - name: universal-crossplane.1.3.3-up.1
        version: 1.3.3-up.1
      - name: universal-crossplane.1.3.1-up.1
        version: 1.3.1-up.1
      - name: universal-crossplane.1.2.3-up.1
        version: 1.2.3-up.1
      - name: universal-crossplane.1.2.2-up.1
        version: 1.2.2-up.1
      - name: universal-crossplane.1.2.1-up.4
        version: 1.2.1-up.4
      - name: universal-crossplane.1.2.1-up.3
        version: 1.2.1-up.3
      name: stable
    defaultChannel: stable
    packageName: universal-crossplane
    provider:
      name: Upbound Inc.
      url: https://upbound.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Palo Alto Networks
      provider-url: ""
    name: pcc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pcc-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "name": "pcc-console",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "ConsoleDefender",
                "metadata": {
                  "name": "pcc-consoledefender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "defenderConfig": {
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Defender",
                "metadata": {
                  "name": "pcc-defender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "defenderConfig": {
                    "clusterAddress": "twistlock-console.example.com",
                    "consoleAddress": "https://twistlock-console.example.com:8083",
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: quay.io/prismacloud/pcc-operator@sha256:51a343d5ea0115e3c43ccaa546492e85ab74cf2cf2658d725c1ea7099fc89191
          description: Deploy Prisma Cloud Compute for cloud-native security in your
            clusters
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator
          support: pcc-integrations@paloaltonetworks.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Install the Prisma Cloud Compute Console and Defenders
            displayName: Console and Defenders
            kind: ConsoleDefender
            name: consoledefenders.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install the Prisma Cloud Compute Console
            displayName: Console
            kind: Console
            name: consoles.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install Prisma Cloud Compute Defenders
            displayName: Defenders
            kind: Defender
            name: defenders.pcc.paloaltonetworks.com
            version: v1alpha1
        description: |
          ## Features
          - Deploy Console
              - Create initial local administrator user
              - Add license
          - Deploy Defenders
          - Support for offline deployments
          - Upgrade Console
          - Upgrade Defenders

          ## Installation
          See the [Prisma Cloud Compute Operator documentation](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator) for installation guides.

          ## Support
          This operator is available "as is," and should be seen as community-supported; however, Palo Alto Networks will continue contributing whenever possible.
          Palo Alto Networks does not provide technical support or help with using or troubleshooting the operator through normal support channels.
          We encourage you to open GitHub [issues](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/issues) and [pull requests](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/pulls) to track bugs and feature requests.

          ## Other links
          [Prisma Cloud Compute product documentation](https://docs.paloaltonetworks.com/prisma/prisma-cloud/prisma-cloud-admin-compute.html)

          [Prisma Cloud Compute API documentation](https://prisma.pan.dev/api/cloud/cwpp)
        displayName: Prisma Cloud Compute Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - palo
        - palo alto
        - palo alto networks
        - paloalto
        - twistlock
        - compute
        - prisma
        - prisma cloud
        - prisma cloud compute
        - security
        - image
        - images
        - container
        - containers
        - docker
        - podman
        - vulnerability
        - vulnerabilities
        - compliance
        - runtime
        - scan
        - scanning
        - monitor
        - monitoring
        - alert
        - alerting
        links:
        - name: Prisma Cloud
          url: https://www.paloaltonetworks.com/prisma/cloud
        maintainers:
        - email: pcc-integrations@paloaltonetworks.com
          name: Wyatt Gill
        maturity: alpha
        provider:
          name: Palo Alto Networks
        relatedImages:
        - quay.io/prismacloud/pcc-operator@sha256:51a343d5ea0115e3c43ccaa546492e85ab74cf2cf2658d725c1ea7099fc89191
        version: 0.2.3
      entries:
      - name: pcc-operator.v0.2.3
        version: 0.2.3
      name: stable
    defaultChannel: stable
    packageName: pcc-operator
    provider:
      name: Palo Alto Networks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-iam-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-iam-operator.v4.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "Authentication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-iam-operator",
                    "app.kubernetes.io/managed-by": "ibm-iam-operator",
                    "app.kubernetes.io/name": "ibm-iam-operator"
                  },
                  "name": "example-authentication"
                },
                "spec": {
                  "auditService": {
                    "imageName": "auditService no longer used - ignore",
                    "imageRegistry": "auditService no longer used - ignore",
                    "imageTag": "auditService no longer used - ignore"
                  },
                  "authService": {
                    "imageName": "icp-platform-auth",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "ldapsCACert": "platform-auth-ldaps-ca-cert",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "650Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "400Mi",
                        "memory": "350Mi"
                      }
                    },
                    "routerCertSecret": "saml-auth-secret"
                  },
                  "clientRegistration": {
                    "imageName": "im-initcontainer",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "158Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "config": {
                    "authUniqueHosts": "internal-ip1 internal-ip2 mycluster.icp",
                    "bootstrapUserId": "kubeadmin",
                    "claimsMap": "name=\"givenName\" family_name=\"givenName\" given_name=\"givenName\" preferred_username=\"displayName\" display_name=\"displayName\"",
                    "claimsSupported": "name,family_name,display_name,given_name,preferred_username",
                    "clusterCADomain": "mycluster.icp",
                    "clusterExternalAddress": "10.0.0.1",
                    "clusterInternalAddress": "10.0.0.1",
                    "clusterName": "mycluster",
                    "defaultAdminPassword": "password",
                    "defaultAdminUser": "cpadmin",
                    "defaultLogin": "",
                    "enableImpersonation": false,
                    "fipsEnabled": true,
                    "ibmCloudSaas": false,
                    "icpPort": 8443,
                    "installType": "fresh",
                    "isOpenshiftEnv": true,
                    "nonceEnabled": true,
                    "oidcIssuerURL": "https://127.0.0.1:443/idauth/oidc/endpoint/OP",
                    "openshiftPort": 443,
                    "preferredLogin": "",
                    "providerIssuerURL": "",
                    "roksEnabled": true,
                    "roksURL": "https://roks.domain.name:443",
                    "roksUserPrefix": "changeme",
                    "saasClientRedirectUrl": "",
                    "scimAdminPassword": "scimpassword",
                    "scimAdminUser": "scimadmin",
                    "scopeClaim": "profile=\"name,family_name,display_name,given_name,preferred_username\"",
                    "wlpClientID": "4444be3a738841016ab76d71b650e836",
                    "wlpClientRegistrationSecret": "f1362ca4d20b8389af2d1ea68042c9af",
                    "wlpClientSecret": "aa73bf39752053bf723d1143fb4cf8a2",
                    "xframeDomain": "",
                    "zenFrontDoor": false
                  },
                  "identityManager": {
                    "imageName": "icp-identity-manager",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "masterNodesList": "10.0.0.1",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "550Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "ephemeral-storage": "300Mi",
                        "memory": "150Mi"
                      }
                    }
                  },
                  "identityProvider": {
                    "imageName": "icp-identity-provider",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "550Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "ephemeral-storage": "300Mi",
                        "memory": "150Mi"
                      }
                    }
                  },
                  "initMongodb": {
                    "imageName": "im-initcontainer",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "178Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "operatorVersion": "0.14.1",
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
          createdAt: "2025-02-20T00:03:25Z"
          description: The IAM operator provides a simple Kubernetes CRD-Based API
            to manage the lifecycle of IAM services. With this operator, you can simply
            deploy and upgrade the IAM services
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.11.0
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-iam-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Authentication is the Schema for the authentications API
            displayName: Authentication
            kind: Authentication
            name: authentications.operator.ibm.com
            version: v1alpha1
          - description: Client is the Schema for the clients API
            displayName: Client
            kind: Client
            name: clients.oidc.security.ibm.com
            version: v1
        description: |-
          The `ibm-iam-operator` installs the IBM Cloud Platform Common Services Identity and access management (IAM) service.

          **Important:** Do not install this operator directly. Install this operator only by using the IBM Common Service Operator. For more information about installing this operator and other Common Services operators, see [Installer documentation](http://ibm.biz/cpcs_opinstall). Additionally, you can exit this panel and navigate to the IBM Common Services tile in OperatorHub to learn more about the operator. If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For more information about IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).

          You can use the `ibm-iam-operator` to install the authentication and authorization services for the IBM Cloud Platform Common Services.

          With these services, you can configure security for IBM Cloud Platform Common Services, IBM Certified Containers (IBM products), or IBM Cloud Paks that are installed.

          For more information about the available IBM Cloud Platform Common Services, see the [IBM Knowledge Center](http://ibm.biz/cpcsdocs).

          ## Supported platforms

           - Red Hat OpenShift Container Platform 4.3 or newer installed on one of the following platforms:

              - Linux x86_64
              - Linux on Power (ppc64le)
              - Linux on IBM Z and LinuxONE

          ## Prerequisites

          Before you install this operator, you need to first install the operator dependencies and prerequisites:

          - For the list of operator dependencies, see the IBM Knowledge Center [Common Services dependencies documentation](http://ibm.biz/cpcs_opdependencies).

          - For the list of prerequisites for installing the operator, see the IBM Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).

          ## Documentation

          To install the operator by using the IBM Common Services Operator, follow the installation and configuration instructions that are in the IBM Knowledge Center.

          - If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).
          - If you are using the operator with an IBM Containerized Software, see the IBM Cloud Platform Common Services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall).
        displayName: IBM IM Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:b8886b9b54e6ee4c6b2981ef092e54348a0286577d9924cf136cee881f5bf1f4
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-bundle@sha256:98cf2a6c7b7af5d7cbfbc61ce140201272b30b12edabaec2b2d4287b0690d89e
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:f289ac0d44803ff12d41c3659e6fdf04c404e0b3b233b6bc44eec671abb1b47e
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:52ce08ac2bce7c02dc30e2c0aa74df4bde8f156debccc59960f02b4b0040ffea
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:97b9ecef8caef156a0eb073b6eb2a34b63209b44511f5694a243f11285d88662
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:0f46b288d060a15a48ecad8f328a8046bf232b8629c3d79da6a82ea712b363f4
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:7cec55c00827b747f23c2f1703d99e1ca5b30a0c445f8ad0471b25b0c9cca57c
        version: 4.11.0
      entries:
      - name: ibm-iam-operator.v4.11.0
        version: 4.11.0
      name: v4.10
    defaultChannel: v4.10
    packageName: ibm-iam-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: vfunction-operator
      app.kubernetes.io/managed-by: operator
      app.kubernetes.io/name: vfunction
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      name: vfunction-server-operator
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: vFunction
      provider-url: ""
    name: vfunction-server-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vfunction-server-operator.v4.3.1762
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vfunction.com/v1",
                "kind": "VfunctionServer",
                "metadata": {
                  "name": "vfunction-server",
                  "namespace": "vfunction",
                  "labels": {
                      "name": "vfunction",
                      "app.kubernetes.io/name": "vfunction",
                      "app.kubernetes.io/instance": "vfunction-server",
                      "app.kubernetes.io/managed-by": "operator"
                  }
                },
                "spec": {
                  "admin": {
                    "email": "admin@mycompany.com",
                    "name": "Admin",
                    "password": "Password1!"
                  },
                  "host": "http://my.domain.com",
                  "org_name": "MyCompany",
                  "upgrade": "Never",
                  "measurements": {
                    "min_number_of_services": "2",
                    "max_number_of_services": "10",
                    "max_pod_memory_capacity": "16G",
                    "max_pod_cpu_capacity": "1",
                    "pod_ephemeral_storage": "1G",
                    "S3": {
                      "bucket": "",
                      "key": "",
                      "secret": "",
                      "region": ""
                    }
                  },
                  "smtp": {
                    "identity": "",
                    "password": "",
                    "url": "",
                    "user": ""
                  },
                  "authentication": {
                    "authority": "",
                    "client_id": "",
                    "client_secret": "",
                    "jwks_url": "",
                    "issuer": "",
                    "ca_root_crt": ""
                  },
                  "tls": {
                    "use_letsencrypt": "No",
                    "crt": "",
                    "key": ""
                  },
                  "nginx": {
                    "force_http": "No",
                    "ipv6_disabled": "No",
                    "service_type": ""
                  },
                  "route": {
                    "use_route": "Yes",
                    "tls_termination": "edge",
                    "use_tls_certificate": "No"
                  },
                  "mysql": {
                    "external_mysql_ip": "",
                    "external_mysql_user": "",
                    "external_mysql_password": "",
                    "max_pod_memory_capacity": "4G",
                    "max_pod_cpu_capacity": "1",
                    "pod_ephemeral_storage": "1G"
                  },
                  "storage": {
                    "storage_class": "",
                    "mysql_persistent_volume": "",
                    "mysql_persistent_volume_claim": "",
                    "mysql_persistent_volume_claim_access_mode": "ReadWriteOnce",
                    "storage_persistent_volume": "",
                    "storage_persistent_volume_claim": "",
                    "storage_persistent_volume_claim_access_mode": "ReadWriteOnce",
                    "lets_encrypt_persistent_volume": "",
                    "lets_encrypt_persistent_volume_claim": "",
                    "lets_encrypt_persistent_volume_claim_access_mode": "ReadWriteOnce"
                  },
                  "distributed": {
                    "otlp": {
                      "min_number_of_services": "1",
                      "max_number_of_services": "10",
                      "max_pod_memory_capacity": "8G",
                      "max_pod_cpu_capacity": "1",
                      "pod_ephemeral_storage": "250M",
                      "max_folder_size_in_mb": "200"
                    },
                    "dm": {
                      "min_number_of_services": "1",
                      "max_number_of_services": "10",
                      "max_pod_memory_capacity": "8G",
                      "max_pod_cpu_capacity": "1",
                      "pod_ephemeral_storage": "1G"
                    }
                  },
                  "backup": {
                    "mysql": {
                      "restore_during_installation": "No",
                      "S3": {
                        "bucket": "",
                        "key": "",
                        "secret": "",
                        "region": ""
                      }
                    }
                  },
                  "security": {
                    "image_pull_policy": "",
                    "force_hashed_images": "No",
                    "set_pod_resources": "Yes",
                    "proxy": {
                      "http_proxy": "",
                      "https_proxy": "",
                      "additional_no_proxy": ""
                    },
                    "use_rate_limiting": "No",
                    "requests_per_second": "10"
                  },
                  "affinity": {
                    "node_affinity": [
                      {
                        "key": "",
                        "operator": "",
                        "values": "",
                        "services": ""
                      }
                    ],
                    "node_selectors": [
                      {
                        "key": "",
                        "value": "",
                        "services": ""
                      }
                    ],
                    "additional_labels": [
                      {
                        "key": "",
                        "value": "",
                        "services": ""
                      }
                    ],
                    "additional_annotations": [
                      {
                        "key": "",
                        "value": "",
                        "services": ""
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, AI/Machine Learning
          certified: Red Hat certified
          containerImage: registry.connect.redhat.com/vfunction/vfunction-server-operator@sha256:a5a0f47f7fb9219139f8d749da14950a8c39698b340644011485dfbec54ebe8b
          createdAt: "2025-08-04T16:10:32Z"
          description: vFunction is a cutting-edge code analysis, machine learning,
            and automation to boost your Java modernization projects.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.3.1762
          support: vFunction
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of vFunction running on your cluster.
            displayName: vFunction
            kind: VfunctionServer
            name: vfunctionservers.vfunction.com
            version: v1
        description: |2-

          ## Introduction

          [vFunction](https://www.vfunction.com/) is a cloud-native modernization platform that combines dynamic and static code analysis, machine learning, and automation to automatically identify and extract services from existing applications. vFunction is the only platform purpose-built for modernization of Java applications.

          The vFunction modernization process starts by learning the running monolithic application, and surfacing the interdependencies within it. Using AI, the platform analyzes and identifies services that can be separated from the application. This decomposition can present a range of micro, mini, or even macro services, depending on your application environment, each being an independently deployable and scalable application component.

          vFunction automates the extraction of these services, enabling you to modernize your monolith, quickly and easily.

          ## The vFunction platform

          The platform consists of 3 basic components; the server, the controller package, and a tools package. The server runs as an operand on an OpenShift environment. The controller package is installed on the machine that runs the monolithic application which can be either a Linux or a Windows machine, and the tools are run on a development machine, with access to the code of the monolithic application.

          The controller package consists of three elements: the vFunction agent, that collects data during the dynamic analysis phase; the vFunction Viper application, that performs static analysis on the binaries of the application; and the vFunction controller that handles all the communication between the agent, Viper, and the vFunction server.

          The vFunction agent is a mix of a Java and native agent, and needs to run on the JVM that is currently running your application. [Refer to the vFunction OOB Support Matrix document](https://drive.google.com/file/d/1ccq8LFab1FrYAimDUxwgjiCOdk4QuVzs/view) for a list of supported application servers and JVMs.

          This operator installs a vFunction server instance to be connected later with one or more vFunction controllers installed on your application machine(s).

          ## Before You Start

          The vFunction operator requires a Red Hat OpenShift Kubernetes Application Platform 4.x.

          The cluster recommended configuration (which is the same as the minimal one) should consist of:

          - *1 master node*

          - *1 worker node*

          - *Storage capacity enough for dynamically provision 2 PVs of 50G each*

          - *A default StorageClass configured*

          vFunction doesn’t provide any encryption mechanism for data stored on any attached storage. If required, encryption can be achieved by the client by using its own external means on the storage itself.

          vFunction supports backup of all critical data as part of the application. Nevertheless, for complete backup of entire data it is recommended that the client apply its own storage based backup mechanism.

          ## Install Prerequisites

          ###Cluster:
          The operator was certified and tested on OCP 4.6 - 4.17.
          If you encounter any issue with other OCP versions or cloud providers please contact info@vfunction.com.

          ###Storage:
          The operator creates two new PersistentVolumeClaims (PVCs) during installation time. For their creation, it relies on the default StorageClass to dynamically provision two PersistentVolumes (PVs). The system must have a default StorageClass in place.
          Both PersistentVolumes are defaulted to access with ReadWriteOnce mode, but you can change this by specifying a different access mode in the YAML file.
          Notes:
          - There is an option to specify a specific StorageClass for the PersistentVolumeClaims.
          - There is an option to specify the use of a specific pre-created PersistentVolume for each PersistentVolumeClaim
          - There is an option to specify a pre-created PersistentVolumeClaim for each Deployment.
          - There is an option to change the access mode for each PersistentVolumeClaim.

          ###Project:
          The operator should be installed in a new and dedicated project (namespace). If you intend to install more than one vFunction operands in the same cluster, please use different projects for each one. Installation of more than one operand on the same cluster will allow scaling by load balancing multiple applications to different vFunction servers.

          ## Installation instructions

          1. Prepare the YAML file to use in the installation:

              a. Copy the following YAML template into a text editor:

              ```yaml
                  apiVersion: vfunction.com/v1
                  kind: VfunctionServer
                  metadata:
                    name: vfunction
                    namespace: vfunction
                    labels:
                      name: vfunction
                      app.kubernetes.io/name: vfunction
                      app.kubernetes.io/instance: vfunction-server
                      app.kubernetes.io/managed-by: operator
                  spec:
                    host: "http://my.domain.com"
                    org_name: "MyCompany"
                    upgrade: "Never"
                    admin:
                      email: "admin@mycompany.com"
                      name: "Admin"
                      password: "Password1!"
                    measurements:
                      min_number_of_services: "2"
                      max_number_of_services: "10"
                      max_pod_memory_capacity: "16G"
                      max_pod_cpu_capacity: "1"
                      S3:
                        bucket: ""
                        key: ""
                        secret: ""
                    smtp:
                      password: ""
                      url: ""
                      identity: ""
                      user: ""
                    authentication:
                      authority: ""
                      client_id: ""
                      client_secret: ""
                      jwks_url: ""
                      issuer: ""
                      ca_root_crt: |
                        -----BEGIN CERTIFICATE-----
                        ...
                        -----END CERTIFICATE-----
                    tls:
                      use_letsencrypt: "No"
                      crt: |
                        -----BEGIN CERTIFICATE-----
                        ...
                        -----END CERTIFICATE-----
                      key: |
                        -----BEGIN PRIVATE KEY-----
                        ...
                        -----END PRIVATE KEY-----
                    nginx:
                      force_http: "No"
                      ipv6_disabled: "No"
                      service_type: ""
                    route:
                      use_route: "Yes"
                      tls_termination: "edge"
                      use_tls_certificate: "No"
                    mysql:
                      external_mysql_ip: ""
                      external_mysql_user: ""
                      external_mysql_password: ""
                      max_pod_memory_capacity: "4G"
                      max_pod_cpu_capacity: "1"
                    storage:
                      storage_class: ""
                      mysql_persistent_volume: ""
                      mysql_persistent_volume_claim: ""
                      mysql_persistent_volume_claim_access_mode: "ReadWriteOnce"
                      storage_persistent_volume: ""
                      storage_persistent_volume_claim: ""
                      storage_persistent_volume_claim_access_mode: "ReadWriteOnce"
                      lets_encrypt_persistent_volume: ""
                      lets_encrypt_persistent_volume_claim: ""
                      lets_encrypt_persistent_volume_claim_access_mode: "ReadWriteOnce"
                    distributed:
                        otlp:
                          min_number_of_services: "1"
                          max_number_of_services: "10"
                          max_pod_memory_capacity: "8G"
                          max_pod_cpu_capacity: "1"
                          max_folder_size_in_mb: "200"
                        dm:
                          min_number_of_services: "1"
                          max_number_of_services: "10"
                          max_pod_memory_capacity: "8G"
                          max_pod_cpu_capacity: "1"
                    backup:
                      mysql:
                        restore_during_installation: "No"
                        S3:
                          bucket: ""
                          key: ""
                          secret: ""
                    security:
                      image_pull_policy: ""
                      force_hashed_images: "No"
                      set_pod_resources: "Yes"
                      proxy:
                        http_proxy: ""
                        https_proxy: ""
                        additional_no_proxy: ""
                      use_rate_limiting: "No"
                      requests_per_second: "10"
                    affinity:
                      node_affinity:
                        - key: ""
                          operator: ""
                          values: ""
                          services: ""
                      node_selectors:
                        - key: ""
                          value: ""
                          services: ""
                      additional_labels:
                        - key: ""
                          value: ""
                          services: ""
                      additional_annotations:
                        - key: ""
                          value: ""
                          services: ""
              ```
              b. Customize the template fields by replacing with your information:

              - **host**: Enter the FQDN for accessing the vFunction dashboard.
                          Make sure that you use a domain name and not an IP address. Writing "https://" will mandate TLS while writing "http://" will expose the server through HTTP.
              - **org_name**: Enter your organization name.
              - **upgrade**: Choose auto-upgrade mode.
                  - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically.
                  - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
                  - **Never**: The automatic upgrading mechanism is off (default).
                  - **Maintenance**: Put the operator in a maintenance mode.
              - **admin.email**: Enter the email address of your administrator.
              - **admin.name**: Enter the name of your administrator.
              - **admin.password**: Enter the password you want to use for the vFunction administrator.
                                    The password should be at least 8 characters long, and consist of at least one lowercase letter, at least one uppercase letter, at least one number, and at least one special character.
              - **measurements.min\_num\_of\_services**: Minimum Number of measurement service pods.
              - **measurements.max\_num\_of\_services**: Maximum Number of measurement service pods.
              - **measurements.max\_pod\_memory\_capacity**: Maximum memory capacity limit for measurement service pods. (Default is "8G").
              - **measurements.max\_pod\_cpu\_capacity**: Maximum CPU capacity limit for measurement service pods. (Default is "1").
              - **measurements.pod\_ephemeral\_storage**: Ephemeral Storage Capacity for measurement service pods. (Default is "1G").
              - **measurements.S3**: Save all measurements data to S3, instead of local PV (optional).
                  - **measurements.S3.bucket**: Measurements S3 bucket name.
                  - **measurements.S3.key**: Measurements S3 key.
                  - **measurements.S3.secret**: Measurements S3 secret.
                  - **measurements.S3.region**: Measurements S3 region.
              - **smtp.user**: Enter the email address for a designated user for the SMTP server (optional).
              - **smtp.password**: Enter this user’s password (optional).
              - **smtp.identity**: Enter this user’s password (optional).
              - **smtp.url**: Enter the SMTP server URL (optional).
              - **authentication.authority**: Authority for authentication (Optional).
              - **authentication.client_id**: Client ID for authentication (Optional).
              - **authentication.client_secret**: Client Secret for authentication (Optional).
              - **authentication.jwks_url**: JWKS URL for authentication (Optional).
              - **authentication.issuer**: Issuer for authentication (Optional).
              - **authentication.ca\_root\_crt**: CA root certificate for authentication (Optional).
              - **tls.use_letsencrypt**: If your host FQDN starts with “https”, choose whether to use Let’s Encrypt to obtain certificates.
              - **tls.crt**: If your host FQDN starts with “https”, paste in the certifications you have for using the TLS connection (Applicable only if you don’t use Let’s Encrypt).
              - **tls.key**: If your host FQDN starts with “https”, enter the key you have for using the TLS connection (Applicable only if you don’t use Let’s Encrypt).
              - **nginx.force_http**: Force internal NGINX to listen on http even if the host FQDN starts with “https”.
                                      This is useful if you want to use a TLS termination proxy in front of vFunction, other then the route provided by the operator.
              - **nginx.ipv6_disabled**: Disable IPv6 on internal NGINX.
              - **nginx.service_type**: Set the service type for the internal NGINX service - ClusterIP, NodePort or LoadBalancer.
              - **route.use_route**: Use OpenShift route to expose vFunction. (Default is "Yes").
              - **route.tls_termination**: Set the TLS termination type for the OpenShift route - edge, passthrough or reencrypt. (Default is "edge").
              - **route.use\_tls\_certificate**: Use TLS certificate for OpenShift route (as provided in tls.crt and tls.key) - instead of the default one defined for the route controller in your cluster. (Default is "No").
              - **mysql.external\_mysql\_ip**: If you want to use an external MySQL database, enter its IP address.
              - **mysql.external\_mysql\_user**: If you want to use an external MySQL database, enter its user name.
              - **mysql.external\_mysql\_password**: If you want to use an external MySQL database, enter its password.
              - **mysql.max\_pod\_memory\_capacity**: Maximum memory capacity limit for MySQL pod. (Default is "2G").
              - **mysql.max\_pod\_cpu\_capacity**: Maximum CPU capacity limit for MySQL pod. (Default is "1").
              - **mysql.pod\_ephemeral\_storage**: Ephemeral Storage Capacity for MySQL pod. (Default is "1G").
              - **storage.storage\_class**: If you want to use a specific StorageClass for all PVCs, other than the default one, enter its name.
              - **storage.mysql\_persistent\_volume**: If you want to use a specific PV for mysql PVC, other than the dynamically provisioned one, enter its name.
              - **storage.mysql\_persistent\_volume\_claim**: If you want to use a specific pre-created PVC for mysql deployment, other than the dynamically provisioned one, enter its name.
              - **storage.mysql\_persistent\_volume\_claim\access\_mode**: If you want to use a specific access mode for mysql PVC, other than the default one, enter its name. (Default is "ReadWriteOnce").
              - **storage.storage\_persistent\_volume**: If you want to use a specific PV for storage PVC, other than the dynamically provisioned one, enter its name.
              - **storage.storage\_persistent\_volume\_claim**: If you want to use a specific pre-created PVC for storage deployment, other than the dynamically provisioned one, enter its name.
              - **storage.storage\_persistent\_volume\_claim\access\_mode**: If you want to use a specific access mode for storage PVC, other than the default one, enter its name. (Default is "ReadWriteOnce").
              - **storage.lets\_encrypt\_persistent\_volume**: If you want to use a specific PV for lets encrypt PVC, other than the dynamically provisioned one, enter its name.
              - **storage.lets\_encrypt\_persistent\_volume\_claim**: If you want to use a specific pre-created PVC for lets encrypt deployment, other than the dynamically provisioned one, enter its name.
              - **storage.lets\_encrypt\_persistent\_volume\_claim\access\_mode**: If you want to use a specific access mode for lets encrypt PVC, other than the default one, enter its name. (Default is "ReadWriteOnce").
              - **distributed.otlp.min\_number\_of\_services**: Minimum Number of OTLP pods.
              - **distributed.otlp.max\_number\_of\_services**: Maximum Number of OTLP pods.
              - **distributed.otlp.max\_pod\_memory\_capacity**: Maximum memory capacity for OTLP pods. (Default is "1G").
              - **distributed.otlp.max\_pod\_cpu\_capacity**: Maximum CPU capacity for OTLP pods. (Default is "1").
              - **distributed.otlp.pod\_ephemeral\_storage**: Ephemeral Storage Capacity for OTLP service pods. (Default is "250M").
              - **distributed.otlp.max\_folder\_size\_in\_mb**: Maximum folder size in MB for OTLP pods. (Default is "200").
              - **distributed.dm.min\_number\_of\_services**: Minimum Number of dm pods.
              - **distributed.dm.max\_number\_of\_services**: Maximum Number of dm pods.
              - **distributed.dm.max\_pod\_memory\_capacity**: Maximum memory capacity for dm pods. (Default is "8G").
              - **distributed.dm.max\_pod\_cpu\_capacity**: Maximum CPU capacity for dm pods. (Default is "1").
              - **distributed.dm.pod\_ephemeral\_storage**: Ephemeral Storage Capacity for dm service pods. (Default is "1G").
              - **backup.mysql.restore\_during\_installation**: If you want, during installation, to restore the backup from the latest kept backup with the same major version (if available on attached DB storage or S3 bucket), choose "Yes".
              - **backup.mysql.S3**: Backup MySQL to S3 (optional).
                  - **backup.mysql.S3.bucket**: MySQL backup S3 bucket name.
                  - **backup.mysql.S3.key**: MySQL backup S3 key.
                  - **backup.mysql.S3.secret**: MySQL backup S3 secret.
                  - **backup.mysql.S3.region**: MySQL backup S3 region.
              - **security.image\_pull\_policy**: If you want to use a custom image pull policy for all deployments, enter it here (optional).
              - **security.force\_hashed\_images**: If you want to use hashed images for all deployments, choose "Yes".
              - **security.set\_pod\_resources**: If you want to not set resource requests and limits for all deployments, choose "No".
              - **security.proxy.http\_proxy**: Enter the address of an HTTP proxy (optional).
              - **security.proxy.https\_proxy**: Enter the address of an HTTPS proxy (optional).
              - **security.proxy.additional\_no\_proxy**: Enter additional addresses to exclude from proxying (optional).
              - **security.use\_rate\_limiting**: If you want to use rate limiting in nginx, choose "Yes".
              - **security.requests\_per\_second**: Enter the number of requests per second for rate limiting.
              - **affinity.node\_affinity**: Enter the node affinity for the deployments (optional). There is only one nodeSelectorTerm as requiredDuringSchedulingIgnoredDuringExecution, but it can contain multiple matchExpressions (which are ANDed together).
                  - **key**: Enter the key for the node affinity.
                  - **operator**: Enter the operator for the node affinity (In, NotIn, Exists, DoesNotExist, Gt, Lt).
                  - **values**: Enter the values for the node affinity. A comma-separated list of values.
                  - **services**: Enter the services for the node affinity. A comma-separated list of services to which the node affinity applies. If not specified, the node affinity will apply to all services. List of possible services: mysql, nginx, storage, vfapi-parser, vfapi-users, vfapi-organizations, vfapi-idp, vfapi-mcp, vfapi-measurements, vfapi-otlp, vfapi-dm. (optional).
              - **affinity.node\_selectors**: Enter the node selectors for the deployments (optional).
                  - **key**: Enter the key for the node selector.
                  - **value**: Enter the value for the node selector.
                  - **services**: Enter the services for the node selector. A comma-separated list of services to which the node selector applies. If not specified, the node selector will apply to all services. List of possible services: mysql, nginx, storage, vfapi-parser, vfapi-users, vfapi-organizations, vfapi-idp, vfapi-mcp, vfapi-measurements, vfapi-otlp, vfapi-dm. (optional).
              - **affinity.additional\_labels**: Enter additional labels for the deployments (optional).
                  - **key**: Enter the key for the additional label.
                  - **value**: Enter the value for the additional label.
                  - **services**: Enter the services for the additional label. See above at the node selectors section for more details. (optional).
              - **affinity.additional\_annotations**: Enter additional annotations for the deployments (optional).
                  - **key**: Enter the key for the additional annotation.
                  - **value**: Enter the value for the additional annotation.
                  - **services**: Enter the services for the additional annotation. See above at the node selectors section for more details. (optional).
          1. In the Installed Operators window choose and click on vFunction Operator and then Create Instance.
          1. In the Create vFunctionServer window, paste the YAML file that you created in step 1.
          1. Click Create.
          1. You can now click on your new created vFunction operand in order to see its details and installation progress.

          ## Verifying the installation

          Check that the operand installed successfully:

          1. Check that the "Operand State" property shows “Working”.
          1. Check that the "Successfully Installed" property shows "Yes".

          If "Operand State" shows "Failed", it indicates the operand wasn’t installed correctly. Check the events and logs for all pods (operator and image containers) for any issues. In the event that you cannot troubleshoot, contact vFunction support.

          ## After installation

          The vFunction site is now accessible via the newly created vFunction application custom address, for example, my.domain.com. There are two ways you can access the vFunction dashboard.

          ### Access using the router canonical hostname

          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address, and to the ”vfunction” subdomain of the OpenShift canonical router hostname as the alias.
          For example, ***my.domain.com.   CNAME   vfunction.apps.ocp4.my-openshift.com.***
          2. Find your cluster Router Canonical Hostname address in the newly created vfunction-route-xxx under your vFunction OpenShift project -> **Networking** -> **Routes** -> **vfunction-route-xxx** route -> **Router Canonical Hostname** field.
          You can now access the dashboard using your defined "host" spec property (as above).

          ### Access using the nginx service location

          You may use this access method if your OpenShift is installed on a provider that supports exposing LoadBalancer-type services.
          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address and to the vfunction-nginx-xxx service location.
          For example, ***my.domain.com.   CNAME   a05951ed7cdf-1394239323.us-east-1.elb.amazonaws.com.***
          1. Point your custom domain to the vfunction-nginx-xxx service's external IP location, which you can find in the OpenShift project > **Networking** -> **Services** -> **vfunction-nginx-xxx** service > **Service Address** > **Location** field.

          ## Upgrade and Rollback

          The vFunction operator includes a built-in auto-upgrade mechanism.

          You can choose one of 4 upgrading modes:
          - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically.
          - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
          - **Never**: The automatic upgrading mechanism is off (default).
          - **Maintenance**: Put the operator in a maintenance mode.

          For an on-demand upgrade,  you can change anytime the upgrading mode from “Never” to “Always”, wait for the operand to be upgraded, and change again to “Never”.

          During the upgrade, the "Operand State" property will change to “Upgrading...” and back again to “Working” after a successful upgrade.

          In case of a failure upgrading the operand, an automatic rollback to the last working version will occur. The failed version will be marked as defective and the operator will not try to upgrade to it again.

          Choosing the "Maintenance" option will tell the operator to take down all pods, so you can safely fix storage issues, return volumes from snapshots, etc.
          After finishing the maintenance, return the upgrade mode to your original desired policy.
        displayName: vFunction Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - vfunction
        - analyzing
        - services
        links:
        - name: vFunction
          url: https://vfunction.com
        - name: Terms
          url: https://www.vfunction.com/eula/
        maintainers:
        - email: raviv@vfunction.com
          name: Raviv Halivni
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: vFunction
        relatedImages:
        - registry.connect.redhat.com/vfunction/vfunction-server-operator@sha256:a5a0f47f7fb9219139f8d749da14950a8c39698b340644011485dfbec54ebe8b
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-users@sha256:c471d450d016aa0801ffb3de4648735245f2cf6901576106aed06cacc948d6f6
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-parser@sha256:b70fe4340d0de33c03a3b3458ea29e4cbff30cdd43f0e8d94265568e9a8d1ac7
        - registry.connect.redhat.com/vfunction/vfunction-storage@sha256:8c36770f1fa5034ba089ae81c36c587cb73a5176f63969023c8ab6f90d8d7abd
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-idp@sha256:17fe2e3ef706a4bbd8fdf0e0e08bd021c49104b331674aae7b8978b31730a03d
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-organizations@sha256:8dd4d043806da1e85a4565b2f14bb9e24671fc158f8c1c2adc3804fda5213571
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-dm@sha256:2f82294b15af873c012094a8984cb0c157bf322c28cbe58b38253adc917b6662
        - registry.connect.redhat.com/vfunction/vfunction-mysql@sha256:b4b6543eb02e396e1339877f61b74cc783977dd607cd3679edce8ac6b627e9e9
        - registry.connect.redhat.com/vfunction/vfunction-nginx@sha256:8370b871d05a24da9841a93fb255bf47945428b0a192b96b4eb95b9c0efb86a0
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-mcp@sha256:d50142c1b7f8a52944f9051a13bb821c0a790f137ebef09b0a84256bf7e5d794
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-measurements@sha256:81edae1c7344c13916530f20c8be09512ac8e20542770949783b0025c37079f8
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-otlp@sha256:80ca6d64b0d74f2d4f1b267002d5666f7c602956c656fd36a5c430f77eef0672
        version: 4.3.1762
      entries:
      - name: vfunction-server-operator.v4.3.1762
        version: 4.3.1762
      - name: vfunction-server-operator.v4.0.1576
        version: 4.0.1576
      name: stable
    defaultChannel: stable
    packageName: vfunction-server-operator
    provider:
      name: vFunction
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent
      provider-url: ""
    name: cilium
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cilium.v1.10.15-x084926b
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/hubble-ui@sha256:dec6e7e541cac91ab859245dd2a7fa02de5363b75caa3509a79cf76d4f622550
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/hubble-ui-backend@sha256:a07e13df3e9d7ec0dcc3b158f243cda5a08f6eca38a25a5aa3171f6b3ec491ed
        - quay.io/cilium/operator-generic@sha256:7f2f7b59b541e4f660d65ed0a21d25ab69e49022172529771dd1134315e95c82
        - quay.io/cilium/clustermesh-apiserver@sha256:e73151c3295589de00dccae75819aabdff8fd907260817020f9e8fe3738e78b2
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - registry.connect.redhat.com/isovalent/cilium-olm:970b1d74dbe2b87c28e8dc1c68936de33123a4df-v1.10.15@sha256:afa3ac140fae24b151ab6ff5a8cdef11566a58f7234d702a6a45a398822e6d34
        - quay.io/cilium/cilium@sha256:adbb60f9d19d641a5496fcfc4769ebb55c1d4cd7699fef752f7c94754194cf83
        - quay.io/cilium/hubble-relay@sha256:67a9183fd1febd9e51f4d02bd35092b81c8612a1716015988f69fffba79d999a
        version: 1.10.15+x084926b
      entries:
      - name: cilium.v1.10.15-x084926b
        version: 1.10.15+x084926b
      - name: cilium.v1.10.13-xa8bb1e5
        version: 1.10.13+xa8bb1e5
      - name: cilium.v1.10.12-x362ea11
        version: 1.10.12+x362ea11
      - name: cilium.v1.10.10-x6adc6ff
        version: 1.10.10+x6adc6ff
      name: "1.10"
    - currentCSV: cilium.v1.11.20-xdb5ed32
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/cilium/cilium@sha256:60df3cb7155886e0b62060c7a4a31e457933c6e35af79febad5fd6e43bab2a99
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:3677840d57e1d74e3ac14db0dfceb83088e6ebd70775d775855cdd3bfed68059
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/operator-generic@sha256:1439954acf620f048ef663524ae70b4a25693c58527a2f2cee51124496e29f90
        - quay.io/cilium/hubble-ui@sha256:bbf8b041152041d3ef000d6074a8dd5d40ed910b5c0067f365b00e52e449d467
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/clustermesh-apiserver@sha256:46760182f8c98227cfac27627275616987b71509227775350573d834133a6d49
        - quay.io/cilium/hubble-relay@sha256:e2f38b901fd8bd5adc9a765a5e68836364ebd1e7dfb85c2bcd8a5488b23c3470
        - quay.io/cilium/hubble-ui-backend@sha256:46070270b8ffe7ac3f1794721be0b5a09f906dff3492f8a199a5f995db362d2a
        version: 1.11.20+xdb5ed32
      entries:
      - name: cilium.v1.11.20-xdb5ed32
        version: 1.11.20+xdb5ed32
      name: "1.11"
    - currentCSV: cilium.v1.12.19-xef55de8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.12.0 <1.12.19+xef55de8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/hubble-ui-backend@sha256:f88a73a120521eeafbcdbd908f517117f6557bf61e115847853fac371f0d774c
        - quay.io/cilium/hubble-relay@sha256:19f2dee371523888985e6e34b30d07ef3f84f2a9206ad79bdfc8dede96e25d5c
        - quay.io/cilium/clustermesh-apiserver@sha256:61f8e871760dceaa0c0459bbf44a3049dd0f4ab15e57e063d9b3284434cb3dd8
        - quay.io/cilium/hubble-ui@sha256:e15af59a2ded739e420be82de6cbdd0ce22f8d3f00e3a10b3d2d2734e074a394
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/cilium/cilium@sha256:b997994ab759443d9fc9807eb39b4784bf03e13730a8132ffe5df3edef22a312
        - quay.io/cilium/operator-generic@sha256:da4d72facecde862c88d08bf99c3f476845beb407052a0cc4c06b68cc93b7570
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:d35994cf5cf319b88ff5a7985cc972edd23df5a282360f9ba5516d610390ebf8
        version: 1.12.19+xef55de8
      entries:
      - name: cilium.v1.12.19-xef55de8
        version: 1.12.19+xef55de8
      - name: cilium.v1.12.18-x63f4f4c
        version: 1.12.18+x63f4f4c
      - name: cilium.v1.12.17-x6cc0c18
        version: 1.12.17+x6cc0c18
      - name: cilium.v1.12.16-x14b1d1d
        version: 1.12.16+x14b1d1d
      - name: cilium.v1.12.15-x7d84617
        version: 1.12.15+x7d84617
      - name: cilium.v1.12.14-x17ed60a
        version: 1.12.14+x17ed60a
      - name: cilium.v1.12.13-xb414403
        version: 1.12.13+xb414403
      - name: cilium.v1.12.12-xd280cc3
        version: 1.12.12+xd280cc3
      - name: cilium.v1.12.11-xc8fb7eb
        version: 1.12.11+xc8fb7eb
      - name: cilium.v1.12.10-xe372776
        version: 1.12.10+xe372776
      - name: cilium.v1.12.9-x2ccb2b6
        version: 1.12.9+x2ccb2b6
      - name: cilium.v1.12.8-xa53bac1
        version: 1.12.8+xa53bac1
      - name: cilium.v1.12.7-xba75145
        version: 1.12.7+xba75145
      - name: cilium.v1.12.0-xb42b669
        version: 1.12.0+xb42b669
      name: "1.12"
    - currentCSV: cilium.v1.13.12-x6e56d47
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.13.12+x6e56d47'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:b0d22d4376b4c2b2b3b988955227f1ed2ad49a363c5367e53f64edae2cec0ff5
        - quay.io/cilium/cilium@sha256:d99204aa7b3b7bd2c9ab47fd398cc9f40290799bc0c7a4386c8dc5c1780cd3d3
        - quay.io/cilium/clustermesh-apiserver@sha256:f1b1d0a85bab65e7d6adc90d000513a56ac58bdb071aa391a8580d73d20e6b6a
        - quay.io/cilium/operator-generic@sha256:f83734bbe270f961d545c7929152785507ce04a05d818ebc9776941723736d02
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/hubble-relay@sha256:01b23ea40bcd81145dde6bfcbfc4d542749d08c2a1c6348954c85123a8d2b1fc
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        version: 1.13.12+x6e56d47
      entries:
      - name: cilium.v1.13.12-x6e56d47
        version: 1.13.12+x6e56d47
      - name: cilium.v1.13.11-xba1ab0b
        version: 1.13.11+xba1ab0b
      - name: cilium.v1.13.10-x412ae64
        version: 1.13.10+x412ae64
      - name: cilium.v1.13.9-x3ae7099
        version: 1.13.9+x3ae7099
      - name: cilium.v1.13.8-xe0355c7
        version: 1.13.8+xe0355c7
      - name: cilium.v1.13.7-x37f9269
        version: 1.13.7+x37f9269
      - name: cilium.v1.13.6-xdb5afcd
        version: 1.13.6+xdb5afcd
      - name: cilium.v1.13.5-xaf5caa6
        version: 1.13.5+xaf5caa6
      - name: cilium.v1.13.4-x129f2a7
        version: 1.13.4+x129f2a7
      - name: cilium.v1.13.3-x5854d64
        version: 1.13.3+x5854d64
      - name: cilium.v1.13.2-x82c79c4
        version: 1.13.2+x82c79c4
      - name: cilium.v1.13.1-x10a5de9
        version: 1.13.1+x10a5de9
      - name: cilium.v1.13.0-x32540df
        version: 1.13.0+x32540df
      name: "1.13"
    - currentCSV: cilium.v1.14.7-x3522df7
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.14.7+x3522df7'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:291a05c58de8d7daefeb0689c23cadd92c9368aea09751de70eaa817f7fcc358
        - quay.io/cilium/operator-generic@sha256:37ef0bd85c27c765c637cd58c3ff4a559f8734ae39f9d1839a3ac7803de7b952
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/cilium@sha256:45ce2b87696082ecf7d53ba1c64ceeb4217578033e5ef28ac479ec049a48bc32
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/hubble-relay@sha256:46762393daf4a0aaef76b106614c2615942f98f10aeacd435ea3fb1a0bdf69e4
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/clustermesh-apiserver@sha256:28f3ffe53365ca79831af600f09a95c0b3e9959f5f891b416dab8cedd90c263d
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        version: 1.14.7+x3522df7
      entries:
      - name: cilium.v1.14.7-x3522df7
        version: 1.14.7+x3522df7
      - name: cilium.v1.14.6-x11464e1
        version: 1.14.6+x11464e1
      - name: cilium.v1.14.5-x3583390
        version: 1.14.5+x3583390
      - name: cilium.v1.14.4-xb89471d
        version: 1.14.4+xb89471d
      - name: cilium.v1.14.3-x041dc10
        version: 1.14.3+x041dc10
      - name: cilium.v1.14.2-xedd5ce2
        version: 1.14.2+xedd5ce2
      - name: cilium.v1.14.1-xdd859b7
        version: 1.14.1+xdd859b7
      name: "1.14"
    - currentCSV: cilium.v1.15.1-x7095b76
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.15.1+x7095b76'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:9ab6be29447125e886300e9258b9a06bedf0a9d87405832aa8b6565ed1ba4215
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        - quay.io/cilium/cilium@sha256:351d6685dc6f6ffbcd5451043167cfa8842c6decf80d8c8e426a417c73fb56d4
        - quay.io/cilium/clustermesh-apiserver@sha256:b353badd255c2ce47eaa8f394ee4cbf70666773d7294bd887693e0c33503dc37
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/operator-generic@sha256:819c7281f5a4f25ee1ce2ec4c76b6fbc69a660c68b7825e9580b1813833fa743
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/hubble-relay@sha256:3254aaf85064bc1567e8ce01ad634b6dd269e91858c83be99e47e685d4bb8012
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        version: 1.15.1+x7095b76
      entries:
      - name: cilium.v1.15.1-x7095b76
        version: 1.15.1+x7095b76
      name: "1.15"
    - currentCSV: cilium.v1.9.14-x70dc065
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Basic Install
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/hubble-relay@sha256:fd6ab1aea260abc5f64eca26c1b1e7009983e4aaa8e5d098e8d442f7659603fb
        - quay.io/cilium/clustermesh-apiserver@sha256:a0da5edf0372899647da51de1b277f0bab8e676d694aee7f939cddfdd3172010
        - quay.io/cilium/hubble-ui-backend@sha256:2cf1c7b1fde6161f575f3147c2ef885739e60b8c59a1ed444fe74edbdf22c7c1
        - quay.io/cilium/cilium@sha256:2c6ce93fa7e625979043a387eb998c17ad57df8768d89facb9b715da42a4c51c
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:3a9ca9a4c2f78109f265c4758f5f693d01865f2d0e31bc5f1ecbc065e878628b
        - docker.io/envoyproxy/envoy@sha256:55a141c0b8e39497b32074cb758b4e5dc2384bd6d9926144969a4452eaa611fd
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/hubble-ui@sha256:c267219946b84f79a5fd2849f5cc71f39453e29d93039e57b9cef2b5802e9c50
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/cilium/operator-generic@sha256:bdcfd2eade99933f2fda55ef79ea697ddfad3512f65b15bcd0ba7702518c1ba3
        version: 1.9.14+x70dc065
      entries:
      - name: cilium.v1.9.14-x70dc065
        version: 1.9.14+x70dc065
      name: "1.9"
    - currentCSV: cilium.v1.10.4-x5bfd7b3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Basic Install
          categories: Networking,Security
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:b45ff6a6552d71cbc26e358e0b25759d33818fd92301f1656453c5bcf826ff13
        version: 1.10.4+x5bfd7b3
      entries:
      - name: cilium.v1.10.4-x5bfd7b3
        version: 1.10.4+x5bfd7b3
      name: stable
    defaultChannel: "1.15"
    packageName: cilium
    provider:
      name: Isovalent
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: metallb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: metallb-operator.v4.16.0-202509230119
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "basic",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64512,
                        "neighbors": [
                          {
                            "address": "172.30.0.3",
                            "asn": 64512,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            },
                            "toReceive": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "two-neighs",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64512,
                        "neighbors": [
                          {
                            "address": "172.30.0.3",
                            "asn": 4200000000,
                            "ebgpMultiHop": true,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          },
                          {
                            "address": "172.18.0.6",
                            "asn": 4200000000,
                            "port": 179,
                            "toAdvertise": {
                              "allowed": {
                                "prefixes": [
                                  "192.168.2.0/24"
                                ]
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "two-routers-vrf",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64513,
                        "neighbors": [
                          {
                            "address": "172.32.0.3",
                            "asn": 4200000000,
                            "ebgpMultiHop": true,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ]
                      },
                      {
                        "asn": 64515,
                        "neighbors": [
                          {
                            "address": "172.31.0.3",
                            "asn": 4200000000,
                            "port": 179,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ],
                        "vrf": "red"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRNodeState",
                "metadata": {
                  "name": "frr-k8s-worker"
                },
                "status": {
                  "lastConversionResult": "success",
                  "lastReloadResult": "success",
                  "runningConfig": "Building configuration...\n\nCurrent configuration:\n!\nfrr version 9.0.2_git\nfrr defaults traditional\nhostname frr-k8s-worker\nlog file /etc/frr/frr.log informational\nlog timestamp precision 3\nno ipv6 forwarding\nservice integrated-vtysh-config\n!\nip nht resolve-via-default\n!\nipv6 nht resolve-via-default\n!\nend\n"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BFDProfile",
                "metadata": {
                  "name": "bfd-profile-sample",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "receiveInterval": 380,
                  "transmitInterval": 270
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "peers": [
                    "ebgp-single-hop0"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPoolSelectors": [
                    {
                      "matchLabels": {
                        "test": "ipv6"
                      }
                    },
                    {
                      "matchLabels": {
                        "test": "ipv4"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "communities": [
                    "65535:65282"
                  ],
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "localPref": 50
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample4",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "communities": [
                    "NO_ADVERTISE"
                  ],
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "localPref": 50
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample5",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "nodeSelectors": [
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-control-plane"
                      }
                    },
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-worker"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-beta1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "Community",
                "metadata": {
                  "name": "community1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "communities": [
                    {
                      "name": "NO_ADVERTISE",
                      "value": "65535:65282"
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "name": "ip-addresspool-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "192.168.10.0/24",
                    "192.168.9.1-192.168.9.5",
                    "fc00:f853:0ccd:e799::/124"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "labels": {
                    "test": "ipv4"
                  },
                  "name": "ip-addresspool-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "172.20.0.100/24"
                  ],
                  "autoAssign": false
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "labels": {
                    "test": "ipv6"
                  },
                  "name": "ip-addresspool-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "2002:2:2::1-2002:2:2::100"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "L2Advertisement",
                "metadata": {
                  "name": "l2-adv-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "L2Advertisement",
                "metadata": {
                  "name": "l2-adv-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "nodeSelectors": [
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-control-plane"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "MetalLB",
                "metadata": {
                  "name": "metallb",
                  "namespace": "metallb-system"
                },
                "spec": {}
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "bfdProfile": "bfd-profile-sample",
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "holdTime": "9m0s",
                  "keepaliveTime": "0s",
                  "myASN": 64512,
                  "passwordSecret": {
                    "name": "secretname",
                    "namespace": "metallb-system"
                  },
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3",
                  "peerPort": 180
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:b07c046ba2ae71b93f27d5cdca7cb93f51f3198cb84bb6e65dd6320bcda268c8
          createdAt: "2023-06-06T15:25:00Z"
          description: An operator for deploying MetalLB on a kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.8.0 <4.16.0'
          operatorframework.io/suggested-namespace: metallb-system
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/metallb-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BFDProfile is the Schema for the bfdprofiles API
            kind: BFDProfile
            name: bfdprofiles.metallb.io
            version: v1beta1
          - description: BGPAdvertisement is the Schema for the bgpadvertisements
              API
            kind: BGPAdvertisement
            name: bgpadvertisements.metallb.io
            version: v1beta1
          - description: BGPPeer is the Schema for the bgppeers API
            kind: BGPPeer
            name: bgppeers.metallb.io
            version: v1beta1
          - description: BGPPeer is the Schema for the bgppeers API
            kind: BGPPeer
            name: bgppeers.metallb.io
            version: v1beta2
          - description: Community is the Schema for the communities API
            kind: Community
            name: communities.metallb.io
            version: v1beta1
          - description: FRRConfiguration is the Schema for the frrconfigurations
              API
            kind: FRRConfiguration
            name: frrconfigurations.frrk8s.metallb.io
            version: v1beta1
          - description: FRRNodeState is the Schema for the frrnodestates API
            kind: FRRNodeState
            name: frrnodestates.frrk8s.metallb.io
            version: v1beta1
          - description: IPAddressPool is the Schema for the ipaddresspools API
            kind: IPAddressPool
            name: ipaddresspools.metallb.io
            version: v1beta1
          - description: L2Advertisement is the Schema for the l2advertisements API
            kind: L2Advertisement
            name: l2advertisements.metallb.io
            version: v1beta1
          - description: MetalLB is the Schema for the metallbs API
            displayName: MetalLB
            kind: MetalLB
            name: metallbs.metallb.io
            version: v1beta1
        description: An operator for deploying MetalLB on a kubernetes cluster.
        displayName: MetalLB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - metallb-operator
        links:
        - name: MetalLB Operator
          url: https://github.com/openshift/metallb-operator
        maintainers:
        - email: fpaoline@redhat.com
          name: Federico Paolinelli
        - email: obraunsh@redhat.com
          name: Ori Braunshtein
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/metallb-rhel9@sha256:95abda97d689078bd124ef59c92cc12e661222cac8e8b0e526579470ab54d05e
        - registry.redhat.io/openshift4/frr-rhel9@sha256:48a361644b2fa8fc632f49117172db2e9c12022774e7e0fd409561d884a6c237
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:b07c046ba2ae71b93f27d5cdca7cb93f51f3198cb84bb6e65dd6320bcda268c8
        version: 4.16.0-202509230119
      entries:
      - name: metallb-operator.v4.16.0-202509230119
        version: 4.16.0-202509230119
      - name: metallb-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: metallb-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: metallb-operator.v4.16.0-202507221936
        version: 4.16.0-202507221936
      - name: metallb-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: metallb-operator.v4.16.0-202506090335
        version: 4.16.0-202506090335
      - name: metallb-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: metallb-operator.v4.16.0-202505051339
        version: 4.16.0-202505051339
      - name: metallb-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: metallb-operator.v4.16.0-202503252033
        version: 4.16.0-202503252033
      - name: metallb-operator.v4.16.0-202502260004
        version: 4.16.0-202502260004
      - name: metallb-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: metallb-operator.v4.16.0-202501232204
        version: 4.16.0-202501232204
      - name: metallb-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: metallb-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: metallb-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: metallb-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: metallb-operator.v4.16.0-202411110236
        version: 4.16.0-202411110236
      - name: metallb-operator.v4.16.0-202410292005
        version: 4.16.0-202410292005
      - name: metallb-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: metallb-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: metallb-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: metallb-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: metallb-operator.v4.16.0-202409161407
        version: 4.16.0-202409161407
      - name: metallb-operator.v4.16.0-202409062036
        version: 4.16.0-202409062036
      - name: metallb-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: metallb-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: metallb-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: metallb-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: metallb-operator.v4.16.0-202408051837
        version: 4.16.0-202408051837
      - name: metallb-operator.v4.16.0-202407221107
        version: 4.16.0-202407221107
      - name: metallb-operator.v4.16.0-202407180936
        version: 4.16.0-202407180936
      - name: metallb-operator.v4.16.0-202407091337
        version: 4.16.0-202407091337
      - name: metallb-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: metallb-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: metallb-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: metallb-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: metallb-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app: enmasse
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq-online
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-online.1.7.4-0.1657725339.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "StandardInfraConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "broker": {
                    "resources": {
                      "memory": "1Gi",
                      "storage": "5Gi"
                    },
                    "addressFullPolicy": "FAIL"
                  },
                  "router": {
                    "linkCapacity": 50,
                    "resources": {
                      "memory": "512Mi"
                    }
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "BrokeredInfraConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "broker": {
                    "resources": {
                      "memory": "4Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta2",
                "kind": "AddressPlan",
                "metadata": {
                  "name": "standard-small-queue"
                },
                "spec": {
                  "addressType": "queue",
                  "shortDescription": "Small Queue",
                  "resources": {
                    "router": 0.01,
                    "broker": 0.1
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta2",
                "kind": "AddressSpacePlan",
                "metadata": {
                  "name": "standard-small"
                },
                "spec": {
                  "addressSpaceType": "standard",
                  "infraConfigRef": "default",
                  "shortDescription": "Small Address Space Plan",
                  "resourceLimits": {
                    "router": 1.0,
                    "broker": 2.0,
                    "aggregate": 2.0
                  },
                  "addressPlans": [
                    "standard-small-queue"
                  ]
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "AuthenticationService",
                "metadata": {
                  "name": "standard-authservice"
                },
                "spec": {
                  "type": "standard",
                  "standard": {
                    "storage": {
                      "claimName": "standard-authservice",
                      "deleteClaim": true,
                      "size": "1Gi",
                      "type": "persistent-claim"
                    }
                  }
                }
              },
              {
                "apiVersion": "enmasse.io/v1beta1",
                "kind": "AddressSpace",
                "metadata": {
                  "name": "myspace"
                },
                "spec": {
                  "plan": "standard-small",
                  "type": "standard"
                }
              },
              {
                "apiVersion": "enmasse.io/v1beta1",
                "kind": "Address",
                "metadata": {
                  "name": "myspace.myqueue"
                },
                "spec": {
                  "address": "myqueue",
                  "plan": "standard-small-queue",
                  "type": "queue"
                }
              },
              {
                "apiVersion": "user.enmasse.io/v1beta1",
                "kind": "MessagingUser",
                "metadata": {
                  "name": "myspace.user"
                },
                "spec": {
                  "authentication": {
                    "password": "ZW5tYXNzZQ==",
                    "type": "password"
                  },
                  "authorization": [
                    {
                      "addresses": [
                        "myqueue"
                      ],
                      "operations": [
                        "send",
                        "recv"
                      ]
                    }
                  ],
                  "username": "user"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator@sha256:2aaf87224a96aa56dc595e96d8e7ec8d138aaf20240afa25a1e805b405941355
          createdAt: "2019-02-19T00:00:00Z"
          description: Red Hat Integration - AMQ Online provides messaging as a managed
            service on OpenShift
          olm.substitutesFor: amq-online.1.7.4
          operators.openshift.io/capability: '["license-required"]'
          operators.openshift.io/valid-licenses: '["Red Hat AMQ", "Red Hat Integration"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: ""
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A messaging user that can connect to an Address Space
            displayName: Messaging User
            kind: MessagingUser
            name: messagingusers.user.enmasse.io
            version: v1beta1
          - description: A resource representing the available schema of plans and
              authentication services
            displayName: AddressSpaceSchema
            kind: AddressSpaceSchema
            name: addressspaceschemas.enmasse.io
            version: v1beta1
          - description: A group of messaging addresses that can be accessed via the
              same endpoint
            displayName: Address Space
            kind: AddressSpace
            name: addressspaces.enmasse.io
            version: v1beta1
          - description: A messaging address that can be used to send/receive messages
              to/from
            displayName: Address
            kind: Address
            name: addresses.enmasse.io
            version: v1beta1
          - description: Infrastructure configuration template for the standard address
              space type
            displayName: Standard Infra Config
            kind: StandardInfraConfig
            name: standardinfraconfigs.admin.enmasse.io
            version: v1beta1
          - description: Infrastructure configuration template for the brokered address
              space type
            displayName: Brokered Infra Config
            kind: BrokeredInfraConfig
            name: brokeredinfraconfigs.admin.enmasse.io
            version: v1beta1
          - description: Plan describing the resource usage of a given address type
            displayName: Address Plan
            kind: AddressPlan
            name: addressplans.admin.enmasse.io
            version: v1beta2
          - description: Plan describing the capabilities and resource limits of a
              given address space type
            displayName: Address Space Plan
            kind: AddressSpacePlan
            name: addressspaceplans.admin.enmasse.io
            version: v1beta2
          - description: Authentication service configuration available to address
              spaces.
            displayName: Authentication Service
            kind: AuthenticationService
            name: authenticationservices.admin.enmasse.io
            version: v1beta1
          - description: Console Service Singleton for deploying global console.
            displayName: Console Service
            kind: ConsoleService
            name: consoleservices.admin.enmasse.io
            version: v1beta1
        description: "**Red Hat Integration - AMQ Online** provides messaging as a
          managed service on OpenShift.\nWith Red Hat Integration - AMQ Online, administrators
          can configure a cloud-native,\nmulti-tenant messaging service either in
          the cloud or on premise.\nDevelopers can provision messaging using the Red
          Hat Integration - AMQ Online Console.\nMultiple development teams can provision
          the brokers and queues from the\nconsole, without requiring each team to
          install, configure, deploy,\nmaintain, or patch any software. \n\n**The
          core capabilities include:**\n\n  * **Built-in authentication and authorization**
          - Use the built-in authentication service or\n    plug in your own authentication
          service for authentication and\n    authorization of messaging clients.\n\n
          \ * **Self-service messaging for applications** - The service administrator
          deploys\n    and manages the messaging infrastructure, while applications
          can request\n    messaging resources regardless of the messaging infrastructure.\n\n
          \ * **Support for a wide variety of messaging patterns** - Choose between\n
          \   JMS-style messaging with strict guarantees, or messaging that supports\n
          \   a larger number of connections and higher throughput.\n\n## Post-installation
          tasks\n\nTo fully use Red Hat Integration - AMQ Online, you need to create
          a few\ninfrastructure components after the installation, including:\n\n
          \ * An authentication service\n  * Infrastructure configuration for the
          standard and brokered address space\n  * Address and address space plans\n
          \ * (Optional) Create RBAC roles to allow users to discover available plans\n
          \ * (Optional) Create RBAC roles to allow users to self-manage addresses
          and\n    address spaces.\n\nFor a complete overview of how to configure
          Red Hat Integration - AMQ Online, see\n[Configuring Red Hat Integration
          - AMQ Online](https://access.redhat.com/documentation/en-us/red_hat_amq/7.7/html/installing_and_managing_amq_online_on_openshift/configuring-messaging).\n\n###
          Quickstart infrastructure configuration\n\nIf you simply want to get started
          quickly, you can import the following\nYAML by saving the content to a file
          and apply it by running the \n`oc apply -f <file>` command. You can also
          split the content (at the `---` marker)\nand import the single YAML document
          using the Web UI: \n\n~~~yaml\n---\napiVersion: admin.enmasse.io/v1beta1\nkind:
          StandardInfraConfig\nmetadata:\n  name: default\n---\napiVersion: admin.enmasse.io/v1beta2\nkind:
          AddressPlan\nmetadata:\n  name: standard-small-queue\nspec:\n  addressType:
          queue\n  resources:\n    router: 0.01\n    broker: 0.1\n---\napiVersion:
          admin.enmasse.io/v1beta2\nkind: AddressSpacePlan\nmetadata:\n  name: standard-small\nspec:\n
          \ addressSpaceType: standard\n  infraConfigRef: default\n  addressPlans:\n
          \ - standard-small-queue\n  resourceLimits:\n    router: 2.0\n    broker:
          3.0\n    aggregate: 4.0\n---\napiVersion: admin.enmasse.io/v1beta1\nkind:
          AuthenticationService\nmetadata:\n  name: standard-authservice\nspec:\n
          \ type: standard\n  standard:\n    storage:\n      claimName: standard-authservice\n
          \     deleteClaim: true\n      size: 1Gi\n      type: persistent-claim\n~~~\n\n###
          Create RBAC roles to allow users to discover available plans\n\nFor users
          to discover the available plans, cluster-wide roles to read the available\nschema
          can be created.  Import the following YAML by saving the content to a file
          and apply it by running the \n`oc apply -f <file>` command. You can also
          split the content (at the `---` marker)\nand import the single YAML document
          using the Web UI: \n\n~~~yaml\n---\napiVersion: rbac.authorization.k8s.io/v1\nkind:
          ClusterRole\nmetadata:\n  name: enmasse.io:schema\n  labels:\n    app: enmasse\nrules:\n
          \ - apiGroups: [ \"enmasse.io\" ]\n    resources: [ \"addressspaceschemas\"
          ]\n    verbs: [ \"get\", \"list\", \"watch\" ]\n---\napiVersion: rbac.authorization.k8s.io/v1\nkind:
          ClusterRoleBinding\nmetadata:\n  name: \"enmasse.io:schema\"\n  labels:\n
          \   app: enmasse\nroleRef:\n  apiGroup: rbac.authorization.k8s.io\n  kind:
          ClusterRole\n  name: enmasse.io:schema\nsubjects:\n- kind: Group\n  name:
          system:authenticated\n~~~\n\n### Creating infrastructure using the Web UI\n\nYou
          must create a new instance of each of the following custom resources. You
          can\nuse the example data directly, which is provided when using the\nWeb
          UI:\n\n  * *Authentication Service* – Create an authentication service.\n
          \ * *Brokered Infra Config* – Create the broker infrastructure configuration.\n
          \ * *Standard Infra Config* – Create the standard infrastructure\n    configuration.\n\nYou
          must also create at least one address space plan and one address plan.\n\n*Note*:
          The name of the address space plan and address plan is required\nlater when
          creating new addresses. Some examples use specific plan\nnames, which might
          not be available in your environment. You can\ncreate those plans, or edit
          the examples to use different plan names.\n\n## Configuration for messaging
          tenants\n\nWhile service administrators perform the infrastructure configuration,
          the following\nresources are for the actual users of the system, the messaging
          tenants.\n\nYou need to create those resources to satisfy your particular
          use case.\n\n  * *Address space* – A container for addresses\n  * *Address*
          – A messaging address (address, topic, queue, and so on)\n  * *Messaging
          user* – Manages access to an address\n"
        displayName: Red Hat Integration - AMQ Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amq
        - messaging
        - amqp
        - iot
        - mqtt
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/7.7/html/using_amq_online_on_openshift/index
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-online-1-topic-forwarder@sha256:de33a701339d91a4e9275f7cf65f917ff75368a59b8ab36dd66c90bc306875aa
        - registry.redhat.io/amq7/amq-online-1-broker-plugin@sha256:2466a336c0d78b196635de3fad636aeff2d9265839f7271b8ef94f3ba97d9bc0
        - registry.redhat.io/amq7/amq-interconnect@sha256:5ebd4c4a0ed0db2725ae466550c915531b32933c420344cc2083c616da3f326f
        - registry.redhat.io/redhat-sso-7/sso73-openshift@sha256:18d15980f28f0d0c6b8582131a1f062d84463577512875b5391a612c3d7d1096
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6d489a7897bd831ec273034f1ff0e223427a40afd01b9835ebb020a8bf9ddc66
        - registry.redhat.io/amq7/amq-broker@sha256:1ec0437bbd82722148fdbb02dc55682b114c311cd6f521a8bc621a8c1cb4b32b
        - registry.redhat.io/amq7/amq-online-1-address-space-controller@sha256:9acc555c8e869aec7d72f53b60320e1a96cc82d748638e0d50970837229fdbee
        - registry.redhat.io/amq7/amq-online-1-none-auth-service@sha256:a013fb1806c002b408a7753ff943cee24fbcdf8a657eff1e9582c42b5a7b0844
        - registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator@sha256:2aaf87224a96aa56dc595e96d8e7ec8d138aaf20240afa25a1e805b405941355
        - registry.redhat.io/amq7/amq-online-1-standard-controller@sha256:070349af30cec52fa8efc185fc940077a64bb40e3e36310d7cc6e39ce0971436
        - registry.redhat.io/amq7/amq-online-1-console-init@sha256:0970804909b82fdc3d4e7d5d88280e41157d743107de8bfa2d1baa1367543176
        - registry.redhat.io/openshift3/oauth-proxy@sha256:59b20f23c03d0012ec5c95c4725892ee64277c22dedc9d0a90a7b6b0bc80fc72
        - registry.redhat.io/amq7/amq-online-1-console-server-rhel7@sha256:1dcb3d0cc7bc6e84f4f96962ff88849511b7a7686a1f5dbd9300c4c2c8c05e2f
        - registry.redhat.io/amq7/amq-online-1-agent@sha256:f98ed8ee97ac59d8fdd40ec8b2c063a5afd360a3e183057c186e7a77a32801ee
        - registry.redhat.io/amq7/amq-online-1-auth-plugin@sha256:68926f434b2f6ce4d9698938bc3c53d514c21c3f76c73b6b4cdf7fcf51f7cb0d
        version: 1.7.4+0.1657725339.p
      entries:
      - name: amq-online.1.7.4
        version: 1.7.4
      - name: amq-online.1.7.4-0.1657725339.p
        version: 1.7.4+0.1657725339.p
      - name: amq-online.1.7.3
        version: 1.7.3
      - name: amq-online.1.7.2
        version: 1.7.2
      - name: amq-online.1.7.1
        version: 1.7.1
      - name: amq-online.1.7.1-0.1628610187.p
        version: 1.7.1+0.1628610187.p
      - name: amq-online.1.7.0
        version: 1.7.0
      name: stable
    defaultChannel: stable
    packageName: amq-online
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: sandboxed-containers-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sandboxed-containers-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:051bd7f1dad8cc3251430fee32184be8d64077aba78580184cef0255d267bdcf
          createdAt: 2021-06-07-13:38:06
          olm.skipRange: <1.0.2
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          OpenShift sandboxed containers based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift’s Defense-in-Depth strategy. For more information
          [see](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5).

          # Features & benefits
          - **Isolated Developer Environments & Priviliges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as CAP_ADMIN or CAP_BPF. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select preview-1.0 from the list of available Update Channel options.
            This ensures that you install the version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              openshift-sandboxed-containers-operator namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              openshift-sandboxed-containers-operator causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a KataConfig CRD instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.8/sandboxed_containers/understanding-sandboxed-containers.html).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kernel-isolated containers
        links:
        - name: OpenShift sandboxed containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        provider:
          name: Red Hat
          url: https://www.github.com/openshift/sandboxed-containers-operator
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:051bd7f1dad8cc3251430fee32184be8d64077aba78580184cef0255d267bdcf
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:c6c589d5e47ba9564c66c84fc2bc7e5e046dae1d56a3dc99d7343f01e42e4d31
        version: 1.0.2
      entries:
      - name: sandboxed-containers-operator.v1.0.2
        version: 1.0.2
      name: preview-1.0
    - currentCSV: sandboxed-containers-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:b1e824e126c579db0f56d04c3d1796d82ed033110c6bc923de66d95b67099611
          createdAt: "2021-10-08T13:20:20.250789Z"
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift’s Defense-in-Depth strategy. For more information
          [see](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5).

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `preview-1.1` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CRD instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.8/sandboxed_containers/understanding-sandboxed-containers.html).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kernel-isolated containers
        links:
        - name: OpenShift sandboxed containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        provider:
          name: Red Hat
          url: https://www.github.com/openshift/sandboxed-containers-operator
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:b1e824e126c579db0f56d04c3d1796d82ed033110c6bc923de66d95b67099611
        version: 1.1.0
      entries:
      - name: sandboxed-containers-operator.v1.1.0
        version: 1.1.0
      - name: sandboxed-containers-operator.v1.0.2
        version: 1.0.2
      name: preview-1.1
    - currentCSV: sandboxed-containers-operator.v1.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-12T14:30:28Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.10.3'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '["peerpods.confidentialcontainers.org","peerpodconfigs.confidentialcontainers.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/sandboxed-containers-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
          - kind: PeerPod
            name: peerpods.confidentialcontainers.org
            version: v1alpha1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `candidate` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the [OpenShift sandboxed containers documentation](https://docs.redhat.com/en/documentation/openshift_sandboxed_containers/).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-cloud-api-adaptor-rhel9@sha256:7d1fa1e85af0d34c23e75868b1388eff17abe633c04a8dc55fbefcc470195c05
        - registry.redhat.io/openshift-sandboxed-containers/osc-cloud-api-adaptor-webhook-rhel9@sha256:6802f00a939f24ecd6dad150433db5ea3fd72d171f0c410cb894d3af910fc7fd
        - registry.redhat.io/openshift-sandboxed-containers/osc-podvm-builder-rhel9@sha256:c15f36a639fa07fee11f74c6563696ec8ef4da3fb1a99b7bd3f308dd04881339
        - registry.redhat.io/openshift-sandboxed-containers/osc-podvm-payload-rhel9@sha256:8d6b1262c6cedd04e33441264539842d82ed057f64eda5c4bce59566bcb72558
        - registry.redhat.io/openshift-sandboxed-containers/osc-dm-verity-image@sha256:e633bc27853f811d919e64461d0c917e659f5430458abe2304a4c85b0d871c80
        - registry.redhat.io/openshift-sandboxed-containers/osc-monitor-rhel9@sha256:037cbc221ae8fda83cdf1093127795ea806ecda53893deb49cdf981e7863774c
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel9-operator@sha256:1ffb432d9a8833854fa5e5262dcbf42530dfbaadfddf6978f7bc7804b84d2e3f
        version: 1.10.3
      entries:
      - name: sandboxed-containers-operator.v1.10.3
        version: 1.10.3
      - name: sandboxed-containers-operator.v1.10.2
        version: 1.10.2
      - name: sandboxed-containers-operator.v1.10.1
        version: 1.10.1
      - name: sandboxed-containers-operator.v1.10.0
        version: 1.10.0
      - name: sandboxed-containers-operator.v1.9.0
        version: 1.9.0
      - name: sandboxed-containers-operator.v1.8.1
        version: 1.8.1
      - name: sandboxed-containers-operator.v1.7.0
        version: 1.7.0
      - name: sandboxed-containers-operator.v1.6.0
        version: 1.6.0
      - name: sandboxed-containers-operator.v1.5.3
        version: 1.5.3
      - name: sandboxed-containers-operator.v1.5.2
        version: 1.5.2
      - name: sandboxed-containers-operator.v1.5.1
        version: 1.5.1
      - name: sandboxed-containers-operator.v1.5.0
        version: 1.5.0
      name: stable
    - currentCSV: sandboxed-containers-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:8f30a9129d817c3f4e404d2c43fb47e196d8c8da3badba4c48f65d440a4d7584
          createdAt: "2021-11-22T14:44:50.817336Z"
          olm.skipRange: '>=1.1.0 <1.2.2'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `stable-1.2` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.9/sandboxed_containers/understanding-sandboxed-containers.html)
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:a660f0b54b9139bed9a3aeef3408001c0d50ba60648364a98a09059b466fbcc1
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:8f30a9129d817c3f4e404d2c43fb47e196d8c8da3badba4c48f65d440a4d7584
        version: 1.2.2
      entries:
      - name: sandboxed-containers-operator.v1.2.2
        version: 1.2.2
      name: stable-1.2
    - currentCSV: sandboxed-containers-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          olm.skipRange: '>=1.1.0 <1.3.3'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/sandboxed-containers-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `stable-1.3` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/latest/sandboxed_containers/understanding-sandboxed-containers.html)
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:5d2b03721043e5221dfb0cf164cf59eba396ba3aae40a56c53aa3496c625eea0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:cb836456974e510eb4bccbffadbc6d99d5f57c36caec54c767a158ffd8a025d5
        version: 1.3.3
      entries:
      - name: sandboxed-containers-operator.v1.3.3
        version: 1.3.3
      name: stable-1.3
    defaultChannel: stable
    packageName: sandboxed-containers-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/stolostron/multicluster-global-hub
    name: multicluster-global-hub-operator-rh
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-global-hub-operator-rh.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:01e8dc62d5dc48af84b92f5315fab22574f648f2f705ac94f0516f4a7dab9237
          createdAt: 27 Aug 2024, 22:21
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.1.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```


          ### Documentation

          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: release-1.1
        provider:
          name: Red Hat
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:01e8dc62d5dc48af84b92f5315fab22574f648f2f705ac94f0516f4a7dab9237
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:78628c4f7a0286417023410261cc3b36797b7cdf72b38d0f4a2533a0f3db89dc
        - registry.redhat.io/rhel9/postgresql-13@sha256:c67d00643295697d196dd70dea457948ebb9e9d9d61a42019490b350c8286b50
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:5c8cea421be39712503378e32c54c332a7a0bf77a1705781d10fc123224bd60e
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:4ea8d3a5317d3b9968fa93dd68416e07e12e00e492860ff20ed90f88ca6ba225
        version: 1.1.2
      entries:
      - name: multicluster-global-hub-operator-rh.v1.1.2
        version: 1.1.2
      - name: multicluster-global-hub-operator-rh.v1.1.1
        version: 1.1.1
      - name: multicluster-global-hub-operator-rh.v1.1.0
        version: 1.1.0
      name: release-1.1
    - currentCSV: multicluster-global-hub-operator-rh.v1.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:98d470c8683c346107ab943776991554e3173c9a6431202bdacc48a75c4cf18f
          createdAt: 15 May 2025, 21:22
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          kafka-zookeeper-0                                       1/1     Running   0          21m
          kafka-zookeeper-1                                       1/1     Running   0          21m
          kafka-zookeeper-2                                       1/1     Running   0          21m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgres-0                      2/2     Running   0          22m
          ```


          ### Documentation

          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: release-1.2
        provider:
          name: Red Hat
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:c83a7b01bfe5a39d6cdf585180a011448e935d109768cb51560308b29d9ffe14
        - registry.redhat.io/rhel9/postgresql-13@sha256:b1af637cbe0e4041cd2fe6fdbcf665c5183e2991d3cc3e5b7384235f801504d9
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:16bb76ad0c68c0d6b3149b2741126a96e154150c1059621b14c2e6c30a214f16
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:98d470c8683c346107ab943776991554e3173c9a6431202bdacc48a75c4cf18f
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:92f44cff160874efe57a985609e599b13c49c0c4cef846dccc80833af2ae74f3
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:64e5ade4fbcd4cb70d62a6a62d775e051ff49838b891b682574646d4e61d9751
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:9b98fc01b08c3262e5ab99096e74fb55982f2a9c0de61e8fe84e419522e725b0
        version: 1.2.3
      entries:
      - name: multicluster-global-hub-operator-rh.v1.2.3
        version: 1.2.3
      - name: multicluster-global-hub-operator-rh.v1.2.2
        version: 1.2.2
      - name: multicluster-global-hub-operator-rh.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator-rh.v1.2.0
        version: 1.2.0
      name: release-1.2
    - currentCSV: multicluster-global-hub-operator-rh.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:0ef24ded54588aa49e366ba1cafc0971162e9d32024045bbcd7ef16e9f38a2d5
          createdAt: 01 Apr 2025, 20:25
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          kafka-zookeeper-0                                       1/1     Running   0          21m
          kafka-zookeeper-1                                       1/1     Running   0          21m
          kafka-zookeeper-2                                       1/1     Running   0          21m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgres-0                      2/2     Running   0          22m
          ```


          ### Documentation

          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: release-1.3
        provider:
          name: Red Hat
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:a139da58c681de987d847b233dd5c1e6ced5d6e86f3aab0f1b4181a77fc583d4
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:a17e90abc5e23bbc3908613c56c4208ad554eb27ed9679c4ccd06c4a1e9fcbcc
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:0ef24ded54588aa49e366ba1cafc0971162e9d32024045bbcd7ef16e9f38a2d5
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:619d8cb6b3811169af183bd8bc26ed98cdb25486ec428d88b25fee0ef9e6c552
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:e848a90d17e4f00d8021cbde47e488b82c447d8a5f6588f85f3d7cfbd6877f0f
        - registry.redhat.io/rhel9/postgresql-13@sha256:76c49b7dbd7c4ed8b430c4931b32f546db6d07ddc2146232f5d17eb17d3243b4
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-kessel-inventory-api-rhel9@sha256:dbd3924eb24db7ef28301b17851b5b5cd20e044cf259f0c556c67a0721561803
        version: 1.3.3
      entries:
      - name: multicluster-global-hub-operator-rh.v1.3.3
        version: 1.3.3
      - name: multicluster-global-hub-operator-rh.v1.3.2
        version: 1.3.2
      - name: multicluster-global-hub-operator-rh.v1.3.1
        version: 1.3.1
      - name: multicluster-global-hub-operator-rh.v1.3.0
        version: 1.3.0
      name: release-1.3
    - currentCSV: multicluster-global-hub-operator-rh.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:c5b3b3907e8135d3eedba40e27ba18a232ea4f30854d652059fcb14f82a8bf15
          createdAt: "2025-07-23T09:04:08Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: release-1.4
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:047b6c378e2b37339dd498c9025060f20812014bde4d0c98183f1a1201c21e66
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:ea906dcf8d45956251df22e269fe0b8fa8d49734e96a61b10db7b25d626b2e6a
        - registry.redhat.io/rhel9/postgresql-16@sha256:2161182c88022f38d6d9b8b467c0d59156edcededc908d383b80caa4eedf57f9
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:2be8bc79667018386235823e322ae10788e4e7e889c10e63566ae27a65dcabf1
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:0f4683d18404b4a9412e42ed417f6d3ac0024b3c57aa72c9162ff72d6ea8d394
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:c5b3b3907e8135d3eedba40e27ba18a232ea4f30854d652059fcb14f82a8bf15
        version: 1.4.2
      entries:
      - name: multicluster-global-hub-operator-rh.v1.4.2
        version: 1.4.2
      - name: multicluster-global-hub-operator-rh.v1.4.1
        version: 1.4.1
      - name: multicluster-global-hub-operator-rh.v1.4.0
        version: 1.4.0
      name: release-1.4
    - currentCSV: multicluster-global-hub-operator-rh.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:8af7e57b8c6285ad232084e7776fef32d026e080eb0bf14b91986ac28d00bb6a
          createdAt: "2025-08-18T08:30:32Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: release-1.5
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:8af7e57b8c6285ad232084e7776fef32d026e080eb0bf14b91986ac28d00bb6a
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:c95052569312d805b6750477c8bc57f7c72e1efd26143f3899ea0eddc2aed37b
        - registry.redhat.io/rhel9/postgresql-16@sha256:2161182c88022f38d6d9b8b467c0d59156edcededc908d383b80caa4eedf57f9
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:921163d139bca9e47bf95df43d7b408ff6fe2bf9b0902fb6af9f4cfc9d4598cc
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:ab3ed25a4448a7d545648ae4cea52b493918cc4d3b0c5adc56bb2510c0838c80
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:eae390ec42c3582f7896a8346df8782b287d64440794983ae93755490639e21e
        version: 1.5.1
      entries:
      - name: multicluster-global-hub-operator-rh.v1.5.1
        version: 1.5.1
      - name: multicluster-global-hub-operator-rh.v1.5.0
        version: 1.5.0
      name: release-1.5
    defaultChannel: release-1.5
    packageName: multicluster-global-hub-operator-rh
    provider:
      name: Red Hat, Inc
      url: https://github.com/stolostron/multicluster-global-hub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: https://www.turbonomic.com/
    name: t8c-tsc-client-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-tsc-client-certified.v1.4.2-candidate-20250806145822000
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "TurbonomicClient",
                "metadata": {
                  "name": "turbonomicclient-release"
                },
                "spec": {
                  "global": {
                    "version": "8.16.1"
                  },
                  "probes": {
                    "actionScript": {
                      "enabled": false
                    },
                    "actionStreamKafka": {
                      "enabled": false
                    },
                    "appDynamics": {
                      "enabled": false
                    },
                    "compellent": {
                      "enabled": false
                    },
                    "datadog": {
                      "enabled": false
                    },
                    "dynatrace": {
                      "enabled": false
                    },
                    "flexera": {
                      "enabled": false
                    },
                    "hds": {
                      "enabled": false
                    },
                    "hpe3par": {
                      "enabled": false
                    },
                    "hyperFlex": {
                      "enabled": false
                    },
                    "hyperV": {
                      "enabled": false
                    },
                    "ibmStorageFlashSystem": {
                      "enabled": false
                    },
                    "instana": {
                      "enabled": false
                    },
                    "jvm": {
                      "enabled": false
                    },
                    "mssql": {
                      "enabled": false
                    },
                    "mysql": {
                      "enabled": false
                    },
                    "netApp": {
                      "enabled": false
                    },
                    "newRelic": {
                      "enabled": false
                    },
                    "nutanix": {
                      "enabled": false
                    },
                    "oneView": {
                      "enabled": false
                    },
                    "oracle": {
                      "enabled": false
                    },
                    "powerVM": {
                      "enabled": false
                    },
                    "pure": {
                      "enabled": false
                    },
                    "scaleIO": {
                      "enabled": false
                    },
                    "serviceNow": {
                      "enabled": false
                    },
                    "snmp": {
                      "enabled": false
                    },
                    "terraform": {
                      "enabled": false
                    },
                    "tomcat": {
                      "enabled": false
                    },
                    "ucs": {
                      "enabled": false
                    },
                    "vCenter": {
                      "enabled": false
                    },
                    "vmax": {
                      "enabled": false
                    },
                    "vmm": {
                      "enabled": false
                    },
                    "vplex": {
                      "enabled": false
                    },
                    "webLogic": {
                      "enabled": false
                    },
                    "webSphere": {
                      "enabled": false
                    },
                    "wmi": {
                      "enabled": false
                    },
                    "xtremIO": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "VersionManager",
                "metadata": {
                  "name": "versionmanager-release"
                },
                "spec": {
                  "url": "http://remote-nginx-tunnel:9080/cluster-manager/clusterConfiguration"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-client-operator@sha256:8db7396c956e7a04aee92e1e90491cd3eb533cebe5f5616ca43ff16d9f3ed6e2
          createdAt: "2025-08-07T07:15:28Z"
          description: Turbonomic Secure Connect for Hybrid environment management
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.4.2-candidate-20250806145822000'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TurbonomicClient is the Schema for the turbonomicclients
              API
            displayName: Turbonomic Client
            kind: TurbonomicClient
            name: turbonomicclients.clients.turbonomic.ibm.com
            version: v1alpha1
          - description: VersionManager is the Schema for the versionmanagers API
            displayName: Version Manager
            kind: VersionManager
            name: versionmanagers.clients.turbonomic.ibm.com
            version: v1alpha1
        description: "### Turbonomic Secure Connect for Hybrid environment management\nTurbonomic
          Workload Automation for Multicloud simultaneously optimizes performance,
          compliance, and cost in real-time. Workloads are precisely resourced, automatically,
          to perform while satisfying business constraints:\n* Continuous placement
          of workload across multiple clouds both on-prem and public clouds providers.\n*
          Continuous scaling for applications and the underlying infrastructure.\n*
          Capability to connect both public cloud and on-prem targets on single saas
          instance\nIt assures application performance by giving workloads the resources
          they need when they need them.\n\n### How do I start?\nTurbonomic documentation
          can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm.\nTo
          setup Turbonomic Secure Connect client, follow the instructions below:\n1.
          Click Install on the landing page\n2. Upon moving to Install Operator page,
          do not modify the default selection for any of the configuration\n3. Select
          the Installed namespace from the drop down or add a new namespace and click
          on create project to create the entered namespace.\n4. Click Install on
          the Install Operator page.\nAt this point operator is installed.\n\nOn the
          Details tab, under Provided APIs: \n1. Click Create instance for the Turbonomic
          Client API. Enable/Disable the probes based on the requirement, and click
          Create.\n2. Click Create instance for the Version Manager API. Do not modify
          the JSON/YAML file, click on Create.\n\n### How does it work?\nTurbonomic
          uses a public APIs already exposed by application and infrastructure instrumentation
          to discover and monitor your environment.\nTurbonomic determines the right
          actions that drive continuous health, including continuous placement and
          continuous scaling for applications and the underlying cluster.\nTurbonomic
          leverages the built-on orchestration provided by the application and infrastructure
          deployment tools and automates the execution of these actions to continiously
          meet the respective service level objective of each application service."
        displayName: Turbonomic Secure Connect Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - turbonomic
        - turbo secure connect
        links:
        - name: T8c Tsc Client Certified
          url: https://t8c-tsc-certified.domain
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        relatedImages:
        - icr.io/cpopen/t8c-client-operator@sha256:8db7396c956e7a04aee92e1e90491cd3eb533cebe5f5616ca43ff16d9f3ed6e2
        version: 1.4.2-candidate-20250806145822000
      entries:
      - name: t8c-tsc-client-certified.v1.4.2-candidate-20250806145822000
        version: 1.4.2-candidate-20250806145822000
      name: candidate
    - currentCSV: t8c-tsc-client-certified.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "TurbonomicClient",
                "metadata": {
                  "name": "turbonomicclient-release"
                },
                "spec": {
                  "global": {
                    "version": "8.16.1"
                  },
                  "probes": {
                    "actionScript": {
                      "enabled": false
                    },
                    "actionStreamKafka": {
                      "enabled": false
                    },
                    "appDynamics": {
                      "enabled": false
                    },
                    "compellent": {
                      "enabled": false
                    },
                    "datadog": {
                      "enabled": false
                    },
                    "dynatrace": {
                      "enabled": false
                    },
                    "flexera": {
                      "enabled": false
                    },
                    "hds": {
                      "enabled": false
                    },
                    "hpe3par": {
                      "enabled": false
                    },
                    "hyperFlex": {
                      "enabled": false
                    },
                    "hyperV": {
                      "enabled": false
                    },
                    "ibmStorageFlashSystem": {
                      "enabled": false
                    },
                    "instana": {
                      "enabled": false
                    },
                    "jvm": {
                      "enabled": false
                    },
                    "mssql": {
                      "enabled": false
                    },
                    "mysql": {
                      "enabled": false
                    },
                    "netApp": {
                      "enabled": false
                    },
                    "newRelic": {
                      "enabled": false
                    },
                    "nutanix": {
                      "enabled": false
                    },
                    "oneView": {
                      "enabled": false
                    },
                    "oracle": {
                      "enabled": false
                    },
                    "powerVM": {
                      "enabled": false
                    },
                    "pure": {
                      "enabled": false
                    },
                    "scaleIO": {
                      "enabled": false
                    },
                    "serviceNow": {
                      "enabled": false
                    },
                    "snmp": {
                      "enabled": false
                    },
                    "terraform": {
                      "enabled": false
                    },
                    "tomcat": {
                      "enabled": false
                    },
                    "ucs": {
                      "enabled": false
                    },
                    "vCenter": {
                      "enabled": false
                    },
                    "vmax": {
                      "enabled": false
                    },
                    "vmm": {
                      "enabled": false
                    },
                    "vplex": {
                      "enabled": false
                    },
                    "webLogic": {
                      "enabled": false
                    },
                    "webSphere": {
                      "enabled": false
                    },
                    "wmi": {
                      "enabled": false
                    },
                    "xtremIO": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "VersionManager",
                "metadata": {
                  "name": "versionmanager-release"
                },
                "spec": {
                  "url": "http://remote-nginx-tunnel:9080/cluster-manager/clusterConfiguration"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-client-operator@sha256:626e024327388410f7a77bd9b8e5a5df29993fcafdb0176cd0bea08b05713c61
          createdAt: "2025-08-12T23:15:08Z"
          description: Turbonomic Secure Connect for Hybrid environment management
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.4.2'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TurbonomicClient is the Schema for the turbonomicclients
              API
            displayName: Turbonomic Client
            kind: TurbonomicClient
            name: turbonomicclients.clients.turbonomic.ibm.com
            version: v1alpha1
          - description: VersionManager is the Schema for the versionmanagers API
            displayName: Version Manager
            kind: VersionManager
            name: versionmanagers.clients.turbonomic.ibm.com
            version: v1alpha1
        description: "### Turbonomic Secure Connect for Hybrid environment management\nTurbonomic
          Workload Automation for Multicloud simultaneously optimizes performance,
          compliance, and cost in real-time. Workloads are precisely resourced, automatically,
          to perform while satisfying business constraints:\n* Continuous placement
          of workload across multiple clouds both on-prem and public clouds providers.\n*
          Continuous scaling for applications and the underlying infrastructure.\n*
          Capability to connect both public cloud and on-prem targets on single saas
          instance\nIt assures application performance by giving workloads the resources
          they need when they need them.\n\n### How do I start?\nTurbonomic documentation
          can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm.\nTo
          setup Turbonomic Secure Connect client, follow the instructions below:\n1.
          Click Install on the landing page\n2. Upon moving to Install Operator page,
          do not modify the default selection for any of the configuration\n3. Select
          the Installed namespace from the drop down or add a new namespace and click
          on create project to create the entered namespace.\n4. Click Install on
          the Install Operator page.\nAt this point operator is installed.\n\nOn the
          Details tab, under Provided APIs: \n1. Click Create instance for the Turbonomic
          Client API. Enable/Disable the probes based on the requirement, and click
          Create.\n2. Click Create instance for the Version Manager API. Do not modify
          the JSON/YAML file, click on Create.\n\n### How does it work?\nTurbonomic
          uses a public APIs already exposed by application and infrastructure instrumentation
          to discover and monitor your environment.\nTurbonomic determines the right
          actions that drive continuous health, including continuous placement and
          continuous scaling for applications and the underlying cluster.\nTurbonomic
          leverages the built-on orchestration provided by the application and infrastructure
          deployment tools and automates the execution of these actions to continiously
          meet the respective service level objective of each application service."
        displayName: Turbonomic Secure Connect Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - turbonomic
        - turbo secure connect
        links:
        - name: T8c Tsc Client Certified
          url: https://t8c-tsc-certified.domain
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        relatedImages:
        - icr.io/cpopen/t8c-client-operator@sha256:626e024327388410f7a77bd9b8e5a5df29993fcafdb0176cd0bea08b05713c61
        version: 1.4.2
      entries:
      - name: t8c-tsc-client-certified.v1.4.2
        version: 1.4.2
      name: stable
    defaultChannel: stable
    packageName: t8c-tsc-client-certified
    provider:
      name: Turbonomic, Inc.
      url: https://www.turbonomic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-recyclebin-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-recyclebin-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "recyclebin.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/recyclebin-controller:1.1.1
          createdAt: "2025-09-20T01:20:03Z"
          description: AWS Recycle Bin controller is a service controller for managing
            Recycle Bin resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Rule represents the state of an AWS recyclebin Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.recyclebin.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Recycle Bin resources in AWS from within your Kubernetes cluster.

          **About Amazon Recycle Bin**

          Recycle Bin is a data recovery feature that enables you to restore accidentally deleted Amazon EBS snapshots and EBS-backed AMIs. When using Recycle Bin, if your resources are deleted, they are retained in the Recycle Bin for a time period that you specify before being permanently deleted.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Recycle Bin
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - recyclebin
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Recycle Bin Developer Resources
          url: https://docs.aws.amazon.com/ebs/latest/userguide/recycle-bin.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: recyclebin maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/recyclebin-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-recyclebin-controller.v1.1.1
        version: 1.1.1
      - name: ack-recyclebin-controller.v1.1.0
        version: 1.1.0
      - name: ack-recyclebin-controller.v1.0.13
        version: 1.0.13
      - name: ack-recyclebin-controller.v1.0.12
        version: 1.0.12
      - name: ack-recyclebin-controller.v1.0.11
        version: 1.0.11
      - name: ack-recyclebin-controller.v1.0.10
        version: 1.0.10
      - name: ack-recyclebin-controller.v1.0.9
        version: 1.0.9
      - name: ack-recyclebin-controller.v1.0.8
        version: 1.0.8
      - name: ack-recyclebin-controller.v1.0.7
        version: 1.0.7
      - name: ack-recyclebin-controller.v1.0.6
        version: 1.0.6
      - name: ack-recyclebin-controller.v1.0.5
        version: 1.0.5
      - name: ack-recyclebin-controller.v1.0.2
        version: 1.0.2
      - name: ack-recyclebin-controller.v1.0.1
        version: 1.0.1
      - name: ack-recyclebin-controller.v1.0.0
        version: 1.0.0
      - name: ack-recyclebin-controller.v0.0.3
        version: 0.0.3
      - name: ack-recyclebin-controller.v0.0.2
        version: 0.0.2
      - name: ack-recyclebin-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-recyclebin-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ian Boden
      provider-url: ""
    name: eclipse-amlen-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eclipse-amlen-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eclipse.org/v1",
                "kind": "Amlen",
                "metadata": {
                  "name": "amlen-sample"
                },
                "spec": {
                  "_amlen_image_tag": "latest",
                  "_amlen_monitor_image_tag": "latest",
                  "_amlen_name": "amlen-sample",
                  "_amlen_namespace": "osdk-test",
                  "_amlen_persistence": "false",
                  "config": "amlen-config",
                  "device_cert_issuer": {
                    "mode": "automatic",
                    "name": "amlen-selfsigned-issuer"
                  },
                  "ha": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "size": 2
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/amlen/operator:v1.0.1a
          createdAt: "2022-11-24T15:31:00Z"
          description: An operator to run the Eclipse Amlen MQTT Broker
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/eclipse/amlen
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Amlen
            name: amlens.eclipse.org
            version: v1
        description: Operator for the eclipse Amlen MQTT Broker\n\n### Supported Feature\n\nCurrently
          supports:\n* Install - simple install process\n* Upgrade - non-disruptive
          upgrade\n*High Availability - can be enabled/disabled via the CRD\n\n###
          Documentation\nThe main Amlen website has links to product documentation,
          forum, blog and the slack channel\nhttp://www.eclipse.org/amlen\n\n### Getting
          help\nThe quickest way to get help is to use the slack channel but the forum
          is monitored as well\n\nIssues can also be added to the github repository
          https://github.com/eclipse/amlen\n\n### License\nEclipse Amlen Operator
          is licensed under the [Eclipse Public License - v 2.0](https://www.eclipse.org/legal/epl-2.0/)
        displayName: eclipse-amlen-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amlen
        - mqtt
        - broker
        links:
        - name: Eclipse Amlen Operator
          url: https://eclipse.org/amlen
        maintainers:
        - email: ianboden@uk.ibm.com
          name: Ian Boden
        maturity: stable
        provider:
          name: Ian Boden
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/amlen/operator:v1.0.1a
        version: 1.0.0
      entries:
      - name: eclipse-amlen-operator.v1.0.0
        version: 1.0.0
      - name: eclipse-amlen-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: eclipse-amlen-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eclipse.org/v1",
                "kind": "Amlen",
                "metadata": {
                  "name": "amlen-sample"
                },
                "spec": {
                  "config": "amlen-config",
                  "device_cert_issuer": {
                    "mode": "automatic",
                    "name": "amlen-selfsigned-issuer"
                  },
                  "ha": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "ldap": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "size": 2
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/amlen/operator:v1.1.3
          createdAt: "2025-05-22T10:00:00Z"
          description: An operator to run the Eclipse Amlen MQTT Broker
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/eclipse-amlen/amlen
          support: Eclipse
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Amlen
            name: amlens.eclipse.org
            version: v1
        description: |-
          Operator for the eclipse Amlen MQTT Broker

          ### Supported Feature

          Currently supports:
          * Install - simple install process
          * Upgrade - non-disruptive upgrade
          * High Availability - can be enabled/disabled via the CRD

          ### Documentation
          The main Amlen website has links to product documentation, forum, blog and the slack channel
          http://www.eclipse.org/amlen

          ### Getting help
          The quickest way to get help is to use the slack channel but the forum is monitored as well
          Issues can also be added to the github repository https://github.com/eclipse-amlen/amlen

          ### License
          Eclipse Amlen Operator is licensed under [Eclipse Public License - v 2.0](https://www.eclipse.org/legal/epl-2.0/)
        displayName: eclipse-amlen-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amlen
        - mqtt
        - broker
        links:
        - name: Eclipse Amlen Operator
          url: https://eclipse.org/amlen
        maintainers:
        - email: ianboden@uk.ibm.com
          name: Ian Boden
        maturity: stable
        provider:
          name: Eclipse Amlen
          url: https://eclipse.org/amlen
        relatedImages:
        - controller:latest
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        version: 1.1.3
      entries:
      - name: eclipse-amlen-operator.v1.1.3
        version: 1.1.3
      - name: eclipse-amlen-operator.v1.1.0
        version: 1.1.0
      - name: eclipse-amlen-operator.v1.0.2
        version: 1.0.2
      - name: eclipse-amlen-operator.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: alpha
    packageName: eclipse-amlen-operator
    provider:
      name: Ian Boden
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: vault-config-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vault-config-operator.v0.8.34
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AuthEngineMount",
                "metadata": {
                  "name": "authenginemount-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kube-authengine-mount-sample",
                  "type": "kubernetes"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azureauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azureauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azureauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureCredentials": {
                    "passwordKey": "clientsecret",
                    "secret": {
                      "name": "aad-credentials"
                    },
                    "usernameKey": "clientid"
                  },
                  "clientID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "environment": "AzurePublicCloud",
                  "maxRetries": 3,
                  "maxRetryDelay": 60,
                  "path": "azure/test",
                  "resource": "https://management.azure.com/",
                  "retryDelay": 4,
                  "tenantID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azureauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azureauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azureauthenginerole-sample"
                },
                "spec": {
                  "BoundResourceGroups": [
                    "resourcegroup1",
                    "resourcegroup2"
                  ],
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "boundGroupIDs": [
                    "group1",
                    "group2"
                  ],
                  "boundLocations": [
                    "location1",
                    "location2"
                  ],
                  "boundScaleSets": [
                    "scaleset1",
                    "scaleset1"
                  ],
                  "boundServicePrincipalIDs": [
                    "sp1",
                    "sp2"
                  ],
                  "boundSubscriptionIDs": [
                    "subscription1",
                    "subscription2"
                  ],
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "dev-role",
                  "path": "azure",
                  "policies": [
                    "policy1",
                    "policy2"
                  ],
                  "tokenBoundCIDRs": [
                    "10.20.102.13",
                    "10.20.102.15"
                  ],
                  "tokenExplicitMaxTTL": "",
                  "tokenMaxTTL": "",
                  "tokenNoDefaultPolicy": false,
                  "tokenNumUses": 0,
                  "tokenPeriod": 0,
                  "tokenPolicies": [
                    "policy1",
                    "policy2"
                  ],
                  "tokenTTL": "",
                  "tokenType": ""
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureSecretEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azuresecretengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azuresecretengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azuresecretengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureCredentials": {
                    "passwordKey": "clientsecret",
                    "secret": {
                      "name": "aad-credentials"
                    },
                    "usernameKey": "clientid"
                  },
                  "clientID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "environment": "AzurePublicCloud",
                  "passwordPolicy": "",
                  "path": "azure",
                  "rootPasswordTTL": "128d",
                  "subscriptionID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "tenantID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureSecretEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azuresecretenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azuresecretenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azuresecretenginerole-sample"
                },
                "spec": {
                  "TTL": "",
                  "applicationObjectID": "",
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureGroups": "",
                  "azureRoles": "",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "maxTTL": "",
                  "name": "azure-role",
                  "path": "azure",
                  "permanentlyDelete": "",
                  "persistApp": "",
                  "signInAudience": "",
                  "tags": ""
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "CertAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "certauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "certauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "certauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "test",
                  "path": "cert"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "CertAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "certauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "certauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "certauthenginerole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "certificate": "-----BEGIN CERTIFICATE-----\nMIIDJTCCAg2gAwIBAgIUF119kmCNlnMwK6yOOpfRs4JGhAAwDQYJKoZIhvcNAQEL\nBQAwFjEUMBIGA1UEAwwLZXhhbXBsZS5jb20wHhcNMjUwMTIxMTM0MjExWhcNMjYw\nMTIxMTM0MjExWjAWMRQwEgYDVQQDDAtleGFtcGxlLmNvbTCCASIwDQYJKoZIhvcN\nAQEBBQADggEPADCCAQoCggEBALad4Sx97VyFSfOepJoXUGYljg7hL+z20pVL2/M8\n8HXAQl4qf/i45b//Qz8VogMxdg3Y9J7O/M0n9hxbL18i1gm270VjEFyzjJCyUq0R\n72WMM4Sr3OjQPnVtJoQkWnH7wR3zsWK2N4NuIPdE3Ofb48Olyachffv7xCI+MLIa\n/adcyRggM4Z5f2nBCMcXL8D/eY03qCgd/U1aaPJ0WyJTZHLL4euL0ewKlKtB1JnU\n4MrRa5mbPQ3B+OhxcaD/7BNV6hPXAe2zftEv8LcimhbVoj3XmLJ+om58YIaEMadk\nJj61xnXbf+F/iV41onzwvTPNWCSWiLowDl8P3zkabuS583sCAwEAAaNrMGkwHQYD\nVR0OBBYEFH971CwrqkTvUz2qrUncWMS5ULwKMB8GA1UdIwQYMBaAFH971CwrqkTv\nUz2qrUncWMS5ULwKMA8GA1UdEwEB/wQFMAMBAf8wFgYDVR0RBA8wDYILZXhhbXBs\nZS5jb20wDQYJKoZIhvcNAQELBQADggEBAIQzT113TQC9BqitOCOoDFCJqcM+BHmf\nITEKlEhlqubPvA40lD4wkz67RP2SdQbF1glzZW5sDTxMjpHL+8nhdmo65CJpsQGP\nyMPwnOvx7W5JWQJa1Yt+48JNhgIymDMun+xwDsjNnlBqp0Kr7HkOxPimMPWN4R6B\nydG96puDGBYqxB1ujx5seOGfGvxPBBr3h8kNg054rtIIDrUG2iUTfpNIILJrd8Q2\nshop9//3VcIpUJO7Z1brvK/Hp4DrcKBZVo7vd8n+5uCx3vBA0tjqNAXYtFnSRvDZ\n6XV+rCZecDBFkaLuJ5YgYEK15gGe6lQAez6DTVLXMAKhZ52kceefnwM=\n-----END CERTIFICATE-----\n",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "test-role",
                  "path": "cert/test"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineConfig",
                "metadata": {
                  "name": "my-postgresql-database"
                },
                "spec": {
                  "allowedRoles": [
                    "read-only"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "connectionURL": "postgresql://{{username}}:{{password}}@my-postgresql-database.test-vault-config-operator.svc:5432",
                  "path": "test-vault-config-operator/database",
                  "pluginName": "postgresql-database-plugin",
                  "rootCredentials": {
                    "passwordKey": "postgresql-password",
                    "secret": {
                      "name": "postgresql-admin-password"
                    }
                  },
                  "username": "postgres"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineRole",
                "metadata": {
                  "name": "read-only"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "creationStatements": [
                    "CREATE ROLE \"{{name}}\" WITH LOGIN PASSWORD '{{password}}' VALID UNTIL '{{expiration}}'; GRANT SELECT ON ALL TABLES IN SCHEMA public TO \"{{name}}\";"
                  ],
                  "dBName": "my-postgresql-database",
                  "defaultTTL": "1h",
                  "maxTTL": "24h",
                  "path": "test-vault-config-operator/database"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineStaticRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "databasesecretenginestaticrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "databasesecretenginestaticrole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "databasesecretenginestaticrole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "dBName": "db-name",
                  "path": "test-vault-config-operator/database",
                  "rotationPeriod": "24h",
                  "rotationStatements": [
                    "ALTER ROLE \"{{name}}\" WITH PASSWORD '{{password}}';git add"
                  ],
                  "username": "db-user"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GCPAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "gcpauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gcpauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "gcpauthengineconfig-sample"
                },
                "spec": {
                  "GCEalias": "role_id",
                  "GCEmetadata": "default",
                  "GCPCredentials": {
                    "passwordKey": "credentials",
                    "secret": {
                      "name": "gcp-serviceaccount-credentials"
                    },
                    "usernameKey": "serviceaccount"
                  },
                  "IAMalias": "default",
                  "IAMmetadata": "default",
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "customEndpoint": {},
                  "path": "azure",
                  "serviceAccount": "vault-sa"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GCPAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "gcpauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gcpauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "gcpauthenginerole-sample"
                },
                "spec": {
                  "addGroupAliases": false,
                  "allowGCEInference": false,
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "boundInstanceGroups": [],
                  "boundLabels": [],
                  "boundProjects": {},
                  "boundRegions": [],
                  "boundServiceAccounts": {},
                  "boundZones": [],
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "maxJWTExp": "",
                  "name": "gcp-iam-role",
                  "path": "gcp",
                  "policies": [],
                  "tokenBoundCIDRs": [],
                  "tokenExplicitMaxTTL": "",
                  "tokenMaxTTL": "",
                  "tokenNoDefaultPolicy": false,
                  "tokenNumUses": 0,
                  "tokenPeriod": 0,
                  "tokenPolicies": [],
                  "tokenTTL": "",
                  "tokenType": "",
                  "type": "iam"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitHubSecretEngineConfig",
                "metadata": {
                  "name": "githubsecretengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitHubSecretEngineRole",
                "metadata": {
                  "name": "githubsecretenginerole-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Group",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "group-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "group",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "group-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "metadata": {
                    "team": "team-abc"
                  },
                  "policies": [
                    "team-abc-access"
                  ],
                  "type": "external"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupAlias",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "groupalias-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "groupalias",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "groupalias-sample"
                },
                "spec": {
                  "authEngineMountPath": "kubernetes",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "groupName": "group-sample"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "JWTOIDCAuthEngineConfig",
                "metadata": {
                  "name": "azure-oidc-config"
                },
                "spec": {
                  "OIDCClientID": "000000000-0000-0000-0000-000000000",
                  "OIDCCredentials": {
                    "passwordKey": "oidc_client_secret",
                    "secret": {
                      "name": "oidccredentials"
                    },
                    "usernameKey": "oidc_client_id"
                  },
                  "OIDCDiscoveryURL": "https://login.microsoftonline.com/000000000-0000-0000-0000-000000000/v2.0",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "oidc/azuread-oidc/"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "JWTOIDCAuthEngineRole",
                "metadata": {
                  "name": "azure-oidc-dev-role"
                },
                "spec": {
                  "OIDCScopes": [
                    "https://graph.microsoft.com/.default"
                  ],
                  "allowedRedirectURIs": [
                    "http://localhost:8250/oidc/callback",
                    "http://localhost:8200/ui/vault/auth/oidc/azuread-oidc/oidc/callback"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "groupsClaim": "groups",
                  "name": "dev-role",
                  "path": "oidc-aad/azuread-oidc",
                  "roleType": "oidc",
                  "tokenPolicies": [
                    "dev",
                    "prod"
                  ],
                  "userClaim": "email"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesAuthEngineConfig",
                "metadata": {
                  "name": "kubernetesauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kube-authengine-mount-sample"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesAuthEngineRole",
                "metadata": {
                  "name": "database-engine-admin"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kubernetes",
                  "policies": [
                    "database-engine-admin"
                  ],
                  "targetNamespaces": {
                    "targetNamespaceSelector": {
                      "matchLabels": {
                        "database-engine-admin": "true"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesSecretEngineConfig",
                "metadata": {
                  "name": "kubernetessecretengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesSecretEngineRole",
                "metadata": {
                  "name": "kubernetessecretenginerole-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "LDAPAuthEngineConfig",
                "metadata": {
                  "name": "ldapauthengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "LDAPAuthEngineGroup",
                "metadata": {
                  "name": "ldapauthenginegroup-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin",
                    "serviceAccount": {
                      "name": "default"
                    }
                  },
                  "name": "test-group",
                  "path": "ldap/test"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PKISecretEngineConfig",
                "metadata": {
                  "name": "pkisecretengineconfig-sample"
                },
                "spec": {
                  "TTL": "8760h",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "commonName": "pki-vault-demo.internal.io",
                  "crlDistributionPoints": [
                    "https://vault-internal.vault.svc:8200/v1/test-vault-config-operator/pki/crl\""
                  ],
                  "issuingCertificates": [
                    "https://vault-internal.vault.svc:8200/v1/test-vault-config-operator/pki/ca"
                  ],
                  "path": "test-vault-config-operator/pki",
                  "privateKeyType": "internal",
                  "type": "root"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PKISecretEngineRole",
                "metadata": {
                  "name": "pkisecretenginerole-sample"
                },
                "spec": {
                  "allowGlobDomains": true,
                  "allowSubdomains": true,
                  "allowedDomains": [
                    "internal.io",
                    "pki-vault-demo.svc",
                    "example.com"
                  ],
                  "allowedOtherSans": "*",
                  "allowedURISans": [
                    "*-pki-vault-demo.apps.example.com"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "maxTTL": "8760h",
                  "path": "test-vault-config-operator/pki"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PasswordPolicy",
                "metadata": {
                  "name": "simple-password-policy"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "passwordPolicy": "length = 20\nrule \"charset\" {\n  charset = \"abcdefghijklmnopqrstuvwxyz\"\n}"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Policy",
                "metadata": {
                  "name": "database-engine-admin"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "policy": "# mount database secrets engines\npath \"/sys/mounts/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n  allowed_parameters = {\n    \"type\" = [\"database\"]\n    \"*\"   = []\n  }\n}\n\n# tune database secrets engines\npath \"/sys/mounts/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/tune\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n\n# Configure database secrets engines\npath \"/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/config/+\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n\n# Configure database roles\npath \"/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/roles/+\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineConfig",
                "metadata": {
                  "name": "quaysecretengineconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineRole",
                "metadata": {
                  "name": "quaysecretenginerole-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineStaticRole",
                "metadata": {
                  "name": "quaysecretenginestaticrole-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RabbitMQSecretEngineConfig",
                "metadata": {
                  "name": "rabbitmqsecretengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "rabbitMQ-engine-admin"
                  },
                  "connectionURI": "https://test.com",
                  "leaseMaxTTL": 86400,
                  "leaseTTL": 86400,
                  "path": "vault-config-operator/rabbitmq"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RabbitMQSecretEngineRole",
                "metadata": {
                  "name": "rabbitmqsecretenginerole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "rabbitmq-engine-admin"
                  },
                  "path": "vault-config-operator/rabbitmq",
                  "tags": "administrator",
                  "vhostTopics": [
                    {
                      "topics": [
                        {
                          "permissions": {
                            "configure": ".*",
                            "read": ".*",
                            "write": ".*"
                          },
                          "topicName": "my-topic"
                        },
                        {
                          "permissions": {
                            "read": ".*"
                          },
                          "topicName": "my-read-topic"
                        }
                      ],
                      "vhostName": "/"
                    }
                  ],
                  "vhosts": [
                    {
                      "permissions": {
                        "configure": ".*",
                        "read": ".*",
                        "write": ".*"
                      },
                      "vhostName": "/"
                    },
                    {
                      "permissions": {
                        "read": "my-queue",
                        "write": "my-exchange"
                      },
                      "vhostName": "my-vhost"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RandomSecret",
                "metadata": {
                  "name": "randomsecret-password"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "secret-writer"
                  },
                  "path": "test-vault-config-operator/kv",
                  "secretFormat": {
                    "passwordPolicyName": "simple-password-policy"
                  },
                  "secretKey": "password"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "SecretEngineMount",
                "metadata": {
                  "name": "database"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "path": "test-vault-config-operator",
                  "type": "database"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "VaultSecret",
                "metadata": {
                  "name": "vaultsecret-sample"
                },
                "spec": {
                  "output": {
                    "annotations": {
                      "refresh": "every-minute"
                    },
                    "labels": {
                      "app": "test-vault-config-operator"
                    },
                    "name": "randomsecret",
                    "stringData": {
                      "anotherpassword": "{{ .anotherrandomsecret.password }}",
                      "password": "{{ .randomsecret.password }}"
                    },
                    "type": "Opaque"
                  },
                  "refreshPeriod": "3m0s",
                  "vaultSecretDefinitions": [
                    {
                      "authentication": {
                        "path": "kubernetes",
                        "role": "secret-reader",
                        "serviceAccount": {
                          "name": "default"
                        }
                      },
                      "name": "randomsecret",
                      "path": "test-vault-config-operator/kv/randomsecret-password"
                    },
                    {
                      "authentication": {
                        "path": "kubernetes",
                        "role": "secret-reader",
                        "serviceAccount": {
                          "name": "default"
                        }
                      },
                      "name": "anotherrandomsecret",
                      "path": "test-vault-config-operator/kv/another-password"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/vault-config-operator@sha256:490852417fac37f3003ab7c1ee797f9035fd2905c354cc46eca5e33d2ad780ce
          createdAt: "2025-09-18T18:56:36Z"
          description: This operator provides primitives to declaratively configure
            Hashicorp Vault.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: vault-config-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/vault-config-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthEngineMount is the Schema for the authenginemounts API
            displayName: Auth Engine Mount
            kind: AuthEngineMount
            name: authenginemounts.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureAuthEngineConfig is the Schema for the azureauthengineconfigs
              API
            displayName: Azure Auth Engine Config
            kind: AzureAuthEngineConfig
            name: azureauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureAuthEngineRole is the Schema for the azureauthengineroles
              API
            displayName: Azure Auth Engine Role
            kind: AzureAuthEngineRole
            name: azureauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - kind: AzureSecretEngineConfig
            name: azuresecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureSecretEngineRole is the Schema for the azuresecretengineroles
              API
            displayName: Azure Secret Engine Role
            kind: AzureSecretEngineRole
            name: azuresecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: CertAuthEngineConfig is the Schema for the certauthengineconfigs
              API
            displayName: Cert Auth Engine Config
            kind: CertAuthEngineConfig
            name: certauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: CertAuthEngineRole is the Schema for the certauthengineroles
              API
            displayName: Cert Auth Engine Role
            kind: CertAuthEngineRole
            name: certauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineConfig is the Schema for the databasesecretengineconfigs
              API
            displayName: Database Secret Engine Config
            kind: DatabaseSecretEngineConfig
            name: databasesecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineRole is the Schema for the databasesecretengineroles
              API
            displayName: Database Secret Engine Role
            kind: DatabaseSecretEngineRole
            name: databasesecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineStaticRole is the Schema for the databasesecretenginestaticroles
              API
            displayName: Database Secret Engine Static Role
            kind: DatabaseSecretEngineStaticRole
            name: databasesecretenginestaticroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GCPAuthEngineConfig is the Schema for the gcpauthengineconfigs
              API
            displayName: GCPAuth Engine Config
            kind: GCPAuthEngineConfig
            name: gcpauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: GCPAuthEngineRole is the Schema for the gcpauthengineroles
              API
            displayName: GCPAuth Engine Role
            kind: GCPAuthEngineRole
            name: gcpauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GitHubSecretEngineConfig is the Schema for the githubsecretengineconfigs
              API
            displayName: Git Hub Secret Engine Config
            kind: GitHubSecretEngineConfig
            name: githubsecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: GitHubSecretEngineRole is the Schema for the githubsecretengineroles
              API
            displayName: Git Hub Secret Engine Role
            kind: GitHubSecretEngineRole
            name: githubsecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GroupAlias is the Schema for the groupaliases API
            displayName: Group Alias
            kind: GroupAlias
            name: groupaliases.redhatcop.redhat.io
            version: v1alpha1
          - description: Group is the Schema for the groups API
            displayName: Group
            kind: Group
            name: groups.redhatcop.redhat.io
            version: v1alpha1
          - description: JWTOIDCAuthEngineConfig is the Schema for the jwtoidcauthengineconfigs
              API
            displayName: JWTOIDCAuth Engine Config
            kind: JWTOIDCAuthEngineConfig
            name: jwtoidcauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: JWTOIDCAuthEngineRole is the Schema for the jwtoidcauthengineroles
              API
            displayName: JWTOIDCAuth Engine Role
            kind: JWTOIDCAuthEngineRole
            name: jwtoidcauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesAuthEngineConfig is the Schema for the kubernetesauthengineconfigs
              API
            displayName: Kubernetes Auth Engine Config
            kind: KubernetesAuthEngineConfig
            name: kubernetesauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesAuthEngineRole can be used to define a KubernetesAuthEngineRole
              for the kube-auth authentication method
            displayName: Kubernetes Auth Engine Role
            kind: KubernetesAuthEngineRole
            name: kubernetesauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesSecretEngineConfig is the Schema for the kubernetessecretengineconfigs
              API
            displayName: Kubernetes Secret Engine Config
            kind: KubernetesSecretEngineConfig
            name: kubernetessecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesSecretEngineRole is the Schema for the kubernetessecretengineroles
              API
            displayName: Kubernetes Secret Engine Role
            kind: KubernetesSecretEngineRole
            name: kubernetessecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: LDAPAuthEngineConfig is the Schema for the ldapauthengineconfigs
              API
            displayName: LDAPAuth Engine Config
            kind: LDAPAuthEngineConfig
            name: ldapauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: LDAPAuthEngineGroup is the Schema for the ldapauthenginegroups
              API
            displayName: LDAPAuth Engine Group
            kind: LDAPAuthEngineGroup
            name: ldapauthenginegroups.redhatcop.redhat.io
            version: v1alpha1
          - description: PasswordPolicy is the Schema for the passwordpolicies API
            displayName: Password Policy
            kind: PasswordPolicy
            name: passwordpolicies.redhatcop.redhat.io
            version: v1alpha1
          - description: PKISecretEngineConfig is the Schema for the pkisecretengineconfigs
              API
            displayName: PKISecret Engine Config
            kind: PKISecretEngineConfig
            name: pkisecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: PKISecretEngineRole is the Schema for the pkisecretengineroles
              API
            displayName: PKISecret Engine Role
            kind: PKISecretEngineRole
            name: pkisecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: Policy is the Schema for the policies API
            displayName: Policy
            kind: Policy
            name: policies.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineConfig is the Schema for the quaysecretengineconfigs
              API
            displayName: Quay Secret Engine Config
            kind: QuaySecretEngineConfig
            name: quaysecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineRole is the Schema for the quaysecretengineroles
              API
            displayName: Quay Secret Engine Role
            kind: QuaySecretEngineRole
            name: quaysecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineStaticRole is the Schema for the quaysecretenginestaticroles
              API
            displayName: Quay Secret Engine Static Role
            kind: QuaySecretEngineStaticRole
            name: quaysecretenginestaticroles.redhatcop.redhat.io
            version: v1alpha1
          - description: RabbitMQSecretEngineConfig is the Schema for the rabbitmqsecretengineconfigs
              API
            displayName: Rabbit MQSecret Engine Config
            kind: RabbitMQSecretEngineConfig
            name: rabbitmqsecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: RabbitMQSecretEngineRole is the Schema for the rabbitmqsecretengineroles
              API
            displayName: Rabbit MQSecret Engine Role
            kind: RabbitMQSecretEngineRole
            name: rabbitmqsecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: RandomSecret is the Schema for the randomsecrets API
            displayName: Random Secret
            kind: RandomSecret
            name: randomsecrets.redhatcop.redhat.io
            version: v1alpha1
          - description: SecretEngineMount is the Schema for the secretenginemounts
              API
            displayName: Secret Engine Mount
            kind: SecretEngineMount
            name: secretenginemounts.redhatcop.redhat.io
            version: v1alpha1
          - description: VaultSecret is the Schema for the vaultsecrets API
            displayName: Vault Secret
            kind: VaultSecret
            name: vaultsecrets.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator helps set up Vault Configurations. The main intent is to do so such that subsequently pods can consume the secrets made available.
          There are two main principles through all of the capabilities of this operator:

          1. high-fidelity API. The CRD exposed by this operator reflect field by field the Vault APIs. This is because we don't want to make any assumption on the kinds of configuration workflow that user will set up. That being said the Vault API is very extensive and we are starting with enough API coverage to support, we think, some simple and very common configuration workflows.
          2. attention to security (after all we are integrating with a security tool). To prevent credential leaks we give no permissions to the operator itself against Vault. All APIs exposed by this operator contains enough information to authenticate to Vault using a local service account (local to the namespace where the API exist). In other word for a namespace user to be abel to successfully configure Vault, a service account in that namespace must have been previously given the needed Vault permissions.

          Currently this operator supports the following CRDs:

          1. [Policy](https://github.com/redhat-cop/vault-config-operator#policy) Configures Vault [Policies](https://www.vaultproject.io/docs/concepts/policies)
          2. [KubernetesAuthEngineRole](https://github.com/redhat-cop/vault-config-operator#KubernetesAuthEngineRole) Configures a Vault [Kubernetes Authentication](https://www.vaultproject.io/docs/auth/kubernetes) Role
          3. [SecretEngineMount](https://github.com/redhat-cop/vault-config-operator#SecretEngineMount) Configures a Mount point for a [SecretEngine](https://www.vaultproject.io/docs/secrets)
          4. [DatabaseSecretEngineConfig](https://github.com/redhat-cop/vault-config-operator#DatabaseSecretEngineConfig) Configures a [Database Secret Engine](https://www.vaultproject.io/docs/secrets/databases) Connection
          5. [DatabaseSecretEngineRole](https://github.com/redhat-cop/vault-config-operator#DatabaseSecretEngineRole) Configures a [Database Secret Engine](https://www.vaultproject.io/docs/secrets/databases) Role
          6. [RandomSecret](https://github.com/redhat-cop/vault-config-operator#RandomSecret) Creates a random secret in a vault [kv Secret Engine](https://www.vaultproject.io/docs/secrets/kv) with one password field generated using a [PasswordPolicy](https://www.vaultproject.io/docs/concepts/password-policies)vault-config-operator
        displayName: Vault Config Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - secrets
        - gitops
        - policy
        - credentials
        links:
        - name: Vault Config Operator
          url: https://github.com/redhat-cop/vault-config-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/vault-config-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/vault-config-operator@sha256:490852417fac37f3003ab7c1ee797f9035fd2905c354cc46eca5e33d2ad780ce
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 0.8.34
      entries:
      - name: vault-config-operator.v0.8.34
        version: 0.8.34
      - name: vault-config-operator.v0.8.33
        version: 0.8.33
      - name: vault-config-operator.v0.8.32
        version: 0.8.32
      - name: vault-config-operator.v0.8.31
        version: 0.8.31
      - name: vault-config-operator.v0.8.30
        version: 0.8.30
      - name: vault-config-operator.v0.8.29
        version: 0.8.29
      - name: vault-config-operator.v0.8.28
        version: 0.8.28
      - name: vault-config-operator.v0.8.27
        version: 0.8.27
      - name: vault-config-operator.v0.8.25
        version: 0.8.25
      - name: vault-config-operator.v0.8.24
        version: 0.8.24
      - name: vault-config-operator.v0.8.23
        version: 0.8.23
      - name: vault-config-operator.v0.8.22
        version: 0.8.22
      - name: vault-config-operator.v0.8.21
        version: 0.8.21
      - name: vault-config-operator.v0.8.20
        version: 0.8.20
      - name: vault-config-operator.v0.8.19
        version: 0.8.19
      - name: vault-config-operator.v0.8.18
        version: 0.8.18
      - name: vault-config-operator.v0.8.17
        version: 0.8.17
      - name: vault-config-operator.v0.8.16
        version: 0.8.16
      - name: vault-config-operator.v0.8.15
        version: 0.8.15
      - name: vault-config-operator.v0.8.14
        version: 0.8.14
      - name: vault-config-operator.v0.8.13
        version: 0.8.13
      - name: vault-config-operator.v0.8.12
        version: 0.8.12
      - name: vault-config-operator.v0.8.11
        version: 0.8.11
      - name: vault-config-operator.v0.8.10
        version: 0.8.10
      - name: vault-config-operator.v0.8.9
        version: 0.8.9
      - name: vault-config-operator.v0.8.7
        version: 0.8.7
      - name: vault-config-operator.v0.8.5
        version: 0.8.5
      - name: vault-config-operator.v0.8.4
        version: 0.8.4
      - name: vault-config-operator.v0.8.3
        version: 0.8.3
      - name: vault-config-operator.v0.8.2
        version: 0.8.2
      - name: vault-config-operator.v0.8.0
        version: 0.8.0
      - name: vault-config-operator.v0.7.1
        version: 0.7.1
      - name: vault-config-operator.v0.6.5
        version: 0.6.5
      - name: vault-config-operator.v0.6.4
        version: 0.6.4
      - name: vault-config-operator.v0.6.3
        version: 0.6.3
      - name: vault-config-operator.v0.6.2
        version: 0.6.2
      - name: vault-config-operator.v0.6.1
        version: 0.6.1
      - name: vault-config-operator.v0.6.0
        version: 0.6.0
      - name: vault-config-operator.v0.5.1
        version: 0.5.1
      - name: vault-config-operator.v0.5.0
        version: 0.5.0
      - name: vault-config-operator.v0.4.1
        version: 0.4.1
      - name: vault-config-operator.v0.4.0
        version: 0.4.0
      - name: vault-config-operator.v0.3.1
        version: 0.3.1
      - name: vault-config-operator.v0.3.0
        version: 0.3.0
      - name: vault-config-operator.v0.2.0
        version: 0.2.0
      - name: vault-config-operator.v0.1.2
        version: 0.1.2
      - name: vault-config-operator.v0.1.1
        version: 0.1.1
      - name: vault-config-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: vault-config-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-rest-api-browser-v1
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-rest-api-browser-v1.v0.1.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rest.loadcore.keysight.com/v1alpha1",
                "kind": "Apibrowser",
                "metadata": {
                  "name": "apibrowser-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-06-29T14:08:05Z"
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Apibrowser
            name: apibrowsers.rest.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Rest API Browser Operator v1
        displayName: Keysight Rest API Browser Operator v1
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Rest Api Browser V1
          url: https://keysight-rest-api-browser-v1.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5d5061dbc2ec5082cf14b6c600fb5400b83cf91d7ccebfa80680a238d275db
        - prod.harbor.keysight.digital/loadcore/keysight-rest-api-browser-v1@sha256:3ada6a0fa630acb0fa962a11d2c272ff679dded1106c4693d2c8f5ec9ab414a4
        version: 0.1.11
      entries:
      - name: keysight-rest-api-browser-v1.v0.1.11
        version: 0.1.11
      name: alpha
    defaultChannel: alpha
    packageName: keysight-rest-api-browser-v1
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Qualcomm Technologies, Inc.
      provider-url: https://github.com/quic/aic-operator
    name: aic-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aic-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aic.quicinc.com/v1",
                "kind": "AIC",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "aic-operator",
                    "app.kubernetes.io/instance": "aic-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "aic",
                    "app.kubernetes.io/part-of": "aic-operator"
                  },
                  "name": "aic-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "false"
          containerImage: ghcr.io/quic/cloud_ai_openshift_operator@sha256:9cf581e9b5c9cd08819fab8f8b6313836a0a20098ad4c99b80d93dbcf5ab32bf
          description: Automate the management of OpenShift Operator for Qualcomm
            AI 100 accelerators.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: Qualcomm Technologies, Inc.
          repository: https://github.com/quic/aic-operator
          support: Qualcomm Technologies, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AIC is the Schema for the aics API
            displayName: AIC
            kind: AIC
            name: aics.aic.quicinc.com
            version: v1
        description: The Openshift Operator enables the Qualcomm® Cloud AI series
          of hardware on OpenShift clusters by automating the configuration and installation
          of their Linux device drivers and its Kubernetes Device Plugin.
        displayName: AIC Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AIC
        - AIC-Operator
        links:
        - name: AIC Operator
          url: https://github.com/quic/aic-operator
        maintainers:
        - email: quic_vmellemp@quicinc.com
          name: quic_vmellemp
        - email: quic_carls@quicinc.com
          name: quic_carls
        maturity: alpha
        provider:
          name: Qualcomm Technologies, Inc.
          url: https://github.com/quic/aic-operator
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:7a943db9bcb22f817d75341621472f4bb30abc5625c53b9a781d97d431bc4d09
        - ghcr.io/quic/cloud_ai_openshift_operator@sha256:9cf581e9b5c9cd08819fab8f8b6313836a0a20098ad4c99b80d93dbcf5ab32bf
        version: 0.3.1
      entries:
      - name: aic-operator.v0.3.1
        version: 0.3.1
      name: alpha
    defaultChannel: alpha
    packageName: aic-operator
    provider:
      name: Qualcomm Technologies, Inc.
      url: https://github.com/quic/aic-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kong Inc.
      provider-url: https://konghq.com
    name: kong-gateway-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kong-gateway-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "ControlPlane",
                "metadata": {
                  "name": "controlplane-example"
                },
                "spec": {
                  "dataplane": "dataplane-example",
                  "gatewayClass": "kong"
                }
              },
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "DataPlane",
                "metadata": {
                  "name": "dataplane-example"
                },
                "spec": {
                  "env": [
                    {
                      "name": "KONG_LOG_LEVEL",
                      "value": "debug"
                    }
                  ]
                }
              },
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "GatewayConfiguration",
                "metadata": {
                  "name": "kong",
                  "namespace": "default"
                },
                "spec": {
                  "controlPlaneDeploymentOptions": {
                    "env": [
                      {
                        "name": "CONTROLLER_LOG_LEVEL",
                        "value": "debug"
                      }
                    ]
                  },
                  "dataPlaneDeploymentOptions": {
                    "containerImage": "kong",
                    "version": "2.8"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking,Integration & Delivery
          operatorframework.io/suggested-namespace: kong-system
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ControlPlane is the Schema for the controlplanes API
            displayName: Control Plane
            kind: ControlPlane
            name: controlplanes.gateway-operator.konghq.com
            version: v1alpha1
          - description: DataPlane is the Schema for the dataplanes API
            displayName: Data Plane
            kind: DataPlane
            name: dataplanes.gateway-operator.konghq.com
            version: v1alpha1
          - description: GatewayConfiguration is the Schema for the gatewayconfigurations
              API
            displayName: Gateway Configuration
            kind: GatewayConfiguration
            name: gatewayconfigurations.gateway-operator.konghq.com
            version: v1alpha1
        description: Kong Operator for Kubernetes. Technical preview.
        displayName: Kong
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - gateway
        - ingress
        - kong
        - kubernetes
        - go
        links:
        - name: Github repository for the operator
          url: https://github.com/kong/gateway-operator
        maintainers:
        - email: support@konghq.com
          name: Kong Inc.
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Kong Inc.
          url: https://konghq.com
        relatedImages:
        - ghcr.io/kong/gateway-operator@sha256:76b227715a3e1915a28f8aaa1e2ff8495349d29d0d61e7b85a72d74b03fbcc7f
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.connect.redhat.com/kong/kong@sha256:54e7c7e57bf93810a7cc36327023b8699f754c4c40c5c289548d54170caa5db2
        - registry.connect.redhat.com/kong/kong-ingress-controller@sha256:f691db070e7c604da9e163800903d0942ec6fda879eb9dcd5d8e90618ae787da
        version: 0.0.2
      entries:
      - name: kong-gateway-operator.v0.0.2
        version: 0.0.2
      - name: kong-gateway-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: kong-gateway-operator
    provider:
      name: Kong Inc.
      url: https://konghq.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: ovms-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openvino-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "ModelServer",
                "metadata": {
                  "name": "model-server-sample"
                },
                "spec": {
                  "image_name": "registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0",
                  "deployment_parameters": {
                    "replicas": 1
                  },
                  "service_parameters": {
                    "grpc_port": 8080,
                    "rest_port": 8081
                  },
                  "models_settings": {
                    "single_model_mode": true,
                    "config_configmap_name": "",
                    "config_path": "",
                    "model_name": "resnet",
                    "model_path": "gs://<bucket_name>/<model_dir>",
                    "nireq": 0,
                    "plugin_config": "{\"PERFORMANCE_HINT\":\"LATENCY\"}",
                    "batch_size": "",
                    "shape": "",
                    "model_version_policy": "{\"latest\": { \"num_versions\":1 }}",
                    "layout": "",
                    "target_device": "CPU",
                    "is_stateful": false,
                    "idle_sequence_cleanup": false,
                    "low_latency_transformation": true,
                    "max_sequence_number": 0
                  },
                  "server_settings": {
                    "file_system_poll_wait_seconds": 0,
                    "sequence_cleaner_poll_wait_minutes": 0,
                    "log_level": "INFO",
                    "grpc_workers": 1,
                    "rest_workers": 0
                  },
                  "models_repository": {
                    "https_proxy": "",
                    "http_proxy": "",
                    "storage_type": "google",
                    "models_host_path": "",
                    "models_volume_claim": "",
                    "aws_secret_access_key": "",
                    "aws_access_key_id": "",
                    "aws_region": "",
                    "s3_compat_api_endpoint": "",
                    "gcp_creds_secret_name": "",
                    "azure_storage_connection_string": ""
                  }
                }
              },
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "Notebook",
                "metadata": {
                  "name": "v2024.1"
                },
                "spec": {
                  "git_ref": "2024.1",
                  "auto_update_image": false
                }
              }
            ]
          alm-examples-metadata: '{ "model-server-sample":{"description":"Example
            of the model server configuration"}, "v2024":{"description":"Default notebook
            configuration"} }'
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
          createdAt: "2024-05-14T09:29:35Z"
          description: OpenVINO Toolkit Operator manages OpenVINO components in OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/ovms-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/ovms-operator-rhmp/support?utm_source=openshift_console
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.34.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Provides inference serving through gRPC & REST
            displayName: Model Server
            kind: ModelServer
            name: modelservers.intel.com
            version: v1alpha1
          - description: Adds a collection of ready-to-run Jupyter notebooks for learning
              and experimenting with the OpenVINO™ Toolkit to Red Hat® OpenShift®
              Data Science
            displayName: Notebook
            kind: Notebook
            name: notebooks.intel.com
            version: v1alpha1
        description: |
          OpenVINO Toolkit Operator manages OpenVINO components in OpenShift.

          Currently there available components are ModelServer and Notebook.

          # Model Server
          [OpenVINO™ Model Server](https://github.com/openvinotoolkit/model_server) (OVMS) is a scalable, high-performance solution for serving machine learning models optimized for Intel® architectures. The server provides an inference service via gRPC or REST API for any models trained in a framework that is supported by [OpenVINO](https://software.intel.com/en-us/openvino-toolki://docs.openvino.ai/latest/index.html).
          Model Server configuration parameters is explained [here](https://github.com/openvinotoolkit/operator/blob/main/docs/modelserver_params.md).
          ## Using the cluster
          OpenVINO Model Server can be consumed as a `Service`. It is called like with `ModelServer` resource with `-ovms` suffix.
          The suffix is omitted when `ovms` phrase is included in the name.
          The service exposes gRPC and REST API interfaces to run inference requests.
          ```
          oc get pods
          NAME                                        READY   STATUS    RESTARTS   AGE
          model-server-sample-ovms-586f6f76df-dpps4   1/1     Running   0          8h

          oc get services
          NAME                       TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)             AGE
          model-server-sample-ovms   ClusterIP   172.25.199.210   <none>        8080/TCP,8081/TCP   8h
          ```

          # Notebook
          The Notebook resource integrates JupyterHub from OpenShift Data Science or Open Data Hub with a container image that includes [developer
          tools](https://github.com/openvinotoolkit/openvino/blob/master/docs/install_guides/pypi-openvino-dev.md)
          from the OpenVINO toolkit and a set of Jupyter notebook tutorials. It enables selecting a defined image `OpenVINO™ Toolkit` from
          the Jupyter Spawner choice list.

          Create the `Notebook` resource in the same project with JupyterHub and RedHat OpenShift Data Science operator as seen in [documentation](https://github.com/openvinotoolkit/operator/blob/main/docs/notebook_in_rhods.md).
          It builds the image in the cluster based on Dockerfile from [openvino_notebooks](https://github.com/openvinotoolkit/openvino_notebooks).

          ## References
          OpenVINO Model Server on [Github](https://github.com/openvinotoolkit/model_server)

          OpenVINO Model Server Operator on [Github](https://github.com/openvinotoolkit/operator)
        displayName: OpenVINO Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Inference OpenVINO
        links:
        - name: OpenVINO Toolkit Operator
          url: https://github.com/openvinotoolkit/operator
        - name: OpenVINO Model Server
          url: https://github.com/openvinotoolkit/model_server
        - name: OpenVINO Notebooks
          url: https://github.com/openvinotoolkit/openvino_notebooks
        maintainers:
        - email: openvino-operator-support@intel.com
          name: Operator Support
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5714b93ac39277affdc10c3ba142d8c032ed0ddb2ba61c04cbddac710464c4
        - registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
        - registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0
        version: 1.2.0
      entries:
      - name: openvino-operator.v1.2.0
        version: 1.2.0
      - name: openvino-operator.v1.1.0
        version: 1.1.0
      - name: openvino-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: ovms-operator-rhmp
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Aerospike
      provider-url: ""
    name: aerospike-kubernetes-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aerospike-kubernetes-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "memory-size": 3000000000,
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:6.3.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: quay.io/aerospike/aerospike-kubernetes-operator@sha256:a3033b50ae434b97de4de1ca03013dabcfac580d810a6ee8c0fe05b6a9301952
          createdAt: 2023-05-04T09:32:47UTC
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: <=2.4.0
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Aerospike
        relatedImages:
        - quay.io/aerospike/aerospike-kubernetes-operator@sha256:a3033b50ae434b97de4de1ca03013dabcfac580d810a6ee8c0fe05b6a9301952
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        version: 2.5.0
      entries:
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      name: alpha
    - currentCSV: aerospike-kubernetes-operator-rhmp.v4.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "data-size": 1073741824,
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:8.1.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackup",
                "metadata": {
                  "name": "aerospikebackup",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "aerospike-cluster": {
                      "aerospike-aerospikebackup-test-cluster": {
                        "credentials": {
                          "password": "admin123",
                          "user": "admin"
                        },
                        "seed-nodes": [
                          {
                            "host-name": "aerocluster.aerospike.svc.cluster.local",
                            "port": 3000
                          }
                        ]
                      }
                    },
                    "backup-routines": {
                      "aerospike-aerospikebackup-test-routine": {
                        "backup-policy": "test-policy",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "local"
                      },
                      "aerospike-aerospikebackup-test-routine1": {
                        "backup-policy": "test-policy1",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "s3Storage"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackupService",
                "metadata": {
                  "name": "aerospikebackupservice-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "config": {
                    "backup-policies": {
                      "test-policy": {
                        "parallel": 3
                      },
                      "test-policy1": {
                        "parallel": 3
                      }
                    },
                    "service": {
                      "http": {
                        "port": 8081
                      }
                    },
                    "storage": {
                      "local": {
                        "local-storage": {
                          "path": "/tmp/localStorage"
                        }
                      },
                      "s3Storage": {
                        "s3-storage": {
                          "bucket": "aerospike-kubernetes-operator-test",
                          "s3-profile": "default",
                          "s3-region": "us-east-1"
                        }
                      }
                    }
                  },
                  "image": "aerospike/aerospike-backup-service:3.2.0",
                  "secrets": [
                    {
                      "secretName": "aws-secret",
                      "volumeMount": {
                        "mountPath": "/root/.aws/credentials",
                        "name": "aws-secret",
                        "subPath": "credentials"
                      }
                    }
                  ],
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeRestore",
                "metadata": {
                  "name": "aerospikerestore-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "backup-data-path": "aerospike-aerospikebackup-test-routine/backup/1733506015827/data/test",
                    "destination": {
                      "credentials": {
                        "password": "admin123",
                        "user": "admin"
                      },
                      "label": "destinationCluster",
                      "seed-nodes": [
                        {
                          "host-name": "aerocluster.test.svc.cluster.local",
                          "port": 3000
                        }
                      ]
                    },
                    "policy": {
                      "no-generation": true,
                      "no-indexes": true,
                      "parallel": 3
                    },
                    "source": {
                      "local-storage": {
                        "path": "/tmp/localStorage"
                      }
                    }
                  },
                  "type": "Full"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/aerospike/aerospike-kubernetes-operator@sha256:e6d6bd7c58e036b3728e52475a9b764ca8c46b83cbc4751cac66294495bf2b31
          createdAt: "2025-09-15T10:14:05Z"
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
          support: Aerospike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeBackup is the Schema for the aerospikebackup API
            displayName: Aerospike Backup
            kind: AerospikeBackup
            name: aerospikebackups.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeBackupService is the Schema for the aerospikebackupservices
              API
            displayName: Aerospike Backup Service
            kind: AerospikeBackupService
            name: aerospikebackupservices.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1
          - description: AerospikeRestore is the Schema for the aerospikerestores
              API
            displayName: Aerospike Restore
            kind: AerospikeRestore
            name: aerospikerestores.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
           -   Aerospike cluster monitoring
           -   Backup and restore Aerospike clusters
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Aerospike
        relatedImages:
        - quay.io/aerospike/aerospike-kubernetes-operator@sha256:e6d6bd7c58e036b3728e52475a9b764ca8c46b83cbc4751cac66294495bf2b31
        version: 4.1.1
      entries:
      - name: aerospike-kubernetes-operator-rhmp.v4.1.1
        version: 4.1.1
      - name: aerospike-kubernetes-operator-rhmp.v4.1.0
        version: 4.1.0
      - name: aerospike-kubernetes-operator-rhmp.v4.0.2
        version: 4.0.2
      - name: aerospike-kubernetes-operator-rhmp.v4.0.1
        version: 4.0.1
      - name: aerospike-kubernetes-operator-rhmp.v4.0.0
        version: 4.0.0
      - name: aerospike-kubernetes-operator-rhmp.v3.4.1
        version: 3.4.1
      - name: aerospike-kubernetes-operator-rhmp.v3.4.0
        version: 3.4.0
      - name: aerospike-kubernetes-operator-rhmp.v3.3.1
        version: 3.3.1
      - name: aerospike-kubernetes-operator-rhmp.v3.3.0
        version: 3.3.0
      - name: aerospike-kubernetes-operator-rhmp.v3.2.2
        version: 3.2.2
      - name: aerospike-kubernetes-operator-rhmp.v3.2.1
        version: 3.2.1
      - name: aerospike-kubernetes-operator.v3.2.0
        version: 3.2.0
      - name: aerospike-kubernetes-operator.v3.1.0
        version: 3.1.0
      - name: aerospike-kubernetes-operator.v3.0.0
        version: 3.0.0
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      name: stable
    defaultChannel: stable
    packageName: aerospike-kubernetes-operator-rhmp
    provider:
      name: Aerospike
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: samsung
      provider-url: ""
    name: socmmd
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: socmmd.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cmmd.samsung.com/v1",
                "kind": "CMMD",
                "metadata": {
                  "name": "my-app",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "allocateMode": "auto",
                  "enable": true,
                  "payload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "metadata": {
                      "name": "my-app"
                    },
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "name": "my-app"
                        }
                      },
                      "template": {
                        "metadata": {
                          "labels": {
                            "name": "my-app"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "quay.io/my-account/my-app:latest",
                              "name": "my-app-01"
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "cmmd.samsung.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "machineconfiguration.openshift.io/role": "cmmd"
                  },
                  "name": "cmmd-node-label",
                  "namespace": "socmmd"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          console.openshift.io/plugins: '["socmmd-console-plugin"]'
          containerImage: quay.io/samsung_sag/socmmd@sha256:2ef361fb15f0fec99e0a8f1ef83ad4dd14f701d579c900cce084c4a62cf805f8
          createdAt: "2024-08-01T05:45:38Z"
          description: Samsung Operator For CMM-D
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "cmmd.samsung.com/v1",
              "kind": "Label",
              "metadata": {
                "name": "cmmd-node-label",
                "namespace": "socmmd",
                "labels": {}
              },
              "spec": {}
            }
          operatorframework.io/suggested-namespace: socmmd-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenMPDK/SMDK/tree/main/lib/redhat/socmmd
          support: samsung
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CMMD is the Schema for the cmmds API.
            displayName: CMMD
            kind: CMMD
            name: cmmds.cmmd.samsung.com
            version: v1
          - description: Label is the Schema for the labels API.
            displayName: Node Label
            kind: Label
            name: labels.cmmd.samsung.com
            version: v1
        description: Samsung Operator For CMM-D
        displayName: socmmd
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - samsung
        - memory
        - extend
        - cmm-d
        links:
        - name: Samsung Operator For CMM-D
          url: https://github.com/OpenMPDK/SMDK/tree/main/lib/redhat/socmmd
        maintainers:
        - email: junhyeok.im@samsung.com
          name: junhyeok-im
        maturity: stable
        provider:
          name: samsung
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/samsung_sag/socmmd@sha256:2ef361fb15f0fec99e0a8f1ef83ad4dd14f701d579c900cce084c4a62cf805f8
        version: 0.0.2
      entries:
      - name: socmmd.v0.0.2
        version: 0.0.2
      - name: socmmd.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: socmmd
    provider:
      name: samsung
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicast-community-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicast-community-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/apicast-operator:v0.8.0
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway built on top of NGINX. It is part
            of the Red Hat 3scale API Management Platform
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |
          The APIcast Operator creates and maintains the Red Hat 3scale API Gateway in several deployment configurations.

          APIcast is an API gateway built on top of [NGINX](https://www.nginx.com/). It is part of the [Red Hat 3scale API Management Platform](https://www.redhat.com/en/technologies/jboss-middleware/3scale).

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator/blob/v0.8.0/doc/operator-user-guide.md#upgrading-APIcast) for more information.

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/apicast-operator/tree/v0.8.0).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: APIcast
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://github.com/3scale/apicast-operator/tree/v0.8.0
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/apicast-operator:v0.8.0
        version: 0.8.0
      entries:
      - name: apicast-community-operator.v0.8.0
        version: 0.8.0
      - name: apicast-community-operator.v0.7.1
        version: 0.7.1
      - name: apicast-community-operator.v0.6.0
        version: 0.6.0
      - name: apicast-community-operator.v0.5.1
        version: 0.5.1
      name: alpha
    - currentCSV: apicast-community-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/apicast-operator:v0.8.0
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway built on top of NGINX. It is part
            of the Red Hat 3scale API Management Platform
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |
          The APIcast Operator creates and maintains the Red Hat 3scale API Gateway in several deployment configurations.

          APIcast is an API gateway built on top of [NGINX](https://www.nginx.com/). It is part of the [Red Hat 3scale API Management Platform](https://www.redhat.com/en/technologies/jboss-middleware/3scale).

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator/blob/v0.8.0/doc/operator-user-guide.md#upgrading-APIcast) for more information.

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/apicast-operator/tree/v0.8.0).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: APIcast
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://github.com/3scale/apicast-operator/tree/v0.8.0
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/apicast-operator:v0.8.0
        version: 0.8.0
      entries:
      - name: apicast-community-operator.v0.8.0
        version: 0.8.0
      - name: apicast-community-operator.v0.7.1
        version: 0.7.1
      - name: apicast-community-operator.v0.6.0
        version: 0.6.0
      - name: apicast-community-operator.v0.5.1
        version: 0.5.1
      name: stable
    defaultChannel: stable
    packageName: apicast-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-appsec
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-appsec.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Appsec",
                "metadata": {
                  "name": "appsec-sample"
                },
                "spec": {
                  "appsecDataModel": {
                    "port": 8080
                  },
                  "appsecResourceService": {
                    "port": 8080
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "remoteRepository": "docker.io",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "reportTemplates": {
                    "enabled": false
                  },
                  "systemMonitor": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:39:34Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Appsec
            name: appsecs.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP APPSEC Operator
        displayName: Keysight WAP APPSEC Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Appsec
          url: https://keysight-wap-appsec.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-appsec@sha256:e39ff225c26b750f86553693dda47d0dcc9834a4ab0975aea154361ee3527314
        version: 0.1.21
      entries:
      - name: keysight-wap-appsec.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-appsec
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-operator.v23.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerify",
                "metadata": {
                  "name": "ibmsecurityverify-sample"
                },
                "spec": {
                  "clientSecret": "--secret--",
                  "sessionLifetime": 3600,
                  "ssoPath": "/verify-sso"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isv-saas/verify-operator@sha256:c3bed2e5db48e54884d2278d24ca4864f389a56a667db481c627d01ab5ffaba7
          createdAt: "2023-03-22T02:49:11Z"
          description: The IBM Security Verify operator can consistently enforce policy-driven
            security, including authentication and authorization, using the Ingress
            networking capability of OpenShift.
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.9"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM-Security/verify-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerify is the Schema for the ibmsecurityverifies
              API.
            displayName: IBMSecurity Verify
            kind: IBMSecurityVerify
            name: ibmsecurityverifies.ibm.com
            version: v1
        description: "[IBM Security Verify](https://www.ibm.com/products/verify-saas)
          is an Identity-as-a-Service platform that allows IT, security and business
          leaders to protect their digital users, assets and data in a hybrid, multi-cloud
          world by enabling technical agility and operational efficiency. IBM Security
          Verify SaaS provides single sign-on (SSO), multi-factor authentication (MFA),
          AI-powered context for risk-based authentication for adaptive access decisions,
          user management, access recertification campaigns and identity analytics.\n\nAs
          part of using Red Hat OpenShift, you are entitled to Verify SaaS SSO for
          unlimited apps and users.  If you do not already have a Verify SaaS SSO
          tenant a new one can be [created](https://www.ibm.com/account/reg/signup?formid=urx-51255).\n\nFor
          a detailed description of IBM Security Verify refer to the [Official documentation](https://www.ibm.com/docs/en/security-verify).\n\nThe
          IBM Security Verify operator can consistently enforce policy-driven security
          by using the Ingress networking capability of OpenShift in conjunction with
          the [Nginx Ingress operator](https://www.nginx.com/blog/getting-started-nginx-ingress-operator-red-hat-openshift/).
          With this approach, you can enforce authentication and authorization policies
          for all of the applications in your cluster at the same time, without ever
          changing your application code. You can also dynamically register your application
          to start protecting them centrally from the cloud via Verify SaaS. \n\nSee
          the project [Readme](https://github.com/IBM-Security/verify-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Security Verify Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Product Information
          url: https://www.ibm.com/products/verify-saas
        - name: Verify Documentation
          url: https://www.ibm.com/docs/en/security-verify
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Development Team
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - icr.io/isv-saas/verify-operator@sha256:c3bed2e5db48e54884d2278d24ca4864f389a56a667db481c627d01ab5ffaba7
        version: 23.3.0
      entries:
      - name: ibm-security-verify-operator.v23.3.0
        version: 23.3.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hitachi
      provider-url: ""
    name: hspc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hspc-operator.v1.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.hitachi.com/v1",
                "kind": "HSPC",
                "metadata": {
                  "name": "hspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.hitachivantara.com/hitachicsi-oci-oss/hspc-operator@sha256:4dda4a43cefc19c8b1992756b499b1f9300920c5250dd77379f523997cfd5d54
          createdAt: "2025-06-13T10:30:39Z"
          description: An operator for managing Hitachi Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: Hitachi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HSPC is the Schema for the hspcs API
            displayName: HSPC
            kind: HSPC
            name: hspcs.csi.hitachi.com
            version: v1
        description: |-
          ## About
          Hitachi Storage Plug-in for Containers is a software plugin component for Kubernetes and OpenShift environments that is used to create and manage persistent block storage from Hitachi Virtual Storage Platform One storages.


          For full documentation, go to our [Product Documentation site](https://docs.hitachivantara.com/home) and refer to the reference guide for HSPC v3.16.0.

          ## Requirements

          ### Supported Driver Version

          * HSPC v3.16.0
        displayName: Hitachi Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hitachi
        - Storage
        - Container Storage Interface
        - CSI
        links:
        - name: Product Documentation
          url: https://docs.hitachivantara.com/home
        maintainers:
        - email: Container-info@hitachivantara.com
          name: Hitachi
        maturity: stable
        provider:
          name: Hitachi
        relatedImages:
        - registry.hitachivantara.com/hitachicsi-oci-oss/hspc-operator@sha256:4dda4a43cefc19c8b1992756b499b1f9300920c5250dd77379f523997cfd5d54
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - registry.hitachivantara.com/hitachicsi-oci-oss/hspc-csi-driver@sha256:2c580d98be55fffd8f6083379b6f23cbfe2aeb42fc4aec9e890072bc831c87c4
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:81d32545fb1b2c319bfeb36a70db4b4d3aea9c51c0574b209b8f64750b32d2f0
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        version: 1.16.0
      entries:
      - name: hspc-operator.v1.16.0
        version: 1.16.0
      name: stable
    defaultChannel: stable
    packageName: hspc-operator
    provider:
      name: Hitachi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CyberArk
      provider-url: https://www.cyberark.com/
    name: conjur-follower-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: conjur-follower-operator.v2.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:75fb5de2c111084ffcc3d973e6c63f1adfae78e29f6fd02152419fbca553ce7b",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:bee2cd31b3620a31e98f9d21326a2cdcecb36479099b06cc9660e4e83354d940",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:383776c23556a0a327a0bacc843d22c4b1516082ddd0cfb8a8275500f2eb841a",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:a9b109c7f3bfba5553aa4366338fa211896564f8e33856d75b4918acf8429f1e",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:b3096c17e7f3b9d723c906619dde293ea084e78c0da48c534335a8eee07b23c7",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:680a2044bc0e94fbc9376cd1eed9d067de91657f566c9752654480c939447841",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:ed53aea1866b1e2262386ac38cb9e1e47cc3f9047ae47f49070e2f506af6ca5a"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0 <2.5.3'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:cf46bdb4238cdb1dc51329822da55b7836958be2344d26f9e1819c0aa97fece6
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:383776c23556a0a327a0bacc843d22c4b1516082ddd0cfb8a8275500f2eb841a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:75fb5de2c111084ffcc3d973e6c63f1adfae78e29f6fd02152419fbca553ce7b
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:bee2cd31b3620a31e98f9d21326a2cdcecb36479099b06cc9660e4e83354d940
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:a9b109c7f3bfba5553aa4366338fa211896564f8e33856d75b4918acf8429f1e
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:b3096c17e7f3b9d723c906619dde293ea084e78c0da48c534335a8eee07b23c7
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:680a2044bc0e94fbc9376cd1eed9d067de91657f566c9752654480c939447841
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:ed53aea1866b1e2262386ac38cb9e1e47cc3f9047ae47f49070e2f506af6ca5a
        version: 2.5.3
      entries:
      - name: conjur-follower-operator.v2.5.3
        version: 2.5.3
      name: "2.5"
    - currentCSV: conjur-follower-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:c5337633eb3cc272f989e3e70c2e9bace9c819bba9e817537c7a691a8b6ee596
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813
        version: 2.6.3
      entries:
      - name: conjur-follower-operator.v2.6.3
        version: 2.6.3
      - name: conjur-follower-operator.v2.6.2
        version: 2.6.2
      - name: conjur-follower-operator.v2.6.1
        version: 2.6.1
      - name: conjur-follower-operator.v2.6.0
        version: 2.6.0
      - name: conjur-follower-operator.v2.5.0
        version: 2.5.0
      - name: conjur-follower-operator.v2.4.4
        version: 2.4.4
      - name: conjur-follower-operator.v2.4.3
        version: 2.4.3
      - name: conjur-follower-operator.v2.4.2
        version: 2.4.2
      - name: conjur-follower-operator.v2.3.5
        version: 2.3.5
      - name: conjur-follower-operator.v2.3.4
        version: 2.3.4
      - name: conjur-follower-operator.v2.3.1-290
        version: 2.3.1
      - name: conjur-follower-operator.v2.3.0
        version: 2.3.0
      - name: conjur-follower-operator.v2.2.4
        version: 2.2.4
      - name: conjur-follower-operator.v2.2.2
        version: 2.2.2
      - name: conjur-follower-operator.v2.2.1
        version: 2.2.1
      name: preview
    - currentCSV: conjur-follower-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:c5337633eb3cc272f989e3e70c2e9bace9c819bba9e817537c7a691a8b6ee596
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f
        version: 2.6.3
      entries:
      - name: conjur-follower-operator.v2.6.3
        version: 2.6.3
      - name: conjur-follower-operator.v2.6.2
        version: 2.6.2
      - name: conjur-follower-operator.v2.6.1
        version: 2.6.1
      - name: conjur-follower-operator.v2.6.0
        version: 2.6.0
      - name: conjur-follower-operator.v2.5.0
        version: 2.5.0
      - name: conjur-follower-operator.v2.4.4
        version: 2.4.4
      - name: conjur-follower-operator.v2.4.3
        version: 2.4.3
      - name: conjur-follower-operator.v2.4.2
        version: 2.4.2
      - name: conjur-follower-operator.v2.3.5
        version: 2.3.5
      - name: conjur-follower-operator.v2.3.4
        version: 2.3.4
      - name: conjur-follower-operator.v2.3.1-290
        version: 2.3.1
      - name: conjur-follower-operator.v2.3.0
        version: 2.3.0
      - name: conjur-follower-operator.v2.2.4
        version: 2.2.4
      - name: conjur-follower-operator.v2.2.2
        version: 2.2.2
      - name: conjur-follower-operator.v2.2.1
        version: 2.2.1
      name: stable
    defaultChannel: stable
    packageName: conjur-follower-operator
    provider:
      name: CyberArk
      url: https://www.cyberark.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/file-integrity-operator
    name: file-integrity-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: file-integrity-operator.v0.1.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <0.1.13'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator
          url: https://file-integrity-operator.domain
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/openshift4/file-integrity-rhel8-operator@sha256:a65f1244636649734763164a5d6155a6f6a7b79b3950ecc03fb3c58188583599
        version: 0.1.13
      entries:
      - name: file-integrity-operator.v0.1.13
        version: 0.1.13
      name: "4.7"
    - currentCSV: file-integrity-operator.v0.1.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <0.1.32'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:c3e34e6198eadb48c007cd949d8be38e26ef7d15d2f9047a8c3c89c03201e4ea
        version: 0.1.32
      entries:
      - name: file-integrity-operator.v0.1.32
        version: 0.1.32
      name: release-0.1
    - currentCSV: file-integrity-operator.v1.3.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          createdAt: "2024-11-21T18:02:08Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.3.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:f7039c6aaaa3e60dc4a3965cc1341c5217837e4b4030eb943771967e0b206578
        version: 1.3.6
      entries:
      - name: file-integrity-operator.v1.3.6
        version: 1.3.6
      - name: file-integrity-operator.v1.3.5
        version: 1.3.5
      - name: file-integrity-operator.v1.3.4
        version: 1.3.4
      - name: file-integrity-operator.v1.3.3
        version: 1.3.3
      name: stable
    - currentCSV: file-integrity-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <1.0.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:dd1dfeccf59214a0de653c3a4c4dbbc8c41640954318d13386cda73891660e4f
        version: 1.0.0
      entries:
      - name: file-integrity-operator.v1.0.0
        version: 1.0.0
      name: v1
    defaultChannel: stable
    packageName: file-integrity-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/file-integrity-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: resource-locker-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: resource-locker-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "ResourceLocker",
                "metadata": {
                  "name": "combined"
                },
                "spec": {
                  "patches": [
                    {
                      "id": "patch1",
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 0).metadata.name }}: {{ (index . 1).metadata.name }}\n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "kind": "Namespace",
                          "name": "resource-locker-test"
                        },
                        {
                          "apiVersion": "v1",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "resource-locker-test"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "resource-locker-test"
                      }
                    },
                    {
                      "id": "patch2",
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 0) }}: {{ (index . 1) }}\n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "fieldPath": "$.metadata.uid",
                          "kind": "Namespace",
                          "name": "resource-locker-test"
                        },
                        {
                          "apiVersion": "v1",
                          "fieldPath": "$.metadata.uid",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "resource-locker-test"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "resource-locker-test"
                      }
                    }
                  ],
                  "resources": [
                    {
                      "object": {
                        "apiVersion": "networking.k8s.io/v1",
                        "kind": "NetworkPolicy",
                        "metadata": {
                          "name": "allow-from-same-namespace",
                          "namespace": "resource-locker-test"
                        },
                        "spec": {
                          "ingress": [
                            {
                              "from": [
                                {
                                  "podSelector": {}
                                }
                              ]
                            }
                          ],
                          "podSelector": {}
                        }
                      }
                    },
                    {
                      "object": {
                        "apiVersion": "networking.k8s.io/v1",
                        "kind": "NetworkPolicy",
                        "metadata": {
                          "name": "allow-from-default-namespace",
                          "namespace": "resource-locker-test"
                        },
                        "spec": {
                          "ingress": [
                            {
                              "from": [
                                {
                                  "namespaceSelector": {
                                    "matchLabels": {
                                      "name": "default"
                                    }
                                  }
                                }
                              ]
                            }
                          ],
                          "podSelector": {}
                        }
                      }
                    }
                  ],
                  "serviceAccountRef": {
                    "name": "default"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/resource-locker-operator@sha256:2176db02f561c24012c5179495c89c3819500dff21e89a62052d465eb3314b44
          createdAt: "2021-11-30T19:38:40Z"
          description: This operator provides a facility to lock resources and / or
            patches into the state described by its custom resource preventing any
            drifts.
          operatorframework.io/suggested-namespace: resource-locker-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/resource-locker-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ResourceLocker Represents the intetion to create and enforce
              a set of resources and/or patches.
            displayName: Resource Locker
            kind: ResourceLocker
            name: resourcelockers.redhatcop.redhat.io
            version: v1alpha1
        description: |
          The resource locker operator allows you to specify a set of configurations that the operator will "keep in place" (lock) preventing any drifts.
          Two types of configurations may be specified:
          * Resources. This will instruct the operator to create and enforce the specified resource. In this case the operator "owns" the created resources.
          * Patches to resources. This will instruct the operator to patch- and enforce the change on- a pre-existing resource. In this case the operator does not "own" the resource.
          Locked resources are defined with the `ResourceLocker` CRD. Here is the high-level structure of this CRD:
          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: ResourceLocker
          metadata:
            name: test-simple-resource
          spec:
            resources:
            - object:
                apiVersion: v1
            ...
            patches:
            - targetObjectRef:
              ...
              patchTemplate: |
                metadata:
                  annotations:
                    ciao: hello
            ...
            serviceAccountRef:
              name: default
          ```
          It contains:
          * `resources`: representing an array of resources
          * `patches`: representing an array of patches
          * `serviceAccountRef`: a reference to a service account defined in the same namespace as the ResourceLocker CR, that will be used to create the resources and apply the patches. If not specified the service account will be defaulted to: `default`
          For each ResourceLocker a manager is dynamically allocated. For each resource and patch a controller with the needed watches is created and associated with the previously created manager.
          **Note**: creating a `ResourceLocker` CR will not work without due preparation (of the service account in particular). Follow these [instructions](https://github.com/redhat-cop/resource-locker-operator/blob/master/EXAMPLES.md) for a step by step walkthrough of the needed prerequisites to make this operator work.
        displayName: Resource Locker Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - resource_management
        - enforcing
        - patches
        - gitops
        links:
        - name: Resource Locker Operator
          url: https://github.com/redhat-cop/resource-locker-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/resource-locker-operator:latest
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/resource-locker-operator@sha256:2176db02f561c24012c5179495c89c3819500dff21e89a62052d465eb3314b44
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.1.3
      entries:
      - name: resource-locker-operator.v1.1.3
        version: 1.1.3
      - name: resource-locker-operator.v1.1.2
        version: 1.1.2
      - name: resource-locker-operator.v1.1.1
        version: 1.1.1
      - name: resource-locker-operator.v1.1.0
        version: 1.1.0
      - name: resource-locker-operator.v1.0.9
        version: 1.0.9
      name: alpha
    defaultChannel: alpha
    packageName: resource-locker-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Jason Skrzypek
      provider-url: https://github.com/shpwrck
    name: openshift-node-upgrade-mutex-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-node-upgrade-mutex-operator.v0.2.66
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexRule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutexrule-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutexrule",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutexrule-sample"
                },
                "spec": {
                  "jobName": "example",
                  "jobNamespace": "node-upgrade-mutex-operator",
                  "jobSpec": {
                    "selector": {},
                    "template": {
                      "metadata": {
                        "name": "mutex"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              "20"
                            ],
                            "image": "ubi9/toolbox",
                            "name": "mutex"
                          }
                        ],
                        "restartPolicy": "Never"
                      }
                    }
                  },
                  "type": "kubernetes"
                }
              },
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexRunConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutexrunconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutexrunconfig",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutexrunconfig-sample"
                },
                "spec": {
                  "canary": {
                    "name": "canary-example"
                  },
                  "mutexRules": [
                    {
                      "name": "mutexrule-sample",
                      "namespace": "node-upgrade-mutex-operator"
                    }
                  ],
                  "mutexTargets": [
                    {
                      "name": "mutextarget-sample",
                      "namespace": "node-upgrade-mutex-operator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexTarget",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutextarget-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutextarget",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutextarget-sample"
                },
                "spec": {
                  "machineConfigPools": [
                    "worker"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/jskrzypek/openshift-node-upgrade-mutex-operator
          createdAt: "2025-05-30T17:49:00Z"
          description: |
            This software injects required operations into the worker node upgrade process.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: node-upgrade-mutex-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          provider: Jason Skrzypek
          repository: https://github.com/shpwrck/openshift-upgrade-rollout-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A node mutex rule definition in the form of a kubernetes
              job spec
            displayName: Mutex Rule
            kind: MutexRule
            name: mutexrules.openshift-optional.skrzypek.io
            version: v1alpha1
          - description: Lists of Targets and Rules to apply to them, as well as an
              upgrade facilitating Machine Config Pool
            displayName: Mutex Run Config
            kind: MutexRunConfig
            name: mutexrunconfigs.openshift-optional.skrzypek.io
            version: v1alpha1
          - description: A list of Machine Config Pools to enforce upgrade procedure.
            displayName: Mutex Target
            kind: MutexTarget
            name: mutextargets.openshift-optional.skrzypek.io
            version: v1alpha1
        description: |
          ## Introduction

          This operator injects required operations in the form of a Kubernetes job into the worker node upgrade process using MachineConfigPools. In short, when provided a list of operations (`MutexRules`) and a list of MachineConfigPools (`MutexTargets`) , the OpenShift Node Upgrade Mutex Operator will execute each operation for a given target, enable updates for the given target by modifying it's MachineConfigPool membership, and wait until the node has been upgraded before proceeding onto subsequent operations and targets.

          ## Dependencies

          ### Release Version Label

          In order to confirm the successful upgrade of each target node, the NFD Operator is used to generate a version feature on every worker node. In order to proceed through given targets, each node must have this label
          `'machineconfiguration.openshift.io/release-image-version'` with a key providing the semantic version of this node (e.g. `'4.16.4'`)

          There are several ways to accomplish this, but an example MachineConfig has been provided [here](https://github.com/shpwrck/openshift-node-upgrade-mutex-operator/blob/main/machineconfig.yaml).

          ### Canary Machine Config Pool

          In addition to node labels, a MachineConfigPool that will be used to accomodate targets subject to upgrades must be provided matching the following criteria:
          * selects only worker nodes
          * is not paused

          Applying the example below will create a compatible canary pool:

          ```yaml
          ---
          apiVersion: machineconfiguration.openshift.io/v1
          kind: MachineConfigPool
          metadata:
            name: canary
          spec:
            machineConfigSelector:
              matchLabels:
                machineconfiguration.openshift.io/role: worker
            nodeSelector:
              matchLabels:
                node-role.kubernetes.io/worker: ''
              matchExpressions:
                - key: 'isCanary'
                  operator: Exists
            paused: false
          ```

          ## Usage / Custom Resources Definitions

          #### MutexRule:

          A resource storing a single rule referring to a template for a Kubernetes job to be created.
          This resource **does not** modify existing resources.

          ```yaml
          ---
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexRule
          metadata:
            name: k8s-example
          spec:
            jobName: "example" # Jobs will be run with this prefix name
            jobNamespace: "node-upgrade-mutex-operator" # Jobs will be run in this namespace
            jobSpec: # Accepts full 'batch/v1' job spec
              selector: {}
              template:
                metadata:
                  name: mutex
                spec:
                  containers:
                  - command:
                    - sleep
                    args:
                    - "30"
                    image: ubi9/toolbox
                    name: mutex
                  restartPolicy: Never
            cleanUpJobSpec: # Accepts full 'batch/v1' job spec (Optional)
              selector: {}
              template:
                metadata:
                  name: mutex
                spec:
                  containers:
                  - command:
                    - sleep
                    args:
                    - "30"
                    image: ubi9/toolbox
                    name: mutex
                  restartPolicy: Never
            type: kubernetes
          ```

          #### MutexTarget:

          A resource storing a set of MachineConfigPools to pause and incrementally upgrade.
          This resource **does not** modify existing resources.

          ```yaml
          ---
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexTarget
          metadata:
            name: worker-example
          spec:
            machineConfigPools:
              - worker # Name of existing machineconfigpool
          ```

          #### MutexRunConfig:

          A resource that references existing MutexRules and MutexTargets as well as an existing MachineConfigPool that will allow members to be upgraded. Every rule defined in the `mutexRules` list will be added as a mandatory task to every node that is a member of the each MachineConfigPool defined in the `mutexTargets` list.
          This resource **does** modify existing resources.

          ```yaml
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexRunConfig
          metadata:
            name: example
          spec:
            skipTLSVerify: false # Defaults to false if omitted
            canary:
              name: canary # Name of unpaused machineconfigpool to modify
            mutexRules: # List of existing mutexRules
              - name: k8s-example
                namespace: node-upgrade-mutex-operator
            mutexTargets: # List of existing mutexTargets
              - name: worker-example
                namespace: node-upgrade-mutex-operator
          ```

          ## Procedure

          #### MutexRule

          1. Upon creation/modification, the schema of each K8s rule is validated.
          2. Status is updated to indicate valid or invalid configuration.

          #### MutexTargets

          1. Upon creation/modification, the existence of each target is validated.
          2. Status is updated to indicate valid or invalid configuration.
          3. The nodes assigned to each target are added to the status field `nodes`.

          #### MutexRunConfigs

          1. Upon creation/modification, the existence as well as validation status of each rule, target, and canary pool is confirmed.
          2. If valid, the existence of an upgrade is queried. An upgrade is available if the following is true:
              * The `status.history` of the `cluster` ClusterVersion equals `Completed`
              * The `release-image-version` of any worker node is less than the `spec.desiredUpdate.version` of the `cluster` ClusterVersion.
          3. If an update exists, each target MachineConfigPool has `spec.paused` set to `true`.
          4. If each pool is paused, each node is updated to include a label for each rule with the following format:
              * `mutexrule.openshift-optional.skrzypek.io/{{MutexRule_Name}}.{{MutexRule_Namespace}}`
          5. If labels have been applied, the provided canary pool is modified to exclude nodes with labels provided by step (4).
          6. Assuming steps (2-5) have completed successfully, then the process for running jobs and updating nodes begins. The process per node is as follows:
              * Launch each K8S job sequentially
              * Remove the current node's rule labels
              * Wait until the `release-image-version` of the node matches the `spec.desiredUpdate.version` of the `cluster` ClusterVersion
              * Launch each K8S cleanup job sequentially
          7. If the update succeeds, all machine config pools are unpaused and all match expressions are removed.

          ## Permissions Granted

          In addition to the permissions required to operate the included CRDs, the following access is granted to the operator's service account.

          | Resource | Permissions |
          | :---: | :---: |
          | Jobs | Create & Read |
          | Nodes | Read & Update |
          | ClusterVersion | Read |
          | MachineConfigPools | Read & Update |

          ## Notes

          * This operator was built using the operator-sdk's ansible plugin and therefore has all of the inherited capabilities and restrictions.
          * This operator has undergone minimal testing and should not be run in production environments without acknowledging the risks.
          * This operator is currently alpha and subject to change based on requirements.

          ## Additional Information
          For additional information refer to the [GitHub Repository](https://github.com/shpwrck/openshift-node-upgrade-mutex-operator)
        displayName: OpenShift Node Upgrade Mutex Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - OpenShift Optional
        - Upgrade
        - Mutex
        links:
        - name: Openshift Node Upgrade Mutex Operator
          url: https://github.com/shpwrck/openshift-node-upgrade-mutex-operator
        maintainers:
        - email: jskrzype@redhat.com
          name: Jason Skrzypek
        maturity: alpha
        minKubeVersion: 1.27.0
        provider:
          name: Jason Skrzypek
          url: https://github.com/shpwrck
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - quay.io/jskrzypek/openshift-node-upgrade-mutex-operator@sha256:d4c28e8b256c00429d9999595c2386f1a25de597ee8e6c33871bb7b22b84c4af
        version: 0.2.66
      entries:
      - name: openshift-node-upgrade-mutex-operator.v0.2.66
        version: 0.2.66
      - name: openshift-node-upgrade-mutex-operator.v0.2.65
        version: 0.2.65
      - name: openshift-node-upgrade-mutex-operator.v0.2.60
        version: 0.2.60
      - name: openshift-node-upgrade-mutex-operator.v0.2.59
        version: 0.2.59
      - name: openshift-node-upgrade-mutex-operator.v0.2.53
        version: 0.2.53
      - name: openshift-node-upgrade-mutex-operator.v0.2.38
        version: 0.2.38
      - name: openshift-node-upgrade-mutex-operator.v0.2.34
        version: 0.2.34
      - name: openshift-node-upgrade-mutex-operator.v0.2.15
        version: 0.2.15
      - name: openshift-node-upgrade-mutex-operator.v0.2.14
        version: 0.2.14
      - name: openshift-node-upgrade-mutex-operator.v0.1.22
        version: 0.1.22
      name: alpha
    defaultChannel: alpha
    packageName: openshift-node-upgrade-mutex-operator
    provider:
      name: Jason Skrzypek
      url: https://github.com/shpwrck
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Digital.ai
      provider-url: https://digital.ai
    name: digitalai-deploy-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: digitalai-deploy-operator.v25.3.0-beta.813
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "quorum",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.0-debian-12-r4"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.0-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.5.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.2-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabsunsupported/deploy-operator@sha256:9ccca8ab999d4b6058c979bbb70568a63cf43a257ae65a787b2ff4d9bf470c39 '
          createdAt: "2025-09-02T05:43:37Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabsunsupported/deploy-operator@sha256:9ccca8ab999d4b6058c979bbb70568a63cf43a257ae65a787b2ff4d9bf470c39
        - docker.io/xebialabsunsupported/xl-deploy@sha256:41feb1d24d70a0c78a560b338d05c96c5383d570e3ef1ff37daff0eb8c92b386
        - docker.io/xebialabsunsupported/deploy-task-engine@sha256:15322f2263c2b9b70336177aec170e59497fa320281c2d73605918b30272b8f1
        - docker.io/xebialabsunsupported/central-configuration@sha256:3a35c2aada71a7f5ab8fe7d5c68441fe58b0b2c778b6ea85ac6f3f7b351f80ad
        version: 25.3.0-beta.813
      entries:
      - name: digitalai-deploy-operator.v25.3.0-beta.813
        version: 25.3.0-beta.813
      - name: digitalai-deploy-operator.v25.3.0-beta.711
        version: 25.3.0-beta.711
      - name: digitalai-deploy-operator.v25.3.0-beta.609
        version: 25.3.0-beta.609
      - name: digitalai-deploy-operator.v25.3.0-beta.512
        version: 25.3.0-beta.512
      name: beta
    - currentCSV: digitalai-deploy-operator.v24.1.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "ingressClass": "haproxy-daid",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "tag": "1.26.0-debian-12-r1"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "tag": "1.10.1-debian-12-r5"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "tag": "15.6.0-debian-12-r7"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "tag": "3.12.13-debian-12-r2"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator without SCC creation and with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:2b3413cd41f2b437e6027c8ae2cff12d1113e0ac1be69fc24840d68a05d3a665 '
          createdAt: "2024-10-16T05:39:21Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind:
          DigitalaiDeploy\nmetadata:\n  name: daid-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  securityContextConstraints:\n
          \   enabled: false\n  centralConfiguration:\n    replicaCount: 1\n    podSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  master:\n    replicaCount: 1\n    persistence:\n
          \     storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n      runAsUser:
          null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  worker:\n    replicaCount: 1\n    persistence:\n
          \     storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n      runAsUser:
          null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  route:\n    enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n        runAsUser: null\n        runAsGroup: null\n
          \       fsGroup: null\n      containerSecurityContext:\n        enabled:
          false\n        runAsUser: null\n        runAsGroup: null\n      securityContextConstraints:\n
          \       enabled: false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n
          \   install: true\n    persistence:\n      size: 1Gi\n      storageClass:
          ''\n    replicaCount: 1\n    podSecurityContext:\n      enabled: false\n
          \     runAsUser: null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     enabled: false\n      runAsUser: null\n      runAsGroup: null\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"daid-doc-digitalai-deploy-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xld-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Embedded PostgreSQL for
          DB management\n- Embedded RabbitMQ for message queue management\n\n### Security
          Configuration Details\n\n- `securityContextConstraints.enabled: false` -
          disables creation of SCCs;\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/deploy-operator@sha256:2b3413cd41f2b437e6027c8ae2cff12d1113e0ac1be69fc24840d68a05d3a665
        - docker.io/xebialabs/deploy-task-engine@sha256:8a381fd84483ea919fa7f8bee5cf19d9dec0d1d61cce73ec6b7c56fa504c6c93
        - docker.io/xebialabs/xl-deploy@sha256:77fdf6a3909367b060eba2ef698baddb77e4086d4d977f9c617d1e26a4427472
        - docker.io/xebialabs/central-configuration@sha256:83359e49c66e4fab502cc75663b456eb37d81d7cb512f2c48667f4d6a324e0df
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:771a9a173e033a3ad8b46f5c00a7036eaa88c8d8d1fbd89217325168998113ea
        version: 24.1.10
      entries:
      - name: digitalai-deploy-operator.v24.1.10
        version: 24.1.10
      name: stable
    - currentCSV: digitalai-deploy-operator.v24.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "image": {
                        "tag": "v0.14.7"
                      },
                      "ingressClass": "haproxy-daid",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.26.2-debian-12-r10"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "15.6.0-debian-12-r20"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.12.14-debian-12-r7"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator without SCC creation and with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:7d505a855728c761433ef7d46f0a64ffe2b4192f4819bad86c23582375099ac9 '
          createdAt: "2025-08-25T04:53:23Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind:
          DigitalaiDeploy\nmetadata:\n  name: daid-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  securityContextConstraints:\n
          \   enabled: false\n  centralConfiguration:\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  master:\n    replicaCount:
          1\n    persistence:\n      storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  worker:\n    replicaCount:
          1\n    persistence:\n      storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  route:\n    enabled: false\n
          \   annotations:\n      haproxy.router.openshift.io/cookie_name: SESSION_XLD\n
          \     haproxy.router.openshift.io/disable_cookies: \"false\"\n      haproxy.router.openshift.io/rewrite-target:
          /\n    hostname: '<mandatory-deploy-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n      securityContextConstraints:\n
          \       enabled: false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n
          \   install: true\n    persistence:\n      size: 1Gi\n      storageClass:
          ''\n    replicaCount: 1\n    podSecurityContext:\n      enabled: false\n
          \   containerSecurityContext:\n      enabled: false\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"daid-doc-digitalai-deploy-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xld-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Embedded PostgreSQL for
          DB management\n- Embedded RabbitMQ for message queue management\n\n### Security
          Configuration Details\n\n- `securityContextConstraints.enabled: false` -
          disables creation of SCCs;\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/xl-deploy@sha256:c2fe209f198fb2f42dc78a2e1544c026a6891d60682b4e47634234bbf7346540
        - docker.io/xebialabs/central-configuration@sha256:581a0bc71ae3652ceb643d81927fcf8cabea92f1d301f5d354076baff32fbcde
        - docker.io/xebialabs/deploy-operator@sha256:7d505a855728c761433ef7d46f0a64ffe2b4192f4819bad86c23582375099ac9
        - docker.io/xebialabs/deploy-task-engine@sha256:a97d67ea53e1b954ffb4d9514fd61426dcc0cab08922009eef02f32162026d6a
        version: 24.1.21
      entries:
      - name: digitalai-deploy-operator.v24.1.21
        version: 24.1.21
      - name: digitalai-deploy-operator.v24.1.20
        version: 24.1.20
      - name: digitalai-deploy-operator.v24.1.19
        version: 24.1.19
      - name: digitalai-deploy-operator.v24.1.18
        version: 24.1.18
      - name: digitalai-deploy-operator.v24.1.17
        version: 24.1.17
      - name: digitalai-deploy-operator.v24.1.14
        version: 24.1.14
      - name: digitalai-deploy-operator.v24.1.10
        version: 24.1.10
      name: stable-v24.1
    - currentCSV: digitalai-deploy-operator.v24.3.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "16.3.0-debian-12-r23"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n{{ if eq .Values.queueType \"quorum\" -}}\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n{{- end }}\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.13.7-debian-12-r5"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "queueType": "classic",
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:41ffc26c1d99a808e537f2366d92473ecf76dbffbac699dfa04785c3016f7470 '
          createdAt: "2025-09-26T18:18:41Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/central-configuration@sha256:4b83d4e9a523ae3615579907c65790478a3b22315ca5afce594793424d1d33b1
        - docker.io/xebialabs/deploy-operator@sha256:41ffc26c1d99a808e537f2366d92473ecf76dbffbac699dfa04785c3016f7470
        - docker.io/xebialabs/xl-deploy@sha256:8536efdf530dde45b84b14191ebfdc7ec5ebb1844f84871032fe2d335b13908d
        - docker.io/xebialabs/deploy-task-engine@sha256:75fe942305bf3bb403eb243b157b88844363eb702f03227a61e76ad4321bf0ee
        version: 24.3.11
      entries:
      - name: digitalai-deploy-operator.v24.3.11
        version: 24.3.11
      - name: digitalai-deploy-operator.v24.3.10
        version: 24.3.10
      - name: digitalai-deploy-operator.v24.3.9
        version: 24.3.9
      - name: digitalai-deploy-operator.v24.3.8
        version: 24.3.8
      - name: digitalai-deploy-operator.v24.3.7
        version: 24.3.7
      - name: digitalai-deploy-operator.v24.3.6
        version: 24.3.6
      - name: digitalai-deploy-operator.v24.3.5
        version: 24.3.5
      name: stable-v24.3
    - currentCSV: digitalai-deploy-operator.v25.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "quorum",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.1-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.4.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.3-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:d192ef9008d72c9792e1b7b988bdae8f1bbce45ed0e44afc48e3e268b657b38b '
          createdAt: "2025-09-26T16:30:53Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/deploy-task-engine@sha256:13bb779d6041836acb95c95f1eed75f4b5b2940d75687306183a9f8a2177cab3
        - docker.io/xebialabs/central-configuration@sha256:0c46275ffa3b251ccb76d2623e0e5a4cd1441bf7f94901845bda743f087bc6e4
        - docker.io/xebialabs/xl-deploy@sha256:e4212d20fad19f09cbe53796371bec7fc368cfe63316f1b7bccecf5faf6ea733
        - docker.io/xebialabs/deploy-operator@sha256:d192ef9008d72c9792e1b7b988bdae8f1bbce45ed0e44afc48e3e268b657b38b
        version: 25.1.5
      entries:
      - name: digitalai-deploy-operator.v25.1.5
        version: 25.1.5
      - name: digitalai-deploy-operator.v25.1.4
        version: 25.1.4
      - name: digitalai-deploy-operator.v25.1.3
        version: 25.1.3
      - name: digitalai-deploy-operator.v25.1.2
        version: 25.1.2
      - name: digitalai-deploy-operator.v25.1.1
        version: 25.1.1
      - name: digitalai-deploy-operator.v25.1.0
        version: 25.1.0
      name: stable-v25.1
    defaultChannel: stable-v24.3
    packageName: digitalai-deploy-operator
    provider:
      name: Digital.ai
      url: https://digital.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://access.redhat.com/products/red-hat-fuse
    name: fuse-online
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-online.v7.10.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:ed4faedcb97bb6a45f568586ff19da19da31ecbc304f12e0ae7978381247744a
          createdAt: 2022-01-31 15:55:46 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:ed4faedcb97bb6a45f568586ff19da19da31ecbc304f12e0ae7978381247744a
        - registry.redhat.io/rhel8/postgresql-12@sha256:44215c5b244b190c82d9d371a8ffd39b0c83576bd85c980cc72782a5cf9e6e1b
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:5048cd4fab2f98e6000bc60835c539cc40b70dbe265e927cfe185595347d8054
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:5dd1bc8f6b7e724736f4aeac86409bda3e0f64f62b736688e9572d14411232f1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:fcdd0292ebfc7e9edc36241f63b54a9d924a7605329c934ff2b10373af26dbbf
        - registry.redhat.io/openshift4/ose-prometheus@sha256:7f679d6c03535e300dac366f4bc6dd8d781d96dc47592836713f8bb76be36fa9
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:4f8849362e9ed6ff95d318a04e287b7ba33254a2cf7abdaf9b90cab7cbba6646
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:df2d55dd8092e9e94d034166890117cf5d1cfd5cf52ef0f90cfa2374cc9e2d06
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:3f0277e2c278740307c9b2d966f539ba12317e306198cb6c4c1c52a2ad9f4f40
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:96810c770f75057b8bba6b78a5ba414c8289315d8af1a387f681ae5e4fb7d6dd
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:ef229a4c8851ec495f28d0eeab83b137cd02698e283f76ea1b32271b7222621b
        version: 7.10.4
      entries:
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.10.x
    - currentCSV: fuse-online.v7.11.1-0.1687357787.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:658af3fc7819efb7cc39cd3a0504dcddc2ce7788c9b1e5178df869f7c9a742a8
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.11.1-0.1684718286.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:a2b1cd0256f30aff8fa37c332419b4483c0bfe34fb4f574ad8f18b0858969342
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:4de91e4527022e1b25df8d3835ddf778f773c20334314be9ba01a885c82287f5
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:8b9bd777bcbd20b1a1bbbc18274a5076a67bea1f928d772cb243c57fb1da7408
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:ac514973ec428d815d7a248d09d27f2a84f762aedfea63b3b48de9e633267ed5
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:658af3fc7819efb7cc39cd3a0504dcddc2ce7788c9b1e5178df869f7c9a742a8
        - registry.redhat.io/rhel8/postgresql-12@sha256:9248c4eaa8aeedacc1c06d7e3141ca1457147eef59e329273eb78e32fcd27e79
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:251c29e81f8e338f980b475b841647a977807908c388f4c86e18da7bff1b198a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:400a6ee03c3200c19b9b15c5d037a849633156a62e13922fb2e62838024ae69c
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:3d98d775bec85704c9a10c2f0d8e4f8116b7a41494f73c0b763e010262e5cd4d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        version: 7.11.1+0.1687357787.p
      entries:
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.11.x
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.12.x
    - currentCSV: fuse-online.v7.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta2",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:c70c5e76dc3db559c722bea2a767eb7dcba1eda786e1ef3a45c4cf32e1aca9ee
          createdAt: 2021-07-21 12:25:15 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta2
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to the product documentation at https://access.redhat.com/documentation/en-us/red_hat_fuse
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:d6df09b15599e776b49e1d46aa01728fd02d5629cee65e77dc277c0166554e0a
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:c668fe1f5d25baddf627acd79df4c8d2ab8b05309d3d41784ddc0a10dd5f34ad
        - registry.redhat.io/openshift4/ose-prometheus@sha256:13e9ea359cbcd3e0e94a6589b86b708dc84197f152701fd4092fffa234d11b58
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:60d5cf00411b03af3f1149a2bb3aeeb196c063d19625f876a094ae6baf46b54b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:f65600506db985689e5b7cd8b838e646d4268a0c4e272e19d7db3577a8dd7ac4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:75d5772c792a7c795281acec3d59e90021c109e9d7d1110868be0b882768aefb
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:c70c5e76dc3db559c722bea2a767eb7dcba1eda786e1ef3a45c4cf32e1aca9ee
        - registry.redhat.io/rhscl/postgresql-12-rhel7@sha256:c3e95e808eecba8ac57632452d3553509131ca783f62c68754a582cf5c55efab
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:eff9f8edf28ccbadcacab2022fff9e6bc93926af97521b49e3968eaf7b6fda65
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:d6f0c804c179fa559d3baed2e2d106c916a0439dc2215c4d34b878bc2e9bc281
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bafd85b0fd1256b8a60a2bb368a7e68f5dbad72718e5059a65660e01a425d8e2
        version: 7.9.3
      entries:
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.9.x
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: candidate
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: latest
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: stable
    defaultChannel: latest
    packageName: fuse-online
    provider:
      name: Red Hat
      url: https://access.redhat.com/products/red-hat-fuse
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: eG Innovations
      provider-url: https://www.eginnovations.com
    name: eginnovations-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eginnovations-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eginnovations.com/v1beta1",
                "kind": "EgUniversalAgent",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eginnovations-operator"
                  },
                  "name": "eguniversalagent",
                  "namespace": "egagent"
                },
                "spec": {
                  "agent": {
                    "env": [
                      {
                        "name": "EG_MANAGER",
                        "value": "Replace with eG Manager IP"
                      },
                      {
                        "name": "EG_MANAGER_PORT",
                        "value": "Replace with eG Manager Port"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "true"
          containerImage: docker.io/eginnovations/universal-agent-operator:0.0.11
          createdAt: "2025-09-11T20:08:00Z"
          description: eG Innovations provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with eG Universal Agent Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eginnovations/universal-agent-operator
          support: eG Innovations
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgUniversalAgent is the Schema for the eguniversalagents
              API
            displayName: Eg Universal Agent
            kind: EgUniversalAgent
            name: eguniversalagents.eginnovations.com
            version: v1beta1
        description: |
          eG Innovations’ converged application and infrastructure performance monitoring capabilities provide end-to-end visibility and correlated analytics for every layer, every tier of your IT landscape.

          The eG Enterprise Universal Agent Operator automatically configures a host agent on every Kubernetes worker node. The host agent auto-discovers the worker nodes and application containers running as Pods on each node and tracks their performance and utilization levels. In-depth monitoring of applications running on containers is also provided using the same host agent. No additional agents are required for the containers.

          ### Capabilities
            * Rolls out the egagent pod per node to monitor its pods and the node itself
            * Automatic BTM profiler injection using webhook based injection for deep application monitoring and transection tracing.

          ### Installation
            1. Create the namespace "egagent" for installing eG Agent on OpenShift Cluster.

               <br />``` $ oc create ns egagent ```
            2. Add "egagent" serviceaccount to scc( Security Context Constraints ) of OpenShift Cluster as privileged using below command.

               <br />``` $ oc adm policy add-scc-to-user privileged -z egagent -n egagent ```
            2. Install eG Innovations Universal Agent operator from [OpenShift Container Platform](https://www.openshift.com/) OperatorHub.
            3. Create the eG Universal Agent CRD (Custom Resource Definition)
               a. Click "View YAML Example" and copy the yaml contents and save it to (filename).yaml on your host.
               b. To apply the CRD run the following command

               <br />``` $ oc apply -f (filename).yaml ```

          ### Required Parameters
            * EG\_MANAGER, EG\_MANAGER\_PORT - provide the ip and port of your eG Manager Installation.

          ### Advanced Options
            * Disable Certificate Checking - disable any certificate validation that may interact poorly with proxies within your cluster
            * Image Override - use a copy of the eG Iniversal Agent container image from a registry other that Docker's or Red Hat's
            * Environment variables - define environment variables for the eG Universal Agent container of for APM profiler.
            * Namespace Selectors- select a subset of namesapces to enable instrumentation for applications.

          For complete list of supported parameters and the detailed documentation available in [eG Innovations Docs](https://docs.eginnovations.com)
        displayName: eG Innovations Universal Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Tracing
        links:
        - name: eG Innovations Kubernetes Monitoring Overview
          url: https://www.eginnovations.com/documentation/Kubernetes/Introduction-to-Monitoring-Kubernetes.htm
        maintainers:
        - email: devops@eginnovations.com
          name: eG Innovations
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: eG Innovations
          url: https://www.eginnovations.com
        relatedImages:
        - docker.io/eginnovations/universal-agent-operator:0.0.11
        version: 0.0.11
      entries:
      - name: eginnovations-operator.v0.0.11
        version: 0.0.11
      - name: eginnovations-operator.v0.0.10
        version: 0.0.10
      - name: eginnovations-operator.v0.0.9
        version: 0.0.9
      - name: eginnovations-operator.v0.0.8
        version: 0.0.8
      name: beta
    defaultChannel: beta
    packageName: eginnovations-operator
    provider:
      name: eG Innovations
      url: https://www.eginnovations.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: APIMatic.io
      provider-url: https://www.apimatic.io/
    name: apimatic-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apimatic-kubernetes-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicodegen.apimatic.io/v1beta1",
                "kind": "APIMatic",
                "metadata": {
                  "name": "apimatic-sample"
                },
                "spec": {
                  "licensespec": {
                    "licenseSourceName": "apimaticlicense",
                    "licenseSourceType": "ConfigMap"
                  },
                  "podspec": {
                    "apimaticContainerSpec": {
                      "image": "apimaticio/apimatic-codegen",
                      "imagePullSecret": "apimaticimagesecret"
                    }
                  },
                  "replicas": 3,
                  "servicespec": {
                    "apimaticserviceport": {
                      "nodePort": 32000,
                      "port": 8070
                    },
                    "servicetype": "NodePort"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: docker.io/apimaticio/apimatic-kubernetes-operator:v0.0.1
          createdAt: "2021-08-24 12:00:00"
          description: Generate client SDKs and interactive Documentation for your
            APIs in minutes
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/apimatic/apimatic-kubernetes-operator
          support: APIMatic.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIMatic is the Schema for the apimatics API
            displayName: APIMatic App
            kind: APIMatic
            name: apimatics.apicodegen.apimatic.io
            version: v1beta1
        description: "##About APIMatic \n\n APIMatic is a Developer Experience   platform
          for Web APIs that empowers users to create multi-language SDKs with API
          reference docs, dynamic code samples, assisted authentication and more -
          all auto-generated to provide a comprehensive developer experience beyond
          any API documentation tool.\n\n## Table of contents\n\n* [Introduction](#introduction)\n*
          [Features](#features)\n* [Running the Sample](#running-the-sample)\n  *
          [Prerequisites](#prerequisites)\n  * [Steps for Direct Deployment](#steps-for-direct-deployment)\n
          \ * [Steps for OLM Deployment](#steps-for-olm-deployment)\n* [Technical
          Support](#technical-support)\n* [Copyrights](#copyrights)\n\n## Introduction\n\nAPIMatic
          Operator simplifies the configuration and lifecycle management of the APIMatic
          code and docs generation solution on different Kubernetes distributions
          and OpenShift. The Operator encapsulates key operational knowledge on how
          to configure and upgrade the APIMatic CodeGen application, making it easy
          to get up and running.\n\n\nMore information about the underlying APIMatic
          CodeGen API that is exposed\nby this operator can be found [here](https://apimatic-core-v3-docs.netlify.app/#/http/getting-started/overview-apimatic-core).\n\n##
          Features\n\nAPIMatic Operator provides the following features:\n- Deploys
          the APIMatic CodeGen Web API service within the Kubernetes or OpenShift
          cluster.\n- Exposes the APIMatic CodeGen API external to the cluster, using
          Service type as [NodePort](https://kubernetes.io/docs/concepts/services-networking/service/#nodeport),
          [LoadBalancer](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          or [ExternalName](https://kubernetes.io/docs/concepts/services-networking/service/#externalname).\n-
          For exposing the service through an ingress resource, create an Ingress
          resource in the namespace of your APIMatic CR and set owned APIMatic service
          created by the operator as a backed service. More information can be found
          [here](https://kubernetes.io/docs/concepts/services-networking/ingress/).\n-
          Manual horizontal scaling of pods.\n  ```sh\n  kubectl scale apm apimatic-sample--replicas=2\n
          \ ```\n\n## Running the Sample \n\n### Prerequisites\n\nPlease contact APIMatic
          at [support@apimatic.io](mailto:support@apimatic.io) to register with the
          APIMatic CodeGen Docker Hub registry and acquire a valid license to run
          the APIMatic CodeGen API.\n\nFurther prerequisites for running the sample
          include:\n\n- [go](https://golang.org/) v1.16.*\n- [git](https://git-scm.com/)\n-
          [make](https://www.gnu.org/software/make/)\n- [Operator SDK](https://sdk.operatorframework.io/docs/overview/)\n-
          A running Kubernetes cluster with [kubectl](https://kubernetes.io/docs/reference/kubectl/overview/)
          on client. For testing purposes, you can use [Minikube](https://minikube.sigs.k8s.io/docs/)
          or [kind](https://kind.sigs.k8s.io/)\n- For checking the service created
          by the APIMatic operator on-prem, you can use [MetalLB](https://metallb.org/)\n\n###
          Steps for Direct Deployment\n\nTo run the sample for checking the APIMatic
          operator:\n\n- Clone the APIMatic repository into your working directory
          using the following command:\n  ```sh\n  git clone https://github.com/apimatic/apimatic-kubernetes-operator.git\n
          \ ```\n- Run `make deploy` to set up the APIMatic operator resources. This
          will deploy the '`apimatic-system`' namespace as well as the CRD and the
          RBAC manifests.\n\n- Create a secret named '`apimaticimagesecret`' to allow
          pulling the APIMatic CodeGen image using the Docker Hub username registered
          with APIMatic. If you have not done so, please contact support@apimatic.io
          for the steps required.\n  ```sh\n  kubectl create secret docker-registry
          apimaticimagesecret --docker-server=https://index.docker.io/v1/ --docker-username=<your-name>
          --docker-password=<your-pword> --docker-email=<your-email>\n  ```\n- Create
          a [configmap](https://kubernetes.io/docs/concepts/configuration/configmap/)
          resource named '`apimaticlicense`' that will serve as the volume storing
          the APIMatic license. An example of this is given below which will create
          the ConfigMap using the *License.lic* file located at */usr/local/apimatic/license/*\n
          \ ```sh\n  kubectl create configmap apimaticlicense --from-file /usr/local/apimatic/license/License.lic\n
          \ ```\n- This will deploy a ConfigMap resource with the following definition:\n
          \ ```sh\n  apiVersion: v1\n  data:\n    License.lic: \\\"<License file contents
          here>\\\"\n  kind: ConfigMap\n  metadata:\n    name: apimaticlicense\n    namespace:
          default\n  ```\n- Now run the sample using the following command:\n  ```sh\n
          \ kubectl apply -f config/samples/apicodegen_v1beta1_apimatic.yaml\n  ```\n-
          You will now see a new [Deployment](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/)
          with replica count of 3 and [Service](https://kubernetes.io/docs/concepts/services-networking/service/)
          of type `NodePort` created, both named ***apimatic-sample***. Accessing
          http://localhost:32000 from your browser (or using curl from within the
          Minikube or Kind cluster, if using that) should now show the APIMatic Titan
          page.\n- You can now use the exposed APIMatic CodeGen API to generate API
          SDKs and Docs using [curl](https://curl.se/), [Postman](https://www.postman.com/)
          or your own custom Web application that consumes the APIMatic CodeGen API
          service.\n\n- Once done, you can remove the APIMatic resources using the
          following command:\n  ```sh\n  make undeploy\n  ```\n\n### Steps for OLM
          Deployment\n\nThe following steps can be used to utilize [Operator LifeCycle
          Manager (OLM)](https://olm.operatorframework.io/docs/) to deploy the operator
          and run the sample. The steps are as follows:\n\n- If not already done so,
          clone the APIMatic repository into your working directory:\n  ```sh\n  git
          clone https://github.com/apimatic/apimatic-kubernetes-operator.git\n  ```
          \n- [Install OLM in your Kubernetes cluster](https://olm.operatorframework.io/docs/getting-started/#installing-olm-in-your-cluster).\n\n-
          Run the following script to install the resources required by OLM to deploy
          the APIMatic operator in the Kubernetes cluster within the `apimatic-system`
          namespace. Information about the different resources required can be found
          using the steps given [here](https://olm.operatorframework.io/docs/tasks/).\n
          \ ```sh\n  kubectl apply -f olm/manifests.yaml\n  ```\n\n- This should spin
          up the `ClusterServiceVersion` of the operator in the `apimatic-system`
          namespace, following which the operator pod will spin up. To ensure the
          that the operator installed successfully, check for the `ClusterServiceVersion`
          and the operator deployment in the namespace it was installed in.\n  ```sh\n
          \ kubectl get csv -n apimatic-system\n  \n  kubectl get deployment -n apimatic-system\n
          \ ```\n\n  - Create a secret named `apimaticimagesecret` to allow pulling
          the APIMatic CodeGen image using the Docker Hub username registered with
          APIMatic. If you have not done so, please contact support@apimatic.io for
          the steps required.\n  ```sh\n  kubectl create secret docker-registry apimaticimagesecret
          --docker-server=https://index.docker.io/v1/ --docker-username=<your-name>
          --docker-password=<your-pword> --docker-email=<your-email>\n  ```\n- Create
          a [configmap](https://kubernetes.io/docs/concepts/configuration/configmap/)
          resource named `apimaticlicense` that will serve as the volume storing the
          APIMatic license. An example of this is given below which will create the
          ConfigMap using the *License.lic* file located at */usr/local/apimatic/license/*\n
          \ ```sh\n  kubectl create configmap apimaticlicense --from-file /usr/local/apimatic/license/License.lic\n
          \ ```\n- This will deploy a ConfigMap resource with the following definition:\n
          \ ```sh\n  apiVersion: v1\n  data:\n    License.lic: \\\"<License file contents
          here>\\\"\n  kind: ConfigMap\n  metadata:\n    name: apimaticlicense\n    namespace:
          default\n  ```\n- Now run the sample using the following command:\n  ```sh\n
          \ kubectl apply -f config/samples/apicodegen_v1beta1_apimatic.yaml\n  ```\n\n-
          You will now see a new [Deployment](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/)
          with replica count of 3 and [Service](https://kubernetes.io/docs/concepts/services-networking/service/)
          of type `NodePort` created, both named ***apimatic-sample***. Accessing
          http://localhost:32000 from your browser (or using curl from within the
          Minikube or Kind cluster, if using that) should now show the APIMatic Titan
          page.\n- You can now use the exposed APIMatic CodeGen API to generate API
          SDKs and Docs using [curl](https://curl.se/), [Postman](https://www.postman.com/)
          or your own custom Web application that consumes the APIMatic CodeGen API
          service.\n- Once done, you can remove the APIMatic operator resources using
          the follow script:\n  ```sh\n  kubectl delete -f olm/manifests.yaml\n  ```\n\n##
          Technical Support\n\n- To request additional features in the future, or
          if you notice any discrepancy regarding this document, please drop an email
          to [support@apimatic.io](mailto:support@apimatic.io).\n\n### Copyrights\n\n&copy;
          2021 APIMatic.io\n"
        displayName: APIMatic Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Rest APIs
        - API SDK Generator
        - API Docs Generator
        - API Developer Portal
        - APIMatic
        - Developer Tools
        links:
        - name: Overview
          url: https://www.apimatic.io/
        - name: Developer Experience Portal
          url: https://www.apimatic.io/developer-experience-portal
        - name: Code-Gen as a Service
          url: https://www.apimatic.io/continuous-code-generation
        - name: API Transformer
          url: https://www.apimatic.io/transformer
        - name: Contact
          url: https://www.apimatic.io/contact
        - name: Documentation
          url: https://docs.apimatic.io/
        - name: Blogs
          url: https://www.apimatic.io/blog/
        maintainers:
        - email: support@apimatic.io
          name: Support
        maturity: alpha
        provider:
          name: APIMatic.io
          url: https://www.apimatic.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - docker.io/apimaticio/apimatic-kubernetes-operator:v0.0.1
        version: 0.0.1
      entries:
      - name: apimatic-kubernetes-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: apimatic-kubernetes-operator
    provider:
      name: APIMatic.io
      url: https://www.apimatic.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-openshift-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.14
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-06-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:154f85cc6721a70909dee755ddd88897ad51038afe2538b6dee4d26e7d8b9afb
        version: 0.0.14
      entries:
      - name: joget-dx8-openshift-operator.v0.0.14
        version: 0.0.14
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-openshift-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sumo Logic
      provider-url: https://www.sumologic.com/
    name: sumologic-kubernetes-collection-helm-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "helm-operator.sumologic.com/v1alpha1",
                    "kind": "SumologicCollection",
                    "metadata": {
                        "name": "collection"
                    },
                    "spec": {
                        "sumologic": {
                            "accessId": "",
                            "accessKey": "",
                            "clusterName": "kubernetes",
                            "scc": {
                                "create": true
                            }
                        },
                        "kube-prometheus-stack": {
                            "prometheus-node-exporter": {
                                "service": {
                                    "port": 9200,
                                    "targetPort": 9200
                                }
                            }
                        },
                        "otellogs": {
                            "daemonset": {
                                "containers": {
                                    "otelcol": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                },
                                "initContainers": {
                                    "changeowner": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                }
                            }
                        }
                    }
                }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing,Monitoring
          containerImage: registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
          createdAt: "2024-07-19T10:48:00Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/sumologic-kubernetes-collection-helm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/sumologic-kubernetes-collection-helm-operator-rhmp/support?utm_source=openshift_console
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator.git
          support: Sumo Logic
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configuration for Sumo Logic Kubernetes Collection
            displayName: SumologicCollection
            kind: SumologicCollection
            name: sumologiccollections.helm-operator.sumologic.com
            version: v1alpha1
        description: Helm Operator for the Sumo Logic Kubernetes Collection Helm Chart
          v4.13.0-2
        displayName: Sumo Logic Kubernetes Collection Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - collection
        - logging
        - metrics
        - monitoring
        links:
        - name: Sumo Logic Kubernetes Collection Helm Operator
          url: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator
        maintainers:
        - email: collection@sumologic.com
          name: Sumo Logic
        maturity: alpha
        provider:
          name: Sumo Logic
          url: https://www.sumologic.com/
        relatedImages:
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
        - registry.connect.redhat.com/sumologic/kube-state-metrics@sha256:3551ededa327a35db798ec8286cd843580e3ae395d2231e041e8c4b7621127ee
        - registry.connect.redhat.com/sumologic/metrics-server@sha256:65d2a3ddc4b4183f438fa18b655079c93c905cb9e3b8c54977336205dc1fff07
        - registry.connect.redhat.com/sumologic/prometheus@sha256:0ef6eaf570a40499dd5de92a79593e950bc771b014e0ee6ff7419b4b82a36fcc
        - registry.connect.redhat.com/sumologic/autoinstrumentation-python@sha256:f0fa45fa45e64c7df88f21181ae18ee96fe83cd3988258663d0806ea6f39e6bb
        - registry.connect.redhat.com/sumologic/kubernetes-setup@sha256:f07eaf68fa11ffc708e4485a83bd3763fc5712f7d73d29edfe5b9c9e08f05df6
        - registry.connect.redhat.com/sumologic/sumologic-otel-collector@sha256:4554d43dba204e3ec8e91b36442a438f10234fa0cb4db7fb7880b257a6485112
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:e1f49fcf5b7e86d9eb49c16dbbf9484cb486f55f999fd51d7c738c3b1f2666cf
        - registry.connect.redhat.com/sumologic/autoinstrumentation-dotnet@sha256:f4f385c4489c74b5c43f2fe6b6620a9eb8d9ecab470567236b3c45f636106640
        - registry.connect.redhat.com/sumologic/tailing-sidecar@sha256:738d653a78dc415243ab3627db9135768175fb7f51d0803566f5d9e824c8744b
        - registry.connect.redhat.com/sumologic/telegraf-operator@sha256:ff958222be01c0c3966474463e43bbd5bcf348790dd34bb761718cb1cea2d793
        - registry.connect.redhat.com/sumologic/autoinstrumentation-nodejs@sha256:c206eef4fdecaa58b4f01e5e7f0dbd74e9606eab2b4debf47461d36191e248d1
        - registry.connect.redhat.com/sumologic/node-exporter@sha256:11e9e604432a66d9b388cb1bf70eac077c7f0372d1264342ef927d76b8ef0272
        - registry.connect.redhat.com/sumologic/kubernetes-tools-kubectl@sha256:d3155b864cc74f327adedf7d013a8f0fcb11c2e73a5655b7cafa848824e80172
        - registry.connect.redhat.com/sumologic/thanos@sha256:16faf7191668e86d8b576bba055dd0cbdeb009635976b9e07bd00f926032f11c
        - registry.connect.redhat.com/sumologic/opentelemetry-operator@sha256:5587fe5b7ebfa57ad320ad2521103f908317388d1741c40e5f7be5ed9e5f9ad7
        - registry.connect.redhat.com/sumologic/nginx-unprivileged@sha256:7260555d16c94c4fe1e016626b4cb8a5c0de6b700e2eea28744331fb298f71ae
        - registry.connect.redhat.com/sumologic/prometheus-operator@sha256:58a329e6582456b32abebb470086191f8366b927af4703741613993465f116b8
        - registry.connect.redhat.com/sumologic/telegraf@sha256:75efd2e509c6017ea0fc5c9103ce068cbc89087b16506a4ef48cb5ff00eab84e
        - registry.connect.redhat.com/sumologic/tailing-sidecar-operator@sha256:c16b8b18e9b263fc3eb24059bba512bcfc3201b32105af6e5518392826ddd66c
        - registry.connect.redhat.com/sumologic/prometheus-config-reloader@sha256:0b605d36782afd8f50fd631d0e9c7b19d39ce962af89b7a737b2f09a8ceff291
        - registry.connect.redhat.com/sumologic/autoinstrumentation-java@sha256:242958b7d8905f4ac02f6ad38897ba58b9badf2a6213a389e67dcc71fee3630c
        - registry.connect.redhat.com/sumologic/busybox@sha256:209bc8e0b1d958fc699f3bb75b1248544e372ad35cdc46c991bd1698f641e1f9
        version: 4.13.0-2
      entries:
      - name: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
        version: 4.13.0-2
      name: alpha
    defaultChannel: alpha
    packageName: sumologic-kubernetes-collection-helm-operator-rhmp
    provider:
      name: Sumo Logic
      url: https://www.sumologic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-bamoe-businessautomation: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: bamoe-businessautomation-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bamoe-businessautomation-operator.8.0.8-1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.kiegroup.org/v2","kind":"KieApp","metadata":{"name":"rhpam-trial"},"spec":{"environment":"rhpam-trial"}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/ibm-bamoe/bamoe-rhel8-operator@sha256:0646da8707d10b3cc2e2344269abd8797db34127a3326bf76bbc1547aa4bbd73
          createdAt: "2025-07-22 17:40:17"
          description: Deploys and manages IBM Business Automation Manager Open Editions
            environment.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM Business Automation Manager
            Open Edition"]'
          repository: https://github.com/kiegroup/kie-cloud-operator
          support: https://www.ibm.com/mysupport
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A project prescription running an IBM BAMOE environment.
            displayName: KieApp
            kind: KieApp
            name: kieapps.app.kiegroup.org
            version: v2
        description: |-
          Deploys and manages IBM Business Automation Manager Open Editions environment.

          * **IBM Process Automation Manager** is a platform for developing containerized microservices and applications that automate business decisions and processes. It includes business process management (BPM), business rules management (BRM), and business resource optimization and complex event processing (CEP) technologies. It also includes a user experience platform to create engaging user interfaces for process and decision services with minimal coding.
        displayName: IBM Business Automation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kieapp
        - pam
        - decision
        - kie
        - cloud
        - bpm
        - process
        - automation
        - operator
        links:
        - name: Product Page
          url: https://ibm.com
        - name: Documentation
          url: https://www.ibm.com/docs/en/ibamoe
        maintainers:
        - email: mysphelp@us.ibm.com
          name: IBM
        maturity: 8.x-stable
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-monitoring-rhel8@sha256:ef1d77bb085172f927ee53d1674c85e2acaf1f580d7cad55bae8bf2c33c49c01
        - registry.redhat.io/openshift4/ose-cli@sha256:b2b93def4eb5edeccfa2d8ba22e493646325842674dff0446b45677bf641a342
        - registry.redhat.io/rhscl/postgresql-13-rhel7@sha256:338dd09681dcf860e6a976a3bc48a1474c2376d698815e4eb611ae9c5aa68966
        - registry.redhat.io/ibm-bamoe/bamoe-process-migration-rhel8@sha256:3fd45f2dabca7220772d823275a5697b42dd35f53c27671b093117933ef2d4f6
        - registry.redhat.io/rhscl/mysql-80-rhel7@sha256:e4f535c78bc2fdb58914c52fb805185035ff31515452b4df714f9ab623ba383a
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-rhel8@sha256:c3553707936cb9eee8a6a622643f72c91e1b0c11add5ec9ad5865242a9e133b4
        - registry.redhat.io/ibm-bamoe/bamoe-kieserver-rhel8@sha256:01bf97beffb3647968aae052cdd401d6d562c35f3e78ac483dbf55c44a71dc2f
        - registry.redhat.io/jboss-datagrid-7/datagrid73-openshift@sha256:4056b1cf8ff5995f56eecfdc087d9ece56a216dba6bd73acc5693f0778e8d573
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-rhel8@sha256:82d8ebde7eb3ecfbf72e5b71e7e392d3a3597fc084b6cf35e8fec23bfe41210d
        - registry.redhat.io/amq7/amq-broker@sha256:792c1cb19e62519cc47d834650fe5a6a6ca589a37d0b53d4f4549e80cabae3f9
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:98e9a6618558bc8a1b9cb91d1c9b3380968d335092f6e953d1abe6af96abf65c
        - registry.redhat.io/ibm-bamoe/bamoe-smartrouter-rhel8@sha256:2f1a7519617b5c0d082f943de0db295370649d968d699d76cfd829f4c9826a0f
        - registry.redhat.io/ibm-bamoe/bamoe-rhel8-operator@sha256:0646da8707d10b3cc2e2344269abd8797db34127a3326bf76bbc1547aa4bbd73
        - registry.redhat.io/ibm-bamoe/bamoe-process-migration-rhel8@sha256:a287a2fbaded371b8fcae5e4124b2d905bb228f94da0fc899ca2600b378dca0d
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-monitoring-rhel8@sha256:25025e489c15108edffe5270d257e1ed9be0a67aa067bcfc184647793aac2617
        - registry.redhat.io/ibm-bamoe/bamoe-dashbuilder-rhel8@sha256:efa55e17190b0ebe1618452f786876ef354b8fbec25c316aed7ed5bb1ce88116
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:457fc73423fa94591bc7e35729a25c58d3465583afccf342fb338b6e11188575
        - registry.redhat.io/ibm-bamoe/bamoe-smartrouter-rhel8@sha256:5ed8821328413ec11859ae317679f0692afff9945f134fc8909e07a3ac387794
        - registry.redhat.io/ibm-bamoe/bamoe-kieserver-rhel8@sha256:c9f08ddc6111cdd74faa0db5721d5730fc7716fccf4dc2a658595f7e0c184c10
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:f714b9c1555e7f27a3c9cf2cf687a74188f36b1283d4e09cec4db3c084d9fe64
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhscl/postgresql-12-rhel7@sha256:191237ccffc476e764714a69e35d9421dcee5ddd7184442d7a8306a06b97e6bb
        - registry.redhat.io/ibm-bamoe/bamoe-dashbuilder-rhel8@sha256:b3b7388b3b116ef37aabbb04b5d2c3e36e58d2fc95f1e59795b423010e071fc3
        - registry.redhat.io/ibm-bamoe/bamoe-controller-rhel8@sha256:3fcd9ae0bc194c198cbb5ee94272e80f302658733e467766bd1789357e9fd015
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:fb326be806b337b0ae1711f974d15e073aebcbad77aaa1123a63cf4f331da5f2
        - registry.redhat.io/ibm-bamoe/bamoe-controller-rhel8@sha256:6cd0f07856d612dca46b0a01a93b796f7a0f112b98cedfd6754f9fd770ee843d
        version: 8.0.8-1
      entries:
      - name: bamoe-businessautomation-operator.8.0.8-1
        version: 8.0.8-1
      - name: bamoe-businessautomation-operator.8.0.7-2
        version: 8.0.7-2
      - name: bamoe-businessautomation-operator.8.0.7-1
        version: 8.0.7-1
      - name: bamoe-businessautomation-operator.8.0.6-3
        version: 8.0.6-3
      - name: bamoe-businessautomation-operator.8.0.6-2
        version: 8.0.6-2
      - name: bamoe-businessautomation-operator.8.0.6-1
        version: 8.0.6-1
      - name: bamoe-businessautomation-operator.8.0.5-3
        version: 8.0.5-3
      - name: bamoe-businessautomation-operator.8.0.5-2
        version: 8.0.5-2
      - name: bamoe-businessautomation-operator.8.0.5-1
        version: 8.0.5-1
      - name: bamoe-businessautomation-operator.8.0.4-3
        version: 8.0.4-3
      - name: bamoe-businessautomation-operator.8.0.4-2
        version: 8.0.4-2
      - name: bamoe-businessautomation-operator.8.0.4-1
        version: 8.0.4-1
      - name: bamoe-businessautomation-operator.8.0.3-4
        version: 8.0.3-4
      - name: bamoe-businessautomation-operator.8.0.3-3
        version: 8.0.3-3
      - name: bamoe-businessautomation-operator.8.0.3-2
        version: 8.0.3-2
      - name: bamoe-businessautomation-operator.8.0.3-1
        version: 8.0.3-1
      - name: bamoe-businessautomation-operator.8.0.2-3
        version: 8.0.2-3
      - name: bamoe-businessautomation-operator.8.0.2-2
        version: 8.0.2-2
      - name: bamoe-businessautomation-operator.8.0.2-1
        version: 8.0.2-1
      - name: bamoe-businessautomation-operator.8.0.1-2
        version: 8.0.1-2
      - name: bamoe-businessautomation-operator.8.0.1-1
        version: 8.0.1-1
      - name: bamoe-businessautomation-operator.8.0.0-1
        version: 8.0.0-1
      name: 8.x-stable
    defaultChannel: 8.x-stable
    packageName: bamoe-businessautomation-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Odoo Community Association
      provider-url: ""
    name: odoo-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odoo-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"odoo-community.org/v1alpha1","kind":"Odoo","metadata":{"name":"example"},"spec":{"postgresqlVolumeSize":"4Gi","odooVolumeSize":"4Gi"}}]'
          capabilities: Full Lifecycle
          categories: Application Runtime
          certified: "false"
          containerImage: ursa/odoo-operator:0.0.2
          createdAt: "2021-04-26T12:00:00Z"
          description: Operator to manage Odoo instances within an OpenShift cluster
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          repository: https://github.com/ursais/odoo-operator
          support: Open Source Integrators
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Odoo
            displayName: Odoo
            kind: Odoo
            name: odoos.odoo-community.org
            version: v1alpha1
        description: |
          ## About Odoo

          [Odoo](https://odoo-community.org), is a suite of business management software tools including CRM, e-commerce, billing, accounting, manufacturing, warehouse, project management, and inventory management to name a few.
          ## About the Odoo Operator

          This is an Ansible-based operator created with operator-sdk v1.5.0.

          The upgrade is handled by the Odoo container itself using [Marabunta](https://pypi.org/project/marabunta/) and [Anthem](https://pypi.org/project/anthem/).

          ## Roadmap
          * Add horizontal pod autoscaler
          * Add cronjob for backup
          ## Prerequisites for enabling this Operator

          Please refer to the README file on the [GitHub project](https://github.com/ursais/odoo-operator).
        displayName: Odoo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ERP
        - Business Software
        links:
        - name: Odoo Community Association
          url: https://odoo-community.org
        - name: GitHub Repository
          url: https://github.com/ursais/odoo-operator
        - name: Open Source Integrators
          url: https://opensourceintegrators.com
        maintainers:
        - email: mchambreuil@opensourceintegrators.com
          name: Maxime Chambreuil
        maturity: alpha
        provider:
          name: Odoo Community Association
        relatedImages:
        - ursa/odoo-operator:0.0.2
        version: 0.0.2
      entries:
      - name: odoo-operator.v0.0.2
        version: 0.0.2
      - name: odoo-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: odoo-operator
    provider:
      name: Odoo Community Association
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fortanix Inc.
      provider-url: https://fortanix.com
    name: ccm-node-agent-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ccm-node-agent-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1",
                "kind": "CcmNodeAgent",
                "metadata": {
                  "name": "ccmnodeagent-sample"
                },
                "spec": {
                  "ccmJoinToken": {
                    "secretName": "ccm-join-token",
                    "token": ""
                  },
                  "fullnameOverride": "",
                  "image": {
                    "ccmNodeAgent": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/ccm-node-agent-ubi7",
                      "tag": "latest"
                    },
                    "ccmNodeAgentInit": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "busybox",
                      "tag": "latest"
                    },
                    "sgxSoftwareInstaller": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/sgx-software-installer-ubi7",
                      "tag": "latest"
                    }
                  },
                  "log": {
                    "debug": false
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "ports": {
                    "containerPort": 9092,
                    "hostPort": 9092
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CCM Node Agent software enables registration of the compute
              nodes to Fortanix CCM when installed in the Openshift Cluster Nodes<U+200B>.
            displayName: CCM Node Agent
            kind: CcmNodeAgent
            name: ccmnodeagents.charts.operatorhub.io
            version: v1
        description: "CCM Node Agent software enables registration of the compute
          nodes to Fortanix CCM when installed in the Openshift Cluster Nodes​. \nIt
          also assists with application attestation and visibility for Fortanix CCM.\n\n##
          About this Operator\nThis Operator is based on a Helm chart for CCM Node
          Agent. It enrolls all the nodes in the openshift cluster to a particular
          CCM account. You have to provide the base64 encode of join token of CCM
          account while creating CcmNodeAgent resource.\n\n## Prerequisites\nAll the
          worker nodes in the cluster must have SGX capabilities.\n\n## Installation\n*
          Choose a namespace to install the Operator. Operator will be available in
          all namespaces.\n* Select Approval Strategy: Available options are Automatic
          or Manual. Automatic approval will automatically upgrade the running instance
          of your Operator without human intervention when a new version is available.
          In case of Manual approval, when a new version is available, OLM will create
          an update request. You must manually approve the request to update the Operator
          to a new version.\n* Generate your CCM Account Join Token: Please log in
          to https://ccm.fortanix.com, and in the Management Console tab, click the
          + ENROLL NODE button. In the ENROLL NODE screen, a Join Token will be generated
          in the text box for \"Get a join token to register an SGX compute node\".
          This Join Token is used by the compute node to authenticate itself.\n* Base64
          encode the Join Token:\n```shell\necho -n <join-token> | openssl base64
          -A\n```\n* Create an instance of CcmNodeAgent by setting the `spec.ccmJoinToken.Token`
          to the base64 encoded CCM account join token.\n* After CcmNodeAgent resource
          is installed successfully you should be able to see all the worker nodes
          of the cluster in your CCM account under Compute Node Tab.\n\n## Getting
          Started\nNow to run your application inside SGX enclaves on your cluster
          follow the [documentation](https://support.fortanix.com/hc/en-us/articles/360043527431-User-s-Guide-Add-and-Edit-an-Application).\n"
        displayName: CCM Node Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dockerfile
        - Helm
        - CCM
        - Confidential Compute
        links:
        - name: CCM Node Agent Operator
          url: https://support.fortanix.com/hc/en-us/categories/360003107511-Confidential-Computing-Manager
        maintainers:
        - email: info@fortanix.com
          name: Fortanix Inc.
        maturity: alpha
        provider:
          name: Fortanix Inc.
          url: https://fortanix.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62575404000cb77fa8d2284c87b31e39e667a55b0fb47821263b23b3d1135a7e
        - docker.io/fortanix/ccm-node-agent-operator@sha256:e2a3d52551198102ddbe7f4ebd9621edc357c57ddeff47b7ed9f2cf56a1bed39
        version: 0.0.2
      entries:
      - name: ccm-node-agent-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ccm-node-agent-operator
    provider:
      name: Fortanix Inc.
      url: https://fortanix.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: https://intel.com
    name: intel-device-plugins-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: intel-device-plugins-operator.v0.29.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "GpuDevicePlugin",
                "metadata": {
                  "name": "gpudeviceplugin-sample"
                },
                "spec": {
                  "enableMonitoring": true,
                  "image": "registry.connect.redhat.com/intel/intel-gpu-plugin@sha256:963caf1e6e16b5e297671c536eec04dc57cf45f5cc5cf458bdcafa56bdb97972",
                  "preferredAllocationPolicy": "none",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/gpu": "true"
                  },
                  "sharedDevNum": 1
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "QatDevicePlugin",
                "metadata": {
                  "name": "qatdeviceplugin-sample"
                },
                "spec": {
                  "dpdkDriver": "vfio-pci",
                  "image": "registry.connect.redhat.com/intel/intel-qat-plugin@sha256:53617c3a72af169031cbaba1380cd6a25c9d2b50c9a2da822cf37b473c6e7a8f",
                  "initImage": "registry.connect.redhat.com/intel/intel-qat-initcontainer@sha256:14f282a0adae43262657f24ce99a620b671c9a2ce12a8216163e6b22ac6e252b",
                  "kernelVfDrivers": [
                    "c6xxvf",
                    "4xxxvf"
                  ],
                  "logLevel": 4,
                  "maxNumDevices": 128,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/qat": "true"
                  }
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "DsaDevicePlugin",
                "metadata": {
                  "name": "dsadeviceplugin-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/intel/intel-dsa-plugin@sha256:092575515739bf9a0f5982c3605b63382bc4d1250128269434c2bfe4a27ec99d",
                  "initImage": "registry.connect.redhat.com/intel/intel-idxd-config-initcontainer@sha256:649e4f4e7f362ba054fb85a67e3baa23ec9649c68df356d1e33721ca3eef2f6b",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/dsa": "true"
                  },
                  "sharedDevNum": 10
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "SgxDevicePlugin",
                "metadata": {
                  "name": "sgxdeviceplugin-sample"
                },
                "spec": {
                  "enclaveLimit": 110,
                  "image": "registry.connect.redhat.com/intel/intel-sgx-plugin@sha256:e54cdf61feb0b67e330ad24d92e7fa4000953c2ec4d4300d052e032d419d5e0a",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/sgx": "true"
                  },
                  "provisionLimit": 110
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Security, AI/Machine Learning
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/intel/intel-technology-enabling-for-openshift
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GpuDevicePlugin is the Schema for the gpudeviceplugins API.
              It represents the GPU device plugin responsible for advertising Intel
              GPU hardware resources to the kubelet.
            displayName: Intel GPU Device Plugin
            kind: GpuDevicePlugin
            name: gpudeviceplugins.deviceplugin.intel.com
            version: v1
          - description: DsaDevicePlugin is the Schema for the dsadeviceplugins API.
              It represents the DSA device plugin responsible for advertising Intel
              DSA hardware resources to the kubelet.
            displayName: Intel DSA Device Plugin
            kind: DsaDevicePlugin
            name: dsadeviceplugins.deviceplugin.intel.com
            version: v1
          - description: QatDevicePlugin is the Schema for the qatdeviceplugins API.
              It represents the QAT device plugin responsible for advertising Intel
              QuickAssist Technology hardware resources to the kubelet.
            displayName: Intel QuickAssist Technology Device Plugin
            kind: QatDevicePlugin
            name: qatdeviceplugins.deviceplugin.intel.com
            version: v1
          - description: SgxDevicePlugin is the Schema for the sgxdeviceplugins API.
              It represents the SGX device plugin responsible for advertising SGX
              device nodes to the kubelet.
            displayName: Intel Software Guard Extensions Device Plugin
            kind: SgxDevicePlugin
            name: sgxdeviceplugins.deviceplugin.intel.com
            version: v1
        description: |
          [Intel Device Plugins for Kubernetes](https://github.com/intel/intel-technology-enabling-for-openshift) is a collection of
          [device plugins](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/) advertising Intel specific hardware resources
          to the kubelet. Currently the operator has basic support for Intel SGX, Intel GPU, Intel QAT and Intel DSA. Intel IAA device plugin will be supported in a future release.
        displayName: Intel Device Plugins Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - qat
        - sgx
        - gpu
        - dsa
        - dlb
        links:
        - name: Intel Device Plugins Operator
          url: https://github.com/intel/intel-technology-enabling-for-openshift
        maintainers:
        - email: martin.xu@intel.com
          name: Martin Xu
        - email: chaitanya.kulkarni@intel.com
          name: Chaitanya Kulkarni
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Intel® Corporation
          url: https://intel.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - registry.connect.redhat.com/intel/intel-sgx-plugin@sha256:e54cdf61feb0b67e330ad24d92e7fa4000953c2ec4d4300d052e032d419d5e0a
        - registry.connect.redhat.com/intel/intel-gpu-plugin@sha256:963caf1e6e16b5e297671c536eec04dc57cf45f5cc5cf458bdcafa56bdb97972
        - registry.connect.redhat.com/intel/intel-qat-plugin@sha256:53617c3a72af169031cbaba1380cd6a25c9d2b50c9a2da822cf37b473c6e7a8f
        - registry.connect.redhat.com/intel/intel-qat-initcontainer@sha256:14f282a0adae43262657f24ce99a620b671c9a2ce12a8216163e6b22ac6e252b
        - registry.connect.redhat.com/intel/intel-dsa-plugin@sha256:092575515739bf9a0f5982c3605b63382bc4d1250128269434c2bfe4a27ec99d
        - registry.connect.redhat.com/intel/intel-idxd-initcontainer@sha256:649e4f4e7f362ba054fb85a67e3baa23ec9649c68df356d1e33721ca3eef2f6b
        - registry.connect.redhat.com/intel/intel-deviceplugin-operator@sha256:283e8b60d7ab8560f8ec8875b905e6446b59a91dd5e0d97da72e3a9aab235cc4
        version: 0.29.1
      entries:
      - name: intel-device-plugins-operator.v0.29.1
        version: 0.29.1
      - name: intel-device-plugins-operator.v0.29.0
        version: 0.29.0
      - name: intel-device-plugins-operator.v0.28.0
        version: 0.28.0
      name: alpha
    defaultChannel: alpha
    packageName: intel-device-plugins-operator
    provider:
      name: Intel® Corporation
      url: https://intel.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: designate-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: designate-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "designate.openstack.org/v1beta1",
                "kind": "Designate",
                "metadata": {
                  "name": "designate"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "designate",
                  "databaseInstance": "openstack",
                  "designateAPI": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate"
                  },
                  "designateBackendbind9": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "networkAttachments": [
                      "designate"
                    ],
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate",
                    "storageClass": "local-storage",
                    "storageRequest": "1G"
                  },
                  "designateCentral": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate"
                  },
                  "designateMdns": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "networkAttachments": [
                      "designate"
                    ],
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate"
                  },
                  "designateProducer": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate"
                  },
                  "designateUnbound": {
                    "replicas": 1
                  },
                  "designateWorker": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "networkAttachments": [
                      "designate"
                    ],
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate"
                  },
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "osp-secret",
                  "serviceUser": "designate"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:48Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DesignateAPI is the Schema for the designateapis API
            displayName: Designate API
            kind: DesignateAPI
            name: designateapis.designate.openstack.org
            version: v1beta1
          - kind: DesignateBackendbind9
            name: designatebackendbind9s.designate.openstack.org
            version: v1beta1
          - description: DesignateCentral is the Schema for the designatecentral API
            displayName: Designate Central
            kind: DesignateCentral
            name: designatecentrals.designate.openstack.org
            version: v1beta1
          - description: DesignateMdns is the Schema for the designatemdnses API
            displayName: Designate Mdns
            kind: DesignateMdns
            name: designatemdnses.designate.openstack.org
            version: v1beta1
          - description: DesignateProducer is the Schema for the designateproducer
              API
            displayName: Designate Producer
            kind: DesignateProducer
            name: designateproducers.designate.openstack.org
            version: v1beta1
          - description: Designate is the Schema for the designates API
            displayName: Designate
            kind: Designate
            name: designates.designate.openstack.org
            version: v1beta1
          - description: DesignateUnbound is the Schema for the designateunbound API
            displayName: Designate Unbound
            kind: DesignateUnbound
            name: designateunbounds.designate.openstack.org
            version: v1beta1
          - description: DesignateWorker is the Schema for the designateworker API
            displayName: Designate Worker
            kind: DesignateWorker
            name: designateworkers.designate.openstack.org
            version: v1beta1
        description: Designate Operator
        displayName: Designate Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Network
        - Designate
        links:
        - name: Designate Operator
          url: https://github.com/openstack-k8s-operators/designate-operator
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso/openstack-designate-api-rhel9@sha256:1a9baa4a0f99029eb8a367ddcce264581aa37fdf1834ba277c30a3f0ae33f400
        - registry.redhat.io/rhoso/openstack-designate-central-rhel9@sha256:ad610cdc29b31cd82a81cea22793176351574f43c0f7705273c8ffc03ad2fdac
        - registry.redhat.io/rhoso/openstack-designate-producer-rhel9@sha256:4beb652a073a65965f96f125d0fcb454255efa5e53a84bd0dd9120b57e59fcf3
        - registry.redhat.io/rhoso-operators/designate-rhel9-operator@sha256:e6c1ba511cfd0278fc8cd2836c6324d97ae3f6e9affc992600978062fa2d4549
        - registry.redhat.io/rhoso/openstack-designate-mdns-rhel9@sha256:a4109755360f340aadf6bc15d24599f882d82bd59739065f87275f9274d9eeb9
        - registry.redhat.io/rhoso/openstack-designate-worker-rhel9@sha256:94ef291ca85b06dffc93d149b09b412d9190c99f7a1b0b300d50bd70346401ad
        - registry.redhat.io/rhoso/openstack-designate-backend-bind9-rhel9@sha256:89ce30fc1eff339867ce422203e6e0c63296667e102ce85d823d44c30f7556ad
        - registry.redhat.io/rhoso/openstack-unbound-rhel9@sha256:ad95d40227a1e52f739f09a2ce8db647d6a2fa7ef63e58a2444688728fa5bdfd
        version: 1.0.6
      entries:
      - name: designate-operator.v1.0.6
        version: 1.0.6
      - name: designate-operator.v1.0.5
        version: 1.0.5
      - name: designate-operator.v1.0.4
        version: 1.0.4
      - name: designate-operator.v1.0.3
        version: 1.0.3
      - name: designate-operator.v1.0.2
        version: 1.0.2
      - name: designate-operator.v1.0.1
        version: 1.0.1
      - name: designate-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: designate-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-cert-manager-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-manager-operator.v1.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
          createdAt: 2025-08-06T08:21:40
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.16.1 <1.17.0'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.17.4](https://github.com/cert-manager/cert-manager/tree/v1.17.4), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:96d51e3a64bf30cbd92836c7cbd82f06edca16eef78ab1432757d34c16628659
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b
        version: 1.17.0
      entries:
      - name: cert-manager-operator.v1.17.0
        version: 1.17.0
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      - name: cert-manager-operator.v1.14.1
        version: 1.14.1
      - name: cert-manager-operator.v1.14.0
        version: 1.14.0
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1
    - currentCSV: cert-manager-operator.v1.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "\u003cACCESS_KEY_ID\u003e",
                            "hostedZoneID": "\u003cHOSTED_ZONE_ID\u003e",
                            "region": "\u003cAWS_REGION\u003e",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4e854c14fecbd27107b91b76164ce2e1575ce60202d835abba5962fb5d386560
          createdAt: 2024-04-30T13:29:33
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.13.0 <1.13.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/infrastructure-features: '["proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.13.6](https://github.com/cert-manager/cert-manager/tree/v1.13.6), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:afc82895c09d75e3d6a932aea2d3e3ef998f765306da92070de1dab26baf0593
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4e854c14fecbd27107b91b76164ce2e1575ce60202d835abba5962fb5d386560
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:578f74c3e96db7fc2f4e94dd9fbc7d6499d87b0bc55576930ee17de08a95022c
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:7915f8049d2451b425e3674eb60e596b9b11946ffda1497f5d5607799e2bf431
        version: 1.13.1
      entries:
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1.13
    - currentCSV: cert-manager-operator.v1.14.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:5c593bfd70309f87cd05716e233e65aa55394eafcdaf85f178698912cb29d79d
          createdAt: 2025-04-09T05:19:10
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.14.1 <1.14.2'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/infrastructure-features: '["proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.14.7](https://github.com/cert-manager/cert-manager/tree/v1.14.7), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:5c593bfd70309f87cd05716e233e65aa55394eafcdaf85f178698912cb29d79d
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:5df082f8e3787e6ac2edf395113806999ae192425ce18b4ddba9ab20c135fa5b
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:49fb8f710319d6f3479b98d2c5dd7d0c9df8c1678b2c00727314dea1a8e933a1
        version: 1.14.2
      entries:
      - name: cert-manager-operator.v1.14.2
        version: 1.14.2
      - name: cert-manager-operator.v1.14.1
        version: 1.14.1
      - name: cert-manager-operator.v1.14.0
        version: 1.14.0
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1.14
    - currentCSV: cert-manager-operator.v1.15.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:f328263e2d29e34ede65e4501f0447b2d9f84e9445a365c2fa2fbb253939e274
          createdAt: 2025-03-13T18:26:37
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.15.0 <1.15.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/infrastructure-features: '["proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.15.5](https://github.com/cert-manager/cert-manager/tree/v1.15.2), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:f328263e2d29e34ede65e4501f0447b2d9f84e9445a365c2fa2fbb253939e274
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9573d74bd2b926ec94af76f813e6358f14c5b2f4e0eedab7c1ff1070b7279a5c
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:6af3ee8b2a5a87042fb7158bda8d6cf2e6324d1e265974acf77214d4cd0ea0d3
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:20efff60a0caf5eafb38986fd21611697b5bc534c2e789da233983a9739938ed
        version: 1.15.1
      entries:
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      - name: cert-manager-operator.v1.14.1
        version: 1.14.1
      - name: cert-manager-operator.v1.14.0
        version: 1.14.0
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1.15
    - currentCSV: cert-manager-operator.v1.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:d30e755a266bb035d50e2b257e7a26c1fe6a5ccb66125576db99d537aa03ec27
          createdAt: 2025-07-10T08:52:43
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.15.1 <1.16.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.16.5](https://github.com/cert-manager/cert-manager/tree/v1.16.5), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:831e4b1e5aa6eed6fbb22f779ddbcb9c73e4bf6cc079d0c6652737a6a44e6320
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:d30e755a266bb035d50e2b257e7a26c1fe6a5ccb66125576db99d537aa03ec27
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:33308217d2226c2555eb18020710d16175295a8e349263a5435e4b158bbc735f
        version: 1.16.1
      entries:
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      - name: cert-manager-operator.v1.14.1
        version: 1.14.1
      - name: cert-manager-operator.v1.14.0
        version: 1.14.0
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1.16
    - currentCSV: cert-manager-operator.v1.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
          createdAt: 2025-08-06T08:21:40
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.16.1 <1.17.0'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.17.4](https://github.com/cert-manager/cert-manager/tree/v1.17.4), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:96d51e3a64bf30cbd92836c7cbd82f06edca16eef78ab1432757d34c16628659
        version: 1.17.0
      entries:
      - name: cert-manager-operator.v1.17.0
        version: 1.17.0
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      - name: cert-manager-operator.v1.14.1
        version: 1.14.1
      - name: cert-manager-operator.v1.14.0
        version: 1.14.0
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1.17
    defaultChannel: stable-v1
    packageName: openshift-cert-manager-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ram-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ram-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ram-controller:1.1.1
          createdAt: "2025-09-19T22:18:09Z"
          description: AWS RAM controller is a service controller for managing RAM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Permission represents the state of an AWS ram Permission
              resource.
            displayName: Permission
            kind: Permission
            name: permissions.ram.services.k8s.aws
            version: v1alpha1
          - description: ResourceShare represents the state of an AWS ram ResourceShare
              resource.
            displayName: ResourceShare
            kind: ResourceShare
            name: resourceshares.ram.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon RAM resources in AWS from within your Kubernetes cluster.

          **About Amazon RAM**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon RAM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ram
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon RAM Developer Resources
          url: https://aws.amazon.com/RAM/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ram maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ram-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ram-controller.v1.1.1
        version: 1.1.1
      - name: ack-ram-controller.v1.1.0
        version: 1.1.0
      - name: ack-ram-controller.v1.0.11
        version: 1.0.11
      - name: ack-ram-controller.v1.0.10
        version: 1.0.10
      name: alpha
    defaultChannel: alpha
    packageName: ack-ram-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operator-kogitocloud: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kogito-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kogito-operator.v1.44.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "process-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogito-infinispan-infra"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "infinispan.org/v1",
                    "kind": "Infinispan"
                  }
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/kiegroup/kogito-operator:1.44.1
          createdAt: "2019-08-22T13:12:22Z"
          description: Kogito Operator for deployment and management of Kogito services.
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.app.kiegroup.org
            version: v1beta1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.app.kiegroup.org
            version: v1beta1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.app.kiegroup.org
            version: v1beta1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.app.kiegroup.org
            version: v1beta1
        description: |-
          Kogito Operator is designed to deploy Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Creates two kinds of build configuration: Native or JVM. For more information refer to [Native build and deployment](https://docs.jboss.org/kogito/release/latest/html_single/#_native_build_and_deployment) section on our docs
          * Builds a Kogito service from a remote git repository and creates a custom image for the given runtime: Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Installs the [Kogito Data Index Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-data-index-service_kogito-deploying-on-openshift) to enable data index feature for Kogito services.
          * Installs the [Kogito Jobs Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-jobs-service_kogito-deploying-on-openshift) to enable jobs scheduling.

          For Kubernetes deployment of Kogito services, please refer to the section [Deploying Kogito services on Kubernetes](https://docs.jboss.org/kogito/release/latest/html_single/#proc-kogito-deploying-on-kubernetes_kogito-deploying-on-openshift) on our docs.

          **IMPORTANT!** Kogito Operator might need Infinispan Operator and Strimzi to implement persistence and messaging use cases if you do not provide these services beforehand.
          If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the Kogito Operator.
        displayName: Kogito
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://kogito.kie.org/
        - name: Documentation
          url: https://github.com/kiegroup/kogito-operator/blob/main/README.md
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/kiegroup/kogito-operator:1.44.1
        version: 1.44.1
      entries:
      - name: kogito-operator.v1.44.1
        version: 1.44.1
      - name: kogito-operator.v1.43.0
        version: 1.43.0
      - name: kogito-operator.v1.42.0
        version: 1.42.0
      - name: kogito-operator.v1.41.0
        version: 1.41.0
      - name: kogito-operator.v1.40.0
        version: 1.40.0
      - name: kogito-operator.v1.39.0
        version: 1.39.0
      - name: kogito-operator.v1.38.0
        version: 1.38.0
      - name: kogito-operator.v1.37.0
        version: 1.37.0
      - name: kogito-operator.v1.36.0
        version: 1.36.0
      - name: kogito-operator.v1.35.0
        version: 1.35.0
      - name: kogito-operator.v1.34.0
        version: 1.34.0
      - name: kogito-operator.v1.33.0
        version: 1.33.0
      - name: kogito-operator.v1.32.0
        version: 1.32.0
      - name: kogito-operator.v1.31.0
        version: 1.31.0
      - name: kogito-operator.v1.30.0
        version: 1.30.0
      - name: kogito-operator.v1.29.0
        version: 1.29.0
      - name: kogito-operator.v1.28.0
        version: 1.28.0
      - name: kogito-operator.v1.27.0
        version: 1.27.0
      - name: kogito-operator.v1.26.0
        version: 1.26.0
      - name: kogito-operator.v1.25.0
        version: 1.25.0
      - name: kogito-operator.v1.24.0
        version: 1.24.0
      - name: kogito-operator.v1.23.0
        version: 1.23.0
      - name: kogito-operator.v1.22.1
        version: 1.22.1
      - name: kogito-operator.v1.22.0
        version: 1.22.0
      - name: kogito-operator.v1.21.0
        version: 1.21.0
      - name: kogito-operator.v1.20.0
        version: 1.20.0
      - name: kogito-operator.v1.19.0
        version: 1.19.0
      - name: kogito-operator.v1.18.0
        version: 1.18.0
      - name: kogito-operator.v1.17.0
        version: 1.17.0
      - name: kogito-operator.v1.16.1
        version: 1.16.1
      - name: kogito-operator.v1.16.0
        version: 1.16.0
      - name: kogito-operator.v1.15.0
        version: 1.15.0
      - name: kogito-operator.v1.14.0
        version: 1.14.0
      - name: kogito-operator.v1.13.0
        version: 1.13.0
      - name: kogito-operator.v1.12.0
        version: 1.12.0
      - name: kogito-operator.v1.11.1
        version: 1.11.1
      - name: kogito-operator.v1.11.0
        version: 1.11.0
      - name: kogito-operator.v1.10.0
        version: 1.10.0
      - name: kogito-operator.v1.9.1
        version: 1.9.1
      - name: kogito-operator.v1.9.0
        version: 1.9.0
      - name: kogito-operator.v1.8.0
        version: 1.8.0
      - name: kogito-operator.v1.7.0
        version: 1.7.0
      - name: kogito-operator.v1.6.0
        version: 1.6.0
      - name: kogito-operator.v1.5.0
        version: 1.5.0
      - name: kogito-operator.v1.4.0
        version: 1.4.0
      name: 1.x
    - currentCSV: kogito-operator.v1.44.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "process-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogito-infinispan-infra"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "infinispan.org/v1",
                    "kind": "Infinispan"
                  }
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/kiegroup/kogito-operator:1.44.1
          createdAt: "2019-08-22T13:12:22Z"
          description: Kogito Operator for deployment and management of Kogito services.
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.app.kiegroup.org
            version: v1beta1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.app.kiegroup.org
            version: v1beta1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.app.kiegroup.org
            version: v1beta1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.app.kiegroup.org
            version: v1beta1
        description: |-
          Kogito Operator is designed to deploy Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Creates two kinds of build configuration: Native or JVM. For more information refer to [Native build and deployment](https://docs.jboss.org/kogito/release/latest/html_single/#_native_build_and_deployment) section on our docs
          * Builds a Kogito service from a remote git repository and creates a custom image for the given runtime: Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Installs the [Kogito Data Index Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-data-index-service_kogito-deploying-on-openshift) to enable data index feature for Kogito services.
          * Installs the [Kogito Jobs Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-jobs-service_kogito-deploying-on-openshift) to enable jobs scheduling.

          For Kubernetes deployment of Kogito services, please refer to the section [Deploying Kogito services on Kubernetes](https://docs.jboss.org/kogito/release/latest/html_single/#proc-kogito-deploying-on-kubernetes_kogito-deploying-on-openshift) on our docs.

          **IMPORTANT!** Kogito Operator might need Infinispan Operator and Strimzi to implement persistence and messaging use cases if you do not provide these services beforehand.
          If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the Kogito Operator.
        displayName: Kogito
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://kogito.kie.org/
        - name: Documentation
          url: https://github.com/kiegroup/kogito-operator/blob/main/README.md
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kiegroup/kogito-operator:1.44.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        version: 1.44.1
      entries:
      - name: kogito-operator.v1.44.1
        version: 1.44.1
      - name: kogito-operator.v1.43.0
        version: 1.43.0
      - name: kogito-operator.v1.42.0
        version: 1.42.0
      - name: kogito-operator.v1.41.0
        version: 1.41.0
      - name: kogito-operator.v1.40.0
        version: 1.40.0
      - name: kogito-operator.v1.39.0
        version: 1.39.0
      - name: kogito-operator.v1.38.0
        version: 1.38.0
      - name: kogito-operator.v1.37.0
        version: 1.37.0
      - name: kogito-operator.v1.36.0
        version: 1.36.0
      - name: kogito-operator.v1.35.0
        version: 1.35.0
      - name: kogito-operator.v1.34.0
        version: 1.34.0
      - name: kogito-operator.v1.33.0
        version: 1.33.0
      - name: kogito-operator.v1.32.0
        version: 1.32.0
      - name: kogito-operator.v1.31.0
        version: 1.31.0
      - name: kogito-operator.v1.30.0
        version: 1.30.0
      - name: kogito-operator.v1.29.0
        version: 1.29.0
      - name: kogito-operator.v1.28.0
        version: 1.28.0
      - name: kogito-operator.v1.27.0
        version: 1.27.0
      - name: kogito-operator.v1.26.0
        version: 1.26.0
      - name: kogito-operator.v1.25.0
        version: 1.25.0
      - name: kogito-operator.v1.24.0
        version: 1.24.0
      - name: kogito-operator.v1.23.0
        version: 1.23.0
      - name: kogito-operator.v1.22.1
        version: 1.22.1
      - name: kogito-operator.v1.22.0
        version: 1.22.0
      - name: kogito-operator.v1.21.0
        version: 1.21.0
      - name: kogito-operator.v1.20.0
        version: 1.20.0
      - name: kogito-operator.v1.19.0
        version: 1.19.0
      - name: kogito-operator.v1.18.0
        version: 1.18.0
      - name: kogito-operator.v1.17.0
        version: 1.17.0
      - name: kogito-operator.v1.16.1
        version: 1.16.1
      - name: kogito-operator.v1.16.0
        version: 1.16.0
      - name: kogito-operator.v1.15.0
        version: 1.15.0
      - name: kogito-operator.v1.14.0
        version: 1.14.0
      - name: kogito-operator.v1.13.0
        version: 1.13.0
      - name: kogito-operator.v1.12.0
        version: 1.12.0
      - name: kogito-operator.v1.11.1
        version: 1.11.1
      - name: kogito-operator.v1.11.0
        version: 1.11.0
      - name: kogito-operator.v1.10.0
        version: 1.10.0
      - name: kogito-operator.v1.9.1
        version: 1.9.1
      - name: kogito-operator.v1.9.0
        version: 1.9.0
      - name: kogito-operator.v1.8.0
        version: 1.8.0
      - name: kogito-operator.v1.7.0
        version: 1.7.0
      - name: kogito-operator.v1.6.0
        version: 1.6.0
      - name: kogito-operator.v1.5.0
        version: 1.5.0
      - name: kogito-operator.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: 1.x
    packageName: kogito-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: neutron-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neutron-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "neutron.openstack.org/v1beta1",
                "kind": "NeutronAPI",
                "metadata": {
                  "name": "neutron",
                  "namespace": "openstack"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "neutron",
                  "databaseInstance": "openstack",
                  "memcachedInstance": "memcached",
                  "preserveJobs": false,
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "neutron-secret",
                  "serviceUser": "neutron"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:57Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NeutronAPI is the Schema for the neutronapis API
            displayName: Neutron API
            kind: NeutronAPI
            name: neutronapis.neutron.openstack.org
            version: v1beta1
        description: Neutron Operator
        displayName: Neutron Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Network
        - Neutron
        links:
        - name: Neutron Operator
          url: https://github.com/openstack-k8s-operators/neutron-operator
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/neutron-rhel9-operator@sha256:16b1a5f83692f3d193754df664992e5b88cb357f158b69261eb802430f3a69f1
        - registry.redhat.io/rhoso/openstack-neutron-server-rhel9@sha256:806e5e009a77a29c82699487e539669e1ec2136514a0332d222b5cb3fed6f452
        version: 1.0.6
      entries:
      - name: neutron-operator.v1.0.6
        version: 1.0.6
      - name: neutron-operator.v1.0.5
        version: 1.0.5
      - name: neutron-operator.v1.0.4
        version: 1.0.4
      - name: neutron-operator.v1.0.3
        version: 1.0.3
      - name: neutron-operator.v1.0.2
        version: 1.0.2
      - name: neutron-operator.v1.0.1
        version: 1.0.1
      - name: neutron-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: neutron-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Tempo Operator SIG
      provider-url: ""
    name: tempo-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tempo-operator.v0.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoMonolithic",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "jaegerui": {
                    "enabled": true,
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    }
                  },
                  "storage": {
                    "traces": {
                      "backend": "memory"
                    }
                  }
                }
              },
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoStack",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "resources": {
                    "total": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2Gi"
                      }
                    }
                  },
                  "storage": {
                    "secret": {
                      "name": "my-storage-secret",
                      "type": "s3"
                    }
                  },
                  "storageSize": "1Gi",
                  "template": {
                    "queryFrontend": {
                      "jaegerQuery": {
                        "enabled": true,
                        "ingress": {
                          "type": "route"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          containerImage: ghcr.io/grafana/tempo-operator/tempo-operator:v0.18.0
          createdAt: "2025-09-17T14:23:28Z"
          description: Create and manage deployments of Tempo, a high-scale distributed
            tracing backend.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: tempo-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/tempo-operator
          support: Grafana Tempo Operator SIG
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TempoMonolithic manages a Tempo deployment in monolithic
              mode.
            displayName: TempoMonolithic
            kind: TempoMonolithic
            name: tempomonolithics.tempo.grafana.com
            version: v1alpha1
          - description: TempoStack manages a Tempo deployment in microservices mode.
            displayName: TempoStack
            kind: TempoStack
            name: tempostacks.tempo.grafana.com
            version: v1alpha1
        description: |-
          Tempo is an open source, easy-to-use, and high-scale distributed tracing backend.
          It can ingest common open source tracing protocols including Jaeger, Zipkin, and OpenTelemetry and requires only object storage to operate.
          Please refer to the [Tempo documentation](https://grafana.com/docs/tempo/latest/) for more information about Tempo.

          The Community Tempo Operator can manage Tempo deployments in Microservices mode (`TempoStack` CR) and Monolithic mode (`TempoMonolithic` CR).

          ### Operator features
          * **Resource Limits** - Specify overall resource requests and limits in the `TempoStack` CR; the operator assigns fractions of it to each component
          * **AuthN and AuthZ** - Supports OpenID Control (OIDC) and role-based access control (RBAC)
          * **Managed upgrades** - Updating the operator will automatically update all managed Tempo clusters
          * **Multitenancy** - Multiple tenants can send traces to the same Tempo cluster
          * **mTLS** - Communication between the Tempo components can be secured via mTLS
          * **Jaeger UI** - Traces can be visualized in Jaeger UI and exposed via Ingress or OpenShift Route
          * **Observability** - The operator and `TempoStack` operands expose telemetry (metrics, traces) and integrate with Prometheus `ServiceMonitor` and `PrometheusRule`

          ### Prerequisites
          `TempoStack` requires object storage to store its traces. `TempoMonolithic` can store traces in-memory, in a Persistent Volume and in object storage.
          Please ensure that an object storage solution is available and configured.
        displayName: Community Tempo Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tempo
        - tracing
        - observability
        - monitoring
        - database
        links:
        - name: Tempo Operator
          url: https://github.com/grafana/tempo-operator
        maintainers:
        - email: ruben.vp8510@gmail.com
          name: Ruben Vargas
        - email: p.loffay@gmail.com
          name: Pavol Loffay
        - email: bongartz@klimlive.de
          name: Benedikt Bongartz
        - email: andreas@gerstmayr.me
          name: Andreas Gerstmayr
        - email: iblancas@redhat.com
          name: Israel Blancas Alvarez
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Grafana Tempo Operator SIG
        relatedImages:
        - docker.io/jaegertracing/jaeger-query:1.68.0
        - docker.io/grafana/tempo-query:2.8.2
        - quay.io/observatorium/api:main-2025-08-25-7489eb0
        - quay.io/observatorium/opa-openshift:main-2025-06-16-ecdeca0
        - quay.io/openshift/origin-oauth-proxy:4.14
        - ghcr.io/grafana/tempo-operator/tempo-operator:v0.18.0
        - docker.io/grafana/tempo:2.8.2
        version: 0.18.0
      entries:
      - name: tempo-operator.v0.18.0
        version: 0.18.0
      - name: tempo-operator.v0.17.1
        version: 0.17.1
      - name: tempo-operator.v0.15.3
        version: 0.15.3
      - name: tempo-operator.v0.15.2
        version: 0.15.2
      - name: tempo-operator.v0.15.1
        version: 0.15.1
      - name: tempo-operator.v0.15.0
        version: 0.15.0
      - name: tempo-operator.v0.14.2
        version: 0.14.2
      - name: tempo-operator.v0.14.1
        version: 0.14.1
      - name: tempo-operator.v0.14.0
        version: 0.14.0
      - name: tempo-operator.v0.13.0
        version: 0.13.0
      - name: tempo-operator.v0.12.0
        version: 0.12.0
      - name: tempo-operator.v0.11.1
        version: 0.11.1
      - name: tempo-operator.v0.11.0
        version: 0.11.0
      - name: tempo-operator.v0.10.0
        version: 0.10.0
      - name: tempo-operator.v0.9.0
        version: 0.9.0
      - name: tempo-operator.v0.8.0
        version: 0.8.0
      - name: tempo-operator.v0.7.0
        version: 0.7.0
      - name: tempo-operator.v0.6.0
        version: 0.6.0
      - name: tempo-operator.v0.5.0
        version: 0.5.0
      - name: tempo-operator.v0.4.0
        version: 0.4.0
      - name: tempo-operator.v0.3.0
        version: 0.3.0
      - name: tempo-operator.v0.2.0
        version: 0.2.0
      - name: tempo-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: tempo-operator
    provider:
      name: Grafana Tempo Operator SIG
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: project-quay
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-operator.v3.10.0-nightly.20231014
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.10.0-nightly.20231014
          createdAt: 2023-10-14 06:26 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.10.0-nightly.20231014'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.10.0-nightly.20231014
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.10.0-nightly.20231014
        version: 3.10.0-nightly.20231014
      entries:
      - name: quay-operator.v3.10.0-nightly.20231014
        version: 3.10.0-nightly.20231014
      - name: quay-operator.v3.10.0-nightly.20231013
        version: 3.10.0-nightly.20231013
      - name: quay-operator.v3.10.0-nightly.20231012
        version: 3.10.0-nightly.20231012
      - name: quay-operator.v3.10.0-nightly.20231011
        version: 3.10.0-nightly.20231011
      - name: quay-operator.v3.10.0-nightly.20231010
        version: 3.10.0-nightly.20231010
      - name: quay-operator.v3.10.0-nightly.20231009
        version: 3.10.0-nightly.20231009
      - name: quay-operator.v3.10.0-nightly.20231008
        version: 3.10.0-nightly.20231008
      - name: quay-operator.v3.10.0-nightly.20231007
        version: 3.10.0-nightly.20231007
      - name: quay-operator.v3.10.0-nightly.20231006
        version: 3.10.0-nightly.20231006
      - name: quay-operator.v3.10.0-nightly.20231005
        version: 3.10.0-nightly.20231005
      - name: quay-operator.v3.10.0-nightly.20231002
        version: 3.10.0-nightly.20231002
      - name: quay-operator.v3.10.0-nightly.20230930
        version: 3.10.0-nightly.20230930
      - name: quay-operator.v3.10.0-nightly.20230929
        version: 3.10.0-nightly.20230929
      - name: quay-operator.v3.10.0-nightly.20230928
        version: 3.10.0-nightly.20230928
      - name: quay-operator.v3.10.0-nightly.20230926
        version: 3.10.0-nightly.20230926
      - name: quay-operator.v3.10.0-nightly.20230925
        version: 3.10.0-nightly.20230925
      - name: quay-operator.v3.10.0-nightly.20230923
        version: 3.10.0-nightly.20230923
      - name: quay-operator.v3.10.0-nightly.20230922
        version: 3.10.0-nightly.20230922
      - name: quay-operator.v3.10.0-nightly.20230921
        version: 3.10.0-nightly.20230921
      - name: quay-operator.v3.10.0-nightly.20230920
        version: 3.10.0-nightly.20230920
      - name: quay-operator.v3.10.0-nightly.20230919
        version: 3.10.0-nightly.20230919
      - name: quay-operator.v3.10.0-nightly.20230918
        version: 3.10.0-nightly.20230918
      - name: quay-operator.v3.10.0-nightly.20230917
        version: 3.10.0-nightly.20230917
      - name: quay-operator.v3.10.0-nightly.20230916
        version: 3.10.0-nightly.20230916
      - name: quay-operator.v3.10.0-nightly.20230915
        version: 3.10.0-nightly.20230915
      - name: quay-operator.v3.10.0-nightly.20230914
        version: 3.10.0-nightly.20230914
      - name: quay-operator.v3.10.0-nightly.20230913
        version: 3.10.0-nightly.20230913
      name: candidate-3.10
    - currentCSV: quay-operator.v3.9.0-nightly.20230812
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.9.0-nightly.20230812
          createdAt: 2023-08-12 06:06 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.9.0-nightly.20230812'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.9.0-nightly.20230812
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.9.0-nightly.20230812
        version: 3.9.0-nightly.20230812
      entries:
      - name: quay-operator.v3.9.0-nightly.20230812
        version: 3.9.0-nightly.20230812
      - name: quay-operator.v3.9.0-nightly.20230724
        version: 3.9.0-nightly.20230724
      - name: quay-operator.v3.9.0-nightly.20230723
        version: 3.9.0-nightly.20230723
      - name: quay-operator.v3.9.0-nightly.20230722
        version: 3.9.0-nightly.20230722
      - name: quay-operator.v3.9.0-nightly.20230721
        version: 3.9.0-nightly.20230721
      - name: quay-operator.v3.9.0-nightly.20230720
        version: 3.9.0-nightly.20230720
      - name: quay-operator.v3.9.0-nightly.20230717
        version: 3.9.0-nightly.20230717
      - name: quay-operator.v3.9.0-nightly.20230716
        version: 3.9.0-nightly.20230716
      - name: quay-operator.v3.9.0-nightly.20230715
        version: 3.9.0-nightly.20230715
      - name: quay-operator.v3.9.0-nightly.20230714
        version: 3.9.0-nightly.20230714
      - name: quay-operator.v3.9.0-nightly.20230712
        version: 3.9.0-nightly.20230712
      - name: quay-operator.v3.9.0-nightly.20230711
        version: 3.9.0-nightly.20230711
      name: candidate-3.9
    - currentCSV: quay-operator.v3.7.0-rc.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.7.0-rc.4
          createdAt: 2022-05-17 14:05 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.3.x <3.7.0-rc.4'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.7.0-rc.4
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.7.0-rc.4
        version: 3.7.0-rc.4
      entries:
      - name: quay-operator.v3.7.0-rc.4
        version: 3.7.0-rc.4
      - name: quay-operator.v3.7.0-rc.3
        version: 3.7.0-rc.3
      - name: quay-operator.v3.7.0-rc.2
        version: 3.7.0-rc.2
      - name: quay-operator.v3.7.0-rc.1
        version: 3.7.0-rc.1
      name: preview-3.7
    - currentCSV: quay-operator.v3.8.0-rc.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.8.0-rc.7
          createdAt: 2022-11-29 15:11 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.8.0-rc.7'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.8.0-rc.7
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.8.0-rc.7
        version: 3.8.0-rc.7
      entries:
      - name: quay-operator.v3.8.0-rc.7
        version: 3.8.0-rc.7
      - name: quay-operator.v3.8.0-rc.6
        version: 3.8.0-rc.6
      - name: quay-operator.v3.8.0-rc.5
        version: 3.8.0-rc.5
      - name: quay-operator.v3.8.0-rc.4
        version: 3.8.0-rc.4
      - name: quay-operator.v3.8.0-rc.3
        version: 3.8.0-rc.3
      - name: quay-operator.v3.8.0-rc.2
        version: 3.8.0-rc.2
      - name: quay-operator.v3.8.0-rc.1
        version: 3.8.0-rc.1
      name: preview-3.8
    - currentCSV: quay-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.10.15
          createdAt: "2025-09-25T14:06:05Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.10.15
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.10.15
        version: 3.10.15
      entries:
      - name: quay-operator.v3.10.15
        version: 3.10.15
      - name: quay-operator.v3.10.14
        version: 3.10.14
      - name: quay-operator.v3.10.13
        version: 3.10.13
      - name: quay-operator.v3.10.12
        version: 3.10.12
      - name: quay-operator.v3.10.11
        version: 3.10.11
      - name: quay-operator.v3.10.10
        version: 3.10.10
      - name: quay-operator.v3.10.9
        version: 3.10.9
      - name: quay-operator.v3.10.8
        version: 3.10.8
      - name: quay-operator.v3.10.7
        version: 3.10.7
      - name: quay-operator.v3.10.6
        version: 3.10.6
      - name: quay-operator.v3.10.5
        version: 3.10.5
      - name: quay-operator.v3.10.4
        version: 3.10.4
      - name: quay-operator.v3.10.3
        version: 3.10.3
      - name: quay-operator.v3.10.2
        version: 3.10.2
      - name: quay-operator.v3.10.1
        version: 3.10.1
      - name: quay-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.11.13
          createdAt: "2025-09-02T06:02:15Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.11.13
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.11.13
        version: 3.11.13
      entries:
      - name: quay-operator.v3.11.13
        version: 3.11.13
      - name: quay-operator.v3.11.12
        version: 3.11.12
      - name: quay-operator.v3.11.11
        version: 3.11.11
      - name: quay-operator.v3.11.10
        version: 3.11.10
      - name: quay-operator.v3.11.9
        version: 3.11.9
      - name: quay-operator.v3.11.8
        version: 3.11.8
      - name: quay-operator.v3.11.7
        version: 3.11.7
      - name: quay-operator.v3.11.6
        version: 3.11.6
      - name: quay-operator.v3.11.5
        version: 3.11.5
      - name: quay-operator.v3.11.4
        version: 3.11.4
      - name: quay-operator.v3.11.3
        version: 3.11.3
      - name: quay-operator.v3.11.2
        version: 3.11.2
      - name: quay-operator.v3.11.1
        version: 3.11.1
      - name: quay-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.12.12
          createdAt: "2025-09-23T09:59:30Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.12.12
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.12.12
        version: 3.12.12
      entries:
      - name: quay-operator.v3.12.12
        version: 3.12.12
      - name: quay-operator.v3.12.11
        version: 3.12.11
      - name: quay-operator.v3.12.10
        version: 3.12.10
      - name: quay-operator.v3.12.9
        version: 3.12.9
      - name: quay-operator.v3.12.8
        version: 3.12.8
      - name: quay-operator.v3.12.7
        version: 3.12.7
      - name: quay-operator.v3.12.6
        version: 3.12.6
      - name: quay-operator.v3.12.5
        version: 3.12.5
      - name: quay-operator.v3.12.4
        version: 3.12.4
      - name: quay-operator.v3.12.3
        version: 3.12.3
      - name: quay-operator.v3.12.2
        version: 3.12.2
      - name: quay-operator.v3.12.1
        version: 3.12.1
      - name: quay-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.13.8
          createdAt: "2025-08-28T03:28:40Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.13.8
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.13.8
        version: 3.13.8
      entries:
      - name: quay-operator.v3.13.8
        version: 3.13.8
      - name: quay-operator.v3.13.7
        version: 3.13.7
      - name: quay-operator.v3.13.6
        version: 3.13.6
      - name: quay-operator.v3.13.5
        version: 3.13.5
      - name: quay-operator.v3.13.4
        version: 3.13.4
      - name: quay-operator.v3.13.3
        version: 3.13.3
      - name: quay-operator.v3.13.2
        version: 3.13.2
      - name: quay-operator.v3.13.1
        version: 3.13.1
      - name: quay-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.14.5
          createdAt: "2025-09-22T14:19:49Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.14.5
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.14.5
        version: 3.14.5
      entries:
      - name: quay-operator.v3.14.5
        version: 3.14.5
      - name: quay-operator.v3.14.4
        version: 3.14.4
      - name: quay-operator.v3.14.3
        version: 3.14.3
      - name: quay-operator.v3.14.2
        version: 3.14.2
      - name: quay-operator.v3.14.1
        version: 3.14.1
      - name: quay-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.15.2
          createdAt: "2025-09-22T13:44:32Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.15.2
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.15.2
        version: 3.15.2
      entries:
      - name: quay-operator.v3.15.2
        version: 3.15.2
      - name: quay-operator.v3.15.1
        version: 3.15.1
      - name: quay-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-operator.v3.6.2
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"quay.redhat.com/v1\",\n    \"kind\":
            \"QuayRegistry\",\n    \"metadata\": {\n      \"name\": \"example-registry\"\n
            \   },\n    \"spec\": {\n      \"components\": [\n        {\"kind\": \"clair\",
            \"managed\": true},\n        {\"kind\": \"postgres\", \"managed\": true},\n
            \       {\"kind\": \"objectstorage\", \"managed\": true},\n        {\"kind\":
            \"redis\", \"managed\": true},\n        {\"kind\": \"horizontalpodautoscaler\",
            \"managed\": true},\n        {\"kind\": \"route\", \"managed\": true},\n
            \       {\"kind\": \"mirror\", \"managed\": true},\n        {\"kind\":
            \"monitoring\", \"managed\": true},\n        {\"kind\": \"tls\", \"managed\":
            true}\n      ]\n    }\n  }, \n  {\n    \"apiVersion\": \"redhatcop.redhat.io/v1alpha1\",\n
            \   \"kind\": \"QuayEcosystem\",\n    \"metadata\": {\n      \"name\":
            \"example-quayecosystem\"\n    },\n    \"spec\": {}\n  }\n]"
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:v3.6.2
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.5.x <3.6.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.6.2
          repository: https://github.com/quay/quay-operator
          support: Quay Team
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Project Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        - quay
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:v3.6.2
        version: 3.6.2
      entries:
      - name: quay-operator.v3.6.2
        version: 3.6.2
      name: stable-3.6
    - currentCSV: quay-operator.v3.7.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.7.11
          createdAt: 2023-02-22 10:02 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.3.x <3.7.11'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.7.11
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.7.11
        version: 3.7.11
      entries:
      - name: quay-operator.v3.7.11
        version: 3.7.11
      - name: quay-operator.v3.7.10
        version: 3.7.10
      - name: quay-operator.v3.7.9
        version: 3.7.9
      - name: quay-operator.v3.7.8
        version: 3.7.8
      - name: quay-operator.v3.7.7
        version: 3.7.7
      - name: quay-operator.v3.7.6
        version: 3.7.6
      - name: quay-operator.v3.7.5
        version: 3.7.5
      - name: quay-operator.v3.7.4
        version: 3.7.4
      - name: quay-operator.v3.7.3
        version: 3.7.3
      - name: quay-operator.v3.7.2
        version: 3.7.2
      - name: quay-operator.v3.7.1
        version: 3.7.1
      - name: quay-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: quay-operator.v3.8.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.8.13
          createdAt: 2023-10-17 16:13 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.8.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.8.13
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.8.13
        version: 3.8.13
      entries:
      - name: quay-operator.v3.8.13
        version: 3.8.13
      - name: quay-operator.v3.8.12
        version: 3.8.12
      - name: quay-operator.v3.8.11
        version: 3.8.11
      - name: quay-operator.v3.8.10
        version: 3.8.10
      - name: quay-operator.v3.8.8
        version: 3.8.8
      - name: quay-operator.v3.8.7
        version: 3.8.7
      - name: quay-operator.v3.8.6
        version: 3.8.6
      - name: quay-operator.v3.8.5
        version: 3.8.5
      - name: quay-operator.v3.8.4
        version: 3.8.4
      - name: quay-operator.v3.8.3
        version: 3.8.3
      - name: quay-operator.v3.8.2
        version: 3.8.2
      - name: quay-operator.v3.8.1
        version: 3.8.1
      - name: quay-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.9.16
          createdAt: "2025-09-25T14:31:58Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.9.16
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.9.16
        version: 3.9.16
      entries:
      - name: quay-operator.v3.9.16
        version: 3.9.16
      - name: quay-operator.v3.9.15
        version: 3.9.15
      - name: quay-operator.v3.9.14
        version: 3.9.14
      - name: quay-operator.v3.9.13
        version: 3.9.13
      - name: quay-operator.v3.9.12
        version: 3.9.12
      - name: quay-operator.v3.9.11
        version: 3.9.11
      - name: quay-operator.v3.9.10
        version: 3.9.10
      - name: quay-operator.v3.9.9
        version: 3.9.9
      - name: quay-operator.v3.9.6
        version: 3.9.6
      - name: quay-operator.v3.9.5
        version: 3.9.5
      - name: quay-operator.v3.9.4
        version: 3.9.4
      - name: quay-operator.v3.9.3
        version: 3.9.3
      - name: quay-operator.v3.9.2
        version: 3.9.2
      - name: quay-operator.v3.9.1
        version: 3.9.1
      - name: quay-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: project-quay
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-ui
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-ui.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Ui",
                "metadata": {
                  "name": "ui-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "volumeMounts": [
                    {
                      "mountPath": "/opt/app-root/src/assets/config.json",
                      "name": "config",
                      "subPath": "config.json"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:37:58Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ui
            name: uis.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP UI Operator
        displayName: Keysight WAP UI Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Ui
          url: https://keysight-wap-ui.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-ui@sha256:876a394f5cc2f745091ad2ee9ca0d08404cfb95284222e1f3a8607e193bd986a
        version: 0.1.21
      entries:
      - name: keysight-wap-ui.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-ui
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator-certified-rhmp.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator-certified-rhmp.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator-certified-rhmp.v2.6.1
        version: 2.6.1
      - name: percona-postgresql-operator-certified-rhmp.v2.6.0
        version: 2.6.0
      - name: percona-postgresql-operator-certified-rhmp.v2.5.0
        version: 2.5.0
      name: stable
    - currentCSV: percona-postgresql-operator-certified-rhmp.v2.7.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        version: 2.7.0-cw
      entries:
      - name: percona-postgresql-operator-certified-rhmp.v2.7.0-cw
        version: 2.7.0-cw
      - name: percona-postgresql-operator-certified-rhmp.v2.6.1-cw
        version: 2.6.1-cw
      - name: percona-postgresql-operator-certified-rhmp.v2.6.0-cw
        version: 2.6.0-cw
      - name: percona-postgresql-operator-certified-rhmp.v2.5.0-cw
        version: 2.5.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-postgresql-operator-certified-rhmp
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: aws-efs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-efs-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"aws-efs.managed.openshift.io/v1alpha1","kind":"SharedVolume","metadata":{"name":"sv1"},"spec":{"accessPointID":
            "fsap-0123456789abcdef", "fileSystemID": "fs-0123cdef"}}]'
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/app-sre/aws-efs-operator:9c87267
          createdAt: "2021-12-20T19:51:50Z"
          description: Manage read-write-many access to AWS EFS volumes in an OpenShift
            cluster.
          repository: https://github.com/openshift/aws-efs-operator
          support: https://github.com/openshift/aws-efs-operator/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Shared Volume
            displayName: SharedVolume
            kind: SharedVolume
            name: sharedvolumes.aws-efs.managed.openshift.io
            version: v1alpha1
        description: |-
          ## About the managed application
          AWS provides Elastic File System, an NFS-based shared storage solution. A couple of mechanisms exist to enable mounting such a file system in containers. Curating these options and making them work in a cluster is nontrivial.
          ## About this Operator
          The aws-efs-operator minimizes the barrier to entry for container use of EFS by:
          - Installing the [AWS EFS CSI Driver](https://github.com/kubernetes-sigs/aws-efs-csi-driver) and supporting infrastructure.
          - Reducing the identification of a shared storage resource to a simple Custom Resource Definition with only two fields.
          - Creating the necessary PersistentVolume and PersistentVolumeClaim resources that can simply be referenced from a Pod specification.

          More information can be found in the operator's [source repository](https://github.com/openshift/aws-efs-operator).
          ## Prerequisites for enabling this Operator
          An OpenShift cluster backed by AWS.
        displayName: AWS EFS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AWS
        - EFS
        - NFS
        - Shared storage
        links:
        - name: README
          url: https://github.com/openshift/aws-efs-operator/blob/master/README.md
        - name: Design
          url: https://github.com/openshift/aws-efs-operator/blob/master/DESIGN.md
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/app-sre/aws-efs-operator:9c87267
        version: 0.0.8
      entries:
      - name: aws-efs-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: aws-efs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-application-gateway-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-application-gateway-operator.v25.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMApplicationGateway",
                "metadata": {
                  "name": "iag-instance"
                },
                "spec": {
                  "configuration": [
                    {
                      "dataKey": "config",
                      "name": "test-config",
                      "type": "configmap"
                    },
                    {
                      "discoveryEndpoint": "https://isam.mmfa.ibm.com/mga/sps/oauth/oauth20/metadata/test",
                      "postData": [
                        {
                          "name": "redirect_uris",
                          "values": [
                            "https://isam.mmfa.ibm.com/pkmsoidc"
                          ]
                        },
                        {
                          "name": "client_name",
                          "value": "OperatorTest"
                        },
                        {
                          "name": "enforce_pkce",
                          "value": "false"
                        },
                        {
                          "name": "all_users_entitled",
                          "value": "true"
                        },
                        {
                          "name": "consent_action",
                          "value": "never_prompt"
                        }
                      ],
                      "secret": "oidc-client",
                      "type": "oidc_registration"
                    },
                    {
                      "headers": [
                        {
                          "name": "Authorization",
                          "secretKey": "value",
                          "type": "secret",
                          "value": "githubsecret"
                        }
                      ],
                      "type": "web",
                      "url": "https://raw.github.com/ibm-security/iag-config/master/test/sample1.yaml"
                    },
                    {
                      "type": "literal",
                      "value": "version: \"22.07\"\n\nserver:\n  local_applications:\n    cred_viewer:\n      path_segment: creds\n      enable_html: true\n\nidentity:\n  oidc:\n    discovery_endpoint: \"https://iag-dev.ibmcloudsecurity.com/oidc/endpoint/default/.well-known/openid-configuration\"\n    client_id: 1cbfe647-9e5f-4d99-8e05-8ed1c862eb47\n    client_secret: uPP8rM7N0e\n"
                    }
                  ],
                  "deployment": {
                    "image": "icr.io/ibmappgateway/ibm-application-gateway@sha256:1144b79a2930fe942ce627b0abcc3bed2f0b0770ae8313cb77ffe6772511c1f9",
                    "imagePullPolicy": "Always",
                    "imagePullSecrets": [
                      {
                        "name": "regcred"
                      }
                    ],
                    "lang": "C",
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 8,
                      "periodSeconds": 9,
                      "successThreshold": 7,
                      "timeoutSeconds": 1
                    },
                    "readinessProbe": {
                      "failureThreshold": 2,
                      "initialDelaySeconds": 7,
                      "periodSeconds": 8,
                      "successThreshold": 4,
                      "timeoutSeconds": 5
                    },
                    "serviceAccountName": "iag"
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/ibmappgateway/ibm-application-gateway-operator@sha256:6a16d4ef802a866942f504b11a948565245d303f322944200c1a1269fe7c371c
          createdAt: "2025-02-05T21:15:37Z"
          description: The IBM Application Gateway operator manages the lifecycle
            of IBM Application Gateway containers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM-Security/ibm-application-gateway-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMApplicationGateway is the Schema for the ibmapplicationgateways
              API
            displayName: IBMApplication Gateway
            kind: IBMApplicationGateway
            name: ibmapplicationgateways.ibm.com
            version: v1
        description: "The [IBM Application Gateway (IAG)](https://ibm.biz/ibm-app-gateway)
          image provides a containerized secure Web Reverse proxy which is designed
          to sit in front of your application, seamlessly adding authentication and
          authorization protection to your application.\n\nAn IAG instance deployed
          on Kubernetes can be a complex deployment.  In particular the configuration
          can be defined externally in one or more locations, and changes to this
          configuration may require all instances to be reloaded for the changes to
          take effect. \n\nThe internal Kubernetes deployment controller does not
          have any knowledge of how an IBM Application Gateway instance should behave
          when the configuration changes. As such this IBM Application Gateway specific
          Kubernetes operator is available to be deployed. Once deployed the operator
          can manage the lifecycle of the IBM Application Gateway instances.\n\nSee
          the project [Readme](https://github.com/IBM-Security/ibm-application-gateway-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Application Gateway Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Security
        - Identity and Access
        links:
        - name: IBM Application Gateway product documentation
          url: https://ibm.biz/ibm-app-gateway
        maintainers:
        - email: isamdev@au1.ibm.com
          name: IBM Application Gateway Development Team
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - icr.io/ibmappgateway/ibm-application-gateway-operator@sha256:6a16d4ef802a866942f504b11a948565245d303f322944200c1a1269fe7c371c
        - icr.io/ibmappgateway/ibm-application-gateway@sha256:1144b79a2930fe942ce627b0abcc3bed2f0b0770ae8313cb77ffe6772511c1f9
        version: 25.2.0
      entries:
      - name: ibm-application-gateway-operator.v25.2.0
        version: 25.2.0
      - name: ibm-application-gateway-operator.v24.4.0
        version: 24.4.0
      name: stable
    defaultChannel: stable
    packageName: ibm-application-gateway-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cockroach Labs
      provider-url: ""
    name: cockroachdb-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroach-operator.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v21.1.10",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cockroachdb/cockroachdb-operator@sha256:9fba7cd5fdf589317911925cb69338508bf9de928b1817c6f117c612635bbab3
          createdAt: "2021-10-19T15:47:03Z"
          description: CockroachDB Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:e67299fc0e79cff5b6caf9c0df154180dbaedc57b84f072acf68b7b2c958668d
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ec1672ce8091b677689a31cf09dbde5d1697f4caddae0151040b0fb156722e9d
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:e39bce1d9b08a7d62eb05d53e29fabc668bf6a4e4ae3eb9994eebf5b642ff6cf
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ab47ddd22328e0a1564814f049616c3e3569d24c0a290abee9308392c6a0de23
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:49421968126f9c98499bc0cf0c65729370ab566da3fbd1588c100b20052b972e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:c56f980b87a066d1e65ac8cff3d75040af9bd13093494cf4b4d30b5234d56edc
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:de94036e9cc8618e0493313512d880ae145e94d3bff54ed51b0de66a88da1319
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:3edda46eccce8350440a0ea78dad2e41a3eaa0838fe57945ad5318739e968516
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:0bd22d69db5ac570f30d3379afc67f6c2e46fc5b027c8aab42c3f562085d4672
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:bcf3f976437c4ee52e922aa5d7e382a86a031d780b711c08c6c158dbf2d26cfa
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:0ef0234b6fd42977b8a9eda2c59ca8194cc1c8cf6425b99d7bf8fde8feb826c5
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:87dcdbdc07904b217880e70484582c9d87dde7e4071e01cf9e2e6da43111190e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:2664d78fbe88c368927ddbf80704f430e6ca2ce1f72cb4b7e1906dfe72be7cd0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:fe0c615b4d1ef4ec6e089e85deae2b6bd85410be8ed525e2317c759669417f47
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1b56420753eac1126039b8ad4be82ddb59c89a4ec72fb7b6198cea88a10f3a91
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1ed18af1d3ffc65c24946c817d7d04d5f01025d3fb6c9eb78940e343b989bed0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:162a356cc8b423926634f83a1372b6321104d8de7a2c5b9b7b2681c769a88cc7
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:288ae92ebdfc848540ff80ef682b74e50809e9742cafce22b028112326d66b65
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:867c46bb4d8ca3f2fb8dc94063facc99758344c7b9d04f10e50932023ef5d262
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ba9da9aa9d662c0fa9fd16a40551eedf100869b008cf9c0c09cbb87a09fda697
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:25791b76b0d4b7220dadab3a405b3fdc330ba264ef6c9fbfdfb924c154c4bb5e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:cdab9fc62e07dd349703ae394adb48ab3d2281610df35e081f4b76e208b38c8e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:a5414e54c4c513c7877f925b09eadc4c28eb592d3e120bcf9c65ab10e07cd607
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:98b76cf63aabc39d026317905d6ba72eec3f0d35c0f770ebb8b789c3fa0709ee
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:9f5dfcde260804cd2d6976a6b3544ddabdd899365be2104210dc78ba13755ec3
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:27ef61d9f320a52ee9617f2ab8a1d1a1eacc576a0add2d85d99f9413912b651f
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:38e940144e34b69dfb7d1cf31456e7a0016c5dd40d5626baa346446c7678ceb0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1cea12847a0d236437f7fc28fba476fe512c343cd020d21bb50316e50da6ffd6
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:c1a07a8686bbadcb8c68ffba4e40e1285e581e24d43e5811f38a22d57fb9cc56
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:d290099496d070f135b5cc1e6bd59607512524d399df81698f11cfa09b8dca4a
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:b02549e5c3673b8154441990afa05a478d1ecc2bd0a70af64fd693cb6b469cb0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:5eb59093ee81f8115d82ec6a8a1c35eefad055cbb3a175978cef7afe9196e6a0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:fec6b0f56f2b94f141c8553a63b2e08b9494d4e6c5dc109d73268621ca445c56
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:d9a756df2ed2536b938d7c24fbf3de149cd1a0f44a3a1a0cc5d042fe8980362e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroachdb-operator@sha256:9fba7cd5fdf589317911925cb69338508bf9de928b1817c6f117c612635bbab3
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:6667919a41d304d5d4ade3ded4f11b42d722a995a4283e11e15320529f7f9abf
        version: 2.3.0
      entries:
      - name: cockroach-operator.v2.3.0
        version: 2.3.0
      name: beta
    - currentCSV: cockroach-operator.v2.18.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v25.2.2",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
          createdAt: "2025-07-22T11:30:56Z"
          description: CockroachDB Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9ab968e4ca62f23bf9848e5475ef4076df7318a8560f63d5b2090a32af6fe4e2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:90cf21c36d19d87b97d06edd1c0912a3bb288c769e6159651877db3206aa7355
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8f1f2fd1aca914a1ef007117c85ac944055bbfb873a468a7374d239ef20b065e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9b5a2e0c006eb4be8e1444734a98743551f278a24e81c20bebc96f0f3b631eb0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6371377718325235a2720dba0c97eb4321ee29b31a28cb893f76abd62baa339
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:68fe64fe3e112288ccfc2a7dffd69b9b2786a5da3d3bdedd6addf41659d460c9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5240f9f591263569fad7ee934e81fa8a1617de3c25da94ddd59ee5b2b3407f90
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f29c31bb267451b389608af3e8a4a28d6ca0691e67b77e8f57bb5e461b064893
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8737b67b91983817e3e4ff3f47f0b070bd36162d309b60993529426c6979cb24
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a3ac544d87a9d8e7c92897c4c23f803753c7ab31467db91b5c51244afc2c6be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf001173bbf00cdf443b8d64a3ad724389f1cf83e3628d0b609019c8cddad35a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce3d94a3fbdcde5858beb3c265987759971d991c397c6486e19904bd4706f44d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7d6320ac7fb0925f10cf6f524a13de629cbf13ccc371912e4fe8c777bc15d1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bd75aa918355e77cd671c3e11ee807861378e96245245f4fdf243cba5749f9a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d590946b2f7c7ec0c7ec057a338e81f864d6a9e3d3ac4e0769aa5756a8e13fc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:32fdca575c334822e4356aab36a7ed97b685c065925fe85f1b8ba8425c57159b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc4b75ed71845d5b18e1d9163d480fd9d0d3cb5ebbcfed00a2e4e1b174c0a5de
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f74c8a5dea4560ad59c6f6f6ab1e699f063334ee8db704460f870508dbeaeff0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4b3aa590623a38f435b075de4c09f8a343d8ece8b673e6bf7c6df5d5eb4bdf4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:46d9287ef11e3b21a36abfa31a26da3b93040bb6163b9223972a5dd1930fecff
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e0f21abea34825e55622c432a306f2e8aafd64d913eb78f2f601c51dc878a1da
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2b5c028f293c6f394e58a5495cb3dbfbedc215d0d03778d0217e045586d55a46
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08cec123813304ab03bd66d72a09949667a874a42320bedafda391ebcdb2e56c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6a2ee8d62c586eb74ebbef6bef256e06e1ba5cb51c1ae55b1654adafc2ca8859
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c2a62179e8ca293d7e30b9da599e48795345cf07a72abd3a83063107b53b03cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:72ea3d0ef72972dba93362dc31366c5946a9898f5c16780b877316a97068ef8f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1304d7f871a95c8b087bad1b548217562792edc66c8fe07c621e05242b8cd53e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9492894eaa5f43d00ac80b60029b32a0b0a85ae88ba609b46f98702c9349a5bb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bf36dab9b3257a4bba24d986f1451fc51a3a37023fbb988281f1941fd3faedd
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d467383af41aa80c26172964c34152abeba45121599804e502984655b72179f0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:022c6e1a0de875474bc953718a481a0af526c49dc910094dcdce0b5677aced18
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2d89398f03736f8ffaa82b10fed0f19fba166d3d2145b26f25cb71f3c39acf07
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e23af8f9d32c3f80a52d79ceb4b359dda1d75ba0dd4cb8ed4b66ac2f4e40c69b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:61e602b0e3582bf00460eb9328808ccf66e3ff1b1f6df07ee6c5c76a91846876
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8443c15eb1653dd9e5ec7aad54d4b8b092d41a0977e42b76e372e5e21280746e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0ec1dc184a2199736b4bd7de3241a357be7427caea5da1c931b08bc06f5c3dc0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c9a31bdf19175880d0981e21dd19a1458b6e8cde8d4acbcb554f3bea3578a64e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3f587e7b9f5a611013af2608e2151ed782859a52501b00adbc95739c56a8ff1e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7ad12df8e258b03c4099d375d02d272bd774bf188d108901d5cf6cb865e57ef3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4f753151af4c3d3f5a1df9014461316249a6d0621b8fba5c4430307b3355251
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:58de9d0b2ce005c7ad3e99152100e59b7677398557028a7a59e56e669fabb81c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1f6b1ec92251397dbcf7ed4a82f1f05619d64676aae2717f4eff1c33e4c2c27a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:266165ab87b19a05d6796e8b15ff0eba632d6ac96e6393400464c6136dd500ec
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:086d7b435993fdf21fd59b0093f52105a9028d6b769398b0033f98a65dfb7e79
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc1ab77a9537ccf236ce6dc370940dfedcfaa91b6e3db37bf2d5eade7ff84708
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4d8e804b35b73d7978c112748aea2d2bb68f65081b90bca62c39f2b6e2989378
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:744193037d6b64cb73fecab15ebe022abc2b79ab0335b76ec1415ac6379b7dfa
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5181196fff3bf8a5ec8fcfe4f1958cc69f285f0beb4f0b476b4b5e0912f3db42
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:40f00476b63cb9e370fe85f01407173693213a9242fc117d28b06378ca0d98e0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fd6c9b8ca82ff4f2d0fe29166d1fc9bca6439cb40907be60f4d8f1b9735de2f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:85a0e085901c2a73192556d5d70795d4b66cbcfe01d24c73b6fe6db8d5f2e5ca
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6bb15b36d64eebb6e4c1db5a5466e108b271d53383c58e0b6c78cec214756a9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b83eaff2e217c06098509709cead525b55b00ccd48019d8357e6231892008055
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:858dc82977b3c9e989a3a20626671b0c67c834c97132094a9212bb227ab55929
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da367cf0ac52045002e1c186f8e6964267ad87d5c25f9e72fd2c9b9a98a32702
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b77b80eed7ce149d2a3f228e5f188e76d73f32bc50ecde20bba9cfa7bdc4714c
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:53c7b960a2ed70a0998a4f8e28a4f78f7f676e01cffebe665e91e3a7629d88ed
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1558a1f5eab79f1a8c6c85d81740cc539dc23025ea457c0019d93a72dc05d57a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db4e5c2f27ba23e95a363c8174c01f0f9c6b6a27e0e581c86a38f3a1f68dc43c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:37bf9e1ec7e3eff47755fe7521223f31ec6aefc6e63901db3881b30898cf78f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:35ede65457d845e8c063ad7d9515d3a69f98e0751424ecbac484963678a0b1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:50393ad005fed19618d04f0716c6d55e2f1afce651e785c396638e852cac83b3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cad60044ad6573bd80b65e892a9ec0510dbe4fcbbfc6b51010ecf419f56f1024
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f42fa22aac3cc41ce72e144a72530ffbee765a392b24593dafd06cc098f8410b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:31b161f9bf9028a1a62cd2eef6abda01257d8889bdc3159822642549a75e3429
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9c85c25d5392817d17b68714dca6195ad53255d43e9c65b021d75b12d97e0579
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a6e2fe24ad52913e2160f85b9fe665a4e4c4d79b164b4b899f69dd9d92f999e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cc20270212f01fc231ff788e20b0476b8228c440bfc40430c0836d5eb702a042
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1374cbdc4ecf38e4b4226328f532e830b8932b34d9449bf2fbb36e60d0310521
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bd336715332f776a72aaf5c3ca52e62dd53a9e8521c1feb6e135cc868d694a39
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7481c0cf3b579c8d98d94d4aa0334ca94bf2ccfcd25ad70b57c42b22bbf4bcee
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4635550222775fc75737be5e5c0359303e3a23c4d1e1c296d7408cd15ad94aad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7daaf1368ab2c25694cdea0a598e1de59247b04a4b4b4b8348ed6edbaa194f9d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fe4a40006d1e8439423de1417cc9bc55716d790b882485074bb2a5ec68d7d1ad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5b3de46526d169fcd1318354545b3a18861843fa2022a23ed04d795493825287
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acc083f7dde1fbc557d858d03fc26b68b2056997fc1a457e4695650214d5b303
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0b0e0588c1aadf0c3826cd584bd2b3a7d6781882dcb5c5e037b47ff5cb84509b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3bf017b2b2506724500f5e29017d1b956825dd0f61a7c0049b0ff084c3554b5b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce1209ca6cc8724baee981c97c66e2852e512423bc0c0a16cdd80a6838b54031
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e82684b828256b3c2814930210031b3b70bc818514d4359a25f87ffdbf22be6e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:19f36f53f7da67755eb86da77098c5eeb84b381f9c18aa29ba573793dc498564
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f5a0ccc02dc9e938e484d5b5282ff650d1890d5f754c30a0c0d717989ed5d600
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9d0bfb642090ff55a3cc3da0f72663cae18063622a4ec785ba99bc71ef4db116
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f045fe5037d44aab1a0b2209241b3931ec7cb9897cb3c6a96a9f8d5280e55dd2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:44a0f40a6c46ba08297b482f3d96defe87b181c944803e5381c9acc42cabf97f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db07461824e73410456ef540b0bce3c7fe550aa67a73d6174300ca74b8f75276
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0754f7901feea9017bc5041cb91b7ac1cdbca10c2064f37c3648a2c395f24f83
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:28ae9661b38a72decf413c6a6c3758e61098debf5162197c27a12a070c7a096d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:743cda1eda6e273c798e3405a9a97e07dd6aa5ad4d5a52789be3b0f39a32a039
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:91f7e59f9e35a0395756eafed906963c535e886125294d671d930ce959a8f511
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d78c2e0709214f96794564f13386861f4b541e55b5bd316ad495cf104f7d40b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da3bd2e9ff2b08d5bb86b1f5faa95a865f85c5e8912df00b6b79190c1e518a2f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f3ffebd812881a830ff46c860984814dd61edcd51b478ebd3667759fb7710682
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5d6901b6b6fd6594d42bfc1842d050c7cffcb0867a5591a67469f0f860b1f21b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e8f26fd132affbaeb8ef6b1807a646c83b301e5dc334a592dfb2cf8dc3dbaf10
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c74cf927a2b2c5098ad74ad0880996f8b042186a9911d888a231ef5b9e2a0715
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5a54d564e08baefc2786f4b742b2507d31984664cd00333406aa7a3aa5b9c913
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:310310515625f099a928545865f7096997871ee4a16650a01c76c3799a18b684
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a0faa52a999289b308dc4caf050c6d9fe2f42e112aa30e4af282ed1f21c5c346
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a59cafff8d4ff2c62204ece15880117011c30978823c232902c7182d747ec827
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1e5d85f5db10fa79d09263db90daa44549bcc3b025e9f3f98cf0d4e51766394b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6ca927f137be28481e3a226e6c6bcb7c5ac27664984f17ffc6c1419cd7d08eb7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e50aab39722af22ccbf3d1db19e37972d52b0fc9a40998a64618dc966b2bac57
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d978678b3c254abd52f69f226b0aacc4b02c2aaca328c54ef10f6f9bb453582d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6408f89b0a88ec154f7ee98c314be574971c4c9aaf906b473d10124d822889be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:468b0a40a33be1dd5a2a89ce745d7e9012bc7371f2f01cc0f8f12c0ac887db43
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b06bd7f3802e8fa2af14bb0b07804cad4cc0fbbc97da18d09a9106bf074ed6d8
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f438fbbf9c87ca0d54c170e2c5105b1e265182e1c2ede48adea370e2be9dc895
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d8cfb221f67516675fdfde061f6ec5dfee1224ca7eb3d8be375c27d483f306af
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e151bded670f4cbd165a8ef864324805667d3e84ce3ace0cdf8f9aad087fccf1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cbc74889e9ca5edbd5aaa243d7649bca0c4da2719239f8d0ec9827868112a771
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a6f1c43486d044fed9f59406118f429c8037b389bd8c8d7a1c7aeb89ed2661f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2af60025ecb3bb933b61328be4c2b2bfd0e7d26f53b72430629208505e45c6d2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:79d688bebe0643ef0dd6ba7108cdb34c0442470c0c20ae0df88b5d02a3476cd0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9097e135159f46396ef4649f724d61bb5b9113c488f989d404586fbb7486fd85
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d14f77f70aea0422a906510c1ac2652ff76e131ed705cb002f80e13e887b0cd9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0fea6ed8a78ea624240ee6ffb3573d5cf1115186aa9180fcd7d1273351deaaa3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e18d73af3257c8cdf00245f2f718b56667524c1cc69d3db28e9b868a539ae16f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:dd521d7bad62da054eec54179be75c643a5492d981e11c0e66db41902cd2182c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:363d8feb560324a848c5771edd224dc19020013af7641979ee81ae8ee536d7e3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7b907d03b001b758d8a4b5e898725c73dda2ec2aa4b7425b685320b947154d11
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4b56c0311af14220a466aa11f25fd10fa38280825539bd02215c945bbafb861c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:42c9930d6861f6a2147ed6e01827eddd730279966763ec68868567a19cc164ab
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:740ba10f1ef0a6c5fa6ad31ab18855fc75f8e024ce4c23b0676c5a9c5b62cd75
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8e8c081c231b5b15eca483ceec16ea9190ba73aae6e421865162446eb584519a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:56109e57ee0379cf48644bcf8226a9238e01139cacc5499002c99f973f121911
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:742a6f69a3a209a7080a1e911fd96467d2a91aacb3f91d6be80a6a6e2e54b2ae
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0e2c7023f1f97acc20c46a4944635a9c5f57bb9bb3dd534cf41fea1a9cfdfdbe
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf7d3bea47638b763b312c7c11ca034a7ac62d993d75a13edceec74799b97e52
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:226d7f1c3f0d70ac6532ad9490bebcdef44a40959b6aa7c11b0f2d1383a46217
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6cd9f9b9fcb6e8f9a931a6ad01cb66329a74d40b22f9029fbc50e57248bad7cb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1ce71e80c0cd0e0063eba275ab91f77982f2e92a8aef8e89cab99b93feab0a88
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1b6d934a7b6c42a61e5ad51380e69af341a575b42205299fd58bc0c6ff306f6a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b046aa0a42dae992a03685ed82a20835532d6d4bf048e2366ff7382be4b3138
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acbfee4492327a720fe7da76929a86192d8af2442bbb76f0a089f6a81a890be3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:818f592c50d4ed5599faa053344f7ecd963003352f1faa3b72dc0f5e00a0e78b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4e5f7df1dc1e1db398c36d590431e7e5782897b209972d8e9e4671971c10d1b6
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5458a8fd89aa664b60e59c35bd933975cd37d18e19b11589e4ca0c2844ea0ed4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2a128b570352a5a7f82b27353f9b774caccce83e0e7bb2673e4802d96ad0bd51
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2171e1c75d0c665fe33e1ea334e605604c51de6dbc22d88a3a0ac604259c448f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ec6eb7c28c213cc83b2d7919cd87988f9a07f12276eb7351d0915f1567a5b095
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e5814748c5ab229a1dea5662a007063c0df06fbbfcfe041fd9a6df9ef67e07cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:82d8104a37349c513447e68ea3790efb256370632353176e5fc13c2e8df1bd44
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b78a4d3864e16d270979d9d7756012abf09e6cdb7f1eb4832f6497c26281099
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9f42e929f2cc7e9b6d24dd61e19c5d93dcfeabd34b9c8b935afbe78577efb18b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8fc8de4f3671a73f9aa747e5b23ec973d192fff6f2869c38985cd656063c1e3d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:02e1ef01a023da0eee653f63b314786fd66a78d184a97a314422f6f9c3ef8e7e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:65923c37fecda2d3174212fa9d9a6f34241d065389336ea69a7be99ba16cedc2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bed27f3f2c2d7627d1af4500d7ea0302da5c3fd3d875a6480fb6d98da481312e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:10b9df672260c023dfb37166e156c0d7e0e13144dc3e485af922e8151fd3ab05
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f05824bb8eaa5a93e54059449eaa236dd83d729049ad00065ef3242f5f74c91e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:87db6e079b96fb392d98ef5fc7063836bff10717efd75061016ce688be0ba38e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:29abb169ffefe36f6d69edd438a7135ee23059cedab8f87ad3b2d52361b96b35
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a4f932c3e9ff3aeb70ba1f7a92aa4642bd14cbb7d37d04ff750ed517b916cdb9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08496b84b3cc691909227be17e67bcd6f4d5185a93712f44d8a0985bd2c5fcd7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f53b07212320c6939d3bb4ddd4e6a0bce311ab5b63858eb1a8728531f0dbd510
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b0057de903654ea0315a2c96affcc8f50a92ede8ad121ed30edb3aaf0fb7b554
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d0014b1c19890b9e8ef1a4f3efe7eb5b269be5610ba6ababd4258b52d2f85e8a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:aeac9cd8e8acdbd2b0c339c48b85079c159b04f0ade24df8e3f9cae3f48b4c70
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6366ba29f301a82215c05e60d78e1de66bcee85440ce2fc3c692712d4da32685
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8a70ba12f610ac0e62066e417a0833bc1d01e91ac99d0dd794a2c0cd2a350164
        version: 2.18.2
      entries:
      - name: cockroach-operator.v2.18.2
        version: 2.18.2
      name: stable
    defaultChannel: stable
    packageName: cockroachdb-certified-rhmp
    provider:
      name: Cockroach Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      control-plane: odfmo-controller-manager
      full_version: 4.15.4-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-compose-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-compose-operator.v4.15.4-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "labels": {
                    "control-plane": "odfmo-controller-manager"
                  },
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              }
            ]
          capabilities: Basic Install
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.4-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          This is a CSV used for mirroring the entire product properly. Do not install it please.
        displayName: ODF Compose (DO NOT INSTALL)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: branto@redhat.com
          name: Boris Ranto
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-operator-bundle@sha256:3773158dbd141530cd851f36c3c8a5f54dc0892d003633118e123f5862923666
        - registry.redhat.io/odf4/odr-cluster-operator-bundle@sha256:4c7049b6137b3d49fa9b5e51e0cf7cd430d70c8790e0ed017e6bd811fcc506c8
        - registry.redhat.io/odf4/odf-csi-addons-operator-bundle@sha256:b2b8144c34df667b13987773b11df2fc8ac96d418999292303ef2c03cd3c79d6
        - registry.redhat.io/odf4/mcg-operator-bundle@sha256:9ce0ab75f56ebeeac63aefd0f48cba671da3377ac628830b3c03be9569f8c87a
        - registry.redhat.io/odf4/ocs-operator-bundle@sha256:1e91c5c6c122fdccab52c37ff2cef76317af80970dbdc06aec752115117b2fa0
        - registry.redhat.io/odf4/ocs-client-operator-bundle@sha256:8904c171ea317249e9a19c3a7638f688357a0b238582330b1bddc23a11a9925e
        - registry.redhat.io/odf4/odf-multicluster-operator-bundle@sha256:f7e1c75ee6b1a62fddbcc1e6a718a9cbd5ecf364e4ef10f29da7a15bc82d5b17
        - registry.redhat.io/odf4/odf-multicluster-rhel8-operator@sha256:861e56167698558109197b515aff93691f8d0fdba19311be18a32512f6896139
        - registry.redhat.io/odf4/odr-hub-operator-bundle@sha256:862a9d983013c16b7dfd5194ab5752c26a473ca5a4575bbcd80be59f97f8ac27
        version: 4.15.4-rhodf
      entries:
      - name: odf-compose-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      name: stable-4.15
    defaultChannel: stable-4.15
    packageName: odf-compose-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: unsupported
      operatorframework.io/arch.ppc64le: unsupported
      operatorframework.io/arch.s390x: unsupported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: mta-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mta-operator.v6.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:f5d10cd2528bebf5094c7744e47d9de6fe827fd09024c5ffc0e1b74f1a6cc3b4
          createdAt: "2023-04-26T14:04:20Z"
          description: MTA contains tools that support the modernization and migration
            of applications to Kubernetes and OpenShift
          olm.skipRange: '>=0.0.0 <6.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/tackle2-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is a collection of tools that supports large-scale application modernization and migration projects to Kubernetes and OpenShift.


          Migration Toolkit for Applications allows users to maintain their portfolio of applications with a full set of metadata and to assess their suitability for modernization leveraging a questionnaire based approach.


          ### Install

          By default, MTA requires a RWX capable storage class for caching purposes, this can be disabled via the rwx_supported parameter in Tackle CR.

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-windup-addon-rhel8@sha256:a4200fc0a39df67eae31c53ee9ae0c05769fb410d0312ebd0bbf8bf1567b2a48
        - registry.redhat.io/mta/mta-hub-rhel8@sha256:c75e1d9d3307d95be392a53440d9123601a6ada3693e3f2573e03022565e4339
        - registry.redhat.io/rhel8/postgresql-12@sha256:3a69efa85af6a6f95665eca69085b007fa1fcb645a37f356895aeb4c9527a621
        - registry.redhat.io/mta/mta-pathfinder-rhel8@sha256:a66dd030eb99207b6aaacc6b7c7cb5182901da836ebf50373c14f8e5ecb7b642
        - registry.redhat.io/mta/mta-ui-rhel8@sha256:de6cb95ce3663922ecc725b186876b043a9c7794e95510459ffbcba7d668fc26
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:f5d10cd2528bebf5094c7744e47d9de6fe827fd09024c5ffc0e1b74f1a6cc3b4
        - registry.redhat.io/mta/mta-admin-addon-rhel8@sha256:812b8dc621a06bb99cf2c59abb8257da3ceeb165c563e14d258e69803a557cc8
        version: 6.0.3
      entries:
      - name: mta-operator.v6.0.3
        version: 6.0.3
      - name: mta-operator.v6.0.2
        version: 6.0.2
      - name: mta-operator.v6.0.1
        version: 6.0.1
      - name: mta-operator.v6.0.0
        version: 6.0.0
      name: stable-v6.0
    - currentCSV: mta-operator.v6.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:6da71d4caecfbe049e63576d4325d4919002196da2e875f79703d8ff0d5a1705
          createdAt: "2023-10-26T15:29:26Z"
          description: MTA contains tools that support the modernization and migration
            of applications to Kubernetes and OpenShift
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0 <6.1.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/tackle2-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is a collection of tools that supports large-scale application modernization and migration projects to Kubernetes and OpenShift.


          Migration Toolkit for Applications allows users to maintain their portfolio of applications with a full set of metadata and to assess their suitability for modernization leveraging a questionnaire based approach.


          ### Install

          By default, MTA requires a RWX capable storage class for caching purposes, this can be disabled via the rwx_supported parameter in Tackle CR.

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:6da71d4caecfbe049e63576d4325d4919002196da2e875f79703d8ff0d5a1705
        - registry.redhat.io/mta/mta-hub-rhel8@sha256:32624f2bddfe8b6b9c6c1c3f2be541c8c61be29680d13b6d446e8874bdf8e219
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/mta/mta-pathfinder-rhel8@sha256:1df881da1e20c2501caf53e1cf59b6b366281d842f0948ca5f91694ec1db829d
        - registry.redhat.io/mta/mta-ui-rhel8@sha256:3845d6d246003ebeb49bcf1d95d6be2eb9cd05a280a5bf8320b201e878a41427
        - registry.redhat.io/mta/mta-windup-addon-rhel8@sha256:6bcaf8e8b63e169652d6568d77e91a289084e621a586105a2d429b5726569058
        version: 6.1.4
      entries:
      - name: mta-operator.v6.1.4
        version: 6.1.4
      - name: mta-operator.v6.1.3
        version: 6.1.3
      - name: mta-operator.v6.1.2
        version: 6.1.2
      - name: mta-operator.v6.1.1
        version: 6.1.1
      - name: mta-operator.v6.1.0
        version: 6.1.0
      name: stable-v6.1
    - currentCSV: mta-operator.v6.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:6938c3fc6dd8fb032e1765fc11d9586dbfd1aaba4b423b0d87da4120ee0bb60d
          createdAt: "2024-06-10T22:12:58Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0 <6.2.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-windup-addon-rhel9@sha256:d1f4939b92dd13c15cb7344ec47a571669f6d108dc2e7d795ce29350f2cb3838
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:6938c3fc6dd8fb032e1765fc11d9586dbfd1aaba4b423b0d87da4120ee0bb60d
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:e59725832547e80cab58077a1b6e42c313a4057621defa841f35c0a7627f149d
        - registry.redhat.io/rhel8/postgresql-12@sha256:c4716bee5e2cf2e43698bc99e5aa42f16a0f2d68c1fbdcd3602696bf88e408dc
        - registry.redhat.io/mta/mta-pathfinder-rhel9@sha256:a2a8b6d204d99494a6a005e8cb1fbd92afdc84dc01e4c1de4ad7d8798bd83741
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:c35f3ed429a7d351cb8c2d07600b4ad35615a5ba08e5cc9f8b9bd504aea3f7ee
        version: 6.2.3
      entries:
      - name: mta-operator.v6.2.3
        version: 6.2.3
      - name: mta-operator.v6.2.2
        version: 6.2.2
      - name: mta-operator.v6.2.1
        version: 6.2.1
      - name: mta-operator.v6.2.0
        version: 6.2.0
      name: stable-v6.2
    - currentCSV: mta-operator.v7.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:268ecf5efc51af71f9b2bc2d01b410f8cfcf81a948b82d2d476fef6c0cd3e033
          createdAt: "2024-05-10T19:37:30Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:268ecf5efc51af71f9b2bc2d01b410f8cfcf81a948b82d2d476fef6c0cd3e033
        - registry.redhat.io/mta/mta-windup-shim-rhel9@sha256:e2ea89055936d7ad229b5e890c6ba234c6cb22d8b688515ed4d6a34848efea83
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:d554cc36748fc05e30eb10e6fe816ff6aab2fe07b61cc60e9d93de5c3cf17748
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:0e1ec99af5666a060176adc52612f20209e3398a9a5a2aa1dcff11ef260b2006
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:1b5e016046d93b6a2674d63f08e0a46f8050e785bd58fd94ba7a70487efe7e2a
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:18c611415589a093e16e9c2f2fba9f3c83d9514f79c8e4fde9704ef0679ea4d5
        - registry.redhat.io/mta/mta-analyzer-lsp-rhel9@sha256:0c8f59f35a950ea8b45f16cf71362128ccc335b58a996a099384fd737dd6ff90
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:af879d8b9b5e396a5308fb675a1d6dba6d70151b9bdbe7829362e081a6adab1c
        - registry.redhat.io/rhel8/postgresql-12@sha256:da619f6617d99790b4fe00fdb23ab4aea0cdb0592590162d51b3ef94aca01a7b
        version: 7.0.3
      entries:
      - name: mta-operator.v7.0.3
        version: 7.0.3
      - name: mta-operator.v7.0.2
        version: 7.0.2
      - name: mta-operator.v7.0.1
        version: 7.0.1
      - name: mta-operator.v7.0.0
        version: 7.0.0
      name: stable-v7.0
    - currentCSV: mta-operator.v7.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:08b33b44140397785f8aa1b736796e09cf6e49ac30cf7c386a2fbd8b678d47b1
          createdAt: "2024-09-16T23:45:42Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.1.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:16cb28d683159ba71a5d7c233404ac1d6b7a07a374b46d24effe551a13a596c3
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:23c97ac122eb7cd1e953df8da968da420f6db8708baf772759517aca4929748b
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:e4e70201df23c9e1168646d3c6013a8ffa22481a85b07df491f240310c13ebe1
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:08b33b44140397785f8aa1b736796e09cf6e49ac30cf7c386a2fbd8b678d47b1
        - registry.redhat.io/rhel8/postgresql-15@sha256:16faf1ee82382263408f713ee4002f47ff13c82d4545e8a69d5b0ebd1d289330
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:30a3b1d5a70dec74a9253ca34f18bd8f2cdfaa68b85b0b25dc395158c2596330
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel8@sha256:582f28e609c1a0466d30c5586d6b6023f1b1de3c28f76cabb498322914ab24d4
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:ac38f1c3d34bc718928eb20743835c99783d846f63a3197d13839ea22f1634c8
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:4ad19b50cdd6be21e010155f2026abf0dc865e72aac1353e35bbb5eeb8320a4e
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:ee2c8ce2af0be097353758d9ee7eaa7e6b3fcd5f2bf5e04280e450e2b619d2d8
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        version: 7.1.1
      entries:
      - name: mta-operator.v7.1.1
        version: 7.1.1
      - name: mta-operator.v7.1.0
        version: 7.1.0
      name: stable-v7.1
    - currentCSV: mta-operator.v7.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:917d6cd534c158f6c4a7e9c83f0c66fa85ba81037c352d8a039e198581126a7f
          createdAt: "2025-06-17T16:04:15Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.2.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:917d6cd534c158f6c4a7e9c83f0c66fa85ba81037c352d8a039e198581126a7f
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel8@sha256:c0643c05833dd1b70ed361bf56c46f56ed6536aac3106b92f0d85b83da1b9d15
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:10ef4b218d603432508b5e1f3685fb32a80cbce6c327e35952cdcccb0465468b
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:9d6c02f6463771e747a30126de95fc0728fd95fff4ddfa6c70d828c55e0b6790
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:2dfc51d5398a90e73273b3e332f70d7ed94bac05e12b142fba22104ecc929436
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:b10fdc78c6f453a34a3d306313dc29fcb3152688a6a1d54cd236d00d9fab88af
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:4cde2c2df482293bd4d6b86a344336e13d2f9f81b29323698f603e656b14b190
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:70ff3b5e429a8fe0c68752ee05c47ead885b59bf9c8d7235337801270e7fc1de
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:c9447659195b9eebb8f8a265fd5052ca9b2611d57080017448297a33409863c6
        - registry.redhat.io/rhel8/postgresql-15@sha256:dd4aec2775da1cf187c0839384cb16d69268e9781d4330874bbc865f98153a7a
        version: 7.2.3
      entries:
      - name: mta-operator.v7.2.3
        version: 7.2.3
      - name: mta-operator.v7.2.2
        version: 7.2.2
      - name: mta-operator.v7.2.1
        version: 7.2.1
      - name: mta-operator.v7.2.0
        version: 7.2.0
      name: stable-v7.2
    - currentCSV: mta-operator.v7.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel9-operator@sha256:6abd6541a955543207c1ff4b903928debcc89dd6ea3d7aeaa3e47465849f7b37
          createdAt: "2025-08-15T16:32:16Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.3.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-15@sha256:854710b428c4c1132c448cbe9ad30087a8a4b3c33ae9c0c3a4de21a15c0ae642
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel9@sha256:9748d954ad8654ac203257286497ccf19cfdfa211fee40449cbe732ff48c1176
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:32328fc0459099ef1de0f92e064c15d29a46d88dfe2f54c88f822d44ab82e3f4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        - registry.redhat.io/mta/mta-rhel9-operator@sha256:6abd6541a955543207c1ff4b903928debcc89dd6ea3d7aeaa3e47465849f7b37
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:b2ff154a4b4ead22311005b7115c02346928b5ebf77d776ef909ad4af1f350bc
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:6118c1db15667851f26eb2acf6060e045ed9065459e3e649bbdd0e2625d825bd
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:072f676323e9abcb2ff22eb5e7a70f6681e9d885d8666ddd81060fbad54c443e
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:8ca00460b2fb64583a247f5b906c19743b7732b60ff8215f1baec12a766c61d1
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:80aa13813fe121da49e166d81d09d63749393f3375c3b96d762d0b09c6685ca9
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:ef57a58e609e244333d35681a164a1d3bf382c0eee37d33c993b018439814198
        version: 7.3.2
      entries:
      - name: mta-operator.v7.3.2
        version: 7.3.2
      - name: mta-operator.v7.3.1
        version: 7.3.1
      - name: mta-operator.v7.3.0
        version: 7.3.0
      name: stable-v7.3
    - currentCSV: mta-operator.v8.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel9-operator@sha256:83c2b804667d28f782d64132a0c393194083d10c820b88df49c2503d3351fd8b
          createdAt: "2025-09-26T03:03:31Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.3.0 <8.0.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Schema
            displayName: Schema
            kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-platform-addon-rhel9@sha256:dcb35178810b7eddc2f2713434d377003f8f4fb7022a40f6c1926b74b13fe23a
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:a77ee9b76225b6cca667fa3402492b0845587fe5e400fa8d141b5ab636c56811
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel9@sha256:2a6f0b392b86fe16244d5bf9fcb84ed9770a22329a96e9861f03872b7209fb04
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ef10b2ecbfa0c4ac8568af9ee64ee4a554e3fdc9089ee5683b6c2c3a74caafbe
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:616d04cfbab92341157c42eee0683f31867d81c5a289723da8da8208d1054774
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:aaf7bf10b1af107ad55289caab2cf9dd052e69211968d2e8d05c1d25f21ed2b3
        - registry.redhat.io/mta/mta-rhel9-operator@sha256:83c2b804667d28f782d64132a0c393194083d10c820b88df49c2503d3351fd8b
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:b1b38e218ea37b529f344eaa1f0f3b55b716ad7953e852c19f34f4a25ab6c5a7
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:af2c1790f2b98ebcfadb5d6bdd43cb9e7e96cdabd5b653d48d0fb67901781c49
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:460d1d45ed44ae9328ed93c5fed240e154549bbec662dc1fc55569ff7256aca9
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:8ae0887fa27d6f4281e5bb9b629bf6b64a9c460948a002839ba42da9b25b5521
        - registry.redhat.io/mta/mta-solution-server-rhel9@sha256:e9d48d11ae203108229d67bc96721daa1c3bb10160b9d6d957cd2e0ba68a2ef7
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        version: 8.0.0
      entries:
      - name: mta-operator.v8.0.0
        version: 8.0.0
      name: stable-v8.0
    defaultChannel: stable-v8.0
    packageName: mta-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubemq.io
      provider-url: ""
    name: kubemq-operator-marketplace-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubemq-operator.v1.7.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
              "apiVersion": "core.k8s.kubemq.io/v1beta1",
              "kind": "KubemqCluster",
              "metadata": {
                "name": "kubemq-cluster"
              },
              "spec": {
                "replicas": 3
              }
            },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqDashboard",
                "metadata": {
                  "name": "kubemq-dashboard"
                }
              },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqConnector",
                "metadata": {
                  "name": "kubemq-bridges"
                },
                "spec": {
                  "type": "bridges",
                  "config": "bindings: null"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/kubemq/kubemq-operator:v1.7.2
          createdAt: 2020-03-11 12:00:00+00:00
          description: Kubernetes Messaging Platform
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/kubemq-operator-marketplace-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/kubemq-operator-marketplace-rhmp/support?utm_source=openshift_console
          support: support@kubemq.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kubemq Cluster
            displayName: Kubemq Cluster
            kind: KubemqCluster
            name: kubemqclusters.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Cluster Dashboard
            displayName: Kubemq Dashboard
            kind: KubemqDashboard
            name: kubemqdashboards.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Connector
            displayName: Kubemq Connector
            kind: KubemqConnector
            name: kubemqconnectors.core.k8s.kubemq.io
            version: v1beta1
        description: |2-

          ## Kubernetes Messaging Platform
          A message-based architecture for microservices providing flexible connectivity for hybrid environments: multi-cloud, on-premises, and at the edge. Powered by enterprise-grade message broker and message queue with ready to use connectors, bridges, and control center.

          Messaging platform components:
            - **KubeMQ Cluster** - Enterprise-grade message broker and message queue, scalable, high available, and secured.
            - **KubeMQ Targets Connector** - connects a KubeMQ cluster with external systems and cloud services.
            - **KubeMQ Bridges Connector** - bridge, replicate, aggregate, and transform messages between KubeMQ clusters.
            - **KubeMQ Sources Connector** - transform external systems and cloud services data to KubeMQ cluster messages.
            - **KubeMQ Dashboard** - monitor KubeMQ cluster metrics.

          ### Kubernetes Native
          Innovative and modern message queue and message broker in a lightweight container developed to run in Kubernetes, certified in the CNCF landscape, and connect natively to the cloud-native ecosystem.

          ### Ease of use

          Simple deployment in Kubernetes in less than 1 minute. Developer friendly by simple to use SDKs and elimination of the many developers and DevOps-centered challenges to define exchanges, brokers, channels, routes, and predefined topics

          ### All messaging patterns

          KubeMQ is available with all messaging patterns such as Queue, Stream, Pub/Sub, and RPC, Saving the need to maintain multiple messaging platforms for different use cases.

          ### Hybrid Environments

          KubeMQ enables the microservices from multiple environments to communicate and build one hybrid infrastructure solution across clouds, on-prem, and at the edge.

          ### KubeMQ Cluster key features

          - All-batteries included Messaging Queue Broker for the Kubernetes environment
          - Deployed with Operator for full life cycle operation
          - Blazing fast (written in Go), small and lightweight Docker container
          - Asynchronous and Synchronous messaging with support for  `At Most Once Delivery` and `At Least Once Delivery` models
          - Supports durable FIFO based Queue, Publish-Subscribe Events, Publish-Subscribe with Persistence (Events Store), RPC Command and Query messaging patterns
          - Supports gRPC, Rest and WebSocket Transport protocols with TLS support (both RPC and Stream modes)
          - Supports Access control Authorization and Authentication
          - Supports message masticating and smart routing
          - No Message broker configuration needed (i.e., queues, exchanges)
          - .Net, Java, Python, Go, and NodeJS SDKs

          ### KubeMQ Targets Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major services types (databases, cache, messaging, serverless, HTTP, etc.)
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-targets)

          ### KubeMQ Bridges Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Build Any Topology - connects KubeMQ clusters in 1:1, 1:n , n:1, n:n
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-bridges)

          ### KubeMQ Sources Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major messaging services such as Kafka, ActiveMQ, SQS, and more.
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-sources)

          ## Getting Started

          [Start with OpenShift](https://docs.kubemq.io/getting-started/create-cluster/openshift)

          ## Get Enterprise Trial License

          [Register](https://account.kubemq.io/login/register)

          ## Request a Demo

          [See KubeMQ in Action](https://kubemq.io/request-a-demo/)
        displayName: Kubemq Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - messaging
        - queue
        - broker
        - pubsub
        - rpc
        links:
        - name: Learn more about Kubemq
          url: https://kubemq.io/
        - name: Register for free license
          url: https://account.kubemq.io/login/register
        - name: Contact Sales
          url: https://kubemq.io/contact-sales/
        - name: Documentation
          url: https://docs.kubemq.io/
        - name: Get Help
          url: https://kubemq.io/contact-support/
        maintainers:
        - email: support@kubemq.io
          name: kubemq
        maturity: alpha
        provider:
          name: Kubemq.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-grafana:v4.3
        - registry.marketplace.redhat.com/rhm/kubemq/kubemq-operator:v1.7.2
        - registry.marketplace.redhat.com/rhm/kubemq/kubemq-uni:v2.2.12
        - registry.redhat.io/openshift4/ose-prometheus:v4.3
        version: 1.7.8
      entries:
      - name: kubemq-operator.v1.7.8
        version: 1.7.8
      name: alpha
    defaultChannel: alpha
    packageName: kubemq-operator-marketplace-rhmp
    provider:
      name: Kubemq.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: crunchy-postgres-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/crunchy-postgres-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/crunchy-postgres-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-exporter@sha256:3b231902b2116cd38de05c7378202716c95813b9961fb8cee811b61d36cddf86
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:0e283616c2d32e53258d62d02702bc4e179ef8c03933592154babed1a70bf666
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:a7f62f367817fc08294a0b8b82ac9826a42c0ae668efdfacaa158a4287b8a688
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:c559c896e174674d151f93543ac92a1b5647eb4676313f7dc6cbb8faee786f18
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:56a5e3a5758982e086c9f78201a7137e0d934316f058a31cf30df44639a569a2
        - registry.connect.redhat.com/crunchydata/crunchy-pgadmin4@sha256:0d84acb1ce6a9b046e3f8b222a583ddcc91dda0a117c9f637ef173d0ac992f57
        - registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:b2c3c74fdcc0e08f555e50c527f0d322292333c5d3dce061a0dee3680a1fbc56
        - registry.connect.redhat.com/crunchydata/crunchy-pgbackrest@sha256:b70da0035b74379f2507cfabdd9d1cc1e3c3eecb9066c22f292981c9a27916e1
        - registry.connect.redhat.com/crunchydata/crunchy-pgbouncer@sha256:17cb58dec78447c2f70a414a2aeff6facbf921b42a0f366d0dbf7be95aa8987a
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:4bdd0588ffbb3f628e7df1c72670e7a5087269438343b68779012bc3a13dea31
        - registry.connect.redhat.com/crunchydata/crunchy-upgrade@sha256:68bc93cce61d9b1e44822ae3e3ac8f7a9c2b712eddfaea090eabefacd60c67f6
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: crunchy-postgres-operator-rhmp
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-networkfirewall-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-networkfirewall-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "Firewall",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "FirewallPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "RuleGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/networkfirewall-controller:1.1.1
          createdAt: "2025-09-19T21:03:12Z"
          description: AWS Network Firewall controller is a service controller for
            managing Network Firewall resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FirewallPolicy represents the state of an AWS networkfirewall
              FirewallPolicy resource.
            displayName: FirewallPolicy
            kind: FirewallPolicy
            name: firewallpolicies.networkfirewall.services.k8s.aws
            version: v1alpha1
          - description: Firewall represents the state of an AWS networkfirewall Firewall
              resource.
            displayName: Firewall
            kind: Firewall
            name: firewalls.networkfirewall.services.k8s.aws
            version: v1alpha1
          - description: RuleGroup represents the state of an AWS networkfirewall
              RuleGroup resource.
            displayName: RuleGroup
            kind: RuleGroup
            name: rulegroups.networkfirewall.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Network Firewall resources in AWS from within your Kubernetes cluster.

          **About Amazon Network Firewall**

          AWS Network Firewall is a stateful, managed, network firewall and intrusion detection and prevention service for your virtual private cloud (VPC) that you create in Amazon Virtual Private Cloud (Amazon VPC).
          With Network Firewall, you can filter traffic at the perimeter of your VPC. This includes filtering traffic going to and coming from an internet gateway, NAT gateway, or over VPN or AWS Direct Connect. Network Firewall uses the open source intrusion prevention system (IPS), Suricata, for stateful inspection. Network Firewall supports Suricata compatible rules. For more information, see [Working with stateful rule groups in AWS Network Firewall](https://docs.aws.amazon.com/network-firewall/latest/developerguide/stateful-rule-groups-ips.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Network Firewall
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networkfirewall
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Network Firewall Developer Resources
          url: https://aws.amazon.com/network-firewall/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: network firewall maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/networkfirewall-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-networkfirewall-controller.v1.1.1
        version: 1.1.1
      - name: ack-networkfirewall-controller.v1.1.0
        version: 1.1.0
      - name: ack-networkfirewall-controller.v1.0.13
        version: 1.0.13
      - name: ack-networkfirewall-controller.v1.0.12
        version: 1.0.12
      - name: ack-networkfirewall-controller.v1.0.11
        version: 1.0.11
      - name: ack-networkfirewall-controller.v1.0.10
        version: 1.0.10
      - name: ack-networkfirewall-controller.v1.0.9
        version: 1.0.9
      - name: ack-networkfirewall-controller.v1.0.8
        version: 1.0.8
      - name: ack-networkfirewall-controller.v1.0.7
        version: 1.0.7
      - name: ack-networkfirewall-controller.v1.0.6
        version: 1.0.6
      - name: ack-networkfirewall-controller.v1.0.5
        version: 1.0.5
      - name: ack-networkfirewall-controller.v1.0.4
        version: 1.0.4
      - name: ack-networkfirewall-controller.v1.0.1
        version: 1.0.1
      - name: ack-networkfirewall-controller.v1.0.0
        version: 1.0.0
      - name: ack-networkfirewall-controller.v0.0.10
        version: 0.0.10
      - name: ack-networkfirewall-controller.v0.0.9
        version: 0.0.9
      - name: ack-networkfirewall-controller.v0.0.8
        version: 0.0.8
      - name: ack-networkfirewall-controller.v0.0.7
        version: 0.0.7
      - name: ack-networkfirewall-controller.v0.0.6
        version: 0.0.6
      - name: ack-networkfirewall-controller.v0.0.5
        version: 0.0.5
      - name: ack-networkfirewall-controller.v0.0.2
        version: 0.0.2
      - name: ack-networkfirewall-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-networkfirewall-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: opdev.github.io
    name: spark-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spark-helm-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apachespark.apache.org/v1alpha1",
                "kind": "Spark",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/install-disable-crds": "true"
                  },
                  "name": "spark-operator"
                },
                "spec": {
                  "affinity": {},
                  "batchScheduler": {
                    "enable": false
                  },
                  "controllerThreads": 10,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/googlecloudplatform/spark-operator",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "ingressUrlFormat": "",
                  "istio": {
                    "enabled": false
                  },
                  "labelSelectorFilter": "",
                  "leaderElection": {
                    "lockName": "spark-operator-lock",
                    "lockNamespace": ""
                  },
                  "logLevel": 2,
                  "metrics": {
                    "enable": true,
                    "endpoint": "/metrics",
                    "port": 10254,
                    "portName": "metrics",
                    "prefix": ""
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podMonitor": {
                    "enable": false,
                    "jobLabel": "spark-operator-podmonitor",
                    "labels": {},
                    "podMetricsEndpoint": {
                      "interval": "5s",
                      "scheme": "http"
                    }
                  },
                  "podSecurityContext": {},
                  "rbac": {
                    "create": false,
                    "createClusterRole": true,
                    "createRole": true
                  },
                  "replicaCount": 1,
                  "resourceQuotaEnforcement": {
                    "enable": false
                  },
                  "resources": {},
                  "resyncInterval": 30,
                  "securityContext": {},
                  "serviceAccounts": {
                    "spark": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "sparkoperator": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    }
                  },
                  "sparkJobNamespace": "",
                  "tolerations": [],
                  "uiService": {
                    "enable": true
                  },
                  "volumeMounts": [],
                  "volumes": [],
                  "webhook": {
                    "cleanupAnnotations": {
                      "helm.sh/hook": "pre-delete, pre-upgrade",
                      "helm.sh/hook-delete-policy": "hook-succeeded"
                    },
                    "enable": false,
                    "initAnnotations": {
                      "helm.sh/hook": "pre-install, pre-upgrade",
                      "helm.sh/hook-weight": "50"
                    },
                    "namespaceSelector": "",
                    "port": 8080,
                    "timeout": 30
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "ScheduledSparkApplication",
                "metadata": {
                  "name": "spark-pi-scheduled",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "concurrencyPolicy": "Allow",
                  "schedule": "@every 5m",
                  "template": {
                    "driver": {
                      "coreLimit": "1200m",
                      "cores": 1,
                      "labels": {
                        "version": "3.1.1"
                      },
                      "memory": "512m",
                      "serviceAccount": "spark-operator-spark",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "executor": {
                      "cores": 1,
                      "instances": 1,
                      "labels": {
                        "version": "3.1.1"
                      },
                      "memory": "512m",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "image": "gcr.io/spark-operator/spark:v3.1.1",
                    "imagePullPolicy": "Always",
                    "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.12-3.1.1.jar",
                    "mainClass": "org.apache.spark.examples.SparkPi",
                    "mode": "cluster",
                    "restartPolicy": {
                      "type": "Never"
                    },
                    "sparkVersion": "3.1.1",
                    "type": "Scala",
                    "volumes": [
                      {
                        "hostPath": {
                          "path": "/tmp",
                          "type": "Directory"
                        },
                        "name": "test-volume"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "pyspark-pi",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "driver": {
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "serviceAccount": "spark-operator-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "gcr.io/spark-operator/spark-py:v3.1.1",
                  "imagePullPolicy": "Always",
                  "mainApplicationFile": "local:///opt/spark/examples/src/main/python/pi.py",
                  "mode": "cluster",
                  "pythonVersion": "3",
                  "restartPolicy": {
                    "onFailureRetries": 3,
                    "onFailureRetryInterval": 10,
                    "onSubmissionFailureRetries": 5,
                    "onSubmissionFailureRetryInterval": 20,
                    "type": "OnFailure"
                  },
                  "sparkVersion": "3.1.1",
                  "type": "Python",
                  "volumes": [
                    {
                      "hostPath": {
                        "path": "/tmp",
                        "type": "Directory"
                      },
                      "name": "test-volume"
                    }
                  ]
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "spark-pi",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "driver": {
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "serviceAccount": "spark-operator-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "gcr.io/spark-operator/spark:v3.1.1",
                  "imagePullPolicy": "Always",
                  "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.12-3.1.1.jar",
                  "mainClass": "org.apache.spark.examples.SparkPi",
                  "mode": "cluster",
                  "restartPolicy": {
                    "type": "Never"
                  },
                  "sparkVersion": "3.1.1",
                  "type": "Scala",
                  "volumes": [
                    {
                      "hostPath": {
                        "path": "/tmp",
                        "type": "Directory"
                      },
                      "name": "test-volume"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/spark-helm-operator-controller:0.0.2
          createdAt: "2023-07-14T17:57:44Z"
          description: An experimental operator that installs the spark on k8s operator
            from its helm chart.
          operatorframework.io/initialization-resource: "{\n  \"apiVersion\": \"apachespark.apache.org/v1alpha1\",\n
            \ \"kind\": \"Spark\",\n  \"metadata\": {\n    \"name\": \"spark-operator\"\n
            \ },\n  \"spec\": {\n    \"affinity\": {},\n    \"batchScheduler\": {\n
            \     \"enable\": false\n    },\n    \"controllerThreads\": 10,\n    \"fullnameOverride\":
            \"\",\n    \"image\": {\n      \"pullPolicy\": \"IfNotPresent\",\n      \"repository\":
            \"ghcr.io/googlecloudplatform/spark-operator\",\n      \"tag\": \"\"\n
            \   },\n    \"imagePullSecrets\": [],\n    \"ingressUrlFormat\": \"\",\n
            \   \"istio\": {\n      \"enabled\": false\n    },\n    \"labelSelectorFilter\":
            \"\",\n    \"leaderElection\": {\n      \"lockName\": \"spark-operator-lock\",\n
            \     \"lockNamespace\": \"\"\n    },\n    \"logLevel\": 2,\n    \"metrics\":
            {\n      \"enable\": true,\n      \"endpoint\": \"/metrics\",\n      \"port\":
            10254,\n      \"portName\": \"metrics\",\n      \"prefix\": \"\"\n    },\n
            \   \"nameOverride\": \"\",\n    \"nodeSelector\": {},\n    \"podAnnotations\":
            {},\n    \"podLabels\": {},\n    \"podMonitor\": {\n      \"enable\":
            false,\n      \"jobLabel\": \"spark-operator-podmonitor\",\n      \"labels\":
            {},\n      \"podMetricsEndpoint\": {\n        \"interval\": \"5s\",\n
            \       \"scheme\": \"http\"\n      }\n    },\n    \"podSecurityContext\":
            {},\n    \"rbac\": {\n      \"create\": false,\n      \"createClusterRole\":
            true,\n      \"createRole\": true\n    },\n    \"replicaCount\": 1,\n
            \   \"resourceQuotaEnforcement\": {\n      \"enable\": false\n    },\n
            \   \"resources\": {},\n    \"resyncInterval\": 30,\n    \"securityContext\":
            {},\n    \"serviceAccounts\": {\n      \"spark\": {\n        \"annotations\":
            {},\n        \"create\": true,\n        \"name\": \"\"\n      },\n      \"sparkoperator\":
            {\n        \"annotations\": {},\n        \"create\": true,\n        \"name\":
            \"\"\n      }\n    },\n    \"sparkJobNamespace\": \"\",\n    \"tolerations\":
            [],\n    \"uiService\": {\n      \"enable\": true\n    },\n    \"volumeMounts\":
            [],\n    \"volumes\": [],\n    \"webhook\": {\n      \"cleanupAnnotations\":
            {\n        \"helm.sh/hook\": \"pre-delete, pre-upgrade\",\n        \"helm.sh/hook-delete-policy\":
            \"hook-succeeded\"\n      },\n      \"enable\": false,\n      \"initAnnotations\":
            {\n        \"helm.sh/hook\": \"pre-install, pre-upgrade\",\n        \"helm.sh/hook-weight\":
            \"50\"\n      },\n      \"namespaceSelector\": \"\",\n      \"port\":
            8080,\n      \"timeout\": 30\n    }\n  }\n}    "
          operatorframework.io/suggested-namespace: spark-operator
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys a scheduled spark application based on the spec provided
            displayName: Scheduled Spark Application
            kind: ScheduledSparkApplication
            name: scheduledsparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - description: Deploys a spark application based on the spec provided
            displayName: Spark Application
            kind: SparkApplication
            name: sparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - description: Deploys the spark operator based on the google cloud platform
              helm chart
            displayName: Spark Operator
            kind: Spark
            name: sparks.apachespark.apache.org
            version: v1alpha1
        description: An experimental helm based operator that wraps the spark on k8s
          helm chart from google cloud platform enabling installation and workload
          management of Apache Spark
        displayName: Spark Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - big data
        - map reduce
        - data science
        links:
        - name: Spark Helm Operator
          url: https://github.com/skattoju/spark-helm-operator
        maintainers:
        - email: skattoju@redhat.com
          name: Sid Kattoju
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: opdev.github.io
        relatedImages:
        - quay.io/opdev/spark-helm-operator-controller:0.0.2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 0.0.2
      entries:
      - name: spark-helm-operator.v0.0.2
        version: 0.0.2
      - name: spark-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: spark-helm-operator
    provider:
      name: opdev
      url: opdev.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-test-core
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-test-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Testcore",
                "metadata": {
                  "name": "testcore-sample"
                },
                "spec": {
                  "agentController": {
                    "port": 8080
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "dataseqService": {
                    "ingressEnabled": true,
                    "port": 8080,
                    "providers": {
                      "loadCore": {
                        "enabled": true
                      }
                    }
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "ingress": {
                    "backwardsCompatibilityEnabled": true
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "systemMonitor": {
                    "enabled": false
                  },
                  "trafficController": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:35:34Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Testcore
            name: testcores.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Test Core Operator
        displayName: Keysight WAP Test Core Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Test Core
          url: https://keysight-wap-test-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-test-core@sha256:9e2d81061abff49112f908e71dace03a748e283f32597aa22ad58471fff30caf
        version: 0.1.21
      entries:
      - name: keysight-wap-test-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-test-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp
      provider-url: https://www.netapp.com/
    name: netapp-spark-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netapp-spark-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.netapp.com/v1alpha1",
                "kind": "SparkOperator",
                "metadata": {
                  "name": "sparkoperator-sample"
                },
                "spec": {
                  "spark-operator": {
                    "affinity": {},
                    "controllerThreads": 20,
                    "disableExecutorReporting": false,
                    "hostNetwork": false,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "public.ecr.aws/f4k1p1n4/spark-operator",
                      "tag": "108810d-ubi9-rel20240916"
                    },
                    "imagePullSecrets": [],
                    "istio": {
                      "enabled": true
                    },
                    "jobAffinity": {},
                    "podLabels": {},
                    "pprof": {
                      "enable": true,
                      "port": 6060,
                      "portName": "pprof"
                    },
                    "rbac": {
                      "createClusterRole": true,
                      "createRole": true
                    },
                    "resources": {
                      "requests": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      }
                    },
                    "securityContext": {},
                    "serviceAccounts": {
                      "spark": {
                        "create": true
                      }
                    },
                    "sparkJobNamespace": "",
                    "tolerations": [],
                    "uiService": {
                      "enable": false
                    },
                    "webhook": {
                      "enable": true,
                      "jobTTLSecondsAfterFinished": 60,
                      "port": 9443
                    },
                    "workqueue": {
                      "maxDelay": {
                        "enabled": false,
                        "value": "6h"
                      },
                      "tokenBucketSize": 500,
                      "tokenRefillRate": 50
                    }
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "ScheduledSparkApplication",
                "metadata": {
                  "name": "spark-pi-scheduled"
                },
                "spec": {
                  "concurrencyPolicy": "Allow",
                  "schedule": "@every 5m",
                  "template": {
                    "arguments": [
                      "1000"
                    ],
                    "driver": {
                      "annotations": {
                        "openshift.io/required-scc": "sparkoperator-sample-spark"
                      },
                      "coreLimit": "1200m",
                      "cores": 1,
                      "labels": {
                        "version": "3.5.0"
                      },
                      "memory": "512m",
                      "serviceAccount": "sparkoperator-sample-spark",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "executor": {
                      "annotations": {
                        "openshift.io/required-scc": "sparkoperator-sample-spark"
                      },
                      "cores": 1,
                      "instances": 1,
                      "labels": {
                        "version": "3.5.0"
                      },
                      "memory": "512m",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "image": "public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce",
                    "imagePullPolicy": "IfNotPresent",
                    "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.13-3.5.0.jar",
                    "mainClass": "org.apache.spark.examples.SparkPi",
                    "mode": "cluster",
                    "restartPolicy": {
                      "type": "Never"
                    },
                    "sparkVersion": "3.5.0",
                    "type": "Scala",
                    "volumes": [
                      {
                        "emptyDir": {},
                        "name": "test-volume"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "spark-pi"
                },
                "spec": {
                  "arguments": [
                    "1000"
                  ],
                  "driver": {
                    "annotations": {
                      "openshift.io/required-scc": "sparkoperator-sample-spark"
                    },
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.5.0"
                    },
                    "memory": "512m",
                    "serviceAccount": "sparkoperator-sample-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "annotations": {
                      "openshift.io/required-scc": "sparkoperator-sample-spark"
                    },
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.5.0"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce",
                  "imagePullPolicy": "IfNotPresent",
                  "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.13-3.5.0.jar",
                  "mainClass": "org.apache.spark.examples.SparkPi",
                  "mode": "cluster",
                  "restartPolicy": {
                    "type": "Never"
                  },
                  "sparkVersion": "3.5.0",
                  "type": "Scala",
                  "volumes": [
                    {
                      "emptyDir": {},
                      "name": "test-volume"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Big Data, Integration & Delivery, AI/Machine Learning, Streaming
            & Messaging
          com.redhat.openshift.versions: v4.14
          containerImage: public.ecr.aws/f4k1p1n4/netapp-spark-operator-controller@sha256:9fe1d0793f11b1c525a9a92cbcde6f4bc1bf3c887617a8496bed3a5a0008a67e
          createdAt: "2024-10-21T13:47:47Z"
          description: Red Hat and NetApp have partnered to offer enterprise-grade
            supported Apache Spark on OpenShift.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ScheduledSparkApplication
            name: scheduledsparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - kind: SparkApplication
            name: sparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - kind: SparkOperator
            name: sparkoperators.charts.netapp.com
            version: v1alpha1
        description: |-
          # NetApp Supported Apache Spark on OpenShift

          Red Hat and NetApp have partnered to offer enterprise-grade supported Apache Spark on OpenShift.
          Use this certified operator to launch Apache Spark applications on OpenShift.

          ## Documentation

          Please find the official documentation at https://www.instaclustr.com/support/documentation/apache-spark/apache-spark-on-red-hat-openshift/getting-started/

          ## Support

          With questions about Apache Spark, contact NetApp's Instaclustr support team on the [web](https://support.instaclustr.com/) or by email at support@instaclustr.com

          With questions about Operators in the OpenShift OperatorHub, contact Red Hat's support team on the [web](https://www.redhat.com/en/services/support) or by email at customerservice@redhat.com

          With questions about pricing, email ng-netapp-apache-spark-support-redhat-openshift@netapp.com
        displayName: NetApp Supported Apache Spark on OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Spark
        - Big Data
        - Apache Spark
        - AI
        - ML
        - Data Engineering
        links:
        - name: NetApp Supported Apache Spark on OpenShift
          url: https://www.instaclustr.com/support-solutions
        maintainers:
        - email: thorsteinn.sigurdsson@netapp.com
          name: Thorsteinn Sigurdsson
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: NetApp
          url: https://www.netapp.com/
        relatedImages:
        - public.ecr.aws/f4k1p1n4/netapp-spark-operator-controller@sha256:9fe1d0793f11b1c525a9a92cbcde6f4bc1bf3c887617a8496bed3a5a0008a67e
        - public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce
        version: 1.1.3
      entries:
      - name: netapp-spark-operator.v1.1.3
        version: 1.1.3
      name: alpha
    defaultChannel: alpha
    packageName: netapp-spark-operator
    provider:
      name: NetApp
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Trilio
      provider-url: ""
    name: k8s-triliovault
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k8s-triliovault-stable.3.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/trilio/k8s-triliovault-operator@sha256:0e3094945427231ecbb884b903dc4d347dfaa73e5a04eb4c5cbaed6fce1cbbb8
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          olm.skipRange: '>=3.0.0 <3.1.3'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          TrilioVault for Kubernetes is an enterprise-grade, cloud-native platform purpose built for data protection and management of Kubernetes applications for IT managers, administrators and developers. TrilioVault supports upstream Kubernetes and OpenShift environments and offers the following features:
            * Application-Centric - protects both data and metadata for Helm, Operator or custom Label based applications.
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any Storage (CSI, NFS, S3), or any Cloud (Private or Public).
            * Application Deployment & Tooling: Helm, Operators, Labels, Prometheus, Fluentd.
          Tutorials
          ------------
          Please click the link below to access the TrilioVault for Kubernetes "How-To" series for deployment, best practice and use-case videos
          <a href="https://www.trilio.io/triliovault-for-kubernetes">TrilioVault for Kubernetes - Tutorials and How-To </a>”
          Licensing
          --------
          Customers can download a 30-day Free Trial or a 10-node Basic Edition at no cost.  You can also connect with the Trilio team for an Enterprise Edition license with no limitations and Premium Support.
          For more information on license plans please vist:
          <a href="www.trilio.io/plans"> Trilio Vault for Kubernetes licensing and plans </a>
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our TrilioVault technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. www.trilio.io and @triliodata on Twitter.
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://www.trilio.io/
        - name: Product Licensing
          url: https://www.trilio.io/plans
        - name: Tutorials
          url: https://www.trilio.io/triliovault-for-kubernetes/
        - name: Support for Free/Basic Users
          url: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Trilio
        relatedImages:
        - registry.connect.redhat.com/trilio/k8s-triliovault-operator@sha256:0e3094945427231ecbb884b903dc4d347dfaa73e5a04eb4c5cbaed6fce1cbbb8
        - registry.connect.redhat.com/trilio/trilio-webhook-init@sha256:c45e9b6b75862451d873d3eba29d6bd11cafc64823855f82ccb8c0b3d779ebb0
        - registry.connect.redhat.com/trilio/web@sha256:76865d8afa8059d40810cb94df29420950a7593a40e87d142d771c57b3319e1d
        - registry.connect.redhat.com/trilio/minio@sha256:cee0f976edc79ff9e7a3f3574a7115dd960042241f3ed5b4636240ad609f9bca
        - registry.connect.redhat.com/trilio/control-plane@sha256:01e36ed5b07104865f922d2996d2211b2b6000adc90266197d9bac74d908db97
        - registry.connect.redhat.com/trilio/datamover@sha256:c7bc15f8217899ed73a812d278eb118633edbee1bf284132054b6f9ca2afd801
        - registry.connect.redhat.com/trilio/dex@sha256:b6331cf3cb3a6a66d9b9ea639cda3c2559fbf1f92f7174032352c1c7bb5176d9
        - registry.connect.redhat.com/trilio/nats@sha256:47396bfcff3e65efb3410fa8369f11f20d794ab543f9ada87b451448b1813178
        - registry.connect.redhat.com/trilio/event-stack@sha256:37c33ae1adddfe739c98fd2b2183c49ec7cfdafc88dec2970ad7c670016ebaac
        version: 3.1.3
      entries:
      - name: k8s-triliovault-stable.3.1.3
        version: 3.1.3
      - name: k8s-triliovault-stable.3.1.2
        version: 3.1.2
      - name: k8s-triliovault-stable.3.1.1
        version: 3.1.1
      - name: k8s-triliovault-stable.3.1.0
        version: 3.1.0
      name: 3.1.x
    - currentCSV: k8s-triliovault-stable.4.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:69871613ceecefff7f691f21bb0b9147236af5a5dc2c7b36ad5e0cd8551f0e6a
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <4.0.5'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/event-stack@sha256:c793418df5a1fc1a49f629d335f4684d9cde1a5f2b462f7c9483ffba904456d2
        - quay.io/triliodata/control-plane@sha256:53368a72a60a38d3ce4293236a841346da6da2380ffd2525e5ffdb8dbcf11c7f
        - quay.io/triliodata/web@sha256:20de1a2f9477ad238bf24f48c8c16a38521c3a86164f2eb6b10d20f09839016e
        - quay.io/triliodata/datamover@sha256:e6848d94ba9f81843bde882a030b0f2556aae4bf16d79e2c5b199417cf626c85
        - quay.io/triliodata/k8s-triliovault-operator@sha256:69871613ceecefff7f691f21bb0b9147236af5a5dc2c7b36ad5e0cd8551f0e6a
        - quay.io/triliodata/operator-webhook-init@sha256:aa85b787f7f37a71e909f049d2abdda746cdd0627e2f74aaf7476d5e5a7f4401
        - quay.io/triliodata/dex@sha256:b4b2372fd03be586b9ff8cd269b9001797d813a34155ab37de31a44fe3cab1ea
        - quay.io/triliodata/nats@sha256:e6723595bcc5674b6201091d45d711d29a5cd1421c8b97c76cf00e468379fddf
        version: 4.0.5
      entries:
      - name: k8s-triliovault-stable.4.0.5
        version: 4.0.5
      - name: k8s-triliovault-stable.4.0.4
        version: 4.0.4
      - name: k8s-triliovault-stable.4.0.3
        version: 4.0.3
      - name: k8s-triliovault-stable.4.0.2
        version: 4.0.2
      - name: k8s-triliovault-stable.4.0.1
        version: 4.0.1
      - name: k8s-triliovault-stable.4.0.0
        version: 4.0.0
      name: 4.0.x
    - currentCSV: k8s-triliovault-stable.5.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:8d582842dbef1c64bbe3b0ce5eeb64790910bdc0416c5e5561557642dede7022
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <5.0.4'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/web@sha256:04c4baadae9c05f4add098ff62774d00102d56581e261626840d17f74effe438
        - quay.io/triliodata/datamover@sha256:104bdca9513c78f81c4340e9a3219df0a3bc8a5b026e4a252363a7e2c10977ba
        - quay.io/triliodata/dex@sha256:b4b2372fd03be586b9ff8cd269b9001797d813a34155ab37de31a44fe3cab1ea
        - quay.io/triliodata/nats@sha256:e6723595bcc5674b6201091d45d711d29a5cd1421c8b97c76cf00e468379fddf
        - quay.io/triliodata/event-stack@sha256:859535e19e0ebebc8261c649c2446f8ca5fe4b4e5b6236e69b780dcc0d0bb5b9
        - quay.io/triliodata/k8s-triliovault-operator@sha256:8d582842dbef1c64bbe3b0ce5eeb64790910bdc0416c5e5561557642dede7022
        - quay.io/triliodata/operator-webhook-init@sha256:ffcc55e85f884b884947b73fa7d2fb257704ff70ccd7f2dc37d07e02292abed6
        - quay.io/triliodata/control-plane@sha256:e60ef0a2c7a5b7879c9718e9a14c223f5290b4710b598fae19466a7558484824
        version: 5.0.4
      entries:
      - name: k8s-triliovault-stable.5.0.4
        version: 5.0.4
      - name: k8s-triliovault-stable.5.0.3
        version: 5.0.3
      - name: k8s-triliovault-stable.5.0.2
        version: 5.0.2
      - name: k8s-triliovault-stable.5.0.1
        version: 5.0.1
      - name: k8s-triliovault-stable.5.0.0
        version: 5.0.0
      name: 5.0.x
    - currentCSV: k8s-triliovault-stable.5.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "2Gi",
                      "cpu": "1"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:f0cf3f586561414671668210c59ac44071fde6abc954917e6ca04df5777a9a2f
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <5.1.1'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/k8s-triliovault-operator@sha256:f0cf3f586561414671668210c59ac44071fde6abc954917e6ca04df5777a9a2f
        - quay.io/triliodata/event-stack@sha256:548ae1a748919f518e811774ec87f128aa8857be234837e09308c295de165726
        - quay.io/triliodata/control-plane@sha256:2e7e44cf6bc70dd9b978dce2cb2d9a36b8886c0f55286d5a6f5eb0a5182c4350
        - quay.io/triliodata/datamover@sha256:8a1350c8975610f23dd0b40d604d218f59dca79db6beb81ffe78a2682e40864f
        - quay.io/triliodata/dex@sha256:6b9d1667916968fb1f187ad5945fcc08db702dcab4c947e309dd9d2881a7b9cf
        - quay.io/triliodata/nats@sha256:72283871b6b9343aeabc3ff090305979422f192d30a6cbc4b0f20b2de7533996
        - quay.io/triliodata/filerecovery-vm-ssh-server@sha256:09186be228bd0b59e83119b28f3239c6173e410a3cf36957fe2d9b56d25200f8
        - quay.io/triliodata/web@sha256:3908a39eed44d702c6b6804d392d08a570b3f5562590d050f17ddc7a5c87b37f
        - quay.io/triliodata/filerecovery@sha256:42559ac5d949d5ecf819cd4768a71fa688c72230502ff0b56539a8bb5e2f5838
        version: 5.1.1
      entries:
      - name: k8s-triliovault-stable.5.1.1
        version: 5.1.1
      - name: k8s-triliovault-stable.5.1.0
        version: 5.1.0
      name: 5.1.x
    defaultChannel: 5.1.x
    packageName: k8s-triliovault
    provider:
      name: Trilio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Couchbase
      provider-url: ""
    name: couchbase-enterprise-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: couchbase-operator.v2.6.4-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator@sha256:1f0a6865c360a14ba98706fd6fe8dff3bdbdede32e3622f93f19c82e377ca33f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        version: 2.6.4-7
      entries:
      - name: couchbase-operator.v2.6.4-7
        version: 2.6.4-7
      - name: couchbase-operator.v2.6.4-6
        version: 2.6.4-6
      name: 2.6.4
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:707571a7881bc12d5f1b0b1a94eb3c675d5c0ea0114c994a47ceb5aa79df4013
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        version: 2.7.0-5
      entries:
      - name: couchbase-operator.v2.7.0-5
        version: 2.7.0-5
      name: 2.7.0
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:a79e1b6781d5f8badf664e7efea6e2889e43fae6e0d37169d766685dcd6ff394
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        version: 2.7.1-1
      entries:
      - name: couchbase-operator.v2.7.1-1
        version: 2.7.1-1
      name: 2.7.1
    - currentCSV: couchbase-operator.v2.8.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator@sha256:df61928b064014d851c6e7dff6dcd63a433a27fa5623254b6e389c3eea170e36
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        version: 2.8.0-2
      entries:
      - name: couchbase-operator.v2.8.0-2
        version: 2.8.0-2
      - name: couchbase-operator.v2.8.0-1
        version: 2.8.0-1
      name: 2.8.0
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        version: 2.8.1-2
      entries:
      - name: couchbase-operator.v2.8.1-2
        version: 2.8.1-2
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: 2.8.1
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        version: 2.8.1-2
      entries:
      - name: couchbase-operator.v2.8.1-2
        version: 2.8.1-2
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: stable
    defaultChannel: stable
    packageName: couchbase-enterprise-certified
    provider:
      name: Couchbase
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: aws-efs-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-efs-csi-driver-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-aws-efs-csi-driver-rhel9-operator@sha256:35ef24108df5f24d24fefe915f501965c98edea934fa6c2895e375ed006e5e7b
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure AWS EFS CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0-0 <4.16.0-202509240123'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/aws-efs-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          AWS EFS CSI Driver Operator provides AWS Elastic File Service (EFS) CSI Driver that enables you to create and mount AWS EFS PersistentVolumes.

          Please follow [AWS EFS CSI Driver Operator documentation](https://docs.openshift.com/container-platform/4.16/storage/container_storage_interface/persistent-storage-csi-aws-efs.html) to:

          1. [Install the operator](https://docs.openshift.com/container-platform/4.16/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#persistent-storage-csi-olm-operator-install_persistent-storage-csi-aws-efs). This section also covers details about configuring an AWS role in OCP cluster that uses Secure Token Services (STS).

          2. [Install the AWS EFS CSI driver using the operator](https://docs.openshift.com/container-platform/4.16/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#persistent-storage-csi-efs-driver-install_persistent-storage-csi-aws-efs)

          3. Finally, [create a StorageClass](https://docs.openshift.com/container-platform/4.16/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#storage-create-storage-class_persistent-storage-csi-aws-efs) that enables dynamic provisioning of PersistentVolumes.
        displayName: AWS EFS CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - efs
        links:
        - name: Documentation
          url: https://github.com/openshift/aws-efs-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/aws-efs-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-aws-efs-csi-driver-rhel9-operator@sha256:35ef24108df5f24d24fefe915f501965c98edea934fa6c2895e375ed006e5e7b
        - registry.redhat.io/openshift4/ose-aws-efs-csi-driver-container-rhel9@sha256:c857dda8f399fa104bd4f392a354b74365657acea6b8fc5b023fdb4bf55057c2
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:9c3c3d2b857ce289324aee56df1ba4ab5e5ea1e76cd86ac988233463915834b4
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0627c09fe3f5a015f117fb30af1f4d17340aefa48b23994737d3901a804b63d3
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:a5704d26d0b562dba9633d1c107f34e05c750c0e3a64dbcace4dab7b964fff8d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        version: 4.16.0-202509240123
      entries:
      - name: aws-efs-csi-driver-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: aws-efs-csi-driver-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: aws-efs-csi-driver-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: aws-efs-csi-driver-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: aws-efs-csi-driver-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: aws-efs-csi-driver-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: aws-efs-csi-driver-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: aws-efs-csi-driver-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: aws-efs-csi-driver-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: aws-efs-csi-driver-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: aws-efs-csi-driver-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: aws-efs-csi-driver-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: aws-efs-csi-driver-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: aws-efs-csi-driver-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: aws-efs-csi-driver-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: aws-efs-csi-driver-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: aws-efs-csi-driver-operator.v4.16.0-202410280205
        version: 4.16.0-202410280205
      - name: aws-efs-csi-driver-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: aws-efs-csi-driver-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: aws-efs-csi-driver-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: aws-efs-csi-driver-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: aws-efs-csi-driver-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: aws-efs-csi-driver-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: aws-efs-csi-driver-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: aws-efs-csi-driver-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: aws-efs-csi-driver-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: aws-efs-csi-driver-operator.v4.16.0-202408061508
        version: 4.16.0-202408061508
      - name: aws-efs-csi-driver-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: aws-efs-csi-driver-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: aws-efs-csi-driver-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: aws-efs-csi-driver-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: aws-efs-csi-driver-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: aws-efs-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: self-node-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: self-node-remediation.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:cb03286b9d9cda4a47e5b1c0058ca315bf082bcc8697506ea88d97ba77452bea
          createdAt: "2023-10-20 16:38:06"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          olm.skipRange: '>=0.4.0 <0.5.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the heath detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-self-node-remediation-operator.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:cb03286b9d9cda4a47e5b1c0058ca315bf082bcc8697506ea88d97ba77452bea
        version: 0.5.1
      entries:
      - name: self-node-remediation.v0.5.1
        version: 0.5.1
      name: 4.12-eus
    - currentCSV: self-node-remediation.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:e2048a8a2ec37e10b57dafeba2e131311e61982242cda4beb989e5e30c1a5537
          createdAt: "2023-10-20 15:32:20"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          olm.skipRange: '>=0.5.0 <0.7.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:e2048a8a2ec37e10b57dafeba2e131311e61982242cda4beb989e5e30c1a5537
        version: 0.7.1
      entries:
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.14-eus
    - currentCSV: self-node-remediation.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:6bd86a47d69f7662e24a93ce205a8179018152a97696a9223d41ed1887c35257
          createdAt: "2025-07-02 16:40:20"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.9.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:6bd86a47d69f7662e24a93ce205a8179018152a97696a9223d41ed1887c35257
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ef91935704bb9f0af4cdad07333e802587e3d458b2cd911e04d557b1defe3412
        version: 0.9.1
      entries:
      - name: self-node-remediation.v0.9.1
        version: 0.9.1
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.16-eus
    - currentCSV: self-node-remediation.v0.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
          createdAt: "2025-07-02 18:34:25"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        version: 0.10.2
      entries:
      - name: self-node-remediation.v0.10.2
        version: 0.10.2
      - name: self-node-remediation.v0.10.1
        version: 0.10.1
      - name: self-node-remediation.v0.10.0
        version: 0.10.0
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.18-eus
    - currentCSV: self-node-remediation.v0.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
          createdAt: "2025-07-02 18:34:25"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        version: 0.10.2
      entries:
      - name: self-node-remediation.v0.10.2
        version: 0.10.2
      - name: self-node-remediation.v0.10.1
        version: 0.10.1
      - name: self-node-remediation.v0.10.0
        version: 0.10.0
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: stable
    defaultChannel: stable
    packageName: self-node-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC Corporation
      provider-url: https://jpn.nec.com/webotx/
    name: webotx-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: webotx-operator.v11.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "webotx.nec.com/v3",
                "kind": "ApplicationServer",
                "metadata": {
                  "name": "webotx-express"
                },
                "spec": {
                  "deployment": {
                    "image": "registry.connect.redhat.com/nec/webotx-express:11.11.00.00@sha256:73a5dc48b4e3340951dbe207d744f5ac68474701dd134fa7aa05e65e3510c8ad",
                    "replicas": 1
                  },
                  "service": {
                    "externalTrafficPolicy": "Cluster",
                    "ports": [
                      {
                        "name": "http",
                        "nodePort": 30080,
                        "port": 8080,
                        "protocol": "TCP",
                        "targetPort": 8080
                      },
                      {
                        "name": "https",
                        "nodePort": 30443,
                        "port": 8443,
                        "protocol": "TCP",
                        "targetPort": 8443
                      }
                    ],
                    "sessionAffinity": "ClientIP",
                    "sessionAffinityConfig": {
                      "clientIP": {
                        "timeoutSeconds": 10800
                      }
                    },
                    "type": "NodePort"
                  },
                  "webotx": {
                    "metering": {
                      "enabled": false,
                      "port": 8443,
                      "scheme": "https"
                    },
                    "sessionReplication": {
                      "enabled": false,
                      "useAllProcessGroups": false,
                      "useJmxAgent": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApplicationServer is the Schema for the applicationservers
              API
            displayName: Application Server
            kind: ApplicationServer
            name: applicationservers.webotx.nec.com
            version: v3
        description: WebOTX Operator is the Operator for WebOTX Application Server.
          WebOTX Operator enables to deploy and administrate WebOTX Application Server
          automatically.
        displayName: WebOTX Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - webotx
        - WebOTX
        - nec
        - NEC
        links:
        - name: WebOTX Operator for Kubernetes
          url: https://github.com/webotx/operator
        maintainers:
        - email: info-webotx@isd.jp.nec.com
          name: NEC Corporation
        maturity: stable
        provider:
          name: NEC Corporation
          url: https://jpn.nec.com/webotx/
        relatedImages:
        - registry.connect.redhat.com/nec/webotx-operator:11.11.00.00@sha256:df0d0315ed88be875ca7a6474500f43f17855f763c5c7d0f4d3ec58a2e46f6fb
        version: 11.11.0
      entries:
      - name: webotx-operator.v11.11.0
        version: 11.11.0
      name: stable
    defaultChannel: stable
    packageName: webotx-operator
    provider:
      name: NEC Corporation
      url: https://jpn.nec.com/webotx/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator-certified-rhmp.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a",
                  "imagePullPolicy": "Always",
                  "initImage": "registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-server-mongodb-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-server-mongodb-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:3c03fb18a5d73876576b80aca2b61e7c84ca052839ccb467d14e15f6d1f9d834
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a57926b53eef862b70a53818a7ccb0e9756f878f30d4128b608090b109bdc585
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator-certified-rhmp.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator-certified-rhmp.v1.20.0
        version: 1.20.0
      - name: percona-server-mongodb-operator-certified-rhmp.v1.19.1
        version: 1.19.1
      - name: percona-server-mongodb-operator-certified-rhmp.v1.19.0
        version: 1.19.0
      - name: percona-server-mongodb-operator-certified-rhmp.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator-certified-rhmp
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kms-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kms-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Alias",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Key",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kms-controller:1.1.1
          createdAt: "2025-09-19T22:02:02Z"
          description: AWS KMS controller is a service controller for managing KMS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alias represents the state of an AWS kms Alias resource.
            displayName: Alias
            kind: Alias
            name: aliases.kms.services.k8s.aws
            version: v1alpha1
          - description: Grant represents the state of an AWS kms Grant resource.
            displayName: Grant
            kind: Grant
            name: grants.kms.services.k8s.aws
            version: v1alpha1
          - description: Key represents the state of an AWS kms Key resource.
            displayName: Key
            kind: Key
            name: keys.kms.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Key Management Service (KMS) resources in AWS from within your Kubernetes cluster.

          **About Amazon KMS**

          AWS Key Management Service (AWS KMS) is a managed service that makes it easy for you to create and control the cryptographic keys that are used to protect your data. AWS KMS uses hardware security modules (HSM) to protect and validate your AWS KMS keys under the [FIPS 140-2 Cryptographic Module Validation Program](https://csrc.nist.gov/projects/cryptographic-module-validation-program/Certificate/3139), except in the China (Beijing) and China (Ningxia) Regions.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon KMS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kms
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon KMS Developer Resources
          url: https://aws.amazon.com/kms/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kms maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/kms-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-kms-controller.v1.1.1
        version: 1.1.1
      - name: ack-kms-controller.v1.1.0
        version: 1.1.0
      - name: ack-kms-controller.v1.0.31
        version: 1.0.31
      - name: ack-kms-controller.v1.0.30
        version: 1.0.30
      - name: ack-kms-controller.v1.0.29
        version: 1.0.29
      - name: ack-kms-controller.v1.0.28
        version: 1.0.28
      - name: ack-kms-controller.v1.0.27
        version: 1.0.27
      - name: ack-kms-controller.v1.0.26
        version: 1.0.26
      - name: ack-kms-controller.v1.0.25
        version: 1.0.25
      - name: ack-kms-controller.v1.0.24
        version: 1.0.24
      - name: ack-kms-controller.v1.0.23
        version: 1.0.23
      - name: ack-kms-controller.v1.0.22
        version: 1.0.22
      - name: ack-kms-controller.v1.0.19
        version: 1.0.19
      - name: ack-kms-controller.v1.0.18
        version: 1.0.18
      - name: ack-kms-controller.v1.0.17
        version: 1.0.17
      - name: ack-kms-controller.v1.0.16
        version: 1.0.16
      - name: ack-kms-controller.v1.0.15
        version: 1.0.15
      - name: ack-kms-controller.v1.0.14
        version: 1.0.14
      - name: ack-kms-controller.v1.0.13
        version: 1.0.13
      - name: ack-kms-controller.v1.0.12
        version: 1.0.12
      - name: ack-kms-controller.v1.0.8
        version: 1.0.8
      - name: ack-kms-controller.v1.0.7
        version: 1.0.7
      - name: ack-kms-controller.v1.0.5
        version: 1.0.5
      - name: ack-kms-controller.v1.0.4
        version: 1.0.4
      - name: ack-kms-controller.v1.0.3
        version: 1.0.3
      - name: ack-kms-controller.v1.0.2
        version: 1.0.2
      - name: ack-kms-controller.v1.0.0
        version: 1.0.0
      - name: ack-kms-controller.v0.1.5
        version: 0.1.5
      - name: ack-kms-controller.v0.1.4
        version: 0.1.4
      - name: ack-kms-controller.v0.1.3
        version: 0.1.3
      - name: ack-kms-controller.v0.1.2
        version: 0.1.2
      - name: ack-kms-controller.v0.1.1
        version: 0.1.1
      - name: ack-kms-controller.v0.1.0
        version: 0.1.0
      - name: ack-kms-controller.v0.0.20
        version: 0.0.20
      - name: ack-kms-controller.v0.0.19
        version: 0.0.19
      - name: ack-kms-controller.v0.0.18
        version: 0.0.18
      - name: ack-kms-controller.v0.0.17
        version: 0.0.17
      - name: ack-kms-controller.v0.0.16
        version: 0.0.16
      - name: ack-kms-controller.v0.0.15
        version: 0.0.15
      - name: ack-kms-controller.v0.0.14
        version: 0.0.14
      - name: ack-kms-controller.v0.0.13
        version: 0.0.13
      - name: ack-kms-controller.v0.0.12
        version: 0.0.12
      - name: ack-kms-controller.v0.0.11
        version: 0.0.11
      - name: ack-kms-controller.v0.0.10
        version: 0.0.10
      - name: ack-kms-controller.v0.0.9
        version: 0.0.9
      - name: ack-kms-controller.v0.0.8
        version: 0.0.8
      - name: ack-kms-controller.v0.0.7
        version: 0.0.7
      - name: ack-kms-controller.v0.0.6
        version: 0.0.6
      - name: ack-kms-controller.v0.0.5
        version: 0.0.5
      - name: ack-kms-controller.v0.0.4
        version: 0.0.4
      - name: ack-kms-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-kms-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: proactive-node-scaling-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: proactive-node-scaling-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NodeScalingWatermark",
                "metadata": {
                  "name": "ai-ml"
                },
                "spec": {
                  "nodeSelector": {
                    "machine.openshift.io/cluster-api-machine-type": "ai-ml"
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "workload",
                      "operator": "Equal",
                      "value": "ai-ml"
                    }
                  ],
                  "watermarkPercentage": 20
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NodeScalingWatermark",
                "metadata": {
                  "name": "us-west-2a"
                },
                "spec": {
                  "nodeSelector": {
                    "topology.kubernetes.io/zone": "us-west-2a"
                  },
                  "watermarkPercentage": 20
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "false"
          containerImage: quay.io/redhat-cop/proactive-node-scaling-operator@sha256:892303babb45090a6458cd4b1c61a01364012c43ea0bd6ed4d10bdcb8041e2ab
          createdAt: "2022-01-22T14:14:23Z"
          description: This operator provides automation proactively scaling nodes
            before pods get stuck waiting for the cluster autoscaler.
          operatorframework.io/suggested-namespace: proactive-node-scaling-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/proactive-node-scaling-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeScalingWatermark is the Schema for the nodescalingwatermarks
              API
            displayName: Node Scaling Watermark
            kind: NodeScalingWatermark
            name: nodescalingwatermarks.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator makes the [cluster autoscaler](https://github.com/kubernetes/autoscaler/tree/master/cluster-autoscaler) more proactive. As of now the cluster auto scaler will create new nodes only when a pod is pending because it cannot be allocated due to lack of capacity. This is not a goos user experience as the pending workload has to wait for several minutes as the new node is create and joins the cluster.

          The Proactive Node Scaling Operator improves the user experience by allocating low priority pods that don't do anything. When the cluster is full and a new user pod is created the following happens:

          1. some of the low priority pods are de-scheduled to make room for the user pod, which can then be scheduled. The user workload does not have to wait in this case.

          2. the de-scheduled low priority pods are rescheduled and in doing so the trigger the cluster autoscaler to add new nodes.

          Essentially this operator allows you to trade wasted resources for faster response time.

          In order for this operator to work correctly [pod priorities](https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/) must be defined. The default name for the priority class used by this operator is "proactive-node-autoscaling-pods" and it should have the lowest possible priority, 0. To ensure your regular workloads get a normal priority you should also define a PriorityClass for those with a higher priority than 0 and set globalDefault to true.

          For example:

          ```yaml
          apiVersion: scheduling.k8s.io/v1
          kind: PriorityClass
          metadata:
            name: proactive-node-autoscaling-pods
          value: 0
          globalDefault: false
          description: "This priority class is the priority class used for Proactive Node Scaling Pods."
          ---
          apiVersion: scheduling.k8s.io/v1
          kind: PriorityClass
          metadata:
            name: normal-workload
          value: 1000
          globalDefault: true
          description: "This priority classis the cluster default and should be used for normal workloads."
          ```

          Also for this operator to work the [cluster autoscaler](https://github.com/kubernetes/autoscaler/tree/master/cluster-autoscaler) must be active, see OpenShift instructions [here](https://docs.openshift.com/container-platform/4.6/machine_management/applying-autoscaling.html) on how to turn it on.

          To activate the proactive autoscaling, a CR must be defined, here is an example:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: NodeScalingWatermark
          metadata:
            name: us-west-2a
          spec:
            priorityClassName: proactive-node-autoscaling-pods
            watermarkPercentage: 20
            nodeSelector:
              topology.kubernetes.io/zone: us-west-2a
          ```

          The `nodeSelector` selects the nodes observed by this operator, which are also the nodes on which the low priority pods will be scheduled. The nodes observed by the cluster autoscaler should coincide with the nodes selected by this operator CR.

          The `watermarkPercentage` define the percentage of capacity of user workload that will be allocated to low priority pods. So in this example 20% of the user allocated capacity will be allocated via low priority pods. This also means that when the user workload reaches 80% capacity of the nodes selected by this CR (and the autoscaler), the cluster will start to scale.
        displayName: Proactive Node Scaling Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Capacity Management
        - Node Autoscaling
        - Proactive Autoscaling
        - Elastic Capacity
        links:
        - name: Proactive Node Scaling Operator
          url: https://github.com/redhat-cop/proactive-node-scaling-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/proactive-node-scaling-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/proactive-node-scaling-operator@sha256:892303babb45090a6458cd4b1c61a01364012c43ea0bd6ed4d10bdcb8041e2ab
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 0.4.2
      entries:
      - name: proactive-node-scaling-operator.v0.4.2
        version: 0.4.2
      - name: proactive-node-scaling-operator.v0.4.0
        version: 0.4.0
      - name: proactive-node-scaling-operator.v0.3.10
        version: 0.3.10
      - name: proactive-node-scaling-operator.v0.3.9
        version: 0.3.9
      - name: proactive-node-scaling-operator.v0.3.8
        version: 0.3.8
      - name: proactive-node-scaling-operator.v0.3.7
        version: 0.3.7
      - name: proactive-node-scaling-operator.v0.3.6
        version: 0.3.6
      - name: proactive-node-scaling-operator.v0.3.5
        version: 0.3.5
      - name: proactive-node-scaling-operator.v0.3.4
        version: 0.3.4
      - name: proactive-node-scaling-operator.v0.3.3
        version: 0.3.3
      - name: proactive-node-scaling-operator.v0.3.2
        version: 0.3.2
      - name: proactive-node-scaling-operator.v0.3.1
        version: 0.3.1
      - name: proactive-node-scaling-operator.v0.3.0
        version: 0.3.0
      - name: proactive-node-scaling-operator.v0.2.0
        version: 0.2.0
      - name: proactive-node-scaling-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: proactive-node-scaling-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Debezium Authors
      provider-url: https://debezium.io/
    name: debezium-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: debezium-operator.v2.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.4.2.Final
          createdAt: "2023-12-13T12:20:56Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.4.2.Final
        version: 2.4.2
      entries:
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.4.x
    - currentCSV: debezium-operator.v2.5.3-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.5.3.Final
          createdAt: "2024-03-19T08:17:17Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.5.3.Final
        version: 2.5.3-final
      entries:
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.5.x
    - currentCSV: debezium-operator.v2.6.1-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.6.1.Final
          createdAt: "2024-04-12T17:18:16Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.5.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.6.1.Final
        version: 2.6.1-final
      entries:
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.6.x
    - currentCSV: debezium-operator.v2.7.3-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.7.3.Final
          createdAt: "2024-09-20T10:31:16Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.7.3.Final
        version: 2.7.3-final
      entries:
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.7.x
    - currentCSV: debezium-operator.v3.0.0-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:3.0.0.Final
          createdAt: "2024-10-02T09:55:38Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.7.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:3.0.0.Final
        version: 3.0.0-final
      entries:
      - name: debezium-operator.v3.0.0-final
        version: 3.0.0-final
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-3.0.x
    - currentCSV: debezium-operator.v3.0.0-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:3.0.0.Final
          createdAt: "2024-10-02T09:55:38Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.7.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:3.0.0.Final
        version: 3.0.0-final
      entries:
      - name: debezium-operator.v3.0.0-final
        version: 3.0.0-final
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-latest
    defaultChannel: debezium-latest
    packageName: debezium-operator
    provider:
      name: Debezium Authors
      url: https://debezium.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-notifications
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-notifications.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Notifications",
                "metadata": {
                  "name": "notifications-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configuration": {
                    "port": 8080
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp",
                    "test": "wap-rhel8-toolbox:8.7.0-3747-36-ocp",
                    "testRepository": "prod.harbor.keysight.digital/loadcore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:34:47Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Notifications
            name: notifications.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Notifications Operator
        displayName: Keysight WAP Notifications Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Notifications
          url: https://keysight-wap-notifications.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-notifications@sha256:50c24180b3c16cf3eb82f6f4eeb61b91a62a198c122f7e59aa680a25bc488c2d
        version: 0.1.21
      entries:
      - name: keysight-wap-notifications.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-notifications
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-stats-dashboards
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-stats-dashboards.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Statsdashboard",
                "metadata": {
                  "name": "statsdashboard-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "statsDashboardService": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:41:08Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Statsdashboard
            name: statsdashboards.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Stats Dashboard Operator
        displayName: Keysight WAP Stats Dashboard Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Stats Dashboards
          url: https://keysight-wap-stats-dashboards.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-stats-dashboards@sha256:bcdc68bf0b9e9c22d775d704320829d258a9ebc77dfe41e2414061740f87f236
        version: 0.1.21
      entries:
      - name: keysight-wap-stats-dashboards.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-stats-dashboards
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Saison Technology Co., Ltd.
      provider-url: https://www.hulft.com/
    name: hulft
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hulft.v10.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hulft.saison-technology.com/v1",
                "kind": "Hulft",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "hulft",
                    "app.kubernetes.io/instance": "sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hulft",
                    "app.kubernetes.io/part-of": "hulft"
                  },
                  "name": "sample",
                  "namespace": "user-apps"
                },
                "spec": {
                  "applicationName": "hulft-apps",
                  "control": {
                    "allowExternalAccess": true,
                    "replicas": {
                      "max": 1,
                      "min": 1
                    }
                  },
                  "db": {
                    "authorizationSecret": "secret-hulft-db-authorization",
                    "connectionRetryCount": 30,
                    "hostname": "database.example.com",
                    "name": "hulft",
                    "port": 3306,
                    "tablePrefix": "hulft"
                  },
                  "licenseSecret": "secret-hulft-license",
                  "resourceIdentifier": "sample",
                  "transfer": {
                    "control": {
                      "retryCount": 10,
                      "retryInterval": 10000,
                      "timeout": 60000
                    },
                    "gracefulShutdownWait": 3600,
                    "observe": {
                      "websocket": {
                        "allowExternalAccess": false
                      }
                    },
                    "replicas": {
                      "max": 5,
                      "min": 1
                    },
                    "s3": {
                      "defaultRegion": "ap-northeast-1"
                    },
                    "websocket": {
                      "allowExternalAccess": false
                    }
                  }
                },
                "status": {
                  "applicationName": "hulft-apps",
                  "control": {
                    "allowExternalAccess": true,
                    "apiClientTlsSecret": "secret-hulft-sample-transfer-control-server-tls",
                    "consoleLoginSecret": "secret-hulft-sample-console-login",
                    "consoleTlsSecret": "secret-hulft-sample-control-console-tls",
                    "hostname": "",
                    "httpPort": 30080,
                    "httpsPort": 30443,
                    "replicas": {
                      "max": 1,
                      "min": 1
                    },
                    "samlSecret": ""
                  },
                  "db": {
                    "authorizationSecret": "secret-hulft-db-authorization",
                    "connectionRetryCount": 30,
                    "encryptionKeySecret": "secret-hulft-sample-db-encryption-key",
                    "hostname": "database.example.com",
                    "name": "hulft",
                    "port": 3306,
                    "tablePrefix": "hulft"
                  },
                  "licenseSecret": "secret-hulft-license",
                  "resourceIdentifier": "sample",
                  "transfer": {
                    "control": {
                      "clientTlsSecret": "secret-hulft-sample-control-console-tls",
                      "port": 30443,
                      "retryCount": 10,
                      "retryInterval": 10000,
                      "serverTlsSecret": "secret-hulft-sample-transfer-control-server-tls",
                      "timeout": 60000
                    },
                    "gracefulShutdownWait": 3600,
                    "hostname": "hulft-sample-transfer-wss-ci.apps.openshift.example.com",
                    "hulft": {
                      "port": 30000
                    },
                    "observe": {
                      "port": 31000,
                      "websocket": {
                        "allowExternalAccess": true,
                        "port": 33443,
                        "serverTlsSecret": "secret-hulft-sample-transfer-observe-wss-server-tls"
                      }
                    },
                    "replicas": {
                      "max": 5,
                      "min": 1
                    },
                    "s3": {
                      "authenticationSecret": "",
                      "defaultRegion": "ap-northeast-1"
                    },
                    "websocket": {
                      "allowExternalAccess": true,
                      "clientTlsSecret": "",
                      "port": 31443,
                      "serverTlsSecret": "secret-hulft-sample-transfer-wss-server-tls"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/saison-technology/hulft-operator@sha256:8ee0a29d0ce96f4ec230f387467b241dccbb2bd350cd642b275ef5f5d3b64972
          createdAt: "2025-09-30T07:54:26Z"
          description: HULFT10 for Container Platform is file integration middleware
            that provides transfer functions for the data accumulated in your IT system
            during the daily operations of your business, such as sales data, customer
            information, and various logs. File transfer from/to on-premises HULFT
            is also available.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: To use HULFT10 for Container
            Platform, you must purchase the product and have a key issued. Please
            refer to [Start Up Guide](https://www.hulft.com/help/ja-jp/HULFT-V10/CONP-SUGOS)
            for the deployment procedure.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Saison Technology Co., Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hulft is the Schema for the hulfts API
            displayName: Hulft
            kind: Hulft
            name: hulfts.hulft.saison-technology.com
            version: v1
        description: |
          HULFT10 for Container Platform is file integration middleware that provides transfer functions for the data accumulated in your IT system during the daily operations of your business, such as sales data, customer information, and various logs. File transfer from/to on-premises HULFT is also available.
          To use HULFT10 for Container Platform, you must purchase the product and have a key issued. Please refer to [Start Up Guide](https://www.hulft.com/help/ja-jp/HULFT-V10/CONP-SUGOS) for the deployment procedure.
        displayName: HULFT10 for Container Platform
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backup & Recovery
        - Storage
        - Migration
        links:
        - name: Product page
          url: https://www.hulft.com/software/hulft-10
        - name: Product manual
          url: https://www.hulft.com/help/ja-jp/HULFT-V10/index.htm
        maintainers:
        - email: salessupport_mp@hulft.com
          name: Saison Technology Co., Ltd.
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Saison Technology Co., Ltd.
          url: https://www.hulft.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fa0acefda43e309e7f9f31f5247c933d0b868ff1573ece2f38e0549d7e2ec942
        - registry.connect.redhat.com/saison-technology/hulft-operator@sha256:8ee0a29d0ce96f4ec230f387467b241dccbb2bd350cd642b275ef5f5d3b64972
        version: 10.5.0
      entries:
      - name: hulft.v10.5.0
        version: 10.5.0
      - name: hulft.v10.4.0
        version: 10.4.0
      - name: hulft.v10.3.0
        version: 10.3.0
      name: stable
    defaultChannel: stable
    packageName: hulft
    provider:
      name: Saison Technology Co., Ltd.
      url: https://www.hulft.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: local-storage-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: local-storage-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "local.storage.openshift.io/v1",
                "kind": "LocalVolume",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "storageClassDevices": [
                    {
                      "devicePaths": [
                        "/dev/disk/by-id/ata-crucial"
                      ],
                      "fsType": "ext4",
                      "storageClassName": "foobar",
                      "volumeMode": "Filesystem"
                    }
                  ]
                }
              },
              {
                "apiVersion": "local.storage.openshift.io/v1alpha1",
                "kind": "LocalVolumeSet",
                "metadata": {
                  "name": "example-localvolumeset"
                },
                "spec": {
                  "deviceInclusionSpec": {
                    "deviceMechanicalProperties": [
                      "Rotational",
                      "NonRotational"
                    ],
                    "deviceTypes": [
                      "disk"
                    ],
                    "maxSize": "100G",
                    "minSize": "10G"
                  },
                  "maxDeviceCount": 10,
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "storageClassName": "example-storageclass",
                  "volumeMode": "Block"
                }
              },
              {
                "apiVersion": "local.storage.openshift.io/v1alpha1",
                "kind": "LocalVolumeDiscovery",
                "metadata": {
                  "name": "auto-discover-devices"
                },
                "spec": {
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-local-storage-rhel9-operator@sha256:6f6491f023ffa21eff6a9b938cb1869f362d6b4a5b04c27dc5f13f49622ca178
          createdAt: "2019-08-14T00:00:00Z"
          description: Configure and use local storage volumes.
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.17"}]'
          olm.skipRange: '>=4.3.0 <4.16.0-202509240123'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-local-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["localvolumediscoveryresults.local.storage.openshift.io"]'
          repository: https://github.com/openshift/local-storage-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manage local storage volumes for OpenShift
            displayName: Local Volume
            kind: LocalVolume
            name: localvolumes.local.storage.openshift.io
            version: v1
          - description: A Local Volume set allows you to filter a set of storage
              volumes, group them and create a dedicated storage class to consume
              storage from the set of volumes.
            displayName: Local Volume Set
            kind: LocalVolumeSet
            name: localvolumesets.local.storage.openshift.io
            version: v1alpha1
          - description: Discover list of potentially usable disks on the chosen set
              of nodes
            displayName: Local Volume Discovery
            kind: LocalVolumeDiscovery
            name: localvolumediscoveries.local.storage.openshift.io
            version: v1alpha1
          - description: Disc inventory of available disks from selected nodes
            displayName: Local Volume Discovery Result
            kind: LocalVolumeDiscoveryResult
            name: localvolumediscoveryresults.local.storage.openshift.io
            version: v1alpha1
        description: |
          Operator that configures local storage volumes for use in Kubernetes and OpenShift. OpenShift 4.2 and above are the only supported OpenShift versions.
        displayName: Local Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - local storage
        links:
        - name: Documentation
          url: https://github.com/openshift/local-storage-operator/tree/master/docs
        - name: Source Repository
          url: https://github.com/openshift/local-storage-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-local-storage-rhel9-operator@sha256:6f6491f023ffa21eff6a9b938cb1869f362d6b4a5b04c27dc5f13f49622ca178
        - registry.redhat.io/openshift4/ose-local-storage-diskmaker-rhel9@sha256:d8d8b21fce5930a36c9755d82037fce1ab81f5ae2d5fb1bfd643d500c4ec7fba
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel9@sha256:ec5847d5006d66a014315decd579e10104b086a06bfefa33430545b4ffcece71
        version: 4.16.0-202509240123
      entries:
      - name: local-storage-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: local-storage-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: local-storage-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: local-storage-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: local-storage-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: local-storage-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: local-storage-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: local-storage-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: local-storage-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: local-storage-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: local-storage-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: local-storage-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: local-storage-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: local-storage-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: local-storage-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: local-storage-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: local-storage-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: local-storage-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: local-storage-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: local-storage-operator.v4.16.0-202410031007
        version: 4.16.0-202410031007
      - name: local-storage-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: local-storage-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: local-storage-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: local-storage-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: local-storage-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: local-storage-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: local-storage-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: local-storage-operator.v4.16.0-202408021139
        version: 4.16.0-202408021139
      - name: local-storage-operator.v4.16.0-202407190036
        version: 4.16.0-202407190036
      - name: local-storage-operator.v4.16.0-202407111006
        version: 4.16.0-202407111006
      - name: local-storage-operator.v4.16.0-202407081937
        version: 4.16.0-202407081937
      - name: local-storage-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: local-storage-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: local-storage-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: local-storage-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: local-storage-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: security-profiles-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: security-profiles-operator.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "AppArmorProfile",
                "metadata": {
                  "annotations": {
                    "description": "Block writing to any files in the disk."
                  },
                  "name": "test-profile"
                },
                "spec": {
                  "abstract": {
                    "filesystem": {
                      "readOnlyPaths": [],
                      "readWritePaths": [],
                      "writeOnlyPaths": []
                    }
                  }
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "SecurityProfilesOperatorDaemon",
                "metadata": {
                  "name": "spod",
                  "namespace": "security-profiles-operator"
                },
                "spec": {
                  "enableAppArmor": false,
                  "enableLogEnricher": false,
                  "enableSelinux": false
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "ProfileRecording",
                "metadata": {
                  "name": "test-recording"
                },
                "spec": {
                  "kind": "SeccompProfile",
                  "podSelector": {
                    "matchLabels": {
                      "app": "alpine"
                    }
                  },
                  "recorder": "bpf"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "ProfileBinding",
                "metadata": {
                  "name": "profile-binding"
                },
                "spec": {
                  "image": "nginx:1.19.1",
                  "profileRef": {
                    "kind": "SeccompProfile",
                    "name": "profile-allow-unsafe"
                  }
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha2",
                "kind": "RawSelinuxProfile",
                "metadata": {
                  "name": "errorlogger"
                },
                "spec": {
                  "policy": "(blockinherit container)\n(allow process var_log_t ( dir ( open read getattr lock search ioctl add_name remove_name write ))) \n(allow process var_log_t ( file ( getattr read write append ioctl lock map open create  ))) \n(allow process var_log_t ( sock_file ( getattr read write append open  ))) \n"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "Blocks all syscalls."
                  },
                  "name": "profile-block-all"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_ERRNO"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "UNSAFE: Allows all syscalls whilst logging their use. Similar to running as unconfined in terms of enforcement."
                  },
                  "name": "profile-complain-unsafe"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_LOG"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "UNSAFE: Allows all syscalls. Similar to running as unconfined as it provides no enforcement."
                  },
                  "name": "profile-allow-unsafe"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_ALLOW"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "Enables complain mode whilst blocking high-risk syscalls. Some essential syscalls are allowed to decrease log noise."
                  },
                  "name": "profile-complain-block-high-risk"
                },
                "spec": {
                  "architectures": [
                    "SCMP_ARCH_X86_64",
                    "SCMP_ARCH_PPC64LE"
                  ],
                  "defaultAction": "SCMP_ACT_LOG",
                  "syscalls": [
                    {
                      "action": "SCMP_ACT_ALLOW",
                      "names": [
                        "exit",
                        "exit_group",
                        "futex",
                        "nanosleep"
                      ]
                    },
                    {
                      "action": "SCMP_ACT_ERRNO",
                      "names": [
                        "acct",
                        "add_key",
                        "bpf",
                        "clock_adjtime",
                        "clock_settime",
                        "create_module",
                        "delete_module",
                        "finit_module",
                        "get_kernel_syms",
                        "get_mempolicy",
                        "init_module",
                        "ioperm",
                        "iopl",
                        "kcmp",
                        "kexec_file_load",
                        "kexec_load",
                        "keyctl",
                        "lookup_dcookie",
                        "mbind",
                        "mount",
                        "move_pages",
                        "name_to_handle_at",
                        "nfsservctl",
                        "open_by_handle_at",
                        "perf_event_open",
                        "personality",
                        "pivot_root",
                        "process_vm_readv",
                        "process_vm_writev",
                        "ptrace",
                        "query_module",
                        "quotactl",
                        "reboot",
                        "request_key",
                        "set_mempolicy",
                        "setns",
                        "settimeofday",
                        "stime",
                        "swapoff",
                        "swapon",
                        "_sysctl",
                        "sysfs",
                        "umount2",
                        "umount",
                        "unshare",
                        "uselib",
                        "userfaultfd",
                        "ustat",
                        "vm86old",
                        "vm86"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha2",
                "kind": "SelinuxProfile",
                "metadata": {
                  "name": "errorlogger"
                },
                "spec": {
                  "allow": {
                    "var_log_t": {
                      "dir": [
                        "open",
                        "read",
                        "getattr",
                        "lock",
                        "search",
                        "ioctl",
                        "add_name",
                        "remove_name",
                        "write"
                      ],
                      "file": [
                        "getattr",
                        "read",
                        "write",
                        "append",
                        "ioctl",
                        "lock",
                        "map",
                        "open",
                        "create"
                      ],
                      "sock_file": [
                        "getattr",
                        "read",
                        "write",
                        "append",
                        "open"
                      ]
                    }
                  },
                  "inherit": [
                    {
                      "name": "container"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          createdAt: "2025-02-25T13:34:54Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.1 <0.9.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-security-profiles
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ProfileBinding is the Schema for the profilebindings API.
            displayName: Profile Binding
            kind: ProfileBinding
            name: profilebindings.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: ProfileRecording is the Schema for the profilerecordings
              API.
            displayName: Profile Recording
            kind: ProfileRecording
            name: profilerecordings.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: SeccompProfile is a cluster level specification for a seccomp
              profile. See https://github.com/opencontainers/runtime-spec/blob/master/config-linux.md#seccomp
            displayName: Seccomp Profile
            kind: SeccompProfile
            name: seccompprofiles.security-profiles-operator.x-k8s.io
            version: v1beta1
          - description: SecurityProfileNodeStatus is a per-node status of a security
              profile
            displayName: Security Profile Node Status
            kind: SecurityProfileNodeStatus
            name: securityprofilenodestatuses.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: SecurityProfilesOperatorDaemon is the Schema to configure
              the spod deployment.
            displayName: Security Profiles Operator Daemon
            kind: SecurityProfilesOperatorDaemon
            name: securityprofilesoperatordaemons.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - kind: RawSelinuxProfile
            name: rawselinuxprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha2
          - description: SelinuxProfile is the Schema for the selinuxprofiles API.
            displayName: Selinux Profile
            kind: SelinuxProfile
            name: selinuxprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha2
          - description: AppArmorProfile is a cluster level specification for an AppArmor
              profile.
            displayName: App Armor Profile
            kind: AppArmorProfile
            name: apparmorprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha1
        description: SPO is an operator which aims to make it easier for users to
          use SELinux, seccomp and AppArmor in Kubernetes clusters
        displayName: Security Profiles Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - selinux
        - seccomp
        - apparmor
        - ebpf
        links:
        - name: Security Profiles Operator
          url: https://github.com/openshift/security-profiles-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-security-profiles-rhel8-operator@sha256:828274d71c65ef820e6a8b652e3f102624d09a0609fb00bae7b32dc62ff37fd1
        - registry.redhat.io/compliance/openshift-selinuxd-rhel9@sha256:3278d69814a8d0d1e31dcdef454c24828aadbb468a3d5d3e15910c76481765bd
        - registry.redhat.io/compliance/openshift-selinuxd-rhel8@sha256:81273ff149482109417064b40d333edb131481c05547e4ffa6485475460a48ba
        version: 0.9.0
      entries:
      - name: security-profiles-operator.v0.9.0
        version: 0.9.0
      - name: security-profiles-operator.v0.8.6
        version: 0.8.6
      - name: security-profiles-operator.v0.8.5
        version: 0.8.5
      - name: security-profiles-operator.v0.8.4
        version: 0.8.4
      - name: security-profiles-operator.v0.8.2
        version: 0.8.2
      name: release-alpha-rhel-8
    defaultChannel: release-alpha-rhel-8
    packageName: security-profiles-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: prometheus-exporter-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometheus-exporter-operator.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.3scale.net/v1alpha1",
                "kind": "PrometheusExporter",
                "metadata": {
                  "name": "example-memcached"
                },
                "spec": {
                  "dbHost": "your-memcached-host",
                  "dbPort": 11211,
                  "grafanaDashboard": {
                    "label": {
                      "key": "autodiscovery",
                      "value": "enabled"
                    }
                  },
                  "type": "memcached"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/3scale/prometheus-exporter-operator:v0.7.0
          createdAt: "2023-09-07 00:00:00"
          description: Operator to setup 3rd party prometheus exporters, with a collection
            of grafana dashboards
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/3scale-ops/prometheus-exporter-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configures a prometheus exporter to monitor a memcached instance
            displayName: PrometheusExporter
            kind: PrometheusExporter
            name: prometheusexporters.monitoring.3scale.net
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK to centralize the setup of 3rd party prometheus exporters on **Kubernetes/OpenShift**, with a collection of grafana dashboards.

          By just providing a few parameters like *dbHost* or *dbPort* (operator manage the container image, port, argument, command, volumes... and also prometheus `ServiceMonitor` and `GrafanaDashboard` k8s objects), you can setup different prometheus exporters to monitor:
          * The **internals from different databases**
          * **HTTP/TCP endpoints** (availability, latency, SSL/TLS certificate expiration...)
          * Any available **cloudwatch metric from any AWS Service**
          * Sendgrid email statistics (delivered, bounces, errors, spam...)

          Current prometheus exporters types supported, managed by same prometheus-exporter-operator:
          * memcached
          * redis
          * mysql
          * postgresql
          * sphinx
          * es (elasticsearch)
          * cloudwatch
          * probe (blackbox)
          * sendgrid

          The operator manages the lifecycle of the following objects:
          * Deployment (one per CR)
          * Service (one per CR)
          * ServiceMonitor (optional, one per CR)
          * GrafanaDashboard (optional, one per Namespace)

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale-ops/prometheus-exporter-operator#documentation).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale-ops/prometheus-exporter-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:
          * Raising any issues you find using Prometheus Exporter Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale-ops/prometheus-exporter-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving [documentation](https://github.com/3scale-ops/prometheus-exporter-operator)
          * Talking about Prometheus Exporter Operator
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale-ops/prometheus-exporter-operator/issues).

          ### License
          Prometheus Exporter Operator is licensed under the [Apache 2.0 license](https://github.com/3scale-ops/prometheus-exporter-operator/blob/main/LICENSE)
        displayName: Prometheus Exporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - metrics
        - observability
        - prometheus
        - exporter
        - grafana
        - mysql
        - postgresql
        - memcached
        - redis
        - sphinx
        - elasticsearch
        - cloudwatch
        - blackbox
        - sendgrid
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/prometheus-exporter-operator
        maintainers:
        - email: 3scale-operations@redhat.com
          name: 3scale Ops
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/3scale/prometheus-exporter-operator:v0.7.0
        version: 0.7.0
      entries:
      - name: prometheus-exporter-operator.v0.7.0
        version: 0.7.0
      - name: prometheus-exporter-operator.v0.3.4
        version: 0.3.4
      name: alpha
    defaultChannel: alpha
    packageName: prometheus-exporter-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: boxboat-github-practice
      provider-url: https://github.com/boxboat-github-practice
    name: github-arc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: github-arc-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "github-practice.boxboat.com/v1alpha1",
                "kind": "ActionsRunnerController",
                "metadata": {
                  "name": "arc-sample"
                },
                "spec": {
                  "authSecret": {
                    "name": "ghauth"
                  },
                  "createRunnerNamespaces": true,
                  "openshift": true,
                  "runnerNamespaces": [
                    "test1",
                    "test2"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery, OpenShift Optional
          containerImage: ghcr.io/boxboat-github-practice/github-arc-operator:1.1.0
          createdAt: "2023-03-03T20:49:16Z"
          description: Operator for deploying the github actions runner controller
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/boxboat-github-practice/github-arc-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActionsRunnerController is the schema for the ActionsRunnerController
              API
            displayName: Actions Runner Controller
            kind: ActionsRunnerController
            name: actionsrunnercontrollers.github-practice.boxboat.com
            version: v1alpha1
        description: helm based operator for github actions runner controller
        displayName: github-arc-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - github
        - actions
        - boxboat
        - ibm
        - self-hosted
        - runners
        links:
        - name: Github Arc Operator
          url: https://github.com/boxboat-github-practice/github-arc-operator
        - name: Actions Runner Controller
          url: https://github.com/actions/actions-runner-controller
        - name: Boxboat
          url: https://boxboat.com/
        maintainers:
        - email: Michael.McMullen@ibm.com
          name: mbmcmullen27
        maturity: alpha
        provider:
          name: boxboat-github-practice
          url: https://github.com/boxboat-github-practice
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - ghcr.io/boxboat-github-practice/github-arc-operator:1.1.0
        version: 1.1.0
      entries:
      - name: github-arc-operator.v1.1.0
        version: 1.1.0
      - name: github-arc-operator.v1.0.4
        version: 1.0.4
      name: alpha
    defaultChannel: alpha
    packageName: github-arc-operator
    provider:
      name: boxboat-github-practice
      url: https://github.com/boxboat-github-practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: special-resource-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: special-resource-operator.v4.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sro.openshift.io/v1beta1",
                "kind": "SpecialResource",
                "metadata": {
                  "name": "simple-kmod"
                },
                "spec": {
                  "chart": {
                    "name": "simple-kmod",
                    "repository": {
                      "name": "example",
                      "url": "file:///charts/example"
                    },
                    "version": "0.0.1"
                  },
                  "driverContainer": {
                    "source": {
                      "git": {
                        "ref": "master",
                        "uri": "https://github.com/openshift-psap/kvc-simple-kmod.git"
                      }
                    }
                  },
                  "namespace": "simple-kmod",
                  "set": {
                    "apiVersion": "sro.openshift.io/v1beta1",
                    "buildArgs": [
                      {
                        "name": "KMODVER",
                        "value": "SRO"
                      }
                    ],
                    "kind": "Values",
                    "kmodNames": [
                      "simple-kmod",
                      "simple-procfs-kmod"
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional, Storage
          containerImage: quay.io/openshift-psap/special-resource-operator:release-4.9
          description: The special resource operator is an orchestrator for resources
            in a cluster required to deploy special hardware and software. Examples
            include hardware accelerators or software defined storage, where extra
            management or kernel modules are needed.
          olm.skipRange: '>=4.6.0 <4.9.0'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpecialResource is the Schema for the specialresources API
            displayName: Special Resource
            kind: SpecialResource
            name: specialresources.sro.openshift.io
            version: v1beta1
          required:
          - description: Discovers special resources and labels the nodes with the
              PCI vendor ID.
            displayName: Node Feature Discovery
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
        description: "## About the managed application \nThe SRO will deploy the complete
          stack for enabling the accelerator in a cluster. \nIt manages the drivers,
          device-plugin, node-exporter, grafana and feature discovery.\n## About this
          Operator\n README: [https://github.com/openshift-psap/special-resource-operator](https://github.com/openshift-psap/special-resource-operator)\n\n##
          Prerequisites for enabling this Operator\nThe SRO has a dependency on NFD
          (Node Feature Discovery)\n"
        displayName: Special Resource Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GPI
        links:
        - name: README
          url: https://github.com/openshift-psap/special-resource-operator
        maintainers:
        - email: zkaiser@redhat.com
          name: Zvonko Kaiser
        - email: dagray@redhat.com
          name: David Gray
        maturity: alpha
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy
        - quay.io/openshift-psap/special-resource-operator:release-4.9
        version: 4.9.0
      entries:
      - name: special-resource-operator.v4.9.0
        version: 4.9.0
      - name: special-resource-operator.v4.8.2
        version: 4.8.2
      - name: special-resource-operator.v4.8.1
        version: 4.8.1
      - name: special-resource-operator.v4.8.0
        version: 4.8.0
      name: alpha
    defaultChannel: alpha
    packageName: special-resource-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Digital.ai
      provider-url: https://digital.ai
    name: digitalai-release-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: digitalai-release-operator.v25.3.0-beta.813
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 5,
                    "metrics": [],
                    "minReplicas": 3
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "cluster": {
                    "management": {
                      "ssl": {
                        "enabled": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    },
                    "manager": "xlr-legacy",
                    "mode": "full",
                    "name": "xl-release_cluster",
                    "node": {
                      "datacenter": "default"
                    },
                    "remote": {
                      "sslEngine": {
                        "enabled": false,
                        "enabledAlgorithms": [
                          "TLS_DHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_DHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_RSA_WITH_AES_128_CBC_SHA"
                        ],
                        "hostnameVerification": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "requireMutualAuthentication": false,
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"${XLR_CLUSTER_NAME}\"\n    manager = \"${XLR_CLUSTER_MANAGER}\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n      {{- if .Values.storage.reportingEngine.downingProviderClass }}\n      cluster {\n        downing-provider-class = {{.Values.storage.reportingEngine.downingProviderClass | quote}}\n      }\n      {{- end }}\n\n      {{- if .Values.cluster.remote.sslEngine.enabled }}\n      remote {\n        artery {  \n          ssl {  \n            config-ssl-engine {  \n              enabled-algorithms = [\n                {{- $last := sub (len .Values.cluster.remote.sslEngine.enabledAlgorithms) 1 }}\n                {{ range $index, $v := .Values.cluster.remote.sslEngine.enabledAlgorithms }}\n                {{ $v | quote }}{{ if ne $index $last }},{{ end }}\n                {{- end }}\n              ]  \n              key-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEY_PASSWORD}\"  \n              key-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.keystore \"default\" (printf \"cluster_remote_keystore.%s\" .Values.cluster.remote.sslEngine.keystoreType) \"context\" $) }}\"  \n              key-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEYSTORE_PASSWORD}\"  \n              trust-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.truststore \"default\" (printf \"cluster_remote_truststore.%s\" .Values.cluster.remote.sslEngine.truststoreType) \"context\" $) }}\"  \n              trust-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_TRUSTSTORE_PASSWORD}\"  \n              hostname-verification = {{ if .Values.cluster.remote.sslEngine.hostnameVerification }}on{{ else }}off{{ end }}\n              require-mutual-authentication = {{ if .Values.cluster.remote.sslEngine.requireMutualAuthentication }}on{{ else }}off{{ end }}  \n            }  \n          }  \n          transport = tls-tcp  \n        }  \n      }\n      {{- end }}\n    }\n    {{- if .Values.cluster.management.ssl.enabled }}\n    xlr-pekko-native {\n      pekko {\n        management {\n          http {  \n            route-providers-read-only = true  \n          }  \n          ssl {  \n            enabled = true\n            keyPassword = \"${?CLUSTER_MANAGEMENT_SSL_KEY_PASSWORD}\"  \n            keyStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.keystore \"default\" (printf \"cluster_management_keystore.%s\" .Values.cluster.management.ssl.keystoreType) \"context\" $) }}\"  \n            keyStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_KEYSTORE_PASSWORD}\"  \n            keystoreType = {{ .Values.cluster.management.ssl.keystoreType }}  \n            trustStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.truststore \"default\" (printf \"cluster_management_truststore.%s\" .Values.cluster.management.ssl.truststoreType) \"context\" $) }}\"  \n            trustStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_TRUSTSTORE_PASSWORD}\"  \n            trustStoreType = {{ .Values.cluster.management.ssl.truststoreType }}  \n          }  \n        }  \n      }\n    }\n    {{- end }}\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\nmanagement.endpoint.prometheus.enabled = {{ .Values.metrics.prometheus.enabled | default false }}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      datacenter = ${XLR_CLUSTER_DATACENTER}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "cd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "livenessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/liveness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "jmx": {
                      "enabled": false
                    },
                    "prometheus": {
                      "enabled": false,
                      "podAnnotations": {
                        "prometheus.io/port": "{{ .Values.service.ports.metrics }}",
                        "prometheus.io/scrape": "true"
                      },
                      "rule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "serviceMonitor": {
                        "annotations": {},
                        "basicAuth": {
                          "password": {
                            "key": "releasePassword",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          },
                          "username": {
                            "key": "releaseUsername",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          }
                        },
                        "enabled": false,
                        "honorLabels": false,
                        "interval": "30s",
                        "jobLabel": "",
                        "labels": {},
                        "metricRelabelings": [],
                        "namespace": "",
                        "params": {},
                        "path": "/s/actuator/prometheus",
                        "podTargetLabels": {},
                        "relabelings": [],
                        "scrapeTimeout": "",
                        "selector": {
                          "service-type": "normal"
                        },
                        "targetLabels": {}
                      }
                    }
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.0-debian-12-r4"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.0-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/extra/lib",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.5.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rbac": {
                    "create": true
                  },
                  "readinessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {
                      "service-type": "normal"
                    },
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "startupProbe": {
                    "enabled": true,
                    "failureThreshold": 12,
                    "initialDelaySeconds": 60,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 450,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnamilegacy/os-shell",
                      "tag": "12-debian-12-r49"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabsunsupported/release-operator@sha256:52914f17dc55dfaef55aed382e7269408fb676a213d12117eafa906d81506824 '
          createdAt: "2025-09-02T05:48:19Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql.\n\n```yaml\napiVersion:
          xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n
          \ global:\n    security:\n      allowInsecureImages: true\n  k8sSetup:\n
          \   platform: Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula:
          true\n  keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n```\n\n#### Configuration Details\n\nThe sample configuration
          uses:\n\n- Autogenerated limited license\n- Simple administrator initial
          password\n- Default keystore setup\n- Only one replica is set for all pods\n-
          All Pods are missing resource limit setup\n- Default storage class and minimal
          size for persistent storage\n- Release pods are using persistence access
          mode `ReadWriteOnce`, for Release pods the requirement is to have `ReadWriteMany`\n-
          Embedded PostgreSQL for DB management\n\n#### Security Configuration Details\n\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabsunsupported/release-operator@sha256:52914f17dc55dfaef55aed382e7269408fb676a213d12117eafa906d81506824
        - docker.io/xebialabsunsupported/xl-release@sha256:0548a1ba63da5ae09c59b67a30240bf48877ab8950e09b0f0f6fd7c88ba34c13
        version: 25.3.0-beta.813
      entries:
      - name: digitalai-release-operator.v25.3.0-beta.813
        version: 25.3.0-beta.813
      - name: digitalai-release-operator.v25.3.0-beta.711
        version: 25.3.0-beta.711
      - name: digitalai-release-operator.v25.3.0-beta.609
        version: 25.3.0-beta.609
      - name: digitalai-release-operator.v25.3.0-beta.512
        version: 25.3.0-beta.512
      name: beta
    - currentCSV: digitalai-release-operator.v24.1.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "fsGroup": null,
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "fsGroup": null,
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  jvm {\n      override-jvm-options = {{ .Values.jvm.overrideOptions }}\n      networkaddress.cache.ttl = {{ index .Values.jvm.options \"networkaddress.cache.ttl\" }}\n  }\n}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "ingressClass": "haproxy-dair",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-client/xl",
                        "kube",
                        "upgrade",
                        "--skip-context-check",
                        "--local-repo",
                        "/opt/xebialabs/xl-op-blueprints",
                        "--answers",
                        "/opt/xebialabs/xl-client/generated_answers.yaml"
                      ],
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    },
                    "overrideOptions": true
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "tag": "1.26.0-debian-12-r1"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "tag": "1.10.1-debian-12-r5"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "tag": "15.6.0-debian-12-r7"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "tag": "3.12.13-debian-12-r2"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnami/os-shell",
                      "tag": "12-debian-12-r16"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator without SCC creation and with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: docker.io/xebialabs/release-operator@sha256:47d6e31f66fce52c6566240b85fb028890409ad3addeea134120d1afa4814883
          createdAt: "2024-10-16T04:25:19Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind:
          DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    runAsUser: null\n    runAsGroup: null\n
          \   fsGroup: null\n  containerSecurityContext:\n    runAsUser: null\n    runAsGroup:
          null\n  securityContextConstraints:\n    enabled: false\n  volumePermissions:\n
          \   enabled: false\n  route:\n    enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n        runAsUser: null\n        runAsGroup: null\n        fsGroup:
          null\n      containerSecurityContext:\n        enabled: false\n        runAsUser:
          null\n        runAsGroup: null\n      securityContextConstraints:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n      runAsUser: null\n
          \     runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     enabled: false\n      runAsUser: null\n      runAsGroup: null\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"dair-doc-digitalai-release-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xlr-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Release pods are using
          persistence access mode `ReadWriteOnce`, for Release pods the requirement
          is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB management\n- Embedded
          RabbitMQ for message queue management\n\n#### Security Configuration Details\n\n-
          `securityContextConstraints.enabled: false` - disables creation of SCCs;\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:771a9a173e033a3ad8b46f5c00a7036eaa88c8d8d1fbd89217325168998113ea
        - docker.io/xebialabs/release-operator@sha256:47d6e31f66fce52c6566240b85fb028890409ad3addeea134120d1afa4814883
        - docker.io/xebialabs/xl-release@sha256:d1b65493ed524d37c60483f44b8bcc877ff398c81d36c7e767f2254dadec98cc
        version: 24.1.10
      entries:
      - name: digitalai-release-operator.v24.1.10
        version: 24.1.10
      name: stable
    - currentCSV: digitalai-release-operator.v24.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  jvm {\n      override-jvm-options = {{ .Values.jvm.overrideOptions }}\n      networkaddress.cache.ttl = {{ index .Values.jvm.options \"networkaddress.cache.ttl\" }}\n  }\n}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "image": {
                        "tag": "v0.14.7"
                      },
                      "ingressClass": "haproxy-dair",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-client/xl",
                        "kube",
                        "upgrade",
                        "--skip-context-check",
                        "--local-repo",
                        "/opt/xebialabs/xl-op-blueprints",
                        "--answers",
                        "/opt/xebialabs/xl-client/generated_answers.yaml"
                      ],
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    },
                    "overrideOptions": true
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.26.2-debian-12-r10"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "15.6.0-debian-12-r20"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.12.14-debian-12-r7"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnamilegacy/os-shell",
                      "tag": "12-debian-12-r49"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator without SCC creation and with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:aca76f0f5db170ddfc6ad692ea2874b9efb83fe732540946d88a0c078907ec8c '
          createdAt: "2025-08-22T08:54:25Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind:
          DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  securityContextConstraints:\n    enabled: false\n
          \ volumePermissions:\n    enabled: false\n  route:\n    enabled: false\n
          \   annotations:\n      haproxy.router.openshift.io/cookie_name: JSESSIONID\n
          \     haproxy.router.openshift.io/disable_cookies: 'false'\n      haproxy.router.openshift.io/rewrite-target:
          /\n      haproxy.router.openshift.io/timeout: 120s\n    hostname: '<mandatory-release-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n      securityContextConstraints:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   securityContextConstraints:\n      enabled: false\n    volumePermissions:\n
          \     enabled: false\n```\n\nIf SCC is required, it must be managed manually.
          The necessary permissions for accessing SecurityContextConstraints must
          be granted when SCC is enabled; otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"dair-doc-digitalai-release-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xlr-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Release pods are using
          persistence access mode `ReadWriteOnce`, for Release pods the requirement
          is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB management\n- Embedded
          RabbitMQ for message queue management\n\n#### Security Configuration Details\n\n-
          `securityContextConstraints.enabled: false` - disables creation of SCCs;\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/release-operator@sha256:aca76f0f5db170ddfc6ad692ea2874b9efb83fe732540946d88a0c078907ec8c
        - docker.io/xebialabs/xl-release@sha256:1cadd1d6ad7389ac9fe398ffba246b78238e54c180045b9dd44fdeefa3cf2af5
        version: 24.1.21
      entries:
      - name: digitalai-release-operator.v24.1.21
        version: 24.1.21
      - name: digitalai-release-operator.v24.1.20
        version: 24.1.20
      - name: digitalai-release-operator.v24.1.19
        version: 24.1.19
      - name: digitalai-release-operator.v24.1.18
        version: 24.1.18
      - name: digitalai-release-operator.v24.1.17
        version: 24.1.17
      - name: digitalai-release-operator.v24.1.14
        version: 24.1.14
      - name: digitalai-release-operator.v24.1.10
        version: 24.1.10
      name: stable-v24.1
    - currentCSV: digitalai-release-operator.v24.3.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "#!/bin/sh\n\ncd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "16.3.0-debian-12-r23"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n{{ if eq .Values.queueType \"quorum\" -}}\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n{{- end }}\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.13.7-debian-12-r5"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "queueType": "classic",
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-release",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:725fb5f7cfa4098c4110a40f54dc646f7663267bcba53603a7c1303c44051803 '
          createdAt: "2025-09-26T18:24:33Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n
          \ name: dair-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n```\n\n#### Configuration
          Details\n\nThe sample configuration uses:\n\n- Autogenerated limited license\n-
          Simple administrator initial password\n- Default keystore setup\n- Only
          one replica is set for all pods\n- All Pods are missing resource limit setup\n-
          Default storage class and minimal size for persistent storage\n- Release
          pods are using persistence access mode `ReadWriteOnce`, for Release pods
          the requirement is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB
          management\n- Embedded RabbitMQ for message queue management\n\n#### Security
          Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined ranges (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/release-operator@sha256:725fb5f7cfa4098c4110a40f54dc646f7663267bcba53603a7c1303c44051803
        - docker.io/xebialabs/xl-release@sha256:e9e87975888fcd4c8c164db90cb0a5c7abde88bf272c59d97e465af7b49bc6be
        version: 24.3.11
      entries:
      - name: digitalai-release-operator.v24.3.11
        version: 24.3.11
      - name: digitalai-release-operator.v24.3.10
        version: 24.3.10
      - name: digitalai-release-operator.v24.3.9
        version: 24.3.9
      - name: digitalai-release-operator.v24.3.8
        version: 24.3.8
      - name: digitalai-release-operator.v24.3.7
        version: 24.3.7
      - name: digitalai-release-operator.v24.3.6
        version: 24.3.6
      - name: digitalai-release-operator.v24.3.5
        version: 24.3.5
      name: stable-v24.3
    - currentCSV: digitalai-release-operator.v25.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "cluster": {
                    "management": {
                      "ssl": {
                        "enabled": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    },
                    "manager": "xlr-legacy",
                    "mode": "full",
                    "name": "xl-release_cluster",
                    "node": {
                      "datacenter": "default"
                    },
                    "remote": {
                      "sslEngine": {
                        "enabled": false,
                        "enabledAlgorithms": [
                          "TLS_DHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_DHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_RSA_WITH_AES_128_CBC_SHA"
                        ],
                        "hostnameVerification": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "requireMutualAuthentication": false,
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"${XLR_CLUSTER_NAME}\"\n    manager = \"${XLR_CLUSTER_MANAGER}\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n      {{- if .Values.storage.reportingEngine.downingProviderClass }}\n      cluster {\n        downing-provider-class = {{.Values.storage.reportingEngine.downingProviderClass | quote}}\n      }\n      {{- end }}\n\n      {{- if .Values.cluster.remote.sslEngine.enabled }}\n      remote {\n        artery {  \n          ssl {  \n            config-ssl-engine {  \n              enabled-algorithms = [\n                {{- $last := sub (len .Values.cluster.remote.sslEngine.enabledAlgorithms) 1 }}\n                {{ range $index, $v := .Values.cluster.remote.sslEngine.enabledAlgorithms }}\n                {{ $v | quote }}{{ if ne $index $last }},{{ end }}\n                {{- end }}\n              ]  \n              key-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEY_PASSWORD}\"  \n              key-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.keystore \"default\" (printf \"cluster_remote_keystore.%s\" .Values.cluster.remote.sslEngine.keystoreType) \"context\" $) }}\"  \n              key-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEYSTORE_PASSWORD}\"  \n              trust-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.truststore \"default\" (printf \"cluster_remote_truststore.%s\" .Values.cluster.remote.sslEngine.truststoreType) \"context\" $) }}\"  \n              trust-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_TRUSTSTORE_PASSWORD}\"  \n              hostname-verification = {{ if .Values.cluster.remote.sslEngine.hostnameVerification }}on{{ else }}off{{ end }}\n              require-mutual-authentication = {{ if .Values.cluster.remote.sslEngine.requireMutualAuthentication }}on{{ else }}off{{ end }}  \n            }  \n          }  \n          transport = tls-tcp  \n        }  \n      }\n      {{- end }}\n    }\n    {{- if .Values.cluster.management.ssl.enabled }}\n    xlr-pekko-native {\n      pekko {\n        management {\n          http {  \n            route-providers-read-only = true  \n          }  \n          ssl {  \n            enabled = true\n            keyPassword = \"${?CLUSTER_MANAGEMENT_SSL_KEY_PASSWORD}\"  \n            keyStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.keystore \"default\" (printf \"cluster_management_keystore.%s\" .Values.cluster.management.ssl.keystoreType) \"context\" $) }}\"  \n            keyStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_KEYSTORE_PASSWORD}\"  \n            keystoreType = {{ .Values.cluster.management.ssl.keystoreType }}  \n            trustStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.truststore \"default\" (printf \"cluster_management_truststore.%s\" .Values.cluster.management.ssl.truststoreType) \"context\" $) }}\"  \n            trustStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_TRUSTSTORE_PASSWORD}\"  \n            trustStoreType = {{ .Values.cluster.management.ssl.truststoreType }}  \n          }  \n        }  \n      }\n    }\n    {{- end }}\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\nmanagement.endpoint.prometheus.enabled = {{ .Values.metrics.prometheus.enabled | default false }}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      datacenter = ${XLR_CLUSTER_DATACENTER}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "#!/bin/sh\n\ncd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "livenessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/liveness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "jmx": {
                      "enabled": false
                    },
                    "prometheus": {
                      "enabled": false,
                      "podAnnotations": {
                        "prometheus.io/port": "{{ .Values.service.ports.metrics }}",
                        "prometheus.io/scrape": "true"
                      },
                      "rule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "serviceMonitor": {
                        "annotations": {},
                        "basicAuth": {
                          "password": {
                            "key": "releasePassword",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          },
                          "username": {
                            "key": "releaseUsername",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          }
                        },
                        "enabled": false,
                        "honorLabels": false,
                        "interval": "30s",
                        "jobLabel": "",
                        "labels": {},
                        "metricRelabelings": [],
                        "namespace": "",
                        "params": {},
                        "path": "/s/actuator/prometheus",
                        "podTargetLabels": {},
                        "relabelings": [],
                        "scrapeTimeout": "",
                        "selector": {
                          "service-type": "normal"
                        },
                        "targetLabels": {}
                      }
                    }
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.1-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.4.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.3-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "readinessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {
                      "service-type": "normal"
                    },
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "startupProbe": {
                    "enabled": true,
                    "failureThreshold": 12,
                    "initialDelaySeconds": 60,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 450,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-release",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:838443baf39514c5e0ac0a09cab411942a3cc4a13e261c8c60138788fe8d96e4 '
          createdAt: "2025-09-26T17:28:37Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n
          \ name: dair-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n```\n\n#### Configuration
          Details\n\nThe sample configuration uses:\n\n- Autogenerated limited license\n-
          Simple administrator initial password\n- Default keystore setup\n- Only
          one replica is set for all pods\n- All Pods are missing resource limit setup\n-
          Default storage class and minimal size for persistent storage\n- Release
          pods are using persistence access mode `ReadWriteOnce`, for Release pods
          the requirement is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB
          management\n- Embedded RabbitMQ for message queue management\n\n#### Security
          Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined ranges (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/release-operator@sha256:838443baf39514c5e0ac0a09cab411942a3cc4a13e261c8c60138788fe8d96e4
        - docker.io/xebialabs/xl-release@sha256:9b63789cdb87745832f1283d3f0e28ac5a4514d767895e0914a6be6672f2d14f
        version: 25.1.5
      entries:
      - name: digitalai-release-operator.v25.1.5
        version: 25.1.5
      - name: digitalai-release-operator.v25.1.4
        version: 25.1.4
      - name: digitalai-release-operator.v25.1.3
        version: 25.1.3
      - name: digitalai-release-operator.v25.1.2
        version: 25.1.2
      - name: digitalai-release-operator.v25.1.1
        version: 25.1.1
      - name: digitalai-release-operator.v25.1.0
        version: 25.1.0
      name: stable-v25.1
    defaultChannel: stable-v24.3
    packageName: digitalai-release-operator
    provider:
      name: Digital.ai
      url: https://digital.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Confluent, Inc.
      provider-url: https://confluent.io
    name: confluent-for-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: confluent-for-kubernetes.v2.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.8.4",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:0ca8769b04d6134ee231af84dfe98fb65161148d0d32e85f18dad6a3a345ed33
          createdAt: "2025-10-01T13:12:32Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0 <2.10.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - confluentinc/confluent-operator@sha256:0ca8769b04d6134ee231af84dfe98fb65161148d0d32e85f18dad6a3a345ed33
        - docker.io/confluentinc/confluent-operator@sha256:0ca8769b04d6134ee231af84dfe98fb65161148d0d32e85f18dad6a3a345ed33
        - docker.io/confluentinc/cp-server@sha256:45eb6e4f9c1201202d90216fbbdfcd8655eb378c937f09ac0129cd259956337d
        - docker.io/confluentinc/cp-server-connect@sha256:24aaa1cc8eb0df643f1efa37999b7ac33d844bfd1e4502da0d678996e4ac5ddd
        - docker.io/confluentinc/confluent-init-container@sha256:175ba68c876eca5804597b1591fdf551b2091d2f1c50e4daed2c32958af5d649
        - docker.io/confluentinc/cp-zookeeper@sha256:3e0af6ac6d34dcb8b9ed8f7a08109a48b2b2acd331c1dc4ea875222685c9afbd
        - docker.io/confluentinc/cp-schema-registry@sha256:bba639a3f9e2bf5f99b5d056ed520d2c268a125bee4f546d6cbca7a2c58f7dbb
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:2024d117dd413e207e49f37558d79dc882fd29f6b39e2d722708f18076813361
        - docker.io/confluentinc/cp-kafka-rest@sha256:8218e8db4731219ceaf9dbc378ead9c49aa86f98bc5106f62fbf70958dfd1ee4
        - docker.io/confluentinc/cp-ksqldb-server@sha256:093e87359c176d13d81317e972888a835cd01092d061bfc03fd60f0d1b4c8428
        version: 2.10.3
      entries:
      - name: confluent-for-kubernetes.v2.10.3
        version: 2.10.3
      name: "2.10"
    - currentCSV: confluent-for-kubernetes.v2.11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.9.3",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:85b512d3f20dc6e4a4da1759ee5ed306a4da0604592563a9af35be2cdf9ba1b6
          createdAt: "2025-10-01T13:10:39Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.10.0 <2.11.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/confluent-operator@sha256:85b512d3f20dc6e4a4da1759ee5ed306a4da0604592563a9af35be2cdf9ba1b6
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:351b68b062824c15db7ef999ba04062defdd4ea1ebe9665932147f99cf74879d
        - docker.io/confluentinc/cp-ksqldb-server@sha256:14a44cbbc13094974e7a5e14b33a829480ec3da75868130fdc7abaeda707dc5e
        - confluentinc/confluent-operator@sha256:85b512d3f20dc6e4a4da1759ee5ed306a4da0604592563a9af35be2cdf9ba1b6
        - docker.io/confluentinc/cp-server@sha256:e61e6c122b50ed345bc3e4e39cda6c93430e669535f8a70957e221563a7af1bd
        - docker.io/confluentinc/confluent-init-container@sha256:bccc7b723afdbe4e04c668720cdb1570835f8fd58ab7dd9a06ecf807f1176019
        - docker.io/confluentinc/cp-kafka-rest@sha256:cc72b3585ef9e7557910ab3efe3082444abe1fee191cdfb8747907fd2643971f
        - docker.io/confluentinc/cp-schema-registry@sha256:bd974dda5abf35fcad66c25fad92ee684e82305435b291b953d221f74d2bc119
        - docker.io/confluentinc/cp-zookeeper@sha256:03eadde2afa390f7339707e6baef2ecf7affb898aaa432378f61da54f5636ba3
        - docker.io/confluentinc/cp-server-connect@sha256:0d3b81e511cdf820daac4a2458b3cada40b4a7c3d142062a48880d38186bbcc4
        version: 2.11.3
      entries:
      - name: confluent-for-kubernetes.v2.11.3
        version: 2.11.3
      name: "2.11"
    - currentCSV: confluent-for-kubernetes.v2.8.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.6.4",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:551b14777b121b1e083524a0d5cce03cbe2fd2833537751bcd6be6cf9801f95e
          createdAt: "2024-12-02T06:51:43Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.0 <2.8.5'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-server@sha256:1b2ae325f87047b83d37d506066f073c27b0c42a0fd3c3409aa079ba736d1975
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:01bcde2aa35c0ad062ece49c63d41bf487618ea3c33e42ea4c23fde7194b26c9
        - docker.io/confluentinc/cp-schema-registry@sha256:dec88d30c8f21bc7fb38d28603c97c81c666ccaee63ff0136ecff48b3b3ec775
        - docker.io/confluentinc/confluent-init-container@sha256:b1ad1aa4daa81967a93b02323172f96db6a50c7135866f016344b2eb1a736a55
        - docker.io/confluentinc/cp-server-connect@sha256:255875bc57d81c322d47d5f6f11bc217e99ae709d1fdc5aec8e11829dd34279a
        - confluentinc/confluent-operator@sha256:551b14777b121b1e083524a0d5cce03cbe2fd2833537751bcd6be6cf9801f95e
        - docker.io/confluentinc/cp-zookeeper@sha256:7e728feebfc90a693d028216da68a5d79e5e8212855c63bac08e5b10b141d143
        - docker.io/confluentinc/cp-kafka-rest@sha256:2a5bc0d39693682e25d79f5036e18b71376df0941fedc2cb962a4b4d5a5eff82
        - docker.io/confluentinc/cp-ksqldb-server@sha256:50e4834c44759ab4a2a0bbab0f20545e124325b7ee12ba1a275af73b57084104
        - docker.io/confluentinc/confluent-operator@sha256:551b14777b121b1e083524a0d5cce03cbe2fd2833537751bcd6be6cf9801f95e
        version: 2.8.5
      entries:
      - name: confluent-for-kubernetes.v2.8.5
        version: 2.8.5
      name: "2.8"
    - currentCSV: confluent-for-kubernetes.v2.9.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.7.5",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:20648758fdb98d0f8913e75fb919ad0b622de8ecda9ab394912103e5e17d4007
          createdAt: "2025-10-03T06:52:55Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0 <2.9.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-server-connect@sha256:560833472dbdf23cf85b780e8306ae1bfe16ce6e2c0a9b326c64a513117525f1
        - docker.io/confluentinc/cp-zookeeper@sha256:2bd46d1f5d1eec0675bb5d1c08634d6458bbac7cacfb233aa61c68137fd4a7ce
        - docker.io/confluentinc/cp-kafka-rest@sha256:b6ac629a880474d81182eaf8459b4d83fe8126db796f1f285798beecda55a50a
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:257c38a75b3f4241780e20f16a7adc026dcca590f2e9c8da7c4e86b961e7611f
        - confluentinc/confluent-operator@sha256:20648758fdb98d0f8913e75fb919ad0b622de8ecda9ab394912103e5e17d4007
        - docker.io/confluentinc/confluent-operator@sha256:20648758fdb98d0f8913e75fb919ad0b622de8ecda9ab394912103e5e17d4007
        - docker.io/confluentinc/cp-ksqldb-server@sha256:6817a555ae4c041869f9b422a115c22894b2cd3988e82dc601694ce713b3ccd4
        - docker.io/confluentinc/confluent-init-container@sha256:b138d954318cc98767c10b4564a1412a22766317a83b9a452285bb0b3d893c63
        - docker.io/confluentinc/cp-server@sha256:79243e7e8b721f9fdf26fc57330ebb27d69c3b4686dbdf5319f6f39c2c7be836
        - docker.io/confluentinc/cp-schema-registry@sha256:6506a8c659d4154ae3e2d702f08b4ccac63513392ba454cbc9cf384d070c46eb
        version: 2.9.7
      entries:
      - name: confluent-for-kubernetes.v2.9.7
        version: 2.9.7
      name: "2.9"
    - currentCSV: confluent-for-kubernetes.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:8.0.1",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-ng-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "alertManagerClient": {
                      "url": "http://controlcenter-next-gen.confluent.svc.cluster.local:9093"
                    },
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    },
                    "prometheusClient": {
                      "url": "http://controlcenter-next-gen.confluent.svc.cluster.local:9090"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center-next-gen:2.2.0",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "replicas": 1,
                  "services": {
                    "alertmanager": {
                      "image": "confluentinc/cp-enterprise-alertmanager:2.2.0"
                    },
                    "prometheus": {
                      "image": "confluentinc/cp-enterprise-prometheus:2.2.0",
                      "pvc": {
                        "dataVolumeCapacity": "10Gi"
                      }
                    }
                  },
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:8.0.1",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:8.0.1",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:8.0.1",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:8.0.1",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:8.0.1",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:55b61717690fe334b5eebf5e7a55590537abddbd404f2eede81d4dd8cd29eda4
          createdAt: "2025-10-03T07:59:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <3.0.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-kafka-rest@sha256:f45f23e891ae073037ed7670659ecb6c3f72f9eea33e109f16243eec24a287a3
        - docker.io/confluentinc/cp-server-connect@sha256:4739f3a07d9f8eadd58f448a8e05b1392ef9b1dc9e9085d25a71f32b5b806a6e
        - docker.io/confluentinc/cp-ksqldb-server@sha256:df94bb805cff4f92315135a042f650c42f887bca70a40df8ecef2603532f11a7
        - docker.io/confluentinc/cp-enterprise-prometheus@sha256:db0989ef9859dbc88bc59f9b8ade86f2ba13e9dc36533e654566218328e5ff57
        - docker.io/confluentinc/cp-enterprise-control-center-next-gen@sha256:6378a0174433685ee52a7fe2e48424d3f598827ed2a5570ec74b916102485e0d
        - confluentinc/confluent-operator@sha256:55b61717690fe334b5eebf5e7a55590537abddbd404f2eede81d4dd8cd29eda4
        - docker.io/confluentinc/confluent-operator@sha256:55b61717690fe334b5eebf5e7a55590537abddbd404f2eede81d4dd8cd29eda4
        - docker.io/confluentinc/cp-schema-registry@sha256:2152975dfb62afbdf1d438dd0eb1f7fc3c929adf5804fe4790eccda0c96b9208
        - docker.io/confluentinc/cp-server@sha256:207cabc12e0d122199da875b2b0e39d1cf6f7fe435d5f184a9c8d7dc57ea81d9
        - docker.io/confluentinc/cp-enterprise-alertmanager@sha256:fcce6f5311c26cd58427a039530a9f8dfd009249ff5a519694aa945d34502252
        - docker.io/confluentinc/confluent-init-container@sha256:bf38b3bd9851bfd3ef0635e93b9e8b9ada570d393f650f60d8ae8943537730dc
        version: 3.0.1
      entries:
      - name: confluent-for-kubernetes.v3.0.1
        version: 3.0.1
      name: "3.0"
    defaultChannel: "3.0"
    packageName: confluent-for-kubernetes
    provider:
      name: Confluent, Inc.
      url: https://confluent.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pachyderm, Inc.
      provider-url: https://pachyderm.com
    name: pachyderm-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pachyderm-operator.v0.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "Pachyderm",
                "metadata": {
                  "name": "pachyderm-sample"
                },
                "spec": {
                  "console": {
                    "disable": true
                  },
                  "pachd": {
                    "metrics": {
                      "disable": false
                    },
                    "storage": {
                      "amazon": {
                        "credentialSecretName": "pachyderm-aws-secret"
                      },
                      "backend": "AMAZON"
                    }
                  },
                  "version": "v2.1.6"
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermExport",
                "metadata": {
                  "name": "pachydermexport-sample"
                },
                "spec": {
                  "backup": {
                    "target": "pachyderm-sample"
                  }
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermImport",
                "metadata": {
                  "name": "pachydermimport-sample"
                },
                "spec": {
                  "backup": "backups/pachyderm-backup-20220610.tar.gz",
                  "destination": {
                    "name": "pachyderm-restore",
                    "namespace": "pachyderm-workloads"
                  },
                  "storageSecret": "s3-credentials-secret"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
          createdAt: "2021-05-20T08:05:00Z"
          description: Pachyderm provides the data layer that allows data science
            teams to productionize and scale their machine learning lifecycle with
            data driven automation, petabyte scalability and end-to-end reproducibility.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/pachyderm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/pachyderm-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pachyderm/openshift-operator
          support: Pachyderm, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PachydermExport is the Schema for the pachydermexports API
            displayName: Pachyderm Export
            kind: PachydermExport
            name: pachydermexports.aiml.pachyderm.com
            version: v1beta1
          - description: PachydermImport is the Schema for the pachydermimports API
            displayName: Pachyderm Import
            kind: PachydermImport
            name: pachydermimports.aiml.pachyderm.com
            version: v1beta1
          - description: Pachyderm is the Schema for the pachyderms API
            displayName: Pachyderm
            kind: Pachyderm
            name: pachyderms.aiml.pachyderm.com
            version: v1beta1
        description: "The Pachyderm operator aims to make deploying and managing Pachyderm
          instances on Openshift container platform easier.\n\nPachyderm is the data
          foundation for machine learning.  Pachyderm provides industry leading data
          versioning, pipelines and lineage that allow data science teams to automate
          the machine learning lifecycle and optimize their machine learning operations
          (MLOps). \n\n## Getting started\n\n1. Pachyderm uses an object store to
          store your data. Set the pachd storage backend at `pachyderm.spec.pachd.storage.backend`
          to `AMAZON`, `MICROSOFT`, `GOOGLE` or `MINIO` depending on your storage
          provider. Additionally, create a user with read/write access to the object
          store.\n\n2. Setup the user credentials in the section, `pachyderm.spec.pachd.storage`.
          Required fields depend on the backend type set in step 1 above.\n\n3. Finally,
          deploy the pachyderm cluster in its own namespace.\n\nFor more information
          on getting started, reference the documentation.\n"
        displayName: Pachyderm
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pachyderm
        - versioning
        - machine learning
        - artificial intelligence
        links:
        - name: Documentation
          url: https://docs.pachyderm.com/latest/
        maintainers:
        - email: kyle@pachyderm.io
          name: Kyle White
        - email: eochieng@redhat.com
          name: Edmund Ochieng
        maturity: beta
        minKubeVersion: v1.19.0
        provider:
          name: Pachyderm, Inc.
          url: https://pachyderm.com
        relatedImages:
        - registry.connect.redhat.com/pachyderm/worker@sha256:374f83d92899707bc87b0318fda6bffa78d2e9e523b827bb3d40693a2d1bdecd
        - registry.connect.redhat.com/pachyderm/etcd@sha256:7f330b51874379b1d4b71d94e37e0ecfa52b86d628296136970688cf204f53b5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
        - registry.connect.redhat.com/pachyderm/backup-handler@sha256:e2ef6bd8191e8ad58c765c0db8f5bb21ea7123caed0c0f58fe681a024201306d
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/pachd@sha256:116477fefdf413ea02c6ac3d013f1f91d3f6e4b16aed2408c9890832cdb41430
        - registry.redhat.io/rhel8/postgresql-13@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/init-utils@sha256:5c0258bfa0203079925648c0a7dda6d05a29b6f461854821a1eec541777ff359
        - registry.connect.redhat.com/pachyderm/pgbouncer@sha256:c3b6949227df24abb8934d111a34b44e87fb16194ac34a5d0fe284cff02e9ac0
        version: 0.1.4
      entries:
      - name: pachyderm-operator.v0.1.4
        version: 0.1.4
      name: stable
    defaultChannel: stable
    packageName: pachyderm-operator-rhmp
    provider:
      name: Pachyderm, Inc.
      url: https://pachyderm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-community-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.5.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "registry": "docker.io",
                  "tag": "5.4.6",
                  "oem": "",
                  "imagePullSecrets": "",
                  "psp": false,
                  "rbac": true,
                  "serviceAccount": "default",
                  "leastPrivilege": false,
                  "global": {
                    "cattle": {
                      "url": ""
                    },
                    "azure": {
                      "enabled": false,
                      "identity": {
                        "clientId": "DONOTMODIFY"
                      },
                      "marketplace": {
                        "planId": "DONOTMODIFY"
                      },
                      "extension": {
                        "resourceId": "DONOTMODIFY"
                      },
                      "serviceAccount": "csp",
                      "imagePullSecrets": "",
                      "images": {
                        "neuvector_csp_pod": {
                          "tag": "latest",
                          "image": "neuvector-billing-azure-by-suse-llc",
                          "registry": "registry.suse.de/suse/sle-15-sp5/update/pubclouds/images",
                          "imagePullPolicy": "IfNotPresent"
                        },
                        "controller": {
                          "tag": "5.2.4",
                          "image": "controller",
                          "registry": "docker.io/neuvector"
                        },
                        "manager": {
                          "tag": "5.2.4",
                          "image": "manager",
                          "registry": "docker.io/neuvector"
                        },
                        "enforcer": {
                          "tag": "5.2.4",
                          "image": "enforcer",
                          "registry": "docker.io/neuvector"
                        }
                      }
                    },
                    "aws": {
                      "enabled": false,
                      "accountNumber": "",
                      "roleName": "",
                      "serviceAccount": "csp",
                      "annotations": {},
                      "imagePullSecrets": "",
                      "image": {
                        "digest": "",
                        "repository": "neuvector/neuvector-csp-adapter",
                        "tag": "latest",
                        "imagePullPolicy": "IfNotPresent"
                      }
                    }
                  },
                  "bootstrapPassword": "",
                  "autoGenerateCert": true,
                  "defaultValidityPeriod": 365,
                  "internal": {
                    "certmanager": {
                      "enabled": false,
                      "secretname": "neuvector-internal"
                    },
                    "autoGenerateCert": true,
                    "autoRotateCert": true
                  },
                  "controller": {
                    "enabled": true,
                    "annotations": {},
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": {
                      "repository": "neuvector/controller",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "searchRegistries": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "nodePort": "",
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      }
                    },
                    "ranchersso": {
                      "enabled": false
                    },
                    "pvc": {
                      "enabled": false,
                      "existingClaim": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {
                        "userinitcfg.yaml": {
                          "users": [
                            {
                              "Fullname": "admin",
                              "Password": "",
                              "Role": "admin"
                            }
                          ]
                        }
                      }
                    },
                    "certupgrader": {
                      "env": [],
                      "schedule": "",
                      "imagePullPolicy": "IfNotPresent",
                      "timeout": 3600,
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "prime": {
                      "enabled": false,
                      "image": {
                        "repository": "neuvector/compliance-config",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "1.0.7",
                        "hash": ""
                      }
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": {
                      "repository": "neuvector/enforcer",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "env": [],
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/etcd"
                      }
                    ],
                    "resources": {},
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "manager": {
                    "enabled": true,
                    "image": {
                      "repository": "neuvector/manager",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "priorityClassName": "",
                    "env": {
                      "ssl": true,
                      "envs": []
                    },
                    "svc": {
                      "type": "ClusterIP",
                      "nodePort": "",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": "",
                      "tls": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "topologySpreadConstraints": [],
                    "affinity": {},
                    "podLabels": {},
                    "podAnnotations": {},
                    "tolerations": [],
                    "nodeSelector": {},
                    "runAsUser": "",
                    "probes": {
                      "enabled": false,
                      "timeout": 1,
                      "periodSeconds": 10,
                      "startupFailureThreshold": 30
                    }
                  },
                  "cve": {
                    "adapter": {
                      "enabled": false,
                      "image": {
                        "repository": "neuvector/registry-adapter",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "0.1.9",
                        "hash": ""
                      },
                      "priorityClassName": "",
                      "resources": {},
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt"
                      },
                      "harbor": {
                        "protocol": "https",
                        "secretName": ""
                      },
                      "svc": {
                        "type": "ClusterIP",
                        "loadBalancerIP": "",
                        "annotations": {}
                      },
                      "route": {
                        "enabled": true,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      },
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ingressClassName": "",
                        "path": "/",
                        "annotations": {
                          "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                        },
                        "tls": false,
                        "secretName": ""
                      },
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      }
                    },
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "cacert": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                      "image": {
                        "registry": "",
                        "repository": "neuvector/updater",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "0.0.5",
                        "hash": ""
                      },
                      "schedule": "0 0 * * *",
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": {
                        "registry": "",
                        "repository": "neuvector/scanner",
                        "imagePullPolicy": "Always",
                        "tag": "6",
                        "hash": ""
                      },
                      "priorityClassName": "",
                      "resources": {},
                      "topologySpreadConstraints": [],
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      },
                      "volumes": "",
                      "volumeMounts": ""
                    }
                  },
                  "resources": {},
                  "runtimePath": "",
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhooksvc": {
                    "enabled": true
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  },
                  "lease": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "false"
          containerImage: docker.io/neuvector/neuvector-operator@sha256:6815df6435eafaf7b2bc9d5e3f0c9b65f0a50ad4e05d2862e56b227e19aeca27
          createdAt: "2019-11-18T02:09:59Z"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: NeuVector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart. NeuVector Operator version 1.5.5 deploys NeuVector app version 5.4.6.

          **Complete below steps to create namespace and Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Privileged SCC is added to Service Account specified in the deploment yaml by Operator version 1.3.4 and above in new deployment. In case of upgrading NeuVector Operator from previous version to 1.3.4, please delete Privileged SCC before upgrading.

                   oc delete rolebinding -nneuvector system:openshift:scc:privileged

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s  neuvector/default


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - wavefront
        - wavefront collector
        - wavefront proxy
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - docker.io/neuvector/neuvector-operator@sha256:6815df6435eafaf7b2bc9d5e3f0c9b65f0a50ad4e05d2862e56b227e19aeca27
        version: 1.5.5
      entries:
      - name: neuvector-operator.v1.5.5
        version: 1.5.5
      - name: neuvector-operator.v1.5.4
        version: 1.5.4
      - name: neuvector-operator.v1.5.3
        version: 1.5.3
      - name: neuvector-operator.v1.5.2
        version: 1.5.2
      - name: neuvector-operator.v1.5.1
        version: 1.5.1
      - name: neuvector-operator.v1.5.0
        version: 1.5.0
      - name: neuvector-operator.v1.4.9
        version: 1.4.9
      - name: neuvector-operator.v1.4.8
        version: 1.4.8
      - name: neuvector-operator.v1.4.7
        version: 1.4.7
      - name: neuvector-operator.v1.4.6
        version: 1.4.6
      - name: neuvector-operator.v1.4.5
        version: 1.4.5
      - name: neuvector-operator.v1.4.4
        version: 1.4.4
      - name: neuvector-operator.v1.4.3
        version: 1.4.3
      - name: neuvector-operator.v1.4.2
        version: 1.4.2
      - name: neuvector-operator.v1.4.1
        version: 1.4.1
      - name: neuvector-operator.v1.4.0
        version: 1.4.0
      - name: neuvector-operator.v1.3.9
        version: 1.3.9
      - name: neuvector-operator.v1.3.8
        version: 1.3.8
      - name: neuvector-operator.v1.3.7
        version: 1.3.7
      - name: neuvector-operator.v1.3.6
        version: 1.3.6
      - name: neuvector-operator.v1.3.5
        version: 1.3.5
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-community-operator
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: mariadb-operator
      provider-url: https://github.com/mariadb-operator/mariadb-operator
    name: mariadb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-operator.v25.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.mariadb.mmontes.io/v1alpha1",
                "kind": "MariadbOperator",
                "metadata": {
                  "name": "mariadb-operator"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "caValidity": "35064h",
                    "certValidity": "8766h",
                    "enabled": true,
                    "extrArgs": [],
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "ha": {
                      "enabled": false,
                      "replicas": 3
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "lookaheadValidity": "2160h",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podSecurityContext": {},
                    "requeueDuration": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "automount": true,
                      "enabled": true,
                      "extraLabels": {},
                      "name": ""
                    },
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    },
                    "tolerations": []
                  },
                  "clusterName": "cluster.local",
                  "extrArgs": [],
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "fullnameOverride": "",
                  "ha": {
                    "enabled": false,
                    "replicas": 3
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "logLevel": "INFO",
                  "metrics": {
                    "enabled": false,
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "rbac": {
                    "enabled": true
                  },
                  "resources": {},
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "automount": true,
                    "enabled": true,
                    "extraLabels": {},
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "annotations": {},
                    "cert": {
                      "caPath": "/tmp/k8s-webhook-server/certificate-authority",
                      "certManager": {
                        "duration": "",
                        "enabled": false,
                        "issuerRef": {},
                        "renewBefore": ""
                      },
                      "path": "/tmp/k8s-webhook-server/serving-certs",
                      "secretAnnotations": {}
                    },
                    "extrArgs": [],
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "ha": {
                      "enabled": false,
                      "replicas": 3
                    },
                    "hostNetwork": false,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podSecurityContext": {},
                    "port": 10250,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "automount": true,
                      "enabled": true,
                      "extraLabels": {},
                      "name": ""
                    },
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    },
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup"
                },
                "spec": {
                  "args": [
                    "--single-transaction",
                    "--all-databases"
                  ],
                  "logLevel": "info",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "backups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Connection",
                "metadata": {
                  "name": "connection"
                },
                "spec": {
                  "database": "mariadb",
                  "healthCheck": {
                    "interval": "10s",
                    "retryInterval": "3s"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "params": {
                    "parseTime": "true"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "secretName": "connection",
                  "secretTemplate": {
                    "annotations": {
                      "k8s.mariadb.com/connection": "sample"
                    },
                    "databaseKey": "database",
                    "hostKey": "host",
                    "key": "dsn",
                    "labels": {
                      "k8s.mariadb.com/connection": "sample"
                    },
                    "passwordKey": "password",
                    "portKey": "port",
                    "usernameKey": "username"
                  },
                  "serviceName": "mariadb",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Database",
                "metadata": {
                  "name": "data-test"
                },
                "spec": {
                  "characterSet": "utf8",
                  "collate": "utf8_general_ci",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "retryInterval": "5s"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "ExternalMariaDB",
                "metadata": {
                  "name": "external-mariadb"
                },
                "spec": {
                  "connection": {
                    "healthCheck": {
                      "interval": "5s"
                    },
                    "secretName": "external-mariadb"
                  },
                  "host": "mariadb.default.svc.cluster.local",
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "port": 3306,
                  "username": "root"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "grant"
                },
                "spec": {
                  "database": "*",
                  "grantOption": true,
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "privileges": [
                    "SELECT",
                    "INSERT",
                    "UPDATE"
                  ],
                  "requeueInterval": "30s",
                  "retryInterval": "5s",
                  "table": "*",
                  "username": "user"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "MariaDB",
                "metadata": {
                  "name": "mariadb"
                },
                "spec": {
                  "connection": {
                    "secretName": "mariadb-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "database": "mariadb",
                  "galera": {
                    "enabled": true
                  },
                  "metrics": {
                    "enabled": true,
                    "passwordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "mariadb-metrics"
                    }
                  },
                  "myCnf": "[mariadb]\nbind-address=*\ndefault_storage_engine=InnoDB\nbinlog_format=row\ninnodb_autoinc_lock_mode=2\ninnodb_buffer_pool_size=1024M\nmax_allowed_packet=256M\n",
                  "passwordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-password"
                  },
                  "primaryConnection": {
                    "secretName": "mariadb-conn-primary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "primaryService": {
                    "type": "ClusterIP"
                  },
                  "replicas": 3,
                  "rootPasswordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-root"
                  },
                  "secondaryConnection": {
                    "secretName": "mariadb-conn-secondary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "secondaryService": {
                    "type": "ClusterIP"
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "storage": {
                    "size": "1Gi"
                  },
                  "updateStrategy": {
                    "type": "ReplicasFirstPrimaryLast"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "MaxScale",
                "metadata": {
                  "name": "maxscale-galera"
                },
                "spec": {
                  "admin": {
                    "guiEnabled": true,
                    "port": 8989
                  },
                  "auth": {
                    "generate": true
                  },
                  "config": {
                    "sync": {
                      "database": "mysql",
                      "interval": "5s",
                      "timeout": "10s"
                    }
                  },
                  "connection": {
                    "port": 3306,
                    "secretName": "mxs-galera-conn"
                  },
                  "guiKubernetesService": {
                    "metadata": {
                      "annotations": {
                        "metallb.universe.tf/loadBalancerIPs": "172.18.0.231"
                      }
                    },
                    "type": "LoadBalancer"
                  },
                  "kubernetesService": {
                    "annotations": {
                      "metallb.universe.tf/loadBalancerIPs": "172.18.0.224"
                    },
                    "type": "LoadBalancer"
                  },
                  "mariaDbRef": {
                    "name": "mariadb-galera"
                  },
                  "monitor": {
                    "cooperativeMonitoring": "majority_of_all",
                    "interval": "2s",
                    "params": {
                      "available_when_donor": "false",
                      "disable_master_failback": "false",
                      "disable_master_role_setting": "false"
                    }
                  },
                  "replicas": 3,
                  "requeueInterval": "10s",
                  "services": [
                    {
                      "listener": {
                        "params": {
                          "connection_metadata": "tx_isolation=auto"
                        },
                        "port": 3306,
                        "protocol": "MariaDBProtocol"
                      },
                      "name": "rw-router",
                      "params": {
                        "master_accept_reads": "true",
                        "max_replication_lag": "3s",
                        "max_slave_connections": "255",
                        "transaction_replay": "true",
                        "transaction_replay_attempts": "10",
                        "transaction_replay_timeout": "5s"
                      },
                      "router": "readwritesplit"
                    },
                    {
                      "listener": {
                        "port": 3307
                      },
                      "name": "rconn-master-router",
                      "params": {
                        "master_accept_reads": "true",
                        "max_replication_lag": "3s",
                        "router_options": "master"
                      },
                      "router": "readconnroute"
                    },
                    {
                      "listener": {
                        "port": 3308
                      },
                      "name": "rconn-slave-router",
                      "params": {
                        "max_replication_lag": "3s",
                        "router_options": "slave"
                      },
                      "router": "readconnroute"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "PhysicalBackup",
                "metadata": {
                  "name": "physicalbackup"
                },
                "spec": {
                  "compression": "bzip2",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "podAffinity": true,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "immediate": true,
                    "suspend": false
                  },
                  "serviceAccountName": "backup",
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "physicalbackups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "region": "us-east-1",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  },
                  "timeout": "1h"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore"
                },
                "spec": {
                  "backupRef": {
                    "name": "backup"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "targetRecoveryTime": "2023-12-19T09:00:00Z"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "SqlJob",
                "metadata": {
                  "name": "03-stars"
                },
                "spec": {
                  "database": "mariadb",
                  "dependsOn": [
                    {
                      "name": "01-users"
                    },
                    {
                      "name": "02-repos"
                    }
                  ],
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "sql": "CREATE TABLE IF NOT EXISTS stars (\n  id bigint PRIMARY KEY AUTO_INCREMENT,\n  user_id bigint NOT NULL,\n  repo_id bigint NOT NULL,\n  FOREIGN KEY (user_id) REFERENCES users(id) ON DELETE CASCADE,\n  FOREIGN KEY (repo_id) REFERENCES repos(id) ON DELETE CASCADE,\n  UNIQUE KEY (user_id, repo_id)\n);\nINSERT INTO stars(user_id, repo_id) \n  VALUES((SELECT id FROM users ORDER BY RAND() LIMIT 1), (SELECT id FROM repos ORDER BY RAND() LIMIT 1))\n  ON DUPLICATE KEY UPDATE id=id;\nDELETE FROM stars WHERE id = (SELECT id FROM stars ORDER BY RAND() LIMIT 1);\nSELECT r.name AS repo, COUNT(*) AS stars\nFROM stars s\nJOIN repos r\nON s.repo_id = r.id\nGROUP BY r.id\nORDER BY stars DESC;",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "user"
                },
                "spec": {
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxUserConnections": 20,
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "retryInterval": "5s"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: ghcr.io/mariadb-operator/mariadb-operator-helm:25.8.4
          createdAt: "2025-09-19T13:24:35Z"
          description: Run and operate MariaDB in a cloud native way
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mariadb-operator/mariadb-operator
          support: mariadb-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configures a backup
            displayName: Backup
            kind: Backup
            name: backups.k8s.mariadb.com
            version: v1alpha1
          - description: Configures a connection
            displayName: Connection
            kind: Connection
            name: connections.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a logical database
            displayName: Database
            kind: Database
            name: databases.k8s.mariadb.com
            version: v1alpha1
          - description: Configures an external MariaDB resources
            displayName: ExternalMariaDB
            kind: ExternalMariaDB
            name: externalmariadbs.k8s.mariadb.com
            version: v1alpha1
          - description: Grants permissions to an user in a database
            displayName: Grant
            kind: Grant
            name: grants.k8s.mariadb.com
            version: v1alpha1
          - description: Configures MariaDB helm chart based operator
            displayName: MariadbOperator
            kind: MariadbOperator
            name: mariadboperators.helm.mariadb.mmontes.io
            version: v1alpha1
          - description: Provisions a MariaDB instance
            displayName: MariaDB
            kind: MariaDB
            name: mariadbs.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a MaxScale database proxy
            displayName: MaxScale
            kind: MaxScale
            name: maxscales.k8s.mariadb.com
            version: v1alpha1
          - description: Configures a physical backup
            displayName: PhysicalBackup
            kind: PhysicalBackup
            name: physicalbackups.k8s.mariadb.com
            version: v1alpha1
          - description: Restores a backup
            displayName: Restore
            kind: Restore
            name: restores.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a SQL job
            displayName: SqlJob
            kind: SqlJob
            name: sqljobs.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a user
            displayName: User
            kind: User
            name: users.k8s.mariadb.com
            version: v1alpha1
        description: |
          Install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) via [OLM](https://olm.operatorframework.io/) using the [helm chart](https://artifacthub.io/packages/helm/mariadb-operator/mariadb-operator).

          This helm operator provides provides a 1:1 mapping between the official helm chart and the [`MariadbOperator`](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/config/samples/helm_v1alpha1_mariadboperator.yaml) CRD, allowing to install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) via OLM without having to do any change in the helm chart.

          Normally, you would install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) providing this `values.yaml` to the helm chart:
          ```yaml
          image:
            repository: ghcr.io/mariadb-operator/mariadb-operator
            pullPolicy: IfNotPresent
          logLevel: INFO
          ha:
            enabled: true
          metrics:
            enabled: true
            serviceMonitor:
              enabled: true
          webhook:
            cert:
              certManager:
                enabled: true
          ```

          This helm chart installation is abstracted in the [`MariadbOperator`](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/config/samples/helm_v1alpha1_mariadboperator.yaml) CRD, which will be reconciled by the helm operator:
          ```yaml
          apiVersion: helm.k8s.mariadb.com/v1alpha1
          kind: MariadbOperator
          metadata:
            name: mariadb-operator
          spec:
            image:
              repository: ghcr.io/mariadb-operator/mariadb-operator
              pullPolicy: IfNotPresent
            logLevel: INFO
            ha:
              enabled: true
            metrics:
              enabled: true
              serviceMonitor:
                enabled: true
            webhook:
              cert:
                certManager:
                  enabled: true
          ```

          Once you have installed the operator, you will able to install a [`MariaDB`](https://github.com/mariadb-operator/mariadb-operator/blob/main/examples/manifests/mariadb_v1alpha1_mariadb.yaml) instance. Refer to the main [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) documentation for getting started with the rest of CRDs.

          ## Documentation
          * [mariadb-operator](https://github.com/mariadb-operator/mariadb-operator/blob/main/README.md)
          * [mariadb-operator-helm](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/README.md)

          ## Releases
          This operator is automatically published in the following repositories whenever a new version of the [helm chart](https://artifacthub.io/packages/helm/mariadb-operator/mariadb-operator) is released:
          - [k8s-operatorhub/community-operators](https://github.com/k8s-operatorhub/community-operators)
          - [redhat-openshift-ecosystem/community-operators-prod](https://github.com/redhat-openshift-ecosystem/community-operators-prod)

          ## Roadmap
          Take a look at our [roadmap](https://github.com/mariadb-operator/mariadb-operator/blob/main/ROADMAP.md) and feel free to open an issue to suggest new features.

          ## Contributing
          We welcome and encourage contributions to this project! Please check our [contributing](https://github.com/mariadb-operator/mariadb-operator/blob/main/CONTRIBUTING.md) and [development](https://github.com/mariadb-operator/mariadb-operator/blob/main/docs/DEVELOPMENT.md) guides. PRs welcome!

          ## Get in touch
          Join us on Slack: **[MariaDB Community Slack](https://r.mariadb.com/join-community-slack)**.
        displayName: MariaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mariadb
        - mysql
        - operator
        - mariadb-operator
        - database
        - maxscale
        links:
        - name: GitHub
          url: https://github.com/mariadb-operator/mariadb-operator
        maintainers:
        - email: mariadb-operator@proton.me
          name: mmontes11
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: mariadb-operator
          url: https://github.com/mariadb-operator/mariadb-operator
        relatedImages:
        - ghcr.io/mariadb-operator/mariadb-operator-helm:25.8.4
        version: 25.8.4
      entries:
      - name: mariadb-operator.v25.8.4
        version: 25.8.4
      - name: mariadb-operator.v25.8.3
        version: 25.8.3
      - name: mariadb-operator.v25.8.2
        version: 25.8.2
      - name: mariadb-operator.v25.8.1
        version: 25.8.1
      - name: mariadb-operator.v25.8.0
        version: 25.8.0
      - name: mariadb-operator.v0.38.1
        version: 0.38.1
      - name: mariadb-operator.v0.38.0
        version: 0.38.0
      - name: mariadb-operator.v0.37.1
        version: 0.37.1
      - name: mariadb-operator.v0.37.0
        version: 0.37.0
      - name: mariadb-operator.v0.36.0
        version: 0.36.0
      - name: mariadb-operator.v0.35.1
        version: 0.35.1
      - name: mariadb-operator.v0.35.0
        version: 0.35.0
      - name: mariadb-operator.v0.34.0
        version: 0.34.0
      - name: mariadb-operator.v0.33.0
        version: 0.33.0
      - name: mariadb-operator.v0.32.0
        version: 0.32.0
      - name: mariadb-operator.v0.31.0
        version: 0.31.0
      - name: mariadb-operator.v0.30.0
        version: 0.30.0
      - name: mariadb-operator.v0.29.0
        version: 0.29.0
      - name: mariadb-operator.v0.28.1
        version: 0.28.1
      - name: mariadb-operator.v0.27.0
        version: 0.27.0
      - name: mariadb-operator.v0.26.0
        version: 0.26.0
      - name: mariadb-operator.v0.24.0
        version: 0.24.0
      - name: mariadb-operator.v0.23.1
        version: 0.23.1
      - name: mariadb-operator.v0.23.0
        version: 0.23.0
      - name: mariadb-operator.v0.22.0
        version: 0.22.0
      - name: mariadb-operator.v0.21.0
        version: 0.21.0
      - name: mariadb-operator.v0.20.0
        version: 0.20.0
      - name: mariadb-operator.v0.19.0
        version: 0.19.0
      - name: mariadb-operator.v0.18.0
        version: 0.18.0
      - name: mariadb-operator.v0.17.0
        version: 0.17.0
      - name: mariadb-operator.v0.16.0
        version: 0.16.0
      - name: mariadb-operator.v0.15.1
        version: 0.15.1
      - name: mariadb-operator.v0.15.0
        version: 0.15.0
      - name: mariadb-operator.v0.14.0
        version: 0.14.0
      - name: mariadb-operator.v0.13.0
        version: 0.13.0
      - name: mariadb-operator.v0.11.0
        version: 0.11.0
      - name: mariadb-operator.v0.10.0
        version: 0.10.0
      - name: mariadb-operator.v0.9.1
        version: 0.9.1
      - name: mariadb-operator.v0.9.0
        version: 0.9.0
      - name: mariadb-operator.v0.8.0
        version: 0.8.0
      - name: mariadb-operator.v0.7.1
        version: 0.7.1
      - name: mariadb-operator.v0.7.0
        version: 0.7.0
      - name: mariadb-operator.v0.6.7
        version: 0.6.7
      - name: mariadb-operator.v0.6.6
        version: 0.6.6
      - name: mariadb-operator.v0.6.5
        version: 0.6.5
      - name: mariadb-operator.v0.6.4
        version: 0.6.4
      name: alpha
    defaultChannel: alpha
    packageName: mariadb-operator
    provider:
      name: mariadb-operator
      url: https://github.com/mariadb-operator/mariadb-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Anchore Inc.
      provider-url: ""
    name: anchore-engine
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anchore-engine.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.anchore.io/v1alpha1",
                "kind": "AnchoreEngine",
                "metadata": {
                  "name": "anchoreengine-sample"
                },
                "spec": {
                  "anchoreGlobal": {
                    "defaultAdminEmail": "example@email.com"
                  },
                  "postgresql": {
                    "extraEnv": [
                      {
                        "name": "POSTGRESQL_USER",
                        "value": "anchoreengine"
                      },
                      {
                        "name": "POSTGRESQL_PASSWORD",
                        "value": "anchore-postgres,123"
                      },
                      {
                        "name": "POSTGRESQL_DATABASE",
                        "value": "anchore"
                      },
                      {
                        "name": "PGUSER",
                        "value": "postgres"
                      }
                    ],
                    "postgresDatabase": "anchore",
                    "postgresPassword": "anchore-postgres,123",
                    "postgresUser": "anchoreengine"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/anchore/engine-operator@sha256:baf0f5acd319bc4d0132c88922ee3f357fdcdc278d931f243d0cbec3f6d416c6
          createdAt: "2021-10-06T23:53:51Z"
          description: Anchore Engine - container image scanning service for policy-based
            security, best-practice and compliance enforcement.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/anchore/engine-operator
          support: anchore.slack.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys Anchore Engine
            displayName: Anchore Engine Operator
            kind: AnchoreEngine
            name: anchoreengines.charts.anchore.io
            version: v1alpha1
        description: "Anchore Engine is an open source software system that provides
          a centralized service for analyzing container images, \n scanning for security
          vulnerabilities, and enforcing deployment policies. Anchore Engine is provided
          as a Docker \n container image that can be run standalone or with an orchestration
          platform such as Kubernetes. Anchore Engine allows \n users to perform detailed
          analysis on their container images, run queries, produce reports, and define
          policies that \n can be used in CI/CD pipelines. Users can extend Anchore
          Engine with plugins that add new queries, image analysis, and \n policies.
          Anchore Engine can be accessed directly through a RESTful API or via the
          Anchore CLI.\n\n The Anchore Engine Operator is based on the official stable
          [Helm Chart](https://github.com/anchore/anchore-charts/tree/master/stable/anchore-engine)."
        displayName: Anchore Engine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - analysis
        - docker
        - anchore
        - anchore-engine
        - image
        - security
        - vulnerability
        - scanner
        links:
        - name: Anchore Inc.
          url: https://www.anchore.com
        - name: Anchore Engine
          url: https://github.com/anchore/anchore-engine
        maintainers:
        - email: bradyt@anchore.com
          name: Brady Todhunter
        - email: zach@anchore.com
          name: Zach Hill
        maturity: alpha
        provider:
          name: Anchore Inc.
        relatedImages:
        - registry.connect.redhat.com/anchore/engine-operator@sha256:baf0f5acd319bc4d0132c88922ee3f357fdcdc278d931f243d0cbec3f6d416c6
        - registry.connect.redhat.com/anchore/engine0@sha256:68e09d5ff695b46f24de3d7610f88dfc14b77b38758bbd68d03350dbae3deb0f
        - registry.redhat.io/rhel8/postgresql-96@sha256:e8177c5de05ccdd6d12e360b65c63c889b2de725d573ffdedd8914230b118639
        version: 1.0.1
      entries:
      - name: anchore-engine.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: anchore-engine
    provider:
      name: Anchore Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: https://github.com/Mellanox/maintenance-operator/
    name: nvidia-maintenance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nvidia-maintenance-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "MaintenanceOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "maintenanceoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "maintenanceoperatorconfig",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "maintenanceoperatorconfig-sample"
                },
                "spec": {
                  "logLevel": "info",
                  "maxNodeMaintenanceTimeSeconds": 1600,
                  "maxParallelOperations": 1,
                  "maxUnavailable": "30%"
                }
              },
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "nodemaintenance-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodemaintenance",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "cordon": true,
                  "drainSpec": {
                    "deleteEmptyDir": true,
                    "force": true,
                    "podEvictionFilters": [
                      {
                        "byResourceNameRegex": "nvidia.com/gpu-*"
                      },
                      {
                        "byResourceNameRegex": "nvidia.com/rdma*"
                      }
                    ],
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  },
                  "nodeName": "worker-node-01",
                  "requestorID": "some-cluster-component.vendor.com",
                  "waitForPodCompletion": {
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
          createdAt: "2025-07-13T12:33:06Z"
          description: Node maintenance in K8s cluster in a coordinated manner
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-maintenance-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: NVIDIA
          repository: https://github.com/Mellanox/maintenance-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MaintenanceOperatorConfig is the Schema for the maintenanceoperatorconfigs
              API
            displayName: Maintenance Operator Config
            kind: MaintenanceOperatorConfig
            name: maintenanceoperatorconfigs.maintenance.nvidia.com
            version: v1alpha1
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.maintenance.nvidia.com
            version: v1alpha1
        description: Coordinates And Performs Common Node Preparations For Maintenance
        displayName: NVIDIA Maintenance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node
        - maintenance
        - drain
        links:
        - name: NVIDIA Maintenance Operator
          url: https://github.com/Mellanox/maintenance-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/maintenance-operator/
        relatedImages:
        - ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
        version: 0.2.3
      entries:
      - name: nvidia-maintenance-operator.v0.2.3
        version: 0.2.3
      name: stable
    - currentCSV: nvidia-maintenance-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "MaintenanceOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "maintenanceoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "maintenanceoperatorconfig",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "maintenanceoperatorconfig-sample"
                },
                "spec": {
                  "logLevel": "info",
                  "maxNodeMaintenanceTimeSeconds": 1600,
                  "maxParallelOperations": 1,
                  "maxUnavailable": "30%"
                }
              },
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "nodemaintenance-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodemaintenance",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "cordon": true,
                  "drainSpec": {
                    "deleteEmptyDir": true,
                    "force": true,
                    "podEvictionFilters": [
                      {
                        "byResourceNameRegex": "nvidia.com/gpu-*"
                      },
                      {
                        "byResourceNameRegex": "nvidia.com/rdma*"
                      }
                    ],
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  },
                  "nodeName": "worker-node-01",
                  "requestorID": "some-cluster-component.vendor.com",
                  "waitForPodCompletion": {
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
          createdAt: "2025-07-13T12:33:06Z"
          description: Node maintenance in K8s cluster in a coordinated manner
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-maintenance-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: NVIDIA
          repository: https://github.com/Mellanox/maintenance-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MaintenanceOperatorConfig is the Schema for the maintenanceoperatorconfigs
              API
            displayName: Maintenance Operator Config
            kind: MaintenanceOperatorConfig
            name: maintenanceoperatorconfigs.maintenance.nvidia.com
            version: v1alpha1
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.maintenance.nvidia.com
            version: v1alpha1
        description: Coordinates And Performs Common Node Preparations For Maintenance
        displayName: NVIDIA Maintenance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node
        - maintenance
        - drain
        links:
        - name: NVIDIA Maintenance Operator
          url: https://github.com/Mellanox/maintenance-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/maintenance-operator/
        relatedImages:
        - ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
        version: 0.2.3
      entries:
      - name: nvidia-maintenance-operator.v0.2.3
        version: 0.2.3
      name: v0.2
    defaultChannel: v0.2
    packageName: nvidia-maintenance-operator
    provider:
      name: NVIDIA
      url: https://github.com/Mellanox/maintenance-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: mariadb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mariadb.openstack.org/v1beta1",
                "kind": "Galera",
                "metadata": {
                  "name": "openstack"
                },
                "spec": {
                  "replicas": 3,
                  "secret": "osp-secret",
                  "storageClass": "local-storage",
                  "storageRequest": "500M"
                }
              },
              {
                "apiVersion": "mariadb.openstack.org/v1beta1",
                "kind": "MariaDBAccount",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "mariadb-operator",
                    "app.kubernetes.io/instance": "mariadbaccount-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mariadbaccount",
                    "app.kubernetes.io/part-of": "mariadb-operator",
                    "mariaDBDatabaseName": "neutron"
                  },
                  "name": "neutron1"
                },
                "spec": {
                  "secret": "neutrondb-secret",
                  "userName": "neutron1"
                }
              },
              {
                "apiVersion": "mariadb.openstack.org/v1beta1",
                "kind": "MariaDBDatabase",
                "metadata": {
                  "labels": {
                    "dbName": "mariadb"
                  },
                  "name": "scott"
                },
                "spec": {
                  "name": "scott",
                  "secret": "tiger"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:39Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Galera is the Schema for the galeras API
            displayName: Galera
            kind: Galera
            name: galeras.mariadb.openstack.org
            version: v1beta1
          - description: MariaDBAccount is the Schema for the mariadbaccounts API
            displayName: Maria DBAccount
            kind: MariaDBAccount
            name: mariadbaccounts.mariadb.openstack.org
            version: v1beta1
          - description: MariaDBDatabase is the Schema for the mariadbdatabases API
            displayName: Maria DBDatabase
            kind: MariaDBDatabase
            name: mariadbdatabases.mariadb.openstack.org
            version: v1beta1
        description: MariaDB Operator
        displayName: MariaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Identity
        - MariaDB
        links:
        - name: MariaDB Operator
          url: https://github.com/openstack-k8s-operators/mariadb-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/mariadb-rhel9-operator@sha256:5ae64a3cf9bf5eaf244548c3bf2e8ef89c9436cb676721c1f1d1c2ff8285e7e5
        - registry.redhat.io/rhoso/openstack-mariadb-rhel9@sha256:13119fe0ec56263a2bff3fc9c3892ea6386b837955280d8488f05d8ea6c4e44d
        version: 1.0.6
      entries:
      - name: mariadb-operator.v1.0.6
        version: 1.0.6
      - name: mariadb-operator.v1.0.5
        version: 1.0.5
      - name: mariadb-operator.v1.0.4
        version: 1.0.4
      - name: mariadb-operator.v1.0.3
        version: 1.0.3
      - name: mariadb-operator.v1.0.2
        version: 1.0.2
      - name: mariadb-operator.v1.0.1
        version: 1.0.1
      - name: mariadb-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: mariadb-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: GitLab, Inc.
      provider-url: https://gitlab.com
    name: gitlab-runner-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-runner-operator.v1.41.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta2",
                "kind": "Runner",
                "metadata": {
                  "name": "gitlab-runner-example",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "gitlabUrl": "https://gitlab.example.com",
                  "token": "runner-token-secret",
                  "tags": "kubernetes,example",
                  "concurrent": 10,
                  "imagePullPolicy": "IfNotPresent"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery, Developer Tools
          containerImage: registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
          createdAt: "2025-09-25T15:08:40Z"
          description: GitLab Runner operator manages lifecycle of GitLab Runner instances
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.11.0 <1.41.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: none
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
          support: GitLab, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Runner is the open source project used to run your jobs and
              send the results back to GitLab
            displayName: GitLab Runner
            kind: Runner
            name: runners.apps.gitlab.com
            version: v1beta2
        description: |-
          GitLab Runner is the lightweight, highly-scalable agent that runs your build jobs and sends the results back to a GitLab instance. GitLab Runner works in conjunction with GitLab CI/CD, the open-source continuous integration service included with GitLab.

              The GitLab Runner operator manages the lifecycle of GitLab Runner in Kubernetes or Openshift clusters. The operator aims to automate the tasks needed to run your CI/CD jobs in your container orchestration platform.

              ## Prerequisites

                For Kubernetes cluster, install cert-manager:

                ```shell
                kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.7.1/cert-manager.yaml
                ```

              ## GitLab Runner version

              This version of **GitLab Runner Operator** ships with **GitLab Runner v18.4.0**.

              To use a different version of **GitLab Runner** change the [`runnerImage` and `helperImage` properties](https://docs.gitlab.com/runner/configuration/configuring_runner_operator.html#operator-properties).

              ## Usage

               To link a GitLab Runner instance to a self-hosted GitLab instance or the hosted [GitLab](https://gitlab.com), you first need to:

               1. Create a secret containing the `runner-token` from your GitLab project.

                ```
                cat > gitlab-runner-secret.yml << EOF
                apiVersion: v1
                kind: Secret
                metadata:
                  name: gitlab-runner-secret
                type: Opaque
                # Only one of the following fields must be set. The Operator will fail to register Runner if both are provided.
                # # NOTE: runner-registration-token is deprecated and will be removed in GitLab 18.0. You should use runner-token instead.
                stringData:
                  runner-token: REPLACE_ME # your project runner token
                  # runner-registration-token: "" # your project runner secret
                EOF
                ```

                ```
                oc apply -f gitlab-runner-secret.yml
                ```

               2. Create the Custom Resource Definition (CRD) file and include the following information. The tags value must be openshift for the job to run.

                 ```
                 cat > gitlab-runner.yml << EOF
                 apiVersion: apps.gitlab.com/v1beta2
                 kind: Runner
                 metadata:
                   name: gitlab-runner
                 spec:
                   gitlabUrl: https://gitlab.example.com
                   buildImage: alpine
                   token: gitlab-runner-secret
                   tags: openshift
                 EOF
                 ```

                ```
                oc apply -f gitlab-runner.yml
                ```

              ## Full documentation

              Visit [Install GitLab Runner Operator](https://docs.gitlab.com/runner/install/operator.html)
        displayName: GitLab Runner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GitLab
        - CI/CD
        - DevOps
        - SAST
        - DAST
        links:
        - name: Gitlab Runner Operator
          url: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
        - name: GitLab Docs
          url: https://docs.gitlab.com
        - name: GitLab Runner
          url: https://docs.gitlab.com/runner/
        - name: GitLab CI/CD
          url: https://docs.gitlab.com/ee/ci/quick_start/
        maintainers:
        - email: support@gitlab.com
          name: Georgi N. Georgiev
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: GitLab, Inc.
          url: https://gitlab.com
        relatedImages:
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-ocp@sha256:0f2a5a3b6779d719eecbf079b99cf06b652603d74b563f57d753e32966d253dd
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-helper-ocp@sha256:29727fae83a226bcf57607240d4046a0b8c208b5ff42ee1f8ca0f115bb188809
        version: 1.41.1
      entries:
      - name: gitlab-runner-operator.v1.41.1
        version: 1.41.1
      - name: gitlab-runner-operator.v1.41.0
        version: 1.41.0
      - name: gitlab-runner-operator.v1.40.1
        version: 1.40.1
      - name: gitlab-runner-operator.v1.40.0
        version: 1.40.0
      - name: gitlab-runner-operator.v1.39.0
        version: 1.39.0
      name: stable
    defaultChannel: stable
    packageName: gitlab-runner-operator
    provider:
      name: GitLab, Inc.
      url: https://gitlab.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "docker.io/percona/percona-postgresql-operator:2.7.0-ppg17.5.2-postgres",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "docker.io/percona/percona-pgbouncer:1.24.1",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "docker.io/percona/percona-pgbackrest:2.55.0",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:3.2.0",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: docker.io/percona/percona-postgresql-operator:2.7.0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          olm.skipRange: <v2.7.0
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native

          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - docker.io/percona/percona-postgresql-operator:2.7.0
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator.v2.6.1
        version: 2.6.1
      - name: percona-postgresql-operator.v2.6.0
        version: 2.6.0
      - name: percona-postgresql-operator.v2.5.0
        version: 2.5.0
      name: stable
    defaultChannel: stable
    packageName: percona-postgresql-operator
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: VMware Tanzu
      provider-url: ""
    name: rabbitmq-messaging-topology-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-messaging-topology-operator.v1.17.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [{"apiVersion":"rabbitmq.com/v1beta1","kind":"Binding","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Exchange","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Federation","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Permission","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Policy","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"OperatorPolicy","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Queue","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"SchemaReplication","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Shovel","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1alpha1","kind":"SuperStream","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"User","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Vhost","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"TopicPermission","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}}]
          capabilities: Full Lifecycle
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/rabbitmqoperator/messaging-topology-operator:1.17.4
          createdAt: 2025-08-21T11:15:42
          description: Kubernetes operator to allow developers to create and manage
            RabbitMQ messaging topologies within a RabbitMQ cluster using a declarative
            Kubernetes API.
          repository: https://github.com/rabbitmq/messaging-topology-operator
          support: VMware Tanzu
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Binding
            displayName: Binding
            kind: Binding
            name: bindings.rabbitmq.com
            version: v1beta1
          - description: Exchange
            displayName: Exchange
            kind: Exchange
            name: exchanges.rabbitmq.com
            version: v1beta1
          - description: Federation
            displayName: Federation
            kind: Federation
            name: federations.rabbitmq.com
            version: v1beta1
          - description: Permission
            displayName: Permission
            kind: Permission
            name: permissions.rabbitmq.com
            version: v1beta1
          - description: TopicPermission
            displayName: TopicPermission
            kind: TopicPermission
            name: topicpermissions.rabbitmq.com
            version: v1beta1
          - description: Policy
            displayName: Policy
            kind: Policy
            name: policies.rabbitmq.com
            version: v1beta1
          - description: OperatorPolicy
            displayName: OperatorPolicy
            kind: OperatorPolicy
            name: operatorpolicies.rabbitmq.com
            version: v1beta1
          - description: Queue
            displayName: Queue
            kind: Queue
            name: queues.rabbitmq.com
            version: v1beta1
          - description: Schema Replication
            displayName: SchemaReplication
            kind: SchemaReplication
            name: schemareplications.rabbitmq.com
            version: v1beta1
          - description: Shovel
            displayName: Shovel
            kind: Shovel
            name: shovels.rabbitmq.com
            version: v1beta1
          - description: Super Stream
            displayName: SuperStream
            kind: SuperStream
            name: superstreams.rabbitmq.com
            version: v1alpha1
          - description: User
            displayName: User
            kind: User
            name: users.rabbitmq.com
            version: v1beta1
          - description: Vhost
            displayName: Vhost
            kind: Vhost
            name: vhosts.rabbitmq.com
            version: v1beta1
          required:
          - description: Rabbitmq Cluster
            displayName: RabbitmqCluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: |
          ## About the managed application
          RabbitMQ is an open-source message-broker software that originally implemented the Advanced Message Queuing Protocol and has since been extended with a plug-in architecture to support Streaming Text Oriented Messaging Protocol, MQ Telemetry Transport, and other protocols.
          ## About this Operator
          Kubernetes operator to allow developers to create and manage RabbitMQ messaging topologies within a RabbitMQ cluster using a declarative Kubernetes API. A Messaging topology is the collection of objects such as exchanges, queues, bindings and policies that provides specific messaging or streaming scenario. This operator is used with RabbitMQ clusters deployed via the RabbitMQ Cluster Kubernetes Operator. This repository contains custom controllers and custom resource definitions (CRDs) enabling a declarative API for RabbitMQ messaging topologies.
          ## Prerequisites for enabling this Operator
          The operator works on top of the RabbitMQ cluster operator.
          Messaging Topology Operator is tested with the latest release of RabbitMQ Cluster Operator. It uses the generated default user secret from RabbitmqCluster (set in rabbitmqcluster.status.binding) to authenticate with RabbitMQ server.
        displayName: rabbitmq-messaging-topology-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        - messaging
        - amqp
        - amqp10
        - stomp
        - mqtt
        - mqtt5
        - queues
        - streaming
        links:
        - name: GitHub
          url: https://github.com/rabbitmq/messaging-topology-operator
        - name: Documentation
          url: https://www.rabbitmq.com/kubernetes/operator/install-topology-operator.html
        maintainers:
        - email: rabbitmq-users@googlegroups.com
          name: VMware Tanzu
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        provider:
          name: VMware Tanzu
        relatedImages:
        - quay.io/rabbitmqoperator/messaging-topology-operator:1.17.4
        version: 1.17.4
      entries:
      - name: rabbitmq-messaging-topology-operator.v1.17.4
        version: 1.17.4
      - name: rabbitmq-messaging-topology-operator.v1.16.0
        version: 1.16.0
      - name: rabbitmq-messaging-topology-operator.v1.15.0
        version: 1.15.0
      - name: rabbitmq-messaging-topology-operator.v1.14.2
        version: 1.14.2
      - name: rabbitmq-messaging-topology-operator.v1.14.1
        version: 1.14.1
      - name: rabbitmq-messaging-topology-operator.v1.13.0
        version: 1.13.0
      - name: rabbitmq-messaging-topology-operator.v1.12.2
        version: 1.12.2
      - name: rabbitmq-messaging-topology-operator.v1.12.1
        version: 1.12.1
      name: stable
    defaultChannel: stable
    packageName: rabbitmq-messaging-topology-operator
    provider:
      name: VMware Tanzu
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Omar Gaye IBM
      provider-url: omarg.net
    name: memcached-operator-ogaye
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: memcached-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.omarg.net/v1alpha1",
                "kind": "Memcached",
                "metadata": {
                  "name": "memcached-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.6.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Memcached is the Schema for the memcacheds API
            displayName: Memcached
            kind: Memcached
            name: memcacheds.cache.omarg.net
            version: v1alpha1
        description: Omar memcached operator
        displayName: memcached-operator-ogaye
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Omar
        - Memcached
        links:
        - name: Memcached Operator OGaye
          url: https://memcached-operator.domain
        maturity: alpha
        provider:
          name: Omar Gaye IBM
          url: omarg.net
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9211b70edbb93d718428e1c9a992e5fc74a20cb213f8a0592ffe340c362890fe
        - quay.io/ogaye/memcached-operator@sha256:5340ac11ce676dd3204bf3807a4c2333a8dddcbcec918db0aeafbc977163c1c6
        version: 0.0.1
      entries:
      - name: memcached-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: memcached-operator-ogaye
    provider:
      name: Omar Gaye IBM
      url: omarg.net
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Aqua Security, Inc.
      provider-url: ""
    name: aqua-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aqua-operator.2022.4.836
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaCsp",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "namespace": "aqua",
                    "version": "2022.4",
                    "requirements": true
                  },
                  "common": {
                    "dbDiskSize": 10,
                    "databaseSecret": {
                      "key": "db-password",
                      "name": "aqua-database-password"
                    }
                  },
                  "database": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "gateway": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "server": {
                    "replicas": 1,
                    "service": "LoadBalancer"
                  },
                  "route": true,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaDatabase",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "splitDB": false
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "diskSize": 10,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "runAsNonRoot": false,
                  "gateway": {
                    "host": "aqua-gateway",
                    "port": 8443
                  },
                  "token": "<<your-token>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaKubeEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "config": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "aqua-secure",
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "service": "ClusterIP"
                  },
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaGateway",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  }
                },
                "runAsNonRoot": false
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaScanner",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "deploy": {
                    "replicas": 1
                  },
                  "runAsNonRoot": false,
                  "login": {
                    "username": "<<YOUR AQUA USER NAME>>",
                    "password": "<<YOUR AQUA USER PASSWORD>>",
                    "host": "http://aqua-server:8080"
                  }
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaServer",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "LoadBalancer"
                  }
                },
                "runAsNonRoot": false
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaLightning",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "global": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "Default-cluster-name"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "kubeEnforcer": {
                    "infra": {
                      "serviceAccount": "aqua-kube-enforcer-sa"
                    },
                    "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                    "deploy": {
                      "service": "ClusterIP"
                    },
                    "starboard": {
                      "infra": {
                        "serviceAccount": "aqua-kube-enforcer-sa"
                      },
                      "config": {
                        "imagePullSecret": "aqua-registry"
                      },
                      "deploy": {
                        "replicas": 1
                      }
                    },
                    "env": []
                  },
                  "enforcer": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "deploy": null,
                    "token": "<<ENFORCER_GROUP_TOKEN>>",
                    "env": [
                    ],
                    "secret": {
                      "name": null,
                      "key": null
                    },
                    "runAsNonRoot": null,
                    "rhcosVersion": null
                  }
                }
              },
            {
                  "apiVersion": "operator.aquasec.com/v1alpha1",
                  "kind": "AquaCloudConnector",
                  "metadata": {
                    "name": "aqua"
                  },
                  "spec": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "common": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    },
                    "runAsNonRoot": false,
                    "gateway": {
                      "host": "aqua-gateway",
                      "port": "8443"
                    },
                    "login": {
                      "host": "aqua-gateway:8443",
                      "password": "<<YOUR AQUA USER PASSWORD>>",
                      "token": "<<YOUR AQUA CLOUD CONNECTOR TOKEN>>",
                      "username": "<<YOUR AQUA USER NAME>>"
                    },
                    "tunnels": [
                      {
                        "host": "<account>.azurecr.io",
                        "port": "443"
                      },
                      {
                        "host": "gke.gcr.io",
                        "port": "443"
                      },
                      {
                        "host": "<account>.jfrog.io",
                        "port": "443"
                      },
                      {
                        "host": "1.1.1.1",
                        "port": "443"
                      },
                      {
                        "region": "us-east1",
                        "type": "ecr"
                      }
                    ]
                  }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/aquasec/aquasec@sha256:c9f2c770a51e9664006c12d124ca18aab76bea8c32a6a253fe31d44c02633642
          createdAt: ""
          description: The Aqua Security Operator runs within a Openshift cluster
            and provides a means to deploy and manage Aqua Security cluster and components.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2022.4.836'
          operators.operatorframework.io/internal-objects: '["aquastarboards.aquasecurity.github.io","configauditreports.aquasecurity.github.io",
            "clusterconfigauditreports.aquasecurity.github.io"]'
          repository: https://github.com/aquasecurity/aqua-operator
          support: Aqua Security, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aqua Security CSP Deployment with Aqua Operator
            displayName: AquaCsp
            kind: AquaCsp
            name: aquacsps.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Lightning Enforcers Deployment with Aqua Operator
            displayName: AquaLightningEnforcers
            kind: AquaLightning
            name: aqualightnings.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Database Deployment with Aqua Operator
            displayName: AquaDatabase
            kind: AquaDatabase
            name: aquadatabases.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Enforcer Deployment with Aqua Operator
            displayName: AquaEnforcer
            kind: AquaEnforcer
            name: aquaenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security KubeEnforcer Deployment with Aqua Operator
            displayName: AquaKubeEnforcer
            kind: AquaKubeEnforcer
            name: aquakubeenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Gateway Deployment with Aqua Operator
            displayName: AquaGateway
            kind: AquaGateway
            name: aquagateways.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Scanner Deployment with Aqua Operator
            displayName: AquaScanner
            kind: AquaScanner
            name: aquascanners.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Server Deployment with Aqua Operator
            displayName: AquaServer
            kind: AquaServer
            name: aquaservers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Starboard for kube-enforcer.
            displayName: AquaStarboard
            kind: AquaStarboard
            name: aquastarboards.aquasecurity.github.io
            version: v1alpha1
          - description: Aqua Cloud Connector Deployment with Aqua Operator
            displayName: AquaCloudConnector
            kind: AquaCloudConnector
            name: aquacloudconnectors.operator.aquasec.com
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ClusterConfigAuditReport
            kind: ClusterConfigAuditReport
            name: clusterconfigauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ConfigAuditReport
            kind: ConfigAuditReport
            name: configauditreports.aquasecurity.github.io
            version: v1alpha1
        description: |-
          The Aqua Security Operator runs within an OpenShift cluster and provides a means to deploy and manage the Aqua Security cluster and components :
          * Server (aka “console”)
          * Database (for production environments we recommend to use an external database and not the Aqua default database)
          * Gateway
          * Enforcer (aka “agent”)
          * KubeEnforcer
          * Scanner
          * CSP (package that contains the Server, Database, and Gateway)
          * LightningEnforcer (package that contains Enforcer and Kube Enforcer) - Do not deploy using Certified Operator.
          * AquaCloudConnector.

          Use the Aqua-Operator to
          * Deploy Aqua Security components on OpenShift
          * Scale up Aqua Security components with extra replicas
          * Assign metadata tags to Aqua Security components

          The Aqua Operator provides a few [Custom Resources](https://github.com/aquasecurity/aqua-operator/tree/master/config/crd) for managing the Aqua CSP platform.

          ## Prerequisites
          There are only 4 prerequisites:
          1. Verify that the OpenShift version is officially supported [Verify Here](https://docs.aquasec.com/saas/workload-protection/product-information/system-requirements/)
          2. Make sure you have a license and to obtain one, please contact Aqua Security at [Contact Us](mailto:cloudsales@aquasec.com).
          3. Create a new project for aqua

              ```oc new-project aqua```

          4. Create the secret for Aqua Database password. If you are using an External Database, make sure you use those credentials.
          NOTE: This step is optional and you can specify the Database password when creating the CRs.

              ```oc create secret generic aqua-database-password --from-literal=db-password=<password> -n aqua```

          Please note that for the certified operator, the Docker registry secret is NOT needed. For the Red Hat certified operator, the Aqua application images are stored in the Red Hat Connect registry and will be automatically pulled by the Aqua operator.

          ## Choosing the right channel

          Channels are a flexible publishing mechanism that offers you different Aqua versions to deploy, at any point. These channels map to their respective Aqua versions, describing the maturity based on where they are in their lifecycle:
          * 4.6: This maps to the older version of Aqua 4.6.0
          * 5.0: This maps to the Aqua version 5.0.0
          * 5.3: This maps to the Aqua version 5.3.0
          * 6.0: This maps to the Aqua version 6.0.0
          * 6.2: This maps to the Aqua version 6.2.0
          * 6.5: This maps to the Aqua version 6.5.0
          * 2022.4: This is the latest and greatest version of Aqua (default)

          ## Deploying the Aqua Operator

          Aqua Operator follows an easy push button deployment that installs the operator in its own aqua namespace.
          Please refer to the instructions [here](https://github.com/aquasecurity/aqua-operator/blob/2022.4.0/docs/DeployOpenShiftOperator.md).

          ## Configuring the Aqua Operator

          Operators leverage CRDs as a configuration mechanism for the application. Aqua offers a number of CRDs that can be tweaked and configured as per your needs. Please refer to this [link](https://github.com/aquasecurity/aqua-operator/blob/6.2.0/docs/DeployOpenShiftOperator.md#aquacsp-crds) for Custom Resource examples for popular scenarios.

          ## Support

          For support please contact support@aquasec.com.
        displayName: Aqua Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aqua-security
        - scanning
        - security
        - runtime-security
        links:
        - name: Aqua Security
          url: https://www.aquasec.com/
        - name: Aqua Operator Github
          url: https://github.com/aquasecurity/aqua-operator
        maintainers:
        - email: support@aquasec.com
          name: Aqua Support, Aqua Security
        maturity: alpha
        provider:
          name: Aqua Security, Inc.
        relatedImages:
        - registry.connect.redhat.com/aquasec/console@sha256:ed371053e46e89ad68ed48282ddf28abc668a613a9c32e3aa45e5111512b8466
        - registry.connect.redhat.com/aquasec/kube-bench@sha256:bbb4e8a61033b303525af6764f685e62a2f441453c3f0eaaf4cbab6997bf93ee
        - registry.connect.redhat.com/aquasec/gateway@sha256:11aa01e681986f6b5bd53b41a07db4456023080a753970d8c872e000f68f3b9b
        - registry.connect.redhat.com/aquasec/kube-enforcer@sha256:c456b65a3f1dc2de7209f3d9cbcf1bb523130021433b8fd740c41d9bd6f7db9f
        - registry.connect.redhat.com/aquasec/aquasec@sha256:c9f2c770a51e9664006c12d124ca18aab76bea8c32a6a253fe31d44c02633642
        - registry.connect.redhat.com/aquasec/enforcer@sha256:50128b750e0a094de266c9a2cb42ce06a757f441a68d206a8a9c17f7b7119066
        - registry.connect.redhat.com/aquasec/scanner@sha256:c086d735c7e6ea842ad1aaeb1eb19d3a951e6e88d397a20f2cd559c6bc042d26
        - registry.connect.redhat.com/aquasec/cloud-connector@sha256:f9da1936ac5e4b8e4537dca69f920c682fb9fac98d091680e01d2122f2b0f85f
        - registry.connect.redhat.com/aquasec/database@sha256:c91ae82b0179c3ec4e156835a9f23b66767ba15514ff33fb268b35d8f04d6fde
        - registry.connect.redhat.com/aquasec/starboard@sha256:61673f7ba1d983ff4bb4bd76036ac55e5b81696c9a3e3aa4ab1a49a27a7a2106
        version: 2022.4.836
      entries:
      - name: aqua-operator.2022.4.836
        version: 2022.4.836
      - name: aqua-operator.2022.4.829
        version: 2022.4.829
      - name: aqua-operator.2022.4.818
        version: 2022.4.818
      - name: aqua-operator.2022.4.805
        version: 2022.4.805
      - name: aqua-operator.2022.4.760
        version: 2022.4.760
      - name: aqua-operator.2022.4.759
        version: 2022.4.759
      - name: aqua-operator.2022.4.720
        version: 2022.4.720
      - name: aqua-operator.2022.4.683
        version: 2022.4.683
      - name: aqua-operator.2022.4.662
        version: 2022.4.662
      - name: aqua-operator.2022.4.588
        version: 2022.4.588
      - name: aqua-operator.2022.4.587
        version: 2022.4.587
      - name: aqua-operator.2022.4.519
        version: 2022.4.519
      - name: aqua-operator.2022.4.518
        version: 2022.4.518
      name: "2022.4"
    defaultChannel: "2022.4"
    packageName: aqua-operator-certified
    provider:
      name: Aqua Security, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: netobserv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: network-observability-operator.v1.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowMetric",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netobserv-operator",
                    "app.kubernetes.io/instance": "flowmetric-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "flowmetric",
                    "app.kubernetes.io/part-of": "netobserv-operator"
                  },
                  "name": "flowmetric-sample"
                },
                "spec": {
                  "charts": [
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "",
                          "promQL": "sum(rate($METRIC[2m]))"
                        }
                      ],
                      "title": "External ingress traffic",
                      "type": "SingleStat",
                      "unit": "Bps"
                    },
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "{{DstK8S_Namespace}} / {{DstK8S_OwnerName}}",
                          "promQL": "sum(rate($METRIC{DstK8S_Namespace!=\"\"}[2m])) by (DstK8S_Namespace, DstK8S_OwnerName)"
                        }
                      ],
                      "sectionName": "External",
                      "title": "Top external ingress traffic per workload",
                      "type": "StackArea",
                      "unit": "Bps"
                    }
                  ],
                  "direction": "Ingress",
                  "filters": [
                    {
                      "field": "SrcSubnetLabel",
                      "matchType": "Absence"
                    }
                  ],
                  "labels": [
                    "DstK8S_HostName",
                    "DstK8S_Namespace",
                    "DstK8S_OwnerName",
                    "DstK8S_OwnerType"
                  ],
                  "metricName": "cluster_external_ingress_bytes_total",
                  "type": "Counter",
                  "valueField": "Bytes"
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "app"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "infra"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "enable": true,
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "includeList": [
                        "node_ingress_bytes_total",
                        "workload_ingress_bytes_total",
                        "namespace_flows_total"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta2",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "metrics": {
                        "server": {
                          "port": 9400
                        }
                      },
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "eBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "Disabled"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "\"app\""
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "\"infra\""
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "\"Pod\"",
                          "src_kind": "\"Pod\""
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "\"Service\""
                        },
                        "name": "Services network"
                      }
                    ],
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    }
                  },
                  "deploymentModel": "Direct",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "enable": true,
                    "lokiStack": {
                      "name": "loki"
                    },
                    "mode": "Monolithic",
                    "monolithic": {
                      "tenantID": "netobserv",
                      "tls": {
                        "caCert": {
                          "certFile": "service-ca.crt",
                          "name": "loki-gateway-ca-bundle",
                          "type": "configmap"
                        },
                        "enable": false
                      },
                      "url": "http://loki.netobserv.svc:3100/"
                    },
                    "readTimeout": "30s",
                    "writeBatchSize": 10485760,
                    "writeBatchWait": "1s",
                    "writeTimeout": "10s"
                  },
                  "namespace": "netobserv",
                  "networkPolicy": {
                    "additionalNamespaces": [],
                    "enable": false
                  },
                  "processor": {
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "Flows",
                    "metrics": {
                      "disableAlerts": [],
                      "server": {
                        "port": 9401
                      }
                    },
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "prometheus": {
                    "querier": {
                      "enable": true,
                      "mode": "Auto",
                      "timeout": "30s"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking, Observability, OpenShift Optional
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:13b6dd0ddf4bdd2836b0d3ec9c79119a902456dcee100cfac6898768e1f22aab
          createdAt: 2025-09-08T21:48:09
          description: Network flows collector and monitoring solution
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: '{"apiVersion":"flows.netobserv.io/v1beta2",
            "kind":"FlowCollector","metadata":{"name":"cluster"},"spec": {}}'
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
          support: NetObserv team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta2
          - description: '`FlowMetric` is the schema for the custom metrics API, which
              allows to generate more metrics out of flow logs. You can find examples
              here: https://github.com/netobserv/network-observability-operator/tree/main/config/samples/flowmetrics'
            displayName: Flow Metric
            kind: FlowMetric
            name: flowmetrics.flows.netobserv.io
            version: v1alpha1
        description: |
          Network Observability is an OpenShift operator that deploys a monitoring pipeline consisting in:
          - an eBPF agent, that generates network flows from captured packets
          - flowlogs-pipeline, a component that collects, enriches and exports these flows
          - a Console plugin for flows visualization with powerful filtering options, a topology representation and more

          Flow data is then available in multiple ways, each optional:

          - As Cluster Monitoring metrics
          - As raw flow logs stored in Grafana Loki
          - As raw flow logs exported to a collector

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, can optionally be used as the backend to store all collected flows. The Network Observability operator does not install Loki directly, however we provide some guidance to help you there.

          - For a production or production-like environment usage, refer to [the operator documentation](https://docs.openshift.com/container-platform/latest/network_observability/installing-operators.html).

          - For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 10GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          oc create namespace netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/1-storage.yaml) -n netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          If you prefer to not use Loki, you must set `spec.loki.enable` to `false` in `FlowCollector`.
          In that case, you can still get the Prometheus metrics or export raw flows to a custom collector. But be aware that some of the Console plugin features will be disabled. For instance, you will not be able to view raw flows there, and the metrics / topology will have a more limited level of details, missing information such as pods or IPs.

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for example [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/config/samples/flows_v1beta2_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          oc edit flowcollector cluster
          ```

          As it operates cluster-wide on every node, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you might have to configure differently if you used another installation method. Make sure to disable it (`spec.loki.enable`) if you don't want to use Loki.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`) an optional list of exporters to which to send enriched flows. KAFKA and IPFIX exporters are supported. This allows you to define any custom storage or processing that can read from Kafka or use the IPFIX standard.

          - To enable availability zones awareness, set `spec.processor.addZone` to `true`.

          ## Resource considerations

          The following table outlines examples of resource considerations for clusters with certain workload sizes.
          The examples outlined in the table demonstrate scenarios that are tailored to specific workloads. Consider each example only as a baseline from which adjustments can be made to accommodate your workload needs.


          | Resource recommendations                        | Extra small (10 nodes) | Small (25 nodes)       | Medium (65 nodes) **    | Large (120 nodes) **          |
          | ----------------------------------------------- | ---------------------- | ---------------------- | ----------------------- | ----------------------------- |
          | *Worker Node vCPU and memory*                   | 4 vCPUs\| 16GiB mem *  | 16 vCPUs\| 64GiB mem * | 16 vCPUs\| 64GiB mem  * |16 vCPUs\| 64GiB Mem *         |
          | *LokiStack size*                                | `1x.extra-small`       | `1x.small`             | `1x.small`              | `1x.medium`                   |
          | *Network Observability controller memory limit* | 400Mi (default)        | 400Mi (default)        | 400Mi (default)         | 800Mi                         |
          | *eBPF sampling rate*                            | 50 (default)           | 50 (default)           | 50 (default)            | 50 (default)                  |
          | *eBPF memory limit*                             | 800Mi (default)        | 800Mi (default)        | 2000Mi                  | 800Mi (default)               |
          | *FLP memory limit*                              | 800Mi (default)        | 800Mi (default)        | 800Mi (default)         | 800Mi (default)               |
          | *FLP Kafka partitions*                          | N/A                    | 48                     | 48                      | 48                            |
          | *Kafka consumer replicas*                       | N/A                    | 24                     | 24                      | 24                            |
          | *Kafka brokers*                                 | N/A                    | 3 (default)            | 3 (default)             | 3 (default)                   |

          *. Tested with AWS M6i instances.
          **. In addition to this worker and its controller, 3 infra nodes (size `M6i.12xlarge`) and 1 workload node (size `M6i.8xlarge`) were tested.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - More information on [configuring metrics](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/Metrics.md).
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: Network Observability
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/network-observability/network-observability-console-plugin-compat-rhel9@sha256:d130f94cdd3cc863a31c59f395a6c4f663e8e0da23747304353c6674b20d81be
        - registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:13b6dd0ddf4bdd2836b0d3ec9c79119a902456dcee100cfac6898768e1f22aab
        - registry.redhat.io/network-observability/network-observability-ebpf-agent-rhel9@sha256:0b4337f4db1e74a2bea1021a617cb42690251036f583bb771cc5ec888f38c97f
        - registry.redhat.io/network-observability/network-observability-flowlogs-pipeline-rhel9@sha256:09454618709a5c83beb4252f1c14c149716d1c09b6567c0a571d5390e8a639a2
        - registry.redhat.io/network-observability/network-observability-console-plugin-rhel9@sha256:7d92eadb3cd08a08672a65a196a5d3181e764c68e8c20cab2554933cdb10fd0c
        version: 1.9.3
      entries:
      - name: network-observability-operator.v1.9.3
        version: 1.9.3
      - name: network-observability-operator.v1.9.2
        version: 1.9.2
      - name: network-observability-operator.v1.9.1
        version: 1.9.1
      - name: network-observability-operator.v1.9.0
        version: 1.9.0
      - name: network-observability-operator.v1.8.1
        version: 1.8.1
      - name: network-observability-operator.v1.8.0
        version: 1.8.0
      - name: network-observability-operator.v1.7.0
        version: 1.7.0
      - name: network-observability-operator.v1.6.2
        version: 1.6.2
      - name: network-observability-operator.v1.6.1
        version: 1.6.1
      - name: network-observability-operator.v1.6.0
        version: 1.6.0
      - name: network-observability-operator.v1.5.0
        version: 1.5.0
      - name: network-observability-operator.v1.4.2
        version: 1.4.2
      - name: network-observability-operator.v1.4.1
        version: 1.4.1
      - name: network-observability-operator.v1.4.0
        version: 1.4.0
      - name: network-observability-operator.v1.3.0
        version: 1.3.0
      - name: network-observability-operator.v1.2.0
        version: 1.2.0
      - name: network-observability-operator.v1.1.0-202302110050
        version: 1.1.0
      - name: network-observability-operator.v1.0.0-202301070345
        version: 1.0.0
      name: stable
    - currentCSV: network-observability-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:fd943746840cba26996dabb47b1beb3445571fa43cb1b859fb073e4aa114b376
          createdAt: 2023-06-22T09:31:50
          description: Network flows collector and monitoring solution
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
          - description: FlowCollector is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
        description: |
          Network Observability is an OpenShift operator that deploys a monitoring pipeline to collect and enrich network flows that are produced by the Network Observability eBPF agent.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When a `FlowCollector` instance is created, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For normal usage, we recommend two options:

          - Installing the [Loki Operator](https://docs.openshift.com/container-platform/4.11//logging/cluster-logging-loki.html). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          oc create namespace netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.0.2/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.0.2/config/samples/flows_v1beta1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          oc edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.0.2/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: Network Observability
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:fd943746840cba26996dabb47b1beb3445571fa43cb1b859fb073e4aa114b376
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:37b3c9b2eed351d9d1a39c99045bbc7993ccb819f67f7b390deaa34930f45946
        - registry.redhat.io/network-observability/network-observability-console-plugin-rhel9@sha256:880c0812a2cd4b7194f417e31fbdcee951ce76776fa06e33c5c9cd479a854daa
        - registry.redhat.io/network-observability/network-observability-ebpf-agent-rhel9@sha256:e31acc12d8690c339e13b5dd22d33e707701362bdd31086c9b4c0f48cad62cc3
        - registry.redhat.io/network-observability/network-observability-flowlogs-pipeline-rhel9@sha256:7c7c42fff63e0599bb943a9e9617b43181360d5a5254b27f368010af275cf0c3
        version: 1.3.0
      entries:
      - name: network-observability-operator.v1.3.0
        version: 1.3.0
      - name: network-observability-operator.v1.2.0
        version: 1.2.0
      - name: network-observability-operator.v1.1.0-202302110050
        version: 1.1.0
      - name: network-observability-operator.v1.0.0-202301070345
        version: 1.0.0
      name: v1.0.x
    defaultChannel: stable
    packageName: netobserv-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: openshift-builds-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-builds-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "openshift-builds"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-rhel8-operator@sha256:7100a83aa053a976ceb77ace79f090ec8601bc9b2152ff6f7d00f0b1deea7f0b
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n\n## Usage\n\nTo deploy and manage [builds
          for Red Hat OpenShift](https://github.com/shipwright-io/build) in your cluster,\nfirst
          make sure this operator is installed and running on your cluster.\n\nNext,
          create the following:\n\n```yaml\n---\napiVersion: operator.shipwright.io/v1alpha1\nkind:
          ShipwrightBuild\nmetadata:\n  name: openshift-builds-operator\nspec:\n  targetNamespace:
          openshift-builds\n```\n\nThe operator will deploy Shipwright Build in the
          provided `targetNamespace`.\nWhen `.spec.targetNamespace` is not set, the
          namespace will default to ` openshift-builds`.\nRefer to the [ShipwrightBuild
          operator documentation](https://github.com/shipwright-io/operator/blob/main/docs/shipwrightbuild.md)
          for more information about this custom resource.\n"
        displayName: builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://shipwright.io
        - name: Shipwright Builds
          url: https://github.com/shipwright-io/build
        maintainers:
        - email: pipeline-integrations-team@redhat.com
          name: Red Hat Pipelines Integration Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-rhel8-operator@sha256:7100a83aa053a976ceb77ace79f090ec8601bc9b2152ff6f7d00f0b1deea7f0b
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel8@sha256:73efcb6ad82f698d7463718b06c86f2b2b3b5326689a62dc39c9f5d68f8943f2
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel8@sha256:0cf079a24b8b3a4b3e1f30710fea85039d9cf323cfedd06e7d1e33e6336fb848
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel8@sha256:22930726a25b11de2571da0395b53352939d142a16f4db6fd5460abd4cc37ba5
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel8@sha256:fd903804672ec08acf9a2b49aa7ecc5f9b543c12f7462637ef756733171b3817
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel8@sha256:f95348e5fc48676d93ebc919af2c27a041dfd632ec285296246d886daab11576
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel8@sha256:d01c1749ce76dd4bf5fefaa6c553725d33533d6e70e884e916f12217b674c3cf
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:97cade2c1ee468261aec5400728c8d44de387b459134aec7a4c3b5ec5a335d2c
        version: 1.0.2
      entries:
      - name: openshift-builds-operator.v1.0.2
        version: 1.0.2
      - name: openshift-builds-operator.v1.0.1
        version: 1.0.1
      - name: openshift-builds-operator.v1.0.0
        version: 1.0.0
      name: builds-1.0
    - currentCSV: openshift-builds-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          createdAt: "2025-01-23T14:48:43Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.openshift.com/builds/latest/about/overview-openshift-builds.html
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:778ad1946bdad68e1430a2fb366cee25cd98d87e44cf5ad34b87603c2f4171d0
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:e8582857566020c71653228355cbadc2c1cda41e36fc767c204a4edb05e7f42a
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:fa039b62ebd1fc6e86db8ec1c7ae5fe0dd3a6f6f642b63b4dc89af32a78cea6b
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:1ec9a1dcfc08b707df48175bba76f8a6cbfcc8396d9e9940cc3d7100e6731701
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:0e139db4370c37e91ef7eb3e0cf9a2587cb81f37709009a37f2724d4461ed2f3
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:949a1a5150642167b6c7d2b8f6d92a2885afe453319c02495f351b2e79eff3dd
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:7cfffdd43fdd33a20d115f35c2ba492da3c1d0e3ab970587bc888c68c47dfacf
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:97cade2c1ee468261aec5400728c8d44de387b459134aec7a4c3b5ec5a335d2c
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:e9f7b8741f265a332008424a2c5b3dae9ea154e623c469085c2abafb53fc62d4
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:5babdf590730b217deb59acbccf142c0500bc692542131810714cc00724ee02c
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:5ec474ab9e62c92e44c1a6635913032aead6246faec0895dcfbdb28ab0dc8a1a
        version: 1.1.1
      entries:
      - name: openshift-builds-operator.v1.1.1
        version: 1.1.1
      - name: openshift-builds-operator.v1.1.0
        version: 1.1.0
      name: builds-1.1
    - currentCSV: openshift-builds-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          createdAt: "2025-01-23T13:47:38Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.2
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:ee2792972dc52d81814603e26103e0d1117194db5ab3b5577736b0fc3285a8ab
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3fa22124916523b958c67af8ad652e73a2c3d68bb5579da1cba1ade537f3b7ae
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:8ace52981058ed51cec5b893e6d773de963b12ab715258ac1f2cbc6430ad52e7
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:775c15b93b19ba855791d34c4a95ef203dd9f8e51110c67ebf131e38eda49422
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:2b288e36e3606cf8a0d7723772edb8d4ac11772ec450539ef3448ad517155ecb
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:4bee8594d17172e5b9893a865a9922a1c7869027633ad64b3f684ec44755bac6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3fcd8e2bf0bcb8ff8c93a87af2c59a3bcae7be8792f9d3236c9b5bbd9b6db3b2
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:e8582857566020c71653228355cbadc2c1cda41e36fc767c204a4edb05e7f42a
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:78ee35ccc70b78638c667152b7b6247417c0aa1bb6d3f817d59b22eb179899ca
        - registry.redhat.io/ubi8/buildah@sha256:56c71ecd450f284325ba8434cbf5b3006078b11f0625c6258689699fa811c95d
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:ebfe4f8a97f6f0f9afe38953ae8cc2fdf2ec0b76401c819bb12c1df7cd671e9e
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:016c93578908f1dbdc3052333a1d5a9ab0c03104070ac36a6ddf99350d7510f4
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:87c4c0e9fe4d9baf8ae4f191c51d90c46729273e182fb5b3dfd951632e7bfad3
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:a91c2bb1239ef7fbce8c1ab9308da03b533df9cea75c7d978fd76f38087bc912
        version: 1.2.2
      entries:
      - name: openshift-builds-operator.v1.2.2
        version: 1.2.2
      - name: openshift-builds-operator.v1.2.1
        version: 1.2.1
      - name: openshift-builds-operator.v1.2.0
        version: 1.2.0
      name: builds-1.2
    - currentCSV: openshift-builds-operator.v1.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-01-29T08:58:54Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.2
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:6e44758fd4a7b34814ac7c71e52d93633781d39f7897ecab98a090859b487b4f
        - registry.redhat.io/ubi9/buildah@sha256:27d837f9bc69ad3c3651cf3315e2501b137c11baa553d9d46140e5cf7fa7873a
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:0783132b14ed8050664810f37e7ad6388fefbc586c45ecc7ba69dd5041a55c9f
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:e8582857566020c71653228355cbadc2c1cda41e36fc767c204a4edb05e7f42a
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:04b8bfc0dc6f260ef367ca1cce2bcc84032bc8be17b4a595be12a344813a2632
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:1e9a54212690f7930d2b681db87a6f79f5a2acc5c982734d4b283a2472b59db5
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:3523e7c2215e156081e5f05443884d65efad9b7b9ca81a35416695020231916a
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:d4b3f451becb7a42341c98dd5b9fc5d2b582c93cac2728070ef4bb83b64abb4c
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:a4ba591c57bbc815c93023f4a080213be32581ba6a1c2d94b0ecbc5c90b07a1e
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:0b1dd35f55181f34775188bc71b3ff8dff66c70acfcfabad19353a02e19351f7
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:6214708f3b517cb901af9f9bd0502ba54f8e5183d0d909a3b559deee6394026e
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:ed3fdd1e85f5c5434de6ec0feaec8c7fe4e680ca6c64258287f18ec2886bb71f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3fcd8e2bf0bcb8ff8c93a87af2c59a3bcae7be8792f9d3236c9b5bbd9b6db3b2
        version: 1.3.2
      entries:
      - name: openshift-builds-operator.v1.3.2
        version: 1.3.2
      - name: openshift-builds-operator.v1.3.1
        version: 1.3.1
      - name: openshift-builds-operator.v1.3.0
        version: 1.3.0
      name: builds-1.3
    - currentCSV: openshift-builds-operator.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2024-11-15T22:28:01Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:9d12e6cbf26e31a1e4458c99a972ed36125d759c5ec57773f203cd8dd0985de7
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:ed3fdd1e85f5c5434de6ec0feaec8c7fe4e680ca6c64258287f18ec2886bb71f
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:538416bc1384eb987e8a7e80b57eda3feddaf4260b196709a6e9a9125e3af39e
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:1a6ca51c2c68950af605ff367356ab6c0ef7d1b2fd6fb8f00aa665196ff7af84
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:f4b935c16dcbbc0393733437747497c1d24f2cc782c40c49738e5f37c037b3e7
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:c60ca44275297d2ee8c499d9d7fb4056d104f4bd5dec8a02e7388241ac466c44
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:1c7c24111f1709d7648156873ced09462638aa4b3cf7e7aee13297e51bcdfb3e
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:6251cb303774e29d137a37d6527875c371fe1d3f6a5fa0383e80f0e9ef5c02be
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:f87c2069105121927fb0c0a2e37f8f72bc6cb5e035fa8d22f7b41ce0e86c0631
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:4e9c152023b2db04ee5d55f102490294b393bf78baa3aa8917f85f8547fbe3c8
        - registry.redhat.io/ubi9/buildah@sha256:6886997da8c92214e99e3e8da47a0f08239172d2544b4edae058d5cb018effab
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:adc4a31d63b7044db7094fddcb1a4b235d848e842257f5baf22b6ead03fd6c91
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:5e87ba403f9e9fff3002a8275aba0d6c36aed54fb4754d90fa0516ad989f187a
        version: 1.4.1
      entries:
      - name: openshift-builds-operator.v1.4.1
        version: 1.4.1
      - name: openshift-builds-operator.v1.4.0
        version: 1.4.0
      name: builds-1.4
    - currentCSV: openshift-builds-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-08-14T12:12:15Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              OpenShiftBuild describes the desired state of Builds for OpenShift, and the status of
              all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I), Buildah, and Buildpacks. You can create and apply
          build resources, view logs of build runs, \nand manage builds in your OpenShift
          Container Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:38560dea7627632d8b24fa670c1d7ac1218cf86975329908786a460f76de558d
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ce7050192d9850ed4c35cb3295ccdeec1aa593e27e3a3da6d6aca97cde782abe
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c1a84feb88d53e93bdcf2cd5f76e2cbb18541c8b0e2979c132a888d6c280b664
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:ddcfcebd5d788d11f756939f07215b2e3507a1d8fe704a1052e15bb4ebe592e8
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:753e24a1cd57ca22552447363aed48f064b525b49e52f774f3e25f41d8dcb329
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:e2a7a42cd5916c05d95d3fb0decc34d0840420532270422aa7f6b7bdd94de1f2
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:bcae84267f582278591d93aecf94c8d7fb1376521d9b5208201c959a0fe56774
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:f4a75c115b5dc8853845f666f3bd063b2736e01e428a605fb5cd0f02736ecce8
        - registry.redhat.io/ubi9/buildah@sha256:8ea2b81b32f70b9502d88722a5f69de8c0cf5efa6f30df041873546e9f9438cb
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:4dc123fa703006662efcc5569a843a043bb31cf69b6868d6a9b6321c66bfca18
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:1d5a0bfe0013ab7602efe95be8a814be1e5586728fdf0e198c97739cc12f1d87
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:504179e8ee9f88a23f1d4c28c7683ab7395f4c1c4868765bf3c1406669fd52fc
        version: 1.5.1
      entries:
      - name: openshift-builds-operator.v1.5.1
        version: 1.5.1
      - name: openshift-builds-operator.v1.5.0
        version: 1.5.0
      name: builds-1.5
    - currentCSV: openshift-builds-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-09-23T13:38:59Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I), Buildah, and Buildpacks. You can create and apply
          build resources, view logs of build runs, \nand manage builds in your OpenShift
          Container Platform namespaces.\n\n## Prerequisites\n\n* OpenShift Pipelines
          operator must be installed before installing this operator\n\nRead more:
          [https://shipwright.io](https://shipwright.io)\n\n## Features\n\n* Standard
          Kubernetes-native API for building container images from source code and
          Dockerfile\n\n* Support for Source-to-Image (S2I) and Buildah build strategies\n\n*
          Extensibility with your own custom build strategies\n\n* Execution of builds
          from source code in a local directory\n\n* Shipwright CLI for creating and
          viewing logs, and managing builds on the cluster\n\n* Integrated user experience
          with the Developer perspective of the OpenShift Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:c6805e85ad4dc29f3a0582a1fbf99647a90c1af0b3924647839ad881cbb199fa
        - registry.redhat.io/ubi9/buildah@sha256:8ea2b81b32f70b9502d88722a5f69de8c0cf5efa6f30df041873546e9f9438cb
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:f9f43795c7ed1e416b92497cf2fe647beb5e45d755f6a6d89c3915b7a9c76570
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:1a61d4a8f69fa845ee678dd96f934685821828e7575852f97b6e7b1dd79a8297
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c1a84feb88d53e93bdcf2cd5f76e2cbb18541c8b0e2979c132a888d6c280b664
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:3d2e501abe1aa757e4e6ba03fe8e7a45bbfdb4303360ed7130c6344051887ed8
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:9f38cd42591b6a6b65e78a3cc5c165f8498ad3c79460a6c47acfb8744d2dd883
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:0ea2b0153ae07ee15972abbbe034ef7feb0dcaa403ac993afcb171906d9cc796
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:23e3f29ae63546db018701abb29b444b9b825fb3a2bda5084542649d9827d4ba
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:dd96e8c857bc52097d4d194db8fa2cb368bbdcbb6594722660d4e906d0bdc420
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ce7050192d9850ed4c35cb3295ccdeec1aa593e27e3a3da6d6aca97cde782abe
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:fd7c805c6d0638f2814503c6779719279119c7268b1eb0eda9b7b58ac775efa6
        version: 1.6.0
      entries:
      - name: openshift-builds-operator.v1.6.0
        version: 1.6.0
      name: builds-1.6
    - currentCSV: openshift-builds-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-09-23T13:38:59Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I), Buildah, and Buildpacks. You can create and apply
          build resources, view logs of build runs, \nand manage builds in your OpenShift
          Container Platform namespaces.\n\n## Prerequisites\n\n* OpenShift Pipelines
          operator must be installed before installing this operator\n\nRead more:
          [https://shipwright.io](https://shipwright.io)\n\n## Features\n\n* Standard
          Kubernetes-native API for building container images from source code and
          Dockerfile\n\n* Support for Source-to-Image (S2I) and Buildah build strategies\n\n*
          Extensibility with your own custom build strategies\n\n* Execution of builds
          from source code in a local directory\n\n* Shipwright CLI for creating and
          viewing logs, and managing builds on the cluster\n\n* Integrated user experience
          with the Developer perspective of the OpenShift Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:0ea2b0153ae07ee15972abbbe034ef7feb0dcaa403ac993afcb171906d9cc796
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:c6805e85ad4dc29f3a0582a1fbf99647a90c1af0b3924647839ad881cbb199fa
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:3d2e501abe1aa757e4e6ba03fe8e7a45bbfdb4303360ed7130c6344051887ed8
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:fd7c805c6d0638f2814503c6779719279119c7268b1eb0eda9b7b58ac775efa6
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:23e3f29ae63546db018701abb29b444b9b825fb3a2bda5084542649d9827d4ba
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ce7050192d9850ed4c35cb3295ccdeec1aa593e27e3a3da6d6aca97cde782abe
        - registry.redhat.io/ubi9/buildah@sha256:8ea2b81b32f70b9502d88722a5f69de8c0cf5efa6f30df041873546e9f9438cb
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:f9f43795c7ed1e416b92497cf2fe647beb5e45d755f6a6d89c3915b7a9c76570
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:1a61d4a8f69fa845ee678dd96f934685821828e7575852f97b6e7b1dd79a8297
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:9f38cd42591b6a6b65e78a3cc5c165f8498ad3c79460a6c47acfb8744d2dd883
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c1a84feb88d53e93bdcf2cd5f76e2cbb18541c8b0e2979c132a888d6c280b664
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:dd96e8c857bc52097d4d194db8fa2cb368bbdcbb6594722660d4e906d0bdc420
        version: 1.6.0
      entries:
      - name: openshift-builds-operator.v1.6.0
        version: 1.6.0
      - name: openshift-builds-operator.v1.5.1
        version: 1.5.1
      - name: openshift-builds-operator.v1.5.0
        version: 1.5.0
      - name: openshift-builds-operator.v1.4.1
        version: 1.4.1
      - name: openshift-builds-operator.v1.4.0
        version: 1.4.0
      - name: openshift-builds-operator.v1.3.2
        version: 1.3.2
      - name: openshift-builds-operator.v1.3.1
        version: 1.3.1
      - name: openshift-builds-operator.v1.3.0
        version: 1.3.0
      - name: openshift-builds-operator.v1.2.2
        version: 1.2.2
      - name: openshift-builds-operator.v1.2.1
        version: 1.2.1
      - name: openshift-builds-operator.v1.2.0
        version: 1.2.0
      - name: openshift-builds-operator.v1.1.1
        version: 1.1.1
      - name: openshift-builds-operator.v1.1.0
        version: 1.1.0
      - name: openshift-builds-operator.v1.0.2
        version: 1.0.2
      - name: openshift-builds-operator.v1.0.1
        version: 1.0.1
      - name: openshift-builds-operator.v1.0.0
        version: 1.0.0
      name: latest
    defaultChannel: latest
    packageName: openshift-builds-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: kaoto.io
      provider-url: https://kaoto.io
    name: kaoto-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kaoto-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "designer.kaoto.io/v1alpha1",
                "kind": "Kaoto",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "kaoto-operator",
                    "app.kubernetes.io/instance": "kaoto-demo",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kaoto",
                    "app.kubernetes.io/part-of": "kaoto-operator"
                  },
                  "name": "kaoto-demo"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          containerImage: quay.io/kaotoio/kaoto-operator@sha256:590d4958f148954ea1edb9825fb7fb4935d041ddcdb4a99302da14a605380071
          createdAt: "2025-09-03T12:48:58Z"
          description: Integration Editor for Apache Camel
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/KaotoIO/kaoto-operator
          support: contact@kaoto.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kaoto is the Schema for the kaotoes API.
            displayName: Kaoto
            kind: Kaoto
            name: kaotoes.designer.kaoto.io
            version: v1alpha1
        description: Kaoto is a low code and no code integration designer to edit
          and deploy based on Apache Camel . It is designed to be extendable and flexible
          and adaptable to different usecases.
        displayName: Kaoto Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - kaoto
        - camel-k
        - designer
        links:
        - name: Kaoto Operator
          url: https://kaoto-operator.domain
        maintainers:
        - email: info@kaoto.io
          name: kaoto
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: kaoto.io
          url: https://kaoto.io
        relatedImages:
        - quay.io/kaotoio/kaoto-operator@sha256:590d4958f148954ea1edb9825fb7fb4935d041ddcdb4a99302da14a605380071
        version: 2.7.0
      entries:
      - name: kaoto-operator.v2.7.0
        version: 2.7.0
      - name: kaoto-operator.v2.6.0
        version: 2.6.0
      - name: kaoto-operator.v2.5.0
        version: 2.5.0
      - name: kaoto-operator.v2.4.0
        version: 2.4.0
      - name: kaoto-operator.v2.3.0
        version: 2.3.0
      - name: kaoto-operator.v2.1.0
        version: 2.1.0
      - name: kaoto-operator.v2.0.1
        version: 2.0.1
      - name: kaoto-operator.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: kaoto-operator
    provider:
      name: kaoto.io
      url: https://kaoto.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/securesign/secure-sign-operator
    name: rhtas-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhtas-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": {
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname-intermediate",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname-leaf",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname-root",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname-intermediate",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname-leaf",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname-root",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": {
                  "database": {
                    "create": true
                  },
                  "server": {
                    "replicas": 1
                  },
                  "signer": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:0ebc6b3bd992309a914cc0da8ad8fc628224f583b66f1ca2cd10d2281b781cde
          createdAt: "2025-09-30T11:33:48Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:e7f5a9a2ea76fdfde34f6aac264b51c596cf4ab53f6922b58514c77a1977677e
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:f2798c21dbcb3334b256bf1370a42c013f8f3eefb0641e1c2f250428a82ddb82
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:2dd696771c0188a119d746711551b1643a14a579d0fc66d70918cca9de0c0107
        - registry.redhat.io/rhtas/createtree-rhel9@sha256:244758721aa1c63adfffbae31df58059ff8a06311f6cb60515eeae14b20e78ce
        - registry.redhat.io/rhtas/rekor-monitor-rhel9@sha256:da3aa5c441653f00c4b558ef6ebf21fef518731e70f152a1fd3a750b1145d901
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:b311fb891203fef071d53c9dab67a45961638af98bdfd8e02026dc2dbc25fd00
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:8f8a5d4abf7878ba34151f8632aa7db25218300dbb9f17b121c5f61e0e5c5e03
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:eaae082559c81c5e1253f1b5cf9e298b14235470c67e9111c18ece46078228d5
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:2d91dceb216d8cb8dacefcfe86f357f4ff4eb5e5e998f264d1ad231821a92fad
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:043103f7c888d150053b873276b92b2ee1907c8a7238652ad0770e36d44c10e2
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:0ebc6b3bd992309a914cc0da8ad8fc628224f583b66f1ca2cd10d2281b781cde
        - registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:08302a0af012c1a1f04ef15de86558c9eef55620ab7b179fbda170025f0584be
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:497203991dcbca71275d691ec970ba480cb8e80c9ada6af876f5559abe0d6155
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:110509f5f24240d79afff7cb0d8a4ead0ed038c6100df8e435b01f270b7d83b2
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:abaa247777eb05bf8c97ba1ffc30c00d0d84590fac2fdfbffa06ba374b924888
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:fcb225019f865dd4b7af1f5134a9aed9c5c8a425f3ee43b74e0517ce602e209c
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:e6a5bf7804f1c4371e4de06805d6e1ddfc98e356e355ac3b55a3be25c5636b4d
        version: 1.3.0
      entries:
      - name: rhtas-operator.v1.3.0
        version: 1.3.0
      - name: rhtas-operator.v1.2.1
        version: 1.2.1
      - name: rhtas-operator.v1.2.0
        version: 1.2.0
      - name: rhtas-operator.v1.1.2
        version: 1.1.2
      - name: rhtas-operator.v1.1.1
        version: 1.1.1
      - name: rhtas-operator.v1.1.0
        version: 1.1.0
      - name: rhtas-operator.v1.0.2
        version: 1.0.2
      - name: rhtas-operator.v1.0.1
        version: 1.0.1
      - name: rhtas-operator.v1.0.0
        version: 1.0.0
      name: stable
    - currentCSV: rhtas-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e1a78636cd7cc8e7227b6b0a5ddfd93d002db0464b15e966e88337d38989ea5a
          createdAt: "2024-06-27T09:45:39Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e1a78636cd7cc8e7227b6b0a5ddfd93d002db0464b15e966e88337d38989ea5a
        version: 1.0.2
      entries:
      - name: rhtas-operator.v1.0.2
        version: 1.0.2
      - name: rhtas-operator.v1.0.1
        version: 1.0.1
      - name: rhtas-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: rhtas-operator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e52d07e2cd8cbdc2ccd0336a4fd29cbb2e5b8a1be628350294a54579cc2636ed
          createdAt: "2025-04-08T09:29:59Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e52d07e2cd8cbdc2ccd0336a4fd29cbb2e5b8a1be628350294a54579cc2636ed
        version: 1.1.2
      entries:
      - name: rhtas-operator.v1.1.2
        version: 1.1.2
      - name: rhtas-operator.v1.1.1
        version: 1.1.1
      - name: rhtas-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: rhtas-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
          createdAt: "2025-08-17T21:52:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:02fbd4772c185edf38fcee4e5c5e1c9dad82b77c9da824addaebc445b67d35ac
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
        - registry.redhat.io/ubi9/httpd-24@sha256:badeee0aa3477e61de64f6a2b93e8a3673edd38fc860125f7536506f3d592132
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:c22cdb9a17f85e13c00412cba83aa644e90a0faea0236029d0f1c53fe212e227
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:12b438ea70f83ee6edb12c96b507f92510f4afba824eac34c8f50793c4199db6
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:fc018a45b7eb48690c46c83f52cac55fa40089e6b02686bc21265dc7c2205b8f
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:1d782a127d2f8bdab633f287cb2df2e400ae91bf5d4b913473539d8512dfaab6
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:09b4aeeb607c88c72f69e6f87cb840c82ef5752c64a58fa13551db749fac2530
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:eb38e98dbb9828fe033a5388609132f7246fe77d2f4d258d015a94ea30752b24
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:6270ba29adf779de5d3916d5e82f1cd9491ba89d4ae2042347b6ad6d1819dd26
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:1c2201d50469d70ec6e21546c7c74bd52251d420dc1dcfa5375c1cf61dd3a9fd
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:030986ea26f33db3a192c67a93cd2bde352f23c68a17b062fc062955675c3c51
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:531603563c69aac6905cd6b8157d30781040527404ac2313acba5a9310e0b713
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:fe31830df2d99359749cb86cd1ac2f32fcc3de07aae427d2f2cd3998d31d2833
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:a93df3206409de75613d398e7dcd8b30ff7cc36eb00349c435784f73f00a2335
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:640633823a54c11fe2a8b3d05e571399fda415b6f0b3adf1ba9806882a94bdd4
        version: 1.2.1
      entries:
      - name: rhtas-operator.v1.2.1
        version: 1.2.1
      - name: rhtas-operator.v1.2.0
        version: 1.2.0
      name: stable-v1.2
    - currentCSV: rhtas-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": {
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname-intermediate",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname-leaf",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname-root",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname-intermediate",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname-leaf",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname-root",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": {
                  "database": {
                    "create": true
                  },
                  "server": {
                    "replicas": 1
                  },
                  "signer": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:0ebc6b3bd992309a914cc0da8ad8fc628224f583b66f1ca2cd10d2281b781cde
          createdAt: "2025-09-30T11:33:48Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:abaa247777eb05bf8c97ba1ffc30c00d0d84590fac2fdfbffa06ba374b924888
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:eaae082559c81c5e1253f1b5cf9e298b14235470c67e9111c18ece46078228d5
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:2d91dceb216d8cb8dacefcfe86f357f4ff4eb5e5e998f264d1ad231821a92fad
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:e6a5bf7804f1c4371e4de06805d6e1ddfc98e356e355ac3b55a3be25c5636b4d
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:8f8a5d4abf7878ba34151f8632aa7db25218300dbb9f17b121c5f61e0e5c5e03
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:2dd696771c0188a119d746711551b1643a14a579d0fc66d70918cca9de0c0107
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:110509f5f24240d79afff7cb0d8a4ead0ed038c6100df8e435b01f270b7d83b2
        - registry.redhat.io/rhtas/createtree-rhel9@sha256:244758721aa1c63adfffbae31df58059ff8a06311f6cb60515eeae14b20e78ce
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:b311fb891203fef071d53c9dab67a45961638af98bdfd8e02026dc2dbc25fd00
        - registry.redhat.io/rhtas/rekor-monitor-rhel9@sha256:da3aa5c441653f00c4b558ef6ebf21fef518731e70f152a1fd3a750b1145d901
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:08302a0af012c1a1f04ef15de86558c9eef55620ab7b179fbda170025f0584be
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:043103f7c888d150053b873276b92b2ee1907c8a7238652ad0770e36d44c10e2
        - registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:e7f5a9a2ea76fdfde34f6aac264b51c596cf4ab53f6922b58514c77a1977677e
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:fcb225019f865dd4b7af1f5134a9aed9c5c8a425f3ee43b74e0517ce602e209c
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:0ebc6b3bd992309a914cc0da8ad8fc628224f583b66f1ca2cd10d2281b781cde
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:497203991dcbca71275d691ec970ba480cb8e80c9ada6af876f5559abe0d6155
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:f2798c21dbcb3334b256bf1370a42c013f8f3eefb0641e1c2f250428a82ddb82
        version: 1.3.0
      entries:
      - name: rhtas-operator.v1.3.0
        version: 1.3.0
      name: stable-v1.3
    defaultChannel: stable
    packageName: rhtas-operator
    provider:
      name: Red Hat
      url: https://github.com/securesign/secure-sign-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-lambda-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-lambda-controller.v1.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "Alias",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "CodeSigningConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "EventSourceMapping",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/lambda-controller:1.9.1
          createdAt: "2025-09-19T19:23:07Z"
          description: AWS Lambda controller is a service controller for managing
            Lambda resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alias represents the state of an AWS lambda Alias resource.
            displayName: Alias
            kind: Alias
            name: aliases.lambda.services.k8s.aws
            version: v1alpha1
          - description: CodeSigningConfig represents the state of an AWS lambda CodeSigningConfig
              resource.
            displayName: CodeSigningConfig
            kind: CodeSigningConfig
            name: codesigningconfigs.lambda.services.k8s.aws
            version: v1alpha1
          - description: EventSourceMapping represents the state of an AWS lambda
              EventSourceMapping resource.
            displayName: EventSourceMapping
            kind: EventSourceMapping
            name: eventsourcemappings.lambda.services.k8s.aws
            version: v1alpha1
          - description: Function represents the state of an AWS lambda Function resource.
            displayName: Function
            kind: Function
            name: functions.lambda.services.k8s.aws
            version: v1alpha1
          - description: FunctionURLConfig represents the state of an AWS lambda FunctionURLConfig
              resource.
            displayName: FunctionURLConfig
            kind: FunctionURLConfig
            name: functionurlconfigs.lambda.services.k8s.aws
            version: v1alpha1
          - description: LayerVersion represents the state of an AWS lambda LayerVersion
              resource.
            displayName: LayerVersion
            kind: LayerVersion
            name: layerversions.lambda.services.k8s.aws
            version: v1alpha1
          - description: Version represents the state of an AWS lambda Version resource.
            displayName: Version
            kind: Version
            name: versions.lambda.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Lambda resources in AWS from within your Kubernetes cluster.

          **About Amazon Lambda**

          Lambda is a compute service that lets you run code without provisioning or managing servers. Lambda runs your code on a high-availability compute infrastructure and performs all of the administration of the compute resources, including server and operating system maintenance, capacity provisioning and automatic scaling, code monitoring and logging. With Lambda, you can run code for virtually any type of application or backend service. All you need to do is supply your code in one of the [languages that Lambda supports](https://docs.aws.amazon.com/lambda/latest/dg/lambda-runtimes.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Lambda
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lambda
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Lambda Developer Resources
          url: https://aws.amazon.com/lambda/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: lambda maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/lambda-controller:1.9.1
        version: 1.9.1
      entries:
      - name: ack-lambda-controller.v1.9.1
        version: 1.9.1
      - name: ack-lambda-controller.v1.9.0
        version: 1.9.0
      - name: ack-lambda-controller.v1.8.0
        version: 1.8.0
      - name: ack-lambda-controller.v1.7.7
        version: 1.7.7
      - name: ack-lambda-controller.v1.7.6
        version: 1.7.6
      - name: ack-lambda-controller.v1.7.5
        version: 1.7.5
      - name: ack-lambda-controller.v1.7.4
        version: 1.7.4
      - name: ack-lambda-controller.v1.7.2
        version: 1.7.2
      - name: ack-lambda-controller.v1.7.1
        version: 1.7.1
      - name: ack-lambda-controller.v1.7.0
        version: 1.7.0
      - name: ack-lambda-controller.v1.6.3
        version: 1.6.3
      - name: ack-lambda-controller.v1.6.2
        version: 1.6.2
      - name: ack-lambda-controller.v1.6.1
        version: 1.6.1
      - name: ack-lambda-controller.v1.6.0
        version: 1.6.0
      - name: ack-lambda-controller.v1.5.8
        version: 1.5.8
      - name: ack-lambda-controller.v1.5.7
        version: 1.5.7
      - name: ack-lambda-controller.v1.5.6
        version: 1.5.6
      - name: ack-lambda-controller.v1.5.5
        version: 1.5.5
      - name: ack-lambda-controller.v1.5.4
        version: 1.5.4
      - name: ack-lambda-controller.v1.5.3
        version: 1.5.3
      - name: ack-lambda-controller.v1.5.2
        version: 1.5.2
      - name: ack-lambda-controller.v1.5.1
        version: 1.5.1
      - name: ack-lambda-controller.v1.5.0
        version: 1.5.0
      - name: ack-lambda-controller.v1.4.2
        version: 1.4.2
      - name: ack-lambda-controller.v1.4.0
        version: 1.4.0
      - name: ack-lambda-controller.v1.3.4
        version: 1.3.4
      - name: ack-lambda-controller.v1.3.2
        version: 1.3.2
      - name: ack-lambda-controller.v1.3.1
        version: 1.3.1
      - name: ack-lambda-controller.v1.3.0
        version: 1.3.0
      - name: ack-lambda-controller.v1.2.1
        version: 1.2.1
      - name: ack-lambda-controller.v1.2.0
        version: 1.2.0
      - name: ack-lambda-controller.v1.1.1
        version: 1.1.1
      - name: ack-lambda-controller.v1.1.0
        version: 1.1.0
      - name: ack-lambda-controller.v1.0.1
        version: 1.0.1
      - name: ack-lambda-controller.v1.0.0
        version: 1.0.0
      - name: ack-lambda-controller.v0.1.7
        version: 0.1.7
      - name: ack-lambda-controller.v0.1.6
        version: 0.1.6
      - name: ack-lambda-controller.v0.1.5
        version: 0.1.5
      - name: ack-lambda-controller.v0.1.3
        version: 0.1.3
      - name: ack-lambda-controller.v0.1.2
        version: 0.1.2
      - name: ack-lambda-controller.v0.1.1
        version: 0.1.1
      - name: ack-lambda-controller.v0.1.0
        version: 0.1.0
      - name: ack-lambda-controller.v0.0.17
        version: 0.0.17
      - name: ack-lambda-controller.v0.0.16
        version: 0.0.16
      - name: ack-lambda-controller.v0.0.15
        version: 0.0.15
      - name: ack-lambda-controller.v0.0.14
        version: 0.0.14
      - name: ack-lambda-controller.v0.0.13
        version: 0.0.13
      - name: ack-lambda-controller.v0.0.12
        version: 0.0.12
      - name: ack-lambda-controller.v0.0.11
        version: 0.0.11
      - name: ack-lambda-controller.v0.0.10
        version: 0.0.10
      - name: ack-lambda-controller.v0.0.9
        version: 0.0.9
      - name: ack-lambda-controller.v0.0.8
        version: 0.0.8
      - name: ack-lambda-controller.v0.0.7
        version: 0.0.7
      - name: ack-lambda-controller.v0.0.6
        version: 0.0.6
      - name: ack-lambda-controller.v0.0.5
        version: 0.0.5
      - name: ack-lambda-controller.v0.0.4
        version: 0.0.4
      - name: ack-lambda-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-lambda-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: fence-agents-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fence-agents-remediation.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/fence-agents-remediation-operator:v0.4.0
          createdAt: "2024-05-23 19:57:04"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://www.medik8s.io/remediation/fence-agents-remediation/fence-agents-remediation/
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/medik8s/fence-agents-remediation-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      - name: fence-agents-remediation.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: fence-agents-remediation
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: hpe-filex-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-filex-csi-operator.v1.0.0-beta3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.hpe.com/v1",
                "kind": "HPEGreenLakeFileCSIDriver",
                "metadata": {
                  "name": "hpegreenlakefilecsidriver-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "disableNodeConformance": false,
                  "imagePullPolicy": "IfNotPresent",
                  "images": {
                    "csiAttacher": "registry.k8s.io/sig-storage/csi-attacher@sha256:6e54dae32284f60a2de1645c527715e4c934e5ce7899560c0d845bac957118dc",
                    "csiControllerDriver": "quay.io/hpestorage/filex-csi-driver@sha256:8e96df8f21db969dcd507c6fef8ecc82b7f4d9b33a9159f16345ebe7e02d8161",
                    "csiNodeDriver": "quay.io/hpestorage/filex-csi-driver@sha256:8e96df8f21db969dcd507c6fef8ecc82b7f4d9b33a9159f16345ebe7e02d8161",
                    "csiNodeDriverRegistrar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f",
                    "csiNodeInit": "quay.io/hpestorage/filex-csi-init@sha256:65a014b344d262f628100dcd96fecd228c42162bbfa67037952a5a67f16a1437",
                    "csiProvisioner": "registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f",
                    "csiResizer": "registry.k8s.io/sig-storage/csi-resizer@sha256:ab774734705a906561e15b67f6a96538f3ad520335d636f793aaafe87a3b5200",
                    "csiSnapshotter": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:dd788d79cf4c1b8edee6d9b80b8a1ebfc51a38a365c5be656986b129be9ac784"
                  },
                  "kubeletRootDir": "/var/lib/kubelet",
                  "node": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/hpestorage/filex-csi-driver-operator@sha256:e00066fccf2c673547ccafffb0a280130b6825ab6a82a52bd43d752535c63775
          createdAt: "2025-01-10T18:20:35Z"
          description: A Container Storage Interface (CSI) driver for HPE GreenLake
            File Storage.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.0.0-beta3'
          operatorhub.io/ui-metadata-max-k8s-version: "1.32"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/hpe-storage/co-deployments
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE GreenLake File Storage CSI Driver
            displayName: HPEGreenLakeFileCSIDriver
            kind: HPEGreenLakeFileCSIDriver
            name: hpegreenlakefilecsidrivers.storage.hpe.com
            version: v1
        description: |
          The HPE GreenLake File Storage CSI Operator deploys the  HPE GreenLake File Storage CSI Driver on OpenShift for dynamic  provisioning of persistent volumes.
          Please refer to our CSI driver [documentation](https://scod.hpedev.io/filex_csi_driver/index.html) for supported HPE platforms and full list of supported CSI features.
          ## Installation
          Refer to the HPE GreenLake File Storage CSI Operator [documentation](https://scod.hpedev.io/filex_csi_driver/deployment.html#operator).
        displayName: HPE GreenLake File Storage CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE CSI
        - HPE Storage
        - CSI
        links:
        - name: Documentation
          url: https://scod.hpedev.io/filex_csi_driver/index.html
        maintainers:
        - email: dev-hi-containers@hpe.com
          name: HPE Storage Containers Team
        maturity: beta
        minKubeVersion: 1.25.0
        provider:
          name: HPE
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - quay.io/hpestorage/filex-csi-init@sha256:65a014b344d262f628100dcd96fecd228c42162bbfa67037952a5a67f16a1437
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:dd788d79cf4c1b8edee6d9b80b8a1ebfc51a38a365c5be656986b129be9ac784
        - registry.k8s.io/sig-storage/csi-resizer@sha256:ab774734705a906561e15b67f6a96538f3ad520335d636f793aaafe87a3b5200
        - quay.io/hpestorage/filex-csi-driver@sha256:8e96df8f21db969dcd507c6fef8ecc82b7f4d9b33a9159f16345ebe7e02d8161
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3fcd8e2bf0bcb8ff8c93a87af2c59a3bcae7be8792f9d3236c9b5bbd9b6db3b2
        - registry.connect.redhat.com/hpestorage/filex-csi-driver-operator@sha256:e00066fccf2c673547ccafffb0a280130b6825ab6a82a52bd43d752535c63775
        - registry.k8s.io/sig-storage/csi-attacher@sha256:6e54dae32284f60a2de1645c527715e4c934e5ce7899560c0d845bac957118dc
        version: 1.0.0-beta3
      entries:
      - name: hpe-filex-csi-operator.v1.0.0-beta3
        version: 1.0.0-beta3
      name: beta
    defaultChannel: beta
    packageName: hpe-filex-csi-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SmartX
      provider-url: https://www.smartx.com
    name: iomesh-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: iomesh-operator.0.10.1-rc4
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
          createdAt: 2021-08-17 00:00:00+00:00
          repository: https://github.com/iomesh/docs
          support: SmartX, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: iomesh
            displayName: iomesh
            kind: IOMeshCluster
            name: iomeshclusters.iomesh.com
            version: v1alpha1
        description: "IOMesh (https://www.iomesh.com/) is a cloud-native storage product
          specifically designed and developed for Kubernetes. It takes ZBS as its
          core, a “production-ready” distributed storage system developed independently
          by SmartX, to provide production-level high performance and reliable persistent
          storage for business-critical applications such as MySQL, Cassandra, MongoDB,
          etc., contributing to the containerized transformation of stateful applications.
          \     \n## Pre-Requisites\n* According to https://docs.iomesh.com/deploy/prerequisites,
          make sure that your openshift cluster meets the \"Installation Requirements\"
          and set up the worker nodes correctly according to the \"Setup Worker Node\"
          chapter.\n* Run IOMesh Operator pre-installation script in an environment
          where oc or kubectl can be used to access the openshift cluster, the script
          will install the dependencies of IOMesh Operator and specific IOMesh settings
          for the openshift cluster:\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-pre-install-openshift.sh
          | sh -\n```\n## Install IOMesh Operator and IOMesh Cluster\n* Click install
          button on the current web page to install IOMesh Operator\n* On the **Installed
          Operators** -> **IOMesh Operator** -> **Create instance** -> **YAML view**,
          fill in an IOMesh Custom Resources according https://iomesh.run/iomesh.yaml,
          change the `spec.*.dataCIDR` to your own data network CIDR:\n## Post Install\n*
          Run IOMesh Operator post-installation script in an environment where oc
          or kubectl can be used to access the openshift cluster, the script will
          install IOMesh CSI Driver\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-post-install-openshift.sh
          | sh -\n```\n## Getting Started\n* Try to quickstart setup and use IOMesh:
          https://docs.iomesh.com/deploy/setup-iomesh\n"
        displayName: IOMesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iomesh-operator
        links:
        - name: IOMesh Operator
          url: https://iomesh.com
        maintainers:
        - email: na.wang@smartx.com
          name: Na Wang
        maturity: alpha
        provider:
          name: SmartX
          url: https://www.smartx.com
        relatedImages:
        - iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
        version: 0.10.1-rc4
      entries:
      - name: iomesh-operator.0.10.1-rc4
        version: 0.10.1-rc4
      name: stable
    defaultChannel: stable
    packageName: iomesh-operator
    provider:
      name: SmartX
      url: https://www.smartx.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-activemqartemis: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/
    name: amq-broker-rhel9
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-broker-operator.v7.13.1-opr-1-0.1758027574.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 9 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-broker-operator.v7.13.1-opr-1-0.1756381365.p
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 9 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel9-operator@sha256:f34672152017a94ed51056435e218341e22787788ceed52439c1ff066480c715
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9cc48eecf1442ae04b8543fa5d4381a13bc2831390850828834d387006d1342b
        - registry.redhat.io/amq7/amq-broker-rhel9@sha256:755d0f7421efba76f52818bb3a8667cd7d2dcff2c7054dac1a4637b74ecac51d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:47fd4ce2533496828aba37bd1f9715e2164d5c90bd0fc6b25e7e0786d723bf01
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b2402179b1d0ff9c950defe5d8808195ef9a4c3e13e3c469cae31ab51d5f5014
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:bb6fbd68475a7852b4d99eea6c4ab313f9267da7963162f0d75375d7063409e7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:40920687402fd747ca02d384e63692a5e1db625723a2bf3056a607c92ee8eee1
        - registry.redhat.io/amq7/amq-broker-rhel9@sha256:4304a5414911fe55ce4c0133723be234ff288463f4916bd18e743d6359489138
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:7f9b921d89b6d4d91b365659afff127d99f1455b94b48a10a6a7f4c38dd8b700
        - registry.redhat.io/amq7/amq-broker-init-rhel9@sha256:ff80e0a7accda6825bf834d9aec27f48205dd6329991890dd017b4d21d0da5c8
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:16253895bc113f395dc6ab5cd511901e43b6ee676369ec80d598a8741c3c9bf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d42d713da0ce6806fdc6492b6342586783e6865a82a8647d3c4288439b1751ee
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:dc13871dfed52402c40fd192900fc3cb5a49da4fa08e4ec4d0e23a879e83db25
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ffffa9875f0379e9373f89f05eb06e5a193273bb04bc3aa5f85b044357b79098
        - registry.redhat.io/amq7/amq-broker-init-rhel9@sha256:e2cb996e0bf131e71fb8e80c128af18ab2b9330c9fca7a5437842796158e2870
        version: 7.13.1-opr-1+0.1758027574.p
      entries:
      - name: amq-broker-operator.v7.13.1-opr-1
        version: 7.13.1-opr-1
      - name: amq-broker-operator.v7.13.1-opr-1-0.1754564924.p
        version: 7.13.1-opr-1+0.1754564924.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1756374447.p
        version: 7.13.1-opr-1+0.1756374447.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1756381365.p
        version: 7.13.1-opr-1+0.1756381365.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1758024989.p
        version: 7.13.1-opr-1+0.1758024989.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1758027574.p
        version: 7.13.1-opr-1+0.1758027574.p
      - name: amq-broker-operator.v7.13.0-opr-1
        version: 7.13.0-opr-1
      - name: amq-broker-operator.v7.13.0-opr-1-0.1747315188.p
        version: 7.13.0-opr-1+0.1747315188.p
      - name: amq-broker-operator.v7.13.0-opr-1-0.1747658456.p
        version: 7.13.0-opr-1+0.1747658456.p
      name: 7.13.x
    defaultChannel: 7.13.x
    packageName: amq-broker-rhel9
    provider:
      name: Red Hat
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: quay-bridge-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-bridge-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "my-quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecretName": "<NAMESPACE>/<SECRET>",
                  "quayHostname": "https://<QUAY_URL>",
                  "insecureRegistry": false,
                  "whitelistNamespaces": ["default"]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:501fe9b003b5066b3936def0ef4de05abd1a12e795e718d6c9a62d2c6c804c2e
          createdAt: 2021-11-02 11:11 UTC
          description: Enhance OCP using Red Hat Quay container registry
          repository: https://github.com/redhat-cop/quay-openshift-registry-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a connected Quay instance.
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.redhatcop.redhat.io
            version: v1alpha1
        description: "## Overview\n\nThe OpenShift Container Platform contains an
          in cluster image registry for pushing and pulling images that either produced
          within the cluster or are references from external sources. Quay is a highly
          available, enterprise grade image registry with enhanced role based access
          control (RBAC) features. The goal of this operator is to support replicating
          the features found within OpenShift intgrated image registry with Quay.\n\n##
          Functionality\n\nThe following capabilities are enabled through this operator:\n\n*
          Synchronization of OpenShift namespaces as Quay organizations\n    * Creation
          of Robot accounts for each default namespace service account\n    * Creation
          of Secrets for each created Robot Account\n        * Associate Robot Secret
          to Service Account as Mountable and Image Pull Secret\n    * Synchronization
          of OpenShift ImageStreams as Quay Repositories\n* Automatically rewrite
          new Builds making use of ImageStreams to output to Quay\n* Automatically
          import ImageStream tag once build completes\n\n## Prerequisites\n\nThe following
          requirements must be satisfied prior to setting and configuring the integration
          between OpenShift and Quay:\n\n* A Red Hat Quay environment with a user
          with superuser permissions\n* A Red Hat OpenShift Container Platform environment
          (at least version 4.0) with cluster-admin permissions on an OpenShift cluster
          \n*\tOpenShift Command Line tool (CLI)\n\n## Setup and Configuration\n\nThe
          setup process involves steps within both Quay and OpenShift as described
          below:\n\n### Quay Setup\n\nThe integration between OpenShift and Quay makes
          extensive use of the Quay RESTFul API. To allow for other methods aside
          from user credentials to be used for authenticating against the API, an
          application is required to be created. When an application is created, an
          OAuth token is generated which can be used by the operator in OpenShift.
          Applications are created within a Quay organization. While an existing organization
          can be utilized, it is recommended that a dedicated organization be used.\n\n1.\tLogin
          to Quay as a user with superuser access and select the organization for
          which the external application will be configured within.\n2.\tOn the left
          hand side of the page, select **Applications**\n3.\tCreate a new application
          by selecting the **Create New Application** button and entering a name for
          the application\n4.\tAfter the new application has been created, select
          the application that was created\n5.\tOn the lefthand navigation bar, select
          the **Generate Token** button to create a new OAuth2 token\n6.\tSelect all
          of the checkboxes to grant the necessary access that is needed for the integration\n7.\tReview
          the permissions that will be assigned and then select the **Authorize Application**
          button\n8.\tTake note of the generated Access Token as it will be needed
          in the following section\n\n### OpenShift Setup\n\nTo complete the process
          to begin to integrate OpenShift and Quay, several steps are required. Before
          beginning, ensure that you have the OpenShift Command Line tool installed
          and that you are logged into OpenShift as a cluster administrator. \n\n####
          Create A Secret for the Quay OAuth Token\n\nThe Operator will use the previously
          obtained Access Token to communicate with Quay. Store this token within
          OpenShift as a secret.\n\nExecute the following command to create a secret
          called `quay-integration` with a key called `token` containing the access
          token:\n\n```\n$ oc create secret generic quay-integration --from-literal=token=<access_token>\n```\n\nThis
          token will be referenced in the final section.\n\n#### MutatingWebhookConfiguration
          Support\n\nSupport for dynamic interception of API requests that are performed
          as part of OpenShift’s typical build process is facilitated through a [MutatingWebhookConfiguration](https://docs.openshift.com/container-platform/3.11/architecture/additional_concepts/dynamic_admission_controllers.html).
          A MutatingWebhookConfiguration allows for invoking an API running within
          a project on OpenShift when certain API requests are received. In particular,
          to support Quay’s integration with OpenShift, any new Build requests should
          be intercepted so that the output can be modified to target Quay instead
          of OpenShift’s integrated registry. \n\nKubernetes requires that the webhook
          endpoint is secured via SSL using a certificate that makes use of the certificate
          authority for the cluster. Fortunately, OpenShift provides support for generating
          a certificate signed by the cluster. A script is available in the documentation
          called `webhook-create-signed-cert.sh` that automates the process for requesting
          a new certificate.\n\nFirst, create a `Service` for the Operator's webhook
          endpoint:\n```yaml\napiVersion: v1\nkind: Service\nmetadata:\n  labels:\n
          \   name: quay-bridge-operator\n  name: quay-bridge-operator\n  namespace:
          openshift-operators\nspec:\n  ports:\n    - name: https\n      port: 443\n
          \     protocol: TCP\n      targetPort: 8443\n  selector:\n    name: quay-bridge-operator\n
          \ sessionAffinity: None\n  type: ClusterIP\n```\n\nExecute the following
          command to request the certificate:\n\n```\n./webhook-create-signed-cert.sh
          --namespace openshift-operators --secret quay-bridge-operator-webhook-certs
          --service quay-bridge-operator\n```\n\nThe result will place the newly created
          private key and certificate within a secret within the secret specified.
          The secret will be mounted into the appropriate located within the operator
          as declared in the Deployment of the Operator. \n\nOnce the secret has been
          created, focus can shift to the `MutatingWebhookConfiguration` (see below).
          The majority of the content should be in place to apply to the cluster.
          The one section that requires configuration is the caBundle property. This
          refers to the Certificate Authority (CA) for the OpenShift environment.
          The OpenShift CA is available as a _ConfigMap_ within the `kube-system`
          namespace.\n\nExecute the following command to retrieve the CA and format
          the result as a single line so that it can be entered into the MutatingWebhookConfiguration
          resource:\n\n```\noc get configmap -n kube-system extension-apiserver-authentication
          -o=jsonpath='{.data.client-ca-file}' | base64 | tr -d '\\n'\n```\n\nReplace
          the `${CA_BUNDLE}` variable in the following `MutatingWebhookConfiguration`
          YAML and create it:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\nkind:
          MutatingWebhookConfiguration \nmetadata:\n  name: quay-bridge-operator\nwebhooks:\n
          \ - name: quayintegration.redhatcop.redhat.io\n    clientConfig: \n      service:\n
          \       namespace: openshift-operators\n        name: quay-bridge-operator\n
          \       path: \"/admissionwebhook\"\n      caBundle: \"${CA_BUNDLE}\"\n
          \   rules: \n    - operations:  [ \"CREATE\" ]\n      apiGroups: [ \"build.openshift.io\"
          ]\n      apiVersions: [\"v1\" ]\n      resources: [ \"builds\" ]\n    failurePolicy:
          Fail\n```\n\n_Note: Until the operator is running, new requests for builds
          will fail since the webserver the MutatingWebhookConfiguration invokes is
          not available and a proper is response is required in order for the object
          to be persisted in etcd._\n\n#### Create the QuayIntegration Custom Resource\n\nFinally,
          to complete the integration between OpenShift and Quay, a `QuayIntegration`
          custom resource needs to be created.\n\nThe following is an example of a
          basic definition of a `QuayIntegration` resource associated from the associated
          CRD.\n\n```\napiVersion: redhatcop.redhat.io/v1alpha1\nkind: QuayIntegration\nmetadata:\n
          \ name: example-quayintegration\nspec:\n  clusterID: openshift\n  credentialsSecretName:
          <NAMESPACE>/<SECRET>\n  quayHostname: https://<QUAY_URL>\n```\n\nThe _clusterID_
          is a value which should be unique across the entire ecosystem. This value
          is optional and defaults to `openshift`.\n\nThe _credentialsSecretName_
          is a NamespacedName value of the secret containing the token that was previously
          created.\n\nNote: If Quay is using self signed certificates, the property
          `insecureRegistry: true`\n\nOrganizations within Quay should be created
          for the related namespaces in OpenShift."
        displayName: Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Source Code
          url: https://github.com/redhat-cop/quay-openshift-registry-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:501fe9b003b5066b3936def0ef4de05abd1a12e795e718d6c9a62d2c6c804c2e
        version: 3.4.7
      entries:
      - name: quay-bridge-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: quay-bridge-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "my-quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecretName": "<NAMESPACE>/<SECRET>",
                  "quayHostname": "https://<QUAY_URL>",
                  "insecureRegistry": false,
                  "whitelistNamespaces": ["default"]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:9149df5ffc83dce6593072b7f72805e07e700b84527e45b4b7486b9c38f70731
          createdAt: 2021-09-15 23:09 UTC
          description: Enhance OCP using Red Hat Quay container registry
          olm.skipRange: '>=3.4.x <3.5.7'
          repository: https://github.com/redhat-cop/quay-openshift-registry-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a connected Quay instance.
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.redhatcop.redhat.io
            version: v1alpha1
        description: "## Overview\n\nThe OpenShift Container Platform contains an
          in cluster image registry for pushing and pulling images that either produced
          within the cluster or are references from external sources. Quay is a highly
          available, enterprise grade image registry with enhanced role based access
          control (RBAC) features. The goal of this operator is to support replicating
          the features found within OpenShift intgrated image registry with Quay.\n\n##
          Functionality\n\nThe following capabilities are enabled through this operator:\n\n*
          Synchronization of OpenShift namespaces as Quay organizations\n    * Creation
          of Robot accounts for each default namespace service account\n    * Creation
          of Secrets for each created Robot Account\n        * Associate Robot Secret
          to Service Account as Mountable and Image Pull Secret\n    * Synchronization
          of OpenShift ImageStreams as Quay Repositories\n* Automatically rewrite
          new Builds making use of ImageStreams to output to Quay\n* Automatically
          import ImageStream tag once build completes\n\n## Prerequisites\n\nThe following
          requirements must be satisfied prior to setting and configuring the integration
          between OpenShift and Quay:\n\n* A Red Hat Quay environment with a user
          with superuser permissions\n* A Red Hat OpenShift Container Platform environment
          (at least version 4.0) with cluster-admin permissions on an OpenShift cluster
          \n*\tOpenShift Command Line tool (CLI)\n\n## Setup and Configuration\n\nThe
          setup process involves steps within both Quay and OpenShift as described
          below:\n\n### Quay Setup\n\nThe integration between OpenShift and Quay makes
          extensive use of the Quay RESTFul API. To allow for other methods aside
          from user credentials to be used for authenticating against the API, an
          application is required to be created. When an application is created, an
          OAuth token is generated which can be used by the operator in OpenShift.
          Applications are created within a Quay organization. While an existing organization
          can be utilized, it is recommended that a dedicated organization be used.\n\n1.\tLogin
          to Quay as a user with superuser access and select the organization for
          which the external application will be configured within.\n2.\tOn the left
          hand side of the page, select **Applications**\n3.\tCreate a new application
          by selecting the **Create New Application** button and entering a name for
          the application\n4.\tAfter the new application has been created, select
          the application that was created\n5.\tOn the lefthand navigation bar, select
          the **Generate Token** button to create a new OAuth2 token\n6.\tSelect all
          of the checkboxes to grant the necessary access that is needed for the integration\n7.\tReview
          the permissions that will be assigned and then select the **Authorize Application**
          button\n8.\tTake note of the generated Access Token as it will be needed
          in the following section\n\n### OpenShift Setup\n\nTo complete the process
          to begin to integrate OpenShift and Quay, several steps are required. Before
          beginning, ensure that you have the OpenShift Command Line tool installed
          and that you are logged into OpenShift as a cluster administrator. \n\n####
          Create A Secret for the Quay OAuth Token\n\nThe Operator will use the previously
          obtained Access Token to communicate with Quay. Store this token within
          OpenShift as a secret.\n\nExecute the following command to create a secret
          called `quay-integration` with a key called `token` containing the access
          token:\n\n```\n$ oc create secret generic quay-integration --from-literal=token=<access_token>\n```\n\nThis
          token will be referenced in the final section.\n\n#### MutatingWebhookConfiguration
          Support\n\nSupport for dynamic interception of API requests that are performed
          as part of OpenShift’s typical build process is facilitated through a [MutatingWebhookConfiguration](https://docs.openshift.com/container-platform/3.11/architecture/additional_concepts/dynamic_admission_controllers.html).
          A MutatingWebhookConfiguration allows for invoking an API running within
          a project on OpenShift when certain API requests are received. In particular,
          to support Quay’s integration with OpenShift, any new Build requests should
          be intercepted so that the output can be modified to target Quay instead
          of OpenShift’s integrated registry. \n\nKubernetes requires that the webhook
          endpoint is secured via SSL using a certificate that makes use of the certificate
          authority for the cluster. Fortunately, OpenShift provides support for generating
          a certificate signed by the cluster. A script is available in the documentation
          called `webhook-create-signed-cert.sh` that automates the process for requesting
          a new certificate.\n\nFirst, create a `Service` for the Operator's webhook
          endpoint:\n```yaml\napiVersion: v1\nkind: Service\nmetadata:\n  labels:\n
          \   name: quay-bridge-operator\n  name: quay-bridge-operator\n  namespace:
          openshift-operators\nspec:\n  ports:\n    - name: https\n      port: 443\n
          \     protocol: TCP\n      targetPort: 8443\n  selector:\n    name: quay-bridge-operator\n
          \ sessionAffinity: None\n  type: ClusterIP\n```\n\nExecute the following
          command to request the certificate:\n\n```\n./webhook-create-signed-cert.sh
          --namespace openshift-operators --secret quay-bridge-operator-webhook-certs
          --service quay-bridge-operator\n```\n\nThe result will place the newly created
          private key and certificate within a secret within the secret specified.
          The secret will be mounted into the appropriate located within the operator
          as declared in the Deployment of the Operator. \n\nOnce the secret has been
          created, focus can shift to the `MutatingWebhookConfiguration` (see below).
          The majority of the content should be in place to apply to the cluster.
          The one section that requires configuration is the caBundle property. This
          refers to the Certificate Authority (CA) for the OpenShift environment.
          The OpenShift CA is available as a _ConfigMap_ within the `kube-system`
          namespace.\n\nExecute the following command to retrieve the CA and format
          the result as a single line so that it can be entered into the MutatingWebhookConfiguration
          resource:\n\n```\noc get configmap -n kube-system extension-apiserver-authentication
          -o=jsonpath='{.data.client-ca-file}' | base64 | tr -d '\\n'\n```\n\nReplace
          the `${CA_BUNDLE}` variable in the following `MutatingWebhookConfiguration`
          YAML and create it:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\nkind:
          MutatingWebhookConfiguration \nmetadata:\n  name: quay-bridge-operator\nwebhooks:\n
          \ - name: quayintegration.redhatcop.redhat.io\n    clientConfig: \n      service:\n
          \       namespace: openshift-operators\n        name: quay-bridge-operator\n
          \       path: \"/admissionwebhook\"\n      caBundle: \"${CA_BUNDLE}\"\n
          \   rules: \n    - operations:  [ \"CREATE\" ]\n      apiGroups: [ \"build.openshift.io\"
          ]\n      apiVersions: [\"v1\" ]\n      resources: [ \"builds\" ]\n    failurePolicy:
          Fail\n    sideEffects: None\n    admissionReviewVersions: [v1beta1]\n    matchPolicy:
          Exact\n    timeoutSeconds: 30  \n```\n\n_Note: Until the operator is running,
          new requests for builds will fail since the webserver the MutatingWebhookConfiguration
          invokes is not available and a proper is response is required in order for
          the object to be persisted in etcd._\n\n#### Create the QuayIntegration
          Custom Resource\n\nFinally, to complete the integration between OpenShift
          and Quay, a `QuayIntegration` custom resource needs to be created.\n\nThe
          following is an example of a basic definition of a `QuayIntegration` resource
          associated from the associated CRD.\n\n```\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          QuayIntegration\nmetadata:\n  name: example-quayintegration\nspec:\n  clusterID:
          openshift\n  credentialsSecretName: <NAMESPACE>/<SECRET>\n  quayHostname:
          https://<QUAY_URL>\n```\n\nThe _clusterID_ is a value which should be unique
          across the entire ecosystem. This value is optional and defaults to `openshift`.\n\nThe
          _credentialsSecretName_ is a NamespacedName value of the secret containing
          the token that was previously created.\n\nNote: If Quay is using self signed
          certificates, the property `insecureRegistry: true`\n\nOrganizations within
          Quay should be created for the related namespaces in OpenShift."
        displayName: Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Source Code
          url: https://github.com/redhat-cop/quay-openshift-registry-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:9149df5ffc83dce6593072b7f72805e07e700b84527e45b4b7486b9c38f70731
        version: 3.5.7
      entries:
      - name: quay-bridge-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: quay-bridge-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:33dc7673155f01728225221768f6b90d87cc7609dcda46b394dba2421c136a23
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.10.15
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:33dc7673155f01728225221768f6b90d87cc7609dcda46b394dba2421c136a23
        version: 3.10.15
      entries:
      - name: quay-bridge-operator.v3.10.15
        version: 3.10.15
      - name: quay-bridge-operator.v3.10.14
        version: 3.10.14
      - name: quay-bridge-operator.v3.10.13
        version: 3.10.13
      - name: quay-bridge-operator.v3.10.12
        version: 3.10.12
      - name: quay-bridge-operator.v3.10.11
        version: 3.10.11
      - name: quay-bridge-operator.v3.10.10
        version: 3.10.10
      - name: quay-bridge-operator.v3.10.9
        version: 3.10.9
      - name: quay-bridge-operator.v3.10.8
        version: 3.10.8
      - name: quay-bridge-operator.v3.10.7
        version: 3.10.7
      - name: quay-bridge-operator.v3.10.6
        version: 3.10.6
      - name: quay-bridge-operator.v3.10.5
        version: 3.10.5
      - name: quay-bridge-operator.v3.10.4
        version: 3.10.4
      - name: quay-bridge-operator.v3.10.3
        version: 3.10.3
      - name: quay-bridge-operator.v3.10.2
        version: 3.10.2
      - name: quay-bridge-operator.v3.10.1
        version: 3.10.1
      - name: quay-bridge-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-bridge-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:511f8471db70e8d5fd9d0b547c73f1416bbac6f0fe6cdeb6c59a9ef8bd845ac2
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.11.13
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:511f8471db70e8d5fd9d0b547c73f1416bbac6f0fe6cdeb6c59a9ef8bd845ac2
        version: 3.11.13
      entries:
      - name: quay-bridge-operator.v3.11.13
        version: 3.11.13
      - name: quay-bridge-operator.v3.11.12
        version: 3.11.12
      - name: quay-bridge-operator.v3.11.11
        version: 3.11.11
      - name: quay-bridge-operator.v3.11.10
        version: 3.11.10
      - name: quay-bridge-operator.v3.11.9
        version: 3.11.9
      - name: quay-bridge-operator.v3.11.8
        version: 3.11.8
      - name: quay-bridge-operator.v3.11.7
        version: 3.11.7
      - name: quay-bridge-operator.v3.11.6
        version: 3.11.6
      - name: quay-bridge-operator.v3.11.5
        version: 3.11.5
      - name: quay-bridge-operator.v3.11.4
        version: 3.11.4
      - name: quay-bridge-operator.v3.11.3
        version: 3.11.3
      - name: quay-bridge-operator.v3.11.2
        version: 3.11.2
      - name: quay-bridge-operator.v3.11.1
        version: 3.11.1
      - name: quay-bridge-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-bridge-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:c6e04de3ad3b8c0ded95146d090e120bffed7f9dc5e74332a183fe1c2e45d16d
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.12.12
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:c6e04de3ad3b8c0ded95146d090e120bffed7f9dc5e74332a183fe1c2e45d16d
        version: 3.12.12
      entries:
      - name: quay-bridge-operator.v3.12.12
        version: 3.12.12
      - name: quay-bridge-operator.v3.12.11
        version: 3.12.11
      - name: quay-bridge-operator.v3.12.10
        version: 3.12.10
      - name: quay-bridge-operator.v3.12.9
        version: 3.12.9
      - name: quay-bridge-operator.v3.12.8
        version: 3.12.8
      - name: quay-bridge-operator.v3.12.7
        version: 3.12.7
      - name: quay-bridge-operator.v3.12.6
        version: 3.12.6
      - name: quay-bridge-operator.v3.12.5
        version: 3.12.5
      - name: quay-bridge-operator.v3.12.4
        version: 3.12.4
      - name: quay-bridge-operator.v3.12.3
        version: 3.12.3
      - name: quay-bridge-operator.v3.12.2
        version: 3.12.2
      - name: quay-bridge-operator.v3.12.1
        version: 3.12.1
      - name: quay-bridge-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-bridge-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:e8ee1944d8e0e9d515a6e63a3e39901c304abd421fcad2abbf9a3c43c7ca6243
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.13.8
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:e8ee1944d8e0e9d515a6e63a3e39901c304abd421fcad2abbf9a3c43c7ca6243
        version: 3.13.8
      entries:
      - name: quay-bridge-operator.v3.13.8
        version: 3.13.8
      - name: quay-bridge-operator.v3.13.7
        version: 3.13.7
      - name: quay-bridge-operator.v3.13.6
        version: 3.13.6
      - name: quay-bridge-operator.v3.13.5
        version: 3.13.5
      - name: quay-bridge-operator.v3.13.4
        version: 3.13.4
      - name: quay-bridge-operator.v3.13.3
        version: 3.13.3
      - name: quay-bridge-operator.v3.13.2
        version: 3.13.2
      - name: quay-bridge-operator.v3.13.1
        version: 3.13.1
      - name: quay-bridge-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-bridge-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b12f0dfb8fdeae3f521e41fcf60ef118bdfd08ad7f68bd87c5494b067b7c7c58
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.14.5
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b12f0dfb8fdeae3f521e41fcf60ef118bdfd08ad7f68bd87c5494b067b7c7c58
        version: 3.14.5
      entries:
      - name: quay-bridge-operator.v3.14.5
        version: 3.14.5
      - name: quay-bridge-operator.v3.14.4
        version: 3.14.4
      - name: quay-bridge-operator.v3.14.3
        version: 3.14.3
      - name: quay-bridge-operator.v3.14.2
        version: 3.14.2
      - name: quay-bridge-operator.v3.14.1
        version: 3.14.1
      - name: quay-bridge-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-bridge-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b2ac4a48c80209d7bebe8b1b2d066571f99a9f47b78604e47a8300d027941b66
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.15.2
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b2ac4a48c80209d7bebe8b1b2d066571f99a9f47b78604e47a8300d027941b66
        version: 3.15.2
      entries:
      - name: quay-bridge-operator.v3.15.2
        version: 3.15.2
      - name: quay-bridge-operator.v3.15.1
        version: 3.15.1
      - name: quay-bridge-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-bridge-operator.v3.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:0ad7af440c6e00fb578ef40e6d600964324fd94ce7835a28d0ec355b7415093f
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.6.10
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:99dc6b5af811bb33892c17334dd4e1a7eddc6982d49a793b1789f505e00f49b1
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:0ad7af440c6e00fb578ef40e6d600964324fd94ce7835a28d0ec355b7415093f
        version: 3.6.10
      entries:
      - name: quay-bridge-operator.v3.6.10
        version: 3.6.10
      - name: quay-bridge-operator.v3.6.9
        version: 3.6.9
      - name: quay-bridge-operator.v3.6.8
        version: 3.6.8
      - name: quay-bridge-operator.v3.6.7
        version: 3.6.7
      - name: quay-bridge-operator.v3.6.6
        version: 3.6.6
      - name: quay-bridge-operator.v3.6.5
        version: 3.6.5
      - name: quay-bridge-operator.v3.6.4
        version: 3.6.4
      - name: quay-bridge-operator.v3.6.2
        version: 3.6.2
      - name: quay-bridge-operator.v3.6.1
        version: 3.6.1
      - name: quay-bridge-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: quay-bridge-operator.v3.7.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:6a0dd80dc48ad0aaa576b3de78b2b265e29a3c10e4996aa6b8f129592fe20f09
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.7.14
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:975f287513039b2244894b9031e590f665d48f9af480dfdc4b85d4f96dd4bc37
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:6a0dd80dc48ad0aaa576b3de78b2b265e29a3c10e4996aa6b8f129592fe20f09
        version: 3.7.14
      entries:
      - name: quay-bridge-operator.v3.7.14
        version: 3.7.14
      - name: quay-bridge-operator.v3.7.13
        version: 3.7.13
      - name: quay-bridge-operator.v3.7.12
        version: 3.7.12
      - name: quay-bridge-operator.v3.7.11
        version: 3.7.11
      - name: quay-bridge-operator.v3.7.10
        version: 3.7.10
      - name: quay-bridge-operator.v3.7.9
        version: 3.7.9
      - name: quay-bridge-operator.v3.7.8
        version: 3.7.8
      - name: quay-bridge-operator.v3.7.7
        version: 3.7.7
      - name: quay-bridge-operator.v3.7.6
        version: 3.7.6
      - name: quay-bridge-operator.v3.7.5
        version: 3.7.5
      - name: quay-bridge-operator.v3.7.4
        version: 3.7.4
      - name: quay-bridge-operator.v3.7.3
        version: 3.7.3
      - name: quay-bridge-operator.v3.7.2
        version: 3.7.2
      - name: quay-bridge-operator.v3.7.1
        version: 3.7.1
      - name: quay-bridge-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: quay-bridge-operator.v3.8.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:f8544a546dd6d209355c3a1cbe35138f6910491fc4584b401634431da4414ec0
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.8.15
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:975f287513039b2244894b9031e590f665d48f9af480dfdc4b85d4f96dd4bc37
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:f8544a546dd6d209355c3a1cbe35138f6910491fc4584b401634431da4414ec0
        version: 3.8.15
      entries:
      - name: quay-bridge-operator.v3.8.15
        version: 3.8.15
      - name: quay-bridge-operator.v3.8.14
        version: 3.8.14
      - name: quay-bridge-operator.v3.8.13
        version: 3.8.13
      - name: quay-bridge-operator.v3.8.12
        version: 3.8.12
      - name: quay-bridge-operator.v3.8.11
        version: 3.8.11
      - name: quay-bridge-operator.v3.8.10
        version: 3.8.10
      - name: quay-bridge-operator.v3.8.9
        version: 3.8.9
      - name: quay-bridge-operator.v3.8.8
        version: 3.8.8
      - name: quay-bridge-operator.v3.8.7
        version: 3.8.7
      - name: quay-bridge-operator.v3.8.6
        version: 3.8.6
      - name: quay-bridge-operator.v3.8.5
        version: 3.8.5
      - name: quay-bridge-operator.v3.8.4
        version: 3.8.4
      - name: quay-bridge-operator.v3.8.3
        version: 3.8.3
      - name: quay-bridge-operator.v3.8.2
        version: 3.8.2
      - name: quay-bridge-operator.v3.8.1
        version: 3.8.1
      - name: quay-bridge-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-bridge-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:875e0b63fdb7d44879650eb2f788ea705d753520841a5d875428f0b3ca70a36c
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.9.16
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:875e0b63fdb7d44879650eb2f788ea705d753520841a5d875428f0b3ca70a36c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 3.9.16
      entries:
      - name: quay-bridge-operator.v3.9.16
        version: 3.9.16
      - name: quay-bridge-operator.v3.9.15
        version: 3.9.15
      - name: quay-bridge-operator.v3.9.14
        version: 3.9.14
      - name: quay-bridge-operator.v3.9.13
        version: 3.9.13
      - name: quay-bridge-operator.v3.9.12
        version: 3.9.12
      - name: quay-bridge-operator.v3.9.11
        version: 3.9.11
      - name: quay-bridge-operator.v3.9.10
        version: 3.9.10
      - name: quay-bridge-operator.v3.9.9
        version: 3.9.9
      - name: quay-bridge-operator.v3.9.8
        version: 3.9.8
      - name: quay-bridge-operator.v3.9.7
        version: 3.9.7
      - name: quay-bridge-operator.v3.9.6
        version: 3.9.6
      - name: quay-bridge-operator.v3.9.5
        version: 3.9.5
      - name: quay-bridge-operator.v3.9.4
        version: 3.9.4
      - name: quay-bridge-operator.v3.9.3
        version: 3.9.3
      - name: quay-bridge-operator.v3.9.2
        version: 3.9.2
      - name: quay-bridge-operator.v3.9.1
        version: 3.9.1
      - name: quay-bridge-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: quay-bridge-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: machine-deletion-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: machine-deletion-remediation.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:420a4d4a8610a3694fd3ef9809c55a4f140ec8cea7f09219647f82a84d1f411b
          createdAt: "2023-10-30 13:59:28"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          olm.skipRange: '>=0.1.0 <0.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:420a4d4a8610a3694fd3ef9809c55a4f140ec8cea7f09219647f82a84d1f411b
        version: 0.2.1
      entries:
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.14-eus
    - currentCSV: machine-deletion-remediation.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:1445e6df8f49672aa729806f0e257043ecc22e19e666bb5a799286c6b878b1dd
          createdAt: "2024-08-11 06:47:20"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.2.0 <0.3.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f4381aaf002b761a2ba614332a2af973f4de5f6a4c84dd96b8808725e561fa89
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:1445e6df8f49672aa729806f0e257043ecc22e19e666bb5a799286c6b878b1dd
        version: 0.3.1
      entries:
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.16-eus
    - currentCSV: machine-deletion-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-29 09:35:10"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.3.1 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.6/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel9-operator@sha256:dbe9095d494ebc7b56982cb1f74329d7a7fd0cbb7544e9f90207eb35de7997dc
        version: 0.4.1
      entries:
      - name: machine-deletion-remediation.v0.4.1
        version: 0.4.1
      - name: machine-deletion-remediation.v0.4.0
        version: 0.4.0
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.18-eus
    - currentCSV: machine-deletion-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:7e1cad02f1e74a1355c168fc53925bd0fbcc55f2535a46f734c8de082eb49400
          createdAt: "2023-04-24 08:22:17"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3af07bfe0fb73b1a7f00487eba302d6da442e742ab4007fc50e52efd4d45dae5
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:7e1cad02f1e74a1355c168fc53925bd0fbcc55f2535a46f734c8de082eb49400
        version: 0.1.0
      entries:
      - name: machine-deletion-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: machine-deletion-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-29 09:35:10"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.3.1 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.6/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel9-operator@sha256:dbe9095d494ebc7b56982cb1f74329d7a7fd0cbb7544e9f90207eb35de7997dc
        version: 0.4.1
      entries:
      - name: machine-deletion-remediation.v0.4.1
        version: 0.4.1
      - name: machine-deletion-remediation.v0.4.0
        version: 0.4.0
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: stable
    defaultChannel: stable
    packageName: machine-deletion-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: sailoperator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sailoperator.v1.28.0-nightly-2025-10-06
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:07ea72de6ff9184f346ee2948228e73cfe4180c8d05fe7ee972c22378ad450ef
          createdAt: "2025-10-06T03:29:39Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0
          - master
          - v1.28-alpha.8de61b9a

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-testing/ztunnel@sha256:38b5902bf36f5163274ce0963dfb1e106000ac639d429f05846448de8454492d
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - quay.io/sail-dev/sail-operator@sha256:07ea72de6ff9184f346ee2948228e73cfe4180c8d05fe7ee972c22378ad450ef
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - gcr.io/istio-testing/install-cni@sha256:9525a3711201bcca04ce2ca9b1576d4094a03779d60db0a36b15a956020bbec1
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-testing/proxyv2@sha256:11cf6a3083a3c1ffa539a21cee30ddc0abe47fb44bb16520cdc2f4141dcfee03
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-testing/pilot@sha256:13e2125870f59f752dc5834b702c2dab28f36edc00dccdeeaf872023018e3df0
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        version: 1.28.0-nightly-2025-10-06
      entries:
      - name: sailoperator.v1.28.0-nightly-2025-10-06
        version: 1.28.0-nightly-2025-10-06
      - name: sailoperator.v1.28.0-nightly-2025-10-05
        version: 1.28.0-nightly-2025-10-05
      - name: sailoperator.v1.28.0-nightly-2025-10-04
        version: 1.28.0-nightly-2025-10-04
      - name: sailoperator.v1.28.0-nightly-2025-10-03
        version: 1.28.0-nightly-2025-10-03
      - name: sailoperator.v1.28.0-nightly-2025-10-02
        version: 1.28.0-nightly-2025-10-02
      - name: sailoperator.v1.28.0-nightly-2025-10-01
        version: 1.28.0-nightly-2025-10-01
      - name: sailoperator.v1.28.0-nightly-2025-09-30
        version: 1.28.0-nightly-2025-09-30
      - name: sailoperator.v1.28.0-nightly-2025-09-29
        version: 1.28.0-nightly-2025-09-29
      - name: sailoperator.v1.28.0-nightly-2025-09-28
        version: 1.28.0-nightly-2025-09-28
      - name: sailoperator.v1.28.0-nightly-2025-09-27
        version: 1.28.0-nightly-2025-09-27
      - name: sailoperator.v1.28.0-nightly-2025-09-26
        version: 1.28.0-nightly-2025-09-26
      - name: sailoperator.v1.28.0-nightly-2025-09-25
        version: 1.28.0-nightly-2025-09-25
      - name: sailoperator.v1.28.0-nightly-2025-09-23
        version: 1.28.0-nightly-2025-09-23
      - name: sailoperator.v1.28.0-nightly-2025-09-22
        version: 1.28.0-nightly-2025-09-22
      name: 1.28-nightly
    - currentCSV: sailoperator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.23.0"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.23.0"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/maistra-dev/sail-operator:0.1-latest
          createdAt: "2024-08-20T11:54:30Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1alpha1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1alpha1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1alpha1
          - description: |-
              RemoteIstio represents a remote Istio Service Mesh deployment consisting of one or more
              remote control plane instances (represented by one or more IstioRevision objects).
            displayName: Remote Istio
            kind: RemoteIstio
            name: remoteistios.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.23.0
          - v1.22.4

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - docker.io/istio/proxyv2:1.22.4
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.16.0
        - quay.io/maistra-dev/sail-operator:0.1-latest
        - docker.io/istio/pilot:1.23.0
        - docker.io/istio/proxyv2:1.23.0
        - docker.io/istio/install-cni:1.22.4
        - docker.io/istio/pilot:1.22.4
        - docker.io/istio/ztunnel:1.23.0
        - docker.io/istio/ztunnel:1.22.4
        - docker.io/istio/install-cni:1.23.0
        version: 0.1.0
      entries:
      - name: sailoperator.v0.1.0
        version: 0.1.0
      - name: sailoperator.v0.1.0-rc.1
        version: 0.1.0-rc.1
      name: candidates
    - currentCSV: sailoperator.v1.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
          createdAt: "2025-09-05T10:12:19Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        version: 1.27.1
      entries:
      - name: sailoperator.v1.27.1
        version: 1.27.1
      - name: sailoperator.v1.27.0
        version: 1.27.0
      - name: sailoperator.v1.26.3
        version: 1.26.3
      - name: sailoperator.v1.26.2
        version: 1.26.2
      - name: sailoperator.v1.26.1
        version: 1.26.1
      - name: sailoperator.v1.26.0
        version: 1.26.0
      - name: sailoperator.v1.25.2
        version: 1.25.2
      - name: sailoperator.v1.25.1
        version: 1.25.1
      - name: sailoperator.v1.25.0
        version: 1.25.0
      name: stable
    - currentCSV: sailoperator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.3"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.3"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator:1.0-latest
          createdAt: "2025-03-07T10:17:30Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.24-latest
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.23-latest
          - v1.23.5
          - v1.23.4
          - v1.23.3
          - v1.23.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/sail-dev/sail-operator:1.0-latest
        - gcr.io/istio-release/ztunnel:1.24.1
        - gcr.io/istio-release/proxyv2:1.24.2
        - gcr.io/istio-release/install-cni:1.24.3
        - gcr.io/istio-release/install-cni:1.23.0
        - gcr.io/istio-release/proxyv2:1.23.5
        - gcr.io/istio-release/proxyv2:1.23.4
        - gcr.io/istio-release/pilot:1.24.2
        - gcr.io/istio-release/pilot:1.23.0
        - gcr.io/istio-release/install-cni:1.24.2
        - gcr.io/istio-release/pilot:1.23.4
        - gcr.io/istio-release/proxyv2:1.23.3
        - gcr.io/istio-release/proxyv2:1.23.0
        - gcr.io/istio-release/pilot:1.23.3
        - gcr.io/istio-release/pilot:1.24.1
        - gcr.io/istio-release/pilot:1.23.5
        - gcr.io/istio-release/proxyv2:1.24.3
        - gcr.io/istio-release/ztunnel:1.24.3
        - gcr.io/istio-release/ztunnel:1.23.4
        - gcr.io/istio-release/install-cni:1.23.5
        - gcr.io/istio-release/ztunnel:1.23.3
        - gcr.io/istio-release/install-cni:1.24.1
        - gcr.io/istio-release/proxyv2:1.24.1
        - gcr.io/istio-release/ztunnel:1.23.0
        - gcr.io/istio-release/install-cni:1.23.4
        - gcr.io/istio-release/ztunnel:1.23.5
        - gcr.io/istio-release/ztunnel:1.24.2
        - gcr.io/istio-release/install-cni:1.23.3
        - gcr.io/istio-release/pilot:1.24.3
        version: 1.0.0
      entries:
      - name: sailoperator.v1.0.0
        version: 1.0.0
      name: stable-1.0
    - currentCSV: sailoperator.v1.25.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.25.2"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.25.2"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator:1.25.2
          createdAt: "2025-05-07T19:44:51Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.25.2'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.25-latest
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0
          - v1.23-latest
          - v1.23.5
          - v1.23.4
          - v1.23.3
          - v1.23.2
          - v1.22-latest
          - v1.22.8
          - v1.22.7
          - v1.22.6
          - v1.22.5
          - v1.21.6

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/ztunnel:1.22.6
        - gcr.io/istio-release/pilot:1.22.8
        - gcr.io/istio-release/install-cni:1.23.5
        - quay.io/sail-dev/sail-operator:1.25.2
        - gcr.io/istio-release/pilot:1.22.6
        - gcr.io/istio-release/install-cni:1.24.2
        - gcr.io/istio-release/install-cni:1.24.3
        - gcr.io/istio-release/install-cni:1.23.3
        - gcr.io/istio-release/ztunnel:1.22.5
        - gcr.io/istio-release/proxyv2:1.23.3
        - gcr.io/istio-release/pilot:1.23.2
        - gcr.io/istio-release/install-cni:1.25.1
        - gcr.io/istio-release/pilot:1.24.3
        - gcr.io/istio-release/ztunnel:1.25.1
        - gcr.io/istio-release/install-cni:1.22.8
        - gcr.io/istio-release/proxyv2:1.22.8
        - gcr.io/istio-release/ztunnel:1.21.6
        - gcr.io/istio-release/pilot:1.23.5
        - gcr.io/istio-release/pilot:1.24.1
        - gcr.io/istio-release/proxyv2:1.21.6
        - gcr.io/istio-release/proxyv2:1.25.1
        - gcr.io/istio-release/ztunnel:1.23.4
        - gcr.io/istio-release/proxyv2:1.23.5
        - gcr.io/istio-release/pilot:1.25.2
        - gcr.io/istio-release/install-cni:1.21.6
        - gcr.io/istio-release/ztunnel:1.23.3
        - gcr.io/istio-release/pilot:1.22.5
        - gcr.io/istio-release/proxyv2:1.25.2
        - gcr.io/istio-release/pilot:1.24.4
        - gcr.io/istio-release/pilot:1.23.3
        - gcr.io/istio-release/pilot:1.25.1
        - gcr.io/istio-release/proxyv2:1.23.4
        - gcr.io/istio-release/install-cni:1.22.7
        - gcr.io/istio-release/ztunnel:1.24.3
        - gcr.io/istio-release/ztunnel:1.24.4
        - gcr.io/istio-release/proxyv2:1.22.6
        - gcr.io/istio-release/proxyv2:1.24.2
        - gcr.io/istio-release/ztunnel:1.23.2
        - gcr.io/istio-release/ztunnel:1.24.2
        - gcr.io/istio-release/ztunnel:1.25.2
        - gcr.io/istio-release/install-cni:1.24.4
        - gcr.io/istio-release/pilot:1.22.7
        - gcr.io/istio-release/install-cni:1.25.2
        - gcr.io/istio-release/ztunnel:1.22.7
        - gcr.io/istio-release/proxyv2:1.22.7
        - gcr.io/istio-release/ztunnel:1.22.8
        - gcr.io/istio-release/pilot:1.24.0
        - gcr.io/istio-release/pilot:1.21.6
        - gcr.io/istio-release/install-cni:1.24.1
        - gcr.io/istio-release/install-cni:1.22.5
        - gcr.io/istio-release/install-cni:1.23.2
        - gcr.io/istio-release/install-cni:1.24.0
        - gcr.io/istio-release/ztunnel:1.23.5
        - gcr.io/istio-release/proxyv2:1.24.0
        - gcr.io/istio-release/install-cni:1.22.6
        - gcr.io/istio-release/proxyv2:1.23.2
        - gcr.io/istio-release/proxyv2:1.24.1
        - gcr.io/istio-release/pilot:1.24.2
        - gcr.io/istio-release/proxyv2:1.22.5
        - gcr.io/istio-release/pilot:1.23.4
        - gcr.io/istio-release/ztunnel:1.24.0
        - gcr.io/istio-release/install-cni:1.23.4
        - gcr.io/istio-release/ztunnel:1.24.1
        - gcr.io/istio-release/proxyv2:1.24.4
        - gcr.io/istio-release/proxyv2:1.24.3
        version: 1.25.2
      entries:
      - name: sailoperator.v1.25.2
        version: 1.25.2
      - name: sailoperator.v1.25.1
        version: 1.25.1
      - name: sailoperator.v1.25.0
        version: 1.25.0
      name: stable-1.25
    - currentCSV: sailoperator.v1.26.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.3"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.3"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:d37258646d76a9f708a69fbaa081a0a95908a2b0c8837a14d7ecbfcf505f9bd4
          createdAt: "2025-08-04T06:23:45Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.26-latest
          - v1.26.3
          - v1.26.2
          - v1.26.1
          - v1.26.0
          - v1.25-latest
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/proxyv2@sha256:fd734e6031566b4fb92be38f0f6bb02fdba6c199c45c2db5dc988bbc4fdee026
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/pilot@sha256:c4e33921e030bb22d3b8fabc693851d52c07a7af79a9534addc2f50005fa462a
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - quay.io/sail-dev/sail-operator@sha256:d37258646d76a9f708a69fbaa081a0a95908a2b0c8837a14d7ecbfcf505f9bd4
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/ztunnel@sha256:2d567893e5b452d5c0b9c43a43efd4f8e28e639a344e28d90498f13858f13d46
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/install-cni@sha256:8e782f5077f63c0a15fd4b9a4fa75f8543033c28ccf2457e25afbf1aefb6a72a
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        version: 1.26.3
      entries:
      - name: sailoperator.v1.26.3
        version: 1.26.3
      - name: sailoperator.v1.26.2
        version: 1.26.2
      - name: sailoperator.v1.26.1
        version: 1.26.1
      - name: sailoperator.v1.26.0
        version: 1.26.0
      name: stable-1.26
    - currentCSV: sailoperator.v1.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
          createdAt: "2025-09-05T10:12:19Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        version: 1.27.1
      entries:
      - name: sailoperator.v1.27.1
        version: 1.27.1
      - name: sailoperator.v1.27.0
        version: 1.27.0
      name: stable-1.27
    defaultChannel: stable
    packageName: sailoperator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-websocket-service
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-websocket-service.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "websocket.loadcore.keysight.com/v1alpha1",
                "kind": "Websocket",
                "metadata": {
                  "name": "websocket-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "websocket": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:26:39Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Websocket
            name: websockets.websocket.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Websocket Operator
        displayName: Keysight Websocket Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Websocket Service
          url: https://keysight-websocket-service.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-websocket-service@sha256:c91da3557ddcb8f052c9caa514d2dbf6c22b4963c080117a9d1859f1458fc404
        version: 0.1.21
      entries:
      - name: keysight-websocket-service.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-websocket-service
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/in-en/products/cloud-pak-for-data
    name: ibm-cpd-datastage-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-datastage-operator.v7.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ds.cpd.ibm.com/v1",
                "kind": "DataStage",
                "metadata": {
                  "name": "datastage"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "fileStorageClass": "managed-nfs-storage",
                  "scaleConfig": "small"
                }
              }
            ]
          capabilities: Seamless Upgrades
          cloudpakId: 49a42b864bb94569bef0188ead948f11
          cloudpakName: IBM DataStage Enterprise Plus Cartridge for IBM Cloud Pak
            for Data
          containerImage: registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
          description: The IBM DataStage Operator is used to deploy IBM DataStage
            services
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-datastage-operator
          olm.skipRange: '>=1.0.0 <7.3.0'
          operatorframework.io/arch.amd64: supported
          operatorframework.io/arch.s390x: supported
          operatorframework.io/os.linux: supported
          operatorframework.io/suggested-namespace: ibm-common-services
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation: [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)
              License: [IBM DataStage Enterprise Cartridge for Cloud Pak for Data](https://ibm.biz/Bdahic),
              [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)'
            kind: DataStage
            name: datastages.ds.cpd.ibm.com
            version: v1
          - description: Documentation [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)
              License [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)
            kind: PXRuntime
            name: pxruntimes.ds.cpd.ibm.com
            version: v1
        description: |-
          # Introduction
          Industry-leading batch and real-time data integration service for data engineers to build trusted pipelines (ETL/ELT) for AI and operational use cases across on-premises and hybrid cloud environments.
          With DataStage, you can design and run data flows that move and transform data anywhere, at any scale.
          No matter how complex your data landscape, DataStage can streamline your data movement costs and increase productivity. DataStage offers:
          <ul><li>A best-in-breed parallel processing engine that enables you to process your data where it resides</li><li>Automated job design</li><li>Simple integration with cloud data lakes, real-time data sources, relational databases, big data, and NoSQL data stores</li></ul>
          The DataStage service uses Cloud Pak for Data platform connections and integration points, with services like Data Virtualization, to simplify the process of connecting to and accessing your data.

          With DataStage, Data Engineers can use the simple user interface to build no-code/low-code data pipelines. The interface offers hundreds of functions and connectors that reduce development time and inconsistencies across pipelines. The interface also makes it easy to collaborate with your peers and control access to specific analytics projects.
          The service also provides automatic workload balancing to provide high performance pipelines that make efficient use of available compute resources.

          ## Details
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=SSQNUZ_4.0/svc-welcome/dsent.html)

          ## Documentation

          [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)

          ## Installing
          To install the operator follow the the installation and configuration instructions within the [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=datastage-installing).

          ## Prerequisites

          - OpenShift Version >= 4.6
          - IBM Cloud Pak for Data >= 4.0.2
          - IBM Cloud Pak for Data Common Core Services (CCS) >= 2.0.2 (The DataStage service automatically installs the CCS if they are not already installed.)

          ### Resources Required

          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=requirements-hardware)

          ## Configuration

          To scale DataStage services, you can specify spec.scaleConfig in the DataStage custom resource to small, medium, or large.

          ## License
          [IBM DataStage Enterprise Cartridge for Cloud Pak for Data](https://ibm.biz/Bdahic)
          [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)

          ## Limitations
          - Only the `x86-64` platform is supported by this operator.
          - See [Limitations and known issues in DataStage](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=issues-datastage)

          ## SecurityContextConstraints Requirements
          Custom SecurityContextConstraints definition:
          ```
          apiVersion: security.openshift.io/v1
          kind: SecurityContextConstraints
          allowHostDirVolumePlugin: false
          allowHostIPC: false
          allowHostNetwork: false
          allowHostPID: false
          allowHostPorts: false
          allowPrivilegeEscalation: true
          allowPrivilegedContainer: false
          allowedCapabilities: null
          defaultAddCapabilities: null
          fsGroup:
            type: MustRunAs
          metadata:
            annotations:
              include.release.openshift.io/self-managed-high-availability: "true"
              kubernetes.io/description: restrict denies access to all host features and requires pods to be run with a UID, and SELinux context that are allocated to the namespace.  This is the most restrictive SCC and it is used by default for authenticated users.
              release.openshift.io/create-only: "true"
            name: restrict
          priority: null
          readOnlyRootFilesystem: false
          requiredDropCapabilities:
          - KILL
          - MKNOD
          - SETUID
          - SETGID
          runAsUser:
            type: MustRunAsRange
          seLinuxContext:
            type: MustRunAs
          supplementalGroups:
            type: RunAsAny
          users: []
          volumes:
          - configMap
          - downwardAPI
          - emptyDir
          - persistentVolumeClaim
          - projected
          - secret
          ```
        displayName: IBM DataStage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DataStage
        - Analytics
        - Cloud Pak
        - Data
        - Cloud Pak for Data
        - IBM
        maintainers:
        - email: cp4d@ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/in-en/products/cloud-pak-for-data
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:046eb99c8f1f37dd4cffa1c6c9a93aea50bb5d10a549aa12c3c90134d03c6d1c
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:4604a076c837600cd4b3e29b31e22470915fe5b6b1191ae9401407f8620f2472
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:5f5faf7d15f46d79eea873c12ff631ed59bc55bc57f1f7aee9fd0a8f35fdd3cf
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:6c569c4687546a03a5f4e127acc13b219ce3efa057fbdfd55e96868e549c70cb
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:8b72cafe9de355634bc3218cc1dc701ffb376a29aecb5cd12a5719d43a1304bd
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:3f8133fcd904befa579b99d3a828fe737faf13ae14e842ed0408eeeee72de5fb
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:7015b410b947338dcafde98b68b95cb4199c08511ae46e733691e8fa835ad988
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:3eecc54f26cd87870c201b5ade1d037d2733ccfda40fc1aedd9a2164e582275d
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:001bf35df08cd397397c3fdd1ddbd7953cdf4bde06965427ae640e5608e60ddf
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:1b39e1d5d7c13451252250ceeb29b91c1e7ef3739fc17baf9958e1bf07c4c2fd
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:ef03bb6ed647d92b9c84d34810376da7049aad88589eb2e35798d82a22f5b994
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:10492010b28f57f8badbc7cadbe15d1447269052b387d878c6d9ecbd2f62923b
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:99fc418fb1a09ee6c45e90dfba47f73ea8e1da1b19f9f570d2a749b68118b7a3
        version: 7.3.0
      entries:
      - name: ibm-cpd-datastage-operator.v7.3.0
        version: 7.3.0
      name: v7.3
    defaultChannel: v7.3
    packageName: ibm-cpd-datastage-operator
    provider:
      name: IBM
      url: https://www.ibm.com/in-en/products/cloud-pak-for-data
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Advanced Micro Devices, Inc.
      provider-url: ""
    name: amd-gpu-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amd-gpu-operator.v1.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "amd.com/v1alpha1",
                "kind": "DeviceConfig",
                "metadata": {
                  "name": "test-deviceconfig",
                  "namespace": "openshift-amd-gpu"
                },
                "spec": {
                  "devicePlugin": {
                    "devicePluginImage": "rocm/k8s-device-plugin:rhubi-latest",
                    "nodeLabellerImage": "rocm/k8s-node-labeller:rhubi-latest"
                  },
                  "driver": {
                    "enable": "true",
                    "version": "6.4.2"
                  },
                  "selector": {
                    "feature.node.kubernetes.io/amd-gpu": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning,Monitoring
          containerImage: docker.io/rocm/gpu-operator@sha256:1583d4ec172d50d4416f4030eb85910ec54633b7e8e061d4bd79b14dbf98d28c
          createdAt: "2025-08-01T01:55:48Z"
          description: |-
            Operator responsible for deploying AMD GPU kernel drivers, device plugin, device test runner and device metrics exporter
            For more information, visit [documentation](https://instinct.docs.amd.com/projects/gpu-operator/en/latest/)
          deviceConfigManagerImage: docker.io/rocm/device-config-manager@sha256:d2896271eacfc664ac7651425f46a85af4f4ef98d713988c2a4fee4a846548ee
          devicePluginImage: docker.io/rocm/k8s-device-plugin@sha256:840c4ea1bf58df60ef3e80e3bf27bbf620a393854e42f58b2224b288767e3dd4
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          metricsExporterImage: docker.io/rocm/device-metrics-exporter@sha256:6c946c6ca10d5c0f40e67b5759759f12d51c45e3b934cb1af912821d5fd76e6e
          nodelabellerImage: docker.io/rocm/k8s-node-labeller@sha256:75f916c5a407e40e9ffdd26cdfc65b7da855710568937e740a969094e2991065
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-amd-gpu
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ROCm/gpu-operator
          support: Advanced Micro Devices, Inc.
          testRunnerImage: docker.io/rocm/test-runner@sha256:98e698263b4df26959341b14b3ddded9df51e504e7c6819d5c7c56c434855e9d
          utilsContainerImage: docker.io/rocm/gpu-operator-utils@sha256:9cbe60dce5b3a2ffd93f7148a363ed5f072a320377d5e5bcbbfb1e05d4959287
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DeviceConfig describes how to enable AMD GPU device
            displayName: DeviceConfig
            kind: DeviceConfig
            name: deviceconfigs.amd.com
            version: v1alpha1
        description: |-
          Operator responsible for deploying AMD GPU kernel drivers, device plugin, device test runner and device metrics exporter
          For more information, visit [documentation](https://instinct.docs.amd.com/projects/gpu-operator/en/latest/)
        displayName: amd-gpu-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AMD
        - GPU
        - AI
        - Deep Learning
        - Hardware
        - Driver
        - Monitoring
        links:
        - name: AMD GPU Operator
          url: https://github.com/ROCm/gpu-operator
        maintainers:
        - email: Yan.Sun3@amd.com
          name: Yan Sun
        - email: farshad.ghodsian@amd.com
          name: Farshad Ghodsian
        - email: shrey.ajmera@amd.com
          name: Shrey Ajmera
        maturity: stable
        provider:
          name: Advanced Micro Devices, Inc.
        relatedImages:
        - docker.io/rocm/gpu-operator@sha256:1583d4ec172d50d4416f4030eb85910ec54633b7e8e061d4bd79b14dbf98d28c
        - docker.io/rocm/device-metrics-exporter@sha256:6c946c6ca10d5c0f40e67b5759759f12d51c45e3b934cb1af912821d5fd76e6e
        - docker.io/rocm/k8s-device-plugin@sha256:840c4ea1bf58df60ef3e80e3bf27bbf620a393854e42f58b2224b288767e3dd4
        - docker.io/rocm/k8s-node-labeller@sha256:75f916c5a407e40e9ffdd26cdfc65b7da855710568937e740a969094e2991065
        - docker.io/rocm/gpu-operator-utils@sha256:9cbe60dce5b3a2ffd93f7148a363ed5f072a320377d5e5bcbbfb1e05d4959287
        - docker.io/rocm/test-runner@sha256:98e698263b4df26959341b14b3ddded9df51e504e7c6819d5c7c56c434855e9d
        - docker.io/rocm/device-config-manager@sha256:d2896271eacfc664ac7651425f46a85af4f4ef98d713988c2a4fee4a846548ee
        version: 1.3.2
      entries:
      - name: amd-gpu-operator.v1.3.2
        version: 1.3.2
      - name: amd-gpu-operator.v1.3.1
        version: 1.3.1
      - name: amd-gpu-operator.v1.2.1
        version: 1.2.1
      - name: amd-gpu-operator.v1.1.1
        version: 1.1.1
      name: alpha
    defaultChannel: alpha
    packageName: amd-gpu-operator
    provider:
      name: Advanced Micro Devices, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-zookeeper-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zookeeper-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:1d36865fe1cac88c9edc63c0e35e42a80456486668de5a575f10cc4b445d47db
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:1d36865fe1cac88c9edc63c0e35e42a80456486668de5a575f10cc4b445d47db
        version: 24.11.1
      entries:
      - name: zookeeper-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: zookeeper-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:6a4729289275e473907a3792c001dbe7dd2dc0498aabc719cf4175d956e3c87f
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:6a4729289275e473907a3792c001dbe7dd2dc0498aabc719cf4175d956e3c87f
        version: 25.3.0
      entries:
      - name: zookeeper-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: zookeeper-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
        version: 25.7.0
      entries:
      - name: zookeeper-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: zookeeper-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
        version: 25.7.0
      entries:
      - name: zookeeper-operator.v25.7.0
        version: 25.7.0
      - name: zookeeper-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-zookeeper-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cluster-kube-descheduler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clusterkubedescheduleroperator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "KubeDescheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-kube-descheduler-operator"
                },
                "spec": {
                  "deschedulingIntervalSeconds": 3600,
                  "managementState": "Managed",
                  "mode": "Predictive",
                  "profiles": ["AffinityAndTaints"]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-cluster-kube-descheduler-operator@sha256:ba0b71ff2a30a069b4a8a8f3c1e0898aaadc6db112e4cc12aff7c77ced7a0405
          createdAt: 2021/10/13
          description: An operator to run descheduler in Openshift cluster.
          olm.skipRange: '>=4.10.0 <4.13.0-202309181427'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-kube-descheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeDescheduler is the Schema for the deschedulers API
            displayName: Kube Descheduler
            kind: KubeDescheduler
            name: kubedeschedulers.operator.openshift.io
            version: v1
        description: |
          The Kube Descheduler Operator provides the ability to evict a running pod so that the pod can be rescheduled onto a more suitable node.

          There are several situations where descheduling can benefit your cluster:

          * Nodes are underutilized or overutilized.
          * Pod and node affinity requirements, such as taints or labels, have changed and the original scheduling decisions are no longer appropriate for certain nodes.
          * Node failure requires pods to be moved.
          * New nodes are added to clusters.

          ## Descheduler Profiles

          Once the operator is installed, you can configure one or more profiles to identify pods to evict. The scheduler will schedule the replacement of the evicted pods.

          The following profiles are available:

          * AffinityAndTaints
          * TopologyAndDuplicates
          * SoftTopologyAndDuplicates
          * LifecycleAndUtilization
          * EvictPodsWithPVC
          * EvictPodsWithLocalStorage

          These profiles are documented in detail in the [descheduler operator README](https://github.com/openshift/cluster-kube-descheduler-operator#profiles).

          ## Additional Parameters

          In addition to the profiles, the following parameters can be configured:

          * `deschedulingIntervalSeconds` - Set the number of seconds between descheduler runs. A value of `0` in this field runs the descheduler once and exits.
          * `profileCustomizations` - Allows certain profile parameters to be tweaked, such as `podLifetime` (see [README](https://github.com/openshift/cluster-kube-descheduler-operator#profile-customizations) for more info).
        displayName: Kube Descheduler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - scheduling
        - descheduler
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/cluster-kube-descheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-cluster-kube-descheduler-operator@sha256:ba0b71ff2a30a069b4a8a8f3c1e0898aaadc6db112e4cc12aff7c77ced7a0405
        - registry.redhat.io/openshift4/ose-descheduler@sha256:45dc69ad93ab50bdf9ce1bb79f6d98f849e320db68af30475b10b7f5497a1b13
        version: 4.13.0-202309181427
      entries:
      - name: clusterkubedescheduleroperator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: "4.13"
    - currentCSV: clusterkubedescheduleroperator.v5.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "KubeDescheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-kube-descheduler-operator"
                },
                "spec": {
                  "deschedulingIntervalSeconds": 3600,
                  "managementState": "Managed",
                  "mode": "Predictive",
                  "profiles": ["AffinityAndTaints"]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/kube-descheduler-operator/kube-descheduler-rhel9-operator@sha256:03404d74da227fdf5805f61657272af2c915ce0014d6a902ef384b7a70bd10e2
          createdAt: 2024/10/30
          description: An operator to run descheduler in Openshift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.12.0 <5.0.2'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-kube-descheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeDescheduler is the Schema for the deschedulers API
            displayName: Kube Descheduler
            kind: KubeDescheduler
            name: kubedeschedulers.operator.openshift.io
            version: v1
        description: |
          The Kube Descheduler Operator provides the ability to evict a running pod so that the pod can be rescheduled onto a more suitable node.

          There are several situations where descheduling can benefit your cluster:

          * Nodes are underutilized or overutilized.
          * Pod and node affinity requirements, such as taints or labels, have changed and the original scheduling decisions are no longer appropriate for certain nodes.
          * Node failure requires pods to be moved.
          * New nodes are added to clusters.

          ## Descheduler Profiles

          Once the operator is installed, you can configure one or more profiles to identify pods to evict. The scheduler will schedule the replacement of the evicted pods.

          The following profiles are available:

          * AffinityAndTaints
          * TopologyAndDuplicates
          * SoftTopologyAndDuplicates
          * LifecycleAndUtilization
          * EvictPodsWithPVC
          * EvictPodsWithLocalStorage

          These profiles are documented in detail in the [descheduler operator README](https://github.com/openshift/cluster-kube-descheduler-operator#profiles).

          ## Additional Parameters

          In addition to the profiles, the following parameters can be configured:

          * `deschedulingIntervalSeconds` - Set the number of seconds between descheduler runs. A value of `0` in this field runs the descheduler once and exits.
          * `profileCustomizations` - Allows certain profile parameters to be tweaked, such as `podLifetime` (see [README](https://github.com/openshift/cluster-kube-descheduler-operator#profile-customizations) for more info).
        displayName: Kube Descheduler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - scheduling
        - descheduler
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/cluster-kube-descheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/kube-descheduler-operator/kube-descheduler-rhel9-operator@sha256:03404d74da227fdf5805f61657272af2c915ce0014d6a902ef384b7a70bd10e2
        - registry.redhat.io/kube-descheduler-operator/descheduler-rhel9@sha256:7ba3f46cbbec2e47edcc255fae405bdbf2b9df3a232fc6809ce65fb1827a3b45
        version: 5.0.2
      entries:
      - name: clusterkubedescheduleroperator.v5.0.2
        version: 5.0.2
      - name: clusterkubedescheduleroperator.4.14.0-202311021650
        version: 4.14.0-202311021650
      - name: clusterkubedescheduleroperator.4.14.0-202310201027
        version: 4.14.0-202310201027
      name: stable
    defaultChannel: stable
    packageName: cluster-kube-descheduler-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Weights & Biases
      provider-url: https://wandb.ai
    name: wandb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: wandb-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.wandb.com/v1",
                "kind": "WeightsAndBiases",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "wandb-operator",
                    "app.kubernetes.io/instance": "weightsandbiases-sample",
                    "app.kubernetes.io/managed-by": "olm",
                    "app.kubernetes.io/name": "weightsandbiases",
                    "app.kubernetes.io/part-of": "wandb-operator"
                  },
                  "name": "weightsandbiases-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning,Developer Tools
          certified: "false"
          containerImage: quay.io/wandb_tools/wandb-k8s-operator:1.0.0
          createdAt: "2024-10-10T14:46:54Z"
          description: Operator for Weights and Biases, enabling seamless integration
            and management of machine learning experiments in Kubernetes environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: wandb-system
          repository: https://github.com/wandb/operator
          support: WeightsAndBiases
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: WeightsAndBiases is a custom resource for managing W&B configurations.
            displayName: Weights And Biases
            kind: WeightsAndBiases
            name: weightsandbiases.apps.wandb.com
            version: v1
        description: |-
          The Weights and Biases Operator simplifies the management of machine learning experiments, tracking, and infrastructure. It integrates with Kubernetes to deploy and manage Weights and Biases instances.
          Features:
          - Experiment Tracking
          - Hyperparameter Tuning
          - Model Management
          - Integration with popular ML frameworks
        displayName: Weights & Biases Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - machine learning
        - experiment tracking
        - ML infrastructure
        - weights and biases
        links:
        - name: Weights & Biases Documentation
          url: https://docs.wandb.ai/
        maintainers:
        - email: abhinav.garg@domain.com
          name: Abhinav Garg
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Weights & Biases
          url: https://wandb.ai
        relatedImages:
        - quay.io/wandb_tools/wandb-k8s-operator:1.0.0
        version: 1.0.0
      entries:
      - name: wandb-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: wandb-operator
    provider:
      name: Weights & Biases
      url: https://wandb.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: keepalived-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keepalived-operator.v1.5.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KeepalivedGroup",
                "metadata": {
                  "name": "keepalivedgroup-workers"
                },
                "spec": {
                  "blacklistRouterIDs": [
                    1,
                    2
                  ],
                  "image": "registry.redhat.io/openshift4/ose-keepalived-ipfailover",
                  "interface": "ens3",
                  "nodeSelector": {
                    "node-role.kubernetes.io/loadbalancer": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Networking, Cloud Provider
          certified: "false"
          containerImage: quay.io/redhat-cop/keepalived-operator@sha256:cd0c2f7df82df1178795e4a18a0942da33adb9670b192a7b7a2b1483a5b4ccf8
          createdAt: "2023-03-11T18:48:22Z"
          description: This operator provides a simple way to create self-hosted load
            balancers using keepalived.
          operatorframework.io/suggested-namespace: keepalived-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/keepalived-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KeepalivedGroup
            name: keepalivedgroups.redhatcop.redhat.io
            version: v1alpha1
        description: "The objective of the keepalived operator is to allow for a way
          to create self-hosted load balancers in an automated way. From a user experience
          point of view the behavior is the same as of when creating [`LoadBalancer`](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          services with a cloud provider able to manage them.\n\nThe keepalived operator
          can be used in all environments that allows nodes to advertise additional
          IPs on their NICs (and at least for now, in networks that allow multicast),
          however it's mainly aimed at supporting LoadBalancer services and ExternalIPs
          on bare metal installations (or other installation environments where a
          cloud provider is not available).\n\nOne possible use of the keepalived
          operator is also to support [OpenShift Ingresses](https://docs.openshift.com/container-platform/4.5/networking/configuring_ingress_cluster_traffic/overview-traffic.html)
          in environments where an external load balancer cannot be provisioned.\n\n##
          How it works\n\nThe keepalived operator will create one or more [VIPs](https://en.wikipedia.org/wiki/Virtual_IP_address)
          (an HA IP that floats between multiple nodes), based on the [`LoadBalancer`](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          services and/or services requesting [`ExternalIPs`](https://kubernetes.io/docs/concepts/services-networking/service/#external-ips).\n\nFor
          `LoadBalancer` services the IPs found at `.Status.LoadBalancer.Ingress[].IP`
          will become VIPs.\n\nFor services requesting a `ExternalIPs`, the IPs found
          at `.Spec.ExternalIPs[]` will become VIPs.\n\nNote that a service can be
          of `LoadBalancer` type and also request `ExternalIPs`, it this case both
          sets of IPs will become VIPs.\n\nDue to a [keepalived](https://www.keepalived.org/manpage.html)
          limitation a single keepalived cluster can manage up to 256 VIP configurations.
          Multiple keepalived clusters can coexists in the same network as long as
          they use different multicast ports [TODO verify this statement].\n\nTo address
          this limitation the `KeepalivedGroup` [CRD](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/)
          has been introduced. This CRD is supposed to be configured by an administrator
          and allows you to specify a node selector to pick on which nodes the keepalived
          pods should be deployed. Here is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          KeepalivedGroup\nmetadata:\n  name: keepalivedgroup-router\nspec:\n  image:
          registry.redhat.io/openshift4/ose-keepalived-ipfailover\n  interface: ens3\n
          \ nodeSelector:\n    node-role.kubernetes.io/loadbalancer: \"\"\n  blacklistRouterIDs:\n
          \ - 1\n  - 2  \n```\n\nThis KeepalivedGroup will be deployed on all the
          nodes with role `loadbalancer`. One must also specify the network device
          on which the VIPs will be exposed, it is assumed that all the nodes have
          the same network device configuration.\n\nServices must be annotated to
          opt-in to being observed by the keepalived operator and to specify which
          KeepalivedGroup they refer to. The annotation looks like this:\n\n`keepalived-operator.redhat-cop.io/keepalivedgroup:
          <keepalivedgroup namespace>/<keepalivedgroup-name>`\n\nThe image used for
          the keepalived containers can be specified with `.Spec.Image` it will default
          to `registry.redhat.io/openshift4/ose-keepalived-ipfailover` if undefined.
          \n\n## Requirements\n\n### Security Context Constraints\n\nEach KeepalivedGroup
          deploys a [daemonset](https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/)
          that requires the [privileged scc](https://docs.openshift.com/container-platform/4.5/authentication/managing-security-context-constraints.html),
          this permission must be given to the `default` service account in the namespace
          where the keepalived group is created by and administrator.\n\n```shell\noc
          adm policy add-scc-to-user privileged -z default -n keepalived-operator\n```\n\n###
          Cluster Network Operator\n\nIn Openshift, use of an external IP address
          is governed by the following fields in the `Network.config.openshift.io`
          CR named `cluster`\n\n* `spec.externalIP.autoAssignCIDRs` defines an IP
          address block used by the load balancer when choosing an external IP address
          for the service. OpenShift supports only a single IP address block for automatic
          assignment.\n\n* `spec.externalIP.policy` defines the permissible IP address
          blocks when manually specifying an IP address. OpenShift does not apply
          policy rules to IP address blocks defined by `spec.externalIP.autoAssignCIDRs`\n\nThe
          following patch can be used to configure the Cluster Network Operator:\n\n```yaml\nspec:\n
          \ externalIP:\n    policy:\n      allowedCIDRs:\n      - ${ALLOWED_CIDR}\n
          \   autoAssignCIDRs:\n      - \"${AUTOASSIGNED_CIDR}\"\n```\n\nHere is an
          example of how to apply the patch:\n\n```shell\nexport ALLOWED_CIDR=\"192.168.131.128/26\"\nexport
          AUTOASSIGNED_CIDR=\"192.168.131.192/26\"\noc patch network cluster -p \"$(envsubst
          < ./network-patch.yaml | yq r -j -)\" --type=merge\n```\nAdditionally, the
          fields can be edited manually via `oc edit Network.config.openshift.io cluster`\n\n##
          Blacklisting router IDs\n\nIf the Keepalived pods are deployed on nodes
          which are in the same network (same broadcast domain to be precise) with
          other keepalived the process, it's necessary to ensure that there is no
          collision between the used routers it.\nFor this purpose it is possible
          to provide a `blacklistRouterIDs` field with a list of black-listed IDs
          that will not be used.\n\n## OpenShift RHV, vSphere, OSP and bare metal
          IPI instructions\n\nWhen IPI is used for RHV, vSphere, OSP or bare metal
          platforms, three keepalived VIPs are deployed. To make sure that keepalived-operator
          can work in these environment we need to discover and blacklist the corresponding
          VRRP router IDs.\n\nTo discover the VRRP router IDs being used, run the
          following command, you can run this command from you laptop:\n\n```shell\npodman
          run quay.io/openshift/origin-baremetal-runtimecfg:4.5 vr-ids <cluster_name>\n```\n\nIf
          you don't know your cluster name, run this command:\n\n```shell\npodman
          run quay.io/openshift/origin-baremetal-runtimecfg:4.5 vr-ids $(oc get cm
          cluster-config-v1 -n kube-system -o jsonpath='{.data.install-config}'| yq
          -r .metadata.name)\n```\n\nThen use these [instructions](#Blacklisting-router-IDs)
          to blacklist those VRRP router IDs.\n\n## Verbatim Configurations\n\nKeepalived
          has dozens of [configurations](https://www.keepalived.org/manpage.html).
          At the early stage of this project it's difficult to tell which one should
          be modeled in the API. Yet, users of this project may still need to use
          them. To account for that there is a way to pass verbatim options both at
          the keepalived group level (which maps to the keepalived config `global_defs`
          section) and at the service level (which maps to the keepalived config `vrrp_instance`
          section).\n\nKeepalivedGroup-level verbatim configurations can be passed
          as in the following example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          KeepalivedGroup\nmetadata:\n  name: keepalivedgroup-router\nspec:\n  interface:
          ens3\n  nodeSelector:\n    node-role.kubernetes.io/loadbalancer: \"\"\n
          \ verbatimConfig:  \n    vrrp_iptables: my-keepalived\n```\n\nthis will
          map to the following `global_defs`:\n\n```\n    global_defs {\n        router_id
          keepalivedgroup-router\n        vrrp_iptables my-keepalived\n    }\n```\n\nService-level
          verbatim configurations can be passed as in the following example:\n\n```yaml\napiVersion:
          v1\nkind: Service\nmetadata:\n  annotations:\n    keepalived-operator.redhat-cop.io/keepalivedgroup:
          keepalived-operator/keepalivedgroup-router\n    keepalived-operator.redhat-cop.io/verbatimconfig:
          '{ \"track_src_ip\": \"\" }'\n```\n\nthis will map to the following `vrrp_instance`
          section\n\n```\n    vrrp_instance openshift-ingress/router-default {\n        interface
          ens3\n        virtual_router_id 1  \n        virtual_ipaddress {\n          192.168.131.129\n
          \       }\n        track_src_ip\n    }\n```\n\n## Metrics collection\n\nEach
          keepalived pod exposes a [Prometheus](https://prometheus.io/) metrics port
          at `9650`. Metrics are collected with [keepalived_exporter](github.com/gen2brain/keepalived_exporter),
          the available metrics are described in the project documentation.\n\nWhen
          a keepalived group is created a [`PodMonitor`](https://github.com/coreos/prometheus-operator/blob/master/Documentation/api.md#podmonitor)
          rule to collect those metrics. All PodMonitor resources created that way
          have the label: `metrics: keepalived`. It is up to you to make sure your
          Prometheus instance watches for those `PodMonitor` rules. Here is an example
          of a fragment of a `Prometheus` CR configured to collect the keepalived
          pod metrics:\n\n```yaml\n  podMonitorSelector:\n    matchLabels:\n      metrics:
          keepalived\n```\n"
        displayName: Keepalived Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadbalancer
        - keepalived
        - ipfailover
        - externalIP
        - self-hosted
        links:
        - name: Keepalived Operator
          url: https://github.com/redhat-cop/keepalived-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/keepalived-operator
        maintainers:
        - email: rspazzol@redhat.com'
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/keepalived-operator@sha256:cd0c2f7df82df1178795e4a18a0942da33adb9670b192a7b7a2b1483a5b4ccf8
        version: 1.5.4
      entries:
      - name: keepalived-operator.v1.5.4
        version: 1.5.4
      - name: keepalived-operator.v1.5.3
        version: 1.5.3
      - name: keepalived-operator.v1.5.2
        version: 1.5.2
      - name: keepalived-operator.v1.5.1
        version: 1.5.1
      - name: keepalived-operator.v1.5.0
        version: 1.5.0
      - name: keepalived-operator.v1.4.2
        version: 1.4.2
      - name: keepalived-operator.v1.4.1
        version: 1.4.1
      - name: keepalived-operator.v1.4.0
        version: 1.4.0
      - name: keepalived-operator.v1.3.4
        version: 1.3.4
      - name: keepalived-operator.v1.3.3
        version: 1.3.3
      - name: keepalived-operator.v1.3.2
        version: 1.3.2
      - name: keepalived-operator.v1.3.1
        version: 1.3.1
      - name: keepalived-operator.v1.3.0
        version: 1.3.0
      - name: keepalived-operator.v1.2.2
        version: 1.2.2
      - name: keepalived-operator.v1.2.1
        version: 1.2.1
      - name: keepalived-operator.v1.2.0
        version: 1.2.0
      - name: keepalived-operator.v1.1.0
        version: 1.1.0
      - name: keepalived-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: keepalived-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel Corporation
      provider-url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
    name: sriov-fec
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sriov-fec.v2.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecClusterConfig",
                "metadata": {
                  "name": "acc100SampleConfig",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "acceleratorSelector": {
                    "deviceID": "someDevice",
                    "driver": "someDriver",
                    "maxVirtualFunctions": 2,
                    "pciAddress": "somePciAddress",
                    "vendorID": "someVendor"
                  },
                  "drainSkip": false,
                  "nodeSelector": {
                    "expectedLabel1": "valueOfExpectedLabel1",
                    "expectedLabelN": "valueOfExpectedLabelN"
                  },
                  "physicalFunction": {
                    "bbDevConfig": {
                      "acc100": {
                        "downlink4G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 0
                        },
                        "downlink5G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 4
                        },
                        "maxQueueSize": 1024,
                        "numVfBundles": 16,
                        "pfMode": true,
                        "uplink4G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 0
                        },
                        "uplink5G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 4
                        }
                      }
                    },
                    "pfDriver": "pci-pf-stub",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  },
                  "priority": 100
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecClusterConfig",
                "metadata": {
                  "name": "n3000SampleConfig",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "acceleratorSelector": {
                    "deviceID": "someDevice",
                    "driver": "someDriver",
                    "maxVirtualFunctions": 2,
                    "pciAddress": "somePciAddress",
                    "vendorID": "someVendor"
                  },
                  "drainSkip": false,
                  "nodeSelector": {
                    "expectedLabel1": "valueOfExpectedLabel1",
                    "expectedLabelN": "valueOfExpectedLabelN"
                  },
                  "physicalFunction": {
                    "bbDevConfig": {
                      "n3000": {
                        "downlink": {
                          "bandwidth": 3,
                          "loadBalance": 128,
                          "queues": {
                            "vf0": 16,
                            "vf1": 16,
                            "vf2": 0,
                            "vf3": 0,
                            "vf4": 0,
                            "vf5": 0,
                            "vf6": 0,
                            "vf7": 0
                          }
                        },
                        "flrTimeout": 610,
                        "networkType": "FPGA_5GNR",
                        "pfMode": true,
                        "uplink": {
                          "bandwidth": 3,
                          "loadBalance": 128,
                          "queues": {
                            "vf0": 16,
                            "vf1": 16,
                            "vf2": 0,
                            "vf3": 0,
                            "vf4": 0,
                            "vf5": 0,
                            "vf6": 0,
                            "vf7": 0
                          }
                        }
                      }
                    },
                    "pfDriver": "pci-pf-stub",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  },
                  "priority": 100
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecNodeConfig",
                "metadata": {
                  "name": "acc100-worker",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "drainSkip": false,
                  "physicalFunctions": [
                    {
                      "bbDevConfig": {
                        "acc100": {
                          "downlink4G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 0
                          },
                          "downlink5G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 4
                          },
                          "maxQueueSize": 1024,
                          "numVfBundles": 16,
                          "pfMode": true,
                          "uplink4G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 0
                          },
                          "uplink5G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 4
                          }
                        }
                      },
                      "pci_addr": "somePciAddress",
                      "pf_driver": "pci-pf-stub",
                      "vf_amount": 2,
                      "vf_driver": "vfio-pci"
                    }
                  ]
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecNodeConfig",
                "metadata": {
                  "name": "n3000-worker",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "drainSkip": false,
                  "physicalFunctions": [
                    {
                      "bbDevConfig": {
                        "n3000": {
                          "downlink": {
                            "bandwidth": 3,
                            "loadBalance": 128,
                            "queues": {
                              "vf0": 16,
                              "vf1": 16,
                              "vf2": 0,
                              "vf3": 0,
                              "vf4": 0,
                              "vf5": 0,
                              "vf6": 0,
                              "vf7": 0
                            }
                          },
                          "flrTimeout": 610,
                          "networkType": "FPGA_5GNR",
                          "pfMode": true,
                          "uplink": {
                            "bandwidth": 3,
                            "loadBalance": 128,
                            "queues": {
                              "vf0": 16,
                              "vf1": 16,
                              "vf2": 0,
                              "vf3": 0,
                              "vf4": 0,
                              "vf5": 0,
                              "vf6": 0,
                              "vf7": 0
                            }
                          }
                        }
                      },
                      "pci_addr": "somePciAddress",
                      "pf_driver": "pci-pf-stub",
                      "vf_amount": 2,
                      "vf_driver": "vfio-pci"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/intel/sriov-fec-operator@sha256:adb6a4c2da6399c8ed40c906f790bce13bfcefae3326a047f6a50592b3f5e485
          createdAt: "2025-01-15"
          description: An operator for Intel Wireless FEC Accelerator to orchestrate
            and manage the resources/devices exposed by a range of Intel's vRAN FEC
            acceleration devices/hardware within the OpenShift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/intel/sriov-fec-operator
          support: Intel Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovFecClusterConfig is the Schema for the sriovfecclusterconfigs
              API
            displayName: SriovFecClusterConfig
            kind: SriovFecClusterConfig
            name: sriovfecclusterconfigs.sriovfec.intel.com
            version: v2
          - description: SriovFecNodeConfig is the Schema for the sriovfecnodeconfigs
              API
            displayName: SriovFecNodeConfig
            kind: SriovFecNodeConfig
            name: sriovfecnodeconfigs.sriovfec.intel.com
            version: v2
          - description: SriovVrbClusterConfig is the Schema for the sriovvrbclusterconfigs
              API
            displayName: SriovVrbClusterConfig
            kind: SriovVrbClusterConfig
            name: sriovvrbclusterconfigs.sriovvrb.intel.com
            version: v1
          - description: SriovVrbNodeConfig is the Schema for the sriovvrbnodeconfigs
              API
            displayName: SriovVrbNodeConfig
            kind: SriovVrbNodeConfig
            name: sriovvrbnodeconfigs.sriovvrb.intel.com
            version: v1
        description: The Intel vRAN Boost family of accelerators ACC100/VRB1/VRB2,
          to offload and accelerate the computing-intensive process of forward error
          correction (FEC) and more for 4G/LTE and 5G technology, freeing up processing
          power. It allows the optimization of data plane performance to reduce total
          cost of ownership while maintaining a high degree of flexibility.  The Intel
          vRAN Boost accelerators ACC100/VRB1/VRB2 plays a key role in accelerating
          5G and network functions virtualization (NFV) adoption for ecosystem partners
          such as telecommunications equipment manufacturers (TEMs) virtual network
          functions (VNF) vendors, system integrators and telcos, to bring scalable
          and high-performance solutions to market. The Intel ACC100 includes a variant
          that is design to be Network Equipment Building System (NEBS)-friendly,
          and features a Root-of-Trust device that helps protect systems from FPGA
          host security exploits.
        displayName: SRIOV-FEC Operator for Intel vRAN Boost Accelerators
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ACC100
        - vRAN
        - ORAN
        links:
        - name: SRIOV-FEC source code
          url: https://github.com/intel/sriov-fec-operator
        maturity: alpha
        provider:
          name: Intel Corporation
          url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
        relatedImages:
        - registry.connect.redhat.com/intel/sriov-fec-daemon@sha256:5b0c60b67bcf5d7b2797deab0d5e3fa079577179f0acc6b0062533584ea0bc23
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1ed3b514bf9af6ac0d386e3739874d1d964e667d54e041c2d67edd769b0009fb
        - registry.connect.redhat.com/intel/sriov-fec-operator@sha256:adb6a4c2da6399c8ed40c906f790bce13bfcefae3326a047f6a50592b3f5e485
        - registry.connect.redhat.com/intel/n3000-labeler@sha256:b417cfcedc998a35491784005cc7eee3d9c647b84eac5ca41cb459076870b8f6
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin@sha256:d0fcb6e9135d9a19ea26851bcdb4ac72c04bd747706fc5123020460e9dbfc21b
        version: 2.11.1
      entries:
      - name: sriov-fec.v2.11.1
        version: 2.11.1
      - name: sriov-fec.v2.11.0
        version: 2.11.0
      - name: sriov-fec.v2.10.0
        version: 2.10.0
      - name: sriov-fec.v2.9.0
        version: 2.9.0
      - name: sriov-fec.v2.8.0
        version: 2.8.0
      - name: sriov-fec.v2.7.2
        version: 2.7.2
      - name: sriov-fec.v2.7.1
        version: 2.7.1
      name: stable
    defaultChannel: stable
    packageName: sriov-fec
    provider:
      name: Intel Corporation
      url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: spyre-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spyre-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreClusterPolicy",
                "metadata": {
                  "name": "spyreclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2",
                      "spyreFilter": ".",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "latest"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "spyre-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "spyre-device-plugin-init",
                      "repository": "icr.io/ibmaiu_internal",
                      "version": "0.1.1"
                    },
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "spyre-exporter",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.2"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "spyre-webhook-validator",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  },
                  "scheduler": {
                    "image": "spyre-scheduler",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  }
                }
              },
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu_internal/spyre-operator@sha256:0f657c98668a8358c37b3fc2ed3f31ec78b7ef7f9a9678e7cd35d383d22ce2da
          createdAt: "2025-07-31T10:32:48Z"
          description: Operator for IBM Spyre devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM Spyre
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["spyrenodestates.spyre.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpyreClusterPolicy is the Schema for the Spyre API
            displayName: Spyre Cluster Policy
            kind: SpyreClusterPolicy
            name: spyreclusterpolicies.spyre.ibm.com
            version: v1alpha1
          - description: SpyreNodeState is the Schema for the SpyreNodeState API
            displayName: Spyre Node State
            kind: SpyreNodeState
            name: spyrenodestates.spyre.ibm.com
            version: v1alpha1
        description: Operator for IBM Spyre devices
        displayName: IBM Spyre Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: IBM Spyre Operator
          url: https://spyre-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: Satyam.Gupta.3@ibm.com
          name: Satyam Gupta
        - email: Rehan.Khan7@ibm.com
          name: Rehan Khan
        - email: Nishan.Acharya@ibm.com
          name: Nishan Acharya
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu_internal/spyre-operator@sha256:0f657c98668a8358c37b3fc2ed3f31ec78b7ef7f9a9678e7cd35d383d22ce2da
        - icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2
        version: 0.1.2
      entries:
      - name: spyre-operator.v0.1.2
        version: 0.1.2
      name: stable-v0.1
    - currentCSV: spyre-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreClusterPolicy",
                "metadata": {
                  "name": "spyreclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb",
                      "spyreFilter": ".",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "latest"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "spyre-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "spyre-device-plugin-init",
                      "repository": "quay.io/ibm-aiu",
                      "version": "0.2.0"
                    },
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "spyre-exporter",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "spyre-webhook-validator",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "scheduler": {
                    "image": "spyre-scheduler",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.1"
                  }
                }
              },
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/ibm-aiu/spyre-operator@sha256:d0cdd449f1d8f0dd9281c4d9a809333db6c4ed201bb14187b6c70ae686a30e99
          createdAt: "2025-09-02T13:25:34Z"
          description: Operator for IBM Spyre devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM Spyre
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["spyrenodestates.spyre.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpyreClusterPolicy is the Schema for the Spyre API
            displayName: Spyre Cluster Policy
            kind: SpyreClusterPolicy
            name: spyreclusterpolicies.spyre.ibm.com
            version: v1alpha1
          - description: SpyreNodeState is the Schema for the SpyreNodeState API
            displayName: Spyre Node State
            kind: SpyreNodeState
            name: spyrenodestates.spyre.ibm.com
            version: v1alpha1
        description: Operator for IBM Spyre devices
        displayName: IBM Spyre Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: IBM Spyre Operator
          url: https://spyre-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: Satyam.Gupta.3@ibm.com
          name: Satyam Gupta
        - email: Rehan.Khan7@ibm.com
          name: Rehan Khan
        - email: Nishan.Acharya@ibm.com
          name: Nishan Acharya
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - quay.io/ibm-aiu/spyre-operator@sha256:d0cdd449f1d8f0dd9281c4d9a809333db6c4ed201bb14187b6c70ae686a30e99
        version: 0.2.1
      entries:
      - name: spyre-operator.v0.2.1
        version: 0.2.1
      name: stable-v0.2
    - currentCSV: spyre-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreClusterPolicy",
                "metadata": {
                  "name": "spyreclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "config": {
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb",
                      "spyreFilter": ".",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "latest"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "spyre-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "spyre-device-plugin-init",
                      "repository": "quay.io/ibm-aiu",
                      "version": "1.0.0"
                    },
                    "repository": "quay.io/ibm-aiu",
                    "version": "1.0.0"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "spyre-exporter",
                    "repository": "quay.io/ibm-aiu",
                    "version": "1.0.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "spyre-webhook-validator",
                    "repository": "quay.io/ibm-aiu",
                    "version": "1.0.0"
                  },
                  "scheduler": {
                    "image": "spyre-scheduler",
                    "repository": "quay.io/ibm-aiu",
                    "version": "1.0.0"
                  }
                }
              },
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/ibm-aiu/spyre-operator@sha256:dd262e8c65f35b12c5d79571e8b5ddd54dd0ee2693f28bea05a6fc02ac1fe6f0
          createdAt: "2025-10-01T00:40:23Z"
          description: Operator for IBM Spyre devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: spyre-operator
          operators.openshift.io/valid-subscription: use with IBM Spyre
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["spyrenodestates.spyre.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpyreClusterPolicy is the Schema for the Spyre API
            displayName: Spyre Cluster Policy
            kind: SpyreClusterPolicy
            name: spyreclusterpolicies.spyre.ibm.com
            version: v1alpha1
          - description: SpyreNodeState is the Schema for the SpyreNodeState API
            displayName: Spyre Node State
            kind: SpyreNodeState
            name: spyrenodestates.spyre.ibm.com
            version: v1alpha1
        description: Operator for IBM Spyre devices
        displayName: IBM Spyre Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: IBM Spyre Operator
          url: https://spyre-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: Satyam.Gupta.3@ibm.com
          name: Satyam Gupta
        - email: Rehan.Khan7@ibm.com
          name: Rehan Khan
        - email: Nishan.Acharya@ibm.com
          name: Nishan Acharya
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - quay.io/ibm-aiu/spyre-operator@sha256:dd262e8c65f35b12c5d79571e8b5ddd54dd0ee2693f28bea05a6fc02ac1fe6f0
        version: 1.0.0
      entries:
      - name: spyre-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: spyre-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hive-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hive-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:e71fc728632622f7ea97357ec361dd2bb73e63c0707e8014ca8ecf8ec4ab93f1
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:e71fc728632622f7ea97357ec361dd2bb73e63c0707e8014ca8ecf8ec4ab93f1
        version: 24.11.1
      entries:
      - name: hive-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: hive-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:5530f5034cd83c2e8b860fce737ac491a2adc6d22246356bee0c79768ed64706
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:5530f5034cd83c2e8b860fce737ac491a2adc6d22246356bee0c79768ed64706
        version: 25.3.0
      entries:
      - name: hive-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: hive-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
        version: 25.7.0
      entries:
      - name: hive-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hive-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
        version: 25.7.0
      entries:
      - name: hive-operator.v25.7.0
        version: 25.7.0
      - name: hive-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hive-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CSI Addons Community
      provider-url: ""
    name: odf-csi-addons-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-csi-addons-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          createdAt: "2025-09-12T07:33:01Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIAddonsNode is the Schema for the csiaddonsnode API
            displayName: CSIAddons Node
            kind: CSIAddonsNode
            name: csiaddonsnodes.csiaddons.openshift.io
            version: v1alpha1
          - description: NetworkFence is the Schema for the networkfences API
            displayName: Network Fence
            kind: NetworkFence
            name: networkfences.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceCronJob is the Schema for the reclaimspacecronjobs
              API
            displayName: Reclaim Space CronJob
            kind: ReclaimSpaceCronJob
            name: reclaimspacecronjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceJob is the Schema for the reclaimspacejobs API
            displayName: Reclaim Space Job
            kind: ReclaimSpaceJob
            name: reclaimspacejobs.csiaddons.openshift.io
            version: v1alpha1
          - description: VolumeReplicationClass is the Schema for the volumereplicationclasses
              API
            displayName: Volume Replication Class
            kind: VolumeReplicationClass
            name: volumereplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplication is the Schema for the volumereplications
              API
            displayName: Volume Replication
            kind: VolumeReplication
            name: volumereplications.replication.storage.openshift.io
            version: v1alpha1
        description: CSI Addons provides the CSI Addons Controller that enables advanced
          storage operations for CSI-drivers.
        displayName: CSI Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: CSI Addons
          url: https://csi-addons.github.io
        maintainers:
        - email: csi-addons@redhat.com
          name: CSI Addons Community
        maturity: alpha
        provider:
          name: CSI Addons Community
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/odf4/odf-csi-addons-rhel9-operator@sha256:8b0c16c21fccca9b674c4d980a793b6cde482b86363b793e173126b030bdcc55
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:9a0008f662b6afa9e2677a8a83bd662d806ad6f4d03fb49bccb78ed487ce4c7a
        version: 4.15.20-rhodf
      entries:
      - name: odf-csi-addons-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odf-csi-addons-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odf-csi-addons-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odf-csi-addons-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odf-csi-addons-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odf-csi-addons-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odf-csi-addons-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odf-csi-addons-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odf-csi-addons-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odf-csi-addons-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odf-csi-addons-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odf-csi-addons-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odf-csi-addons-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odf-csi-addons-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odf-csi-addons-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odf-csi-addons-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odf-csi-addons-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odf-csi-addons-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odf-csi-addons-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odf-csi-addons-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odf-csi-addons-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          createdAt: "2025-09-12T06:32:07Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIAddonsNode is the Schema for the csiaddonsnode API
            displayName: CSIAddons Node
            kind: CSIAddonsNode
            name: csiaddonsnodes.csiaddons.openshift.io
            version: v1alpha1
          - description: NetworkFence is the Schema for the networkfences API
            displayName: Network Fence
            kind: NetworkFence
            name: networkfences.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceCronJob is the Schema for the reclaimspacecronjobs
              API
            displayName: Reclaim Space CronJob
            kind: ReclaimSpaceCronJob
            name: reclaimspacecronjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceJob is the Schema for the reclaimspacejobs API
            displayName: Reclaim Space Job
            kind: ReclaimSpaceJob
            name: reclaimspacejobs.csiaddons.openshift.io
            version: v1alpha1
          - description: VolumeReplicationClass is the Schema for the volumereplicationclasses
              API
            displayName: Volume Replication Class
            kind: VolumeReplicationClass
            name: volumereplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplication is the Schema for the volumereplications
              API
            displayName: Volume Replication
            kind: VolumeReplication
            name: volumereplications.replication.storage.openshift.io
            version: v1alpha1
        description: CSI Addons provides the CSI Addons Controller that enables advanced
          storage operations for CSI-drivers.
        displayName: CSI Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: CSI Addons
          url: https://csi-addons.github.io
        maintainers:
        - email: csi-addons@redhat.com
          name: CSI Addons Community
        maturity: alpha
        provider:
          name: CSI Addons Community
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/odf4/odf-csi-addons-rhel9-operator@sha256:d370f19578d1c466d39d00da82f1f3fe3d9f13d7021ed0c39f759b24b0ed7573
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:0afcb516eb528857887c6ee1a9cba081dc8f6b7d47d9f26a0dbb2dca0cefe07c
        version: 4.16.16-rhodf
      entries:
      - name: odf-csi-addons-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odf-csi-addons-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odf-csi-addons-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odf-csi-addons-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odf-csi-addons-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odf-csi-addons-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odf-csi-addons-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odf-csi-addons-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odf-csi-addons-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odf-csi-addons-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odf-csi-addons-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odf-csi-addons-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odf-csi-addons-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odf-csi-addons-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odf-csi-addons-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odf-csi-addons-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odf-csi-addons-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odf-csi-addons-operator
    provider:
      name: CSI Addons Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp, Inc.
      provider-url: https://www.netapp.com/
    name: trident-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trident-operator.v25.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "trident.netapp.io/v1",
                "kind": "TridentOrchestrator",
                "metadata": {
                  "name": "trident"
                },
                "spec": {
                  "IPv6": false,
                  "debug": true,
                  "nodePrep": [],
                  "imagePullSecrets": [],
                  "imageRegistry": "",
                  "namespace": "trident",
                  "silenceAutosupport": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: docker.io/netapp/trident-operator:25.06.2
          createdAt: "2025-08-03"
          description: Trident Operator, to manage NetApp Trident installations
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          repository: https://github.com/netapp/trident
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Used to deploy Astra Trident.
            displayName: Trident Orchestrator
            kind: TridentOrchestrator
            name: tridentorchestrators.trident.netapp.io
            version: v1
          - description: Automates AWS FSxN backend configuration
            displayName: Trident Configurator
            kind: TridentConfigurator
            name: tridentconfigurators.trident.netapp.io
            version: v1
        description: |
          NetApp Trident is an open source storage provisioner and orchestrator maintained by NetApp. It enables you to create storage volumes for containerized applications managed by Docker and Kubernetes. For full release information, including patch release changes, see https://docs.netapp.com/us-en/trident/trident-rn.html.
        displayName: NetApp Trident
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - NetApp
        - ONTAP
        - Trident
        - storage
        - performance
        - open source
        - Astra
        - Cloud Native
        - AWS
        - Azure
        - GCP
        - Cloud Volumes
        links:
        - name: GitHub Repository
          url: https://github.com/NetApp/trident
        - name: Trident documentation
          url: https://docs.netapp.com/us-en/trident/index.html
        - name: Support policy
          url: https://mysupport.netapp.com/site/info/version-support
        - name: Release Notes
          url: https://docs.netapp.com/us-en/trident/trident-rn.html
        maintainers:
        - email: ng-astra-product@netapp.com
          name: NetApp
        provider:
          name: NetApp, Inc.
          url: https://www.netapp.com/
        relatedImages:
        - docker.io/netapp/trident-operator:25.06.2
        version: 25.6.2
      entries:
      - name: trident-operator.v25.6.2
        version: 25.6.2
      - name: trident-operator.v25.6.1
        version: 25.6.1
      - name: trident-operator.v25.6.0
        version: 25.6.0
      - name: trident-operator.v25.2.1
        version: 25.2.1
      - name: trident-operator.v25.2.0
        version: 25.2.0
      - name: trident-operator.v24.10.1
        version: 24.10.1
      - name: trident-operator.v24.10.0
        version: 24.10.0
      - name: trident-operator.v24.6.0
        version: 24.6.0
      - name: trident-operator.v24.2.0
        version: 24.2.0
      name: stable
    defaultChannel: stable
    packageName: trident-operator
    provider:
      name: NetApp, Inc.
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: multicluster-operators-subscription
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-operators-subscription.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.3.0
          createdAt: "2021-07-19T18:00:00Z"
          description: An operator to subscribe resources from object Bucket, helm
            and github types of channels
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-application:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-channel:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-placementrule:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-deployable:2.3.0
        version: 0.3.0
      entries:
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.3
    - currentCSV: multicluster-operators-subscription.v0.3.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.4.0
          createdAt: "2021-09-29T10:23:00Z"
          description: An operator to subscribe resources from object Bucket, helm
            and github types of channels
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-placementrule:2.4.0
        - quay.io/open-cluster-management/multicloud-integrations:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-deployable:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-application:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-channel:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.4.0
        version: 0.3.6
      entries:
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.4
    - currentCSV: multicluster-operators-subscription.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.5.0
          createdAt: "2022-03-23T18:00:00Z"
          description: An operator for deploying resources from object Bucket, helm
            repo or github repo to multiple managed clusters
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management-io/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: subscription status per cluster per app
            displayName: App Subscription Report
            kind: SubscriptionReport
            name: subscriptionreports.apps.open-cluster-management.io
            version: v1alpha1
          - description: subscription status per package
            displayName: App Subscription Status
            kind: SubscriptionStatus
            name: subscriptionstatuses.apps.open-cluster-management.io
            version: v1alpha1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.5.0
        - quay.io/open-cluster-management/multicloud-integrations:2.5.0
        - quay.io/open-cluster-management/multicluster-operators-application:2.5.0
        - quay.io/open-cluster-management/multicluster-operators-channel:2.5.0
        version: 0.4.0
      entries:
      - name: multicluster-operators-subscription.v0.4.0
        version: 0.4.0
      - name: multicluster-operators-subscription.v0.3.9
        version: 0.3.9
      - name: multicluster-operators-subscription.v0.3.8
        version: 0.3.8
      - name: multicluster-operators-subscription.v0.3.7
        version: 0.3.7
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.5
    - currentCSV: multicluster-operators-subscription.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-operators-subscription:2.6.0
          createdAt: "2022-06-06T18:00:00Z"
          description: An operator for deploying resources from object Bucket, helm
            repo or github repo to multiple managed clusters
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management-io/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: subscription status per cluster per app
            displayName: App Subscription Report
            kind: SubscriptionReport
            name: subscriptionreports.apps.open-cluster-management.io
            version: v1alpha1
          - description: subscription status per package
            displayName: App Subscription Status
            kind: SubscriptionStatus
            name: subscriptionstatuses.apps.open-cluster-management.io
            version: v1alpha1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/multicluster-operators-subscription:2.6.0
        - quay.io/stolostron/multicloud-integrations:2.6.0
        - quay.io/stolostron/multicluster-operators-application:2.6.0
        - quay.io/stolostron/multicluster-operators-channel:2.6.0
        version: 0.4.1
      entries:
      - name: multicluster-operators-subscription.v0.4.1
        version: 0.4.1
      - name: multicluster-operators-subscription.v0.4.0
        version: 0.4.0
      - name: multicluster-operators-subscription.v0.3.9
        version: 0.3.9
      - name: multicluster-operators-subscription.v0.3.8
        version: 0.3.8
      - name: multicluster-operators-subscription.v0.3.7
        version: 0.3.7
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.6
    defaultChannel: release-2.5
    packageName: multicluster-operators-subscription
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.90.0-beta.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"externalDbIP":
            "10.0.2.15","customImageNames": false,"repository": "icr.io/cpopen/turbonomic","tag":"8.17.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
          createdAt: "2023-03-24T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=42.53.0-beta.1 <42.90.0-beta.2'
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints. Turbonomic platform install and configuration
              documentation can be found online here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How do I start?

          Turbonomic platform installation and configuration documentation can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
          Carefully read through the prerequisites before beginning the installation.  Once ready, on the `Details` tab for the Turbonomic Platform Operator, under `Provided APIs`, click `Create instance`.
          For samples of common configurations, refer to sample custom resource YAMLs provided here: https://github.com/turbonomic/t8c-install/tree/master/samples
          Looking to upgrade your current deployment? Refer to documentation here: https://www.ibm.com/docs/en/tarm/latest?topic=version-updating-using-red-hat-openshift-operatorhub

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
        version: 42.90.0-beta.2
      entries:
      - name: t8c-operator.v42.90.0-beta.2
        version: 42.90.0-beta.2
      name: beta
    - currentCSV: t8c-operator.v42.90.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"externalDbIP":
            "10.0.2.15","customImageNames": false,"repository": "icr.io/cpopen/turbonomic","tag":"8.17.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
          createdAt: "2023-03-24T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=42.53.0 <42.90.0'
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints. Turbonomic platform install and configuration
              documentation can be found online here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How do I start?

          Turbonomic platform installation and configuration documentation can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
          Carefully read through the prerequisites before beginning the installation.  Once ready, on the `Details` tab for the Turbonomic Platform Operator, under `Provided APIs`, click `Create instance`.
          For samples of common configurations, refer to sample custom resource YAMLs provided here: https://github.com/turbonomic/t8c-install/tree/master/samples
          Looking to upgrade your current deployment? Refer to documentation here: https://www.ibm.com/docs/en/tarm/latest?topic=version-updating-using-red-hat-openshift-operatorhub

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
        version: 42.90.0
      entries:
      - name: t8c-operator.v42.90.0
        version: 42.90.0
      name: stable
    defaultChannel: stable
    packageName: t8c-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: https://stakater.com
    name: tenant-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tenant-operator.v0.12.66
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: =0.12.65
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:9f806a065b25229741dac8437a32ec45a99acecf4e6bcbe6e7d1fa4192bc560d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a35b96eacb4e46114b39354f66429f8f05df09c55c98c9416b4ffa2768310ef7
        version: 0.12.66
      entries:
      - name: tenant-operator.v0.12.66
        version: 0.12.66
      name: release-0.12
    - currentCSV: tenant-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2024-10-16T08:15:54Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.0
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:feb71a5045e621f8a8cbfe7e41831100b6b85af9812138071b6217cc4396860a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a35b96eacb4e46114b39354f66429f8f05df09c55c98c9416b4ffa2768310ef7
        version: 1.0.0
      entries:
      - name: tenant-operator.v1.0.0
        version: 1.0.0
      name: release-1.0
    - currentCSV: tenant-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-01-27T13:23:27Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.0
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:f412b6d05038a185f5fe35f9c917388194a5ea66afb538a513423c24390983ce
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a35b96eacb4e46114b39354f66429f8f05df09c55c98c9416b4ffa2768310ef7
        version: 1.1.0
      entries:
      - name: tenant-operator.v1.1.0
        version: 1.1.0
      name: release-1.1
    - currentCSV: tenant-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-03-19T13:30:50Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.2.1
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a35b96eacb4e46114b39354f66429f8f05df09c55c98c9416b4ffa2768310ef7
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:a899d60b31efd5cf2358efcd48737b1c4f4a0356ab6c0abe7a5f62ea825a52cd
        version: 1.2.1
      entries:
      - name: tenant-operator.v1.2.1
        version: 1.2.1
      name: release-1.2
    - currentCSV: tenant-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-06-11T15:25:04Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.3.0
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:e88aa5b008ece406bca35ec22fdc87c842544049d613d4d81c8bc5dd2fb50812
        version: 1.3.0
      entries:
      - name: tenant-operator.v1.3.0
        version: 1.3.0
      name: release-1.3
    - currentCSV: tenant-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-09-08T08:14:28Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.4.2
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:7ee14c0974d1eda849d2c199a56786b0ae777ef6569085be917449f1998b1158
        version: 1.4.2
      entries:
      - name: tenant-operator.v1.4.2
        version: 1.4.2
      name: release-1.4
    defaultChannel: release-1.4
    packageName: tenant-operator
    provider:
      name: Stakater
      url: https://stakater.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.0.4
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
        version: 1.0.4
      entries:
      - name: datadog-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      name: stable
    defaultChannel: stable
    packageName: datadog-operator-certified-rhmp
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sfn-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sfn-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sfn.services.k8s.aws/v1alpha1",
                "kind": "Activity",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sfn.services.k8s.aws/v1alpha1",
                "kind": "StateMachine",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sfn-controller:1.1.1
          createdAt: "2025-09-19T18:41:33Z"
          description: AWS Step Functions controller is a service controller for managing
            Step Functions resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Activity represents the state of an AWS sfn Activity resource.
            displayName: Activity
            kind: Activity
            name: activities.sfn.services.k8s.aws
            version: v1alpha1
          - description: StateMachine represents the state of an AWS sfn StateMachine
              resource.
            displayName: StateMachine
            kind: StateMachine
            name: statemachines.sfn.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Step Functions resources in AWS from within your Kubernetes cluster.

          **About Amazon Step FUnctions**

          Step Functions is a serverless orchestration service that lets you combine AWS Lambda functions and other AWS services to build business-critical applications. Through Step Functions' graphical console, you see your application’s workflow as a series of event-driven steps.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Step Functions
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sfn
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Step Functions Developer Resources
          url: https://aws.amazon.com/step-functions/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: step functions maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sfn-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-sfn-controller.v1.1.1
        version: 1.1.1
      - name: ack-sfn-controller.v1.1.0
        version: 1.1.0
      - name: ack-sfn-controller.v1.0.29
        version: 1.0.29
      - name: ack-sfn-controller.v1.0.28
        version: 1.0.28
      - name: ack-sfn-controller.v1.0.27
        version: 1.0.27
      - name: ack-sfn-controller.v1.0.26
        version: 1.0.26
      - name: ack-sfn-controller.v1.0.25
        version: 1.0.25
      - name: ack-sfn-controller.v1.0.24
        version: 1.0.24
      - name: ack-sfn-controller.v1.0.23
        version: 1.0.23
      - name: ack-sfn-controller.v1.0.22
        version: 1.0.22
      - name: ack-sfn-controller.v1.0.21
        version: 1.0.21
      - name: ack-sfn-controller.v1.0.20
        version: 1.0.20
      - name: ack-sfn-controller.v1.0.19
        version: 1.0.19
      - name: ack-sfn-controller.v1.0.18
        version: 1.0.18
      - name: ack-sfn-controller.v1.0.17
        version: 1.0.17
      - name: ack-sfn-controller.v1.0.16
        version: 1.0.16
      - name: ack-sfn-controller.v1.0.15
        version: 1.0.15
      - name: ack-sfn-controller.v1.0.14
        version: 1.0.14
      - name: ack-sfn-controller.v1.0.13
        version: 1.0.13
      - name: ack-sfn-controller.v1.0.11
        version: 1.0.11
      - name: ack-sfn-controller.v1.0.10
        version: 1.0.10
      - name: ack-sfn-controller.v1.0.9
        version: 1.0.9
      - name: ack-sfn-controller.v1.0.8
        version: 1.0.8
      - name: ack-sfn-controller.v1.0.6
        version: 1.0.6
      - name: ack-sfn-controller.v1.0.5
        version: 1.0.5
      - name: ack-sfn-controller.v1.0.2
        version: 1.0.2
      - name: ack-sfn-controller.v1.0.1
        version: 1.0.1
      - name: ack-sfn-controller.v1.0.0
        version: 1.0.0
      - name: ack-sfn-controller.v0.1.3
        version: 0.1.3
      - name: ack-sfn-controller.v0.1.2
        version: 0.1.2
      - name: ack-sfn-controller.v0.1.1
        version: 0.1.1
      - name: ack-sfn-controller.v0.1.0
        version: 0.1.0
      - name: ack-sfn-controller.v0.0.19
        version: 0.0.19
      - name: ack-sfn-controller.v0.0.18
        version: 0.0.18
      - name: ack-sfn-controller.v0.0.17
        version: 0.0.17
      - name: ack-sfn-controller.v0.0.16
        version: 0.0.16
      - name: ack-sfn-controller.v0.0.15
        version: 0.0.15
      - name: ack-sfn-controller.v0.0.14
        version: 0.0.14
      - name: ack-sfn-controller.v0.0.13
        version: 0.0.13
      - name: ack-sfn-controller.v0.0.12
        version: 0.0.12
      - name: ack-sfn-controller.v0.0.11
        version: 0.0.11
      - name: ack-sfn-controller.v0.0.10
        version: 0.0.10
      - name: ack-sfn-controller.v0.0.9
        version: 0.0.9
      - name: ack-sfn-controller.v0.0.8
        version: 0.0.8
      - name: ack-sfn-controller.v0.0.7
        version: 0.0.7
      - name: ack-sfn-controller.v0.0.6
        version: 0.0.6
      - name: ack-sfn-controller.v0.0.5
        version: 0.0.5
      name: alpha
    defaultChannel: alpha
    packageName: ack-sfn-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeVirt project
      provider-url: ""
    name: community-kubevirt-hyperconverged
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubevirt-hyperconverged-operator.v1.10.7
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployTektonTaskResources":true,"deployVmConsoleProxy":true},"tektonPipelines":{"namespace":"kubevirt"},"tektonTasks":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:0ee30557d5d5a42dec19121e70ed77fbdacd8932f55b11e2c7c1a3d5ad5df4aa
          createdAt: "2024-02-09 14:34:30"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:38293a1050bce9c7cb786cbe8719be1edb91a922a6c583eedb61e56729ed4c72
        - quay.io/kubevirt/virt-controller@sha256:0789fafed2913b35a771e3db882748502b3250be04ece86d97f30201779b4e54
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:e79027973b09aac75860f267c7d7f830599978c38a081001f404b9a1c3d2990f
        - quay.io/kubevirt/mtq-lock-server@sha256:9614df0d5af7c998bcceaf73b548f2e3b29ff8ee4ca64c32613bfbdd6e452f6c
        - quay.io/kubevirt/cdi-uploadproxy@sha256:551221d79902a5053d1c734b81163d69f087217e2ac13c49bdf6900336ef0786
        - quay.io/kubevirt/virt-operator@sha256:cbca7f93afbb18a840f7e763e4c03aeeb1fd97967f381c12c443cc179b51da29
        - quay.io/kubevirt/hostpath-provisioner@sha256:eae30f37b3e837478a4500144b7bc8eb4f5ddf15f49fa7354bc9d5e240ea8aea
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5ff5392cb762a41ee9bc4767e3a693249ae08b46a4322f623f24ad6b63dd06b5
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:57573a24923e5588bca6bc337a8b2b08406c5b77583974365d2cf063c0dd5d06
        - quay.io/kubevirt/cni-default-plugins@sha256:825e3f9fec1996c54a52cec806154945b38f76476b160d554c36e38dfffe5e61
        - quay.io/kubevirt/kubevirt-template-validator@sha256:a148b6e812b70f326b48d0439b1e44a967a6e1f8cf2f98b1a08e9294a93674eb
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:f64ab5981d9ddef1662dd0a05888783f5dbadda8ede596585ba80bfc97a0f51e
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:f7567bea7f24184da06c3ed87b396c804342fcfd89f54989622780ce0b7b0724
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:3fbcc32bd4e4d15bd93c96def784a229cd84cca27942bf4858b581f31c97ee02
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4dc0b87ccd69f9865b89234d8555d3a614ab0a16ed94a3016ffd27f8106132ce
        - quay.io/kubevirt/virtio-container-disk@sha256:f8e71a3a930ce95d646d80862d5ff6f9d3ed7ef2541dd2e2acec0bce26274734
        - quay.io/kubevirt/virt-api@sha256:707003b221496b4432da2f507d1e36e528b45888b5d321e06d460f0678da44ae
        - quay.io/kubevirt/virt-artifacts-server:1.10.7-unstable
        - quay.io/kubevirt/virt-handler@sha256:138dfda5fea8622f3da0d6413fe214fef80c2fd6a6f9533592a0dbfa7e1865b5
        - quay.io/kubevirt/hostpath-csi-driver@sha256:0ca8b50b0adb20afd4e20da9974200f3290b6d51aa7c1129de6e1b953bf26594
        - quay.io/kubevirt/mtq-controller@sha256:96dd73ea0d7847d9f6322192559747bd85655e25ee25482d112e6e3c6fef21b2
        - quay.io/kubevirt/libguestfs-tools@sha256:b7e248be783691daad1326cf41641d5042be80c5c057ebc5315ca7a1619b0581
        - quay.io/kubevirt/cdi-apiserver@sha256:e9e39408413b1478d2e98eba68913f9e20c93000558b190b47de73bdfd1d9ac4
        - quay.io/kubevirt/ovs-cni-plugin@sha256:08f72edf2bef876bba0b0f5513d30225304ad5e7ad6912a61c083664acdb99ff
        - quay.io/kubevirt/macvtap-cni@sha256:434420511e09b2b5ede785a2c9062b6658ffbc26fbdd4629ce06110f9039c600
        - quay.io/kubevirt/cdi-importer@sha256:3143bbc67cdc6267eb48b7eaac664b8551ac4c11401dfbf4921efd3f233e6ce9
        - ghcr.io/kubevirt/kubesecondarydns@sha256:e87e829380a1e576384145f78ccaa885ba1d5690d5de7d0b73d40cfb804ea24d
        - quay.io/kubevirt/virt-exportproxy@sha256:f14444b0200a85efb4b3c176caefe70aabba7d295e33d2af14027a59ce297e24
        - quay.io/kubevirt/cdi-operator@sha256:6c63521d835578b0fbb77d3145b648f95ab0932d5d1b36ff2068ed8fcd91bc5a
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:16ecabb1d733aa17084cd16d88963c5255d04f389e0f4d41941d67e8be973594
        - quay.io/kubevirt/virt-launcher@sha256:4c5fce3de2e2589197de72fb0c9436490ea318aca952c05a622c43e067023f35
        - quay.io/kubevirt/bridge-marker@sha256:bba066e3b5ff3fb8c5e20861fe8abe51e3c9b50ad6ce3b2616af9cb5479a06d0
        - quay.io/kubevirt/cdi-uploadserver@sha256:30f1827d3696cf996b081c22c3267ca78e7219c872fdb54950198fa54359f6ee
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - quay.io/kubevirt/cdi-controller@sha256:27c47883a08226f83757971d3adafb0cd9bcb26e58fbcf7208236070e0adf37e
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:0ee30557d5d5a42dec19121e70ed77fbdacd8932f55b11e2c7c1a3d5ad5df4aa
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:86c10c122767661e31ee26a9e9de7ea606c2c1168b32793b4ea567ede2ae3b21
        - quay.io/kubevirt/virt-exportserver@sha256:73311f79a9c71007f8572b3cc40cd6f6da404c7ef0a9c6509fb717d979546582
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f6717ce72a2615c7fbc746b4068f788e78579c54c43b8716e5ce650d97af2df1
        - quay.io/kubevirt/kubemacpool@sha256:3214020c21204c189cdf2d767beb8efb61ee3ef65dd8de7f0f1edc61e13e8adf
        - quay.io/kubevirt/ssp-operator@sha256:49cac00844f091d6e80301573e00235d685415185ec29c89b4bd7361f938711e
        - quay.io/kubevirt/mtq-operator@sha256:ef5c4bb1fa72f9e6e75d5585ca352f7563720389e6c9876aa8b07d12d1be6194
        - quay.io/kubevirt/cdi-cloner@sha256:9d31b14f23259398c5bac636f5ead13ad0afd6fe8eeab4499e8e047b4d85074f
        version: 1.10.7
      entries:
      - name: kubevirt-hyperconverged-operator.v1.10.7
        version: 1.10.7
      name: 1.10.7
    - currentCSV: kubevirt-hyperconverged-operator.v1.11.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployTektonTaskResources":true,"deployVmConsoleProxy":true},"tektonPipelines":{"namespace":"kubevirt"},"tektonTasks":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:dd6538e720f18b723b0807037ca7cc1014cba09e57b20706e10e6eb641ac6acc
          createdAt: "2024-03-06 15:11:00"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: quay.io/kubevirt/must-gather
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - quay.io/kubevirt/aaq-server@sha256:0c6b01c48030e03e99b3ff2b319658c2bf2b2001f4e3e030ba70d1143b25717a
        - quay.io/kubevirt/virt-exportserver@sha256:f8473fc824f8bac5f21c64b71cdb4227de28f9bb25c049932f1259216b874a86
        - quay.io/kubevirt/cdi-importer@sha256:7320dce01f4628363992997b592200b4804daf15a9f237cd50a4a6acab7798e8
        - quay.io/kubevirt/virt-artifacts-server@sha256:22fa3218deca38dc98fa0e905df16173a47ccec85707eb91404a367da6480817
        - quay.io/kubevirt/virt-operator@sha256:0e6d56ea0ebdf9bb832f29ab0dc4603f0f619474bb8bc02a01e0a06e9d8d6f4a
        - quay.io/kubevirt/mtq-operator@sha256:90c7608b0ba085582a65ccdc11d089049c5e6b0438d018dd9dc76eb3490383c7
        - quay.io/kubevirt/cni-default-plugins@sha256:c884d6d08f8c0db98964f1eb3877b44ade41fa106083802a9914775df17d5291
        - quay.io/kubevirt/bridge-marker@sha256:bba066e3b5ff3fb8c5e20861fe8abe51e3c9b50ad6ce3b2616af9cb5479a06d0
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:dd6538e720f18b723b0807037ca7cc1014cba09e57b20706e10e6eb641ac6acc
        - quay.io/kubevirt/virt-handler@sha256:238550bb263608ba334b44a15647cbf875482ccfb46049713c14af5d5f0cc788
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:f1bd6ee18c4021c1c94f29edfab89b49b6a4d1b800936c19dbef2d75f8202f2d
        - quay.io/kubevirt/kubemacpool@sha256:cf8daa57ae6603b776d3af512331b143fa03bc2f4b72f28420fddcf5e4156d0a
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:4ed1f37116cbaf883edd8783201d091fbbe121fcb84e7184ee531d3b1abd77ab
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - registry.k8s.io/sig-storage/livenessprobe@sha256:5baeb4a6d7d517434292758928bb33efc6397368cbb48c8a4cf29496abf4e987
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:bf17da674ed6dac95aa4fd7637173fd4e026ebb4523b66f529c0eff394a25bcc
        - quay.io/kubevirt/cdi-uploadserver@sha256:16d257e575b2bd504f148f63b57dc181fa6058d33a11ada64888183c3e4d4863
        - quay.io/kubevirt/virt-exportproxy@sha256:942c548d4848a87945eba51683002797eb4e1f5c463a1873115b2ec0e1e6e169
        - quay.io/kubevirt/ovs-cni-plugin@sha256:e16ac74343da21abb8fb668ce71e728053d00503a992dae2164b9e94a280113e
        - quay.io/kubevirt/hostpath-csi-driver@sha256:8d84b9a33e0ec3749aece7c5f9879d1e0d248e8028ab3f776528a044d387033c
        - ghcr.io/kubevirt/kubesecondarydns@sha256:e87e829380a1e576384145f78ccaa885ba1d5690d5de7d0b73d40cfb804ea24d
        - quay.io/kubevirt/mtq-controller@sha256:ae7dd07e4018c63414b021eba34aa4776a6c7d6ff4fcdf9fd2d7ca30aae98a29
        - quay.io/kubevirt/cdi-apiserver@sha256:95f6e71b36883abcb9ec463cfea2332f5490ec9eb3e4868bdf0a4da769279094
        - quay.io/kubevirt/cdi-controller@sha256:606de89e68be2ac911466e6713345932bda496c1e4f20e34ef5efa270456c9fd
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:dd2948ca8e62d17eafba40bdcaad8367a97f092fd3b8d9849b2d62aec3152a72
        - quay.io/kubevirt/aaq-operator@sha256:590d24742401eeff3fb6a9c8ed977c63442f8f0a0505509fae69aa1bf8aa84ae
        - quay.io/kubevirt/macvtap-cni@sha256:850b89343ace7c7ea6b18dd8e11964613974e9d1f7377af03854d407fb15230a
        - quay.io/kubevirt/virtio-container-disk@sha256:f8e71a3a930ce95d646d80862d5ff6f9d3ed7ef2541dd2e2acec0bce26274734
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:3fbcc32bd4e4d15bd93c96def784a229cd84cca27942bf4858b581f31c97ee02
        - quay.io/kubevirt/virt-controller@sha256:4f576e30b83db7b9f96dde77896898bde87db3517963c4227634d0c3e96bb86d
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:70efc16ab2bfec04f73ccc292064d080870659cfd47c2e0c9017ff7b6fd78863
        - quay.io/kubevirt/cdi-operator@sha256:8a27e1450c3bf8c1d8924e74470a62419d02ed2202924b082a74e2414b1fc21a
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:c53535af8a7f7e3164609838c4b191b42b2d81238d75c1b2a2b582ada62a9780
        - quay.io/kubevirt/libguestfs-tools@sha256:4971af1ee42959d366a3db6129ab684d11064e00bbd3c2b36324ad14e90cb903
        - quay.io/kubevirt/cdi-cloner@sha256:b22985af0ee7b1ce4e1dedded08c853d90d9e7209f98c5d470897bbefce2a063
        - quay.io/kubevirt/ssp-operator@sha256:d57ef05fea09c4690dbe728ebe6a39d9df1a3574dfa7d18346bf3fe66afed702
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:57573a24923e5588bca6bc337a8b2b08406c5b77583974365d2cf063c0dd5d06
        - quay.io/kubevirt/virt-launcher@sha256:4fbfd76561667e98fafc75c0b804f77df251f96471d1121d69f149cf8fb9a4d8
        - quay.io/kubevirt/kubevirt-template-validator@sha256:f6db539fe4d4b752644c6bc8a308f5a80ae09a71da5c5d535114dc4c3cf18e0a
        - quay.io/kubevirt/aaq-controller@sha256:cb2fa52ae3f6a1a99b2de77b58c5db54c43ba8ee86c6540b4844c901180ce611
        - quay.io/kubevirt/virt-api@sha256:735662c824da3d6069624d40674163d758c2256f8043cf6f02f505c3a724afa3
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:e2def4213ec0657e72eb790ae8a115511d5b8f164a62d3568d2f1bff189917e8
        - quay.io/kubevirt/mtq-lock-server@sha256:8e686107dd9990dab1a3d90ee1068541938f673bf54ea9e1308b2ed7218546b0
        - quay.io/kubevirt/hostpath-provisioner@sha256:adb1d33e95e8991dc47107926172469b5ecbeace509391ce021ca42d305a41d3
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:38293a1050bce9c7cb786cbe8719be1edb91a922a6c583eedb61e56729ed4c72
        - quay.io/kubevirt/cdi-uploadproxy@sha256:b89104ef5513d63cab43b34ac52f752c3d935cb986d5527d920c2c31d3b6f239
        version: 1.11.0
      entries:
      - name: kubevirt-hyperconverged-operator.v1.11.0
        version: 1.11.0
      name: 1.11.0
    - currentCSV: kubevirt-hyperconverged-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:7afeef450521fb7bbc918e874d9297dc3762cdeb28db0bef96d66d48b8752a9d
          createdAt: "2024-08-07 15:33:27"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: quay.io/kubevirt/must-gather
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:7afeef450521fb7bbc918e874d9297dc3762cdeb28db0bef96d66d48b8752a9d
        - quay.io/kubevirt/virt-artifacts-server@sha256:cc3e1e17bc3fa2a895f934210c9106bee01c3f9bf001f988d920eb58628e347b
        - quay.io/kubevirt/virt-operator@sha256:1ecc62c25f869e17027c930150d5c9b52b72d7e1fb434376bc0be0d886952b2e
        - quay.io/kubevirt/aaq-operator@sha256:002513b78dfd795f5195534b651eace136ead3d73c8ac8cae145607e79c8a8ba
        - quay.io/kubevirt/hostpath-csi-driver@sha256:b7ccc469ad7e640fed2e98b50eae57a64e81c7b9862a799b7a9892d14ade8937
        - quay.io/kubevirt/virt-controller@sha256:3471b00eb9f4774ae33867c7d0b83e1d26a065a2bd19d305b5d3946ab665b5e9
        - quay.io/kubevirt/aaq-controller@sha256:170abd7a75397ddce07ba0d64886ead25183c49cacbccf0d191a4dd37ca013fd
        - quay.io/kubevirt/ssp-operator@sha256:a15a53d0f3ef530fe469acf3b7673486011605d0021d4dbe29af94096b3b65ee
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:413a6c229ebb608868e8b2c7c06b251c6b1d43c07af12a76b893292cbed8bef9
        - quay.io/kubevirt/ovs-cni-plugin@sha256:e16ac74343da21abb8fb668ce71e728053d00503a992dae2164b9e94a280113e
        - quay.io/kubevirt/kubevirt-template-validator@sha256:8f7c389a7cd7ef34f2edc84245bc1d0ebcf13e86da5e0b8893f56733c482b10d
        - quay.io/kubevirt/hostpath-provisioner@sha256:2b93b6c4ce0307a8f0dbdae2226da4acb1bae781788fbde7f0c32d6ad6421aa0
        - quay.io/kubevirt/mtq-lock-server@sha256:8b5189518ac6f68161d6a259e2a6f70e7902855ea1510aa260e79093520a7803
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac
        - quay.io/kubevirt/libguestfs-tools@sha256:0de3de89720d4e5433a33944f00dac590daed6ab12e4d26a657ef082d0573747
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:e2def4213ec0657e72eb790ae8a115511d5b8f164a62d3568d2f1bff189917e8
        - quay.io/kubevirt/mtq-operator@sha256:859524721ae1a3b1e5dd6e29036b4c80acff13f74a4ee6769d867283e9ccbba5
        - registry.k8s.io/sig-storage/livenessprobe@sha256:5baeb4a6d7d517434292758928bb33efc6397368cbb48c8a4cf29496abf4e987
        - quay.io/kubevirt/virtio-container-disk@sha256:f8e71a3a930ce95d646d80862d5ff6f9d3ed7ef2541dd2e2acec0bce26274734
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:935475c2850466aa5ac57e4de627fb177515cb2c402a95842ead095d82b6df5f
        - quay.io/kubevirt/mtq-controller@sha256:04b38720b65479046270c8132cff5ab0f2dabc9cae91de1cc95fecf167efc356
        - quay.io/kubevirt/virt-exportserver@sha256:acac33a19528592cc3ef0cd3a25fa4c72bcf992e4c2c247437007fcb9de0aa6d
        - quay.io/kubevirt/cdi-uploadproxy@sha256:86285df5cf52533732ddf4fdb74674cbce877ba69833cb8ed058c2a157cb5573
        - quay.io/kubevirt/cdi-cloner@sha256:05da8df0b64024f9e2a48a1a54f06c821d79a3efe7828fad2c4fda6a01d0cd86
        - quay.io/kubevirt/cni-default-plugins@sha256:0c354fa9d695b8cab97b459e8afea2f7662407a987e83f6f6f1a8af4b45726be
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:3c70cb833819d8346e5531b10ac516283bab7098e782708768a6fe00ae0ddfac
        - quay.io/kubevirt/cdi-operator@sha256:f68e4f2a082fc72d92bc1df0340285af03cd12d8c7721b38c619d1de0f4c175a
        - ghcr.io/kubevirt/kubesecondarydns@sha256:6268d84154e2483fbce8c1adacbdaf6f0839117b2d48d9fa4687cc8f76bd5130
        - quay.io/kubevirt/virt-api@sha256:aeeae94c9dd3b352ff2fbcdba77a6b084fb4ebe0fed35fe7af9481e9e2b1f24c
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:83b460502671fb4f34116363a1a39b2ddfc9d14a920ee0a6413bfc3bd0580404
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - quay.io/kubevirt/cdi-apiserver@sha256:4ba004f57e7d589ddb5372a9729360460a97effecc6e5ecaccd7fbf16c6fcf13
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:d51821f001c959a79943cb22f5ea47589d7cf5c032e11ec1d571e8db7bf29b78
        - quay.io/kubevirt/virt-exportproxy@sha256:dd3b5c12314fa76b88d55d978210de81562fb888351748ea3997551c8c96a877
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:c4b6b02737bc24906fcce57fe6626d1a36cb2b91baa971af2a5e5a919093c34e
        - quay.io/kubevirt/cdi-importer@sha256:de5b89c879a531165482d540deffdc2dec0c188fd0f7b5fbd49c0298cf9a5b1d
        - quay.io/kubevirt/kubemacpool@sha256:20e156be33e6d3692c456081acbb91a4349f94de448f7d1f1cddd0228931b31f
        - quay.io/kubevirt/virt-launcher@sha256:cd37543ed2bc6baadbcbffa6951d740e4c65dc0ea366dec0a8593193c384e6de
        - quay.io/kubevirt/virt-handler@sha256:2a7d96da4de3ea51339b3a047abb4b05e088e3d1080110b42a85ea5baff4a0cc
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:246b3819cfa3833f9a8cb01e8cb859aa61a1924a85ea4a112e43e8414f32c429
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:3fbcc32bd4e4d15bd93c96def784a229cd84cca27942bf4858b581f31c97ee02
        - quay.io/kubevirt/cdi-controller@sha256:392ead2fdc5050c584765ebf99706cc5c0ab9c2b424b692ad35005e77d110b58
        - quay.io/kubevirt/cdi-uploadserver@sha256:d7fb653bb33dade84583fff6048092346639532d6675aaf9f88070171cb0fe7c
        - quay.io/kubevirt/bridge-marker@sha256:bba066e3b5ff3fb8c5e20861fe8abe51e3c9b50ad6ce3b2616af9cb5479a06d0
        - quay.io/kubevirt/aaq-server@sha256:9db6ca9147b1d75ba4d6bb1326c4ec289327eb9f8aadbcd951fcc689e584cd0d
        - quay.io/kubevirt/macvtap-cni@sha256:850b89343ace7c7ea6b18dd8e11964613974e9d1f7377af03854d407fb15230a
        version: 1.12.0
      entries:
      - name: kubevirt-hyperconverged-operator.v1.12.0
        version: 1.12.0
      - name: kubevirt-hyperconverged-operator.v1.11.1
        version: 1.11.1
      - name: kubevirt-hyperconverged-operator.v1.10.0
        version: 1.10.0
      - name: kubevirt-hyperconverged-operator.v1.9.0
        version: 1.9.0
      - name: kubevirt-hyperconverged-operator.v1.8.2
        version: 1.8.2
      - name: kubevirt-hyperconverged-operator.v1.8.1
        version: 1.8.1
      - name: kubevirt-hyperconverged-operator.v1.8.0
        version: 1.8.0
      - name: kubevirt-hyperconverged-operator.v1.7.0
        version: 1.7.0
      - name: kubevirt-hyperconverged-operator.v1.6.0
        version: 1.6.0
      - name: kubevirt-hyperconverged-operator.v1.5.2
        version: 1.5.2
      - name: kubevirt-hyperconverged-operator.v1.5.1
        version: 1.5.1
      - name: kubevirt-hyperconverged-operator.v1.5.0
        version: 1.5.0
      - name: kubevirt-hyperconverged-operator.v1.4.4
        version: 1.4.4
      - name: kubevirt-hyperconverged-operator.v1.4.3
        version: 1.4.3
      - name: kubevirt-hyperconverged-operator.v1.4.2
        version: 1.4.2
      - name: kubevirt-hyperconverged-operator.v1.4.1
        version: 1.4.1
      - name: kubevirt-hyperconverged-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: community-kubevirt-hyperconverged
    provider:
      name: KubeVirt project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: StreamNative
      provider-url: https://streamnative.io
    name: sn-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
        - docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21
        - docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      name: alpha
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
        - docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21
        - docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      name: beta
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21
        - quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
        - docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      name: stable
    defaultChannel: alpha
    packageName: sn-operator
    provider:
      name: StreamNative
      url: https://streamnative.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzo-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzo-operator.v2.0.101
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzo.cambridgesemantics.com/v1",
                "kind": "Anzo",
                "metadata": {
                  "name": "agent01"
                },
                "spec": {
                  "nodeConfig": {
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "app": "anzo"
                        }
                      },
                      "serviceName": "anzo-agent01",
                      "template": {
                        "metadata": {
                          "labels": {
                            "app": "anzo"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo@sha256:d1c2d54ca205d669c5ea1bb6aae105afe2385cc5695cecd78423ffa1d98929f8",
                              "name": "anzo",
                              "resources": {
                                "limits": {
                                  "cpu": "4000m",
                                  "memory": "12Gi"
                                },
                                "requests": {
                                  "cpu": "4000m",
                                  "memory": "12Gi"
                                }
                              }
                            }
                          ],
                          "serviceAccountName": "anzo-operator"
                        }
                      }
                    }
                  },
                  "role": "AnzoAgent"
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-operator@sha256:563e964f5771b9bfd860ea993b6e0936a54504908f5b2d1dff6f99ade74bb7a0
          createdAt: "2021-08-02"
          description: kubernetes operator for Anzo
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzo-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzo-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Anzo is the Schema for the anzos API
            displayName: Anzo Operator
            kind: Anzo
            name: anzos.anzo.cambridgesemantics.com
            version: v1
        description: |-
          The Anzo Operator provides the way to install and configure an anzo agent setup on Red Hat K8S environment.
          Currently, this is supported only through when deployed as an Anzo Agent as part of an Anzo Unstructured deployment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v2.0.0/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers through [ Cambridge Semantics Customer Center ]( https://customercenter.cambridgesemantics.com/ ).
        displayName: Anzo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        links:
        - name: Anzo Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v2.0.0/README_openshift.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-operator@sha256:563e964f5771b9bfd860ea993b6e0936a54504908f5b2d1dff6f99ade74bb7a0
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo@sha256:d1c2d54ca205d669c5ea1bb6aae105afe2385cc5695cecd78423ffa1d98929f8
        version: 2.0.0
      entries:
      - name: anzo-operator.v2.0.101
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anzo-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: odr-hub-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odr-hub-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east"
                },
                "spec": {
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east1"
                },
                "spec": {
                  "cidrs": [
                    "192.168.39.45/32"
                  ],
                  "clusterFence": "ManuallyFenced",
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east1"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPlacementControl",
                "metadata": {
                  "name": "drplacementcontrol-sample",
                  "namespace": "application-namespace"
                },
                "spec": {
                  "drPolicyRef": {
                    "name": "drpolicy-sample"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "excludedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "includedResources": [
                          "*"
                        ]
                      }
                    ]
                  },
                  "placementRef": {
                    "kind": "PlacementRule",
                    "name": "application-placement-rule"
                  },
                  "preferredCluster": "east",
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east",
                    "west"
                  ],
                  "replicationClassSelector": {
                    "matchLabels": {
                      "class": "ramen"
                    }
                  },
                  "schedulingInterval": "1h"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "metrodr-drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east1",
                    "east2"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DRCluster is the Schema for the drclusters API
            displayName: DRCluster
            kind: DRCluster
            name: drclusters.ramendr.openshift.io
            version: v1alpha1
          - description: DRPlacementControl is the Schema for the drplacementcontrols
              API
            displayName: DRPlacement Control
            kind: DRPlacementControl
            name: drplacementcontrols.ramendr.openshift.io
            version: v1alpha1
          - description: DRPolicy is the Schema for the drpolicies API
            displayName: DRPolicy
            kind: DRPolicy
            name: drpolicies.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Hub is a disaster-recovery orchestrator for stateful
          applications. It operates from an Advanced Cluster Management (ACM) hub
          cluster to orchestrate the recovery of application state, and scheduling
          of ACM PlacementRule for disaster recovery operations.
        displayName: Openshift DR Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:769acf15d016d4992793c9c527d7c268d2a67502dd7e0fadc240e4896398195b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 4.15.20-rhodf
      entries:
      - name: odr-hub-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odr-hub-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odr-hub-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odr-hub-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odr-hub-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odr-hub-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odr-hub-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odr-hub-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odr-hub-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odr-hub-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odr-hub-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odr-hub-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odr-hub-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odr-hub-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odr-hub-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odr-hub-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odr-hub-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odr-hub-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odr-hub-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odr-hub-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odr-hub-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east"
                },
                "spec": {
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east1"
                },
                "spec": {
                  "cidrs": [
                    "192.168.39.45/32"
                  ],
                  "clusterFence": "ManuallyFenced",
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east1"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPlacementControl",
                "metadata": {
                  "name": "drplacementcontrol-sample",
                  "namespace": "application-namespace"
                },
                "spec": {
                  "drPolicyRef": {
                    "name": "drpolicy-sample"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "excludedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "includedResources": [
                          "*"
                        ]
                      }
                    ]
                  },
                  "placementRef": {
                    "kind": "PlacementRule",
                    "name": "application-placement-rule"
                  },
                  "preferredCluster": "east",
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east",
                    "west"
                  ],
                  "replicationClassSelector": {
                    "matchLabels": {
                      "class": "ramen-replication-class"
                    }
                  },
                  "schedulingInterval": "1h",
                  "volumeSnapshotClassSelector": {
                    "matchLabels": {
                      "class": "ramen-snapshot-class"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "metrodr-drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east1",
                    "east2"
                  ],
                  "replicationClassSelector": {},
                  "schedulingInterval": {},
                  "volumeSnapshotClassSelector": {}
                }
              }
            ]
          capabilities: Deep Insights
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DRCluster is the Schema for the drclusters API
            displayName: DRCluster
            kind: DRCluster
            name: drclusters.ramendr.openshift.io
            version: v1alpha1
          - description: DRPlacementControl is the Schema for the drplacementcontrols
              API
            displayName: DRPlacement Control
            kind: DRPlacementControl
            name: drplacementcontrols.ramendr.openshift.io
            version: v1alpha1
          - description: DRPolicy is the Schema for the drpolicies API
            displayName: DRPolicy
            kind: DRPolicy
            name: drpolicies.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Hub is a disaster-recovery orchestrator for stateful
          applications. It operates from an Advanced Cluster Management (ACM) hub
          cluster to orchestrate the recovery of application state, and scheduling
          of ACM PlacementRule for disaster recovery operations.
        displayName: Openshift DR Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:af262b0890300183db1766bc392017b0bda723a9ba841dc25e3d928aa413db9a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        version: 4.16.16-rhodf
      entries:
      - name: odr-hub-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odr-hub-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odr-hub-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odr-hub-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odr-hub-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odr-hub-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odr-hub-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odr-hub-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odr-hub-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odr-hub-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odr-hub-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odr-hub-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odr-hub-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odr-hub-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odr-hub-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odr-hub-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odr-hub-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odr-hub-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: container-security-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: container-security-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secscan.quay.redhat.com/v1alpha1",
                "kind": "ImageManifestVuln",
                "metadata": {
                  "selfLink": "/apis/secscan.quay.redhat.com/v1alpha1/namespaces/openshift-cluster-version/imagemanifestvulns/sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18",
                  "resourceVersion": "14565395",
                  "name": "sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18",
                  "uid": "3f16a188-e69a-406c-aa06-df340195409c",
                  "creationTimestamp": "2021-03-17T12:07:16Z",
                  "generation": 2,
                  "managedFields": [
                    {
                      "apiVersion": "secscan.quay.redhat.com/v1alpha1",
                      "fieldsType": "FieldsV1",
                      "fieldsV1": {
                        "f:metadata": {
                          "f:labels": {
                            ".": {},
                            "f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": {}
                          }
                        },
                        "f:spec": {
                          ".": {},
                          "f:features": {},
                          "f:image": {},
                          "f:manifest": {}
                        },
                        "f:status": {
                          ".": {},
                          "f:affectedPods": {
                            ".": {},
                            "f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": {}
                          },
                          "f:fixableCount": {},
                          "f:highCount": {},
                          "f:highestSeverity": {},
                          "f:lastUpdate": {},
                          "f:lowCount": {},
                          "f:mediumCount": {}
                        }
                      },
                      "manager": "security-labeller",
                      "operation": "Update",
                      "time": "2021-03-17T13:07:20Z"
                    }
                  ],
                  "namespace": "openshift-cluster-version",
                  "labels": {
                    "openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": "true"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:447e05859134a0455ca55f8aae16c8a27ce06af4c18c191b41f6339df553f1bf
          createdAt: 2021-11-02 11:11 UTC
          description: Identify image vulnerabilities in Kubernetes pods
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Quay Container Security Operator (CSO) brings Quay and Clair
          metadata to Kubernetes / OpenShift. Starting with vulnerability information
          the scope will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Quay Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Quay Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:447e05859134a0455ca55f8aae16c8a27ce06af4c18c191b41f6339df553f1bf
        version: 3.4.7
      entries:
      - name: container-security-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: container-security-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"secscan.quay.redhat.com/v1alpha1\",\n
            \   \"kind\": \"ImageManifestVuln\",\n    \"metadata\": {\n      \"selfLink\":
            \"/apis/secscan.quay.redhat.com/v1alpha1/namespaces/openshift-cluster-version/imagemanifestvulns/sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18\",\n
            \     \"resourceVersion\": \"14565395\",\n      \"name\": \"sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18\",\n
            \     \"uid\": \"3f16a188-e69a-406c-aa06-df340195409c\",\n      \"creationTimestamp\":
            \"2021-03-17T12:07:16Z\",\n      \"generation\": 2,\n      \"managedFields\":
            [\n        {\n          \"apiVersion\": \"secscan.quay.redhat.com/v1alpha1\",\n
            \         \"fieldsType\": \"FieldsV1\",\n          \"fieldsV1\": {\n            \"f:metadata\":
            {\n              \"f:labels\": {\n                \".\": {},\n                \"f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            {}\n              }\n            },\n            \"f:spec\": {\n              \".\":
            {},\n              \"f:features\": {},\n              \"f:image\": {},\n
            \             \"f:manifest\": {}\n            },\n            \"f:status\":
            {\n              \".\": {},\n              \"f:affectedPods\": {\n                \".\":
            {},\n                \"f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            {}\n              },\n              \"f:fixableCount\": {},\n              \"f:highCount\":
            {},\n              \"f:highestSeverity\": {},\n              \"f:lastUpdate\":
            {},\n              \"f:lowCount\": {},\n              \"f:mediumCount\":
            {}\n            }\n          },\n          \"manager\": \"security-labeller\",\n
            \         \"operation\": \"Update\",\n          \"time\": \"2021-03-17T13:07:20Z\"\n
            \       }\n      ],\n      \"namespace\": \"openshift-cluster-version\",\n
            \     \"labels\": {\n        \"openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            \"true\"\n      }\n    },\n    \"spec\": {\n      \"features\": [\n        {\n
            \         \"name\": \"musl\",\n          \"namespaceName\": \"alpine:v3.14\",\n
            \         \"version\": \"1.2.2-r3\",\n          \"versionformat\": \"dpkg\",\n
            \         \"vulnerabilities\": [\n            {\n              \"fixedby\":
            \"1.2.2_pre2-r0\",\n              \"link\": \"https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28928\",\n
            \             \"metadata\": \"{\\\"NVD\\\": {\\\"CVSSv3\\\": {\\\"Vectors\\\":
            \\\"CVSS:3.0/AV:L/AC:L/PR:L/UI:N/S:U/C:N/I:N/A:H\\\", \\\"ImpactScore\\\":
            3.6, \\\"Score\\\": 5.5, \\\"ExploitabilityScore\\\": 1.8}, \\\"CVSSv2\\\":
            {\\\"Score\\\": 2.1, \\\"PublishedDateTime\\\": \\\"2020-11-24T18:15Z\\\",
            \\\"Vectors\\\": \\\"AV:L/AC:L/Au:N/C:N/I:N/A:P\\\"}}}\",\n              \"name\":
            \"CVE-2020-28928\",\n              \"namespaceName\": \"alpine:v3.14\",\n
            \             \"severity\": \"Low\"\n            }\n          ]\n        }\n
            \     ],\n      \"image\": \"quay.io/harishg/pruned-redhat-operators-index\",\n
            \     \"manifest\": \"sha256:19631c86642609254777ff59fc9e128aa5d0a94a7dbb504ef1fcfb2bbc6e486e\"\n
            \   }\n  }\n]    "
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:6b3581e62550d6a615d36eee6eaf4c89b9c35ba53cab8433bd93c68393a9d115
          createdAt: 2021-09-10 23:09 UTC
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.4.x <3.5.7'
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Quay Container Security Operator (CSO) brings Quay and Clair
          metadata to Kubernetes / OpenShift. Starting with vulnerability information
          the scope will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Quay Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Quay Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:6b3581e62550d6a615d36eee6eaf4c89b9c35ba53cab8433bd93c68393a9d115
        version: 3.5.7
      entries:
      - name: container-security-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: container-security-operator.v3.10.15
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:de7c66a0004f3c0643c09d4c3bbb4a0f1f8f2cbf264250802cdc58028fc72b50
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.10.15
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:de7c66a0004f3c0643c09d4c3bbb4a0f1f8f2cbf264250802cdc58028fc72b50
        version: 3.10.15
      entries:
      - name: container-security-operator.v3.10.15
        version: 3.10.15
      - name: container-security-operator.v3.10.14
        version: 3.10.14
      - name: container-security-operator.v3.10.13
        version: 3.10.13
      - name: container-security-operator.v3.10.12
        version: 3.10.12
      - name: container-security-operator.v3.10.11
        version: 3.10.11
      - name: container-security-operator.v3.10.10
        version: 3.10.10
      - name: container-security-operator.v3.10.9
        version: 3.10.9
      - name: container-security-operator.v3.10.8
        version: 3.10.8
      - name: container-security-operator.v3.10.7
        version: 3.10.7
      - name: container-security-operator.v3.10.6
        version: 3.10.6
      - name: container-security-operator.v3.10.5
        version: 3.10.5
      - name: container-security-operator.v3.10.4
        version: 3.10.4
      - name: container-security-operator.v3.10.3
        version: 3.10.3
      - name: container-security-operator.v3.10.2
        version: 3.10.2
      - name: container-security-operator.v3.10.1
        version: 3.10.1
      - name: container-security-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: container-security-operator.v3.11.13
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:3914389fd53c371627921298c5dc3564999d92bc720f27a240f74c034ebbf5ec
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.11.13
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:3914389fd53c371627921298c5dc3564999d92bc720f27a240f74c034ebbf5ec
        version: 3.11.13
      entries:
      - name: container-security-operator.v3.11.13
        version: 3.11.13
      - name: container-security-operator.v3.11.12
        version: 3.11.12
      - name: container-security-operator.v3.11.11
        version: 3.11.11
      - name: container-security-operator.v3.11.10
        version: 3.11.10
      - name: container-security-operator.v3.11.9
        version: 3.11.9
      - name: container-security-operator.v3.11.8
        version: 3.11.8
      - name: container-security-operator.v3.11.7
        version: 3.11.7
      - name: container-security-operator.v3.11.6
        version: 3.11.6
      - name: container-security-operator.v3.11.5
        version: 3.11.5
      - name: container-security-operator.v3.11.4
        version: 3.11.4
      - name: container-security-operator.v3.11.3
        version: 3.11.3
      - name: container-security-operator.v3.11.2
        version: 3.11.2
      - name: container-security-operator.v3.11.1
        version: 3.11.1
      - name: container-security-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: container-security-operator.v3.12.12
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:1a15328f0b2c64d3a857f2f155ad4def095027e809610de98545e303eec63225
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.12.12
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:1a15328f0b2c64d3a857f2f155ad4def095027e809610de98545e303eec63225
        version: 3.12.12
      entries:
      - name: container-security-operator.v3.12.12
        version: 3.12.12
      - name: container-security-operator.v3.12.11
        version: 3.12.11
      - name: container-security-operator.v3.12.10
        version: 3.12.10
      - name: container-security-operator.v3.12.9
        version: 3.12.9
      - name: container-security-operator.v3.12.8
        version: 3.12.8
      - name: container-security-operator.v3.12.7
        version: 3.12.7
      - name: container-security-operator.v3.12.6
        version: 3.12.6
      - name: container-security-operator.v3.12.5
        version: 3.12.5
      - name: container-security-operator.v3.12.4
        version: 3.12.4
      - name: container-security-operator.v3.12.3
        version: 3.12.3
      - name: container-security-operator.v3.12.2
        version: 3.12.2
      - name: container-security-operator.v3.12.1
        version: 3.12.1
      - name: container-security-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: container-security-operator.v3.13.8
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:ba50621739581da9a661c2c2e355849c328cfa33b7311b06abb148e1183a6321
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.13.8
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:ba50621739581da9a661c2c2e355849c328cfa33b7311b06abb148e1183a6321
        version: 3.13.8
      entries:
      - name: container-security-operator.v3.13.8
        version: 3.13.8
      - name: container-security-operator.v3.13.7
        version: 3.13.7
      - name: container-security-operator.v3.13.6
        version: 3.13.6
      - name: container-security-operator.v3.13.5
        version: 3.13.5
      - name: container-security-operator.v3.13.4
        version: 3.13.4
      - name: container-security-operator.v3.13.3
        version: 3.13.3
      - name: container-security-operator.v3.13.2
        version: 3.13.2
      - name: container-security-operator.v3.13.1
        version: 3.13.1
      - name: container-security-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: container-security-operator.v3.14.5
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:c42487ccf098df3ec6dbeb5d409346bdd30e92781cf3ffea5fdb6dd0b196d0cf
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.14.5
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:c42487ccf098df3ec6dbeb5d409346bdd30e92781cf3ffea5fdb6dd0b196d0cf
        version: 3.14.5
      entries:
      - name: container-security-operator.v3.14.5
        version: 3.14.5
      - name: container-security-operator.v3.14.4
        version: 3.14.4
      - name: container-security-operator.v3.14.3
        version: 3.14.3
      - name: container-security-operator.v3.14.2
        version: 3.14.2
      - name: container-security-operator.v3.14.1
        version: 3.14.1
      - name: container-security-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: container-security-operator.v3.15.2
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:81e929ec095f791a83310ee7fcda239d5c348ba69a3cf5adcf1b3a77c349ce77
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.15.2
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:81e929ec095f791a83310ee7fcda239d5c348ba69a3cf5adcf1b3a77c349ce77
        version: 3.15.2
      entries:
      - name: container-security-operator.v3.15.2
        version: 3.15.2
      - name: container-security-operator.v3.15.1
        version: 3.15.1
      - name: container-security-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: container-security-operator.v3.6.10
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:641b80df8d7102d49a5dfd5b99245669f60393880a6f1bc8ba8450b88a640d96
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.6.10
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:641b80df8d7102d49a5dfd5b99245669f60393880a6f1bc8ba8450b88a640d96
        version: 3.6.10
      entries:
      - name: container-security-operator.v3.6.10
        version: 3.6.10
      - name: container-security-operator.v3.6.9
        version: 3.6.9
      - name: container-security-operator.v3.6.8
        version: 3.6.8
      - name: container-security-operator.v3.6.7
        version: 3.6.7
      - name: container-security-operator.v3.6.6
        version: 3.6.6
      - name: container-security-operator.v3.6.5
        version: 3.6.5
      - name: container-security-operator.v3.6.4
        version: 3.6.4
      - name: container-security-operator.v3.6.2
        version: 3.6.2
      - name: container-security-operator.v3.6.1
        version: 3.6.1
      - name: container-security-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: container-security-operator.v3.7.14
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:10736a06beb503b4eb73260a87a2cde9bd83353374ee0ac9e8dd0a6260cf2665
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.7.14
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:10736a06beb503b4eb73260a87a2cde9bd83353374ee0ac9e8dd0a6260cf2665
        version: 3.7.14
      entries:
      - name: container-security-operator.v3.7.14
        version: 3.7.14
      - name: container-security-operator.v3.7.13
        version: 3.7.13
      - name: container-security-operator.v3.7.12
        version: 3.7.12
      - name: container-security-operator.v3.7.11
        version: 3.7.11
      - name: container-security-operator.v3.7.10
        version: 3.7.10
      - name: container-security-operator.v3.7.9
        version: 3.7.9
      - name: container-security-operator.v3.7.8
        version: 3.7.8
      - name: container-security-operator.v3.7.7
        version: 3.7.7
      - name: container-security-operator.v3.7.6
        version: 3.7.6
      - name: container-security-operator.v3.7.5
        version: 3.7.5
      - name: container-security-operator.v3.7.4
        version: 3.7.4
      - name: container-security-operator.v3.7.3
        version: 3.7.3
      - name: container-security-operator.v3.7.2
        version: 3.7.2
      - name: container-security-operator.v3.7.1
        version: 3.7.1
      - name: container-security-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: container-security-operator.v3.8.15
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:108bd033543beb7cce077130c96f992c202f7a258229c239795f8f19e5fae5e5
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.8.15
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:108bd033543beb7cce077130c96f992c202f7a258229c239795f8f19e5fae5e5
        version: 3.8.15
      entries:
      - name: container-security-operator.v3.8.15
        version: 3.8.15
      - name: container-security-operator.v3.8.14
        version: 3.8.14
      - name: container-security-operator.v3.8.13
        version: 3.8.13
      - name: container-security-operator.v3.8.12
        version: 3.8.12
      - name: container-security-operator.v3.8.11
        version: 3.8.11
      - name: container-security-operator.v3.8.10
        version: 3.8.10
      - name: container-security-operator.v3.8.9
        version: 3.8.9
      - name: container-security-operator.v3.8.8
        version: 3.8.8
      - name: container-security-operator.v3.8.7
        version: 3.8.7
      - name: container-security-operator.v3.8.6
        version: 3.8.6
      - name: container-security-operator.v3.8.5
        version: 3.8.5
      - name: container-security-operator.v3.8.4
        version: 3.8.4
      - name: container-security-operator.v3.8.3
        version: 3.8.3
      - name: container-security-operator.v3.8.2
        version: 3.8.2
      - name: container-security-operator.v3.8.1
        version: 3.8.1
      - name: container-security-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: container-security-operator.v3.9.16
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:df8e1d246273d21049c6bb8169a2e5f3fc8d69c85c564359abea7364969c6acc
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.9.16
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:df8e1d246273d21049c6bb8169a2e5f3fc8d69c85c564359abea7364969c6acc
        version: 3.9.16
      entries:
      - name: container-security-operator.v3.9.16
        version: 3.9.16
      - name: container-security-operator.v3.9.15
        version: 3.9.15
      - name: container-security-operator.v3.9.14
        version: 3.9.14
      - name: container-security-operator.v3.9.13
        version: 3.9.13
      - name: container-security-operator.v3.9.12
        version: 3.9.12
      - name: container-security-operator.v3.9.11
        version: 3.9.11
      - name: container-security-operator.v3.9.10
        version: 3.9.10
      - name: container-security-operator.v3.9.9
        version: 3.9.9
      - name: container-security-operator.v3.9.8
        version: 3.9.8
      - name: container-security-operator.v3.9.7
        version: 3.9.7
      - name: container-security-operator.v3.9.6
        version: 3.9.6
      - name: container-security-operator.v3.9.5
        version: 3.9.5
      - name: container-security-operator.v3.9.4
        version: 3.9.4
      - name: container-security-operator.v3.9.3
        version: 3.9.3
      - name: container-security-operator.v3.9.2
        version: 3.9.2
      - name: container-security-operator.v3.9.1
        version: 3.9.1
      - name: container-security-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: container-security-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: prometurbo-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometurbo-operator.v8.17.40-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5-beta.1 <8.17.40-beta.1'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
          - kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.

          ### Get Started
          Please follow [this](https://www.ibm.com/docs/en/tarm/8.15.x?topic=configuration-prometheus) for an overview of Prometurbo and deployment instructions.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/prometurbo
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/prometurbo-operator@sha256:a1605506884f6bce7471d79c6243728bce991ceb381d7fe0c51f1cdcbeb96e22
        version: 8.17.40-beta.1
      entries:
      - name: prometurbo-operator.v8.17.40-beta.1
        version: 8.17.40-beta.1
      name: beta
    - currentCSV: prometurbo-operator.v8.17.40
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5 <8.17.40'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
          - kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.

          ### Get Started
          Please follow [this](https://www.ibm.com/docs/en/tarm/8.15.x?topic=configuration-prometheus) for an overview of Prometurbo and deployment instructions.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/prometurbo
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/prometurbo-operator@sha256:cbcf90e44ffcedec8e3f161eff4c95b581924a1db195bf55d13da3714346b26f
        version: 8.17.40
      entries:
      - name: prometurbo-operator.v8.17.40
        version: 8.17.40
      name: stable
    defaultChannel: stable
    packageName: prometurbo-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: AppDynamics LLC
      provider-url: https://www.appdynamics.com/
    name: appdynamics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appdynamics-operator.v24.7.0-861
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clusteragent",
                "metadata": {
                  "name": "k8s-cluster-agent",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "appName": "<app-name>",
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/cluster-agent:tag"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "InfraViz",
                "metadata": {
                  "name": "appdynamics-infraviz",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/machine-agent:latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3c420cce8b626d8f5cf424410527c6a7a17ca3348e903dd8cdfcd74819b8de41
          createdAt: "2023-02-06T14:26:07Z"
          description: AppDynamics Operator to install AppDynamics Cluster Agent &
            AppDynamics Machine Agent for K8s Monitoring for various Cloud Providers
            like EKS, AKS etc.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: support@appdynamics.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clusteragent is the Schema for the clusteragents API
            displayName: Clusteragent
            kind: Clusteragent
            name: clusteragents.cluster.appdynamics.com
            version: v1alpha1
          - description: InfraViz is the Schema for the infravizs API
            displayName: Infra Viz
            kind: InfraViz
            name: infravizs.cluster.appdynamics.com
            version: v1alpha1
        description: |-
          AppDynamics Operator to install AppDynamics Cluster Agent & Machine Agent for K8s Monitoring on various Cloud Providers like EKS, AKS, GCP, Openshift etc.
          **About AppDynamics Operator**
          AppDynamics Operator simplifies the installation and lifecycle management of the AppDynamics ClusterAgent and the AppDynamics Machine Agent for K8s Monitoring on different Cloud Providers like EKS, AKS, GCP, OpenShift etc.
          **Installation Steps**
          Please refer below links for installation guide:
          [Install AppDynamics Cluster Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli)
          [Install AppDynamics Machine Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli)
          [Install AppDynamics Cluster Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts)
          [Install Appdynamics Machine Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts)
        displayName: AppDynamics Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AppDynamics
        - Operator
        - AppDynamics Cluster Agent
        - AppDynamics Machine Agent
        - AppDynamics K8s Monitoring
        links:
        - name: AppDynamics Kubernetes Monitoring Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent
        - name: AppDynamics Cluster Agent Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/overview-of-cluster-monitoring
        - name: AppDynamics Cluster Agent Requirements
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/cluster-agent-requirements-and-supported-environments
        - name: Install AppDynamics Cluster Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli
        - name: Install AppDynamics Cluster Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts
        - name: Install AppDynamics Machine Agent & NetViz Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli
        - name: Install AppDynamics Machine Agent & NetViz Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: AppDynamics LLC
          url: https://www.appdynamics.com/
        relatedImages:
        - registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3c420cce8b626d8f5cf424410527c6a7a17ca3348e903dd8cdfcd74819b8de41
        version: 24.7.0-861
      entries:
      - name: appdynamics-operator.v24.7.0-861
        version: 24.7.0-861
      name: alpha
    - currentCSV: appdynamics-operator.v25.9.0-1218
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clusteragent",
                "metadata": {
                  "name": "k8s-cluster-agent",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "appName": "<app-name>",
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/cluster-agent:tag"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "InfraViz",
                "metadata": {
                  "name": "appdynamics-infraviz",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/machine-agent:latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3320c2c8d5f60adc541befd099dac1ef96257b16e0df28b727b0e589f78d981a
          createdAt: "2023-02-06T14:26:07Z"
          description: AppDynamics Operator to install AppDynamics Cluster Agent &
            AppDynamics Machine Agent for K8s Monitoring for various Cloud Providers
            like EKS, AKS etc.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: support@appdynamics.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clusteragent is the Schema for the clusteragents API
            displayName: Clusteragent
            kind: Clusteragent
            name: clusteragents.cluster.appdynamics.com
            version: v1alpha1
          - description: InfraViz is the Schema for the infravizs API
            displayName: Infra Viz
            kind: InfraViz
            name: infravizs.cluster.appdynamics.com
            version: v1alpha1
        description: |-
          AppDynamics Operator to install AppDynamics Cluster Agent & Machine Agent for K8s Monitoring on various Cloud Providers like EKS, AKS, GCP, Openshift etc.
          **About AppDynamics Operator**
          AppDynamics Operator simplifies the installation and lifecycle management of the AppDynamics ClusterAgent and the AppDynamics Machine Agent for K8s Monitoring on different Cloud Providers like EKS, AKS, GCP, OpenShift etc.
          **Installation Steps**
          Please refer below links for installation guide:
          [Install AppDynamics Cluster Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli)
          [Install AppDynamics Machine Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli)
          [Install AppDynamics Cluster Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts)
          [Install Appdynamics Machine Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts)
        displayName: AppDynamics Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - AppDynamics
        - Operator
        - AppDynamics Cluster Agent
        - AppDynamics Machine Agent
        - AppDynamics K8s Monitoring
        links:
        - name: AppDynamics Kubernetes Monitoring Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent
        - name: AppDynamics Cluster Agent Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/overview-of-cluster-monitoring
        - name: AppDynamics Cluster Agent Requirements
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/cluster-agent-requirements-and-supported-environments
        - name: Install AppDynamics Cluster Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli
        - name: Install AppDynamics Cluster Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts
        - name: Install AppDynamics Machine Agent & NetViz Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli
        - name: Install AppDynamics Machine Agent & NetViz Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: AppDynamics LLC
          url: https://www.appdynamics.com/
        relatedImages:
        - registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3320c2c8d5f60adc541befd099dac1ef96257b16e0df28b727b0e589f78d981a
        version: 25.9.0-1218
      entries:
      - name: appdynamics-operator.v25.9.0-1218
        version: 25.9.0-1218
      name: stable
    defaultChannel: alpha
    packageName: appdynamics-operator
    provider:
      name: AppDynamics LLC
      url: https://www.appdynamics.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: www.redhat.com
    name: odf-node-recovery-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-node-recovery-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "NodeRecovery",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "odf-node-recovery-operator",
                    "app.kubernetes.io/instance": "noderecovery-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "noderecovery",
                    "app.kubernetes.io/part-of": "odf-node-recovery-operator"
                  },
                  "name": "noderecovery-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Storage,OpenShift Optional
          description: ODF Node Recovery is an operator that assist in the recovery
            of and ODF cluster that has a device or a node replaced
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/jordigilh/odf-node-recovery-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeRecovery is the Schema for the noderecoveries API
            displayName: Node Recovery
            kind: NodeRecovery
            name: noderecoveries.odf.openshift.io
            version: v1alpha1
        description: |
          ODF Node Recovery is an operator that assist in the recovery of an
          ODF cluster in these 2 scenarios:
          * <a href="https://docs.redhat.com/en/documentation/red_hat_openshift_data_foundation/4.14/html-single/replacing_devices/index"> Device replacement<a/>
          * <a href="https://docs.redhat.com/en/documentation/red_hat_openshift_data_foundation/4.14/html-single/replacing_nodes/index"> Node replacement</a>
        displayName: ODF Node Recovery Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - node recovery
        links:
        - name: GitHub Repository
          url: https://github.com/jordigilh/odf-node-recovery-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
          url: www.redhat.com
        relatedImages:
        - quay.io/jordigilh/odf-node-recovery-controller@sha256:4164fafa9bb1c4d6121f4c5fc700fa2827c3994ca98a65d641471a2a46d4f59c
        version: 1.1.0
      entries:
      - name: odf-node-recovery-operator.v1.1.0
        version: 1.1.0
      - name: odf-node-recovery-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: odf-node-recovery-operator
    provider:
      name: Red Hat, Inc.
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: authorino-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: authorino-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/authorino-operator:v0.8.0
          createdAt: 2021-12-08T10-00-00Z
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta1
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: The operator to manage instances of Authorino
        displayName: Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: Authorino Operator
          url: https://github.com/Kuadrant/authorino-operator
        - name: Authorino
          url: https://github.com/Kuadrant/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/authorino-operator:v0.8.0
        version: 0.8.0
      entries:
      - name: authorino-operator.v0.8.0
        version: 0.8.0
      - name: authorino-operator.v0.7.0
        version: 0.7.0
      - name: authorino-operator.v0.6.0
        version: 0.6.0
      - name: authorino-operator.v0.5.0
        version: 0.5.0
      - name: authorino-operator.v0.4.1
        version: 0.4.1
      - name: authorino-operator.v0.4.0
        version: 0.4.0
      - name: authorino-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: authorino-operator.v0.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta3",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/authorino-operator:v0.16.0
          createdAt: "2024-11-25T13:22:24Z"
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta3
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: The operator to manage instances of Authorino
        displayName: Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: Authorino Operator
          url: https://github.com/Kuadrant/authorino-operator
        - name: Authorino
          url: https://github.com/Kuadrant/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/authorino-operator:v0.16.0
        - quay.io/kuadrant/authorino:v0.20.0
        version: 0.16.0
      entries:
      - name: authorino-operator.v0.16.0
        version: 0.16.0
      - name: authorino-operator.v0.15.1
        version: 0.15.1
      - name: authorino-operator.v0.15.0
        version: 0.15.0
      - name: authorino-operator.v0.14.0
        version: 0.14.0
      - name: authorino-operator.v0.13.1
        version: 0.13.1
      - name: authorino-operator.v0.13.0
        version: 0.13.0
      - name: authorino-operator.v0.12.0
        version: 0.12.0
      - name: authorino-operator.v0.11.1
        version: 0.11.1
      - name: authorino-operator.v0.10.0
        version: 0.10.0
      - name: authorino-operator.v0.9.0
        version: 0.9.0
      - name: authorino-operator.v0.8.0
        version: 0.8.0
      - name: authorino-operator.v0.7.0
        version: 0.7.0
      - name: authorino-operator.v0.6.0
        version: 0.6.0
      - name: authorino-operator.v0.5.0
        version: 0.5.0
      - name: authorino-operator.v0.4.1
        version: 0.4.1
      - name: authorino-operator.v0.4.0
        version: 0.4.0
      - name: authorino-operator.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: authorino-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Eclipse Foundation
      provider-url: ""
    name: eclipse-che
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eclipse-che.v7.110.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "org.eclipse.che/v1",
                "kind": "CheCluster",
                "metadata": {
                  "name": "eclipse-che",
                  "namespace": "eclipse-che"
                },
                "spec": {
                  "auth": {
                    "identityProviderURL": "",
                    "oAuthClientName": "",
                    "oAuthSecret": ""
                  },
                  "database": {
                    "externalDb": false
                  },
                  "k8s": {
                    "ingressDomain": null,
                    "tlsSecretName": null
                  },
                  "metrics": {
                    "enable": true
                  },
                  "server": {
                    "externalDevfileRegistries": [
                      {
                        "url": "https://registry.devfile.io"
                      }
                    ],
                    "externalDevfileRegistry": true,
                    "externalPluginRegistry": true,
                    "workspaceNamespaceDefault": "<username>-che"
                  },
                  "storage": {
                    "pvcStrategy": "common"
                  }
                }
              },
              {
                "apiVersion": "org.eclipse.che/v2",
                "kind": "CheCluster",
                "metadata": {
                  "name": "eclipse-che",
                  "namespace": "eclipse-che"
                },
                "spec": {
                  "components": {
                    "devfileRegistry": {
                      "disableInternalRegistry": true,
                      "externalDevfileRegistries": [
                        {
                          "url": "https://registry.devfile.io"
                        }
                      ]
                    },
                    "pluginRegistry": {
                      "disableInternalRegistry": true
                    }
                  },
                  "containerRegistry": {},
                  "devEnvironments": {},
                  "gitServices": {},
                  "networking": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/eclipse/che-operator@sha256:10c4d05c2cb848d04b97caf752c7f4d181e6973683263e45adec417342164e51
          createdAt: "2025-10-02T14:44:20Z"
          description: A Kube-native development solution that delivers portable and
            collaborative developer workspaces.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eclipse-che/che-operator
          support: Eclipse Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              The `CheCluster` custom resource allows defining and managing Eclipse Che server installation.
              Based on these settings, the  Operator automatically creates and maintains several ConfigMaps:
              `che`, `plugin-registry` that will contain the appropriate environment variables
              of the various components of the installation. These generated ConfigMaps must NOT be updated manually.
            displayName: Eclipse Che instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v2
          - description: The `CheCluster` custom resource allows defining and managing
              a Che server installation
            displayName: Eclipse Che instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v1
        description: |
          A collaborative Kubernetes-native development solution that delivers OpenShift workspaces and in-browser IDE for rapid cloud application development.
          This operator installs the Plugin and Devfile registries, Dashboard, Gateway and the Eclipse Che server, and configures these services.
          OpenShift OAuth is used directly for authentication. TLS mode is on.

          ## How to Install
          Press the **Install** button, choose the channel and the upgrade strategy, and wait for the **Installed** Operator status.
          When the operator is installed, create a new Custom Resource (CR) of Kind CheCluster (click the **Create New** button).
          The CR spec contains all defaults. You can start using Eclipse Che when the CR status is set to **Available**, and you see a URL to Eclipse Che.

          ## Defaults
          By default, the operator deploys Eclipse Che with:
          * 10Gi storage
          * Auto-generated passwords
          * Bundled Plugin and Devfile registries

          Use `oc edit checluster/eclipse-che -n eclipse-che` to update Eclipse Che default installation options.
          See more in the [Installation guide](https://www.eclipse.org/che/docs/stable/administration-guide/configuring-che/).

          ### Certificates
          Operator uses a default router certificate to secure Eclipse Che routes.
          Follow the [guide](https://www.eclipse.org/che/docs/stable/administration-guide/importing-untrusted-tls-certificates/)
          to import certificates into Eclipse Che.
        displayName: Eclipse Che
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eclipse che
        - workspaces
        - devtools
        - developer
        - ide
        - java
        links:
        - name: Product Page
          url: https://www.eclipse.org/che
        - name: Documentation
          url: https://www.eclipse.org/che/docs
        - name: Operator GitHub Repo
          url: https://github.com/eclipse-che/che-operator
        maintainers:
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Eclipse Foundation
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy@sha256:738c854322f56d63ebab75de5210abcdd5e0782ce2d30c0ecd4620f63b24694d
        - quay.io/devfile/universal-developer-image@sha256:3cf53fb9eab60ee3156e9edd4ec5af72b70e55d5ffb28f41ecb5b26e01dc9838
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:354fc75eb7a21a934381e93d03ef9d42bc2c8ae8989fdcacecfb39b863b96ced
        - quay.io/eclipse/che-operator@sha256:10c4d05c2cb848d04b97caf752c7f4d181e6973683263e45adec417342164e51
        - quay.io/eclipse/che--traefik@sha256:4cf907247939b5d20bf4eff73abd21cb413c339600dde76dbc94a874b2578a27
        - quay.io/che-incubator/che-code@sha256:cfcfbe4d665b18b11401c3c35b23b04d292c31e7f4be6d78a5e6d60fce776ed0
        - quay.io/che-incubator/che-code-server@sha256:357c95297ce2182ef0a792038ea122075fdeb6c8e835099be09e8d4f105eac2a
        - quay.io/eclipse/che-tls-secret-creator@sha256:cbb2fd8ce903d822377f2093af1aa2543672ed8c516d607a7a9dd857379dac5d
        - quay.io/che-incubator/configbump@sha256:396714521307ae9f79453c492fb8064cf940cd6b4e6183a825f1056045623874
        - quay.io/che-incubator/che-idea-dev-server@sha256:87c8b53cc295e3d7eb35b326e221a949553be3345135a4b6c1999d66f345a8b2
        - quay.io/che-incubator/che-code-sshd@sha256:7eb5381fba8878b8f252523caf70c9957b9f9e699db4564061017fb0a023cc1a
        - quay.io/che-incubator/che-idea@sha256:8aae69dc4b0c122491a75400639af0fe92b5e214c6e68ac97cda29fb58b44151
        - quay.io/eclipse/che-plugin-registry@sha256:a70a0d98505ba5aeed691fc70c3c59aed3304155fba0ab3ad70a6d0453b22093
        - quay.io/che-incubator/che-code-server@sha256:2093baab1a73e9682f5892f80582481abfbef99d744472ee6098e2d82fb2899f
        - quay.io/che-incubator/header-rewrite-proxy@sha256:bd7873b8feef35f218f54c6251ea224bea2c8bf202a328230019a0ba2941245d
        - quay.io/oauth2-proxy/oauth2-proxy@sha256:dcb6ff8dd21bf3058f6a22c6fa385fa5b897a9cd3914c88a2cc2bb0a85f8065d
        - quay.io/che-incubator/che-code@sha256:ac62a2f42ae6d6f6628620a00133afece954fcbed2df309fa3de79e4c28cc029
        - quay.io/openshift/origin-oauth-proxy@sha256:870bfe92a4663720775c0dfe5728ecbb10a17f0644eef5f57276ec135034c6a1
        - quay.io/eclipse/che-dashboard@sha256:2db3255752aebd75dceea578cd27b7e8ecb41c2290827ff7104ed77e1b61feb9
        - quay.io/eclipse/che-server@sha256:b83d536cdd3cb2b86def5599495c40870dde20c194a003736a2b47a5e5e17a06
        version: 7.110.0
      entries:
      - name: eclipse-che.v7.110.0
        version: 7.110.0
      - name: eclipse-che.v7.109.0
        version: 7.109.0
      - name: eclipse-che.v7.108.0
        version: 7.108.0
      - name: eclipse-che.v7.107.0
        version: 7.107.0
      - name: eclipse-che.v7.106.0
        version: 7.106.0
      - name: eclipse-che.v7.105.0
        version: 7.105.0
      - name: eclipse-che.v7.104.0
        version: 7.104.0
      - name: eclipse-che.v7.103.0
        version: 7.103.0
      - name: eclipse-che.v7.102.0
        version: 7.102.0
      - name: eclipse-che.v7.101.0
        version: 7.101.0
      - name: eclipse-che.v7.100.0
        version: 7.100.0
      - name: eclipse-che.v7.99.0
        version: 7.99.0
      - name: eclipse-che.v7.98.0
        version: 7.98.0
      - name: eclipse-che.v7.97.0
        version: 7.97.0
      - name: eclipse-che.v7.96.0
        version: 7.96.0
      - name: eclipse-che.v7.95.0
        version: 7.95.0
      - name: eclipse-che.v7.94.0
        version: 7.94.0
      - name: eclipse-che.v7.93.0
        version: 7.93.0
      - name: eclipse-che.v7.92.0
        version: 7.92.0
      - name: eclipse-che.v7.91.0
        version: 7.91.0
      - name: eclipse-che.v7.90.0
        version: 7.90.0
      - name: eclipse-che.v7.89.0
        version: 7.89.0
      - name: eclipse-che.v7.88.0
        version: 7.88.0
      - name: eclipse-che.v7.87.0
        version: 7.87.0
      - name: eclipse-che.v7.86.0
        version: 7.86.0
      - name: eclipse-che.v7.85.0
        version: 7.85.0
      - name: eclipse-che.v7.84.0
        version: 7.84.0
      - name: eclipse-che.v7.83.0
        version: 7.83.0
      - name: eclipse-che.v7.82.0
        version: 7.82.0
      - name: eclipse-che.v7.81.0
        version: 7.81.0
      - name: eclipse-che.v7.80.0
        version: 7.80.0
      - name: eclipse-che.v7.79.0
        version: 7.79.0
      - name: eclipse-che.v7.78.0
        version: 7.78.0
      - name: eclipse-che.v7.77.0
        version: 7.77.0
      - name: eclipse-che.v7.76.0
        version: 7.76.0
      - name: eclipse-che.v7.75.0
        version: 7.75.0
      - name: eclipse-che.v7.74.0
        version: 7.74.0
      - name: eclipse-che.v7.73.0
        version: 7.73.0
      - name: eclipse-che.v7.72.0
        version: 7.72.0
      - name: eclipse-che.v7.71.0
        version: 7.71.0
      - name: eclipse-che.v7.70.0
        version: 7.70.0
      - name: eclipse-che.v7.69.0
        version: 7.69.0
      - name: eclipse-che.v7.68.0
        version: 7.68.0
      - name: eclipse-che.v7.67.0
        version: 7.67.0
      - name: eclipse-che.v7.66.0
        version: 7.66.0
      - name: eclipse-che.v7.65.0
        version: 7.65.0
      - name: eclipse-che.v7.64.0
        version: 7.64.0
      - name: eclipse-che.v7.63.0
        version: 7.63.0
      - name: eclipse-che.v7.62.0
        version: 7.62.0
      - name: eclipse-che.v7.61.0
        version: 7.61.0
      - name: eclipse-che.v7.60.1
        version: 7.60.1
      - name: eclipse-che.v7.60.0
        version: 7.60.0
      - name: eclipse-che.v7.59.0
        version: 7.59.0
      - name: eclipse-che.v7.58.0
        version: 7.58.0
      - name: eclipse-che.v7.57.0
        version: 7.57.0
      - name: eclipse-che.v7.56.0
        version: 7.56.0
      - name: eclipse-che.v7.55.0
        version: 7.55.0
      - name: eclipse-che.v7.54.0
        version: 7.54.0
      - name: eclipse-che.v7.53.0
        version: 7.53.0
      - name: eclipse-che.v7.52.0
        version: 7.52.0
      - name: eclipse-che.v7.51.0
        version: 7.51.0
      - name: eclipse-che.v7.50.0
        version: 7.50.0
      - name: eclipse-che.v7.49.0
        version: 7.49.0
      - name: eclipse-che.v7.48.1
        version: 7.48.1
      - name: eclipse-che.v7.47.0
        version: 7.47.0
      - name: eclipse-che.v7.46.0
        version: 7.46.0
      - name: eclipse-che.v7.45.0
        version: 7.45.0
      - name: eclipse-che.v7.44.0
        version: 7.44.0
      - name: eclipse-che.v7.42.0
        version: 7.42.0
      - name: eclipse-che.v7.41.2
        version: 7.41.2
      - name: eclipse-che.v7.41.1
        version: 7.41.1
      - name: eclipse-che.v7.41.0
        version: 7.41.0
      - name: eclipse-che.v7.40.2
        version: 7.40.2
      - name: eclipse-che.v7.40.1
        version: 7.40.1
      name: stable
    defaultChannel: stable
    packageName: eclipse-che
    provider:
      name: Eclipse Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solution-Soft
      provider-url: www.solution-soft.com
    name: timemachine-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: timemachine-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tm.solution-soft.com/v1alpha1",
                "kind": "TimeMachine",
                "metadata": {
                  "name": "timemachine-sample"
                },
                "spec": {
                  "cluster_provider": "ibm",
                  "aws": {
                    "shared_volume_name": ""
                  },
                  "ibm": {
                    "persistentvolume": {
                      "storageclass": "ibmc-file-silver-gid"
                    }
                  },
                  "tm_lichost": "52.152.148.40",
                  "tm_licpass": "qwaszx",
                  "tm_licport": "57777",
                  "imagePullSecrets": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
          createdAt: "2024-01-30T10:20:30Z"
          description: Time Machine Operator enables time travel of OpenShift pods,
            without the need to change the system clock on any of the cluster nodes
          olm.skipRange: '>=0.0.4 <1.0.17'
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: SolutionSoft Systems, Inc. - support@solution-soft.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Time Machine
            displayName: Time Machine
            kind: TimeMachine
            name: timemachines.tm.solution-soft.com
            version: v1alpha1
        description: |-
          Time Machine operator enables time travel of OpenShift pods/deployments.

          To time travel a pod where your application is, you just need to install the Time Machine operator and create a Time Machine custom resource in the same namespace, without the need of changing the container images of the application you’re using, or adding containers to the pod with your application.

          Time travelling is done on the namespace level where all the (configured) target pods/deployments see the same virtual time.
        displayName: Time Machine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - time shift
        - time travel
        - virtual clocks
        - time machine
        - virtual time
        - solution-soft
        links:
        - name: Timemachine Operator
          url: https://catalog.redhat.com/software/operators/detail/5e98727a6c5dcb34dfbb19e3
        maintainers:
        - email: dtrajkovic@solution-soft.com
          name: Dragan Trajkovic
        maturity: alpha
        provider:
          name: Solution-Soft
          url: www.solution-soft.com
        relatedImages:
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
        - registry.connect.redhat.com/rhel-timemachine-container/virtual-clock-testing-tool@sha256:a81a823328b0ee795c40668c89a08bcf545c5192d9eed0146071e3721a0a8372
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0c304c11986790f446604319dfc29db494b15f5f9fa13f2c436688ec29c92dbe
        version: 1.0.17
      entries:
      - name: timemachine-operator.v1.0.17
        version: 1.0.17
      - name: timemachine-operator.v1.0.16
        version: 1.0.16
      name: alpha
    defaultChannel: alpha
    packageName: timemachine-operator
    provider:
      name: Solution-Soft
      url: www.solution-soft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: antrea.io
      provider-url: ""
    name: antrea-by-broadcom-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: antrea-by-broadcom-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.antrea.vmware.com/v1",
                "kind": "AntreaInstall",
                "metadata": {
                  "name": "antrea-install",
                  "namespace": "antrea-operator"
                },
                "spec": {
                  "antreaAgentConfig": "# FeatureGates is a map of feature names to bools that enable or disable experimental features.\nfeatureGates:\n# AllAlpha is a global toggle for alpha features. Per-feature key values override the default set by AllAlpha.\n#  AllAlpha: false\n\n# AllBeta is a global toggle for beta features. Per-feature key values override the default set by AllBeta.\n#  AllBeta: false\n\n# Enable AntreaProxy which provides ServiceLB for in-cluster Services in antrea-agent.\n# It should be enabled on Windows, otherwise NetworkPolicy will not take effect on\n# Service traffic.\n#  AntreaProxy: true\n\n# Enable EndpointSlice support in AntreaProxy. Don't enable this feature unless that EndpointSlice\n# API version v1 is supported and set as enabled in Kubernetes. If AntreaProxy is not enabled,\n# this flag will not take effect.\n#  EndpointSlice: false\n\n# Enable TopologyAwareHints in AntreaProxy. This requires AntreaProxy and EndpointSlice to be\n# enabled, otherwise this flag will not take effect.\n#  TopologyAwareHints: false\n\n# Enable traceflow which provides packet tracing feature to diagnose network issue.\n#  Traceflow: true\n\n# Enable NodePortLocal feature to make the Pods reachable externally through NodePort\n#  NodePortLocal: true\n\n# Enable Antrea ClusterNetworkPolicy feature to complement K8s NetworkPolicy for cluster admins\n# to define security policies which apply to the entire cluster, and Antrea NetworkPolicy\n# feature that supports priorities, rule actions and externalEntities in the future.\n#  AntreaPolicy: true\n\n# Enable flowexporter which exports polled conntrack connections as IPFIX flow records from each\n# agent to a configured collector.\n#  FlowExporter: false\n\n# Enable collecting and exposing NetworkPolicy statistics.\n#  NetworkPolicyStats: true\n\n# Enable controlling SNAT IPs of Pod egress traffic.\n#  Egress: true\n\n# Enable AntreaIPAM, which can allocate IP addresses from IPPools. AntreaIPAM is required by the\n# bridging mode and allocates IPs to Pods in bridging mode. It is also required to use Antrea for\n# IPAM when configuring secondary network interfaces with Multus.\n#  AntreaIPAM: false\n\n# Enable multicast traffic.\n#  Multicast: false\n\n# Enable Antrea Multi-cluster Gateway to support cross-cluster traffic.\n# This feature is supported only with encap mode.\n#  Multicluster: false\n\n# Enable support for provisioning secondary network interfaces for Pods (using\n# Pod annotations). At the moment, Antrea can only create secondary network\n# interfaces using SR-IOV VFs on baremetal Nodes.\n#  SecondaryNetwork: false\n\n# Enable managing external IPs of Services of LoadBalancer type.\n#  ServiceExternalIP: false\n\n# Enable mirroring or redirecting the traffic Pods send or receive.\n#  TrafficControl: false\n\n# Enable certificated-based authentication for IPsec.\n#  IPsecCertAuth: false\n\n# Enable collecting support bundle files with SupportBundleCollection CRD.\n#  SupportBundleCollection: false\n\n# Enable users to protect their applications by specifying how they are allowed to communicate with others, taking\n# into account application context.\n#  L7NetworkPolicy: false\n\n# Name of the OpenVSwitch bridge antrea-agent will create and use.\n# Make sure it doesn't conflict with your existing OpenVSwitch bridges.\novsBridge: \"br-int\"\n\n# Datapath type to use for the OpenVSwitch bridge created by Antrea. At the moment, the only\n# supported value is 'system', which corresponds to the kernel datapath.\n#ovsDatapathType: system\n\n# Name of the interface antrea-agent will create and use for host \u003c--\u003e pod communication.\n# Make sure it doesn't conflict with your existing interfaces.\nhostGateway: \"antrea-gw0\"\n\n# Determines how traffic is encapsulated. It has the following options:\n# encap(default):    Inter-node Pod traffic is always encapsulated and Pod to external network\n#                    traffic is SNAT'd.\n# noEncap:           Inter-node Pod traffic is not encapsulated; Pod to external network traffic is\n#                    SNAT'd if noSNAT is not set to true. Underlying network must be capable of\n#                    supporting Pod traffic across IP subnets.\n# hybrid:            noEncap if source and destination Nodes are on the same subnet, otherwise encap.\n# networkPolicyOnly: Antrea enforces NetworkPolicy only, and utilizes CNI chaining and delegates Pod\n#                    IPAM and connectivity to the primary CNI.\n#\ntrafficEncapMode: \"encap\"\n\n# Whether or not to SNAT (using the Node IP) the egress traffic from a Pod to the external network.\n# This option is for the noEncap traffic mode only, and the default value is false. In the noEncap\n# mode, if the cluster's Pod CIDR is reachable from the external network, then the Pod traffic to\n# the external network needs not be SNAT'd. In the networkPolicyOnly mode, antrea-agent never\n# performs SNAT and this option will be ignored; for other modes it must be set to false.\nnoSNAT: false\n\n# Tunnel protocols used for encapsulating traffic across Nodes. If WireGuard is enabled in trafficEncryptionMode,\n# this option will not take effect. Supported values:\n# - geneve (default)\n# - vxlan\n# - gre\n# - stt\n# Note that \"gre\" is not supported for IPv6 clusters (IPv6-only or dual-stack clusters).\ntunnelType: \"geneve\"\n\n# TunnelPort is the destination port for UDP and TCP based tunnel protocols (Geneve, VXLAN, and STT).\n# If zero, it will use the assigned IANA port for the protocol, i.e. 6081 for Geneve, 4789 for VXLAN,\n# and 7471 for STT.\ntunnelPort: 0\n\n# TunnelCsum determines whether to compute UDP encapsulation header (Geneve or VXLAN) checksums on outgoing\n# packets. For Linux kernel before Mar 2021, UDP checksum must be present to trigger GRO on the receiver for better\n# performance of Geneve and VXLAN tunnels. The issue has been fixed by\n# https://github.com/torvalds/linux/commit/89e5c58fc1e2857ccdaae506fb8bc5fed57ee063, thus computing UDP checksum is\n# no longer necessary.\n# It should only be set to true when you are using an unpatched Linux kernel and observing poor transfer performance.\ntunnelCsum: false\n\n# Determines how tunnel traffic is encrypted. Currently encryption only works with encap mode.\n# It has the following options:\n# - none (default):  Inter-node Pod traffic will not be encrypted.\n# - ipsec:           Enable IPsec (ESP) encryption for Pod traffic across Nodes. Antrea uses\n#                    Preshared Key (PSK) for IKE authentication. When IPsec tunnel is enabled,\n#                    the PSK value must be passed to Antrea Agent through an environment\n#                    variable: ANTREA_IPSEC_PSK.\n# - wireGuard:       Enable WireGuard for tunnel traffic encryption.\ntrafficEncryptionMode: \"none\"\n\n# Enable bridging mode of Pod network on Nodes, in which the Node's transport interface is connected\n# to the OVS bridge, and cross-Node/VLAN traffic of AntreaIPAM Pods (Pods whose IP addresses are\n# allocated by AntreaIPAM from IPPools) is sent to the underlay network, and forwarded/routed by the\n# underlay network.\n# This option requires the `AntreaIPAM` feature gate to be enabled. At this moment, it supports only\n# IPv4 and Linux Nodes, and can be enabled only when `ovsDatapathType` is `system`,\n# `trafficEncapMode` is `noEncap`, and `noSNAT` is true.\nenableBridgingMode: false\n\n# Disable TX checksum offloading for container network interfaces. It's supposed to be set to true when the\n# datapath doesn't support TX checksum offloading, which causes packets to be dropped due to bad checksum.\n# It affects Pods running on Linux Nodes only.\ndisableTXChecksumOffload: false\n\n# Default MTU to use for the host gateway interface and the network interface of each Pod.\n# If omitted, antrea-agent will discover the MTU of the Node's primary interface and\n# also adjust MTU to accommodate for tunnel encapsulation overhead (if applicable).\ndefaultMTU: 0\n\n# wireGuard specifies WireGuard related configurations.\nwireGuard:\n  # The port for WireGuard to receive traffic.\n  port: 51820\n\negress:\n  # exceptCIDRs is the CIDR ranges to which outbound Pod traffic will not be SNAT'd by Egresses.\n  exceptCIDRs:\n\n# ClusterIP CIDR range for Services. It's required when AntreaProxy is not enabled, and should be\n# set to the same value as the one specified by --service-cluster-ip-range for kube-apiserver. When\n# AntreaProxy is enabled, this parameter is not needed and will be ignored if provided.\nserviceCIDR: \"\"\n\n# ClusterIP CIDR range for IPv6 Services. It's required when using kube-proxy to provide IPv6 Service in a Dual-Stack\n# cluster or an IPv6 only cluster. The value should be the same as the configuration for kube-apiserver specified by\n# --service-cluster-ip-range. When AntreaProxy is enabled, this parameter is not needed.\n# No default value for this field.\nserviceCIDRv6: \"\"\n\n# The port for the antrea-agent APIServer to serve on.\n# Note that if it's set to another value, the `containerPort` of the `api` port of the\n# `antrea-agent` container must be set to the same value.\napiPort: 10350\n\n# Enable metrics exposure via Prometheus. Initializes Prometheus metrics listener.\nenablePrometheusMetrics: true\n\n# Provide the IPFIX collector address as a string with format \u003cHOST\u003e:[\u003cPORT\u003e][:\u003cPROTO\u003e].\n# HOST can either be the DNS name or the IP of the Flow Collector. For example,\n# \"flow-aggregator.flow-aggregator.svc\" can be provided as DNS name to connect\n# to the Antrea Flow Aggregator service. If IP, it can be either IPv4 or IPv6.\n# However, IPv6 address should be wrapped with [].\n# If PORT is empty, we default to 4739, the standard IPFIX port.\n# If no PROTO is given, we consider \"tls\" as default. We support \"tls\", \"tcp\" and\n# \"udp\" protocols. \"tls\" is used for securing communication between flow exporter and\n# flow aggregator.\nflowCollectorAddr: \"flow-aggregator.flow-aggregator.svc:4739:tls\"\n\n# Provide flow poll interval as a duration string. This determines how often the\n# flow exporter dumps connections from the conntrack module. Flow poll interval\n# should be greater than or equal to 1s (one second).\n# Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\", \"m\", \"h\".\nflowPollInterval: \"5s\"\n\n# Provide the active flow export timeout, which is the timeout after which a flow\n# record is sent to the collector for active flows. Thus, for flows with a continuous\n# stream of packets, a flow record will be exported to the collector once the elapsed\n# time since the last export event is equal to the value of this timeout.\n# Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\", \"m\", \"h\".\nactiveFlowExportTimeout: \"5s\"\n\n# Provide the idle flow export timeout, which is the timeout after which a flow\n# record is sent to the collector for idle flows. A flow is considered idle if no\n# packet matching this flow has been observed since the last export event.\n# Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\", \"m\", \"h\".\nidleFlowExportTimeout: \"15s\"\n\nnodePortLocal:\n# Enable NodePortLocal, a feature used to make Pods reachable using port forwarding on the host. To\n# enable this feature, you need to set \"enable\" to true, and ensure that the NodePortLocal feature\n# gate is also enabled (which is the default).\n  enable: false\n# Provide the port range used by NodePortLocal. When the NodePortLocal feature is enabled, a port\n# from that range will be assigned whenever a Pod's container defines a specific port to be exposed\n# (each container can define a list of ports as pod.spec.containers[].ports), and all Node traffic\n# directed to that port will be forwarded to the Pod.\n  portRange: \"61000-62000\"\n\n# Provide the address of Kubernetes apiserver, to override any value provided in kubeconfig or InClusterConfig.\n# Defaults to \"\". It must be a host string, a host:port pair, or a URL to the base of the apiserver.\nkubeAPIServerOverride: \"\"\n\n# Provide the address of DNS server, to override the kube-dns service. It's used to resolve hostname in FQDN policy.\n# Defaults to \"\". It must be a host string or a host:port pair of the DNS server (e.g. 10.96.0.10, 10.96.0.10:53,\n# [fd00:10:96::a]:53).\ndnsServerOverride: \"\"\n\n# Comma-separated list of Cipher Suites. If omitted, the default Go Cipher Suites will be used.\n# https://golang.org/pkg/crypto/tls/#pkg-constants\n# Note that TLS1.3 Cipher Suites cannot be added to the list. But the apiserver will always\n# prefer TLS1.3 Cipher Suites whenever possible.\ntlsCipherSuites: \"\"\n\n# TLS min version from: VersionTLS10, VersionTLS11, VersionTLS12, VersionTLS13.\ntlsMinVersion: \"\"\n\n# The name of the interface on Node which is used for tunneling or routing the traffic across Nodes.\n# If there are multiple IP addresses configured on the interface, the first one is used. The IP\n# address used for tunneling or routing traffic to remote Nodes is decided in the following order of\n# preference (from highest to lowest):\n# 1. transportInterface\n# 2. transportInterfaceCIDRs\n# 3. The Node IP\ntransportInterface: \"\"\n\nmulticast:\n# The names of the interfaces on Nodes that are used to forward multicast traffic.\n# Defaults to transport interface if not set.\n  multicastInterfaces:\n\n# The interval at which the antrea-agent sends IGMP queries to Pods.\n# Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\", \"m\", \"h\".\n  igmpQueryInterval: \"125s\"\n\n# The network CIDRs of the interface on Node which is used for tunneling or routing the traffic across\n# Nodes. If there are multiple interfaces configured the same network CIDR, the first one is used. The\n# IP address used for tunneling or routing traffic to remote Nodes is decided in the following order of\n# preference (from highest to lowest):\n# 1. transportInterface\n# 2. transportInterfaceCIDRs\n# 3. The Node IP\ntransportInterfaceCIDRs:\n\n# Option antreaProxy contains AntreaProxy related configuration options.\nantreaProxy:\n  # ProxyAll tells antrea-agent to proxy all Service traffic, including NodePort, LoadBalancer, and ClusterIP traffic,\n  # regardless of where they come from. Therefore, running kube-proxy is no longer required. This requires the AntreaProxy\n  # feature to be enabled.\n  # Note that this option is experimental. If kube-proxy is removed, option kubeAPIServerOverride must be used to access\n  # apiserver directly.\n  proxyAll: false\n  # A string array of values which specifies the host IPv4/IPv6 addresses for NodePort. Values can be valid IP blocks.\n  # (e.g. 1.2.3.0/24, 1.2.3.4/32). An empty string slice is meant to select all host IPv4/IPv6 addresses.\n  # Note that the option is only valid when proxyAll is true.\n  nodePortAddresses:\n  # An array of string values to specify a list of Services which should be ignored by AntreaProxy (traffic to these\n  # Services will not be load-balanced). Values can be a valid ClusterIP (e.g. 10.11.1.2) or a Service name\n  # with Namespace (e.g. kube-system/kube-dns)\n  skipServices:\n  # When ProxyLoadBalancerIPs is set to false, AntreaProxy no longer load-balances traffic destined to the\n  # External IPs of LoadBalancer Services. This is useful when the external LoadBalancer provides additional\n  # capabilities (e.g. TLS termination) and it is desirable for Pod-to-ExternalIP traffic to be sent to the\n  # external LoadBalancer instead of being load-balanced to an Endpoint directly by AntreaProxy.\n  # Note that setting ProxyLoadBalancerIPs to false usually only makes sense when ProxyAll is set to true and\n  # kube-proxy is removed from the cluser, otherwise kube-proxy will still load-balance this traffic.\n  proxyLoadBalancerIPs: true\n\n# IPsec tunnel related configurations.\nipsec:\n  # The authentication mode of IPsec tunnel. It has the following options:\n  # - psk (default): Use pre-shared key (PSK) for IKE authentication.\n  # - cert:          Use CA-signed certificates for IKE authentication. This option requires the `IPsecCertAuth`\n  #                  feature gate to be enabled.\n  authenticationMode: \"psk\"\n\nmulticluster:\n# Enable Antrea Multi-cluster Gateway to support cross-cluster traffic.\n# This feature is supported only with encap mode.\n  enable: false\n# The Namespace where Antrea Multi-cluster Controller is running.\n# The default is antrea-agent's Namespace.\n  namespace: \"\"\n# Enable StretchedNetworkPolicy which could be enforced on cross-cluster traffic.\n# This feature is supported only with encap mode.\n  enableStretchedNetworkPolicy: false\n",
                  "antreaCNIConfig": "{\n    \"cniVersion\":\"0.3.0\",\n    \"name\": \"antrea\",\n    \"plugins\": [\n        {\n            \"type\": \"antrea\",\n            \"ipam\": {\n                \"type\": \"host-local\"\n            }\n        }\n        ,\n        {\n            \"type\": \"portmap\",\n            \"capabilities\": {\"portMappings\": true}\n        }\n        ,\n        {\n            \"type\": \"bandwidth\",\n            \"capabilities\": {\"bandwidth\": true}\n        }\n    ]\n}\n",
                  "antreaControllerConfig": "# FeatureGates is a map of feature names to bools that enable or disable experimental features.\nfeatureGates:\n# AllAlpha is a global toggle for alpha features. Per-feature key values override the default set by AllAlpha.\n#  AllAlpha: false\n\n# AllBeta is a global toggle for beta features. Per-feature key values override the default set by AllBeta.\n#  AllBeta: false\n\n# Enable traceflow which provides packet tracing feature to diagnose network issue.\n#  Traceflow: true\n\n# Enable Antrea ClusterNetworkPolicy feature to complement K8s NetworkPolicy for cluster admins\n# to define security policies which apply to the entire cluster, and Antrea NetworkPolicy\n# feature that supports priorities, rule actions and externalEntities in the future.\n#  AntreaPolicy: true\n\n# Enable collecting and exposing NetworkPolicy statistics.\n#  NetworkPolicyStats: true\n\n# Enable multicast traffic.\n#  Multicast: false\n\n# Enable controlling SNAT IPs of Pod egress traffic.\n#  Egress: true\n\n# Run Kubernetes NodeIPAMController with Antrea.\n#  NodeIPAM: false\n\n# Enable AntreaIPAM, which can allocate IP addresses from IPPools. AntreaIPAM is required by the\n# bridging mode and allocates IPs to Pods in bridging mode. It is also required to use Antrea for\n# IPAM when configuring secondary network interfaces with Multus.\n#  AntreaIPAM: false\n\n# Enable managing external IPs of Services of LoadBalancer type.\n#  ServiceExternalIP: false\n\n# Enable certificated-based authentication for IPsec.\n#  IPsecCertAuth: false\n\n# Enable managing ExternalNode for unmanaged VM/BM.\n#  ExternalNode: false\n\n# Enable collecting support bundle files with SupportBundleCollection CRD.\n#  SupportBundleCollection: false\n\n# Enable multi-cluster features.\n#  Multicluster: false\n\n# Enable users to protect their applications by specifying how they are allowed to communicate with others, taking\n# into account application context.\n#  L7NetworkPolicy: false\n\n# The port for the antrea-controller APIServer to serve on.\n# Note that if it's set to another value, the `containerPort` of the `api` port of the\n# `antrea-controller` container must be set to the same value.\napiPort: 10349\n\n# Enable metrics exposure via Prometheus. Initializes Prometheus metrics listener.\nenablePrometheusMetrics: true\n\n# Indicates whether to use auto-generated self-signed TLS certificate.\n# If false, a Secret named \"antrea-controller-tls\" must be provided with the following keys:\n#   ca.crt: \u003cCA certificate\u003e\n#   tls.crt: \u003cTLS certificate\u003e\n#   tls.key: \u003cTLS private key\u003e\nselfSignedCert: true\n\n# Comma-separated list of Cipher Suites. If omitted, the default Go Cipher Suites will be used.\n# https://golang.org/pkg/crypto/tls/#pkg-constants\n# Note that TLS1.3 Cipher Suites cannot be added to the list. But the apiserver will always\n# prefer TLS1.3 Cipher Suites whenever possible.\ntlsCipherSuites: \"\"\n\n# TLS min version from: VersionTLS10, VersionTLS11, VersionTLS12, VersionTLS13.\ntlsMinVersion: \"\"\n\nnodeIPAM:\n  # Enable the integrated Node IPAM controller within the Antrea controller.\n  enableNodeIPAM: false\n  # CIDR ranges for Pods in cluster. String array containing single CIDR range, or multiple ranges.\n  # The CIDRs could be either IPv4 or IPv6. At most one CIDR may be specified for each IP family.\n  # Value ignored when enableNodeIPAM is false.\n  clusterCIDRs:\n  # CIDR ranges for Services in cluster. It is not necessary to specify it when there is no overlap with clusterCIDRs.\n  # Value ignored when enableNodeIPAM is false.\n  serviceCIDR: \"\"\n  serviceCIDRv6: \"\"\n  # Mask size for IPv4 Node CIDR in IPv4 or dual-stack cluster. Value ignored when enableNodeIPAM is false\n  # or when IPv4 Pod CIDR is not configured. Valid range is 16 to 30.\n  nodeCIDRMaskSizeIPv4: 24\n  # Mask size for IPv6 Node CIDR in IPv6 or dual-stack cluster. Value ignored when enableNodeIPAM is false\n  # or when IPv6 Pod CIDR is not configured. Valid range is 64 to 126.\n  nodeCIDRMaskSizeIPv6: 64\n\nipsecCSRSigner:\n  # Determines the auto-approve policy of Antrea CSR signer for IPsec certificates management.\n  # If enabled, Antrea will auto-approve the CertificateSingingRequest (CSR) if its subject and x509 extensions\n  # are permitted, and the requestor can be validated. If K8s `BoundServiceAccountTokenVolume` feature is enabled,\n  # the Pod identity will also be validated to provide maximum security.\n  # If set to false, Antrea will not auto-approve CertificateSingingRequests and they need to be approved\n  # manually by `kubectl certificate approve`.\n  autoApprove: true\n  # Indicates whether to use auto-generated self-signed CA certificate.\n  # If false, a Secret named \"antrea-ipsec-ca\" must be provided with the following keys:\n  #   tls.crt: \u003cCA certificate\u003e\n  #   tls.key: \u003cCA private key\u003e\n  selfSignedCA: true\n\nmulticluster:\n  # Enable StretchedNetworkPolicy which allow Antrea-native policies to select peers\n  # from other clusters in a ClusterSet.\n  enableStretchedNetworkPolicy: false\n",
                  "antreaImage": "antrea/antrea-ubi:v2.1.0",
                  "antreaPlatform": "openshift"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: projects.packages.broadcom.com/antreainterworking/antrea-operator:v2.1.0_vmware.3@sha256:70f129499c801f164adf8772e7ccd5d1fc659df5f298d7851e6bf02ee760f2a9
          createdAt: "2025-04-16T12:00:00Z"
          description: An operator which installs Antrea network CNI plugin on the
            Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          support: projectantrea-maintainers@googlegroups.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AntreaInstall is the Schema for the antreainstalls API
            displayName: Antrea Install
            kind: AntreaInstall
            name: antreainstalls.operator.antrea.vmware.com
            version: v1
        description: An operator which installs Antrea network CNI plugin on the Kubernetes
          cluster.
        displayName: Antrea-by-Broadcom Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        links:
        - name: Antrea Operator For Kubernetes
          url: https://github.com/vmware/antrea-operator-for-kubernetes
        maintainers:
        - email: projectantrea-maintainers@googlegroups.com
          name: Project Antrea Maintainers
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: antrea.io
        relatedImages:
        - projects.packages.broadcom.com/antreainterworking/antrea-operator:v2.1.0_vmware.3@sha256:70f129499c801f164adf8772e7ccd5d1fc659df5f298d7851e6bf02ee760f2a9
        version: 1.10.0
      entries:
      - name: antrea-by-broadcom-operator.v1.10.0
        version: 1.10.0
      name: alpha
    defaultChannel: alpha
    packageName: antrea-by-broadcom-operator
    provider:
      name: antrea.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Nokia
      provider-url: https://docs.eda.dev
    name: eda-k8s-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eda-openshift-operator.v4.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.eda.nokia.com/v1",
                "kind": "ConnectControllerConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "operator"
                  },
                  "name": "connectcontrollerconfig-sample",
                  "namespace": "eda-connect-k8s"
                },
                "spec": {
                  "connectPluginName": "",
                  "controllerImagePullSecret": "",
                  "edaNamespace": "eda",
                  "heartBeat": "",
                  "logLevel": "",
                  "skipTLSVerify": false,
                  "tlsCertificateData": "",
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:c43a7fa7475db83011cc08b728982511fc0279a259e1aacc8ddde9e15f947725
          createdAt: "2025-06-10T15:47:27Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: eda-connect-k8s
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["connectnetworkattachmentstatuses.connect.eda.nokia.com",
            connectnetworkdefinitions.connect.eda.nokia.com]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConnectNetworkAttachmentStatus is the Schema for the connectnetworkattachmentstatuses
              API
            displayName: Connect NetworkAttachment Status
            kind: ConnectNetworkAttachmentStatus
            name: connectnetworkattachmentstatuses.connect.eda.nokia.com
            version: v1
          - description: ConnectNetworkDefinition is the Schema for the connectnetworkdefinitions
              API
            displayName: Connect Network Definition
            kind: ConnectNetworkDefinition
            name: connectnetworkdefinitions.connect.eda.nokia.com
            version: v1
          - description: ConnectControllerConfig is the Schema for the connectcontrollerconfigs
              API
            displayName: Connect Controller Config
            kind: ConnectControllerConfig
            name: connectcontrollerconfigs.config.eda.nokia.com
            version: v1
        description: "K8s operator to deploy eda connect controller to configure EDA
          Connect.\eb"
        displayName: eda-openshift-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - EDA Connect
        - K8s Controller
        - EDA K8s Controller Operator
        links:
        - name: EDA Connect K8s Controller Operator
          url: https://docs.eda.dev/connect/openshift-plugin/
        maintainers:
        - email: support@nokia.com
          name: nokia
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Nokia
          url: https://docs.eda.dev
        relatedImages:
        - ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:c43a7fa7475db83011cc08b728982511fc0279a259e1aacc8ddde9e15f947725
        version: 4.0.0
      entries:
      - name: eda-openshift-operator.v4.0.0
        version: 4.0.0
      name: alpha
    - currentCSV: eda-openshift-operator.v4.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.eda.nokia.com/v1",
                "kind": "ConnectControllerConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "operator"
                  },
                  "name": "connectcontrollerconfig-sample",
                  "namespace": "eda-connect-k8s"
                },
                "spec": {
                  "connectPluginName": "",
                  "controllerImagePullSecret": "",
                  "edaNamespace": "eda",
                  "heartBeat": "",
                  "logLevel": "",
                  "skipTLSVerify": false,
                  "tlsCertificateData": "",
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:ec264b52d581b8fd76e737b9811ab894957a4edf2f627f90bd9fe707890e6bf5
          createdAt: "2025-06-10T15:47:27Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: eda-connect-k8s
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["connectnetworkattachmentstatuses.connect.eda.nokia.com",
            connectnetworkdefinitions.connect.eda.nokia.com]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConnectNetworkAttachmentStatus is the Schema for the connectnetworkattachmentstatuses
              API
            displayName: Connect NetworkAttachment Status
            kind: ConnectNetworkAttachmentStatus
            name: connectnetworkattachmentstatuses.connect.eda.nokia.com
            version: v1
          - description: ConnectNetworkDefinition is the Schema for the connectnetworkdefinitions
              API
            displayName: Connect Network Definition
            kind: ConnectNetworkDefinition
            name: connectnetworkdefinitions.connect.eda.nokia.com
            version: v1
          - description: ConnectControllerConfig is the Schema for the connectcontrollerconfigs
              API
            displayName: Connect Controller Config
            kind: ConnectControllerConfig
            name: connectcontrollerconfigs.config.eda.nokia.com
            version: v1
        description: "K8s operator to deploy eda connect controller to configure EDA
          Connect.\eb"
        displayName: eda-openshift-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - EDA Connect
        - K8s Controller
        - EDA K8s Controller Operator
        links:
        - name: EDA Connect K8s Controller Operator
          url: https://docs.eda.dev/connect/openshift-plugin/
        maintainers:
        - email: support@nokia.com
          name: nokia
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Nokia
          url: https://docs.eda.dev
        relatedImages:
        - ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:ec264b52d581b8fd76e737b9811ab894957a4edf2f627f90bd9fe707890e6bf5
        version: 4.0.1
      entries:
      - name: eda-openshift-operator.v4.0.1
        version: 4.0.1
      name: stable
    defaultChannel: alpha
    packageName: eda-k8s-operator
    provider:
      name: Nokia
      url: https://docs.eda.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-prometheus-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-prometheus-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:b6071d6a1ec7cdd5f65b6d7296c1e9cd5660972a8ed614da4277e3a4ab5d1df1
          createdAt: "2024-03-01T23:43:00Z"
          description: Manage the full lifecycle of configuring and managing Prometheus
            and Alertmanager servers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/prometheus-operator/prometheus-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Prometheus Operator for Kubernetes provides easy monitoring definitions for Kubernetes services and deployment and management of Prometheus instances.

          Once installed, the Prometheus Operator provides the following features:

          * **Create/Destroy**: Easily launch a Prometheus instance for your Kubernetes namespace, a specific application or team easily using the Operator.
          * **Simple Configuration**: Configure the fundamentals of Prometheus like versions, persistence, retention policies, and replicas from a native Kubernetes resource.
          * **Target Services via Labels**: Automatically generate monitoring target configurations based on familiar Kubernetes label queries; no need to learn a Prometheus specific configuration language.

          ### Other Supported Features

          **High availability**

          Multiple instances are run across failure zones and data is replicated. This keeps your monitoring available during an outage, when you need it most.

          **Updates via automated operations**

          New Prometheus versions are deployed using a rolling update with no downtime, making it easy to stay up to date.

          **Handles the dynamic nature of containers**

          Alerting rules are attached to groups of containers instead of individual instances, which is ideal for the highly dynamic nature of container deployment.
        displayName: Prometheus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - prometheus
        - monitoring
        - tsdb
        - alerting
        - observability
        links:
        - name: Prometheus
          url: https://www.prometheus.io/
        - name: Documentation
          url: https://github.com/prometheus-operator/prometheus-operator/tree/master/Documentation
        - name: Prometheus Operator
          url: https://github.com/prometheus-operator/prometheus-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:b6071d6a1ec7cdd5f65b6d7296c1e9cd5660972a8ed614da4277e3a4ab5d1df1
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:413f7da5cb2e78abf33eb4e43f5fd483c19e791416eec7f5d4c3dbeb54db33a0
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader-rhel9@sha256:eb7614528884f2eefe1ae5e33a09624590c23e0e75510b4d88a274880cf090bc
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager-rhel9@sha256:7d1a3392872fd3990ae3acce79da659ba2238b30421824b7be533dcf8b93e89c
        version: 4.16.16-rhodf
      entries:
      - name: odf-prometheus-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odf-prometheus-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odf-prometheus-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odf-prometheus-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odf-prometheus-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odf-prometheus-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odf-prometheus-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odf-prometheus-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odf-prometheus-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odf-prometheus-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odf-prometheus-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odf-prometheus-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odf-prometheus-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odf-prometheus-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odf-prometheus-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odf-prometheus-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odf-prometheus-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odf-prometheus-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Alvearie
      provider-url: https://alvearie.github.io
    name: alvearie-imaging-ingestion
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: imaging-ingestion-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomEventBridge",
                "metadata": {
                  "name": "events"
                },
                "spec": {
                  "dicomEventDrivenIngestionName": "core",
                  "eventBridge": {},
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "events",
                  "natsTokenSecret": "nats-events-secure-bound-token",
                  "natsUrl": "jetstream.imaging-ingestion.svc.cluster.local:4222",
                  "role": "hub"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomEventDrivenIngestion",
                "metadata": {
                  "name": "core"
                },
                "spec": {
                  "databaseConfigName": "db-config",
                  "databaseSecretName": "db-secret",
                  "eventProcessor": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  },
                  "imagePullPolicy": "Always"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomInstanceBinding",
                "metadata": {
                  "name": "stow"
                },
                "spec": {
                  "bindingConfigName": "instance-binding-config",
                  "bindingSecretName": "instance-binding-secret",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "instanceBinding": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomStudyBinding",
                "metadata": {
                  "name": "fhir"
                },
                "spec": {
                  "bindingConfigName": "study-binding-config",
                  "bindingSecretName": "study-binding-secret",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "studyBinding": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomwebIngestionService",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "bucketConfigName": "imaging-ingestion",
                  "bucketSecretName": "imaging-ingestion",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "providerName": "provider",
                  "stowService": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  },
                  "wadoService": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DimseIngestionService",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "applicationEntityTitle": "DICOM-INGEST",
                  "bucketConfigName": "imaging-ingestion",
                  "bucketSecretName": "imaging-ingestion",
                  "dicomEventDrivenIngestionName": "core",
                  "dimseService": {},
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "DIMSE",
                  "natsTokenSecret": "ingestion-nats-secure-bound-token",
                  "natsUrl": "nats-secure.imaging-ingestion.svc.cluster.local:4222",
                  "providerName": "provider"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DimseProxy",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "applicationEntityTitle": "DCM4CHEE",
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "DIMSE",
                  "natsTokenSecret": "ingestion-nats-secure-bound-token",
                  "natsUrl": "nats-secure.imaging-ingestion.svc.cluster.local:4222",
                  "proxy": {},
                  "targetDimseHost": "arc.dcm4chee.svc.cluster.local",
                  "targetDimsePort": 11112
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: alvearie/imaging-ingestion-operator:0.0.3
          createdAt: "2021-04-08T15:29:55Z"
          description: The Alvearie Imaging Ingestion provides a collection of components
            for extending enterprise medical imaging informatics systems to Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Alvearie/imaging-ingestion
          support: alvearie.slack.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DicomEventBridge is the Schema for the dicomeventbridges
              API
            displayName: Dicom Event Bridge
            kind: DicomEventBridge
            name: dicomeventbridges.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Event driven manifest of all DICOM data across all associated
              S3 buckets
            displayName: Dicom Event Driven Ingestion
            kind: DicomEventDrivenIngestion
            name: dicomeventdriveningestions.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Fan-out notification of DICOM instance data (STOW-RS)
            displayName: Dicom Instance Binding
            kind: DicomInstanceBinding
            name: dicominstancebindings.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Fan-out notification of DICOM studies (FHIRv4 ImagingStudy)
            displayName: Dicom Study Binding
            kind: DicomStudyBinding
            name: dicomstudybindings.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides DICOMweb WADO-RS and STOW-RS access to a S3 bucket
            displayName: Dicomweb Ingestion Service
            kind: DicomwebIngestionService
            name: dicomwebingestionservices.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides a proxied DIMSE Application Entity (AE) in the cluster
              for C-STORE operations to a storage space
            displayName: Dimse Ingestion Service
            kind: DimseIngestionService
            name: dimseingestionservices.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides a bidirectional proxied DIMSE Application Entity
              (AE) in the cluster
            displayName: Dimse Proxy
            kind: DimseProxy
            name: dimseproxies.imaging-ingestion.alvearie.org
            version: v1alpha1
        description: |
          The Alvearie Imaging Ingestion provides a collection of components for extending enterprise medical imaging informatics systems to Kubernetes.

          ## About this Operator
          This operator enables standards-based medical imaging application interfaces such as DICOM Message Service Element (DIMSE) and DICOMweb.
          The API's provided allow for extending enterprise imaging systems to the cloud with custom, fit for purpose deployments.
          This is possible at speed, low cost, and little or zero integration effort.or extending enterprise medical imaging informatics systems to a modern cloud.


          Some potential usages:

          - Perform lightweight secure bridging of DICOM from the enterprise imaging zone to a cloud availability zone.
          - Utilize S3 buckets for segregating DICOM data to be used for clinical or scientific purposes.
          - Fan-out DICOM data to a number of different DICOM subscribers
          - Generate FHIRv4 *ImagingStudy* resources from ingested content and publish to a FHIRv4 based patient logitudinal record.
          - Distribute imaging study arrival and update notifications to imaging study subscribers


          ## Prerequisites

          - [Knative](https://operatorhub.io/operator/knative-operator)
          - ProgresSQL such as [Crunchy](https://operatorhub.io/operator/postgresql)
          - S3 storage such as [MinIO](https://operatorhub.io/operator/minio-operator)
          - [NATS operator](https://github.com/nats-io/nats-operator/) (Optional:  only required for DIMSE Proxy component)



          ## Documentation

          Visit the [official github repository](https://github.com/Alvearie/imaging-ingestion) for documentation and examples.
        displayName: Alvearie Imaging Ingestion Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - dicom
        - medical imaging
        - fhir
        - dimse-c
        - dimse
        - imaging study
        links:
        - name: Alvearie Imaging Ingestion Operator
          url: https://github.com/Alvearie/imaging-ingestion
        maintainers:
        - email: rduggan@ca.ibm.com
          name: Richard Duggan
        - email: jjacob@us.ibm.com
          name: Jeesmon Jacob
        maturity: alpha
        provider:
          name: Alvearie
          url: https://alvearie.github.io
        relatedImages:
        - alvearie/imaging-ingestion-operator:0.0.3
        version: 0.0.3
      entries:
      - name: imaging-ingestion-operator.v0.0.3
        version: 0.0.3
      - name: imaging-ingestion-operator.v0.0.2
        version: 0.0.2
      - name: imaging-ingestion-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: alvearie-imaging-ingestion
    provider:
      name: Alvearie
      url: https://alvearie.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solace Corporation
      provider-url: www.solace.com
    name: pubsubplus-eventbroker-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pubsubplus-eventbroker-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pubsubplus.solace.com/v1beta1",
                "kind": "PubSubPlusEventBroker",
                "metadata": {
                  "name": "non-ha-standalone-example"
                },
                "spec": {
                  "redundancy": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "true"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.10
          containerImage: registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator@sha256:d68db92bcb217d94dc0ad0a1cea625a0a9c9299d886d325a903163f9f887dcd5
          createdAt: "2025-07-31T19:48:36Z"
          description: The Solace PubSub+ Event Broker Operator deploys and manages
            the lifecycle of PubSub+ Event Brokers
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
          support: Solace Products
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PubSub+ Event Broker
            displayName: PubSub+ Event Broker
            kind: PubSubPlusEventBroker
            name: pubsubpluseventbrokers.pubsubplus.solace.com
            version: v1beta1
        description: |
          ## Solace PubSub+ Platform
          Solace [PubSub+ Platform](https://solace.com/products/platform/) is a complete event streaming and management platform
          for real-time enterprises. The [PubSub+ Software Event Broker](https://solace.com/products/event-broker/software/) efficiently streams
          event-driven information between applications, IoT devices, and user interfaces running in the cloud, on-premises,
          and in hybrid environments using open APIs and protocols like AMQP, JMS, MQTT, REST and WebSocket.
          It can be installed into a variety of public and private clouds, PaaS, and on-premises environments.
          Event brokers in multiple locations can be linked together in an [Event Mesh](https://solace.com/what-is-an-event-mesh/)
          to dynamically share events across the distributed enterprise.

          ## Solace PubSub+ Event Broker Operator
          Solace Pubsub+ Event Broker Operator automatically deploys and manages PubSub+ Software Event Brokers on Kubernetes and OpenShift environments.

          Features include:
          * Initial deployment configuration: Production-ready HA, non-HA or for developers
          * Rolling update for configuration changes or broker upgrades
          * Prometheus monitoring support

          ## Getting started
          Refer to the [Solace PubSub+ Event Broker Operator Quick Start](https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/README.md)
        displayName: Solace PubSub+ Event Broker Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - solace
        - pubsubplus
        - pubsub+
        - pubsub
        - messaging
        - advanced event broker
        - event broker
        - event mesh
        - event streaming
        - data streaming
        - event integration
        - middleware
        links:
        - name: Operator QuickStart
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
        - name: Operator Documentation
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/docs/EventBrokerOperatorUserGuide.md
        - name: Solace.com
          url: https://solace.com
        - name: Understanding Solace Technology
          url: https://solace.com/products/platform/
        - name: Solace PubSub+ Technical Documentation
          url: https://docs.solace.com/
        - name: Solace Developer Portal
          url: https://www.solace.dev/
        - name: Solace Community
          url: https://solace.community/
        maintainers:
        - email: support@solace.com
          name: Solace R&D
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Solace Corporation
          url: www.solace.com
        relatedImages:
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator@sha256:d68db92bcb217d94dc0ad0a1cea625a0a9c9299d886d325a903163f9f887dcd5
        version: 1.4.0
      entries:
      - name: pubsubplus-eventbroker-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: pubsubplus-eventbroker-operator
    provider:
      name: Solace Corporation
      url: www.solace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: StreamNative
      provider-url: https://streamnative.io
    name: function-mesh
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: function-mesh.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "java-function-sample",
                  "namespace": "default"
                },
                "spec": {
                  "autoAck": true,
                  "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                  "clusterName": "test-pulsar",
                  "forwardSourceMessageProperty": true,
                  "input": {
                    "topics": [
                      "persistent://public/default/java-function-input-topic"
                    ],
                    "typeClassName": "java.lang.String"
                  },
                  "java": {
                    "extraDependenciesDir": "random-dir/",
                    "jar": "pulsar-functions-api-examples.jar",
                    "jarLocation": "public/default/nlu-test-java-function"
                  },
                  "logTopic": "persistent://public/default/logging-function-logs",
                  "maxPendingAsyncRequests": 1000,
                  "maxReplicas": 5,
                  "output": {
                    "topic": "persistent://public/default/java-function-output-topic",
                    "typeClassName": "java.lang.String"
                  },
                  "pod": {
                    "annotations": {
                      "managed-function": "true"
                    },
                    "env": [
                      {
                        "name": "EXAMPLE_VARIABLE",
                        "value": "exampleValue"
                      }
                    ],
                    "imagePullSecrets": [
                      {
                        "name": "regcred"
                      }
                    ],
                    "initContainers": [
                      {
                        "command": [
                          "sh",
                          "-c",
                          "echo The app is running! \u0026\u0026 sleep 30"
                        ],
                        "image": "busybox:1.28",
                        "name": "init-function"
                      }
                    ],
                    "labels": {
                      "locaction": "mtv"
                    },
                    "sidecars": [
                      {
                        "command": [
                          "sh",
                          "-c",
                          "echo The app is running! \u0026\u0026 sleep 30000"
                        ],
                        "image": "busybox:1.28",
                        "name": "sidecar-function"
                      }
                    ],
                    "volumes": [
                      {
                        "emptyDir": {},
                        "name": "cache-volume"
                      }
                    ]
                  },
                  "pulsar": {
                    "authSecret": "test-auth",
                    "pulsarConfig": "test-pulsar",
                    "tlsSecret": "test-tls"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "secretsMap": {
                    "name": {
                      "key": "username",
                      "path": "test-secret"
                    },
                    "pwd": {
                      "key": "password",
                      "path": "test-secret"
                    }
                  },
                  "volumeMounts": [
                    {
                      "mountPath": "/cache",
                      "name": "cache-volume"
                    }
                  ]
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "FunctionMesh",
                "metadata": {
                  "name": "functionmesh-sample"
                },
                "spec": {
                  "functions": [
                    {
                      "autoAck": true,
                      "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                      "clusterName": "test-pulsar",
                      "forwardSourceMessageProperty": true,
                      "input": {
                        "topics": [
                          "persistent://public/default/functionmesh-input-topic"
                        ],
                        "typeClassName": "java.lang.String"
                      },
                      "java": {
                        "jar": "pulsar-functions-api-examples.jar",
                        "jarLocation": "public/default/nlu-test-functionmesh-ex1"
                      },
                      "logTopic": "persistent://public/default/logging-function-log",
                      "maxReplicas": 1,
                      "name": "ex1",
                      "output": {
                        "topic": "persistent://public/default/mid-topic",
                        "typeClassName": "java.lang.String"
                      },
                      "pulsar": {
                        "pulsarConfig": "mesh-test-pulsar"
                      },
                      "replicas": 1,
                      "resources": {
                        "limits": {
                          "cpu": "0.2",
                          "memory": "1.1G"
                        },
                        "requests": {
                          "cpu": "0.1",
                          "memory": "1G"
                        }
                      }
                    },
                    {
                      "autoAck": true,
                      "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                      "clusterName": "test-pulsar",
                      "forwardSourceMessageProperty": true,
                      "input": {
                        "topics": [
                          "persistent://public/default/mid-topic"
                        ],
                        "typeClassName": "java.lang.String"
                      },
                      "java": {
                        "jar": "pulsar-functions-api-examples.jar",
                        "jarLocation": "public/default/nlu-test-functionmesh-ex2"
                      },
                      "logTopic": "persistent://public/default/logging-function-logs",
                      "maxReplicas": 1,
                      "name": "ex2",
                      "output": {
                        "topic": "persistent://public/default/functionmesh-output-topic",
                        "typeClassName": "java.lang.String"
                      },
                      "pulsar": {
                        "pulsarConfig": "mesh-test-pulsar"
                      },
                      "replicas": 1,
                      "resources": {
                        "limits": {
                          "cpu": "0.2",
                          "memory": "1.1G"
                        },
                        "requests": {
                          "cpu": "0.1",
                          "memory": "1G"
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Sink",
                "metadata": {
                  "name": "sink-sample"
                },
                "spec": {
                  "autoAck": true,
                  "className": "org.apache.pulsar.io.elasticsearch.ElasticSearchSink",
                  "clusterName": "test-pulsar",
                  "image": "streamnative/pulsar-io-elastic-search:2.10.0.0-rc10",
                  "input": {
                    "topics": [
                      "persistent://public/default/input"
                    ],
                    "typeClassName": "[B"
                  },
                  "java": {
                    "jar": "connectors/pulsar-io-elastic-search-2.10.0.0-rc10.nar",
                    "jarLocation": ""
                  },
                  "maxReplicas": 1,
                  "pulsar": {
                    "pulsarConfig": "test-sink"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "sinkConfig": {
                    "elasticSearchUrl": "http://quickstart-es-http.default.svc.cluster.local:9200",
                    "indexName": "my_index",
                    "password": "wJ757TmoXEd941kXm07Z2GW3",
                    "typeName": "doc",
                    "username": "elastic"
                  }
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Source",
                "metadata": {
                  "name": "source-sample"
                },
                "spec": {
                  "className": "org.apache.pulsar.io.debezium.mongodb.DebeziumMongoDbSource",
                  "clusterName": "test-pulsar",
                  "forwardSourceMessageProperty": true,
                  "image": "streamnative/pulsar-io-debezium-mongodb:2.10.0.0-rc10",
                  "java": {
                    "jar": "connectors/pulsar-io-debezium-mongodb-2.10.0.0-rc10.nar",
                    "jarLocation": ""
                  },
                  "maxReplicas": 1,
                  "output": {
                    "producerConf": {
                      "maxPendingMessages": 1000,
                      "maxPendingMessagesAcrossPartitions": 50000,
                      "useThreadLocalProducers": true
                    },
                    "topic": "persistent://public/default/destination",
                    "typeClassName": "org.apache.pulsar.common.schema.KeyValue"
                  },
                  "pulsar": {
                    "pulsarConfig": "test-source"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "sourceConfig": {
                    "database.whitelist": "inventory",
                    "mongodb.hosts": "rs0/mongo-dbz-0.mongo.default.svc.cluster.local:27017,rs0/mongo-dbz-1.mongo.default.svc.cluster.local:27017,rs0/mongo-dbz-2.mongo.default.svc.cluster.local:27017",
                    "mongodb.name": "dbserver1",
                    "mongodb.password": "dbz",
                    "mongodb.task.id": "1",
                    "mongodb.user": "debezium",
                    "pulsar.service.url": "pulsar://test-pulsar-broker.default.svc.cluster.local:6650"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/function-mesh@sha256:3e13e5cc3ac4d7387ee0adc02cb5183aad6cf2301f76ab230482609c4da1c0b7
          createdAt: "2022-10-27T07:49:28Z"
          description: The Function Mesh Operator manages the Pulsar Functions and
            Connectors deployed on a Kubernetes cluster.
          olm.properties: '[{"type":"olm.maxOpenShiftVersion","value":"4.8"}]'
          olm.skipRange: <0.7.0
          operatorhub.io/ui-metadata-max-k8s-version: "1.21"
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/streamnative/function-mesh
          support: StreamNative
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FunctionMesh is the Schema for the functionmeshes API
            displayName: Mesh
            kind: FunctionMesh
            name: functionmeshes.compute.functionmesh.io
            version: v1alpha1
          - description: Function is the Schema for the functions API
            displayName: Function
            kind: Function
            name: functions.compute.functionmesh.io
            version: v1alpha1
          - description: Sink is the Schema for the sinks API
            displayName: Sink
            kind: Sink
            name: sinks.compute.functionmesh.io
            version: v1alpha1
          - description: Source is the Schema for the sources API
            displayName: Source
            kind: Source
            name: sources.compute.functionmesh.io
            version: v1alpha1
        description: |
          [Function Mesh](https://functionmesh.io/) is a serverless framework purpose-built for stream processing applications. It brings powerful event-streaming capabilities to your applications by orchestrating multiple [Pulsar Functions](http://pulsar.apache.org/docs/en/next/functions-overview/) and [Pulsar IO connectors](http://pulsar.apache.org/docs/en/next/io-overview/) for complex stream processing jobs on Kubernetes.
          Prerequisites:
          - Install cert-manager operator first.
        displayName: Function Mesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pulsar
        - function-mesh
        - serverless
        - pulsar-functions
        - pulsar-io-connectors
        - StreamNative
        links:
        - name: Function Mesh
          url: https://function-mesh.io
        maintainers:
        - email: function-mesh@streamnative.io
          name: Function Mesh
        maturity: alpha
        minKubeVersion: v1.17.0
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/function-mesh@sha256:3e13e5cc3ac4d7387ee0adc02cb5183aad6cf2301f76ab230482609c4da1c0b7
        - docker.cloudsmith.io/streamnative/mirrors/gcr.io/kubebuilder/kube-rbac-proxy@sha256:67ecb332573384515406ebd71816781366b70adb0eb66345e5980e92603373e1
        version: 0.7.0
      entries:
      - name: function-mesh.v0.7.0
        version: 0.7.0
      name: alpha
    defaultChannel: alpha
    packageName: function-mesh
    provider:
      name: StreamNative
      url: https://streamnative.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.3.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo-operator@sha256:32178a16dd2115e0a11a2e3dfbfc130ae1d0afd8e293131e76b6fc1a79b36c12
          createdAt: "2019-05-01T00:00:00Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/kubeturbo-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/kubeturbo-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo-operator@sha256:32178a16dd2115e0a11a2e3dfbfc130ae1d0afd8e293131e76b6fc1a79b36c12
        - registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo@sha256:671be694176d9cc31b8a22c11bbfefc2504ae72d27fd010b315b6839f9ab59d5
        version: 8.3.3
      entries:
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: stable
    defaultChannel: stable
    packageName: kubeturbo-certified-rhmp
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: devspaces
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devspacesoperator.v3.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "org.eclipse.che/v1",
                "kind": "CheCluster",
                "metadata": {
                  "name": "devspaces",
                  "namespace": "openshift-devspaces"
                },
                "spec": {
                  "auth": {
                    "identityProviderURL": "",
                    "oAuthClientName": "",
                    "oAuthSecret": ""
                  },
                  "database": {
                    "externalDb": false
                  },
                  "k8s": {
                    "ingressDomain": null,
                    "tlsSecretName": null
                  },
                  "metrics": {
                    "enable": true
                  },
                  "server": {
                    "workspaceNamespaceDefault": "<username>-devspaces"
                  },
                  "storage": {
                    "pvcStrategy": "common"
                  }
                }
              },
              {
                "apiVersion": "org.eclipse.che/v2",
                "kind": "CheCluster",
                "metadata": {
                  "name": "devspaces",
                  "namespace": "openshift-devspaces"
                },
                "spec": {
                  "components": {},
                  "containerRegistry": {},
                  "devEnvironments": {},
                  "gitServices": {},
                  "networking": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/devspaces/devspaces-rhel9-operator@sha256:39c95279bced68928d028ef026f82eafe53eddb3dbe57703c8007cb5a170a70a
          createdAt: "2025-09-11T22:30:49+00:00"
          description: Devfile v2 and v1 development solution, 1 instance per cluster,
            for portable, collaborative k8s workspaces.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/devspaces-images/
          skipRange: '>=3.0.0 <3.23.0'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              The `CheCluster` custom resource allows defining and managing Red Hat OpenShift Dev Spaces server installation.
              Based on these settings, the  Operator automatically creates and maintains several ConfigMaps:
              `che`, `plugin-registry` that will contain the appropriate environment variables
              of the various components of the installation. These generated ConfigMaps must NOT be updated manually.
            displayName: Red Hat OpenShift Dev Spaces instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v2
          - description: The `CheCluster` custom resource allows defining and managing
              a Che server installation
            displayName: Red Hat OpenShift Dev Spaces instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v1
        description: |
          A collaborative Kubernetes-native development solution that delivers OpenShift workspaces and in-browser IDE for rapid cloud application development.
          This operator installs the Plugin and Devfile registries, Dashboard, Gateway and the Red Hat OpenShift Dev Spaces server, and configures these services.
          OpenShift OAuth is used directly for authentication. TLS mode is on.

          ## How to Install
          Press the **Install** button, choose the channel and the upgrade strategy, and wait for the **Installed** Operator status.
          When the operator is installed, create a new Custom Resource (CR) of Kind CheCluster (click the **Create New** button).
          The CR spec contains all defaults. You can start using Red Hat OpenShift Dev Spaces when the CR status is set to **Available**, and you see a URL to Red Hat OpenShift Dev Spaces.

          ## Defaults
          By default, the operator deploys Red Hat OpenShift Dev Spaces with:
          * 10Gi storage
          * Auto-generated passwords
          * Bundled Plugin and Devfile registries

          Use `oc edit checluster/devspaces -n openshift-devspaces` to update Red Hat OpenShift Dev Spaces default installation options.
          See more in the [Installation guide](https://docs.redhat.com/en/documentation/red_hat_openshift_dev_spaces/3.23/html/administration_guide/configuring-devspaces).

          ### Certificates
          Operator uses a default router certificate to secure Red Hat OpenShift Dev Spaces routes.
          Follow the [guide](https://docs.redhat.com/en/documentation/red_hat_openshift_dev_spaces/3.23/html/administration_guide/configuring-devspaces#importing-untrusted-tls-certificates)
          to import certificates into Red Hat OpenShift Dev Spaces.
        displayName: Red Hat OpenShift Dev Spaces
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eclipse che
        - workspaces
        - devtools
        - developer
        - ide
        - java
        - devspaces
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/openshift-dev-spaces/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_openshift_dev_spaces
        - name: Operator GitHub Repo
          url: https://github.com/redhat-developer/devspaces-images/
        maintainers:
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/devspaces/pluginregistry-rhel9@sha256:3ee131041761f5250b29d497043ecf7820e02e03e73ed0f929c17e2247c0d7a3
        - registry.redhat.io/devspaces/traefik-rhel9@sha256:95cf6a28028ec7a546a3d8d78377f204bb559c3d1022a73f08c646e76b85ce85
        - registry.redhat.io/devspaces-tech-preview/jetbrains-ide-rhel9@sha256:46bf7bea716df74f14d2629fb7e0e2ab692564cec348ffb8a6f4fea1739b23c2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5a6edfecf1636805e6f0bcad02df10bceec764f0580f3ac8fe62aba978d50006
        - registry.redhat.io/devspaces/server-rhel9@sha256:c0ec8c792f88aacf6ee31a4b6d2d847d864707a084148b48e4e402388989f868
        - registry.redhat.io/devspaces/devspaces-rhel9-operator@sha256:39c95279bced68928d028ef026f82eafe53eddb3dbe57703c8007cb5a170a70a
        - registry.redhat.io/devspaces/code-rhel9@sha256:abc9b12131c8cdca0719e0404d354310f53769990b089f1a3adea31193798a8e
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6bcfa981a7a75c242a16afa6dab36bf6e3b8a1442fa863875919ad551f48eceb
        - registry.redhat.io/devspaces/configbump-rhel9@sha256:cecbc3796058618da9ad0473a6f937aae5aca25776bb434894286f22baa4677c
        - registry.redhat.io/devspaces/dashboard-rhel9@sha256:c34cf5ac9bf95c1c5082afd5a2941b5e412be8ca43aef6a72e8fe5a279a43271
        - registry.redhat.io/devspaces/udi-rhel9@sha256:f7325ee7259a0585e8017d4442d778a37f4c8a67d81fe70c241a014001e37539
        - registry.redhat.io/devspaces-tech-preview/idea-rhel9@sha256:df24ebb239a2cc0d4543309f51a7268c7d524d428a536e0b3443c51089c3f87e
        version: 3.23.0
      entries:
      - name: devspacesoperator.v3.23.0
        version: 3.23.0
      - name: devspacesoperator.v3.22.0
        version: 3.22.0
      - name: devspacesoperator.v3.21.0
        version: 3.21.0
      - name: devspacesoperator.v3.20.0
        version: 3.20.0
      - name: devspacesoperator.v3.19.0
        version: 3.19.0
      - name: devspacesoperator.v3.18.1-0.1739475915.p
        version: 3.18.1+0.1739475915.p
      - name: devspacesoperator.v3.18.1
        version: 3.18.1
      - name: devspacesoperator.v3.18.0
        version: 3.18.0
      - name: devspacesoperator.v3.17.0
        version: 3.17.0
      - name: devspacesoperator.v3.16.1
        version: 3.16.1
      - name: devspacesoperator.v3.16.0
        version: 3.16.0
      - name: devspacesoperator.v3.15.0
        version: 3.15.0
      - name: devspacesoperator.v3.14.0
        version: 3.14.0
      - name: devspacesoperator.v3.13.0
        version: 3.13.0
      - name: devspacesoperator.v3.12.0-0.1712003563.p
        version: 3.12.0+0.1712003563.p
      - name: devspacesoperator.v3.12.0
        version: 3.12.0
      - name: devspacesoperator.v3.11.0
        version: 3.11.0
      - name: devspacesoperator.v3.10.1
        version: 3.10.1
      - name: devspacesoperator.v3.10.0
        version: 3.10.0
      - name: devspacesoperator.v3.9.1
        version: 3.9.1
      - name: devspacesoperator.v3.9.0
        version: 3.9.0
      - name: devspacesoperator.v3.8.0
        version: 3.8.0
      - name: devspacesoperator.v3.7.2
        version: 3.7.2
      - name: devspacesoperator.v3.7.1-0.1691407091.p
        version: 3.7.1+0.1691407091.p
      - name: devspacesoperator.v3.7.1
        version: 3.7.1
      - name: devspacesoperator.v3.7.1-0.1691548593.p
        version: 3.7.1+0.1691548593.p
      - name: devspacesoperator.v3.7.0
        version: 3.7.0
      - name: devspacesoperator.v3.6.0
        version: 3.6.0
      - name: devspacesoperator.v3.6.0-0.1685320856.p
        version: 3.6.0+0.1685320856.p
      - name: devspacesoperator.v3.6.0-0.1685322634.p
        version: 3.6.0+0.1685322634.p
      - name: devspacesoperator.v3.5.0
        version: 3.5.0
      - name: devspacesoperator.v3.4.0-0.1675914643.p
        version: 3.4.0+0.1675914643.p
      - name: devspacesoperator.v3.4.0
        version: 3.4.0
      - name: devspacesoperator.v3.3.0-0.1673528130.p
        version: 3.3.0+0.1673528130.p
      - name: devspacesoperator.v3.3.0
        version: 3.3.0
      name: stable
    defaultChannel: stable
    packageName: devspaces
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Jooho Lee
      provider-url: https://github.com/jooho/nfs-provisioner-operator
    name: nfs-provisioner-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfs-provisioner-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.jhouse.com/v1alpha1",
                "kind": "NFSProvisioner",
                "metadata": {
                  "name": "nfsprovisioner-sample"
                },
                "spec": {
                  "nfsImageConfiguration": {
                    "image": "k8s.gcr.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439",
                    "imagePullPolicy": "IfNotPresent"
                  },
                  "scForNFS": "nfs",
                  "scForNFSPvc": "local-sc",
                  "storageSize": "1G"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: quay.io/jooholee/nfs-provisioner-operator@sha256:d9c013967421ec72644a588a155975bf856a856f1ef38ba71dfea306fdb47acd
          createdAt: "2025-08-08T20:23:50Z"
          description: Create and manage NFS Server and Provisioner
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/jooho/nfs-provisioner-operator
          support: NFS
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NFSProvisioner is the Schema for the nfsprovisioners API
            displayName: NFS Provisioner App
            kind: NFSProvisioner
            name: nfsprovisioners.cache.jhouse.com
            version: v1alpha1
        description: "This operator deploy NFS server with local storage and also
          provide provisioner for storageClass.\n### Core Capabilities\n* **NFS Server:**
          Deployed\n* **NFS Provisioner:** Help customers to create PV using StorageClass\n*
          **StorageClass:** Dynamically create PV for requested PVC\n### NFS Provisioner
          Operator Features\n* NFS Server can use localStorage PVC or HostPath on
          the node\n### Must-Gather \n> isv-cli must-gather --image=quay.io/jooholee/nfs-provisioner-operator-must-gather:0.0.1\n"
        displayName: NFS Provisioner Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nfs
        - storage
        links:
        - name: Nfs Provisioner Operator
          url: https://github.com/jooho/nfs-provisioner-operator
        maintainers:
        - email: ljhiyh@gmail.com
          name: Jooho
        maturity: alpha
        provider:
          name: Jooho Lee
          url: https://github.com/jooho/nfs-provisioner-operator
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - quay.io/jooholee/nfs-provisioner-operator@sha256:d9c013967421ec72644a588a155975bf856a856f1ef38ba71dfea306fdb47acd
        version: 0.0.8
      entries:
      - name: nfs-provisioner-operator.v0.0.8
        version: 0.0.8
      - name: nfs-provisioner-operator.v0.0.7
        version: 0.0.7
      - name: nfs-provisioner-operator.v0.0.6
        version: 0.0.6
      - name: nfs-provisioner-operator.v0.0.5
        version: 0.0.5
      - name: nfs-provisioner-operator.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: nfs-provisioner-operator
    provider:
      name: Jooho Lee
      url: https://github.com/jooho/nfs-provisioner-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-licensing
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-licensing.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Licensing",
                "metadata": {
                  "name": "licensing-sample"
                },
                "spec": {
                  "environment": {
                    "deploymentType": "openshift"
                  },
                  "hostid": {
                    "enabled": true,
                    "environment": {},
                    "hostMounts": {},
                    "image": "prod.harbor.keysight.digital/loadcore/kcos-licensing-hostid-vmuuid@sha256:076a2c6aeec875470414af7905613a25b56bcb2aca5edcb1c4c53037cd01b332",
                    "imagePullPolicy": "IfNotPresent"
                  },
                  "image": {
                    "dbimage": "prod.harbor.keysight.digital/loadcore/postgresql-ubi-image@sha256:17f6a8f775d06650939d2c73e82b45d73d50190cc735ca70695d082f7790834b",
                    "initOwnership": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf",
                    "repository": "prod.harbor.keysight.digital/loadcore"
                  },
                  "ingress": {
                    "annotations": {
                      "nginx.ingress.kubernetes.io/auth-method": "POST",
                      "nginx.ingress.kubernetes.io/auth-response-headers": "X-AUTH-USER",
                      "nginx.ingress.kubernetes.io/auth-url": "https://127.0.0.1/token/validate"
                    },
                    "enabled": true
                  },
                  "nodeSelector": {
                    "node-role.kubernetes.io/kcos-licensing": "kcos-licensing"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:29:04Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Licensing
            name: licensings.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Licensing Operator
        displayName: Keysight KCOS Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Licensing
          url: https://keysight-kcos-licensing.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-licensing@sha256:a09291923ad8d7533b33e40e1a3c4e735c2a4ae3b2b87be7c9d1a1a9244b3dad
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - prod.harbor.keysight.digital/loadcore/postgresql-ubi-image@sha256:17f6a8f775d06650939d2c73e82b45d73d50190cc735ca70695d082f7790834b
        - prod.harbor.keysight.digital/loadcore/kcos-licensing-hostid-vmuuid@sha256:076a2c6aeec875470414af7905613a25b56bcb2aca5edcb1c4c53037cd01b332
        version: 0.1.21
      entries:
      - name: keysight-kcos-licensing.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-licensing
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: multicluster-engine
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-engine.v2.5.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-08-06T12:21:10Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0 <2.5.10'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '[]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine is the Schema for the multiclusterengines
              API
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:80f33a762535ff69f57ecb04de043b8fe249a63767712f4942ac6685450e8f28
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:1af88c3703b668020137f90657e6262b73e407aa858bbc6d9c12ba3a1cab8043
        - registry.redhat.io/multicluster-engine/agent-service-rhel8@sha256:92464c57700dcb13ccf5431c104a3fba6845f651776aa5beb1ce7171087dc1ea
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:cb2c57c51edc8a2320a10ac217cca6a401db09de55c0c9d0c9a7b4545e2636e9
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:7f32ed997e03dba1746b3627d5a33fc69282b2e831edd21ec74020fe4fa1ca5d
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:90e792c92ef8c75ffb97545394206c9197997d474493d93f09fca273f65133e4
        - registry.redhat.io/multicluster-engine/hive-rhel8@sha256:27eefe6ad7d93e752f4d76b776884129615c2e87ff24116c401a886e98abfcc5
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:1e9eff48ea837a1d19c595986d765d74810d6d1397602ad79e2e820b1cbb5c6e
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:297d8d0d5855db90ad40af702d6ad5f3f32a4aab486686919d86bbd21c367ca9
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:b0a70eee404ba1267c0ae1a8eb140251d054e722305d680685223cfda360aa37
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:53b8d48e540c3d9bb5d948af0375c26377271e595fffdb3d5f921e3a358fa298
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:36abf06619cd3971da8f42f4193621a1384970365fce0f48e86a65e2840bca38
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel8@sha256:360d3dfb823d8e199574cf1c3fbf890fa74c71fc48644e22b47339363e941b3e
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:c8d48e7ccdce43ce72154e464e74c32477efb0094aab327ed0eaab6c1244774a
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:515eda1e4e1ced222709103789af14f7b4e99d4da037deddd8499aad3b45451b
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:3e3f7ee898e0f6b0f74026399753d7fde41013ade76f7963d5c25e578eb1906e
        - registry.redhat.io/rhel8/postgresql-12@sha256:da619f6617d99790b4fe00fdb23ab4aea0cdb0592590162d51b3ef94aca01a7b
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel8@sha256:6bf850d8fb250740113bfb004b19edbd5cf560c8175b81e2f3766280b3482e27
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:43cc171404571bb3c87275a86e9ab4409861c1c01b5f1441c455565d56e1921d
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:d71c1441577b1cd360d074581ad19c2b0cfa99cdfe1869af1777ff7de6821735
        - registry.redhat.io/multicluster-engine/cluster-api-provider-aws-rhel9@sha256:7f65119dc5d81b8ba485b651efe1d9013b62ec57e4725dbf8bdc2e3b7748f677
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:1a5ff852d96bd15c631a6f89e2b0047cd8dde6264646f51106f2b523decc64ba
        - registry.redhat.io/multicluster-engine/aws-encryption-provider-rhel9@sha256:22cefc43e205e4bdd7874c7e1fccba60fe556df4b2bab055d01c834d62f194bd
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:53ad9402a0a13c894b485f418dd1353c8870792893a5ce33f9d6c3cc9839ebc4
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:2497e077a00dfa94b913359be0c696126e932db1a45e066ead65dee76a1db8fd
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel8@sha256:1e755b64358c10309804a3bf44ca640a7d5bb6e0474d33774a8d76095f2b8c3d
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:25b30d4f4ccc23fda707db88969c0ae036ecb4a2baf15b82221e3351c23292f9
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:a4f58b94d45c1ae1c4ba69b12030b9945b37f3ce6e96c27919f256ae887ec28b
        - registry.redhat.io/multicluster-engine/cluster-api-rhel9@sha256:0e329ca39b1f18520da093a5194ef6e1d5527270c74bf260517620577a21b6c7
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:19dba8be2295f54a93853dbe566132011f0120b2c4b48de2816509f5d201e420
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:bd6d0189ea2bbc87b5745f9d7f9c82118f8dcc9ee63fe0a36bb5e7dfc0e56021
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:424ca238d9c8b20dc434a07aa6a27fe6e7c63b3e2480eead284570864d682175
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:3d420ba8bc5e1dbe629c0d4aa65c0a79a7914ac349828046994e245f4f2adb2b
        - registry.redhat.io/multicluster-engine/assisted-installer-reporter-rhel8@sha256:5d41ca13f4ae102760b65375578c2f632aa6ce41092732300d6727f195da3d55
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:460d37a0032813aa85a3b18bd6f152e4ed6b89f802d17d787c3365082420d379
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:d7ccc8ddbeb11ac8a3a0d4946920a0ab87d9e9595d7c61740bd57f3f4a19a711
        version: 2.5.10
      entries:
      - name: multicluster-engine.v2.5.10
        version: 2.5.10
      - name: multicluster-engine.v2.5.9
        version: 2.5.9
      - name: multicluster-engine.v2.5.8
        version: 2.5.8
      - name: multicluster-engine.v2.5.7
        version: 2.5.7
      - name: multicluster-engine.v2.5.6
        version: 2.5.6
      - name: multicluster-engine.v2.5.5
        version: 2.5.5
      - name: multicluster-engine.v2.5.4
        version: 2.5.4
      - name: multicluster-engine.v2.5.3
        version: 2.5.3
      - name: multicluster-engine.v2.5.2
        version: 2.5.2
      - name: multicluster-engine.v2.5.1
        version: 2.5.1
      - name: multicluster-engine.v2.5.0-103
        version: 2.5.0-103
      name: stable-2.5
    - currentCSV: multicluster-engine.v2.6.8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-08-04T11:16:16Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0 <2.6.8'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:69a40339bf3b3dd3c3b6359721895313147a5f682b62faf5151288ba63b65812
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:f6b15f827aeabc33074de1ff9ed20edaf04271b4e4d2eb16abbe786c78e033ca
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:9e1ea96b998b0c7854ca3c7c8a505fc96cbc1b35e6ce9f49b1dbec703fdc8070
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:7dde47f7fe16a26676a6e5b1739954d76cad1cddbefce2d6869ffb2af399e403
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:f563ff0acb4ea9d1bb112765b67bee535063304a98a2aa2cb7f21b0155d49262
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:54b8e17cacae2d355d82677ed334b4981b50e825579ff30406039933833e828b
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:a8117c99b33eedc702baeab32f0347892b587dfc1860f467cb1900565c3fdb3b
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:00ea8fe4945e93d2cbeceda4440acd6a53a48b4dd420dfe2cbd5287e20f0362f
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel8@sha256:42fee1066ff00066a172f82754ad9f08e8c29029e75f76712cb283d9fe5d1d9a
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:1d098eab26421f99ef71ffce82e4fa59ef0f9043de04ee74ca43486cbdb82006
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:b39819aa916974f4a408d23177e99cfb71e21af8ff9b0eacf6c37159c354ba70
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:0737e57ae8af6a62ac580710276d2c1620ac1b1283d3b84369db43e76e518610
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:5f9d5bdfd45f931989d069dcb1dea91a69020d83011fe0846ddfe22bba14b001
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel8@sha256:a630a20a633d057e475b4c39b342512119a71021e199aa0becf80899d8840ad7
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:c3c654be0b20cac501af1e31631464c32aecdbd16cbfb88f865900c70ccf77c2
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:7c5f339e762db1714b3f6cf1fc1bc8f08892cac433ab5a46e73b530fa08486fb
        - registry.redhat.io/multicluster-engine/cluster-api-rhel9@sha256:bd4eb05427ba2ea35f01227600cebd0fe8598d4de456348c24bb5e88d9ed8720
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:aff06db527e4ddc713a2fb8885824f0482647cbdcd844b311ccc3880a0dd2c3f
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:5ef8acc612817d478cede3b1bda55abafac00d0debebd68d948d157ee5b09286
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:294928c58b33ae8e776f86df211dac3b63a2cd9fc00aa995a0de037a29cb2c7d
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:2d807348774b8c6c0255c4bc931a956fc987c961fae6be689835410e8e301fd0
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel8@sha256:4d96379ad99d299d2a89eb1b0eb8a22db8bdada9a16248503cd8a9a5900eab35
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:d2c1d90a4e829b181fdf3d2838764c8f2b61c22b761279f53cb22316ad46ca03
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:93d65bfa0102c7ad7ec6ebfaad9ab062a31098b7ad7026313bcd8c6a495d5ca7
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:cd23f237c91dd8585946b31e17e38193f2e5ac8993d65459bb703a68f6fa8ca0
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:80a6615500949cc8c7a5b2a1e8c8c5f4564e9cdb9f995787a44f462e5bbd62cb
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel8@sha256:a703f2b1f4eb538aa02dbc0f4ff793506076480e94905c367b0d4ae352b87b65
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:5d488c8a644f2bfce97444d770516de10ceecc988cb3e9fbf0072d0b0a2eeb83
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:be1aa4cf5e09914774dcf69ab7ccdaf67504309d40c7e8cf4d91e9c1080e706f
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:82c0af2d285ba943dfa1da9f0856b26ec7e4c2c431577fcd4ab031ae040cfaea
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:87f5db53ad18f6d84501a5c2294cff73a5425276cdbde92225909826091b5ffa
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:46c4ac0cacef587b6e678b6cb0f68c29f5ac63bd59209d42a3ae0b7807e99972
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:d7c8bef4c3dec710ccb299482a10bfdf6efacd4e2ed2e6e323d448ed505000b3
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:d71ef80913d4e34389bce56c4a282cf7d2be705f003575dab87c07ca71f8c2d6
        version: 2.6.8
      entries:
      - name: multicluster-engine.v2.6.8
        version: 2.6.8
      - name: multicluster-engine.v2.6.7
        version: 2.6.7
      - name: multicluster-engine.v2.6.6
        version: 2.6.6
      - name: multicluster-engine.v2.6.5
        version: 2.6.5
      - name: multicluster-engine.v2.6.4
        version: 2.6.4
      - name: multicluster-engine.v2.6.3
        version: 2.6.3
      - name: multicluster-engine.v2.6.2
        version: 2.6.2
      - name: multicluster-engine.v2.6.1
        version: 2.6.1
      - name: multicluster-engine.v2.6.0
        version: 2.6.0
      name: stable-2.6
    - currentCSV: multicluster-engine.v2.7.5
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-21T22:32:21Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0 <2.7.5'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:44b07a18116ad827c23ca5a98b9fe8ef74a99abe4787cd01d848de42df44178d
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:15c2873703afcc74e41ae27dbb4761ad19c4bd2594ca6de82d3b6c02752e0ce3
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:816712bb3ccb58b705a85e5f4a55e519316fd8b9d7d9295d1042420ce44a4bad
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:e6a5e71f0aee8a682179cfd8f85b45a2ee04a1de680fc03dabdb659a2d2c0396
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:2a78ab192a554e78f3fc367c47e571f96b650ee9fb5ba0d06674f340f4320d94
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:9d1be5eb4fa357fe6bd97a52a53b3eeb537c431daf7af8a30efc3e0b6c5d5d66
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:2e3fe5e3b39721cc77b7d32b5682300eb60b1bc6940b8e83662b458d88489140
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:a24e9855a11d879d5308e29af6e3b2947365b7e466e7fa12dc57997140753604
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:b124242a55b1b2a2258cb12094617a0e009cf1cf58a60ab1b42d48d4ea2514ac
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:b66f2f4b5e0d77d501f9cc9da483f976f3c0eaf0ca72a5e9c9435c92d58e1388
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:00d415071f4c10575f9e5068868a7f80b0c0fb9c8614da86bc515923bed136a3
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:e2e16ff6213c0274126c9d64d495c0abbe17379e841ab730de66520908f1a190
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:4a0f1d3e696310f7a5bccbb358c538023a28f5bbc23b5743d10fd61b452d45d0
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:041044c08a1d526b349b5ad5fc4c4457f3632b1353530ed925c65ef2cce8e9c5
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:2025e424c11c6fe4a4395945c2e7d124d07cde43bb8d8cbe70286fe6ae345754
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:8c420720a48ea7c3aa096ae1929dcaa979c39e7865c5d43cba899e294745fc94
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:081151ca906b0f1afbcd3ee9c6d9e5bfe234494223c660ef376aa4004b0508f4
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:3defab1abc7395c9f9c22912ec6825be11b120a1416404eb0599e91f7f9c679a
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:c0fb7985620d5bd0dfadd57cb75664ee6777b8dac5a3463967e9b649e9eb836c
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:9e5c66ffdf39ae923ecccf79be538b883f2bdc8019169d55057565f077997d67
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:4fa34ab0aa8854eb5e942f2ed315041bf2026ce0bed86f57125878af0c73dc8b
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:7369bc79a283ff8b34c3189ea83e55f88a00d619cc78f8dad03d302dcb34a550
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:fedc699c0f786fb303d3956769f9c5fca1473ec4bf014ec87c5a4d40e34dcdb3
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:b82b1837dfc24a367f68741dece6d19c89c576706f006ee8d5123a0c6782c492
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:90c19d2aaa50e7cc9184f2f220b568c34c9fdcfdcba314da555885c999078057
        - registry.redhat.io/multicluster-engine/cluster-api-rhel9@sha256:cd8d6e0056fb0436275c12cc6b42f11c0ccd06729c7108e964b03d6c48d063bd
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:0684bcde2c018ae893bd71be8df3448afa09d199415e83285d2c740cff66192d
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:2216b015943eb8d9f25bf4ac37fbad2b7215a5c0003bd7c3b3d97b3b0a5ba708
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:555fcfce226e3b059f8f911d44a48578b898e085b6ad2575479feee71ec8baaf
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:a6736c7cac00f78c0e8877549ae2b74b7207cc348f24f175133db49fd38f240a
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:8e41d489e1def9b540ddd6387a1cc151b24e91a5fa44a74fbbaa0d1978588d57
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:45571dbb55bd213400eec6d2065b40615185bf3e52efdf495ab73201ec94faff
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:1c176688f13052783662115d5cabe8d614b8d9f79507dc2bdd4be1604e2eecc0
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:2cbd8c470dc4b263fd11952aa792094a7550f4b99e10f22186d51621bc8d3b3f
        version: 2.7.5
      entries:
      - name: multicluster-engine.v2.7.5
        version: 2.7.5
      - name: multicluster-engine.v2.7.4
        version: 2.7.4
      - name: multicluster-engine.v2.7.3
        version: 2.7.3
      - name: multicluster-engine.v2.7.2
        version: 2.7.2
      - name: multicluster-engine.v2.7.1
        version: 2.7.1
      - name: multicluster-engine.v2.7.0
        version: 2.7.0
      name: stable-2.7
    - currentCSV: multicluster-engine.v2.8.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-09-08T13:08:30Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0 <2.8.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:f869555beec0875c74f30a88428ac9641029dfed0365d45b23d5cae469cf7d0f
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:7ca7e87615e12b9d7fa75befd05f4f45bdc4bf8ef1b576c69f36a23c66c90cd6
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:988eb4aa01ba0d7e6728b23d043e4c0babe872ac23b6520f4daac714619a360b
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:f809a8774b0e3f891af46931dd4d2c11b95da4e7f5a8aee1814152d9a468d357
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:c5ac848e3b7eb88cdbdbf0b5f4aa96698d852b3a8118cf7146eb7e1948afaeba
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:a7f63077799a62fe6ece76a12346a87b5cbae966ab69df6ec6d951259dbf0e4a
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:c3b62ab1234257d30f41fa1ab33cb319321777a18de8ed751be1fcd48b204f2e
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:318ca0c7e9cd9a7336178519cadbf21eaeb474cd4a8eb5cf411a640d1c77372e
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:5ea86c20383a6932c593e020fd3b7b1a2a8e666bc374bff8b4bca0ee2d4ecc33
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:aa2754d35c8460a2879d4d523ebf32351d36e8fd7886335b5c0d2ca4befb8e8e
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:bee8f0c047cf8dbfa3f28712ce91bea55e926b2edb26c394f18a2d780e142083
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:7c86c89ff57091d4112783d82d1add47f2d56dba584418d5664c198ba048e0c2
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:f708f3e1007ed2b0e420e51d8b7ff9dd9bccaea9234755f4375f5514018211b4
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:3788b2674dda0b2cb4e9ca21bdce0b9e44702d72458b2396ef0acdad37fa866b
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:18b0e86305afe2eae3c26e637755fa9e8ab2abf30bc15e5428467a9a81ec99b0
        - registry.redhat.io/openshift4/ose-aws-cluster-api-controllers-rhel9@sha256:ada8b480a86d0a9e3ecc58f8b1cb53dbd7452b2b9183b3ba2b15a6a58c4436d1
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:f694b6e9e76941b11cc8a15153d9005cffe9bbbf58d0962f8ff546e90a63d93c
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:806b6c956f7bad824f0fe9acc3b534964e8f5f3399a1813f75511bc0f3904414
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:cf457b5a13cdd54c45a7b88f0f150a04e756bfac727c4365cb12f113d904937d
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:5461643a61850d95409d989e81c304053372f1dc5dbba291f3f3386de9ecab1c
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:261b99b93d98825ade0a0d3b9c4e465a785f35d2a310dedd4bbd6e4aa1f33ef9
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:0392c81a404be343d658dbc64153cbc46a64d0adf69aa8024c3f727351102d75
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:8379827635dfc4b51e04145d8048f7dcf9c6956ea1a22371bd22f948ae01b183
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:ee014c576263e19798df98460ebb5f1a21cb6249f09fe8a267b2dfa98b07d1eb
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:d4d11d0aeb3e1c9ff2a80e769d92dce509a0202a2ef1c9d2430030acdbf85bce
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:83a82f9558c406899e573308947c15b1f36de454006bb0e2ac44709f0c2c28d7
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:ea2ba066abd170e1087558c7b701faeb1be2e1667943dc14440f19c2ee5b0c30
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:fb2b8cef9916f387d26e65d9b53c7567ae1433d43d16e879ddb582ddc248eaa3
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:5ad96dad74c6f4967bd5fd397d6697e01b3b25bb3b8bae123a38e96964ddfa17
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:b9afa4d5328f17c7c3ce1dc105d177ecacf3447092d51a50f250a4b0ea692e2c
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:c6cd578123c1c86fcb3305ecca117543106f943cbcc4baf9065683988b2e10ff
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:e5e459e665afc14a1540e300015222b7c1ac896e2bee889283d67c3e19ca01ac
        - registry.redhat.io/openshift4/ose-cluster-api-rhel9@sha256:ec58463239eaa9bd15b31cbcfea0a34fd650201bac6120aa7ddaf72734451e06
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:ab0898d513b9b5ca2f5ca5879adc4e73da3232d515c0647a2a0ec28792264db1
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:0d1dfcb06eaa26de74994891371a5657aded7a59972f006f8c4a3859369877f3
        version: 2.8.3
      entries:
      - name: multicluster-engine.v2.8.3
        version: 2.8.3
      - name: multicluster-engine.v2.8.2
        version: 2.8.2
      - name: multicluster-engine.v2.8.1
        version: 2.8.1
      - name: multicluster-engine.v2.8.0
        version: 2.8.0
      name: stable-2.8
    - currentCSV: multicluster-engine.v2.9.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-30T19:22:43Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0 <2.9.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/open-cluster-management/backplane-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational components for managing multiple Kubernetes-based clusters. The deployment of the multicluster engine components is
              determined based on the configuration that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:8901e984b3b47f0650cddc1c36bdd89af22390f3d4a562d562fde7774d8ba6db
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:f3d65f68538e94260fad54f145433a36366f8c88f9c9f80b7fc5545ba62e1773
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:0a9c0ccb6a75273474fb04b785aa8fc58aec2f9ffbff8c232d359e14892bbae2
        - registry.redhat.io/multicluster-engine/capoa-control-plane-rhel9@sha256:b510a25a6370ca6b8d2f74165814ebfdae0ce9898cc343e73fab6344eebf2bed
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:8db1386d12d40dfbc80054c023897e5bfc0a987f6d07792b0f9f50f59f1a68d8
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:57699ecb529cb46843de847deef3e3757ca252910249d8deac16fcb00c01b636
        - registry.redhat.io/multicluster-engine/capoa-bootstrap-rhel9@sha256:a8b1278a2df9b17cee66813715995385bb3539af05e4df9b5bf7d9ccd508c313
        - registry.redhat.io/multicluster-engine/mce-capi-webhook-config-rhel9@sha256:7f1f70813e6d753b4acb04c7e9d7acb0047ad91f8a9a10e0fe2fff6c4411c8b4
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:a91276fcf5a589ef11f01b39df7d33ceb58ebcb47f95dfab772114bce7c1b4a8
        - registry.redhat.io/openshift4/ose-cluster-api-rhel9@sha256:9a72b8081a57846d942981ced07da86ad8e197309a1bc3c9d53876528b58f8ec
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:6a75df76fb0d9516b13fa6b079bc6c4d49dc6774f0d4f31343abfb8d70edcc23
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:ce0fd7bdc73cd2c84d1a9c5af528916488ca8116dd05d83dacf72f2af4248e65
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:ac43feaf079fa0d250d53cd9685fb68677d888dd091730faa32757dcdf2c95e4
        - registry.redhat.io/rhel8/postgresql-12@sha256:4773a29755ea826c0d3b58a8b8297c1d7388019516195cae9bd815f28ec4bca0
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:5e6594464bf7ddc2a66f127f645d93ad35d8cd95dd70544fa7fcec791f67f45c
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:36a6000b6482966f8bbf86e45c25335c066736ae1f8695a8b68e6b76b7c491ac
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:1b218f0aa0ea1f494094e3bc783ef40968cbff780e22de186d578a1d64c33bf8
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:3c450cb6a3ba6f2e3028856f14967803c1a691d95b8f54bd9d29d93691d6e795
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:8076db8ccd5578cb5aba2d66f53790563d3f42b93a6cfd14c7094d3a25707982
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:ad382d5eb347c448c46495a0a17503d789db4eb1096ad71a588b316b235e437f
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:6b95560a6a6dc1ddd41c76f275f83baec52ec9d79de04bfe54feaed02db123a3
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:ee75e805f0f9d73f073ce870f5d46962c0033d63158017841d17410febfc2c57
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:25ec9a0cc4ac92d4bed75c38fa37ef8d5ff9478cc475049b303d74eb58687a81
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:f8c3898e29f8c0a20c3be92bc4ee5a9443b9fc8218db95ba541fe3e57a89c40d
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:bf343592bf2f4399b5ac61709f56528435990e903812e264cbef845ff4433231
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:9555fd5f577f32a1d8ebe25f0d03e453c8ea925e4a12dbbe1015c58d3c8528b0
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:1cc6fddaf9e55637e90d5c3e598467df8967b8a362f8b8d46286b256bad6dd3f
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:a41cafc65e30deb073e051542702e258901b552ad00e42eacab05d149d6f0565
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:ea57247cf11f6ab966893ce4e6d2948232a4314588af2ca33c1d9689da7482dd
        - registry.redhat.io/openshift4/ose-baremetal-cluster-api-controllers-rhel9@sha256:f855d5ce17d3ac4560afc73c35d03f045f84551fdccc00312c890288a889b9ce
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:72a4ac73bb5e7797329e1856f785751d48815a64e82350a5d694bb5a8e121956
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:cd9dc89b933b394b08fe05ab5f0bbecea727bfb2c29770f11ec1db4aa5740b3c
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:37087ad8678f3e8ec8e2b2d035d3a12087ce9e0c993194a04a9874fd07e80d91
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:cecec1f7d7f0d37cf17fb25ccf34602728f2a09c61af7e212967b4ee28d2dec3
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:d068d73f41d8f8ecbec36752d0a432913bfbe2580f489abefe513c06233e80c9
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:edbb174cb69dd31339a53542d24c4fd00552e41fa09d7e2553baa7a1ad8dd810
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:1bbd134d44839258172e013fe84bfd174d36d6b7232e2cc8a3d8556998181d9d
        - registry.redhat.io/openshift4/ose-aws-cluster-api-controllers-rhel9@sha256:cf13b0658271cd0dbd1ce6dcc350e64c5459ef02ba4dc88a0031de9d6f26ffd1
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:8d189b565e4c3c434988aa4ff0ecad4c5f86b2a94120325c142ff111b25facde
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:902c5b3b4a4e8cb0a14fa5aba8451ec02830d5167ff1b91323ded62255983cd9
        version: 2.9.0
      entries:
      - name: multicluster-engine.v2.9.0
        version: 2.9.0
      name: stable-2.9
    defaultChannel: stable-2.9
    packageName: multicluster-engine
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JohnSnowLabs
      provider-url: ""
    name: nlp-server
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nlp-server.v0.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.johnsnowlabs.com/v1",
                "kind": "Nlpserver",
                "metadata": {
                  "name": "nlpserver-sample"
                },
                "spec": {
                  "ExtraEnv": {
                    "johnsnowlabs_client_id": "sI4MKSmLHOX2Pg7XhM3McJS2oyKG5PHcp0BlANEW",
                    "nb_uid": "1000",
                    "nb_user": "johnsnowlabs",
                    "pyspark_version": "3.0.2",
                    "spark_nlp_version": "3.4.4"
                  },
                  "affinity": {},
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "johnsnowlabs/nlp-server",
                    "tag": "0.8.0rc1"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {
                      "kubernetes.io/ingress.class": "nginx",
                      "nginx.ingress.kubernetes.io/ssl-redirect": "true"
                    },
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "nlp-server.jsl.aiplatform.eks.test.johnsnowlabs.com",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": [
                      {
                        "hosts": [
                          "nlp-server.jsl.aiplatform.eks.test.johnsnowlabs.com"
                        ],
                        "secretName": "annotationlab-tls-secret"
                      }
                    ]
                  },
                  "livenessProbe": "httpGet:\n  path: /\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 3\n",
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "readinessProbe": "httpGet:\n  path: /\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 3\n",
                  "replicaCount": 1,
                  "resources": {},
                  "securityContext": {},
                  "service": {
                    "port": 5000,
                    "portName": "http",
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": "nlp-server-sa"
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.johnsnowlabs.com/nlp-server/nlp-server-operator:0.8.6
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nlpserver
            name: nlpservers.apps.johnsnowlabs.com
            version: v1
        description: NLPServer from JohnSnow labs
        displayName: nlp-server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nlp-server
        links:
        - name: Nlp Server
          url: https://johnsnowlabs.com
        maturity: alpha
        provider:
          name: JohnSnowLabs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - registry.johnsnowlabs.com/nlp-server/nlp-server-operator:0.8.6
        version: 0.8.6
      entries:
      - name: nlp-server.v0.8.6
        version: 0.8.6
      name: alpha
    defaultChannel: alpha
    packageName: nlp-server
    provider:
      name: JohnSnowLabs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator-certified.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator-certified.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator-certified.v2.6.1
        version: 2.6.1
      - name: percona-postgresql-operator-certified.v2.6.0
        version: 2.6.0
      - name: percona-postgresql-operator-certified.v2.5.0
        version: 2.5.0
      name: stable
    - currentCSV: percona-postgresql-operator-certified.v2.7.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        version: 2.7.0-cw
      entries:
      - name: percona-postgresql-operator-certified.v2.7.0-cw
        version: 2.7.0-cw
      - name: percona-postgresql-operator-certified.v2.6.1-cw
        version: 2.6.1-cw
      - name: percona-postgresql-operator-certified.v2.6.0-cw
        version: 2.6.0-cw
      - name: percona-postgresql-operator-certified.v2.5.0-cw
        version: 2.5.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-postgresql-operator-certified
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Starburst Data, Inc.
      provider-url: https://www.starburstdata.com
    name: starburst-enterprise-helm-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: starburst-enterprise-helm-operator.v402.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstEnterprise",
                "metadata": {
                  "name": "starburstenterprise-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "cache": {
                    "diskUsagePercentage": 80,
                    "enabled": false,
                    "ttl": "7d",
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "catalogs": {
                    "tpch": "connector.name=tpch\ntpch.splits-per-node=4"
                  },
                  "commonLabels": {},
                  "coordinator": {
                    "additionalProperties": "",
                    "affinity": {},
                    "deploymentAnnotations": {},
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=true\nnode-scheduler.include-coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://localhost:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "resources": {
                      "cpu": 16,
                      "memory": "60Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "tolerations": []
                  },
                  "environment": null,
                  "expose": {
                    "clusterIp": {
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      }
                    },
                    "ingress": {
                      "annotations": {},
                      "host": null,
                      "ingressClassName": null,
                      "ingressName": "coordinator-ingress",
                      "path": "/",
                      "pathType": "ImplementationSpecific",
                      "serviceName": "starburst",
                      "servicePort": 8080,
                      "tls": {
                        "enabled": true,
                        "secretName": null
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "nodePort": 30080,
                          "port": 8080
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "externalSecrets": {
                    "enabled": false,
                    "goDaddy": {
                      "backendType": "secretsManager"
                    },
                    "secretPrefix": "external0",
                    "type": "goDaddy"
                  },
                  "extraArguments": [],
                  "extraSecret": {
                    "file": null,
                    "name": null
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise",
                    "tag": "402-e.1"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "initImage": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise-init",
                    "tag": "1.3.2"
                  },
                  "internal": {
                    "ports": {
                      "http": {
                        "port": 8080
                      },
                      "https": {
                        "port": 8443
                      }
                    }
                  },
                  "internalTls": false,
                  "licenseManager": {
                    "cronjob": {
                      "affinity": {},
                      "nodeSelector": {},
                      "priorityClassName": null,
                      "securityContext": {},
                      "tolerations": []
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "registry.connect.redhat.com/starburst/license-manager",
                      "tag": "0.0.2"
                    },
                    "url": "https://license-api.dogfood.eng.starburstdata.net/test/license/checkout"
                  },
                  "livenessProbe": null,
                  "prometheus": {
                    "agent": {
                      "config": "/etc/starburst/telemetry/prometheus.yaml",
                      "port": 8081,
                      "version": "0.16.1"
                    },
                    "enabled": true,
                    "rules": [
                      {
                        "attrNameSnakeCase": true,
                        "name": "$1",
                        "pattern": "trino.execution<name=QueryManager><>(running_queries|queued_queries)",
                        "type": "GAUGE"
                      },
                      {
                        "name": "failed_queries",
                        "pattern": "trino.execution<name=QueryManager><>FailedQueries\\.TotalCount",
                        "type": "COUNTER"
                      }
                    ]
                  },
                  "query": {
                    "maxConcurrentQueries": 3
                  },
                  "readinessProbe": null,
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "securityContext": {},
                  "sharedSecret": null,
                  "siac": {
                    "additionalHeapSizePercentage": 15,
                    "enabled": false
                  },
                  "spilling": {
                    "enabled": false,
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "startupProbe": null,
                  "usageMetrics": {
                    "enabled": false,
                    "usageClient": {
                      "initialDelay": "1m",
                      "interval": "1m"
                    }
                  },
                  "userDatabase": {
                    "enabled": false,
                    "users": [
                      {
                        "password": "46991b33f7a75ff79213c0dc0e610610",
                        "username": "admin"
                      }
                    ]
                  },
                  "worker": {
                    "additionalProperties": "",
                    "affinity": {},
                    "autoscaling": {
                      "enabled": false,
                      "maxReplicas": 100,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "deploymentAnnotations": {},
                    "deploymentTerminationGracePeriodSeconds": 300,
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://coordinator:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "kedaScaler": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.connect.redhat.com/starburst/keda-trino-scaler",
                        "tag": "0.1.7"
                      },
                      "port": 8021,
                      "scaledObjectSpec": {
                        "advanced": {},
                        "cooldownPeriod": 300,
                        "idleReplicaCount": 0,
                        "maxReplicaCount": 100,
                        "minReplicaCount": 1,
                        "pollingInterval": 30,
                        "scaleTargetRef": {
                          "name": "worker"
                        },
                        "triggers": [
                          {
                            "metadata": {
                              "numberOfQueriesPerWorker": "10",
                              "scaleInToIdleReplicaCountIfNoQueuedQueriesLeft": "true",
                              "scaleMethod": "query_queue"
                            },
                            "type": "external"
                          }
                        ]
                      }
                    },
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "replicas": 2,
                    "resources": {
                      "cpu": 16,
                      "memory": "100Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "starburstWorkerShutdownGracePeriodSeconds": 120,
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstHive",
                "metadata": {
                  "name": "starbursthive-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "affinity": {},
                  "commonLabels": {},
                  "database": {
                    "external": {
                      "driver": null,
                      "jdbcUrl": null,
                      "password": null,
                      "user": null
                    },
                    "internal": {
                      "databaseName": "hive",
                      "databasePassword": "HivePass1234",
                      "databaseUser": "hive",
                      "driver": "org.postgresql.Driver",
                      "env": [],
                      "envFrom": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.redhat.io/rhel8/postgresql-10",
                        "tag": "1"
                      },
                      "port": 5432,
                      "resources": {
                        "limits": {
                          "cpu": 2,
                          "memory": "1Gi"
                        },
                        "requests": {
                          "cpu": 2,
                          "memory": "1Gi"
                        }
                      },
                      "volume": {
                        "emptyDir": {}
                      }
                    },
                    "type": "internal"
                  },
                  "deploymentAnnotations": {},
                  "env": [],
                  "envFrom": [],
                  "expose": {
                    "clusterIp": {
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "nodePort": 30083,
                          "port": 9083
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "extraArguments": [],
                  "hdfs": {
                    "hadoopUserName": null
                  },
                  "heapSizePercentage": 85,
                  "hiveMetastoreStorageAuthorization": true,
                  "hiveMetastoreWarehouseDir": null,
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/hive",
                    "tag": "3.1.2-e.22"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "nodeSelector": {},
                  "objectStorage": {
                    "adl": {
                      "oauth2": {
                        "clientId": null,
                        "credential": null,
                        "refreshUrl": null
                      }
                    },
                    "awsS3": {
                      "accessKey": null,
                      "endpoint": null,
                      "pathStyleAccess": false,
                      "region": null,
                      "secretKey": null
                    },
                    "azure": {
                      "abfs": {
                        "accessKey": {
                          "accessKey": null,
                          "storageAccount": null
                        },
                        "authType": "accessKey",
                        "oauth": {
                          "clientId": null,
                          "endpoint": null,
                          "secret": null
                        }
                      },
                      "wasb": {
                        "accessKey": null,
                        "storageAccount": null
                      }
                    },
                    "gs": {
                      "cloudKeyFileSecret": null
                    }
                  },
                  "podAnnotations": {},
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "resources": {
                    "limits": {
                      "cpu": 1,
                      "memory": "1Gi"
                    },
                    "requests": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "securityContext": {},
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
          description: Deploys StarburstEnterprise Helm Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Starburst Data
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: StarburstEnterprise
            name: starburstenterprises.charts.starburstdata.com
            version: v1alpha1
          - kind: StarburstHive
            name: starbursthives.charts.starburstdata.com
            version: v1alpha1
        description: CRs for deploying starburst enterprise and hive
        displayName: Starburst Enterprise Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - starburst
        - trino
        - sep
        - bigdata
        - data
        links:
        - name: Starburst Enterprise Helm Operator
          url: https://www.starburstdata.com
        maintainers:
        - email: hello@starburstdata.com
          name: Starburst Data, Inc.
        maturity: alpha
        provider:
          name: Starburst Data, Inc.
          url: https://www.starburstdata.com
        relatedImages:
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
        version: 402.1.0
      entries:
      - name: starburst-enterprise-helm-operator.v402.1.0
        version: 402.1.0
      name: alpha
    defaultChannel: alpha
    packageName: starburst-enterprise-helm-operator-rhmp
    provider:
      name: Starburst Data, Inc.
      url: https://www.starburstdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: keystone-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keystone-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keystone.openstack.org/v1beta1",
                "kind": "KeystoneAPI",
                "metadata": {
                  "name": "keystone"
                },
                "spec": {
                  "adminProject": "admin",
                  "adminUser": "admin",
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "keystone",
                  "databaseInstance": "openstack",
                  "preserveJobs": false,
                  "region": "regionOne",
                  "resources": {
                    "requests": {
                      "cpu": "1.0",
                      "memory": "500Mi"
                    }
                  },
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "keystone.openstack.org/v1beta1",
                "kind": "KeystoneEndpoint",
                "metadata": {
                  "name": "placement"
                },
                "spec": {
                  "endpoints": {
                    "admin": "http://placement-admin-openstack.apps-crc.testing",
                    "internal": "http://placement-internal-openstack.apps-crc.testing",
                    "public": "http://placement-public-openstack.apps-crc.testing"
                  },
                  "serviceName": "placement"
                }
              },
              {
                "apiVersion": "keystone.openstack.org/v1beta1",
                "kind": "KeystoneService",
                "metadata": {
                  "name": "placement"
                },
                "spec": {
                  "enabled": true,
                  "passwordSelector": "PlacementPassword",
                  "secret": "osp-secret",
                  "serviceDescription": "Placement service",
                  "serviceName": "placement",
                  "serviceType": "placement",
                  "serviceUser": "placement"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:01Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KeystoneAPI is the Schema for the keystoneapis API
            displayName: Keystone API
            kind: KeystoneAPI
            name: keystoneapis.keystone.openstack.org
            version: v1beta1
          - description: KeystoneEndpoint is the Schema for the keystoneendpoints
              API
            displayName: Keystone Endpoint
            kind: KeystoneEndpoint
            name: keystoneendpoints.keystone.openstack.org
            version: v1beta1
          - description: KeystoneService is the Schema for the keystoneservices API
            displayName: Keystone Service
            kind: KeystoneService
            name: keystoneservices.keystone.openstack.org
            version: v1beta1
        description: Keystone Operator
        displayName: Keystone Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Identity
        - Keystone
        links:
        - name: Keystone Operator
          url: https://github.com/openstack-k8s-operators/keystone-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-keystone-rhel9@sha256:740e92a78d1e3874d2a40e37c970c6163d9e295dbc236914f1204baed070c595
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/keystone-rhel9-operator@sha256:8b055945a3808f7397265a10e6a6463af0ee3ef80fdff9d66402b488e5b70523
        version: 1.0.6
      entries:
      - name: keystone-operator.v1.0.6
        version: 1.0.6
      - name: keystone-operator.v1.0.5
        version: 1.0.5
      - name: keystone-operator.v1.0.4
        version: 1.0.4
      - name: keystone-operator.v1.0.3
        version: 1.0.3
      - name: keystone-operator.v1.0.2
        version: 1.0.2
      - name: keystone-operator.v1.0.1
        version: 1.0.1
      - name: keystone-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: keystone-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: ""
    name: ingress-node-firewall
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ingress-node-firewall.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ingressnodefirewall.openshift.io/v1alpha1",
                "kind": "IngressNodeFirewall",
                "metadata": {
                  "name": "ingressnodefirewall-demo2"
                },
                "spec": {
                  "ingress": [
                    {
                      "rules": [
                        {
                          "action": "Deny",
                          "order": 10,
                          "protocolConfig": {
                            "icmp": {
                              "icmpType": 8
                            },
                            "protocol": "ICMP"
                          }
                        },
                        {
                          "action": "Deny",
                          "order": 20,
                          "protocolConfig": {
                            "protocol": "TCP",
                            "tcp": {
                              "ports": "8000-9000"
                            }
                          }
                        }
                      ],
                      "sourceCIDRs": [
                        "172.16.0.0/12"
                      ]
                    },
                    {
                      "rules": [
                        {
                          "action": "Deny",
                          "order": 10,
                          "protocolConfig": {
                            "icmpv6": {
                              "icmpType": 128
                            },
                            "protocol": "ICMPv6"
                          }
                        }
                      ],
                      "sourceCIDRs": [
                        "fc00:f853:ccd:e793::0/64"
                      ]
                    }
                  ],
                  "interfaces": [
                    "eth0"
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "do-node-ingress-firewall": "true"
                    }
                  }
                }
              },
              {
                "apiVersion": "ingressnodefirewall.openshift.io/v1alpha1",
                "kind": "IngressNodeFirewallConfig",
                "metadata": {
                  "name": "ingressnodefirewallconfig",
                  "namespace": "ingress-node-firewall-system"
                },
                "spec": {
                  "nodeSelector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ingress-node-firewall-rhel9-operator@sha256:144403b009d1b5f2b2ee25ed3cebc525f3588f21bfa671b25c662cf800798839
          createdAt: "2024-03-13T12:36:01Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.11.0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-ingress-node-firewall
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/ingress-node-firewall
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: IngressNodeFirewallConfig
            name: ingressnodefirewallconfigs.ingressnodefirewall.openshift.io
            version: v1alpha1
          - kind: IngressNodeFirewallNodeState
            name: ingressnodefirewallnodestates.ingressnodefirewall.openshift.io
            version: v1alpha1
          - kind: IngressNodeFirewall
            name: ingressnodefirewalls.ingressnodefirewall.openshift.io
            version: v1alpha1
        description: The Ingress node firewall operator, manages the configuration
          of cluster scoped Ingress Node firewall rules for Kubernetes.
        displayName: Ingress Node Firewall Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ingressnodefirewall
        links:
        - name: Ingress Node Firewall
          url: https://github.com/openshift/ingress-node-firewall
        maintainers:
        - email: mmahmoud@redhat.com
          name: msherif1234
        maturity: alpha
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ingress-node-firewall-rhel9@sha256:4e3d69793f9280f6910a6749e80e8f6fe5d20c7a2c160d5682f310309c497e64
        - registry.redhat.io/openshift4/ingress-node-firewall-rhel9-operator@sha256:144403b009d1b5f2b2ee25ed3cebc525f3588f21bfa671b25c662cf800798839
        version: 4.16.0-202509240123
      entries:
      - name: ingress-node-firewall.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: ingress-node-firewall.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: ingress-node-firewall.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: ingress-node-firewall.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: ingress-node-firewall.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: ingress-node-firewall.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: ingress-node-firewall.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: ingress-node-firewall.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: ingress-node-firewall.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: ingress-node-firewall.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: ingress-node-firewall.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: ingress-node-firewall.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: ingress-node-firewall.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: ingress-node-firewall.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: ingress-node-firewall.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: ingress-node-firewall.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: ingress-node-firewall.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: ingress-node-firewall.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: ingress-node-firewall.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: ingress-node-firewall.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: ingress-node-firewall.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: ingress-node-firewall.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: ingress-node-firewall.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: ingress-node-firewall.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: ingress-node-firewall.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: ingress-node-firewall.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: ingress-node-firewall.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: ingress-node-firewall.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: ingress-node-firewall.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: ingress-node-firewall.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: ingress-node-firewall.v4.16.0-202405281506
        version: 4.16.0-202405281506
      name: stable
    defaultChannel: stable
    packageName: ingress-node-firewall
    provider:
      name: Red Hat, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cluster-observability-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-observability-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e
          createdAt: "2025-06-24T11:30:26Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=..0 <1.2.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-cluster-observability-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs",
              "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs",
              "prometheusagents.monitoring.rhobs",
              "perses.perses.dev"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
          support: Cluster Observability (https://issues.redhat.com/projects/COO/)
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: Perses is the Schema for the perses API
            displayName: Perses
            kind: Perses
            name: perses.perses.dev
            version: v1alpha1
          - description: A Perses Dashboard
            displayName: Perses Dashboard
            kind: PersesDashboard
            name: persesdashboards.perses.dev
            version: v1alpha1
          - description: A Perses Datasource
            displayName: Perses Datasource
            kind: PersesDatasource
            name: persesdatasources.perses.dev
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
          - description: UIPlugin defines a console plugin for observability
            displayName: UIPlugin
            kind: UIPlugin
            name: uiplugins.observability.openshift.io
            version: v1alpha1
        description: |-
          Cluster Observability Operator is a Go based Kubernetes operator to easily setup and manage various observability tools.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          - Setup console plugins
          - Setup korrel8r
          - Setup Perses
          - Setup Cluster Health Analyzer
          ### Documentation
          - **[Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/cluster_observability_operator/index)**
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)
        displayName: Cluster Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad
        - registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf
        - registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd
        - registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a
        - registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552
        - registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:e2681bce57dc9c15701f5591532c2dfe8f19778606661339553a28dc003dbca5
        - registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e
        - registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:e54c1e1301be66933f3ecb01d5a0ca27f58aabfd905b18b7d057bbf23bdb7b0d
        - registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:8597c48fc71fc6ec8e87dbe40dace4dbb7b817c1039db608af76a0d90f7ac2d0
        - registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404
        - registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223
        - registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00
        - registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1
        - registry.redhat.io/cluster-observability-operator/perses-0-1-rhel9-operator@sha256:bfed9f442aea6e8165644f1dc615beea06ec7fd84ea3f8ca393a63d3627c6a7c
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed
        version: 1.2.2
      entries:
      - name: cluster-observability-operator.v1.2.2
        version: 1.2.2
      - name: cluster-observability-operator.v1.2.1
        version: 1.2.1
      - name: cluster-observability-operator.v1.2.0
        version: 1.2.0
      - name: cluster-observability-operator.v1.1.1
        version: 1.1.1
      - name: cluster-observability-operator.v1.1.0
        version: 1.1.0
      - name: cluster-observability-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: cluster-observability-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kinesis-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kinesis-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kinesis.services.k8s.aws/v1alpha1",
                "kind": "Stream",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kinesis-controller:1.1.1
          createdAt: "2025-09-19T20:56:43Z"
          description: AWS Kinesis controller is a service controller for managing
            Kinesis resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Stream represents the state of an AWS kinesis Stream resource.
            displayName: Stream
            kind: Stream
            name: streams.kinesis.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Kinesis resources in AWS from within your Kubernetes cluster.

          **About Amazon Kinesis**

          You can use Amazon Kinesis Data Streams to collect and process large streams of data records in real time. You can create data-processing applications, known as Kinesis Data Streams applications. A typical Kinesis Data Streams application reads data from a data stream as data records. These applications can use the Kinesis Client Library, and they can run on Amazon EC2 instances.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Kinesis
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kinesis
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Kinesis Developer Resources
          url: https://aws.amazon.com/kinesis/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kinesis maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/kinesis-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-kinesis-controller.v1.1.1
        version: 1.1.1
      - name: ack-kinesis-controller.v1.1.0
        version: 1.1.0
      - name: ack-kinesis-controller.v1.0.16
        version: 1.0.16
      - name: ack-kinesis-controller.v1.0.15
        version: 1.0.15
      - name: ack-kinesis-controller.v1.0.14
        version: 1.0.14
      - name: ack-kinesis-controller.v1.0.13
        version: 1.0.13
      - name: ack-kinesis-controller.v1.0.12
        version: 1.0.12
      - name: ack-kinesis-controller.v1.0.11
        version: 1.0.11
      - name: ack-kinesis-controller.v1.0.10
        version: 1.0.10
      - name: ack-kinesis-controller.v1.0.9
        version: 1.0.9
      - name: ack-kinesis-controller.v1.0.8
        version: 1.0.8
      - name: ack-kinesis-controller.v1.0.7
        version: 1.0.7
      - name: ack-kinesis-controller.v1.0.6
        version: 1.0.6
      - name: ack-kinesis-controller.v1.0.5
        version: 1.0.5
      - name: ack-kinesis-controller.v1.0.4
        version: 1.0.4
      - name: ack-kinesis-controller.v1.0.3
        version: 1.0.3
      - name: ack-kinesis-controller.v1.0.2
        version: 1.0.2
      - name: ack-kinesis-controller.v1.0.1
        version: 1.0.1
      - name: ack-kinesis-controller.v1.0.0
        version: 1.0.0
      - name: ack-kinesis-controller.v0.0.19
        version: 0.0.19
      - name: ack-kinesis-controller.v0.0.18
        version: 0.0.18
      - name: ack-kinesis-controller.v0.0.17
        version: 0.0.17
      - name: ack-kinesis-controller.v0.0.16
        version: 0.0.16
      - name: ack-kinesis-controller.v0.0.15
        version: 0.0.15
      - name: ack-kinesis-controller.v0.0.10
        version: 0.0.10
      - name: ack-kinesis-controller.v0.0.9
        version: 0.0.9
      - name: ack-kinesis-controller.v0.0.8
        version: 0.0.8
      - name: ack-kinesis-controller.v0.0.6
        version: 0.0.6
      - name: ack-kinesis-controller.v0.0.5
        version: 0.0.5
      - name: ack-kinesis-controller.v0.0.4
        version: 0.0.4
      - name: ack-kinesis-controller.v0.0.3
        version: 0.0.3
      - name: ack-kinesis-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-kinesis-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Seldon Technologies
      provider-url: ""
    name: seldon-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: seldon-operator.v1.14.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1",
                "kind": "SeldonDeployment",
                "metadata": {
                  "labels": {
                    "app": "seldon",
                    "app.kubernetes.io/instance": "seldon1",
                    "app.kubernetes.io/name": "seldon",
                    "app.kubernetes.io/version": "v0.5"
                  },
                  "name": "seldon-model"
                },
                "spec": {
                  "name": "test-deployment",
                  "predictors": [
                    {
                      "componentSpecs": [
                        {
                          "spec": {
                            "containers": [
                              {
                                "image": "seldonio/mock_classifier:1.6.0",
                                "name": "classifier"
                              }
                            ]
                          }
                        }
                      ],
                      "graph": {
                        "children": [],
                        "name": "classifier",
                        "type": "MODEL"
                      },
                      "name": "example",
                      "replicas": 1
                    }
                  ]
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: docker.io/seldonio/seldon-core-operator:1.14.1
          createdAt: "2019-05-21 15:00:00"
          description: The Seldon operator for management, monitoring and operations
            of machine learning systems through the Seldon Engine. Once installed,
            the Seldon Operator provides multiple functions which facilitate the productisation,
            monitoring and maintenance of machine learning systems at scale.
          olm.skipRange: <1.13.1
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/SeldonIO/seldon-core
          support: Clive Cox
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A seldon engine deployment
            displayName: Seldon Deployment
            kind: SeldonDeployment
            name: seldondeployments.machinelearning.seldon.io
            version: v1
        description: "The Seldon operator enables for native operation of production
          machine learning workloads, including monitoring and operations of language-agnostic
          models with the benefits of real-time metrics and log analysis.\n   \n##
          Overview\nSeldon Core is an open source platform for deploying machine learning
          models on a Kubernetes cluster.\n\n* Deploy machine learning models in the
          cloud or on-premise.\n* Get metrics and ensure proper governance and compliance
          for your running machine learning models.\n* Create powerful inference graphs
          made up of multiple components.\n* Provide a consistent serving layer for
          models built using heterogeneous ML toolkits.\n\nYou can get started by
          following the guides in our documentation at https://docs.seldon.io/projects/seldon-core/en/latest/workflow/README.html\n"
        displayName: Seldon Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - aiops
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://docs.seldon.io/projects/seldon-core/en/latest/
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Seldon Technologies
        relatedImages:
        - docker.io/seldonio/seldon-core-operator:1.14.1
        - ""
        version: 1.14.1
      entries:
      - name: seldon-operator.v1.14.1
        version: 1.14.1
      name: stable
    defaultChannel: stable
    packageName: seldon-operator
    provider:
      name: Seldon Technologies
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: keycloak-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keycloak-operator.v19.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "False"
          containerImage: quay.io/keycloak/keycloak-operator:19.0.3-legacy
          createdAt: "2022-10-07 17:00:00"
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
          * Install Extensions
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/docs/latest/server_installation/index.html#_operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@lists.jboss.org
          name: Keycloak DEV mailing list
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/keycloak/keycloak-operator:19.0.3-legacy
        version: 19.0.3
      entries:
      - name: keycloak-operator.v19.0.3
        version: 19.0.3
      - name: keycloak-operator.v19.0.2
        version: 19.0.2
      - name: keycloak-operator.v19.0.1
        version: 19.0.1
      - name: keycloak-operator.v19.0.0
        version: 19.0.0
      - name: keycloak-operator.v18.0.2
        version: 18.0.2
      - name: keycloak-operator.v18.0.1
        version: 18.0.1
      - name: keycloak-operator.v18.0.0
        version: 18.0.0
      - name: keycloak-operator.v17.0.1
        version: 17.0.1
      - name: keycloak-operator.v17.0.0
        version: 17.0.0
      - name: keycloak-operator.v16.1.0
        version: 16.1.0
      - name: keycloak-operator.v16.0.0
        version: 16.0.0
      - name: keycloak-operator.v15.1.1
        version: 15.1.1
      - name: keycloak-operator.v15.1.0
        version: 15.1.0
      - name: keycloak-operator.v15.0.2
        version: 15.0.2
      - name: keycloak-operator.v15.0.1
        version: 15.0.1
      - name: keycloak-operator.v15.0.0
        version: 15.0.0
      - name: keycloak-operator.v14.0.0
        version: 14.0.0
      name: alpha
    - currentCSV: keycloak-operator.v20.0.0-alpha.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "False"
          containerImage: quay.io/keycloak/keycloak-operator:19.0.3
          createdAt: 10/07/22 17:00:00
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/docs/latest/server_installation/index.html#_operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@lists.jboss.org
          name: Keycloak DEV mailing list
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/keycloak/keycloak-operator:19.0.3
        version: 20.0.0-alpha.7
      entries:
      - name: keycloak-operator.v20.0.0-alpha.7
        version: 20.0.0-alpha.7
      - name: keycloak-operator.v20.0.0-alpha.6
        version: 20.0.0-alpha.6
      - name: keycloak-operator.v20.0.0-alpha.5
        version: 20.0.0-alpha.5
      - name: keycloak-operator.v20.0.0-alpha.4
        version: 20.0.0-alpha.4
      - name: keycloak-operator.v20.0.0-alpha.3
        version: 20.0.0-alpha.3
      - name: keycloak-operator.v20.0.0-alpha.2
        version: 20.0.0-alpha.2
      - name: keycloak-operator.v20.0.0-alpha.1
        version: 20.0.0-alpha.1
      name: candidate
    - currentCSV: keycloak-operator.v26.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/keycloak/keycloak-operator:26.4.0
          createdAt: "2025-09-30T11:58:06Z"
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/keycloak/keycloak-operator:26.4.0
        version: 26.4.0
      entries:
      - name: keycloak-operator.v26.4.0
        version: 26.4.0
      - name: keycloak-operator.v26.3.5
        version: 26.3.5
      - name: keycloak-operator.v26.3.4
        version: 26.3.4
      - name: keycloak-operator.v26.3.3
        version: 26.3.3
      - name: keycloak-operator.v26.3.2
        version: 26.3.2
      - name: keycloak-operator.v26.3.1
        version: 26.3.1
      - name: keycloak-operator.v26.3.0
        version: 26.3.0
      - name: keycloak-operator.v26.2.5
        version: 26.2.5
      - name: keycloak-operator.v26.2.4
        version: 26.2.4
      - name: keycloak-operator.v26.2.3
        version: 26.2.3
      - name: keycloak-operator.v26.2.2
        version: 26.2.2
      - name: keycloak-operator.v26.2.1
        version: 26.2.1
      - name: keycloak-operator.v26.2.0
        version: 26.2.0
      - name: keycloak-operator.v26.1.5
        version: 26.1.5
      - name: keycloak-operator.v26.1.4
        version: 26.1.4
      - name: keycloak-operator.v26.1.3
        version: 26.1.3
      - name: keycloak-operator.v26.1.2
        version: 26.1.2
      - name: keycloak-operator.v26.1.1
        version: 26.1.1
      - name: keycloak-operator.v26.1.0
        version: 26.1.0
      - name: keycloak-operator.v26.0.8
        version: 26.0.8
      - name: keycloak-operator.v26.0.7
        version: 26.0.7
      - name: keycloak-operator.v26.0.6
        version: 26.0.6
      - name: keycloak-operator.v26.0.5
        version: 26.0.5
      - name: keycloak-operator.v26.0.4
        version: 26.0.4
      - name: keycloak-operator.v26.0.2
        version: 26.0.2
      - name: keycloak-operator.v26.0.1
        version: 26.0.1
      - name: keycloak-operator.v26.0.0
        version: 26.0.0
      - name: keycloak-operator.v25.0.6
        version: 25.0.6
      - name: keycloak-operator.v25.0.5
        version: 25.0.5
      - name: keycloak-operator.v25.0.4
        version: 25.0.4
      - name: keycloak-operator.v25.0.2
        version: 25.0.2
      - name: keycloak-operator.v25.0.1
        version: 25.0.1
      - name: keycloak-operator.v25.0.0
        version: 25.0.0
      - name: keycloak-operator.v24.0.5
        version: 24.0.5
      - name: keycloak-operator.v24.0.4
        version: 24.0.4
      - name: keycloak-operator.v24.0.3
        version: 24.0.3
      - name: keycloak-operator.v24.0.2
        version: 24.0.2
      - name: keycloak-operator.v24.0.1
        version: 24.0.1
      - name: keycloak-operator.v24.0.0
        version: 24.0.0
      - name: keycloak-operator.v23.0.7
        version: 23.0.7
      - name: keycloak-operator.v23.0.6
        version: 23.0.6
      - name: keycloak-operator.v23.0.5
        version: 23.0.5
      - name: keycloak-operator.v23.0.4
        version: 23.0.4
      - name: keycloak-operator.v23.0.3
        version: 23.0.3
      - name: keycloak-operator.v23.0.2
        version: 23.0.2
      - name: keycloak-operator.v23.0.1
        version: 23.0.1
      - name: keycloak-operator.v23.0.0
        version: 23.0.0
      - name: keycloak-operator.v22.0.5
        version: 22.0.5
      - name: keycloak-operator.v22.0.4
        version: 22.0.4
      - name: keycloak-operator.v22.0.3
        version: 22.0.3
      - name: keycloak-operator.v22.0.2
        version: 22.0.2
      - name: keycloak-operator.v22.0.1
        version: 22.0.1
      - name: keycloak-operator.v22.0.0
        version: 22.0.0
      - name: keycloak-operator.v21.1.2
        version: 21.1.2
      - name: keycloak-operator.v21.1.1
        version: 21.1.1
      - name: keycloak-operator.v21.1.0
        version: 21.1.0
      - name: keycloak-operator.v21.0.2
        version: 21.0.2
      - name: keycloak-operator.v21.0.1
        version: 21.0.1
      - name: keycloak-operator.v21.0.0
        version: 21.0.0
      - name: keycloak-operator.v20.0.3
        version: 20.0.3
      - name: keycloak-operator.v20.0.2
        version: 20.0.2
      - name: keycloak-operator.v20.0.1
        version: 20.0.1
      - name: keycloak-operator.v20.0.0
        version: 20.0.0
      name: fast
    defaultChannel: fast
    packageName: keycloak-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: redis-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"Redis","metadata":{"name":"redis"},"spec":{"mode":"cluster","size":3,"global":{"image":"quay.io/opstree/redis:v6.2","imagePullPolicy":"IfNotPresent","password":"Opstree@1234","resources":{"requests":{"cpu":"100m","memory":"128Mi"},"limits":{"cpu":"100m","memory":"128Mi"}}},"master":{"service":{"type":"ClusterIP"},"redisConfig":{}},"slave":{"service":{"type":"ClusterIP"},"redisConfig":{}},"service":{"type":"ClusterIP"},"redisConfig":{},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:1.0","imagePullPolicy":"Always","resources":{"requests":{"cpu":"100m","memory":"128Mi"},"limits":{"cpu":"100m","memory":"128Mi"}}},"storage":{"volumeClaimTemplate":{"spec":{"storageClassName":"buildpiper-storage","accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}}}}]'
          capabilities: Seamless Upgrades
          categories: Database, Storage
          certified: "false"
          containerImage: quay.io/opstree/redis-operator:v0.4.0
          createdAt: 06-02-2021
          description: A Golang based redis operator that will make/oversee Redis
            standalone/cluster mode setup on top of the Kubernetes.
          repository: https://github.com/ot-container-kit/redis-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Redis
            displayName: Redis
            kind: Redis
            name: redis.redis.redis.opstreelabs.in
            version: v1beta1
        description: |+
          A Golang based redis operator that will make/oversee Redis standalone/cluster mode setup on top of the Kubernetes. It can create a redis cluster setup with best practices on Cloud as well as the Bare metal environment. Also, it provides an in-built monitoring capability using redis-exporter.

          ### Supported Features

          Here the features which are supported by this operator:-
          - Redis cluster/standalone mode setup
          - Inbuilt monitoring with prometheus exporter
          - Dynamic storage provisioning with pvc template
          - Resources restrictions with k8s requests and limits
          - Password/Password-less setup
          - Node selector and affinity
          - Priority class to manage setup priority
          - SecurityContext to manipulate kernel parameters
          ### Documentation

          [Documentation](https://docs.opstreelabs.in/redis-operator/)

          ### Getting Help

          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.

          ### Contributing

          You can contribute to this project by:-
          - Raising Bugs and Feature Requests
          - Fixing issue by opening a Pull Request
          - Improving Documentation

          ### License

          Logging Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/redis-operator/blob/master/LICENSE)


        displayName: Redis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - Redis
        - Opstree
        - CRD
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://docs.opstreelabs.in/redis-operator/
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: beta
        minKubeVersion: 1.11.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/opstree/redis-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: redis-operator.v0.4.0
        version: 0.4.0
      - name: redis-operator.v0.3.0
        version: 0.3.0
      name: preview
    - currentCSV: redis-operator.v0.15.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisCluster","metadata":{"name":"redis-cluster"},"spec":{"clusterSize":3,"clusterVersion":"v7","securityContext":{"runAsUser":1000,"fsGroup":1000},"persistenceEnabled":true,"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"Redis","metadata":{"name":"redis-standalone"},"spec":{"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"IfNotPresent"},"securityContext":{"runAsUser":1000,"fsGroup":1000}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisSentinel","metadata":{"name":"redis-sentinel"},"spec":{"clusterSize":3,"podSecurityContext":{"runAsUser":1000,"fsGroup":1000},"redisSentinelConfig":{"redisReplicationName":"redis-replication"},"kubernetesConfig":{"image":"quay.io/opstree/redis-sentinel:v7.0.7","imagePullPolicy":"IfNotPresent","resources":{"requests":{"cpu":"101m","memory":"128Mi"},"limits":{"cpu":"101m","memory":"128Mi"}}}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisReplication","metadata":{"name":"redis-replication"},"spec":{"clusterSize":3,"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"Always"},"podSecurityContext":{"runAsUser":1000,"fsGroup":1000}}}]'
          capabilities: Seamless Upgrades
          categories: Database, Storage
          certified: "false"
          containerImage: quay.io/opstree/redis-operator:v0.15.1
          createdAt: 24-09-2023
          description: A Golang based redis operator that will make/oversee Redis
            standalone/cluster mode setup on top of the Kubernetes.
          repository: https://github.com/ot-container-kit/redis-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Redis Cluster
            displayName: RedisCluster
            kind: RedisCluster
            name: redisclusters.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis Replication
            displayName: RedisReplication
            kind: RedisReplication
            name: redisreplications.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis Sentinel
            displayName: RedisSentinel
            kind: RedisSentinel
            name: redissentinels.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis
            displayName: Redis
            kind: Redis
            name: redis.redis.redis.opstreelabs.in
            version: v1beta1
        description: "A Golang based redis operator that will make/oversee Redis\nstandalone/cluster
          mode setup on top of the Kubernetes. It can create a\nredis cluster setup
          with best practices on Cloud as well as the Bare metal\nenvironment. Also,
          it provides an in-built monitoring capability using\nredis-exporter.\n\nOrganizations
          that are using Redis Operator to manage their redis workload \ncan be found
          here. If your organization is also using Redis Operator, please \nfree to
          add by creating a PR https://github.com/OT-CONTAINER-KIT/redis-operator/pulls.\n\nThis
          operator only supports versions of redis =>6.\n\n### Supported Features\n\nHere
          the features which are supported by this operator:-\n- Redis cluster and
          standalone mode setup\n- Redis cluster failover and recovery\n- Redis leader
          and follower setup with Sentinel for High Availability\n- Inbuilt monitoring
          with redis exporter\n- Password and password-less setup of redis\n- TLS
          support for additional security layer\n- Ipv4 and Ipv6 support for redis
          setup\n- External configuration support for leader and follower\n- Detailed
          monitoring grafana dashboard\n\n### Documentation\n\n[Documentation](https://ot-redis-operator.netlify.app/)\n\n###
          Getting Help\nThis project is managed by OpsTree Solutions. If you have
          any queries or\nsuggestions, mail us at opensource@opstree.com.\n\n### Contributing\n\nYou
          can contribute to this project by:-\n- Raising Bugs and Feature Requests\n-
          Fixing issue by opening a Pull Request\n- Improving Documentation\n\n###
          License\n\nLogging Operator is licensed under [Apache License, Version\n2.0](https://github.com/OT-CONTAINER-KIT/redis-operator/blob/master/LICENSE)\n"
        displayName: Redis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - Redis
        - Opstree
        - CRD
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-redis-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/opstree/redis-operator:v0.15.1
        version: 0.15.1
      entries:
      - name: redis-operator.v0.15.1
        version: 0.15.1
      - name: redis-operator.v0.15.0
        version: 0.15.0
      - name: redis-operator.v0.14.0
        version: 0.14.0
      - name: redis-operator.v0.13.0
        version: 0.13.0
      - name: redis-operator.v0.8.0
        version: 0.8.0
      - name: redis-operator.v0.6.0
        version: 0.6.0
      - name: redis-operator.v0.5.0
        version: 0.5.0
      - name: redis-operator.v0.4.0
        version: 0.4.0
      - name: redis-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: redis-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Veeam Kasten
      provider-url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
    name: k10-kasten-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k10-kasten-operator-rhmp.v8.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apik10.kasten.io/v1alpha1",
                "kind": "K10",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/reconcile-period": "2m",
                    "helm.sdk.operatorframework.io/rollback-force": "false"
                  },
                  "name": "k10"
                },
                "spec": {
                  "auth": {
                    "basicAuth": {
                      "enabled": false,
                      "htpasswd": "",
                      "secretName": ""
                    },
                    "tokenAuth": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "persistence": {
                      "catalog": {
                        "size": ""
                      },
                      "storageClass": ""
                    }
                  },
                  "metering": {
                    "mode": ""
                  },
                  "route": {
                    "enabled": false,
                    "host": "",
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "apik10.kasten.io/v1alpha1",
                "kind": "K10restore",
                "metadata": {
                  "name": "k10restore"
                },
                "spec": {
                  "pointInTime": null,
                  "profileName": "",
                  "skipResource": null,
                  "sourceClusterID": ""
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          console.openshift.io/plugins: '["veeam-kasten-console-plugin"]'
          containerImage: registry.connect.redhat.com/kasten/k10-operator@sha256:4e8ccf38e47f92d2a308587591f85fb4f15aea1e7fd5c248803764c94eaecd4a
          createdAt: Wed Sep 24 02:11:59 PM UTC 2025
          description: Kasten data management platform for OpenShift, The Market Leader
            for Kubernetes Backup and DR
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/k10-kasten-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/k10-kasten-operator-rhmp/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: kasten-io
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: contact@kasten.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              k10restore can recover the K10 platform in case of a variety of disasters such as
              the accidental deletion of K10, failure of underlying storage that K10 uses for its catalog,
              or even the accidental destruction of the Kubernetes cluster on which K10 is deployed.

              **Before creating k10restore**: [K10 Disaster Recovery documentation](https://docs.kasten.io/latest/operating/dr.html#recovering-with-the-operator)
            kind: K10restore
            name: k10restores.apik10.kasten.io
            version: v1alpha1
          - description: |-
              [Kasten's k10](https://docs.kasten.io/) is a data lifecycle management system for all your persistence enabled container-based applications.

              **Note: The complete list of values supported by the K10 spec field in the "YAML View" are documented [here](https://docs.kasten.io/latest/install/advanced.html#complete-list-of-k10-helm-options).**
            kind: K10
            name: k10s.apik10.kasten.io
            version: v1alpha1
        description: |-
          Kasten data management platform for OpenShift, The Market Leader for Kubernetes Backup and DR.

          The supported Kubernetes versions are listed here - https://docs.kasten.io/latest/operating/support.html
          Please review release notes before an upgrade - https://docs.kasten.io/latest/releasenotes.html
        displayName: Veeam Kasten (Free)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kubernetes
        - storage
        - data-management
        links:
        - name: Kasten Documentation
          url: https://docs.kasten.io
        - name: Veeam Kasten
          url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
        maintainers:
        - email: contact@kasten.io
          name: Veeam Kasten
        maturity: stable
        provider:
          name: Veeam Kasten
          url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:503d5ee0fbcd5bd3492785f4246e19cbdb92960be0d05f3c6f123a32fe0fe84d
        - registry.connect.redhat.com/kasten/bloblifecyclemanager@sha256:7b209314b96bc8ebaa77e97426bb34176d05844cc3e8d4b462cfba5c4a8d3798
        - registry.connect.redhat.com/kasten/kanister-tools@sha256:7af2e37c9027f46c565901d35a3a7da4d9c7ddd2812628b67275d4acf006b80c
        - registry.connect.redhat.com/kasten/kanister@sha256:1d69ff91cec8e7c131dfbe1ad3e282aab854d3412f5f6e3b808d25eb7b8c38bc
        - registry.connect.redhat.com/kasten/metric-sidecar@sha256:20c142258de56cf8c60d4b0afa5e72b9317023ebf4ab109fb84e0d57fae854a4
        - registry.connect.redhat.com/kasten/controllermanager@sha256:ab3b815bb12cd3c263cbfdba3939a95097450f1864431cca7380827986d704ed
        - registry.connect.redhat.com/kasten/configmap-reload@sha256:01010f2bb580eb131d63d3271efe40eadd8269b84cdb02bdea36750d1359a43f
        - registry.connect.redhat.com/kasten/aggregatedapis@sha256:db868afd5c6d4a8d2e7c8376ed6943963c4be299fca6b26931f907430763d244
        - registry.connect.redhat.com/kasten/jobs@sha256:c6358878e981baee4ac72abcc1269c206ee014af2c16584d1b41fe18e0f0e57f
        - registry.connect.redhat.com/kasten/upgrade@sha256:bf2200697d38f3ee95ba2863dfbae4d982273d2fd8450c9ffbc3a85be7d6d87d
        - registry.connect.redhat.com/kasten/prometheus@sha256:29d913bb06c3245cca5798772dba7d7d5dc9b76ed2e0700f5eb4d20416e24fa2
        - registry.connect.redhat.com/kasten/k10tools@sha256:59caa1920c76a1bc86e7733731602e3f0fc08026d9a16000bb748503a4d9cd9b
        - registry.connect.redhat.com/kasten/garbagecollector@sha256:6f2972bb255f5a2de33ace5ce805b15d5e04d3c259c9c8d018c1257c74562878
        - registry.connect.redhat.com/kasten/events@sha256:7cae1c0909ea7a468e96484d99300677bdad49f34cc57c412ebaf342cb176264
        - registry.connect.redhat.com/kasten/logging@sha256:014bfdfc3620cc2fe6540b9993b9b9e7d70c9b969714477db4de91a90bdb4a38
        - registry.connect.redhat.com/kasten/k10-operator@sha256:4e8ccf38e47f92d2a308587591f85fb4f15aea1e7fd5c248803764c94eaecd4a
        - registry.connect.redhat.com/kasten/auth@sha256:f8543e67de3d5276c597c26f16053bf9bbcbd9cbdad4652214935a1e40a26817
        - registry.connect.redhat.com/kasten/frontend@sha256:a021964ed85133897e4956014b5c5013da3332a36a7e96a0de2e2e1b67888491
        - registry.connect.redhat.com/kasten/init@sha256:cc7fbbda1b6946e93dd083fb7092f38789f626c26ac68c226b61fe57064a8f11
        - registry.connect.redhat.com/kasten/dex@sha256:4c8522a27c0d4f0977baa074800bba96a9bd9aefc57473de283b1d95586f7517
        - registry.connect.redhat.com/kasten/catalog@sha256:11a77ac968634850c72cd10f2bde55afc632c68cdc0881aeae83c100288c6aec
        - registry.connect.redhat.com/kasten/dashboardbff@sha256:e1208c60146faa8c3c7ef73b2290663840276ed8d3b3c990f495cee2967104ee
        - registry.connect.redhat.com/kasten/state@sha256:191383c76c4ff22805c06217e9f32a2f41a1206afab5f16acec4a5dd56e4a6b4
        - registry.connect.redhat.com/kasten/repositories@sha256:924a607874e1807a877cb454e00539d64f4487e7435d4bfb524f3e569ad1f942
        - registry.connect.redhat.com/kasten/metering@sha256:f809696ac1309fc54fde87e5d2030267d266386fb73a56287868bbac433756d8
        - registry.connect.redhat.com/kasten/restorectl@sha256:9dc3c834bd3d4543115ae1ad586ee1248fc2b637dbc2d1e2c66ecc2a459db67e
        - registry.connect.redhat.com/kasten/datamover@sha256:c9e2dc5b646aa0c465a7b2daf011462273a6f5b53fdff772d879293208e2d57c
        - registry.connect.redhat.com/kasten/vbrintegrationapi@sha256:1a20b0da4efb7e802516224675b37751bd56d406e57b16216388d2b5167c648d
        - registry.connect.redhat.com/kasten/executor@sha256:32c4ac1aa2ca802ff77e913991949e6841149963863ad527a67f6bab6c94a5af
        - registry.connect.redhat.com/kasten/gateway@sha256:dc2d7bdeb8abaf9454470afd53af2cd5430505351a97b6501ff65a60dd2a6db1
        - registry.connect.redhat.com/kasten/ocpconsoleplugin@sha256:69f6bfb9421f4264dd7d066d809311022c0db57c41319a6242532f6acfbd2145
        - registry.connect.redhat.com/kasten/crypto@sha256:86bfbbf4c4f06bab1e4b091f9f312d536813e2c11aa899690b3277ff73110a09
        version: 8.0.9
      entries:
      - name: k10-kasten-operator-rhmp.v8.0.9
        version: 8.0.9
      - name: k10-kasten-operator-rhmp.v8.0.8
        version: 8.0.8
      - name: k10-kasten-operator-rhmp.v8.0.7
        version: 8.0.7
      - name: k10-kasten-operator-rhmp.v8.0.6
        version: 8.0.6
      - name: k10-kasten-operator-rhmp.v8.0.5
        version: 8.0.5
      - name: k10-kasten-operator-rhmp.v8.0.4
        version: 8.0.4
      - name: k10-kasten-operator-rhmp.v8.0.3
        version: 8.0.3
      - name: k10-kasten-operator-rhmp.v8.0.2
        version: 8.0.2
      - name: k10-kasten-operator-rhmp.v8.0.1
        version: 8.0.1
      - name: k10-kasten-operator-rhmp.v8.0.0
        version: 8.0.0
      - name: k10-kasten-operator-rhmp.v7.5.10
        version: 7.5.10
      - name: k10-kasten-operator-rhmp.v7.5.9
        version: 7.5.9
      - name: k10-kasten-operator-rhmp.v7.5.8
        version: 7.5.8
      - name: k10-kasten-operator-rhmp.v7.5.7
        version: 7.5.7
      - name: k10-kasten-operator-rhmp.v7.5.4
        version: 7.5.4
      - name: k10-kasten-operator-rhmp.v7.5.3
        version: 7.5.3
      - name: k10-kasten-operator-rhmp.v7.5.2
        version: 7.5.2
      - name: k10-kasten-operator-rhmp.v7.5.1
        version: 7.5.1
      - name: k10-kasten-operator-rhmp.v7.5.0
        version: 7.5.0
      - name: k10-kasten-operator-rhmp.v7.0.14
        version: 7.0.14
      - name: k10-kasten-operator-rhmp.v7.0.13
        version: 7.0.13
      - name: k10-kasten-operator-rhmp.v7.0.12
        version: 7.0.12
      - name: k10-kasten-operator-rhmp.v7.0.11
        version: 7.0.11
      - name: k10-kasten-operator-rhmp.v7.0.9
        version: 7.0.9
      - name: k10-kasten-operator-rhmp.v7.0.8
        version: 7.0.8
      - name: k10-kasten-operator-rhmp.v7.0.7
        version: 7.0.7
      - name: k10-kasten-operator-rhmp.v7.0.6
        version: 7.0.6
      - name: k10-kasten-operator-rhmp.v7.0.5
        version: 7.0.5
      - name: k10-kasten-operator-rhmp.v7.0.4
        version: 7.0.4
      - name: k10-kasten-operator-rhmp.v7.0.3
        version: 7.0.3
      - name: k10-kasten-operator-rhmp.v7.0.1
        version: 7.0.1
      - name: k10-kasten-operator-rhmp.v6.5.14
        version: 6.5.14
      - name: k10-kasten-operator-rhmp.v6.5.13
        version: 6.5.13
      - name: k10-kasten-operator-rhmp.v6.5.12
        version: 6.5.12
      - name: k10-kasten-operator-rhmp.v6.5.11
        version: 6.5.11
      - name: k10-kasten-operator-rhmp.v6.5.10
        version: 6.5.10
      - name: k10-kasten-operator-rhmp.v6.5.8
        version: 6.5.8
      - name: k10-kasten-operator-rhmp.v6.5.6
        version: 6.5.6
      - name: k10-kasten-operator-rhmp.v6.5.5
        version: 6.5.5
      - name: k10-kasten-operator-rhmp.v6.5.4
        version: 6.5.4
      - name: k10-kasten-operator-rhmp.v6.5.3
        version: 6.5.3
      - name: k10-kasten-operator-rhmp.v6.5.1
        version: 6.5.1
      - name: k10-kasten-operator-rhmp.v6.5.0
        version: 6.5.0
      - name: k10-kasten-operator-rhmp.v6.0.12
        version: 6.0.12
      - name: k10-kasten-operator-rhmp.v6.0.11
        version: 6.0.11
      - name: k10-kasten-operator-rhmp.v6.0.9
        version: 6.0.9
      - name: k10-kasten-operator-rhmp.v6.0.8
        version: 6.0.8
      - name: k10-kasten-operator-rhmp.v6.0.7
        version: 6.0.7
      - name: k10-kasten-operator-rhmp.v6.0.6
        version: 6.0.6
      - name: k10-kasten-operator-rhmp.v6.0.5
        version: 6.0.5
      - name: k10-kasten-operator-rhmp.v6.0.4
        version: 6.0.4
      - name: k10-kasten-operator-rhmp.v6.0.3
        version: 6.0.3
      - name: k10-kasten-operator-rhmp.v6.0.2
        version: 6.0.2
      - name: k10-kasten-operator-rhmp.v6.0.1
        version: 6.0.1
      - name: k10-kasten-operator-rhmp.v6.0.0
        version: 6.0.0
      name: stable
    defaultChannel: stable
    packageName: k10-kasten-operator-rhmp
    provider:
      name: Veeam Kasten
      url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SusQL Operator Contributors
      provider-url: https://github.com/sustainable-computing-io/susql-operator
    name: susql-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: susql-operator.v0.0.35
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "susql.ibm.com/v1",
                "kind": "LabelGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "susql-operator",
                    "app.kubernetes.io/instance": "labelgroup-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "susql-operator",
                    "app.kubernetes.io/part-of": "susql-operator"
                  },
                  "name": "labelgroup-sample"
                },
                "spec": {
                  "labels": [
                    "labelgroup-sample-1",
                    "labelgroup-sample-2"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: quay.io/sustainable_computing_io/susql_operator:0.0.34
          createdAt: "2025-07-10T01:20:37Z"
          description: 'Aggregates energy and CO2 emission data for pods tagged with
            SusQL labels '
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/sustainable-computing-io/susql-operator
          support: https://github.com/sustainable-computing-io/susql-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LabelGroup is the Schema for the LabelGroups API
            displayName: Label Group
            kind: LabelGroup
            name: labelgroups.susql.ibm.com
            version: v1
        description: |-
          ### About this Operator

          SusQL is a Kubernetes operator that aggregates energy and estimated
          carbon dioxide emission data for pods tagged with SusQL specific
          labels. The energy measurements are taken from Kepler which should
          be deployed in the cluster before using SusQL.

          ### Prerequisites

          1. Deployment of [Kepler](https://sustainable-computing.io/) on the cluster
          2. Ensure that [User Project Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          is enabled to monitor energy consumed in user projects.
          3. Creation of a SusQL Service Monitor:
                `oc apply -n <NAMESPACE> -f https://raw.githubusercontent.com/sustainable-computing-io/susql-operator/main/hack/susql-servicemonitor.yaml`

          ### API Backward Compatibility

          Until Operator supports **seamless upgrades**, a new version of the operator may
          introduce a change that is **NOT** backwards compatible with the previous version.
          Thus, to upgrade the operator, uninstall the already installed version completely
          (including CRDs) and install the new version. For example to upgrade from 0.1.0
          to 0.2.0, you must first uninstall 0.1.0.

          ### LabelGroup spec Example

          ```
          apiVersion: susql.ibm.com/v1
          kind: LabelGroup
          metadata:
            name: labelgroup-name
            namespace: susql-operator
          spec:
            labels:
              - my-label-1
              - my-label-2
          ```

          ### Documentation

          Documentation can be found at [SusQL Operator](https://github.com/sustainable-computing-io/susql-operator#readme).

          ### Getting help

          If you encounter any issues while using the SusQL Operator, you can create an
          [issue](https://github.com/sustainable-computing-io/susql-operator/issues)
          on our GitHub repo for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by raising [issues](https://github.com/sustainable-computing-io/susql-operator/issues)
          related to susql-operator.

          ### License
          susql-operator is licensed under the Apache 2.0 license
        displayName: SusQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - aggregation
        - sustainability
        - energy
        - kepler
        - susql
        - carbon dioxide (CO2) emissions
        - green computing
        links:
        - name: SusQL Operator
          url: https://github.com/sustainable-computing-io/susql-operator
        - name: Sustainable Computing
          url: https://github.com/sustainable-computing-io
        maintainers:
        - email: trent@jp.ibm.com
          name: Scott Trent
        - email: mohammed.munir.abdi@ibm.com
          name: Mohammed Abdi
        maturity: alpha
        provider:
          name: SusQL Operator Contributors
          url: https://github.com/sustainable-computing-io/susql-operator
        relatedImages:
        - quay.io/sustainable_computing_io/susql_operator:0.0.34
        version: 0.0.35
      entries:
      - name: susql-operator.v0.0.35
        version: 0.0.35
      - name: susql-operator.v0.0.34
        version: 0.0.34
      - name: susql-operator.v0.0.33
        version: 0.0.33
      - name: susql-operator.v0.0.32
        version: 0.0.32
      - name: susql-operator.v0.0.30
        version: 0.0.30
      - name: susql-operator.v0.0.29
        version: 0.0.29
      - name: susql-operator.v0.0.28
        version: 0.0.28
      - name: susql-operator.v0.0.27
        version: 0.0.27
      name: alpha
    defaultChannel: alpha
    packageName: susql-operator
    provider:
      name: SusQL Operator Contributors
      url: https://github.com/sustainable-computing-io/susql-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cockroach Labs
      provider-url: ""
    name: cockroachdb-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroach-operator.v2.18.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v25.2.2",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
          createdAt: "2025-07-22T11:30:55Z"
          description: CockroachDB Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1304d7f871a95c8b087bad1b548217562792edc66c8fe07c621e05242b8cd53e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6408f89b0a88ec154f7ee98c314be574971c4c9aaf906b473d10124d822889be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:65923c37fecda2d3174212fa9d9a6f34241d065389336ea69a7be99ba16cedc2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3bf017b2b2506724500f5e29017d1b956825dd0f61a7c0049b0ff084c3554b5b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:85a0e085901c2a73192556d5d70795d4b66cbcfe01d24c73b6fe6db8d5f2e5ca
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bd75aa918355e77cd671c3e11ee807861378e96245245f4fdf243cba5749f9a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e8f26fd132affbaeb8ef6b1807a646c83b301e5dc334a592dfb2cf8dc3dbaf10
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6ca927f137be28481e3a226e6c6bcb7c5ac27664984f17ffc6c1419cd7d08eb7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da367cf0ac52045002e1c186f8e6964267ad87d5c25f9e72fd2c9b9a98a32702
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc1ab77a9537ccf236ce6dc370940dfedcfaa91b6e3db37bf2d5eade7ff84708
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e0f21abea34825e55622c432a306f2e8aafd64d913eb78f2f601c51dc878a1da
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1b6d934a7b6c42a61e5ad51380e69af341a575b42205299fd58bc0c6ff306f6a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a6f1c43486d044fed9f59406118f429c8037b389bd8c8d7a1c7aeb89ed2661f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b06bd7f3802e8fa2af14bb0b07804cad4cc0fbbc97da18d09a9106bf074ed6d8
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:740ba10f1ef0a6c5fa6ad31ab18855fc75f8e024ce4c23b0676c5a9c5b62cd75
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c9a31bdf19175880d0981e21dd19a1458b6e8cde8d4acbcb554f3bea3578a64e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf7d3bea47638b763b312c7c11ca034a7ac62d993d75a13edceec74799b97e52
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f438fbbf9c87ca0d54c170e2c5105b1e265182e1c2ede48adea370e2be9dc895
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9ab968e4ca62f23bf9848e5475ef4076df7318a8560f63d5b2090a32af6fe4e2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d978678b3c254abd52f69f226b0aacc4b02c2aaca328c54ef10f6f9bb453582d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:dd521d7bad62da054eec54179be75c643a5492d981e11c0e66db41902cd2182c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d0014b1c19890b9e8ef1a4f3efe7eb5b269be5610ba6ababd4258b52d2f85e8a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5a54d564e08baefc2786f4b742b2507d31984664cd00333406aa7a3aa5b9c913
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5181196fff3bf8a5ec8fcfe4f1958cc69f285f0beb4f0b476b4b5e0912f3db42
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9f42e929f2cc7e9b6d24dd61e19c5d93dcfeabd34b9c8b935afbe78577efb18b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:266165ab87b19a05d6796e8b15ff0eba632d6ac96e6393400464c6136dd500ec
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:46d9287ef11e3b21a36abfa31a26da3b93040bb6163b9223972a5dd1930fecff
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acbfee4492327a720fe7da76929a86192d8af2442bbb76f0a089f6a81a890be3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6371377718325235a2720dba0c97eb4321ee29b31a28cb893f76abd62baa339
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:79d688bebe0643ef0dd6ba7108cdb34c0442470c0c20ae0df88b5d02a3476cd0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e23af8f9d32c3f80a52d79ceb4b359dda1d75ba0dd4cb8ed4b66ac2f4e40c69b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:58de9d0b2ce005c7ad3e99152100e59b7677398557028a7a59e56e669fabb81c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7daaf1368ab2c25694cdea0a598e1de59247b04a4b4b4b8348ed6edbaa194f9d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:35ede65457d845e8c063ad7d9515d3a69f98e0751424ecbac484963678a0b1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f42fa22aac3cc41ce72e144a72530ffbee765a392b24593dafd06cc098f8410b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a4f932c3e9ff3aeb70ba1f7a92aa4642bd14cbb7d37d04ff750ed517b916cdb9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:19f36f53f7da67755eb86da77098c5eeb84b381f9c18aa29ba573793dc498564
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4635550222775fc75737be5e5c0359303e3a23c4d1e1c296d7408cd15ad94aad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f3ffebd812881a830ff46c860984814dd61edcd51b478ebd3667759fb7710682
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2a128b570352a5a7f82b27353f9b774caccce83e0e7bb2673e4802d96ad0bd51
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f74c8a5dea4560ad59c6f6f6ab1e699f063334ee8db704460f870508dbeaeff0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bd336715332f776a72aaf5c3ca52e62dd53a9e8521c1feb6e135cc868d694a39
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5240f9f591263569fad7ee934e81fa8a1617de3c25da94ddd59ee5b2b3407f90
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d14f77f70aea0422a906510c1ac2652ff76e131ed705cb002f80e13e887b0cd9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:10b9df672260c023dfb37166e156c0d7e0e13144dc3e485af922e8151fd3ab05
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2af60025ecb3bb933b61328be4c2b2bfd0e7d26f53b72430629208505e45c6d2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:32fdca575c334822e4356aab36a7ed97b685c065925fe85f1b8ba8425c57159b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:42c9930d6861f6a2147ed6e01827eddd730279966763ec68868567a19cc164ab
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4b3aa590623a38f435b075de4c09f8a343d8ece8b673e6bf7c6df5d5eb4bdf4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:02e1ef01a023da0eee653f63b314786fd66a78d184a97a314422f6f9c3ef8e7e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:310310515625f099a928545865f7096997871ee4a16650a01c76c3799a18b684
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6a2ee8d62c586eb74ebbef6bef256e06e1ba5cb51c1ae55b1654adafc2ca8859
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acc083f7dde1fbc557d858d03fc26b68b2056997fc1a457e4695650214d5b303
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b78a4d3864e16d270979d9d7756012abf09e6cdb7f1eb4832f6497c26281099
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:61e602b0e3582bf00460eb9328808ccf66e3ff1b1f6df07ee6c5c76a91846876
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2d89398f03736f8ffaa82b10fed0f19fba166d3d2145b26f25cb71f3c39acf07
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5458a8fd89aa664b60e59c35bd933975cd37d18e19b11589e4ca0c2844ea0ed4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:50393ad005fed19618d04f0716c6d55e2f1afce651e785c396638e852cac83b3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ec6eb7c28c213cc83b2d7919cd87988f9a07f12276eb7351d0915f1567a5b095
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6366ba29f301a82215c05e60d78e1de66bcee85440ce2fc3c692712d4da32685
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5d6901b6b6fd6594d42bfc1842d050c7cffcb0867a5591a67469f0f860b1f21b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4d8e804b35b73d7978c112748aea2d2bb68f65081b90bca62c39f2b6e2989378
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:743cda1eda6e273c798e3405a9a97e07dd6aa5ad4d5a52789be3b0f39a32a039
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4f753151af4c3d3f5a1df9014461316249a6d0621b8fba5c4430307b3355251
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce1209ca6cc8724baee981c97c66e2852e512423bc0c0a16cdd80a6838b54031
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f5a0ccc02dc9e938e484d5b5282ff650d1890d5f754c30a0c0d717989ed5d600
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f045fe5037d44aab1a0b2209241b3931ec7cb9897cb3c6a96a9f8d5280e55dd2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08cec123813304ab03bd66d72a09949667a874a42320bedafda391ebcdb2e56c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5b3de46526d169fcd1318354545b3a18861843fa2022a23ed04d795493825287
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d590946b2f7c7ec0c7ec057a338e81f864d6a9e3d3ac4e0769aa5756a8e13fc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0754f7901feea9017bc5041cb91b7ac1cdbca10c2064f37c3648a2c395f24f83
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db07461824e73410456ef540b0bce3c7fe550aa67a73d6174300ca74b8f75276
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c74cf927a2b2c5098ad74ad0880996f8b042186a9911d888a231ef5b9e2a0715
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7d6320ac7fb0925f10cf6f524a13de629cbf13ccc371912e4fe8c777bc15d1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:28ae9661b38a72decf413c6a6c3758e61098debf5162197c27a12a070c7a096d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:226d7f1c3f0d70ac6532ad9490bebcdef44a40959b6aa7c11b0f2d1383a46217
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:44a0f40a6c46ba08297b482f3d96defe87b181c944803e5381c9acc42cabf97f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b046aa0a42dae992a03685ed82a20835532d6d4bf048e2366ff7382be4b3138
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1ce71e80c0cd0e0063eba275ab91f77982f2e92a8aef8e89cab99b93feab0a88
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:29abb169ffefe36f6d69edd438a7135ee23059cedab8f87ad3b2d52361b96b35
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:68fe64fe3e112288ccfc2a7dffd69b9b2786a5da3d3bdedd6addf41659d460c9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8443c15eb1653dd9e5ec7aad54d4b8b092d41a0977e42b76e372e5e21280746e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b83eaff2e217c06098509709cead525b55b00ccd48019d8357e6231892008055
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bed27f3f2c2d7627d1af4500d7ea0302da5c3fd3d875a6480fb6d98da481312e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b0057de903654ea0315a2c96affcc8f50a92ede8ad121ed30edb3aaf0fb7b554
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cbc74889e9ca5edbd5aaa243d7649bca0c4da2719239f8d0ec9827868112a771
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1558a1f5eab79f1a8c6c85d81740cc539dc23025ea457c0019d93a72dc05d57a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf001173bbf00cdf443b8d64a3ad724389f1cf83e3628d0b609019c8cddad35a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:87db6e079b96fb392d98ef5fc7063836bff10717efd75061016ce688be0ba38e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:72ea3d0ef72972dba93362dc31366c5946a9898f5c16780b877316a97068ef8f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1374cbdc4ecf38e4b4226328f532e830b8932b34d9449bf2fbb36e60d0310521
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2171e1c75d0c665fe33e1ea334e605604c51de6dbc22d88a3a0ac604259c448f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db4e5c2f27ba23e95a363c8174c01f0f9c6b6a27e0e581c86a38f3a1f68dc43c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d8cfb221f67516675fdfde061f6ec5dfee1224ca7eb3d8be375c27d483f306af
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:aeac9cd8e8acdbd2b0c339c48b85079c159b04f0ade24df8e3f9cae3f48b4c70
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:086d7b435993fdf21fd59b0093f52105a9028d6b769398b0033f98a65dfb7e79
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9d0bfb642090ff55a3cc3da0f72663cae18063622a4ec785ba99bc71ef4db116
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3f587e7b9f5a611013af2608e2151ed782859a52501b00adbc95739c56a8ff1e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:82d8104a37349c513447e68ea3790efb256370632353176e5fc13c2e8df1bd44
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c2a62179e8ca293d7e30b9da599e48795345cf07a72abd3a83063107b53b03cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d467383af41aa80c26172964c34152abeba45121599804e502984655b72179f0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e5814748c5ab229a1dea5662a007063c0df06fbbfcfe041fd9a6df9ef67e07cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a6e2fe24ad52913e2160f85b9fe665a4e4c4d79b164b4b899f69dd9d92f999e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6cd9f9b9fcb6e8f9a931a6ad01cb66329a74d40b22f9029fbc50e57248bad7cb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:56109e57ee0379cf48644bcf8226a9238e01139cacc5499002c99f973f121911
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f53b07212320c6939d3bb4ddd4e6a0bce311ab5b63858eb1a8728531f0dbd510
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce3d94a3fbdcde5858beb3c265987759971d991c397c6486e19904bd4706f44d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e151bded670f4cbd165a8ef864324805667d3e84ce3ace0cdf8f9aad087fccf1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6bb15b36d64eebb6e4c1db5a5466e108b271d53383c58e0b6c78cec214756a9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0e2c7023f1f97acc20c46a4944635a9c5f57bb9bb3dd534cf41fea1a9cfdfdbe
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:91f7e59f9e35a0395756eafed906963c535e886125294d671d930ce959a8f511
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f05824bb8eaa5a93e54059449eaa236dd83d729049ad00065ef3242f5f74c91e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b77b80eed7ce149d2a3f228e5f188e76d73f32bc50ecde20bba9cfa7bdc4714c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cad60044ad6573bd80b65e892a9ec0510dbe4fcbbfc6b51010ecf419f56f1024
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8fc8de4f3671a73f9aa747e5b23ec973d192fff6f2869c38985cd656063c1e3d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:022c6e1a0de875474bc953718a481a0af526c49dc910094dcdce0b5677aced18
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:468b0a40a33be1dd5a2a89ce745d7e9012bc7371f2f01cc0f8f12c0ac887db43
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4e5f7df1dc1e1db398c36d590431e7e5782897b209972d8e9e4671971c10d1b6
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7481c0cf3b579c8d98d94d4aa0334ca94bf2ccfcd25ad70b57c42b22bbf4bcee
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08496b84b3cc691909227be17e67bcd6f4d5185a93712f44d8a0985bd2c5fcd7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fd6c9b8ca82ff4f2d0fe29166d1fc9bca6439cb40907be60f4d8f1b9735de2f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:742a6f69a3a209a7080a1e911fd96467d2a91aacb3f91d6be80a6a6e2e54b2ae
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e82684b828256b3c2814930210031b3b70bc818514d4359a25f87ffdbf22be6e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8737b67b91983817e3e4ff3f47f0b070bd36162d309b60993529426c6979cb24
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4b56c0311af14220a466aa11f25fd10fa38280825539bd02215c945bbafb861c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9097e135159f46396ef4649f724d61bb5b9113c488f989d404586fbb7486fd85
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1e5d85f5db10fa79d09263db90daa44549bcc3b025e9f3f98cf0d4e51766394b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a0faa52a999289b308dc4caf050c6d9fe2f42e112aa30e4af282ed1f21c5c346
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1f6b1ec92251397dbcf7ed4a82f1f05619d64676aae2717f4eff1c33e4c2c27a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a59cafff8d4ff2c62204ece15880117011c30978823c232902c7182d747ec827
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:858dc82977b3c9e989a3a20626671b0c67c834c97132094a9212bb227ab55929
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:744193037d6b64cb73fecab15ebe022abc2b79ab0335b76ec1415ac6379b7dfa
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d78c2e0709214f96794564f13386861f4b541e55b5bd316ad495cf104f7d40b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9b5a2e0c006eb4be8e1444734a98743551f278a24e81c20bebc96f0f3b631eb0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e50aab39722af22ccbf3d1db19e37972d52b0fc9a40998a64618dc966b2bac57
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:818f592c50d4ed5599faa053344f7ecd963003352f1faa3b72dc0f5e00a0e78b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0b0e0588c1aadf0c3826cd584bd2b3a7d6781882dcb5c5e037b47ff5cb84509b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:53c7b960a2ed70a0998a4f8e28a4f78f7f676e01cffebe665e91e3a7629d88ed
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f29c31bb267451b389608af3e8a4a28d6ca0691e67b77e8f57bb5e461b064893
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bf36dab9b3257a4bba24d986f1451fc51a3a37023fbb988281f1941fd3faedd
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7b907d03b001b758d8a4b5e898725c73dda2ec2aa4b7425b685320b947154d11
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8f1f2fd1aca914a1ef007117c85ac944055bbfb873a468a7374d239ef20b065e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:37bf9e1ec7e3eff47755fe7521223f31ec6aefc6e63901db3881b30898cf78f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:40f00476b63cb9e370fe85f01407173693213a9242fc117d28b06378ca0d98e0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8a70ba12f610ac0e62066e417a0833bc1d01e91ac99d0dd794a2c0cd2a350164
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2b5c028f293c6f394e58a5495cb3dbfbedc215d0d03778d0217e045586d55a46
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da3bd2e9ff2b08d5bb86b1f5faa95a865f85c5e8912df00b6b79190c1e518a2f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7ad12df8e258b03c4099d375d02d272bd774bf188d108901d5cf6cb865e57ef3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cc20270212f01fc231ff788e20b0476b8228c440bfc40430c0836d5eb702a042
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:31b161f9bf9028a1a62cd2eef6abda01257d8889bdc3159822642549a75e3429
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0fea6ed8a78ea624240ee6ffb3573d5cf1115186aa9180fcd7d1273351deaaa3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9c85c25d5392817d17b68714dca6195ad53255d43e9c65b021d75b12d97e0579
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a3ac544d87a9d8e7c92897c4c23f803753c7ab31467db91b5c51244afc2c6be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e18d73af3257c8cdf00245f2f718b56667524c1cc69d3db28e9b868a539ae16f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:363d8feb560324a848c5771edd224dc19020013af7641979ee81ae8ee536d7e3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fe4a40006d1e8439423de1417cc9bc55716d790b882485074bb2a5ec68d7d1ad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:90cf21c36d19d87b97d06edd1c0912a3bb288c769e6159651877db3206aa7355
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0ec1dc184a2199736b4bd7de3241a357be7427caea5da1c931b08bc06f5c3dc0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8e8c081c231b5b15eca483ceec16ea9190ba73aae6e421865162446eb584519a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc4b75ed71845d5b18e1d9163d480fd9d0d3cb5ebbcfed00a2e4e1b174c0a5de
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9492894eaa5f43d00ac80b60029b32a0b0a85ae88ba609b46f98702c9349a5bb
        version: 2.18.2
      entries:
      - name: cockroach-operator.v2.18.2
        version: 2.18.2
      name: stable
    defaultChannel: stable
    packageName: cockroachdb-certified
    provider:
      name: Cockroach Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Elastic
      provider-url: ""
    name: elasticsearch-eck-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: elasticsearch-eck-operator-certified.v3.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "elasticsearch.k8s.elastic.co/v1",
                    "kind": "Elasticsearch",
                    "metadata": {
                        "name": "elasticsearch-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "nodeSets": [
                            {
                                "name": "default",
                                "config": {
                                    "node.roles": ["master", "data"],
                                    "node.attr.attr_name": "attr_value",
                                    "node.store.allow_mmap": false
                                },
                                "podTemplate": {
                                    "metadata": {
                                        "labels": {
                                            "foo": "bar"
                                        }
                                    },
                                    "spec": {
                                        "containers": [
                                            {
                                                "name": "elasticsearch",
                                                "resources": {
                                                    "requests": {
                                                        "memory": "4Gi",
                                                        "cpu": 1
                                                    },
                                                    "limits": {
                                                        "memory": "4Gi",
                                                        "cpu": 2
                                                    }
                                                }
                                            }
                                        ]
                                    }
                                },
                                "count": 3
                            }
                        ]
                    }
                },
                {
                    "apiVersion": "kibana.k8s.elastic.co/v1",
                    "kind": "Kibana",
                    "metadata": {
                        "name": "kibana-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        },
                        "podTemplate": {
                            "metadata": {
                                "labels": {
                                    "foo": "bar"
                                }
                            },
                            "spec": {
                                "containers": [
                                    {
                                        "name": "kibana",
                                        "resources": {
                                            "requests": {
                                                "memory": "1Gi",
                                                "cpu": 0.5
                                            },
                                            "limits": {
                                                "memory": "2Gi",
                                                "cpu": 2
                                            }
                                        }
                                    }
                                ]
                            }
                        }
                    }
                },
                {
                    "apiVersion": "apm.k8s.elastic.co/v1",
                    "kind": "ApmServer",
                    "metadata": {
                        "name": "apmserver-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                    "apiVersion": "enterprisesearch.k8s.elastic.co/v1",
                    "kind": "EnterpriseSearch",
                    "metadata": {
                        "name": "ent-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "config": {
                            "ent_search.external_url": "https://localhost:3002"
                        },
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                  "apiVersion": "beat.k8s.elastic.co/v1beta1",
                  "kind": "Beat",
                  "metadata": {
                    "name": "heartbeat-sample"
                  },
                  "spec": {
                    "type": "heartbeat",
                    "version": "9.1.0",
                    "elasticsearchRef": {
                      "name": "elasticsearch-sample"
                    },
                    "config": {
                      "heartbeat.monitors": [
                        {
                          "type": "tcp",
                          "schedule": "@every 5s",
                          "hosts": [
                            "elasticsearch-sample-es-http.default.svc:9200"
                          ]
                        }
                      ]
                    },
                    "deployment": {
                      "replicas": 1,
                      "podTemplate": {
                        "spec": {
                          "securityContext": {
                            "runAsUser": 0
                          }
                        }
                      }
                    }
                  }
                },
                {
                  "apiVersion": "agent.k8s.elastic.co/v1alpha1",
                  "kind": "Agent",
                  "metadata": {
                    "name": "agent-sample"
                  },
                  "spec": {
                    "version": "9.1.0",
                    "elasticsearchRefs": [
                      {
                        "name": "elasticsearch-sample"
                      }
                    ],
                    "daemonSet": {},
                    "config": {
                      "inputs": [
                        {
                          "name": "system-1",
                          "revision": 1,
                          "type": "system/metrics",
                          "use_output": "default",
                          "meta": {
                            "package": {
                              "name": "system",
                              "version": "0.9.1"
                            }
                          },
                          "data_stream": {
                            "namespace": "default"
                          },
                          "streams": [
                            {
                              "id": "system/metrics-system.cpu",
                              "data_stream": {
                                "dataset": "system.cpu",
                                "type": "metrics"
                              },
                              "metricsets": [
                                "cpu"
                              ],
                              "cpu.metrics": [
                                "percentages",
                                "normalized_percentages"
                              ],
                              "period": "10s"
                            }
                          ]
                        }
                      ]
                    }
                  }
                },
                {
                    "apiVersion": "maps.k8s.elastic.co/v1alpha1",
                    "kind": "ElasticMapsServer",
                    "metadata": {
                        "name": "ems-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                    "apiVersion": "logstash.k8s.elastic.co/v1alpha1",
                    "kind": "Logstash",
                    "metadata" : {
                        "name": "logstash-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.connect.redhat.com/elastic/eck-operator@sha256:1750178c5dd716c031155f2e7058e19d4b7f1d3fdbb3e8a806b177a71be88500
          createdAt: "2025-07-30 11:10:34"
          description: Run Elasticsearch, Kibana, APM Server, Beats, Enterprise Search,
            Elastic Agent, Elastic Maps Server and Logstash on Kubernetes and OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Elastic Basic license
          repository: https://github.com/elastic/cloud-on-k8s
          support: elastic.co
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Elastic Agent instance
            displayName: Elastic Agent
            kind: Agent
            name: agents.agent.k8s.elastic.co
            version: v1alpha1
          - description: APM Server instance
            displayName: APM Server
            kind: ApmServer
            name: apmservers.apm.k8s.elastic.co
            version: v1
          - description: Beats instance
            displayName: Beats
            kind: Beat
            name: beats.beat.k8s.elastic.co
            version: v1beta1
          - description: Elastic Maps Server instance
            displayName: Elastic Maps Server
            kind: ElasticMapsServer
            name: elasticmapsservers.maps.k8s.elastic.co
            version: v1alpha1
          - description: Instance of an Elasticsearch autoscaler
            displayName: Elasticsearch Autoscaler
            kind: ElasticsearchAutoscaler
            name: elasticsearchautoscalers.autoscaling.k8s.elastic.co
            version: v1alpha1
          - description: Instance of an Elasticsearch cluster
            displayName: Elasticsearch Cluster
            kind: Elasticsearch
            name: elasticsearches.elasticsearch.k8s.elastic.co
            version: v1
          - description: Enterprise Search instance
            displayName: Enterprise Search
            kind: EnterpriseSearch
            name: enterprisesearches.enterprisesearch.k8s.elastic.co
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.kibana.k8s.elastic.co
            version: v1
          - description: Logstash instance
            displayName: Logstash
            kind: Logstash
            name: logstashes.logstash.k8s.elastic.co
            version: v1alpha1
          - description: Elastic Stack Config Policy
            displayName: Elastic Stack Config Policy
            kind: StackConfigPolicy
            name: stackconfigpolicies.stackconfigpolicy.k8s.elastic.co
            version: v1alpha1
        description: |-
          Elastic Cloud on Kubernetes (ECK) is the official operator by Elastic for automating the deployment, provisioning, management, and orchestration of Elasticsearch, Kibana, APM Server, Beats, Enterprise Search, Elastic Agent, Elastic Maps Server, and Logstash on Kubernetes.

          Current features:

          *  Elasticsearch, Kibana, APM Server, Enterprise Search, Beats, Elastic Agent, Elastic Maps Server, and Logstash deployments
          *  TLS Certificates management
          *  Safe Elasticsearch cluster configuration and topology changes
          *  Persistent volumes usage
          *  Custom node configuration and attributes
          *  Secure settings keystore updates

          Supported versions:

          * Kubernetes 1.29-1.33
          * OpenShift 4.15-4.19
          * Google Kubernetes Engine (GKE), Azure Kubernetes Service (AKS), and Amazon Elastic Kubernetes Service (EKS)
          * Elasticsearch, Kibana, APM Server: 7.17+, 8+, 9+
          * Enterprise Search: 7.17+, 8+
          * Beats: 7.17+, 8+, 9+
          * Elastic Agent: 7.17+, 8+, 9+
          * Elastic Maps Server: 7.17+, 8+, 9+
          * Logstash 8.12+, 9+

          ECK should work with all conformant installers as listed in these [FAQs](https://github.com/cncf/k8s-conformance/blob/master/faq.md#what-is-a-distribution-hosted-platform-and-an-installer). Distributions include source patches and so may not work as-is with ECK.
          Alpha, beta, and stable API versions follow the same [conventions used by Kubernetes](https://kubernetes.io/docs/concepts/overview/kubernetes-api/#api-versioning).
          See the full [Elastic support matrix](https://www.elastic.co/support/matrix#matrix_kubernetes) for more information.
          See the [Quickstart](https://www.elastic.co/guide/en/cloud-on-k8s/3.1/k8s-quickstart.html) to get started with ECK.
        displayName: Elasticsearch (ECK) Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - analytics
        - search
        - database
        - apm
        - logstash
        links:
        - name: Documentation
          url: https://www.elastic.co/guide/en/cloud-on-k8s/3.1/index.html
        maintainers:
        - email: eck@elastic.co
          name: Elastic
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Elastic
        relatedImages:
        - registry.connect.redhat.com/elastic/eck-operator@sha256:1750178c5dd716c031155f2e7058e19d4b7f1d3fdbb3e8a806b177a71be88500
        version: 3.1.0
      entries:
      - name: elasticsearch-eck-operator-certified.v3.1.0
        version: 3.1.0
      - name: elasticsearch-eck-operator-certified.v3.0.0
        version: 3.0.0
      - name: elasticsearch-eck-operator-certified.v2.16.1
        version: 2.16.1
      - name: elasticsearch-eck-operator-certified.v2.16.0
        version: 2.16.0
      - name: elasticsearch-eck-operator-certified.v2.15.0
        version: 2.15.0
      - name: elasticsearch-eck-operator-certified.v2.14.0
        version: 2.14.0
      - name: elasticsearch-eck-operator-certified.v2.13.0
        version: 2.13.0
      - name: elasticsearch-eck-operator-certified.v2.12.1
        version: 2.12.1
      - name: elasticsearch-eck-operator-certified.v2.12.0
        version: 2.12.0
      - name: elasticsearch-eck-operator-certified.v2.11.1
        version: 2.11.1
      - name: elasticsearch-eck-operator-certified.v2.11.0
        version: 2.11.0
      - name: elasticsearch-eck-operator-certified.v2.10.0
        version: 2.10.0
      - name: elasticsearch-eck-operator-certified.v2.9.0
        version: 2.9.0
      - name: elasticsearch-eck-operator-certified.v2.8.0
        version: 2.8.0
      - name: elasticsearch-eck-operator-certified.v2.7.0
        version: 2.7.0
      - name: elasticsearch-eck-operator-certified.v2.6.2
        version: 2.6.2
      - name: elasticsearch-eck-operator-certified.v2.6.1
        version: 2.6.1
      - name: elasticsearch-eck-operator-certified.v2.5.0
        version: 2.5.0
      - name: elasticsearch-eck-operator-certified.v2.4.0
        version: 2.4.0
      - name: elasticsearch-eck-operator-certified.v2.3.0
        version: 2.3.0
      - name: elasticsearch-eck-operator-certified.v2.2.0
        version: 2.2.0
      - name: elasticsearch-eck-operator-certified.v2.1.0
        version: 2.1.0
      - name: elasticsearch-eck-operator-certified.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: elasticsearch-eck-operator-certified
    provider:
      name: Elastic
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Portworx by PureStorage
      provider-url: ""
    name: portworx-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: portworx-operator.v25.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageCluster",
                "metadata": {
                  "name": "portworx",
                  "namespace": "test-operator",
                  "annotations": {
                    "portworx.io/is-openshift": "true"
                  }
                },
                "spec": {}
              },
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageNode",
                "metadata": {
                  "name": "example",
                  "namespace": "test-operator"
                },
                "spec": {}
              }
            ]
          capabilities: Auto Pilot
          categories: Storage
          certified: "true"
          console.openshift.io/plugins: '["portworx"]'
          containerImage: registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
          createdAt: "2025-09-06T19:38:43Z"
          description: Cloud native storage solution for production workloads
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "core.libopenstorage.org/v1",
              "kind": "StorageCluster",
              "metadata": {
                "name": "portworx",
                "annotations": {
                  "portworx.io/is-openshift": "true"
                }
              }
            }
          support: Portworx, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Storage Cluster installs Portworx in the cluster. It has
              all the necessary configurations to setup and update a Portworx cluster.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.core.libopenstorage.org
            version: v1
          - description: Do not create Storage Node as it is internally created by
              the operator. It represents the status of a Portworx node.
            displayName: Storage Node
            kind: StorageNode
            name: storagenodes.core.libopenstorage.org
            version: v1
        description: |
          Portworx-Enterprise is the most widely-used and reliable cloud-native
          storage solution for production workloads and provides high-availability,
          data protection and security for containerized applications.

          Portworx Enterprise enables you to migrate entire applications, including
          data, between clusters in a single data center or cloud, or between clouds,
          with a single kubectl command.

          The cloud native storage and data management platform that enterprises trust
          to manage data in containers now has an operator which simplifies the install,
          configuration, upgrades and manages the Portworx Enterprise cluster lifecycle.

          Learn more about the Portworx Enterprise
          [the data platform for Kubernetes](https://portworx.com/products/introduction)

          To learn more about the platform features, please visit our
          [product features page](https://portworx.com/products/features)

          ### About Portworx

          Portworx is the solution for running stateful containers in production,
          designed with DevOps in mind. With Portworx, users can manage any database
          or stateful service on any infrastructure using any container scheduler,
          including Kubernetes, Mesosphere DC/OS, and Docker Swarm. Portworx solves
          the five most common problems DevOps teams encounter when running stateful
          services in production: persistence, high availability, data automation,
          security, and support for multiple data stores and infrastructure.

          ### How to install StorageCluster

          To customize your cluster's configuration (specification), use the
          [Spec Generator](https://central.portworx.com/) from PX-Central.

          ### Prerequisite

          Ensure ports 17001-17020 on worker nodes are reachable from master and other worker nodes.

          ### Tutorials

          * [Portworx Enterprise on Openshift](https://portworx.com/openshift)

          * [Stateful applications on Kubernetes](https://docs.portworx.com/portworx-enterprise/concepts/kubernetes-storage-101/applications)

          * [Portworx Enterprise on Kubernetes](https://docs.portworx.com/portworx-install-with-kubernetes)

          * [Kafka on Kubernetes](https://portworx.com/kafka-kubernetes)

          * [Elastisearch on Kubernetes](https://portworx.com/elasticsearch-kubernetes)

          * [PostgreSQL on Kubernetes](https://portworx.com/postgres-kubernetes/)

          * [MongoDB on Kubernetes](https://portworx.com/mongodb-kubernetes/)

          * [Cassandra on Kubernetes](https://portworx.com/cassandra-kubernetes/)

          * [Kubernetes backup and recovery](https://portworx.com/kubernetes-backup/)

          * [Disaster Recovery for Kubernetes](https://portworx.com/kubernetes-disaster-recovery/)

          ### Uninstall

          Deleting the StorageCluster object for Portworx cluster does not stop Portworx
          service running on the nodes, to avoid application downtime.

          To uninstall Portworx completely without wiping the data, you should add the
          following delete strategy to the StorageCluster spec:
          ```
          spec:
            deleteStrategy:
              type: Uninstall
          ```
          **Caution:** To uninstall Portworx and **wipe all the data**, you should use the following
          delete strategy:
          ```
          spec:
            deleteStrategy:
              type: UninstallAndWipe
          ```
        displayName: Portworx Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - portworx
        - persistent storage
        - storage
        - cloud native
        - open source
        links:
        - name: Product Features
          url: https://portworx.com/products/features
        - name: Documentation
          url: https://docs.portworx.com/portworx-enterprise/platform/openshift
        - name: Support
          url: https://docs.portworx.com/portworx-enterprise/support/contact-support
        - name: Source Code
          url: https://github.com/libopenstorage/operator
        maintainers:
        - email: support@portworx.com
          name: Portworx
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Portworx by PureStorage
        relatedImages:
        - registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
        version: 25.3.1
      entries:
      - name: portworx-operator.v25.3.1
        version: 25.3.1
      - name: portworx-operator.v25.3.0
        version: 25.3.0
      - name: portworx-operator.v25.2.2
        version: 25.2.2
      - name: portworx-operator.v25.2.1
        version: 25.2.1
      - name: portworx-operator.v25.2.0
        version: 25.2.0
      - name: portworx-operator.v24.2.4
        version: 24.2.4
      - name: portworx-operator.v24.2.3
        version: 24.2.3
      - name: portworx-operator.v24.2.2
        version: 24.2.2
      - name: portworx-operator.v24.2.1
        version: 24.2.1
      - name: portworx-operator.v24.2.0
        version: 24.2.0
      - name: portworx-operator.v24.1.3
        version: 24.1.3
      - name: portworx-operator.v24.1.2
        version: 24.1.2
      - name: portworx-operator.v24.1.1
        version: 24.1.1
      - name: portworx-operator.v24.1.0
        version: 24.1.0
      - name: portworx-operator.v23.10.5
        version: 23.10.5
      - name: portworx-operator.v23.10.4
        version: 23.10.4
      - name: portworx-operator.v23.10.3
        version: 23.10.3
      - name: portworx-operator.v23.10.2
        version: 23.10.2
      - name: portworx-operator.v23.10.1
        version: 23.10.1
      - name: portworx-operator.v23.7.0
        version: 23.7.0
      - name: portworx-operator.v23.5.1
        version: 23.5.1
      - name: portworx-operator.v23.5.0
        version: 23.5.0
      - name: portworx-operator.v23.4.0
        version: 23.4.0
      - name: portworx-operator.v23.3.1
        version: 23.3.1
      - name: portworx-operator.v23.3.0
        version: 23.3.0
      - name: portworx-operator.v1.10.5
        version: 1.10.5
      - name: portworx-operator.v1.10.4
        version: 1.10.4
      - name: portworx-operator.v1.10.3
        version: 1.10.3
      - name: portworx-operator.v1.10.2
        version: 1.10.2
      - name: portworx-operator.v1.10.1
        version: 1.10.1
      - name: portworx-operator.v1.10.0
        version: 1.10.0
      - name: portworx-operator.v1.9.1
        version: 1.9.1
      - name: portworx-operator.v1.9.0
        version: 1.9.0
      - name: portworx-operator.v1.8.1
        version: 1.8.1
      - name: portworx-operator.v1.8.0
        version: 1.8.0
      - name: portworx-operator.v1.7.1
        version: 1.7.1
      - name: portworx-operator.v1.6.1
        version: 1.6.1
      name: alpha
    - currentCSV: portworx-operator.v25.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageCluster",
                "metadata": {
                  "name": "portworx",
                  "namespace": "test-operator",
                  "annotations": {
                    "portworx.io/is-openshift": "true"
                  }
                },
                "spec": {}
              },
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageNode",
                "metadata": {
                  "name": "example",
                  "namespace": "test-operator"
                },
                "spec": {}
              }
            ]
          capabilities: Auto Pilot
          categories: Storage
          certified: "true"
          console.openshift.io/plugins: '["portworx"]'
          containerImage: registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
          createdAt: "2025-09-06T19:38:43Z"
          description: Cloud native storage solution for production workloads
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "core.libopenstorage.org/v1",
              "kind": "StorageCluster",
              "metadata": {
                "name": "portworx",
                "annotations": {
                  "portworx.io/is-openshift": "true"
                }
              }
            }
          support: Portworx, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Storage Cluster installs Portworx in the cluster. It has
              all the necessary configurations to setup and update a Portworx cluster.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.core.libopenstorage.org
            version: v1
          - description: Do not create Storage Node as it is internally created by
              the operator. It represents the status of a Portworx node.
            displayName: Storage Node
            kind: StorageNode
            name: storagenodes.core.libopenstorage.org
            version: v1
        description: |
          Portworx-Enterprise is the most widely-used and reliable cloud-native
          storage solution for production workloads and provides high-availability,
          data protection and security for containerized applications.

          Portworx Enterprise enables you to migrate entire applications, including
          data, between clusters in a single data center or cloud, or between clouds,
          with a single kubectl command.

          The cloud native storage and data management platform that enterprises trust
          to manage data in containers now has an operator which simplifies the install,
          configuration, upgrades and manages the Portworx Enterprise cluster lifecycle.

          Learn more about the Portworx Enterprise
          [the data platform for Kubernetes](https://portworx.com/products/introduction)

          To learn more about the platform features, please visit our
          [product features page](https://portworx.com/products/features)

          ### About Portworx

          Portworx is the solution for running stateful containers in production,
          designed with DevOps in mind. With Portworx, users can manage any database
          or stateful service on any infrastructure using any container scheduler,
          including Kubernetes, Mesosphere DC/OS, and Docker Swarm. Portworx solves
          the five most common problems DevOps teams encounter when running stateful
          services in production: persistence, high availability, data automation,
          security, and support for multiple data stores and infrastructure.

          ### How to install StorageCluster

          To customize your cluster's configuration (specification), use the
          [Spec Generator](https://central.portworx.com/) from PX-Central.

          ### Prerequisite

          Ensure ports 17001-17020 on worker nodes are reachable from master and other worker nodes.

          ### Tutorials

          * [Portworx Enterprise on Openshift](https://portworx.com/openshift)

          * [Stateful applications on Kubernetes](https://docs.portworx.com/portworx-enterprise/concepts/kubernetes-storage-101/applications)

          * [Portworx Enterprise on Kubernetes](https://docs.portworx.com/portworx-install-with-kubernetes)

          * [Kafka on Kubernetes](https://portworx.com/kafka-kubernetes)

          * [Elastisearch on Kubernetes](https://portworx.com/elasticsearch-kubernetes)

          * [PostgreSQL on Kubernetes](https://portworx.com/postgres-kubernetes/)

          * [MongoDB on Kubernetes](https://portworx.com/mongodb-kubernetes/)

          * [Cassandra on Kubernetes](https://portworx.com/cassandra-kubernetes/)

          * [Kubernetes backup and recovery](https://portworx.com/kubernetes-backup/)

          * [Disaster Recovery for Kubernetes](https://portworx.com/kubernetes-disaster-recovery/)

          ### Uninstall

          Deleting the StorageCluster object for Portworx cluster does not stop Portworx
          service running on the nodes, to avoid application downtime.

          To uninstall Portworx completely without wiping the data, you should add the
          following delete strategy to the StorageCluster spec:
          ```
          spec:
            deleteStrategy:
              type: Uninstall
          ```
          **Caution:** To uninstall Portworx and **wipe all the data**, you should use the following
          delete strategy:
          ```
          spec:
            deleteStrategy:
              type: UninstallAndWipe
          ```
        displayName: Portworx Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - portworx
        - persistent storage
        - storage
        - cloud native
        - open source
        links:
        - name: Product Features
          url: https://portworx.com/products/features
        - name: Documentation
          url: https://docs.portworx.com/portworx-enterprise/platform/openshift
        - name: Support
          url: https://docs.portworx.com/portworx-enterprise/support/contact-support
        - name: Source Code
          url: https://github.com/libopenstorage/operator
        maintainers:
        - email: support@portworx.com
          name: Portworx
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Portworx by PureStorage
        relatedImages:
        - registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
        version: 25.3.1
      entries:
      - name: portworx-operator.v25.3.1
        version: 25.3.1
      - name: portworx-operator.v25.3.0
        version: 25.3.0
      - name: portworx-operator.v25.2.2
        version: 25.2.2
      - name: portworx-operator.v25.2.1
        version: 25.2.1
      - name: portworx-operator.v25.2.0
        version: 25.2.0
      - name: portworx-operator.v24.2.4
        version: 24.2.4
      - name: portworx-operator.v24.2.3
        version: 24.2.3
      - name: portworx-operator.v24.2.2
        version: 24.2.2
      - name: portworx-operator.v24.2.1
        version: 24.2.1
      - name: portworx-operator.v24.2.0
        version: 24.2.0
      - name: portworx-operator.v24.1.3
        version: 24.1.3
      - name: portworx-operator.v24.1.2
        version: 24.1.2
      - name: portworx-operator.v24.1.1
        version: 24.1.1
      - name: portworx-operator.v24.1.0
        version: 24.1.0
      - name: portworx-operator.v23.10.5
        version: 23.10.5
      - name: portworx-operator.v23.10.4
        version: 23.10.4
      - name: portworx-operator.v23.10.3
        version: 23.10.3
      - name: portworx-operator.v23.10.2
        version: 23.10.2
      - name: portworx-operator.v23.10.1
        version: 23.10.1
      - name: portworx-operator.v23.7.0
        version: 23.7.0
      - name: portworx-operator.v23.5.1
        version: 23.5.1
      - name: portworx-operator.v23.5.0
        version: 23.5.0
      - name: portworx-operator.v23.4.0
        version: 23.4.0
      - name: portworx-operator.v23.3.1
        version: 23.3.1
      - name: portworx-operator.v23.3.0
        version: 23.3.0
      - name: portworx-operator.v1.10.5
        version: 1.10.5
      - name: portworx-operator.v1.10.4
        version: 1.10.4
      - name: portworx-operator.v1.10.3
        version: 1.10.3
      - name: portworx-operator.v1.10.2
        version: 1.10.2
      - name: portworx-operator.v1.10.1
        version: 1.10.1
      - name: portworx-operator.v1.10.0
        version: 1.10.0
      - name: portworx-operator.v1.9.1
        version: 1.9.1
      - name: portworx-operator.v1.9.0
        version: 1.9.0
      - name: portworx-operator.v1.8.1
        version: 1.8.1
      - name: portworx-operator.v1.8.0
        version: 1.8.0
      - name: portworx-operator.v1.7.1
        version: 1.7.1
      - name: portworx-operator.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: portworx-certified
    provider:
      name: Portworx by PureStorage
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: clusterresourceoverride
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clusterresourceoverride-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.autoscaling.openshift.io/v1",
                "kind": "ClusterResourceOverride",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "podResourceOverride": {
                    "spec": {
                      "memoryRequestToLimitPercent": 50,
                      "cpuRequestToLimitPercent": 25,
                      "limitCPUToMemoryPercent": 200
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certifiedLevel: Primed
          containerImage: registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9-operator@sha256:40b19c94f688c3a58189fcb404df35a40a8acbbdbf73fadb8818a86fd14a9ab9
          createdAt: 2019/11/15
          description: An operator to manage the OpenShift ClusterResourceOverride
            Mutating Admission Webhook Server
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          healthIndex: B
          olm.skipRange: '>=4.3.0 <4.16.0-202509240123'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-resource-override-admission-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of ClusterResourceOverride Admission
              Webhook
            displayName: ClusterResourceOverride
            kind: ClusterResourceOverride
            name: clusterresourceoverrides.operator.autoscaling.openshift.io
            version: v1
        description: "ClusterResourceOverride\n==============\n\nContainers can specify
          compute resource requests and limits. Requests are used for scheduling your
          container and provide a minimum service guarantee. Limits constrain the
          amount of compute resource that may be consumed on your node.\n\nThe scheduler
          attempts to optimize the compute resource use across all nodes in your cluster.
          It places pods onto specific nodes, taking the pods' compute resource requests
          and nodes' available capacity into consideration.\n\nRequests and limits
          enable administrators to allow and manage the overcommitment of resources
          on a node, which may be desirable in development environments where a trade
          off of guaranteed performance for capacity is acceptable.\n\n### Requests
          and Limits\n\nFor each compute resource, a container may specify a resource
          request and limit. Scheduling decisions are made based on the request to
          ensure that a node has enough capacity available to meet the requested value.
          If a container specifies limits, but omits requests, the requests are defaulted
          to the limits. A container is not able to exceed the specified limit on
          the node.\n\nThe enforcement of limits is dependent upon the compute resource
          type. If a container makes no request or limit, the container is scheduled
          to a node with no resource guarantees. In practice, the container is able
          to consume as much of the specified resource as is available with the lowest
          local priority. In low resource situations, containers that specify no resource
          requests are given the lowest quality of service.\n\n### Compute Resources\n\nThe
          node-enforced behavior for compute resources is specific to the resource
          type.\n\n#### CPU\n\nA container is guaranteed the amount of CPU it requests
          and is additionally able to consume excess CPU available on the node, up
          to any limit specified by the container. If multiple containers are attempting
          to use excess CPU, CPU time is distributed based on the amount of CPU requested
          by each container.\n\nFor example, if one container requested 500m of CPU
          time and another container requested 250m of CPU time, then any extra CPU
          time available on the node is distributed among the containers in a 2:1
          ratio. If a container specified a limit, it will be throttled not to use
          more CPU than the specified limit.\n\nCPU requests are enforced using the
          CFS shares support in the Linux kernel. By default, CPU limits are enforced
          using the CFS quota support in the Linux kernel over a 100ms measuring interval,
          though this can be disabled.\n\n#### Memory\n\nA container is guaranteed
          the amount of memory it requests. A container may use more memory than requested,
          but once it exceeds its requested amount, it could be killed in a low memory
          situation on the node.\n\nIf a container uses less memory than requested,
          it will not be killed unless system tasks or daemons need more memory than
          was accounted for in the node’s resource reservation. If a container specifies
          a limit on memory, it is immediately killed if it exceeds the limit amount.\n\n###
          Configuring the Cluster for Overcommitment\n\nScheduling is based on resources
          requested, while quota and hard limits refer to resource limits, which can
          be set higher than requested resources. The difference between request and
          limit determines the level of overcommit; for instance, if a container is
          given a memory request of 1Gi and a memory limit of 2Gi, it is scheduled
          based on the 1Gi request being available on the node, but could use up to
          2Gi; so it is 200% overcommitted. If OpenShift Container Platform administrators
          would like to control the level of overcommit and manage container density
          on nodes, ClusterResourceOverride Admission Webhook can be configured to
          override the ratio between request and limit set on developer containers.
          In conjunction with a per-project LimitRange specifying limits and defaults,
          this adjusts the container limit and request to achieve the desired level
          of overcommit.\n\nThis requires creating a custom resource of `ClusterResourceOverride`
          type as in the following example:\n\n    \n    - apiVersion: operator.autoscaling.openshift.io/v1\n
          \   - kind: ClusterResourceOverride\n    - metadata:\n    -   name: cluster\n
          \   - spec:\n    -   podResourceOverride:\n    -     spec:\n    -       memoryRequestToLimitPercent:
          25\n    -       cpuRequestToLimitPercent: 25\n    -       limitCPUToMemoryPercent:
          200\n    \n **memoryRequestToLimitPercent**: (optional, 1-100) If a container
          memory limit has been specified or defaulted, the memory request is overridden
          to this percentage of the limit.\n\n **cpuRequestToLimitPercent**: (optional,
          1-100) If a container CPU limit has been specified or defaulted, the CPU
          request is overridden to this percentage of the limit.\n\n **limitCPUToMemoryPercent**:
          (optional, positive integer) If a container memory limit has been specified
          or defaulted, the CPU limit is overridden to a percentage of the memory
          limit, with a 100 percentage scaling 1Gi of RAM to equal 1 CPU core. This
          is processed prior to overriding CPU request (if configured).\n\n Note that
          these overrides have no effect if no limits have been set on containers.
          [Create a LimitRange object] (https://docs.openshift.com/container-platform/3.3/admin_guide/limits.html#admin-guide-limits)
          with default limits (per individual project, or in the [project template](https://docs.openshift.com/container-platform/3.3/admin_guide/managing_projects.html#modifying-the-template-for-new-projects))
          in order to ensure that the overrides apply.\n\nWhen configured, overrides
          can be enabled per-project by applying the following label.\n    ```\n    clusterresourceoverrides.admission.autoscaling.openshift.io/enabled:
          \"true\"\n    ```\n\n"
        displayName: ClusterResourceOverride Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - deschedule
        - scale
        - binpack
        - efficiency
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9-operator@sha256:40b19c94f688c3a58189fcb404df35a40a8acbbdbf73fadb8818a86fd14a9ab9
        - registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9@sha256:e4992c556fe6be50fa30a45b05cffd6029ebd8914eca7c1c683564ff869e0e7d
        version: 4.16.0-202509240123
      entries:
      - name: clusterresourceoverride-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: clusterresourceoverride-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: clusterresourceoverride-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: clusterresourceoverride-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: clusterresourceoverride-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: clusterresourceoverride-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: clusterresourceoverride-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: clusterresourceoverride-operator.v4.16.0-202505092110
        version: 4.16.0-202505092110
      - name: clusterresourceoverride-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: clusterresourceoverride-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: clusterresourceoverride-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: clusterresourceoverride-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: clusterresourceoverride-operator.v4.16.0-202501241104
        version: 4.16.0-202501241104
      - name: clusterresourceoverride-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: clusterresourceoverride-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: clusterresourceoverride-operator.v4.16.0-202412021735
        version: 4.16.0-202412021735
      - name: clusterresourceoverride-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: clusterresourceoverride-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: clusterresourceoverride-operator.v4.16.0-202411081636
        version: 4.16.0-202411081636
      - name: clusterresourceoverride-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: clusterresourceoverride-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: clusterresourceoverride-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: clusterresourceoverride-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: clusterresourceoverride-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: clusterresourceoverride-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: clusterresourceoverride-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: clusterresourceoverride-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: clusterresourceoverride-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: clusterresourceoverride-operator.v4.16.0-202407040206
        version: 4.16.0-202407040206
      - name: clusterresourceoverride-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: clusterresourceoverride-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: clusterresourceoverride-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: clusterresourceoverride-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: clusterresourceoverride
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicurio-registry-3
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-3.v3.0.15-r1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "simple"
                },
                "spec": {
                  "app": {
                    "ingress": {
                      "host": "simple-app.apps.cluster.example"
                    }
                  },
                  "ui": {
                    "ingress": {
                      "host": "simple-ui.apps.cluster.example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/apicurio/apicurio-registry-rhel9-operator@sha256:e6d2ceab8accdb6e1b9766733039e24df620d7128eb1e13e60a61339130d52a7
          createdAt: "2025-09-29"
          description: Deploy and manage Red Hat build of Apicurio Registry 3 on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations",
            "Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Red Hat build
              of Apicurio Registry version 3.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Red Hat build of Apicurio Registry 3

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Allows you manage event schemas and API designs using the Apicurio Registry web console, REST API, Maven plug-in, or Java client
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, MySQL\*, SQL Server\*, in-memory\*
          - Compatible with Confluent APIs
          - Runs on a lightweight Quarkus platform

          _Note: Storage options marked with * have not yet been tested by Red Hat for production deployments._

          ## Red Hat build of Apicurio Registry 3 Operator

          Provides a quick and easy way to deploy and manage Red Hat build of Apicurio Registry 3 on OpenShift.

          **Features**
          - Supports installation and configuration of Red Hat build of Apicurio Registry 3
          - Perform a rolling upgrade of Red Hat build of Apicurio Registry 3

          ## Prerequisites

          This operator does not deploy storage for Red Hat build of Apicurio Registry 3. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Red Hat build of Apicurio Registry 3 and Apicurio Registry 3 Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Red Hat build of Apicurio Registry 3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/apicurio/apicurio-registry-ui-rhel9@sha256:e4791546fd371b34c7580c6aca2faed8527c985c8029fa8087fb46fd94ab8feb
        - registry.redhat.io/apicurio/apicurio-registry-rhel9-operator@sha256:e6d2ceab8accdb6e1b9766733039e24df620d7128eb1e13e60a61339130d52a7
        - registry.redhat.io/apicurio/apicurio-registry-rhel9@sha256:c86d73d5e2b4146f016ce7ec67265fe80bd596f472247eb9f5590cc0c89dba75
        version: 3.0.15-r1
      entries:
      - name: apicurio-registry-3.v3.0.15-r1
        version: 3.0.15-r1
      - name: apicurio-registry-3.v3.0.9-r3
        version: 3.0.9-r3
      - name: apicurio-registry-3.v3.0.9-r2
        version: 3.0.9-r2
      - name: apicurio-registry-3.v3.0.9-r1
        version: 3.0.9-r1
      - name: apicurio-registry-3.v3.0.7-r2
        version: 3.0.7-r2
      - name: apicurio-registry-3.v3.0.7-r1
        version: 3.0.7-r1
      - name: apicurio-registry-3.v3.0.6-beta.r1
        version: 3.0.6-beta.r1
      - name: apicurio-registry-3.v3.0.6-beta.r1-0.1740495889.p
        version: 3.0.6-beta.r1+0.1740495889.p
      - name: apicurio-registry-3.v3.0.6-beta
        version: 3.0.6-beta
      name: 3.x
    defaultChannel: 3.x
    packageName: apicurio-registry-3
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: MariaDB Corporation
      provider-url: ""
    name: mariadb-enterprise-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-enterprise-operator.v25.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup"
                },
                "spec": {
                  "args": [
                    "--single-transaction",
                    "--all-databases"
                  ],
                  "logLevel": "info",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "0 0 * * *",
                    "suspend": false
                  },
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "backups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Connection",
                "metadata": {
                  "name": "connection"
                },
                "spec": {
                  "database": "mariadb",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "secretName": "connection",
                  "secretTemplate": {
                    "key": "dsn"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Database",
                "metadata": {
                  "name": "database"
                },
                "spec": {
                  "characterSet": "utf8",
                  "collate": "utf8_general_ci",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "name": "my-logical-database"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "grant"
                },
                "spec": {
                  "database": "*",
                  "grantOption": true,
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "privileges": [
                    "SELECT",
                    "INSERT",
                    "UPDATE"
                  ],
                  "table": "*",
                  "username": "my-user"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "MariaDB",
                "metadata": {
                  "name": "mariadb"
                },
                "spec": {
                  "connection": {
                    "secretName": "mariadb-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "database": "mariadb",
                  "galera": {
                    "enabled": true
                  },
                  "metrics": {
                    "enabled": true,
                    "passwordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "mariadb-metrics"
                    }
                  },
                  "myCnf": "[mariadb]\nbind-address=*\ndefault_storage_engine=InnoDB\nbinlog_format=row\ninnodb_autoinc_lock_mode=2\ninnodb_buffer_pool_size=1024M\nmax_allowed_packet=256M\n",
                  "passwordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-password"
                  },
                  "primaryConnection": {
                    "secretName": "mariadb-conn-primary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "primaryService": {
                    "type": "ClusterIP"
                  },
                  "replicas": 3,
                  "rootPasswordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-root"
                  },
                  "secondaryConnection": {
                    "secretName": "mariadb-conn-secondary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "secondaryService": {
                    "type": "ClusterIP"
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "storage": {
                    "size": "1Gi"
                  },
                  "updateStrategy": {
                    "type": "ReplicasFirstPrimaryLast"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "MaxScale",
                "metadata": {
                  "name": "maxscale"
                },
                "spec": {
                  "admin": {
                    "guiEnabled": true,
                    "port": 8989
                  },
                  "auth": {
                    "adminPasswordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "maxscale-admin"
                    },
                    "adminUsername": "mariadb-enterprise-operator",
                    "generate": true
                  },
                  "config": {
                    "sync": {
                      "database": "mysql",
                      "interval": "5s",
                      "timeout": "10s"
                    }
                  },
                  "connection": {
                    "port": 3306,
                    "secretName": "maxscale-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "guiKubernetesService": {
                    "type": "ClusterIP"
                  },
                  "kubernetesService": {
                    "type": "ClusterIP"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "metrics": {
                    "enabled": true
                  },
                  "monitor": {
                    "cooperativeMonitoring": "majority_of_all",
                    "interval": "2s",
                    "params": {
                      "available_when_donor": "false",
                      "disable_master_failback": "false",
                      "disable_master_role_setting": "false"
                    }
                  },
                  "replicas": 2,
                  "requeueInterval": "10s",
                  "services": [
                    {
                      "listener": {
                        "port": 3306,
                        "protocol": "MariaDBProtocol"
                      },
                      "name": "rw-router",
                      "params": {
                        "transaction_replay": "true",
                        "transaction_replay_attempts": "10",
                        "transaction_replay_timeout": "5s"
                      },
                      "router": "readwritesplit"
                    }
                  ]
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "PhysicalBackup",
                "metadata": {
                  "name": "physicalbackup"
                },
                "spec": {
                  "compression": "bzip2",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "podAffinity": true,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "0 0 * * *",
                    "immediate": true,
                    "suspend": false
                  },
                  "serviceAccountName": "backup",
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "physicalbackups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "region": "us-east-1",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  },
                  "timeout": "1h"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore"
                },
                "spec": {
                  "backupRef": {
                    "name": "backup"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "targetRecoveryTime": "2023-12-19T09:00:00Z"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "SqlJob",
                "metadata": {
                  "name": "sqljob"
                },
                "spec": {
                  "database": "mariadb",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "sql": "CREATE TABLE IF NOT EXISTS orders (\n  id bigint PRIMARY KEY AUTO_INCREMENT,\n  customer_id INT,\n  order_date DATE,\n  total_amount_cents INT\n);\nINSERT INTO orders (customer_id, order_date, total_amount_cents) VALUES\n  (101, '2024-05-15', 5000),\n  (102, '2024-05-14', 7525),\n  (103, '2024-05-13', 10050),\n  (104, '2024-05-12', 2575);",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "user"
                },
                "spec": {
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxUserConnections": 20,
                  "name": "my-user",
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.mariadb.com/mariadb-enterprise-operator@sha256:1effbbfa7632be92c17ef61c18a25f010b187840beec42c31d2410d0209e6985
          createdAt: "2025-08-28T11:35:03Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["MariaDB Enterprise"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: MariaDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API. It is used to define
              backup jobs and its storage.
            displayName: Backup
            kind: Backup
            name: backups.enterprise.mariadb.com
            version: v1alpha1
          - description: Connection is the Schema for the connections API. It is used
              to configure connection strings for the applications connecting to MariaDB.
            displayName: Connection
            kind: Connection
            name: connections.enterprise.mariadb.com
            version: v1alpha1
          - description: Database is the Schema for the databases API. It is used
              to define a logical database as if you were running a 'CREATE DATABASE'
              statement.
            displayName: Database
            kind: Database
            name: databases.enterprise.mariadb.com
            version: v1alpha1
          - description: Grant is the Schema for the grants API. It is used to define
              grants as if you were running a 'GRANT' statement.
            displayName: Grant
            kind: Grant
            name: grants.enterprise.mariadb.com
            version: v1alpha1
          - description: MariaDB is the Schema for the mariadbs API. It is used to
              define MariaDB clusters.
            displayName: Maria DB
            kind: MariaDB
            name: mariadbs.enterprise.mariadb.com
            version: v1alpha1
          - description: MaxScale is the Schema for the maxscales API. It is used
              to define MaxScale clusters.
            displayName: Max Scale
            kind: MaxScale
            name: maxscales.enterprise.mariadb.com
            version: v1alpha1
          - description: PhysicalBackup is the Schema for the physicalbackups API.
              It is used to define physical backup jobs and its storage.
            displayName: Physical Backup
            kind: PhysicalBackup
            name: physicalbackups.enterprise.mariadb.com
            version: v1alpha1
          - description: Restore is the Schema for the restores API. It is used to
              define restore jobs and its restoration source.
            displayName: Restore
            kind: Restore
            name: restores.enterprise.mariadb.com
            version: v1alpha1
          - description: SqlJob is the Schema for the sqljobs API. It is used to run
              sql scripts as jobs.
            displayName: Sql Job
            kind: SqlJob
            name: sqljobs.enterprise.mariadb.com
            version: v1alpha1
          - description: User is the Schema for the users API.  It is used to define
              grants as if you were running a 'CREATE USER' statement.
            displayName: User
            kind: User
            name: users.enterprise.mariadb.com
            version: v1alpha1
        description: |
          MariaDB Enterprise Operator for Openshift is an operator that manages MariaDB Enterprise operands declaratively via Custom Resources, with support for the following features:

          * Provision and Configure MariaDB and MaxScale Declaratively: Define MariaDB Enterprise Server and MaxScale clusters in YAML manifests and deploy them with ease in Kubernetes.
          * High Availability with Galera: Ensure availability with MariaDB Enterprise Cluster, providing synchronous multi-master replication.
          * Query and Connection-Based Routing with MaxScale: MaxScale provides query routing and connection load balancing for improved application performance.
          * Cluster-Aware Rolling Updates: Perform rolling updates on MariaDB and MaxScale clusters, ensuring zero-downtime upgrades with no disruptions to your applications.
          * Flexible Storage Configuration and Volume Expansion: Easily configure storage for MariaDB instances, including the ability to expand volumes as needed.
          * Physical Backups based on [mariadb-backup](https://mariadb.com/docs/server/server-usage/backup-and-restore/mariadb-backup/full-backup-and-restore-with-mariadb-backup) and [Kubernetes VolumeSnapshots](https://kubernetes.io/docs/concepts/storage/volume-snapshots/). By leveraging the [BACKUP STAGE](https://mariadb.com/docs/server/reference/sql-statements/administrative-sql-statements/backup-commands/backup-stage) feature, backups are taken without long read locks or service interruptions.
          * Logical Backups based on [mariadb-dump](https://mariadb.com/docs/server/clients-and-utilities/backup-restore-and-import-clients/mariadb-dump).
          * Backup Management: Take, restore, and schedule backups with multiple storage types supported: S3, PVCs, Kubernetes volumes and VolumeSnapshots..
          * Policy-Driven Backup Retention: Implement backup retention policies with bzip2 and gzip compression.
          * Target Recovery Time: Restore your database to the closest available backup based on a specified recovery time.
          * Bootstrap New Instances: Initialize new MariaDB instances from backups, S3, PVCs or VolumeSnapshots to quickly spin up new clusters.
          * TLS Certificate Management: Issue, configure, and rotate TLS certificates and Certificate Authorities (CAs) for secure connections.
          * Advanced TLS Support: customize certificate lifetime, private key algorithm and TLS version.
          * Native Integration with cert-manager: Leverage [cert-manager](https://cert-manager.io/docs/), the de-facto standard for managing certificates in Kubernetes, to enable issuance with private CAs, public CAs and HashiCorp Vault.
          * Prometheus Metrics: Expose metrics using the MariaDB and MaxScale Prometheus exporters.
          * Native Integration with prometheus-operator: Leverage [prometheus-operator](https://github.com/prometheus-operator/prometheus-operator) to scrape metrics from MariaDB and MaxScale instances.
          * Declarative User and Database Management: Manage users, grants, and logical databases in a declarative manner using Kubernetes resources.
          * Secure, immutable and lightweight images based on Red Hat UBI, available for multiple architectires (amd64, arm64 and ppc64le).
          * [Operator certified ](https://catalog.redhat.com/en/software/container-stacks/detail/65789bcbe17f1b31944acb1d#overview) by Red Hat.

          ## Installation

          In order to use this product, you will need a [MariaDB Enterprise subscription](https://mariadb.com/products/enterprise/).

          To be able to pull MariaDB Enterprise images, you will need to configure the MariaDB registry credentials in Openshift using the [MariaDB customer download token](https://customers.mariadb.com/downloads/token/) as described below:

          - Extract your [Openshift global pull secret](https://docs.redhat.com/en/documentation/openshift_container_platform/4.18/html/images/managing-images#images-update-global-pull-secret_using-image-pull-secrets):
            ```bash
            oc extract secret/pull-secret -n openshift-config --confirm
            ```

          - Login in the MariaDB registry providing the customer download token as password:
            ```bash
            oc registry login \
              --registry="docker.mariadb.com" \
              --auth-basic="<email>:<customer-download-token>" \
              --to=.dockerconfigjson
            ```

          - Update the global pull secret:
            ```bash
            oc set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson
            ```

          Alternatively, you can also create a `Secret` named `mariadb-enterprise` in the namespace where are you installing the operator by using the following command:
            ```bash
            oc create secret docker-registry mariadb-enterprise \
              --docker-server=docker.mariadb.com \
              --docker-username=<email> \
              --docker-password=<customer-download-token> \
              -n <operator-namespace>
            ```

          At this point, you can proceed to install the MariaDB Enterprise Operator, either via the Openshift console or by creating a `Subscription` object as described in the [Openshift installation docs](https://mariadb.com/kb/en/openshift/).

          Refer to the [quickstart guide](https://mariadb.com/kb/en/quickstart/), the [examples catalog](https://mariadb.com/kb/en/examples-catalog/) and the [API reference](https://mariadb.com/kb/en/api-reference/) for getting started with the Custom Resources offered by this operator.
        displayName: MariaDB Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mariadb
        - mysql
        - database
        - operator
        - mariadb-enterprise-operator
        - database
        links:
        - name: MariaDB Enterprise
          url: https://mariadb.com/products/enterprise/
        - name: Documentation
          url: https://mariadb.com/kb/en/mariadb-enterprise-operator/
        maintainers:
        - email: operator@mariadb.com
          name: MariaDB Corporation
        - email: support@mariadb.com
          name: Support
        - email: sales-EMEA@mariadb.com
          name: Sales EMEA
        - email: sales-APAC@mariadb.com
          name: Sales APAC
        - email: sales-nam@mariadb.com
          name: Sales Americas
        maturity: beta
        minKubeVersion: 1.29.0
        provider:
          name: MariaDB Corporation
        relatedImages:
        - docker.mariadb.com/enterprise-server@sha256:9872721d35ebe8d820b1aa9c6dd83b363a7b2df9af11a65bff6b80ff4144460a
        - docker.mariadb.com/maxscale@sha256:d9e04a7a406353dab2051e190be0c32ed20b780ed71e56e55cedd4ccb61af416
        - mariadb/mariadb-prometheus-exporter-ubi@sha256:f072175198cad050e80617d13e63f415bacdfe47e2dd177fde7283a96ee56c07
        - mariadb/maxscale-prometheus-exporter-ubi@sha256:cc17c1554795ad01b4d8db799f6aa90e00e41ede431f09d30eec64d9dd4479d9
        - docker.mariadb.com/mariadb-enterprise-operator@sha256:1effbbfa7632be92c17ef61c18a25f010b187840beec42c31d2410d0209e6985
        version: 25.8.1-1
      entries:
      - name: mariadb-enterprise-operator.v25.8.1-1
        version: 25.8.1-1
      - name: mariadb-enterprise-operator.v25.8.0
        version: 25.8.0
      - name: mariadb-enterprise-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: mariadb-enterprise-operator
    provider:
      name: MariaDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Instana
      provider-url: ""
    name: instana-agent-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: instana-agent-operator.v2.1.36
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instana.io/v1",
                "kind": "InstanaAgent",
                "metadata": {
                  "name": "instana-agent",
                  "namespace": "instana-agent"
                },
                "spec": {
                  "agent": {
                    "configuration_yaml": "# You can leave this empty, or use this to configure your instana agent.\n# See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.\n",
                    "endpointHost": "ingress-red-saas.instana.io",
                    "endpointPort": "443",
                    "env": {},
                    "key": "replace-key"
                  },
                  "cluster": {
                    "name": "my-cluster"
                  },
                  "opentelemetry": {
                    "grpc": {
                      "enabled": true
                    },
                    "http": {
                      "enabled": true
                    }
                  },
                  "zone": {
                    "name": "edited-zone"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring,Logging & Tracing,OpenShift Optional
          certified: "false"
          containerImage: icr.io/instana/instana-agent-operator@sha256:ab8b35f0b45c68f69210d1b17c5dd609799f6bb0bc9f12d0d4e5302c910607a1
          description: Fully automated Application Performance Monitoring (APM) for
            microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/instana-agent-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/instana-agent-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/instana/instana-agent-operator
          support: Instana
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InstanaAgent is the Schema for the agents API
            displayName: Instana Agent
            kind: InstanaAgent
            name: agents.instana.io
            version: v1
          - description: InstanaAgentRemote is the Schema for the agents API
            displayName: Instana Agent Remote
            kind: InstanaAgentRemote
            name: agentsremote.instana.io
            version: v1
        description: |
          # Instana

          Instana is an [APM solution](https://www.instana.com/product-overview/) built for microservices that enables IT Ops to build applications faster and deliver higher quality services by automating monitoring, tracing and root cause analysis. The solution is optimized for [Kubernetes](https://www.instana.com/automatic-kubernetes-monitoring/) and [OpenShift](https://www.instana.com/blog/automatic-root-cause-analysis-for-openshift-applications/).

          ## Instana Agent Operator

          This is the Kubernetes Operator for installing the Instana Agent on Kubernetes or OpenShift.

          ## Prerequisites for OpenShift

          Before the agent will be able to run in OpenShift, you need to perform a couple of extra configuration steps.

          You need to set up a project for the Instana Agent and configure it's permissions.

          The project you create here needs to be the namespace where you create the Instana Agent custom resource that the operator will use to deploy the agent.

          For example, create the `instana-agent` project:

              oc new-project instana-agent

          Then, ensure the `instana-agent` service account is in the privileged security context:

              oc adm policy add-scc-to-user privileged -z instana-agent

          The `instana-agent-remote` service account must be allowed to use any userid:

              oc adm policy add-scc-to-user anyuid -z instana-agent-remote -n instana-agent

          This service account will be created by the operator.

          Now you can proceed with installing the operator for the Instana agent.

          ## Installation and Configuration

          First, install this operator from [OperatorHub.io](https://operatorhub.io/), [OpenShift Container Platform](https://www.openshift.com/), or [OKD](https://www.okd.io/).

          Second, create a custom resource with the agent configuration in the target namespace (for now, this MUST always be the `instana-agent` namespace). The operator will pick up the custom resource and install the Instana agent accordingly.

          The following is a minimal template of the custom resource:

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgent
          metadata:
            name: instana-agent
            namespace: instana-agent
          spec:
            zone:
              name: my-zone # (optional) name of the zone of the host
            cluster:
              name: replace-me # replace with the name of your Kubernetes cluster
            agent:
              key: replace-me # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io # the monitoring ingress endpoint
              endpointPort: "443" # the monitoring ingress endpoint port, wrapped in quotes
              env:
                INSTANA_AGENT_TAGS: example
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://ibm.biz/monitoring-k8s
          ```

          Save the template in a file `instana-agent.yaml` and edit the following values:

          * If your target namespace is not `instana-agent`, replace the `namespace:` accordingly.
          * `agent.key` must be set with your Instana agent key.
          * `agent.endpointHost` must be set with the monitoring ingress endpoint, generally either `saas-us-west-2.instana.io` or `saas-eu-west-1.instana.io`.
          * `agent.endpointPort` must be set with the monitoring ingress port, generally "443" (wrapped in quotes).
          * `zone.name` should be set with the name of the Kubernetes cluster that is be displayed in Instana.

          For advanced configuration, you can edit the contents of the `configuration.yaml` file. View documentation [here](https://ibm.biz/monitoring-k8s).

          Apply the custom resource with `kubectl apply -f instana-agent.yaml`. After some time, you should see `instana-agent` Pods being created on each node of your cluster, and your cluster should show on the infrastructure map on your Instana Web interface.

          In addition to the InstanaAgent custom resource, a remote agent can be deployed. This agent will not monitor workload running inside the cluster, but can deploy a single replica of an agent which can get an independent configuration for running remote monitoring for certain sensors.

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgentRemote
          metadata:
            name: remote-agent
            namespace: "instana-agent"
          spec:
            zone:
              name: edited-zone # name of the zone of the host
            hostname:
              name: use-custom-hostname-for-remote-agent # (optional) give your remote agent pod a custom hostname
            agent:
              key: replace-key # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io
              endpointPort: "443"
              env: {}
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.
          ```

          ## Uninstalling

          In order to uninstall the Instana agent, simply remove the custom resource with `kubectl delete -f instana-agent.yaml`.

          ## Source Code

          The Instana agent operator is an open source project hosted on [https://github.com/instana/instana-agent-operator](https://github.com/instana/instana-agent-operator/).
        displayName: Instana Agent Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - APM
        - Instana
        links:
        - name: GitHub Repository
          url: https://github.com/instana/instana-agent-operator
        maintainers:
        - email: support@instana.com
          name: Instana
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: Instana
        relatedImages:
        - icr.io/instana/instana-agent-operator@sha256:ab8b35f0b45c68f69210d1b17c5dd609799f6bb0bc9f12d0d4e5302c910607a1
        - icr.io/instana/agent@sha256:ed71e6c3e5bdb155853a2db8c916460e6a867fb02627e3a332186188cbbd98dd
        version: 2.1.36
      entries:
      - name: instana-agent-operator.v2.1.36
        version: 2.1.36
      - name: instana-agent-operator.v2.1.35
        version: 2.1.35
      - name: instana-agent-operator.v2.1.34
        version: 2.1.34
      - name: instana-agent-operator.v2.1.33
        version: 2.1.33
      - name: instana-agent-operator.v2.1.32
        version: 2.1.32
      - name: instana-agent-operator.v2.1.31
        version: 2.1.31
      - name: instana-agent-operator.v2.1.30
        version: 2.1.30
      - name: instana-agent-operator.v2.1.28
        version: 2.1.28
      - name: instana-agent-operator.v2.1.27
        version: 2.1.27
      - name: instana-agent-operator.v2.1.24
        version: 2.1.24
      - name: instana-agent-operator.v2.1.23
        version: 2.1.23
      - name: instana-agent-operator.v2.1.22
        version: 2.1.22
      - name: instana-agent-operator.v2.1.21
        version: 2.1.21
      - name: instana-agent-operator.v2.1.20
        version: 2.1.20
      - name: instana-agent-operator.v2.1.19
        version: 2.1.19
      - name: instana-agent-operator.v2.1.18
        version: 2.1.18
      - name: instana-agent-operator.v2.1.17
        version: 2.1.17
      - name: instana-agent-operator.v2.1.16
        version: 2.1.16
      - name: instana-agent-operator.v2.1.15
        version: 2.1.15
      - name: instana-agent-operator.v2.1.14
        version: 2.1.14
      - name: instana-agent-operator.v2.1.13
        version: 2.1.13
      - name: instana-agent-operator.v2.1.12
        version: 2.1.12
      - name: instana-agent-operator.v2.1.11
        version: 2.1.11
      - name: instana-agent-operator.v2.1.10
        version: 2.1.10
      - name: instana-agent-operator.v2.1.9
        version: 2.1.9
      - name: instana-agent-operator.v2.1.8
        version: 2.1.8
      - name: instana-agent-operator.v2.1.7
        version: 2.1.7
      - name: instana-agent-operator.v2.1.6
        version: 2.1.6
      - name: instana-agent-operator.v2.1.5
        version: 2.1.5
      - name: instana-agent-operator.v2.1.4
        version: 2.1.4
      - name: instana-agent-operator.v2.1.3
        version: 2.1.3
      - name: instana-agent-operator.v2.1.2
        version: 2.1.2
      - name: instana-agent-operator.v2.1.1
        version: 2.1.1
      - name: instana-agent-operator.v2.1.0
        version: 2.1.0
      - name: instana-agent-operator.v2.0.32
        version: 2.0.32
      - name: instana-agent-operator.v2.0.31
        version: 2.0.31
      - name: instana-agent-operator-rhmp.v2.0.30
        version: 2.0.30
      - name: instana-agent-operator.v2.0.27
        version: 2.0.27
      - name: instana-agent-operator.v2.0.19
        version: 2.0.19
      - name: instana-agent-operator.v2.0.18
        version: 2.0.18
      - name: instana-agent-operator.v2.0.17
        version: 2.0.17
      - name: instana-agent-operator.v2.0.13
        version: 2.0.13
      - name: instana-agent-operator.v2.0.12
        version: 2.0.12
      - name: instana-agent-operator.v2.0.11
        version: 2.0.11
      - name: instana-agent-operator.v2.0.10
        version: 2.0.10
      - name: instana-agent-operator.v2.0.9
        version: 2.0.9
      - name: instana-agent-operator.v2.0.8
        version: 2.0.8
      - name: instana-agent-operator.v2.0.7
        version: 2.0.7
      - name: instana-agent-operator.v2.0.5
        version: 2.0.5
      - name: instana-agent-operator.v2.0.4
        version: 2.0.4
      name: stable
    defaultChannel: stable
    packageName: instana-agent-operator-rhmp
    provider:
      name: Instana
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: pelorus-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pelorus-operator.v0.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.pelorus.dora-metrics.io/v1alpha1",
                "kind": "Pelorus",
                "metadata": {
                  "name": "pelorus-sample"
                },
                "spec": {
                  "exporters": {
                    "global": {},
                    "instances": [
                      {
                        "app_name": "deploytime-exporter",
                        "exporter_type": "deploytime"
                      },
                      {
                        "app_name": "failuretime-exporter",
                        "enabled": false,
                        "exporter_type": "failure"
                      },
                      {
                        "app_name": "committime-exporter",
                        "exporter_type": "committime"
                      },
                      {
                        "app_name": "webhook-exporter",
                        "exporter_type": "webhook"
                      }
                    ]
                  },
                  "federate_openshift_monitoring": {
                    "enabled": false
                  },
                  "openshift_prometheus_basic_auth_pass": "changeme",
                  "openshift_prometheus_htpasswd_auth": "internal:{SHA}+pvrmeQCmtWmYVOZ57uuITVghrM=",
                  "prometheus_retention": "1y",
                  "prometheus_retention_size": "1GB",
                  "prometheus_storage": false,
                  "prometheus_storage_pvc_capacity": "2Gi",
                  "prometheus_storage_pvc_storageclass": "gp2"
                }
              }
            ]
          capabilities: Basic Install
          categories: |
            Modernization & Migration,Developer Tools,Monitoring,Integration & Delivery
          containerImage: quay.io/pelorus/pelorus-operator:0.0.9
          createdAt: "2024-05-27T20:04:01Z"
          description: |
            Tool that helps IT organizations measure their impact on the overall performance of their organization
          operatorframework.io/suggested-namespace: pelorus
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/dora-metrics/pelorus/
          support: Pelorus Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pelorus is the Schema for pelorus API to adopt instance to
              the requested workflow
            displayName: Pelorus
            kind: Pelorus
            name: peloruses.charts.pelorus.dora-metrics.io
            version: v1alpha1
        description: |
          Pelorus is a tool that helps IT organizations measure their impact on the overall performance of their organization. It does this by gathering metrics about team and organizational behaviors over time in some key areas of IT that have been shown to impact the value they deliver to the organization as a whole. Some of the key outcomes Pelorus can focus on are:

          - Software Delivery Performance
          - Product Quality and Sustainability
          - Customer experience

          For more background on the project you can read [@trevorquinn](https://github.com/trevorquinn)'s blog post on [Metrics Driven Transformation](https://www.openshift.com/blog/exploring-a-metrics-driven-approach-to-transformation).

          ## Software Delivery Performance as an outcome

          Currently, Pelorus functionality can capture proven metrics that measure Software Delivery Performance -- a significant outcome that IT organizations aim to deliver.

          Pelorus is a Grafana dashboard that can easily be deployed to an OpenShift cluster, and provides an organizational-level view of the [four critical measures of software delivery performance](https://blog.openshift.com/exploring-a-metrics-driven-approach-to-transformation/).

          ## Software Delivery Metrics Dashboard

          A short video describing each of metrics provided by Pelorus is available [here](https://www.youtube.com/watch?v=7-iB_KhUaQg).

          ## Demo

          [YouTube Video](https://www.youtube.com/watch?v=VPCOIfDcgso) with the Pelorus in action recorded during online Konveyor Community event.

          ## Documentation

          Pelorus documentation is available at [pelorus.readthedocs.io](https://pelorus.readthedocs.io/en/latest/).

          ## Contributing to Pelorus

          If you are interested in contributing to the Pelorus project, please review our Contribution guide which can be found in the [contribution guide](https://github.com/dora-metrics/pelorus/blob/master/CONTRIBUTING.md)

          ## Statement of Support

          Our support policy can be found in the [Upstream Support statement](https://github.com/dora-metrics/pelorus/blob/master/docs/UpstreamSupport.md)

          ## Code of Conduct
          Refer to dora-metrics's Code of Conduct [here](https://github.com/dora-metrics/pelorus/blob/master/CODE_OF_CONDUCT.md).

          ## License

          This repository is licensed under the terms of [Apache-2.0 License](https://raw.githubusercontent.com/dora-metrics/pelorus/master/LICENSE).
        displayName: Pelorus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pelorus-operator
        - dora
        - dora metrics
        - pelorus
        - metrics
        - transformation
        - devops
        links:
        - name: Pelorus Docs
          url: https://pelorus.readthedocs.io/en/latest/
        - name: Pelorus Getting Started
          url: https://pelorus.readthedocs.io/en/latest/GettingStarted/Overview/
        - name: Pelorus GIT repository
          url: https://github.com/dora-metrics/pelorus/
        maintainers:
        - email: kgranger@redhat.com
          name: Kevin Granger
        - email: whayutin@redhat.com
          name: Wesley Hayutin
        - email: mpryc@redhat.com
          name: Michal Pryc
        - email: msouzaol@redhat.com
          name: Mateus Souza Oliveira
        - email: esauer@redhat.com
          name: Eric Sauer
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - quay.io/pelorus/pelorus-operator:0.0.9
        version: 0.0.9
      entries:
      - name: pelorus-operator.v0.0.9
        version: 0.0.9
      - name: pelorus-operator.v0.0.8
        version: 0.0.8
      - name: pelorus-operator.v0.0.7
        version: 0.0.7
      - name: pelorus-operator.v0.0.6
        version: 0.0.6
      - name: pelorus-operator.v0.0.5
        version: 0.0.5
      - name: pelorus-operator.v0.0.4
        version: 0.0.4
      - name: pelorus-operator.v0.0.3
        version: 0.0.3
      - name: pelorus-operator.v0.0.2
        version: 0.0.2
      - name: pelorus-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: pelorus-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CodeFlare
      provider-url: https://github.com/project-codeflare
    name: codeflare-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: codeflare-operator.v1.16.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: quay.io/project-codeflare/codeflare-operator:v1.16.0
          createdAt: "2025-06-16T10:53:51Z"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/project-codeflare/codeflare-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AppWrapper
            name: appwrappers.workload.codeflare.dev
            version: v1beta2
        description: CodeFlare allows you to scale complex pipelines anywhere
        displayName: CodeFlare Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Pipelines
        - Scaling
        - App
        - Jobs
        links:
        - name: Codeflare Operator
          url: https://codeflare.dev/
        maintainers:
        - email: aasthana@redhat.com
          name: Anish Asthana
        - email: kpostlet@redhat.com
          name: Kevin Postlethwait
        - email: meyceoz@redhat.com
          name: Mustafa Eyceoz
        maturity: alpha
        provider:
          name: CodeFlare
          url: https://github.com/project-codeflare
        relatedImages:
        - quay.io/project-codeflare/codeflare-operator:v1.16.0
        version: 1.16.0
      entries:
      - name: codeflare-operator.v1.16.0
        version: 1.16.0
      - name: codeflare-operator.v1.15.0
        version: 1.15.0
      - name: codeflare-operator.v1.14.0
        version: 1.14.0
      - name: codeflare-operator.v1.13.0
        version: 1.13.0
      - name: codeflare-operator.v1.12.0
        version: 1.12.0
      - name: codeflare-operator.v1.11.0
        version: 1.11.0
      - name: codeflare-operator.v1.10.0
        version: 1.10.0
      - name: codeflare-operator.v1.9.0
        version: 1.9.0
      - name: codeflare-operator.v1.8.0
        version: 1.8.0
      - name: codeflare-operator.v1.7.0
        version: 1.7.0
      - name: codeflare-operator.v1.6.0
        version: 1.6.0
      - name: codeflare-operator.v1.5.0
        version: 1.5.0
      - name: codeflare-operator.v1.4.4
        version: 1.4.4
      - name: codeflare-operator.v1.4.3
        version: 1.4.3
      - name: codeflare-operator.v1.4.2
        version: 1.4.2
      - name: codeflare-operator.v1.4.1
        version: 1.4.1
      - name: codeflare-operator.v1.4.0
        version: 1.4.0
      - name: codeflare-operator.v1.3.1
        version: 1.3.1
      - name: codeflare-operator.v1.3.0
        version: 1.3.0
      - name: codeflare-operator.v1.2.0
        version: 1.2.0
      - name: codeflare-operator.v1.1.0
        version: 1.1.0
      - name: codeflare-operator.v1.0.1
        version: 1.0.1
      - name: codeflare-operator.v1.0.0
        version: 1.0.0
      - name: codeflare-operator.v1.0.0-rc.4
        version: 1.0.0-rc.4
      - name: codeflare-operator.v1.0.0-rc.3
        version: 1.0.0-rc.3
      - name: codeflare-operator.v1.0.0-rc.2
        version: 1.0.0-rc.2
      - name: codeflare-operator.v1.0.0-rc.1
        version: 1.0.0-rc.1
      - name: codeflare-operator.v0.2.3
        version: 0.2.3
      - name: codeflare-operator.v0.2.2
        version: 0.2.2
      - name: codeflare-operator.v0.2.1
        version: 0.2.1
      - name: codeflare-operator.v0.2.0
        version: 0.2.0
      - name: codeflare-operator.v0.1.0
        version: 0.1.0
      - name: codeflare-operator.v0.0.6
        version: 0.0.6
      - name: codeflare-operator.v0.0.5
        version: 0.0.5
      - name: codeflare-operator.v0.0.4
        version: 0.0.4
      - name: codeflare-operator.v0.0.3
        version: 0.0.3
      - name: codeflare-operator.v0.0.2
        version: 0.0.2
      - name: codeflare-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: codeflare-operator
    provider:
      name: CodeFlare
      url: https://github.com/project-codeflare
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: opdev.github.io
    name: airflow-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airflow-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "workflow.apache.org/v1alpha1",
                "kind": "AirFlow",
                "metadata": {
                  "name": "airflow-helm"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/airflow-helm-operator:0.0.1
          createdAt: "2023-06-28T16:57:21Z"
          description: An experimental operator that installs Apache Airflow.
          operatorframework.io/suggested-namespace: airflow-helm
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys Apache Airflow based on the official helm chart in
              openshift
            kind: AirFlow
            name: airflows.workflow.apache.org
            version: v1alpha1
        description: Deploys Apache Airflow based on the official helm chart with
          openshift specific modifications applied by default
        displayName: Airflow Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        - data
        - batch
        - scheduler
        - data workflow
        links:
        - name: Airflow Helm Operator
          url: https://airflow-helm-operator.domain
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: opdev.github.io
        relatedImages:
        - quay.io/opdev/airflow-helm-operator:0.0.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 0.0.1
      entries:
      - name: airflow-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: airflow-helm-operator
    provider:
      name: opdev
      url: opdev.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-sso
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-sso.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Sso",
                "metadata": {
                  "name": "sso-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "kcos-keycloak": {
                    "auth": {
                      "baseURL": "https://keycloak.kcos-sso.svc.cluster.local:8443",
                      "enabled": true,
                      "sessionCookie": "AT"
                    },
                    "authproxy": {
                      "image": "prod.harbor.keysight.digital/loadcore/kcos-auth-proxy@sha256:a777ee5116280fbeb1e1342445fef5135f9ca61e983a3e2821f9ca4b2038b044",
                      "initContainer": {
                        "image": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf"
                      }
                    },
                    "crsCleanupJob": {
                      "clientExec": "oc",
                      "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "realm": {
                      "accessTokenLifespan": 266400,
                      "accessTokenLifespanForImplicitFlow": 266400,
                      "admin": {
                        "password": "admin"
                      },
                      "offlineSessionIdleTimeout": 2592000,
                      "offlineSessionMaxLifespan": 5184000,
                      "registrationAllowed": true,
                      "ssoSessionIdleTimeout": 266400,
                      "ssoSessionMaxLifespan": 266400
                    }
                  },
                  "subcharts": {
                    "enable": {
                      "kcos-keycloak": true,
                      "keycloak-operator": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:29:57Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Sso
            name: ssoes.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS SSO Operator
        displayName: Keysight KCOS SSO Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Sso
          url: https://keysight-kcos-sso.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-sso@sha256:298417fac2fc053bdc352f77b59b14e39eadd66c871cfb547304a517489f294e
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - prod.harbor.keysight.digital/loadcore/kcos-auth-proxy@sha256:a777ee5116280fbeb1e1342445fef5135f9ca61e983a3e2821f9ca4b2038b044
        version: 0.1.21
      entries:
      - name: keysight-kcos-sso.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-sso
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-cpd-wspipelines
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-wspipelines.v10.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wspipelines.cpd.ibm.com/v1beta1",
                "kind": "WSPipelines",
                "metadata": {
                  "name": "wspipelines-cr",
                  "labels": {
                    "app.kubernetes.io/name": "wspipelines",
                    "app.kubernetes.io/instance": "wspipelines.wspipelines.cpd.ibm.com",
                    "app.kubernetes.io/managed-by": "wspipelines-operator",
                    "wspipelinesBuildNumber": "2017"
                }
                },
                "spec": {
                  "version": "5.1.3",
                  "scaleConfig": "small",
                  "license": {
                    "accept": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:f9f616d17f816468957b69bbb09e6c92d678acb4ba7a40f1cb493d08fb98deb8
          createdAt: "2025-02-11"
          description: WS Pipelines CSV file
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-wspipelines
          olm.skipRange: '>=1.0.0 <10.3.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM watsonx.ai"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation For additional details regarding install parameters
              check: https://ibm.biz/BdaaL9 or https://ibm.biz/BdaaLC. License By
              installing this product you accept the license terms https://ibm.biz/BdaaL9
              or https://ibm.biz/BdaaLC.'
            kind: WSPipelines
            name: wspipelines.wspipelines.cpd.ibm.com
            version: v1beta1
        description: |-
          # Name
          IBM&reg; Watson Pipelines
          # Introduction
          ## Summary
          The Watson Pipelines editor provides a graphical interface for orchestrating an end-to-end flow of assets from creation through deployment. It enables to automate an end-to-end flow to prepare data, then create, train, deploy, monitor and update machine learning models and Python scripts. To design a pipeline you drag nodes onto the canvas, specify objects and parameters, then run and monitor the pipeline.
          ## Features
          Watson Pipelines are built off of Kubeflow pipelines on Tekton runtime and are fully integrated into the Watson Studio platform, allowing users to combine tools including:
          - Notebooks
          - Data refinery flows
          - AutoAI experiments
          - Web service / online deployments
          - Batch deployments
          - Import and export of Project / Space assets
          into repeatable and automated pipelines.
          # Details
          ## Prerequisites
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-pre-installation-tasks)
          ### Resources Required
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=planning-system-requirements)
          # Installing
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing)
          ## Configuration
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing)
          ## Storage
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=planning-storage-considerations)
          ## Limitations
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=issues-watson-studio)
          ## Documentation
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=overview)
          ## PodSecurityPolicy Requirements
          Custom PodSecurityPolicy definition:
          ```
          Not Applicable
          ```
          ## SecurityContextConstraints Requirements
          Custom SecurityContextConstraints definition:
          ```
          apiVersion: security.openshift.io/v1
          kind: SecurityContextConstraints
          allowHostDirVolumePlugin: false
          allowHostIPC: false
          allowHostNetwork: false
          allowHostPID: false
          allowHostPorts: false
          allowPrivilegeEscalation: false
          allowPrivilegedContainer: false
          allowedCapabilities: null
          defaultAddCapabilities: null
          fsGroup:
            type: MustRunAs
          metadata:
            annotations:
              include.release.openshift.io/self-managed-high-availability: "true"
              kubernetes.io/description: restrict denies access to all host features and requires pods to be run with a UID, and SELinux context that are allocated to the namespace.  This is the most restrictive SCC and it is used by default for authenticated users.
              release.openshift.io/create-only: "true"
            name: restrict
          priority: null
          readOnlyRootFilesystem: false
          requiredDropCapabilities:
          - ALL
          runAsUser:
            type: MustRunAsRange
          seLinuxContext:
            type: MustRunAs
          supplementalGroups:
            type: RunAsAny
          users: []
          volumes:
          - configMap
          - downwardAPI
          - emptyDir
          - persistentVolumeClaim
          - projected
          - secret
          ```
        displayName: Watson Pipelines
        installModes:
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: OwnNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - wspipelines
        - WSPipelines
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: v1
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:a477d47eaf8f8e569601d3c23c1def66f31c098ac3f5a1ae350618c460917dd8
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:0164d8fbea591bbd9d67d84603162c3781b08551eddcc2a8a842aafc828af18c
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:164f6063f3a50a36e28a240ae7c060c4f90266b7ed799bfdc075f90628bea7ff
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:ebcfc8941330da6991480dc18ac34d6df4fe879e562ee537e5c36df73f4bc536
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a24f4cb635e8b5759a5e2bda20ddefc6d5ea9b16fb1081592885c54aefb5480c
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:29cd5c9cd60bbcfe211337786616622b911e2319db9870eb9d38d4f630c03e75
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:f0c68715f78bd20c352c97dd48ac5a912cad58e19a00159b5e6878f6beeb00fb
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:a2949c3868c2c91bf35e50c64b51fcc24372986744b55520e21e9ce31fde59f0
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-bundle@sha256:842f2ac568db4db79cf0450b413b4b09b5f2180ef3541385b46fb650ac7066c0
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:f9f616d17f816468957b69bbb09e6c92d678acb4ba7a40f1cb493d08fb98deb8
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:61155acdb8982648ae5748be9f5944ba913db1590ec3e773206d6fbb1d8952cb
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:8d7625a0ba656425f601b6ec9a398a6d30511fe1de119cf0e29f680c704ea225
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:e4888fadfadb3462e3a9ad0395a526a2631872f70caf291dc18b60269ded199f
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:48b7cfa31b9a185176c20afc119e37346a6f9380d74000edf6c729b4677ca44e
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:19bb32e0389361e8314e627bcad2e9bd93984289485cf632466834353d9c3449
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:e29885c1ee083a73b952b0de8d1cd675b2b2830abf44d69a8577f56972a601fb
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:00f8bcee4ba7e0c5cf5f0c52e9e7707df367041c9f4dd5508dbefe0dbdba8877
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:99d0b3192e6e46167423505f48655e7d49af1a461ae00b95c9c7c9d4209e5b65
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:25a4bb9cea1c3bdbeeeb205622512bcba98163f9629eba35625756346ce9d652
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:6fc976bc622c2165dd8bf93c9244e1c29b7c973a4da7a5ce4d0a097a1c1ea4fb
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:ad2215adfb8dd2dc600bf5097b7f88c257d5fbaf541409c01b36cdb67c130236
        version: 10.3.0
      entries:
      - name: ibm-cpd-wspipelines.v10.3.0
        version: 10.3.0
      name: v10.3
    defaultChannel: v10.3
    packageName: ibm-cpd-wspipelines
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-enterprise
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-enterprise.v1.33.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "my-replica-set"
                },
                "spec": {
                  "credentials": "my-credentials",
                  "members": 3,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "sample-sharded-cluster"
                },
                "spec": {
                  "configServerCount": 3,
                  "credentials": "my-credentials",
                  "mongodsPerShardCount": 3,
                  "mongosCount": 2,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "shardCount": 1,
                  "type": "ShardedCluster",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBMultiCluster",
                "metadata": {
                  "name": "multi-replica-set"
                },
                "spec": {
                  "clusterSpecList": [
                    {
                      "clusterName": "e2e.cluster1.mongokubernetes.com",
                      "members": 2
                    },
                    {
                      "clusterName": "e2e.cluster2.mongokubernetes.com",
                      "members": 1
                    },
                    {
                      "clusterName": "e2e.cluster3.mongokubernetes.com",
                      "members": 2
                    }
                  ],
                  "credentials": "my-credentials",
                  "duplicateServiceObjects": false,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBOpsManager",
                "metadata": {
                  "name": "ops-manager"
                },
                "spec": {
                  "adminCredentials": "ops-manager-admin",
                  "applicationDatabase": {
                    "members": 3,
                    "podSpec": {
                      "cpu": 1
                    }
                  },
                  "configuration": {
                    "mms.fromEmailAddr": "admin@thecompany.com"
                  },
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "version": "6.0.3"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBUser",
                "metadata": {
                  "name": "my-replica-set-x509-user"
                },
                "spec": {
                  "db": "$external",
                  "mongodbResourceRef": {
                    "name": "my-replica-set"
                  },
                  "roles": [
                    {
                      "db": "admin",
                      "name": "dbOwner"
                    }
                  ],
                  "username": "CN=my-replica-set-x509-user,OU=cloud,O=MongoDB,L=New York,ST=New York,C=US"
                },
                "status": {
                  "phase": "Updated"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/mongodb/mongodb-enterprise-operator-ubi@sha256:6d394a5247cc346c4421bd9b37948c0378970dd3f4fc5d6eaac681f4c9655957
          createdAt: "2025-04-15T11:32:47Z"
          description: The MongoDB Enterprise Kubernetes Operator enables easy deploys
            of MongoDB into Kubernetes clusters, using our management, monitoring
            and backup platforms, Ops Manager and Cloud Manager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mongodb/mongodb-enterprise-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MongoDB Deployment
            displayName: MongoDB Deployment
            kind: MongoDB
            name: mongodb.mongodb.com
            version: v1
          - description: MongoDB Multi Deployment
            displayName: MongoDB Multi Deployment
            kind: MongoDBMultiCluster
            name: mongodbmulticluster.mongodb.com
            version: v1
          - description: MongoDB x509 User
            displayName: MongoDB User
            kind: MongoDBUser
            name: mongodbusers.mongodb.com
            version: v1
          - description: MongoDB Ops Manager
            displayName: MongoDB Ops Manager
            kind: MongoDBOpsManager
            name: opsmanagers.mongodb.com
            version: v1
        description: |
          The MongoDB Enterprise Kubernetes Operator enables easy deploys of MongoDB
          into Kubernetes clusters, using our management, monitoring and backup
          platforms, Ops Manager and Cloud Manager.

          ## Before You Start

          To start using the operator you''ll need an account in MongoDB Cloud Manager or
          a MongoDB Ops Manager deployment.

          * [Create a Secret with your OpsManager API key](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-operator-credentials/#procedure)

          * [Create a ConfigMap with your OpsManager project ID and URL](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-project-using-configmap/)

          By installing this integration, you will be able to deploy MongoDB instances
          with a single simple command.

          ## Required Parameters

          * `opsManager` or `cloudManager` - Enter the name of the ConfigMap containing project information
          * `credentials` - Enter the name of the Secret containing your OpsManager credentials
          * `type` - Enter MongoDB Deployment Types ("Standalone", "ReplicaSet", "ShardedCluster"

          ## Supported MongoDB Deployment Types ##

          * Standalone: An instance of mongod that is running as a single server and
          not as part of a replica set, this is, it does not do any kind of
          replication.

          * Replica Set: A replica set in MongoDB is a group of mongod processes that
          maintain the same data set. Replica sets provide redundancy and high
          availability, and are the basis for all production deployments. This section
          introduces replication in MongoDB as well as the components and architecture
          of replica sets. The section also provides tutorials for common tasks
          related to replica sets.

          * Sharded Cluster: The set of nodes comprising a sharded MongoDB deployment.
          A sharded cluster consists of config servers, shards, and one or more mongos
          routing processes. Sharding is a A database architecture that partitions
          data by key ranges and distributes the data among two or more database
          instances. Sharding enables horizontal scaling.

          ## Requirements for deploying MongoDB OpsManager

          * In order to deploy resources of type MongoDB OpsManager, you will need to
          create a secret containing the [credentials](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/plan-om-resource/#om-rsrc-prereqs)
          for the Global Owner user

          ## Security ##

          The operator can enable TLS for all traffic between servers and also between
          clients and servers. Before enabling `security.tls.enabled` to `true` you
          should create your certificates.  or you can leave the operator to create all
          the certificates for you. The operator ability to create certs is been
          deprecated due to Kubernetes API changes.

          For more information, please read the official MongoDB
          Kubernetes Operator  [docs](https://docs.mongodb.com/kubernetes-operator/stable/).
        displayName: MongoDB Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mongodb
        - databases
        - nosql
        links:
        - name: Documentation
          url: https://docs.opsmanager.mongodb.com/current/tutorial/install-k8s-operator/index.html
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: stable
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-enterprise-server@sha256:fb8397cbcdbf7c33f6ad142492792771127f0d9a8c46388f350b6927405ee590
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa8683500e8d3632c4588bbcdddb4e3186a01f48aefe1b6d95f8860dae8b6c76
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:7d93aee462c60cacc4e5e66729f5131c5c0fa434e6e6400c8e2ccac4bd5e20ba
        - quay.io/mongodb/mongodb-agent-ubi@sha256:85f883076bc8edc516800704e2290c54b1bee14a9a533df84928c7ce1ae807d2
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0b2d01d543d38feddcc9d351e711df33a0a5187b22d95cdb926f6015ee1093bd
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ca2aaf57d70a0b9d743a6533c4e1b88324056197453de540d3352e679a3c9c92
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7e5361c2a1604dec3fc671d629806fe6ddc6df524b69a441b86933a4b30093
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b2bd8eb7f8b96adb874468ff1b0ebae2db4ebe490b6e1da4c9a6c5108f813ff5
        - quay.io/mongodb/mongodb-agent-ubi@sha256:766e6cbeec27f995211aa509ff7f118b6945f3e728141f28f9c4ea0dd2297af5
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3add8b68bfbaa42596fe5dbbee4623ba5ff1a826b18bb71301525adf574dcb94
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0fde236fe26d5b20210d6f246aadacd7845874712def516cbe2eec11d79d5181
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:dfec259209995bb040a1eff1fcafb1d6c3fdd145129f12c04a56ec74834b80a6
        - quay.io/mongodb/mongodb-enterprise-server@sha256:110a560f5a0ebe1fde1b2c20ae5ff5638682229d1c8656b993bdaf7b71eb94bc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:19925e1297ab652db59c35f53e7b7c8b2b277f8ecdc4215682961aaf2b3e924e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa300abf739e26350a258f643fa8e57aedacd041055c7b923bd22b21ef554cdb
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:83af86f8098e3c81b6e0879eabd582592ef6cbb1d6592f5e175c0794b9df2a0e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5421e00c0935b7a9ced51a11e935773ef5df4d259f3e1b1930fbf2f1b9d46586
        - quay.io/mongodb/mongodb-agent-ubi@sha256:38175af108178d89e4d3359ee2754534e5e9b24e843f12711fc14fe1dead516a
        - quay.io/mongodb/mongodb-agent-ubi@sha256:6ebe232ed0ea87045792e578ecb11dd46411333ad0ea61383c247704b4a0efe9
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:d3248506b6bd6b01c5973ab7bdac51ee41efa0ec8dd6211be4cc0b76a53a40c1
        - quay.io/mongodb/mongodb-enterprise-init-database-ubi@sha256:ab12a330e6b40735ee516168073887a4f37b58694bd10026a451eb9972e55500
        - quay.io/mongodb/mongodb-agent-ubi@sha256:5119a1ceeae6440dbe88b89550a7ae18de8e85c9a2059f467ad66adf7ec5d75f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:11ab324698f919f84a5dab5a964d93fe63f98a48146bb3b86974cba7606d30ec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c752a7f0726d21d55f204a359b85fa29cd727a5fe571b37289ae3ea52d6206b0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8f8735d377d214f64961381b5b1ed7cd3c306e6354dbc6b45c51134a25444b93
        - quay.io/mongodb/mongodb-enterprise-server@sha256:db63f6d93e03a3e644ef6ef4846ed7d24576c688ffb584226c78e1d1fba38878
        - quay.io/mongodb/mongodb-agent-ubi@sha256:575a43be4c667f3249f46f1fdcc06dd2e68cacc9eed989da0f6edd7e214d51b7
        - quay.io/mongodb/mongodb-agent-ubi@sha256:23d962d038888c6aba5540c8f5d81d7e86c4ccfaa07b9e159825d98676609f8e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0ee80032abe873b330537d4b64511ee76c5efb512af95287f28ddeffab13bb6d
        - quay.io/mongodb/mongodb-agent-ubi@sha256:8c11348eac180f5faa113db2804817302ec78847472e1891b063744adba4aa60
        - quay.io/mongodb/mongodb-agent-ubi@sha256:6d007271faed24bdf06eb64f2b54e5cdc3e11ce71840a7f586dabe2ee02c640e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:83db493ac6158d654d3ce55619f312d53cfbac258e0faf22e9110ca004e174bd
        - quay.io/mongodb/mongodb-agent-ubi@sha256:2395fe9d38a1a9160f7cb2f400b776ea3fc2e72187d23b6dee7fd1e4fb9ca721
        - quay.io/mongodb/mongodb-agent-ubi@sha256:66b49d9dce414dcd657c3550b26f71964b82edfb67a41a3ce8d2386b04ef07e8
        - quay.io/mongodb/mongodb-agent-ubi@sha256:d927afa9d3d569d943518f0a684e19afd4adb0875f55df7c904b9dd019589e5e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:c96ad084edbdc6a6d907f57795fa38f39199b730328dbc64997d56ea50d55dfc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:67238129df09d33ee4746ff1acb5515d699a29d9f42350eae3f3f69ccb7fbf54
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:f64d99a77d6a6b9ed10de9a28b1b8b8155414eca2e561f3fdb04bd66914fb3c5
        - quay.io/mongodb/mongodb-enterprise-server@sha256:15e31285c8949a4666f7702df53416b0905d76945a28575a6028a876cc8a6f0f
        - quay.io/mongodb/mongodb-agent-ubi@sha256:31cfefd2660da3efe1b6d63e39090464c28ae006d6ed1dffb2189a09183b9c61
        - quay.io/mongodb/mongodb-agent-ubi@sha256:97bb6bb5ecd5a802e8a91e69aa10d8bde0f648b106dc5591d49ee94209822daa
        - quay.io/mongodb/mongodb-agent-ubi@sha256:f5f3b14ee6533b21d89d9ed6d4af47dbdee9e071f909dc686dfcdf6742179e46
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:985f770667e6c4c083e0c4bb79a948b83efa0cdc4ed3012b1dbf72fe48b38986
        - quay.io/mongodb/mongodb-agent-ubi@sha256:9d1f21df008b14ca721c9ed0ff905af23e1a56b64cd4eddab4df1e8e9b63df8c
        - quay.io/mongodb/mongodb-enterprise-server@sha256:523566f44cbb6bbf0cd12b1adae939a883f2254c78dc4b8c9d0098581337b1db
        - quay.io/mongodb/mongodb-agent-ubi@sha256:2dbaaece4dd48087370e43d9bcf6b0a0312f86cb5dfe916e8725323bd5401992
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9a3a6496be0d2d446e6d654eb56386c7726e0cd2d7232f6dcec579923edf0014
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1d418c65d82ccad205b7889be4ba161d3cbfa8b1fb75b21c8fcf02de38fad9ca
        - quay.io/mongodb/mongodb-enterprise-server@sha256:172e338a43df01ee5408d8d1e3cb038d2c55e77761c1a2ac6ed37959d1525140
        - quay.io/mongodb/mongodb-enterprise-server@sha256:655896bac9e82de8e9729d59a36c3a31f66235f22b93449260a2ccb7709de084
        - quay.io/mongodb/mongodb-agent-ubi@sha256:695074fbbfcadf454230b80fbf1094260e631f6108cefb2050473edaee2a217a
        - quay.io/mongodb/mongodb-agent-ubi@sha256:b175972792aa1be268d5fb3cf3ad0156f4aecffa323e1b473360fc2a3a8d8891
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d876f57ea45582e84748b02459748e440af443a7eb0882585638973409c08f6a
        - quay.io/mongodb/mongodb-agent-ubi@sha256:411f85772e7a4b2c76e21d26d320b7743612080bf0cb4c7900d9a626cfeea33b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a48a4d5c275fd669a27469649fb39ea38c728dbcf9f0a9da4e7cf27a2ce912ed
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:2be0c6c89898b18618932b6a6303aeefdde60f48ec37d5c1fca5b8d04a00fd78
        - quay.io/mongodb/mongodb-agent-ubi@sha256:bd7edb8d0931dd952524f1f0ef99f61b0f69385d9e61efd034a4f45fd315a40b
        - quay.io/mongodb/mongodb-agent-ubi@sha256:ee150e969c0bb786969c7e0a45fb1812766b752aee539072e28648817826433b
        - quay.io/mongodb/mongodb-agent-ubi@sha256:7144ac5e5038e92a6d142bf39eba46554c1b5fafeb8ea59f0390896a0d35ed17
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:2127d7f2684ae36b123422f55237055aad02fdb71c4c21a5b8f0b0f6d8baedfc
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:c9a211a2c31b97eed3a25a814198547f2aa3470df93d661ca3700b828100d8de
        - quay.io/mongodb/mongodb-agent-ubi@sha256:c03e4a7663206f54c5ef39974502816f66d8d64936c19849f0337ea4547e614e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:35b263861d6ea66b541abc190ce1aaf4d4e36f128e5d8ba2a5e9fa8e25d84446
        - quay.io/mongodb/mongodb-enterprise-server@sha256:85c1d669c073caf9cbaac815e4c8ea2ca18b71deefa03bf84cd6d8b8ebd53e12
        - quay.io/mongodb/mongodb-enterprise-server@sha256:37af0913742b6de339c7e664b30f35dcef04506529e4a09a9bbbdd57643f7149
        - quay.io/mongodb/mongodb-agent-ubi@sha256:442f5d99d02674099525abaa69daa42cec758ce5554198f978927aa5c49006c8
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f1bae6375f846af6fca7faeeff4aedf9fab41a9d45b515ba74e9fd7336408617
        - quay.io/mongodb/mongodb-enterprise-server@sha256:605dcca06a45e5ea127edb9ff262a1c3ddafdab2f30b6d3a18f73bfb8a6acdb0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ef732e3eea91cc86caf2eabedc0c9115ed235a8d204f33e74bbf090a0b0259fb
        - quay.io/mongodb/mongodb-agent-ubi@sha256:0f8d102780a9d2ce54b7453f9575d0e24b15f014cd9b74816b1e6a4090b979b2
        - quay.io/mongodb/mongodb-enterprise-init-appdb-ubi@sha256:be4295843d92ce0f325b8daaeb3a5490c11772d0348cbd9ea35a6126d038b7fb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:33a30d11969eb21952681108092fe7f916e25c8e1de43e6a4e40e126025a2bec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7aa00dca44559318356627e88440765576dd6706187044edd1f21d0d0dcb0bf6
        - quay.io/mongodb/mongodb-enterprise-server@sha256:df6081325ea2fe01d1b27d0c70005dec8232aaf51a43d91661b722b8a1761263
        - quay.io/mongodb/mongodb-enterprise-server@sha256:dfd4277f0c16ed476ac7de51894739e8c8d13a65ea5024b9a97b3dce0f6c6975
        - quay.io/mongodb/mongodb-enterprise-server@sha256:584255582db31d0bb1f9593feb3b675d934d8750766fc98d3a1b3340feae4995
        - quay.io/mongodb/mongodb-enterprise-database-ubi@sha256:e2ca129f04ad6bf4e0dbb988badb083187725f28bac01bb7080cc80cad9ac316
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7885ab6e88214f9289d8babf5189136dde0dd4c84a15a5481ccd3ef512b2e66f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f2d9a3264a8814f33143b73c56624572fb80afeab184133d3a901713a31c1a87
        - quay.io/mongodb/mongodb-agent-ubi@sha256:03763de5018fd5fa14843444d24adfc5e387605970b531801ef2d0916233a010
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d21c65780cadc9d8668d2c28935d492625d6993f13ba562722200a819281cbcc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b00ac773623e1bf422b47e6d783263d2d08737865abe9abbe106d86d1ea25165
        - quay.io/mongodb/mongodb-enterprise-operator-ubi@sha256:6d394a5247cc346c4421bd9b37948c0378970dd3f4fc5d6eaac681f4c9655957
        - quay.io/mongodb/mongodb-agent-ubi@sha256:b23cfa639c5a4d20381245772dfe396ec915af3a3046df5d9b10016b900431f5
        - quay.io/mongodb/mongodb-enterprise-server@sha256:cd0cdf1b2a58f4c3b514e602cc1d96543410f631825045bbf65bd3fb84113f1f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8445d59d491a6d3890ea9c8b421214bcd0fdcfbc420ed40497a790b79f0bc89e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c54eeb910855594ed4bf5dcbbbcc14aa6f733e97313896a8381fb4be6aeac583
        - quay.io/mongodb/mongodb-agent-ubi@sha256:36c84115c6c645fd7d65d1004d2507ef3592aa1f07176645b1ee8e8b2a4f5f73
        - quay.io/mongodb/mongodb-enterprise-init-ops-manager-ubi@sha256:6ccc8148c266a9037ea744e8d19875146b1c231909f29a8f01c7ce415ca15936
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d49417ea85ee1ddb84429552a5aac87a82c2843904d8c0300a43b657b7a49e0e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:bd868284d9dfcb7c5a20e1160218d0172358384229b556c37e0789af0030eb24
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c07f20f8e7a5479e43cac7c74fc9f18cb4d4a1bd8b6b46f7dcc6f7ec7525fe28
        - quay.io/mongodb/mongodb-enterprise-server@sha256:13186ad45a2ba3ab55ed5c2a008eca494ee3937c8519475379476cc8e6f1d055
        - quay.io/mongodb/mongodb-agent-ubi@sha256:3b57a03dc17a301fda86c8afed1917c10d9b98947680116399c975ece944f5de
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d323c1bcd14118cc537ac282996ea5a58756d20b6b73f1eec0efa8710d246077
        - quay.io/mongodb/mongodb-agent-ubi@sha256:8467b1c2924a19929fc101ef2ff8d7e53cb0115a0cdc2e371dc0733bd0262395
        version: 1.33.0
      entries:
      - name: mongodb-enterprise.v1.33.0
        version: 1.33.0
      - name: mongodb-enterprise.v1.32.0
        version: 1.32.0
      - name: mongodb-enterprise.v1.31.0
        version: 1.31.0
      - name: mongodb-enterprise.v1.30.0
        version: 1.30.0
      - name: mongodb-enterprise.v1.29.0
        version: 1.29.0
      - name: mongodb-enterprise.v1.28.0
        version: 1.28.0
      - name: mongodb-enterprise.v1.27.0
        version: 1.27.0
      - name: mongodb-enterprise.v1.26.0
        version: 1.26.0
      - name: mongodb-enterprise.v1.25.0
        version: 1.25.0
      - name: mongodb-enterprise.v1.24.0
        version: 1.24.0
      - name: mongodb-enterprise.v1.23.0
        version: 1.23.0
      - name: mongodb-enterprise.v1.22.0
        version: 1.22.0
      - name: mongodb-enterprise.v1.21.0
        version: 1.21.0
      - name: mongodb-enterprise.v1.20.1
        version: 1.20.1
      - name: mongodb-enterprise.v1.19.1
        version: 1.19.1
      - name: mongodb-enterprise.v1.18.0
        version: 1.18.0
      - name: mongodb-enterprise.v1.17.2
        version: 1.17.2
      - name: mongodb-enterprise.v1.17.1
        version: 1.17.1
      - name: mongodb-enterprise.v1.17.0
        version: 1.17.0
      - name: mongodb-enterprise.v1.16.4
        version: 1.16.4
      - name: mongodb-enterprise.v1.16.3
        version: 1.16.3
      - name: mongodb-enterprise.v1.16.2
        version: 1.16.2
      - name: mongodb-enterprise.v1.16.1
        version: 1.16.1
      name: stable
    defaultChannel: stable
    packageName: mongodb-enterprise
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: manila-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: manila-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "manila.openstack.org/v1beta1",
                "kind": "Manila",
                "metadata": {
                  "name": "manila",
                  "namespace": "openstack"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "manila",
                  "databaseInstance": "openstack",
                  "manilaAPI": {},
                  "manilaScheduler": {},
                  "manilaShares": {
                    "share1": {}
                  },
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "osp-secret",
                  "serviceUser": "manila"
                }
              },
              {
                "apiVersion": "manila.openstack.org/v1beta1",
                "kind": "ManilaAPI",
                "metadata": {
                  "name": "manilaapi-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "manila.openstack.org/v1beta1",
                "kind": "ManilaScheduler",
                "metadata": {
                  "name": "manilascheduler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "manila.openstack.org/v1beta1",
                "kind": "ManilaShare",
                "metadata": {
                  "name": "manilashare-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:57Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManilaAPI is the Schema for the manilaapis API
            displayName: Manila API
            kind: ManilaAPI
            name: manilaapis.manila.openstack.org
            version: v1beta1
          - description: Manila is the Schema for the manilas API
            displayName: Manila
            kind: Manila
            name: manilas.manila.openstack.org
            version: v1beta1
          - description: ManilaScheduler is the Schema for the manilaschedulers API
            displayName: Manila Scheduler
            kind: ManilaScheduler
            name: manilaschedulers.manila.openstack.org
            version: v1beta1
          - description: ManilaShare is the Schema for the manilashares API
            displayName: Manila Share
            kind: ManilaShare
            name: manilashares.manila.openstack.org
            version: v1beta1
        description: Manila Operator
        displayName: Manila Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Share
        - Manila
        links:
        - name: Manila Operator
          url: https://github.com/openstack-k8s-operators/manila-operator
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/manila-rhel9-operator@sha256:1208676b94d8a726c277088591d898e08a3a8331a0125ba7bff101152afba934
        - registry.redhat.io/rhoso/openstack-manila-scheduler-rhel9@sha256:7a8f645126c073213ce066cdcac2164df51276f8aa4a3f89954664d3010e6fae
        - registry.redhat.io/rhoso/openstack-manila-share-rhel9@sha256:83a6d292785805a0ac2895810ea466cfb6e1cb7aa1a53ae8f4e22671381490df
        - registry.redhat.io/rhoso/openstack-manila-api-rhel9@sha256:4cbf725169614c2974d6d1a683cfc0179bfe06ca29cf30f7dca88caeb368c51a
        version: 1.0.6
      entries:
      - name: manila-operator.v1.0.6
        version: 1.0.6
      - name: manila-operator.v1.0.5
        version: 1.0.5
      - name: manila-operator.v1.0.4
        version: 1.0.4
      - name: manila-operator.v1.0.3
        version: 1.0.3
      - name: manila-operator.v1.0.2
        version: 1.0.2
      - name: manila-operator.v1.0.1
        version: 1.0.1
      - name: manila-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: manila-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: horizon-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: horizon-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "horizon.openstack.org/v1beta1",
                "kind": "Horizon",
                "metadata": {
                  "name": "horizon"
                },
                "spec": {
                  "customServiceConfig": "DEBUG = True\nLOGGING['handlers']['console']['level'] = 'DEBUG'\n",
                  "replicas": 1,
                  "secret": "osp-secret"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:00Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Horizon is the Schema for the horizons API
            displayName: Horizon
            kind: Horizon
            name: horizons.horizon.openstack.org
            version: v1beta1
        description: Horizon Operator
        displayName: Horizon Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - UI
        - Horizon
        links:
        - name: Horizon Operator
          url: https://github.com/openstack-k8s-operators/horizon-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-horizon-rhel9@sha256:b02c115d40aacc30e6763e295da90d69681b24fba20ebbc53133be852393b399
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/horizon-rhel9-operator@sha256:b76038b96a064c275721c1d45b95d7b0dfdb9fc626a2b3250d0adb1f97f89319
        version: 1.0.6
      entries:
      - name: horizon-operator.v1.0.6
        version: 1.0.6
      - name: horizon-operator.v1.0.5
        version: 1.0.5
      - name: horizon-operator.v1.0.4
        version: 1.0.4
      - name: horizon-operator.v1.0.3
        version: 1.0.3
      - name: horizon-operator.v1.0.2
        version: 1.0.2
      - name: horizon-operator.v1.0.1
        version: 1.0.1
      - name: horizon-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: horizon-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq-streams
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amqstreams.v2.2.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.3",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.3",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:f908d848b5e67de4b162597b190ef18d449ffd4b6540457b24b491445d5c5ea3
          createdAt: "2022-07-01 10:33:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.1.0-0 <2.2.2-2'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ", "Red Hat Application Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from AMQ Streams 1.7, migration to `v1beta2` needs to be completed for all AMQ Streams CRDs and CRs before the upgrade to 2.2 is done! If upgrading from AMQ Streams version earlier than 1.7, you need to first install the CRDs from AMQ Streams 1.7 and complete the migration to `v1beta2` for all AMQ Streams CRDs and CRs before the upgrade to 2.2 is done! For more details about the CRD upgrades, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.2/html/deploying_and_upgrading_amq_streams_on_openshift/assembly-upgrade-str#assembly-upgrade-resources-str).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.2/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: Red Hat Integration - AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        - registry.redhat.io/amq7/amq-streams-kafka-31-rhel8@sha256:d0607285fae95365b9d34b9474466824fb5ad012871204e8b13e52e438b66bf5
        - registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:f908d848b5e67de4b162597b190ef18d449ffd4b6540457b24b491445d5c5ea3
        - registry.redhat.io/amq7/amq-streams-kafka-32-rhel8@sha256:e3bbfff505f12c0360e175f4640c2dd9a7adcf3dce1ae2f586f2bde805ed8b1f
        - registry.redhat.io/amq7/amq-streams-bridge-rhel8@sha256:5d49bd23da40590e74422424b9457398bad89da8276718ff6a8a088e901c8345
        version: 2.2.2-2
      entries:
      - name: amqstreams.v2.2.2-2
        version: 2.2.2-2
      - name: amqstreams.v2.2.2-1
        version: 2.2.2-1
      - name: amqstreams.v2.2.2-0
        version: 2.2.2-0
      - name: amqstreams.v2.2.1-7
        version: 2.2.1-7
      - name: amqstreams.v2.2.1-6
        version: 2.2.1-6
      - name: amqstreams.v2.2.1-5
        version: 2.2.1-5
      - name: amqstreams.v2.2.1-4
        version: 2.2.1-4
      - name: amqstreams.v2.2.1-3
        version: 2.2.1-3
      - name: amqstreams.v2.2.1-2
        version: 2.2.1-2
      - name: amqstreams.v2.2.1-1
        version: 2.2.1-1
      - name: amqstreams.v2.2.1-0
        version: 2.2.1-0
      - name: amqstreams.v2.2.0-5
        version: 2.2.0-5
      name: amq-streams-2.2.x
    - currentCSV: amqstreams.v2.3.0-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.3.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.3.1",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:2e46a25144d697761f0119deaf528c6f1c2ccd04e2acd95c661b0df4f32e1a9f
          createdAt: "2022-10-31 17:05:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.2.0-0 <2.3.0-4'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.3.0 release, AMQ Streams supports only OCP 4.8 to 4.12!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 2.2 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more details, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.3/html/deploying_and_upgrading_amq_streams_on_openshift/assembly-upgrade-str#assembly-upgrade-resources-str

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.3/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-streams-kafka-32-rhel8@sha256:0152a85f2d0bc85881dacaee62c6e8b0b8b04f57a542b9b82dc93ad5303d6bf9
        - registry.redhat.io/amq7/amq-streams-kafka-33-rhel8@sha256:046c2539fef32bd928fc1f4f4c41e63f9a143b371fcc201d78c66f48fc34ea32
        - registry.redhat.io/amq7/amq-streams-bridge-rhel8@sha256:c822b3d87f7dea233fc523238d8d0516e42f9a3d252f50f4d427c934995f9b22
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        - registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:2e46a25144d697761f0119deaf528c6f1c2ccd04e2acd95c661b0df4f32e1a9f
        version: 2.3.0-4
      entries:
      - name: amqstreams.v2.3.0-4
        version: 2.3.0-4
      name: amq-streams-2.3.x
    - currentCSV: amqstreams.v2.4.0-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:cd1fefcd199400a504b4b9942c24ba6becafcecc952a9e39fa2729228be79e77
          createdAt: "2022-10-31 17:05:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.3.0-0 <2.4.0-3'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.4.0 release, AMQ Streams supports only OCP 4.8 to 4.12!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more details, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.4/html/deploying_and_managing_amq_streams_on_openshift/assembly-upgrade-str)

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.4/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:8d467fc746b78440df06f6d88d266f6ac9b5a4574a2b17fd870c173a107f8d65
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:cd1fefcd199400a504b4b9942c24ba6becafcecc952a9e39fa2729228be79e77
        - registry.redhat.io/amq-streams/kafka-33-rhel8@sha256:aa780a250444c2426be79e95366050a33f4a4c9d68c08b58743413ea67a2d6d7
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:50b3be789993cfeedc4923a8e760c1cf52498b2178be3abfa3719fc2d15ce723
        version: 2.4.0-3
      entries:
      - name: amqstreams.v2.4.0-3
        version: 2.4.0-3
      - name: amqstreams.v2.4.0-2
        version: 2.4.0-2
      - name: amqstreams.v2.4.0-1
        version: 2.4.0-1
      name: amq-streams-2.4.x
    - currentCSV: amqstreams.v2.5.2-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
          createdAt: "2023-07-10 21:47:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.2.0-0 <2.5.2-8'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.5.2 release, AMQ Streams supports only OCP 4.12 to 4.16!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more information, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:8c7e8ccb6c2d5da32b70d9d0c3919c303ad73a25447f6993d964f7b19cca1072
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:4ede5eeed159bddacfe830a768bab4996003d4c57a496fc8ec9b41df916d4acc
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:95bc35951eb545b918114af1e92a7139e2e3323b7d054dcb2a060843c7326d28
        version: 2.5.2-8
      entries:
      - name: amqstreams.v2.5.2-8
        version: 2.5.2-8
      - name: amqstreams.v2.5.2-7
        version: 2.5.2-7
      - name: amqstreams.v2.5.2-6
        version: 2.5.2-6
      - name: amqstreams.v2.5.2-5
        version: 2.5.2-5
      - name: amqstreams.v2.5.2-4
        version: 2.5.2-4
      - name: amqstreams.v2.5.2-3
        version: 2.5.2-3
      - name: amqstreams.v2.5.2-2
        version: 2.5.2-2
      - name: amqstreams.v2.5.2-1
        version: 2.5.2-1
      - name: amqstreams.v2.5.2-0
        version: 2.5.2-0
      - name: amqstreams.v2.5.1-5
        version: 2.5.1-5
      - name: amqstreams.v2.5.1-4
        version: 2.5.1-4
      - name: amqstreams.v2.5.1-3
        version: 2.5.1-3
      - name: amqstreams.v2.5.1-2
        version: 2.5.1-2
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.5.x
    - currentCSV: amqstreams.v2.6.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.6.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:dacf7ddffe7cbe2dc35e7d3290637b3334117e09dde986e6eaeaceded7069fbf
          createdAt: "2023-10-14 18:18:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.6.0-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: "**Red Hat AMQ Streams** is a massively scalable, distributed,
          and high performance data streaming platform based on the Apache Kafka®
          project. AMQ Streams provides an event streaming backbone that allows microservices
          and other application components to exchange data with extremely high throughput
          and low latency.\n**From the 2.6.0 release, AMQ Streams supports only OCP
          4.11 to 4.14! \n**The core capabilities include:** * A pub/sub messaging
          model, similar to a traditional enterprise messaging system, in which application
          components publish and consume events to/from an ordered stream\n* The long
          term, fault-tolerant storage of events\n* The ability for a consumer to
          replay streams of events\n* The ability to partition topics for horizontal
          scalability\n### Upgrades\n**!!! IMPORTANT !!!** **Direct upgrade from AMQ
          Stream 1.7 or earlier is not supported!** You have to upgrade first to one
          of the previous versions of AMQ Streams. You will also need to convert the
          CRD resources. For more information, see the [documentation for AMQ Streams
          on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).\n###
          Supported Features\n* **Manages the Kafka Cluster** - Deploys and manages
          all of the components of this complex application, including dependencies
          like Apache ZooKeeper® that are traditionally hard to administer.\n* **Includes
          Kafka Connect** - Allows for configuration of common data sources and sinks
          to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for mirroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashboards\n### Upgrading your Clusters\nThe
          AMQ Streams Operator understands how to run and upgrade between a set of
          Kafka versions. When specifying a new version in your config, check to make
          sure you aren't using any features that may have been removed. For more
          information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).\n###
          Storage\nAn efficient data storage infrastructure is essential to the optimal
          performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires
          block storage. The use of file storage (for example, NFS) is not recommended.\nThe
          AMQ Streams Operator supports three types of data storage:\n* Ephemeral
          (Recommended for development only!)\n* Persistent\n* JBOD (Just a Bunch
          of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nAMQ Streams
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation for the current release
          can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).\n"
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:dacf7ddffe7cbe2dc35e7d3290637b3334117e09dde986e6eaeaceded7069fbf
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:be7466ffd7f624f769eedca2e4106c597d2bf1d95cc9679f235e6dcdec66cfa9
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:fa8361e5089ab4449261f5a0ad57bd0bd059dc254b94bd9ab68340faae386947
        - registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:8b10bfb697b48ba3ed246fb846384f4cc67f05f670e9521edf0f47f829869404
        version: 2.6.0-2
      entries:
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.6.x
    - currentCSV: amqstreams.v2.7.0-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:671c1a2fcb0856cb208c0ead539fff9571fa2281a1e3d45b3df9ebd3df8c787e
          createdAt: "2024-03-09 21:27:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0-0 <2.7.0-7'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.6.0 release, Streams supports only OCP 4.12 to 4.14!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:671c1a2fcb0856cb208c0ead539fff9571fa2281a1e3d45b3df9ebd3df8c787e
        - registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:66362c800b78003d26df9471b7f21f5c9ad7fca8e8fd2511428dfc549cdf46c2
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:be9b82af13cbc120bb415abf4fcc97420e543172d6450b836c47344e4dcacd00
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:a7620559f3e5a0a4d5ecee44538619f698158847d9d0a06c7168404ed6f7cd36
        - registry.redhat.io/amq-streams/kafka-36-rhel9@sha256:8890c3217343d9aa3dd2ab9c7fa1742f960b5a26e09906581ce6075e61dc2888
        version: 2.7.0-7
      entries:
      - name: amqstreams.v2.7.0-7
        version: 2.7.0-7
      - name: amqstreams.v2.7.0-6
        version: 2.7.0-6
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.7.x
    - currentCSV: amqstreams.v2.8.0-0-0.1738265624.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:675d4acc375fb20e1a3fd5755184a318341c15f9337a11ea41e902fcda6d0bec
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.0-0 <2.8.0-0'
          olm.substitutesFor: amqstreams.v2.8.0-0
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.8.0 release, Streams supports only OCP 4.12 to 4.14!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:675d4acc375fb20e1a3fd5755184a318341c15f9337a11ea41e902fcda6d0bec
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:b46457a0a7f63518bc12b2789f75afb0e44496e62de918b9b584129c7078de86
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:4df906cf1490922a4587d96250f3be6ff605fe2caf87dfba7b812c144c5c5774
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:e771a84a458388c07ecf970ec625e4203c18e5d2bd91e7571d9c44ebb09313de
        - registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:17b2d255439f38c6ef54c0a56929ae636743b12c688d42223ac3fb58729e51ad
        version: 2.8.0-0+0.1738265624.p
      entries:
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.8.x
    - currentCSV: amqstreams.v2.9.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.9.2-1'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.9.0 release, Streams supports only OCP 4.14 to 4.18!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:f5985592b7f05c9b3c64caa0ac90f34f26b9b4277b342b6bd4a6415d2c563f9a
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:86ea2e823cd83c476521f2a1e9da5619c4810661c60d28295d0f5482c8aa519d
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:ddfb50ec2f1cf45ccccc632182bf0ff549c14969da35b7248802eab67d9c2e9d
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:67647bff1a76adb5b51f43c99d5f88a8aa38055c1e87190add8972bb621ddc15
        version: 2.9.2-1
      entries:
      - name: amqstreams.v2.9.2-1
        version: 2.9.2-1
      - name: amqstreams.v2.9.2-0
        version: 2.9.2-0
      - name: amqstreams.v2.9.1-2
        version: 2.9.1-2
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.9.x
    - currentCSV: amqstreams.v2.9.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.9.2-1'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.9.0 release, Streams supports only OCP 4.14 to 4.18!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:86ea2e823cd83c476521f2a1e9da5619c4810661c60d28295d0f5482c8aa519d
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:ddfb50ec2f1cf45ccccc632182bf0ff549c14969da35b7248802eab67d9c2e9d
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:67647bff1a76adb5b51f43c99d5f88a8aa38055c1e87190add8972bb621ddc15
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:f5985592b7f05c9b3c64caa0ac90f34f26b9b4277b342b6bd4a6415d2c563f9a
        version: 2.9.2-1
      entries:
      - name: amqstreams.v2.9.2-1
        version: 2.9.2-1
      - name: amqstreams.v2.9.2-0
        version: 2.9.2-0
      - name: amqstreams.v2.9.1-2
        version: 2.9.1-2
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.x
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-3.0.x
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-3.x
    - currentCSV: amqstreams.v2.5.2-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
          createdAt: "2023-07-10 21:47:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.2.0-0 <2.5.2-8'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.5.2 release, AMQ Streams supports only OCP 4.12 to 4.16!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more information, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:4ede5eeed159bddacfe830a768bab4996003d4c57a496fc8ec9b41df916d4acc
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:95bc35951eb545b918114af1e92a7139e2e3323b7d054dcb2a060843c7326d28
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:8c7e8ccb6c2d5da32b70d9d0c3919c303ad73a25447f6993d964f7b19cca1072
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
        version: 2.5.2-8
      entries:
      - name: amqstreams.v2.5.2-8
        version: 2.5.2-8
      - name: amqstreams.v2.5.2-7
        version: 2.5.2-7
      - name: amqstreams.v2.5.2-6
        version: 2.5.2-6
      - name: amqstreams.v2.5.2-5
        version: 2.5.2-5
      - name: amqstreams.v2.5.2-4
        version: 2.5.2-4
      - name: amqstreams.v2.5.2-3
        version: 2.5.2-3
      - name: amqstreams.v2.5.2-2
        version: 2.5.2-2
      - name: amqstreams.v2.5.2-1
        version: 2.5.2-1
      - name: amqstreams.v2.5.2-0
        version: 2.5.2-0
      - name: amqstreams.v2.5.1-5
        version: 2.5.1-5
      - name: amqstreams.v2.5.1-4
        version: 2.5.1-4
      - name: amqstreams.v2.5.1-3
        version: 2.5.1-3
      - name: amqstreams.v2.5.1-2
        version: 2.5.1-2
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-lts
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: stable
    defaultChannel: stable
    packageName: amq-streams
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JFrog
      provider-url: ""
    name: openshiftxray-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: xray-operator.v2.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftXray",
                "metadata": {
                  "name": "openshiftxray"
                },
                "spec": {
                  "xray": {
                    "common": {
                      "xrayUserId": "1000721035",
                      "xrayGroupId": "1000721035"
                    },
                    "analysis": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-analysis"
                      },
                      "name": "xray-analysis",
                      "podManagementPolicy": "Parallel",
                      "preStartCommand": null,
                      "updateStrategy": "RollingUpdate"
                    },
                    "database": {
                      "password": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "global": {
                      "postgresqlTlsSecret": null
                    },
                    "indexer": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-indexer"
                      },
                      "name": "xray-indexer",
                      "podManagementPolicy": "Parallel",
                      "updateStrategy": "RollingUpdate"
                    },
                    "persist": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-persist"
                      },
                      "name": "xray-persist",
                      "persistence": {
                        "size": "10Gi"
                      },
                      "podManagementPolicy": "Parallel",
                      "preStartCommand": null,
                      "updateStrategy": "RollingUpdate"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "rabbitmq": {
                      "enabled": true,
                      "replicaCount": 1,
                      "rbac": {
                        "create": true
                      },
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-rabbitmq"
                      },
                      "podSecurityContext": {
                        "fsGroup": 1000721001,
                        "runAsUser": 1000721001
                      },
                      "auth": {
                        "username": "xray",
                        "password": "xray",
                        "erlangCookie": "XRAYRABBITMQCLUSTER",
                        "tls": {
                          "enabled": false,
                          "failIfNoPeerCert": true,
                          "sslOptionsVerify": "verify_peer",
                          "caCertificate": "",
                          "serverCertificate": "",
                          "serverKey": ""
                        }
                      }
                    },
                    "rabbitmq-ha": {
                      "enabled": false,
                      "image": {
                        "repository": "registry.connect.redhat.com/jfrog/xray-rabbitmq"
                      },
                      "rabbitmqUsername": "guest",
                      "rabbitmqPassword": "guest",
                      "managementUsername": "guest",
                      "managementPassword": "guest",
                      "initContainer": {
                        "enabled": false
                      },
                      "securityContext": {
                        "fsGroup": 1000721035,
                        "runAsUser": 1000721035,
                        "runAsGroup": 1000721035
                      },
                      "replicaCount": 1
                    },
                    "replicaCount": 1,
                    "unifiedUpgradeAllowed":  "true",
                    "router": {
                      "image": {
                        "imagePullPolicy": "IfNotPresent",
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-router"
                      },
                      "name": "router"
                    },
                    "server": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-server"
                      },
                      "name": "xray-server",
                      "podManagementPolicy": "Parallel",
                      "replicaCount": 1,
                      "updateStrategy": "RollingUpdate"
                    },
                    "xray": {
                      "consoleLog": false,
                      "jfrogUrl": "OVERRIDE",
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/xray-operator@sha256:da7f6977170876b335875f4074ddfc0266b665351249295ce3c53f0e95891cc2
          createdAt: 2020-05-22 00:00:00+00:00
          description: JFrog Xray Enterprise deploys Xray continuous security scanner
            into Openshift (Requires Jfrog Artifactory)
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Xray Instances
            displayName: Xray
            kind: OpenshiftXray
            name: openshiftxrays.charts.helm.k8s.io
            version: v1
        description: "## Overview Openshift Operator to deploy JFrog Xray Continuous
          Security scanner into your Openshift cluster. NOTE: Artifactory is required
          for JFrog Xray to connect with and scan artifacts against.\n### Major version
          v2.0.4 Breaking change notification!!\nThis release has major bug fix for
          Rabbitmq Pod. Please update to version 2.0.4 of the operator.\n### Major
          version v2.0.0 Breaking change notification!! Xray is dependent upon Rabbitmq.
          The version of Rabbitmq that this chart uses has significantly changed from
          version 1.1.8 of the operator. For full details on how to migrate please
          visit: `https://github.com/jfrog/charts/blob/master/stable/xray/RABBITMQ_MIGRATION_NOTES.md`\nIf
          you are using operator version < 2.1.0 there is a issue with upgrade where
          xray version is not updated if you update operator version.As workaround
          you can update the Custom Resource OpenshiftXray which has hardcored image
          tag value.After upgrading to version 2.1.0 you can remove the image tags
          parameters from your CR so that future upgrade works without workaround.\n###
          Security Context Constraints To deploy this helm chart you will need to
          be a cluster admin w/ access to the anyuid scc and add the operator service
          account to the anyuid scc. ``` oc adm policy add-scc-to-user anyuid -z xray-operator
          ``` ``` oc adm policy add-scc-to-user anyuid -z openshiftxray ``` ``` oc
          adm policy add-scc-to-user anyuid -z openshiftxray-rabbitmq-ha ``` ## Usage
          An external DB is required. The operator will not deploy a DB but will require
          you to specify the configuration values to connect to it. Search for JFrog
          and click JFrog Xray Operator to install. Go to the Installed Operators.
          Wait for the JFrog Xray Operator to complete the installation. Open the
          Operator and click on the provided API: Xray Click Create New Instance and
          provide the following parameters for your DB configuration: ``` JFROG_URL
          DATABASE_URL DATABASE_USER DATABASE_PASSWORD ``` JFROG_URL is the external
          ip or DNS of your Artifactory to connect Xray to. Artifactory is required
          to use this operator. You can set your JFROG_URL to the service name of
          your Artifactory Nginx: ``` oc get svc -n my_namespace | grep nginx ```
          DATABASE_URL must be a Postgresql URL in the format: `` postgres://postgres-postgresql:5432/xraydb?sslmode=disable
          `` DATABASE_USER and DATABASE_PASSWORD must supply a valid user on Postgresql.
          Click Create for Xray to deploy into OpenShift. Open Artifactory in a web
          browser to complete the onboarding wizard for Xray! ### Air gap environments
          To use Xray Operator in an air gap environment you will need to download
          the images as image streams into your Openshift air gap cluster manually.
          Use the image overrides to then specify the image stream locations that
          are local to your cluster. Next you will need to setup the database of index
          data for xray to use when scanning artifacts. Follow the link below for
          instructions on setup: https://www.jfrog.com/confluence/display/JFROG/Configuring+Xray#ConfiguringXray-SynchronizingtheDatabase "
        displayName: JFrog Xray Continuous Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Xray Continuous Security
          url: https://jfrog.com/xray/
        - name: JFrog Xray Continuous Security Video
          url: https://youtu.be/wGzyKlQklKQ
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: alpha
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/xray-indexer@sha256:7f40d2dfe4a35b25d16dfcbfa04b4a58f227c132d4c3c86345e7b2bd19a97311
        - registry.connect.redhat.com/jfrog/xray-router@sha256:019235289e17d0e22dcd23a1505f71a78710827dfab338d8fa980f64c14f162f
        - registry.connect.redhat.com/jfrog/xray-rabbitmq@sha256:459d7d5e78192a6f45a5527e8fbaf1d8b30299b10ad4c42515d3d2bb5fcc75f4
        - registry.connect.redhat.com/jfrog/init@sha256:3f27e44b44c3556c50318110dbd542e34ef1061e28d7f0e0e7b8baa73c5f3e87
        - registry.connect.redhat.com/jfrog/xray-server@sha256:c5584c4e9528f0199e840b2b1ad1144220850ca9595ebf3e270bc6f86de4a71a
        - registry.connect.redhat.com/jfrog/xray-analysis@sha256:bcd8211f1bc753bfd7735165ce39b476e4efff198bffa877883620b41fcb33b5
        - registry.connect.redhat.com/jfrog/xray-persist@sha256:93fae0c124919a57d5162e32e88d3456ef6254889e454f89c21c89e69e6ae323
        - registry.connect.redhat.com/jfrog/xray-operator@sha256:da7f6977170876b335875f4074ddfc0266b665351249295ce3c53f0e95891cc2
        version: 2.1.6
      entries:
      - name: xray-operator.v2.1.6
        version: 2.1.6
      name: alpha
    defaultChannel: alpha
    packageName: openshiftxray-operator
    provider:
      name: JFrog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: eap
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eap-operator.v3.2.2
      currentCSVDesc:
        annotations:
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "false"
          containerImage: registry.redhat.io/jboss-eap/eap-rhel9-operator@sha256:2361e73b8c517177e80f60d204996ebf65b166e4e8b60b329323a2b409c54f6c
          createdAt: "2025-07-10T14:38:00Z"
          description: Operator that creates and manages Java applications running
            on JBoss EAP.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat JBoss Enterprise Application
            Platform", "Red Hat Runtimes", "Red Hat Integration", "Red Hat Process
            Automation"]'
          repository: https://github.com/wildfly/wildfly-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An application running on EAP application runtime.
            displayName: WildFlyServer
            kind: WildFlyServer
            name: wildflyservers.wildfly.org
            version: v1alpha1
        description: "# Red Hat JBoss Enterprise Application Platform (JBoss EAP)\n\nRed
          Hat® JBoss® Enterprise Application Platform (JBoss EAP) delivers enterprise-grade
          security, performance, and scalability in any environment.  Some features
          of JBoss EAP are:\n\n**Optimized for cloud and containers**: with simplified
          deployment and full Jakarta EE performance for applications in any environment.\n\n**Lightweight,
          flexible architecture**: which is well-suited for microservices and traditional
          applications.\n\n**Support and standardize microservices development**:
          With the JBoss Enterprise Application Platform expansion pack that supports
          Eclipse MicroProfile application programming interfaces (APIs).\n\n# JBoss
          EAP Operator\nOperators are software components that extend the Kubernetes
          API to customize how to manage applications and their components.\n\nThe
          JBoss EAP Operator facilitates the deployment and management of JBoss EAP
          on container orchestration platforms such as Red Hat OpenShift. The operator
          acts as an automated management tool, simplifying the process of provisioning,
          scaling, and updating JBoss EAP applications.\n\n# JBoss EAP Operator capabilities\nThe
          JBoss EAP Operator deploys and manages multiple instances of the JBoss EAP
          Java application in a Red Hat OpenShift cluster. \n\n**Simplifies management**:
          Creates, configures, manages, and seamlessly upgrades instances of complex
          stateful applications. No more manual configuration, making management a
          breeze.\n\n**Efficient scaling**: Effortlessly scales the instances of your
          JBoss EAP Java application based on demand.\n\n**Faster updates**: Seamlessly
          updates, enabling you to roll out new operator features and bug fixes without
          downtime\n\n**Safe transaction recovery**: Provides safe transaction recovery
          in your application cluster by verifying all transactions are completed
          before scaling down the replicas and marking a pod as clean for termination.\n\n#
          Operator installation and configuration\nLet's walk through the steps to
          install the JBoss EAP Operator on your OpenShift Cluster and deploy your
          first example application:\n\n**Step 1: Install JBoss EAP from OperatorHub**\n1.
          Click Install in the JBoss EAP operator window.\n2. In the Install Operator
          window, for Installation mode, select All namespaces on the cluster The
          operator is available in openshift-operators namespace.\n\n**Step 2: Build
          an application image**\n\nThe JBoss EAP Operator deploys an application
          image. If you do not have your JBoss EAP application image built yet, you
          can use the following example to build the image using Helm Charts.\n\n1.
          In the main navigation, click the dropdown menu (perspective switcher) and
          select Developer.\n2. Click Add > Helm Chart.\n3. Search for \"JBoss EAP
          8\" (Provided by Red Hat) and create a new Helm Chart release.\n4. On the
          Release name text box, type “eap-application”. \n5. Switch to the YAML view
          and replace the existing code with the following:\n\n    ```\n    build:\n
          \     uri: https://github.com/jboss-eap-up-and-running/eap8-getting-started\n
          \   deploy:\n      enabled: false\n    ```\n\n6. Click Create to build the
          application image.\n\n    The application image is created with the name
          eap-application:latest.\n\n**Step 3: Deploy your application with JBoss
          EAP Operator**\n\n1. In the main navigation, click the dropdown menu (perspective
          switcher) and select Developer.\n2. Click Add > Operator Backed > WildFlyServer
          > Create.\n3. Switch to the YAML view and replace the existing code with
          the following:\n\n    ```\n    apiVersion: wildfly.org/v1alpha1\n    kind:
          WildFlyServer\n    metadata:\n      name: eap-application\n    spec:\n      applicationImage:
          eap-application:latest\n      replicas: 1\n    ```\n\n 4. Click Create to
          deploy the application with JBoss EAP Operator.\n"
        displayName: JBoss EAP
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eap
        - java
        - open source
        - application runtime
        links:
        - name: JBoss EAP Homepage
          url: https://www.redhat.com/en/technologies/jboss-middleware/application-platform
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_jboss_enterprise_application_platform/
        - name: Blog
          url: https://developers.redhat.com/blog/
        - name: Quickstart
          url: https://github.com/jboss-developer/jboss-eap-quickstarts
        - name: Red Hat Customer Portal
          url: https://access.redhat.com/support
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat, Inc.
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/jboss-eap/eap-rhel9-operator@sha256:2361e73b8c517177e80f60d204996ebf65b166e4e8b60b329323a2b409c54f6c
        version: 3.2.2
      entries:
      - name: eap-operator.v3.2.2
        version: 3.2.2
      - name: eap-operator.v3.2.1
        version: 3.2.1
      - name: eap-operator.v3.2.0
        version: 3.2.0
      - name: eap-operator.v3.1.9
        version: 3.1.9
      - name: eap-operator.v3.1.8
        version: 3.1.8
      - name: eap-operator.v3.1.7
        version: 3.1.7
      - name: eap-operator.v3.1.6
        version: 3.1.6
      - name: eap-operator.v3.1.5
        version: 3.1.5
      - name: eap-operator.v3.1.4
        version: 3.1.4
      - name: eap-operator.v3.1.3
        version: 3.1.3
      - name: eap-operator.v3.1.2
        version: 3.1.2
      - name: eap-operator.v3.1.1
        version: 3.1.1
      - name: eap-operator.v3.1.0
        version: 3.1.0
      - name: eap-operator.v3.0.2
        version: 3.0.2
      - name: eap-operator.v3.0.1
        version: 3.0.1
      - name: eap-operator.v3.0.0
        version: 3.0.0
      - name: eap-operator.v2.4.3
        version: 2.4.3
      - name: eap-operator.v2.4.2
        version: 2.4.2
      - name: eap-operator.v2.4.1
        version: 2.4.1
      - name: eap-operator.v2.4.0
        version: 2.4.0
      - name: eap-operator.v2.3.10
        version: 2.3.10
      - name: eap-operator.v2.3.9
        version: 2.3.9
      - name: eap-operator.v2.3.8
        version: 2.3.8
      - name: eap-operator.v2.3.7
        version: 2.3.7
      - name: eap-operator.v2.3.6
        version: 2.3.6
      - name: eap-operator.v2.3.5
        version: 2.3.5
      - name: eap-operator.v2.3.4
        version: 2.3.4
      - name: eap-operator.v2.3.3
        version: 2.3.3
      - name: eap-operator.v2.3.2
        version: 2.3.2
      - name: eap-operator.v2.3.1
        version: 2.3.1
      - name: eap-operator.v2.3.0
        version: 2.3.0
      - name: eap-operator.v2.2.2
        version: 2.2.2
      - name: eap-operator.v2.2.1
        version: 2.2.1
      - name: eap-operator.v2.2.0
        version: 2.2.0
      name: stable
    defaultChannel: stable
    packageName: eap
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: m88i Labs
      provider-url: ""
    name: nexus-operator-m88i
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.m88i.io/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "name": "nexus3"
                },
                "spec": {
                  "networking": {
                    "expose": false
                  },
                  "persistence": {
                    "persistent": false
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    },
                    "requests": {
                      "cpu": "1",
                      "memory": "2Gi"
                    }
                  },
                  "useRedHatImage": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/m88i/nexus-operator:0.6.0
          createdAt: "2019-11-16T13:12:22Z"
          description: Nexus Operator to deploy and manage Nexus 3.x servers
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/m88i/nexus-operator
          support: m88i Labs
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: Nexus
            kind: Nexus
            name: nexus.apps.m88i.io
            version: v1alpha1
        description: |-
          Creates a new Nexus 3.x deployment in a Kubernetes cluster. Will help DevOps to have a quick Nexus application exposed to the world that can be used in a CI/CD process:

          * Deploys a new Nexus 3.x server based on either Community or Red Hat images
          * Creates an [Ingress](https://kubernetes.io/docs/concepts/services-networking/ingress/) in Kubernetes (1.14+) environments to expose the application to the world
          * On OpenShift, creates a Route to expose the service outside the cluster
          * Automatically creates Apache, Red Hat and JBoss Maven repositories
          * Automatically updates Nexus within the same minor version

          [See our documentation](https://github.com/m88i/nexus-operator/blob/main/README.md) for more installation and usage scenarios.

          If you experience any issues or have any ideas for new features, please [file an issue in our Github repository](https://github.com/m88i/nexus-operator/issues) or send an email to our maillist: [nexus-operator@googlegroups.com](mailto:nexus-operator@googlegroups.com)

          *Please note that the operator is an individual work and it's not provided nor supported by Sonatype.*
        displayName: Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nexus
        - sonatype
        - maven
        - docker
        - ci
        - continuous integration
        - continuous delivery
        - repository
        - repository manager
        - dev tools
        links:
        - name: Documentation
          url: https://github.com/m88i/nexus-operator/blob/main/README.md
        - name: Nexus Operator source code repository
          url: https://github.com/m88i/nexus-operator
        maintainers:
        - email: nexus-operator@googlegroups.com
          name: m88i Labs
        maturity: alpha
        provider:
          name: m88i Labs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/m88i/nexus-operator:0.6.0
        version: 0.6.0
      entries:
      - name: nexus-operator.v0.6.0
        version: 0.6.0
      name: alpha
    defaultChannel: alpha
    packageName: nexus-operator-m88i
    provider:
      name: m88i Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Aerospike
      provider-url: ""
    name: aerospike-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aerospike-kubernetes-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "memory-size": 3000000000,
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:6.3.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: docker.io/aerospike/aerospike-kubernetes-operator:2.5.0
          createdAt: 2023-05-04T09:32:47UTC
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Aerospike
        relatedImages:
        - docker.io/aerospike/aerospike-kubernetes-operator:2.5.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 2.5.0
      entries:
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      - name: aerospike-kubernetes-operator.v2.4.0
        version: 2.4.0
      - name: aerospike-kubernetes-operator.v2.3.0
        version: 2.3.0
      - name: aerospike-kubernetes-operator.v2.2.1
        version: 2.2.1
      - name: aerospike-kubernetes-operator.v2.2.0
        version: 2.2.0
      - name: aerospike-kubernetes-operator.v2.1.0
        version: 2.1.0
      name: alpha
    - currentCSV: aerospike-kubernetes-operator.v4.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "data-size": 1073741824,
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:8.1.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackup",
                "metadata": {
                  "name": "aerospikebackup",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "aerospike-cluster": {
                      "aerospike-aerospikebackup-test-cluster": {
                        "credentials": {
                          "password": "admin123",
                          "user": "admin"
                        },
                        "seed-nodes": [
                          {
                            "host-name": "aerocluster.aerospike.svc.cluster.local",
                            "port": 3000
                          }
                        ]
                      }
                    },
                    "backup-routines": {
                      "aerospike-aerospikebackup-test-routine": {
                        "backup-policy": "test-policy",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "local"
                      },
                      "aerospike-aerospikebackup-test-routine1": {
                        "backup-policy": "test-policy1",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "s3Storage"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackupService",
                "metadata": {
                  "name": "aerospikebackupservice-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "config": {
                    "backup-policies": {
                      "test-policy": {
                        "parallel": 3
                      },
                      "test-policy1": {
                        "parallel": 3
                      }
                    },
                    "service": {
                      "http": {
                        "port": 8081
                      }
                    },
                    "storage": {
                      "local": {
                        "local-storage": {
                          "path": "/tmp/localStorage"
                        }
                      },
                      "s3Storage": {
                        "s3-storage": {
                          "bucket": "aerospike-kubernetes-operator-test",
                          "s3-profile": "default",
                          "s3-region": "us-east-1"
                        }
                      }
                    }
                  },
                  "image": "aerospike/aerospike-backup-service:3.2.0",
                  "secrets": [
                    {
                      "secretName": "aws-secret",
                      "volumeMount": {
                        "mountPath": "/root/.aws/credentials",
                        "name": "aws-secret",
                        "subPath": "credentials"
                      }
                    }
                  ],
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeRestore",
                "metadata": {
                  "name": "aerospikerestore-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "backup-data-path": "aerospike-aerospikebackup-test-routine/backup/1733506015827/data/test",
                    "destination": {
                      "credentials": {
                        "password": "admin123",
                        "user": "admin"
                      },
                      "label": "destinationCluster",
                      "seed-nodes": [
                        {
                          "host-name": "aerocluster.test.svc.cluster.local",
                          "port": 3000
                        }
                      ]
                    },
                    "policy": {
                      "no-generation": true,
                      "no-indexes": true,
                      "parallel": 3
                    },
                    "source": {
                      "local-storage": {
                        "path": "/tmp/localStorage"
                      }
                    }
                  },
                  "type": "Full"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: docker.io/aerospike/aerospike-kubernetes-operator:4.1.1
          createdAt: "2025-09-15T10:14:05Z"
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          olm.skipRange: '>=4.0.0 <4.1.1'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
          support: Aerospike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeBackup is the Schema for the aerospikebackup API
            displayName: Aerospike Backup
            kind: AerospikeBackup
            name: aerospikebackups.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeBackupService is the Schema for the aerospikebackupservices
              API
            displayName: Aerospike Backup Service
            kind: AerospikeBackupService
            name: aerospikebackupservices.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1
          - description: AerospikeRestore is the Schema for the aerospikerestores
              API
            displayName: Aerospike Restore
            kind: AerospikeRestore
            name: aerospikerestores.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
           -   Aerospike cluster monitoring
           -   Backup and restore Aerospike clusters
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Aerospike
        relatedImages:
        - docker.io/aerospike/aerospike-kubernetes-operator:4.1.1
        version: 4.1.1
      entries:
      - name: aerospike-kubernetes-operator.v4.1.1
        version: 4.1.1
      - name: aerospike-kubernetes-operator.v4.1.0
        version: 4.1.0
      - name: aerospike-kubernetes-operator.v4.0.2
        version: 4.0.2
      - name: aerospike-kubernetes-operator.v4.0.1
        version: 4.0.1
      - name: aerospike-kubernetes-operator.v4.0.0
        version: 4.0.0
      - name: aerospike-kubernetes-operator.v3.4.1
        version: 3.4.1
      - name: aerospike-kubernetes-operator.v3.4.0
        version: 3.4.0
      - name: aerospike-kubernetes-operator.v3.3.1
        version: 3.3.1
      - name: aerospike-kubernetes-operator.v3.3.0
        version: 3.3.0
      - name: aerospike-kubernetes-operator.v3.2.2
        version: 3.2.2
      - name: aerospike-kubernetes-operator.v3.2.1
        version: 3.2.1
      - name: aerospike-kubernetes-operator.v3.2.0
        version: 3.2.0
      - name: aerospike-kubernetes-operator.v3.1.0
        version: 3.1.0
      - name: aerospike-kubernetes-operator.v3.0.0
        version: 3.0.0
      name: stable
    defaultChannel: stable
    packageName: aerospike-kubernetes-operator
    provider:
      name: Aerospike
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: 3scale-community-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: 3scale-community-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.8.2
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Improving [documentation](https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.8.2
        version: 0.8.2
      entries:
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.11
    - currentCSV: 3scale-community-operator.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.9.0
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.9.0/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.9.0/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.9.0
        version: 0.9.0
      entries:
      - name: 3scale-community-operator.v0.9.0
        version: 0.9.0
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.12
    - currentCSV: 3scale-community-operator.v0.10.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "ProductCRName": "product1-sample",
                  "Production": true
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.10.1
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.10.1/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Improving [documentation](https://github.com/3scale/3scale-operator/blob/master/doc/operator-user-guide.md)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.10.1/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.10.1
        version: 0.10.1
      entries:
      - name: 3scale-community-operator.v0.10.1
        version: 0.10.1
      - name: 3scale-community-operator.v0.9.0
        version: 0.9.0
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.13
    defaultChannel: threescale-2.13
    packageName: 3scale-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: https://www.intel.com
    name: ipm-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ipm-controller.v25.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ipm.intel.com/v1",
                "kind": "CPUPowerProv",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ipm-controller",
                    "node-label": "ipm-provisioning"
                  },
                  "name": "cpupowerprov-sample"
                },
                "spec": {
                  "cstates": {
                    "cores": {
                      "5": {
                        "C1": true,
                        "C1E": true,
                        "C6": false
                      },
                      "9,12": {
                        "C1": true,
                        "C6": true
                      }
                    },
                    "idleGovernor": "ladder"
                  },
                  "pstates": {
                    "cores": {
                      "5": {
                        "energyPerformanceBias": "normal",
                        "energyPerformancePreference": "performance",
                        "freqGovernor": "performance",
                        "maxFreqMHz": "3500",
                        "minFreqMHz": "800"
                      },
                      "9,12": {
                        "freqGovernor": "powersave",
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    },
                    "global": {
                      "conservative": {
                        "downThreshold": 20,
                        "freqStep": 5,
                        "ignoreNiceLoad": 0,
                        "samplingDownFactor": 1,
                        "samplingRate": 8888,
                        "upThreshold": 88
                      },
                      "maxPerfPct": 95,
                      "minPerfPct": 45,
                      "ondemand": {
                        "ignoreNiceLoad": 0,
                        "ioIsBusy": 1,
                        "powersaveBias": 0,
                        "samplingDownFactor": 1,
                        "samplingRate": 8888,
                        "upThreshold": 88
                      },
                      "schedutil": {
                        "rateLimitUs": 5000
                      },
                      "turbo": true
                    }
                  },
                  "uncoreFreqCPU": {
                    "cpu0": [
                      {
                        "maxFreqMHz": "1600",
                        "minFreqMHz": "800"
                      }
                    ],
                    "cpu1": [
                      {
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    ]
                  },
                  "uncoreFreqIO": {
                    "cpu0": [
                      {
                        "maxFreqMHz": "1600",
                        "minFreqMHz": "800"
                      }
                    ],
                    "cpu1": [
                      {
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/ipm-controller@sha256:32135c888f86a2c725d093ada9a59d222229bee5640d37f6e19a9e9812b4b67a
          createdAt: "2025-08-07T13:13:25Z"
          description: IPM Controller manages resources via REST API interactions.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.operatorframework.io/required-sccs: '["privileged"]'
          olm.securityContextConstraints.required: privileged
          operators.openshift.io/infrastructure-features: RestrictedNetwork
          operators.openshift.io/valid-subscription: Intel OBL Limited Distribution
            (Commercial Use) License Agreement [v2022.12.20]
          operators.operatorframework.io.bundle.package.v1: ipm-controller
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Intel® Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CPUPowerProv is the Schema for the cpupowerprovs API
            displayName: CPUPower Prov
            kind: CPUPowerProv
            name: cpupowerprovs.ipm.intel.com
            version: v1
        description: Infrastructure Power Manager (IPM) governs platform power technologies
          based on input from application workloads, the platform itself and customer
          input
        displayName: Intel® Infrastructure Power Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - intel
        - infrastructure
        - power
        - manager
        - management
        links:
        - name: IPM
          url: https://builders.intel.com/intel-technologies/power-management
        maintainers:
        - email: infrastructure.power.manager.operator@intel.com
          name: Infrastructure Power Manager Operator
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Intel® Corporation
          url: https://www.intel.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.13.0@sha256:471549d9cc07fcec0f285a05745b3d39a48e9a503b01b4047156a86ec1a30c00
        - registry.connect.redhat.com/intel/ipm-controller@sha256:32135c888f86a2c725d093ada9a59d222229bee5640d37f6e19a9e9812b4b67a
        - registry.connect.redhat.com/intel/ipm-agent@sha256:54de45cdc2e34768f6cdbf1cdee219406684143dddc4d236877709dc2536fe9c
        - registry.connect.redhat.com/intel/ipm-watcher@sha256:51eb773d5a4d1f401d4cbf2b42b8df4d0359beb85d5e274b8392dd7c5165e6ff
        version: 25.7.0
      entries:
      - name: ipm-controller.v25.7.0
        version: 25.7.0
      name: stable
    defaultChannel: stable
    packageName: ipm-controller
    provider:
      name: Intel® Corporation
      url: https://www.intel.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: ""
    name: nim-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nim-operator-certified.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMCache",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "source": {
                    "ngc": {
                      "modelPuller": "nvcr.io/nim/meta/llama-3.2-1b-instruct@sha256:1ff21ad32124d2382f8c733879394e7849cc3df5172651a24fcb671ac4fe77c6",
                      "pullSecret": "ngc-secret",
                      "authSecret": "ngc-api-secret",
                      "model": {
                        "lora": false,
                        "precision": "fp16",
                        "engine": "tensorrt_llm",
                        "qosProfile": "throughput",
                        "gpus": [
                          {
                            "product": "a100",
                            "ids": []
                          }
                        ]
                      }
                    }
                  },
                  "storage": {
                    "pvc": {
                      "create": true,
                      "storageClass": "local-path",
                      "size": "50Gi",
                      "volumeAccessMode": "ReadWriteOnce"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMService",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "image": {
                    "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                    "tag": "1.12.0",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [
                      "ngc-secret"
                    ]
                  },
                  "authSecret": "ngc-api-secret",
                  "storage": {
                    "nimCache": {
                      "name": "meta-llama3-1b-instruct",
                      "profile": ""
                    }
                  },
                  "scale": {
                    "enabled": false,
                    "hpa": {
                      "minReplicas": 1,
                      "maxReplicas": 10
                    }
                  },
                  "resources": {
                    "limits": {
                      "nvidia.com/gpu": 1
                    }
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMPipeline",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "k8s-nim-operator",
                    "app.kubernetes.io/managed-by": "k8s-nim-operator"
                  },
                  "name": "nimpipeline-sample"
                },
                "spec": {
                  "services": [
                    {
                      "name": "meta-llama3-1b-instruct",
                      "enabled": true,
                      "spec": {
                        "image": {
                          "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                          "tag": "1.12.0",
                          "pullPolicy": "IfNotPresent",
                          "pullSecrets": [
                            "ngc-secret"
                          ]
                        },
                        "authSecret": "ngc-api-secret",
                        "storage": {
                          "nimCache": {
                            "name": "meta-llama3-1b-instruct",
                            "profile": ""
                          }
                        },
                        "replicas": 1,
                        "resources": {
                          "limits": {
                            "nvidia.com/gpu": 1
                          }
                        },
                        "expose": {
                          "service": {
                            "type": "ClusterIP",
                            "port": 8000
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMBuild",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "nimCache": {
                    "name": "meta-llama3-1b-instruct",
                    "profile": "7cc8597690a35aba19a3636f35e7f1c7e7dbc005fe88ce9394cad4a4adeed414"
                  },
                  "image": {
                    "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                    "tag": "1.12.0",
                    "pullSecrets": [
                     "ngc-secret"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoCustomizer",
                "metadata": {
                  "name": "nemocustomizer-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "scheduler": {
                    "type": "volcano"
                  },
                  "wandb": {
                    "secretName": "wandb-secret",
                    "apiKeyKey": "apiKey",
                    "encryptionKey": "encryptionKey"
                  },
                  "otel": {
                    "enabled": true,
                    "exporterOtlpEndpoint": "http://customizer-otel-opentelemetry-collector.nemo.svc.cluster.local:4317"
                  },
                  "databaseConfig": {
                    "credentials": {
                      "user": "ncsuser",
                      "secretName": "customizer-pg-existing-secret",
                      "passwordKey": "password"
                    },
                    "host": "customizer-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "ncsdb"
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/customizer-api",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "entitystore": {
                    "endpoint": "http://nemoentitystore-sample.nemo.svc.cluster.local:8000"
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000"
                  },
                  "mlflow": {
                    "endpoint": "http://mlflow-tracking.nemo.svc.cluster.local:80"
                  },
                  "nemoDatastoreTools": {
                    "image": "nvcr.io/nvidia/nemo-microservices/nds-v2-huggingface-cli@sha256:7958e5244217de8af0ed723dfe7ea8e291b3afbb2a0aa73f5c27b218a1149b11"
                  },
                  "modelDownloadJobs": {
                    "image": "nvcr.io/nvidia/nemo-microservices/customizer-api@sha256:c590fa19121af2aac0379933deace3991f27c3ef88a8043560d6e3a172273b0b",
                    "ngcAPISecret": {
                      "name": "ngc-api-secret",
                      "key": "NGC_API_KEY"
                    },
                    "securityContext": {
                      "fsGroup": 1000,
                      "runAsNonRoot": true,
                      "runAsUser": 1000,
                      "runAsGroup": 1000
                    },
                    "ttlSecondsAfterFinished": 600,
                    "pollIntervalSeconds": 15
                  },
                  "modelConfig": {
                    "name": "nemo-model-config"
                  },
                  "trainingConfig": {
                    "configMap": {
                      "name": "nemo-training-config"
                    },
                    "modelPVC": {
                      "create": true,
                      "name": "finetuning-ms-models-pvc",
                      "storageClass": "",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": 50
                    },
                    "workspacePVC": {
                      "storageClass": "local-path",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": 10,
                      "mountPath": "/pvc/workspace"
                    },
                    "image": {
                      "repository": "nvcr.io/nvidia/nemo-microservices/customizer",
                      "tag": "25.08"
                    },
                    "env": [
                      { "name": "LOG_LEVEL", "value": "INFO" }
                    ],
                    "networkConfig": [
                      { "name": "NCCL_IB_SL", "value": "0" },
                      { "name": "NCCL_IB_TC", "value": "41" },
                      { "name": "NCCL_IB_QPS_PER_CONNECTION", "value": "4" },
                      { "name": "UCX_TLS", "value": "TCP" },
                      { "name": "UCX_NET_DEVICES", "value": "eth0" },
                      { "name": "HCOLL_ENABLE_MCAST_ALL", "value": "0" },
                      { "name": "NCCL_IB_GID_INDEX", "value": "3" }
                    ],
                    "ttlSecondsAfterFinished": 3600,
                    "timeout": 3600,
                    "tolerations": [
                      {
                        "key": "nvidia.com/gpu",
                        "operator": "Exists",
                        "effect": "NoSchedule"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoGuardrail",
                "metadata": {
                  "name": "nemoguardrails-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "configStore": {
                    "pvc": {
                      "name": "pvc-guardrail-config",
                      "create": true,
                      "storageClass": "",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": "1Gi"
                    }
                  },
                  "nimEndpoint": {
                    "baseURL": "http://meta-llama3-1b-instruct.nemo.svc.cluster.local:8000/v1"
                  },
                  "databaseConfig": {
                    "host": "guardrail-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "guardraildb",
                    "credentials": {
                      "user": "guardrailuser",
                      "secretName": "guardrail-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/guardrails",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [
                      "ngc-secret"
                    ]
                  },
                  "metrics": {
                    "serviceMonitor": {}
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "ephemeral-storage": "10Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoEvaluator",
                "metadata": {
                  "name": "nemoevaluator-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "evaluationImages": {
                    "bigcodeEvalHarness": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-bigcode@sha256:62f74c32d1ed07925ddb906de8975025bf63d73e82e7faefb840339aaf21a9dc",
                    "lmEvalHarness": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-lm-eval-harness@sha256:58e7c3da7f1edfed2ff508fa8abab7660366c2c17bb55d17aa68159d6e16be6d",
                    "similarityMetrics": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-custom-eval@sha256:eb51edf668481d1f5d2b100c850958dadf184fd8779433e45da40fa2a35ec4b9",
                    "llmAsJudge": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055",
                    "mtBench": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055",
                    "retriever": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-retriever@sha256:cb99fb3ea5b4e9989832679d4c6028573c28b99e43744e008ac5c9748b7dd250",
                    "rag": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-rag@sha256:b5e35cf277185e0901a4bbba35780b3673be71ae2c5e8efcdbab7710a8a6f783",
                    "bfcl": "nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-bfcl@sha256:95f45f66140d481d544b36b16bc3de7ad5e090d76225fd73ed70461b9bd9907c",
                    "agenticEval": "nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-agentic-eval@sha256:b38bbc5e61df0930204efa096cf4f40db4655b92d03741190513b019d67fc7dc"
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/evaluator",
                    "tag": "25.06",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "argoWorkflows": {
                    "endpoint": "https://argo-workflows-server.nemo.svc.cluster.local:2746",
                    "serviceAccount": "argo-workflows-executor"
                  },
                  "vectorDB": {
                    "endpoint": "http://milvus.nemo.svc.cluster.local:19530"
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000/v1/hf"
                  },
                  "entitystore": {
                    "endpoint": "http://nemoentitystore-sample.nemo.svc.cluster.local:8000"
                  },
                  "databaseConfig": {
                    "host": "evaluator-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "evaldb",
                    "credentials": {
                      "user": "evaluser",
                      "secretName": "evaluator-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "otel": {
                    "enabled": true,
                    "exporterOtlpEndpoint": "http://evaluator-otel-opentelemetry-collector.nemo.svc.cluster.local:4317"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoEntitystore",
                "metadata": {
                  "name": "nemoentitystore-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/entity-store",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "databaseConfig": {
                    "databaseName": "nesdb",
                    "host": "entity-store-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "credentials": {
                      "user": "nesuser",
                      "secretName": "entity-store-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000"
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoDatastore",
                "metadata": {
                  "name": "nemodatastore-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "secrets": {
                    "datastoreConfigSecret": "nemo-ms-nemo-datastore",
                    "datastoreInitSecret": "nemo-ms-nemo-datastore-init",
                    "datastoreInlineConfigSecret": "nemo-ms-nemo-datastore-inline-config",
                    "giteaAdminSecret": "gitea-admin-credentials",
                    "lfsJwtSecret": "nemo-ms-nemo-datastore--lfs-jwt"
                  },
                  "databaseConfig": {
                    "credentials": {
                      "user": "ndsuser",
                      "secretName": "datastore-pg-existing-secret",
                      "passwordKey": "password"
                    },
                    "host": "datastore-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "ndsdb"
                  },
                  "pvc": {
                    "name": "pvc-shared-data",
                    "create": true,
                    "storageClass": "",
                    "volumeAccessMode": "ReadWriteOnce",
                    "size": "10Gi"
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/datastore",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "replicas": 1,
                  "resources": {
                    "requests": {
                      "memory": "256Mi",
                      "cpu": "500m"
                    },
                    "limits": {
                      "memory": "512Mi",
                      "cpu": "1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: nvcr.io/nvidia/cloud-native/k8s-nim-operator@sha256:f42dd5c3dc269553ee46725ed4b1ab2d2f253ef06117f1d9d52faaa22d8fe60d
          createdAt: Wed Sep 10 05:49:48 PDT 2025
          description: An Operator for deployment and maintenance of NVIDIA NIMs and
            NeMo microservices in a Kubernetes environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          k8sMinVersion: 1.26.0
          repository: github.com/NVIDIA/k8s-nim-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NIM Cache
            displayName: NIMCache
            kind: NIMCache
            name: nimcaches.apps.nvidia.com
            version: v1alpha1
          - description: NIM Service
            displayName: NIMService
            kind: NIMService
            name: nimservices.apps.nvidia.com
            version: v1alpha1
          - description: NIM Pipeline
            displayName: NIMPipeline
            kind: NIMPipeline
            name: nimpipelines.apps.nvidia.com
            version: v1alpha1
          - description: NIM Build
            displayName: NIMBuild
            kind: NIMBuild
            name: nimbuilds.apps.nvidia.com
            version: v1alpha1
          - description: NEMO guardrail Service
            displayName: NemoGuardrail
            kind: NemoGuardrail
            name: nemoguardrails.apps.nvidia.com
            version: v1alpha1
          - description: NEMO datastore Service
            displayName: NemoDatastore
            kind: NemoDatastore
            name: nemodatastores.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Eval Service
            displayName: NemoEvaluator
            kind: NemoEvaluator
            name: nemoevaluators.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Customizer Service
            displayName: NemoCustomizer
            kind: NemoCustomizer
            name: nemocustomizers.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Entitystore Service
            displayName: NemoEntitystore
            kind: NemoEntitystore
            name: nemoentitystores.apps.nvidia.com
            version: v1alpha1
        description: |-
          ## NVIDIA NIM Operator
          The NVIDIA NIM Operator enables Kubernetes cluster administrators to operate the software components and services necessary to run NVIDIA NIMs in various domains such as reasoning, retrieval, speech, and biology. Additionally, it allows the use of NeMo Microservices to fine-tune, evaluate, or apply guardrails to your models.
          The Operator manages the life cycle of the following microservices and the models they use:
          ### NVIDIA NIM models, such as:
          * Reasoning LLMs
          * Retrieval - Embedding, Reranking, etc.
          * Speech
          * Biology
          ### NeMo core microservices:
          * NeMo Customizer
          * NeMo Evaluator
          * NeMo Guardrails
          ### NeMo platform component microservices:
          * NeMo Data Store
          * NeMo Entity Store
        displayName: The NVIDIA NIM Operator for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        links:
        - name: NVIDIA NIM
          url: https://docs.nvidia.com/nim/index.html
        - name: Documentation
          url: https://docs.nvidia.com/nim-operator/latest
        maintainers:
        - email: operator_feedback@nvidia.com
          name: shivamerla
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: NVIDIA
        relatedImages:
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-retriever@sha256:cb99fb3ea5b4e9989832679d4c6028573c28b99e43744e008ac5c9748b7dd250
        - nvcr.io/nvidia/nemo-microservices/customizer-api@sha256:c590fa19121af2aac0379933deace3991f27c3ef88a8043560d6e3a172273b0b
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-lm-eval-harness@sha256:58e7c3da7f1edfed2ff508fa8abab7660366c2c17bb55d17aa68159d6e16be6d
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-custom-eval@sha256:eb51edf668481d1f5d2b100c850958dadf184fd8779433e45da40fa2a35ec4b9
        - nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-bfcl@sha256:95f45f66140d481d544b36b16bc3de7ad5e090d76225fd73ed70461b9bd9907c
        - nvcr.io/nvidia/cloud-native/k8s-nim-operator@sha256:f42dd5c3dc269553ee46725ed4b1ab2d2f253ef06117f1d9d52faaa22d8fe60d
        - nvcr.io/nvidia/nemo-microservices/nds-v2-huggingface-cli@sha256:7958e5244217de8af0ed723dfe7ea8e291b3afbb2a0aa73f5c27b218a1149b11
        - nvcr.io/nim/meta/llama-3.2-1b-instruct@sha256:1ff21ad32124d2382f8c733879394e7849cc3df5172651a24fcb671ac4fe77c6
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-bigcode@sha256:62f74c32d1ed07925ddb906de8975025bf63d73e82e7faefb840339aaf21a9dc
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-rag@sha256:b5e35cf277185e0901a4bbba35780b3673be71ae2c5e8efcdbab7710a8a6f783
        - nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-agentic-eval@sha256:b38bbc5e61df0930204efa096cf4f40db4655b92d03741190513b019d67fc7dc
        version: 3.0.0
      entries:
      - name: nim-operator-certified.v3.0.0
        version: 3.0.0
      - name: nim-operator-certified.v2.0.2
        version: 2.0.2
      - name: nim-operator-certified.v2.0.1
        version: 2.0.1
      - name: nim-operator-certified.v2.0.0
        version: 2.0.0
      - name: nim-operator-certified.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: stable
    packageName: nim-operator-certified
    provider:
      name: NVIDIA
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-druid-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: druid-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:909a03d95f8173c418f6547092421a5531ff8957aa4319ce5fc890f32fe1f7be
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:909a03d95f8173c418f6547092421a5531ff8957aa4319ce5fc890f32fe1f7be
        version: 24.11.1
      entries:
      - name: druid-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: druid-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:253239268b8d32dbea10233fe6ab202bbe8a074c369a929630681dea37935383
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:253239268b8d32dbea10233fe6ab202bbe8a074c369a929630681dea37935383
        version: 25.3.0
      entries:
      - name: druid-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: druid-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
        version: 25.7.0
      entries:
      - name: druid-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: druid-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
        version: 25.7.0
      entries:
      - name: druid-operator.v25.7.0
        version: 25.7.0
      - name: druid-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-druid-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: aiu-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aiu-operator.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "labels": {
                    "operator-version": "2.1.0-dev"
                  },
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "aiu": {
                    "strategy": "single"
                  },
                  "devicePlugin": {
                    "aiuSelector": {
                      "deviceID": "06a7",
                      "vendor": "1014"
                    },
                    "image": "aiu-device-plugin",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "deviceType": "vfio-pci",
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation"
                  ],
                  "featureDiscovery": {
                    "image": "aiu-feature-discovery",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "isRdma": false,
                  "loglevel": "info",
                  "metricsExporter": {
                    "enabled": true,
                    "image": "aiu-exporter",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "numVfs": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
          createdAt: "2024-10-07T13:07:51Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 2.1.1
      entries:
      - name: aiu-operator.v2.1.1
        version: 2.1.1
      name: candidate-2.1
    - currentCSV: aiu-operator.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "labels": {
                    "operator-version": "2.1.0-dev"
                  },
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "aiu": {
                    "strategy": "single"
                  },
                  "devicePlugin": {
                    "aiuSelector": {
                      "deviceID": "06a7",
                      "vendor": "1014"
                    },
                    "image": "aiu-device-plugin",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "deviceType": "vfio-pci",
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation"
                  ],
                  "featureDiscovery": {
                    "image": "aiu-feature-discovery",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "isRdma": false,
                  "loglevel": "info",
                  "metricsExporter": {
                    "enabled": true,
                    "image": "aiu-exporter",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "numVfs": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
          createdAt: "2024-10-07T13:07:51Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
        version: 2.1.1
      entries:
      - name: aiu-operator.v2.1.1
        version: 2.1.1
      name: fast-2.1
    - currentCSV: aiu-operator.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "labels": {
                    "operator-version": "2.1.0-dev"
                  },
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "aiu": {
                    "strategy": "single"
                  },
                  "devicePlugin": {
                    "aiuSelector": {
                      "deviceID": "06a7",
                      "vendor": "1014"
                    },
                    "image": "aiu-device-plugin",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "deviceType": "vfio-pci",
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation"
                  ],
                  "featureDiscovery": {
                    "image": "aiu-feature-discovery",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "isRdma": false,
                  "loglevel": "info",
                  "metricsExporter": {
                    "enabled": true,
                    "image": "aiu-exporter",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "numVfs": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
          createdAt: "2024-10-07T13:07:51Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 2.1.1
      entries:
      - name: aiu-operator.v2.1.1
        version: 2.1.1
      name: stable-2.1
    - currentCSV: aiu-operator.v2.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "labels": {
                    "operator-version": "2.3.3"
                  },
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "aiu-device-plugin",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.1"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "featureDiscovery": {
                    "image": "aiu-feature-discovery",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "aiu-exporter",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "aiu-webhook-validator",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "scheduler": {
                    "image": "aiu-kube-scheduler-ocp-4.16",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.2"
                  }
                }
              },
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuNodeState",
                "metadata": {
                  "labels": {
                    "operator-version": "2.3.3"
                  },
                  "name": "aiunodestate"
                },
                "spec": {
                  "nodeName": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2025_03/aiu-operator@sha256:2ea8268e94d44581b3490cd4bbbd5525359073129bae42c7f007c05727afe1c3
          createdAt: "2025-04-03T01:19:38Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["aiunodestates.aiu.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu/release_2025_03/aiu-operator@sha256:2ea8268e94d44581b3490cd4bbbd5525359073129bae42c7f007c05727afe1c3
        version: 2.3.3
      entries:
      - name: aiu-operator.v2.3.3
        version: 2.3.3
      name: stable-v2.3
    - currentCSV: aiu-operator.v2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "aiuFilter": ".",
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "release_2025_05"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "aiu-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "e2e_stable",
                      "repository": "icr.io/ibmaiu_internal/x86_64/dd2",
                      "version": "release_2025_05"
                    },
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "aiu-exporter",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "aiu-webhook-validator",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "scheduler": {
                    "image": "aiu-kube-scheduler-ocp-4.16",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  }
                }
              },
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2025_05/aiu-operator@sha256:9796ab1c39d1dd55f8aa79c9f585be0f988b5026745184bdc890dfd1846909df
          createdAt: "2025-05-28T05:53:45Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["aiunodestates.aiu.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu/release_2025_05/aiu-operator@sha256:9796ab1c39d1dd55f8aa79c9f585be0f988b5026745184bdc890dfd1846909df
        - icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453
        version: 2.4.0
      entries:
      - name: aiu-operator.v2.4.0
        version: 2.4.0
      name: stable-v2.4
    defaultChannel: stable-v2.4
    packageName: aiu-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lifecycle-agent
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lifecycle-agent.v4.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "ImageBasedUpgrade",
                "metadata": {
                  "name": "upgrade"
                },
                "spec": {
                  "autoRollbackOnFailure": {},
                  "extraManifests": [
                    {
                      "name": "sno-extramanifests",
                      "namespace": "openshift-lifecycle-agent"
                    }
                  ],
                  "oadpContent": [
                    {
                      "name": "oadp-cm-sno-backup",
                      "namespace": "openshift-adp"
                    }
                  ],
                  "seedImageRef": {
                    "image": "quay.io/xyz",
                    "version": "4.16.0"
                  },
                  "stage": "Idle"
                }
              },
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "SeedGenerator",
                "metadata": {
                  "name": "seedimage"
                },
                "spec": {
                  "seedImage": "quay.io/xyz"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:6c603c89401823086d7a02d7ad221da44d8c84000ba63dd30675ae0cbcacc95f
          description: The Lifecycle Agent for OpenShift provides local lifecycle
            management services for Single Node Openshift (SNO) clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.3'
          operatorframework.io/suggested-namespace: openshift-lifecycle-agent
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-lifecycle-agent",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/lifecycle-agent
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ImageBasedUpgrade is the Schema for the ImageBasedUpgrades
              API
            displayName: Image-based Cluster Upgrade
            kind: ImageBasedUpgrade
            name: imagebasedupgrades.lca.openshift.io
            version: v1
          - description: SeedGenerator is the Schema for the seedgenerators API
            displayName: Seed Generator
            kind: SeedGenerator
            name: seedgenerators.lca.openshift.io
            version: v1
        description: "# Lifecycle Agent for OpenShift\nThe Lifecycle Agent for OpenShift
          provides local lifecycle management services \nfor Single Node Openshift
          (SNO) clusters.\n\n## Where to find more information\nYou can find additional
          guidance in the [agent repository](https://github.com/openshift-kni/lifecycle-agent).\n"
        displayName: Lifecycle Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - upgrade
        links:
        - name: Lifecycle Agent
          url: https://lifecycle-agent.domain
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift4/recert-rhel9@sha256:c91def5d3da4e777d2555c354e39bf021d9f499993f86614969dee5ecc00fc49
        - registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:6c603c89401823086d7a02d7ad221da44d8c84000ba63dd30675ae0cbcacc95f
        version: 4.16.3
      entries:
      - name: lifecycle-agent.v4.16.3
        version: 4.16.3
      - name: lifecycle-agent.v4.16.2
        version: 4.16.2
      - name: lifecycle-agent.v4.16.1
        version: 4.16.1
      - name: lifecycle-agent.v4.16.0
        version: 4.16.0
      name: "4.16"
    - currentCSV: lifecycle-agent.v4.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "ImageBasedUpgrade",
                "metadata": {
                  "name": "upgrade"
                },
                "spec": {
                  "autoRollbackOnFailure": {},
                  "extraManifests": [
                    {
                      "name": "sno-extramanifests",
                      "namespace": "openshift-lifecycle-agent"
                    }
                  ],
                  "oadpContent": [
                    {
                      "name": "oadp-cm-sno-backup",
                      "namespace": "openshift-adp"
                    }
                  ],
                  "seedImageRef": {
                    "image": "quay.io/xyz",
                    "version": "4.16.0"
                  },
                  "stage": "Idle"
                }
              },
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "SeedGenerator",
                "metadata": {
                  "name": "seedimage"
                },
                "spec": {
                  "seedImage": "quay.io/xyz"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:6c603c89401823086d7a02d7ad221da44d8c84000ba63dd30675ae0cbcacc95f
          description: The Lifecycle Agent for OpenShift provides local lifecycle
            management services for Single Node Openshift (SNO) clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.3'
          operatorframework.io/suggested-namespace: openshift-lifecycle-agent
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-lifecycle-agent",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/lifecycle-agent
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ImageBasedUpgrade is the Schema for the ImageBasedUpgrades
              API
            displayName: Image-based Cluster Upgrade
            kind: ImageBasedUpgrade
            name: imagebasedupgrades.lca.openshift.io
            version: v1
          - description: SeedGenerator is the Schema for the seedgenerators API
            displayName: Seed Generator
            kind: SeedGenerator
            name: seedgenerators.lca.openshift.io
            version: v1
        description: "# Lifecycle Agent for OpenShift\nThe Lifecycle Agent for OpenShift
          provides local lifecycle management services \nfor Single Node Openshift
          (SNO) clusters.\n\n## Where to find more information\nYou can find additional
          guidance in the [agent repository](https://github.com/openshift-kni/lifecycle-agent).\n"
        displayName: Lifecycle Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - upgrade
        links:
        - name: Lifecycle Agent
          url: https://lifecycle-agent.domain
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:6c603c89401823086d7a02d7ad221da44d8c84000ba63dd30675ae0cbcacc95f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift4/recert-rhel9@sha256:c91def5d3da4e777d2555c354e39bf021d9f499993f86614969dee5ecc00fc49
        version: 4.16.3
      entries:
      - name: lifecycle-agent.v4.16.3
        version: 4.16.3
      - name: lifecycle-agent.v4.16.2
        version: 4.16.2
      - name: lifecycle-agent.v4.16.1
        version: 4.16.1
      - name: lifecycle-agent.v4.16.0
        version: 4.16.0
      name: stable
    defaultChannel: stable
    packageName: lifecycle-agent
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-opensearchservice-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-opensearchservice-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opensearchservice.services.k8s.aws/v1alpha1",
                "kind": "Domain",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/opensearchservice-controller:1.1.1
          createdAt: "2025-09-19T18:05:50Z"
          description: AWS OpenSearch Service controller is a service controller for
            managing OpenSearch Service resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Domain represents the state of an AWS opensearchservice Domain
              resource.
            displayName: Domain
            kind: Domain
            name: domains.opensearchservice.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage OpenSearch Service resources in AWS from within your Kubernetes cluster.

          **About Amazon OpenSearch Service**

          Amazon OpenSearch Service (successor to Amazon Elasticsearch Service) is a managed service that makes it easy to deploy, operate, and scale OpenSearch clusters in the AWS Cloud. Amazon OpenSearch Service supports OpenSearch and legacy Elasticsearch OSS. When you create a cluster, you have the option of which search engine to use. OpenSearch Service offers broad compatibility with Elasticsearch OSS 7.10, the final open source version of the software. For information about what changed with the recent service renaming, and the actions you might need to take, see [Amazon OpenSearch Service - Summary of changes](https://docs.aws.amazon.com/opensearch-service/latest/developerguide/rename.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon OpenSearch Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opensearchservice
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon OpenSearch Service Developer Resources
          url: https://aws.amazon.com/opensearch-service/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: opensearch service maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/opensearchservice-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-opensearchservice-controller.v1.1.1
        version: 1.1.1
      - name: ack-opensearchservice-controller.v1.1.0
        version: 1.1.0
      - name: ack-opensearchservice-controller.v1.0.15
        version: 1.0.15
      - name: ack-opensearchservice-controller.v1.0.12
        version: 1.0.12
      - name: ack-opensearchservice-controller.v1.0.10
        version: 1.0.10
      - name: ack-opensearchservice-controller.v1.0.9
        version: 1.0.9
      - name: ack-opensearchservice-controller.v1.0.8
        version: 1.0.8
      - name: ack-opensearchservice-controller.v1.0.7
        version: 1.0.7
      - name: ack-opensearchservice-controller.v1.0.6
        version: 1.0.6
      - name: ack-opensearchservice-controller.v1.0.5
        version: 1.0.5
      - name: ack-opensearchservice-controller.v1.0.4
        version: 1.0.4
      - name: ack-opensearchservice-controller.v1.0.2
        version: 1.0.2
      - name: ack-opensearchservice-controller.v1.0.1
        version: 1.0.1
      - name: ack-opensearchservice-controller.v1.0.0
        version: 1.0.0
      - name: ack-opensearchservice-controller.v0.0.28
        version: 0.0.28
      - name: ack-opensearchservice-controller.v0.0.27
        version: 0.0.27
      - name: ack-opensearchservice-controller.v0.0.26
        version: 0.0.26
      - name: ack-opensearchservice-controller.v0.0.24
        version: 0.0.24
      - name: ack-opensearchservice-controller.v0.0.23
        version: 0.0.23
      - name: ack-opensearchservice-controller.v0.0.19
        version: 0.0.19
      - name: ack-opensearchservice-controller.v0.0.17
        version: 0.0.17
      - name: ack-opensearchservice-controller.v0.0.15
        version: 0.0.15
      - name: ack-opensearchservice-controller.v0.0.14
        version: 0.0.14
      - name: ack-opensearchservice-controller.v0.0.13
        version: 0.0.13
      - name: ack-opensearchservice-controller.v0.0.12
        version: 0.0.12
      - name: ack-opensearchservice-controller.v0.0.11
        version: 0.0.11
      - name: ack-opensearchservice-controller.v0.0.10
        version: 0.0.10
      - name: ack-opensearchservice-controller.v0.0.9
        version: 0.0.9
      - name: ack-opensearchservice-controller.v0.0.8
        version: 0.0.8
      - name: ack-opensearchservice-controller.v0.0.7
        version: 0.0.7
      - name: ack-opensearchservice-controller.v0.0.6
        version: 0.0.6
      - name: ack-opensearchservice-controller.v0.0.5
        version: 0.0.5
      - name: ack-opensearchservice-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-opensearchservice-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
    name: flink-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flink-kubernetes-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"flink.apache.org/v1beta1",
                "kind":"FlinkDeployment",
                "metadata":{
                    "name":"basic-example"
                },
                "spec":{
                    "image":"flink:1.16",
                    "flinkVersion":"v1_16",
                    "flinkConfiguration":{
                        "taskmanager.numberOfTaskSlots":"2"
                    },
                    "serviceAccount":"flink",
                    "jobManager":{
                        "resource":{
                            "memory":"2048m",
                            "cpu":1
                        }
                    },
                    "taskManager":{
                        "resource":{
                            "memory":"2048m",
                            "cpu":1
                        }
                    },
                    "job":{
                        "jarURI":"local:///opt/flink/examples/streaming/StateMachineExample.jar",
                        "parallelism":2,
                        "upgradeMode":"stateless"
                    }
                }
              },
              {
                "apiVersion":"flink.apache.org/v1beta1",
                "kind":"FlinkSessionJob",
                "metadata":{
                  "name":"basic-session-job-example2"
                },
                "spec":{
                  "deploymentName":"basic-session-cluster",
                  "job":{
                      "jarURI":"https://repo1.maven.org/maven2/org/apache/flink/flink-examples-streaming_2.12/1.16.1/flink-examples-streaming_2.12-1.16.1.jar",
                      "parallelism":2,
                      "upgradeMode":"stateless",
                      "entryClass":"org.apache.flink.streaming.examples.statemachine.StateMachineExample"
                  }
                }
              },
              {
                "apiVersion": "flink.apache.org/v1beta1",
                "kind": "FlinkStateSnapshot",
                "metadata": {
                  "name": "example-savepoint"
                },
                "spec": {
                  "savepoint": {
                    "alreadyExists": true,
                    "disposeOnDelete": false,
                    "path": "/flink-data/savepoints/savepoint-98682a-21931bf10c39"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "basic-example": {
                "description": "Example flink deployment"
              },
              "basic-session-job-example2": {
                "description": "Example flink session job"
              }
            }
          capabilities: Deep Insights
          categories: Big Data, Streaming & Messaging
          containerImage: ghcr.io/apache/flink-kubernetes-operator:1.12.0
          createdAt: "2025-06-23 18:42:54"
          description: Flink Kubernetes Operator manages the complete deployment lifecycle
            of Apache Flink applications.
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/apache/flink-kubernetes-operator.git
          support: Apache Software Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The FlinkDeployment CR defines Flink Application and Session
              cluster deployments
            displayName: Flink Deployment
            kind: FlinkDeployment
            name: flinkdeployments.flink.apache.org
            version: v1beta1
          - description: The FlinkSessionJob CR defines the session job on the Session
              cluster and each Session cluster can run multiple FlinkSessionJob.
            displayName: Flink Session Job
            kind: FlinkSessionJob
            name: flinksessionjobs.flink.apache.org
            version: v1beta1
          - description: The FlinkStateSnapshots CR is used to describe savepoint
              or checkpoint for a Flink job.
            displayName: Flink State Snapshot
            kind: FlinkStateSnapshot
            name: flinkstatesnapshots.flink.apache.org
            version: v1beta1
        description: |-
          Flink Kubernetes Operator acts as a control plane to manage the complete deployment lifecycle of Apache Flink applications.
          ## Installation
          This operator can be installed from [OperatorHub.io](https://operatorhub.io/operator/flink-kubernetes-operator).

          By default, the Flink operator monitors and run Flink applications in the same namespace as the operator.
          To run Flink jobs in another namespace, users are responsible for setting up the RBAC in that namespace.

          For example, users need to do the following for each additional namespace that runs the Flink applications:

          1. Switch to the namespace by running:

              ```sh
              kubectl config set-context --current --namespace=CHANGEIT
              ```

          2. Create the service account, role, and role binding in the namespace using the commands below:

              ```sh
              kubectl apply -f - <<EOF
              apiVersion: v1
              kind: ServiceAccount
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink
              ---
              apiVersion: rbac.authorization.k8s.io/v1
              kind: Role
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink
              rules:
              - apiGroups:
                - ""
                resources:
                - pods
                - configmaps
                verbs:
                - '*'
              - apiGroups:
                - apps
                resources:
                - deployments
                - deployments/finalizers
                verbs:
                - '*'
              ---
              apiVersion: rbac.authorization.k8s.io/v1
              kind: RoleBinding
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink-role-binding
              roleRef:
                apiGroup: rbac.authorization.k8s.io
                kind: Role
                name: flink
              subjects:
              - kind: ServiceAccount
                name: flink
              EOF
              ```

          3. Proceed deploying Flink applications using the the Custom Resource Definition below.

              ```sh
              kubectl create -f https://raw.githubusercontent.com/apache/flink-kubernetes-operator/release-1.12.0/examples/basic.yaml
              ```

          See [Flink custom resources](https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/custom-resource/overview/#overview) for more detail.
        displayName: Flink Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flink
        - streaming
        links:
        - name: Website
          url: https://flink.apache.org/
        - name: Documentation
          url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/
        - name: Mailing list
          url: https://lists.apache.org/list.html?dev@flink.apache.org
        - name: Slack
          url: https://apache-flink.slack.com/join/shared_invite/zt-1llkzbgyt-K2nNGGg88rfsDGLkT09Qzg#/shared-invite/email
        - name: GitHub
          url: https://github.com/apache/flink-kubernetes-operator
        maintainers:
        - email: mbalassi@apache.org
          name: Marton Balassi
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: jbusche@us.ibm.com
          name: James Busche
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Community
          url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
        relatedImages:
        - ghcr.io/apache/flink-kubernetes-operator:1.12.0
        version: 1.12.0
      entries:
      - name: flink-kubernetes-operator.v1.12.0
        version: 1.12.0
      - name: flink-kubernetes-operator.v1.11.0
        version: 1.11.0
      - name: flink-kubernetes-operator.v1.10.0
        version: 1.10.0
      - name: flink-kubernetes-operator.v1.6.0
        version: 1.6.0
      - name: flink-kubernetes-operator.v1.5.0
        version: 1.5.0
      - name: flink-kubernetes-operator.v1.4.0
        version: 1.4.0
      - name: flink-kubernetes-operator.v1.3.1
        version: 1.3.1
      - name: flink-kubernetes-operator.v1.3.0
        version: 1.3.0
      name: alpha
    defaultChannel: alpha
    packageName: flink-kubernetes-operator
    provider:
      name: Community
      url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeRocketCI
      provider-url: https://docs.kuberocketci.io/
    name: edp-keycloak-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: edp-keycloak-operator.v1.19.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "keycloak-sample"
                },
                "spec": {
                  "secret": "my-keycloak-secret",
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakAuthFlow",
                "metadata": {
                  "name": "keycloakauthflow-sample"
                },
                "spec": {
                  "alias": "MyBrowser",
                  "authenticationExecutions": [
                    {
                      "authenticator": "auth-cookie",
                      "priority": 0,
                      "requirement": "ALTERNATIVE"
                    },
                    {
                      "authenticator": "identity-provider-redirector",
                      "authenticatorConfig": {
                        "alias": "my-alias",
                        "config": {
                          "defaultProvider": "my-alias"
                        }
                      },
                      "priority": 1,
                      "requirement": "REQUIRED"
                    }
                  ],
                  "builtIn": false,
                  "description": "browser with idp",
                  "providerId": "basic-flow",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "topLevel": true
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-sample"
                },
                "spec": {
                  "advancedProtocolMappers": true,
                  "clientId": "agocd",
                  "defaultClientScopes": [
                    "argocd_groups"
                  ],
                  "directAccess": true,
                  "public": false,
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-name:client-secret-key",
                  "webUrl": "https://argocd.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClientScope",
                "metadata": {
                  "name": "keycloakclientscope-sample"
                },
                "spec": {
                  "description": "Group Membership",
                  "name": "groups",
                  "protocol": "openid-connect",
                  "protocolMappers": [
                    {
                      "config": {
                        "access.token.claim": "true",
                        "claim.name": "groups",
                        "full.path": "false",
                        "id.token.claim": "true",
                        "userinfo.token.claim": "true"
                      },
                      "name": "groups",
                      "protocol": "openid-connect",
                      "protocolMapper": "oidc-group-membership-mapper"
                    }
                  ],
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "keycloakrealm-sample"
                },
                "spec": {
                  "id": "d1-id-kc-realm-name",
                  "keycloakRef": {
                    "kind": "Keycloak",
                    "name": "keycloak-sample"
                  },
                  "passwordPolicy": [
                    {
                      "type": "forceExpiredPasswordChange",
                      "value": "365"
                    },
                    {
                      "type": "length",
                      "value": "8"
                    }
                  ],
                  "realmEventConfig": {
                    "adminEventsDetailsEnabled": false,
                    "adminEventsEnabled": true,
                    "enabledEventTypes": [
                      "UPDATE_CONSENT_ERROR",
                      "CLIENT_LOGIN"
                    ],
                    "eventsEnabled": true,
                    "eventsExpiration": 15000,
                    "eventsListeners": [
                      "jboss-logging"
                    ]
                  },
                  "realmName": "d2-id-kc-realm-name",
                  "ssoAutoRedirectEnabled": false,
                  "ssoRealmEnabled": false,
                  "ssoRealmName": "openshift"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmComponent",
                "metadata": {
                  "name": "keycloakrealmcomponent-sample"
                },
                "spec": {
                  "config": {
                    "allowPasswordAuthentication": [
                      "true"
                    ],
                    "cachePolicy": [
                      "EVICT_WEEKLY"
                    ],
                    "debug": [
                      "true"
                    ],
                    "editMode": [
                      "READ_ONLY"
                    ],
                    "enabled": [
                      "true"
                    ],
                    "evictionDay": [
                      "3"
                    ],
                    "evictionHour": [
                      "5"
                    ],
                    "evictionMinute": [
                      "7"
                    ],
                    "kerberosRealm": [
                      "test-realm"
                    ],
                    "keyTab": [
                      "test-key-tab"
                    ],
                    "priority": [
                      "0"
                    ],
                    "serverPrincipal": [
                      "srv-principal-test"
                    ],
                    "updateProfileFirstLogin": [
                      "true"
                    ]
                  },
                  "name": "cr-kerb-test",
                  "providerId": "kerberos",
                  "providerType": "org.keycloak.storage.UserStorageProvider",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmGroup",
                "metadata": {
                  "name": "keycloakrealmgroup-sample"
                },
                "spec": {
                  "name": "ArgoCDAdmins",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "keycloakrealmidentityprovider-sample"
                },
                "spec": {
                  "alias": "instagram",
                  "authenticateByDefault": false,
                  "config": {
                    "clientId": "foo",
                    "clientSecret": "$secretName:secretKey",
                    "hideOnLoginPage": "true",
                    "syncMode": "IMPORT",
                    "useJwksUrl": "true"
                  },
                  "enabled": true,
                  "firstBrokerLoginFlowAlias": "first broker login",
                  "mappers": [
                    {
                      "config": {
                        "role": "role-tr",
                        "syncMode": "INHERIT"
                      },
                      "identityProviderAlias": "instagram",
                      "identityProviderMapper": "oidc-hardcoded-role-idp-mapper",
                      "name": "test3212"
                    },
                    {
                      "config": {
                        "attribute": "foo",
                        "attribute.value": "bar",
                        "syncMode": "IMPORT"
                      },
                      "identityProviderAlias": "instagram",
                      "identityProviderMapper": "hardcoded-attribute-idp-mapper",
                      "name": "test-33221"
                    }
                  ],
                  "providerId": "instagram",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRole",
                "metadata": {
                  "name": "keycloakrealmrole-sample"
                },
                "spec": {
                  "composite": true,
                  "description": "default developer role",
                  "name": "developer",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRoleBatch",
                "metadata": {
                  "name": "keycloakrealmrolebatch-sample"
                },
                "spec": {
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "roles": [
                    {
                      "composite": true,
                      "description": "default developer role",
                      "isDefault": false,
                      "name": "developer"
                    },
                    {
                      "composite": true,
                      "description": "default administrator role",
                      "isDefault": false,
                      "name": "administrator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmUser",
                "metadata": {
                  "name": "keycloakrealmuser-sample"
                },
                "spec": {
                  "attributes": {
                    "baz": "jazz",
                    "foo": "bar"
                  },
                  "email": "john.snow13@example.com",
                  "emailVerified": true,
                  "enabled": true,
                  "firstName": "John",
                  "keepResource": true,
                  "lastName": "Snow",
                  "password": "12345678",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "requiredUserActions": [
                    "UPDATE_PASSWORD"
                  ],
                  "username": "john.snow13"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloak",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloak-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloak",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloak-sample"
                },
                "spec": {
                  "secret": "keycloak-access",
                  "url": "https://keycloak.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloakRealm",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloakrealm-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloakrealm",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloakrealm-sample"
                },
                "spec": {
                  "clusterKeycloakRef": "clusterkeycloak-sample",
                  "realmName": "realm-sample"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          containerImage: docker.io/epamedp/keycloak-operator:1.19.0
          createdAt: "2023-11-15T16:04:45Z"
          description: An Operator for managing Keycloak
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-keycloak-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterKeycloakRealm is the Schema for the clusterkeycloakrealms
              API.
            displayName: Cluster Keycloak Realm
            kind: ClusterKeycloakRealm
            name: clusterkeycloakrealms.v1.edp.epam.com
            version: v1alpha1
          - description: ClusterKeycloak is the Schema for the clusterkeycloaks API.
            displayName: Cluster Keycloak
            kind: ClusterKeycloak
            name: clusterkeycloaks.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakAuthFlow is the Schema for the keycloak authentication
              flow API.
            displayName: Keycloak Auth Flow
            kind: KeycloakAuthFlow
            name: keycloakauthflows.v1.edp.epam.com
            version: v1
          - description: KeycloakClient is the Schema for the keycloak clients API.
            displayName: Keycloak Client
            kind: KeycloakClient
            name: keycloakclients.v1.edp.epam.com
            version: v1
          - description: KeycloakClientScope is the Schema for the keycloakclientscopes
              API.
            displayName: Keycloak Client Scope
            kind: KeycloakClientScope
            name: keycloakclientscopes.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmComponent is the Schema for the keycloak component
              API.
            displayName: Keycloak Realm Component
            kind: KeycloakRealmComponent
            name: keycloakrealmcomponents.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmGroup is the Schema for the keycloak group API.
            displayName: Keycloak Realm Group
            kind: KeycloakRealmGroup
            name: keycloakrealmgroups.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmIdentityProvider is the Schema for the keycloak
              realm identity provider API.
            displayName: Keycloak Realm Identity Provider
            kind: KeycloakRealmIdentityProvider
            name: keycloakrealmidentityproviders.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRoleBatch is the Schema for the keycloak roles
              API.
            displayName: Keycloak Realm Role Batch
            kind: KeycloakRealmRoleBatch
            name: keycloakrealmrolebatches.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRole is the Schema for the keycloak group API.
            displayName: Keycloak Realm Role
            kind: KeycloakRealmRole
            name: keycloakrealmroles.v1.edp.epam.com
            version: v1
          - description: KeycloakRealm is the Schema for the keycloak realms API.
            displayName: Keycloak Realm
            kind: KeycloakRealm
            name: keycloakrealms.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmUser is the Schema for the keycloak user API.
            displayName: Keycloak Realm User
            kind: KeycloakRealmUser
            name: keycloakrealmusers.v1.edp.epam.com
            version: v1
          - description: Keycloak is the Schema for the keycloaks API.
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.v1.edp.epam.com
            version: v1
        description: |
          Keycloak Operator is an operator that is responsible for establishing
          a connection to provided Keycloak Server, reconciling Keycloak entities (realms,
          roles, groups, users, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Keycloak `Master` realm, and assign a `create-realm` role.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  keycloak-access
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Keycloak` with Keycloak instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: Keycloak
              metadata:
                name: keycloak-sample
              spec:
                secret: keycloak-access             # Secret name
                url: https://keycloak.example.com   # Keycloak URL
              ```

              Wait for the `.status` field with  `status.connected: true`

          4. Create Keycloak realm and group using Custom Resources:

            ```yaml
            apiVersion: v1.edp.epam.com/v1
            kind: KeycloakRealm
            metadata:
              name: keycloakrealm-sample
            spec:
              realmName: realm-sample
              keycloakOwner: keycloak-sample   # the name of `kind: Keycloak`
              ```

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: KeycloakRealmGroup
              metadata:
                name: argocd-admins
              spec:
                name: ArgoCDAdmins
                realm: keycloakrealm-sample   # the name of `kind: KeycloakRealm`
              ```
        displayName: EDP Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - authentication
        - authorization
        - edp
        - idp
        - keycloak
        - oauth
        - oidc
        - operator
        - saml
        - sso
        links:
        - name: Edp Keycloak Operator
          url: https://github.com/epam/edp-keycloak-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/keycloak-operator:1.19.0
        version: 1.19.0
      entries:
      - name: edp-keycloak-operator.v1.19.0
        version: 1.19.0
      - name: edp-keycloak-operator.v1.18.2
        version: 1.18.2
      - name: edp-keycloak-operator.v1.17.1
        version: 1.17.1
      - name: edp-keycloak-operator.v1.13.0
        version: 1.13.0
      name: alpha
    - currentCSV: edp-keycloak-operator.v1.29.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "keycloak-sample"
                },
                "spec": {
                  "secret": "my-keycloak-secret",
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakAuthFlow",
                "metadata": {
                  "name": "keycloakauthflow-sample"
                },
                "spec": {
                  "alias": "MyBrowser",
                  "authenticationExecutions": [
                    {
                      "authenticator": "auth-cookie",
                      "priority": 0,
                      "requirement": "ALTERNATIVE"
                    },
                    {
                      "authenticator": "identity-provider-redirector",
                      "authenticatorConfig": {
                        "alias": "my-alias",
                        "config": {
                          "defaultProvider": "my-alias"
                        }
                      },
                      "priority": 1,
                      "requirement": "REQUIRED"
                    }
                  ],
                  "builtIn": false,
                  "description": "browser with idp",
                  "providerId": "basic-flow",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "topLevel": true
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-authorization-sample"
                },
                "spec": {
                  "adminFineGrainedPermissionsEnabled": true,
                  "authorization": {
                    "permissions": [
                      {
                        "decisionStrategy": "AFFIRMATIVE",
                        "description": "Resource permission",
                        "logic": "POSITIVE",
                        "name": "resource-permission",
                        "policies": [
                          "role-policy"
                        ],
                        "resources": [
                          "resource1"
                        ],
                        "type": "resource"
                      },
                      {
                        "decisionStrategy": "CONSENSUS",
                        "description": "Scope permission",
                        "logic": "POSITIVE",
                        "name": "scope-permission",
                        "policies": [
                          "role-policy"
                        ],
                        "scopes": [
                          "scope1"
                        ],
                        "type": "scope"
                      }
                    ],
                    "policies": [
                      {
                        "decisionStrategy": "AFFIRMATIVE",
                        "description": "Role policy",
                        "logic": "POSITIVE",
                        "name": "role-policy",
                        "rolePolicy": {
                          "roles": [
                            {
                              "name": "developer",
                              "required": true
                            }
                          ]
                        },
                        "type": "role"
                      },
                      {
                        "aggregatedPolicy": {
                          "policies": [
                            "policy1",
                            "policy2"
                          ]
                        },
                        "description": "Aggregate policy",
                        "name": "aggregate-policy",
                        "type": "aggregate"
                      },
                      {
                        "clientPolicy": {
                          "clients": [
                            "client1",
                            "client2"
                          ]
                        },
                        "description": "Client policy",
                        "name": "client-policy",
                        "type": "client"
                      },
                      {
                        "description": "Group policy",
                        "groupPolicy": {
                          "groups": [
                            {
                              "extendChildren": true,
                              "name": "group1"
                            }
                          ]
                        },
                        "name": "group-policy",
                        "type": "group"
                      },
                      {
                        "description": "Role policy",
                        "name": "role-policy",
                        "rolePolicy": {
                          "roles": [
                            {
                              "name": "developer",
                              "required": true
                            }
                          ]
                        },
                        "type": "role"
                      },
                      {
                        "description": "Time policy",
                        "name": "time-policy",
                        "timePolicy": {
                          "notBefore": "2021-01-01T00:00:00Z",
                          "notOnOrAfter": "2021-12-31T23:59:59Z"
                        },
                        "type": "time"
                      },
                      {
                        "description": "User policy",
                        "name": "user-policy",
                        "type": "user",
                        "userPolicy": {
                          "users": [
                            "user1",
                            "user2"
                          ]
                        }
                      }
                    ],
                    "resources": [
                      {
                        "displayName": "Resource 1",
                        "iconUri": "https://example.com/icon.png",
                        "name": "resource1",
                        "scopes": [
                          "scope1"
                        ],
                        "type": "test"
                      }
                    ],
                    "scopes": [
                      "scope1"
                    ]
                  },
                  "authorizationServicesEnabled": true,
                  "clientId": "authorization-sample",
                  "directAccess": true,
                  "permission": {
                    "scopePermissions": [
                      {
                        "name": "token-exchange",
                        "policies": [
                          "policy1"
                        ]
                      }
                    ]
                  },
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-authorization-sample:client-secret-key",
                  "serviceAccount": {
                    "enabled": true
                  },
                  "webUrl": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-sample"
                },
                "spec": {
                  "advancedProtocolMappers": true,
                  "authenticationFlowBindingOverrides": {
                    "browser": "browser",
                    "directGrant": "direct grant"
                  },
                  "clientId": "agocd",
                  "defaultClientScopes": [
                    "argocd_groups"
                  ],
                  "directAccess": true,
                  "public": false,
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-name:client-secret-key",
                  "webUrl": "https://argocd.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClientScope",
                "metadata": {
                  "name": "keycloakclientscope-sample"
                },
                "spec": {
                  "description": "Group Membership",
                  "name": "groups",
                  "protocol": "openid-connect",
                  "protocolMappers": [
                    {
                      "config": {
                        "access.token.claim": "true",
                        "claim.name": "groups",
                        "full.path": "false",
                        "id.token.claim": "true",
                        "userinfo.token.claim": "true"
                      },
                      "name": "groups",
                      "protocol": "openid-connect",
                      "protocolMapper": "oidc-group-membership-mapper"
                    }
                  ],
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "keycloakrealm-sample"
                },
                "spec": {
                  "id": "d1-id-kc-realm-name",
                  "keycloakRef": {
                    "kind": "Keycloak",
                    "name": "keycloak-sample"
                  },
                  "organizationsEnabled": true,
                  "passwordPolicy": [
                    {
                      "type": "forceExpiredPasswordChange",
                      "value": "365"
                    },
                    {
                      "type": "length",
                      "value": "8"
                    }
                  ],
                  "realmEventConfig": {
                    "adminEventsDetailsEnabled": false,
                    "adminEventsEnabled": true,
                    "enabledEventTypes": [
                      "UPDATE_CONSENT_ERROR",
                      "CLIENT_LOGIN"
                    ],
                    "eventsEnabled": true,
                    "eventsExpiration": 15000,
                    "eventsListeners": [
                      "jboss-logging"
                    ]
                  },
                  "realmName": "d2-id-kc-realm-name",
                  "userProfileConfig": {
                    "attributes": [
                      {
                        "annotations": {
                          "inputType": "text"
                        },
                        "displayName": "Test Attribute",
                        "group": "test-group",
                        "multivalued": true,
                        "name": "test-attribute",
                        "permissions": {
                          "edit": [
                            "admin"
                          ],
                          "view": [
                            "admin",
                            "user"
                          ]
                        },
                        "required": {
                          "roles": [
                            "admin"
                          ],
                          "scopes": [
                            "profile"
                          ]
                        },
                        "selector": {
                          "scopes": [
                            "profile"
                          ]
                        },
                        "validations": {
                          "email": {
                            "max-local-length": {
                              "intVal": 64
                            }
                          },
                          "local-date": {},
                          "multivalued": {
                            "max": {
                              "stringVal": "10"
                            },
                            "min": {
                              "stringVal": "1"
                            }
                          },
                          "options": {
                            "options": {
                              "sliceVal": [
                                "option1",
                                "option2"
                              ]
                            }
                          }
                        }
                      }
                    ],
                    "groups": [
                      {
                        "annotations": {
                          "groupAnnotation": "groupAnnotation"
                        },
                        "displayDescription": "Test Group",
                        "displayHeader": "Test Group",
                        "name": "test-group"
                      }
                    ],
                    "unmanagedAttributePolicy": "ENABLED"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmComponent",
                "metadata": {
                  "name": "keycloakrealmcomponent-sample"
                },
                "spec": {
                  "config": {
                    "allowPasswordAuthentication": [
                      "true"
                    ],
                    "cachePolicy": [
                      "EVICT_WEEKLY"
                    ],
                    "debug": [
                      "true"
                    ],
                    "editMode": [
                      "READ_ONLY"
                    ],
                    "enabled": [
                      "true"
                    ],
                    "evictionDay": [
                      "3"
                    ],
                    "evictionHour": [
                      "5"
                    ],
                    "evictionMinute": [
                      "7"
                    ],
                    "kerberosRealm": [
                      "test-realm"
                    ],
                    "keyTab": [
                      "test-key-tab"
                    ],
                    "priority": [
                      "0"
                    ],
                    "serverPrincipal": [
                      "srv-principal-test"
                    ],
                    "updateProfileFirstLogin": [
                      "true"
                    ]
                  },
                  "name": "cr-kerb-test",
                  "providerId": "kerberos",
                  "providerType": "org.keycloak.storage.UserStorageProvider",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmGroup",
                "metadata": {
                  "name": "keycloakrealmgroup-sample"
                },
                "spec": {
                  "name": "ArgoCDAdmins",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "test-org-idp",
                  "namespace": "default"
                },
                "spec": {
                  "alias": "test-org-idp",
                  "config": {
                    "clientId": "test-org-client-id",
                    "clientSecret": "test-org-client-secret",
                    "kc.org.broker.redirect.mode.email-matches": "true",
                    "kc.org.domain": "example.com"
                  },
                  "enabled": true,
                  "providerId": "github",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "test-org-realm"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "keycloakrealmidentityprovider-sample"
                },
                "spec": {
                  "alias": "github",
                  "authenticateByDefault": false,
                  "config": {
                    "clientId": "foo",
                    "clientSecret": "$secretName:secretKey",
                    "syncMode": "IMPORT",
                    "useJwksUrl": "true"
                  },
                  "enabled": true,
                  "firstBrokerLoginFlowAlias": "first broker login",
                  "mappers": [
                    {
                      "config": {
                        "role": "role-tr",
                        "syncMode": "INHERIT"
                      },
                      "identityProviderAlias": "github",
                      "identityProviderMapper": "oidc-hardcoded-role-idp-mapper",
                      "name": "test3212"
                    },
                    {
                      "config": {
                        "attribute": "foo",
                        "attribute.value": "bar",
                        "syncMode": "IMPORT"
                      },
                      "identityProviderAlias": "github",
                      "identityProviderMapper": "hardcoded-attribute-idp-mapper",
                      "name": "test-33221"
                    }
                  ],
                  "providerId": "github",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRole",
                "metadata": {
                  "name": "keycloakrealmrole-sample"
                },
                "spec": {
                  "composite": true,
                  "description": "default developer role",
                  "name": "developer",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRoleBatch",
                "metadata": {
                  "name": "keycloakrealmrolebatch-sample"
                },
                "spec": {
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "roles": [
                    {
                      "composite": true,
                      "description": "default developer role",
                      "isDefault": false,
                      "name": "developer"
                    },
                    {
                      "composite": true,
                      "description": "default administrator role",
                      "isDefault": false,
                      "name": "administrator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmUser",
                "metadata": {
                  "name": "keycloakrealmuser-sample"
                },
                "spec": {
                  "attributes": {
                    "baz": "jazz",
                    "foo": "bar"
                  },
                  "email": "john.snow13@example.com",
                  "emailVerified": true,
                  "enabled": true,
                  "firstName": "John",
                  "keepResource": true,
                  "lastName": "Snow",
                  "password": "12345678",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "requiredUserActions": [
                    "UPDATE_PASSWORD"
                  ],
                  "username": "john.snow13"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloak",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloak-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloak",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloak-sample"
                },
                "spec": {
                  "secret": "keycloak-access",
                  "url": "https://keycloak.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloakRealm",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloakrealm-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloakrealm",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloakrealm-sample"
                },
                "spec": {
                  "authenticationFlows": {
                    "browserFlow": "browserFlow-sample"
                  },
                  "clusterKeycloakRef": "clusterkeycloak-sample",
                  "organizationsEnabled": false,
                  "realmName": "realm-sample"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "KeycloakOrganization",
                "metadata": {
                  "name": "test-keycloak-organization",
                  "namespace": "default"
                },
                "spec": {
                  "alias": "test-org",
                  "attributes": {
                    "department": [
                      "engineering",
                      "qa"
                    ],
                    "location": [
                      "us-east"
                    ]
                  },
                  "description": "Test organization",
                  "domains": [
                    "example.com",
                    "test.com"
                  ],
                  "identityProviders": [
                    {
                      "alias": "test-org-idp"
                    }
                  ],
                  "name": "Test Organization",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "test-org-realm"
                  },
                  "redirectUrl": "https://example.com/redirect"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          containerImage: docker.io/epamedp/keycloak-operator:1.29.0
          createdAt: "2025-09-26T12:16:55Z"
          description: An Operator for managing Keycloak
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/epam/edp-keycloak-operator
          support: KubeRocketCI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterKeycloakRealm is the Schema for the clusterkeycloakrealms
              API.
            displayName: Cluster Keycloak Realm
            kind: ClusterKeycloakRealm
            name: clusterkeycloakrealms.v1.edp.epam.com
            version: v1alpha1
          - description: ClusterKeycloak is the Schema for the clusterkeycloaks API.
            displayName: Cluster Keycloak
            kind: ClusterKeycloak
            name: clusterkeycloaks.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakAuthFlow is the Schema for the keycloak authentication
              flow API.
            displayName: Keycloak Auth Flow
            kind: KeycloakAuthFlow
            name: keycloakauthflows.v1.edp.epam.com
            version: v1
          - description: KeycloakClient is the Schema for the keycloak clients API.
            displayName: Keycloak Client
            kind: KeycloakClient
            name: keycloakclients.v1.edp.epam.com
            version: v1
          - description: KeycloakClientScope is the Schema for the keycloakclientscopes
              API.
            displayName: Keycloak Client Scope
            kind: KeycloakClientScope
            name: keycloakclientscopes.v1.edp.epam.com
            version: v1
          - description: KeycloakOrganization is the Schema for the organizations
              API.
            displayName: Keycloak Organization
            kind: KeycloakOrganization
            name: keycloakorganizations.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakRealmComponent is the Schema for the keycloak component
              API.
            displayName: Keycloak Realm Component
            kind: KeycloakRealmComponent
            name: keycloakrealmcomponents.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmGroup is the Schema for the keycloak group API.
            displayName: Keycloak Realm Group
            kind: KeycloakRealmGroup
            name: keycloakrealmgroups.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmIdentityProvider is the Schema for the keycloak
              realm identity provider API.
            displayName: Keycloak Realm Identity Provider
            kind: KeycloakRealmIdentityProvider
            name: keycloakrealmidentityproviders.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRoleBatch is the Schema for the keycloak roles
              API.
            displayName: Keycloak Realm Role Batch
            kind: KeycloakRealmRoleBatch
            name: keycloakrealmrolebatches.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRole is the Schema for the keycloak group API.
            displayName: Keycloak Realm Role
            kind: KeycloakRealmRole
            name: keycloakrealmroles.v1.edp.epam.com
            version: v1
          - description: KeycloakRealm is the Schema for the keycloak realms API.
            displayName: Keycloak Realm
            kind: KeycloakRealm
            name: keycloakrealms.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmUser is the Schema for the keycloak user API.
            displayName: Keycloak Realm User
            kind: KeycloakRealmUser
            name: keycloakrealmusers.v1.edp.epam.com
            version: v1
          - description: Keycloak is the Schema for the keycloaks API.
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.v1.edp.epam.com
            version: v1
        description: |
          Keycloak Operator is an operator that is responsible for establishing
          a connection to provided Keycloak Server, reconciling Keycloak entities (realms,
          roles, groups, users, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Keycloak `Master` realm, and assign a `create-realm` role.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  keycloak-access
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Keycloak` with Keycloak instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: Keycloak
              metadata:
                name: keycloak-sample
              spec:
                secret: keycloak-access             # Secret name
                url: https://keycloak.example.com   # Keycloak URL
              ```

              Wait for the `.status` field with  `status.connected: true`

          4. Create Keycloak realm and group using Custom Resources:

            ```yaml
            apiVersion: v1.edp.epam.com/v1
            kind: KeycloakRealm
            metadata:
              name: keycloakrealm-sample
            spec:
              realmName: realm-sample
              keycloakOwner: keycloak-sample   # the name of `kind: Keycloak`
              ```

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: KeycloakRealmGroup
              metadata:
                name: argocd-admins
              spec:
                name: ArgoCDAdmins
                realm: keycloakrealm-sample   # the name of `kind: KeycloakRealm`
              ```
        displayName: EDP Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - authentication
        - authorization
        - edp
        - idp
        - keycloak
        - kuberocketci
        - oauth
        - oidc
        - operator
        - saml
        - sso
        links:
        - name: Edp Keycloak Operator
          url: https://github.com/epam/edp-keycloak-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: KubeRocketCI
          url: https://docs.kuberocketci.io/
        relatedImages:
        - docker.io/epamedp/keycloak-operator:1.29.0
        version: 1.29.0
      entries:
      - name: edp-keycloak-operator.v1.29.0
        version: 1.29.0
      - name: edp-keycloak-operator.v1.28.0
        version: 1.28.0
      - name: edp-keycloak-operator.v1.27.1
        version: 1.27.1
      - name: edp-keycloak-operator.v1.26.0
        version: 1.26.0
      - name: edp-keycloak-operator.v1.25.0
        version: 1.25.0
      - name: edp-keycloak-operator.v1.24.0
        version: 1.24.0
      - name: edp-keycloak-operator.v1.23.0
        version: 1.23.0
      name: stable
    defaultChannel: stable
    packageName: edp-keycloak-operator
    provider:
      name: KubeRocketCI
      url: https://docs.kuberocketci.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: minhokim
      provider-url: ""
    name: node-discovery-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-discovery-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.wunicorns.com/v1alpha1",
                "kind": "NodeDiscovery",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "node-discovery-operator",
                    "app.kubernetes.io/instance": "nodediscovery-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodediscovery",
                    "app.kubernetes.io/part-of": "node-discovery-operator"
                  },
                  "name": "nodediscovery-sample"
                },
                "spec": {
                  "containers": [
                    {
                      "args": [
                        "tail",
                        "-f",
                        "/dev/null"
                      ],
                      "image": "alicek106/stress",
                      "imagePullPolicy": "IfNotPresent",
                      "name": "stress",
                      "resources": {
                        "limits": {
                          "cpu": "100m",
                          "memory": "1000Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "1000Mi"
                        }
                      },
                      "terminationGracePeriodSeconds": 10
                    }
                  ],
                  "cpus": "0",
                  "enabled": "true",
                  "mems": "0",
                  "nodeName": "worker3"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-18T01:32:51Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeDiscovery is the Schema for the nodediscoveries API
            displayName: Node Discovery
            kind: NodeDiscovery
            name: nodediscoveries.cache.wunicorns.com
            version: v1alpha1
        description: find cgroup infor and set value.
        displayName: node-discovery-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cgroup
        - cpulist
        - mems
        - cpus
        - numa
        links:
        - name: Node Discovery Operator
          url: https://node-discovery-operator.domain
        maintainers:
        - email: wunicorns1436@gmail.com
          name: minhokim
        maturity: alpha
        provider:
          name: minhokim
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/caspell/cgroup-discovery@sha256:b437dcde0107124e43a4e2fb8026ec191f44b5e761bedc2c295f47ff489cf343
        version: 0.0.1
      entries:
      - name: node-discovery-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: node-discovery-operator
    provider:
      name: minhokim
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HashiCorp
      provider-url: https://www.hashicorp.com/
    name: hcp-terraform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hcp-terraform-operator.v2.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "AgentPool",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "agentTokens": [
                    {
                      "name": "token"
                    }
                  ],
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Module",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "module": {
                    "source": "redeux/terraform-cloud-agent/kubernetes",
                    "version": "1.0.1"
                  },
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  },
                  "workspace": {
                    "id": "WORKSPACE_ID"
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Project",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Workspace",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.connect.redhat.com/hashicorp/hcp-terraform-operator@sha256:8885a6a5eae092117a670ec5541a3e4a24c6b23807fd17b0446d91b01eb6e421
          createdAt: "2025-05-28T13:58:52Z"
          description: HCP Terraform Operator for Kubernetes allows managing HCP Terraform
            / Terraform Enterprise resources via Kubernetes Custom Resources.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Valid HCP Terraform / Terraform
            Enterprise subscription is required
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hashicorp/hcp-terraform-operator
          support: HashiCorp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              AgentPool manages HCP Terraform Agent Pools, HCP Terraform Agent Tokens and can perform HCP Terraform Agent scaling.
              More infromation:
                - https://developer.hashicorp.com/terraform/cloud-docs/agents/agent-pools
                - https://developer.hashicorp.com/terraform/cloud-docs/users-teams-organizations/api-tokens#agent-api-tokens
                - https://developer.hashicorp.com/terraform/cloud-docs/agents
            displayName: Agent Pool
            kind: AgentPool
            name: agentpools.app.terraform.io
            version: v1alpha2
          - description: |-
              Module implements API-driven Run Workflows.
              More information:
                - https://developer.hashicorp.com/terraform/cloud-docs/run/api
            displayName: Module
            kind: Module
            name: modules.app.terraform.io
            version: v1alpha2
          - description: |-
              Project manages HCP Terraform Projects.
              More information:
              - https://developer.hashicorp.com/terraform/cloud-docs/projects/manage
            displayName: Project
            kind: Project
            name: projects.app.terraform.io
            version: v1alpha2
          - description: |-
              Workspace manages HCP Terraform Workspaces.
              More information:
                - https://developer.hashicorp.com/terraform/cloud-docs/workspaces
            displayName: Workspace
            kind: Workspace
            name: workspaces.app.terraform.io
            version: v1alpha2
        description: HCP Terraform Operator for Kubernetes allows managing HCP Terraform
          / Terraform Enterprise resources via Kubernetes Custom Resources.
        displayName: HCP Terraform Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - terraform
        links:
        - name: HCP Terraform Operator
          url: https://github.com/hashicorp/hcp-terraform-operator
        maintainers:
        - email: hcp-tf-operator@hashicorp.com
          name: HashiCorp
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: HashiCorp
          url: https://www.hashicorp.com/
        relatedImages:
        - registry.connect.redhat.com/hashicorp/hcp-terraform-operator@sha256:8885a6a5eae092117a670ec5541a3e4a24c6b23807fd17b0446d91b01eb6e421
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        version: 2.9.2
      entries:
      - name: hcp-terraform-operator.v2.9.2
        version: 2.9.2
      - name: hcp-terraform-operator.v2.9.1
        version: 2.9.1
      - name: hcp-terraform-operator.v2.9.0
        version: 2.9.0
      - name: hcp-terraform-operator.v2.8.1
        version: 2.8.1
      - name: hcp-terraform-operator.v2.8.0
        version: 2.8.0
      name: alpha
    defaultChannel: alpha
    packageName: hcp-terraform-operator
    provider:
      name: HashiCorp
      url: https://www.hashicorp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Appian Corporation
      provider-url: https://appian.com/
    name: appian-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appian-operator.v0.184.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: AI/Machine Learning, Application Runtime, Big Data, Integration
            & Delivery, Modernization & Migration
          createdAt: "2025-08-07T15:50:26Z"
          description: The Appian operator facilitates managing instances of Appian
            on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appian-operator
          operators.openshift.io/valid-subscription: '["Appian License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Appian Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Appian
            name: appians.crd.k8s.appian.com
            version: v1beta1
        description: The Appian operator facilitates managing instances of Appian
          on Kubernetes.
        displayName: Appian operator
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - Appian
        links:
        - name: Appian
          url: https://appian.com/
        - name: Documentation
          url: https://docs.appian.com/suite/help/latest/k8s
        maintainers:
        - email: support@appian.com
          name: Appian Corporation
        maturity: planning
        provider:
          name: Appian Corporation
          url: https://appian.com/
        relatedImages:
        - 937620572175.dkr.ecr.us-east-1.amazonaws.com/appian/appian-operator@sha256:26c2ba27f9360226f34f289109cb0d10e9bf44777146ccfd5f367adf2367a400
        version: 0.184.0
      entries:
      - name: appian-operator.v0.184.0
        version: 0.184.0
      - name: appian-operator.v0.171.0
        version: 0.171.0
      name: alpha
    defaultChannel: alpha
    packageName: appian-operator
    provider:
      name: Appian Corporation
      url: https://appian.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-ingress
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-ingress.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Ingress",
                "metadata": {
                  "name": "ingress-sample"
                },
                "spec": {
                  "admissionWebhooks": {
                    "patch": {
                      "image": {
                        "repository": "docker.io/jettech/kube-webhook-certgen"
                      }
                    }
                  },
                  "configMapMerger": {
                    "registry": "prod.harbor.keysight.digital/loadcore",
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "environment": "openshift",
                  "ingress-nginx": {
                    "controller": {
                      "config": {
                        "global-auth-signin": "https://$host/eula/static/",
                        "global-auth-url": "http://kcos-framework-v1-kcos-eula.kcos-framework.svc.cluster.local/v1/check"
                      },
                      "daemonset": {
                        "useHostPort": false
                      },
                      "hostNetwork": false,
                      "image": {
                        "registry": "prod.harbor.keysight.digital/loadcore"
                      },
                      "ingressClassResource": {
                        "default": true
                      },
                      "nodeSelector": "",
                      "service": {
                        "nodePorts": {
                          "http": 80,
                          "https": 443
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "defaultBackend": {
                      "image": {
                        "repository": "k8s.gcr.io/defaultbackend-amd64"
                      }
                    }
                  },
                  "jobDelWebhook": {
                    "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:28:18Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ingress
            name: ingresses.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Ingress Operator
        displayName: Keysight KCOS Ingress Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Ingress
          url: https://keysight-kcos-ingress.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-ingress@sha256:c6f6f8b44276efd8f5ffaa099b393e7e97d9cfa627952c0d049b98d92dc3ab59
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        version: 0.1.21
      entries:
      - name: keysight-kcos-ingress.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-ingress
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-migration
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-migration.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Migration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {
                  "auth": {
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configuration": {
                    "port": 8080
                  },
                  "database": {
                    "image": {
                      "name": "postgresql-ubi-image",
                      "tag": "14.2.0-3747-32"
                    }
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "esinfo": {
                    "batchSize": 10000,
                    "inputIndex": "wireless_stats"
                  },
                  "image": {
                    "remoteRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "systemMonitor": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:36:21Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Migration
            name: migrations.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Migration Operator
        displayName: Keysight WAP Migration Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Migration
          url: https://keysight-wap-migration.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-migration@sha256:118b466f1c59832a3620fa9b153507cc66ba38ee08b337b3a1a4225b158bebc5
        version: 0.1.21
      entries:
      - name: keysight-wap-migration.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-migration
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-xtradb-cluster-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator-certified.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator-certified.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator-certified.v1.17.0
        version: 1.17.0
      - name: percona-xtradb-cluster-operator-certified.v1.16.1
        version: 1.16.1
      - name: percona-xtradb-cluster-operator-certified.v1.16.0
        version: 1.16.0
      - name: percona-xtradb-cluster-operator-certified.v1.15.1
        version: 1.15.1
      - name: percona-xtradb-cluster-operator-certified.v1.15.0
        version: 1.15.0
      name: stable
    - currentCSV: percona-xtradb-cluster-operator-certified.v1.18.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        version: 1.18.0-cw
      entries:
      - name: percona-xtradb-cluster-operator-certified.v1.18.0-cw
        version: 1.18.0-cw
      - name: percona-xtradb-cluster-operator-certified.v1.17.0-cw
        version: 1.17.0-cw
      - name: percona-xtradb-cluster-operator-certified.v1.16.1-cw
        version: 1.16.1-cw
      - name: percona-xtradb-cluster-operator-certified.v1.16.0-cw
        version: 1.16.0-cw
      - name: percona-xtradb-cluster-operator-certified.v1.15.0-cw
        version: 1.15.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator-certified
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-storage-odf-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-storage-odf-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:962ca15393dd2c48f8a7189c1801b43e9fd1cd073a1971286d485f4b96060426
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.6.0'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:962ca15393dd2c48f8a7189c1801b43e9fd1cd073a1971286d485f4b96060426
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:2f785096aaa63b042e415aa0b64ace2a085cb7c219853f089d803bc37d8295ce
        version: 1.6.0
      entries:
      - name: ibm-storage-odf-operator.v1.6.0
        version: 1.6.0
      name: stable-v1.6
    - currentCSV: ibm-storage-odf-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:2dd8608645b13f81330466e25664b132d46af480a3ab2b0dd48fb1feba38870a
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.7.1'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:2dd8608645b13f81330466e25664b132d46af480a3ab2b0dd48fb1feba38870a
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:24eae9d1502ce98bb024d3eba154ddec07fdac0d78ecad0c3a64b55f5b3600f7
        version: 1.7.1
      entries:
      - name: ibm-storage-odf-operator.v1.7.1
        version: 1.7.1
      name: stable-v1.7
    - currentCSV: ibm-storage-odf-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:7432340ac156e620240eea73a19321a871fa89cc513c340a91d97d3f3fd4ded2
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.8.0'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:7432340ac156e620240eea73a19321a871fa89cc513c340a91d97d3f3fd4ded2
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:3344c7e987674fb98173bbc6ae1b189613e3ccefb468f1af9866734ddbf0408f
        version: 1.8.0
      entries:
      - name: ibm-storage-odf-operator.v1.8.0
        version: 1.8.0
      name: stable-v1.8
    defaultChannel: stable-v1.8
    packageName: ibm-storage-odf-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Strimzi
      provider-url: ""
    name: strimzi-kafka-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: strimzi-cluster-operator.v0.48.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.1.0",
                      "metadataVersion": "4.1-IV1",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.1.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
          createdAt: "2025-09-25 10:32:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.48.0 supports only Kubernetes 1.27 and newer.**
            Kubernetes 1.25 and 1.26 are not supported anymore

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.48.0
          * Support for Kafka 4.1.0
          * Add monitoring of custom resources using kubernetes-state-metrics (KSM)
          * Support for type: custom client authentication
          * Added support for Strimzi Metrics Reporter to Kafka Connect, Mirror Maker 2 and Kafka Bridge.
          * Added support for KIP-745 in Kafka Connector and Mirror Maker 2, allowing the usage of includeTasks and onlyFailed arguments in Kafka connectors restart.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
        - quay.io/strimzi/kafka@sha256:34450afc1a3399a9026fc3c6ec90e9f6adfec9faf6520545b4d84e90cda34964
        - quay.io/strimzi/kafka-bridge@sha256:53034f64f0b672f10b5bacea1c7a25132f118df7fd5c9032c4dbf702ed93796a
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:64380efcb8da87fdc0e0de238a20456be9abb830174c2f2331882b7a84657e57
        - quay.io/strimzi/kafka@sha256:48db05ad2e9fa39eb388089e7e25654bcbf4da1ddc3a5635d675056c28bbcb96
        version: 0.48.0
      entries:
      - name: strimzi-cluster-operator.v0.48.0
        version: 0.48.0
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: stable
    - currentCSV: strimzi-cluster-operator.v0.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnectS2I",
                "metadata":{
                  "name":"my-connect-s2i-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "whitelist":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 2,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": 1,
                        "offset.storage.replication.factor": 1,
                        "status.storage.replication.factor": 1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "NetworkOutboundCapacityGoal",
                    "CpuCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:59d939f342f96d0e8a54c96fd187832c4b19cd4137bf55ace53d486728ae6c5f
          createdAt: "2021-05-14 11:37:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster with Source 2 Image support
            displayName: Kafka Connect Source to Image
            kind: KafkaConnectS2I
            name: kafkaconnects2is.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  Starting with Strimzi 0.23, we support only the `v1beta2` API version of all Strimzi custom resources.  This is a required for migration to `apiextensions/v1` which is needed because Kubernetes 1.22 will remove support for `apiextensions/v1beta1`.  Migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs **before upgrading to Strimzi 0.23**. For more details about the CRD upgrades, see the [documentation](https://strimzi.io/docs/operators/0.23.0/deploying.html#assembly-upgrade-resources-str).
          ### New in 0.23
          * Migrate to CRD v1 (required by Kubernetes 1.22+)
          * Strimzi API versions `v1alpha1` and `v1beta1` were removed from all Strimzi custom resources apart from `KafkaTopic` and `KafkaUser` (use `v1beta2` versions instead)
          * Add support for Kafka 2.8.0 and 2.6.2, remove support for Kafka 2.5.x
          * Make it possible to configure maximum number of connections and maximum connection creation rate in listener configuration
          * Add support for configuring finalizers for `loadbalancer` type listeners
          * Use dedicated Service Account for Kafka Connect Build on Kubernetes
          * Remove direct ZooKeeper access for handling user quotas in the User Operator. Add usage of Admin Client API instead.
          * Support for configuring custom Authorizer implementation
          * Changed Reconciliation interval for Topic Operator from 90 to 120 seconds (to keep it the same as for other operators)
          * Changed Zookeeper session timeout default value to 18 seconds for Topic and User Operators (for improved resiliency)
          * Removed requirement for replicas and partitions KafkaTopic spec making these parameters optional
          * Support to configure a custom filter for parent CR's labels propagation into subresources
          * Allow disabling service links (environment variables describing Kubernetes services) in Pod template
          * Add support for separate control plane listener (disabled by default, available via the `ControlPlaneListener` feature gate)
          * Support for Dual Stack networking
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _master_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/master/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/master/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:59d939f342f96d0e8a54c96fd187832c4b19cd4137bf55ace53d486728ae6c5f
        - quay.io/strimzi/kafka@sha256:699bff48ee445d8ba381a791cd0340015e3ecede8d169f41bc151095ab7fb5cc
        - quay.io/strimzi/kafka@sha256:d2745f48ed262290ce66c3362ffb9b9a8cfeb9c76beac40c995f72f3f06ad10e
        - quay.io/strimzi/kafka@sha256:dbc53bb374d2cb9513a5bb8f80fea751c91b249cabd36d8b3c04af5e5c961738
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/kafka@sha256:e90695b35d6dd5db20c9900dd0dab95d1ef908f3145b4f9f436b903303c88549
        - quay.io/strimzi/kafka-bridge@sha256:bed07186355a4c04ee50e90048e3bd3a023264093a9ef28f0278cce44640e93f
        - quay.io/strimzi/kafka@sha256:641f46114c59d61577bcb3f4ef5af16aea3ed42c2a575f8d3641c82848df85e8
        - quay.io/strimzi/jmxtrans@sha256:b4ec1844f4cc5eb74cd0aac83d74b81b977e178bf670d1136d1af0b68663d0a3
        version: 0.23.0
      entries:
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.23.x
    - currentCSV: strimzi-cluster-operator.v0.24.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnectS2I",
                "metadata":{
                  "name":"my-connect-s2i-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 2,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:786ea69c77e0651f009a73a8b3ce0b3d41b3cd71b917dc0d9bdc61c725f494a6
          createdAt: "2021-06-24 13:13:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster with Source 2 Image support
            displayName: Kafka Connect Source to Image
            kind: KafkaConnectS2I
            name: kafkaconnects2is.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: "Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org)
          cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/)
          in various deployment configurations. See our [website](https://strimzi.io)
          for more details about the project.\n### CRD Upgrades\n**!!! IMPORTANT !!!**
          \ This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.
          \ If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed
          for all Strimzi CRDs and CRs before the upgrade to 0.24 is done!  If upgrading
          from Strimzi version earlier than 0.22, you need to first install the CRDs
          from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi
          CRDs and CRs before the upgrade to 0.24 is done!  For more details about
          the CRD upgrades, see the documentation.\n### New in 0.24\n* Add support
          for Kafka 2.7.1 and remove support for 2.6.0, 2.6.1, and 2.6.2\n* Add support
          for [Kubernetes Configuration Provider for Apache Kafka](https://github.com/strimzi/kafka-kubernetes-config-provider)\n*
          Use Red Hat UBI8 base image\n* Support for patching of service accounts
          and configuring their labels and annotations. The feature is disabled by
          default and enabled using the new `ServiceAccountPatching` feature gate.\n*
          Added support for configuring cluster-operator's worker thread pool size
          that is used for various sync and async tasks\n* Add Kafka Quotas plugin
          with produce, consume, and storage quotas\n* Support pausing reconciliation
          of KafkaTopic CR with annotation `strimzi.io/pause-reconciliation`\n* Update
          cruise control to 2.5.57\n* Update to Strimzi Kafka Bridge to 0.20.1\n*
          Support for broker load information added to the rebalance optimization
          proposal. Information on the load difference, before and after a rebalance
          is stored in a ConfigMap\n* Add support for selectively changing the verbosity
          of logging for individual CRs, using markers.\n* Added support for `controller_mutation_rate'
          quota. Creation/Deletion of topics and creation of partitions can be configured
          through this.\n* Use newer version of Kafka Exporter with different bugfixes
          \n### Supported Features\n* **Manages the Kafka Cluster** - Deploys and
          manages all of the components of this complex application, including dependencies
          like Apache ZooKeeper® that are traditionally hard to administer.\n* **Includes
          Kafka Connect** - Allows for configuration of common data sources and sinks
          to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for morroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashabords\n### Upgrading your Clusters\nThe
          Strimzi Operator understands how to run and upgrade between a set of Kafka
          versions. When specifying a new version in your config, check to make sure
          you aren't using any features that may have been removed. See [the upgrade
          guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str)
          for more information.\n### Storage\nAn efficient data storage infrastructure
          is essential to the optimal performance of Apache Kafka®. Apache Kafka®
          deployed via Strimzi requires block storage. The use of file storage (for
          example, NFS) is not recommended.\nThe Strimzi Operator supports three types
          of data storage:\n* Ephemeral (Recommended for development only!)\n* Persistent\n*
          JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nStrimzi
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation to the current _main_
          branch as well as all releases can be found on our [website](https://strimzi.io/documentation).\n###
          Getting help\nIf you encounter any issues while using Strimzi, you can get
          help using:\n* [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          Contributing\nYou can contribute by:\n* Raising any issues you find using
          Strimzi\n* Fixing issues by opening Pull Requests\n* Improving documentation\n*
          Talking about Strimzi\n\nAll bugs, tasks or enhancements are tracked as
          [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues).
          Issues which  might be a good start for new contributors are marked with
          [\"good-start\"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start)
          label.\n\nThe [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md)
          describes how to build Strimzi and how to  test your changes before submitting
          a patch or opening a PR.\n\nThe [Documentation Contributor Guide](https://strimzi.io/contributing/guide/)
          describes how to contribute to Strimzi documentation.\n\nIf you want to
          get in touch with us first before contributing, you can use:\n* [Strimzi
          mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          License\nStrimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).\n"
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/kafka@sha256:95cfe9000afda2f7def269ca46472d3803ee85146c521884884d8505a7187daf
        - quay.io/strimzi/kafka@sha256:8959b7968ab8b3306906cdbff2ebb8d63329af37e58124a601843795c4ef5bd6
        - quay.io/strimzi/kafka@sha256:fbb08410d9595029bc4a02ed859971264e6ce2dc85dd6a9855eaa7bb58b52a25
        - quay.io/strimzi/kafka-bridge@sha256:537ea10e87619251204747ebc1005c632a6c6ed270f5ef06f86c0c3ae4a7553b
        - quay.io/strimzi/operator@sha256:786ea69c77e0651f009a73a8b3ce0b3d41b3cd71b917dc0d9bdc61c725f494a6
        - quay.io/strimzi/jmxtrans@sha256:2c54adabf6f31c087dbb898057f9c5e200aaf4c5b39fb3a47bbaf7f6587df72d
        version: 0.24.0
      entries:
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.24.x
    - currentCSV: strimzi-cluster-operator.v0.25.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:afaf1dd1f2f696ae80674d74159661ea239ae609f78a73a1423c7a17a39e5044
          createdAt: "2021-06-24 13:13:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.25 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.25 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.25
          * Move from Scala 2.12 to Scala 2.13
          * Open Policy Agent authorizer updated to a new version supporting Scala 2.13.
            The new version of the authroizer uses different format of the input data sent to the OPA server and might require you to update your policies.
            See [release notes](https://github.com/strimzi/strimzi-kafka-operator/releases/tag/0.25.0) for more details.

          * Allow a custom password to be set for SCRAM-SHA-512 users by referencing a secret in the `KafkaUser` resource
          * Add support for [EnvVar Configuration Provider for Apache Kafka](https://github.com/strimzi/kafka-env-var-config-provider)
          * Add support for `tls-external` authentication to User Operator to allow management of ACLs and Quotas for TLS users with user certificates generated externally
          * Support for disabling the automatic generation of network policies by the Cluster Operator
          * Configure fixed size limit for `emptyDir` volumes used for temporary files
          * Update Strimzi Kafka Bridge to 0.20.2
          * The `KafkaConnectS2I` resource has been removed and is no longer supported by the operator.
            Please use the [migration guide](https://strimzi.io/docs/operators/0.24.0/full/using.html#proc-migrating-kafka-connect-s2i-str) to migrate your `KafkaConnectS2I` deployments to [`KafkaConnect` Build](https://strimzi.io/docs/operators/latest/full/deploying.html#creating-new-image-using-kafka-connect-build-str) instead.

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:d8991608ec71f17e850c33714ee02ca87ba94bc0f077be62ad1ed127ca244823
        - quay.io/strimzi/kafka@sha256:ee17eb3611019b0cfa5e9e196610edc1dbd7084471fb20990f055c89da8eef45
        - quay.io/strimzi/kafka@sha256:56a33815f7d50a97ea9bb4cce1e3a6d04d918b841b5aa285f16b792d8930059e
        - quay.io/strimzi/kafka-bridge@sha256:a005e20d7fbe04d6f587f2093e21a00556ec62b850df757b9f7d9d57e3141c40
        - quay.io/strimzi/jmxtrans@sha256:af5eb8459ac9a0acbca51e67c6e0217f62ba404cba9f2251968c927dcef75252
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/operator@sha256:afaf1dd1f2f696ae80674d74159661ea239ae609f78a73a1423c7a17a39e5044
        version: 0.25.0
      entries:
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.25.x
    - currentCSV: strimzi-cluster-operator.v0.26.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.0.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"3.0",
                        "inter.broker.protocol.version":"3.0"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.0.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:5d90caa66f57939ba62a9a50efa00fce319ba596e719a3f8a42170d7635ee3f4
          createdAt: "2021-12-14 00:16:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.26 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.26 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.26.1
          * **Updated Log4j2 to fix CVE-2021-44228**
          ### New in 0.26.0
          * Add support for Kafka 2.8.1 and 3.0.0; remove Kafka 2.7.0 and 2.7.1
          * Add support for exposing JMX in Zookeeper
          * Allow configuring labels and annotations for JMX authentication secrets
          * Enable Cruise Control anomaly.detection configurations
          * Add support for building connector images from the Maven coordinates
          * Allow Kafka Connect Build artifacts to be downloaded from insecure servers
          * Add option to specify pull secret in Kafka Connect Build on OpenShift
          * Configurable authentication, authorization, and SSL for Cruise Control API
          * Allow to configure `/tmp` volume size via Pod template. By default `1Mi` is used.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:d3899c79747e4cb3a6df5a9fcd80b81f038ca0634a97329c1d160499469876dd
        - quay.io/strimzi/jmxtrans@sha256:bff5d3bab575d0b9ed95dd79d2d22518cc00c4d1c9dbb5614ed16149d4920a1b
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/maven-builder@sha256:663eb81388ae8f824e7920c272f6d2e2274cf6c140d61416607261cdce9d50e2
        - quay.io/strimzi/kafka@sha256:8c022a8535742b2eea691737f863af5ddef1e8df08b415b8139ecd02a79052a5
        - quay.io/strimzi/kafka@sha256:bd22a134b0c351ddfc78a3263850c2f16733c52f57ac92984cbe0ffff477c724
        - quay.io/strimzi/operator@sha256:5d90caa66f57939ba62a9a50efa00fce319ba596e719a3f8a42170d7635ee3f4
        - quay.io/strimzi/kafka@sha256:6367b0fbba9593e3ac736040ea2d7ef71296b1ec144d3b54d9370352141b90e4
        version: 0.26.1
      entries:
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.26.x
    - currentCSV: strimzi-cluster-operator.v0.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.0.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"3.0",
                        "inter.broker.protocol.version":"3.0"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.0.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:9ea49407a04006d0bddfaaee822adfb95f5857af32ffcee81ba60ce891b07ba8
          createdAt: "2022-01-14 20:00:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.27 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.27 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.27.1
          * Updated Log4j2 to version 2.17.1 and Fabric8 Kubernetes Client to 5.10.2
          ### New in 0.27.0
          * Updated Log4j2 to version 2.17.0 (adresses CVE-2021-44228, CVE-2021-45046, CVE-2021-45105)
          * (Experimental) support for AArch64 / ARM64 architecture
          * `ControlPlaneListener` and `ServiceAccountPatching` feature gates are now in the beta phase and are enabled by default
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a07ebe6600d6baba09980abdde2931b3b6abd2f75aeb2cac5c6a2d22a37aa807
        - quay.io/strimzi/maven-builder@sha256:9acd178cc504be8acaa70a39f3c17830aae1f0e93a656110790f9476b09260d5
        - quay.io/strimzi/kafka@sha256:f4d68bb94447c6612f70de7f6587e9e7ef712c83769ea1f11a8dbda0d241a059
        - quay.io/strimzi/kafka@sha256:a3ec567ca529df2fa41f8986d21e86dde66be94aa4baa56faa66db4dcb3b258e
        - quay.io/strimzi/kafka@sha256:5682a6bfcc4ee13d53a30d587c8f96bc40abe3618880989336c5af274ea3cefa
        - quay.io/strimzi/kafka-bridge@sha256:55a52316aead02efa2004174e11ad4728bb8b6f1c1415f725f597f1d2f2da188
        - quay.io/strimzi/operator@sha256:9ea49407a04006d0bddfaaee822adfb95f5857af32ffcee81ba60ce891b07ba8
        - quay.io/strimzi/jmxtrans@sha256:d2d2477517eb1ac98b51ad6d64d8c1c578141637afa8d46037076f0013183c0b
        version: 0.27.1
      entries:
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.27.x
    - currentCSV: strimzi-cluster-operator.v0.28.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.1.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.1"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.1.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.1.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:8b010c5fb8244b742a37f264feefc116d75bb2abd2d6895375cbeea250727b39
          createdAt: "2022-02-12 23:02:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.28 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.28 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.28.0
          * Support for Kafka 3.1.0
          * Support for `StrimziPodSet` resources (disabled by default through the `UseStrimziPodSets` feature gate)
          * Support custom authentication mechanisms in Kafka listeners
          * Intra-broker disk balancing using Cruise Control
          * Add support for disabling the FIPS mode in OpenJDK
          * Fix renewing your own CA certificates
          * Added the option `createBootstrapService` to disable the creation of the bootstrap service for the Load Balancer Type Listener
          * Dependency updates (Strimzi Kafka Bridge, Strimzi OAuth, Open Policy Agent Authroizer plugin and others)
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a07ebe6600d6baba09980abdde2931b3b6abd2f75aeb2cac5c6a2d22a37aa807
        - quay.io/strimzi/maven-builder@sha256:3d5d664e089651944712361a6d8ba42b578dbd9bb3dfd9972f614520c9cfab75
        - quay.io/strimzi/kafka@sha256:7c04f65e216bb941458aeadffb85e9eeee43d5ca67da566017c2c93a0664c98e
        - quay.io/strimzi/kafka@sha256:06e9bda15495feb2b92d9950e834cdeed6ef20064f021a879610f1b59dd46950
        - quay.io/strimzi/kafka-bridge@sha256:c345c783d8fbbee4edeca2ebc47e7c454fcc2e26686820b8037de97369280ba0
        - quay.io/strimzi/jmxtrans@sha256:3a55a81790dc88b90d1bf858e2cfd7e591ad0e79f763c6d1667b2d3eb0d0c79d
        - quay.io/strimzi/operator@sha256:8b010c5fb8244b742a37f264feefc116d75bb2abd2d6895375cbeea250727b39
        version: 0.28.0
      entries:
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.28.x
    - currentCSV: strimzi-cluster-operator.v0.29.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:afde94c6d2544207cb7f3c94b54455643558a9d5daaa13b9f736cc8da20e319c
          createdAt: "2022-05-15 00:01:07"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.29 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.29 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.29.0
          * Add support for Apache Kafka 3.0.1, 3.1.1 and 3.2.0
          * Renew user certificates in User Operator only during maintenance windows
          * Remove TLS sidecar from Cruise Control pod. Cruise Control is now configured to not using ZooKeeper, so the TLS sidecar is not needed anymore.
          * Allow Cruise Control topic names to be configured
          * Add support for `spec.rack.topologyKey` property in Mirror Maker 2 to enable "fetch from the closest replica" feature.
          * Experiemntal support for the s390x platform
          * Added new rebalancing modes on the `KafkaRebalance` custom resource to add or remove brokers
          * Experimental KRaft mode (ZooKeeper-less Kafka). Use it for development and testing only!
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:544c4987f7f3ca558b04e9ef4357be204f088212b7b5b9ed0f909401a8bc5238
        - quay.io/strimzi/maven-builder@sha256:0c096a59faf8fd224298c01415a895043db4647534c060cb7e53207e9448bf65
        - quay.io/strimzi/operator@sha256:afde94c6d2544207cb7f3c94b54455643558a9d5daaa13b9f736cc8da20e319c
        - quay.io/strimzi/kafka@sha256:95678e843bc87cbe571f3f299045dd075725816b59e1b87821cd6025876ec559
        - quay.io/strimzi/kafka@sha256:935caab03f1d0365b47ddefec0c1d8fa449fc3de715ccdd298aedef5ce968c73
        - quay.io/strimzi/kafka@sha256:ac190141120c1800cc7131d3a1f701e64efdcd4a5d87f0b2e0e2cadfdef8b438
        - quay.io/strimzi/kafka@sha256:9dfc09273ad8fc4bdc75cbe93b4a061a05a5defb5884a221ae1e178ba602787f
        - quay.io/strimzi/kafka@sha256:284cde40784b6ff7bfb81b6dd28928e1e8c598fe6a72504f025592e5b43fa94c
        - quay.io/strimzi/kafka-bridge@sha256:b7bc3dd90e87223766f608a0087b521f0bba79d1c21eaa833bc2708c71148b7f
        - quay.io/strimzi/jmxtrans@sha256:f89e3c9618430908a8b390f2490f5aa43687b12889b434472a962ce0170c6463
        version: 0.29.0
      entries:
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.29.x
    - currentCSV: strimzi-cluster-operator.v0.30.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:587db6f4389f56209f48853beedbfe249f6d590a4cb0000be2c42b57e43c2adb
          createdAt: "2022-07-16 00:00:10"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: "Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org)
          cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/)
          in various deployment configurations. See our [website](https://strimzi.io)
          for more details about the project.\n### CRD Upgrades\n**!!! IMPORTANT !!!**
          \ This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.
          \ If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed
          for all Strimzi CRDs and CRs before the upgrade to 0.30 is done!  If upgrading
          from Strimzi version earlier than 0.22, you need to first install the CRDs
          from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi
          CRDs and CRs before the upgrade to 0.30 is done!  For more details about
          the CRD upgrades, see the documentation.\n### New in 0.30.0\n* The `UseStrimziPodSets`
          feature gate moves to beta stage. By default, StrimziPodSets are used instead
          of StatefulSets. If needed, `UseStrimziPodSets` can be disabled in the feature
          gates configuration in the Cluster Operator.\n* The `ServiceAccountPatching`
          feature gate moves to GA. It cannot be disabled anymore and will be permanently
          enabled.\n* Remove support for Kafka 3.0.0 and 3.0.1\n* Add support for
          `simple` authorization and for the User Operator to the experimental `UseKRaft`
          feature gate. _(Note: Due to [KAFKA-13909](https://issues.apache.org/jira/browse/KAFKA-13909),
          broker restarts currently don't work when authorization is enabled.)_\n*
          Add CPU and network capacity overrides for Cruise Control capacity config\n*
          Use better encryption and digest algorithms when creating the PKCS12 stores.
          For existing clusters, the certificates will not be updated during upgrade
          but only next time the PKCS12 store is created. \n* Operator emits Kafka
          events to explain why it restarted a Kafka broker\n* Better configurability
          of the Kafka Admin client in the User Operator\n* Update Strimzi Kafka Bridge
          to 0.21.6\n### Supported Features\n* **Manages the Kafka Cluster** - Deploys
          and manages all of the components of this complex application, including
          dependencies like Apache ZooKeeper® that are traditionally hard to administer.\n*
          **Includes Kafka Connect** - Allows for configuration of common data sources
          and sinks to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for mirroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashboards\n### Upgrading your Clusters\nThe
          Strimzi Operator understands how to run and upgrade between a set of Kafka
          versions. When specifying a new version in your config, check to make sure
          you aren't using any features that may have been removed. See [the upgrade
          guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str)
          for more information.\n### Storage\nAn efficient data storage infrastructure
          is essential to the optimal performance of Apache Kafka®. Apache Kafka®
          deployed via Strimzi requires block storage. The use of file storage (for
          example, NFS) is not recommended.\nThe Strimzi Operator supports three types
          of data storage:\n* Ephemeral (Recommended for development only!)\n* Persistent\n*
          JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nStrimzi
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation to the current _main_
          branch as well as all releases can be found on our [website](https://strimzi.io/documentation).\n###
          Getting help\nIf you encounter any issues while using Strimzi, you can get
          help using:\n* [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          Contributing\nYou can contribute by:\n* Raising any issues you find using
          Strimzi\n* Fixing issues by opening Pull Requests\n* Improving documentation\n*
          Talking about Strimzi\nAll bugs, tasks or enhancements are tracked as [GitHub
          issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues
          which  might be a good start for new contributors are marked with [\"good-start\"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start)
          label.\nThe [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md)
          describes how to build Strimzi and how to  test your changes before submitting
          a patch or opening a PR.\nThe [Documentation Contributor Guide](https://strimzi.io/contributing/guide/)
          describes how to contribute to Strimzi documentation.\nIf you want to get
          in touch with us first before contributing, you can use:\n* [Strimzi mailing
          list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n* [Strimzi
          Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          License\nStrimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).\n"
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:544c4987f7f3ca558b04e9ef4357be204f088212b7b5b9ed0f909401a8bc5238
        - quay.io/strimzi/maven-builder@sha256:f2cb45234f3a7a010824a370d03032b706271cd0d1d260da929f046f9259d7e0
        - quay.io/strimzi/kafka@sha256:68d6f4cb5cc5ccad44de799bd9cf417c469a09e1024dbe42298942f25a09e004
        - quay.io/strimzi/kafka@sha256:d9258ddd453c3a53ca8e737b989d27725818da95f7a4fc879c9f07fe743c0cef
        - quay.io/strimzi/kafka@sha256:5d881605425ab0b37d0b9cac6957d00ba904df0db185ca6181c810151bdab8e7
        - quay.io/strimzi/operator@sha256:587db6f4389f56209f48853beedbfe249f6d590a4cb0000be2c42b57e43c2adb
        - quay.io/strimzi/kafka-bridge@sha256:77fc68da171c74d3a4b75b1030eaaa5eae12e200002f7aa2ee1969385eba1d66
        - quay.io/strimzi/jmxtrans@sha256:de1ca88b713b926b72df535aca2cfad24fb1b2bf40481065d6fa0c421b872320
        version: 0.30.0
      entries:
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.30.x
    - currentCSV: strimzi-cluster-operator.v0.31.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.3",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.3",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:6e5e207c70c25c595d3ce60268c0214c7518e95836c9b1212ef1eb2f008c7a76
          createdAt: "2022-09-20 21:07:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          **This is the last version of Strimzi with support for Kubernetes 1.16, 1.17 and 1.18!**
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.31 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.31 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.31.1
          * Kafka 3.1.2 and 3.2.3 (fixes CVE-2022-34917)
          * Make `sasl.server.max.receive.size` broker option user configurable
          * Update Strimzi Kafka Bridge to 0.22.1
          ### New in 0.31.0
          * Support for Apache Kafka 3.2.1
          * Pluggable Pod Security Profiles with built-in support for _restricted_ Kubernetes Security Profile
          * Support for leader election and running multiple operator replicas (1 active leader replicas and one or more stand-by replicas)
          * Support for IPv6 addresses being used in Strimzi issued certificates
          * `StrimziPodSet` reconciliation metrics
          * Updated Strimzi Kafka Bridge to 0.22.0 and Kafka Exporter to 1.6.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:03cc48bc5cf367f898338665b253855bed7d6096fe577adfe344da2f2cddcfe4
        - quay.io/strimzi/operator@sha256:6e5e207c70c25c595d3ce60268c0214c7518e95836c9b1212ef1eb2f008c7a76
        - quay.io/strimzi/kaniko-executor@sha256:06581fb615f0b13ddfe27fae26ba5836b12c6e28afd55a81cba7685c9ca4bf79
        - quay.io/strimzi/maven-builder@sha256:670ca70900f2229ef140a592b77f2107934601f8b050b79250eab17ff62099b6
        - quay.io/strimzi/kafka-bridge@sha256:3ddac62a4f8c8a3ebb72d5841d3826f7fd3fbb0e00eb75fb40bced0a576b86a9
        - quay.io/strimzi/kafka@sha256:b7c059bd72d8b063a96dde788d91d1ad87498bad3e73acba600929d595aeb7d6
        - quay.io/strimzi/kafka@sha256:430b3eb70e47a91d7b5b6f230d5be024c2de853099cc0437e219b443e0ead9f5
        - quay.io/strimzi/kafka@sha256:9553aae22d35448681cbbf7a831a28149fb159f3df58ccf8c8ec4ca7eb9c1fdf
        - quay.io/strimzi/kafka@sha256:83acc1e2cc1e64a88bdbb54b9818a6ad3658c3da60e72c4be365f824c5fcf57c
        - quay.io/strimzi/jmxtrans@sha256:52dd107062d78c29bb58c905ac654c39df72bbe3c25613023aa33e834930606d
        - quay.io/strimzi/kafka@sha256:652924a1dba46aa716caba8381f1e791c09035e31ce62fdcc81f9b69af36cd02
        version: 0.31.1
      entries:
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.31.x
    - currentCSV: strimzi-cluster-operator.v0.32.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.3.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.3.1",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:c5e0e5dca7503a8371b8d366db0f49e3fc68f62ea480262e39b6cc277991c97c
          createdAt: "2022-10-31 17:05:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **From the 0.32.0 release, Strimzi supports only Kubernetes 1.19 and newer! Kubernetes 1.16, 1.17 and 1.18 are not supported anymore.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.32.0/deploying.html#assembly-upgrade-str).
          ### New in 0.32.0
          * Add support for Kafka 3.3.1 and remove support for Kafka 3.1.0, 3.1.1, and 3.1.2
          * Update KafkaConnector CR status so the 'NotReady' condition is added if the connector or any tasks are reporting a 'FAILED' state.
          * Add auto-approval mechanism on KafkaRebalance resource when an optimization proposal is ready
          * The ControlPlaneListener feature gate moves to GA
          * Add client rack-awareness support to Strimzi Bridge pods
          * Add support for OpenTelemetry for distributed tracing
          * ZookeeperRoller considers unready pods
          * Support multiple operations per ACLRule
          * Add new listener type "cluster-ip" which is using per-broker services instead of the pod DNS names
          * Update Strimzi OAuth library to 0.11.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:8c9f38b45dd814e617c9c51cc90fd51e23d7dd23e8efe7e9eff013522bd6a5a0
        - quay.io/strimzi/kafka@sha256:bb97008b7f928ed98939b2008a7baf8909c88e726a59aaf607f2eeba55375cc0
        - quay.io/strimzi/kafka@sha256:680ae1958dbcb9da8ee4128a67c1163a6ee2744221f7d29f73d8bcc237fd0173
        - quay.io/strimzi/maven-builder@sha256:593e9980b786ca8ad3223151c3e17f16c7652ee0e86d722af4c34b023038ecc2
        - quay.io/strimzi/operator@sha256:c5e0e5dca7503a8371b8d366db0f49e3fc68f62ea480262e39b6cc277991c97c
        - quay.io/strimzi/jmxtrans@sha256:208928133b2854f60bb34912e07561a3511de895b815960e73d565be2d553918
        - quay.io/strimzi/kafka@sha256:4bacf5b54041bd0ab9199c8c5d51e8058758139db4b7abc380cf0400737716ef
        - quay.io/strimzi/kafka-bridge@sha256:aad0cee4381cc1c8c6638ab5732b135acd271b5889486ca9a33312cce67e4a02
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        version: 0.32.0
      entries:
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.32.x
    - currentCSV: strimzi-cluster-operator.v0.33.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:cf0d41fdb99b20e91063e23eb1d664817c18866c9a2f5c4430eac897e78d5d53
          createdAt: "2023-02-18 21:21:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.33 supports only Kubernetes 1.19 and newer! Kubernetes versions 1.16, 1.17 and 1.18 are not supported anymore since Strimzi 0.32.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.33.2/deploying.html#assembly-upgrade-str).
          ### New in 0.33
          _Note: Strimzi 0.33.2 add support for Kafka 3.4.0 which fixes CVE-2023-25194._
          * Add support for Kafka 3.3.2
          * Automatic restarts of failed Connect or Mirror Maker 2 connectors
          * Redesign of Strimzi User Operator to improve its scalability
          * Move to Java 17
          * Improved FIPS support
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/maven-builder@sha256:de9e51365e7acc63feae745618804e6286fd730f41c3b31582cbb7319a93b483
        - quay.io/strimzi/kafka@sha256:116cf543fbadadd0b45007ce24d44a0c79a62280e8faa1fc80df40291fbdf05d
        - quay.io/strimzi/kafka@sha256:9eca149cfc436580e3cdd6cee0ca318366b8d0ec53c2cf007e4e2c614b988921
        - quay.io/strimzi/kafka@sha256:301dc52ed5891ed4fa33e79ca3c344b1ac0eec7469dd46a5a7ac22064e053fc2
        - quay.io/strimzi/kafka-bridge@sha256:3a380d5f3ea509f90cdf95280d83b42f5de4f86ee3a512e02234e7639f50c9eb
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/kafka@sha256:30945aecb9d63cf8863c672222f130469c89aefa07078c95aaed52c40fd16146
        - quay.io/strimzi/jmxtrans@sha256:5ebdafa7a5e3718ed3274d6254c9e655613c272340094b997b1c97fde2598b4a
        - quay.io/strimzi/operator@sha256:cf0d41fdb99b20e91063e23eb1d664817c18866c9a2f5c4430eac897e78d5d53
        - quay.io/strimzi/kafka@sha256:fe9d3f8b029ff32d5d725e57384592914957314d63e92fda3abf4028fe4e66f6
        - quay.io/strimzi/kafka@sha256:40a847f8455ec6ea9ff9fe1fff20b1ed01f848326dc13977d13ac911f0cb40b7
        version: 0.33.2
      entries:
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.33.x
    - currentCSV: strimzi-cluster-operator.v0.34.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:f4ee3ca482a2dd19785f59634578a0850273444c1173a5699a99cf02f3a018fe
          createdAt: "2023-03-17 13:02:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.34 supports only Kubernetes 1.19 and newer! Kubernetes versions 1.16, 1.17 and 1.18 are not supported anymore since Strimzi 0.32.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.34.0/deploying.html#assembly-upgrade-str).
          ### New in 0.34
          * Add support for Apache Kafka 3.4.0 (fixes CVE-2023-25194) and remove support for Apache Kafka 3.2.x
          * Stable Pod identities for Kafka Connect and MirrorMaker 2 (disabled by default)
          * Use JDK HTTP client in the Kubernetes client instead of the OkHttp client
          * Add truststore configuration for HTTPS connections to Open Policy Agent server
          * Update Strimzi HTTP Bridge to 0.25.0 and strimzi Oauth to 0.12.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:37cf8d0fea1d2078511920190a72b49977ff4047b6ecbfed6b394c6f9e1e2ff0
        - quay.io/strimzi/kafka-bridge@sha256:51babca0f7a328f4db005c6a16daa8ac4a4306c46fe3ca89382919134aa09179
        - quay.io/strimzi/jmxtrans@sha256:62272e8a03da2144340bd7267f7a824d1ed70f4f3e30dc7aadfba9ee1b9baf76
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/maven-builder@sha256:f781b6b857d437c2de62585462e97b9fae357912c9f6711c1df39f431278878d
        - quay.io/strimzi/operator@sha256:f4ee3ca482a2dd19785f59634578a0850273444c1173a5699a99cf02f3a018fe
        - quay.io/strimzi/kafka@sha256:1da35fd7b36b113d1343e254fdc6efd97ca8660cc7977d8985aa3a3547d5ebb1
        - quay.io/strimzi/kafka@sha256:1ace2cc0db9653ac1b388477fa56c3c1cd97c7b3846eba3c7824dc58ea179476
        version: 0.34.0
      entries:
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.34.x
    - currentCSV: strimzi-cluster-operator.v0.35.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:06a94a3021cf028ccc1a49271f35f79216029e344536e664f196c1725ff2c663
          createdAt: "2023-06-07 16:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.35 is the last version with support for Kubernetes 1.19 and 1.20. From Strimzi 0.36, only Kubernetes 1.21 and newer will be supported.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.35.1/deploying.html#assembly-upgrade-str).
          ### New in 0.35
          * Move feature gate `UseStrimziPodSets` to GA and remove support for StatefulSets
          * Remove support for JMX Trans
          * Redesigned the Cluster and User Operator configuration to make it more efficient and flexible
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:39778b90c2b2afc30261e4ad5135805e1a10a2b60e2e53108fb9f80487f1208a
        - quay.io/strimzi/maven-builder@sha256:88a79eff3b3a386880a630658964b7754caed9e99dd6e645a4c0d23d0fdb47ee
        - quay.io/strimzi/operator@sha256:06a94a3021cf028ccc1a49271f35f79216029e344536e664f196c1725ff2c663
        - quay.io/strimzi/kafka@sha256:4de4874a7b722ad813f4dcc58acf509527bca0609999b81e70d81e3b38534d9d
        - quay.io/strimzi/kafka@sha256:0d910e7138cb49e1cd8cd84cef88bce35698b93ddd683a3398f1d485a3162693
        - quay.io/strimzi/kafka@sha256:54c6b25b31f51ef401c1b6e2a1b27432911e819bf3e502e05186f01be3f798e5
        - quay.io/strimzi/kafka-bridge@sha256:d6be183e492f8f88157ab9fe0af53950df8b6711a8a8c33da465de6064f6f86e
        version: 0.35.1
      entries:
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.35.x
    - currentCSV: strimzi-cluster-operator.v0.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:e9e03b31007c0f9b67a6a33e6b644e8c1ba35bd9976804ee3d22a04e4bf511b1
          createdAt: "2023-07-10 21:47:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!! Strimzi 0.36.1 supports only Kubernetes 1.21 and newer! Kubernetes versions 1.19 and 1.20 are not supported anymore since Strimzi 0.36.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.35.1/deploying.html#assembly-upgrade-str).
          ### New in 0.36
          * Add support for Apache Kafka 3.4.1, 3.5.0, and 3.5.1
          * Remove support for 3.3.1 and 3.3.2
          * Add support for _Kafka node pools_
          * Add support for _Unidirectional Topic Operator_
          * Enable SCRAM-SHA authentication in KRaft mode (supported in Apache Kafka 3.5.0 and newer)
          * Add support for insecure flag in Maven artifacts in Kafka Connect Build
          * Improve Kafka rolling update to avoid rolling broker in log recovery
          * Added support for Kafka Exporter topic exclude and consumer group exclude parameters
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:9f35f40be3d47b0cb1f6628f7ffd9c6a687ab9f564829f341dd66fd0759c064f
        - quay.io/strimzi/kafka@sha256:67c2b563a65c85c9393608f9af7eaf12ea48dcc0ef5f726d51f27273a3845186
        - quay.io/strimzi/operator@sha256:e9e03b31007c0f9b67a6a33e6b644e8c1ba35bd9976804ee3d22a04e4bf511b1
        - quay.io/strimzi/kafka@sha256:ea35817ff1ef295c682d70562f209ef658fcd87c61f467a4549d7a2f39e71700
        - quay.io/strimzi/kafka-bridge@sha256:fefee252895e9f94756ed5c9ea71749e3fa89281df01bc55878a244ca7d17697
        - quay.io/strimzi/kaniko-executor@sha256:5ff9fc7d0b7b694875ad0150fc787c582d42da3b4864d807ee8a6524e47b0e4e
        - quay.io/strimzi/maven-builder@sha256:2a469b40122027b981914db5b5c102be48280e039c3089963422c181174af226
        - quay.io/strimzi/kafka@sha256:a0da6a48b5bcbff355365f07df2d99ef6443d22c5690495f10a6417e761b3799
        version: 0.36.1
      entries:
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.36.x
    - currentCSV: strimzi-cluster-operator.v0.37.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:52f376e64b9b31d74162e26248c278e9bd301d2d8b4240ee5bcf780dfe2f8ef8
          createdAt: "2023-09-03 17:45:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.37.0/deploying.html#assembly-upgrade-str).
          ### New in 0.37
          * The StableConnectIdentites feature gate moves to a beta stage.
          * Support for the ppc64le platform
          * Version fields to the Kafka custom resource status to track installation and upgrade state
          * Infinite auto-restarts of Kafka Connect and Kafka Mirror Maker 2 connectors
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:1de6e073e8625a193fcfd11dd1a6717cc17596911c10bcebd792c00b08a7ebaf
        - quay.io/strimzi/kafka@sha256:2997ba1c169eb792a6004d7a41b674560515f3ea6e62ab8b9a1fdb747eab40c1
        - quay.io/strimzi/kafka@sha256:dfeb1eb7b94107c443b6bc82fcfee7fcf782516c83c6fc64451fb6a7e2de74a6
        - quay.io/strimzi/kafka-bridge@sha256:5b821f6f3396002ffa7c5804391b627ff9da3b494b2777dde518c55c2b73cefc
        - quay.io/strimzi/kaniko-executor@sha256:5729f1f0511da95e64e358cfabb3980ad9a660af2968118a87d6755f5d652a67
        - quay.io/strimzi/maven-builder@sha256:995be06a50b86c992fcee9f00e44f6a21c3406975e4e35802f10e88b72ffb9df
        - quay.io/strimzi/operator@sha256:52f376e64b9b31d74162e26248c278e9bd301d2d8b4240ee5bcf780dfe2f8ef8
        - quay.io/strimzi/kafka@sha256:38bbadcdb196a373e1698be890caa1dc01134ce91763d8f95c71768a117e1932
        version: 0.37.0
      entries:
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.37.x
    - currentCSV: strimzi-cluster-operator.v0.38.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:aee45bbe439335e45a0423e6565bae0d0727d85c7b807812949bfa127d467fe6
          createdAt: "2023-10-14 18:18:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.38.0/deploying.html#assembly-upgrade-str).
          ### New in 0.38
          * Add support for Apache Kafka 3.6.0 and remove support for 3.4.0 and 3.4.1
          * Container signing and SBOMs
          * Stopping Connect and MirrorMaker 2 connectors
          * Manual rolling updates of Kafka Connect and Kafka Mirror Maker 2 pods using annotation
          * Prevent scale-down of brokers that still contain any replicas
          * Run ZooKeeper- and KRaft-based clusters in parallel
          * Unidirectional Topic Operator improvements
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:0cc01438e8596643407b6b8ff05fafcea7ef54fa9410d4ae2492a9dbf41118b1
        - quay.io/strimzi/kafka@sha256:20d451fb32ec82eab3c16891198ace46bb9b9193f1f5511bb60ce249080dd7d4
        - quay.io/strimzi/kafka-bridge@sha256:13838889cc20a3a382ea64e749cad94f997c0b928840588f2d1befcf19247bea
        - quay.io/strimzi/kaniko-executor@sha256:4fe7a3f3f78e93369284a5ab456873c911230926ca071d1d88167e63d8ce376c
        - quay.io/strimzi/maven-builder@sha256:922d91dab9e2b8e14b6db25d8b48d7f4e3521bedd435468cc22039599ac4f888
        - quay.io/strimzi/operator@sha256:aee45bbe439335e45a0423e6565bae0d0727d85c7b807812949bfa127d467fe6
        - quay.io/strimzi/kafka@sha256:50e24c797455f9cdbb3cdde8e4f26bb8e1f0503469456da5d179017fd3e03ffe
        version: 0.38.0
      entries:
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.38.x
    - currentCSV: strimzi-cluster-operator.v0.39.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.6.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:50b49f9b5e554cca517fc2794e11319d5aaba1676494b285baa3a74fc79eb501
          createdAt: "2023-12-14 20:28:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.39.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.39 is the last minor release with support for Kubernetes 1.21 and 1.22.** From Strimzi 0.40 on, only Kubernetes 1.23 and newer will be supported.
          ### New in 0.39
          * Add support for Apache Kafka 3.5.2 and 3.6.1
          * The `StableConnectIdentities` feature gate moves to GA stage and is now permanently enabled without the possibility to disable it.
            All Connect and Mirror Maker 2 operands will now use StrimziPodSets.

          * The `KafkaNodePools` feature gate moves to the beta stage and is enabled by default.
            If needed, `KafkaNodePools` can be disabled in the feature gates configuration in the Cluster Operator.

          * The `UnidirectionalTopicOperator` feature gate moves to the beta stage and is enabled by default.
            If needed, `UnidirectionalTopicOperator` can be disabled in the feature gates configuration in the Cluster Operator.

          * Improved Kafka Connect metrics and dashboard example files
          * Allow specifying and managing KRaft metadata version
          * Add support for KRaft to KRaft upgrades (Apache Kafka upgrades for the KRaft-based clusters)
          * Improved Kafka Mirror Maker 2 dashboard example file
          * Support for rolling updates of KRaft controller nodes
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:dd1706f36031c96dfcd0da7b3f804168e5f076220caa302973455fdcd76bd03e
        - quay.io/strimzi/kaniko-executor@sha256:723b7add2473e73c45c81361e1708af4971c321b0391c1298a0557fb8ca4f7f6
        - quay.io/strimzi/maven-builder@sha256:6ce3c16284379d5db9e42a6f3e2b765c981091164a28fdd6bf927b6c5d4ab557
        - quay.io/strimzi/kafka@sha256:49b7530ada1175750ad5a706e545ba3594d85e1e3a90f98b0200a73fa9875057
        - quay.io/strimzi/kafka@sha256:84691b3afb67b312400e01572ff08b64c1e1b1d32a4276e0c21bba10c71c36fe
        - quay.io/strimzi/kafka-bridge@sha256:13838889cc20a3a382ea64e749cad94f997c0b928840588f2d1befcf19247bea
        - quay.io/strimzi/kafka@sha256:53b960411e852f5cb182ecc5d395a3562865f83e15e6251f3a9af0eb413feeed
        - quay.io/strimzi/kafka@sha256:2badc65532b483bba6c1f6a8b0c9895e993ce24e9ebd3cd953d0d17fbe36d27a
        - quay.io/strimzi/operator@sha256:50b49f9b5e554cca517fc2794e11319d5aaba1676494b285baa3a74fc79eb501
        version: 0.39.0
      entries:
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.39.x
    - currentCSV: strimzi-cluster-operator.v0.40.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:2228d89b53e3524e69d1bcf5c2cd373884ce696d9005389a6c78dad23310e837
          createdAt: "2024-03-09 21:27:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.40.0/deploying.html#assembly-upgrade-str).

          **From Strimzi 0.40 on, we support only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.40
          * Add support for Apache Kafka 3.7.0. Remove support for Apache Kafka 3.5.0, 3.5.1, and 3.5.2.
          * The `UseKRaft` feature gate moves to beta stage and is enabled by default. If needed, `UseKRaft` can be disabled in the feature gates configuration in the Cluster Operator.
          * Add support for ZooKeeper to KRaft migration by enabling the users to move from using ZooKeeper to store metadata to KRaft.
          * Add support for moving from dedicated controller-only KRaft nodes to mixed KRaft nodes
          * Added support for Kafka Exporter `offset.show-all` parameter
          * Prevent removal of the `broker` process role from KRaft mixed-nodes that have assigned partition-replicas
          * Improve broker scale-down prevention to continue in reconciliation when scale-down cannot be executed
          * Added support for Tiered Storage by enabling the configuration of custom storage plugins through the Kafka custom resource.
          * Update HTTP bridge to the latest 0.28.0 release
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:73deb1af0d93e9fa8955a5a412d0e7138384bdc80e7574e815d21796e446b636
        - quay.io/strimzi/kaniko-executor@sha256:3fb4ace22589fcf8d3f01b74c37b10b32d87e15d103ea59e0933b61c4e46ff4a
        - quay.io/strimzi/maven-builder@sha256:d02a9b48c77a44eeab6426c9af4adf75405217994bcb01ce0196e616041350d8
        - quay.io/strimzi/operator@sha256:2228d89b53e3524e69d1bcf5c2cd373884ce696d9005389a6c78dad23310e837
        - quay.io/strimzi/kafka@sha256:2c143eb10b66037ef877311eb767457e4b877d5ff2c47d393c76babec155caff
        - quay.io/strimzi/kafka@sha256:e2b9f2835498831bbf30ac2d651de143d57088bcb23943101b066e941db92aca
        - quay.io/strimzi/kafka@sha256:c491a95f9b9b58f406461faf222ef612ad8d2f4d7b53f7f3645b20c99d7d1583
        version: 0.40.0
      entries:
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.40.x
    - currentCSV: strimzi-cluster-operator.v0.41.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:4c687a4fac96ffc22433ca6d00fb1bd31722767b44c9d87c465f95b57fb70089
          createdAt: "2024-05-09 22:23:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.41.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.41 supports only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.41
          * Support for Apache Kafka 3.6.2
          * Metrics to monitor CA expiration
          * Support for JBOD storage in KRaft mode. (Note: JBOD support in KRaft mode is considered early access in Apache Kafka 3.7.x)
          * Support for changing topic replication factor by changing the `KafkaTopic` resources.
          * The `KafkaNodePools` feature gate moves to GA stage and is permanently enabled without the possibility to disable it. To use the Kafka Node Pool resources, you still need to use the `strimzi.io/node-pools: enabled` annotation on the `Kafka` custom resources.
          * Support for configuring the `externalIPs` field in node port type services.
          * The `UnidirectionalTopicOperator` feature gate moves to GA stage and is permanently enabled without the possibility to disable it. If the topics whose names start with `strimzi-store-topic` and `strimzi-topic-operator` still exist, you can delete them.
          * Improve validation of `KafkaMirrorMaker2` resource to make sure the Connect cluster is set to the target cluster
          * Continue reconciliation after failed manual rolling update using the `strimzi.io/manual-rolling-update` annotation (when the `ContinueReconciliationOnManualRollingUpdateFailure` feature gate is enabled)
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/maven-builder@sha256:17f09b07fc41a059e6a048cb766046d43a072cfd3a8266c97388a37543dadee8
        - quay.io/strimzi/kafka@sha256:c3a346235ce4dea42986648ca07f2482d8fc0bf654c1d441b1301a23999304db
        - quay.io/strimzi/kafka@sha256:bf49499980bbe5d4dd399de0dd396089b0fd9dd29bc46e14d39b20f1e826aec1
        - quay.io/strimzi/kafka@sha256:41811bb106d6182978a889eaedffeadffaa5ab1599c4cd8e4f5015b131c7ac30
        - quay.io/strimzi/kafka@sha256:6243d98890c665ef65b5f5fa2ffd046412281ffef31bc7a07cac37bd77b2579d
        - quay.io/strimzi/kafka-bridge@sha256:c833f930ead2d1e9054912a20689ee979d335f5e5ca8769e3c6b8a5514a0df96
        - quay.io/strimzi/operator@sha256:4c687a4fac96ffc22433ca6d00fb1bd31722767b44c9d87c465f95b57fb70089
        - quay.io/strimzi/kaniko-executor@sha256:9404037bd5431b012c8c633ec3345907667453c5532e5bfaaedc78286de22c62
        version: 0.41.0
      entries:
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.41.x
    - currentCSV: strimzi-cluster-operator.v0.42.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:ca5c1b70e1196deec4bf519921bfa55a967aa1ea15bd4c182bc56bea08d2e114
          createdAt: "2024-07-03 12:34:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.42.0/deploying.html#assembly-upgrade-str).

          **From version 0.40.0, Strimzi supports only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.42
          * Support for Apache Kafka 3.7.1
          * The UseKRaft feature gate moves to GA stage and is permanently enabled without the possibility of disabling it. To use KRaft (ZooKeeper-less Apache Kafka), you still need to use the strimzi.io/kraft: enabled annotation on the Kafka custom resources or migrate from an existing ZooKeeper-based cluster.
          * Update the base image used by Strimzi containers from UBI8 to UBI9
          * Add support for filename patterns when configuring trusted certificates
          * Enhance KafkaBridge resource with consumer inactivity timeout and HTTP consumer/producer enablement.
          * Add support for setting publishNotReadyAddresses on services for listener types other than internal.
          * Added support for configuring the quotas plugin with type strimzi or kafka in the Kafka custom resource. The Strimzi Quotas plugin version was updated to 0.3.1.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:6527c426f753521c07f3a3df2ee432273a41b109594096b4c6cae2e8e609d669
        - quay.io/strimzi/kafka@sha256:6c519531cdd5f40f326b295d8827344140efe2607a1826f7f1cab8c3427f43f7
        - quay.io/strimzi/kafka-bridge@sha256:fdddffb3de59ba48c05964bb0e760593088d24d1ee35b090112fe6cc6b553cd8
        - quay.io/strimzi/kafka@sha256:bc49c10b3e839e161e515c3a1e411d1f8df9de5e2c8d71aa7b63ccdbbae9e8d2
        - quay.io/strimzi/kaniko-executor@sha256:11821b0078a270612f1bcc37f5dc716ab273c974648c40f95fa4804352e647a7
        - quay.io/strimzi/operator@sha256:ca5c1b70e1196deec4bf519921bfa55a967aa1ea15bd4c182bc56bea08d2e114
        - quay.io/strimzi/maven-builder@sha256:342e8d0086c53cf7334127798e5bd3764b28f5e4909ba67d65f6d937dc3dba05
        - quay.io/strimzi/kafka@sha256:841a46e9c10481df3ef702412caf362b55ae07c30a3a5927073356cd36dc049f
        - quay.io/strimzi/kafka@sha256:c56c5abf94897f20f976603cba0678a33cdfbbcd6f9030a193f7684cb9894a6c
        version: 0.42.0
      entries:
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.42.x
    - currentCSV: strimzi-cluster-operator.v0.43.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:d3732245f34afaa6b571f1f912e1cd017891b22003d6367ae8ca059b2a5cb13c
          createdAt: "2024-08-22 15:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.43.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.43 is the last version with support for Kubernetes 1.23 and 1.24.** From Strimzi 0.44, only Kubernetes versions 1.25 and newer will be supported.
          ### New in 0.43
          * Support for Apache Kafka 3.8.0
          * Removed support for Apache Kafka 3.6.0, 3.6.1, and 3.6.2
          * Specifying additional volumes and volume mounts in Strimzi custom resources
          * Support for custom Cruise Control API users
          * Unregistration of KRaft nodes after scale-down
          * Kafka Exporter updated to 1.8.0 (and the Grafana dashboard updated to work with it)
          * Service account based authentication
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:80f329ecd3ba96fac12e75be874dce0debc8395337737581de34238ce11c8fb1
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:6414f8f3985f221b0eda53867c00494140108d17579c4dcbbc93934bb00e5d49
        - quay.io/strimzi/kafka@sha256:d27cf5bf3624e18ba705246b14f92d12b8a15eb8551b13abf027cf71b549baf0
        - quay.io/strimzi/kafka@sha256:ad06ead5125bce87498b1582a76e78143a6d21408ac4b12bb7e6007aa4cabd84
        - quay.io/strimzi/operator@sha256:d3732245f34afaa6b571f1f912e1cd017891b22003d6367ae8ca059b2a5cb13c
        - quay.io/strimzi/kafka@sha256:b2f4e8895e9ed66228512ce14aec685101fa75dc329128d47cb1c15e3aef3c9c
        version: 0.43.0
      entries:
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.43.x
    - currentCSV: strimzi-cluster-operator.v0.44.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:b07b81f7e282dea2e4e29a7c93cfdd911d4715a5c32fe4c4e3d7c71cba3091e8
          createdAt: "2024-08-22 15:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.44.0/deploying.html#assembly-upgrade-str).

          **From version 0.44, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**

          **ZooKeeper support will be soon removed from Apache Kafka and Strimzi.** Currently, the last Strimzi version with ZooKeeper support is expected to be Strimzi 0.45. Please plan your migration to KRaft (ZooKeeper-less Apache Kafka) accordingly. Follow the [documentation](https://strimzi.io/docs/operators/latest/full/deploying.html#assembly-kraft-mode-str) for more details.

          **Kafka Mirror Maker 1 support will be soon removed from Apache Kafka and Strimzi.** Currently, the last Strimzi version with Mirror Maker 1 support is expected to be Strimzi 0.45. Please plan your migration to Mirror Maker 2 or another mirroring tool.

          ### New in 0.44
          * Managing offsets in Connect and Mirror Maker 2 connectors through custom resources
          * Templating `advertisedHost` and `host` fields in listener configuration
          * Configuring environment variables in operands based on Secrets and ConfigMaps
          * Disabling Pod disruption Budget management
          * Auto-rebalancing at scale-up/scale-down
          * `ContinueReconciliationOnManualRollingUpdateFailure` feature gate moves to the beta stage and is enabled by default
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:9a4224635e09ece2c1d3ffe9d715813cf63037461c9f06c1b5f53554aed5ecdc
        - quay.io/strimzi/kafka-bridge@sha256:cc95ce9f496057f3b18d5e8297b40a410cc66f9fd4a2227615971458a26bac25
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:7b55f971a3b62923a9a61a61c82f542e07deeb6ffdbaede68244ba358a0853f2
        - quay.io/strimzi/operator@sha256:b07b81f7e282dea2e4e29a7c93cfdd911d4715a5c32fe4c4e3d7c71cba3091e8
        - quay.io/strimzi/kafka@sha256:238bb080825b7ecebaa819be09788bc3d056ae8162ebcd41444076dc22f85740
        - quay.io/strimzi/kafka@sha256:98d332b7fa5c9e48f60b516e00ed5812bd9268e014fd2337284e7cc3e664f1eb
        version: 0.44.0
      entries:
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.44.x
    - currentCSV: strimzi-cluster-operator.v0.45.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:58e8722216b0b0e8b1dd79109725238fa4016a16cc6a9fca0e327bccb1506810
          createdAt: "2025-06-27 01:38:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.45.1/deploying.html#assembly-upgrade-str).

          **From version 0.45, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**

          **ZooKeeper and Kafka Mirror Maker 1 support will be removed from Apache Kafka and Strimzi.** Strimzi 0.45 is the last minor Strimzi version with support for ZooKeeper-based Apache Kafka clusters and MirrorMaker 1 deployments. Please make sure to [migrate to KRaft](https://strimzi.io/docs/operators/latest/full/deploying.html#assembly-kraft-mode-str) and MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          **Strimzi EnvVar Configuration Provider and Strimzi MirrorMaker 2 Extensions will be removed from Strimzi.** Strimzi 0.45 is the last Strimzi version to include the Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0). Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.
          ### New in 0.45.1
          * Add support for Kafka 3.9.1
          * Update HTTP bridge to 0.31.2
          * Update Strimzi Kafka OAuth library to 0.15.1
          * Fixed bug which may lead to metadata loss within the cluster when restarting a KRaft migration after a previous rollback due to missing `/migration` znode deletion.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:e98fbe49cadbc5d3afbd8f17abea218f216c95d52f4fa1fe514a7a988624da24
        - quay.io/strimzi/kafka@sha256:ba52ed046b1dccdbd96f4e68057ce014d862a7c9c1fc670760c023b9aa09f23f
        - quay.io/strimzi/kafka-bridge@sha256:c462441a966f0732855e684b4840563bfad45391345c9511fa49e1630ba9c3e3
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:770e410f1e7e0e92c31bae6c19831ba36d59dd11b44e90c159aef16a965aa998
        - quay.io/strimzi/kafka@sha256:173cb5793756f75739c51a83bbe06d7f0094fd003a2982160a5cc6e216cd6d28
        - quay.io/strimzi/kafka@sha256:6a05924839a2c7da0431408e6d6223a055e75d88b3f057ac5caa7f9097227f56
        - quay.io/strimzi/operator@sha256:58e8722216b0b0e8b1dd79109725238fa4016a16cc6a9fca0e327bccb1506810
        version: 0.45.1
      entries:
      - name: strimzi-cluster-operator.v0.45.1
        version: 0.45.1
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.45.x
    - currentCSV: strimzi-cluster-operator.v0.46.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.0.0",
                      "metadataVersion": "4.0-IV3",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:6a22e4968e9d409405423d12c04a0e3fecaea2ee2794bd27de7cf21f59453265
          createdAt: "2025-06-26 11:58:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.46.0/deploying.html#assembly-upgrade-str).

          **From version 0.46, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**
          **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed. Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!
          **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images. Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.
          **Support for MirrorMaker 1 has been removed. Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.
          ### New in 0.46.1
          * Update CSV file for OLM installation to add missing permission for broker Pods to "get" nodes Kubernetes resource.
          * Allow the injection of annotations on the deployment object in Helm charts.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:3e1815aed3f74b8573cebfced66b2c78ca0cf7c746c26db2002380cd39d09ae6
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:b532234c234fc30fecec992ad1a6154c5ef08ebf9a20ec8bc62e0c88d791ff36
        - quay.io/strimzi/kafka@sha256:f3837f8cb2a60e906e84486128c41c398424f4085df2348e9e0486bfc12089fd
        - quay.io/strimzi/kafka@sha256:82273d873ce21a73f8f3118bf61627cb15734b43ea7ca92d43b290340b44e485
        - quay.io/strimzi/operator@sha256:6a22e4968e9d409405423d12c04a0e3fecaea2ee2794bd27de7cf21f59453265
        version: 0.46.1
      entries:
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.46.x
    - currentCSV: strimzi-cluster-operator.v0.47.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.0.0",
                      "metadataVersion": "4.0-IV3",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:589c4d63641d9a944462dd682f6e5febe8b38ff55b9253949b061aca16feb154
          createdAt: "2025-07-12 17:21:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.47 is the last Strimzi version with support for Kubernetes 1.25 and 1.26.**
            From Strimzi 0.48.0 on, we will support only Kubernetes 1.27 and newer.

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.47.0
          * Support for Kafka 3.9.1
          * Progress tracking for Cruise Control rebalances
          * Support for Kubernetes Image Volumes to mount custom plugins
          * Support for [JsonTemplateLayout](https://logging.apache.org/log4j/2.x/manual/json-template-layout.html) in Operators, Cruise Control, and Kafka 4.0.0
          * Support for Strimzi Metrics Reporter for the Kafka brokers / controllers
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/operator@sha256:589c4d63641d9a944462dd682f6e5febe8b38ff55b9253949b061aca16feb154
        - quay.io/strimzi/maven-builder@sha256:c67a3ad3d92420ac8b7f8435eeec53ebffaea4e894e65414a7c47c20e3faba27
        - quay.io/strimzi/kafka@sha256:978e2ed5f0d315e8bf0ab0503ee839acbd9da01ca427e6581b59ac25d36bc9c9
        - quay.io/strimzi/kafka@sha256:8a793cdf5084cdb72c6663419530985158b16330a518720e9e9d1d96babbaf7e
        - quay.io/strimzi/kafka@sha256:551882f47a6d2dc8300bcfca251fffffc3379acb2724a7efe16f17c0c4f529fb
        - quay.io/strimzi/kafka-bridge@sha256:428c5587f547dabfc995a73745569d113380c7e6b8bf0b2504c6a9a00d6c4608
        version: 0.47.0
      entries:
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.47.x
    - currentCSV: strimzi-cluster-operator.v0.48.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.1.0",
                      "metadataVersion": "4.1-IV1",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.1.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
          createdAt: "2025-09-25 10:32:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.48.0 supports only Kubernetes 1.27 and newer.**
            Kubernetes 1.25 and 1.26 are not supported anymore

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.48.0
          * Support for Kafka 4.1.0
          * Add monitoring of custom resources using kubernetes-state-metrics (KSM)
          * Support for type: custom client authentication
          * Added support for Strimzi Metrics Reporter to Kafka Connect, Mirror Maker 2 and Kafka Bridge.
          * Added support for KIP-745 in Kafka Connector and Mirror Maker 2, allowing the usage of includeTasks and onlyFailed arguments in Kafka connectors restart.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:64380efcb8da87fdc0e0de238a20456be9abb830174c2f2331882b7a84657e57
        - quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
        - quay.io/strimzi/kafka@sha256:48db05ad2e9fa39eb388089e7e25654bcbf4da1ddc3a5635d675056c28bbcb96
        - quay.io/strimzi/kafka@sha256:34450afc1a3399a9026fc3c6ec90e9f6adfec9faf6520545b4d84e90cda34964
        - quay.io/strimzi/kafka-bridge@sha256:53034f64f0b672f10b5bacea1c7a25132f118df7fd5c9032c4dbf702ed93796a
        version: 0.48.0
      entries:
      - name: strimzi-cluster-operator.v0.48.0
        version: 0.48.0
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.48.x
    defaultChannel: stable
    packageName: strimzi-kafka-operator
    provider:
      name: Strimzi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-nifi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nifi-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:05946a5e8f040e8ee46d56f2d4cca3d97e02d086e6d8224200204de4d76851fa
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:05946a5e8f040e8ee46d56f2d4cca3d97e02d086e6d8224200204de4d76851fa
        version: 24.11.1
      entries:
      - name: nifi-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: nifi-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:e5c3c4e2ad1c455fbb04a60658936bfb44c2bc9d2b616737ce1a0eba201b7a3f
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiClusterSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:e5c3c4e2ad1c455fbb04a60658936bfb44c2bc9d2b616737ce1a0eba201b7a3f
        version: 25.3.0
      entries:
      - name: nifi-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: nifi-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiClusterSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
        version: 25.7.0
      entries:
      - name: nifi-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: nifi-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiClusterSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
        version: 25.7.0
      entries:
      - name: nifi-operator.v25.7.0
        version: 25.7.0
      - name: nifi-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-nifi-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/en/products/edge/validated-patterns
    name: patterns-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: patterns-operator.v0.0.62
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gitops.hybrid-cloud-patterns.io/v1alpha1",
                "kind": "Pattern",
                "metadata": {
                  "name": "pattern-sample"
                },
                "spec": {
                  "clusterGroupName": "hub",
                  "gitSpec": {
                    "targetRepo": "https://github.com/validatedpatterns/multicloud-gitops",
                    "targetRevision": "main"
                  },
                  "multiSourceConfig": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/hybridcloudpatterns/patterns-operator:0.0.62
          createdAt: "2025-09-02 16:14:40"
          description: "An operator to deploy and manage architecture patterns from
            https://validatedpatterns.io. \nThis operator collects some Analytics
            like cluster version, cloud type, etc. To disable this\nadd \"ANALYTICS:
            false\" to the environment variables in the subscription."
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/validatedpatterns/patterns-operator
          support: validatedpatterns@googlegroups.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pattern is the Schema for the patterns API
            displayName: Pattern
            kind: Pattern
            name: patterns.gitops.hybrid-cloud-patterns.io
            version: v1alpha1
        description: Deploys and manages architecture patterns from https://validatedpatterns.io
        displayName: Validated Patterns Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Validated
        - Pattern
        - Blueprint
        - Architecture
        links:
        - name: Hybrid Cloud Patterns
          url: https://validatedpatterns.io
        maintainers:
        - email: validatedpatterns@googlegroups.com
          name: patterns-team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/en/products/edge/validated-patterns
        relatedImages:
        - quay.io/hybridcloudpatterns/kube-rbac-proxy@sha256:409a464692c7db3cf68893376479dca1b53802412237637725927f7e78aa0844
        - quay.io/hybridcloudpatterns/patterns-operator@sha256:15ed44938b112bcf9b3a9a566749b39ba2b9053a6a8b9b499ef247977c18496c
        version: 0.0.62
      entries:
      - name: patterns-operator.v0.0.62
        version: 0.0.62
      - name: patterns-operator.v0.0.61
        version: 0.0.61
      - name: patterns-operator.v0.0.60
        version: 0.0.60
      - name: patterns-operator.v0.0.59
        version: 0.0.59
      - name: patterns-operator.v0.0.58
        version: 0.0.58
      - name: patterns-operator.v0.0.57
        version: 0.0.57
      - name: patterns-operator.v0.0.56
        version: 0.0.56
      - name: patterns-operator.v0.0.55
        version: 0.0.55
      - name: patterns-operator.v0.0.54
        version: 0.0.54
      - name: patterns-operator.v0.0.53
        version: 0.0.53
      - name: patterns-operator.v0.0.52
        version: 0.0.52
      - name: patterns-operator.v0.0.51
        version: 0.0.51
      - name: patterns-operator.v0.0.50
        version: 0.0.50
      - name: patterns-operator.v0.0.49
        version: 0.0.49
      - name: patterns-operator.v0.0.48
        version: 0.0.48
      - name: patterns-operator.v0.0.47
        version: 0.0.47
      - name: patterns-operator.v0.0.46
        version: 0.0.46
      - name: patterns-operator.v0.0.45
        version: 0.0.45
      - name: patterns-operator.v0.0.44
        version: 0.0.44
      - name: patterns-operator.v0.0.43
        version: 0.0.43
      - name: patterns-operator.v0.0.42
        version: 0.0.42
      - name: patterns-operator.v0.0.41
        version: 0.0.41
      - name: patterns-operator.v0.0.40
        version: 0.0.40
      - name: patterns-operator.v0.0.39
        version: 0.0.39
      - name: patterns-operator.v0.0.38
        version: 0.0.38
      - name: patterns-operator.v0.0.37
        version: 0.0.37
      - name: patterns-operator.v0.0.36
        version: 0.0.36
      - name: patterns-operator.v0.0.35
        version: 0.0.35
      - name: patterns-operator.v0.0.34
        version: 0.0.34
      - name: patterns-operator.v0.0.33
        version: 0.0.33
      - name: patterns-operator.v0.0.32
        version: 0.0.32
      - name: patterns-operator.v0.0.31
        version: 0.0.31
      - name: patterns-operator.v0.0.30
        version: 0.0.30
      - name: patterns-operator.v0.0.29
        version: 0.0.29
      - name: patterns-operator.v0.0.28
        version: 0.0.28
      - name: patterns-operator.v0.0.27
        version: 0.0.27
      - name: patterns-operator.v0.0.26
        version: 0.0.26
      - name: patterns-operator.v0.0.25
        version: 0.0.25
      - name: patterns-operator.v0.0.24
        version: 0.0.24
      - name: patterns-operator.v0.0.23
        version: 0.0.23
      - name: patterns-operator.v0.0.22
        version: 0.0.22
      - name: patterns-operator.v0.0.21
        version: 0.0.21
      - name: patterns-operator.v0.0.20
        version: 0.0.20
      - name: patterns-operator.v0.0.19
        version: 0.0.19
      - name: patterns-operator.v0.0.18
        version: 0.0.18
      - name: patterns-operator.v0.0.17
        version: 0.0.17
      - name: patterns-operator.v0.0.16
        version: 0.0.16
      - name: patterns-operator.v0.0.15
        version: 0.0.15
      - name: patterns-operator.v0.0.14
        version: 0.0.14
      - name: patterns-operator.v0.0.13
        version: 0.0.13
      - name: patterns-operator.v0.0.12
        version: 0.0.12
      - name: patterns-operator.v0.0.11
        version: 0.0.11
      - name: patterns-operator.v0.0.10
        version: 0.0.10
      - name: patterns-operator.v0.0.9
        version: 0.0.9
      - name: patterns-operator.v0.0.8
        version: 0.0.8
      - name: patterns-operator.v0.0.7
        version: 0.0.7
      - name: patterns-operator.v0.0.6
        version: 0.0.6
      - name: patterns-operator.v0.0.5
        version: 0.0.5
      - name: patterns-operator.v0.0.4
        version: 0.0.4
      - name: patterns-operator.v0.0.3
        version: 0.0.3
      - name: patterns-operator.v0.0.2
        version: 0.0.2
      - name: patterns-operator.v0.0.1
        version: 0.0.1
      name: fast
    defaultChannel: fast
    packageName: patterns-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com/en/products/edge/validated-patterns
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sns-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sns-controller.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "PlatformApplication",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "PlatformEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sns-controller:1.2.2
          createdAt: "2025-09-30T21:19:34Z"
          description: AWS SNS controller is a service controller for managing SNS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PlatformApplication represents the state of an AWS sns PlatformApplication
              resource.
            displayName: PlatformApplication
            kind: PlatformApplication
            name: platformapplications.sns.services.k8s.aws
            version: v1alpha1
          - description: PlatformEndpoint represents the state of an AWS sns PlatformEndpoint
              resource.
            displayName: PlatformEndpoint
            kind: PlatformEndpoint
            name: platformendpoints.sns.services.k8s.aws
            version: v1alpha1
          - description: Subscription represents the state of an AWS sns Subscription
              resource.
            displayName: Subscription
            kind: Subscription
            name: subscriptions.sns.services.k8s.aws
            version: v1alpha1
          - description: Topic represents the state of an AWS sns Topic resource.
            displayName: Topic
            kind: Topic
            name: topics.sns.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Simple Notification Service (SNS) resources in AWS from within your Kubernetes cluster.

          **About Amazon SNS**

          Amazon Simple Notification Service (Amazon SNS) is a managed service that provides message delivery from publishers to subscribers (also known as producers and consumers). Publishers communicate asynchronously with subscribers by sending messages to a topic, which is a logical access point and communication channel. Clients can subscribe to the SNS topic and receive published messages using a supported endpoint type, such as Amazon Kinesis Data Firehose, Amazon SQS, AWS Lambda, HTTP, email, mobile push notifications, and mobile text messages (SMS).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SNS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sns
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SNS Developer Resources
          url: https://aws.amazon.com/sns/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sns maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sns-controller:1.2.2
        version: 1.2.2
      entries:
      - name: ack-sns-controller.v1.2.2
        version: 1.2.2
      - name: ack-sns-controller.v1.2.1
        version: 1.2.1
      - name: ack-sns-controller.v1.2.0
        version: 1.2.0
      - name: ack-sns-controller.v1.1.16
        version: 1.1.16
      - name: ack-sns-controller.v1.1.15
        version: 1.1.15
      - name: ack-sns-controller.v1.1.14
        version: 1.1.14
      - name: ack-sns-controller.v1.1.13
        version: 1.1.13
      - name: ack-sns-controller.v1.1.12
        version: 1.1.12
      - name: ack-sns-controller.v1.1.11
        version: 1.1.11
      - name: ack-sns-controller.v1.1.10
        version: 1.1.10
      - name: ack-sns-controller.v1.1.9
        version: 1.1.9
      - name: ack-sns-controller.v1.1.8
        version: 1.1.8
      - name: ack-sns-controller.v1.1.7
        version: 1.1.7
      - name: ack-sns-controller.v1.1.6
        version: 1.1.6
      - name: ack-sns-controller.v1.1.5
        version: 1.1.5
      - name: ack-sns-controller.v1.1.4
        version: 1.1.4
      - name: ack-sns-controller.v1.1.3
        version: 1.1.3
      - name: ack-sns-controller.v1.1.2
        version: 1.1.2
      - name: ack-sns-controller.v1.1.1
        version: 1.1.1
      - name: ack-sns-controller.v1.1.0
        version: 1.1.0
      - name: ack-sns-controller.v1.0.14
        version: 1.0.14
      - name: ack-sns-controller.v1.0.13
        version: 1.0.13
      - name: ack-sns-controller.v1.0.12
        version: 1.0.12
      - name: ack-sns-controller.v1.0.11
        version: 1.0.11
      - name: ack-sns-controller.v1.0.9
        version: 1.0.9
      - name: ack-sns-controller.v1.0.6
        version: 1.0.6
      - name: ack-sns-controller.v1.0.5
        version: 1.0.5
      - name: ack-sns-controller.v1.0.4
        version: 1.0.4
      - name: ack-sns-controller.v1.0.3
        version: 1.0.3
      - name: ack-sns-controller.v1.0.2
        version: 1.0.2
      - name: ack-sns-controller.v1.0.1
        version: 1.0.1
      - name: ack-sns-controller.v1.0.0
        version: 1.0.0
      - name: ack-sns-controller.v0.0.12
        version: 0.0.12
      - name: ack-sns-controller.v0.0.11
        version: 0.0.11
      - name: ack-sns-controller.v0.0.10
        version: 0.0.10
      - name: ack-sns-controller.v0.0.9
        version: 0.0.9
      - name: ack-sns-controller.v0.0.8
        version: 0.0.8
      - name: ack-sns-controller.v0.0.7
        version: 0.0.7
      - name: ack-sns-controller.v0.0.5
        version: 0.0.5
      - name: ack-sns-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-sns-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Helm Community
      provider-url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
    name: cockroachdb
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroachdb.v5.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1alpha1",
                "kind": "Cockroachdb",
                "metadata": {
                  "name": "cockroachdb-sample"
                },
                "spec": {
                  "clusterDomain": "cluster.local",
                  "conf": {
                    "attrs": [],
                    "cache": "25%",
                    "cluster-name": "",
                    "disable-cluster-name-verification": false,
                    "http-port": 8080,
                    "join": [],
                    "locality": "",
                    "logtostderr": "INFO",
                    "max-sql-memory": "25%",
                    "port": 26257,
                    "single-node": false,
                    "sql-audit-dir": ""
                  },
                  "image": {
                    "credentials": {},
                    "pullPolicy": "IfNotPresent",
                    "repository": "cockroachdb/cockroach",
                    "tag": "v20.2.4"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [],
                    "labels": {},
                    "paths": [
                      "/"
                    ],
                    "tls": []
                  },
                  "init": {
                    "affinity": {},
                    "annotations": {},
                    "labels": {
                      "app.kubernetes.io/component": "init"
                    },
                    "nodeSelector": {},
                    "resources": {},
                    "tolerations": []
                  },
                  "labels": {},
                  "networkPolicy": {
                    "enabled": false,
                    "ingress": {
                      "grpc": [],
                      "http": []
                    }
                  },
                  "service": {
                    "discovery": {
                      "annotations": {},
                      "labels": {
                        "app.kubernetes.io/component": "cockroachdb"
                      }
                    },
                    "ports": {
                      "grpc": {
                        "external": {
                          "name": "grpc",
                          "port": 26257
                        },
                        "internal": {
                          "name": "grpc-internal",
                          "port": 26257
                        }
                      },
                      "http": {
                        "name": "http",
                        "port": 8080
                      }
                    },
                    "public": {
                      "annotations": {},
                      "labels": {
                        "app.kubernetes.io/component": "cockroachdb"
                      },
                      "type": "ClusterIP"
                    }
                  },
                  "serviceMonitor": {
                    "annotations": {},
                    "enabled": false,
                    "interval": "10s",
                    "labels": {}
                  },
                  "statefulset": {
                    "annotations": {},
                    "args": [],
                    "budget": {
                      "maxUnavailable": 1
                    },
                    "env": [],
                    "labels": {
                      "app.kubernetes.io/component": "cockroachdb"
                    },
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "podAffinity": {},
                    "podAntiAffinity": {
                      "topologyKey": "kubernetes.io/hostname",
                      "type": "soft",
                      "weight": 100
                    },
                    "podManagementPolicy": "Parallel",
                    "priorityClassName": "",
                    "replicas": 3,
                    "resources": {},
                    "secretMounts": [],
                    "tolerations": [],
                    "topologySpreadConstraints": {
                      "maxSkew": 1,
                      "topologyKey": "topology.kubernetes.io/zone",
                      "whenUnsatisfiable": "ScheduleAnyway"
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    }
                  },
                  "storage": {
                    "hostPath": "",
                    "persistentVolume": {
                      "annotations": {},
                      "enabled": true,
                      "labels": {},
                      "size": "100Gi",
                      "storageClass": ""
                    }
                  },
                  "tls": {
                    "certs": {
                      "clientRootSecret": "cockroachdb-root",
                      "nodeSecret": "cockroachdb-node",
                      "provided": false,
                      "tlsSecret": false
                    },
                    "enabled": false,
                    "init": {
                      "image": {
                        "credentials": {},
                        "pullPolicy": "IfNotPresent",
                        "repository": "cockroachdb/cockroach-k8s-request-cert",
                        "tag": "0.4"
                      }
                    },
                    "serviceAccount": {
                      "create": true,
                      "name": ""
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          certified: "false"
          containerImage: quay.io/helmoperators/cockroachdb:v5.0.4
          createdAt: 2021-01-23T10-00-00Z
          description: CockroachDB Operator based on the CockroachDB helm chart
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/dmesser/cockroachdb-operator
          support: dmesser
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a CockroachDB cluster
            displayName: CockroachDB
            kind: Cockroachdb
            name: cockroachdbs.charts.operatorhub.io
            version: v1alpha1
        description: |
          CockroachDB is a scalable, survivable, strongly-consistent SQL database.
          ## About this Operator
          This Operator is based on a Helm chart for CockroachDB. It supports the same parameters like the official [CockroachDB Helm Chart](https://github.com/cockroachdb/helm-charts/tree/master/cockroachdb).

          ## Core capabilities
          * **StatefulSet** - Sets up a dynamically scalable CockroachDB cluster using a Kubernetes StatefulSet
          * **Expand Replicas** - Supports expanding the set of replicas by simply editing your object
          * **Dashboard** - Installs the CockroachDB user interface to administer your cluster. Easily expose it via an Ingress rule.
          Review all of the [confiuguration options](https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration) to best run your database instance. The example configuration is derived from the chart's [`values.yaml`](https://github.com/helm/charts/blob/master/stable/cockroachdb/values.yaml).
          ## Using the cluster
          The resulting cluster endpoint can be consumed from a `Service` that follows the pattern: `<StatefulSet-name>-public`. For example to connect using the command line client, use something like the following to obtain the name of the service:
          ```
          kubectl get service -l chart=cockroachdb-5.0.4
          NAME                                           TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)              AGE
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb          ClusterIP   None             <none>        26257/TCP,8080/TCP   24m
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public   ClusterIP   10.106.249.134   <none>        26257/TCP,8080/TCP   24m
          ```
          Then you can use the CockroachDB command line client to connect to the database cluster:
          ```
          kubectl run -it --rm cockroach-client --image=cockroachdb/cockroach --restart=Never --command -- ./cockroach sql --insecure --host example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public
          ```
          ## Before you start
          This Operator requires a cluster with PV support in order to run correctly.

          Note that for a production cluster, you will likely want to override the following parameters in [`values.yaml`](values.yaml) with your own values.

          - `statefulset.resources.requests.memory` and `statefulset.resources.limits.memory` allocate memory resources to CockroachDB pods in your cluster.
          - `conf.cache` and `conf.max-sql-memory` are memory limits that we recommend setting to 1/4 of the above resource allocation. When running CockroachDB, you must set these limits explicitly to avoid running out of memory.
          - `storage.persistentVolume.size` defaults to `100Gi` of disk space per pod, which you may increase or decrease for your use case.
          - `storage.persistentVolume.storageClass` uses the default storage class for your environment. We strongly recommend that you specify a storage class which uses an SSD.
          - `tls.enabled` must be set to `yes`/`true` to deploy in secure mode.

          For more information on overriding the `values.yaml` parameters, please see:
          > <https://www.cockroachlabs.com/docs/stable/orchestrate-cockroachdb-with-kubernetes.html#step-2-start-cockroachdb>

          If you are running in secure mode (with configuration parameter `tls.enabled` set to `yes`/`true`) and `tls.certs.provided` set to `no`/`false`), then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          When `tls.certs.provided` is set to `yes`/`true`, this chart will use certificates created outside of Kubernetes. You may want to use this if you want to use a different certificate authority from the one being used by Kubernetes or if your Kubernetes cluster doesn't fully support certificate-signing requests. To use this, first set up your certificates and load them into your Kubernetes cluster as Secrets using the commands below:

          ```
          mkdir certs
          mkdir my-safe-directory
          cockroach cert create-ca --certs-dir=certs --ca-key=my-safe-directory/ca.key
          cockroach cert create-client root --certs-dir=certs --ca-key=my-safe-directory/ca.key
          kubectl create secret generic cockroachdb-root --from-file=certs
          cockroach cert create-node --certs-dir=certs --ca-key=my-safe-directory/ca.key localhost 127.0.0.1 eerie-horse-cockroachdb-public eerie-horse-cockroachdb-public.default eerie-horse-cockroachdb-public.default.svc.cluster.local *.eerie-horse-cockroachdb *.eerie-horse-cockroachdb.default *.eerie-horse-cockroachdb.default.svc.cluster.local
          kubectl create secret generic cockroachdb-node --from-file=certs
          ```

          Set `tls.certs.tlsSecret` to `yes/true` if you make use of [cert-manager][3] in your cluster.

          [cert-manager][3] stores generated certificates in dedicated TLS secrets. Thus, they are always named:

          * `ca.crt`
          * `tls.crt`
          * `tls.key`

          On the other hand, CockroachDB also demands dedicated certificate filenames:

          * `ca.crt`
          * `node.crt`
          * `node.key`
          * `client.root.crt`
          * `client.root.key`

          By activating `tls.certs.tlsSecret` we benefit from projected secrets and convert the TLS secret filenames to their according CockroachDB filenames.

          If you are running in secure mode, then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          Confirm that all pods are `Running` successfully and init has been completed:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS      RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running     0          1m
          my-release-cockroachdb-1            1/1       Running     0          1m
          my-release-cockroachdb-2            1/1       Running     0          1m
          my-release-cockroachdb-init-k6jcr   0/1       Completed   0          1m
          ```

          Confirm that persistent volumes are created and claimed for each pod:

          ```shell
          kubectl get pv
          ```

          ```
          NAME                                       CAPACITY   ACCESS MODES   RECLAIM POLICY   STATUS    CLAIM                                      STORAGECLASS   REASON    AGE
          pvc-64878ebf-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-0   standard                 51s
          pvc-64945b4f-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-1   standard                 51s
          pvc-649d920d-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-2   standard                 51s
          ```

          ## Upgrading the cluster

          ### Chart version 3.0.0 and after

          Launch a temporary interactive pod and start the built-in SQL client:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          Set `cluster.preserve_downgrade_option`, where `$current_version` is the CockroachDB version currently running (e.g., `19.2`):

          ```sql
          > SET CLUSTER SETTING cluster.preserve_downgrade_option = '$current_version';
          ```

          Exit the shell and delete the temporary pod:

          ```sql
          > \q
          ```

          Kick off the upgrade process by changing the new Docker image, where `$new_version` is the CockroachDB version to which you are upgrading:

          ```shell
          kubectl delete job my-release-cockroachdb-init
          ```

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"image":{"tag":"<new version tag here>"}}}' --type merge
          ```

          Kubernetes will carry out a safe [rolling upgrade](https://kubernetes.io/docs/tutorials/stateful-application/basic-stateful-set/#updating-statefulsets) of your CockroachDB nodes one-by-one. Monitor the cluster's pods until all have been successfully restarted:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS              RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running             0          2m
          my-release-cockroachdb-1            1/1       Running             0          3m
          my-release-cockroachdb-2            1/1       Running             0          3m
          my-release-cockroachdb-3            0/1       ContainerCreating   0          25s
          my-release-cockroachdb-init-nwjkh   0/1       ContainerCreating   0          6s
          ```

          ```shell
          kubectl get pods \
          -o jsonpath='{range .items[*]}{.metadata.name}{"\t"}{.spec.containers[0].image}{"\n"}'
          ```

          ```
          my-release-cockroachdb-0    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-1    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-2    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-3    cockroachdb/cockroach:v20.2.4
          ```

          Resume normal operations. Once you are comfortable that the stability and performance of the cluster is what you'd expect post-upgrade, finalize the upgrade:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          ```sql
          > RESET CLUSTER SETTING cluster.preserve_downgrade_option;
          > \q
          ```

          ## Configuration

          The following table lists the configurable parameters of the CockroachDB chart and their default values.
          For details see the [`values.yaml`](values.yaml) file.

          | Parameter                                                 | Description                                                     | Default                                          |
          | ---------                                                 | -----------                                                     | -------                                          |
          | `clusterDomain`                                           | Cluster's default DNS domain                                    | `cluster.local`                                  |
          | `conf.attrs`                                              | CockroachDB node attributes                                     | `[]`                                             |
          | `conf.cache`                                              | Size of CockroachDB's in-memory cache                           | `25%`                                            |
          | `conf.cluster-name`                                       | Name of CockroachDB cluster                                     | `""`                                             |
          | `conf.disable-cluster-name-verification`                  | Disable CockroachDB cluster name verification                   | `no`                                             |
          | `conf.join`                                               | List of already-existing CockroachDB instances                  | `[]`                                             |
          | `conf.max-disk-temp-storage`                              | Max storage capacity for temp data                              | `0`                                              |
          | `conf.max-offset`                                         | Max allowed clock offset for CockroachDB cluster                | `500ms`                                          |
          | `conf.max-sql-memory`                                     | Max memory to use processing SQL querie                         | `25%`                                            |
          | `conf.locality`                                           | Locality attribute for this deployment                          | `""`                                             |
          | `conf.single-node`                                        | Disable CockroachDB clustering (standalone mode)                | `no`                                             |
          | `conf.sql-audit-dir`                                      | Directory for SQL audit log                                     | `""`                                             |
          | `conf.port`                                               | CockroachDB primary serving port in Pods                        | `26257`                                          |
          | `conf.http-port`                                          | CockroachDB HTTP port in Pods                                   | `8080`                                           |
          | `image.repository`                                        | Container image name                                            | `cockroachdb/cockroach`                          |
          | `image.tag`                                               | Container image tag                                             | `v20.2.4`                                        |
          | `image.pullPolicy`                                        | Container pull policy                                           | `IfNotPresent`                                   |
          | `image.credentials`                                       | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `statefulset.replicas`                                    | StatefulSet replicas number                                     | `3`                                              |
          | `statefulset.updateStrategy`                              | Update strategy for StatefulSet Pods                            | `{"type": "RollingUpdate"}`                      |
          | `statefulset.podManagementPolicy`                         | `OrderedReady`/`Parallel` Pods creation/deletion order          | `Parallel`                                       |
          | `statefulset.budget.maxUnavailable`                       | k8s PodDisruptionBudget parameter                               | `1`                                              |
          | `statefulset.args`                                        | Extra command-line arguments                                    | `[]`                                             |
          | `statefulset.env`                                         | Extra env vars                                                  | `[]`                                             |
          | `statefulset.secretMounts`                                | Additional Secrets to mount at cluster members                  | `[]`                                             |
          | `statefulset.labels`                                      | Additional labels of StatefulSet and its Pods                   | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `statefulset.annotations`                                 | Additional annotations of StatefulSet Pods                      | `{}`                                             |
          | `statefulset.nodeAffinity`                                | [Node affinity rules][2] of StatefulSet Pods                    | `{}`                                             |
          | `statefulset.podAffinity`                                 | [Inter-Pod affinity rules][1] of StatefulSet Pods               | `{}`                                             |
          | `statefulset.podAntiAffinity`                             | [Anti-affinity rules][1] of StatefulSet Pods                    | auto                                             |
          | `statefulset.podAntiAffinity.topologyKey`                 | The topologyKey for auto [anti-affinity rules][1]               | `kubernetes.io/hostname`                         |
          | `statefulset.podAntiAffinity.type`                        | Type of auto [anti-affinity rules][1]                           | `soft`                                           |
          | `statefulset.podAntiAffinity.weight`                      | Weight for `soft` auto [anti-affinity rules][1]                 | `100`                                            |
          | `statefulset.nodeSelector`                                | Node labels for StatefulSet Pods assignment                     | `{}`                                             |
          | `statefulset.priorityClassName`                           | [PriorityClassName][4] for StatefulSet Pods                     | `""`                                             |
          | `statefulset.tolerations`                                 | Node taints to tolerate by StatefulSet Pods                     | `[]`                                             |
          | `statefulset.topologySpreadConstraints`                   | [Topology Spread Constraints rules][5] of StatefulSet Pods      | auto                                             |
          | `statefulset.topologySpreadConstraints.maxSkew`           | Degree to which Pods may be unevenly distributed                | `1`                                              |
          | `statefulset.topologySpreadConstraints.topologyKey`       | The key of node labels                                          | `topology.kubernetes.io/zone`                    |
          | `statefulset.topologySpreadConstraints.whenUnsatisfiable` | `ScheduleAnyway`/`DoNotSchedule` for unsatisfiable constraints  | `ScheduleAnyway`                                 |
          | `statefulset.resources`                                   | Resource requests and limits for StatefulSet Pods               | `{}`                                             |
          | `service.ports.grpc.external.port`                        | CockroachDB primary serving port in Services                    | `26257`                                          |
          | `service.ports.grpc.external.name`                        | CockroachDB primary serving port name in Services               | `grpc`                                           |
          | `service.ports.grpc.internal.port`                        | CockroachDB inter-communication port in Services                | `26257`                                          |
          | `service.ports.grpc.internal.name`                        | CockroachDB inter-communication port name in Services           | `grpc-internal`                                  |
          | `service.ports.http.port`                                 | CockroachDB HTTP port in Services                               | `8080`                                           |
          | `service.ports.http.name`                                 | CockroachDB HTTP port name in Services                          | `http`                                           |
          | `service.public.type`                                     | Public Service type                                             | `ClusterIP`                                      |
          | `service.public.labels`                                   | Additional labels of public Service                             | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.public.annotations`                              | Additional annotations of public Service                        | `{}`                                             |
          | `service.discovery.labels`                                | Additional labels of discovery Service                          | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.discovery.annotations`                           | Additional annotations of discovery Service                     | `{}`                                             |
          | `ingress.enabled`                                         | Enable ingress resource for CockroachDB                         | `false`                                          |
          | `ingress.labels`                                          | Additional labels of Ingress                                    | `{}`                                             |
          | `ingress.annotations`                                     | Additional annotations of Ingress                               | `{}`                                             |
          | `ingress.paths`                                           | Paths for the default host                                      | `[/]`                                            |
          | `ingress.hosts`                                           | CockroachDB Ingress hostnames                                   | `[]`                                             |
          | `ingress.tls[0].hosts`                                    | CockroachDB Ingress tls hostnames                               | `nil`                                            |
          | `ingress.tls[0].secretName`                               | CockroachDB Ingress tls secret name                             | `nil`                                            |
          | `serviceMonitor.enabled`                                  | Create [ServiceMonitor](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/design.md#servicemonitor) Resource for scraping metrics using [PrometheusOperator](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/user-guides/getting-started.md#prometheus-operator)                     | `false`                                             |
          | `serviceMonitor.labels`                                   | Additional labels of ServiceMonitor                             | `{}`                                             |
          | `serviceMonitor.annotations`                              | Additional annotations of ServiceMonitor                        | `{}`                                             |
          | `serviceMonitor.interval`                                 | ServiceMonitor scrape metrics interval                          | `10s`                                            |
          | `serviceMonitor.scrapeTimeout`                            | ServiceMonitor scrape timeout                                   | `nil`                                            |
          | `storage.hostPath`                                        | Absolute path on host to store data                             | `""`                                             |
          | `storage.persistentVolume.enabled`                        | Whether to use PersistentVolume to store data                   | `yes`                                            |
          | `storage.persistentVolume.size`                           | PersistentVolume size                                           | `100Gi`                                          |
          | `storage.persistentVolume.storageClass`                   | PersistentVolume class                                          | `""`                                             |
          | `storage.persistentVolume.labels`                         | Additional labels of PersistentVolumeClaim                      | `{}`                                             |
          | `storage.persistentVolume.annotations`                    | Additional annotations of PersistentVolumeClaim                 | `{}`                                             |
          | `init.labels`                                             | Additional labels of init Job and its Pod                       | `{"app.kubernetes.io/component": "init"}`        |
          | `init.annotations`                                        | Additional labels of the Pod of init Job                        | `{}`                                             |
          | `init.affinity`                                           | [Affinity rules][2] of init Job Pod                             | `{}`                                             |
          | `init.nodeSelector`                                       | Node labels for init Job Pod assignment                         | `{}`                                             |
          | `init.tolerations`                                        | Node taints to tolerate by init Job Pod                         | `[]`                                             |
          | `init.resources`                                          | Resource requests and limits for the Pod of init Job            | `{}`                                             |
          | `tls.enabled`                                             | Whether to run securely using TLS certificates                  | `no`                                             |
          | `tls.serviceAccount.create`                               | Whether to create a new RBAC service account                    | `yes`                                            |
          | `tls.serviceAccount.name`                                 | Name of RBAC service account to use                             | `""`                                             |
          | `tls.certs.provided`                                      | Bring your own certs scenario, i.e certificates are provided    | `no`                                             |
          | `tls.certs.clientRootSecret`                              | If certs are provided, secret name for client root cert         | `cockroachdb-root`                               |
          | `tls.certs.nodeSecret`                                    | If certs are provided, secret name for node cert                | `cockroachdb-node`                               |
          | `tls.certs.tlsSecret`                                     | Own certs are stored in TLS secret                              | `no`                                             |
          | `tls.init.image.repository`                               | Image to use for requesting TLS certificates                    | `cockroachdb/cockroach-k8s-request-cert`         |
          | `tls.init.image.tag`                                      | Image tag to use for requesting TLS certificates                | `0.4`                                            |
          | `tls.init.image.pullPolicy`                               | Requesting TLS certificates container pull policy               | `IfNotPresent`                                   |
          | `tls.init.image.credentials`                              | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `networkPolicy.enabled`                                   | Enable NetworkPolicy for CockroachDB's Pods                     | `no`                                             |
          | `networkPolicy.ingress.grpc`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |
          | `networkPolicy.ingress.http`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |


          Override the default parameters using the `CockroachDB` custom resource.

          ## Deep dive

          ### Connecting to the CockroachDB cluster

          Once you've created the cluster, you can start talking to it by connecting to its `-public` Service. CockroachDB is PostgreSQL wire protocol compatible, so there's a [wide variety of supported clients](https://www.cockroachlabs.com/docs/install-client-drivers.html). As an example, we'll open up a SQL shell using CockroachDB's built-in shell and play around with it a bit, like this (likely needing to replace `my-release-cockroachdb-public` with the name of the `-public` Service that was created with your installed chart):

          ```shell
          kubectl run cockroach-client --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host my-release-cockroachdb-public
          ```

          ```
          Waiting for pod default/cockroach-client to be running, status is Pending,
          pod ready: false
          If you don't see a command prompt, try pressing enter.
          root@my-release-cockroachdb-public:26257> SHOW DATABASES;
          +--------------------+
          |      Database      |
          +--------------------+
          | information_schema |
          | pg_catalog         |
          | system             |
          +--------------------+
          (3 rows)
          root@my-release-cockroachdb-public:26257> CREATE DATABASE bank;
          CREATE DATABASE
          root@my-release-cockroachdb-public:26257> CREATE TABLE bank.accounts (id INT
          PRIMARY KEY, balance DECIMAL);
          CREATE TABLE
          root@my-release-cockroachdb-public:26257> INSERT INTO bank.accounts VALUES
          (1234, 10000.50);
          INSERT 1
          root@my-release-cockroachdb-public:26257> SELECT * FROM bank.accounts;
          +------+---------+
          |  id  | balance |
          +------+---------+
          | 1234 | 10000.5 |
          +------+---------+
          (1 row)
          root@my-release-cockroachdb-public:26257> \q
          Waiting for pod default/cockroach-client to terminate, status is Running
          pod "cockroach-client" deleted
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          ### Cluster health

          Because our pod spec includes regular health checks of the CockroachDB processes, simply running `kubectl get pods` and looking at the `STATUS` column is sufficient to determine the health of each instance in the cluster.

          If you want more detailed information about the cluster, the best place to look is the Admin UI.

          ### Accessing the Admin UI

          If you want to see information about how the cluster is doing, you can try pulling up the CockroachDB Admin UI by port-forwarding from your local machine to one of the pods (replacing `my-release-cockroachdb-0` with the name of one of your pods:

          ```shell
          kubectl port-forward my-release-cockroachdb-0 8080
          ```

          You should then be able to access the Admin UI by visiting <http://localhost:8080/> in your web browser.

          ### Failover

          If any CockroachDB member fails, it is restarted or recreated automatically by the Kubernetes infrastructure, and will re-join the cluster automatically when it comes back up. You can test this scenario by killing any of the CockroachDB pods:

          ```shell
          kubectl delete pod my-release-cockroachdb-1
          ```

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          After a while:

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-1  1/1       Running       0          20s
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          You can check the state of re-joining from the new pod's logs:

          ```shell
          kubectl logs my-release-cockroachdb-1
          ```

          ```
          [...]
          I161028 19:32:09.754026 1 server/node.go:586  [n1] node connected via gossip and
          verified as part of cluster {"35ecbc27-3f67-4e7d-9b8f-27c31aae17d6"}
          [...]
          cockroachdb-0.my-release-cockroachdb.default.svc.cluster.local:26257
          build:      beta-20161027-55-gd2d3c7f @ 2016/10/28 19:27:25 (go1.7.3)
          admin:      http://0.0.0.0:8080
          sql:
          postgresql://root@my-release-cockroachdb-1.my-release-cockroachdb.default.svc.cluster.local:26257?sslmode=disable
          logs:       cockroach-data/logs
          store[0]:   path=cockroach-data
          status:     restarted pre-existing node
          clusterID:  {35ecbc27-3f67-4e7d-9b8f-27c31aae17d6}
          nodeID:     2
          [...]
          ```

          ### NetworkPolicy

          To enable NetworkPolicy for CockroachDB, install [a networking plugin that implements the Kubernetes NetworkPolicy spec](https://kubernetes.io/docs/tasks/administer-cluster/declare-network-policy#before-you-begin), and set `networkPolicy.enabled` to `yes`/`true`.

          For Kubernetes v1.5 & v1.6, you must also turn on NetworkPolicy by setting the `DefaultDeny` Namespace annotation. Note: this will enforce policy for _all_ pods in the Namespace:

          ```shell
          kubectl annotate namespace default "net.beta.kubernetes.io/network-policy={\"ingress\":{\"isolation\":\"DefaultDeny\"}}"
          ```

          For more precise policy, set `networkPolicy.ingress.grpc` and `networkPolicy.ingress.http` rules. This will only allow pods that match the provided rules to connect to CockroachDB.

          ### Scaling

          Scaling should be managed via the `helm upgrade` command. After resizing your cluster on your cloud environment (e.g., GKE or EKS), run the following command to add a pod. This assumes you scaled from 3 to 4 nodes:

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"statefulset":{"replicas":"4"}}}' --type merge
          ```

          Note, that if you are running in secure mode (`tls.enabled` is `yes`/`true`) and increase the size of your cluster, you will also have to approve the CSR (certificate-signing request) of each new node (using `kubectl get csr` and `kubectl certificate approve`).

          [1]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [2]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#node-affinity
          [3]: https://cert-manager.io/
          [4]: https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass
          [5]: https://kubernetes.io/docs/concepts/workloads/pods/pod-topology-spread-constraints/
        displayName: CockroachDB Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cockroach
        - cockroachdb
        links:
        - name: Helm Chart Source
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb
        - name: Configuration Options
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration
        - name: CockroachDB Source
          url: https://github.com/cockroachdb/cockroach
        maintainers:
        - email: dmesser@redhat.com
          name: Daniel Messer
        maturity: stable
        minKubeVersion: 1.8.0
        provider:
          name: Helm Community
          url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/helmoperators/cockroachdb:v5.0.4
        version: 5.0.4
      entries:
      - name: cockroachdb.v5.0.4
        version: 5.0.4
      - name: cockroachdb.v5.0.3
        version: 5.0.3
      name: stable-5.x
    - currentCSV: cockroachdb.v6.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1alpha1",
                "kind": "Cockroachdb",
                "metadata": {
                  "name": "cockroachdb-sample"
                },
                "spec": {"image":{"repository":"cockroachdb/cockroach","tag":"v22.2.4","pullPolicy":"IfNotPresent","credentials":{}},"labels":{},"clusterDomain":"cluster.local","conf":{"attrs":[],"cache":"25%","cluster-name":"","disable-cluster-name-verification":false,"join":[],"log":{"enabled":false,"config":{}},"logtostderr":"INFO","max-sql-memory":"25%","locality":"","single-node":false,"sql-audit-dir":"","port":26257,"http-port":8080,"path":"cockroach-data","store":{"enabled":false,"type":null,"size":null,"attrs":null}},"statefulset":{"replicas":3,"updateStrategy":{"type":"RollingUpdate"},"podManagementPolicy":"Parallel","budget":{"maxUnavailable":1},"args":[],"env":[],"secretMounts":[],"labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{},"nodeAffinity":{},"podAffinity":{},"podAntiAffinity":{"topologyKey":"kubernetes.io/hostname","type":"soft","weight":100},"nodeSelector":{},"priorityClassName":"","tolerations":[],"topologySpreadConstraints":{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"},"resources":{},"customLivenessProbe":{},"customReadinessProbe":{}},"service":{"ports":{"grpc":{"external":{"port":26257,"name":"grpc"},"internal":{"port":26257,"name":"grpc-internal"}},"http":{"port":8080,"name":"http"}},"public":{"type":"ClusterIP","labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{}},"discovery":{"labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{}}},"ingress":{"enabled":false,"labels":{},"annotations":{},"paths":["/"],"hosts":[],"tls":[]},"prometheus":{"enabled":true},"serviceMonitor":{"enabled":false,"labels":{},"annotations":{},"interval":"10s","namespaced":false},"storage":{"hostPath":"","persistentVolume":{"enabled":true,"size":"100Gi","storageClass":"","labels":{},"annotations":{}}},"init":{"labels":{"app.kubernetes.io/component":"init"},"jobAnnotations":{},"annotations":{},"affinity":{},"nodeSelector":{},"tolerations":[],"resources":{},"provisioning":{"enabled":false,"clusterSettings":null,"users":[],"databases":[]}},"tls":{"enabled":true,"serviceAccount":{"create":true,"name":""},"copyCerts":{"image":"busybox"},"certs":{"provided":false,"clientRootSecret":"cockroachdb-root","nodeSecret":"cockroachdb-node","tlsSecret":false,"selfSigner":{"enabled":true,"caProvided":false,"caSecret":"","minimumCertDuration":"624h","caCertDuration":"43800h","caCertExpiryWindow":"648h","clientCertDuration":"672h","clientCertExpiryWindow":"48h","nodeCertDuration":"8760h","nodeCertExpiryWindow":"168h","rotateCerts":true,"readinessWait":"30s","podUpdateTimeout":"2m"},"certManager":false,"certManagerIssuer":{"group":"cert-manager.io","kind":"Issuer","name":"cockroachdb","clientCertDuration":"672h","clientCertExpiryWindow":"48h","nodeCertDuration":"8760h","nodeCertExpiryWindow":"168h"},"useCertManagerV1CRDs":false},"selfSigner":{"image":{"repository":"cockroachlabs-helm-charts/cockroach-self-signer-cert","tag":"1.3","pullPolicy":"IfNotPresent","credentials":{},"registry":"gcr.io"}}},"networkPolicy":{"enabled":false,"ingress":{"grpc":[],"http":[]}},"iap":{"enabled":false}}
              }
            ]
          capabilities: Basic Install
          categories: Database
          certified: "false"
          containerImage: quay.io/cockroachdb/cockroach-helm-operator:6.0.0
          createdAt: 2021-01-23T10-00-00Z
          description: CockroachDB Operator based on the CockroachDB helm chart
          olm.skipRange: <6.0.0
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/cockroachdb/helm-charts
          support: CockroachDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a CockroachDB cluster
            displayName: CockroachDB
            kind: Cockroachdb
            name: cockroachdbs.charts.operatorhub.io
            version: v1alpha1
        description: |
          CockroachDB is a scalable, survivable, strongly-consistent SQL database.
          ## About this Operator
          This Operator is based on a Helm chart for CockroachDB. It supports reconfiguration for some parameters, but notably does not handle scale down of the replica count in a seamless manner. Scale up works great.

          ## Core capabilities
          * **StatefulSet** - Sets up a dynamically scalable CockroachDB cluster using a Kubernetes StatefulSet
          * **Expand Replicas** - Supports expanding the set of replicas by simply editing your object
          * **Dashboard** - Installs the CockroachDB user interface to administer your cluster. Easily expose it via an Ingress rule.
          Review all of the [confiuguration options](https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration) to best run your database instance. The example configuration is derived from the chart's [`values.yaml`](https://github.com/helm/charts/blob/master/stable/cockroachdb/values.yaml).
          ## Using the cluster
          The resulting cluster endpoint can be consumed from a `Service` that follows the pattern: `<StatefulSet-name>-public`. For example to connect using the command line client, use something like the following to obtain the name of the service:
          ```
          kubectl get service -l chart=cockroachdb-6.0.0
          NAME                                           TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)              AGE
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb          ClusterIP   None             <none>        26257/TCP,8080/TCP   24m
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public   ClusterIP   10.106.249.134   <none>        26257/TCP,8080/TCP   24m
          ```
          Then you can use the CockroachDB command line client to connect to the database cluster:
          ```
          kubectl run -it --rm cockroach-client --image=cockroachdb/cockroach --restart=Never --command -- ./cockroach sql --insecure --host example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public
          ```
          ## Before you start
          This Operator requires a cluster with PV support in order to run correctly.

          Note that for a production cluster, you will likely want to override the following parameters in [`values.yaml`](values.yaml) with your own values.

          - `statefulset.resources.requests.memory` and `statefulset.resources.limits.memory` allocate memory resources to CockroachDB pods in your cluster.
          - `conf.cache` and `conf.max-sql-memory` are memory limits that we recommend setting to 1/4 of the above resource allocation. When running CockroachDB, you must set these limits explicitly to avoid running out of memory.
          - `storage.persistentVolume.size` defaults to `100Gi` of disk space per pod, which you may increase or decrease for your use case.
          - `storage.persistentVolume.storageClass` uses the default storage class for your environment. We strongly recommend that you specify a storage class which uses an SSD.
          - `tls.enabled` must be set to `yes`/`true` to deploy in secure mode.

          For more information on overriding the `values.yaml` parameters, please see:
          > <https://www.cockroachlabs.com/docs/stable/orchestrate-cockroachdb-with-kubernetes.html#step-2-start-cockroachdb>

          If you are running in secure mode (with configuration parameter `tls.enabled` set to `yes`/`true`) and `tls.certs.provided` set to `no`/`false`), then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          When `tls.certs.provided` is set to `yes`/`true`, this chart will use certificates created outside of Kubernetes. You may want to use this if you want to use a different certificate authority from the one being used by Kubernetes or if your Kubernetes cluster doesn't fully support certificate-signing requests. To use this, first set up your certificates and load them into your Kubernetes cluster as Secrets using the commands below:

          ```
          mkdir certs
          mkdir my-safe-directory
          cockroach cert create-ca --certs-dir=certs --ca-key=my-safe-directory/ca.key
          cockroach cert create-client root --certs-dir=certs --ca-key=my-safe-directory/ca.key
          kubectl create secret generic cockroachdb-root --from-file=certs
          cockroach cert create-node --certs-dir=certs --ca-key=my-safe-directory/ca.key localhost 127.0.0.1 eerie-horse-cockroachdb-public eerie-horse-cockroachdb-public.default eerie-horse-cockroachdb-public.default.svc.cluster.local *.eerie-horse-cockroachdb *.eerie-horse-cockroachdb.default *.eerie-horse-cockroachdb.default.svc.cluster.local
          kubectl create secret generic cockroachdb-node --from-file=certs
          ```

          Set `tls.certs.tlsSecret` to `yes/true` if you make use of [cert-manager][3] in your cluster.

          [cert-manager][3] stores generated certificates in dedicated TLS secrets. Thus, they are always named:

          * `ca.crt`
          * `tls.crt`
          * `tls.key`

          On the other hand, CockroachDB also demands dedicated certificate filenames:

          * `ca.crt`
          * `node.crt`
          * `node.key`
          * `client.root.crt`
          * `client.root.key`

          By activating `tls.certs.tlsSecret` we benefit from projected secrets and convert the TLS secret filenames to their according CockroachDB filenames.

          If you are running in secure mode, then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          Confirm that all pods are `Running` successfully and init has been completed:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS      RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running     0          1m
          my-release-cockroachdb-1            1/1       Running     0          1m
          my-release-cockroachdb-2            1/1       Running     0          1m
          my-release-cockroachdb-init-k6jcr   0/1       Completed   0          1m
          ```

          Confirm that persistent volumes are created and claimed for each pod:

          ```shell
          kubectl get pv
          ```

          ```
          NAME                                       CAPACITY   ACCESS MODES   RECLAIM POLICY   STATUS    CLAIM                                      STORAGECLASS   REASON    AGE
          pvc-64878ebf-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-0   standard                 51s
          pvc-64945b4f-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-1   standard                 51s
          pvc-649d920d-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-2   standard                 51s
          ```

          ## Upgrading the cluster

          ### Chart version 3.0.0 and after

          Launch a temporary interactive pod and start the built-in SQL client:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          Set `cluster.preserve_downgrade_option`, where `$current_version` is the CockroachDB version currently running (e.g., `19.2`):

          ```sql
          > SET CLUSTER SETTING cluster.preserve_downgrade_option = '$current_version';
          ```

          Exit the shell and delete the temporary pod:

          ```sql
          > \q
          ```

          Kick off the upgrade process by changing the new Docker image, where `$new_version` is the CockroachDB version to which you are upgrading:

          ```shell
          kubectl delete job my-release-cockroachdb-init
          ```

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"image":{"tag":"<new version tag here>"}}}' --type merge
          ```

          Kubernetes will carry out a safe [rolling upgrade](https://kubernetes.io/docs/tutorials/stateful-application/basic-stateful-set/#updating-statefulsets) of your CockroachDB nodes one-by-one. Monitor the cluster's pods until all have been successfully restarted:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS              RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running             0          2m
          my-release-cockroachdb-1            1/1       Running             0          3m
          my-release-cockroachdb-2            1/1       Running             0          3m
          my-release-cockroachdb-3            0/1       ContainerCreating   0          25s
          my-release-cockroachdb-init-nwjkh   0/1       ContainerCreating   0          6s
          ```

          ```shell
          kubectl get pods \
          -o jsonpath='{range .items[*]}{.metadata.name}{"\t"}{.spec.containers[0].image}{"\n"}'
          ```

          ```
          my-release-cockroachdb-0    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-1    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-2    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-3    cockroachdb/cockroach:v20.2.4
          ```

          Resume normal operations. Once you are comfortable that the stability and performance of the cluster is what you'd expect post-upgrade, finalize the upgrade:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          ```sql
          > RESET CLUSTER SETTING cluster.preserve_downgrade_option;
          > \q
          ```

          ## Configuration

          The following table lists the configurable parameters of the CockroachDB chart and their default values.
          For details see the [`values.yaml`](values.yaml) file.

          | Parameter                                                 | Description                                                     | Default                                          |
          | ---------                                                 | -----------                                                     | -------                                          |
          | `clusterDomain`                                           | Cluster's default DNS domain                                    | `cluster.local`                                  |
          | `conf.attrs`                                              | CockroachDB node attributes                                     | `[]`                                             |
          | `conf.cache`                                              | Size of CockroachDB's in-memory cache                           | `25%`                                            |
          | `conf.cluster-name`                                       | Name of CockroachDB cluster                                     | `""`                                             |
          | `conf.disable-cluster-name-verification`                  | Disable CockroachDB cluster name verification                   | `no`                                             |
          | `conf.join`                                               | List of already-existing CockroachDB instances                  | `[]`                                             |
          | `conf.max-disk-temp-storage`                              | Max storage capacity for temp data                              | `0`                                              |
          | `conf.max-offset`                                         | Max allowed clock offset for CockroachDB cluster                | `500ms`                                          |
          | `conf.max-sql-memory`                                     | Max memory to use processing SQL querie                         | `25%`                                            |
          | `conf.locality`                                           | Locality attribute for this deployment                          | `""`                                             |
          | `conf.single-node`                                        | Disable CockroachDB clustering (standalone mode)                | `no`                                             |
          | `conf.sql-audit-dir`                                      | Directory for SQL audit log                                     | `""`                                             |
          | `conf.port`                                               | CockroachDB primary serving port in Pods                        | `26257`                                          |
          | `conf.http-port`                                          | CockroachDB HTTP port in Pods                                   | `8080`                                           |
          | `image.repository`                                        | Container image name                                            | `cockroachdb/cockroach`                          |
          | `image.tag`                                               | Container image tag                                             | `v20.2.4`                                        |
          | `image.pullPolicy`                                        | Container pull policy                                           | `IfNotPresent`                                   |
          | `image.credentials`                                       | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `statefulset.replicas`                                    | StatefulSet replicas number                                     | `3`                                              |
          | `statefulset.updateStrategy`                              | Update strategy for StatefulSet Pods                            | `{"type": "RollingUpdate"}`                      |
          | `statefulset.podManagementPolicy`                         | `OrderedReady`/`Parallel` Pods creation/deletion order          | `Parallel`                                       |
          | `statefulset.budget.maxUnavailable`                       | k8s PodDisruptionBudget parameter                               | `1`                                              |
          | `statefulset.args`                                        | Extra command-line arguments                                    | `[]`                                             |
          | `statefulset.env`                                         | Extra env vars                                                  | `[]`                                             |
          | `statefulset.secretMounts`                                | Additional Secrets to mount at cluster members                  | `[]`                                             |
          | `statefulset.labels`                                      | Additional labels of StatefulSet and its Pods                   | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `statefulset.annotations`                                 | Additional annotations of StatefulSet Pods                      | `{}`                                             |
          | `statefulset.nodeAffinity`                                | [Node affinity rules][2] of StatefulSet Pods                    | `{}`                                             |
          | `statefulset.podAffinity`                                 | [Inter-Pod affinity rules][1] of StatefulSet Pods               | `{}`                                             |
          | `statefulset.podAntiAffinity`                             | [Anti-affinity rules][1] of StatefulSet Pods                    | auto                                             |
          | `statefulset.podAntiAffinity.topologyKey`                 | The topologyKey for auto [anti-affinity rules][1]               | `kubernetes.io/hostname`                         |
          | `statefulset.podAntiAffinity.type`                        | Type of auto [anti-affinity rules][1]                           | `soft`                                           |
          | `statefulset.podAntiAffinity.weight`                      | Weight for `soft` auto [anti-affinity rules][1]                 | `100`                                            |
          | `statefulset.nodeSelector`                                | Node labels for StatefulSet Pods assignment                     | `{}`                                             |
          | `statefulset.priorityClassName`                           | [PriorityClassName][4] for StatefulSet Pods                     | `""`                                             |
          | `statefulset.tolerations`                                 | Node taints to tolerate by StatefulSet Pods                     | `[]`                                             |
          | `statefulset.topologySpreadConstraints`                   | [Topology Spread Constraints rules][5] of StatefulSet Pods      | auto                                             |
          | `statefulset.topologySpreadConstraints.maxSkew`           | Degree to which Pods may be unevenly distributed                | `1`                                              |
          | `statefulset.topologySpreadConstraints.topologyKey`       | The key of node labels                                          | `topology.kubernetes.io/zone`                    |
          | `statefulset.topologySpreadConstraints.whenUnsatisfiable` | `ScheduleAnyway`/`DoNotSchedule` for unsatisfiable constraints  | `ScheduleAnyway`                                 |
          | `statefulset.resources`                                   | Resource requests and limits for StatefulSet Pods               | `{}`                                             |
          | `service.ports.grpc.external.port`                        | CockroachDB primary serving port in Services                    | `26257`                                          |
          | `service.ports.grpc.external.name`                        | CockroachDB primary serving port name in Services               | `grpc`                                           |
          | `service.ports.grpc.internal.port`                        | CockroachDB inter-communication port in Services                | `26257`                                          |
          | `service.ports.grpc.internal.name`                        | CockroachDB inter-communication port name in Services           | `grpc-internal`                                  |
          | `service.ports.http.port`                                 | CockroachDB HTTP port in Services                               | `8080`                                           |
          | `service.ports.http.name`                                 | CockroachDB HTTP port name in Services                          | `http`                                           |
          | `service.public.type`                                     | Public Service type                                             | `ClusterIP`                                      |
          | `service.public.labels`                                   | Additional labels of public Service                             | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.public.annotations`                              | Additional annotations of public Service                        | `{}`                                             |
          | `service.discovery.labels`                                | Additional labels of discovery Service                          | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.discovery.annotations`                           | Additional annotations of discovery Service                     | `{}`                                             |
          | `ingress.enabled`                                         | Enable ingress resource for CockroachDB                         | `false`                                          |
          | `ingress.labels`                                          | Additional labels of Ingress                                    | `{}`                                             |
          | `ingress.annotations`                                     | Additional annotations of Ingress                               | `{}`                                             |
          | `ingress.paths`                                           | Paths for the default host                                      | `[/]`                                            |
          | `ingress.hosts`                                           | CockroachDB Ingress hostnames                                   | `[]`                                             |
          | `ingress.tls[0].hosts`                                    | CockroachDB Ingress tls hostnames                               | `nil`                                            |
          | `ingress.tls[0].secretName`                               | CockroachDB Ingress tls secret name                             | `nil`                                            |
          | `serviceMonitor.enabled`                                  | Create [ServiceMonitor](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/design.md#servicemonitor) Resource for scraping metrics using [PrometheusOperator](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/user-guides/getting-started.md#prometheus-operator)                     | `false`                                             |
          | `serviceMonitor.labels`                                   | Additional labels of ServiceMonitor                             | `{}`                                             |
          | `serviceMonitor.annotations`                              | Additional annotations of ServiceMonitor                        | `{}`                                             |
          | `serviceMonitor.interval`                                 | ServiceMonitor scrape metrics interval                          | `10s`                                            |
          | `serviceMonitor.scrapeTimeout`                            | ServiceMonitor scrape timeout                                   | `nil`                                            |
          | `storage.hostPath`                                        | Absolute path on host to store data                             | `""`                                             |
          | `storage.persistentVolume.enabled`                        | Whether to use PersistentVolume to store data                   | `yes`                                            |
          | `storage.persistentVolume.size`                           | PersistentVolume size                                           | `100Gi`                                          |
          | `storage.persistentVolume.storageClass`                   | PersistentVolume class                                          | `""`                                             |
          | `storage.persistentVolume.labels`                         | Additional labels of PersistentVolumeClaim                      | `{}`                                             |
          | `storage.persistentVolume.annotations`                    | Additional annotations of PersistentVolumeClaim                 | `{}`                                             |
          | `init.labels`                                             | Additional labels of init Job and its Pod                       | `{"app.kubernetes.io/component": "init"}`        |
          | `init.annotations`                                        | Additional labels of the Pod of init Job                        | `{}`                                             |
          | `init.affinity`                                           | [Affinity rules][2] of init Job Pod                             | `{}`                                             |
          | `init.nodeSelector`                                       | Node labels for init Job Pod assignment                         | `{}`                                             |
          | `init.tolerations`                                        | Node taints to tolerate by init Job Pod                         | `[]`                                             |
          | `init.resources`                                          | Resource requests and limits for the Pod of init Job            | `{}`                                             |
          | `tls.enabled`                                             | Whether to run securely using TLS certificates                  | `no`                                             |
          | `tls.serviceAccount.create`                               | Whether to create a new RBAC service account                    | `yes`                                            |
          | `tls.serviceAccount.name`                                 | Name of RBAC service account to use                             | `""`                                             |
          | `tls.certs.provided`                                      | Bring your own certs scenario, i.e certificates are provided    | `no`                                             |
          | `tls.certs.clientRootSecret`                              | If certs are provided, secret name for client root cert         | `cockroachdb-root`                               |
          | `tls.certs.nodeSecret`                                    | If certs are provided, secret name for node cert                | `cockroachdb-node`                               |
          | `tls.certs.tlsSecret`                                     | Own certs are stored in TLS secret                              | `no`                                             |
          | `tls.init.image.repository`                               | Image to use for requesting TLS certificates                    | `cockroachdb/cockroach-k8s-request-cert`         |
          | `tls.init.image.tag`                                      | Image tag to use for requesting TLS certificates                | `0.4`                                            |
          | `tls.init.image.pullPolicy`                               | Requesting TLS certificates container pull policy               | `IfNotPresent`                                   |
          | `tls.init.image.credentials`                              | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `networkPolicy.enabled`                                   | Enable NetworkPolicy for CockroachDB's Pods                     | `no`                                             |
          | `networkPolicy.ingress.grpc`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |
          | `networkPolicy.ingress.http`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |


          Override the default parameters using the `CockroachDB` custom resource.

          ## Deep dive

          ### Connecting to the CockroachDB cluster

          Once you've created the cluster, you can start talking to it by connecting to its `-public` Service. CockroachDB is PostgreSQL wire protocol compatible, so there's a [wide variety of supported clients](https://www.cockroachlabs.com/docs/install-client-drivers.html). As an example, we'll open up a SQL shell using CockroachDB's built-in shell and play around with it a bit, like this (likely needing to replace `my-release-cockroachdb-public` with the name of the `-public` Service that was created with your installed chart):

          ```shell
          kubectl run cockroach-client --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host my-release-cockroachdb-public
          ```

          ```
          Waiting for pod default/cockroach-client to be running, status is Pending,
          pod ready: false
          If you don't see a command prompt, try pressing enter.
          root@my-release-cockroachdb-public:26257> SHOW DATABASES;
          +--------------------+
          |      Database      |
          +--------------------+
          | information_schema |
          | pg_catalog         |
          | system             |
          +--------------------+
          (3 rows)
          root@my-release-cockroachdb-public:26257> CREATE DATABASE bank;
          CREATE DATABASE
          root@my-release-cockroachdb-public:26257> CREATE TABLE bank.accounts (id INT
          PRIMARY KEY, balance DECIMAL);
          CREATE TABLE
          root@my-release-cockroachdb-public:26257> INSERT INTO bank.accounts VALUES
          (1234, 10000.50);
          INSERT 1
          root@my-release-cockroachdb-public:26257> SELECT * FROM bank.accounts;
          +------+---------+
          |  id  | balance |
          +------+---------+
          | 1234 | 10000.5 |
          +------+---------+
          (1 row)
          root@my-release-cockroachdb-public:26257> \q
          Waiting for pod default/cockroach-client to terminate, status is Running
          pod "cockroach-client" deleted
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          ### Cluster health

          Because our pod spec includes regular health checks of the CockroachDB processes, simply running `kubectl get pods` and looking at the `STATUS` column is sufficient to determine the health of each instance in the cluster.

          If you want more detailed information about the cluster, the best place to look is the Admin UI.

          ### Accessing the Admin UI

          If you want to see information about how the cluster is doing, you can try pulling up the CockroachDB Admin UI by port-forwarding from your local machine to one of the pods (replacing `my-release-cockroachdb-0` with the name of one of your pods:

          ```shell
          kubectl port-forward my-release-cockroachdb-0 8080
          ```

          You should then be able to access the Admin UI by visiting <http://localhost:8080/> in your web browser.

          ### Failover

          If any CockroachDB member fails, it is restarted or recreated automatically by the Kubernetes infrastructure, and will re-join the cluster automatically when it comes back up. You can test this scenario by killing any of the CockroachDB pods:

          ```shell
          kubectl delete pod my-release-cockroachdb-1
          ```

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          After a while:

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-1  1/1       Running       0          20s
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          You can check the state of re-joining from the new pod's logs:

          ```shell
          kubectl logs my-release-cockroachdb-1
          ```

          ```
          [...]
          I161028 19:32:09.754026 1 server/node.go:586  [n1] node connected via gossip and
          verified as part of cluster {"35ecbc27-3f67-4e7d-9b8f-27c31aae17d6"}
          [...]
          cockroachdb-0.my-release-cockroachdb.default.svc.cluster.local:26257
          build:      beta-20161027-55-gd2d3c7f @ 2016/10/28 19:27:25 (go1.7.3)
          admin:      http://0.0.0.0:8080
          sql:
          postgresql://root@my-release-cockroachdb-1.my-release-cockroachdb.default.svc.cluster.local:26257?sslmode=disable
          logs:       cockroach-data/logs
          store[0]:   path=cockroach-data
          status:     restarted pre-existing node
          clusterID:  {35ecbc27-3f67-4e7d-9b8f-27c31aae17d6}
          nodeID:     2
          [...]
          ```

          ### NetworkPolicy

          To enable NetworkPolicy for CockroachDB, install [a networking plugin that implements the Kubernetes NetworkPolicy spec](https://kubernetes.io/docs/tasks/administer-cluster/declare-network-policy#before-you-begin), and set `networkPolicy.enabled` to `yes`/`true`.

          For Kubernetes v1.5 & v1.6, you must also turn on NetworkPolicy by setting the `DefaultDeny` Namespace annotation. Note: this will enforce policy for _all_ pods in the Namespace:

          ```shell
          kubectl annotate namespace default "net.beta.kubernetes.io/network-policy={\"ingress\":{\"isolation\":\"DefaultDeny\"}}"
          ```

          For more precise policy, set `networkPolicy.ingress.grpc` and `networkPolicy.ingress.http` rules. This will only allow pods that match the provided rules to connect to CockroachDB.

          ### Scaling

          Scaling should be managed via the `helm upgrade` command. After resizing your cluster on your cloud environment (e.g., GKE or EKS), run the following command to add a pod. This assumes you scaled from 3 to 4 nodes:

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"statefulset":{"replicas":"4"}}}' --type merge
          ```

          Note, that if you are running in secure mode (`tls.enabled` is `yes`/`true`) and increase the size of your cluster, you will also have to approve the CSR (certificate-signing request) of each new node (using `kubectl get csr` and `kubectl certificate approve`).

          [1]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [2]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#node-affinity
          [3]: https://cert-manager.io/
          [4]: https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass
          [5]: https://kubernetes.io/docs/concepts/workloads/pods/pod-topology-spread-constraints/
        displayName: CockroachDB Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cockroach
        - cockroachdb
        links:
        - name: Helm Chart Source
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb
        - name: Configuration Options
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration
        - name: CockroachDB Source
          url: https://github.com/cockroachdb/cockroach
        maintainers:
        - email: yandu@cockroachlabs.com
          name: Yandu Oppacher
        maturity: stable
        minKubeVersion: 1.8.0
        provider:
          name: Helm Community
          url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/cockroachdb/cockroach-helm-operator:6.0.0
        version: 6.0.0
      entries:
      - name: cockroachdb.v6.0.0
        version: 6.0.0
      name: stable-v6.x
    defaultChannel: stable-v6.x
    packageName: cockroachdb
    provider:
      name: Helm Community
      url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CrowdStrike
      provider-url: https://crowdStrike.com
    name: falcon-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: falcon-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconContainer",
                "metadata": {
                  "name": "falcon-sidecar-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "sidecar"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconNodeSensor",
                "metadata": {
                  "name": "falcon-node-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "daemonset"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security,Monitoring
          containerImage: quay.io/crowdstrike/falcon-operator:0.8.1
          createdAt: "2024-06-06T10:00:00Z"
          description: '[DEPRECATED] Use the CrowdStrike Falcon Operator from the
            certified channel instead'
          operatorframework.io/suggested-namespace: falcon-operator
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/CrowdStrike/falcon-operator
          support: Community Only
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FalconContainer is the Schema for the falconcontainers API
            displayName: Falcon Container
            kind: FalconContainer
            name: falconcontainers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconNodeSensor is the Schema for the falconnodesensors
              API
            displayName: Falcon Node Sensor
            kind: FalconNodeSensor
            name: falconnodesensors.falcon.crowdstrike.com
            version: v1alpha1
        description: |-
          ## Deprecated
          This CrowdStrike community operator is deprecated, use the CrowdStrike Falcon Operator from the certified channel instead.

          ## About the CrowdStrike Falcon Operator
          The CrowdStrike Falcon Operator deploys CrowdStrike Falcon Workload Protection to the cluster. The operator exposes 2 custom resources that allows you to deploy either the Falcon Container Sensor or Falcon Node Sensor.

          ## About Custom Resources

          | Custom Resource                                       | Description                                                      |
          | :--------                                             | :------------                                                    |
          | [FalconContainer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/container/README.md) | Manages installation of Falcon Container Sensor on the cluster   |
          | [FalconNodeSensor](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/node/README.md)     | Manages installation of Falcon Linux Sensor on the cluster nodes |

          ## Installation and Deployment

          For installation and deployment of the CrowdStrike Falcon Operator and its Custom Resources, please read the [Installation and Deployment Guide](./docs/install_guide.md) and choose the deployment method that is right for your target environment.

          ## Getting Help
          If you encounter any issues while using the Falcon Operator, you can create an issue on our [Github repo](https://github.com/CrowdStrike/falcon-operator) for bugs, enhancements, or other requests.

          ## Contributing
          You can contribute by:

          * Raising any issues you find using Falcon Operator
          * Fixing issues by opening [Pull Requests](https://github.com/CrowdStrike/falcon-operator/pulls)
          * Improving documentation
          * Talking about the CrowdStrike Falcon Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/CrowdStrike/falcon-operator/issues).

          ## Additional Resources
           - CrowdStrike Container Security: [Product Page](https://www.crowdstrike.com/products/cloud-security/falcon-cloud-workload-protection/container-security/)
           - So You Think Your Containers Are Secure? Four Steps to Ensure a Secure Container Deployment: [Blog Post](https://www.crowdstrike.com/blog/four-steps-to-ensure-a-secure-containter-deployment/)
           - Container Security With CrowdStrike: [Blog Post](https://www.crowdstrike.com/blog/tech-center/container-security/)
           - To learn more about Falcon Container Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/support/documentation/146/falcon-container-sensor-for-linux), [Release Notes](https://falcon.crowdstrike.com/support/news/release-notes-falcon-container-sensor-for-linux)
           - To learn more about Falcon Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/documentation/20/falcon-sensor-for-linux#kubernetes-support)
           - [Developer Documentation](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/developer_guide.md)
        displayName: '[DEPRECATED] CrowdStrike Operator'
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - runtime-protection
        links:
        - name: Falcon Operator
          url: https://github.com/CrowdStrike/falcon-operator
        maintainers:
        - email: integrations@crowdstrike.com
          name: CrowdStrike Solutions Architecture
        maturity: alpha
        minKubeVersion: 1.17.0
        provider:
          name: CrowdStrike
          url: https://crowdStrike.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/crowdstrike/falcon-operator:0.8.1
        version: 0.8.2
      entries:
      - name: falcon-operator.v0.8.2
        version: 0.8.2
      - name: falcon-operator.v0.8.1
        version: 0.8.1
      name: alpha
    defaultChannel: alpha
    packageName: falcon-operator
    provider:
      name: CrowdStrike
      url: https://crowdStrike.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Foundation Model Stack
      provider-url: ""
    name: multi-nic-cni-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multi-nic-cni-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.3.0",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.3.0",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.3.0
          createdAt: "2024-04-17T04:51:40Z"
          description: Automate container multi-network configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: http://github.com/foundation-model-stack/multi-nic-cni
          support: Foundation Model Stack Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Foundation Model Stack
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.3.0
        version: 1.3.0
      entries:
      - name: multi-nic-cni-operator.v1.3.0
        version: 1.3.0
      - name: multi-nic-cni-operator.v1.2.9
        version: 1.2.9
      - name: multi-nic-cni-operator.v1.2.8
        version: 1.2.8
      - name: multi-nic-cni-operator.v1.2.2
        version: 1.2.2
      - name: multi-nic-cni-operator.v1.2.1
        version: 1.2.1
      - name: multi-nic-cni-operator.v1.0.2
        version: 1.0.2
      name: alpha
    - currentCSV: multi-nic-cni-operator.v1.2.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.7",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.7",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.7
          createdAt: "2025-05-09T12:45:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: minzhang@redhat.com
          name: Min Zhang
        - email: ebendror@redhat.com
          name: Eran Ben Dror
        maturity: alpha
        provider:
          name: Foundation Model Stack
        relatedImages:
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.7
        version: 1.2.7
      entries:
      - name: multi-nic-cni-operator.v1.2.7
        version: 1.2.7
      - name: multi-nic-cni-operator.v1.2.3
        version: 1.2.3
      - name: multi-nic-cni-operator.v1.2.0
        version: 1.2.0
      - name: multi-nic-cni-operator.v1.1.0
        version: 1.1.0
      name: beta
    - currentCSV: multi-nic-cni-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.6",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.6",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.6
          createdAt: "2025-03-24T14:49:54Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: minzhang@redhat.com
          name: Min Zhang
        - email: ebendror@redhat.com
          name: Eran Ben Dror
        maturity: alpha
        provider:
          name: Foundation Model Stack
        relatedImages:
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.6
        version: 1.2.6
      entries:
      - name: multi-nic-cni-operator.v1.2.6
        version: 1.2.6
      - name: multi-nic-cni-operator.v1.2.5
        version: 1.2.5
      - name: multi-nic-cni-operator.v1.2.4
        version: 1.2.4
      - name: multi-nic-cni-operator.v1.0.5
        version: 1.0.5
      - name: multi-nic-cni-operator.v1.0.4
        version: 1.0.4
      - name: multi-nic-cni-operator.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: multi-nic-cni-operator
    provider:
      name: Foundation Model Stack
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: assisted-service-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: assisted-service-operator.v0.7.74
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-x86_64-live.x86_64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-aarch64-live.aarch64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-s390x-live.s390x.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-x86_64-live.x86_64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-aarch64-live.aarch64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-s390x-live.s390x.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-x86_64-live.x86_64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-aarch64-live.aarch64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-ppc64le-live.ppc64le.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-s390x-live.s390x.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-x86_64-live.x86_64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-aarch64-live.aarch64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-ppc64le-live.ppc64le.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-s390x-live.s390x.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-s390x-live.s390x.iso",
                      "version": "418.94.202409162337-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service:latest
          createdAt: "2024-10-16T17:43:34Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.74'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AgentClassification is the Schema for the AgentClassifications
              API
            displayName: Agent Classification
            kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - description: Agent is the Schema for the hosts API
            displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: Infra Env
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service:latest
        version: 0.7.74
      entries:
      - name: assisted-service-operator.v0.7.74
        version: 0.7.74
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: assisted-service-operator.v0.7.71
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-x86_64-live.x86_64.iso",
                      "version": "414.92.202310170514-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-aarch64-live.aarch64.iso",
                      "version": "414.92.202310170514-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202309201615-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-s390x-live.s390x.iso",
                      "version": "414.92.202309201615-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-s390x-live.s390x.iso",
                      "version": "415.92.202310310037-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:cffbdc5fd51c08f16ec2aa484cf226e3b29510919a8aa9668d766007de8f3d79
          createdAt: "2023-12-15T15:26:08Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.71'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service@sha256:cffbdc5fd51c08f16ec2aa484cf226e3b29510919a8aa9668d766007de8f3d79
        - quay.io/edge-infrastructure/assisted-image-service@sha256:027fa9c5a881848786745aafeac5e48575eefe22417abe29d9e8eb17940b8f80
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/edge-infrastructure/assisted-installer@sha256:0f4ecf9a75b6755d02a45f9b70357ac45b6b4db66f24f0e83f888269d64f4293
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:40bafed94547a5fe96deb5492f10225501b965198227c9634869931692bda492
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:98740646fbfc0520f9d85ac5b483d1a898f22e164c5234901be51970bef7fcf3
        version: 0.7.71
      entries:
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.10
    - currentCSV: assisted-service-operator.v0.7.73
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-x86_64-live.x86_64.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-aarch64-live.aarch64.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-s390x-live.s390x.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-x86_64-live.x86_64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-aarch64-live.aarch64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-s390x-live.s390x.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.16.0-ec.4/rhcos-4.16.0-ec.4-x86_64-live.x86_64.iso",
                      "version": "416.94.202403071059-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-aarch64-live.aarch64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-s390x-live.s390x.iso",
                      "version": "415.92.202402130021-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:641fb65d8f0117cc6007567337ae38d5e912add454a76905617b60db99b60a64
          createdAt: "2024-03-20T10:08:49Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.73'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:2cc14304ba27c51d61b86db05445f3f0acc68593a4d76081da8eb9d6bae41757
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:d0da07ec81877302fc376e8049dca69c1acdbf56a37aa3e320cd75f2b85a2ed0
        - quay.io/edge-infrastructure/assisted-service@sha256:641fb65d8f0117cc6007567337ae38d5e912add454a76905617b60db99b60a64
        - quay.io/edge-infrastructure/assisted-image-service@sha256:b33068bbe88e38c0444fb158e233e5644fb127416198d52fdc809a3f2e684fde
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/edge-infrastructure/assisted-installer@sha256:234e4680b0ff4376f2c11d6f94a75a5dd6ae16c3218d6b63d2003cf43c65e185
        version: 0.7.73
      entries:
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.11
    - currentCSV: assisted-service-operator.v0.7.74
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-x86_64-live.x86_64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-aarch64-live.aarch64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-s390x-live.s390x.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-x86_64-live.x86_64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-aarch64-live.aarch64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-s390x-live.s390x.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-x86_64-live.x86_64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-aarch64-live.aarch64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-ppc64le-live.ppc64le.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-s390x-live.s390x.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-x86_64-live.x86_64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-aarch64-live.aarch64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-ppc64le-live.ppc64le.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-s390x-live.s390x.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-s390x-live.s390x.iso",
                      "version": "418.94.202409162337-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service:latest
          createdAt: "2024-10-16T17:43:34Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.74'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AgentClassification is the Schema for the AgentClassifications
              API
            displayName: Agent Classification
            kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - description: Agent is the Schema for the hosts API
            displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: Infra Env
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service:latest
        version: 0.7.74
      entries:
      - name: assisted-service-operator.v0.7.74
        version: 0.7.74
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.13
    - currentCSV: assisted-service-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "osImages": [
                    {
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202105062123-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/ocpmetal/assisted-service@sha256:64e88c5aa41c606bc2db2b0991acdbb1166559306474e13f5e2810f325616d64
          createdAt: "2021-09-23T11:11:20Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.0 <0.0.8'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled, for
              all other names a "Warning" event will be emitted + message logged and
              it will be ignored forever.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent".
          No other name will be accepted.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/ocpmetal/assisted-service@sha256:64e88c5aa41c606bc2db2b0991acdbb1166559306474e13f5e2810f325616d64
        - quay.io/ocpmetal/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/ocpmetal/assisted-installer@sha256:db97f7a565a8d5af61f3876d6465996e90c7f79dad6ec71c5bb5cce837444974
        - quay.io/ocpmetal/assisted-installer-agent@sha256:a391892766d1ce81300bf82943318e7c3be0d2945fc19065527ab9f64fc24cf8
        - quay.io/ocpmetal/assisted-installer-controller@sha256:43b65c310eee985ee0fc20ad8ee684f7557ffe704fd1a2fa53831e3dbc41e849
        version: 0.0.8
      entries:
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.3
    - currentCSV: assisted-service-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.2/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.2/rhcos-4.8.2-x86_64-live.x86_64.iso",
                      "version": "48.84.202107202156-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-x86_64-live.x86_64.iso",
                      "version": "49.84.202110081407-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:c3a852842b88f47c51cf1c62d3a19b102729fb355df19ced0e03e8265739c39c
          createdAt: "2022-04-07T11:23:06Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.1.3'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled, for
              all other names a "Warning" event will be emitted + message logged and
              it will be ignored forever.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent".
          No other name will be accepted.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-image-service@sha256:add19e251cafdfc4f34264d0aec5b660f0ea2dcc13326f4a2c8d50a2e465c0a2
        - quay.io/edge-infrastructure/assisted-service@sha256:c3a852842b88f47c51cf1c62d3a19b102729fb355df19ced0e03e8265739c39c
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/edge-infrastructure/assisted-installer@sha256:461d60b21d1d3ccb15a170f937f4871088cd0f83108b2271e84bdf2883a66870
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:979650f2e03e457d71958e8813285fcb7185fe0035f21a6ba956b4277cb29b70
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:658787d1e6c1f733c5f03863655aff315258c0b38e1972573c3878df46e4e463
        version: 0.1.3
      entries:
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.4
    - currentCSV: assisted-service-operator.v0.3.30
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-x86_64-live.x86_64.iso",
                      "version": "49.84.202110081407-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-aarch64-live-rootfs.aarch64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-aarch64-live.aarch64.iso",
                      "version": "49.84.202110080947-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.3/rhcos-4.10.3-x86_64-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.3/rhcos-4.10.3-x86_64-live.x86_64.iso",
                      "version": "410.84.202201251210-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:7c13ce52d82f761aec47bc5a6639d7baabe99d0f4f2e2106c4341018c606b458
          createdAt: "2022-06-13T23:16:28Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.3.30'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service@sha256:7c13ce52d82f761aec47bc5a6639d7baabe99d0f4f2e2106c4341018c606b458
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/edge-infrastructure/assisted-installer@sha256:57cf8d5fe49067a8c1345bedaf90f2b63aededdd7d615878496182c35a13892e
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:6271d639e4c0647bd28fbc85cea54f31143a876c3237376b9c670ec254a9c2fc
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:2267d2d7be38f81d530fed977b0c46f6c976369cd89c123c8e036c3b1dd988d5
        - quay.io/edge-infrastructure/assisted-image-service@sha256:d6396f9360b877a4f57139ba5b4969f243278db709be4de3573f2fba75e50018
        version: 0.3.30
      entries:
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.5
    - currentCSV: assisted-service-operator.v0.4.41
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.40/rhcos-4.9.40-x86_64-live.x86_64.iso",
                      "version": "49.84.202206171736-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.16/rhcos-4.10.16-x86_64-live.x86_64.iso",
                      "version": "410.84.202205191234-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.16/rhcos-4.10.16-aarch64-live.aarch64.iso",
                      "version": "410.84.202205191023-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.11.0-rc.2/rhcos-4.11.0-rc.2-x86_64-live.x86_64.iso",
                      "version": "411.86.202207090519-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.11.0-rc.2/rhcos-4.11.0-rc.2-aarch64-live.aarch64.iso",
                      "version": "411.86.202207090159-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:0fa869095185759c26c391f0e353f293c903ea739576f160d6668f5f20404bdb
          createdAt: "2022-09-11T14:56:07Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.4.41'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:56484f99799d22f345215bb65c6b81b6ee14e07d26b625c3f283a2ddfdc37238
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:951b9d625bc300b5b72a3916cb382015527ea786f10b7acbeebf83ae854ca51a
        - quay.io/edge-infrastructure/assisted-service@sha256:0fa869095185759c26c391f0e353f293c903ea739576f160d6668f5f20404bdb
        - quay.io/edge-infrastructure/assisted-image-service@sha256:104d2aae313c724fb10d3ef3bba78b2370fc0597de31335453894b679a450ab8
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:4ca00219efbc0d9dcce415dadc1039e9495f3a3b00ab3bd575b41554d7934899
        - quay.io/edge-infrastructure/assisted-installer@sha256:803168aeb7e01db77e3b2ff7418ceea3d93daccaceeb95f8fc02027144ebac75
        version: 0.4.41
      entries:
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.6
    - currentCSV: assisted-service-operator.v0.4.50
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-x86_64-live.x86_64.iso",
                      "version": "411.86.202210041459-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-aarch64-live.aarch64.iso",
                      "version": "411.86.202210032347-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.12.0-ec.2/rhcos-4.12.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "412.86.202208101039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.12.0-ec.2/rhcos-4.12.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "412.86.202208101040-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:0ca1cfe04f088cd4c7318cb91978754d6472ada29e3436d05e22335740bfff49
          createdAt: "2022-11-23T19:17:47Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.4.50'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service@sha256:0ca1cfe04f088cd4c7318cb91978754d6472ada29e3436d05e22335740bfff49
        - quay.io/edge-infrastructure/assisted-image-service@sha256:71e8c9c76caa8fe01255fdb4a1172cda372969f7fe1399a3a54fb93e967e68cf
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:54b39c6cafefb70d258ec1d491b6a0f1dd888cd85bd442c8f40a4805cf4721ca
        - quay.io/edge-infrastructure/assisted-installer@sha256:f225c86242b0bfbf58753123ef33b7c07f367a5099641d19cfda5a7121382586
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:34e0341c0471b1772db11535107db06e511c8d5d9816edc3971e745928164ecd
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:c1c7281b9f06e56330cc55ac8d74185d42ee82dff105ae747db677b8eb059aea
        version: 0.4.50
      entries:
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.7
    - currentCSV: assisted-service-operator.v0.6.65
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-x86_64-live.x86_64.iso",
                      "version": "411.86.202210041459-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-aarch64-live.aarch64.iso",
                      "version": "411.86.202210032347-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-s390x-live.s390x.iso",
                      "version": "411.86.202210171842-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202210171844-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-x86_64-live.x86_64.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-aarch64-live.aarch64.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-s390x-live.s390x.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-x86_64-live.x86_64.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-aarch64-live.aarch64.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-s390x-live.s390x.iso",
                      "version": "413.92.202303281804-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:2dfaf2c3ac58b2987b0088df302c7cf717d0b6640cdb8b2327370aa110120784
          createdAt: "2023-05-24T11:15:57Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.6.65'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-image-service@sha256:b1ba1076fe16b2f868ed31a660b16583a957a83634c7ec3cf03574b56e02144c
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:1287d947464fbf5275feb31d106d2feb856c6117f0eaec8d947f2e03bd8618c5
        - quay.io/edge-infrastructure/assisted-installer@sha256:f8a7e7cb6ff1a4a7daa0cb26f0fab8da17f109653af5979c9d53c35b7ca2c1bb
        - quay.io/edge-infrastructure/assisted-service@sha256:2dfaf2c3ac58b2987b0088df302c7cf717d0b6640cdb8b2327370aa110120784
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:1ad91b13c763437863480df68f5b65576e95c558d4537886432a81c0b8d3cedc
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:a6f99547564cc2e047df74003019817b2dbd3dad992a0c8e53aaaba4f5b02ab6
        version: 0.6.65
      entries:
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.8
    - currentCSV: assisted-service-operator.v0.7.67
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-s390x-live.s390x.iso",
                      "version": "414.92.202305090606-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:fb0397a1402d42a64e3128e66a0817aa9466e67eb407cb792dda0d380bd5a189
          createdAt: "2023-10-08T07:14:20Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.67'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service@sha256:fb0397a1402d42a64e3128e66a0817aa9466e67eb407cb792dda0d380bd5a189
        - quay.io/edge-infrastructure/assisted-image-service@sha256:69ff087be8e5d8458a97f49d88b8b2025ee341d362dcf5618988c011b66688e0
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/edge-infrastructure/assisted-installer@sha256:d85be59bc379927f46d0a88ef04172c5c6733e31e1e4a8188de2d434d5bedb66
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:f2aeb4f7d193ea2c5702cae0f6039c02991b9f417d091468d46974408c97efe4
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:35c118d8c07dd72fe3307424cc051d9e0426e916412171449a05e41445b9f599
        version: 0.7.67
      entries:
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.9
    defaultChannel: alpha
    packageName: assisted-service-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-commons-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: commons-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:6768e4189fa8deebf337e4732e5f12bb2ad4acdc112d9d05f5dc1a4ed81f08de
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for S3ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for S3BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/commons-operator@sha256:6768e4189fa8deebf337e4732e5f12bb2ad4acdc112d9d05f5dc1a4ed81f08de
        version: 24.11.1
      entries:
      - name: commons-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: commons-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:adc2b25a676f7357ca3971c13ecf5c9ed3cb1af24ffc902b03733cc814e8946f
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for S3ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for S3BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/commons-operator@sha256:adc2b25a676f7357ca3971c13ecf5c9ed3cb1af24ffc902b03733cc814e8946f
        version: 25.3.0
      entries:
      - name: commons-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: commons-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
        version: 25.7.0
      entries:
      - name: commons-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: commons-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
        version: 25.7.0
      entries:
      - name: commons-operator.v25.7.0
        version: 25.7.0
      - name: commons-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-commons-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: web-terminal
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: web-terminal.v1.11.1-0.1747215995.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: registry.redhat.io/web-terminal/web-terminal-rhel9-operator@sha256:0478d14e92df84fdfbf4584384d34dc9a71427a6487c2564d2eb7815ba1ac12b
          createdAt: "2021-10-26T07:24:32Z"
          description: Start a web terminal in your browser with common CLI tools
            for interacting with the cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.substitutesFor: web-terminal.v1.11.1
          operatorframework.io/suggested-namespace: openshift-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/redhat-developer/web-terminal-operator/
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - kind: DevWorkspaceRouting
            name: devworkspaceroutings.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha1
        description: |
          Start a web terminal in your browser with common CLI tools for interacting with
          the cluster.

          **Note:** The Web Terminal Operator integrates with the OpenShift Console in
          OpenShift 4.5.3 and higher to simplify web terminal instance creation and
          automate OpenShift login. In earlier versions of OpenShift, the operator can
          be installed but web terminals will have to be created and accessed manually.

          ## Description
          The Web Terminal Operator leverages the
          [DevWorkspace Operator](https://github.com/devfile/devworkspace-operator)
          to provision enviroments which support common cloud CLI tools. When this
          operator is installed, the DevWorkspace Operator will be installed as a
          dependency.

          ## How to Install
          Press the **Install** button, choose the upgrade strategy, and wait for the
          **Installed** Operator status.

          When the operator is installed, you will see a terminal button appear on the
          top right of the console after refreshing the OpenShift console window.

          ## How to Uninstall
          The Web Terminal Operator requires manual steps to fully uninstall the operator.
          As the Web Terminal Operator is designed as a way to access the OpenShift
          cluster, web terminal instances store user credentials. To avoid exposing these
          credentials to unwanted parties, the operator deploys webhooks and finalizers
          that aren't removed when the operator is uninstalled. See the
          [uninstall guide](https://docs.openshift.com/container-platform/latest/web_console/web_terminal/uninstalling-web-terminal.html)
          for more details.

          ## Documentation
          Documentation for this Operator is available at https://docs.openshift.com/container-platform/latest/web\_console/web\_terminal/installing-web-terminal.html
        displayName: Web Terminal
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - workspace
        - devtools
        - developer
        - ide
        - terminal
        links:
        - name: Web Terminal Repo
          url: https://github.com/redhat-developer/web-terminal-operator/
        maintainers:
        - email: aobuchow@redhat.com
          name: Andrew Obuchowicz
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/web-terminal/web-terminal-tooling-rhel9@sha256:5c24220f884dcdf1b1e5ac1e20dc6b7c8c4300bb89e8f118c0b11c331a56ab3f
        - registry.redhat.io/web-terminal/web-terminal-exec-rhel9@sha256:cfc8200340655a045f45d02fa327538f87e98d6369bdef2b46cf447053c44426
        - registry.redhat.io/web-terminal/web-terminal-rhel9-operator@sha256:0478d14e92df84fdfbf4584384d34dc9a71427a6487c2564d2eb7815ba1ac12b
        version: 1.11.1+0.1747215995.p
      entries:
      - name: web-terminal.v1.11.1-0.1747215995.p
        version: 1.11.1+0.1747215995.p
      - name: web-terminal.v1.11.1
        version: 1.11.1
      - name: web-terminal.v1.11.0
        version: 1.11.0
      - name: web-terminal.v1.10.1
        version: 1.10.1
      - name: web-terminal.v1.10.1-0.1740684238.p
        version: 1.10.1+0.1740684238.p
      - name: web-terminal.v1.10.0-0.1731481377.p
        version: 1.10.0+0.1731481377.p
      - name: web-terminal.v1.10.0-0.1732652667.p
        version: 1.10.0+0.1732652667.p
      - name: web-terminal.v1.10.0
        version: 1.10.0
      - name: web-terminal.v1.10.0-0.1727169028.p
        version: 1.10.0+0.1727169028.p
      - name: web-terminal.v1.10.0-0.1720435222.p
        version: 1.10.0+0.1720435222.p
      - name: web-terminal.v1.10.0-0.1720402943.p
        version: 1.10.0+0.1720402943.p
      - name: web-terminal.v1.9.0-0.1708477317.p
        version: 1.9.0+0.1708477317.p
      - name: web-terminal.v1.9.0
        version: 1.9.0
      - name: web-terminal.v1.8.0-0.1708477299.p
        version: 1.8.0+0.1708477299.p
      - name: web-terminal.v1.8.0-0.1701199376.p
        version: 1.8.0+0.1701199376.p
      - name: web-terminal.v1.8.0-0.1692219820.p
        version: 1.8.0+0.1692219820.p
      - name: web-terminal.v1.8.0
        version: 1.8.0
      - name: web-terminal.v1.7.0-0.1682321121.p
        version: 1.7.0+0.1682321121.p
      - name: web-terminal.v1.7.0-0.1692219820.p
        version: 1.7.0+0.1692219820.p
      - name: web-terminal.v1.7.0
        version: 1.7.0
      - name: web-terminal.v1.7.0-0.1681197295.p
        version: 1.7.0+0.1681197295.p
      - name: web-terminal.v1.7.0-0.1708477265.p
        version: 1.7.0+0.1708477265.p
      - name: web-terminal.v1.7.0-0.1684429884.p
        version: 1.7.0+0.1684429884.p
      - name: web-terminal.v1.6.0
        version: 1.6.0
      - name: web-terminal.v1.6.0-0.1692219820.p
        version: 1.6.0+0.1692219820.p
      - name: web-terminal.v1.5.1-0.1661829403.p
        version: 1.5.1+0.1661829403.p
      - name: web-terminal.v1.5.1
        version: 1.5.1
      - name: web-terminal.v1.5.0-0.1657220207.p
        version: 1.5.0+0.1657220207.p
      - name: web-terminal.v1.5.0
        version: 1.5.0
      - name: web-terminal.v1.4.0
        version: 1.4.0
      - name: web-terminal.v1.3.0
        version: 1.3.0
      name: fast
    defaultChannel: fast
    packageName: web-terminal
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM Quantum
      provider-url: ""
    name: openshift-qiskit-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-qiskit-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "qiskit.ibm.com/v1alpha1",
                "kind": "QiskitPlayground",
                "metadata": {
                  "name": "qiskitplayground-sample"
                },
                "spec": {
                  "image": "qiskit/jupyter:0.1",
                  "imagePullPolicy": "Always",
                  "resources": {
                    "requests": {
                      "memory": "2G",
                      "cpu": "1"
                    },
                    "limits": {
                      "memory": "3G",
                      "cpu": "1.5"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: docker.io/qiskit/operator:0.1
          createdAt: 2021-004-23 22:25:53
          description: |-
            Launches a developement environment Jupyter notebok development environment for
            developing quantum applications using Qiskit.
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator
          support: https://github.com/qiskit-community/openshift-quantum-operators/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QiskitPlayground is the Schema for the qiskitplaygrounds
              API
            displayName: Qiskit Playground
            kind: QiskitPlayground
            name: qiskitplaygrounds.qiskit.ibm.com
            version: v1alpha1
        description: |-
          ### Introduction
          Launches a developement Jupiter lab environment with Qiskit.

          QiskitPlayground is a Jupyter notebook with the pre installed qiskit libraries for implementing quantum application development.
          It comes with pre installed python packages for visualising results. QiskitPlayground can run quantum
          applications on simulator or on real quantum machines using IBM Q account.

          ### Documentation
          For detailed installation guide [follow](https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator)

          ### Contributing
          You can contribute by
            * Raising any [issues](https://github.com/qiskit-community/openshift-quantum-operators/issues) you find using ibm-quantum-operator community operator
            * Fixing issues by opening [Pull Requests](https://github.com/qiskit-community/openshift-quantum-operators/pulls)
            * Talking about openshift-qiskit-operator

          ### License
          kubernetes-qiskit-operator is licensed under the [Apache 2.0 license](https://github.com/qiskit-community/openshift-quantum-operators/blob/master/LICENSE)

          #### Note
          Any api could be changed any time without any notice. That said, your feedback is very important and appreciated to make this project more stable and useful.
        displayName: QiskitPlayground
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Qiskit
        - Open Source
        - Quantum
        links:
        - name: GitHub
          url: https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator
        maintainers:
        - email: blublinsky@ibm.com
          name: boris
        maturity: alpha
        provider:
          name: IBM Quantum
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - docker.io/qiskit/operator:0.1
        version: 0.2.1
      entries:
      - name: openshift-qiskit-operator.v0.2.1
        version: 0.2.1
      - name: openshift-qiskit-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: openshift-qiskit-operator
    provider:
      name: IBM Quantum
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: ""
    name: hawtio-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hawtio-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.14.0",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/hawtio/operator:0.5.0
          createdAt: "2022-04-08T00:00:00Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Hawtio Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - docker.io/hawtio/operator:0.5.0
        version: 0.5.0
      entries:
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: alpha
    - currentCSV: hawtio-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/hawtio/operator:1.2.0
          createdAt: "2024-12-02T17:22:48Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          olm.skipRange: '>=1.0.0 <1.0.2'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Hawtio console that eases the discovery and management
              of hawtio-enabled applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#jvm-agent).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/hawtio/operator:1.2.0
        version: 1.2.0
      entries:
      - name: hawtio-operator.v1.2.0
        version: 1.2.0
      - name: hawtio-operator.v1.1.1
        version: 1.1.1
      - name: hawtio-operator.v1.1.0
        version: 1.1.0
      - name: hawtio-operator.v1.0.1
        version: 1.0.1
      - name: hawtio-operator.v1.0.0
        version: 1.0.0
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: latest
    - currentCSV: hawtio-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/hawtio/operator:1.2.0
          createdAt: "2024-12-02T17:22:48Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          olm.skipRange: '>=1.0.0 <1.0.2'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Hawtio console that eases the discovery and management
              of hawtio-enabled applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#jvm-agent).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/hawtio/operator:1.2.0
        version: 1.2.0
      entries:
      - name: hawtio-operator.v1.2.0
        version: 1.2.0
      - name: hawtio-operator.v1.1.1
        version: 1.1.1
      - name: hawtio-operator.v1.1.0
        version: 1.1.0
      - name: hawtio-operator.v1.0.1
        version: 1.0.1
      - name: hawtio-operator.v1.0.0
        version: 1.0.0
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: stable-v1
    defaultChannel: stable-v1
    packageName: hawtio-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "docker.io/percona/percona-server-mongodb:7.0.18-11",
                  "imagePullPolicy": "Always",
                  "initImage": "docker.io/percona/percona-server-mongodb-operator:1.20.1",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:2.44.1",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "docker.io/percona/percona-backup-mongodb:2.9.1",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.io/percona/percona-server-mongodb-operator:1.20.1
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.20.1
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - docker.io/percona/percona-server-mongodb-operator:1.20.1
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator.v1.20.0
        version: 1.20.0
      - name: percona-server-mongodb-operator.v1.19.1
        version: 1.19.1
      - name: percona-server-mongodb-operator.v1.19.0
        version: 1.19.0
      - name: percona-server-mongodb-operator.v1.18.0
        version: 1.18.0
      - name: percona-server-mongodb-operator.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: directpv
      provider-url: ""
    name: minio-volume-manager-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-volume-manager-operator.v5.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.quay.io/v1alpha1",
                "kind": "AistorVolumemanager",
                "metadata": {
                  "name": "aistorvolumemanager-sample"
                },
                "spec": {
                  "annotations": {},
                  "controller": {
                    "podAnnotations": {},
                    "podLabels": {},
                    "replicas": 3
                  },
                  "images": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": []
                  },
                  "labels": {},
                  "license": "",
                  "nodeserver": {
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-10-05T23:18:24Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AistorVolumemanager
            name: aistorvolumemanagers.charts.quay.io
            version: v1alpha1
        description: directpv
        displayName: directpv
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - directpv
        links:
        - name: Private Dpv Operator
          url: https://private-dpv-operator.domain
        maintainers:
        - email: cesarcelis@minio.io
          name: celis
        maturity: alpha
        provider:
          name: directpv
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - quay.io/cniackz4/directpv-operator@sha256:af06e57661a7e05809e44c3fac571ed949cc04fd656fdb0c8324b533778eb604
        version: 5.0.3
      entries:
      - name: minio-volume-manager-operator.v5.0.3
        version: 5.0.3
      - name: private-dpv-operator.v5.0.0
        version: 5.0.0
      name: stable
    defaultChannel: stable
    packageName: minio-volume-manager-operator
    provider:
      name: directpv
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-commonui-operator-app
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-commonui-operator.v4.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operators.ibm.com/v1alpha1",
                "kind": "CommonWebUI",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "example-commonwebui",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator"
                  },
                  "name": "example-commonwebui"
                },
                "spec": {
                  "commonWebUIConfig": {
                    "cpuLimits": "300",
                    "cpuMemory": "256",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.8.0",
                    "ingressPath": "/common-nav",
                    "landingPage": "",
                    "requestLimits": "300",
                    "requestMemory": "256",
                    "serviceName": "common-web-ui"
                  },
                  "globalUIConfig": {
                    "cloudPakVersion": "3.6.0",
                    "defaultAdminUser": "admin",
                    "defaultAuth": "",
                    "enterpriseLDAP": "",
                    "enterpriseSAML": "",
                    "osAuth": "",
                    "sessionPollingInterval": 5000
                  },
                  "operatorVersion": "4.8.0",
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "256Mi"
                    },
                    "requests": {
                      "cpu": "300m",
                      "memory": "256Mi",
                      "ephemeral-storage": "251Mi"
                    }
                  },
                  "version": "4.8.0"
                }
              },
              {
                "apiVersion": "foundation.ibm.com/v1",
                "kind": "NavConfiguration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "common-web-ui-config",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator",
                    "default": "true",
                    "name": "common-web-ui-config"
                  },
                  "name": "common-web-ui-config"
                },
                "spec": {
                  "about": {
                    "copyright": "© 2018, 2020 IBM. All rights reserved.",
                    "licenses": [
                      "yq, version 3.3.0, MIT+GPL",
                      "MongoDB, version 4.0.16 Community Edition, SSPL",
                      "Ansible: © 2017 Red Hat, Inc., http://www.redhat.com; © Henry Graham (hzgraham) \u003cHenry.Graham@mail.wvu.edu\u003e",
                      "calico-bird: © 1998–2008, Martin Mares \u003cmj@ucw.cz\u003e; © 1998–2000 Pavel Machek \u003cpavel@ucw.cz\u003e;",
                      "© 1998–2008, Ondrej Filip \u003cfeela@network.cz\u003e; © 2009–2013,  CZ.NIC z.s.p.o.",
                      "chrony: © Richard P. Curnow  1997-2003, GPL v2",
                      "collectd, © 2017-2018, version 5.7.2, GPL v2, \u003chttps://github.com/collectd/collectd/tree/collectd-5.7.2\u003e",
                      "crudini: © Pádraig Brady \u003cP@draigBrady.com\u003e",
                      "Galera-3: © 2007–2014 Codership Oy \u003cinfo@codership.com\u003e",
                      "glusterfs: © 2010–2013+ James Shubin \u003chttps://ttboj.wordpress.com/\u003e",
                      "haproxy: © 2000–2013  Willy Tarreau \u003cw@1wt.eu\u003e",
                      "heketi v6.0.0: © 2015 The heketi Authors, GPL v2",
                      "heketi v8.0.0: © 2015 The heketi Authors, GPL v2",
                      "heketi-master/apps.app.go: © 2015 The heketi Authors",
                      "heketi-master/client/api/go-client/backup.go: © 2016 The heketi Authors",
                      "heketi-master/doc/man/heketi-cli.8: © 2016 The heketi Authors",
                      "heketi-master/extras/docker/gluster/gluster-setup.sh: © 2016 Red Hat, Inc. \u003chttp://www.redhat.com\u003e",
                      "ieee-data: © 2013 Luciano Bello \u003cluciano@debian.org\u003e",
                      "javax.mail: © 2017 Oracle and/or its affiliates. All rights reserved.",
                      "keepalived: © 2001-2017 Alexandre Cassen \u003cacassen@gmail.com\u003e",
                      "libonig2: © 2006–2008 Max Kellermann \u003cmax@duempel.org\u003e; © 2014–2015 Jörg Frings-Fürst \u003cdebian@jff-webhosting.net\u003e",
                      "libtomcrypt: © 2004 Sam Hocevar \u003csam@hocevar.net\u003e, GPL v2",
                      "mariadb-common: © 2018 MariaDB. All rights reserved.",
                      "mariaDB: © 2018 MariaDB. All rights reserved. \u003chttps://mariadb.com/\u003e",
                      "mariadb-server: © 2018 MariaDB. All rights reserved.",
                      "minitar: © 2004 Mauricio Julio Fernandez Pradier and Austin Ziegler",
                      "MongoDB: © 2007 Free Software Foundation, Inc. \u003chttp://fsf.org/\u003e",
                      "nvmi-cli: © 1989, 1991 Free Software Foundation, Inc., GPL v2",
                      "OpenJDK: © 2018 Oracle Corporation and/or its affiliates",
                      "openshift-mariadb-galera: © 2007 Free Software Foundation, Inc. \u003chttp://fsf.org/\u003e",
                      "percona-xtrabackup: © 2006–2018 Percona LLC.",
                      "pwgen: © Christian Thöing \u003cc.thoeing@web.de\u003e",
                      "rdoc: © 2001–2003 Dave Thomas, The Pragmatic Programmers",
                      "readline: © Chet Ramey \u003cchet.ramey@case.edu\u003e",
                      "John the Ripper password cracker: © 1996–2013 by Solar Designer \u003csolar@openwall.com\u003e",
                      "spdx-exceptions: © 2018 SPDX Workgroup a Linux Foundation Project. All rights reserved.",
                      "socat: © 2001–2010 Gerhard Rieger",
                      "sshpass: © 2006, 2008 Lingnu Open Source Consulting Ltd. \u003chttp://www.lingnu.com\u003e",
                      "timelimit: © 2001, 2007 - 2010  Peter Pentchev, GPL v2",
                      "ua-parser-js: © 2012-2018 Faisal Salman \u003cf@faisalman.com\u003e, GPL v2",
                      "ubuntu-cloud-keyring: © 2010 Michael Vogt \u003cmichael.vogt@canonical.com\u003e",
                      "unboundid-ldapsdk: © 2015 UnboundID. The LDAP SDK for Java is developed by UnboundID. \u003cinfo@unboundid.com\u003e",
                      "xmpp4r: © Lucas Nussbaum \u003clucas@lucas-nussbaum.net\u003e, Stephan Maka \u003cstephan@spaceboyz.net\u003e, and others.",
                      "module-assistant: © 2003-2008 Eduard Bloch \u003cblade@debian.org\u003e, version 0.11.8, GPL v2;",
                      "© 2009 Cyril Brulebois \u003ckibi@debian.org\u003e, version 0.11.8, GPL v2; © 2013-2018 Andreas Beckmann \u003canbe@debian.org\u003e, version 0.11.8, GPL v2",
                      "module-init-tools: © 2011 ProFUSION embedded systems, version 22, GPL v2",
                      "thin: © 2017 Marc-Andre Cournoyer \u003cmacournoyer@gmail.com\u003e, version 1.7.2, GPL v2",
                      "gosu, © 1999-2014, version 1.1, GPL v3",
                      "mercurial (Python), © 2006-2018 ,version v4.5.3, GPL v2",
                      "garden-runc, © 2015-Present CloudFoundry.org Foundation, Inc. All Rights Reserved, version 1.17.0, GPLv2",
                      "libtomcrypt0, © 2003-2007 Tom St Denis \u003ctomstdenis@gmail.com\u003e, version 1.17-7, GPLv2",
                      "console-setup-min, © 1999,2000,2001,2002,2003,2006,2007,2008,2009,2010,2011 Anton Zinoviev, \u003canton@lml.bas.bg\u003e,version 1.108, GPLv2",
                      "dracut, © 2009 Harald Hoyer \u003charald@redhat.com\u003e, version 044+3-3, GPLv2",
                      "dracut-core, © 2009 Harald Hoyer \u003charald@redhat.com\u003e, version 044+3-3, GPLv2",
                      "g++, version 5.4.0-6ubuntu, GPL v2",
                      "libstdc++6, version 5.4.0-6ubuntu, GPL v3",
                      "libstdc++-5-dev, version 5.4.0-6ubuntu, GPL v3",
                      "docker-engine-selinux, version 3b5fac4, GPLv2",
                      "unorm, version 1.5.0, GPL v2",
                      "psmisc, version 22.20, GPL v2",
                      "lvm2-devel, version 2.0.2, GPL v2",
                      "nfs-utils, version 1.3, GPL v2",
                      "popt-static, version 1.13, GPL v2",
                      "sysvinit-tools, version 2.88, GPL v2",
                      "stunnel, version 5.53, GPL v2",
                      "stunnel, version 5.39, GPL v2",
                      "LVM2, version 2.02.180-10.el7_6.2, GPL v2",
                      "sysdig, version 2c43237, GPL",
                      "chisels, version 9722dbc, GPL",
                      "MongoDB, version 4.0.12, SSPL",
                      "ffi (Ruby Gem), 1.11.1, GPL",
                      "inotify-tools, v3.14, GPL v2",
                      "logrotate, v3.8.6, GPL v2",
                      "checker-qual, version 2.0.0, GPLv2",
                      "ocrad-bower, v1.0.0, GPL v3",
                      "Grafana, v7.5.12, AGPL",
                      "IBM Semeru Runtime Open Edition binaries, GPL v2",
                      "checker-compat-qual 2.5.5, GPL-2.0",
                      "javax.annotation-api 1.3.1, GPL-2.0"
                    ],
                    "logoUrl": "IBM Cloud Pak | Administration Hub",
                    "logoAltText": "IBM Cloud Pak | Administration Hub"
                  },
                  "header": {
                    "disabledItems": [
                      "createResource",
                      "catalog"
                    ],
                    "docUrlMapping": "http://ibm.biz/cpcs_adminui",
                    "logoAltText": "IBM Cloud Pak Administration Hub",
                    "logoHeight": "47px",
                    "logoUrl": "/common-nav/graphics/ibm-cloudpack-logo.svg",
                    "logoWidth": "190px"
                  },
                  "login": {
                    "loginDialog": {
                      "acceptText": "Your acceptance text here",
                      "dialogText": "You must set your dialog for this environment",
                      "enable": false,
                      "headerText": "Header text here"
                    },
                    "logoAltText": "Cloud Pak",
                    "logoHeight": "47px",
                    "logoUrl": "/common-nav/api/graphics/logincloudpak.svg",
                    "logoWidth": "190px"
                  },
                  "navItems": [
                    {
                      "id": "providers",
                      "label": "Identity providers",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/realms"
                    }
                  ]
                }
              },
              {
                "apiVersion": "foundation.ibm.com/v1",
                "kind": "NavConfiguration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "icp4i",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator",
                    "name": "icp4i"
                  },
                  "name": "icp4i"
                },
                "spec": {
                  "header": {
                    "disabledItems": [
                      "createResource",
                      "catalog",
                      "bookmark"
                    ],
                    "logoAltText": "Cloud Pak for Integration",
                    "logoUrl": "/common-nav/graphics/ibm-cloudpak-integration.svg"
                  },
                  "navItems": [
                    {
                      "detectionServiceName": true,
                      "id": "metering",
                      "label": "Metering",
                      "serviceId": "metering-ui",
                      "serviceName": "metering-ui",
                      "url": "/metering/dashboard?ace_config={ 'showClusterData': false }\u0026dashboard=cpi.icp.main"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "monitoring",
                      "isAuthorized": [
                        "Administrator",
                        "ClusterAdministrator",
                        "Operator"
                      ],
                      "label": "Monitoring",
                      "serviceId": "monitoring-ui",
                      "serviceName": "ibm-monitoring-grafana",
                      "target": "_blank",
                      "url": "/grafana"
                    },
                    {
                      "id": "id-access",
                      "label": "Identity and access",
                      "serviceId": "webui-nav"
                    },
                    {
                      "id": "providers",
                      "parentId": "id-access",
                      "label": "Identity providers",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/realms?useNav=icp4i"
                    },
                    {
                      "id": "teams-ids",
                      "parentId": "id-access",
                      "label": "Teams and service IDs",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/teams?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "logging",
                      "label": "Logging",
                      "serviceId": "kibana",
                      "serviceName": "kibana",
                      "target": "_blank",
                      "url": "/kibana"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "releases",
                      "label": "Helm Releases",
                      "serviceId": "catalog-ui",
                      "serviceName": "catalog-ui",
                      "url": "/catalog/instances?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "repos",
                      "label": "Helm Repositories",
                      "serviceId": "catalog-ui",
                      "serviceName": "catalog-ui",
                      "url": "/catalog/repositories?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "licensing",
                      "label": "Licensing",
                      "serviceId": "ibm-license-service-reporter",
                      "url": "/license-service-reporter",
                      "isAuthorized": [
                        "ClusterAdministrator",
                        "CloudPakAdministrator"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operators.ibm.com/v1alpha1",
                "kind": "SwitcherItem",
                "metadata": {
                  "name": "example-switcheritem"
                },
                "spec": {
                  "cloudPakInfo": {
                    "display": "Administration",
                    "label": "administration-hub",
                    "landingPage": "/common-nav/dashboard",
                    "logoURL": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
          description: The IBM Common Web UI delivers the common header API and the
            identity and access pages for IBM Cloud Platform Common Services.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.8.0
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-commonui-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            displayName: CommonWebUI service
            kind: CommonWebUI
            name: commonwebuis.operators.ibm.com
            version: v1alpha1
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            kind: NavConfiguration
            name: navconfigurations.foundation.ibm.com
            version: v1
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            kind: SwitcherItem
            name: switcheritems.operators.ibm.com
            version: v1alpha1
        description: "**Important:** Do not install this operator directly. Only install
          this operator using the IBM Common Services Operator. For more information
          about installing this operator and other Common Services operators, see
          [Installer documentation](http://ibm.biz/cpcs_opinstall). Additionally,
          you can exit this panel and navigate to the IBM Common Services tile in
          OperatorHub to learn more about the operator.\n\n If you are using this
          operator as part of an IBM Cloud Pak, see the documentation for that IBM
          Cloud Pak to learn more about how to install and use the operator service.
          For more information about IBM Cloud Paks, see [IBM Cloud Paks that use
          Common Services](http://ibm.biz/cpcs_cloudpaks).\n\n You can use the ibm-commonui-operator
          to install the Common Web UI service for the IBM Cloud Platform Common Services
          and access the Common Web UI console. You can use the Common Web UI console
          to access information and features from other IBM Cloud Platform Common
          Services or IBM Cloud Paks that you install. \n\nFor more information about
          the available IBM Cloud Platform Common Services, see the [IBM Knowledge
          Center](http://ibm.biz/cpcsdocs). \n## Supported platforms \n\n Red Hat
          OpenShift Container Platform 4.2 or newer installed on one of the following
          platforms: \n\n- Linux x86_64 \n- Linux on Power (ppc64le) \n- Linux on
          IBM Z and LinuxONE \n## Prerequisites\n\n The Common Web UI service has
          dependencies on other IBM Cloud Platform Common Services. Before you install
          this operator, you need to first install the operator dependencies and prerequisites:
          \n For the list of operator dependencies, see the IBM Knowledge Center [Common
          Services dependencies documentation](http://ibm.biz/cpcs_opdependencies).
          \n For the list of prerequisites for installing the operator, see the IBM
          Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).
          \n## Documentation \n\n To install the operator with the IBM Common Services
          Operator follow the the installation and configuration instructions within
          the IBM Knowledge Center. \n- If you are using the operator as part of an
          IBM Cloud Pak, see the documentation for that IBM Cloud Pak, for a list
          of IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).
          \n- If you are using the operator with an IBM Containerized Software, see
          the IBM Cloud Platform Common Services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall)."
        displayName: Ibm Common UI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        - Web Console
        - Common Services
        links:
        - name: GitHub
          url: https://github.com/IBM/ibm-commonui-operator
        maintainers:
        - email: siddarth.patel@ibm.com
          name: Siddarth Patel
        - email: sgrube@us.ibm.com
          name: Steve Grube
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:e4eb6a75e0d207232e479693d9c99ba863f8e62830f3191ca4d1e7ceaf3586d1
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator-bundle@sha256:e4340bd9481da998a6b89e94c884c46e17cfb50d9d0c1099f39a3c54e500d2b2
        version: 4.8.0
      entries:
      - name: ibm-commonui-operator.v4.8.0
        version: 4.8.0
      name: v4.7
    defaultChannel: v4.7
    packageName: ibm-commonui-operator-app
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent
      provider-url: ""
    name: cilium-enterprise
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cilium.v1.10.13-xed86eae
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"prometheus":{"serviceMonitor":{"enabled":false}},"mode":"cluster-pool"},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium Enterprise - eBPF-based Networking, Security, and Observability
        displayName: Cilium Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Enterprise Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Cilium Enterprise
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-ui-enterprise@sha256:c9921c65892a5293a4f1bef787636d6ae58cec1af65da06862a47e3450afdc6e
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/isovalent/hubble-relay@sha256:9145b03acc954efde54e1ab9da878462526c5e820b83a92e32cc24291f784d33
        - quay.io/isovalent/clustermesh-apiserver@sha256:eb3ee3d2cb07e6c5966e4bc2e50615949303be5d5128e156d5c0dbe8404dc093
        - quay.io/isovalent/hubble-enterprise-metadata@sha256:72c7aa3961f14df413569cd156820a45553bb24d393b11246a18bf8b6553e022
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        - quay.io/isovalent/cilium@sha256:6887537789603946543aaa370a3ec592287a235231bf16aa386e091823285515
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:cc1f0e6b27ccda8a843dfb7fa4a8fef052ba29716646557815677c5d1bea2a1a
        - quay.io/isovalent/hubble-export-fluentd@sha256:05cb13f533f2c89b98a47cd58badda07add54077cd22283ca5028cb39f2e4486
        - quay.io/isovalent/hubble-export-s3@sha256:ce4f71f7f13301133b7d471b87fb81108b2c2da7fc29865ef96bc8431b2be72b
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/oauth2-proxy/oauth2-proxy@sha256:ecd26b74a01f2b547ddaed4d32d35f8f5e09c378d5c1fc6cfa63f0adf659ac2b
        - quay.io/isovalent/hubble-enterprise-operator@sha256:0d5e6a5d9c07780c18e5aa81d8f1c95c399bff2d36a44fa5f61ac89a788ca83b
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/hubble-ui-enterprise-backend@sha256:e892cba6dafcdf6704df3135a0b38d11894a5f276ace40683e74cb62b9c311d8
        - quay.io/isovalent/operator-generic@sha256:1cc0b0d2e71564ad8759be326c66ea13b4955b9441fb67e906da534e1b38aefa
        - quay.io/cilium/hubble-export-stdout@sha256:f1b1475a3c89950481264f5f8ca63eea3b911a45dc83fdf15122d695a0fba9b3
        - quay.io/isovalent/hubble-enterprise@sha256:1e1d3e6c199546b2398f4166ff64210c04075bb407c8f07c5be8d3db08560d4a
        - quay.io/isovalent/hubble-rbac@sha256:03f3304c157024356a6de13c3d91b4506b8e54d587920d80ebcc76cbcfeaeda6
        version: 1.10.13+xed86eae
      entries:
      - name: cilium.v1.10.13-xed86eae
        version: 1.10.13+xed86eae
      - name: cilium.v1.10.12-xb203fae
        version: 1.10.12+xb203fae
      - name: cilium.v1.10.11-x10ae342
        version: 1.10.11+x10ae342
      name: "1.10"
    - currentCSV: cilium.v1.11.21-xd0d07b1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium Enterprise - eBPF-based Networking, Security, and Observability
        displayName: Cilium Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Enterprise Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Cilium Enterprise
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - cgr.dev/chainguard/oauth2-proxy:7.4.0@sha256:54857c70321004d3e0b8b8d7208ce5422c1f71c18db5be02d2543f187ea8e8e2
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:372511f6fa3523dadfe9128caf77520cf74ce24df4a19e1583ee197375e06014
        - quay.io/isovalent/hubble-ui-enterprise:v0.23.0@sha256:02635d1392cca7e09ab7797f69147880068ba7e3b0d865be97eeab33edd646a0
        - quay.io/cilium/certgen:v0.1.5@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/startup-script:62093c5c233ea914bfa26a10ba41f8780d9b737f@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/isovalent/clustermesh-apiserver:v1.11.21-cee.1@sha256:e16d8375ed8a1c3a872047b3aacf3469bde43552f1a51ee3e159ef2e9fea3e97
        - quay.io/coreos/etcd:v3.4.13@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/isovalent/hubble-relay:v1.11.21-cee.1@sha256:5933dd1f8111d6bca0c68db9247385b5d215d5b613095bafab8244098f2c5a26
        - quay.io/isovalent/hubble-ui-enterprise-backend:v0.23.0@sha256:2fbdad86bb5244075f31487262d790f6fa75cdd993f868b952252905c118621c
        - quay.io/isovalent/hubble-rbac:v1.3.0@sha256:4c106dd133861a678eb24cf98b35d5023cccca0834c954c49b68514c366aedf0
        - quay.io/isovalent/hubble-enterprise-operator:v1.10.5@sha256:31b18a1ea7002936f04dd070549e0c9a02f8a33394666fc5b0523ab1350c6120
        - quay.io/isovalent/hubble-export-stdout:v1.0.3@sha256:bef600089c1671fce11bb4bcb98b81f25e65152a1a2f13b64ea637cfc5779c11
        - quay.io/isovalent/hubble-export-s3:v1.0.0@sha256:89c178c0a010166c68819ab5de32ab59b64a88aa23a05d8e4f78adb1a4b6c3d3
        - quay.io/isovalent/hubble-export-fluentd:v1.5.1@sha256:eaf42a84821509b69e2ef2b658796062e79a90e63045a98689da3da99a12e492
        - quay.io/isovalent/hubble-enterprise:v1.10.5@sha256:340700d4e28ded2aa9c50ff9f938996bb43faa84e0b2ba38ec79452648d7e5d7
        - quay.io/isovalent/operator-generic:v1.11.21-cee.1@sha256:25b668c1febe1d1a70b63dfe926be4e55fbf77fd186e9df7581e41d7e9813469
        - quay.io/isovalent/cilium:v1.11.21-cee.1@sha256:f41241a15cd96d26727aa1ec413ace643728ddbe42f7fb7b234dbda22bf3950b
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:d3148bf3b9765467936ee067b25b5ef3683d3e55260f517c7fe271ac0e19cb47
        version: 1.11.21+xd0d07b1
      entries:
      - name: cilium.v1.11.21-xd0d07b1
        version: 1.11.21+xd0d07b1
      - name: cilium.v1.11.20-x0f6d966
        version: 1.11.20+x0f6d966
      name: "1.11"
    - currentCSV: cilium.v1.12.21-x3861686
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.12.14 <1.12.21'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/cilium:v1.12.21-cee.1@sha256:3a462308cf12b572db987c03d190ee33691ccb04775d01beb192703a52f8d47b
        - quay.io/isovalent/hubble-relay:v1.12.21-cee.1@sha256:9d8dadc3f330acf0fa169c428d5fdf1d138d0294065eec5c628f6647106fe8b6
        - quay.io/coreos/etcd:v3.5.4@sha256:795d8660c48c439a7c3764c2330ed9222ab5db5bb524d8d0607cac76f7ba82a3
        - quay.io/isovalent/hubble-export-s3:v1.0.0@sha256:89c178c0a010166c68819ab5de32ab59b64a88aa23a05d8e4f78adb1a4b6c3d3
        - quay.io/isovalent/hubble-enterprise-operator:v1.10.5@sha256:31b18a1ea7002936f04dd070549e0c9a02f8a33394666fc5b0523ab1350c6120
        - quay.io/isovalent/clustermesh-apiserver:v1.12.21-cee.1@sha256:3f6549409966086a49e338c7829b9141a73ef2ebeebaac91db430aeb3ccd2e22
        - quay.io/isovalent/hubble-ui-enterprise:v0.28.1@sha256:11dfba0e4fcf43415aa6e8b3b6e98a5c4f468ec67b377328ad48b92037bbfbb0
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:f95c9e8002d8e381d604752b1d59b75595badbdb0b11237ac1cac4ecd04f4800
        - quay.io/isovalent/hubble-ui-enterprise-backend:v0.28.1@sha256:515aeb8166dd6995b04d4cd3333c94a740182943ba4e90a38871ee89c0b64e92
        - quay.io/isovalent/hubble-export-fluentd:v1.5.1@sha256:eaf42a84821509b69e2ef2b658796062e79a90e63045a98689da3da99a12e492
        - quay.io/cilium/certgen:v0.1.8@sha256:4a456552a5f192992a6edcec2febb1c54870d665173a33dc7d876129b199ddbd
        - quay.io/isovalent/hubble-enterprise:v1.10.5@sha256:340700d4e28ded2aa9c50ff9f938996bb43faa84e0b2ba38ec79452648d7e5d7
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/isovalent/hubble-rbac:v1.3.0@sha256:4c106dd133861a678eb24cf98b35d5023cccca0834c954c49b68514c366aedf0
        - quay.io/isovalent/operator-generic:v1.12.21-cee.1@sha256:4e19cda799a747b69f53b5e5a860ba256399d18221d5a4bbf243142ba1c69bb8
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/hubble-export-stdout:v1.0.3@sha256:bef600089c1671fce11bb4bcb98b81f25e65152a1a2f13b64ea637cfc5779c11
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:b6e40d76f0925e76f0a5b181be3f6c62728fa3e4aaef979caf9ea245f4ed03da
        - quay.io/cilium/startup-script:62093c5c233ea914bfa26a10ba41f8780d9b737f@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        version: 1.12.21+x3861686
      entries:
      - name: cilium.v1.12.21-x3861686
        version: 1.12.21+x3861686
      - name: cilium.v1.12.19-xca684e0
        version: 1.12.19+xca684e0
      - name: cilium.v1.12.18-xcb1adcf
        version: 1.12.18+xcb1adcf
      - name: cilium.v1.12.17-xae21c14
        version: 1.12.17+xae21c14
      - name: cilium.v1.12.16-x38e0b62
        version: 1.12.16+x38e0b62
      - name: cilium.v1.12.15-x8ca64a3
        version: 1.12.15+x8ca64a3
      - name: cilium.v1.12.14-xc86b260
        version: 1.12.14+xc86b260
      - name: cilium.v1.12.13-x2b5cec9
        version: 1.12.13+x2b5cec9
      - name: cilium.v1.12.12-xbd7b835
        version: 1.12.12+xbd7b835
      - name: cilium.v1.12.11-xffdea76
        version: 1.12.11+xffdea76
      - name: cilium.v1.12.10-x4a0e9e7
        version: 1.12.10+x4a0e9e7
      - name: cilium.v1.12.9-x4be66e4
        version: 1.12.9+x4be66e4
      - name: cilium.v1.12.8-x87b90a9
        version: 1.12.8+x87b90a9
      - name: cilium.v1.12.7-x53a2aab
        version: 1.12.7+x53a2aab
      name: "1.12"
    - currentCSV: cilium-enterprise.v1.13.21-x96449fb
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.7 <1.13.21+x96449fb'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/startup-script:c54c7edeab7fde4da68e59acd319ab24af242c3f@sha256:a9d0f451353f1b2f29ed3868e230ecfae10150eab9fd521de67adb3c8ca9c33a
        - quay.io/cilium/certgen:v0.1.13@sha256:01802e6a153a9473b06ebade7ee5730f8f2c6cc8db8768508161da3cdd778641
        - quay.io/isovalent/hubble-export-fluentd:v1.11.0@sha256:50cd2e4c2d04e8befcc2ae1a1463f07cba01edb9b546fb455fe7976311165385
        - quay.io/isovalent/hubble-enterprise:v1.11.6@sha256:eb60d59d1fb1b136855829274b31805c4206f8d050c3bf068e63a552576c1b38
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:b04e75456359b50b18e75464da5d7b467fdfeb058fc08ce3a10f057ba5f60107
        - quay.io/isovalent/cilium:v1.13.21-cee.2@sha256:f760cd85fa64eb8a197ec282644a8db630745a8a2854ee08aa2662cf67188db5
        - quay.io/isovalent/hubble-export-s3:v1.0.2@sha256:518daea1074d260d824c3e03a93ab6b2a7a5f4c22b8e07938770add8257f4a52
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:fc7046dee1a204163e68a6d9a42fec200b02129493a7d166f899dd12e2763364
        - quay.io/isovalent/hubble-rbac:v1.5.6@sha256:739ca45ea09829b0fa8a6198901b34eb5254df55be071bc854bbd37cae12a557
        - quay.io/isovalent/hubble-relay:v1.13.21-cee.2@sha256:aaf5f959a76d6b517e627d2d0b09dab86ae942616819a9a95f057ed4f45ea480
        - quay.io/isovalent/operator-generic:v1.13.21-cee.2@sha256:6f2d0d575e3f74395c959833992cd5fbb6971f4e292df1b7cf45be8b07be5e7a
        - quay.io/isovalent/clustermesh-apiserver:v1.13.21-cee.2@sha256:980b0235fd16817ecc4457ad4317a4601993b76df2e81b07a53f6f437c2705af
        - quay.io/isovalent/hubble-export-stdout:v1.0.4@sha256:a0ff77efa890478f18f5deeba2cd1e2dfd6c5c77981adac57b14d7734de71ff1
        - quay.io/isovalent/hubble-ui-enterprise:v1.0.3@sha256:55bf787c96561f4a460e2bfe5966421e2d85f3f9a38ac2487c6f6486c97cd5af
        - quay.io/isovalent/hubble-ui-enterprise-backend:v1.0.3@sha256:6756d2b808c38efce4765f90641d26aa2e0e79329ecc378a123cd3dc2c439640
        - quay.io/isovalent/hubble-enterprise-operator:v1.11.6@sha256:38253c2e17a8b2e88a226f0976c9cf4a00445212e2886e6f07cbb3a52757324f
        - quay.io/coreos/etcd:v3.5.4@sha256:795d8660c48c439a7c3764c2330ed9222ab5db5bb524d8d0607cac76f7ba82a3
        version: 1.13.21+x96449fb
      entries:
      - name: cilium-enterprise.v1.13.21-x96449fb
        version: 1.13.21+x96449fb
      - name: cilium-enterprise.v1.13.20-x5b70650
        version: 1.13.20+x5b70650
      - name: cilium-enterprise.v1.13.19-xdb27142
        version: 1.13.19+xdb27142
      - name: cilium-enterprise.v1.13.18-x0c21a94
        version: 1.13.18+x0c21a94
      - name: cilium-enterprise.v1.13.17-xc8f905f
        version: 1.13.17+xc8f905f
      - name: cilium-enterprise.v1.13.16-xcfe62ed
        version: 1.13.16+xcfe62ed
      - name: cilium-enterprise.v1.13.15-x1dfd26f
        version: 1.13.15+x1dfd26f
      - name: cilium.v1.13.14-xb70933f
        version: 1.13.14+xb70933f
      - name: cilium.v1.13.12-x453f6ad
        version: 1.13.12+x453f6ad
      - name: cilium.v1.13.11-xf6d6935
        version: 1.13.11+xf6d6935
      - name: cilium.v1.13.10-x9d57fc7
        version: 1.13.10+x9d57fc7
      - name: cilium.v1.13.9-xe8a359b
        version: 1.13.9+xe8a359b
      - name: cilium.v1.13.8-x96e1ca9
        version: 1.13.8+x96e1ca9
      - name: cilium.v1.13.7-x7320e7d
        version: 1.13.7+x7320e7d
      - name: cilium.v1.13.6-xd3714ef
        version: 1.13.6+xd3714ef
      - name: cilium.v1.13.5-x8386cae
        version: 1.13.5+x8386cae
      - name: cilium.v1.13.4-xc8b22ea
        version: 1.13.4+xc8b22ea
      - name: cilium.v1.13.3-x69e9561
        version: 1.13.3+x69e9561
      - name: cilium.v1.13.2-x4b1c6c8
        version: 1.13.2+x4b1c6c8
      name: "1.13"
    - currentCSV: cilium-enterprise.v1.14.22-x6d86c88
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.14.22+x6d86c88'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-enterprise:v1.11.6@sha256:eb60d59d1fb1b136855829274b31805c4206f8d050c3bf068e63a552576c1b38
        - quay.io/isovalent/hubble-rbac:v1.5.2@sha256:29309837e2246a1b6849d7a09d01ceebfdb43fa0ade973d14584f0e52b856b95
        - quay.io/coreos/etcd:v3.5.4@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:148b1029eac14049e2ba84d4d75be1ca43ef2d89951adbdeddec392ad7c9fe1e
        - quay.io/isovalent/hubble-ui-enterprise:v1.2.2@sha256:3e03a58f6eea5d2df195f42ee2817f7b28fa20bbe140f5086937eee8da1b8d56
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/isovalent/hubble-enterprise-operator:v1.11.6@sha256:38253c2e17a8b2e88a226f0976c9cf4a00445212e2886e6f07cbb3a52757324f
        - quay.io/cilium/startup-script:c54c7edeab7fde4da68e59acd319ab24af242c3f@sha256:a9d0f451353f1b2f29ed3868e230ecfae10150eab9fd521de67adb3c8ca9c33a
        - quay.io/isovalent/operator-generic:v1.14.22-cee.1@sha256:a997fe24ff0b7f7df3ee12748413ae53d6e07c803162b0db4d812b916b0cfc74
        - quay.io/isovalent/hubble-export-fluentd:v1.8.1@sha256:ce0cce63a8473d5ae73872fc65d82c54f2875511bb03f0bc8ee8480148c43ee1
        - quay.io/cilium/certgen:v0.1.17@sha256:07959f3125dae659ab9036c14df046e2536f7a4f6c92dabda837388b0344dc74
        - quay.io/isovalent/hubble-relay:v1.14.22-cee.1@sha256:d7ffc880d2e42e0876417ec26115b0e9155ad9a2e7f47c02c434746afa42b29a
        - quay.io/isovalent/clustermesh-apiserver:v1.14.22-cee.1@sha256:588e2b173ab409288eda2d404571184562d3d9fd12e2d9afe7282a5a7d79113a
        - quay.io/cilium/cilium-envoy:v1.32.5-1743506100-0821ef0acdf9f824d47d34e02932be522b3e7233@sha256:305c8b7da56ef1b64e4deb0642acf93a3944f2712e40c99a40bb005e6dc9761b
        - quay.io/isovalent/hubble-ui-enterprise-backend:v1.2.2@sha256:4258f7e0bff24ff1e0b214f62b37097032b6426ba28edd3dfc4f451a34de0110
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:fc7046dee1a204163e68a6d9a42fec200b02129493a7d166f899dd12e2763364
        - quay.io/isovalent/hubble-export-stdout:v1.0.4@sha256:a0ff77efa890478f18f5deeba2cd1e2dfd6c5c77981adac57b14d7734de71ff1
        - quay.io/isovalent/cilium:v1.14.22-cee.1@sha256:0e664a5a9e0edba38ab1a25eb7fa50861bb15f69ed0bbe3f3c792bd909240da3
        - quay.io/isovalent/hubble-export-s3:v1.0.2@sha256:518daea1074d260d824c3e03a93ab6b2a7a5f4c22b8e07938770add8257f4a52
        version: 1.14.22+x6d86c88
      entries:
      - name: cilium-enterprise.v1.14.22-x6d86c88
        version: 1.14.22+x6d86c88
      - name: cilium-enterprise.v1.14.21-xe0c9553
        version: 1.14.21+xe0c9553
      - name: cilium-enterprise.v1.14.20-xe01e743
        version: 1.14.20+xe01e743
      - name: cilium-enterprise.v1.14.19-xc1b9aff
        version: 1.14.19+xc1b9aff
      name: "1.14"
    - currentCSV: cilium-enterprise.v1.15.21-x2661a01
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.15.21+x2661a01'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:bd1a4d47ef48a6eb84c49d979e8f5ce867af5932828dfd3233fa9f78c2eac1f4
        - quay.io/isovalent/cilium-envoy-ubi:v1.33.4-1752151664-7c2edb0b44cf95f326d628b837fcdd845102ba68@sha256:e81fdde48c3d5cd93111e1d51d9f8554adb32ee937601a9124354e0894933b99
        - quay.io/isovalent/operator-generic-ubi:v1.15.21-cee.1@sha256:97daac94735e497e1e897beaa01eb50866abace1aa9a3919fc57971e45f538fe
        - quay.io/isovalent/hubble-ui-frontend-ubi:v1.3.7@sha256:f2b4e935abe3eb94f82dab41c9d3d929fca71d82fa39a6c8002683375b5303a4
        - quay.io/isovalent/oauth2-proxy:7.12.0-20250819@sha256:fe8f76728c36536d498b2a5c11172cb18dc30bbb555df2a6d14c036b29cb9f10
        - quay.io/isovalent/cilium:v1.15.21-cee.1@sha256:39ef22d467783481eccb376fcb5d6bc254ed8f4b3b25032121498db40278f0b8
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.15.21-cee.1@sha256:36d00eb9f8ed45a3d8892942cfae097782dd6509889ef7d1a451f9b0334d17e1
        - quay.io/isovalent/certgen-ubi:v0.1.19@sha256:c32e96b0972c9bcd3896945c46ea23e330c9dea407afee3a637daae637296f4c
        - quay.io/isovalent/hubble-ui-backend-ubi:v1.3.7@sha256:a4b5e7a032c1bbcd969a6873bc12935f8b6aa4c2d3a2ed87f5494fcfcdfd02d4
        - quay.io/isovalent/hubble-relay-ubi:v1.15.21-cee.1@sha256:8461f720ca95a9c543a298bee1c1e56241f56d5426030c13afed8262f2f5c5ff
        - quay.io/isovalent/startup-script-ubi:1755531540-60ee83e@sha256:2e1d5048cbfcd3c77f732da8203a6c792cbf19897b5cf137b9ad23a6a267f6b2
        version: 1.15.21+x2661a01
      entries:
      - name: cilium-enterprise.v1.15.21-x2661a01
        version: 1.15.21+x2661a01
      - name: cilium-enterprise.v1.15.20-x6df37bb
        version: 1.15.20+x6df37bb
      - name: cilium-enterprise.v1.15.19-x0be49ab
        version: 1.15.19+x0be49ab
      - name: cilium-enterprise.v1.15.18-xeeaa70b
        version: 1.15.18+xeeaa70b
      - name: cilium-enterprise.v1.15.17-xdfd9482
        version: 1.15.17+xdfd9482
      - name: cilium-enterprise.v1.15.16-xc7e31ca
        version: 1.15.16+xc7e31ca
      - name: cilium-enterprise.v1.15.15-xf37f22c
        version: 1.15.15+xf37f22c
      - name: cilium-enterprise.v1.15.14-x0e512e3
        version: 1.15.14+x0e512e3
      - name: cilium-enterprise.v1.15.13-x134fa29
        version: 1.15.13+x134fa29
      name: "1.15"
    - currentCSV: cilium-enterprise.v1.16.15-x6419e26
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.16.0 <1.16.15+x6419e26'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/startup-script-ubi:v1.16.15-cee.1@sha256:5d59a46be44c4d899576e69e0f20e148245e17d905a0769a73fd01f2432d1bab
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:ec8a17956e75e91b60e2999d10e67ef76f9b4cfd754db415aaf387d21690da87
        - quay.io/isovalent/cilium-envoy-ubi:v1.16.15-cee.1@sha256:843e6fdf6845f3d5f841dbd2dce75224b16c9d34c2b0ce7f329d9666466bfc6b
        - quay.io/isovalent/cilium-ubi:v1.16.15-cee.1@sha256:1d7bbbcb613ef20cd1719bd77473e71c198960ee5d8e563aec9c65cfd3c2efe2
        - quay.io/isovalent/hubble-relay-ubi:v1.16.15-cee.1@sha256:c992d549627e8f9180c272e7d96ff3783de184796633ed5d0b6fe2cf7e74055d
        - quay.io/isovalent/operator-generic-ubi:v1.16.15-cee.1@sha256:970aa156b6f5c390b6835a626190e18656f02db6a5cf27aaa1ac9c7872f4b15a
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.16.15-cee.1@sha256:f7573ab4145d306051b5d6fa7ff1af395fc71935f5204839ab1c9f574aee17ad
        - quay.io/isovalent/certgen-ubi:v1.16.15-cee.1@sha256:e3f1fa5ee59887d3a15c43947a853c8a3f4fc650ed72bcd891b277037ea16a9d
        version: 1.16.15+x6419e26
      entries:
      - name: cilium-enterprise.v1.16.15-x6419e26
        version: 1.16.15+x6419e26
      - name: cilium-enterprise.v1.16.13-xb2de075
        version: 1.16.13+xb2de075
      - name: cilium-enterprise.v1.16.12-xd373871
        version: 1.16.12+xd373871
      - name: cilium-enterprise.v1.16.11-xbd1cd23
        version: 1.16.11+xbd1cd23
      - name: cilium-enterprise.v1.16.10-x89ab4b0
        version: 1.16.10+x89ab4b0
      - name: cilium-enterprise.v1.16.9-x8ec4137
        version: 1.16.9+x8ec4137
      - name: cilium-enterprise.v1.16.8-yd2b50b4
        version: 1.16.8+yd2b50b4
      - name: cilium-enterprise.v1.16.7-x4867421
        version: 1.16.7+x4867421
      - name: cilium-enterprise.v1.16.6-x7e6eb1c
        version: 1.16.6+x7e6eb1c
      name: "1.16"
    defaultChannel: "1.16"
    packageName: cilium-enterprise
    provider:
      name: Isovalent
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: heat-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: heat-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "heat.openstack.org/v1beta1",
                "kind": "Heat",
                "metadata": {
                  "name": "heat"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = True\n",
                  "databaseInstance": "openstack",
                  "heatAPI": {},
                  "heatCfnAPI": {},
                  "heatEngine": {},
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "heat.openstack.org/v1beta1",
                "kind": "HeatAPI",
                "metadata": {
                  "name": "heat-api"
                },
                "spec": {
                  "databaseAccount": "heat",
                  "databaseHostname": "openstack",
                  "passwordSelectors": {
                    "authEncryptionKey": "HeatAuthEncryptionKey",
                    "service": "HeatPassword"
                  },
                  "replicas": 1,
                  "resources": {},
                  "secret": "osp-secret",
                  "serviceAccount": "heat-heat",
                  "serviceUser": "heat",
                  "transportURLSecret": "rabbitmq-transport-url-heat-heat-transport"
                }
              },
              {
                "apiVersion": "heat.openstack.org/v1beta1",
                "kind": "HeatCfnAPI",
                "metadata": {
                  "name": "heat-cfnapi"
                },
                "spec": {
                  "databaseAccount": "heat",
                  "databaseHostname": "openstack",
                  "passwordSelectors": {
                    "authEncryptionKey": "HeatAuthEncryptionKey",
                    "service": "HeatPassword"
                  },
                  "replicas": 1,
                  "resources": {},
                  "secret": "osp-secret",
                  "serviceAccount": "heat-heat",
                  "serviceUser": "heat",
                  "transportURLSecret": "rabbitmq-transport-url-heat-heat-transport"
                }
              },
              {
                "apiVersion": "heat.openstack.org/v1beta1",
                "kind": "HeatEngine",
                "metadata": {
                  "name": "heat-engine"
                },
                "spec": {
                  "databaseAccount": "heat",
                  "databaseHostname": "openstack",
                  "passwordSelectors": {
                    "authEncryptionKey": "HeatAuthEncryptionKey",
                    "service": "HeatPassword"
                  },
                  "replicas": 1,
                  "resources": {},
                  "secret": "osp-secret",
                  "serviceAccount": "heat-heat",
                  "serviceUser": "heat",
                  "transportURLSecret": "rabbitmq-transport-url-heat-heat-transport"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:16Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HeatAPI is the Schema for the heatapis API
            displayName: Heat API
            kind: HeatAPI
            name: heatapis.heat.openstack.org
            version: v1beta1
          - description: HeatCfnAPI is the Schema for the heatcfnapis API
            displayName: Heat Cfn API
            kind: HeatCfnAPI
            name: heatcfnapis.heat.openstack.org
            version: v1beta1
          - description: HeatEngine is the Schema for the heatengines API
            displayName: Heat Engine
            kind: HeatEngine
            name: heatengines.heat.openstack.org
            version: v1beta1
          - description: Heat is the Schema for the heats API
            displayName: Heat
            kind: Heat
            name: heats.heat.openstack.org
            version: v1beta1
        description: Heat Operator
        displayName: Heat Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Orchestration
        - Heat
        links:
        - name: Heat Operator
          url: https://github.com/openstack-k8s-operators/heat-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/heat-rhel9-operator@sha256:fa61aeee8d63087d150a7641d1e8d4b67dd067bcfed6a140401fe396f1e32403
        - registry.redhat.io/rhoso/openstack-heat-api-rhel9@sha256:dba236e5fed89fbe46397b6ab01de5795e66e9194f5cd99fdf137c37a7b96135
        - registry.redhat.io/rhoso/openstack-heat-api-cfn-rhel9@sha256:95548184e68325a5fe1604d5c4d35d1e9cca7f39b84303fce50479e0e35f8280
        - registry.redhat.io/rhoso/openstack-heat-engine-rhel9@sha256:0b6fc782986bf11b26c0839fd81c52257f4a31aa5e5a6728ce251f3670537aba
        version: 1.0.6
      entries:
      - name: heat-operator.v1.0.6
        version: 1.0.6
      - name: heat-operator.v1.0.5
        version: 1.0.5
      - name: heat-operator.v1.0.4
        version: 1.0.4
      - name: heat-operator.v1.0.3
        version: 1.0.3
      - name: heat-operator.v1.0.2
        version: 1.0.2
      - name: heat-operator.v1.0.1
        version: 1.0.1
      - name: heat-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: heat-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: servicemeshoperator3
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: servicemeshoperator3.v3.0.0-tp.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh-tech-preview/istio-rhel9-operator@sha256:f5dfe495ba92ecb880485b66cfafbf6451f8142b8e0d5fd7f29db94171c51336
          createdAt: "2024-12-12T10:28:38Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1alpha1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1alpha1
          - description: IstioRevisionTag references a Istio or IstioRevision object
              and serves as an alias for sidecar injection.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1alpha1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-rhel9-operator@sha256:f5dfe495ba92ecb880485b66cfafbf6451f8142b8e0d5fd7f29db94171c51336
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-pilot-rhel9@sha256:376f64eab0ebeb06035983a60d1e7647c5d41354f31a49cb34f4deb7b6708e3e
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-pilot-rhel9@sha256:ea9372fabe06cd8e65ea7b4567ce3c8fbddd9d673dc34a12743fd95c548059d1
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-proxyv2-rhel9@sha256:c21481d6dae58b82dac93b1b47b9d9f53a30617fa9d930a0b5d075620cffa74b
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-must-gather-rhel9@sha256:2bf3704ba8d5945973ccf6bd13418457e929836352acf0ea6a47976ffa246616
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-proxyv2-rhel9@sha256:cbd6bc47341ab8618609f89d063552e17232541c727429855c01deab73e4d06c
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-must-gather-rhel9@sha256:e3978aecaafe2678355e21a7303f8620f53d8b774f6592f10527c15bae957224
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-cni-rhel9@sha256:d0ab36246c556249b6d67346eb0c72ebd22483a6502caa0fc75112a8d327b053
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-cni-rhel9@sha256:f16cd8a29ddcbb3de77cbfcf7e1da78e42cc0c1c502132d477a01c52689cdf23
        version: 3.0.0-tp.2
      entries:
      - name: servicemeshoperator3.v3.0.0-tp.2
        version: 3.0.0-tp.2
      - name: servicemeshoperator3.v3.0.0-tp.1
        version: 3.0.0-tp.1
      name: candidates
    - currentCSV: servicemeshoperator3.v3.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.4"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.4"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
          createdAt: "2025-09-10T06:22:42Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:366266f10e658c4cea86bddf6ee847e1908ea4dd780cb5f7fb0e466bac8995f1
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:0d6930075b8906bee0ab1aeac82f6e2431c16163cf87f7f0c68c10b2c40552a1
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:26747627ae22bbdffcf9de58077454fc0c890cda83659d7834b15dea2b5aaaf2
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2ea112ab90b8540f11e9949d77c3e7e3b3ef57ac3bf23f6cf1e883a88430e1f9
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:028e10651db0d1ddb769a27c9483c6d41be6ac597f253afd9d599f395d9c82d8
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:ecc6a22a471f4f4b836648f3a72ea7f57a4ea960ebcbc7dbe3fe729500902d0b
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d518f3d1539f45e1253c5c9fa22062802804601d4998cd50344e476a3cc388fe
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:33feac0c9c888b375e207e86b3f9a1a289126ca46b64632a7f693e2191cfbbfd
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:2ca96716812d339775d3dfd396c863834dd3810720e98bfcf87bc67c5fbd29b5
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:6bb44c84363a4113c5bce2ae0d6528b345fddd12bc559bbfaacf6222060ddcb0
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:6881fb9f5a4b447be4c468b8585fff86171bd3d82671a48559f3ceabcb9dad24
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:14c5a52faf20267baa43d9a72ee6416f56fbc41dd640adc2aba3c4043802a0e9
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:b757b81d34711cb8abfcb0e8896c023d39d4778f237dd80b637ba4bf613df893
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:f7ab868455b5e887e2d89b7078678531290365f9166cae99ce56119f31fa2ba4
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d757e700a124d7583b9456ec9e08e5e7e249c1de202bc4e973102922b9f69191
        version: 3.1.2
      entries:
      - name: servicemeshoperator3.v3.1.2
        version: 3.1.2
      - name: servicemeshoperator3.v3.1.1
        version: 3.1.1
      - name: servicemeshoperator3.v3.1.0
        version: 3.1.0
      - name: servicemeshoperator3.v3.0.5
        version: 3.0.5
      - name: servicemeshoperator3.v3.0.4
        version: 3.0.4
      - name: servicemeshoperator3.v3.0.3
        version: 3.0.3
      - name: servicemeshoperator3.v3.0.2
        version: 3.0.2
      - name: servicemeshoperator3.v3.0.1
        version: 3.0.1
      - name: servicemeshoperator3.v3.0.0
        version: 3.0.0
      name: stable
    - currentCSV: servicemeshoperator3.v3.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.6"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.6"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:d716f3558fbdf8514f225de708135ba87fdcac4c718ef101f394daf546514afb
          createdAt: "2025-09-03T07:53:34Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:d716f3558fbdf8514f225de708135ba87fdcac4c718ef101f394daf546514afb
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        version: 3.0.5
      entries:
      - name: servicemeshoperator3.v3.0.5
        version: 3.0.5
      - name: servicemeshoperator3.v3.0.4
        version: 3.0.4
      - name: servicemeshoperator3.v3.0.3
        version: 3.0.3
      - name: servicemeshoperator3.v3.0.2
        version: 3.0.2
      - name: servicemeshoperator3.v3.0.1
        version: 3.0.1
      - name: servicemeshoperator3.v3.0.0
        version: 3.0.0
      name: stable-3.0
    - currentCSV: servicemeshoperator3.v3.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.4"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.4"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
          createdAt: "2025-09-10T06:22:42Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:028e10651db0d1ddb769a27c9483c6d41be6ac597f253afd9d599f395d9c82d8
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d518f3d1539f45e1253c5c9fa22062802804601d4998cd50344e476a3cc388fe
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:366266f10e658c4cea86bddf6ee847e1908ea4dd780cb5f7fb0e466bac8995f1
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:33feac0c9c888b375e207e86b3f9a1a289126ca46b64632a7f693e2191cfbbfd
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:ecc6a22a471f4f4b836648f3a72ea7f57a4ea960ebcbc7dbe3fe729500902d0b
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:b757b81d34711cb8abfcb0e8896c023d39d4778f237dd80b637ba4bf613df893
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:14c5a52faf20267baa43d9a72ee6416f56fbc41dd640adc2aba3c4043802a0e9
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:6bb44c84363a4113c5bce2ae0d6528b345fddd12bc559bbfaacf6222060ddcb0
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:0d6930075b8906bee0ab1aeac82f6e2431c16163cf87f7f0c68c10b2c40552a1
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:6881fb9f5a4b447be4c468b8585fff86171bd3d82671a48559f3ceabcb9dad24
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d757e700a124d7583b9456ec9e08e5e7e249c1de202bc4e973102922b9f69191
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2ea112ab90b8540f11e9949d77c3e7e3b3ef57ac3bf23f6cf1e883a88430e1f9
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:26747627ae22bbdffcf9de58077454fc0c890cda83659d7834b15dea2b5aaaf2
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:f7ab868455b5e887e2d89b7078678531290365f9166cae99ce56119f31fa2ba4
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:2ca96716812d339775d3dfd396c863834dd3810720e98bfcf87bc67c5fbd29b5
        version: 3.1.2
      entries:
      - name: servicemeshoperator3.v3.1.2
        version: 3.1.2
      - name: servicemeshoperator3.v3.1.1
        version: 3.1.1
      - name: servicemeshoperator3.v3.1.0
        version: 3.1.0
      name: stable-3.1
    defaultChannel: stable
    packageName: servicemeshoperator3
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: must-gather-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: must-gather-operator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "MustGather",
                "metadata": {
                  "name": "example-mustgather"
                },
                "spec": {
                  "caseID": "02527285",
                  "caseManagementAccountSecretRef": {
                    "name": "case-management-creds"
                  },
                  "serviceAccountRef": {
                    "name": "must-gather-admin"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security, Monitoring
          certified: "false"
          containerImage: quay.io/redhat-cop/must-gather-operator@sha256:aebef122cdb8409f5a5d70d3cdc585e02272704d57e103553a84bc6fc238e646
          createdAt: "2021-12-02T17:13:53Z"
          description: This operator provides a facility to easily upload must-gather
            reports to a Red Hat case.
          operatorframework.io/suggested-namespace: global-load-balancer-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/must-gather-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: MustGather
            name: mustgathers.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The must gather operator helps collecting must gather information on a cluster and uploading it to a case.
          To use the operator a cluster administrator can create the following must gather CR:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: MustGather
          metadata:
            name: example-mustgather
          spec:
            caseID: '02527285'
            caseManagementAccountSecretRef:
              name: case-management-creds
          ```

          this request will collect the standard must gather info and upload it to case `#02527285` using the credentials found in the `caseManagementCreds` secret.

          A more complex example:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: MustGather
          metadata:
            name: full-mustgather
          spec:
            caseID: '02527285'
            caseManagementAccountSecretRef:
              name: case-management-creds
            serviceAccountRef:
              name: must-gather-admin
            mustGatherImages:
            - quay.io/kubevirt/must-gather:latest
            - quay.io/ocs-dev/ocs-must-gather
          ```

          in this example we are using a specific service account (which must have cluster admin permissions as per must-gather requirements) and we are specifying a couple of additional must gather images to be run for the `kubevirt` and `ocs` subsystem. If not specified serviceAccountRef.Name will default to `default`. Also the standard must gather image: `quay.io/openshift/origin-must-gather:latest` is always added by default.
        displayName: Must Gather Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - case_management
        - support
        - must_gather
        links:
        - name: Must Gather Operator
          url: https://github.com/redhat-cop/must-gather-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/must-gather-operator
        maintainers:
        - email: rspazzol@redhat.com'
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/must-gather-operator@sha256:aebef122cdb8409f5a5d70d3cdc585e02272704d57e103553a84bc6fc238e646
        version: 1.1.2
      entries:
      - name: must-gather-operator.v1.1.2
        version: 1.1.2
      - name: must-gather-operator.v1.1.1
        version: 1.1.1
      - name: must-gather-operator.v1.1.0
        version: 1.1.0
      - name: must-gather-operator.v1.0.6
        version: 1.0.6
      - name: must-gather-operator.v1.0.5
        version: 1.0.5
      - name: must-gather-operator.v1.0.4
        version: 1.0.4
      - name: must-gather-operator.v1.0.3
        version: 1.0.3
      - name: must-gather-operator.v1.0.2
        version: 1.0.2
      - name: must-gather-operator.v1.0.1
        version: 1.0.1
      - name: must-gather-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: must-gather-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CrowdStrike
      provider-url: https://crowdStrike.com
    name: falcon-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: falcon-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconAdmission",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-admission",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconadmission",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-kac"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "admission_controller"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconContainer",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-sidecar-sensor",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconcontainer",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-container-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "sidecar"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconDeployment",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcondeployment-sample",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falcon-deployment",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-deployment"
                },
                "spec": {
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconImageAnalyzer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "falcon-operator",
                    "app.kubernetes.io/instance": "falconimageanalyzer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "falconimageanalyzer",
                    "app.kubernetes.io/part-of": "falcon-operator"
                  },
                  "name": "falcon-image-analyzer"
                },
                "spec": {
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconNodeSensor",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-node-sensor",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconnodesensor",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-node-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "daemonset"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security,Monitoring
          containerImage: registry.connect.redhat.com/crowdstrike/falcon-operator@sha256:ffe00e9e22e209d1e08f29c3d257bbd51937fcbde68a1ed93b542a26dd436dee
          createdAt: "2025-06-03T22:34:20Z"
          description: Falcon Operator installs CrowdStrike Falcon Sensors on the
            cluster
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: falcon-operator
          operators.openshift.io/valid-subscription: Not required to use this operator
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: registry.connect.redhat.com/crowdstrike/falcon-operator
          support: CrowdStrike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FalconAdmission is the Schema for the falconadmissions API
            displayName: Falcon Admission
            kind: FalconAdmission
            name: falconadmissions.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconContainer is the Schema for the falconcontainers API
            displayName: Falcon Container
            kind: FalconContainer
            name: falconcontainers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconDeployment is the Schema for the falcondeployments
              API
            displayName: Falcon Deployment
            kind: FalconDeployment
            name: falcondeployments.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconImageAnalyzer is the Schema for the falconImageAnalyzers
              API
            displayName: Falcon Image Analyzer
            kind: FalconImageAnalyzer
            name: falconimageanalyzers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconNodeSensor is the Schema for the falconnodesensors
              API
            displayName: Falcon Node Sensor
            kind: FalconNodeSensor
            name: falconnodesensors.falcon.crowdstrike.com
            version: v1alpha1
        description: |-
          The CrowdStrike Falcon Operator installs CrowdStrike Falcon custom resources on a Kubernetes cluster.

          ## About the CrowdStrike Falcon Operator
          The CrowdStrike Falcon Operator deploys CrowdStrike Falcon to the cluster. The operator exposes custom resources that allow you to protect your Kubernetes clusters when deployed.

          ## About Custom Resources

          | Custom Resource                                       | Description                                                      |
          | :--                                             | :--                                                    |
          | [FalconAdmission](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/admission/README.md) | Manages installation of Falcon Admission Controller on the cluster |
          | [FalconImageAnalyzer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/imageanalyzer/README.md) | Manages installation of Falcon Image Assessment at Runtime on the cluster |
          | [FalconContainer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/container/README.md) | Manages installation of Falcon Container Sensor on the cluster   |
          | [FalconNodeSensor](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/node/README.md)     | Manages installation of Falcon Linux Sensor on the cluster nodes |
          | [FalconDeployment](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/falcondeployment/README.md)       | Deploys FalconAdmission, FalconImageAnalyzer, FalconContainer, and FalconNodeSensor CRs from a single manifest |

          ## Installation and Deployment

          For installation and deployment of the CrowdStrike Falcon Operator and its Custom Resources, please read the [Installation and Deployment Guide](./docs/install_guide.md) and choose the deployment method that is right for your target environment.

          ## Getting Help
          If you encounter any issues while using the Falcon Operator, you can create an issue on our [Github repo](https://github.com/CrowdStrike/falcon-operator) for bugs, enhancements, or other requests.

          ## Contributing
          You can contribute by:

          * Raising any issues you find using Falcon Operator
          * Fixing issues by opening [Pull Requests](https://github.com/CrowdStrike/falcon-operator/pulls)
          * Improving documentation
          * Talking about the CrowdStrike Falcon Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/CrowdStrike/falcon-operator/issues).

          ## Additional Resources
           - CrowdStrike Container Security: [Product Page](https://www.crowdstrike.com/products/cloud-security/falcon-cloud-workload-protection/container-security/)
           - So You Think Your Containers Are Secure? Four Steps to Ensure a Secure Container Deployment: [Blog Post](https://www.crowdstrike.com/blog/four-steps-to-ensure-a-secure-containter-deployment/)
           - Container Security With CrowdStrike: [Blog Post](https://www.crowdstrike.com/blog/tech-center/container-security/)
           - To learn more about Falcon Container Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/support/documentation/146/falcon-container-sensor-for-linux), [Release Notes](https://falcon.crowdstrike.com/support/news/release-notes-falcon-container-sensor-for-linux)
           - To learn more about Falcon Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/documentation/20/falcon-sensor-for-linux#kubernetes-support)
           - [Developer Documentation](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/developer_guide.md)
        displayName: CrowdStrike Falcon Platform - Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - runtime-protection
        links:
        - name: Falcon Operator
          url: https://github.com/CrowdStrike/falcon-operator
        maintainers:
        - email: integrations@crowdstrike.com
          name: CrowdStrike Solutions Architecture
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: CrowdStrike
          url: https://crowdStrike.com
        relatedImages:
        - registry.connect.redhat.com/crowdstrike/falcon-operator@sha256:ffe00e9e22e209d1e08f29c3d257bbd51937fcbde68a1ed93b542a26dd436dee
        - registry.crowdstrike.com/falcon-container/us-1/release/falcon-sensor@sha256:92d222cd604c9f2c9e9f2c9b402b9fb0858fcb04dac38c15e25125978f43c99b
        - registry.crowdstrike.com/falcon-kac/us-1/release/falcon-kac@sha256:5049128dcfe9565f39fc6e28fb94f5bf05fb7c51cf787fc6fe5c63a2b04f800d
        - registry.crowdstrike.com/falcon-imageanalyzer/us-1/release/falcon-imageanalyzer@sha256:f08603648658041ad52ccf0cfd058da40b385904830030aa82dc50609aed7ceb
        - registry.crowdstrike.com/falcon-sensor/us-1/release/falcon-sensor@sha256:ed4f3fe8d9f5f491600f18d4440dc1d8e3a07d4b9c08a1f17382a23a3a5cf523
        version: 1.8.0
      entries:
      - name: falcon-operator.v1.8.0
        version: 1.8.0
      - name: falcon-operator.v1.7.0
        version: 1.7.0
      - name: falcon-operator.v1.6.0
        version: 1.6.0
      - name: falcon-operator.v1.5.0
        version: 1.5.0
      - name: falcon-operator.v1.4.0
        version: 1.4.0
      - name: falcon-operator.v1.3.1
        version: 1.3.1
      - name: falcon-operator.v1.3.0
        version: 1.3.0
      - name: falcon-operator.v1.2.0
        version: 1.2.0
      - name: falcon-operator.v1.1.0
        version: 1.1.0
      - name: falcon-operator.v1.0.1
        version: 1.0.1
      - name: falcon-operator.v1.0.0
        version: 1.0.0
      name: certified-1.0
    defaultChannel: certified-1.0
    packageName: falcon-operator
    provider:
      name: CrowdStrike
      url: https://crowdStrike.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Airlock
      provider-url: ""
    name: airlock-microgateway
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airlock-microgateway.v4.5.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControl","metadata":{"name":"access-control-example"},"spec":{"policies":[{"requestConditions":{"path":{"matcher":{"exact":"/secure"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}},{"oidc":{"claim":{"name":"roles","value":{"matcher":{"contains":"\"admin\""}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"oidc-example"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"identity-propagation-example"}}],"onFailure":"Pass"}},{"authorization":{}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurity","metadata":{"name":"content-security-example"},"spec":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1"},"spec":{"value":{"name":"ext-authz","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}}}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"lifetime":"16h"},"prefix":"session_prefix_example"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SidecarGateway","metadata":{"name":"sidecar-gateway-example"},"spec":{"podSelector":{"matchLabels":{"app":"example-app"}},"sessionHandlingRef":{"name":"session-handling-example"},"applications":[{"containerPort":8443,"routes":[{"pathPrefix":"/metrics","unsecured":{}},{"pathPrefix":"/","secured":{"accessControlRef":{"name":"access-control-example"},"contentSecurityRef":{"name":"content-security-example"}}}],"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}},"restrictions":{"http":{"headersLength":"80Ki"}},"protocol":{"auto":{}},"tls":{"enable":true,"secretRef":{"name":"example-tls-secret"},"protocol":{"minimum":"TLSv1_3"},"clientCertificate":{"ignored":{}},"xfcc":"AlwaysForwardOnly"}},"upstream":{"tls":{"enable":true,"protocol":{"minimum":"TLSv1_2"}}},"envoyHTTPFilterRefs":{"prepend":[{"name":"envoy-http-filter-example-1"}]},"telemetryRef":{"name":"telemetry-example"}}],"envoyClusterRefs":[{"name":"envoy-cluster-example-1"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-06-13T17:33:31Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: AccessControl
            name: accesscontrols.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurity
            name: contentsecurities.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: SidecarGateway
            name: sidecargateways.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056)
          * Gateway API
          * (Optional) Airlock Microgateway CNI (Required for [data plane mode sidecar](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000137)).

          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ### (Optional) Install Airlock Microgateway CNI
          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          Refer to [our documentation](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000140) for installation instructions

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD_MONITOR\_CREATE | "false" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.5/
        - name: Airlock Microgateway Labs
          url: https://play.instruqt.com/airlock/invite/hyi9fy4b4jzc?icp_referrer=openshift
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:806366a0b3f78844fe14775fc290fbfbde8d2a701d34a6478b019a0e7140b699
        version: 4.5.4
      entries:
      - name: airlock-microgateway.v4.5.4
        version: 4.5.4
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.5"
    - currentCSV: airlock-microgateway.v4.6.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControl","metadata":{"name":"access-control-example"},"spec":{"policies":[{"requestConditions":{"path":{"matcher":{"exact":"/secure"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}},{"oidc":{"claim":{"name":"roles","value":{"matcher":{"contains":"\"admin\""}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"oidc-example"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"token-exchange-example"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"identity-propagation-example"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"jwt-example"}}}}},{"authorization":{}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurity","metadata":{"name":"content-security-example"},"spec":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1"},"spec":{"value":{"name":"ext-authz","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}}}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"lifetime":"16h"},"prefix":"session_prefix_example"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SidecarGateway","metadata":{"name":"sidecar-gateway-example"},"spec":{"podSelector":{"matchLabels":{"app":"example-app"}},"sessionHandlingRef":{"name":"session-handling-example"},"applications":[{"containerPort":8443,"routes":[{"pathPrefix":"/metrics","unsecured":{}},{"pathPrefix":"/","secured":{"accessControlRef":{"name":"access-control-example"},"contentSecurityRef":{"name":"content-security-example"}}}],"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}},"restrictions":{"http":{"headersLength":"80Ki"}},"protocol":{"auto":{}},"tls":{"enable":true,"secretRef":{"name":"example-tls-secret"},"protocol":{"minimum":"TLSv1_3"},"clientCertificate":{"ignored":{}},"xfcc":"AlwaysForwardOnly"}},"upstream":{"tls":{"enable":true,"protocol":{"minimum":"TLSv1_2"}}},"envoyHTTPFilterRefs":{"prepend":[{"name":"envoy-http-filter-example-1"}]},"telemetryRef":{"name":"telemetry-example"}}],"envoyClusterRefs":[{"name":"envoy-cluster-example-1"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-23T06:27:21Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: AccessControl
            name: accesscontrols.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurity
            name: contentsecurities.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: SidecarGateway
            name: sidecargateways.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056)
          * Gateway API
          * (Optional) Airlock Microgateway CNI (Required for [data plane mode sidecar](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000137)).

          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.3.0/standard-install.yaml
          ```

          ### (Optional) Install Airlock Microgateway CNI
          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          Refer to [our documentation](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000140) for installation instructions

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD_MONITOR\_CREATE | "false" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.6/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:8fa300ced5b69acbad824c037d2b344fd1926bcfb39e264e9f0f055ceb9bd68d
        version: 4.6.4
      entries:
      - name: airlock-microgateway.v4.6.4
        version: 4.6.4
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.6"
    - currentCSV: airlock-microgateway.v4.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1","namespace":"default"},"spec":{"value":{"name":"envoycluster/default/envoy-cluster-example-1","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyExtensionPolicy","metadata":{"name":"envoy-extension-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"envoyHTTPFilterRefs":{"prepend":[{"name":"geoip-blocker"}],"append":[{"name":"lua-transform-request"}]},"envoyClusterRefs":[{"name":"geoip-server"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"volumes":[{"name":"vol-1","emptyDir":{}}],"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}},"volumeMounts":[{"name":"vol-1","mountPath":"/tmp/data"}]}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"telemetryRef":{"name":"telemetry-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"mode":"Enforce","persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"idle":"30s","lifetime":"16h"},"prefix":"session_prefix_example","cookie":{"name":"SESS","attributes":{"domain":"example.org","path":"/","sameSite":"Strict","secure":true}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"},"session":{"id":"%SESSION_ID%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-03T13:02:23Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyExtensionPolicy
            name: envoyextensionpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ***Note:*** Sidecar-based data plane mode is *not* supported by this OLM Bundle.
          If you want to use the Sidecar-based data plane mode, use Helm instead. Please note that the Airlock Microgateway CNI and the Helm charts are not certified for RedHat OpenShift.
          Refer to https://github.com/airlock/microgateway for installation instruction with Helm.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056)
          * [Gateway API](https://gateway-api.sigs.k8s.io/)

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ***Note:*** *This step is only required for OpenShift versions below 4.19, as Gateway API is included by default in OpenShift 4.19+.*

          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD\_MONITOR\_CREATE | "true" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gateway API
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.7/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:d4ce35184922953c209a9ebb234ca3370dbfc5054de411b5503c26ac2730b7ae
        version: 4.7.1
      entries:
      - name: airlock-microgateway.v4.7.1
        version: 4.7.1
      - name: airlock-microgateway.v4.7.0
        version: 4.7.0
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.7"
    - currentCSV: airlock-microgateway.v4.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1","namespace":"default"},"spec":{"value":{"name":"envoycluster/default/envoy-cluster-example-1","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyExtensionPolicy","metadata":{"name":"envoy-extension-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"envoyHTTPFilterRefs":{"prepend":[{"name":"geoip-blocker"}],"append":[{"name":"lua-transform-request"}]},"envoyClusterRefs":[{"name":"geoip-server"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"volumes":[{"name":"vol-1","emptyDir":{}}],"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}},"volumeMounts":[{"name":"vol-1","mountPath":"/tmp/data"}]}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"telemetryRef":{"name":"telemetry-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"mode":"Enforce","persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"idle":"30s","lifetime":"16h"},"prefix":"session_prefix_example","cookie":{"name":"SESS","attributes":{"domain":"example.org","path":"/","sameSite":"Strict","secure":true}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"},"session":{"id":"%SESSION_ID%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-03T13:02:23Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyExtensionPolicy
            name: envoyextensionpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ***Note:*** Sidecar-based data plane mode is *not* supported by this OLM Bundle.
          If you want to use the Sidecar-based data plane mode, use Helm instead. Please note that the Airlock Microgateway CNI and the Helm charts are not certified for RedHat OpenShift.
          Refer to https://github.com/airlock/microgateway for installation instruction with Helm.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056)
          * [Gateway API](https://gateway-api.sigs.k8s.io/)

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ***Note:*** *This step is only required for OpenShift versions below 4.19, as Gateway API is included by default in OpenShift 4.19+.*

          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD\_MONITOR\_CREATE | "true" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gateway API
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.7/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:d4ce35184922953c209a9ebb234ca3370dbfc5054de411b5503c26ac2730b7ae
        version: 4.7.1
      entries:
      - name: airlock-microgateway.v4.7.1
        version: 4.7.1
      - name: airlock-microgateway.v4.7.0
        version: 4.7.0
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: stable
    defaultChannel: "4.7"
    packageName: airlock-microgateway
    provider:
      name: Airlock
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: marin3r
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marin3r.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyConfig",
                "metadata": {
                  "name": "envoyconfig-example",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "nodeID": "example",
                  "resources": [
                    {
                      "type": "cluster",
                      "value": {
                        "connect_timeout": "0.01s",
                        "dns_lookup_family": "V4_ONLY",
                        "lb_policy": "ROUND_ROBIN",
                        "load_assignment": {
                          "cluster_name": "example",
                          "endpoints": [
                            {
                              "lb_endpoints": [
                                {
                                  "endpoint": {
                                    "address": {
                                      "socket_address": {
                                        "address": "example",
                                        "port_value": 8080
                                      }
                                    }
                                  }
                                }
                              ]
                            }
                          ]
                        },
                        "name": "example",
                        "type": "STRICT_DNS"
                      }
                    },
                    {
                      "type": "route",
                      "value": {
                        "name": "local",
                        "virtual_hosts": [
                          {
                            "domains": [
                              "*"
                            ],
                            "name": "all",
                            "routes": [
                              {
                                "match": {
                                  "prefix": "/"
                                },
                                "route": {
                                  "cluster": "example"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    {
                      "type": "listener",
                      "value": {
                        "address": {
                          "socket_address": {
                            "address": "0.0.0.0",
                            "port_value": 8443
                          }
                        },
                        "filter_chains": [
                          {
                            "filters": [
                              {
                                "name": "envoy.filters.network.http_connection_manager",
                                "typed_config": {
                                  "@type": "type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager",
                                  "access_log": [
                                    {
                                      "name": "envoy.access_loggers.file",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.access_loggers.file.v3.FileAccessLog",
                                        "path": "/dev/stdout"
                                      }
                                    }
                                  ],
                                  "http_filters": [
                                    {
                                      "name": "envoy.filters.http.router",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.filters.http.router.v3.Router"
                                      }
                                    }
                                  ],
                                  "rds": {
                                    "config_source": {
                                      "ads": {},
                                      "resource_api_version": "V3"
                                    },
                                    "route_config_name": "local"
                                  },
                                  "stat_prefix": "https"
                                }
                              }
                            ],
                            "transport_socket": {
                              "name": "envoy.transport_sockets.tls",
                              "typed_config": {
                                "@type": "type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext",
                                "common_tls_context": {
                                  "tls_certificate_sds_secret_configs": [
                                    {
                                      "name": "example.default.svc",
                                      "sds_config": {
                                        "ads": {},
                                        "resource_api_version": "V3"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          }
                        ],
                        "name": "https"
                      }
                    },
                    {
                      "blueprint": "tlsCertificate",
                      "generateFromTlsSecret": "example.default.svc",
                      "type": "secret"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "DiscoveryService",
                "metadata": {
                  "name": "discoveryservice-example",
                  "namespace": "my-namespace"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyDeployment",
                "metadata": {
                  "name": "envoydeployment-sample",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "discoveryServiceRef": "discoveryservice-example",
                  "envoyConfigRef": "envoyconfig-example",
                  "ports": [
                    {
                      "name": "https",
                      "port": 8443
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: quay.io/3scale/marin3r
          createdAt: "2024-08-05T16:08:25Z"
          description: Lighweight, CRD based Envoy control plane for Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["envoyconfigrevisions.marin3r.3scale.net","discoveryservicecertificates.operator.marin3r.3scale.net"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/3scale-ops/marin3r
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DiscoveryServiceCertificate is an internal resource used
              to create certificates. This resource is used by the DiscoveryService
              controller to create the required certificates for the different components.
              Direct use of DiscoveryServiceCertificate objects is discouraged.
            displayName: DiscoveryServiceCertificate
            kind: DiscoveryServiceCertificate
            name: discoveryservicecertificates.operator.marin3r.3scale.net
            version: v1alpha1
          - description: DiscoveryService represents an envoy discovery service server.
              Only one instance per namespace is currently supported.
            displayName: DiscoveryService
            kind: DiscoveryService
            name: discoveryservices.operator.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfigRevision is an internal resource that stores a
              specific version of an EnvoyConfig resource. EnvoyConfigRevisions are
              automatically created and deleted by the EnvoyConfig controller and
              are not intended to be directly used. Use EnvoyConfig objects instead.
            displayName: EnvoyConfigRevision
            kind: EnvoyConfigRevision
            name: envoyconfigrevisions.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfig holds the configuration for a given envoy nodeID.
              The spec of an EnvoyConfig object holds the Envoy resources that conform
              the desired configuration for the given nodeID and that the discovery
              service will send to any envoy client that identifies itself with that
              nodeID.
            displayName: EnvoyConfig
            kind: EnvoyConfig
            name: envoyconfigs.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyDeployment is a resource to deploy and manage a Kubernetes
              Deployment of Envoy Pods.
            displayName: EnvoyDeployment
            kind: EnvoyDeployment
            name: envoydeployments.operator.marin3r.3scale.net
            version: v1alpha1
        description: |
          MARIN3R implements a control plane to deploy, configure and operate a fleet of envoy instances within a Kubernetes cluster.
          It operates a group of gateways and Pod sidecar containers to provide general purpose functionality required to operate production systems.

          Features:

          * Deploy and manage an Envoy xDS server using the DiscoveryService custom resource.
          * Inject Envoy sidecar containers based on Pod annotations.
          * Deploy Envoy as a Kubernetes Deployment using the EnvoyDeployment custom resource.
          * Dynamic Envoy configuration and re-configuration using the EnvoyConfig custom resource.
          * Use Kubernetes Secrets as certificate sources.
          * Syntactic validation of Envoy configurations.
          * Lifecycle management of Envoy containers (graceful shutdown and connection draining).
          * Self-healing.

          Check the [docs](https://github.com/3scale-ops/marin3r#readme) for more information.

          ## License
          MARIN3R is licensed under the [Apache 2.0 license](https://github.com/3scale/prometheus-exporter-operator/blob/master/LICENSE)
        displayName: MARIN3R
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - envoy
        - discovery
        - controlplane
        - proxy
        - sidecar
        - xds
        - network
        - gateway
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/marin3r
        maintainers:
        - email: 3scale-operations+marin3r@redhat.com
          name: 3scale Operations Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/marin3r:v0.13.1
        version: 0.13.1
      entries:
      - name: marin3r.v0.13.1
        version: 0.13.1
      - name: marin3r.v0.12.3
        version: 0.12.3
      - name: marin3r.v0.11.1
        version: 0.11.1
      - name: marin3r.v0.10.0
        version: 0.10.0
      name: alpha
    - currentCSV: marin3r.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyConfig",
                "metadata": {
                  "name": "envoyconfig-example",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "nodeID": "example",
                  "resources": [
                    {
                      "type": "cluster",
                      "value": {
                        "connect_timeout": "0.01s",
                        "dns_lookup_family": "V4_ONLY",
                        "lb_policy": "ROUND_ROBIN",
                        "load_assignment": {
                          "cluster_name": "example",
                          "endpoints": [
                            {
                              "lb_endpoints": [
                                {
                                  "endpoint": {
                                    "address": {
                                      "socket_address": {
                                        "address": "example",
                                        "port_value": 8080
                                      }
                                    }
                                  }
                                }
                              ]
                            }
                          ]
                        },
                        "name": "example",
                        "type": "STRICT_DNS"
                      }
                    },
                    {
                      "type": "route",
                      "value": {
                        "name": "local",
                        "virtual_hosts": [
                          {
                            "domains": [
                              "*"
                            ],
                            "name": "all",
                            "routes": [
                              {
                                "match": {
                                  "prefix": "/"
                                },
                                "route": {
                                  "cluster": "example"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    {
                      "type": "listener",
                      "value": {
                        "address": {
                          "socket_address": {
                            "address": "0.0.0.0",
                            "port_value": 8443
                          }
                        },
                        "filter_chains": [
                          {
                            "filters": [
                              {
                                "name": "envoy.filters.network.http_connection_manager",
                                "typed_config": {
                                  "@type": "type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager",
                                  "access_log": [
                                    {
                                      "name": "envoy.access_loggers.file",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.access_loggers.file.v3.FileAccessLog",
                                        "path": "/dev/stdout"
                                      }
                                    }
                                  ],
                                  "http_filters": [
                                    {
                                      "name": "envoy.filters.http.router",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.filters.http.router.v3.Router"
                                      }
                                    }
                                  ],
                                  "rds": {
                                    "config_source": {
                                      "ads": {},
                                      "resource_api_version": "V3"
                                    },
                                    "route_config_name": "local"
                                  },
                                  "stat_prefix": "https"
                                }
                              }
                            ],
                            "transport_socket": {
                              "name": "envoy.transport_sockets.tls",
                              "typed_config": {
                                "@type": "type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext",
                                "common_tls_context": {
                                  "tls_certificate_sds_secret_configs": [
                                    {
                                      "name": "example.default.svc",
                                      "sds_config": {
                                        "ads": {},
                                        "resource_api_version": "V3"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          }
                        ],
                        "name": "https"
                      }
                    },
                    {
                      "blueprint": "tlsCertificate",
                      "generateFromTlsSecret": "example.default.svc",
                      "type": "secret"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "DiscoveryService",
                "metadata": {
                  "name": "discoveryservice-example",
                  "namespace": "my-namespace"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyDeployment",
                "metadata": {
                  "name": "envoydeployment-sample",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "discoveryServiceRef": "discoveryservice-example",
                  "envoyConfigRef": "envoyconfig-example",
                  "ports": [
                    {
                      "name": "https",
                      "port": 8443
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: quay.io/3scale/marin3r
          createdAt: "2024-08-05T16:08:25Z"
          description: Lighweight, CRD based Envoy control plane for Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["envoyconfigrevisions.marin3r.3scale.net","discoveryservicecertificates.operator.marin3r.3scale.net"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/3scale-ops/marin3r
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DiscoveryServiceCertificate is an internal resource used
              to create certificates. This resource is used by the DiscoveryService
              controller to create the required certificates for the different components.
              Direct use of DiscoveryServiceCertificate objects is discouraged.
            displayName: DiscoveryServiceCertificate
            kind: DiscoveryServiceCertificate
            name: discoveryservicecertificates.operator.marin3r.3scale.net
            version: v1alpha1
          - description: DiscoveryService represents an envoy discovery service server.
              Only one instance per namespace is currently supported.
            displayName: DiscoveryService
            kind: DiscoveryService
            name: discoveryservices.operator.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfigRevision is an internal resource that stores a
              specific version of an EnvoyConfig resource. EnvoyConfigRevisions are
              automatically created and deleted by the EnvoyConfig controller and
              are not intended to be directly used. Use EnvoyConfig objects instead.
            displayName: EnvoyConfigRevision
            kind: EnvoyConfigRevision
            name: envoyconfigrevisions.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfig holds the configuration for a given envoy nodeID.
              The spec of an EnvoyConfig object holds the Envoy resources that conform
              the desired configuration for the given nodeID and that the discovery
              service will send to any envoy client that identifies itself with that
              nodeID.
            displayName: EnvoyConfig
            kind: EnvoyConfig
            name: envoyconfigs.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyDeployment is a resource to deploy and manage a Kubernetes
              Deployment of Envoy Pods.
            displayName: EnvoyDeployment
            kind: EnvoyDeployment
            name: envoydeployments.operator.marin3r.3scale.net
            version: v1alpha1
        description: |
          MARIN3R implements a control plane to deploy, configure and operate a fleet of envoy instances within a Kubernetes cluster.
          It operates a group of gateways and Pod sidecar containers to provide general purpose functionality required to operate production systems.

          Features:

          * Deploy and manage an Envoy xDS server using the DiscoveryService custom resource.
          * Inject Envoy sidecar containers based on Pod annotations.
          * Deploy Envoy as a Kubernetes Deployment using the EnvoyDeployment custom resource.
          * Dynamic Envoy configuration and re-configuration using the EnvoyConfig custom resource.
          * Use Kubernetes Secrets as certificate sources.
          * Syntactic validation of Envoy configurations.
          * Lifecycle management of Envoy containers (graceful shutdown and connection draining).
          * Self-healing.

          Check the [docs](https://github.com/3scale-ops/marin3r#readme) for more information.

          ## License
          MARIN3R is licensed under the [Apache 2.0 license](https://github.com/3scale/prometheus-exporter-operator/blob/master/LICENSE)
        displayName: MARIN3R
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - envoy
        - discovery
        - controlplane
        - proxy
        - sidecar
        - xds
        - network
        - gateway
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/marin3r
        maintainers:
        - email: 3scale-operations+marin3r@redhat.com
          name: 3scale Operations Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/marin3r:v0.13.1
        version: 0.13.1
      entries:
      - name: marin3r.v0.13.1
        version: 0.13.1
      - name: marin3r.v0.12.3
        version: 0.12.3
      - name: marin3r.v0.11.1
        version: 0.11.1
      - name: marin3r.v0.10.0
        version: 0.10.0
      name: stable
    defaultChannel: stable
    packageName: marin3r
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: machine-deletion-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: machine-deletion-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/machine-deletion-remediation-operator:v0.1.0
          createdAt: "2023-05-11 10:46:12"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/machine-deletion-remediation-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: machine-deletion-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: machine-deletion-remediation.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/machine-deletion-remediation-operator:v0.3.0
          createdAt: "2024-03-05 18:45:16"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/medik8s/machine-deletion-remediation-operator:v0.3.0
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        version: 0.3.0
      entries:
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      - name: machine-deletion-remediation.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: machine-deletion-remediation
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-nimbusmosaic-dusim
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-nimbusmosaic-dusim.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Nimbusmosaicdusim",
                "metadata": {
                  "name": "nimbusmosaicdusim-sample"
                },
                "spec": {
                  "development": false,
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "microservices": {
                    "jvm": "-Xmx300M -Xms64M -Xss256k"
                  },
                  "service": {
                    "type": "ClusterIP"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:41:56Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nimbusmosaicdusim
            name: nimbusmosaicdusims.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Nimbus Mosaic DUSim Operator
        displayName: Keysight Nimbus Mosaic DUSim Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Nimbusmosaic Dusim
          url: https://keysight-nimbusmosaic-dusim.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-nimbusmosaic-dusim@sha256:655dd6595a9c8e82947bed24991ac52f8fa55be99ca410e7d9f0026c51ef4520
        version: 0.1.21
      entries:
      - name: keysight-nimbusmosaic-dusim.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-nimbusmosaic-dusim
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-namespace-scope-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-namespace-scope-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1",
                "kind": "NamespaceScope",
                "metadata": {
                  "name": "namespacescope"
                },
                "spec": {
                  "namespaceMembers": [
                    "ibm-common-services",
                    "default"
                  ],
                  "restartLabels": {
                    "intent": "projected"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
          createdAt: "2025-02-19T00:54:24Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.2.14
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/IBM/ibm-namespace-scope-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NamespaceScope is the Schema for the namespacescopes API
            displayName: Namespace Scope
            kind: NamespaceScope
            name: namespacescopes.operator.ibm.com
            version: v1
        description: This operator automates the extension of operator watch and service
          account permission scope to other namespaces in an openshift cluster.
        displayName: IBM NamespaceScope Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Namespace Scope Operator
          url: https://github.com/IBM/ibm-namespace-scope-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:0774b1d916ebc354050612474800dda3a501c66a7a5aab6d18698c8c4e81ac4e
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-bundle@sha256:b6cb9e5976b749dd7818ac19d1cfd5e24673f3c052a41d6e2fe1d2d7094694b4
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:7a2cd72736bf9d30341763fd1652ff4b9ce3c785da5123645ee7664ee38c7283
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-catalog@sha256:c016904e8c5cfcf8270241ad5e28b52f528859984ed894cdd5e1283fa90cd6e8
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator-bundle@sha256:c743dd7de2edbd08c6c73e20cf15f3b147491ad77a81677339e88caffb93e84b
        version: 4.2.14
      entries:
      - name: ibm-namespace-scope-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-namespace-scope-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hitachi
      provider-url: ""
    name: hnas-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hnas-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.csi.hitachi.com/v1",
                "kind": "HNAS",
                "metadata": {
                  "name": "hnas",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-operator@sha256:9d3f9f10fc7172ee9cebbd9ee318906756d7e96909f2e66e0534e98bb53e0a7b
          createdAt: "2025-07-30T15:12:23Z"
          description: An operator for managing Hitachi NAS CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HNAS is the Schema for the hnas API
            displayName: HNAS
            kind: HNAS
            name: hnas.charts.csi.hitachi.com
            version: v1
        description: |-
          The Hitachi NAS CSI driver operator deploys the Hitachi NAS CSI driver on Kubernetes and OpenShift environments.  The driver can be used to create and manage persistent storage hosted on Hitachi NAS products, including those in the Hitachi VSP One File 3x range and the HNAS range of products.

          For the complete documentation, go to our [Product Documentation site](https://docs.hitachivantara.com/home) and refer to the Hitachi NAS CSI Driver User Guide - MK-92HNAS091.  The guide covers prerequisites for the NAS system and covers the driver functionality and all configuration options available for the driver.
        displayName: Hitachi NAS CSI Driver Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hitachi
        - Storage
        - Container Storage Interface
        - CSI
        - NAS
        links:
        - name: Product Documentation
          url: https://docs.hitachivantara.com/home
        maturity: stable
        minKubeVersion: 1.28.0
        provider:
          name: Hitachi
        relatedImages:
        - registry.k8s.io/sig-storage/csi-resizer@sha256:ab774734705a906561e15b67f6a96538f3ad520335d636f793aaafe87a3b5200
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:dd788d79cf4c1b8edee6d9b80b8a1ebfc51a38a365c5be656986b129be9ac784
        - registry.k8s.io/sig-storage/livenessprobe@sha256:2c5f9dc4ea5ac5509d93c664ae7982d4ecdec40ca7b0638c24e5b16243b8360f
        - registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-operator@sha256:9d3f9f10fc7172ee9cebbd9ee318906756d7e96909f2e66e0534e98bb53e0a7b
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:d7138bcc3aa5f267403d45ad4292c95397e421ea17a0035888850f424c7de25d
        - registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-driver@sha256:6b6f6816c9f4e608e93e7fbc0347255415426a0ff0c9a53b7750e5ba54d91262
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d5e46da8aff7d73d6f00c761dae94472bcda6e78f4f17b3802dc89d44de0111b
        - registry.k8s.io/sig-storage/csi-attacher@sha256:a399393ff5bd156277c56bae0c08389b1a1b95b7fd6ea44a316ce55e0dd559d7
        version: 1.0.0
      entries:
      - name: hnas-csi-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: hnas-csi-operator
    provider:
      name: Hitachi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-memorydb-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-memorydb-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "ACL",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "ParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "Snapshot",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "SubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/memorydb-controller:1.1.1
          createdAt: "2025-09-19T19:36:49Z"
          description: Amazon MemoryDB for Redis controller is a service controller
            for managing MemoryDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ACL represents the state of an AWS memorydb ACL resource.
            displayName: ACL
            kind: ACL
            name: acls.memorydb.services.k8s.aws
            version: v1alpha1
          - description: Cluster represents the state of an AWS memorydb Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.memorydb.services.k8s.aws
            version: v1alpha1
          - description: ParameterGroup represents the state of an AWS memorydb ParameterGroup
              resource.
            displayName: ParameterGroup
            kind: ParameterGroup
            name: parametergroups.memorydb.services.k8s.aws
            version: v1alpha1
          - description: Snapshot represents the state of an AWS memorydb Snapshot
              resource.
            displayName: Snapshot
            kind: Snapshot
            name: snapshots.memorydb.services.k8s.aws
            version: v1alpha1
          - description: SubnetGroup represents the state of an AWS memorydb SubnetGroup
              resource.
            displayName: SubnetGroup
            kind: SubnetGroup
            name: subnetgroups.memorydb.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS memorydb User resource.
            displayName: User
            kind: User
            name: users.memorydb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon MemoryDB for Redis resources in AWS from within your Kubernetes cluster.

          **About Amazon MemoryDB for Redis**

          MemoryDB for Redis is a durable, in-memory database service that delivers ultra-fast performance. It is purpose-built for modern applications with microservices architectures.
          MemoryDB is compatible with Redis, a popular open source data store, enabling you to quickly build applications using the same flexible and friendly Redis data structures, APIs, and commands that they already use today. With MemoryDB, all of your data is stored in memory, which enables you to achieve microsecond read and single-digit millisecond write latency and high throughput. MemoryDB also stores data durably across multiple Availability Zones (AZs) using a Multi-AZ transactional log to enable fast failover, database recovery, and node restarts.
          Delivering both in-memory performance and Multi-AZ durability, MemoryDB can be used as a high-performance primary database for your microservices applications, eliminating the need to separately manage both a cache and durable database.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - MemoryDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - memorydb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon MemoryDB for Redis Developer Resources
          url: https://docs.aws.amazon.com/memorydb/?id=docs_gateway
        maintainers:
        - email: ack-maintainers@amazon.com
          name: memorydb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/memorydb-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-memorydb-controller.v1.1.1
        version: 1.1.1
      - name: ack-memorydb-controller.v1.1.0
        version: 1.1.0
      - name: ack-memorydb-controller.v1.0.15
        version: 1.0.15
      - name: ack-memorydb-controller.v1.0.13
        version: 1.0.13
      - name: ack-memorydb-controller.v1.0.12
        version: 1.0.12
      - name: ack-memorydb-controller.v1.0.11
        version: 1.0.11
      - name: ack-memorydb-controller.v1.0.10
        version: 1.0.10
      - name: ack-memorydb-controller.v1.0.9
        version: 1.0.9
      - name: ack-memorydb-controller.v1.0.8
        version: 1.0.8
      - name: ack-memorydb-controller.v1.0.7
        version: 1.0.7
      - name: ack-memorydb-controller.v1.0.6
        version: 1.0.6
      - name: ack-memorydb-controller.v1.0.5
        version: 1.0.5
      - name: ack-memorydb-controller.v1.0.4
        version: 1.0.4
      - name: ack-memorydb-controller.v1.0.3
        version: 1.0.3
      - name: ack-memorydb-controller.v1.0.2
        version: 1.0.2
      - name: ack-memorydb-controller.v1.0.1
        version: 1.0.1
      - name: ack-memorydb-controller.v1.0.0
        version: 1.0.0
      - name: ack-memorydb-controller.v0.0.12
        version: 0.0.12
      - name: ack-memorydb-controller.v0.0.11
        version: 0.0.11
      - name: ack-memorydb-controller.v0.0.10
        version: 0.0.10
      - name: ack-memorydb-controller.v0.0.9
        version: 0.0.9
      - name: ack-memorydb-controller.v0.0.8
        version: 0.0.8
      - name: ack-memorydb-controller.v0.0.7
        version: 0.0.7
      - name: ack-memorydb-controller.v0.0.6
        version: 0.0.6
      - name: ack-memorydb-controller.v0.0.5
        version: 0.0.5
      - name: ack-memorydb-controller.v0.0.4
        version: 0.0.4
      - name: ack-memorydb-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-memorydb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: GitLab, Inc.
      provider-url: https://gitlab.com
    name: gitlab-runner-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-runner-operator.v1.41.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta2",
                "kind": "Runner",
                "metadata": {
                  "name": "gitlab-runner-example",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "gitlabUrl": "https://gitlab.example.com",
                  "token": "runner-token-secret",
                  "tags": "kubernetes,example",
                  "concurrent": 10,
                  "imagePullPolicy": "IfNotPresent"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery, Developer Tools
          containerImage: registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
          createdAt: "2025-09-25T15:08:40Z"
          description: GitLab Runner operator manages lifecycle of GitLab Runner instances
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.11.0 <1.41.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: none
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
          support: GitLab, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Runner is the open source project used to run your jobs and
              send the results back to GitLab
            displayName: GitLab Runner
            kind: Runner
            name: runners.apps.gitlab.com
            version: v1beta2
        description: |-
          GitLab Runner is the lightweight, highly-scalable agent that runs your build jobs and sends the results back to a GitLab instance. GitLab Runner works in conjunction with GitLab CI/CD, the open-source continuous integration service included with GitLab.

              The GitLab Runner operator manages the lifecycle of GitLab Runner in Kubernetes or Openshift clusters. The operator aims to automate the tasks needed to run your CI/CD jobs in your container orchestration platform.

              ## Prerequisites

                For Kubernetes cluster, install cert-manager:

                ```shell
                kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.7.1/cert-manager.yaml
                ```

              ## GitLab Runner version

              This version of **GitLab Runner Operator** ships with **GitLab Runner v18.4.0**.

              To use a different version of **GitLab Runner** change the [`runnerImage` and `helperImage` properties](https://docs.gitlab.com/runner/configuration/configuring_runner_operator.html#operator-properties).

              ## Usage

               To link a GitLab Runner instance to a self-hosted GitLab instance or the hosted [GitLab](https://gitlab.com), you first need to:

               1. Create a secret containing the `runner-token` from your GitLab project.

                ```
                cat > gitlab-runner-secret.yml << EOF
                apiVersion: v1
                kind: Secret
                metadata:
                  name: gitlab-runner-secret
                type: Opaque
                # Only one of the following fields must be set. The Operator will fail to register Runner if both are provided.
                # # NOTE: runner-registration-token is deprecated and will be removed in GitLab 18.0. You should use runner-token instead.
                stringData:
                  runner-token: REPLACE_ME # your project runner token
                  # runner-registration-token: "" # your project runner secret
                EOF
                ```

                ```
                oc apply -f gitlab-runner-secret.yml
                ```

               2. Create the Custom Resource Definition (CRD) file and include the following information. The tags value must be openshift for the job to run.

                 ```
                 cat > gitlab-runner.yml << EOF
                 apiVersion: apps.gitlab.com/v1beta2
                 kind: Runner
                 metadata:
                   name: gitlab-runner
                 spec:
                   gitlabUrl: https://gitlab.example.com
                   buildImage: alpine
                   token: gitlab-runner-secret
                   tags: openshift
                 EOF
                 ```

                ```
                oc apply -f gitlab-runner.yml
                ```

              ## Full documentation

              Visit [Install GitLab Runner Operator](https://docs.gitlab.com/runner/install/operator.html)
        displayName: GitLab Runner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GitLab
        - CI/CD
        - DevOps
        - SAST
        - DAST
        links:
        - name: Gitlab Runner Operator
          url: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
        - name: GitLab Docs
          url: https://docs.gitlab.com
        - name: GitLab Runner
          url: https://docs.gitlab.com/runner/
        - name: GitLab CI/CD
          url: https://docs.gitlab.com/ee/ci/quick_start/
        maintainers:
        - email: support@gitlab.com
          name: Georgi N. Georgiev
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: GitLab, Inc.
          url: https://gitlab.com
        relatedImages:
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-ocp@sha256:0f2a5a3b6779d719eecbf079b99cf06b652603d74b563f57d753e32966d253dd
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-helper-ocp@sha256:29727fae83a226bcf57607240d4046a0b8c208b5ff42ee1f8ca0f115bb188809
        version: 1.41.1
      entries:
      - name: gitlab-runner-operator.v1.41.1
        version: 1.41.1
      - name: gitlab-runner-operator.v1.41.0
        version: 1.41.0
      - name: gitlab-runner-operator.v1.40.1
        version: 1.40.1
      - name: gitlab-runner-operator.v1.40.0
        version: 1.40.0
      - name: gitlab-runner-operator.v1.39.0
        version: 1.39.0
      name: stable
    defaultChannel: stable
    packageName: gitlab-runner-operator
    provider:
      name: GitLab, Inc.
      url: https://gitlab.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Anjuna Security
      provider-url: https://www.anjuna.io
    name: anjuna-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anjuna-operator.v2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anjuna.io/v1",
                "kind": "AnjunaConfig",
                "metadata": {
                  "name": "anjuna-config-remote"
                },
                "spec": {
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "runtimeImage": "registry.connect.redhat.com/anjuna/anjuna-k8s-sev-runtime@sha256:1531b1ba955b2d3ea5d9238d2e20617a534bb1bab73f5fadadf8c9d3aee410b4"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-02-19T01:11:18Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          operatorframework.io/suggested-namespace: anjuna-operator
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnjunaConfig operand instructs the Anjuna Seaglass Operator
              to know (1) what needs to be installed into the cluster and (2) what
              nodes will be affected.
            displayName: Anjuna Config
            kind: AnjunaConfig
            name: anjunaconfigs.anjuna.io
            version: v1
        description: "The Anjuna Seaglass Operator for OpenShift allows running any
          Pod in\nOpenShift in a Confidential Pod leveraging the full power of the
          Confidential\nComputing technology.\n\n# Anjuna Seaglass\nAnjuna Seaglass™
          is the Universal Confidential Computing Platform to run applications in
          any cloud, and on-premises, with complete data security and privacy. \nIt
          isolates workloads in a hardware-backed environment that creates a zero-trust
          model for code and data.\n\nThe Anjuna Seaglass platform allows users to
          leverage the power of Confidential Computing:\n\nFor any Confidential Computing
          technology\nIn any Cloud Service Provider, as well as on-premises\nWith
          no changes to the application\n\nRunning a Confidential Pod with Anjuna
          Seaglass protects the Pod's memory with hardware-enforced encryption and
          integrity protection, and enables you to authenticate and trust the workload
          using Remote Attestation.\n\n# Data-in-use protection\nWhen protecting sensitive
          data managed by a Pod in OpenShift, one can protect the data at rest by
          encrypting the data written to the persistent storage, and protecting the
          data in transit when sending data over the network.\n\nHowever, when the
          data is used in the Pod's memory, it cannot be both encrypted and used at
          the same time.\nThis means that if a malicious insider (either human or
          software) had access to the OpenShift Node, they could dump the Pod's memory
          and access all of its sensitive data. This includes accessing the secrets
          used to encrypt the data at rest and in transit.\n\nRunning the Pod in a
          Confidential Pod with the Anjuna Seaglass platform solves this problem.
          Anjuna Seaglass runs the Pod inside a Trusted Execution Environment, where
          the Pod memory is encrypted by the hardware so that no privileged human
          or software on the Pod can access this memory in the clear.\n\n# Enabling
          zero trust for the Pod\nWhen running a Pod in OpenShift, the Pod may need
          to create trust with other Pods or with software running outside of the
          OpenShift cluster.\nThe standard way to create this trust is to identify
          with a token or credentials that could be easily read and stolen from the
          Pod disk or by providing a machine identity, a process name, or an IP that
          could be spoofed or reused.\n\nConfidential Computing technology provides
          a mechanism called Remote Attestation, which provides an identity to the
          software via the Hardware—the Software Measurement.\n\nWhen running a Pod
          as a Confidential Pod with the Anjuna Seaglass platform, a user can leverage
          the Remote Attestation to perform trusted authentication with the Confidential
          Pod, in a way that cannot be spoofed.\nThe Anjuna Policy Manager allows
          you to protect the first secret of a Confidential Pod and remove any secret
          from the Pod image or Kubernetes control plane that could be read by malicious
          insiders.\n\n# Quick start and configuration instructions\nFollow the directions
          in Installing the [Anjuna Kubernetes Toolset to OpenShift on GCP](https://docs.anjuna.io/k8s-sev/latest/installation/openshift_gcp.html)
          \nto learn how to deploy the Operator into your OpenShift cluster.\n\nFollow
          the directions in [Quickstart guides for OpenShift on GCP](https://docs.anjuna.io/k8s-sev/latest/applications/gcp.html)
          \nto learn how to deploy your first Confidential Pod in your OpenShift cluster.\n\nFollow
          the directions in [Attestation with Anjuna Policy Manager](https://docs.anjuna.io/k8s-sev/latest/attestation.html)
          \nto learn how to leverage Remote Attestation to protect your Pod's first
          secret."
        displayName: Anjuna Seaglass™ Operator for OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Security
        links:
        - name: Anjuna Operator Documentation
          url: https://docs.anjuna.io/k8s-sev/latest/index.html
        - name: Anjuna Seaglass
          url: https://www.anjuna.io/essence-of-anjuna-seaglass
        - name: Anjuna Security
          url: https://www.anjuna.io
        maintainers:
        - email: support@anjuna.io
          name: Anjuna Security
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Anjuna Security
          url: https://www.anjuna.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - registry.connect.redhat.com/anjuna/anjuna-k8s-sev-tools@sha256:e5ba31c240aaff8ca8f6addf12577fd6045bf6619a1d2cb2987b2ff5d75f4028
        - registry.connect.redhat.com/anjuna/anjuna-k8s-sev-runtime@sha256:1531b1ba955b2d3ea5d9238d2e20617a534bb1bab73f5fadadf8c9d3aee410b4
        version: 2.0.0
      entries:
      - name: anjuna-operator.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anjuna-operator
    provider:
      name: Anjuna Security
      url: https://www.anjuna.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzograph-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzograph-operator.v3.2.500
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzograph.clusters.cambridgesemantics.com/v2",
                "kind": "AnzoGraph",
                "metadata": {
                  "name": "azg01"
                },
                "spec": {
                  "db": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_data": "anzograph-data-grpc",
                            "app_mgmt": "anzograph-mgmt-grpc"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_data": "anzograph-data-grpc",
                              "app_mgmt": "anzograph-mgmt-grpc"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzograph-db@sha256:30ed2ec337135701ab8b54e75bb48ad8928a2c48818643c23b9dba4484255a2e",
                                "name": "db",
                                "resources": {
                                  "limits": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  },
                                  "requests": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  }
                                }
                              }
                            ],
                            "securityContext": {
                              "fsGroup": 1000,
                              "runAsNonRoot": true,
                              "runAsUser": 1000,
                              "seccompProfile": {
                                "type": "RuntimeDefault"
                              }
                            },
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  },
                  "deployFrontend": false,
                  "frontend": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_mgmt": "anzograph-frontend"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_mgmt": "anzograph-frontend"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzograph-frontend@sha256:e5446b3c04889f2189e955702c4165eb243eb89f47ac797eb417b60b1459647f",
                                "name": "frontend",
                                "resources": {
                                  "limits": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "securityContext": {
                              "fsGroup": 1000,
                              "runAsNonRoot": true,
                              "runAsUser": 1000,
                              "seccompProfile": {
                                "type": "RuntimeDefault"
                              }
                            },
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/anzograph-operator@sha256:2333154a14ab7f3a03a601f00c017feeafb9d9814aff5e59c95cef615ab2abd6
          createdAt: "2025-09-22T12:03:18Z"
          description: kubernetes operator for Altair Graph Lakehouse DB
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Altair Engineering Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoGraph is the Schema for the anzographs API
            displayName: Altair Graph Lakehouse Operator
            kind: AnzoGraph
            name: anzographs.anzograph.clusters.cambridgesemantics.com
            version: v2
        description: |-
          The Altair Graph Lakehouse Operator provides the way to install and configure an Altair Graph Lakehouse
          cluster on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v3.2.5/README_openshift_marketplace.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzograph/userdoc/ )

          ### Support

          We offer Support to our customers with the Altair Graph Lakehouse db Enterprise Edition License [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Lakehouse Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        - Database
        links:
        - name: Anzograph Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v3.2.5/README_openshift_marketplace.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzograph-operator@sha256:2333154a14ab7f3a03a601f00c017feeafb9d9814aff5e59c95cef615ab2abd6
        - registry.connect.redhat.com/cambridgesemantics/anzograph-db@sha256:30ed2ec337135701ab8b54e75bb48ad8928a2c48818643c23b9dba4484255a2e
        - registry.connect.redhat.com/cambridgesemantics/anzograph-frontend@sha256:e5446b3c04889f2189e955702c4165eb243eb89f47ac797eb417b60b1459647f
        version: 3.2.500
      entries:
      - name: anzograph-operator.v3.2.500
        version: 3.2.500
      - name: anzograph-operator.v3.2.300
        version: 3.2.300
      - name: anzograph-operator.v3.2.200
        version: 3.2.200
      - name: anzograph-operator.v3.2.104
        version: 3.2.104
      name: stable
    defaultChannel: stable
    packageName: anzograph-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: kubernetes-nmstate-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubernetes-nmstate-operator.4.13.0-202309261702
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nmstate.io/v1",
                "kind": "NMState",
                "metadata": {
                  "name": "nmstate"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift4/kubernetes-nmstate-rhel8-operator@sha256:bdcb80591b839ec6b4b94efa838ca41585f3044826534413b91efd2acdca4b4d
          createdAt: "2022-02-21 08:46:16"
          description: |
            Kubernetes NMState is a declaritive means of configuring NetworkManager.
          olm.skipRange: '>=4.3.0 <4.13.0'
          operatorframework.io/suggested-namespace: openshift-nmstate
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/kubernetes-nmstate
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an NMState deployment.
            displayName: NMState
            kind: NMState
            name: nmstates.nmstate.io
            version: v1
        description: A Kubernetes Operator to install Kubernetes NMState
        displayName: Kubernetes NMState Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - nmstate
        - networking
        - NetworkManager
        links:
        - name: Kubernetes Nmstate Operator
          url: https://github.com/nmstate/kubernetes-nmstate
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: GA
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/kubernetes-nmstate-rhel8-operator@sha256:bdcb80591b839ec6b4b94efa838ca41585f3044826534413b91efd2acdca4b4d
        - registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel8@sha256:a96434b959c1f61c1cfb4352284aaae453d2e41ca0824d016975a1216451b665
        version: 4.13.0-202309261702
      entries:
      - name: kubernetes-nmstate-operator.4.13.0-202309261702
        version: 4.13.0-202309261702
      name: "4.13"
    - currentCSV: kubernetes-nmstate-operator.4.16.0-202509241326
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nmstate.io/v1",
                "kind": "NMState",
                "metadata": {
                  "name": "nmstate"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:14750ac40ba2a9a92e2b37351eed94c22f2663747ea09f241a3a80607a9fca09
          createdAt: "2022-02-21 08:46:16"
          description: |
            Kubernetes NMState is a declaritive means of configuring NetworkManager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-nmstate
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/kubernetes-nmstate
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an NMState deployment.
            displayName: NMState
            kind: NMState
            name: nmstates.nmstate.io
            version: v1
        description: A Kubernetes Operator to install Kubernetes NMState
        displayName: Kubernetes NMState Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - nmstate
        - networking
        - NetworkManager
        links:
        - name: Kubernetes Nmstate Operator
          url: https://github.com/nmstate/kubernetes-nmstate
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: GA
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:14750ac40ba2a9a92e2b37351eed94c22f2663747ea09f241a3a80607a9fca09
        - registry.redhat.io/openshift4/nmstate-console-plugin-rhel9@sha256:8d03a6c414da30521e0794d3d61336bfbb1d9d4c3fec5ab5dac4bd24c897a540
        - registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel9@sha256:c592fdb65d99de2c3d274d7e72621408e578bdba39753bb425d0765ac016b11a
        version: 4.16.0-202509241326
      entries:
      - name: kubernetes-nmstate-operator.4.16.0-202509241326
        version: 4.16.0-202509241326
      - name: kubernetes-nmstate-operator.4.16.0-202509101329
        version: 4.16.0-202509101329
      - name: kubernetes-nmstate-operator.4.16.0-202508251030
        version: 4.16.0-202508251030
      - name: kubernetes-nmstate-operator.4.16.0-202507221936
        version: 4.16.0-202507221936
      - name: kubernetes-nmstate-operator.4.16.0-202507100305
        version: 4.16.0-202507100305
      - name: kubernetes-nmstate-operator.4.16.0-202506250035
        version: 4.16.0-202506250035
      - name: kubernetes-nmstate-operator.4.16.0-202506041738
        version: 4.16.0-202506041738
      - name: kubernetes-nmstate-operator.4.16.0-202505150035
        version: 4.16.0-202505150035
      - name: kubernetes-nmstate-operator.4.16.0-202505060805
        version: 4.16.0-202505060805
      - name: kubernetes-nmstate-operator.4.16.0-202504141106
        version: 4.16.0-202504141106
      - name: kubernetes-nmstate-operator.4.16.0-202503260136
        version: 4.16.0-202503260136
      - name: kubernetes-nmstate-operator.4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: kubernetes-nmstate-operator.4.16.0-202502190034
        version: 4.16.0-202502190034
      - name: kubernetes-nmstate-operator.4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: kubernetes-nmstate-operator.4.16.0-202501281135
        version: 4.16.0-202501281135
      - name: kubernetes-nmstate-operator.4.16.0-202501211505
        version: 4.16.0-202501211505
      - name: kubernetes-nmstate-operator.4.16.0-202501080105
        version: 4.16.0-202501080105
      - name: kubernetes-nmstate-operator.4.16.0-202412180105
        version: 4.16.0-202412180105
      - name: kubernetes-nmstate-operator.4.16.0-202412110604
        version: 4.16.0-202412110604
      - name: kubernetes-nmstate-operator.4.16.0-202412040032
        version: 4.16.0-202412040032
      - name: kubernetes-nmstate-operator.4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: kubernetes-nmstate-operator.4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: kubernetes-nmstate-operator.4.16.0-202410300036
        version: 4.16.0-202410300036
      - name: kubernetes-nmstate-operator.4.16.0-202410231737
        version: 4.16.0-202410231737
      - name: kubernetes-nmstate-operator.4.16.0-202410180404
        version: 4.16.0-202410180404
      - name: kubernetes-nmstate-operator.4.16.0-202410091137
        version: 4.16.0-202410091137
      - name: kubernetes-nmstate-operator.4.16.0-202410021935
        version: 4.16.0-202410021935
      - name: kubernetes-nmstate-operator.4.16.0-202409251605
        version: 4.16.0-202409251605
      - name: kubernetes-nmstate-operator.4.16.0-202409111235
        version: 4.16.0-202409111235
      - name: kubernetes-nmstate-operator.4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: kubernetes-nmstate-operator.4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: kubernetes-nmstate-operator.4.16.0-202408211407
        version: 4.16.0-202408211407
      - name: kubernetes-nmstate-operator.4.16.0-202408140143
        version: 4.16.0-202408140143
      - name: kubernetes-nmstate-operator.4.16.0-202408070540
        version: 4.16.0-202408070540
      - name: kubernetes-nmstate-operator.4.16.0-202407251436
        version: 4.16.0-202407251436
      - name: kubernetes-nmstate-operator.4.16.0-202407240806
        version: 4.16.0-202407240806
      - name: kubernetes-nmstate-operator.4.16.0-202407181806
        version: 4.16.0-202407181806
      - name: kubernetes-nmstate-operator.4.16.0-202407101306
        version: 4.16.0-202407101306
      - name: kubernetes-nmstate-operator.4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: kubernetes-nmstate-operator.4.16.0-202406260236
        version: 4.16.0-202406260236
      - name: kubernetes-nmstate-operator.4.16.0-202406140306
        version: 4.16.0-202406140306
      - name: kubernetes-nmstate-operator.4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: kubernetes-nmstate-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: istio-workspace-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: istio-workspace-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "workspace.maistra.io/v1alpha1",
                "kind": "Session",
                "metadata": {
                  "name": "session-sample"
                },
                "spec": {
                  "ref": [
                    {
                      "args": {
                        "image": "quay.io/maistra-dev/istio-workspace-test:latest"
                      },
                      "name": "ratings-v1",
                      "strategy": "prepared-image"
                    }
                  ],
                  "route": {
                    "name": "x-workspace-route",
                    "type": "header",
                    "value": "feature-x"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/maistra/istio-workspace:v0.5.3
          createdAt: "2022-11-16T13:50:35Z"
          description: Safely develop and test on any Kubernetes cluster without affecting
            others.
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/maistra/istio-workspace
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Session controls the creation of the specialized hidden routes.
            displayName: Session
            kind: Session
            name: sessions.workspace.maistra.io
            version: v1alpha1
        description: "Istio Workspace enables developers to:\n\n * run one or more
          services locally during development but make it appear like it would be
          developed in the actual cluster\n   * this way you can use your favourite
          development tools\n   * have instant feedback loop\n   * don't make your
          machine overheating trying to run the whole cluster locally\n * safely preview
          incoming changes in pull requests without affecting regular users \n * have
          confidence in testing or troubleshooting your services directly in the cluster\n\nIstio
          Workspace supports testing on multi-user environment in an unobtrusive way.\nIt
          doesn’t really matter if it is QE cluster or actual production. We give
          you confidence that your changes won’t blow up the cluster, and your users
          won’t even notice a glitch.\n\nYou will need [Kubernetes](https://k8s.io)
          or [Openshift](https://openshift.com) cluster with  [Istio](https://istio.io/)
          installed. \n\nYou can read more about our vision on [Red Hat Developer’s
          blog](https://developers.redhat.com/blog/2020/07/14/developing-and-testing-on-production-with-kubernetes-and-istio-workspace/)
          or ...\n\n## See it in action!\n\n[![https://youtu.be/XTNVadUzMCc](https://img.youtube.com/vi/XTNVadUzMCc/hqdefault.jpg)](https://youtu.be/XTNVadUzMCc)\n\nWatch
          demo: [\"How to develop on production: An introduction to Istio-Workspaces\"](https://youtu.be/XTNVadUzMCc).\n\n##
          Documentation\n\nMore details can be found on our [documentation page](https://istio-workspace-docs.netlify.com/).\n"
        displayName: Istio Workspace
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - telepresence
        links:
        - name: Istio Workspace Documentation
          url: https://istio-workspace-docs.netlify.com/
        - name: How to install CLI tool
          url: https://github.com/maistra/istio-workspace#install-in-two-easy-steps
        maintainers:
        - email: aslak@redhat.com
          name: Aslak Knutsen
        - email: bartosz@redhat.com
          name: Bartosz Majsak
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/maistra/istio-workspace:v0.5.3
        version: 0.5.3
      entries:
      - name: istio-workspace-operator.v0.5.3
        version: 0.5.3
      - name: istio-workspace-operator.v0.5.2
        version: 0.5.2
      - name: istio-workspace-operator.v0.5.1
        version: 0.5.1
      - name: istio-workspace-operator.v0.5.0
        version: 0.5.0
      - name: istio-workspace-operator.v0.4.0
        version: 0.4.0
      - name: istio-workspace-operator.v0.3.0
        version: 0.3.0
      - name: istio-workspace-operator.v0.2.0
        version: 0.2.0
      - name: istio-workspace-operator.v0.1.0
        version: 0.1.0
      - name: istio-workspace-operator.v0.0.10
        version: 0.0.10
      - name: istio-workspace-operator.v0.0.9
        version: 0.0.9
      - name: istio-workspace-operator.v0.0.8
        version: 0.0.8
      - name: istio-workspace-operator.v0.0.7
        version: 0.0.7
      - name: istio-workspace-operator.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: istio-workspace-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibmcloud-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibmcloud-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"ibmcloud.ibm.com/v1","kind":"Service","metadata":{"annotations":{"ibmcloud.ibm.com/self-healing":"enabled"},"name":"mytranslator"},"spec":{"plan":"lite","serviceClass":"language-translator"}},{"apiVersion":"ibmcloud.ibm.com/v1","kind":"Binding","metadata":{"name":"binding-translator"},"spec":{"parameters":[{"name":"serviceid_crn","value":"crn:v1:bluemix:public:iam-identity::a/ACCOUNT_ID::serviceid:SERVICE_ID"}],"secretName":"translator-secret","serviceName":"mytranslator"}}]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: cloudoperators/ibmcloud-operator:1.1.0
          createdAt: "2022-03-17T04:17:40Z"
          description: The IBM Cloud Operator provides a Kubernetes CRD-Based API
            to manage the lifecycle of IBM public cloud services.
          repository: https://github.com/IBM/cloud-operators
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Binding is an instance of a service binding resource on IBM
              Cloud. A Binding creates a secret with the service instance credentials.
            displayName: Binding
            kind: Binding
            name: bindings.ibmcloud.ibm.com
            version: v1
          - description: Service is an instance of a Service resource on IBM Cloud.
            displayName: Service
            kind: Service
            name: services.ibmcloud.ibm.com
            version: v1
        description: |-
          With the IBM Cloud Operator, you can provision and bind [IBM public cloud services](https://cloud.ibm.com/catalog#services) to your Kubernetes cluster in a Kubernetes-native way. The IBM Cloud Operator is based on the [Kubernetes custom resource definition (CRD) API](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/) so that your applications can create, update, and delete IBM Cloud services from within the cluster by calling Kubnernetes APIs, instead of needing to use several IBM Cloud APIs in addition to configuring your app for Kubernetes.


          ## Features

          * **Service provisioning**: Create any service with any plan that is available in the [IBM Cloud catalog](https://cloud.ibm.com/catalog#services).

          * **Bindings management**: Automatically create Kubernetes secrets in your Kubernetes cluster with the credentials to bind to any provisioned service to the cluster.


          ## Upgrading the operator


          To upgrade an OperatorHub installation, use the OpenShift web console to perform the upgrade. If upgrading from the `alpha` channel subscription to `stable`, follow the additional upgrade instructions below.

          ### Upgrading to version v0.3 or v1.0
          **IMPORTANT NOTICE:** v0.1 and v0.2 used a different naming scheme for secrets and configmaps. Before you update the IBM Cloud Operator, create secret and configmap resources with these names and copy the contents of your previous resources to the new resources. Then, the upgraded operator recognizes and continues to update the resources.

          Existing `Service` and `Binding` resources **do not** need modification for the upgrade.

          | Previous names (v0.1 and v0.2)         | **Current names (v0.3 and v1.0)**             | Description                                                                           |
          |:---------------------------------------|:--------------------------------------------|:--------------------------------------------------------------------------------------|
          | secret-ibm-cloud-operator              | **ibmcloud-operator-secret**                | Secret with the API key, scoped to the namespace.                                     |
          | config-ibm-cloud-operator              | **ibmcloud-operator-defaults**              | ConfigMap with the default values for new resources.                                  |
          | ibm-cloud-operator                     | **ibmcloud-operator-config**                | ConfigMap with the management namespace configuration.                                |
          | ${namespace}-secret-ibm-cloud-operator | **${namespace}-ibmcloud-operator-secret**   | Management namespace Secret with the API key for ${namespace}.                        |
          | ${namespace}-config-ibm-cloud-operator | **${namespace}-ibmcloud-operator-defaults** | Management namespace ConfigMap with default values for new resources in ${namespace}. |

          **Tip:** Forgot to update the secret and configmap names before upgrading? The operator will not take action on Service and Binding resources until the new secrets and configmaps have been created. Creating these after the upgrade will also work.


          ### OperatorHub stable channel
          Continuous updates and bug fixes are provided in the latest stable channel. Subscribing to the `stable` channel in OperatorHub means you automatically get the latest updates without breaking backward compatibility.

          ### Changelog

          For a list of recent changes, see the [releases page](https://github.com/IBM/cloud-operators/releases).


          ## Prerequisites

          1.  Have an [IBM Cloud account](https://cloud.ibm.com/registration).
          2.  Have a cluster that runs Kubernetes version 1.11 or later (OpenShift 3.11 or later).
          3.  Install the required command line tools.
              *   [IBM Cloud CLI](https://cloud.ibm.com/docs/cli?topic=cloud-cli-getting-started) (`ibmcloud`)
              *   [Kubernetes CLI](https://kubernetes.io/docs/tasks/tools/install-kubectl/) (`kubectl`)
          4.  Log in to your IBM Cloud account from the CLI.

              ```bash
              ibmcloud login
              ```

          5.  Target the appropriate resource group (`-g`) and default region (`-r`) for provisioning services:

              ```bash
              ibmcloud target -g default -r us-south
              ```

          6.  If you also want to provision Cloud Foundry services, target the appropriate `org` and `space` (`--cf`).

              ```bash
              ibmcloud target --cf
              ```

          7.  Set the Kubernetes context of your CLI to your cluster so that you can run `kubectl` commands. For example, if your cluster runs OpenShift, use the `oc login` command.


          ## Setting up the operator

          You can use an installation script to set up the IBM Cloud Operator.
          By default, the installation script stores an API key in a Kubernetes secret in your cluster that can be accessed by the IBM Cloud Operator.
          Next, the script sets default values that are used to provision IBM Cloud services, like the resource group and region to provision the services in. Later, you can override any default value in the `Service` custom resource. Finally, the script deploys the operator in your cluster.

          To use your own API key, set the `IBMCLOUD_API_KEY` environment variable to the key before running the installation script:

              export IBMCLOUD_API_KEY="CUSTOM_API_KEY"

          If installed with Operator Hub, the operator will run in the `openshift-operators` namespace. Otherwise, it will run in the `ibmcloud-operator-system` namespace.

          Prefer to create the secrets and defaults yourself? See the [IBM Cloud Operator installation guide](https://github.com/IBM/cloud-operators/blob/master/docs/install.md).

          ### Using a service ID

          By default, the installation script creates an IBM Cloud API key that impersonates your user credentials, to use to set up the IBM Cloud Operator. However, you might want to create a service ID in IBM Cloud Identity and Access Managment (IAM). By using a service ID, you can control access for the IBM Cloud Operator without having the permissions tied to a particular user, such as if that user leaves the company. For more information, see the [IBM Cloud docs](https://cloud.ibm.com/docs/account?topic=account-serviceids).

          1.  Create a service ID in IBM Cloud IAM. If possible, do not use spaces in the names for your IAM credentials. When you use the operator binding feature, any spaces are replaced with underscores.

              ```bash
              ibmcloud iam service-id-create serviceid-ico -d service-ID-for-ibm-cloud-operator
              ```

          2.  Assign the service ID access to the required permissions to work with the IBM Cloud services that you want the operator to manage. The required permissions vary with each IBM Cloud service. You can also scope an access policy to different regions or resource groups. For example, the following command grants the service ID the **Administrator** platform role in the default resource group in the US South (Dallas) region. For more information, see the [IBM Cloud docs](https://cloud.ibm.com/docs/account?topic=account-userroles).

              ```bash
              ibmcloud iam service-policy-create serviceid-ico --roles Administrator --resource-group-name default --region us-south
              ```

          3.  Create an API key for the service ID.

              ```bash
              ibmcloud iam service-api-key-create apikey-ico serviceid-ico -d api-key-for-ibm-cloud-operator
              ```

          4.  Set the API key of the service ID as your CLI environment variable. Now, when you run the installation script, the script uses the service ID's API key. The following command is an example for macOS.

              ```bash
              export IBMCLOUD_API_KEY=<apikey-ico-value>
              ```

          5.  Confirm that the API key environment variable is set in your CLI.

              ```bash
              echo $IBMCLOUD_API_KEY
              ```

          6.  Follow the [prerequisite steps](https://github.com/IBM/cloud-operators/blob/master/README.md#prerequisites) to log in to the IBM Cloud account that owns the service ID.


          ### Installing the operator for OpenShift clusters

          Before you begin, complete the [prerequisite steps](https://github.com/IBM/cloud-operators/blob/master/README.md#prerequisites) to log in to IBM Cloud and your cluster, and optionally set up a [service ID API key](https://github.com/IBM/cloud-operators/blob/master/README.md#using-a-service-id).

          To configure the latest release for OpenShift before installing via the OperatorHub, run the following script:

          *   **Latest release**:

              ```bash
              curl -sL https://raw.githubusercontent.com/IBM/cloud-operators/master/hack/configure-operator.sh | bash
              ```

          *   **Specific release**: Replace `-v 0.0.0` with the specific version that you want to install.

              ```bash
              curl -sL https://raw.githubusercontent.com/IBM/cloud-operators/master/hack/configure-operator.sh | bash -s -- -v 0.0.0 store-creds
              ```


          ## Using the IBM Cloud Operator

          To use the IBM Cloud Operator, create a service instance and then bind the service to your cluster. For more information, see the [sample configuration files](https://github.com/IBM/cloud-operators/blob/master/config/samples), [user guide](https://github.com/IBM/cloud-operators/blob/master/docs/user-guide.md), and [reference documentation](https://github.com/IBM/cloud-operators/blob/master/README.md#reference).

          #### Step 1: Creating a service instance

          1.  To create an instance of an IBM public cloud service, first create a `Service` custom resource file. For more options, see the [Service properties](https://github.com/IBM/cloud-operators/blob/master/README.md#service-properties) reference doc.
              *   `<SERVICE_CLASS>` is the IBM Cloud service that you want to create. To list IBM Cloud services, run `ibmcloud catalog service-marketplace` and use the **Name** value from the output.
              *   `<PLAN>` is the plan for the IBM Cloud service that you want to create, such as `free` or `standard`. To list available plans, run `ibmcloud catalog service <SERVICE_CLASS> | grep plan`.

              ```yaml
              apiVersion: ibmcloud.ibm.com/v1
              kind: Service
              metadata:
                  name: myservice
              spec:
                  plan: <PLAN>
                  serviceClass: <SERVICE_CLASS>
              ```

          2.  Create the service instance in your cluster.

              ```bash
              kubectl apply -f filepath/myservice.yaml
              ```

          3.  Check that your service status is **Online** in your cluster.

              ```bash
              kubectl get services.ibmcloud
              NAME           STATUS   AGE
              myservice      Online   12s
              ```

          4.  Verify that your service instance is created in IBM Cloud.

              ```bash
              ibmcloud resource service-instances | grep myservice
              ```

          #### Step 2: Binding the service instance

          1.  To bind your service to the cluster so that your apps can use the service, create a `Binding` custom resource, where the `serviceName` field is the name of the `Service` custom resource that you previously created. For more options, see [Binding properties](https://github.com/IBM/cloud-operators/blob/master/README.md#binding-properties).

              ```yaml
              apiVersion: ibmcloud.ibm.com/v1
              kind: Binding
              metadata:
                  name: mybinding
              spec:
                  serviceName: myservice
              ```

          2.  Create the binding in your cluster.

              ```bash
              kubectl apply -f filepath/mybinding.yaml
              ```

          3.  Check that your service status is **Online**.

              ```bash
              kubectl get bindings.ibmcloud
              NAME         STATUS   AGE
              mybinding    Online   25s
              ```

          4.  Check that a secret of the same name as your binding is created. The secret contains the service credentials that apps in your cluster can use to access the service.

              ```bash
              kubectl get secrets
              NAME         TYPE      DATA   AGE
              mybinding    Opaque    6      102s
              ```
        displayName: IBM Cloud Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: Source on GitHub
          url: https://github.com/ibm/cloud-operators
        maintainers:
        - email: johnstarich@johnstarich.com
          name: John Starich
        - email: mvaziri@us.ibm.com
          name: vazirim
        - email: greg.lanthier@ca.ibm.com
          name: Greg Lanthier
        - email: halacs@users.noreply.github.com
          name: Halacs
        - email: artberger@users.noreply.github.com
          name: Art
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: IBM
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - cloudoperators/ibmcloud-operator:1.1.0
        version: 1.1.0
      entries:
      - name: ibmcloud-operator.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: ibmcloud-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Infinispan
      provider-url: ""
    name: infinispan
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infinispan-operator.v2.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.2.5.Final
          description: Create and manage Infinispan clusters.
          olm.skipRange: '>=2.1.x <2.2.1'
          operators.operatorframework.io/builder: operator-sdk-v1.3.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/infinispan/operator:2.2.5.Final
        version: 2.2.5
      entries:
      - name: infinispan-operator.v2.2.5
        version: 2.2.5
      - name: infinispan-operator.v2.2.4
        version: 2.2.4
      - name: infinispan-operator.v2.2.3
        version: 2.2.3
      - name: infinispan-operator.v2.2.2
        version: 2.2.2
      - name: infinispan-operator.v2.2.1
        version: 2.2.1
      - name: infinispan-operator.v2.2.0
        version: 2.2.0
      name: 2.2.x
    - currentCSV: infinispan-operator.v2.3.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.3.7.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/infinispan/operator:2.3.7.Final
        version: 2.3.8
      entries:
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: 2.3.x
    - currentCSV: infinispan-operator.v2.4.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.4.17.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/infinispan/operator:2.4.17.Final
        version: 2.4.17
      entries:
      - name: infinispan-operator.v2.4.17
        version: 2.4.17
      - name: infinispan-operator.v2.4.16
        version: 2.4.16
      - name: infinispan-operator.v2.4.15
        version: 2.4.15
      - name: infinispan-operator.v2.4.14
        version: 2.4.14
      - name: infinispan-operator.v2.4.13
        version: 2.4.13
      - name: infinispan-operator.v2.4.12
        version: 2.4.12
      - name: infinispan-operator.v2.4.11
        version: 2.4.11
      - name: infinispan-operator.v2.4.10
        version: 2.4.10
      - name: infinispan-operator.v2.4.9
        version: 2.4.9
      - name: infinispan-operator.v2.4.8
        version: 2.4.8
      - name: infinispan-operator.v2.4.7
        version: 2.4.7
      - name: infinispan-operator.v2.4.6
        version: 2.4.6
      - name: infinispan-operator.v2.4.5
        version: 2.4.5
      - name: infinispan-operator.v2.4.4
        version: 2.4.4
      - name: infinispan-operator.v2.4.3
        version: 2.4.3
      - name: infinispan-operator.v2.4.2
        version: 2.4.2
      - name: infinispan-operator.v2.4.1
        version: 2.4.1
      - name: infinispan-operator.v2.4.0
        version: 2.4.0
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: 2.4.x
    - currentCSV: infinispan-operator.v2.4.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.4.18.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/infinispan/operator:2.4.18.Final
        version: 2.4.18
      entries:
      - name: infinispan-operator.v2.4.18
        version: 2.4.18
      - name: infinispan-operator.v2.4.16
        version: 2.4.16
      - name: infinispan-operator.v2.4.15
        version: 2.4.15
      - name: infinispan-operator.v2.4.14
        version: 2.4.14
      - name: infinispan-operator.v2.4.13
        version: 2.4.13
      - name: infinispan-operator.v2.4.12
        version: 2.4.12
      - name: infinispan-operator.v2.4.11
        version: 2.4.11
      - name: infinispan-operator.v2.4.10
        version: 2.4.10
      - name: infinispan-operator.v2.4.9
        version: 2.4.9
      - name: infinispan-operator.v2.4.8
        version: 2.4.8
      - name: infinispan-operator.v2.4.7
        version: 2.4.7
      - name: infinispan-operator.v2.4.6
        version: 2.4.6
      - name: infinispan-operator.v2.4.5
        version: 2.4.5
      - name: infinispan-operator.v2.4.4
        version: 2.4.4
      - name: infinispan-operator.v2.4.3
        version: 2.4.3
      - name: infinispan-operator.v2.4.2
        version: 2.4.2
      - name: infinispan-operator.v2.4.1
        version: 2.4.1
      - name: infinispan-operator.v2.4.0
        version: 2.4.0
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: stable
    defaultChannel: stable
    packageName: infinispan
    provider:
      name: Infinispan
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent, part of Cisco
      provider-url: https://isovalent.com/
    name: clife
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clife.v1.17.8-cee.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cilium.io/v1alpha1",
                "kind": "CiliumConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "clife"
                  },
                  "name": "ciliumconfig"
                },
                "spec": {
                  "cluster": {
                    "name": "default"
                  },
                  "clusterHealthPort": 9940,
                  "cni": {
                    "binPath": "/var/lib/cni/bin",
                    "chainingMode": "portmap",
                    "confPath": "/var/run/multus/cni/net.d",
                    "exclusive": false
                  },
                  "enterprise": {
                    "featureGate": {
                      "approved": [
                        "CNIChainingMode"
                      ]
                    }
                  },
                  "hubble": {
                    "enabled": true
                  },
                  "ipam": {
                    "mode": "cluster-pool"
                  },
                  "kubeProxyReplacement": false,
                  "operator": {
                    "prometheus": {
                      "enabled": true,
                      "serviceMonitor": {
                        "enabled": true
                      }
                    }
                  },
                  "prometheus": {
                    "enabled": true,
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "securityContext": {
                    "privileged": true
                  },
                  "sessionAffinity": true,
                  "tunnelPort": 4789
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Networking,Security
          createdAt: "2025-09-23T07:55:19Z"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CiliumConfig defines the configuration of Isovalent Networking
              for Kubernetes and all its components
            displayName: CiliumConfig
            kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Networking for Kubernetes, built on Cilium, provides
          eBPF-based networking, observability, and security with optimal scale and
          performance for platform teams operating Kubernetes environments across
          cloud and on-prem infrastructure.
        displayName: Isovalent Networking for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cilium
        - networking
        - observability
        - security
        - eBPF
        - CNI
        links:
        - name: Isovalent Networking for Kubernetes
          url: https://isovalent.com/product/
        maintainers:
        - email: support@isovalent.com
          name: Isovalent, part of Cisco
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Isovalent, part of Cisco
          url: https://isovalent.com/
        relatedImages:
        - quay.io/isovalent/startup-script-ubi:v1.17.8-cee.1@sha256:9a6a413262ec59026345aefc743de847891d80e31ef2f41db954acdcee11c2ff
        - quay.io/isovalent/certgen-ubi:v1.17.8-cee.1@sha256:fb63bed641a279cc09114e93821c7b7555c61393ab594ea2b823ce34fa73335e
        - quay.io/isovalent/cilium-envoy-ubi:v1.17.8-cee.1@sha256:c3de0334f90f9b2540cdec7e2356d813f2df93f0d7d8701005eab7161c7663b5
        - quay.io/isovalent/clife@sha256:ba9bd4a5c35e1d26e395afbc2a4de89f42df1e5e61c9b23959c23194368b7b0e
        - quay.io/isovalent/operator-generic-ubi:v1.17.8-cee.1@sha256:37e4b9f65b2c67d24afeeebebe5d0cb59c09280179de57cd77da2b68e19c3c1b
        - quay.io/isovalent/cilium-ubi:v1.17.8-cee.1@sha256:e9816096e481fad5d95209f902e4a7018aeb65d1e3be1574e6b289e21229f87a
        - quay.io/isovalent/hubble-relay-ubi:v1.17.8-cee.1@sha256:88964d4836d59420dc7e0abb758654ceaf2abcfb53e19bbe2c7962aa68874e37
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.17.8-cee.1@sha256:b097c823638b4fdbe5bdd2b6040a17fef0c70a020e67a18f8231d60eeac02dd5
        version: 1.17.8-cee.1
      entries:
      - name: clife.v1.17.8-cee.1
        version: 1.17.8-cee.1
      - name: clife.v1.17.7-cee.1
        version: 1.17.7-cee.1
      - name: clife.v1.17.5-cee.1
        version: 1.17.5-cee.1
      - name: clife.v1.17.4-cee.1
        version: 1.17.4-cee.1
      - name: clife.v0.0.1
        version: 0.0.1
      name: "1.17"
    defaultChannel: "1.17"
    packageName: clife
    provider:
      name: Isovalent, part of Cisco
      url: https://isovalent.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microsafe
      provider-url: https://microsafe.com.mx/
    name: operator-aims-mobile
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operator-aims-mobile.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.microsafe.com.mx/v1",
                "kind": "AimsMobile",
                "metadata": {
                  "name": "aimsmobile-sample"
                },
                "spec": {
                  "adminApp": {
                    "autoscaling": {
                      "enabled": false
                    },
                    "config": {
                      "NODE_ENV": "production",
                      "VITE_EXPRESS_HOST": "aims-mob-service",
                      "VITE_EXPRESS_PORT": "8080",
                      "VITE_EXPRESS_PROTOCOL": "http",
                      "VITE_KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "VITE_KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "VITE_KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "VITE_VERSION": "1.0.0"
                    },
                    "image": "quay.io/microsafe_app_ban/aims-mob-admin-private@sha256:8e333f900885b0d3f6442c8ffcb43bd93540b8c38ad5bf17cbbd78929e0b1be0",
                    "ingress": {
                      "enabled": true,
                      "isOpenshiftRoute": true
                    },
                    "name": "aims-mob-admin",
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "secretName": "aims-mob-admin-secrets",
                    "service": {
                      "enabled": true,
                      "port": 8083,
                      "targetPort": 8080
                    }
                  },
                  "dbService": {
                    "autoscaling": {
                      "enabled": false
                    },
                    "config": {
                      "MSSQL_HOST": "mssql-database.prod.host",
                      "MSSQL_PORT": "mssql-database.prod.port"
                    },
                    "image": "quay.io/microsafe_app_ban/aims-db-service-private@sha256:bcb4d8624626436526549427f9f169cff714b39cbcc530dd2b3470c6712abaf9",
                    "ingress": {
                      "enabled": false
                    },
                    "name": "aims-db-service",
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "300m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "128Mi"
                      }
                    },
                    "secretName": "aims-db-service-secrets",
                    "service": {
                      "enabled": true,
                      "port": 3000,
                      "targetPort": 3000
                    }
                  },
                  "mainService": {
                    "annotations": {
                      "instrumentation.opentelemetry.io/inject-nodejs": "true"
                    },
                    "autoscaling": {
                      "enabled": true,
                      "maxReplicas": 2,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "config": {
                      "AIMS1_API_URL": "http://aims-db-service:3000",
                      "EXPRESS_PORT": "4000",
                      "FRONTEND_EXPRESS_HOST": "aims-mob-admin",
                      "FRONTEND_EXPRESS_PORT": "4000",
                      "FRONTEND_EXPRESS_PROTOCOL": "http",
                      "FRONTEND_IP": "aims-mob-admin",
                      "FRONTEND_KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "FRONTEND_KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "FRONTEND_KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "FRONTEND_PORT": "80",
                      "FRONTEND_PROTOCOL": "http",
                      "FRONTEND_VERSION": "1.0.0",
                      "KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "MSSQL_HOST": "mssql-database.prod.host",
                      "MSSQL_PORT": "mssql-database.prod.port",
                      "NODE_ENV": "production",
                      "OTLP_LOG_EXPORTER_URL": "opentelemetry-logs.prod.url",
                      "OTLP_METRIC_EXPORTER_URL": "opentelemetry-metrics.prod.url",
                      "OTLP_TRACE_EXPORTER_URL": "opentelemetry-traces.prod.url",
                      "REMEDY_API_URL": "remedy-aims.prod.svc",
                      "REMEDY_USER": "RMDYUSR"
                    },
                    "ingress": {
                      "enabled": true,
                      "isOpenshiftRoute": true
                    },
                    "name": "aims-mob-service",
                    "proxyContainer": {
                      "image": "quay.io/microsafe_app_ban/aims-mob-proxy-private@sha256:29b82de987b28c23f17d5f2176877c5aecea1a0ef182bb3a63612290ff469949",
                      "name": "aims-mob-service-proxy",
                      "resources": {
                        "limits": {
                          "cpu": "2000m",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "replicas": 2,
                    "secretName": "aims-mob-service-secrets",
                    "service": {
                      "enabled": true,
                      "port": 8085,
                      "targetPort": 8081
                    },
                    "serviceContainer": {
                      "image": "quay.io/microsafe_app_ban/aims-mob-service-private@sha256:f9586200106aa4de259f5cae2d22d13dc147a76dfd06e4f4bea09797071dad17",
                      "name": "aims-mob-service-main",
                      "resources": {
                        "limits": {
                          "cpu": "2500m",
                          "memory": "2048Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "128Mi"
                        }
                      }
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": false,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          createdAt: "2025-09-25T19:23:26Z"
          description: Este operador despliega los servicios AIMS Mobile.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/network-aware: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          instrumentation.opentelemetry.io/inject-nodejs: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: Valid Subscription Microsafe
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://quay.io/repository/microsafe_app_ban/operador-aims-mobile-private
          support: Microsafe
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AimsMobile
            name: aimsmobiles.cache.microsafe.com.mx
            version: v1
        description: Este operador despliega la aplicación AIMS Mobile.
        displayName: AIMS Mobile Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aims mobile
        - aims
        links:
        - name: AIMS Mobile Operator
          url: https://microsafe.com.mx/
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Microsafe
          url: https://microsafe.com.mx/
        relatedImages:
        - quay.io/microsafe_app_ban/aims-db-service-private@sha256:bcb4d8624626436526549427f9f169cff714b39cbcc530dd2b3470c6712abaf9
        - quay.io/microsafe_app_ban/aims-mob-admin-private@sha256:8e333f900885b0d3f6442c8ffcb43bd93540b8c38ad5bf17cbbd78929e0b1be0
        - quay.io/microsafe_app_ban/operador-aims-mobile-private@sha256:731f2d731c79ade02f4b6c39b380590833cacddfd795b703cec10ad55f730fad
        - quay.io/microsafe_app_ban/aims-mob-proxy-private@sha256:29b82de987b28c23f17d5f2176877c5aecea1a0ef182bb3a63612290ff469949
        - quay.io/microsafe_app_ban/aims-mob-service-private@sha256:f9586200106aa4de259f5cae2d22d13dc147a76dfd06e4f4bea09797071dad17
        version: 0.0.10
      entries:
      - name: operator-aims-mobile.v0.0.10
        version: 0.0.10
      - name: operator-aims-mobile.v0.0.9
        version: 0.0.9
      - name: operator-aims-mobile.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: operator-aims-mobile
    provider:
      name: Microsafe
      url: https://microsafe.com.mx/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzograph-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzograph-operator.v2.0.102
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzograph.clusters.cambridgesemantics.com/v2",
                "kind": "AnzoGraph",
                "metadata": {
                  "name": "azg01"
                },
                "spec": {
                  "db": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_data": "anzograph-data-grpc",
                            "app_mgmt": "anzograph-mgmt-grpc"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_data": "anzograph-data-grpc",
                              "app_mgmt": "anzograph-mgmt-grpc"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-db@sha256:8a5bd26551990929fdb776d6b421a9713cb84e03b19c18f81b625582f7f41af5",
                                "name": "db",
                                "resources": {
                                  "limits": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  },
                                  "requests": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  },
                  "deployFrontend": false,
                  "frontend": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_mgmt": "anzograph-frontend"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_mgmt": "anzograph-frontend"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-frontend@sha256:c1443bcdc8e998fc699f61980a665f1a083350adf2d3d1d55a7cb55e31ca50b3",
                                "name": "frontend",
                                "resources": {
                                  "limits": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-operator@sha256:a5b212558f748ddc8156541b546290320b656506e039a18903ad9fa6e638e732
          createdAt: "2021-08-11"
          description: kubernetes operator for AnzoGraph DB
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzograph-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzograph-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoGraph is the Schema for the anzographs API
            displayName: AnzoGraph Operator
            kind: AnzoGraph
            name: anzographs.anzograph.clusters.cambridgesemantics.com
            version: v2
        description: |-
          The AnzoGraph Operator provides the way to install and configure an AnzoGraph
          cluster on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v2.0.1/README_openshift_marketplace.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzograph/userdoc/ )

          ### Support

          We offer Support to our customers with the AnzoGraph db Enterprise Edition License [here]( https://customercenter.cambridgesemantics.com/ ). For AnzoGraph db Free Edition questions, get help from our Anzograph User Community at Stack Overflow. When submitting a question, include the tag 'anzograph'.
        displayName: AnzoGraph Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        - Database
        links:
        - name: Anzograph Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v2.0.1/README_openshift_marketplace.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-operator@sha256:a5b212558f748ddc8156541b546290320b656506e039a18903ad9fa6e638e732
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph@sha256:e3bdddf811b5ba69f388d023aea1ba538edd00e9f415c33ec555321d39129a36
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-db@sha256:8a5bd26551990929fdb776d6b421a9713cb84e03b19c18f81b625582f7f41af5
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-frontend@sha256:c1443bcdc8e998fc699f61980a665f1a083350adf2d3d1d55a7cb55e31ca50b3
        version: 2.0.1
      entries:
      - name: anzograph-operator.v2.0.102
        version: 2.0.1
      name: stable
    defaultChannel: stable
    packageName: anzograph-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fortanix Inc.
      provider-url: https://fortanix.com
    name: ccm-node-agent-dcap-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ccm-node-agent-dcap-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1",
                "kind": "CcmNodeAgent",
                "metadata": {
                  "name": "ccmnodeagent-sample"
                },
                "spec": {
                  "ccmJoinToken": {
                    "secretName": "ccm-join-token",
                    "token": ""
                  },
                  "fullnameOverride": "",
                  "image": {
                    "aesmd": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/aesmd-dcap",
                      "tag": "latest"
                    },
                    "ccmNodeAgent": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/ccm-node-agent-ubi7",
                      "tag": "latest"
                    },
                    "ccmNodeAgentInit": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "busybox",
                      "tag": "latest"
                    },
                    "multiPlatformRegistration": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/multi-platform-registration-ubi8",
                      "tag": "latest"
                    }
                  },
                  "log": {
                    "debug": false
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "ports": {
                    "containerPort": 9092,
                    "hostPort": 9092
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CCM Node Agent software enables registration of the compute
              nodes to Fortanix CCM when installed in the Openshift Cluster Nodes<U+200B>.
            displayName: CCM Node Agent
            kind: CcmNodeAgent
            name: ccmnodeagents.charts.operatorhub.io
            version: v1
        description: "CCM Node Agent software enables registration of the compute
          nodes to Fortanix CCM when installed in the Openshift Cluster Nodes​. \nIt
          also assists with application attestation and visibility for Fortanix CCM.\n\n##
          About this Operator\nThis Operator is based on a Helm chart for CCM Node
          Agent. It enrolls all the nodes in the openshift cluster to a particular
          CCM account. You have to provide the base64 encode of join token of CCM
          account while creating CcmNodeAgent resource.\n\n## Prerequisites\nAll the
          worker nodes in the cluster must have SGX capabilities, in-kernel driver
          installed and fsgsbase support.\n\n## Installation\n * Choose a namespace
          to install the Operator. \n Operator will be available in all namespaces.\n*
          Select Approval Strategy: Available options are Automatic or Manual. Automatic
          approval will automatically upgrade the running instance of your Operator
          without human intervention when a new version is available. In case of Manual
          approval, when a new version is available, OLM will create an update request.
          You must manually approve the request to update the Operator to a new version.\n*
          Generate your CCM Account Join Token: Please log in to https://ccm.fortanix.com,
          and in the Management Console tab, click the + ENROLL NODE button. In the
          ENROLL NODE screen, a Join Token will be generated in the text box for \"Get
          a join token to register an SGX compute node\". This Join Token is used
          by the compute node to authenticate itself.\n* Base64 encode the Join Token:\n```shell\necho
          -n <join-token> | openssl base64 -A\n```\n* Create an instance of CcmNodeAgent
          by setting the `spec.ccmJoinToken.Token` to the base64 encoded CCM account
          join token.\n* After CcmNodeAgent resource is installed successfully you
          should be able to see all the worker nodes of the cluster in your CCM account
          under Compute Node Tab.\n\n## Getting Started\nNow to run your application
          inside SGX enclaves on your cluster follow the [documentation](https://support.fortanix.com/hc/en-us/articles/360043527431-User-s-Guide-Add-and-Edit-an-Application).\n"
        displayName: CCM Node Agent DCAP Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dockerfile
        - Helm
        - CCM
        - Confidential Compute
        links:
        - name: CCM Node Agent DCAP Operator
          url: https://support.fortanix.com/hc/en-us/categories/360003107511-Confidential-Computing-Manager
        maintainers:
        - email: info@fortanix.com
          name: Fortanix Inc.
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Fortanix Inc.
          url: https://fortanix.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62575404000cb77fa8d2284c87b31e39e667a55b0fb47821263b23b3d1135a7e
        - docker.io/fortanix/ccm-node-agent-dcap-operator@sha256:4fd28bea26aee461eb9c1800cdf5f041b64ae3ea6652203f8ad9feb680aaefee
        version: 0.0.2
      entries:
      - name: ccm-node-agent-dcap-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ccm-node-agent-dcap-operator
    provider:
      name: Fortanix Inc.
      url: https://fortanix.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Infinidat
      provider-url: https://infinidat.com
    name: infinibox-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infinibox-operator-certified.v2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csidriver.infinidat.com/v1alpha1",
                "kind": "InfiniboxCsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infinibox-operator-certified",
                    "app.kubernetes.io/instance": "infiniboxcsidriver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infiniboxcsidriver",
                    "app.kubernetes.io/part-of": "infinibox-operator-certified"
                  },
                  "name": "infiniboxcsidriver-sample"
                },
                "spec": {
                  "Ibox_Creds": [
                    {
                      "SecretName": "infinibox-creds",
                      "hostname": "ibox0000",
                      "inbound_secret": "0.0000000000000",
                      "inbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonout",
                      "outbound_secret": "0.0000000000001",
                      "outbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonin",
                      "password": "my-pool-admin-passwd",
                      "username": "my-pool-admin"
                    }
                  ],
                  "Infinibox_Protocol_Secrets": [
                    {
                      "SecretName": "infinibox-protocol-secret",
                      "iscsi": {
                        "network_space": "default_iscsi_space",
                        "useCHAP": "none"
                      },
                      "nfs": {
                        "network_space": "default_nas_space",
                        "nfs_export_permissions": ""
                      },
                      "nvme": {
                        "network_space": "default_nvme_space"
                      },
                      "preferred_auto_order": "fc,nvme,iscsi",
                      "storage_protocol": "fc"
                    }
                  ],
                  "autoUpdate": false,
                  "cleanupNFSPerms": false,
                  "controllerserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "controllerservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "createEvents": true,
                  "csiDriverName": "infinibox-csi-driver",
                  "csiDriverVersion": "v2.23.0",
                  "e2etesting": false,
                  "images": {
                    "attachersidecar": "registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901",
                    "attachersidecar_pull_policy": "IfNotPresent",
                    "csidriver": "registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61",
                    "csidriver_pull_policy": "Always",
                    "livenesssidecar": "registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc",
                    "livenesssidecar_pull_policy": "IfNotPresent",
                    "provisionersidecar": "registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2",
                    "provisionersidecar_pull_policy": "IfNotPresent",
                    "registrarsidecar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24",
                    "registrarsidecar_pull_policy": "IfNotPresent",
                    "resizersidecar": "registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d",
                    "resizersidecar_pull_policy": "IfNotPresent",
                    "snapshottersidecar": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a",
                    "snapshottersidecar_pull_policy": "IfNotPresent"
                  },
                  "instanceCount": 1,
                  "logLevel": "debug",
                  "nodeSelector": {
                    "kubernetes.io/os": "linux"
                  },
                  "nodeserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "nodeservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "protocolSecret": "",
                  "removeDomainName": false,
                  "replicaCount": 1,
                  "skipCredentialsCreation": false,
                  "volumeNamePrefix": "ibox"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
          createdAt: "2025-09-27T16:12:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2.23.0
          operatorframework.io/suggested-namespace: infinidat-csi
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Infinidat/infinibox-csi-driver
          support: Infinidat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Iboxreplica
            name: iboxreplicas.csidriver.infinidat.com
            version: v1
          - kind: InfiniboxCsiDriver
            name: infiniboxcsidrivers.csidriver.infinidat.com
            version: v1alpha1
        description: "Infinidat InfiniBox Container Storage Interface (CSI) Driver
          is a\n    CNCF-compliant Kubernetes integration for InfiniBox storage systems,
          offering\n    advanced enterprise functionality for petabyte-scale Kubernetes
          deployments including\n    Red Hat OpenShift.\n \n## Features and Benefits\n
          \ \n* **Multi-protocol flexibility** - manage\n    Kubernetes Persistent
          Volumes attached via block and file protocols,\n    including Fibre Channel,
          iSCSI, and NFS, with all Kubernetes PV access modes\n    \n* **Multi-petabyte
          scalability** - support hundreds of thousands of PVs per InfiniBox\n    system
          and control multiple InfiniBox arrays within a single Kubernetes cluster\n
          \   \\ \n* **Advanced enterprise features** - manage native InfiniBox snapshots
          and\n    clones, including restoring from snapshots, and import PVs created
          outside of\n    InfiniBox CSI Driver\n  \n## Required Parameters\n  \n*
          `hostname` - IP address\n    or hostname of the InfiniBox management interface\n*
          `username` / `password` -\n    InfiniBox credentials\n* `SecretName` - secret
          name, to be used in the StorageClass\n    to define a specific InfiniBox
          for persistent volumes\n\n## Optional Parameters\n*\n    `inbound_user`
          / `inbound_secret` / `outbound_user` / `outbound_secret` - credentials \n
          \   for iSCSI CHAP authentication\n\n    \n## Installation Instructions\n
          \n1. Create *infinidat-csi* namespace to install operator into.\n    \n2.
          Delete the following *clusterrolebindings*, if they exist, before installing
          operator\n     * infinidat-csi-operator-infinidat-csi-attacher\n     * infinidat-csi-operator-infinidat-csi-controller
          \                    \n     * infinidat-csi-operator-infinidat-csi-driver
          \                    \n     * infinidat-csi-operator-infinidat-csi-node
          \                              \n     * infinidat-csi-operator-infinidat-csi-provisioner
          \                       \n     * infinidat-csi-operator-infinidat-csi-resizer
          \          \n     * infinidat-csi-operator-infinidat-csi-snapshotter \n
          \   \n3. Install Infinidat Infinibox CSI Driver - Operator (this) into the
          *infinidat-csi* namespace.\n    \n4. Apply the following Security Context
          Constraints to the created service accounts in the *infinidat-csi* \n    namespace\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    \n5. Recreate or edit any existing storage classes, pointing to the
          new namespace\n\n     \n6. If you choose a different namespace other than
          *infinidat-csi* you will need to edit the clusterrolebindings listed above
          to specify your custom namespace\n\n     \n7. Install an instance of Infinidat
          Driver into the *infinidat-csi* namespace"
        displayName: InfiniBox CSI Driver - Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - csi
        - nas
        - san
        - storage
        - infinidat
        - infinibox
        links:
        - name: Infinidat
          url: https://infinidat.com
        maintainers:
        - email: info@infinidat.com
          name: Infinidat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Infinidat
          url: https://infinidat.com
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified@sha256:42ff34a019bc67076515e89735b4bb9921fbd4610c127ac8b61e1fa6a6a08194
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24
        version: 2.23.0
      entries:
      - name: infinibox-operator-certified.v2.23.0
        version: 2.23.0
      - name: infinibox-operator-certified.v2.22.0
        version: 2.22.0
      - name: infinibox-operator-certified.v2.21.0
        version: 2.21.0
      name: alpha
    - currentCSV: infinibox-operator-certified.v2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csidriver.infinidat.com/v1alpha1",
                "kind": "InfiniboxCsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infinibox-operator-certified",
                    "app.kubernetes.io/instance": "infiniboxcsidriver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infiniboxcsidriver",
                    "app.kubernetes.io/part-of": "infinibox-operator-certified"
                  },
                  "name": "infiniboxcsidriver-sample"
                },
                "spec": {
                  "Ibox_Creds": [
                    {
                      "SecretName": "infinibox-creds",
                      "hostname": "ibox0000",
                      "inbound_secret": "0.0000000000000",
                      "inbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonout",
                      "outbound_secret": "0.0000000000001",
                      "outbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonin",
                      "password": "my-pool-admin-passwd",
                      "username": "my-pool-admin"
                    }
                  ],
                  "Infinibox_Protocol_Secrets": [
                    {
                      "SecretName": "infinibox-protocol-secret",
                      "iscsi": {
                        "network_space": "default_iscsi_space",
                        "useCHAP": "none"
                      },
                      "nfs": {
                        "network_space": "default_nas_space",
                        "nfs_export_permissions": ""
                      },
                      "nvme": {
                        "network_space": "default_nvme_space"
                      },
                      "preferred_auto_order": "fc,nvme,iscsi",
                      "storage_protocol": "fc"
                    }
                  ],
                  "autoUpdate": false,
                  "cleanupNFSPerms": false,
                  "controllerserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "controllerservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "createEvents": true,
                  "csiDriverName": "infinibox-csi-driver",
                  "csiDriverVersion": "v2.23.0",
                  "e2etesting": false,
                  "images": {
                    "attachersidecar": "registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901",
                    "attachersidecar_pull_policy": "IfNotPresent",
                    "csidriver": "registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61",
                    "csidriver_pull_policy": "Always",
                    "livenesssidecar": "registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc",
                    "livenesssidecar_pull_policy": "IfNotPresent",
                    "provisionersidecar": "registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2",
                    "provisionersidecar_pull_policy": "IfNotPresent",
                    "registrarsidecar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24",
                    "registrarsidecar_pull_policy": "IfNotPresent",
                    "resizersidecar": "registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d",
                    "resizersidecar_pull_policy": "IfNotPresent",
                    "snapshottersidecar": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a",
                    "snapshottersidecar_pull_policy": "IfNotPresent"
                  },
                  "instanceCount": 1,
                  "logLevel": "debug",
                  "nodeSelector": {
                    "kubernetes.io/os": "linux"
                  },
                  "nodeserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "nodeservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "protocolSecret": "",
                  "removeDomainName": false,
                  "replicaCount": 1,
                  "skipCredentialsCreation": false,
                  "volumeNamePrefix": "ibox"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
          createdAt: "2025-09-27T16:12:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2.23.0
          operatorframework.io/suggested-namespace: infinidat-csi
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Infinidat/infinibox-csi-driver
          support: Infinidat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Iboxreplica
            name: iboxreplicas.csidriver.infinidat.com
            version: v1
          - kind: InfiniboxCsiDriver
            name: infiniboxcsidrivers.csidriver.infinidat.com
            version: v1alpha1
        description: "Infinidat InfiniBox Container Storage Interface (CSI) Driver
          is a\n    CNCF-compliant Kubernetes integration for InfiniBox storage systems,
          offering\n    advanced enterprise functionality for petabyte-scale Kubernetes
          deployments including\n    Red Hat OpenShift.\n \n## Features and Benefits\n
          \ \n* **Multi-protocol flexibility** - manage\n    Kubernetes Persistent
          Volumes attached via block and file protocols,\n    including Fibre Channel,
          iSCSI, and NFS, with all Kubernetes PV access modes\n    \n* **Multi-petabyte
          scalability** - support hundreds of thousands of PVs per InfiniBox\n    system
          and control multiple InfiniBox arrays within a single Kubernetes cluster\n
          \   \\ \n* **Advanced enterprise features** - manage native InfiniBox snapshots
          and\n    clones, including restoring from snapshots, and import PVs created
          outside of\n    InfiniBox CSI Driver\n  \n## Required Parameters\n  \n*
          `hostname` - IP address\n    or hostname of the InfiniBox management interface\n*
          `username` / `password` -\n    InfiniBox credentials\n* `SecretName` - secret
          name, to be used in the StorageClass\n    to define a specific InfiniBox
          for persistent volumes\n\n## Optional Parameters\n*\n    `inbound_user`
          / `inbound_secret` / `outbound_user` / `outbound_secret` - credentials \n
          \   for iSCSI CHAP authentication\n\n    \n## Installation Instructions\n
          \n1. Create *infinidat-csi* namespace to install operator into.\n    \n2.
          Delete the following *clusterrolebindings*, if they exist, before installing
          operator\n     * infinidat-csi-operator-infinidat-csi-attacher\n     * infinidat-csi-operator-infinidat-csi-controller
          \                    \n     * infinidat-csi-operator-infinidat-csi-driver
          \                    \n     * infinidat-csi-operator-infinidat-csi-node
          \                              \n     * infinidat-csi-operator-infinidat-csi-provisioner
          \                       \n     * infinidat-csi-operator-infinidat-csi-resizer
          \          \n     * infinidat-csi-operator-infinidat-csi-snapshotter \n
          \   \n3. Install Infinidat Infinibox CSI Driver - Operator (this) into the
          *infinidat-csi* namespace.\n    \n4. Apply the following Security Context
          Constraints to the created service accounts in the *infinidat-csi* \n    namespace\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    \n5. Recreate or edit any existing storage classes, pointing to the
          new namespace\n\n     \n6. If you choose a different namespace other than
          *infinidat-csi* you will need to edit the clusterrolebindings listed above
          to specify your custom namespace\n\n     \n7. Install an instance of Infinidat
          Driver into the *infinidat-csi* namespace"
        displayName: InfiniBox CSI Driver - Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - csi
        - nas
        - san
        - storage
        - infinidat
        - infinibox
        links:
        - name: Infinidat
          url: https://infinidat.com
        maintainers:
        - email: info@infinidat.com
          name: Infinidat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Infinidat
          url: https://infinidat.com
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified@sha256:42ff34a019bc67076515e89735b4bb9921fbd4610c127ac8b61e1fa6a6a08194
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24
        version: 2.23.0
      entries:
      - name: infinibox-operator-certified.v2.23.0
        version: 2.23.0
      - name: infinibox-operator-certified.v2.22.0
        version: 2.22.0
      - name: infinibox-operator-certified.v2.21.0
        version: 2.21.0
      name: stable
    defaultChannel: stable
    packageName: infinibox-operator-certified
    provider:
      name: Infinidat
      url: https://infinidat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Shipwright Contributors
      provider-url: https://shipwright.io
    name: shipwright-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: shipwright-operator.v0.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "shipwright-build"
                },
                "spec": {
                  "targetNamespace": "shipwright-build"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: ghcr.io/shipwright-io/operator/operator:0.13.0@sha256:12cd5f936043c96f31cf571d7a4f8b6f2e240852cd9ad1a5423dbab0c208b542
          description: Shipwright is a framework for building container images on
            Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/shipwright-io/operator
          support: The Shipwright Contributors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
        description: |
          Shipwright is a framework for building container images on Kubernetes.

          Read more: [https://shipwright.io](https://shipwright.io)

          ## Usage

          To deploy and manage [Shipwright Builds](https://github.com/shipwright-io/build) in your cluster,
          first make sure this operator is installed and running on your cluster.

          Next, create the following:

          ```yaml
          ---
          apiVersion: operator.shipwright.io/v1alpha1
          kind: ShipwrightBuild
          metadata:
            name: shipwright-operator
          spec:
            targetNamespace: shipwright-build
          ```

          The operator will deploy Shipwright Builds in the provided `targetNamespace`.
          When `.spec.targetNamespace` is not set, the namespace will default to `shipwright-build`.
          Refer to the [ShipwrightBuild documentation](https://github.com/shipwright-io/operator/blob/main/docs/shipwrightbuild.md) for more information about this custom resource.
        displayName: Shipwright Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://shipwright.io
        - name: Shipwright Builds
          url: https://github.com/shipwright-io/build
        maintainers:
        - email: shipwright-dev@lists.shipwright.io
          name: The Shipwright Contributors
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Shipwright Contributors
          url: https://shipwright.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - ghcr.io/shipwright-io/operator/operator:0.13.0@sha256:12cd5f936043c96f31cf571d7a4f8b6f2e240852cd9ad1a5423dbab0c208b542
        version: 0.13.0
      entries:
      - name: shipwright-operator.v0.13.0
        version: 0.13.0
      - name: shipwright-operator.v0.9.0
        version: 0.9.0
      name: alpha
    defaultChannel: alpha
    packageName: shipwright-operator
    provider:
      name: The Shipwright Contributors
      url: https://shipwright.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ThinkParQ
      provider-url: https://www.thinkparq.com
    name: beegfs-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: beegfs-csi-driver-operator.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "beegfs.csi.netapp.com/v1",
                "kind": "BeegfsDriver",
                "metadata": {
                  "name": "csi-beegfs-cr"
                },
                "spec": {
                  "nodeAffinityControllerService": {
                    "preferredDuringSchedulingIgnoredDuringExecution": [
                      {
                        "preference": {
                          "matchExpressions": [
                            {
                              "key": "node-role.kubernetes.io/master",
                              "operator": "Exists"
                            }
                          ]
                        },
                        "weight": 50
                      }
                    ],
                    "requiredDuringSchedulingIgnoredDuringExecution": {
                      "nodeSelectorTerms": [
                        {
                          "matchExpressions": [
                            {
                              "key": "node.openshift.io/os_id",
                              "operator": "NotIn",
                              "values": [
                                "rhcos"
                              ]
                            }
                          ]
                        }
                      ]
                    }
                  },
                  "nodeAffinityNodeService": {
                    "requiredDuringSchedulingIgnoredDuringExecution": {
                      "nodeSelectorTerms": [
                        {
                          "matchExpressions": [
                            {
                              "key": "node.openshift.io/os_id",
                              "operator": "NotIn",
                              "values": [
                                "rhcos"
                              ]
                            }
                          ]
                        }
                      ]
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: ghcr.io/thinkparq/beegfs-csi-driver-operator:v1.7.0
          createdAt: 2021-06-30:03:11:00
          description: Install and manage the BeeGFS CSI driver
          operatorframework.io/suggested-namespace: beegfs-csi
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ThinkParQ/beegfs-csi-driver
          support: ThinkParQ
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys the BeeGFS CSI driver
            displayName: BeeGFS Driver
            kind: BeegfsDriver
            name: beegfsdrivers.beegfs.csi.netapp.com
            version: v1
        description: |
          The BeeGFS Container Storage Interface (CSI) driver provides high performing and scalable storage for workloads
          running in container orchestrators like Kubernetes. It allows containers to access existing datasets or request
          on-demand ephemeral or persistent high speed storage backed by BeeGFS parallel file systems.

          This operator eases the installation and management of the BeeGFS CSI driver in an OpenShift or other Operator
          Lifecycle Manager enabled cluster.

          ### Documentation

          Compatibility information and a quick start guide for the BeeGFS CSI driver itself are available on our [main GitHub
          page](https://www.github.com/ThinkParQ/beegfs-csi-driver). Additional detailed documentation for the driver is
          available in the [docs directory](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/docs) and for the
          operator is available in the [operator
          README](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/operator/README.md).

          ### Important Notes

          The BeeGFS client packages and kernel module must be pre-installed on all nodes that will run the driver. See the
          driver ["Getting Started" guide](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/README.md#getting-started)
          for additional information.

          The BeeGFS CSI driver is only supported on nodes running a [BeeGFS supported Linux
          distribution](https://doc.beegfs.io/latest/release_notes.html#supported-linux-distributions-and-kernels). Red Hat
          CoreOS (RHCOS), the default distribution in OpenShift environments, is NOT supported. However, Red Hat Enterprise
          Linux (RHEL) nodes can be added to an OpenShift cluster and RHEL is supported for running BeeGFS. By default, the
          operator will install the driver with a node affinity that ensures it does not run on RHCOS nodes. See the driver
          [compatibility
          matrix](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/README.md#interoperability-and-csi-feature-matrix)
          for additional information.

          We have recently published experimental manifests that enable the BeeGFS CSI driver on OpenShift nodes running
          RHCOS. These manifests are not intended for production use. See [this
          README](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/deploy/openshift-beegfs-client/README.md) for more
          details.

          All driver configuration can be specified through the BeegfsDrivers API (in a BeegfsDriver Custom Resource)
          except for BeeGFS connAuth information, which must be placed separately into a Kubernetes Secret. See the [ConnAuth
          Configuration
          section](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/operator/README.md#connauth-configuration) of the
          operator documentation for additional information.

          Starting with version 1.4.0 of the BeeGFS CSI driver the driver and operator images will be signed using the Cosign
          tool. Refer to the [BeeGFS CSI driver operator
          README](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/operator/README.md)
          for details on how to verify the operator and driver image signatures.
        displayName: BeeGFS CSI Driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CSI
        - BeeGFS
        - ThinkParQ
        - HPC
        - AI
        - High-Performance Computing
        links:
        - name: BeeGFS introduction
          url: https://www.beegfs.io/c/
        - name: BeeGFS documentation
          url: https://doc.beegfs.io/latest/index.html
        - name: BeeGFS CSI driver source code
          url: https://github.com/ThinkParQ/beegfs-csi-driver
        - name: BeeGFS blog posts
          url: https://www.beegfs.io/c/blog/
        maintainers:
        - email: csi-driver@beegfs.io
          name: ThinkParQ
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: ThinkParQ
          url: https://www.thinkparq.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - ghcr.io/thinkparq/beegfs-csi-driver-operator:v1.7.0
        version: 1.7.0
      entries:
      - name: beegfs-csi-driver-operator.v1.7.0
        version: 1.7.0
      - name: beegfs-csi-driver-operator.v1.6.0
        version: 1.6.0
      - name: beegfs-csi-driver-operator.v1.5.0
        version: 1.5.0
      - name: beegfs-csi-driver-operator.v1.4.0
        version: 1.4.0
      - name: beegfs-csi-driver-operator.v1.3.0
        version: 1.3.0
      - name: beegfs-csi-driver-operator.v1.2.2
        version: 1.2.2
      - name: beegfs-csi-driver-operator.v1.2.1
        version: 1.2.1
      - name: beegfs-csi-driver-operator.v1.2.0
        version: 1.2.0
      name: stable
    defaultChannel: stable
    packageName: beegfs-csi-driver-operator
    provider:
      name: ThinkParQ
      url: https://www.thinkparq.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stackwatch
      provider-url: www.kubecost.com
    name: kubecost
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubecost.v2.2.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kubecost.com/v1alpha1",
                "kind": "CostAnalyzer",
                "metadata": {
                  "name": "costanalyzer-sample"
                },
                "spec": {
                  "global": {
                    "prometheus": {
                      "enabled": true,
                      "fqdn": "http://cost-analyzer-prometheus-server.default.svc"
                    },
                    "grafana": {
                      "enabled": false,
                      "domainName": "cost-analyzer-grafana.default.svc",
                      "scheme": "http",
                      "proxy": false
                    },
                    "gcpstore": {
                      "enabled": false
                    },
                    "gmp": {
                      "enabled": false,
                      "prometheusServerEndpoint": "http://localhost:8085/",
                      "gmpProxy": {
                        "enabled": false,
                        "image": "gke.gcr.io/prometheus-engine/frontend:v0.4.1-gke.0",
                        "imagePullPolicy": "Always",
                        "name": "gmp-proxy",
                        "port": 8085,
                        "projectId": "YOUR_PROJECT_ID"
                      }
                    },
                    "amp": {
                      "enabled": false,
                      "prometheusServerEndpoint": "http://localhost:8005/workspaces/<workspaceId>/",
                      "remoteWriteService": "https://aps-workspaces.us-west-2.amazonaws.com/workspaces/<workspaceId>/api/v1/remote_write",
                      "sigv4": {
                        "region": "us-west-2"
                      }
                    },
                    "mimirProxy": {
                      "enabled": false,
                      "name": "mimir-proxy",
                      "image": "nginxinc/nginx-unprivileged",
                      "port": 8085,
                      "mimirEndpoint": "$mimir_endpoint",
                      "orgIdentifier": "$your_tenant_ID"
                    },
                    "notifications": {
                      "alertmanager": {
                        "enabled": false,
                        "fqdn": "http://cost-analyzer-prometheus-server.default.svc"
                      }
                    },
                    "savedReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Saved Report 0",
                          "window": "today",
                          "aggregateBy": "namespace",
                          "chartDisplay": "category",
                          "idle": "separate",
                          "rate": "cumulative",
                          "accumulate": false,
                          "filters": [
                            {
                              "key": "cluster",
                              "operator": ":",
                              "value": "dev"
                            }
                          ]
                        },
                        {
                          "title": "Example Saved Report 1",
                          "window": "month",
                          "aggregateBy": "controllerKind",
                          "chartDisplay": "category",
                          "idle": "share",
                          "rate": "monthly",
                          "accumulate": false,
                          "filters": [
                            {
                              "key": "namespace",
                              "operator": "!:",
                              "value": "kubecost"
                            }
                          ]
                        },
                        {
                          "title": "Example Saved Report 2",
                          "window": "2020-11-11T00:00:00Z,2020-12-09T23:59:59Z",
                          "aggregateBy": "service",
                          "chartDisplay": "category",
                          "idle": "hide",
                          "rate": "daily",
                          "accumulate": true,
                          "filters": []
                        }
                      ]
                    },
                    "assetReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Asset Report 0",
                          "window": "today",
                          "aggregateBy": "type",
                          "accumulate": false,
                          "filters": [
                            {
                              "property": "cluster",
                              "value": "cluster-one"
                            }
                          ]
                        }
                      ]
                    },
                    "advancedReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Advanced Report 0",
                          "window": "7d",
                          "aggregateBy": "namespace",
                          "filters": [
                            {
                              "key": "cluster",
                              "operator": ":",
                              "value": "dev"
                            }
                          ],
                          "cloudBreakdown": "service",
                          "cloudJoin": "label:kubernetes_namespace"
                        }
                      ]
                    },
                    "cloudCostReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Cloud Cost Report 0",
                          "window": "today",
                          "aggregateBy": "service",
                          "accumulate": false
                        }
                      ]
                    },
                    "podAnnotations": {},
                    "additionalLabels": {},
                    "securityContext": {
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      },
                      "fsGroup": 1001,
                      "runAsGroup": 1001,
                      "runAsUser": 1001,
                      "fsGroupChangePolicy": "OnRootMismatch"
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "privileged": false,
                      "readOnlyRootFilesystem": true,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "platforms": {
                      "openshift": {
                        "enabled": true,
                        "route": {
                          "enabled": false,
                          "annotations": {}
                        },
                        "scc": {
                          "nodeExporter": false,
                          "networkCosts": false
                        },
                        "securityContext": {
                          "runAsNonRoot": true,
                          "seccompProfile": {
                            "type": "RuntimeDefault"
                          }
                        }
                      },
                      "cicd": {
                        "enabled": true,
                        "skipSanityChecks": true
                      }
                    }
                  },
                  "upgrade": {
                    "toV2": false
                  },
                  "kubecostToken": null,
                  "pricingCsv": {
                    "enabled": false,
                    "location": {
                      "provider": "AWS",
                      "region": "us-east-1",
                      "URI": "s3://kc-csv-test/pricing_schema.csv",
                      "csvAccessCredentials": "pricing-schema-access-secret"
                    }
                  },
                  "saml": {
                    "enabled": false,
                    "rbac": {
                      "enabled": false
                    }
                  },
                  "oidc": {
                    "enabled": false,
                    "clientID": "",
                    "clientSecret": "",
                    "existingCustomSecret": {
                      "enabled": false,
                      "name": ""
                    },
                    "skipOnlineTokenValidation": false,
                    "rbac": {
                      "enabled": false
                    }
                  },
                  "systemProxy": {
                    "enabled": false,
                    "httpProxyUrl": "",
                    "httpsProxyUrl": "",
                    "noProxy": ""
                  },
                  "kubecostFrontend": {
                    "enabled": true,
                    "deployMethod": "singlepod",
                    "haReplicas": 2,
                    "image": "gcr.io/kubecost1/frontend",
                    "imagePullPolicy": "Always",
                    "resources": {
                      "requests": {
                        "cpu": "10m",
                        "memory": "55Mi"
                      }
                    },
                    "deploymentStrategy": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 1,
                      "periodSeconds": 5,
                      "failureThreshold": 6
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 1,
                      "periodSeconds": 5,
                      "failureThreshold": 6
                    },
                    "ipv6": {
                      "enabled": true
                    },
                    "useDefaultFqdn": false
                  },
                  "kubecostMetrics": {
                    "exporter": {
                      "enabled": false,
                      "port": 9005,
                      "prometheusScrape": true,
                      "resources": {},
                      "tolerations": [],
                      "affinity": {},
                      "service": {
                        "annotations": {}
                      },
                      "serviceMonitor": {
                        "enabled": false,
                        "additionalLabels": {},
                        "metricRelabelings": [],
                        "relabelings": []
                      },
                      "priorityClassName": "",
                      "additionalLabels": {},
                      "nodeSelector": {},
                      "extraArgs": []
                    }
                  },
                  "sigV4Proxy": {
                    "image": "public.ecr.aws/aws-observability/aws-sigv4-proxy:latest",
                    "imagePullPolicy": "Always",
                    "name": "aps",
                    "port": 8005,
                    "region": "us-west-2",
                    "host": "aps-workspaces.us-west-2.amazonaws.com",
                    "extraEnv": null,
                    "resources": {}
                  },
                  "kubecostModel": {
                    "image": "gcr.io/kubecost1/cost-model",
                    "imagePullPolicy": "Always",
                    "outOfClusterPromMetricsEnabled": false,
                    "warmCache": false,
                    "etl": true,
                    "etlFileStoreEnabled": true,
                    "etlDailyStoreDurationDays": 91,
                    "etlHourlyStoreDurationHours": 49,
                    "etlWeeklyStoreDurationWeeks": 53,
                    "etlReadOnlyMode": false,
                    "plugins": {
                      "enabled": false,
                      "install": {
                        "enabled": false,
                        "fullImageName": "curlimages/curl:latest",
                        "securityContext": {
                          "allowPrivilegeEscalation": false,
                          "seccompProfile": {
                            "type": "RuntimeDefault"
                          },
                          "capabilities": {
                            "drop": [
                              "ALL"
                            ]
                          },
                          "readOnlyRootFilesystem": true,
                          "runAsNonRoot": true,
                          "runAsUser": 1001
                        }
                      },
                      "folder": "/opt/opencost/plugin",
                      "enabledPlugins": [],
                      "configSecret": "kubecost-plugin-secret"
                    },
                    "cloudCost": {
                      "labelList": {
                        "IsIncludeList": false,
                        "labels": ""
                      },
                      "topNItems": 1000
                    },
                    "allocation": null,
                    "containerStatsEnabled": true,
                    "maxQueryConcurrency": 5,
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "55Mi"
                      }
                    },
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "extraArgs": [],
                    "ingress": {
                      "enabled": false,
                      "labels": null,
                      "annotations": null,
                      "paths": [
                        "/"
                      ],
                      "pathType": "ImplementationSpecific",
                      "hosts": [
                        "cost-analyzer-model.local"
                      ],
                      "tls": []
                    },
                    "utcOffset": "+00:00",
                    "extraPorts": []
                  },
                  "etlUtils": {
                    "enabled": false,
                    "fullImageName": null,
                    "resources": {},
                    "env": {},
                    "nodeSelector": {},
                    "tolerations": {},
                    "affinity": {}
                  },
                  "ingress": {
                    "enabled": false,
                    "labels": null,
                    "annotations": null,
                    "paths": [
                      "/"
                    ],
                    "pathType": "ImplementationSpecific",
                    "hosts": [
                      "cost-analyzer.local"
                    ],
                    "tls": []
                  },
                  "nodeSelector": {},
                  "tolerations": [],
                  "affinity": {},
                  "topologySpreadConstraints": [],
                  "priority": {
                    "enabled": false,
                    "name": ""
                  },
                  "networkPolicy": {
                    "enabled": false,
                    "denyEgress": true,
                    "sameNamespace": true,
                    "costAnalyzer": {
                      "enabled": false,
                      "annotations": {},
                      "additionalLabels": {}
                    }
                  },
                  "extraVolumes": [],
                  "extraVolumeMounts": [],
                  "persistentVolume": {
                    "size": "32Gi",
                    "dbSize": "32.0Gi",
                    "enabled": true,
                    "labels": {},
                    "annotations": {},
                    "dbPVEnabled": false
                  },
                  "service": {
                    "type": "ClusterIP",
                    "port": 9090,
                    "targetPort": 9090,
                    "nodePort": {},
                    "labels": {},
                    "annotations": {},
                    "sessionAffinity": {
                      "enabled": false,
                      "timeoutSeconds": 10800
                    }
                  },
                  "prometheus": {
                    "rbac": {
                      "create": true
                    },
                    "serviceAccounts": {
                      "alertmanager": {
                        "create": true,
                        "name": null
                      },
                      "nodeExporter": {
                        "create": true,
                        "name": null
                      },
                      "pushgateway": {
                        "create": true,
                        "name": null
                      },
                      "server": {
                        "create": true,
                        "name": null,
                        "annotations": {}
                      }
                    },
                    "imagePullSecrets": null,
                    "extraScrapeConfigs": "- job_name: kubecost\n  honor_labels: true\n  scrape_interval: 1m\n  scrape_timeout: 60s\n  metrics_path: /metrics\n  scheme: http\n  dns_sd_configs:\n  - names:\n    - {{ template \"cost-analyzer.serviceName\" . }}\n    type: 'A'\n    port: 9003\n- job_name: kubecost-networking\n  kubernetes_sd_configs:\n    - role: pod\n  relabel_configs:\n  # Scrape only the the targets matching the following metadata\n    - source_labels: [__meta_kubernetes_pod_label_app_kubernetes_io_instance]\n      action: keep\n      regex:  kubecost\n    - source_labels: [__meta_kubernetes_pod_label_app_kubernetes_io_name]\n      action: keep\n      regex:  network-costs\n",
                    "server": {
                      "enabled": true,
                      "name": "server",
                      "sidecarContainers": null,
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "image": {
                        "repository": "quay.io/prometheus/prometheus",
                        "tag": "v2.50.1",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "prefixURL": "",
                      "baseURL": "",
                      "env": [],
                      "extraFlags": [
                        "web.enable-lifecycle"
                      ],
                      "configPath": "/etc/config/prometheus.yml",
                      "global": {
                        "scrape_interval": "1m",
                        "scrape_timeout": "60s",
                        "evaluation_interval": "1m",
                        "external_labels": {
                          "cluster_id": "cluster-one"
                        }
                      },
                      "remoteWrite": {},
                      "remoteRead": {},
                      "extraArgs": {
                        "query.max-concurrency": 1,
                        "query.max-samples": 100000000
                      },
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "extraHostPathMounts": [],
                      "extraConfigmapMounts": [],
                      "extraSecretMounts": [],
                      "configMapOverrideName": "",
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "extraLabels": {},
                        "hosts": [],
                        "pathType": "Prefix",
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "affinity": {},
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "32Gi",
                        "subPath": ""
                      },
                      "emptyDir": {
                        "sizeLimit": ""
                      },
                      "podAnnotations": {},
                      "deploymentAnnotations": {},
                      "podLabels": {},
                      "alertmanagers": [],
                      "replicaCount": 1,
                      "statefulSet": {
                        "enabled": false,
                        "annotations": {},
                        "labels": {},
                        "podManagementPolicy": "OrderedReady",
                        "headless": {
                          "annotations": {},
                          "labels": {},
                          "servicePort": 80
                        }
                      },
                      "readinessProbeInitialDelay": 5,
                      "readinessProbeTimeout": 3,
                      "readinessProbeFailureThreshold": 3,
                      "readinessProbeSuccessThreshold": 1,
                      "livenessProbeInitialDelay": 5,
                      "livenessProbeTimeout": 3,
                      "livenessProbeFailureThreshold": 3,
                      "livenessProbeSuccessThreshold": 1,
                      "resources": {},
                      "verticalAutoscaler": {
                        "enabled": false
                      },
                      "securityContext": {},
                      "containerSecurityContext": {},
                      "service": {
                        "annotations": {},
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 80,
                        "sessionAffinity": "None",
                        "type": "ClusterIP",
                        "gRPC": {
                          "enabled": false,
                          "servicePort": 10901
                        },
                        "statefulsetReplica": {
                          "enabled": false,
                          "replica": 0
                        }
                      },
                      "terminationGracePeriodSeconds": 300,
                      "retention": "97h"
                    },
                    "alertmanager": {
                      "enabled": false,
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "name": "alertmanager",
                      "image": {
                        "repository": "quay.io/prometheus/alertmanager",
                        "tag": "v0.27.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "extraArgs": {},
                      "prefixURL": "",
                      "baseURL": "http://localhost:9093",
                      "extraEnv": {},
                      "extraSecretMounts": [],
                      "configMapOverrideName": "",
                      "configFromSecret": "",
                      "configFileName": "alertmanager.yml",
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "extraLabels": {},
                        "hosts": [],
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "affinity": {},
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "2Gi",
                        "subPath": ""
                      },
                      "podAnnotations": {},
                      "podLabels": {},
                      "replicaCount": 1,
                      "statefulSet": {
                        "enabled": false,
                        "podManagementPolicy": "OrderedReady",
                        "headless": {
                          "annotations": {},
                          "labels": {},
                          "servicePort": 80
                        }
                      },
                      "resources": {},
                      "securityContext": {
                        "runAsUser": 1001,
                        "runAsNonRoot": true,
                        "runAsGroup": 1001,
                        "fsGroup": 1001
                      },
                      "service": {
                        "annotations": {},
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 80,
                        "sessionAffinity": "None",
                        "type": "ClusterIP"
                      }
                    },
                    "alertmanagerFiles": {
                      "alertmanager.yml": {
                        "global": {},
                        "receivers": [
                          {
                            "name": "default-receiver"
                          }
                        ],
                        "route": {
                          "group_wait": "10s",
                          "group_interval": "5m",
                          "receiver": "default-receiver",
                          "repeat_interval": "3h"
                        }
                      }
                    },
                    "configmapReload": {
                      "prometheus": {
                        "enabled": false,
                        "name": "configmap-reload",
                        "image": {
                          "repository": "quay.io/prometheus-operator/prometheus-config-reloader",
                          "tag": "v0.72.0",
                          "pullPolicy": "IfNotPresent"
                        },
                        "extraArgs": {},
                        "extraVolumeDirs": [],
                        "extraConfigmapMounts": [],
                        "resources": {},
                        "containerSecurityContext": {}
                      },
                      "alertmanager": {
                        "enabled": false,
                        "name": "configmap-reload",
                        "image": {
                          "repository": "quay.io/prometheus-operator/prometheus-config-reloader",
                          "tag": "v0.72.0",
                          "pullPolicy": "IfNotPresent"
                        },
                        "extraArgs": {},
                        "extraVolumeDirs": [],
                        "extraConfigmapMounts": [],
                        "resources": {}
                      }
                    },
                    "nodeExporter": {
                      "enabled": false,
                      "hostNetwork": true,
                      "hostPID": true,
                      "dnsPolicy": "ClusterFirstWithHostNet",
                      "name": "node-exporter",
                      "image": {
                        "repository": "prom/node-exporter",
                        "tag": "v1.7.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "extraArgs": {},
                      "extraHostPathMounts": [],
                      "extraConfigmapMounts": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "deploymentAnnotations": {},
                      "pod": {
                        "labels": {}
                      },
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "resources": {},
                      "securityContext": {},
                      "service": {
                        "annotations": {
                          "prometheus.io/scrape": "true"
                        },
                        "labels": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "hostPort": 9100,
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 9100,
                        "type": "ClusterIP"
                      }
                    },
                    "pushgateway": {
                      "enabled": false,
                      "name": "pushgateway",
                      "image": {
                        "repository": "prom/pushgateway",
                        "tag": "v1.6.2",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "extraArgs": {},
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "hosts": [],
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "replicaCount": 1,
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "resources": {},
                      "securityContext": {
                        "runAsUser": 1001,
                        "runAsNonRoot": true
                      },
                      "service": {
                        "annotations": {
                          "prometheus.io/probe": "pushgateway"
                        },
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 9091,
                        "type": "ClusterIP"
                      },
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "2Gi",
                        "subPath": ""
                      }
                    },
                    "serverFiles": {
                      "alerting_rules.yml": {},
                      "alerts": {},
                      "recording_rules.yml": {},
                      "prometheus.yml": {
                        "rule_files": [
                          "/etc/config/recording_rules.yml",
                          "/etc/config/alerting_rules.yml",
                          "/etc/config/rules",
                          "/etc/config/alerts"
                        ],
                        "scrape_configs": [
                          {
                            "job_name": "prometheus",
                            "static_configs": [
                              {
                                "targets": [
                                  "localhost:9090"
                                ]
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-nodes-cadvisor",
                            "scheme": "https",
                            "tls_config": {
                              "ca_file": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                              "insecure_skip_verify": true
                            },
                            "bearer_token_file": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                            "kubernetes_sd_configs": [
                              {
                                "role": "node"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_node_label_(.+)"
                              },
                              {
                                "target_label": "__address__",
                                "replacement": "kubernetes.default.svc:443"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_node_name"
                                ],
                                "regex": "(.+)",
                                "target_label": "__metrics_path__",
                                "replacement": "/api/v1/nodes/$1/proxy/metrics/cadvisor"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(container_cpu_usage_seconds_total|container_memory_working_set_bytes|container_network_receive_errors_total|container_network_transmit_errors_total|container_network_receive_packets_dropped_total|container_network_transmit_packets_dropped_total|container_memory_usage_bytes|container_cpu_cfs_throttled_periods_total|container_cpu_cfs_periods_total|container_fs_usage_bytes|container_fs_limit_bytes|container_cpu_cfs_periods_total|container_fs_inodes_free|container_fs_inodes_total|container_fs_usage_bytes|container_fs_limit_bytes|container_cpu_cfs_throttled_periods_total|container_cpu_cfs_periods_total|container_network_receive_bytes_total|container_network_transmit_bytes_total|container_fs_inodes_free|container_fs_inodes_total|container_fs_usage_bytes|container_fs_limit_bytes|container_spec_cpu_shares|container_spec_memory_limit_bytes|container_network_receive_bytes_total|container_network_transmit_bytes_total|container_fs_reads_bytes_total|container_network_receive_bytes_total|container_fs_writes_bytes_total|container_fs_reads_bytes_total|cadvisor_version_info|kubecost_pv_info)",
                                "action": "keep"
                              },
                              {
                                "source_labels": [
                                  "container"
                                ],
                                "target_label": "container_name",
                                "regex": "(.+)",
                                "action": "replace"
                              },
                              {
                                "source_labels": [
                                  "pod"
                                ],
                                "target_label": "pod_name",
                                "regex": "(.+)",
                                "action": "replace"
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-nodes",
                            "scheme": "https",
                            "tls_config": {
                              "ca_file": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                              "insecure_skip_verify": true
                            },
                            "bearer_token_file": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                            "kubernetes_sd_configs": [
                              {
                                "role": "node"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_node_label_(.+)"
                              },
                              {
                                "target_label": "__address__",
                                "replacement": "kubernetes.default.svc:443"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_node_name"
                                ],
                                "regex": "(.+)",
                                "target_label": "__metrics_path__",
                                "replacement": "/api/v1/nodes/$1/proxy/metrics"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(kubelet_volume_stats_used_bytes)",
                                "action": "keep"
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-service-endpoints",
                            "kubernetes_sd_configs": [
                              {
                                "role": "endpoints"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_scrape"
                                ],
                                "action": "keep",
                                "regex": true
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_endpoints_name"
                                ],
                                "action": "keep",
                                "regex": "(.*node-exporter|kubecost-network-costs)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_scheme"
                                ],
                                "action": "replace",
                                "target_label": "__scheme__",
                                "regex": "(https?)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_path"
                                ],
                                "action": "replace",
                                "target_label": "__metrics_path__",
                                "regex": "(.+)"
                              },
                              {
                                "source_labels": [
                                  "__address__",
                                  "__meta_kubernetes_service_annotation_prometheus_io_port"
                                ],
                                "action": "replace",
                                "target_label": "__address__",
                                "regex": "([^:]+)(?::\\d+)?;(\\d+)",
                                "replacement": "$1:$2"
                              },
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_service_label_(.+)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_namespace"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_namespace"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_name"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_name"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_pod_node_name"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_node"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(container_cpu_allocation|container_cpu_usage_seconds_total|container_fs_limit_bytes|container_fs_writes_bytes_total|container_gpu_allocation|container_memory_allocation_bytes|container_memory_usage_bytes|container_memory_working_set_bytes|container_network_receive_bytes_total|container_network_transmit_bytes_total|DCGM_FI_DEV_GPU_UTIL|deployment_match_labels|kube_daemonset_status_desired_number_scheduled|kube_daemonset_status_number_ready|kube_deployment_spec_replicas|kube_deployment_status_replicas|kube_deployment_status_replicas_available|kube_job_status_failed|kube_namespace_annotations|kube_namespace_labels|kube_node_info|kube_node_labels|kube_node_status_allocatable|kube_node_status_allocatable_cpu_cores|kube_node_status_allocatable_memory_bytes|kube_node_status_capacity|kube_node_status_capacity_cpu_cores|kube_node_status_capacity_memory_bytes|kube_node_status_condition|kube_persistentvolume_capacity_bytes|kube_persistentvolume_status_phase|kube_persistentvolumeclaim_info|kube_persistentvolumeclaim_resource_requests_storage_bytes|kube_pod_container_info|kube_pod_container_resource_limits|kube_pod_container_resource_limits_cpu_cores|kube_pod_container_resource_limits_memory_bytes|kube_pod_container_resource_requests|kube_pod_container_resource_requests_cpu_cores|kube_pod_container_resource_requests_memory_bytes|kube_pod_container_status_restarts_total|kube_pod_container_status_running|kube_pod_container_status_terminated_reason|kube_pod_labels|kube_pod_owner|kube_pod_status_phase|kube_replicaset_owner|kube_statefulset_replicas|kube_statefulset_status_replicas|kubecost_cluster_info|kubecost_cluster_management_cost|kubecost_cluster_memory_working_set_bytes|kubecost_load_balancer_cost|kubecost_network_internet_egress_cost|kubecost_network_region_egress_cost|kubecost_network_zone_egress_cost|kubecost_node_is_spot|kubecost_pod_network_egress_bytes_total|node_cpu_hourly_cost|node_cpu_seconds_total|node_disk_reads_completed|node_disk_reads_completed_total|node_disk_writes_completed|node_disk_writes_completed_total|node_filesystem_device_error|node_gpu_count|node_gpu_hourly_cost|node_memory_Buffers_bytes|node_memory_Cached_bytes|node_memory_MemAvailable_bytes|node_memory_MemFree_bytes|node_memory_MemTotal_bytes|node_network_transmit_bytes_total|node_ram_hourly_cost|node_total_hourly_cost|pod_pvc_allocation|pv_hourly_cost|service_selector_labels|statefulSet_match_labels|kubecost_pv_info|up)",
                                "action": "keep"
                              }
                            ]
                          }
                        ]
                      },
                      "rules": {
                        "groups": [
                          {
                            "name": "CPU",
                            "rules": [
                              {
                                "expr": "sum(rate(container_cpu_usage_seconds_total{container!=\"\"}[5m]))",
                                "record": "cluster:cpu_usage:rate5m"
                              },
                              {
                                "expr": "rate(container_cpu_usage_seconds_total{container!=\"\"}[5m])",
                                "record": "cluster:cpu_usage_nosum:rate5m"
                              },
                              {
                                "expr": "avg(irate(container_cpu_usage_seconds_total{container!=\"POD\", container!=\"\"}[5m])) by (container,pod,namespace)",
                                "record": "kubecost_container_cpu_usage_irate"
                              },
                              {
                                "expr": "sum(container_memory_working_set_bytes{container!=\"POD\",container!=\"\"}) by (container,pod,namespace)",
                                "record": "kubecost_container_memory_working_set_bytes"
                              },
                              {
                                "expr": "sum(container_memory_working_set_bytes{container!=\"POD\",container!=\"\"})",
                                "record": "kubecost_cluster_memory_working_set_bytes"
                              }
                            ]
                          },
                          {
                            "name": "Savings",
                            "rules": [
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind!=\"DaemonSet\"}) by (pod) * sum(container_cpu_allocation) by (pod))",
                                "record": "kubecost_savings_cpu_allocation",
                                "labels": {
                                  "daemonset": "false"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind=\"DaemonSet\"}) by (pod) * sum(container_cpu_allocation) by (pod)) / sum(kube_node_info)",
                                "record": "kubecost_savings_cpu_allocation",
                                "labels": {
                                  "daemonset": "true"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind!=\"DaemonSet\"}) by (pod) * sum(container_memory_allocation_bytes) by (pod))",
                                "record": "kubecost_savings_memory_allocation_bytes",
                                "labels": {
                                  "daemonset": "false"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind=\"DaemonSet\"}) by (pod) * sum(container_memory_allocation_bytes) by (pod)) / sum(kube_node_info)",
                                "record": "kubecost_savings_memory_allocation_bytes",
                                "labels": {
                                  "daemonset": "true"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    "alertRelabelConfigs": null,
                    "networkPolicy": {
                      "enabled": false
                    }
                  },
                  "networkCosts": {
                    "enabled": false,
                    "image": {
                      "repository": "gcr.io/kubecost1/kubecost-network-costs",
                      "tag": "v0.17.3"
                    },
                    "imagePullPolicy": "Always",
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "prometheusScrape": false,
                    "trafficLogging": true,
                    "logLevel": "info",
                    "port": 3001,
                    "resources": {
                      "limits": {
                        "cpu": "500m"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "20Mi"
                      }
                    },
                    "extraArgs": [],
                    "config": {
                      "destinations": {
                        "in-zone": [
                          "127.0.0.0/8",
                          "169.254.0.0/16",
                          "10.0.0.0/8",
                          "172.16.0.0/12",
                          "192.168.0.0/16"
                        ],
                        "in-region": [],
                        "cross-region": [],
                        "internet": [],
                        "direct-classification": []
                      },
                      "services": {
                        "google-cloud-services": false,
                        "amazon-web-services": false,
                        "azure-cloud-services": false
                      }
                    },
                    "tolerations": [],
                    "affinity": {},
                    "service": {
                      "annotations": {},
                      "labels": {}
                    },
                    "priorityClassName": "",
                    "podMonitor": {
                      "enabled": false,
                      "additionalLabels": {}
                    },
                    "additionalLabels": {},
                    "nodeSelector": {},
                    "annotations": {},
                    "healthCheckProbes": {},
                    "additionalSecurityContext": {}
                  },
                  "kubecostDeployment": {
                    "replicas": 1,
                    "labels": {},
                    "annotations": {}
                  },
                  "forecasting": {
                    "enabled": true,
                    "fullImageName": "gcr.io/kubecost1/kubecost-modeling:v0.1.6",
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "limits": {
                        "cpu": "1500m",
                        "memory": "1Gi"
                      }
                    },
                    "env": {
                      "GUNICORN_CMD_ARGS": "--log-level info -t 1200"
                    },
                    "priority": {
                      "enabled": false,
                      "name": ""
                    },
                    "nodeSelector": {},
                    "tolerations": {},
                    "affinity": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    }
                  },
                  "kubecostAggregator": {
                    "deployMethod": "singlepod",
                    "enabled": false,
                    "replicas": 1,
                    "stagingEmptyDirSizeLimit": "2Gi",
                    "numDBCopyPartitions": 25,
                    "env": {
                      "LOG_LEVEL": "info",
                      "DB_READ_THREADS": "1",
                      "DB_WRITE_THREADS": "1",
                      "DB_CONCURRENT_INGESTION_COUNT": "3"
                    },
                    "persistentConfigsStorage": {
                      "storageClass": "",
                      "storageRequest": "1Gi"
                    },
                    "aggregatorDbStorage": {
                      "storageClass": "",
                      "storageRequest": "128Gi"
                    },
                    "resources": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "cloudCost": {
                      "enabled": false,
                      "resources": {},
                      "readinessProbe": {
                        "enabled": true,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 10,
                        "failureThreshold": 200
                      }
                    },
                    "jaeger": {
                      "enabled": false,
                      "image": "jaegertracing/all-in-one",
                      "imageVersion": "latest"
                    }
                  },
                  "diagnostics": {
                    "enabled": true,
                    "primary": {
                      "enabled": false,
                      "retention": "7d",
                      "readonly": false
                    },
                    "pollingInterval": "300s",
                    "keepDiagnosticHistory": false,
                    "collectHelmValues": false,
                    "deployment": {
                      "enabled": false,
                      "resources": {
                        "requests": {
                          "cpu": "10m",
                          "memory": "20Mi"
                        }
                      },
                      "env": {},
                      "labels": {},
                      "securityContext": {},
                      "containerSecurityContext": {},
                      "nodeSelector": {},
                      "tolerations": {},
                      "affinity": {}
                    }
                  },
                  "clusterController": {
                    "enabled": false,
                    "image": {
                      "repository": "gcr.io/kubecost1/cluster-controller",
                      "tag": "v0.16.0"
                    },
                    "imagePullPolicy": "Always",
                    "priorityClassName": "",
                    "tolerations": [],
                    "actionConfigs": {
                      "clusterTurndown": [],
                      "namespaceTurndown": null,
                      "clusterRightsize": null,
                      "containerRightsize": null
                    },
                    "kubescaler": {
                      "defaultResizeAll": false
                    },
                    "namespaceTurndown": {
                      "rbac": {
                        "enabled": true
                      }
                    }
                  },
                  "reporting": {
                    "logCollection": true,
                    "productAnalytics": true,
                    "errorReporting": true,
                    "valuesReporting": true
                  },
                  "serviceMonitor": {
                    "enabled": false,
                    "additionalLabels": {},
                    "metricRelabelings": [],
                    "relabelings": [],
                    "networkCosts": {
                      "enabled": false,
                      "scrapeTimeout": "10s",
                      "additionalLabels": {},
                      "metricRelabelings": [],
                      "relabelings": []
                    }
                  },
                  "prometheusRule": {
                    "enabled": false,
                    "additionalLabels": {}
                  },
                  "supportNFS": false,
                  "initChownDataImage": "busybox",
                  "initChownData": {
                    "resources": {}
                  },
                  "grafana": {
                    "rbac": {
                      "create": true
                    },
                    "serviceAccount": {
                      "create": true,
                      "name": ""
                    },
                    "replicas": 1,
                    "deploymentStrategy": "RollingUpdate",
                    "readinessProbe": {
                      "httpGet": {
                        "path": "/api/health",
                        "port": 3000
                      }
                    },
                    "livenessProbe": {
                      "httpGet": {
                        "path": "/api/health",
                        "port": 3000
                      },
                      "initialDelaySeconds": 60,
                      "timeoutSeconds": 30,
                      "failureThreshold": 10
                    },
                    "image": {
                      "repository": "grafana/grafana",
                      "tag": "10.3.4",
                      "pullPolicy": "IfNotPresent"
                    },
                    "securityContext": {},
                    "priorityClassName": "",
                    "downloadDashboardsImage": {
                      "repository": "curlimages/curl",
                      "tag": "latest",
                      "pullPolicy": "IfNotPresent"
                    },
                    "podAnnotations": {},
                    "annotations": {},
                    "service": {
                      "type": "ClusterIP",
                      "port": 80,
                      "annotations": {},
                      "labels": {}
                    },
                    "ingress": {
                      "enabled": false,
                      "annotations": {},
                      "labels": {},
                      "path": "/",
                      "pathType": "Prefix",
                      "hosts": [
                        "chart-example.local"
                      ],
                      "tls": []
                    },
                    "resources": {},
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "persistence": {
                      "enabled": false
                    },
                    "adminUser": "admin",
                    "adminPassword": "strongpassword",
                    "env": {},
                    "envFromSecret": "",
                    "extraSecretMounts": [],
                    "plugins": [],
                    "dashboardProviders": {},
                    "dashboards": {},
                    "dashboardsConfigMaps": {},
                    "ldap": {
                      "existingSecret": "",
                      "config": ""
                    },
                    "smtp": {
                      "existingSecret": ""
                    },
                    "sidecar": {
                      "image": {
                        "repository": "kiwigrid/k8s-sidecar",
                        "tag": "1.26.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "resources": {},
                      "dashboards": {
                        "enabled": true,
                        "label": "grafana_dashboard",
                        "labelValue": "1",
                        "annotations": {},
                        "error_throttle_sleep": 0,
                        "folder": "/tmp/dashboards"
                      },
                      "datasources": {
                        "enabled": false,
                        "error_throttle_sleep": 0,
                        "label": "grafana_datasource"
                      }
                    },
                    "grafana.ini": {
                      "server": {
                        "serve_from_sub_path": false,
                        "root_url": "%(protocol)s://%(domain)s:%(http_port)s/grafana"
                      },
                      "paths": {
                        "data": "/var/lib/grafana/data",
                        "logs": "/var/log/grafana",
                        "plugins": "/var/lib/grafana/plugins",
                        "provisioning": "/etc/grafana/provisioning"
                      },
                      "analytics": {
                        "check_for_updates": true
                      },
                      "log": {
                        "mode": "console"
                      },
                      "grafana_net": {
                        "url": "https://grafana.net"
                      },
                      "auth.anonymous": {
                        "enabled": true,
                        "org_role": "Editor",
                        "org_name": "Main Org."
                      }
                    }
                  },
                  "serviceAccount": {
                    "create": true,
                    "annotations": {}
                  },
                  "awsstore": {
                    "useAwsStore": false,
                    "imageNameAndVersion": "gcr.io/kubecost1/awsstore:latest",
                    "createServiceAccount": false,
                    "priorityClassName": "",
                    "nodeSelector": {},
                    "annotations": {}
                  },
                  "federatedETL": {
                    "agentOnly": false,
                    "federatedCluster": false,
                    "redirectS3Backup": false,
                    "useMultiClusterDB": false
                  },
                  "kubecostAdmissionController": {
                    "enabled": false,
                    "secretName": "webhook-server-tls",
                    "caBundle": "${CA_BUNDLE}"
                  },
                  "costEventsAudit": {
                    "enabled": false
                  },
                  "extraObjects": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: gcr.io/kubecost1/operator/kubecost-operator-controller:v2.2.4
          createdAt: "2024-05-19T20:46:35Z"
          description: An operator that deploys Kubecost 2.2.4 via the Helm chart.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Stackwatch
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The CostAnalyzer API describes a deployment of Kubecost and
              its dependencies.
            displayName: CostAnalyzer
            kind: CostAnalyzer
            name: costanalyzers.charts.kubecost.com
            version: v1alpha1
        description: |-
          Kubecost provides real-time cost visibility and insights for teams using Kubernetes, helping you continuously reduce your cloud costs.

          This OpenShift operator deploys Kubecost from the official Helm chart.

          Kubecost allows you to:

          * See allocated spend across all native Kubernetes concepts, so you can provide your teams with transparent, accurate cost data reconciled with your actual cloud bill.
          * Join in-cluster costs like CPU and memory with out-of-cluster spend from cloud infrastructure services for complete cost visibility—all services from AWS, GCP, and Microsoft Azure are supported.
          * Gain insights you can use to save 30-50% or more on your infrastructure spend, without exposing your private information—your data never gets shared externally, even with us.
          * Have real-time alerts and recurring reports, empowering teams to take control of their Kubernetes-enabled infrastructure, stay within budgeted limits, and address monitoring interruptions immediately.

          Visit the Kubecost website for more details including features, pricing, documentation, and more at www.kubecost.com.
        displayName: Kubecost
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubecost
        - cost
        - kubernetes
        links:
        - name: Kubecost
          url: https://www.kubecost.com
        maintainers:
        - email: czoller@kubecost.com
          name: Chip Zoller
        maturity: alpha
        provider:
          name: Stackwatch
          url: www.kubecost.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - gcr.io/kubecost1/operator/kubecost-operator-controller:v2.2.4
        version: 2.2.4
      entries:
      - name: kubecost.v2.2.4
        version: 2.2.4
      - name: kubecost.v2.1.0
        version: 2.1.0
      - name: kubecost-operator.v1.107.1
        version: 1.107.1
      - name: kubecost-operator.v1.106.2
        version: 1.106.2
      - name: kubecost-operator.v1.106.0
        version: 1.106.0
      - name: kubecost.v1.99.0
        version: 1.99.0
      - name: kubecost.v1.97.0
        version: 1.97.0
      name: alpha
    defaultChannel: alpha
    packageName: kubecost
    provider:
      name: Stackwatch
      url: www.kubecost.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: prometurbo
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometurbo-operator.v8.7.1-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.7.1-SNAPSHOT
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.7.1-SNAPSHOT
        version: 8.7.1-beta.1
      entries:
      - name: prometurbo-operator.v8.7.1-beta.1
        version: 8.7.1-beta.1
      name: beta
    - currentCSV: prometurbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.6
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.6
        version: 8.6.0
      entries:
      - name: prometurbo-operator.v8.6.0
        version: 8.6.0
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: stable
    - currentCSV: prometurbo-operator.v8.5.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.5
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.5
        version: 8.5.0
      entries:
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: v8.5.0
    - currentCSV: prometurbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.6
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.6
        version: 8.6.0
      entries:
      - name: prometurbo-operator.v8.6.0
        version: 8.6.0
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: v8.6.0
    defaultChannel: stable
    packageName: prometurbo
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apache KIE
      provider-url: ""
    name: sonataflow-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sonataflow-operator.v10.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "docker.io/apache",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: docker.io/apache/incubator-kie-sonataflow-operator:10.0.0
          description: SonataFlow Kubernetes Operator for deploying workflow applications
            based on the CNCF Serverless Workflow specification
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/apache/incubator-kie-tools/packages/sonataflow-operator
          support: Apache Software Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: Sonata Flow Build
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: Sonata Flow Cluster Platform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: Sonata Flow Platform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: Sonata Flow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          SonataFlow Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification](https://serverlessworkflow.io/):

          * Deploy workflow applications using the [dev profile](https://sonataflow.org/serverlessworkflow/latest/cloud/operator/developing-workflows.html), suited for the your development cycle
          * Build workflow applications based on the platform you're currently working on.
        displayName: SonataFlow Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        links:
        - name: Product Page
          url: https://github.com/apache/incubator-kie-tools
        maintainers:
        - email: dev@kie.apache.org
          name: Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Apache Software Foundation
        relatedImages:
        - docker.io/apache/incubator-kie-sonataflow-operator:10.0.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 10.0.0
      entries:
      - name: sonataflow-operator.v10.0.0
        version: 10.0.0
      - name: sonataflow-operator.v1.40.1
        version: 1.40.1
      - name: sonataflow-operator.v1.40.0
        version: 1.40.0
      name: alpha
    - currentCSV: sonataflow-operator.v10.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "host/namespace",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: docker.io/apache/incubator-kie-sonataflow-operator:10.1.0
          createdAt: "2025-09-11T14:44:56Z"
          description: SonataFlow Kubernetes Operator for deploying workflow applications
            based on the CNCF Serverless Workflow specification
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/apache/incubator-kie-tools/packages/sonataflow-operator
          support: Apache KIE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: SonataFlowBuild
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: SonataFlowClusterPlatform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: SonataFlowPlatform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: SonataFlow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          SonataFlow Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification](https://serverlessworkflow.io/):

          * Deploy workflow applications using the [dev profile](https://sonataflow.org/serverlessworkflow/latest/cloud/operator/developing-workflows.html), suited for the your development cycle
          * Build workflow applications based on the platform you're currently working on.
        displayName: SonataFlow Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        links:
        - name: Product Page
          url: https://sonataflow.org/serverlessworkflow/latest/index.html
        maintainers:
        - email: dev@kie.apache.org
          name: Apache KIE
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Apache KIE
        relatedImages:
        - docker.io/apache/incubator-kie-sonataflow-operator:10.1.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 10.1.0
      entries:
      - name: sonataflow-operator.v10.1.0
        version: 10.1.0
      - name: sonataflow-operator.v10.0.0
        version: 10.0.0
      - name: sonataflow-operator.v1.40.1
        version: 1.40.1
      - name: sonataflow-operator.v1.40.0
        version: 1.40.0
      name: stable
    defaultChannel: stable
    packageName: sonataflow-operator
    provider:
      name: Apache KIE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: directpv
      provider-url: ""
    name: private-dpv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: private-dpv-operator.v5.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.quay.io/v1alpha1",
                "kind": "AistorVolumemanager",
                "metadata": {
                  "name": "aistorvolumemanager-sample"
                },
                "spec": {
                  "annotations": {},
                  "controller": {
                    "podAnnotations": {},
                    "podLabels": {},
                    "replicas": 3
                  },
                  "images": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": []
                  },
                  "labels": {},
                  "license": "",
                  "nodeserver": {
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-10-05T15:47:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AistorVolumemanager
            name: aistorvolumemanagers.charts.quay.io
            version: v1alpha1
        description: directpv
        displayName: directpv
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - directpv
        links:
        - name: Private Dpv Operator
          url: https://private-dpv-operator.domain
        maintainers:
        - email: cesarcelis@minio.io
          name: celis
        maturity: alpha
        provider:
          name: directpv
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - quay.io/cniackz4/directpv-operator@sha256:af06e57661a7e05809e44c3fac571ed949cc04fd656fdb0c8324b533778eb604
        version: 5.0.3
      entries:
      - name: private-dpv-operator.v5.0.3
        version: 5.0.3
      - name: private-dpv-operator.v5.0.2
        version: 5.0.2
      name: alpha
    defaultChannel: alpha
    packageName: private-dpv-operator
    provider:
      name: directpv
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: service-telemetry-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-telemetry-operator.v1.5.1757650583
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infra.watch/v1beta1",
                "kind": "ServiceTelemetry",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "alerting": {
                    "alertmanager": {
                      "receivers": {
                        "snmpTraps": {
                          "alertOidLabel": "oid",
                          "community": "public",
                          "enabled": false,
                          "port": 162,
                          "retries": 5,
                          "target": "192.168.24.254",
                          "timeout": 1,
                          "trapDefaultOid": "1.3.6.1.4.1.50495.15.1.2.1",
                          "trapDefaultSeverity": "",
                          "trapOidPrefix": "1.3.6.1.4.1.50495.15"
                        }
                      },
                      "storage": {
                        "persistent": {
                          "pvcStorageRequest": "20G"
                        },
                        "strategy": "persistent"
                      }
                    },
                    "enabled": true
                  },
                  "backends": {
                    "events": {
                      "elasticsearch": {
                        "certificates": {
                          "caCertDuration": "70080h",
                          "endpointCertDuration": "70080h"
                        },
                        "enabled": false,
                        "forwarding": {
                          "hostUrl": "https://elasticsearch-es-http:9200",
                          "tlsSecretName": "elasticsearch-es-cert",
                          "tlsServerName": "",
                          "useBasicAuth": true,
                          "useTls": true,
                          "userSecretName": "elasticsearch-es-elastic-user"
                        },
                        "storage": {
                          "persistent": {
                            "pvcStorageRequest": "20Gi"
                          },
                          "strategy": "persistent"
                        },
                        "version": "7.16.1"
                      }
                    },
                    "metrics": {
                      "prometheus": {
                        "enabled": true,
                        "scrapeInterval": "30s",
                        "storage": {
                          "persistent": {
                            "pvcStorageRequest": "20G"
                          },
                          "retention": "24h",
                          "strategy": "persistent"
                        }
                      }
                    }
                  },
                  "clouds": [
                    {
                      "metrics": {
                        "collectors": [
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 16384,
                              "verbose": false
                            },
                            "collectorType": "collectd",
                            "debugEnabled": false,
                            "subscriptionAddress": "collectd/cloud1-telemetry"
                          },
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 16384,
                              "verbose": false
                            },
                            "collectorType": "ceilometer",
                            "debugEnabled": false,
                            "subscriptionAddress": "anycast/ceilometer/cloud1-metering.sample"
                          },
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 65535,
                              "verbose": false
                            },
                            "collectorType": "sensubility",
                            "debugEnabled": false,
                            "subscriptionAddress": "sensubility/cloud1-telemetry"
                          }
                        ]
                      },
                      "name": "cloud1"
                    }
                  ],
                  "graphing": {
                    "enabled": false,
                    "grafana": {
                      "baseImage": "registry.redhat.io/rhel8/grafana:9",
                      "dashboards": {
                        "enabled": true
                      },
                      "disableSignoutMenu": false,
                      "ingressEnabled": true
                    }
                  },
                  "highAvailability": {
                    "enabled": false
                  },
                  "observabilityStrategy": "use_redhat",
                  "transports": {
                    "qdr": {
                      "auth": "basic",
                      "certificates": {
                        "caCertDuration": "70080h",
                        "endpointCertDuration": "70080h"
                      },
                      "enabled": true,
                      "web": {
                        "enabled": false
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/stf/service-telemetry-rhel9-operator@sha256:822becd77fa8835e753c88b3e307f62e42e090bdedc6e1fbcbe9fd6110310846
          createdAt: "2025-09-12T04:17:40Z"
          description: Service Telemetry Framework. Umbrella Operator for instantiating
            the required dependencies and configuration of various components to build
            a Service Telemetry platform for telco grade monitoring.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: service-telemetry
          operators.openshift.io/valid-subscription: '["OpenStack Platform", "Cloud
            Infrastructure", "Cloud Suite"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/infrawatch/service-telemetry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of the Service Telemetry Framework
            displayName: STF Cluster
            kind: ServiceTelemetry
            name: servicetelemetrys.infra.watch
            version: v1beta1
        description: Service Telemetry Operator for monitoring clouds
        displayName: Service Telemetry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - servicetelemetry
        - monitoring
        - telemetry
        - notifications
        - telecommunications
        links:
        - name: Source Code
          url: https://github.com/infrawatch/service-telemetry-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_openstack_platform/17.1/html/service_telemetry_framework_1.5/index
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy:latest@sha256:dc24aa382e02c4ac058d4b824a21f0e90a6259d6d817d0341b2a41809b5daa78
        - registry.redhat.io/cluster-observability-operator/coo-prometheus-rhel8:0.2.0@sha256:6bbb1aebb1cfdc3310e398ec7720ca1d4a9a89cecf7406412071bfaceb0d3cb3
        - registry.redhat.io/cluster-observability-operator/coo-prometheus-alertmanager-rhel8:0.2.0@sha256:cfd18e83d64a55ac310937eefe83fc5b9e98660962a127cd4f8d9a382764c171
        - registry.redhat.io/stf/service-telemetry-rhel9-operator@sha256:822becd77fa8835e753c88b3e307f62e42e090bdedc6e1fbcbe9fd6110310846
        - registry.redhat.io/stf/prometheus-webhook-snmp-rhel9@sha256:03199f8a3b6b22c82e01ede3adab6e98b276369d642029ae03641aac3b25581a
        version: 1.5.1757650583
      entries:
      - name: service-telemetry-operator.v1.5.1757650583
        version: 1.5.1757650583
      - name: service-telemetry-operator.v1.5.1743671601
        version: 1.5.1743671601
      - name: service-telemetry-operator.v1.5.1740051851
        version: 1.5.1740051851
      - name: service-telemetry-operator.v1.5.1731336911
        version: 1.5.1731336911
      - name: service-telemetry-operator.v1.5.1725952694
        version: 1.5.1725952694
      name: stable-1.5
    defaultChannel: stable-1.5
    packageName: service-telemetry-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: barbican-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: barbican-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "barbican.openstack.org/v1beta1",
                "kind": "Barbican",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "barbican-operator",
                    "app.kubernetes.io/instance": "barbican",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "barbican",
                    "app.kubernetes.io/part-of": "barbican-operator"
                  },
                  "name": "barbican"
                },
                "spec": {
                  "barbicanAPI": {
                    "customServiceConfig": "[optional]\noverrides = True\n",
                    "defautlConfigOverwrite": {
                      "optional_policy.json": "{\"some\": \"custom policy\"}\n"
                    },
                    "passwordSelectors": {
                      "database": "BarbicanDatabasePassword",
                      "service": "BarbicanPassword"
                    }
                  },
                  "barbicanKeystoneListener": {
                    "customServiceConfig": "[optional]\noverrides = True\n",
                    "defautlConfigOverwrite": {
                      "optional_policy.json": "{\"some\": \"custom policy\"}\n"
                    }
                  },
                  "barbicanWorker": {
                    "customServiceConfig": "[optional]\noverrides = True\n",
                    "defautlConfigOverwrite": {
                      "optional_policy.json": "{\"some\": \"custom policy\"}\n"
                    }
                  },
                  "customServiceConfig": "[DEFAULT]\ndebug = True\n",
                  "databaseAccount": "barbican",
                  "databaseInstance": "openstack",
                  "defaultConfigOverwrite": {
                    "policy.json": "{\"some\": \"custom policy\"}\n"
                  },
                  "passwordSelectors": {
                    "database": "BarbicanDatabasePassword",
                    "service": "BarbicanPassword",
                    "simplecryptokek": "BarbicanSimpleCryptoKEK"
                  },
                  "preserveJobs": true,
                  "rabbitMqCusterName": "barbican_rabbit",
                  "secret": "osp-secret",
                  "serviceAccount": "barbican",
                  "serviceUser": "barbican"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:32Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BarbicanAPI is the Schema for the barbicanapis API
            displayName: Barbican API
            kind: BarbicanAPI
            name: barbicanapis.barbican.openstack.org
            version: v1beta1
          - kind: BarbicanKeystoneListener
            name: barbicankeystonelisteners.barbican.openstack.org
            version: v1beta1
          - description: Barbican is the Schema for the barbicans API
            displayName: Barbican
            kind: Barbican
            name: barbicans.barbican.openstack.org
            version: v1beta1
          - description: BarbicanWorker is the Schema for the barbicanworkers API
            displayName: Barbican Worker
            kind: BarbicanWorker
            name: barbicanworkers.barbican.openstack.org
            version: v1beta1
        description: Barbican Operator
        displayName: Barbican Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - KeyManager
        - Barbican
        links:
        - name: Barbican Operator
          url: https://github.com/openstack-k8s-operators/barbican-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-barbican-worker-rhel9@sha256:b5133f539658ed5d532e0c3cf63d90bc0f76c7f0368963a8ebec49123f27f7a7
        - registry.redhat.io/rhoso/openstack-barbican-keystone-listener-rhel9@sha256:ac112444c0a12664e8a7900e72cd97fa795df167c0d95437a029644f0819f30d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/barbican-rhel9-operator@sha256:1275a3f7f09999e9dc25bc38f917a3b5d780d9d309442cddd0e7adf313d6f576
        - registry.redhat.io/rhoso/openstack-barbican-api-rhel9@sha256:0db6c1ab5b4ca302e38338a3ab9cee9bd4ae7c8e1e7bd3419e47b65df9ef1f4f
        version: 1.0.6
      entries:
      - name: barbican-operator.v1.0.6
        version: 1.0.6
      - name: barbican-operator.v1.0.5
        version: 1.0.5
      - name: barbican-operator.v1.0.4
        version: 1.0.4
      - name: barbican-operator.v1.0.3
        version: 1.0.3
      - name: barbican-operator.v1.0.2
        version: 1.0.2
      - name: barbican-operator.v1.0.1
        version: 1.0.1
      - name: barbican-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: barbican-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: ansible-cloud-addons-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aca-operator.v2.4.0-0.1737660894
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Cloud Addons Operator manages everything Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.1.1-0.1645500829 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy an AAP AWS Billing instance
            displayName: AnsibleAutomationPlatformAwsBilling
            kind: AnsibleAutomationPlatformAwsBilling
            name: ansibleautomationplatformawsbillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Azure Billing instance
            displayName: AnsibleAutomationPlatformAzureBilling
            kind: AnsibleAutomationPlatformAzureBilling
            name: ansibleautomationplatformazurebillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Metrics instance
            displayName: AnsibleAutomationPlatformMetrics
            kind: AnsibleAutomationPlatformMetrics
            name: ansibleautomationplatformmetrics.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP UI instance
            displayName: AutomationPlatform
            kind: AutomationPlatform
            name: automationplatforms.platform.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Cloud Addons is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Cloud Addons Operator provides for push-button deployments of Cloud Add-ons into your OpenShift environment. Ansible Cloud Addons deployments include:
            * Cloud Billing - Cloud billing connector for Ansible Automation Platform
            * Cloud UI - Unified welcome page UI for Ansible Automation Platform

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Cloud Addons operator by following instructions presented when you click the Install button.
        displayName: Ansible Cloud Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-platform
        - ansible-automation-platform
        - aap
        - cloud
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:38ee3484ba60bb474e7231e9d0718157386716cea7ffd6da6f32ddd551a5243d
        - registry.redhat.io/ansible-automation-platform-24/aap-cloud-billing-rhel8-operator@sha256:c461b0c4781e95d2b2e31e7bbd6d6d192806e8a8dd2e8b9225b5735da93dd6a2
        - registry.redhat.io/ansible-automation-platform-24/aap-cloud-ui-rhel8-operator@sha256:06bf313d4daf080ce6f2b1ba859fefd84431c7a2a0618feae55b92c9f5fff594
        - registry.redhat.io/ansible-automation-platform-24/aap-cloud-billing-rhel8@sha256:84fa86d312663a74502516f2ba838ac80e500c7bc17dc518940984a603da2548
        - registry.redhat.io/ansible-automation-platform-24/aap-cloud-metrics-collector-rhel8@sha256:6bd27ad06d0a65c7fa1cd2419bfc30976d53a5b89698d309cabbbd65b531a1ab
        - registry.redhat.io/ansible-automation-platform-24/aap-cloud-ui-rhel8@sha256:8489c253072186db064289f5ecc0a08ed3654a73be3c3f2130be479098e59047
        version: 2.4.0+0.1737660894
      entries:
      - name: aca-operator.v2.4.0-0.1708974547
        version: 2.4.0+0.1708974547
      - name: aca-operator.v2.4.0-0.1711587287
        version: 2.4.0+0.1711587287
      - name: aca-operator.v2.4.0-0.1713463331
        version: 2.4.0+0.1713463331
      - name: aca-operator.v2.4.0-0.1715891874
        version: 2.4.0+0.1715891874
      - name: aca-operator.v2.4.0-0.1717108266
        version: 2.4.0+0.1717108266
      - name: aca-operator.v2.4.0-0.1717558193
        version: 2.4.0+0.1717558193
      - name: aca-operator.v2.4.0-0.1718149348
        version: 2.4.0+0.1718149348
      - name: aca-operator.v2.4.0-0.1720478479
        version: 2.4.0+0.1720478479
      - name: aca-operator.v2.4.0-0.1725253934
        version: 2.4.0+0.1725253934
      - name: aca-operator.v2.4.0-0.1726108417
        version: 2.4.0+0.1726108417
      - name: aca-operator.v2.4.0-0.1727255735
        version: 2.4.0+0.1727255735
      - name: aca-operator.v2.4.0-0.1730152463
        version: 2.4.0+0.1730152463
      - name: aca-operator.v2.4.0-0.1733181569
        version: 2.4.0+0.1733181569
      - name: aca-operator.v2.4.0-0.1733937635
        version: 2.4.0+0.1733937635
      - name: aca-operator.v2.4.0-0.1737660894
        version: 2.4.0+0.1737660894
      name: stable-2.4-cluster-scoped
    - currentCSV: aca-operator.v2.5.0-0.1758143647
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Cloud Addons Operator manages everything Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.1.1-0.1645500829 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy an AAP AWS Billing instance
            displayName: AnsibleAutomationPlatformAwsBilling
            kind: AnsibleAutomationPlatformAwsBilling
            name: ansibleautomationplatformawsbillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Azure Billing instance
            displayName: AnsibleAutomationPlatformAzureBilling
            kind: AnsibleAutomationPlatformAzureBilling
            name: ansibleautomationplatformazurebillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Metrics instance
            displayName: AnsibleAutomationPlatformMetrics
            kind: AnsibleAutomationPlatformMetrics
            name: ansibleautomationplatformmetrics.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP UI instance
            displayName: AutomationPlatform
            kind: AutomationPlatform
            name: automationplatforms.platform.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Cloud Addons is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Cloud Addons Operator provides for push-button deployments of Cloud Add-ons into your OpenShift environment. Ansible Cloud Addons deployments include:
            * Cloud Billing - Cloud billing connector for Ansible Automation Platform
            * Cloud UI - Unified welcome page UI for Ansible Automation Platform

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Cloud Addons operator by following instructions presented when you click the Install button.
        displayName: Ansible Cloud Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-platform
        - ansible-automation-platform
        - aap
        - cloud
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-billing-rhel8-operator@sha256:f4b29ec59498bb6c9fdb2790ffc9d8c878efa3fd9da3416f7fe9f2a253171508
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-ui-rhel8-operator@sha256:47e2c40cdcfe218d5bf7abbf6fc35340bacdc2d02dac7f81deee7876b5570c53
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-billing-rhel8@sha256:0e50f539bfeaec91f0b18409cbc1888297a26af5114252a8ca07e1b97436001d
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-metrics-collector-rhel8@sha256:c6ab25b3839f8c3134e7d5e1f0b7716ea5fd66fe8d131261ff08aabd97c3cd2d
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-ui-rhel8@sha256:869efe830cdf23f9344263f0e9b5a209cb13b84d9d7588a1744fd3ea4cd3f83b
        version: 2.5.0+0.1758143647
      entries:
      - name: aca-operator.v2.5.0-0.1727240231
        version: 2.5.0+0.1727240231
      - name: aca-operator.v2.5.0-0.1727874815
        version: 2.5.0+0.1727874815
      - name: aca-operator.v2.5.0-0.1728516369
        version: 2.5.0+0.1728516369
      - name: aca-operator.v2.5.0-0.1729737962
        version: 2.5.0+0.1729737962
      - name: aca-operator.v2.5.0-0.1731452082
        version: 2.5.0+0.1731452082
      - name: aca-operator.v2.5.0-0.1733188778
        version: 2.5.0+0.1733188778
      - name: aca-operator.v2.5.0-0.1733964232
        version: 2.5.0+0.1733964232
      - name: aca-operator.v2.5.0-0.1736892777
        version: 2.5.0+0.1736892777
      - name: aca-operator.v2.5.0-0.1737095650
        version: 2.5.0+0.1737095650
      - name: aca-operator.v2.5.0-0.1737671171
        version: 2.5.0+0.1737671171
      - name: aca-operator.v2.5.0-0.1739244024
        version: 2.5.0+0.1739244024
      - name: aca-operator.v2.5.0-0.1740088272
        version: 2.5.0+0.1740088272
      - name: aca-operator.v2.5.0-0.1740769078
        version: 2.5.0+0.1740769078
      - name: aca-operator.v2.5.0-0.1741367998
        version: 2.5.0+0.1741367998
      - name: aca-operator.v2.5.0-0.1742427652
        version: 2.5.0+0.1742427652
      - name: aca-operator.v2.5.0-0.1743655743
        version: 2.5.0+0.1743655743
      - name: aca-operator.v2.5.0-0.1746134365
        version: 2.5.0+0.1746134365
      - name: aca-operator.v2.5.0-0.1747339736
        version: 2.5.0+0.1747339736
      - name: aca-operator.v2.5.0-0.1749072064
        version: 2.5.0+0.1749072064
      - name: aca-operator.v2.5.0-0.1750895274
        version: 2.5.0+0.1750895274
      - name: aca-operator.v2.5.0-0.1753899647
        version: 2.5.0+0.1753899647
      - name: aca-operator.v2.5.0-0.1755834478
        version: 2.5.0+0.1755834478
      - name: aca-operator.v2.5.0-0.1758143647
        version: 2.5.0+0.1758143647
      name: stable-2.5-cluster-scoped
    defaultChannel: stable-2.5-cluster-scoped
    packageName: ansible-cloud-addons-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SAP LinuxLab - Red Hat
      provider-url: https://github.com/sap-linuxlab
    name: sap-data-intelligence-observer-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: observer-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sdi.sap-redhat.io/v1alpha1",
                "kind": "SDIObserver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "observer-operator",
                    "app.kubernetes.io/instance": "sdiobserver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sdiobserver",
                    "app.kubernetes.io/part-of": "observer-operator"
                  },
                  "name": "sdiobserver-sample"
                },
                "spec": {
                  "manageSDINodeConfig": true,
                  "sdiNamespace": "sdi",
                  "sdiVSystemRoute": {
                    "managementState": "Managed"
                  },
                  "slcbNamespace": "sap-slcbridge",
                  "slcbRoute": {
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/redhat-sap-cop/sdi-observer-operator:0.1.5
          description: Configure and enable SAP Data Intelligence to run on OpenShift
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SDIObserver is the Schema for the sdiobservers API
            displayName: SDIObserver
            kind: SDIObserver
            name: sdiobservers.sdi.sap-redhat.io
            version: v1alpha1
        description: Operator for monitoring SAP Data Intelligence (SDI) namespace
          and modifying objects in there that enable running of SDI on top of OpenShift.
          The observer shall be run in a dedicated namespace. It must be deployed
          before the SDI installation is started. This is the prerequisite for the
          successful installation of the SDI on OpenShift. Please refer to https://access.redhat.com/articles/5100521
          for detailed information.
        displayName: SAP Data Intelligence 3 - Observer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sdi
        - sap data intelligence
        links:
        - name: SAP Data Intelligence 3 - Observer Operator
          url: https://github.com/redhat-sap/sap-data-intelligence/tree/master/observer-operator
        maintainers:
        - email: mjiao@redhat.com
          name: Manjun Jiao
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/redhat-sap-cop/sdi-observer-operator:0.1.5
        version: 0.0.3
      entries:
      - name: observer-operator.v0.0.3
        version: 0.0.3
      name: alpha
    - currentCSV: observer-operator.v1.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sdi.sap-redhat.io/v1alpha1",
                "kind": "SDIObserver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/part-of": "observer-operator"
                  },
                  "name": "sdiobserver-sample",
                  "namespace": "sdi-observer"
                },
                "spec": {
                  "manageSDINodeConfig": true,
                  "sdiNamespace": "sdi",
                  "sdiVSystemRoute": {
                    "managementState": "Managed"
                  },
                  "slcbNamespace": "sap-slcbridge",
                  "slcbRoute": {
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/redhat-sap-cop/sdi-observer-operator:0.1.32
          createdAt: "20240726"
          description: Configure and enable SAP Data Intelligence 3 to run on OpenShift
            4
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mjiao/sap-data-intelligence/tree/master/observer-operator
          support: RedHat-SAPLinuxLab
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SDIObserver is the Schema for the sdiobservers API
            displayName: SDIObserver
            kind: SDIObserver
            name: sdiobservers.sdi.sap-redhat.io
            version: v1alpha1
        description: Operator for monitoring SAP Data Intelligence (SDI) namespace
          and modifying objects in there that enable running of SDI on top of OpenShift.
          It must be deployed before the SDI installation is started. This is the
          prerequisite for the successful installation of the SDI on OpenShift. Please
          refer to https://access.redhat.com/articles/7042265#ocp-install-sdi-observer-operator
          for detailed information.
        displayName: SAP Data Intelligence 3 - Observer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sdi
        - sap data intelligence
        links:
        - name: SAP Data Intelligence 3 - Observer Operator
          url: https://github.com/redhat-sap/sap-data-intelligence/tree/master/observer-operator
        maintainers:
        - email: mjiao@redhat.com
          name: Manjun Jiao
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/redhat-sap-cop/sdi-observer-operator:0.1.32
        version: 1.0.9
      entries:
      - name: observer-operator.v1.0.9
        version: 1.0.9
      - name: observer-operator.v1.0.8
        version: 1.0.8
      - name: observer-operator.v1.0.7
        version: 1.0.7
      - name: observer-operator.v1.0.6
        version: 1.0.6
      - name: observer-operator.v1.0.5
        version: 1.0.5
      - name: observer-operator.v1.0.4
        version: 1.0.4
      - name: observer-operator.v1.0.3
        version: 1.0.3
      - name: observer-operator.v1.0.2
        version: 1.0.2
      - name: observer-operator.v1.0.1
        version: 1.0.1
      - name: observer-operator.v0.0.3
        version: 0.0.3
      name: stable
    defaultChannel: stable
    packageName: sap-data-intelligence-observer-operator
    provider:
      name: SAP LinuxLab - Red Hat
      url: https://github.com/sap-linuxlab
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: node-healthcheck-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-healthcheck-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "maxUnhealthy": "49%",
                  "remediationTemplate": {
                    "apiVersion": "poison-pill.medik8s.io/v1alpha1",
                    "kind": "PoisonPillRemediationTemplate",
                    "name": "group-x",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-healthcheck-operator:0.1.0-latest
          createdAt: "2021-04-20 12:00:00"
          description: An operator to monitor node healthcheck and remdiate via 3rd
            party providers like poison-pill.
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: This operator deploys the Node Health Check controller and its
          artifacts
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://node-healthcheck-operator.domain
        maintainers:
        - email: medik8s@googlegroups.com
          name: medik8s team
        maturity: alpha
        provider:
          name: medik8s
          url: https://github.com/medik8s
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/node-healthcheck-operator:0.1.0-latest
        version: 0.1.1
      entries:
      - name: node-healthcheck-operator.v0.1.1
        version: 0.1.1
      name: alpha
    - currentCSV: node-healthcheck-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-healthcheck-operator:v0.3.0
          createdAt: "2022-08-24 11:40:38"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.2.0 <0.3.0'
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: medik8s team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: medik8s
          url: https://github.com/medik8s
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/node-healthcheck-operator:v0.3.0
        version: 0.3.0
      entries:
      - name: node-healthcheck-operator.v0.3.0
        version: 0.3.0
      name: candidate
    - currentCSV: node-healthcheck-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: quay.io/medik8s/node-healthcheck-operator:v0.6.1
          createdAt: "2023-11-14T15:45:34Z"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.1.0 <0.6.1'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/medik8s/node-healthcheck-operator:v0.6.1
        - quay.io/medik8s/node-remediation-console:v0.6.1
        version: 0.6.1
      entries:
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      - name: node-healthcheck-operator.v0.6.0
        version: 0.6.0
      - name: node-healthcheck-operator.v0.5.0
        version: 0.5.0
      - name: node-healthcheck-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: node-healthcheck-operator
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: GitLab Inc
      provider-url: https://about.gitlab.com/
    name: gitlab-operator-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "gitlab": {
                        "gitlab-shell": {
                          "maxReplicas": 2,
                          "minReplicas": 1
                        },
                        "sidekiq": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1000M"
                            }
                          }
                        },
                        "webservice": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1500M"
                            }
                          }
                        }
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "installCertmanager": false,
                      "minio": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      },
                      "nginx-ingress": {
                        "controller": {
                          "electionID": "ingress-controller-leader",
                          "replicaCount": 1
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      },
                      "redis": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-29T12:42:23Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      name: stable
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "gitlab": {
                        "gitlab-shell": {
                          "maxReplicas": 2,
                          "minReplicas": 1
                        },
                        "sidekiq": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1000M"
                            }
                          }
                        },
                        "webservice": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1500M"
                            }
                          }
                        }
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "installCertmanager": false,
                      "minio": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      },
                      "nginx-ingress": {
                        "controller": {
                          "electionID": "ingress-controller-leader",
                          "replicaCount": 1
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      },
                      "redis": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-29T12:42:23Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      name: unstable
    defaultChannel: stable
    packageName: gitlab-operator-kubernetes
    provider:
      name: GitLab Inc
      url: https://about.gitlab.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: edb-hcp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: edb-hcp-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edbpgai.edb.com/v1alpha1",
                "kind": "HybridControlPlane",
                "metadata": {
                  "name": "placeholder"
                },
                "spec": {
                  "beaconAgent": {
                    "applianceProviderEnabled": true,
                    "beaconDBConnected": true,
                    "beaconServerAddress": "upm-beacon-server-ingestion.upm-beacon.svc.cluster.local:8080",
                    "inCluster": true,
                    "location": "devspatcher",
                    "logLevel": "debug",
                    "plaintext": true,
                    "providers": [
                      "appliance"
                    ],
                    "provisioning": {
                      "imageDiscovery": true,
                      "imagesetDiscoveryAuthenticationType": "token",
                      "imagesetDiscoveryContainerRegistryURL": "placeholder",
                      "loadBalancersEnabled": true,
                      "nodePortDomain": "nodeportdomain.com",
                      "openshift": true,
                      "provider": "aws"
                    },
                    "tlsInsecure": false,
                    "transparentDataEncryptionMethods": [
                      "passphrase",
                      "hashicorp_vault"
                    ]
                  },
                  "beaconServer": {
                    "logLevel": "debug"
                  },
                  "componentsParameters": {
                    "edb-migration-portal": {
                      "enable_pdb": "false"
                    },
                    "transporter-data-operator": {
                      "fips_enabled": "false"
                    },
                    "transporter-dp-agent": {
                      "rw_service_url": "placeholder"
                    },
                    "transporter-rw-service": {
                      "domain_name": "placeholder"
                    },
                    "upm-api-pnxt-metric": {
                      "log_level": "info",
                      "tsdb_host": "thanos-query.monitoring.svc.cluster.local",
                      "tsdb_port": "9090"
                    },
                    "upm-api-pnxt-pd": {
                      "log_level": "info"
                    },
                    "upm-beaco-ff-base": {
                      "usage_generator_interval": "24h"
                    },
                    "upm-beaco-usages": {
                      "usage_generator_interval": "24h"
                    },
                    "upm-beacon": {
                      "beacon_db_hibernation": "\"off\"",
                      "beacon_server_replicas": "1",
                      "external_metrics_replicas": "1",
                      "image_discovery_credentials_secret_name": "placeholder",
                      "server_host": "placeholder"
                    },
                    "upm-istio": {
                      "ingress_grpc_tls_node_port": "30290",
                      "ingress_http_node_port": "32542",
                      "ingress_https_node_port": "30288"
                    },
                    "upm-istio-gateway": {
                      "cookie_aeskey": "rzkutHl8NJNztPMEJYykZouHslNiA7xmIXH+58ISUVo=",
                      "ingress_grpc_tls_node_port": "30290",
                      "ingress_http_node_port": "32542",
                      "ingress_https_node_port": "30288",
                      "openshift_console_domain_name": "placeholder"
                    },
                    "upm-prometheus": {
                      "cpu": "200m",
                      "memory": "6Gi",
                      "prometheus_retention_size": "15GB",
                      "prometheus_storage_size": "20Gi"
                    },
                    "upm-thanos": {
                      "memcached_cpu": "500m",
                      "memcached_memory": "2Gi",
                      "query_frontend_cpu": "200m",
                      "query_frontend_memory": "128Mi"
                    },
                    "upm-trust-manager": {
                      "cert_manager_namespace": "cert-manager"
                    }
                  },
                  "flavour": "placeholder",
                  "globalParameters": {
                    "default_ingress_domain": "placeholder",
                    "internal_backup_folder": "a7462dbc7106",
                    "portal_certificate_secret": "beaco-istio-gw-cert",
                    "portal_domain_name": "placeholder",
                    "portal_port": "443",
                    "storage_class": "placeholder"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "placeholder",
                      "namespace": "default"
                    }
                  ],
                  "imageRegistry": "placeholder",
                  "pgai": {
                    "portal": {
                      "authentication": {
                        "clientSecret": "edb-postgres-ai-secret",
                        "idpConnectors": [],
                        "staticPasswords": [
                          {
                            "email": "owner@mycompany.com",
                            "hash": "$2y$10/JLTqiuGblH.RemIutWxkztN5m",
                            "userID": "c5998173-a605-449a-a9a5-4a9c33e26df7",
                            "username": "Owner MyCompany"
                          }
                        ]
                      }
                    }
                  },
                  "source": {
                    "useLocalKustomizations": true
                  },
                  "version": "placeholder"
                }
              }
            ]
          capabilities: Basic Install
          categories: Database,AI/Machine Learning
          containerImage: docker.enterprisedb.com/pgai-platform/edb-hcp-operator/manager@sha256:82610c624e956a67df4858929e1aabd561a7537ba68d40e6512b68f55e2649e0
          createdAt: "2025-06-17T11:00:42Z"
          description: Unify transactional, analytical, and AI workloads and bring
            cloud agility to your hybrid data infrastructure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/EnterpriseDB/edb-hcp-operator
          support: techsupport@enterprisedb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HybridControlPlane is the Schema for the hybridcontrolplanes
              API
            displayName: Hybrid Control Plane
            kind: HybridControlPlane
            name: hybridcontrolplanes.edbpgai.edb.com
            version: v1alpha1
        description: |
          EDB Postgres AI is a sovereign data and AI platform that brings analytics and AI closer to core operational data and enables cloud agility in private environments.
          Unify your structured and unstructured data in one multi-model data platform with centralized management and observability to:
          - Accelerate enterprise-grade transactional databases
          - Enable rapid analytics and powerful lakehouse integrations
          - Automate AI pipelines and semantic search for faster GenAI development
          Using EDB’s unmatched Postgres expertise, EDB Postgres AI automates database operations, ensuring expert-level execution across deployments – boosting operational efficiency by up to 30% and accelerating innovation.
        displayName: EDB Postgres AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Postgres
        - Database
        - DBaaS
        - EDB Postgres AI
        links:
        - name: EDB Postgres AI Software Deployment
          url: https://www.enterprisedb.com/products/software-deployment
        maintainers:
        - email: techsupport@enterprisedb.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - docker.enterprisedb.com/pgai-platform/edb-hcp-operator/manager@sha256:82610c624e956a67df4858929e1aabd561a7537ba68d40e6512b68f55e2649e0
        version: 1.2.0
      entries:
      - name: edb-hcp-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: edb-hcp-operator
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-documentdb-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-documentdb-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBInstance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/documentdb-controller:1.1.1
          createdAt: "2025-09-19T22:13:37Z"
          description: AWS DocumentDB controller is a service controller for managing
            DocumentDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DBCluster represents the state of an AWS documentdb DBCluster
              resource.
            displayName: DBCluster
            kind: DBCluster
            name: dbclusters.documentdb.services.k8s.aws
            version: v1alpha1
          - description: DBInstance represents the state of an AWS documentdb DBInstance
              resource.
            displayName: DBInstance
            kind: DBInstance
            name: dbinstances.documentdb.services.k8s.aws
            version: v1alpha1
          - description: DBSubnetGroup represents the state of an AWS documentdb DBSubnetGroup
              resource.
            displayName: DBSubnetGroup
            kind: DBSubnetGroup
            name: dbsubnetgroups.documentdb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon DocumentDB resources in AWS from within your Kubernetes cluster.

          **About Amazon DocumentDB**

          Amazon DocumentDB (with MongoDB compatibility) is a fast, reliable, and fully managed database service. Amazon DocumentDB makes it easy to set up, operate, and scale MongoDB-compatible databases in the cloud. With Amazon DocumentDB, you can run the same application code and use the same drivers and tools that you use with MongoDB.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon DocumentDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - documentdb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon DocumentDB Developer Resources
          url: https://aws.amazon.com/documentdb/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: documentdb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/documentdb-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-documentdb-controller.v1.1.1
        version: 1.1.1
      - name: ack-documentdb-controller.v1.1.0
        version: 1.1.0
      - name: ack-documentdb-controller.v1.0.14
        version: 1.0.14
      - name: ack-documentdb-controller.v1.0.13
        version: 1.0.13
      - name: ack-documentdb-controller.v1.0.12
        version: 1.0.12
      - name: ack-documentdb-controller.v1.0.11
        version: 1.0.11
      - name: ack-documentdb-controller.v1.0.10
        version: 1.0.10
      - name: ack-documentdb-controller.v1.0.9
        version: 1.0.9
      - name: ack-documentdb-controller.v1.0.7
        version: 1.0.7
      - name: ack-documentdb-controller.v1.0.6
        version: 1.0.6
      - name: ack-documentdb-controller.v1.0.5
        version: 1.0.5
      - name: ack-documentdb-controller.v1.0.2
        version: 1.0.2
      - name: ack-documentdb-controller.v1.0.1
        version: 1.0.1
      - name: ack-documentdb-controller.v1.0.0
        version: 1.0.0
      - name: ack-documentdb-controller.v0.0.9
        version: 0.0.9
      - name: ack-documentdb-controller.v0.0.8
        version: 0.0.8
      - name: ack-documentdb-controller.v0.0.7
        version: 0.0.7
      - name: ack-documentdb-controller.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: ack-documentdb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: k8gb.io
      provider-url: https://github.com/k8gb-io/k8gb
    name: k8gb
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k8gb.v0.11.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"k8gb.absa.oss/v1beta1","kind":"Gslb","metadata":{"name":"test-gslb-failover","namespace":"test-gslb"},"spec":{"ingress":{"rules":[{"host":"failover.test.k8gb.io","http":{"paths":[{"backend":{"service":{"name":"frontend-podinfo","port":{"name":"http"}}},"path":"/"}]}}]},"strategy":{"primaryGeoTag":"eu-west-1","type":"failover"}}}]'
          capabilities: Seamless Upgrades
          categories: Networking
          certified: "false"
          containerImage: docker.io/absaoss/k8gb:v0.11.4
          createdAt: "2021-09-24 12:00:00"
          description: A cloud native Kubernetes Global Balancer
          operatorframework.io/suggested-namespace: k8gb
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: k8gb
          operators.operatorframework.io/builder: operator-sdk-v1.12.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/k8gb-io/k8gb
          support: cncf-k8gb-maintainers@lists.cncf.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gslb resource for global load balancing strategy configuration
            displayName: Gslb
            kind: Gslb
            name: gslbs.k8gb.absa.oss
            version: v1beta1
          - description: Using ExternalDNS it synchronizes exposed Kubernetes Services
              and Ingresses with DNS providers
            displayName: DNSEndpoint
            kind: DNSEndpoint
            name: dnsendpoints.externaldns.k8s.io
            version: v1alpha1
        description: |
          **A cloud native Kubernetes Global Balancer**

          A Global Service Load Balancing solution with a focus on having cloud native qualities and work natively in a Kubernetes context.

          Key Differentiators:
           - Load balancing is based on timeproof DNS protocol which is perfect for global scope and extremely reliable
           - No dedicated management cluster and no single point of failure
           - Kubernetes native application health checks utilizing status of Liveness and Readiness probes for load balancing decisions
           - Configuration with a single Kubernetes CRD of `Gslb` kind

          Operator needs to be configured by setting couple of environment variables. This can be done by modifying
          the subscription yaml in the web ui or via kubectl. Here is an example of changing the `EDGE_DNS_ZONE`:

          ```yaml
          kind: Subscription
            ...
          spec:
            ...
            config:
              env:
              - name: CLUSTER_GEO_TAG
                value: us
              - name: EDGE_DNS_ZONE
                value: mycloud.example.com
                ...
           ```
           For more deployment variables see [helm code](https://github.com/k8gb-io/k8gb/blob/v0.8.3/chart/k8gb/templates/operator.yaml#L53:L68).

           For more information see [k8gb.io](https://k8gb.io).
        displayName: k8gb
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gslb
        - dns-lb
        - kuberneters-global-balancer
        - kubernetes-operator
        - balancer
        - multi-cluster
        links:
        - name: Home
          url: https://www.k8gb.io/
        - name: Source
          url: https://github.com/k8gb-io/k8gb
        maintainers:
        - email: dinar.valeev@absa.africa
          name: Dinar Valeev
        - email: jiri.kremser@gmail.com
          name: Jiri Kremser
        - email: kuritka@gmail.com
          name: Michal Kuritka
        - email: timofey.ilinykh@absa.africa
          name: Timofey Ilinykh
        - email: yury.tsarev@absa.africa
          name: Yury Tsarev
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: k8gb.io
          url: https://github.com/k8gb-io/k8gb
        relatedImages:
        - absaoss/k8s_crd:v0.0.11
        - docker.io/absaoss/k8gb:v0.11.4
        version: 0.11.4
      entries:
      - name: k8gb.v0.11.4
        version: 0.11.4
      name: alpha
    defaultChannel: alpha
    packageName: k8gb
    provider:
      name: k8gb.io
      url: https://github.com/k8gb-io/k8gb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Internal
      provider-url: redhat.com
    name: l5-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: l5-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pets.bestie.com/v1",
                "kind": "Bestie",
                "metadata": {
                  "name": "bestie"
                },
                "spec": {
                  "image": "quay.io/opdev/bestie",
                  "maxReplicas": 10,
                  "size": 3,
                  "version": "1.3"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Application Runtime
          certified: "false"
          containerImage: quay.io/opdev/l5-operator:v0.0.3
          createdAt: "2023-06-27T16:56:10Z"
          description: L5 Operator is a demo operator that aims to demonstrate level
            5 capabilities for web applications.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          relatedImages: quay.io/opdev/bestie@sha256:0e99ada2b1de19043845dd9076789584aeddc2646402e5f0987dd00424be1ee7
          repository: https://github.com/opdev/l5-operator-demo
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Bestie is the Schema for the besties API.
            displayName: Bestie
            kind: Bestie
            name: besties.pets.bestie.com
            version: v1
        description: |
          ## Usage
          The L5 operator is meant to demonstrate a minimum viable prototype of all capability levels outilned in the operator framework. The L5 operator is compatible with both Kubernetes and OpenShift clusters. When running on vanilla kubernetes a few prerequisites need to be kept in mind which may not be available out of the box. These include an ingress controller and a prometheus operator for level 4 capability. The L5 operator also depends on the Crunchy Data PostgreSQL operator. On openshift when installing via the Operator Lifecycle Manager, this will automatically be installed when installing the operator via the embdeded operator hub. The Operator Lifecycle Manager is also available for vanilla kubernetes however it may not be installed out of the box. In this case the Crunchy Data PostgreSQL operator must be installed manually. The Crunchy Data PostgreSQL operator can be installed on vanilla kuberentes using the [instructions provided by Crunchy Data](https://access.crunchydata.com/documentation/postgres-operator/4.7.5/installation/postgres-operator/) More documentation and examples can be found on [github](https://github.com/opdev/l5-operator-demo).
        displayName: L5 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - l5-operator
        links:
        - name: L5 Operator
          url: https://opdev.github.io/l5-operator-demo/
        maintainers:
        - email: partnereng@redhat.com
          name: OpDev Team
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat Internal
          url: redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/opdev/l5-operator:v0.0.3
        version: 0.0.3
      entries:
      - name: l5-operator.v0.0.3
        version: 0.0.3
      - name: l5-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: l5-operator
    provider:
      name: Red Hat Internal
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
    name: aikit-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aikit-operator.v2021.2.102120
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aikit.intel/v1alpha1",
                "kind": "AIKitContainer",
                "metadata": {
                  "name": "intel-aikit-container",
                  "namespace": "redhat-ods-applications"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagestream": {
                    "name": "oneapi-aikit",
                    "namespace": "redhat-ods-applications",
                    "registry": {
                      "root": "registry.connect.redhat.com",
                      "repo": "intel",
                      "name": "oneapi-aikit",
                      "version": "latest"
                    }
                  },
                  "nameOverride": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AIKitContainer
            name: aikitcontainers.aikit.intel
            version: v1alpha1
        description: |
          The Intel® oneAPI AI Analytics Toolkit gives data scientists, AI developers, and researchers familiar Python* tools and frameworks to accelerate end-to-end data science and machine learning pipelines on Intel® architectures. The components are built using oneAPI libraries for low-level compute optimizations. This toolkit maximizes performance from preprocessing through machine and deep learning phases and provides interoperability for efficient model development and deployment across single and multinodes.
        displayName: Intel® oneAPI AI Analytics Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aikit-operator
        links:
        - name: Aikit Operator
          url: https://aikit-operator.domain
        maintainers:
        - email: kam.d.kasravi@intel.com
          name: Kam Kasravi
        - email: abolfazl.shahbazi@intel.com
          name: Abolfazl Shahbazi
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Intel
          url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
        relatedImages:
        - registry.connect.redhat.com/intel/aikit-operator@sha256:7abb5e6164ffb3e4b87c39061f45d09848218fcfb74b48bf5e1035d56b41ee0c
        version: 2021.2.102120
      entries:
      - name: aikit-operator.v2021.2.102120
        version: 2021.2.102120
      - name: aikit-operator.v2021.2.100410
        version: 2021.2.100410
      - name: aikit-operator.v2021.2.92911
        version: 2021.2.92911
      - name: aikit-operator.v2021.2.92814
        version: 2021.2.92814
      - name: aikit-operator.v2021.2.92808-dev-a60e5ba
        version: 2021.2.92808-dev-a60e5ba
      name: alpha
    defaultChannel: alpha
    packageName: aikit-operator
    provider:
      name: Intel
      url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-airflow-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airflow-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:d8f8c439f33e5a80e16c7867a008d71693fd4928357ac6919c77bd0ff0bb1339
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:d8f8c439f33e5a80e16c7867a008d71693fd4928357ac6919c77bd0ff0bb1339
        version: 24.11.1
      entries:
      - name: airflow-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: airflow-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:173363db3cb0997aff09fe4b7f3a54c04f22bc56db83d2b30757e3cc745c4587
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:173363db3cb0997aff09fe4b7f3a54c04f22bc56db83d2b30757e3cc745c4587
        version: 25.3.0
      entries:
      - name: airflow-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: airflow-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
        version: 25.7.0
      entries:
      - name: airflow-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: airflow-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
        version: 25.7.0
      entries:
      - name: airflow-operator.v25.7.0
        version: 25.7.0
      - name: airflow-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-airflow-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: windows-machine-config-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: windows-machine-config-operator.v10.16.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          createdAt: "2025-06-06T22:08:07Z"
          description: An operator that enables Windows container workloads on OCP
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=10.15.0 <10.16.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.openshift.io/valid-subscription: '["Red Hat OpenShift support
            for Windows Containers"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ### Introduction
          The Windows Machine Config Operator configures Windows Machines into nodes, enabling Windows container workloads to
          be run on OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/4.16/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://docs.openshift.com/container-platform/4.16/windows_containers/byoh-windows-instance.html)
          The operator completes all the necessary steps to configure the Windows instance so that it can join the cluster as a worker node.

          Usage of this operator requires an OpenShift Container Platform for Windows Containers subscription. Users looking to
          deploy Windows containers workloads in production clusters should acquire a subscription before attempting to
          install this operator. Users without a subscription can try the community operator, a distribution which lacks
          official support.

          ### Pre-requisites
          * A Red Hat OpenShift Container Platform for Windows Containers [subscription](https://access.redhat.com/support/policy/updates/openshift#windows)
          * OCP 4.16 cluster running on Azure, AWS, GCP or vSphere configured with hybrid OVN Kubernetes networking
          * [WMCO prerequisites](https://docs.openshift.com/container-platform/4.16/windows_containers/understanding-windows-container-workloads.html#wmco-prerequisites__understanding-windows-container-workloads)

          ### Usage
          Once the `openshift-windows-machine-config-operator` namespace has been created, a secret must be created containing
          the private key that will be used to access the Windows VMs:
          ```
          # Create secret containing the private key in the openshift-windows-machine-config-operator namespace
          oc create secret generic cloud-private-key --from-file=private-key.pem=/path/to/key -n openshift-windows-machine-config-operator
          ```
          We strongly recommend not using the same
          [private key](https://docs.openshift.com/container-platform/4.16/installing/installing_azure/installing-azure-default.html#ssh-agent-using_installing-azure-default)
          used when installing the cluster

          Below is an example of a vSphere Windows MachineSet which can create Windows Machines that the WMCO can react upon.
          Please note that the windows-user-data secret will be created by the WMCO lazily when it is configuring the first
          Windows Machine. After that, the windows-user-data will be available for the subsequent MachineSets to be consumed.
          It might take around 10 minutes for the Windows VM to be configured so that it joins the cluster. Please note that
          the MachineSet should have following labels:
          * `machine.openshift.io/os-id: Windows`
          * `machine.openshift.io/cluster-api-machine-role: worker`
          * `machine.openshift.io/cluster-api-machine-type: worker`

          The following label must be added to the Machine spec within the MachineSet spec:
          * `node-role.kubernetes.io/worker: ""`

          Not having these labels will result in the Windows node not being marked as a worker.

          `<infrastructureID>` should be replaced with the output of:
          ```
          oc get -o jsonpath='{.status.infrastructureName}{"\n"}' infrastructure cluster
          ```

          The following template variables need to be replaced as follows with values from your vSphere environment:
          * `<Windows_VM_template>`: template name
          * `<VM Network Name>`: network name
          * `<vCenter DataCenter Name>`: datacenter name
          * `<Path to VM Folder in vCenter>`: path where your OpenShift cluster is running
          * `<vCenter Datastore Name>`: datastore name
          * `<vCenter Server FQDN/IP>`: IP address or FQDN of the vCenter server

          Please note that on vSphere, Windows Machine names cannot be more than 15 characters long. The MachineSet name therefore
          cannot be more than 9 characters long, due to the way Machine names are generated from it.
          ```
          apiVersion: machine.openshift.io/v1beta1
          kind: MachineSet
          metadata:
            labels:
              machine.openshift.io/cluster-api-cluster: <infrastructureID>
            name: winworker
            namespace: openshift-machine-api
          spec:
            replicas: 1
            selector:
              matchLabels:
                machine.openshift.io/cluster-api-cluster: <infrastructureID>
                machine.openshift.io/cluster-api-machineset: winworker
            template:
              metadata:
                labels:
                  machine.openshift.io/cluster-api-cluster: <infrastructureID>
                  machine.openshift.io/cluster-api-machine-role: worker
                  machine.openshift.io/cluster-api-machine-type: worker
                  machine.openshift.io/cluster-api-machineset: winworker
                  machine.openshift.io/os-id: Windows
              spec:
                metadata:
                  labels:
                    node-role.kubernetes.io/worker: ""
                providerSpec:
                  value:
                    apiVersion: vsphereprovider.openshift.io/v1beta1
                    credentialsSecret:
                      name: vsphere-cloud-credentials
                      namespace: openshift-machine-api
                    diskGiB: 128
                    kind: VSphereMachineProviderSpec
                    memoryMiB: 16384
                    metadata:
                      creationTimestamp: null
                    network:
                      devices:
                      - networkName: "<VM Network Name>"
                    numCPUs: 4
                    numCoresPerSocket: 1
                    snapshot: ""
                    template: <Windows_VM_template
                    userDataSecret:
                      name: windows-user-data
                    workspace:
                      datacenter: <vCenter DataCenter Name>
                      datastore: <vCenter Datastore Name>
                      folder: <Path to VM Folder in vCenter> # e.g. /DC/vm/ocp45-2tdrm
                      server: <vCenter Server FQDN/IP>
          ```
          Example MachineSet for other cloud providers:
          - [AWS](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-aws.html)
          - [Azure](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-azure.html)
          - [GCP](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-gcp.html)

          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.

          ### Reporting issues
          Support for this distribution of WMCO requires a Red Hat OpenShfit subscription. Support should be requested through the Red Hat Customer Portal.
          Please read through the [troubleshooting document](https://docs.openshift.com/container-platform/4.16/support/troubleshooting/troubleshooting-windows-container-workload-issues.html)
          before opening a support case.
        displayName: Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4-wincw/windows-machine-config-rhel9-operator@sha256:c9a6166158e1ccb1764007d79c6e0f879505c3546eb45cbbe79ebeefcc78a728
        version: 10.16.2
      entries:
      - name: windows-machine-config-operator.v10.16.2
        version: 10.16.2
      - name: windows-machine-config-operator.v10.16.1
        version: 10.16.1
      - name: windows-machine-config-operator.v10.16.0
        version: 10.16.0
      name: preview
    - currentCSV: windows-machine-config-operator.v10.16.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          createdAt: "2025-06-06T22:08:07Z"
          description: An operator that enables Windows container workloads on OCP
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=10.15.0 <10.16.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.openshift.io/valid-subscription: '["Red Hat OpenShift support
            for Windows Containers"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ### Introduction
          The Windows Machine Config Operator configures Windows Machines into nodes, enabling Windows container workloads to
          be run on OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/4.16/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://docs.openshift.com/container-platform/4.16/windows_containers/byoh-windows-instance.html)
          The operator completes all the necessary steps to configure the Windows instance so that it can join the cluster as a worker node.

          Usage of this operator requires an OpenShift Container Platform for Windows Containers subscription. Users looking to
          deploy Windows containers workloads in production clusters should acquire a subscription before attempting to
          install this operator. Users without a subscription can try the community operator, a distribution which lacks
          official support.

          ### Pre-requisites
          * A Red Hat OpenShift Container Platform for Windows Containers [subscription](https://access.redhat.com/support/policy/updates/openshift#windows)
          * OCP 4.16 cluster running on Azure, AWS, GCP or vSphere configured with hybrid OVN Kubernetes networking
          * [WMCO prerequisites](https://docs.openshift.com/container-platform/4.16/windows_containers/understanding-windows-container-workloads.html#wmco-prerequisites__understanding-windows-container-workloads)

          ### Usage
          Once the `openshift-windows-machine-config-operator` namespace has been created, a secret must be created containing
          the private key that will be used to access the Windows VMs:
          ```
          # Create secret containing the private key in the openshift-windows-machine-config-operator namespace
          oc create secret generic cloud-private-key --from-file=private-key.pem=/path/to/key -n openshift-windows-machine-config-operator
          ```
          We strongly recommend not using the same
          [private key](https://docs.openshift.com/container-platform/4.16/installing/installing_azure/installing-azure-default.html#ssh-agent-using_installing-azure-default)
          used when installing the cluster

          Below is an example of a vSphere Windows MachineSet which can create Windows Machines that the WMCO can react upon.
          Please note that the windows-user-data secret will be created by the WMCO lazily when it is configuring the first
          Windows Machine. After that, the windows-user-data will be available for the subsequent MachineSets to be consumed.
          It might take around 10 minutes for the Windows VM to be configured so that it joins the cluster. Please note that
          the MachineSet should have following labels:
          * `machine.openshift.io/os-id: Windows`
          * `machine.openshift.io/cluster-api-machine-role: worker`
          * `machine.openshift.io/cluster-api-machine-type: worker`

          The following label must be added to the Machine spec within the MachineSet spec:
          * `node-role.kubernetes.io/worker: ""`

          Not having these labels will result in the Windows node not being marked as a worker.

          `<infrastructureID>` should be replaced with the output of:
          ```
          oc get -o jsonpath='{.status.infrastructureName}{"\n"}' infrastructure cluster
          ```

          The following template variables need to be replaced as follows with values from your vSphere environment:
          * `<Windows_VM_template>`: template name
          * `<VM Network Name>`: network name
          * `<vCenter DataCenter Name>`: datacenter name
          * `<Path to VM Folder in vCenter>`: path where your OpenShift cluster is running
          * `<vCenter Datastore Name>`: datastore name
          * `<vCenter Server FQDN/IP>`: IP address or FQDN of the vCenter server

          Please note that on vSphere, Windows Machine names cannot be more than 15 characters long. The MachineSet name therefore
          cannot be more than 9 characters long, due to the way Machine names are generated from it.
          ```
          apiVersion: machine.openshift.io/v1beta1
          kind: MachineSet
          metadata:
            labels:
              machine.openshift.io/cluster-api-cluster: <infrastructureID>
            name: winworker
            namespace: openshift-machine-api
          spec:
            replicas: 1
            selector:
              matchLabels:
                machine.openshift.io/cluster-api-cluster: <infrastructureID>
                machine.openshift.io/cluster-api-machineset: winworker
            template:
              metadata:
                labels:
                  machine.openshift.io/cluster-api-cluster: <infrastructureID>
                  machine.openshift.io/cluster-api-machine-role: worker
                  machine.openshift.io/cluster-api-machine-type: worker
                  machine.openshift.io/cluster-api-machineset: winworker
                  machine.openshift.io/os-id: Windows
              spec:
                metadata:
                  labels:
                    node-role.kubernetes.io/worker: ""
                providerSpec:
                  value:
                    apiVersion: vsphereprovider.openshift.io/v1beta1
                    credentialsSecret:
                      name: vsphere-cloud-credentials
                      namespace: openshift-machine-api
                    diskGiB: 128
                    kind: VSphereMachineProviderSpec
                    memoryMiB: 16384
                    metadata:
                      creationTimestamp: null
                    network:
                      devices:
                      - networkName: "<VM Network Name>"
                    numCPUs: 4
                    numCoresPerSocket: 1
                    snapshot: ""
                    template: <Windows_VM_template
                    userDataSecret:
                      name: windows-user-data
                    workspace:
                      datacenter: <vCenter DataCenter Name>
                      datastore: <vCenter Datastore Name>
                      folder: <Path to VM Folder in vCenter> # e.g. /DC/vm/ocp45-2tdrm
                      server: <vCenter Server FQDN/IP>
          ```
          Example MachineSet for other cloud providers:
          - [AWS](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-aws.html)
          - [Azure](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-azure.html)
          - [GCP](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-gcp.html)

          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.

          ### Reporting issues
          Support for this distribution of WMCO requires a Red Hat OpenShfit subscription. Support should be requested through the Red Hat Customer Portal.
          Please read through the [troubleshooting document](https://docs.openshift.com/container-platform/4.16/support/troubleshooting/troubleshooting-windows-container-workload-issues.html)
          before opening a support case.
        displayName: Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4-wincw/windows-machine-config-rhel9-operator@sha256:c9a6166158e1ccb1764007d79c6e0f879505c3546eb45cbbe79ebeefcc78a728
        version: 10.16.2
      entries:
      - name: windows-machine-config-operator.v10.16.2
        version: 10.16.2
      - name: windows-machine-config-operator.v10.16.1
        version: 10.16.1
      - name: windows-machine-config-operator.v10.16.0
        version: 10.16.0
      name: stable
    defaultChannel: stable
    packageName: windows-machine-config-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JohnSnowLABS
      provider-url: nlp.johnsnowlabs.com
    name: annotationlab
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: annotationlab.v5.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.johnsnowlabs.com/v1",
                "kind": "Annotationlab",
                "metadata": {
                  "name": "annotationlab"
                },
                "spec": {
                  "activelearning": {
                    "extraEnv": "",
                    "extraVolumeMounts": "",
                    "extraVolumes": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/johnsnowlabs/annotationlab",
                      "tag": "active-learning-5.7.0"
                    },
                    "lifecycleHooks": ""
                  },
                  "affinity": "podAntiAffinity:\n  requiredDuringSchedulingIgnoredDuringExecution:\n    - labelSelector:\n        matchLabels:\n          {{- include \"annotationlab.selectorLabels\" . | nindent 10 }}\n        matchExpressions:\n          - key: role\n            operator: NotIn\n            values:\n              - test\n      topologyKey: kubernetes.io/hostname\n  preferredDuringSchedulingIgnoredDuringExecution:\n    - weight: 100\n      podAffinityTerm:\n        labelSelector:\n          matchLabels:\n            {{- include \"annotationlab.selectorLabels\" . | nindent 12 }}\n          matchExpressions:\n            - key: role\n              operator: NotIn\n              values:\n                - test\n        topologyKey: failure-domain.beta.kubernetes.io/zone\n",
                  "airflow": {
                    "activelearning": {
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/johnsnowlabs/annotationlab",
                        "tag": "active-learning-5.7.0"
                      }
                    },
                    "affinity": {},
                    "airflowConfigAnnotations": {},
                    "airflowHome": "/opt/airflow",
                    "airflowLocalSettings": "{{- if semverCompare \"\u003e=2.2.0\" .Values.airflowVersion }}\n{{- if not (or .Values.webserverSecretKey .Values.webserverSecretKeySecretName) }}\nfrom airflow.www.utils import UIAlert\n\nDASHBOARD_UIALERTS = [\n  UIAlert(\n    'Usage of a dynamic webserver secret key detected. We recommend a static webserver secret key instead.'\n    ' See the \u003ca href='\n    '\"https://airflow.apache.org/docs/helm-chart/stable/production-guide.html#webserver-secret-key\"\u003e'\n    'Helm Chart Production Guide\u003c/a\u003e for more details.',\n    category=\"warning\",\n    roles=[\"Admin\"],\n    html=True,\n  )\n]\n{{- end }}\n{{- end }}",
                    "airflowPodAnnotations": {},
                    "airflowVersion": "2.5.1",
                    "allowPodLaunching": true,
                    "cleanup": {
                      "affinity": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow kubernetes cleanup-pods --namespace={{ .Release.Namespace }}"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "failedJobsHistoryLimit": null,
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "resources": {},
                      "schedule": "*/15 * * * *",
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "successfulJobsHistoryLimit": null,
                      "tolerations": [],
                      "topologySpreadConstraints": []
                    },
                    "config": {
                      "api": {
                        "auth_backend": "airflow.api.auth.backend.default"
                      },
                      "celery": {
                        "flower_url_prefix": "{{ .Values.ingress.flower.path }}",
                        "worker_concurrency": 16
                      },
                      "celery_kubernetes_executor": {
                        "kubernetes_queue": "kubernetes"
                      },
                      "core": {
                        "colored_console_log": "False",
                        "dags_are_paused_at_creation": "False",
                        "dags_folder": "{{ include \"airflow_dags\" . }}",
                        "executor": "{{ .Values.executor }}",
                        "load_examples": "False",
                        "remote_logging": "{{- ternary \"True\" \"False\" .Values.elasticsearch.enabled }}"
                      },
                      "elasticsearch": {
                        "json_format": "True",
                        "log_id_template": "{dag_id}_{task_id}_{execution_date}_{try_number}"
                      },
                      "elasticsearch_configs": {
                        "max_retries": 3,
                        "retry_timeout": "True",
                        "timeout": 30
                      },
                      "kerberos": {
                        "ccache": "{{ .Values.kerberos.ccacheMountPath }}/{{ .Values.kerberos.ccacheFileName }}",
                        "keytab": "{{ .Values.kerberos.keytabPath }}",
                        "principal": "{{ .Values.kerberos.principal }}",
                        "reinit_frequency": "{{ .Values.kerberos.reinitFrequency }}"
                      },
                      "kubernetes": {
                        "airflow_configmap": "{{ include \"airflow_config\" . }}",
                        "airflow_local_settings_configmap": "{{ include \"airflow_config\" . }}",
                        "multi_namespace_mode": "{{ ternary \"True\" \"False\" .Values.multiNamespaceMode }}",
                        "namespace": "{{ .Release.Namespace }}",
                        "pod_template_file": "{{ include \"airflow_pod_template_file\" . }}/pod_template_file.yaml",
                        "worker_container_repository": "{{ .Values.images.airflow.repository | default .Values.defaultAirflowRepository }}",
                        "worker_container_tag": "{{ .Values.images.airflow.tag | default .Values.defaultAirflowTag }}"
                      },
                      "logging": {
                        "colored_console_log": "False",
                        "remote_logging": "{{- ternary \"True\" \"False\" .Values.elasticsearch.enabled }}"
                      },
                      "metrics": {
                        "statsd_host": "{{ printf \"%s-statsd\" .Release.Name }}",
                        "statsd_on": "{{ ternary \"True\" \"False\" .Values.statsd.enabled }}",
                        "statsd_port": 9125,
                        "statsd_prefix": "airflow"
                      },
                      "scheduler": {
                        "run_duration": 41460,
                        "standalone_dag_processor": "{{ ternary \"True\" \"False\" .Values.dagProcessor.enabled }}",
                        "statsd_host": "{{ printf \"%s-statsd\" .Release.Name }}",
                        "statsd_on": "{{ ternary \"True\" \"False\" .Values.statsd.enabled }}",
                        "statsd_port": 9125,
                        "statsd_prefix": "airflow"
                      },
                      "webserver": {
                        "authenticate": "False",
                        "base_url": "http://0.0.0.0:8080",
                        "enable_proxy_fix": "True",
                        "rbac": "False"
                      }
                    },
                    "createUserJob": {
                      "affinity": {},
                      "annotations": {},
                      "applyCustomEnv": true,
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"users create\" \"create_user\" }} \"$@\"",
                        "--",
                        "-r",
                        "{{ .Values.webserver.defaultUser.role }}",
                        "-u",
                        "{{ .Values.webserver.defaultUser.username }}",
                        "-e",
                        "{{ .Values.webserver.defaultUser.email }}",
                        "-f",
                        "{{ .Values.webserver.defaultUser.firstName }}",
                        "-l",
                        "{{ .Values.webserver.defaultUser.lastName }}",
                        "-p",
                        "{{ .Values.webserver.defaultUser.password }}"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "jobAnnotations": {},
                      "labels": {},
                      "nodeSelector": {},
                      "resources": {},
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-create-user-job"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "useHelmHooks": true
                    },
                    "dagProcessor": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow dag-processor"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 60,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "dags": {
                      "gitSync": {
                        "branch": "v2-2-stable",
                        "containerName": "git-sync",
                        "depth": 1,
                        "enabled": false,
                        "env": [],
                        "extraVolumeMounts": [],
                        "maxFailures": 0,
                        "repo": "https://github.com/apache/airflow.git",
                        "resources": {},
                        "rev": "HEAD",
                        "securityContext": {},
                        "subPath": "tests/dags",
                        "uid": 65533,
                        "wait": 5
                      },
                      "persistence": {
                        "accessMode": "ReadWriteOnce",
                        "enabled": false,
                        "existingClaim": null,
                        "size": "1Gi",
                        "storageClassName": null,
                        "subPath": null
                      }
                    },
                    "data": {
                      "brokerUrl": null,
                      "brokerUrlSecretName": null,
                      "metadataConnection": {
                        "db": "airflow",
                        "pass": "airflow",
                        "port": 5432,
                        "protocol": "postgresql",
                        "sslmode": "disable",
                        "user": "airflow"
                      },
                      "metadataSecretName": null,
                      "resultBackendConnection": null,
                      "resultBackendSecretName": null
                    },
                    "defaultAirflowRepository": "apache/airflow",
                    "defaultAirflowTag": "2.5.1",
                    "elasticsearch": {
                      "connection": {},
                      "enabled": false,
                      "secretName": null
                    },
                    "enableBuiltInSecretEnvVars": {
                      "AIRFLOW_CONN_AIRFLOW_DB": true,
                      "AIRFLOW__CELERY__BROKER_URL": true,
                      "AIRFLOW__CELERY__CELERY_RESULT_BACKEND": true,
                      "AIRFLOW__CELERY__RESULT_BACKEND": true,
                      "AIRFLOW__CORE__FERNET_KEY": true,
                      "AIRFLOW__CORE__SQL_ALCHEMY_CONN": true,
                      "AIRFLOW__DATABASE__SQL_ALCHEMY_CONN": true,
                      "AIRFLOW__ELASTICSEARCH__ELASTICSEARCH_HOST": true,
                      "AIRFLOW__ELASTICSEARCH__HOST": true,
                      "AIRFLOW__WEBSERVER__SECRET_KEY": true
                    },
                    "env": [],
                    "executor": "CeleryExecutor",
                    "extraConfigMaps": {},
                    "extraEnv": "- name: ACTIVE_LEARNING_IMAGE_PULL_POLICY\n  value: \"{{ .Values.activelearning.image.pullPolicy }}\"\n- name: ACTIVE_LEARNING_IMAGE\n  value: \"{{ .Values.activelearning.image.repository }}:{{ .Values.activelearning.image.tag }}\"\n- name: NUMBER_OF_PREANNOTATOR_DAG\n  value: {{ .Values.preannotator_queues.count | quote }}\n- name: AIRFLOW__LOGGING__LOGGING_LEVEL\n  value: INFO\n- name: ANNOTATIONLAB_FULLNAME\n  value: {{ include \"annotationlab.annotationlab.fullname\" . }}\n- name: ANNOTATIONLAB_URL\n  value: http://{{ include \"annotationlab.annotationlab.fullname\" . }}:8200\n- name: AIRFLOW_NAMESPACE\n  value: {{ .Release.Namespace }}\n- name: DATABASE_CONNECTION_STRING\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"annotationlab.annotationlab.fullname\" . }}-secret\n      key: DATABASE_CONNECTION_STRING\n",
                    "extraEnvFrom": null,
                    "extraSecrets": {},
                    "fernetKey": null,
                    "fernetKeySecretName": null,
                    "flower": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"celery flower\" \"flower\" }}"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "networkPolicy": {
                        "ingress": {
                          "from": [],
                          "ports": [
                            {
                              "port": "{{ .Values.ports.flowerUI }}"
                            }
                          ]
                        }
                      },
                      "nodeSelector": {},
                      "password": null,
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "secretName": null,
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "loadBalancerIP": null,
                        "loadBalancerSourceRanges": [],
                        "ports": [
                          {
                            "name": "flower-ui",
                            "port": "{{ .Values.ports.flowerUI }}"
                          }
                        ],
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "username": null
                    },
                    "fullnameOverride": "",
                    "gid": 0,
                    "global": {},
                    "image": {
                      "repository": "docker.io/johnsnowlabs/annotationlab"
                    },
                    "images": {
                      "airflow": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": "dataflows-5.7.0"
                      },
                      "flower": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": null
                      },
                      "gitSync": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "k8s.gcr.io/git-sync/git-sync",
                        "tag": "v3.6.3"
                      },
                      "migrationsWaitTimeout": 60,
                      "pgbouncer": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "apache/airflow",
                        "tag": "airflow-pgbouncer-2021.04.28-1.14.0"
                      },
                      "pgbouncerExporter": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "apache/airflow",
                        "tag": "airflow-pgbouncer-exporter-2021.09.22-0.12.0"
                      },
                      "pod_template": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": null
                      },
                      "redis": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/redis",
                        "tag": "alpine3.16"
                      },
                      "statsd": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "quay.io/prometheus/statsd-exporter",
                        "tag": "v0.22.8"
                      },
                      "useDefaultImageForMigration": false
                    },
                    "ingress": {
                      "enabled": false,
                      "flower": {
                        "annotations": {},
                        "enabled": false,
                        "host": "",
                        "hosts": [],
                        "ingressClassName": "",
                        "path": "/",
                        "pathType": "ImplementationSpecific",
                        "tls": {
                          "enabled": false,
                          "secretName": ""
                        }
                      },
                      "web": {
                        "annotations": {},
                        "enabled": false,
                        "host": "",
                        "hosts": [],
                        "ingressClassName": "",
                        "path": "/",
                        "pathType": "ImplementationSpecific",
                        "precedingPaths": [],
                        "succeedingPaths": [],
                        "tls": {
                          "enabled": false,
                          "secretName": ""
                        }
                      }
                    },
                    "kerberos": {
                      "ccacheFileName": "cache",
                      "ccacheMountPath": "/var/kerberos-ccache",
                      "config": "# This is an example config showing how you can use templating and how \"example\" config\n# might look like. It works with the test kerberos server that we are using during integration\n# testing at Apache Airflow (see `scripts/ci/docker-compose/integration-kerberos.yml` but in\n# order to make it production-ready you must replace it with your own configuration that\n# Matches your kerberos deployment. Administrators of your Kerberos instance should\n# provide the right configuration.\n\n[logging]\ndefault = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kerberos_libs.log\"\nkdc = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kerberos_kdc.log\"\nadmin_server = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kadmind.log\"\n\n[libdefaults]\ndefault_realm = FOO.COM\nticket_lifetime = 10h\nrenew_lifetime = 7d\nforwardable = true\n\n[realms]\nFOO.COM = {\n  kdc = kdc-server.foo.com\n  admin_server = admin_server.foo.com\n}\n",
                      "configPath": "/etc/krb5.conf",
                      "enabled": false,
                      "keytabBase64Content": null,
                      "keytabPath": "/etc/airflow.keytab",
                      "principal": "airflow@FOO.COM",
                      "reinitFrequency": 3600
                    },
                    "kubeVersionOverride": "",
                    "labels": {},
                    "limits": [],
                    "logs": {
                      "persistence": {
                        "enabled": false,
                        "existingClaim": null,
                        "size": "100Gi",
                        "storageClassName": null
                      }
                    },
                    "migrateDatabaseJob": {
                      "affinity": {},
                      "annotations": {},
                      "applyCustomEnv": true,
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"db upgrade\" \"upgradedb\" }}"
                      ],
                      "command": null,
                      "enabled": true,
                      "extraContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "jobAnnotations": {},
                      "nodeSelector": {},
                      "resources": {},
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-migrate-database-job"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "useHelmHooks": true
                    },
                    "multiNamespaceMode": false,
                    "nameOverride": "",
                    "networkPolicies": {
                      "enabled": false
                    },
                    "nodeSelector": {},
                    "pgbouncer": {
                      "affinity": {},
                      "annotations": {},
                      "args": null,
                      "auth_file": "/etc/pgbouncer/users.txt",
                      "auth_type": "md5",
                      "ciphers": "normal",
                      "command": [
                        "pgbouncer",
                        "-u",
                        "nobody",
                        "/etc/pgbouncer/pgbouncer.ini"
                      ],
                      "configSecretName": null,
                      "enabled": false,
                      "extraIni": null,
                      "extraIniMetadata": null,
                      "extraIniResultBackend": null,
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "logConnections": 0,
                      "logDisconnections": 0,
                      "maxClientConn": 100,
                      "metadataPoolSize": 10,
                      "metricsExporterSidecar": {
                        "resources": {},
                        "sslmode": "disable"
                      },
                      "nodeSelector": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "resultBackendPoolSize": 5,
                      "revisionHistoryLimit": null,
                      "service": {
                        "extraAnnotations": {}
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "ssl": {
                        "ca": null,
                        "cert": null,
                        "key": null
                      },
                      "sslmode": "prefer",
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 65534,
                      "verbose": 0
                    },
                    "podTemplate": null,
                    "ports": {
                      "airflowUI": 8080,
                      "flowerUI": 5555,
                      "pgbouncer": 6543,
                      "pgbouncerScrape": 9127,
                      "redisDB": 6379,
                      "statsdIngest": 9125,
                      "statsdScrape": 9102,
                      "workerLogs": 8793
                    },
                    "postgresql": {
                      "auth": {
                        "enablePostgresUser": true,
                        "password": null,
                        "postgresPassword": "postgres",
                        "username": null
                      },
                      "common": {
                        "exampleValue": "common-chart",
                        "global": {
                          "postgresql": {}
                        }
                      },
                      "commonAnnotations": {},
                      "enabled": true,
                      "extraEnv": [],
                      "global": {
                        "postgresql": {}
                      },
                      "image": {
                        "debug": false,
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgresql",
                        "tag": "11"
                      },
                      "ldap": {
                        "baseDN": "",
                        "bindDN": "",
                        "enabled": false,
                        "port": "",
                        "prefix": "",
                        "scheme": "",
                        "search_attr": "",
                        "search_filter": "",
                        "server": "",
                        "suffix": "",
                        "tls": false,
                        "url": ""
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 30,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "master": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": [],
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "metrics": {
                        "enabled": false,
                        "extraEnvVars": {},
                        "image": {
                          "pullPolicy": "IfNotPresent",
                          "registry": "docker.io",
                          "repository": "bitnami/postgres-exporter",
                          "tag": "0.8.0-debian-10-r166"
                        },
                        "livenessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "prometheusRule": {
                          "additionalLabels": {},
                          "enabled": false,
                          "namespace": "",
                          "rules": []
                        },
                        "readinessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "securityContext": {},
                        "service": {
                          "annotations": {
                            "prometheus.io/port": "9187",
                            "prometheus.io/scrape": "true"
                          },
                          "type": "ClusterIP"
                        },
                        "serviceMonitor": {
                          "additionalLabels": {},
                          "enabled": false
                        }
                      },
                      "nameOverride": "airflow-postgresql",
                      "networkPolicy": {
                        "allowExternal": true,
                        "enabled": false,
                        "explicitNamespacesSelector": {}
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "enabled": true,
                        "mountPath": "/bitnami/postgresql",
                        "size": "8Gi",
                        "subPath": ""
                      },
                      "postgresqlDataDir": "/bitnami/postgresql/data",
                      "postgresqlDatabase": "airflow",
                      "postgresqlPassword": "airflow",
                      "postgresqlUsername": "airflow",
                      "psp": {
                        "create": false
                      },
                      "rbac": {
                        "create": false
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "replication": {
                        "applicationName": "my_application",
                        "enabled": false,
                        "numSynchronousReplicas": 0,
                        "password": "repl_password",
                        "slaveReplicas": 1,
                        "synchronousCommit": "off",
                        "user": "repl_user"
                      },
                      "resources": {
                        "requests": {
                          "cpu": "250m",
                          "memory": "256Mi"
                        }
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "port": 5432,
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "enabled": true,
                        "name": "annotationlab-airflow-postgresql"
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": true
                        },
                        "enabled": true
                      },
                      "slave": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "tls": {
                        "certFilename": "",
                        "certKeyFilename": "",
                        "certificatesSecret": "",
                        "enabled": false,
                        "preferServerCiphers": true
                      },
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "volumePermissions": {
                        "enabled": false,
                        "image": {
                          "pullPolicy": "Always",
                          "registry": "docker.io",
                          "repository": "bitnami/minideb",
                          "tag": "buster"
                        },
                        "securityContext": {}
                      }
                    },
                    "preannotator_queues": {
                      "count": 5
                    },
                    "quotas": {},
                    "rbac": {
                      "create": true,
                      "createSCCRoleBinding": false
                    },
                    "redis": {
                      "affinity": {},
                      "enabled": true,
                      "nodeSelector": {},
                      "password": "DefaulTRediS-P0ss",
                      "passwordSecretName": null,
                      "persistence": {
                        "annotations": {},
                        "enabled": true,
                        "size": "1Gi",
                        "storageClassName": null
                      },
                      "podAnnotations": {},
                      "resources": {},
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "annotationlab-airflow-redis"
                      },
                      "terminationGracePeriodSeconds": 600,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 0
                    },
                    "registry": {
                      "connection": {},
                      "secretName": "kubernetesdockerhub"
                    },
                    "registryCredentials": {
                      "pullPolicy": "IfNotPresent",
                      "registryUrl": "docker.io/annotationlab"
                    },
                    "revisionHistoryLimit": null,
                    "scheduler": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow scheduler"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "logGroomerSidecar": {
                        "args": [
                          "bash",
                          "/clean-logs"
                        ],
                        "command": null,
                        "enabled": true,
                        "resources": {},
                        "retentionDays": 15
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-scheduler"
                      },
                      "strategy": null,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "updateStrategy": null,
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "secret": [],
                    "securityContext": {},
                    "statsd": {
                      "affinity": {},
                      "args": [
                        "--statsd.mapping-config=/etc/statsd-exporter/mappings.yml"
                      ],
                      "enabled": false,
                      "extraMappings": [],
                      "extraNetworkPolicies": [],
                      "nodeSelector": {},
                      "overrideMappings": [],
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "securityContext": {},
                      "service": {
                        "extraAnnotations": {}
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-statsd"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 65534
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "triggerer": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow triggerer"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "annotationlab-airflow-triggerer"
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 60,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "uid": 50000,
                    "volumeMounts": [],
                    "volumes": [],
                    "webserver": {
                      "affinity": {},
                      "allowPodLogReading": true,
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow webserver"
                      ],
                      "command": null,
                      "defaultUser": {
                        "email": "admin@example.com",
                        "enabled": false,
                        "firstName": "admin",
                        "lastName": "user",
                        "password": "admin",
                        "role": "Admin",
                        "username": "admin"
                      },
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "failureThreshold": 20,
                        "initialDelaySeconds": 15,
                        "periodSeconds": 5,
                        "scheme": "HTTP",
                        "timeoutSeconds": 30
                      },
                      "networkPolicy": {
                        "ingress": {
                          "from": [],
                          "ports": [
                            {
                              "port": "{{ .Values.ports.airflowUI }}"
                            }
                          ]
                        }
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "readinessProbe": {
                        "failureThreshold": 20,
                        "initialDelaySeconds": 15,
                        "periodSeconds": 5,
                        "scheme": "HTTP",
                        "timeoutSeconds": 30
                      },
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "loadBalancerIP": null,
                        "loadBalancerSourceRanges": [],
                        "ports": [
                          {
                            "name": "airflow-ui",
                            "port": "{{ .Values.ports.airflowUI }}"
                          }
                        ],
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-webserver"
                      },
                      "strategy": null,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      },
                      "webserverConfig": null,
                      "webserverConfigConfigMapName": null
                    },
                    "webserverSecretKey": "f565cd64401775f93c820dc63fc3d4f6",
                    "webserverSecretKeySecretName": null,
                    "workers": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"celery worker\" \"worker\" }}"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "hostAliases": [],
                      "keda": {
                        "advanced": {},
                        "cooldownPeriod": 30,
                        "enabled": false,
                        "maxReplicaCount": 10,
                        "minReplicaCount": 0,
                        "namespaceLabels": {},
                        "pollingInterval": 5
                      },
                      "kerberosSidecar": {
                        "enabled": false,
                        "resources": {}
                      },
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "enabled": true,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "logGroomerSidecar": {
                        "args": [
                          "bash",
                          "/clean-logs"
                        ],
                        "command": null,
                        "enabled": true,
                        "resources": {},
                        "retentionDays": 15
                      },
                      "nodeSelector": {},
                      "persistence": {
                        "annotations": {},
                        "enabled": true,
                        "fixPermissions": false,
                        "size": "5Gi",
                        "storageClassName": null
                      },
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-worker"
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 600,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "updateStrategy": null,
                      "waitForMigrations": {
                        "env": []
                      }
                    }
                  },
                  "backup": {
                    "enable": false,
                    "files": true,
                    "image": {
                      "repository": null,
                      "tag": null
                    },
                    "s3_access_key": "",
                    "s3_bucket_fullpath": "",
                    "s3_secret_key": ""
                  },
                  "configuration": {
                    "FLASK_SECRET_KEY": "rpaQghA-9lmpLwB8qxW0FsW1TJyZZYuB60ppzswpTzM=",
                    "KEYCLOAK_CLIENT_ID": "annotationlab",
                    "KEYCLOAK_CLIENT_SECRET_KEY": "09a71c59-0351-4ce6-bc8f-8fd3feb9d2ff",
                    "KEYCLOAK_REALM_NAME": "master",
                    "PERSISTENT_STORAGE": "/projects"
                  },
                  "extraContainers": "",
                  "extraEnv": "",
                  "extraNetworkPolicies": "- namespaceSelector: {}\n",
                  "extraVolumeMounts": "",
                  "extraVolumes": "",
                  "github_oauth_origin": "https://github.oauth.johnsnowlabs.com",
                  "google_analytics": {
                    "api_secret": "WENfd00yeUxRRWl4N0Utbml5SWt3Zw==",
                    "measurement_id": "Ry1XODBIMDhTTFBZ"
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "docker.io/johnsnowlabs/annotationlab",
                    "tag": "5.7.0"
                  },
                  "ingress": {
                    "annotations": {},
                    "defaultBackend": false,
                    "enabled": true,
                    "hosts": [
                      {
                        "host": "annotationlab",
                        "path": "/"
                      }
                    ],
                    "ingressroute": {
                      "enabled": false
                    },
                    "servicePort": "http",
                    "tls": [],
                    "uploadLimitInMegabytes": 16
                  },
                  "installAirflow": true,
                  "installKeycloak": true,
                  "installPostgresql": true,
                  "keycloak": {
                    "affinity": "podAntiAffinity:\n  requiredDuringSchedulingIgnoredDuringExecution:\n    - labelSelector:\n        matchLabels:\n          {{- include \"keycloak.selectorLabels\" . | nindent 10 }}\n        matchExpressions:\n          - key: app.kubernetes.io/component\n            operator: NotIn\n            values:\n              - test\n      topologyKey: kubernetes.io/hostname\n  preferredDuringSchedulingIgnoredDuringExecution:\n    - weight: 100\n      podAffinityTerm:\n        labelSelector:\n          matchLabels:\n            {{- include \"keycloak.selectorLabels\" . | nindent 12 }}\n          matchExpressions:\n            - key: app.kubernetes.io/component\n              operator: NotIn\n              values:\n                - test\n        topologyKey: topology.kubernetes.io/zone\n",
                    "args": [],
                    "auththeme": {
                      "image": {
                        "repository": "docker.io/johnsnowlabs/annotationlab",
                        "tag": "auth-theme-5.7.0"
                      }
                    },
                    "autoscaling": {
                      "behavior": {
                        "scaleDown": {
                          "policies": [
                            {
                              "periodSeconds": 300,
                              "type": "Pods",
                              "value": 1
                            }
                          ],
                          "stabilizationWindowSeconds": 300
                        }
                      },
                      "enabled": false,
                      "labels": {},
                      "maxReplicas": 10,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 80,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 3
                    },
                    "cache": {
                      "stack": "default"
                    },
                    "clusterDomain": "cluster.local",
                    "command": [
                      "/opt/keycloak/bin/kc.sh",
                      "--verbose",
                      "start",
                      "--http-enabled=true",
                      "--http-port=8080",
                      "--hostname-strict=false",
                      "--hostname-strict-https=false",
                      "--spi-events-listener-jboss-logging-success-level=info",
                      "--spi-events-listener-jboss-logging-error-level=warn"
                    ],
                    "database": {
                      "enabled": true,
                      "postgresqlDatabase": "keycloak",
                      "postgresqlPassword": "keycloak",
                      "postgresqlUsername": "keycloak"
                    },
                    "dbchecker": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/busybox",
                        "tag": 1.32
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "32Mi"
                        },
                        "requests": {
                          "cpu": "20m",
                          "memory": "32Mi"
                        }
                      },
                      "securityContext": {}
                    },
                    "enableServiceLinks": true,
                    "extraContainers": "",
                    "extraEnv": "- name: KC_FEATURES\n  value: \"authorization,account-api,admin-api,admin-fine-grained-authz,ciba,impersonation,kerberos,preview,step-up-authentication,token-exchange,web-authn\"\n- name: KEYCLOAK_ADMIN\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"keycloak.fullname\" . }}-admincreds\n      key: user\n- name: KEYCLOAK_ADMIN_PASSWORD\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"keycloak.fullname\" . }}-admincreds\n      key: password\n- name: JAVA_OPTS_APPEND\n  value: \u003e-\n    -XX:+UseContainerSupport\n    -XX:MaxRAMPercentage=50.0\n    -Djava.awt.headless=true\n    -Djgroups.dns.query={{ include \"keycloak.fullname\" . }}-headless\n",
                    "extraEnvFrom": "",
                    "extraInitContainers": "- name: theme-provider\n  image: {{ .Values.auththeme.image.repository }}:{{ .Values.auththeme.image.tag }}\n  imagePullPolicy: IfNotPresent\n  command:\n    - sh\n  args:\n    - -c\n    - |\n      echo \"Copying theme...\"\n      cp -R /johnsnowlabs/* /theme\n  volumeMounts:\n    - name: theme\n      mountPath: /theme\n",
                    "extraPorts": [],
                    "extraServiceMonitor": {
                      "annotations": {},
                      "enabled": false,
                      "interval": "10s",
                      "labels": {},
                      "namespace": "",
                      "namespaceSelector": {},
                      "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/realms/master/metrics",
                      "port": "http",
                      "scrapeTimeout": "10s"
                    },
                    "extraVolumeMounts": "- name: admincreds\n  mountPath: /secrets/admincreds\n  readOnly: true\n- name: theme\n  mountPath: /opt/keycloak/themes/johnsnowlabs\n",
                    "extraVolumes": "- name: admincreds\n  secret:\n    secretName: '{{ include \"annotationlab.keycloak.fullname\" . }}-admincreds'\n- name: theme\n  emptyDir: {}\n",
                    "fullnameOverride": "",
                    "global": {},
                    "health": {
                      "enabled": true
                    },
                    "hostAliases": [],
                    "http": {
                      "relativePath": "/auth"
                    },
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/keycloak/keycloak",
                      "tag": "20.0.3"
                    },
                    "imagePullSecrets": [
                      {
                        "name": "kubernetesdockerhub"
                      }
                    ],
                    "ingress": {
                      "annotations": {
                        "nginx.ingress.kubernetes.io/proxy-buffer-size": "16k"
                      },
                      "console": {
                        "annotations": {},
                        "enabled": false,
                        "ingressClassName": "",
                        "rules": [
                          {
                            "host": "{{ .Release.Name }}.keycloak.example.com",
                            "paths": [
                              {
                                "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/admin",
                                "pathType": "Prefix"
                              }
                            ]
                          }
                        ],
                        "tls": []
                      },
                      "enabled": false,
                      "ingressClassName": "",
                      "labels": {},
                      "rules": [
                        {
                          "host": "keycloak.example.com",
                          "paths": [
                            {
                              "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/",
                              "pathType": "Prefix"
                            }
                          ]
                        }
                      ],
                      "servicePort": "http",
                      "tls": []
                    },
                    "lifecycleHooks": "",
                    "livenessProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health/live'\n  port: http\ninitialDelaySeconds: 0\ntimeoutSeconds: 5\n",
                    "metrics": {
                      "enabled": true
                    },
                    "nameOverride": "",
                    "networkPolicy": {
                      "egress": [],
                      "enabled": false,
                      "extraFrom": [],
                      "labels": {}
                    },
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podDisruptionBudget": {},
                    "podLabels": {},
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {},
                    "postgresql": {
                      "common": {
                        "exampleValue": "common-chart",
                        "global": {
                          "postgresql": {}
                        }
                      },
                      "commonAnnotations": {},
                      "enabled": true,
                      "extraEnv": [],
                      "global": {
                        "postgresql": {}
                      },
                      "image": {
                        "debug": false,
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgresql",
                        "tag": "11.20.0-debian-11-r12"
                      },
                      "ldap": {
                        "baseDN": "",
                        "bindDN": "",
                        "bind_password": null,
                        "enabled": false,
                        "port": "",
                        "prefix": "",
                        "scheme": "",
                        "search_attr": "",
                        "search_filter": "",
                        "server": "",
                        "suffix": "",
                        "tls": false,
                        "url": ""
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 30,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "master": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": [],
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "metrics": {
                        "enabled": false,
                        "extraEnvVars": {},
                        "image": {
                          "pullPolicy": "IfNotPresent",
                          "registry": "docker.io",
                          "repository": "bitnami/postgres-exporter",
                          "tag": "0.8.0-debian-10-r166"
                        },
                        "livenessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "prometheusRule": {
                          "additionalLabels": {},
                          "enabled": false,
                          "namespace": "",
                          "rules": []
                        },
                        "readinessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "securityContext": {},
                        "service": {
                          "annotations": {
                            "prometheus.io/port": "9187",
                            "prometheus.io/scrape": "true"
                          },
                          "loadBalancerIP": null,
                          "type": "ClusterIP"
                        },
                        "serviceMonitor": {
                          "additionalLabels": {},
                          "enabled": false
                        }
                      },
                      "nameOverride": "keycloak-postgres",
                      "networkPolicy": {
                        "allowExternal": true,
                        "enabled": false,
                        "explicitNamespacesSelector": {}
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "enabled": true,
                        "mountPath": "/bitnami/postgresql",
                        "size": "8Gi",
                        "subPath": ""
                      },
                      "postgresqlDataDir": "/bitnami/postgresql/data",
                      "postgresqlDatabase": "keycloak",
                      "postgresqlPassword": "keycloak",
                      "postgresqlUsername": "keycloak",
                      "psp": {
                        "create": false
                      },
                      "rbac": {
                        "create": false
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "replication": {
                        "applicationName": "my_application",
                        "enabled": false,
                        "numSynchronousReplicas": 0,
                        "password": "repl_password",
                        "slaveReplicas": 1,
                        "synchronousCommit": "off",
                        "user": "repl_user"
                      },
                      "resources": {
                        "requests": {
                          "cpu": "250m",
                          "memory": "256Mi"
                        }
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "port": 5432,
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "enabled": true,
                        "name": "annotationlab-keycloak-postgresql"
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": true
                        },
                        "enabled": true
                      },
                      "slave": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "tls": {
                        "certCAFilename": null,
                        "certFilename": "",
                        "certKeyFilename": "",
                        "certificatesSecret": "",
                        "crlFilename": null,
                        "enabled": false,
                        "preferServerCiphers": true
                      },
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "volumePermissions": {
                        "enabled": false,
                        "image": {
                          "pullPolicy": "Always",
                          "registry": "docker.io",
                          "repository": "bitnami/minideb",
                          "tag": "buster"
                        },
                        "securityContext": {}
                      }
                    },
                    "priorityClassName": "",
                    "prometheusRule": {
                      "annotations": {},
                      "enabled": false,
                      "labels": {},
                      "namespace": "",
                      "rules": []
                    },
                    "proxy": {
                      "enabled": true,
                      "mode": "edge"
                    },
                    "rbac": {
                      "create": false,
                      "rules": []
                    },
                    "readinessProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health/ready'\n  port: http\ninitialDelaySeconds: 10\ntimeoutSeconds: 1\n",
                    "replicas": 1,
                    "resources": {},
                    "restartPolicy": "Always",
                    "route": {
                      "annotations": {},
                      "enabled": false,
                      "host": "",
                      "labels": {},
                      "path": "/",
                      "tls": {
                        "enabled": true,
                        "insecureEdgeTerminationPolicy": "Redirect",
                        "termination": "edge"
                      }
                    },
                    "secrets": {
                      "admincreds": {
                        "stringData": {
                          "password": "secret",
                          "user": "admin"
                        }
                      }
                    },
                    "securityContext": {},
                    "service": {
                      "annotations": {},
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "httpNodePort": null,
                      "httpPort": 80,
                      "httpsNodePort": null,
                      "httpsPort": 8443,
                      "labels": {},
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "sessionAffinity": "",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "serviceAccount": {
                      "allowReadPods": false,
                      "annotations": {},
                      "automountServiceAccountToken": true,
                      "create": false,
                      "imagePullSecrets": [],
                      "labels": {},
                      "name": "annotationlab-keyclo"
                    },
                    "serviceHeadless": {
                      "annotations": {}
                    },
                    "serviceMonitor": {
                      "annotations": {},
                      "enabled": false,
                      "interval": "10s",
                      "labels": {},
                      "namespace": "",
                      "namespaceSelector": {},
                      "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/metrics",
                      "port": "http",
                      "scrapeTimeout": "10s"
                    },
                    "skipInitContainers": false,
                    "startupProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health'\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 1\nfailureThreshold: 60\nperiodSeconds: 5\n",
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {},
                    "terminationGracePeriodSeconds": 60,
                    "test": {
                      "deletionPolicy": "before-hook-creation",
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/seleniarm/standalone-chromium",
                        "tag": "111.0"
                      },
                      "podSecurityContext": {},
                      "securityContext": {}
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": null,
                    "updateStrategy": "RollingUpdate"
                  },
                  "licenseserver": {
                    "origin": "https://license.johnsnowlabs.com"
                  },
                  "lifecycleHooks": "# postStart:\n#   exec:\n#     command: [\"/bin/sh\", \"-c\", \"ls\"]\n",
                  "modelshub_origin": "https://modelshub.johnsnowlabs.com",
                  "myjohnsnowlabs": {
                    "client_id": "UIGgoiIL9szXW8Nnik0O7tPT4to87oVtWvYWfc6r",
                    "origin": "https://my.johnsnowlabs.com"
                  },
                  "networkPolicy": {
                    "enabled": false
                  },
                  "nodeSelector": {},
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "annotations": {},
                    "enabled": true,
                    "size": "8Gi"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "postgresql": {
                    "common": {
                      "exampleValue": "common-chart",
                      "global": {
                        "postgresql": {}
                      }
                    },
                    "commonAnnotations": {},
                    "extraEnv": [],
                    "global": {
                      "postgresql": {}
                    },
                    "image": {
                      "debug": false,
                      "pullPolicy": "IfNotPresent",
                      "registry": "docker.io",
                      "repository": "bitnami/postgresql",
                      "tag": "11.20.0-debian-11-r12"
                    },
                    "ldap": {
                      "baseDN": "",
                      "bindDN": "",
                      "bind_password": null,
                      "enabled": false,
                      "port": "",
                      "prefix": "",
                      "scheme": "",
                      "search_attr": "",
                      "search_filter": "",
                      "server": "",
                      "suffix": "",
                      "tls": false,
                      "url": ""
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "master": {
                      "affinity": {},
                      "annotations": {},
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podLabels": {},
                      "priorityClassName": "",
                      "service": {},
                      "sidecars": [],
                      "tolerations": []
                    },
                    "metrics": {
                      "enabled": false,
                      "extraEnvVars": {},
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgres-exporter",
                        "tag": "0.8.0-debian-10-r166"
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "prometheusRule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {
                          "prometheus.io/port": "9187",
                          "prometheus.io/scrape": "true"
                        },
                        "loadBalancerIP": null,
                        "type": "ClusterIP"
                      },
                      "serviceMonitor": {
                        "additionalLabels": {},
                        "enabled": false
                      }
                    },
                    "networkPolicy": {
                      "allowExternal": true,
                      "enabled": false,
                      "explicitNamespacesSelector": {}
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {},
                      "enabled": true,
                      "mountPath": "/bitnami/postgresql",
                      "size": "8Gi",
                      "subPath": ""
                    },
                    "postgresqlDataDir": "/bitnami/postgresql/data",
                    "postgresqlDatabase": "annotationlab",
                    "postgresqlPassword": "annotationlab",
                    "postgresqlUsername": "annotationlab",
                    "psp": {
                      "create": false
                    },
                    "rbac": {
                      "create": false
                    },
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 6,
                      "initialDelaySeconds": 5,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "replication": {
                      "applicationName": "my_application",
                      "enabled": false,
                      "numSynchronousReplicas": 0,
                      "password": "repl_password",
                      "slaveReplicas": 1,
                      "synchronousCommit": "off",
                      "user": "repl_user"
                    },
                    "resources": {
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {},
                    "service": {
                      "annotations": {},
                      "port": 5432,
                      "type": "ClusterIP"
                    },
                    "serviceAccount": {
                      "enabled": false
                    },
                    "shmVolume": {
                      "chmod": {
                        "enabled": true
                      },
                      "enabled": true
                    },
                    "slave": {
                      "affinity": {},
                      "annotations": {},
                      "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podLabels": {},
                      "priorityClassName": "",
                      "service": {},
                      "sidecars": [],
                      "tolerations": []
                    },
                    "tls": {
                      "certCAFilename": null,
                      "certFilename": "",
                      "certKeyFilename": "",
                      "certificatesSecret": "",
                      "crlFilename": null,
                      "enabled": false,
                      "preferServerCiphers": true
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "volumePermissions": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "Always",
                        "registry": "docker.io",
                        "repository": "bitnami/minideb",
                        "tag": "buster"
                      },
                      "securityContext": {}
                    }
                  },
                  "preannotator_queues": {
                    "count": 5,
                    "service": {
                      "port": 5000,
                      "portName": "http",
                      "type": "ClusterIP"
                    }
                  },
                  "priorityClassName": "",
                  "readinessProbe": "httpGet:\n  path: /healthz\n  port: http\ninitialDelaySeconds: 10\ntimeoutSeconds: 1\n",
                  "registryCredentials": {
                    "create": false,
                    "password": "XXXXX",
                    "secretName": "kubernetesdockerhub",
                    "username": "XXXXX"
                  },
                  "replicas": 1,
                  "resources": {},
                  "restartPolicy": "Always",
                  "securityContext": {
                    "enabled": false,
                    "fsGroup": 65534,
                    "fsGroupChangePolicy": "OnRootMismatch",
                    "runAsUser": 65534
                  },
                  "service": {
                    "port": 8200,
                    "portName": "http",
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": false,
                    "name": "annotationlab"
                  },
                  "sharedData": {
                    "enabled": true,
                    "images": {
                      "storageSize": "5Gi"
                    },
                    "storageSize": "100Gi"
                  },
                  "sqlalchemy_pool_size": 10,
                  "telemetry": {
                    "enabled": true
                  },
                  "terminationGracePeriodSeconds": 60,
                  "tolerations": [],
                  "updateStrategy": {
                    "type": "Recreate"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.johnsnowlabs.com/annotationlab/annotationlab-operator:v5.7.0
          createdAt: "2023-12-19T10:26:54Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/annotationlab/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/annotationlab/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: johnsnowlabs-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: NLP Lab
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Annotationlab
            name: annotationlabs.apps.johnsnowlabs.com
            version: v1
        description: "NLP Lab is a data labeling and annotation tool that helps you\nlabel
          different types of data using a simple interface with a standardized output\nformat.\n\nIt
          contains enterprise features like Project Management and sharing, SSO\nintegration,
          OpenAPI Specification 3.0 compliant API, task management, and active\nlearning/preannotation
          using Spark NLP.\n\nFurthermore, the production deployment of the application
          is done using a helm chart.\nThe same helm chart is also used to create
          an AWS AMI for an EC2 based single server \ndeployment using packer.\n\nMake
          sure you have defined shared storage such as efs/nfs/cephfs prior to installing
          the NLP Lab Operator.\nUpdate the host property in the YAML configuration
          to define the required domain name to use instead of the \ndefault hostname
          annotationlab.\n\nNLP Lab Operator Installtion guide [NLP Lab Operator][]:\n[NLP
          Lab Operator]: https://nlp.johnsnowlabs.com/docs/en/alab/install    \n"
        displayName: NLPLab
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nlplab
        links:
        - name: Annotationlab
          url: https://annotationlab.domain
        maintainers:
        - email: bhandari@johnsnowlabs.com
          name: bikash
        maturity: alpha
        provider:
          name: JohnSnowLABS
          url: nlp.johnsnowlabs.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - registry.johnsnowlabs.com/annotationlab/annotationlab-operator:v5.7.0
        version: 5.7.0
      entries:
      - name: annotationlab.v5.7.0
        version: 5.7.0
      - name: annotationlab.v5.4.1
        version: 5.4.1
      - name: annotationlab.v4.9.2
        version: 4.9.2
      name: alpha
    defaultChannel: alpha
    packageName: annotationlab
    provider:
      name: JohnSnowLABS
      url: nlp.johnsnowlabs.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NGINX Inc
      provider-url: ""
    name: nginx-ingress-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nginx-ingress-operator.v3.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.nginx.org/v1alpha1",
                "kind": "NginxIngress",
                "metadata": {
                  "name": "nginxingress-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "annotations": {},
                    "appprotect": {
                      "enable": false
                    },
                    "appprotectdos": {
                      "debug": false,
                      "enable": false,
                      "maxDaemons": 0,
                      "maxWorkers": 0,
                      "memory": 0
                    },
                    "autoscaling": {
                      "annotations": {},
                      "behavior": {},
                      "enabled": false,
                      "maxReplicas": 3,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 50,
                      "targetMemoryUtilizationPercentage": 50
                    },
                    "config": {
                      "annotations": {},
                      "entries": {}
                    },
                    "containerPort": {
                      "http": 80,
                      "https": 443
                    },
                    "customConfigMap": "",
                    "customPorts": [],
                    "defaultHTTPListenerPort": 80,
                    "defaultHTTPSListenerPort": 443,
                    "defaultTLS": {
                      "cert": "",
                      "key": "",
                      "secret": ""
                    },
                    "disableIPV6": false,
                    "dnsPolicy": "ClusterFirst",
                    "enableCertManager": false,
                    "enableCustomResources": true,
                    "enableExternalDNS": false,
                    "enableLatencyMetrics": false,
                    "enableOIDC": false,
                    "enableSSLDynamicReload": true,
                    "enableSnippets": false,
                    "enableTLSPassthrough": false,
                    "env": [],
                    "extraContainers": [],
                    "globalConfiguration": {
                      "create": false,
                      "spec": {}
                    },
                    "healthStatus": false,
                    "healthStatusURI": "/nginx-health",
                    "hostNetwork": false,
                    "hostPort": {
                      "enable": false,
                      "http": 80,
                      "https": 443
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "nginx/nginx-ingress",
                      "tag": "5.2.0-ubi"
                    },
                    "ingressClass": {
                      "create": true,
                      "name": "nginx",
                      "setAsDefaultIngress": false
                    },
                    "initContainerResources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "initContainers": [],
                    "kind": "deployment",
                    "lifecycle": {},
                    "logFormat": "glog",
                    "logLevel": "info",
                    "mgmt": {
                      "licenseTokenSecretName": "license-token"
                    },
                    "minReadySeconds": 0,
                    "name": "controller",
                    "nginxDebug": false,
                    "nginxReloadTimeout": 60000,
                    "nginxStatus": {
                      "allowCidrs": "127.0.0.1",
                      "enable": true,
                      "port": 8080
                    },
                    "nginxplus": false,
                    "pod": {
                      "annotations": {},
                      "extraLabels": {}
                    },
                    "podDisruptionBudget": {
                      "annotations": {},
                      "enabled": false
                    },
                    "readOnlyRootFilesystem": false,
                    "readyStatus": {
                      "enable": true,
                      "initialDelaySeconds": 0,
                      "port": 8081
                    },
                    "replicaCount": 1,
                    "reportIngressStatus": {
                      "annotations": {},
                      "enable": true,
                      "enableLeaderElection": true,
                      "ingressLink": "",
                      "leaderElectionLockName": "nginx-ingress-leader"
                    },
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "selectorLabels": {},
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "create": true,
                      "customPorts": [],
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraLabels": {},
                      "httpPort": {
                        "enable": true,
                        "port": 80,
                        "targetPort": 80
                      },
                      "httpsPort": {
                        "enable": true,
                        "port": 443,
                        "targetPort": 443
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "type": "LoadBalancer"
                    },
                    "serviceAccount": {
                      "annotations": {},
                      "imagePullSecretName": "",
                      "imagePullSecretsNames": []
                    },
                    "shareProcessNamespace": false,
                    "strategy": {},
                    "terminationGracePeriodSeconds": 30,
                    "tlsPassthroughPort": 443,
                    "tolerations": [],
                    "volumeMounts": [],
                    "volumes": [],
                    "watchNamespace": "",
                    "watchNamespaceLabel": "",
                    "watchSecretNamespace": "",
                    "wildcardTLS": {
                      "cert": "",
                      "key": "",
                      "secret": ""
                    }
                  },
                  "nginxServiceMesh": {
                    "enable": false,
                    "enableEgress": false
                  },
                  "prometheus": {
                    "create": true,
                    "port": 9113,
                    "scheme": "http",
                    "secret": "",
                    "service": {
                      "create": false,
                      "labels": {
                        "service": "nginx-ingress-prometheus-service"
                      }
                    },
                    "serviceMonitor": {
                      "create": false,
                      "endpoints": [
                        {
                          "port": "prometheus"
                        }
                      ],
                      "labels": {},
                      "selectorMatchLabels": {
                        "service": "nginx-ingress-prometheus-service"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "serviceInsight": {
                    "create": false,
                    "port": 9114,
                    "scheme": "http",
                    "secret": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking
          certified: "true"
          containerImage: quay.io/nginx/nginx-ingress-operator@sha256:fb372438634b4a667218e725091bd448a253e7a9c40e7dda1044a318780d995f
          createdAt: "2025-09-15T16:08:18Z"
          description: The NGINX Ingress Operator is a Kubernetes/OpenShift component
            which deploys and manages one or more NGINX/NGINX Plus Ingress Controllers
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nginx-ingress
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/nginx/nginx-ingress-helm-operator
          support: NGINX Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The `NginxIngress` Custom Resource is the definition of a
              deployment of the Ingress Controller.
            displayName: Nginx Ingress Controller
            kind: NginxIngress
            name: nginxingresses.charts.nginx.org
            version: v1alpha1
        description: The NGINX Ingress Operator is a Kubernetes/OpenShift component
          which deploys and manages one or more NGINX/NGINX Plus Ingress Controllers
        displayName: Nginx Ingress Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nginx
        - ingress-controller
        - ingress
        - controller
        - kubernetes
        - openshift
        links:
        - name: Nginx Ingress Operator
          url: https://github.com/nginx/nginx-ingress-helm-operator
        maintainers:
        - email: kubernetes@nginx.com
          name: NGINX Inc
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: NGINX Inc
        relatedImages:
        - quay.io/nginx/nginx-ingress-operator@sha256:fb372438634b4a667218e725091bd448a253e7a9c40e7dda1044a318780d995f
        - quay.io/brancz/kube-rbac-proxy@sha256:754ab2a723c8fc7ea8e09ae1224f8af5be91849b8b2f2a3735c6237f1c534187
        version: 3.3.0
      entries:
      - name: nginx-ingress-operator.v3.3.0
        version: 3.3.0
      name: alpha
    defaultChannel: alpha
    packageName: nginx-ingress-operator
    provider:
      name: NGINX Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: QuestLabs
      provider-url: https://www.questlabs.in
    name: alloy
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: alloy.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.alloys.com/v1alpha1",
                "kind": "Alloy",
                "metadata": {
                  "name": "alloy"
                },
                "spec": {
                  "acmeEmail": "",
                  "dnsFQDN": "",
                  "sslTermination": "true"
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Alloy
            name: alloys.charts.alloys.com
            version: v1alpha1
        description: |-
          Alloy is platform to build trusted data in OSDU, available on-premise or in the cloud. It is equipped with built-in data quality rules derived from PPDM and OSDU standards that control the ingestion of data.

              • Applicable for conventional and new energy streams
              • Highly Scalable

              Benefits
              • Helps to build robust applications/solutions on stable data foundation using full data lifecycle APIs & built-in data/business rules
              • Help data governance team to monitor/oversea data across organisation and intervene when required
              • Substantially low implementation time for any new data entity or new domain (solar, wind etc.) because of modular rules packages
              • Capability to deploy workflows for complex corporate specific data quality and enrichment requirements
              • Spend less time & resources to deploy fit-for-purpose data
              • Schema based checks out of the box for conventional and new energy streams.
              • Sustainable and future proof


              Prerequisites
              1. Please create alloy project before installing alloy operator

          Getting Help

          Once Alloy is successfully installed you can also browse documentation at https://your.domain.com/alloy/docs/introduction
          Please visit https://www.questlabs.in/alloy for more details and if you have any queries or suggestions, mail us at info@questlabs.in
        displayName: Alloy
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - alloy
        links:
        - name: alloy
          url: https://www.questlabs.in
        maintainers:
        - email: info@questlabs.in
          name: QuestLabs
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: QuestLabs
          url: https://www.questlabs.in
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d291e0d38bd95eb2731a10bd5ff06bfa07c6621528d3c2bbe888397d87b3feaa
        - alloy.azurecr.io/alloy@sha256:bf29f69889d408507cf1a9a940beb80d95ba9faff1da2d70a4aea82cc235d96d
        version: 1.0.4
      entries:
      - name: alloy.v1.0.4
        version: 1.0.4
      - name: alloy.v1.0.3
        version: 1.0.3
      - name: alloy.v1.0.2
        version: 1.0.2
      - name: alloy.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: alloy
    provider:
      name: QuestLabs
      url: https://www.questlabs.in
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: SynaXG Engineering Team
      provider-url: ""
    name: synaxg-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: synaxg-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "card.synaxg.app.com/v1",
                "kind": "SG3Reset",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "synaxg-operator",
                    "app.kubernetes.io/instance": "sg3reset-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sg3reset",
                    "app.kubernetes.io/part-of": "synaxg-operator"
                  },
                  "name": "sg3reset-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/synaxg/synaxg-reset-utility@sha256:4050d5731a21bbec6e26165b8dc2b7632577ffa376e7eb056aa3245cf9574753",
                  "sg3ResetList": [
                    {
                      "nodeName": "test",
                      "pciList": [
                        {
                          "enable": "false",
                          "pciAddr": "000d:01:00.0"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "card.synaxg.app.com/v1",
                "kind": "SG3Update",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "synaxg-operator",
                    "app.kubernetes.io/instance": "sg3update-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sg3update",
                    "app.kubernetes.io/part-of": "synaxg-operator"
                  },
                  "name": "sg3update-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/synaxg/synaxg-upgrade-utility@sha256:7ffec3249edd2467f2414c2ce4232e8e436278fd1b43a51b66f54b6ea10a60d1",
                  "sg3UpdateList": [
                    {
                      "nodeName": "test",
                      "pciList": [
                        {
                          "enable": "false",
                          "pciAddr": "000d:01:00.0",
                          "sdkImageAddr": "quay.io/synaxgcom/sdk-img@sha256:1ed57013f278e5a858f71f343649652dbd52b6f3cf451374e45b6ac427fa76f9"
                        }
                      ]
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/synaxg/synaxg-operator-utility@sha256:3b22a8022a97cb1ffdb47ffcc9d9d562601f48669aa14fd58cd8ba6128b39bd7
          createdAt: "2025-07-01T03:42:07Z"
          description: SynaXG Operator will work with the SuperMicro Server + SG3
            card  and Openshift Container Platform.The Marvell kernel module “octeon_ep”
            and “octeon_ep_vf” are necessary to the Operator. The Operator can send
            command of “reset” and “update” to SG3 board via the VF channel.As a result,
            users can realize the process of “reset card” and “update firmware” on
            the web GUI, and monitor the executing result.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: free text
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: SynaXG
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SG3Reset is the Schema for the sg3resets API
            displayName: SG3 Reset
            kind: SG3Reset
            name: sg3resets.card.synaxg.app.com
            version: v1
          - description: SG3Update is the Schema for the sg3updates API
            displayName: SG3 Update
            kind: SG3Update
            name: sg3updates.card.synaxg.app.com
            version: v1
        description: SynaXG Operator will work with the SuperMicro Server + SG3 card  and
          Openshift Container Platform.The Marvell kernel module “octeon_ep” and “octeon_ep_vf”
          are necessary to the Operator. The Operator can send command of “reset”
          and “update” to SG3 board via the VF channel.As a result, users can realize
          the process of “reset card” and “update firmware” on the web GUI, and monitor
          the executing result.
        displayName: SynaXG vDU card 102X Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vDU
        - 102X
        - SynaXG
        links:
        - name: Synaxg Operator
          url: https://synaxg-operator.domain
        maintainers:
        - email: notice@synaxg.com
          name: Mantosh Malhotra
        maturity: alpha
        provider:
          name: SynaXG Engineering Team
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:38ee3484ba60bb474e7231e9d0718157386716cea7ffd6da6f32ddd551a5243d
        - registry.connect.redhat.com/synaxg/synaxg-operator-utility@sha256:3b22a8022a97cb1ffdb47ffcc9d9d562601f48669aa14fd58cd8ba6128b39bd7
        - registry.connect.redhat.com/synaxg/synaxg-reset-utility@sha256:4050d5731a21bbec6e26165b8dc2b7632577ffa376e7eb056aa3245cf9574753
        - registry.connect.redhat.com/synaxg/synaxg-upgrade-utility@sha256:7ffec3249edd2467f2414c2ce4232e8e436278fd1b43a51b66f54b6ea10a60d1
        - quay.io/synaxgcom/sdk-img@sha256:1ed57013f278e5a858f71f343649652dbd52b6f3cf451374e45b6ac427fa76f9
        version: 1.2.0
      entries:
      - name: synaxg-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: synaxg-operator
    provider:
      name: SynaXG Engineering Team
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-opa-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opa-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:08f995a39db5c2c6a17008cffa5216bb9bebd784d13d2b94154474d727fc796f
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/opa-operator@sha256:08f995a39db5c2c6a17008cffa5216bb9bebd784d13d2b94154474d727fc796f
        version: 24.11.1
      entries:
      - name: opa-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: opa-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:826aa12df4531d93a337ba70bd0c0d1e2e8062efe6954fc22797f6474c7732ca
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/opa-operator@sha256:826aa12df4531d93a337ba70bd0c0d1e2e8062efe6954fc22797f6474c7732ca
        version: 25.3.0
      entries:
      - name: opa-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: opa-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaClusterSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
        version: 25.7.0
      entries:
      - name: opa-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: opa-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaClusterSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
        version: 25.7.0
      entries:
      - name: opa-operator.v25.7.0
        version: 25.7.0
      - name: opa-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-opa-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: datagrid
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datagrid-operator.v8.2.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "adminAuth": {
                    "secretName": "basic-auth"
                  },
                  "clusterName": "example-infinispan",
                  "name": "mycache"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC batch-cache"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:1435aa60e8ab4d25c86daae68dcb93d14999f50d7c1e11a3c5726a86d3d30453
          createdAt: "2021-08-24T17:39:44Z"
          description: Create and manage Red Hat Data Grid clusters.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Restore API schema.
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Interoperability:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Data Grid Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
          Red Hat Data Grid is a distributed, in-memory data store that increases application performance and delivers open-source capabilities to handle demanding use cases.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.2/html/data_grid_operator_8.2_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.2
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:1435aa60e8ab4d25c86daae68dcb93d14999f50d7c1e11a3c5726a86d3d30453
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4d8057709958ef60bcabdcec52f05eae8df121fd3f4762c3218abca38130f638
        - registry.redhat.io/datagrid/datagrid-8-openj9-11-rhel8@sha256:c8607d6538a7528e27e6ecb138e0a93a03b3c9fa1f8a9a5aad3fd9f4a6814448
        version: 8.2.8
      entries:
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.2.x
    - currentCSV: datagrid-operator.v8.3.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:4d9d7221b029866837add5f3cb52dd0b7956f57996ce2d8b76344ce2de678716
          createdAt: "2022-11-03T08:40:57Z"
          description: Create and manage Red Hat Data Grid clusters.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.3/html/data_grid_operator_8.3_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.3
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:4d9d7221b029866837add5f3cb52dd0b7956f57996ce2d8b76344ce2de678716
        version: 8.3.9
      entries:
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.3.x
    - currentCSV: datagrid-operator.v8.4.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:e5315682869de72514e3e027bfb02bf0bb64919453ca738c7ecfa1e34634ac2b
          createdAt: 2024-05-30T08:05:19 EDT
          description: Create and manage Red Hat Data Grid clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.4/html/data_grid_operator_8.4_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.4
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:985193caf9bc3698c4d85aeedd1cfea09bccda881636d67885387b2a83c65ebd
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9d9e466b7c367d0406406550c74bc790aa68f3f875ea4f652eb28f2e25a05458
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:b56269cb849afca34e7f82cc8e01a44bb32af6126f24db28019a58e556d4cb3c
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:d522a20ca6909bd464f85967d8b86fe42b83f773478a80785f9a85a0cb87dfed
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bf7760520f5bea151a531d538d525c4591c47166995aefb4d77fef138de11544
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bcb040bf640691568a93a4c71c8072c37ffb76b4296ade83186f279862b7f197
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:24cf79efa89bbfa495665b1711a27531b7ab0813695a89b2c38c96c2e792eb24
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:532dca6facb584c1261c0480c499af9f857c70dc390e0d46e46afec2561468af
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:78a6bed6fcb6962e5689b92806789b72a66db21b7c607b392a573c98593f2c09
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:ae852e4679978725bb7d24c1bdfe25b40bd883130ed04a36d4fca49dbea024ef
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:424b30dccd773f69c079b35a42e8c3e339246c5d490e14e3fcfb8bd8144a9160
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9223b4b5c1c866a91033dac40f30a4caae0f6946cdb64935a3fdfffdb9ede660
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:c15d4fd827f554291c14851f187947f4109c0e20257fa2f1e141b32669569358
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:54bc26a0fddc160a4e0d1a813aec78abb3c9cb77db5039b74c23de759faa67f5
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:554b831a0803d55a4e3b7022011b5a5f6d2b4b65009065a9b28bc8e55fc3ad46
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:e5315682869de72514e3e027bfb02bf0bb64919453ca738c7ecfa1e34634ac2b
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4dbe1b7a1a36990c0a31d5a66f3cc4ca2954c9e1ec3b87431113a4111530277d
        version: 8.4.16
      entries:
      - name: datagrid-operator.v8.4.16
        version: 8.4.16
      - name: datagrid-operator.v8.4.15
        version: 8.4.15
      - name: datagrid-operator.v8.4.14
        version: 8.4.14
      - name: datagrid-operator.v8.4.13
        version: 8.4.13
      - name: datagrid-operator.v8.4.12
        version: 8.4.12
      - name: datagrid-operator.v8.4.11
        version: 8.4.11
      - name: datagrid-operator.v8.4.10
        version: 8.4.9
      - name: datagrid-operator.v8.4.8
        version: 8.4.8
      - name: datagrid-operator.v8.4.7
        version: 8.4.7
      - name: datagrid-operator.v8.4.6
        version: 8.4.6
      - name: datagrid-operator.v8.4.5
        version: 8.4.5
      - name: datagrid-operator.v8.4.4
        version: 8.4.4
      - name: datagrid-operator.v8.4.3
        version: 8.4.3
      - name: datagrid-operator.v8.4.2
        version: 8.4.2
      - name: datagrid-operator.v8.4.1
        version: 8.4.1
      - name: datagrid-operator.v8.4.0
        version: 8.4.0
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.4.x
    - currentCSV: datagrid-operator.v8.5.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel9-operator@sha256:d719d778b0848588a7e61d3c3fa689763f51a6533af4cf9e1c738c4ed8fac760
          createdAt: 2025-08-29T05:44:37 UTC
          description: Create and manage Red Hat Data Grid clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.5/html/data_grid_operator_8.5_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.5
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:a711c50e489995fb60e5dec23ce286360743feda5066e4a0861d448aa4f2e903
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bf7760520f5bea151a531d538d525c4591c47166995aefb4d77fef138de11544
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bcb040bf640691568a93a4c71c8072c37ffb76b4296ade83186f279862b7f197
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:4b9bf39d2ff14bcfbd4628a399093703d87a2b2b9c8e74c3d7cfdff4b5167fa4
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:ae852e4679978725bb7d24c1bdfe25b40bd883130ed04a36d4fca49dbea024ef
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:d522a20ca6909bd464f85967d8b86fe42b83f773478a80785f9a85a0cb87dfed
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9223b4b5c1c866a91033dac40f30a4caae0f6946cdb64935a3fdfffdb9ede660
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:11bfe5c8109d26164cff428232b2001ab8df464f14a382661b936543336e2960
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:ad4f91f984b4b898012963d52e6ebcd166ec03e76e26eba3191ecb87f51fb18d
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:c15d4fd827f554291c14851f187947f4109c0e20257fa2f1e141b32669569358
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:5656cd188d054551489d45be996266d03e3faa8814b8bfc0f360cf0dceb83cbe
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:424b30dccd773f69c079b35a42e8c3e339246c5d490e14e3fcfb8bd8144a9160
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:879f3dcc24ffe795472779f777d955666ba4041c17e717d5e165cd7f12d7955a
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:985193caf9bc3698c4d85aeedd1cfea09bccda881636d67885387b2a83c65ebd
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:b56269cb849afca34e7f82cc8e01a44bb32af6126f24db28019a58e556d4cb3c
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:9bc0b730570ab55fbeca83075949f4caa96ed5c37637420a59cd4a7db01ea840
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:532dca6facb584c1261c0480c499af9f857c70dc390e0d46e46afec2561468af
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9d9e466b7c367d0406406550c74bc790aa68f3f875ea4f652eb28f2e25a05458
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:54bc26a0fddc160a4e0d1a813aec78abb3c9cb77db5039b74c23de759faa67f5
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:78a6bed6fcb6962e5689b92806789b72a66db21b7c607b392a573c98593f2c09
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4dbe1b7a1a36990c0a31d5a66f3cc4ca2954c9e1ec3b87431113a4111530277d
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:48cf13e71953cea6f62354289f0d8ece7a8a481ee0bb2c76eb9352d438db9ca3
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:0d368459c98c5788dd55d2add5b4b7d438b9a8c814ea8aee2222ea956699cbb6
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:c25e3152d6a6b6d5cd86d311bf567e594b9948440785209cd8d7986076a5a10d
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:f59ae21773dfc207454a709ac5f7aad2be6a6a20ac88da26c379edeed54537b8
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:24cf79efa89bbfa495665b1711a27531b7ab0813695a89b2c38c96c2e792eb24
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:6e1766784b684561666134d41dca5331b80d50e63695fe7b1a1d6d4cdb73f0bd
        - registry.redhat.io/datagrid/datagrid-8-rhel9-operator@sha256:d719d778b0848588a7e61d3c3fa689763f51a6533af4cf9e1c738c4ed8fac760
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:554b831a0803d55a4e3b7022011b5a5f6d2b4b65009065a9b28bc8e55fc3ad46
        version: 8.5.11
      entries:
      - name: datagrid-operator.v8.5.11
        version: 8.5.11
      - name: datagrid-operator.v8.5.10
        version: 8.5.10
      - name: datagrid-operator.v8.5.9
        version: 8.5.9
      - name: datagrid-operator.v8.5.8
        version: 8.5.8
      - name: datagrid-operator.v8.5.7
        version: 8.5.7
      - name: datagrid-operator.v8.5.6
        version: 8.5.6
      - name: datagrid-operator.v8.5.5
        version: 8.5.5
      - name: datagrid-operator.v8.5.4
        version: 8.5.4
      - name: datagrid-operator.v8.5.3
        version: 8.5.3
      - name: datagrid-operator.v8.5.2
        version: 8.5.2
      - name: datagrid-operator.v8.5.1
        version: 8.5.1
      - name: datagrid-operator.v8.5.0
        version: 8.5.0
      - name: datagrid-operator.v8.4.16
        version: 8.4.16
      - name: datagrid-operator.v8.4.15
        version: 8.4.15
      - name: datagrid-operator.v8.4.14
        version: 8.4.14
      - name: datagrid-operator.v8.4.13
        version: 8.4.13
      - name: datagrid-operator.v8.4.12
        version: 8.4.12
      - name: datagrid-operator.v8.4.11
        version: 8.4.11
      - name: datagrid-operator.v8.4.10
        version: 8.4.9
      - name: datagrid-operator.v8.4.8
        version: 8.4.8
      - name: datagrid-operator.v8.4.7
        version: 8.4.7
      - name: datagrid-operator.v8.4.6
        version: 8.4.6
      - name: datagrid-operator.v8.4.5
        version: 8.4.5
      - name: datagrid-operator.v8.4.4
        version: 8.4.4
      - name: datagrid-operator.v8.4.3
        version: 8.4.3
      - name: datagrid-operator.v8.4.2
        version: 8.4.2
      - name: datagrid-operator.v8.4.1
        version: 8.4.1
      - name: datagrid-operator.v8.4.0
        version: 8.4.0
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: stable
    defaultChannel: stable
    packageName: datagrid
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeRocketCI
      provider-url: https://docs.kuberocketci.io/
    name: sonar-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sonar-operator.v3.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Sonar",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonar-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonar",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonar-sample"
                },
                "spec": {
                  "defaultPermissionTemplate": "edp-default",
                  "secret": "sonar-secret",
                  "settings": [
                    {
                      "key": "sonar.global.exclusions",
                      "values": [
                        "**/*.jsx",
                        "**/*.tsx"
                      ]
                    },
                    {
                      "fieldValues": {
                        "beginBlockRegexp": ".*",
                        "endBlockRegexp": ".*"
                      },
                      "key": "sonar.issue.ignore.block"
                    },
                    {
                      "key": "sonar.dbcleaner.hoursBeforeKeepingOnlyOneSnapshotByDay",
                      "value": "20"
                    }
                  ],
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonargroup-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonargroup",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonargroup-sample"
                },
                "spec": {
                  "description": "Developers",
                  "name": "sonar-developers",
                  "permissions": [
                    "admin"
                  ],
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarPermissionTemplate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarpermissiontemplate-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarpermissiontemplate",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarpermissiontemplate-sample"
                },
                "spec": {
                  "default": true,
                  "description": "EDP permission templates",
                  "groupsPermissions": {
                    "sonar-developers": [
                      "scan"
                    ],
                    "sonar-users": [
                      "issueadmin"
                    ]
                  },
                  "name": "edp-default",
                  "projectKeyPattern": ".+",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityGate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarqualitygate-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualitygate",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarqualitygate-sample"
                },
                "spec": {
                  "conditions": {
                    "new_duplicated_lines_density": {
                      "error": "1",
                      "op": "GT"
                    }
                  },
                  "default": true,
                  "name": "sonarqualitygate-sample",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityProfile",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarqualityprofile-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualityprofile",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarqualityprofile-sample"
                },
                "spec": {
                  "default": true,
                  "language": "go",
                  "name": "sonarqualityprofile-sample",
                  "rules": {
                    "go:S1151": {
                      "params": "max=\"6\"",
                      "severity": "MAJOR"
                    }
                  },
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonaruser-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonaruser",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonaruser-sample"
                },
                "spec": {
                  "email": "sonaruser@mail.com",
                  "groups": [
                    "sonar-administrators"
                  ],
                  "login": "sonaruser",
                  "name": "sonar user",
                  "permissions": [
                    "scan"
                  ],
                  "secret": "sonaruser-secret",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/sonar-operator:3.1.0
          createdAt: "2023-09-08T12:53:57Z"
          description: An Operator for managing Sonar entities.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-sonar-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonarGroup is the Schema for the sonar group API.
            displayName: Sonar Group
            kind: SonarGroup
            name: sonargroups.edp.epam.com
            version: v1alpha1
          - description: SonarPermissionTemplate is the Schema for the sonar permission
              template API.
            displayName: Sonar Permission Template
            kind: SonarPermissionTemplate
            name: sonarpermissiontemplates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityGate is the Schema for the sonarqualitygates
              API
            displayName: Sonar Quality Gate
            kind: SonarQualityGate
            name: sonarqualitygates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityProfile is the Schema for the sonarqualityprofiles
              API
            displayName: Sonar Quality Profile
            kind: SonarQualityProfile
            name: sonarqualityprofiles.edp.epam.com
            version: v1alpha1
          - description: Sonar is the Schema for the sonars API.
            displayName: Sonar
            kind: Sonar
            name: sonars.edp.epam.com
            version: v1alpha1
          - description: SonarUser is the Schema for the sonarusers API
            displayName: Sonar User
            kind: SonarUser
            name: sonarusers.edp.epam.com
            version: v1alpha1
        description: |
          Sonar Operator is an operator that is responsible for establishing
          a connection to provided SonarQube server, reconciling Sonar entities (users,
          groups, permission templates, etc) according to the created CRs.
        displayName: EDP Sonar Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - analysis
        - automation
        - code
        - continuous
        - devops
        - edp
        - inspection
        - metrics
        - metrics
        - monitoring
        - operator
        - quality
        - security
        - software
        - static
        - technical
        - vulnerabilities
        links:
        - name: Sonar Operator
          url: https://github.com/epam/edp-sonar-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/sonar-operator:3.1.0
        version: 3.1.0
      entries:
      - name: sonar-operator.v3.1.0
        version: 3.1.0
      name: alpha
    - currentCSV: sonar-operator.v3.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Sonar",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonar"
                  },
                  "name": "sonar-sample"
                },
                "spec": {
                  "defaultPermissionTemplate": "edp-default",
                  "secret": "sonar-secret",
                  "settings": [
                    {
                      "key": "sonar.global.exclusions",
                      "values": [
                        "**/*.jsx",
                        "**/*.tsx"
                      ]
                    },
                    {
                      "fieldValues": {
                        "beginBlockRegexp": ".*",
                        "endBlockRegexp": ".*"
                      },
                      "key": "sonar.issue.ignore.block"
                    },
                    {
                      "key": "sonar.dbcleaner.hoursBeforeKeepingOnlyOneSnapshotByDay",
                      "value": "20"
                    }
                  ],
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonargroup"
                  },
                  "name": "sonargroup-sample"
                },
                "spec": {
                  "description": "Developers",
                  "name": "sonar-developers",
                  "permissions": [
                    "admin"
                  ],
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarPermissionTemplate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarpermissiontemplate"
                  },
                  "name": "sonarpermissiontemplate-sample"
                },
                "spec": {
                  "default": true,
                  "description": "EDP permission templates",
                  "groupsPermissions": {
                    "sonar-developers": [
                      "scan"
                    ],
                    "sonar-users": [
                      "issueadmin"
                    ]
                  },
                  "name": "edp-default",
                  "projectKeyPattern": ".+",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityGate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualitygate"
                  },
                  "name": "sonarqualitygate-sample"
                },
                "spec": {
                  "conditions": {
                    "new_duplicated_lines_density": {
                      "error": "1",
                      "op": "GT"
                    }
                  },
                  "default": true,
                  "name": "sonarqualitygate-sample",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityProfile",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualityprofile"
                  },
                  "name": "sonarqualityprofile-sample"
                },
                "spec": {
                  "default": true,
                  "language": "go",
                  "name": "sonarqualityprofile-sample",
                  "rules": {
                    "go:S1151": {
                      "params": "max=\"6\"",
                      "severity": "MAJOR"
                    }
                  },
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonaruser"
                  },
                  "name": "sonaruser-sample"
                },
                "spec": {
                  "email": "sonaruser@mail.com",
                  "groups": [
                    "sonar-administrators"
                  ],
                  "login": "sonaruser",
                  "name": "sonar user",
                  "permissions": [
                    "scan"
                  ],
                  "secret": "sonaruser-secret",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/sonar-operator:3.3.0
          createdAt: "2025-04-22T11:39:32Z"
          description: An Operator for managing Sonar entities.
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/epam/edp-sonar-operator
          support: KubeRocketCI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonarGroup is the Schema for the sonar group API.
            displayName: Sonar Group
            kind: SonarGroup
            name: sonargroups.edp.epam.com
            version: v1alpha1
          - description: SonarPermissionTemplate is the Schema for the sonar permission
              template API.
            displayName: Sonar Permission Template
            kind: SonarPermissionTemplate
            name: sonarpermissiontemplates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityGate is the Schema for the sonarqualitygates
              API
            displayName: Sonar Quality Gate
            kind: SonarQualityGate
            name: sonarqualitygates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityProfile is the Schema for the sonarqualityprofiles
              API
            displayName: Sonar Quality Profile
            kind: SonarQualityProfile
            name: sonarqualityprofiles.edp.epam.com
            version: v1alpha1
          - description: Sonar is the Schema for the sonars API.
            displayName: Sonar
            kind: Sonar
            name: sonars.edp.epam.com
            version: v1alpha1
          - description: SonarUser is the Schema for the sonarusers API.
            displayName: Sonar User
            kind: SonarUser
            name: sonarusers.edp.epam.com
            version: v1alpha1
        description: |
          Sonar Operator is an operator that is responsible for establishing
          a connection to provided SonarQube server, reconciling Sonar entities (users,
          groups, permission templates, etc) according to the created CRs.
        displayName: EDP Sonar Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - analysis
        - automation
        - code
        - continuous
        - devops
        - edp
        - inspection
        - metrics
        - metrics
        - monitoring
        - operator
        - quality
        - security
        - software
        - static
        - technical
        - vulnerabilities
        links:
        - name: Sonar Operator
          url: https://github.com/epam/edp-sonar-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: KubeRocketCI
          url: https://docs.kuberocketci.io/
        relatedImages:
        - docker.io/epamedp/sonar-operator:3.3.0
        version: 3.3.0
      entries:
      - name: sonar-operator.v3.3.0
        version: 3.3.0
      - name: sonar-operator.v3.2.0
        version: 3.2.0
      name: stable
    defaultChannel: stable
    packageName: sonar-operator
    provider:
      name: KubeRocketCI
      url: https://docs.kuberocketci.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.6.5-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6.5
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/kubeturbo-operator:8.6.5
        version: 8.6.5-beta.1
      entries:
      - name: kubeturbo-operator.v8.6.5-beta.1
        version: 8.6.5-beta.1
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: beta
    - currentCSV: kubeturbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/kubeturbo-operator:8.6
        version: 8.6.0
      entries:
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: stable
    - currentCSV: kubeturbo-operator.v8.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.4
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/kubeturbo-operator:8.4
        version: 8.4.0
      entries:
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.4.0
    - currentCSV: kubeturbo-operator.v8.5.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.5
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/kubeturbo-operator:8.5
        version: 8.5.0
      entries:
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.5.0
    - currentCSV: kubeturbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/kubeturbo-operator:8.6
        version: 8.6.0
      entries:
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.6.0
    defaultChannel: stable
    packageName: kubeturbo
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Silicom Ltd.
      provider-url: http://www.silicom-usa.com
    name: silicom-sts-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: silicom-sts-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 3,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-BC-8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/mode": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "hp-ocp",
                  "namespase": "openshift-operators"
                },
                "spec": {
                  "dummy": 0
                },
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicom-sts/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicom-sts/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/rmr/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - quay.io/silicom/gpsd@sha256:e479c542e061b5422865f2617ed0563794f265c6e6660e5a9d204c9e1cb2f346
        - quay.io/silicom/tsyncd@sha256:3f1cbd4a8b5d62d7d3e525af5f26d9f27b8c08e886b0639a1e368163e166b537
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - quay.io/silicom/sts-operator@sha256:7d474d0aa91ac57d8f5817b033b67b511c8601e0dbd83d808a69004bd24b4e60
        - quay.io/silicom/grpc-tsyncd@sha256:631bb754742e1a24376913baed60e480a6e6ae988ad9a3892254ff96996c15d1
        - quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
        - quay.io/silicom/phc2sys@sha256:b067167a1a5d93e9fe715c7774f6cdbbcd93cac27780f022e44dcb8a4d4d7bb9
        - quay.io/silicom/sts-plugin@sha256:233bc95f8617a06bd452548ec116d91551dd4933970860d8d5ccb81f75ad2220
        - quay.io/silicom/ice-driver-src@sha256:569c8333e21c7c022ea71fa5b418404899040d5b3a3a4128eeac21f62aacaf3e
        version: 0.1.3
      entries:
      - name: silicom-sts-operator.v0.1.3
        version: 0.1.3
      - name: silicom-sts-operator.v0.0.9
        version: 0.0.9
      - name: silicom-sts-operator.v0.0.8
        version: 0.0.8
      name: alpha
    - currentCSV: silicom-sts-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 3,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-BC-8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/mode": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "hp-ocp",
                  "namespase": "openshift-operators"
                },
                "spec": {
                  "dummy": 0
                },
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicom-sts/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicom-sts/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/rmr/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - quay.io/silicom/sts-plugin@sha256:233bc95f8617a06bd452548ec116d91551dd4933970860d8d5ccb81f75ad2220
        - quay.io/silicom/sts-operator@sha256:7d474d0aa91ac57d8f5817b033b67b511c8601e0dbd83d808a69004bd24b4e60
        - quay.io/silicom/tsyncd@sha256:3f1cbd4a8b5d62d7d3e525af5f26d9f27b8c08e886b0639a1e368163e166b537
        - quay.io/silicom/grpc-tsyncd@sha256:631bb754742e1a24376913baed60e480a6e6ae988ad9a3892254ff96996c15d1
        - quay.io/silicom/ice-driver-src@sha256:569c8333e21c7c022ea71fa5b418404899040d5b3a3a4128eeac21f62aacaf3e
        - quay.io/silicom/gpsd@sha256:e479c542e061b5422865f2617ed0563794f265c6e6660e5a9d204c9e1cb2f346
        - quay.io/silicom/phc2sys@sha256:b067167a1a5d93e9fe715c7774f6cdbbcd93cac27780f022e44dcb8a4d4d7bb9
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
        version: 0.1.3
      entries:
      - name: silicom-sts-operator.v0.1.3
        version: 0.1.3
      - name: silicom-sts-operator.v0.0.9
        version: 0.0.9
      - name: silicom-sts-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: silicom-sts-operator
    provider:
      name: Silicom Ltd.
      url: http://www.silicom-usa.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-flashsystemcluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "flashsystemcluster.odf.ibm.com/v1alpha1",
                  "name": "my-flashsystemcluster",
                  "namespace": "openshift-storage"
                }
              },
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-storagecluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "storagecluster.ocs.openshift.io/v1",
                  "name": "my-storagecluster",
                  "namespace": "openshift-storage"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-console"]'
          containerImage: registry.redhat.io/odf4/odf-rhel9-operator@sha256:8d377d2d8a94e461a6eca8d1c3b4f5cf1f9ec1ba6a03044304403b58bbbabaa4
          createdAt: "2025-09-12T07:34:03Z"
          description: OpenShift Data Foundation provides a common control plane for
            storage solutions on OpenShift Container Platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.16"}]'
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operator.openshift.io/uninstall-message: 'Warning: deleting the operand
            (StorageSystem) will result in data loss'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "odf.openshift.io/v1alpha1",
              "kind": "StorageSystem",
              "metadata": {
                "name": "ocs-storagecluster-storagesystem",
                "namespace": "openshift-storage"
              },
              "spec": {
                "kind": "storagecluster.ocs.openshift.io/v1",
                "name": "ocs-storagecluster",
                "namespace": "openshift-storage"
              }
            }
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/odf-operator
          support: Red Hat
          vendors.odf.openshift.io/kind: '["storagecluster.ocs.openshift.io/v1", "flashsystemcluster.odf.ibm.com/v1alpha1"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageSystem is the Schema for the storagesystems API
            displayName: Storage System
            kind: StorageSystem
            name: storagesystems.odf.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Data Foundation** deploys three operators.

          ### OpenShift Data Foundation
          The OpenShift Data Foundation operator is the primary operator for
          OpenShift Data Foundation. It serves to facilitate the other operators in
          OpenShift Data Foundation by performing administrative tasks outside their
          scope as well as watching and configuring their CustomResources.

          ### Rook
          [Rook][1] deploys and manages Ceph on OpenShift, which provides block,
          file and object storage.

          ### NooBaa operator
          The NooBaa operator deploys and manages the [NooBaa][2] Multi-Cloud Gateway
          on OpenShift, which provides object storage.

          # Core Capabilities
          * **Self-managing service:** No matter which supported storage technologies
          you choose, OpenShift Data Foundation ensures that resources can be
          deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Data Foundation you
          can either build dedicated storage clusters or hyper-converged clusters
          where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Data Foundation:**
          OpenShift Data Foundation integrates Ceph with multiple storage
          presentations including object storage (compatible with S3), block storage,
          and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Data Foundation efficiently
          distributes and replicates your data across your cluster to minimize the
          risk of data loss. With snapshots, cloning, and versioning, no more losing
          sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your
          datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data
          storage for your workloads, using a single namespace.

          # Subscriptions
          OpenShift Data Foundation requires a matching 1:1 ratio with the OpenShift
          subscription used. The Essential subscription is included in the OpenShift
          Platform Plus subscription. Disaster Recovery features, external mode, and
          KMS integration for encryption at rest require a bundle of OpenShift
          Platform Plus with OpenShift Data Foundation Advanced subscription.

          [1]: https://rook.io
          [2]: https://www.noobaa.io
        displayName: OpenShift Data Foundation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/odf4/odf-rhel9-operator@sha256:8d377d2d8a94e461a6eca8d1c3b4f5cf1f9ec1ba6a03044304403b58bbbabaa4
        - registry.redhat.io/odf4/odf-console-rhel9@sha256:5d0d27a091a74e65985d2f047d3f6322043f2ae80fa5ff3192ec84cce92f143d
        version: 4.15.20-rhodf
      entries:
      - name: odf-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odf-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odf-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odf-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odf-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odf-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odf-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odf-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odf-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odf-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odf-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odf-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odf-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odf-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odf-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odf-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odf-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odf-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odf-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odf-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odf-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-flashsystemcluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "flashsystemcluster.odf.ibm.com/v1alpha1",
                  "name": "my-flashsystemcluster",
                  "namespace": "openshift-storage"
                }
              },
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-storagecluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "storagecluster.ocs.openshift.io/v1",
                  "name": "my-storagecluster",
                  "namespace": "openshift-storage"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-console"]'
          containerImage: registry.redhat.io/odf4/odf-rhel9-operator@sha256:20cb12259667539a0244cfde37297b975e2682833fbe537caecb5107ebecf15f
          createdAt: "2025-09-12T06:33:04Z"
          description: OpenShift Data Foundation provides a common control plane for
            storage solutions on OpenShift Container Platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.17"}]'
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operator.openshift.io/uninstall-message: 'Warning: deleting the operand
            (StorageSystem) will result in data loss'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "odf.openshift.io/v1alpha1",
              "kind": "StorageSystem",
              "metadata": {
                "name": "ocs-storagecluster-storagesystem",
                "namespace": "openshift-storage"
              },
              "spec": {
                "kind": "storagecluster.ocs.openshift.io/v1",
                "name": "ocs-storagecluster",
                "namespace": "openshift-storage"
              }
            }
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/odf-operator
          support: Red Hat
          vendors.odf.openshift.io/kind: '["storagecluster.ocs.openshift.io/v1", "flashsystemcluster.odf.ibm.com/v1alpha1"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageSystem is the Schema for the storagesystems API
            displayName: Storage System
            kind: StorageSystem
            name: storagesystems.odf.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Data Foundation** deploys three operators.

          ### OpenShift Data Foundation
          The OpenShift Data Foundation operator is the primary operator for
          OpenShift Data Foundation. It serves to facilitate the other operators in
          OpenShift Data Foundation by performing administrative tasks outside their
          scope as well as watching and configuring their CustomResources.

          ### Rook
          [Rook][1] deploys and manages Ceph on OpenShift, which provides block,
          file and object storage.

          ### NooBaa operator
          The NooBaa operator deploys and manages the [NooBaa][2] Multi-Cloud Gateway
          on OpenShift, which provides object storage.

          # Core Capabilities
          * **Self-managing service:** No matter which supported storage technologies
          you choose, OpenShift Data Foundation ensures that resources can be
          deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Data Foundation you
          can either build dedicated storage clusters or hyper-converged clusters
          where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Data Foundation:**
          OpenShift Data Foundation integrates Ceph with multiple storage
          presentations including object storage (compatible with S3), block storage,
          and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Data Foundation efficiently
          distributes and replicates your data across your cluster to minimize the
          risk of data loss. With snapshots, cloning, and versioning, no more losing
          sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your
          datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data
          storage for your workloads, using a single namespace.

          # Subscriptions
          OpenShift Data Foundation requires a matching 1:1 ratio with the OpenShift
          subscription used. The Essential subscription is included in the OpenShift
          Platform Plus subscription. Disaster Recovery features, external mode, and
          KMS integration for encryption at rest require a bundle of OpenShift
          Platform Plus with OpenShift Data Foundation Advanced subscription.

          [1]: https://rook.io
          [2]: https://www.noobaa.io
        displayName: OpenShift Data Foundation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/odf4/odf-rhel9-operator@sha256:20cb12259667539a0244cfde37297b975e2682833fbe537caecb5107ebecf15f
        - registry.redhat.io/odf4/odf-console-rhel9@sha256:8bd822a3410e520fc57ac69c7599831173003e2c34b980c517f79aca97210257
        version: 4.16.16-rhodf
      entries:
      - name: odf-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odf-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odf-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odf-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odf-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odf-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odf-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odf-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odf-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odf-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odf-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odf-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odf-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odf-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odf-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odf-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odf-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odf-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: jaeger-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jaeger-operator.v1.65.0-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "jaegertracing.io/v1",
                "kind": "Jaeger",
                "metadata": {
                  "name": "my-jaeger"
                },
                "spec": {
                  "strategy": "allInOne"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/rhosdt/jaeger-rhel8-operator@sha256:51a8685e4dc70888c991996f149cd42233049d18f702ec72b066d6dad77a07c9
          createdAt: 25 Aug 2025, 13:30
          description: Provides tracing, monitoring and troubleshooting for microservices-based
            distributed systems
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.65.0-4'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-distributed-tracing
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/os-observability/konflux-jaeger
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Jaeger is the Schema for the jaegers API
            displayName: Jaeger
            kind: Jaeger
            name: jaegers.jaegertracing.io
            version: v1
        description: |2-

          ## Operator deprecation warning

          This Operator is deprecated and not recommended for new deployments. This Operator will be removed from the `redhat-operators` catalog in a future release.

          If this is not the first time you are installing it, migrate to the Red Hat build of OpenTelemetry Operator and the Tempo Operator for distributed tracing collection and storage. For more information, see "Migrating" in the Red Hat build of OpenTelemetry documentation, "Installing" in the Red Hat build of OpenTelemetry documentation, and "Installing" in the distributed tracing platform (Tempo) documentation.

          ## About

          Red Hat OpenShift distributed tracing platform based on Jaeger. Jaeger is project inspired by [Dapper](https://research.google.com/pubs/pub36356.html) and [OpenZipkin](http://zipkin.io/). It is a distributed tracing system released as open source by Uber Technologies. It is used for monitoring and troubleshooting microservices-based distributed systems.

          ### Core capabilities
          Jaeger is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed context propagation
          * Distributed transaction monitoring
          * Root cause analysis
          * Service dependency analysis
          * Performance / latency optimization
          * OpenTracing compatible data model
          * Multiple storage backends: Elasticsearch, Memory.

          ### Operator features

          * **Multiple modes** - Provides `allInOne`, `production` and `streaming` [modes of deployment](https://www.jaegertracing.io/docs/latest/operator/#deployment-strategies).

          * **Configuration** - The Operator manages [configuration information](https://www.jaegertracing.io/docs/latest/operator/#configuring-the-custom-resource) when installing Jaeger instances.

          * **Storage** - [Configure storage](https://www.jaegertracing.io/docs/latest/operator/#storage-options) used by Jaeger. By default, `memory` is used. Other options include `elasticsearch`. The operator can delegate creation of an Elasticsearch cluster to the Elasticsearch Operator if deployed.

          * **Agent** - can be deployed as [sidecar](https://www.jaegertracing.io/docs/latest/operator/#auto-injecting-jaeger-agent-sidecars) (default) and/or [daemonset](https://www.jaegertracing.io/docs/latest/operator/#installing-the-agent-as-daemonset).

          * **UI** - Optionally setup secure route to provide [access to the Jaeger UI](https://www.jaegertracing.io/docs/latest/operator/#accessing-the-jaeger-console-ui).

          ### Before you start
          1. Ensure that the appropriate storage solution, that will be used by the Jaeger instance, is available and configured.
          2. If intending to deploy an Elasticsearch cluster via the Jaeger custom resource, then the Elasticsearch Operator must first be installed.

          ### Support & Troubleshooting

          Red Hat OpenShift distributed tracing Jaeger is available and supported as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/4.17/html/distributed_tracing/distributed-tracing-platform-jaeger) is recommended when installing, configuring, and managing the Operator and instances.
          Support is provided to Red Hat OpenShift entitled customers subject to the [Production Scope for Coverage](https://access.redhat.com/support/offerings/production/soc) and the [Red Hat OpenShift distributed tracing Life Cycle](https://access.redhat.com/support/policy/updates/openshift#jaeger).
        displayName: Red Hat OpenShift distributed tracing platform
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tracing
        - monitoring
        - troubleshooting
        links:
        - name: Jaeger Operator Source Code
          url: https://github.com/jaegertracing/jaeger-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:667d6545b92aa8381f65369de3b5a6781d1de5b4bcfa58ccde4721b6b0406f3d
        - registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:d92003d6cf0b4f3dae11f849355c9179e78e768f3b7dbeaf3224a4a019ec5d5d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e77efd450e757772d0753a73062753dabd08e53f21315e9b87f4cc1efa332661
        - registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:1d586f9aafd1cf847a14f509692ce54956b0e754a9a04beaa7a75c480f86717c
        - registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:7b021ed467d35b2bcae87765dfc64f639b03bf71f7e34bbea1dfad16626847b9
        - registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:1f4a9daf152420d9ca7cf89c38ae5550b12dd274bada389c7b85ea4cbcbc4fb7
        - registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:fd8110529daa8cf3e7f9fcf71346adc5f48cb05717ebbad1c25bc61c94f31656
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        - registry.redhat.io/rhosdt/jaeger-rhel8-operator@sha256:51a8685e4dc70888c991996f149cd42233049d18f702ec72b066d6dad77a07c9
        version: 1.65.0-4
      entries:
      - name: jaeger-operator.v1.65.0-4
        version: 1.65.0-4
      - name: jaeger-operator.v1.65.0-3
        version: 1.65.0-3
      - name: jaeger-operator.v1.65.0-2
        version: 1.65.0-2
      - name: jaeger-operator.v1.65.0-1
        version: 1.65.0-1
      - name: jaeger-operator.v1.62.0-2
        version: 1.62.0-2
      - name: jaeger-operator.v1.62.0-1
        version: 1.62.0-1
      - name: jaeger-operator.v1.57.0-10
        version: 1.57.0-10
      - name: jaeger-operator.v1.57.0-10-0.1730817420.p
        version: 1.57.0-10+0.1730817420.p
      - name: jaeger-operator.v1.57.0-7
        version: 1.57.0-7
      - name: jaeger-operator.v1.57.0-6
        version: 1.57.0-6
      - name: jaeger-operator.v1.57.0-5
        version: 1.57.0-5
      - name: jaeger-operator.v1.53.0-4
        version: 1.53.0-4
      - name: jaeger-operator.v1.53.0-3
        version: 1.53.0-3
      - name: jaeger-operator.v1.51.0-1
        version: 1.51.0-1
      - name: jaeger-operator.v1.47.1-5
        version: 1.47.1-5
      - name: jaeger-operator.v1.42.0-5-0.1687199951.p
        version: 1.42.0-5+0.1687199951.p
      - name: jaeger-operator.v1.42.0-5
        version: 1.42.0-5
      - name: jaeger-operator.v1.34.1-5
        version: 1.34.1-5
      - name: jaeger-operator.v1.30.2
        version: 1.30.2
      name: stable
    defaultChannel: stable
    packageName: jaeger-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator.v1.6.3
        version: 1.6.3
      - name: dynatrace-operator.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator.v1.4.0
        version: 1.4.0
      - name: dynatrace-operator.v1.3.2
        version: 1.3.2
      - name: dynatrace-operator.v1.3.1
        version: 1.3.1
      - name: dynatrace-operator.v1.3.0
        version: 1.3.0
      - name: dynatrace-operator.v0.6.0
        version: 0.6.0
      - name: dynatrace-operator.v0.5.1
        version: 0.5.1
      - name: dynatrace-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco
      provider-url: ""
    name: aci-containers-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aci-containers-operator.v1.0.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"aci.ctrl/v1alpha1","kind":"AciContainersOperator","metadata":{"name":"acicnioperator","namespace":"aci-containers-system"},"spec":{}}]'
          capabilities: Seamless Upgrades
          categories: Networking
          certified: "false"
          containerImage: quay.io/noiro/aci-containers-operator@sha256:2b2d685e254c206be5371285852a8f95d0318af69cb85b786427969e035c125c
          createdAt: "2025-04-10T12:00:00Z"
          description: The Cisco Application Centric Infrastructure (ACI) Containers
            Operator brings the ACI CNI Plugin and Policy model to Kubernetes clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://github.com/noironetworks/aci-containers/blob/master/README.md
          repository: https://github.com/noironetworks/aci-containers
          support: acisupport@cisco.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aci Containers Operator
            displayName: AciContainersOperator
            kind: AciContainersOperator
            name: acicontainersoperators.aci.ctrl
            version: v1alpha1
        description: |
          ## ACI CNI Plugin

          The Cisco Application Centric Infrastructure (ACI) CNI plugin brings the ACI Networking and Policy model to Kubernetes clusters that reside on-prem or in the cloud. It is fully open source and relies on the Opflex Protocol to program Open vSwitch instances running on the Kubernetes nodes. It provides IP address management, L2/L3 networking, load balancing, and security functions for container workloads.

          The Cisco ACI CNI plugin for the Red Hat OpenShift Container Platform provides a single, programmable network infrastructure that eliminates bottlenecks between OpenShift workloads and the rest of the infrastructure while providing enterprise-grade security and offers flexible micro segmentation possibilities.

          Unified Networking - By extending the integrated overlay of the ACI fabric to OpenShift workloads, Pod and Service endpoints are managed as native ACI network resources at the same level as Bare Metal or Virtual Machines.

          No Egress Router Required - Applications running in OpenShift Pods that need high-bandwidth, low-latency traffic for data transfers external to the cluster suffer the bottleneck imposed by the egress router implementation. Pods can easily connect with other fabric or non-fabric endpoints leveraging the combined hardware and software distributed routing and switching provided by ACI and the ACI CNI plugin.

          Simplified Operations and Enhanced Visibility - Whether policies and networking are defined by the devops or fabric or security admins, the APIC becomes the single source of truth for operations and security governance. Kubernetes workload visibility is provided to Network Operators within their familiar Cisco tools.

          Hardware-assisted Load Balancing - Ingress connections to LoadBalancer-type services using ACI’s Policy Based Redirect technology moves load balancing into the fabric's hardware.

          Granular Security - Security can be implemented by using native NetworkPolicy or by using ACI EPGs and contracts, or both models complementing each other. Network policy of container workloads becomes a shared responsibility while ensuring global compliance and tenant constraints are enforced.
        displayName: Cisco ACI Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cisco
        - aci
        links:
        - name: Cisco ACI Homepage
          url: https://www.cisco.com/c/en/us/td/docs/switches/datacenter/aci/apic/sw/kb/b_Kubernetes_Integration_with_ACI.html
        maintainers:
        - email: abhishs9@cisco.com
          name: 'Abhishek Sharma '
        - email: sumitnaiksatam@gmail.com
          name: Sumit Naiksatam
        - email: vlella@cisco.com
          name: Vishal Lella
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Cisco
        relatedImages:
        - quay.io/noiro/aci-containers-operator@sha256:2b2d685e254c206be5371285852a8f95d0318af69cb85b786427969e035c125c
        - quay.io/noiro/acc-provision-operator@sha256:9c8fa6d705602af13d539fb919e72e7ee5064c6998fc9ca95d10108e986c68fe
        version: 1.0.10
      entries:
      - name: aci-containers-operator.v1.0.10
        version: 1.0.10
      name: stable
    defaultChannel: stable
    packageName: aci-containers-operator
    provider:
      name: Cisco
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Accuknox
      provider-url: https://accuknox.com
    name: kubearmor-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubearmor-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          com.redhat.openshift.versions: v4.12
          createdAt: "2023-05-01T05:17:07Z"
          description: The KubeArmor Operator deloys and manages KubeArmor daemonset
            and its componenets within a cluster.
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.28.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/kubearmor/KubeArmor
          support: Accuknox
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          The KubeArmor Operator deploys and manages KubeArmor and its components:
          - KubeArmor Daemonset
          - KubeArmor Relay Server
          - KubeArmor Controller to manage CRDs (KubeArmorPolicy, KubeArmorHostPolicy)

          KubeArmor is a cloud-native runtime security enforcement system that restricts the behavior (such as process
          execution, file access, and networking operations) of pods, containers, and nodes (VMs) at the system level.

          KubeArmor leverages Linux security modules (LSMs) such as AppArmor, SELinux, or BPF-LSM to enforce the user-specified policies. KubeArmor generates rich alerts/telemetry events with container/pod/namespace identities by leveraging eBPF.

          ## Documentation

          github: [https://github.com/kubearmor/KubeArmor](https://github.com/kubearmor/KubeArmor)
          website: [https://kubearmor.io/](https://kubearmor.io/)
        displayName: Kubearmor Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubearmor-operator
        - runtime-security
        - security
        - cloud-native
        links:
        - name: Kubearmor
          url: https://kubearmor.io
        maintainers:
        - email: ramakant@accuknox.com
          name: Ramakant Sharma
        maturity: alpha
        provider:
          name: Accuknox
          url: https://www.accuknox.com/
        relatedImages:
        - docker.io/kubearmor/kubearmor-init@sha256:44fa534587972c5f85e3e6f362dd8268e9088a33c5657a21b72a7e8563164fb5
        - docker.io/kubearmor/kubearmor-relay-server@sha256:695f9ade3382bcf858077e2a7b9b01e0713071517603778d72f7aca5a6d71323
        - docker.io/kubearmor/kubearmor-controller@sha256:ccec0fb8bb1e91bd48631cc89baf752b842f19c33b50edab08f8b33e57af64d4
        - docker.io/kubearmor/kubearmor-operator@sha256:8c2d56781a2adf78ee58f956e01417f38b6ad2f319feada6470e93f4f9bf0808
        - docker.io/kubearmor/kubearmor@sha256:52f7049ecdbfa779663d2cd44ddd7f36aecd8ccd2ca91e69709b628600029f13
        version: 1.0.0
      entries:
      - name: kubearmor-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: kubearmor-operator.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.kubearmor.com/v1",
                "kind": "KubeArmorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kubearmorconfig",
                    "app.kubernetes.io/part-of": "kubearmoroperator"
                  },
                    "name": "kubearmorconfig-default",
                    "namespace": "placeholder"
                },
                "spec": {
                  "adapters": {
                    "elasticsearch": {
                      "url": "",
                      "enabled": false,
                      "auth": {
                        "secretName": "elastic-secret",
                        "usernameKey": "username",
                        "passwordKey": "password",
                        "allowInsecureTLS": false,
                        "caCertSecretName": "",
                        "caCertKey": "ca.crt"
                      },
                      "alertsIndex": "kubearmor-alerts"
                    }
                  },
                  "recommendedPolicies": {
                      "enable": true,
                      "matchExpressions": [
                          {
                              "key": "namespace",
                              "operator": "NotIn",
                              "values": [
                                  "kube-system",
                                  "kubearmor",
                                  "openshift-infra",
                                  "openshift-ingress",
                                  "openshift-ingress-canary",
                                  "openshift-ingress-operator",
                                  "openshift-insights",
                                  "openshift-kni-infra",
                                  "openshift-kube-apiserver",
                                  "openshift-kube-apiserver-operator",
                                  "openshift-kube-controller-manager",
                                  "openshift-kube-controller-manager-operator",
                                  "openshift-kube-scheduler",
                                  "openshift-kube-scheduler-operator",
                                  "openshift-kube-storage-version-migrator",
                                  "openshift-kube-storage-version-migrator-operator",
                                  "openshift-logging",
                                  "openshift-machine-api",
                                  "openshift-machine-config-operator",
                                  "openshift-managed-node-metadata-operator",
                                  "openshift-managed-upgrade-operator",
                                  "openshift-marketplace",
                                  "openshift-monitoring",
                                  "openshift-multus",
                                  "openshift-must-gather-operator",
                                  "openshift-network-diagnostics",
                                  "openshift-network-node-identity",
                                  "openshift-network-operator",
                                  "openshift-node",
                                  "openshift-nutanix-infra",
                                  "openshift-oauth-apiserver",
                                  "openshift-observability-operator",
                                  "openshift-ocm-agent-operator",
                                  "openshift-openstack-infra",
                                  "openshift-operator-lifecycle-manager",
                                  "openshift-operators",
                                  "openshift-operators-redhat",
                                  "openshift-osd-metrics",
                                  "openshift-ovirt-infra",
                                  "openshift-ovn-kubernetes",
                                  "openshift-package-operator",
                                  "openshift-rbac-permissions",
                                  "openshift-route-controller-manager",
                                  "openshift-route-monitor-operator",
                                  "openshift-security",
                                  "openshift-service-ca",
                                  "openshift-service-ca-operator",
                                  "openshift-splunk-forwarder-operator",
                                  "openshift-sre-pruning",
                                  "openshift-user-workload-monitoring",
                                  "openshift-validation-webhook",
                                  "openshift-velero",
                                  "openshift-vsphere-infra",
                                  "openshift",
                                  "openshift-addon-operator",
                                  "openshift-apiserver",
                                  "openshift-apiserver-operator",
                                  "openshift-aqua",
                                  "openshift-authentication",
                                  "openshift-authentication-operator",
                                  "openshift-backplane",
                                  "openshift-backplane-cee",
                                  "openshift-backplane-csa",
                                  "openshift-backplane-cse",
                                  "openshift-backplane-csm",
                                  "openshift-backplane-managed-scripts",
                                  "openshift-backplane-mcs-tier-two",
                                  "openshift-backplane-mobb",
                                  "openshift-backplane-srep",
                                  "openshift-backplane-tam",
                                  "openshift-cloud-controller-manager",
                                  "openshift-cloud-controller-manager-operator",
                                  "openshift-cloud-credential-operator",
                                  "openshift-cloud-ingress-operator",
                                  "openshift-cloud-network-config-controller",
                                  "openshift-cloud-platform-infra",
                                  "openshift-cluster-csi-drivers",
                                  "openshift-cluster-machine-approver",
                                  "openshift-cluster-node-tuning-operator",
                                  "openshift-cluster-samples-operator",
                                  "openshift-cluster-storage-operator",
                                  "openshift-cluster-version",
                                  "openshift-codeready-workspaces",
                                  "openshift-config",
                                  "openshift-config-managed",
                                  "openshift-config-operator",
                                  "openshift-console",
                                  "openshift-console-operator",
                                  "openshift-console-user-settings",
                                  "openshift-controller-manager",
                                  "openshift-controller-manager-operator",
                                  "openshift-custom-domains-operator",
                                  "openshift-customer-monitoring",
                                  "openshift-deployment-validation-operator",
                                  "openshift-dns",
                                  "openshift-dns-operator",
                                  "openshift-etcd",
                                  "openshift-etcd-operator",
                                  "openshift-host-network",
                                  "openshift-image-registry"
                              ]
                          }
                      ]
                  },
                  "defaultCapabilitiesPosture": "audit",
                  "defaultFilePosture": "audit",
                  "defaultNetworkPosture": "audit",
                  "defaultVisibility": "process,network",
                  "enableStdOutLogs": false,
                  "enableStdOutAlerts": false,
                  "enableStdOutMsgs": false,
                  "seccompEnabled": false
                }
              }
            ]
          alm-examples-metadata: '{"kubearmorconfig-default":{"description":"Default
            KubeArmorConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Security, Monitoring
          certified: "true"
          com.redhat.openshift.versions: v4.9
          containerImage: docker.io/kubearmor/kubearmor-operator@sha256:d6c6b16e636474568835ded97cecb5ed17bc6004501227f6b6d63522ed57b01c
          createdAt: "2024-03-20T17:43:15Z"
          description: KubeArmor Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: kubearmor
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kubearmor/KubeArmor/blob/main/pkg/KubeArmorOperator/
          support: Accuknox
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeArmorConfig is the Schema for the kubearmorconfigs API
            kind: KubeArmorConfig
            name: kubearmorconfigs.operator.kubearmor.com
            version: v1
        description: |2-

          ## Prerequisites

          1. Create a new project for KubeArmor (Recommended)

             ```
             oc new-project kubearmor
             ```

           2. Create [KubeArmor SCC](https://github.com/kubearmor/KubeArmor/blob/main/pkg/KubeArmorOperator/config/rbac/kubearmor-scc.yaml)

             ```
             oc create -f https://raw.githubusercontent.com/kubearmor/KubeArmor/main/pkg/KubeArmorOperator/config/rbac/kubearmor-scc.yaml
             ```

             Note: please change the serviceaccount namespace if you are using some other namespace than the kubearmor.

          KubeArmor is a cloud-native runtime security enforcement system that restricts the behavior
          (such as process execution, file access, and networking operations) of pods, containers,
          and nodes (VMs) at the system level.
          KubeArmor leverages Linux security modules (LSMs) such as AppArmor, SELinux, or BPF-LSM to enforce the
          user-specified policies. KubeArmor generates rich alerts/telemetry events with container/pod/namespace
          identities by leveraging eBPF.

          ## Features and Benefits

          ### Harden Infrastructure

          safeguards vital paths like certificate bundles, enforces security rules based on MITRE, STIGs, and CIS guidelines,
          and restricts access to raw database tables, enhancing overall cluster security.

          ### Least Permissive Access

          controls sensitive asset access through process and network whitelisting,
          reducing security risks and maintaining system integrity.

          ### Application Behavior

          in-depth monitoring of process executions, file system access, service bindings,
          ingress and egress connections, and sensitive system call profiling

          ## Documentation

          github: [https://github.com/kubearmor/KubeArmor](https://github.com/kubearmor/KubeArmor)

          website: [https://kubearmor.io/](https://kubearmor.io/)

          ## Installing the KubeArmor Operator

          The KubeArmor Operator, available through Red Hat's Operator catalog, offers stable update channel.
          **Stable Channel**: This channel provides the latest stable version of KubeArmor along with its patches.

          The KubeArmor operator is responsible for deploying various components of the KubeArmor application, including the KubeArmor
          Daemonset, KubeArmor Controller, KubeArmor Relay Server, and KubeArmor Snitch. The role of KubeArmor Snitch is to gather information
          about the enabled Linux security modules on the host, which is utilized by KubeArmor to enforce runtime security policies.
          Additionally, KubeArmor Snitch collects details about the container runtime socket and storage volume cluster configurations.
          This deployment is executed as a short-term job and necessitates root user privileges. KubeArmor Snitch is certified by RedHat
          as a container with root privileges.

          KubeArmor Operator comes with a custom resource:

          - **KubeArmorConfig** - User will provide the initial configurations of the KubeArmor using KubeArmorConfig CR.

          ### KubeArmorConfig Customization Options:

          | Parameter                                 | Description                                          | Default                                 |
          | ---                                       | ---                                                  | ---                                     |
          | `alertThrottling`                         | Enable throttling for the KubeArmor generated alerts | False                                   |
          | `defaultFilePosture`                      | Default file posture for KubeArmor                   | audit                                   |
          | `defaultNetworkPosture`                   | Default network posture for KubeArmor                | audit                                   |
          | `defaultCapabilitiesPosture`              | Default capabilities posture for KubeArmor           | audit                                   |
          | `defaultVisibility`                       | Default visibility for KubeArmor                     | process,file,network                    |
          | `enableStdOutLogs` |  show stdout logs for relay server | false |
          | `enableStdOutAlerts` |  show stdout alerts for relay server | false |
          | `enableStdOutMsgs` |  show stdout messages for relay server | false |
          | `kubearmorImage.image`                    | Default image for KubeArmor Daemonset                | kubearmor/kubearmor-ubi:stable          |
          | `kubearmorImage.imagePullPolicy`          | Default imagePullPolicy for KubeArmor Daemonset      | Always                                  |
          | `kubearmorInitImage.image`                | Default image for KubeArmor Init Container           | kubearmor/kubearmor-init:stable         |
          | `kubearmorInitImage.imagePullPolicy`      | Default imagePullPolicy for KubeArmor Init Container | Always                                  |
          | `kubearmorRelayImage.image`               | Default image for KubeArmor Relay Server             | kubearmor/kubearmor-relay-server:latest |
          | `kubearmorRelayImage.imagePullPolicy`     | Default imagePullPolicy for KubeArmor Relay Server   | Always                                  |
          | `kubearmorControllerImage.image`          | Default image for KubeArmor Controller               | kubearmor/kubearmor-controller:latest   |
          | `kubearmorControllerImage.imagePullPolicy`| Default imagePullPolicy for KubeArmor Controller     | Always                                  |
          | `maxAlertPerSec`                          | Maximum no of alerts per second to trigger throttling| 10                                      |
          | `recommendedPolicies`                     | Recommended Policies Spec                            | -                                       |
          | `recommendedPolicies.matchExpressions`    | Target Namespaces for recommended policies           | -                                       |
          | `recommendedPolicies.excludePolicy`       | Exclude policies from recommended policies           | -                                       |
          | `seccompEnabled`                          | Deploy KubeArmor Daemon with default seccomp profile | True                                    |
          | `throttleSec`                             | Thottling duration in seconds                        | 30                                      |

          Users are not required to specify the images in the KubearmorConfig custom resource (CR). By default, the operator will utilize the images published with the operator release.

          For detailed information on Default Posture refer [this guide](https://github.com/kubearmor/KubeArmor/blob/main/getting-started/default_posture.md)

          similarly for default visibility refer [this guide](https://github.com/kubearmor/KubeArmor/blob/main/getting-started/kubearmor_visibility.md)

          ## Support

          For Support please join KubeArmor Slack [Here](https://join.slack.com/t/kubearmor/shared_invite/zt-1ltmqdbc6-rSHw~LM6MesZZasmP2hAcA)
        displayName: KubeArmor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - security
        - alerting
        - runtime-security
        - cloud native
        - container security
        links:
        - name: Website
          url: https://kubearmor.io
        - name: Github
          url: https://github.com/kubearmor/KubeArmor.git
        - name: Slack
          url: https://join.slack.com/t/kubearmor/shared_invite/zt-1ltmqdbc6-rSHw~LM6MesZZasmP2hAcA
        maintainers:
        - email: ramakant@accuknox.com
          name: Ramakant Sharma
        maturity: stable
        provider:
          name: Accuknox
          url: https://accuknox.com
        relatedImages:
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:1153a4592746b05e019bde4d818d176ff9350c013f84d49829032540de882841
        - docker.io/kubearmor/kubearmor-operator@sha256:d6c6b16e636474568835ded97cecb5ed17bc6004501227f6b6d63522ed57b01c
        - docker.io/kubearmor/kubearmor-snitch@sha256:c3d0f45590f9a0ebb50254047f12a7468844bcae7961eccb11f5ef578832c0eb
        - docker.io/kubearmor/kubearmor-ubi@sha256:629e295a2e64953bf95f3f3e842c42ff665ed6e4db32db8e43dcde4befdd3fc2
        - docker.io/kubearmor/kubearmor-init@sha256:e30ee508aa95d273f74a543e1b412b98c1714d04f093b59991c0ec2abaa450be
        - docker.io/kubearmor/kubearmor-relay-server@sha256:0a1a20ef02bd05eaa4a900948cadee6ae7650b0dde70abe1935608eaa5f931b7
        - docker.io/kubearmor/kubearmor-controller@sha256:2d369138cdbb31468c41e9ce70eda986a16bbeb9dd1192cd8a011c9df8a83857
        version: 1.5.2
      entries:
      - name: kubearmor-operator.v1.5.2
        version: 1.5.2
      name: stable
    defaultChannel: stable
    packageName: kubearmor-operator-certified
    provider:
      name: Accuknox
      url: https://accuknox.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Tigera
      provider-url: https://tigera.io/
    name: tigera-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tigera-operator.v1.20.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:bd93e66be5cfbc123fe4e8c3f6e8ddf7654217bf26136f06dcf23f28b53063d2
          createdAt: "2021-10-05T20:02:14.564507Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.20.4
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:bd93e66be5cfbc123fe4e8c3f6e8ddf7654217bf26136f06dcf23f28b53063d2
        version: 1.20.4
      entries:
      - name: tigera-operator.v1.20.4
        version: 1.20.4
      - name: tigera-operator.v1.20.3
        version: 1.20.3
      - name: tigera-operator.v1.20.2
        version: 1.20.2
      - name: tigera-operator.v1.20.1
        version: 1.20.1
      name: release-v1.20
    - currentCSV: tigera-operator.v1.23.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:f258275fe3d48c2ae16346a57f628523a6f2fb7b854b6d43b169056e158537f1
          createdAt: "2022-01-12T22:55:57.409244Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.23.4
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:f258275fe3d48c2ae16346a57f628523a6f2fb7b854b6d43b169056e158537f1
        version: 1.23.4
      entries:
      - name: tigera-operator.v1.23.4
        version: 1.23.4
      - name: tigera-operator.v1.23.3
        version: 1.23.3
      - name: tigera-operator.v1.23.2
        version: 1.23.2
      - name: tigera-operator.v1.23.1
        version: 1.23.1
      - name: tigera-operator.v1.23.0
        version: 1.23.0
      name: release-v1.23
    - currentCSV: tigera-operator.v1.25.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:e862708c700df55f99a9c9bb3c453e31f22aef1993d6079d80433eb8da831c94
          createdAt: "2022-02-17T05:39:13.720885Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.25.1
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:e862708c700df55f99a9c9bb3c453e31f22aef1993d6079d80433eb8da831c94
        version: 1.25.1
      entries:
      - name: tigera-operator.v1.25.1
        version: 1.25.1
      name: release-v1.25
    - currentCSV: tigera-operator.v1.28.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:9e8e47045fc67a2b00f9f2fd575fa526d7ddcac50af0b8b6d583b7518ccce120
          createdAt: "2023-05-17T16:13:52.148441156Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.28.13
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:9e8e47045fc67a2b00f9f2fd575fa526d7ddcac50af0b8b6d583b7518ccce120
        version: 1.28.13
      entries:
      - name: tigera-operator.v1.28.13
        version: 1.28.13
      name: release-v1.28
    - currentCSV: tigera-operator.v1.29.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:2a5acd29d8dca6adb318af04dbb15d93d899ed2b28b2ee9aaac09147395efb58
          createdAt: "2023-03-31T20:49:52.518230079Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.29.3
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:2a5acd29d8dca6adb318af04dbb15d93d899ed2b28b2ee9aaac09147395efb58
        version: 1.29.3
      entries:
      - name: tigera-operator.v1.29.3
        version: 1.29.3
      name: release-v1.29
    - currentCSV: tigera-operator.v1.30.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:431f037ff18b5c867d01312e42671effc55602421aeed25dd3f6109f70596b4a
          createdAt: "2023-11-17T00:52:31.263564294Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.30.9
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: APIServer
            name: apiservers.operator.tigera.io
            version: v1
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:431f037ff18b5c867d01312e42671effc55602421aeed25dd3f6109f70596b4a
        version: 1.30.9
      entries:
      - name: tigera-operator.v1.30.9
        version: 1.30.9
      name: release-v1.30
    - currentCSV: tigera-operator.v1.32.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:715ac9a30f8a9579e44258af20de354715429e11836b493918e9e1a696e9b028
          createdAt: "2023-12-15T23:45:03.964812504Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.32.3
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:715ac9a30f8a9579e44258af20de354715429e11836b493918e9e1a696e9b028
        version: 1.32.3
      entries:
      - name: tigera-operator.v1.32.3
        version: 1.32.3
      name: release-v1.32
    - currentCSV: tigera-operator.v1.36.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:3341fa9475c0325b86228c8726389f9bae9fd6c430c66fe5cd5dc39d7bb6ad4b
          createdAt: "2025-02-06T02:39:01.007218336Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.36.5
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:3341fa9475c0325b86228c8726389f9bae9fd6c430c66fe5cd5dc39d7bb6ad4b
        version: 1.36.5
      entries:
      - name: tigera-operator.v1.36.5
        version: 1.36.5
      name: release-v1.36
    - currentCSV: tigera-operator.v1.38.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "calicoNetwork": {
                    "linuxDataplane": "BPF"
                  },
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing, Monitoring, Networking, Security
          containerImage: quay.io/tigera/operator@sha256:e0a34b265aebce1a2db906d8dad99190706e8bf3910cae626b9c2eb6bbb21775
          createdAt: "2025-05-06T13:27:07.802897329Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          features.operators.openshift.io/valid-subscription: No subscription required
          olm.skipRange: <1.38.0
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift.
        displayName: Tigera Operator v1.38
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tigera-operator
        - networking
        - security
        - monitoring
        links:
        - name: Tigera
          url: https://tigera.io/
        - name: Calico Introduction
          url: https://docs.tigera.io/calico/latest/about/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.tigera.io/calico/latest/getting-started/kubernetes/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
          url: https://tigera.io/
        relatedImages:
        - quay.io/tigera/operator@sha256:e0a34b265aebce1a2db906d8dad99190706e8bf3910cae626b9c2eb6bbb21775
        version: 1.38.0
      entries:
      - name: tigera-operator.v1.38.0
        version: 1.38.0
      name: release-v1.38
    defaultChannel: release-v1.38
    packageName: tigera-operator
    provider:
      name: Tigera
      url: https://tigera.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solution-Soft
      provider-url: www.solution-soft.com
    name: timemachine-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: timemachine-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tm.solution-soft.com/v1alpha1",
                "kind": "TimeMachine",
                "metadata": {
                  "name": "timemachine-sample"
                },
                "spec": {
                  "cluster_provider": "ibm",
                  "aws": {
                    "shared_volume_name": ""
                  },
                  "ibm": {
                    "persistentvolume": {
                      "storageclass": "ibmc-file-silver-gid"
                    }
                  },
                  "tm_lichost": "52.152.148.40",
                  "tm_licpass": "qwaszx",
                  "tm_licport": "57777",
                  "imagePullSecrets": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
          createdAt: "2024-02-05T09:46:17Z"
          description: Time Machine Operator enables time travel of OpenShift pods,
            without the need to change the system clock on any of the cluster nodes
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/timemachine-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/timemachine-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=0.0.4 <1.0.17'
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: SolutionSoft Systems, Inc. - support@solution-soft.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Time Machine
            displayName: Time Machine
            kind: TimeMachine
            name: timemachines.tm.solution-soft.com
            version: v1alpha1
        description: |-
          Time Machine operator enables time travel of OpenShift pods/deployments.

          To time travel a pod where your application is, you just need to install Time Machine operator and create a Time Machine custom resource in the same namespace, without the need of changing the container images of the application you’re using, or adding containers to the pod with your application.

          Time travelling is done on the namespace level where all the (configured) target pods/deployments see the same virtual time.
        displayName: Time Machine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - time shift
        - time travel
        - virtual clocks
        - time machine
        - virtual time
        - solution-soft
        links:
        - name: Timemachine Operator
          url: https://catalog.redhat.com/software/operators/detail/5e98727a6c5dcb34dfbb19e3
        maintainers:
        - email: dtrajkovic@solution-soft.com
          name: Dragan Trajkovic
        maturity: alpha
        provider:
          name: Solution-Soft
          url: www.solution-soft.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0c304c11986790f446604319dfc29db494b15f5f9fa13f2c436688ec29c92dbe
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
        - registry.connect.redhat.com/rhel-timemachine-container/virtual-clock-testing-tool@sha256:a81a823328b0ee795c40668c89a08bcf545c5192d9eed0146071e3721a0a8372
        version: 1.0.17
      entries:
      - name: timemachine-operator.v1.0.17
        version: 1.0.17
      - name: timemachine-operator.v1.0.16
        version: 1.0.16
      name: alpha
    defaultChannel: alpha
    packageName: timemachine-operator-rhmp
    provider:
      name: Solution-Soft
      url: www.solution-soft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: ironic-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ironic-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "Ironic",
                "metadata": {
                  "name": "ironic"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseInstance": "openstack",
                  "ironicAPI": {},
                  "ironicConductors": [
                    {
                      "storageRequest": "10G"
                    }
                  ],
                  "ironicInspector": {},
                  "ironicNeutronAgent": {},
                  "secret": "osp-secret",
                  "serviceUser": "ironic",
                  "storageClass": "local-storage"
                }
              },
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "IronicAPI",
                "metadata": {
                  "name": "ironicapi-sample"
                },
                "spec": {
                  "databaseHostname": "openstack.ironic-kuttl-tests.svc"
                }
              },
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "IronicConductor",
                "metadata": {
                  "name": "ironicconductor-sample"
                },
                "spec": {
                  "databaseHostname": "openstack.ironic-kuttl-tests.svc",
                  "storageRequest": "10G"
                }
              },
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "IronicInspector",
                "metadata": {
                  "name": "ironicinspector-sample"
                },
                "spec": {
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "IronicNeutronAgent",
                "metadata": {
                  "name": "ironic-neutron-agent-sample",
                  "namespace": "openstack"
                },
                "spec": {
                  "secret": "osp-secret"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:02Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IronicAPI is the Schema for the ironicapis API
            displayName: Ironic API
            kind: IronicAPI
            name: ironicapis.ironic.openstack.org
            version: v1beta1
          - description: IronicConductor is the Schema for the ironicconductors Conductor
            displayName: Ironic Conductor
            kind: IronicConductor
            name: ironicconductors.ironic.openstack.org
            version: v1beta1
          - description: IronicInspector is the Schema for the ironicinspectors
            displayName: Ironic Inspector
            kind: IronicInspector
            name: ironicinspectors.ironic.openstack.org
            version: v1beta1
          - description: IronicNeutronAgent is the Schema for the ML2 networking-baremetal's
              agent
            displayName: Ironic Neutron Agent
            kind: IronicNeutronAgent
            name: ironicneutronagents.ironic.openstack.org
            version: v1beta1
          - description: Ironic is the Schema for the ironics API
            displayName: Ironic
            kind: Ironic
            name: ironics.ironic.openstack.org
            version: v1beta1
        description: Ironic Operator
        displayName: Ironic Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Identity
        - Ironic
        links:
        - name: Ironic Operator
          url: https://github.com/openstack-k8s-operators/ironic-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/ironic-rhel9-operator@sha256:184a1783a3b2a38bdb23f9811243414a2b21c16d78ca79ed83e2bcaa720d199b
        - registry.redhat.io/rhoso/openstack-ironic-api-rhel9@sha256:86c5330a3065a11f251e8c333bd855d37294c4e9d36cc68208fd16b3f95ef15a
        - registry.redhat.io/rhoso/openstack-ironic-conductor-rhel9@sha256:4950d74335eb48ed04f1a4f602730f1dfd42f139233a4edfebef0d8403c00b2d
        - registry.redhat.io/rhoso/openstack-ironic-inspector-rhel9@sha256:34ff13efc254d4390fbef0a5c088c569241d44d063ec16aed42e36d8179e0c85
        - registry.redhat.io/rhoso/openstack-ironic-pxe-rhel9@sha256:d49af96baa8d90dad3b244ed8c006a9945c41d05e5fa197cc35cc662b49cdd7c
        - registry.redhat.io/rhoso/openstack-ironic-neutron-agent-rhel9@sha256:132627eaf483353c5daefbbc84c6b06f51c50d0813a229ddc558e9df18bb5578
        - registry.redhat.io/rhoso/ironic-python-agent-rhel9@sha256:f41166809d2e7d76446c01f81d325a320df21e64f4b776d0f83cadcb9387eb7f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: ironic-operator.v1.0.6
        version: 1.0.6
      - name: ironic-operator.v1.0.5
        version: 1.0.5
      - name: ironic-operator.v1.0.4
        version: 1.0.4
      - name: ironic-operator.v1.0.3
        version: 1.0.3
      - name: ironic-operator.v1.0.2
        version: 1.0.2
      - name: ironic-operator.v1.0.1
        version: 1.0.1
      - name: ironic-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: ironic-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kove
      provider-url: https://kove.com
    name: kove-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kove-operator.v3.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "containers.kove.com/v1",
                "kind": "KoveMemoryConfig",
                "metadata": {
                  "labels": {
                    "KoveMemoryConfig": ""
                  },
                  "name": "example-memory-config"
                },
                "spec": {
                  "container": "",
                  "namespace": "",
                  "size": ""
                }
              },
              {
                "apiVersion": "containers.kove.com/v1",
                "kind": "KoveSystem",
                "metadata": {
                  "labels": {
                    "KoveSystem": ""
                  },
                  "name": "kove-system"
                },
                "spec": {
                  "autogrowInitialIncrement": "1 GiB",
                  "autogrowMaxIncrement": "1 TiB",
                  "l4MaxMemoryPercent": 75,
                  "managementConsoleAddress": "",
                  "memoryPlusLocalMemoryMaxSize": "0 MiB"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: quay.io/kove/kove-operator@sha256:0fb5fa2e6702d5e5bf692c25f190f65d306fc72cdceebf0814ff64adcb5f6994
          createdAt: "2025-09-25T16:31:36Z"
          description: Configure containers with Kove® pooled memory
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: kove-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Kove
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configure containers with Kove® pooled memory
            displayName: Kove Memory Configuration
            kind: KoveMemoryConfig
            name: kovememoryconfigs.containers.kove.com
            version: v1
          - description: Configure cluster with Kove:SDM™
            displayName: Kove System Configuration
            kind: KoveSystem
            name: kovesystems.containers.kove.com
            version: v1
        description: |
          ## About the Operator
          The Kove:SDM™ operator integrates Kove's software-defined memory
          solution with Kubernetes and OpenShift clusters. The operator enables
          containers to transparently provision any amount of memory, enabling total
          container memory capacity to exceed physical hardware limitations.

          Use the Kove:SDM™ operator to:

          * Create custom resources to automatically allocate memory to specific
            containers
          * Increase container flexibility with dynamic memory capacity that
            auto-grows as needed
          * Scale up container density by removing memory as an orchestration
            bottleneck

          ## Using the Operator
          Using the operator is as simple as creating a custom resource which
          specifies an initial Kove memory allocation size, a namespace, and a
          container name. Containers in the specified namespace with a matching
          container name will automatically be provided with a Kove memory allocation
          which will auto-grow as needed.

          ## About Kove
          Kove makes memory a provisionable resource, just like storage, CPUs, and
          networking. Kove:SDM™ provides a transparent, linearly scalable
          memory resource accessible to any application without code changes, from
          bare-metal apps to virtual machines and containers.
        displayName: Kove:SDM™ Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kove
        - Pooled Memory
        - Memory Pool
        - Memory Target
        - External Memory
        - Memory
        links:
        - name: Kove
          url: https://kove.com
        maintainers:
        - email: contact-kove@kove.com
          name: Kove
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Kove
          url: https://kove.com
        relatedImages:
        - registry.access.redhat.com/ubi8/pause@sha256:eacbaaadc932633330e799eda24e9b731b5a4ba7ed823ee70791d21029724a9a
        - quay.io/kove/kove-agent@sha256:9d026b925e1d414a531530a30ed8b5f10b421ee12c7c13949ea0ee9907d36b75
        - quay.io/kove/kove-client-drivers@sha256:1c1786e1bbf16609218154126b4340b2f7c4b71cbd9045589402f75c032ea38d
        - quay.io/kove/kove-compatibility-detection@sha256:0303830a1f785b5e4ce6824db7c50f8feee2678fd5e87ded983686ab3f60b232
        - quay.io/kove/kove-operator@sha256:0fb5fa2e6702d5e5bf692c25f190f65d306fc72cdceebf0814ff64adcb5f6994
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fa0acefda43e309e7f9f31f5247c933d0b868ff1573ece2f38e0549d7e2ec942
        - quay.io/kove/kove-container-hook@sha256:66ee33d59ced63406d9b40f3a3226fab7f52071f8bed9aa1ba7d859837038528
        - quay.io/kove/kove-xmem@sha256:41bfc536a278153b0a4460d2836b56762b56cf50b94d17e123a63be066c28fe7
        - quay.io/kove/kove-client-drivers@sha256:716ccda2d8f082a954f2042fd189fec257bd2dcb2c50e497bad42b9a9e5f0e3d
        version: 3.0.2
      entries:
      - name: kove-operator.v3.0.2
        version: 3.0.2
      - name: kove-operator.v3.0.1
        version: 3.0.1
      - name: kove-operator.v3.0.0
        version: 3.0.0
      name: stable
    defaultChannel: stable
    packageName: kove-operator
    provider:
      name: Kove
      url: https://kove.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Fujitsu
      provider-url: https://www.postgresql.fastware.com/
    name: fujitsu-enterprise-postgres-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fujitsu-enterprise-postgres-operator.v7.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPAction",
                "metadata": {
                  "name": "new-fep-action"
                },
                "spec": {
                  "fepAction": {
                    "args": [
                      "new-fep-sts-0",
                      "new-fep-sts-1"
                    ],
                    "type": "reload"
                  },
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "targetClusterName": "new-fep"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPExporter",
                "metadata": {
                  "name": "new-fep-exporter"
                },
                "spec": {
                  "fepExporter": {
                    "exporterLogLevel": "error",
                    "fepClusterList": [
                      "new-fep1"
                    ],
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "userCustomQueries": "usr_example:\n  query: \"SELECT EXTRACT(EPOCH FROM (now() - pg_last_xact_replay_timestamp())) as lag\"\n  master: true\n  metrics:\n    - lag:\n        usage: \"GAUGE\"\n        description: \"Replication lag behind master in seconds\""
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPLogging",
                "metadata": {
                  "name": "new-fep-logging"
                },
                "spec": {
                  "fepLogging": {
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPPgpool2",
                "metadata": {
                  "name": "new-fep-pgpool2"
                },
                "spec": {
                  "clientAuthMethod": "scram",
                  "count": 2,
                  "customhba": "local   all         all                               trust\nhost    all         all         127.0.0.1/32          trust\nhost    all         all         ::1/128               trust\n",
                  "customlogsize": "128Mi",
                  "customparams": "listen_addresses = '*'\npcp_listen_addresses = '*'\nnum_init_children = 32\nreserved_connections = 0\nenable_pool_hba = off\nallow_clear_text_frontend_auth = off\nauthentication_timeout = 80\nbackend_weight0 = 1\nbackend_weight1 = 1\nbackend_flag0 = 'ALWAYS_PRIMARY'\nbackend_flag1 = 'DISALLOW_TO_FAILOVER'\nconnection_cache = on\nmax_pool = 4\nlisten_backlog_multiplier = 2\nserialize_accept = off\nchild_life_time = 300\nclient_idle_limit = 0\nchild_max_connections = 0\nconnection_life_time = 0\nreset_query_list = 'ABORT; DISCARD ALL'\nclient_min_messages = info\nlog_min_messages = debug1\nlog_statement = on\nlog_per_node_statement = on\nlog_client_messages = on\nlog_hostname = on\nlog_connections = on\nlog_line_prefix = '%t: pid %p: '\nload_balance_mode = on\nignore_leading_white_space = on\nwhite_function_list = ''\nblack_function_list = 'currval,lastval,nextval,setval'\nblack_query_pattern_list = ''\ndatabase_redirect_preference_list = ''\napp_name_redirect_preference_list = ''\nallow_sql_comments = off\ndisable_load_balance_on_write = 'transaction'\nstatement_level_load_balance = on\nsr_check_period = 0\nsr_check_user = 'postgres'\ndelay_threshold = 0\nlog_standby_delay = 'none'\nssl = on\nssl_ciphers = 'HIGH:MEDIUM:+3DES:!aNULL'\nssl_prefer_server_ciphers = off\nssl_ecdh_curve = 'prime256v1'\nssl_dh_params_file = ''\nrelcache_expire = 0\nrelcache_size = 256\ncheck_temp_table = catalog\ncheck_unlogged_table = on\nenable_shared_relcache = off\nrelcache_query_target = primary\nwd_port0 = 9000\nfailover_on_backend_error = off\n",
                  "custompcp": "none",
                  "customsslcacert": "none",
                  "customsslcert": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                  "customsslkey": "none",
                  "fepclustername": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "limits": {
                    "cpu": "400m",
                    "memory": "512Mi"
                  },
                  "requests": {
                    "cpu": "200m",
                    "memory": "256Mi"
                  },
                  "serviceport": 9999,
                  "statusport": 9898
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPRestore",
                "metadata": {
                  "name": "new-fep-restore"
                },
                "spec": {
                  "fromFEPcluster": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "mcSpec": {
                    "limits": {
                      "cpu": "200m",
                      "memory": "300Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "restoretype": "latest",
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "toFEPcluster": "new-fep-2"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v2",
                "kind": "FEPCluster",
                "metadata": {
                  "name": "new-fep"
                },
                "spec": {
                  "fep": {
                    "customAnnotations": {
                      "allDeployments": {}
                    },
                    "forceSsl": true,
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "instances": 1,
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "podAntiAffinity": false,
                    "podDisruptionBudget": false,
                    "servicePort": 27500,
                    "syncMode": "off",
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "usePodName": true
                  },
                  "fepChildCrVal": {
                    "backup": {
                      "image": {
                        "pullPolicy": "IfNotPresent"
                      },
                      "mcSpec": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      },
                      "pgbackrestParams": "# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[global]\nrepo1-retention-full=7\nrepo1-retention-full-type=time\nlog-path=/database/log/backup\n",
                      "postScript": " ",
                      "preScript": " ",
                      "schedule": {
                        "num": 2
                      },
                      "schedule1": {
                        "schedule": "15 0 * * 0",
                        "type": "full"
                      },
                      "schedule2": {
                        "schedule": "15 0 * * 1-6",
                        "type": "incr"
                      },
                      "schedule3": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule4": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule5": {
                        "schedule": " ",
                        "type": " "
                      }
                    },
                    "customPgAudit": "# define pg audit custom params here to override defaults.\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[output]\nlogger = 'auditlog'\nlog_directory = '/database/log/audit'\nlog_truncate_on_rotation = on\nlog_filename = 'pgaudit-%a.log'\nlog_rotation_age = 1d\nlog_rotation_size = 0\n[rule]\n",
                    "customPgHba": "# define pg_hba custom rules here to be merged with default rules.\n# TYPE     DATABASE        USER        ADDRESS        METHOD\n",
                    "customPgParams": "# define custom postgresql.conf parameters below to override defaults.\n# Current values are as per default FEP deployment\n# If you add a library to shared_preload_libraries, add it after the default value.\n# fsep_operator_security is available in FEP15 and later container images.\nshared_preload_libraries='pgx_datamasking,pg_prewarm,pg_stat_statements,fsep_operator_security'\nsession_preload_libraries='pg_prewarm'\nmax_prepared_transactions = 100\nmax_worker_processes = 30\nmax_connections = 100\nwork_mem = 1MB\nmaintenance_work_mem = 12MB\nshared_buffers = 128MB\neffective_cache_size = 384MB\ncheckpoint_completion_target = 0.8\n# tcp parameters\ntcp_keepalives_idle = 30\ntcp_keepalives_interval = 10\ntcp_keepalives_count = 3\n# logging parameters in default fep installation\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\nlog_directory = '/database/log'\nlog_filename = 'logfile-%a.log'\nlog_file_mode = 0600\nlog_truncate_on_rotation = on\nlog_rotation_age = 1d\nlog_rotation_size = 0\nlog_checkpoints = on\nlog_line_prefix = '%e %t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h'\nlog_lock_waits = on\nlog_autovacuum_min_duration = 60s\nlogging_collector = on\npgaudit.config_file='/opt/app-root/src/pgaudit-cfg/pgaudit.conf'\nlog_replication_commands = on\nlog_min_messages = WARNING\nlog_destination = csvlog\n# wal_archive parameters in default fep installation\narchive_mode = on\narchive_command = 'pgbackrest --stanza=backupstanza --config=/database/userdata/pgbackrest.conf archive-push %p'\nwal_level = replica\nmax_wal_senders = 12\nwal_keep_segments = 64\ntrack_activities = on\ntrack_counts = on\npassword_encryption = 'scram-sha-256'\n",
                    "storage": {
                      "archivewalVol": {
                        "size": "1Gi"
                      },
                      "backupVol": {
                        "size": "2Gi"
                      },
                      "dataVol": {
                        "size": "2Gi"
                      },
                      "logVol": {
                        "size": "1Gi"
                      },
                      "tablespaceVol": {
                        "size": "512Mi"
                      },
                      "walVol": {
                        "size": "1200Mi"
                      }
                    },
                    "sysUsers": {
                      "pgRewindPassword": "rewind_password",
                      "pgRewindUser": "rewind_user",
                      "pgdb": "mydb",
                      "pgpassword": "mydbpassword",
                      "pgreplpassword": "repluserpwd",
                      "pgrepluser": "repluser",
                      "pguser": "mydbuser",
                      "tdepassphrase": "tde-passphrase"
                    },
                    "systemCertificates": {
                      "cacrt": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                      "crt": "-----BEGIN CERTIFICATE-----\nMIIDUTCCAjmgAwIBAgIRAMocW3qMoHrD6qRvMPppMkMwDQYJKoZIhvcNAQELBQAw\nNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9yIEt1\nYmVybmV0ZXMwHhcNMjEwMjA2MDQzMjM2WhcNMjYwMjA1MDQzMjM2WjA/MRAwDgYD\nVQQKEwdGdWppdHN1MSswKQYDVQQDEyJGVUpJVFNVIEVudGVycHJpc2UgUG9zdGdy\nZXMgU2VydmVyMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA4AI33yvH\nZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I2e4SceTKi6O3C/I1XuvWlpng\n5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4ANB5JWWqDOjrRT3o7nRPGXfila\nbP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYpmjdbfxabTz07ig0+6/cwKoRR\nxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTFYvmAH7gcdssSFBt8NPlUATHE\nsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6Wdgmu5H2pDml8CDNLDv98Aj7i\n+I5SRKKcVPlnuQIDAQABo1AwTjAdBgNVHSUEFjAUBggrBgEFBQcDAQYIKwYBBQUH\nAwIwDAYDVR0TAQH/BAIwADAfBgNVHSMEGDAWgBQcwrrUO0u+FhIUuVdrDRCQRsi6\nZjANBgkqhkiG9w0BAQsFAAOCAQEAm5dxBoI9pScOCvRAchg4CprdRDSJb9K6yB3O\nnCAxnM47iHeXnY3WlnI388kHu8DU7O4ba1tJbGs3KY9KzioPk43pU12jWkO1onoF\n+mTDjx/Ef1cYWA9r5q/LtgTa6Q2sxV4O2x67QW82aAnaxO34dV5zWCPIvAoovZBV\nHRT+BgCg3r2vD1RGKK2nl1aYJtWhO1SZubam+VttdZ/vbM9oOJctxmImsEtBXjkY\nKteePdQtLL5o03JhyXWyRshCq+HMmKf2KgyY8gvydGcP4eLQdBWcW40LcnVq6UjT\n0kJycJEKngMVademq1ZWHGaiYB7hyT6GhgIcHUJ2cKrPgbEh1Q==\n-----END CERTIFICATE-----",
                      "key": "-----BEGIN RSA PRIVATE KEY-----\nMIIEowIBAAKCAQEA4AI33yvHZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I\n2e4SceTKi6O3C/I1XuvWlpng5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4AN\nB5JWWqDOjrRT3o7nRPGXfilabP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYp\nmjdbfxabTz07ig0+6/cwKoRRxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTF\nYvmAH7gcdssSFBt8NPlUATHEsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6W\ndgmu5H2pDml8CDNLDv98Aj7i+I5SRKKcVPlnuQIDAQABAoIBAFPQYKlOzw/+BA0b\nyMIUpdctIMb/54CR/xR0mVw1DbSjigNVPjHUQvB8Y1B2FAITQObgJO06bAv0QdWN\nRb0/v/yYiNJDFjaLjaIAHlO/2+oWrXbFaZqgpVDJhB+e1xaZr2x7XGxm+p925k30\nl6pvIRY+I8JRKvZiV1VZHwL/R3JOtPr++xMZtLVjVOI+f+ySqJ+TZHuAjm49EKxj\ncEmmJ28b7QcziXsvKy00f+zbqLIBKXQdZAFU5eEr1BsDRXdRW+Kf0XIvftuy4BJZ\nvoKT+VGhEvF/qysswL4+6IAO6tpuYnnM0Y2d3sOGoWPkTcQK0MekYKzL/WmtCjNs\n9hodJtECgYEA5EWyhEOf4uOKe5TDp697UCUvXLoOR58FDe/S8XNvScn29jjOkqIg\nOMoqo9xAkJTNTzqn5UUdt1x/pgM2NxlPLFijrc0zQlX3SoOO2ryDd9WNi7YKtN16\nKJqa536WeZu2OEbuAZ+S3GALVy1RPeTNPnUOmKnF06DjDUGzLNCZy10CgYEA+zfw\n952DWuz1U0Z4wvAEqqcgUKXPKrkTXV/iUnjkDkrLYVr0ZofDNTXrdHl+UedFmaOC\ncieZn6DNhcdz5tKtyysGMH3g/qs9PfoGUngvcXsy0Egk04l3x1jc8TTCLqXZXYaQ\nHMsx51n+R58oncPtzYSUOr9qQ6PbC2CstTbFJA0CgYEAjGEsUliAB/jknfEzjXjG\nPdhQUxb8VyE864Az2lah9t/kJzFyIAziAeqZ5GE7t247AGFTBRTHHI8e1Qoemi3P\nWbc9GVIbFs1lIYbcIDpUIyrKPEP8O5QEXtoNLxXTFgAjRGKiVY87spjCAJ+W2ZhO\ne/1it5GYXfgQCYQA2yuBmOUCgYANRkR2YR1axaCk+NlSu6oTdmdPu6M5x7PNQE7O\nOtMaKjua9lppvIzFGAdMDUtueoEEAE7ZR1xnwfB6PDLUpJdIYAqgr1YfPt8qkjaZ\nTv56yZ7CwL0pbF8m6nwqRrZoDp1wwraEvvvxFKFKGY/k3kCHlpTakdjEoDjn3gDi\nRnWeVQKBgCEneMSzucei5LRppRtRaJw/Btll8qlPMlX3W7dxQ3cLwpmLOn0m51Fp\nPIZ44zYK8R6fu4+/sSrlfaIg86Ugeufp6YNxyNROKxUGza5vDIu5OftwWtBeg+UK\nZ8lLWNdX6pp7WMujmF3H1DrkBbauYMUKZ4UxUYtelgHERMePIxwb\n-----END RSA PRIVATE KEY-----"
                    }
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "false"
          containerImage: quay.io/fujitsu/fujitsu-enterprise-postgres-operator@sha256:b7dd55c0bfda84e398cbfbb87345fcec218cb866e286e18a111330321ab6ad54
          createdAt: "2025-09-10T11:50:37Z"
          description: OpenShift Operator for Fujitsu Enterprise Postgres
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.1.0 <7.3.0'
          operators.openshift.io/valid-subscription: '["Fujitsu Enterprise Postgres
            Advanced Edition Annual Subscription License per Core 17 Operator for
            Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/internal-objects: '["fepautoscales.fep.fujitsu.io","fepconfigs.fep.fujitsu.io","fepusers.fep.fujitsu.io","fepcerts.fep.fujitsu.io","fepvolumes.fep.fujitsu.io","fepbackups.fep.fujitsu.io"]'
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: fj-dbaas-tec@dl.jp.fujitsu.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: FEPAction
            name: fepactions.fep.fujitsu.io
            version: v1
          - kind: FEPAutoscale
            name: fepautoscales.fep.fujitsu.io
            version: v1
          - kind: FEPBackup
            name: fepbackups.fep.fujitsu.io
            version: v1
          - kind: FEPCert
            name: fepcerts.fep.fujitsu.io
            version: v1
          - kind: FEPCluster
            name: fepclusters.fep.fujitsu.io
            version: v2
          - kind: FEPConfig
            name: fepconfigs.fep.fujitsu.io
            version: v1
          - kind: FEPExporter
            name: fepexporters.fep.fujitsu.io
            version: v1
          - kind: FEPLogging
            name: feploggings.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2Cert
            name: feppgpool2certs.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2
            name: feppgpool2s.fep.fujitsu.io
            version: v1
          - kind: FEPRestore
            name: feprestores.fep.fujitsu.io
            version: v1
          - kind: FEPUpgrade
            name: fepupgrades.fep.fujitsu.io
            version: v1
          - kind: FEPUser
            name: fepusers.fep.fujitsu.io
            version: v1
          - kind: FEPVolume
            name: fepvolumes.fep.fujitsu.io
            version: v1
        description: "Fujitsu Enterprise Postgres 17 delivers an enterprise-grade
          PostgreSQL on OpenShift Container Platform.\n\nThis solution provides the
          flexibility of a hybrid cloud solution while delivering an enhanced distribution
          of PostgreSQL to support enterprise-level workloads and provide improved
          deployment and management, availability, performance, data governance and
          security.  \n\nAvailable as a multi-architecture container built for both
          amd64, s390x and ppc64le.\n\nThe download and Use of the Product is strictly
          subject to the terms of the End User License Agreement with Fujitsu Limited
          found at https://www.fast.fujitsu.com/fujitsu-enterprise-postgres-license-agreements.
          Where the Product that has been embedded as a whole or part into a third
          party program, only Authorised Customers may download and use the Product.\n"
        displayName: Fujitsu Enterprise Postgres Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - Fujitsu
        - fep
        links:
        - name: Fujitsu Enterprise Postgres
          url: https://www.postgresql.fastware.com/
        maintainers:
        - email: fj-dbaas-tec@dl.jp.fujitsu.com
          name: Fujitsu
        maturity: stable
        provider:
          name: Fujitsu
          url: https://www.postgresql.fastware.com/
        relatedImages:
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-upgrade@sha256:bf3b069a71202ea78c7c0c7b36315c8bef8e6a1c11cd0668dcd888329e26ecfb
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-backup@sha256:dd8b84e200c9d0fb32cd567a33968e15e15efc115ff120f54fdd8862c7ea3111
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentbit@sha256:9622d2191525da39e5100e1882dd2cfa3eda740d8094f3fbb71ca8a5734d2843
        - quay.io/fujitsu/fujitsu-enterprise-postgres-utils@sha256:57a3fb792480a0d3a8f075c48dee2d02f1d79cf459761e3305cf62799575a885
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-backup@sha256:42252444e2dcfe5435532f318cce041990d2b71efa26007924717924d30a46ca
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-utils@sha256:edeec4d00fd8282aaa7ec6fe3ec4cd84070a6402611002a18addcac87fb8ee94
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-server@sha256:fa4e8849d75daf331ba7d488bcb52c083cd92bb5f288e4b62c980754e299f618
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-pgpool2@sha256:185af2e4fd1701aded1ff5e44d4336562f5a219c30ba535c2a3097b914c24efe
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-upgrade@sha256:04af997b4f548ffec711b63a06c35715b4e7863537c83697b3ee0c1d95a6aec3
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-pgpool2@sha256:8eaace2f9c8e5e844992fe5e008cd1061af33cc79b4d442c4a83a65610bcc502
        - quay.io/fujitsu/fujitsu-enterprise-postgres-exporter@sha256:1c0e22aeab65ff3848772424fc3ea6baa6b58b0d88db5413bda160a08435bcff
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-upgrade@sha256:8a0f7ca1f59ba162424a308e8a70c4717135e2847e61210753b71876c9b0a75f
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-restore@sha256:09be02274427e3022849fb635a4c9abdc34696838cae39f3afafc8009567db9d
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-backup@sha256:78a3ad639d0c63858b9db0a5de36ce11a9ea121f4535e9ca07ac7defb689bc0d
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-server@sha256:3e93a368a6198818e1b7c5525324c8130c5ed8d7b4cb499a47d4bec45ef0347b
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-pgpool2@sha256:c878935ffc796c405d811a6e8a9194aaaaa8652ea7d7b879c82919ce09794b35
        - quay.io/fujitsu/fujitsu-enterprise-postgres-cronjob@sha256:7582964eef14f89f2ca9046298357afd0132fb236f30f44f84cee14692227b0c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-restore@sha256:a81a0ce926572b75aa3151e2d2cc45a0ca411cc0158d69b89247ef70418f3740
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-server@sha256:1c5823afc7e77d46940214693c12c0ad6ffd6c7d9a156e26399ab55cd5ea1dfa
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-restore@sha256:17301c2191c0d75be82d12c5a35c0f74b8d65304bcfb6d4ab312a10ffa286eda
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-restore@sha256:a6e90dd77634033f2009942b635078cceb0f4c6b15b031624c5dce820713b14c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-server@sha256:5c65f8e76f70341c512100760de6792a880ad0d6eebc9e68154b5bbd1332a896
        - quay.io/fujitsu/fujitsu-enterprise-postgres-operator@sha256:b7dd55c0bfda84e398cbfbb87345fcec218cb866e286e18a111330321ab6ad54
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-utils@sha256:b7ae535f5e5fa5eb51c6f4cdb406c7b7a9dabe8573939b0431dada6f7d27b87e
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentd@sha256:3c5534b62695867c5973c9106fc46895ed73a4b9136ad83b2586bf7e71c427d0
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-pgpool2@sha256:38aac8108dd3590abcfd854a6ec2c025a68d2170a1ba68d359b45fea6c83cec1
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-backup@sha256:ac268e2af652bdabb8ee5d764a2b335dc73c48b9022eee8c9f3cd422227a4d98
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-upgrade@sha256:c0e44447349f02eb9f6105001136402f849d876a288d61997c8c05625facff45
        version: 7.3.0
      entries:
      - name: fujitsu-enterprise-postgres-operator.v7.3.0
        version: 7.3.0
      - name: fujitsu-enterprise-postgres-operator.v7.2.7
        version: 7.2.7
      name: stable
    defaultChannel: stable
    packageName: fujitsu-enterprise-postgres-operator
    provider:
      name: Fujitsu
      url: https://www.postgresql.fastware.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-block-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12
    - currentCSV: ibm-block-csi-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "k8s.gcr.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
        version: 1.12.0
      entries:
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.0
    - currentCSV: ibm-block-csi-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
        version: 1.12.1
      entries:
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.1
    - currentCSV: ibm-block-csi-operator.v1.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:af9215670675fe9c4fa2dac7a14d547c249f3aa0c0e72f1e5390d237e411735b
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:af9215670675fe9c4fa2dac7a14d547c249f3aa0c0e72f1e5390d237e411735b
        version: 1.12.2
      entries:
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.2
    - currentCSV: ibm-block-csi-operator.v1.12.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:6e7e13d259f4d743524ebcd546af5c73d1a3ba90edaf4b5edd2005a3f8adcdfe
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:6e7e13d259f4d743524ebcd546af5c73d1a3ba90edaf4b5edd2005a3f8adcdfe
        version: 1.12.3
      entries:
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.3
    - currentCSV: ibm-block-csi-operator.v1.12.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:42728e7fd1d2302511ab08f51335ab188eef125870fa27f322df83ebb8e1f447
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:42728e7fd1d2302511ab08f51335ab188eef125870fa27f322df83ebb8e1f447
        version: 1.12.4
      entries:
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.4
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.5
    defaultChannel: stable
    packageName: ibm-block-csi-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cli-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cli-manager-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "CliManager",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-cli-manager-operator"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed",
                  "operatorLogLevel": "Normal"
                }
              }
            ]
          capabilities: Basic Install
          categories: Openshift Optional
          createdAt: "2024-06-04"
          description: "The CLI Manager is a comprehensive tool designed to simplify
            the management of \nOpenShift CLI plugins within the OpenShift environment.
            \nModeled after the popular krew plugin manager, it offers seamless integration,
            \neasy installation, and efficient handling of a wide array of plugins,
            \nenhancing your OpenShift command-line experience.\n"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          repository: https://github.com/openshift/cli-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: CLI Manager
            kind: CliManager
            name: climanagers.operator.openshift.io
            version: v1
          - displayName: CLI Plugin
            kind: Plugin
            name: plugins.config.openshift.io
            version: v1alpha1
        description: "The CLI Manager is a comprehensive tool designed to simplify
          the management of \nOpenShift CLI plugins within the OpenShift environment.
          \nModeled after the popular krew plugin manager, it offers seamless integration,
          \neasy installation, and efficient handling of a wide array of plugins,
          \nenhancing your OpenShift command-line experience.\n"
        displayName: CLI Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cli-manager-operator
        - cli-manager
        links:
        - name: Source Code
          url: https://github.com/openshift/cli-manager
        - name: Source Code
          url: https://github.com/openshift/cli-manager-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/cli-manager-operator/cli-manager-rhel9-operator@sha256:31973d7f43d052ddfb9554dd623e37fb541a8c1f08ea981558eaf4bb5ab22d5a
        - registry.redhat.io/cli-manager-operator/cli-manager-rhel9@sha256:07853bc33a0ce783d8d45c08c161e454a3abe09a8a8dadeca70c3a199466b58b
        version: 0.1.0
      entries:
      - name: cli-manager-operator.v0.1.0
        version: 0.1.0
      name: tech-preview
    defaultChannel: tech-preview
    packageName: cli-manager
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cryostat-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cryostat-operator.v4.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.cryostat.io/v1beta2",
                "kind": "Cryostat",
                "metadata": {
                  "name": "cryostat-sample"
                },
                "spec": {
                  "enableCertManager": true,
                  "eventTemplates": [],
                  "networkPolicies": {},
                  "reportOptions": {
                    "replicas": 0
                  },
                  "trustedCertSecrets": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Developer Tools
          containerImage: registry.redhat.io/cryostat/cryostat-rhel9-operator@sha256:867a649b54ad144295c7aceda6b90631841aa31a62062df62d4980313723f359
          createdAt: "2025-09-30T15:10:08Z"
          description: JVM monitoring and profiling tool
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.0 <4.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "operator.cryostat.io/v1beta2",
              "kind": "Cryostat",
              "metadata": {
                "name": "cryostat-sample"
              },
              "spec": {
                "enableCertManager": true,
                "reportOptions": {
                  "replicas": 0
                }
              }
            }
          operators.openshift.io/valid-subscription: '["Red Hat build of OpenJDK"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/cryostatio/cryostat-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cryostat allows you to install Cryostat for a single namespace,
              or multiple namespaces. It contains configuration options for controlling
              the Deployment of the Cryostat application and its related components.
              A Cryostat instance must be created to instruct the operator to deploy
              the Cryostat application.
            displayName: Cryostat
            kind: Cryostat
            name: cryostats.operator.cryostat.io
            version: v1beta2
          - description: Cryostat allows you to install Cryostat for a single namespace.
              It contains configuration options for controlling the Deployment of
              the Cryostat application and its related components. A Cryostat instance
              must be created to instruct the operator to deploy the Cryostat application.
            displayName: Cryostat
            kind: Cryostat
            name: cryostats.operator.cryostat.io
            version: v1beta1
        description: |
          The Red Hat build of Cryostat provides a cloud-based solution for interacting with the JDK Flight Recorder already present in OpenJDK 11+ JVMs. With Cryostat, users can remotely start, stop, retrieve, and even analyze JFR event data, providing the capability to easily take advantage of Flight Recorder's extremely low runtime cost and overhead and the flexibility to monitor applications and analyze recording data without transferring data outside of the cluster the application runs within.
          ##Prerequisites
          Cryostat requires [cert-manager](https://cert-manager.io/) to run. If not already installed in your cluster, please [install](https://cert-manager.io/docs/installation/) it using your preferred method.
          When running on Kubernetes, the operator also requires [Ingress configurations](https://github.com/cryostatio/cryostat-operator/blob/v4.0.0/docs/config.md#network-options) for each of its services to make them available outside of the cluster. The user is responsible for providing the hostnames for each Ingress.
        displayName: Red Hat build of Cryostat
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flightrecorder
        - java
        - jdk
        - jfr
        - jmc
        - missioncontrol
        - monitoring
        - profiling
        - diagnostic
        links:
        - name: Upstream Project
          url: https://github.com/cryostatio/cryostat
        - name: Website
          url: https://cryostat.io/
        maintainers:
        - email: cryostat-development@googlegroups.com
          name: The Cryostat Authors
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi9/nginx-124@sha256:7c1d7172880a3649591a0dc1ceb670c133ef88eed3d8192d8a2e80e3878d04c3
        - registry.redhat.io/cryostat/cryostat-reports-rhel9@sha256:5535a539b185fa2781621741ccd7f10a6b9947adc6170a95e930f2d9d7a9f608
        - registry.redhat.io/insights-runtimes-tech-preview/runtimes-inventory-rhel8-operator@sha256:58e01d11c3d30abc5b0ee3795511bd04e39aea21e2853b4f4cc6d1c8a2bd7bb0
        - registry.redhat.io/cryostat/cryostat-rhel9@sha256:fadae8713daf82c329573191381f8f9e4ebf3fc852ac5f1a68ce0ea1da376469
        - registry.redhat.io/cryostat/jfr-datasource-rhel9@sha256:1e5f29aa096d783fd0a18b3ba31c391dc203761ba3f4af8e5aeeef1f7cc86cee
        - registry.redhat.io/cryostat/cryostat-db-rhel9@sha256:23d61204ecaf037cf17bf9a5cc4428461a60fecce077884d9458802fb071daab
        - registry.redhat.io/cryostat/cryostat-grafana-dashboard-rhel9@sha256:5715e897f9a8969aedd0066b2ba597264495b4c85e4ae8ae7d8871f93d13f4c1
        - registry.redhat.io/cryostat/cryostat-agent-init-rhel9@sha256:273d6c3d3194a4e9ec50c1d8e147cd9bc6d277d8d66878d885a18528b9ba7dc6
        - registry.redhat.io/cryostat/cryostat-rhel9-operator@sha256:867a649b54ad144295c7aceda6b90631841aa31a62062df62d4980313723f359
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        - registry.redhat.io/cryostat/cryostat-ose-oauth-proxy-rhel9@sha256:d38ce4dcac5b073b001d886f0494e0fc9b750f8bd68b855fc103767a1ed8b7df
        - registry.redhat.io/cryostat/cryostat-storage-rhel9@sha256:f642ca8a0ca0aa130462c0f423ce4c775cb296f8a40c9270e5369176e87c7ee6
        - registry.redhat.io/cryostat/cryostat-openshift-console-plugin-rhel9@sha256:ca748f96414ad648eb1e3953aeeeb828b187f3c5c30e014513adc3619a698683
        version: 4.0.3
      entries:
      - name: cryostat-operator.v4.0.3
        version: 4.0.3
      - name: cryostat-operator.v4.0.2-4
        version: 4.0.2+4
      - name: cryostat-operator.v4.0.2
        version: 4.0.2
      - name: cryostat-operator.v4.0.1
        version: 4.0.1
      - name: cryostat-operator.v4.0.1-5
        version: 4.0.1+5
      - name: cryostat-operator.v4.0.1-4
        version: 4.0.1+4
      - name: cryostat-operator.v4.0.1-3
        version: 4.0.1+3
      - name: cryostat-operator.v4.0.1-2
        version: 4.0.1+2
      - name: cryostat-operator.v4.0.0-11
        version: 4.0.0+11
      - name: cryostat-operator.v4.0.0
        version: 4.0.0
      - name: cryostat-operator.v4.0.0-9
        version: 4.0.0+9
      - name: cryostat-operator.v4.0.0-10
        version: 4.0.0+10
      - name: cryostat-operator.v3.0.1-7
        version: 3.0.1+7
      - name: cryostat-operator.v3.0.1-9
        version: 3.0.1+9
      - name: cryostat-operator.v3.0.1-8
        version: 3.0.1+8
      - name: cryostat-operator.v3.0.1
        version: 3.0.1
      - name: cryostat-operator.v3.0.1-6
        version: 3.0.1+6
      - name: cryostat-operator.v3.0.1-5
        version: 3.0.1+5
      - name: cryostat-operator.v3.0.1-3
        version: 3.0.1+3
      - name: cryostat-operator.v3.0.1-13
        version: 3.0.1+13
      - name: cryostat-operator.v3.0.1-12
        version: 3.0.1+12
      - name: cryostat-operator.v3.0.1-11
        version: 3.0.1+11
      - name: cryostat-operator.v3.0.1-10
        version: 3.0.1+10
      - name: cryostat-operator.v3.0.0-12
        version: 3.0.0+12
      - name: cryostat-operator.v3.0.0
        version: 3.0.0
      - name: cryostat-operator.v3.0.0-6
        version: 3.0.0+6
      - name: cryostat-operator.v3.0.0-8
        version: 3.0.0+8
      - name: cryostat-operator.v2.4.0
        version: 2.4.0
      - name: cryostat-operator.v2.4.0-8
        version: 2.4.0+8
      - name: cryostat-operator.v2.4.0-7
        version: 2.4.0+7
      - name: cryostat-operator.v2.4.0-6
        version: 2.4.0+6
      - name: cryostat-operator.v2.4.0-4
        version: 2.4.0+4
      - name: cryostat-operator.v2.4.0-3
        version: 2.4.0+3
      - name: cryostat-operator.v2.3.1-6
        version: 2.3.1+6
      - name: cryostat-operator.v2.3.1-8
        version: 2.3.1+8
      - name: cryostat-operator.v2.3.1-9
        version: 2.3.1+9
      - name: cryostat-operator.v2.3.1
        version: 2.3.1
      - name: cryostat-operator.v2.3.0
        version: 2.3.0
      - name: cryostat-operator.v2.3.0-6
        version: 2.3.0+6
      - name: cryostat-operator.v2.2.1-14
        version: 2.2.1+14
      - name: cryostat-operator.v2.2.1-8
        version: 2.2.1+8
      - name: cryostat-operator.v2.2.1-5
        version: 2.2.1+5
      - name: cryostat-operator.v2.2.1-13
        version: 2.2.1+13
      - name: cryostat-operator.v2.2.1-12
        version: 2.2.1+12
      - name: cryostat-operator.v2.2.1-10
        version: 2.2.1+10
      - name: cryostat-operator.v2.2.1
        version: 2.2.1
      - name: cryostat-operator.v2.2.0
        version: 2.2.0
      - name: cryostat-operator.v2.1.1-2.1659469835.p
        version: 2.1.1+2.1659469835.p
      - name: cryostat-operator.v2.1.1-2.1655334614.p
        version: 2.1.1+2.1655334614.p
      - name: cryostat-operator.v2.1.1
        version: 2.1.1
      - name: cryostat-operator.v2.1.0
        version: 2.1.0
      name: stable
    defaultChannel: stable
    packageName: cryostat-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-xtradb-cluster-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.17.0
        version: 1.17.0
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.16.1
        version: 1.16.1
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.16.0
        version: 1.16.0
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.15.1
        version: 1.15.1
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.15.0
        version: 1.15.0
      name: stable
    - currentCSV: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        version: 1.18.0-cw
      entries:
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0-cw
        version: 1.18.0-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.17.0-cw
        version: 1.17.0-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.16.1-cw
        version: 1.16.1-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.16.0-cw
        version: 1.16.0-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.15.0-cw
        version: 1.15.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator-certified-rhmp
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: community-windows-machine-config-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: community-windows-machine-config-operator.v6.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-windows/community-windows-machine-config-operator:community-4.11-6c3601f
          createdAt: "2022-08-31T18:57:22Z"
          description: An operator that enables Windows container workloads on OCP
          olm.skipRange: '>=5.0.0 <6.0.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ## WARNING
          Community distribution of the Windows Machine Config Operator.
          This is a preview build, and is not meant for production workloads.
          Issues with this distribution of WMCO can be opened against the [WMCO repository](https://github.com/openshift/windows-machine-config-operator).
          Please read through the [troubleshooting doc](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/docs/TROUBLESHOOTING.md)
          before opening an issue.
          Please ensure that when installing this operator the starting CSV you subscribe to is supported on the
          version of OKD/OCP you are using. This CSV is meant for OKD/OCP community-4.11.
          ## Documentation
          ### Introduction
          The Windows Machine Config Operator configures Windows instances into nodes, enabling Windows container workloads
          to be ran within OKD/OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/latest/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/README.md#configuring-byoh-bring-your-own-host-windows-instances).
          The operator will do all the necessary steps to configure the instance so that it can join the cluster as a worker node.
          ### Pre-requisites
          - [Cluster and OS pre-requisites](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/docs/wmco-prerequisites.md)
          ### Usage
          Please see the usage section of [README.md](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/README.md#usage).
          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.
        displayName: Community Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: preview
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-windows/community-windows-machine-config-operator:community-4.11-6c3601f
        version: 6.0.0
      entries:
      - name: community-windows-machine-config-operator.v6.0.0
        version: 6.0.0
      name: preview
    defaultChannel: preview
    packageName: community-windows-machine-config-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom
      provider-url: https://www.broadcom.com/
    name: layer7-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: layer7-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "security.brcmlabs.com/v1",
                "kind": "Gateway",
                "metadata": {
                  "name": "ssg"
                },
                "spec": {
                  "app": {
                    "image": "docker.io/caapim/gateway:11.1.2",
                    "management": {
                      "cluster": {
                        "hostname": "gateway.brcmlabs.com",
                        "password": "7layer"
                      },
                      "password": "7layer",
                      "username": "admin"
                    },
                    "replicas": 1,
                    "service": {
                      "ports": [
                        {
                          "name": "https",
                          "port": 8443,
                          "protocol": "TCP",
                          "targetPort": 8443
                        },
                        {
                          "name": "management",
                          "port": 9443,
                          "protocol": "TCP",
                          "targetPort": 9443
                        }
                      ],
                      "type": "LoadBalancer"
                    }
                  },
                  "license": {
                    "accept": false,
                    "secretName": "gateway-license"
                  },
                  "version": "11.1.2"
                },
                "status": {}
              },
              {
                "apiVersion": "security.brcmlabs.com/v1",
                "kind": "Repository",
                "metadata": {
                  "name": "my-repository"
                },
                "spec": {
                  "auth": {},
                  "branch": "main",
                  "enabled": true,
                  "endpoint": "https://github.com/\u003cusername\u003e/\u003crepository\u003e",
                  "type": "git"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7Api",
                "metadata": {
                  "name": "l7api-sample"
                },
                "spec": {
                  "deploymentTags": [
                    "ssg",
                    "ssg1"
                  ],
                  "graphmanBundle": "jsongz",
                  "portalPublished": false,
                  "serviceUrl": "/v1/api"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7Portal",
                "metadata": {
                  "name": "dev-portal"
                },
                "spec": {
                  "auth": {},
                  "deploymentTags": [
                    "ssg"
                  ],
                  "enabled": true,
                  "endpoint": "dev-ssg.brcmlabs.com",
                  "enrollmentBundle": "",
                  "portalTenant": "portal"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7StateStore",
                "metadata": {
                  "name": "state-store-example"
                },
                "spec": {
                  "redis": {
                    "database": 0,
                    "groupName": "l7gw",
                    "masterPassword": "7layer",
                    "standalone": {
                      "host": "redis-standalone",
                      "port": 6379
                    },
                    "storeId": "test",
                    "type": "standalone"
                  },
                  "type": "redis"
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: docker.io/caapim/layer7-operator:v1.2.1
          createdAt: "2025-07-04T17:22:57Z"
          description: The Layer7 Operator covers all aspects of deploying, maintaining
            and upgrading Layer7 API Gateways in Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: layer7-operator-system
          operators.openshift.io/valid-subscription: Layer7 (Broadcom) API Gateway
            license
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/caapim/layer7-operator
          support: Broadcom Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gateway is the Schema for the Gateway Custom Resource
            displayName: Gateway
            kind: Gateway
            name: gateways.security.brcmlabs.com
            version: v1
          - description: L7Api is the Schema for the l7apis API
            displayName: L7 Api
            kind: L7Api
            name: l7apis.security.brcmlabs.com
            version: v1alpha1
          - description: L7Portal is the Schema for the l7portals API
            displayName: L7 Portal
            kind: L7Portal
            name: l7portals.security.brcmlabs.com
            version: v1alpha1
          - description: L7StateStore is the Schema for the l7statestores API
            displayName: L7 State Store
            kind: L7StateStore
            name: l7statestores.security.brcmlabs.com
            version: v1alpha1
          - description: Repository is the Schema for the repositories API
            displayName: Repository
            kind: Repository
            name: repositories.security.brcmlabs.com
            version: v1
        description: |
          The [Layer7 Operator](https://github.com/CAAPIM/layer7-operator) covers all aspects of deploying, maintaining and upgrading Layer7 API Gateways in Kubernetes. The Layer7 Operator is exclusive to Kubernetes and utilizes the Container Gateway with specific emphasis on running in Ephemeral mode (without an external MySQL policy database).

          ## Getting Started
          This [Getting Started](https://github.com/CAAPIM/layer7-operator/wiki/Getting-Started) guide covers a simple introduction to the Layer7 Operator. You can find more documentation [here](https://github.com/CAAPIM/layer7-operator/wiki).
        displayName: Layer7 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - layer7-operator
        links:
        - name: Layer7 Operator
          url: https://github.com/caapim/layer7-operator
        maintainers:
        - email: gary.vermeulen@broadcom.com
          name: Gary Vermeulen
        maturity: alpha
        provider:
          name: Broadcom
          url: https://www.broadcom.com/
        relatedImages:
        - docker.io/caapim/graphman-static-init:1.0.3
        - docker.io/caapim/gateway:11.1.2
        - docker.io/caapim/layer7-operator:v1.2.1
        version: 1.2.1
      entries:
      - name: layer7-operator.v1.2.1
        version: 1.2.1
      - name: layer7-operator.v1.2.0
        version: 1.2.0
      name: preview
    defaultChannel: preview
    packageName: layer7-operator
    provider:
      name: Broadcom
      url: https://www.broadcom.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: eG Innovations
      provider-url: https://www.eginnovations.com
    name: eginnovations-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eginnovations-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eginnovations.com/v1beta1",
                "kind": "EgUniversalAgent",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eginnovations-operator"
                  },
                  "name": "eguniversalagent",
                  "namespace": "egagent"
                },
                "spec": {
                  "agent": {
                    "env": [
                      {
                        "name": "EG_MANAGER",
                        "value": "Replace with eG Manager IP"
                      },
                      {
                        "name": "EG_MANAGER_PORT",
                        "value": "Replace with eG Manager Port"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "true"
          containerImage: docker.io/eginnovations/universal-agent-operator@sha256:b8e3e26dca1b94fabfd79496feb36f3f2de4434331ef9ca810125892a936dca7
          createdAt: "2025-09-11T20:08:00Z"
          description: eG Innovations provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with eG Universal Agent Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eginnovations/universal-agent-operator
          support: eG Innovations
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgUniversalAgent is the Schema for the eguniversalagents
              API
            displayName: Eg Universal Agent
            kind: EgUniversalAgent
            name: eguniversalagents.eginnovations.com
            version: v1beta1
        description: |
          eG Innovations’ converged application and infrastructure performance monitoring capabilities provide end-to-end visibility and correlated analytics for every layer, every tier of your IT landscape.

          The eG Enterprise Universal Agent Operator automatically configures a host agent on every Kubernetes worker node. The host agent auto-discovers the worker nodes and application containers running as Pods on each node and tracks their performance and utilization levels. In-depth monitoring of applications running on containers is also provided using the same host agent. No additional agents are required for the containers.

          ### Capabilities
            * Rolls out the egagent pod per node to monitor its pods and the node itself
            * Automatic BTM profiler injection using webhook based injection for deep application monitoring and transection tracing.

          ### Installation
            1. Create the namespace "egagent" for installing eG Agent on OpenShift Cluster.

               <br />``` $ oc create ns egagent ```
            2. Add "egagent" serviceaccount to scc( Security Context Constraints ) of OpenShift Cluster as privileged using below command.

               <br />``` $ oc adm policy add-scc-to-user privileged -z egagent -n egagent ```
            2. Install eG Innovations Universal Agent operator from [OpenShift Container Platform](https://www.openshift.com/) OperatorHub.
            3. Create the eG Universal Agent CRD (Custom Resource Definition)
               a. Click "View YAML Example" and copy the yaml contents and save it to (filename).yaml on your host.
               b. To apply the CRD run the following command

               <br />``` $ oc apply -f (filename).yaml ```

          ### Required Parameters
            * EG\_MANAGER, EG\_MANAGER\_PORT - provide the ip and port of your eG Manager Installation.

          ### Advanced Options
            * Disable Certificate Checking - disable any certificate validation that may interact poorly with proxies within your cluster
            * Image Override - use a copy of the eG Iniversal Agent container image from a registry other that Docker's or Red Hat's
            * Environment variables - define environment variables for the eG Universal Agent container of for APM profiler.
            * Namespace Selectors- select a subset of namesapces to enable instrumentation for applications.

          For complete list of supported parameters and the detailed documentation available in [eG Innovations Docs](https://docs.eginnovations.com)
        displayName: eG Innovations Universal Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Tracing
        links:
        - name: eG Innovations Kubernetes Monitoring Overview
          url: https://www.eginnovations.com/documentation/Kubernetes/Introduction-to-Monitoring-Kubernetes.htm
        maintainers:
        - email: devops@eginnovations.com
          name: eG Innovations
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: eG Innovations
          url: https://www.eginnovations.com
        relatedImages:
        - docker.io/eginnovations/universal-agent-operator@sha256:b8e3e26dca1b94fabfd79496feb36f3f2de4434331ef9ca810125892a936dca7
        version: 0.0.11
      entries:
      - name: eginnovations-operator.v0.0.11
        version: 0.0.11
      name: beta
    defaultChannel: beta
    packageName: eginnovations-operator
    provider:
      name: eG Innovations
      url: https://www.eginnovations.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-trino-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trino-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:904eb8ac055171c75c8c86e4923df85461b81e6d5d9269d9cddfb03fc615f9f4
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:904eb8ac055171c75c8c86e4923df85461b81e6d5d9269d9cddfb03fc615f9f4
        version: 24.11.1
      entries:
      - name: trino-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: trino-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b50b247ed8f0ad262ed59e8aafd099919e98156472e5663957f9922616140bd9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b50b247ed8f0ad262ed59e8aafd099919e98156472e5663957f9922616140bd9
        version: 25.3.0
      entries:
      - name: trino-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: trino-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
        version: 25.7.0
      entries:
      - name: trino-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: trino-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
        version: 25.7.0
      entries:
      - name: trino-operator.v25.7.0
        version: 25.7.0
      - name: trino-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-trino-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-kafka-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kafka-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:f76b5558088bead829be8018a89e3efc1bb7a84d40429a0a1df1c88552db169b
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:f76b5558088bead829be8018a89e3efc1bb7a84d40429a0a1df1c88552db169b
        version: 24.11.1
      entries:
      - name: kafka-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: kafka-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:72b390c2d382af0a9cbe2d9ae394fc61c62b9a3f6c7310cf5f325d0f17257f1f
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:72b390c2d382af0a9cbe2d9ae394fc61c62b9a3f6c7310cf5f325d0f17257f1f
        version: 25.3.0
      entries:
      - name: kafka-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: kafka-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
        version: 25.7.0
      entries:
      - name: kafka-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: kafka-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
        version: 25.7.0
      entries:
      - name: kafka-operator.v25.7.0
        version: 25.7.0
      - name: kafka-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-kafka-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: vfunction-operator
      app.kubernetes.io/managed-by: operator
      app.kubernetes.io/name: vfunction
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      name: vfunction-server-operator
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: vFunction
      provider-url: ""
    name: vfunction-server-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vfunction-server-operator.v2.2.469
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vfunction.com/v1",
                "kind": "VfunctionServer",
                "metadata": {
                  "name": "example-vfunction-server",
                  "namespace": "vfunction",
                  "labels": {
                      "name": "vfunction",
                      "app.kubernetes.io/name": "vfunction",
                      "app.kubernetes.io/instance": "vfunction-server",
                      "app.kubernetes.io/managed-by": "operator"
                  }
                },
                "spec": {
                  "admin": {
                    "email": "admin@mycompany.com",
                    "name": "Admin",
                    "password": "Password1!"
                  },
                  "host": "http://my.domain.com",
                  "org_name": "MyCompany",
                  "upgrade": "Daily",
                  "measurements": {
                    "auto_scaling": "No",
                    "min_number_of_services": "2",
                    "max_number_of_services": "10"
                  },
                  "smtp": {
                    "identity": "",
                    "password": "",
                    "url": "",
                    "user": ""
                  },
                  "tls": {
                    "crt": "",
                    "key": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, AI/Machine Learning
          certified: Red Hat certified
          containerImage: registry.marketplace.redhat.com/rhm/vfunction/vfunction-server-operator@sha256:60d1965460101d0fd90ecc9499d0a6d4c9684a54ada61a43a253d30661fab22f
          createdAt: "2021-08-03T19:08:54Z"
          description: vFunction is a cutting-edge code analysis, machine learning,
            and automation to boost your Java modernization projects.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/vfunction-server-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/vfunction-server-operator-rhmp/support?utm_source=openshift_console
          support: vFunction
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of vFunction running on your cluster.
            displayName: vFunction
            kind: VfunctionServer
            name: vfunctionservers.vfunction.com
            version: v1
        description: |2-

          ## Introduction

          [vFunction](https://www.vfunction.com/) is a cloud-native modernization platform that combines dynamic and static code analysis, machine learning, and automation to automatically identify and extract services from existing applications. vFunction is the only platform purpose-built for modernization of Java applications.

          The vFunction modernization process starts by learning the running monolithic application, and surfacing the interdependencies within it. Using AI, the platform analyzes and identifies services that can be separated from the application. This decomposition can present a range of micro, mini, or even macro services, depending on your application environment, each being an independently deployable and scalable application component.

          vFunction automates the extraction of these services, enabling you to modernize your monolith, quickly and easily.

          ## The vFunction platform

          The platform consists of 3 basic components; the server, the controller package, and a tools package. The server runs as an operand on an OpenShift environment. The controller package is installed on the machine that runs the monolithic application which can be either a Linux or a Windows machine, and the tools are run on a development machine, with access to the code of the monolithic application.

          The controller package consists of three elements: the vFunction agent, that collects data during the dynamic analysis phase; the vFunction Viper application, that performs static analysis on the binaries of the application; and the vFunction controller that handles all the communication between the agent, Viper, and the vFunction server.

          The vFunction agent is a mix of a Java and native agent, and needs to run on the JVM that is currently running your application. [Refer to the vFunction OOB Support Matrix document](https://drive.google.com/file/d/1ccq8LFab1FrYAimDUxwgjiCOdk4QuVzs/view) for a list of supported application servers and JVMs.

          This operator installs a vFunction server instance to be connected later with one or more vFunction controllers installed on your application machine(s).

          ## Before You Start

          The vFunction operator requires a Red Hat OpenShift Kubernetes Application Platform 4.x.

          The cluster recommended configuration (which is the same as the minimal one) should consist of:

          - *1 master node*

          - *1 worker node*

          - *Storage capacity enough for dynamically provision 2 PVs of 50G each*

          - *A default StorageClass configured*

          vFunction doesn’t provide any encryption mechanism for data stored on any attached storage. If required, encryption can be achieved by the client by using its own external means on the storage itself.

          vFunction supports backup of all critical data as part of the application. Nevertheless, for complete backup of entire data it is recommended that the client apply its own storage based backup mechanism.

          ## Install Prerequisites

          ###Cluster:
          The operator was certified and tested on OCP 4.6 and 4.7.
          If you encounter any issue with other OCP versions or cloud providers please contact info@vfunction.com.

          ###Storage:
          The operator creates two new PersistentVolumeClaims (PVCs) during installation time. For their creation, it relies on the default StorageClass to dynamically provision two PersistentVolumes (PVs). The system must have a default StorageClass in place.

          Both PersistentVolumes are accessed with ReadWriteOnce mode.

          ###Project:
          The operator should be installed in a new and dedicated project (namespace). If you intend to install more than one vFunction operands in the same cluster, please use different projects for each one. Installation of more than one operand on the same cluster will allow scaling by load balancing multiple applications to different vFunction servers.

          ## Installation instructions

          1. Prepare the YAML file to use in the installation:

              a. Copy the following YAML template into a text editor:

              ```yaml
                  apiVersion: vfunction.com/v1
                  kind: VfunctionServer
                  metadata:
                    name: vfunction
                    namespace: vfunction
                    labels:
                      name: vfunction
                      app.kubernetes.io/name: vfunction
                      app.kubernetes.io/instance: vfunction-server
                      app.kubernetes.io/managed-by: operator
                  spec:
                    host: "http://my.domain.com"
                    org_name: "MyCompany"
                    upgrade: "Daily"
                    admin:
                      email: "admin@mycompany.com"
                      name: "Admin"
                      password: "Password1!"
                    smtp:
                      password: ""
                      url: ""
                      identity: ""
                      user: ""
                    measurements:
                      auto_scaling: "No"
                      min_number_of_services: "2"
                      max_number_of_services: "10"
                    tls:
                      crt: |
                        -----BEGIN CERTIFICATE-----
                        ...
                        -----END CERTIFICATE-----
                      key: |
                        -----BEGIN PRIVATE KEY-----
                        ...
                        -----END PRIVATE KEY-----
              ```
              b. Customize the template fields by replacing with your information:

              - **host**: Enter the FQDN for accessing the vFunction dashboard.
                          Make sure that you use a domain name and not an IP address. Writing "https://" will mandate TLS while writing "http://" will expose the server through HTTP.
              - **org_name**: Enter your organization name.
              - **upgrade**: Choose auto-upgrade mode
                  - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically (default).
                  - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
                  - **Never**: The automatic upgrading mechanism is off.
                  - **Maintenance**: Put the operator in a maintenance mode.

              - **admin.email**: Enter the email address of your administrator.
              - **admin.name**: Enter the name of your administrator.
              - **admin.password**: Enter the password you want to use for the vFunction administrator.
                                    The password should be at least 8 characters long, and consist of at least one lowercase letter, at least one uppercase letter, at least one number, and at least one special character.
              - **smtp.user**: Enter the email address for a designated user for the SMTP server (optional).
              - **smtp.password**: Enter this user’s password (optional).
              - **smtp.identity**: Enter this user’s password (optional).
              - **smtp.url**: Enter the SMTP server URL (optional).
              - **tls.crt**: If your host FQDN starts with “https”, paste in the certifications you have for using the TLS connection.
              - **tls.key**: If your host FQDN starts with “https”, enter the key you have for using the TLS connection.
              - **measurements.auto_scaling**: Indicates if the measurement services auto scalling is active.
              - **measurements.min_num_of_services**: Minimum Number of measurement service pods.
              - **measurements.max_num_of_services**: Maximum Number of measurement service pods.
              - **measurements.S3**: Save all measurements data to S3, instead of local PV (optional).
                  - **measurements.S3.bucket**: Measurements S3 bucket name.
                  - **measurements.S3.key**: Measurements S3 key.
                  - **measurements.S3.secret**: Measurements S3 secret.
                  - **measurements.S3.region**: Measurements S3 region.
              - **backup.mysql.S3**: Backup MySQL to S3 (optional).
                  - **backup.mysql.S3.bucket**: MySQL backup S3 bucket name.
                  - **backup.mysql.S3.key**: MySQL backup S3 key.
                  - **backup.mysql.S3.secret**: MySQL backup S3 secret.
                  - **backup.mysql.S3.region**: MySQL backup S3 region.

          1. In the Installed Operators window choose and click on vFunction Operator and then Create Instance.
          1. In the Create vFunctionServer window, paste the YAML file that you created in step 1.
          1. Click Create.
          1. You can now click on your new created vFunction operand in order to see its details and installation progress.

          ## Verifying the installation

          Check that the operand installed successfully:

          1. Check that the "Operand State" property shows “Working”.
          1. Check that the "Successfully Installed" property shows "Yes".

          If "Operand State" shows "Failed", it indicates the operand wasn’t installed correctly. Check the events and logs for all pods (operator and image containers) for any issues. In the event that you cannot troubleshoot, contact vFunction support.

          ## After installation

          The vFunction site is now accessible via the newly created vFunction application custom address, for example, my.domain.com. There are two ways you can access the vFunction dashboard.

          ### Access using the router canonical hostname

          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address, and to the ”vfunction” subdomain of the  OpenShift canonical router hostname as the alias.
          For example, ***my.domain.com.   CNAME   vfunction.apps.ocp4.my-openshift.com.***
          2. Find your cluster Router Canonical Hostname address in the newly created vfunction-route-xxx under your vFunction OpenShift project -> **Networking** -> **Routes** -> **vfunction-route-xxx** route -> **Router Canonical Hostname** field.
          You can now access the dashboard using your defined "host" spec property (as above).

          ### Access using the nginx service location

          You may use this access method if your OpenShift is installed on a provider that supports exposing LoadBalancer-type services.
          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address and to the vfunction-nginx-xxx service location.
          For example, ***my.domain.com.   CNAME   a05951ed7cdf-1394239323.us-east-1.elb.amazonaws.com.***
          1. Point your custom domain to the vfunction-nginx-xxx service's external IP location, which you can find in the OpenShift project > **Networking** -> **Services** -> **vfunction-nginx-xxx** service > **Service Address** > **Location** field.

          ## Upgrade and Rollback

          The vFunction operator includes a built-in auto-upgrade mechanism.

          You can choose one of 4 upgrading modes:
          - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically (default).
          - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
          - **Never**: The automatic upgrading mechanism is off.
          - **Maintenance**: Put the operator in a maintenance mode.

          For an on-demand upgrade,  you can change anytime the upgrading mode from “Never” to “Always”, wait for the operand to be upgraded, and change again to “Never”.

          During the upgrade, the "Operand State" property will change to “Upgrading...” and back again to “Working” after a successful upgrade.

          In case of a failure upgrading the operand, an automatic rollback to the last working version will occur. The failed version will be marked as defective and the operator will not try to upgrade to it again.

          Choosing the "Maintenance" option will tell the operator to take down all pods, so you can safely fix storage issues, return volumes from snapshots, etc.
          After finishing the maintenance, return the upgrade mode to your original desired policy.
        displayName: vFunction Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - vfunction
        - analyzing
        - services
        links:
        - name: vFunction
          url: https://vfunction.com
        - name: Terms
          url: https://www.vfunction.com/eula/
        maintainers:
        - email: raviv@vfunction.com
          name: Raviv Halivni
        maturity: alpha
        provider:
          name: vFunction
        relatedImages:
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-organizations@sha256:bcc2c3d6222af95e614318daea3108fffc6bc77aa17b23031cce296a3811b79f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-users@sha256:e47cc051953976a0e075e8cfb457cd4749e9b9e4394f2fbecab2458c2ade92ca
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-mysql@sha256:a6e06ea08fffa33b709d4d1240587bb3ab5b4cc6a28f493a227555e88e645798
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-nginx@sha256:435d1489aaf29a811081ddb51f40ae692e5833f42c31dbced6bc189c0ddc40ab
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-storage@sha256:b778cb9c724a59e51b4cb9088047be471cdf338f1138c31b81f79e50a491203f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-idp@sha256:21ca95a0209f3c557a45658452d154777f928539a93129148470562c79e68d70
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-measurements@sha256:35ffbee4b0eead58706b23cf5a702b1d55d2a7913cd0019cefbff94957d4e5d6
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-server-operator@sha256:60d1965460101d0fd90ecc9499d0a6d4c9684a54ada61a43a253d30661fab22f
        version: 2.2.469
      entries:
      - name: vfunction-server-operator.v2.2.469
        version: 2.2.469
      name: stable
    defaultChannel: stable
    packageName: vfunction-server-operator-rhmp
    provider:
      name: vFunction
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CNCF
      provider-url: ""
    name: jaeger
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jaeger-operator.v1.65.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "jaegertracing.io/v1",
                "kind": "Jaeger",
                "metadata": {
                  "name": "my-jaeger"
                },
                "spec": {
                  "strategy": "allInOne"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing
          certified: "false"
          containerImage: quay.io/jaegertracing/jaeger-operator:1.65.0
          createdAt: "2025-01-22T20:40:19Z"
          description: Provides tracing, monitoring and troubleshooting for microservices-based
            distributed systems
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/jaegertracing/jaeger-operator
          support: Jaeger Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Jaeger is the Schema for the jaegers API
            displayName: Jaeger
            kind: Jaeger
            name: jaegers.jaegertracing.io
            version: v1
        description: |-
          Jaeger, inspired by [Dapper](https://research.google.com/pubs/pub36356.html) and [OpenZipkin](http://zipkin.io/), is a distributed tracing system released as open source by Uber Technologies. It is used for monitoring and troubleshooting microservices-based distributed systems.
          ### Core capabilities
          Jaeger is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed context propagation
          * Distributed transaction monitoring
          * Root cause analysis
          * Service dependency analysis
          * Performance / latency optimization
          * OpenTracing compatible data model
          * Multiple storage backends: Badger, Cassandra, Elasticsearch, Memory.
          ### Operator features
          * **Multiple modes** - Supports `allInOne`, `production` and `streaming` [modes of deployment](https://www.jaegertracing.io/docs/latest/operator/#deployment-strategies).
          * **Configuration** - The Operator manages [configuration information](https://www.jaegertracing.io/docs/latest/operator/#configuring-the-custom-resource) when installing Jaeger instances.
          * **Storage** - [Configure storage](https://www.jaegertracing.io/docs/latest/operator/#storage-options) used by Jaeger. By default, `memory` is used. Other options include `badger`, `cassandra` or `elasticsearch`. On OpenShift, the operator can delegate creation of an Elasticsearch cluster to the Elasticsearch Operator if deployed.
          * **Agent** - can be deployed as [sidecar](https://www.jaegertracing.io/docs/latest/operator/#auto-injecting-jaeger-agent-sidecars) (default) and/or [daemonset](https://www.jaegertracing.io/docs/latest/operator/#installing-the-agent-as-daemonset).
          * **UI** - Optionally setup ingress (Kubernetes) or secure route (OpenShift) to provide [access to the Jaeger UI](https://www.jaegertracing.io/docs/latest/operator/#accessing-the-jaeger-console-ui).
          ### Before you start
          1. Ensure that the appropriate storage solution, that will be used by the Jaeger instance, is available and configured.
          2. If intending to deploy an Elasticsearch cluster via the Jaeger custom resource, then the Elasticsearch Operator must first be installed.
          ### Troubleshooting
          * https://www.jaegertracing.io/docs/latest/troubleshooting/
        displayName: Community Jaeger Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tracing
        - monitoring
        - troubleshooting
        links:
        - name: Jaeger Operator Source Code
          url: https://github.com/jaegertracing/jaeger-operator
        maintainers:
        - email: jaeger-tracing@googlegroups.com
          name: Jaeger Google Group
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: CNCF
        relatedImages:
        - quay.io/jaegertracing/jaeger-operator:1.65.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 1.65.0
      entries:
      - name: jaeger-operator.v1.65.0
        version: 1.65.0
      - name: jaeger-operator.v1.62.0
        version: 1.62.0
      - name: jaeger-operator.v1.61.0
        version: 1.61.0
      - name: jaeger-operator.v1.60.0
        version: 1.60.0
      - name: jaeger-operator.v1.59.0
        version: 1.59.0
      - name: jaeger-operator.v1.57.0
        version: 1.57.0
      - name: jaeger-operator.v1.56.0
        version: 1.56.0
      - name: jaeger-operator.v1.55.0
        version: 1.55.0
      - name: jaeger-operator.v1.53.0
        version: 1.53.0
      - name: jaeger-operator.v1.52.0
        version: 1.52.0
      - name: jaeger-operator.v1.51.0
        version: 1.51.0
      - name: jaeger-operator.v1.49.0
        version: 1.49.0
      - name: jaeger-operator.v1.48.1
        version: 1.48.1
      - name: jaeger-operator.v1.47.0
        version: 1.47.0
      - name: jaeger-operator.v1.46.0
        version: 1.46.0
      - name: jaeger-operator.v1.45.0
        version: 1.45.0
      - name: jaeger-operator.v1.44.0
        version: 1.44.0
      - name: jaeger-operator.v1.43.0
        version: 1.43.0
      - name: jaeger-operator.v1.42.0
        version: 1.42.0
      - name: jaeger-operator.v1.41.1
        version: 1.41.1
      - name: jaeger-operator.v1.41.0
        version: 1.41.0
      - name: jaeger-operator.v1.40.0
        version: 1.40.0
      - name: jaeger-operator.v1.39.0
        version: 1.39.0
      - name: jaeger-operator.v1.38.1
        version: 1.38.1
      - name: jaeger-operator.v1.38.0
        version: 1.38.0
      - name: jaeger-operator.v1.37.0
        version: 1.37.0
      - name: jaeger-operator.v1.36.0
        version: 1.36.0
      - name: jaeger-operator.v1.35.0
        version: 1.35.0
      - name: jaeger-operator.v1.34.1
        version: 1.34.1
      - name: jaeger-operator.v1.34.0
        version: 1.34.0
      - name: jaeger-operator.v1.33.0
        version: 1.33.0
      - name: jaeger-operator.v1.32.0
        version: 1.32.0
      - name: jaeger-operator.v1.31.0
        version: 1.31.0
      - name: jaeger-operator.v1.30.0
        version: 1.30.0
      - name: jaeger-operator.v1.29.1
        version: 1.29.1
      - name: jaeger-operator.v1.25.0
        version: 1.25.0
      - name: jaeger-operator.v1.24.0
        version: 1.24.0
      name: stable
    defaultChannel: stable
    packageName: jaeger
    provider:
      name: CNCF
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: JWS team
      provider-url: ""
    name: jws-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jws-operator.v2.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "web.servers.org/v1alpha1",
                "kind": "WebServer",
                "metadata": {
                  "name": "webserver-example"
                },
                "spec": {
                  "applicationName": "webapp",
                  "replicas": 1,
                  "useInsightsClient": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/jboss-webserver-5/jws5-rhel8-operator@sha256:ea042277bdcdb70b2dc32083ea868efd4bf474d5a8ec0f9b47f5bea35aaeae1e
          createdAt: "2025-09-12T08:31:52.558719Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat JBoss Web Server",
            "Red Hat Runtimes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/web-servers/jws-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Web Server is the schema for the webservers API
            displayName: Web Server
            kind: WebServer
            name: webservers.web.servers.org
            version: v1alpha1
        description: WebServer operator can manage JBoss Web Server images.
        displayName: WebServer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WebServer
        - JBoss
        links:
        - name: Jws Operator
          url: https://jws-operator.domain
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: JWS team
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:021ff671b0fb72eed635b05489ec86342fcb7ada3ba0353591fdb9328871d067
        - registry.redhat.io/jboss-webserver-5/jws5-rhel8-operator@sha256:ea042277bdcdb70b2dc32083ea868efd4bf474d5a8ec0f9b47f5bea35aaeae1e
        - registry.redhat.io/insights-runtimes-tech-preview/runtimes-inventory-rhel8-operator@sha256:58e01d11c3d30abc5b0ee3795511bd04e39aea21e2853b4f4cc6d1c8a2bd7bb0
        version: 2.2.0
      entries:
      - name: jws-operator.v2.2.0
        version: 2.2.0
      - name: jws-operator.v2.1.3
        version: 2.1.3
      - name: jws-operator.v2.1.2
        version: 2.1.2
      - name: jws-operator.v2.1.1
        version: 2.1.1
      - name: jws-operator.v2.1.0
        version: 2.1.0
      - name: jws-operator.v2.0.16
        version: 2.0.16
      - name: jws-operator.v2.0.15
        version: 2.0.15
      - name: jws-operator.v2.0.14
        version: 2.0.14
      - name: jws-operator.v2.0.13
        version: 2.0.13
      - name: jws-operator.v2.0.12
        version: 2.0.12
      - name: jws-operator.v2.0.11
        version: 2.0.11
      - name: jws-operator.v2.0.10
        version: 2.0.10
      - name: jws-operator.v2.0.9
        version: 2.0.9
      - name: jws-operator.v2.0.8
        version: 2.0.8
      - name: jws-operator.v2.0.7
        version: 2.0.7
      - name: jws-operator.v2.0.6
        version: 2.0.6
      - name: jws-operator.v2.0.5
        version: 2.0.5
      - name: jws-operator.v2.0.4
        version: 2.0.4
      - name: jws-operator.v2.0.3
        version: 2.0.3
      - name: jws-operator.v2.0.2
        version: 2.0.2
      - name: jws-operator.v2.0.1
        version: 2.0.1
      name: alpha
    defaultChannel: alpha
    packageName: jws-operator
    provider:
      name: JWS team
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Windup
      provider-url: ""
    name: windup-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: windup-operator.v0.0.25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
             {
               "apiVersion": "windup.jboss.org/v1",
               "kind": "Windup",
               "metadata": {
                 "name": "windup",
                 "labels": {
                   "application": "windup"
                 }
               },
               "spec": {}
             }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/windupeng/windup-operator-native:0.0.25
          createdAt: 01/22/24
          description: Windup is an analysis tool that supports the modernization
            and migration of Java applications.
          imagePullPolicy: Always
          operatorframework.io/suggested-namespace: windup
          repository: https://github.com/windup/windup-operator/
          support: https://issues.redhat.com/projects/WINDUP
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "Create a Windup Web Console application.  \n\nTo access
              the Windup Web Console application from the \"Developer\" perspective,
              please go to Topology and click on the \"Open URL\" icon on the web-console
              pod.  \n\nAlternatively from the \"Administrator\" perspective, please
              go to Networking->Routes and click on the Location hyperlink.  \n\nKnown
              issue  \nIf you want to customize the Windup Web Console instance's
              parameters and you can not see any in the `Form view`, please switch
              to the `YAML view` and change them as needed.\n"
            displayName: Windup
            kind: Windup
            name: windups.windup.jboss.org
            version: v1
        description: |
          Windup is a web console application that supports large-scale Java application modernization and migration projects across a broad range of transformations and use cases.

          It analyzes application code, supports effort estimation, accelerates code migration, and enables users to move applications to containers.

          For more information please refer to the [https://windup.github.io/](https://windup.github.io/) page.
        displayName: Windup Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - modernization
        - windup
        links:
        - name: Website
          url: https://windup.github.io
        - name: GitHub
          url: https://github.com/windup/windup-operator
        maintainers:
        - email: migrate@redhat.com
          name: Windup
        maturity: alpha
        provider:
          name: Windup
        relatedImages:
        - quay.io/windupeng/windup-operator-native:0.0.25
        version: 0.0.25
      entries:
      - name: windup-operator.v0.0.25
        version: 0.0.25
      - name: windup-operator.v0.0.24
        version: 0.0.24
      - name: windup-operator.v0.0.23
        version: 0.0.23
      - name: windup-operator.v0.0.22
        version: 0.0.22
      - name: windup-operator.v0.0.21
        version: 0.0.21
      - name: windup-operator.v0.0.20
        version: 0.0.20
      - name: windup-operator.v0.0.19
        version: 0.0.19
      - name: windup-operator.v0.0.18
        version: 0.0.18
      - name: windup-operator.v0.0.17
        version: 0.0.17
      - name: windup-operator.v0.0.16
        version: 0.0.16
      - name: windup-operator.v0.0.15
        version: 0.0.15
      - name: windup-operator.v0.0.14
        version: 0.0.14
      - name: windup-operator.v0.0.13
        version: 0.0.13
      - name: windup-operator.v0.0.12
        version: 0.0.12
      - name: windup-operator.v0.0.11
        version: 0.0.11
      name: alpha
    defaultChannel: alpha
    packageName: windup-operator
    provider:
      name: Windup
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stolostron
      provider-url: ""
    name: cluster-aas-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-aas-operator.v0.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplate",
                "metadata": {
                  "name": "clustertemplate-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplateInstance",
                "metadata": {
                  "name": "clustertemplateinstance-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplateQuota",
                "metadata": {
                  "name": "clustertemplatequota-sample"
                },
                "spec": {}
              }
            ]
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/cluster-templates-operator:2.11.0-0c3131154065397d37e3b66ae635d7ea88650bf9
          createdAt: "2024-06-06T19:34:44Z"
          description: Easily install fully configured clusters with guard-rails.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/cluster-templates-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents instance of a cluster
            displayName: Cluster template instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.clustertemplate.openshift.io
            version: v1alpha1
          - description: Defines which ClusterTemplates can be used in a given namespace
            displayName: Cluster template quota
            kind: ClusterTemplateQuota
            name: clustertemplatequotas.clustertemplate.openshift.io
            version: v1alpha1
          - description: Template of a cluster - both installation and post-install
              setup are defined as ArgoCD application spec. Any application source
              is supported - typically a Helm chart
            displayName: Cluster template
            kind: ClusterTemplate
            name: clustertemplates.clustertemplate.openshift.io
            version: v1alpha1
          - kind: ClusterTemplateSetup
            name: clustertemplatesetup.clustertemplate.openshift.io
            version: v1alpha1
          - kind: Config
            name: config.clustertemplate.openshift.io
            version: v1alpha1
        description: |
          **Self-service clusters with guardrails!**
          Cluster as a service operator provides an easy way to define clusters as templates and allows creating instances of those templates even for non-privileged developer/devops engineers. Cluster templates operator also allows specifing quotas for the developer/devops engineers.

          ### User Interface
          The User interface is installed but is not enabled by default. To enable go to `Home` -> `Overview` -> in the status card click `Dynamic Plugins` -> click `View all` -> click the `pencil` icon -> pick `Enable` and hit `Save`. You can access the UI in the `All Clusters` perspective under `Infrastructure` -> `Cluster Templates`.

          ### Documentation
          Documentation can be found on our [website](https://github.com/stolostron/cluster-templates-operator).

          ### Getting help
          To report any issues or ask questions, create an issue on our [Github repo](https://github.com/stolostron/cluster-templates-operator/issues)
        displayName: Cluster as a service operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cluster
        - templates
        - hive
        - hypershift
        - selfservice
        - openshift
        - multicluster
        - quota
        - helm
        - argocd
        links:
        - name: Cluster as a service operator
          url: https://github.com/stolostron/cluster-templates-operator
        maintainers:
        - email: rawagner@redhat.com
          name: Rastislav Wagner
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Stolostron
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/stolostron/cluster-templates-operator:2.11.0-0c3131154065397d37e3b66ae635d7ea88650bf9
        version: 0.1.5
      entries:
      - name: cluster-aas-operator.v0.1.5
        version: 0.1.5
      - name: cluster-aas-operator.v0.1.4
        version: 0.1.4
      - name: cluster-aas-operator.v0.1.3
        version: 0.1.3
      - name: cluster-aas-operator.v0.1.2
        version: 0.1.2
      - name: cluster-aas-operator.v0.1.1
        version: 0.1.1
      - name: cluster-aas-operator.v0.1.0
        version: 0.1.0
      - name: cluster-aas-operator.v0.0.6
        version: 0.0.6
      - name: cluster-aas-operator.v0.0.5
        version: 0.0.5
      - name: cluster-aas-operator.v0.0.4
        version: 0.0.4
      - name: cluster-aas-operator.v0.0.3
        version: 0.0.3
      - name: cluster-aas-operator.v0.0.2
        version: 0.0.2
      - name: cluster-aas-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cluster-aas-operator
    provider:
      name: Stolostron
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: ibm-spectrum-scale-csi-operator
      app.kubernetes.io/managed-by: ibm-spectrum-scale-csi-operator
      app.kubernetes.io/name: ibm-spectrum-scale-csi-operator
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-spectrum-scale-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-spectrum-scale-csi-operator.v2.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "CSIScaleOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-spectrum-scale-csi-operator",
                    "app.kubernetes.io/managed-by": "ibm-spectrum-scale-csi-operator",
                    "app.kubernetes.io/name": "ibm-spectrum-scale-csi-operator",
                    "release": "ibm-spectrum-scale-csi-operator"
                  },
                  "name": "ibm-spectrum-scale-csi",
                  "namespace": "ibm-spectrum-scale-csi-driver"
                },
                "spec": {
                  "attacherNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "clusters": [
                    {
                      "id": "\u003cPrimary Cluster ID\u003e",
                      "primary": {
                        "primaryFs": "\u003c Primary Filesystem \u003e"
                      },
                      "restApi": [
                        {
                          "guiHost": "\u003c Primary cluster GUI IP/Hostname \u003e"
                        }
                      ],
                      "secrets": "secret1",
                      "secureSslMode": false
                    }
                  ],
                  "pluginNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "provisionerNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "resizerNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "snapshotterNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ]
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/ibm-spectrum-scale/ibm-spectrum-scale-csi-operator@sha256:b62ad588fbc338090f405740fd8bceaf59b3267cdb532c60ae965643ccae4d5d
          createdAt: Thu May 5 04:42:15 IST 2022
          description: An operator for deploying and managing the IBM Spectrum Scale
            CSI Driver.
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-spectrum-scale-csi/
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIScaleOperator is the Schema for the csiscaleoperators
              API
            displayName: IBM CSI Spectrum Scale Driver
            kind: CSIScaleOperator
            name: csiscaleoperators.csi.ibm.com
            version: v1
        description: |
          The IBM Spectrum Scale CSI Operator for Kubernetes installs, manages,
          upgrades the IBM Spectrum Scale CSI Driver on OpenShift and Kubernetes
          for provisioning of persistent volumes on IBM Spectrum Scale.

          ### Documentation Documentation for IBM Spectrum Scale CSI Driver and Operator can be found at [IBM Documentation](https://www.ibm.com/docs/en/spectrum-scale-csi).

          ### Getting help
          IBM Spectrum Scale CSI driver is part of the IBM Spectrum Scale offering.
          Please follow the [IBM support procedure](https://www.ibm.com/mysupport/s/?language=en_US) for any issues with the driver.

          For non-urgent issues, suggestions, recommendations feel free to open an
          issue in [github](https://github.com/IBM/ibm-spectrum-scale-csi). Issues will be addressed as team availability permits.

          **Note:** Custom Resource example represents the minimum settings needed to run
          the operator. If your environment needs more advanced settings (e.g. remote clusters,
          node mapping, image pull secret, Taint and Toleration etc.) please refer to  the sample [Custom Resource](https://raw.githubusercontent.com/IBM/ibm-spectrum-scale-csi/v2.6.0/operator/config/samples/csiscaleoperators.csi.ibm.com_cr.yaml).
        displayName: IBM Spectrum Scale CSI Plugin Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - GPFS
        - Spectrum Scale
        - CSI Driver
        links:
        - name: IBM CSI Spectrum Scale Documentation
          url: https://www.ibm.com/docs/en/spectrum-scale-csi
        - name: CSI Developer Documentation
          url: https://kubernetes-csi.github.io/docs/
        maintainers:
        - email: madhu.punjabi@in.ibm.com
          name: Madhu Thorat
        - email: amdabhad@in.ibm.com
          name: Amol Dabhade
        - email: Nitish.Kumar4@ibm.com
          name: Nitish Kumar
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibm-spectrum-scale/ibm-spectrum-scale-csi-operator@sha256:b62ad588fbc338090f405740fd8bceaf59b3267cdb532c60ae965643ccae4d5d
        version: 2.6.0
      entries:
      - name: ibm-spectrum-scale-csi-operator.v2.6.0
        version: 2.6.0
      - name: ibm-spectrum-scale-csi-operator.v2.5.1
        version: 2.5.1
      - name: ibm-spectrum-scale-csi-operator.v2.5.0
        version: 2.5.0
      - name: ibm-spectrum-scale-csi-operator.v2.4.0
        version: 2.4.0
      - name: ibm-spectrum-scale-csi-operator.v2.3.1
        version: 2.3.1
      - name: ibm-spectrum-scale-csi-operator.v2.3.0
        version: 2.3.0
      name: stable
    defaultChannel: stable
    packageName: ibm-spectrum-scale-csi-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Managed Openshift Black Belt
      provider-url: ""
    name: ecr-secret-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ecr-secret-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "ecr-secret-sample",
                  "namespace": "test-ecr-secret-operator"
                },
                "spec": {
                  "ecr_registry": "ACCOUNT_ID.dkr.ecr.us-east-2.amazonaws.com",
                  "frequency": "10h",
                  "generated_secret_name": "ecr-docker-secret",
                  "region": "us-east-2"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ecr-secret-operator:v0.3.2
          operatorframework.io/suggested-namespace: ecr-secret-operator
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ecr-secret-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret is the Schema for the secrets API
            displayName: Secret
            kind: Secret
            name: secrets.ecr.mobb.redhat.com
            version: v1alpha1
        description: "# ECR Secret Operator\n\nAmazon Elastic Container Registry [Private
          Registry Authentication](https://docs.aws.amazon.com/AmazonECR/latest/userguide/registry_auth.html)
          provides a temporary token that is valid only for 12 hours. It is a challenge
          for automatic container image build process to refresh the token or secret
          in a timely manner.\n\nThis operators frequently talks with AWS ECR GetAuthroization
          Token and create/update the secret, so that the service account can perform
          docker image build.\n\n\n## How to use this operator\n\n### Prerequisites\n\n*
          [Create an ECR private repository](https://docs.aws.amazon.com/AmazonECR/latest/userguide/repository-create.html)\n*
          Create An Openshift Cluster\n* Provide AWS Authentication to the operator.
          Two Options:\n  **Note: You have to provide a secret aws-ecr-cloud-credentials
          in the installed namespace**\n  * [IAM User](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_user.md)\n
          \ * [STS Assume Role](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_assume_role.md)\n\n###
          Install the operator\n\n### Create the ECR Secret CRD\n\n```\napiVersion:
          ecr.mobb.redhat.com/v1alpha1\nkind: Secret\nmetadata:\n  name: ecr-secret\n
          \ namespace: test-ecr-secret-operator\nspec:\n  generated_secret_name: ecr-docker-secret\n
          \ ecr_registry: [ACCOUNT_ID].dkr.ecr.us-east-2.amazonaws.com\n  frequency:
          10h\n  region: us-east-2\n```\n\n```\noc create -f CRD_FILE\n```\n\nA docker
          registry secret is created by the operator momentally and the token is patched
          every 10 hours\n\n```\noc get secret ecr-docker-secret   \nNAME                TYPE
          \                            DATA   AGE\necr-docker-secret   kubernetes.io/dockerconfigjson
          \  1      16h\n```\n\n### A sample build process with generated secret\n\n\nLink
          the secret to builder\n\n```\noc secrets link builder ecr-docker-secret
          \n```\n\nConfigure [build config](https://github.com/rh-mobb/ecr-secret-operator/tree/main/samples/build-config.yaml)
          to point to your ECR Container repository\n\n```\noc create imagestream
          ruby\noc tag openshift/ruby:2.5-ubi8 ruby:2.5\noc create -f samples/build-config.yaml\noc
          start-build ruby-sample-build --wait\n```\n\nBuild should succeed and push
          the image to the the private ECR Container repository\n\n"
        displayName: ECR Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - redhat
        links:
        - name: Ecr Secret Operator
          url: https://github.com/rh-mobb/ecr-secret-operator
        maintainers:
        - email: shading@redhat.com
          name: Shaozhen Ding
        maturity: alpha
        provider:
          name: Managed Openshift Black Belt
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/mobb/ecr-secret-operator:v0.3.2
        version: 0.3.2
      entries:
      - name: ecr-secret-operator.v0.3.2
        version: 0.3.2
      name: v0.3
    - currentCSV: ecr-secret-operator.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "ecr-secret-sample",
                  "namespace": "test-ecr-secret-operator"
                },
                "spec": {
                  "ecr_registry": "ACCOUNT_ID.dkr.ecr.us-east-2.amazonaws.com",
                  "frequency": "10h",
                  "generated_secret_name": "ecr-docker-secret",
                  "region": "us-east-2"
                }
              },
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "ArgoHelmRepoSecret",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ecr-secret-operator",
                    "app.kubernetes.io/instance": "argohelmreposecret-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "argohelmreposecret",
                    "app.kubernetes.io/part-of": "ecr-secret-operator"
                  },
                  "name": "argohelmreposecret-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ecr-secret-operator:v0.4.1
          createdAt: "2023-04-14T14:52:13Z"
          operatorframework.io/suggested-namespace: ecr-secret-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ecr-secret-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ArgoHelmRepoSecret is the Schema for the argohelmreposecrets
              API
            displayName: Argo Helm Repo Secret
            kind: ArgoHelmRepoSecret
            name: argohelmreposecrets.ecr.mobb.redhat.com
            version: v1alpha1
          - description: Secret is the Schema for the secrets API
            displayName: Secret
            kind: Secret
            name: secrets.ecr.mobb.redhat.com
            version: v1alpha1
        description: "# ECR Secret Operator\n\nAmazon Elastic Container Registry [Private
          Registry Authentication](https://docs.aws.amazon.com/AmazonECR/latest/userguide/registry_auth.html)
          provides a temporary token that is valid only for 12 hours. It is a challenge
          for automatic container image build process to refresh the token or secret
          in a timely manner.\n\nThis operators frequently talks with AWS ECR GetAuthroization
          Token and create/update the secret, so that the service account can perform
          docker image build.\n\n\n## How to use this operator\n\n### Prerequisites\n\n*
          [Create an ECR private repository](https://docs.aws.amazon.com/AmazonECR/latest/userguide/repository-create.html)\n*
          Create An Openshift Cluster\n* Provide AWS Authentication to the operator.
          Two Options:\n  **Note: You have to provide a secret aws-ecr-cloud-credentials
          in the installed namespace**\n  * [IAM User](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_user.md)\n
          \ * [STS Assume Role](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_assume_role.md)\n\n###
          Install the operator\n\n### Create the ECR Secret CRD\n\n```\napiVersion:
          ecr.mobb.redhat.com/v1alpha1\nkind: Secret\nmetadata:\n  name: ecr-secret\n
          \ namespace: test-ecr-secret-operator\nspec:\n  generated_secret_name: ecr-docker-secret\n
          \ ecr_registry: [ACCOUNT_ID].dkr.ecr.us-east-2.amazonaws.com\n  frequency:
          10h\n  region: us-east-2\n```\n\n```\noc create -f CRD_FILE\n```\n\nA docker
          registry secret is created by the operator momentally and the token is patched
          every 10 hours\n\n```\noc get secret ecr-docker-secret   \nNAME                TYPE
          \                            DATA   AGE\necr-docker-secret   kubernetes.io/dockerconfigjson
          \  1      16h\n```\n\n### A sample build process with generated secret\n\n\nLink
          the secret to builder\n\n```\noc secrets link builder ecr-docker-secret
          \n```\n\nConfigure [build config](https://github.com/rh-mobb/ecr-secret-operator/tree/main/samples/build-config.yaml)
          to point to your ECR Container repository\n\n```\noc create imagestream
          ruby\noc tag openshift/ruby:2.5-ubi8 ruby:2.5\noc create -f samples/build-config.yaml\noc
          start-build ruby-sample-build --wait\n```\n\nBuild should succeed and push
          the image to the the private ECR Container repository\n\n"
        displayName: ECR Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - redhat
        links:
        - name: Ecr Secret Operator
          url: https://github.com/rh-mobb/ecr-secret-operator
        maintainers:
        - email: shading@redhat.com
          name: Shaozhen Ding
        maturity: alpha
        provider:
          name: Managed Openshift Black Belt
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/mobb/ecr-secret-operator:v0.4.1
        version: 0.4.1
      entries:
      - name: ecr-secret-operator.v0.4.1
        version: 0.4.1
      - name: ecr-secret-operator.v0.4.0
        version: 0.4.0
      name: v0.4
    defaultChannel: v0.4
    packageName: ecr-secret-operator
    provider:
      name: Managed Openshift Black Belt
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-apigatewayv2-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-apigatewayv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Deployment",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "API",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Authorizer",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Route",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Stage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "VPCLink",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/apigatewayv2-controller:1.1.1
          createdAt: "2025-09-19T21:58:07Z"
          description: AWS API Gateway v2 controller is a service controller for managing
            API Gateway v2 resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIMapping represents the state of an AWS apigatewayv2 APIMapping
              resource.
            displayName: APIMapping
            kind: APIMapping
            name: apimappings.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: API represents the state of an AWS apigatewayv2 API resource.
            displayName: API
            kind: API
            name: apis.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Authorizer represents the state of an AWS apigatewayv2 Authorizer
              resource.
            displayName: Authorizer
            kind: Authorizer
            name: authorizers.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Deployment represents the state of an AWS apigatewayv2 Deployment
              resource.
            displayName: Deployment
            kind: Deployment
            name: deployments.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: DomainName represents the state of an AWS apigatewayv2 DomainName
              resource.
            displayName: DomainName
            kind: DomainName
            name: domainnames.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Integration represents the state of an AWS apigatewayv2 Integration
              resource.
            displayName: Integration
            kind: Integration
            name: integrations.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Route represents the state of an AWS apigatewayv2 Route resource.
            displayName: Route
            kind: Route
            name: routes.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Stage represents the state of an AWS apigatewayv2 Stage resource.
            displayName: Stage
            kind: Stage
            name: stages.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: VPCLink represents the state of an AWS apigatewayv2 VPCLink
              resource.
            displayName: VPCLink
            kind: VPCLink
            name: vpclinks.apigatewayv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon API Gateway v2 resources in AWS from within your Kubernetes cluster. Currently only HTTP APIs are supported. WebSocket API support will be added soon. For HTTP APIs, DomainName and APIMapping resource support will be added soon.

          **About Amazon API Gateway v2**

          Amazon API Gateway is a fully managed service that makes it easy for developers to create, publish, maintain, monitor, and secure APIs at any scale. APIs act as the "front door" for applications to access data, business logic, or functionality from your backend services. Using API Gateway, you can create RESTful APIs and WebSocket APIs that enable real-time two-way communication applications. API Gateway supports containerized and serverless workloads, as well as web applications.

          API Gateway handles all the tasks involved in accepting and processing up to hundreds of thousands of concurrent API calls, including traffic management, CORS support, authorization and access control, throttling, monitoring, and API version management. API Gateway has no minimum fees or startup costs. You pay for the API calls you receive and the amount of data transferred out and, with the API Gateway tiered pricing model, you can reduce your cost as your API usage scales.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon API Gateway v2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apigatewayv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon API Gateway v2 Developer Resources
          url: https://aws.amazon.com/api-gateway/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: api gateway maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/apigatewayv2-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-apigatewayv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-apigatewayv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-apigatewayv2-controller.v1.0.30
        version: 1.0.30
      - name: ack-apigatewayv2-controller.v1.0.29
        version: 1.0.29
      - name: ack-apigatewayv2-controller.v1.0.28
        version: 1.0.28
      - name: ack-apigatewayv2-controller.v1.0.26
        version: 1.0.26
      - name: ack-apigatewayv2-controller.v1.0.25
        version: 1.0.25
      - name: ack-apigatewayv2-controller.v1.0.23
        version: 1.0.23
      - name: ack-apigatewayv2-controller.v1.0.21
        version: 1.0.21
      - name: ack-apigatewayv2-controller.v1.0.17
        version: 1.0.17
      - name: ack-apigatewayv2-controller.v1.0.16
        version: 1.0.16
      - name: ack-apigatewayv2-controller.v1.0.15
        version: 1.0.15
      - name: ack-apigatewayv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-apigatewayv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-apigatewayv2-controller.v1.0.7
        version: 1.0.7
      - name: ack-apigatewayv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-apigatewayv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-apigatewayv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-apigatewayv2-controller.v1.0.3
        version: 1.0.3
      - name: ack-apigatewayv2-controller.v1.0.2
        version: 1.0.2
      - name: ack-apigatewayv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-apigatewayv2-controller.v0.1.6
        version: 0.1.6
      - name: ack-apigatewayv2-controller.v0.1.4
        version: 0.1.4
      - name: ack-apigatewayv2-controller.v0.1.3
        version: 0.1.3
      - name: ack-apigatewayv2-controller.v0.1.2
        version: 0.1.2
      - name: ack-apigatewayv2-controller.v0.1.1
        version: 0.1.1
      - name: ack-apigatewayv2-controller.v0.1.0
        version: 0.1.0
      - name: ack-apigatewayv2-controller.v0.0.22
        version: 0.0.22
      - name: ack-apigatewayv2-controller.v0.0.21
        version: 0.0.21
      - name: ack-apigatewayv2-controller.v0.0.20
        version: 0.0.20
      - name: ack-apigatewayv2-controller.v0.0.19
        version: 0.0.19
      - name: ack-apigatewayv2-controller.v0.0.18
        version: 0.0.18
      - name: ack-apigatewayv2-controller.v0.0.17
        version: 0.0.17
      - name: ack-apigatewayv2-controller.v0.0.16
        version: 0.0.16
      - name: ack-apigatewayv2-controller.v0.0.15
        version: 0.0.15
      - name: ack-apigatewayv2-controller.v0.0.14
        version: 0.0.14
      - name: ack-apigatewayv2-controller.v0.0.13
        version: 0.0.13
      - name: ack-apigatewayv2-controller.v0.0.12
        version: 0.0.12
      - name: ack-apigatewayv2-controller.v0.0.11
        version: 0.0.11
      - name: ack-apigatewayv2-controller.v0.0.10
        version: 0.0.10
      - name: ack-apigatewayv2-controller.v0.0.9
        version: 0.0.9
      name: alpha
    defaultChannel: alpha
    packageName: ack-apigatewayv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Backube
      provider-url: https://github.com/backube
    name: snapscheduler
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: snapscheduler.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "snapscheduler.backube/v1",
                "kind": "SnapshotSchedule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "snapscheduler",
                    "app.kubernetes.io/instance": "daily",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "snapshotschedule",
                    "app.kubernetes.io/part-of": "snapscheduler"
                  },
                  "name": "daily"
                },
                "spec": {
                  "retention": {
                    "maxCount": 7
                  },
                  "schedule": "0 0 * * *"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/backube/snapscheduler:3.5.0
          createdAt: "2025-05-14T14:46:43Z"
          description: The SnapScheduler operator takes snapshots of Kubernetes CSI-based
            persistent volumes according to user-supplied schedules.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/snapscheduler
          support: https://github.com/backube/snapscheduler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SnapshotSchedule defines a schedule for taking automated
              snapshots of PVC(s)
            displayName: Snapshot Schedule
            kind: SnapshotSchedule
            name: snapshotschedules.snapscheduler.backube
            version: v1
        description: |-
          ## About this operator

          The SnapScheduler operator takes snapshots of CSI-based PersistentVolumes
          according to a configurable
          [Cron-like](https://en.wikipedia.org/wiki/Cron#Overview) schedule. The
          schedules include configurable retention policies for snapshots as well as
          selectors to limit the volumes that are snapshotted. An example schedule
          could be:

          > *Snapshot **all volumes** in a namespace **daily at midnight**, retaining
          > the most recent **7** snapshots for each volume.*

          Multiple schedules can be combined to provide more elaborate protection
          schemes. For example, a given volume (or collection of volumes) could be
          protected with:

          - 6 hourly snapshots
          - 7 daily snapshots
          - 4 weekly snapshots
          - 12 monthly snapshots

          ### How it works

          The operator watches for `SnapshotSchedule` CRs in each namespace. When the
          current time matches the schedule's cronspec, the operator creates a
          `VolumeSnapshot` object for each `PersistentVolumeClaim` in the namespace
          (or subset thereof if a label selector is provided). The `VolumeSnapshot`
          objects are named according to the template:
          `<pvcname>-<schedulename>-<timestamp>`. After creating the new snapshots,
          the oldest snapshots are removed if necessary, according to the retention
          policy of the schedule.

          Please see the [full
          documentation](https://backube.github.io/snapscheduler/) for more
          information.

          ## Requirements

          - OpenShift >= 4.7 (Kubernetes >= 1.20)
          - CSI-based storage driver that supports snapshots (i.e. has the
            `CREATE_DELETE_SNAPSHOT` capability)

          ## Examples

          The schedule for snapshotting is controlled by the
          `snapshotschedules.snapscheduler.backube` Custom Resource. This is a
          namespaced resource that applies only to the PersistentVolumeClaims in its
          namespace.

          Below is a simple example that keeps 7 daily (taken at midnight) snapshots
          of all PVCs in a given namespace:

          ```yaml
          ---
          apiVersion: snapscheduler.backube/v1
          kind: SnapshotSchedule
          metadata:
            name: daily
          spec:
            retention:
              maxCount: 7
            schedule: "0 0 * * *"
          ```

          See the [usage
          documentation](https://backube.github.io/snapscheduler/usage.html) for full
          details, including how to:

          - add label selectors to restrict which PVCs this schedule applies to
          - set the VolumeSnapshotClass used by the schedule
          - apply custom labels to the automatically created VolumeSnapshot objects
        displayName: SnapScheduler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - snapshot
        - storage
        - scheduler
        links:
        - name: Documentation
          url: https://backube.github.io/snapscheduler/
        - name: Source
          url: https://github.com/backube/snapscheduler
        - name: Blog
          url: https://backube.github.io/
        maintainers:
        - email: jstrunk@redhat.com
          name: John Strunk
        maturity: stable
        minKubeVersion: 1.20.0-0
        provider:
          name: Backube
          url: https://github.com/backube
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.19.1
        - quay.io/backube/snapscheduler:3.5.0
        version: 3.5.0
      entries:
      - name: snapscheduler.v3.5.0
        version: 3.5.0
      - name: snapscheduler.v3.4.0
        version: 3.4.0
      - name: snapscheduler.v3.3.0
        version: 3.3.0
      - name: snapscheduler.v3.2.0
        version: 3.2.0
      - name: snapscheduler.v3.1.0
        version: 3.1.0
      - name: snapscheduler.v3.0.0
        version: 3.0.0
      - name: snapscheduler.v2.1.0
        version: 2.1.0
      - name: snapscheduler.v2.0.0
        version: 2.0.0
      name: candidate
    - currentCSV: snapscheduler.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "snapscheduler.backube/v1",
                "kind": "SnapshotSchedule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "snapscheduler",
                    "app.kubernetes.io/instance": "daily",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "snapshotschedule",
                    "app.kubernetes.io/part-of": "snapscheduler"
                  },
                  "name": "daily"
                },
                "spec": {
                  "retention": {
                    "maxCount": 7
                  },
                  "schedule": "0 0 * * *"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/backube/snapscheduler:3.5.0
          createdAt: "2025-05-14T14:46:43Z"
          description: The SnapScheduler operator takes snapshots of Kubernetes CSI-based
            persistent volumes according to user-supplied schedules.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/snapscheduler
          support: https://github.com/backube/snapscheduler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SnapshotSchedule defines a schedule for taking automated
              snapshots of PVC(s)
            displayName: Snapshot Schedule
            kind: SnapshotSchedule
            name: snapshotschedules.snapscheduler.backube
            version: v1
        description: |-
          ## About this operator

          The SnapScheduler operator takes snapshots of CSI-based PersistentVolumes
          according to a configurable
          [Cron-like](https://en.wikipedia.org/wiki/Cron#Overview) schedule. The
          schedules include configurable retention policies for snapshots as well as
          selectors to limit the volumes that are snapshotted. An example schedule
          could be:

          > *Snapshot **all volumes** in a namespace **daily at midnight**, retaining
          > the most recent **7** snapshots for each volume.*

          Multiple schedules can be combined to provide more elaborate protection
          schemes. For example, a given volume (or collection of volumes) could be
          protected with:

          - 6 hourly snapshots
          - 7 daily snapshots
          - 4 weekly snapshots
          - 12 monthly snapshots

          ### How it works

          The operator watches for `SnapshotSchedule` CRs in each namespace. When the
          current time matches the schedule's cronspec, the operator creates a
          `VolumeSnapshot` object for each `PersistentVolumeClaim` in the namespace
          (or subset thereof if a label selector is provided). The `VolumeSnapshot`
          objects are named according to the template:
          `<pvcname>-<schedulename>-<timestamp>`. After creating the new snapshots,
          the oldest snapshots are removed if necessary, according to the retention
          policy of the schedule.

          Please see the [full
          documentation](https://backube.github.io/snapscheduler/) for more
          information.

          ## Requirements

          - OpenShift >= 4.7 (Kubernetes >= 1.20)
          - CSI-based storage driver that supports snapshots (i.e. has the
            `CREATE_DELETE_SNAPSHOT` capability)

          ## Examples

          The schedule for snapshotting is controlled by the
          `snapshotschedules.snapscheduler.backube` Custom Resource. This is a
          namespaced resource that applies only to the PersistentVolumeClaims in its
          namespace.

          Below is a simple example that keeps 7 daily (taken at midnight) snapshots
          of all PVCs in a given namespace:

          ```yaml
          ---
          apiVersion: snapscheduler.backube/v1
          kind: SnapshotSchedule
          metadata:
            name: daily
          spec:
            retention:
              maxCount: 7
            schedule: "0 0 * * *"
          ```

          See the [usage
          documentation](https://backube.github.io/snapscheduler/usage.html) for full
          details, including how to:

          - add label selectors to restrict which PVCs this schedule applies to
          - set the VolumeSnapshotClass used by the schedule
          - apply custom labels to the automatically created VolumeSnapshot objects
        displayName: SnapScheduler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - snapshot
        - storage
        - scheduler
        links:
        - name: Documentation
          url: https://backube.github.io/snapscheduler/
        - name: Source
          url: https://github.com/backube/snapscheduler
        - name: Blog
          url: https://backube.github.io/
        maintainers:
        - email: jstrunk@redhat.com
          name: John Strunk
        maturity: stable
        minKubeVersion: 1.20.0-0
        provider:
          name: Backube
          url: https://github.com/backube
        relatedImages:
        - quay.io/backube/snapscheduler:3.5.0
        - quay.io/brancz/kube-rbac-proxy:v0.19.1
        version: 3.5.0
      entries:
      - name: snapscheduler.v3.5.0
        version: 3.5.0
      - name: snapscheduler.v3.4.0
        version: 3.4.0
      - name: snapscheduler.v3.3.0
        version: 3.3.0
      - name: snapscheduler.v3.2.0
        version: 3.2.0
      - name: snapscheduler.v3.1.0
        version: 3.1.0
      - name: snapscheduler.v3.0.0
        version: 3.0.0
      - name: snapscheduler.v2.1.0
        version: 2.1.0
      - name: snapscheduler.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: snapscheduler
    provider:
      name: Backube
      url: https://github.com/backube
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: watcher-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: watcher-operator.v1.0.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "Watcher",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcher",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcher",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcher"
                },
                "spec": {
                  "apiServiceTemplate": {
                    "tls": {
                      "caBundleSecretName": "combined-ca-bundle"
                    }
                  },
                  "databaseInstance": "openstack"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherAPI",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcher-operator"
                  },
                  "name": "watcherapi-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherApplier",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcherapplier-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcherapplier",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcherapplier-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherDecisionEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcherdecisionengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcherdecisionengine",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcherdecisionengine-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-09-02T13:32:31Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.14
          operatorframework.io/suggested-namespace: openstack-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: WatcherAPI is the Schema for the watcherapis API
            displayName: Watcher API
            kind: WatcherAPI
            name: watcherapis.watcher.openstack.org
            version: v1beta1
          - description: WatcherApplier is the Schema for the watcherappliers API
            displayName: Watcher Applier
            kind: WatcherApplier
            name: watcherappliers.watcher.openstack.org
            version: v1beta1
          - description: WatcherDecisionEngine is the Schema for the watcherdecisionengines
              API
            displayName: Watcher Decision Engine
            kind: WatcherDecisionEngine
            name: watcherdecisionengines.watcher.openstack.org
            version: v1beta1
          - description: Watcher is the Schema for the watchers API
            displayName: Watcher
            kind: Watcher
            name: watchers.watcher.openstack.org
            version: v1beta1
        description: The Watcher Operator project
        displayName: Watcher Operator (Technology Preview)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        links:
        - name: Watcher Operator
          url: https://github.com/openstack-k8s-operators/watcher-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/watcher-rhel9-operator@sha256:63c341a3e20a522c601098502a8fd8f1164d23c773ae8f55782c43ff4b6c8c18
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:eed8234b0fe47083ab941f513709f778bfcb80e32747b71a65fd5f62328a5ca0
        - registry.redhat.io/rhoso/openstack-watcher-api-rhel9@sha256:e97369e3a47bb882f1920c6af3f6345a842af45c58f0f5c5df9ac2933cf3098a
        - registry.redhat.io/rhoso/openstack-watcher-decision-engine-rhel9@sha256:c759d66cf0986c0036be91c008cbd059b4412c94160ed82dd12cd4e00d0685a9
        - registry.redhat.io/rhoso/openstack-watcher-applier-rhel9@sha256:c92a82f901613991926367bb8aab4f875c27a3e4a5282d92847a8463d98cd9e3
        version: 1.0.14
      entries:
      - name: watcher-operator.v1.0.14
        version: 1.0.14
      - name: watcher-operator.v1.0.13
        version: 1.0.13
      - name: watcher-operator.v1.0.12
        version: 1.0.12
      - name: watcher-operator.v1.0.11
        version: 1.0.11
      - name: watcher-operator.v1.0.10
        version: 1.0.10
      - name: watcher-operator.v1.0.9
        version: 1.0.9
      - name: watcher-operator.v1.0.8
        version: 1.0.8
      - name: watcher-operator.v1.0.7
        version: 1.0.7
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: watcher-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NdmSpc project
      provider-url: ""
    name: ndmspc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ndmspc-operator.v0.20250209.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.ndmspc.io/v1alpha1",
                "kind": "NdmSpcConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ndmspc-operator",
                    "app.kubernetes.io/instance": "ndmspcconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ndmspcconfig",
                    "app.kubernetes.io/part-of": "ndmspc-operator"
                  },
                  "name": "ndmspc-sample"
                },
                "spec": {
                  "executor": {
                    "image": "registry.gitlab.com/ndmspc/api:v0.20240215.2",
                    "name": "executor"
                  },
                  "executors": 1,
                  "ingressclasstype": "nginx",
                  "ingressenabled": true,
                  "ingresshostdomain": "127.0.0.1.sslip.io",
                  "ingressnostname": "ndmspc",
                  "ingressssltraefik": false,
                  "naf": {
                    "image": "registry.gitlab.com/ndmspc/naf:v0.20250129.0",
                    "name": "naf"
                  },
                  "ndmvr": {
                    "image": "registry.gitlab.com/ndmspc/ndmvr/shared:v0.20230320.0",
                    "name": "ndmvr"
                  },
                  "salsa": {
                    "image": "registry.gitlab.com/ndmspc/user/ndmspc:v0.20240412.0"
                  },
                  "storage": {
                    "image": "registry.gitlab.com/ndmspc/storage-server:v0.20250205.0",
                    "mongodb": {
                      "image": "docker.io/library/mongo:latest",
                      "password": "admin",
                      "port": 27017,
                      "username": "admin"
                    },
                    "name": "storage",
                    "port": 8080
                  },
                  "webapp": {
                    "image": "registry.gitlab.com/ndmspc/react-ndmspc:v0.20240412.0",
                    "name": ""
                  },
                  "webappcontainerport": 8080,
                  "webapps": 1,
                  "worker": {
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "1Gi"
                      }
                    }
                  },
                  "workers": 2,
                  "workerslots": 1,
                  "zmq2ws": {
                    "image": "registry.gitlab.com/ndmspc/zmq2ws:v1.20231012.0",
                    "name": "zmq2ws"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: registry.gitlab.com/ndmspc/ndmspc-operator:0.20250209.0
          createdAt: "2025-02-09T12:49:06Z"
          description: The NdmSpc operator automates the deployment and management
            of Ndmspc suite on Kubernetes cluster.
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://gitlab.com/ndmspc/ndmspc-operator
          support: NDMSPC Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NdmSpc configuration
            displayName: NdmSpc Config
            kind: NdmSpcConfig
            name: ndmspcconfigs.apps.ndmspc.io
            version: v1alpha1
        description: "The NdmSpc operator automates the deployment and management
          of Ndmspc suite on Kubernetes cluster. This operator allows user to deploy
          SALSA cluster with ZMQ2WS and job executor service and specify custom webapp
          page.\n\n### Capabilities\n\n- SALSA cluster\n- ZMQ2WS api service\n- Job
          executor service \n- Custom WebApp page\n"
        displayName: NdmSpc operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ndmspc
        links:
        - name: Ndmspc Operator
          url: https://ndmspc.gitlab.io/docs
        maintainers:
        - email: ndmspc@gmail.com
          name: NdmSpc Support
        maturity: alpha
        provider:
          name: NdmSpc project
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - registry.gitlab.com/ndmspc/ndmspc-operator:0.20250209.0
        version: 0.20250209.0
      entries:
      - name: ndmspc-operator.v0.20250209.0
        version: 0.20250209.0
      - name: ndmspc-operator.v0.20250131.0
        version: 0.20250131.0
      - name: ndmspc-operator.v0.20240412.0
        version: 0.20240412.0
      - name: ndmspc-operator.v0.11.6
        version: 0.11.6
      - name: ndmspc-operator.v0.11.4
        version: 0.11.4
      - name: ndmspc-operator.v0.11.3
        version: 0.11.3
      - name: ndmspc-operator.v0.11.2
        version: 0.11.2
      - name: ndmspc-operator.v0.11.1
        version: 0.11.1
      - name: ndmspc-operator.v0.10.2
        version: 0.10.2
      - name: ndmspc-operator.v0.10.1
        version: 0.10.1
      - name: ndmspc-operator.v0.10.0
        version: 0.10.0
      - name: ndmspc-operator.v0.9.0
        version: 0.9.0
      - name: ndmspc-operator.v0.7.4
        version: 0.7.4
      - name: ndmspc-operator.v0.7.1
        version: 0.7.1
      - name: ndmspc-operator.v0.6.0
        version: 0.6.0
      - name: ndmspc-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: ndmspc-operator
    provider:
      name: NdmSpc project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: VMware Tanzu
      provider-url: ""
    name: rabbitmq-cluster-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-cluster-operator.v2.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "rabbitmq.com/v1beta1",
                "kind": "RabbitmqCluster",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "override": {
                    "statefulset": {
                      "spec": {
                        "template": {
                          "spec": {
                            "containers": [],
                            "securityContext": {}
                          }
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/rabbitmqoperator/cluster-operator:2.16.0
          createdAt: 2025-08-12T09:34:33
          repository: https://github.com/rabbitmq/cluster-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Rabbitmq Cluster
            displayName: RabbitmqCluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: "## About the managed application\nRabbitMQ is an open-source
          message broker that originally implemented the\nAdvanced Message Queuing
          Protocol (AMQP). Nowadays features a plug-in\narchitecture supporting protocols
          such as Streaming Text Oriented Messaging\nProtocol (STOMP), MQ Telemetry
          Transport (MQTT), and others. RabbitMQ\nsupports native streaming and AMQP
          1.0.\n\nVisit [rabbitmq.com](rabbitmq.com) for more information.\n\n## About
          this Operator\nRabbitMQ Cluster Kubernetes Operator is developed and maintened
          by the VMware \nRabbitMQ team and provides a consistent and easy way to
          deploy RabbitMQ\nclusters to Kubernetes and run them, including \"day two\"
          (continuous)\noperations. RabbitMQ clusters deployed using the Operator
          can be used by\napplications running on Kubernetes or outside of Kubernetes.\n\nRead
          [RabbitMQ operator docs](https://www.rabbitmq.com/kubernetes/operator/operator-overview)\nfor
          more information.\n"
        displayName: RabbitMQ-cluster-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        - messaging
        - amqp
        - amqp10
        - stomp
        - mqtt
        - mqtt5
        - queues
        - streaming
        links:
        - name: Documentation
          url: https://www.rabbitmq.com/kubernetes/operator/operator-overview.html
        - name: GitHub
          url: https://github.com/rabbitmq/cluster-operator
        maintainers:
        - email: rabbitmq-users@googlegroups.com
          name: VMware Tanzu
        maturity: stable
        minKubeVersion: 1.26.0
        nativeApis:
        - group: apps
          kind: Deployment
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: ReplicaSet
          version: v1
        - group: ""
          kind: Pod
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: ConfigMap
          version: v1
        provider:
          name: VMware Tanzu
        relatedImages:
        - quay.io/rabbitmqoperator/cluster-operator:2.16.0
        version: 2.16.0
      entries:
      - name: rabbitmq-cluster-operator.v2.16.0
        version: 2.16.0
      - name: rabbitmq-cluster-operator.v2.15.0
        version: 2.15.0
      - name: rabbitmq-cluster-operator.v2.14.0
        version: 2.14.0
      - name: rabbitmq-cluster-operator.v2.13.0
        version: 2.13.0
      - name: rabbitmq-cluster-operator.v2.12.1
        version: 2.12.1
      - name: rabbitmq-cluster-operator.v2.10.0
        version: 2.10.0
      - name: rabbitmq-cluster-operator.v2.9.0
        version: 2.9.0
      - name: rabbitmq-cluster-operator.v2.8.0
        version: 2.8.0
      - name: rabbitmq-cluster-operator.v2.7.0
        version: 2.7.0
      - name: rabbitmq-cluster-operator.v2.6.0
        version: 2.6.0
      - name: rabbitmq-cluster-operator.v2.5.0
        version: 2.5.0
      - name: rabbitmq-cluster-operator.v2.4.0
        version: 2.4.0
      - name: rabbitmq-cluster-operator.v2.3.0
        version: 2.3.0
      - name: rabbitmq-cluster-operator.v2.2.0
        version: 2.2.0
      - name: rabbitmq-cluster-operator.v2.1.0
        version: 2.1.0
      - name: rabbitmq-cluster-operator.v2.0.0
        version: 2.0.0
      - name: rabbitmq-cluster-operator.v1.14.0
        version: 1.14.0
      name: stable
    defaultChannel: stable
    packageName: rabbitmq-cluster-operator
    provider:
      name: VMware Tanzu
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: ""
    name: ibm-metrics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-metrics-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MarketplaceConfig",
                "metadata": {
                  "name": "marketplaceconfig"
                },
                "spec": {
                  "features": {
                    "deployment": true,
                    "meterDefinitionCatalogServer": false,
                    "registration": true
                  },
                  "installIBMCatalogSource": true,
                  "isDisconnected": false,
                  "license": {
                    "accept": false
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterBase",
                "metadata": {
                  "name": "meterbase"
                },
                "spec": {
                  "enabled": true,
                  "prometheus": {
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "1",
                        "memory": "1G"
                      }
                    },
                    "storage": {
                      "size": "20Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterReport",
                "metadata": {
                  "name": "example-meterreport"
                },
                "spec": {
                  "endTime": "2020-08-05T00:00:00Z",
                  "prometheusService": {
                    "name": "rhm-prometheus-meterbase",
                    "namespace": "openshift-redhat-marketplace",
                    "targetPort": "rbac"
                  },
                  "startTime": "2020-08-04T00:00:00Z"
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "example-meterdefinition-4"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "meterdefinition-sample"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
          createdAt: "2025-08-28T14:18:55Z"
          description: The IBM Metrics Operator provides workload metering and reporting
            for IBM and Red Hat Marketplace customers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
          - description: MeterBase is the resource that sets up Metering for Red Hat
              Marketplace. This is an internal resource not meant to be modified directly.
            displayName: Meter Base
            kind: MeterBase
            name: meterbases.marketplace.redhat.com
            version: v1alpha1
          - description: MeterDefinition defines the meter workloads used to enable
              pay for use billing.
            displayName: Meter Definition
            kind: MeterDefinition
            name: meterdefinitions.marketplace.redhat.com
            version: v1beta1
          - kind: MeterReport
            name: meterreports.marketplace.redhat.com
            version: v1alpha1
        description: "The IBM Metrics Operator provides workload metering and reporting
          for IBM Software Central and Red Hat Marketplace customers.\n### **Important
          Note**\nA set of instructions for onboarding is provided here. For more
          detailed onboarding instructions or information about what is installed
          please visit [swc.saas.ibm.com](https://swc.saas.ibm.com).\n\nUsage metrics
          may be monitored through [https://swc.saas.ibm.com](https://swc.saas.ibm.com)
          with only IBM Metrics Operator and a Red Hat Marketplace account.\n\n###
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n### Prerequisites\n1. Installations are required
          to [enable monitoring for user-defined projects](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          as the Prometheus provider.\n2. Edit the cluster-monitoring-config ConfigMap
          object:\n\n   ```sh\n   $ oc -n openshift-monitoring edit configmap cluster-monitoring-config\n
          \   ```\n\n3. Add enableUserWorkload: true under data/config.yaml:\n  \n
          \   ```sh\n    apiVersion: v1\n    kind: ConfigMap\n    metadata:\n      name:
          cluster-monitoring-config\n        namespace: openshift-monitoring\n    data:\n
          \     config.yaml: |\n        enableUserWorkload: true\n    ```\n\n4. Configure
          the user-workload-monitoring-config ConfigMap object:\n\n    ```sh\n    $
          oc -n openshift-user-workload-monitoring edit configmap user-workload-monitoring-config\n
          \   ```\n\n5. Configure a minimum retention time of 168h and minimum storage
          capacity of 40Gi\n  \n    ```sh\n    apiVersion: v1\n    kind: ConfigMap\n
          \   metadata:\n      name: user-workload-monitoring-config\n      namespace:
          openshift-user-workload-monitoring\n\n    data:\n      config.yaml: |\n
          \       prometheus:\n          retention: 168h\n          volumeClaimTemplate:\n
          \           spec:\n              resources:\n                requests:\n
          \                 storage: 40Gi\n    ```\n\n### Resources Required\n\nMinimum
          system resources required:\n\n| Operator                | Memory (MB) |
          CPU (cores) | Disk (GB) | Nodes |\n| ----------------------- | -----------
          | ----------- | --------- | ----- |\n| **Metrics**   |        750  |     0.25
          \   | 3x1       |    3  |\n| **Deployment** |        250  |     0.25    |
          -         |    1  |\n\n| Prometheus Provider  | Memory (GB) | CPU (cores)
          | Disk (GB) | Nodes |\n| --------- | ----------- | ----------- | ---------
          | ----- |\n| **[Openshift User Workload Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)**
          |          1  |     0.1       | 2x40        |   2    |\n\nMultiple nodes
          are required to provide pod scheduling for high availability for IBM Metrics
          Operator Data Service and Prometheus.\n\nThe IBM Metrics Operator automatically
          creates 3 x 1Gi PersistentVolumeClaims to store reports as part of the data
          service, with _ReadWriteOnce_ access mode. Te PersistentVolumeClaims are
          automatically created by the ibm-metrics-operator after creating a `redhat-marketplace-pull-secret`
          and accepting the license in `marketplaceconfig`.\n\n| NAME                                |
          CAPACITY | ACCESS MODES |\n| ----------------------------------- | --------
          | ------------ |\n| rhm-data-service-rhm-data-service-0 | 1Gi | RWO |\n|
          rhm-data-service-rhm-data-service-1 | 1Gi | RWO |\n| rhm-data-service-rhm-data-service-2
          | 1Gi | RWO |\n\n### Supported Storage Providers\n\n- OpenShift Container
          Storage / OpenShift Data Foundation version 4.x, from version 4.2 or higher\n-
          IBM Cloud Block storage and IBM Cloud File storage\n- IBM Storage Suite
          for IBM Cloud Paks:\n  - File storage from IBM Spectrum Fusion/Scale \n
          \ - Block storage from IBM Spectrum Virtualize, FlashSystem or DS8K\n- Portworx
          Storage, version 2.5.5 or above\n- Amazon Elastic File Storage\n\n### Access
          Modes required\n\n - ReadWriteOnce (RWO)\n\n### Provisioning Options supported\n\nChoose
          one of the following options to provision storage for the ibm-metrics-operator
          data-service\n\n#### Dynamic provisioning using a default StorageClass\n
          \  - A StorageClass is defined with a `metadata.annotations: storageclass.kubernetes.io/is-default-class:
          \"true\"`\n   - PersistentVolumes will be provisioned automatically for
          the generated PersistentVolumeClaims\n--- \n#### Manually create each PersistentVolumeClaim
          with a specific StorageClass\n   - Must be performed before creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`. Otherwise, the automatically
          generated PersistentVolumeClaims are immutable.\n```\napiVersion: v1\nkind:
          PersistentVolumeClaim\nmetadata:\n  labels:\n    app: rhm-data-service\n
          \ name: rhm-data-service-rhm-data-service-0\n  namespace: ibm-software-central\nspec:\n
          \ storageClassName: rook-cephfs\n  accessModes:\n  - ReadWriteOnce\n  resources:\n
          \   requests:\n      storage: 1Gi\n```\n---\n#### Manually provision each
          PersistentVolume for the generated PersistentVolumeClaims with a specific
          StorageClass\n  - May be performed before or after creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`.  \n```\napiVersion: v1\nkind:
          PersistentVolume\nmetadata:\n  name: rhm-data-service-rhm-data-service-0\nspec:\n
          \ csi:\n    driver: rook-ceph.cephfs.csi.ceph.com\n    volumeHandle: rhm-data-service-rhm-data-service-0\n
          \ capacity:\n    storage: 1Gi\n  accessModes:\n    - ReadWriteOnce\n  persistentVolumeReclaimPolicy:
          Delete\n  storageClassName: rook-cephfs\n  volumeMode: Filesystem\n  claimRef:\n
          \   kind: PersistentVolumeClaim\n    namespace: ibm-software-central\n    name:
          rhm-data-service-rhm-data-service-0\n```\n\n### Installation\n1. Create
          or get your [pull secret](https://swc.saas.ibm.com/en-us/documentation/clusters#get-pull-secret).\n2.
          Install the IBM Metrics Operator\n3. Create a Kubernetes secret in the installed
          namespace with the name `redhat-marketplace-pull-secret` and key `PULL_SECRET`
          with the value of the pull secret.\n    ```sh\n    # Replace ${PULL_SECRET}
          with your pull secret\n    oc create secret generic redhat-marketplace-pull-secret
          -n ibm-software-central --from-literal=PULL_SECRET=${PULL_SECRET}\n    ```\n4.
          Use of the platform is governed by the:\n\n    [IBM Cloud Services Agreement](https://www.ibm.com/support/customer/csol/terms/?id=Z126-6304_WS&_ga=2.116312197.2046730452.1684328846-812467790.1684328846)
          (or other base agreement between you and IBM such as a [Passport Advantage
          Agreement](https://www.ibm.com/software/passportadvantage/pa_agreements.html?_ga=2.116312197.2046730452.1684328846-812467790.1684328846))
          and the [Service Description for the Red Hat Marketplace](https://www.ibm.com/support/customer/csol/terms/?id=i126-8719&_ga=2.83289621.2046730452.1684328846-812467790.1684328846).\n
          \   \n5. Update MarketplaceConfig to accept the license.\n    ```\n    oc
          patch marketplaceconfig marketplaceconfig -n ibm-software-central --type='merge'
          -p '{\"spec\": {\"license\": {\"accept\": true}}}'\n    ```\n6. Install
          the pull secret as a global pull secret on the cluster.\n\n    These steps
          require `oc`, `jq`, and `base64` to be available on your machine.\n\n    ```sh\n
          \   # Create the docker pull secret file using your pull secret.\n    #
          Store it in a file called entitledregistryconfigjson.\n    oc create secret
          docker-registry entitled-registry --docker-server=registry.marketplace.redhat.com
          --docker-username \"cp\" --docker-password \"${PULL_SECRET}\" --dry-run=client
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > entitledregistryconfigjson\n
          \   # Get the current global secrets on the cluster and store it as a file
          named dockerconfigjson\n    oc get secret pull-secret -n openshift-config
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > dockerconfigjson\n
          \   # Merge the two dockerconfigs together into a file called dockerconfigjson-merged.\n
          \   jq -s '.[0] * .[1]' dockerconfigjson entitledregistryconfigjson > dockerconfigjson-merged\n
          \   # Set the cluster's dockerconfig file to the new merged version.\n    oc
          set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson=dockerconfigjson-merged\n
          \   ```\n  7. (Optional) Your IBM product may have installed IBM Metrics
          Operator using an `ibm-entitlement-key` for authentication to IBM Software
          Central. To authorize the automatic creation of an IBM Software Central
          account for the IBM account owning the `ibm-entitlement-key`\n      ```\n
          \     oc patch marketplaceconfig marketplaceconfig -n ibm-software-central
          --type='merge' -p '{\"spec\": {\"authorizeAccountCreation\": true}}'\n      ```\n\n###
          Why is a global pull secret required?\nIn order to successfully install
          the products hosted by the container image registry, you will need to make
          the pull secret available across the cluster. This can be achieved by applying
          the pull as a [global pull secret](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html#images-update-global-pull-secret_using-image-pull-secrets).
          For alternative approaches, please see the official OpenShift [documentation](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html).\n\n\n###
          SecurityContextConstraints requirements\n\nThe Operators and their components
          support running under the OpenShift Container Platform default restricted
          and restricted-v2 security context constraints.\n\n### Installation Namespace
          and ClusterRoleBinding requirements\n\nThe IBM Metrics Operator components
          require specific ClusterRoleBindings.\n- The metric-state component requires
          a ClusterRoleBinding for the the `view` ClusterRole.\n  - Ability to read
          non-sensitive CustomResources. \n  - The `view` ClusterRole dynamically
          adds new CustomResourceDefinitions.\n- The operator & reporter component
          requires a ClusterRoleBinding for the the `cluster-monitoring-view` ClusterRole.\n
          \ - Ability to view Prometheus metrics. \n  - The underlying ClusterRole
          RBAC often updated between OpenShift versions.\n\nDue to limitations of
          Operator Lifecycle Manager (OLM), this ClusterRoleBinding can not be provided
          dynamically for arbitrary installation target namespaces.\n\nA static ClusterRoleBinding
          is included for installation to the default namespace of `ibm-software-central`,
          and namespaces `openshift-redhat-marketplace`,  `redhat-marketplace`, `ibm-common-services`.\n\nTo
          create the ClusterRoleBindings for installation to an alternate namespace\n```\noc
          project INSTALL-NAMESPACE\noc adm policy add-cluster-role-to-user view -z
          ibm-metrics-operator-metric-state\noc adm policy add-cluster-role-to-user
          cluster-monitoring-view -z ibm-metrics-operator-controller-manager,ibm-metrics-operator-reporter\n```\n\n###
          Metric State scoping requirements\nThe metric-state Deployment obtains `get/list/watch`
          access to metered resources via the `view` ClusterRole. For operators deployed
          using Operator Lifecycle Manager (OLM), permissions are added to `clusterrole/view`
          dynamically via a generated and annotated `-view` ClusterRole. If you wish
          to meter an operator, and its Custom Resource Definitions (CRDs) are not
          deployed through OLM, one of two options are required\n1. Add the following
          label to a clusterrole that has get/list/watch access to your CRD: `rbac.authorization.k8s.io/aggregate-to-view:
          \"true\"`, thereby dynamically adding it to `clusterrole/view`\n2. Create
          a ClusterRole that has get/list/watch access to your CRD, and create a ClusterRoleBinding
          for the metric-state ServiceAccount\n\nAttempting to meter a resource with
          a MeterDefinition without the required permissions will log an `AccessDeniedError`
          in metric-state.\n\n### Disaster Recovery\n\nTo plan for disaster recovery,
          note the PhysicalVolumeClaims `rhm-data-service-rhm-data-service-N`. \n-
          In connected environments, MeterReport data upload attempts occur hourly,
          and are then removed from data-service. There is a low risk of losing much
          unreported data.\n- In an airgap environment, MeterReport data must be pulled
          from data-service and uploaded manually using [datactl](https://github.com/redhat-marketplace/datactl/).
          To prevent data loss in a disaster scenario, the data-service volumes should
          be considered in a recovery plan.\n\n### Subscription Config\n\nIt is possible
          to [configure](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md)
          how OLM deploys an Operator via the `config` field in the [Subscription](https://github.com/operator-framework/olm-book/blob/master/docs/subscriptions.md)
          object.\n\nThe IBM Metrics Operator will also read the `config` and append
          it to the operands. The primary use case is to control scheduling using
          [Tolerations](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/)
          and [NodeSelectors](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/).\n\nA
          limitation is that the `config` elements are only appended to the operands.
          The elements in the operands are not removed if the `config` is removed
          from the `Subscripion`. The operand must be modified manually, or deleted
          and recreated by the controller.\n\n### Cluster permission requirements\n\n|API
          group             |Resources              |Verbs                                     |\n|----------------------|-----------------------|------------------------------------------|\n|apps
          \                 |statefulsets           |get;list;watch                            |\n|authentication.k8s.io
          |tokenreviews           |create                                    |\n|authorization.k8s.io
          \ |subjectaccessreviews   |create                                    |\n|config.openshift.io
          \  |clusterversions        |get;list;watch                            |\n|marketplace.redhat.com|meterdefinitions
          \      |get;list;watch;create;update;patch;delete |\n|operators.coreos.com
          \ |clusterserviceversions |get;list;watch                            |\n|operators.coreos.com
          \ |subscriptions          |get;list;watch                            |\n|operators.coreos.com
          \ |operatorgroups         |get;list                                  |\n|storage.k8s.io
          \       |storageclasses         |get;list;watch                            |\n|
          \                     |namespaces             |get;list;watch                            |\n|
          \                     |configmaps             |get;list;watch                            |\n|
          \                     |services               |get;list;watch                            |\n|
          \                     |**clusterrole/view**   |                                          |\n\n\n###
          Documentation\nYou can find our documentation [here.](https://swc.saas.ibm.com/en-us/documentation/)\n\n###
          Getting help\nIf you encounter any issues while using the operators, you
          can create an issue on our [Github\nrepo](https://github.com/redhat-marketplace/redhat-marketplace-operator)
          for bugs, enhancements, or other requests. You can also visit our main page
          and\nreview our [support](https://swc.saas.ibm.com/en-us/support) and [documentation](https://swc.saas.ibm.com/en-us/documentation/).\n\n###
          Readme\nYou can find our readme [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/README.md)\n\n###
          License information\nYou can find our license information [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/LICENSE)\n"
        displayName: IBM Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        - buy
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/redhat-marketplace-authcheck@sha256:13a81b8aca79ae8c89ee24dcc30f3322a4cfdedf929e7bb984295c1e7e631fa9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/redhat-marketplace-data-service@sha256:30fd8d2996b921cfce51f697ee4490a6d9a836021952429b877336f5edf2749d
        - quay.io/rh-marketplace/redhat-marketplace-reporter@sha256:2bd098f057ca07f2b14be9b60ea6ad8ca6b0cae1d226b1ca3b27e23d04ccdf20
        - quay.io/rh-marketplace/redhat-marketplace-metric-state@sha256:e4014c9a546c10e84b13a09db6a18431ee9489ad1296808635e48d9ef4963cf1
        - quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
        version: 2.20.0
      entries:
      - name: ibm-metrics-operator.v2.20.0
        version: 2.20.0
      name: beta
    - currentCSV: ibm-metrics-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MarketplaceConfig",
                "metadata": {
                  "name": "marketplaceconfig"
                },
                "spec": {
                  "features": {
                    "deployment": true,
                    "meterDefinitionCatalogServer": false,
                    "registration": true
                  },
                  "installIBMCatalogSource": true,
                  "isDisconnected": false,
                  "license": {
                    "accept": false
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterBase",
                "metadata": {
                  "name": "meterbase"
                },
                "spec": {
                  "enabled": true,
                  "prometheus": {
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "1",
                        "memory": "1G"
                      }
                    },
                    "storage": {
                      "size": "20Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterReport",
                "metadata": {
                  "name": "example-meterreport"
                },
                "spec": {
                  "endTime": "2020-08-05T00:00:00Z",
                  "prometheusService": {
                    "name": "rhm-prometheus-meterbase",
                    "namespace": "openshift-redhat-marketplace",
                    "targetPort": "rbac"
                  },
                  "startTime": "2020-08-04T00:00:00Z"
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "example-meterdefinition-4"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "meterdefinition-sample"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
          createdAt: "2025-08-28T14:18:55Z"
          description: The IBM Metrics Operator provides workload metering and reporting
            for IBM and Red Hat Marketplace customers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
          - description: MeterBase is the resource that sets up Metering for Red Hat
              Marketplace. This is an internal resource not meant to be modified directly.
            displayName: Meter Base
            kind: MeterBase
            name: meterbases.marketplace.redhat.com
            version: v1alpha1
          - description: MeterDefinition defines the meter workloads used to enable
              pay for use billing.
            displayName: Meter Definition
            kind: MeterDefinition
            name: meterdefinitions.marketplace.redhat.com
            version: v1beta1
          - kind: MeterReport
            name: meterreports.marketplace.redhat.com
            version: v1alpha1
        description: "The IBM Metrics Operator provides workload metering and reporting
          for IBM Software Central and Red Hat Marketplace customers.\n### **Important
          Note**\nA set of instructions for onboarding is provided here. For more
          detailed onboarding instructions or information about what is installed
          please visit [swc.saas.ibm.com](https://swc.saas.ibm.com).\n\nUsage metrics
          may be monitored through [https://swc.saas.ibm.com](https://swc.saas.ibm.com)
          with only IBM Metrics Operator and a Red Hat Marketplace account.\n\n###
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n### Prerequisites\n1. Installations are required
          to [enable monitoring for user-defined projects](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          as the Prometheus provider.\n2. Edit the cluster-monitoring-config ConfigMap
          object:\n\n   ```sh\n   $ oc -n openshift-monitoring edit configmap cluster-monitoring-config\n
          \   ```\n\n3. Add enableUserWorkload: true under data/config.yaml:\n  \n
          \   ```sh\n    apiVersion: v1\n    kind: ConfigMap\n    metadata:\n      name:
          cluster-monitoring-config\n        namespace: openshift-monitoring\n    data:\n
          \     config.yaml: |\n        enableUserWorkload: true\n    ```\n\n4. Configure
          the user-workload-monitoring-config ConfigMap object:\n\n    ```sh\n    $
          oc -n openshift-user-workload-monitoring edit configmap user-workload-monitoring-config\n
          \   ```\n\n5. Configure a minimum retention time of 168h and minimum storage
          capacity of 40Gi\n  \n    ```sh\n    apiVersion: v1\n    kind: ConfigMap\n
          \   metadata:\n      name: user-workload-monitoring-config\n      namespace:
          openshift-user-workload-monitoring\n\n    data:\n      config.yaml: |\n
          \       prometheus:\n          retention: 168h\n          volumeClaimTemplate:\n
          \           spec:\n              resources:\n                requests:\n
          \                 storage: 40Gi\n    ```\n\n### Resources Required\n\nMinimum
          system resources required:\n\n| Operator                | Memory (MB) |
          CPU (cores) | Disk (GB) | Nodes |\n| ----------------------- | -----------
          | ----------- | --------- | ----- |\n| **Metrics**   |        750  |     0.25
          \   | 3x1       |    3  |\n| **Deployment** |        250  |     0.25    |
          -         |    1  |\n\n| Prometheus Provider  | Memory (GB) | CPU (cores)
          | Disk (GB) | Nodes |\n| --------- | ----------- | ----------- | ---------
          | ----- |\n| **[Openshift User Workload Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)**
          |          1  |     0.1       | 2x40        |   2    |\n\nMultiple nodes
          are required to provide pod scheduling for high availability for IBM Metrics
          Operator Data Service and Prometheus.\n\nThe IBM Metrics Operator automatically
          creates 3 x 1Gi PersistentVolumeClaims to store reports as part of the data
          service, with _ReadWriteOnce_ access mode. Te PersistentVolumeClaims are
          automatically created by the ibm-metrics-operator after creating a `redhat-marketplace-pull-secret`
          and accepting the license in `marketplaceconfig`.\n\n| NAME                                |
          CAPACITY | ACCESS MODES |\n| ----------------------------------- | --------
          | ------------ |\n| rhm-data-service-rhm-data-service-0 | 1Gi | RWO |\n|
          rhm-data-service-rhm-data-service-1 | 1Gi | RWO |\n| rhm-data-service-rhm-data-service-2
          | 1Gi | RWO |\n\n### Supported Storage Providers\n\n- OpenShift Container
          Storage / OpenShift Data Foundation version 4.x, from version 4.2 or higher\n-
          IBM Cloud Block storage and IBM Cloud File storage\n- IBM Storage Suite
          for IBM Cloud Paks:\n  - File storage from IBM Spectrum Fusion/Scale \n
          \ - Block storage from IBM Spectrum Virtualize, FlashSystem or DS8K\n- Portworx
          Storage, version 2.5.5 or above\n- Amazon Elastic File Storage\n\n### Access
          Modes required\n\n - ReadWriteOnce (RWO)\n\n### Provisioning Options supported\n\nChoose
          one of the following options to provision storage for the ibm-metrics-operator
          data-service\n\n#### Dynamic provisioning using a default StorageClass\n
          \  - A StorageClass is defined with a `metadata.annotations: storageclass.kubernetes.io/is-default-class:
          \"true\"`\n   - PersistentVolumes will be provisioned automatically for
          the generated PersistentVolumeClaims\n--- \n#### Manually create each PersistentVolumeClaim
          with a specific StorageClass\n   - Must be performed before creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`. Otherwise, the automatically
          generated PersistentVolumeClaims are immutable.\n```\napiVersion: v1\nkind:
          PersistentVolumeClaim\nmetadata:\n  labels:\n    app: rhm-data-service\n
          \ name: rhm-data-service-rhm-data-service-0\n  namespace: ibm-software-central\nspec:\n
          \ storageClassName: rook-cephfs\n  accessModes:\n  - ReadWriteOnce\n  resources:\n
          \   requests:\n      storage: 1Gi\n```\n---\n#### Manually provision each
          PersistentVolume for the generated PersistentVolumeClaims with a specific
          StorageClass\n  - May be performed before or after creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`.  \n```\napiVersion: v1\nkind:
          PersistentVolume\nmetadata:\n  name: rhm-data-service-rhm-data-service-0\nspec:\n
          \ csi:\n    driver: rook-ceph.cephfs.csi.ceph.com\n    volumeHandle: rhm-data-service-rhm-data-service-0\n
          \ capacity:\n    storage: 1Gi\n  accessModes:\n    - ReadWriteOnce\n  persistentVolumeReclaimPolicy:
          Delete\n  storageClassName: rook-cephfs\n  volumeMode: Filesystem\n  claimRef:\n
          \   kind: PersistentVolumeClaim\n    namespace: ibm-software-central\n    name:
          rhm-data-service-rhm-data-service-0\n```\n\n### Installation\n1. Create
          or get your [pull secret](https://swc.saas.ibm.com/en-us/documentation/clusters#get-pull-secret).\n2.
          Install the IBM Metrics Operator\n3. Create a Kubernetes secret in the installed
          namespace with the name `redhat-marketplace-pull-secret` and key `PULL_SECRET`
          with the value of the pull secret.\n    ```sh\n    # Replace ${PULL_SECRET}
          with your pull secret\n    oc create secret generic redhat-marketplace-pull-secret
          -n ibm-software-central --from-literal=PULL_SECRET=${PULL_SECRET}\n    ```\n4.
          Use of the platform is governed by the:\n\n    [IBM Cloud Services Agreement](https://www.ibm.com/support/customer/csol/terms/?id=Z126-6304_WS&_ga=2.116312197.2046730452.1684328846-812467790.1684328846)
          (or other base agreement between you and IBM such as a [Passport Advantage
          Agreement](https://www.ibm.com/software/passportadvantage/pa_agreements.html?_ga=2.116312197.2046730452.1684328846-812467790.1684328846))
          and the [Service Description for the Red Hat Marketplace](https://www.ibm.com/support/customer/csol/terms/?id=i126-8719&_ga=2.83289621.2046730452.1684328846-812467790.1684328846).\n
          \   \n5. Update MarketplaceConfig to accept the license.\n    ```\n    oc
          patch marketplaceconfig marketplaceconfig -n ibm-software-central --type='merge'
          -p '{\"spec\": {\"license\": {\"accept\": true}}}'\n    ```\n6. Install
          the pull secret as a global pull secret on the cluster.\n\n    These steps
          require `oc`, `jq`, and `base64` to be available on your machine.\n\n    ```sh\n
          \   # Create the docker pull secret file using your pull secret.\n    #
          Store it in a file called entitledregistryconfigjson.\n    oc create secret
          docker-registry entitled-registry --docker-server=registry.marketplace.redhat.com
          --docker-username \"cp\" --docker-password \"${PULL_SECRET}\" --dry-run=client
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > entitledregistryconfigjson\n
          \   # Get the current global secrets on the cluster and store it as a file
          named dockerconfigjson\n    oc get secret pull-secret -n openshift-config
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > dockerconfigjson\n
          \   # Merge the two dockerconfigs together into a file called dockerconfigjson-merged.\n
          \   jq -s '.[0] * .[1]' dockerconfigjson entitledregistryconfigjson > dockerconfigjson-merged\n
          \   # Set the cluster's dockerconfig file to the new merged version.\n    oc
          set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson=dockerconfigjson-merged\n
          \   ```\n  7. (Optional) Your IBM product may have installed IBM Metrics
          Operator using an `ibm-entitlement-key` for authentication to IBM Software
          Central. To authorize the automatic creation of an IBM Software Central
          account for the IBM account owning the `ibm-entitlement-key`\n      ```\n
          \     oc patch marketplaceconfig marketplaceconfig -n ibm-software-central
          --type='merge' -p '{\"spec\": {\"authorizeAccountCreation\": true}}'\n      ```\n\n###
          Why is a global pull secret required?\nIn order to successfully install
          the products hosted by the container image registry, you will need to make
          the pull secret available across the cluster. This can be achieved by applying
          the pull as a [global pull secret](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html#images-update-global-pull-secret_using-image-pull-secrets).
          For alternative approaches, please see the official OpenShift [documentation](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html).\n\n\n###
          SecurityContextConstraints requirements\n\nThe Operators and their components
          support running under the OpenShift Container Platform default restricted
          and restricted-v2 security context constraints.\n\n### Installation Namespace
          and ClusterRoleBinding requirements\n\nThe IBM Metrics Operator components
          require specific ClusterRoleBindings.\n- The metric-state component requires
          a ClusterRoleBinding for the the `view` ClusterRole.\n  - Ability to read
          non-sensitive CustomResources. \n  - The `view` ClusterRole dynamically
          adds new CustomResourceDefinitions.\n- The operator & reporter component
          requires a ClusterRoleBinding for the the `cluster-monitoring-view` ClusterRole.\n
          \ - Ability to view Prometheus metrics. \n  - The underlying ClusterRole
          RBAC often updated between OpenShift versions.\n\nDue to limitations of
          Operator Lifecycle Manager (OLM), this ClusterRoleBinding can not be provided
          dynamically for arbitrary installation target namespaces.\n\nA static ClusterRoleBinding
          is included for installation to the default namespace of `ibm-software-central`,
          and namespaces `openshift-redhat-marketplace`,  `redhat-marketplace`, `ibm-common-services`.\n\nTo
          create the ClusterRoleBindings for installation to an alternate namespace\n```\noc
          project INSTALL-NAMESPACE\noc adm policy add-cluster-role-to-user view -z
          ibm-metrics-operator-metric-state\noc adm policy add-cluster-role-to-user
          cluster-monitoring-view -z ibm-metrics-operator-controller-manager,ibm-metrics-operator-reporter\n```\n\n###
          Metric State scoping requirements\nThe metric-state Deployment obtains `get/list/watch`
          access to metered resources via the `view` ClusterRole. For operators deployed
          using Operator Lifecycle Manager (OLM), permissions are added to `clusterrole/view`
          dynamically via a generated and annotated `-view` ClusterRole. If you wish
          to meter an operator, and its Custom Resource Definitions (CRDs) are not
          deployed through OLM, one of two options are required\n1. Add the following
          label to a clusterrole that has get/list/watch access to your CRD: `rbac.authorization.k8s.io/aggregate-to-view:
          \"true\"`, thereby dynamically adding it to `clusterrole/view`\n2. Create
          a ClusterRole that has get/list/watch access to your CRD, and create a ClusterRoleBinding
          for the metric-state ServiceAccount\n\nAttempting to meter a resource with
          a MeterDefinition without the required permissions will log an `AccessDeniedError`
          in metric-state.\n\n### Disaster Recovery\n\nTo plan for disaster recovery,
          note the PhysicalVolumeClaims `rhm-data-service-rhm-data-service-N`. \n-
          In connected environments, MeterReport data upload attempts occur hourly,
          and are then removed from data-service. There is a low risk of losing much
          unreported data.\n- In an airgap environment, MeterReport data must be pulled
          from data-service and uploaded manually using [datactl](https://github.com/redhat-marketplace/datactl/).
          To prevent data loss in a disaster scenario, the data-service volumes should
          be considered in a recovery plan.\n\n### Subscription Config\n\nIt is possible
          to [configure](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md)
          how OLM deploys an Operator via the `config` field in the [Subscription](https://github.com/operator-framework/olm-book/blob/master/docs/subscriptions.md)
          object.\n\nThe IBM Metrics Operator will also read the `config` and append
          it to the operands. The primary use case is to control scheduling using
          [Tolerations](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/)
          and [NodeSelectors](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/).\n\nA
          limitation is that the `config` elements are only appended to the operands.
          The elements in the operands are not removed if the `config` is removed
          from the `Subscripion`. The operand must be modified manually, or deleted
          and recreated by the controller.\n\n### Cluster permission requirements\n\n|API
          group             |Resources              |Verbs                                     |\n|----------------------|-----------------------|------------------------------------------|\n|apps
          \                 |statefulsets           |get;list;watch                            |\n|authentication.k8s.io
          |tokenreviews           |create                                    |\n|authorization.k8s.io
          \ |subjectaccessreviews   |create                                    |\n|config.openshift.io
          \  |clusterversions        |get;list;watch                            |\n|marketplace.redhat.com|meterdefinitions
          \      |get;list;watch;create;update;patch;delete |\n|operators.coreos.com
          \ |clusterserviceversions |get;list;watch                            |\n|operators.coreos.com
          \ |subscriptions          |get;list;watch                            |\n|operators.coreos.com
          \ |operatorgroups         |get;list                                  |\n|storage.k8s.io
          \       |storageclasses         |get;list;watch                            |\n|
          \                     |namespaces             |get;list;watch                            |\n|
          \                     |configmaps             |get;list;watch                            |\n|
          \                     |services               |get;list;watch                            |\n|
          \                     |**clusterrole/view**   |                                          |\n\n\n###
          Documentation\nYou can find our documentation [here.](https://swc.saas.ibm.com/en-us/documentation/)\n\n###
          Getting help\nIf you encounter any issues while using the operators, you
          can create an issue on our [Github\nrepo](https://github.com/redhat-marketplace/redhat-marketplace-operator)
          for bugs, enhancements, or other requests. You can also visit our main page
          and\nreview our [support](https://swc.saas.ibm.com/en-us/support) and [documentation](https://swc.saas.ibm.com/en-us/documentation/).\n\n###
          Readme\nYou can find our readme [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/README.md)\n\n###
          License information\nYou can find our license information [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/LICENSE)\n"
        displayName: IBM Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        - buy
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/redhat-marketplace-metric-state@sha256:e4014c9a546c10e84b13a09db6a18431ee9489ad1296808635e48d9ef4963cf1
        - quay.io/rh-marketplace/redhat-marketplace-authcheck@sha256:13a81b8aca79ae8c89ee24dcc30f3322a4cfdedf929e7bb984295c1e7e631fa9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/redhat-marketplace-data-service@sha256:30fd8d2996b921cfce51f697ee4490a6d9a836021952429b877336f5edf2749d
        - quay.io/rh-marketplace/redhat-marketplace-reporter@sha256:2bd098f057ca07f2b14be9b60ea6ad8ca6b0cae1d226b1ca3b27e23d04ccdf20
        - quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
        version: 2.20.0
      entries:
      - name: ibm-metrics-operator.v2.20.0
        version: 2.20.0
      name: stable
    defaultChannel: stable
    packageName: ibm-metrics-operator
    provider:
      name: Red Hat Marketplace
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-dr-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-dr-manager.v8.0.220
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com",
                  "resRecord": {
                    "HostName_1": "100.200.30.41",
                    "HostName_2": "100.200.30.43",
                    "abc": "HostName_2",
                    "www": "HostName_1",
                    "xyz": "HostName_3"
                  },
                  "stealthMasters": [
                    "1.2.3.4"
                  ],
                  "tsigKeyFile": "/Kvcslnxdns.com.+157+10641.private"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "dataReplicationPointer": "test-datareplication",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 10,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          olm.skipRange: <8.0.220
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
            - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:a98f9cc51f2c7baeec6dbb4933a42ae8ae20ff1a42c9d48a37174cf5211bbcf8
        version: 8.0.220
      entries:
      - name: infoscale-dr-manager.v8.0.220
        version: 8.0.220
      - name: infoscale-dr-manager.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-dr-manager.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          createdAt: "2024-06-17T10:14:13Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:c46cbf78631fc790c77b6c13c1d1d4b5ab35f49164f7c2e36c7ec59220cbdce2
        version: 8.0.330
      entries:
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-dr-manager.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
          createdAt: "2024-02-22T14:14:42Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721648-168743109)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721654-168743109)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
        version: 8.0.410
      entries:
      - name: infoscale-dr-manager.v8.0.410
        version: 8.0.410
      - name: infoscale-dr-manager.v8.0.409
        version: 8.0.409
      - name: infoscale-dr-manager.v8.0.400
        version: 8.0.400
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-dr-manager.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
          createdAt: "2024-02-22T14:14:42Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721648-168743109)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721654-168743109)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
        version: 8.0.410
      entries:
      - name: infoscale-dr-manager.v8.0.410
        version: 8.0.410
      - name: infoscale-dr-manager.v8.0.409
        version: 8.0.409
      - name: infoscale-dr-manager.v8.0.400
        version: 8.0.400
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-dr-manager.v8.0.100
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com",
                  "resRecord": {
                    "abc": "r7515-054-vm9",
                    "r7515-054-vm10": "10.221.85.83",
                    "r7515-054-vm8": "10.221.85.81",
                    "r7515-054-vm9": "10.221.85.82",
                    "www": "r7515-054-vm8",
                    "xyz": "r7515-054-vm10"
                  },
                  "stealthMasters": [
                    "1.2.3.4"
                  ],
                  "tsigKeyFile": "/Kvcslnxdns.com.+157+10641.private"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "currentPrimary": "Clus1",
                  "force": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "fail",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "dataReplicationPointer": "test-datareplication",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 10,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:eb999187a1f9a9ce5566089e24adf598b1c229d31c5e57ddcfe4439ba25e97dc
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
            - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membershrip
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membershrip
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponsding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:eb999187a1f9a9ce5566089e24adf598b1c229d31c5e57ddcfe4439ba25e97dc
        version: 8.0.100
      entries:
      - name: infoscale-dr-manager.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-dr-manager
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-s3control-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-s3control-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/s3control-controller:1.1.1
          createdAt: "2025-09-19T20:52:02Z"
          description: AWS S3 Control controller is a service controller for managing
            S3 Control resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessPoint represents the state of an AWS s3control AccessPoint
              resource.
            displayName: AccessPoint
            kind: AccessPoint
            name: accesspoints.s3control.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon S3 Control resources in AWS from within your Kubernetes cluster.

          **About Amazon S3 Control**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon S3 Control
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - s3control
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon S3 Control Developer Resources
          url: https://aws.amazon.com/S3 Control/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: s3control maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/s3control-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-s3control-controller.v1.1.1
        version: 1.1.1
      - name: ack-s3control-controller.v1.1.0
        version: 1.1.0
      - name: ack-s3control-controller.v1.0.11
        version: 1.0.11
      - name: ack-s3control-controller.v1.0.10
        version: 1.0.10
      name: alpha
    defaultChannel: alpha
    packageName: ack-s3control-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: sosivio
      provider-url: https://www.sosiv.io
    name: sosivio
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sosivio-operator.v1.5.1-97
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sosivio.sosiv.io/v1alpha1",
                "kind": "Sosivio",
                "metadata": {
                  "name": "sosivio"
                },
                "spec": {
                  "platform": "ocp4",
                  "useSha": false
                }
              },
              {
                "apiVersion": "sosivio.sosiv.io/v1alpha1",
                "kind": "Sosivio",
                "metadata": {
                  "name": "sosivio"
                },
                "spec": {
                  "platform": "ocp4",
                  "useSha": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Monitoring,Integration & Delivery,Logging & Tracing,Developer
            Tools
          com.redhat.openshift.versions: v4.1-v4.1
          containerImage: quay.io/sosivio/sosivio-operator:v1.5.1-97
          createdAt: "2021-10-28 01:01:01"
          description: Sosivio | Predictive Troubleshooting for Kubernetes
          operatorhub.io/ui-metadata-max-k8s-version: "1.25"
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Sosivio/operator
          support: Sosivio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Sosivio is the Schema for the sosivios API
            displayName: Sosivio
            kind: Sosivio
            name: sosivios.sosivio.sosiv.io
            version: v1alpha1
        description: "# Sosivio Operator\nThe Sosivio Operator for RedHat’s Openshift
          Container Platform allows users to quickly and easily deploy Sosivio’s Kubernetes
          Predictive Troubleshooting Platform on your cluster.\n\n# Sosivio Overview\n\nSosivio
          is the first and only non-intrusive and predictive autonomous cloud-native
          platform built specifically for Kubernetes from the ground up. Sosivio provides
          predictive troubleshooting tools, root cause determination and instant remediation
          for any failures in Kubernetes environments.\n\nPredict and Prevent!\n\nThe
          Operator will deploy and manage Sosivio on Kubernetes and OpenShift. Included
          are the following features;\n-   **Root Cause Analysis and Remediation**
          - predictive troubleshooting to prevent failures before they occur. Instantly
          understand where the root cause is and how to fix it.\n-   **Application
          Profiling** - analyze container resource consumption and recommendations
          for requests and limits.\n-   **Real-time Metrics** - in-depth, granular,
          and accurate real-time data from the process level inside containers (and
          pods) with failure correlation.\n-   **Cluster Health Checks** - proactive
          health checks for cluster infrastructure and applications running on the
          cluster.\n-   **Multi-Cluster View** - management, monitoring, and maintenance
          capabilities for all clusters from one location. \n-   **Kubernetes Components
          Alerting** - notifications sent through Slack, email and other 3rd party
          platforms.\n\n# Installation\nSosivio is completely **non-intrusive** and
          **does not require any persistent storage**. There is **no data offloading**
          outside of your Kubernetes cluster. Applying our Data Swirling methodology
          enables the product to perform at a **remarkably low resource footprint**
          and requires **no persistent volume configurations**.\n\n\nSosivio is installed
          under one namespace (labeled: \"sosivio\") which creates all the required
          components for the product.\n\nAll users will start with 4 weeks of Sosivio
          Premium after which Sosivio will automatically convert to the Community
          Version (free forever).\n\n# Help\nMore information about Sosivio and our
          product usage instructions can be found in our documentation [here](https://docs.sosiv.io/)
          \ \n"
        displayName: Sosivio | Predictive Troubleshooting for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sosivio
        - kubernetes
        - observanility
        - monitoring
        - root cause
        links:
        - name: Sosivio
          url: https://www.sosiv.io
        - name: Documentation
          url: https://docs.sosiv.io
        maintainers:
        - email: info@sosiv.io
          name: Sosivio
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: sosivio
          url: https://www.sosiv.io
        relatedImages:
        - quay.io/sosivio/discovery-engine@sha256:b87b3f7c9ed3aed80d0a77788fd8925789450d9dc05d25017e89dff611cd82da
        - quay.io/sosivio/nsq@sha256:781aac23ef0a1ed0af2b8e93d625ae49bd11dbdd98787d3c79dde8d128f3810d
        - quay.io/sosivio/actuator@sha256:a9abe9730800ca2b0a1bbd7a361b2ab3a818cbf127798314e5477fb1cb16ba4c
        - quay.io/sosivio/sosivio-node-collector@sha256:61ed8dd52de5edc09d16a31b18f2f65ad9344c0b6c9cfab323c3eb3d7e7ed060
        - quay.io/sosivio/crud-manager@sha256:38289a28829fa737917d76dd4810ba6140e123061ead0a4b79fa79b2e246a293
        - quay.io/sosivio/sosivio-operator:v1.5.1-97
        - quay.io/sosivio/sosivio-dashboard@sha256:fdfca5bb2123d1008cec2b656cde73c7e9364f84c5ff4f4a6fee5751b1bfb2f8
        - quay.io/sosivio/sosivio-node-netpacket@sha256:84dd159b1dfdb52dee3eb1c4c93f323bae49ce517ff44663b69978e14bdd8390
        - quay.io/sosivio/classifier@sha256:dd8bf145524db23ff333efa1bcdff258dab116d3534f95f73775d8dd339e6963
        - quay.io/sosivio/communicator@sha256:fffec9a6c66594fde5f05e2dfb84361544ca9d53a345188c197a0ea6f53bc4b9
        - quay.io/sosivio/authentication@sha256:0ea6025f333576792f5b1fc4d96d9ab1a404749f185830410e71248683004da2
        - quay.io/sosivio/sosiviodb@sha256:a3c6c3975bc2b0316135ee5c547f1eb215f3a1f9c8b0a9c8eceff23415f93597
        - quay.io/sosivio/kube-rbac-proxy:v0.8.0
        - quay.io/sosivio/sequence-recognition@sha256:7335cf8f3e40bc46e8ca0de525d99f0655833b7c52d36dc273aa7b096eb70fa2
        - quay.io/sosivio/network-aggregator@sha256:e59d937f4d9c52acad0adadab39c6849d20b4ee72a83e239a33ea137f08323f4
        - quay.io/sosivio/correlation-entities@sha256:c00af09b2589bad4740fd44d7a81c69b2e29138c360939b2a45f65edd44b6901
        - quay.io/sosivio/contract-testing@sha256:cad6fac6ed3fdf2597ceb09ba5dbb7d81f70cf45f483a7064e3968215dd982e3
        - quay.io/sosivio/kube-rbac-proxy@sha256:34e8724e0f47e31eb2ec3279ac398b657db5f60f167426ee73138e2e84af6486
        - quay.io/sosivio/analyzer@sha256:3825608fe7c43ef059108e76665acc99fe72bea77d4d9e3852f6ad45e74379b1
        - quay.io/sosivio/notifier@sha256:35f5f3e8e5ccad914fcfe34fb4b4a1b044ee87f5a6141284cd5273f0b2b3bf22
        - quay.io/sosivio/sosivio-node-pinger@sha256:937de52ad7b76e6cff4ad5197de9705988d61f285ebadac5ebadb5856f59ee58
        - quay.io/sosivio/sosivio-operator@sha256:11bc627773bb5e08b629e0b013e7f600fcb4aaefbd97f2de337d339475616763
        version: 1.5.1
      entries:
      - name: sosivio-operator.v1.5.1-97
        version: 1.5.1
      - name: sosivio-operator.v1.5.0-95
        version: 1.5.0
      - name: sosivio-operator.v1.4.4-85
        version: 1.4.4
      - name: sosivio-operator.v1.4.3-83
        version: 1.4.3
      - name: sosivio-operator.v1.4.2-82
        version: 1.4.2
      - name: sosivio-operator.v1.4.1-81
        version: 1.4.1-81
      - name: sosivio-operator.v1.4.1.1-80
        version: 1.4.1-80
      - name: sosivio-operator.v1.4.1-78
        version: 1.4.1-79
      - name: sosivio-operator.v1.3.1-76
        version: 1.3.1-76
      - name: sosivio-operator.v1.3.0-75
        version: 1.3.0-75
      - name: sosivio-operator.v1.2.1-1
        version: 1.2.1-1
      - name: sosivio-operator.v1.2.0-1
        version: 1.2.0-1
      name: stable
    defaultChannel: stable
    packageName: sosivio
    provider:
      name: sosivio
      url: https://www.sosiv.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-keycloak-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-keycloak-operator.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2023-09-19T12:43:29Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KCOSClientRoleMapping
            name: kcosclientrolemappings.keycloak.org
            version: v1alpha1
          - kind: KCOSRealmRole
            name: kcosrealmroles.keycloak.org
            version: v1alpha1
          - kind: KCOSUserRoleMapping
            name: kcosuserrolemappings.keycloak.org
            version: v1alpha1
          - kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: Keysight Keycloak Operator
        displayName: Keysight Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Keycloak Operator
          url: https://keysight-keycloak-operator.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak@sha256:772f73cce6a77b9c46f8035faac99a28602219657d18b75b19b4d905d91942c9
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak-postgresql@sha256:b636955199065e18f9d4679835ae78214843258e5abb4ae758288d6f00df2f8e
        - prod.harbor.keysight.digital/loadcore/keysight-keycloak-operator@sha256:d74e7e98a114768a6f3327edabce9336743a078844fffee985053b6a90100da8
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak-init-container@sha256:90965beaf41c760cb8c8bf906acf08f8fe4770d313de8dfd92d18d7c25348a38
        version: 0.1.21
      entries:
      - name: keysight-keycloak-operator.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-keycloak-operator
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: fence-agents-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fence-agents-remediation.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  }
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:00cc420250d8e25d9da7c43d9c4cff35c7edeb7d390af754c48c0ce00e9e7770
          createdAt: "2023-10-30 14:06:57"
          description: Fence Agents Remediation Operator for remediating nodes using
            upstream fence-agents.
          olm.skipRange: '>=0.2.0 <0.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: "Fence Agents Remediation Operator (FAR)\n\nThe operator will
          remediate/fence a node when a FenceAgentsRemediation (far) custom resource
          exists with the node's name.\nFAR uses an existing set of upstream fencing
          agents for environments with a \ntraditional API end-point (e.g., IPMI)
          for power cycling cluster nodes.\n\nThe operator is recommended to be used
          with Node Healthcheck Operator for easier and smoother experience,\nbut
          it can also work as a standalonde remediator for the more advanced user.\n"
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:00cc420250d8e25d9da7c43d9c4cff35c7edeb7d390af754c48c0ce00e9e7770
        version: 0.2.1
      entries:
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.14-eus
    - currentCSV: fence-agents-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:a367e8fdd80938e97ac195631a996e0c537bf3666aef0199147534762dcf739f
          createdAt: "2024-08-10 22:49:23"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.2.0 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:a367e8fdd80938e97ac195631a996e0c537bf3666aef0199147534762dcf739f
        version: 0.4.1
      entries:
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.16-eus
    - currentCSV: fence-agents-remediation.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-16 10:25:57"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.5.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.5/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel9-operator@sha256:da5975557078fbfa6882df87cff5bda5671b9838b5fb0a738fea724469d03576
        version: 0.5.1
      entries:
      - name: fence-agents-remediation.v0.5.1
        version: 0.5.1
      - name: fence-agents-remediation.v0.5.0
        version: 0.5.0
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.18-eus
    - currentCSV: fence-agents-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  }
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:5c3778c7aae45f54e949fbae3d133777f7aa78fe1251ac3d5510aa28b71e2ba4
          createdAt: "2023-04-20 15:36:05"
          description: Fence Agents Remediation Operator for remediating nodes using
            upstream fence-agents.
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: "Fence Agents Remediation Operator (FAR)\n\nThe operator will
          remediate/fence a node when a FenceAgentsRemediation (far) custom resource
          exists with the node's name.\nFAR uses an existing set of upstream fencing
          agents for environments with a \ntraditional API end-point (e.g., IPMI)
          for power cycling cluster nodes.\n\nThe operator is recommended to be used
          with Node Healthcheck Operator for easier and smoother experience,\nbut
          it can also work as a standalonde remediator for the more advanced user.\n"
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3af07bfe0fb73b1a7f00487eba302d6da442e742ab4007fc50e52efd4d45dae5
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:5c3778c7aae45f54e949fbae3d133777f7aa78fe1251ac3d5510aa28b71e2ba4
        version: 0.1.0
      entries:
      - name: fence-agents-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: fence-agents-remediation.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeSecretNames": {
                    "master-0": "fence-agents-credentials-master0",
                    "master-1": "fence-agents-credentials-master1",
                    "master-2": "fence-agents-credentials-master2",
                    "worker-0": "fence-agents-credentials-worker0",
                    "worker-1": "fence-agents-credentials-worker1",
                    "worker-2": "fence-agents-credentials-worker2"
                  },
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedSecretName": "fence-agents-credentials-shared",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {
                    "spec": {
                      "agent": "fence_ipmilan",
                      "nodeSecretNames": {
                        "master-0": "fence-agents-credentials-master0",
                        "master-1": "fence-agents-credentials-master1",
                        "master-2": "fence-agents-credentials-master2",
                        "worker-0": "fence-agents-credentials-worker0",
                        "worker-1": "fence-agents-credentials-worker1",
                        "worker-2": "fence-agents-credentials-worker2"
                      },
                      "nodeparameters": {
                        "--ipport": {
                          "master-0": "6230",
                          "master-1": "6231",
                          "master-2": "6232",
                          "worker-0": "6233",
                          "worker-1": "6234",
                          "worker-2": "6235"
                        }
                      },
                      "remediationStrategy": "ResourceDeletion",
                      "retrycount": 5,
                      "retryinterval": "5s",
                      "sharedSecretName": "fence-agents-credentials-shared",
                      "sharedparameters": {
                        "--action": "reboot",
                        "--ip": "192.168.111.1",
                        "--lanplus": "",
                        "--username": "admin"
                      },
                      "timeout": "60s"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-09-08 09:19:00"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.5.0 <0.6.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.8/html/remediation_fencing_and_maintenance/about-remediation-fencing-maintenance
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel9-operator@sha256:eef1a298718650a22a3da4a07c140ce21b6f287d87af79801610317fa8f63d52
        version: 0.6.0
      entries:
      - name: fence-agents-remediation.v0.6.0
        version: 0.6.0
      - name: fence-agents-remediation.v0.5.1
        version: 0.5.1
      - name: fence-agents-remediation.v0.5.0
        version: 0.5.0
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: stable
    defaultChannel: stable
    packageName: fence-agents-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ssm-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ssm-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ssm-controller:1.1.1
          createdAt: "2025-09-19T20:57:07Z"
          description: AWS SSM controller is a service controller for managing SSM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Document represents the state of an AWS ssm Document resource.
            displayName: Document
            kind: Document
            name: documents.ssm.services.k8s.aws
            version: v1alpha1
          - description: PatchBaseline represents the state of an AWS ssm PatchBaseline
              resource.
            displayName: PatchBaseline
            kind: PatchBaseline
            name: patchbaselines.ssm.services.k8s.aws
            version: v1alpha1
          - description: ResourceDataSync represents the state of an AWS ssm ResourceDataSync
              resource.
            displayName: ResourceDataSync
            kind: ResourceDataSync
            name: resourcedatasyncs.ssm.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon SSM resources in AWS from within your Kubernetes cluster.

          **About Amazon SSM**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon SSM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ssm
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SSM Developer Resources
          url: https://aws.amazon.com/SSM/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ssm maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ssm-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ssm-controller.v1.1.1
        version: 1.1.1
      - name: ack-ssm-controller.v1.1.0
        version: 1.1.0
      - name: ack-ssm-controller.v1.0.12
        version: 1.0.12
      - name: ack-ssm-controller.v1.0.10
        version: 1.0.10
      - name: ack-ssm-controller.v1.0.9
        version: 1.0.9
      - name: ack-ssm-controller.v1.0.8
        version: 1.0.8
      - name: ack-ssm-controller.v1.0.7
        version: 1.0.7
      - name: ack-ssm-controller.v1.0.6
        version: 1.0.6
      - name: ack-ssm-controller.v1.0.5
        version: 1.0.5
      - name: ack-ssm-controller.v1.0.4
        version: 1.0.4
      - name: ack-ssm-controller.v1.0.3
        version: 1.0.3
      - name: ack-ssm-controller.v1.0.2
        version: 1.0.2
      - name: ack-ssm-controller.v1.0.1
        version: 1.0.1
      - name: ack-ssm-controller.v1.0.0
        version: 1.0.0
      - name: ack-ssm-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-ssm-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-efs-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-efs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "AccessPoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "FileSystem",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "MountTarget",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/efs-controller:1.1.1
          createdAt: "2025-09-19T18:38:13Z"
          description: AWS EFS controller is a service controller for managing EFS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessPoint represents the state of an AWS efs AccessPoint
              resource.
            displayName: AccessPoint
            kind: AccessPoint
            name: accesspoints.efs.services.k8s.aws
            version: v1alpha1
          - description: FileSystem represents the state of an AWS efs FileSystem
              resource.
            displayName: FileSystem
            kind: FileSystem
            name: filesystems.efs.services.k8s.aws
            version: v1alpha1
          - description: MountTarget represents the state of an AWS efs MountTarget
              resource.
            displayName: MountTarget
            kind: MountTarget
            name: mounttargets.efs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EFS resources in AWS from within your Kubernetes cluster.

          **About Amazon Elastic File System (Amazon EFS)**

          Amazon Elastic File System (Amazon EFS) provides serverless, fully elastic file storage so that you can share file data without provisioning or managing storage capacity and performance. Amazon EFS is built to scale on demand to petabytes without disrupting applications, growing and shrinking automatically as you add and remove files. Because Amazon EFS has a simple web services interface, you can create and configure file systems quickly and easily. The service manages all the file storage infrastructure for you, meaning that you can avoid the complexity of deploying, patching, and maintaining complex file system configurations.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Elastic File System (Amazon
          EFS)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - efs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EFS Developer Resources
          url: https://aws.amazon.com/efs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: efs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/efs-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-efs-controller.v1.1.1
        version: 1.1.1
      - name: ack-efs-controller.v1.1.0
        version: 1.1.0
      - name: ack-efs-controller.v1.0.19
        version: 1.0.19
      - name: ack-efs-controller.v1.0.18
        version: 1.0.18
      - name: ack-efs-controller.v1.0.17
        version: 1.0.17
      - name: ack-efs-controller.v1.0.16
        version: 1.0.16
      - name: ack-efs-controller.v1.0.15
        version: 1.0.15
      - name: ack-efs-controller.v1.0.14
        version: 1.0.14
      - name: ack-efs-controller.v1.0.13
        version: 1.0.13
      - name: ack-efs-controller.v1.0.12
        version: 1.0.12
      - name: ack-efs-controller.v1.0.11
        version: 1.0.11
      - name: ack-efs-controller.v1.0.10
        version: 1.0.10
      - name: ack-efs-controller.v1.0.8
        version: 1.0.8
      - name: ack-efs-controller.v1.0.6
        version: 1.0.6
      - name: ack-efs-controller.v1.0.5
        version: 1.0.5
      - name: ack-efs-controller.v1.0.3
        version: 1.0.3
      - name: ack-efs-controller.v1.0.2
        version: 1.0.2
      - name: ack-efs-controller.v1.0.0
        version: 1.0.0
      - name: ack-efs-controller.v0.0.9
        version: 0.0.9
      - name: ack-efs-controller.v0.0.8
        version: 0.0.8
      - name: ack-efs-controller.v0.0.7
        version: 0.0.7
      - name: ack-efs-controller.v0.0.6
        version: 0.0.6
      - name: ack-efs-controller.v0.0.5
        version: 0.0.5
      - name: ack-efs-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-efs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solace Corporation
      provider-url: www.solace.com
    name: pubsubplus-eventbroker-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pubsubplus-eventbroker-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pubsubplus.solace.com/v1beta1",
                "kind": "PubSubPlusEventBroker",
                "metadata": {
                  "name": "non-ha-standalone-example"
                },
                "spec": {
                  "redundancy": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "true"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.10
          containerImage: registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator:1.4.0
          createdAt: "2025-07-31T19:48:36Z"
          description: The Solace PubSub+ Event Broker Operator deploys and manages
            the lifecycle of PubSub+ Event Brokers
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
          support: Solace Products
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PubSub+ Event Broker
            displayName: PubSub+ Event Broker
            kind: PubSubPlusEventBroker
            name: pubsubpluseventbrokers.pubsubplus.solace.com
            version: v1beta1
        description: |
          ## Solace PubSub+ Platform
          Solace [PubSub+ Platform](https://solace.com/products/platform/) is a complete event streaming and management platform
          for real-time enterprises. The [PubSub+ Software Event Broker](https://solace.com/products/event-broker/software/) efficiently streams
          event-driven information between applications, IoT devices, and user interfaces running in the cloud, on-premises,
          and in hybrid environments using open APIs and protocols like AMQP, JMS, MQTT, REST and WebSocket.
          It can be installed into a variety of public and private clouds, PaaS, and on-premises environments.
          Event brokers in multiple locations can be linked together in an [Event Mesh](https://solace.com/what-is-an-event-mesh/)
          to dynamically share events across the distributed enterprise.

          ## Solace PubSub+ Event Broker Operator
          Solace Pubsub+ Event Broker Operator automatically deploys and manages PubSub+ Software Event Brokers on Kubernetes and OpenShift environments.

          Features include:
          * Initial deployment configuration: Production-ready HA, non-HA or for developers
          * Rolling update for configuration changes or broker upgrades
          * Prometheus monitoring support

          ## Getting started
          Refer to the [Solace PubSub+ Event Broker Operator Quick Start](https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/README.md)
        displayName: Solace PubSub+ Event Broker Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - solace
        - pubsubplus
        - pubsub+
        - pubsub
        - messaging
        - advanced event broker
        - event broker
        - event mesh
        - event streaming
        - data streaming
        - event integration
        - middleware
        links:
        - name: Operator QuickStart
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
        - name: Operator Documentation
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/docs/EventBrokerOperatorUserGuide.md
        - name: Solace.com
          url: https://solace.com
        - name: Understanding Solace Technology
          url: https://solace.com/products/platform/
        - name: Solace PubSub+ Technical Documentation
          url: https://docs.solace.com/
        - name: Solace Developer Portal
          url: https://www.solace.dev/
        - name: Solace Community
          url: https://solace.community/
        maintainers:
        - email: support@solace.com
          name: Solace R&D
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Solace Corporation
          url: www.solace.com
        relatedImages:
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator:1.4.0
        version: 1.4.0
      entries:
      - name: pubsubplus-eventbroker-operator.v1.4.0
        version: 1.4.0
      - name: pubsubplus-eventbroker-operator.v1.3.0
        version: 1.3.0
      - name: pubsubplus-eventbroker-operator.v1.2.0
        version: 1.2.0
      - name: pubsubplus-eventbroker-operator.v1.1.2
        version: 1.1.2
      - name: pubsubplus-eventbroker-operator.v1.1.1
        version: 1.1.1
      - name: pubsubplus-eventbroker-operator.v1.1.0
        version: 1.1.0
      - name: pubsubplus-eventbroker-operator.v1.0.3
        version: 1.0.3
      - name: pubsubplus-eventbroker-operator.v1.0.2
        version: 1.0.2
      - name: pubsubplus-eventbroker-operator.v1.0.1
        version: 1.0.1
      - name: pubsubplus-eventbroker-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pubsubplus-eventbroker-operator
    provider:
      name: Solace Corporation
      url: www.solace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: ibm-data-reporter-operator
      app.kubernetes.io/managed-by: ibm-data-reporter-operator
      app.kubernetes.io/name: ibm-data-reporter-operator
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: ""
    name: ibm-data-reporter-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-data-reporter-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "DataReporterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-data-reporter-operator",
                    "app.kubernetes.io/instance": "datareporterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datareporterconfig",
                    "app.kubernetes.io/part-of": "ibm-data-reporter-operator"
                  },
                  "name": "datareporterconfig"
                },
                "spec": {
                  "userConfig": [
                    {
                      "metadata": {
                        "ameta1": "ametadata1",
                        "bmeta1": "bmetadata1",
                        "cmeta1": "cmetadata1",
                        "dmeta1": "dmetadata1"
                      },
                      "userName": "system:serviceaccount:redhat-marketplace:ibm-data-reporter-operator-api"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
          createdAt: ""
          description: The IBM Data Reporter Operator accepts events and transforms
            them into reports submitted to the Data Service of the IBM Metrics Operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReporterConfig is the Schema for the datareporterconfigs
              API
            displayName: Data Reporter Config
            kind: DataReporterConfig
            name: datareporterconfigs.marketplace.redhat.com
            version: v1alpha1
          required:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
        description: "# IBM Data Reporter Operator\n\n# Introduction\n\nThe IBM Data
          Reporter Operator accepts events and transforms them into reports submitted
          to the IBM Metrics Operator Data Service.\n\n# Details\n\nThe IBM Data Reporter
          Operator deploys a service that provides an endpoint to which callers can
          send raw json event data. The event data is transformed into a report and
          is sent to the IBM Metrics Operator's Data Service. The IBM Metrics Operator
          Data Service periodically uploads the reports to IBM Software Central. It
          can also filter, transform and forward events to alternate endpoints.\n\n##
          Prerequisites\n\n- OpenShift Container Platform, major version 4 with any
          available supported minor version\n- Install IBM Metrics Operator\n  - IBM
          Data Reporter Operator prerequisties the IBM Metrics Operator's Data Service\n
          \ - Register the Cluster by creating a `redhat-marketplace-pull-secret`,
          as per the instructions\n  - Once `rhm-data-service` has started IBM Data
          Reporter Operator will accept requests\n\n## SecurityContextConstraints
          Requirements\n\n- The operator runs under Red Hat restricted SCC\n\n## Resources
          Required\n\n- The operator requires at least 85 Mi memory.\n\n## Limitations\n\n-
          Only runs on amd64, s390x and ppc64le architectures\n\n## Installing\n\n-
          A user with the cluster administrator role.\n- Install this operator in
          the same namespace as the IBM Metrics Operator\n  - default namespace: `ibm-software-central`\n\n##
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n# Configuration\n\nOptional:\n- Configure the
          Custom Resource Definition `DataReporterConfig` named `datareporterconfig`
          as per the following sample\n  - Configuration that can not be reconciled
          successfully will be reported on the DataReporterConfig Status\n  - jsonPath
          expressions are handled by: https://github.com/ohler55/ojg\n    - jsonPath
          comparison: https://cburgmer.github.io/json-path-comparison/\n  - Transformer
          type currently supported is `kazaam`\n    - https://github.com/willie68/kazaam\n\nSample
          DataReporterConfig:\n```YAML\napiVersion: marketplace.redhat.com/v1alpha1\nkind:
          DataReporterConfig\nmetadata:\n  name: datareporterconfig\nspec:\n  # true:
          report generated for 1 event and provide return code from data-service,
          confirming delivery\n  # false: events will be accumulated before building
          report, and immediately return 200, delivery unfconfirmed\n  confirmDelivery:
          false\n  dataFilters:   # dataFilters matches the first dataFilter selector,
          applies transforms and forwards to data-service and destinations\n  - altDestinations:
          \  # List of desintations to transform and foward events to\n    - authorization:
          \  # optional endpoint if an authorization token must first be requested
          \n        authDestHeader: Authorization   # http header key to be appended
          to destination request\n        authDestHeaderPrefix: 'Bearer '   # optional
          prefix for constructing destination request http header value\n        bodyData:
          \  # body data to send to authorization request, often an apiKey\n          secretKeyRef:\n
          \           name: auth-body-data-secret\n            key: bodydata\n        header:
          \  # secret map to use for authorization request headers\n          secret:\n
          \           name: auth-header-map-secret\n        tokenExpr: $.token   #
          optionally extract the returned body data using a jsonPath expression\n
          \       url: https://127.0.0.1:8443/api/2.0/accounts/my_account_id/apikeys/token
          \  # the url to make an authorization request to\n      header:   # secret
          map to use for destination request headers\n        secret:\n          name:
          dest-header-map-secret\n      transformer:   # The transformer to apply
          to event data for this destination\n        configMapKeyRef:\n          key:
          kazaam.json\n          name: kazaam-configmap\n        type: kazaam\n      url:
          https://127.0.0.1:8443/metering/api/1.0/usage   # the url to send the event
          to\n      urlSuffixExpr: $.properties.productId   # optional jsonPath expression
          on the event to extract and set a suffix to the url path\n    manifestType:
          dataReporter   # override manifestType of report sent to data-service\n
          \   selector:   # Matches against event data and a user. Empty selector
          matches all.\n      matchExpressions:   # jsonPath expression must return
          a result for all expressions to match (AND)\n      - $[?($.event == \"Account
          Contractual Usage\")]\n      - $.properties.productId\n      - $[?($.properties.source
          != null)]\n      - $[?($.properties.unit == \"AppPoints\")]\n      - $[?($.properties.quantity
          >= 0)]\n      - $[?($.timestamp != null)]\n      matchUsers:   # must match
          one of these users (OR). Omitting users matches any user\n      - system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n
          \   transformer:   # The transformer to apply to the event for data-service\n
          \     configMapKeyRef:\n        key: kazaam.json\n        name: kazaam-configmap\n
          \     type: kazaam\n  tlsConfig:   # TLS configuration for requests outbound
          to destinations\n    caCerts:\n    - name: tls-ca-certificates\n      key:
          ca.crt\n    certificates:\n    - clientCert:\n        secretKeyRef:\n          name:
          tls-ca-certificates\n          key: tls.crt\n      clientKey:\n        secretKeyRef:\n
          \         name: tls-ca-certificates\n          key: tls.key\n    cipherSuites:
          \n    - TLS_AES_128_GCM_SHA256\n    - TLS_AES_256_GCM_SHA384\n    - TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\n    - TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\n    insecureSkipVerify: false\n
          \   minVersion: VersionTLS12\n  userConfig:   # Optional metadata to apply
          to report sent to data-service for a specific user\n  - metadata:\n      ameta1:
          ametadata1\n      bmeta1: bmetadata1\n      cmeta1: cmetadata1\n      dmeta1:
          dmetadata1\n    userName: system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n```\n\nSample
          Header Secrets:\n```YAML\napiVersion: v1\nstringData:\n  accept: | \n    */*\nkind:
          Secret\nmetadata:\n  name: dest-header-map-secret\ntype: Opaque\n---\napiVersion:
          v1\nstringData:\n  accept: application/json\n  Content-Type: application/json\nkind:
          Secret\nmetadata:\n  name: auth-header-map-secret\ntype: Opaque\n```\n\nSample
          Authentication Body Secret:\n```YAML\napiVersion: v1\nstringData:\n  bodydata:
          |\n    {\"apikey\": \"myapikey\"}\nkind: Secret\nmetadata:\n  name: auth-body-data-secret\ntype:
          Opaque\n```\n\nSample TLS Secret:\n```YAML\napiVersion: v1\ndata:\n  ca.crt:
          LS0t...\n  tls.key: LS0t...\n  tls.crt: LS0t...\nkind: Secret\nmetadata:\n
          \ name: tls-ca-certificates\n```\n\nSample Kazaam ConfigMap:\n```YAML\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: kazaam-configmap\ndata:\n  kazaam.json:
          |\n    [\n        {\n            \"operation\": \"timestamp\",\n            \"spec\":
          {\n                \"timestamp\": {\n                \"inputFormat\": \"2006-01-02T15:04:05.999999+00:00\",\n
          \               \"outputFormat\": \"$unixmilli\"\n                }\n            }\n
          \       },\n        {\n            \"operation\": \"shift\",\n            \"spec\":
          {\n                \"instances[0].instanceId\": \"properties.source\",\n
          \               \"instances[0].startTime\": \"timestamp\",\n                \"instances[0].endTime\":
          \"timestamp\",\n                \"instances[0].metricUsage[0].metricId\":
          \"properties.unit\",\n                \"instances[0].metricUsage[0].quantity\":
          \"properties.quantity\"\n            }\n        },\n        {\n            \"operation\":
          \"default\",\n            \"spec\": {\n                \"meteringModel\":
          \"point-in-time\",\n                \"meteringPlan\": \"contract\"\n            }\n
          \       }\n    ]\n```\n\n### API Service User Configuration\n\n- The ClusterRole
          for api access is `clusterrole/ibm-data-reporter-operator-api`\n- The default
          ServiceAccount provided as an api user is `system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api`\n
          \ - The default ClusterRoleBinding for this user is `clusterrolebinding/ibm-data-reporter-operator-api`\n\nOptional:\n\n-
          To create an additional ServiceAccount and ClusterRoleBinding for api access\n\n```SHELL\nNAMESPACE=$(oc
          config view --minify -o jsonpath='{..namespace}') && \\\noc create serviceaccount
          my-api-service-account && \\\noc create clusterrolebinding ibm-data-reporter-operator-api
          --clusterrole=ibm-data-reporter-operator-api --serviceaccount=${NAMESPACE}:my-api-service-account\n```\n\n-
          Update datareporterconfig to attach metadata to reports associated with
          this user\n\n## API Service Usage\n\n- Get Token & Host\n\n```SHELL\noc
          project ibm-software-central\nexport DRTOKEN=$(oc create token ibm-data-reporter-operator-api
          --namespace ibm-software-central --duration 1h)\nexport DRHOST=$(oc get
          route ibm-data-reporter --template='{{ .spec.host }}')\n```\n\n- Get the
          Status\n\n```SHELL\ncurl -k -H \"Authorization: Bearer ${DRTOKEN}\" https://${DRHOST}/v1/status
          \n```\n\n- Post an Event\n\n```SHELL\ncurl -k -H \"Authorization: Bearer
          ${DRTOKEN}\" -X POST -d '{\"event\":\"myevent\"}' https://${DRHOST}/v1/event\n```\n\n##
          Storage\n\n- The operator temporarily stores events in pod memory, and writes
          to the IBM Metrics Operator data-service, which requires a PersistentVolume\n\n##
          License\n\nCopyright IBM Corporation 2023. All Rights Reserved.\n\nLicensed
          under the Apache License, Version 2.0 (the \"License\");\nyou may not use
          this file except in compliance with the License.\nYou may obtain a copy
          of the License at\n\n    http://www.apache.org/licenses/LICENSE-2.0\n\nUnless
          required by applicable law or agreed to in writing, software\ndistributed
          under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES
          OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for
          the specific language governing permissions and\nlimitations under the License.\n"
        displayName: IBM Data Reporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 2.20.0
      entries:
      - name: ibm-data-reporter-operator.v2.20.0
        version: 2.20.0
      name: beta
    - currentCSV: ibm-data-reporter-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "DataReporterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-data-reporter-operator",
                    "app.kubernetes.io/instance": "datareporterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datareporterconfig",
                    "app.kubernetes.io/part-of": "ibm-data-reporter-operator"
                  },
                  "name": "datareporterconfig"
                },
                "spec": {
                  "userConfig": [
                    {
                      "metadata": {
                        "ameta1": "ametadata1",
                        "bmeta1": "bmetadata1",
                        "cmeta1": "cmetadata1",
                        "dmeta1": "dmetadata1"
                      },
                      "userName": "system:serviceaccount:redhat-marketplace:ibm-data-reporter-operator-api"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
          createdAt: ""
          description: The IBM Data Reporter Operator accepts events and transforms
            them into reports submitted to the Data Service of the IBM Metrics Operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReporterConfig is the Schema for the datareporterconfigs
              API
            displayName: Data Reporter Config
            kind: DataReporterConfig
            name: datareporterconfigs.marketplace.redhat.com
            version: v1alpha1
          required:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
        description: "# IBM Data Reporter Operator\n\n# Introduction\n\nThe IBM Data
          Reporter Operator accepts events and transforms them into reports submitted
          to the IBM Metrics Operator Data Service.\n\n# Details\n\nThe IBM Data Reporter
          Operator deploys a service that provides an endpoint to which callers can
          send raw json event data. The event data is transformed into a report and
          is sent to the IBM Metrics Operator's Data Service. The IBM Metrics Operator
          Data Service periodically uploads the reports to IBM Software Central. It
          can also filter, transform and forward events to alternate endpoints.\n\n##
          Prerequisites\n\n- OpenShift Container Platform, major version 4 with any
          available supported minor version\n- Install IBM Metrics Operator\n  - IBM
          Data Reporter Operator prerequisties the IBM Metrics Operator's Data Service\n
          \ - Register the Cluster by creating a `redhat-marketplace-pull-secret`,
          as per the instructions\n  - Once `rhm-data-service` has started IBM Data
          Reporter Operator will accept requests\n\n## SecurityContextConstraints
          Requirements\n\n- The operator runs under Red Hat restricted SCC\n\n## Resources
          Required\n\n- The operator requires at least 85 Mi memory.\n\n## Limitations\n\n-
          Only runs on amd64, s390x and ppc64le architectures\n\n## Installing\n\n-
          A user with the cluster administrator role.\n- Install this operator in
          the same namespace as the IBM Metrics Operator\n  - default namespace: `ibm-software-central`\n\n##
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n# Configuration\n\nOptional:\n- Configure the
          Custom Resource Definition `DataReporterConfig` named `datareporterconfig`
          as per the following sample\n  - Configuration that can not be reconciled
          successfully will be reported on the DataReporterConfig Status\n  - jsonPath
          expressions are handled by: https://github.com/ohler55/ojg\n    - jsonPath
          comparison: https://cburgmer.github.io/json-path-comparison/\n  - Transformer
          type currently supported is `kazaam`\n    - https://github.com/willie68/kazaam\n\nSample
          DataReporterConfig:\n```YAML\napiVersion: marketplace.redhat.com/v1alpha1\nkind:
          DataReporterConfig\nmetadata:\n  name: datareporterconfig\nspec:\n  # true:
          report generated for 1 event and provide return code from data-service,
          confirming delivery\n  # false: events will be accumulated before building
          report, and immediately return 200, delivery unfconfirmed\n  confirmDelivery:
          false\n  dataFilters:   # dataFilters matches the first dataFilter selector,
          applies transforms and forwards to data-service and destinations\n  - altDestinations:
          \  # List of desintations to transform and foward events to\n    - authorization:
          \  # optional endpoint if an authorization token must first be requested
          \n        authDestHeader: Authorization   # http header key to be appended
          to destination request\n        authDestHeaderPrefix: 'Bearer '   # optional
          prefix for constructing destination request http header value\n        bodyData:
          \  # body data to send to authorization request, often an apiKey\n          secretKeyRef:\n
          \           name: auth-body-data-secret\n            key: bodydata\n        header:
          \  # secret map to use for authorization request headers\n          secret:\n
          \           name: auth-header-map-secret\n        tokenExpr: $.token   #
          optionally extract the returned body data using a jsonPath expression\n
          \       url: https://127.0.0.1:8443/api/2.0/accounts/my_account_id/apikeys/token
          \  # the url to make an authorization request to\n      header:   # secret
          map to use for destination request headers\n        secret:\n          name:
          dest-header-map-secret\n      transformer:   # The transformer to apply
          to event data for this destination\n        configMapKeyRef:\n          key:
          kazaam.json\n          name: kazaam-configmap\n        type: kazaam\n      url:
          https://127.0.0.1:8443/metering/api/1.0/usage   # the url to send the event
          to\n      urlSuffixExpr: $.properties.productId   # optional jsonPath expression
          on the event to extract and set a suffix to the url path\n    manifestType:
          dataReporter   # override manifestType of report sent to data-service\n
          \   selector:   # Matches against event data and a user. Empty selector
          matches all.\n      matchExpressions:   # jsonPath expression must return
          a result for all expressions to match (AND)\n      - $[?($.event == \"Account
          Contractual Usage\")]\n      - $.properties.productId\n      - $[?($.properties.source
          != null)]\n      - $[?($.properties.unit == \"AppPoints\")]\n      - $[?($.properties.quantity
          >= 0)]\n      - $[?($.timestamp != null)]\n      matchUsers:   # must match
          one of these users (OR). Omitting users matches any user\n      - system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n
          \   transformer:   # The transformer to apply to the event for data-service\n
          \     configMapKeyRef:\n        key: kazaam.json\n        name: kazaam-configmap\n
          \     type: kazaam\n  tlsConfig:   # TLS configuration for requests outbound
          to destinations\n    caCerts:\n    - name: tls-ca-certificates\n      key:
          ca.crt\n    certificates:\n    - clientCert:\n        secretKeyRef:\n          name:
          tls-ca-certificates\n          key: tls.crt\n      clientKey:\n        secretKeyRef:\n
          \         name: tls-ca-certificates\n          key: tls.key\n    cipherSuites:
          \n    - TLS_AES_128_GCM_SHA256\n    - TLS_AES_256_GCM_SHA384\n    - TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\n    - TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\n    insecureSkipVerify: false\n
          \   minVersion: VersionTLS12\n  userConfig:   # Optional metadata to apply
          to report sent to data-service for a specific user\n  - metadata:\n      ameta1:
          ametadata1\n      bmeta1: bmetadata1\n      cmeta1: cmetadata1\n      dmeta1:
          dmetadata1\n    userName: system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n```\n\nSample
          Header Secrets:\n```YAML\napiVersion: v1\nstringData:\n  accept: | \n    */*\nkind:
          Secret\nmetadata:\n  name: dest-header-map-secret\ntype: Opaque\n---\napiVersion:
          v1\nstringData:\n  accept: application/json\n  Content-Type: application/json\nkind:
          Secret\nmetadata:\n  name: auth-header-map-secret\ntype: Opaque\n```\n\nSample
          Authentication Body Secret:\n```YAML\napiVersion: v1\nstringData:\n  bodydata:
          |\n    {\"apikey\": \"myapikey\"}\nkind: Secret\nmetadata:\n  name: auth-body-data-secret\ntype:
          Opaque\n```\n\nSample TLS Secret:\n```YAML\napiVersion: v1\ndata:\n  ca.crt:
          LS0t...\n  tls.key: LS0t...\n  tls.crt: LS0t...\nkind: Secret\nmetadata:\n
          \ name: tls-ca-certificates\n```\n\nSample Kazaam ConfigMap:\n```YAML\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: kazaam-configmap\ndata:\n  kazaam.json:
          |\n    [\n        {\n            \"operation\": \"timestamp\",\n            \"spec\":
          {\n                \"timestamp\": {\n                \"inputFormat\": \"2006-01-02T15:04:05.999999+00:00\",\n
          \               \"outputFormat\": \"$unixmilli\"\n                }\n            }\n
          \       },\n        {\n            \"operation\": \"shift\",\n            \"spec\":
          {\n                \"instances[0].instanceId\": \"properties.source\",\n
          \               \"instances[0].startTime\": \"timestamp\",\n                \"instances[0].endTime\":
          \"timestamp\",\n                \"instances[0].metricUsage[0].metricId\":
          \"properties.unit\",\n                \"instances[0].metricUsage[0].quantity\":
          \"properties.quantity\"\n            }\n        },\n        {\n            \"operation\":
          \"default\",\n            \"spec\": {\n                \"meteringModel\":
          \"point-in-time\",\n                \"meteringPlan\": \"contract\"\n            }\n
          \       }\n    ]\n```\n\n### API Service User Configuration\n\n- The ClusterRole
          for api access is `clusterrole/ibm-data-reporter-operator-api`\n- The default
          ServiceAccount provided as an api user is `system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api`\n
          \ - The default ClusterRoleBinding for this user is `clusterrolebinding/ibm-data-reporter-operator-api`\n\nOptional:\n\n-
          To create an additional ServiceAccount and ClusterRoleBinding for api access\n\n```SHELL\nNAMESPACE=$(oc
          config view --minify -o jsonpath='{..namespace}') && \\\noc create serviceaccount
          my-api-service-account && \\\noc create clusterrolebinding ibm-data-reporter-operator-api
          --clusterrole=ibm-data-reporter-operator-api --serviceaccount=${NAMESPACE}:my-api-service-account\n```\n\n-
          Update datareporterconfig to attach metadata to reports associated with
          this user\n\n## API Service Usage\n\n- Get Token & Host\n\n```SHELL\noc
          project ibm-software-central\nexport DRTOKEN=$(oc create token ibm-data-reporter-operator-api
          --namespace ibm-software-central --duration 1h)\nexport DRHOST=$(oc get
          route ibm-data-reporter --template='{{ .spec.host }}')\n```\n\n- Get the
          Status\n\n```SHELL\ncurl -k -H \"Authorization: Bearer ${DRTOKEN}\" https://${DRHOST}/v1/status
          \n```\n\n- Post an Event\n\n```SHELL\ncurl -k -H \"Authorization: Bearer
          ${DRTOKEN}\" -X POST -d '{\"event\":\"myevent\"}' https://${DRHOST}/v1/event\n```\n\n##
          Storage\n\n- The operator temporarily stores events in pod memory, and writes
          to the IBM Metrics Operator data-service, which requires a PersistentVolume\n\n##
          License\n\nCopyright IBM Corporation 2023. All Rights Reserved.\n\nLicensed
          under the Apache License, Version 2.0 (the \"License\");\nyou may not use
          this file except in compliance with the License.\nYou may obtain a copy
          of the License at\n\n    http://www.apache.org/licenses/LICENSE-2.0\n\nUnless
          required by applicable law or agreed to in writing, software\ndistributed
          under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES
          OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for
          the specific language governing permissions and\nlimitations under the License.\n"
        displayName: IBM Data Reporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
        version: 2.20.0
      entries:
      - name: ibm-data-reporter-operator.v2.20.0
        version: 2.20.0
      name: stable
    defaultChannel: stable
    packageName: ibm-data-reporter-operator
    provider:
      name: Red Hat Marketplace
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-xtradb-cluster-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "docker.io/percona/percona-xtradb-cluster:8.0.42-33.1",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "docker.io/percona/haproxy:2.8.15",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "docker.io/percona/proxysql2:2.7.3",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "docker.io/percona/fluentbit:4.0.1",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:3.3.1",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "docker.io/percona/percona-xtrabackup:8.0.35-34.1",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.io/percona/percona-xtradb-cluster-operator:1.18.0
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <v1.18.0
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            clustercheck: Y2x1c3RlcmNoZWNrcGFzc3dvcmQ=
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            clustercheck: Y2x1c3RlcmNoZWNrcGFzc3dvcmQ=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ```
          ###  Release Highlights
          * **PMM3 support**

          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**

          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**

          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**

          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.

        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - database
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - docker.io/percona/percona-xtradb-cluster-operator:1.18.0
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator.v1.17.0
        version: 1.17.0
      - name: percona-xtradb-cluster-operator.v1.16.1
        version: 1.16.1
      - name: percona-xtradb-cluster-operator.v1.16.0
        version: 1.16.0
      - name: percona-xtradb-cluster-operator.v1.15.1
        version: 1.15.1
      - name: percona-xtradb-cluster-operator.v1.15.0
        version: 1.15.0
      - name: percona-xtradb-cluster-operator.v1.9.0
        version: 1.9.0
      name: stable
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: authorino-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: authorino-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta3",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/authorino-rhel9-operator@sha256:2d81613856df9725ad20928a7a9895d14482c42effce0a3a0de88f40ec35d411
          createdAt: "2025-03-25T22:39:23Z"
          description: Enables authentication and authorization for Gateways and applications
            in a Gateway API network.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta3
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: |-
          [Authorino](https://docs.kuadrant.io/authorino/) is Red Hat's Kubernetes-native lightweight external authorization service for tailor-made Zero Trust API security.

          Install this Red Hat official distribution of [Authorino Operator](https://docs.kuadrant.io/authorino-operator/) to manage instances of Authorino in this cluster.

          The current state of this distribution of the operator is: **Generally Available**.

          The Community version of Authorino Operator, based on upstream public images, is available in [OperatorHub.io](https://operatorhub.io/operator/authorino-operator).

          ## Getting started

          After installing the Operator, request an instance of the external authorization service by creating an `Authorino` custom resource.

          **Minimal example (namespaced)**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            listener:
              tls:
                enabled: false
            oidcServer:
              tls:
                enabled: false
          ```

          **Extended example**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            clusterWide: true
            authConfigLabelSelectors: environment=production
            secretLabelSelectors: authorino.kuadrant.io/component=authorino,environment=production

            replicas: 2

            evaluatorCacheSize: 2 # mb

            logLevel: info
            logMode: production

            listener:
              ports:
                grpc: 50001
                http: 5001
              tls:
                certSecretRef:
                  name: authorino-server-cert # secret must contain `tls.crt` and `tls.key` entries
              timeout: 2

            oidcServer:
              port: 8083
              tls:
                certSecretRef:
                  name: authorino-oidc-server-cert # secret must contain `tls.crt` and `tls.key` entries

            metrics:
              port: 8080
              deep: true

            healthz:
              port: 8081

            tracing:
              endpoint: rpc://otel-collector.observability.svc.cluster.local:4317
              insecure: true

            volumes:
              items:
                - name: keycloak-tls-cert
                  mountPath: /etc/ssl/certs
                  configMaps:
                    - keycloak-tls-cert
                  items: # details to mount the k8s configmap in the authorino pods
                    - key: keycloak.crt
                      path: keycloak.crt
              defaultMode: 420
          ```

          ### Cluster-wide vs Namespaced

          Namespaced instances only watch auth resources (`AuthConfig` and `Secrets`) created in the same namespace as the Authorino service. Use this mode for dedicated instances that do not require elevated privileges.

          Cluster-wide instances watch resources across the entire cluster (all namespaces.) Deploying and running Authorino in this mode requires elevated privileges.

          ### Multi-tenancy

          Use the `authConfigLabelSelectors` field of the `Authorino` custom resource to narrow the scope of the Authorino instance.

          Only `AuthConfig` custom resources whose labels match the value of this field will be handled by the Authorino instance.

          ## Protect a host

          To protect a host, create an `AuthConfig` custom resource for the host. E.g.:

          ```yaml
          apiVersion: authorino.kuadrant.io/v1beta3
          kind: AuthConfig
          metadata:
            name: my-api-protection
          spec:
            hosts:
            - my-api.io

            authentication:
              "keycloak":
                jwt:
                  issuerUrl: https://keycloak.keycloak.svc.cluster.local:8080/realms/my-realm

            authorization:
              "k8s-rbac":
                kubernetesSubjectAccessReview:
                  user:
                    selector: auth.identity.user.username
                  resourceAttributes:
                    resource:
                      value: my-api
                    verb:
                      selector: request.method
                cache:
                  key:
                    selector: auth.identity.user.username
                  ttl: 30
              "after-2am-only":
                rego: |
                  allow {
                    [hour, _, _] := time.clock(time.now_ns())
                    hour >= 2
                  }
          ```

          Make sure all requests to the host are fisrt checked with the Authorino instance, by configuring an Envoy proxy for external authz:

          ```yaml
          clusters:
          - name: my-api
            …
          - name: authorino
            connect_timeout: 0.25s
            type: STRICT_DNS
            lb_policy: ROUND_ROBIN
            http2_protocol_options: {}
            load_assignment:
              cluster_name: authorino
              endpoints:
              - lb_endpoints:
                - endpoint:
                    address:
                      socket_address:
                        address: authorino-authorino-authorization
                        port_value: 50051
          listeners:
          - filter_chains:
            - filters:
              name: envoy.http_connection_manager
              typed_config:
                "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
                use_remote_address: true
                route_config:
                  name: my-api-config
                  virtual_hosts:
                  - name: my-api-vs
                    domains:
                    - my-api.io
                    routes:
                    - match:
                        prefix: /
                      route:
                        cluster: my-api
                http_filters:
                - name: envoy.filters.http.ext_authz
                  typed_config:
                    "@type": type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz
                    transport_api_version: V3
                    failure_mode_allow: false
                    include_peer_certificate: true
                    grpc_service:
                      envoy_grpc:
                        cluster_name: authorino
                      timeout: 1s
          ```

          ...or, if using Istio, by creating an [`AuthorizationPolicy`](https://istio.io/latest/docs/reference/config/security/authorization-policy/#AuthorizationPolicy) custom resource. Use `action: CUSTOM` in the resource and the Authorino authorization service configured in the mesh extension provider settings.

          ## Features

          **Authentication**

          * JWT validation (with OpenID Connect Discovery)
          * OAuth 2.0 Token Introspection (opaque tokens)
          * Kubernetes TokenReview (ServiceAccount tokens)
          * API key authentication
          * X.509 client certificate authentication
          * Anonymous access
          * Proxy-handled (authentication performed by the proxy)

          **Authorization**

          * Built-in simple pattern matching (e.g. JWT claims, request attributes checking)
          * OPA policies (inline Rego and fetch from external registry)
          * Kubernetes SubjectAccessReview (resource and non-resource attributes)
          * Authzed SpiceDB

          **External metadata**

          * HTTP request
          * OpenID Connect User Info
          * UMA-protected resource attributes

          **Custom responses**

          * Header injection (Festival Wristbands tokens, JSON, plain text)
          * Envoy Dynamic Metadata
          * Custom HTTP response (status code, headers, messages, body, etc)

          **Callbacks**

          * HTTP webhooks

          **Caching**

          * OpenID Connect and User-Managed Access configs
          * JSON Web Keys (JWKs) and JSON Web Key Sets (JWKS)
          * Access tokens
          * External metadata
          * Precompiled Rego policies
          * Policy evaluation

          Check out the full [Feature Specification](https://docs.kuadrant.io/authorino/docs/features/) and how-to guides in the [Kuadrant Docs](https://docs.kuadrant.io) website.
        displayName: Red Hat - Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: GitHub
          url: https://github.com/kuadrant/authorino-operator
        - name: Documentation
          url: https://docs.kuadrant.io/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhcl-1/authorino-rhel9-operator@sha256:2d81613856df9725ad20928a7a9895d14482c42effce0a3a0de88f40ec35d411
        - registry.redhat.io/rhcl-1/authorino-rhel9@sha256:413d3796ad0101bf14d3341d31d9367e67fea1e73ebb3fb79124ef4c5dde0c09
        version: 1.2.2
      entries:
      - name: authorino-operator.v1.2.2
        version: 1.2.2
      - name: authorino-operator.v1.2.1
        version: 1.2.1
      - name: authorino-operator.v1.2.0
        version: 1.2.0
      - name: authorino-operator.v1.1.3
        version: 1.1.3
      - name: authorino-operator.v1.1.2
        version: 1.1.2
      - name: authorino-operator.v1.1.1
        version: 1.1.1
      - name: authorino-operator.v1.1.0
        version: 1.1.0
      - name: authorino-operator.v1.0.2
        version: 1.0.2
      - name: authorino-operator.v0.16.1
        version: 0.16.1
      - name: authorino-operator.v0.16.0
        version: 0.16.0
      name: stable
    - currentCSV: authorino-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta1",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "hosts": [
                    "my-api.io"
                  ],
                  "identity": [
                    {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "in": "authorization_header",
                        "keySelector": "APIKEY"
                      },
                      "name": "api-key-users"
                    }
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/3scale-tech-preview/authorino-rhel9-operator@sha256:eb574781cf39917f8f5d9e5c806f520534601f819ee1ebcbb751832da8938008
          createdAt: "2024-11-08T11:06:33Z"
          description: Authorino
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta1
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: |-
          [Authorino](https://docs.kuadrant.io/authorino/) is Red Hat's Kubernetes-native lightweight external authorization service for tailor-made Zero Trust API security.

          Install this Red Hat official distribution of [Authorino Operator](https://docs.kuadrant.io/authorino-operator/) to manage instances of Authorino in this cluster.

          The current state of this distribution of the operator is: **_Technical Preview_**.

          The Community version of Authorino Operator, based on upstream public images, is available in [OperatorHub.io](https://operatorhub.io/operator/authorino-operator).

          ## Getting started

          After installing the Operator, request an instance of the external authorization service by creating an `Authorino` custom resource.

          **Minimal example (namespaced)**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            listener:
              tls:
                enabled: false
            oidcServer:
              tls:
                enabled: false
          ```

          **Extended example**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            clusterWide: true
            authConfigLabelSelectors: environment=production
            secretLabelSelectors: authorino.kuadrant.io/component=authorino,environment=production

            replicas: 2

            evaluatorCacheSize: 2 # mb

            logLevel: info
            logMode: production

            listener:
              ports:
                grpc: 50001
                http: 5001
              tls:
                certSecretRef:
                  name: authorino-server-cert # secret must contain `tls.crt` and `tls.key` entries
              timeout: 2

            oidcServer:
              port: 8083
              tls:
                certSecretRef:
                  name: authorino-oidc-server-cert # secret must contain `tls.crt` and `tls.key` entries

            metrics:
              port: 8080
              deep: true

            healthz:
              port: 8081

            tracing:
              endpoint: rpc://otel-collector.observability.svc.cluster.local:4317
              insecure: true

            volumes:
              items:
                - name: keycloak-tls-cert
                  mountPath: /etc/ssl/certs
                  configMaps:
                    - keycloak-tls-cert
                  items: # details to mount the k8s configmap in the authorino pods
                    - key: keycloak.crt
                      path: keycloak.crt
              defaultMode: 420
          ```

          ### Cluster-wide vs Namespaced

          Namespaced instances only watch auth resources (`AuthConfig` and `Secrets`) created in the same namespace as the Authorino service. Use this mode for dedicated instances that do not require elevated privileges.

          Cluster-wide instances watch resources across the entire cluster (all namespaces.) Deploying and running Authorino in this mode requires elevated privileges.

          ### Multi-tenancy

          Use the `authConfigLabelSelectors` field of the `Authorino` custom resource to narrow the scope of the Authorino instance.

          Only `AuthConfig` custom resources whose labels match the value of this field will be handled by the Authorino instance.

          ## Protect a host

          To protect a host, create an `AuthConfig` custom resource for the host. E.g.:

          ```yaml
          apiVersion: authorino.kuadrant.io/v1beta2
          kind: AuthConfig
          metadata:
            name: my-api-protection
          spec:
            hosts:
            - my-api.io

            authentication:
              "keycloak":
                jwt:
                  issuerUrl: https://keycloak.keycloak.svc.cluster.local:8080/realms/my-realm

            authorization:
              "k8s-rbac":
                kubernetesSubjectAccessReview:
                  user:
                    selector: auth.identity.user.username
                  resourceAttributes:
                    resource:
                      value: my-api
                    verb:
                      selector: request.method
                cache:
                  key:
                    selector: auth.identity.user.username
                  ttl: 30
              "after-2am-only":
                rego: |
                  allow {
                    [hour, _, _] := time.clock(time.now_ns())
                    hour >= 2
                  }
          ```

          Make sure all requests to the host are fisrt checked with the Authorino instance, by configuring an Envoy proxy for external authz:

          ```yaml
          clusters:
          - name: my-api
            …
          - name: authorino
            connect_timeout: 0.25s
            type: STRICT_DNS
            lb_policy: ROUND_ROBIN
            http2_protocol_options: {}
            load_assignment:
              cluster_name: authorino
              endpoints:
              - lb_endpoints:
                - endpoint:
                    address:
                      socket_address:
                        address: authorino-authorino-authorization
                        port_value: 50051
          listeners:
          - filter_chains:
            - filters:
              name: envoy.http_connection_manager
              typed_config:
                "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
                use_remote_address: true
                route_config:
                  name: my-api-config
                  virtual_hosts:
                  - name: my-api-vs
                    domains:
                    - my-api.io
                    routes:
                    - match:
                        prefix: /
                      route:
                        cluster: my-api
                http_filters:
                - name: envoy.filters.http.ext_authz
                  typed_config:
                    "@type": type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz
                    transport_api_version: V3
                    failure_mode_allow: false
                    include_peer_certificate: true
                    grpc_service:
                      envoy_grpc:
                        cluster_name: authorino
                      timeout: 1s
          ```

          ...or, if using Istio, by creating an [`AuthorizationPolicy`](https://istio.io/latest/docs/reference/config/security/authorization-policy/#AuthorizationPolicy) custom resource. Use `action: CUSTOM` in the resource and the Authorino authorization service configured in the mesh extension provider settings.

          ## Features

          **Authentication**

          * JWT validation (with OpenID Connect Discovery)
          * OAuth 2.0 Token Introspection (opaque tokens)
          * Kubernetes TokenReview (ServiceAccount tokens)
          * API key authentication
          * X.509 client certificate authentication
          * Anonymous access
          * Proxy-handled (authentication performed by the proxy)

          **Authorization**

          * Built-in simple pattern matching (e.g. JWT claims, request attributes checking)
          * OPA policies (inline Rego and fetch from external registry)
          * Kubernetes SubjectAccessReview (resource and non-resource attributes)
          * Authzed SpiceDB

          **External metadata**

          * HTTP request
          * OpenID Connect User Info
          * UMA-protected resource attributes

          **Custom responses**

          * Header injection (Festival Wristbands tokens, JSON, plain text)
          * Envoy Dynamic Metadata
          * Custom HTTP response (status code, headers, messages, body, etc)

          **Callbacks**

          * HTTP webhooks

          **Caching**

          * OpenID Connect and User-Managed Access configs
          * JSON Web Keys (JWKs) and JSON Web Key Sets (JWKS)
          * Access tokens
          * External metadata
          * Precompiled Rego policies
          * Policy evaluation

          Check out the full [Feature Specification](https://docs.kuadrant.io/authorino/docs/features/) and how-to guides in the [Kuadrant Docs](https://docs.kuadrant.io) website.
        displayName: Red Hat - Authorino (Technical Preview)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: GitHub
          url: https://github.com/kuadrant/authorino-operator
        - name: Documentation
          url: https://docs.kuadrant.io/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-tech-preview/authorino-rhel9@sha256:a473dae20e71e3e813ac30ba978f2ab3c5e19d7d39b501ae9103dca892107c87
        - registry.redhat.io/3scale-tech-preview/authorino-rhel9-operator@sha256:eb574781cf39917f8f5d9e5c806f520534601f819ee1ebcbb751832da8938008
        - registry.redhat.io/openshift4/ose-cli-rhel9@sha256:3a398158d10018bfac5df291d6998b513cf9a30e5477bcaefddcf66d229627c7
        version: 1.1.3
      entries:
      - name: authorino-operator.v1.1.3
        version: 1.1.3
      - name: authorino-operator.v1.1.2
        version: 1.1.2
      - name: authorino-operator.v1.1.1
        version: 1.1.1
      - name: authorino-operator.v1.1.0
        version: 1.1.0
      - name: authorino-operator.v1.0.2
        version: 1.0.2
      name: tech-preview-v1
    defaultChannel: stable
    packageName: authorino-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-athena-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-athena-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "athena.services.k8s.aws/v1alpha1",
                "kind": "WorkGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/athena-controller:1.1.1
          createdAt: "2025-09-19T21:06:02Z"
          description: AWS Athena controller is a service controller for managing
            Athena resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PreparedStatement represents the state of an AWS athena PreparedStatement
              resource.
            displayName: PreparedStatement
            kind: PreparedStatement
            name: preparedstatements.athena.services.k8s.aws
            version: v1alpha1
          - description: WorkGroup represents the state of an AWS athena WorkGroup
              resource.
            displayName: WorkGroup
            kind: WorkGroup
            name: workgroups.athena.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Athena resources in AWS from within your Kubernetes cluster.

          **About Amazon Athena**

          Amazon Athena is an interactive query service that makes it easy to analyze data directly in Amazon Simple Storage Service (Amazon S3) using standard SQL. With a few actions in the AWS Management Console, you can point Athena at your data stored in Amazon S3 and begin using standard SQL to run ad-hoc queries and get results in seconds.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Athena
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - athena
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Athena Developer Resources
          url: https://aws.amazon.com/athena/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: athena maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/athena-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-athena-controller.v1.1.1
        version: 1.1.1
      - name: ack-athena-controller.v1.1.0
        version: 1.1.0
      - name: ack-athena-controller.v1.0.14
        version: 1.0.14
      - name: ack-athena-controller.v1.0.13
        version: 1.0.13
      - name: ack-athena-controller.v1.0.12
        version: 1.0.12
      - name: ack-athena-controller.v1.0.11
        version: 1.0.11
      - name: ack-athena-controller.v1.0.10
        version: 1.0.10
      - name: ack-athena-controller.v1.0.9
        version: 1.0.9
      - name: ack-athena-controller.v1.0.8
        version: 1.0.8
      - name: ack-athena-controller.v1.0.7
        version: 1.0.7
      - name: ack-athena-controller.v1.0.6
        version: 1.0.6
      - name: ack-athena-controller.v1.0.5
        version: 1.0.5
      - name: ack-athena-controller.v1.0.4
        version: 1.0.4
      - name: ack-athena-controller.v1.0.1
        version: 1.0.1
      - name: ack-athena-controller.v1.0.0
        version: 1.0.0
      - name: ack-athena-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-athena-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Flux project
      provider-url: ""
    name: flux
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.toolkit.fluxcd.io/v2beta1",
                "kind": "HelmRelease",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "5m",
                    "chart": {
                      "spec": {
                          "chart": "podinfo",
                          "version": "4.0.x",
                          "sourceRef": {
                            "kind": "HelmRepository",
                            "name": "podinfo",
                            "namespace": "flux-system"
                          },
                          "interval": "1m"
                      }
                    },
                    "values": {
                      "replicaCount": 2
                    }
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "GitRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "1m",
                    "ref": {
                      "branch": "master"
                    },
                    "url": "https://github.com/stefanprodan/podinfo"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmRepository",
                "metadata": {
                    "name": "stable",
                    "namespace": "default"
                },
                "spec": {
                    "url": "https://kubernetes-charts.storage.googleapis.com/",
                    "interval": "10m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmChart",
                "metadata": {
                    "name": "redis",
                    "namespace": "default"
                },
                "spec": {
                    "chart": "redis",
                    "version": "10.5.7",
                    "sourceRef": {
                      "name": "stable",
                      "kind": "HelmRepository"
                    },
                    "interval": "5m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "Bucket",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "1m",
                    "provider": "generic",
                    "bucketName": "podinfo",
                    "endpoint": "minio.minio.svc.cluster.local:9000",
                    "insecure": true,
                    "secretRef": {
                      "name": "minio-credentials"
                    }
                }
              },
              {
                "apiVersion": "kustomize.toolkit.fluxcd.io/v1beta2",
                "kind": "Kustomization",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "5m",
                    "path": "./deploy/overlays/dev",
                    "prune": true,
                    "sourceRef": {
                      "kind": "GitRepository",
                      "name": "podinfo"
                    },
                    "timeout": "2m",
                    "wait": true
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "address": "https://github.com/stefanprodan/podinfo",
                    "secretRef": {
                      "name": "api-token"
                    }
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Alert",
                "metadata": {
                    "name": "webapp",
                    "namespace": "default"
                },
                "spec": {
                    "providerRef": {
                      "name": "on-call-slack"
                    },
                    "eventSeverity": "info",
                    "eventSources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "Bucket",
                          "name": "secrets"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-backend"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-frontend"
                      }
                    ]
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Receiver",
                "metadata": {
                    "name": "github-receiver",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "events": [
                      "ping",
                      "push"
                    ],
                    "secretRef": {
                      "name": "webhook-token"
                    },
                    "resources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "HelmRepository",
                          "name": "webapp"
                      }
                    ]
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "image": "ghcr.io/stefanprodan/podinfo",
                    "interval": "1m0s"
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImagePolicy",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "imageRepositoryRef": {
                      "name": "podinfo"
                    },
                    "policy": {
                      "semver": {
                          "range": "5.0.x"
                      }
                    }
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageUpdateAutomation",
                "metadata": {
                  "name": "flux-system",
                  "namespace": "flux-system"
                },
                "spec": {
                  "interval": "1m0s",
                  "sourceRef": {
                    "kind": "GitRepository",
                    "name": "flux-system"
                  },
                  "git": {
                    "checkout": {
                      "ref": {
                        "branch": "main"
                      }
                    },
                    "commit": {
                      "author": {
                        "email": "fluxcdbot@users.noreply.github.com",
                        "name": "fluxcdbot"
                      },
                      "messageTemplate": "{{range .Updated.Images}}{{println .}}{{end}}"
                    },
                    "push": {
                      "branch": "main"
                    }
                  },
                  "update": {
                    "path": "./clusters/my-cluster",
                    "strategy": "Setters"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/fluxcd/source-controller:v1.3.0
          description: Flux is a Continuous Delivery solution for Kubernetes.
          operatorframework.io/suggested-namespace: flux-system
          support: The Flux community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta3
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta1
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta2
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta1
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta2
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta2
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta2
          - description: OCIRepository
            displayName: OCIRepository
            kind: OCIRepository
            name: ocirepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta3
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta2
        description: |-
          [Flux](https://fluxcd.io/) is a Continuous Delivery solution for Kubernetes.
          Flux is constructed with the [GitOps Toolkit](https://toolkit.fluxcd.io/components/), a set of composable APIs and specialized tools for keeping Kubernetes clusters in sync with sources of configuration (like Git & Helm repositories), and automating updates to configuration when there is new code to deploy.

          Flux is a CNCF incubation project made for:

          - **cluster operators** who automate provision and configuration of clusters
          - **platform engineers** who build continuous delivery for developer teams
          - **app developers** who rely on continuous delivery to get their code live

          Features:

          - Source configuration from Git and Helm repositories, and S3-compatible buckets (e.g., Minio)
          - Kustomize and Helm support
          - Event-triggered and periodic reconciliation
          - Integration with Kubernetes RBAC
          - Health assessment (clusters and workloads)
          - Dependency management (infrastructure and workloads)
          - Secrets management (builtin support for Mozilla SOPS and Cloud KMS)
          - Alerting to external systems (webhook senders)
          - External events handling (webhook receivers)
          - Automated container image updates to Git (image scanning and patching)
          - Policy-driven validation (OPA, Kyverno, admission controllers)
          - Seamless integration with Git providers (GitHub, GitLab, Bitbucket)
          - Interoperability with Cluster API (CAPI) providers

          ### Steps to support bootstraping with Flux CLI

          SecComp Profile should be removed as it requires a special privilege to get the access.
          Also, it requires to enforce `runUserAs` to the same UID provided by the images to prevent OpenShift
          to alter the user id expected by our controllers.

          To do so, please specifiy a patch using the following Kustomization file, for example.

          ```yaml
          apiVersion: kustomize.config.k8s.io/v1beta1
          kind: Kustomization
          resources:
            - gotk-components.yaml
            - gotk-sync.yaml
          patches:
            - patch: |
                apiVersion: apps/v1
                kind: Deployment
                metadata:
                  name: all
                spec:
                  template:
                    spec:
                      containers:
                        - name: manager
                          securityContext:
                            runAsUser: 65534
                            seccompProfile:
                              $patch: delete
              target:
                kind: Deployment
                labelSelector: app.kubernetes.io/part-of=flux
          ```

          Then before installing Flux, please also perform the following steps to bind Flux to the `nonroot` SecurityContextConstraint (SCC). This applies to installation via OC's OLM or via Flux CLI.

          ```sh
          # replace default "flux-system" with the project/namespace name for your flux installation
          namespace="flux-system"
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:source-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:kustomize-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:helm-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:notification-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-automation-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-reflector-controller

          ```

          To see complete steps please refer to this [documentation](https://github.com/fluxcd/website/blob/main/content/en/docs/use-cases/openshift.md).

          If you've already installed Flux, you might need to uninstall it after applying the above pre-requisite steps.

          ### Network Policies

          This version of Flux is shipped without the normally provided network Policies.
          By default all traffic is enabled, and anyone can access the source-controller
          artifacts.

          It is not recommended to run it in production as is.

          Please visit [Flux's policies](https://github.com/fluxcd/flux2/tree/main/manifests/policies)
          to deploy the needed network policies and secure your Flux setup.

          ### Breaking Changes

          - Flux 0.18.x ships with v1beta2 of the kustomize.toolkit.fluxcd.io/Kustomization CRD.
          - Flux 0.25.x is the last Flux release where Kubernetes 1.19 is supported.

          ### Flux on OpenShift Use Cases

          - [The GitOps with Flux + OpenShift](https://www.youtube.com/watch?v=8ybCtNVP5To) talk by Andrew Block (Red Hat) showing how to setup multi-tenacy on OpenShift using Flux.
          - [Flux OpenShift Use Case](https://fluxcd.io/docs/use-cases/openshift/) on FluxCD.io
        displayName: Flux
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - gitops
        - cd
        - continuous delivery
        - fluxcd
        - helm
        - kustomize
        links:
        - name: Homepage
          url: https://fluxcd.io/
        - name: Documentation
          url: https://fluxcd.io/docs/
        - name: Slack
          url: https://cloud-native.slack.com/messages/flux
        - name: Mailing List
          url: https://lists.cncf.io/g/cncf-flux-dev
        - name: Source Code
          url: https://github.com/fluxcd/flux2
        maintainers:
        - email: support@weave.works
          name: The Flux maintainers
        maturity: stable
        provider:
          name: The Flux project
        relatedImages:
        - ghcr.io/fluxcd/helm-controller:v1.0.1
        - ghcr.io/fluxcd/image-automation-controller:v0.38.0
        - ghcr.io/fluxcd/image-reflector-controller:v0.32.0
        - ghcr.io/fluxcd/kustomize-controller:v1.3.0
        - ghcr.io/fluxcd/notification-controller:v1.3.0
        - ghcr.io/fluxcd/source-controller:v1.3.0
        version: 2.3.0
      entries:
      - name: flux.v2.3.0
        version: 2.3.0
      - name: flux.v2.2.2
        version: 2.2.2
      - name: flux.v2.2.0
        version: 2.2.0
      - name: flux.v2.0.1
        version: 2.0.1
      - name: flux.v0.41.2
        version: 0.41.2
      - name: flux.v0.40.0
        version: 0.40.0
      - name: flux.v0.39.0
        version: 0.39.0
      - name: flux.v0.38.3
        version: 0.38.3
      - name: flux.v0.37.0
        version: 0.37.0
      - name: flux.v0.36.0
        version: 0.36.0
      - name: flux.v0.35.0
        version: 0.35.0
      - name: flux.v0.31.4
        version: 0.31.4
      - name: flux.v0.31.3
        version: 0.31.3
      - name: flux.v0.30.2
        version: 0.30.2
      - name: flux.v0.28.5
        version: 0.28.5
      - name: flux.v0.28.4
        version: 0.28.4
      - name: flux.v0.28.3
        version: 0.28.3
      - name: flux.v0.27.4
        version: 0.27.4
      - name: flux.v0.27.3
        version: 0.27.3
      - name: flux.v0.27.2
        version: 0.27.2
      - name: flux.v0.27.0
        version: 0.27.0
      - name: flux.v0.26.1
        version: 0.26.1
      - name: flux.v0.26.0
        version: 0.26.0
      - name: flux.v0.25.3
        version: 0.25.3
      - name: flux.v0.25.2
        version: 0.25.2
      - name: flux.v0.24.1
        version: 0.24.1
      - name: flux.v0.24.0
        version: 0.24.0
      - name: flux.v0.23.0
        version: 0.23.0
      - name: flux.v0.22.1
        version: 0.22.1
      - name: flux.v0.22.0
        version: 0.22.0
      - name: flux.v0.21.1
        version: 0.21.1
      - name: flux.v0.21.0
        version: 0.21.0
      - name: flux.v0.20.1
        version: 0.20.1
      - name: flux.v0.20.0
        version: 0.20.0
      - name: flux.v0.19.1
        version: 0.19.1
      - name: flux.v0.18.3
        version: 0.18.3
      - name: flux.v0.18.1
        version: 0.18.1
      - name: flux.v0.17.2
        version: 0.17.2
      - name: flux.v0.17.1
        version: 0.17.1
      - name: flux.v0.17.0
        version: 0.17.0
      - name: flux.v0.16.2
        version: 0.16.2
      - name: flux.v0.16.1
        version: 0.16.1
      - name: flux.v0.16.0
        version: 0.16.0
      - name: flux.v0.15.3
        version: 0.15.3
      - name: flux.v0.15.2
        version: 0.15.2
      - name: flux.v0.14.2
        version: 0.14.2
      - name: flux.v0.14.1
        version: 0.14.1
      - name: flux.v0.14.0
        version: 0.14.0
      - name: flux.v0.13.4
        version: 0.13.4
      name: stable
    defaultChannel: stable
    packageName: flux
    provider:
      name: The Flux project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          olm.skipRange: <1.0.4
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
        version: 1.0.4
      entries:
      - name: datadog-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      name: stable
    defaultChannel: stable
    packageName: datadog-operator-certified
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kiali-ossm
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kiali-operator.v2.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "Kiali",
                "metadata": {
                  "name": "kiali"
                },
                "spec": {
                  "version": "default",
                  "istio_namespace": "istio-system",
                  "deployment": {
                    "logger": {
                      "log_level": "info"
                    },
                    "view_only_mode": false
                  }
                }
              },
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "OSSMConsole",
                "metadata": {
                  "name": "ossmconsole"
                },
                "spec": {
                  "version": "default",
                  "kiali": {
                    "serviceName": "",
                    "serviceNamespace": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9-operator@sha256:0612dfa52f2bf39a1d87e640409313585fedc7cfcff6d00e77bb14809d27d646
          createdAt: "2024-12-12T10:44:15Z"
          description: This productized operator provides Kiali and OSSMC. Kiali is
            the Istio observability and management Console. OSSMC is the OpenShift
            Service Mesh Console plugin, powered by Kiali.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.1.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kiali/kiali
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A configuration file for a Kiali installation.
            displayName: Kiali
            kind: Kiali
            name: kialis.kiali.io
            version: v1alpha1
          - description: A configuration file for a OpenShift Service Mesh Console
              installation.
            displayName: OpenShift Service Mesh Console
            kind: OSSMConsole
            name: ossmconsoles.kiali.io
            version: v1alpha1
        description: |-
          ## About the managed application

          A Microservice Architecture breaks up the monolith into many smaller pieces
          that are composed together. Patterns to secure the communication between
          services like fault tolerance (via timeout, retry, circuit breaking, etc.)
          have come up as well as distributed tracing to be able to see where calls
          are going.

          A service mesh can now provide these services on a platform level and frees
          the application writers from those tasks. Routing decisions are done at the
          mesh level.

          Kiali works with OpenShift Service Mesh to visualize the service
          mesh topology, to provide visibility into features like circuit breakers,
          request rates and more. It offers insights about the mesh components at
          different levels, from abstract Applications to Services and Workloads.

          See [https://www.kiali.io](https://www.kiali.io) to read more.

          ### Accessing the UI

          By default, the Kiali operator exposes the Kiali UI as an OpenShift Route.

          If on OpenShift, you can create an OSSMConsole CR to have the operator
          install the OpenShift ServiceMesh Console plugin to the OpenShift Console
          thus providing an interface directly integrated with the OpenShift Console.

          ## About this Operator

          ### Kiali Custom Resource Configuration Settings

          For quick descriptions of all the settings you can configure in the Kiali
          Custom Resource (CR), see
          [the kiali.io docs](https://kiali.io/docs/configuration/kialis.kiali.io/).

          ## Prerequisites for enabling this Operator

          Kiali is a companion tool for OpenShift Service Mesh. So before you install Kiali, you must have
          already installed Service Mesh.
        displayName: Kiali Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - service-mesh
        - observability
        - monitoring
        - istio
        links:
        - name: Installation Guide
          url: https://www.kiali.io/docs/installation
        - name: Features
          url: https://www.kiali.io/docs/features
        - name: Documentation Home
          url: https://www.kiali.io
        - name: Blogs and Articles
          url: https://medium.com/kialiproject
        - name: Server and UI Source Code
          url: https://github.com/kiali/kiali
        - name: Operator Source Code
          url: https://github.com/kiali/kiali-operator
        - name: OpenShift Service Mesh Console Code
          url: https://github.com/kiali/openshift-servicemesh-plugin
        maintainers:
        - email: kiali-dev@googlegroups.com
          name: Kiali Developers Google Group
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel8@sha256:e37d76dd7c7a2db79ff1e96afef19b07c3beaa8ebf0c7bd85b808908b8f0ce77
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9-operator@sha256:0612dfa52f2bf39a1d87e640409313585fedc7cfcff6d00e77bb14809d27d646
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9@sha256:06fde6c696a33d17b0d2f7620ec46422c82277fd17a3fe6d7319de05c732aace
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:976a31ca8cf17ac0887b8a586edfcc29eb30175802461b3114758af525e7b800
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:66e4d933a9d69158bf87dd45835183bf2308b77c48908ab64496c5d22c816293
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:27a6aebffa3d049ee48050126e6107db2e07c49e9ce9297175facb6368e1e792
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-ossmc-rhel9@sha256:99643f294c3f92a6fc82f321cab7c756e4879fefb48fa824f1bf49ff7182658d
        version: 2.1.2
      entries:
      - name: kiali-operator.v2.1.2
        version: 2.1.2
      - name: kiali-operator.v1.89.8
        version: 1.89.8
      - name: kiali-operator.v1.89.7
        version: 1.89.7
      - name: kiali-operator.v1.89.6
        version: 1.89.6
      - name: kiali-operator.v1.89.1
        version: 1.89.1
      - name: kiali-operator.v1.73.10
        version: 1.73.10
      - name: kiali-operator.v1.73.8
        version: 1.73.8
      - name: kiali-operator.v1.73.7
        version: 1.73.7
      - name: kiali-operator.v1.73.4
        version: 1.73.4
      - name: kiali-operator.v1.65.11
        version: 1.65.11
      - name: kiali-operator.v1.65.10
        version: 1.65.10
      - name: kiali-operator.v1.65.9
        version: 1.65.9
      - name: kiali-operator.v1.65.8
        version: 1.65.8
      - name: kiali-operator.v1.65.7
        version: 1.65.7
      - name: kiali-operator.v1.65.6
        version: 1.65.6
      - name: kiali-operator.v1.57.7
        version: 1.57.7
      name: candidate
    - currentCSV: kiali-operator.v2.11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "Kiali",
                "metadata": {
                  "name": "kiali"
                },
                "spec": {
                  "version": "default",
                  "istio_namespace": "istio-system",
                  "deployment": {
                    "logger": {
                      "log_level": "info"
                    },
                    "view_only_mode": false
                  }
                }
              },
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "OSSMConsole",
                "metadata": {
                  "name": "ossmconsole"
                },
                "spec": {
                  "version": "default",
                  "kiali": {
                    "serviceName": "",
                    "serviceNamespace": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/kiali-rhel9-operator@sha256:2970861bd8e65e36d5080dc4d56340d08a52dbb1217be352998d135eb854732f
          createdAt: "2025-09-10T00:28:49Z"
          description: This productized operator provides Kiali and OSSMC. Kiali is
            the Istio observability and management Console. OSSMC is the OpenShift
            Service Mesh Console plugin, powered by Kiali.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kiali/kiali
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A configuration file for a Kiali installation.
            displayName: Kiali
            kind: Kiali
            name: kialis.kiali.io
            version: v1alpha1
          - description: A configuration file for a OpenShift Service Mesh Console
              installation.
            displayName: OpenShift Service Mesh Console
            kind: OSSMConsole
            name: ossmconsoles.kiali.io
            version: v1alpha1
        description: |-
          ## About the managed application

          A Microservice Architecture breaks up the monolith into many smaller pieces
          that are composed together. Patterns to secure the communication between
          services like fault tolerance (via timeout, retry, circuit breaking, etc.)
          have come up as well as distributed tracing to be able to see where calls
          are going.

          A service mesh can now provide these services on a platform level and frees
          the application writers from those tasks. Routing decisions are done at the
          mesh level.

          Kiali works with OpenShift Service Mesh to visualize the service
          mesh topology, to provide visibility into features like circuit breakers,
          request rates and more. It offers insights about the mesh components at
          different levels, from abstract Applications to Services and Workloads.

          See [https://www.kiali.io](https://www.kiali.io) to read more.

          ### Accessing the UI

          By default, the Kiali operator exposes the Kiali UI as an OpenShift Route.

          If on OpenShift, you can create an OSSMConsole CR to have the operator
          install the OpenShift ServiceMesh Console plugin to the OpenShift Console
          thus providing an interface directly integrated with the OpenShift Console.

          ## About this Operator

          ### Kiali Custom Resource Configuration Settings

          For quick descriptions of all the settings you can configure in the Kiali
          Custom Resource (CR), see
          [the kiali.io docs](https://kiali.io/docs/configuration/kialis.kiali.io/).

          ## Prerequisites for enabling this Operator

          Kiali is a companion tool for OpenShift Service Mesh. So before you install Kiali, you must have
          already installed Service Mesh.
        displayName: Kiali Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - service-mesh
        - observability
        - monitoring
        - istio
        links:
        - name: Installation Guide
          url: https://www.kiali.io/docs/installation
        - name: Features
          url: https://www.kiali.io/docs/features
        - name: Documentation Home
          url: https://www.kiali.io
        - name: Blogs and Articles
          url: https://medium.com/kialiproject
        - name: Server and UI Source Code
          url: https://github.com/kiali/kiali
        - name: Operator Source Code
          url: https://github.com/kiali/kiali-operator
        - name: OpenShift Service Mesh Console Code
          url: https://github.com/kiali/openshift-servicemesh-plugin
        maintainers:
        - email: kiali-dev@googlegroups.com
          name: Kiali Developers Google Group
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9-operator@sha256:2970861bd8e65e36d5080dc4d56340d08a52dbb1217be352998d135eb854732f
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:7487fbb4b776dfca28ad09311ba857e07ad98bc91ca5ee6fcaaa5b7783d04fbd
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel9@sha256:9a0d4862a25c2999a3f85b9a7a0332dab4ec1744548191d7a00c838b06ba26d2
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel9@sha256:fe25be563d50e5ad4e5475f153e03d4b24c7db094d7e824c51ebaae79418b1f7
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel8@sha256:81db94df331799f4087596378914b712d85d117dde3df2d4902079a82d4a9c55
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9@sha256:a00ec00e466e080592dca5c1cc153698b1412a54e1979647f74281f79b5691cf
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9@sha256:85eeb972e46f26ff30e04991e76ec3050d98b334e7b7f201a094f8d50b06854e
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:d729ef8d74c2f6e02d8c68ca761b1fb3c7c3706f1290b684468f0acb0ef313e0
        version: 2.11.3
      entries:
      - name: kiali-operator.v2.11.3
        version: 2.11.3
      - name: kiali-operator.v2.11.2
        version: 2.11.2
      - name: kiali-operator.v2.11.1
        version: 2.11.1
      - name: kiali-operator.v2.4.7
        version: 2.4.7
      - name: kiali-operator.v2.4.6
        version: 2.4.6
      - name: kiali-operator.v2.4.5
        version: 2.4.5
      - name: kiali-operator.v2.4.4
        version: 2.4.4
      - name: kiali-operator.v2.4.3
        version: 2.4.3
      - name: kiali-operator.v1.89.10
        version: 1.89.10
      - name: kiali-operator.v1.89.9
        version: 1.89.9
      - name: kiali-operator.v1.89.8
        version: 1.89.8
      - name: kiali-operator.v1.89.7
        version: 1.89.7
      - name: kiali-operator.v1.89.6
        version: 1.89.6
      - name: kiali-operator.v1.89.1
        version: 1.89.1
      - name: kiali-operator.v1.73.10
        version: 1.73.10
      - name: kiali-operator.v1.73.8
        version: 1.73.8
      - name: kiali-operator.v1.73.7
        version: 1.73.7
      - name: kiali-operator.v1.73.4
        version: 1.73.4
      - name: kiali-operator.v1.65.11
        version: 1.65.11
      - name: kiali-operator.v1.65.10
        version: 1.65.10
      - name: kiali-operator.v1.65.9
        version: 1.65.9
      - name: kiali-operator.v1.65.8
        version: 1.65.8
      - name: kiali-operator.v1.65.7
        version: 1.65.7
      - name: kiali-operator.v1.65.6
        version: 1.65.6
      - name: kiali-operator.v1.57.7
        version: 1.57.7
      name: stable
    defaultChannel: stable
    packageName: kiali-ossm
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: test-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: test-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "test.openstack.org/v1beta1",
                "kind": "HorizonTest",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "test-operator",
                    "app.kubernetes.io/instance": "horizontest-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "horizontest",
                    "app.kubernetes.io/part-of": "test-operator"
                  },
                  "name": "horizontest-sample"
                },
                "spec": {
                  "adminPassword": "12345678",
                  "adminUsername": "admin",
                  "authUrl": "https://keystone-public-openstack.apps.ocp.openstack.lab",
                  "backoffLimit": 0,
                  "containerImage": "",
                  "dashboardUrl": "https://horizon-openstack.apps.ocp.openstack.lab/",
                  "flavorName": "m1.tiny",
                  "horizonRepoBranch": "master",
                  "horizonTestDir": "/var/lib/horizontest",
                  "imageUrl": "http://download.cirros-cloud.net/0.6.2/cirros-0.6.2-x86_64-disk.img",
                  "logsDirectoryName": "horizon",
                  "parallel": false,
                  "password": "horizontest",
                  "projectName": "horizontest",
                  "repoUrl": "https://review.opendev.org/openstack/horizon",
                  "storageClass": "local-storage",
                  "user": "horizontest"
                }
              },
              {
                "apiVersion": "test.openstack.org/v1beta1",
                "kind": "Tempest",
                "metadata": {
                  "name": "tempest-tests",
                  "namespace": "openstack"
                },
                "spec": {
                  "containerImage": "",
                  "tempestRun": {
                    "concurrency": 8,
                    "includeList": "tempest.api.identity.v3.*\n"
                  },
                  "tempestconfRun": null
                }
              },
              {
                "apiVersion": "test.openstack.org/v1beta1",
                "kind": "Tobiko",
                "metadata": {
                  "name": "tobiko-tests",
                  "namespace": "openstack"
                },
                "spec": {
                  "config": "[DEFAULT]\nlog_file = value\n\n[testcase]\ntimeout = value\ntest_runner_timeout = value\n\n[ubuntu]\ninterface_name = value\n\n[keystone]\ninterface_name = value\n",
                  "containerImage": "",
                  "testenv": "py3",
                  "version": "master",
                  "workflow": [
                    {
                      "stepName": "unit-tests",
                      "testenv": "py3"
                    },
                    {
                      "stepName": "functional-tests",
                      "testenv": "functional"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:43Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: Ansible Test
            kind: AnsibleTest
            name: ansibletests.test.openstack.org
            version: v1beta1
          - displayName: Horizon Test
            kind: HorizonTest
            name: horizontests.test.openstack.org
            version: v1beta1
          - displayName: Tempest
            kind: Tempest
            name: tempests.test.openstack.org
            version: v1beta1
          - kind: Tobiko
            name: tobikoes.test.openstack.org
            version: v1beta1
        description: Test Operator
        displayName: OpenStack Test Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Tempest
        - Tobiko
        - Tests
        links:
        - name: Test Operator
          url: https://github.com/openstack-k8s-operators/test-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/test-rhel9-operator@sha256:17949ad6a69a14ea38870d471fd4f5a51122927f795976ac7e2b0189e2d0da5a
        - registry.redhat.io/rhoso/openstack-tempest-all-rhel9@sha256:2f07239d932a67a6589844f4380f613cc16614ff4de7699f64c17359bbdc5534
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: test-operator.v1.0.6
        version: 1.0.6
      - name: test-operator.v1.0.5
        version: 1.0.5
      - name: test-operator.v1.0.4
        version: 1.0.4
      - name: test-operator.v1.0.3
        version: 1.0.3
      - name: test-operator.v1.0.2
        version: 1.0.2
      - name: test-operator.v1.0.1
        version: 1.0.1
      - name: test-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: test-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhsso-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhsso-operator.7.5.1-opr-005
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm"
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "False"
          containerImage: registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:66577ec32a9a975f304e52b8ab49b9f7602158dcab1ef5d8c28a5d3455491336
          createdAt: "2021-08-09 00:00:00"
          description: An Operator for installing and managing Red Hat Single Sign-On
          olm.skipRange: <7.5.0
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Red Hat Single Sign-On Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Red Hat Single Sign-On.

          Red Hat Single Sign-On lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
        displayName: Red Hat Single Sign-On Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://access.redhat.com/products/red-hat-single-sign-on
        - name: Keycloak
          url: https://www.keycloak.org/
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat Customer Service
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso7-rhel8-init-container@sha256:614a9b1d7975769e8559e0b8c1578d16f6e3f7fee0778283797f4d62f5fc2100
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:5f0840a69642d13016430151892ec987fd67e711563b1dfb1a57353037c3bf79
        - registry.redhat.io/rh-sso-7/sso75-openj9-openshift-rhel8@sha256:2d4ea7226f43e7e8e8d3fcef6d425eb5104a15acf07791355033300b80db2901
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:66577ec32a9a975f304e52b8ab49b9f7602158dcab1ef5d8c28a5d3455491336
        version: 7.5.1-opr-005
      entries:
      - name: rhsso-operator.7.5.1-opr-005
        version: 7.5.1-opr-005
      - name: rhsso-operator.7.5.1-opr-004
        version: 7.5.1-opr-004
      - name: rhsso-operator.7.5.1-opr-003
        version: 7.5.1-opr-003
      - name: rhsso-operator.7.5.1-opr-002
        version: 7.5.1-opr-002
      - name: rhsso-operator.7.5.1-opr-001
        version: 7.5.1-opr-001
      - name: rhsso-operator.7.5.0
        version: 7.5.0
      name: alpha
    - currentCSV: rhsso-operator.7.6.12-opr-004
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm"
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "False"
          containerImage: registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:229265579ec77f2bd18c936ba19e1f9dbc6b2f0a13d963d5eae5b52c260010d1
          createdAt: "2021-08-09 00:00:00"
          description: An Operator for installing and managing Red Hat Single Sign-On
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Red Hat Single Sign-On Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Red Hat Single Sign-On.

          Red Hat Single Sign-On lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
        displayName: Red Hat Single Sign-On Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://access.redhat.com/products/red-hat-single-sign-on
        - name: Keycloak
          url: https://www.keycloak.org/
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat Customer Service
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:229265579ec77f2bd18c936ba19e1f9dbc6b2f0a13d963d5eae5b52c260010d1
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:923206112ed0873ee58edd0817dbce7bb4c9a89dc75d9b3ba2affe9aed4811de
        - registry.redhat.io/rh-sso-7/sso7-rhel8-init-container@sha256:282daf78934ca495ad4051371a58e3af0917bb9241d043a66e23fc8b12fbd1ce
        version: 7.6.12-opr-004
      entries:
      - name: rhsso-operator.7.6.12-opr-004
        version: 7.6.12-opr-004
      - name: rhsso-operator.7.6.12-opr-003
        version: 7.6.12-opr-003
      - name: rhsso-operator.7.6.12-opr-002
        version: 7.6.12-opr-002
      - name: rhsso-operator.7.6.12-opr-001
        version: 7.6.12-opr-001
      - name: rhsso-operator.7.6.11-opr-006
        version: 7.6.11-opr-006
      - name: rhsso-operator.7.6.11-opr-005
        version: 7.6.11-opr-005
      - name: rhsso-operator.7.6.11-opr-004
        version: 7.6.11-opr-004
      - name: rhsso-operator.7.6.11-opr-003
        version: 7.6.11-opr-003
      - name: rhsso-operator.7.6.11-opr-002
        version: 7.6.11-opr-002
      - name: rhsso-operator.7.6.11-opr-001
        version: 7.6.11-opr-001
      - name: rhsso-operator.7.6.10-opr-001
        version: 7.6.10-opr-001
      - name: rhsso-operator.7.6.9-opr-003
        version: 7.6.9-opr-003
      - name: rhsso-operator.7.6.9-opr-002
        version: 7.6.9-opr-002
      - name: rhsso-operator.7.6.9-opr-001
        version: 7.6.9-opr-001
      - name: rhsso-operator.7.6.8-opr-002
        version: 7.6.8-opr-002
      - name: rhsso-operator.7.6.8-opr-001
        version: 7.6.8-opr-001
      - name: rhsso-operator.7.6.7-opr-002
        version: 7.6.7-opr-002
      - name: rhsso-operator.7.6.7-opr-001
        version: 7.6.7-opr-001
      - name: rhsso-operator.7.6.6-opr-004
        version: 7.6.6-opr-004
      - name: rhsso-operator.7.6.6-opr-003
        version: 7.6.6-opr-003
      - name: rhsso-operator.7.6.6-opr-002
        version: 7.6.6-opr-002
      - name: rhsso-operator.7.6.6-opr-001
        version: 7.6.6-opr-001
      - name: rhsso-operator.7.6.5-opr-004
        version: 7.6.5-opr-004
      - name: rhsso-operator.7.6.5-opr-003
        version: 7.6.5-opr-003
      - name: rhsso-operator.7.6.5-opr-002
        version: 7.6.5-opr-002
      - name: rhsso-operator.7.6.5-opr-001
        version: 7.6.5-opr-001
      - name: rhsso-operator.7.6.4-opr-003
        version: 7.6.4-opr-003
      - name: rhsso-operator.7.6.4-opr-002
        version: 7.6.4-opr-002
      - name: rhsso-operator.7.6.4-opr-001
        version: 7.6.4-opr-001
      - name: rhsso-operator.7.6.3-opr-002
        version: 7.6.3-opr-002
      - name: rhsso-operator.7.6.3-opr-001
        version: 7.6.3-opr-001
      - name: rhsso-operator.7.6.2-opr-003
        version: 7.6.2-opr-003
      - name: rhsso-operator.7.6.2-opr-002
        version: 7.6.2-opr-002
      - name: rhsso-operator.7.6.2-opr-001
        version: 7.6.2-opr-001
      - name: rhsso-operator.7.6.1-opr-005
        version: 7.6.1-opr-005
      - name: rhsso-operator.7.6.1-opr-004
        version: 7.6.1-opr-004
      - name: rhsso-operator.7.6.1-opr-003
        version: 7.6.1-opr-003
      - name: rhsso-operator.7.6.1-opr-002
        version: 7.6.1-opr-002
      - name: rhsso-operator.7.6.1-opr-001
        version: 7.6.1-opr-001
      - name: rhsso-operator.7.6.0-opr-005
        version: 7.6.0-opr-005
      - name: rhsso-operator.7.6.0-opr-004
        version: 7.6.0-opr-004
      - name: rhsso-operator.7.6.0-opr-003
        version: 7.6.0-opr-003
      - name: rhsso-operator.7.6.0-opr-002
        version: 7.6.0-opr-002
      - name: rhsso-operator.7.6.0-opr-001
        version: 7.6.0-opr-001
      - name: rhsso-operator.7.5.2-opr-003
        version: 7.5.2-opr-003
      - name: rhsso-operator.7.5.2-opr-002
        version: 7.5.2-opr-002
      - name: rhsso-operator.7.5.2-opr-001
        version: 7.5.2-opr-001
      - name: rhsso-operator.7.5.1-opr-011
        version: 7.5.1-opr-011
      - name: rhsso-operator.7.5.1-opr-010
        version: 7.5.1-opr-010
      - name: rhsso-operator.7.5.1-opr-009
        version: 7.5.1-opr-009
      - name: rhsso-operator.7.5.1-opr-009-0.1646836600.p
        version: 7.5.1-opr-009+0.1646836600.p
      - name: rhsso-operator.7.5.1-opr-008
        version: 7.5.1-opr-008
      - name: rhsso-operator.7.5.1-opr-007
        version: 7.5.1-opr-007
      - name: rhsso-operator.7.5.1-opr-006
        version: 7.5.1-opr-006
      - name: rhsso-operator.7.5.1-opr-005
        version: 7.5.1-opr-005
      - name: rhsso-operator.7.5.1-opr-004
        version: 7.5.1-opr-004
      - name: rhsso-operator.7.5.1-opr-003
        version: 7.5.1-opr-003
      - name: rhsso-operator.7.5.1-opr-002
        version: 7.5.1-opr-002
      - name: rhsso-operator.7.5.1-opr-001
        version: 7.5.1-opr-001
      - name: rhsso-operator.7.5.0
        version: 7.5.0
      name: stable
    defaultChannel: stable
    packageName: rhsso-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-glue-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-glue-controller.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/glue-controller:0.2.1
          createdAt: "2025-09-19T20:55:49Z"
          description: AWS Glue controller is a service controller for managing Glue
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Job represents the state of an AWS glue Job resource.
            displayName: Job
            kind: Job
            name: jobs.glue.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Glue resources in AWS from within your Kubernetes cluster.

          **About Amazon Glue**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Glue
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - glue
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Glue Developer Resources
          url: https://aws.amazon.com/Glue/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: glue maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/glue-controller:0.2.1
        version: 0.2.1
      entries:
      - name: ack-glue-controller.v0.2.1
        version: 0.2.1
      - name: ack-glue-controller.v0.2.0
        version: 0.2.0
      - name: ack-glue-controller.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ack-glue-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Protocol Labs
      provider-url: ""
    name: ipfs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ipfs-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.ipfs.io/v1alpha1",
                "kind": "CircuitRelay",
                "metadata": {
                  "name": "circuitrelay-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cluster.ipfs.io/v1alpha1",
                "kind": "IpfsCluster",
                "metadata": {
                  "name": "ipfscluster-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Storage, Networking, Big Data
          containerImage: quay.io/redhat-et-ipfs/ipfs-operator:v0.0.1
          createdAt: 2022-04-05T15:50:37-0400
          operatorframework.io/suggested-namespace: ipfs-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ipfs-cluster/ipfs-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CircuitRelay is the Schema for the circuitrelays API.
            displayName: Circuit Relay
            kind: CircuitRelay
            name: circuitrelays.cluster.ipfs.io
            version: v1alpha1
          - description: IpfsCluster is the Schema for the ipfs API.
            displayName: Ipfs Cluster
            kind: IpfsCluster
            name: ipfsclusters.cluster.ipfs.io
            version: v1alpha1
        description: "The IPFS Operator enables users to easily create IPFS Cluster
          instances\nwithin their Kubernetes or OpenShift clusters. \n\n### Supported
          Features\n* Horizontal scaling\n* Joining collaborative clusters\n* Controlling
          IPFS node resource usage\n* Specifying reprovider settings  \n\n### Documentation
          \nDocumentation can be found on our [read-the-docs page](https://ipfs-operator.readthedocs.io/en/latest/),
          as well as our [GitHub repository](https://github.com/ipfs-cluster/ipfs-operator).\n\n###
          Getting Help\nIn case you run into issues while using the operator, please
          create an issue in our [GitHub\nrepository](https://github.com/ipfs-cluster/ipfs-operator).\n\n###
          Contributing\nWe welcome contributions of all types and sizes! You can find
          open issues\nin our [GitHub Repository](https://github.com/ipfs-cluster/ipfs-operator/issues),\nas
          well as make your own issue if you feel that something is missing.\n\n###
          License\nIPFS Operator is licensed under the [Apache 2.0 License](https://github.com/ipfs-cluster/ipfs-operator/blob/main/LICENSE)"
        displayName: IPFS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipfs
        - ipfs cluster
        - decentralized storage
        - peer-to-peer
        - distributed hash table
        - content-addressed
        links:
        - name: IPFS Operator GitHub page
          url: https://github.com/ipfs-cluster/ipfs-operator
        - name: IPFS Cluster
          url: https://ipfscluster.io/
        - name: IPFS
          url: https://ipfs.io/
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Protocol Labs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/redhat-et-ipfs/ipfs-operator:v0.0.1
        version: 0.0.2
      entries:
      - name: ipfs-operator.v0.0.2
        version: 0.0.2
      - name: ipfs-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ipfs-operator
    provider:
      name: Protocol Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: gaudi-ai-sw-tools-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gaudi-ai-sw-tools-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aitools.intel/v1",
                "kind": "GaudiAIToolsContainer",
                "metadata": {
                  "name": "intel-gaudi-1.17.0-495"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagespec": {
                    "registry": "registry.connect.redhat.com",
                    "repo": "intel/gaudi-notebooks",
                    "tags": [
                      {
                        "gaudi_software": "1.17.0-495",
                        "rhel_os": "9.2"
                      }
                    ]
                  },
                  "nameOverride": "",
                  "namespace": "redhat-ods-applications"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: registry.connect.redhat.com/intel/gaudi-ai-sw-tools-operator@sha256:0c286f1e4cec57a189d749501f82345afbe0a1dbbbf92b906ad3c549e47b89cd
          createdAt: "2024-09-11T21:27:02Z"
          description: Operator to enable Gaudi AI SW Tools into OpenShift AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: GaudiAIToolsContainer
            name: gaudiaitoolscontainers.aitools.intel
            version: v1
        description: Operator to enable Gaudi AI SW Tools into OpenShift AI
        displayName: Intel® Gaudi AI SW Tools Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gaudi
        - opendatahub
        - openshiftai
        - notebooks
        links:
        - name: Intel® Gaudi AI SW Tools Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/66d7aa630eb66a02febc8103
        maintainers:
        - email: sharvil.shah@intel.com
          name: sharvil
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.connect.redhat.com/intel/gaudi-ai-sw-tools-operator@sha256:0c286f1e4cec57a189d749501f82345afbe0a1dbbbf92b906ad3c549e47b89cd
        version: 0.0.2
      entries:
      - name: gaudi-ai-sw-tools-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: gaudi-ai-sw-tools-operator
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-certified-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "psp": false,
                  "serviceAccount": "default",
                  "controller": {
                    "enabled": true,
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": "registry.connect.redhat.com/neuvector/controller",
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": ""
                      }
                    },
                    "pvc": {
                      "enabled": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "path": "/",
                          "annotations": {
                            "ingress.kubernetes.io/protocol": "https"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "path": "/",
                          "annotations": {
                            "ingress.kubernetes.io/protocol": "https"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "path": "/",
                      "annotations": {
                        "ingress.kubernetes.io/protocol": "https"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {}
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/enforcer",
                    "priorityClassName": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      }
                    ],
                    "resources": {}
                  },
                  "manager": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/manager",
                    "priorityClassName": "",
                    "env": {
                      "ssl": true
                    },
                    "svc": {
                      "type": "NodePort",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "path": "/",
                      "annotations": {},
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "affinity": {},
                    "tolerations": [],
                    "nodeSelector": {}
                  },
                  "cve": {
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "image": "registry.access.redhat.com/ubi8@sha256:228824aa581f3b31bf79411f8448b798291c667a37155bdea61cfa128b2833f2",
                      "schedule": "0 0 * * *",
                      "priorityClassName": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": "registry.connect.redhat.com/neuvector/scanner@sha256:42004e13b348bb43b58ae8538c983a70b237e3aca870758ac629fb4fc43e8130",
                      "priorityClassName": "",
                      "resources": {},
                      "affinity": {},
                      "tolerations": [],
                      "nodeSelector": {}
                    }
                  },
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "resources": {},
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "true"
          containerImage: registry.connect.redhat.com/neuvector/neuvector-operator@sha256:6155ac3c95aa6471c9857e0a6f5e5d19e6fc1d8cc7d014108d7d4a10d14f96bb
          createdAt: "2020-12-02 14:19:21"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/neuvector-certified-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/neuvector-certified-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: Neuvector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm/tree/master/charts/core) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart.  NeuVector Operator versions are tied to NeuVector product versions. Version 1.3.4 of the NeuVector Certified Operator deploys version 4.4.3 of NeuVector.

          **Complete below steps to Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Grant Service Account Access to the Privileged SCC

                   oc -n neuvector adm policy add-scc-to-user privileged -z default

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s                    neuvector/default


          **Add NeuVector license from NeuVector WebUI->setting**


          #Deploying the NeuVector Operator#


          Please refer to the instructions [here](https://github.com/neuvector/neuvector-operator/blob/master/README.md)


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Networking
        - Security
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - registry.connect.redhat.com/neuvector/scanner@sha256:42004e13b348bb43b58ae8538c983a70b237e3aca870758ac629fb4fc43e8130
        - registry.access.redhat.com/ubi8@sha256:228824aa581f3b31bf79411f8448b798291c667a37155bdea61cfa128b2833f2
        - registry.connect.redhat.com/neuvector/neuvector-operator@sha256:6155ac3c95aa6471c9857e0a6f5e5d19e6fc1d8cc7d014108d7d4a10d14f96bb
        - registry.connect.redhat.com/neuvector/controller@sha256:acc943568520282061547b310c50080dc32924a531bbb543d959b281eb6664f4
        - registry.connect.redhat.com/neuvector/enforcer@sha256:be1d299ec3532d55444681a882e9fd1238949da6598987308e2800ed26291b7a
        - registry.connect.redhat.com/neuvector/manager@sha256:8245cd179d11fff279690c5a422f46b4ff252ed72b9b998f0c9b5453bc7d8fc0
        version: 1.3.4
      entries:
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-certified-operator-rhmp
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dpu-network-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dpu-network-operator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dpu.openshift.io/v1alpha1",
                "kind": "OVNKubeConfig",
                "metadata": {
                  "name": "ovnkubeconfig-sample"
                },
                "spec": {
                  "kubeConfigFile": "tenant-cluster-1-kubeconf",
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/dpu-worker": ""
                    }
                  },
                  "poolName": "dpu"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional, Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/dpu-network-rhel8-operator@sha256:d07b8ecb5d1625599878b5c1e34b0de94188b81d90c829787a2d13969f6d5971
          createdAt: "2023-07-13T18:03:11Z"
          description: The operator is responsible for the life-cycle management of
            the ovn-kube components and the necessary host network initialization
            on DPU cards.
          olm.skipRange: '>=4.10.0-0 <4.13.0-202309181427'
          operatorframework.io/suggested-namespace: openshift-dpu-network-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/dpu-network-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OVNKubeConfig is the Schema for the ovnkubeconfigs API
            displayName: OVNKube Config
            kind: OVNKubeConfig
            name: ovnkubeconfigs.dpu.openshift.io
            version: v1alpha1
        description: The operator to be responsible for the life-cycle management
          of the ovn-kube components and the necessary host network initialization
          on DPU cards.
        displayName: DPU Network Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dpu
        - network
        links:
        - name: Dpu Network Operator
          url: https://dpu-network-operator.domain
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:767682dd3bd65651a8204c9fb732b9c48b99127189992b43abc6ecce027f4589
        - registry.redhat.io/openshift4/dpu-network-rhel8-operator@sha256:d07b8ecb5d1625599878b5c1e34b0de94188b81d90c829787a2d13969f6d5971
        version: 4.13.0-202309181427
      entries:
      - name: dpu-network-operator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: stable
    defaultChannel: stable
    packageName: dpu-network-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-odlm
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operand-deployment-lifecycle-manager.v4.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandBindInfo",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "bindings": {
                    "public": {
                      "configmap": "mongodb-configmap",
                      "secret": "mongodb-secret"
                    }
                  },
                  "description": "Binding information that should be accessible to MongoDB adopters",
                  "operand": "mongodb-atlas-kubernetes",
                  "registry": "example-service"
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "services": [
                    {
                      "name": "jaeger",
                      "spec": {
                        "jaeger": {
                          "strategy": "allinone"
                        }
                      }
                    },
                    {
                      "name": "mongodb-atlas-kubernetes",
                      "spec": {
                        "atlasDeployment": {
                          "deploymentSpec": {
                            "name": "test-deployment"
                          },
                          "projectRef": {
                            "name": "my-project"
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "operators": [
                    {
                      "channel": "stable",
                      "installMode": "cluster",
                      "name": "jaeger",
                      "namespace": "default",
                      "packageName": "jaeger",
                      "sourceName": "community-operators",
                      "sourceNamespace": "openshift-marketplace"
                    },
                    {
                      "channel": "stable",
                      "name": "mongodb-atlas-kubernetes",
                      "namespace": "default",
                      "packageName": "mongodb-atlas-kubernetes",
                      "sourceName": "community-operators",
                      "sourceNamespace": "openshift-marketplace"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandRequest",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "requests": [
                    {
                      "operands": [
                        {
                          "name": "jaeger"
                        },
                        {
                          "name": "mongodb-atlas-kubernetes"
                        }
                      ],
                      "registry": "example-service"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, Monitoring, Logging & Tracing, Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
          createdAt: "2025-02-14T21:16:49Z"
          description: The Operand Deployment Lifecycle Manager provides a Kubernetes
            CRD-based API to manage the lifecycle of operands.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-odlm
          olm.skipRange: '>=1.2.0 <4.5.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/operand-deployment-lifecycle-manager
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OperandBindInfo is the Schema for the operandbindinfoes API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandBindInfo
            kind: OperandBindInfo
            name: operandbindinfos.operator.ibm.com
            version: v1alpha1
          - description: OperandConfig is the Schema for the operandconfigs API. Documentation
              For additional details regarding install parameters check https://ibm.biz/icpfs39install.
              License By installing this product you accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandConfig
            kind: OperandConfig
            name: operandconfigs.operator.ibm.com
            version: v1alpha1
          - description: OperandRegistry is the Schema for the operandregistries API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandRegistry
            kind: OperandRegistry
            name: operandregistries.operator.ibm.com
            version: v1alpha1
          - description: OperandRequest is the Schema for the operandrequests API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandRequest
            kind: OperandRequest
            name: operandrequests.operator.ibm.com
            version: v1alpha1
          - description: OperatorConfig is the Schema for the operatorconfigs API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.ibm.com
            version: v1alpha1
        description: |-
          # Introduction

          The operand is the instance managed by the operator. ODLM is used to manage the lifecycle of for a group of operands, compared with operator lifecycle manager, ODLM focus on the management of operands but not operators.

          # Details

          ### What are operands?

          Operator is a method of packaging, deploying and managing a Kubernetes application.
          Operands are the services and applications that Operator manage.

          For example, cert-manager operator deploys a cert-manager deployment, then the cert-manager-operator is an operator and the cert-manager deployment is an operand.

          The ODLM will have four CRDs:

          | Resource                 | Short Name | Description                                                                                |
          |--------------------------|------------|--------------------------------------------------------------------------------------------|
          | OperandRequest | opreq | It defines which operator/operand want to be installed in the cluster |
          | OperandRegistry | opreg | It defines the OLM information, like channel and catalog source, for each operator |
          | OperandConfig | opcon | It defines the parameters that should be used to install the operator's operand |
          | OperandBindInfo | opbi | It identifies secrets and/or configmaps that should be shared with requests |

          ### Goal

          1. A single entry point to manage a group of operands
          1. User can select a set of operands to install
          1. The install can be invoked through either OCP UI or CLI

          ## Prerequisites
          - [operator-sdk] version v1.3.0.
          - [go] version 1.15.7+
          - [oc] version v3.11+ or [kubectl] v1.11.3+
          - Access to an Openshift v4.3+ cluster

          ### Resources Required

          #### Minimum scheduling capacity

          | Software                   | Memory (MB) | CPU (cores) |
          | -------------------------- | ----------- | ----------- |
          | Operand Deployment Lifecycle Manager | 200          | 0.2        |
          | **Total**                  | 200         | 0.2         |

          ## Installing

          ### Create Operator NS, Group, Subscription

          ```yaml
          apiVersion: v1
          kind: Namespace
          metadata:
            name: odlm

          ---
          apiVersion: operators.coreos.com/v1alpha2
          kind: OperatorGroup
          metadata:
            name: operatorgroup
            namespace: odlm
          spec:
            targetNamespaces:
            - odlm

          ---
          apiVersion: v1
          data:
            namespaces: odlm
          kind: ConfigMap
          metadata:
            name: namespace-scope
            namespace: odlm

          ---
          apiVersion: operators.coreos.com/v1alpha1
          kind: Subscription
          metadata:
            name: operand-deployment-lifecycle-manager
            namespace: odlm
          spec:
            channel: v3.20
            name: ibm-odlm
            source: opencloud-operators
            sourceNamespace: olm
            config:
              env:
              - name: INSTALL_SCOPE
                value: namespaced
          END
          ```

          ## Configuration

          OperandRegistry defines the OLM information used for installation, like package name and catalog source, for each operator.

          Following is an example of the OperandRegistry CR:

          **NOTE:** The "spec.operators[*].name" parameter must be unique for each entry.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRegistry
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            operators:
            - name: jenkins [3]
              namespace: default [4]
              channel: alpha [5]
              packageName: jenkins-operator [6]
              scope: public [7]
              sourceName: community-operators [8]
              sourceNamespace: openshift-marketplace [9]
              installMode: cluster [10]
              installPlanApproval: Manual [11]
          ```

          The OperandRegistry Custom Resource (CR) lists OLM Operator information for operands that may be requested for installation and/or access by an application that runs in a namespace. The registry CR specifies:

          1. `name` of the OperandRegistry
          2. `namespace` of the OperandRegistry
          3. `name` is the name of the operator, which should be the same as the services name in the OperandConfig and OperandRequest.
          4. `namespace` defines the namespace where the operator and its CR will be deployed. (1) When InstallMode is `cluster`, the operator will be deployed into the `openshift-operators` namespace and the operator CRs will be deployed into the namespace this parameter defines. (2) When InstallMode is empty or set to `namespace`, it is the namespace where both operator and operator CR will be deployed.
          5. `channel` is the name of OLM channel that is subscribed for the operator.
          6. `packageName` is the name of the package in CatalogSource that is subscribed for the operator.
          7. (optional) `scope` is an indicator, either public or private, that dictates whether deployment can be requested from other namespaces (public) or only from the namespace of this OperandRegistry (private). The default value is private.
          8. `sourceName` is the name of the CatalogSource.
          9. `sourceNamespace` is the namespace of the CatalogSource.
          10. (optional) `installMode` is the install mode of the operator, can be either `namespace` (OLM one namespace) or `cluster` (OLM all namespaces). The default value is `namespace`. Operator is deployed in `openshift-operators` namespace when InstallMode is set to `cluster`.
          11. (optional) `installPlanApproval` is the approval mode for emitted installplan. The default value is `Automatic`.

          ### OperandConfig Spec

          OperandConfig defines the individual operand configuration. The OperandConfig Custom Resource (CR) defines the parameters for each operator that is listed in the OperandRegistry that should be used to install the operator instance by specifying an installation CR.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          Kind: OperandConfigs
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            services:
            - name: jenkins [3]
              spec: [4]
                jenkins:
                  port: 8081
          ```

          OperandConfig defines the individual operand deployment config:

          1. `name` of the OperandConfig
          2. `namespace` of the OperandConfig
          3. `name` is the name of the operator, which should be the same as the services name in the OperandRegistry and OperandRequest.
          4. `spec` defines a map. Its key is the kind name of the custom resource. Its value is merged to the spec field of custom resource. For more details, you can check the following topic **How does ODLM create the individual operator CR?**

          #### How does Operator create the individual operator CR

          Jenkins Operator has one CRD: Jenkins:

          The OperandConfig CR has

          ```yaml
          - name: jenkins
            spec:
              jenkins:
                service:
                  port: 8081
          ```

          The Jenkins Operator CSV has

          ```yaml
          apiVersion: operators.coreos.com/v1alpha1
          kind: ClusterServiceVersion
          metadata:
            annotations:
            alm-examples: |-
              [
              {
                "apiVersion":"jenkins.io/v1alpha2",
                "kind":"Jenkins",
                "metadata": {
                  "name":"example"
                },
                "spec":{
                  ...
                  "service":{"port":8080,"type":"ClusterIP"},
                  ...
                }
              }
            ]
          ```

          The ODLM will deep merge the OperandConfig CR spec and Jenkins Operator CSV alm-examples to create the Jenkins CR.

          ```yaml
          apiVersion: jenkins.io/v1alpha2
          kind: Jenkins
          metadata:
            name: example
          spec:
            ...
            service:
              port: 8081
              type: ClusterIP
            ...
          ```

          For day2 operations, the ODLM will patch the OperandConfigs CR spec to the existing Jenkins CR.

          ### OperandRequest Spec

          OperandRequest defines which operator/operand you want to install in the cluster.

          **NOTE:** OperandRequest custom resource is used to trigger a deployment for Operators and Operands.

          #### OperandRequest sample to create custom resource via OperandConfig

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRequest
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            requests:
            - registry: example-service [3]
              registryNamespace: example-service-ns [4]
              operands: [5]
              - name: jenkins [6]
                bindings: [7]
                  public:
                    secret: jenkins-operator-credential [8]
                    configmap: jenkins-operator-base-configuration [9]
          ```

          1. `name` of the OperandRequest
          2. `namespace` of the OperandRequest
          3. `registry` identifies the name of the OperandRegistry CR from which this operand deployment is being requested.
          4. (optional) `registryNamespace` identifies the namespace in which the OperandRegistry CR is defined. **Note:** If the `registryNamespace` is not specified then it is assumed that the OperandRegistry CR is in the current (OperandRequest's) namespace.
          5. `operands` in the CR is a list of operands.
          6. `name` of operands in the CR must match a name specification in an OperandRegistry's CR.
          7. (optional) The `bindings` of the operands is a map to get and rename the secret and/or configmap from the provider and create them in the requester's namespace. If the requester wants to rename the secret and/or configmap, they need to know the key of the binding in the OperandBindInfo. If the key of the bindings map is prefixed with public, it means the secret and/or configmap can be shared with the requester in the other namespace. If the key of the bindings map is prefixed with private, it means the secret and/or configmap can only be shared within its own namespace.
          8. (optional) `secret` names a secret that should be created in the requester's namespace with formatted data that can be used to interact with the service.
          9. (optional) `configmap` names a configmap that should be created in the requester's namespace with formatted data that can be used to interact with the service.

          #### OperandRequest sample to create custom resource via OperandRequest

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRequest
          metadata:
            name: example-service
            namespace: example-service-ns
          spec:
            requests:
            - registry: example-service
              registryNamespace: example-service-ns
              operands:
              - name: jenkins
                kind: Jenkins [1]
                apiVersion: "jenkins.io/v1alpha2" [2]
                instanceName: "example" [3]
                spec: [4]
                  service:
                    port: 8081
          ```

          1. `kind` of the target custom resource. If it is set in the operand item, ODLM will create CR according to OperandRequest only and get rid of the alm-example and OperandConfig.
          2. `apiVersion` of the target custom resource.
          3. `instanceName` is the name of the custom resource. If `instanceName` is not set, the name of the custom resource will be created with the name of the OperandRequest as a prefix.
          4. `spec` is the spec field of the target CR.

          ### OperandBindInfo Spec

          The ODLM will use the OperandBindInfo to copy the generated secret and/or configmap to a requester's namespace when a service is requested with the OperandRequest CR. An example specification for an OperandBindInfo CR is shown below.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandBindInfo
          metadata:
            name: publicjenkinsbinding [1]
            namespace: example-service-ns [2]
          spec:
            operand: jenkins [3]
            registry: example-service [4]
            description: "Binding information that should be accessible to jenkins adopters" [5]
            bindings: [6]
              public:
                secret: jenkins-operator-credentials-example [7]
                configmap: jenkins-operator-base-configuration-example [8]
          ```

          Fields in this CR are described below.

          1. `name` of the OperandBindInfo
          2. `namespace` of the OperandBindInfo
          3. The `operand` should be the the individual operator name.
          4. The `registry` section must match the name in the OperandRegistry CR in the current namespace.
          5. `description` is used to add a detailed description of a service.
          6. The `bindings` section is used to specify information about the access/configuration data that is to be shared. If the key of the bindings map is prefixed with public, it means the secret and/or configmap can be shared with the requester in the other namespace. If the key of the bindings map is prefixed with private, it means the secret and/or configmap can only be shared within its own namespace. If the key of the bindings map is prefixed with protected, it means the secret and/or configmap can only be shared if it is explicitly declared in the OperandRequest.
          7. The `secret` field names an existing secret, if any, that has been created and holds information that is to be shared with the requester.
          8. The `configmap` field identifies a configmap object, if any, that should be shared with the requester

          ODLM will use the OperandBindInfo CR to pass information to an adopter when they create a OperandRequest to access the service, assuming that both have compatible scopes. ODLM will copy the information from the shared service's "OperandBindInfo.bindinfo[].secret" and/or "OperandBindInfo.bindinfo[].configmap" to the requester namespace.

          **NOTE:** If in the OperandRequest, there is no secret and/or configmap name specified in the bindings or no bindings field in the element of operands, ODLM will copy the secret and/or configmap to the requester's namespace and rename them to the name of the OperandBindInfo + secret/configmap name.

          ### E2E Use Case

          1. User installs ODLM from OLM

              The ODLM will automatically generate two default CRD CRs, since OperandRegistry and OperandConfigs CRs don't define the state, so it should be fine.

          2. **Optional:** Users can update the OperandConfigs CR with their own parameter values

              Users can skip this step if the default configurations match the requirement.

          3. User creates the OperandRequest CR from OLM

              This tells ODLM that users want to install some of the individual operator/operand.

          4. The rest works will be done by OLM and ODLM

              Finally, users will get what they want.

          ### Operator/Operand Upgrade

          - For operator/operand upgrade, you only need to publish your operator OLM to your operator channel, and OLM will handle the upgrade automatically.
          - If there are major version, then you may want to update `channel` in `OperandRegistry` to trigger upgrade.

          ### Limitations

          ODLM depends on NamespaceScope Operator

          ### SecurityContextConstraints Requirements

          The Operand Deployment Lifecycle Manager supports running under the OpenShift Container Platform default restricted security context constraints.
        displayName: Operand Deployment Lifecycle Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Operand Deployment Lifecycle Manager Project
          url: https://github.com/IBM/operand-deployment-lifecycle-manager
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d4dff1a8968c18d984e16c4c9f1517c3a10e23a11bf166c64d734773324d4c9d
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
        - registry.marketplace.redhat.com/cpopen/operand-deployment-lifecycle-manager-bundle@sha256:68f71f2767a40ca9333f0a670e07493e8a8c13a5c7ca877ea5bd8a544b352d9a
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:27195e9758afdf8521971866fd60ea3dafed91a146c3c4165876e5ea476a4622
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bb074bdf89cbc7c703fadb17639608599ab4f66db9be5617e5f79a0550676790
        version: 4.5.0
      entries:
      - name: operand-deployment-lifecycle-manager.v4.5.0
        version: 4.5.0
      name: v4.5
    defaultChannel: v4.5
    packageName: ibm-odlm
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fujitsu
      provider-url: https://www.postgresql.fastware.com/
    name: fujitsu-enterprise-postgres-operator-with-cg-module
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPAction",
                "metadata": {
                  "name": "new-fep-action"
                },
                "spec": {
                  "fepAction": {
                    "args": [
                      "new-fep-sts-0",
                      "new-fep-sts-1"
                    ],
                    "type": "reload"
                  },
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "targetClusterName": "new-fep"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPExporter",
                "metadata": {
                  "name": "new-fep-exporter"
                },
                "spec": {
                  "fepExporter": {
                    "exporterLogLevel": "error",
                    "fepClusterList": [
                      "new-fep1"
                    ],
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "userCustomQueries": "usr_example:\n  query: \"SELECT EXTRACT(EPOCH FROM (now() - pg_last_xact_replay_timestamp())) as lag\"\n  master: true\n  metrics:\n    - lag:\n        usage: \"GAUGE\"\n        description: \"Replication lag behind master in seconds\""
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPLogging",
                "metadata": {
                  "name": "new-fep-logging"
                },
                "spec": {
                  "fepLogging": {
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPPgpool2",
                "metadata": {
                  "name": "new-fep-pgpool2"
                },
                "spec": {
                  "clientAuthMethod": "scram",
                  "count": 2,
                  "customhba": "local   all         all                               trust\nhost    all         all         127.0.0.1/32          trust\nhost    all         all         ::1/128               trust\n",
                  "customlogsize": "128Mi",
                  "customparams": "listen_addresses = '*'\npcp_listen_addresses = '*'\nnum_init_children = 32\nreserved_connections = 0\nenable_pool_hba = off\nallow_clear_text_frontend_auth = off\nauthentication_timeout = 80\nbackend_weight0 = 1\nbackend_weight1 = 1\nbackend_flag0 = 'ALWAYS_PRIMARY'\nbackend_flag1 = 'DISALLOW_TO_FAILOVER'\nconnection_cache = on\nmax_pool = 4\nlisten_backlog_multiplier = 2\nserialize_accept = off\nchild_life_time = 300\nclient_idle_limit = 0\nchild_max_connections = 0\nconnection_life_time = 0\nreset_query_list = 'ABORT; DISCARD ALL'\nclient_min_messages = info\nlog_min_messages = debug1\nlog_statement = on\nlog_per_node_statement = on\nlog_client_messages = on\nlog_hostname = on\nlog_connections = on\nlog_line_prefix = '%t: pid %p: '\nload_balance_mode = on\nignore_leading_white_space = on\nwhite_function_list = ''\nblack_function_list = 'currval,lastval,nextval,setval'\nblack_query_pattern_list = ''\ndatabase_redirect_preference_list = ''\napp_name_redirect_preference_list = ''\nallow_sql_comments = off\ndisable_load_balance_on_write = 'transaction'\nstatement_level_load_balance = on\nsr_check_period = 0\nsr_check_user = 'postgres'\ndelay_threshold = 0\nlog_standby_delay = 'none'\nssl = on\nssl_ciphers = 'HIGH:MEDIUM:+3DES:!aNULL'\nssl_prefer_server_ciphers = off\nssl_ecdh_curve = 'prime256v1'\nssl_dh_params_file = ''\nrelcache_expire = 0\nrelcache_size = 256\ncheck_temp_table = catalog\ncheck_unlogged_table = on\nenable_shared_relcache = off\nrelcache_query_target = primary\nwd_port0 = 9000\nfailover_on_backend_error = off\n",
                  "custompcp": "none",
                  "customsslcacert": "none",
                  "customsslcert": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                  "customsslkey": "none",
                  "fepclustername": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "limits": {
                    "cpu": "400m",
                    "memory": "512Mi"
                  },
                  "requests": {
                    "cpu": "200m",
                    "memory": "256Mi"
                  },
                  "serviceport": 9999,
                  "statusport": 9898
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPRestore",
                "metadata": {
                  "name": "new-fep-restore"
                },
                "spec": {
                  "fromFEPcluster": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "mcSpec": {
                    "limits": {
                      "cpu": "200m",
                      "memory": "300Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "restoretype": "latest",
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "toFEPcluster": "new-fep-2"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v2",
                "kind": "FEPCluster",
                "metadata": {
                  "name": "new-fep"
                },
                "spec": {
                  "fep": {
                    "customAnnotations": {
                      "allDeployments": {}
                    },
                    "forceSsl": true,
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "instances": 1,
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "podAntiAffinity": false,
                    "podDisruptionBudget": false,
                    "servicePort": 27500,
                    "syncMode": "off",
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "usePodName": true
                  },
                  "fepChildCrVal": {
                    "backup": {
                      "image": {
                        "pullPolicy": "IfNotPresent"
                      },
                      "mcSpec": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      },
                      "pgbackrestParams": "# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[global]\nrepo1-retention-full=7\nrepo1-retention-full-type=time\nlog-path=/database/log/backup\n",
                      "postScript": " ",
                      "preScript": " ",
                      "schedule": {
                        "num": 2
                      },
                      "schedule1": {
                        "schedule": "15 0 * * 0",
                        "type": "full"
                      },
                      "schedule2": {
                        "schedule": "15 0 * * 1-6",
                        "type": "incr"
                      },
                      "schedule3": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule4": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule5": {
                        "schedule": " ",
                        "type": " "
                      }
                    },
                    "customPgAudit": "# define pg audit custom params here to override defaults.\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[output]\nlogger = 'auditlog'\nlog_directory = '/database/log/audit'\nlog_truncate_on_rotation = on\nlog_filename = 'pgaudit-%a.log'\nlog_rotation_age = 1d\nlog_rotation_size = 0\n[rule]\n",
                    "customPgHba": "# define pg_hba custom rules here to be merged with default rules.\n# TYPE     DATABASE        USER        ADDRESS        METHOD\n",
                    "customPgParams": "# define custom postgresql.conf parameters below to override defaults.\n# Current values are as per default FEP deployment\n# If you add a library to shared_preload_libraries, add it after the default value.\n# fsep_operator_security is available in FEP15 and later container images.\nshared_preload_libraries='pgx_datamasking,pg_prewarm,pg_stat_statements,fsep_operator_security'\nsession_preload_libraries='pg_prewarm'\nmax_prepared_transactions = 100\nmax_worker_processes = 30\nmax_connections = 100\nwork_mem = 1MB\nmaintenance_work_mem = 12MB\nshared_buffers = 128MB\neffective_cache_size = 384MB\ncheckpoint_completion_target = 0.8\n# tcp parameters\ntcp_keepalives_idle = 30\ntcp_keepalives_interval = 10\ntcp_keepalives_count = 3\n# logging parameters in default fep installation\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\nlog_directory = '/database/log'\nlog_filename = 'logfile-%a.log'\nlog_file_mode = 0600\nlog_truncate_on_rotation = on\nlog_rotation_age = 1d\nlog_rotation_size = 0\nlog_checkpoints = on\nlog_line_prefix = '%e %t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h'\nlog_lock_waits = on\nlog_autovacuum_min_duration = 60s\nlogging_collector = on\npgaudit.config_file='/opt/app-root/src/pgaudit-cfg/pgaudit.conf'\nlog_replication_commands = on\nlog_min_messages = WARNING\nlog_destination = csvlog\n# wal_archive parameters in default fep installation\narchive_mode = on\narchive_command = 'pgbackrest --stanza=backupstanza --config=/database/userdata/pgbackrest.conf archive-push %p'\nwal_level = replica\nmax_wal_senders = 12\nwal_keep_segments = 64\ntrack_activities = on\ntrack_counts = on\npassword_encryption = 'scram-sha-256'\n",
                    "storage": {
                      "archivewalVol": {
                        "size": "1Gi"
                      },
                      "backupVol": {
                        "size": "2Gi"
                      },
                      "dataVol": {
                        "size": "2Gi"
                      },
                      "logVol": {
                        "size": "1Gi"
                      },
                      "tablespaceVol": {
                        "size": "512Mi"
                      },
                      "walVol": {
                        "size": "1200Mi"
                      }
                    },
                    "sysUsers": {
                      "pgRewindPassword": "rewind_password",
                      "pgRewindUser": "rewind_user",
                      "pgdb": "mydb",
                      "pgpassword": "mydbpassword",
                      "pgreplpassword": "repluserpwd",
                      "pgrepluser": "repluser",
                      "pguser": "mydbuser",
                      "tdepassphrase": "tde-passphrase"
                    },
                    "systemCertificates": {
                      "cacrt": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                      "crt": "-----BEGIN CERTIFICATE-----\nMIIDUTCCAjmgAwIBAgIRAMocW3qMoHrD6qRvMPppMkMwDQYJKoZIhvcNAQELBQAw\nNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9yIEt1\nYmVybmV0ZXMwHhcNMjEwMjA2MDQzMjM2WhcNMjYwMjA1MDQzMjM2WjA/MRAwDgYD\nVQQKEwdGdWppdHN1MSswKQYDVQQDEyJGVUpJVFNVIEVudGVycHJpc2UgUG9zdGdy\nZXMgU2VydmVyMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA4AI33yvH\nZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I2e4SceTKi6O3C/I1XuvWlpng\n5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4ANB5JWWqDOjrRT3o7nRPGXfila\nbP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYpmjdbfxabTz07ig0+6/cwKoRR\nxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTFYvmAH7gcdssSFBt8NPlUATHE\nsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6Wdgmu5H2pDml8CDNLDv98Aj7i\n+I5SRKKcVPlnuQIDAQABo1AwTjAdBgNVHSUEFjAUBggrBgEFBQcDAQYIKwYBBQUH\nAwIwDAYDVR0TAQH/BAIwADAfBgNVHSMEGDAWgBQcwrrUO0u+FhIUuVdrDRCQRsi6\nZjANBgkqhkiG9w0BAQsFAAOCAQEAm5dxBoI9pScOCvRAchg4CprdRDSJb9K6yB3O\nnCAxnM47iHeXnY3WlnI388kHu8DU7O4ba1tJbGs3KY9KzioPk43pU12jWkO1onoF\n+mTDjx/Ef1cYWA9r5q/LtgTa6Q2sxV4O2x67QW82aAnaxO34dV5zWCPIvAoovZBV\nHRT+BgCg3r2vD1RGKK2nl1aYJtWhO1SZubam+VttdZ/vbM9oOJctxmImsEtBXjkY\nKteePdQtLL5o03JhyXWyRshCq+HMmKf2KgyY8gvydGcP4eLQdBWcW40LcnVq6UjT\n0kJycJEKngMVademq1ZWHGaiYB7hyT6GhgIcHUJ2cKrPgbEh1Q==\n-----END CERTIFICATE-----",
                      "key": "-----BEGIN RSA PRIVATE KEY-----\nMIIEowIBAAKCAQEA4AI33yvHZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I\n2e4SceTKi6O3C/I1XuvWlpng5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4AN\nB5JWWqDOjrRT3o7nRPGXfilabP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYp\nmjdbfxabTz07ig0+6/cwKoRRxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTF\nYvmAH7gcdssSFBt8NPlUATHEsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6W\ndgmu5H2pDml8CDNLDv98Aj7i+I5SRKKcVPlnuQIDAQABAoIBAFPQYKlOzw/+BA0b\nyMIUpdctIMb/54CR/xR0mVw1DbSjigNVPjHUQvB8Y1B2FAITQObgJO06bAv0QdWN\nRb0/v/yYiNJDFjaLjaIAHlO/2+oWrXbFaZqgpVDJhB+e1xaZr2x7XGxm+p925k30\nl6pvIRY+I8JRKvZiV1VZHwL/R3JOtPr++xMZtLVjVOI+f+ySqJ+TZHuAjm49EKxj\ncEmmJ28b7QcziXsvKy00f+zbqLIBKXQdZAFU5eEr1BsDRXdRW+Kf0XIvftuy4BJZ\nvoKT+VGhEvF/qysswL4+6IAO6tpuYnnM0Y2d3sOGoWPkTcQK0MekYKzL/WmtCjNs\n9hodJtECgYEA5EWyhEOf4uOKe5TDp697UCUvXLoOR58FDe/S8XNvScn29jjOkqIg\nOMoqo9xAkJTNTzqn5UUdt1x/pgM2NxlPLFijrc0zQlX3SoOO2ryDd9WNi7YKtN16\nKJqa536WeZu2OEbuAZ+S3GALVy1RPeTNPnUOmKnF06DjDUGzLNCZy10CgYEA+zfw\n952DWuz1U0Z4wvAEqqcgUKXPKrkTXV/iUnjkDkrLYVr0ZofDNTXrdHl+UedFmaOC\ncieZn6DNhcdz5tKtyysGMH3g/qs9PfoGUngvcXsy0Egk04l3x1jc8TTCLqXZXYaQ\nHMsx51n+R58oncPtzYSUOr9qQ6PbC2CstTbFJA0CgYEAjGEsUliAB/jknfEzjXjG\nPdhQUxb8VyE864Az2lah9t/kJzFyIAziAeqZ5GE7t247AGFTBRTHHI8e1Qoemi3P\nWbc9GVIbFs1lIYbcIDpUIyrKPEP8O5QEXtoNLxXTFgAjRGKiVY87spjCAJ+W2ZhO\ne/1it5GYXfgQCYQA2yuBmOUCgYANRkR2YR1axaCk+NlSu6oTdmdPu6M5x7PNQE7O\nOtMaKjua9lppvIzFGAdMDUtueoEEAE7ZR1xnwfB6PDLUpJdIYAqgr1YfPt8qkjaZ\nTv56yZ7CwL0pbF8m6nwqRrZoDp1wwraEvvvxFKFKGY/k3kCHlpTakdjEoDjn3gDi\nRnWeVQKBgCEneMSzucei5LRppRtRaJw/Btll8qlPMlX3W7dxQ3cLwpmLOn0m51Fp\nPIZ44zYK8R6fu4+/sSrlfaIg86Ugeufp6YNxyNROKxUGza5vDIu5OftwWtBeg+UK\nZ8lLWNdX6pp7WMujmF3H1DrkBbauYMUKZ4UxUYtelgHERMePIxwb\n-----END RSA PRIVATE KEY-----"
                    }
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "false"
          containerImage: quay.io/fujitsu/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:eaf601e7aef525ebcd4b769c859ed74c7a8077f843b3a75d76713eabf0295745
          createdAt: "2025-09-11T09:32:52Z"
          description: OpenShift Operator for Fujitsu Enterprise Postgres
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.1.0 <7.1.6'
          operators.openshift.io/valid-subscription: '["Fujitsu Enterprise Postgres
            Advanced Edition with Cryptographic Module Annual Subscription License
            per Core 17 Operator for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/internal-objects: '["fepautoscales.fep.fujitsu.io","fepconfigs.fep.fujitsu.io","fepusers.fep.fujitsu.io","fepcerts.fep.fujitsu.io","fepvolumes.fep.fujitsu.io","fepbackups.fep.fujitsu.io"]'
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: fj-dbaas-tec@dl.jp.fujitsu.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: FEPAction
            name: fepactions.fep.fujitsu.io
            version: v1
          - kind: FEPAutoscale
            name: fepautoscales.fep.fujitsu.io
            version: v1
          - kind: FEPBackup
            name: fepbackups.fep.fujitsu.io
            version: v1
          - kind: FEPCert
            name: fepcerts.fep.fujitsu.io
            version: v1
          - kind: FEPCluster
            name: fepclusters.fep.fujitsu.io
            version: v2
          - kind: FEPConfig
            name: fepconfigs.fep.fujitsu.io
            version: v1
          - kind: FEPExporter
            name: fepexporters.fep.fujitsu.io
            version: v1
          - kind: FEPLogging
            name: feploggings.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2Cert
            name: feppgpool2certs.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2
            name: feppgpool2s.fep.fujitsu.io
            version: v1
          - kind: FEPRestore
            name: feprestores.fep.fujitsu.io
            version: v1
          - kind: FEPUpgrade
            name: fepupgrades.fep.fujitsu.io
            version: v1
          - kind: FEPUser
            name: fepusers.fep.fujitsu.io
            version: v1
          - kind: FEPVolume
            name: fepvolumes.fep.fujitsu.io
            version: v1
        description: "Fujitsu Enterprise Postgres 17 delivers an enterprise-grade
          PostgreSQL on OpenShift Container Platform.\n\nThis solution provides the
          flexibility of a hybrid cloud solution while delivering an enhanced distribution
          of PostgreSQL to support enterprise-level workloads and provide improved
          deployment and management, availability, performance, data governance and
          security.  \n\nAvailable as a multi-architecture container built for both
          amd64.\n\nThe download and Use of the Product is strictly subject to the
          terms of the End User License Agreement with Fujitsu Limited found at https://www.fast.fujitsu.com/fujitsu-enterprise-postgres-license-agreements.
          Where the Product that has been embedded as a whole or part into a third
          party program, only Authorised Customers may download and use the Product.\n"
        displayName: Fujitsu Enterprise Postgres Operator with Cryptographic Module
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - Fujitsu
        - fep
        links:
        - name: Fujitsu Enterprise Postgres
          url: https://www.postgresql.fastware.com/
        maintainers:
        - email: fj-dbaas-tec@dl.jp.fujitsu.com
          name: Fujitsu
        maturity: stable
        provider:
          name: Fujitsu
          url: https://www.postgresql.fastware.com/
        relatedImages:
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-restore-with-cryptographic-module@sha256:edeb7b72a89e05995661e6a17e75b5b477a3ca351e0db146a939a33e88e16c97
        - quay.io/fujitsu/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:eaf601e7aef525ebcd4b769c859ed74c7a8077f843b3a75d76713eabf0295745
        - quay.io/fujitsu/fujitsu-enterprise-postgres-utils-with-cryptographic-module@sha256:7d169df0b8fe47000891afde79be8dd7978da3f893d1e92f01f57fd8c0b3b4c6
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-pgpool2-with-cryptographic-module@sha256:e510bc2d30df33e4c9d3babb898ff9cbd13704fa3fa69a284bf672765fb03083
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-server-with-cryptographic-module@sha256:7778f4b03861b56371d18f7726b68ad21d0faa7c1a25cb38fe7cac483d81c095
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-restore-with-cryptographic-module@sha256:6c0d6e10cfe1868be2986fdde14c6f55f86b12cea01a82199b1e2e0c3154ef75
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-upgrade-with-cryptographic-module@sha256:b8be0344b6cf99b2c5de2b16f04f6f2b05bde861d9b38afd319bbbc4dd72a620
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-upgrade-with-cryptographic-module@sha256:caf6e1d2cc31d01958e4adcb4c3c307d761f350d5b0963ce19214f041c14184a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-restore-with-cryptographic-module@sha256:98a3b0bfa6ec7a9a4a08fb524c331c3de6a9b580d6386a7b5430955ed84b2fc5
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-backup-with-cryptographic-module@sha256:dd15cf3b6b4a5c959efef8b602abe5d64b4f5a424fb6a3e55079a2040e67cd7a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-backup-with-cryptographic-module@sha256:03b7f3c76e3eb775e4ec7f66f7506ee1feeab56dca96e868ac3f38b9c1ca5567
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentd-with-cryptographic-module@sha256:f743f35012e6fd3d54a6b54baeeee98bd7c65fe52923acd87546066fa9b54d7e
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-upgrade-with-cryptographic-module@sha256:1fe690a9c04a55213062ad3bef8c0cff73849f8777c3945b9a8ae2dcff6330fa
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-utils-with-cryptographic-module@sha256:8b2b6a981c863f7075ced1235c614317b6526626b42405295a9c664736f0a80f
        - quay.io/fujitsu/fujitsu-enterprise-postgres-cronjob-with-cryptographic-module@sha256:809a64600396378ff8c825595d40f72aedbc2b57f394420548262b085a3f2ea5
        - quay.io/fujitsu/fujitsu-enterprise-postgres-exporter-with-cryptographic-module@sha256:71e1d1e715e2d89757e80d90b8d5b11baa729110cd5f7898df920a4c85f2bf8a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentbit-with-cryptographic-module@sha256:bb661b3aad0b825d56a9696d6d2d9f5c04add2a83e793fc3d4a73980f4a7a0ac
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-utils-with-cryptographic-module@sha256:4caf1dc5211ac03893ce9a62b417f412ffd411f13569f3ac75e12ca5128af9e4
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-pgpool2-with-cryptographic-module@sha256:a4c5a59ca18f6fa295b78d570f03991f1a4b4bd6df52f8526a1254ac1a567990
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-pgpool2-with-cryptographic-module@sha256:e8aa27cbb57503ec5f4f37beeb2e833ace08d4c583d4361b7ce945c7461a904c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-server-with-cryptographic-module@sha256:cad4e36dc2b1e5ca6bf9308d9f93958d4c8f2bbdc14e000ca0b5700fa335df44
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-backup-with-cryptographic-module@sha256:5b49aa46386ca954ebcbea28fbabd15929cb64fc2a4d2dd2796c9d2e15b212c6
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-server-with-cryptographic-module@sha256:a012f21f63729ba667af097375e08874156e2cf1b94ae203985d45f11031fea9
        version: 7.1.6
      entries:
      - name: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.6
        version: 7.1.6
      - name: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.5
        version: 7.1.5
      name: stable
    defaultChannel: stable
    packageName: fujitsu-enterprise-postgres-operator-with-cg-module
    provider:
      name: Fujitsu
      url: https://www.postgresql.fastware.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-05-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx8-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx8-operator-rhmp/support?utm_source=openshift_console
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:154f85cc6721a70909dee755ddd88897ad51038afe2538b6dee4d26e7d8b9afb
        version: 0.0.10
      entries:
      - name: joget-dx8-openshift-operator.v0.0.10
        version: 0.0.10
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-operator-rhmp
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://www.redhat.com/
    name: rhdh
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhdh-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
          createdAt: "2025-09-12T06:25:11Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.7.1'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.7/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:62f32e50727c5006766a34daa5b6472b649cd9894f3fe9543b8ecc67e6760e8e
        version: 1.7.1
      entries:
      - name: rhdh-operator.v1.7.1
        version: 1.7.1
      - name: rhdh-operator.v1.7.0
        version: 1.7.0
      - name: rhdh-operator.v1.6.5
        version: 1.6.5
      - name: rhdh-operator.v1.6.4
        version: 1.6.4
      - name: rhdh-operator.v1.6.3
        version: 1.6.3
      - name: rhdh-operator.v1.6.2
        version: 1.6.2
      - name: rhdh-operator.v1.6.1
        version: 1.6.1
      - name: rhdh-operator.v1.6.0
        version: 1.6.0
      - name: rhdh-operator.v1.5.3
        version: 1.5.3
      - name: rhdh-operator.v1.5.2
        version: 1.5.2
      - name: rhdh-operator.v1.5.1
        version: 1.5.1
      - name: rhdh-operator.v1.4.3
        version: 1.4.3
      - name: rhdh-operator.v1.4.2
        version: 1.4.2
      - name: rhdh-operator.v1.4.1
        version: 1.4.1
      - name: rhdh-operator.v1.4.0
        version: 1.4.0
      - name: rhdh-operator.v1.3.5
        version: 1.3.5
      - name: rhdh-operator.v1.3.4
        version: 1.3.4
      - name: rhdh-operator.v1.3.3
        version: 1.3.3
      - name: rhdh-operator.v1.3.1
        version: 1.3.1
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast
    - currentCSV: rhdh-operator.v1.1.2-0.1714688890.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:9bea7eabdea44342248dfd6091a8f5c3b6e65884c916b3c2a073a4a64481aa6f
          createdAt: "2023-03-20T16:11:34Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins, configuration settings, and deployment
            mechanisms, which can help streamline the process of setting up a self-managed
            internal developer portal for adopters who are just starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: rhdh-operator.v1.1.2
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.1.0 <1.2.0'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/janus-idp/operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: asoro@redhat.com
          name: Armel Soro
        - email: gazarenk@redhat.com
          name: Gennady Azarenkov
        - email: jianrzha@redhat.com
          name: Jianrong Zhang
        - email: nboldt@redhat.com
          name: Nick Boldt
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:95ed239eb427c7b1e422ece28c5277b5018efb751c2b472bf502b8216219c047
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:9bea7eabdea44342248dfd6091a8f5c3b6e65884c916b3c2a073a4a64481aa6f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:d7ae2456857cab09a9e9acf69c9990ef18deb7f920046b8c2483fbfd0d42bbfc
        - registry.redhat.io/rhel9/postgresql-15@sha256:d5f3d496ecd55688e384cc0627700a97af6924582123bcf00749c375724af3de
        version: 1.1.2+0.1714688890.p
      entries:
      - name: rhdh-operator.v1.1.2
        version: 1.1.2
      - name: rhdh-operator.v1.1.2-0.1714608881.p
        version: 1.1.2+0.1714608881.p
      - name: rhdh-operator.v1.1.2-0.1714688618.p
        version: 1.1.2+0.1714688618.p
      - name: rhdh-operator.v1.1.2-0.1714688890.p
        version: 1.1.2+0.1714688890.p
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.1
    - currentCSV: rhdh-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:c9051a6c339cd40cbc7474a33b32a6b94913af964e12ca5da0c0b7e1c6f444d5
          createdAt: "2024-11-28T17:55:19Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.2.6'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.2/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.2/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: asoro@redhat.com
          name: Armel Soro
        - email: cdaley@redhat.com
          name: Corey Daley
        - email: gazarenk@redhat.com
          name: Gennady Azarenkov
        - email: nboldt@redhat.com
          name: Nick Boldt
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:ef29f2b340a6dc52c5014239876d54a54f7c0f0bf4ea917391438a69c94a8532
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:c9051a6c339cd40cbc7474a33b32a6b94913af964e12ca5da0c0b7e1c6f444d5
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:025eb58c69b9cf5992ecf2fccb29a3d36061cbcf62652a1c6878f0c4d767c1fb
        - registry.redhat.io/rhel9/postgresql-15@sha256:24fb4e7914a6e1464d015be9e5582cc4b9da224137408bd429e7ea4f391aa198
        version: 1.2.6
      entries:
      - name: rhdh-operator.v1.2.6
        version: 1.2.6
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.2
    - currentCSV: rhdh-operator.v1.3.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:5ce579edc10bde5c1c400104f7e0d9f86ca277cd565aaf3dd84b8636f5b949a8
          createdAt: "2025-02-10T18:12:52Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.3.5'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.3/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:5ce579edc10bde5c1c400104f7e0d9f86ca277cd565aaf3dd84b8636f5b949a8
        - registry.redhat.io/rhel9/postgresql-15@sha256:24fb4e7914a6e1464d015be9e5582cc4b9da224137408bd429e7ea4f391aa198
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:c044fed8485a62a5051ac726525da546936be22f268db05491bf81582e0e6194
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b4934c45ba595b36e65c73b734c58d1517e7f43b7cfc309ca8bc853733582f76
        version: 1.3.5
      entries:
      - name: rhdh-operator.v1.3.5
        version: 1.3.5
      - name: rhdh-operator.v1.3.4
        version: 1.3.4
      - name: rhdh-operator.v1.3.3
        version: 1.3.3
      - name: rhdh-operator.v1.3.1
        version: 1.3.1
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.3
    - currentCSV: rhdh-operator.v1.4.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:ae5b52c1b35e82f3c564402e367a2a2028e16806867bba0868643089ba7ed223
          createdAt: "2025-04-03T05:42:32Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.4.3'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.4/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:ae5b52c1b35e82f3c564402e367a2a2028e16806867bba0868643089ba7ed223
        version: 1.4.3
      entries:
      - name: rhdh-operator.v1.4.3
        version: 1.4.3
      - name: rhdh-operator.v1.4.2
        version: 1.4.2
      - name: rhdh-operator.v1.4.1
        version: 1.4.1
      - name: rhdh-operator.v1.4.0
        version: 1.4.0
      name: fast-1.4
    - currentCSV: rhdh-operator.v1.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:8fe167cc649a1c53994c376f3cf3eed8c6e318741a0c08e5aaf07cb0b74e7f8a
          createdAt: "2025-06-05T14:31:21Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.5.3'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.5/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:8fe167cc649a1c53994c376f3cf3eed8c6e318741a0c08e5aaf07cb0b74e7f8a
        - registry.redhat.io/rhel9/postgresql-15@sha256:74377f5a61dd7e3e0ba5d2fa880b7107025433a026396567c79eac927b4212bc
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:2ad8c87e82df27c2b6ceba94179d1f0a40fe775368663c1b5572611888e0e02b
        version: 1.5.3
      entries:
      - name: rhdh-operator.v1.5.3
        version: 1.5.3
      - name: rhdh-operator.v1.5.2
        version: 1.5.2
      - name: rhdh-operator.v1.5.1
        version: 1.5.1
      name: fast-1.5
    - currentCSV: rhdh-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:50f20300c81621221260d03d22b1442750e3f1d2e62f3e511d63925d84c4d1f9
          createdAt: "2025-08-27T18:43:12Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.6.5'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.6/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:50f20300c81621221260d03d22b1442750e3f1d2e62f3e511d63925d84c4d1f9
        - registry.redhat.io/rhel9/postgresql-15@sha256:5e225c1646251d2509b3a6e45fad23abfb11dd5c45d3224b59178f90deb38a70
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:177a7308f61a6b1ba21bc0f7bf550a93d4478e52479e2c94a2437f2152e3d8ff
        version: 1.6.5
      entries:
      - name: rhdh-operator.v1.6.5
        version: 1.6.5
      - name: rhdh-operator.v1.6.4
        version: 1.6.4
      - name: rhdh-operator.v1.6.3
        version: 1.6.3
      - name: rhdh-operator.v1.6.2
        version: 1.6.2
      - name: rhdh-operator.v1.6.1
        version: 1.6.1
      - name: rhdh-operator.v1.6.0
        version: 1.6.0
      name: fast-1.6
    - currentCSV: rhdh-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
          createdAt: "2025-09-12T06:25:11Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.7.1'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.7/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:62f32e50727c5006766a34daa5b6472b649cd9894f3fe9543b8ecc67e6760e8e
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
        version: 1.7.1
      entries:
      - name: rhdh-operator.v1.7.1
        version: 1.7.1
      - name: rhdh-operator.v1.7.0
        version: 1.7.0
      name: fast-1.7
    defaultChannel: fast
    packageName: rhdh
    provider:
      name: Red Hat Inc.
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-activemqartemis: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/
    name: amq-broker-rhel8
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-broker-operator.v7.10.7-opr-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                    "name": "ex-aao",
                    "application": "ex-aao-app"
                },
                "spec": {
                    "deploymentPlan": {
                        "image": "placeholder",
                        "size": 2,
                        "requireLogin": false,
                        "persistenceEnabled": false,
                        "journalType": "nio",
                        "messageMigration": false,
                        "jolokiaAgentEnabled": false,
                        "managementRBACEnabled": true
                    }
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                    "name": "ex-aaoaddress"
                },
                "spec": {
                    "addressName": "myAddress0",
                    "queueName": "myQueue0",
                    "routingType": "anycast"
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                    "name": "ex-aaoscaledown"
                },
                "spec": {
                    "localOnly": "true"
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                    "name": "ex-keycloak"
                },
                "spec": {
                    "loginModules": {
                        "propertiesLoginModules": [
                            {
                                "name": "prop-module",
                                "users": [
                                    {
                                        "name": "superman",
                                        "roles": [
                                            "root"
                                        ]
                                    }
                                ]
                            }
                        ]
                    },
                    "securityDomains": {
                        "brokerDomain": {
                            "name": "activemq",
                            "loginModules": [
                                {
                                    "name": "prop-module",
                                    "flag": "sufficient"
                                }
                            ]
                        }
                    }
                }
            }]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:1a7aa54d2799d238eb5f49f7a95a78a896f6bf8d222567e9118e0e3963cc9aad
          createdAt: "2021-06-18T00:02:04Z"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A stateful deployment of one or more brokers
            displayName: AMQ Broker
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: Adding and removing addresses via custom resource definitions
            displayName: AMQ Broker Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: Provides message migration on clustered broker scaledown
            displayName: AMQ Broker Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: Security configuration for the broker
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
          to deploy and manage stateful AMQ Broker broker clusters
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: rkieley@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:1a7aa54d2799d238eb5f49f7a95a78a896f6bf8d222567e9118e0e3963cc9aad
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ba21779c29b431a22c824e91820705c9501fb0b7d6f7e1ae22fcd3becc8c8a2f
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:d4010af53429502f41d722c39d69a7bb147bb06d19e3d4a69de70dbe88e3990e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b575d6eebf44dc9049445e357488efb17a690953852419b534df30e0105b9ca4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ff6fefad518a6c997d4c5a6e475ba89640260167f0bc27715daf3cc30116fad1
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:56282c6a461e945a83bffb13d3f3d8d9621800f0fc9cd755a75d21eacd8f5fbf
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:979b59325aa0f34eb05625201beba53fccbb83bd5eb80a89dcb5261ae358138f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:83faefa67b33d93836a1be3cb438bb147242d602565dcbe59b4ed42b26072191
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:967d840b831b310abd3a468a8652d1fa48b2ce28b342bfab53999245deb004f1
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:2cb14c917246ffcbd5d861ca6f3d58cb1146a4e55715ee08160e115d3dfc1c3e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f37f98c809c6f29a83e3d5a3ac4494e28efe9b25d33c54f533c6a08662244622
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:4e09ee3bd8ff3479fdd00e221b7b1505ae04ca6c64fd6bc9e2f1a34b8e98d137
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0d857b95084e39c011651be89482d4dcc77c2d1c3ac3c6a819f3c9d5a57a425f
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:9465e398a55c89c121b3ae50374414190a991839fa5e1018fc02551f6aa3b9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3c00b524722b500ff0922fe7a8722868557512183a8f69248f37a68c45c6a770
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a94fa8135d3765363e442bda45df6e02a3e22bee7ea198bcb0d3584331a4320c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c5f4c08e068b9721967cf7c7cbd9a9e93fb5e39b264dd13b653e99d8f3fa9e0e
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:aa7af7e038a12b5202108b362c8af243954603e84034a16c0192684c86d5d28b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b74d03ed852a3731467ffda95266ce49f2065972f1c37bf254f3d52b34c11991
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:7abfdcfd7b2d3d12b9234f7b057bd199bceefa47bdf02f62236f82a13fdbf5b7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:869cda365447b0b7b53ed21d2496884c15dc9c07295a9dfa0337c2c9974fe018
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:5ff28e3062d3eabb9f52d8271b837fdab06cafbd8b8f732b1bc00de49ad69aad
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f58678f03a250dbb296ebeb112c9150340155799ba4dad9e90bf347b9ed0ca7b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3219f90789690a3c5cfcc86e5f602c443e75f1fdb3a19660f43ee3996bfa8f62
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:f2651cb097eb631fc6af90e4a8404d36516aa2205c546125bf0c37739ab525fa
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:35b7d96d5065f280a528626abd2c3df84e79ccea93ab6bd8b8837af66c330f30
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:45a5e054546a2c1ee338518f8477f728f5f7108c9dd8b516b4c79d4a7d6f5ba7
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:1a9f13fd565113055b2aa2022bd0709a63c1dc058a8570e0fd7eba5c7b101334
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:97c3aa5673610875a53571449f7d5210ae8d675072d044f17b416725ed0dba88
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:ba0f5df7da08bc030ff5e0e68ecb52a931c25f6b67c6556aa8a9661c82d69d42
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:83ed89fa942ccbf4451954956db42138c83f9b015e7de07fc159ddc4bb6dc843
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9a02decbc460ddfc501d8a4a9b41cd084776090951906b027838c0829e1fc15a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ff819e913406911735da39c51699405f747bd697116a273a00897ff176bf05ff
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:5d39f26f06769fe5c989ea0e0ee8f9166120e152ce2c9bc3336fc8d7cab2ff92
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:e6f71001dccfbbe1cd7cf4fded69e181c10f46cc5c8bd3a8d3045a3f3ac41e2f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:61e5a8c1e46f1c18648e0406e2a57c52f911979079020cc57c7a4b06a56940bb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:e03945756d86bd189e8c33027fa668d2065c168510085dfd06defd9f40d3e0ab
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:34739366153b825dcb113e0ed6115ef5656085cf8d47492789ff17464042b576
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:34ab1f66d8daa544e3823464383e9788e307665c875499838fbed95f3a74a4dd
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0b2126cfb6054fdf428c1f43b69e36e93a09a49ce15350e9273c98cc08c6598b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a8b2a9364fd06c8e1c29b8f8f28d3e4844bf7840a62b67137b663c558982d4a3
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:6e9e0e3fddc44fbf87591d8904176407f1ec409d578a7d108c13cdd48d85131b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:67ae74e72927557c55611d0248347b2099b117ac1aa245dac05856fe782fb59c
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:c3ac4602b48867f3d4f7cc0efd5c9f133098e409626cfb402f80386ff460dbf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c9eba0d1f9fba588711774f4db75221f057578407065127ed60c33eef564f499
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:86ff145bd68e797a0507495570063b54cbe4186fcb4acd83c900406619ef9524
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:d4a9176f49ffd9db6051fbda15d47a1aa32c282fb5010d0caf5b9243e80bf055
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:4440b20349032bb5c0bb2a8deb89832b37767af6f0b3208b067867a85641913e
        version: 7.10.7-opr-1
      entries:
      - name: amq-broker-operator.v7.10.7-opr-1
        version: 7.10.7-opr-1
      - name: amq-broker-operator.v7.10.6-opr-1
        version: 7.10.6-opr-1
      - name: amq-broker-operator.v7.10.5-opr-1
        version: 7.10.5-opr-1
      - name: amq-broker-operator.v7.10.4-opr-1
        version: 7.10.4-opr-1
      - name: amq-broker-operator.v7.10.3-opr-3
        version: 7.10.3-opr-3
      - name: amq-broker-operator.v7.10.3-opr-2
        version: 7.10.3-opr-2
      - name: amq-broker-operator.v7.10.3-opr-1
        version: 7.10.3-opr-1
      - name: amq-broker-operator.v7.10.2-opr-2-0.1676475747.p
        version: 7.10.2-opr-2+0.1676475747.p
      - name: amq-broker-operator.v7.10.2-opr-2-0.1680622941.p
        version: 7.10.2-opr-2+0.1680622941.p
      - name: amq-broker-operator.v7.10.2-opr-2
        version: 7.10.2-opr-2
      - name: amq-broker-operator.v7.10.2-opr-1
        version: 7.10.2-opr-1
      - name: amq-broker-operator.v7.10.1-opr-2
        version: 7.10.1-opr-2
      - name: amq-broker-operator.v7.10.1-opr-1
        version: 7.10.1-opr-1
      - name: amq-broker-operator.v7.10.0-opr-4
        version: 7.10.0-opr-4
      - name: amq-broker-operator.v7.10.0-opr-3
        version: 7.10.0-opr-3
      - name: amq-broker-operator.v7.10.0-opr-2
        version: 7.10.0-opr-2
      - name: amq-broker-operator.v7.10.0-opr-1
        version: 7.10.0-opr-1
      name: 7.10.x
    - currentCSV: amq-broker-operator.v7.11.8-opr-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisSecurity is the Schema for the activemqartemissecurities
              API
            displayName: Active MQArtemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:b9aecae6b0ce0b241c7a5b9e6e096395940bb6a384373a4d121822cf10c26d99
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b74d03ed852a3731467ffda95266ce49f2065972f1c37bf254f3d52b34c11991
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:442b6638133d42fa7faa41d86e47be95cdca56c2e16efcd6e7e22991c3812a36
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f37f98c809c6f29a83e3d5a3ac4494e28efe9b25d33c54f533c6a08662244622
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0b2126cfb6054fdf428c1f43b69e36e93a09a49ce15350e9273c98cc08c6598b
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:4440b20349032bb5c0bb2a8deb89832b37767af6f0b3208b067867a85641913e
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:7abfdcfd7b2d3d12b9234f7b057bd199bceefa47bdf02f62236f82a13fdbf5b7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b9f93c243ead118651083880786a81c624f390fcc3596c7ec5ebd55f34f95ea3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:35b7d96d5065f280a528626abd2c3df84e79ccea93ab6bd8b8837af66c330f30
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b575d6eebf44dc9049445e357488efb17a690953852419b534df30e0105b9ca4
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:f2651cb097eb631fc6af90e4a8404d36516aa2205c546125bf0c37739ab525fa
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:83ed89fa942ccbf4451954956db42138c83f9b015e7de07fc159ddc4bb6dc843
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:967d840b831b310abd3a468a8652d1fa48b2ce28b342bfab53999245deb004f1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0d857b95084e39c011651be89482d4dcc77c2d1c3ac3c6a819f3c9d5a57a425f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:e03945756d86bd189e8c33027fa668d2065c168510085dfd06defd9f40d3e0ab
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:56282c6a461e945a83bffb13d3f3d8d9621800f0fc9cd755a75d21eacd8f5fbf
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5e5d8b551ab8ca377e0bc939037970a2f31ee14e9f555853ebea2759d673895a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9a02decbc460ddfc501d8a4a9b41cd084776090951906b027838c0829e1fc15a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0f443f5aa818dd85bfa93e51e37bdac3909628a283aa52a89125b527dd75d9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:5d39f26f06769fe5c989ea0e0ee8f9166120e152ce2c9bc3336fc8d7cab2ff92
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:9465e398a55c89c121b3ae50374414190a991839fa5e1018fc02551f6aa3b9aa
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:1a9f13fd565113055b2aa2022bd0709a63c1dc058a8570e0fd7eba5c7b101334
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:45a5e054546a2c1ee338518f8477f728f5f7108c9dd8b516b4c79d4a7d6f5ba7
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c9eba0d1f9fba588711774f4db75221f057578407065127ed60c33eef564f499
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a94fa8135d3765363e442bda45df6e02a3e22bee7ea198bcb0d3584331a4320c
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:34ab1f66d8daa544e3823464383e9788e307665c875499838fbed95f3a74a4dd
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:86ff145bd68e797a0507495570063b54cbe4186fcb4acd83c900406619ef9524
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c5f4c08e068b9721967cf7c7cbd9a9e93fb5e39b264dd13b653e99d8f3fa9e0e
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:d4a9176f49ffd9db6051fbda15d47a1aa32c282fb5010d0caf5b9243e80bf055
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:29f07006ad62023c20d34971cb3ae900349efae5ab2918aaa8e5647d788032be
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ff819e913406911735da39c51699405f747bd697116a273a00897ff176bf05ff
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b0b3684ef1ebf8b44d3a4c46391984173792dde1bd7363ea06d43a8f95a717d0
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:cc438b4bc2b14f6bff4e5d32e7dc0fad6d2ca391c4e84b5ae925b7b40af76729
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ba21779c29b431a22c824e91820705c9501fb0b7d6f7e1ae22fcd3becc8c8a2f
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:ba0f5df7da08bc030ff5e0e68ecb52a931c25f6b67c6556aa8a9661c82d69d42
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:e6f71001dccfbbe1cd7cf4fded69e181c10f46cc5c8bd3a8d3045a3f3ac41e2f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:67ae74e72927557c55611d0248347b2099b117ac1aa245dac05856fe782fb59c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:979b59325aa0f34eb05625201beba53fccbb83bd5eb80a89dcb5261ae358138f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ade37271426ab8f543d72f8da3e9f663d665f9427379a2539fc9323f6897cc44
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:34739366153b825dcb113e0ed6115ef5656085cf8d47492789ff17464042b576
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:869cda365447b0b7b53ed21d2496884c15dc9c07295a9dfa0337c2c9974fe018
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ff6fefad518a6c997d4c5a6e475ba89640260167f0bc27715daf3cc30116fad1
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:61e5a8c1e46f1c18648e0406e2a57c52f911979079020cc57c7a4b06a56940bb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0dde839c2dce7cb684094bf26523c8e16677de03149a0fff468b8c3f106e1f4f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5f842ebe9a6a4294a056ad10189a5f277cfd3fb4e6e0b3225e194bd5c131151d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3c00b524722b500ff0922fe7a8722868557512183a8f69248f37a68c45c6a770
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:4e09ee3bd8ff3479fdd00e221b7b1505ae04ca6c64fd6bc9e2f1a34b8e98d137
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:d4010af53429502f41d722c39d69a7bb147bb06d19e3d4a69de70dbe88e3990e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a8b2a9364fd06c8e1c29b8f8f28d3e4844bf7840a62b67137b663c558982d4a3
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:5ff28e3062d3eabb9f52d8271b837fdab06cafbd8b8f732b1bc00de49ad69aad
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:d2ae3defac36f00112c6b773e13626336ba15d6f50f905ad1c33f122bfe1a6a9
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:97c3aa5673610875a53571449f7d5210ae8d675072d044f17b416725ed0dba88
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f58678f03a250dbb296ebeb112c9150340155799ba4dad9e90bf347b9ed0ca7b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:40ccaf2bb2e0f8132f2541b39cb835f17b38050b3ca9232f0861fd5fcaf4324a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:6e9e0e3fddc44fbf87591d8904176407f1ec409d578a7d108c13cdd48d85131b
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:c3ac4602b48867f3d4f7cc0efd5c9f133098e409626cfb402f80386ff460dbf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:137b893ad15e0cd5353bae134c8fae833257a12641b12d0f774f1565ce66f63c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:89bb769b4c33eb931cc7866eeabec9a315b913851cfa84bc6e11ae8d1e8b2cbb
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:2cb14c917246ffcbd5d861ca6f3d58cb1146a4e55715ee08160e115d3dfc1c3e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:25c3732e945b4bc3e9aca721ce709d0217339b065d9dac9afdb6202e9fd8a799
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:8b4c4400a02c6cacfd978147b158e2ccbe5da955578b4b970a5ee7edac865326
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3219f90789690a3c5cfcc86e5f602c443e75f1fdb3a19660f43ee3996bfa8f62
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:7d53744ae8c77aea09c59165486c9ed0eaa131140f6b346042293bce0bd8a745
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:aa7af7e038a12b5202108b362c8af243954603e84034a16c0192684c86d5d28b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9c1ac38bf8eec67a3da57c3cc61105513776a749408cb287fe6448a17889de4e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:83faefa67b33d93836a1be3cb438bb147242d602565dcbe59b4ed42b26072191
        version: 7.11.8-opr-1
      entries:
      - name: amq-broker-operator.v7.11.8-opr-1
        version: 7.11.8-opr-1
      - name: amq-broker-operator.v7.11.7-opr-2-1742200979
        version: 7.11.7-opr-2-1742200979
      - name: amq-broker-operator.v7.11.7-opr-2-1740583644
        version: 7.11.7-opr-2-1740583644
      - name: amq-broker-operator.v7.11.7-opr-2-1736956309
        version: 7.11.7-opr-2-1736956309
      - name: amq-broker-operator.v7.11.7-opr-2-1731330819
        version: 7.11.7-opr-2-1731330819
      - name: amq-broker-operator.v7.11.7-opr-2
        version: 7.11.7-opr-2
      - name: amq-broker-operator.v7.11.7-opr-1-1726157430
        version: 7.11.7-opr-1-1726157430
      - name: amq-broker-operator.v7.11.7-opr-1-1723204800
        version: 7.11.7-opr-1-1723204800
      - name: amq-broker-operator.v7.11.7-opr-1
        version: 7.11.7-opr-1
      - name: amq-broker-operator.v7.11.6-opr-2
        version: 7.11.6-opr-2
      - name: amq-broker-operator.v7.11.5-opr-1
        version: 7.11.5-opr-1
      - name: amq-broker-operator.v7.11.4-opr-1
        version: 7.11.4-opr-1
      - name: amq-broker-operator.v7.11.3-opr-1
        version: 7.11.3-opr-1
      - name: amq-broker-operator.v7.11.2-opr-1
        version: 7.11.2-opr-1
      - name: amq-broker-operator.v7.11.1-opr-2
        version: 7.11.1-opr-2
      - name: amq-broker-operator.v7.11.1-opr-1
        version: 7.11.1-opr-1
      - name: amq-broker-operator.v7.11.0-opr-3
        version: 7.11.0-opr-3
      - name: amq-broker-operator.v7.11.0-opr-2
        version: 7.11.0-opr-2
      name: 7.11.x
    - currentCSV: amq-broker-operator.v7.12.5-opr-1-0.1757928550.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-broker-operator.v7.12.5-opr-1-0.1756216162.p
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:7d53744ae8c77aea09c59165486c9ed0eaa131140f6b346042293bce0bd8a745
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9c1ac38bf8eec67a3da57c3cc61105513776a749408cb287fe6448a17889de4e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:29f07006ad62023c20d34971cb3ae900349efae5ab2918aaa8e5647d788032be
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:dc13871dfed52402c40fd192900fc3cb5a49da4fa08e4ec4d0e23a879e83db25
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5e5d8b551ab8ca377e0bc939037970a2f31ee14e9f555853ebea2759d673895a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:cc438b4bc2b14f6bff4e5d32e7dc0fad6d2ca391c4e84b5ae925b7b40af76729
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d71506472636d940b5dbd3e708d3e0448d777eeecfd38de61ba49e281d2979c6
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ffffa9875f0379e9373f89f05eb06e5a193273bb04bc3aa5f85b044357b79098
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:b9aecae6b0ce0b241c7a5b9e6e096395940bb6a384373a4d121822cf10c26d99
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:442b6638133d42fa7faa41d86e47be95cdca56c2e16efcd6e7e22991c3812a36
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:25c3732e945b4bc3e9aca721ce709d0217339b065d9dac9afdb6202e9fd8a799
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ade37271426ab8f543d72f8da3e9f663d665f9427379a2539fc9323f6897cc44
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:16253895bc113f395dc6ab5cd511901e43b6ee676369ec80d598a8741c3c9bf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:89bb769b4c33eb931cc7866eeabec9a315b913851cfa84bc6e11ae8d1e8b2cbb
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:bb6fbd68475a7852b4d99eea6c4ab313f9267da7963162f0d75375d7063409e7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b9f93c243ead118651083880786a81c624f390fcc3596c7ec5ebd55f34f95ea3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9cc48eecf1442ae04b8543fa5d4381a13bc2831390850828834d387006d1342b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:47fd4ce2533496828aba37bd1f9715e2164d5c90bd0fc6b25e7e0786d723bf01
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:023478b773a0c42739b86e822a4d435e01ef341b8136566ea52463f1b7e447ed
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:eacc4fb55e8c4c30894a51944960d536508001a49fda62782c6cb2ca8767961d
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0dde839c2dce7cb684094bf26523c8e16677de03149a0fff468b8c3f106e1f4f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:8b4c4400a02c6cacfd978147b158e2ccbe5da955578b4b970a5ee7edac865326
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d42d713da0ce6806fdc6492b6342586783e6865a82a8647d3c4288439b1751ee
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:7f9b921d89b6d4d91b365659afff127d99f1455b94b48a10a6a7f4c38dd8b700
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:90f1678c19c74454b77aed14fa6b3074a1c6f80fe884f1c6136e7e907f402aec
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b2402179b1d0ff9c950defe5d8808195ef9a4c3e13e3c469cae31ab51d5f5014
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:8329acdff170c5a8ae6ae16f6c37d3f7ab68d106a6e8be50c3fe131ec5ecceeb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5f842ebe9a6a4294a056ad10189a5f277cfd3fb4e6e0b3225e194bd5c131151d
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:40920687402fd747ca02d384e63692a5e1db625723a2bf3056a607c92ee8eee1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0f443f5aa818dd85bfa93e51e37bdac3909628a283aa52a89125b527dd75d9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:40ccaf2bb2e0f8132f2541b39cb835f17b38050b3ca9232f0861fd5fcaf4324a
        version: 7.12.5-opr-1+0.1757928550.p
      entries:
      - name: amq-broker-operator.v7.12.5-opr-1-0.1757928550.p
        version: 7.12.5-opr-1+0.1757928550.p
      - name: amq-broker-operator.v7.12.5-opr-1-0.1756216162.p
        version: 7.12.5-opr-1+0.1756216162.p
      - name: amq-broker-operator.v7.12.5-opr-1-0.1754272113.p
        version: 7.12.5-opr-1+0.1754272113.p
      - name: amq-broker-operator.v7.12.5-opr-1
        version: 7.12.5-opr-1
      - name: amq-broker-operator.v7.12.4-opr-1
        version: 7.12.4-opr-1
      - name: amq-broker-operator.v7.12.4-opr-1-0.1747217191.p
        version: 7.12.4-opr-1+0.1747217191.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741881887.p
        version: 7.12.3-opr-1+0.1741881887.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739991235.p
        version: 7.12.3-opr-1+0.1739991235.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1730142550.p
        version: 7.12.3-opr-1+0.1730142550.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1730813578.p
        version: 7.12.3-opr-1+0.1730813578.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1731483798.p
        version: 7.12.3-opr-1+0.1731483798.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1732654016.p
        version: 7.12.3-opr-1+0.1732654016.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1733315503.p
        version: 7.12.3-opr-1+0.1733315503.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1737643631.p
        version: 7.12.3-opr-1+0.1737643631.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1738596517.p
        version: 7.12.3-opr-1+0.1738596517.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739390944.p
        version: 7.12.3-opr-1+0.1739390944.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739889884.p
        version: 7.12.3-opr-1+0.1739889884.p
      - name: amq-broker-operator.v7.12.3-opr-1
        version: 7.12.3-opr-1
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741884937.p
        version: 7.12.3-opr-1+0.1741884937.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741883943.p
        version: 7.12.3-opr-1+0.1741883943.p
      - name: amq-broker-operator.v7.12.2-opr-1
        version: 7.12.2-opr-1
      - name: amq-broker-operator.v7.12.1-opr-1-0.1724054375.p
        version: 7.12.1-opr-1+0.1724054375.p
      - name: amq-broker-operator.v7.12.1-opr-1-0.1723669176.p
        version: 7.12.1-opr-1+0.1723669176.p
      - name: amq-broker-operator.v7.12.1-opr-1-0.1721247907.p
        version: 7.12.1-opr-1+0.1721247907.p
      - name: amq-broker-operator.v7.12.1-opr-1
        version: 7.12.1-opr-1
      - name: amq-broker-operator.v7.12.0-opr-2
        version: 7.12.0-opr-2
      - name: amq-broker-operator.v7.12.0-opr-1
        version: 7.12.0-opr-1
      name: 7.12.x
    defaultChannel: 7.12.x
    packageName: amq-broker-rhel8
    provider:
      name: Red Hat
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecrpublic-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecrpublic-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecrpublic-controller:1.1.1
          createdAt: "2025-09-20T01:21:52Z"
          description: AWS ECR PUBLIC controller is a service controller for managing
            ECR PUBLIC resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Repository represents the state of an AWS ecrpublic Repository
              resource.
            displayName: Repository
            kind: Repository
            name: repositories.ecrpublic.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ECR PUBLIC resources in AWS from within your Kubernetes cluster.

          **About Amazon ECR PUBLIC**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECR PUBLIC
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecrpublic
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECR PUBLIC Developer Resources
          url: https://aws.amazon.com/ECR PUBLIC/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecrpublic maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecrpublic-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ecrpublic-controller.v1.1.1
        version: 1.1.1
      - name: ack-ecrpublic-controller.v1.1.0
        version: 1.1.0
      - name: ack-ecrpublic-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecrpublic-controller.v1.0.12
        version: 1.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecrpublic-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: egressip-ipam-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: egressip-ipam-operator.v1.2.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "EgressIPAM",
                "metadata": {
                  "name": "egressipam-aws"
                },
                "spec": {
                  "cidrAssignments": [
                    {
                      "CIDR": "10.0.128.0/19",
                      "labelValue": "us-west-2a"
                    },
                    {
                      "CIDR": "10.0.160.0/19",
                      "labelValue": "us-west-2b",
                      "reservedIPs": [
                        "10.0.160.126"
                      ]
                    },
                    {
                      "CIDR": "10.0.192.0/19",
                      "labelValue": "us-west-2c"
                    }
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "topologyLabel": "failure-domain.beta.kubernetes.io/zone"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "EgressIPAM",
                "metadata": {
                  "name": "egressipam-baremetal"
                },
                "spec": {
                  "cidrAssignments": [
                    {
                      "CIDR": "192.169.0.0/24",
                      "labelValue": "true"
                    }
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "topologyLabel": "egressGateway"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Networking, Cloud Provider, Security
          certified: "false"
          containerImage: quay.io/redhat-cop/egressip-ipam-operator@sha256:d6129bd287b87696a23e0da64dcd598958f779dc88ae33b46b3c909dbab80adf
          createdAt: "2022-07-25T15:07:22Z"
          description: This operator provides automation around managing egressIPs.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: egressip-ipam-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/egressip-ipam-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgressIPAM is the Schema for the egressipams API
            displayName: Egress IPAM
            kind: EgressIPAM
            name: egressipams.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator automates the assignment of egressIPs to namespaces.
          Namespaces can opt in to receiving one or more egressIPs with the following annotation `egressip-ipam-operator.redhat-cop.io/egressipam:<egressIPAM>`, where `egressIPAM` is the CRD that controls how egressIPs are assigned.
          IPs assigned to the namespace can be looked up in the following annotation: `egressip-ipam-operator.redhat-cop.io/egressips`.

          EgressIP assignments is managed by the EgressIPAM CRD. Here is an example of it:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: EgressIPAM
          metadata:
            name: example-egressipam
          spec:
            cidrAssignments:
              - labelValue: "true"
                CIDR: 192.169.0.0/24
                reservedIPs:
                - "192.159.0.5"
            topologyLabel: egressGateway
            nodeSelector:
              node-role.kubernetes.io/worker: ""
          ```

          This EgressIPAM specifies that all nodes that comply with the specified node selector and that also have labels `egressGateway: "true"` will be assigned egressIP from the specified CIDR.

          Note that the `cidrAssigments` field is an array and therefore, multiple groups of nodes can be identified with the labelValue and different CIDRs can be assigned to them. This is usually not necessary on a bare metal deployment.

          In the bare metal scenario, when this egressCRD is created, all the `hostsubnet` relative to the nodes selected by this EgressIPAM will be update to have the EgressCIDRs field equal to the specified CIDR (see below for cloud provider scenarios).

          When a namespace is created with the opt-in annotation: `egressip-ipam-operator.redhat-cop.io/egressipam=<egressIPAM>`, an available egressIP is selected from the CIDR and assigned to the namespace.The `netnamespace` associated with this namespace is updated to use that egressIP.

          It is possible to specify a set of reserved IPs. These IPs must belong to the CIDR and will never be assigned.

          ## Passing EgressIPs as input

          The normal mode of operation of this operator is to pick a random IP from the configured CIDR. However, it also supports a scenario where egressIPs are picked by an external process and passed as input.

          In this case IPs must me passed as an annotation to the namespace, like this: `egressip-ipam-operator.redhat-cop.io/egressips=IP1,IP2...`. The value of the annotation is a comma separated array of ip with no spaces.

          There must be exactly one IP per CIDR defined in the referenced egressIPAM. Moreover, each IP must belong to the corresponding CIDR. Because this situation can lead to inconsistencies, failing to have correspondence between IPs in the namespace annotation and CIDRs in the egressIPAM CR will cause the operator to error out and stop processing all the namespaces associated with the given EgressIPAM CR.

          It also also a responsibility of the progress picking the IPs to ensure that those IPs are actually available.

          ## Assumptions

          The following assumptions apply when using this operator:

          1. If multiple EgressIPAMs are defined, their selected nodes MUST NOT overlap.
          2. When using a cloud provider the topology label MUST be `failure-domain.beta.kubernetes.io/zone`.

          ## Support for AWS

          In AWS as well as other cloud providers, one cannot freely assign IPs to machines. Additional steps need to be performed in this case. Considering this EgressIPAM

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: EgressIPAM
          metadata:
            name: egressipam-aws
          spec:
            cidrAssignments:
              - labelValue: "eu-central-1a"
                CIDR: 10.0.128.0/20
              - labelValue: "eu-central-1b"
                CIDR: 10.0.144.0/20
              - labelValue: "eu-central-1c"
                CIDR: 10.0.160.0/20
            topologyLabel: failure-domain.beta.kubernetes.io/zone
            nodeSelector:
              matchLabels:
                node-role.kubernetes.io/worker: ""
          ```

          When a namespace with the opt-in annotation is created, the following happens:

          1. for each of the CIDRs, an available IP is selected and assigned to the namespace
          2. the relative `netnamespace` is update to reflect the assignment (multiple IPs will be assigned in this case).
          3. one node per zone is selected to carry the egressIP
          4. the relative aws machines are assigned the additional IP on the main interface (support for secondary interfaces in not available)
          5. the relative `hostsubnets` are updated to reflect the assigned IP, the `egressIP` field is updated.
        displayName: Egressip Ipam Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - egressIPs
        - high availability
        - firewall support
        - traffic management
        - ipam
        links:
        - name: Egressip Ipam Operator
          url: https://github.com/redhat-cop/egressip-ipam-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/egressip-ipam-operator:v0.1.2
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/egressip-ipam-operator@sha256:d6129bd287b87696a23e0da64dcd598958f779dc88ae33b46b3c909dbab80adf
        version: 1.2.4
      entries:
      - name: egressip-ipam-operator.v1.2.4
        version: 1.2.4
      - name: egressip-ipam-operator.v1.2.3
        version: 1.2.3
      - name: egressip-ipam-operator.v1.2.2
        version: 1.2.2
      - name: egressip-ipam-operator.v1.2.1
        version: 1.2.1
      - name: egressip-ipam-operator.v1.2.0
        version: 1.2.0
      - name: egressip-ipam-operator.v1.1.0
        version: 1.1.0
      - name: egressip-ipam-operator.v1.0.8
        version: 1.0.8
      - name: egressip-ipam-operator.v1.0.7
        version: 1.0.7
      - name: egressip-ipam-operator.v1.0.6
        version: 1.0.6
      - name: egressip-ipam-operator.v1.0.5
        version: 1.0.5
      - name: egressip-ipam-operator.v1.0.4
        version: 1.0.4
      - name: egressip-ipam-operator.v1.0.3
        version: 1.0.3
      - name: egressip-ipam-operator.v1.0.2
        version: 1.0.2
      - name: egressip-ipam-operator.v1.0.1
        version: 1.0.1
      - name: egressip-ipam-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: egressip-ipam-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: hive-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hive-operator.v1.2.4953-c248a51
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:c248a51ae1
          createdAt: "2025-09-22T13:41:19Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeploymentCustomization is the Schema for clusterdeploymentcustomizations
              API.
            displayName: ClusterDeploymentCustomization
            kind: ClusterDeploymentCustomization
            name: clusterdeploymentcustomizations.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: |-
              ClusterPool represents a pool of clusters that should be kept ready to be given out to users. Clusters are removed
              from the pool once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: |-
              MachinePoolNameLease is the Schema for the MachinePoolNameLeases API. This resource is mostly empty
              as we're primarily relying on the name to determine if a lease is available.
              Note that not all cloud providers require the use of a lease for naming, at present this
              is only required for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * IBM Cloud
          * Red Hat OpenStack
          * vSphere

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Troubleshooting](https://github.com/openshift/hive/blob/master/docs/troubleshooting.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
              logLevel: debug
              targetNamespace: hive
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:c248a51ae1
        version: 1.2.4953-c248a51
      entries:
      - name: hive-operator.v1.2.4953-c248a51
        version: 1.2.4953-c248a51
      - name: hive-operator.v1.2.4928-76a65b6
        version: 1.2.4928-76a65b6
      - name: hive-operator.v1.2.4902-4929f16
        version: 1.2.4902-4929f16
      - name: hive-operator.v1.2.4896-a29a238
        version: 1.2.4896-a29a238
      - name: hive-operator.v1.2.4868-45ae912
        version: 1.2.4868-45ae912
      - name: hive-operator.v1.2.4863-e9ad0fa
        version: 1.2.4863-e9ad0fa
      - name: hive-operator.v1.2.4774-87bff59
        version: 1.2.4774-87bff59
      - name: hive-operator.v1.2.4766-6a82857
        version: 1.2.4766-6a82857
      - name: hive-operator.v1.2.4746-d25e0db
        version: 1.2.4746-d25e0db
      - name: hive-operator.v1.2.4738-b19773e
        version: 1.2.4738-b19773e
      - name: hive-operator.v1.2.4731-41f153f
        version: 1.2.4731-41f153f
      - name: hive-operator.v1.2.4703-02f3f00
        version: 1.2.4703-02f3f00
      - name: hive-operator.v1.2.4693-ec3bc72
        version: 1.2.4693-ec3bc72
      - name: hive-operator.v1.2.4686-553c6c2
        version: 1.2.4686-553c6c2
      - name: hive-operator.v1.2.4682-406d6ca
        version: 1.2.4682-406d6ca
      - name: hive-operator.v1.2.4672-af54e2f
        version: 1.2.4672-af54e2f
      - name: hive-operator.v1.2.4668-3ccd44d
        version: 1.2.4668-3ccd44d
      - name: hive-operator.v1.2.4664-00f84cc
        version: 1.2.4664-00f84cc
      - name: hive-operator.v1.2.4658-c13e207
        version: 1.2.4658-c13e207
      - name: hive-operator.v1.2.4654-3268a18
        version: 1.2.4654-3268a18
      - name: hive-operator.v1.2.4650-531d549
        version: 1.2.4650-531d549
      - name: hive-operator.v1.2.4644-365f074
        version: 1.2.4644-365f074
      - name: hive-operator.v1.2.4638-56d6d64
        version: 1.2.4638-56d6d64
      - name: hive-operator.v1.2.4625-389d717
        version: 1.2.4625-389d717
      - name: hive-operator.v1.2.4615-52100dd
        version: 1.2.4615-52100dd
      - name: hive-operator.v1.2.4605-b41ca3f
        version: 1.2.4605-b41ca3f
      - name: hive-operator.v1.2.4584-95efcb7
        version: 1.2.4584-95efcb7
      - name: hive-operator.v1.2.4574-1ec27ad
        version: 1.2.4574-1ec27ad
      - name: hive-operator.v1.2.4568-b3eca52
        version: 1.2.4568-b3eca52
      - name: hive-operator.v1.2.4564-bdf9d08
        version: 1.2.4564-bdf9d08
      - name: hive-operator.v1.2.4552-bf2d689
        version: 1.2.4552-bf2d689
      - name: hive-operator.v1.2.4532-cddd4f9
        version: 1.2.4532-cddd4f9
      - name: hive-operator.v1.2.4511-1fce770
        version: 1.2.4511-1fce770
      - name: hive-operator.v1.2.4500-b1ac27b
        version: 1.2.4500-b1ac27b
      - name: hive-operator.v1.2.4485-47df0d9
        version: 1.2.4485-47df0d9
      - name: hive-operator.v1.2.4469-7be6433
        version: 1.2.4469-7be6433
      - name: hive-operator.v1.2.4463-d7ead60
        version: 1.2.4463-d7ead60
      - name: hive-operator.v1.2.4461-97686bb
        version: 1.2.4461-97686bb
      - name: hive-operator.v1.2.4459-f910c9b
        version: 1.2.4459-f910c9b
      - name: hive-operator.v1.2.4451-83aedb9
        version: 1.2.4451-83aedb9
      - name: hive-operator.v1.2.4446-e11a23b
        version: 1.2.4446-e11a23b
      - name: hive-operator.v1.2.4441-fdbfaf0
        version: 1.2.4441-fdbfaf0
      - name: hive-operator.v1.2.4436-44323bc
        version: 1.2.4436-44323bc
      - name: hive-operator.v1.2.4428-ceaafa8
        version: 1.2.4428-ceaafa8
      - name: hive-operator.v1.2.4426-fb27114
        version: 1.2.4426-fb27114
      - name: hive-operator.v1.2.4420-fe0e302
        version: 1.2.4420-fe0e302
      - name: hive-operator.v1.2.4418-eb5403f
        version: 1.2.4418-eb5403f
      - name: hive-operator.v1.2.4416-fba8d52
        version: 1.2.4416-fba8d52
      - name: hive-operator.v1.2.4408-9c5806a
        version: 1.2.4408-9c5806a
      - name: hive-operator.v1.2.4404-410fed9
        version: 1.2.4404-410fed9
      - name: hive-operator.v1.2.4400-8c54fc9
        version: 1.2.4400-8c54fc9
      - name: hive-operator.v1.2.4386-6f1fd3e
        version: 1.2.4386-6f1fd3e
      - name: hive-operator.v1.2.4382-0e95d5b
        version: 1.2.4382-0e95d5b
      - name: hive-operator.v1.2.4366-df65e7e
        version: 1.2.4366-df65e7e
      - name: hive-operator.v1.2.4360-de169dd
        version: 1.2.4360-de169dd
      - name: hive-operator.v1.2.4355-e5f809f
        version: 1.2.4355-e5f809f
      name: alpha
    - currentCSV: hive-operator.v2.5.3516-a2ed9b3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.5.3516-a2ed9b3
          createdAt: "2022-06-03T13:24:12Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:v2.5.3516-a2ed9b3
        version: 2.5.3516-a2ed9b3
      entries:
      - name: hive-operator.v2.5.3516-a2ed9b3
        version: 2.5.3516-a2ed9b3
      - name: hive-operator.v2.5.3514-4052426
        version: 2.5.3514-4052426
      - name: hive-operator.v2.5.3508-6cb94c6
        version: 2.5.3508-6cb94c6
      - name: hive-operator.v2.5.3506-f42617c
        version: 2.5.3506-f42617c
      - name: hive-operator.v2.5.3495-5cfdd44
        version: 2.5.3495-5cfdd44
      - name: hive-operator.v2.5.3491-6feded1
        version: 2.5.3491-6feded1
      - name: hive-operator.v2.5.3489-6dec9c3
        version: 2.5.3489-6dec9c3
      name: mce-2.0
    - currentCSV: hive-operator.v2.3.3039-60b8a9a
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.3.3039-60b8a9a
          createdAt: "2022-08-15T17:23:39Z"
          description: OpenShift cluster provisioning and management at scale.
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:v2.3.3039-60b8a9a
        version: 2.3.3039-60b8a9a
      entries:
      - name: hive-operator.v2.3.3039-60b8a9a
        version: 2.3.3039-60b8a9a
      name: ocm-2.3
    - currentCSV: hive-operator.v2.4.3231-ba51985
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.4.3231-ba51985
          createdAt: "2022-08-15T11:53:28Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:v2.4.3231-ba51985
        version: 2.4.3231-ba51985
      entries:
      - name: hive-operator.v2.4.3231-ba51985
        version: 2.4.3231-ba51985
      - name: hive-operator.v2.4.3227-74831db
        version: 2.4.3227-74831db
      - name: hive-operator.v2.4.3222-d68346c
        version: 2.4.3222-d68346c
      - name: hive-operator.v2.4.3206-b0714a7
        version: 2.4.3206-b0714a7
      - name: hive-operator.v2.4.3203-672cfe7
        version: 2.4.3203-672cfe7
      - name: hive-operator.v2.4.3201-8c50440
        version: 2.4.3201-8c50440
      - name: hive-operator.v2.4.3199-cce4cb9
        version: 2.4.3199-cce4cb9
      - name: hive-operator.v2.4.3197-6c9c8e4
        version: 2.4.3197-6c9c8e4
      - name: hive-operator.v2.4.3194-f82b7d8
        version: 2.4.3194-f82b7d8
      - name: hive-operator.v2.4.3192-fae0111
        version: 2.4.3192-fae0111
      - name: hive-operator.v2.4.3189-026033b
        version: 2.4.3189-026033b
      - name: hive-operator.v2.4.3180-7299056
        version: 2.4.3180-7299056
      - name: hive-operator.v1.2.3179-0756b70e6
        version: 1.2.3179-0756b70e6
      - name: hive-operator.v1.1.16
        version: 1.1.16
      - name: hive-operator.v1.1.15
        version: 1.1.15
      - name: hive-operator.v1.1.14
        version: 1.1.14
      - name: hive-operator.v1.1.13
        version: 1.1.13
      - name: hive-operator.v1.1.12
        version: 1.1.12
      - name: hive-operator.v1.1.11
        version: 1.1.11
      - name: hive-operator.v1.1.10
        version: 1.1.10
      - name: hive-operator.v1.1.9
        version: 1.1.9
      - name: hive-operator.v1.1.8
        version: 1.1.8
      - name: hive-operator.v1.1.7
        version: 1.1.7
      - name: hive-operator.v1.1.6
        version: 1.1.6
      name: ocm-2.4
    defaultChannel: alpha
    packageName: hive-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-admin-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-admin-operator.v2024.10.2-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": null
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:ed279ff5416cce855118f66a73d55385aba26af19735eed052010ba653f64b75
          createdAt: "2024-10-25T01:47:18Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2024.10.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2023.9.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:7c4868db021beaffa762eb0de43739bbc865a4bf5900f1d17f78caed8612c073
        - registry.connect.redhat.com/ca/uma-operator@sha256:ed279ff5416cce855118f66a73d55385aba26af19735eed052010ba653f64b75
        version: 2024.10.2-12
      entries:
      - name: uma-admin-operator.v2024.10.2-12
        version: 2024.10.2-12
      name: "2024.10"
    - currentCSV: uma-admin-operator.v2025.1.3-33
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.1.1.33,990033",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": null
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:6a932a065add9f36f9615a93037315603f2d9cd8a2e3c64b1b1fc390bb4f1dcf
          createdAt: "2025-02-18T05:34:34Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.1.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2023.9.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.1.1.26,990026"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:6a932a065add9f36f9615a93037315603f2d9cd8a2e3c64b1b1fc390bb4f1dcf
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:ae91a28d57c8a359f81757e4bf79472f2ac189ac0ae3ee0d87e8ff9c433f1ccc
        version: 2025.1.3-33
      entries:
      - name: uma-admin-operator.v2025.1.3-33
        version: 2025.1.3-33
      name: "2025.1"
    - currentCSV: uma-admin-operator.v2025.2.1-25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.2.1.25,990025",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverCollectorCert": "",
                    "serverCollectorKey": "",
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:d201790c2d9ded6bc7a1c5f68a267574e7b5d39c39ce2f98e20ef630b0f7365f
          createdAt: "2025-03-05T05:51:02Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.2.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2023.9.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.1.1.26,990026"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:d201790c2d9ded6bc7a1c5f68a267574e7b5d39c39ce2f98e20ef630b0f7365f
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:8bcf0dac889f38a0bf77eb9babe3f726e3b2564d5988c05bdc5c2d91ae189c61
        version: 2025.2.1-25
      entries:
      - name: uma-admin-operator.v2025.2.1-25
        version: 2025.2.1-25
      name: "2025.2"
    - currentCSV: uma-admin-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:4c2a7f68a47428f3e386107254f74fa8c34e477eb40fc17c173df92f3c66c559
          createdAt: "2025-04-16T06:55:39Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.4.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.4.2.5,990005"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        - registry.connect.redhat.com/ca/uma-operator@sha256:4c2a7f68a47428f3e386107254f74fa8c34e477eb40fc17c173df92f3c66c559
        version: 2025.4.5-5
      entries:
      - name: uma-admin-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-admin-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:514b953eaf0959bc120f80bdc0a2e19d8f059d3fe2883a4f6843a398470a56b1
          createdAt: "2025-07-24T14:54:21Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.5.2.12,990012"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:514b953eaf0959bc120f80bdc0a2e19d8f059d3fe2883a4f6843a398470a56b1
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        version: 2025.5.4-12
      entries:
      - name: uma-admin-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-admin-operator.v2025.7.3-21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:188a2b8d2255ded6141a4b795f1b6d7bd98017b4fd57c9c1a1d38fc701358057
          createdAt: "2025-08-23T09:00:37Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.3
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.7.1.23,990023"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:188a2b8d2255ded6141a4b795f1b6d7bd98017b4fd57c9c1a1d38fc701358057
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.3-21
      entries:
      - name: uma-admin-operator.v2025.7.3-21
        version: 2025.7.3-21
      name: "2025.7"
    - currentCSV: uma-admin-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:0cfb63c62258e54d67e7baca874003a04477f30bcf7f56861f565ad068070580
          createdAt: "2025-08-15T03:20:28Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-admin-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.8.1.9,990009"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:0cfb63c62258e54d67e7baca874003a04477f30bcf7f56861f565ad068070580
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        version: 2025.8.2-9
      entries:
      - name: uma-admin-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-admin-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
          createdAt: "2025-09-24T12:23:56Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.12,990012"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-admin-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-admin-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:7b93de818fa5160e72f878b0350d13fc1d5943eac95b0794d3d4f930cf21b59a
          createdAt: "2025-07-30T08:07:39Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-admin-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "99.99.0.serp_F152813,19"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:7b93de818fa5160e72f878b0350d13fc1d5943eac95b0794d3d4f930cf21b59a
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        version: 2025.7.2-19
      entries:
      - name: uma-admin-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-admin-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1cf3909dd2caeabc53209d6f8565dcb6c3562b9c6eb715dda8ba1953f264668e
          createdAt: "2025-09-11T11:16:06Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.8,990008"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:1cf3909dd2caeabc53209d6f8565dcb6c3562b9c6eb715dda8ba1953f264668e
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-admin-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-admin-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
          createdAt: "2025-09-24T12:23:56Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.12,990012"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-admin-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-admin-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: octavia-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: octavia-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "octavia.openstack.org/v1beta1",
                "kind": "Octavia",
                "metadata": {
                  "name": "octavia"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "octavia",
                  "databaseInstance": "openstack",
                  "lbMgmtNetwork": {
                    "availabilityZones": [
                      "zone-1"
                    ]
                  },
                  "octaviaAPI": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "octavia",
                    "databaseInstance": "openstack",
                    "persistenceDatabaseAccount": "octavia-persistence",
                    "preserveJobs": false,
                    "secret": "osp-secret",
                    "serviceAccount": "octavia",
                    "serviceUser": "octavia"
                  },
                  "octaviaHealthManager": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "octavia",
                    "databaseInstance": "openstack",
                    "networkAttachments": [
                      "octavia"
                    ],
                    "persistenceDatabaseAccount": "octavia-persistence",
                    "preserveJobs": false,
                    "role": "healthmanager",
                    "secret": "osp-secret",
                    "serviceAccount": "octavia",
                    "serviceUser": "octavia"
                  },
                  "octaviaHousekeeping": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "octavia",
                    "databaseInstance": "openstack",
                    "networkAttachments": [
                      "octavia"
                    ],
                    "persistenceDatabaseAccount": "octavia-persistence",
                    "preserveJobs": false,
                    "role": "housekeeping",
                    "secret": "osp-secret",
                    "serviceAccount": "octavia",
                    "serviceUser": "octavia"
                  },
                  "octaviaWorker": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "octavia",
                    "databaseInstance": "openstack",
                    "networkAttachments": [
                      "octavia"
                    ],
                    "persistenceDatabaseAccount": "octavia-persistence",
                    "preserveJobs": false,
                    "role": "worker",
                    "secret": "osp-secret",
                    "serviceAccount": "octavia",
                    "serviceUser": "octavia"
                  },
                  "persistenceDatabaseAccount": "octavia-persistence",
                  "preserveJobs": false,
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "osp-secret",
                  "serviceUser": "octavia"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:01Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: Octavia Amphora Controller
            kind: OctaviaAmphoraController
            name: octaviaamphoracontrollers.octavia.openstack.org
            version: v1beta1
          - description: OctaviaAPI is the Schema for the octaviaapis API
            displayName: Octavia API
            kind: OctaviaAPI
            name: octaviaapis.octavia.openstack.org
            version: v1beta1
          - kind: OctaviaRsyslog
            name: octaviarsyslogs.octavia.openstack.org
            version: v1beta1
          - description: Octavia is the Schema for the octavia API
            displayName: Octavia
            kind: Octavia
            name: octavias.octavia.openstack.org
            version: v1beta1
        description: Octavia Operator
        displayName: Octavia Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Load-Balancer
        - Octavia
        links:
        - name: Octavia Operator
          url: https://github.com/openstack-k8s-operators/octavia-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/octavia-rhel9-operator@sha256:ddd280ec099cde63e38bd964d7f0608c1c34e5c9f6b140f86dde2126b0357ef5
        - registry.redhat.io/rhoso/openstack-octavia-health-manager-rhel9@sha256:7d305af6e9e6c636510186dd80037849d19bb1e6d2b2507c025d1d0d9254874a
        - registry.redhat.io/rhoso/openstack-octavia-worker-rhel9@sha256:0919ddff6421430f8ad24cb21a89c65f02f98df424267f7e436c97d555112e2a
        - registry.redhat.io/ubi9/httpd-24@sha256:dbfbe3383e12d63df7178e4cdd5b1e6e51a84c781041b9192026ea6edb72fbf3
        - registry.redhat.io/rhoso/openstack-octavia-api-rhel9@sha256:ae9aae94c2523933a346bdade16158b4302b3a0bee93dec89bae5d15606c8dfb
        - registry.redhat.io/rhoso/openstack-octavia-housekeeping-rhel9@sha256:d54868077feb205199b76e83fc9220d9b4215dbeba037cf13fb68dc6577a07dc
        version: 1.0.6
      entries:
      - name: octavia-operator.v1.0.6
        version: 1.0.6
      - name: octavia-operator.v1.0.5
        version: 1.0.5
      - name: octavia-operator.v1.0.4
        version: 1.0.4
      - name: octavia-operator.v1.0.3
        version: 1.0.3
      - name: octavia-operator.v1.0.2
        version: 1.0.2
      - name: octavia-operator.v1.0.1
        version: 1.0.1
      - name: octavia-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: octavia-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: gatekeeper-operator-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gatekeeper-operator-product.v3.11.2-0.1725401426.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel8-operator@sha256:c8894e4d02fb2c822ab2e02e489dd9b67c953d96bd3b96a8305219a2245dfd99
          createdAt: 21 Mar 2024, 01:59
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.11.0
          olm.substitutesFor: gatekeeper-operator-product.v3.11.2-0.1721233953.p
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/gatekeeper/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the Schema for the gatekeepers API
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |-
          Gatekeeper allows administrators to detect and reject non-compliant commits to an infrastructure-as-code system's source-of-truth. This strengthens compliance efforts and prevents a bad state from slowing down the organization.

          ## Support & Troubleshooting
          Gatekeeper is available and supported as part of a valid Red Hat software subscription when it is used with Red Hat Advanced Cluster Management for Kubernetes. Troubleshooting information is available in the Gatekeeper documentation.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/gatekeeper/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - registry.redhat.io/gatekeeper/gatekeeper-rhel8-operator@sha256:c8894e4d02fb2c822ab2e02e489dd9b67c953d96bd3b96a8305219a2245dfd99
        - registry.redhat.io/gatekeeper/gatekeeper-rhel8@sha256:701b165def9e4b92d91b8eac4917721e76e634d2159b8efc69bef23e10b51b9d
        version: 3.11.2+0.1725401426.p
      entries:
      - name: gatekeeper-operator-product.v3.11.2-0.1725401426.p
        version: 3.11.2+0.1725401426.p
      - name: gatekeeper-operator-product.v3.11.2-0.1721233953.p
        version: 3.11.2+0.1721233953.p
      - name: gatekeeper-operator-product.v3.11.2-0.1718224960.p
        version: 3.11.2+0.1718224960.p
      - name: gatekeeper-operator-product.v3.11.2
        version: 3.11.2
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.11"
    - currentCSV: gatekeeper-operator-product.v3.14.3-0.1746550072.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e47f0e9da800eb0cf4a0da6af98986508a35798a41581158ee7dbc977cb651dc
          createdAt: 16 Jan 2025, 06:57
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.14.3
          olm.substitutesFor: gatekeeper-operator-product.v3.14.3-0.1744033158.p
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the Schema for the gatekeepers API
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |-
          Gatekeeper allows administrators to detect and reject non-compliant commits to an infrastructure-as-code system's source-of-truth. This strengthens compliance efforts and prevents a bad state from slowing down the organization.

          ## Support & Troubleshooting
          Gatekeeper is available and supported as part of a valid Red Hat software subscription when it is used with Red Hat Advanced Cluster Management for Kubernetes. Troubleshooting information is available in the Gatekeeper documentation.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e47f0e9da800eb0cf4a0da6af98986508a35798a41581158ee7dbc977cb651dc
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:6255f7a673e6989617176842d8f0cc897e4630b5283e812a7a2d52dbb2a1f335
        version: 3.14.3+0.1746550072.p
      entries:
      - name: gatekeeper-operator-product.v3.14.3-0.1746550072.p
        version: 3.14.3+0.1746550072.p
      - name: gatekeeper-operator-product.v3.14.3-0.1744033158.p
        version: 3.14.3+0.1744033158.p
      - name: gatekeeper-operator-product.v3.14.3-0.1742934403.p
        version: 3.14.3+0.1742934403.p
      - name: gatekeeper-operator-product.v3.14.3-0.1740676608.p
        version: 3.14.3+0.1740676608.p
      - name: gatekeeper-operator-product.v3.14.3
        version: 3.14.3
      - name: gatekeeper-operator-product.v3.14.2
        version: 3.14.2
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.14"
    - currentCSV: gatekeeper-operator-product.v3.15.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:8b43248aed06563e06f7a0347ff1c45fa5892e3a9adb2b5c363962a0b152eff0
          createdAt: 13 Mar 2025, 17:07
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.15.4
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: Gatekeeper allows administrators to detect and reject non-compliant
          commits to an infrastructure-as-code system's source-of-truth. This strengthens
          compliance efforts and prevents a bad state from slowing down the organization.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:8b43248aed06563e06f7a0347ff1c45fa5892e3a9adb2b5c363962a0b152eff0
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:29690144667b3753a7e76095e136d6fc42325270e7b9f22f97dfd411439e9f54
        version: 3.15.4
      entries:
      - name: gatekeeper-operator-product.v3.15.4
        version: 3.15.4
      - name: gatekeeper-operator-product.v3.15.3
        version: 3.15.3
      - name: gatekeeper-operator-product.v3.15.2
        version: 3.15.2
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.15"
    - currentCSV: gatekeeper-operator-product.v3.17.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:3c027ec78398df0a8fc09eb79e222a9508b4f18154d1e1b477d4f156181ae51a
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.17.3
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e44078fc6073a79168ed52922b370773b451d69bf335aeb91d26a1462df21a9b
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:3c027ec78398df0a8fc09eb79e222a9508b4f18154d1e1b477d4f156181ae51a
        version: 3.17.3
      entries:
      - name: gatekeeper-operator-product.v3.17.3
        version: 3.17.3
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.17"
    - currentCSV: gatekeeper-operator-product.v3.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:0803b7219ceff7aa676488eeba58607a19c9c18a2bb8eae8b3b16cfec411cb6e
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.18.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:49726cb2702441305e8047b24054575170fb595cfe5ebf815cdbfdaa957b02f4
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:0803b7219ceff7aa676488eeba58607a19c9c18a2bb8eae8b3b16cfec411cb6e
        version: 3.18.1
      entries:
      - name: gatekeeper-operator-product.v3.18.1
        version: 3.18.1
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.18"
    - currentCSV: gatekeeper-operator-product.v3.19.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.19.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:60ee347425c6cfb36f0e1e834b33cbb3d77aa2d6c89ac513216d0b9fc55d4b52
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
        version: 3.19.1
      entries:
      - name: gatekeeper-operator-product.v3.19.1
        version: 3.19.1
      - name: gatekeeper-operator-product.v3.19.0
        version: 3.19.0
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.19"
    - currentCSV: gatekeeper-operator-product.v3.19.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.19.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:60ee347425c6cfb36f0e1e834b33cbb3d77aa2d6c89ac513216d0b9fc55d4b52
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
        version: 3.19.1
      entries:
      - name: gatekeeper-operator-product.v3.19.1
        version: 3.19.1
      - name: gatekeeper-operator-product.v3.19.0
        version: 3.19.0
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: stable
    defaultChannel: stable
    packageName: gatekeeper-operator-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://github.com/redhat-developer/app-services-operator/
    name: rhoas-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhoas-operator.0.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "ServiceRegistryConnection",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace",
                  "labels": {
                    "app.kubernetes.io/component": "external-service",
                    "app.kubernetes.io/managed-by": "rhoas"
                  }
                },
                "spec": {
                  "accessTokenSecretName": "rh-managed-services-api-accesstoken",
                  "serviceRegistryId": "exampleId",
                  "credentials": {
                    "serviceAccountSecretName": "service-account-secret"
                  }
                }
              },
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "CloudServiceAccountRequest",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace"
                },
                "spec": {
                  "serviceAccountName": "rhoas-sa",
                  "serviceAccountDescription": "Operator created service account",
                  "serviceAccountSecretName": "service-account-credentials",
                  "accessTokenSecretName": "rh-managed-services-api-accesstoken"
                }
              },
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "CloudServicesRequest",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace",
                  "labels": {
                    "app.kubernetes.io/component": "external-service",
                    "app.kubernetes.io/managed-by": "rhoas"
                  }
                },
                "spec": {
                  "accessTokenSecretName": "rh-cloud-services-api-accesstoken"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: quay.io/rhoas/service-operator:0.13.0
          createdAt: "2023-01-20 16:09:37"
          description: Red Hat OpenShift Application Services Operator
          olm.skipRange: '>=0.0.1 <0.13.0'
          repository: https://github.com/redhat-developer/app-services-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Creates RHOAS Service Accounts
            displayName: Cloud Service Account Request
            kind: CloudServiceAccountRequest
            name: cloudserviceaccountrequests.rhoas.redhat.com
            version: v1alpha1
          - description: Cloud Services Request
            displayName: Cloud Services Request
            kind: CloudServicesRequest
            name: cloudservicesrequests.rhoas.redhat.com
            version: v1alpha1
          - description: Binds Kafka Instance
            displayName: Kafka Connection
            kind: KafkaConnection
            name: kafkaconnections.rhoas.redhat.com
            version: v1alpha1
          - description: Binds Service Registry Instance
            displayName: Service Registry Connection
            kind: ServiceRegistryConnection
            name: serviceregistryconnections.rhoas.redhat.com
            version: v1alpha1
        description: "**OpenShift Application Services Operator (RHOAS-Operator)**
          creates instances of Red Hat OpenShift Application Services in your cluster.
          Operator is currently available as limited Beta and it cannot be consumed
          without invitation.\n\nRed Hat OpenShift Application Services, when added
          to Red Hat OpenShift Dedicated,  provide a streamlined developer experience
          for building, deploying,  and scaling cloud-native applications in open
          hybrid-cloud environments.   \n\nRHOAS-Operator enables developers to connect
          with their Application Services directly in their own cluster by using [RHOAS
          cli](https://github.com/redhat-developer/app-services-cli) and OpenShift
          Console UI.  \n\nRHOAS-Operator requires developers to connect with their
          Red Hat OpenShift Streams for Apache Kafka (RHOSAK) instances.  "
        displayName: OpenShift Application Services (RHOAS)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - rhoas
        - kafka
        - serviceregistry
        - app-services
        - application
        - services
        links:
        - name: RHOAS Operator Documentation
          url: https://github.com/redhat-developer/app-services-operator/tree/main/docs
        maintainers:
        - email: rhosak-eval-support@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://github.com/redhat-developer/app-services-operator/
        relatedImages:
        - quay.io/rhoas/service-operator:0.13.0
        version: 0.13.0
      entries:
      - name: rhoas-operator.0.13.0
        version: 0.13.0
      - name: rhoas-operator.0.12.0
        version: 0.12.0
      - name: rhoas-operator.0.11.0
        version: 0.11.0
      - name: rhoas-operator.0.9.6
        version: 0.9.6
      - name: rhoas-operator.0.9.5
        version: 0.9.5
      - name: rhoas-operator.0.9.3
        version: 0.9.3
      - name: rhoas-operator.0.9.2
        version: 0.9.2
      - name: rhoas-operator.0.9.1
        version: 0.9.1
      - name: rhoas-operator.0.9.0
        version: 0.9.0
      - name: rhoas-operator.0.8.2
        version: 0.8.2
      - name: rhoas-operator.0.7.10
        version: 0.7.10
      - name: rhoas-operator.0.7.9
        version: 0.7.9
      - name: rhoas-operator.0.7.8
        version: 0.7.8
      - name: rhoas-operator.0.7.6
        version: 0.7.6
      - name: rhoas-operator.0.7.2
        version: 0.7.2
      - name: rhoas-operator.0.7.1
        version: 0.7.1
      - name: rhoas-operator.0.6.8
        version: 0.6.8
      name: beta
    defaultChannel: beta
    packageName: rhoas-operator
    provider:
      name: Red Hat Inc.
      url: https://github.com/redhat-developer/app-services-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-s3-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-s3-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "s3.services.k8s.aws/v1alpha1",
                "kind": "Bucket",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/s3-controller:1.1.1
          createdAt: "2025-09-19T19:23:45Z"
          description: AWS S3 controller is a service controller for managing S3 resources
            in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Bucket represents the state of an AWS s3 Bucket resource.
            displayName: Bucket
            kind: Bucket
            name: buckets.s3.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Simple Storage Service (S3) resources in AWS from within your Kubernetes cluster.

          **About Amazon S3**

          Amazon Simple Storage Service (Amazon S3) is an object storage service that offers industry-leading scalability, data availability, security, and performance. This means customers of all sizes and industries can use it to store and protect any amount of data for a range of use cases, such as data lakes, websites, mobile applications, backup and restore, archive, enterprise applications, IoT devices, and big data analytics. Amazon S3 provides easy-to-use management features so you can organize your data and configure finely-tuned access controls to meet your specific business, organizational, and compliance requirements. Amazon S3 is designed for 99.999999999% (11 9s) of durability, and stores data for millions of applications for companies all around the world.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon S3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - s3
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon S3 Developer Resources
          url: https://aws.amazon.com/s3/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: s3 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/s3-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-s3-controller.v1.1.1
        version: 1.1.1
      - name: ack-s3-controller.v1.1.0
        version: 1.1.0
      - name: ack-s3-controller.v1.0.35
        version: 1.0.35
      - name: ack-s3-controller.v1.0.33
        version: 1.0.33
      - name: ack-s3-controller.v1.0.32
        version: 1.0.32
      - name: ack-s3-controller.v1.0.31
        version: 1.0.31
      - name: ack-s3-controller.v1.0.30
        version: 1.0.30
      - name: ack-s3-controller.v1.0.29
        version: 1.0.29
      - name: ack-s3-controller.v1.0.28
        version: 1.0.28
      - name: ack-s3-controller.v1.0.27
        version: 1.0.27
      - name: ack-s3-controller.v1.0.26
        version: 1.0.26
      - name: ack-s3-controller.v1.0.25
        version: 1.0.25
      - name: ack-s3-controller.v1.0.24
        version: 1.0.24
      - name: ack-s3-controller.v1.0.23
        version: 1.0.23
      - name: ack-s3-controller.v1.0.21
        version: 1.0.21
      - name: ack-s3-controller.v1.0.20
        version: 1.0.20
      - name: ack-s3-controller.v1.0.19
        version: 1.0.19
      - name: ack-s3-controller.v1.0.18
        version: 1.0.18
      - name: ack-s3-controller.v1.0.17
        version: 1.0.17
      - name: ack-s3-controller.v1.0.16
        version: 1.0.16
      - name: ack-s3-controller.v1.0.15
        version: 1.0.15
      - name: ack-s3-controller.v1.0.14
        version: 1.0.14
      - name: ack-s3-controller.v1.0.13
        version: 1.0.13
      - name: ack-s3-controller.v1.0.12
        version: 1.0.12
      - name: ack-s3-controller.v1.0.11
        version: 1.0.11
      - name: ack-s3-controller.v1.0.9
        version: 1.0.9
      - name: ack-s3-controller.v1.0.7
        version: 1.0.7
      - name: ack-s3-controller.v1.0.6
        version: 1.0.6
      - name: ack-s3-controller.v1.0.5
        version: 1.0.5
      - name: ack-s3-controller.v1.0.4
        version: 1.0.4
      - name: ack-s3-controller.v1.0.3
        version: 1.0.3
      - name: ack-s3-controller.v1.0.2
        version: 1.0.2
      - name: ack-s3-controller.v1.0.1
        version: 1.0.1
      - name: ack-s3-controller.v1.0.0
        version: 1.0.0
      - name: ack-s3-controller.v0.1.8
        version: 0.1.8
      - name: ack-s3-controller.v0.1.7
        version: 0.1.7
      - name: ack-s3-controller.v0.1.6
        version: 0.1.6
      - name: ack-s3-controller.v0.1.5
        version: 0.1.5
      - name: ack-s3-controller.v0.1.4
        version: 0.1.4
      - name: ack-s3-controller.v0.1.3
        version: 0.1.3
      - name: ack-s3-controller.v0.1.2
        version: 0.1.2
      - name: ack-s3-controller.v0.1.1
        version: 0.1.1
      - name: ack-s3-controller.v0.1.0
        version: 0.1.0
      - name: ack-s3-controller.v0.0.20
        version: 0.0.20
      - name: ack-s3-controller.v0.0.19
        version: 0.0.19
      - name: ack-s3-controller.v0.0.18
        version: 0.0.18
      - name: ack-s3-controller.v0.0.17
        version: 0.0.17
      - name: ack-s3-controller.v0.0.16
        version: 0.0.16
      - name: ack-s3-controller.v0.0.15
        version: 0.0.15
      - name: ack-s3-controller.v0.0.14
        version: 0.0.14
      - name: ack-s3-controller.v0.0.13
        version: 0.0.13
      - name: ack-s3-controller.v0.0.12
        version: 0.0.12
      - name: ack-s3-controller.v0.0.11
        version: 0.0.11
      - name: ack-s3-controller.v0.0.10
        version: 0.0.10
      - name: ack-s3-controller.v0.0.9
        version: 0.0.9
      - name: ack-s3-controller.v0.0.8
        version: 0.0.8
      - name: ack-s3-controller.v0.0.7
        version: 0.0.7
      name: alpha
    defaultChannel: alpha
    packageName: ack-s3-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: numaresources-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: numaresources-operator.v4.16.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodetopology.openshift.io/v1alpha1",
                "kind": "NUMAResourcesOperator",
                "metadata": {
                  "name": "numaresourcesoperator"
                },
                "spec": {
                  "nodeGroups": [
                    {
                      "machineConfigPoolSelector": {
                        "matchLabels": {
                          "pools.operator.machineconfiguration.openshift.io/worker": ""
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "nodetopology.openshift.io/v1alpha1",
                "kind": "NUMAResourcesScheduler",
                "metadata": {
                  "name": "numaresourcesscheduler"
                },
                "spec": {
                  "imageSpec": "URL_OF_SCHEDULER_IMAGE_FROM_REDHAT_REGISTRY",
                  "logLevel": "Normal",
                  "schedulerName": "topo-aware-scheduler"
                }
              }
            ]
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "true"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.15.0 <4.16.6'
          operators.openshift.io/valid-subscription: "[\n\t\"OpenShift Kubernetes
            Engine\",\n\t\"OpenShift Container Platform\",\n\t\"OpenShift Platform
            Plus\"\n]"
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NUMAResourcesOperator is the Schema for the numaresourcesoperators
              API
            displayName: NUMA Resources Operator
            kind: NUMAResourcesOperator
            name: numaresourcesoperators.nodetopology.openshift.io
            version: v1
          - description: NUMAResourcesOperator is the Schema for the numaresourcesoperators
              API
            displayName: NUMA Resources Operator
            kind: NUMAResourcesOperator
            name: numaresourcesoperators.nodetopology.openshift.io
            version: v1alpha1
          - description: NUMAResourcesScheduler is the Schema for the numaresourcesschedulers
              API
            displayName: NUMA Aware Scheduler
            kind: NUMAResourcesScheduler
            name: numaresourcesschedulers.nodetopology.openshift.io
            version: v1
          - description: NUMAResourcesScheduler is the Schema for the numaresourcesschedulers
              API
            displayName: NUMA Aware Scheduler
            kind: NUMAResourcesScheduler
            name: numaresourcesschedulers.nodetopology.openshift.io
            version: v1alpha1
        description: NUMA resources exporter operator
        displayName: numaresources-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - RTE
        - NUMA
        links:
        - name: Numaresources Operator
          url: https://github.com/openshift-kni/numaresources-operator
        maintainers:
        - email: fromani@redhat.com
          name: fromani
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/numaresources-rhel9-operator@sha256:aca8232e89172ca9fd9ce4a66cae9e8bb027b6d16482ce6b8e27fb7779c398e2
        version: 4.16.6
      entries:
      - name: numaresources-operator.v4.16.6
        version: 4.16.6
      - name: numaresources-operator.v4.16.5
        version: 4.16.5
      - name: numaresources-operator.v4.16.4
        version: 4.16.4
      - name: numaresources-operator.v4.16.3
        version: 4.16.3
      - name: numaresources-operator.v4.16.2
        version: 4.16.2
      - name: numaresources-operator.v4.16.1
        version: 4.16.1
      - name: numaresources-operator.v4.16.0
        version: 4.16.0
      name: "4.16"
    defaultChannel: "4.16"
    packageName: numaresources-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.9-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lvms-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lvms-operator.v4.15.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "my-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "default": true,
                        "fstype": "xfs",
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:696edeb4c168646b49b8b0ca5b3d06071ab66fdb010e7cd46888d8b2f3a0c7a8
          description: Logical volume manager storage provides dynamically provisioned
            local storage for container workloads
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.11'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "test-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
          operatorframework.io/suggested-namespace: openshift-storage
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-storage",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/internal-objects: '["logicalvolumes.topolvm.io",
            "lvmvolumegroups.lvm.topolvm.io", "lvmvolumegroupnodestatuses.lvm.topolvm.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/lvm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LogicalVolume
            name: logicalvolumes.topolvm.io
            version: v1
          - description: LVMCluster is the Schema for the lvmclusters API
            displayName: LVMCluster
            kind: LVMCluster
            name: lvmclusters.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroupNodeStatus
            name: lvmvolumegroupnodestatuses.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroup
            name: lvmvolumegroups.lvm.topolvm.io
            version: v1alpha1
        description: Logical volume manager storage provides dynamically provisioned
          local storage.
        displayName: LVM Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - local storage
        - operator
        - LVM
        links:
        - name: Source Repository
          url: https://github.com/openshift/lvm-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:771d513c4d4e5d77d797b4c18f86e23162a20c6d39955e93d24f62b94e62e785
        - registry.redhat.io/lvms4/lvms-must-gather-rhel9@sha256:382ceea1151da56d0931403207c9f85721d58a66b39b1cd79790eca6219807d0
        - registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:696edeb4c168646b49b8b0ca5b3d06071ab66fdb010e7cd46888d8b2f3a0c7a8
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:82c01635b32821f04a7cfcf80da8b3f7e097786f220e0c85aab47bda4536a1d6
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:2db4f32824a8f9f6780dc3d236f73684cb20db41466ffa8f9a89fd3f5eb62fdd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:7ce50398cd29df6e07ccd89790a53a711a7e5b484698c48f6581f5d4e9efb05a
        - registry.redhat.io/lvms4/topolvm-rhel9@sha256:074a206d36c9e4198d90baed86f3046abd2ec12f9b36a715b60900011366b3d4
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:05dceaff7f0b2b769ddfd9261cc66d42aab9158ce99ed3dbd6c229e784072635
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:479454d0d4de6609d8a4696b9c7f0097115afb45db3bdd46631d58afd77dcc58
        version: 4.15.11
      entries:
      - name: lvms-operator.v4.15.11
        version: 4.15.11
      - name: lvms-operator.v4.15.10
        version: 4.15.10
      - name: lvms-operator.v4.15.9
        version: 4.15.9
      - name: lvms-operator.v4.15.8
        version: 4.15.8
      - name: lvms-operator.v4.15.7
        version: 4.15.7
      - name: lvms-operator.v4.15.6
        version: 4.15.6
      - name: lvms-operator.v4.15.5
        version: 4.15.5
      - name: lvms-operator.v4.15.4
        version: 4.15.4
      - name: lvms-operator.v4.15.3
        version: 4.15.3
      - name: lvms-operator.v4.15.2
        version: 4.15.2
      - name: lvms-operator.v4.15.1
        version: 4.15.1
      name: stable-4.15
    - currentCSV: lvms-operator.v4.16.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "my-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "default": true,
                        "fstype": "xfs",
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:b9d80bb263a54192501d5ba9366a4f0c19b34875b51bad4fa75ef22c60bb3ee7
          description: Logical volume manager storage provides dynamically provisioned
            local storage for container workloads
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.9'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "test-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
          operatorframework.io/suggested-namespace: openshift-storage
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-storage",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["logicalvolumes.topolvm.io",
            "lvmvolumegroups.lvm.topolvm.io", "lvmvolumegroupnodestatuses.lvm.topolvm.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lvm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LogicalVolume
            name: logicalvolumes.topolvm.io
            version: v1
          - description: LVMCluster is the Schema for the lvmclusters API
            displayName: LVMCluster
            kind: LVMCluster
            name: lvmclusters.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroupNodeStatus
            name: lvmvolumegroupnodestatuses.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroup
            name: lvmvolumegroups.lvm.topolvm.io
            version: v1alpha1
        description: Logical volume manager storage provides dynamically provisioned
          local storage.
        displayName: LVM Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - local storage
        - operator
        - LVM
        links:
        - name: Source Repository
          url: https://github.com/openshift/lvm-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/lvms4/topolvm-rhel9@sha256:227748103049d2cced921ba9fc4e2d3671f08c897cc2a5627d755fe5f9d2767d
        - registry.redhat.io/lvms4/lvms-must-gather-rhel9@sha256:20b28b87b3fe6d6bf739181376231d6f8b7454c34b59eaf2e06c04a9fca8f41d
        - registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:b9d80bb263a54192501d5ba9366a4f0c19b34875b51bad4fa75ef22c60bb3ee7
        version: 4.16.9
      entries:
      - name: lvms-operator.v4.16.9
        version: 4.16.9
      - name: lvms-operator.v4.16.8
        version: 4.16.8
      - name: lvms-operator.v4.16.7
        version: 4.16.7
      - name: lvms-operator.v4.16.6
        version: 4.16.6
      - name: lvms-operator.v4.16.5
        version: 4.16.5
      - name: lvms-operator.v4.16.4
        version: 4.16.4
      - name: lvms-operator.v4.16.3
        version: 4.16.3
      - name: lvms-operator.v4.16.2
        version: 4.16.2
      - name: lvms-operator.v4.16.1
        version: 4.16.1
      - name: lvms-operator.v4.16.0
        version: 4.16.0
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: lvms-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-custom-metrics-autoscaler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: custom-metrics-autoscaler.v2.17.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eventing.keda.sh/v1alpha1",
                "kind": "CloudEventSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keda",
                    "app.kubernetes.io/instance": "cloudeventsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cloudeventsource",
                    "app.kubernetes.io/part-of": "keda"
                  },
                  "name": "cloudeventsource-sample"
                },
                "spec": {
                  "clusterName": "cluster-sample",
                  "destination": {
                    "http": {
                      "uri": "http://foo.bar"
                    }
                  }
                }
              },
              {
                "apiVersion": "eventing.keda.sh/v1alpha1",
                "kind": "ClusterCloudEventSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keda",
                    "app.kubernetes.io/instance": "clustercloudeventsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clustercloudeventsource",
                    "app.kubernetes.io/part-of": "keda"
                  },
                  "name": "clustercloudeventsource-sample"
                },
                "spec": {
                  "clusterName": "clustercluster-sample",
                  "destination": {
                    "http": {
                      "uri": "http://foo.bar"
                    }
                  }
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ClusterTriggerAuthentication",
                "metadata": {
                  "name": "example-clustertriggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "example-secret-parameter"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "KedaController",
                "metadata": {
                  "name": "keda",
                  "namespace": "openshift-keda"
                },
                "spec": {
                  "admissionWebhooks": {
                    "logEncoder": "console",
                    "logLevel": "info"
                  },
                  "metricsServer": {
                    "logLevel": "0"
                  },
                  "operator": {
                    "logEncoder": "console",
                    "logLevel": "info"
                  },
                  "serviceAccount": null,
                  "watchNamespace": ""
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledJob",
                "metadata": {
                  "name": "scaledjob-sample"
                },
                "spec": {
                  "failedJobsHistoryLimit": 5,
                  "jobTargetRef": {
                    "activeDeadlineSeconds": 600,
                    "backoffLimit": 6,
                    "completions": 1,
                    "parallelism": 1,
                    "template": null
                  },
                  "maxReplicaCount": 100,
                  "pollingInterval": 30,
                  "successfulJobsHistoryLimit": 5,
                  "triggers": [
                    {
                      "metadata": {
                        "property": "example-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledObject",
                "metadata": {
                  "name": "example-scaledobject"
                },
                "spec": {
                  "cooldownPeriod": 300,
                  "maxReplicaCount": 100,
                  "minReplicaCount": 0,
                  "pollingInterval": 30,
                  "scaleTargetRef": {
                    "name": "example-deployment"
                  },
                  "triggers": [
                    {
                      "metadata": {
                        "property": "example-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "TriggerAuthentication",
                "metadata": {
                  "name": "example-triggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "example-secret-parameter"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Cloud Provider
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9-operator@sha256:9a359ae6612d2f22fe983653418143873cb9939f83dca6f008ec3cb8f3d53990
          createdAt: 20 Aug 2025, 21:31
          description: Custom Metrics Autoscaler Operator, an event-driven autoscaler
            based upon KEDA
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.1 <2.17.2-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-keda
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/custom-metrics-autoscaler-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudEventSource
            name: cloudeventsources.eventing.keda.sh
            version: v1alpha1
          - kind: ClusterCloudEventSource
            name: clustercloudeventsources.eventing.keda.sh
            version: v1alpha1
          - description: Describes cluster wide authentication parameters
            displayName: ClusterTriggerAuthentication
            kind: ClusterTriggerAuthentication
            name: clustertriggerauthentications.keda.sh
            version: v1alpha1
          - description: |
              Represents an installation of a particular version of KEDA Controller.
            displayName: KedaController
            kind: KedaController
            name: kedacontrollers.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your jobs and what the triggers
            displayName: ScaledJob
            kind: ScaledJob
            name: scaledjobs.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your application and what the
              triggers are
            displayName: ScaledObject
            kind: ScaledObject
            name: scaledobjects.keda.sh
            version: v1alpha1
          - description: Describes authentication parameters
            displayName: TriggerAuthentication
            kind: TriggerAuthentication
            name: triggerauthentications.keda.sh
            version: v1alpha1
        description: |
          ## About the managed application
          Custom Metrics Autoscaler for OpenShift is an event driven autoscaler based upon KEDA.  Custom Metrics Autoscaler can monitor event sources like Kafka, RabbitMQ, or cloud event sources and feed the metrics from those sources into the Kubernetes horizontal pod autoscaler.  With Custom Metrics Autoscaler, you can have event driven and serverless scale of deployments within any Kubernetes cluster.
          ## About this Operator
          The Custom Metrics Autoscaler Operator deploys and manages installation of KEDA Controller in the cluster. Install this operator and follow installation instructions on how to install Custom Metrics Autoscaler in your cluster.

          ## Prerequisites for enabling this Operator
          ## How to install Custom Metrics Autoscaler in the cluster
          The installation of Custom Metrics Autoscaler is triggered by the creation of `KedaController` resource. Please refer to the [KedaController Spec](https://github.com/openshift/custom-metrics-autoscaler-operator/blob/main/README.md#the-kedacontroller-custom-resource) for more details on available options.

          Only a resource named `keda` in the namespace where the operator is installed will trigger the installation, reconfiguration or removal of the KEDA Controller resource.

          There should be only one KEDA Controller in the cluster.

          The operator will behave in this manner whether it is installed with the `AllNamespaces` or `OwnNamespace` install mode. While the operator more closely matches the `OwnNamespace` semantics, `AllNamespaces` is a supported installation mode to allow it to be installed to namespaces with existing `OperatorGroups` which require that installation mode, such as `openshift-operators`.
        displayName: Custom Metrics Autoscaler
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keda
        - scaling
        - event
        links:
        - name: Custom Metrics Autoscaler Documentation
          url: https://docs.openshift.com/container-platform/latest/nodes/cma/nodes-cma-autoscaling-custom.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9-operator@sha256:9a359ae6612d2f22fe983653418143873cb9939f83dca6f008ec3cb8f3d53990
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9@sha256:ace9358fbd2d81a6e16bfa049c763ca16acf6cc2a59cdc0b6282e5e0dfc902cc
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-adapter-rhel9@sha256:0c8ae9323f927346e168ba829f6c7679af6b2dd359c8d0a020dd0ae3954481ce
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-admission-webhooks-rhel9@sha256:75631a21de057844e9ae45ae2496a22e16f3a32a6fd5f6cb5f90e173e3be2180
        version: 2.17.2-1
      entries:
      - name: custom-metrics-autoscaler.v2.17.2-1
        version: 2.17.2-1
      - name: custom-metrics-autoscaler.v2.15.1-6
        version: 2.15.1-6
      - name: custom-metrics-autoscaler.v2.15.1-4
        version: 2.15.1-4
      - name: custom-metrics-autoscaler.v2.14.1-467
        version: 2.14.1-467
      - name: custom-metrics-autoscaler.v2.14.1-454
        version: 2.14.1-454
      - name: custom-metrics-autoscaler.v2.13.1-427
        version: 2.13.1-427
      - name: custom-metrics-autoscaler.v2.12.1-394
        version: 2.12.1-394
      - name: custom-metrics-autoscaler.v2.11.2-322
        version: 2.11.2-322
      - name: custom-metrics-autoscaler.v2.11.2-311
        version: 2.11.2-311
      - name: custom-metrics-autoscaler.v2.10.1-267
        version: 2.10.1-267
      name: stable
    defaultChannel: stable
    packageName: openshift-custom-metrics-autoscaler-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com
    name: aws-load-balancer-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-load-balancer-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
          createdAt: 2022-06-28T17:56:41
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:d7bc364512178c36671d8a4b5a76cf7cb10f8e56997106187b0fe1f032670ece
        version: 0.0.1
      entries:
      - name: aws-load-balancer-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: aws-load-balancer-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
          createdAt: 2022-12-09T15:32:32
          olm.skipRange: <0.2.0
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:cfd210a2cb705156883fda126d052e05eb84345cc27ba04688067f8ed9a802ea
        version: 0.2.0
      entries:
      - name: aws-load-balancer-operator.v0.2.0
        version: 0.2.0
      name: stable-v0
    - currentCSV: aws-load-balancer-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
          createdAt: 2022-06-28T17:56:41
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:d7bc364512178c36671d8a4b5a76cf7cb10f8e56997106187b0fe1f032670ece
        version: 0.0.1
      entries:
      - name: aws-load-balancer-operator.v0.0.1
        version: 0.0.1
      name: stable-v0.1
    - currentCSV: aws-load-balancer-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
          createdAt: 2022-12-09T15:32:32
          olm.skipRange: <0.2.0
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:cfd210a2cb705156883fda126d052e05eb84345cc27ba04688067f8ed9a802ea
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        version: 0.2.0
      entries:
      - name: aws-load-balancer-operator.v0.2.0
        version: 0.2.0
      name: stable-v0.2
    - currentCSV: aws-load-balancer-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
          createdAt: 2024-01-23T11:21:00
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:2e0b9332a44d8d9c23e19c7accab0813a651f39210257820db508cac28876595
        version: 1.1.1
      entries:
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1
    - currentCSV: aws-load-balancer-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:6e15df39a3d25755a79888f739e1af8b052c3fba57bf8f6b3434ef58fd4288cf
          createdAt: 2024-01-24T09:13:21
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat <<EOF | oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
          EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note about UPI:__ For UPI based installs, additional documentation can be found in
          [VPC and subnets](https://github.com/openshift/aws-load-balancer-operator/blob/release-1.0/docs/prerequisites.md#vpc-and-subnets)

          __Note about STS:__ For clusters using STS, use the [STS documentation](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#for-sts-clusters) instead.
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9f9f8f69ab98c1f25cc301572339e0cc5c5b76f2e29ed26e10ac91ad20e0090c
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:6e15df39a3d25755a79888f739e1af8b052c3fba57bf8f6b3434ef58fd4288cf
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:73db97d20ee9415e9a8941cc6a36ed81d901d760d9949e95e10c90866ef8bf64
        version: 1.0.1
      entries:
      - name: aws-load-balancer-operator.v1.0.1
        version: 1.0.1
      - name: aws-load-balancer-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: aws-load-balancer-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
          createdAt: 2024-01-23T11:21:00
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:2e0b9332a44d8d9c23e19c7accab0813a651f39210257820db508cac28876595
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
        version: 1.1.1
      entries:
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    defaultChannel: stable-v1
    packageName: aws-load-balancer-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres-community
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.17.0-rc1
        version: 1.17.0-rc1
      - name: stackgres-community.v1.16.3
        version: 1.16.3
      - name: stackgres-community.v1.16.2
        version: 1.16.2
      - name: stackgres-community.v1.16.1
        version: 1.16.1
      - name: stackgres-community.v1.16.0
        version: 1.16.0
      - name: stackgres-community.v1.16.0-rc3
        version: 1.16.0-rc3
      - name: stackgres-community.v1.16.0-rc2
        version: 1.16.0-rc2
      - name: stackgres-community.v1.16.0-rc1
        version: 1.16.0-rc1
      - name: stackgres-community.v1.15.2
        version: 1.15.2
      - name: stackgres-community.v1.15.1
        version: 1.15.1
      - name: stackgres-community.v1.15.0
        version: 1.15.0
      - name: stackgres-community.v1.15.0-rc4
        version: 1.15.0-rc4
      - name: stackgres-community.v1.15.0-rc3
        version: 1.15.0-rc3
      - name: stackgres-community.v1.15.0-rc2
        version: 1.15.0-rc2
      - name: stackgres-community.v1.15.0-rc1
        version: 1.15.0-rc1
      - name: stackgres-community.v1.14.3
        version: 1.14.3
      - name: stackgres-community.v1.14.2
        version: 1.14.2
      - name: stackgres-community.v1.14.1
        version: 1.14.1
      - name: stackgres-community.v1.14.0
        version: 1.14.0
      - name: stackgres-community.v1.14.0-rc2
        version: 1.14.0-rc2
      - name: stackgres-community.v1.14.0-rc1
        version: 1.14.0-rc1
      - name: stackgres-community.v1.13.0
        version: 1.13.0
      - name: stackgres-community.v1.13.0-rc1
        version: 1.13.0-rc1
      - name: stackgres-community.v1.12.0
        version: 1.12.0
      - name: stackgres-community.v1.12.0-rc1
        version: 1.12.0-rc1
      name: candidate
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.17.0-rc1
        version: 1.17.0-rc1
      - name: stackgres-community.v1.16.3
        version: 1.16.3
      - name: stackgres-community.v1.16.2
        version: 1.16.2
      - name: stackgres-community.v1.16.1
        version: 1.16.1
      - name: stackgres-community.v1.16.0
        version: 1.16.0
      - name: stackgres-community.v1.16.0-rc3
        version: 1.16.0-rc3
      - name: stackgres-community.v1.16.0-rc2
        version: 1.16.0-rc2
      - name: stackgres-community.v1.16.0-rc1
        version: 1.16.0-rc1
      - name: stackgres-community.v1.15.2
        version: 1.15.2
      - name: stackgres-community.v1.15.1
        version: 1.15.1
      - name: stackgres-community.v1.15.0
        version: 1.15.0
      - name: stackgres-community.v1.15.0-rc4
        version: 1.15.0-rc4
      - name: stackgres-community.v1.15.0-rc3
        version: 1.15.0-rc3
      - name: stackgres-community.v1.15.0-rc2
        version: 1.15.0-rc2
      - name: stackgres-community.v1.15.0-rc1
        version: 1.15.0-rc1
      - name: stackgres-community.v1.14.3
        version: 1.14.3
      - name: stackgres-community.v1.14.2
        version: 1.14.2
      - name: stackgres-community.v1.14.1
        version: 1.14.1
      - name: stackgres-community.v1.14.0
        version: 1.14.0
      - name: stackgres-community.v1.14.0-rc2
        version: 1.14.0-rc2
      - name: stackgres-community.v1.14.0-rc1
        version: 1.14.0-rc1
      - name: stackgres-community.v1.13.0
        version: 1.13.0
      - name: stackgres-community.v1.13.0-rc1
        version: 1.13.0-rc1
      - name: stackgres-community.v1.12.0
        version: 1.12.0
      - name: stackgres-community.v1.12.0-rc1
        version: 1.12.0-rc1
      name: fast
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.16.3
        version: 1.16.3
      - name: stackgres-community.v1.16.2
        version: 1.16.2
      - name: stackgres-community.v1.16.1
        version: 1.16.1
      - name: stackgres-community.v1.16.0
        version: 1.16.0
      - name: stackgres-community.v1.15.2
        version: 1.15.2
      - name: stackgres-community.v1.15.1
        version: 1.15.1
      - name: stackgres-community.v1.15.0
        version: 1.15.0
      - name: stackgres-community.v1.14.3
        version: 1.14.3
      - name: stackgres-community.v1.14.2
        version: 1.14.2
      - name: stackgres-community.v1.14.1
        version: 1.14.1
      - name: stackgres-community.v1.14.0
        version: 1.14.0
      - name: stackgres-community.v1.13.0
        version: 1.13.0
      - name: stackgres-community.v1.12.0
        version: 1.12.0
      name: stable
    defaultChannel: stable
    packageName: stackgres-community
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Rebaca Technologies Pvt Ltd
      provider-url: https://www.rebaca.com/telecom/solutions
    name: abot-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: abot-operator-rhmp.3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "abot.com/v3",
                "kind": "Abot",
                "metadata": {
                  "name": "abot"
                },
                "spec": {
                  "abotanalytics": {
                    "image": "registry.connect.redhat.com/rebaca/abot-analytics@sha256:c854d9eafbe26f834e10f5a97a587b2a2bf05558554897817bb8e2556124b75d",
                    "resources": {
                      "limits": {
                        "cpu": "4000m",
                        "ephemeral-storage": "8Gi",
                        "memory": "6000Mi"
                      },
                      "requests": {
                        "cpu": "2000m",
                        "ephemeral-storage": "4Gi",
                        "memory": "4000Mi"
                      }
                    }
                  },
                  "abotanalyticsbase": {
                    "image": "registry.connect.redhat.com/rebaca/abot-analyticsbase@sha256:dcc2a417a889acd6ec63538a6dfa393d4c1942dac178a04e5d65189a1ae054dc",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotbaseimage": "registry.connect.redhat.com/rebaca/abot-baseimage@sha256:2cbe38298d26be5896e0dc071aae111dd921d0dbd05c3525290f6ff31c73e926",
                  "abotbdd": {
                    "image": "registry.connect.redhat.com/rebaca/abot-bdd@sha256:d5c6bfe12a548cbdb4e0d83d5f18e5ab0c17c3af5340e437b2d0c23740c9c9e6",
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "ephemeral-storage": "8Gi",
                        "memory": "2000Mi"
                      },
                      "requests": {
                        "cpu": "1000m",
                        "ephemeral-storage": "4Gi",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotgrafana": {
                    "image": "registry.redhat.io/rhel8/grafana@sha256:d894cd34f2d8cac30e0fab69bd869f4b0f258b5086a8fd7ead0c5886ce2c5f7a",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "1500Mi"
                      },
                      "requests": {
                        "cpu": "300m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotgui": {
                    "image": "registry.connect.redhat.com/rebaca/abot-gui@sha256:78b512d6bd223a1f5d35f4cbd221f6637e2c9267726ce8899ee5d86aaac82b6f",
                    "resources": {
                      "limits": {
                        "cpu": "4000m",
                        "ephemeral-storage": "8Gi",
                        "memory": "6000Mi"
                      },
                      "requests": {
                        "cpu": "2000m",
                        "ephemeral-storage": "4Gi",
                        "memory": "4000Mi"
                      }
                    }
                  },
                  "abotkibana": {
                    "image": "registry.connect.redhat.com/elastic/kibana@sha256:2be3542d50f51ad786ead958a399f26852339e9417279688aad194c6794336ce",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "abotlogstash": {
                    "image": "registry.connect.redhat.com/elastic/logstash@sha256:dbc25dc5c31dfee08e86221af9f549b50075b2ccf7f255779c5f022c54cd892e",
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "2000Mi"
                      }
                    }
                  },
                  "abotmongo": {
                    "image": "registry.connect.redhat.com/rebaca/abot-mongodb@sha256:be9059e7e8bc1636a84575416b823c20463228785598662ee81c27e4c3c987dc",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "abotnginx": {
                    "image": "registry.connect.redhat.com/rebaca/abot-nginx@sha256:3edc22b40cd30230768115efdceec3d054d71f78214b4100b7425d8484154378",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "abotrabbitmq": {
                    "image": "registry.connect.redhat.com/rebaca/abot-rabbitmq@sha256:e609bec067fe7740590bb272db27d1a7770db47a0fc802e2288c022e101c9e25",
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "abotsidecar": {
                    "alertmanagerimage": "registry.redhat.io/rhacm2/prometheus-alertmanager-rhel8@sha256:ecaee8ad24a9c7cb3b945ee6d2c1275b9de2a676e1868abc2687e687385b3f6a",
                    "kubestatemetricsimage": "registry.redhat.io/rhacm2/kube-state-metrics-rhel8@sha256:e6bb6fa0c155bebf64749368dd951c15f1dbfa3559e2ba6734d3661d3a95c00c",
                    "prometheusimage": "registry.connect.redhat.com/kasten/prometheus@sha256:1e490fcccdd50c1026150ba4cad7c29297d2b73b3d5c80240ff31509e666363d",
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "8000Mi"
                      },
                      "requests": {
                        "cpu": "1000m",
                        "memory": "6000Mi"
                      }
                    }
                  },
                  "elasticsearch": {
                    "image": "registry.connect.redhat.com/elastic/elasticsearch@sha256:f47fd7df80ab05635e31d733f6969d38205f9f81cf95da8594a047f3d07c59ae",
                    "replicas": 3,
                    "resources": {
                      "limits": {
                        "cpu": "1200m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "800m",
                        "memory": "1000Mi"
                      }
                    },
                    "volumeClaim": {
                      "storage": "50Gi"
                    }
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "initContainers": {
                    "image": "registry.access.redhat.com/ubi8/ubi@sha256:a965f33ee4ee57dc8e40a1f9350ddf28ed0727b6cf80db46cdad0486a7580f9d",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "protocol": "https",
                  "pvc": {
                    "abotvol": "100Gi",
                    "analyticsvol": "100Gi",
                    "databak": "10Gi",
                    "mongovol": "10Gi",
                    "prometheusvol": "10Gi"
                  },
                  "replicas": 1,
                  "security": {
                    "fsGroup": 1018450000,
                    "seLinuxOptions": {
                      "level": "s0:c123,c456"
                    }
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "name": "abot"
                  },
                  "storageclass": "ocs-storagecluster-cephfs"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/rebaca/abot-operator@sha256:d3e8e1b6988827b3c636bdfcb57cda26870089d90cf2a732be522f5474c0a734
          createdAt: "2024-10-03T20:32:24Z"
          description: The ABot Test Bench is a test automation tool for 4G, 5G and
            ORAN Telecom networks that follow specifications and use standards driven
            design and compliance as specified by the 3GPP consortium.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/abot-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/abot-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=1.0.0 <3.0.0'
          operatorframework.io/suggested-namespace: abot
          operators.openshift.io/valid-subscription: '["ABot Basic License for 15
            days"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: https://abot.rebaca.com/support/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Abot
            name: abots.abot.com
            version: v3
        description: A Helm based operator for deploying Abot application in Openshift
          cluster. The ABot Test Bench is a test automation tool for 4G, 5G and ORAN
          Telecom networks that follow specifications and use standards driven design
          and compliance as specified by the 3GPP consortium. It provides a platform
          for a user to quickly check for protocol compliance and test for load characteristics
          of a particular system under test.
        displayName: Abot Operator-v3.0.0
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 5G
        - 4G
        - ORAN
        - test tool
        - compliance
        - interop
        links:
        - name: Abot Operator
          url: https://abot-operator.domain
        maintainers:
        - email: suchismita.nag@rebaca.com
          name: Suchismita Nag
        - email: samir@rebaca.com
          name: Samir Chatterjee
        maturity: alpha
        provider:
          name: Rebaca Technologies Pvt Ltd
          url: https://www.rebaca.com/telecom/solutions
        relatedImages:
        - registry.connect.redhat.com/rebaca/abot-rabbitmq@sha256:e609bec067fe7740590bb272db27d1a7770db47a0fc802e2288c022e101c9e25
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel8@sha256:e6bb6fa0c155bebf64749368dd951c15f1dbfa3559e2ba6734d3661d3a95c00c
        - registry.connect.redhat.com/rebaca/abot-bdd@sha256:d5c6bfe12a548cbdb4e0d83d5f18e5ab0c17c3af5340e437b2d0c23740c9c9e6
        - registry.connect.redhat.com/rebaca/abot-baseimage@sha256:2cbe38298d26be5896e0dc071aae111dd921d0dbd05c3525290f6ff31c73e926
        - registry.connect.redhat.com/elastic/elasticsearch@sha256:f47fd7df80ab05635e31d733f6969d38205f9f81cf95da8594a047f3d07c59ae
        - registry.connect.redhat.com/rebaca/abot-analyticsbase@sha256:dcc2a417a889acd6ec63538a6dfa393d4c1942dac178a04e5d65189a1ae054dc
        - registry.connect.redhat.com/rebaca/abot-mongodb@sha256:be9059e7e8bc1636a84575416b823c20463228785598662ee81c27e4c3c987dc
        - registry.connect.redhat.com/rebaca/abot-nginx@sha256:3edc22b40cd30230768115efdceec3d054d71f78214b4100b7425d8484154378
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e34f912158985525f6efb3c4b675f20854fc677c867d12fe3cd2e971309428dc
        - registry.connect.redhat.com/rebaca/abot-operator@sha256:d3e8e1b6988827b3c636bdfcb57cda26870089d90cf2a732be522f5474c0a734
        - registry.redhat.io/rhel8/grafana@sha256:d894cd34f2d8cac30e0fab69bd869f4b0f258b5086a8fd7ead0c5886ce2c5f7a
        - registry.connect.redhat.com/elastic/logstash@sha256:dbc25dc5c31dfee08e86221af9f549b50075b2ccf7f255779c5f022c54cd892e
        - registry.connect.redhat.com/kasten/prometheus@sha256:1e490fcccdd50c1026150ba4cad7c29297d2b73b3d5c80240ff31509e666363d
        - registry.connect.redhat.com/rebaca/abot-analytics@sha256:c854d9eafbe26f834e10f5a97a587b2a2bf05558554897817bb8e2556124b75d
        - registry.connect.redhat.com/rebaca/abot-gui@sha256:78b512d6bd223a1f5d35f4cbd221f6637e2c9267726ce8899ee5d86aaac82b6f
        - registry.connect.redhat.com/elastic/kibana@sha256:2be3542d50f51ad786ead958a399f26852339e9417279688aad194c6794336ce
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel8@sha256:ecaee8ad24a9c7cb3b945ee6d2c1275b9de2a676e1868abc2687e687385b3f6a
        - registry.access.redhat.com/ubi8/ubi@sha256:a965f33ee4ee57dc8e40a1f9350ddf28ed0727b6cf80db46cdad0486a7580f9d
        version: 3.0.0
      entries:
      - name: abot-operator-rhmp.3.0.0
        version: 3.0.0
      - name: abot-operator.2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: abot-operator-rhmp
    provider:
      name: Rebaca Technologies Pvt Ltd
      url: https://www.rebaca.com/telecom/solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: secrets-store-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secrets-store-csi-driver-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9-operator@sha256:96f3a8df9ac5ff3828b0add052af00f7d9df0ebb622a6adfd9772658d55baeb8
          createdAt: "2023-06-12T00:00:00Z"
          description: Install and configure Secrets Store CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.13.0-0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/secrets-store-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SecretProviderClass is the Schema for the secretproviderclasses
              API
            displayName: Secret Provider Class
            kind: SecretProviderClass
            name: secretproviderclasses.secrets-store.csi.x-k8s.io
            version: v1
          - description: SecretProviderClassPodStatus is the Schema for the secretproviderclassespodstatus
              API
            displayName: Secret Provider Class Pod Status
            kind: SecretProviderClassPodStatus
            name: secretproviderclasspodstatuses.secrets-store.csi.x-k8s.io
            version: v1
        description: |
          Operator that installs and configures the CSI driver for Secrets Store. The Secrets Store CSI Driver Operator is a Technology Preview feature. For more information on Red Hat Technology Preview features support scope, see https://access.redhat.com/support/offerings/techpreview/.
        displayName: Secrets Store CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: Documentation
          url: https://github.com/openshift/secrets-store-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/secrets-store-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: preview
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0627c09fe3f5a015f117fb30af1f4d17340aefa48b23994737d3901a804b63d3
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:a5704d26d0b562dba9633d1c107f34e05c750c0e3a64dbcace4dab7b964fff8d
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9-operator@sha256:96f3a8df9ac5ff3828b0add052af00f7d9df0ebb622a6adfd9772658d55baeb8
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9@sha256:148672ccd882a43c5c1d52763da465558b5ed54310418ec40c7ca0b56e408b1a
        version: 4.16.0-202509240123
      entries:
      - name: secrets-store-csi-driver-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: secrets-store-csi-driver-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: secrets-store-csi-driver-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: secrets-store-csi-driver-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: secrets-store-csi-driver-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: secrets-store-csi-driver-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: secrets-store-csi-driver-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: secrets-store-csi-driver-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: secrets-store-csi-driver-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: secrets-store-csi-driver-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: secrets-store-csi-driver-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: secrets-store-csi-driver-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: secrets-store-csi-driver-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: secrets-store-csi-driver-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: secrets-store-csi-driver-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: secrets-store-csi-driver-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: secrets-store-csi-driver-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: secrets-store-csi-driver-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: secrets-store-csi-driver-operator.v4.16.0-202410031007
        version: 4.16.0-202410031007
      - name: secrets-store-csi-driver-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: secrets-store-csi-driver-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: secrets-store-csi-driver-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: secrets-store-csi-driver-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: secrets-store-csi-driver-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: secrets-store-csi-driver-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: secrets-store-csi-driver-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: secrets-store-csi-driver-operator.v4.16.0-202408021139
        version: 4.16.0-202408021139
      - name: secrets-store-csi-driver-operator.v4.16.0-202407190036
        version: 4.16.0-202407190036
      - name: secrets-store-csi-driver-operator.v4.16.0-202407111006
        version: 4.16.0-202407111006
      - name: secrets-store-csi-driver-operator.v4.16.0-202407081937
        version: 4.16.0-202407081937
      - name: secrets-store-csi-driver-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: secrets-store-csi-driver-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: secrets-store-csi-driver-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: secrets-store-csi-driver-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: preview
    defaultChannel: preview
    packageName: secrets-store-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/limitador-operator
    name: limitador-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: limitador-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "name": "toy_get_route",
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/limitador-rhel9-operator@sha256:edb28f112a0d3536f2f52467f38df685d953f46f1999cbd61f4fd4fa4805cd34
          createdAt: "2025-04-07T15:52:04Z"
          description: Enables rate limiting for Gateways and applications in a Gateway
            API network
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Limitador is the Schema for the limitadors API
            displayName: Limitador
            kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: Enables rate limiting for Gateways and applications in a Gateway
          API network. - Part of Red Hat Connectivity Link
        displayName: Red Hat - Limitador Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        - kuadrant
        links:
        - name: GitHub
          url: https://github.com/kuadrant/limitador-operator
        - name: Documentation
          url: https://github.com/Kuadrant/limitador/tree/main/docs
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - registry.redhat.io/rhcl-1/limitador-rhel9-operator@sha256:edb28f112a0d3536f2f52467f38df685d953f46f1999cbd61f4fd4fa4805cd34
        - registry.redhat.io/rhcl-1/limitador-rhel9@sha256:98f85a05da53fbf7f20845d1969441620407cc13a185fd52714308679f116425
        version: 1.1.0
      entries:
      - name: limitador-operator.v1.1.0
        version: 1.1.0
      - name: limitador-operator.v1.0.2
        version: 1.0.2
      - name: limitador-operator.v1.0.1
        version: 1.0.1
      - name: limitador-operator.v0.12.1
        version: 0.12.1
      name: stable
    defaultChannel: stable
    packageName: limitador-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/limitador-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: run-once-duration-override-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runoncedurationoverrideoperator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "RunOnceDurationOverride",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "runOnceDurationOverride": {
                    "spec": {
                      "activeDeadlineSeconds": 3600
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certifiedLevel: "false"
          containerImage: registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9-operator@sha256:54993506a3a2fc1e88c9b5aac97ae90d2a04a06ae6e3a97bf487b5e8996b95fb
          createdAt: 2024/07/01
          description: An operator to manage the OpenShift RunOnceDurationOverride
            Mutating Admission Webhook Server
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.1.2'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/run-once-duration-override-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RunOnceDurationOverride is the schema for the runoncedurationoverrides
              API
            displayName: Run Once Duration Override
            kind: RunOnceDurationOverride
            name: runoncedurationoverrides.operator.openshift.io
            version: v1
        displayName: Run Once Duration Override Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mutating
        - webhook
        - workload
        - run-once
        links:
        - name: Source Code
          url: https://github.com/openshift/run-once-duration-override
        - name: Source Code
          url: https://github.com/openshift/run-once-duration-override-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9-operator@sha256:54993506a3a2fc1e88c9b5aac97ae90d2a04a06ae6e3a97bf487b5e8996b95fb
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9@sha256:3dc6e2a5900472e3e083abcb659a13fea3d08f7a9ff167df524b6f890a16dc8e
        version: 1.1.2
      entries:
      - name: runoncedurationoverrideoperator.v1.1.2
        version: 1.1.2
      - name: runoncedurationoverrideoperator.v1.1.1
        version: 1.1.1
      - name: runoncedurationoverrideoperator.v1.1.0
        version: 1.1.0
      - name: runoncedurationoverrideoperator.v1.0.2
        version: 1.0.2
      - name: runoncedurationoverrideoperator.v1.0.1
        version: 1.0.1
      - name: runoncedurationoverrideoperator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: run-once-duration-override-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicast-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicast-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:ccd64090b85536ddaf1b81ca07d40279c9f277b1c5911628f11d678b98dc3bad
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:ccd64090b85536ddaf1b81ca07d40279c9f277b1c5911628f11d678b98dc3bad
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:f0e9f64c353fdbee3bf7ec0eaf070c3b75330ef659a881ec4ebc874ba584d085
        version: 0.5.3
      entries:
      - name: apicast-operator.v0.5.3
        version: 0.5.3
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      name: threescale-2.11
    - currentCSV: apicast-operator.v0.6.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:8c7fe66b8d4044ad965e7b3d149e9d8e3846ac4e55465873f7095a25b297b3df
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:8c7fe66b8d4044ad965e7b3d149e9d8e3846ac4e55465873f7095a25b297b3df
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:12f7a8e69a1369a51bbcd2b34d4ec8c66dd1fc9fb515349a04e7b202953423e3
        version: 0.6.4
      entries:
      - name: apicast-operator.v0.6.4
        version: 0.6.4
      - name: apicast-operator.v0.6.3
        version: 0.6.3
      - name: apicast-operator.v0.6.2
        version: 0.6.2
      - name: apicast-operator.v0.6.1
        version: 0.6.1
      - name: apicast-operator.v0.6.0
        version: 0.6.0
      - name: apicast-operator.v0.5.3
        version: 0.5.3
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      name: threescale-2.12
    - currentCSV: apicast-operator.v0.7.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:bde44c0fb9f13f212184b24e0bef45db734a8439a92179a2089e5aebd6b87b63
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          olm.skipRange: '>=0.6.2 <0.7.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.13/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:bde44c0fb9f13f212184b24e0bef45db734a8439a92179a2089e5aebd6b87b63
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:88dc7a289c803248efbe151999707325b1bc0aac2d26827ea1707e46e5c71652
        version: 0.7.5
      entries:
      - name: apicast-operator.v0.7.5
        version: 0.7.5
      name: threescale-2.13
    - currentCSV: apicast-operator.v0.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:c5e49be7b43a0058f974a5496976255ab3fbd0677bb7bcf9c3438341cf55965a
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.2 <0.8.3'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.14/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:c5e49be7b43a0058f974a5496976255ab3fbd0677bb7bcf9c3438341cf55965a
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:f8cd25658fa6a154e7ea6e5adf0f9dae3fc596ec9989021de6a021ec94eef226
        version: 0.8.3
      entries:
      - name: apicast-operator.v0.8.3
        version: 0.8.3
      - name: apicast-operator.v0.8.2
        version: 0.8.2
      - name: apicast-operator.v0.8.1
        version: 0.8.1
      name: threescale-2.14
    - currentCSV: apicast-operator.v0.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel9-operator@sha256:c4d6acae4e4cc3236c5b8d8dcfb339b072688ab114785d8598bbe68b1f3571a9
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=0.8.0 <0.12.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.14/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel9-operator@sha256:c4d6acae4e4cc3236c5b8d8dcfb339b072688ab114785d8598bbe68b1f3571a9
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        version: 0.12.5
      entries:
      - name: apicast-operator.v0.12.5
        version: 0.12.5
      - name: apicast-operator.v0.12.4
        version: 0.12.4
      - name: apicast-operator.v0.12.3
        version: 0.12.3
      - name: apicast-operator.v0.12.2
        version: 0.12.2
      - name: apicast-operator.v0.12.1
        version: 0.12.1
      - name: apicast-operator.v0.12.0
        version: 0.12.0
      name: threescale-2.15
    - currentCSV: apicast-operator.v0.12.1-mas
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-mas/apicast-rhel9-operator@sha256:e9f22a8af53f881f1948f5d6a033c35a530895633f2eae522b2fb67f9dc6088b
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.17"}]'
          olm.skipRange: '>=0.7.9 <0.12.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-mas/apicast-gateway-rhel8@sha256:f174636a9ad43a2419fff5a9e6797e3185c540a9509188fd1606aef4acea7bdf
        - registry.redhat.io/3scale-mas/apicast-rhel9-operator@sha256:e9f22a8af53f881f1948f5d6a033c35a530895633f2eae522b2fb67f9dc6088b
        version: 0.12.1-mas
      entries:
      - name: apicast-operator.v0.12.1-mas
        version: 0.12.1-mas
      - name: apicast-operator.v0.7.9-mas
        version: 0.7.9-mas
      - name: apicast-operator.v0.7.8-mas
        version: 0.7.8-mas
      - name: apicast-operator.v0.7.7-mas
        version: 0.7.7-mas
      - name: apicast-operator.v0.7.6-mas
        version: 0.7.6-mas
      - name: apicast-operator.v0.7.5-mas
        version: 0.7.5-mas
      - name: apicast-operator.v0.7.4-mas
        version: 0.7.4-mas
      - name: apicast-operator.v0.7.3-mas
        version: 0.7.3-mas
      - name: apicast-operator.v0.7.2-mas
        version: 0.7.2-mas
      - name: apicast-operator.v0.7.1-mas
        version: 0.7.1-mas
      - name: apicast-operator.v0.7.0-mas
        version: 0.7.0-mas
      - name: apicast-operator.v0.6.0-mas
        version: 0.6.0-mas
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      name: threescale-mas
    defaultChannel: threescale-2.15
    packageName: apicast-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: cloud-bulldozer
      provider-url: ""
    name: ripsaw
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ripsaw.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery, Cloud Provider
          containerImage: quay.io/cloud-bulldozer/benchmark-operator:latest
          description: '[Benchmark-operator](https://github.com/cloud-bulldozer/ripsaw/)
            is a benchmarking tool to benchmark your kubernetes cluster. The intent
            of this Operator is to deploy common workloads to establish a performance
            baseline of Kubernetes cluster'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/cloud-bulldozer/benchmark-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Benchmark
            name: benchmarks.ripsaw.cloudbulldozer.io
            version: v1alpha1
        description: '[Benchmark-operator](https://github.com/cloud-bulldozer/benchmark-operator/)
          is a benchmarking tool to benchmark your kubernetes cluster. The intent
          of this Operator is to deploy common workloads to establish a performance
          baseline of Kubernetes cluster'
        displayName: benchmark-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - performance
        - benchmark
        links:
        - name: Site
          url: https://github.com/cloud-bulldozer/benchmark-operator
        - name: Documentation
          url: https://github.com/cloud-bulldozer/benchmark-operator/tree/master/docs
        maintainers:
        - email: rsevilla@redhat.com
          name: rsevilla87
        - email: jtaleric@redhat.com
          name: jtaleric
        - email: rzaleski@redhat.com
          name: dry939
        - email: kwhitley@redhat.com
          name: whitleykeith
        maturity: alpha
        provider:
          name: cloud-bulldozer
        relatedImages:
        - quay.io/cloud-bulldozer/benchmark-operator:latest
        - k8s.gcr.io/redis:v1
        version: 0.1.0
      entries:
      - name: ripsaw.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ripsaw
    provider:
      name: cloud-bulldozer
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: stolostron-engine
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stolostron-engine.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:b7d0c8413a80c251779fe3c1c9200f9fcfbbe085ccee3ab968ebd4a39209946c
          createdAt: "2024-06-28T02:42:26Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/work@sha256:4ef2518aa107ebd28d5fb1a98d63d82a8e0610544600169e0dd5f345599e518b
        - quay.io/stolostron/assisted-installer@sha256:a31ed7b8e4ff2f9e5ade2e61191cc759e1d47ddbba95a2af41bfa53fa6ddf855
        - quay.io/stolostron/cluster-api-provider-agent@sha256:c074ca1f28d03e86efe4f09200a38af4ac947247a19c46c5300eefcd4f337006
        - quay.io/stolostron/placement@sha256:0dbcc7c62569920afcad1443529f10e38d651a5fe5ae14ade89d6e42fb522e60
        - quay.io/stolostron/managed-serviceaccount@sha256:4c3f1e0b9c4f2937ceb46e6b3f63e3a07061f9b29fdda7dcfdf9ebd74c3bb30b
        - quay.io/stolostron/backplane-operator@sha256:b7d0c8413a80c251779fe3c1c9200f9fcfbbe085ccee3ab968ebd4a39209946c
        - quay.io/stolostron/assisted-installer-controller@sha256:213dc6fde247a5b9862f5076d40394164dd596cde7ca2061e58b73f4d2dbc89e
        - quay.io/stolostron/hive@sha256:fd39b8340cfaeb9849e83bd910e8a3575791fcc58dbbdeb3e66528f009b33d65
        - quay.io/stolostron/cluster-image-set-controller@sha256:632661c6cbfce7df0545bc2fccf1ef87257abb3cc92dcede054806613a3540e3
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:f0d6d46d3a4d65fba6b74bdc9eab99e8149133f83a5c8486f22a73ab92f78ed6
        - quay.io/stolostron/assisted-service@sha256:653728f209070c6adad846a4780449f56985b19006024cfe8137a3a1854fdc85
        - quay.io/stolostron/registration@sha256:8dca354dd63739d7578c24d57b2a72c0676c979344e05c6f2634a8a9fae3218d
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:26f8bb3c9b531e96e6538a5bff96569628f45c06eb3b991ebb421d4e90d7c7e3
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/registration-operator@sha256:4b6c31592c405db11375f4694a7f15bb13c5c9a770786a1607ea0be615396435
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/discovery-testserver@sha256:83d7d6cfaef629a4e3d5d0e9fe35456aa220f1eb68f6fe611d85232acf46bc73
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/discovery-operator-tests@sha256:c98c42b0c16306de86d7b365096cb809a0c797b0a4b7dc04a7de877d0ba5733a
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:b76a201edcad4805e32c40d3f0ad3f61584a879aa00c9e49aae3b2985bfb5fbc
        - quay.io/stolostron/addon-manager@sha256:87c829911438061a53aff2b297416bf43080fdd8d93558caefc198e6f9cb382f
        - quay.io/stolostron/backplane-operator-test@sha256:961d8c40aa3526ca368fd6833202800ba2eb344c97788a646d6fb350a5338397
        - quay.io/stolostron/hypershift-operator@sha256:7dd5eaf195e6b1ca4cecd50514f5b7045948184f9ee617341bc90edc5f9914f2
        - quay.io/stolostron/backplane-must-gather@sha256:2676ebb7438ca42f0d852100612715c429fe9fbe072297fc178d51c34378d1af
        - quay.io/stolostron/discovery-operator@sha256:acb2c4033da62055feb49ff3023b2850a2768996e1fa80f52588b3e8e458e8f5
        - quay.io/stolostron/assisted-image-service@sha256:0c29a74e0f82d1782d5170ff674f5bfcd89ecc5d180fe5b583a5e0f388b9504b
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:a39c52097faf7a21cc717768401851466f8c9341f08103e4d1fe8459ae7f5ea2
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/helloworld@sha256:abc0d8002c5fe6d95a1f24fce1b7899ca24fac576c894ea7443cf7e15e6a42c2
        - quay.io/stolostron/image-based-install-operator@sha256:b9af0ddb7ae8e6fd29b953dfb917546dfd676206e2ad143e892172b6e1cea5a1
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:b269d88e30551636e024aa1472ea4dd71b873b5a2778a199b32732488fd776fb
        - quay.io/stolostron/multicloud-manager@sha256:05da692b2780453115ef6c195658cadb2f28eb0f3a86dc0a586e459dd7c55afe
        - quay.io/stolostron/cluster-curator-controller@sha256:e3a91a4b06dd73fc4e401ddf034e61d4d5b2d324d2c5cea92941e1a56c4023c1
        - quay.io/stolostron/hypershift-addon-operator@sha256:47ed6a7c00036599051718d6015bee35818ea5bbaa14f1c01bedb338dce90668
        - quay.io/stolostron/assisted-installer-agent@sha256:a2efc62ca0080779acfea2e413e524608bc0e8f72b2b917e6c7e76864d08b2b6
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/provider-credential-controller@sha256:d3bd513c06591786216f2baab46c9f57e373d1939f038edacecb2180f673d888
        - quay.io/stolostron/cluster-proxy@sha256:6cd53c588bb9c5064ba0eb8eab03cd12c86976b353209102059f7d76ec59d7c3
        - quay.io/stolostron/cluster-proxy-addon@sha256:f70a26b276c67ad791d337e35cd102afa82f1005bc6ecbc60309932d2d163a6b
        - quay.io/stolostron/managedcluster-import-controller@sha256:925bbc6173501bdb60d329a1f5a134ea3e9230034509320f93382f1c570be275
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/console-mce@sha256:4e49f2b2fdb0491c225e4d5dc143a6d7c0a25cd855057b3792c7ee94a868ac85
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/assisted-service-el8@sha256:31422ffce8b94d662131986bd084fcba503809cb939389096f9c9dbb27cb58db
        - quay.io/stolostron/clusterclaims-controller@sha256:a0613147b0ec010b8a4442a3c515f7ffc48ec7d8137436a0e8b935443cac2f96
        version: 0.4.0
      entries:
      - name: stolostron-engine.v0.4.0
        version: 0.4.0
      name: community-0.4
    - currentCSV: stolostron-engine.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:30d5bcccc542988ac788dbca305e6f29fb678470b19e48adada9ca6f5803f04a
          createdAt: "2024-10-02T05:51:59Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.5.0 <0.5.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/assisted-installer@sha256:dcc4e28c7eea55dd2ac22dd75c2b7b96ee3552fcbeba8128ad8528fda235d879
        - quay.io/stolostron/cluster-curator-controller@sha256:fc2ad8e0b6701638ab89bb62a49700fdaadfe1149228cd0b3cd4a41810020c6d
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/work@sha256:4467a4bd00586ab837cb7e8bf56ea1faff82ed5ce2fffdeb7e51ee5fafb433b0
        - quay.io/stolostron/cluster-image-set-controller@sha256:654e218ecf0a3f67520c0521b544eea9d4552e92f54bf424e88ccab24e7c151e
        - quay.io/stolostron/assisted-installer-agent@sha256:16b040dd0690b62bc051aa6369272d004d2157a8e5230d318ba827adba8e8d63
        - quay.io/stolostron/hypershift-addon-operator@sha256:ba199ebfe0eea3ef6b07123f51116831e7b0a73860b0410ece723bcc9a3230af
        - quay.io/stolostron/placement@sha256:dc0ad451b085b3a747ee648475f8e8102b645df69b898fc88f1cb190d1b95a83
        - quay.io/stolostron/cluster-proxy@sha256:be09bb293a990600ec48aba2f9bf8b457d5bdbc6b136be319e5feebadb232a8c
        - quay.io/stolostron/discovery-operator@sha256:db56e2a78ed27a46b47b519e29db66636b9abb499fd507d99451bae47438e32b
        - quay.io/stolostron/hive@sha256:707ac9a2ad18dbc4c4686f1b72946077c7232afd05505d5e50e13f806a452392
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:4b5b2bea2aad6c7e16fe1520a2fb8454b0312f2749d2f398ca54e644e813055c
        - quay.io/stolostron/cluster-api-provider-agent@sha256:1458de75311675a5acec7361e507edb24b970ffa3403e916de3413f7061bfe36
        - quay.io/stolostron/clusterclaims-controller@sha256:1c0b3c4730ef1ab628bf2c1b7cf12b1b93b149abdefc4d908dbdc90e4b6c1c19
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:2fe53994fa5bde93112a2cc75758e0f68be013a34c8206066d0b315c2e6a6f47
        - quay.io/stolostron/assisted-service-el8@sha256:c8248f9a7ac02f47373d097953cee38b66125e98b98922f18c578a591f04bcca
        - quay.io/stolostron/console-mce@sha256:10a54c5012ef6439539297919c71d9b53aaaddc6342d90677113d6ce7b42e3fe
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/backplane-operator-test@sha256:048f21cf89fb5ed2557ffc90ae5d7104cdd5ce0e6a2a865ce6f4230d09400115
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:b76a201edcad4805e32c40d3f0ad3f61584a879aa00c9e49aae3b2985bfb5fbc
        - quay.io/stolostron/addon-manager@sha256:9384ae0e9bcc9f85d3fe36e44ab7b3a61777197846559b5add0c56f513e35333
        - quay.io/stolostron/assisted-image-service@sha256:703c12a918d3c66b62c29a5abd0e8c5032299289cfed459671d6ca82c4f35c73
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/assisted-service@sha256:6383ba508a0cac030a46df96ac37daaeddc27ad1fb41d8b4dca5a87cf4ab9c31
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:14a0648b76301de182154a8d7d6b7e7eca0bfbed226e6dcebddb75af5152451f
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/registration-operator@sha256:b42341143b1e1693748f4b7d69e97c56c8a5f9408abc5db1de96ba246b7b98a9
        - quay.io/stolostron/backplane-operator@sha256:30d5bcccc542988ac788dbca305e6f29fb678470b19e48adada9ca6f5803f04a
        - quay.io/stolostron/discovery-operator-tests@sha256:e8d3d1675242272cc139b0dba15605d00e25b17b4b1cc1b5b4f3ea8913ffb2da
        - quay.io/stolostron/discovery-testserver@sha256:0b23021d6c0c7d6c7aa97a92fd0c40f31fdff6363a5b10a1edc864b5d8d54108
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/multicloud-manager@sha256:6dd212840ea7fadd427b503ce69e6e32464881a0047e2af86f9f2c7562bd4703
        - quay.io/stolostron/managedcluster-import-controller@sha256:01ab0c8b664692d69c2a9b3363e007a5611c332ac0e3491234dd0cbfba31c88a
        - quay.io/stolostron/registration@sha256:4779cf430d479806efd2f3cb7e2a9ed3dce749287609c39ad45f1b06e220d023
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/managed-serviceaccount@sha256:667f823799984404c0c1589df7b857a6a35218c2653254905dc22ca90ce1542b
        - quay.io/stolostron/backplane-must-gather@sha256:92e992e23855375a8bae542480d789d1bb25a48959896bfb75f659df2a37dd98
        - quay.io/stolostron/image-based-install-operator@sha256:bd76503406e6d20693e636461411e158c8dde3d03c4904ecb8a1850deeefa4f5
        - quay.io/stolostron/hypershift-operator@sha256:a30fb6cf4d5e103820813019ce452d8134b0fee8ecd3c2166d5681199d4a0cad
        - quay.io/stolostron/cluster-proxy-addon@sha256:7aa09b0d53fdc0f5cec9c8a13c160277a9274466025f11b5b04da95c9a10e59c
        - quay.io/stolostron/provider-credential-controller@sha256:d240ad43c23de822f91a33849448ae70a4cf27570da3387bdecd101ca2f0b65d
        - quay.io/stolostron/assisted-installer-controller@sha256:01dd7c7fb9cb335215afefde204309d3127847b900658ee9744b9d31df847565
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/helloworld@sha256:96a9b8855fb8800a85aed6fdc4911295dc2c12629bbdcc34ed8dc8061827b17e
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        version: 0.5.3
      entries:
      - name: stolostron-engine.v0.5.3
        version: 0.5.3
      - name: stolostron-engine.v0.5.2
        version: 0.5.2
      - name: stolostron-engine.v0.5.1
        version: 0.5.1
      - name: stolostron-engine.v0.5.0
        version: 0.5.0
      name: community-0.5
    - currentCSV: stolostron-engine.v0.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:fa1fa6ef3958fd9d50bfb38197f033a635a05b47c32e653822246e708d8598cc
          createdAt: "2025-02-24T01:37:53Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.6.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/discovery-operator-tests@sha256:21449ecd291cf7203cc8e27a2a00ac60e74f5993708f9e7465e8efdfdca1c89f
        - quay.io/stolostron/assisted-installer@sha256:c88976ab4b35d458a39044e522d2d5ab1916d8485af07bbef373af1bcbbd6af0
        - quay.io/stolostron/assisted-service-el8@sha256:f50f4dece91100f45ef9e20b1e52557a55db45b17a380f32d365ee013d695d5c
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:b016a001a53bc37ae534b74cc832552759800746c27e16adcf0626ff9ad2e4c0
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:2781d87c1568f2e40409f5431bfd1a9a4c9ca35655537e7d4db65d9b881edd08
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        - quay.io/stolostron/placement@sha256:543f09446652bff8c4149945d94d715eff7a76d82783f2bad89e712798542c73
        - quay.io/stolostron/provider-credential-controller@sha256:a5ac3999d0aefb1dd9a239f9394485f93e4252a45eb4a43a43e603aaff80b3bb
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:44f8bb72b975456d4151a88a560fc525ad6f4da7968a38e94a86b6dd77d3c543
        - quay.io/stolostron/registration@sha256:9c221a40ffa119ac154d76f578e0919b5a63137c908afb0583519a9b8a399b2b
        - quay.io/stolostron/hypershift-operator@sha256:2d1809223ce5e75b4133a39b78a269703c6b5234f7ec2a7c0072124c345fae0a
        - quay.io/stolostron/clusterclaims-controller@sha256:b796cb2edd3d4c22b29fc6d2b45e5328a59b7ec1bdc52d5500e1a4ee74725d6c
        - quay.io/stolostron/cluster-api-provider-agent@sha256:dda3e9aef611c744c4c89c2689a12d0649a173c1a64954c3037de55ec8eacec9
        - quay.io/stolostron/multicloud-manager@sha256:0293f0852206562b891375b167cfe17843e2e17585f04c9bb6482d1fe78089d0
        - quay.io/stolostron/backplane-operator@sha256:fa1fa6ef3958fd9d50bfb38197f033a635a05b47c32e653822246e708d8598cc
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/discovery-testserver@sha256:f757a7a11108a88de4bc0ecc5e16f74b3bd9b5110ec94bfeef2074121ec1e47f
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/cluster-proxy-addon@sha256:7da9da3e4ecd1e0df636c9f72f9f30cc7ff24b81e247ba9825cef39f18e86182
        - quay.io/stolostron/image-based-install-operator@sha256:fecbff73eb97b28bedd7c17d5adbef4826f181a6838ad4cc72e7478acf9b4254
        - quay.io/stolostron/hive@sha256:1e1f60492c6a3b43b69fe8a0640f9b81f5a1cfafc0afcd6854c3cfe362ed063a
        - quay.io/stolostron/cluster-proxy@sha256:b341dd7058cdffccb00a026908031fa123fec5ba58ba9307dbb873b75d250e4d
        - quay.io/stolostron/cluster-image-set-controller@sha256:629361c1c0ec50fe6410fc6fe6417cba89957c49ab89201c6470c0f97c8fc91d
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/assisted-image-service@sha256:6846719f91dd954ce4b4c6e7c9654c7d79db72990270acd3c34bdd2a89af34fc
        - quay.io/stolostron/console-mce@sha256:bfaf9180cd1e5af4d34daf4f1dcdee89951067a605f21bfe6e1ce6bda85b0334
        - quay.io/stolostron/helloworld@sha256:4838f699ea2461f8358dcda32b3ac96ed05b55d4e1958eac2bdacae4807501d1
        - quay.io/stolostron/ose_cluster_api_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/assisted-installer-controller@sha256:3b0b43fdbbbbd5c1e7588fc449b002e1ed9ad0ea6daa774925a962bd499a4195
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/work@sha256:6ec081826f32467ea1e1a91f6624df349a382fb8a1cbbda906141281ac2bad5c
        - quay.io/stolostron/ose_aws_cluster_api_controllers_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/backplane-operator-test@sha256:b04661ad658423c9c7557b722adb2159d988c10445de1c59f39c44583262cb42
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/assisted-installer-agent@sha256:cd74c8b8888c100bf567dd07cfff40097c458d18fcc29a11ed3830f697156267
        - quay.io/stolostron/backplane-must-gather@sha256:99b558d687e1c17036b164d819cb1c9b96a814be79e3c87960ef7cdb521eee7d
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/discovery-operator@sha256:43f40b9c20b112248fcdb42d114fe4af73fdd45b05d9ff7855d827cdd5e971b4
        - quay.io/stolostron/managed-serviceaccount@sha256:0e4edd9ab95c1558fba40a0c41c5cc326996e9e19e296324ea10b5a442a7ef63
        - quay.io/stolostron/hypershift-addon-operator@sha256:65f2b7a356c22413307ce7b7de38ce00c31e4c43b6ee3b4b484c0bc61dc483e9
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:65180432218d035ffc7996a7643ddb5300eb8b09e47e09cb270cb9b961aecf7a
        - quay.io/stolostron/managedcluster-import-controller@sha256:ecfe7c1f2d24a75bfb8db2a0a2765fe5e9136f81f341882de896431570fd10e9
        - quay.io/stolostron/addon-manager@sha256:7e347a177d8c5b80a78411a3e8dbf04e49d157a4ca67e7fd5170970c2ba62da9
        - quay.io/stolostron/registration-operator@sha256:52289e6cf5f159d344717b49e6226ff9d4ee33942f63cc543d1b820d7ee24fc5
        - quay.io/stolostron/assisted-service@sha256:58271e54eda979c43fb474c3bb255f7328e380a7af9ba833e5f7558f323bd239
        - quay.io/stolostron/cluster-curator-controller@sha256:5d05462c2c9d93b21656a436c01fbda4256757bd057e5de7014eb6f320310a1e
        version: 0.6.3
      entries:
      - name: stolostron-engine.v0.6.3
        version: 0.6.3
      - name: stolostron-engine.v0.6.2
        version: 0.6.2
      - name: stolostron-engine.v0.6.1
        version: 0.6.1
      - name: stolostron-engine.v0.6.0
        version: 0.6.0
      name: community-0.6
    - currentCSV: stolostron-engine.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:9b39a5b6a502778d7884e276b276a9b44e6dbc88e7e82470e26d91ded3cb3839
          createdAt: "2025-06-09T15:04:58Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.7.1'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational components for managing multiple Kubernetes-based clusters. The deployment of the multicluster engine components is
              determined based on the configuration that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/cluster-image-set-controller@sha256:7f05ad13ffaa11daec066c5889b87df1b631e6eb3dc1c2913a2766c8e65218aa
        - quay.io/stolostron/cluster-proxy@sha256:1528e36dfa4fa2c9063d84c0f2b9aa8d2b549a750e812d4a2e2b8c7515bfe5cb
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:1e95f55d6918263e79de1c1e2be8a96cd3d35a346dde6d3e2fb484f48e567811
        - quay.io/stolostron/hive@sha256:86ff94d3e7c174ac4d6818183c66791edc7638965dd09e2d41172609c81f451f
        - quay.io/stolostron/cluster-api-provider-openshift-assisted-bootstrap@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/console-mce@sha256:7f538a21eb22ccd57765f47692da23be9cecfd3b7b65986f1413234bdf709530
        - quay.io/stolostron/assisted-installer-agent@sha256:ffceb06ce038013291e092324d3c49091cd1b2ff1569abe77c29484075e28107
        - quay.io/stolostron/hypershift-addon-operator@sha256:ed1c1580fee28310fc1b137c5c70271448e58ed29af4371515fe9c74512a2c7e
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/placement@sha256:6276e411c9f7cf7072d3d364569b6c3cd472742ab4ad48912f18cb4091d27a6f
        - quay.io/stolostron/ose_cluster_api_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/cluster-proxy-addon@sha256:c776405c4b1b623500c69a12b7d76563646b27cf23bab74dc8f1361534cfb81e
        - quay.io/stolostron/addon-manager@sha256:a245847f135f96c4b4fcaa673fc32013c8e407252b3259175c796e3c16111c71
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/discovery-testserver@sha256:faef24208a65283b4157521079a7bf23bc4ebd553d70a83a78c324b5cb0e24e4
        - quay.io/stolostron/discovery-operator@sha256:a6a37237bc06f0635a936b803d3c53f0f0caf5426c76b37951c4e157b181ecb3
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:1d4221cfc364743b58e15bad3afe9548f4151f06b0dbf0aec7c00715ba75abba
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/ip-address-manager@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/assisted-installer@sha256:17b78d73775a93b31ca872db79520ddc49d2be042fdf1599f419fbcbfd9a98ac
        - quay.io/stolostron/assisted-service-el8@sha256:f52c55691850a1f2f254b0e9377ba97493be337e012223e5a2b3190f03e1a349
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:44f8bb72b975456d4151a88a560fc525ad6f4da7968a38e94a86b6dd77d3c543
        - quay.io/stolostron/discovery-operator-tests@sha256:d96fe1f4f49ecfd4a27bedd87d7070a3c7f3f82b6ee8bb59e98a262d8cedcfc0
        - quay.io/stolostron/clusterclaims-controller@sha256:afab145b12a52f81e1f2eaa189731e3a6287b54f2f1f1ad2976ad378fa7dfa3a
        - quay.io/stolostron/ose_aws_cluster_api_controllers_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/backplane-operator@sha256:9b39a5b6a502778d7884e276b276a9b44e6dbc88e7e82470e26d91ded3cb3839
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        - quay.io/stolostron/assisted-image-service@sha256:b6df9c9fc8bbfdc58e2169b9e199e1eed188d3d847511eba01eb4920fabe3992
        - quay.io/stolostron/multicloud-manager@sha256:5a95a5d1dbb17456d7ce62f1a517ae49ca780e25f3ebb74b32346de1df23b4c9
        - quay.io/stolostron/helloworld@sha256:a790795409f6e781f506b6a63f39f7d344f01efa30c3de87b1727dc2010d5bbf
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/backplane-must-gather@sha256:f4f9596eade78333628199f62f8e71e4c2b3957c20da5bad807c12a2a02560db
        - quay.io/stolostron/managedcluster-import-controller@sha256:4d0be693db0576eb2e9e3d409bf1479cc60784f5747ed1b25b6a2443c5e91e36
        - quay.io/stolostron/cluster-api-provider-agent@sha256:2b22890b948ad1d597c424d73d4fe34c7c4d265ad6e54fa0ed87a7dec920ac57
        - quay.io/stolostron/ose-baremetal-cluster-api-controllers-rhel9@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/cluster-curator-controller@sha256:d667bd0892812eb4e1f724679759469db335db4ab801f9d57ea8818d8cd13188
        - quay.io/stolostron/image-based-install-operator@sha256:9d6796007b616de60560f55f14546295c4e48b3768d059098fce9367674a3368
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/registration@sha256:2db2c28af25007c77979cc456d929f5f077cd671cc4e0562c43efce2c42c4cd8
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/provider-credential-controller@sha256:3484cb0c2102f4a522efe31bc4c77910a202b42e1ad082a0a8709cc1a7698649
        - quay.io/stolostron/work@sha256:709c41d1a7a1eaf10b334dd0c0105a261d6db811693d5ae67eb55b9a26daed87
        - quay.io/stolostron/assisted-installer-controller@sha256:8ea9d12fa30825fe4f965a7d506af687e8de709f8d07dd75204747da2a6e09a8
        - quay.io/stolostron/assisted-service@sha256:af3532d487e03a077ce23e4178abbcb41cf45726b646d7660840c036677715e8
        - quay.io/stolostron/hypershift-operator@sha256:c9a8fc943ef35126434cc767f152abd09943a4d2d194a9966495fdc0870b417c
        - quay.io/stolostron/registration-operator@sha256:a9fdf837e149c8659fc99deab6a284975c8eb8785549381fa082303cdb172e44
        - quay.io/stolostron/managed-serviceaccount@sha256:4bcd5c9b2044254c28215fd78bf1e3fcd919578238986831fcea8055b2764696
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:1757b9ad635839087c4bef198d5e81f4abb7aee38bc74690021f9e8887366bcf
        - quay.io/stolostron/backplane-operator-test@sha256:8d6514fc0db5ee79c346ddc31bb094ad87fa375519da67e32398300d70d72f6e
        - quay.io/stolostron/cluster-api-provider-openshift-assisted-control-plane@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        version: 0.7.1
      entries:
      - name: stolostron-engine.v0.7.1
        version: 0.7.1
      - name: stolostron-engine.v0.7.0
        version: 0.7.0
      name: community-0.7
    defaultChannel: community-0.7
    packageName: stolostron-engine
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pulp Community
      provider-url: https://github.com/pulp/pulp-operator
    name: pulp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pulp-operator.v1.0.0-alpha.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache_enabled": true,
                  "content": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "ingress_type": "Route",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/"
                  },
                  "storage_type": "File",
                  "worker": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example-pulp"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true,
                    "redis_storage_class": "standard"
                  },
                  "content": {
                    "replicas": 1
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "pulp",
                  "file_storage_access_mode": "ReadWriteOnce",
                  "file_storage_size": "2Gi",
                  "file_storage_storage_class": "standard",
                  "image_version": "nightly",
                  "image_web_version": "nightly",
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/",
                    "telemetry": false
                  },
                  "web": {
                    "replicas": 1
                  },
                  "worker": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "galaxy-example"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "api": {
                    "replicas": 1
                  },
                  "content": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "galaxy",
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "file_storage_storage_class": "standard",
                  "image": "quay.io/pulp/galaxy-minimal",
                  "image_version": "nightly",
                  "image_web": "quay.io/pulp/galaxy-web",
                  "image_web_version": "nightly",
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/api/galaxy/pulp/",
                    "telemetry": false
                  },
                  "signing_scripts_configmap": "signing-scripts",
                  "signing_secret": "signing-galaxy",
                  "storage_type": "File",
                  "web": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "worker": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "galaxy-example",
                  "deployment_type": "galaxy",
                  "instance_name": "pulp",
                  "postgres_configuration_secret": "galaxy-example-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "galaxy-example"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.0.0-alpha.1
          createdAt: "2021-03-26 16:57:40"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1alpha1
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1alpha1
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1alpha1
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://docs.pulpproject.org/](https://docs.pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          Note that Pulp operator works with three different types of service containers (the operator itself, the main service and the web service):

          |           | Operator | Main | Web |
          | --------- | -------- | ---- | --- |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[pulp-minimal](https://quay.io/repository/pulp/pulp-minimal?tab=tags) | [pulp-web](https://quay.io/repository/pulp/pulp-web?tab=tags) |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[galaxy-minimal](https://quay.io/repository/pulp/galaxy-minimal?tab=tags) | [galaxy-web](https://quay.io/repository/pulp/galaxy-web?tab=tags) |

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://docs.pulpproject.org/pulp_operator/container/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://docs.pulpproject.org/pulp_operator/](https://docs.pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://docs.pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: beta
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - docker.io/library/postgres:13
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/pulp/pulp-operator:v1.0.0-alpha.1
        - quay.io/pulp/pulp-minimal:stable
        - quay.io/pulp/pulp-web:stable
        - docker.io/library/redis:latest
        version: 1.0.0-alpha.1
      entries:
      - name: pulp-operator.v1.0.0-alpha.1
        version: 1.0.0-alpha.1
      name: alpha
    - currentCSV: pulp-operator.v1.0.0-beta.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "Pulp",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true
                  },
                  "content": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/"
                  },
                  "worker": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "Pulp",
                "metadata": {
                  "name": "galaxy-example"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "api": {
                    "replicas": 1
                  },
                  "content": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "galaxy",
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "file_storage_storage_class": "standard",
                  "image": "quay.io/pulp/galaxy-minimal",
                  "image_version": "4.6.3",
                  "image_web": "quay.io/pulp/galaxy-web",
                  "image_web_version": "4.6.3",
                  "ingress_type": "nodeport",
                  "nodeport_port": 30000,
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "analytics": false,
                    "ansible_api_hostname": "http://nodeport.local:30000",
                    "api_root": "/api/galaxy/pulp/",
                    "content_origin": "http://nodeport.local:30000",
                    "galaxy_feature_flags": {
                      "ai_deny_index": true,
                      "execution_environments": true
                    },
                    "token_server": "http://nodeport.local:30000/token/"
                  },
                  "signing_scripts": "signing-scripts",
                  "signing_secret": "signing-galaxy",
                  "web": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "worker": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "galaxy-example",
                  "deployment_type": "galaxy",
                  "postgres_configuration_secret": "galaxy-example-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "galaxy-example"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.0.0-beta.4
          createdAt: "2024-01-18T20:33:37Z"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1beta2
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1beta2
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1beta2
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://docs.pulpproject.org/](https://docs.pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          Note that Pulp operator works with three different types of service containers (the operator itself, the main service and the web service):

          |           | Operator | Main | Web |
          | --------- | -------- | ---- | --- |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[pulp-minimal](https://quay.io/repository/pulp/pulp-minimal?tab=tags) | [pulp-web](https://quay.io/repository/pulp/pulp-web?tab=tags) |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[galaxy-minimal](https://quay.io/repository/pulp/galaxy-minimal?tab=tags) | [galaxy-web](https://quay.io/repository/pulp/galaxy-web?tab=tags) |

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://docs.pulpproject.org/pulp_operator/container/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://docs.pulpproject.org/pulp_operator/](https://docs.pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://docs.pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: beta
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.4
        - quay.io/pulp/pulp-operator:v1.0.0-beta.4
        - quay.io/pulp/pulp-minimal:stable
        - quay.io/pulp/pulp-web:stable
        - docker.io/library/redis:latest
        - docker.io/library/postgres:13
        version: 1.0.0-beta.4
      entries:
      - name: pulp-operator.v1.0.0-beta.4
        version: 1.0.0-beta.4
      - name: pulp-operator.v1.0.0-beta.3
        version: 1.0.0-beta.3
      - name: pulp-operator.v1.0.0-beta.2
        version: 1.0.0-beta.2
      - name: pulp-operator.v1.0.0-beta.1
        version: 1.0.0-beta.1
      - name: pulp-operator.v1.0.0-alpha.9
        version: 1.0.0-alpha.9
      - name: pulp-operator.v1.0.0-alpha.8
        version: 1.0.0-alpha.8
      - name: pulp-operator.v1.0.0-alpha.7
        version: 1.0.0-alpha.7
      - name: pulp-operator.v1.0.0-alpha.6
        version: 1.0.0-alpha.6
      - name: pulp-operator.v1.0.0-alpha.5
        version: 1.0.0-alpha.5
      - name: pulp-operator.v1.0.0-alpha.4
        version: 1.0.0-alpha.4
      - name: pulp-operator.v1.0.0-alpha.3
        version: 1.0.0-alpha.3
      - name: pulp-operator.v1.0.0-alpha.2
        version: 1.0.0-alpha.2
      - name: pulp-operator.v1.0.0-alpha.1
        version: 1.0.0-alpha.1
      name: beta
    defaultChannel: beta
    packageName: pulp-operator
    provider:
      name: Pulp Community
      url: https://github.com/pulp/pulp-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-common-service-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-common-service-operator.v4.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v3",
                "kind": "CommonService",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-common-service-operator",
                    "app.kubernetes.io/managed-by": "ibm-common-service-operator",
                    "app.kubernetes.io/name": "ibm-common-service-operator"
                  },
                  "name": "example-commonservice"
                },
                "spec": {
                  "size": "starterset"
                }
              }
            ]
          capabilities: Seamless Upgrades
          cloudPakThemesVersion: styles4100.css
          containerImage: registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
          createdAt: "2025-03-18T21:27:31Z"
          description: The IBM Cloud Pak foundational services operator is used to
            deploy IBM foundational services.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-common-service-operator
          nss.operator.ibm.com/managed-webhooks: ""
          olm.skipRange: '>=3.3.0 <4.12.0'
          operatorChannel: v4.12
          operatorVersion: 4.12.0
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/ibm-common-service-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CommonService is the Schema for the commonservices API. Documentation
              For additional details regarding install parameters check https://ibm.biz/icpfs39install.
              License By installing this product you accept the license terms https://ibm.biz/icpfs39license
            displayName: CommonService
            kind: CommonService
            name: commonservices.operator.ibm.com
            version: v3
        description: |-
          # Introduction

          The `IBM Cloud Pak foundational services` is a bridge to connect IBM Cloud Paks and Operand Deployment Lifecycle Manager (ODLM) with IBM Cloud Pak foundational services. You can also install the `IBM Cloud Pak foundational services` in stand-alone mode.

          When you install this operator, the operator completes the following tasks:

            - Installs ODLM operator namespace
            - Creates the Common Services `OperandRegistry` and `OperandConfig` in the service namespace

          # Details
          For more information about installing this operator and other Common Services operators, see [Installer documentation](http://ibm.biz/cpcs_opinstall). If you are using this operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For more information about IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).

          For more information about the available IBM Cloud Pak foundational services, see the [IBM Knowledge Center](http://ibm.biz/cpcsdocs).

          ## Supported platforms

          Red Hat OpenShift Container Platform 4.5 or newer installed on one of the following platforms:

           - Linux x86_64
           - Linux on Power (ppc64le)
           - Linux on IBM Z and LinuxONE

          ## Operator versions

           - 4.12.0

          ## Prerequisites

          Before you install this operator, you need to first install the operator prerequisites:

          - For the list of prerequisites for installing the operator, see the IBM Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).

          ### Resources Required

          #### Minimum scheduling capacity

          | Software                   | Memory (MB) | CPU (cores) | Disk (GB) | Nodes  |
          | -------------------------- | ----------- | ----------- | --------- | ------ |
          | ibm common service operator | 200          | 0.2        | 1          | worker |
          | **Total**                  | 200         | 0.2         | 1         |        |

          ## Documentation

          - If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak. For a list of IBM Cloud Paks, see [IBM Cloud Paks that use IBM Cloud Pak foundational services](http://ibm.biz/cpcs_cloudpaks).
          - If you are using the operator in stand-alone mode or with an IBM Containerized Software, see the IBM Cloud Pak foundational services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall).

          ## Installing

          For installation, see the [IBM Cloud Platform Common Services documentation](http://ibm.biz/cpcsdocs).

          ## Configuration

          For configuration, see the [IBM Cloud Platform Common Services documentation](http://ibm.biz/cpcsdocs).

          ### SecurityContextConstraints Requirements

          The IBM Common Service Operator supports running with the OpenShift Container Platform 4.3 default restricted Security Context Constraints (SCCs) and IBM Cloud Pak Security Context Constraints (SCCs).

          For more information about the OpenShift Container Platform Security Context Constraints, see [Managing Security Context Constraints](https://docs.openshift.com/container-platform/4.3/authentication/managing-security-context-constraints.html).

          For more information about the IBM Cloud Pak Security Context Constraints, see [Managing Security Context Constraints](https://ibm.biz/cpkspec-scc).

          ### Limitations

          No
        displayName: IBM Cloud Pak foundational services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Common Service Operator
          url: https://github.com/IBM/ibm-common-service-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90046eb6170e12439ccc5aa343f19a4d4fc99a10d98b383cf1dc0e020ebeed55
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:18f44ab15289625fa2c6b72ece6e06afb51058dcba30b8f38b4fdeedf8c9e51b
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:0fdf1beeb4888f648f6dbd17a29dcc53a715d2c4301bc714a4172b2daf8d13be
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:4209e78f12120b9c9f320496b780c3993a9317825ea17dc7d554452b28c99014
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:9a537eae7e581322e2f0764ed38a797ba7cd260724a285f24d5095058b7b47a0
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-operator-bundle@sha256:086eaeaaaf440a1d4a254984afacd8d04583aafb9ec11ac072feb42a87b986f7
        version: 4.12.0
      entries:
      - name: ibm-common-service-operator.v4.12.0
        version: 4.12.0
      name: v4.12
    defaultChannel: v4.12
    packageName: ibm-common-service-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: rebellions
      provider-url: https://rebellions.ai/
    name: rbln-npu-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: candidate
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: fast
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: rbln-npu-operator
    provider:
      name: rebellions
      url: https://rebellions.ai/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/component: operator
      app.kubernetes.io/managed-by: kustomize
      app.kubernetes.io/name: zabbix-operator-certified
      app.kubernetes.io/part-of: zabbix-operator-certified
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zabbix LLC
      provider-url: ""
    name: zabbix-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zabbix-operator-certified-rhmp.v6.0.40
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent",
                "metadata": {
                  "labels": {
                    "app": "agent",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "log_remote_commands": true,
                  "max_lines_per_second": 20,
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent2",
                "metadata": {
                  "labels": {
                    "app": "agent2",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent2"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "enable_persistent_buffer": false,
                  "enable_status_port": false,
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "persistent_buffer_claim": "",
                  "persistent_buffer_period": "1h",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixFull",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-full"
                },
                "spec": {
                  "builtin_report_service": true,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysql_volumeclaim": "mysql-full-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxyMysql",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-mysql"
                },
                "spec": {
                  "db_server_port": 3306,
                  "enable_snmp_traps": false,
                  "internal_db": true,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix_proxy",
                  "proxy": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "zabbix_mysql_volumeclaim": "mysql-proxy-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxySqlite",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-sqlite"
                },
                "spec": {
                  "db_data_volumeclaim": "zabbix-proxy-sqlite",
                  "enable_snmp_traps": false,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "proxy": {
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "node_name_as_db_name": true,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "proxy_size": 1,
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  }
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixServer",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-server"
                },
                "spec": {
                  "builtin_report_service": true,
                  "db_server_host": "mysql-server",
                  "db_server_port": 3306,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix",
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              }
            ]
          ansible.operator-sdk/reconcile-period: 5s
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          certified: "true"
          containerImage: registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
          createdAt: "2025-05-26T07:41:22Z"
          description: Zabbix operator with multiple deployment variants and different
            components
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/zabbix-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/zabbix-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <6.0.40
          operatorframework.io/suggested-namespace: zabbix
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: ""
          support: Zabbix
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Zabbix agent 2 is a new generation of Zabbix agent and may
              be used in place of Zabbix agent
            displayName: Zabbix agent 2
            kind: ZabbixAgent2
            name: zabbixagent2s.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix agent is deployed on a monitoring nodes to actively
              monitor local resources and applications
            displayName: Zabbix agent
            kind: ZabbixAgent
            name: zabbixagents.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database server, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Full
            kind: ZabbixFull
            name: zabbixfulls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with MySQL database server
            displayName: Zabbix proxy (MySQL)
            kind: ZabbixProxyMysql
            name: zabbixproxymysqls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with SQLite database
            displayName: Zabbix proxy (SQLite3)
            kind: ZabbixProxySqlite
            name: zabbixproxysqlites.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database support, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Server
            kind: ZabbixServer
            name: zabbixservers.kubernetes.zabbix.com
            version: v1alpha1
        description: "## About this Operator\n\nZabbix helps you to real-time monitoring
          of millions of metrics collected from tens of thousands of servers, virtual
          machines and network devices.\nThe Zabbix Operator allows users to easily
          deploy, manage, and maintain Zabbix deployments on OpenShift. By installing
          this integration you will be able to deploy Zabbix server / proxies and
          other components with a single command.\n## Supported Features\n* **Zabbix
          Server** - Simple Zabbix installation with included Zabbix server, Zabbix
          web-interface and Zabbix Java Gateway with MySQL database support. The feature
          does not provide MySQL service and requires an external MySQL database.
          \n* **Zabbix Server (Full)** - Zabbix installation with included Zabbix
          server, Zabbix web-interface, Zabbix Java Gateway and MySQL server instance.\n*
          **Zabbix proxy (SQLite3)** - Very simple way to gain power of Zabbix proxy.
          The feature has  SQLite3 support for Zabbix proxies and allow to specify
          amount of proxies. \n* **Zabbix proxy (MySQL)** - Another option of Zabbix
          proxy. The option support and deliver MySQL database.\n* **Zabbix agent**
          - Zabbix agent can be deployed on each available node for stability and
          performance monitoring on remote nodes. It allows to gather metrics with
          full automation!* \n## Prerequisites\nAll deployment options are require
          additional information during deployment. Please, check the following instructions
          and provide required configuration:\n* **Zabbix Server** - MySQL database
          host information and MySQL database credentials in specially formatted *Secret*.
          Additionally it is possible to specify SSL certificates for HTTPS support
          in *Secret*.\n* **Zabbix Server (Full)** - MySQL database credentials in
          specially formatted *Secret*. MySQL database volume name information. Additionally
          it is possible to specify SSL certificates for HTTPS support in *Secret*.\n*
          **Zabbix proxy (SQLite3)** - Zabbix server host information only.\n* **Zabbix
          proxy (MySQL)** - MySQL database credentials in specially formatted *Secrets*
          and Zabbix server host information.\n* **Zabbix agent** - Zabbix server
          host information only for outgoing and incoming connections. Zabbix agent
          uses \"privileged\" mode to monitor node resources! For example, running
          processes.* \n### MySQL credentials\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-full-secrets\ndata:\n  mysql_root_password: emFiYml4X3N1cGVyX3Jvb3Q=
          [1]\n  mysql_zabbix_password: emFiYml4X3VzZXJfcGFzc3dvcmQ= [2]\n  mysql_zabbix_username:
          emFiYml4 [3]\ntype: Opaque\n```\nUsing MySQL root password (*mysqlrootpassword*
          [1]) Zabbix server / proxy will try to create MySQL database schema with
          grant permissions to *mysqlzabbixusername* [2] and *mysqlzabbixpassword*
          [3].\n### SSL certificates for HTTPS\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-web-sslsecret\ndata:\n  ssl.crt: >-\n   < ssl.crt data>\n
          \ ssl.key: >-\n\t < ssl.key data >\n  dhparam.pem: >-\n   <  dhparam.pem
          data >\n```\nFiles *ssl.crt*, *ssl.key* and *dhparam.perm* are required
          for Zabbix web-interface for SSL support.\n"
        displayName: Zabbix Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - zabbix
        - appliance
        - agent
        - metrics
        links:
        - name: Zabbix
          url: https://www.zabbix.com
        - name: Zabbix Official Documentation
          url: https://www.zabbix.com/documentation/6.0/en/manual/quickstart
        - name: Zabbix Operator Installation
          url: https://www.zabbix.com/documentation/6.0/en/manual/installation/containers/openshift
        - name: Downloads
          url: https://www.zabbix.com/download
        maintainers:
        - email: alexey.pustovalov@zabbix.com
          name: Alexey Pustovalov
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Zabbix LLC
        relatedImages:
        - registry.connect.redhat.com/zabbix/zabbix-agent2-60@sha256:424003d52f3cb4360adf2b3619d8b9f8de9196c630ee7e4ceb5d177c1c5a36e4
        - registry.connect.redhat.com/zabbix/zabbix-server-mysql-60@sha256:637484faec6e9a55049c13dbd1f100c03d757c3a21749388c0605040f02d05ce
        - registry.connect.redhat.com/zabbix/zabbix-proxy-mysql-60@sha256:c2ae74d97c083ff348dc9c55932a33da4779cd9efdd9d7c5a79acfffae5b9ed0
        - registry.connect.redhat.com/zabbix/zabbix-java-gateway-60@sha256:a4feb725c12c4ac90eb9244c6fafdb6708679f2e0e7f6aa683c765127306412e
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
        - registry.connect.redhat.com/zabbix/zabbix-web-mysql-60@sha256:611d2670e1bf95b81226976c0a1f64434b2a7766f196dadbd646a7d8319f4cb1
        - registry.connect.redhat.com/zabbix/zabbix-web-service-60@sha256:4a333094e110713470a897b6406c5e340e962caf32a429dddf19865041fcc448
        - registry.redhat.io/rhel9/mysql-80@sha256:c4b5246d0c0d089a766ea07768c22d9d6be8049ac1379bd345db99390ffe2a36
        - registry.connect.redhat.com/zabbix/zabbix-snmptraps-60@sha256:e61e6b1c551dec8488cce0998f94d83d6d10b5d89a465d7119b088313032cb81
        - registry.connect.redhat.com/zabbix/zabbix-proxy-sqlite-60@sha256:62e62ebc8c1bc13fb610f7a83412e65d852f74b3c2c212baa75ea27fc7fa03eb
        - registry.connect.redhat.com/zabbix/zabbix-agent-60@sha256:2895b5e45785aaf528d4a6f78c8af0a0d494dbe9362403fdd470673d911d5004
        - quay.io/zabbix/zabbix-operator@sha256:43d26488a19967bdb2a8aafde5f43820be95011811da9cc5ec623c9029801675
        version: 6.0.40
      entries:
      - name: zabbix-operator-certified-rhmp.v6.0.40
        version: 6.0.40
      - name: zabbix-operator-certified-rhmp.v6.0.38
        version: 6.0.38
      - name: zabbix-operator-certified-rhmp.v6.0.36
        version: 6.0.36
      - name: zabbix-operator-certified-rhmp.v6.0.33
        version: 6.0.33
      - name: zabbix-operator-certified-rhmp.v6.0.29
        version: 6.0.29
      - name: zabbix-operator-certified-rhmp.v6.0.27
        version: 6.0.27
      - name: zabbix-operator-certified-rhmp.v6.0.26
        version: 6.0.26
      - name: zabbix-operator-certified-rhmp.v6.0.25
        version: 6.0.25
      - name: zabbix-operator-certified-rhmp.v6.0.17
        version: 6.0.17
      - name: zabbix-operator-certified-rhmp.v6.0.14
        version: 6.0.14
      - name: zabbix-operator-certified-rhmp.v6.0.13
        version: 6.0.13
      name: lts
    defaultChannel: lts
    packageName: zabbix-operator-certified-rhmp
    provider:
      name: Zabbix LLC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/streamshub
    name: amq-streams-console
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: alpha
    - currentCSV: amq-streams-console.v2.8.0-13-0.1738265629.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:1b1b80710885b43600e22ff2a24930b5cab06cd91fad439ad8a282c00b4564b9
          createdAt: "2024-11-12T15:34:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-streams-console.v2.8.0-13
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.8
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.23.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:1b1b80710885b43600e22ff2a24930b5cab06cd91fad439ad8a282c00b4564b9
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:38b80d81e54d59667c1f7dfa4da4125d5c7430816da130ddc84678a4e3cff6cf
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:366d60ff48fb9ba32dc73d92a775046d7b1eb1df21c93e31605372c7858f16ca
        version: 2.8.0-13+0.1738265629.p
      entries:
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.8.x
    - currentCSV: amq-streams-console.v2.9.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
          createdAt: "2025-09-22T10:41:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0-0 <2.9.2-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.9
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:839368767bec735cd12f002874917c23f58550fe21df976858bdc92946985ec6
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:b7444b70a5dc4543a9976c08a619ff6978c9ff86ac7085d300b4870b7c6b7851
        version: 2.9.2-2
      entries:
      - name: amq-streams-console.v2.9.2-2
        version: 2.9.2-2
      - name: amq-streams-console.v2.9.2-1
        version: 2.9.2-1
      - name: amq-streams-console.v2.9.1-7
        version: 2.9.1-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.9.x
    - currentCSV: amq-streams-console.v2.9.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
          createdAt: "2025-09-22T10:41:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0-0 <2.9.2-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.9
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:839368767bec735cd12f002874917c23f58550fe21df976858bdc92946985ec6
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:b7444b70a5dc4543a9976c08a619ff6978c9ff86ac7085d300b4870b7c6b7851
        version: 2.9.2-2
      entries:
      - name: amq-streams-console.v2.9.2-2
        version: 2.9.2-2
      - name: amq-streams-console.v2.9.2-1
        version: 2.9.2-1
      - name: amq-streams-console.v2.9.1-7
        version: 2.9.1-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.x
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-3.0.x
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-3.x
    defaultChannel: alpha
    packageName: amq-streams-console
    provider:
      name: Red Hat
      url: https://github.com/streamshub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: 3scale-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: 3scale-operator.v0.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "ProductCRName": "product1-sample",
                  "Production": true
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/3scale-rhel7-operator@sha256:0751f9f5343a1264224c36db3db292e5a9e28c40cd96649ef48b5399c34668dd
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.16"}]'
          olm.skipRange: '>=0.10.0 <0.11.13'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/3scale-rhel7-operator@sha256:0751f9f5343a1264224c36db3db292e5a9e28c40cd96649ef48b5399c34668dd
        - registry.redhat.io/3scale-amp2/backend-rhel8@sha256:cdc53799ff9d022462a9a4afcbcd8043497cd40ccdc6fafdcb9ccf7ce69f43d1
        - registry.redhat.io/3scale-amp2/system-rhel7@sha256:61dcc91e5d3c91a0ea85c02f2599cadcb5496b1c09a8838b4e1c8e4efe126e8f
        - registry.redhat.io/3scale-amp2/memcached-rhel7@sha256:7e2ab1e24646fbfef878eb2b5a74ed865ae812b60aa0ba505045f838f26ebbfc
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:a95d09fc3b224f550dec3de3d23fd2dbfc0a220dc869b4ad9559ee2f85327daa
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:f8cd25658fa6a154e7ea6e5adf0f9dae3fc596ec9989021de6a021ec94eef226
        - registry.redhat.io/rhel8/redis-6@sha256:5250d644e80aea794b0f732e9285883a4348dace83557f177f64d78fe4dcabd0
        - registry.redhat.io/openshift4/ose-cli@sha256:1a8caae40be21f23ba3cb48b75809a73333907033a6041927aa568b5c9290f3c
        - registry.redhat.io/3scale-amp2/searchd-rhel7@sha256:8e2001c021fad554cb8a96bb34e9973010fe4206526f19aa040fa196709fb688
        - registry.redhat.io/3scale-amp2/zync-rhel9@sha256:0d2205d9f61a58414c4ba472ce8f19b12d2df722772847d5fb7442e664759627
        - registry.redhat.io/rhel8/mysql-80@sha256:08547a93659ad804d637b91e8e1eb3464b75783d494e6c40e1c921bd74f31645
        version: 0.11.13
      entries:
      - name: 3scale-operator.v0.11.13
        version: 0.11.13
      - name: 3scale-operator.v0.11.12
        version: 0.11.12
      name: threescale-2.14
    - currentCSV: 3scale-operator.v0.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                },
                "status": {
                  "deployments": {
                    "ready": [
                      "deployment x ready"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                },
                "status": {
                  "adminId": 1,
                  "tenantId": 2
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ApplicationAuth",
                "metadata": {
                  "name": "applicationauth-sample"
                },
                "spec": {
                  "applicationCRName": "application-cr-name",
                  "authSecretRef": {
                    "name": "auth-secret-reference"
                  },
                  "generateSecret": true
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "productCRName": "product1-sample",
                  "production": true
                },
                "status": {}
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/3scale-rhel9-operator@sha256:fcea6a77d0a48f35f4c3ce63e72dde104aac6b580f340950b311eb418fff5ec9
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=0.11.0 <0.12.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          rht_backend_redis_requirements: 6.2.0
          rht_mysql_requirements: 8.0.0
          rht_postgres_requirements: 10.0.0
          rht_system_redis_requirements: 6.2.0
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: ApplicationAuth is the Schema for the applicationauths API
            displayName: Application Auth
            kind: ApplicationAuth
            name: applicationauths.capabilities.3scale.net
            version: v1beta1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/3scale-rhel9-operator@sha256:fcea6a77d0a48f35f4c3ce63e72dde104aac6b580f340950b311eb418fff5ec9
        - registry.redhat.io/3scale-amp2/manticore-rhel9@sha256:ccf6baf5f2e9494a34739a79abe0f46e36d8ab9dc81c426f7a3a8964e8f39859
        - registry.redhat.io/3scale-amp2/backend-rhel8@sha256:eb2288de900dc42ead755b28cdb2509b25ecc8dd256781beaa4a2a86e7cbe532
        - registry.redhat.io/rhel9/memcached@sha256:8738110ef16c9e35deeb07c0577ffe09d6ef029761b504a8d20253fa715da394
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:a95d09fc3b224f550dec3de3d23fd2dbfc0a220dc869b4ad9559ee2f85327daa
        - registry.redhat.io/3scale-amp2/zync-rhel9@sha256:df9d62897ff286165772150a8b5e6952d2ce3b52a8afa07bd693a2d95c2c67ff
        - registry.redhat.io/openshift4/ose-cli@sha256:1a8caae40be21f23ba3cb48b75809a73333907033a6041927aa568b5c9290f3c
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/rhel8/mysql-80@sha256:ee9aca9096473da004f9140305742eb7b9a77ed81ce55d5a6503388cccab030f
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        - registry.redhat.io/3scale-amp2/system-rhel8@sha256:4c72dfeb1919a1de4fbc47f4611d0c830ef89cf9f17255ad55a9b370618e0e8f
        version: 0.12.5
      entries:
      - name: 3scale-operator.v0.12.5
        version: 0.12.5
      - name: 3scale-operator.v0.12.4
        version: 0.12.4
      - name: 3scale-operator.v0.12.3
        version: 0.12.3
      - name: 3scale-operator.v0.12.2
        version: 0.12.2
      - name: 3scale-operator.v0.12.1
        version: 0.12.1
      - name: 3scale-operator.v0.12.0
        version: 0.12.0
      name: threescale-2.15
    - currentCSV: 3scale-operator.v0.12.1-mas
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                },
                "status": {
                  "deployments": {
                    "ready": [
                      "deployment x ready"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                },
                "status": {
                  "adminId": 1,
                  "tenantId": 2
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ApplicationAuth",
                "metadata": {
                  "name": "applicationauth-sample"
                },
                "spec": {
                  "applicationCRName": "application-cr-name",
                  "authSecretRef": {
                    "name": "auth-secret-reference"
                  },
                  "generateSecret": true
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "productCRName": "product1-sample",
                  "production": true
                },
                "status": {}
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-mas/3scale-rhel9-operator@sha256:8f41ab6027616e4bb65b1fddbcf44cc8a25a79cdda19d5cc354ac4228a15141e
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.17"}]'
          olm.skipRange: '>=0.11.8 <0.12.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          rht_backend_redis_requirements: 6.2.0
          rht_mysql_requirements: 8.0.0
          rht_postgres_requirements: 10.0.0
          rht_system_redis_requirements: 6.2.0
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: ApplicationAuth is the Schema for the applicationauths API
            displayName: Application Auth
            kind: ApplicationAuth
            name: applicationauths.capabilities.3scale.net
            version: v1beta1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-mas/3scale-manticore-search-rhel9@sha256:7379186175c9c0f0a6d922cd2a0754b429c9728dfbe8979a9d6025789ebae0da
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:a95d09fc3b224f550dec3de3d23fd2dbfc0a220dc869b4ad9559ee2f85327daa
        - registry.redhat.io/3scale-mas/backend-rhel8@sha256:d0c561c6b9dfe9741722ba4a257fbd2df4d2a709457a0b0657ef7bcca1f2a5ca
        - registry.redhat.io/3scale-mas/zync-rhel9@sha256:78e5223a1a878f6527bd88d382948c125b7cb4a81540648244553af426d1a944
        - registry.redhat.io/3scale-mas/memcached-rhel7@sha256:7a1abee11b4120f399485b438bda1e5262f04de28d3e788b207eecc3a9de9ed2
        - registry.redhat.io/3scale-mas/3scale-rhel9-operator@sha256:8f41ab6027616e4bb65b1fddbcf44cc8a25a79cdda19d5cc354ac4228a15141e
        - registry.redhat.io/rhel8/mysql-80@sha256:754f6830e0bc4aadb973b1f613f2bc7278ec13ad0f0e1e87655a2bb2f76cb9b3
        - registry.redhat.io/openshift4/ose-cli@sha256:a688994ea404ce6ce4bb5ec1249f0e779de0d6c187b5078e2ff067e60b50937d
        - registry.redhat.io/3scale-mas/apicast-gateway-rhel8@sha256:f174636a9ad43a2419fff5a9e6797e3185c540a9509188fd1606aef4acea7bdf
        - registry.redhat.io/rhel8/redis-6@sha256:218f7fa2458c274cc34f79190423e0aba03027f55ef6b3334d916951a668b7b7
        - registry.redhat.io/3scale-mas/system-rhel8@sha256:0fa28aaf3467fbd22fc1335dc6a712043e4dead694c3f14b60e89740bb7eb13e
        version: 0.12.1-mas
      entries:
      - name: 3scale-operator.v0.12.1-mas
        version: 0.12.1-mas
      - name: 3scale-operator.v0.11.8-mas
        version: 0.11.8-mas
      - name: 3scale-operator.v0.11.7-mas
        version: 0.11.7-mas
      - name: 3scale-operator.v0.11.6-mas
        version: 0.11.6-mas
      - name: 3scale-operator.v0.11.5-mas
        version: 0.11.5-mas
      - name: 3scale-operator.v0.11.3-mas
        version: 0.11.3-mas
      - name: 3scale-operator.v0.11.2-mas
        version: 0.11.2-mas
      - name: 3scale-operator.v0.11.1-mas
        version: 0.11.1-mas
      - name: 3scale-operator.v0.11.0-mas
        version: 0.11.0-mas
      - name: 3scale-operator.v0.10.0-mas
        version: 0.10.0-mas
      name: threescale-mas
    defaultChannel: threescale-2.15
    packageName: 3scale-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: operator-certification-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operator-certification-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "certification.redhat.com/v1alpha1",
                "kind": "OperatorPipeline",
                "metadata": {
                  "name": "operatorpipeline-sample"
                },
                "spec": {
                  "applyCIPipeline": true,
                  "applyHostedPipeline": false,
                  "applyReleasePipeline": false,
                  "gitHubSecretName": "github-api-token",
                  "kubeconfigSecretName": "kubeconfig",
                  "operatorPipelinesRelease": "main",
                  "pyxisSecretName": "pyxis-api-secret"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-06-23T16:02:00Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io.bundle.channel.default.v1: alpha
          operators.operatorframework.io.bundle.channels: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-openshift-ecosystem/operator-certification-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OperatorPipeline is the Schema for the operatorpipelines
              API
            displayName: Operator Pipeline
            kind: OperatorPipeline
            name: operatorpipelines.certification.redhat.com
            version: v1alpha1
        description: |-
          A Kubernetes operator to provision resources for the operator certification pipeline. This operator is installed in all namespaces which can support multi-tenant scenarios. **Note:** This operator should only be used by Red Hat partners attempting to certify their operator(s).


          **Requirements**


          The certification operator requires that you have the following tools installed, functional, and in your path.
          - [Install](https://docs.openshift.com/container-platform/4.8/cli_reference/openshift_cli/getting-started-cli.html#installing-openshift-cli) oc, the OpenShift CLI tool (tested with version 4.7.13)
          - [Install](https://tekton.dev/docs/cli/) tkn, the Tekton CLI tool (tested with version 0.19.1)
          - [Install](https://git-scm.com/downloads) git, the Git CLI tool (tested with 2.32.0)
          - The certification pipeline expects you to have the source files of your Operator bundle


          **Pre - Installation**


          The below steps exist for simplicity and to make the installation clearer. The operator watches all namespaces, so if secrets/configs/etc already exist in another namespace, feel free to use the existing namespace when following the operator installation steps.


          **Create a new namespace where the following secrets will be applied.**
          ```
          oc new-project oco
          ```


          **Add the kubeconfig secret which will be used to deploy the operator under test and run the certification checks.**
          * Open a terminal window
          * Set the `KUBECONFIG` environment variable
          ```
          export KUBECONFIG=/path/to/your/cluster/kubeconfig
          ```
          > *This kubeconfig will be used to deploy the Operator under test and run the certification checks.*
          ```
          oc create secret generic kubeconfig --from-file=kubeconfig=$KUBECONFIG
          ```


          **Configuring steps for submitting the results**
          - Add the github API token to the repo where the PR will be created
          ```
          oc create secret generic github-api-token --from-literal GITHUB_TOKEN=<github token>
          ```
          - Add RedHat Container API access key

            This API access key is specifically related to your unique partner account for Red Hat Connect portal. Instructions to obtain your API key can be found: [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/operator-cert-workflow.md#step-b---get-api-key)
          ```
          oc create secret generic pyxis-api-secret --from-literal pyxis_api_key=< API KEY >
          ```

          - Optional pipeline configurations can be found [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md#optional-configuration)


          **Execute the Pipeline (Development Iterations)**


          A pre-requisite to running a pipeline is that a `workspace-template.yaml` exists in the directory you want to execute the `tkn` commands from.

          To create a workspace-template.yaml
          ```
          cat <<EOF > workspace-template.yaml
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 5Gi
          EOF
          ```

          There are multiple ways to execute the Pipeline which can be found [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md#execute-the-pipeline-development-iterations)
        displayName: Operator Certification Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - certification
        - pipeline
        - openshift
        links:
        - name: Operator Certification CI Pipeline
          url: https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md
        - name: Red Hat OpenShift Pipelines
          url: https://github.com/openshift/tektoncd-operator
        maintainers:
        - email: certification-operator-devel@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/opdev/operator-certification-operator@sha256:7c84772f96650ff757c2552ba421355d54da41a5ed2155f72a14911b8a3f18ff
        version: 1.3.1
      entries:
      - name: operator-certification-operator.v1.3.1
        version: 1.3.1
      name: alpha
    defaultChannel: alpha
    packageName: operator-certification-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: LINBIT
      provider-url: https://linbit.com
    name: linstor-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: linstor-operator.v2.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorCluster",
                "metadata": {
                  "name": "linstorcluster"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorNodeConnection",
                "metadata": {
                  "name": "example-connection"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatellite",
                "metadata": {
                  "name": "example-satellite"
                },
                "spec": {
                  "clusterRef": {
                    "name": "linstorcluster"
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatelliteConfiguration",
                "metadata": {
                  "name": "all-satellites"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          com.redhat.openshift.versions: v4.7
          containerImage: drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
          createdAt: "2025-09-25T13:44:22Z"
          description: LINSTOR Operator deploys LINBIT SDS in the cluster.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: linbit-sds
          operators.openshift.io/valid-subscription: '[''LINBIT SDS Subscription'']'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/internal-objects: '["linstorsatellites.piraeus.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster-wide configuration for LINSTOR.
            displayName: Linstor Cluster
            kind: LinstorCluster
            name: linstorclusters.piraeus.io
            version: v1
          - description: Configuration for connections between specific nodes.
            displayName: Linstor Node Connection
            kind: LinstorNodeConnection
            name: linstornodeconnections.piraeus.io
            version: v1
          - description: Configuration for specific Satellites/Nodes in the LINSTOR
              Cluster.
            displayName: Linstor Satellite Configuration
            kind: LinstorSatelliteConfiguration
            name: linstorsatelliteconfigurations.piraeus.io
            version: v1
          - description: Resource to configure a single satellite, managed by the
              Operator itself.
            displayName: Linstor Satellite
            kind: LinstorSatellite
            name: linstorsatellites.piraeus.io
            version: v1
          required:
          - kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
        description: |
          The LINSTOR Operator manages [LINSTOR](https://github.com/LINBIT/linstor-server) clusters in Kubernetes.

          All components of the LINSTOR software stack can be managed by the operator:
          * DRBD
          * LINSTOR
          * LINSTOR CSI driver
          * LINSTOR HA Controller
        displayName: LINSTOR Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: User Guide
          url: https://linbit.com/drbd-user-guide/linstor-guide-1_0-en/
        - name: LINSTOR Operator
          url: https://charts.linstor.io/
        - name: Source Code
          url: https://github.com/LINBIT/linstor-operator-builder
        maintainers:
        - email: support@linbit.com
          name: LINBIT
        maturity: stable
        minKubeVersion: 1.20.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: DaemonSet
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: storage.k8s.io
          kind: CSIDriver
          version: v1
        provider:
          name: LINBIT
          url: https://linbit.com
        relatedImages:
        - drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:f8cde5b5c289f7b3d233773f4e79990ba392fe6e14e0177de80714e7d3e98862
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ac3963bfd4246da386b82d810a348e9d42a9e4b746acaf4dbcbe0766643cc2fa
        - drbd.io/drbd-shutdown-guard@sha256:7ba049ecb205a0086e425d2ff37285b00449549d89e179b4b943ace6e59a9156
        - drbd.io/drbd9-rhel8@sha256:7cddfa770de66cefcb42c2b33da57d0f1adfcc5f1cc12522977aa6993722bb2a
        - drbd.io/linstor-k8s-ha-controller@sha256:ed0f0f86dad88feadf8c3c32d7d143eaa52b1130af3576315a562909b39d2afa
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:2669cde3db2268d37bc671f494eaa06480478a90b8943054af14cbffc050debf
        - drbd.io/drbd-reactor@sha256:77c51aeeefdcd788cb1583738f93778032fcc03c36fb289b692d3cf195c8b184
        - drbd.io/linstor-csi@sha256:9c7ef452b0997d3373803201614c3b8075687e75c397359738a9e4ef5530cab3
        - drbd.io/linstor-controller@sha256:841b3aa84cdf13aaaf1243bc5da8ac2205ca3fe0186fdfc7a04b69397f0020d6
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:624f61b1ae9d32f5a074ee8c12839d975fba9efcd3bfa7981cfe38febf5c4495
        - drbd.io/drbd9-rhel9@sha256:4eb553336ed64e127068a4566370b040166bf11301e21ac47ac5b92237ec9d2a
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:5a650dc21a6a038de6267f117d01adfd904ff3f5c94f428f6a13735f18bf94d6
        - drbd.io/drbd9-rhel7@sha256:0c96b216e55a128a9944276af9080aced8bd72fdf1fd424f024ed9ace2c1feb5
        - drbd.io/ktls-utils@sha256:3351a6e0fc048fe1544499095cdb8fdd221e832cecfa8a9a01f046c271991138
        - drbd.io/linstor-satellite@sha256:ce0a0b05494c6e6d5f16121e5762bf58e7b78f40c00b84780e2abba2a690af1b
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:700be620923fc99ed03f9f1c4a911e4084beb30685361270c49165b27ad1d374
        - drbd.io/drbd9-rhel10@sha256:da0a7015cf15724a7c73100d505891e62262a726141347df7f2d51ec79a0a1e1
        - drbd.io/linstor-affinity-controller@sha256:78e5b299ef1702114528678387b5e74821d7658bc02f05f0bd7edd9a176fba39
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:cd58f1789027e32eae08f00dac7b2275496c41020fb7cdebc10e9986c625d4f4
        version: 2.9.1
      entries:
      - name: linstor-operator.v2.9.1
        version: 2.9.1
      name: dev
    - currentCSV: linstor-operator.v2.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorCluster",
                "metadata": {
                  "name": "linstorcluster"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorNodeConnection",
                "metadata": {
                  "name": "example-connection"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatellite",
                "metadata": {
                  "name": "example-satellite"
                },
                "spec": {
                  "clusterRef": {
                    "name": "linstorcluster"
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatelliteConfiguration",
                "metadata": {
                  "name": "all-satellites"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          com.redhat.openshift.versions: v4.7
          containerImage: drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
          createdAt: "2025-09-25T13:44:22Z"
          description: LINSTOR Operator deploys LINBIT SDS in the cluster.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: linbit-sds
          operators.openshift.io/valid-subscription: '[''LINBIT SDS Subscription'']'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/internal-objects: '["linstorsatellites.piraeus.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster-wide configuration for LINSTOR.
            displayName: Linstor Cluster
            kind: LinstorCluster
            name: linstorclusters.piraeus.io
            version: v1
          - description: Configuration for connections between specific nodes.
            displayName: Linstor Node Connection
            kind: LinstorNodeConnection
            name: linstornodeconnections.piraeus.io
            version: v1
          - description: Configuration for specific Satellites/Nodes in the LINSTOR
              Cluster.
            displayName: Linstor Satellite Configuration
            kind: LinstorSatelliteConfiguration
            name: linstorsatelliteconfigurations.piraeus.io
            version: v1
          - description: Resource to configure a single satellite, managed by the
              Operator itself.
            displayName: Linstor Satellite
            kind: LinstorSatellite
            name: linstorsatellites.piraeus.io
            version: v1
          required:
          - kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
        description: |
          The LINSTOR Operator manages [LINSTOR](https://github.com/LINBIT/linstor-server) clusters in Kubernetes.

          All components of the LINSTOR software stack can be managed by the operator:
          * DRBD
          * LINSTOR
          * LINSTOR CSI driver
          * LINSTOR HA Controller
        displayName: LINSTOR Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: User Guide
          url: https://linbit.com/drbd-user-guide/linstor-guide-1_0-en/
        - name: LINSTOR Operator
          url: https://charts.linstor.io/
        - name: Source Code
          url: https://github.com/LINBIT/linstor-operator-builder
        maintainers:
        - email: support@linbit.com
          name: LINBIT
        maturity: stable
        minKubeVersion: 1.20.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: DaemonSet
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: storage.k8s.io
          kind: CSIDriver
          version: v1
        provider:
          name: LINBIT
          url: https://linbit.com
        relatedImages:
        - drbd.io/drbd9-rhel7@sha256:0c96b216e55a128a9944276af9080aced8bd72fdf1fd424f024ed9ace2c1feb5
        - drbd.io/linstor-controller@sha256:841b3aa84cdf13aaaf1243bc5da8ac2205ca3fe0186fdfc7a04b69397f0020d6
        - drbd.io/drbd9-rhel9@sha256:4eb553336ed64e127068a4566370b040166bf11301e21ac47ac5b92237ec9d2a
        - drbd.io/linstor-k8s-ha-controller@sha256:ed0f0f86dad88feadf8c3c32d7d143eaa52b1130af3576315a562909b39d2afa
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:f8cde5b5c289f7b3d233773f4e79990ba392fe6e14e0177de80714e7d3e98862
        - drbd.io/drbd-shutdown-guard@sha256:7ba049ecb205a0086e425d2ff37285b00449549d89e179b4b943ace6e59a9156
        - drbd.io/linstor-affinity-controller@sha256:78e5b299ef1702114528678387b5e74821d7658bc02f05f0bd7edd9a176fba39
        - drbd.io/linstor-satellite@sha256:ce0a0b05494c6e6d5f16121e5762bf58e7b78f40c00b84780e2abba2a690af1b
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:5a650dc21a6a038de6267f117d01adfd904ff3f5c94f428f6a13735f18bf94d6
        - drbd.io/linstor-csi@sha256:9c7ef452b0997d3373803201614c3b8075687e75c397359738a9e4ef5530cab3
        - drbd.io/ktls-utils@sha256:3351a6e0fc048fe1544499095cdb8fdd221e832cecfa8a9a01f046c271991138
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ac3963bfd4246da386b82d810a348e9d42a9e4b746acaf4dbcbe0766643cc2fa
        - drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:cd58f1789027e32eae08f00dac7b2275496c41020fb7cdebc10e9986c625d4f4
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:624f61b1ae9d32f5a074ee8c12839d975fba9efcd3bfa7981cfe38febf5c4495
        - drbd.io/drbd9-rhel8@sha256:7cddfa770de66cefcb42c2b33da57d0f1adfcc5f1cc12522977aa6993722bb2a
        - drbd.io/drbd-reactor@sha256:77c51aeeefdcd788cb1583738f93778032fcc03c36fb289b692d3cf195c8b184
        - drbd.io/drbd9-rhel10@sha256:da0a7015cf15724a7c73100d505891e62262a726141347df7f2d51ec79a0a1e1
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:700be620923fc99ed03f9f1c4a911e4084beb30685361270c49165b27ad1d374
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:2669cde3db2268d37bc671f494eaa06480478a90b8943054af14cbffc050debf
        version: 2.9.1
      entries:
      - name: linstor-operator.v2.9.1
        version: 2.9.1
      name: stable
    defaultChannel: dev
    packageName: linstor-operator
    provider:
      name: LINBIT
      url: https://linbit.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: https://opdev.github.io/
    name: loki-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.example.com/v1alpha1",
                "kind": "Loki",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/install-disable-crds": "true"
                  },
                  "name": "loki-sample"
                },
                "spec": {
                  "global": {
                    "clusterDomain": "cluster.local",
                    "dnsNamespace": "openshift-dns",
                    "dnsService": "dns-default"
                  },
                  "loki": {
                    "auth_enabled": false,
                    "commonConfig": {
                      "replication_factor": 1
                    },
                    "storage": {
                      "type": "filesystem"
                    }
                  },
                  "monitoring": {
                    "dashboards": {
                      "enabled": false
                    },
                    "lokiCanary": {
                      "enabled": false
                    },
                    "rules": {
                      "alerting": false,
                      "enabled": false
                    },
                    "selfMonitoring": {
                      "enabled": false,
                      "grafanaAgent": {
                        "installOperator": false
                      }
                    },
                    "serviceMonitor": {
                      "enabled": false,
                      "metricsInstance": {
                        "enabled": false
                      }
                    }
                  },
                  "rbac": {
                    "sccEnabled": true
                  },
                  "singleBinary": {
                    "replicas": 1
                  },
                  "test": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/loki-helm-operator-controller:0.0.1
          createdAt: "2023-07-14T15:06:17Z"
          description: An experimental operator that installs loki from its official
            helmchart.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys the loki operator helm chart
            displayName: Loki
            kind: Loki
            name: lokis.charts.example.com
            version: v1alpha1
        description: Loki operator based on the helm chart with single binary mode
          deployment optionality
        displayName: Loki Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loki
        - single-binary
        - demo
        - helm
        links:
        - name: Loki Helm Operator
          url: https://github.com/skattoju/loki-helm-operator
        maintainers:
        - email: skattoju@redhat.com
          name: Sid Kattoju
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: https://opdev.github.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/opdev/loki-helm-operator-controller:0.0.1
        version: 0.0.1
      entries:
      - name: loki-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: loki-helm-operator
    provider:
      name: opdev
      url: https://opdev.github.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat CCIT
      provider-url: ""
    name: reportportal-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: reportportal-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ReportPortal",
                "metadata": {
                  "name": "reportportal-sample"
                },
                "spec": {
                  "accesskey": "AKIAIOSFODNN7EXAMPLE",
                  "analyzer_image": "quay.io/waynesun09/service-auto-analyzer:5.3.5",
                  "api_image": "quay.io/waynesun09/service-api:5.3.5-rootless",
                  "api_replicas": 1,
                  "app_domain": "apps.test-example.com",
                  "bucket_prefix": null,
                  "default_bucket_name": null,
                  "enable_pg_restic_backup": "yes",
                  "es_backup_schedule": "@daily",
                  "es_s3_backup_dir": "s3_backup",
                  "es_snapshot_bucket": "es-snapshot-123123",
                  "gateway_replicas": 1,
                  "index_image": "quay.io/waynesun09/service-index:5.0.10",
                  "index_replicas": 1,
                  "integration_salt_path": null,
                  "migration_image": "quay.io/waynesun09/migrations:5.3.5",
                  "pg_restic_password": "rp_user",
                  "pg_restic_s3_bucket_init": "yes",
                  "pg_s3_bucket_name": "pgbackup-123123",
                  "s3_enabled": false,
                  "s3_endpoint": null,
                  "secretkey": "wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY",
                  "uat_image": "quay.io/waynesun09/service-authorization:5.3.5",
                  "ui_image": "quay.io/waynesun09/service-ui:5.3.5",
                  "ui_replicas": 1
                }
              },
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ReportPortalRestore",
                "metadata": {
                  "name": "reportportalrestore-sample"
                },
                "spec": {
                  "accesskey": "AKIAIOSFODNN7EXAMPLE",
                  "bucket_prefix": null,
                  "es_restore": "yes",
                  "es_restore_date": "2020-08-07",
                  "es_s3_backup_dir": "s3_backup",
                  "es_snapshot_bucket": "es-snapshot-123123",
                  "pg_recovery": "yes",
                  "pg_recovery_target_time": "2020-08-07 00:00:00",
                  "pg_restic_password": "rp_user",
                  "pg_s3_bucket_name": "pgbackup-123123",
                  "s3_endpoint": null,
                  "secretkey": "wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY"
                }
              },
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ServiceApiServiceMonitor",
                "metadata": {
                  "name": "serviceapiservicemonitor-sample"
                },
                "spec": {
                  "bearer_token_secret_key": "token",
                  "bearer_token_secret_name": "rp-user-auth"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/waynesun09/rp5-operator:v0.0.8
          createdAt: "2021-04-22 14:39:00"
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/waynesun09/reportportal-operator
          support: waynesun09
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ReportPortal Restore from s3 backup
            kind: ReportPortalRestore
            name: reportportalrestores.rp5.reportportal.io
            version: v1
          - description: ReportPortal v5 Deployment with ReportPortal Operator
            kind: ReportPortal
            name: reportportals.rp5.reportportal.io
            version: v1
          - description: ReportPortal service-api ServiceMonitor for metrics
            kind: ServiceApiServiceMonitor
            name: serviceapiservicemonitors.rp5.reportportal.io
            version: v1
        description: |-
          ReportPortal is a test automation analytics platform. ReportPortal Operator support instantiating the ReportPortal services and required dependencies.

          What services are included:

              - Report Portal v5.3
              - PostgreSQL
              - Elasticsearch
              - RabbitMQ
              - Minio

          The Operator also support data backup & restore and metrics monitoring.

          ## Install

          Install the Operator to a namespace, then create ReportPortal v5 instance under Provided APIs on the UI, all Report Portal v5 services and dependencies will be installed.

          ## Data backup and restore

          The Report Portal v5 data backup and restore in the operator include two parts, PostgreSQL database data and Elasticsearch index data.

          The Operator support automatically data backup and restore for both.

          ### PostgreSQL database WAL archiving and recovery

          The Operator supports PostgreSQL WAL archiving and recovery point-in-time via Minio S3 using restic.

          #### WAL archiving

          When deploy Report Portal v5 instance, make sure `enable_pg_restic_backup` option is set as `yes`, then the WAL archiving will be enabled when initiate the database, the WAL archive data will be automatically backup to S3 storage via Minio using restic command.

          Parameter `pg_restic_s3_bucket_init` is for init the bucket if not exist, `pg_s3_bucket_name` is the bucket name and `pg_restic_password` is the restic password for encryption.

          #### WAL recovery

          WAL archive recovery will download all WAL archive files to the Postgres container and restore point-in-time.

          To restore PostgreSQL database point-in-time, on Operator UI select `ReportPortalRestore` under Provided APIs to create ReportPortalRestore operand.

          After the ReportPortalRestore operand created, the PG WAL restore will get started, as restore time will base on the database size, the process will last in minutes, hours or days.

          **Note:** Current Postgres container liveless probe timeout is set as one hour in the postgresql statefulset, it might cause restore fail if restore time is longer than one hour, so it might need be adjusted to be longer.

          ### Elasticsearch snapshot and restore

          The operator supports elasticsearch snapshot and restore with data saved in S3 bucket via Minio, the elasticsearch container have installed the s3 repository plugin.

          #### Snapshot

          When deploy Report Portal v5 instance, make sure following elasticsearch parameters are set:

              es_s3_backup_dir: s3_backup
              es_snapshot_bucket: es-snapshot-123123
              es_backup_schedule: '@daily'

          #### Restore

          Elasticsearch restore support restore to certain date as default snapshot schedule is daily.

          To restore elasticsearch from snapshot, on Operator UI select `ReportPortalRestore` under Provided APIs to create ReportPortalRestore operand.

          The backup dir and bucket name should be same as when create the RP instance, `es_restore` need set to `yes` to enable the restore and `es_restore_date` is the restore target date.

          The restore time will also be depend on the snapshot size.

          ## Metrics

          The operator by default have enabled metrics, the default ansible operator port 8080 is used for the operator container.

          All Report Portal v5 depended services, rabbitmq, minio, postgresql and elasticsearch have metrics enabled in this operator.

          The Report Portal gatway service have metrics enabled.

          By default several ServiceMonitors have been defined when create ReportPortal v5 operand:

              $ oc get servicemonitor
              NAME                    AGE
              elasticsearch-metrics   1d21h
              gateway-metrics         1d21h
              minio-metrics           1d21h
              postgres-metrics        1d21h
              rabbitmq-metrics        1d21h

          To export Report Portal v5 service-api metrics with ServiceMonitor, create a `secret` with RP user api token. The api access token could be found with login to Report Portal -> Click user icon -> Profile.

          Then select `ServiceApiServiceMonitor` under provided APIs to Create Instance on operator UI.

          Input the secret name and key name that have been created, and click create, a api-metrics servicemonitor will be created.

          To check the metrics, if [user workload monitoring](https://docs.openshift.com/container-platform/4.5/monitoring/monitoring-your-own-services.html) is enabled on the cluster, click to Monitoring on OpenShift UI and navigate to Metrics, where you could query the metrics.

          You could also deploy Promethues Operator in the current project and query via it.

          ## Doc

          More detail with installation, data backup & restore and metrics could be found at https://github.com/waynesun09/reportportal-operator#reportportal-operator
        displayName: reportportal-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - reportportal
        - failure analysis
        - test automation
        links:
        - name: Documentation
          url: https://github.com/waynesun09/reportportal-operator/blob/master/README.md
        - name: ReportPortal.io
          url: https://reportportal.io
        maintainers:
        - email: gsun@redhat.com
          name: Wayne Sun
        maturity: alpha
        provider:
          name: Red Hat CCIT
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.7.0
        - quay.io/waynesun09/rp5-operator:v0.0.8
        version: 0.0.8
      entries:
      - name: reportportal-operator.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: reportportal-operator
    provider:
      name: Red Hat CCIT
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://www.redhat.com/
    name: analytics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: analytics-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyData",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalydata-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalydata",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalydata-sample"
                },
                "spec": {
                  "anomalyname": "etcd_object_namespaces",
                  "config": {
                    "max": 100,
                    "min": 60,
                    "query": "max(apiserver_storage_objects{resource=~\"namespaces\"}) by (resource)"
                  },
                  "method": "min_max",
                  "metricdata": {
                    "groupeddata": "{ \"resource\": \"namespaces\" }",
                    "latestvalue": 121,
                    "timestamp": 1695804362
                  }
                }
              },
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalyengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalyengine",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalyengine-sample"
                },
                "spec": {
                  "anomalyqueryconfiguration": "etcd_objects_all:\n  method: percentage_change\n  query: max(apiserver_storage_objects{resource!~\"certificatesigningrequests.certificates.k8s.io|installplans.operators.coreos.com|operators.operators.coreos.com|subscriptions.operators.coreos.com|anomalydata.observability-analytics.redhat.com|anomalyengines.observability-analytics.redhat.com\"}) by (resource)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: True\nfailed_pods:\n  method: min_max\n  query: count(kube_pod_status_phase{phase!~\"Succeeded|Pending|Running|Unknown\"}) by (phase, namespace)\n  min: 0\n  max: 10\n  have_multi_result_data: True\nworkload_memory_usage_bytes:\n  method: percentage_change\n  query: sum(container_memory_working_set_bytes{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nworkload_cpu_usage_cores:\n  method: percentage_change\n  query: sum(container_cpu_usage_seconds_total{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nrest_client_requests_total:\n  method: percentage_change\n  query: sum(rest_client_requests_total)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\n",
                  "cronjobconfig": {
                    "anomalyqueries": "etcd_objects_all,failed_pods,workload_memory_usage_bytes,workload_cpu_usage_cores,rest_client_requests_total",
                    "name": "osa-anomaly-detection",
                    "resource": {
                      "cpulimit": "256m",
                      "cpurequest": "128m",
                      "memorylimit": "512Mi",
                      "memoryrequest": "256Mi"
                    },
                    "schedule": "*/5 * * * *"
                  },
                  "namespace": "osa-anomaly-detection",
                  "serviceaccountrolebinding": {
                    "clusterrolename": "cluster-monitoring-view",
                    "name": "osa-in-cluster-thanos-querier"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
          createdAt: "2024-03-12T08:02:25Z"
          description: Deploys and Manages Observability Analytics Operator
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4-alpha
          repository: https://github.com/openshift/analytics-operator
          support: RedHat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnomalyData is the Schema for the anomalydata API
            displayName: Anomaly Data
            kind: AnomalyData
            name: anomalydata.observability-analytics.redhat.com
            version: v1alpha1
          - description: AnomalyEngine is the Schema for the anomalyengines API
            displayName: Anomaly Engine
            kind: AnomalyEngine
            name: anomalyengines.observability-analytics.redhat.com
            version: v1alpha1
        description: "### Introduction\nObservability Analytics for Red Hat OpenShift
          runs OpenShift analytics processes like anomaly detection.\n\n### API Backward
          Compatibility\n**NOTE:** Until the Operator supports **seamless upgrades**,
          a new version of the operator may introduce a change that is **NOT** backward-compatible
          with the previous version. Thus, to upgrade the operator, uninstall the
          already installed version completely (including CRDs) and install the new
          version.\n\n### Installtion Steps for the Anomaly Engine\n Create namespace
          if you would like to install this operator into separate namespace (You
          can utilize existing namespace as well). ```bash kubectl create ns <Namespace>
          ``` Install the Observability Analytics operator to specific `<Namespace>`,
          wait until you see the operator in the **Installed Operator** list. Click
          on the Operator name and go to Operator details.\n\nHere you should see
          a tab called **Anomaly Engine**. Click on it, and then click on the **Create
          AnomalyEngine** button, which will redirect you to a form where you need
          to provide the required inputs for the Anomaly Engine. Once successfully
          created, you should see a cronjob in the given namespace. This cronjob will
          identify anomalies and store them into **anomalydata**.\n\n### How to See
          Anomaly Data Captured by Anomaly Engine?\nWe are storing anomaly data into
          CR.\n ####Option 1:Through `kubectl`` command\nYou can check detected anomalies
          using the following command from your terminal. \n\n```bash\nkubectl get
          anomalydata -n <Namespace>\nkubectl describe anomalydata <NAME of the anomalydata>
          -n <Namespace> \n```\n #### Option 2: From Installed Operator UI Go to installed
          operator and select **Observability Analytics** operator, now go to **Anomaly
          Data** tab where you can able to see the detected anomaly, click on specific
          anomaly and you should be able to see details of the same. ### Documentation\nDocumentation
          around how to test Operator codebase as a developer can be found [here](https://github.com/openshift/analytics-operator#readme)\n\n###
          Contributing\nYou can contribute by:\n  * Raising [issues](https://github.com/openshift/analytics-operator/issues)\n
          \   related to observability-analytics-operator\n  * Fixing issues by opening
          [Pull Requests](https://github.com/openshift/analytics-operator/pulls)\n
          \ * Discussing observability-analytics-operator\n\n### License\nobservability-analytics-operator
          is licensed under the Apache 2.0 license.\n"
        displayName: Observability Analytics
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - observability-analytics-operator
        - analytics-operator
        - analytics
        - anomaly
        links:
        - name: Analytics Operator
          url: https://analytics-operator.domain
        maintainers:
        - email: rzalavad@redhat.com
          name: Raj
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
        version: 0.1.1
      entries:
      - name: analytics-operator.v0.1.1
        version: 0.1.1
      - name: analytics-operator.v0.1.0
        version: 0.1.0
      name: alpha
    - currentCSV: analytics-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyData",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalydata-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalydata",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalydata-sample"
                },
                "spec": {
                  "anomalyname": "etcd_object_namespaces",
                  "config": {
                    "max": 100,
                    "min": 60,
                    "query": "max(apiserver_storage_objects{resource=~\"namespaces\"}) by (resource)"
                  },
                  "method": "min_max",
                  "metricdata": {
                    "groupeddata": "{ \"resource\": \"namespaces\" }",
                    "latestvalue": 121,
                    "timestamp": 1695804362
                  }
                }
              },
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalyengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalyengine",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalyengine-sample"
                },
                "spec": {
                  "anomalyqueryconfiguration": "etcd_objects_all:\n  method: percentage_change\n  query: max(apiserver_storage_objects{resource!~\"certificatesigningrequests.certificates.k8s.io|installplans.operators.coreos.com|operators.operators.coreos.com|subscriptions.operators.coreos.com|anomalydata.observability-analytics.redhat.com|anomalyengines.observability-analytics.redhat.com\"}) by (resource)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: True\nfailed_pods:\n  method: min_max\n  query: count(kube_pod_status_phase{phase!~\"Succeeded|Pending|Running|Unknown\"}) by (phase, namespace)\n  min: 0\n  max: 10\n  have_multi_result_data: True\nworkload_memory_usage_bytes:\n  method: percentage_change\n  query: sum(container_memory_working_set_bytes{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nworkload_cpu_usage_cores:\n  method: percentage_change\n  query: sum(container_cpu_usage_seconds_total{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nrest_client_requests_total:\n  method: percentage_change\n  query: sum(rest_client_requests_total)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\n",
                  "cronjobconfig": {
                    "anomalyqueries": "etcd_objects_all,failed_pods,workload_memory_usage_bytes,workload_cpu_usage_cores,rest_client_requests_total",
                    "name": "osa-anomaly-detection",
                    "resource": {
                      "cpulimit": "256m",
                      "cpurequest": "128m",
                      "memorylimit": "512Mi",
                      "memoryrequest": "256Mi"
                    },
                    "schedule": "*/5 * * * *"
                  },
                  "namespace": "osa-anomaly-detection",
                  "serviceaccountrolebinding": {
                    "clusterrolename": "cluster-monitoring-view",
                    "name": "osa-in-cluster-thanos-querier"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
          createdAt: "2024-03-12T08:02:25Z"
          description: Deploys and Manages Observability Analytics Operator
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4-alpha
          repository: https://github.com/openshift/analytics-operator
          support: RedHat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnomalyData is the Schema for the anomalydata API
            displayName: Anomaly Data
            kind: AnomalyData
            name: anomalydata.observability-analytics.redhat.com
            version: v1alpha1
          - description: AnomalyEngine is the Schema for the anomalyengines API
            displayName: Anomaly Engine
            kind: AnomalyEngine
            name: anomalyengines.observability-analytics.redhat.com
            version: v1alpha1
        description: "### Introduction\nObservability Analytics for Red Hat OpenShift
          runs OpenShift analytics processes like anomaly detection.\n\n### API Backward
          Compatibility\n**NOTE:** Until the Operator supports **seamless upgrades**,
          a new version of the operator may introduce a change that is **NOT** backward-compatible
          with the previous version. Thus, to upgrade the operator, uninstall the
          already installed version completely (including CRDs) and install the new
          version.\n\n### Installtion Steps for the Anomaly Engine\n Create namespace
          if you would like to install this operator into separate namespace (You
          can utilize existing namespace as well). ```bash kubectl create ns <Namespace>
          ``` Install the Observability Analytics operator to specific `<Namespace>`,
          wait until you see the operator in the **Installed Operator** list. Click
          on the Operator name and go to Operator details.\n\nHere you should see
          a tab called **Anomaly Engine**. Click on it, and then click on the **Create
          AnomalyEngine** button, which will redirect you to a form where you need
          to provide the required inputs for the Anomaly Engine. Once successfully
          created, you should see a cronjob in the given namespace. This cronjob will
          identify anomalies and store them into **anomalydata**.\n\n### How to See
          Anomaly Data Captured by Anomaly Engine?\nWe are storing anomaly data into
          CR.\n ####Option 1:Through `kubectl`` command\nYou can check detected anomalies
          using the following command from your terminal. \n\n```bash\nkubectl get
          anomalydata -n <Namespace>\nkubectl describe anomalydata <NAME of the anomalydata>
          -n <Namespace> \n```\n #### Option 2: From Installed Operator UI Go to installed
          operator and select **Observability Analytics** operator, now go to **Anomaly
          Data** tab where you can able to see the detected anomaly, click on specific
          anomaly and you should be able to see details of the same. ### Documentation\nDocumentation
          around how to test Operator codebase as a developer can be found [here](https://github.com/openshift/analytics-operator#readme)\n\n###
          Contributing\nYou can contribute by:\n  * Raising [issues](https://github.com/openshift/analytics-operator/issues)\n
          \   related to observability-analytics-operator\n  * Fixing issues by opening
          [Pull Requests](https://github.com/openshift/analytics-operator/pulls)\n
          \ * Discussing observability-analytics-operator\n\n### License\nobservability-analytics-operator
          is licensed under the Apache 2.0 license.\n"
        displayName: Observability Analytics
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - observability-analytics-operator
        - analytics-operator
        - analytics
        - anomaly
        links:
        - name: Analytics Operator
          url: https://analytics-operator.domain
        maintainers:
        - email: rzalavad@redhat.com
          name: Raj
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
        version: 0.1.1
      entries:
      - name: analytics-operator.v0.1.1
        version: 0.1.1
      - name: analytics-operator.v0.1.0
        version: 0.1.0
      name: dev-preview
    defaultChannel: alpha
    packageName: analytics-operator
    provider:
      name: Red Hat Inc.
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-atlas-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-atlas-kubernetes.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupCompliancePolicy",
                "metadata": {
                  "name": "my-backup-compliance-policy"
                },
                "spec": {
                  "authorizedEmail": "john.doe@example.com",
                  "authorizedUserFirstName": "John",
                  "authorizedUserLastName": "Doe",
                  "copyProtectionEnabled": false,
                  "encryptionAtRestEnabled": false,
                  "onDemandPolicy": {
                    "retentionUnit": "weeks",
                    "retentionValue": 3
                  },
                  "overwriteBackupPolicies": false,
                  "pointInTimeEnabled": true,
                  "restoreWindowDays": 42,
                  "scheduledPolicyItems": [
                    {
                      "frequencyInterval": 2,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "hourly",
                      "retentionUnit": "days",
                      "retentionValue": 2
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "weekly",
                      "retentionUnit": "weeks",
                      "retentionValue": 4
                    },
                    {
                      "frequencyInterval": 15,
                      "frequencyType": "monthly",
                      "retentionUnit": "months",
                      "retentionValue": 12
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": false,
                  "copySettings": [
                    {
                      "cloudProvider": "AWS",
                      "frequencies": [
                        "WEEKLY"
                      ],
                      "regionName": "US_WEST_1",
                      "shouldCopyOplogs": false
                    }
                  ],
                  "policy": {
                    "name": "atlasbackuppolicy-sample"
                  },
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2,
                  "useOrgAndGroupNamesInExportPrefix": true
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasCustomRole",
                "metadata": {
                  "name": "shard-operator-role"
                },
                "spec": {
                  "projectRef": {
                    "name": "my-project"
                  },
                  "role": {
                    "actions": [
                      {
                        "name": "getShardMap",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "shardingState",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "connPoolStats",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "getLog",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      }
                    ],
                    "inheritedRoles": [
                      {
                        "database": "admin",
                        "name": "operator-role-1"
                      }
                    ],
                    "name": "my-role"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDataFederation",
                "metadata": {
                  "name": "my-data-federation",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "name": "my-data-federation",
                  "privateEndpoints": [
                    {
                      "endpointId": "vpce-03f9eeaa764e32454",
                      "provider": "AWS",
                      "type": "DATA_LAKE"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project",
                    "namespace": "mongodb-atlas-system"
                  },
                  "storage": {
                    "databases": [
                      {
                        "collections": [
                          {
                            "dataSources": [
                              {
                                "storeName": "http-test",
                                "urls": [
                                  "https://data.cityofnewyork.us/api/views/vfnx-vebw/rows.csv"
                                ]
                              }
                            ],
                            "name": "test-collection-1"
                          }
                        ],
                        "name": "test-db-1"
                      }
                    ],
                    "stores": [
                      {
                        "name": "http-test",
                        "provider": "http"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "the-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "theuser"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-deployment"
                },
                "spec": {
                  "backupRef": {
                    "name": "atlasbackupschedule-sample"
                  },
                  "deploymentSpec": {
                    "backupEnabled": true,
                    "clusterType": "REPLICASET",
                    "mongoDBMajorVersion": "7.0",
                    "name": "test-deployment",
                    "replicationSpecs": [
                      {
                        "regionConfigs": [
                          {
                            "electableSpecs": {
                              "instanceSize": "M10",
                              "nodeCount": 3
                            },
                            "priority": 7,
                            "providerName": "AWS",
                            "regionName": "US_EAST_1"
                          }
                        ],
                        "zoneName": "Zone 1"
                      }
                    ],
                    "tags": [
                      {
                        "key": "team",
                        "value": "atlas-operator"
                      }
                    ]
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasFederatedAuth",
                "metadata": {
                  "name": "atlas-default-federated-auth",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "connectionSecretRef": {
                    "name": "my-org-secret",
                    "namespace": "mongodb-atlas-system"
                  },
                  "domainAllowList": [
                    "my-org-domain.com"
                  ],
                  "domainRestrictionEnabled": true,
                  "enabled": true,
                  "postAuthRoleGrants": [
                    "ORG_MEMBER"
                  ],
                  "roleMappings": [
                    {
                      "externalGroupName": "org-admin",
                      "roleAssignments": [
                        {
                          "role": "ORG_OWNER"
                        }
                      ]
                    },
                    {
                      "externalGroupName": "dev-team",
                      "roleAssignments": [
                        {
                          "role": "ORG_GROUP_CREATOR"
                        },
                        {
                          "projectName": "dev-project",
                          "role": "GROUP_OWNER"
                        }
                      ]
                    }
                  ],
                  "ssoDebugEnabled": false
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasIPAccessList",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasipaccesslist-sample"
                },
                "spec": {
                  "entries": [
                    {
                      "cidrBlock": "192.168.1.0/24"
                    },
                    {
                      "deleteAfterDate": "2025-01-20T23:59:59+02:00",
                      "ipAddress": "10.0.0.1"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkContainer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasnetworkcontainer-sample"
                },
                "spec": {
                  "cidrBlock": "10.11.0.0/16",
                  "connectionSecret": {
                    "name": "atlas-connection-secret"
                  },
                  "externalProjectRef": {
                    "projectId": "66e2f2b621571b7e69a89b66"
                  },
                  "provider": "AWS",
                  "region": "US_EAST_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkPeering",
                "metadata": {
                  "name": "atlasnetworkpeering-sample"
                },
                "spec": {
                  "atlasCidrBlock": "10.12.0.0/16",
                  "awsConfiguration": {
                    "accepterRegionName": "US_EAST_1",
                    "awsAccountId": "23214235145",
                    "routeTableCidrBlock": "10.11.0.0/16",
                    "vpcId": "vpcid-7123782134"
                  },
                  "containerId": "623412394512350",
                  "containerRegion": "US_EAST_1",
                  "projectRef": {
                    "name": "atlas-project",
                    "namespace": "namespace"
                  },
                  "provider": "AWS"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasPrivateEndpoint",
                "metadata": {
                  "name": "atlasprivateendpoint-sample"
                },
                "spec": {
                  "awsConfiguration": {
                    "id": "vpce-f4k34w51d"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "provider": "AWS",
                  "region": "EU_CENTRAL_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    },
                    {
                      "cidrBlock": "203.0.113.0/24",
                      "comment": "CIDR block for Application Server B - D"
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasSearchIndexConfig",
                "metadata": {
                  "name": "atlassearchindexconfig-sample"
                },
                "spec": {
                  "analyzer": "lucene.standard",
                  "analyzers": [
                    {
                      "name": "test-analyzer",
                      "tokenFilters": [
                        {
                          "asciiFolding": {
                            "originalTokens": "omit"
                          },
                          "type": "asciiFolding"
                        }
                      ],
                      "tokenizer": {
                        "type": "whiteSpace",
                        "whitespace": {
                          "maxTokenLenght": 200
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamConnection",
                "metadata": {
                  "name": "my-streamconnection-kafka"
                },
                "spec": {
                  "kafkaConfig": {
                    "authentication": {
                      "credentials": {
                        "name": "kafka-userpass"
                      },
                      "mechanism": "SCRAM-512"
                    },
                    "bootstrapServers": "kafka.server1:9001,kafka.server2:9002,kafka.server3:9003",
                    "security": {
                      "certificate": {
                        "name": "kafka-cert"
                      },
                      "protocol": "SSL"
                    }
                  },
                  "name": "kafka-config",
                  "type": "Kafka"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamInstance",
                "metadata": {
                  "name": "my-streaminstance-sample"
                },
                "spec": {
                  "clusterConfig": {
                    "provider": "AWS",
                    "region": "VIRGINIA_USA",
                    "tier": "SP30"
                  },
                  "connectionRegistry": [
                    {
                      "name": "my-streamconnection-sample"
                    }
                  ],
                  "name": "my-instance",
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasTeam",
                "metadata": {
                  "name": "atlasteam-sample"
                },
                "spec": {
                  "name": "my-team",
                  "usernames": [
                    "username@mongodb.com"
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasThirdPartyIntegration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasintegration-sample"
                },
                "spec": {
                  "datadog": {
                    "apiKeySecretRef": {
                      "name": "secret-ref"
                    },
                    "region": "string",
                    "sendCollectionLatencyMetrics": "string",
                    "sendDatabaseMetrics": "string"
                  },
                  "projectRef": {
                    "name": "my-atlas-project"
                  },
                  "type": "DATADOG"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/mongodb/mongodb-atlas-kubernetes-operator:2.11.0
          createdAt: "2025-09-25T14:34:34Z"
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/mongodb/mongodb-atlas-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The AtlasBackupCompliancePolicy is a configuration that enforces
              specific backup and retention requirements
            displayName: Atlas Backup Compliance Policy
            kind: AtlasBackupCompliancePolicy
            name: atlasbackupcompliancepolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasCustomRole is the Schema for the AtlasCustomRole API
            displayName: Atlas Custom Role
            kind: AtlasCustomRole
            name: atlascustomroles.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDataFederation is the Schema for the Atlas Data Federation
              API
            displayName: Atlas Data Federation
            kind: AtlasDataFederation
            name: atlasdatafederations.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasdeployments API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasFederatedAuth is the Schema for the Atlasfederatedauth
              API
            displayName: Atlas Federated Auth
            kind: AtlasFederatedAuth
            name: atlasfederatedauths.atlas.mongodb.com
            version: v1
          - description: AtlasIPAccessList is the Schema for the atlasipaccesslists
              API.
            displayName: Atlas IPAccess List
            kind: AtlasIPAccessList
            name: atlasipaccesslists.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkContainer is the Schema for the AtlasNetworkContainer
              API
            displayName: Atlas Network Container
            kind: AtlasNetworkContainer
            name: atlasnetworkcontainers.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkPeering is the Schema for the AtlasNetworkPeering
              API
            displayName: Atlas Network Peering
            kind: AtlasNetworkPeering
            name: atlasnetworkpeerings.atlas.mongodb.com
            version: v1
          - displayName: Atlas Org Settings
            kind: AtlasOrgSettings
            name: atlasorgsettings.atlas.mongodb.com
            version: v1
          - description: "The AtlasPrivateEndpoint custom resource definition (CRD)
              defines a desired [Private Endpoint](https://www.mongodb.com/docs/atlas/security-private-endpoint/#std-label-private-endpoint-overview)
              configuration for an Atlas project. It allows a private connection between
              your cloud provider and Atlas that doesn't send information through
              a public network. \n You can use private endpoints to create a unidirectional
              connection to Atlas clusters from your virtual network."
            displayName: Atlas Private Endpoint
            kind: AtlasPrivateEndpoint
            name: atlasprivateendpoints.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
          - description: Atlas Search Index Config is the Schema for the Atlas Search
              API
            displayName: Atlas Search Index Config
            kind: AtlasSearchIndexConfig
            name: atlassearchindexconfigs.atlas.mongodb.com
            version: v1
          - description: AtlasStreamConnection is the Schema for the atlasstreamconnections
              API
            displayName: Atlas Stream Connection
            kind: AtlasStreamConnection
            name: atlasstreamconnections.atlas.mongodb.com
            version: v1
          - description: AtlasStreamInstance is the Schema for the atlasstreaminstances
              API
            displayName: Atlas Stream Instance
            kind: AtlasStreamInstance
            name: atlasstreaminstances.atlas.mongodb.com
            version: v1
          - description: AtlasTeam is the Schema for the Atlas Teams API
            displayName: Atlas Team
            kind: AtlasTeam
            name: atlasteams.atlas.mongodb.com
            version: v1
          - description: AtlasThirdPartyIntegration is the Schema for the atlas 3rd
              party integrations API.
            displayName: Atlas Third Party Integration
            kind: AtlasThirdPartyIntegration
            name: atlasthirdpartyintegrations.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB
          Atlas — the only multi-cloud document database service that gives you the versatility you need to build sophisticated
          and resilient applications that can adapt to changing customer demands and market trends.

          The full documentation for the Operator can be found [here](https://docs.atlas.mongodb.com/atlas-operator/)

          ## Getting Started

          ### Supported features

          * Create and configure a project, or connect to an existing one.
          * Deploy, manage, scale, and tear down clusters.
          * Support for serverless instances.
          * Create and edit database users.
          * Manage IP Access Lists, network peering and private endpoints.
          * Configure and control Atlas’s fully managed cloud backup.
          * Configure federated authentication for your Atlas organization
          * Integrate Atlas monitoring with Prometheus.

          ... and more.

          To view the list of custom resources and each CR's schema, visit our [reference](https://www.mongodb.com/docs/atlas/operator/stable/custom-resources/)
          documentation. See the [Quickstart](https://www.mongodb.com/docs/atlas/operator/stable/ak8so-quick-start/) to get
          started with Atlas Kubernetes Operator.
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator:2.11.0
        version: 2.11.0
      entries:
      - name: mongodb-atlas-kubernetes.v2.11.0
        version: 2.11.0
      - name: mongodb-atlas-kubernetes.v2.10.0
        version: 2.10.0
      - name: mongodb-atlas-kubernetes.v2.9.1
        version: 2.9.1
      - name: mongodb-atlas-kubernetes.v2.9.0
        version: 2.9.0
      - name: mongodb-atlas-kubernetes.v2.8.2
        version: 2.8.2
      - name: mongodb-atlas-kubernetes.v2.8.1
        version: 2.8.1
      - name: mongodb-atlas-kubernetes.v2.8.0
        version: 2.8.0
      - name: mongodb-atlas-kubernetes.v2.7.1
        version: 2.7.1
      - name: mongodb-atlas-kubernetes.v2.7.0
        version: 2.7.0
      - name: mongodb-atlas-kubernetes.v2.6.1
        version: 2.6.1
      - name: mongodb-atlas-kubernetes.v2.6.0
        version: 2.6.0
      - name: mongodb-atlas-kubernetes.v2.5.0
        version: 2.5.0
      - name: mongodb-atlas-kubernetes.v2.4.1
        version: 2.4.1
      - name: mongodb-atlas-kubernetes.v2.4.0
        version: 2.4.0
      - name: mongodb-atlas-kubernetes.v2.3.1
        version: 2.3.1
      - name: mongodb-atlas-kubernetes.v2.3.0
        version: 2.3.0
      - name: mongodb-atlas-kubernetes.v2.2.2
        version: 2.2.2
      - name: mongodb-atlas-kubernetes.v2.2.1
        version: 2.2.1
      - name: mongodb-atlas-kubernetes.v2.2.0
        version: 2.2.0
      - name: mongodb-atlas-kubernetes.v2.1.0
        version: 2.1.0
      - name: mongodb-atlas-kubernetes.v2.0.1
        version: 2.0.1
      - name: mongodb-atlas-kubernetes.v2.0.0
        version: 2.0.0
      - name: mongodb-atlas-kubernetes.v1.9.1
        version: 1.9.1
      - name: mongodb-atlas-kubernetes.v1.9.0
        version: 1.9.0
      - name: mongodb-atlas-kubernetes.v1.8.2
        version: 1.8.2
      - name: mongodb-atlas-kubernetes.v1.8.1
        version: 1.8.1
      - name: mongodb-atlas-kubernetes.v1.8.0
        version: 1.8.0
      - name: mongodb-atlas-kubernetes.v1.7.3
        version: 1.7.3
      - name: mongodb-atlas-kubernetes.v1.7.2
        version: 1.7.2
      - name: mongodb-atlas-kubernetes.v1.7.1
        version: 1.7.1
      - name: mongodb-atlas-kubernetes.v1.7.0
        version: 1.7.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-atlas-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OpenTelemetry Community
      provider-url: ""
    name: opentelemetry-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opentelemetry-operator.v0.136.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "Instrumentation",
                "metadata": {
                  "name": "instrumentation"
                },
                "spec": {
                  "exporter": {
                    "endpoint": "http://otel-collector-headless:4317"
                  }
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpAMPBridge",
                "metadata": {
                  "name": "opampbridge-sample"
                },
                "spec": {
                  "capabilities": {
                    "AcceptsOpAMPConnectionSettings": true,
                    "AcceptsOtherConnectionSettings": true,
                    "AcceptsRemoteConfig": true,
                    "AcceptsRestartCommand": true,
                    "ReportsEffectiveConfig": true,
                    "ReportsHealth": true,
                    "ReportsOwnLogs": true,
                    "ReportsOwnMetrics": true,
                    "ReportsOwnTraces": true,
                    "ReportsRemoteConfig": true,
                    "ReportsStatus": true
                  },
                  "componentsAllowed": {
                    "exporters": [
                      "debug"
                    ],
                    "processors": [
                      "memory_limiter"
                    ],
                    "receivers": [
                      "otlp"
                    ]
                  },
                  "endpoint": "ws://opamp-server:4320/v1/opamp"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": "receivers:\n  otlp:\n    protocols: \n      grpc: {}\n      http: {}\n\nexporters:\n  debug: {}\n\nservice:\n  pipelines:\n    traces:\n      receivers: [otlp]\n      exporters: [debug]\n"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1beta1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": {
                    "exporters": {
                      "debug": {}
                    },
                    "receivers": {
                      "otlp": {
                        "protocols": {
                          "grpc": {},
                          "http": {}
                        }
                      }
                    },
                    "service": {
                      "pipelines": {
                        "traces": {
                          "exporters": [
                            "debug"
                          ],
                          "receivers": [
                            "otlp"
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          certified: "false"
          containerImage: ghcr.io/open-telemetry/opentelemetry-operator/opentelemetry-operator
          createdAt: "2025-09-29T13:00:19Z"
          description: Provides the OpenTelemetry components, including the Collector
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/open-telemetry/opentelemetry-operator
          support: OpenTelemetry Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instrumentation is the spec for OpenTelemetry instrumentation.
            displayName: OpenTelemetry Instrumentation
            kind: Instrumentation
            name: instrumentations.opentelemetry.io
            version: v1alpha1
          - description: OpAMPBridge is the Schema for the opampbridges API.
            displayName: OpAMP Bridge
            kind: OpAMPBridge
            name: opampbridges.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1beta1
          - description: TargetAllocator is the Schema for the targetallocators API.
            displayName: Target Allocator
            kind: TargetAllocator
            name: targetallocators.opentelemetry.io
            version: v1alpha1
        description: |-
          OpenTelemetry is a collection of tools, APIs, and SDKs. You use it to instrument, generate, collect, and export telemetry data (metrics, logs, and traces) for analysis in order to understand your software's performance and behavior.

          ### Operator features

          * **Sidecar injection** - annotate your pods and let the operator inject a sidecar.
          * **Managed upgrades** - updating the operator will automatically update your OpenTelemetry collectors.
          * **Deployment modes** - your collector can be deployed as sidecar, daemon set, or regular deployment.
          * **Service port management** - the operator detects which ports need to be exposed based on the provided configuration.
        displayName: Community OpenTelemetry Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opentelemetry
        - tracing
        - logging
        - metrics
        - monitoring
        - troubleshooting
        links:
        - name: OpenTelemetry Operator
          url: https://github.com/open-telemetry/opentelemetry-operator
        maintainers:
        - email: jpkroehling@redhat.com
          name: Juraci Paixão Kröhling
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: OpenTelemetry Community
        relatedImages:
        - ghcr.io/open-telemetry/opentelemetry-operator/opentelemetry-operator:0.136.0
        - quay.io/brancz/kube-rbac-proxy:v0.13.1
        version: 0.136.0
      entries:
      - name: opentelemetry-operator.v0.136.0
        version: 0.136.0
      - name: opentelemetry-operator.v0.135.0
        version: 0.135.0
      - name: opentelemetry-operator.v0.134.0
        version: 0.134.0
      - name: opentelemetry-operator.v0.132.0
        version: 0.132.0
      - name: opentelemetry-operator.v0.131.0
        version: 0.131.0
      - name: opentelemetry-operator.v0.129.1
        version: 0.129.1
      - name: opentelemetry-operator.v0.127.0
        version: 0.127.0
      - name: opentelemetry-operator.v0.126.0
        version: 0.126.0
      - name: opentelemetry-operator.v0.125.0
        version: 0.125.0
      - name: opentelemetry-operator.v0.124.0
        version: 0.124.0
      - name: opentelemetry-operator.v0.123.0
        version: 0.123.0
      - name: opentelemetry-operator.v0.122.0
        version: 0.122.0
      - name: opentelemetry-operator.v0.121.0
        version: 0.121.0
      - name: opentelemetry-operator.v0.120.0
        version: 0.120.0
      - name: opentelemetry-operator.v0.119.0
        version: 0.119.0
      - name: opentelemetry-operator.v0.118.0
        version: 0.118.0
      - name: opentelemetry-operator.v0.117.0
        version: 0.117.0
      - name: opentelemetry-operator.v0.116.0
        version: 0.116.0
      - name: opentelemetry-operator.v0.115.0
        version: 0.115.0
      - name: opentelemetry-operator.v0.114.1
        version: 0.114.1
      - name: opentelemetry-operator.v0.112.0
        version: 0.112.0
      - name: opentelemetry-operator.v0.111.0
        version: 0.111.0
      - name: opentelemetry-operator.v0.110.0
        version: 0.110.0
      - name: opentelemetry-operator.v0.109.0
        version: 0.109.0
      - name: opentelemetry-operator.v0.108.0
        version: 0.108.0
      - name: opentelemetry-operator.v0.107.0
        version: 0.107.0
      - name: opentelemetry-operator.v0.106.0
        version: 0.106.0
      - name: opentelemetry-operator.v0.105.0
        version: 0.105.0
      name: alpha
    defaultChannel: alpha
    packageName: opentelemetry-operator
    provider:
      name: OpenTelemetry Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: StreamNative
      provider-url: https://streamnative.io
    name: sn-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc:latest",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway:0.10.0"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator:v0.10.0
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator:v0.10.0
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      - name: sn-operator.v0.9.0
        version: 0.9.0
      - name: sn-operator.v0.8.5
        version: 0.8.5
      - name: sn-operator.v0.7.10
        version: 0.7.10
      name: alpha
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc:latest",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway:0.10.0"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator:v0.10.0
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator:v0.10.0
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      - name: sn-operator.v0.9.0
        version: 0.9.0
      - name: sn-operator.v0.8.5
        version: 0.8.5
      - name: sn-operator.v0.7.10
        version: 0.7.10
      name: beta
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc:latest",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway:0.10.0"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator:v0.10.0
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator:v0.10.0
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      - name: sn-operator.v0.9.0
        version: 0.9.0
      - name: sn-operator.v0.8.5
        version: 0.8.5
      - name: sn-operator.v0.7.10
        version: 0.7.10
      name: stable
    defaultChannel: alpha
    packageName: sn-operator
    provider:
      name: StreamNative
      url: https://streamnative.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Instana
      provider-url: ""
    name: instana-agent-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: instana-agent-operator.v2.1.35
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instana.io/v1",
                "kind": "InstanaAgent",
                "metadata": {
                  "name": "instana-agent",
                  "namespace": "instana-agent"
                },
                "spec": {
                  "agent": {
                    "configuration_yaml": "# You can leave this empty, or use this to configure your instana agent.\n# See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.\n",
                    "endpointHost": "ingress-red-saas.instana.io",
                    "endpointPort": "443",
                    "env": {},
                    "key": "replace-key"
                  },
                  "cluster": {
                    "name": "my-cluster"
                  },
                  "opentelemetry": {
                    "grpc": {
                      "enabled": true
                    },
                    "http": {
                      "enabled": true
                    }
                  },
                  "zone": {
                    "name": "edited-zone"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring,Logging & Tracing,OpenShift Optional
          certified: "false"
          containerImage: icr.io/instana/instana-agent-operator@sha256:fe949ae0ca9cc05ffe3569cd7a6f4fde48f5e3ee64f27e20e0b3102733a6e809
          description: Fully automated Application Performance Monitoring (APM) for
            microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/instana/instana-agent-operator
          support: Instana
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InstanaAgent is the Schema for the agents API
            displayName: Instana Agent
            kind: InstanaAgent
            name: agents.instana.io
            version: v1
          - description: InstanaAgentRemote is the Schema for the agents API
            displayName: Instana Agent Remote
            kind: InstanaAgentRemote
            name: agentsremote.instana.io
            version: v1
        description: |
          # Instana

          Instana is an [APM solution](https://www.instana.com/product-overview/) built for microservices that enables IT Ops to build applications faster and deliver higher quality services by automating monitoring, tracing and root cause analysis. The solution is optimized for [Kubernetes](https://www.instana.com/automatic-kubernetes-monitoring/) and [OpenShift](https://www.instana.com/blog/automatic-root-cause-analysis-for-openshift-applications/).

          ## Instana Agent Operator

          This is the Kubernetes Operator for installing the Instana Agent on Kubernetes or OpenShift.

          ## Prerequisites for OpenShift

          Before the agent will be able to run in OpenShift, you need to perform a couple of extra configuration steps.

          You need to set up a project for the Instana Agent and configure it's permissions.

          The project you create here needs to be the namespace where you create the Instana Agent custom resource that the operator will use to deploy the agent.

          For example, create the `instana-agent` project:

              oc new-project instana-agent

          Then, ensure the `instana-agent` service account is in the privileged security context:

              oc adm policy add-scc-to-user privileged -z instana-agent

          The `instana-agent-remote` service account must be allowed to use any userid:

              oc adm policy add-scc-to-user anyuid -z instana-agent-remote -n instana-agent

          This service account will be created by the operator.

          Now you can proceed with installing the operator for the Instana agent.

          ## Installation and Configuration

          First, install this operator from [OperatorHub.io](https://operatorhub.io/), [OpenShift Container Platform](https://www.openshift.com/), or [OKD](https://www.okd.io/).

          Second, create a custom resource with the agent configuration in the target namespace (for now, this MUST always be the `instana-agent` namespace). The operator will pick up the custom resource and install the Instana agent accordingly.

          The following is a minimal template of the custom resource:

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgent
          metadata:
            name: instana-agent
            namespace: instana-agent
          spec:
            zone:
              name: my-zone # (optional) name of the zone of the host
            cluster:
              name: replace-me # replace with the name of your Kubernetes cluster
            agent:
              key: replace-me # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io # the monitoring ingress endpoint
              endpointPort: "443" # the monitoring ingress endpoint port, wrapped in quotes
              env:
                INSTANA_AGENT_TAGS: example
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://ibm.biz/monitoring-k8s
          ```

          Save the template in a file `instana-agent.yaml` and edit the following values:

          * If your target namespace is not `instana-agent`, replace the `namespace:` accordingly.
          * `agent.key` must be set with your Instana agent key.
          * `agent.endpointHost` must be set with the monitoring ingress endpoint, generally either `saas-us-west-2.instana.io` or `saas-eu-west-1.instana.io`.
          * `agent.endpointPort` must be set with the monitoring ingress port, generally "443" (wrapped in quotes).
          * `zone.name` should be set with the name of the Kubernetes cluster that is be displayed in Instana.

          For advanced configuration, you can edit the contents of the `configuration.yaml` file. View documentation [here](https://ibm.biz/monitoring-k8s).

          Apply the custom resource with `kubectl apply -f instana-agent.yaml`. After some time, you should see `instana-agent` Pods being created on each node of your cluster, and your cluster should show on the infrastructure map on your Instana Web interface.

          In addition to the InstanaAgent custom resource, a remote agent can be deployed. This agent will not monitor workload running inside the cluster, but can deploy a single replica of an agent which can get an independent configuration for running remote monitoring for certain sensors.

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgentRemote
          metadata:
            name: remote-agent
            namespace: "instana-agent"
          spec:
            zone:
              name: edited-zone # name of the zone of the host
            hostname:
              name: use-custom-hostname-for-remote-agent # (optional) give your remote agent pod a custom hostname
            agent:
              key: replace-key # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io
              endpointPort: "443"
              env: {}
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.
          ```

          ## Uninstalling

          In order to uninstall the Instana agent, simply remove the custom resource with `kubectl delete -f instana-agent.yaml`.

          ## Source Code

          The Instana agent operator is an open source project hosted on [https://github.com/instana/instana-agent-operator](https://github.com/instana/instana-agent-operator/).
        displayName: Instana Agent Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - APM
        - Instana
        links:
        - name: GitHub Repository
          url: https://github.com/instana/instana-agent-operator
        maintainers:
        - email: support@instana.com
          name: Instana
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: Instana
        relatedImages:
        - icr.io/instana/instana-agent-operator@sha256:fe949ae0ca9cc05ffe3569cd7a6f4fde48f5e3ee64f27e20e0b3102733a6e809
        - icr.io/instana/agent@sha256:ed71e6c3e5bdb155853a2db8c916460e6a867fb02627e3a332186188cbbd98dd
        version: 2.1.35
      entries:
      - name: instana-agent-operator.v2.1.35
        version: 2.1.35
      - name: instana-agent-operator.v2.1.34
        version: 2.1.34
      - name: instana-agent-operator.v2.1.33
        version: 2.1.33
      - name: instana-agent-operator.v2.1.32
        version: 2.1.32
      - name: instana-agent-operator.v2.1.31
        version: 2.1.31
      - name: instana-agent-operator.v2.1.30
        version: 2.1.30
      - name: instana-agent-operator.v2.1.29
        version: 2.1.29
      - name: instana-agent-operator.v2.1.28
        version: 2.1.28
      - name: instana-agent-operator.v2.1.27
        version: 2.1.27
      - name: instana-agent-operator.v2.1.25
        version: 2.1.25
      - name: instana-agent-operator.v2.1.24
        version: 2.1.24
      - name: instana-agent-operator.v2.1.23
        version: 2.1.23
      - name: instana-agent-operator.v2.1.22
        version: 2.1.22
      - name: instana-agent-operator.v2.1.21
        version: 2.1.21
      - name: instana-agent-operator.v2.1.20
        version: 2.1.20
      - name: instana-agent-operator.v2.1.19
        version: 2.1.19
      - name: instana-agent-operator.v2.1.18
        version: 2.1.18
      - name: instana-agent-operator.v2.1.17
        version: 2.1.17
      - name: instana-agent-operator.v2.1.16
        version: 2.1.16
      - name: instana-agent-operator.v2.1.15
        version: 2.1.15
      - name: instana-agent-operator.v2.1.14
        version: 2.1.14
      - name: instana-agent-operator.v2.1.13
        version: 2.1.13
      - name: instana-agent-operator.v2.1.12
        version: 2.1.12
      - name: instana-agent-operator.v2.1.11
        version: 2.1.11
      - name: instana-agent-operator.v2.1.10
        version: 2.1.10
      - name: instana-agent-operator.v2.1.9
        version: 2.1.9
      - name: instana-agent-operator.v2.1.8
        version: 2.1.8
      - name: instana-agent-operator.v2.1.7
        version: 2.1.7
      - name: instana-agent-operator.v2.1.6
        version: 2.1.6
      - name: instana-agent-operator.v2.1.5
        version: 2.1.5
      - name: instana-agent-operator.v2.1.4
        version: 2.1.4
      - name: instana-agent-operator.v2.1.3
        version: 2.1.3
      - name: instana-agent-operator.v2.1.2
        version: 2.1.2
      - name: instana-agent-operator.v2.1.1
        version: 2.1.1
      - name: instana-agent-operator.v2.1.0
        version: 2.1.0
      - name: instana-agent-operator.v2.0.32
        version: 2.0.32
      - name: instana-agent-operator.v2.0.31
        version: 2.0.31
      - name: instana-agent-operator.v2.0.30
        version: 2.0.30
      - name: instana-agent-operator.v2.0.27
        version: 2.0.27
      - name: instana-agent-operator.v2.0.19
        version: 2.0.19
      - name: instana-agent-operator.v2.0.18
        version: 2.0.18
      - name: instana-agent-operator.v2.0.17
        version: 2.0.17
      - name: instana-agent-operator.v2.0.13
        version: 2.0.13
      - name: instana-agent-operator.v2.0.12
        version: 2.0.12
      - name: instana-agent-operator.v2.0.11
        version: 2.0.11
      - name: instana-agent-operator.v2.0.10
        version: 2.0.10
      - name: instana-agent-operator.v2.0.9
        version: 2.0.9
      - name: instana-agent-operator.v2.0.8
        version: 2.0.8
      - name: instana-agent-operator.v2.0.7
        version: 2.0.7
      - name: instana-agent-operator.v2.0.5
        version: 2.0.5
      - name: instana-agent-operator.v2.0.4
        version: 2.0.4
      name: stable
    defaultChannel: stable
    packageName: instana-agent-operator
    provider:
      name: Instana
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hdfs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hdfs-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:e8b4ad2c3b127537239b003ae19a3f58b541d599fe3d5a74a6718612fe645d50
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:e8b4ad2c3b127537239b003ae19a3f58b541d599fe3d5a74a6718612fe645d50
        version: 24.11.1
      entries:
      - name: hdfs-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: hdfs-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:00c79036eed6dd6749736b964787acb638f1a6e0156c0bba5651b93738534377
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:00c79036eed6dd6749736b964787acb638f1a6e0156c0bba5651b93738534377
        version: 25.3.0
      entries:
      - name: hdfs-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: hdfs-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
        version: 25.7.0
      entries:
      - name: hdfs-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hdfs-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
        version: 25.7.0
      entries:
      - name: hdfs-operator.v25.7.0
        version: 25.7.0
      - name: hdfs-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hdfs-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: seldon
      provider-url: ""
    name: seldon-deploy-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: seldon-deploy-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1alpha1",
                "kind": "SeldonDeploy",
                "metadata": {
                  "name": "seldondeploy-sample"
                },
                "spec": {
                  "affinity": {},
                  "alibidetect": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306"
                  },
                  "batchjobs": {
                    "mc": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48"
                    },
                    "processor": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d"
                    },
                    "pvc": {
                      "defaultSize": "1Gi"
                    },
                    "serviceAccount": "workflow"
                  },
                  "defaultUserID": "",
                  "docker": {
                    "user": "unknown"
                  },
                  "elasticsearch": {
                    "basicAuth": true,
                    "secret": {
                      "name": "elastic-credentials",
                      "passwordKey": "password",
                      "userKey": "username"
                    },
                    "url": "https://elasticsearch-seldon-es-http.seldon-logs:9200"
                  },
                  "enableAppAnalytics": false,
                  "enableAppAuth": false,
                  "env": {
                    "USERID_CLAIM_KEY": "name"
                  },
                  "external": {
                    "protocol": "http"
                  },
                  "fullnameOverride": "",
                  "gitops": {
                    "argocd": {
                      "enabled": false,
                      "namespace": "argocd"
                    },
                    "fileFormat": "json",
                    "git": {
                      "email": "",
                      "secret": "git-creds",
                      "skipVerifyGit": true,
                      "token": "",
                      "user": "",
                      "webhook": {
                        "service": {
                          "create": true,
                          "loadBalancerSourceRanges": {}
                        }
                      }
                    }
                  },
                  "image": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e",
                    "pullPolicy": "Always"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": []
                      }
                    ],
                    "tls": []
                  },
                  "ingressGateway": {
                    "ingressNamespace": "istio-system",
                    "kfServingIngressService": "istio-ingressgateway",
                    "seldonIngressService": "istio-ingressgateway"
                  },
                  "kfserving": {
                    "curlForm": "MODEL_NAME={{ .ModelName }}\u003cbr\u003e\nCLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\nSERVICE_HOSTNAME=$(kubectl get inferenceservice {{ .ModelName }} -o jsonpath='{.status.url}' | cut -d \"/\" -f 3)\u003cbr\u003e\ncurl -v -H \"Host: ${SERVICE_HOSTNAME}\" {{ .KfServingProtocol }}://$CLUSTER_IP/v1/models/$MODEL_NAME:predict -d '{{ .Payload }}'\n",
                    "enabled": false,
                    "explainForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:explain",
                    "protocol": "http",
                    "requestForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:predict"
                  },
                  "loadtest": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c"
                  },
                  "metadata": {
                    "pg": {
                      "enabled": false,
                      "secret": "metadata-postgres"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "openshiftMarketplace": {
                    "cleanupClusterServiceVersions": false,
                    "kubectlCleanupImage": "registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382",
                    "prometheus": {
                      "monitorSpecs": {
                        "create": true
                      }
                    },
                    "seldonCore": {
                      "istioGateway": {
                        "create": true,
                        "name": "seldon-gateway",
                        "namespace": "istio-system"
                      },
                      "subscription": {
                        "apiVersion": "operators.coreos.com/v1alpha1",
                        "channel": "stable",
                        "create": true,
                        "istioEnabled": true,
                        "metricsPath": "/metrics",
                        "requestLoggerEndpoint": "http://broker-ingress.knative-eventing.svc.cluster.local/seldon-logs/default"
                      }
                    }
                  },
                  "prometheus": {
                    "knative": {
                      "url": "http://prometheus-system-np.knative-monitoring.svc.cluster.local:8080/api/v1/"
                    },
                    "seldon": {
                      "jwtSecretKey": "jwt-seldon.txt",
                      "jwtSecretName": "jwt-seldon",
                      "namespaceMetricName": "namespace",
                      "resourceMetricsUrl": "https://prometheus-k8s.openshift-monitoring:9091/api/v1/",
                      "serviceMetricName": "exported_service",
                      "url": "https://thanos-querier.openshift-monitoring.svc:9091/api/v1/"
                    }
                  },
                  "rbac": {
                    "clusterWide": true,
                    "create": true,
                    "readNamespaces": true
                  },
                  "replicaCount": 1,
                  "requestLogger": {
                    "create": true,
                    "elasticsearch": {
                      "host": "elasticsearch-seldon-es-http.seldon-logs",
                      "port": "9200",
                      "protocol": "https"
                    },
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398",
                    "imagePullPolicy": "IfNotPresent",
                    "namespace": {
                      "create": false,
                      "name": "seldon-logs"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "600m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "200Mi"
                      }
                    },
                    "trigger": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "broker": "default"
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "800m",
                      "memory": "800Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "seldon": {
                    "curlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions -d '{{ .Payload }}'\n",
                    "enabled": true,
                    "explainForm": "http://{{ .ModelName }}-{{ .Predictor }}-explainer.{{ .Namespace }}:9000/v1/models/{{ .ModelName }}:explain",
                    "kfservingV2CurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .GraphModelName }}/infer -d '{{ .Payload }}'\n",
                    "protocol": "http",
                    "seldonRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions",
                    "tensorFlowCurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict -d '{{ .Payload }}'\n",
                    "tensorflowRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict",
                    "v2RequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .ModelName }}/infer"
                  },
                  "service": {
                    "port": 80,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "create": true
                  },
                  "serviceAccountName": "seldon-deploy",
                  "skipVerifyHttpCalls": true,
                  "tolerations": [],
                  "virtualService": {
                    "create": true,
                    "gateways": [
                      "istio-system/seldon-gateway"
                    ],
                    "prefix": "/seldon-deploy/"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
          createdAt: "2020-12-11 15:00:00"
          description: The Seldon deploy operator allows the creation of Seldon Deploy
            applications to provide management, monitoring and updating of machine
            learning models from a range of technologies. It includes the ability
            to run explanations as well as outlier, drift and adversarial detection
            against running models. There is a full metrics and logging integration
            via Prometheus and ELK stacks as well as full GitOps operations to allow
            full auditing and recovery of updated models.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SeldonIO/seldon-deploy-operator
          support: Seldon Dev
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: SeldonDeploy
            name: seldondeploys.machinelearning.seldon.io
            version: v1alpha1
        description: The Seldon deploy operator allows the creation of Seldon Deploy
          applications to provide management, monitoring and updating of machine learning
          models from a range of technologies. It includes the ability to run explanations
          as well as outlier, drift and adversarial detection against running models.
          There is a full metrics and logging integration via Prometheus and ELK stacks
          as well as full GitOps operations to allow full auditing and recovery of
          updated models.
        displayName: Seldon Deploy Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - machine learning
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://deploy.seldon.io
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        provider:
          name: seldon
        relatedImages:
        - registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398
        - registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
        - registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c
        version: 1.2.1
      entries:
      - name: seldon-deploy-operator.v1.2.1
        version: 1.2.1
      - name: seldon-deploy-operator.v1.2.0
        version: 1.2.0
      - name: seldon-deploy-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: seldon-deploy-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1alpha1",
                "kind": "SeldonDeploy",
                "metadata": {
                  "name": "seldondeploy-sample"
                },
                "spec": {
                  "affinity": {},
                  "alibidetect": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306"
                  },
                  "batchjobs": {
                    "mc": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48"
                    },
                    "processor": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d"
                    },
                    "pvc": {
                      "defaultSize": "1Gi"
                    },
                    "serviceAccount": "workflow"
                  },
                  "defaultUserID": "",
                  "docker": {
                    "user": "unknown"
                  },
                  "elasticsearch": {
                    "basicAuth": true,
                    "secret": {
                      "name": "elastic-credentials",
                      "passwordKey": "password",
                      "userKey": "username"
                    },
                    "url": "https://elasticsearch-seldon-es-http.seldon-logs:9200"
                  },
                  "enableAppAnalytics": false,
                  "enableAppAuth": false,
                  "env": {
                    "USERID_CLAIM_KEY": "name"
                  },
                  "external": {
                    "protocol": "http"
                  },
                  "fullnameOverride": "",
                  "gitops": {
                    "argocd": {
                      "enabled": false,
                      "namespace": "argocd"
                    },
                    "fileFormat": "json",
                    "git": {
                      "email": "",
                      "secret": "git-creds",
                      "skipVerifyGit": true,
                      "token": "",
                      "user": "",
                      "webhook": {
                        "service": {
                          "create": true,
                          "loadBalancerSourceRanges": {}
                        }
                      }
                    }
                  },
                  "image": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e",
                    "pullPolicy": "Always"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": []
                      }
                    ],
                    "tls": []
                  },
                  "ingressGateway": {
                    "ingressNamespace": "istio-system",
                    "kfServingIngressService": "istio-ingressgateway",
                    "seldonIngressService": "istio-ingressgateway"
                  },
                  "kfserving": {
                    "curlForm": "MODEL_NAME={{ .ModelName }}\u003cbr\u003e\nCLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\nSERVICE_HOSTNAME=$(kubectl get inferenceservice {{ .ModelName }} -o jsonpath='{.status.url}' | cut -d \"/\" -f 3)\u003cbr\u003e\ncurl -v -H \"Host: ${SERVICE_HOSTNAME}\" {{ .KfServingProtocol }}://$CLUSTER_IP/v1/models/$MODEL_NAME:predict -d '{{ .Payload }}'\n",
                    "enabled": false,
                    "explainForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:explain",
                    "protocol": "http",
                    "requestForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:predict"
                  },
                  "loadtest": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c"
                  },
                  "metadata": {
                    "pg": {
                      "enabled": false,
                      "secret": "metadata-postgres"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "openshiftMarketplace": {
                    "cleanupClusterServiceVersions": false,
                    "kubectlCleanupImage": "registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382",
                    "prometheus": {
                      "monitorSpecs": {
                        "create": true
                      }
                    },
                    "seldonCore": {
                      "istioGateway": {
                        "create": true,
                        "name": "seldon-gateway",
                        "namespace": "istio-system"
                      },
                      "subscription": {
                        "apiVersion": "operators.coreos.com/v1alpha1",
                        "channel": "stable",
                        "create": true,
                        "istioEnabled": true,
                        "metricsPath": "/metrics",
                        "requestLoggerEndpoint": "http://broker-ingress.knative-eventing.svc.cluster.local/seldon-logs/default"
                      }
                    }
                  },
                  "prometheus": {
                    "knative": {
                      "url": "http://prometheus-system-np.knative-monitoring.svc.cluster.local:8080/api/v1/"
                    },
                    "seldon": {
                      "jwtSecretKey": "jwt-seldon.txt",
                      "jwtSecretName": "jwt-seldon",
                      "namespaceMetricName": "namespace",
                      "resourceMetricsUrl": "https://prometheus-k8s.openshift-monitoring:9091/api/v1/",
                      "serviceMetricName": "exported_service",
                      "url": "https://thanos-querier.openshift-monitoring.svc:9091/api/v1/"
                    }
                  },
                  "rbac": {
                    "clusterWide": true,
                    "create": true,
                    "readNamespaces": true
                  },
                  "replicaCount": 1,
                  "requestLogger": {
                    "create": true,
                    "elasticsearch": {
                      "host": "elasticsearch-seldon-es-http.seldon-logs",
                      "port": "9200",
                      "protocol": "https"
                    },
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398",
                    "imagePullPolicy": "IfNotPresent",
                    "namespace": {
                      "create": false,
                      "name": "seldon-logs"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "600m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "200Mi"
                      }
                    },
                    "trigger": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "broker": "default"
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "800m",
                      "memory": "800Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "seldon": {
                    "curlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions -d '{{ .Payload }}'\n",
                    "enabled": true,
                    "explainForm": "http://{{ .ModelName }}-{{ .Predictor }}-explainer.{{ .Namespace }}:9000/v1/models/{{ .ModelName }}:explain",
                    "kfservingV2CurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .GraphModelName }}/infer -d '{{ .Payload }}'\n",
                    "protocol": "http",
                    "seldonRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions",
                    "tensorFlowCurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict -d '{{ .Payload }}'\n",
                    "tensorflowRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict",
                    "v2RequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .ModelName }}/infer"
                  },
                  "service": {
                    "port": 80,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "create": true
                  },
                  "serviceAccountName": "seldon-deploy",
                  "skipVerifyHttpCalls": true,
                  "tolerations": [],
                  "virtualService": {
                    "create": true,
                    "gateways": [
                      "istio-system/seldon-gateway"
                    ],
                    "prefix": "/seldon-deploy/"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
          createdAt: "2020-12-11 15:00:00"
          description: The Seldon deploy operator allows the creation of Seldon Deploy
            applications to provide management, monitoring and updating of machine
            learning models from a range of technologies. It includes the ability
            to run explanations as well as outlier, drift and adversarial detection
            against running models. There is a full metrics and logging integration
            via Prometheus and ELK stacks as well as full GitOps operations to allow
            full auditing and recovery of updated models.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SeldonIO/seldon-deploy-operator
          support: Seldon Dev
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: SeldonDeploy
            name: seldondeploys.machinelearning.seldon.io
            version: v1alpha1
        description: The Seldon deploy operator allows the creation of Seldon Deploy
          applications to provide management, monitoring and updating of machine learning
          models from a range of technologies. It includes the ability to run explanations
          as well as outlier, drift and adversarial detection against running models.
          There is a full metrics and logging integration via Prometheus and ELK stacks
          as well as full GitOps operations to allow full auditing and recovery of
          updated models.
        displayName: Seldon Deploy Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - machine learning
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://deploy.seldon.io
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        provider:
          name: seldon
        relatedImages:
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d
        - registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48
        - registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c
        - registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e
        version: 1.2.1
      entries:
      - name: seldon-deploy-operator.v1.2.1
        version: 1.2.1
      - name: seldon-deploy-operator.v1.2.0
        version: 1.2.0
      - name: seldon-deploy-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: seldon-deploy-operator-rhmp
    provider:
      name: seldon
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: rhods-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhods-operator.2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
          createdAt: "2025-08-12T18:04:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "llamastackoperator": {
                      "managementState": "Removed"
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (including PyTorch, TensorFlow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack – unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.23
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fbe10282ada2fa6bf07ae97b497f46ebe735de130b4ce0182bf34301429999cb
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:2d2e7803b2b1e15017ca04961f9dcd25e89c47b07b101ca54cb9e793e6027202
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:e50adae13e9d773e7be8bf81b7525dac3f5e3864c9cdf4c287342bd568ee1d6b
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:bb7c352ccc10a8c60ef8d573368d0ada59e73cd131653d11bcebdf9f564acc10
        - registry.redhat.io/rhel9/mariadb-105@sha256:050dd5a7a32395b73b8680570e967e55050b152727412fdd73a25d8816e62d53
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:ccfc84028a95f1cc6a16e06ff6e9ed2ace74f74e5e9963a4fb33544c2ce005c2
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:2886e40f94fde72eb5959ad564bbdd466657208c09bac731b3860b0b54f104c6
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:b038d20def504c140e4365f2eb117a979534629332bec75a2ba750ebafdefe3d
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5fb7e6c3ee1e3b170cd146ea9cd172fa953326b08163897605ca74a3644f2819
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:795ab1842ef675e3d8b47deb8a7485a69064428e823dfeb930a6a7e7017cde0f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:80d284cac2ef6638fe028102b71a70f26041576c11a49414e77430954a3c1b99
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f23f8051886d0f0240937507471c818f440bca8703b8209a836d5dbe3ea2f7c8
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:16ff53f4f1a1e1e66826c8355c654038da83df545427a931485ec7a2d3764b16
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:23c098f7e791441cfcfa6bf37de51763511bd8bf4d543747b7d37ed591d68886
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:edcce7cddfa4c250108532e56541ddb95099612c180b2297e5934bc89d2bb46e
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:ff740fc3615edaa90dae976da1aa6455d64811e239eb66b422836a3aae791b41
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:240692b86bc8a4fe0b545e6e4cd7cf4ff0ca990406bc55f0446d565fd67e37f6
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:a4649e8ad4970bc3d5e967811c9694fa82d78ee9f8f025b8b2fbd37237eb14fc
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:68e74bdf518fc624cc25a59a18b2a87d0aa2a9b8aad30750ace7115d297a21ca
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:ca21e218e26c46e3c63d926241846f8f307fd4a586cc4b04147da49af6018ef5
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:857560240bb45bfc2db6d7743abf11a47f929a0d3035940256a6527afd313412
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:84c3213e11299ec032aa0e5a4dfa9b3055c4246b04b029e888d8d9f1caf5c30c
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:4847ac1be55284815670c2048c1edfc56c7227494280d3c6094aa23f1ed1a9d2
        - registry.redhat.io/ubi9/toolbox@sha256:4213516b6fe7957a01072720efd9ecbc7d4443238e151b5fe0263bd0c36792b1
        - registry.redhat.io/rhoai/odh-llama-stack-core-rhel9@sha256:43b60b1ee6f66fec38fe2ffbbe08dca8541ef162332e4bd8e422ecd24ee02646
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:467b04f4ee4118bbe0934d1ab8b80f72df8703dcdac61c2b53fdb5bd1b404c14
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:a1497e64349634b2567d28f685052de118f64f6b32e7d3ea469212e2004d7eb3
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:e0c9ac73891781ef421e0a020cfd5c962f7632d38020897183cf048b1819326a
        - registry.redhat.io/rhoai/odh-llama-stack-k8s-operator-rhel9@sha256:626b9ef09ee82c8d425ddb7ed3f2ffbb3d1019f541e31604db891787c881d5c9
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:3243caa4c8aac75069d19225dcee361bf9c012782328a6266b7016b3026f6d56
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:756b7bdaa7f6e7c4a18488e690188dd91153a33913d225bab001b0085382e514
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:da8debeb2b84a42aba0b98785ccc20e02ed520e125428fe012df3ca473cbb755
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:208005da389b7d4176ad8bffc703312011a1299954df151e2abbe8f3dec01d19
        version: 2.23.0
      entries:
      - name: rhods-operator.2.23.0
        version: 2.23.0
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.20.0
        version: 2.20.0
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.18.0
        version: 2.18.0
      - name: rhods-operator.2.17.0
        version: 2.17.0
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.15.0
        version: 2.15.0
      - name: rhods-operator.2.14.0
        version: 2.14.0
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: alpha
    - currentCSV: rhods-operator.2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhods/odh-rhel8-operator@sha256:a4d7f020509b9720063e87e1c2f19af04e4c9b4986b199be4dc617831b971434
          createdAt: "2023-11-21T08:37:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            Data Science
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:08c90a8f133b4cdb1477a742bc44be4d8fc4e88153b8bc2a2a148cbdf19f4e7d
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:6a56010de933f172b195a1a575855d37b70a4968be8edb35157f6ca193969ad2
          olm.skipRange: '>=1.33.0 <2.4.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "rhods",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "rhods",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Data Science"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:eb96f4b753606528750b3c28f7f9256582f81fa030450b05afdef684634903be
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift Data Science
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift Data Science (RHODS) is a complete platform for the entire lifecycle of your AI/ML projects. It is the flagship product of OpenShift AI.

          When using RHODS, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. RHODS can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * RHODS dashboard
          * Curated Notebook Images (incl CUDA, PyTorch, Tensorflow)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh (and Kserve) with a provided OpenVino Model Serving Runtime
          * Ability to use other runtimes for serving
          * Model Monitoring
        displayName: Red Hat OpenShift Data Science
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift Data Science
        - RHODS
        - rhods
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift Data Science
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-data-science
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhods/odh-ml-pipelines-api-server-rhel8@sha256:be57f237df156474a709667fa4fb4aa142118e3459963fed56648c1cb2e38cd6
        - registry.redhat.io/rhods/odh-kuberay-operator-controller-rhel8@sha256:bfc805aa1689a4bd6f7e617acf9eb1c2c307624dd49956e9e7442afac328151a
        - registry.redhat.io/rhods/odh-data-science-pipelines-operator-controller-rhel8@sha256:250db47cb873207cfb75c7e8a2f0de08b75b056a72db73362d2d79c08cb7b185
        - registry.redhat.io/rhods/odh-trustyai-service-operator-rhel8@sha256:68471555d0b2bb8150ebfae8d4f3d1dbdff9c5b961ff2305ef23ab7514cbbea1
        - registry.redhat.io/rhods/odh-codeflare-operator-rhel8@sha256:2841fd3051a50a1b0d9e3caafa20a8084b5c472d49393b372828960dbf156674
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhods/odh-model-controller-rhel8@sha256:fde989817cd1a6ec8eb408a7f6c73e28ba043ab2d88035586b1b7e55fb3405cb
        - registry.redhat.io/rhods/odh-modelmesh-runtime-adapter-rhel8@sha256:08478a1647d0413783bcbcd966a73b444b580464de46efedb39504420f22053e
        - registry.redhat.io/rhel8/mariadb-103@sha256:08c90a8f133b4cdb1477a742bc44be4d8fc4e88153b8bc2a2a148cbdf19f4e7d
        - registry.redhat.io/rhods/odh-trustyai-service-rhel8@sha256:fe17d6e07ebaa088f274dd8b5e83892dacc11cb336065b06582a3c055095dba5
        - registry.redhat.io/rhods/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:c5758102609821a6fdabec0277e03c4c728f82176093d1ade0570aba5f3cd8f9
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhods/odh-ml-pipelines-persistenceagent-rhel8@sha256:857a8eb444375505ccb34a11cce69c1794ceabbfb00df06933d7f12cd187319b
        - registry.redhat.io/ubi8/ubi-micro@sha256:6a56010de933f172b195a1a575855d37b70a4968be8edb35157f6ca193969ad2
        - registry.redhat.io/rhods/odh-rhel8-operator@sha256:a4d7f020509b9720063e87e1c2f19af04e4c9b4986b199be4dc617831b971434
        - registry.redhat.io/rhods/odh-mm-rest-proxy-rhel8@sha256:8d989f54009484d7f43fe79d7c37757418fc3c15cd4d7e80a2076677a9645bdf
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhods/odh-ml-pipelines-cache-rhel8@sha256:33aae32e3d348d382906f6e926281852f54006a94d7bdac188bb33debadec059
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhods/odh-modelmesh-rhel8@sha256:3827221f84d3aa22440d6b364b990a4ec0666dfb5963f174f7ed518b3650d223
        - registry.redhat.io/rhods/odh-kf-notebook-controller-rhel8@sha256:cd0ab8d70559be6074fbaf1caa1d50f8eb75e1bff1a22d6d69380f19190df0b6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:eb96f4b753606528750b3c28f7f9256582f81fa030450b05afdef684634903be
        - registry.redhat.io/rhods/odh-ml-pipelines-artifact-manager-rhel8@sha256:76c67d7015a2db6965219db6533e0cddb5b7cf699c88df3ef81155f9c192a3cc
        - registry.redhat.io/rhods/odh-modelmesh-serving-controller-rhel8@sha256:d94395fb251e45aaa73d09bc6106a2947c16b90d87fc3b64bfe4f2f600be5841
        - registry.redhat.io/rhods/odh-dashboard-rhel8@sha256:b98171c5543815abf7c095fba92796ceed9ea44bd8b78db143d2815c6bcfdcf6
        - registry.redhat.io/rhods/odh-notebook-controller-rhel8@sha256:753994fce63802f1e1121990a5c1dd722216446abee54cc1a7c50ab195134414
        version: 2.4.0
      entries:
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: beta
    - currentCSV: rhods-operator.2.21.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:7d0b8d69b6defb5f232da066573b0067ec810a8ccfb2f3cb91d472d2979ab7a7
          createdAt: "2025-09-24T14:42:33Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.21
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:687c8eeed55f021ecaab1307f0e88b5b16d91f72d63b3d7168d7bbee90e8947b
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:04859c878ffcc8c42550a98be625ac52212f9b79a6fd5b3baeafb3592926cbe9
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:2a16c353fd13dc1f3f5c4266f1b9786a8e7a100dc05f2d554bdaef082271063f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:239f226328e5ac77d6dee498217d55103a6d562f4829e38abf65ba55fda0aca2
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:7da004e158006dcbdcf0c3b78c40411b82c9d56ebf54b1c30c245b6550b8ee7d
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:2b4136569e10ad7d47c82bedf55c1705a833dc67c8f833656996bfe1fddbef3e
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:52e1ba8ebf4d55c64c94c9076185c5612b537e6e5ed4944fcd6ea87ccb9cd182
        - registry.redhat.io/ubi9/toolbox@sha256:7656f815a0d13ed37f56d11635559a87a3d00840253755c71b644b858f67d556
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:7d0b8d69b6defb5f232da066573b0067ec810a8ccfb2f3cb91d472d2979ab7a7
        - registry.redhat.io/rhel9/mariadb-105@sha256:c390b12f5c2c3ea7cd897b59eba7127ce22f474130cd8c1bbbbaadae870baca4
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:1b5452b87b36de2b62febc777efdc6233a168ffcc09e2067e482cc444463a57d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:cb47e165935f0d9d3c88088689c37e68a69519357a20d3c52ab6f71240dab507
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:003fdc6cd1821d620e1ce4050ee608e2d2a32f2b08c8fa9c147e2e14418a8d91
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:80fd53a580ce7bc4d02023d38a04912e84c2cd50679fc5ea034afba935623157
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:fe6593bc73e77d315d82b7921bb1975c19fda1e7630ef5b30b69f5901ce30178
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:e68492a6a75cb83be0a968b832c1478b97f14e661376328085fd5416a08cbc31
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:c3e283c0536266195dd461042ca0036ed98cc33650d806be132db4fba45239e4
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:152b425725228953f43b57a4a034a6725993a5ca556af328ae5e06aa3f1acbfc
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:a88696cff8cd2cc2e295c9ec082b7f25f61dfc9ac12f1c208529ff7036d5af29
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fc98ed817a4e9ab84e2a66063987304f0cdbe6e5620a58a7f25cccd61b34ebf9
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:5c0d56f434298f6e8bf6cf7061cd1ae6df9915225cbf04b010b6b59a64775cc9
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:cfaf4eb86ddb9c717340ab17fa4b83323e5b40977ac70265376c87329a3f6f20
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:4c57dcf4df09fe0d03578cc8b9a3ca2ea485b6af0ca0203bfb85e640058b7daf
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:4c192208c618b49f31cf31c9de1a33c622ce3f120204c2eec8ff7c439ea6785c
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:a936d87646843edea2a84eb429bb44eb730bf3e04e8900fecef66721426dba89
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:9c2d6ce05775ac34ee5695b3367d12bf2a9c14e318a9544d00a56653ea5f2007
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:dc0db686b7688f98b5f8595661022cdffa5d26a078289cc7aed528f0ee1d08a4
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:6a388ee3cbe3813ac955d319f0ab65a81da321d982c9881c6c81fe135311fe1d
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5f581957e4f8deebf6104a394272e6fbb1c54d813bf069e95875d865eac209ad
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:2638e3a90bbb242dd2483fefe9fc8b472f334c510d0139b9b7c875c30dab7984
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:85202cb75250a1ba2691d65b52b306fd77566660dd29cdf56bf8962e8bb4939c
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:1b7befcd29c9c8f783bdd3c5654b0412ea45cd88cd18eb5f8f5a411f1323c6bd
        version: 2.21.1
      entries:
      - name: rhods-operator.2.21.1
        version: 2.21.1
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: embedded
    - currentCSV: rhods-operator.2.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
          createdAt: "2025-09-24T13:27:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.16
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:22557a6fa52d2f311750a9ba253860f423ba697d26efa02ef8524a8258d2a909
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:d7d4fa406e0fcf0507894a7676532b27f45be742467e603a86f98ea5d2615df8
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:2ba2b2c4db8bb334c50f4dfb54059f060361186900a44c06eed00b7a3c43977e
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:7b70af8847b0806d8b43c399e2b3109f016fd864f5e9d30c44e2baca5d1359dc
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a0f72ffefb2b74b488dd949493f5d295a39bb9c97f578bf219d0138601f65468
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:6cf74044ae8d5308a2dfe03fa5d81086c89302db7f4cdbe2f4174a1c48b77869
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:de5f91180ead7d73a1825fe8b032fde9b8b01392569c9789f5ce1b4b9c08a98f
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:ee01e89f98feb185f6cd59c564e590a13e4d8d9ea760cca8de51426eb71b83a1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:04f305c44413ae7dcb7017e53570ee49a509701792c5f50efadd64f47395730b
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:c8456d98e90b6505957ab3686e9fd2f156e29f123c5558e581c206daf1e7d93a
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:f738aab1eab25854c93e1d8d4d98100a8ae7bb45a6b83f0326774e4220b1183b
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:5dcdcc2424602a69451f16d31dbfa1d43cb72c095ba561eb9076f0cd1e8182ed
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:efd039012559786deb3c521a130886c265e88d635e08baace2e573a9df270134
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:27b113d96453c2054d4c965a963d2badba6daac235eddaf23234c76e87bcb069
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:4c5ff3496b2a2a739939d94ee9dafc02b682100785d228dc2fde480fb597b7a5
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:4a7599f8a866eb169c9a62885906adbf6df0417c0be15857df1eef20cd9b1be2
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:633984b538f027c93c5886ea1045dcbe81eeda74acf80001d5fc5f765bdbe0be
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:23d307a36b69e0df04f72a7d3b35e28d8417a8bbe23dba31e8e977569785c078
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:11a301728573adf2b64ea72d0cb2d83ea5d4dbebea759f346e99f18c3d368c6e
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:5192f1dbd2a9ab92ae390c4ae506efbed0970545b6122e95b014728ac937e777
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:2eedd1e2465d2e78fafb39e234dbb413a82713f187c64d70a80340ec94807d30
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f37e4048f3a152798286793f3abfc6ed814453fcbe2667255a7e78eee483c5a3
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:97e2bd9b587f08e135a9aeb9b3e0dc6eafa1a9bdacbb5ecb681ce9bd5aa37fc9
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:65482864055021272a18b5b26792ab00cfa5fc9cc005d8d3a884cc82991506f3
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:4a212ec634225c14beac09be24ddf336e562f2aa9a13555fb1196f366ddae23c
        version: 2.16.3
      entries:
      - name: rhods-operator.2.16.3
        version: 2.16.3
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.8.5
        version: 2.8.5
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: eus-2.16
    - currentCSV: rhods-operator.2.8.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:3b9e8a71ba795a2ea91f4129253ccb4be3a79c7c1c6a87082f96ccc67eabbbac
          createdAt: "2024-11-07T07:17:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.8.0 <2.8.5'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:eeb8bdc94e7fa2b454fedf72eb75756052f8bfb9d92b116a8267a5584e2528c1
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:f9c474d454d0b006e06859601ec9a64b5e5c91e9a47bfab6ad6d8cce5aa6b89d
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:45498198fadfbc1457bf2a6084ae7ea515d454d86d148783aae3abdaca6e999d
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:60507a3dddbd88560fd85204699147facd2b6371fb25e0bf6701dad02c0b4dd4
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:df34e9d2d121a3def2dbb86f108b3fb6948bc7ba7cb44c70801dfd2e81d3fbb1
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:13dff890ec67fa56864c244a40fc404e9eac05f33573dd16d088b37f9340ec41
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:28ec01e514c15cae145499840597d1be1b750123bb6014de2ecc3ef0d60fa134
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:9aa5a93831992588766ac8594ae3c4dc7111d08a90fdbc8751c40484339e46d4
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:038ce76571217d6c624d614652e8b96ef6484562cf52ec1be88392b8ef8a3bbd
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:1a1074263ae26e4c8b0e3863228c35e50133169f377ece6c47b58a12e0fcf453
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:81ae2b95f2da740cb5264b9ba8b85f0f58bb4ab80d6431f66d52848ed0c21afb
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:1dbf09cb1624c9f961f8ac8e371c97aa7c3fbfcad4c4d02ab4fad0eb86d2aa95
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:61a506fbf263cd72951ac8bd151127e0c74fc74f0a05f7aaecaa54fde47ee490
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:3b9e8a71ba795a2ea91f4129253ccb4be3a79c7c1c6a87082f96ccc67eabbbac
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:13b850a5a3f8776c6c8db887c666d88df72901895b5ef1be9a8bd372cc02fcb8
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:8ffc5da798e783c65d5be7f344219b2951d8c2233d6494a4ae2bb5c86eccd699
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:d87e4b1cba72eb4f50b3f2d59863799929c22ca37b01a1ff4828a9ce0be31f25
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:2f6db789ceb3f90f9c9911c6392e62d0a8e4d61be68a7f7cbd781f47faab6389
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:4b1b9f05a94713292a7a2f6782fe1d3eea6f36214c36b38f402aa91a554ab18a
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:5c48474563d98df32e7243e33302428ea8836be5e1a2e30d20b4d796a62cbf6b
        version: 2.8.5
      entries:
      - name: rhods-operator.2.8.5
        version: 2.8.5
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: eus-2.8
    - currentCSV: rhods-operator.2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
          createdAt: "2025-08-12T18:04:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "llamastackoperator": {
                      "managementState": "Removed"
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (including PyTorch, TensorFlow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack – unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.23
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:edcce7cddfa4c250108532e56541ddb95099612c180b2297e5934bc89d2bb46e
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-llama-stack-k8s-operator-rhel9@sha256:626b9ef09ee82c8d425ddb7ed3f2ffbb3d1019f541e31604db891787c881d5c9
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:ff740fc3615edaa90dae976da1aa6455d64811e239eb66b422836a3aae791b41
        - registry.redhat.io/ubi9/toolbox@sha256:4213516b6fe7957a01072720efd9ecbc7d4443238e151b5fe0263bd0c36792b1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhel9/mariadb-105@sha256:050dd5a7a32395b73b8680570e967e55050b152727412fdd73a25d8816e62d53
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:68e74bdf518fc624cc25a59a18b2a87d0aa2a9b8aad30750ace7115d297a21ca
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:a1497e64349634b2567d28f685052de118f64f6b32e7d3ea469212e2004d7eb3
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:e50adae13e9d773e7be8bf81b7525dac3f5e3864c9cdf4c287342bd568ee1d6b
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:84c3213e11299ec032aa0e5a4dfa9b3055c4246b04b029e888d8d9f1caf5c30c
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:ccfc84028a95f1cc6a16e06ff6e9ed2ace74f74e5e9963a4fb33544c2ce005c2
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:bb7c352ccc10a8c60ef8d573368d0ada59e73cd131653d11bcebdf9f564acc10
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:2d2e7803b2b1e15017ca04961f9dcd25e89c47b07b101ca54cb9e793e6027202
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:b038d20def504c140e4365f2eb117a979534629332bec75a2ba750ebafdefe3d
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:3243caa4c8aac75069d19225dcee361bf9c012782328a6266b7016b3026f6d56
        - registry.redhat.io/rhoai/odh-llama-stack-core-rhel9@sha256:43b60b1ee6f66fec38fe2ffbbe08dca8541ef162332e4bd8e422ecd24ee02646
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:23c098f7e791441cfcfa6bf37de51763511bd8bf4d543747b7d37ed591d68886
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5fb7e6c3ee1e3b170cd146ea9cd172fa953326b08163897605ca74a3644f2819
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:857560240bb45bfc2db6d7743abf11a47f929a0d3035940256a6527afd313412
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:795ab1842ef675e3d8b47deb8a7485a69064428e823dfeb930a6a7e7017cde0f
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f23f8051886d0f0240937507471c818f440bca8703b8209a836d5dbe3ea2f7c8
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fbe10282ada2fa6bf07ae97b497f46ebe735de130b4ce0182bf34301429999cb
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:80d284cac2ef6638fe028102b71a70f26041576c11a49414e77430954a3c1b99
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:467b04f4ee4118bbe0934d1ab8b80f72df8703dcdac61c2b53fdb5bd1b404c14
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:ca21e218e26c46e3c63d926241846f8f307fd4a586cc4b04147da49af6018ef5
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:e0c9ac73891781ef421e0a020cfd5c962f7632d38020897183cf048b1819326a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:a4649e8ad4970bc3d5e967811c9694fa82d78ee9f8f025b8b2fbd37237eb14fc
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:208005da389b7d4176ad8bffc703312011a1299954df151e2abbe8f3dec01d19
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:240692b86bc8a4fe0b545e6e4cd7cf4ff0ca990406bc55f0446d565fd67e37f6
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:756b7bdaa7f6e7c4a18488e690188dd91153a33913d225bab001b0085382e514
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:2886e40f94fde72eb5959ad564bbdd466657208c09bac731b3860b0b54f104c6
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:16ff53f4f1a1e1e66826c8355c654038da83df545427a931485ec7a2d3764b16
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:4847ac1be55284815670c2048c1edfc56c7227494280d3c6094aa23f1ed1a9d2
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:da8debeb2b84a42aba0b98785ccc20e02ed520e125428fe012df3ca473cbb755
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        version: 2.23.0
      entries:
      - name: rhods-operator.2.23.0
        version: 2.23.0
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.20.0
        version: 2.20.0
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.18.0
        version: 2.18.0
      - name: rhods-operator.2.17.0
        version: 2.17.0
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.15.0
        version: 2.15.0
      - name: rhods-operator.2.14.0
        version: 2.14.0
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: fast
    - currentCSV: rhods-operator.2.22.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
          createdAt: "2025-09-25T11:28:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.22
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:7b3d7787ba95156de9c284ad78f2d63a0d3bcb525f4d8d11ebf0203d4c5c1fe0
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:7fd540586a1ca111370100a473923b3dc6cec4255e9e4f3654bfbabd9158a2bb
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:a44d32a18b1e906c1411cafc82102881573fcd04e040a4e2ba6bd1bd617b70fa
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:6234e15f001122ab222481321153fb0eba7f4d86bf3dfd64edf3f6cecbe7d62f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:abe1f9dd11c7313fdd55b678be511a5cdc3546654a68e2b53d42c1013fec8d00
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:b2ca229bb4c4dbf04b96766b8af8d9f55f2edef59ec4110de0d20f8c857e7d27
        - registry.redhat.io/rhel9/mariadb-105@sha256:5e77b33ce04c1eafc2e8840586ad6a5b51148dea8a46dda151071e28858331f0
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:2f3f305c97e50057781553416fb4026cde306ead0fbe0213627b9b5bbcc8a0c3
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:e6fbe6723b4801f6854303b3fd7d4a4a36727bdd28dc28002b7de4929739ccc0
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:2f1a0a0803b2f5118b167e9d9c2e86660add5f79ee6f47044f63b0e8abe0b75f
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:cb5be3dd7098f4c8d66f9e85ad04656460df6102652deddc6cccb884a12be733
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:9104cbc2d595933f0a2acb7313e41ed69905e76a9fae6bd9820e7717b56a632b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e3b90530c6e38f6940d7f9408e9fa2c81eb06a857ec2b0bab5d006b1e9a63a4f
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:32c9511846077ef6035eb19be7653f6a78e0f207eb420ab93b18f742ab5779b6
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:27415b1c763cf89b79391f0b039936c3e078a078bda4420f13bf10ef9be31d5e
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:463e9be9c4409ba1e5a42713bd65844a023e7a3e9032d4e862edbaa5fe35300a
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:1709fa3c79aad4ba7eb9be8299949396092c8e20210124e0c0936385bc04e839
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:b2c2ea9d35edbcd16e1ae559c88a453129b877fc049ab7a7cd6c2ebb8f0bcb59
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:bd6d1599b4ff9eb890be9808014715b28ca19247222d65f40df024db4dd3525a
        - registry.redhat.io/ubi9/toolbox@sha256:ea4ce8cb8f59cf1b0709f1606d7cab074c3806d4016964eedc0555410cb5dfeb
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:d59ed40de694bc0e0564bab159a5b586cffa0721b6d761fb2d81dd42e6225d1c
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:74d4e0174655cf6f6273857325a1e714a1531c5a06ad38bf139bee9c7d8f5f19
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:5d79f0bcca0a213b53aee6d88834e503d343821596b857b13d3a6362de8ebeab
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:5914f71e37c815ce2567826896e051303bdef52fae1acf8d02e672fc24c7eb9d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:2065b96bd6f0a830eab66e36217a927448cfc5a7dabf2a5e7d907aa6c40657a4
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:2b3aace0d80474470e71c2c8291b25738e096109229fe6d5d3e06e9885ee58b6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:13f3679ab289eea9de54fe129a7c83f6c8ec5d26b61b430637a7c4cfcd31a74a
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:ccba67a6a5aa3403e0018f29dcdfef483220dea8065d8b572e80c653ca992add
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:2fe9473cd1d6c445d2ca7cc357ee4bb009aa1838f9cb69ea418296b2a044816b
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:d7c4ac937add511460a36375d9b6b3c475a3a9a7cc96c00df2fbad6219aee0e9
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f9965177de24f3165ec4c001c14cf1e2c2c01ffe847839af837c442e3c96beae
        version: 2.22.2
      entries:
      - name: rhods-operator.2.22.2
        version: 2.22.2
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable
    - currentCSV: rhods-operator.2.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:8643f7ee8afacfe6ae08cbbf81da76ca97032dc3ce3da4fddf16e330e9e33a08
          createdAt: "2024-11-14T21:18:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.10.0 <2.10.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:a1d590d0e2b3fae9ecbd843a664dab083061e797f66a9f0a021b0c2c487d7e45
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:0af6a91553f40ea40e86d85edc204658af3e5a45e8ab94cdb6b767e545df59b9
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:214ca0608fcbc0fd585bf01768d28c7ba17082d1043349ec8445afd663095383
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:d2514bf4b4b2d9abc19facac3f1a55405eb3ad0d9e9203e8816bfcfa08c9a6c5
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:4876f14ca9cff0e2b998a56f2f56d0775edf3963de610f4b810b4d94fb12d719
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:24238abd46633709feb75f2a3e41cf6c9e8528d2c0e4be2fac6f56de44e3fb9b
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:d9b55e8f25b81050380b4aab040243dc18a7432c48912c9b660c97dcc5eea008
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:8643f7ee8afacfe6ae08cbbf81da76ca97032dc3ce3da4fddf16e330e9e33a08
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:9f7b1cb8623de8b0550b0ba124d3b50f5e960ed011b33a3262c6346dbe2dff95
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:807ec05e9e107afc969e718d0c2c980959e2fd462e777a5ff63e55156ea150a4
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:fdc2bc06b0dfa79ebc6451ac3793b5e8d9f165171633831dafe46ebc18c10c35
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:4da77afafde99b15fd357a4ab6609644bf6330a1d9c0c03e64726ed1b9633230
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:cba9f25ac6bb2734bb990a622b515c838b881edb688bf44ff196ebe7684ab74b
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:8911afbb65c8c4f10c3597624485be54b9103e87a5020b718b35ec3df2c9bf00
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:d709f65e1194bc877fa4f9ec681b598467703d94139d57b58c24490d3a59fb82
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:220cad7a22d40998063c6b4fb2f63e69330e64ca7118490a6680b05fc99b7733
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:8fed8f58732bff512289fff738d0a94be67286c41b24eaf5928bbcb30d88c88a
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:232ebac25084db056fdde49477b5c1e6a03a0228fea3758bcf6234b3d0eaa582
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:2007da5abf840712cd2edb50462b2d8d571a0193c3ae1550d4a279eddf2b335f
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:a4b9553fde02d9c67a9c98f063eaf82fb1d2689f4b1269e57bf5f12c8d4adecf
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:5d15175f934017c04d6a6057200ff55b2172a1f4bcbeb0f2cd9e9e545631d93f
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:f4d7d3ba4a0d3322bc9f327e095e9ec56073758f6dcd8da8c7bb185cf952ead5
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:7df2f8bc2c5c0f4f347d009b62a058b5107413a2abd789f6010eb4d6c9e4f8ee
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:a9c6e3afdf605465766fa93334cc0f6bb5e46a5fdfd7ec64d648e2ed72b05ccf
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:267d6795bf53f9a1955d73445776a428dd042af0cd3cb8bb5b91692a1bc1e5d2
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:65db5e8cb2fa46d020ed4837d4479d133b52167c373e65f81cb590eec5391af2
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:0ff5664c2eaadf3d3073b3d7fb711da47fb78d21f67710d8b8aad5452cb0355d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:12741379ab24d2286b5fae77f4383fb65713af7b3356fb43d8254ea85d4b0ee3
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:52e8deeeeaa21b17370a5d5cd644745da745b6e1cee8c6295d6df75c8d2aec04
        version: 2.10.2
      entries:
      - name: rhods-operator.2.10.2
        version: 2.10.2
      - name: rhods-operator.2.10.1
        version: 2.10.1
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.10
    - currentCSV: rhods-operator.2.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:28b5fff602ad6fb36d462e1905f453509ec22fc251a1e0cf64fedba86ff420d7
          createdAt: "2024-10-17T10:53:18Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.13.0 <2.13.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:cd6b535d95eb1b44fe261a5336270fd898b966f431c14e173944484381c26d25
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:c2bd493e9345eccd124bbe0440e45cce8cf57588827d8eb9533f8bbadaa3ebbe
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:74e6cfa0bc481588cb8dc61c206ad567fbd2b28f43c2a86bee4a8e0eff55447c
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:28b5fff602ad6fb36d462e1905f453509ec22fc251a1e0cf64fedba86ff420d7
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:77bca19180f30ef08f4d1598ac6c365e190bb7909ac4a1fb35d9ba9ba8d6ff7a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a37529eff3656a570d796083404e5a3f63603afecfe9e738da1a243c0cb3d7c2
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:5d6354e65106460ab9285b42124faf288298387a8b16317565bd591f009c39a6
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:fb630d6f5817512733b8ea05f7147121928917a24ac8d98239f43d920880c359
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:6b2c0d1dd40e2c276684e17818b14375bc778e6c88df0bdc3cc24bfc9294804f
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:8883bf71abf39b4a6985940c287eb6ac7dc137d491264805d48bb112b65bd008
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:9ccdd7fe8223057a2234a3a2eb3b492e7b69e6ce714cc26a805fe2dd82935195
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:850901511bde065cf73e4c18d787dc10b6e91ca6139fc3ba68da28c41dd28236
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:688389195aebf3f43e74e40d591d2fd5f9ae41994d8ac2db661f292f6eda012a
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:2566b73e98b783871fb500bb1a8cc4e3d58a513c1bb18abfb525da0818e6138b
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:fd4ce524acc1769475079c7231f5ea79ac29551c513855ef34e5694d2c9fbbd3
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:a55f73d72ad2b07f01c49949cdb00969bfe4789375b3d7cdf9814b969fcc808d
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:7d085942e4da73c2ff4885c62f10b904dc70fd08e819e62f925eb1b4ddfa1554
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:8faab013568ee023968a5ec5b3a79977d7ddae3deafff6328e6fd56a27198e24
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:71abac4fb3d613edef7f3ba117a73bbeea61f341d6182e8310360a76576c789e
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:030c53690fe25e15bc9907f51c3f04fab39b4d085ff2098d26316ba32cc8b09f
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:8c19c7fa535f36bdce3b73cf3300e321909d44954cbc63a8aef6321c5f2fa693
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:59e9cc9cd57ba48ca7e87ae1e797a902f826bb890cb198ccf549f76ea581ba6e
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:bf25fce160dd7b8cce11e68a501006e4fa81a7858645038cdefdbba03df7ecdb
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:a3048eaa8924ce3571c6e78014438183fae70d9e8c833a7c07774c675104e3d4
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:5c2ee4136ee0014d1ef395b5e3bcab125c846f5e258ece3631d912f431c8d208
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:7c895490d49beb6bd724f47ad1b17ce501650557b0889314799c4072632f2b99
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f5b17f44c5ab72d6cf658f72f24f15e06a5ac9be1790336f7228ab8e7cfc15c3
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:cf3779c64bbaef1988886ed629c516e7d6ef7bdc73b9c95ad4c087b62e75750d
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:363a35c332fc80b068157951d5e50e6b7209947649cc64803ffc02726e7581ab
        version: 2.13.1
      entries:
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.13
    - currentCSV: rhods-operator.2.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
          createdAt: "2025-09-24T13:27:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.16
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:97e2bd9b587f08e135a9aeb9b3e0dc6eafa1a9bdacbb5ecb681ce9bd5aa37fc9
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:efd039012559786deb3c521a130886c265e88d635e08baace2e573a9df270134
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:d7d4fa406e0fcf0507894a7676532b27f45be742467e603a86f98ea5d2615df8
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:7b70af8847b0806d8b43c399e2b3109f016fd864f5e9d30c44e2baca5d1359dc
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f37e4048f3a152798286793f3abfc6ed814453fcbe2667255a7e78eee483c5a3
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:4a7599f8a866eb169c9a62885906adbf6df0417c0be15857df1eef20cd9b1be2
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:4c5ff3496b2a2a739939d94ee9dafc02b682100785d228dc2fde480fb597b7a5
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:4a212ec634225c14beac09be24ddf336e562f2aa9a13555fb1196f366ddae23c
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:2ba2b2c4db8bb334c50f4dfb54059f060361186900a44c06eed00b7a3c43977e
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:f738aab1eab25854c93e1d8d4d98100a8ae7bb45a6b83f0326774e4220b1183b
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:23d307a36b69e0df04f72a7d3b35e28d8417a8bbe23dba31e8e977569785c078
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:04f305c44413ae7dcb7017e53570ee49a509701792c5f50efadd64f47395730b
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:2eedd1e2465d2e78fafb39e234dbb413a82713f187c64d70a80340ec94807d30
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:5dcdcc2424602a69451f16d31dbfa1d43cb72c095ba561eb9076f0cd1e8182ed
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:11a301728573adf2b64ea72d0cb2d83ea5d4dbebea759f346e99f18c3d368c6e
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:27b113d96453c2054d4c965a963d2badba6daac235eddaf23234c76e87bcb069
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:c8456d98e90b6505957ab3686e9fd2f156e29f123c5558e581c206daf1e7d93a
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:de5f91180ead7d73a1825fe8b032fde9b8b01392569c9789f5ce1b4b9c08a98f
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a0f72ffefb2b74b488dd949493f5d295a39bb9c97f578bf219d0138601f65468
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:6cf74044ae8d5308a2dfe03fa5d81086c89302db7f4cdbe2f4174a1c48b77869
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:633984b538f027c93c5886ea1045dcbe81eeda74acf80001d5fc5f765bdbe0be
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:5192f1dbd2a9ab92ae390c4ae506efbed0970545b6122e95b014728ac937e777
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:65482864055021272a18b5b26792ab00cfa5fc9cc005d8d3a884cc82991506f3
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:22557a6fa52d2f311750a9ba253860f423ba697d26efa02ef8524a8258d2a909
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:ee01e89f98feb185f6cd59c564e590a13e4d8d9ea760cca8de51426eb71b83a1
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        version: 2.16.3
      entries:
      - name: rhods-operator.2.16.3
        version: 2.16.3
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.16
    - currentCSV: rhods-operator.2.19.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:c605e5fdb3d66302781146347df6da5a91dee97e4f8f5d9ac33f105b8cdee383
          createdAt: "2025-09-24T13:27:42Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:b30d60cd458133430d4c92bf84911e03cecd02f60e88a58d1c6c003543cf833a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:f0ee0d27bb784e289f7d88cc8ee0e085ca70e88a5d126562105542f259a1ac01
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:8ce44de8c683f198bf24ba36cd17e89708153d11f5b42c0a27e77f8fdb233551
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.19
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:8a6732ac3ca60f13dcb1fd8f186d2fc5dc601b9b5981a58f039af9e237f20044
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:b20ee16b776871d7c49b23f2bc91726812998561129f5dbd14eddbfe57964c93
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:c605e5fdb3d66302781146347df6da5a91dee97e4f8f5d9ac33f105b8cdee383
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:b30d60cd458133430d4c92bf84911e03cecd02f60e88a58d1c6c003543cf833a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:f49483df1d342d8e39f9a8085920aad79313a475df02a2292a136ecb784402b5
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:c8c23c07038f59db848dea0d4247b62840dbe58e1c581dc0722fb9d7f9be3b75
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:4e8f6c6ca0ed516b8310af0332637683b5bfc8f2946eb935c647698dffbbcc79
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhel8/mariadb-103@sha256:f0ee0d27bb784e289f7d88cc8ee0e085ca70e88a5d126562105542f259a1ac01
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:d24e277f468b0a48f38801e8e0bcfeb8264c3f7c128073ddaa17934b8572cc3c
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:520ed21b80e5487b3a0a13b7e5c7d791d4dd28e0b6799d57abb65d53fca5a03f
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:6275a19ef06114b33d79833a94ef10cc8203368eee98e5d415e73d20cae1db85
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:87685c5cab105d876ff0d14b0e24d8dc2a1ff5b55bebe36028c5f3376b2a5271
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:6ad35cf5c4c62ed5ad4e192afe8cebead9796cb24a964401c76ff8656c47742f
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:b6039c3c33a5dc345ceb8e757437bb13023d2bf85cc2b4c10b9ab3887e78e0c0
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:6f01ac7cc62ab961b9c8a630a558338a61f6d7902301c4a1ad337495ac7ce85a
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:4313cae03c041d461bc276abc23d706b582b5d16feab316796ca43478b9dc705
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:b24c25dfdcefb4ec26e7f311c00798df108f7685def2b1762b7624fdfb160ebf
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:bbef4839f24facce111d21aca43898dcf4e6cf57adb16200e3f1590bd5f89eae
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:49c02285fcef6949892cecca20eeee49d03a9bb4d17bb9ceff7c7a0dbbb8ab40
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel8@sha256:6166f525948a018c41d735014d97fdfa52024f7be31e18976d85d1a0d88901a0
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:baefbe4e7dd5aa74c71cb81da4a7a2b888a02d3eb1398ec4419bd8003b77f0c2
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:8ce44de8c683f198bf24ba36cd17e89708153d11f5b42c0a27e77f8fdb233551
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:af4721d61dfb61c75488a1bd6a5f62b23e20d7b7586a9cf51ba062ece2837715
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:bf6231bfdbe52b597d86e03e4730ec105ebe7a639fcd529268b7c0423f2a890c
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:71cd15f0b1acba3a48ebedff037bf8a846b426c18363a4de8896a376b6b7b3d5
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:53ac36baa374159b9065c718a9ede821bbb61d9ebe9502b2243e0a9f7aca0d16
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:d47514cbceb3cc96e9d33b7db6fc1df60893e5c4118bf6086cdd27f04c472773
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:876ca7ce7b0b975688fb4a77d6cd1a6f87cf9eb40c2a9a5b84d8cffdc80eeebd
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:de0e4b7fb822101198817e83cd7e9f4f7406e75ec91d07a833a27a417cffe66b
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:c1c97b6beb759d3f6f2dd53c4258e4ccbb3bf35cf11873bd446b135defaa69d6
        version: 2.19.3
      entries:
      - name: rhods-operator.2.19.3
        version: 2.19.3
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.19
    - currentCSV: rhods-operator.2.22.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
          createdAt: "2025-09-25T11:28:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.22
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:13f3679ab289eea9de54fe129a7c83f6c8ec5d26b61b430637a7c4cfcd31a74a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:d59ed40de694bc0e0564bab159a5b586cffa0721b6d761fb2d81dd42e6225d1c
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:d7c4ac937add511460a36375d9b6b3c475a3a9a7cc96c00df2fbad6219aee0e9
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:b2ca229bb4c4dbf04b96766b8af8d9f55f2edef59ec4110de0d20f8c857e7d27
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:9104cbc2d595933f0a2acb7313e41ed69905e76a9fae6bd9820e7717b56a632b
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f9965177de24f3165ec4c001c14cf1e2c2c01ffe847839af837c442e3c96beae
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:2f3f305c97e50057781553416fb4026cde306ead0fbe0213627b9b5bbcc8a0c3
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:b2c2ea9d35edbcd16e1ae559c88a453129b877fc049ab7a7cd6c2ebb8f0bcb59
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:6234e15f001122ab222481321153fb0eba7f4d86bf3dfd64edf3f6cecbe7d62f
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:5d79f0bcca0a213b53aee6d88834e503d343821596b857b13d3a6362de8ebeab
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:a44d32a18b1e906c1411cafc82102881573fcd04e040a4e2ba6bd1bd617b70fa
        - registry.redhat.io/rhel9/mariadb-105@sha256:5e77b33ce04c1eafc2e8840586ad6a5b51148dea8a46dda151071e28858331f0
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:7fd540586a1ca111370100a473923b3dc6cec4255e9e4f3654bfbabd9158a2bb
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:2fe9473cd1d6c445d2ca7cc357ee4bb009aa1838f9cb69ea418296b2a044816b
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:27415b1c763cf89b79391f0b039936c3e078a078bda4420f13bf10ef9be31d5e
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:7b3d7787ba95156de9c284ad78f2d63a0d3bcb525f4d8d11ebf0203d4c5c1fe0
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:2065b96bd6f0a830eab66e36217a927448cfc5a7dabf2a5e7d907aa6c40657a4
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e3b90530c6e38f6940d7f9408e9fa2c81eb06a857ec2b0bab5d006b1e9a63a4f
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:e6fbe6723b4801f6854303b3fd7d4a4a36727bdd28dc28002b7de4929739ccc0
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/ubi9/toolbox@sha256:ea4ce8cb8f59cf1b0709f1606d7cab074c3806d4016964eedc0555410cb5dfeb
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:5914f71e37c815ce2567826896e051303bdef52fae1acf8d02e672fc24c7eb9d
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:1709fa3c79aad4ba7eb9be8299949396092c8e20210124e0c0936385bc04e839
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:cb5be3dd7098f4c8d66f9e85ad04656460df6102652deddc6cccb884a12be733
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:2b3aace0d80474470e71c2c8291b25738e096109229fe6d5d3e06e9885ee58b6
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:2f1a0a0803b2f5118b167e9d9c2e86660add5f79ee6f47044f63b0e8abe0b75f
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:ccba67a6a5aa3403e0018f29dcdfef483220dea8065d8b572e80c653ca992add
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:463e9be9c4409ba1e5a42713bd65844a023e7a3e9032d4e862edbaa5fe35300a
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:abe1f9dd11c7313fdd55b678be511a5cdc3546654a68e2b53d42c1013fec8d00
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:bd6d1599b4ff9eb890be9808014715b28ca19247222d65f40df024db4dd3525a
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:74d4e0174655cf6f6273857325a1e714a1531c5a06ad38bf139bee9c7d8f5f19
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:32c9511846077ef6035eb19be7653f6a78e0f207eb420ab93b18f742ab5779b6
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        version: 2.22.2
      entries:
      - name: rhods-operator.2.22.2
        version: 2.22.2
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.22
    - currentCSV: rhods-operator.2.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:b3563e201bf2720f60e73c0b04cb70566ea52eda7d09a28753f3a1487993c2e8
          createdAt: "2024-08-01T06:23:03Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.8.0 <2.8.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:eb4da3a645b53acfd6ce8758c043ae19b4126e62bcca09165114a141c2c8d014
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:92d9e4b53d580df6cc1f14a849b477749a1306941a2e8f9d1b6160f4911c3cff
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:b3563e201bf2720f60e73c0b04cb70566ea52eda7d09a28753f3a1487993c2e8
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:5f39b11f0cd6db2d7d6e51ee666c030e2643aee92a090b49656c1e7dfd8a5430
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:90833e85e0de3d7de870074c8b04319cd11b913b9652af6a5727aae41a91458c
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:91525fac5fadece75fbdafc40b94445fb92c88cfc8c2f06b596f8c6377e59c9e
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:ab8792fcd16b1aa0065444ba54335cb3e747379740c3ec2d56beb46ff80cb044
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:987c768c0688431a53e48a80591f2a41e074994169f9e691b28b89528d8e736f
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:1a6e29214e06a2665baadbed598ba758d6141bfa522984eb05f05a80558a403c
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:850b78a201effcafe33705dcd2670f1c470d8b46ba6e307032403627b55a7cfe
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:080377d42d782bf525b4072384d0cf8409b48019b3748ca7c855f340bcb94251
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:b6ecb3858c201498024e3bfc4a93e630699be4b1fdc3a65fd2d70e4b375eb9c2
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:df2edcaf4402b11ce24439263956e3b0cc43efbac2fdf152afacb1a398dcc59a
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:889f046b255b13d99801108588d3abf577611b828657b5240a18ea21f0e3cbbe
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:1c49888ba2e9449c04d5588040c099290fce4793c2ba8874c1037316e550d6e1
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:0b8665e8978deb73be2cc65830f08a74a5a48f3e3ee7b2b817f7bbacba257d22
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:68e1331f13dfaf6641a1e7d54f4cf29c0ef42663abc98b6dc69167892d2d8adb
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:06a6d5cbbe80d49bca9e4db9929571a69e35ae7951fcae84dcd6460f7ddf622c
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:9799003449bfb61ae8f59efd2a611b0da3bccadd9b20cc1519f6c9279d085337
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:9c90873fbd5af8fac8efc3c7c27fba742a06fad2d47f336319dd05d2b116867b
        version: 2.8.4
      entries:
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.8
    defaultChannel: stable
    packageName: rhods-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: ansible-automation-platform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aap-operator.v2.4.0-0.1758079821
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.1-0.1635279521 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the hub, including all hosted Ansible
              content, secrets and the database. By default, a persistent volume claim
              will be created using the default StorageClass on your cluster to store
              the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous hub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Automation Hub
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-24

          ## How to Upgrade

          Starting with AAP 2.4 in the May 2025 async release, the managed database was upgraded to PostgreSQL 15. It is recommended to take a backup of your deployment before upgrading.

          To take a backup, create AutomationControllerBackup and AutomationHubBackup objects for your deployments.
          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel.

          Note: Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8-operator@sha256:5b957031609c8aabf7575e675bc636c1152babc9564d503802301aff4642a404
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-runner-rhel8@sha256:e14286dc9605be3a8c6cb144d0b2e3f3bb3211129005009e7566ea6c16c91f52
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8@sha256:7ae8db2b5d12b619f40db1e581db1e982a695bc1a29a2628222fff09e7c25249
        - registry.redhat.io/ansible-automation-platform-24/ee-minimal-rhel8@sha256:7670e41037f023e3fc6ee95d2342787a80077273ab8d30ba6804029ff05b4edc
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-ui-rhel8@sha256:d229d4e09b2b788d1ba3413ea7ec15374bd1613d507c285ac1ddc10570b9473f
        - registry.redhat.io/ansible-automation-platform-24/hub-web-rhel8@sha256:6ff1f602f0d9cad201b3addb65acc570a9d527de005ca85fc861d3cb98138710
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8@sha256:bfd13d6b39c5ec3a67adbafb05a0482328f0b9b97a517b45c093f6c35a3a92d3
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8@sha256:04fb5cb6b0d7c79cf5d4b754fe917224ddfcb2b9a4994462db404c833ed69483
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-rhel8-operator@sha256:b120e6a5ef376662b4324f0458eae31627087dc91fc613788143bc3954a15318
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-24/ee-supported-rhel8@sha256:f1384eb558f7da1e4a35d51408b3cf5a5a9be08f4d50503f473b26909bc246fc
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8@sha256:cccb4db08735551c697a7b2cbf75812f34500e691f9cdbaea02b30edea292e47
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8-operator@sha256:5253c9fd587ff2ca94d251f2b5fc7f284080adc5eb64643df5299beb2015592f
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8-operator@sha256:3c28a1326ecb7495952a5bb8be13b0972a4378af82611b0b9d990c5fa9fa82c3
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8-operator@sha256:e20106a8a04abd568a4c56861cea39ae444c0bd204bc807ac01f1705e12937d3
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        version: 2.4.0+0.1758079821
      entries:
      - name: aap-operator.v2.4.0-0.1702352195
        version: 2.4.0+0.1702352195
      - name: aap-operator.v2.4.0-0.1707506910
        version: 2.4.0+0.1707506910
      - name: aap-operator.v2.4.0-0.1708974895
        version: 2.4.0+0.1708974895
      - name: aap-operator.v2.4.0-0.1711590540
        version: 2.4.0+0.1711590540
      - name: aap-operator.v2.4.0-0.1713466670
        version: 2.4.0+0.1713466670
      - name: aap-operator.v2.4.0-0.1715898994
        version: 2.4.0+0.1715898994
      - name: aap-operator.v2.4.0-0.1717558474
        version: 2.4.0+0.1717558474
      - name: aap-operator.v2.4.0-0.1718152210
        version: 2.4.0+0.1718152210
      - name: aap-operator.v2.4.0-0.1720482179
        version: 2.4.0+0.1720482179
      - name: aap-operator.v2.4.0-0.1725256739
        version: 2.4.0+0.1725256739
      - name: aap-operator.v2.4.0-0.1726111413
        version: 2.4.0+0.1726111413
      - name: aap-operator.v2.4.0-0.1727257740
        version: 2.4.0+0.1727257740
      - name: aap-operator.v2.4.0-0.1730152858
        version: 2.4.0+0.1730152858
      - name: aap-operator.v2.4.0-0.1733185647
        version: 2.4.0+0.1733185647
      - name: aap-operator.v2.4.0-0.1733943951
        version: 2.4.0+0.1733943951
      - name: aap-operator.v2.4.0-0.1737661496
        version: 2.4.0+0.1737661496
      - name: aap-operator.v2.4.0-0.1741194076
        version: 2.4.0+0.1741194076
      - name: aap-operator.v2.4.0-0.1742444794
        version: 2.4.0+0.1742444794
      - name: aap-operator.v2.4.0-0.1743629049
        version: 2.4.0+0.1743629049
      - name: aap-operator.v2.4.0-0.1746132523
        version: 2.4.0+0.1746132523
      - name: aap-operator.v2.4.0-0.1749069319
        version: 2.4.0+0.1749069319
      - name: aap-operator.v2.4.0-0.1750877268
        version: 2.4.0+0.1750877268
      - name: aap-operator.v2.4.0-0.1753232084
        version: 2.4.0+0.1753232084
      - name: aap-operator.v2.4.0-0.1755833431
        version: 2.4.0+0.1755833431
      - name: aap-operator.v2.4.0-0.1758079821
        version: 2.4.0+0.1758079821
      name: stable-2.4
    - currentCSV: aap-operator.v2.4.0-0.1758080293
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.1-0.1635283332 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the hub, including all hosted Ansible
              content, secrets and the database. By default, a persistent volume claim
              will be created using the default StorageClass on your cluster to store
              the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous hub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Automation Hub
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-24

          ## How to Upgrade

          Starting with AAP 2.4 in the May 2025 async release, the managed database was upgraded to PostgreSQL 15. It is recommended to take a backup of your deployment before upgrading.

          To take a backup, create AutomationControllerBackup and AutomationHubBackup objects for your deployments.
          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel.

          Note: Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8@sha256:bfd13d6b39c5ec3a67adbafb05a0482328f0b9b97a517b45c093f6c35a3a92d3
        - registry.redhat.io/ansible-automation-platform-24/ee-minimal-rhel8@sha256:7670e41037f023e3fc6ee95d2342787a80077273ab8d30ba6804029ff05b4edc
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8@sha256:04fb5cb6b0d7c79cf5d4b754fe917224ddfcb2b9a4994462db404c833ed69483
        - registry.redhat.io/ansible-automation-platform-24/ee-supported-rhel8@sha256:f1384eb558f7da1e4a35d51408b3cf5a5a9be08f4d50503f473b26909bc246fc
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-rhel8-operator@sha256:b120e6a5ef376662b4324f0458eae31627087dc91fc613788143bc3954a15318
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8-operator@sha256:5253c9fd587ff2ca94d251f2b5fc7f284080adc5eb64643df5299beb2015592f
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8-operator@sha256:3c28a1326ecb7495952a5bb8be13b0972a4378af82611b0b9d990c5fa9fa82c3
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8@sha256:7ae8db2b5d12b619f40db1e581db1e982a695bc1a29a2628222fff09e7c25249
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8@sha256:cccb4db08735551c697a7b2cbf75812f34500e691f9cdbaea02b30edea292e47
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8-operator@sha256:5b957031609c8aabf7575e675bc636c1152babc9564d503802301aff4642a404
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8-operator@sha256:e20106a8a04abd568a4c56861cea39ae444c0bd204bc807ac01f1705e12937d3
        - registry.redhat.io/ansible-automation-platform-24/hub-web-rhel8@sha256:6ff1f602f0d9cad201b3addb65acc570a9d527de005ca85fc861d3cb98138710
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-runner-rhel8@sha256:e14286dc9605be3a8c6cb144d0b2e3f3bb3211129005009e7566ea6c16c91f52
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-ui-rhel8@sha256:d229d4e09b2b788d1ba3413ea7ec15374bd1613d507c285ac1ddc10570b9473f
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        version: 2.4.0+0.1758080293
      entries:
      - name: aap-operator.v2.4.0-0.1702352614
        version: 2.4.0+0.1702352614
      - name: aap-operator.v2.4.0-0.1707507304
        version: 2.4.0+0.1707507304
      - name: aap-operator.v2.4.0-0.1708975279
        version: 2.4.0+0.1708975279
      - name: aap-operator.v2.4.0-0.1711591330
        version: 2.4.0+0.1711591330
      - name: aap-operator.v2.4.0-0.1713467078
        version: 2.4.0+0.1713467078
      - name: aap-operator.v2.4.0-0.1715899379
        version: 2.4.0+0.1715899379
      - name: aap-operator.v2.4.0-0.1717558885
        version: 2.4.0+0.1717558885
      - name: aap-operator.v2.4.0-0.1718152680
        version: 2.4.0+0.1718152680
      - name: aap-operator.v2.4.0-0.1720482740
        version: 2.4.0+0.1720482740
      - name: aap-operator.v2.4.0-0.1725257213
        version: 2.4.0+0.1725257213
      - name: aap-operator.v2.4.0-0.1726111804
        version: 2.4.0+0.1726111804
      - name: aap-operator.v2.4.0-0.1727259094
        version: 2.4.0+0.1727259094
      - name: aap-operator.v2.4.0-0.1730153340
        version: 2.4.0+0.1730153340
      - name: aap-operator.v2.4.0-0.1733186325
        version: 2.4.0+0.1733186325
      - name: aap-operator.v2.4.0-0.1733945743
        version: 2.4.0+0.1733945743
      - name: aap-operator.v2.4.0-0.1737662040
        version: 2.4.0+0.1737662040
      - name: aap-operator.v2.4.0-0.1741194557
        version: 2.4.0+0.1741194557
      - name: aap-operator.v2.4.0-0.1742445409
        version: 2.4.0+0.1742445409
      - name: aap-operator.v2.4.0-0.1743630487
        version: 2.4.0+0.1743630487
      - name: aap-operator.v2.4.0-0.1746133932
        version: 2.4.0+0.1746133932
      - name: aap-operator.v2.4.0-0.1749069693
        version: 2.4.0+0.1749069693
      - name: aap-operator.v2.4.0-0.1750878254
        version: 2.4.0+0.1750878254
      - name: aap-operator.v2.4.0-0.1753232791
        version: 2.4.0+0.1753232791
      - name: aap-operator.v2.4.0-0.1755833968
        version: 2.4.0+0.1755833968
      - name: aap-operator.v2.4.0-0.1758080293
        version: 2.4.0+0.1758080293
      name: stable-2.4-cluster-scoped
    - currentCSV: aap-operator.v2.5.0-0.1758147230
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727257740 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation gateway - Unify the management of AAP from a central location.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner


          ## New in AAP 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-25

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.5

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console. For example, change the channel from `stable-2.4` to `stable-2.5`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.

          After changing to the new 2.5 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place.

          #### Create the AnsibleAutomationPlatform custom resource

          Once that is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: myaap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8@sha256:3a989f148648a5a62e687058ec3ca772fa22aa10faec89453626d2b17b3fd9c2
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8-operator@sha256:43e731b9a435b3a412108170b4b39e6b6215be51d362a9f03c33fa412e97da53
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8-operator@sha256:8d8182e3a87c55eacc222e71df4a9bf7a2e51102d913a5824cc91e36056d4249
        - registry.redhat.io/ansible-automation-platform-25/hub-web-rhel8@sha256:e96832efcc03c6754031cb70f56c808135f62d1b80a882dd6d6fc9e1a4cfc69b
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8@sha256:697343aaea153ab22d7efa8f1bb5aed37a55cbb353fc206f4a355e8b878ae418
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8@sha256:ac0de2b7d354eb0c15a6c12861a4de3577a982c84af661fbab25517d8105c085
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8@sha256:aed3973eb953dda3571260dcf1b33bc7dcbc18c41e3350b8c361b7e635fb9527
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-runner-rhel8@sha256:ace1d2ab49ba5ff9f11ab18e8edce27c9e73ae7298bf7335569fa1740a44899b
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8-operator@sha256:acc0fcd509eeac46db307ab744dd0a0c26770a93e495cf2ec0ccb61a91c6fbc5
        - registry.redhat.io/ansible-automation-platform-25/ee-minimal-rhel8@sha256:41f944eb1450db56c3b14fc7baece8f148d5f10886a57bad730bec60fb28ac04
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8-operator@sha256:07aa7e257d04ae0eda6fcaa4af9121f1aece439231fce4f924e6467da91f0022
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-ui-rhel8@sha256:9ab35e80a6172018ed3f5cd179e0bcb6befc1378583d9e73c3a64a7cf9e8050f
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8@sha256:9f2867603b9747c2fe37862c2c5efac4a4530902f323a8bf7cc0717de4cdb5aa
        - registry.redhat.io/ansible-automation-platform-25/gateway-proxy-rhel8@sha256:789d55f355fa0d90f9509a912d199123bfc5a4041930a72a8fd4ed1d819f0d84
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8-operator@sha256:503fc7a20e2dd7de2cb55b9be5f97efa0d2b31f054ce326f23f3a11b4d683e64
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-rhel8-operator@sha256:74be9182723b41f87b3d13b3d86b6a9049bba88e2902cc9453264aa8a0f584e4
        - registry.redhat.io/ansible-automation-platform-25/ee-supported-rhel8@sha256:52d2c73fa2015af5f1b9a88ee48085c7c7bab44330fc9f95914f48b58e4b48bc
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-chatbot-rhel8@sha256:9a37c9739c488ccf211ec5efc6debb045223d508c9cb3fa5a8256eca98ff2f0c
        version: 2.5.0+0.1758147230
      entries:
      - name: aap-operator.v2.5.0-0.1727240614
        version: 2.5.0+0.1727240614
      - name: aap-operator.v2.5.0-0.1727875185
        version: 2.5.0+0.1727875185
      - name: aap-operator.v2.5.0-0.1728520175
        version: 2.5.0+0.1728520175
      - name: aap-operator.v2.5.0-0.1729741547
        version: 2.5.0+0.1729741547
      - name: aap-operator.v2.5.0-0.1731456196
        version: 2.5.0+0.1731456196
      - name: aap-operator.v2.5.0-0.1733193164
        version: 2.5.0+0.1733193164
      - name: aap-operator.v2.5.0-0.1733969526
        version: 2.5.0+0.1733969526
      - name: aap-operator.v2.5.0-0.1736893446
        version: 2.5.0+0.1736893446
      - name: aap-operator.v2.5.0-0.1737101025
        version: 2.5.0+0.1737101025
      - name: aap-operator.v2.5.0-0.1737675968
        version: 2.5.0+0.1737675968
      - name: aap-operator.v2.5.0-0.1739248042
        version: 2.5.0+0.1739248042
      - name: aap-operator.v2.5.0-0.1740093573
        version: 2.5.0+0.1740093573
      - name: aap-operator.v2.5.0-0.1740773472
        version: 2.5.0+0.1740773472
      - name: aap-operator.v2.5.0-0.1741368605
        version: 2.5.0+0.1741368605
      - name: aap-operator.v2.5.0-0.1742434024
        version: 2.5.0+0.1742434024
      - name: aap-operator.v2.5.0-0.1743660124
        version: 2.5.0+0.1743660124
      - name: aap-operator.v2.5.0-0.1746137767
        version: 2.5.0+0.1746137767
      - name: aap-operator.v2.5.0-0.1747343762
        version: 2.5.0+0.1747343762
      - name: aap-operator.v2.5.0-0.1749074128
        version: 2.5.0+0.1749074128
      - name: aap-operator.v2.5.0-0.1749604727
        version: 2.5.0+0.1749604727
      - name: aap-operator.v2.5.0-0.1750901111
        version: 2.5.0+0.1750901111
      - name: aap-operator.v2.5.0-0.1753900156
        version: 2.5.0+0.1753900156
      - name: aap-operator.v2.5.0-0.1755835086
        version: 2.5.0+0.1755835086
      - name: aap-operator.v2.5.0-0.1758147230
        version: 2.5.0+0.1758147230
      name: stable-2.5
    - currentCSV: aap-operator.v2.5.0-0.1758147817
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727259094 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation gateway - Unify the management of AAP from a central location.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner


          ## New in AAP 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-25

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.5

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console. For example, change the channel from `stable-2.4` to `stable-2.5`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.

          After changing to the new 2.5 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place.

          #### Create the AnsibleAutomationPlatform custom resource

          Once that is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: myaap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8-operator@sha256:acc0fcd509eeac46db307ab744dd0a0c26770a93e495cf2ec0ccb61a91c6fbc5
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-ui-rhel8@sha256:9ab35e80a6172018ed3f5cd179e0bcb6befc1378583d9e73c3a64a7cf9e8050f
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8-operator@sha256:43e731b9a435b3a412108170b4b39e6b6215be51d362a9f03c33fa412e97da53
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8-operator@sha256:8d8182e3a87c55eacc222e71df4a9bf7a2e51102d913a5824cc91e36056d4249
        - registry.redhat.io/ansible-automation-platform-25/hub-web-rhel8@sha256:e96832efcc03c6754031cb70f56c808135f62d1b80a882dd6d6fc9e1a4cfc69b
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-chatbot-rhel8@sha256:9a37c9739c488ccf211ec5efc6debb045223d508c9cb3fa5a8256eca98ff2f0c
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8@sha256:3a989f148648a5a62e687058ec3ca772fa22aa10faec89453626d2b17b3fd9c2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/ee-minimal-rhel8@sha256:41f944eb1450db56c3b14fc7baece8f148d5f10886a57bad730bec60fb28ac04
        - registry.redhat.io/ansible-automation-platform-25/gateway-proxy-rhel8@sha256:789d55f355fa0d90f9509a912d199123bfc5a4041930a72a8fd4ed1d819f0d84
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8@sha256:aed3973eb953dda3571260dcf1b33bc7dcbc18c41e3350b8c361b7e635fb9527
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8@sha256:697343aaea153ab22d7efa8f1bb5aed37a55cbb353fc206f4a355e8b878ae418
        - registry.redhat.io/ansible-automation-platform-25/ee-supported-rhel8@sha256:52d2c73fa2015af5f1b9a88ee48085c7c7bab44330fc9f95914f48b58e4b48bc
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8@sha256:ac0de2b7d354eb0c15a6c12861a4de3577a982c84af661fbab25517d8105c085
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-rhel8-operator@sha256:74be9182723b41f87b3d13b3d86b6a9049bba88e2902cc9453264aa8a0f584e4
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-runner-rhel8@sha256:ace1d2ab49ba5ff9f11ab18e8edce27c9e73ae7298bf7335569fa1740a44899b
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8@sha256:9f2867603b9747c2fe37862c2c5efac4a4530902f323a8bf7cc0717de4cdb5aa
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8-operator@sha256:07aa7e257d04ae0eda6fcaa4af9121f1aece439231fce4f924e6467da91f0022
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8-operator@sha256:503fc7a20e2dd7de2cb55b9be5f97efa0d2b31f054ce326f23f3a11b4d683e64
        version: 2.5.0+0.1758147817
      entries:
      - name: aap-operator.v2.5.0-0.1727241223
        version: 2.5.0+0.1727241223
      - name: aap-operator.v2.5.0-0.1727875896
        version: 2.5.0+0.1727875896
      - name: aap-operator.v2.5.0-0.1728520687
        version: 2.5.0+0.1728520687
      - name: aap-operator.v2.5.0-0.1729742145
        version: 2.5.0+0.1729742145
      - name: aap-operator.v2.5.0-0.1731456854
        version: 2.5.0+0.1731456854
      - name: aap-operator.v2.5.0-0.1733193761
        version: 2.5.0+0.1733193761
      - name: aap-operator.v2.5.0-0.1733970210
        version: 2.5.0+0.1733970210
      - name: aap-operator.v2.5.0-0.1736894220
        version: 2.5.0+0.1736894220
      - name: aap-operator.v2.5.0-0.1737101781
        version: 2.5.0+0.1737101781
      - name: aap-operator.v2.5.0-0.1737676792
        version: 2.5.0+0.1737676792
      - name: aap-operator.v2.5.0-0.1739248649
        version: 2.5.0+0.1739248649
      - name: aap-operator.v2.5.0-0.1740094176
        version: 2.5.0+0.1740094176
      - name: aap-operator.v2.5.0-0.1740774104
        version: 2.5.0+0.1740774104
      - name: aap-operator.v2.5.0-0.1741369251
        version: 2.5.0+0.1741369251
      - name: aap-operator.v2.5.0-0.1742434756
        version: 2.5.0+0.1742434756
      - name: aap-operator.v2.5.0-0.1743660958
        version: 2.5.0+0.1743660958
      - name: aap-operator.v2.5.0-0.1746138413
        version: 2.5.0+0.1746138413
      - name: aap-operator.v2.5.0-0.1747345055
        version: 2.5.0+0.1747345055
      - name: aap-operator.v2.5.0-0.1749074612
        version: 2.5.0+0.1749074612
      - name: aap-operator.v2.5.0-0.1749607543
        version: 2.5.0+0.1749607543
      - name: aap-operator.v2.5.0-0.1750901870
        version: 2.5.0+0.1750901870
      - name: aap-operator.v2.5.0-0.1753900674
        version: 2.5.0+0.1753900674
      - name: aap-operator.v2.5.0-0.1755835623
        version: 2.5.0+0.1755835623
      - name: aap-operator.v2.5.0-0.1758147817
        version: 2.5.0+0.1758147817
      name: stable-2.5-cluster-scoped
    - currentCSV: aap-operator.v2.6.0-0.1758684783
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727257740 <2.6.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Automation Gateway - Unify the management of AAP from a central location.
            * Automation Controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private Automation Hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Ansible Lightspeed: Deploy and manage the lifecycle of an Ansible Lightspeed application.
            * Platform Resource Operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## New in AAP 2.6 and 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.


          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-26

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5 and above, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.6

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console.

          **For customers on AAP 2.4:** Change the channel directly from `stable-2.4` to `stable-2.6` to upgrade directly to AAP 2.6, which will be released in October.

          **For customers on AAP 2.5:** Change the channel from `stable-2.5` to `stable-2.6`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.

          After changing to the new 2.6 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place. When upgrading from AAP 2.4 to 2.6, your external authentication configuration in Automation Controller will automatically be migrated to the new unified Platform UI.

          **Database:** AAP 2.6 continues to use PostgreSQL 15.

          #### Create the AnsibleAutomationPlatform custom resource (Required for upgrades from 2.4 only)

          **If upgrading from AAP 2.4:** Once the operator upgrade is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created.

          **If upgrading from AAP 2.5:** This step is not required as your AnsibleAutomationPlatform custom resource already exists.

          Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          Next, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda (lightspeed is disabled by default):


          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: aap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-26/gateway-rhel9-operator@sha256:fa1b102a0690767a1e08935a0e164c1eabb2d4cf9afccd310262ee9185544de7
        - registry.redhat.io/ansible-automation-platform-26/hub-rhel9@sha256:8041b141cd8921aaaa20cce15eab4c6189da9f449d99057189284891c071ac3b
        - registry.redhat.io/ansible-automation-platform-26/gateway-proxy-rhel9@sha256:13284d20c6fb4f2de24ec1715e7abdb5c4b0e4298c5034a7795f494b93634a02
        - registry.redhat.io/rhel9/postgresql-15@sha256:5e225c1646251d2509b3a6e45fad23abfb11dd5c45d3224b59178f90deb38a70
        - registry.redhat.io/ansible-automation-platform-26/controller-rhel9-operator@sha256:863e77e52ab01cec2a71672a224080a59d963d4fc08b9ed6f3cc7950729b1746
        - registry.redhat.io/ansible-automation-platform-26/hub-rhel9-operator@sha256:18114aab6db0d00780fdeaf9efa74ee44cab94824c2fd35f4b9998b6378ad8b3
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-rhel9-operator@sha256:cfbb98e17297134c187673f355342ac5ad60a56d73f4b0c918be2ff244286cc0
        - registry.redhat.io/ansible-automation-platform-26/platform-resource-rhel9-operator@sha256:969bbd033fbba33624329ee9347b3b89fb5a1f64e71892f5c6c7d74ed14a6d39
        - registry.redhat.io/ansible-automation-platform-26/controller-rhel9@sha256:64f5e77a7842965e9c6f7bc7990354b71bb19ffb06ab8ef7793dde956cb6c0f0
        - registry.redhat.io/ansible-automation-platform-26/ee-minimal-rhel9@sha256:261c324cac081e47d559c5fccbd48e7ec8f5bf64e3f27287fbe4fb8aff6b7d99
        - registry.redhat.io/ansible-automation-platform-26/hub-web-rhel9@sha256:aa4041672542a7f828d8a51c79e9c4dcea2a185388e4f20f828104db37e03a9b
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-rhel9@sha256:01e56bee772d15b120d6e4b02a59024b21eb8b266127b1a9b7f2c535ecdb2855
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-ui-rhel9@sha256:5b5f38a5f76781a042eec9cbf70b8be026878e4fa61c7fb90d1b0ad64bd21592
        - registry.redhat.io/ubi9/ubi-minimal@sha256:7c5495d5fad59aaee12abc3cbbd2b283818ee1e814b00dbc7f25bf2d14fa4f0c
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-chatbot-rhel9@sha256:8901fba14b21569b02108d2ab9050d558dee29f8177fae9d1bfcd5de3f4ce7c1
        - registry.redhat.io/ansible-automation-platform-26/mcp-tools-rhel9@sha256:0c1f6f6958f2570c8cd073287555d28c70ef9dded702ec014d3ead8ce572463e
        - registry.redhat.io/ansible-automation-platform-26/platform-resource-runner-rhel9@sha256:9983043386df38a17ff6fffef64dc187691e0b1fefec01eef3af39aa08e532ef
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-rhel9-operator@sha256:bfd55bffd8239e224d4ebba500d8517e8e66915078d6b59244f5e910fc5e10c8
        - registry.redhat.io/ansible-automation-platform-26/gateway-rhel9@sha256:c85dbbf6beaa264afec480dc1f06576aada531cdfd15f247f4db37dc9702be33
        - registry.redhat.io/rhel9/redis-6@sha256:0962d18bb5292c0e524354f319b63287c2298a4a2fd34556c68fa285df381427
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-rhel9@sha256:c5d4e2e76944e0991729e5c3b99f43b7f29ea08ad112e5e6513feb8fb1d899a1
        - registry.redhat.io/ansible-automation-platform-26/ee-supported-rhel9@sha256:2534d07f1ae41af32adb681e283cbaea465cbe0f9d234ed5478a7bfad7676e7f
        version: 2.6.0+0.1758684783
      entries:
      - name: aap-operator.v2.6.0-0.1758684783
        version: 2.6.0+0.1758684783
      name: stable-2.6
    - currentCSV: aap-operator.v2.6.0-0.1758685241
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727259094 <2.6.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Automation Gateway - Unify the management of AAP from a central location.
            * Automation Controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private Automation Hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Ansible Lightspeed: Deploy and manage the lifecycle of an Ansible Lightspeed application.
            * Platform Resource Operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## New in AAP 2.6 and 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.


          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-26

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5 and above, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.6

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console.

          **For customers on AAP 2.4:** Change the channel directly from `stable-2.4` to `stable-2.6` to upgrade directly to AAP 2.6, which will be released in October.

          **For customers on AAP 2.5:** Change the channel from `stable-2.5` to `stable-2.6`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.

          After changing to the new 2.6 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place. When upgrading from AAP 2.4 to 2.6, your external authentication configuration in Automation Controller will automatically be migrated to the new unified Platform UI.

          **Database:** AAP 2.6 continues to use PostgreSQL 15.

          #### Create the AnsibleAutomationPlatform custom resource (Required for upgrades from 2.4 only)

          **If upgrading from AAP 2.4:** Once the operator upgrade is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created.

          **If upgrading from AAP 2.5:** This step is not required as your AnsibleAutomationPlatform custom resource already exists.

          Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          Next, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda (lightspeed is disabled by default):


          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: aap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-26/gateway-rhel9@sha256:c85dbbf6beaa264afec480dc1f06576aada531cdfd15f247f4db37dc9702be33
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-ui-rhel9@sha256:5b5f38a5f76781a042eec9cbf70b8be026878e4fa61c7fb90d1b0ad64bd21592
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-rhel9-operator@sha256:cfbb98e17297134c187673f355342ac5ad60a56d73f4b0c918be2ff244286cc0
        - registry.redhat.io/ansible-automation-platform-26/controller-rhel9@sha256:64f5e77a7842965e9c6f7bc7990354b71bb19ffb06ab8ef7793dde956cb6c0f0
        - registry.redhat.io/ansible-automation-platform-26/ee-supported-rhel9@sha256:2534d07f1ae41af32adb681e283cbaea465cbe0f9d234ed5478a7bfad7676e7f
        - registry.redhat.io/ansible-automation-platform-26/ee-minimal-rhel9@sha256:261c324cac081e47d559c5fccbd48e7ec8f5bf64e3f27287fbe4fb8aff6b7d99
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-rhel9@sha256:c5d4e2e76944e0991729e5c3b99f43b7f29ea08ad112e5e6513feb8fb1d899a1
        - registry.redhat.io/ansible-automation-platform-26/hub-web-rhel9@sha256:aa4041672542a7f828d8a51c79e9c4dcea2a185388e4f20f828104db37e03a9b
        - registry.redhat.io/ansible-automation-platform-26/platform-resource-runner-rhel9@sha256:9983043386df38a17ff6fffef64dc187691e0b1fefec01eef3af39aa08e532ef
        - registry.redhat.io/ansible-automation-platform-26/gateway-rhel9-operator@sha256:fa1b102a0690767a1e08935a0e164c1eabb2d4cf9afccd310262ee9185544de7
        - registry.redhat.io/ansible-automation-platform-26/platform-resource-rhel9-operator@sha256:969bbd033fbba33624329ee9347b3b89fb5a1f64e71892f5c6c7d74ed14a6d39
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-chatbot-rhel9@sha256:8901fba14b21569b02108d2ab9050d558dee29f8177fae9d1bfcd5de3f4ce7c1
        - registry.redhat.io/ansible-automation-platform-26/mcp-tools-rhel9@sha256:0c1f6f6958f2570c8cd073287555d28c70ef9dded702ec014d3ead8ce572463e
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-rhel9-operator@sha256:bfd55bffd8239e224d4ebba500d8517e8e66915078d6b59244f5e910fc5e10c8
        - registry.redhat.io/ansible-automation-platform-26/hub-rhel9-operator@sha256:18114aab6db0d00780fdeaf9efa74ee44cab94824c2fd35f4b9998b6378ad8b3
        - registry.redhat.io/rhel9/postgresql-15@sha256:5e225c1646251d2509b3a6e45fad23abfb11dd5c45d3224b59178f90deb38a70
        - registry.redhat.io/ubi9/ubi-minimal@sha256:7c5495d5fad59aaee12abc3cbbd2b283818ee1e814b00dbc7f25bf2d14fa4f0c
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-rhel9@sha256:01e56bee772d15b120d6e4b02a59024b21eb8b266127b1a9b7f2c535ecdb2855
        - registry.redhat.io/ansible-automation-platform-26/gateway-proxy-rhel9@sha256:13284d20c6fb4f2de24ec1715e7abdb5c4b0e4298c5034a7795f494b93634a02
        - registry.redhat.io/ansible-automation-platform-26/hub-rhel9@sha256:8041b141cd8921aaaa20cce15eab4c6189da9f449d99057189284891c071ac3b
        - registry.redhat.io/rhel9/redis-6@sha256:0962d18bb5292c0e524354f319b63287c2298a4a2fd34556c68fa285df381427
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/ansible-automation-platform-26/controller-rhel9-operator@sha256:863e77e52ab01cec2a71672a224080a59d963d4fc08b9ed6f3cc7950729b1746
        version: 2.6.0+0.1758685241
      entries:
      - name: aap-operator.v2.6.0-0.1758685241
        version: 2.6.0+0.1758685241
      name: stable-2.6-cluster-scoped
    defaultChannel: stable-2.6-cluster-scoped
    packageName: ansible-automation-platform-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/component: kmm-hub
      app.kubernetes.io/name: kmm-hub
      app.kubernetes.io/part-of: kmm
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: kernel-module-management-hub
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kernel-module-management-hub.v2.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:888a3acad339cf38a324c718e0cbe1237ca5bf2f4636b01d001752658e4cca1b
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.0.2'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:888a3acad339cf38a324c718e0cbe1237ca5bf2f4636b01d001752658e4cca1b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62c44057a27abd41cf3daf60b8736da59a82dadb65fa6c32bef65d19bd49ea49
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:192e08b121a04ae69aea60c35199e8f5e3c42305e1ebd024a605a371de4ab472
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:94b4ddc45cb34d4ffe794a51b99c868cb709f90cc4eb8b28284d70c526a0a226
        version: 2.0.2
      entries:
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.0
    - currentCSV: kernel-module-management-hub.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:d1c1490d4c20dc2ec31effa1af067676c4d264e3b7d4670fc11c2baf8a400001
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.1.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:14e100e5853444114029116255ebfb7a54e5ed51503ba969740ad48b5ca80f70
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:1a9a19df17b464d1708b2905d03b8726e4480ca1f11aa97a15ff4c049a24ea4b
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:d1c1490d4c20dc2ec31effa1af067676c4d264e3b7d4670fc11c2baf8a400001
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:611556d999ea97f9dbb78fec5963baa02de7a26626d7096ba63b177e3bcf9707
        version: 2.1.1
      entries:
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.1
    - currentCSV: kernel-module-management-hub.v2.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:90c1946b3f577cde0435402b04baa52dc54e5aab3bb84bd834ac026a57453296
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.2.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:b1d98c0f14448ae8eb80821cbec236150c338674aae2a2b37f8d0ab0d5f1430c
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:53fddf0b226eb579c71c86112a8d9b6b1d90a1498d6f6985932b16b8be532f35
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:c3dc9ad4ebc68cd85a074b520903a9310e72a90a73bdf4aa7c42b7ed66b7bcba
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:90c1946b3f577cde0435402b04baa52dc54e5aab3bb84bd834ac026a57453296
        version: 2.2.1
      entries:
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.2
    - currentCSV: kernel-module-management-hub.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:f1d8f3c95f621c3faf41867c5dd4bc13fa2b9b291422d6184e6e42f1d1460259
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.3.0'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:f1d8f3c95f621c3faf41867c5dd4bc13fa2b9b291422d6184e6e42f1d1460259
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:c5005b74402f680d11c9ea682a9a0d3f74b6ab0d01d25298d232f910f1a29c34
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:f9fde334af68220943017a7491130e6a82f15e63522f5bb3931aa7d9b00ccc40
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:06ac089cc89b303e525fd71d6439b958f34ab944a72209f1cc8cb562ecfd1112
        version: 2.3.0
      entries:
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.3
    - currentCSV: kernel-module-management-hub.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm-hub",
                    "app.kubernetes.io/name": "kmm-hub",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:2efb81137ad2981b994c62d460c917749308133a7dcad5b9a2b251e18f3052ec
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        version: 2.4.1
      entries:
      - name: kernel-module-management-hub.v2.4.1
        version: 2.4.1
      - name: kernel-module-management-hub.v2.4.0
        version: 2.4.0
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.4
    - currentCSV: kernel-module-management-hub.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm-hub",
                    "app.kubernetes.io/name": "kmm-hub",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:2efb81137ad2981b994c62d460c917749308133a7dcad5b9a2b251e18f3052ec
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        version: 2.4.1
      entries:
      - name: kernel-module-management-hub.v2.4.1
        version: 2.4.1
      - name: kernel-module-management-hub.v2.4.0
        version: 2.4.0
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: kernel-module-management-hub
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: cert-utils-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-utils-operator.v1.3.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/cert-utils-operator@sha256:0290e7b2800a5aa8eb57b0268908f2a20a93b16403c76df4c61021048161d6ff
          createdAt: "2024-01-23T12:29:04Z"
          description: Set of utilities for TLS certificates
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-utils-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/cert-utils-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Cert utils operator is a set of functionalities around certificates packaged in a [Kubernetes operator](https://github.com/operator-framework/operator-sdk).

          Certificates are assumed to be available in a [secret](https://kubernetes.io/docs/concepts/configuration/secret/) of type `kubernetes.io/tls` (other types of secrets are *ignored* by this operator).
          By convention this type of secrets have three optional entries:

          1. `tls.key`: the private key of the certificate.
          2. `tls.crt`: the actual certificate.
          3. `ca.crt`: the CA bundle that validates the certificate.

          The functionalities are the following:

          1. [Ability to populate route certificates](https://github.com/redhat-cop/cert-utils-operator#Populating-route-certificates)
          2. [Ability to create java keystore and truststore from the certificates](https://github.com/redhat-cop/cert-utils-operator#Creating-java-keystore-and-truststore)
          3. [Ability to show info regarding the certificates](https://github.com/redhat-cop/cert-utils-operator#Showing-info-on-the-certificates)
          4. [Ability to alert when a certificate is about to expire](https://github.com/redhat-cop/cert-utils-operator#Alerting-when-a-certificate-is-about-to-expire)
          5. [Ability to inject ca bundles in ValidatingWebhookConfiguration, MutatingWebhookConfiguration, CustomResourceDefinition object](https://github.com/redhat-cop/cert-utils-operator#CA-injection)

          All these feature are activated via opt-in annotations.
        displayName: Cert Utils Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Cert Utils Operator
          url: https://github.com/redhat-cop/cert-utils-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/cert-utils-operator
        maintainers:
        - email: rspazzoli@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/cert-utils-operator@sha256:0290e7b2800a5aa8eb57b0268908f2a20a93b16403c76df4c61021048161d6ff
        version: 1.3.12
      entries:
      - name: cert-utils-operator.v1.3.12
        version: 1.3.12
      - name: cert-utils-operator.v1.3.11
        version: 1.3.11
      - name: cert-utils-operator.v1.3.10
        version: 1.3.10
      - name: cert-utils-operator.v1.3.9
        version: 1.3.9
      - name: cert-utils-operator.v1.3.8
        version: 1.3.8
      - name: cert-utils-operator.v1.3.7
        version: 1.3.7
      - name: cert-utils-operator.v1.3.6
        version: 1.3.6
      - name: cert-utils-operator.v1.3.5
        version: 1.3.5
      - name: cert-utils-operator.v1.3.4
        version: 1.3.4
      - name: cert-utils-operator.v1.3.3
        version: 1.3.3
      - name: cert-utils-operator.v1.3.2
        version: 1.3.2
      - name: cert-utils-operator.v1.3.1
        version: 1.3.1
      - name: cert-utils-operator.v1.3.0
        version: 1.3.0
      - name: cert-utils-operator.v1.2.0
        version: 1.2.0
      - name: cert-utils-operator.v1.1.0
        version: 1.1.0
      - name: cert-utils-operator.v1.0.7
        version: 1.0.7
      - name: cert-utils-operator.v1.0.6
        version: 1.0.6
      - name: cert-utils-operator.v1.0.5
        version: 1.0.5
      - name: cert-utils-operator.v1.0.4
        version: 1.0.4
      - name: cert-utils-operator.v1.0.3
        version: 1.0.3
      - name: cert-utils-operator.v1.0.2
        version: 1.0.2
      - name: cert-utils-operator.v1.0.1
        version: 1.0.1
      - name: cert-utils-operator.v1.0.0
        version: 1.0.0
      - name: cert-utils-operator.v0.2.2
        version: 0.2.2
      - name: cert-utils-operator.v0.2.1
        version: 0.2.1
      - name: cert-utils-operator.v0.2.0
        version: 0.2.0
      - name: cert-utils-operator.v0.1.1
        version: 0.1.1
      - name: cert-utils-operator.v0.1.0
        version: 0.1.0
      - name: cert-utils-operator.v0.0.3
        version: 0.0.3
      - name: cert-utils-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cert-utils-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ses-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ses-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ses-controller:1.1.1
          createdAt: "2025-09-19T19:37:25Z"
          description: AWS SES controller is a service controller for managing SES
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConfigurationSet represents the state of an AWS ses ConfigurationSet
              resource.
            displayName: ConfigurationSet
            kind: ConfigurationSet
            name: configurationsets.ses.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Simple Email Service (SES) resources in AWS from within your Kubernetes cluster.

          **About Amazon SES**

          Amazon Simple Email Service (SES) is an email platform that provides an easy, cost-effective way for you to send and receive email using your own email addresses and domains.
          For example, you can send marketing emails such as special offers, transactional emails such as order confirmations, and other types of correspondence such as newsletters. When you use Amazon SES to receive mail, you can develop software solutions such as email autoresponders, email unsubscribe systems, and applications that generate customer support tickets from incoming emails.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SES
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ses
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SES Developer Resources
          url: https://aws.amazon.com/ses/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ses maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ses-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ses-controller.v1.1.1
        version: 1.1.1
      - name: ack-ses-controller.v1.1.0
        version: 1.1.0
      - name: ack-ses-controller.v1.0.13
        version: 1.0.13
      - name: ack-ses-controller.v1.0.12
        version: 1.0.12
      - name: ack-ses-controller.v1.0.11
        version: 1.0.11
      - name: ack-ses-controller.v1.0.10
        version: 1.0.10
      - name: ack-ses-controller.v1.0.9
        version: 1.0.9
      - name: ack-ses-controller.v1.0.8
        version: 1.0.8
      - name: ack-ses-controller.v1.0.7
        version: 1.0.7
      - name: ack-ses-controller.v1.0.6
        version: 1.0.6
      - name: ack-ses-controller.v1.0.5
        version: 1.0.5
      - name: ack-ses-controller.v1.0.3
        version: 1.0.3
      - name: ack-ses-controller.v1.0.2
        version: 1.0.2
      - name: ack-ses-controller.v1.0.1
        version: 1.0.1
      - name: ack-ses-controller.v1.0.0
        version: 1.0.0
      - name: ack-ses-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-ses-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco Networking - Compute Solutions
      provider-url: ""
    name: ucs-ci-solutions-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ucs-ci-solutions-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intersight.cisco.com/v1alpha1",
                "kind": "UCSCISolutions",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ucs-ci-solutions-operator",
                    "app.kubernetes.io/instance": "ucscisolutions-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ucscisolutions",
                    "app.kubernetes.io/part-of": "ucs-ci-solutions-operator"
                  },
                  "name": "ucscisolutions-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: docker.io/pkoppa/ucs-ci-solutions-operator:latest
          createdAt: "2023-07-15T15:19:39Z"
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: UCSCISolutions
            name: ucscisolutions.intersight.cisco.com
            version: v1alpha1
        description: The operator is used to automate the creation of  Pools, Policies
          and Server Profile Templates required to deploy Cisco UCS Converged Infrastructure
          Solution in Intersight.
        displayName: UCS-CI-Solutions-Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ucscisolutions
        - intersight
        - ucs
        - ci
        - flexpod
        - flashstack
        links:
        - name: Ucs Ci Solutions Operator
          url: https://ucs-ci-solutions-operator.domain
        maintainers:
        - email: pkoppa@cisco.com
          name: Paniraj Koppa
        - email: chrobrie@cisco.com
          name: Chris O'Brien
        maturity: alpha
        provider:
          name: Cisco Networking - Compute Solutions
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - docker.io/pkoppa/ucs-ci-solutions-operator:latest
        version: 0.0.1
      entries:
      - name: ucs-ci-solutions-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ucs-ci-solutions-operator
    provider:
      name: Cisco Networking - Compute Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-diagnostics
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-diagnostics.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Diagnostics",
                "metadata": {
                  "name": "diagnostics-sample"
                },
                "spec": {
                  "agentDiagnosticsService": {
                    "port": 8080
                  },
                  "auth": {
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "wapdbDiagnosticsService": {
                    "dumpedTables": "[ \"--regexpr cluster_logs_week_*\", \"current_week_cluster_logs\", \"config_metadata_dbs\", \"config_metadata_dbs_id_seq\", \"config_metadata_tags\", \"config_tags\", \"csv_reports_meta\", \"custom_import_handler_dbs\", \"data_point_db_models\", \"db_test_entries\", \"license_server_metadata\", \"operations\", \"result_file_metadata_dbs\", \"result_metadata_dbs\", \"result_metadata_tags\", \"result_tags\", \"session_meta\", \"session_types\", \"sessions\", \"stat_view_query_data\", \"test_run_configs\", \"test_run_infos\", \"test_runs\", \"wireless_globalsettings\", \"wireless_sessionconfigs\", \"wireless_templates\"]",
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:37:10Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Diagnostics
            name: diagnostics.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Diagnostics Operator
        displayName: Keysight WAP Diagnostics Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Diagnostics
          url: https://keysight-wap-diagnostics.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-diagnostics@sha256:97bf8affa7bf1f3a391f5f036313e7cd8196be2eb8e95debeab06c1b6c964c5e
        version: 0.1.21
      entries:
      - name: keysight-wap-diagnostics.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-diagnostics
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JFrog
      provider-url: ""
    name: openshiftartifactoryha-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: artifactory-ha-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftArtifactoryHa",
                "metadata": {
                  "name": "openshiftartifactoryha"
                },
                "spec": {
                  "artifactory-ha": {
                    "artifactory": {
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE",
                      "node": {
                        "replicaCount": 2,
                        "waitForPrimaryStartup": {
                          "enabled": false
                        }
                      }
                    },
                    "databaseUpgradeReady": true,
                    "database": {
                      "driver": "OVERRIDE",
                      "password": "OVERRIDE",
                      "type": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "nginx": {
                      "http": {
                        "externalPort": 80,
                        "internalPort": 8080
                      },
                      "https": {
                        "externalPort": 443,
                        "internalPort": 8443
                      },
                      "service": {
                        "ssloffload": false
                      },
                      "tlsSecretName": "OVERRIDE"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "waitForDatabase": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/artifactory-operator@sha256:de69bb16052d5ac6fed9882eb65d49446b6fa1cc5ba830dd2a39e882712b1347
          createdAt: 2024-03-23 00:00:00+00:00
          description: JFrog Artifactory Enterprise deploys Artifactory in a high
            availability environment across multiple pods
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Artifactory HA Instances
            displayName: Artifactory HA
            kind: OpenshiftArtifactoryHa
            name: openshiftartifactoryhas.charts.helm.k8s.io
            version: v1
        description: |2+
           ## Breaking change
          Please update to version 1.1.10 of the operator.
          Version 1.1.8 and 1.1.9 have issues. Please skip these versions.
          If you are using operator version < 1.2.0 there is a issue with upgrade where artifactory version is not updated if you update operator version. As workaround you can update the Custom Resource OpenshiftArtifactoryHa which has hardcored image value.After upgrading to version 1.2.0 you can remove the below parameters from your CR so that future upgrade works without workaround. ``` artifactory-ha.artifactory.image.registry artifactory-ha.artifactory.image.repository artifactory-ha.artifactory.image.tag ```
          ## Overview
          Openshift Operator to deploy JFrog Artifactory Enterprise into your Openshift cluster.
          ## Security Context Constraints
          To deploy this helm chart you will need to be a cluster admin w/ access to the anyuid scc and add the operator service account to the anyuid scc.
          ```
          oc adm policy add-scc-to-user anyuid -z artifactory-ha-operator -n openshift-operators
          ```
          Add the service account for the Artifactory chart to deploy successfully:
          ``` oc adm policy add-scc-to-user anyuid -z openshiftartifactoryha-artifactory-ha -n artifactory ```
          ### Usage

          An external DB is required. The operator will not deploy a DB but will require you to specify the configuration values to connect to it.

          Search for JFrog and click JFrog Artifactory Enterprise Operator to install.

          Go to the Installed Operators.

          Wait for the JFrog Artifactory Enterprise Operator to complete the installation.

          Open the Operator and click on the provided API: Artifactory HA.

          Click Create New Instance and provide the following parameters for your DB configuration:

          ```
          DATABASE_TYPE
          DATABASE_DRIVER
          DATABASE_URL
          DATABASE_USER
          DATABASE_PASSWORD
          ```
          Master key and Join key must be supplied. To generate a new key for each run the command below:

          ```
          # Create a key
          export JOIN_KEY=$(openssl rand -hex 32)
          echo ${JOIN_KEY}
          ```

          To use TLS you will need to first create a k8s tls secret to store your .crt and .key file into.
          Then supply the value of this k8s secret into the TLS_SECRET field.
          ``` oc create secret tls my_tls_secret --cert=tls.crt --key=tls.key --namespace=my_namespace ```
          Click Create for Artifactory Enterprise to deploy into OpenShift and connect to it on the external IP exposed by the load balancer.
          ### Create a route
          To expose Artifactory from Openshift we recommend you create a new route in Openshift.
          You can either use the oc command line tool or the Openshift web console to generate a new route.
          Depending upon where you terminate TLS you will need to either specify pass through or edge.
          Command Line (Edge):
          ``` oc create route edge --service=openshiftartifactory-ha --cert=tls.crt --key=tls.key --ca-cert=ca.crt --hostname=www.example.com ```
          Or for more information visit the official Openshift documentation on routes here:
          https://docs.openshift.com/container-platform/4.6/networking/routes/route-configuration.html

        displayName: JFrog Artifactory Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Artifact Repository
          url: https://jfrog.com/artifactory/
        - name: Artifactory Video
          url: https://www.youtube.com/watch?v=r2_A5CPo43U
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: alpha
        provider:
          name: JFrog
        relatedImages:
        - registry.redhat.io/rhel8/nginx-120@sha256:87f658744af695273c434fac8246b716728b691a17d5d0c5b89247362e61e45c
        - registry.connect.redhat.com/jfrog/artifactory-operator@sha256:de69bb16052d5ac6fed9882eb65d49446b6fa1cc5ba830dd2a39e882712b1347
        - registry.connect.redhat.com/jfrog/artifactory-pro@sha256:0986ca71419ae628892db11e7ca7da5ac1c60219c144ce9aeb2de6e683a45262
        version: 1.3.3
      entries:
      - name: artifactory-ha-operator.v1.3.3
        version: 1.3.3
      name: alpha
    - currentCSV: artifactory-ha-operator.v1.1.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftArtifactoryHa",
                "metadata": {
                  "name": "openshiftartifactoryha"
                },
                "spec": {
                  "artifactory-ha": {
                    "artifactory": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/artifactory-pro",
                        "tag": "7.27.3-1"
                      },
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE",
                      "uid": "1000721030",
                      "node": {
                        "replicaCount": 2,
                        "waitForPrimaryStartup": {
                          "enabled": false
                        }
                      }
                    },
                    "databaseUpgradeReady": true,
                    "database": {
                      "driver": "OVERRIDE",
                      "password": "OVERRIDE",
                      "type": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "initContainerImage": "registry.connect.redhat.com/jfrog/init@sha256:164b41720a37f75c6129c8a28d05cc56f821022313d44ff918cce4a2146d93e8",
                    "nginx": {
                      "uid": "1000720104",
                      "gid": "1000720107",
                      "http": {
                        "externalPort": 80,
                        "internalPort": 8080
                      },
                      "https": {
                        "externalPort": 443,
                        "internalPort": 8443
                      },
                      "image": {
                        "registry": "registry.redhat.io",
                        "repository": "rhel8/nginx-116",
                        "tag": "latest"
                      },
                      "service": {
                        "ssloffload": false
                      },
                      "tlsSecretName": "OVERRIDE"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "waitForDatabase": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/artifactory-operator@sha256:e292553e1b6e792ff19276c7f6565d7ed61dd95289eb176066329954bc93e371
          createdAt: 2020-03-25 00:00:00+00:00
          description: JFrog Artifactory Enterprise deploys Artifactory in a high
            availability environment across multiple pods
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Artifactory HA Instances
            displayName: Artifactory HA
            kind: OpenshiftArtifactoryHa
            name: openshiftartifactoryhas.charts.helm.k8s.io
            version: v1
        description: |2+
           ## Breaking change
          Please update to version 1.1.10 of the operator.
          Version 1.1.8 and 1.1.9 have issues. Please skip these versions.

          ## Overview
          Openshift Operator to deploy JFrog Artifactory Enterprise into your Openshift cluster.
          ## Security Context Constraints
          To deploy this helm chart you will need to be a cluster admin w/ access to the anyuid scc and add the operator service account to the anyuid scc.
          ```
          oc adm policy add-scc-to-user anyuid -z artifactory-ha-operator -n openshift-operators
          ```
          Add the service account for the Artifactory chart to deploy successfully:
          ``` oc adm policy add-scc-to-user anyuid -z openshiftartifactoryha-artifactory-ha -n artifactory ```
          ### Usage

          An external DB is required. The operator will not deploy a DB but will require you to specify the configuration values to connect to it.

          Search for JFrog and click JFrog Artifactory Enterprise Operator to install.

          Go to the Installed Operators.

          Wait for the JFrog Artifactory Enterprise Operator to complete the installation.

          Open the Operator and click on the provided API: Artifactory HA.

          Click Create New Instance and provide the following parameters for your DB configuration:

          ```
          DATABASE_TYPE
          DATABASE_DRIVER
          DATABASE_URL
          DATABASE_USER
          DATABASE_PASSWORD
          ```
          Master key and Join key must be supplied. To generate a new key for each run the command below:

          ```
          # Create a key
          export JOIN_KEY=$(openssl rand -hex 32)
          echo ${JOIN_KEY}
          ```

          To use TLS you will need to first create a k8s tls secret to store your .crt and .key file into.
          Then supply the value of this k8s secret into the TLS_SECRET field.
          ``` oc create secret tls my_tls_secret --cert=tls.crt --key=tls.key --namespace=my_namespace ```
          Click Create for Artifactory Enterprise to deploy into OpenShift and connect to it on the external IP exposed by the load balancer.
          ### Create a route
          To expose Artifactory from Openshift we recommend you create a new route in Openshift.
          You can either use the oc command line tool or the Openshift web console to generate a new route.
          Depending upon where you terminate TLS you will need to either specify pass through or edge.
          Command Line (Edge):
          ``` oc create route edge --service=openshiftartifactory-ha --cert=tls.crt --key=tls.key --ca-cert=ca.crt --hostname=www.example.com ```
          Or for more information visit the official Openshift documentation on routes here:
          https://docs.openshift.com/container-platform/4.6/networking/routes/route-configuration.html

        displayName: JFrog Artifactory Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Artifact Repository
          url: https://jfrog.com/artifactory/
        - name: Artifactory Video
          url: https://www.youtube.com/watch?v=r2_A5CPo43U
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: stable
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/init@sha256:164b41720a37f75c6129c8a28d05cc56f821022313d44ff918cce4a2146d93e8
        - registry.connect.redhat.com/jfrog/artifactory-operator@sha256:e292553e1b6e792ff19276c7f6565d7ed61dd95289eb176066329954bc93e371
        - registry.connect.redhat.com/jfrog/artifactory-pro@sha256:7289dde95d1e4d8e161a28cffd9c4a0cc96f92f7536d069ab63ec304a92254fd
        - registry.redhat.io/rhel8/nginx-116@sha256:d71d368d6094a5777000a8d99164959ccb97fca355e855e24bd92c13a8df04f6
        version: 1.1.18
      entries:
      - name: artifactory-ha-operator.v1.1.18
        version: 1.1.18
      name: stable
    defaultChannel: alpha
    packageName: openshiftartifactoryha-operator
    provider:
      name: JFrog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: advanced-cluster-management
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: advanced-cluster-management.v2.10.9
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-08-06T03:17:11Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0 <2.10.9'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/internal-objects: '[]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterHub defines the configuration for an instance
              of the MultiCluster Hub
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:95df18f4479b7f555a41508486624bde714778730921c20c665e9f23baadbabf
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:8f1bb3a8422d29960d8b90dd24a813a56875abf1f0393f78b862eaa48ddda783
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:50331cdd61332c8f4128b3dab237255463b307c5c40b0607730a7d3b33b9c6b5
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:5b162aeab4aa51185c880696d1472e152a20ada926a3baea0fb8494793ad238e
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:0ba2471b7a67532bdb90e02904671f0e724cdb39f74a45392ae6c693c25b354d
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:4b16da3070e30d879421aaa0d4a73aa2a1b0b852f62b8faabd801c03f22d5319
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:ced03229f208739867a43b9ed9bd158f19f4242dd1f5f045d3e5db81b2b1d171
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:4ad5ccf9ce682024b6e3025a9bcadd6f32692aa18d892afd670eacb7b4031860
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:1c2dbcf179508ea1c602f7817f85b0d3f58d6d85ae93ccc523e88a3936c9c690
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:23713ebb12d510d37de7e060d5ce4ff2198990f83cebbd2302e74bd83a55cecd
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:1aaba9ca65e363c7f9c07ef2f66f23cb8fff677204a56373634d319113d2c84b
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:e3c80e8e3af5c02c202dfb288c63b20c44c08f2a0474d2568cfb0b878f7fe888
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:9e78db48ff0b0fc0ac0347ca055a4b66e2f917073499adc6e2db2a9c16a70e01
        - registry.redhat.io/rhacm2/console-rhel9@sha256:1538a3a935aa01d8800d8145c45037eaab6987ad6e6e927017ca1cfea4bb111c
        - registry.redhat.io/rhel9/postgresql-13@sha256:31c77c19480ab82cb33525d32419b68cf59daccf0b9936e88e8238e48a2d47e5
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:fbc5b1456a37951c9f70f64357919222739f4a7e533fb6dde5b93672535fc87e
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:95890bb1aad544d97ae1b80e801a800c95ef51e1034b60e680b67bdb2f5aeb69
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:9e931cab5fcbf3df0ffe6fff8d12303800423ace3d749ee0e33b55300c816b19
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:dcd6a420467929b9af8c52c77c8717d22f3977e6b6b392b5dc90cbb04005b9f1
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:14094f09dfb166a6b342bd91e2db9162f0de88d55655ccce44d345047c700187
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:23b4d7fbc776c63dcd7007442a557124581939ce110257822e8233261d31fb7d
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:bbe57064cb1c40c35929526a082f11b2da6fd3746b6eef8af51ca63a4fd7d98b
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:46e3b8ccf3e7069b44eb3ad407329c03048b5219ef2886e237e1ab0dc3f7bae4
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:bb4b72c1510da84aad2d67516c58f37495d7682c51c8f2088e2c4a350fa36c50
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:dcbc4c5e2d46ec2b4911d8cd9f96a4c141bf7dd18b79aa4c25b77141e4f60b48
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:69a8faa2f4da2281c6c124508634628510d91ad40f725a9b41790e0821d64c14
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:500718f834e05c7a4531eb4f9d8ce560053c5125f76f6921c38c9cf150046a23
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:9db39230c528bfc675bc422a7acd40c394169b9632b088ca16358751994c9cac
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:335c903dde6f21e9c54f025c45666f6b4b6e7d455dfd5f85a6373b3947069f0a
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:3ccc45b0e9483f33be0d3b32a6d46fa794976131ad2f9861d96e319c8e35b9be
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:2f68a16a14f192b2cceaa07ed4840c08fbf413718a39bd584cacde866be7f2b1
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:9c4eb0ef8758d943796e13f65bc92b190db614eabd0fe7375ebfecc90e96e32c
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:da09087ba7d558611d733551b6ebc27ccf441581fc264efe7a5933db3816ef89
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:e02992cec8edcc9374c4cd30dfd7bccbda77d6d5cdd309862ab5265494f1278e
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:f4474ed21a71ad490778a6a378a856bec1884b4b9dc2c0e6c6140cc6c9d03b61
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:805e443fbfca4f816dc2dcea71f4cf723fd753ccea6d13f04d25945f29cb47dc
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:d8b8d130f75647cfeb9036db50d66fc4f1fca56fead6add4ba24c52f9213543e
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:13c33a997be0ddef59d1fb31835751367d4f901a1a6eac41284931bf7988035c
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:814f34f9e55b59d96021d3160b1dad7a68ffe208b3f4a999840b228aa7b3300b
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:47a8fa9541275f849a24e2ef1df87b7e9051dedd0ee8bd833d60defbd8724b46
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:3803d82d2b829c0e0883c0504176a96d2012cb7096cfcace316831a2b55bca3f
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:de68d598587a967d4b3f5e0e28c4675f74fd79cae7a5bb197c1b4b44bb4b31d9
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:35ff861165fccd848b8d328fae46cb8b3ac0c853c63062ec8672a6d41a23f333
        - registry.redhat.io/rhacm2/iam-policy-controller-rhel9@sha256:8c2b356bd9ff4ebfba9059a5f20cad8bd336af237d84cdc21f5c0aa35fabeabe
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:d989e9d2078f8a1e4b18de64f917e6d5e290c5cac769899f829b5f6bbfe6f0f2
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:87093f85c6235c99d4cca9d503b0e4198e3731e5fc3171506485091570e70bb5
        version: 2.10.9
      entries:
      - name: advanced-cluster-management.v2.10.9
        version: 2.10.9
      - name: advanced-cluster-management.v2.10.8
        version: 2.10.8
      - name: advanced-cluster-management.v2.10.7
        version: 2.10.7
      - name: advanced-cluster-management.v2.10.6
        version: 2.10.6
      - name: advanced-cluster-management.v2.10.5
        version: 2.10.5
      - name: advanced-cluster-management.v2.10.4
        version: 2.10.4
      - name: advanced-cluster-management.v2.10.3
        version: 2.10.3
      - name: advanced-cluster-management.v2.10.2
        version: 2.10.2
      - name: advanced-cluster-management.v2.10.1
        version: 2.10.1
      - name: advanced-cluster-management.v2.10.0
        version: 2.10.0
      name: release-2.10
    - currentCSV: advanced-cluster-management.v2.11.8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-08-01T05:25:37Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.10.0 <2.11.8'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:956d549e64c412f86f2083c8ddf97fea269d7806d8826aab93178bec9ed5d7af
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:9d821f4373722d9cb2b679cf5793ef22cb0b9ebcbb5a38060504c93c05c492c4
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:aef5ffbaa578d006bc7caec8fb35446dc263a51813f84c756aa57903b428329b
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:d248a5fad4ae3ee3210eb6dc4f58406d2f1e8f5c390f3c491040b437fe1adeba
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:447b1f6067a82476e14dfbb74a8f324fb86956420e55575f0017b91784f303a5
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:a081f531efb5b54d7d14d9ba0519552ecfc87c3a80d9d3ec4e598d2b445fc689
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:ffd3d9c363d8e15232bd6d12c5f1a77f7cb1ca18f94b7aea79d226568be4586e
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:95890bb1aad544d97ae1b80e801a800c95ef51e1034b60e680b67bdb2f5aeb69
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:f5c56a7381eb80c02a7c140f7c3f0ed08bb93697cd51732ee57498c080d8d3e5
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:8d15be5f40d135fd7069ac768dcf9188d8d8299c65db56df04e66f8a28941966
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:94c89ebd9257797ddc264511399689feb70d56e53332d9894c6d95fbdc51de40
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:6d4e3114494e051b3b875eeeae4bad4d48665fbcbbd3574a98f87b5a5742b41d
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:9be2d97fa3c1140db7d86157503a96ca2d3bdfb57125ab2230c8cf0649ea1e90
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:d4feeb3d3322210765d916ef1f49b47a1527f1c61041f94834c4395319a9ec32
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:49f0ec1aced72e6a2f0c862b0c6c0cd430d997a09ceb9eb0a7d25aaa1d9798b7
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:a9e0533af92ebadca27c9fe93fe4db9c68a1a8bc0060f4d13f67f26eb5691cb3
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:b3e0427341e3b1cbb1fe9166c56184eb0b0ec526d8a3992306e4861f65e4f0c0
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:ded47519fcf30e60cf38d44a101494b0318757969675782365de88b51cab24f0
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:968eb25c5cfcbdcea87b92dbff734d01b5954c13c13f5e794ad1db7bd886e2b4
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:d7552b2a3c864a3fc0d279a2519535e170fd8608195fe1f9abb4b538dfe2a3df
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:d6666c572294c7c85990304d21eaae3040b9c4e47fcb7b446289cb09aabc2d26
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:4a62d8706772ede443cffa005b984c26d137c35a33aafedaa7f01050b618c2a2
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:d4bfef313ff9044ae36aa5f832651a4dd09228d4b9b186803009aafe9774ee2c
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:1e8a26ebfcc33e60f60da4ba8d58cf7c152a99752eac6ebc9e114187e4cc2267
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:4dce3cebeb4ea0f1d3fc1c7e036dc6a3fd09e02f8f2ae4f3a753d9f4f97defe7
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:b713264826124df2c2b3a99dc268a3f3bd7eb24a9010b93c9876d83656e70e1a
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:e36832aa4852ea807d44f2b1f2c402bdaac6b7e2246bf9dd811b41f0b5e8e647
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:eaa719f9d9db8f91eced81e0e164bbc12bb502bf4760fc8c673cc542b4414771
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:45d4404abe02596611b29316bd10985f7a0684b1d305a57d3b24068ac776cf6e
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:251d3ee881c3e902e87f1752648ab084e8780aef79eb88a253ba79f984a01bcf
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:e347cd0f57e0691c61263e6554495c76ce56c7443bce3fd010cf76d068083fee
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:ad10d15efb6833b2daaf153210a080420395a7debfcb873dd757bb924197a654
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:36c90a52b12af24faaf28301d0d1a0c522723bd6cadd7391e5ee1f43f3019eef
        - registry.redhat.io/rhacm2/console-rhel9@sha256:2ed2f4add387caf790b2b41a65a3986e0046e08d0b8406ad17e4f0440ba0ac4a
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:5ba07381111651906ac3c93344e705152eb006826cb36bf5c05bd88723c4c999
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:7d39d0c29b02c399ffdcd83278693a8dd9c5ca3931e660af056030b126d10055
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:eec0e9bd6ea6f146939516da6c7ca9c3755888d96e81954f2f282fae63c1847c
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:827ca1beb707c02908914c517c8ab14b77b1c65ca7690dbbe1bcff600b7551d5
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:4e0c2cf5229fd68102a9e6d0f3155c651b1159b7af7b057151ce52990f747b08
        - registry.redhat.io/rhel9/postgresql-13@sha256:31c77c19480ab82cb33525d32419b68cf59daccf0b9936e88e8238e48a2d47e5
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:ef66846a3f178234618fd1e265067f863bbb56f14cbca0dfc96b0d741885dc9b
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:4b16da3070e30d879421aaa0d4a73aa2a1b0b852f62b8faabd801c03f22d5319
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:028b7258ec88d04397798552ed19a459110660ddbdf3ab375a7a1f74049c455a
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:147c567ce95673ba7151bcd290474756e13018ebc9f7275d61677ffe586a8a76
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:2878b5c41f8d58e986e86803bfd7eb3a05ccf23f7e49ebba56ad30245609b23d
        version: 2.11.8
      entries:
      - name: advanced-cluster-management.v2.11.8
        version: 2.11.8
      - name: advanced-cluster-management.v2.11.7
        version: 2.11.7
      - name: advanced-cluster-management.v2.11.6
        version: 2.11.6
      - name: advanced-cluster-management.v2.11.5
        version: 2.11.5
      - name: advanced-cluster-management.v2.11.4
        version: 2.11.4
      - name: advanced-cluster-management.v2.11.3
        version: 2.11.3
      - name: advanced-cluster-management.v2.11.2
        version: 2.11.2
      - name: advanced-cluster-management.v2.11.1
        version: 2.11.1
      - name: advanced-cluster-management.v2.11.0
        version: 2.11.0
      name: release-2.11
    - currentCSV: advanced-cluster-management.v2.12.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-21T21:16:17Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <2.12.4'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:b83e630e77bd011c374d73d801f10a3230f6798ede10d7f39bf2e8ff6437cf2c
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:3d3260ff206c5ee249443b1f8da49562584f7ee1abd284691b42f1728b6ed5dc
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:b85f3de5231f2f1bc3b4e9c95917f562f39e09dc9b7db35690024163e3a5afe5
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:6f2c4f1abc4388e76414544e65f08238c96fc4d6fbf71fa79330d9dca117d221
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:ec65a5744945ca20f804d94999a4002efab9ddad360a58f9498800bffa9fa0f4
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:187879345174e515863bc1b29a961a0a0d9ad44006a6e12de715b4508dec38a8
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:93df0586f9281fa2879eed0fdf2fb115be355d572cc24dc5c5c9a92137044877
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:58f08a93cd35bd2ac19f94a1779205c81a2fc3a552d1e77b7c659fd51c9dd34f
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:bd24cfabb81b464a31c3d8c21fa6371e46b83094d5c750acc33346203fe15e84
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:c2259b90decf24f8fb1cb058d4751c75fd1558e3c9fd1079857ad9c47ca9e8d2
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:22e262c0353c6f0d5d1b916074836c53437e22697f67c85a43e6d717daf651fa
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:c402a4ac61e11dc9e45fdc7bc7fcb5bf5ca3475a66d782aea60d9f6524e51a24
        - registry.redhat.io/rhel8/postgresql-13@sha256:826a1a75f4186f36217b5c1f1aa270838c4b55b1021d33a4876566bfc3b8e629
        - registry.redhat.io/rhacm2/console-rhel9@sha256:d239df44dafac3bde3eb7ae8c80cb342fa4bb4eabe22f27d2c92d283324e6e10
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:9c297b4c1f9ba17908594f1843ea6de6c5a5e0f3500258c23fbb8b5684249c9e
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:cdaf7be69a5d66cfa4777feb8266421dafd55353d6703e7e591bb9fb23017abc
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:7c548dd869f3458fb70a536708306cb615b9166c7e866cc1215e82f183b39529
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:1419e274b4e2e4c37985208c3b1b9ecd0cdef493427ace3c71ff35ed8fcc394b
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:a5c65dee1f7e4177050244109c10e384899e01d1924b41a540373aa7daf4f274
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:c46657831915ed5f42f4a55e725bc9d17e0671191f4509e610468fda160710c0
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:1e5c12ed0de090198e097cc5deb133ad075fba224f239b8b892e6e1672ec9883
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:fff4fb0924f866b305ad0ecc5d2682da6cb20de15448a2a8c640df182719b4e2
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:28730c8f2344e51b0b9ca603ac63b60885b39be86582943bd1cc200779eee0c5
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:1ccd17b482df844ec42936993ff162f665b5a473ad344270ff2425cd63d428fc
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:6a7c54cb7e6c3cc506ffb8a4f3ec7443ef80798e2c917e20b5f06d39bde02bc4
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:7375991ce3b9e64cb7542dc176414f0974047667d08beac184e2adadd9f83532
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:3ac71b363aef8db127b88e0a63ee616d0d06342142f83522e34c9cac4dd77b2e
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:c4c3bc4f45ae194b884d89319625f2a0970dada823877a35feecd8c733cb2c8d
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:7050f71c3a038fec16db8d0eee69d7d5a4294c043c13e9cf69c3cb2e28897f4c
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:70ddf96707d398d5444bb7bbde6c684275cff62b8a83c7cf6143ff529daf0dee
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:f3a00f55dee152e0345e2616ac14f652dcb4f37f3b9dfca6f049ac5522874c6e
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:f1c371fef66e96199c4a96e21192b47bca423b282d6f99e927644840aaedfd80
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:8ccbbc058d324429aef0c31c99de8c7b3a5abc02ec6f7a14cc60eef63a8f5c69
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:7b64803ddd3b7886ad3ec3de943a38cec5aeb9604732ab7a5ea5061748bf1411
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:5156129064dbe4e9989b41018b1bb250b2eb3f02e0c0afd75a874da9b54ad3c2
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:875e4bf42ad192cfb954e38c8474a529aa938855d69dc0d83c4ee4a4777c1028
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:3faa2318f159d264a9194fc01239d3a2098c184b365b153e3c56e555d5d4992e
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:7968e96b48b2ed63e7657793cf6cc2564055809df19cb43b7e284ac03f2cc26c
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:dca845ad05b02619b81f231eb57dd826720c221d22282fe7422ceebe57191371
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:794ef15591f217be629d6c680c56e84962923993f6dd77514ca5d2bdbd96217b
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:c10a72bf8c0d76b9b4d9d7c0bfca4d83a3f7dd7f6a37b4cc24ee7591638a2482
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:08dd3f6805fa02478671927982a2c77bdec0a5b7e7fd1c9faf20c8af2bb7b428
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:fbfade039739179995828b2ce4f4411b1315ac9a0e4a98e96a115012676a859d
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:b782190b3a285003d319321be09a25ee1376c951cb0cd146cbb634c9b1ca97b3
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:8e62fa4decf01a3beae7080f19469f33822f5010a688ada5db8d2d23b008e498
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:3f286e3c24fe73b4b42e6cf4b6769f91e333014450e1ab135b1e5b4a271cbaec
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:c2ef90b4704251dc9f7ef5288f5bad7aeb4cd2b3706ec531c60f6f4346073a77
        version: 2.12.4
      entries:
      - name: advanced-cluster-management.v2.12.4
        version: 2.12.4
      - name: advanced-cluster-management.v2.12.3
        version: 2.12.3
      - name: advanced-cluster-management.v2.12.2
        version: 2.12.2
      - name: advanced-cluster-management.v2.12.1
        version: 2.12.1
      - name: advanced-cluster-management.v2.12.0
        version: 2.12.0
      name: release-2.12
    - currentCSV: advanced-cluster-management.v2.13.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-09-08T18:37:07Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <2.13.4'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:dc271885425ab74e0554b531a2078133ce02006d0bb9688f3b54a850cfefd8f8
        - registry.redhat.io/rhel8/postgresql-16@sha256:89dfeaefc7e9e974e47a3365e22f354314f8a5e906e93a0152afe4e42f6a82d0
        - registry.redhat.io/rhacm2/console-rhel9@sha256:5526edf70b784bddfc2b7e1c154a44201c346fc20117635500bb54d325921c25
        - registry.redhat.io/rhacm2/acm-flightctl-worker-rhel9@sha256:57f69874b2df77c185374ddcea530af6783410a0f35528ee1a44cc2dee6f7fa2
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:0040791500337f88af7d922895487549915faa3b69dacba1a23540e7d6f2e580
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:ba51f7d077858a0abe32a254970390b50cd7d75433b5661e17f887c48c56a30f
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:65e1f0eb4e6c797c6dc6518674b28370db3fdaa6d7e17f4233c07207224dcb48
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:3028d011e7b88dd9597821128275f606fdba639e5b83e4e6a8783fa2616cba6f
        - registry.redhat.io/rhacm2/acm-flightctl-api-rhel9@sha256:8302b536c896290533f3f50084cb9047cb55af3a4404887b70704d23862619fa
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:d29ceda561f36faa581d76d8b889f7e01d253d33c88709078855ae26cada4917
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:0272215ab43b5688295d94cce0de4490a2fbabc1dd09ed41645be7acef1e3c9b
        - registry.redhat.io/rhacm2/acm-flightctl-ocp-ui-rhel9@sha256:89863ce0eb5fb73272bda8762d74461552f8502fbe262b16cbde01064cc533b3
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:84109be1814719c343b8b4676981cdbcea0356f012b6c854221c271bdb518fea
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:fc16203de469dd1ba16e428a9d7efbb16a5f511d866ef6fcc253530de1f032a6
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:ab1e4f8685896e2016a0b4c4479378058314f365e4f09ef42cba67f3649bb306
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:b3535378919524df259b85587ca83f5620433914418c9941374951c36e0530ae
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:1b7e6e6a4ae36209f98b1539c37c5975a82363b33fe5ac1e3a320e58b99be239
        - registry.redhat.io/rhacm2/acm-flightctl-periodic-rhel9@sha256:2c9b3c3d6cd7e435da337447580dfbfde8c4932308405eeae5d97ed5ec57fbfc
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:588678bea97237d87cd2e55a0e8000e73461cfc421cc91d554a2f1e352b68d14
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:7f742f024c787932932b8a4100a5329e70aad39537e0c20c2a4e245fc6245e25
        - registry.redhat.io/rhel9/redis-7@sha256:848f4298a9465dafb7ce9790e991bd8a11de2558e3a6685e1d7c4a6e0fc5f371
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:ccc9c0524149820d701c39355af89287ad749832053ea55d1585e77347fa0ec0
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:7b15c6b492e5cb2725b9b345aa3bfae4108e2815bc233ac063b81a67c6550fb6
        - registry.redhat.io/rhel8/postgresql-13@sha256:826a1a75f4186f36217b5c1f1aa270838c4b55b1021d33a4876566bfc3b8e629
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:b1620cebc62240f93b12e6937d15760440be119c421c127214974bdfa9d08ebd
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:0d8202cd10170bb0a138d52a1d71e630ff74df466541897ba86ed324a75ea1d8
        - registry.redhat.io/openshift4/ose-cli-artifacts-rhel9@sha256:67ff9f19edac8d623ce9080a8c05cde3da73afd68e7352e4ee5b22a9ccb596bc
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:3490daccc203fade26e7df348c74a907ab0ffc8901f4965606c8227849087c44
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:f74fb6b4d42f1d7431bfc9d2424f953732fed06d2abd8d1fc27b878b4d1322d5
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:ce342242c50c103fc5e4b0126551231977b35bd477eb627fb8f5d2a6510ddaa0
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:e23115b10a92c323b4cdaf0412919c273dbca33f19910120498bd49cb5810ff5
        - registry.redhat.io/rhacm2/acm-flightctl-ui-rhel9@sha256:86800d67ec8ed63e5408947ba6a441e5afc1050b72a86487b9357674ed495a7e
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:f1db9b10481319dc142ca6046e4e3141addb4c0afab482eeaf80643fad9ad039
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:96bdb2e25ae0f381de0c94a09a3cd11f3c0bdc1d8cfb4fd5ced2cd7b24fbe2c2
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:4d068813e3eb704ea7b2d342d5a2ce8fa2ac71793b1b37634c18704ef0a5e8d7
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:5fc35a944895cf7ea6afa5929c45304aaa9f17397b4f1e42cb8c799d9aead9d5
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:f9ba60c2a415a863c6bd6030fb4663913368a9c271bb40b8acfdc294dcf3a7eb
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:8b152051ef2c4c077379e72aa510c0bd613b219254cad66fab0cd3ec0d1d4bf1
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:04bcddae590e2debf35090055dd52ae4bcaac4b369727266be53ac64670203d0
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:49e99f41cbea3faefc3ae93385a10b544ce91df0d6639cfee65760be6ae0df00
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:81cc71faaf20a986d233fee799a7d2b954cd3f43875a2689f76c53920ccf4df0
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:67939d6e4342372820d7af6777f0ffd123e86e4e8186132c3d4892cb717a5e7c
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:6f95308a508499d2d995653d3fd2def94ac5170b4c3f391d32a1e0594250c33a
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:669e9e2178476bc7a4688f3ebf4c20402608bb4b10e4df6863ec821c7e572299
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:ae1f175ffbd3698f2b3036bf9b62b5c9b24729cfb8248345b8bdd1525ccc5bb4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:ab0192a51ee37b7508f3e4db66ffa53b904dcb1f76d9aa1ac0e80447d8ff9e6a
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:5aec572fc18c8ed533620553aba2c676453297ad2440336a5228a5a4f50f5782
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:a0218b9d5f1803f60d9b295d1504a870780895cffe8f1c5b3ce4c3553da5e4f9
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:a2ab651af1c97e54d2c14de78ff554c0f25e8d000af93a74fc7b7ed913f4e338
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:72c6896a5a4dbb7717d76d9eeade96130d97909eee07e8d3cc17af0658b306ab
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:9ef9f1e2010d77211126482c03e8c72dfb7458940dc95ac952ee8dd1dedd2991
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:6ed6f62f5c05a24938335287b785b80c253424e1ee20d83311cd87c6a2090895
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:2da264ca59278f63990e579939371d45afec3bc7d9922a608520e7647a0377ac
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:1393744757874e79d1fb058cb5d3fe7dcbbcd2f187bdcdcdf34e67227c61fbb9
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:296a31e1aad9f6288bdf80fb3191642e3dd404e9e8fcad198016bedbd3abe9b4
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:3cbb9d271c2f23d430a623c83606bbc8a80379c24d13e610886ad51fd21ce1e4
        version: 2.13.4
      entries:
      - name: advanced-cluster-management.v2.13.4
        version: 2.13.4
      - name: advanced-cluster-management.v2.13.3
        version: 2.13.3
      - name: advanced-cluster-management.v2.13.2
        version: 2.13.2
      - name: advanced-cluster-management.v2.13.1
        version: 2.13.1
      - name: advanced-cluster-management.v2.13.0
        version: 2.13.0
      name: release-2.13
    - currentCSV: advanced-cluster-management.v2.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-25T02:10:33Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.13.0 <2.14.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          repository: https://github.com/open-cluster-management/multiclusterhub-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MulticlusterHub defines the configuration
              for an instance of a multicluster hub, a central point for managing multiple
              Kubernetes-based clusters. The deployment of multicluster hub components
              is determined based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n### Where to find more installation
          information\nYou can find additional installation guidance in the [install
          documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:364c7416a9642dfb37f05e2eac4427918959b12b159ac90a9ca310c584d6434a
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:69f0b9a7fb494572adf2795be865d6d6ba4b67db0eed0e9d61f5dcb5e9bf6efe
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:c2ef90b4704251dc9f7ef5288f5bad7aeb4cd2b3706ec531c60f6f4346073a77
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:0e1eba23e0e8454f2dec240704e7ccc922351d58880f72514d3b2bc30c9e995e
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:c6ff6cae5bc6447d0ed24e458fc5d5fc709b2106fe93b143f9546bd962ea7f87
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:ad7f30c61bb11c0d32686e75b01f3b286a1fb06edd8d2ef6a9e2d924fba2b565
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:b489229190f9a3da66bd154e65490c3a35d1999d2db6707759f06eadb02b968d
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:8e6ee423acfafe28c002911e832d47dffc00a5227325bd6a5b1ab359c7eee5f4
        - registry.redhat.io/rhacm2/console-rhel9@sha256:d4c42a203110050a3834be8d5a0eee384ad7e9fbb991da589d75acd15055b8b1
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:0cecc3d87afcd0d63bdeae3ecae37f8a38534fbdc8bad5c548e0d4f4fb0156a4
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:8f4f59e237f366f08367cabbea3eb51e766ad52db0d35c5efc56182c13fc5c09
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:8253a9225f30b0f0a405e2c3293924acbfbf81f3d9c50bc4431fe4052a9fe908
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:55ef2c163b0c0deb3402e39acf37ff6bb05e2a5d268dcae5f8a400128d67c584
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:b7b7581e7743148d19283250d3300d292b96c48ecae96f6e6202a7ca8204e2ab
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:c329cfd004c58a7fcc40f4fbb986399dc2fe08d7922d2b4f210664f58751b1a6
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:143f49d2ce1f58282e276396722ba37165cd1bf2051d537bb4e6208786e5aed6
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:53deacec64238607afda29c27463d72ac3b582c9670e2e77a9c7e48851214795
        - registry.redhat.io/openshift4/ose-cli-rhel9@sha256:a48758b69b9a7451b7e0caea465a181f854b41f1e74ce86c15216ee1bb31fcb9
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:f9a9e10fde9065d650a7dc8eba83349e5db0a21727027464cf144724d0de8cfa
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:935baf67627d5e76c0766d5d2a9e36b398a12535f097749a987dd0c5de788570
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:0252cf95019717f03c7ac72fb034b5bd9ad7aef2da0a92dd21c56cf47f03b812
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:d56b997a9084d56c63f539147052e64ba30258afe91a9ff832f3673cbccbe467
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ee742e7baa8bcbbfccf79abcf15d22347d14f280eafaf22c42cef71051b10fba
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:0ab21d4189669901fe085c7f3a4b4c9778c853f8abd706bbfffcb59942fbc8e6
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:9c79b1153fc927e7582c5346ea0408e03f794eb983597f02a5e816ae961ba86b
        - registry.redhat.io/rhem/flightctl-periodic-rhel9@sha256:ee59f307e468e2b7b38f7c040898d587d441f8fb3d59cb022ae23e4717739e3f
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:2b1b1c707e0b5217c66a96fb82070de47fc0342747081d4be0049c4542944c8a
        - registry.redhat.io/rhel8/postgresql-16@sha256:d4abbd9492347e0aa7e1b0ff967e74d4b6a4edd6fd20d896f2f2ae03c623181c
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:46c5835a4bfc0e4767d0959bc47e939c195be197e9715aae17788d4d0e799d56
        - registry.redhat.io/rhel9/redis-7@sha256:ff31822cba556e40b6b00720839cc56f474fdb9257fb39b7604fafc11761d379
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:0ea80a1039d217e899ea012b77e99a7c5832e5b7e292dca473759c449bf2d1ea
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:e063c54c7182b70d78d24b5026f174436885383980c9b1e3e7664fa2adfad883
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:d550013d385552d6242b56c5b60f15b3ebb488c2033f808de54bcc639abc2f30
        - registry.redhat.io/rhem/flightctl-ui-ocp-rhel9@sha256:973eb3ca3da32c87ba308928a21fbd25514b2f6a6662300d4928479cb554cff3
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:bcb5949ebc31bbf6b45dfd34aae9c80c82a08531e8a3d9f6eef3354f942a5a5d
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:cee29bdb03927dc1866c267232d391ea5ee762b23c6b04a777e7d61d14a61bba
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:0cc23b5063c0d3264ed345dad8453901e4db18cc9bed1656f2799f72de75bd53
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:8151a89ebfebf2e6d336849fdd329b123b7f486938808ee80bf6bd8d44f6985c
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:7ee0aea84bf00b7b42027fa3e970c1f4b279e345a698b4050c9f413eb0383bb1
        - registry.redhat.io/rhem/flightctl-ui-rhel9@sha256:940c0564ced75b7912d3ffc923a1f6bab9aed0f4a52962f7f894d08937b7d37e
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:ce5ff191023b5e14c339b2a03aa1be05a12d024efa5302652065597e630996b6
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:90d399736ca81bb244415efaf05738beff806912a6ac7444a7d4839089b7a081
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:556d737195a4878877c650a93619af1528e52694f2a8e47e13463a603d2f4900
        - registry.redhat.io/rhem/flightctl-cli-artifacts-rhel9@sha256:3491ec932d90112e89f3a4cebe5bdd9ed6a7811f116df43bab8bb5f827129361
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:d9fa60b5b1d6b6c94c5d8b610c9a7b979202de8b46a344e0a65d06f6ef21a7ef
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:40b3e063435eea6e526a57d6bd1ae489ca703e81eeca82dda1c2e6aedec1262c
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:95085fc33a0150a39b6ef4e616fd23f0bddd9dcad0e75c4a89ed27113626e6eb
        - registry.redhat.io/rhel9/memcached@sha256:955138db9a7a23ee31198aabebd0811438fee11bb0b624c38ddb90efe52bd07c
        - registry.redhat.io/rhem/flightctl-worker-rhel9@sha256:1b1f6f85af0910f1acd029f54a608b25d3a2977c3d8b323266391543324305b4
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:f3e34b200e9480363c56da1cd88b2de37fc6a490691e4596a132217f8ea286d8
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:c70242a037e36690d8dd1030254dde778917356b4f1f61db2cecb03b6c2d7380
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:cb5346afe1c7b58eb758cdced94f2c1161c1621033b52f29757efe3a50320a65
        - registry.redhat.io/rhem/flightctl-api-rhel9@sha256:83c6c7819e04aa58d961518367ce6ad3159a28300ca097967f630044d181be34
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:8e57af77cc5dee3aaa83f6ae6e62e232c5fc4fbe4dca8eb755226351963058e3
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:44e305ca15e7c9b50bd7fa8bb391f5ee66906aa26ecc5f612d029aaee5bb4cb8
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:d65aef370a7ed01c0c7cc0ad7f0800359db7863db4ec8845de5b450992f26933
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:ad534f7b0efea45b9f4a0a026c76a2a030cb608acdb835dfea3852332acfcae1
        version: 2.14.0
      entries:
      - name: advanced-cluster-management.v2.14.0
        version: 2.14.0
      name: release-2.14
    defaultChannel: release-2.14
    packageName: advanced-cluster-management
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.17.40-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5-beta.1 <8.17.40-beta.1'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBM Turbonomic software automates and simultaneously optimizes
              performance, compliance, and cost in real time. Turbonomic discovers
              and automates resource optimization for the complete application stack.
              Applications are continually resourced to perform while satisfying technical
              and business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
          - kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Optimization for Kubernetes
          IBM Turbonomic software automates and simultaneously optimizes performance, compliance, and cost in real time.
          Turbonomic discovers and automates resource optimization for the complete application stack. Applications are continually resourced to perform while satisfying technical and business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (proactive rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          Turbonomic assures application performance by giving workloads the resources they need when they need them.

          ### How Does It Work?
          Turbonomic uses a client agent called — KubeTurbo — that runs in your cluster to discover, monitor and optimize resource usage within your environment.
          KubeTurbo runs together with the default Kubernetes scheduler and sends data back to the Turbonomic for analysis.
          Turbonomic determines the right actions that drive continuous health for your workloads and your cluster.

          Turbonomic discovers Red Hat OpenShift Virtualization resources through a new or existing target configuration with Red Hat OpenShift. For more information refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=targets-red-hat-openshift-virtualization)

          ### Get Started
          Please follow the [Kubeturbo Documentation Page](https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets) for all deployment options and an overview of typical use cases.

          ### Known Issues
          For known issues refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=notes-known-issues)
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Kubeturbo Documentation Page
          url: https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets
        - name: Kubeturbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
        - name: IBM Turbonomic Application Resource Management
          url: https://www.ibm.com/products/turbonomic
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/kubeturbo-operator@sha256:7111c7c4002a447dc6977dbb73e96507bae56766fc0552fb2d036b20583722ba
        version: 8.17.40-beta.1
      entries:
      - name: kubeturbo-operator.v8.17.40-beta.1
        version: 8.17.40-beta.1
      name: beta
    - currentCSV: kubeturbo-operator.v8.17.40
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5 <8.17.40'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBM Turbonomic software automates and simultaneously optimizes
              performance, compliance, and cost in real time. Turbonomic discovers
              and automates resource optimization for the complete application stack.
              Applications are continually resourced to perform while satisfying technical
              and business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
          - kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Optimization for Kubernetes
          IBM Turbonomic software automates and simultaneously optimizes performance, compliance, and cost in real time.
          Turbonomic discovers and automates resource optimization for the complete application stack. Applications are continually resourced to perform while satisfying technical and business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (proactive rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          Turbonomic assures application performance by giving workloads the resources they need when they need them.

          ### How Does It Work?
          Turbonomic uses a client agent called — KubeTurbo — that runs in your cluster to discover, monitor and optimize resource usage within your environment.
          KubeTurbo runs together with the default Kubernetes scheduler and sends data back to the Turbonomic for analysis.
          Turbonomic determines the right actions that drive continuous health for your workloads and your cluster.

          Turbonomic discovers Red Hat OpenShift Virtualization resources through a new or existing target configuration with Red Hat OpenShift. For more information refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=targets-red-hat-openshift-virtualization)

          ### Get Started
          Please follow the [Kubeturbo Documentation Page](https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets) for all deployment options and an overview of typical use cases.

          ### Known Issues
          For known issues refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=notes-known-issues)
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Kubeturbo Documentation Page
          url: https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets
        - name: Kubeturbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
        - name: IBM Turbonomic Application Resource Management
          url: https://www.ibm.com/products/turbonomic
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/kubeturbo-operator@sha256:829dc6b9800efd3fd77d07f955a53610c7b2888b93495b2db32f1f48e93a05a6
        version: 8.17.40
      entries:
      - name: kubeturbo-operator.v8.17.40
        version: 8.17.40
      name: stable
    defaultChannel: stable
    packageName: kubeturbo-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://access.redhat.com/products/red-hat-integration
    name: red-hat-camel-k
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: red-hat-camel-k-operator.v1.10.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
          createdAt: "2025-01-08T17:00:43Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
        version: 1.10.10
      entries:
      - name: red-hat-camel-k-operator.v1.10.10
        version: 1.10.10
      - name: red-hat-camel-k-operator.v1.10.9
        version: 1.10.9
      - name: red-hat-camel-k-operator.v1.10.9-0.1743522834.p
        version: 1.10.9+0.1743522834.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1742828111.p
        version: 1.10.9+0.1742828111.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741884247.p
        version: 1.10.9+0.1741884247.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741875993.p
        version: 1.10.9+0.1741875993.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741580860.p
        version: 1.10.9+0.1741580860.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739988175.p
        version: 1.10.9+0.1739988175.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739388437.p
        version: 1.10.9+0.1739388437.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739387844.p
        version: 1.10.9+0.1739387844.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.10.x
    - currentCSV: red-hat-camel-k-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1beta1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:9e4fb5369e0479d4a68f60725e7d12abdc66c552babe59f1358f6e6b0f26357d
          createdAt: "2020-10-06T02:45:00Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:9e4fb5369e0479d4a68f60725e7d12abdc66c552babe59f1358f6e6b0f26357d
        version: 1.4.2
      entries:
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.4.x
    - currentCSV: red-hat-camel-k-operator.v1.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
          createdAt: "2022-08-15T20:17:25Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <=1.6.8'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
        version: 1.6.10
      entries:
      - name: red-hat-camel-k-operator.v1.6.10
        version: 1.6.10
      - name: red-hat-camel-k-operator.v1.6.9
        version: 1.6.9
      name: 1.6.x
    - currentCSV: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:e916f9040d3f4feb26cc1534d3ab2722890b34c7fe146cdce8b6085b47206e22
          createdAt: "2022-11-28T10:43:18Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <1.8.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:e916f9040d3f4feb26cc1534d3ab2722890b34c7fe146cdce8b6085b47206e22
        version: 1.8.2+0.1675913507.p
      entries:
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.8.x
    - currentCSV: red-hat-camel-k-operator.v1.10.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
          createdAt: "2025-01-08T17:00:43Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
        version: 1.10.10
      entries:
      - name: red-hat-camel-k-operator.v1.10.10
        version: 1.10.10
      - name: red-hat-camel-k-operator.v1.10.9
        version: 1.10.9
      - name: red-hat-camel-k-operator.v1.10.9-0.1743522834.p
        version: 1.10.9+0.1743522834.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1742828111.p
        version: 1.10.9+0.1742828111.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741884247.p
        version: 1.10.9+0.1741884247.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741875993.p
        version: 1.10.9+0.1741875993.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741580860.p
        version: 1.10.9+0.1741580860.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739988175.p
        version: 1.10.9+0.1739988175.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739388437.p
        version: 1.10.9+0.1739388437.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739387844.p
        version: 1.10.9+0.1739387844.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: candidate
    - currentCSV: red-hat-camel-k-operator.v1.10.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
          createdAt: "2025-01-08T17:00:43Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
        version: 1.10.10
      entries:
      - name: red-hat-camel-k-operator.v1.10.10
        version: 1.10.10
      - name: red-hat-camel-k-operator.v1.10.9
        version: 1.10.9
      - name: red-hat-camel-k-operator.v1.10.9-0.1743522834.p
        version: 1.10.9+0.1743522834.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1742828111.p
        version: 1.10.9+0.1742828111.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741884247.p
        version: 1.10.9+0.1741884247.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741875993.p
        version: 1.10.9+0.1741875993.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741580860.p
        version: 1.10.9+0.1741580860.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739988175.p
        version: 1.10.9+0.1739988175.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739388437.p
        version: 1.10.9+0.1739388437.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739387844.p
        version: 1.10.9+0.1739387844.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: latest
    - currentCSV: red-hat-camel-k-operator.v1.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
          createdAt: "2022-08-15T20:17:25Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <=1.6.8'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
        version: 1.6.10
      entries:
      - name: red-hat-camel-k-operator.v1.6.10
        version: 1.6.10
      - name: red-hat-camel-k-operator.v1.6.9
        version: 1.6.9
      name: stable
    defaultChannel: 1.10.x
    packageName: red-hat-camel-k
    provider:
      name: Red Hat
      url: https://access.redhat.com/products/red-hat-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MySQL
      provider-url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
    name: ndb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ndb-operator.v1.6.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "mysql.oracle.com/v1","kind": "NdbCluster","metadata":
            {"name": "example-ndb"},"spec": {"redundancyLevel": 2,"dataNode": {"nodeCount":
            2},"mysqlNode": {"nodeCount":2}}}]'
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: container-registry.oracle.com/mysql/community-ndb-operator@sha256:b562dc3303ffae9f4dab1d73fd0fe05833bdaa8b731af4227da81af617c75c6b
          createdAt: "2024-12-18T15:05:46.000Z"
          description: The MySQL NDB Operator is a Kubernetes operator for managing
            a MySQL NDB Cluster setup inside a Kubernetes Cluster.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Ndb Cluster
            displayName: NdbCluster
            kind: NdbCluster
            name: ndbclusters.mysql.oracle.com
            version: v1
        description: |
          ## NdbCluster Custom Resource
          The NdbCluster custom resource is defined with custom fields that map to the configuration of a MySQL Cluster. To deploy a MySQL Cluster inside the K8s Cluster, an NdbCluster object with the desired configuration of the MySQL Cluster has to be created in the K8s Cluster. The NDB Operator watches for any changes made to any NdbCluster objects and will receive an event when the NdbCluster object is created in the K8s Cluster. It will then start deploying the MySQL Cluster, with the specified configuration, inside the K8s Cluster using native Kubernetes workloads and resources. Once the MySQL Cluster is ready, the NdbCluster resource can be further modified to make changes to the MySQL Cluster configuration. The Operator will receive this event as well and will update the configuration of the already running MySQL Cluster.

          The documentation at [NdbCluster-CRD.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md) explains the various custom fields of the NdbCluster CRD and the effect they have on the MySQL Cluster configuration. The [examples](https://github.com/mysql/mysql-ndb-operator/tree/main/docs/examples) folder has a few examples of the NdbCluster resource objects which can be used to deploy MySQL Cluster inside the K8s Cluster.
          ## Deploy a simple MySQL Cluster

          The [examples/example-ndb.yaml](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/examples/example-ndb.yaml) file has a NdbCluster object example-ndb that defines a simple MySQL Cluster with 2 data nodes and 2 MySQL Servers. To create this object in the default namespace of the K8s Cluster, run this command from the root of the repo :
          ```
          kubectl apply -f docs/examples/example-ndb.yaml
          ```

          Various NdbCluster status fields report back the actual status of the MySQL Cluster running inside the K8s Cluster. Wait for the MySQL Cluster nodes to be started and become ready using the [UpToDate](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md#ndbclusterconditiontypestring-alias) Condition defined in the NdbCluster object status :
          ```
          kubectl wait --for=condition=UpToDate ndb example-ndb --timeout=10m
          ```
          The MySQL Cluster is ready for transactions once the NdbCluster is UpToDate, i.e. the UpToDate condition becomes true.

          At any point, a brief status of an NdbCluster object can be viewed using kubectl get ndb command.
          ```
          kubectl get ndb example-ndb
          ```

          The output of the command gives a short summary the MySQL Cluster nodes controlled by the respective NdbCluster object.
          ```
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE   UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:2/2    Ready:2/2 3m    True
          ```

          To list all the pods created by the NDB Operator, run :
          ```
          kubectl get pods -l mysql.oracle.com/v1=example-ndb
          ```

          For example-ndb NdbCluster object, two management nodes (*-mgmd-* pods), two multi-threaded data nodes (*-ndbmtd-* pods) and two MySQL Servers (*-mysqld-*) would have been created by the NDB operator :
          ```
          NAME                   READY   STATUS    RESTARTS   AGE
          example-ndb-mgmd-0     1/1     Running   0          3m53s
          example-ndb-mgmd-1     1/1     Running   0          2m5s
          example-ndb-mysqld-0   1/1     Running   0          78s
          example-ndb-mysqld-1   1/1     Running   0          78s
          example-ndb-ndbmtd-0   1/1     Running   0          3m53s
          example-ndb-ndbmtd-1   1/1     Running   0          3m53s
          ```
          If running inside a openshift cluster, pods from statefulset "example-ndb-mysqld" will not start up. By default openshift runs all the pod with "restricted" scc, but the statefulset "example-ndb-mysqld" required higher privilege. Use the below command to check the exact scc required
          ```
          oc get pods example-ndb-mysqld-0 -o yaml | oc adm policy scc-subject-review -f -
          ```
          Create a user account with the scc and set the user to the statefulset "example-ndb-mysqld". Below commands creates an user with "anyuid" and assign it to the statefulset.
          ```
          oc create sa any-uid-user
          oc adm policy add-scc-to-user anyuid -z any-uid-user
          oc set sa sts example-ndb-mysqld any-uid-user
          ```
          ## Connect to the MySQL Cluster

          The NDB Operator, by default, creates few Services to expose the services offered by the MySQL Cluster nodes within the Kubernetes Cluster.
          ```
          kubectl get services -l mysql.oracle.com/v1=example-ndb
          ```

          The above command will generate an output similar to :
          ```
          NAME                 TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S) AGE
          example-ndb-mgmd     ClusterIP   10.100.207.255   <none>        1186/TCP 5m
          example-ndb-mysqld   ClusterIP   10.103.182.157   <none>        3306/TCP 5m
          example-ndb-ndbmtd   ClusterIP   None             <none>        1186/TCP 5m
          ```

          The NDB Operator will also create a root user with a random password on all the MySQL Servers. The root password will be stored in a K8s Secret which can be retrieved from the NdbCluster object status :
          ```
          kubectl get ndb example-ndb -o jsonpath={.status.generatedRootPasswordSecretName}
          ```

          To extract and decode the password from the Secret :
          ```
          base64 -d <<< \
          $(kubectl get secret example-ndb-mysqld-root-password \
          -o jsonpath={.data.password})
          ```

          One can also specify the password and host to be used with the root account via the NdbCluster spec. See the [CRD documentation](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md#ndbmysqldspec) for more details on this.
          ## Access MySQL Cluster from inside K8s

          To connect to the Management and MySQL Servers from within the Kubernetes Cluster, you can use the example-ndb-mgmd and example-ndb-mysqld services as the MySQL Cluster connectstring and MySQL host respectively. You can also use the mgmd and mysqld pods' hostnames to directly connect to that particular pod. The mysql, ndb_mgm clients and any other ndb tools will work as expected.

          A demonstration is available at [connect-from-inside-k8s-demo.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/connect-from-inside-k8s-demo.md).
          ## Access MySQL Cluster from outside K8s

          By default, the Management and MySQL services created by the NDB Operator are of type [ClusterIP](https://kubernetes.io/docs/concepts/services-networking/service/#publishing-services-service-types) and are accessible only from within the K8s Cluster. To expose them outside the K8s Cluster, the services have to be upgraded into a LoadBalancer type. This can be done by setting the spec.managementNode.enableLoadBalancer and spec.mysqlNode.enableLoadBalancer fields of the NdbCluster resource to true. These options can be enabled when the NdbCluster resource object is created or via an update to the object when the MySQL Cluster is already running.

          Once enabled, the previously ClusterIP type example-ndb-mgmd and example-ndb-mysqld services will be upgraded to LoadBalancer type, and they will be available at the external IP assigned to them by the cloud provider.

          Another way to access these services without enabling the LoadBalancers support is to use the kubectl port-forward command.

          In both ways, only the mysql and ndb_mgm clients work. Any NDB tool which uses the NDBAPI to connect to the MySQL data nodes will not work as expected from outside the K8s Cluster.

          A demonstration is available at [connect-from-outside-k8s-demo.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/connect-from-outside-k8s-demo.md).
          ## Updating MySQL Cluster Configuration

          Once a MySQL Cluster has been deployed by the NDB Operator inside the K8s Cluster, its configuration can be updated by updating the spec of the NdbCluster resource object that represents it. The NDB Operator will pick up any changes made to the NdbCluster resource object and will apply the updated configuration to the respective MySQL Cluster. The NDB Operator takes care of updating the MySQL Cluster config file and restarting all the management/data nodes if required. When the update is being handled by the NdbOperator, it will set the UpToDate condition of the NdbCluster resource object to false and will not accept any further updates until it completes the current one.

          The kubectl get ndb example-ndb command will report the UpToDate status of an NdbCluster. It can be used to check if the current update has been completed.
          ```
          $ kubectl get ndb example-ndb
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE   UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:1/2    Ready:2/2 10m   False
          ```

          The UpToDate condition can also be used to wait for the operator to complete the update.
          ```
          kubectl wait --for=condition=UpToDate ndb example-ndb --timeout=10m
          ```

          Once the update is complete, the UpToDate condition will be set back to true by the operator.
          ```
          $ kubectl get ndb example-ndb
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE      UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:2/2    Ready:2/2 10m50s   True
          ```

          ## Delete a MySQL Cluster

          To stop and remove the MySQL Cluster running inside the K8s Cluster, delete the NdbCluster resource object.
          ```
          kubectl delete -f docs/examples/example-ndb.yaml
          ```

          (or)
          ```
          kubectl delete ndb example-ndb
          ```

          This will delete all the data, the pods running the MySQL Cluster nodes and also delete all other associated K8s resources created by the NDB Operator.
        displayName: Ndb Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - ndb
        - in-memory
        links:
        - name: Document
          url: https://dev.mysql.com/doc/ndb-operator/en/
        maintainers:
        - email: paulo.p.silva@oracle.com
          name: paulofasilva
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: MySQL
          url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
        relatedImages:
        - container-registry.oracle.com/mysql/community-ndb-operator@sha256:b562dc3303ffae9f4dab1d73fd0fe05833bdaa8b731af4227da81af617c75c6b
        version: 1.6.2
      entries:
      - name: ndb-operator.v1.6.2
        version: 1.6.2
      - name: ndb-operator.v1.6.1
        version: 1.6.1
      - name: ndb-operator.v1.6.0
        version: 1.6.0
      name: alpha
    defaultChannel: alpha
    packageName: ndb-operator
    provider:
      name: MySQL
      url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: VASTData
      provider-url: https://www.vastdata.com
    name: vast-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vast-csi-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"storage.vastdata.com/v1","kind":"VastCSIDriver","metadata":{"name":"csi.vastdata.com"},"spec":{"applySecurityContextConstraints":true,"attachRequired":true,"controller":{"dnsPolicy":"Default","extraArgs":{"csiAttacher":[],"csiProvisioner":[],"csiResizer":[],"csiSnapshotter":[]},"nodeAffinity":{},"nodeSelector":{},"podAffinity":{},"podAntiAffinity":{},"resources":{"csiAttacher":{"limits":{"memory":"200Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiProvisioner":{"limits":{"memory":"400Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiResizer":{"limits":{"memory":"300Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiSnapshotter":{"limits":{"memory":"200Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiVastPlugin":{"limits":{"memory":"400Mi"},"requests":{"cpu":"100m","memory":"50Mi"}}},"runOnControlPlane":false,"runOnMaster":false,"tolerations":[]},"deletionViewPolicy":null,"deletionVipPool":null,"dontUseTrashApi":false,"driverType":"nfs","endpoint":null,"image":{"csiAttacher":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-attacher@sha256:b4d611100ece2f9bc980d1cb19c2285b8868da261e3b1ee8f45448ab5512ab94"},"csiNodeDriverRegistrar":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac"},"csiProvisioner":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-provisioner@sha256:405a14e1aa702f7ea133cea459e8395fe40a6125c088c55569e696d48e1bd385"},"csiResizer":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-resizer@sha256:a541e6cc2d8b011bb21b1d4ffec6b090e85270cce6276ee302d86153eec0af43"},"csiSnapshotter":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-snapshotter@sha256:2e04046334baf9be425bb0fa1d04c2d1720d770825eedbdbcdb10d430da4ad8c"},"csiVastPlugin":{"imagePullPolicy":"IfNotPresent","repository":"registry.connect.redhat.com/vastdataorg/csi@sha256:7169f279d0dd7130535d74df5b938ed6b7944bfd4fe52d242ceee5178f2ea156"}},"imagePullSecrets":[],"kubeletPath":"/var/lib/kubelet","logLevel":5,"node":{"dnsPolicy":"Default","nodeAffinity":{},"nodeSelector":{},"podAffinity":{},"podAntiAffinity":{},"propagateHostMountOptions":true,"resources":{"csiVastPlugin":{"limits":{"memory":"400Mi"},"requests":{"cpu":"100m","memory":"50Mi"}},"nodeDriverRegistrar":{"limits":{"memory":"100Mi"},"requests":{"cpu":"10m","memory":"20Mi"}}},"tolerations":[]},"numWorkers":10,"operationRetryIntervalMax":60,"operationRetryIntervalStart":10,"operationTimeout":15,"secretName":null,"sslCertsSecretName":null,"truncateVolumeName":64,"useLocalIpForMount":null,"verifySsl":false}},{"apiVersion":"storage.vastdata.com/v1","kind":"VastStorage","metadata":{"name":"vastdata-filesystem"},"spec":{"allowVolumeExpansion":true,"blockingClones":false,"clusterName":"","createSnapshotClass":true,"driverType":"nfs","ephemeralVolumeNameFormat":"csi:{namespace}:{name}:{id}","fsType":"","hostEncryption":null,"mountOptions":[],"provisioner":"","qosPolicy":"","reclaimPolicy":"Delete","secretName":"","secretNamespace":"","setDefaultStorageClass":false,"snapshotClass":{"deletionPolicy":"Delete","setDefaultSnapshotClass":false,"snapshotNameFormat":"csi:{namespace}:{name}:{id}"},"storagePath":"","subsystem":"","tenantName":"","viewPolicy":"","vipPool":"","vipPoolFQDN":null,"volumeGroup":"","volumeNameFormat":"csi:{namespace}:{name}:{id}"}},{"apiVersion":"storage.vastdata.com/v1","kind":"VastCluster","metadata":{"name":"vastcluster-sample"},"spec":{"endpoint":null,"passphrase":null,"password":null,"sslCert":null,"tenant":null,"token":null,"username":null}}]'
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/vastdataorg/csi-operator@sha256:ba7f30a42e327004ecbc81ff5ccc453bf3046a29bab5785b9f8a9746e5c6a3fd
          createdAt: "2024-07-01T11:59:59Z"
          description: |-
            VAST CSI Operator supports seamless installation and upgrade of VAST Data's CSI Driver.
            The VAST CSI Driver allows container orchestration frameworks such as Kubernetes to dynamically provision, manage and scale volumes on the [VAST Data Platform](https://www.vastdata.com/platform/overview). VAST provides scalable, reliable, and fast persistent storage that can be accessed remotely by any Kubernetes application container.
            ### Installation
            Refer to the VASTData Operator for Kubernetes [official documentation](https://support.vastdata.com/s/document-item?bundleId=vast-csi-driver-2.6-administrator-s-guide&topicId=deploying-vast-csi-driver/deploying-vast-csi-driver-on-openshift-with-vast-csi-operator.html&_LANG=enus).
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: vast-csi
          operators.operatorframework.io/builder: operator-sdk-v1.3.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/vast-data/vast-csi
          support: VastData Infra Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VastCSIDriver is a common specification for VAST CSI Controller
              and VAST CSI Node plugins, typically intended for creation in a single
              instance.
            displayName: VastCSIDriver
            kind: VastCSIDriver
            name: vastcsidrivers.storage.vastdata.com
            version: v1
          - description: VastCluster encapsulates the connection details required
              for accessing a VAST cluster. This information is ultimately stored
              in a Kubernetes Secret.
            displayName: VastCluster
            kind: VastCluster
            name: vastclusters.storage.vastdata.com
            version: v1
          - description: Represents VAST storage. Each VastStorage instance generates
              a corresponding storage class that can be used in PVCs to create workloads.
            displayName: VastStorage
            kind: VastStorage
            name: vaststorages.storage.vastdata.com
            version: v1
        description: |-
          VAST CSI Operator supports seamless installation and upgrade of VAST Data's CSI Driver.
          The VAST CSI Driver allows container orchestration frameworks such as Kubernetes to dynamically provision, manage and scale volumes on the [VAST Data Platform](https://www.vastdata.com/platform/overview). VAST provides scalable, reliable, and fast persistent storage that can be accessed remotely by any Kubernetes application container.
          ### Installation
          Refer to the VASTData Operator for Kubernetes [official documentation](https://support.vastdata.com/s/document-item?bundleId=vast-csi-driver-2.6-administrator-s-guide&topicId=deploying-vast-csi-driver/deploying-vast-csi-driver-on-openshift-with-vast-csi-operator.html&_LANG=enus).
        displayName: VAST CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - VAST
        - VASTData
        - CSI
        - Multi Cluster
        - AI
        - High-Performance Computing
        links:
        - name: VAST CSI Plugin
          url: https://github.com/vast-data/vast-csi
        maintainers:
        - email: volodymyr.boiko@vastdata.com
          name: VAST Infrastructure Team
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: VASTData
          url: https://www.vastdata.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.connect.redhat.com/vastdataorg/csi-operator@sha256:ba7f30a42e327004ecbc81ff5ccc453bf3046a29bab5785b9f8a9746e5c6a3fd
        - registry.connect.redhat.com/vastdataorg/csi@sha256:7169f279d0dd7130535d74df5b938ed6b7944bfd4fe52d242ceee5178f2ea156
        - registry.k8s.io/sig-storage/csi-resizer@sha256:a541e6cc2d8b011bb21b1d4ffec6b090e85270cce6276ee302d86153eec0af43
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:2e04046334baf9be425bb0fa1d04c2d1720d770825eedbdbcdb10d430da4ad8c
        - registry.k8s.io/sig-storage/csi-attacher@sha256:b4d611100ece2f9bc980d1cb19c2285b8868da261e3b1ee8f45448ab5512ab94
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:405a14e1aa702f7ea133cea459e8395fe40a6125c088c55569e696d48e1bd385
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac
        version: 2.6.3
      entries:
      - name: vast-csi-operator.v2.6.3
        version: 2.6.3
      name: stable
    defaultChannel: stable
    packageName: vast-csi-operator
    provider:
      name: VASTData
      url: https://www.vastdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: ""
    name: power-monitoring-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: power-monitoring-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "Kepler",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "kepler",
                    "app.kubernetes.io/name": "kepler",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "kepler"
                },
                "spec": {
                  "exporter": {
                    "deployment": {
                      "port": 9103
                    },
                    "redfish": null
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: registry.redhat.io/openshift-power-monitoring/power-monitoring-rhel9-operator@sha256:5eec15f6fe9bd4bdb1f5fb8057dc67c448c3711bcc7a0a4760f0d90c07688d1e
          createdAt: "2025-02-14T06:23:56Z"
          description: Deploys and Manages Kepler on OpenShift
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/internal-objects: |-
            [
              "keplerinternals.kepler.system.sustainable.computing.io"
            ]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift-power-monitoring/power-monitoring-operator
          support: Team Power Monitoring (team-powermon@redhat.com)
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KeplerInternal is the Schema for internal/unsupported API
            displayName: KeplerInternal
            kind: KeplerInternal
            name: keplerinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: Kepler is the Schema for the keplers API
            displayName: Kepler
            kind: Kepler
            name: keplers.kepler.system.sustainable.computing.io
            version: v1alpha1
        description: |
          ### Introduction
          Kepler (Kubernetes Efficient Power Level Exporter) uses eBPF to probe
          energy-related system stats and exports them as Prometheus metrics.

          ### NOTE
          To view the metrics exposed by Kepler, ensure that
          [User Project Monitoring](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/monitoring/configuring-user-workload-monitoring)
          is enabled in your cluster.

          ### Documentation
          Documentation and installation guide can be found below:
            * [Installation Guide](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/power_monitoring/installing-power-monitoring)
        displayName: Power monitoring for Red Hat OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kepler
        - power
        - energy
        - monitoring
        links:
        - name: Installation Guide
          url: https://docs.redhat.com/documentation/openshift_container_platform/latest/html/power_monitoring/installing-power-monitoring
        maintainers:
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: vimalkum@redhat.com
          name: Vimal Kumar
        - email: vprashar@redhat.com
          name: Vibhu Prashar
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift-power-monitoring/power-monitoring-rhel9-operator@sha256:5eec15f6fe9bd4bdb1f5fb8057dc67c448c3711bcc7a0a4760f0d90c07688d1e
        - registry.redhat.io/openshift-power-monitoring/kepler-rhel9@sha256:27d3cb0a98f1cffd649745e3886ad92995d83a0de6823ddc2d770bf036e2eeaf
        version: 0.4.0
      entries:
      - name: power-monitoring-operator.v0.4.0
        version: 0.4.0
      - name: power-monitoring-operator.v0.3.0
        version: 0.3.0
      - name: power-monitoring-operator.v0.2.0
        version: 0.2.0
      - name: power-monitoring-operator.v0.1.0
        version: 0.1.0
      name: tech-preview
    defaultChannel: tech-preview
    packageName: power-monitoring-operator
    provider:
      name: Red Hat, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtc-operator.v1.7.19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigrationController",
                  "metadata": {
                    "name": "migration-controller",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "azure_resource_group": "",
                    "cluster_name": "host",
                    "migration_velero": true,
                    "migration_controller": true,
                    "migration_ui": true,
                    "migration_log_reader": true,
                    "olm_managed": true,
                    "restic_timeout": "1h",
                    "mig_pv_limit": "100",
                    "mig_pod_limit": "100",
                    "mig_namespace_limit": "10"
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageMigration",
                  "metadata": {
                    "name": "directimagemigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageStreamMigration",
                  "metadata": {
                    "name": "directimagestreammigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigAnalytic",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "plan": {
                      "name": "foo",
                      "namespace": "openshift-migration"
                    },
                    "analyzeK8SResources": true,
                    "analyzeImageCount": true,
                    "analyzePVCapacity": true
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigCluster",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectVolumeMigration",
                  "metadata": {
                    "name": "direct",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                      "kind": "DirectVolumeMigrationProgress",
                      "metadata": {
                        "name": "direct-progress",
                        "namespace": "openshift-migration"
                      },
                      "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigPlan",
                  "metadata": {
                    "name": "migplan",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigMigration",
                  "metadata": {
                    "name": "migmigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigStorage",
                  "metadata": {
                    "name": "migstorage",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:89ba5ccb1cd9c44b8c8d2189f1fc6b04e780d46b19ee8a22e20c0b8c9d94d91f
          createdAt: "2025-04-28T19:55:38Z"
          description: Facilitates migration of container workloads from OpenShift
            3.x to OpenShift 4.x
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.5.0 <1.7.19'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-migration
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:3b6168c6df0d86d76ae4de91755c631b7f6a1f84286257d18c97d5a6af312291
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/mig-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: a request to migrate PVCs from a set of namespaces directly
              to a target
            displayName: DirectVolumeMigration
            kind: DirectVolumeMigration
            name: directvolumemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate watch the progress of rsync pod migrating
              a PVC to a target
            displayName: DirectVolumeMigrationProgress
            kind: DirectVolumeMigrationProgress
            name: directvolumemigrationprogresses.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a set of namespaces
              directly to a target
            displayName: DirectImageMigration
            kind: DirectImageMigration
            name: directimagemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a single imagestream
              directly to a target
            displayName: DirectImageStreamMigration
            kind: DirectImageStreamMigration
            name: directimagestreammigrations.migration.openshift.io
            version: v1alpha1
          - description: A request to Analyze a MigPlan and its namespaces
            displayName: MigAnalytic
            kind: MigAnalytic
            name: miganalytics.migration.openshift.io
            version: v1alpha1
          - description: A cluster defined for migration
            displayName: MigCluster
            kind: MigCluster
            name: migclusters.migration.openshift.io
            version: v1alpha1
          - description: A migration process
            displayName: MigMigration
            kind: MigMigration
            name: migmigrations.migration.openshift.io
            version: v1alpha1
          - description: A migration plan
            displayName: MigPlan
            kind: MigPlan
            name: migplans.migration.openshift.io
            version: v1alpha1
          - description: A migration contoller deployment
            displayName: MigrationController
            kind: MigrationController
            name: migrationcontrollers.migration.openshift.io
            version: v1alpha1
          - description: A migration hook
            displayName: MigHook
            kind: MigHook
            name: mighooks.migration.openshift.io
            version: v1alpha1
          - description: A migration storage location
            displayName: MigStorage
            kind: MigStorage
            name: migstorages.migration.openshift.io
            version: v1alpha1
        description: |-
          The Migration Toolkit for Containers Operator enables installation of the application migration tool components.

          After you have installed the Operator, you can install components by creating a MigrationController CR.

          By default, the Operator installs the following components on an OpenShift Container Platform 4 target cluster:

          * velero and restic, to back up data on the source cluster and restore it on the target cluster
          * mig-controller, to coordinate the backup and restore processes
          * mig-ui, the web console for managing migrations

          If you experience any issues or have feature requests, please file an [issue](https://github.com/konveyor/mig-operator/issues)

          For more information, see the [OpenShift Container Platform Migration Guide](https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html).
        displayName: Migration Toolkit for Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - crane
        - mtc
        links:
        - name: Migrations
          url: https://github.com/konveyor/mig-controller
        - name: Documentation
          url: https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html
        - name: Crane Operator
          url: https://github.com/konveyor/mig-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhmtc/openshift-migration-hook-runner-rhel8@sha256:c8c932a486d9b83e5a93290058d9e84691b87cd4161f8132a1bb6de9a360605e
        - registry.redhat.io/rhmtc/openshift-migration-velero-restic-restore-helper-rhel8@sha256:932216fd08d8ad97baf48ac7b8c989806d30a191337c650df37da537c153dc17
        - registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:3b6168c6df0d86d76ae4de91755c631b7f6a1f84286257d18c97d5a6af312291
        - registry.redhat.io/rhmtc/openshift-migration-velero-rhel8@sha256:bc010bc31e8b553f86aacbca39e28a95e74994c56e85c9fc1b8b80d0791ecb16
        - registry.redhat.io/rhmtc/openshift-migration-log-reader-rhel8@sha256:a0c8ab3a5bbc78c158c2ccccb1d06430a20f40358649196aa4e7311e54f9bdf8
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-aws-rhel8@sha256:023465788745625ad3b4dc0574e35b93d6ac8eea17be2ad1fc88959149a0df2c
        - registry.redhat.io/rhmtc/openshift-migration-ui-rhel8@sha256:66b288170dca8007382d13dff3f2e1868e409fb8002e04ec1fc5c480f868d9b6
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-gcp-rhel8@sha256:cc70c5a63af77948760bc42bf695a5e4eb0bce6285d6b350bba89a53f4b4e7e9
        - registry.redhat.io/rhmtc/openshift-migration-registry-rhel8@sha256:49485dfe708fe1294dec3f0bf34fedc38845a22f91915bba523d728a586b0c30
        - registry.redhat.io/rhmtc/openshift-migration-controller-rhel8@sha256:9b1490b04fc235191515fcd07d7af2e81f846d865016831e17872bb82bf4eaf5
        - registry.redhat.io/rhmtc/openshift-migration-rsync-transfer-rhel8@sha256:23a2bc5ad0f99fc0e7517a0172fdf3597fa619543b0192757983219a578f4ee2
        - registry.redhat.io/rhmtc/openshift-velero-plugin-rhel8@sha256:02698c655a01df2237aee6c570db96f13c0018cee2ae58b96b9570d93c354617
        - registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:89ba5ccb1cd9c44b8c8d2189f1fc6b04e780d46b19ee8a22e20c0b8c9d94d91f
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-microsoft-azure-rhel8@sha256:dad8fa4dbce3b4562bf06756fd74c2007812ec4edbc308e4517ff123c5519767
        version: 1.7.19
      entries:
      - name: mtc-operator.v1.7.19
        version: 1.7.19
      - name: mtc-operator.v1.7.18
        version: 1.7.18
      - name: mtc-operator.v1.7.17
        version: 1.7.17
      - name: mtc-operator.v1.7.16
        version: 1.7.16
      - name: mtc-operator.v1.7.15
        version: 1.7.15
      - name: mtc-operator.v1.7.14
        version: 1.7.14
      - name: mtc-operator.v1.7.13
        version: 1.7.13
      - name: mtc-operator.v1.7.12
        version: 1.7.12
      - name: mtc-operator.v1.7.11
        version: 1.7.11
      - name: mtc-operator.v1.7.10
        version: 1.7.10
      - name: mtc-operator.v1.7.9
        version: 1.7.9
      - name: mtc-operator.v1.7.8
        version: 1.7.8
      - name: mtc-operator.v1.7.7
        version: 1.7.7
      - name: mtc-operator.v1.7.6
        version: 1.7.6
      - name: mtc-operator.v1.7.5
        version: 1.7.5
      - name: mtc-operator.v1.7.4
        version: 1.7.4
      - name: mtc-operator.v1.7.3
        version: 1.7.3
      - name: mtc-operator.v1.7.2
        version: 1.7.2
      - name: mtc-operator.v1.7.1
        version: 1.7.1
      - name: mtc-operator.v1.7.0
        version: 1.7.0
      name: release-v1.7
    - currentCSV: mtc-operator.v1.8.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigrationController",
                  "metadata": {
                    "name": "migration-controller",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "azure_resource_group": "",
                    "cluster_name": "host",
                    "migration_velero": true,
                    "migration_controller": true,
                    "migration_ui": true,
                    "migration_log_reader": true,
                    "olm_managed": true,
                    "restic_timeout": "1h",
                    "mig_pv_limit": "100",
                    "mig_pod_limit": "100",
                    "mig_namespace_limit": "10"
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageMigration",
                  "metadata": {
                    "name": "directimagemigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageStreamMigration",
                  "metadata": {
                    "name": "directimagestreammigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigAnalytic",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "plan": {
                      "name": "foo",
                      "namespace": "openshift-migration"
                    },
                    "analyzeK8SResources": true,
                    "analyzeImageCount": true,
                    "analyzePVCapacity": true
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigCluster",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectVolumeMigration",
                  "metadata": {
                    "name": "direct",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                      "kind": "DirectVolumeMigrationProgress",
                      "metadata": {
                        "name": "direct-progress",
                        "namespace": "openshift-migration"
                      },
                      "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigPlan",
                  "metadata": {
                    "name": "migplan",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigMigration",
                  "metadata": {
                    "name": "migmigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigStorage",
                  "metadata": {
                    "name": "migstorage",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:8a44fe2aea3e842dcccb1e20df4818f4ca573fe13f51427ae8b603d0000a9411
          createdAt: "2025-08-22T15:27:29Z"
          description: Facilitates migration of container workloads from OpenShift
            3.x to OpenShift 4.x
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.5.0 <1.8.10'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-migration
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:f9492269b75608057fbc5edd53603807f1f80cd8581dea12032c88e04cf8f4cb
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/mig-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: a request to migrate PVCs from a set of namespaces directly
              to a target
            displayName: DirectVolumeMigration
            kind: DirectVolumeMigration
            name: directvolumemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate watch the progress of rsync pod migrating
              a PVC to a target
            displayName: DirectVolumeMigrationProgress
            kind: DirectVolumeMigrationProgress
            name: directvolumemigrationprogresses.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a set of namespaces
              directly to a target
            displayName: DirectImageMigration
            kind: DirectImageMigration
            name: directimagemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a single imagestream
              directly to a target
            displayName: DirectImageStreamMigration
            kind: DirectImageStreamMigration
            name: directimagestreammigrations.migration.openshift.io
            version: v1alpha1
          - description: A request to Analyze a MigPlan and its namespaces
            displayName: MigAnalytic
            kind: MigAnalytic
            name: miganalytics.migration.openshift.io
            version: v1alpha1
          - description: A cluster defined for migration
            displayName: MigCluster
            kind: MigCluster
            name: migclusters.migration.openshift.io
            version: v1alpha1
          - description: A migration process
            displayName: MigMigration
            kind: MigMigration
            name: migmigrations.migration.openshift.io
            version: v1alpha1
          - description: A migration plan
            displayName: MigPlan
            kind: MigPlan
            name: migplans.migration.openshift.io
            version: v1alpha1
          - description: A migration contoller deployment
            displayName: MigrationController
            kind: MigrationController
            name: migrationcontrollers.migration.openshift.io
            version: v1alpha1
          - description: A migration hook
            displayName: MigHook
            kind: MigHook
            name: mighooks.migration.openshift.io
            version: v1alpha1
          - description: A migration storage location
            displayName: MigStorage
            kind: MigStorage
            name: migstorages.migration.openshift.io
            version: v1alpha1
        description: |-
          The Migration Toolkit for Containers Operator enables installation of the application migration tool components.

          After you have installed the Operator, you can install components by creating a MigrationController CR.

          By default, the Operator installs the following components on an OpenShift Container Platform 4 target cluster:

          * velero and restic, to back up data on the source cluster and restore it on the target cluster
          * mig-controller, to coordinate the backup and restore processes
          * mig-ui, the web console for managing migrations

          If you experience any issues or have feature requests, please file an [issue](https://github.com/konveyor/mig-operator/issues)

          For more information, see the [OpenShift Container Platform Migration Guide](https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html).
        displayName: Migration Toolkit for Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - crane
        - mtc
        links:
        - name: Migrations
          url: https://github.com/konveyor/mig-controller
        - name: Documentation
          url: https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html
        - name: Crane Operator
          url: https://github.com/konveyor/mig-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhmtc/openshift-migration-registry-rhel8@sha256:accef2a0dd69b254bd120232e5dc86a56aa8f25bcfdba3f9bab7e4127c157762
        - registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:f9492269b75608057fbc5edd53603807f1f80cd8581dea12032c88e04cf8f4cb
        - registry.redhat.io/rhmtc/openshift-migration-controller-rhel8@sha256:c776387c662e21aa61378d0136b2495826e1c157548f9ca9c95d044af38ee169
        - registry.redhat.io/rhmtc/openshift-migration-hook-runner-rhel8@sha256:adaa4f1afefb4d5853713821cb6e7de0d9f5a640f0243f20b41a44be84707caf
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-mtc-rhel8@sha256:5f46ba04146438f2abd6905468d2ecbeb19374a83771d364904ac2ee6afdbefa
        - registry.redhat.io/rhmtc/openshift-migration-openvpn-rhel8@sha256:d21b99d8edefa7a1bab30c85969ec7d7560d06be308bd787e698e409d54cbc19
        - registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:8a44fe2aea3e842dcccb1e20df4818f4ca573fe13f51427ae8b603d0000a9411
        - registry.redhat.io/rhmtc/openshift-migration-log-reader-rhel8@sha256:4f76bcf3df3bb7073b8aabd3da072cacfec05ecf4bfb47e23340074756f7fc30
        - registry.redhat.io/rhmtc/openshift-migration-rsync-transfer-rhel8@sha256:442e7c7f2cc075bf468403ace8846c96ad2fe43eb7f6562d566caa24e4f86873
        - registry.redhat.io/rhmtc/openshift-migration-ui-rhel8@sha256:86e747e0330b0156379d273d0c29c13a65dafbf4f536656723e2bc746872286c
        version: 1.8.10
      entries:
      - name: mtc-operator.v1.8.10
        version: 1.8.10
      - name: mtc-operator.v1.8.9
        version: 1.8.9
      - name: mtc-operator.v1.8.8
        version: 1.8.8
      - name: mtc-operator.v1.8.7
        version: 1.8.7
      - name: mtc-operator.v1.8.6
        version: 1.8.6
      - name: mtc-operator.v1.8.5
        version: 1.8.5
      - name: mtc-operator.v1.8.4
        version: 1.8.4
      - name: mtc-operator.v1.8.3
        version: 1.8.3
      - name: mtc-operator.v1.8.2
        version: 1.8.2
      - name: mtc-operator.v1.8.1
        version: 1.8.1
      - name: mtc-operator.v1.8.0
        version: 1.8.0
      name: release-v1.8
    defaultChannel: release-v1.8
    packageName: mtc-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Performance
      provider-url: ""
    name: cluster-impairment-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-impairment-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "all-impairments"
                },
                "spec": {
                  "duration": 480,
                  "egress": {
                    "bandwidth": 10000,
                    "latency": 50,
                    "loss": 0.02
                  },
                  "ingress": {
                    "bandwidth": 10000,
                    "latency": 50,
                    "loss": 0.02
                  },
                  "interfaces": [
                    "ens2f0"
                  ],
                  "linkFlapping": {
                    "downTime": 30,
                    "enable": true,
                    "upTime": 120
                  },
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "duplication"
                },
                "spec": {
                  "duration": 60,
                  "egress": {
                    "duplication": 25
                  },
                  "ingress": {
                    "duplication": 25
                  },
                  "interfaces": [
                    "ens3"
                  ],
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "two-min-flap"
                },
                "spec": {
                  "duration": 480,
                  "interfaces": [
                    "ens2f0"
                  ],
                  "linkFlapping": {
                    "downTime": 120,
                    "enable": true,
                    "upTime": 120
                  },
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "typical-scenario"
                },
                "spec": {
                  "duration": 30,
                  "egress": {
                    "latency": 50
                  },
                  "ingress": {
                    "latency": 50
                  },
                  "interfaces": [
                    "ens2f0",
                    "eno1"
                  ],
                  "loss": 0.02,
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "uneven-latency"
                },
                "spec": {
                  "duration": 60,
                  "egress": {
                    "latency": 100
                  },
                  "ingress": {
                    "latency": 10
                  },
                  "interfaces": [
                    "ens2f0"
                  ],
                  "nodeSelector": {
                    "key": "node-role.kubernetes.io/worker",
                    "value": ""
                  },
                  "startDelay": 5
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Developer Tools
          containerImage: quay.io/redhat-performance/cluster-impairment-operator:1.0.4
          createdAt: "2021-08-11 12:15:00"
          description: Run Impairments on Kubernetes/Openshift Nodes
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/redhat-performance/cluster-impairment-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterImpairment
            name: clusterimpairments.apps.redhat.com
            version: v1alpha1
        description: "cluster-impairment-operator is an operator designed to automate
          the application of impairments to the nodes in a cluster.\n\n## Features\n\nTraffic
          Direction:\n\n* Egress impairments affect the traffic going out.    \n*
          Ingress impairments affect the traffic coming in. (Ingress impairments require
          the kernel module IFB.)\n\n### Impairments\n\n| Impairment    | Description
          \                               | Unit    | Uses Netem | Correlation Supported
          |\n|---------------|--------------------------------------------|---------|------------|-----------------------|\n|
          Bandwidth     | The bandwidth limit                        | kbit/s  | Yes
          \       | No                    |\n| Latency       | The delay of the packets
          \                  | ms      | Yes        | Yes                   |\n| Packet
          Loss   | The percent of packets that are dropped    | percent | Yes        |
          Yes                   |\n| Corruption    | The percent of packets that are
          corrupted  | percent | Yes        | Yes                   |\n| Duplication
          \  | The percent of packets that are duplicated | percent | Yes        |
          Yes                   |\n| Link Flapping | Turns the interface on and off
          \            | bool    | No         | No                    |\n\n\nOn the
          tested environment (RHEL CoreOS 48.84), the impairments can be used alongside
          link flapping.\n\n\n#### Latency Options\n\nIn addition to simply delaying
          the packets, there are advanced latency options.\n\n##### Jitter\n\nSemi-randomly
          adds or subtracts from the latency according to the distribution up to the
          amount specified.\nIf latency is 100ms, and jitter is 10ms, the actual latency
          will vary 100±10ms.\n\n\n##### Distribution\n\nThe distribution of the jitter.
          The options are:\n* Normal\n* Uniform\n* Pareto\n* Paretonormal\n\n#####
          Jitter correlation\n\nThe percent chance that the next latency's jitter
          value will correlate with the preceeding latency.\n\n##### Reorder\n\nThe
          percentage of packets that are not delayed, causing a reorder due to them
          being sent before the delayed ones.\nApplying jitter itself has the potential
          to also cause reordering.\n\n##### Reorder correlation\n\nThe percet chance
          that the value for the next reorder will correlate with the preceeding value.\n\n####
          Loss options\n\n##### Correlation\n\nThe percent chance that the previous
          loss value for a packet correlates with the loss value for the next packet.\n\n####
          Corruption options\n\n###### Corruption\n\nThe percent chance that each
          packet will be corrupted.\n\n##### Corruption correlation\n\nThe percet
          chance that the value for the next corruption will correlate with the preceeding
          value.\n\n#### Duplication options\n\n##### Duplication\n\nThe percent chance
          that each packet will be duplicated.\n\n##### Duplication correlation\n\nThe
          percet chance that the value for the next duplication will correlate with
          the preceeding value.\n\n\n\n## Configuration\n\nHere is an example of the
          ClusterImpairment custom resource.\n\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: test-impairment-cr\nspec:\n  duration:
          30 # seconds\n  startDelay: 5 # seconds. It typically takes about 2-3 seconds
          for the Daemonset to run\n  interfaces:\n  - \"ens2f0\"\n  ingress: # uses
          ifb\n    bandwidth: 0 # kbit\n    latency: 10 # ms\n    latencyOptions:\n
          \     jitter: 5 # ms\n      jitterCorrelation: 25 # percent\n      distribution:
          normal\n      reorder: 25 # percent of packets that will skip the delay\n
          \     reorderCorrelation: 25 # percent\n    loss: 0 # percent\n    lossOptions:\n
          \     correlation: 25 # percent\n    corruption: 0.1 # percent\n    corruptionOptions:\n
          \     correlation: 25 # percent\n    duplication: 1 # percent\n    duplicationOptions:\n
          \     correlation: 25 # percent\n  egress:\n    bandwidth: 0 # kbit\n    latency:
          100 # ms\n    loss: 0 # percent\n  linkFlapping:\n    enable: false\n    downTime:
          3 # Seconds\n    upTime: 3 # Seconds\n  nodeSelector:\n    key: \"node-role.kubernetes.io/worker\"\n
          \   value: \"\"\n    \n```\n\n#### Interfaces\n\nThe interfaces configuration
          option is a list of all interfaces that the impairments should be applied
          to. The current implementation will error out once it hits an invalid interface.\n\nIf
          an invalid interface is found, it will print out the list of interfaces.\n\n####
          Node Selector\n\nThe default node selector is all worker nodes, but you
          can change it to whatever node selector you want by setting the key and
          value.\nThere is a toleration to allow it to run on master nodes if the
          node selector includes master nodes.\n\n#### Duration\n\nThe number of seconds
          the impairments should run. It will try to sync the start and end time between
          all pods.\nIf set to -1, it will run indefinitely (a year), until you remove
          the ClusterImpairment custom resource. This is good for testing that requires
          steady impairments.\n\nIf the script is link flapping, set this value to
          be short enough so that if the link flap interrupts communication between
          the nodes, the daemonset will remove itself.\n\n#### Start Delay\n\nThe
          delay before starting the script. If you want the pods to be in sync, a
          minimum of a few seconds should be used because the pods take time to start
          up.\n\nYou can also utilize this feature to run an separate impairment after
          another. Just apply two resources at the same time, but add the duration
          and start delay of the first to the start delay of the second.\n\n#### Ingress
          and Egress\n\nThe configuration sections \"ingress\" and \"egress\" apply
          to each direction.\n\n##### Examples:\n\n##### Example 1\n\n\nIn this example,
          egress latency is set to 100ms, but the ingress latency is set to 10ms.
          So the latency to the interface will end up being 10ms, but 100ms going
          out. When pinging, this will result in 110ms of latency.\n\n```\n\napiVersion:
          apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n  name: uneven-latency\nspec:\n
          \ duration: 60\n  startDelay: 5\n  interfaces:\n  - \"ens2f0\"\n  ingress:\n
          \   latency: 10 # ms\n  egress:\n    latency: 100 # ms\n\n```\n\n##### Example
          2\n\n\nIn this example, link flapping will be enabled, and it will turn
          the interface on and off every 2 minutes. That is long enough for kubernetes
          to determine that a node is unavailable.\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: two-min-flap\nspec:\n  duration: 480\n
          \ startDelay: 5\n  interfaces:\n  - \"ens2f0\"\n  linkFlapping:\n    enable:
          true\n    downTime: 120 # Seconds\n    upTime: 120 # Seconds\n\n```\n\n#####
          Example 3\n\n\nIn this example, a realistic set of impairments is applied
          to `ens2f0` and `eno1` for 30 seconds:\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: typical-scenario\nspec:\n  duration:
          30 # seconds\n  startDelay: 5 # seconds\n  interfaces:\n  - \"ens2f0\"\n
          \ - \"eno1\"\n  egress:\n    latency: 50 # ms. Bidirectional, so total of
          100ms\n  ingress:\n    latency: 50 # ms. Bidirectional, so total of 100ms\n
          \ loss: 0.02 # percent\n\n```\n\n##### Example 4\n\n\nIn this example, a
          combination of latency, packet loss, bandwidth, and link flapping is applied.\n\n```\n\napiVersion:
          apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n  name: all-impairments\nspec:\n
          \ duration: 480 # seconds\n  startDelay: 5 # seconds\n  interfaces:\n  -
          \"ens2f0\"\n  egress:\n    latency: 50 # ms. Bidirectional, so total of
          100ms\n    loss: 0.02 # percent\n    bandwidth: 1000 # 1000 kbit/s, about
          1 mbit/s\n  ingress:\n    latency: 50 # ms. Bidirectional, so total of 100ms\n
          \   loss: 0.02 # percent\n    bandwidth: 1000 # 1000 kbit/s, about 1 mbit/s\n
          \ linkFlapping:\n    enable: true\n    downTime: 30 # Seconds\n    upTime:
          120 # Seconds\n\n```\n\n##### Example 5\n\n\nIn this example, packet duplicaiton
          is applied.\n\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n
          \ name: duplication\nspec:\n  duration: 60\n  startDelay: 5\n  interfaces:\n
          \ - \"ens3\"\n  ingress:\n    duplication: 25 # percent\n  egress:\n    duplication:
          25 # percent\n\n```\n\n## Setup\n\n### Requirements\n\n1. You need `make`
          installed.\n2. You need access to the kubernetes cluster with a kubeconfig.\n\n###
          Installation\n\nTo run using the current latest image:\n1. Clone the repository.\n2.
          Run `make deploy` with KUBECONFIG environment variable set to the location
          of the kubeconfig.\n\nTo run with your own code, there are more steps.\n1.
          Fork the repository.\n2. Clone to a machine that has access to the Kubernetes
          cluster and the kubeconfig.\n3. Modify the makefile to change the `IMG`
          variable to your image repository. If you do not have podman installed,
          also change podman to docker.\n4. Run `make docker-build` then `make docker-push`.\n5.
          Then run `make deploy`\n\n### Deploying from operator-hub\n\n*Important*
          You must manually apply the serviceaccount!\n\nFirst, Follow the instructions
          on this site to deploy from here.\n\nNext, apply the manifest \"serviceaccount.yaml\"
          from the manifests folder.\n\nThen you can apply your ClusterImpairment
          resources.\n\n## Running impairments\n\nFirst, configure a ClusterImpairment
          type's spec with the information for the impairment you want to run. Most
          importantly, set the interface(s). You can list the interfaces with `ifconfig`.
          If an invalid interface is set, the pod's logs will show `ifconfig` output.\n\n\nOnce
          the clusterimpairment type is set, apply it and it will work.\n\n\nTo re-run
          an impairment, delete the impairment with `oc delete -f path/to/manifest.yaml`,
          then apply it again.\n\n\n## Limitations\n\n\n### Multiple Impairments\n\n\nYou
          should avoid any configurations that apply impairments to the same interface
          on the same node. There are potential conflicts. The worker pod will attempt
          to remove all impairments on that interface before applying new ones.\n\n\nInstead,
          take advantage of the full control of both ingress and egress impairments
          from within the same ClusterImpairment resource.\n\n\nThere is currently
          a configured limit of 20 ClusterImpairment resources working at the same
          time. If that is a limitation for your use case, you can change it in `config/manager/manager.yaml`
          or you can try to find a way to make each ClusterImpairment do more.\n\n###
          Traffic Control (TC)\n\nTraffic control is how cluster-impairment-operator
          applies the latency, bandwidth, and packet loss impairments. The limitation
          is due to Linux not being realtime operating system, so the impairment will
          not be perfectly consistent.\n\n### Link Flapping\n\nWhen link flapping,
          if you flap the link that Kubernetes uses to communicate with the pods,
          you may be unable to remove the pod until the link is on long enough for
          Kubernetes to communicate with the impaired node.\n\n\nIn this case, it
          is helpful to set the duration properly instead of running for an indefinite
          or large amount of time, because the node will properly go back to the unimpaired
          state at that time.\n\n\nIf the cluster becomes offline due to the link
          flapping when you do not want it to be offline, soft restarting the nodes
          after removing the custom resource should remove all impairments.\n"
        displayName: cluster-impairment-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network
        - impairment
        - kubernetes
        - openshift
        - tc
        - trafficcontrol
        links:
        - name: Cluster Impairment Operator
          url: https://github.com/redhat-performance/cluster-impairment-operator
        maintainers:
        - email: joconnel@redhat.com
          name: Jared O'Connell
        - email: akrzos@redhat.com
          name: Alex Krzos
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat Performance
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/redhat-performance/cluster-impairment-operator:1.0.4
        version: 1.0.4
      entries:
      - name: cluster-impairment-operator.v1.0.4
        version: 1.0.4
      - name: cluster-impairment-operator.v1.0.3
        version: 1.0.3
      - name: cluster-impairment-operator.v1.0.2
        version: 1.0.2
      - name: cluster-impairment-operator.v1.0.1
        version: 1.0.1
      name: beta
    defaultChannel: beta
    packageName: cluster-impairment-operator
    provider:
      name: Red Hat Performance
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: mercury-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mercury-operator.1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:1.0.1",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CustomerCreditRating",
                  "expose": ["http"]
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:1.0.1
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:1.0.1
        version: 1.0.1
      entries:
      - name: mercury-operator.1.0.1
        version: 1.0.1
      name: 1.0.x
    - currentCSV: mercury-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_CREDIT_RATING",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-offer",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-offer-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_OFFER",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-party-routing-profile",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/party-routing-profile-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "PARTY_ROUTING_PROFILE",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-example-sdi": {
                "description": "Example Service Domain Infrastructure"
            },
            {
             "example-customer-credit-rating": {
               "description": "Example Customer Credit Rating Service Domain"
            },
            {
             "example-customer-offer": {
               "description": "Example Customer Offer Service Domain"
            },
            {
             "example-party-routing-profile": {
               "description": "Example Party Routing Profile Service Domain"
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:v1.0.2
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - description: Represents a Kafka cluster
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Integration is the Schema for the Camel K integrations API
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:v1.0.2
        version: 1.0.2
      entries:
      - name: mercury-operator.v1.0.2
        version: 1.0.2
      name: dev
    - currentCSV: mercury-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_CREDIT_RATING",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-offer",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-offer-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_OFFER",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-party-routing-profile",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/party-routing-profile-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "PARTY_ROUTING_PROFILE",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-example-sdi": {
                "description": "Example Service Domain Infrastructure"
            },
            {
             "example-customer-credit-rating": {
               "description": "Example Customer Credit Rating Service Domain"
            },
            {
             "example-customer-offer": {
               "description": "Example Customer Offer Service Domain"
            },
            {
             "example-party-routing-profile": {
               "description": "Example Party Routing Profile Service Domain"
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:v1.0.2
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - description: Represents a Kafka cluster
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Integration is the Schema for the Camel K integrations API
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:v1.0.2
        version: 1.0.2
      entries:
      - name: mercury-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: 1.0.x
    packageName: mercury-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-ati-updates
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-ati-updates.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ati.loadcore.keysight.com/v1alpha1",
                "kind": "Updates",
                "metadata": {
                  "name": "updates-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "installWait": {
                    "enabled": true,
                    "image": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:40:21Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Updates
            name: updates.ati.loadcore.keysight.com
            version: v1alpha1
        description: Keysight ATI Updates Operator
        displayName: Keysight ATI Updates Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Ati Updates
          url: https://keysight-ati-updates.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-ati-updates@sha256:abda76eb50d77e5a95191c9dbb21eefe64af5b794511ceafdfda007c15449c6e
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        version: 0.1.21
      entries:
      - name: keysight-ati-updates.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-ati-updates
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: CyberArk
      provider-url: https://www.cyberark.com/
    name: vcp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vcp-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "installer.venafi.com/v1alpha1",
                "kind": "VenafiInstall",
                "metadata": {
                  "name": "install-venafi-enhanced-issuer"
                },
                "spec": {
                  "certManager": {
                    "install": true
                  },
                  "globals": {
                    "imagePullSecretNames": [
                      "venafi"
                    ]
                  },
                  "venafiConnection": {
                    "install": true
                  },
                  "venafiEnhancedIssuer": {
                    "install": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:9da79fd0941c21e433f22aaaccc43ebbd4646e685aea0050194954d8b4019c05
          createdAt: "2024-05-30T11:35:41Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              VenafiInstall is an object for installing Venafi components onto a Kubernetes
              cluster
            displayName: VenafiInstall
            kind: VenafiInstall
            name: venafiinstalls.installer.venafi.com
            version: v1alpha1
        description: |
          Venafi Control Plane Operator for Red Hat OpenShift simplifies deployment of Venafi's enterprise components and dependencies.

          The enterprise components include:
          - Enterprise cert-manager: An enhanced version of cert-manager designed for application certificate management within your clusters.
          - Venafi Enhanced Issuer: Enables seamless certificate enrollment for your clusters directly from the Venafi Control Plane.
          - Venafi Firefly: A high-performance, lightweight issuer for certificates, centrally managed by the Venafi Control Plane.
          - Venafi Kubernetes Agent: Provides real-time visibility into certificates and Kubernetes resources directly within the Venafi Control Plane.
        displayName: Venafi Control Plane Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - venafi
        - vcp-operator
        links:
        - name: Documentation
          url: https://docs.venafi.cloud/
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Venafi
          url: https://venafi.com
        relatedImages:
        - registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:9da79fd0941c21e433f22aaaccc43ebbd4646e685aea0050194954d8b4019c05
        version: 1.0.0
      entries:
      - name: vcp-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: vcp-operator.v1.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "installer.venafi.com/v1alpha1",
                "kind": "VenafiInstall",
                "metadata": {
                  "name": "install-venafi-enhanced-issuer"
                },
                "spec": {
                  "certManager": {
                    "install": true
                  },
                  "globals": {
                    "imagePullSecretNames": [
                      "venafi"
                    ]
                  },
                  "venafiConnection": {
                    "install": true
                  },
                  "venafiEnhancedIssuer": {
                    "install": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:eb2b06d191dbfbb02332375a7bca6bc010d5410e53b3bc5fb8c320101292408c
          createdAt: "2025-07-22T06:44:30Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              VenafiInstall is an object for installing Venafi components onto a Kubernetes
              cluster
            displayName: VenafiInstall
            kind: VenafiInstall
            name: venafiinstalls.installer.venafi.com
            version: v1alpha1
        description: |
          The operator, formerly known as the Venafi Control Plane Operator, simplifies the deployment of enterprise components part of the CyberArk Certificate Manager for Kubernetes.

          The components include:
          - cert-manager: An enterprise distribution of cert-manager designed for application certificate management within your clusters.
          - Enterprise Issuer: Enables seamless certificate enrollment directly from the Certificate Manager.
          - Workload Identity Manager: A high-performance issuer for certificates, centrally managed by the Certificate Manager.
          - Agent: Provides real-time visibility into cluster resources directly within the Certificate Manager.
        displayName: CyberArk Certificate Manager Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cyberark
        - vcp-operator
        links:
        - name: Documentation
          url: https://docs.venafi.cloud/vaas/k8s-components/c-vcpo-overview/
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:eb2b06d191dbfbb02332375a7bca6bc010d5410e53b3bc5fb8c320101292408c
        version: 1.8.1
      entries:
      - name: vcp-operator.v1.8.1
        version: 1.8.1
      - name: vcp-operator.v1.8.0
        version: 1.8.0
      - name: vcp-operator.v1.7.0
        version: 1.7.0
      - name: vcp-operator.v1.6.0
        version: 1.6.0
      - name: vcp-operator.v1.5.0
        version: 1.5.0
      - name: vcp-operator.v1.4.0
        version: 1.4.0
      - name: vcp-operator.v1.3.0
        version: 1.3.0
      - name: vcp-operator.v1.2.1
        version: 1.2.1
      - name: vcp-operator.v1.2.0
        version: 1.2.0
      - name: vcp-operator.v1.1.1
        version: 1.1.1
      name: stable
    defaultChannel: stable
    packageName: vcp-operator
    provider:
      name: CyberArk
      url: https://www.cyberark.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE Ezmeral
      provider-url: ""
    name: hpe-ezmeral-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-ezmeral-csi-operator.v1.0.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ezmeral.hpe.com/v1",
                "kind": "HPEEzmeralCSIDriver",
                "metadata": {
                  "name": "hpeezmeralcsidriver"
                },
                "spec": {
                  "controllerImage": "registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03",
                  "nodeImage": "registry.connect.redhat.com/maprtech/csi-kdfplugin@sha256:ffd273ee2e608d2ef64fd3b2bcbb02d5f0a6c173ed96a15eaf5549ea45fea8d1",
                  "pullPolicy": "Always"
                }
              },
              {
                "apiVersion": "ezmeral.hpe.com/v1",
                "kind": "HPEEzmeralNFSCSIDriver",
                "metadata": {
                  "name": "hpeezmeralnfscsidriver"
                },
                "spec": {
                  "controllerImage": "registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03",
                  "nodeImage": "registry.connect.redhat.com/maprtech/csi-nfsplugin@sha256:a5c954f62f1395bafe2d68454ac604bca4be97573a3482fcbd8e76b6361cc08a",
                  "pullPolicy": "Always"
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator@sha256:c9c243c51cd760c2c2a66b489807a5e22053fcbeb94fa29afd10d9c91e2d3908
          createdAt: "2025-06-25T11:51:05Z"
          description: A Container Storage Interface (CSI) driver for HPE Ezmeral
            Data Fabric platform. The CSI driver allows you to use HPE Ezmeral Data
            Fabric with your preferred container orchestrator.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mapr/mapr-csi
          support: HPE Ezmeral Data Fabric
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE Ezmeral CSI Driver (FUSE)
            displayName: HPEEzmeralCSIDriver
            kind: HPEEzmeralCSIDriver
            name: hpeezmeralcsidrivers.ezmeral.hpe.com
            version: v1
          - description: HPE Ezmeral CSI Driver (NFS)
            displayName: HPEEzmeralNFSCSIDriver
            kind: HPEEzmeralNFSCSIDriver
            name: hpeezmeralnfscsidrivers.ezmeral.hpe.com
            version: v1
        description: |
          The HPE Ezmeral CSI Operator for Kubernetes packages, deploys, manages, upgrades the HPE Ezmeral CSI Driver on Kubernetes and OpenShift for dynamic provisioning of persistent volumes on HPE Ezmeral Data Fabric platform.
          The HPE Ezmeral CSI Driver for Kubernetes leverages HPE Ezmeral Data Fabric platform to provide scalable and persistent storage for stateful applications. Please refer to our CSI driver [documentation](https://docs.datafabric.hpe.com/home/CSIdriver/csi_overview.html) for full list of supported CSI features.
          ## Installation
          Refer to the HPE Ezmeral CSI Operator for Kubernetes [documentation](https://docs.datafabric.hpe.com/home/CSIdriver/csi_installing_operator.html)
        displayName: HPE Ezmeral Data Fabric CSI Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE Ezmeral
        - HPE Ezmeral Data Fabric
        - Storage
        - FileSystem
        - BlockVolume
        links:
        - name: Documentation
          url: https://docs.datafabric.hpe.com/home/CSIdriver/csi_overview.html
        maintainers:
        - email: ezmeral-csi-support@hpe.com
          name: HPE Ezmeral Data Fabric
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: HPE Ezmeral
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6ba961c2c2a29750c0132fe6dd6fa9f6001010afbc5f19b98add87b31b54bcf6
        - registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator@sha256:c9c243c51cd760c2c2a66b489807a5e22053fcbeb94fa29afd10d9c91e2d3908
        - registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03
        - registry.connect.redhat.com/maprtech/csi-kdfplugin@sha256:ffd273ee2e608d2ef64fd3b2bcbb02d5f0a6c173ed96a15eaf5549ea45fea8d1
        - registry.connect.redhat.com/maprtech/csi-nfsplugin@sha256:a5c954f62f1395bafe2d68454ac604bca4be97573a3482fcbd8e76b6361cc08a
        version: 1.0.18
      entries:
      - name: hpe-ezmeral-csi-operator.v1.0.18
        version: 1.0.18
      name: stable
    defaultChannel: stable
    packageName: hpe-ezmeral-csi-operator
    provider:
      name: HPE Ezmeral
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzo-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzo-operator.v3.1.200
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzo.cambridgesemantics.com/v2",
                "kind": "Anzo",
                "metadata": {
                  "name": "no-persistence"
                },
                "spec": {
                  "jsonActivation": true,
                  "jvmMemory": "4096m",
                  "nodeConfig": {
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "app": "anzo"
                        }
                      },
                      "serviceName": "anzo-no-persistence",
                      "template": {
                        "metadata": {
                          "annotations": {
                            "cluster-autoscaler.kubernetes.io/safe-to-evict": "false"
                          },
                          "labels": {
                            "app": "anzo"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "registry.connect.redhat.com/cambridgesemantics/anzo@sha256:5b192604338f602fdabe55fecefd9eb82400a77c5aa7f04bd8e54aead8644c6d",
                              "imagePullPolicy": "Always",
                              "name": "anzo",
                              "resources": {
                                "limits": {
                                  "cpu": "2.0",
                                  "memory": "8.0Gi"
                                },
                                "requests": {
                                  "cpu": "2.0",
                                  "memory": "8.0Gi"
                                }
                              }
                            }
                          ],
                          "initContainers": [
                            {
                              "name": "anzo-init",
                              "resources": {
                                "limits": {
                                  "cpu": "0.5",
                                  "memory": "2Gi"
                                },
                                "requests": {
                                  "cpu": "0.5",
                                  "memory": "2Gi"
                                }
                              }
                            }
                          ],
                          "securityContext": {
                            "fsGroup": 1001
                          },
                          "serviceAccountName": "anzo-operator"
                        }
                      }
                    }
                  },
                  "role": "AnzoServer"
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/anzo-operator@sha256:4c0c8bf7acbfbb96f15459a20364608e5d0a0933c6326a89d219666ee0d88ca2
          createdAt: "2025-07-09T06:52:12Z"
          description: kubernetes operator for Altair Graph Studio
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Altair Customer Support
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoBackup is the Schema for the anzobackups API
            displayName: Altair Graph Studio Backup
            kind: AnzoBackup
            name: anzobackups.anzobackup.cambridgesemantics.com
            version: v1beta1
          - description: Altair Graph Studio is the Schema for the anzos API
            displayName: Altair Graph Studio Operator
            kind: Anzo
            name: anzos.anzo.cambridgesemantics.com
            version: v2
        description: |-
          The Altair Graph Studio Operator provides the way to install and configure an altair graph studio server and agent setup on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v3.1.2/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/deploy-k8s.htm )

          ### Support

          We offer Support to our customers [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Studio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        links:
        - name: Altair Graph Studio Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v3.1.2/README_openshift.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzo-operator@sha256:4c0c8bf7acbfbb96f15459a20364608e5d0a0933c6326a89d219666ee0d88ca2
        - registry.connect.redhat.com/cambridgesemantics/anzo@sha256:5b192604338f602fdabe55fecefd9eb82400a77c5aa7f04bd8e54aead8644c6d
        version: 3.1.200
      entries:
      - name: anzo-operator.v3.1.200
        version: 3.1.200
      - name: anzo-operator.v3.1.106
        version: 3.1.106
      - name: anzo-operator.v3.1.105
        version: 3.1.105
      name: stable
    defaultChannel: stable
    packageName: anzo-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq7-interconnect-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: 1.10.x
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: 1.2.0
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: latest
    defaultChannel: 1.10.x
    packageName: amq7-interconnect-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-registry
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-operator.v1.1.3-v2.6.4.final
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-mem"
                },
                "spec": {
                  "configuration": {
                    "persistence": "mem"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-operator:1.1.3
          createdAt: "2024-09-17"
          description: Deploy and manage Apicurio Registry on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents an Apicurio Registry instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Apicurio Registry

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Provides a REST API and web UI to manage the artifacts and artifact meta-data
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, in-memory
          - Compatible with Confluent and IBM APIs
          - Runs on a lightweight Quarkus platform
          - Includes Maven plugin to integrate with Maven based builds

          ## Apicurio Registry Operator

          Provides a quick and easy way to deploy and manage Apicurio Registry on Kubernetes.

          **Features**
          - Supports installation and configuration of the Registry for all storage options
          - Easily perform a rolling upgrade of Apicurio Registry

          ## Prerequisites

          This operator does not deploy storage for Apicurio Registry. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/main/LICENSE)
        displayName: Apicurio Registry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-operator:1.1.3
        version: 1.1.3-v2.6.4.final
      entries:
      - name: apicurio-registry-operator.v1.1.3-v2.6.4.final
        version: 1.1.3-v2.6.4.final
      - name: apicurio-registry-operator.v1.1.2-v2.5.11.final
        version: 1.1.2-v2.5.11.final
      - name: apicurio-registry-operator.v1.1.1-v2.5.9.final
        version: 1.1.1-v2.5.9.final
      - name: apicurio-registry-operator.v1.1.0-v2.4.12.final
        version: 1.1.0-v2.4.12.final
      - name: apicurio-registry-operator.v1.0.0-v2.0.0.final
        version: 1.0.0-v2.0.0.final
      name: 2.x
    defaultChannel: 2.x
    packageName: apicurio-registry
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zettaset
      provider-url: ""
    name: xcrypt-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zts-xcrypt-operator.v0.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xcrypt.zettaset.com/v1alpha1",
                "kind": "XCrypt",
                "metadata": {
                  "name": "zts-masterset"
                },
                "spec": {
                  "status": {
                    "podNames": "zts-ca",
                    "replicas": "1,"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.connect.redhat.com/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
          createdAt: "2021-06-16T12:29:13.97Z"
          description: XCrypt Operator deploys Zettaset XCrypt Container Encryption  -
            the only software-defined solution that offers transparent high performance
            data-at-rest encryption of critical data in container environments.
          support: Zettaset
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XCrypt is the Schema for the xcrypts API
            kind: XCrypt
            name: xcrypts.xcrypt.zettaset.com
            version: v1alpha1
        description: "Zettaset’s RedHat Certified XCrypt Operator allows users to
          deploy multiple XCrypt components on OpenShift with just a few simple commands.
          The XCrypt Operator deployment provides software-defined encryption that
          transparently protects container data throughout your OpenShift environment.\n\n##
          Core Features\n* **Automated XCrypt provisioning** - Deploying a Zettaset
          XCrypt has never been easier. Adjust cluster specific configuration and
          let the Zettaset Operator take care of provisioning XCrypt services and
          setting up the XCrypt encryption to your exact specification. \n* **CSI
          Driver deployment**: Zettaset XCrypt Operator automatically deploys Zettaset
          XCrypt CSI Driver in order to provide automated and transparent data-at-rest
          \ encryption functionality compatible with Openshift and any other container
          orchestrator. Once the Operator is provisioned, the driver runs fine tuned
          and ready to work with other XCrypt components.\n* **Multi-node topology**:
          Encrypt as many nodes and volumes as you need to. XCrypt components can
          be installed on multiple nodes across the cluster. This is defined by two
          labels: zts-master and zts-worker. Normally zts-master node is a single
          node that runs Zettaset XCrypt major services required for the product operation.
          Multiple nodes can be marked as zts-worker and those are the nodes that
          can have encrypted volumes.\n* **Seamless Encrypted Volume mount**: Volumes
          are not only being encrypted, but also are mounted/unmounted to containers
          with Persistent Volume Claims. This happens as seamlessly as regular volumes
          provision and mount/unmount.\n* **Robust failover**: A node failure doesn’t
          affect XCrypt operations on other nodes of the cluster. Encryption on a
          restarted/recovered node gets back to normal functioning if the node stays
          healthy and data hasn’t been corrupted during the restart\n\n## About Zettaset
          XCrypt Kubernetes Encryption for OpenShift\nXCrypt Kubernetes Encryption
          for OpenShift is a software-only solution that enables encryption of data
          at rest stored in Red Hat OpenShift container environments without slowing
          down the DevSecOps process. In addition to providing a transparent and high-performance
          layer of security that flexibly protects container data across on-premise,
          in hybrid, and multi-cloud environments, it also helps your organization
          efficiently transition from DevOps to DevSecOps.  \n\nMore information about
          Zettaset XCrypt Kubernetes Encryption for OpenShift can be found in the
          [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)\n\n##
          RedHat OpenShift + Zettaset: The next generation of secure enterprise infrastructure
          for DevSecOps  \nBy combining RedHat’s Container Platform with Zettaset’s
          certified software-defined encryption, you can create the next generation
          of secure enterprise infrastructure that DevSecOps requires. Together they
          create the stable, consistent, and supported base that your organization
          needs to develop applications faster, simplify updates, drive innovation,
          and reduce the risk of potential security breaches and data theft.\n###
          Core Features\n* Software-only for simple and scalable deployments \n* Negligible
          performance impact\n* Transparent to developers and administrators \n* Works
          with AWS, Google, and Azure \n* Unique encryption key per container volume
          \n* Encrypted volumes are only available when in use\n* Direct integration
          with OpenShift \n* Automated encryption policy management Secure erase of
          volumes\n* Ability to securely decommission a node\n\n## How does XCrypt
          Kubernetes Encryption fit into OpenShift? \nXCrypt XCrypt Kubernetes Encryption
          for OpenShift Encryption makes it incredibly easy to protect container data
          in your OpenShift environment. \n### Use cases include:\n* Transitioning
          from DevOps to DevSecOps \n* Ensuring data protection in OpenShift container
          environments with negligible performance impact \n* Achieving compliance
          in regulated industries like healthcare and finance\n\n## Other Information\nRole-based
          access control needs to be configured prior to the Operator installation.
          Please find details in the [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)
          \   "
        displayName: Zettaset XCrypt Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - encryption
        - storage
        links:
        - name: Zettaset
          url: https://www.zettaset.com/
        maintainers:
        - email: info@zettaset.com
          name: Zettaset
        maturity: alpha
        provider:
          name: Zettaset
        relatedImages:
        - registry.connect.redhat.com/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
        - registry.connect.redhat.com/zts/xcrypt-license-server-1@sha256:bc5486892a41e70f5dc4c4e1d82fca8db2149598fb8b5950ba5da1ee2d9d245d
        - registry.connect.redhat.com/zts/xcrypt-host-manager-1@sha256:9faa4b597eb74df412761dffa057b64fb69dbf9fa016414962eb2099a343e734
        - registry.connect.redhat.com/zts/zts-csi-driver-1@sha256:c15fa9a7cc40891103f426206663dcebe1210c8ee3ab6037cbf64e1f86b9c7e5
        - registry.connect.redhat.com/zts/xcrypt-ca-1@sha256:9f1015dfb459b69f5dbf449efa0517452095a6cf4f20826ebae278f9f800ffe7
        - registry.connect.redhat.com/zts/xcrypt-kmip-server-1@sha256:9af5d21dd748dfbf4e107f27a5f10fa3c0ba928ade9b989a744b620ce34ec757
        version: 0.0.17
      entries:
      - name: zts-xcrypt-operator.v0.0.17
        version: 0.0.17
      name: alpha
    defaultChannel: alpha
    packageName: xcrypt-operator
    provider:
      name: Zettaset
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-spark-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spark-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:a6c82a6d6ebaa0d2572d644ac9013451e020f134c6fe46222fef37b1e9aa3cbb
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:a6c82a6d6ebaa0d2572d644ac9013451e020f134c6fe46222fef37b1e9aa3cbb
        version: 24.11.1
      entries:
      - name: spark-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: spark-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:8ad9948bf8264924a0eedd92c2e58269117487b9d57816384447bb6c999be681
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:8ad9948bf8264924a0eedd92c2e58269117487b9d57816384447bb6c999be681
        version: 25.3.0
      entries:
      - name: spark-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: spark-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkConnectServerSpec via
              `CustomResource`
            displayName: sparkconnectservers.spark.stackable.tech
            kind: SparkConnectServer
            name: sparkconnectservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
        version: 25.7.0
      entries:
      - name: spark-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: spark-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkConnectServerSpec via
              `CustomResource`
            displayName: sparkconnectservers.spark.stackable.tech
            kind: SparkConnectServer
            name: sparkconnectservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
        version: 25.7.0
      entries:
      - name: spark-operator.v25.7.0
        version: 25.7.0
      - name: spark-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-spark-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://skupper.io
    name: skupper-netobs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-netobs-operator.v2.1.1-rh-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability.skupper.io/v2alpha1",
                "kind": "NetworkObserver",
                "metadata": {
                  "name": "networkobserver-sample"
                },
                "spec": {
                  "auth": {
                    "basic": {
                      "create": true,
                      "secretName": ""
                    },
                    "openshift": {
                      "cookieSecretName": "",
                      "createCookieSecret": true,
                      "serviceAccount": {
                        "create": true,
                        "nameOverride": ""
                      }
                    },
                    "strategy": "openshift"
                  },
                  "extraArgs": null,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "registry.redhat.io/service-interconnect/skupper-network-observer-rhel9",
                    "tag": "2.1.1"
                  },
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nginx": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/ubi9/nginx-124",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": 9.6
                  },
                  "openshiftOauthProxy": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-oauth-proxy-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.19.0"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "prometheus": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-prometheus-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.16.0"
                  },
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "router": {
                    "certificate": {
                      "create": true,
                      "nameOverride": ""
                    },
                    "endpoint": "amqps://skupper-router-local"
                  },
                  "securityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    }
                  },
                  "service": {
                    "port": 443,
                    "type": "ClusterIP"
                  },
                  "skipManagementLabels": false,
                  "tls": {
                    "openshiftIssued": true,
                    "secretName": "",
                    "skupperIssued": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
          createdAt: "2025-08-27T20:21:18Z"
          description: Red Hat Service Interconnect Network Observer Operator provides
            the ability to monitor and observe a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkObserver
            name: networkobservers.observability.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect Network Observer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/ubi9/nginx-124@sha256:cbcca4743015b4ef167383a080f4711a6049e275e0b66f80a74f8b087b0dfc7f
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:d3056b35d9a205b9f2c48d924f199c5ac23904eb18d526e4bff229e7c7181415
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
        version: 2.1.1-rh-8
      entries:
      - name: skupper-netobs-operator.v2.1.1-rh-8
        version: 2.1.1-rh-8
      name: stable-2
    - currentCSV: skupper-netobs-operator.v2.1.1-rh-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability.skupper.io/v2alpha1",
                "kind": "NetworkObserver",
                "metadata": {
                  "name": "networkobserver-sample"
                },
                "spec": {
                  "auth": {
                    "basic": {
                      "create": true,
                      "secretName": ""
                    },
                    "openshift": {
                      "cookieSecretName": "",
                      "createCookieSecret": true,
                      "serviceAccount": {
                        "create": true,
                        "nameOverride": ""
                      }
                    },
                    "strategy": "openshift"
                  },
                  "extraArgs": null,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "registry.redhat.io/service-interconnect/skupper-network-observer-rhel9",
                    "tag": "2.1.1"
                  },
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nginx": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/ubi9/nginx-124",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": 9.6
                  },
                  "openshiftOauthProxy": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-oauth-proxy-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.19.0"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "prometheus": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-prometheus-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.16.0"
                  },
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "router": {
                    "certificate": {
                      "create": true,
                      "nameOverride": ""
                    },
                    "endpoint": "amqps://skupper-router-local"
                  },
                  "securityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    }
                  },
                  "service": {
                    "port": 443,
                    "type": "ClusterIP"
                  },
                  "skipManagementLabels": false,
                  "tls": {
                    "openshiftIssued": true,
                    "secretName": "",
                    "skupperIssued": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
          createdAt: "2025-08-27T20:21:18Z"
          description: Red Hat Service Interconnect Network Observer Operator provides
            the ability to monitor and observe a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkObserver
            name: networkobservers.observability.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect Network Observer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:d3056b35d9a205b9f2c48d924f199c5ac23904eb18d526e4bff229e7c7181415
        - registry.redhat.io/ubi9/nginx-124@sha256:cbcca4743015b4ef167383a080f4711a6049e275e0b66f80a74f8b087b0dfc7f
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        version: 2.1.1-rh-8
      entries:
      - name: skupper-netobs-operator.v2.1.1-rh-8
        version: 2.1.1-rh-8
      name: stable-2.1
    defaultChannel: stable-2
    packageName: skupper-netobs-operator
    provider:
      name: Red Hat
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/trustification/trusted-profile-analyzer-operator
    name: rhtpa-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhtpa-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtpa.io/v1",
                "kind": "TrustedProfileAnalyzer",
                "metadata": {
                  "name": "trustedprofileanalyzer-sample"
                },
                "spec": {
                  "appDomain": "change-me",
                  "collector": {},
                  "database": {},
                  "image": {
                    "fullName": "registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e",
                    "pullPolicy": "IfNotPresent"
                  },
                  "infrastructure": {
                    "port": 9010
                  },
                  "ingress": {},
                  "metrics": {
                    "enabled": false
                  },
                  "modules": {
                    "createDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "createImporters": {
                      "enabled": false,
                      "image": {},
                      "importers": {}
                    },
                    "importer": {
                      "concurrency": 1,
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {},
                      "workingDirectory": {
                        "size": "32Gi"
                      }
                    },
                    "migrateDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "server": {
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "ingress": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {}
                    }
                  },
                  "oidc": {
                    "clients": {
                      "cli": {
                        "clientSecret": null
                      },
                      "frontend": {}
                    }
                  },
                  "openshift": {
                    "useServiceCa": true
                  },
                  "partOf": "trustify",
                  "replicas": 1,
                  "rust": {},
                  "storage": {},
                  "tls": {},
                  "tracing": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-09-19T15:04:16Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Profile Analyzer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trustification/trusted-profile-analyzer-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: TrustedProfileAnalyzer
            name: trustedprofileanalyzers.rhtpa.io
            version: v1
        description: |
          [Tech Preview]
          Red Hat Trusted Profile Analyzer (RHTPA) provides a single source of truth for your software inventory across all your environments. You can verify the integrity of your in-house and third-party applications against their respective Software Bill of Materials (SBOM) document.  Trusted Profile Analyzer provides vulnerability, component and license transparency, strengthens your software supply chain security, and enables an accurate, up-to-date understanding of your software inventory.

          Features and Benefits:

          Enhanced Software Supply Chain Security: Offers critical transparency and integrity verification across the entire software lifecycle.

          Reduced Operational Risk: Proactively identifies and helps remediate vulnerabilities minimizing risk and compliance violations.

          Improved Compliance and Auditability: Simplifies the process of meeting regulatory requirements for software transparency and integrity by providing verifiable documentation and continuous monitoring.

          Faster Remediation and Incident Response: Provides targeted, actionable insights to prioritize and accelerate the patching and remediation of vulnerabilities.

          Increased Developer Productivity & Trust: Integrates security into existing workflows without significant overhead, allowing developers to build and deploy with greater confidence in the integrity of their code and components.

          Centralized Visibility and Control: Consolidates disparate security information into a single pane of glass, offering a holistic view of your application security posture.

          Red Hat Trusted Profile Analyzer is a crucial tool for any organization running applications on OpenShift that needs to confidently understand, verify, and secure the software they deploy, especially in an era of complex software supply chains and increasing regulatory scrutiny.

          [Documentation](https://docs.redhat.com/en/documentation/red_hat_trusted_profile_analyzer/2.1).
        displayName: Red Hat Trusted Profile Analyzer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rhtpa
        - trusted-profile-analyzer
        links:
        - name: Trusted Profile Analyzer Operator
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        maintainers:
        - email: rhtpa-support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        relatedImages:
        - registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
        - registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e
        version: 1.0.2
      entries:
      - name: rhtpa-operator.v1.0.2
        version: 1.0.2
      - name: rhtpa-operator.v1.0.1
        version: 1.0.1
      - name: rhtpa-operator.v1.0.0
        version: 1.0.0
      name: stable
    - currentCSV: rhtpa-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtpa.io/v1",
                "kind": "TrustedProfileAnalyzer",
                "metadata": {
                  "name": "trustedprofileanalyzer-sample"
                },
                "spec": {
                  "appDomain": "change-me",
                  "collector": {},
                  "database": {},
                  "image": {
                    "fullName": "registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e",
                    "pullPolicy": "IfNotPresent"
                  },
                  "infrastructure": {
                    "port": 9010
                  },
                  "ingress": {},
                  "metrics": {
                    "enabled": false
                  },
                  "modules": {
                    "createDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "createImporters": {
                      "enabled": false,
                      "image": {},
                      "importers": {}
                    },
                    "importer": {
                      "concurrency": 1,
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {},
                      "workingDirectory": {
                        "size": "32Gi"
                      }
                    },
                    "migrateDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "server": {
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "ingress": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {}
                    }
                  },
                  "oidc": {
                    "clients": {
                      "cli": {
                        "clientSecret": null
                      },
                      "frontend": {}
                    }
                  },
                  "openshift": {
                    "useServiceCa": true
                  },
                  "partOf": "trustify",
                  "replicas": 1,
                  "rust": {},
                  "storage": {},
                  "tls": {},
                  "tracing": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-09-19T15:04:16Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Profile Analyzer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trustification/trusted-profile-analyzer-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: TrustedProfileAnalyzer
            name: trustedprofileanalyzers.rhtpa.io
            version: v1
        description: |
          [Tech Preview]
          Red Hat Trusted Profile Analyzer (RHTPA) provides a single source of truth for your software inventory across all your environments. You can verify the integrity of your in-house and third-party applications against their respective Software Bill of Materials (SBOM) document.  Trusted Profile Analyzer provides vulnerability, component and license transparency, strengthens your software supply chain security, and enables an accurate, up-to-date understanding of your software inventory.

          Features and Benefits:

          Enhanced Software Supply Chain Security: Offers critical transparency and integrity verification across the entire software lifecycle.

          Reduced Operational Risk: Proactively identifies and helps remediate vulnerabilities minimizing risk and compliance violations.

          Improved Compliance and Auditability: Simplifies the process of meeting regulatory requirements for software transparency and integrity by providing verifiable documentation and continuous monitoring.

          Faster Remediation and Incident Response: Provides targeted, actionable insights to prioritize and accelerate the patching and remediation of vulnerabilities.

          Increased Developer Productivity & Trust: Integrates security into existing workflows without significant overhead, allowing developers to build and deploy with greater confidence in the integrity of their code and components.

          Centralized Visibility and Control: Consolidates disparate security information into a single pane of glass, offering a holistic view of your application security posture.

          Red Hat Trusted Profile Analyzer is a crucial tool for any organization running applications on OpenShift that needs to confidently understand, verify, and secure the software they deploy, especially in an era of complex software supply chains and increasing regulatory scrutiny.

          [Documentation](https://docs.redhat.com/en/documentation/red_hat_trusted_profile_analyzer/2.1).
        displayName: Red Hat Trusted Profile Analyzer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rhtpa
        - trusted-profile-analyzer
        links:
        - name: Trusted Profile Analyzer Operator
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        maintainers:
        - email: rhtpa-support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        relatedImages:
        - registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
        - registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e
        version: 1.0.2
      entries:
      - name: rhtpa-operator.v1.0.2
        version: 1.0.2
      - name: rhtpa-operator.v1.0.1
        version: 1.0.1
      - name: rhtpa-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable
    packageName: rhtpa-operator
    provider:
      name: Red Hat
      url: https://github.com/trustification/trusted-profile-analyzer-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: nova-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nova-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "Nova",
                "metadata": {
                  "name": "nova"
                },
                "spec": {
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaAPI",
                "metadata": {
                  "name": "novaapi-sample"
                },
                "spec": {
                  "apiDatabaseHostname": "openstack",
                  "cell0DatabaseHostname": "openstack",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "registeredCells": {},
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaCell",
                "metadata": {
                  "name": "novacell1-with-upcall-sample"
                },
                "spec": {
                  "apiDatabaseHostname": "openstack",
                  "cellDatabaseHostname": "openstack",
                  "cellName": "cell1",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaCompute",
                "metadata": {
                  "name": "nova-compute-ironic"
                },
                "spec": {
                  "novaComputeTemplates": {
                    "compute-ironic": {
                      "computeDriver": "ironic.IronicDriver",
                      "customServiceConfig": "",
                      "networkAttachments": [
                        "internalapi"
                      ],
                      "replicas": 1,
                      "resources": {}
                    }
                  }
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaConductor",
                "metadata": {
                  "name": "novaconductor-sample"
                },
                "spec": {
                  "cellName": "cell1",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaMetadata",
                "metadata": {
                  "name": "novametadata-sample"
                },
                "spec": {
                  "apiDatabaseAccount": "nova-api",
                  "apiDatabaseHostname": "openstack",
                  "cellDatabaseAccount": "nova-cell0",
                  "cellDatabaseHostname": "openstack",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "registeredCells": {},
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaNoVNCProxy",
                "metadata": {
                  "name": "novanovncproxy-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaScheduler",
                "metadata": {
                  "name": "novascheduler-sample"
                },
                "spec": {
                  "apiDatabaseHostname": "openstack",
                  "cell0DatabaseHostname": "openstack",
                  "customServiceConfig": "[DEFAULT]\ndebug=true\n",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "registeredCells": {},
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:21Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nova
            name: nova.nova.openstack.org
            version: v1beta1
          - description: NovaAPI is the Schema for the novaapis API
            displayName: Nova API
            kind: NovaAPI
            name: novaapis.nova.openstack.org
            version: v1beta1
          - description: NovaCell is the Schema for the novacells API
            displayName: Nova Cell
            kind: NovaCell
            name: novacells.nova.openstack.org
            version: v1beta1
          - description: NovaCompute is the Schema for the NovaCompute
            displayName: Nova Compute
            kind: NovaCompute
            name: novacomputes.nova.openstack.org
            version: v1beta1
          - description: NovaConductor is the Schema for the novaconductors API
            displayName: Nova Conductor
            kind: NovaConductor
            name: novaconductors.nova.openstack.org
            version: v1beta1
          - description: NovaMetadata is the Schema for the novametadata API
            displayName: Nova Metadata
            kind: NovaMetadata
            name: novametadata.nova.openstack.org
            version: v1beta1
          - description: NovaNoVNCProxy is the Schema for the novanovncproxies API
            displayName: Nova No VNCProxy
            kind: NovaNoVNCProxy
            name: novanovncproxies.nova.openstack.org
            version: v1beta1
          - description: NovaScheduler is the Schema for the novaschedulers API
            displayName: Nova Scheduler
            kind: NovaScheduler
            name: novaschedulers.nova.openstack.org
            version: v1beta1
        description: Nova Operator
        displayName: Nova Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Nova
        links:
        - name: Nova Operator
          url: https://github.com/openstack-k8s-operators/nova-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-nova-scheduler-rhel9@sha256:ba91196cd4f9b6e6f49868c91343b8eca7cc0d5016c36e83e2cc76f45cbe777b
        - registry.redhat.io/rhoso/openstack-nova-compute-rhel9@sha256:43cbe46ab263403fc303b6bc2388863029aad33465b237db18da7cd6d42078e1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/nova-rhel9-operator@sha256:f058c56df02a108505d51b9d19973894c2ee59055776c9f3fd7f7afe23a5106a
        - registry.redhat.io/rhoso/openstack-nova-api-rhel9@sha256:bbb8ebceb6f30b6c8beb120429e69f323bd231c09aaaddc5be9f15a527e928ed
        - registry.redhat.io/rhoso/openstack-nova-conductor-rhel9@sha256:500c36fd519611be8662e97ca31b1e59269c14d8aa94286dad9925e957462155
        - registry.redhat.io/rhoso/openstack-nova-novncproxy-rhel9@sha256:bf534f0c72f188d5f51e40b0392ace182db2055dff9ad14b01eeef13056febbf
        version: 1.0.6
      entries:
      - name: nova-operator.v1.0.6
        version: 1.0.6
      - name: nova-operator.v1.0.5
        version: 1.0.5
      - name: nova-operator.v1.0.4
        version: 1.0.4
      - name: nova-operator.v1.0.3
        version: 1.0.3
      - name: nova-operator.v1.0.2
        version: 1.0.2
      - name: nova-operator.v1.0.1
        version: 1.0.1
      - name: nova-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: nova-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Cloudera
      provider-url: https://cloudera.com
    name: cfm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cfm-operator.v2.9.0-b96
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "Nifi",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifi-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifi",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifi-sample"
                },
                "spec": {
                  "hostName": "nifi.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifi-k8s",
                    "tag": "2.8.0-b94-nifi_1.25.0.2.3.13.0-36"
                  },
                  "replicas": 1,
                  "stateManagement": {
                    "clusterProvider": {
                      "class": "org.apache.nifi.kubernetes.state.provider.KubernetesConfigMapStateProvider",
                      "id": "kubernetes-provider"
                    }
                  },
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.8.0-b94"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              },
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "NifiRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifiregistry-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifiregistry",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifiregistry-sample"
                },
                "spec": {
                  "hostName": "registry.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s",
                    "tag": "2.8.0-b94-nifi_1.25.0.2.3.13.0-36"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.8.0-b94"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning,Big Data
          com.redhat.openshift.versions: v4.10
          containerImage: container.repository.cloudera.com/cloudera/cfm-operator@sha256:890d81188143ccf9ebe2c2fe5e039cdff3a2df53f7512555673f0468147e991e
          createdAt: "2024-12-06T19:03:24Z"
          description: Cloudera Flow Management - Kubernetes Operator for Apache NiFideploys
            and operates Apache NiFi
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: A valid Cloudera license is required.
            Contact a Cloudera account team to learn more.
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NifiRegistry is the Schema for the nifiregistries API.
            displayName: Nifi Registry
            kind: NifiRegistry
            name: nifiregistries.cfm.cloudera.com
            version: v1alpha1
          - description: Nifi is the Schema for the nifis API.
            displayName: Nifi
            kind: Nifi
            name: nifis.cfm.cloudera.com
            version: v1alpha1
        description: |-
          Cloudera Flow Management - Kubernetes Operator for Apache NiFi brings enterprise-grade NiFi and NiFi Registry deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of NiFi and NiFi Registry for enterprise deployments.
          To learn more about Cloudera Flow Management - Kubernetes Operator for Apache NiFi read this
          [deep dive blog post](https://community.cloudera.com/t5/Community-Articles/Cloudera-Flow-Management-Operator-A-technical-deep-dive/ta-p/390473).

          Cloudera Flow Management - Kubernetes Operator for Apache NiFi consists of and ships multiple components including Apache NiFi, Apache NiFi Registry,
          and a custom sidecar image for logging aggregation.

          ## Installation prerequisites

          Installing Cloudera Flow Management - Kubernetes Operator for Apache NiFi requires the a valid license as well as Cloudera credentials
          (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly.
          The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) from which installation artifacts are pulled.

          ### Obtaining a license
          Contact [Cloudera sales](https://cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets
          Before installation, create two `Secrets`. One containing your Cloudera credentials and one containing your license. Cloudera credentials are provided to you
          together with your license.

          1. Create a `Secret` containing your Cloudera credentials.
              `oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***]
               --docker-server=container.repository.cloudera.com`

          2. Create a `Secret` containing your license.
              `oc create secret generic cfm-operator-license --from-file=license.txt=[***PATH TO LICENSE FILE***]`

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces in which you
            plan to deploy NiFi clusters or NiFi Registry deployments. Cloudera recommends that you create the `Secret` in all required namespaces before installation
            if you know which namespaces you will be using to deploy NiFi clusters or NiFi Registry deployments.

          ## Learn more
          To learn more, visit the [Cloudera Flow Management - Kubernetes Operator for Apache NiFi documentation library](https://docs.cloudera.com/cfm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Flow Management - Kubernetes Operator for Apache NiFi, visit the following resources:
          * [Apache NiFi](https://nifi.apache.org/)
          * [Apache NiFi Registry](https://nifi.apache.org/registry.html)
        displayName: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        - nifiregistry
        - cloudera
        - cfm
        links:
        - name: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
          url: https://docs.cloudera.com/cfm-operator
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s@sha256:914dcde9ee97894fa6f5deec285e53bf81d196e071a997cabb40fedd52deccc3
        - container.repository.cloudera.com/cloudera/cfm-tini@sha256:6493d23e20fca8daf31823403fca3dd9b42da983464fbe788a6863e141e8b66b
        - container.repository.cloudera.com/cloudera/cfm-nifi-k8s@sha256:0d700b7eff096dfc362868345826c3a6460fa43a700273b1422e95015b7727ee
        - container.repository.cloudera.com/cloudera/cfm-operator@sha256:890d81188143ccf9ebe2c2fe5e039cdff3a2df53f7512555673f0468147e991e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86af06fb3b6c17b22be47118b0a8c797fa7fdd3c00dbe99a2d2a593ca8fc3d63
        version: 2.9.0-b96
      entries:
      - name: cfm-operator.v2.9.0-b96
        version: 2.9.0-b96
      name: alpha
    - currentCSV: cfm-operator.v2.10.0-b135
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "Nifi",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifi-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifi",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifi-sample"
                },
                "spec": {
                  "hostName": "nifi.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifi-k8s",
                    "tag": "2.9.0-b96-nifi_1.27.0.2.3.14.0-14"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.9.0-b96"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              },
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "NifiRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifiregistry-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifiregistry",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifiregistry-sample"
                },
                "spec": {
                  "hostName": "registry.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s",
                    "tag": "2.9.0-b96-nifi_1.27.0.2.3.14.0-14"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.9.0-b96"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning,Big Data
          com.redhat.openshift.versions: v4.10
          containerImage: container.repository.cloudera.com/cloudera/cfm-operator@sha256:0312b6dadd5904ea9507bb857ab89d7266bf244c93b86855e0528e72813b7a85
          createdAt: "2025-05-05T14:39:51Z"
          description: Cloudera Flow Management - Kubernetes Operator for Apache NiFideploys
            and operates Apache NiFi
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: A valid Cloudera license is required.
            Contact a Cloudera account team to learn more.
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NifiRegistry is the Schema for the nifiregistries API.
            displayName: Nifi Registry
            kind: NifiRegistry
            name: nifiregistries.cfm.cloudera.com
            version: v1alpha1
          - description: Nifi is the Schema for the nifis API.
            displayName: Nifi
            kind: Nifi
            name: nifis.cfm.cloudera.com
            version: v1alpha1
        description: |-
          Cloudera Flow Management - Kubernetes Operator for Apache NiFi brings enterprise-grade NiFi and NiFi Registry deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of NiFi and NiFi Registry for enterprise deployments.
          To learn more about Cloudera Flow Management - Kubernetes Operator for Apache NiFi read this
          [deep dive blog post](https://community.cloudera.com/t5/Community-Articles/Cloudera-Flow-Management-Operator-A-technical-deep-dive/ta-p/390473).

          Cloudera Flow Management - Kubernetes Operator for Apache NiFi consists of and ships multiple components including Apache NiFi, Apache NiFi Registry,
          and a custom sidecar image for logging aggregation.

          ## Installation prerequisites

          Installing Cloudera Flow Management - Kubernetes Operator for Apache NiFi requires the a valid license as well as Cloudera credentials
          (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly.
          The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) from which installation artifacts are pulled.

          ### Obtaining a license
          Contact [Cloudera sales](https://cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets
          Before installation, create two `Secrets`. One containing your Cloudera credentials and one containing your license. Cloudera credentials are provided to you
          together with your license.

          1. Create a `Secret` containing your Cloudera credentials.
              `oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***]
               --docker-server=container.repository.cloudera.com --namespace [***OPERATOR INSTALL NAMESPACE***]`

          2. Create a `Secret` containing your license.
              `oc create secret generic cfm-operator-license --from-file=license.txt=[***PATH TO LICENSE FILE***] --namespace [***OPERATOR INSTALL NAMESPACE***]`

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces in which you
            plan to deploy NiFi clusters or NiFi Registry deployments. Cloudera recommends that you create the `Secret` in all required namespaces before installation
            if you know which namespaces you will be using to deploy NiFi clusters or NiFi Registry deployments.

          ## Learn more
          To learn more, visit the [Cloudera Flow Management - Kubernetes Operator for Apache NiFi documentation library](https://docs.cloudera.com/cfm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Flow Management - Kubernetes Operator for Apache NiFi, visit the following resources:
          * [Apache NiFi](https://nifi.apache.org/)
          * [Apache NiFi Registry](https://nifi.apache.org/registry.html)
        displayName: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        - nifiregistry
        - cloudera
        - cfm
        links:
        - name: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
          url: https://docs.cloudera.com/cfm-operator
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/cfm-nifi-k8s@sha256:8607c81906422d65784c6dca74a37af3b7c87012bca65ce28a995366014aec86
        - container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s@sha256:52e00f102579b972500eb187e64a7662a6a46caa883df743e454b1d5111b34e5
        - container.repository.cloudera.com/cloudera/cfm-operator@sha256:0312b6dadd5904ea9507bb857ab89d7266bf244c93b86855e0528e72813b7a85
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86af06fb3b6c17b22be47118b0a8c797fa7fdd3c00dbe99a2d2a593ca8fc3d63
        - container.repository.cloudera.com/cloudera/cfm-tini@sha256:d42ad5f05a9bb283fed6dbbd820855714a31662f5b851e691f13703153e90d13
        version: 2.10.0-b135
      entries:
      - name: cfm-operator.v2.10.0-b135
        version: 2.10.0-b135
      name: stable
    defaultChannel: stable
    packageName: cfm-operator
    provider:
      name: Cloudera
      url: https://cloudera.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Splunk Inc.
      provider-url: www.splunk.com
    name: splunk-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: splunk-operator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "IndexerCluster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "indexercluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "LicenseMaster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "licensemaster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "MonitoringConsole",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "monitoringconsole-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "SearchHeadCluster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "searchheadcluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "Standalone",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "standalone-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "clustermanager-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "IndexerCluster",
                "metadata": {
                  "name": "indexercluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "LicenseManager",
                "metadata": {
                  "name": "licensemanager-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "MonitoringConsole",
                "metadata": {
                  "name": "monitoringconsole-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "SearchHeadCluster",
                "metadata": {
                  "name": "searchheadcluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "Standalone",
                "metadata": {
                  "name": "standalone-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Big Data, Logging & Tracing, Monitoring, Security, AI/Machine
            Learning
          containerImage: splunk/splunk-operator@sha256:ed02561995d46ff8eea143fa78f5da118ee6fdb500832e9a33b42d059a67e229
          createdAt: "2025-09-08T10:17:59Z"
          description: The Splunk Operator for Kubernetes enables you to quickly and
            easily deploy Splunk Enterprise on your choice of private or public cloud
            provider. The Operator simplifies scaling and management of Splunk Enterprise
            by automating administrative workflows using Kubernetes best practices.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=2.4.0 <3.0.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/splunk/splunk-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterManager is the Schema for the cluster manager API
            displayName: Cluster Manager
            kind: ClusterManager
            name: clustermanagers.enterprise.splunk.com
            version: v4
          - description: ClusterMaster is the Schema for the cluster manager API
            displayName: Cluster Master
            kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v3
          - kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v1
          - kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v2
          - description: IndexerCluster is the Schema for a Splunk Enterprise indexer
              cluster
            displayName: Indexer Cluster
            kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v3
          - description: IndexerCluster is the Schema for a Splunk Enterprise indexer
              cluster
            displayName: Indexer Cluster
            kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v4
          - kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v1
          - kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v2
          - description: LicenseManager is the Schema for a Splunk Enterprise license
              manager.
            displayName: License Manager
            kind: LicenseManager
            name: licensemanagers.enterprise.splunk.com
            version: v4
          - description: LicenseMaster is the Schema for a Splunk Enterprise license
              master.
            displayName: License Master
            kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v3
          - kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v1
          - kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v2
          - description: MonitoringConsole is the Schema for the monitoringconsole
              API
            displayName: Monitoring Console
            kind: MonitoringConsole
            name: monitoringconsoles.enterprise.splunk.com
            version: v3
          - description: MonitoringConsole is the Schema for the monitoringconsole
              API
            displayName: Monitoring Console
            kind: MonitoringConsole
            name: monitoringconsoles.enterprise.splunk.com
            version: v4
          - description: SearchHeadCluster is the Schema for a Splunk Enterprise search
              head cluster
            displayName: Search Head Cluster
            kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v3
          - description: SearchHeadCluster is the Schema for a Splunk Enterprise search
              head cluster
            displayName: Search Head Cluster
            kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v4
          - kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v1
          - kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v2
          - description: Standalone is the Schema for a Splunk Enterprise standalone
              instances.
            displayName: Standalone
            kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v3
          - description: Standalone is the Schema for a Splunk Enterprise standalone
              instances.
            displayName: Standalone
            kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v4
          - kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v1
          - kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v2
        description: The Splunk Operator for Kubernetes enables you to quickly and
          easily deploy Splunk Enterprise on your choice of private or public cloud
          provider. The Operator simplifies scaling and management of Splunk Enterprise
          by automating administrative workflows using Kubernetes best practices.
        displayName: Splunk Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Splunk
        - Enterprise
        - Operator
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://github.com/splunk/splunk-operator/tree/main/docs
        maintainers:
        - email: support@splunk.com
          name: Splunk Operator Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Splunk Inc.
          url: www.splunk.com
        relatedImages:
        - splunk/splunk-operator@sha256:ed02561995d46ff8eea143fa78f5da118ee6fdb500832e9a33b42d059a67e229
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - splunk/splunk@sha256:4a5ce5a7b4e84a87396f0664895095c6510a0f2bf52e8747a373c7357f3ee313
        version: 3.0.0
      entries:
      - name: splunk-operator.v3.0.0
        version: 3.0.0
      - name: splunk-operator.v2.8.1
        version: 2.8.1
      - name: splunk-operator.v2.8.0
        version: 2.8.0
      - name: splunk-operator.v2.7.1
        version: 2.7.1
      - name: splunk-operator.v2.7.0
        version: 2.7.0
      name: stable
    defaultChannel: stable
    packageName: splunk-operator
    provider:
      name: Splunk Inc.
      url: www.splunk.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-emrcontainers-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-emrcontainers-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "emrcontainers.services.k8s.aws/v1alpha1",
                "kind": "VirtualCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/emrcontainers-controller:1.1.1
          createdAt: "2025-09-19T19:36:26Z"
          description: AWS EMR on EKS controller is a service controller for managing
            EMR on EKS resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: JobRun represents the state of an AWS emrcontainers JobRun
              resource.
            displayName: JobRun
            kind: JobRun
            name: jobruns.emrcontainers.services.k8s.aws
            version: v1alpha1
          - description: VirtualCluster represents the state of an AWS emrcontainers
              VirtualCluster resource.
            displayName: VirtualCluster
            kind: VirtualCluster
            name: virtualclusters.emrcontainers.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EMR on EKS resources in AWS from within your Kubernetes cluster.

          **About Amazon EMR on EKS**

          Amazon EMR on EKS provides a deployment option for Amazon EMR that allows you to run open-source big data frameworks on Amazon Elastic Kubernetes Service (Amazon EKS). With this deployment option, you can focus on running analytics workloads while Amazon EMR on EKS builds, configures, and manages containers for open-source applications.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EMR on EKS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - emrcontainers
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EMR on EKS Developer Resources
          url: https://docs.aws.amazon.com/emr/latest/EMR-on-EKS-DevelopmentGuide/emr-eks.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: emr on eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/emrcontainers-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-emrcontainers-controller.v1.1.1
        version: 1.1.1
      - name: ack-emrcontainers-controller.v1.1.0
        version: 1.1.0
      - name: ack-emrcontainers-controller.v1.0.27
        version: 1.0.27
      - name: ack-emrcontainers-controller.v1.0.25
        version: 1.0.25
      - name: ack-emrcontainers-controller.v1.0.24
        version: 1.0.24
      - name: ack-emrcontainers-controller.v1.0.23
        version: 1.0.23
      - name: ack-emrcontainers-controller.v1.0.22
        version: 1.0.22
      - name: ack-emrcontainers-controller.v1.0.21
        version: 1.0.21
      - name: ack-emrcontainers-controller.v1.0.18
        version: 1.0.18
      - name: ack-emrcontainers-controller.v1.0.16
        version: 1.0.16
      - name: ack-emrcontainers-controller.v1.0.15
        version: 1.0.15
      - name: ack-emrcontainers-controller.v1.0.14
        version: 1.0.14
      - name: ack-emrcontainers-controller.v1.0.13
        version: 1.0.13
      - name: ack-emrcontainers-controller.v1.0.12
        version: 1.0.12
      - name: ack-emrcontainers-controller.v1.0.11
        version: 1.0.11
      - name: ack-emrcontainers-controller.v1.0.9
        version: 1.0.9
      - name: ack-emrcontainers-controller.v1.0.8
        version: 1.0.8
      - name: ack-emrcontainers-controller.v1.0.4
        version: 1.0.4
      - name: ack-emrcontainers-controller.v1.0.3
        version: 1.0.3
      - name: ack-emrcontainers-controller.v1.0.1
        version: 1.0.1
      - name: ack-emrcontainers-controller.v1.0.0
        version: 1.0.0
      - name: ack-emrcontainers-controller.v0.1.2
        version: 0.1.2
      - name: ack-emrcontainers-controller.v0.1.0
        version: 0.1.0
      - name: ack-emrcontainers-controller.v0.0.10
        version: 0.0.10
      - name: ack-emrcontainers-controller.v0.0.9
        version: 0.0.9
      - name: ack-emrcontainers-controller.v0.0.8
        version: 0.0.8
      - name: ack-emrcontainers-controller.v0.0.7
        version: 0.0.7
      - name: ack-emrcontainers-controller.v0.0.6
        version: 0.0.6
      - name: ack-emrcontainers-controller.v0.0.5
        version: 0.0.5
      - name: ack-emrcontainers-controller.v0.0.4
        version: 0.0.4
      - name: ack-emrcontainers-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-emrcontainers-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: volume-expander-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: volume-expander-operator.v0.3.6
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage,Monitoring
          certified: "false"
          containerImage: quay.io/redhat-cop/volume-expander-operator@sha256:e4059a66a9af7c786c8c651abb8a57a55a0e528342e3a59f8c264b40f87d6e25
          createdAt: "2022-05-05T13:25:53Z"
          description: This operator provides automation expanding pod's volumes when
            they are about to run out of space.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: volume-expander-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/volume-expander-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The purpose of the volume-expander-operator is to expand volumes when they are running out of space.
          This is achieved by using the [volume expansion feature](https://kubernetes.io/docs/concepts/storage/persistent-volumes/#expanding-persistent-volumes-claims).

          The operator periodically checks the `kubelet_volume_stats_used_bytes` and `kubelet_volume_stats_capacity_bytes published` by the kubelets to decide when to expand a volume.
          Notice that these metrics are generated only when a volume is mounted to a pod. Also the kubelet takes a minute or two to start generating accurate values for these metrics. The operator accounts for that.

          This operator works based on the following annotations to PersistentVolumeClaim resources:

          | Annotation | Default &nbsp;  &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;  | Description  |
          | :--- | :--- | :--- |
          | `volume-expander-operator.redhat-cop.io/autoexpand`  | N/A  | if set to "true" enables the volume-expander-operator to watch on this PVC  |
          | `volume-expander-operator.redhat-cop.io/polling-frequency`  | `"30s"` | How frequently to poll the volume metrics. Express this value as a valid golang [Duration](https://golang.org/pkg/time/#ParseDuration)  |
          | `volume-expander-operator.redhat-cop.io/expand-threshold-percent` | `"80"` | the percentage of used storage after which the volume will be expanded. This must be a positive integer. |
          | `volume-expander-operator.redhat-cop.io/expand-by-percent` | `"25"` | the percentage by which the volume will be expanded, relative to the current size. This must be an integer between 0 and 100 |
          | `volume-expander-operator.redhat-cop.io/expand-up-to` | MaxInt64 | the upper bound for this volume to be expanded to. The default value is the largest quantity representable and is intended to be interpreted as infinite. If the default is used it is recommend to ensure the namespace has a quota on the used storage class. |

          Note that not all of the storage driver implementations support volume expansion. It is a responsibility of the user/platform administrator to ensure that storage class and the persistent volume claim meet all the requirements needed for the volume expansion feature to work properly.

          This operator was tested with [OCS](https://www.redhat.com/en/technologies/cloud-computing/openshift-container-storage), but should work with any other storage driver that supports volume expansion.
        displayName: Volume Expander Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - volume
        - storage
        - csi
        - expansion
        links:
        - name: Volume Expander Operator
          url: https://github.com/redhat-cop/volume-expander-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/volume-expander-operator@sha256:e4059a66a9af7c786c8c651abb8a57a55a0e528342e3a59f8c264b40f87d6e25
        version: 0.3.6
      entries:
      - name: volume-expander-operator.v0.3.6
        version: 0.3.6
      - name: volume-expander-operator.v0.3.5
        version: 0.3.5
      - name: volume-expander-operator.v0.3.4
        version: 0.3.4
      - name: volume-expander-operator.v0.3.3
        version: 0.3.3
      - name: volume-expander-operator.v0.3.2
        version: 0.3.2
      - name: volume-expander-operator.v0.3.1
        version: 0.3.1
      - name: volume-expander-operator.v0.3.0
        version: 0.3.0
      - name: volume-expander-operator.v0.2.0
        version: 0.2.0
      - name: volume-expander-operator.v0.1.8
        version: 0.1.8
      - name: volume-expander-operator.v0.1.7
        version: 0.1.7
      - name: volume-expander-operator.v0.1.6
        version: 0.1.6
      - name: volume-expander-operator.v0.1.5
        version: 0.1.5
      - name: volume-expander-operator.v0.1.4
        version: 0.1.4
      - name: volume-expander-operator.v0.1.3
        version: 0.1.3
      - name: volume-expander-operator.v0.1.2
        version: 0.1.2
      - name: volume-expander-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: volume-expander-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-eventbridge-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-eventbridge-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Archive",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Endpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "EventBus",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/eventbridge-controller:1.1.1
          createdAt: "2025-09-19T20:51:55Z"
          description: AWS EventBridge controller is a service controller for managing
            EventBridge resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Archive represents the state of an AWS eventbridge Archive
              resource.
            displayName: Archive
            kind: Archive
            name: archives.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: Endpoint represents the state of an AWS eventbridge Endpoint
              resource.
            displayName: Endpoint
            kind: Endpoint
            name: endpoints.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: EventBus represents the state of an AWS eventbridge EventBus
              resource.
            displayName: EventBus
            kind: EventBus
            name: eventbuses.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: Rule represents the state of an AWS eventbridge Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.eventbridge.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EventBridge resources in AWS from within your Kubernetes cluster.

          **About Amazon EventBridge**

          EventBridge is a serverless service that uses events to connect application components together, making it easier for you to build scalable event-driven applications. Use it to route events from sources such as home-grown applications, AWS services, and third- party software to consumer applications across your organization. EventBridge provides a simple and consistent way to ingest, filter, transform, and deliver events so you can build new applications quickly.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EventBridge
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eventbridge
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EventBridge Developer Resources
          url: https://aws.amazon.com/eventbridge/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eventbridge maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/eventbridge-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-eventbridge-controller.v1.1.1
        version: 1.1.1
      - name: ack-eventbridge-controller.v1.1.0
        version: 1.1.0
      - name: ack-eventbridge-controller.v1.0.28
        version: 1.0.28
      - name: ack-eventbridge-controller.v1.0.27
        version: 1.0.27
      - name: ack-eventbridge-controller.v1.0.26
        version: 1.0.26
      - name: ack-eventbridge-controller.v1.0.25
        version: 1.0.25
      - name: ack-eventbridge-controller.v1.0.24
        version: 1.0.24
      - name: ack-eventbridge-controller.v1.0.23
        version: 1.0.23
      - name: ack-eventbridge-controller.v1.0.22
        version: 1.0.22
      - name: ack-eventbridge-controller.v1.0.21
        version: 1.0.21
      - name: ack-eventbridge-controller.v1.0.20
        version: 1.0.20
      - name: ack-eventbridge-controller.v1.0.18
        version: 1.0.18
      - name: ack-eventbridge-controller.v1.0.17
        version: 1.0.17
      - name: ack-eventbridge-controller.v1.0.16
        version: 1.0.16
      - name: ack-eventbridge-controller.v1.0.15
        version: 1.0.15
      - name: ack-eventbridge-controller.v1.0.14
        version: 1.0.14
      - name: ack-eventbridge-controller.v1.0.13
        version: 1.0.13
      - name: ack-eventbridge-controller.v1.0.12
        version: 1.0.12
      - name: ack-eventbridge-controller.v1.0.11
        version: 1.0.11
      - name: ack-eventbridge-controller.v1.0.6
        version: 1.0.6
      - name: ack-eventbridge-controller.v1.0.5
        version: 1.0.5
      - name: ack-eventbridge-controller.v1.0.4
        version: 1.0.4
      - name: ack-eventbridge-controller.v1.0.3
        version: 1.0.3
      - name: ack-eventbridge-controller.v1.0.1
        version: 1.0.1
      - name: ack-eventbridge-controller.v1.0.0
        version: 1.0.0
      - name: ack-eventbridge-controller.v0.0.7
        version: 0.0.7
      - name: ack-eventbridge-controller.v0.0.6
        version: 0.0.6
      - name: ack-eventbridge-controller.v0.0.5
        version: 0.0.5
      - name: ack-eventbridge-controller.v0.0.4
        version: 0.0.4
      - name: ack-eventbridge-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-eventbridge-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Oracle Corporation
      provider-url: https://github.com/oracle/coherence-operator
    name: coherence-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: coherence-operator.v3.5.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "coherence.oracle.com/v1",
                "kind": "Coherence",
                "metadata": {
                  "name": "coherence-sample"
                },
                "spec": {
                  "replicas": 3
                }
              },
              {
                "apiVersion": "coherence.oracle.com/v1",
                "kind": "CoherenceJob",
                "metadata": {
                  "name": "coherence-job-sample"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "false"
          containerImage: container-registry.oracle.com/middleware/coherence-operator:3.5.4
          createdAt: "2025-08-28T16:15:09Z"
          description: The Oracle Coherence Kubernetes Operator enables easy management
            of Coherence clusters in a Kubernetes environment.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/arch.amd64: supported
          operatorframework.io/arch.arm64: supported
          operatorframework.io/os.linux: supported
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/oracle/coherence-operator
          support: Oracle Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Coherence is the top level schema for the Coherence API and
              custom resource definition (CRD).
            displayName: Coherence
            kind: Coherence
            name: coherence.coherence.oracle.com
            version: v1
          - description: |-
              CoherenceJob is the top level schema for the CoherenceJob API and custom resource definition (CRD)
              for configuring Coherence Job workloads.
            displayName: Coherence Job
            kind: CoherenceJob
            name: coherencejob.coherence.oracle.com
            version: v1
        description: |
          The Oracle Coherence Kubernetes Operator enables easy management of Coherence clusters in a Kubernetes environment.

          ### Installation Guide & Configuration

          For the complete installation guide and all configuration options please refer to
          [Coherence Operator Documentation](https://oracle.github.io/coherence-operator/docs/latest/#/about/01_overview).
        displayName: Coherence Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Coherence
        - Oracle
        - In Memory Data
        - Caching
        links:
        - name: Coherence Kubernetes Operator
          url: https://oracle.github.io/coherence-operator/docs/latest
        - name: Oracle Coherence
          url: https://coherence.community
        maintainers:
        - email: jonathan.knight@oracle.com
          name: Jonathan Knight
        maturity: mature
        minKubeVersion: 1.26.0
        provider:
          name: Oracle Corporation
          url: https://github.com/oracle/coherence-operator
        relatedImages:
        - container-registry.oracle.com/middleware/coherence-operator:3.5.4
        version: 3.5.4
      entries:
      - name: coherence-operator.v3.5.4
        version: 3.5.4
      - name: coherence-operator.v3.5.2
        version: 3.5.2
      - name: coherence-operator.v3.5.0
        version: 3.5.0
      - name: coherence-operator.v3.4.3
        version: 3.4.3
      name: stable
    defaultChannel: stable
    packageName: coherence-operator
    provider:
      name: Oracle Corporation
      url: https://github.com/oracle/coherence-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nxrm-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nxrm-operator-certified.v3.58.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusRepo",
                "metadata": {
                  "name": "example-nexusrepo"
                },
                "spec": {
                  "config": {
                    "data": null,
                    "enabled": false,
                    "mountPath": "/sonatype-nexus-conf"
                  },
                  "deployment": {
                    "additionalContainers": null,
                    "additionalVolumeMounts": null,
                    "additionalVolumes": null,
                    "annotations": {},
                    "initContainers": null,
                    "postStart": {
                      "command": null
                    },
                    "preStart": {
                      "command": null
                    },
                    "terminationGracePeriodSeconds": 120
                  },
                  "deploymentStrategy": {
                    "type": "Recreate"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "path": "/",
                    "rules": null,
                    "tls": {
                      "enabled": true,
                      "secretName": "nexus-tls"
                    }
                  },
                  "nexus": {
                    "dockerPort": 5003,
                    "env": [
                      {
                        "name": "INSTALL4J_ADD_VM_PARAMS",
                        "value": "-Xms2703M -Xmx2703M -XX:MaxDirectMemorySize=2703M -XX:+UnlockExperimentalVMOptions -XX:+UseCGroupMemoryLimitForHeap"
                      }
                    ],
                    "hostAliases": [],
                    "imageName": "registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:06fe31f40ff3fd7556fc7b8f6b2efe80ada1d9079ec9aaad60d1fdc0ee0677f0",
                    "imagePullPolicy": "IfNotPresent",
                    "imagePullSecret": "",
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "path": "/",
                      "periodSeconds": 30
                    },
                    "nexusPort": 8081,
                    "podAnnotations": {},
                    "readinessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "path": "/",
                      "periodSeconds": 30
                    },
                    "resources": {},
                    "securityContext": {
                    },
                    "service": {
                      "type": "NodePort"
                    }
                  },
                  "nexusProxyRoute": {
                    "annotations": null,
                    "enabled": false,
                    "labels": null
                  },
                  "persistence": {
                    "accessMode": "ReadWriteOnce",
                    "enabled": true,
                    "storageSize": "8Gi"
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": null,
                    "enabled": false,
                    "labels": null,
                    "name": "docker",
                    "portName": "docker"
                  },
                  "secret": {
                    "data": null,
                    "enabled": false,
                    "mountPath": "/etc/secret-volume",
                    "readOnly": true
                  },
                  "service": {
                    "annotations": {},
                    "enabled": false,
                    "labels": {},
                    "ports": [
                      {
                        "name": "nexus-service",
                        "port": 80,
                        "targetPort": 80
                      }
                    ]
                  },
                  "statefulset": {
                    "enabled": false
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nxrm-operator-certified@sha256:4945ad993a10e27b358b2d9869d05f23ca761162ec088e5a833939022bc23192
          createdAt: "2023-07-14"
          description: |-
            Nexus Repository is the central source of control to efficiently manage all binaries
            and build artifacts across your DevOps pipeline.
          olm.skipRange: <3.58.1-2
          repository: https://github.com/sonatype/operator-nxrm3
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus Repository
            displayName: NexusRepo
            kind: NexusRepo
            name: nexusrepos.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Repository is the central source of control to efficiently manage all binaries
          and build artifacts across your DevOps pipeline.
          The flow of open source and third-party components into and through an organization
          creates a complex software supply chain.
          Nexus Repository delivers speed, efficiency, and quality to the governance
          and management of all dependencies, libraries, and applications for your DevOps teams.

          ## Core Capabilities

          * **Dependency Management**:
            Improves reliability with repeatable, fast access to secure dependencies
          * **Developer Productivity**:
            Streamline developer workflows by enabling the sharing of components and applications across teams
          * **Supply Chain Performance**:
            Improve speed-to-market and reduced build times with release advanced staging and component tagging
          * **CI/CD Integrations**:
            Increase DevOps scalability with integrations to the most popular build and deployment tools

          Version control systems and package registries do not scale when managing proprietary,
          open source, and third-party components.
          Organizations need a central binary and build artifact repository to manage dependencies
          across the entire software supply chain.

          ## Limitations

          High Availability Clustering (HA-C) is not supported for Nexus Repository Pro for OpenShift.

          This operator will be released on a quarterly basis.

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus Repository Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage

          Once the server instance is created by the operator and running,
          you'll want to expose the service as you see fit:
          1. Create a Route to that service for nexus.port (8081).

          By default, the Nexus Repository starts up in OSS mode until a license is installed.

          The Nexus Repository can be further configured via the NexusRepo custom resource definition:

          | Parameter                                   | Description                         | Default                                 |
          | ------------------------------------------  | ----------------------------------  | ----------------------------------------|
          | `statefulset.enabled`                       | Use statefulset instead of deployment | `false` |
          | `deploymentStrategy.type`                   | Deployment Strategy     |  `Recreate` |
          | `nexus.env`                                 | Nexus environment variables         | `See example.` |
          | `nexus.resources`                           | Nexus resource requests and limits  | `{}`                                    |
          | `nexus.dockerPort`                          | Port to access docker               | `5003`                                  |
          | `nexus.nexusPort`                           | Internal port for Nexus service     | `8081`                                  |
          | `nexus.service.type`                        | Service for Nexus                   |`NodePort`                                |
          | `nexus.service.clusterIp`                   | Specific cluster IP when service type is cluster IP. Use None for headless service |`nil`   |
          | `nexus.securityContext`                     | Security Context |
          | `nexus.labels`                              | Service labels                      | `{}`                                    |
          | `nexus.podAnnotations`                      | Pod Annotations                     | `{}`
          | `nexus.livenessProbe.initialDelaySeconds`   | LivenessProbe initial delay         | 30                                      |
          | `nexus.livenessProbe.periodSeconds`         | Seconds between polls               | 30                                      |
          | `nexus.livenessProbe.failureThreshold`      | Number of attempts before failure   | 6                                       |
          | `nexus.livenessProbe.timeoutSeconds`        | Time in seconds after liveness probe times out    | `nil`                     |
          | `nexus.livenessProbe.path`                  | Path for LivenessProbe              | /                                       |
          | `nexus.readinessProbe.initialDelaySeconds`  | ReadinessProbe initial delay        | 30                                      |
          | `nexus.readinessProbe.periodSeconds`        | Seconds between polls               | 30                                      |
          | `nexus.readinessProbe.failureThreshold`     | Number of attempts before failure   | 6                                       |
          | `nexus.readinessProbe.timeoutSeconds`       | Time in seconds after readiness probe times out    | `nil`                    |
          | `nexus.readinessProbe.path`                 | Path for ReadinessProbe             | /                                       |
          | `nexus.hostAliases`                         | Aliases for IPs in /etc/hosts       | []                                      |
          | `ingress.enabled`                           | Create an ingress for Nexus         | `true`                                  |
          | `ingress.annotations`                       | Annotations to enhance ingress configuration  | `{}`                          |
          | `ingress.tls.enabled`                       | Enable TLS                          | `true`                                 |
          | `ingress.tls.secretName`                    | Name of the secret storing TLS cert, `false` to use the Ingress' default certificate | `nexus-tls`                             |
          | `ingress.path`                              | Path for ingress rules. GCP users should set to `/*` | `/`                    |
          | `tolerations`                               | tolerations list                    | `[]`                                    |
          | `config.enabled`                            | Enable configmap                    | `false`                                 |
          | `config.mountPath`                          | Path to mount the config            | `/sonatype-nexus-conf`                  |
          | `config.data`                               | Configmap data                      | `nil`                                   |
          | `deployment.terminationGracePeriodSeconds`  | Time to allow for clean shutdown    | 120                                     |
          | `deployment.annotations`                    | Annotations to enhance deployment configuration  | `{}`                       |
          | `deployment.initContainers`                 | Init containers to run before main containers  | `nil`                        |
          | `deployment.postStart.command`              | Command to run after starting the nexus container  | `nil`                    |
          | `deployment.preStart.command`               | Command to run before starting the nexus container  | `nil`                   |
          | `deployment.additionalContainers`           | Add additional Container         | `nil`                                      |
          | `deployment.additionalVolumes`              | Add additional Volumes           | `nil`                                      |
          | `deployment.additionalVolumeMounts`         | Add additional Volume mounts     | `nil`                                      |
          | `secret.enabled`                            | Enable secret                    | `false`                                    |
          | `secret.mountPath`                          | Path to mount the secret         | `/etc/secret-volume`                       |
          | `secret.readOnly`                           | Secret readonly state            | `true`                                     |
          | `secret.data`                               | Secret data                      | `nil`                                      |
          | `service.enabled`                           | Enable additional service        | `nil`                                      |
          | `service.name`                              | Service name                     | `nil`                                      |
          | `service.portName`                          | Service port name                | `nil`                                      |
          | `service.labels`                            | Service labels                   | `nil`                                      |
          | `service.annotations`                       | Service annotations              | `nil`                                      |
          | `service.loadBalancerSourceRanges`          | Service LoadBalancer source IP whitelist | `nil`                              |
          | `service.targetPort`                        | Service port                     | `nil`                                      |
          | `service.port`                              | Port for exposing service        | `nil`                                      |
          | `route.enabled`         | Set to true to create route for additional service | `false` |
          | `route.name`            | Name of route                                      | `docker` |
          | `route.portName`        | Target port name of service                        | `docker` |
          | `route.labels`          | Labels to be added to route                        | `{}` |
          | `route.annotations`     | Annotations to be added to route                   | `{}` |
          | `route.path`            | Host name of Route e.g jenkins.example.com         | nil |
        displayName: Nexus Repository Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - repository
        - sonatype
        links:
        - name: Nexus Repository Manager
          url: https://www.sonatype.com/product-nexus-repository
        maintainers:
        - email: support@sonatype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:06fe31f40ff3fd7556fc7b8f6b2efe80ada1d9079ec9aaad60d1fdc0ee0677f0
        - registry.connect.redhat.com/sonatype/nxrm-operator-certified@sha256:4945ad993a10e27b358b2d9869d05f23ca761162ec088e5a833939022bc23192
        version: 3.58.1-2
      entries:
      - name: nxrm-operator-certified.v3.58.1-2
        version: 3.58.1-2
      name: stable
    defaultChannel: stable
    packageName: nxrm-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: AMD
      provider-url: ""
    name: amd-t2-january
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amd-t2-january.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovfect2.amd.com/v1",
                "kind": "SriovT2Card",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "amd-t2-january",
                    "app.kubernetes.io/instance": "sriovt2card-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sriovt2card",
                    "app.kubernetes.io/part-of": "amd-t2-january"
                  },
                  "name": "sriovt2card",
                  "namespace": "amd-t2"
                },
                "spec": {
                  "acceleratorSelector": {
                    "pciAddress": "0000:01:00.0"
                  },
                  "nodeSelector": {
                    "kubernetes.io/hostname": "7c-c2-55-85-ba-fe.amd.com"
                  },
                  "physicalFunction": {
                    "pfDriver": "qdma-pf",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-07-14T09:46:54Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovT2Card is the Schema for the sriovt2cards API
            displayName: Sriov T2 Card
            kind: SriovT2Card
            name: sriovt2cards.sriovfect2.amd.com
            version: v1
        description: The AMD-T2 Operator for accelerator enables the seamless integration
          and management of AMD T2 accelerator within Openshift/Kubernetes envirnoment.
        displayName: AMD-T2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - SRIOV
        - Xilinx
        - FPGA
        - Telco
        - T2
        - MultiArch
        - AMD
        - TelcoAccelerator
        links:
        - name: Amd T2 January
          url: https://amd-t2-january.domain
        maintainers:
        - email: dhirajs.lakhane@amd.com
          name: DhirajLakhane
        maturity: alpha
        provider:
          name: AMD
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:928e64203edad8f1bba23593c7be04f0f8410c6e4feb98d9e9c2d00a8ff59048
        - quay.io/amdt2operator/amd-t2-january@sha256:318f56e2670582c217b8973f69f0675e7fff7dc19d07875294f2ff2fc5403902
        version: 1.0.1
      entries:
      - name: amd-t2-january.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: amd-t2-january
    provider:
      name: AMD
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: placement-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: placement-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "placement.openstack.org/v1beta1",
                "kind": "PlacementAPI",
                "metadata": {
                  "name": "placement"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "placement",
                  "databaseInstance": "openstack",
                  "preserveJobs": false,
                  "replicas": 1,
                  "secret": "placement-secret",
                  "serviceUser": "placement"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:43Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PlacementAPI is the Schema for the placementapis API
            displayName: Placement API
            kind: PlacementAPI
            name: placementapis.placement.openstack.org
            version: v1beta1
        description: Placement Operator
        displayName: Placement Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Placement
        links:
        - name: Placement Operator
          url: https://github.com/openstack-k8s-operators/placement-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/placement-rhel9-operator@sha256:d3be48618b1f0f717aea162ae41876137fee8b35329d2e3e66e9d4eaccf62376
        - registry.redhat.io/rhoso/openstack-placement-api-rhel9@sha256:3308f2eccfbe7416a85443696273417bfb816df36b810ebbefa108dda12a5b19
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: placement-operator.v1.0.6
        version: 1.0.6
      - name: placement-operator.v1.0.5
        version: 1.0.5
      - name: placement-operator.v1.0.4
        version: 1.0.4
      - name: placement-operator.v1.0.3
        version: 1.0.3
      - name: placement-operator.v1.0.2
        version: 1.0.2
      - name: placement-operator.v1.0.1
        version: 1.0.1
      - name: placement-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: placement-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/cluster-nfd-operator
    name: nfd
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfd.4.16.0-202507211806
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nfd.openshift.io/v1",
                "kind": "NodeFeatureDiscovery",
                "metadata": {
                  "name": "nfd-instance",
                  "namespace": "openshift-nfd"
                },
                "spec": {
                  "customConfig": {
                    "configData": "#    - name: \"more.kernel.features\"\n#      matchOn:\n#      - loadedKMod: [\"example_kmod3\"]\n#    - name: \"more.features.by.nodename\"\n#      value: customValue\n#      matchOn:\n#      - nodename: [\"special-.*-node-.*\"]\n"
                  },
                  "operand": {
                    "image": "registry.redhat.io/openshift4/ose-node-feature-discovery-rhel9@sha256:bb65ec6a6f88688630305e081cb4bcb88aa538b5f6a3385ee28c61522dcd1662",
                    "imagePullPolicy": "IfNotPresent",
                    "servicePort": 12000
                  },
                  "workerConfig": {
                    "configData": "core:\n#  labelWhiteList:\n#  noPublish: false\n  sleepInterval: 60s\n#  sources: [all]\n#  klog:\n#    addDirHeader: false\n#    alsologtostderr: false\n#    logBacktraceAt:\n#    logtostderr: true\n#    skipHeaders: false\n#    stderrthreshold: 2\n#    v: 0\n#    vmodule:\n##   NOTE: the following options are not dynamically run-time \n##          configurable and require a nfd-worker restart to take effect\n##          after being changed\n#    logDir:\n#    logFile:\n#    logFileMaxSize: 1800\n#    skipLogHeaders: false\nsources:\n#  cpu:\n#    cpuid:\n##     NOTE: whitelist has priority over blacklist\n#      attributeBlacklist:\n#        - \"BMI1\"\n#        - \"BMI2\"\n#        - \"CLMUL\"\n#        - \"CMOV\"\n#        - \"CX16\"\n#        - \"ERMS\"\n#        - \"F16C\"\n#        - \"HTT\"\n#        - \"LZCNT\"\n#        - \"MMX\"\n#        - \"MMXEXT\"\n#        - \"NX\"\n#        - \"POPCNT\"\n#        - \"RDRAND\"\n#        - \"RDSEED\"\n#        - \"RDTSCP\"\n#        - \"SGX\"\n#        - \"SSE\"\n#        - \"SSE2\"\n#        - \"SSE3\"\n#        - \"SSE4.1\"\n#        - \"SSE4.2\"\n#        - \"SSSE3\"\n#      attributeWhitelist:\n#  kernel:\n#    kconfigFile: \"/path/to/kconfig\"\n#    configOpts:\n#      - \"NO_HZ\"\n#      - \"X86\"\n#      - \"DMI\"\n  pci:\n    deviceClassWhitelist:\n      - \"0200\"\n      - \"03\"\n      - \"12\"\n    deviceLabelFields:\n#      - \"class\"\n      - \"vendor\"\n#      - \"device\"\n#      - \"subsystem_vendor\"\n#      - \"subsystem_device\"\n#  usb:\n#    deviceClassWhitelist:\n#      - \"0e\"\n#      - \"ef\"\n#      - \"fe\"\n#      - \"ff\"\n#    deviceLabelFields:\n#      - \"class\"\n#      - \"vendor\"\n#      - \"device\"\n#  custom:\n#    - name: \"my.kernel.feature\"\n#      matchOn:\n#        - loadedKMod: [\"example_kmod1\", \"example_kmod2\"]\n#    - name: \"my.pci.feature\"\n#      matchOn:\n#        - pciId:\n#            class: [\"0200\"]\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#        - pciId :\n#            vendor: [\"8086\"]\n#            device: [\"1000\", \"1100\"]\n#    - name: \"my.usb.feature\"\n#      matchOn:\n#        - usbId:\n#          class: [\"ff\"]\n#          vendor: [\"03e7\"]\n#          device: [\"2485\"]\n#        - usbId:\n#          class: [\"fe\"]\n#          vendor: [\"1a6e\"]\n#          device: [\"089a\"]\n#    - name: \"my.combined.feature\"\n#      matchOn:\n#        - pciId:\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#          loadedKMod : [\"vendor_kmod1\", \"vendor_kmod2\"]\n"
                  }
                }
              },
              {
                "apiVersion": "nfd.openshift.io/v1alpha1",
                "kind": "NodeFeatureRule",
                "metadata": {
                  "name": "my-sample-rule-object"
                },
                "spec": {
                  "rules": [
                    {
                      "labels": {
                        "my-sample-feature": "true"
                      },
                      "matchFeatures": [
                        {
                          "feature": "kernel.loadedmodule",
                          "matchExpressions": {
                            "dummy": {
                              "op": "Exists"
                            }
                          }
                        },
                        {
                          "feature": "kernel.config",
                          "matchExpressions": {
                            "X86": {
                              "op": "In",
                              "value": [
                                "y"
                              ]
                            }
                          }
                        }
                      ],
                      "name": "my sample rule"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery,OpenShift Optional
          containerImage: registry.redhat.io/openshift4/ose-cluster-nfd-rhel9-operator@sha256:c1b6a5b8a59c08458d57c0fc2bb2bc9b056f006d15edb4d3161a3d1c0bc97ae6
          description: |-
            The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

            NFD consists  of the following software components:

            The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

            ##NFD-Master
            NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

            ##NFD-Worker
            NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

            ##NFD-Topology-Updater
            NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.0-202507211806'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-nfd
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/internal-objects: '["noderesourcetopologies.topology.node.k8s.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/cluster-nfd-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              The NodeFeatureDiscovery instance is the CustomResource being watched by the NFD-Operator, and holds all the needed information to setup the behaviour of the master and worker pods
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
          - description: |
              NodeFeatureRule resource specifies a configuration for feature-based customization of node objects, such as node labeling.
            kind: NodeFeatureRule
            name: nodefeaturerules.nfd.openshift.io
            version: v1alpha1
          - kind: NodeFeature
            name: nodefeatures.nfd.openshift.io
            version: v1alpha1
        description: |-
          The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

          NFD consists  of the following software components:

          The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

          ##NFD-Master
          NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

          ##NFD-Worker
          NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

          ##NFD-Topology-Updater
          NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
        displayName: Node Feature Discovery Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - feature-discovery
        - feature-detection
        - node-labels
        links:
        - name: Node Feature Discovery Operator
          url: https://docs.openshift.com/container-platform/4.16/hardware_enablement/psap-node-feature-discovery-operator.html
        - name: Node Feature Discovery Documentation
          url: https://kubernetes-sigs.github.io/node-feature-discovery/stable/get-started/index.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/cluster-nfd-operator
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:bcca89fb000b0fe8722ad7807b62a43ef3ae8fa76db1d790c285a975bfecb1aa
        - registry.redhat.io/openshift4/ose-cluster-nfd-rhel9-operator@sha256:c1b6a5b8a59c08458d57c0fc2bb2bc9b056f006d15edb4d3161a3d1c0bc97ae6
        - registry.redhat.io/openshift4/ose-node-feature-discovery-rhel9@sha256:bb65ec6a6f88688630305e081cb4bcb88aa538b5f6a3385ee28c61522dcd1662
        version: 4.16.0-202507211806
      entries:
      - name: nfd.4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: nfd.4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: nfd.4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: nfd.4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: nfd.4.16.0-202504271139
        version: 4.16.0-202504271139
      - name: nfd.4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: nfd.4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: nfd.4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: nfd.4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: nfd.4.16.0-202501271512
        version: 4.16.0-202501271512
      - name: nfd.4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: nfd.4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: nfd.4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: nfd.4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: nfd.4.16.0-202410251436
        version: 4.16.0-202410251436
      - name: nfd.4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: nfd.4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: nfd.4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: nfd.4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: nfd.4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: nfd.4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: nfd.4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: nfd.4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: nfd.4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: nfd.4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: nfd.4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: nfd.4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: nfd.4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: nfd.4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: nfd
    provider:
      name: Red Hat
      url: https://github.com/openshift/cluster-nfd-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhpam-kogito-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhpam-kogito-operator.7.13.5-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "ruleunit-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogitoinfra-sample"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "kafka.strimzi.io/v1beta2",
                    "kind": "Kafka",
                    "name": "kogito-kafka"
                  }
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator@sha256:19497d4aff0ea3ea114dccb0d0e265458bb14ba6d325353eb82b554547dbd2e4
          description: RHPAM Kogito Operator for deployment and management of Kogito
            services.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Process Automation
            Manager", "Red Hat Decision Manager"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/rhpam-kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.rhpam.kiegroup.org
            version: v1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.rhpam.kiegroup.org
            version: v1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.rhpam.kiegroup.org
            version: v1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.rhpam.kiegroup.org
            version: v1
        description: |-
          RHPAM Kogito Operator is designed for deployment and management of Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Builds a RHPAM Kogito service from a remote git repository and creates a custom image for the given runtime: Red Hat build of Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Deploy RHPAM Kogito service from user provided image.
          * Setup Grafana Dashboard when Grafana Operator is available in cluster.
          * Create Prometheus monitor object when Prometheus operator is available in cluster.

          **IMPORTANT!** RHPAM Kogito Operator may need Grafana and Prometheus Operator to configure monitoring in cases if you do not provide these services beforehand. If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the RHPAM Kogito Operator.
        displayName: RHPAM Kogito Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-process-automation-manager
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_process_automation_manager/7.13/html/getting_started_with_red_hat_build_of_kogito_in_red_hat_process_automation_manager/assembly-deploying-kogito-microservices-on-openshift
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator@sha256:19497d4aff0ea3ea114dccb0d0e265458bb14ba6d325353eb82b554547dbd2e4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9492c67eaa7ab37df6f5fa623c4aac955b7d7ffaab9ae2e57dea745e82907e40
        version: 7.13.5-4
      entries:
      - name: rhpam-kogito-operator.7.13.5-4
        version: 7.13.5-4
      - name: rhpam-kogito-operator.7.13.5-3
        version: 7.13.5-3
      - name: rhpam-kogito-operator.7.13.5-2
        version: 7.13.5-2
      - name: rhpam-kogito-operator.7.13.5-1
        version: 7.13.5-1
      - name: rhpam-kogito-operator.7.13.4-3
        version: 7.13.4-3
      - name: rhpam-kogito-operator.7.13.4-2
        version: 7.13.4-2
      - name: rhpam-kogito-operator.7.13.4-1
        version: 7.13.4-1
      - name: rhpam-kogito-operator.7.13.3-1
        version: 7.13.3-1
      - name: rhpam-kogito-operator.7.13.2-3
        version: 7.13.2-3
      - name: rhpam-kogito-operator.7.13.2-2
        version: 7.13.2-2
      - name: rhpam-kogito-operator.7.13.2-1
        version: 7.13.2-1
      - name: rhpam-kogito-operator.7.13.1-2
        version: 7.13.1-2
      - name: rhpam-kogito-operator.7.13.1-1
        version: 7.13.1-1
      - name: rhpam-kogito-operator.7.13.0-1
        version: 7.13.0-1
      - name: rhpam-kogito-operator.v7.12.0-4
        version: 7.12.0-4
      - name: rhpam-kogito-operator.v7.12.0-3
        version: 7.12.0-3
      - name: rhpam-kogito-operator.v7.12.0-2
        version: 7.12.0-2
      name: 7.x
    defaultChannel: 7.x
    packageName: rhpam-kogito-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://skupper.io
    name: skupper-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-operator.v1.4.8-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus@sha256:65602221cdff7f718d4149f4dead167d05dbbe0cea1eaa8bf4b98be588135583
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:c665677c639699d6a5ef24aed5a1a3d9c0a5ee25f077a6db42a9fc6247d93132
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:eedde9f2019d4e919612bcc18f427c0d88b4846e773751af231ba0e560d48869
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:1fd908e80ef21fe7d7051c400cb099f2f3ca51c8cdd698705ef9784d34688911
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:4c9cc9df83baa1ef37cfb429babe5b86a962b8bb096209b3fb74291c67c94000
        version: 1.4.8-rh-1
      entries:
      - name: skupper-operator.v1.4.8-rh-1
        version: 1.4.8-rh-1
      - name: skupper-operator.v1.4.7-rh-3
        version: 1.4.7-rh-3
      - name: skupper-operator.v1.4.7-rh-2
        version: 1.4.7-rh-2
      - name: skupper-operator.v1.4.7-rh-1
        version: 1.4.7-rh-1
      - name: skupper-operator.v1.4.5-rh-4
        version: 1.4.5-rh-4
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1
    - currentCSV: skupper-operator.v1.4.8-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:eedde9f2019d4e919612bcc18f427c0d88b4846e773751af231ba0e560d48869
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:1fd908e80ef21fe7d7051c400cb099f2f3ca51c8cdd698705ef9784d34688911
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:4c9cc9df83baa1ef37cfb429babe5b86a962b8bb096209b3fb74291c67c94000
        - registry.redhat.io/openshift4/ose-prometheus@sha256:65602221cdff7f718d4149f4dead167d05dbbe0cea1eaa8bf4b98be588135583
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:c665677c639699d6a5ef24aed5a1a3d9c0a5ee25f077a6db42a9fc6247d93132
        version: 1.4.8-rh-1
      entries:
      - name: skupper-operator.v1.4.8-rh-1
        version: 1.4.8-rh-1
      - name: skupper-operator.v1.4.7-rh-3
        version: 1.4.7-rh-3
      - name: skupper-operator.v1.4.7-rh-2
        version: 1.4.7-rh-2
      - name: skupper-operator.v1.4.7-rh-1
        version: 1.4.7-rh-1
      - name: skupper-operator.v1.4.5-rh-4
        version: 1.4.5-rh-4
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.4
    - currentCSV: skupper-operator.v1.5.5-rh-6
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:9e9e1f2288f0218e3ba2abf4bebac9731d7cf9f9c567e0f6de6767b0b66cbee4
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:a5d7c1b2b9db672548126c0873d062a55667d93291fd3e32809d3c9c900c548e
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:9e9e1f2288f0218e3ba2abf4bebac9731d7cf9f9c567e0f6de6767b0b66cbee4
        - registry.redhat.io/openshift4/ose-prometheus@sha256:53afaf53975807765fc3777314edd218c789c17680c1f57a8e33a8f9a76f38bf
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:3e78ec8cdb729a704d5a26e54697f285ba7c0b9afee6160b7ecc703876120dea
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:be9dbcd2a3a1e1e536921390dcfd283fc45dc4423d2705376ef37a9fb53f7cf1
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:1de071dabf620265252f3922c632f17f676ba716ce4c3df931a0ccb3871dbb6c
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:5f8621d0013febcecbed4d7c82db11a4dded99ad8ce6b6d83858b9bdbe79ee5a
        version: 1.5.5-rh-6
      entries:
      - name: skupper-operator.v1.5.5-rh-6
        version: 1.5.5-rh-6
      - name: skupper-operator.v1.5.5-rh-5
        version: 1.5.5-rh-5
      - name: skupper-operator.v1.5.5-rh-4
        version: 1.5.5-rh-4
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.5
    - currentCSV: skupper-operator.v1.8.7-rh-2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:b0d43628b779a101a24e4b271318a34c21f3b71bcba2bf20804f941a93d27455
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.8.7-rh-2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:56b8fde75a6e7df235c560a0f4b9114398da5ba8b20f75f1ea7338f48ce19506
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:d7d252cff08ba1589afd8707ed349f87d3b6214906a4fb49932bda0a4d19fee4
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:5ea36e96f878577104ab7601e57e7a2a233f11dbdc533454f95e2e14a066d21f
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:3adeef827eca0382cd410f7c3e99a0fee9b50791cb0e0c6fc130b72a2550684f
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:b0d43628b779a101a24e4b271318a34c21f3b71bcba2bf20804f941a93d27455
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:f798745e799515065af099a816057b9c98a741e951d30a0f45ed47ca50e6724e
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:61542c5f0989f51ac5d6c96f5d0e29aca75614b79e7626299c77c5a74636b564
        version: 1.8.7-rh-2
      entries:
      - name: skupper-operator.v1.8.7-rh-2
        version: 1.8.7-rh-2
      - name: skupper-operator.v1.8.7-rh-1
        version: 1.8.7-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.8
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.9
    - currentCSV: skupper-operator.v2.1.1-rh-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
          createdAt: "2025-08-19T20:43:19Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:36872037066fae806cbf88cab673229f25e592295972aa18227827233362eec5
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:c01d0eab0841b1a5fddb58a0ae110beb464b91390f02662590bebc7b2a7213b6
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e8df53454cf12a14338790af6ed209798010990ca3320096e8b44acf901c4123
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:1a290a5ac8b9d1b082e740054675fe432aa424c2f55dd662caf67c740044f229
        version: 2.1.1-rh-3
      entries:
      - name: skupper-operator.v2.1.1-rh-3
        version: 2.1.1-rh-3
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2
    - currentCSV: skupper-operator.v2.0.1-rh-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:b5c73dc59a9184cd120c2ee35256cc097c790b8ef488af0c732d5efe3192b768
          createdAt: "2025-05-29T13:35:40Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.0.1-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:92f44cff160874efe57a985609e599b13c49c0c4cef846dccc80833af2ae74f3
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:e239c859a1f49816ae3068455e703b8a12919f11158b0f6ee0e6bfb31e8a8082
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:f2a41674ed196c089748d6386c18c14580e502de33235ef23c37eddeb6350f41
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:95062ff1256135cfa17907e82b09b2a211371ea0e76ccc711b7c737e040d87e9
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:498b5bb648e5bbaad07663941600173abac7d60a10744305f619509ddf7563e0
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:bd3f72544841c91c2d573a65ac3c988756f3523e17e0988027793b0f5f3ec3d1
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:b5c73dc59a9184cd120c2ee35256cc097c790b8ef488af0c732d5efe3192b768
        version: 2.0.1-rh-2
      entries:
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2.0
    - currentCSV: skupper-operator.v2.1.1-rh-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
          createdAt: "2025-08-19T20:43:19Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:c01d0eab0841b1a5fddb58a0ae110beb464b91390f02662590bebc7b2a7213b6
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e8df53454cf12a14338790af6ed209798010990ca3320096e8b44acf901c4123
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:1a290a5ac8b9d1b082e740054675fe432aa424c2f55dd662caf67c740044f229
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:36872037066fae806cbf88cab673229f25e592295972aa18227827233362eec5
        version: 2.1.1-rh-3
      entries:
      - name: skupper-operator.v2.1.1-rh-3
        version: 2.1.1-rh-3
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2.1
    defaultChannel: stable-2
    packageName: skupper-operator
    provider:
      name: Red Hat
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://horreum.hyperfoil.io
    name: horreum-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: horreum-operator.v0.7.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hyperfoil.io/v1alpha1",
                "kind": "Horreum",
                "metadata": {
                  "name": "horreum"
                },
                "spec": {
                  "nodeHost": "127.0.0.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/hyperfoil/horreum-operator:0.7.9
          createdAt: "2023-06-13T10:42:39Z"
          description: Performance results repository
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/hyperfoil/horreum-operator
          support: Red Hat Application Runtimes Performance Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Horreum is the object configuring Horreum performance results
              repository
            displayName: Horreum
            kind: Horreum
            name: horreums.hyperfoil.io
            version: v1alpha1
        description: Performance results repository
        displayName: Horreum
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - performance
        - repository
        - database
        - results
        - hyperfoil
        - benchmark
        links:
        - name: Horreum Operator
          url: https://horreum-operator.domain
        maintainers:
        - email: rvansa@redhat.com
          name: Radim Vansa
        - email: spederse@redhat.com
          name: Ståle W. Pedersen
        - email: wreicher@redhat.com
          name: Will Reichert
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat, Inc.
          url: https://horreum.hyperfoil.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/hyperfoil/horreum-operator:0.7.9
        version: 0.7.9
      entries:
      - name: horreum-operator.v0.7.9
        version: 0.7.9
      - name: horreum-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: horreum-operator
    provider:
      name: Red Hat, Inc.
      url: https://horreum.hyperfoil.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/stolostron/multicluster-global-hub
    name: multicluster-global-hub-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-global-hub-operator.v1.4.0-dev.20250302
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2025-02-14T08:24:47Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: planning
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/postgresql-16:9.5-1732622748
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-inventory/inventory-api:latest
        - quay.io/prometheuscommunity/postgres-exporter:v0.15.0
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-relations/spicedb-operator:latest
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-relations/spicedb:latest
        - quay.io/stolostron/grafana:2.12.0-SNAPSHOT-2024-09-03-21-11-25
        - quay.io/stolostron/multicluster-global-hub-operator:v1.4.0
        - quay.io/stolostron/multicluster-global-hub-agent:v1.4.0
        - quay.io/stolostron/multicluster-global-hub-manager:v1.4.0
        version: 1.4.0-dev.20250302
      entries:
      - name: multicluster-global-hub-operator.v1.4.0-dev.20250302
        version: 1.4.0-dev.20250302
      - name: multicluster-global-hub-operator.v1.4.0-dev.20250225
        version: 1.4.0-dev.20250225
      - name: multicluster-global-hub-operator.v1.3.0-dev.20240911
        version: 1.3.0-dev.20240911
      - name: multicluster-global-hub-operator.v1.3.0-dev.20240724
        version: 1.3.0-dev.20240724
      - name: multicluster-global-hub-operator.v1.2.0-dev.20240516
        version: 1.2.0-dev.20240516
      - name: multicluster-global-hub-operator.v1.2.0-dev
        version: 1.2.0-dev
      - name: multicluster-global-hub-operator.v1.1.0-dev
        version: 1.1.0-dev
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: multicluster-global-hub-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-operator-globalhub-1-5@sha256:4db1251323bd64930461cb5025da41baffcf5c00c51ea4e49940b7130971b510
          createdAt: "2025-06-10T06:58:35Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-operator-globalhub-1-5@sha256:4db1251323bd64930461cb5025da41baffcf5c00c51ea4e49940b7130971b510
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-manager-globalhub-1-5@sha256:01253426d65f1297a1955a8393051e6a4b1df5e5249774fc5189a75f35e7d644
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-agent-globalhub-1-5@sha256:6f7520eac83abec7780f831d8473544d0b8db43fc6e77d59e59317b09eb1f6c1
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/glo-grafana-globalhub-1-5@sha256:73fdf00538e63595b7cc92d484e087ff815579a95d7259c579c6e7dc1beacff6
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/postgres-exporter-globalhub-1-5@sha256:1108572a8bdafde926254f1ff94c2b22e048706be660b07c793e9f9e038a0f4a
        version: 1.5.0
      entries:
      - name: multicluster-global-hub-operator.v1.5.0
        version: 1.5.0
      - name: multicluster-global-hub-operator.v1.4.0
        version: 1.4.0
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-1.5
    - currentCSV: multicluster-global-hub-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2024-03-05T02:13:54Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.32.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:latest
        version: 1.1.0
      entries:
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.10
    - currentCSV: multicluster-global-hub-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2024-07-03T06:52:02Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v1.2.0
        version: 1.2.1
      entries:
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.11
    - currentCSV: multicluster-global-hub-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator@sha256:636aea861da80fc36fb27a419a18c5b3fbb1bc072e8ae33c75e1cc47e2c05f75
          createdAt: "2024-10-10T05:42:48Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          kafka-zookeeper-0                                       1/1     Running   0          21m
          kafka-zookeeper-1                                       1/1     Running   0          21m
          kafka-zookeeper-2                                       1/1     Running   0          21m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgres-0                      2/2     Running   0          22m
          ```
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:636aea861da80fc36fb27a419a18c5b3fbb1bc072e8ae33c75e1cc47e2c05f75
        - quay.io/prometheuscommunity/postgres-exporter@sha256:386b12d19eab2a37d7cd8ca8b4c7491cc7a830d9581f49af6c98a393da9605e6
        - quay.io/stolostron/inventory-api@sha256:a15e62ee070dba5d74cf336b3b89b467593c5a33cfe3623d922f0af4fd7df392
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/release-globalhub-1-3/multicluster-global-hub-manager-globalhub-1-3@sha256:7656d01a4d7b39c61c3249695d68011d9e0df308627b9efd117686e901281ab1
        - quay.io/stolostron/multicluster-global-hub-manager@sha256:715cb5d0c6c407ec5c0303fc7e6d83fc77ce7e06775e169a475f7c075f1e7000
        - quay.io/stolostron/grafana@sha256:719b320a05d10158fadce5ae2fc6e4254a18957a6608b6f81e876cf46d3a6897
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        version: 1.3.0
      entries:
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.12
    - currentCSV: multicluster-global-hub-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator@sha256:085e9fc6cdc6a42df10a38856c8ccff25e5acffaba67ef95da72027e90e507fc
          createdAt: "2025-02-14T08:24:47Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:085e9fc6cdc6a42df10a38856c8ccff25e5acffaba67ef95da72027e90e507fc
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-inventory/inventory-api@sha256:6a55a1919ecd92d94e05c12be1b6450956f5fda2ce17af7027cac3806c4970a4
        - quay.io/prometheuscommunity/postgres-exporter@sha256:386b12d19eab2a37d7cd8ca8b4c7491cc7a830d9581f49af6c98a393da9605e6
        - quay.io/stolostron/multicluster-global-hub-manager@sha256:4cd602ff6f8dd951848dc8a95c856004295c9f4710f77d4dafcc0735c1492395
        - quay.io/stolostron/multicluster-global-hub-agent@sha256:8dfc33dfed33a85733ee5e02dcd4f05662c75263791e0961ba17cbb3d0c340c9
        - quay.io/stolostron/grafana@sha256:719b320a05d10158fadce5ae2fc6e4254a18957a6608b6f81e876cf46d3a6897
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        version: 1.4.0
      entries:
      - name: multicluster-global-hub-operator.v1.4.0
        version: 1.4.0
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.13
    - currentCSV: multicluster-global-hub-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "storage": {
                    "name": "storage-secret"
                  },
                  "transport": {
                    "name": "transport-secret"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.5.0
          createdAt: "2022-08-23T20:00:31Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha1
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.5 or later needs to be installed.\n- PostgreSQL needs to be installed
          and you must create database for the multicluster global hub. You must create
          a secret named postgresql-secret that contains the database access credential
          in `open-cluster-management` namespace. You can run the following command
          to create the secret: \n\n```\nkubectl create secret generic storage-secret
          -n open-cluster-management \\\n  --from-literal=database_uri=<postgresql-uri>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)
          to install postgres in `hoh-postgres` namespace and create the secret `storage-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n- Kafka needs to be installed,
          and you must create two sections, `spec` and `status`. You need to create
          a secret named kafka-secret that contains the kafka access information should
          be created in `open-cluster-management` namespace. You can run the following
          command to create the secret: \n\n```\nkubectl create secret generic transport-secret
          -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n  --from-literal=CA=<CA-for-kafka-server>\n```\n >_Note:_ There is a
          sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `transport-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n## How to Install\n\nInstall the
          Multicluster Global Hub Operator by following the instructions that are
          displayed after you select the Install button. The operator must be installed
          in the same namespace as Red Hat Advanced Cluster Management for Kubernetes
          (RHACM), by default, in `open-cluster-management` namespace.\n\nA pod will
          be created in `open-cluster-management` namespace\n\n```\n$ kubectl get
          pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          19s\n```\n\nThe operator is
          now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installing the operator,
          create an instance of the MulticlusterGlobalHub resource to install the
          Multicluster Global Hub.\n\nThe pods of Multicluster Global Hub are created
          in the `open-cluster-management` namespace:\n\n```\n$ kubectl get pods -n
          open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\nmulticluster-global-hub-rbac-77888b7dc4-z9vbp
          \                   1/1     Running   0          10s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.5.0
        version: 0.5.0
      entries:
      - name: multicluster-global-hub-operator.v0.5.0
        version: 0.5.0
      name: release-2.6
    - currentCSV: multicluster-global-hub-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha2",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "largeScale": {
                      "kafka": {
                        "name": "transport-secret"
                      },
                      "postgres": {
                        "name": "storage-secret"
                      }
                    },
                    "type": "largeScale"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.6.0
          createdAt: "2023-02-09T20:00:31Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha2
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.7 or later needs to be installed.\n- PostgreSQL needs to be installed
          and you must create database for the multicluster global hub. You must create
          a secret named postgresql-secret that contains the database access credential
          in `open-cluster-management` namespace. You can run the following command
          to create the secret: \n\n```\nkubectl create secret generic storage-secret
          -n open-cluster-management \\\n  --from-literal=database_uri=<postgresql-uri>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)
          to install postgres in `hoh-postgres` namespace and create the secret `storage-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n- Kafka needs to be installed,
          and you must create two sections, `spec` and `status`. You need to create
          a secret named kafka-secret that contains the kafka access information should
          be created in `open-cluster-management` namespace. You can run the following
          command to create the secret: \n\n```\nkubectl create secret generic transport-secret
          -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n  --from-literal=CA=<CA-for-kafka-server>\n```\n >_Note:_ There is a
          sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `transport-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n## How to Install\n\nInstall the
          Multicluster Global Hub Operator by following the instructions that are
          displayed after you select the Install button. The operator must be installed
          in the same namespace as Red Hat Advanced Cluster Management for Kubernetes
          (RHACM), by default, in `open-cluster-management` namespace.\n\nA pod will
          be created in `open-cluster-management` namespace\n\n```\n$ kubectl get
          pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          19s\n```\n\nThe operator is
          now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installing the operator,
          create an instance of the MulticlusterGlobalHub resource to install the
          Multicluster Global Hub.\n\nThe pods of Multicluster Global Hub are created
          in the `open-cluster-management` namespace:\n\n```\n$ kubectl get pods -n
          open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: multicluster-global-hub-operator.v0.6.0
        version: 0.6.0
      name: release-2.7
    - currentCSV: multicluster-global-hub-operator.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha3",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "largeScale": {
                      "kafka": {
                        "transportFormat": "cloudEvents"
                      }
                    },
                    "type": "largeScale"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.7.0
          createdAt: "2023-07-19T07:53:07Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha3
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.7 or later needs to be installed.\n## How to Install\n\nInstall the Multicluster
          Global Hub Operator by following the instructions that are displayed after
          you select the Install button. The operator must be installed in the same
          namespace as Red Hat Advanced Cluster Management for Kubernetes (RHACM),
          by default, in `open-cluster-management` namespace.\n\nA pod will be created
          in `open-cluster-management` namespace\n\n```\n$ kubectl get pods -n open-cluster-management\nNAME
          \                                                            READY   STATUS
          \   RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp                1/1
          \    Running   0          19s\n```\n\nThe operator is now providing new
          Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installed the operator,
          create an instance of the MulticlusterGlobalHub resource to instantiate
          the Multicluster Global Hub.\n\nYou need to prepare:\n\n- PostgreSQL needs
          to be installed and you must create database for the multicluster global
          hub. You must create a secret named multicluster-global-hub-storage that
          contains the database access credential in `open-cluster-management` namespace.
          You can run the following command to create the secret: \n\n```\nkubectl
          create secret generic multicluster-global-hub-storage -n open-cluster-management
          \\\n  --from-literal=database_uri=<postgresql-uri> \\\n --from-file=ca.crt=<CA-for-postgres-server>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)(Note:
          the client version of kubectl must be v1.21+) to install postgres in `hoh-postgres`
          namespace and create the secret `multicluster-global-hub-storage` in namespace
          `open-cluster-management` automatically. To override the secret namespace,
          set `TARGET_NAMESPACE` environment variable to the RHACM installation namespace
          before executing the script.\n\n- Kafka needs to be installed, and you must
          create three topics, `spec` `status` and `event`. You need to create a secret
          named multicluster-global-hub-transport that contains the kafka access information
          should be created in `open-cluster-management` namespace. You can run the
          following command to create the secret: \n\n```\nkubectl create secret generic
          multicluster-global-hub-transport -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n      --from-file=ca.crt=<CA-cert-for-kafka-server> --from-file=client.crt=<Client-cert-for-kafka-server>
          --from-file=client.key=<Client-key-for-kafka-server> \n```\n >_Note:_ There
          is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `multicluster-global-hub-transport`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\nOnce an instance of the MulticlusterGlobalHub
          is created, the following pods are created in the `open-cluster-management`
          namespace:\n\n```\n$ kubectl get pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-grafana-79f8d59f58-qwcjl
          \                1/1     Running   0          22s\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: beta
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.7.0
        version: 0.7.0
      entries:
      - name: multicluster-global-hub-operator.v0.7.0
        version: 0.7.0
      name: release-2.8
    - currentCSV: multicluster-global-hub-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v1.0.0
          createdAt: "2023-10-16T03:07:33Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the MultiCluster Global Hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.7 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```

          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v1.0.0
        version: 1.0.0
      entries:
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.9
    defaultChannel: release-1.5
    packageName: multicluster-global-hub-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/stolostron/multicluster-global-hub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhbk-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhbk-operator.v22.0.13-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
          createdAt: "2024-09-17T11:20:22Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:884afe0ca1698d85a02aa47b6d60b8222e50d70ad280db6f98e1a4bbdbe13c40
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
        version: 22.0.13-opr.1
      entries:
      - name: rhbk-operator.v22.0.13-opr.1
        version: 22.0.13-opr.1
      - name: rhbk-operator.v22.0.12-opr.1
        version: 22.0.12-opr.1
      - name: rhbk-operator.v22.0.11-opr.2
        version: 22.0.11-opr.2
      - name: rhbk-operator.v22.0.11-opr.1
        version: 22.0.11-opr.1
      - name: rhbk-operator.v22.0.10-opr.1
        version: 22.0.10-opr.1
      - name: rhbk-operator.v22.0.9-opr.1
        version: 22.0.9-opr.1
      - name: rhbk-operator.v22.0.8-opr.2
        version: 22.0.8-opr.2
      - name: rhbk-operator.v22.0.8-opr.1
        version: 22.0.8-opr.1
      - name: rhbk-operator.v22.0.7-opr.1
        version: 22.0.7-opr.1
      - name: rhbk-operator.v22.0.6-opr.1
        version: 22.0.6-opr.1
      name: stable-v22
    - currentCSV: rhbk-operator.v22.0.13-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
          createdAt: "2024-09-17T11:20:22Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:884afe0ca1698d85a02aa47b6d60b8222e50d70ad280db6f98e1a4bbdbe13c40
        version: 22.0.13-opr.1
      entries:
      - name: rhbk-operator.v22.0.13-opr.1
        version: 22.0.13-opr.1
      - name: rhbk-operator.v22.0.12-opr.1
        version: 22.0.12-opr.1
      - name: rhbk-operator.v22.0.11-opr.2
        version: 22.0.11-opr.2
      - name: rhbk-operator.v22.0.11-opr.1
        version: 22.0.11-opr.1
      - name: rhbk-operator.v22.0.10-opr.1
        version: 22.0.10-opr.1
      - name: rhbk-operator.v22.0.9-opr.1
        version: 22.0.9-opr.1
      - name: rhbk-operator.v22.0.8-opr.2
        version: 22.0.8-opr.2
      - name: rhbk-operator.v22.0.8-opr.1
        version: 22.0.8-opr.1
      - name: rhbk-operator.v22.0.7-opr.1
        version: 22.0.7-opr.1
      - name: rhbk-operator.v22.0.6-opr.1
        version: 22.0.6-opr.1
      name: stable-v22.0
    - currentCSV: rhbk-operator.v24.0.11-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
          createdAt: "2025-03-26T02:27:51Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:d339ecf788d1681ba9945a2ea63ca675caec863ec75f497ef916103f4c898003
        version: 24.0.11-opr.1
      entries:
      - name: rhbk-operator.v24.0.11-opr.1
        version: 24.0.11-opr.1
      - name: rhbk-operator.v24.0.10-opr.1
        version: 24.0.10-opr.1
      - name: rhbk-operator.v24.0.9-opr.1
        version: 24.0.9-opr.1
      - name: rhbk-operator.v24.0.8-opr.1
        version: 24.0.8-opr.1
      - name: rhbk-operator.v24.0.7-opr.1
        version: 24.0.7-opr.1
      - name: rhbk-operator.v24.0.6-opr.1
        version: 24.0.6-opr.1
      - name: rhbk-operator.v24.0.5-opr.1
        version: 24.0.5-opr.1
      - name: rhbk-operator.v24.0.4-opr.1
        version: 24.0.4-opr.1
      - name: rhbk-operator.v24.0.3-opr.1
        version: 24.0.3-opr.1
      name: stable-v24
    - currentCSV: rhbk-operator.v24.0.11-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
          createdAt: "2025-03-26T02:27:51Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:d339ecf788d1681ba9945a2ea63ca675caec863ec75f497ef916103f4c898003
        version: 24.0.11-opr.1
      entries:
      - name: rhbk-operator.v24.0.11-opr.1
        version: 24.0.11-opr.1
      - name: rhbk-operator.v24.0.10-opr.1
        version: 24.0.10-opr.1
      - name: rhbk-operator.v24.0.9-opr.1
        version: 24.0.9-opr.1
      - name: rhbk-operator.v24.0.8-opr.1
        version: 24.0.8-opr.1
      - name: rhbk-operator.v24.0.7-opr.1
        version: 24.0.7-opr.1
      - name: rhbk-operator.v24.0.6-opr.1
        version: 24.0.6-opr.1
      - name: rhbk-operator.v24.0.5-opr.1
        version: 24.0.5-opr.1
      - name: rhbk-operator.v24.0.4-opr.1
        version: 24.0.4-opr.1
      - name: rhbk-operator.v24.0.3-opr.1
        version: 24.0.3-opr.1
      name: stable-v24.0
    - currentCSV: rhbk-operator.v26.2.9-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
          createdAt: "2025-09-18T15:03:55Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:50a62b46af44ce0ff7e6c3a15e1bfb2a9611ba4a8b65b8d0ab65817064692c55
        version: 26.2.9-opr.1
      entries:
      - name: rhbk-operator.v26.2.9-opr.1
        version: 26.2.9-opr.1
      - name: rhbk-operator.v26.2.8-opr.1
        version: 26.2.8-opr.1
      - name: rhbk-operator.v26.2.7-opr.1
        version: 26.2.7-opr.1
      - name: rhbk-operator.v26.2.6-opr.1
        version: 26.2.6-opr.1
      - name: rhbk-operator.v26.2.5-opr.1
        version: 26.2.5-opr.1
      - name: rhbk-operator.v26.2.4-opr.2
        version: 26.2.4-opr.2
      - name: rhbk-operator.v26.2.4-opr.1
        version: 26.2.4-opr.1
      - name: rhbk-operator.v26.0.15-opr.1
        version: 26.0.15-opr.1
      - name: rhbk-operator.v26.0.14-opr.1
        version: 26.0.14-opr.1
      - name: rhbk-operator.v26.0.13-opr.1
        version: 26.0.13-opr.1
      - name: rhbk-operator.v26.0.12-opr.1
        version: 26.0.12-opr.1
      - name: rhbk-operator.v26.0.11-opr.1
        version: 26.0.11-opr.1
      - name: rhbk-operator.v26.0.10-opr.1
        version: 26.0.10-opr.1
      - name: rhbk-operator.v26.0.9-opr.1
        version: 26.0.9-opr.1
      - name: rhbk-operator.v26.0.8-opr.1
        version: 26.0.8-opr.1
      - name: rhbk-operator.v26.0.7-opr.1
        version: 26.0.7-opr.1
      - name: rhbk-operator.v26.0.6-opr.1
        version: 26.0.6-opr.1
      - name: rhbk-operator.v26.0.5-opr.1
        version: 26.0.5-opr.1
      name: stable-v26
    - currentCSV: rhbk-operator.v26.0.15-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:fc5fc8cfb2c100c5b3a33fc11a78dfca569f5f34f45d88439340392d2c60afb2
          createdAt: "2025-08-28T19:34:30Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:fc5fc8cfb2c100c5b3a33fc11a78dfca569f5f34f45d88439340392d2c60afb2
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:4bb4d3070fe118bdd7de24e8b68133b8b6e158cea48c9f1abad2b4297be0dfdb
        version: 26.0.15-opr.1
      entries:
      - name: rhbk-operator.v26.0.15-opr.1
        version: 26.0.15-opr.1
      - name: rhbk-operator.v26.0.14-opr.1
        version: 26.0.14-opr.1
      - name: rhbk-operator.v26.0.13-opr.1
        version: 26.0.13-opr.1
      - name: rhbk-operator.v26.0.12-opr.1
        version: 26.0.12-opr.1
      - name: rhbk-operator.v26.0.11-opr.1
        version: 26.0.11-opr.1
      - name: rhbk-operator.v26.0.10-opr.1
        version: 26.0.10-opr.1
      - name: rhbk-operator.v26.0.9-opr.1
        version: 26.0.9-opr.1
      - name: rhbk-operator.v26.0.8-opr.1
        version: 26.0.8-opr.1
      - name: rhbk-operator.v26.0.7-opr.1
        version: 26.0.7-opr.1
      - name: rhbk-operator.v26.0.6-opr.1
        version: 26.0.6-opr.1
      - name: rhbk-operator.v26.0.5-opr.1
        version: 26.0.5-opr.1
      name: stable-v26.0
    - currentCSV: rhbk-operator.v26.2.9-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
          createdAt: "2025-09-18T15:03:55Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:50a62b46af44ce0ff7e6c3a15e1bfb2a9611ba4a8b65b8d0ab65817064692c55
        version: 26.2.9-opr.1
      entries:
      - name: rhbk-operator.v26.2.9-opr.1
        version: 26.2.9-opr.1
      - name: rhbk-operator.v26.2.8-opr.1
        version: 26.2.8-opr.1
      - name: rhbk-operator.v26.2.7-opr.1
        version: 26.2.7-opr.1
      - name: rhbk-operator.v26.2.6-opr.1
        version: 26.2.6-opr.1
      - name: rhbk-operator.v26.2.5-opr.1
        version: 26.2.5-opr.1
      - name: rhbk-operator.v26.2.4-opr.2
        version: 26.2.4-opr.2
      - name: rhbk-operator.v26.2.4-opr.1
        version: 26.2.4-opr.1
      name: stable-v26.2
    defaultChannel: stable-v26.2
    packageName: rhbk-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-opensearchserverless-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-opensearchserverless-controller.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/opensearchserverless-controller:0.2.1
          createdAt: "2025-09-20T21:43:17Z"
          description: AWS OpenSearchServerless controller is a service controller
            for managing OpenSearchServerless resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Collection represents the state of an AWS opensearchserverless
              Collection resource.
            displayName: Collection
            kind: Collection
            name: collections.opensearchserverless.services.k8s.aws
            version: v1alpha1
          - description: SecurityPolicy represents the state of an AWS opensearchserverless
              SecurityPolicy resource.
            displayName: SecurityPolicy
            kind: SecurityPolicy
            name: securitypolicies.opensearchserverless.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon OpenSearchServerless resources in AWS from within your Kubernetes cluster.

          **About Amazon OpenSearchServerless**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon OpenSearchServerless
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opensearchserverless
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon OpenSearchServerless Developer Resources
          url: https://aws.amazon.com/OpenSearchServerless/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: opensearchserverless maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/opensearchserverless-controller:0.2.1
        version: 0.2.1
      entries:
      - name: ack-opensearchserverless-controller.v0.2.1
        version: 0.2.1
      - name: ack-opensearchserverless-controller.v0.2.0
        version: 0.2.0
      - name: ack-opensearchserverless-controller.v0.1.2
        version: 0.1.2
      - name: ack-opensearchserverless-controller.v0.1.1
        version: 0.1.1
      - name: ack-opensearchserverless-controller.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ack-opensearchserverless-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Automotive SIG CentOS
      provider-url: ""
    name: automotive-infra
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: automotive-infra.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "ContainerBuild",
                "metadata": {
                  "name": "cb-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "ImageBuild",
                "metadata": {
                  "name": "imagebuild-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "OperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "automotive-infra"
                  },
                  "name": "operatorconfig-sample",
                  "namespace": "automotive-infra-operator"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/fedora-sig-robotics/infra-ci-operator:v0.0.10
          createdAt: "2025-03-31T14:06:33Z"
          description: Streamline CI/CD for RHIVOS and its ecosystem using OpenShift
            Pipelines
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: automotive-infra-operator
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator
          support: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Client
            name: clients.jumpstarter.dev
            version: v1alpha1
          - kind: ContainerBuild
            name: containerbuilds.automotive.sig.centos.org
            version: v1
          - kind: ExporterAccessPolicy
            name: exporteraccesspolicies.jumpstarter.dev
            version: v1alpha1
          - kind: Exporter
            name: exporters.jumpstarter.dev
            version: v1alpha1
          - kind: ImageBuild
            name: imagebuilds.automotive.sig.centos.org
            version: v1
          - kind: Lease
            name: leases.jumpstarter.dev
            version: v1alpha1
          - kind: OperatorConfig
            name: operatorconfigs.automotive.sig.centos.org
            version: v1
        description: This operator is designed to streamline CI/CD for RHIVOS and
          its ecosystem using OpenShift Pipelines
        displayName: Automotive-CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Automotive
        - RHIVOS
        - Pipelines
        - CI
        links:
        - name: Automotive infra Operator
          url: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator
        maintainers:
        - email: automotive-sig@lists.centos.org
          name: Automotive SIG CentOS Community
        maturity: alpha
        provider:
          name: Automotive SIG CentOS
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/fedora-sig-robotics/infra-ci-operator:v0.0.10
        version: 0.0.10
      entries:
      - name: automotive-infra.v0.0.10
        version: 0.0.10
      - name: automotive-infra.v0.0.9
        version: 0.0.9
      - name: automotive-infra.v0.0.8
        version: 0.0.8
      - name: automotive-infra.v0.0.7
        version: 0.0.7
      - name: automotive-infra.v0.0.6
        version: 0.0.6
      - name: automotive-infra.v0.0.5
        version: 0.0.5
      - name: automotive-infra.v0.0.4
        version: 0.0.4
      - name: automotive-infra.v0.0.3
        version: 0.0.3
      - name: automotive-infra.v0.0.2
        version: 0.0.2
      - name: automotive-infra.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: automotive-infra
    provider:
      name: Automotive SIG CentOS
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: ibm-licensing-operator
      app.kubernetes.io/managed-by: ibm-licensing-operator
      app.kubernetes.io/name: ibm-licensing
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      intent: projected
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-licensing-operator-app
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-licensing-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "IBMLicensing",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-licensing-operator",
                    "app.kubernetes.io/managed-by": "ibm-licensing-operator",
                    "app.kubernetes.io/name": "ibm-licensing"
                  },
                  "name": "instance"
                },
                "spec": {
                  "version": "4.2.14",
                  "apiSecretToken": "ibm-licensing-token",
                  "datasource": "datacollector",
                  "httpsEnable": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
          createdAt: "2025-02-25T12:16:11Z"
          description: The IBM Licensing Operator provides a Kubernetes CRD-Based
            API to monitor the license usage of products.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-licensing-operator-app
          olm.skipRange: '>=1.0.0 <4.2.14'
          operatorframework.io/suggested-namespace: ibm-licensing
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/internal-objects: '["ibmlicensingmetadatas.operator.ibm.com",
            "ibmlicensingdefinitions.operator.ibm.com", "ibmlicensingquerysources.operator.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'IBMLicensing custom resource is used to create an instance
              of the License Service, used to collect information about license usage
              of IBM containerized products and IBM Cloud Paks per cluster. You can
              retrieve license usage data through a dedicated API call and generate
              an audit snapshot on demand. Documentation: For additional details regarding
              install parameters check: https://ibm.biz/icpfs39install. License: Please
              refer to the IBM Terms website (ibm.biz/lsvc-lic) to find the license
              terms for the particular IBM product for which you are deploying this
              component.'
            displayName: IBM License Service
            kind: IBMLicensing
            name: ibmlicensings.operator.ibm.com
            version: v1alpha1
          - kind: IBMLicensingDefinition
            name: ibmlicensingdefinitions.operator.ibm.com
            version: v1
          - description: 'IBMLicensingMetadata is the schema for IBM License Service.
              Thanks to IBMLicensingMetadata, you can track the license usage of Virtual
              Processor Core (VPC) metric by Pods that are managed by automation,
              for which licensing annotations are not defined at the time of deployment,
              such as the open source OpenLiberty. / For more information, see documentation:
              https://ibm.biz/icpfs39install. By installing the operator for IBM License
              Service, you accept the license terms: https://ibm.biz/icpfs39license.'
            displayName: IBMLicensing Metadata
            kind: IBMLicensingMetadata
            name: ibmlicensingmetadatas.operator.ibm.com
            version: v1alpha1
          - kind: IBMLicensingQuerySource
            name: ibmlicensingquerysources.operator.ibm.com
            version: v1
        description: |-
          **Important:**
          - If you are using the IBM Licensing Operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For the link to your IBM Cloud Pak documentation, see [IBM Cloud Paks that use Common Services](https://ibm.biz/BdyGwb).
          - If you are using a stand-alone IBM Container Software, you can use the IBM Licensing Operator directly. For more information, see [ibm-licensing-operator for stand-alone IBM Containerized Software](https://ibm.biz/BdyGwh).

          **IBM Licensing Operator overview**

          IBM Licensing Operator installs License Service. You can use License Service to collect information about license usage of IBM Containerized products and IBM Cloud Paks per cluster. You can retrieve license usage data through a dedicated API call and generate an audit snapshot on demand.

          **Supported platforms**

           Red Hat OpenShift Container Platform 4.10 or newer installed on Linux x86_64, Linux on Power (ppc64le), Linux on IBM Z and LinuxONE.

          **Prerequisites**

           Prerequisites depend on the integration of the License Service with an IBM Cloud Pak or IBM Containerized Software. For more information, see the applicable IBM Cloud Pak documentation or [ibm-licensing-operator for stand-alone IBM Containerized Software](https://ibm.biz/BdyGwh).
        displayName: IBM Licensing
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBMLicensing
        - IBM
        - Cloud
        - License
        - Service
        - License Service
        links:
        - name: GitHub
          url: https://github.com/IBM/ibm-licensing-operator
        maintainers:
        - email: talk2sam@us.ibm.com
          name: talk2sam
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9186c350fa54e8b792d72e47be6b54a67b610f726175c9641795010d45ad51f9
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:ec137115210dae9b69cb2899ed8a5437d94c8a658ae1ae5eb295232b34b0925e
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:96e6db93b2d16ba590252d6568b68f254548aaed73ec17e4bdd8d530ea668be6
        version: 4.2.14
      entries:
      - name: ibm-licensing-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-licensing-operator-app
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator-rhmp.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/dynatrace-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/dynatrace-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator-rhmp.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator-rhmp.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator-rhmp.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator-rhmp.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator-rhmp.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator-rhmp.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator-rhmp.v1.5.0
        version: 1.5.0
      - name: dynatrace-operator-rhmp.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator-rhmp.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator-rhmp.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator-rhmp
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Argo CD Community
      provider-url: ""
    name: argocd-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: argocd-operator.v0.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd-sample"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCDExport",
                "metadata": {
                  "name": "argocdexport-sample"
                },
                "spec": {
                  "argocd": "argocd-sample"
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd-sample"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/argoprojlabs/argocd-operator@sha256:223ad82fb697471ede79201c4a042115c9ca5c00a172d5bdcd5a780e58c46c79
          createdAt: "2025-09-11T06:29:33Z"
          description: Argo CD is a declarative, GitOps continuous delivery tool for
            Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/argoproj-labs/argocd-operator
          support: Argo CD
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: An ApplicationSet is a group or set of Application resources.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: ArgoCDExport is the Schema for the argocdexports API
            displayName: Argo CDExport
            kind: ArgoCDExport
            name: argocdexports.argoproj.io
            version: v1alpha1
          - description: ArgoCD is the Schema for the argocds API
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: ArgoCD is the Schema for the argocds API
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: NotificationsConfiguration is the Schema for the notificationsconfiguration
              API
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
        description: |
          ## Overview

          The Argo CD Operator manages the full lifecycle for [Argo CD](https://argoproj.github.io/argo-cd/) and it's
          components. The operator's goal is to automate the tasks required when operating an Argo CD cluster.

          Beyond installation, the operator helps to automate the process of upgrading, backing up and restoring as needed and
          remove the human as much as possible. In addition, the operator aims to provide deep insights into the Argo CD
          environment by configuring Prometheus and Grafana to aggregate, visualize and expose the metrics already exported by
          Argo CD.

          The operator aims to provide the following, and is a work in progress.

          * Easy configuration and installation of the Argo CD components with sane defaults to get up and running quickly.
          * Provide seamless upgrades to the Argo CD components.
          * Ability to back up and restore an Argo CD cluster from a point in time or on a recurring schedule.
          * Aggregate and expose the metrics for Argo CD and the operator itself using Prometheus and Grafana.
          * Autoscale the Argo CD components as necessary to handle variability in demand.

          ## Usage

          Deploy a basic Argo CD cluster by creating a new ArgoCD resource in the namespace where the operator is installed.

          ```
          apiVersion: argoproj.io/v1alpha1
          kind: ArgoCD
          metadata:
            name: example-argocd
          spec: {}
          ```

          ## Backup

          Backup the cluster above by creating a new ArgoCDExport resource in the namespace where the operator is installed.

          ```
          apiVersion: argoproj.io/v1alpha1
          kind: ArgoCDExport
          metadata:
            name: example-argocdexport
          spec:
            argocd: example-argocd
          ```

          See the [documentation](https://argocd-operator.readthedocs.io) and examples on
          [GitHub](https://github.com/argoproj-labs/argocd-operator) for more information.
        displayName: Argo CD
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - kubernetes
        links:
        - name: Argo CD Project
          url: https://argoproj.github.io/argo-cd/
        - name: Operator Documentation
          url: https://argocd-operator.readthedocs.io
        - name: Operator Source Code
          url: https://github.com/argoproj-labs/argocd-operator
        maintainers:
        - email: aveerama@redhat.com
          name: Abhishek Veeramalla
        maturity: alpha
        provider:
          name: Argo CD Community
        relatedImages:
        - quay.io/argoprojlabs/argocd-operator@sha256:223ad82fb697471ede79201c4a042115c9ca5c00a172d5bdcd5a780e58c46c79
        version: 0.15.0
      entries:
      - name: argocd-operator.v0.15.0
        version: 0.15.0
      - name: argocd-operator.v0.14.1
        version: 0.14.1
      - name: argocd-operator.v0.14.0
        version: 0.14.0
      - name: argocd-operator.v0.13.0
        version: 0.13.0
      - name: argocd-operator.v0.12.0
        version: 0.12.0
      - name: argocd-operator.v0.11.0
        version: 0.11.0
      - name: argocd-operator.v0.10.1
        version: 0.10.1
      - name: argocd-operator.v0.10.0
        version: 0.10.0
      - name: argocd-operator.v0.9.1
        version: 0.9.1
      - name: argocd-operator.v0.9.0
        version: 0.9.0
      - name: argocd-operator.v0.8.0
        version: 0.8.0
      - name: argocd-operator.v0.7.0
        version: 0.7.0
      - name: argocd-operator.v0.5.0
        version: 0.5.0
      - name: argocd-operator.v0.4.0
        version: 0.4.0
      - name: argocd-operator.v0.3.0
        version: 0.3.0
      - name: argocd-operator.v0.2.1
        version: 0.2.1
      - name: argocd-operator.v0.2.0
        version: 0.2.0
      - name: argocd-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: argocd-operator
    provider:
      name: Argo CD Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-framework
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-framework.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Framework",
                "metadata": {
                  "name": "framework-sample"
                },
                "spec": {
                  "kcos-eula": {
                    "image": {
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    },
                    "service": {
                      "nodePort": 0,
                      "type": "ClusterIP"
                    }
                  },
                  "kcos-logging-framework": {
                    "image": {
                      "remoteRepository": "docker.io",
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    },
                    "ingress": {
                      "annotations": {
                        "nginx.ingress.kubernetes.io/auth-method": "POST",
                        "nginx.ingress.kubernetes.io/auth-response-headers": "X-AUTH-USER",
                        "nginx.ingress.kubernetes.io/auth-url": "https://127.0.0.1/token/validate",
                        "nginx.ingress.kubernetes.io/cors-allow-credentials": "true",
                        "nginx.ingress.kubernetes.io/cors-allow-methods": "PUT, GET, POST, OPTIONS, PATCH",
                        "nginx.ingress.kubernetes.io/cors-allow-origin": "https://admin.example.com,https://localhost:4200",
                        "nginx.ingress.kubernetes.io/enable-cors": "true",
                        "nginx.ingress.kubernetes.io/proxy-body-size": "0",
                        "nginx.ingress.kubernetes.io/ssl-redirect": "true"
                      },
                      "enabled": true
                    }
                  },
                  "kcos-postgresql": {
                    "jobCopySecret": {
                      "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                    },
                    "postgresql": {
                      "image": {
                        "registry": "prod.harbor.keysight.digital/loadcore"
                      },
                      "initdbScripts": {
                        "init_script.sh": "#!/bin/sh\nexport PGUSER=postgres\npsql -c \"SELECT 1 FROM pg_database WHERE datname = 'db'\" | grep -q 1 || psql -c \"CREATE DATABASE db\"\npsql -c \"SELECT 1 FROM pg_database WHERE datname = 'kcoslicensingdb'\" | grep -q 1 || psql -c \"CREATE DATABASE kcoslicensingdb\"\n"
                      },
                      "metrics": {
                        "image": {
                          "registry": "docker.io"
                        }
                      },
                      "persistence": {
                        "storageClass": "managed-loadcore-storage"
                      },
                      "pgHbaConfiguration": "local all all trust\nhost all all 0.0.0.0/0 trust",
                      "securityContext": {
                        "enabled": false
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": false
                        }
                      },
                      "volumePermissions": {
                        "image": {
                          "registry": "docker.io"
                        },
                        "securityContext": {
                          "runAsUser": "auto"
                        }
                      }
                    }
                  },
                  "kcos-system-diagnostics": {
                    "environment": {
                      "ELASTICSEARCH_ENDPOINT": "elasticsearch.keysight-wap.svc.cluster.local.",
                      "IGNORE_DATA_PREFIX": "wireless_stats"
                    },
                    "image": {
                      "initContainer": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf",
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    }
                  },
                  "subcharts": {
                    "enable": {
                      "kcos-coredns-fixup": false,
                      "kcos-eula": true,
                      "kcos-local-storage": false,
                      "kcos-logging-framework": true,
                      "kcos-logging-rotate": false,
                      "kcos-naas": false,
                      "kcos-postgresql": true,
                      "kcos-shell": false,
                      "kcos-system-diagnostics": true,
                      "kcos-vital-service": false,
                      "nfs-server-provisioner": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:27:28Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Framework
            name: frameworks.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Framework Operator
        displayName: Keysight KCOS Framework Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Framework
          url: https://keysight-kcos-framework.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-framework@sha256:531f278def85e33d453479fce74127e12e3ed62a251872eab421c6561bedd0c5
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        version: 0.1.21
      entries:
      - name: keysight-kcos-framework.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-framework
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx-openshift-operator.v0.0.36
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
          createdAt: "2019-07-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX on JBoss EAP 7
            displayName: Joget DX on JBoss EAP 7
            kind: JogetDX
            name: jogetdx.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX7/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
        - registry.connect.redhat.com/joget/joget-dx7-eap7@sha256:7245501f5fccf011d75c034b7d55542cfe3bd2c64cc7e48b3224a48b623af282
        version: 0.0.36
      entries:
      - name: joget-dx-openshift-operator.v0.0.36
        version: 0.0.36
      - name: joget-dx-openshift-operator.v0.0.35
        version: 0.0.35
      - name: joget-dx-openshift-operator.v0.0.34
        version: 0.0.34
      - name: joget-dx-openshift-operator.v0.0.33
        version: 0.0.33
      - name: joget-dx-openshift-operator.v0.0.32
        version: 0.0.32
      - name: joget-dx-openshift-operator.v0.0.31
        version: 0.0.31
      - name: joget-dx-openshift-operator.v0.0.30
        version: 0.0.30
      - name: joget-dx-openshift-operator.v0.0.29
        version: 0.0.29
      - name: joget-dx-openshift-operator.v0.0.28
        version: 0.0.28
      - name: joget-dx-openshift-operator.v0.0.27
        version: 0.0.27
      - name: joget-dx-openshift-operator.v0.0.26
        version: 0.0.26
      - name: joget-dx-openshift-operator.v0.0.25
        version: 0.0.25
      - name: joget-dx-openshift-operator.v0.0.24
        version: 0.0.24
      - name: joget-dx-openshift-operator.v0.0.23
        version: 0.0.23
      - name: joget-dx-openshift-operator.v0.0.22
        version: 0.0.22
      - name: joget-dx-openshift-operator.v0.0.21
        version: 0.0.21
      - name: joget-dx-openshift-operator.v0.0.20
        version: 0.0.20
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Habana Labs Ltd.
      provider-url: https://habana.ai
    name: habana-ai-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: habana-ai-operator.v1.19.2-32-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.19.2-32"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.19.2"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.19.2-32"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.19.2-32"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.19.2-32"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.19.2-32"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.19.2-32"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.19.2-32"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.19.2-32"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.19.2-32"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.19.2-32"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:975ed2135061d3bdc145c83abaa15f99e4c9f6efb15c708de41aa00fbcb72333
          createdAt: "2025-02-10T07:33:30Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.19.2-32-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi AI accelerator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:ac16752b53df69573da257d3a3a717a214b2eb1656779a7ef94084d74363fb46
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:b12802e7ef05273fceaef3a9987c3c4ae760714e6f333363f665eb26be9175e9
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:f05cd78d7593b877ec8a75175a9faa72d9b991d81c74af9ecd92208fde296eca
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:170a15041fc3e3901b4e8e0b9a32d53fde11026776c17a6560d500dacb976c56
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:d728a91a36503f07dfacd49fdbfefe8083607ca33babb5f2dfc1e9de05386454
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:e667695b845cd2be887f117dac66861d2c5e74def009d2939559ae7c30574dc0
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:975ed2135061d3bdc145c83abaa15f99e4c9f6efb15c708de41aa00fbcb72333
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:eeb4573de20d3341a085c75e669b7d37c424e348421d7916440fbbfff12ea482
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:71d4c7181930c71c0ca4e6b696ea416a687eec6d798a43e854fcd1bfdc224a0c
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:72cc75353615369d12a07f5488bcdc041a8435906ca6e16c53b001016108e099
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:373b9caa27a96bb7979f7a6aec246795142ce3a28c6b97a8902035ec20814fa2
        version: 1.19.2-32-1
      entries:
      - name: habana-ai-operator.v1.19.2-32-1
        version: 1.19.2-32-1
      name: "1.19"
    - currentCSV: habana-ai-operator.v1.20.0-543-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.20.0-543"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.20.0-4"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.20.0-543-2"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.20.0-543"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.20.0-543"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.20.0-543-4"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.20.0-543"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.20.0-543-4"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:d5b37eafe48446e1ffa492e195f4519b77d4a11f6750e5c71f27c1536bdfd2cc
          createdAt: "2025-02-26T09:03:44Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.20.0-543-5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:9ab54a84a00e8afc078eec2a7d643f09b1bd674e0e230733fbce7108e392e009
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:d5b37eafe48446e1ffa492e195f4519b77d4a11f6750e5c71f27c1536bdfd2cc
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:317765eac4e0332a0c4f6bead5563b343a8c22e6e0eaa62276ff66be08608c30
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:c5beac6209d0671d3a49b63f1cf95401055b042430eafaefc3859b5c633f0e84
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:68d0f44170cbf3ea15b793e0503fb07ce7e32488011fc85ccb6c9234b3b9302f
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:3ea80aa8994e8b19fbbdcd18ba2c35800774da3ca4e27a2514d0f5f9676786fa
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:957702b58ee66d7339a20b2a5ece783606d6ccebde627c58951d5ecd6bb1b13f
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:cba03259b33c4cb9bcd22622ab0dda962846ea0a80c727779365cf55129ad308
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:d14c457106b6133fc41ef727ac40a6bf2f0e5c97f21659969d1daebdc1310b43
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:c662de244c5900c89b918b16982298676d9a6d553360a5716174b28d4d64a635
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:a673900f08c38c9bb40c90b3cddea580afcc6a99195d816cd0b5cae9831ca5c9
        version: 1.20.0-543-5
      entries:
      - name: habana-ai-operator.v1.20.0-543-5
        version: 1.20.0-543-5
      name: "1.20"
    - currentCSV: habana-ai-operator.v1.21.5-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.21.5-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.21.5-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.21.5-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:0741da99140d137474b9e7edba921672eab8ff6277e1dd8bf3a4179e87f01dd4
          createdAt: "2025-09-17T11:55:32Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.21.5-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:0741da99140d137474b9e7edba921672eab8ff6277e1dd8bf3a4179e87f01dd4
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:a216ec6a6e7df50e067c557e9ed5871a3308b2ba42a4e41c9642f0c2c4503d88
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:9692d079ef18a364ed235c187482f329b3a5764a773e363acb1238044ad4a884
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:c95d3ffd6c91be3418415740258d33a8dbcedad6fe7d6bfe68d4dd122027447b
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:bc3749ff153fff71e252820cbf22fcd0753b195b2204f9731e42af29fb7ddaa0
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:bd57e7059467f0ca1667d19a70bd9a6ad96b83bbcb53c09fdad5a262fee76af9
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:c84e997c7d8c22c4d4e902a29e45e5364f0ad306f17d03a7d8ddb4391142e167
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:256cd8bebec3ed835e187e925a14a6b605ce9b3a17fe7d8886e0b32cde163d90
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:1e488c8ad31784dab0fe70144f5146d5809e2fc7c86424d1b512ab887787e28f
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:c7147359dcd6a2e5adee95259d37743ff302b1deaeb13b184044f4029d33f44a
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:d7f64d15c379524597aa2b2c5b4586c890442fa32356692013cf665b01b0f092
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:a0b45ccfb6af54b1043ccc2336ad3d7990cc447af5ba3621f3eb1953cd40ffc5
        version: 1.21.5-6
      entries:
      - name: habana-ai-operator.v1.21.5-6
        version: 1.21.5-6
      name: "1.21"
    - currentCSV: habana-ai-operator.v1.22.1-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.22.1-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
          createdAt: "2025-09-17T10:15:21Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.22.1-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:1e15f3de21c2b1e20f54d1f8a973290c6d940624506880c5bc3eb880011bcc02
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:d1974d72774c626e77707e8f44bcc631ac142bfe2a5395714969a964c963e36e
        - vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer@sha256:2cf151c2a518e3e1c7c09efb8301e469d266d1f030fff2e6180b086cc2e04ee7
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:704b4bebdfe6b85693df4b213f67862859a554c21ce4e614736172f64b541ea7
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:c9ff6f3fe24ef87c7bfc71bdb1cd7900941df52f9ad1293badd7e14783ce7ef6
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:e34a6820a58e7e8090b521781c76acc3867c94f8291634de7b138f928e0bdb74
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:53c1844b5e8eb08d0cc29865a732100208c5dfc0fe55df8b32e589e319bcab6c
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:38c21eae6249e5571921998660418e59870bc0880e7b8c9b95a4a8db30ef2e7d
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:1925586fda34c6fa444a59b1fa7be20b3cf16e845bf56566ad2a41c276b0cfe9
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:392fcd2a4730299e6e3bab9f7e96fc3b710de8adb374a2c64b7316e097ed05f9
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:7140877b23291a5db3e5e28def3c9d5e7cb7be62f046e2fb7ac45eff369cea84
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:8f413edfff1978b35ba3998ed5f1c6293e6764b1d2cd9faa08f0bd81dc6e3e10
        version: 1.22.1-6
      entries:
      - name: habana-ai-operator.v1.22.1-6
        version: 1.22.1-6
      name: "1.22"
    - currentCSV: habana-ai-operator.v1.22.1-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.22.1-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
          createdAt: "2025-09-17T10:15:21Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.22.1-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:c9ff6f3fe24ef87c7bfc71bdb1cd7900941df52f9ad1293badd7e14783ce7ef6
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:392fcd2a4730299e6e3bab9f7e96fc3b710de8adb374a2c64b7316e097ed05f9
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:38c21eae6249e5571921998660418e59870bc0880e7b8c9b95a4a8db30ef2e7d
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:1925586fda34c6fa444a59b1fa7be20b3cf16e845bf56566ad2a41c276b0cfe9
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:704b4bebdfe6b85693df4b213f67862859a554c21ce4e614736172f64b541ea7
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:7140877b23291a5db3e5e28def3c9d5e7cb7be62f046e2fb7ac45eff369cea84
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:8f413edfff1978b35ba3998ed5f1c6293e6764b1d2cd9faa08f0bd81dc6e3e10
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:1e15f3de21c2b1e20f54d1f8a973290c6d940624506880c5bc3eb880011bcc02
        - vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer@sha256:2cf151c2a518e3e1c7c09efb8301e469d266d1f030fff2e6180b086cc2e04ee7
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:e34a6820a58e7e8090b521781c76acc3867c94f8291634de7b138f928e0bdb74
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:53c1844b5e8eb08d0cc29865a732100208c5dfc0fe55df8b32e589e319bcab6c
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:d1974d72774c626e77707e8f44bcc631ac142bfe2a5395714969a964c963e36e
        version: 1.22.1-6
      entries:
      - name: habana-ai-operator.v1.22.1-6
        version: 1.22.1-6
      name: stable
    defaultChannel: stable
    packageName: habana-ai-operator
    provider:
      name: Habana Labs Ltd.
      url: https://habana.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-multicluster-orchestrator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-multicluster-orchestrator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "items": [
                    {
                      "clusterName": "local-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    },
                    {
                      "clusterName": "spoke-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    }
                  ],
                  "manageS3": false,
                  "schedulingIntervals": [
                    "5m"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["odf-multicluster-console"]'
          createdAt: "2025-09-12T07:33:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          Orchestrator for OpenShift Data Foundation clusters running across multiple OpenShift clusters.
          It uses Red Hat Advanced Cluster Management for Kubernetes as the multicluster control plane.
        displayName: ODF Multicluster Orchestrator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: ODF Multicluster Orchestrator
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: uchapaga@redhat.com
          name: Umanga Chapagain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-multicluster-rhel9-operator@sha256:09e2c7a20ab92ddc5d589072bb0ab9ce2ed8706ed8364f1b11192d7a7396e138
        - registry.redhat.io/odf4/odf-multicluster-console-rhel9@sha256:7527f4a14aed0386b8f160fa8a92b7ff1f58114970e07c3ec6c433d2a53459b5
        version: 4.15.20-rhodf
      entries:
      - name: odf-multicluster-orchestrator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odf-multicluster-orchestrator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odf-multicluster-orchestrator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odf-multicluster-orchestrator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odf-multicluster-orchestrator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odf-multicluster-orchestrator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odf-multicluster-orchestrator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odf-multicluster-orchestrator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odf-multicluster-orchestrator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odf-multicluster-orchestrator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odf-multicluster-orchestrator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odf-multicluster-orchestrator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odf-multicluster-orchestrator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odf-multicluster-orchestrator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odf-multicluster-orchestrator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odf-multicluster-orchestrator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odf-multicluster-orchestrator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odf-multicluster-orchestrator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odf-multicluster-orchestrator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odf-multicluster-orchestrator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odf-multicluster-orchestrator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "items": [
                    {
                      "clusterName": "local-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    },
                    {
                      "clusterName": "spoke-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    }
                  ],
                  "manageS3": false,
                  "schedulingIntervals": [
                    "5m"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["odf-multicluster-console"]'
          createdAt: "2025-09-12T06:32:35Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          Orchestrator for OpenShift Data Foundation clusters running across multiple OpenShift clusters.
          It uses Red Hat Advanced Cluster Management for Kubernetes as the multicluster control plane.
        displayName: ODF Multicluster Orchestrator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: ODF Multicluster Orchestrator
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: uchapaga@redhat.com
          name: Umanga Chapagain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-multicluster-rhel9-operator@sha256:78d6bc14b95d1ee91ed529422062ec191afc489d6ab4e75906ae3c7fe58749cb
        - registry.redhat.io/odf4/odf-multicluster-console-rhel9@sha256:97d3645560388f6ae08fc12cafb58b6c766e38d83de04b645b3e6fb03fbba231
        version: 4.16.16-rhodf
      entries:
      - name: odf-multicluster-orchestrator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odf-multicluster-orchestrator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odf-multicluster-orchestrator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odf-multicluster-orchestrator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odf-multicluster-orchestrator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odf-multicluster-orchestrator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odf-multicluster-orchestrator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odf-multicluster-orchestrator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odf-multicluster-orchestrator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odf-multicluster-orchestrator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odf-multicluster-orchestrator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odf-multicluster-orchestrator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odf-multicluster-orchestrator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odf-multicluster-orchestrator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odf-multicluster-orchestrator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odf-multicluster-orchestrator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odf-multicluster-orchestrator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odf-multicluster-orchestrator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-access-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-access-operator.v23.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyAccess",
                "metadata": {
                  "name": "isva-sample"
                },
                "spec": {
                  "image": "icr.io/isva/verify-access-wrp:10.0.5.0",
                  "instance": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isva/verify-access-operator:23.3.0
          createdAt: Mon Mar  20 00:59:34 UTC 2023
          description: The IBM Security Verify Access Operator manages the lifecycle
            of IBM Security Verify Access worker containers.
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM-Security/verify-access-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyAccess is the Schema for the ibmsecurityverifyaccesses
              API.
            displayName: IBMSecurity Verify Access
            kind: IBMSecurityVerifyAccess
            name: ibmsecurityverifyaccesses.ibm.com
            version: v1
        description: |+
          In a world of highly fragmented access management environments, [IBM Security Verify Access](https://www.ibm.com/au-en/products/verify-access) helps you simplify your users' access while more securely adopting web, mobile and cloud technologies. This solution helps you strike a balance between usability and security through the use of risk-based access, single sign-on, integrated access management control, identity federation and its mobile multi-factor authentication capability, IBM Verify. Take back control of your access management with IBM Security Verify Access.

          The IBM Security Verify Access operator provides lifecycle management of the lightweight containers which are used to protect an environment, namely:
          * [Web Reverse Proxy](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-web-reverse-proxy)
          * [Runtime](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-runtime)
          * [Distributed Session Cache](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-distributed-session-cache)

          The operator will manage the deployment of these lightweight IBM Security Verify Access worker containers, and also control the rolling restart of these containers when a configuration snapshot is updated.

          See the project [Readme](https://github.com/IBM-Security/verify-access-operator/blob/master/README.md) for further information and details.

        displayName: IBM Security Verify Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Access Product Information
          url: https://www.ibm.com/au-en/products/verify-access
        - name: Verify Access Documentation
          url: https://www.ibm.com/docs/en/sva
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Access Development Team
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - icr.io/isva/verify-access-operator:23.3.0
        version: 23.3.0
      entries:
      - name: ibm-security-verify-access-operator.v23.3.0
        version: 23.3.0
      - name: ibm-security-verify-access-operator.v22.10.0
        version: 22.10.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-access-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: oadp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oadp-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "Velero",
                "metadata": {
                  "name": "example-velero",
                  "namespace":"openshift-adp"
                },
                "spec": {
                  "backupStorageLocations": [
                    {
                      "config": {
                        "profile": "default",
                        "region": "us-east-1"
                      },
                      "name": "default",
                      "objectStorage": {
                        "bucket": "myBucket",
                        "prefix": "velero"
                      },
                      "provider": "aws"
                    }
                  ],
                  "defaultVeleroPlugins": [
                    "aws",
                    "csi",
                    "openshift"
                  ],
                  "veleroFeatureFlags": [
                    "EnableCSI"
                  ],
                  "enableRestic": true,
                  "volumeSnapshotLocations": [
                    {
                      "config": {
                        "profile": "default",
                        "region": "us-west-1"
                      },
                      "name": "default",
                      "provider": "aws"
                    }
                  ]
                }
               },
                {
                  "apiVersion": "velero.io/v1",
                  "kind": "Backup",
                  "metadata": {
                    "name": "backup",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "BackupStorageLocation",
                  "metadata": {
                    "name": "backupstoragelocation",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "DeleteBackupRequest",
                  "metadata": {
                    "name": "deletebackuprequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "DownloadRequest",
                  "metadata": {
                    "name": "downloadrequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "PodVolumeBackup",
                  "metadata": {
                    "name": "podvolumebackup",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "PodVolumeRestore",
                  "metadata": {
                    "name": "podvolumerestore",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "ResticRepository",
                  "metadata": {
                    "name": "resticrepository",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "Restore",
                  "metadata": {
                    "name": "restore",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "Schedule",
                  "metadata": {
                    "name": "schedule",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "ServerStatusRequest",
                  "metadata": {
                    "name": "serverstatusrequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "VolumeSnapshotLocation",
                  "metadata": {
                    "name": "volumesnapshotlocation",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/konveyor/oadp-operator@sha256:ecb8534fbc245e890b9c7766946d01e1a8b4ce9be359c4ada4c95fd908b679b3
          createdAt: "2020-09-08T12:21:00Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          olm.skipRange: '>=0.0.0 <0.4.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Velero is a tool to safely backup and restore, perform disaster
              recovery and migrate Kubernetes cluster resources and persistent volumes.
            displayName: Velero
            kind: Velero
            name: veleroes.oadp.openshift.io
            version: v1alpha1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - description: Backup is a Velero resource that respresents the capture
              of Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Velero on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://oadp.konveyor.io/
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/konveyor/velero-restic-restore-helper@sha256:c4d12829e42cabe70b662a659875900677dfb415f912f931c3bf4c09a26aed5b
        - quay.io/konveyor/openshift-velero-plugin@sha256:aec8b4fc66cfcebd5a35ce99d8c985453ebf25c11453d30256089a8b425990b5
        - quay.io/konveyor/velero-plugin-for-aws@sha256:2197c6fef32910747526d33c1462b75cb7940aeace65791d0a5ddaa038480e52
        - quay.io/konveyor/velero-plugin-for-csi@sha256:5c2261effc902d8baa16e38265f9ad1fbe5284a3e6abd2bbc97761c161fa92cf
        - quay.io/konveyor/registry@sha256:b74ab3278b79eb89f99b6970c6aa0f5306221a8495b5496faad6cab2f74ff2cd
        - quay.io/konveyor/oadp-operator@sha256:ecb8534fbc245e890b9c7766946d01e1a8b4ce9be359c4ada4c95fd908b679b3
        - quay.io/konveyor/velero-plugin-for-microsoft-azure@sha256:62a344e5cf196e55dae92e537648987d758ccc75985bc45e9be6d0646a9e64d0
        - quay.io/konveyor/velero-plugin-for-gcp@sha256:3300458b1d6c1514dca7d24353ad03bd92db396b2a29f7d3522563160a5f5219
        - quay.io/konveyor/velero@sha256:e7865cc707664eb6ee767e77ccd933420604c2562736b53e16c8043a51df9d86
        version: 0.4.2
      entries:
      - name: oadp-operator.v0.4.2
        version: 0.4.2
      - name: oadp-operator.v0.4.1
        version: 0.4.1
      - name: oadp-operator.v0.4.0
        version: 0.4.0
      - name: oadp-operator.v0.3.0
        version: 0.3.0
      name: beta
    - currentCSV: oadp-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "restic": {
                      "enable": true
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ResticRepository",
                "metadata": {
                  "name": "resticrepository",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/konveyor/oadp-operator@sha256:90609044d52173619ea1e63ddef0ef3e7c1c939ac737b64bd427cbed5374dd27
          createdAt: "2020-09-08T12:21:00Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          olm.skipRange: '>=0.0.0 <0.5.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is a Velero resource that respresents the capture
              of Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Data Protection Applications on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://oadp.konveyor.io/
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/konveyor/registry@sha256:1be970879206258335357e3cf630138c24b8a5ccab35aae05eccb260e5e8ffb4
        - quay.io/konveyor/openshift-velero-plugin@sha256:221ab0bcd1d95fc166ef7319d187e196b59d2fbf9eefd29d91bd6435d45b5c3c
        - quay.io/konveyor/oadp-operator@sha256:90609044d52173619ea1e63ddef0ef3e7c1c939ac737b64bd427cbed5374dd27
        - quay.io/konveyor/velero@sha256:0ea5336dcbf0a1fab561697b24da01a0e5b533ee23553fbfd5f8e002f5a31c3e
        - quay.io/konveyor/velero-restic-restore-helper@sha256:c16248d027c09cd5b7360ae38484df88ca1cc01f9e470e3eade95076ee1a6b84
        - quay.io/konveyor/velero-plugin-for-aws@sha256:1e8bc27d6b60000e1406085d582910cb530af3fb901ef9bc780628df6c766288
        - quay.io/konveyor/velero-plugin-for-microsoft-azure@sha256:2d0089cbe715b482fb42e1acd9add70b975a8575636d8b143e6f90cef317ea13
        - quay.io/konveyor/velero-plugin-for-gcp@sha256:ee463ed2a8019c560961b118f8a87b587ed302f24fd25480c5d4ddad2557e92f
        - quay.io/konveyor/velero-plugin-for-csi@sha256:8c36ccfd294dafc415a7f113770cbbd66dbf6ecf8fef3bc4e73adc9878c606b0
        version: 0.5.6
      entries:
      - name: oadp-operator.v0.5.6
        version: 0.5.6
      - name: oadp-operator.v0.5.5
        version: 0.5.5
      - name: oadp-operator.v0.5.4
        version: 0.5.4
      - name: oadp-operator.v0.5.3
        version: 0.5.3
      - name: oadp-operator.v0.5.2
        version: 0.5.2
      - name: oadp-operator.v0.5.1
        version: 0.5.1
      - name: oadp-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: oadp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Trustification
      provider-url: ""
    name: trustify-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trustify-operator.v0.1.0-alpha.9
      currentCSVDesc:
        annotations:
          alm-examples: |
            [{
              "apiVersion": "org.trustify/v1alpha1",
              "kind": "Trustify",
              "metadata": {
                "name": "myapp"
              },
              "spec": { }
            }]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          containerImage: ghcr.io/trustification/trustify-operator:v0.1.0-alpha.9
          createdAt: "2025-03-07T11:57:13+00:00"
          description: An Operator for installing and managing Trustify
          repository: https://github.com/trustification/trustify-operator
          support: https://github.com/trustification/trustify-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Trustify instance
            displayName: Trustify
            kind: Trustify
            name: trustifies.org.trustify
            version: v1alpha1
        description: |+
          Trustify is vendor-neutral, thought-leadering, mostly informational collection of resources devoted to making Software Supply Chains easier to create, manage, consume and ultimately... to trust!


          You can use Trustify for:

          - Store and Manage all the SBOM (Software Bill of Materials) files of your company.
          - Understand which are the Vulnerabilities that affect each of your SBOM files
          - Understand exactly which Packages are included/shipped within each SBOM

          Trustify is a project within the [Trustification community](https://trustification.io/).


          ### Install
          Once you have successfully installed the Operator, proceed to deploy components by creating the required CR.

          By default, the Operator installs the following components on a target cluster:

          * Server
          * UI

          ### Documentation
          Documentation can be found on our [website](https://trustification.io/).

          ### Getting help
          If you encounter any issues while using Trustify, you can create an issue on our [Github repo](https://github.com/trustification/trustify/issues), for bugs, enhancements or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find
          * Fixing issues by opening [Pull Requests](https://github.com/trustification/trustify/pulls)

        displayName: Trustify Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trust
        links:
        - name: Website
          url: https://trustification.io/
        - name: Github
          url: https://github.com/trustification/trustify
        maintainers:
        - email: trustification@googlegroups.com
          name: Trustification
        maturity: alpha
        minKubeVersion: 1.23.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Trustification
        relatedImages:
        - ghcr.io/trustification/trustify-operator:v0.1.0-alpha.9
        version: 0.1.0-alpha.9
      entries:
      - name: trustify-operator.v0.1.0-alpha.9
        version: 0.1.0-alpha.9
      - name: trustify-operator.v0.1.0-alpha.8
        version: 0.1.0-alpha.8
      - name: trustify-operator.v0.1.0-alpha.7
        version: 0.1.0-alpha.7
      - name: trustify-operator.v0.1.0-alpha.6
        version: 0.1.0-alpha.6
      - name: trustify-operator.v0.1.0-alpha.5
        version: 0.1.0-alpha.5
      - name: trustify-operator.v0.1.0-alpha.4
        version: 0.1.0-alpha.4
      - name: trustify-operator.v0.1.0-alpha.3
        version: 0.1.0-alpha.3
      - name: trustify-operator.v0.1.0-alpha.2
        version: 0.1.0-alpha.2
      name: alpha
    defaultChannel: alpha
    packageName: trustify-operator
    provider:
      name: Trustification
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-directory-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-directory-operator.v23.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyDirectory",
                "metadata": {
                  "name": "ibmsecurityverifydirectory-sample"
                },
                "spec": {
                  "pods": {
                    "configMap": {
                      "proxy": {
                        "key": "config.yaml",
                        "name": "isvd-proxy-config"
                      },
                      "server": {
                        "key": "config.yaml",
                        "name": "isvd-server-config"
                      }
                    },
                    "image": {
                      "label": "latest",
                      "repo": "icr.io/isvd"
                    },
                    "proxy": {
                      "replicas": 1
                    },
                    "serviceAccountName": "isvd"
                  },
                  "replicas": {
                    "pvcs": [
                      "replica-1",
                      "replica-2"
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: icr.io/isvd/verify-directory-operator@sha256:d1dd7037584c6818dcf8d5aa69472278833bb8e37fba2aec5a69185ba88dcc34
          createdAt: "2023-04-18T22:10:32Z"
          description: The IBM Security Verify Directory operator is used to scale
            an IBM Security Verify Directory Server environment.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ibm-security/verify-directory-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyDirectory is the Schema for the ibmsecurityverifydirectories
              API
            displayName: IBMSecurity Verify Directory
            kind: IBMSecurityVerifyDirectory
            name: ibmsecurityverifydirectories.ibm.com
            version: v1
        description: "IBM Security Verify Directory is a scalable, standards-based
          identity directory that helps simplify identity and directory management.
          Verify Directory helps consolidate identity silos into a single identity
          source. Verify Directory is purpose-built to provide a directory foundation
          that can help provide a trusted identity data infrastructure that assists
          in enabling mission-critical security and authentication. It is designed
          to deliver a reliable, scalable, standards-based identity data platform
          that interoperates with a broad range of operating systems and applications.
          Verify Directory supports Lightweight Directory Access Protocol (LDAP) V3,
          offering a flexible and highly scalable LDAP infrastructure. \n\nThe operator
          can be used to manage deployed instances of the IBM Security Verify Directory
          server, providing an easy mechanism to scale the number of servers which
          are used to handle the workload.\n\nSee the project [Readme](https://github.com/ibm-security/verify-directory-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Security Verify Directory
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Security
        - Identity and Access
        links:
        - name: IBM Security Verify Directory Documentation
          url: https://www.ibm.com/docs/en/svd
        maintainers:
        - email: iamdev@us.ibm.com
          name: IBM Security Verify Directory Development Team
        maturity: stable
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - icr.io/isvd/verify-directory-operator@sha256:d1dd7037584c6818dcf8d5aa69472278833bb8e37fba2aec5a69185ba88dcc34
        version: 23.4.1
      entries:
      - name: ibm-security-verify-directory-operator.v23.4.1
        version: 23.4.1
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-directory-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Flux project
      provider-url: ""
    name: flux
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.toolkit.fluxcd.io/v2beta1",
                "kind": "HelmRelease",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "5m",
                    "chart": {
                      "spec": {
                          "chart": "podinfo",
                          "version": "4.0.x",
                          "sourceRef": {
                            "kind": "HelmRepository",
                            "name": "podinfo",
                            "namespace": "flux-system"
                          },
                          "interval": "1m"
                      }
                    },
                    "values": {
                      "replicaCount": 2
                    }
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "GitRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "1m",
                    "ref": {
                      "branch": "master"
                    },
                    "url": "https://github.com/stefanprodan/podinfo"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmRepository",
                "metadata": {
                    "name": "stable",
                    "namespace": "default"
                },
                "spec": {
                    "url": "https://kubernetes-charts.storage.googleapis.com/",
                    "interval": "10m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmChart",
                "metadata": {
                    "name": "redis",
                    "namespace": "default"
                },
                "spec": {
                    "chart": "redis",
                    "version": "10.5.7",
                    "sourceRef": {
                      "name": "stable",
                      "kind": "HelmRepository"
                    },
                    "interval": "5m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "Bucket",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "1m",
                    "provider": "generic",
                    "bucketName": "podinfo",
                    "endpoint": "minio.minio.svc.cluster.local:9000",
                    "insecure": true,
                    "secretRef": {
                      "name": "minio-credentials"
                    }
                }
              },
              {
                "apiVersion": "kustomize.toolkit.fluxcd.io/v1beta2",
                "kind": "Kustomization",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "5m",
                    "path": "./deploy/overlays/dev",
                    "prune": true,
                    "sourceRef": {
                      "kind": "GitRepository",
                      "name": "podinfo"
                    },
                    "timeout": "2m",
                    "wait": true
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "address": "https://github.com/stefanprodan/podinfo",
                    "secretRef": {
                      "name": "api-token"
                    }
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Alert",
                "metadata": {
                    "name": "webapp",
                    "namespace": "default"
                },
                "spec": {
                    "providerRef": {
                      "name": "on-call-slack"
                    },
                    "eventSeverity": "info",
                    "eventSources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "Bucket",
                          "name": "secrets"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-backend"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-frontend"
                      }
                    ]
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Receiver",
                "metadata": {
                    "name": "github-receiver",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "events": [
                      "ping",
                      "push"
                    ],
                    "secretRef": {
                      "name": "webhook-token"
                    },
                    "resources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "HelmRepository",
                          "name": "webapp"
                      }
                    ]
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "image": "ghcr.io/stefanprodan/podinfo",
                    "interval": "1m0s"
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImagePolicy",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "imageRepositoryRef": {
                      "name": "podinfo"
                    },
                    "policy": {
                      "semver": {
                          "range": "5.0.x"
                      }
                    }
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageUpdateAutomation",
                "metadata": {
                  "name": "flux-system",
                  "namespace": "flux-system"
                },
                "spec": {
                  "interval": "1m0s",
                  "sourceRef": {
                    "kind": "GitRepository",
                    "name": "flux-system"
                  },
                  "git": {
                    "checkout": {
                      "ref": {
                        "branch": "main"
                      }
                    },
                    "commit": {
                      "author": {
                        "email": "fluxcdbot@users.noreply.github.com",
                        "name": "fluxcdbot"
                      },
                      "messageTemplate": "{{range .Updated.Images}}{{println .}}{{end}}"
                    },
                    "push": {
                      "branch": "main"
                    }
                  },
                  "update": {
                    "path": "./clusters/my-cluster",
                    "strategy": "Setters"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.connect.redhat.com/flux/source-controller@sha256:b576e3f010e3ff07e716064ab0dad042aba93f0ea5c250535dd18bd678c45337
          description: Flux is a Continuous Delivery solution for Kubernetes.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/flux/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/flux/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: flux-system
          support: The Flux community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta1
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta2
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta2
          - description: OCIRepository
            displayName: OCIRepository
            kind: OCIRepository
            name: ocirepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta2
        description: |-
          [Flux](https://fluxcd.io/) is a Continuous Delivery solution for Kubernetes.
          Flux is constructed with the [GitOps Toolkit](https://toolkit.fluxcd.io/components/), a set of composable APIs and specialized tools for keeping Kubernetes clusters in sync with sources of configuration (like Git & Helm repositories), and automating updates to configuration when there is new code to deploy.

          Flux is a CNCF incubation project made for:

          - **cluster operators** who automate provision and configuration of clusters
          - **platform engineers** who build continuous delivery for developer teams
          - **app developers** who rely on continuous delivery to get their code live

          Features:

          - Source configuration from Git and Helm repositories, and S3-compatible buckets (e.g., Minio)
          - Kustomize and Helm support
          - Event-triggered and periodic reconciliation
          - Integration with Kubernetes RBAC
          - Health assessment (clusters and workloads)
          - Dependency management (infrastructure and workloads)
          - Secrets management (builtin support for Mozilla SOPS and Cloud KMS)
          - Alerting to external systems (webhook senders)
          - External events handling (webhook receivers)
          - Automated container image updates to Git (image scanning and patching)
          - Policy-driven validation (OPA, Kyverno, admission controllers)
          - Seamless integration with Git providers (GitHub, GitLab, Bitbucket)
          - Interoperability with Cluster API (CAPI) providers

          ### Steps to support bootstraping with Flux CLI

          SecComp Profile should be removed as it requires a special privilege to get the access.
          Also, it requires to enforce `runUserAs` to the same UID provided by the images to prevent OpenShift
          to alter the user id expected by our controllers.

          To do so, please specifiy a patch using the following Kustomization file, for example.

          ```yaml
          apiVersion: kustomize.config.k8s.io/v1beta1
          kind: Kustomization
          resources:
            - gotk-components.yaml
            - gotk-sync.yaml
          patches:
            - patch: |
                apiVersion: apps/v1
                kind: Deployment
                metadata:
                  name: all
                spec:
                  template:
                    spec:
                      containers:
                        - name: manager
                          securityContext:
                            runAsUser: 65534
                            seccompProfile:
                              $patch: delete
              target:
                kind: Deployment
                labelSelector: app.kubernetes.io/part-of=flux
          ```

          Then before installing Flux, please also perform the following steps to bind Flux to the `nonroot` SecurityContextConstraint (SCC). This applies to installation via OC's OLM or via Flux CLI.

          ```sh
          # replace default "flux-system" with the project/namespace name for your flux installation
          namespace="flux-system"
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:source-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:kustomize-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:helm-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:notification-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-automation-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-reflector-controller

          ```

          To see complete steps please refer to this [documentation](https://github.com/fluxcd/website/blob/main/content/en/docs/use-cases/openshift.md).

          If you've already installed Flux, you might need to uninstall it after applying the above pre-requisite steps.

          ### Network Policies

          This version of Flux is shipped without the normally provided network Policies.
          By default all traffic is enabled, and anyone can access the source-controller
          artifacts.

          It is not recommended to run it in production as is.

          Please visit [Flux's policies](https://github.com/fluxcd/flux2/tree/main/manifests/policies)
          to deploy the needed network policies and secure your Flux setup.

          ### Breaking Changes

          - Flux 0.18.x ships with v1beta2 of the kustomize.toolkit.fluxcd.io/Kustomization CRD.
          - Flux 0.25.x is the last Flux release where Kubernetes 1.19 is supported.

          ### Flux on OpenShift Use Cases

          - [The GitOps with Flux + OpenShift](https://www.youtube.com/watch?v=8ybCtNVP5To) talk by Andrew Block (Red Hat) showing how to setup multi-tenacy on OpenShift using Flux.
          - [Flux OpenShift Use Case](https://fluxcd.io/docs/use-cases/openshift/) on FluxCD.io
        displayName: Flux
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - gitops
        - cd
        - continuous delivery
        - fluxcd
        - helm
        - kustomize
        links:
        - name: Homepage
          url: https://fluxcd.io/
        - name: Documentation
          url: https://fluxcd.io/docs/
        - name: Slack
          url: https://cloud-native.slack.com/messages/flux
        - name: Mailing List
          url: https://lists.cncf.io/g/cncf-flux-dev
        - name: Source Code
          url: https://github.com/fluxcd/flux2
        maintainers:
        - email: support@weave.works
          name: The Flux maintainers
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: The Flux project
        relatedImages:
        - registry.connect.redhat.com/flux/helm-controller@sha256:cc24a7b73b023ad86a6c3a845088d153ea290dda7cab1ec03eb85aaf814e6049
        - registry.connect.redhat.com/flux/image-automation-controller@sha256:0c51595307e14acc15f22d50b414a87513ebb03530ef4a8f7d6cd9f8b757243e
        - registry.connect.redhat.com/flux/image-reflector-controller@sha256:3bef7fde22be641304e5ca69f77150e416eba5141f8aa13f68be7f810e55aa59
        - registry.connect.redhat.com/flux/kustomize-controller@sha256:246290908d899b9f6feb920f88913bd2d4673cf852a3020714fd4152c7f78237
        - registry.connect.redhat.com/flux/notification-controller@sha256:b3d81234ba3312e0106f2f7cabc73fbe1d7bf683eda36b9ed5bbdaa43789be40
        - registry.connect.redhat.com/flux/source-controller@sha256:b576e3f010e3ff07e716064ab0dad042aba93f0ea5c250535dd18bd678c45337
        version: 2.0.1
      entries:
      - name: flux.v2.0.1
        version: 2.0.1
      name: stable
    defaultChannel: stable
    packageName: flux
    provider:
      name: The Flux project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: telemetry-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: telemetry-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "Autoscaling",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "telemetry-operator",
                    "app.kubernetes.io/instance": "autoscaling-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "autoscaling",
                    "app.kubernetes.io/part-of": "telemetry-operator"
                  },
                  "name": "autoscaling"
                },
                "spec": {
                  "aodh": {
                    "databaseAccount": "aodh",
                    "databaseInstance": "openstack",
                    "memcachedInstance": "memcached",
                    "passwordSelectors": null,
                    "secret": "osp-secret"
                  }
                }
              },
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "Ceilometer",
                "metadata": {
                  "name": "ceilometer",
                  "namespace": "openstack"
                },
                "spec": {
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "Logging",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "telemetry-operator",
                    "app.kubernetes.io/instance": "logging-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "logging",
                    "app.kubernetes.io/part-of": "telemetry-operator"
                  },
                  "name": "logging"
                },
                "spec": {
                  "cloNamespace": "openshift-logging",
                  "port": 10514
                }
              },
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "MetricStorage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "telemetry-operator",
                    "app.kubernetes.io/instance": "metricstorage-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "metricstorage",
                    "app.kubernetes.io/part-of": "telemetry-operator"
                  },
                  "name": "metricstorage-sample"
                },
                "spec": {
                  "monitoringStack": {
                    "alertingEnabled": true,
                    "scrapeInterval": "30s",
                    "storage": {
                      "persistent": {
                        "pvcStorageRequest": "20G"
                      },
                      "retention": "24h",
                      "strategy": "persistent"
                    }
                  }
                }
              },
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "Telemetry",
                "metadata": {
                  "name": "telemetry"
                },
                "spec": {
                  "autoscaling": {
                    "aodh": {
                      "databaseAccount": "aodh",
                      "databaseInstance": "openstack",
                      "memcachedInstance": "memcached",
                      "passwordSelectors": null,
                      "secret": "osp-secret"
                    },
                    "enabled": false,
                    "heatInstance": "heat"
                  },
                  "ceilometer": {
                    "enabled": true,
                    "secret": "osp-secret"
                  },
                  "logging": {
                    "enabled": false,
                    "port": 10514
                  },
                  "metricStorage": {
                    "enabled": false,
                    "monitoringStack": {
                      "alertingEnabled": true,
                      "scrapeInterval": "30s",
                      "storage": {
                        "persistent": {
                          "pvcStorageRequest": "20G"
                        },
                        "retention": "24h",
                        "strategy": "persistent"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:47Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Autoscaling is the Schema for the autoscalings API
            displayName: Autoscaling
            kind: Autoscaling
            name: autoscalings.telemetry.openstack.org
            version: v1beta1
          - description: Ceilometer is the Schema for the ceilometers API
            displayName: Ceilometer
            kind: Ceilometer
            name: ceilometers.telemetry.openstack.org
            version: v1beta1
          - description: Logging is the Schema for the loggings API
            displayName: Logging
            kind: Logging
            name: loggings.telemetry.openstack.org
            version: v1beta1
          - description: MetricStorage is the Schema for the metricstorages API
            displayName: Metric Storage
            kind: MetricStorage
            name: metricstorages.telemetry.openstack.org
            version: v1beta1
          - description: Telemetry is the Schema for the telemetry API
            displayName: Telemetry
            kind: Telemetry
            name: telemetries.telemetry.openstack.org
            version: v1beta1
        description: Telemetry Operator
        displayName: Telemetry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Telemetry
        - Ceilometer
        links:
        - name: Telemetry Operator
          url: https://github.com/openstack-k8s-operators/telemetry-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-ceilometer-notification-rhel9@sha256:239e5376f65682fd644e8d9b10b27ad879747a056c8fd26ff19882e34c5bb056
        - registry.redhat.io/rhoso/openstack-ceilometer-ipmi-rhel9@sha256:86618d8f54df09210349e382188f3b6563ab96ee511fb4da6eb6dc80e4027aae
        - registry.redhat.io/rhoso/openstack-aodh-listener-rhel9@sha256:3599a25515a96b3ff4768f85fecb4bf955789fb9b5d8abc3181678062020ae64
        - registry.redhat.io/rhoso/openstack-ceilometer-compute-rhel9@sha256:b8a0412041c50e312bef39c17a9c2600e9dbb42ea380d55e58243cda641cef4f
        - registry.redhat.io/rhoso/openstack-aodh-evaluator-rhel9@sha256:16027b7d4f07385c87f97e1581d4454d4ce7ed654499720d9ba5bc6fc74fc246
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/telemetry-rhel9-operator@sha256:83962f19193f98b87c54136f02d18797a8047f65e0d75c11ce9518e50aabc050
        - registry.redhat.io/rhoso/openstack-ceilometer-central-rhel9@sha256:1c79d1edb2404a8614ea29f24e2a869695d75b1695ea0937eebc9c57250fc4ab
        - registry.redhat.io/rhoso/openstack-aodh-api-rhel9@sha256:5ab39a468ab014f2fec4cb9cd210b430ae1c1262f52b9e0f22881d6d8e9def97
        - registry.redhat.io/rhoso/openstack-aodh-notifier-rhel9@sha256:f729548ac3d718f2f5b6c85045e7c676e1656207ee54f43540d246bd97426894
        - registry.redhat.io/rhoso-operators/sg-core-rhel9@sha256:a6e4edc24409361cb770bb44fe5b475dc58748426ccc4003abad7e70f01f1b97
        - registry.redhat.io/ubi9/httpd-24@sha256:dbfbe3383e12d63df7178e4cdd5b1e6e51a84c781041b9192026ea6edb72fbf3
        version: 1.0.6
      entries:
      - name: telemetry-operator.v1.0.6
        version: 1.0.6
      - name: telemetry-operator.v1.0.5
        version: 1.0.5
      - name: telemetry-operator.v1.0.4
        version: 1.0.4
      - name: telemetry-operator.v1.0.3
        version: 1.0.3
      - name: telemetry-operator.v1.0.2
        version: 1.0.2
      - name: telemetry-operator.v1.0.1
        version: 1.0.1
      - name: telemetry-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: telemetry-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Snyk Ltd.
      provider-url: ""
    name: snyk-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: snyk-operator.v1.92.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1alpha1",
                "kind": "SnykMonitor",
                "metadata": {
                  "name": "snyk-monitor"
                },
                "spec": {
                  "clusterName": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "snyk/kubernetes-monitor",
                    "tag": "1.92.0-ubi8"
                  },
                  "integrationApi": "",
                  "monitorSecrets": "snyk-monitor",
                  "scope": "Cluster",
                  "temporaryStorageSize": "50Gi",
                  "pvc": {
                    "enabled": false,
                    "create": false,
                    "name": "snyk-monitor-pvc",
                    "storageClassName": null
                  },
                  "initContainerImage": {
                    "repository": "busybox",
                    "tag": "latest"
                  },
                  "requests": {
                    "memory": "400Mi"
                  },
                  "limits": {
                    "memory": "2Gi"
                  },
                  "nodeAffinity": {
                    "disableBetaArchNodeSelector": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Security
          certified: "false"
          containerImage: docker.io/snyk/kubernetes-operator:1.92.0
          createdAt: "2022-06-09T17:49:36Z"
          description: A Kubernetes Operator for creating and managing Snyk controller
            instances.
          olm.skipRange: <1.68.2
          repository: https://github.com/snyk/kubernetes-monitor
          support: Snyk Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a deployment of the Snyk controller, which scans
              container images in the cluster for vulnerabilities.
            displayName: Snyk Monitor
            kind: SnykMonitor
            name: snykmonitors.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          A Kubernetes Operator for creating and managing Snyk Kubernetes controller instances.

          Snyk integrates with Kubernetes, enabling you to import and test your running workloads and identify vulnerabilities in their associated images and configurations that might make those workloads less secure. Once imported, Snyk continues to monitor those workloads, identifying additional security issues as new images are deployed and the workload configuration changes.

          ## Prerequisites

          To start using the Operator and the Snyk controller that it manages, you will need to create a secret containing your Snyk integrationId and Docker config file:

          ```
          kubectl create namespace snyk-monitor
          kubectl create secret generic snyk-monitor -n snyk-monitor --from-literal=dockercfg.json={} --from-literal=integrationId=${SNYK_INTEGRATION_ID}
          ```

          Refer to the [Snyk documentation](https://support.snyk.io/hc/en-us/articles/360003916138-Kubernetes-integration-overview) for more details.
        displayName: Snyk Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - snyk
        - controller
        - monitor
        - operator
        - security
        - cluster
        - container
        - image
        - security
        - vulnerabilities
        - advisories
        - audit
        - scan
        - scanning
        links:
        - name: Website
          url: https://snyk.io
        - name: Documentation
          url: https://support.snyk.io/hc/en-us/articles/360003916138-Kubernetes-integration-overview
        maintainers:
        - email: support@snyk.io
          name: Snyk Ltd.
        maturity: stable
        provider:
          name: Snyk Ltd.
        relatedImages:
        - docker.io/snyk/kubernetes-operator:1.92.0
        version: 1.92.0
      entries:
      - name: snyk-operator.v1.92.0
        version: 1.92.0
      - name: snyk-operator.v1.90.2
        version: 1.90.2
      - name: snyk-operator.v1.90.1
        version: 1.90.1
      - name: snyk-operator.v1.88.4
        version: 1.88.4
      - name: snyk-operator.v1.85.2
        version: 1.85.2
      - name: snyk-operator.v1.82.1
        version: 1.82.1
      - name: snyk-operator.v1.81.1
        version: 1.81.1
      - name: snyk-operator.v1.80.0
        version: 1.80.0
      - name: snyk-operator.v1.73.0
        version: 1.73.0
      - name: snyk-operator.v1.72.0
        version: 1.72.0
      - name: snyk-operator.v1.70.4
        version: 1.70.4
      - name: snyk-operator.v1.70.1
        version: 1.70.1
      name: stable
    defaultChannel: stable
    packageName: snyk-operator
    provider:
      name: Snyk Ltd.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Aqua Security, Inc.
      provider-url: ""
    name: aqua
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aqua-operator.v2022.4.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaCsp",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "namespace": "aqua",
                    "version": "2022.4",
                    "requirements": true
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry",
                    "dbDiskSize": 10,
                    "databaseSecret": {
                      "key": "db-password",
                      "name": "aqua-database-password"
                    }
                  },
                  "database": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "database",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "gateway": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "gateway",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "server": {
                    "replicas": 1,
                    "service": "LoadBalancer",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "console",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "route": true,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaDatabase",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "database",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "diskSize": 10,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "image": {
                      "repository": "enforcer",
                      "registry": "registry.aquasec.com",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "IfNotPresent"
                    }
                  },
                  "runAsNonRoot": false,
                  "gateway": {
                    "host": "aqua-gateway",
                    "port": 8443
                  },
                  "token": "<<your-token>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaKubeEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "config": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "aqua-secure",
                    "imagePullSecret": "aqua-registry",
                    "kubebench_image": "docker.io/aquasec/kube-bench:v0.7.1"
                  },
                  "deploy": {
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "tag": "<<KUBE_ENFORCER_TAG>>",
                      "repository": "kube-enforcer",
                      "pullPolicy": "Always"
                    }
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "starboard-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  }
                }
              },
              {
              "apiVersion": "operator.aquasec.com/v1alpha1",
              "kind": "AquaLightning",
              "metadata": {
                "name": "aqua"
              },
              "spec": {
                "global": {
                  "gateway_address": "aqua-gateway.aqua:8443",
                  "cluster_name": "Default-cluster-name"
                },
                "common": {
                  "imagePullSecret": "aqua-registry"
                },
                "kubeEnforcer": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                  "allowAnyVersion": null,
                  "deploy": {
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "tag": "2022.4",
                      "repository": "kube-enforcer",
                      "pullPolicy": "Always"
                    }
                  },
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "starboard-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  },
                  "env": []
                },
                "enforcer": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "deploy": {
                    "image": {
                      "repository": "enforcer",
                      "registry": "registry.aquasec.com",
                      "tag": "2022.4",
                      "pullPolicy": "IfNotPresent"
                    }
                  },
                  "token": "<<ENFORCER_GROUP_TOKEN>>",
                  "env": [],
                  "secret": {
                    "name": null,
                    "key": null
                  },
                  "runAsNonRoot": null,
                  "rhcosVersion": null
                }
              }
            },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaGateway",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "gateway",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaScanner",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "replicas": 1,
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "scanner",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "runAsNonRoot": false,
                  "login": {
                    "username": "<<YOUR AQUA USER NAME>>",
                    "password": "<<YOUR AQUA USER PASSWORD>>",
                    "host": "http://aqua-server:8080",
                    "token": ""
                  }
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaServer",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "LoadBalancer",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "console",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "runAsNonRoot": false
                }
              },
              {
                  "apiVersion": "operator.aquasec.com/v1alpha1",
                  "kind": "AquaCloudConnector",
                  "metadata": {
                    "name": "aqua"
                  },
                  "spec": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "common": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "image": {
                        "repository": "aqua-cloud-connector",
                        "registry": "registry.aquasec.com",
                        "tag": "<<IMAGE TAG>>"
                      },
                      "replicas": 1
                    },
                    "runAsNonRoot": false,
                    "gateway": {
                      "host": "aqua-gateway",
                      "port": "8443"
                    },
                    "login": {
                      "host": "aqua-gateway:8443",
                      "password": "<<YOUR AQUA USER PASSWORD>>",
                      "token": "<<YOUR AQUA CLOUD CONNECTOR TOKEN>>",
                      "username": "<<YOUR AQUA USER NAME>>"
                    },
                    "tunnels": [
                      {
                        "host": "<account>.azurecr.io",
                        "port": "443"
                      },
                      {
                        "host": "gke.gcr.io",
                        "port": "443"
                      },
                      {
                        "host": "<account>.jfrog.io",
                        "port": "443"
                      },
                      {
                        "host": "1.1.1.1",
                        "port": "443"
                      },
                      {
                        "region": "us-east1",
                        "type": "ecr"
                      }
                    ]
                  }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: docker.io/aquasec/aqua-operator:2022.4.18
          createdAt: "2020-03-25T08:00:00Z"
          description: The Aqua Security Operator runs within a Openshift cluster
            and provides a means to deploy and manage Aqua Security cluster and components.
          operators.operatorframework.io/internal-objects: '["aquastarboards.aquasecurity.github.io","configauditreports.aquasecurity.github.io",
            "clusterconfigauditreports.aquasecurity.github.io"]'
          repository: https://github.com/aquasecurity/aqua-operator
          support: Aqua Security, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aqua Security CSP Deployment with Aqua Operator
            displayName: AquaCsp
            kind: AquaCsp
            name: aquacsps.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Lightning Enforcers Deployment with Aqua Operator
            displayName: AquaLightningEnforcers
            kind: AquaLightning
            name: aqualightnings.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Database Deployment with Aqua Operator
            displayName: AquaDatabase
            kind: AquaDatabase
            name: aquadatabases.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Enforcer Deployment with Aqua Operator
            displayName: AquaEnforcer
            kind: AquaEnforcer
            name: aquaenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security KubeEnforcer Deployment with Aqua Operator
            displayName: AquaKubeEnforcer
            kind: AquaKubeEnforcer
            name: aquakubeenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Gateway Deployment with Aqua Operator
            displayName: AquaGateway
            kind: AquaGateway
            name: aquagateways.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Scanner Deployment with Aqua Operator
            displayName: AquaScanner
            kind: AquaScanner
            name: aquascanners.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Server Deployment with Aqua Operator
            displayName: AquaServer
            kind: AquaServer
            name: aquaservers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Starboard for kube-enforcer.
            displayName: AquaStarboard
            kind: AquaStarboard
            name: aquastarboards.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ClusterConfigAuditReport
            kind: ClusterConfigAuditReport
            name: clusterconfigauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ConfigAuditReport
            kind: ConfigAuditReport
            name: configauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: Aqua Security Cloud Connector Deployment with Aqua Operator
            displayName: AquaCloudConnector
            kind: AquaCloudConnector
            name: aquacloudconnector.operator.aquasec.com
            version: v1alpha1
        description: |-
          The Aqua Security Operator runs within an OpenShift cluster and provides a means to deploy and manage the Aqua Security cluster and components :
          * Server (aka “console”)
          * Database (for production environments we recommend to use an external database and not the Aqua default database)
          * Gateway
          * Enforcer (aka “agent”)
          * KubeEnforcer
          * Scanner
          * CSP (package that contains the Server, Database, and Gateway)
          * LightningEnforcer (package that contains Enforcer and Kube Enforcer)

          Use the Aqua-Operator to
          * Deploy Aqua Security components on OpenShift
          * Scale up Aqua Security components with extra replicas
          * Assign metadata tags to Aqua Security components

          ## Before You Begin Using the Operator CRDs

          Obtain access to the Aqua registry - https://www.aquasec.com/about-us/contact-us/


          ## Instructions and exmaples

          Follow instructions and exmaples
          [here](https://github.com/aquasecurity/aqua-operator/blob/2022.4.0/docs/DeployOpenShiftOperator.md)
        displayName: Aqua Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aqua-security
        - scanning
        - security
        - runtime-security
        links:
        - name: Aqua Security
          url: https://www.aquasec.com/
        - name: Aqua Operator Github
          url: https://github.com/aquasecurity/aqua-operator
        maintainers:
        - email: support@aquasec.com
          name: Aqua Support, Aqua Security
        maturity: alpha
        provider:
          name: Aqua Security, Inc.
        relatedImages:
        - docker.io/aquasec/aqua-operator:2022.4.18
        version: 2022.4.18
      entries:
      - name: aqua-operator.v2022.4.18
        version: 2022.4.18
      - name: aqua-operator.v2022.4.17
        version: 2022.4.17
      name: 2022.4.0
    defaultChannel: 2022.4.0
    packageName: aqua
    provider:
      name: Aqua Security, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: The bpfman Community
      provider-url: https://bpfman.io/
    name: bpfman-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bpfman-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfapplication"
                  },
                  "name": "bpfapplication-sample"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:v0.5.6"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "kprobe": {
                        "bpffunctionname": "kprobe_counter",
                        "func_name": "try_to_wake_up",
                        "offset": 0,
                        "retprobe": false
                      },
                      "type": "Kprobe"
                    },
                    {
                      "tracepoint": {
                        "bpffunctionname": "tracepoint_kill_recorder",
                        "names": [
                          "syscalls/sys_enter_kill"
                        ]
                      },
                      "type": "Tracepoint"
                    },
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "containernames": [
                            "bpfman",
                            "bpfman-agent"
                          ],
                          "namespace": "bpfman",
                          "pods": {
                            "matchLabels": {
                              "name": "bpfman-daemon"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfNsApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfnsapplication"
                  },
                  "name": "bpfapplication-sample",
                  "namespace": "acme"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:v0.5.6"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FentryProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fentryprogram"
                  },
                  "name": "fentry-example"
                },
                "spec": {
                  "bpffunctionname": "test_fentry",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fentry:v0.5.6"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FexitProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fexitprogram"
                  },
                  "name": "fexit-example"
                },
                "spec": {
                  "bpffunctionname": "test_fexit",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fexit:v0.5.6"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "KprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kprobeprogram"
                  },
                  "name": "kprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_kprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/kprobe:v0.5.6"
                    }
                  },
                  "func_name": "try_to_wake_up",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "offset": 0,
                  "retprobe": false
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcnsprogram"
                  },
                  "name": "tc-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcprogram"
                  },
                  "name": "tc-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:v0.5.6"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxnsprogram"
                  },
                  "name": "tcx-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "tcx_next",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxprogram"
                  },
                  "name": "tcx-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "tcx_pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:v0.5.6"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TracepointProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tracepointprogram"
                  },
                  "name": "tracepoint-example"
                },
                "spec": {
                  "bpffunctionname": "enter_openat",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tracepoint:v0.5.6"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "names": [
                    "syscalls/sys_enter_openat"
                  ],
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobensprogram"
                  },
                  "name": "uprobe-example",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobeprogram"
                  },
                  "name": "uprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:v0.5.6"
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpnsprogram"
                  },
                  "name": "xdp-ns-pass-all-nodes",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpprogram"
                  },
                  "name": "xdp-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:v0.5.6"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/bpfman/bpfman-operator:v0.5.6
          createdAt: "2025-02-18T13:44:53Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: bpfman
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "bpfman",
                "labels": {
                  "pod-security.kubernetes.io/enforce": "privileged",
                  "pod-security.kubernetes.io/audit": "privileged",
                  "pod-security.kubernetes.io/warn": "privileged",
                },
                "annotations": {
                  "openshift.io/node-selector": ""
                },
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/bpfman/bpfman
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BpfApplication is the Schema for the bpfapplications API
            displayName: Bpf Application
            kind: BpfApplication
            name: bpfapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsApplication is the Schema for the BpfNsApplications
              API
            displayName: Bpf Namespaced Application
            kind: BpfNsApplication
            name: bpfnsapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsProgram is the Schema for the BpfNsProgram API
            displayName: Bpf Namespaced Program
            kind: BpfNsProgram
            name: bpfnsprograms.bpfman.io
            version: v1alpha1
          - description: BpfProgram is the Schema for the BpfProgram API
            displayName: Bpf Program
            kind: BpfProgram
            name: bpfprograms.bpfman.io
            version: v1alpha1
          - description: FentryProgram is the Schema for the Fentryprograms API
            displayName: Fentry Program
            kind: FentryProgram
            name: fentryprograms.bpfman.io
            version: v1alpha1
          - description: FexitProgram is the Schema for the Fexitprograms API
            displayName: Fexit Program
            kind: FexitProgram
            name: fexitprograms.bpfman.io
            version: v1alpha1
          - description: KprobeProgram is the Schema for the Kprobeprograms API
            displayName: Kprobe Program
            kind: KprobeProgram
            name: kprobeprograms.bpfman.io
            version: v1alpha1
          - description: TcNsProgram is the Schema for the Tcnsprograms API
            displayName: TC Namespaced Program
            kind: TcNsProgram
            name: tcnsprograms.bpfman.io
            version: v1alpha1
          - description: TcProgram is the Schema for the Tcprograms API
            displayName: TC Program
            kind: TcProgram
            name: tcprograms.bpfman.io
            version: v1alpha1
          - description: TcxNsProgram is the Schema for the Tcxnsprograms API
            displayName: TCX Namespaced Program
            kind: TcxNsProgram
            name: tcxnsprograms.bpfman.io
            version: v1alpha1
          - description: TcxProgram is the Schema for the Tcxprograms API
            displayName: TCX Program
            kind: TcxProgram
            name: tcxprograms.bpfman.io
            version: v1alpha1
          - description: TracepointProgram is the Schema for the Tracepointprograms
              API
            displayName: Tracepoint Program
            kind: TracepointProgram
            name: tracepointprograms.bpfman.io
            version: v1alpha1
          - description: UprobeNsProgram is the Schema for the Uprobensprograms API
            displayName: Uprobe Namespaced Program
            kind: UprobeNsProgram
            name: uprobensprograms.bpfman.io
            version: v1alpha1
          - description: UprobeProgram is the Schema for the Uprobeprograms API
            displayName: Uprobe Program
            kind: UprobeProgram
            name: uprobeprograms.bpfman.io
            version: v1alpha1
          - description: XdpNsProgram is the Schema for the Xdpnsprograms API
            displayName: Xdp Namespaced Program
            kind: XdpNsProgram
            name: xdpnsprograms.bpfman.io
            version: v1alpha1
          - description: XdpProgram is the Schema for the Xdpprograms API
            displayName: Xdp Program
            kind: XdpProgram
            name: xdpprograms.bpfman.io
            version: v1alpha1
        description: "The bpfman Operator is a Kubernetes Operator for deploying [bpfman](https://bpfman.netlify.app/),
          a system daemon\nfor managing eBPF programs. It deploys bpfman itself along
          with CRDs to make deploying\neBPF programs in Kubernetes much easier.\n\n##
          Quick Start\n\nTo get bpfman up and running quickly simply click 'install'
          to deploy the bpfman-operator in the bpfman namespace via operator-hub.\n##
          Configuration\n\nThe `bpfman-config` configmap is automatically created
          in the `bpfman` namespace and used to configure the bpfman deployment.\n\nTo
          edit the config simply run\n\n```bash\nkubectl edit cm bpfman-config\n```\n\nThe
          following fields are adjustable\n\n- `bpfman.agent.image`: The image used
          for the bpfman-agent`\n- `bpfman.image`: The image used for bpfman`\n- `bpfman.log.level`:
          the log level for bpfman, currently supports `debug`, `info`, `warn`, `error`,
          and `fatal`, defaults to `info`\n- `bpfman.agent.log.level`: the log level
          for the bpfman-agent currently supports `info`, `debug`, and `trace` \n\nThe
          bpfman operator deploys eBPF programs via CRDs. The following CRDs are currently
          available, \n\n- BpfApplication\n- FentryProgram\n- FexitProgram\n - KprobeProgram\n-
          TcProgram\n- TcxProgram\n- TracepointProgram\n- UprobeProgram\n- XdpProgram\n
          - BpfNsApplication\n- TcNsProgram\n- TcxNsProgram\n- UprobeNsProgram\n-
          XdpNsProgram\n- BpfProgram\n - BpfNsProgram\n\n ## More information\n\nPlease
          checkout the [bpfman community website](https://bpfman.io/) for more information."
        displayName: Bpfman Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ebpf
        - kubernetes
        links:
        - name: bpfman website
          url: https://bpfman.io/
        maintainers:
        - email: astoycos@redhat.com
          name: Andrew Stoycos
        - email: afredette@redhat.com
          name: Andre Fredette
        maturity: alpha
        provider:
          name: The bpfman Community
          url: https://bpfman.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/bpfman/bpfman-operator:v0.5.6
        version: 0.5.6
      entries:
      - name: bpfman-operator.v0.5.6
        version: 0.5.6
      - name: bpfman-operator.v0.5.5
        version: 0.5.5
      - name: bpfman-operator.v0.5.4
        version: 0.5.4
      - name: bpfman-operator.v0.5.2
        version: 0.5.2
      - name: bpfman-operator.v0.5.1
        version: 0.5.1
      - name: bpfman-operator.v0.5.0
        version: 0.5.0
      - name: bpfman-operator.v0.4.2
        version: 0.4.2
      - name: bpfman-operator.v0.4.1
        version: 0.4.1
      name: alpha
    defaultChannel: alpha
    packageName: bpfman-operator
    provider:
      name: The bpfman Community
      url: https://bpfman.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-organizations-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-organizations-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "organizations.services.k8s.aws/v1alpha1",
                "kind": "OrganizationalUnit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/organizations-controller:1.1.1
          createdAt: "2025-09-19T19:35:20Z"
          description: AWS Organizations controller is a service controller for managing
            Organizations resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Account represents the state of an AWS organizations Account
              resource.
            displayName: Account
            kind: Account
            name: accounts.organizations.services.k8s.aws
            version: v1alpha1
          - description: OrganizationalUnit represents the state of an AWS organizations
              OrganizationalUnit resource.
            displayName: OrganizationalUnit
            kind: OrganizationalUnit
            name: organizationalunits.organizations.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Organizations resources in AWS from within your Kubernetes cluster.

          **About Amazon Organizations**

          With AWS Organizations you can perform account management activities at scale by consolidating multiple AWS accounts into a single organization. Consolidating accounts simplifies how you use other AWS services. You can leverage the multi-account management services available in AWS Organizations with select AWS services to perform tasks on all accounts that are members of your organization.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Organizations
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - organizations
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Organizations Developer Resources
          url: https://aws.amazon.com/organizations/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: organizations maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/organizations-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-organizations-controller.v1.1.1
        version: 1.1.1
      - name: ack-organizations-controller.v1.0.13
        version: 1.0.13
      - name: ack-organizations-controller.v1.0.11
        version: 1.0.11
      - name: ack-organizations-controller.v1.0.10
        version: 1.0.10
      - name: ack-organizations-controller.v1.0.9
        version: 1.0.9
      - name: ack-organizations-controller.v1.0.8
        version: 1.0.8
      - name: ack-organizations-controller.v1.0.7
        version: 1.0.7
      - name: ack-organizations-controller.v1.0.6
        version: 1.0.6
      - name: ack-organizations-controller.v1.0.5
        version: 1.0.5
      - name: ack-organizations-controller.v1.0.4
        version: 1.0.4
      - name: ack-organizations-controller.v1.0.1
        version: 1.0.1
      - name: ack-organizations-controller.v1.0.0
        version: 1.0.0
      - name: ack-organizations-controller.v0.0.10
        version: 0.0.10
      - name: ack-organizations-controller.v0.0.9
        version: 0.0.9
      - name: ack-organizations-controller.v0.0.8
        version: 0.0.8
      - name: ack-organizations-controller.v0.0.7
        version: 0.0.7
      - name: ack-organizations-controller.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: ack-organizations-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-sds-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-sds-operator.v8.0.230
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.230"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleStoragePool",
                "metadata": {
                  "name": "infoscalestoragepool-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:f4947eab2b31fccc86b5d5d0d3fd4eb1f7513fd8325e379cc259f5ccd1dfc3e4
          createdAt: "2023-10-26T20:36:59Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          olm.skipRange: <8.0.230
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleStoragePool is the Schema for the infoscalestoragepools
              API
            displayName: InfoScale Storage Pool
            kind: InfoScaleStoragePool
            name: infoscalestoragepools.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.230"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:f4947eab2b31fccc86b5d5d0d3fd4eb1f7513fd8325e379cc259f5ccd1dfc3e4
        version: 8.0.230
      entries:
      - name: infoscale-sds-operator.v8.0.230
        version: 8.0.230
      - name: infoscale-sds-operator.v8.0.220
        version: 8.0.220
      - name: infoscale-sds-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-sds-operator.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-sds-operator.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.330"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:dab501b99373c840c7e8e3ab479fa273b39f9bf8a896f3f1e4e945f4ab81f64e
          createdAt: "2024-06-24T09:03:32Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.330"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:dab501b99373c840c7e8e3ab479fa273b39f9bf8a896f3f1e4e945f4ab81f64e
        version: 8.0.330
      entries:
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-sds-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.410"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
          createdAt: "2024-07-30T09:00:57Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Arctera InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/168721626-168743109-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.410"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Arctera recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
        version: 8.0.410
      entries:
      - name: infoscale-sds-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-sds-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-sds-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-sds-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.410"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
          createdAt: "2024-07-30T09:00:57Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Arctera InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/168721626-168743109-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.410"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Arctera recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
        version: 8.0.410
      entries:
      - name: infoscale-sds-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-sds-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-sds-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-sds-operator.v8.0.210
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.210"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:26573b98743245c0b607abbe0d981abce6ce9fb1c5c0cd9fe735351b7136f4e8
          createdAt: "2022-11-26T16:36:59Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleStoragePool is the Schema for the infoscalestoragepools
              API
            displayName: InfoScale Storage Pool
            kind: InfoScaleStoragePool
            name: infoscalestoragepools.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.210"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:1667227b5111fe52ed304ab19ff9fadfc677239e56b2c93a3b89b1bfbc31354d
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:26573b98743245c0b607abbe0d981abce6ce9fb1c5c0cd9fe735351b7136f4e8
        version: 8.0.210
      entries:
      - name: infoscale-sds-operator.v8.0.210
        version: 8.0.210
      - name: infoscale-sds-operator.v8.0.201
        version: 8.0.201
      - name: infoscale-sds-operator.v8.0.200
        version: 8.0.200
      - name: infoscale-sds-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-sds-operator.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-sds-operator
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: ovn-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ovn-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ovn.openstack.org/v1beta1",
                "kind": "OVNController",
                "metadata": {
                  "name": "ovncontroller-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "ovn.openstack.org/v1beta1",
                "kind": "OVNDBCluster",
                "metadata": {
                  "name": "ovndbcluster-nb-sample"
                },
                "spec": {
                  "dbType": "NB",
                  "logLevel": "info",
                  "storageClass": "local-storage",
                  "storageRequest": "10G"
                }
              },
              {
                "apiVersion": "ovn.openstack.org/v1beta1",
                "kind": "OVNDBCluster",
                "metadata": {
                  "name": "ovndbcluster-sb-sample"
                },
                "spec": {
                  "dbType": "SB",
                  "logLevel": "info",
                  "storageClass": "local-storage",
                  "storageRequest": "10G"
                }
              },
              {
                "apiVersion": "ovn.openstack.org/v1beta1",
                "kind": "OVNNorthd",
                "metadata": {
                  "name": "ovnnorthd-sample"
                },
                "spec": {
                  "logLevel": "info"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:45Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OVNController is the Schema for the ovncontrollers API
            displayName: OVNController
            kind: OVNController
            name: ovncontrollers.ovn.openstack.org
            version: v1beta1
          - description: OVNDBCluster is the Schema for the ovndbclusters API
            displayName: OVNDBCluster
            kind: OVNDBCluster
            name: ovndbclusters.ovn.openstack.org
            version: v1beta1
          - description: OVNNorthd is the Schema for the ovnnorthds API
            displayName: OVNNorthd
            kind: OVNNorthd
            name: ovnnorthds.ovn.openstack.org
            version: v1beta1
        description: OVN Operator
        displayName: OVN Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Network
        - Ovn
        links:
        - name: Ovn Operator
          url: https://github.com/openstack-k8s-operators/ovn-operator
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/ovn-rhel9-operator@sha256:bce19011b3623c66a59efe6ea874e2403440813bdc970cd6170853a8093076f3
        - registry.redhat.io/rhoso/openstack-ovn-base-rhel9@sha256:b0d424f6ea6de8192382541a95aa56fa32b9b95e8d433c41c4a6dadbe4127fae
        - registry.redhat.io/rhoso/openstack-ovn-nb-db-server-rhel9@sha256:e71c8aac00a78e3d786332aaada3657276f69483df1e441b70589601ca038a5b
        - registry.redhat.io/rhoso/openstack-ovn-sb-db-server-rhel9@sha256:7aa2832f80846d5ac71733a1f186d39dce3654ebb661d32b8e4580d92be8e338
        - registry.redhat.io/rhoso/openstack-ovn-northd-rhel9@sha256:f1e4ce04b67d888fa950b83d1f00459c67c602397336da81de3ff71d4052a5d8
        - registry.redhat.io/rhoso/openstack-ovn-controller-rhel9@sha256:6c7827dbe25c168d28250f197d5ea6f3f60cfbab4c974267ec5ea0f8d631c3f5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: ovn-operator.v1.0.6
        version: 1.0.6
      - name: ovn-operator.v1.0.5
        version: 1.0.5
      - name: ovn-operator.v1.0.4
        version: 1.0.4
      - name: ovn-operator.v1.0.3
        version: 1.0.3
      - name: ovn-operator.v1.0.2
        version: 1.0.2
      - name: ovn-operator.v1.0.1
        version: 1.0.1
      - name: ovn-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: ovn-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: fuse-apicurito
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-apicurito.v7.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1alpha1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:8d19ad579f874d6b9a08f127c8900906913a5a7c1e69cbe4b5cb0333b8d86152
          createdAt: 2022-02-28 09:35:38 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.10.3'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:8d19ad579f874d6b9a08f127c8900906913a5a7c1e69cbe4b5cb0333b8d86152
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:8dbeeb8b04c81f41d896f71db90813de57e67493fa85389b63ac23f390785043
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:c02c218ccfbb385ea29af98409995a3d9c26cb0b4c76caf2e0779fa233b5ff8c
        version: 7.10.3
      entries:
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.10.x
    - currentCSV: fuse-apicurito.v7.11.1-0.1687357785.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:2f8d4450187839f0e0c761e5ddcf61b03f5996acfe3d75b60120e76e5b0167c1
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.11.2'
          olm.substitutesFor: fuse-apicurito.v7.11.1-0.1682509045.p
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:2f8d4450187839f0e0c761e5ddcf61b03f5996acfe3d75b60120e76e5b0167c1
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:df61aafc4023b9068a8636c35eabb0d5fda400f523b64cb1fc5d8c480272c435
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:4c493a2d690d0d7c1684c53d15c06b47d691aeffdde6dd471b6e75fed2fec6a9
        version: 7.11.1+0.1687357785.p
      entries:
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.11.x
    - currentCSV: fuse-apicurito.v7.12.1-0.1706187670.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:7ccefee63165e7eaf9a609e8a9f978a06f27deff69a318fe399a29292ad99106
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.12.1'
          olm.substitutesFor: fuse-apicurito.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:7ccefee63165e7eaf9a609e8a9f978a06f27deff69a318fe399a29292ad99106
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:4cd612283e8136839f1f48a3a220dac7c6c38adf14b61bc03a981e514baefe9b
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:bee0eeaa7898af835f80c5217059d73361daa42197fa6cbbaf416e1ba373eb54
        version: 7.12.1+0.1706187670.p
      entries:
      - name: fuse-apicurito.v7.12.1-0.1706187670.p
        version: 7.12.1+0.1706187670.p
      - name: fuse-apicurito.v7.12.1
        version: 7.12.1
      - name: fuse-apicurito.v7.12.0
        version: 7.12.0
      - name: fuse-apicurito.v7.12.0-0.1697748407.p
        version: 7.12.0+0.1697748407.p
      - name: fuse-apicurito.v7.12.0-0.1697419531.p
        version: 7.12.0+0.1697419531.p
      - name: fuse-apicurito.v7.12.0-0.1694087055.p
        version: 7.12.0+0.1694087055.p
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.12.x
    - currentCSV: fuse-apicurito.v7.13.0-0.1753785079.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:d58d345d74059242d57d5c1a75f31517769230d3014a42243fda20aa93ebd5c8
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.7.0 <7.13.0'
          olm.substitutesFor: fuse-apicurito.v7.13.0-0.1751291353.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:d58d345d74059242d57d5c1a75f31517769230d3014a42243fda20aa93ebd5c8
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:f5bf9240b28b3664bdb11b1f57f611e6194ff8643ae939b6ab3b2b176705f487
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:4e64a5930a10b76147ac128a216cd7c0346da2ff3e9f31ed86311941ed193cc1
        version: 7.13.0+0.1753785079.p
      entries:
      - name: fuse-apicurito.v7.13.0-0.1739885035.p
        version: 7.13.0+0.1739885035.p
      - name: fuse-apicurito.v7.13.0-0.1741884705.p
        version: 7.13.0+0.1741884705.p
      - name: fuse-apicurito.v7.13.0-0.1753785079.p
        version: 7.13.0+0.1753785079.p
      - name: fuse-apicurito.v7.13.0-0.1721224330.p
        version: 7.13.0+0.1721224330.p
      - name: fuse-apicurito.v7.13.0-0.1753778625.p
        version: 7.13.0+0.1753778625.p
      - name: fuse-apicurito.v7.13.0-0.1753702605.p
        version: 7.13.0+0.1753702605.p
      - name: fuse-apicurito.v7.13.0-0.1721607497.p
        version: 7.13.0+0.1721607497.p
      - name: fuse-apicurito.v7.13.0-0.1751291353.p
        version: 7.13.0+0.1751291353.p
      - name: fuse-apicurito.v7.13.0-0.1748960166.p
        version: 7.13.0+0.1748960166.p
      - name: fuse-apicurito.v7.13.0-0.1745523192.p
        version: 7.13.0+0.1745523192.p
      - name: fuse-apicurito.v7.13.0-0.1744737027.p
        version: 7.13.0+0.1744737027.p
      - name: fuse-apicurito.v7.13.0-0.1744280419.p
        version: 7.13.0+0.1744280419.p
      - name: fuse-apicurito.v7.13.0-0.1741875975.p
        version: 7.13.0+0.1741875975.p
      - name: fuse-apicurito.v7.13.0-0.1739473863.p
        version: 7.13.0+0.1739473863.p
      - name: fuse-apicurito.v7.13.0-0.1738056248.p
        version: 7.13.0+0.1738056248.p
      - name: fuse-apicurito.v7.13.0-0.1731460815.p
        version: 7.13.0+0.1731460815.p
      - name: fuse-apicurito.v7.13.0-0.1729189341.p
        version: 7.13.0+0.1729189341.p
      - name: fuse-apicurito.v7.13.0-0.1727957847.p
        version: 7.13.0+0.1727957847.p
      - name: fuse-apicurito.v7.13.0-0.1727193594.p
        version: 7.13.0+0.1727193594.p
      - name: fuse-apicurito.v7.13.0-0.1725022690.p
        version: 7.13.0+0.1725022690.p
      - name: fuse-apicurito.v7.13.0-0.1724659732.p
        version: 7.13.0+0.1724659732.p
      - name: fuse-apicurito.v7.13.0-0.1724053718.p
        version: 7.13.0+0.1724053718.p
      - name: fuse-apicurito.v7.13.0-0.1723709434.p
        version: 7.13.0+0.1723709434.p
      - name: fuse-apicurito.v7.13.0
        version: 7.13.0
      - name: fuse-apicurito.v7.13.0-0.1720707249.p
        version: 7.13.0+0.1720707249.p
      - name: fuse-apicurito.v7.13.0-0.1718285973.p
        version: 7.13.0+0.1718285973.p
      - name: fuse-apicurito.v7.13.0-0.1717672390.p
        version: 7.13.0+0.1717672390.p
      - name: fuse-apicurito.v7.13.0-0.1753778936.p
        version: 7.13.0+0.1753778936.p
      - name: fuse-apicurito.v7.13.0-0.1753701210.p
        version: 7.13.0+0.1753701210.p
      - name: fuse-apicurito.v7.12.1-0.1706187670.p
        version: 7.12.1+0.1706187670.p
      - name: fuse-apicurito.v7.12.1
        version: 7.12.1
      - name: fuse-apicurito.v7.12.0-0.1697419531.p
        version: 7.12.0+0.1697419531.p
      - name: fuse-apicurito.v7.12.0-0.1697748407.p
        version: 7.12.0+0.1697748407.p
      - name: fuse-apicurito.v7.12.0-0.1694087055.p
        version: 7.12.0+0.1694087055.p
      - name: fuse-apicurito.v7.12.0
        version: 7.12.0
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.13.x
    - currentCSV: fuse-apicurito.v7.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
                    "apiVersion": "apicur.io/v1alpha1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
            }]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:05c9676aada2d34a985a59abcf55105a3765d75b0d3764089d2b32e9ca5d367c
          createdAt: "2020-10-06 12:48:47"
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.9.3'
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito Project
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CRD for Apicurito
            displayName: Apicurito CRD
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:05c9676aada2d34a985a59abcf55105a3765d75b0d3764089d2b32e9ca5d367c
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:5535864ce393ae0ae5ee52ed0c7661e0351fb7df2f9235da70608245519530b2
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:dc39b79c360511ff999eba5b2c299db3e08c9c244a0a1f3da8c90c02695e506d
        version: 7.9.3
      entries:
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.9.x
    defaultChannel: fuse-apicurito-7.13.x
    packageName: fuse-apicurito
    provider:
      name: Red Hat
kind: List
metadata:
  resourceVersion: ""
apiVersion: v1
items:
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://github.com/che-incubator/kubernetes-image-puller-operator
    name: kubernetes-imagepuller-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubernetes-imagepuller-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "che.eclipse.org/v1alpha1",
                "kind": "KubernetesImagePuller",
                "metadata": {
                  "name": "image-puller"
                },
                "spec": {
                  "daemonsetName": "k8s-image-puller",
                  "images": "java11-maven=quay.io/eclipse/che-java11-maven:next;che-theia=quay.io/eclipse/che-theia:next"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/eclipse/kubernetes-image-puller-operator:1.1.0
          createdAt: "2024-01-15T16:23:04Z"
          description: Create and manage kubernetes-image-puller instances.
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/che-incubator/kubernetes-image-puller-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubernetesImagePuller is the Schema for the kubernetesimagepullers
              API
            displayName: Kubernetes Image Puller
            kind: KubernetesImagePuller
            name: kubernetesimagepullers.che.eclipse.org
            version: v1alpha1
        description: "## About the managed application\n\nCreate a `KubernetesImagePuller`
          custom resource to automatically configure and run an instance of the [kubernetes-image-puller.](https://github.com/che-incubator/kubernetes-image-puller)\n##
          About this Operator\n\nThe `KubernetesImagePuller` custom resource understands
          the following fields in the `spec`:\n\n1. `configMapName` - The name of
          the `ConfigMap` to create\t\n2. `daemonsetName` - The name of the `DaemonSet`
          to be create.\n3. `deploymentName` - The name of the `kubernetes-image-puller`
          `Deployment` to create.\n4. `images` - A list of key-value pairs separated
          by semicolons of images to pull.  For example: `java=quay.io/eclipse/che-java8-maven:latest;theia=quay.io/eclipse/che-theia:next`\n5.
          `cachingIntervalHours` - The amount of time, in hours, between `DaemonSet`
          health checks.\n6. `cachingMemoryRequest` - The memory request for each
          cached image when the puller is running.\n7. `cachingMemoryLimit` - The
          memory limit for each cached image when the puller is running.\n8. `nodeSelector`
          - Node selector applied to pods created by the `DaemonSet`.\n## Prerequisites
          for enabling this Operator\n\nThe operator requires an existing namespace
          to be installed in.\nCheck the [README](https://github.com/che-incubator/kubernetes-image-puller-operator#installing-the-operator)
          for more details.\n"
        displayName: Kubernetes Image Puller Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - che
        - kubernetes-image-puller
        links:
        - name: Repository
          url: https://github.com/che-incubator/kubernetes-image-puller-operator
        maintainers:
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        provider:
          name: Red Hat, Inc.
          url: https://github.com/che-incubator/kubernetes-image-puller-operator
        relatedImages:
        - quay.io/eclipse/kubernetes-image-puller-operator:1.1.0
        - quay.io/brancz/kube-rbac-proxy:v0.11.0
        version: 1.1.0
      entries:
      - name: kubernetes-imagepuller-operator.v1.1.0
        version: 1.1.0
      - name: kubernetes-imagepuller-operator.v1.0.6
        version: 1.0.6
      - name: kubernetes-imagepuller-operator.v1.0.5
        version: 1.0.5
      - name: kubernetes-imagepuller-operator.v1.0.4
        version: 1.0.4
      - name: kubernetes-imagepuller-operator.v1.0.3
        version: 1.0.3
      - name: kubernetes-imagepuller-operator.v1.0.2
        version: 1.0.2
      - name: kubernetes-imagepuller-operator.v1.0.1
        version: 1.0.1
      - name: kubernetes-imagepuller-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: kubernetes-imagepuller-operator
    provider:
      name: Red Hat, Inc.
      url: https://github.com/che-incubator/kubernetes-image-puller-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Palo Alto Networks
      provider-url: ""
    name: pcc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pcc-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "name": "pcc-console",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "ConsoleDefender",
                "metadata": {
                  "name": "pcc-consoledefender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "defenderConfig": {
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Defender",
                "metadata": {
                  "name": "pcc-defender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "defenderConfig": {
                    "clusterAddress": "twistlock-console.example.com",
                    "consoleAddress": "https://twistlock-console.example.com:8083",
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: quay.io/prismacloud/pcc-operator@sha256:b8fcfbd6c51286c874e00db1bd35523386cec406fa4050ef44c0a887730cf9b8
          createdAt: "2021-11-15"
          description: Deploy Prisma Cloud Compute for cloud-native security in your
            clusters
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator
          support: pcc-integrations@paloaltonetworks.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Install the Prisma Cloud Compute Console and Defenders
            displayName: Console and Defenders
            kind: ConsoleDefender
            name: consoledefenders.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install the Prisma Cloud Compute Console
            displayName: Console
            kind: Console
            name: consoles.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install Prisma Cloud Compute Defenders
            displayName: Defenders
            kind: Defender
            name: defenders.pcc.paloaltonetworks.com
            version: v1alpha1
        description: |
          ## Features
          - Deploy Console
              - Create initial local administrator user
              - Add license
          - Deploy Defenders
          - Support for offline deployments
          - Upgrade Console
          - Upgrade Defenders

          ## Installation
          See the [Prisma Cloud Compute Operator documentation](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator) for installation guides.

          ## Support
          This operator is available "as is," and should be seen as community-supported; however, Palo Alto Networks will continue contributing whenever possible.
          Palo Alto Networks does not provide technical support or help with using or troubleshooting the operator through normal support channels.
          We encourage you to open GitHub [issues](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/issues) and [pull requests](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/pulls) to track bugs and feature requests.

          ## Other links
          [Prisma Cloud Compute product documentation](https://docs.paloaltonetworks.com/prisma/prisma-cloud/prisma-cloud-admin-compute.html)

          [Prisma Cloud Compute API documentation](https://prisma.pan.dev/api/cloud/cwpp)
        displayName: Prisma Cloud Compute Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - palo
        - palo alto
        - palo alto networks
        - paloalto
        - twistlock
        - compute
        - prisma
        - prisma cloud
        - prisma cloud compute
        - security
        - image
        - images
        - container
        - containers
        - docker
        - podman
        - vulnerability
        - vulnerabilities
        - compliance
        - runtime
        - scan
        - scanning
        - monitor
        - monitoring
        - alert
        - alerting
        links:
        - name: Prisma Cloud
          url: https://www.paloaltonetworks.com/prisma/cloud
        maintainers:
        - email: pcc-integrations@paloaltonetworks.com
          name: Wyatt Gill
        maturity: alpha
        provider:
          name: Palo Alto Networks
        relatedImages:
        - quay.io/prismacloud/pcc-operator@sha256:b8fcfbd6c51286c874e00db1bd35523386cec406fa4050ef44c0a887730cf9b8
        version: 0.2.0
      entries:
      - name: pcc-operator.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: pcc-operator
    provider:
      name: Palo Alto Networks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: koku-metrics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: koku-metrics-operator.v4.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
          createdAt: "2025-09-25T19:04:35Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: koku-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Cost Management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Koku Metrics Operator
          ## Introduction
          The `koku-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Koku Metrics Operator (`koku-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created KokuMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The KokuMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.2.0:
          * Enabled gathering and reporting NVIDIA GPU metrics for resource optimization.
          * (Bugfix) Fixed custom CA certificate validation and configuration issues.
          * (Bugfix) Fixed issue where ROS queries were missing workload metadata for namespaces labeled only with the legacy `insights_cost_management_optimizations` label.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **Progress towards FIPS 140 Compliance:** The operator is using the Go Cryptographic Module v1.0.0 to progress toward FIPS 140 compliance. Although this module is not validated at the time of this release, the operator aims to meet stricter security standards when the module does successfully achieve FIPS validation.
          * **API Name Change:** The Custom Resource Definition (CRD) for configuring the upstream `Koku Metrics Operator` was renamed from `KokuMetricsConfig` to `CostManagementMetricsConfig`.

          ### NOTE:
            * The API name change impacts only users of the upstream (community) Koku Metrics Operator.
            * The downstream (Red Hat-supported) Cost Management Metrics Operator is not impacted by this change and users should continue using their existing configurations.

          **Important Upgrade Instructions:**

            If you are upgrading the upstream Koku Metrics Operator to version 4.0.0 or higher, you must manually migrate your configuration. The operator will no longer recognize existing `KokuMetricsConfig` resources.

            To successfully upgrade and retain your operator's configuration, complete the following steps. You can use the provided `oc` commands replacing the angle brackets (`< >`) with your specific values:

            1. Backup existing configuration (recommended):

               Before you upgrade, retrieve the details of your current `KokuMetricsConfig` instance to help you recreate it with the new API name.

               ```
               oc get kokumetricsconfig -n koku-metrics-operator <your-config-name> -o yaml > koku-metrics-config-backup.yaml
               ```

            2. Delete the previous `KokuMetricsConfig` instance (recommended):

                After the operator is upgraded to `version 4.0.0`, delete any existing `KokuMetricsConfig` resources.

                ```
                oc delete kokumetricsconfig -n koku-metrics-operator <your-config-name>
                ```

            3. Create the new `CostManagementMetricsConfig` instance:

                Using the information from your backup (or a new configuration), create a new `CostManagementMetricsConfig` instance. Adjust the spec section based on your specific needs. For more information about configurations, [refer to the configurable parameters](#configurable-parameters).

            4. Apply your new configuration:

                ```
                oc apply -f <your-costmanagementmetricsconfig-file>.yaml -n koku-metrics-operator
                ```

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created KokuMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the KokuMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `KokuMetricsConfig`, creating a new `KokuMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `koku-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: koku-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the KokuMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the KokuMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon KokuMetricsConfig creation** (This parameter will start to appear in KokuMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the KokuMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the koku-metrics-operator
          **Note** There are separate instructions for configuring the `koku-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `koku-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the koku-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `koku-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `koku-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the koku-metrics-operator for a restricted network
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `koku-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: koku-metrics-operator
                labels:
                  app: koku-metrics-operator
              spec:
                volumes:
                - name: koku-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: koku-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: koku-metrics-operator-reports
                    mountPath: /tmp/koku-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/koku-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/koku-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `koku-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the KokuMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Koku Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Koku Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
        version: 4.2.0
      entries:
      - name: koku-metrics-operator.v4.2.0
        version: 4.2.0
      - name: koku-metrics-operator.v4.1.0
        version: 4.1.0
      - name: koku-metrics-operator.v4.0.0
        version: 4.0.0
      - name: koku-metrics-operator.v3.3.2
        version: 3.3.2
      - name: koku-metrics-operator.v3.3.1
        version: 3.3.1
      - name: koku-metrics-operator.v3.3.0
        version: 3.3.0
      - name: koku-metrics-operator.v3.2.1
        version: 3.2.1
      - name: koku-metrics-operator.v3.2.0
        version: 3.2.0
      - name: koku-metrics-operator.v3.1.0
        version: 3.1.0
      - name: koku-metrics-operator.v3.0.1
        version: 3.0.1
      - name: koku-metrics-operator.v3.0.0
        version: 3.0.0
      - name: koku-metrics-operator.v2.0.0
        version: 2.0.0
      - name: koku-metrics-operator.v1.2.0
        version: 1.2.0
      - name: koku-metrics-operator.v1.1.9
        version: 1.1.9
      - name: koku-metrics-operator.v1.1.8
        version: 1.1.8
      - name: koku-metrics-operator.v1.1.7
        version: 1.1.7
      - name: koku-metrics-operator.v1.1.6
        version: 1.1.6
      - name: koku-metrics-operator.v1.1.5
        version: 1.1.5
      - name: koku-metrics-operator.v1.1.4
        version: 1.1.4
      - name: koku-metrics-operator.v1.1.3
        version: 1.1.3
      - name: koku-metrics-operator.v1.1.2
        version: 1.1.2
      - name: koku-metrics-operator.v1.1.1
        version: 1.1.1
      - name: koku-metrics-operator.v0.9.8
        version: 0.9.8
      - name: koku-metrics-operator.v0.9.7
        version: 0.9.7
      - name: koku-metrics-operator.v0.9.6
        version: 0.9.6
      - name: koku-metrics-operator.v0.9.5
        version: 0.9.5
      - name: koku-metrics-operator.v0.9.4
        version: 0.9.4
      - name: koku-metrics-operator.v0.9.3
        version: 0.9.3
      - name: koku-metrics-operator.v0.9.2
        version: 0.9.2
      - name: koku-metrics-operator.v0.9.1
        version: 0.9.1
      - name: koku-metrics-operator.v0.9.0
        version: 0.9.0
      name: alpha
    - currentCSV: koku-metrics-operator.v4.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
          createdAt: "2025-09-25T19:04:35Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: koku-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Cost Management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Koku Metrics Operator
          ## Introduction
          The `koku-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Koku Metrics Operator (`koku-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created KokuMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The KokuMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.2.0:
          * Enabled gathering and reporting NVIDIA GPU metrics for resource optimization.
          * (Bugfix) Fixed custom CA certificate validation and configuration issues.
          * (Bugfix) Fixed issue where ROS queries were missing workload metadata for namespaces labeled only with the legacy `insights_cost_management_optimizations` label.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **Progress towards FIPS 140 Compliance:** The operator is using the Go Cryptographic Module v1.0.0 to progress toward FIPS 140 compliance. Although this module is not validated at the time of this release, the operator aims to meet stricter security standards when the module does successfully achieve FIPS validation.
          * **API Name Change:** The Custom Resource Definition (CRD) for configuring the upstream `Koku Metrics Operator` was renamed from `KokuMetricsConfig` to `CostManagementMetricsConfig`.

          ### NOTE:
            * The API name change impacts only users of the upstream (community) Koku Metrics Operator.
            * The downstream (Red Hat-supported) Cost Management Metrics Operator is not impacted by this change and users should continue using their existing configurations.

          **Important Upgrade Instructions:**

            If you are upgrading the upstream Koku Metrics Operator to version 4.0.0 or higher, you must manually migrate your configuration. The operator will no longer recognize existing `KokuMetricsConfig` resources.

            To successfully upgrade and retain your operator's configuration, complete the following steps. You can use the provided `oc` commands replacing the angle brackets (`< >`) with your specific values:

            1. Backup existing configuration (recommended):

               Before you upgrade, retrieve the details of your current `KokuMetricsConfig` instance to help you recreate it with the new API name.

               ```
               oc get kokumetricsconfig -n koku-metrics-operator <your-config-name> -o yaml > koku-metrics-config-backup.yaml
               ```

            2. Delete the previous `KokuMetricsConfig` instance (recommended):

                After the operator is upgraded to `version 4.0.0`, delete any existing `KokuMetricsConfig` resources.

                ```
                oc delete kokumetricsconfig -n koku-metrics-operator <your-config-name>
                ```

            3. Create the new `CostManagementMetricsConfig` instance:

                Using the information from your backup (or a new configuration), create a new `CostManagementMetricsConfig` instance. Adjust the spec section based on your specific needs. For more information about configurations, [refer to the configurable parameters](#configurable-parameters).

            4. Apply your new configuration:

                ```
                oc apply -f <your-costmanagementmetricsconfig-file>.yaml -n koku-metrics-operator
                ```

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created KokuMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the KokuMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `KokuMetricsConfig`, creating a new `KokuMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `koku-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: koku-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the KokuMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the KokuMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon KokuMetricsConfig creation** (This parameter will start to appear in KokuMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the KokuMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the koku-metrics-operator
          **Note** There are separate instructions for configuring the `koku-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `koku-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the koku-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `koku-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `koku-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the koku-metrics-operator for a restricted network
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `koku-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: koku-metrics-operator
                labels:
                  app: koku-metrics-operator
              spec:
                volumes:
                - name: koku-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: koku-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: koku-metrics-operator-reports
                    mountPath: /tmp/koku-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/koku-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/koku-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `koku-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the KokuMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Koku Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Koku Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
        version: 4.2.0
      entries:
      - name: koku-metrics-operator.v4.2.0
        version: 4.2.0
      - name: koku-metrics-operator.v4.1.0
        version: 4.1.0
      - name: koku-metrics-operator.v4.0.0
        version: 4.0.0
      - name: koku-metrics-operator.v3.3.2
        version: 3.3.2
      - name: koku-metrics-operator.v3.3.1
        version: 3.3.1
      - name: koku-metrics-operator.v3.3.0
        version: 3.3.0
      - name: koku-metrics-operator.v3.2.1
        version: 3.2.1
      - name: koku-metrics-operator.v3.2.0
        version: 3.2.0
      - name: koku-metrics-operator.v3.1.0
        version: 3.1.0
      - name: koku-metrics-operator.v3.0.1
        version: 3.0.1
      - name: koku-metrics-operator.v3.0.0
        version: 3.0.0
      - name: koku-metrics-operator.v2.0.0
        version: 2.0.0
      - name: koku-metrics-operator.v1.2.0
        version: 1.2.0
      - name: koku-metrics-operator.v1.1.9
        version: 1.1.9
      - name: koku-metrics-operator.v1.1.8
        version: 1.1.8
      - name: koku-metrics-operator.v1.1.7
        version: 1.1.7
      - name: koku-metrics-operator.v1.1.6
        version: 1.1.6
      - name: koku-metrics-operator.v1.1.5
        version: 1.1.5
      - name: koku-metrics-operator.v1.1.4
        version: 1.1.4
      - name: koku-metrics-operator.v1.1.3
        version: 1.1.3
      - name: koku-metrics-operator.v1.1.2
        version: 1.1.2
      - name: koku-metrics-operator.v1.1.1
        version: 1.1.1
      - name: koku-metrics-operator.v0.9.8
        version: 0.9.8
      - name: koku-metrics-operator.v0.9.7
        version: 0.9.7
      - name: koku-metrics-operator.v0.9.6
        version: 0.9.6
      - name: koku-metrics-operator.v0.9.5
        version: 0.9.5
      - name: koku-metrics-operator.v0.9.4
        version: 0.9.4
      - name: koku-metrics-operator.v0.9.3
        version: 0.9.3
      - name: koku-metrics-operator.v0.9.2
        version: 0.9.2
      - name: koku-metrics-operator.v0.9.1
        version: 0.9.1
      - name: koku-metrics-operator.v0.9.0
        version: 0.9.0
      name: beta
    defaultChannel: beta
    packageName: koku-metrics-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Lenovo (Shanghai) Computer Technology Co., Ltd.
      provider-url: ""
    name: lenovo-locd-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lenovo-locd-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "locd.lenovo.com/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample",
                  "namespace": "locd-operator-system"
                },
                "spec": {
                  "affinity": {},
                  "argocd": {
                    "password": "xxxxxxxxxxx",
                    "server": "argocd.svc.cluster.local",
                    "user": "admin"
                  },
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "core": {
                    "image": "registry.connect.redhat.com/loc-devops/core@sha256:f771217b3f7849e1b63cc6abb93b8b0fc9636fd4a201e31807a15f773138c609",
                    "pullPolicy": "Always"
                  },
                  "fullnameOverride": "",
                  "global": {
                    "managementNamespace": "loc-devops",
                    "serviceEndpoint": "loc-devops.domain",
                    "serviceaccount": {
                      "annotations": {
                        "managedBy": "LOC DevOps"
                      },
                      "create": true,
                      "name": "loc-devops-admin"
                    }
                  },
                  "gui": {
                    "image": "registry.connect.redhat.com/loc-devops/gui@sha256:65142a36095188e879467038936b7b6b6ee39dcea6d42a437fa7066fb2bc0f44",
                    "pullPolicy": "Always"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "className": "nginx",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "loc-devops.domain",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "Prefix",
                            "servicePort": 8080
                          },
                          {
                            "path": "/api",
                            "pathType": "Prefix",
                            "servicePort": 8000
                          },
                          {
                            "path": "/swagger",
                            "pathType": "Prefix",
                            "servicePort": 8000
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "replicaCount": 1,
                  "resources": {},
                  "securityContext": {},
                  "service": {
                    "port": {
                      "core": 8000,
                      "gui": 8080
                    },
                    "type": "ClusterIP"
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Cluster
            name: clusters.locd.lenovo.com
            version: v1alpha1
        description: This is operator used to manage components of Lenovo Open Cloud
          DevOps
        displayName: lenovo-locd-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Application management
        links:
        - name: LOCD Operator
          url: https://locd-operator.domain
        maintainers:
        - email: xujt5@lenovo.com
          name: jintaoxu
        maturity: alpha
        provider:
          name: Lenovo (Shanghai) Computer Technology Co., Ltd.
        relatedImages:
        - registry.connect.redhat.com/loc-devops/gui@sha256:65142a36095188e879467038936b7b6b6ee39dcea6d42a437fa7066fb2bc0f44
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b5786bbbef725badf3dfcc2c2c7a86ead5ebb584c978c47aae8b9a62e241b80d
        - registry.connect.redhat.com/loc-devops/operator@sha256:4c17a3688ae3098e4b5c2c0530f49b1da8dedcc6fc1d3cd848513a1f189812a7
        - registry.connect.redhat.com/loc-devops/core@sha256:f771217b3f7849e1b63cc6abb93b8b0fc9636fd4a201e31807a15f773138c609
        version: 0.1.0
      entries:
      - name: lenovo-locd-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: lenovo-locd-operator
    provider:
      name: Lenovo (Shanghai) Computer Technology Co., Ltd.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtr-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtr-operator.1.2.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
             {
               "apiVersion": "windup.jboss.org/v1",
               "kind": "Windup",
               "metadata": {
                 "namespace": "mtr",
                 "name": "mtr",
                 "labels": {
                   "application": "windup"
                 }
               },
               "spec": {}
             }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mtr/mtr-rhel8-operator@sha256:4df7c4ad353339e8b958ffc651e811667767c8a0dcdd61f89352c926da52a513
          createdAt: "2024-09-03T17:22:23Z"
          description: Migration Toolkit for Runtimes is an analysis tool that supports
            the modernization and migration of Java applications.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          imagePullPolicy: Always
          operatorframework.io/suggested-namespace: mtr
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/windup/windup-operator/
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "Create an MTR Web Console application.  \n\nTo access the
              MTR Web Console application from the \"Developer\" perspective, please
              go to Topology and click on the \"Open URL\" icon on the web-console
              pod.  \n\nAlternatively from the \"Administrator\" perspective, please
              go to Networking->Routes and click on the Location hyperlink.  \n\nIn
              order to connect with default login credentials, please use \"migration\"
              as username and \"password\" as password.  \n\nKnown issue  \nIf you
              want to customize the MTR Web Console instance's parameters and you
              can not see any in the `Form view`, please switch to the `YAML view`
              and change them as needed.\n"
            displayName: Migration Toolkit for Runtimes
            kind: Windup
            name: windups.windup.jboss.org
            version: v1
        description: "The Migration Toolkit for Runtimes (MTR) is a web console application
          that supports large-scale Java application modernization and migration projects
          across a broad range of transformations and use cases.  \n\nIt analyzes
          application code, supports effort estimation, accelerates code migration,
          and enables users to move applications to containers.  \n\nFor more information
          please refer to the [https://developers.redhat.com/products/mtr/overview](https://developers.redhat.com/products/mtr/overview)
          page.\n"
        displayName: Migration Toolkit for Runtimes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - security
        - alerting
        - metric
        - troubleshooting
        - run-time
        - migration
        - modernization
        links:
        - name: Website
          url: https://developers.redhat.com/products/mtr/overview
        - name: GitHub
          url: https://github.com/windup/windup-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_runtimes
        maintainers:
        - email: migrate@redhat.com
          name: Migration Toolkit for Runtimes
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mtr/mtr-rhel8-operator@sha256:4df7c4ad353339e8b958ffc651e811667767c8a0dcdd61f89352c926da52a513
        - registry.redhat.io/mtr/mtr-web-container-rhel8@sha256:cb88a5453b506ca70563537ec6db2f44ff8cb4050c9652ec86eb5128aa313003
        - registry.redhat.io/mtr/mtr-web-executor-container-rhel8@sha256:2f0622774367f33efc321926196681322ab1535144e6a0785608c8050fc59f2e
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        version: 1.2.7
      entries:
      - name: mtr-operator.1.2.7
        version: 1.2.7
      - name: mtr-operator.1.2.6
        version: 1.2.6
      - name: mtr-operator.1.2.5
        version: 1.2.5
      - name: mtr-operator.1.2.4
        version: 1.2.4
      - name: mtr-operator.1.2.3
        version: 1.2.3
      - name: mtr-operator.1.2.2
        version: 1.2.2
      - name: mtr-operator.1.2.1
        version: 1.2.1
      - name: mtr-operator.1.2.0
        version: 1.2.0
      - name: mtr-operator.1.1.1
        version: 1.1.1
      - name: mtr-operator.1.1.0
        version: 1.1.0
      - name: mtr-operator.1.0.3
        version: 1.0.3
      - name: mtr-operator.1.0.2
        version: 1.0.2
      - name: mtr-operator.1.0.1
        version: 1.0.1
      - name: mtr-operator.1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: mtr-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: GitLab Inc
      provider-url: https://about.gitlab.com/
    name: gitlab-operator-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "certmanager": {
                        "install": false
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-26T07:16:35Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3281fe12aaae428fd7102332e7c0d7033515042d7830eba3953891ab0c3e3a29
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      - name: gitlab-operator-kubernetes.v2.1.1
        version: 2.1.1
      - name: gitlab-operator-kubernetes.v2.1.0
        version: 2.1.0
      - name: gitlab-operator-kubernetes.v2.0.2
        version: 2.0.2
      - name: gitlab-operator-kubernetes.v2.0.1
        version: 2.0.1
      - name: gitlab-operator-kubernetes.v2.0.0
        version: 2.0.0
      - name: gitlab-operator-kubernetes.v1.12.2
        version: 1.12.2
      - name: gitlab-operator-kubernetes.v1.12.1
        version: 1.12.1
      - name: gitlab-operator-kubernetes.v1.12.0
        version: 1.12.0
      - name: gitlab-operator-kubernetes.v1.11.3
        version: 1.11.3
      - name: gitlab-operator-kubernetes.v1.11.2
        version: 1.11.2
      - name: gitlab-operator-kubernetes.v1.11.1
        version: 1.11.1
      - name: gitlab-operator-kubernetes.v1.11.0
        version: 1.11.0
      - name: gitlab-operator-kubernetes.v1.10.2
        version: 1.10.2
      - name: gitlab-operator-kubernetes.v1.10.1
        version: 1.10.1
      - name: gitlab-operator-kubernetes.v1.10.0
        version: 1.10.0
      - name: gitlab-operator-kubernetes.v1.9.2
        version: 1.9.2
      - name: gitlab-operator-kubernetes.v1.9.1
        version: 1.9.1
      - name: gitlab-operator-kubernetes.v1.9.0
        version: 1.9.0
      - name: gitlab-operator-kubernetes.v1.8.3
        version: 1.8.3
      - name: gitlab-operator-kubernetes.v1.8.2
        version: 1.8.2
      - name: gitlab-operator-kubernetes.v1.8.1
        version: 1.8.1
      - name: gitlab-operator-kubernetes.v1.8.0
        version: 1.8.0
      - name: gitlab-operator-kubernetes.v1.7.2
        version: 1.7.2
      - name: gitlab-operator-kubernetes.v1.7.1
        version: 1.7.1
      - name: gitlab-operator-kubernetes.v1.7.0
        version: 1.7.0
      - name: gitlab-operator-kubernetes.v1.6.1
        version: 1.6.1
      - name: gitlab-operator-kubernetes.v1.6.0
        version: 1.6.0
      - name: gitlab-operator-kubernetes.v1.5.1
        version: 1.5.1
      - name: gitlab-operator-kubernetes.v1.4.2
        version: 1.4.2
      - name: gitlab-operator-kubernetes.v1.4.1
        version: 1.4.1
      - name: gitlab-operator-kubernetes.v1.4.0
        version: 1.4.0
      - name: gitlab-operator-kubernetes.v1.3.3
        version: 1.3.3
      - name: gitlab-operator-kubernetes.v1.3.2
        version: 1.3.2
      - name: gitlab-operator-kubernetes.v1.3.1
        version: 1.3.1
      - name: gitlab-operator-kubernetes.v1.3.0
        version: 1.3.0
      name: stable
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "certmanager": {
                        "install": false
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-26T07:16:35Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3281fe12aaae428fd7102332e7c0d7033515042d7830eba3953891ab0c3e3a29
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      - name: gitlab-operator-kubernetes.v2.1.1
        version: 2.1.1
      - name: gitlab-operator-kubernetes.v2.1.0
        version: 2.1.0
      - name: gitlab-operator-kubernetes.v2.0.2
        version: 2.0.2
      - name: gitlab-operator-kubernetes.v2.0.1
        version: 2.0.1
      - name: gitlab-operator-kubernetes.v2.0.0
        version: 2.0.0
      - name: gitlab-operator-kubernetes.v1.12.2
        version: 1.12.2
      - name: gitlab-operator-kubernetes.v1.12.1
        version: 1.12.1
      - name: gitlab-operator-kubernetes.v1.12.0
        version: 1.12.0
      - name: gitlab-operator-kubernetes.v1.11.3
        version: 1.11.3
      - name: gitlab-operator-kubernetes.v1.11.2
        version: 1.11.2
      - name: gitlab-operator-kubernetes.v1.11.1
        version: 1.11.1
      - name: gitlab-operator-kubernetes.v1.11.0
        version: 1.11.0
      - name: gitlab-operator-kubernetes.v1.10.2
        version: 1.10.2
      - name: gitlab-operator-kubernetes.v1.10.1
        version: 1.10.1
      - name: gitlab-operator-kubernetes.v1.10.0
        version: 1.10.0
      - name: gitlab-operator-kubernetes.v1.9.2
        version: 1.9.2
      - name: gitlab-operator-kubernetes.v1.9.1
        version: 1.9.1
      - name: gitlab-operator-kubernetes.v1.9.0
        version: 1.9.0
      - name: gitlab-operator-kubernetes.v1.8.3
        version: 1.8.3
      - name: gitlab-operator-kubernetes.v1.8.2
        version: 1.8.2
      - name: gitlab-operator-kubernetes.v1.8.1
        version: 1.8.1
      - name: gitlab-operator-kubernetes.v1.8.0
        version: 1.8.0
      - name: gitlab-operator-kubernetes.v1.7.2
        version: 1.7.2
      - name: gitlab-operator-kubernetes.v1.7.1
        version: 1.7.1
      - name: gitlab-operator-kubernetes.v1.7.0
        version: 1.7.0
      - name: gitlab-operator-kubernetes.v1.6.1
        version: 1.6.1
      - name: gitlab-operator-kubernetes.v1.6.0
        version: 1.6.0
      - name: gitlab-operator-kubernetes.v1.5.1
        version: 1.5.1
      - name: gitlab-operator-kubernetes.v1.4.2
        version: 1.4.2
      - name: gitlab-operator-kubernetes.v1.4.1
        version: 1.4.1
      - name: gitlab-operator-kubernetes.v1.4.0
        version: 1.4.0
      - name: gitlab-operator-kubernetes.v1.3.3
        version: 1.3.3
      - name: gitlab-operator-kubernetes.v1.3.2
        version: 1.3.2
      - name: gitlab-operator-kubernetes.v1.3.1
        version: 1.3.1
      - name: gitlab-operator-kubernetes.v1.3.0
        version: 1.3.0
      name: unstable
    defaultChannel: stable
    packageName: gitlab-operator-kubernetes
    provider:
      name: GitLab Inc
      url: https://about.gitlab.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Silicom Ltd.
      provider-url: http://www.silicom-usa.com
    name: silicom-sts-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: silicom-sts-operator.v0.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "namespace": "openshift-operators",
                  "nodeSelector": {
                    "sts.silicom.com/config": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1"
                },
                "spec": {
                  "aprLevel": 0,
                  "domainNumber": 24,
                  "esmcMode": 1,
                  "forwardable": 0,
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-TBC.8275.1",
                  "mode10MHz": 2,
                  "modePPS": 2,
                  "namespace": "openshift-operators",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  },
                  "phyLedsCtl": 0,
                  "priority2": 128,
                  "src10MHz": 1,
                  "srcPPS": 1,
                  "syncOption": 1,
                  "synceCpu": 10,
                  "synceRecClkPort": 0,
                  "tracePtpMsg": -1,
                  "twoStep": 0
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "worker2"
                },
                "spec": null,
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "images": null,
                  "sro": null
                }
              }
            ]
          capabilities: Basic Install
          categories: Drivers and plugins
          certified: "false"
          containerImage: quay.io/silicom/sts-operator:0.0.6
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicomdk/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicomDK/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/silicomDK/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - quay.io/silicom/sts-plugin@sha256:44339fd400d46d9dddb02c2da47c6f9f96ad2605a5c6de8a7dce2f8c3da3bb57
        - quay.io/silicom/phc2sys@sha256:0f3b63a3059b7d6158a916814281097f0d418b99a5fac9e8e5fd9bae58fb0b74
        - quay.io/silicom/tsyncd@sha256:78be78ec8939f025f9a8b899693ab77b3990116d1c7ffb2115c620b7c3382d9d
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/silicom/sts-operator:0.0.6
        - quay.io/silicom/grpc-tsyncd@sha256:aac7233423439e352d8eb39cef63e479c30c2f80b063c84ffc8797cc5a045117
        - quay.io/silicom/tsync_extts@sha256:be6a5950a5fdc6fdf8015767901856bf173d01bdc803be7141e4f9795ab1478a
        - quay.io/silicom/ice-driver-src@sha256:ca728f648769d31501cdedda92554bcfa890754405b0b06d41629cea47707da8
        - quay.io/silicom/gpsd@sha256:649e8a6c524920e1c564d8f797b8d64235b2527aa923f94bc9c7a0b7aba8ae7f
        version: 0.0.6
      entries:
      - name: silicom-sts-operator.v0.0.6
        version: 0.0.6
      - name: silicom-sts-operator.v0.0.5
        version: 0.0.5
      - name: silicom-sts-operator.v0.0.4
        version: 0.0.4
      - name: silicom-sts-operator.v0.0.3
        version: 0.0.3
      - name: silicom-sts-operator.v0.0.2
        version: 0.0.2
      - name: silicom-sts-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: silicom-sts-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "sts-silicom"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "namespace": "sts-silicom",
                  "nodeSelector": {
                    "sts.silicom.com/config": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1"
                },
                "spec": {
                  "aprLevel": 0,
                  "domainNumber": 24,
                  "esmcMode": 1,
                  "forwardable": 0,
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-TBC.8275.1",
                  "mode10MHz": 2,
                  "modePPS": 2,
                  "namespace": "sts-silicom",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  },
                  "phyLedsCtl": 0,
                  "priority2": 128,
                  "src10MHz": 1,
                  "srcPPS": 1,
                  "syncOption": 1,
                  "synceCpu": 10,
                  "synceRecClkPort": 0,
                  "tracePtpMsg": -1,
                  "twoStep": 0
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "worker2"
                },
                "spec": null,
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config"
                },
                "spec": {
                  "gpsSvcPort": 2947,
                  "grpcSvcPort": 50051,
                  "images": {
                    "gpsd": "quay.io/silicom/gpsd:3.23.1",
                    "grpcTsyncd": "quay.io/silicom/grpc-tsyncd:2.1.0.0",
                    "phcs2Sys": "quay.io/silicom/phcs2sys:3.1.1",
                    "stsPlugin": "quay.io/silicom/sts-plugin:0.0.2",
                    "ts2Phcs": "quay.io/silicom/ts2phcs:1.0.0",
                    "tsyncd": "quay.io/silicom/tsyncd:2.1.0.0"
                  },
                  "namespace": "sts-silicom",
                  "sro": {
                    "chart": {
                      "name": "ice-special-resource",
                      "repository": {
                        "name": "ice-special-resource",
                        "url": "http://ice-driver-src"
                      },
                      "version": "0.0.1"
                    },
                    "iceVersion": "1.7.16",
                    "namespace": "openshift-operators",
                    "srcImage": "quay.io/silicom/ice-driver-src:1.7.16",
                    "srcSvcPort": 3000
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Drivers and plugins
          certified: "false"
          containerImage: quay.io/silicom/sts-operator:0.0.2
          createdAt: 2021-11-17T10:21:24 CET
          description: This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of Silicom STS device
            plugins for Kubernetes.
          operators.operatorframework.io.bundle.channels.v1: stable,development
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicomdk/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicomDK/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/silicomDK/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/silicom/sts-operator:0.0.2
        version: 0.0.2
      entries:
      - name: silicom-sts-operator.v0.0.2
        version: 0.0.2
      - name: silicom-sts-operator.v0.0.1
        version: 0.0.1
      name: beta
    defaultChannel: alpha
    packageName: silicom-sts-operator
    provider:
      name: Silicom Ltd.
      url: http://www.silicom-usa.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-logging
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-logging.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Logging",
                "metadata": {
                  "name": "logging-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "elasticsearch": {
                    "storage": "16Gi",
                    "tag": "7.17.6-3747-35"
                  },
                  "environment": "openshift",
                  "fluentbit": {
                    "runOnMaster": false,
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "grafana": {
                    "dashboards": [
                      "wap",
                      "load-core"
                    ],
                    "dataSources": [
                      "wap",
                      "load-core"
                    ]
                  },
                  "image": {
                    "elasticInitImage": "rhel8/toolbox@sha256:a753c24cb8b03066d494f010ad654d23fb0559f14d7121542298da3e0d48340d",
                    "elasticInitRepository": "registry.access.redhat.com",
                    "fbImage": "fluent-bit:1.8.15-3747-20",
                    "fbRepository": "prod.harbor.keysight.digital/loadcore",
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "createClusterRoles": true,
                    "serviceAccountName": "load-core-service-account"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:33:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Logging
            name: loggings.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Logging Operator
        displayName: Keysight WAP Logging Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Logging
          url: https://keysight-wap-logging.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-logging@sha256:f1e3b3ad84cd0f96c92cdaf4609e29257035cb982a761e832dae2095e636c721
        version: 0.1.21
      entries:
      - name: keysight-wap-logging.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-logging
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hazelcast, Inc
      provider-url: ""
    name: hazelcast-platform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hazelcast-platform-operator.v5.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "cache-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "CronHotBackup",
                "metadata": {
                  "name": "cronhotbackup-sample"
                },
                "spec": {
                  "hotBackupTemplate": {
                    "metadata": {
                      "labels": {
                        "cron-hotbackup": "true"
                      }
                    },
                    "spec": {
                      "hazelcastResourceName": "hazelcast"
                    }
                  },
                  "schedule": "*/30 * * * *"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Flow",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hazelcast-platform-operator"
                  },
                  "name": "flow"
                },
                "spec": {
                  "database": {
                    "host": "flow-postgresql",
                    "secretName": "flow-db-secret"
                  },
                  "env": [
                    {
                      "name": "OPTIONS",
                      "value": "--flow.analytics.persistRemoteCallResponses=true --flow.analytics.persistResults=false --flow.config.custom.managementCenterUrl=http://34.140.160.158/mc\n"
                    }
                  ],
                  "externalConnectivity": {
                    "ingress": {
                      "annotations": {
                        "kubernetes.io/ingress.class": "nginx",
                        "nginx.ingress.kubernetes.io/enable-opentelemetry": "true",
                        "nginx.ingress.kubernetes.io/ssl-redirect": "false"
                      },
                      "hostname": "",
                      "ingressClassName": "nginx",
                      "path": "/"
                    }
                  },
                  "imagePullSecrets": [
                    {
                      "name": "hazelcast-cloud-hazelcast-connect-demo-pull-secret"
                    }
                  ],
                  "licenseKeySecretName": "flow-mc-license-key",
                  "repository": "quay.io/hazelcast_cloud/hazelcast-flow-internal",
                  "size": 3,
                  "version": "next"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast"
                },
                "spec": {
                  "clusterSize": 3,
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast-sample"
                },
                "spec": {
                  "clusterSize": 3,
                  "jet": {
                    "bucketConfig": {
                      "bucketURI": "gs://operator-user-code/jetJobs",
                      "secretName": "br-secret-gcp"
                    },
                    "enabled": true,
                    "resourceUploadEnabled": true
                  },
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HazelcastEndpoint",
                "metadata": {
                  "name": "hazelcastendpoint-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "port": 5701,
                  "type": "Discovery"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HotBackup",
                "metadata": {
                  "name": "hot-backup"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-longrun"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-longrun-2.0.0.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-1.0.2.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJobSnapshot",
                "metadata": {
                  "name": "jetjobsnapshot-sample"
                },
                "spec": {
                  "cancelJob": false,
                  "jetJobResourceName": "jet-job"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ManagementCenter",
                "metadata": {
                  "name": "managementcenter"
                },
                "spec": {
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "hazelcastClusters": [
                    {
                      "address": "hazelcast",
                      "name": "dev"
                    }
                  ],
                  "licenseKeySecretName": "hazelcast-license-key",
                  "persistence": {
                    "enabled": true,
                    "size": "10Gi"
                  },
                  "repository": "hazelcast/management-center"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Map",
                "metadata": {
                  "name": "map"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "MultiMap",
                "metadata": {
                  "name": "multimap-sample"
                },
                "spec": {
                  "backupCount": 3,
                  "binary": true,
                  "collectionType": "LIST",
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "queue-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ReplicatedMap",
                "metadata": {
                  "name": "replicatedmap-sample"
                },
                "spec": {
                  "asyncFillup": true,
                  "hazelcastResourceName": "hazelcast",
                  "inMemoryFormat": "OBJECT"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "topic-sample"
                },
                "spec": {
                  "globalOrderingEnabled": true,
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "UserCodeNamespace",
                "metadata": {
                  "name": "ucn-sample"
                },
                "spec": {
                  "bucketConfig": {
                    "bucketURI": "gs://operator-user-code/userCodeNamespace",
                    "secretName": "br-secret-gcp"
                  },
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "VectorCollection",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hazelcast-platform-operator"
                  },
                  "name": "vectorcollection-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "indexes": [
                    {
                      "dimension": 10,
                      "efConstruction": 256,
                      "maxDegree": 32,
                      "metric": "Dot",
                      "useDeduplication": false
                    }
                  ]
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanReplication",
                "metadata": {
                  "name": "wanreplication-sample"
                },
                "spec": {
                  "endpoints": "35.192.33.252",
                  "resources": [
                    {
                      "kind": "Map",
                      "name": "map-sample"
                    },
                    {
                      "kind": "Hazelcast",
                      "name": "hz-sample"
                    }
                  ],
                  "targetClusterName": "dev"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanSync",
                "metadata": {
                  "name": "wansync-sample"
                },
                "spec": {
                  "wanReplicationResourceName": "wan-replication-sample"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: docker.io/hazelcast/hazelcast-platform-operator@sha256:59328a4ea7a4e2110e0750364db792cf266f1e7231e3bb321e3de09d535dab23
          createdAt: "2025-03-13T11:59:59Z"
          description: Install Hazelcast clusters in Kubernetes environments.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hazelcast/hazelcast-platform-operator/
          support: Hazelcast, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.hazelcast.com
            version: v1alpha1
          - description: CronHotBackup is the Schema for the cronhotbackups API
            displayName: Cron Hot Backup
            kind: CronHotBackup
            name: cronhotbackups.hazelcast.com
            version: v1alpha1
          - description: Flow is the Schema for the flows API
            displayName: Flow
            kind: Flow
            name: flows.hazelcast.com
            version: v1alpha1
          - description: HazelcastEndpoint is the Schema for the hazelcastendpoints
              API
            displayName: Hazelcast Endpoint
            kind: HazelcastEndpoint
            name: hazelcastendpoints.hazelcast.com
            version: v1alpha1
          - description: Hazelcast is the Schema for the hazelcasts API
            displayName: Hazelcast
            kind: Hazelcast
            name: hazelcasts.hazelcast.com
            version: v1alpha1
          - description: HotBackup is the Schema for the hot backup API
            displayName: Hot Backup
            kind: HotBackup
            name: hotbackups.hazelcast.com
            version: v1alpha1
          - description: JetJob is the Schema for the jetjobs API
            displayName: Jet Job
            kind: JetJob
            name: jetjobs.hazelcast.com
            version: v1alpha1
          - description: JetJobSnapshot is the Schema for the jetjobsnapshots API
            displayName: Jet Job Snapshot
            kind: JetJobSnapshot
            name: jetjobsnapshots.hazelcast.com
            version: v1alpha1
          - description: ManagementCenter is the Schema for the managementcenters
              API
            displayName: Management Center
            kind: ManagementCenter
            name: managementcenters.hazelcast.com
            version: v1alpha1
          - description: Map is the Schema for the maps API
            displayName: Map
            kind: Map
            name: maps.hazelcast.com
            version: v1alpha1
          - description: MultiMap is the Schema for the multimaps API
            displayName: Multi Map
            kind: MultiMap
            name: multimaps.hazelcast.com
            version: v1alpha1
          - description: Queue is the Schema for the queues API
            displayName: Queue
            kind: Queue
            name: queues.hazelcast.com
            version: v1alpha1
          - description: ReplicatedMap is the Schema for the replicatedmaps API
            displayName: Replicated Map
            kind: ReplicatedMap
            name: replicatedmaps.hazelcast.com
            version: v1alpha1
          - description: Topic is the Schema for the topics API
            displayName: Topic
            kind: Topic
            name: topics.hazelcast.com
            version: v1alpha1
          - description: UserCodeNamespace is the Schema for the usercodenamespaces
              API
            displayName: User Code Namespace
            kind: UserCodeNamespace
            name: usercodenamespaces.hazelcast.com
            version: v1alpha1
          - description: VectorCollection is the Schema for the vectorcollections
              API
            displayName: Vector Collection
            kind: VectorCollection
            name: vectorcollections.hazelcast.com
            version: v1alpha1
          - description: WanReplication is the Schema for the wanreplications API
            displayName: Wan Replication
            kind: WanReplication
            name: wanreplications.hazelcast.com
            version: v1alpha1
          - description: WanSync is the Schema for the wansyncs API
            displayName: Wan Sync
            kind: WanSync
            name: wansyncs.hazelcast.com
            version: v1alpha1
        description: |
          Easily deploy Hazelcast clusters and Management Center into Kubernetes environments and manage their lifecycles.

          ## Before You Start

          If you are planning to create Hazelcast Platform Enterprise clusters, you need to [create a secret](https://docs.hazelcast.com/operator/latest/get-started#step-2-start-the-hazelcast-cluster) for the license. You can request a trial license key from [here](https://trialrequest.hazelcast.com).

          For Hazelcast Platform clusters, you can simply continue.

          ## Documentation

          1. [Get started](https://docs.hazelcast.com/operator/latest/get-started) with the Operator
          2. [Connect to the cluster from outside Kubernetes](https://guides.hazelcast.org/hazelcast-platform-operator-expose-externally)
          3. [Restore a Cluster from Cloud Storage with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-external-backup-restore)
          4. [Replicate Data between Two Hazelcast Clusters with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-wan-replication)

          ## Features

          Hazelcast Platform Operator supports the features below:

          * Custom resource for Hazelcast Platform (Enterprise) and Management Center
          * Observe status of Hazelcast clusters and Management Center instances
          * High Availability Mode configuration to create clusters that are resilient to node and zone failures
          * Support for TLS connections between members using self-signed certificates
          * Scale up and down Hazelcast clusters
          * Expose Hazelcast cluster to external clients ([Smart & Unisocket](https://docs.hazelcast.com/hazelcast/latest/clients/java#java-client-operation-modes))
          * Backup Hazelcast persistence data to cloud storage with the possibility of scheduling it and restoring the data accordingly
          * WAN Replication feature when you need to synchronize multiple Hazelcast clusters, which are connected by WANs
          * Full and Delta WAN Sync support
          * Tiered Storage support for Hazelcast and Map CRs
          * CP Subsystem configuration support for Hazelcast CR
          * User Code Deployment feature, which allows you to deploy custom and domain classes from cloud storages and URLs to Hazelcast members
          * The User Code Namespaces feature allows you to deploy custom and domain classes at runtime without a members restart
          * Support the configuration of advanced networking options
          * Support Multi-namespace configuration
          * ExecutorService and EntryProcessor support
          * Support several data structures like Map, Topic, MultiMap, ReplicatedMap, Queue and Cache which can be created dynamically via specific Custom Resources
          * MapStore, Near Cache and off-heap memory (HD memory and native memory) support for the Map CR
          * Native Memory support for the Cache CR
          * Support Jet configuration and Jet Job submission using the JetJob CR
          * Support for exporting the snapshots of JetJob CRs using JetJobSnapshot CR
          * Support for custom configurations using ConfigMap
        displayName: Hazelcast Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hazelcast
        - keyvalue
        - in-memory
        - database
        - caching
        links:
        - name: Documentation
          url: https://docs.hazelcast.com/operator/latest/
        - name: Source Code
          url: https://github.com/hazelcast/hazelcast-platform-operator
        maintainers:
        - email: cloudnative@hazelcast.com
          name: Hazelcast Cloud Native Team
        maturity: alpha
        minKubeVersion: 1.22.7
        provider:
          name: Hazelcast, Inc
        relatedImages:
        - docker.io/hazelcast/hazelcast-platform-operator@sha256:59328a4ea7a4e2110e0750364db792cf266f1e7231e3bb321e3de09d535dab23
        version: 5.15.0
      entries:
      - name: hazelcast-platform-operator.v5.15.0
        version: 5.15.0
      name: alpha
    defaultChannel: alpha
    packageName: hazelcast-platform-operator
    provider:
      name: Hazelcast, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: tang-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tang-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "daemons.redhat.com/v1alpha1",
                "kind": "TangServer",
                "metadata": {
                  "finalizers": [
                    "finalizer.daemons.tangserver.redhat.com"
                  ],
                  "name": "tangserver",
                  "namespace": "nbde"
                },
                "spec": {
                  "healthScript": "/usr/bin/tangd-health-check",
                  "image": "registry.redhat.io/rhel9/tang",
                  "keyPath": "/var/db/tang",
                  "persistentVolumeClaim": "tangserver-pvc",
                  "replicas": 1,
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:f43c96035f43040b3644e37119a3b35fc433f366d4b636459d1b9b02a05b5d3d
          createdAt: 2023-11-23T12:49:40
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TangServer is the Schema for the tangservers API
            displayName: Tang Server
            kind: TangServer
            name: tangservers.daemons.redhat.com
            version: v1alpha1
        description: NBDE Tang Server operator allows Tang Server deployment on OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nbde
        - tang
        links:
        - name: Tang Operator
          url: https://github.com/latchset/tang-operator
        maintainers:
        - email: sarroutb@redhat.com
          name: Sergio Arroutbi Braojos
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:97cade2c1ee468261aec5400728c8d44de387b459134aec7a4c3b5ec5a335d2c
        - registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:f43c96035f43040b3644e37119a3b35fc433f366d4b636459d1b9b02a05b5d3d
        version: 1.0.0
      entries:
      - name: tang-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: tang-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "daemons.redhat.com/v1alpha1",
                "kind": "TangServer",
                "metadata": {
                  "finalizers": [
                    "finalizer.daemons.tangserver.redhat.com"
                  ],
                  "name": "tangserver",
                  "namespace": "nbde"
                },
                "spec": {
                  "healthScript": "/usr/bin/tangd-health-check",
                  "image": "registry.redhat.io/rhel9/tang",
                  "keyPath": "/var/db/tang",
                  "persistentVolumeClaim": "tangserver-pvc",
                  "replicas": 1,
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:1cb2f794689817282ac570cacb106d75a2751a1926d89d3b83b619e02d869e5e
          createdAt: 2024-12-04T19:01:27
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TangServer is the Schema for the tangservers API
            displayName: Tang Server
            kind: TangServer
            name: tangservers.daemons.redhat.com
            version: v1alpha1
        description: NBDE Tang Server operator allows Tang Server deployment on OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nbde
        - tang
        links:
        - name: Tang Operator
          url: https://github.com/latchset/tang-operator
        maintainers:
        - email: sarroutb@redhat.com
          name: Sergio Arroutbi Braojos
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1172e150fff22c5eeab572f26961f3f53fbf896ee76d08c7503cfe2777c55458
        - registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:1cb2f794689817282ac570cacb106d75a2751a1926d89d3b83b619e02d869e5e
        version: 1.0.3
      entries:
      - name: tang-operator.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: alpha
    packageName: tang-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: www.redhat.com
    name: compliance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: compliance-operator.v0.1.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.17 <0.1.32'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          repository: https://github.com/openshift/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repo
          url: https://github.com/openshift/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/4.7/security/compliance_operator/compliance-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/compliance-rhel8-operator@sha256:ed68a0574ddeaf7d676403cd843f28cb586894f35dbb3ca5e06a5264bd42f030
        - registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c
        - registry.redhat.io/openshift4/compliance-openscap-rhel8@sha256:867c6d4205653d8a954d234cbaf470313009eb983d4f455fb45f49c4a7764422
        version: 0.1.32
      entries:
      - name: compliance-operator.v0.1.32
        version: 0.1.32
      name: "4.7"
    - currentCSV: compliance-operator.v0.1.61
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-node"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-platform"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "description": "Example of a tailoredProfile that extends OCP4 FedRAMP Moderate",
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.17 <0.1.61'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            displayName: Compliance Remediation
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            displayName: Compliance Scan
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            displayName: Compliance Suite
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            displayName: Profile Bundle
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            displayName: Scan Setting Binding
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            displayName: Tailored Profile
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repo
          url: https://github.com/openshift/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/compliance_operator/compliance-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b
        - registry.redhat.io/compliance/openshift-compliance-openscap-rhel8@sha256:83ad3113964ec48c5cf28315d923adb6edb95e45960e2756665d617dcd9b2e47
        - registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:5578e2225365ac1c15e4c189c5c7a8d14fd3fd26189414b36028f7f0879bb1dc
        version: 0.1.61
      entries:
      - name: compliance-operator.v0.1.61
        version: 0.1.61
      name: release-0.1
    - currentCSV: compliance-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-node"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-platform"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "description": "Example of a tailoredProfile that extends OCP4 FedRAMP Moderate",
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          must-gather-image: registry.redhat.io/compliance/openshift-compliance-must-gather-rhel8@sha256:e6468ccc55d72539b7fb980d0bce5c072e85879d4e9db39974d53eaab4d32b25
          olm.skipRange: '>=0.1.17 <1.7.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ComplianceAsCode/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            displayName: Compliance Remediation
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            displayName: Compliance Scan
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            displayName: Compliance Suite
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            displayName: Profile Bundle
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            displayName: Scan Setting Binding
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            displayName: Tailored Profile
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repository
          url: https://github.com/ComplianceAsCode/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/compliance_operator/co-overview.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4
        - registry.redhat.io/compliance/openshift-compliance-openscap-rhel8@sha256:d0f89d82845c6245a18e7db345ac3be92b2863d57f55f63e21652086128920c9
        - registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:8d3257bd6a020df53f8d7a400ae877ae029c78ffcd8ec9e35270ee3ff5c8e509
        - registry.redhat.io/compliance/openshift-compliance-must-gather-rhel8@sha256:e6468ccc55d72539b7fb980d0bce5c072e85879d4e9db39974d53eaab4d32b25
        version: 1.7.1
      entries:
      - name: compliance-operator.v1.7.1
        version: 1.7.1
      - name: compliance-operator.v1.7.0
        version: 1.7.0
      - name: compliance-operator.v1.6.2
        version: 1.6.2
      - name: compliance-operator.v1.6.1
        version: 1.6.1
      - name: compliance-operator.v1.6.0
        version: 1.6.0
      - name: compliance-operator.v1.5.1
        version: 1.5.1
      - name: compliance-operator.v1.5.0
        version: 1.5.0
      - name: compliance-operator.v1.4.1
        version: 1.4.1
      - name: compliance-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: compliance-operator
    provider:
      name: Red Hat Inc.
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: hyperfoil-bundle
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hyperfoil-operator.v0.26.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hyperfoil.io/v1alpha2",
                "kind": "Hyperfoil",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "hyperfoil-operator",
                    "app.kubernetes.io/instance": "hyperfoil-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hyperfoil",
                    "app.kubernetes.io/part-of": "hyperfoil-operator"
                  },
                  "name": "hyperfoil-sample"
                },
                "spec": {
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/hyperfoil/hyperfoil-operator:0.26.0
          createdAt: "2024-06-11T17:39:20Z"
          description: Microservice-oriented distributed benchmark framework.
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Hyperfoil/hyperfoil-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hyperfoil is the Schema for the hyperfoils API
            displayName: Hyperfoil
            kind: Hyperfoil
            name: hyperfoils.hyperfoil.io
            version: v1alpha2
        description: |
          Hyperfoil is a modern benchmark framework oriented promising
          accurate results with more flexibility and distributed architecture.

          See more on [hyperfoil.io](http://hyperfoil.io).

          Hyperfoil is licensed under Apache License 2.0
        displayName: Hyperfoil
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hyperfoil
        - benchmark
        - test
        - load
        - driver
        links:
        - name: Website
          url: http://hyperfoil.io
        - name: GitHub
          url: http://github.com/Hyperfoil
        maintainers:
        - email: rvansa@redhat.com
          name: Radim Vansa
        - email: spederse@redhat.com
          name: Ståle W. Pedersen
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/hyperfoil/hyperfoil-operator:0.26.0
        version: 0.26.0
      entries:
      - name: hyperfoil-operator.v0.26.0
        version: 0.26.0
      - name: hyperfoil-operator.v0.24.2
        version: 0.24.2
      - name: hyperfoil-operator.v0.21.0
        version: 0.21.0
      - name: hyperfoil-operator.v0.15.1
        version: 0.15.1
      - name: hyperfoil-operator.v0.15.0
        version: 0.15.0
      name: alpha
    defaultChannel: alpha
    packageName: hyperfoil-bundle
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Antline-Soft
      provider-url: ""
    name: cxl-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cxl-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-05-08T08:36:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
        description: memory extend operator
        displayName: extend-community-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - memory
        - extend
        - community
        - operator
        links:
        - name: Cxl Operator
          url: https://cxl-operator.domain
        maintainers:
        - email: tomboy@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        provider:
          name: Antline-Soft
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - rasen708/cxl-operator:latest
        version: 0.0.1
      entries:
      - name: cxl-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cxl-operator
    provider:
      name: Antline-Soft
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-msg-broker
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-msg-broker.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Messagebroker",
                "metadata": {
                  "name": "messagebroker-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "chiselServer": {
                    "enabled": false,
                    "port": 8080
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": false
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "natsCore": {
                    "reloader": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats-server-config-reloader",
                      "tag": "0.8.3-3747-30"
                    },
                    "server": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats",
                      "tag": "2.9.0-3747-21"
                    }
                  },
                  "natsHttpProxy": {
                    "enabled": false,
                    "port": 8080
                  },
                  "natsStreaming": {
                    "image": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats-streaming",
                      "tag": "0.23.2-3747-31"
                    }
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "tagSuffix": "-ocp"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:31:32Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Messagebroker
            name: messagebrokers.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Message Broker Operator
        displayName: Keysight WAP Message Broker Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Msg Broker
          url: https://keysight-wap-msg-broker.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-msg-broker@sha256:4701b6e456b099697bf1e25aeb4bc8e32b1892bb0956e36d84f9f80666b3a1fc
        version: 0.1.21
      entries:
      - name: keysight-wap-msg-broker.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-msg-broker
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC
      provider-url: https://www.nec.com
    name: sanstoragecsi-operator-bundle
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sanstoragecsi-operator-bundle.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.nec.com/v1",
                "kind": "SANStorageCSI",
                "metadata": {
                  "name": "sanstoragecsi"
                },
                "spec": {
                  "driverName": "san.csi.nec.com",
                  "images": {
                    "csiDriver": "",
                    "externalAttacher": "registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:bd8262cc4b6c405e175c39623323dd98a74bbf70e6aa90851a72612056c184ff",
                    "externalProvisioner": "registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:85f1cc8f21568e7885e492b29660f29e09796fb23ebdf5fd526aa2e99d6b2fa8",
                    "livenessProbe": "registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:1a0d15012ba70955af754b03a7488b7c6ace813646ec7c3ce845f80597050c3c",
                    "nodeRegistrar": "registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:dc7f0992be6c2853dc0e798afda368e36005a94ad76b21fa2f09790621318864"
                  },
                  "parameters": {
                    "controller": {
                      "logLevel": 5
                    },
                    "node": {
                      "livenessProbePort": 9808,
                      "logLevel": 5,
                      "maxVolumesPerNode": 1024
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          createdAt: "2022-06-15T00:00:00Z"
          description: An operator for managing the NEC Storage M Series CSI Driver
          operatorframework.io/suggested-namespace: nec-san-csi-operator-system
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: NEC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SANStorageCSI is the Schema for the sanstoragecsis API
            displayName: NEC Storage M Series CSI Driver
            kind: SANStorageCSI
            name: sanstoragecsis.csi.nec.com
            version: v1
        description: |
          This operator deploys the NEC Storage M Series CSI Driver on OpenShift Container Platform.

          NEC Storage M Series CSI Driver provides persistent storage for stateful applications using NEC Storage M Series.
          Please refer to the documentation of the NEC Storage M Series CSI Driver for detail information about supported NEC Storage Platforms and supported CSI features.
        displayName: NEC Storage M Series CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - NEC
        - M Series
        - SAN Storage
        - CSI Driver
        links:
        - name: NEC Storage M Series
          url: https://www.nec.com/en/global/prod/storage/product/san/index.html
        maturity: stable
        provider:
          name: NEC
          url: https://www.nec.com
        relatedImages:
        - registry.connect.redhat.com/nec/sanstoragecsi-operator@sha256:6656c2b5036bdc013712ecb640bbbebe0e3e90cf625c73697dbb49e3e201e2ac
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:dc7f0992be6c2853dc0e798afda368e36005a94ad76b21fa2f09790621318864
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:bd8262cc4b6c405e175c39623323dd98a74bbf70e6aa90851a72612056c184ff
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:85f1cc8f21568e7885e492b29660f29e09796fb23ebdf5fd526aa2e99d6b2fa8
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:1a0d15012ba70955af754b03a7488b7c6ace813646ec7c3ce845f80597050c3c
        version: 1.1.0
      entries:
      - name: sanstoragecsi-operator-bundle.v1.1.0
        version: 1.1.0
      - name: sanstoragecsi-operator-bundle.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: sanstoragecsi-operator-bundle
    provider:
      name: NEC
      url: https://www.nec.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nxiq-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nxiq-operator-certified.v1.193.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusIQ",
                "metadata": {
                  "name": "example-nexusiq"
                },
                "spec": {
                  "deployment": {
                    "postStart": {
                      "command": null
                    }
                  },
                  "deploymentStrategy": {
                    "type": "Recreate"
                  },
                  "fullnameOverride": "",
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [],
                    "path": "/",
                    "tls": []
                  },
                  "iq": {
                    "adminPort": 8071,
                    "applicationPort": 8070,
                    "env": [
                      {
                        "name": "JAVA_OPTS",
                        "value": "-Djava.util.prefs.userRoot=$(SONATYPE_WORK)/javaprefs -Djavax.net.ssl.trustStore=/etc/nexus-iq-server/jks/keystore.jks"
                      }
                    ],
                    "configYaml": {
                      "sonatypeWork": "/sonatype-work",
                      "createSampleData": true,
                      "server": {
                        "adminConnectors": [
                          {
                            "port": 8071,
                            "type": "http"
                          }
                        ],
                        "applicationConnectors": [
                          {
                            "port": 8070,
                            "type": "http"
                          }
                        ],
                        "requestLog": {
                          "appenders": [
                            {
                              "type": "file",
                              "currentLogFilename": "/var/log/nexus-iq-server/request.log",
                              "logFormat": "%clientHost %l %user [%date] \"%requestURL\" %statusCode %bytesSent %elapsedTime \"%header{User-Agent}\"",
                              "archivedLogFilenamePattern": "/var/log/nexus-iq-server/request-%d.log.gz",
                              "archivedFileCount": 50
                            }
                          ]
                        }
                      },
                      "logging": {
                        "level": "DEBUG",
                        "loggers": {
                          "com.sonatype.insight.scan": "INFO",
                          "eu.medsea.mimeutil.MimeUtil2": "INFO",
                          "org.apache.http": "INFO",
                          "org.apache.http.wire": "ERROR",
                          "org.eclipse.birt.report.engine.layout.pdf.font.FontConfigReader": "WARN",
                          "org.eclipse.jetty": "INFO",
                          "org.apache.shiro.web.filter.authc.BasicHttpAuthenticationFilter": "INFO",
                          "com.sonatype.insight.audit": {
                            "appenders": [
                              {
                                "type": "file",
                                "currentLogFilename": "/var/log/nexus-iq-server/audit.log",
                                "archivedLogFilenamePattern": "/var/log/nexus-iq-server/audit-%d.log.gz",
                                "archivedFileCount": 50
                              }
                            ]
                          },
                          "com.sonatype.insight.policy.violation": {
                            "appenders": [
                              {
                                "type": "console",
                                "threshold": "INFO"
                              }
                            ]
                          }
                        },
                        "appenders": [
                          {
                            "type": "console",
                            "threshold": "INFO",
                            "logFormat": "%d{'yyyy-MM-dd HH:mm:ss,SSSZ'} %level [%thread] %X{username} %logger - %msg%n"
                          },
                          {
                            "type": "file",
                            "threshold": "ALL",
                            "logFormat": "%d{'yyyy-MM-dd HH:mm:ss,SSSZ'} %level [%thread] %X{username} %logger - %msg%n",
                            "currentLogFilename": "/var/log/nexus-iq-server/clm-server.log",
                            "archivedLogFilenamePattern": "/var/log/nexus-iq-server/clm-server-%d.log.gz",
                            "archivedFileCount": 50
                          }
                        ]
                      }
                    },
                    "imageName": "registry.connect.redhat.com/sonatype/nexus-iq-server@sha256:d65df6bdff1f32bf32a87e5cb38f5e9f98b75b81e9acccbffbd967e1f71691f0",
                    "imagePullPolicy": "IfNotPresent",
                    "imagePullSecret": "",
                    "licenseSecret": "",
                    "memory": "1Gi",
                    "name": "nxiq",
                    "certificates": {}
                  },
                  "nameOverride": "",
                  "persistence": {
                    "accessMode": "ReadWriteOnce",
                    "storageSize": "1Gi"
                  },
                  "service": {
                    "annotations": {},
                    "enabled": false,
                    "labels": {},
                    "ports": [
                      {
                        "name": "nexus-service",
                        "port": 80,
                        "targetPort": 80
                      }
                    ]
                  },
                  "serviceAccount": {
                    "create": true,
                    "name": null
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nxiq-operator-certified@sha256:59c52729189da87d9b5c3de8ef85326662d94c99835ab99d6e8d7fefdcf53ef0
          createdAt: "2020-07-20"
          description: |-
            Nexus Lifecycle is an open source governance platform that continuously identifies risk,
            automatically enforces policy, and provides visibility throughout the entire SDLC.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.193.0-1
          repository: https://github.com/sonatype/operator-iq
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus IQ Server
            displayName: NexusIQ
            kind: NexusIQ
            name: nexusiqs.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Lifecycle is an open source governance platform that continuously identifies risk,
          automatically enforces policy, and provides visibility throughout the entire SDLC.
          Developers leveraging open source require a solution that helps them make safer choices
          while still delivering at DevOps speed.
          Nexus Lifecycle empowers developers and security teams to choose the best components
          and govern open source usage, ensuring organizations continue to innovate with less risk.

          ## Core Capabilities

          * **Dependency Management**:
            Generate a precise software bill of materials for all applications
            to identify open source components and gather contextual direct
            and transitive dependency information for faster action.
          * **One-Click Remediation**:
            Upgrade to the best components and versions based on real-time intelligence.
            Available in your preferred IDE or a GitHub pull request.
          * **Automated Open Source Policy Enforcement**:
            Create custom security, license, and architectural policies based
            on application type or organization and contextually enforce those policies
            across every stage of the SDLC.
          * **Continuous Monitoring**:
            Real-time notifications of newly discovered defects,
            as well as vulnerabilities based on component,
            risk level or application/container affected.

          ## Extras

          * IDE Integrations: Eclipse, IntelliJ, and Visual Studio
          * Source Control Integrations: GitHub, GitLab, and Bitbucket
          * Chrome Browser Extension
          * Red Hat Clair for Container Scanning

          ## Limitations

          This operator will be released on a quarterly basis.

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus IQ Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage

          Once the server instance is created by the operator and running,
          you'll want to expose the service as you see fit:
          1. Create a Route to the new service for iq.applicationPort (8070).
          2. Visit the URL provided by the Route, login, and set new credentials.
            The default credentials are `admin`/`admin123`.

          The Nexus IQ Server can be further configured via the NexusIQ custom resource definition:

          | Parameter            | Description                                                  | Default           |
          | -------------------- | ------------------------------------------------------------ | ----------------- |
          | `iq.applicationPort` | Port of the application connector. Must match the value in the `configYaml` property | `8070`            |
          | `iq.adminPort`       | Port of the application connector. Must match the value in the `configYaml` property | `8071`            |
          | `iq.memory`          | The amount of RAM to allocate                                | `1Gi`             |
          | `iq.env`             | Customize the server environment, including `JAVA_OPTS`      | See example YAML show when ceating a NexusIQ. |
          | `iq.licenseSecret`   | The base-64 encoded license file to be installed at startup  | `""`              |
          | `iq.configYaml`      | A YAML block which will be used as a configuration block for IQ Server. | See example YAML shown when creating a NexusIQ. |
          | `iq.certificates`   | Add your own certificates for your own trusted Certificate Authorities with names and plain armored text. | `{}` |
          | `ingress.enabled`                           | Create an ingress for Nexus         | `true`                                  |
          | `ingress.annotations`                       | Annotations to enhance ingress configuration  | `{}`                          |
          | `ingress.tls.enabled`                       | Enable TLS                          | `true`                                 |
          | `ingress.tls.secretName`                    | Name of the secret storing TLS cert, `false` to use the Ingress' default certificate | `nexus-tls`                             |
          | `ingress.path`                              | Path for ingress rules. GCP users should set to `/*` | `/`                    |
          | `deploymentStrategy.type`                   | Deployment Strategy                 |  `Recreate` |
          | `initContainers`                            | Containers to run before starting the server | `nil` |
          | `deployment.postStart.command`              | Command to run after starting the IQ Server container  | `nil`                    |
          | `deployment.terminationGracePeriodSeconds` | Time to allow for clean shutdown                        | 120                      |
          | `persistence.storageClass` | The provisioner class                        | `-` (disables dynamic provisioning)            |
          | `persistence.storageSize` | The amount of drive space to allocate                        | `1Gi`             |
          | `persistence.accessMode` | Default access mode                        | `ReadWriteOnce`             |
          | `persistence.volumeConfiguration` | A YAML block to configure the persistent volume type. Defaults to `hostPath` which should not be used in production | `hostPath`             |

          ## Configuring IQ Server

          You can define the `config.yml` for IQ Server in your CRD on startup.
          It is the `iq.configYaml` property. For more details, see the [Configuring IQ Server](https://help.sonatype.com/iqserver/configuring) help page.

          ## Installing the License

          The license file can be installed via the UI when IQ server is running, or it can be done as a part of the deploy.
          If you leave the `licenseFile` field empty/commented, IQ Server will start and prompt you to manually install the license
          when you first enter the GUI.

          ### Installing the License Automatically
          To do it automatically, first encode your `.lic` file in Base 64 with no line breaks, eg:

          ```bash
          base64 --wrap=0 mylicense.lic > lic.base64
          ```

          Then add this value to your CRD file as `iq.licenseSecret`, eg:

          ```yaml
          iq:
            licenseSecret: bXkgc2FtcGxlIGxpY2Vuc2U=
          ```

          Specify the `licenseFile` path in `iq.configYaml` as:

          ```yaml
          iq:
            configYaml:
              sonatypeWork: /sonatype-work
              createSampleData: true
              # add this line and the `licenseSecret` above to autoconfigure licensing
              licenseFile: /etc/nexus-iq-license/license_lic
              server:
                applicationConnectors:
                  - type: http
                    port: 8070
                adminConnectors:
                  - type: http
                    port: 8071
          ```

          ## Adding Certificates for Your Own Trusted Certificate Authorities

          Specify a set of certificates to be imported into the Java trust store
          of the running IQ Server. More than one certificate can be provided.

          ```yaml
          iq:
            certificates:
              my-company-cert: |
                -----BEGIN CERTIFICATE-----
                ...
                -----END CERTIFICATE-----
          ```

          ## Running Commands Before Initialization of the Server

          If you need custom initialization, you can use standard Kubernetes
          [`initContainers`](https://kubernetes.io/docs/concepts/workloads/pods/init-containers/)
          to run any series of containers and commands you need.

          ```yaml
          initContainers:
            - name: an-init-container
              image: busybox:1.28
              command: ['/bin/sh', '-c', 'echo "initializing"']
          ```
        displayName: Nexus IQ Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - sonatype
        - licensing
        - policy
        - components
        links:
        - name: Nexus Lifecycle
          url: https://www.sonatype.com/product-nexus-lifecycle
        maintainers:
        - email: support@sonaype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nxiq-operator-certified@sha256:59c52729189da87d9b5c3de8ef85326662d94c99835ab99d6e8d7fefdcf53ef0
        - registry.connect.redhat.com/sonatype/nexus-iq-server@sha256:d65df6bdff1f32bf32a87e5cb38f5e9f98b75b81e9acccbffbd967e1f71691f0
        version: 1.193.0-1
      entries:
      - name: nxiq-operator-certified.v1.193.0-1
        version: 1.193.0-1
      name: stable
    defaultChannel: stable
    packageName: nxiq-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com/
    name: openstack-baremetal-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openstack-baremetal-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "baremetal.openstack.org/v1beta1",
                "kind": "OpenStackBaremetalSet",
                "metadata": {
                  "name": "openstackbaremetalset"
                },
                "spec": {
                  "baremetalHosts": {
                    "compute-0": "172.22.0.100/24",
                    "compute-1": "172.22.0.101/24"
                  },
                  "ctlplaneGateway": "172.22.0.3",
                  "ctlplaneInterface": "enp1s0",
                  "deploymentSSHSecret": "baremetalset-deploy-secret",
                  "domainName": "osptest.openstack.org",
                  "osImage": "edpm-hardened-uefi.qcow2",
                  "passwordSecret": "baremetalset-password-secret"
                }
              },
              {
                "apiVersion": "baremetal.openstack.org/v1beta1",
                "kind": "OpenStackProvisionServer",
                "metadata": {
                  "name": "openstackprovisionserver"
                },
                "spec": {
                  "interface": "enp1s0",
                  "osImage": "edpm-hardened-uefi.qcow2",
                  "port": 8080
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:07Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenStackBaremetalSet is the Schema for the openstackbaremetalsets
              API
            displayName: OpenStack BaremetalSet
            kind: OpenStackBaremetalSet
            name: openstackbaremetalsets.baremetal.openstack.org
            version: v1beta1
          - description: OpenStackProvisionServer used to serve custom images for
              baremetal provisioning with Metal3
            displayName: OpenStackProvisionServer
            kind: OpenStackProvisionServer
            name: openstackprovisionservers.baremetal.openstack.org
            version: v1beta1
        description: An operator for handling baremetal provisioning concerns for
          OpenStack
        displayName: OpenStack Baremetal Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cn-openstack
        - openstack
        links:
        - name: Openstack Baremetal Operator
          url: https://github.com/openstack-k8s-operators/openstack-baremetal-operator
        maturity: beta
        provider:
          name: Red Hat
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/ubi9/httpd-24@sha256:dbfbe3383e12d63df7178e4cdd5b1e6e51a84c781041b9192026ea6edb72fbf3
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/openstack-baremetal-rhel9-operator@sha256:a6c69c03e59faf50973d07ccea32b80cb9c46ceb60ed00ffe7311ca42d501296
        - registry.redhat.io/rhoso/edpm-hardened-uefi-rhel9@sha256:e4bfacb9d8de056a251bee16a66b69631b96fb19fb926d8e33acd379d40a92d6
        - registry.redhat.io/rhoso-operators/openstack-baremetal-agent-rhel9@sha256:8f6f8e99042710819aa639eae6513f36c4642190ecd85d58a2ed819f1b55e38f
        version: 1.0.6
      entries:
      - name: openstack-baremetal-operator.v1.0.6
        version: 1.0.6
      - name: openstack-baremetal-operator.v1.0.5
        version: 1.0.5
      - name: openstack-baremetal-operator.v1.0.4
        version: 1.0.4
      - name: openstack-baremetal-operator.v1.0.3
        version: 1.0.3
      - name: openstack-baremetal-operator.v1.0.2
        version: 1.0.2
      - name: openstack-baremetal-operator.v1.0.1
        version: 1.0.1
      - name: openstack-baremetal-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: openstack-baremetal-operator
    provider:
      name: Red Hat
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecr-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecr-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.services.k8s.aws/v1alpha1",
                "kind": "Repository",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecr-controller:1.3.1
          createdAt: "2025-09-20T01:21:24Z"
          description: AWS ECR controller is a service controller for managing ECR
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PullThroughCacheRule represents the state of an AWS ecr PullThroughCacheRule
              resource.
            displayName: PullThroughCacheRule
            kind: PullThroughCacheRule
            name: pullthroughcacherules.ecr.services.k8s.aws
            version: v1alpha1
          - description: Repository represents the state of an AWS ecr Repository
              resource.
            displayName: Repository
            kind: Repository
            name: repositories.ecr.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Container Registry (ECR) resources in AWS from within your Kubernetes cluster.

          **About Amazon ECR**

          Amazon Elastic Container Registry (Amazon ECR) is an AWS managed container image registry service that is secure, scalable, and reliable. Amazon ECR supports private repositories with resource-based permissions using AWS IAM. This is so that specified users or Amazon EC2 instances can access your container repositories and images. You can use your preferred CLI to push, pull, and manage Docker images, Open Container Initiative (OCI) images, and OCI compatible artifacts.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECR
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECR Developer Resources
          url: https://aws.amazon.com/ecr/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecr maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecr-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-ecr-controller.v1.3.1
        version: 1.3.1
      - name: ack-ecr-controller.v1.3.0
        version: 1.3.0
      - name: ack-ecr-controller.v1.2.0
        version: 1.2.0
      - name: ack-ecr-controller.v1.1.0
        version: 1.1.0
      - name: ack-ecr-controller.v1.0.33
        version: 1.0.33
      - name: ack-ecr-controller.v1.0.32
        version: 1.0.32
      - name: ack-ecr-controller.v1.0.31
        version: 1.0.31
      - name: ack-ecr-controller.v1.0.30
        version: 1.0.30
      - name: ack-ecr-controller.v1.0.29
        version: 1.0.29
      - name: ack-ecr-controller.v1.0.28
        version: 1.0.28
      - name: ack-ecr-controller.v1.0.27
        version: 1.0.27
      - name: ack-ecr-controller.v1.0.25
        version: 1.0.25
      - name: ack-ecr-controller.v1.0.24
        version: 1.0.24
      - name: ack-ecr-controller.v1.0.23
        version: 1.0.23
      - name: ack-ecr-controller.v1.0.22
        version: 1.0.22
      - name: ack-ecr-controller.v1.0.21
        version: 1.0.21
      - name: ack-ecr-controller.v1.0.20
        version: 1.0.20
      - name: ack-ecr-controller.v1.0.19
        version: 1.0.19
      - name: ack-ecr-controller.v1.0.17
        version: 1.0.17
      - name: ack-ecr-controller.v1.0.16
        version: 1.0.16
      - name: ack-ecr-controller.v1.0.15
        version: 1.0.15
      - name: ack-ecr-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecr-controller.v1.0.12
        version: 1.0.12
      - name: ack-ecr-controller.v1.0.10
        version: 1.0.10
      - name: ack-ecr-controller.v1.0.9
        version: 1.0.9
      - name: ack-ecr-controller.v1.0.8
        version: 1.0.8
      - name: ack-ecr-controller.v1.0.7
        version: 1.0.7
      - name: ack-ecr-controller.v1.0.6
        version: 1.0.6
      - name: ack-ecr-controller.v1.0.4
        version: 1.0.4
      - name: ack-ecr-controller.v1.0.3
        version: 1.0.3
      - name: ack-ecr-controller.v1.0.2
        version: 1.0.2
      - name: ack-ecr-controller.v1.0.0
        version: 1.0.0
      - name: ack-ecr-controller.v0.1.9
        version: 0.1.9
      - name: ack-ecr-controller.v0.1.7
        version: 0.1.7
      - name: ack-ecr-controller.v0.1.6
        version: 0.1.6
      - name: ack-ecr-controller.v0.1.5
        version: 0.1.5
      - name: ack-ecr-controller.v0.1.4
        version: 0.1.4
      - name: ack-ecr-controller.v0.1.3
        version: 0.1.3
      - name: ack-ecr-controller.v0.1.2
        version: 0.1.2
      - name: ack-ecr-controller.v0.1.1
        version: 0.1.1
      - name: ack-ecr-controller.v0.1.0
        version: 0.1.0
      - name: ack-ecr-controller.v0.0.27
        version: 0.0.27
      - name: ack-ecr-controller.v0.0.26
        version: 0.0.26
      - name: ack-ecr-controller.v0.0.25
        version: 0.0.25
      - name: ack-ecr-controller.v0.0.24
        version: 0.0.24
      - name: ack-ecr-controller.v0.0.23
        version: 0.0.23
      - name: ack-ecr-controller.v0.0.22
        version: 0.0.22
      - name: ack-ecr-controller.v0.0.21
        version: 0.0.21
      - name: ack-ecr-controller.v0.0.20
        version: 0.0.20
      - name: ack-ecr-controller.v0.0.19
        version: 0.0.19
      - name: ack-ecr-controller.v0.0.18
        version: 0.0.18
      - name: ack-ecr-controller.v0.0.17
        version: 0.0.17
      - name: ack-ecr-controller.v0.0.16
        version: 0.0.16
      - name: ack-ecr-controller.v0.0.15
        version: 0.0.15
      - name: ack-ecr-controller.v0.0.14
        version: 0.0.14
      - name: ack-ecr-controller.v0.0.13
        version: 0.0.13
      - name: ack-ecr-controller.v0.0.12
        version: 0.0.12
      - name: ack-ecr-controller.v0.0.11
        version: 0.0.11
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecr-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: VictoriaMetrics
      provider-url: ""
    name: victoriametrics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: victoriametrics-operator.v0.63.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLAgent",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "persistentVolumeClaimRetentionPolicy": {
                    "whenDeleted": "Delete"
                  },
                  "remoteWrite": [
                    {
                      "maxDiskUsage": "10GB",
                      "url": "http://vlsingle-example-0.default.svc:9428/internal/insert"
                    }
                  ],
                  "remoteWriteSettings": {
                    "maxBlockSize": "30MB"
                  },
                  "replicaCount": 2,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "350Mi"
                    }
                  },
                  "storage": {
                    "volumeClaimTemplate": {
                      "spec": {
                        "resources": {
                          "requests": {
                            "storage": "50Gi"
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "victoriametrics-operator"
                  },
                  "name": "example"
                },
                "spec": {
                  "requestsLoadBalancer": {
                    "enabled": true,
                    "spec": {}
                  },
                  "vlinsert": {
                    "replicaCount": 1
                  },
                  "vlselect": {
                    "replicaCount": 1
                  },
                  "vlstorage": {
                    "replicaCount": 2,
                    "retentionPeriod": "1y",
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "50Gi"
                            }
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLSingle",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "victoriametrics-operator"
                  },
                  "name": "sample"
                },
                "spec": {
                  "retentionMaxDiskSpaceUsageBytes": "20GB",
                  "retentionPeriod": "1y",
                  "storage": {
                    "resources": {
                      "requests": {
                        "storage": "50Gi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VLogs",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "retentionPeriod": "2"
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAgent",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "extraArgs": {
                    "memory.allowedPercent": "40"
                  },
                  "remoteWrite": [
                    {
                      "maxDiskUsage": "10GB",
                      "url": "http://vmsingle-dc-1.default.svc:8429/api/v1/write"
                    },
                    {
                      "maxDiskUsage": "15GB",
                      "url": "http://vmsingle-dc-2.default.svc:8429/api/v1/write"
                    },
                    {
                      "url": "http://vminsert-main.default.svc:8480/insert/1:1/prometheus/api/v1/write"
                    }
                  ],
                  "remoteWriteSettings": {
                    "maxDiskUsagePerURL": "5GB"
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "350Mi"
                    }
                  },
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlert",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "datasource": {
                    "url": "http://vmsingle-example-pvc.default.svc:8429"
                  },
                  "evaluationInterval": "30s",
                  "notifier": {
                    "url": "http://vmalertmanager-example.default.svc:9093"
                  },
                  "replicaCount": 1,
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlertmanager",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "configSecret": "am-global-config",
                  "replicaCount": 1,
                  "selectAllByDefault": true,
                  "storage": {
                    "volumeClaimTemplate": {
                      "spec": {
                        "resources": {
                          "requests": {
                            "storage": "2Gi"
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlertmanagerConfig",
                "metadata": {
                  "name": "example-email-web"
                },
                "spec": {
                  "receivers": [
                    {
                      "email_configs": [
                        {
                          "from": "alerting@example.com",
                          "smarthost": "example.com:25",
                          "text": "ALARM",
                          "to": "some-email@example.com"
                        }
                      ],
                      "name": "email",
                      "webhook_configs": [
                        {
                          "url": "http://some-other-wh"
                        }
                      ]
                    }
                  ],
                  "route": {
                    "group_interval": "1m",
                    "receiver": "email",
                    "routes": [
                      {
                        "receiver": "email"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAuth",
                "metadata": {
                  "name": "vmauth-sample"
                },
                "spec": {
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "victoriametrics/vmauth"
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "250m",
                      "memory": "350Mi"
                    }
                  },
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMCluster",
                "metadata": {
                  "name": "example-persistent"
                },
                "spec": {
                  "replicationFactor": 2,
                  "retentionPeriod": "4",
                  "vminsert": {
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "0.5",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "vmselect": {
                    "cacheMountPath": "/select-cache",
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "0.5",
                        "memory": "500Mi"
                      }
                    },
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "2Gi"
                            }
                          }
                        }
                      }
                    }
                  },
                  "vmstorage": {
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1500Mi"
                      }
                    },
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "10Gi"
                            }
                          }
                        }
                      }
                    },
                    "storageDataPath": "/vm-data"
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMNodeScrape",
                "metadata": {
                  "name": "cadvisor-metrics"
                },
                "spec": {
                  "bearerTokenFile": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                  "interval": "10s",
                  "relabelConfigs": [
                    {
                      "action": "labelmap",
                      "regex": "__meta_kubernetes_node_label_(.+)"
                    },
                    {
                      "replacement": "kubernetes.default.svc:443",
                      "targetLabel": "__address__"
                    },
                    {
                      "regex": "(.+)",
                      "replacement": "/api/v1/nodes/$1/proxy/metrics/cadvisor",
                      "sourceLabels": [
                        "__meta_kubernetes_node_name"
                      ],
                      "targetLabel": "__metrics_path__"
                    }
                  ],
                  "scheme": "https",
                  "scrapeTimeout": "2s",
                  "selector": {},
                  "tlsConfig": {
                    "caFile": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                    "insecureSkipVerify": true
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMPodScrape",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "podMetricsEndpoints": [
                    {
                      "port": "metrics",
                      "scheme": "http"
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app.kubernetes.io/name": "example-application",
                      "monitored-by": "vm-operator"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMProbe",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "interval": "2s",
                  "jobName": "static-probe",
                  "module": "http_2xx",
                  "targets": {
                    "staticConfig": {
                      "targets": [
                        "vmagent-example.default.svc:8429/health"
                      ]
                    }
                  },
                  "vmProberSpec": {
                    "url": "prometheus-blackbox-exporter.default.svc:9115"
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMRule",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "groups": [
                    {
                      "name": "kafka",
                      "rules": [
                        {
                          "alert": "coordinator down",
                          "annotations": {
                            "description": "kafka coordinator is down",
                            "value": "{{ $value }}"
                          },
                          "expr": "ml_app_gauge{exec_context=\"consumer_group_state\"} == 0",
                          "for": "60s",
                          "labels": {
                            "job": "{{ $labels.job }}",
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMScrapeConfig",
                "metadata": {
                  "name": "mongodb"
                },
                "spec": {
                  "consulSDConfigs": [
                    {
                      "server": "https://consul-dns:8500",
                      "services": [
                        "mongodb"
                      ]
                    }
                  ],
                  "relabelConfigs": [
                    {
                      "action": "replace",
                      "sourceLabels": [
                        "__meta_consul_service"
                      ],
                      "targetLabel": "job"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMServiceScrape",
                "metadata": {
                  "name": "example-scrape"
                },
                "spec": {
                  "endpoints": [
                    {
                      "port": "http"
                    }
                  ],
                  "namespaceSelector": {},
                  "selector": {
                    "matchLabels": {
                      "app.kubernetes.io/name": "example-application",
                      "monitored-by": "vm-operator"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMSingle",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "retentionPeriod": "1"
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMSingle",
                "metadata": {
                  "name": "example-pvc"
                },
                "spec": {
                  "removePvcAfterDelete": true,
                  "retentionPeriod": "1",
                  "storage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMStaticScrape",
                "metadata": {
                  "name": "vmstaticscrape-sample"
                },
                "spec": {
                  "jobName": "static",
                  "targetEndpoints": [
                    {
                      "labels": {
                        "env": "dev",
                        "project": "operator"
                      },
                      "targets": [
                        "192.168.0.1:9100",
                        "196.168.0.50:9100"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMUser",
                "metadata": {
                  "name": "vmuser-tenant-1"
                },
                "spec": {
                  "bearerToken": "some-token",
                  "targetRefs": [
                    {
                      "crd": {
                        "kind": "VMCluster/vminsert",
                        "name": "test-persistent",
                        "namespace": "vm"
                      },
                      "target_path_suffix": "/insert/1"
                    },
                    {
                      "crd": {
                        "kind": "VMCluster/vmselect",
                        "name": "test-persistent",
                        "namespace": "vm"
                      },
                      "target_path_suffix": "/select/1"
                    },
                    {
                      "paths": [
                        "/internal/resetRollupResultCache"
                      ],
                      "static": {
                        "url": "http://vmselect-test-persistent.default.svc:8481/"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: docker.io/victoriametrics/operator:v0.63.0
          createdAt: "2025-09-12T12:58:29Z"
          description: Provides monitoring capabilities for kubernetes clusters and
            applications
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: "1.30"
          operators.openshift.io/infrastructure-features: '[fips]'
          operators.operatorframework.io.bundle.channel.default.v1: beta
          operators.operatorframework.io.bundle.channels.v1: beta
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/VictoriaMetrics/operator
          support: VictoriaMetrics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VLAgent - is a tiny but brave agent, which helps you collect
              logs from various sources and stores them in VictoriaLogs.
            displayName: VLAgent
            kind: VLAgent
            name: vlagents.operator.victoriametrics.com
            version: v1
          - description: VLCluster is fast, cost-effective and scalable logs database.
            displayName: VLCluster
            kind: VLCluster
            name: vlclusters.operator.victoriametrics.com
            version: v1
          - description: |-
              VLogs is fast, cost-effective and scalable logs database.
              VLogs is the Schema for the vlogs API
            displayName: VLogs
            kind: VLogs
            name: vlogs.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VLSingle is fast, cost-effective and scalable logs database.
              VLSingle is the Schema for the API
            displayName: VLSingle
            kind: VLSingle
            name: vlsingles.operator.victoriametrics.com
            version: v1
          - description: |-
              VMAgent - is a tiny but brave agent, which helps you collect metrics from various sources and stores them in VictoriaMetrics
              or any other Prometheus-compatible storage system that supports the remote_write protocol.
            displayName: VMAgent
            kind: VMAgent
            name: vmagents.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlertmanagerConfig is the Schema for the vmalertmanagerconfigs
              API
            displayName: VMAlertmanager Config
            kind: VMAlertmanagerConfig
            name: vmalertmanagerconfigs.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlertmanager represents Victoria-Metrics deployment for
              Alertmanager.
            displayName: VMAlertmanager
            kind: VMAlertmanager
            name: vmalertmanagers.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlert  executes a list of given alerting or recording rules
              against configured address.
            displayName: VMAlert
            kind: VMAlert
            name: vmalerts.operator.victoriametrics.com
            version: v1beta1
          - description: VMAnomaly is the Schema for the vmanomalies API.
            displayName: VMAnomaly
            kind: VMAnomaly
            name: vmanomalies.operator.victoriametrics.com
            version: v1
          - description: VMAuth is the Schema for the vmauths API
            displayName: VMAuth
            kind: VMAuth
            name: vmauths.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMCluster is fast, cost-effective and scalable time-series database.
              Cluster version with
            displayName: VMCluster
            kind: VMCluster
            name: vmclusters.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMNodeScrape defines discovery for targets placed on kubernetes nodes,
              usually its node-exporters and other host services.
              InternalIP is used as __address__ for scraping.
            displayName: VMNode Scrape
            kind: VMNodeScrape
            name: vmnodescrapes.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMPodScrape is scrape configuration for pods,
              it generates vmagent's config for scraping pod targets
              based on selectors.
            displayName: VMPod Scrape
            kind: VMPodScrape
            name: vmpodscrapes.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMProbe defines a probe for targets, that will be executed with prober,
              like blackbox exporter.
              It helps to monitor reachability of target with various checks.
            displayName: VMProbe
            kind: VMProbe
            name: vmprobes.operator.victoriametrics.com
            version: v1beta1
          - description: VMRule defines rule records for vmalert application
            displayName: VMRule
            kind: VMRule
            name: vmrules.operator.victoriametrics.com
            version: v1beta1
          - description: VMScrapeConfig specifies a set of targets and parameters
              describing how to scrape them.
            displayName: VMScrape Config
            kind: VMScrapeConfig
            name: vmscrapeconfigs.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMServiceScrape is scrape configuration for endpoints associated with
              kubernetes service,
              it generates scrape configuration for vmagent based on selectors.
              result config will scrape service endpoints
            displayName: VMService Scrape
            kind: VMServiceScrape
            name: vmservicescrapes.operator.victoriametrics.com
            version: v1beta1
          - description: VMSingle  is fast, cost-effective and scalable time-series
              database.
            displayName: VMSingle
            kind: VMSingle
            name: vmsingles.operator.victoriametrics.com
            version: v1beta1
          - description: VMStaticScrape  defines static targets configuration for
              scraping.
            displayName: VMStatic Scrape
            kind: VMStaticScrape
            name: vmstaticscrapes.operator.victoriametrics.com
            version: v1beta1
          - description: VMUser is the Schema for the vmusers API
            displayName: VMUser
            kind: VMUser
            name: vmusers.operator.victoriametrics.com
            version: v1beta1
          - description: VTCluster is fast, cost-effective and scalable traces database.
            displayName: VTCluster
            kind: VTCluster
            name: vtclusters.operator.victoriametrics.com
            version: v1
          - description: |-
              VTSingle is fast, cost-effective and scalable traces database.
              VTSingle is the Schema for the API
            displayName: VTSingle
            kind: VTSingle
            name: vtsingles.operator.victoriametrics.com
            version: v1
        description: |
          Operator manages VictoriaMetrics applications and provides monitoring features for applications running inside and outside kubernetes cluster. It has support for prometheus-operator objects and
          provides migration mechanism.

          Operator designed to use existing installation of VictoriaMetrics inside or outside kubernetes cluster or it could
          bring own resources and fully manages them.

           You can read more about operator at docs:
            - quick start [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/quick-start.md)
            - high availability [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/high-availability.md)
            - design and description of implementation [design](https://github.com/VictoriaMetrics/operator/blob/master/docs/design.md)
            - operator objects description [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/api.md)

          # VictoriaMetrics
          * VictoriaMetrics can be used as long-term storage for Prometheus or for [vmagent](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmagent/README.md).
          See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#prometheus-setup) for details.
          * Supports [Prometheus querying API](https://prometheus.io/docs/prometheus/latest/querying/api/), so it can be used as Prometheus drop-in replacement in Grafana.
          VictoriaMetrics implements [MetricsQL](https://github.com/VictoriaMetrics/VictoriaMetrics/wiki/MetricsQL) query language, which is inspired by PromQL.
          * Supports global query view. Multiple Prometheus instances may write data into VictoriaMetrics. Later this data may be used in a single query.
          * High performance and good scalability for both [inserts](https://medium.com/@valyala/high-cardinality-tsdb-benchmarks-victoriametrics-vs-timescaledb-vs-influxdb-13e6ee64dd6b)
          and [selects](https://medium.com/@valyala/when-size-matters-benchmarking-victoriametrics-vs-timescale-and-influxdb-6035811952d4).
          [Outperforms InfluxDB and TimescaleDB by up to 20x](https://medium.com/@valyala/measuring-vertical-scalability-for-time-series-databases-in-google-cloud-92550d78d8ae).
          * [Uses 10x less RAM than InfluxDB](https://medium.com/@valyala/insert-benchmarks-with-inch-influxdb-vs-victoriametrics-e31a41ae2893) when working with millions of unique time series (aka high cardinality).
          * Optimized for time series with high churn rate. Think about [prometheus-operator](https://github.com/coreos/prometheus-operator) metrics from frequent deployments in Kubernetes.
          * High data compression, so [up to 70x more data points](https://medium.com/@valyala/when-size-matters-benchmarking-victoriametrics-vs-timescale-and-influxdb-6035811952d4)
          may be crammed into limited storage comparing to TimescaleDB.
          * Optimized for storage with high-latency IO and low IOPS (HDD and network storage in AWS, Google Cloud, Microsoft Azure, etc). See [graphs from these benchmarks](https://medium.com/@valyala/high-cardinality-tsdb-benchmarks-victoriametrics-vs-timescaledb-vs-influxdb-13e6ee64dd6b).
          * A single-node VictoriaMetrics may substitute moderately sized clusters built with competing solutions such as Thanos, M3DB, Cortex, InfluxDB or TimescaleDB.
          See [vertical scalability benchmarks](https://medium.com/@valyala/measuring-vertical-scalability-for-time-series-databases-in-google-cloud-92550d78d8ae),
          [comparing Thanos to VictoriaMetrics cluster](https://medium.com/@valyala/comparing-thanos-to-victoriametrics-cluster-b193bea1683)
          and [Remote Write Storage Wars](https://promcon.io/2019-munich/talks/remote-write-storage-wars/) talk
          from [PromCon 2019](https://promcon.io/2019-munich/talks/remote-write-storage-wars/).
          * Easy operation:
              * VictoriaMetrics consists of a single [small executable](https://medium.com/@valyala/stripping-dependency-bloat-in-victoriametrics-docker-image-983fb5912b0d) without external dependencies.
              * All the configuration is done via explicit command-line flags with reasonable defaults.
              * All the data is stored in a single directory pointed by `-storageDataPath` flag.
              * Easy and fast backups from [instant snapshots](https://medium.com/@valyala/how-victoriametrics-makes-instant-snapshots-for-multi-terabyte-time-series-data-e1f3fb0e0282)
              to S3 or GCS with [vmbackup](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmbackup/README.md) / [vmrestore](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmrestore/README.md).
              See [this article](https://medium.com/@valyala/speeding-up-backups-for-big-time-series-databases-533c1a927883) for more details.
          * Storage is protected from corruption on unclean shutdown (i.e. OOM, hardware reset or `kill -9`) thanks to [the storage architecture](https://medium.com/@valyala/how-victoriametrics-makes-instant-snapshots-for-multi-terabyte-time-series-data-e1f3fb0e0282).
          * Supports metrics' scraping, ingestion and [backfilling](https://github.com/VictoriaMetrics/VictoriaMetrics#backfilling) via the following protocols:
            * [Metrics from Prometheus exporters](https://github.com/prometheus/docs/blob/master/content/docs/instrumenting/exposition_formats.md#text-based-format)
              such as [node_exporter](https://github.com/prometheus/node_exporter). See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-scrape-prometheus-exporters-such-as-node-exporter) for details.
              * [Prometheus remote write API](https://prometheus.io/docs/prometheus/latest/configuration/configuration/#remote_write)
              * [InfluxDB line protocol](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-send-data-from-influxdb-compatible-agents-such-as-telegraf) over HTTP, TCP and UDP.
              * [Graphite plaintext protocol](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-send-data-from-graphite-compatible-agents-such-as-statsd) with [tags](https://graphite.readthedocs.io/en/latest/tags.html#carbon)
              if `-graphiteListenAddr` is set.
              * [OpenTSDB put message](https://github.com/VictoriaMetrics/VictoriaMetrics#sending-data-via-telnet-put-protocol) if `-opentsdbListenAddr` is set.
              * [HTTP OpenTSDB /api/put requests](https://github.com/VictoriaMetrics/VictoriaMetrics#sending-opentsdb-data-via-http-apiput-requests) if `-opentsdbHTTPListenAddr` is set.
              * [How to import time series data](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-time-series-data).
              * [Prometheus exposition format](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-data-in-prometheus-exposition-format).
              * [Arbitrary CSV data](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-csv-data).
          * Supports metrics' relabeling. See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#relabeling) for details.
          * Ideally works with big amounts of time series data from Kubernetes, IoT sensors, connected cars, industrial telemetry, financial data and various Enterprise workloads.
        displayName: VictoriaMetrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - prometheus
        - victoriametrics
        links:
        - name: VictoriaMetrics
          url: https://victoriametrics.com
        maintainers:
        - email: nik@victoriametrics.com
          name: f41gh7
        - email: info@victoriametrics.com
          name: info
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: VictoriaMetrics
        relatedImages:
        - docker.io/victoriametrics/operator:v0.63.0
        version: 0.63.0
      entries:
      - name: victoriametrics-operator.v0.63.0
        version: 0.63.0
      - name: victoriametrics-operator.v0.62.0
        version: 0.62.0
      - name: victoriametrics-operator.v0.61.2
        version: 0.61.2
      - name: victoriametrics-operator.v0.61.1
        version: 0.61.1
      - name: victoriametrics-operator.v0.61.0
        version: 0.61.0
      - name: victoriametrics-operator.v0.60.2
        version: 0.60.2
      - name: victoriametrics-operator.v0.60.1
        version: 0.60.1
      - name: victoriametrics-operator.v0.60.0
        version: 0.60.0
      - name: victoriametrics-operator.v0.59.2
        version: 0.59.2
      - name: victoriametrics-operator.v0.59.1
        version: 0.59.1
      - name: victoriametrics-operator.v0.59.0
        version: 0.59.0
      - name: victoriametrics-operator.v0.58.1
        version: 0.58.1
      - name: victoriametrics-operator.v0.58.0
        version: 0.58.0
      - name: victoriametrics-operator.v0.57.0
        version: 0.57.0
      - name: victoriametrics-operator.v0.56.0
        version: 0.56.0
      - name: victoriametrics-operator.v0.55.0
        version: 0.55.0
      - name: victoriametrics-operator.v0.54.1
        version: 0.54.1
      - name: victoriametrics-operator.v0.53.0
        version: 0.53.0
      - name: victoriametrics-operator.v0.52.0
        version: 0.52.0
      - name: victoriametrics-operator.v0.51.1
        version: 0.51.1
      - name: victoriametrics-operator.v0.49.0
        version: 0.49.0
      - name: victoriametrics-operator.v0.48.4
        version: 0.48.4
      - name: victoriametrics-operator.v0.48.3
        version: 0.48.3
      - name: victoriametrics-operator.v0.48.1
        version: 0.48.1
      - name: victoriametrics-operator.v0.47.3
        version: 0.47.3
      - name: victoriametrics-operator.v0.47.1
        version: 0.47.1
      - name: victoriametrics-operator.v0.47.0
        version: 0.47.0
      - name: victoriametrics-operator.v0.46.4
        version: 0.46.4
      - name: victoriametrics-operator.v0.46.3
        version: 0.46.3
      - name: victoriametrics-operator.v0.39.4
        version: 0.39.4
      - name: victoriametrics-operator.v0.33.0
        version: 0.33.0
      - name: victoriametrics-operator.v0.32.1
        version: 0.32.1
      - name: victoriametrics-operator.v0.32.0
        version: 0.32.0
      - name: victoriametrics-operator.v0.29.1
        version: 0.29.1
      - name: victoriametrics-operator.v0.27.2
        version: 0.27.2
      - name: victoriametrics-operator.v0.26.2
        version: 0.26.2
      - name: victoriametrics-operator.v0.26.0
        version: 0.26.0
      - name: victoriametrics-operator.v0.25.1
        version: 0.25.1
      - name: victoriametrics-operator.v0.25.0
        version: 0.25.0
      - name: victoriametrics-operator.v0.24.0
        version: 0.24.0
      - name: victoriametrics-operator.v0.23.3
        version: 0.23.3
      - name: victoriametrics-operator.v0.22.1
        version: 0.22.1
      - name: victoriametrics-operator.v0.22.0
        version: 0.22.0
      - name: victoriametrics-operator.v0.21.0
        version: 0.21.0
      - name: victoriametrics-operator.v0.20.1
        version: 0.20.1
      - name: victoriametrics-operator.v0.19.1
        version: 0.19.1
      - name: victoriametrics-operator.v0.19.0
        version: 0.19.0
      - name: victoriametrics-operator.v0.18.2
        version: 0.18.2
      - name: victoriametrics-operator.v0.18.1
        version: 0.18.1
      - name: victoriametrics-operator.v0.18.0
        version: 0.18.0
      name: beta
    defaultChannel: beta
    packageName: victoriametrics-operator
    provider:
      name: VictoriaMetrics
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: F5 Networks Inc.
      provider-url: ""
    name: f5-bigip-ctlr-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: alpha
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: beta
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: stable
    defaultChannel: stable
    packageName: f5-bigip-ctlr-operator
    provider:
      name: F5 Networks Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dns-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dns-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSHealthCheckProbe",
                "metadata": {
                  "name": "$NAME"
                },
                "spec": {
                  "additionalHeadersRef": {
                    "name": "headers"
                  },
                  "address": "192.168.0.16",
                  "allowInsecureCertificate": true,
                  "failureThreshold": 5,
                  "hostname": "test.com",
                  "interval": "60s",
                  "path": "/healthz",
                  "port": 443,
                  "protocol": "HTTPS"
                }
              },
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSRecord",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dns-operator",
                    "app.kubernetes.io/instance": "dnsrecord-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dnsrecord",
                    "app.kubernetes.io/part-of": "dns-operator"
                  },
                  "name": "dnsrecord-sample"
                },
                "spec": {
                  "endpoints": [
                    {
                      "dnsName": "dnsrecord-simple.example.com",
                      "recordTTL": 60,
                      "recordType": "A",
                      "targets": [
                        "52.215.108.61",
                        "52.30.101.221"
                      ]
                    }
                  ],
                  "providerRef": {
                    "name": "dns-provider-creds"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/dns-rhel9-operator@sha256:8ad3876e36ba2f76355499d98a97672f54f26754fc9d219d9cb1c5b9f9874430
          createdAt: "2025-04-07T08:13:58Z"
          description: Configures how north-south traffic from outside the network
            is balanced and reaches Gateways
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kuadrant/dns-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSHealthCheckProbe is the Schema for the dnshealthcheckprobes
              API
            displayName: DNSHealthCheckProbe
            kind: DNSHealthCheckProbe
            name: dnshealthcheckprobes.kuadrant.io
            version: v1alpha1
          - description: DNSRecord is the Schema for the dnsrecords API
            displayName: DNSRecord
            kind: DNSRecord
            name: dnsrecords.kuadrant.io
            version: v1alpha1
        description: Configures how north-south traffic from outside the network is
          balanced and reaches Gateways - Part of Red Hat Connectivity Link
        displayName: Red Hat - DNS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dns
        - kuadrant
        links:
        - name: GitHub
          url: https://github.com/kuadrant/dns-operator
        - name: Documentation
          url: https://github.com/Kuadrant/dns-operator/tree/main/docs
        maintainers:
        - email: mnairn@redhat.com
          name: Michael Nairn
        - email: pbrookes@redhat.com
          name: Phil Brookes
        - email: cbrookes@redhat.com
          name: Craig Brookes
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhcl-1/dns-rhel9-operator@sha256:8ad3876e36ba2f76355499d98a97672f54f26754fc9d219d9cb1c5b9f9874430
        version: 1.1.0
      entries:
      - name: dns-operator.v1.1.0
        version: 1.1.0
      - name: dns-operator.v1.0.2
        version: 1.0.2
      - name: dns-operator.v1.0.1
        version: 1.0.1
      - name: dns-operator.v0.12.0
        version: 0.12.0
      name: stable
    defaultChannel: stable
    packageName: dns-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: patch-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: patch-operator.v0.1.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Patch",
                "metadata": {
                  "name": "test-complex-patch"
                },
                "spec": {
                  "patches": {
                    "test-complex-patch": {
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 1).metadata.name }}: {{ (index . 2).metadata.name }}    \n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "kind": "Namespace",
                          "name": "test-patch-operator"
                        },
                        {
                          "apiVersion": "v1",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "test-patch-operator"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "test-patch-operator"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/patch-operator@sha256:030ade9b9428deca3ac6390618dc4d6e6d728577d1467a22dc4d4b68c9848886
          createdAt: "2024-05-10T14:07:19Z"
          description: An operator to apply patches to Kubernetes objects in a declarative
            way.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: patch-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/patch-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Patch
            name: patches.redhatcop.redhat.io
            version: v1alpha1
        description: "The patch operator helps with defining patches in a declarative
          way. This operator has two main features:\n\n1. [ability to patch an object
          at creation time via a mutating webhook](#creation-time-patch-injection)\n2.
          [ability to enforce patches on one or more objects via a controller](#runtime-patch-enforcement)\n
          \n## Creation-time patch injection\n\nWhy apply a patch at creation time
          when you could directly create the correct object? The reason is that sometime
          the correct value depends on configuration set on the specific cluster in
          which the object is being deployed. For example, an ingress/route hostname
          might depend on the specific cluster. Consider the following example based
          on cert-manager:\n \n```yaml\napiVersion: cert-manager.io/v1\nkind: ClusterIssuer\nmetadata:\n
          \ name: letsencrypt-issuer\nspec:\n  acme:\n    server: 'https://acme-v02.api.letsencrypt.org/directory'\n
          \   email: {{ .Values.letsencrypt.email }}\n    privateKeySecretRef:\n      name:
          letsencrypt-staging\n    solvers:  \n    - dns01:\n        route53:\n          accessKeyID:
          << access_key >>\n          secretAccessKeySecretRef:\n            name:
          cert-manager-dns-credentials\n            key: aws_secret_access_key\n          region:
          << region >>\n          hostedZoneID: << hosted_zone_id >>\n ```\n\nIn this
          example the fields: `<< access_key >>`, `<< region >>` and `<< hosted_zone_id
          >>` are dependent on the specific region in which the cluster is being deployed
          and in many cases they are discoverable from other configurations already
          present in the cluster. If you want to deploy the above Cluster Issuer object
          with a gitops approach, then there is no easy way to discover those values.
          The solution so far is to manually discover those values and create a different
          gitops configuration for each cluster. But consider if you could look up
          values at deploy time based on the cluster you are deploying to. Here is
          how this object might look:\n\n```yaml\napiVersion: cert-manager.io/v1\nkind:
          ClusterIssuer\nmetadata:\n  name: letsencrypt-issuer\n  namespace: {{ .Release.Namespace
          }}\n  annotations:\n    \"redhat-cop.redhat.io/patch\": |\n      spec:\n
          \       acme:\n        - dns01:\n            route53:\n              accessKeyID:
          {{ (lookup \"v1\" \"Secret\" .metadata.namespace \"cert-manager-dns-credentials\").data.aws_access_key_id
          | b64dec }}\n              secretAccessKeySecretRef:\n                name:
          cert-manager-dns-credentials\n                key: aws_secret_access_key\n
          \             region: {{ (lookup \"config.openshift.io/v1\" \"Infrastructure\"
          \"\" \"cluster\").status.platformStatus.aws.region }}\n              hostedZoneID:
          {{ (lookup \"config.openshift.io/v1\" \"DNS\" \"\" \"cluster\").spec.publicZone.id
          }} \nspec:\n  acme:\n    server: 'https://acme-v02.api.letsencrypt.org/directory'\n
          \   email: {{ .Values.letsencrypt.email }}\n    privateKeySecretRef:\n      name:
          letsencrypt-staging\n    solvers:  \n    - dns01:\n        route53:\n          accessKeyID:
          << access_key >>\n          secretAccessKeySecretRef:\n            name:
          cert-manager-dns-credentials\n            key: aws_secret_access_key\n          region:
          << region >>\n          hostedZoneID: << hosted_zone_id >>\n ```\n\nThe
          annotation specifies a patch that will be applied by a [MutatingWebhook](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/),
          as you can see the three values are now looked up from different configurations
          in the cluster.\n\nTwo annotations influences the behavior of this MutatingWebhook\n
          \n1. \"redhat-cop.redhat.io/patch\" : this is the patch itself. The patch
          is evaluated as a template with the object itself as it's only parameter.
          The template is expressed in golang template notation and supports the same
          functions as helm template including the [lookup](https://helm.sh/docs/chart_template_guide/functions_and_pipelines/#using-the-lookup-function)
          function which plays a major role here. The patch must be expressed in yaml
          for readability. It will be converted to json by the webhook logic.\n2.
          \"redhat-cop.redhat.io/patch-type\" : this is the type of json patch. The
          possible values are: `application/json-patch+json`, `application/merge-patch+json`
          and `application/strategic-merge-patch+json`. If this annotation is omitted
          it defaults to strategic merge.\n\n### Security Considerations\n \nThe lookup
          function, if used by the template, is executed with a client which impersonates
          the user issuing the object creation/update request. This should prevent
          security permission leakage.\n\n### Installing the creation time webhook\n
          \nThe creation time webhook is not installed by the operator. This is because
          there is no way to know which specific object type should be intercepted
          and intercepting all of the types would be too inefficient. It's up to the
          administrator then to install the webhook. Here is some guidance.\n\nIf
          you installed the operator via OLM, use the following webhook template:\n\n```yaml\napiVersion:
          admissionregistration.k8s.io/v1\n kind: MutatingWebhookConfiguration\nmetadata:\n
          \ name: patch-operator-inject\n  annotations:\n    service.beta.openshift.io/inject-cabundle:
          \"true\"\nwebhooks:\n - admissionReviewVersions:\n  - v1\n  clientConfig:\n
          \   service:\n      name: patch-operator-webhook-service\n      namespace:
          patch-operator\n      path: /inject\n  failurePolicy: Fail\n  name: patch-operator-inject.redhatcop.redhat.io\n
          \ rules:\n  - << add your intercepted objects here >>\n  sideEffects: None\n```\n\nIf
          you installed the operator via the Helm chart and are using cert-manager,
          use the following webhook template:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\n
          kind: MutatingWebhookConfiguration\nmetadata:\n  name: patch-operator-inject\n
          \ annotations:\n    cert-manager.io/inject-ca-from: '{{ .Release.Namespace
          }}/webhook-server-cert'\n webhooks:\n- admissionReviewVersions:\n  - v1\n
          \ clientConfig:\n    service:\n      name: patch-operator-webhook-service\n
          \     namespace: patch-operator\n      path: /inject\n  failurePolicy: Fail\n
          \ name: patch-operator-inject.redhatcop.redhat.io\n  rules:\n  - << add
          your intercepted objects here >>\n  sideEffects: None\n```  \n\nYou should
          need to enable the webhook only for `CREATE` operations. So for example
          to enable the webhook on configmaps:\n\n```yaml\n  rules:\n  - apiGroups:\n
          \   - \"\"\n    apiVersions:\n    - v1\n    operations:\n    - CREATE\n
          \   resources:\n    - configmaps\n```\n\n## Runtime patch enforcement\n\nThere
          are situations when we need to patch pre-existing objects. Again this is
          a use case that is hard to model with gitops operators which will work only
          on object that they own. Especially with sophisticated Kubernetes distributions,
          it is not uncommon that a Kubernetes instance, at installation time, is
          configured with some default settings. Changing those configurations means
          patching those objects. For example, let's take the case of OpenShift Oauth
          configuration. This object is present by default and it is expected to be
          patched with any newly enabled authentication mechanism. This is how it
          looks like after installation:\n\n```yaml\napiVersion: config.openshift.io/v1\n
          kind: OAuth\nmetadata:\n  name: cluster\n  ownerReferences:\n    - apiVersion:
          config.openshift.io/v1\n      kind: ClusterVersion\n      name: version\n
          \     uid: 9a9d450b-3076-4e30-ac05-a889d6341fc3\n  resourceVersion: '20405124'\n
          spec: []\n```\n\nIf we need to patch it we can use the patch controller
          and the `Patch` object as follows:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: Patch\nmetadata:\n  name: gitlab-ocp-oauth-provider\n  namespace:
          openshift-config\nspec:\n  serviceAccountRef:\n    name: default\n  patches:\n
          \   gitlab-ocp-oauth-provider:\n      targetObjectRef:\n        apiVersion:
          config.openshift.io/v1\n        kind: OAuth\n        name: cluster\n      patchTemplate:
          |\n        spec:\n          identityProviders:\n          - name: my-github
          \n            mappingMethod: claim \n            type: GitHub\n            github:\n
          \             clientID: \"{{ (index . 1).data.client_id | b64dec }}\" \n
          \             clientSecret: \n                name: ocp-github-app-credentials\n
          \             organizations: \n              - my-org\n              teams:
          []            \n      patchType: application/merge-patch+json\n      sourceObjectRefs:\n
          \     - apiVersion: v1\n        kind: Secret\n        name: ocp-github-app-credentials\n
          \       namespace: openshift-config\n```\n\nThis will cause the OAuth object
          to be patched and the patch to be enforced. That means that if anything
          changes on the secret that we use a parameter (which may be rotated) or
          in Oauth object itself, the patch will be reapplied. In this case we are
          adding a gitlab authentication provider.\n\n A `patch` has the following
          fields:\n\n`targetObjectRef` this refers to the object(s) receiving the
          patch. Mutliple object can be selected based on the following rules:\n \n|
          Namespaced Type | Namespace | Name | Selection type |\n| --- | --- | ---
          | --- |\n| yes | null | null | multiple selection across namespaces |\n|
          yes | null | not null | multiple selection across namespaces where the name
          corresponds to the passed name |\n| yes | not null | null | multiple selection
          within a namespace |\n| yes | not null | not nul | single selection |\n|
          no | N/A | null | multiple selection  |\n| no | N/A | not null | single
          selection |\n\nSelection can be further narrowed down by filtering by labels
          and/or annotations using the `labelSelector` and `annotationSelector` fields.
          The patch will be applied to all of the selected instances.\n\n`sourceObjectRefs`
          these are the objects that will be watched and become part of the parameters
          of the patch template. Name and Namespace of sourceRefObjects are interpreted
          as golang templates with the current target instance and the only parameter.
          This allows to select different source object for each target object.\n
          \nSo, for example, with this patch:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: Patch\nmetadata:\n  name: multiple-namespaced-targets-patch\nspec:\n
          \ serviceAccountRef:\n    name: default\n  patches:\n    multiple-namespaced-targets-patch:\n
          \     targetObjectRef:\n        apiVersion: v1\n        kind: ServiceAccount\n
          \       name: deployer\n      patchTemplate: |\n        metadata:\n          annotations:\n
          \           {{ (index . 1).metadata.uid }}: {{ (index . 1) }}\n      patchType:
          application/strategic-merge-patch+json\n      sourceObjectRefs:\n      -
          apiVersion: v1\n        kind: ServiceAccount\n        name: default\n        namespace:
          \"{{ .metadata.namespace }}\"\n        fieldPath: $.metadata.uid\n```\n\nThe
          `deployer` service accounts from all namespaces are selected as target of
          this patch, each patch template will receive a different parameter and that
          is the `default` service account of the same namespace as the namespace
          of the `deployer` service account being processed.\n\n`sourceObjectRefs`
          also have the `fieldPath` field which can contain a jsonpath expression.
          If a value is passed the jsonpath expression will be calculate for the current
          source object and the result will be passed as parameter of the template.\n\n`patchTemplate`
          This is the the template that will be evaluated. The result must be a valid
          patch compatible with the requested type and expressed in yaml for readability.
          The parameters passed to the template are the target object and then the
          all of the source object. So if you want to refer to the target object in
          the template you can use this expression `(index . 0)`. Higher indexes refer
          to the sourceObjectRef array. The template is expressed in golang template
          notation and supports the same functions as helm template.\n\n`patchType`
          is the type of the json patch. The possible values are: `application/json-patch+json`,
          `application/merge-patch+json` and `application/strategic-merge-patch+json`.
          If this annotation is omitted it defaults to strategic merge.\n\n### Patch
          Controller Security Considerations\n\nThe patch enforcement enacted by the
          patch controller is executed with a client which uses the service account
          referenced by the `serviceAccountRef` field. So before a patch object can
          actually work an administrator must have granted the needed permissions
          to a service account in the same namespace. The `serviceAccountRef` will
          default to the `default` service account if not specified.\n\n### Patch
          Controller Performance Considerations\n\nThe patch controller will create
          a controller-manager and per `Patch` object and a reconciler for each of
          the `PatchSpec` defined in the array on patches in the `Patch` object.\nThese
          reconcilers share the same cached client. In order to be able to watch changes
          on target and source objects of a `PatchSpec`, all of the target and source
          object type instances will be cached by the client. This is a normal behavior
          of a controller-manager client, but it implies that if you create patches
          on object types that have many instances in etcd (Secrets, ServiceAccounts,
          Namespaces for example), the patch operator instance will require a significant
          amount of memory. A way to contain this issue is to try to aggregate together
          `PatchSpec` that deal with the same object types. This will cause those
          object type instances to cached only once.  \n"
        displayName: Patch Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - patch
        - enforcing
        - declarative
        links:
        - name: Patch Operator
          url: https://github.com/redhat-cop/patch-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/patch-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/patch-operator@sha256:030ade9b9428deca3ac6390618dc4d6e6d728577d1467a22dc4d4b68c9848886
        version: 0.1.11
      entries:
      - name: patch-operator.v0.1.11
        version: 0.1.11
      - name: patch-operator.v0.1.10
        version: 0.1.10
      - name: patch-operator.v0.1.9
        version: 0.1.9
      - name: patch-operator.v0.1.8
        version: 0.1.8
      - name: patch-operator.v0.1.7
        version: 0.1.7
      - name: patch-operator.v0.1.6
        version: 0.1.6
      - name: patch-operator.v0.1.5
        version: 0.1.5
      - name: patch-operator.v0.1.4
        version: 0.1.4
      - name: patch-operator.v0.1.3
        version: 0.1.3
      - name: patch-operator.v0.1.2
        version: 0.1.2
      - name: patch-operator.v0.1.1
        version: 0.1.1
      - name: patch-operator.v0.1.0
        version: 0.1.0
      - name: patch-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: patch-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-elbv2-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-elbv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "LoadBalancer",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "TargetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/elbv2-controller:1.1.1
          createdAt: "2025-09-19T22:16:38Z"
          description: AWS ELB controller is a service controller for managing ELB
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Listener represents the state of an AWS elbv2 Listener resource.
            displayName: Listener
            kind: Listener
            name: listeners.elbv2.services.k8s.aws
            version: v1alpha1
          - description: LoadBalancer represents the state of an AWS elbv2 LoadBalancer
              resource.
            displayName: LoadBalancer
            kind: LoadBalancer
            name: loadbalancers.elbv2.services.k8s.aws
            version: v1alpha1
          - description: Rule represents the state of an AWS elbv2 Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.elbv2.services.k8s.aws
            version: v1alpha1
          - description: TargetGroup represents the state of an AWS elbv2 TargetGroup
              resource.
            displayName: TargetGroup
            kind: TargetGroup
            name: targetgroups.elbv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Load Balancing (ELB) resources in AWS from within your Kubernetes cluster.

          **About Amazon ELB**

          Elastic Load Balancing automatically distributes your incoming traffic across multiple targets, such as EC2 instances, containers, and IP addresses, in one or more Availability Zones. It monitors the health of its registered targets, and routes traffic only to the healthy targets. Elastic Load Balancing scales your load balancer capacity automatically in response to changes in incoming traffic.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ELB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elbv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ELB Developer Resources
          url: https://aws.amazon.com/elasticloadbalancing
        maintainers:
        - email: ack-maintainers@amazon.com
          name: elb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/elbv2-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-elbv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-elbv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-elbv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-elbv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-elbv2-controller.v1.0.10
        version: 1.0.10
      - name: ack-elbv2-controller.v1.0.9
        version: 1.0.9
      - name: ack-elbv2-controller.v1.0.8
        version: 1.0.8
      - name: ack-elbv2-controller.v1.0.7
        version: 1.0.7
      - name: ack-elbv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-elbv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-elbv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-elbv2-controller.v1.0.3
        version: 1.0.3
      - name: ack-elbv2-controller.v1.0.2
        version: 1.0.2
      - name: ack-elbv2-controller.v1.0.1
        version: 1.0.1
      - name: ack-elbv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-elbv2-controller.v0.0.11
        version: 0.0.11
      - name: ack-elbv2-controller.v0.0.10
        version: 0.0.10
      - name: ack-elbv2-controller.v0.0.9
        version: 0.0.9
      - name: ack-elbv2-controller.v0.0.8
        version: 0.0.8
      - name: ack-elbv2-controller.v0.0.6
        version: 0.0.6
      - name: ack-elbv2-controller.v0.0.5
        version: 0.0.5
      - name: ack-elbv2-controller.v0.0.4
        version: 0.0.4
      - name: ack-elbv2-controller.v0.0.3
        version: 0.0.3
      - name: ack-elbv2-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-elbv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hazelcast, Inc
      provider-url: ""
    name: hazelcast-platform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hazelcast-platform-operator.v5.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "cache-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "CronHotBackup",
                "metadata": {
                  "name": "cronhotbackup-sample"
                },
                "spec": {
                  "hotBackupTemplate": {
                    "metadata": {
                      "labels": {
                        "cron-hotbackup": "true"
                      }
                    },
                    "spec": {
                      "hazelcastResourceName": "hazelcast"
                    }
                  },
                  "schedule": "*/30 * * * *"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast"
                },
                "spec": {
                  "clusterSize": 3,
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast-sample"
                },
                "spec": {
                  "clusterSize": 3,
                  "jet": {
                    "bucketConfig": {
                      "bucketURI": "gs://operator-user-code/jetJobs",
                      "secretName": "br-secret-gcp"
                    },
                    "enabled": true,
                    "resourceUploadEnabled": true
                  },
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HazelcastEndpoint",
                "metadata": {
                  "name": "hazelcastendpoint-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "port": 5701,
                  "type": "Discovery"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HotBackup",
                "metadata": {
                  "name": "hot-backup"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-longrun"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-longrun-2.0.0.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-1.0.2.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJobSnapshot",
                "metadata": {
                  "name": "jetjobsnapshot-sample"
                },
                "spec": {
                  "cancelJob": false,
                  "jetJobResourceName": "jet-job"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ManagementCenter",
                "metadata": {
                  "name": "managementcenter"
                },
                "spec": {
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "hazelcastClusters": [
                    {
                      "address": "hazelcast",
                      "name": "dev"
                    }
                  ],
                  "licenseKeySecretName": "hazelcast-license-key",
                  "persistence": {
                    "enabled": true,
                    "size": "10Gi"
                  },
                  "repository": "hazelcast/management-center"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Map",
                "metadata": {
                  "name": "map"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "MultiMap",
                "metadata": {
                  "name": "multimap-sample"
                },
                "spec": {
                  "backupCount": 3,
                  "binary": true,
                  "collectionType": "LIST",
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "queue-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ReplicatedMap",
                "metadata": {
                  "name": "replicatedmap-sample"
                },
                "spec": {
                  "asyncFillup": true,
                  "hazelcastResourceName": "hazelcast",
                  "inMemoryFormat": "OBJECT"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "topic-sample"
                },
                "spec": {
                  "globalOrderingEnabled": true,
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "UserCodeNamespace",
                "metadata": {
                  "name": "ucn-sample"
                },
                "spec": {
                  "bucketConfig": {
                    "bucketURI": "gs://operator-user-code/userCodeNamespace",
                    "secretName": "br-secret-gcp"
                  },
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanReplication",
                "metadata": {
                  "name": "wanreplication-sample"
                },
                "spec": {
                  "endpoints": "35.192.33.252",
                  "resources": [
                    {
                      "kind": "Map",
                      "name": "map-sample"
                    },
                    {
                      "kind": "Hazelcast",
                      "name": "hz-sample"
                    }
                  ],
                  "targetClusterName": "dev"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanSync",
                "metadata": {
                  "name": "wansync-sample"
                },
                "spec": {
                  "wanReplicationResourceName": "wan-replication-sample"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: docker.io/hazelcast/hazelcast-platform-operator@sha256:99005afda5c71ceeac200928dfed4ea35480f871053101e3aeaa44005e3fda14
          createdAt: "2024-07-29T15:35:56Z"
          description: '[DEPRECATED] Use the certified Hazelcast Operator instead'
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hazelcast/hazelcast-platform-operator/
          support: Hazelcast, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.hazelcast.com
            version: v1alpha1
          - description: CronHotBackup is the Schema for the cronhotbackups API
            displayName: Cron Hot Backup
            kind: CronHotBackup
            name: cronhotbackups.hazelcast.com
            version: v1alpha1
          - description: HazelcastEndpoint is the Schema for the hazelcastendpoints
              API
            displayName: Hazelcast Endpoint
            kind: HazelcastEndpoint
            name: hazelcastendpoints.hazelcast.com
            version: v1alpha1
          - description: Hazelcast is the Schema for the hazelcasts API
            displayName: Hazelcast
            kind: Hazelcast
            name: hazelcasts.hazelcast.com
            version: v1alpha1
          - description: HotBackup is the Schema for the hot backup API
            displayName: Hot Backup
            kind: HotBackup
            name: hotbackups.hazelcast.com
            version: v1alpha1
          - description: JetJob is the Schema for the jetjobs API
            displayName: Jet Job
            kind: JetJob
            name: jetjobs.hazelcast.com
            version: v1alpha1
          - description: JetJobSnapshot is the Schema for the jetjobsnapshots API
            displayName: Jet Job Snapshot
            kind: JetJobSnapshot
            name: jetjobsnapshots.hazelcast.com
            version: v1alpha1
          - description: ManagementCenter is the Schema for the managementcenters
              API
            displayName: Management Center
            kind: ManagementCenter
            name: managementcenters.hazelcast.com
            version: v1alpha1
          - description: Map is the Schema for the maps API
            displayName: Map
            kind: Map
            name: maps.hazelcast.com
            version: v1alpha1
          - description: MultiMap is the Schema for the multimaps API
            displayName: Multi Map
            kind: MultiMap
            name: multimaps.hazelcast.com
            version: v1alpha1
          - description: Queue is the Schema for the queues API
            displayName: Queue
            kind: Queue
            name: queues.hazelcast.com
            version: v1alpha1
          - description: ReplicatedMap is the Schema for the replicatedmaps API
            displayName: Replicated Map
            kind: ReplicatedMap
            name: replicatedmaps.hazelcast.com
            version: v1alpha1
          - description: Topic is the Schema for the topics API
            displayName: Topic
            kind: Topic
            name: topics.hazelcast.com
            version: v1alpha1
          - description: UserCodeNamespace is the Schema for the usercodenamespaces
              API
            displayName: User Code Namespace
            kind: UserCodeNamespace
            name: usercodenamespaces.hazelcast.com
            version: v1alpha1
          - description: WanReplication is the Schema for the wanreplications API
            displayName: Wan Replication
            kind: WanReplication
            name: wanreplications.hazelcast.com
            version: v1alpha1
          - description: WanSync is the Schema for the wansyncs API
            displayName: Wan Sync
            kind: WanSync
            name: wansyncs.hazelcast.com
            version: v1alpha1
        description: |
          ## Deprecated
          Use the certified Hazelcast Operator instead.
          # Hazelcast Platform Operator #

          Easily deploy Hazelcast clusters and Management Center into Kubernetes environments and manage their lifecycles.

          ## Before You Start

          If you are planning to create Hazelcast Platform Enterprise clusters, you need to [create a secret](https://docs.hazelcast.com/operator/latest/get-started#step-2-start-the-hazelcast-cluster) for the license. You can request a trial license key from [here](https://trialrequest.hazelcast.com).

          For Hazelcast Platform clusters, you can simply continue.

          ## Documentation

          1. [Get started](https://docs.hazelcast.com/operator/latest/get-started) with the Operator
          2. [Connect to the cluster from outside Kubernetes](https://guides.hazelcast.org/hazelcast-platform-operator-expose-externally)
          3. [Restore a Cluster from Cloud Storage with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-external-backup-restore)
          4. [Replicate Data between Two Hazelcast Clusters with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-wan-replication)

          ## Features

          Hazelcast Platform Operator supports the features below:

          * Custom resource for Hazelcast Platform (Enterprise) and Management Center
          * Observe status of Hazelcast clusters and Management Center instances
          * High Availability Mode configuration to create clusters that are resilient to node and zone failures
          * Support for TLS connections between members using self-signed certificates
          * Scale up and down Hazelcast clusters
          * Expose Hazelcast cluster to external clients ([Smart & Unisocket](https://docs.hazelcast.com/hazelcast/latest/clients/java#java-client-operation-modes))
          * Backup Hazelcast persistence data to cloud storage with the possibility of scheduling it and restoring the data accordingly
          * WAN Replication feature when you need to synchronize multiple Hazelcast clusters, which are connected by WANs
          * Full and Delta WAN Sync support
          * Tiered Storage support for Hazelcast and Map CRs
          * CP Subsystem configuration support for Hazelcast CR
          * User Code Deployment feature, which allows you to deploy custom and domain classes from cloud storages and URLs to Hazelcast members
          * The User Code Namespaces feature allows you to deploy custom and domain classes at runtime without a members restart
          * Support the configuration of advanced networking options
          * Support Multi-namespace configuration
          * ExecutorService and EntryProcessor support
          * Support several data structures like Map, Topic, MultiMap, ReplicatedMap, Queue and Cache which can be created dynamically via specific Custom Resources
          * MapStore, Near Cache and off-heap memory (HD memory and native memory) support for the Map CR
          * Native Memory support for the Cache CR
          * Support Jet configuration and Jet Job submission using the JetJob CR
          * Support for exporting the snapshots of JetJob CRs using JetJobSnapshot CR
          * Support for custom configurations using ConfigMap
        displayName: '[DEPRECATED] Hazelcast Platform Operator'
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hazelcast
        - keyvalue
        - in-memory
        - database
        - caching
        links:
        - name: Documentation
          url: https://docs.hazelcast.com/operator/latest/
        - name: Source Code
          url: https://github.com/hazelcast/hazelcast-platform-operator
        maintainers:
        - email: cloudnative@hazelcast.com
          name: Hazelcast Cloud Native Team
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Hazelcast, Inc
        relatedImages:
        - docker.io/hazelcast/hazelcast-platform-operator@sha256:99005afda5c71ceeac200928dfed4ea35480f871053101e3aeaa44005e3fda14
        version: 5.13.1
      entries:
      - name: hazelcast-platform-operator.v5.13.1
        version: 5.13.1
      - name: hazelcast-platform-operator.v5.13.0
        version: 5.13.0
      - name: hazelcast-platform-operator.v5.12.0
        version: 5.12.0
      - name: hazelcast-platform-operator.v5.11.0
        version: 5.11.0
      - name: hazelcast-platform-operator.v5.10.0
        version: 5.10.0
      - name: hazelcast-platform-operator.v5.9.0
        version: 5.9.0
      - name: hazelcast-platform-operator.v5.8.0
        version: 5.8.0
      - name: hazelcast-platform-operator.v5.7.0
        version: 5.7.0
      - name: hazelcast-platform-operator.v5.6.0
        version: 5.6.0
      - name: hazelcast-platform-operator.v5.5.0
        version: 5.5.0
      - name: hazelcast-platform-operator.v5.4.0
        version: 5.4.0
      - name: hazelcast-platform-operator.v5.3.0
        version: 5.3.0
      - name: hazelcast-platform-operator.v5.2.0
        version: 5.2.0
      - name: hazelcast-platform-operator.v5.1.0
        version: 5.1.0
      - name: hazelcast-platform-operator.v5.0.0
        version: 5.0.0
      name: alpha
    defaultChannel: alpha
    packageName: hazelcast-platform-operator
    provider:
      name: Hazelcast, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-storage
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-storage.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Storage",
                "metadata": {
                  "name": "storage-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore"
                  },
                  "minio": {
                    "clusterDomain": "cluster.local",
                    "helmKubectlJqImage": {
                      "repository": "docker.io/bskim45/helm-kubectl-jq"
                    },
                    "image": {
                      "repository": "prod.harbor.keysight.digital/loadcore/wap-minio",
                      "tag": "RELEASE.2021-12-29T06-49-06Z-3747-37-ocp"
                    },
                    "mcImage": {
                      "repository": "prod.harbor.keysight.digital/loadcore/wap-minio-mc",
                      "tag": "RELEASE.2021-12-29T06-52-55Z-3747-37-ocp"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "securityContext": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:32:20Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Storage
            name: storages.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Storage Operator
        displayName: Keysight WAP Storage Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Storage
          url: https://keysight-wap-storage.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-storage@sha256:9f7710cfe72658cd5456fd54f86ca6be5c9a4de2a01114c0f7531b70712bd981
        version: 0.1.21
      entries:
      - name: keysight-wap-storage.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-storage
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/cluster-nfd-operator
    name: openshift-nfd-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfd.v4.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nfd.openshift.io/v1",
                "kind": "NodeFeatureDiscovery",
                "metadata": {
                  "name": "nfd-instance",
                  "namespace": "openshift-nfd"
                },
                "spec": {
                  "customConfig": {
                    "configData": "#    - name: \"more.kernel.features\"\n#      matchOn:\n#      - loadedKMod: [\"example_kmod3\"]\n#    - name: \"more.features.by.nodename\"\n#      value: customValue\n#      matchOn:\n#      - nodename: [\"special-.*-node-.*\"]\n"
                  },
                  "operand": {
                    "image": "quay.io/openshift/origin-node-feature-discovery:4.10",
                    "servicePort": 12000
                  },
                  "workerConfig": {
                    "configData": "core:\n#  labelWhiteList:\n#  noPublish: false\n  sleepInterval: 60s\n#  sources: [all]\n#  klog:\n#    addDirHeader: false\n#    alsologtostderr: false\n#    logBacktraceAt:\n#    logtostderr: true\n#    skipHeaders: false\n#    stderrthreshold: 2\n#    v: 0\n#    vmodule:\n##   NOTE: the following options are not dynamically run-time \n##          configurable and require a nfd-worker restart to take effect\n##          after being changed\n#    logDir:\n#    logFile:\n#    logFileMaxSize: 1800\n#    skipLogHeaders: false\nsources:\n#  cpu:\n#    cpuid:\n##     NOTE: whitelist has priority over blacklist\n#      attributeBlacklist:\n#        - \"BMI1\"\n#        - \"BMI2\"\n#        - \"CLMUL\"\n#        - \"CMOV\"\n#        - \"CX16\"\n#        - \"ERMS\"\n#        - \"F16C\"\n#        - \"HTT\"\n#        - \"LZCNT\"\n#        - \"MMX\"\n#        - \"MMXEXT\"\n#        - \"NX\"\n#        - \"POPCNT\"\n#        - \"RDRAND\"\n#        - \"RDSEED\"\n#        - \"RDTSCP\"\n#        - \"SGX\"\n#        - \"SSE\"\n#        - \"SSE2\"\n#        - \"SSE3\"\n#        - \"SSE4.1\"\n#        - \"SSE4.2\"\n#        - \"SSSE3\"\n#      attributeWhitelist:\n#  kernel:\n#    kconfigFile: \"/path/to/kconfig\"\n#    configOpts:\n#      - \"NO_HZ\"\n#      - \"X86\"\n#      - \"DMI\"\n  pci:\n    deviceClassWhitelist:\n      - \"0200\"\n      - \"03\"\n      - \"12\"\n    deviceLabelFields:\n#      - \"class\"\n      - \"vendor\"\n#      - \"device\"\n#      - \"subsystem_vendor\"\n#      - \"subsystem_device\"\n#  usb:\n#    deviceClassWhitelist:\n#      - \"0e\"\n#      - \"ef\"\n#      - \"fe\"\n#      - \"ff\"\n#    deviceLabelFields:\n#      - \"class\"\n#      - \"vendor\"\n#      - \"device\"\n#  custom:\n#    - name: \"my.kernel.feature\"\n#      matchOn:\n#        - loadedKMod: [\"example_kmod1\", \"example_kmod2\"]\n#    - name: \"my.pci.feature\"\n#      matchOn:\n#        - pciId:\n#            class: [\"0200\"]\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#        - pciId :\n#            vendor: [\"8086\"]\n#            device: [\"1000\", \"1100\"]\n#    - name: \"my.usb.feature\"\n#      matchOn:\n#        - usbId:\n#          class: [\"ff\"]\n#          vendor: [\"03e7\"]\n#          device: [\"2485\"]\n#        - usbId:\n#          class: [\"fe\"]\n#          vendor: [\"1a6e\"]\n#          device: [\"089a\"]\n#    - name: \"my.combined.feature\"\n#      matchOn:\n#        - pciId:\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#          loadedKMod : [\"vendor_kmod1\", \"vendor_kmod2\"]\n"
                  }
                }
              },
              {
                "apiVersion": "nfd.openshift.io/v1alpha1",
                "kind": "NodeFeatureRule",
                "metadata": {
                  "name": "my-sample-rule-object"
                },
                "spec": {
                  "rules": [
                    {
                      "labels": {
                        "my-sample-feature": "true"
                      },
                      "matchFeatures": [
                        {
                          "feature": "kernel.loadedmodule",
                          "matchExpressions": {
                            "dummy": {
                              "op": "Exists"
                            }
                          }
                        },
                        {
                          "feature": "kernel.config",
                          "matchExpressions": {
                            "X86": {
                              "op": "In",
                              "value": [
                                "y"
                              ]
                            }
                          }
                        }
                      ],
                      "name": "my sample rule"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery,OpenShift Optional
          containerImage: quay.io/openshift/origin-cluster-nfd-operator:4.10
          description: |-
            The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

            NFD consists  of the following software components:

            The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

            ##NFD-Master
            NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

            ##NFD-Worker
            NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

            ##NFD-Topology-Updater
            NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
          olm.skipRange: '>=4.6.0 <4.11.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-nfd
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/internal-objects: '["noderesourcetopologies.topology.node.k8s.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/cluster-nfd-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              The NodeFeatureDiscovery instance is the CustomResource being watched by the NFD-Operator, and holds all the needed information to setup the behaviour of the master and worker pods
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
          - description: |
              NodeFeatureRule resource specifies a configuration for feature-based customization of node objects, such as node labeling.
            kind: NodeFeatureRule
            name: nodefeaturerules.nfd.openshift.io
            version: v1alpha1
          - description: |
              NodeResourceTopology resource describes node resources and their topology.
            kind: NodeResourceTopology
            name: noderesourcetopologies.topology.node.k8s.io
            version: v1alpha1
        description: |-
          The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

          NFD consists  of the following software components:

          The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

          ##NFD-Master
          NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

          ##NFD-Worker
          NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

          ##NFD-Topology-Updater
          NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
        displayName: Node Feature Discovery Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - feature-discovery
        - feature-detection
        - node-labels
        links:
        - name: Node Feature Discovery Operator
          url: https://docs.openshift.com/container-platform/4.9/hardware_enablement/psap-node-feature-discovery-operator.html
        - name: Node Feature Discovery Documentation
          url: https://kubernetes-sigs.github.io/node-feature-discovery/stable/get-started/index.html
        maintainers:
        - email: openshift-psap@redhat.com
          name: openshift-psap
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/cluster-nfd-operator
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy
        - quay.io/openshift/origin-cluster-nfd-operator:4.10
        version: 4.10.0
      entries:
      - name: nfd.v4.10.0
        version: 4.10.0
      - name: nfd.v4.9.0
        version: 4.9.0
      - name: nfd.v4.8.0
        version: 4.8.0
      name: stable
    defaultChannel: stable
    packageName: openshift-nfd-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/cluster-nfd-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-bedrockagent-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-bedrockagent-controller.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bedrockagent.services.k8s.aws/v1alpha1",
                "kind": "Agent",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/bedrockagent-controller:0.1.1
          createdAt: "2025-09-20T01:18:56Z"
          description: AWS Bedrock Agent controller is a service controller for managing
            Bedrock Agent resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Agent represents the state of an AWS bedrockagent Agent resource.
            displayName: Agent
            kind: Agent
            name: agents.bedrockagent.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Bedrock Agent resources in AWS from within your Kubernetes cluster.

          **About Amazon Bedrock Agent**

          Amazon Bedrock is a fully managed service that makes high-performing foundation models (FMs) from leading AI companies and Amazon available for your use through a unified API. You can choose from a wide range of foundation models to find the model that is best suited for your use case. Amazon Bedrock also offers a broad set of capabilities to build generative AI applications with security, privacy, and responsible AI. Using Amazon Bedrock, you can easily experiment with and evaluate top foundation models for your use cases, privately customize them with your data using techniques such as fine-tuning and Retrieval Augmented Generation (RAG), and build agents that execute tasks using your enterprise systems and data sources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Bedrock Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bedrockagent
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Bedrock Agent Developer Resources
          url: https://aws.amazon.com/bedrock/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: bedrock-agent maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/bedrockagent-controller:0.1.1
        version: 0.1.1
      entries:
      - name: ack-bedrockagent-controller.v0.1.1
        version: 0.1.1
      - name: ack-bedrockagent-controller.v0.1.0
        version: 0.1.0
      - name: ack-bedrockagent-controller.v0.0.6
        version: 0.0.6
      - name: ack-bedrockagent-controller.v0.0.5
        version: 0.0.5
      - name: ack-bedrockagent-controller.v0.0.4
        version: 0.0.4
      - name: ack-bedrockagent-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-bedrockagent-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: devopstales
      provider-url: devopstales.github.io
    name: community-trivy-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: community-trivy-operator.v2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "trivy-operator.devopstales.io/v1",
                "kind": "NamespaceScanner",
                "metadata": {
                  "name": "trivy-operator-main-config"
                },
                "spec": {
                  "crontab": "*/5 * * * *",
                  "namespace_selector": "trivy-scan",
                  "clusterWide": "False",
                  "policyreport": "False",
                  "registry": [
                    {
                      "name": "docker.io",
                      "user": "",
                      "password": ""
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: devopstales/trivy-operator:2.4.1
          createdAt: "2022-07-11T15:00:00Z"
          description: Trivy Operator for scheduled imagescans and an Admission Control.
          repository: https://github.com/devopstales/trivy-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Confifuration for trivy-operator.
            displayName: NamespaceScanner
            kind: NamespaceScanner
            name: namespace-scanners.trivy-operator.devopstales.io
            version: v1
          - description: |
              VulnerabilityReport summarizes vulnerabilities in application dependencies and operating system packages
              built into container images.
            displayName: VulnerabilityReport
            kind: VulnerabilityReport
            name: vulnerabilityreports.trivy-operator.devopstales.io
            version: v1
        description: |
          Trivy Operator for scheduled imagescans and an Admission Control.

          ### Add Github token to extend your Github API Rate Limit
          The Operator use Github to download Trivy database. If your Github API Rate Limit Exceeded add your Github token to the operator deployment as GITHUB_TOKEN variable.

          ```yaml
          ...
          env:
          - name: GITHUB_TOKEN
            value: "githubToken"
          ```

          ### Scheduled Image scans
          Default every 5 minutes execute a scan script. It will get image list from all namespaces with the label `trivy-scan=true`, and then scan this images with trivy, finally we will get metrics on `http://[pod-ip]:9115/metrics`

          You can chane the schedule of the image scanning by setting the `crontab` variable in the namespace-scanner object.


          ### Cache Scan results
          From the 2.4 version you can use an external redis to store scanning results and speed up the scanning process.

          ```bash
          ...
          env:
          - name: REDIS_ENABLED
            value: "True"
          - name: REDIS_BACKEND
            value: redis://redis-cache:6379
          ```

          ### Persist Trivy database cache
          To Persist Trivy cache you need to create a pvc and mount to the Deployment under `/home/trivy-operator/trivy-cache`

          ```yaml
          ...
            volumeMounts:
            - name: cache
              mountPath: "/home/trivy-operator/trivy-cache"
          volumes:
          - name: cache
            persistentVolumeClaim:
              claimName: trivy-cache
          ```

          ### Trivy Image Validator
          The admission controller function can be configured as a ValidatingWebhook in a k8s cluster. Kubernetes will send requests to the admission server when a Pod creation is initiated. The admission controller checks the image using trivy if it is in a namespace with the lable `trivy-operator-validation=true`.

          To enable the admission controller you need to set the `clusterWide` variable to `true` in the namespace-scanner object.

          Enable the scheduled scan and the admission controller for a specific namespace:
          ```bash
          kubectl label namespaces test-apps trivy-scan=true
          # or
          kubectl label namespaces test-apps trivy-operator-validation=true
          ```

          You can define policy to the Admission Controller, by adding annotation to the pod trough the deployment:

          ```bash
          spec:
            ...
            template:
              metadata:
                annotations:
                  trivy.security.devopstales.io/medium: "5"
                  trivy.security.devopstales.io/low: "10"
                  trivy.security.devopstales.io/critical: "2"
          ...
          ```

          ### Monitorng
          Trivy operator has a prometheus endpoint wher yo can get the resoults of the namespace scans and the adission controller.

          ```bash
          curl -s http://10.43.179.39:9115/metrics | grep so_vulnerabilities

          # HELP so_vulnerabilities Container vulnerabilities
          # TYPE so_vulnerabilities gauge
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="UNKNOWN"} 0
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="LOW"} 23
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="MEDIUM"} 93
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="HIGH"} 76
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="CRITICAL"} 25
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="UNKNOWN"} 0
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="LOW"} 23
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="MEDIUM"} 88
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="HIGH"} 60
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="CRITICAL"} 8
          ```

          ```bash
          curl -s http://10.43.179.39:9115/metrics | grep ac_vulnerabilities

          # HELP ac_vulnerabilities Admission Controller vulnerabilities
          # TYPE ac_vulnerabilities gauge
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="UNKNOWN"} 0.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="LOW"} 83.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="MEDIUM"} 6.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="HIGH"} 6.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="CRITICAL"} 4.0
          ```

          ### Policy Report reation
          The PolicyReport object is a protopype object probosed by the Kubernetes policy work group. The Policy Report Custom Resource Definition (CRD) can be used as a common way to provide policy results to Kubernetes cluster administrators and users, using native tools.

          Add the PolicyReport CRDs to your cluster (v1alpha2):
          ```bash
          kubectl create -f https://github.com/kubernetes-sigs/wg-policy-prototypes/raw/master/policy-report/crd/v1alpha2/wgpolicyk8s.io_policyreports.yaml
          ```

          You can enable the Policy Report creatin by setting the `policyreport` to `True` in the NamespaceScanner object.

          This objects can be visualized by the Policy Reporter UI. The Policy Reporter UI is a monitoring and Observability Tool for the PolicyReport CRD with an optional UI. It is created by Kyverno. The main goal was a tool to visualize the resoluts of the Kyverno policies, but because it uses the PolicyReports CRD it can visualize the resoults of the trivy-operator scans.
        displayName: Community Trivy Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trivy
        - security
        links:
        - name: Documentation
          url: https://devopstales.github.io/trivy-operator
        - name: Blog
          url: https://devopstales.github.io
        maintainers:
        - email: devopstales@protonmail.com
          name: devopstales
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: devopstales
          url: devopstales.github.io
        relatedImages:
        - devopstales/trivy-operator:2.4.1
        version: 2.4.0
      entries:
      - name: community-trivy-operator.v2.4.0
        version: 2.4.0
      - name: community-trivy-operator.v2.3.0
        version: 2.3.0
      - name: community-trivy-operator.v2.2.0
        version: 2.2.0
      - name: community-trivy-operator.v2.1.1
        version: 2.1.1
      name: stable
    defaultChannel: stable
    packageName: community-trivy-operator
    provider:
      name: devopstales
      url: devopstales.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: www.ibm.com
    name: ibm-elasticsearch-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-elasticsearch-operator.v1.1.2667
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
          createdAt: "2025-01-30"
          description: Elasticsearch operator for use within Cloudpak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-elasticsearch-operator
          olm.skipRange: <0.0.0 >1.1.0
          operators.openshift.io/valid-subscription: '["IBM Elasticsearch"]'
          operators.operatorframework.io/internal-objects: '["elasticsearchclusters.elasticsearch.opencontent.ibm.com"]'
          operators.operatorframework.io/operator-type: non-standalone
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ElasticsearchCluster
            name: elasticsearchclusters.elasticsearch.opencontent.ibm.com
            version: v1
        description: |-
          # Introduction
          An Elasticsearch operator for use with IBM Cloudpak services.
          # Details
          ibm-elasticsearch-operator supports installing Elasticsearch major version 6 and 7. This operator supports two CRDs Elasticsearch and Elasticsearchdetails. Currently Elasticsearch is an abstraction of Elasticsearchdetails, but there is no difference between both the CRDs. Creating either of them provisions elasticsearch instances. This abstraction layer has been created to support any changes to the operator with out breaking the CRD/API for adopting products.

          ## Prerequisites
          - Openshift 3.11+
          ## Resources Required
          In short the answer is it depends. Especially if you are running Elasticsearch 6 you need 3 or a greater odd number of replicas to make a cluster and not have split brain. Each replica will have a sidecar that needs 0.1 cpu and 2gb of ram. You should size Elasticsearch based on your use-case. The default is a full cpu and 4 cpu and 64 gb ram. To assist sizing requirements we have included a copy of [esrally](https://esrally.readthedocs.io/en/stable/) on all the docker images.
          ## Installing
          ElasticsearchCluster is the strategic Kind moving forward. There are two flavors of CRD supported that will soon be deprecated. A simple Elasticseach "All in one" crd and a more detailed CRD for teams who need greater control. Upon deploy the elasticsearch is available from the service (not the headless) created by the operator. Only https will be available on port 443. the service will also be in the ElasticsearchDetails status.

          ### Case install
          To install the component using case you must either use the native install or catalog install. Upon installing the catalog you can then use either OLM or ODLM to install but must create the OperatorGroup & Subscription for either method to avoid conflict.

          # Configuration
          Architecture:
            - x86
            - s390x
            - ppc64le
          Storage: RWO Encrypted Block Storage with 120+ IOPS/GB
          Clouds:
          - IBM CLoud (ibm-block-gold, Gen2 custom iops storage class)
          - AWS (EBS)
          - Azure (For Azure please ensure you are using a higher tier of AzureDisk storage)
          - OCS (RBD not Ceph)
          - Portworx (Non-shared replacated)
          - Please do not use with NFS

          ### Limitations
          This product is not deployable standalone and it is part of the LICCR already accepted
          ## SecurityContextConstraints Requirements
          This component requires a restricted SCC.
        displayName: IBM OpenContent Elasticsearch
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - internal
        - Elasticsearch
        maintainers:
        - email: noreply@ibm.com
          name: IBM CloudpakOpen Community
        maturity: alpha
        provider:
          name: IBM
          url: www.ibm.com
        relatedImages:
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:47a3745c78280f98f2ee352bb9192cdcd2fcd6d4c1a3627e7aa42178a0f107fd
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:30a5b2c4e3c231373fa18d8aaab16d3505cc11a8a5d4be36eaf99d7c0dee35f2
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:2668a562eb8df1117c40acbc3600d6735469d210e5aae4aa03bf9f54d0ebfe46
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:2c39ff1850b7a9d32a11e99393039ef74ab363f3fffec1e4fd542f2ae148a6b4
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:130872119947fe91cbfb8dd82b3baaf78f7adcaca5933ca394eca3f9ee84c8cc
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
        version: 1.1.2667
      entries:
      - name: ibm-elasticsearch-operator.v1.1.2667
        version: 1.1.2667
      name: v1.1
    defaultChannel: v1.1
    packageName: ibm-elasticsearch-operator
    provider:
      name: IBM
      url: www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: ovms-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openvino-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "ModelServer",
                "metadata": {
                  "name": "model-server-sample"
                },
                "spec": {
                  "image_name": "registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0",
                  "deployment_parameters": {
                    "replicas": 1
                  },
                  "service_parameters": {
                    "grpc_port": 8080,
                    "rest_port": 8081
                  },
                  "models_settings": {
                    "single_model_mode": true,
                    "config_configmap_name": "",
                    "config_path": "",
                    "model_name": "resnet",
                    "model_path": "gs://<bucket_name>/<model_dir>",
                    "nireq": 0,
                    "plugin_config": "{\"PERFORMANCE_HINT\":\"LATENCY\"}",
                    "batch_size": "",
                    "shape": "",
                    "model_version_policy": "{\"latest\": { \"num_versions\":1 }}",
                    "layout": "",
                    "target_device": "CPU",
                    "is_stateful": false,
                    "idle_sequence_cleanup": false,
                    "low_latency_transformation": true,
                    "max_sequence_number": 0
                  },
                  "server_settings": {
                    "file_system_poll_wait_seconds": 0,
                    "sequence_cleaner_poll_wait_minutes": 0,
                    "log_level": "INFO",
                    "grpc_workers": 1,
                    "rest_workers": 0
                  },
                  "models_repository": {
                    "https_proxy": "",
                    "http_proxy": "",
                    "storage_type": "google",
                    "models_host_path": "",
                    "models_volume_claim": "",
                    "aws_secret_access_key": "",
                    "aws_access_key_id": "",
                    "aws_region": "",
                    "s3_compat_api_endpoint": "",
                    "gcp_creds_secret_name": "",
                    "azure_storage_connection_string": ""
                  }
                }
              },
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "Notebook",
                "metadata": {
                  "name": "v2024.1"
                },
                "spec": {
                  "git_ref": "2024.1",
                  "auto_update_image": false
                }
              }
            ]
          alm-examples-metadata: '{ "model-server-sample":{"description":"Example
            of the model server configuration"}, "v2024":{"description":"Default notebook
            configuration"} }'
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
          createdAt: "2024-05-14T09:29:35Z"
          description: OpenVINO Toolkit Operator manages OpenVINO components in OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.34.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Provides inference serving through gRPC & REST
            displayName: Model Server
            kind: ModelServer
            name: modelservers.intel.com
            version: v1alpha1
          - description: Adds a collection of ready-to-run Jupyter notebooks for learning
              and experimenting with the OpenVINO™ Toolkit to Red Hat® OpenShift®
              Data Science
            displayName: Notebook
            kind: Notebook
            name: notebooks.intel.com
            version: v1alpha1
        description: |
          OpenVINO Toolkit Operator manages OpenVINO components in OpenShift.

          Currently there available components are ModelServer and Notebook.

          # Model Server
          [OpenVINO™ Model Server](https://github.com/openvinotoolkit/model_server) (OVMS) is a scalable, high-performance solution for serving machine learning models optimized for Intel® architectures. The server provides an inference service via gRPC or REST API for any models trained in a framework that is supported by [OpenVINO](https://software.intel.com/en-us/openvino-toolki://docs.openvino.ai/latest/index.html).
          Model Server configuration parameters is explained [here](https://github.com/openvinotoolkit/operator/blob/main/docs/modelserver_params.md).
          ## Using the cluster
          OpenVINO Model Server can be consumed as a `Service`. It is called like with `ModelServer` resource with `-ovms` suffix.
          The suffix is omitted when `ovms` phrase is included in the name.
          The service exposes gRPC and REST API interfaces to run inference requests.
          ```
          oc get pods
          NAME                                        READY   STATUS    RESTARTS   AGE
          model-server-sample-ovms-586f6f76df-dpps4   1/1     Running   0          8h

          oc get services
          NAME                       TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)             AGE
          model-server-sample-ovms   ClusterIP   172.25.199.210   <none>        8080/TCP,8081/TCP   8h
          ```

          # Notebook
          The Notebook resource integrates JupyterHub from OpenShift Data Science or Open Data Hub with a container image that includes [developer
          tools](https://github.com/openvinotoolkit/openvino/blob/master/docs/install_guides/pypi-openvino-dev.md)
          from the OpenVINO toolkit and a set of Jupyter notebook tutorials. It enables selecting a defined image `OpenVINO™ Toolkit` from
          the Jupyter Spawner choice list.

          Create the `Notebook` resource in the same project with JupyterHub and RedHat OpenShift Data Science operator as seen in [documentation](https://github.com/openvinotoolkit/operator/blob/main/docs/notebook_in_rhods.md).
          It builds the image in the cluster based on Dockerfile from [openvino_notebooks](https://github.com/openvinotoolkit/openvino_notebooks).

          ## References
          OpenVINO Model Server on [Github](https://github.com/openvinotoolkit/model_server)

          OpenVINO Model Server Operator on [Github](https://github.com/openvinotoolkit/operator)
        displayName: OpenVINO Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Inference OpenVINO
        links:
        - name: OpenVINO Toolkit Operator
          url: https://github.com/openvinotoolkit/operator
        - name: OpenVINO Model Server
          url: https://github.com/openvinotoolkit/model_server
        - name: OpenVINO Notebooks
          url: https://github.com/openvinotoolkit/openvino_notebooks
        maintainers:
        - email: openvino-operator-support@intel.com
          name: Operator Support
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5714b93ac39277affdc10c3ba142d8c032ed0ddb2ba61c04cbddac710464c4
        - registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
        - registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0
        version: 1.2.0
      entries:
      - name: openvino-operator.v1.2.0
        version: 1.2.0
      - name: openvino-operator.v1.1.0
        version: 1.1.0
      - name: openvino-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: ovms-operator
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx-openshift-operator.v0.0.33
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
          createdAt: "2019-07-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx-operator-rhmp/support?utm_source=openshift_console
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX on JBoss EAP 7
            displayName: Joget DX on JBoss EAP 7
            kind: JogetDX
            name: jogetdx.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX7/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
        - registry.connect.redhat.com/joget/joget-dx7-eap7@sha256:7245501f5fccf011d75c034b7d55542cfe3bd2c64cc7e48b3224a48b623af282
        version: 0.0.33
      entries:
      - name: joget-dx-openshift-operator.v0.0.33
        version: 0.0.33
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx-operator-rhmp
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-codeartifact-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-codeartifact-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/codeartifact-controller:1.1.1
          createdAt: "2025-09-19T22:00:24Z"
          description: AWS codeartifact controller is a service controller for managing
            codeartifact resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Domain represents the state of an AWS codeartifact Domain
              resource.
            displayName: Domain
            kind: Domain
            name: domains.codeartifact.services.k8s.aws
            version: v1alpha1
          - description: PackageGroup represents the state of an AWS codeartifact
              PackageGroup resource.
            displayName: PackageGroup
            kind: PackageGroup
            name: packagegroups.codeartifact.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon codeartifact resources in AWS from within your Kubernetes cluster.

          **About Amazon codeartifact**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon codeartifact
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - codeartifact
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon codeartifact Developer Resources
          url: https://aws.amazon.com/codeartifact/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: codeartifact maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/codeartifact-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-codeartifact-controller.v1.1.1
        version: 1.1.1
      - name: ack-codeartifact-controller.v1.1.0
        version: 1.1.0
      - name: ack-codeartifact-controller.v1.0.13
        version: 1.0.13
      - name: ack-codeartifact-controller.v1.0.11
        version: 1.0.11
      name: alpha
    defaultChannel: alpha
    packageName: ack-codeartifact-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kepler Operator Contributors
      provider-url: https://sustainable-computing.io/
    name: kepler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kepler-operator.v0.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "Kepler",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "kepler",
                    "app.kubernetes.io/name": "kepler",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "kepler"
                },
                "spec": {
                  "exporter": {
                    "deployment": {
                      "port": 9103
                    },
                    "redfish": null
                  }
                }
              },
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "PowerMonitor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "powermonitor",
                    "app.kubernetes.io/name": "powermonitor",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "power-monitor"
                },
                "spec": {
                  "kepler": {
                    "config": {
                      "logLevel": "info"
                    },
                    "deployment": {
                      "security": {
                        "mode": "none"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: quay.io/sustainable_computing_io/kepler-operator:0.20.0
          createdAt: "2025-07-18T06:13:30Z"
          description: Deploys and Manages Kepler on Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: |-
            [
              "keplerinternals.kepler.system.sustainable.computing.io",
              "powermonitorinternals.kepler.system.sustainable.computing.io"
            ]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/sustainable-computing-io/kepler-operator
          support: https://github.com/sustainable-computing-io/kepler-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KeplerInternal is the Schema for internal/unsupported API
            displayName: KeplerInternal
            kind: KeplerInternal
            name: keplerinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: Kepler is deprecated and will be removed in a future release.
              Please use PowerMonitor instead.
            displayName: Kepler
            kind: Kepler
            name: keplers.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitorInternal is the Schema for internal/unsupported
              API
            displayName: PowerMonitorInternal
            kind: PowerMonitorInternal
            name: powermonitorinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitor is the Schema for the powermonitors API
            displayName: PowerMonitor
            kind: PowerMonitor
            name: powermonitors.kepler.system.sustainable.computing.io
            version: v1alpha1
        description: |-
          Kepler (Kubernetes-based Efficient Power Level Exporter) is a Prometheus exporter that
          measures energy consumption metrics at the container, pod, and node level in Kubernetes clusters.

          ### Deprecation Notice
          **Important:** The Kepler CR (Custom Resource) is deprecated and will be removed in a future release. Please use the new PowerMonitor instead.

          ### Prerequisites
          To view the metrics exposed by Kepler, **[User Workload Monitoring](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/monitoring/configuring-user-workload-monitoring)**
          MUST be enabled in your cluster.

          ### Documentation
          * [User Documentation](https://github.com/sustainable-computing-io/kepler-operator/blob/v1alpha1/README.md)
          * [Developer Guide](https://github.com/sustainable-computing-io/kepler-operator/tree/v1alpha1/docs/developer)

          ### Contributing
          * [Report Bugs](https://github.com/sustainable-computing-io/kepler-operator/issues/new?template=bug_report.md) - Found a bug? Let us know!
          * [Browse Issues](https://github.com/sustainable-computing-io/kepler-operator/issues) - See what we're working on

          ### License
          Licensed under [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)
        displayName: Kepler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kepler
        - power
        - energy
        - monitoring
        links:
        - name: Installation Guide
          url: https://sustainable-computing.io/installation/community-operator/
        - name: Kepler Project
          url: https://sustainable-computing.io/
        maintainers:
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: vprashar@redhat.com
          name: Vibhu Prashar
        - email: vimalkum@redhat.com
          name: Vimal Kumar
        - email: kaliu@redhat.com
          name: Kaiyi Liu
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Kepler Operator Contributors
          url: https://sustainable-computing.io/
        relatedImages:
        - quay.io/sustainable_computing_io/kepler-operator:0.20.0
        - quay.io/sustainable_computing_io/kepler:v0.10.2
        - quay.io/brancz/kube-rbac-proxy:v0.19.0
        version: 0.20.0
      entries:
      - name: kepler-operator.v0.20.0
        version: 0.20.0
      - name: kepler-operator.v0.19.0
        version: 0.19.0
      - name: kepler-operator.v0.18.0
        version: 0.18.0
      - name: kepler-operator.v0.17.0
        version: 0.17.0
      - name: kepler-operator.v0.16.0
        version: 0.16.0
      - name: kepler-operator.v0.15.0
        version: 0.15.0
      name: alpha
    defaultChannel: alpha
    packageName: kepler-operator
    provider:
      name: Kepler Operator Contributors
      url: https://sustainable-computing.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MuleSoft
      provider-url: anypoint.mulesoft.com
    name: runtime-fabric-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runtime-fabric-operator.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rtf.operators.mulesoft.com/v1",
                "kind": "RuntimeFabric",
                "metadata": {
                  "name": "runtime-fabric"
                },
                "spec": {
                  "activationData": null,
                  "agent": {
                    "replicas": 1,
                    "app": {
                      "resources": {
                        "limits": {
                          "cpu": "1000m",
                          "memory": "500Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      }
                    },
                    "rtfd": {
                      "resources": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "150m",
                          "memory": "200Mi"
                        }
                      },
                      "watch": {
                        "enabled": true,
                        "components": {
                          "deployments": {
                            "enabled": true
                          },
                          "ingress": {
                            "enabled": true
                          },
                          "secrets": {
                            "enabled": true
                          },
                          "persistencegateways": {
                            "enabled": true
                          },
                          "nodes": {
                            "enabled": true
                          },
                          "jobs": {
                            "enabled": true
                          },
                          "routes": {
                            "enabled": true
                          }
                        }
                      }
                    }
                  },
                  "resource-cache": {
                    "replicas": 1,
                    "nginx": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "250Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    },
                    "fetcher": {
                      "resources": {
                        "limits": {
                          "cpu": "50m",
                          "memory": "100Mi"
                        },
                        "requests": {
                          "cpu": "50m",
                          "memory": "50Mi"
                        }
                      }
                    }
                  },
                  "mule-clusterip-service": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "global": {
                    "cluster": {
                      "flavor": "byok",
                      "type": "openshift"
                    },
                    "containerLogPaths": [
                      "/var/lib/docker/containers",
                      "/var/log/containers",
                      "/var/log/pods"
                    ],
                    "core": {
                      "affinity": {},
                      "tolerations": [],
                      "topologySpreadConstraints": [
                        {
                          "maxSkew": 1,
                          "topologyKey": "topology.kubernetes.io/zone",
                          "whenUnsatisfiable": "ScheduleAnyway"
                        }
                      ]
                    },
                    "worker": {
                      "affinity": {},
                      "tolerations": []
                    },
                    "deployment": {
                      "nodeSelector": ""
                    },
                    "image": {
                      "provider": "aws",
                      "pullSecretName": "rtf-pull-secret",
                      "rtfRegistry": "rtf-runtime-registry.kprod.msap.io",
                      "repositoryNamespace": "mulesoft"
                    },
                    "service": {
                      "clusterIp": null
                    },
                    "authorizedNamespaces": false,
                    "fipsEnabled": false,
                    "crds": {
                      "install": true,
                      "keep": true,
                      "flex": {
                        "enabled": false
                      }
                    },
                    "nodeWatcherEnabled": true,
                    "nodeReconciliation": true,
                    "deploymentRateLimitPerSecond": 1
                  },
                  "muleLicense": null,
                  "proxy": {
                    "http_no_proxy": null,
                    "http_proxy": null,
                    "monitoring_proxy": null
                  },
                  "customLog4jEnabled": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
          createdAt: "2025-07-31T21:12:41Z"
          description: Runtime Fabric Operator Application from Mulesoft
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mulesoft/rtf-ocp-operator
          support: MuleSoft
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Runtime Fabric Operator from Mulesoft
              **Activation data**
              Used to install Runtime Fabric on a Kubernetes service. Copy this value from the Runtime Fabric setup page in Runtime Manager.
              **Registry URL**
              The registry from which to retrieve the images needed to install and run Runtime Fabric. If you're using a local registry, specify the URL here.
                * Default registry URLs:
                    - US control plane: rtf-runtime-registry.kprod.msap.io
                    - EU control plane: rtf-runtime-registry.kprod-eu.msap.io
              **Registry pull secret**
              Used to retrieve images from the registry. Change this value only if you did so during the operator installation.
              **Mule license key**
              Your company’s Mule license key. The key is not required during install, but it is required to deploy applications to Runtime Fabric. Must be Base64 encoded.
            displayName: Runtime Fabric
            kind: RuntimeFabric
            name: runtimefabrics.rtf.operators.mulesoft.com
            version: v1
        description: |-
          Anypoint Runtime Fabric is a container service that you install in your existing Red Hat OpenShift cluster that enables you to run Mule applications, composite APIs, and API gateways across any environment with centralized management via Anypoint Platform.
          Capabilities of Runtime Fabric include:
          * Isolation between applications by running a separate Mule runtime server per application.
          * Ability to run multiple versions of Mule runtime server on the same set of resources.
          * Scaling applications across multiple replicas.
          * Automated application fail-over.
          * Application management with Anypoint Runtime Manager.
          Runtime Fabric supports the following Red Hat OpenShift deployment options:
          * Red Hat OpenShift Service on AWS
          * Microsoft Azure Red Hat OpenShift
          * Red Hat OpenShift Dedicated
          * Red Hat OpenShift on IBM Cloud
          * Self-managed Red Hat OpenShift editions (Performance Plus, OCP, Kubernetes engine)
          Runtime Fabric supports Red Hat OpenShift versions 4.16 to 4.19.
        displayName: Runtime Fabric Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - agent
        - rtf
        - runtime
        - fabric
        - mulesoft
        links:
        - name: Runtime Fabric Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/62901512fefb1b137ba23946
        - name: Documentation
          url: https://docs.mulesoft.com/release-notes/runtime-fabric/runtime-fabric
        maintainers:
        - email: support@mulesoft.com
          name: Mulesoft
        maturity: stable
        provider:
          name: MuleSoft
          url: anypoint.mulesoft.com
        relatedImages:
        - quay.io/rtf/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
        version: 2.11.0
      entries:
      - name: runtime-fabric-operator.v2.11.0
        version: 2.11.0
      - name: runtime-fabric-operator.v2.10.34
        version: 2.10.34
      - name: runtime-fabric-operator.v2.10.17
        version: 2.10.17
      - name: runtime-fabric-operator.v2.10.0
        version: 2.10.0
      - name: runtime-fabric-operator.v2.9.2
        version: 2.9.2
      - name: runtime-fabric-operator.v2.8.2
        version: 2.8.2
      - name: runtime-fabric-operator.v2.7.0
        version: 2.7.0
      - name: runtime-fabric-operator.v2.6.52
        version: 2.6.52
      name: alpha
    - currentCSV: runtime-fabric-operator.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rtf.operators.mulesoft.com/v1",
                "kind": "RuntimeFabric",
                "metadata": {
                  "name": "runtime-fabric"
                },
                "spec": {
                  "activationData": null,
                  "agent": {
                    "replicas": 1,
                    "app": {
                      "resources": {
                        "limits": {
                          "cpu": "1000m",
                          "memory": "500Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      }
                    },
                    "rtfd": {
                      "resources": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "150m",
                          "memory": "200Mi"
                        }
                      },
                      "watch": {
                        "enabled": true,
                        "components": {
                          "deployments": {
                            "enabled": true
                          },
                          "ingress": {
                            "enabled": true
                          },
                          "secrets": {
                            "enabled": true
                          },
                          "persistencegateways": {
                            "enabled": true
                          },
                          "nodes": {
                            "enabled": true
                          },
                          "jobs": {
                            "enabled": true
                          },
                          "routes": {
                            "enabled": true
                          }
                        }
                      }
                    }
                  },
                  "resource-cache": {
                    "replicas": 1,
                    "nginx": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "250Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    },
                    "fetcher": {
                      "resources": {
                        "limits": {
                          "cpu": "50m",
                          "memory": "100Mi"
                        },
                        "requests": {
                          "cpu": "50m",
                          "memory": "50Mi"
                        }
                      }
                    }
                  },
                  "mule-clusterip-service": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "global": {
                    "cluster": {
                      "flavor": "byok",
                      "type": "openshift"
                    },
                    "containerLogPaths": [
                      "/var/lib/docker/containers",
                      "/var/log/containers",
                      "/var/log/pods"
                    ],
                    "core": {
                      "affinity": {},
                      "tolerations": [],
                      "topologySpreadConstraints": [
                        {
                          "maxSkew": 1,
                          "topologyKey": "topology.kubernetes.io/zone",
                          "whenUnsatisfiable": "ScheduleAnyway"
                        }
                      ]
                    },
                    "worker": {
                      "affinity": {},
                      "tolerations": []
                    },
                    "deployment": {
                      "nodeSelector": ""
                    },
                    "image": {
                      "provider": "aws",
                      "pullSecretName": "rtf-pull-secret",
                      "rtfRegistry": "rtf-runtime-registry.kprod.msap.io",
                      "repositoryNamespace": "mulesoft"
                    },
                    "service": {
                      "clusterIp": null
                    },
                    "authorizedNamespaces": false,
                    "fipsEnabled": false,
                    "crds": {
                      "install": true,
                      "keep": true,
                      "flex": {
                        "enabled": false
                      }
                    },
                    "nodeWatcherEnabled": true,
                    "nodeReconciliation": true,
                    "deploymentRateLimitPerSecond": 1
                  },
                  "muleLicense": null,
                  "proxy": {
                    "http_no_proxy": null,
                    "http_proxy": null,
                    "monitoring_proxy": null
                  },
                  "customLog4jEnabled": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
          createdAt: "2025-07-31T21:12:41Z"
          description: Runtime Fabric Operator Application from Mulesoft
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mulesoft/rtf-ocp-operator
          support: MuleSoft
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Runtime Fabric Operator from Mulesoft
              **Activation data**
              Used to install Runtime Fabric on a Kubernetes service. Copy this value from the Runtime Fabric setup page in Runtime Manager.
              **Registry URL**
              The registry from which to retrieve the images needed to install and run Runtime Fabric. If you're using a local registry, specify the URL here.
                * Default registry URLs:
                    - US control plane: rtf-runtime-registry.kprod.msap.io
                    - EU control plane: rtf-runtime-registry.kprod-eu.msap.io
              **Registry pull secret**
              Used to retrieve images from the registry. Change this value only if you did so during the operator installation.
              **Mule license key**
              Your company’s Mule license key. The key is not required during install, but it is required to deploy applications to Runtime Fabric. Must be Base64 encoded.
            displayName: Runtime Fabric
            kind: RuntimeFabric
            name: runtimefabrics.rtf.operators.mulesoft.com
            version: v1
        description: |-
          Anypoint Runtime Fabric is a container service that you install in your existing Red Hat OpenShift cluster that enables you to run Mule applications, composite APIs, and API gateways across any environment with centralized management via Anypoint Platform.
          Capabilities of Runtime Fabric include:
          * Isolation between applications by running a separate Mule runtime server per application.
          * Ability to run multiple versions of Mule runtime server on the same set of resources.
          * Scaling applications across multiple replicas.
          * Automated application fail-over.
          * Application management with Anypoint Runtime Manager.
          Runtime Fabric supports the following Red Hat OpenShift deployment options:
          * Red Hat OpenShift Service on AWS
          * Microsoft Azure Red Hat OpenShift
          * Red Hat OpenShift Dedicated
          * Red Hat OpenShift on IBM Cloud
          * Self-managed Red Hat OpenShift editions (Performance Plus, OCP, Kubernetes engine)
          Runtime Fabric supports Red Hat OpenShift versions 4.16 to 4.19.
        displayName: Runtime Fabric Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - agent
        - rtf
        - runtime
        - fabric
        - mulesoft
        links:
        - name: Runtime Fabric Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/62901512fefb1b137ba23946
        - name: Documentation
          url: https://docs.mulesoft.com/release-notes/runtime-fabric/runtime-fabric
        maintainers:
        - email: support@mulesoft.com
          name: Mulesoft
        maturity: stable
        provider:
          name: MuleSoft
          url: anypoint.mulesoft.com
        relatedImages:
        - quay.io/rtf/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
        version: 2.11.0
      entries:
      - name: runtime-fabric-operator.v2.11.0
        version: 2.11.0
      - name: runtime-fabric-operator.v2.10.34
        version: 2.10.34
      - name: runtime-fabric-operator.v2.10.17
        version: 2.10.17
      - name: runtime-fabric-operator.v2.10.0
        version: 2.10.0
      - name: runtime-fabric-operator.v2.9.2
        version: 2.9.2
      - name: runtime-fabric-operator.v2.8.2
        version: 2.8.2
      - name: runtime-fabric-operator.v2.7.0
        version: 2.7.0
      - name: runtime-fabric-operator.v2.6.52
        version: 2.6.52
      name: stable
    defaultChannel: stable
    packageName: runtime-fabric-operator
    provider:
      name: MuleSoft
      url: anypoint.mulesoft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-dynamodb-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-dynamodb-controller.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "GlobalTable",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "Table",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/dynamodb-controller:1.5.2
          createdAt: "2025-09-19T19:19:47Z"
          description: AWS DynamoDB controller is a service controller for managing
            DynamoDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup represents the state of an AWS dynamodb Backup resource.
            displayName: Backup
            kind: Backup
            name: backups.dynamodb.services.k8s.aws
            version: v1alpha1
          - description: GlobalTable represents the state of an AWS dynamodb GlobalTable
              resource.
            displayName: GlobalTable
            kind: GlobalTable
            name: globaltables.dynamodb.services.k8s.aws
            version: v1alpha1
          - description: Table represents the state of an AWS dynamodb Table resource.
            displayName: Table
            kind: Table
            name: tables.dynamodb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon DynamoDB resources in AWS from within your Kubernetes cluster.

          **About Amazon DynamoDB**

          Amazon DynamoDB is a fully managed NoSQL database service that provides fast and predictable performance with seamless scalability. DynamoDB lets you offload the administrative burdens of operating and scaling a distributed database so that you don't have to worry about hardware provisioning, setup and configuration, replication, software patching, or cluster scaling. DynamoDB also offers encryption at rest, which eliminates the operational burden and complexity involved in protecting sensitive data. For more information, see [DynamoDB Encryption at Rest](https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/EncryptionAtRest.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon DynamoDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dynamodb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon DynamoDB Developer Resources
          url: https://aws.amazon.com/dynamodb/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: dynamodb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/dynamodb-controller:1.5.2
        version: 1.5.2
      entries:
      - name: ack-dynamodb-controller.v1.5.2
        version: 1.5.2
      - name: ack-dynamodb-controller.v1.5.1
        version: 1.5.1
      - name: ack-dynamodb-controller.v1.5.0
        version: 1.5.0
      - name: ack-dynamodb-controller.v1.4.10
        version: 1.4.10
      - name: ack-dynamodb-controller.v1.4.9
        version: 1.4.9
      - name: ack-dynamodb-controller.v1.4.7
        version: 1.4.7
      - name: ack-dynamodb-controller.v1.4.6
        version: 1.4.6
      - name: ack-dynamodb-controller.v1.4.5
        version: 1.4.5
      - name: ack-dynamodb-controller.v1.4.4
        version: 1.4.4
      - name: ack-dynamodb-controller.v1.4.3
        version: 1.4.3
      - name: ack-dynamodb-controller.v1.4.2
        version: 1.4.2
      - name: ack-dynamodb-controller.v1.4.1
        version: 1.4.1
      - name: ack-dynamodb-controller.v1.3.0
        version: 1.3.0
      - name: ack-dynamodb-controller.v1.2.23
        version: 1.2.23
      - name: ack-dynamodb-controller.v1.2.22
        version: 1.2.22
      - name: ack-dynamodb-controller.v1.2.21
        version: 1.2.21
      - name: ack-dynamodb-controller.v1.2.20
        version: 1.2.20
      - name: ack-dynamodb-controller.v1.2.15
        version: 1.2.15
      - name: ack-dynamodb-controller.v1.2.14
        version: 1.2.14
      - name: ack-dynamodb-controller.v1.2.13
        version: 1.2.13
      - name: ack-dynamodb-controller.v1.2.12
        version: 1.2.12
      - name: ack-dynamodb-controller.v1.2.11
        version: 1.2.11
      - name: ack-dynamodb-controller.v1.2.9
        version: 1.2.9
      - name: ack-dynamodb-controller.v1.2.8
        version: 1.2.8
      - name: ack-dynamodb-controller.v1.2.6
        version: 1.2.6
      - name: ack-dynamodb-controller.v1.2.5
        version: 1.2.5
      - name: ack-dynamodb-controller.v1.2.4
        version: 1.2.4
      - name: ack-dynamodb-controller.v1.2.2
        version: 1.2.2
      - name: ack-dynamodb-controller.v1.2.1
        version: 1.2.1
      - name: ack-dynamodb-controller.v1.2.0
        version: 1.2.0
      - name: ack-dynamodb-controller.v1.1.1
        version: 1.1.1
      - name: ack-dynamodb-controller.v1.1.0
        version: 1.1.0
      - name: ack-dynamodb-controller.v1.0.3
        version: 1.0.3
      - name: ack-dynamodb-controller.v1.0.2
        version: 1.0.2
      - name: ack-dynamodb-controller.v1.0.1
        version: 1.0.1
      - name: ack-dynamodb-controller.v1.0.0
        version: 1.0.0
      - name: ack-dynamodb-controller.v0.1.10
        version: 0.1.10
      - name: ack-dynamodb-controller.v0.1.9
        version: 0.1.9
      - name: ack-dynamodb-controller.v0.1.8
        version: 0.1.8
      - name: ack-dynamodb-controller.v0.1.7
        version: 0.1.7
      - name: ack-dynamodb-controller.v0.1.6
        version: 0.1.6
      - name: ack-dynamodb-controller.v0.1.5
        version: 0.1.5
      - name: ack-dynamodb-controller.v0.1.4
        version: 0.1.4
      - name: ack-dynamodb-controller.v0.1.3
        version: 0.1.3
      - name: ack-dynamodb-controller.v0.1.1
        version: 0.1.1
      - name: ack-dynamodb-controller.v0.1.0
        version: 0.1.0
      - name: ack-dynamodb-controller.v0.0.20
        version: 0.0.20
      - name: ack-dynamodb-controller.v0.0.19
        version: 0.0.19
      - name: ack-dynamodb-controller.v0.0.18
        version: 0.0.18
      - name: ack-dynamodb-controller.v0.0.17
        version: 0.0.17
      - name: ack-dynamodb-controller.v0.0.16
        version: 0.0.16
      - name: ack-dynamodb-controller.v0.0.15
        version: 0.0.15
      - name: ack-dynamodb-controller.v0.0.14
        version: 0.0.14
      - name: ack-dynamodb-controller.v0.0.13
        version: 0.0.13
      - name: ack-dynamodb-controller.v0.0.12
        version: 0.0.12
      - name: ack-dynamodb-controller.v0.0.11
        version: 0.0.11
      - name: ack-dynamodb-controller.v0.0.10
        version: 0.0.10
      - name: ack-dynamodb-controller.v0.0.9
        version: 0.0.9
      - name: ack-dynamodb-controller.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: ack-dynamodb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-elasticache-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-elasticache-controller.v1.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "CacheParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "CacheSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "ReplicationGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "Snapshot",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "UserGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/elasticache-controller:1.2.3
          createdAt: "2025-09-19T19:19:33Z"
          description: AWS ElastiCache controller is a service controller for managing
            ElastiCache resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CacheCluster represents the state of an AWS elasticache CacheCluster
              resource.
            displayName: CacheCluster
            kind: CacheCluster
            name: cacheclusters.elasticache.services.k8s.aws
            version: v1alpha1
          - description: CacheParameterGroup represents the state of an AWS elasticache
              CacheParameterGroup resource.
            displayName: CacheParameterGroup
            kind: CacheParameterGroup
            name: cacheparametergroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: CacheSubnetGroup represents the state of an AWS elasticache
              CacheSubnetGroup resource.
            displayName: CacheSubnetGroup
            kind: CacheSubnetGroup
            name: cachesubnetgroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ReplicationGroup represents the state of an AWS elasticache
              ReplicationGroup resource.
            displayName: ReplicationGroup
            kind: ReplicationGroup
            name: replicationgroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCache represents the state of an AWS elasticache
              ServerlessCache resource.
            displayName: ServerlessCache
            kind: ServerlessCache
            name: serverlesscaches.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCacheSnapshot represents the state of an AWS elasticache
              ServerlessCacheSnapshot resource.
            displayName: ServerlessCacheSnapshot
            kind: ServerlessCacheSnapshot
            name: serverlesscachesnapshots.elasticache.services.k8s.aws
            version: v1alpha1
          - description: Snapshot represents the state of an AWS elasticache Snapshot
              resource.
            displayName: Snapshot
            kind: Snapshot
            name: snapshots.elasticache.services.k8s.aws
            version: v1alpha1
          - description: UserGroup represents the state of an AWS elasticache UserGroup
              resource.
            displayName: UserGroup
            kind: UserGroup
            name: usergroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS elasticache User resource.
            displayName: User
            kind: User
            name: users.elasticache.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ElastiCache resources in AWS from within your Kubernetes cluster.

          **About Amazon ElastiCache**

          Amazon ElastiCache allows you to seamlessly set up, run, and scale popular open-source compatible in-memory data stores in the cloud. Build data-intensive apps or boost the performance of your existing databases by retrieving data from high throughput and low latency in-memory data stores. Amazon ElastiCache is a popular choice for real-time use cases like Caching, Session Stores, Gaming, Geospatial Services, Real-Time Analytics, and Queuing.

          Amazon ElastiCache offers fully managed Redis, voted the most loved database by developers in the Stack Overflow 2020 Developer Survey, and Memcached for your most demanding applications that require sub-millisecond response times.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ElastiCache
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticache
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ElastiCache Developer Resources
          url: https://aws.amazon.com/elasticache/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: elasticache maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/elasticache-controller:1.2.3
        version: 1.2.3
      entries:
      - name: ack-elasticache-controller.v1.2.3
        version: 1.2.3
      - name: ack-elasticache-controller.v1.2.2
        version: 1.2.2
      - name: ack-elasticache-controller.v1.2.1
        version: 1.2.1
      - name: ack-elasticache-controller.v1.2.0
        version: 1.2.0
      - name: ack-elasticache-controller.v1.1.0
        version: 1.1.0
      - name: ack-elasticache-controller.v1.0.7
        version: 1.0.7
      - name: ack-elasticache-controller.v1.0.6
        version: 1.0.6
      - name: ack-elasticache-controller.v1.0.5
        version: 1.0.5
      - name: ack-elasticache-controller.v1.0.4
        version: 1.0.4
      - name: ack-elasticache-controller.v1.0.3
        version: 1.0.3
      - name: ack-elasticache-controller.v1.0.2
        version: 1.0.2
      - name: ack-elasticache-controller.v1.0.1
        version: 1.0.1
      - name: ack-elasticache-controller.v1.0.0
        version: 1.0.0
      - name: ack-elasticache-controller.v0.2.6
        version: 0.2.6
      - name: ack-elasticache-controller.v0.2.5
        version: 0.2.5
      - name: ack-elasticache-controller.v0.2.4
        version: 0.2.4
      - name: ack-elasticache-controller.v0.2.3
        version: 0.2.3
      - name: ack-elasticache-controller.v0.2.2
        version: 0.2.2
      - name: ack-elasticache-controller.v0.2.1
        version: 0.2.1
      - name: ack-elasticache-controller.v0.2.0
        version: 0.2.0
      - name: ack-elasticache-controller.v0.1.1
        version: 0.1.1
      - name: ack-elasticache-controller.v0.1.0
        version: 0.1.0
      - name: ack-elasticache-controller.v0.0.31
        version: 0.0.31
      - name: ack-elasticache-controller.v0.0.30
        version: 0.0.30
      - name: ack-elasticache-controller.v0.0.29
        version: 0.0.29
      - name: ack-elasticache-controller.v0.0.28
        version: 0.0.28
      - name: ack-elasticache-controller.v0.0.27
        version: 0.0.27
      - name: ack-elasticache-controller.v0.0.26
        version: 0.0.26
      - name: ack-elasticache-controller.v0.0.25
        version: 0.0.25
      - name: ack-elasticache-controller.v0.0.24
        version: 0.0.24
      - name: ack-elasticache-controller.v0.0.23
        version: 0.0.23
      - name: ack-elasticache-controller.v0.0.22
        version: 0.0.22
      - name: ack-elasticache-controller.v0.0.21
        version: 0.0.21
      - name: ack-elasticache-controller.v0.0.20
        version: 0.0.20
      - name: ack-elasticache-controller.v0.0.19
        version: 0.0.19
      - name: ack-elasticache-controller.v0.0.18
        version: 0.0.18
      - name: ack-elasticache-controller.v0.0.17
        version: 0.0.17
      - name: ack-elasticache-controller.v0.0.16
        version: 0.0.16
      - name: ack-elasticache-controller.v0.0.15
        version: 0.0.15
      - name: ack-elasticache-controller.v0.0.14
        version: 0.0.14
      - name: ack-elasticache-controller.v0.0.13
        version: 0.0.13
      - name: ack-elasticache-controller.v0.0.12
        version: 0.0.12
      - name: ack-elasticache-controller.v0.0.11
        version: 0.0.11
      - name: ack-elasticache-controller.v0.0.10
        version: 0.0.10
      - name: ack-elasticache-controller.v0.0.9
        version: 0.0.9
      - name: ack-elasticache-controller.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: ack-elasticache-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ember-csi-community-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ember-csi-community-operator.v0.9.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ember-csi.io/v1alpha1",
                "kind": "EmberStorageBackend",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "config": {
                    "envVars": {
                      "X_CSI_PERSISTENCE_CONFIG": {
                        "storage": "crd"
                      },
                      "X_CSI_DEBUG_MODE": "",
                      "X_CSI_DEFAULT_MOUNT_FS": "ext4",
                      "X_CSI_EMBER_CONFIG": {
                        "grpc_workers": 30,
                        "disable_logs": false,
                        "plugin_name": "",
                        "slow_operations": true,
                        "disabled": [],
                        "enable_probe": false,
                        "debug": false,
                        "user_id": "ember_csi.io",
                        "project_id": "ember_csi.io"
                      },
                      "X_CSI_BACKEND_CONFIG": {
                        "enable_unsupported_driver": false,
                        "driver": "ACCESSIscsi",
                        "name": "",
                        "driver__ACCESSIscsi__vrts_lun_sparse": true,
                        "driver__ACCESSIscsi__vrts_target_config": "/etc/cinder/vrts_target.xml",
                        "driver__AS13000__as13000_ipsan_pools__transform_csv": "Pool0",
                        "driver__AS13000__as13000_meta_pool": "",
                        "driver__AS13000__as13000_token_available_time": 3300,
                        "driver__Acs5000ISCSI__acs5000_copy_interval": 5,
                        "driver__Acs5000ISCSI__acs5000_target": 0,
                        "driver__Acs5000ISCSI__acs5000_volpool_name__transform_csv": "pool01",
                        "driver__DPLFC__dpl_pool": "",
                        "driver__DPLFC__dpl_port": 8357,
                        "driver__DPLISCSI__dpl_pool": "",
                        "driver__DPLISCSI__dpl_port": 8357,
                        "driver__DSWARE__dsware_rest_url": "",
                        "driver__DSWARE__dsware_storage_pools": "",
                        "driver__DSWARE__manager_ips__transform_csv_kvs": "",
                        "driver__Datera__backend_availability_zone": "",
                        "driver__Datera__chap_password": "",
                        "driver__Datera__chap_username": "",
                        "driver__Datera__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__Datera__driver_client_cert": "",
                        "driver__Datera__driver_client_cert_key": "",
                        "driver__Datera__driver_data_namespace": "",
                        "driver__Datera__driver_ssl_cert_path": "",
                        "driver__Datera__driver_ssl_cert_verify": false,
                        "driver__Datera__driver_use_ssl": false,
                        "driver__Datera__filter_function": "",
                        "driver__Datera__goodness_function": "",
                        "driver__Datera__iscsi_iotype": "fileio",
                        "driver__Datera__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__Datera__iscsi_target_flags": "",
                        "driver__Datera__iscsi_write_cache": "on",
                        "driver__Datera__num_shell_tries": 3,
                        "driver__Datera__num_volume_device_scan_tries": 3,
                        "driver__Datera__report_discard_supported": false,
                        "driver__Datera__storage_protocol": "iscsi",
                        "driver__Datera__target_helper": "tgtadm",
                        "driver__Datera__target_ip_address": "$my_ip",
                        "driver__Datera__target_port": 3260,
                        "driver__Datera__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__Datera__target_protocol": "iscsi",
                        "driver__Datera__trace_flags__transform_csv": "",
                        "driver__Datera__use_chap_auth": false,
                        "driver__Datera__volume_backend_name": "",
                        "driver__Datera__volume_clear": "zero",
                        "driver__Datera__volume_clear_ionice": "",
                        "driver__Datera__volume_clear_size": 0,
                        "driver__Datera__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__Datera__volume_copy_bps_limit": 0,
                        "driver__Datera__volume_dd_blocksize": "1M",
                        "driver__Datera__volumes_dir": "$state_path/volumes",
                        "driver__FJDXFC__cinder_eternus_config_file": "/etc/cinder/cinder_fujitsu_eternus_dx.xml",
                        "driver__FJDXISCSI__cinder_eternus_config_file": "/etc/cinder/cinder_fujitsu_eternus_dx.xml",
                        "driver__FlashSystemFC__flashsystem_connection_protocol": "FC",
                        "driver__FlashSystemFC__flashsystem_multihostmap_enabled": true,
                        "driver__FlashSystemISCSI__flashsystem_connection_protocol": "FC",
                        "driver__FlashSystemISCSI__flashsystem_multihostmap_enabled": true,
                        "driver__GPFS__gpfs_images_dir": "",
                        "driver__GPFS__gpfs_images_share_mode__transform_empty_none": "",
                        "driver__GPFS__gpfs_max_clone_depth": 0,
                        "driver__GPFS__gpfs_mount_point_base": "",
                        "driver__GPFS__gpfs_sparse_volumes": true,
                        "driver__GPFS__gpfs_storage_pool": "system",
                        "driver__GPFSRemote__gpfs_hosts__transform_csv": "",
                        "driver__GPFSRemote__gpfs_hosts_key_file": "$state_path/ssh_known_hosts",
                        "driver__GPFSRemote__gpfs_images_dir": "",
                        "driver__GPFSRemote__gpfs_images_share_mode__transform_empty_none": "",
                        "driver__GPFSRemote__gpfs_max_clone_depth": 0,
                        "driver__GPFSRemote__gpfs_mount_point_base": "",
                        "driver__GPFSRemote__gpfs_private_key": "",
                        "driver__GPFSRemote__gpfs_sparse_volumes": true,
                        "driver__GPFSRemote__gpfs_ssh_port": 22,
                        "driver__GPFSRemote__gpfs_storage_pool": "system",
                        "driver__GPFSRemote__gpfs_strict_host_key_policy": false,
                        "driver__GPFSRemote__gpfs_user_login": "root",
                        "driver__GPFSRemote__gpfs_user_password": "",
                        "driver__HBSDFC__backend_availability_zone": "",
                        "driver__HBSDFC__chap_password": "",
                        "driver__HBSDFC__chap_username": "",
                        "driver__HBSDFC__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__HBSDFC__driver_client_cert": "",
                        "driver__HBSDFC__driver_client_cert_key": "",
                        "driver__HBSDFC__driver_data_namespace": "",
                        "driver__HBSDFC__driver_ssl_cert_path": "",
                        "driver__HBSDFC__driver_ssl_cert_verify": false,
                        "driver__HBSDFC__driver_use_ssl": false,
                        "driver__HBSDFC__filter_function": "",
                        "driver__HBSDFC__goodness_function": "",
                        "driver__HBSDFC__iscsi_iotype": "fileio",
                        "driver__HBSDFC__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__HBSDFC__iscsi_target_flags": "",
                        "driver__HBSDFC__iscsi_write_cache": "on",
                        "driver__HBSDFC__num_shell_tries": 3,
                        "driver__HBSDFC__num_volume_device_scan_tries": 3,
                        "driver__HBSDFC__report_discard_supported": false,
                        "driver__HBSDFC__storage_protocol": "iscsi",
                        "driver__HBSDFC__target_helper": "tgtadm",
                        "driver__HBSDFC__target_ip_address": "$my_ip",
                        "driver__HBSDFC__target_port": 3260,
                        "driver__HBSDFC__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__HBSDFC__target_protocol": "iscsi",
                        "driver__HBSDFC__trace_flags__transform_csv": "",
                        "driver__HBSDFC__use_chap_auth": false,
                        "driver__HBSDFC__volume_backend_name": "",
                        "driver__HBSDFC__volume_clear": "zero",
                        "driver__HBSDFC__volume_clear_ionice": "",
                        "driver__HBSDFC__volume_clear_size": 0,
                        "driver__HBSDFC__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__HBSDFC__volume_copy_bps_limit": 0,
                        "driver__HBSDFC__volume_dd_blocksize": "1M",
                        "driver__HBSDFC__volumes_dir": "$state_path/volumes",
                        "driver__HBSDISCSI__backend_availability_zone": "",
                        "driver__HBSDISCSI__chap_password": "",
                        "driver__HBSDISCSI__chap_username": "",
                        "driver__HBSDISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__HBSDISCSI__driver_client_cert": "",
                        "driver__HBSDISCSI__driver_client_cert_key": "",
                        "driver__HBSDISCSI__driver_data_namespace": "",
                        "driver__HBSDISCSI__driver_ssl_cert_path": "",
                        "driver__HBSDISCSI__driver_ssl_cert_verify": false,
                        "driver__HBSDISCSI__driver_use_ssl": false,
                        "driver__HBSDISCSI__filter_function": "",
                        "driver__HBSDISCSI__goodness_function": "",
                        "driver__HBSDISCSI__iscsi_iotype": "fileio",
                        "driver__HBSDISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__HBSDISCSI__iscsi_target_flags": "",
                        "driver__HBSDISCSI__iscsi_write_cache": "on",
                        "driver__HBSDISCSI__num_shell_tries": 3,
                        "driver__HBSDISCSI__num_volume_device_scan_tries": 3,
                        "driver__HBSDISCSI__report_discard_supported": false,
                        "driver__HBSDISCSI__storage_protocol": "iscsi",
                        "driver__HBSDISCSI__target_helper": "tgtadm",
                        "driver__HBSDISCSI__target_ip_address": "$my_ip",
                        "driver__HBSDISCSI__target_port": 3260,
                        "driver__HBSDISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__HBSDISCSI__target_protocol": "iscsi",
                        "driver__HBSDISCSI__trace_flags__transform_csv": "",
                        "driver__HBSDISCSI__use_chap_auth": false,
                        "driver__HBSDISCSI__volume_backend_name": "",
                        "driver__HBSDISCSI__volume_clear": "zero",
                        "driver__HBSDISCSI__volume_clear_ionice": "",
                        "driver__HBSDISCSI__volume_clear_size": 0,
                        "driver__HBSDISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__HBSDISCSI__volume_copy_bps_limit": 0,
                        "driver__HBSDISCSI__volume_dd_blocksize": "1M",
                        "driver__HBSDISCSI__volumes_dir": "$state_path/volumes",
                        "driver__HPE3PARFC__hpe3par_api_url": "",
                        "driver__HPE3PARFC__hpe3par_cpg__transform_csv": "OpenStack",
                        "driver__HPE3PARFC__hpe3par_cpg_snap": "",
                        "driver__HPE3PARFC__hpe3par_debug": false,
                        "driver__HPE3PARFC__hpe3par_iscsi_chap_enabled": false,
                        "driver__HPE3PARFC__hpe3par_iscsi_ips__transform_csv": "",
                        "driver__HPE3PARFC__hpe3par_password": "",
                        "driver__HPE3PARFC__hpe3par_snapshot_expiration": "",
                        "driver__HPE3PARFC__hpe3par_snapshot_retention": "",
                        "driver__HPE3PARFC__hpe3par_target_nsp": "",
                        "driver__HPE3PARFC__hpe3par_username": "",
                        "driver__HPE3PARFC__san_ip": "",
                        "driver__HPE3PARFC__san_login": "admin",
                        "driver__HPE3PARFC__san_password": "",
                        "driver__HPE3PARFC__san_private_key": "",
                        "driver__HPE3PARFC__san_ssh_port": 22,
                        "driver__HPE3PARFC__ssh_conn_timeout": 30,
                        "driver__HPE3PARFC__target_ip_address": "$my_ip",
                        "driver__HPE3PARFC__target_port": 3260,
                        "driver__HPE3PARFC__unique_fqdn_network": true,
                        "driver__HPE3PARISCSI__hpe3par_api_url": "",
                        "driver__HPE3PARISCSI__hpe3par_cpg__transform_csv": "OpenStack",
                        "driver__HPE3PARISCSI__hpe3par_cpg_snap": "",
                        "driver__HPE3PARISCSI__hpe3par_debug": false,
                        "driver__HPE3PARISCSI__hpe3par_iscsi_chap_enabled": false,
                        "driver__HPE3PARISCSI__hpe3par_iscsi_ips__transform_csv": "",
                        "driver__HPE3PARISCSI__hpe3par_password": "",
                        "driver__HPE3PARISCSI__hpe3par_snapshot_expiration": "",
                        "driver__HPE3PARISCSI__hpe3par_snapshot_retention": "",
                        "driver__HPE3PARISCSI__hpe3par_target_nsp": "",
                        "driver__HPE3PARISCSI__hpe3par_username": "",
                        "driver__HPE3PARISCSI__san_ip": "",
                        "driver__HPE3PARISCSI__san_login": "admin",
                        "driver__HPE3PARISCSI__san_password": "",
                        "driver__HPE3PARISCSI__san_private_key": "",
                        "driver__HPE3PARISCSI__san_ssh_port": 22,
                        "driver__HPE3PARISCSI__ssh_conn_timeout": 30,
                        "driver__HPE3PARISCSI__target_ip_address": "$my_ip",
                        "driver__HPE3PARISCSI__target_port": 3260,
                        "driver__HPE3PARISCSI__unique_fqdn_network": true,
                        "driver__HPMSAFC__driver_ssl_cert_path": "",
                        "driver__HPMSAFC__driver_ssl_cert_verify": false,
                        "driver__HPMSAFC__driver_use_ssl": false,
                        "driver__HPMSAFC__hpmsa_pool_name": "A",
                        "driver__HPMSAFC__hpmsa_pool_type": "virtual",
                        "driver__HPMSAFC__san_ip": "",
                        "driver__HPMSAFC__san_login": "admin",
                        "driver__HPMSAFC__san_password": "",
                        "driver__HPMSAISCSI__driver_ssl_cert_path": "",
                        "driver__HPMSAISCSI__driver_ssl_cert_verify": false,
                        "driver__HPMSAISCSI__driver_use_ssl": false,
                        "driver__HPMSAISCSI__hpmsa_iscsi_ips__transform_csv": "",
                        "driver__HPMSAISCSI__hpmsa_pool_name": "A",
                        "driver__HPMSAISCSI__hpmsa_pool_type": "virtual",
                        "driver__HPMSAISCSI__san_ip": "",
                        "driver__HPMSAISCSI__san_login": "admin",
                        "driver__HPMSAISCSI__san_password": "",
                        "driver__HuaweiFC__cinder_huawei_conf_file": "/etc/cinder/cinder_huawei_conf.xml",
                        "driver__HuaweiFC__hypermetro_devices": "",
                        "driver__HuaweiFC__metro_domain_name": "",
                        "driver__HuaweiFC__metro_san_address": "",
                        "driver__HuaweiFC__metro_san_password": "",
                        "driver__HuaweiFC__metro_san_user": "",
                        "driver__HuaweiFC__metro_storage_pools": "",
                        "driver__HuaweiISCSI__cinder_huawei_conf_file": "/etc/cinder/cinder_huawei_conf.xml",
                        "driver__HuaweiISCSI__hypermetro_devices": "",
                        "driver__HuaweiISCSI__metro_domain_name": "",
                        "driver__HuaweiISCSI__metro_san_address": "",
                        "driver__HuaweiISCSI__metro_san_password": "",
                        "driver__HuaweiISCSI__metro_san_user": "",
                        "driver__HuaweiISCSI__metro_storage_pools": "",
                        "driver__IBMStorage__chap": "disabled",
                        "driver__IBMStorage__connection_type": "iscsi",
                        "driver__IBMStorage__management_ips": "",
                        "driver__IBMStorage__proxy": "cinder.volume.drivers.ibm.ibm_storage.proxy.IBMStorageProxy",
                        "driver__InStorageMCSFC__instorage_mcs_allow_tenant_qos": false,
                        "driver__InStorageMCSFC__instorage_mcs_localcopy_rate": 50,
                        "driver__InStorageMCSFC__instorage_mcs_localcopy_timeout": 120,
                        "driver__InStorageMCSFC__instorage_mcs_vol_autoexpand": true,
                        "driver__InStorageMCSFC__instorage_mcs_vol_compression": false,
                        "driver__InStorageMCSFC__instorage_mcs_vol_grainsize": 256,
                        "driver__InStorageMCSFC__instorage_mcs_vol_intier": true,
                        "driver__InStorageMCSFC__instorage_mcs_vol_iogrp": "0",
                        "driver__InStorageMCSFC__instorage_mcs_vol_rsize": 2,
                        "driver__InStorageMCSFC__instorage_mcs_vol_warning": 0,
                        "driver__InStorageMCSFC__instorage_mcs_volpool_name__transform_csv": "volpool",
                        "driver__InStorageMCSFC__instorage_san_secondary_ip": "",
                        "driver__InStorageMCSISCSI__instorage_mcs_allow_tenant_qos": false,
                        "driver__InStorageMCSISCSI__instorage_mcs_localcopy_rate": 50,
                        "driver__InStorageMCSISCSI__instorage_mcs_localcopy_timeout": 120,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_autoexpand": true,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_compression": false,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_grainsize": 256,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_intier": true,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_iogrp": "0",
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_rsize": 2,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_warning": 0,
                        "driver__InStorageMCSISCSI__instorage_mcs_volpool_name__transform_csv": "volpool",
                        "driver__InStorageMCSISCSI__instorage_san_secondary_ip": "",
                        "driver__InfiniboxVolume__chap_password": "",
                        "driver__InfiniboxVolume__chap_username": "",
                        "driver__InfiniboxVolume__infinidat_iscsi_netspaces__transform_csv": "",
                        "driver__InfiniboxVolume__infinidat_pool_name": "",
                        "driver__InfiniboxVolume__infinidat_storage_protocol": "fc",
                        "driver__InfiniboxVolume__infinidat_use_compression": false,
                        "driver__InfiniboxVolume__num_volume_device_scan_tries": 3,
                        "driver__InfiniboxVolume__san_ip": "",
                        "driver__InfiniboxVolume__san_login": "admin",
                        "driver__InfiniboxVolume__san_password": "",
                        "driver__InfiniboxVolume__san_thin_provision": true,
                        "driver__InfiniboxVolume__use_chap_auth": false,
                        "driver__InfiniboxVolume__volume_dd_blocksize": "1M",
                        "driver__InfortrendCLIFC__infortrend_cli_cache": false,
                        "driver__InfortrendCLIFC__infortrend_cli_max_retries": 5,
                        "driver__InfortrendCLIFC__infortrend_cli_path": "/opt/bin/Infortrend/raidcmd_ESDS10.jar",
                        "driver__InfortrendCLIFC__infortrend_cli_timeout": 60,
                        "driver__InfortrendCLIFC__infortrend_iqn_prefix": "iqn.2002-10.com.infortrend",
                        "driver__InfortrendCLIFC__infortrend_pools_name__transform_csv": "",
                        "driver__InfortrendCLIFC__infortrend_slots_a_channels_id__transform_csv": "",
                        "driver__InfortrendCLIFC__infortrend_slots_b_channels_id__transform_csv": "",
                        "driver__InfortrendCLIFC__java_path": "/usr/bin/java",
                        "driver__InfortrendCLIISCSI__infortrend_cli_cache": false,
                        "driver__InfortrendCLIISCSI__infortrend_cli_max_retries": 5,
                        "driver__InfortrendCLIISCSI__infortrend_cli_path": "/opt/bin/Infortrend/raidcmd_ESDS10.jar",
                        "driver__InfortrendCLIISCSI__infortrend_cli_timeout": 60,
                        "driver__InfortrendCLIISCSI__infortrend_iqn_prefix": "iqn.2002-10.com.infortrend",
                        "driver__InfortrendCLIISCSI__infortrend_pools_name__transform_csv": "",
                        "driver__InfortrendCLIISCSI__infortrend_slots_a_channels_id__transform_csv": "",
                        "driver__InfortrendCLIISCSI__infortrend_slots_b_channels_id__transform_csv": "",
                        "driver__InfortrendCLIISCSI__java_path": "/usr/bin/java",
                        "driver__JovianISCSI__backend_availability_zone": "",
                        "driver__JovianISCSI__chap_password": "",
                        "driver__JovianISCSI__chap_username": "",
                        "driver__JovianISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__JovianISCSI__driver_client_cert": "",
                        "driver__JovianISCSI__driver_client_cert_key": "",
                        "driver__JovianISCSI__driver_data_namespace": "",
                        "driver__JovianISCSI__driver_ssl_cert_path": "",
                        "driver__JovianISCSI__driver_ssl_cert_verify": false,
                        "driver__JovianISCSI__driver_use_ssl": false,
                        "driver__JovianISCSI__filter_function": "",
                        "driver__JovianISCSI__goodness_function": "",
                        "driver__JovianISCSI__iscsi_iotype": "fileio",
                        "driver__JovianISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__JovianISCSI__iscsi_target_flags": "",
                        "driver__JovianISCSI__iscsi_write_cache": "on",
                        "driver__JovianISCSI__num_shell_tries": 3,
                        "driver__JovianISCSI__num_volume_device_scan_tries": 3,
                        "driver__JovianISCSI__report_discard_supported": false,
                        "driver__JovianISCSI__storage_protocol": "iscsi",
                        "driver__JovianISCSI__target_helper": "tgtadm",
                        "driver__JovianISCSI__target_ip_address": "$my_ip",
                        "driver__JovianISCSI__target_port": 3260,
                        "driver__JovianISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__JovianISCSI__target_protocol": "iscsi",
                        "driver__JovianISCSI__trace_flags__transform_csv": "",
                        "driver__JovianISCSI__use_chap_auth": false,
                        "driver__JovianISCSI__volume_backend_name": "",
                        "driver__JovianISCSI__volume_clear": "zero",
                        "driver__JovianISCSI__volume_clear_ionice": "",
                        "driver__JovianISCSI__volume_clear_size": 0,
                        "driver__JovianISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__JovianISCSI__volume_copy_bps_limit": 0,
                        "driver__JovianISCSI__volume_dd_blocksize": "1M",
                        "driver__JovianISCSI__volumes_dir": "$state_path/volumes",
                        "driver__KaminarioISCSI__disable_discovery": false,
                        "driver__KaminarioISCSI__san_ip": "",
                        "driver__KaminarioISCSI__san_login": "admin",
                        "driver__KaminarioISCSI__san_password": "",
                        "driver__KaminarioISCSI__unique_fqdn_network": true,
                        "driver__KaminarioISCSI__volume_dd_blocksize": "1M",
                        "driver__LVMVolume__iscsi_iotype": "fileio",
                        "driver__LVMVolume__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__LVMVolume__iscsi_target_flags": "",
                        "driver__LVMVolume__iscsi_write_cache": "on",
                        "driver__LVMVolume__lvm_conf_file": "/etc/cinder/lvm.conf",
                        "driver__LVMVolume__lvm_mirrors": 0,
                        "driver__LVMVolume__lvm_suppress_fd_warnings": false,
                        "driver__LVMVolume__lvm_type": "auto",
                        "driver__LVMVolume__nvmet_port_id": 1,
                        "driver__LVMVolume__scst_target_driver": "iscsi",
                        "driver__LVMVolume__scst_target_iqn_name": "",
                        "driver__LVMVolume__spdk_max_queue_depth": 64,
                        "driver__LVMVolume__spdk_rpc_ip": "",
                        "driver__LVMVolume__spdk_rpc_password": "",
                        "driver__LVMVolume__spdk_rpc_port": 8000,
                        "driver__LVMVolume__spdk_rpc_username": "",
                        "driver__LVMVolume__target_helper": "tgtadm",
                        "driver__LVMVolume__target_ip_address": "$my_ip",
                        "driver__LVMVolume__target_port": 3260,
                        "driver__LVMVolume__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__LVMVolume__target_protocol": "iscsi",
                        "driver__LVMVolume__volume_clear": "zero",
                        "driver__LVMVolume__volume_clear_size": 0,
                        "driver__LVMVolume__volume_dd_blocksize": "1M",
                        "driver__LVMVolume__volume_group": "cinder-volumes",
                        "driver__LVMVolume__volumes_dir": "$state_path/volumes",
                        "driver__LenovoFC__lenovo_pool_name": "A",
                        "driver__LenovoFC__lenovo_pool_type": "virtual",
                        "driver__LenovoISCSI__lenovo_iscsi_ips__transform_csv": "",
                        "driver__LenovoISCSI__lenovo_pool_name": "A",
                        "driver__LenovoISCSI__lenovo_pool_type": "virtual",
                        "driver__LinstorDrbd__linstor_autoplace_count": 0,
                        "driver__LinstorDrbd__linstor_controller_diskless": true,
                        "driver__LinstorDrbd__linstor_default_blocksize": 4096,
                        "driver__LinstorDrbd__linstor_default_storage_pool_name": "DfltStorPool",
                        "driver__LinstorDrbd__linstor_default_uri": "linstor://localhost",
                        "driver__LinstorDrbd__linstor_default_volume_group_name": "drbd-vg",
                        "driver__LinstorDrbd__linstor_volume_downsize_factor__transform_string_float": "4096",
                        "driver__LinstorIscsi__linstor_autoplace_count": 0,
                        "driver__LinstorIscsi__linstor_controller_diskless": true,
                        "driver__LinstorIscsi__linstor_default_blocksize": 4096,
                        "driver__LinstorIscsi__linstor_default_storage_pool_name": "DfltStorPool",
                        "driver__LinstorIscsi__linstor_default_uri": "linstor://localhost",
                        "driver__LinstorIscsi__linstor_default_volume_group_name": "drbd-vg",
                        "driver__LinstorIscsi__linstor_volume_downsize_factor__transform_string_float": "4096",
                        "driver__MStorageFC__nec_actual_free_capacity": false,
                        "driver__MStorageFC__nec_auto_accesscontrol": true,
                        "driver__MStorageFC__nec_backend_max_ld_count": 1024,
                        "driver__MStorageFC__nec_backup_ldname_format": "LX:%s",
                        "driver__MStorageFC__nec_backup_pools__transform_csv": "",
                        "driver__MStorageFC__nec_cv_ldname_format": "LX:__ControlVolume_%xh",
                        "driver__MStorageFC__nec_diskarray_name": "",
                        "driver__MStorageFC__nec_ismcli_fip": "",
                        "driver__MStorageFC__nec_ismcli_password": "",
                        "driver__MStorageFC__nec_ismcli_privkey": "",
                        "driver__MStorageFC__nec_ismcli_user": "",
                        "driver__MStorageFC__nec_ismview_alloptimize": false,
                        "driver__MStorageFC__nec_ismview_dir": "/tmp/nec/cinder",
                        "driver__MStorageFC__nec_ldname_format": "LX:%s",
                        "driver__MStorageFC__nec_ldset": "",
                        "driver__MStorageFC__nec_pools__transform_csv": "",
                        "driver__MStorageFC__nec_queryconfig_view": false,
                        "driver__MStorageFC__nec_ssh_pool_port_number": 22,
                        "driver__MStorageFC__nec_unpairthread_timeout": 3600,
                        "driver__MStorageISCSI__nec_actual_free_capacity": false,
                        "driver__MStorageISCSI__nec_auto_accesscontrol": true,
                        "driver__MStorageISCSI__nec_backend_max_ld_count": 1024,
                        "driver__MStorageISCSI__nec_backup_ldname_format": "LX:%s",
                        "driver__MStorageISCSI__nec_backup_pools__transform_csv": "",
                        "driver__MStorageISCSI__nec_cv_ldname_format": "LX:__ControlVolume_%xh",
                        "driver__MStorageISCSI__nec_diskarray_name": "",
                        "driver__MStorageISCSI__nec_ismcli_fip": "",
                        "driver__MStorageISCSI__nec_ismcli_password": "",
                        "driver__MStorageISCSI__nec_ismcli_privkey": "",
                        "driver__MStorageISCSI__nec_ismcli_user": "",
                        "driver__MStorageISCSI__nec_ismview_alloptimize": false,
                        "driver__MStorageISCSI__nec_ismview_dir": "/tmp/nec/cinder",
                        "driver__MStorageISCSI__nec_ldname_format": "LX:%s",
                        "driver__MStorageISCSI__nec_ldset": "",
                        "driver__MStorageISCSI__nec_pools__transform_csv": "",
                        "driver__MStorageISCSI__nec_queryconfig_view": false,
                        "driver__MStorageISCSI__nec_ssh_pool_port_number": 22,
                        "driver__MStorageISCSI__nec_unpairthread_timeout": 3600,
                        "driver__MacroSANFC__backend_availability_zone": "",
                        "driver__MacroSANFC__chap_password": "",
                        "driver__MacroSANFC__chap_username": "",
                        "driver__MacroSANFC__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__MacroSANFC__driver_client_cert": "",
                        "driver__MacroSANFC__driver_client_cert_key": "",
                        "driver__MacroSANFC__driver_data_namespace": "",
                        "driver__MacroSANFC__driver_ssl_cert_path": "",
                        "driver__MacroSANFC__driver_ssl_cert_verify": false,
                        "driver__MacroSANFC__driver_use_ssl": false,
                        "driver__MacroSANFC__filter_function": "",
                        "driver__MacroSANFC__goodness_function": "",
                        "driver__MacroSANFC__iscsi_iotype": "fileio",
                        "driver__MacroSANFC__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__MacroSANFC__iscsi_target_flags": "",
                        "driver__MacroSANFC__iscsi_write_cache": "on",
                        "driver__MacroSANFC__num_shell_tries": 3,
                        "driver__MacroSANFC__num_volume_device_scan_tries": 3,
                        "driver__MacroSANFC__report_discard_supported": false,
                        "driver__MacroSANFC__storage_protocol": "iscsi",
                        "driver__MacroSANFC__target_helper": "tgtadm",
                        "driver__MacroSANFC__target_ip_address": "$my_ip",
                        "driver__MacroSANFC__target_port": 3260,
                        "driver__MacroSANFC__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__MacroSANFC__target_protocol": "iscsi",
                        "driver__MacroSANFC__trace_flags__transform_csv": "",
                        "driver__MacroSANFC__use_chap_auth": false,
                        "driver__MacroSANFC__volume_backend_name": "",
                        "driver__MacroSANFC__volume_clear": "zero",
                        "driver__MacroSANFC__volume_clear_ionice": "",
                        "driver__MacroSANFC__volume_clear_size": 0,
                        "driver__MacroSANFC__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__MacroSANFC__volume_copy_bps_limit": 0,
                        "driver__MacroSANFC__volume_dd_blocksize": "1M",
                        "driver__MacroSANFC__volumes_dir": "$state_path/volumes",
                        "driver__MacroSANISCSI__backend_availability_zone": "",
                        "driver__MacroSANISCSI__chap_password": "",
                        "driver__MacroSANISCSI__chap_username": "",
                        "driver__MacroSANISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__MacroSANISCSI__driver_client_cert": "",
                        "driver__MacroSANISCSI__driver_client_cert_key": "",
                        "driver__MacroSANISCSI__driver_data_namespace": "",
                        "driver__MacroSANISCSI__driver_ssl_cert_path": "",
                        "driver__MacroSANISCSI__driver_ssl_cert_verify": false,
                        "driver__MacroSANISCSI__driver_use_ssl": false,
                        "driver__MacroSANISCSI__filter_function": "",
                        "driver__MacroSANISCSI__goodness_function": "",
                        "driver__MacroSANISCSI__iscsi_iotype": "fileio",
                        "driver__MacroSANISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__MacroSANISCSI__iscsi_target_flags": "",
                        "driver__MacroSANISCSI__iscsi_write_cache": "on",
                        "driver__MacroSANISCSI__num_shell_tries": 3,
                        "driver__MacroSANISCSI__num_volume_device_scan_tries": 3,
                        "driver__MacroSANISCSI__report_discard_supported": false,
                        "driver__MacroSANISCSI__storage_protocol": "iscsi",
                        "driver__MacroSANISCSI__target_helper": "tgtadm",
                        "driver__MacroSANISCSI__target_ip_address": "$my_ip",
                        "driver__MacroSANISCSI__target_port": 3260,
                        "driver__MacroSANISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__MacroSANISCSI__target_protocol": "iscsi",
                        "driver__MacroSANISCSI__trace_flags__transform_csv": "",
                        "driver__MacroSANISCSI__use_chap_auth": false,
                        "driver__MacroSANISCSI__volume_backend_name": "",
                        "driver__MacroSANISCSI__volume_clear": "zero",
                        "driver__MacroSANISCSI__volume_clear_ionice": "",
                        "driver__MacroSANISCSI__volume_clear_size": 0,
                        "driver__MacroSANISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__MacroSANISCSI__volume_copy_bps_limit": 0,
                        "driver__MacroSANISCSI__volume_dd_blocksize": "1M",
                        "driver__MacroSANISCSI__volumes_dir": "$state_path/volumes",
                        "driver__NetAppCmodeFibreChannel__netapp_vserver": "",
                        "driver__NetAppCmodeISCSI__netapp_vserver": "",
                        "driver__NexentaISCSI__nexenta_blocksize": 4096,
                        "driver__NexentaISCSI__nexenta_dataset_compression": "on",
                        "driver__NexentaISCSI__nexenta_dataset_dedup": "off",
                        "driver__NexentaISCSI__nexenta_dataset_description": "",
                        "driver__NexentaISCSI__nexenta_folder": "",
                        "driver__NexentaISCSI__nexenta_group_snapshot_template": "group-snapshot-%s",
                        "driver__NexentaISCSI__nexenta_host": "",
                        "driver__NexentaISCSI__nexenta_host_group_prefix": "cinder",
                        "driver__NexentaISCSI__nexenta_iscsi_target_host_group": "all",
                        "driver__NexentaISCSI__nexenta_iscsi_target_portal_groups": "",
                        "driver__NexentaISCSI__nexenta_iscsi_target_portal_port": 3260,
                        "driver__NexentaISCSI__nexenta_iscsi_target_portals": "",
                        "driver__NexentaISCSI__nexenta_lu_writebackcache_disabled": false,
                        "driver__NexentaISCSI__nexenta_luns_per_target": 100,
                        "driver__NexentaISCSI__nexenta_ns5_blocksize": 32,
                        "driver__NexentaISCSI__nexenta_origin_snapshot_template": "origin-snapshot-%s",
                        "driver__NexentaISCSI__nexenta_rest_backoff_factor__transform_string_float": "0.5",
                        "driver__NexentaISCSI__nexenta_rest_connect_timeout__transform_string_float": "30",
                        "driver__NexentaISCSI__nexenta_rest_protocol": "auto",
                        "driver__NexentaISCSI__nexenta_rest_read_timeout__transform_string_float": "300",
                        "driver__NexentaISCSI__nexenta_rest_retry_count": 3,
                        "driver__NexentaISCSI__nexenta_sparse": false,
                        "driver__NexentaISCSI__nexenta_target_group_prefix": "cinder",
                        "driver__NexentaISCSI__nexenta_target_prefix": "iqn.1986-03.com.sun:02:cinder",
                        "driver__NexentaISCSI__nexenta_use_https": true,
                        "driver__NexentaISCSI__nexenta_volume": "cinder",
                        "driver__NexentaISCSI__nexenta_volume_group": "iscsi",
                        "driver__NimbleFC__nimble_pool_name": "default",
                        "driver__NimbleFC__nimble_subnet_label": "*",
                        "driver__NimbleFC__nimble_verify_cert_path": "",
                        "driver__NimbleFC__nimble_verify_certificate": false,
                        "driver__NimbleISCSI__nimble_pool_name": "default",
                        "driver__NimbleISCSI__nimble_subnet_label": "*",
                        "driver__NimbleISCSI__nimble_verify_cert_path": "",
                        "driver__NimbleISCSI__nimble_verify_certificate": false,
                        "driver__PVMEFC__driver_ssl_cert_path": "",
                        "driver__PVMEFC__driver_ssl_cert_verify": false,
                        "driver__PVMEFC__driver_use_ssl": false,
                        "driver__PVMEFC__pvme_pool_name": "A",
                        "driver__PVMEFC__san_ip": "",
                        "driver__PVMEFC__san_login": "admin",
                        "driver__PVMEFC__san_password": "",
                        "driver__PVMEISCSI__driver_ssl_cert_path": "",
                        "driver__PVMEISCSI__driver_ssl_cert_verify": false,
                        "driver__PVMEISCSI__driver_use_ssl": false,
                        "driver__PVMEISCSI__pvme_iscsi_ips__transform_csv": "",
                        "driver__PVMEISCSI__pvme_pool_name": "A",
                        "driver__PVMEISCSI__san_ip": "",
                        "driver__PVMEISCSI__san_login": "admin",
                        "driver__PVMEISCSI__san_password": "",
                        "driver__PowerFlex__powerflex_allow_migration_during_rebuild": false,
                        "driver__PowerFlex__powerflex_allow_non_padded_volumes": false,
                        "driver__PowerFlex__powerflex_max_over_subscription_ratio__transform_string_float": "10.0",
                        "driver__PowerFlex__powerflex_rest_server_port": 443,
                        "driver__PowerFlex__powerflex_round_volume_capacity": true,
                        "driver__PowerFlex__powerflex_server_api_version": "",
                        "driver__PowerFlex__powerflex_storage_pools": "",
                        "driver__PowerFlex__powerflex_unmap_volume_before_deletion": false,
                        "driver__PowerMaxFC__driver_ssl_cert_verify": false,
                        "driver__PowerMaxFC__initiator_check": false,
                        "driver__PowerMaxFC__interval": 3,
                        "driver__PowerMaxFC__load_balance": false,
                        "driver__PowerMaxFC__load_balance_real_time": false,
                        "driver__PowerMaxFC__load_data_format": "Avg",
                        "driver__PowerMaxFC__load_look_back": 60,
                        "driver__PowerMaxFC__load_look_back_real_time": 1,
                        "driver__PowerMaxFC__port_group_load_metric": "PercentBusy",
                        "driver__PowerMaxFC__port_load_metric": "PercentBusy",
                        "driver__PowerMaxFC__powermax_array": "",
                        "driver__PowerMaxFC__powermax_array_tag_list__transform_csv": "",
                        "driver__PowerMaxFC__powermax_port_group_name_template": "portGroupName",
                        "driver__PowerMaxFC__powermax_port_groups__transform_csv": "",
                        "driver__PowerMaxFC__powermax_service_level": "",
                        "driver__PowerMaxFC__powermax_short_host_name_template": "shortHostName",
                        "driver__PowerMaxFC__powermax_snapvx_unlink_limit": 3,
                        "driver__PowerMaxFC__powermax_srp": "",
                        "driver__PowerMaxFC__retries": 200,
                        "driver__PowerMaxFC__san_ip": "",
                        "driver__PowerMaxFC__san_login": "admin",
                        "driver__PowerMaxFC__san_password": "",
                        "driver__PowerMaxFC__u4p_failover_autofailback": true,
                        "driver__PowerMaxFC__u4p_failover_backoff_factor": 1,
                        "driver__PowerMaxFC__u4p_failover_retries": 3,
                        "driver__PowerMaxFC__u4p_failover_target__transform_csv_kvs": "",
                        "driver__PowerMaxFC__u4p_failover_timeout": 20,
                        "driver__PowerMaxFC__vmax_workload": "",
                        "driver__PowerMaxISCSI__chap_password": "",
                        "driver__PowerMaxISCSI__chap_username": "",
                        "driver__PowerMaxISCSI__driver_ssl_cert_verify": false,
                        "driver__PowerMaxISCSI__initiator_check": false,
                        "driver__PowerMaxISCSI__interval": 3,
                        "driver__PowerMaxISCSI__load_balance": false,
                        "driver__PowerMaxISCSI__load_balance_real_time": false,
                        "driver__PowerMaxISCSI__load_data_format": "Avg",
                        "driver__PowerMaxISCSI__load_look_back": 60,
                        "driver__PowerMaxISCSI__load_look_back_real_time": 1,
                        "driver__PowerMaxISCSI__port_group_load_metric": "PercentBusy",
                        "driver__PowerMaxISCSI__port_load_metric": "PercentBusy",
                        "driver__PowerMaxISCSI__powermax_array": "",
                        "driver__PowerMaxISCSI__powermax_array_tag_list__transform_csv": "",
                        "driver__PowerMaxISCSI__powermax_port_group_name_template": "portGroupName",
                        "driver__PowerMaxISCSI__powermax_port_groups__transform_csv": "",
                        "driver__PowerMaxISCSI__powermax_service_level": "",
                        "driver__PowerMaxISCSI__powermax_short_host_name_template": "shortHostName",
                        "driver__PowerMaxISCSI__powermax_snapvx_unlink_limit": 3,
                        "driver__PowerMaxISCSI__powermax_srp": "",
                        "driver__PowerMaxISCSI__retries": 200,
                        "driver__PowerMaxISCSI__san_ip": "",
                        "driver__PowerMaxISCSI__san_login": "admin",
                        "driver__PowerMaxISCSI__san_password": "",
                        "driver__PowerMaxISCSI__u4p_failover_autofailback": true,
                        "driver__PowerMaxISCSI__u4p_failover_backoff_factor": 1,
                        "driver__PowerMaxISCSI__u4p_failover_retries": 3,
                        "driver__PowerMaxISCSI__u4p_failover_target__transform_csv_kvs": "",
                        "driver__PowerMaxISCSI__u4p_failover_timeout": 20,
                        "driver__PowerMaxISCSI__use_chap_auth": false,
                        "driver__PowerMaxISCSI__vmax_workload": "",
                        "driver__PowerStore__powerstore_appliances__transform_csv": "",
                        "driver__PowerStore__powerstore_ports__transform_csv": "",
                        "driver__PureFC__driver_ssl_cert_path": "",
                        "driver__PureFC__driver_ssl_cert_verify": false,
                        "driver__PureFC__pure_api_token": "",
                        "driver__PureFC__pure_automatic_max_oversubscription_ratio": true,
                        "driver__PureFC__pure_eradicate_on_delete": false,
                        "driver__PureFC__pure_host_personality__transform_empty_none": "",
                        "driver__PureFC__pure_iscsi_cidr": "0.0.0.0/0",
                        "driver__PureFC__pure_replica_interval_default": 3600,
                        "driver__PureFC__pure_replica_retention_long_term_default": 7,
                        "driver__PureFC__pure_replica_retention_long_term_per_day_default": 3,
                        "driver__PureFC__pure_replica_retention_short_term_default": 14400,
                        "driver__PureFC__pure_replication_pg_name": "cinder-group",
                        "driver__PureFC__pure_replication_pod_name": "cinder-pod",
                        "driver__PureFC__san_ip": "",
                        "driver__PureFC__use_chap_auth": false,
                        "driver__PureISCSI__driver_ssl_cert_path": "",
                        "driver__PureISCSI__driver_ssl_cert_verify": false,
                        "driver__PureISCSI__pure_api_token": "",
                        "driver__PureISCSI__pure_automatic_max_oversubscription_ratio": true,
                        "driver__PureISCSI__pure_eradicate_on_delete": false,
                        "driver__PureISCSI__pure_host_personality__transform_empty_none": "",
                        "driver__PureISCSI__pure_iscsi_cidr": "0.0.0.0/0",
                        "driver__PureISCSI__pure_replica_interval_default": 3600,
                        "driver__PureISCSI__pure_replica_retention_long_term_default": 7,
                        "driver__PureISCSI__pure_replica_retention_long_term_per_day_default": 3,
                        "driver__PureISCSI__pure_replica_retention_short_term_default": 14400,
                        "driver__PureISCSI__pure_replication_pg_name": "cinder-group",
                        "driver__PureISCSI__pure_replication_pod_name": "cinder-pod",
                        "driver__PureISCSI__san_ip": "",
                        "driver__PureISCSI__use_chap_auth": false,
                        "driver__QnapISCSI__chap_password": "",
                        "driver__QnapISCSI__chap_username": "",
                        "driver__QnapISCSI__driver_ssl_cert_verify": false,
                        "driver__QnapISCSI__qnap_management_url": "",
                        "driver__QnapISCSI__qnap_poolname": "",
                        "driver__QnapISCSI__qnap_storage_protocol": "iscsi",
                        "driver__QnapISCSI__san_login": "admin",
                        "driver__QnapISCSI__san_password": "",
                        "driver__QnapISCSI__target_ip_address": "$my_ip",
                        "driver__QnapISCSI__use_chap_auth": false,
                        "driver__Quobyte__quobyte_client_cfg": "",
                        "driver__Quobyte__quobyte_mount_point_base": "$state_path/mnt",
                        "driver__Quobyte__quobyte_overlay_volumes": false,
                        "driver__Quobyte__quobyte_qcow2_volumes": true,
                        "driver__Quobyte__quobyte_sparsed_volumes": true,
                        "driver__Quobyte__quobyte_volume_from_snapshot_cache": false,
                        "driver__Quobyte__quobyte_volume_url": "",
                        "driver__RBD__deferred_deletion_delay": 0,
                        "driver__RBD__deferred_deletion_purge_interval": 60,
                        "driver__RBD__enable_deferred_deletion": false,
                        "driver__RBD__rados_connect_timeout": -1,
                        "driver__RBD__rados_connection_interval": 5,
                        "driver__RBD__rados_connection_retries": 3,
                        "driver__RBD__rbd_ceph_conf": "",
                        "driver__RBD__rbd_cluster_name": "ceph",
                        "driver__RBD__rbd_exclusive_cinder_pool": false,
                        "driver__RBD__rbd_flatten_volume_from_snapshot": false,
                        "driver__RBD__rbd_keyring_conf": "",
                        "driver__RBD__rbd_max_clone_depth": 5,
                        "driver__RBD__rbd_pool": "rbd",
                        "driver__RBD__rbd_secret_uuid": "",
                        "driver__RBD__rbd_store_chunk_size": 4,
                        "driver__RBD__rbd_user": "",
                        "driver__RBD__report_dynamic_total_capacity": true,
                        "driver__RBD__volume_dd_blocksize": "1M",
                        "driver__RBDISCSI__deferred_deletion_delay": 0,
                        "driver__RBDISCSI__deferred_deletion_purge_interval": 60,
                        "driver__RBDISCSI__driver_ssl_cert_verify": false,
                        "driver__RBDISCSI__enable_deferred_deletion": false,
                        "driver__RBDISCSI__rados_connect_timeout": -1,
                        "driver__RBDISCSI__rados_connection_interval": 5,
                        "driver__RBDISCSI__rados_connection_retries": 3,
                        "driver__RBDISCSI__rbd_ceph_conf": "",
                        "driver__RBDISCSI__rbd_cluster_name": "ceph",
                        "driver__RBDISCSI__rbd_exclusive_cinder_pool": false,
                        "driver__RBDISCSI__rbd_flatten_volume_from_snapshot": false,
                        "driver__RBDISCSI__rbd_iscsi_api_debug": false,
                        "driver__RBDISCSI__rbd_iscsi_api_password": "",
                        "driver__RBDISCSI__rbd_iscsi_api_url": "",
                        "driver__RBDISCSI__rbd_iscsi_api_user": "",
                        "driver__RBDISCSI__rbd_iscsi_target_iqn": "",
                        "driver__RBDISCSI__rbd_max_clone_depth": 5,
                        "driver__RBDISCSI__rbd_pool": "rbd",
                        "driver__RBDISCSI__rbd_secret_uuid": "",
                        "driver__RBDISCSI__rbd_store_chunk_size": 4,
                        "driver__RBDISCSI__rbd_user": "",
                        "driver__RBDISCSI__report_dynamic_total_capacity": true,
                        "driver__RBDISCSI__suppress_requests_ssl_warnings": false,
                        "driver__RBDISCSI__volume_dd_blocksize": "1M",
                        "driver__RSD__podm_password": "",
                        "driver__RSD__podm_url": "",
                        "driver__RSD__podm_username": "",
                        "driver__SCFC__dell_api_async_rest_timeout": 15,
                        "driver__SCFC__dell_api_sync_rest_timeout": 30,
                        "driver__SCFC__dell_sc_api_port": 3033,
                        "driver__SCFC__dell_sc_server_folder": "openstack",
                        "driver__SCFC__dell_sc_ssn": 64702,
                        "driver__SCFC__dell_sc_verify_cert": false,
                        "driver__SCFC__dell_sc_volume_folder": "openstack",
                        "driver__SCFC__dell_server_os": "Red Hat Linux 6.x",
                        "driver__SCFC__excluded_domain_ips__transform_csv": "",
                        "driver__SCFC__included_domain_ips__transform_csv": "",
                        "driver__SCFC__secondary_san_ip": "",
                        "driver__SCFC__secondary_san_login": "Admin",
                        "driver__SCFC__secondary_san_password": "",
                        "driver__SCFC__secondary_sc_api_port": 3033,
                        "driver__SCISCSI__dell_api_async_rest_timeout": 15,
                        "driver__SCISCSI__dell_api_sync_rest_timeout": 30,
                        "driver__SCISCSI__dell_sc_api_port": 3033,
                        "driver__SCISCSI__dell_sc_server_folder": "openstack",
                        "driver__SCISCSI__dell_sc_ssn": 64702,
                        "driver__SCISCSI__dell_sc_verify_cert": false,
                        "driver__SCISCSI__dell_sc_volume_folder": "openstack",
                        "driver__SCISCSI__dell_server_os": "Red Hat Linux 6.x",
                        "driver__SCISCSI__excluded_domain_ips__transform_csv": "",
                        "driver__SCISCSI__included_domain_ips__transform_csv": "",
                        "driver__SCISCSI__secondary_san_ip": "",
                        "driver__SCISCSI__secondary_san_login": "Admin",
                        "driver__SCISCSI__secondary_san_password": "",
                        "driver__SCISCSI__secondary_sc_api_port": 3033,
                        "driver__SdsISCSI__backend_availability_zone": "",
                        "driver__SdsISCSI__chap_password": "",
                        "driver__SdsISCSI__chap_username": "",
                        "driver__SdsISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__SdsISCSI__driver_client_cert": "",
                        "driver__SdsISCSI__driver_client_cert_key": "",
                        "driver__SdsISCSI__driver_data_namespace": "",
                        "driver__SdsISCSI__driver_ssl_cert_path": "",
                        "driver__SdsISCSI__driver_ssl_cert_verify": false,
                        "driver__SdsISCSI__driver_use_ssl": false,
                        "driver__SdsISCSI__filter_function": "",
                        "driver__SdsISCSI__goodness_function": "",
                        "driver__SdsISCSI__iscsi_iotype": "fileio",
                        "driver__SdsISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__SdsISCSI__iscsi_target_flags": "",
                        "driver__SdsISCSI__iscsi_write_cache": "on",
                        "driver__SdsISCSI__num_shell_tries": 3,
                        "driver__SdsISCSI__num_volume_device_scan_tries": 3,
                        "driver__SdsISCSI__report_discard_supported": false,
                        "driver__SdsISCSI__storage_protocol": "iscsi",
                        "driver__SdsISCSI__target_helper": "tgtadm",
                        "driver__SdsISCSI__target_ip_address": "$my_ip",
                        "driver__SdsISCSI__target_port": 3260,
                        "driver__SdsISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__SdsISCSI__target_protocol": "iscsi",
                        "driver__SdsISCSI__trace_flags__transform_csv": "",
                        "driver__SdsISCSI__use_chap_auth": false,
                        "driver__SdsISCSI__volume_backend_name": "",
                        "driver__SdsISCSI__volume_clear": "zero",
                        "driver__SdsISCSI__volume_clear_ionice": "",
                        "driver__SdsISCSI__volume_clear_size": 0,
                        "driver__SdsISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__SdsISCSI__volume_copy_bps_limit": 0,
                        "driver__SdsISCSI__volume_dd_blocksize": "1M",
                        "driver__SdsISCSI__volumes_dir": "$state_path/volumes",
                        "driver__SolidFire__driver_ssl_cert_verify": false,
                        "driver__SolidFire__san_ip": "",
                        "driver__SolidFire__san_login": "admin",
                        "driver__SolidFire__san_password": "",
                        "driver__SolidFire__sf_account_prefix": "",
                        "driver__SolidFire__sf_allow_tenant_qos": false,
                        "driver__SolidFire__sf_api_port": 443,
                        "driver__SolidFire__sf_api_request_timeout": 30,
                        "driver__SolidFire__sf_cluster_pairing_timeout": 60,
                        "driver__SolidFire__sf_emulate_512": true,
                        "driver__SolidFire__sf_enable_vag": false,
                        "driver__SolidFire__sf_provisioning_calc": "maxProvisionedSpace",
                        "driver__SolidFire__sf_svip": "",
                        "driver__SolidFire__sf_volume_clone_timeout": 600,
                        "driver__SolidFire__sf_volume_create_timeout": 60,
                        "driver__SolidFire__sf_volume_pairing_timeout": 3600,
                        "driver__SolidFire__sf_volume_prefix": "UUID-",
                        "driver__StorPool__storpool_replication": 3,
                        "driver__StorPool__storpool_template": "",
                        "driver__StorwizeSVCFC__cycle_period_seconds": 300,
                        "driver__StorwizeSVCFC__storwize_peer_pool": "",
                        "driver__StorwizeSVCFC__storwize_preferred_host_site__transform_csv_kvs": "",
                        "driver__StorwizeSVCFC__storwize_san_secondary_ip": "",
                        "driver__StorwizeSVCFC__storwize_svc_allow_tenant_qos": false,
                        "driver__StorwizeSVCFC__storwize_svc_flashcopy_rate": 50,
                        "driver__StorwizeSVCFC__storwize_svc_flashcopy_timeout": 120,
                        "driver__StorwizeSVCFC__storwize_svc_mirror_pool": "",
                        "driver__StorwizeSVCFC__storwize_svc_multipath_enabled": false,
                        "driver__StorwizeSVCFC__storwize_svc_retain_aux_volume": false,
                        "driver__StorwizeSVCFC__storwize_svc_stretched_cluster_partner": "",
                        "driver__StorwizeSVCFC__storwize_svc_vol_autoexpand": true,
                        "driver__StorwizeSVCFC__storwize_svc_vol_compression": false,
                        "driver__StorwizeSVCFC__storwize_svc_vol_easytier": true,
                        "driver__StorwizeSVCFC__storwize_svc_vol_grainsize": 256,
                        "driver__StorwizeSVCFC__storwize_svc_vol_iogrp": "0",
                        "driver__StorwizeSVCFC__storwize_svc_vol_nofmtdisk": false,
                        "driver__StorwizeSVCFC__storwize_svc_vol_rsize": 2,
                        "driver__StorwizeSVCFC__storwize_svc_vol_warning": 0,
                        "driver__StorwizeSVCFC__storwize_svc_volpool_name__transform_csv": "volpool",
                        "driver__StorwizeSVCISCSI__cycle_period_seconds": 300,
                        "driver__StorwizeSVCISCSI__storwize_peer_pool": "",
                        "driver__StorwizeSVCISCSI__storwize_preferred_host_site__transform_csv_kvs": "",
                        "driver__StorwizeSVCISCSI__storwize_san_secondary_ip": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_allow_tenant_qos": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_flashcopy_rate": 50,
                        "driver__StorwizeSVCISCSI__storwize_svc_flashcopy_timeout": 120,
                        "driver__StorwizeSVCISCSI__storwize_svc_iscsi_chap_enabled": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_mirror_pool": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_retain_aux_volume": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_stretched_cluster_partner": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_autoexpand": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_compression": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_easytier": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_grainsize": 256,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_iogrp": "0",
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_nofmtdisk": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_rsize": 2,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_warning": 0,
                        "driver__StorwizeSVCISCSI__storwize_svc_volpool_name__transform_csv": "volpool",
                        "driver__SynoISCSI__chap_password": "",
                        "driver__SynoISCSI__chap_username": "",
                        "driver__SynoISCSI__driver_use_ssl": false,
                        "driver__SynoISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__SynoISCSI__synology_admin_port": 5000,
                        "driver__SynoISCSI__synology_device_id": "",
                        "driver__SynoISCSI__synology_one_time_pass": "",
                        "driver__SynoISCSI__synology_password": "",
                        "driver__SynoISCSI__synology_pool_name": "",
                        "driver__SynoISCSI__synology_ssl_verify": true,
                        "driver__SynoISCSI__synology_username": "admin",
                        "driver__SynoISCSI__target_ip_address": "$my_ip",
                        "driver__SynoISCSI__target_port": 3260,
                        "driver__SynoISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__SynoISCSI__target_protocol": "iscsi",
                        "driver__SynoISCSI__use_chap_auth": false,
                        "driver__Unity__remove_empty_host": false,
                        "driver__Unity__unity_io_ports__transform_csv": "",
                        "driver__Unity__unity_storage_pool_names__transform_csv": "",
                        "driver__VNX__default_timeout": 31536000,
                        "driver__VNX__destroy_empty_storage_group": false,
                        "driver__VNX__force_delete_lun_in_storagegroup": true,
                        "driver__VNX__ignore_pool_full_threshold": false,
                        "driver__VNX__initiator_auto_deregistration": false,
                        "driver__VNX__initiator_auto_registration": false,
                        "driver__VNX__io_port_list__transform_csv": "",
                        "driver__VNX__iscsi_initiators": "",
                        "driver__VNX__max_luns_per_storage_group": 255,
                        "driver__VNX__naviseccli_path": "",
                        "driver__VNX__storage_vnx_authentication_type": "global",
                        "driver__VNX__storage_vnx_pool_names__transform_csv": "",
                        "driver__VNX__storage_vnx_security_file_dir": "",
                        "driver__VNX__vnx_async_migrate": true,
                        "driver__VZStorage__vzstorage_default_volume_format": "raw",
                        "driver__VZStorage__vzstorage_mount_options__transform_csv": "",
                        "driver__VZStorage__vzstorage_mount_point_base": "$state_path/mnt",
                        "driver__VZStorage__vzstorage_shares_config": "/etc/cinder/vzstorage_shares",
                        "driver__VZStorage__vzstorage_sparsed_volumes": true,
                        "driver__VZStorage__vzstorage_used_ratio__transform_string_float": "0.95",
                        "driver__WindowsISCSI__windows_iscsi_lun_path": "C:\\iSCSIVirtualDisks",
                        "driver__WindowsSmbfs__smbfs_default_volume_format": "vhd",
                        "driver__WindowsSmbfs__smbfs_mount_point_base": "C:\\OpenStack\\_mnt",
                        "driver__WindowsSmbfs__smbfs_pool_mappings__transform_csv_kvs": "",
                        "driver__WindowsSmbfs__smbfs_shares_config": "C:\\OpenStack\\smbfs_shares.txt",
                        "driver__XtremIOFC__driver_ssl_cert_path": "",
                        "driver__XtremIOFC__driver_ssl_cert_verify": false,
                        "driver__XtremIOFC__san_ip": "",
                        "driver__XtremIOFC__san_login": "admin",
                        "driver__XtremIOFC__san_password": "",
                        "driver__XtremIOFC__xtremio_array_busy_retry_count": 5,
                        "driver__XtremIOFC__xtremio_array_busy_retry_interval": 5,
                        "driver__XtremIOFC__xtremio_clean_unused_ig": false,
                        "driver__XtremIOFC__xtremio_cluster_name": "",
                        "driver__XtremIOISCSI__driver_ssl_cert_path": "",
                        "driver__XtremIOISCSI__driver_ssl_cert_verify": false,
                        "driver__XtremIOISCSI__san_ip": "",
                        "driver__XtremIOISCSI__san_login": "admin",
                        "driver__XtremIOISCSI__san_password": "",
                        "driver__XtremIOISCSI__xtremio_array_busy_retry_count": 5,
                        "driver__XtremIOISCSI__xtremio_array_busy_retry_interval": 5,
                        "driver__XtremIOISCSI__xtremio_clean_unused_ig": false,
                        "driver__XtremIOISCSI__xtremio_cluster_name": "",
                        "driver__ZadaraVPSAISCSI__zadara_access_key": "",
                        "driver__ZadaraVPSAISCSI__zadara_default_snap_policy": false,
                        "driver__ZadaraVPSAISCSI__zadara_ssl_cert_verify": true,
                        "driver__ZadaraVPSAISCSI__zadara_use_iser": true,
                        "driver__ZadaraVPSAISCSI__zadara_vol_encrypt": false,
                        "driver__ZadaraVPSAISCSI__zadara_vol_name_template": "OS_%s",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_host": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_poolname": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_port": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_use_ssl": false
                      }
                    },
                    "sysFiles": {
                      "name": ""
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/embercsi/ember-csi-operator:latest
          createdAt: 2021-02-22:17:47:00
          description: Multi-vendor CSI plugin supporting 80+ storage drivers
          repository: https://github.com/embercsi/ember-csi-operator
          support: http://readthedocs.org/projects/ember-csi/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an Ember CSI storage backend
            displayName: Storage Backends
            kind: EmberStorageBackend
            name: embercsis.ember-csi.io
            version: v1alpha1
        description: |
          Ember-CSI is a multi-vendor CSI plugin driver supporting
          over 80 storage drivers in a single plugin to provide block and mount
          storage to container orchestration systems.

          The operator adds support for management of multiple Ember CSI instances
          within a cluster and simplifies storage backend deployments.

          ### Supported drivers
          Ember-CSI includes a good number of storage drivers, but due to limitation on hardware availability only a small number of them have been validated at one point or another. In alphabetical order they are:

          * HPE3PARFC
          * HPE3PARISCSI
          * KaminarioISCSI
          * LVMVolume
          * PowerMaxFC
          * PowerMaxISCSI
          * PureFC
          * PureISCSI
          * QnapISCSI
          * RBD
          * SolidFire
          * SynoISCSI
          * XtremIOFC
          * XtremIOISCSI

          ### Getting started
          Please have a look at our quick start documentation on how to get started:
          [github.com/embercsi/ember-csi-operator#quick-start](https://github.com/embercsi/ember-csi-operator#quick-start).
          We recommend to use the user friendly "Edit Form" when creating a new
          EmberStorageBackend. You'll find the link in the upper right corner of the
          editor.

          ### Documentation
          Documentation on Ember CSI itself can be found at [docs.ember-csi.io](http://docs.ember-csi.io/).

          ### Getting help
          Please contact us if you encounter any issues while using Ember CSI.
          Get in touch through the #ember-csi channel on Freenode, also accessible through
          [Kiwi IRC](https://kiwiirc.com/nextclient/irc.freenode.net/ember-csi).

          You can also create an issue on the [Github project](https://github.com/embercsi/ember-csi-operator)
          for bugs, enhancements, or other requests.
        displayName: Ember CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Ember-CSI
        - CSI
        links:
        - name: Learn more about the project
          url: http://ember-csi.io/
        - name: Documentation
          url: http://readthedocs.org/projects/ember-csi/
        - name: Ember-CSI Source Code
          url: https://github.com/embercsi/ember-csi
        - name: Ember-CSI Operator Source Code
          url: https://github.com/embercsi/ember-csi-operator
        maintainers:
        - email: geguileo@redhat.com
          name: Gorka Eguileor
        - email: cschwede@redhat.com
          name: Christian Schwede
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/embercsi/ember-csi-operator:latest
        version: 0.9.9
      entries:
      - name: ember-csi-community-operator.v0.9.9
        version: 0.9.9
      - name: ember-csi-community-operator.v0.9.8
        version: 0.9.8
      - name: ember-csi-community-operator.v0.9.7
        version: 0.9.7
      name: beta
    defaultChannel: beta
    packageName: ember-csi-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sqs-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sqs-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sqs.services.k8s.aws/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sqs-controller:1.2.1
          createdAt: "2025-09-19T19:14:25Z"
          description: AWS SQS controller is a service controller for managing SQS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Queue represents the state of an AWS sqs Queue resource.
            displayName: Queue
            kind: Queue
            name: queues.sqs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Simple Queue Service (SQS) resources in AWS from within your Kubernetes cluster.

          **About Amazon SQS**

          Amazon Simple Queue Service (Amazon SQS) offers a secure, durable, and available hosted queue that lets you integrate and decouple distributed software systems and components. Amazon SQS offers common constructs such as [dead-letter queues](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGuide/sqs-dead-letter-queues.html) and [cost allocation tags](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGuide/sqs-queue-tags.html). It provides a generic web services API that you can access using any programming language that the AWS SDK supports.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SQS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sqs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SQS Developer Resources
          url: https://aws.amazon.com/sqs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sqs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sqs-controller:1.2.1
        version: 1.2.1
      entries:
      - name: ack-sqs-controller.v1.2.1
        version: 1.2.1
      - name: ack-sqs-controller.v1.2.0
        version: 1.2.0
      - name: ack-sqs-controller.v1.1.16
        version: 1.1.16
      - name: ack-sqs-controller.v1.1.14
        version: 1.1.14
      - name: ack-sqs-controller.v1.1.13
        version: 1.1.13
      - name: ack-sqs-controller.v1.1.12
        version: 1.1.12
      - name: ack-sqs-controller.v1.1.11
        version: 1.1.11
      - name: ack-sqs-controller.v1.1.10
        version: 1.1.10
      - name: ack-sqs-controller.v1.1.9
        version: 1.1.9
      - name: ack-sqs-controller.v1.1.8
        version: 1.1.8
      - name: ack-sqs-controller.v1.1.7
        version: 1.1.7
      - name: ack-sqs-controller.v1.1.6
        version: 1.1.6
      - name: ack-sqs-controller.v1.1.5
        version: 1.1.5
      - name: ack-sqs-controller.v1.1.4
        version: 1.1.4
      - name: ack-sqs-controller.v1.1.3
        version: 1.1.3
      - name: ack-sqs-controller.v1.1.2
        version: 1.1.2
      - name: ack-sqs-controller.v1.1.1
        version: 1.1.1
      - name: ack-sqs-controller.v1.1.0
        version: 1.1.0
      - name: ack-sqs-controller.v1.0.17
        version: 1.0.17
      - name: ack-sqs-controller.v1.0.16
        version: 1.0.16
      - name: ack-sqs-controller.v1.0.15
        version: 1.0.15
      - name: ack-sqs-controller.v1.0.14
        version: 1.0.14
      - name: ack-sqs-controller.v1.0.13
        version: 1.0.13
      - name: ack-sqs-controller.v1.0.12
        version: 1.0.12
      - name: ack-sqs-controller.v1.0.10
        version: 1.0.10
      - name: ack-sqs-controller.v1.0.7
        version: 1.0.7
      - name: ack-sqs-controller.v1.0.5
        version: 1.0.5
      - name: ack-sqs-controller.v1.0.3
        version: 1.0.3
      - name: ack-sqs-controller.v1.0.2
        version: 1.0.2
      - name: ack-sqs-controller.v1.0.1
        version: 1.0.1
      - name: ack-sqs-controller.v1.0.0
        version: 1.0.0
      - name: ack-sqs-controller.v0.0.7
        version: 0.0.7
      - name: ack-sqs-controller.v0.0.6
        version: 0.0.6
      - name: ack-sqs-controller.v0.0.5
        version: 0.0.5
      - name: ack-sqs-controller.v0.0.4
        version: 0.0.4
      - name: ack-sqs-controller.v0.0.2
        version: 0.0.2
      - name: ack-sqs-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-sqs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/limitador-operator
    name: limitador-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: limitador-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/limitador-operator:v0.5.0
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: The Limitador operator installs and maintains limitador instances
        displayName: Limitador
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        links:
        - name: Limitador Operator
          url: https://github.com/Kuadrant/limitador-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - quay.io/kuadrant/limitador-operator:v0.5.0
        - quay.io/kuadrant/limitador:v1.2.0
        version: 0.5.0
      entries:
      - name: limitador-operator.v0.5.0
        version: 0.5.0
      - name: limitador-operator.v0.4.0
        version: 0.4.0
      - name: limitador-operator.v0.3.0
        version: 0.3.0
      name: alpha
    - currentCSV: limitador-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "name": "toy_get_route",
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/limitador-operator:v0.11.0
          createdAt: "2024-10-02T10:46:31Z"
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: The Limitador operator installs and maintains limitador instances
        displayName: Limitador
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        links:
        - name: Limitador Operator
          url: https://github.com/Kuadrant/limitador-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - quay.io/kuadrant/limitador-operator:v0.11.0
        - quay.io/kuadrant/limitador:v1.6.0
        version: 0.11.0
      entries:
      - name: limitador-operator.v0.11.0
        version: 0.11.0
      - name: limitador-operator.v0.10.0
        version: 0.10.0
      - name: limitador-operator.v0.6.0
        version: 0.6.0
      - name: limitador-operator.v0.5.0
        version: 0.5.0
      - name: limitador-operator.v0.4.0
        version: 0.4.0
      - name: limitador-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: limitador-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/limitador-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Commvault
      provider-url: ""
    name: hedvig-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hedvig-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\":\"hedvig.io/v1\",\n    \"kind\":\"HedvigDeploy\",\n
            \   \"metadata\": {\n      \"name\": \"testdeploy\",\n      \"namespace\":
            \"hedvig\"\n    },\n    \"spec\": {\n      \"cluster\": {\n        \"name\":
            \"hedvigcl\",\n        \"seeds\": [\n          \"hedvigcl1.hedvig.io\",\n
            \         \"hedvigcl2.hedvig.io\",\n          \"hedvigcl3.hedvig.io\"\n
            \       ]\n      },\n      \"k8s\": {\n        \"name\": \"k8scl\"\n      },\n
            \     \"proxy\": {\n        \"repository\": \"hedvig/hedvig-proxy\",\n
            \       \"tag\": \"v-4.3.1\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"controller\": {\n        \"repository\": \"hedvig/hedvig-csi-volume-driver\",\n
            \       \"tag\": \"v-2.0\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"node\": {\n        \"repository\": \"hedvig/hedvig-csi-volume-driver\",\n
            \       \"tag\": \"v-2.0\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"sidecars\": [\n        {\n          \"name\": \"csi-provisioner\",\n
            \         \"repository\": \"quay.io/k8scsi/csi-provisioner\",\n          \"tag\":
            \"v1.6.0\",\n          \"imagePullPolicy\": \"IfNotPresent\"\n        },\n
            \       {\n          \"name\": \"csi-snapshotter\",\n          \"repository\":
            \"quay.io/k8scsi/csi-snapshotter\",\n          \"tag\": \"v2.1.0\",\n
            \         \"imagePullPolicy\": \"IfNotPresent\"\n        },              \n
            \       {\n          \"name\": \"csi-resizer\",\n          \"repository\":
            \"quay.io/k8scsi/csi-resizer\",\n          \"tag\": \"v0.5.0\",\n          \"imagePullPolicy\":
            \"IfNotPresent\"\n        },              \n        {\n          \"name\":
            \"csi-node-driver-registrar\",\n          \"repository\": \"quay.io/k8scsi/csi-node-driver-registrar\",\n
            \         \"tag\": \"v1.0-canary\",\n          \"imagePullPolicy\": \"IfNotPresent\"\n
            \       }\n      ]\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: hedvig/hedvig-operator:2.0.0-ubi
          createdAt: ""
          description: Hedvig Operator is the official operator to deploy and manage
            Hedvig components in container orchestrators such as Kubernetes and Openshift.
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hedvig Deploy
            displayName: HedvigDeploy
            kind: HedvigDeploy
            name: hedvigdeploys.hedvig.io
            version: v1
        description: "The Hedvig Distributed Storage Platform is a software-defined,
          scale-out storage solution. It distributes data simultaneously across multiple
          locations – from on-premises data centers to the cloud – and scales capacity
          on-demand by leveraging the storage of commodity x86/ARM servers. The distributed
          write maximizes availability and protects data from hardware failures ranging
          from a single disk to an entire site, improving disaster recovery planning.\n\nHedvig
          provides a CSI Driver for accelerating your journey into the container ecosystem.
          CSI Driver enables containerized applications to maintain state by dynamically
          provisioning and consuming Hedvig virtual disks as persistent volumes. \n###
          \n\nHedvig Operator deploys the Hedvig Storage Proxy and the Hedvig CSI
          Driver to manage the life cycle of persistent storage in Kubernetes and
          Openshift.\n\n**Supported Container Orchestrators:**\n* Kubernetes 1.16+\n*
          Red Hat OpenShift 4.6+\n\n**Supported Hedvig Releases:**\n* Hedvig 4.0+\nPlease
          refer to the [documentation](https://documentation.commvault.com/commvault/hedvig/others/pdf/Hedvig_Operator_User_Guide.pdf)
          for steps to install and use the operator.\n\n\n\n\n\n"
        displayName: Hedvig Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - CSI
        - Hedvig
        - Cloud
        - Block
        - NFS
        - iSCSI
        - Snapshot
        - Volume
        links:
        - name: Documentation
          url: https://documentation.commvault.com/commvault/hedvig/article
        - name: Hedvig CSI User Guide
          url: https://documentation.commvault.com/commvault/hedvig/others/pdf/Hedvig_CSI_User_Guide.pdf
        maintainers:
        - email: ashenoy@commvault.com
          name: Abhijith Shenoy
        maturity: stable
        provider:
          name: Commvault
        relatedImages:
        - hedvig/hedvig-operator:2.0.0-ubi
        version: 1.0.2
      entries:
      - name: hedvig-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: hedvig-operator
    provider:
      name: Commvault
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: External Secrets
      provider-url: https://external-secrets.io
    name: external-secrets-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: external-secrets-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": "1234"
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "key": "provider-key",
                      "property": "provider-key-property",
                      "version": "provider-key-version"
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterExternalSecret",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "externalSecretName": "hello-world-es",
                  "externalSecretSpec": {
                    "data": [
                      {
                        "remoteRef": {
                          "key": "provider-key",
                          "property": "provider-key-property",
                          "version": "provider-key-version"
                        },
                        "secretKey": "secret-key-to-be-managed"
                      }
                    ],
                    "dataFrom": [
                      {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    ],
                    "refreshInterval": "1h",
                    "secretStoreRef": {
                      "kind": "SecretStore",
                      "name": "secret-store-name"
                    },
                    "target": {
                      "creationPolicy": "Merge",
                      "name": "my-secret",
                      "template": {
                        "data": {
                          "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                        },
                        "metadata": {
                          "annotations": {},
                          "labels": {}
                        },
                        "templateFrom": [
                          {
                            "configMap": {
                              "items": [
                                {
                                  "key": "alertmanager.yaml"
                                }
                              ],
                              "name": "alertmanager"
                            }
                          }
                        ],
                        "type": "kubernetes.io/dockerconfigjson"
                      }
                    }
                  },
                  "namespaceSelector": {
                    "matchLabels": {
                      "cool": "label"
                    }
                  },
                  "refreshTime": "1m"
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": 1234
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "extract": {
                        "conversionStrategy": "Default",
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    },
                    {
                      "find": {
                        "conversionStrategy": "Unicode",
                        "name": {
                          "regexp": ".*foobar.*"
                        },
                        "path": "path-to-filter",
                        "tags": {
                          "foo": "bar"
                        }
                      }
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "deletionPolicy": "Retain",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.external-secrets.io/v1alpha1",
                "kind": "OperatorConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "requeueInterval": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  },
                  "concurrent": 1,
                  "controllerClass": "",
                  "crds": {
                    "createClusterExternalSecret": true,
                    "createClusterSecretStore": true
                  },
                  "createOperator": true,
                  "deploymentAnnotations": {},
                  "extraArgs": {},
                  "extraEnv": [],
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/external-secrets/external-secrets",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "installCRDs": false,
                  "leaderElect": false,
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "priorityClassName": "",
                  "processClusterExternalSecret": true,
                  "processClusterStore": true,
                  "prometheus": {
                    "enabled": false,
                    "service": {
                      "port": 8080
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "scopedNamespace": "",
                  "scopedRBAC": false,
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "certCheckInterval": "5m",
                    "certDir": "/tmp/certs",
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
          createdAt: "2024-12-03T10:44:19Z"
          description: Operator to configure external-secrets helm-chart based operator
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/external-secrets/external-secrets-helm-operator
          support: external-secrets
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ACRAccessToken
            name: acraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster scoped resource that allows you to manage ExternalSecrets
              across namespaces
            displayName: ClusterExternalSecret
            kind: ClusterExternalSecret
            name: clusterexternalsecrets.external-secrets.io
            version: v1beta1
          - kind: ClusterGenerator
            name: clustergenerators.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1alpha1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1beta1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1beta1
          - kind: ECRAuthorizationToken
            name: ecrauthorizationtokens.generators.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1alpha1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1beta1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1beta1
          - kind: Fake
            name: fakes.generators.external-secrets.io
            version: v1alpha1
          - kind: GCRAccessToken
            name: gcraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - kind: GithubAccessToken
            name: githubaccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Configures external-secrets helm-chart based operator
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.external-secrets.io
            version: v1alpha1
          - kind: Password
            name: passwords.generators.external-secrets.io
            version: v1alpha1
          - kind: PushSecret
            name: pushsecrets.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1alpha1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1beta1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1beta1
          - kind: STSSessionToken
            name: stssessiontokens.generators.external-secrets.io
            version: v1alpha1
          - kind: UUID
            name: uuids.generators.external-secrets.io
            version: v1alpha1
          - kind: VaultDynamicSecret
            name: vaultdynamicsecrets.generators.external-secrets.io
            version: v1alpha1
          - kind: Webhook
            name: webhooks.generators.external-secrets.io
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK (Helm version) to configure **[official external-secrets operator helm chart](https://github.com/external-secrets/external-secrets)**, so it can be installed via OLM without having to do any change on current Helm Charts.

          Before any other resources provided by this Operator can be deployed, it is essential to create an
          OperatorConfig resource.

          The usual Helm Chart file `values.yaml`, like:
          ```yaml
          prometheus:
            enabled: true
            service:
              port: 8080
          resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          need to be encapsulated into a new custom resource called `OperatorConfig`:
          ```yaml
          apiVersion: operator.external-secrets.io/v1alpha1
          kind: OperatorConfig
          metadata:
            name: cluster
          spec:
            prometheus:
              enabled: true
              service:
                port: 8080
            resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          Once the OperatorConfig resource is deployed, the operator will create all helm chart resources, using the custom resource name as a preffix for all resources names, like for example a `Deployment` called `cluster-external-secrets`.

          ## Documentation

          * [Upstream Documentation](https://external-secrets.io)
          * [Helm Operator Install](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/install.md)
          * [Helm Operator Development](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/development.md)
          * [Helm Operator Release](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/release.md)

          ## Contributing

          You can contribute by:

          * Raising any issues you find using External Secrets Operator
          * Fixing issues by opening [Pull Requests](https://github.com/external-secrets/external-secrets-helm-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving documentation
          * Talking about External Secrets Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/external-secrets/external-secrets-helm-operator/issues).

          ## License

          External Secrets Operator is under Apache 2.0 license. See the [LICENSE](LICENSE) file for details.
        displayName: External Secrets Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secrets
        - secrets-manager
        - vault
        - aws
        - azure
        - google
        - ibm
        - akeyless
        - yandex
        - gitlab
        - oracle
        links:
        - name: External Secrets Operator
          url: https://external-secrets.io
        - name: GitHub
          url: https://github.com/external-secrets/external-secrets
        - name: Slack
          url: https://kubernetes.slack.com/messages/external-secrets
        maintainers:
        - email: contact@external-secrets.io
          name: external-secrets community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: External Secrets
          url: https://external-secrets.io
        relatedImages:
        - ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
        version: 0.11.0
      entries:
      - name: external-secrets-operator.v0.11.0
        version: 0.11.0
      - name: external-secrets-operator.v0.10.7
        version: 0.10.7
      - name: external-secrets-operator.v0.10.6
        version: 0.10.6
      - name: external-secrets-operator.v0.10.5
        version: 0.10.5
      - name: external-secrets-operator.v0.10.4
        version: 0.10.4
      - name: external-secrets-operator.v0.10.3
        version: 0.10.3
      - name: external-secrets-operator.v0.10.2
        version: 0.10.2
      - name: external-secrets-operator.v0.10.1
        version: 0.10.1
      - name: external-secrets-operator.v0.10.0
        version: 0.10.0
      - name: external-secrets-operator.v0.9.20
        version: 0.9.20
      - name: external-secrets-operator.v0.9.19
        version: 0.9.19
      - name: external-secrets-operator.v0.9.18
        version: 0.9.18
      - name: external-secrets-operator.v0.9.17
        version: 0.9.17
      - name: external-secrets-operator.v0.9.16
        version: 0.9.16
      - name: external-secrets-operator.v0.9.15-2
        version: 0.9.15-2
      - name: external-secrets-operator.v0.9.14
        version: 0.9.14
      - name: external-secrets-operator.v0.9.13
        version: 0.9.13
      - name: external-secrets-operator.v0.9.12
        version: 0.9.12
      - name: external-secrets-operator.v0.9.11
        version: 0.9.11
      - name: external-secrets-operator.v0.9.10
        version: 0.9.10
      - name: external-secrets-operator.v0.9.9
        version: 0.9.9
      - name: external-secrets-operator.v0.9.8
        version: 0.9.8
      - name: external-secrets-operator.v0.9.7
        version: 0.9.7
      - name: external-secrets-operator.v0.9.6
        version: 0.9.6
      - name: external-secrets-operator.v0.9.5
        version: 0.9.5
      - name: external-secrets-operator.v0.9.4
        version: 0.9.4
      - name: external-secrets-operator.v0.9.3
        version: 0.9.3
      - name: external-secrets-operator.v0.9.2
        version: 0.9.2
      - name: external-secrets-operator.v0.9.1
        version: 0.9.1
      - name: external-secrets-operator.v0.9.0
        version: 0.9.0
      - name: external-secrets-operator.v0.8.7
        version: 0.8.7
      - name: external-secrets-operator.v0.8.6
        version: 0.8.6
      - name: external-secrets-operator.v0.8.5
        version: 0.8.5
      - name: external-secrets-operator.v0.8.4
        version: 0.8.4
      - name: external-secrets-operator.v0.8.3
        version: 0.8.3
      - name: external-secrets-operator.v0.8.2
        version: 0.8.2
      - name: external-secrets-operator.v0.8.1
        version: 0.8.1
      - name: external-secrets-operator.v0.8.0
        version: 0.8.0
      - name: external-secrets-operator.v0.7.2
        version: 0.7.2
      - name: external-secrets-operator.v0.7.1
        version: 0.7.1
      - name: external-secrets-operator.v0.7.0
        version: 0.7.0
      - name: external-secrets-operator.v0.7.0-rc1
        version: 0.7.0-rc1
      - name: external-secrets-operator.v0.6.1
        version: 0.6.1
      - name: external-secrets-operator.v0.6.0
        version: 0.6.0
      - name: external-secrets-operator.v0.6.0-rc1
        version: 0.6.0-rc1
      - name: external-secrets-operator.v0.5.9
        version: 0.5.9
      - name: external-secrets-operator.v0.5.8
        version: 0.5.8
      - name: external-secrets-operator.v0.5.7
        version: 0.5.7
      - name: external-secrets-operator.v0.5.6
        version: 0.5.6
      - name: external-secrets-operator.v0.5.5
        version: 0.5.5
      - name: external-secrets-operator.v0.5.4
        version: 0.5.4
      - name: external-secrets-operator.v0.5.3
        version: 0.5.3
      - name: external-secrets-operator.v0.5.2
        version: 0.5.2
      - name: external-secrets-operator.v0.5.1
        version: 0.5.1
      - name: external-secrets-operator.v0.5.0
        version: 0.5.0
      - name: external-secrets-operator.v0.4.4
        version: 0.4.4
      - name: external-secrets-operator.v0.4.3
        version: 0.4.3
      - name: external-secrets-operator.v0.4.2
        version: 0.4.2
      - name: external-secrets-operator.v0.4.1
        version: 0.4.1
      - name: external-secrets-operator.v0.3.11
        version: 0.3.11
      - name: external-secrets-operator.v0.3.10
        version: 0.3.10
      - name: external-secrets-operator.v0.3.9
        version: 0.3.9
      - name: external-secrets-operator.v0.3.8
        version: 0.3.8
      name: alpha
    - currentCSV: external-secrets-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": "1234"
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "key": "provider-key",
                      "property": "provider-key-property",
                      "version": "provider-key-version"
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterExternalSecret",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "externalSecretName": "hello-world-es",
                  "externalSecretSpec": {
                    "data": [
                      {
                        "remoteRef": {
                          "key": "provider-key",
                          "property": "provider-key-property",
                          "version": "provider-key-version"
                        },
                        "secretKey": "secret-key-to-be-managed"
                      }
                    ],
                    "dataFrom": [
                      {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    ],
                    "refreshInterval": "1h",
                    "secretStoreRef": {
                      "kind": "SecretStore",
                      "name": "secret-store-name"
                    },
                    "target": {
                      "creationPolicy": "Merge",
                      "name": "my-secret",
                      "template": {
                        "data": {
                          "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                        },
                        "metadata": {
                          "annotations": {},
                          "labels": {}
                        },
                        "templateFrom": [
                          {
                            "configMap": {
                              "items": [
                                {
                                  "key": "alertmanager.yaml"
                                }
                              ],
                              "name": "alertmanager"
                            }
                          }
                        ],
                        "type": "kubernetes.io/dockerconfigjson"
                      }
                    }
                  },
                  "namespaceSelector": {
                    "matchLabels": {
                      "cool": "label"
                    }
                  },
                  "refreshTime": "1m"
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": 1234
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "extract": {
                        "conversionStrategy": "Default",
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    },
                    {
                      "find": {
                        "conversionStrategy": "Unicode",
                        "name": {
                          "regexp": ".*foobar.*"
                        },
                        "path": "path-to-filter",
                        "tags": {
                          "foo": "bar"
                        }
                      }
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "deletionPolicy": "Retain",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.external-secrets.io/v1alpha1",
                "kind": "OperatorConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "requeueInterval": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  },
                  "concurrent": 1,
                  "controllerClass": "",
                  "crds": {
                    "createClusterExternalSecret": true,
                    "createClusterSecretStore": true
                  },
                  "createOperator": true,
                  "deploymentAnnotations": {},
                  "extraArgs": {},
                  "extraEnv": [],
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/external-secrets/external-secrets",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "installCRDs": false,
                  "leaderElect": false,
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "priorityClassName": "",
                  "processClusterExternalSecret": true,
                  "processClusterStore": true,
                  "prometheus": {
                    "enabled": false,
                    "service": {
                      "port": 8080
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "scopedNamespace": "",
                  "scopedRBAC": false,
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "certCheckInterval": "5m",
                    "certDir": "/tmp/certs",
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
          createdAt: "2024-12-03T10:44:19Z"
          description: Operator to configure external-secrets helm-chart based operator
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/external-secrets/external-secrets-helm-operator
          support: external-secrets
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ACRAccessToken
            name: acraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster scoped resource that allows you to manage ExternalSecrets
              across namespaces
            displayName: ClusterExternalSecret
            kind: ClusterExternalSecret
            name: clusterexternalsecrets.external-secrets.io
            version: v1beta1
          - kind: ClusterGenerator
            name: clustergenerators.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1alpha1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1beta1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1beta1
          - kind: ECRAuthorizationToken
            name: ecrauthorizationtokens.generators.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1alpha1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1beta1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1beta1
          - kind: Fake
            name: fakes.generators.external-secrets.io
            version: v1alpha1
          - kind: GCRAccessToken
            name: gcraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - kind: GithubAccessToken
            name: githubaccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Configures external-secrets helm-chart based operator
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.external-secrets.io
            version: v1alpha1
          - kind: Password
            name: passwords.generators.external-secrets.io
            version: v1alpha1
          - kind: PushSecret
            name: pushsecrets.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1alpha1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1beta1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1beta1
          - kind: STSSessionToken
            name: stssessiontokens.generators.external-secrets.io
            version: v1alpha1
          - kind: UUID
            name: uuids.generators.external-secrets.io
            version: v1alpha1
          - kind: VaultDynamicSecret
            name: vaultdynamicsecrets.generators.external-secrets.io
            version: v1alpha1
          - kind: Webhook
            name: webhooks.generators.external-secrets.io
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK (Helm version) to configure **[official external-secrets operator helm chart](https://github.com/external-secrets/external-secrets)**, so it can be installed via OLM without having to do any change on current Helm Charts.

          Before any other resources provided by this Operator can be deployed, it is essential to create an
          OperatorConfig resource.

          The usual Helm Chart file `values.yaml`, like:
          ```yaml
          prometheus:
            enabled: true
            service:
              port: 8080
          resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          need to be encapsulated into a new custom resource called `OperatorConfig`:
          ```yaml
          apiVersion: operator.external-secrets.io/v1alpha1
          kind: OperatorConfig
          metadata:
            name: cluster
          spec:
            prometheus:
              enabled: true
              service:
                port: 8080
            resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          Once the OperatorConfig resource is deployed, the operator will create all helm chart resources, using the custom resource name as a preffix for all resources names, like for example a `Deployment` called `cluster-external-secrets`.

          ## Documentation

          * [Upstream Documentation](https://external-secrets.io)
          * [Helm Operator Install](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/install.md)
          * [Helm Operator Development](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/development.md)
          * [Helm Operator Release](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/release.md)

          ## Contributing

          You can contribute by:

          * Raising any issues you find using External Secrets Operator
          * Fixing issues by opening [Pull Requests](https://github.com/external-secrets/external-secrets-helm-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving documentation
          * Talking about External Secrets Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/external-secrets/external-secrets-helm-operator/issues).

          ## License

          External Secrets Operator is under Apache 2.0 license. See the [LICENSE](LICENSE) file for details.
        displayName: External Secrets Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secrets
        - secrets-manager
        - vault
        - aws
        - azure
        - google
        - ibm
        - akeyless
        - yandex
        - gitlab
        - oracle
        links:
        - name: External Secrets Operator
          url: https://external-secrets.io
        - name: GitHub
          url: https://github.com/external-secrets/external-secrets
        - name: Slack
          url: https://kubernetes.slack.com/messages/external-secrets
        maintainers:
        - email: contact@external-secrets.io
          name: external-secrets community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: External Secrets
          url: https://external-secrets.io
        relatedImages:
        - ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
        version: 0.11.0
      entries:
      - name: external-secrets-operator.v0.11.0
        version: 0.11.0
      - name: external-secrets-operator.v0.10.7
        version: 0.10.7
      - name: external-secrets-operator.v0.10.6
        version: 0.10.6
      - name: external-secrets-operator.v0.10.5
        version: 0.10.5
      - name: external-secrets-operator.v0.10.4
        version: 0.10.4
      - name: external-secrets-operator.v0.10.3
        version: 0.10.3
      - name: external-secrets-operator.v0.10.2
        version: 0.10.2
      - name: external-secrets-operator.v0.10.1
        version: 0.10.1
      - name: external-secrets-operator.v0.10.0
        version: 0.10.0
      - name: external-secrets-operator.v0.9.20
        version: 0.9.20
      - name: external-secrets-operator.v0.9.19
        version: 0.9.19
      - name: external-secrets-operator.v0.9.18
        version: 0.9.18
      - name: external-secrets-operator.v0.9.17
        version: 0.9.17
      - name: external-secrets-operator.v0.9.16
        version: 0.9.16
      - name: external-secrets-operator.v0.9.15-2
        version: 0.9.15-2
      - name: external-secrets-operator.v0.9.14
        version: 0.9.14
      - name: external-secrets-operator.v0.9.13
        version: 0.9.13
      - name: external-secrets-operator.v0.9.12
        version: 0.9.12
      - name: external-secrets-operator.v0.9.11
        version: 0.9.11
      - name: external-secrets-operator.v0.9.10
        version: 0.9.10
      - name: external-secrets-operator.v0.9.9
        version: 0.9.9
      - name: external-secrets-operator.v0.9.8
        version: 0.9.8
      - name: external-secrets-operator.v0.9.7
        version: 0.9.7
      - name: external-secrets-operator.v0.9.6
        version: 0.9.6
      - name: external-secrets-operator.v0.9.5
        version: 0.9.5
      - name: external-secrets-operator.v0.9.4
        version: 0.9.4
      - name: external-secrets-operator.v0.9.3
        version: 0.9.3
      - name: external-secrets-operator.v0.9.2
        version: 0.9.2
      - name: external-secrets-operator.v0.9.1
        version: 0.9.1
      - name: external-secrets-operator.v0.9.0
        version: 0.9.0
      - name: external-secrets-operator.v0.8.7
        version: 0.8.7
      - name: external-secrets-operator.v0.8.6
        version: 0.8.6
      - name: external-secrets-operator.v0.8.5
        version: 0.8.5
      - name: external-secrets-operator.v0.8.4
        version: 0.8.4
      - name: external-secrets-operator.v0.8.3
        version: 0.8.3
      - name: external-secrets-operator.v0.8.2
        version: 0.8.2
      - name: external-secrets-operator.v0.8.1
        version: 0.8.1
      - name: external-secrets-operator.v0.8.0
        version: 0.8.0
      - name: external-secrets-operator.v0.7.2
        version: 0.7.2
      - name: external-secrets-operator.v0.7.1
        version: 0.7.1
      - name: external-secrets-operator.v0.7.0
        version: 0.7.0
      - name: external-secrets-operator.v0.7.0-rc1
        version: 0.7.0-rc1
      - name: external-secrets-operator.v0.6.1
        version: 0.6.1
      - name: external-secrets-operator.v0.6.0
        version: 0.6.0
      - name: external-secrets-operator.v0.5.9
        version: 0.5.9
      - name: external-secrets-operator.v0.5.8
        version: 0.5.8
      - name: external-secrets-operator.v0.5.7
        version: 0.5.7
      - name: external-secrets-operator.v0.5.6
        version: 0.5.6
      - name: external-secrets-operator.v0.5.5
        version: 0.5.5
      - name: external-secrets-operator.v0.5.4
        version: 0.5.4
      - name: external-secrets-operator.v0.5.3
        version: 0.5.3
      - name: external-secrets-operator.v0.5.2
        version: 0.5.2
      - name: external-secrets-operator.v0.5.1
        version: 0.5.1
      - name: external-secrets-operator.v0.5.0
        version: 0.5.0
      - name: external-secrets-operator.v0.4.4
        version: 0.4.4
      - name: external-secrets-operator.v0.4.2
        version: 0.4.2
      - name: external-secrets-operator.v0.4.1
        version: 0.4.1
      - name: external-secrets-operator.v0.3.11
        version: 0.3.11
      - name: external-secrets-operator.v0.3.10
        version: 0.3.10
      - name: external-secrets-operator.v0.3.9
        version: 0.3.9
      name: stable
    defaultChannel: alpha
    packageName: external-secrets-operator
    provider:
      name: External Secrets
      url: https://external-secrets.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/kuadrant-operator
    name: rhcl-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhcl-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "authpolicy-sample"
                },
                "spec": {
                  "rules": {
                    "authentication": {
                      "apikey": {
                        "apiKey": {
                          "selector": {}
                        },
                        "credentials": {
                          "authorizationHeader": {
                            "prefix": "APIKEY"
                          }
                        }
                      }
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "DNSPolicy",
                "metadata": {
                  "name": "dnspolicy-sample"
                },
                "spec": {
                  "healthCheck": {
                    "protocol": "HTTP"
                  },
                  "providerRefs": [
                    {
                      "name": "provider-ref"
                    }
                  ],
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "ratelimitpolicy-sample"
                },
                "spec": {
                  "limits": {
                    "toys": {
                      "rates": [
                        {
                          "limit": 50,
                          "window": "1m"
                        }
                      ]
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "TLSPolicy",
                "metadata": {
                  "name": "tlspolicy-sample"
                },
                "spec": {
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "self-signed-ca"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/rhcl-rhel9-operator@sha256:92f14fd8de74bc72c407947cf5fa4739a7b455e5d351baeb64521e4c07f02a9e
          createdAt: "2025-04-28T13:10:44Z"
          description: Red Hat Connectivity Link enables you to secure, protect, and
            connect your APIs and applications in multicluster, multicloud, and hybrid
            cloud environments
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthPolicy enables authentication and authorization for service
              workloads in a Gateway API network
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1
          - description: DNSPolicy configures how North-South based traffic should
              be balanced and reach the gateways
            displayName: DNSPolicy
            kind: DNSPolicy
            name: dnspolicies.kuadrant.io
            version: v1
          - description: Kuadrant configures installations of Kuadrant Service Protection
              components
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: RateLimitPolicy enables rate limiting for service workloads
              in a Gateway API network
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1
          - description: TLSPolicy provides tls for gateway listeners by managing
              the lifecycle of tls certificates
            displayName: TLSPolicy
            kind: TLSPolicy
            name: tlspolicies.kuadrant.io
            version: v1
        description: Red Hat Connectivity Link enables you to secure, protect, and
          connect your APIs and applications in multicluster, multicloud, and hybrid
          cloud environments
        displayName: Red Hat Connectivity Link
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: GitHub
          url: https://github.com/kuadrant/kuadrant-operator
        - name: Documentation
          url: https://github.com/Kuadrant/kuadrant-operator/tree/main/docs
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - registry.redhat.io/rhcl-1/rhcl-console-plugin-rhel9@sha256:ae92d638d993d782986e761d5aa3c7c5da2b8af0aabc1decb332a4f1487dacd3
        - registry.redhat.io/rhcl-1/rhcl-rhel9-operator@sha256:92f14fd8de74bc72c407947cf5fa4739a7b455e5d351baeb64521e4c07f02a9e
        - registry.redhat.io/rhcl-1/wasm-shim-rhel9@sha256:845bb8af57f3d219aa09b9c0bb20fa945306f3b413e34e196760c8cdf624a532
        version: 1.1.0
      entries:
      - name: rhcl-operator.v1.1.0
        version: 1.1.0
      - name: rhcl-operator.v1.0.2
        version: 1.0.2
      - name: rhcl-operator.v1.0.1
        version: 1.0.1
      - name: rhcl-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: rhcl-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/kuadrant-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: akaris
      provider-url: https://github.com/andreaskaris/sosreport-operator
    name: sosreport-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sosreport-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "support.openshift.io/v1alpha1",
                "kind": "Sosreport",
                "metadata": {
                  "name": "sosreport-sample"
                },
                "spec": {
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Sosreport is the Schema for the sosreports API
            displayName: Sosreport
            kind: Sosreport
            name: sosreports.support.openshift.io
            version: v1alpha1
        description: Operator that facilitates the generation of sosreports
        displayName: sosreport-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sosreport
        - debugging
        - troubleshooting
        links:
        - name: Sosreport Operator
          url: https://github.com/andreaskaris/sosreport-operator
        maintainers:
        - email: ak.karis@gmail.com
          name: Andreas Karis
        maturity: alpha
        provider:
          name: akaris
          url: https://github.com/andreaskaris/sosreport-operator
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/akaris/sosreport-operator:0.0.2
        version: 0.0.2
      entries:
      - name: sosreport-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: sosreport-operator
    provider:
      name: akaris
      url: https://github.com/andreaskaris/sosreport-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-rds-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-rds-controller.v1.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rds.services.k8s.aws/v1alpha1",
                "kind": "DBParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "rds.services.k8s.aws/v1alpha1",
                "kind": "DBSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/rds-controller:1.6.2
          createdAt: "2025-09-22T17:27:42Z"
          description: AWS RDS controller is a service controller for managing RDS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DBClusterEndpoint represents the state of an AWS rds DBClusterEndpoint
              resource.
            displayName: DBClusterEndpoint
            kind: DBClusterEndpoint
            name: dbclusterendpoints.rds.services.k8s.aws
            version: v1alpha1
          - description: DBClusterParameterGroup represents the state of an AWS rds
              DBClusterParameterGroup resource.
            displayName: DBClusterParameterGroup
            kind: DBClusterParameterGroup
            name: dbclusterparametergroups.rds.services.k8s.aws
            version: v1alpha1
          - description: DBCluster represents the state of an AWS rds DBCluster resource.
            displayName: DBCluster
            kind: DBCluster
            name: dbclusters.rds.services.k8s.aws
            version: v1alpha1
          - description: DBClusterSnapshot represents the state of an AWS rds DBClusterSnapshot
              resource.
            displayName: DBClusterSnapshot
            kind: DBClusterSnapshot
            name: dbclustersnapshots.rds.services.k8s.aws
            version: v1alpha1
          - description: DBInstance represents the state of an AWS rds DBInstance
              resource.
            displayName: DBInstance
            kind: DBInstance
            name: dbinstances.rds.services.k8s.aws
            version: v1alpha1
          - description: DBParameterGroup represents the state of an AWS rds DBParameterGroup
              resource.
            displayName: DBParameterGroup
            kind: DBParameterGroup
            name: dbparametergroups.rds.services.k8s.aws
            version: v1alpha1
          - description: DBProxy represents the state of an AWS rds DBProxy resource.
            displayName: DBProxy
            kind: DBProxy
            name: dbproxies.rds.services.k8s.aws
            version: v1alpha1
          - description: DBSnapshot represents the state of an AWS rds DBSnapshot
              resource.
            displayName: DBSnapshot
            kind: DBSnapshot
            name: dbsnapshots.rds.services.k8s.aws
            version: v1alpha1
          - description: DBSubnetGroup represents the state of an AWS rds DBSubnetGroup
              resource.
            displayName: DBSubnetGroup
            kind: DBSubnetGroup
            name: dbsubnetgroups.rds.services.k8s.aws
            version: v1alpha1
          - description: GlobalCluster represents the state of an AWS rds GlobalCluster
              resource.
            displayName: GlobalCluster
            kind: GlobalCluster
            name: globalclusters.rds.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Relational Database Service ("RDS") resources in AWS from within your Kubernetes cluster.

          **About Amazon RDS**

          Amazon Relational Database Service (Amazon RDS) makes it easy to set up, operate, and scale a relational database in the cloud. It provides cost-efficient and resizable capacity while automating time-consuming administration tasks such as hardware provisioning, database setup, patching and backups. It frees you to focus on your applications so you can give them the fast performance, high availability, security and compatibility they need.

          Amazon RDS is available on several database instance types - optimized for memory, performance or I/O -and provides you with six familiar database engines to choose from, including Amazon Aurora, PostgreSQL, MySQL, MariaDB, Oracle Database, and SQL Server. You can use the AWS Database Migration Service to easily migrate or replicate your existing databases to Amazon RDS.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon RDS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rds
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon RDS Developer Resources
          url: https://aws.amazon.com/rds/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: rds maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/rds-controller:1.6.2
        version: 1.6.2
      entries:
      - name: ack-rds-controller.v1.6.2
        version: 1.6.2
      - name: ack-rds-controller.v1.6.1
        version: 1.6.1
      - name: ack-rds-controller.v1.6.0
        version: 1.6.0
      - name: ack-rds-controller.v1.5.0
        version: 1.5.0
      - name: ack-rds-controller.v1.4.26
        version: 1.4.26
      - name: ack-rds-controller.v1.4.25
        version: 1.4.25
      - name: ack-rds-controller.v1.4.24
        version: 1.4.24
      - name: ack-rds-controller.v1.4.22
        version: 1.4.22
      - name: ack-rds-controller.v1.4.21
        version: 1.4.21
      - name: ack-rds-controller.v1.4.20
        version: 1.4.20
      - name: ack-rds-controller.v1.4.19
        version: 1.4.19
      - name: ack-rds-controller.v1.4.18
        version: 1.4.18
      - name: ack-rds-controller.v1.4.17
        version: 1.4.17
      - name: ack-rds-controller.v1.4.16
        version: 1.4.16
      - name: ack-rds-controller.v1.4.15
        version: 1.4.15
      - name: ack-rds-controller.v1.4.14
        version: 1.4.14
      - name: ack-rds-controller.v1.4.13
        version: 1.4.13
      - name: ack-rds-controller.v1.4.12
        version: 1.4.12
      - name: ack-rds-controller.v1.4.11
        version: 1.4.11
      - name: ack-rds-controller.v1.4.10
        version: 1.4.10
      - name: ack-rds-controller.v1.4.8
        version: 1.4.8
      - name: ack-rds-controller.v1.4.7
        version: 1.4.7
      - name: ack-rds-controller.v1.4.6
        version: 1.4.6
      - name: ack-rds-controller.v1.4.4
        version: 1.4.4
      - name: ack-rds-controller.v1.4.3
        version: 1.4.3
      - name: ack-rds-controller.v1.4.2
        version: 1.4.2
      - name: ack-rds-controller.v1.4.1
        version: 1.4.1
      - name: ack-rds-controller.v1.4.0
        version: 1.4.0
      - name: ack-rds-controller.v1.3.0
        version: 1.3.0
      - name: ack-rds-controller.v1.2.5
        version: 1.2.5
      - name: ack-rds-controller.v1.2.4
        version: 1.2.4
      - name: ack-rds-controller.v1.2.3
        version: 1.2.3
      - name: ack-rds-controller.v1.2.2
        version: 1.2.2
      - name: ack-rds-controller.v1.2.1
        version: 1.2.1
      - name: ack-rds-controller.v1.2.0
        version: 1.2.0
      - name: ack-rds-controller.v1.1.11
        version: 1.1.11
      - name: ack-rds-controller.v1.1.9
        version: 1.1.9
      - name: ack-rds-controller.v1.1.8
        version: 1.1.8
      - name: ack-rds-controller.v1.1.7
        version: 1.1.7
      - name: ack-rds-controller.v1.1.6
        version: 1.1.6
      - name: ack-rds-controller.v1.1.5
        version: 1.1.5
      - name: ack-rds-controller.v1.1.4
        version: 1.1.4
      - name: ack-rds-controller.v1.1.3
        version: 1.1.3
      - name: ack-rds-controller.v1.1.2
        version: 1.1.2
      - name: ack-rds-controller.v1.1.1
        version: 1.1.1
      - name: ack-rds-controller.v1.1.0
        version: 1.1.0
      - name: ack-rds-controller.v1.0.0
        version: 1.0.0
      - name: ack-rds-controller.v0.1.3
        version: 0.1.3
      - name: ack-rds-controller.v0.1.1
        version: 0.1.1
      - name: ack-rds-controller.v0.1.0
        version: 0.1.0
      - name: ack-rds-controller.v0.0.30
        version: 0.0.30
      - name: ack-rds-controller.v0.0.29
        version: 0.0.29
      - name: ack-rds-controller.v0.0.28
        version: 0.0.28
      - name: ack-rds-controller.v0.0.27
        version: 0.0.27
      - name: ack-rds-controller.v0.0.26
        version: 0.0.26
      - name: ack-rds-controller.v0.0.25
        version: 0.0.25
      - name: ack-rds-controller.v0.0.24
        version: 0.0.24
      - name: ack-rds-controller.v0.0.23
        version: 0.0.23
      - name: ack-rds-controller.v0.0.22
        version: 0.0.22
      - name: ack-rds-controller.v0.0.21
        version: 0.0.21
      - name: ack-rds-controller.v0.0.20
        version: 0.0.20
      - name: ack-rds-controller.v0.0.19
        version: 0.0.19
      - name: ack-rds-controller.v0.0.18
        version: 0.0.18
      - name: ack-rds-controller.v0.0.17
        version: 0.0.17
      - name: ack-rds-controller.v0.0.16
        version: 0.0.16
      - name: ack-rds-controller.v0.0.15
        version: 0.0.15
      - name: ack-rds-controller.v0.0.14
        version: 0.0.14
      - name: ack-rds-controller.v0.0.13
        version: 0.0.13
      - name: ack-rds-controller.v0.0.12
        version: 0.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-rds-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: integrity-shield-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: integrity-shield-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/open-cluster-management/integrity-shield-observer",
                    "imagePullPolicy": "Always",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "provenanece": true,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  resp.status_code == 200\n  result := json.unmarshal(resp.raw_body)\n  result.allow == false\n  not is_detect_mode\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"tls_insecure_skip_verify\": true\n  })\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "requestHandlerConfig": "defaultConstraintAction:\n  audit:\n    inform: true\n  admissionControl:\n    enforce: false\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/open-cluster-management/integrity-shield-api",
                    "imagePullPolicy": "Always",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/open-cluster-management/integrity-shield-operator:0.3.0
          createdAt: "2021-10-14T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.

          [K8s Integrity Shield](https://github.com/open-cluster-management/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.

          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Detect (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster

          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).

          ## Preparations before installation

          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).


          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.

          If you want to change the settings such as default run mode (detection/enforcement) or audit interval,  please check [here](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).

          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-7b7f768bf7-fhrpg                           1/1     Running   0          20s
          integrity-shield-observer-5bc66f75f7-tn8fw                      1/1     Running   0          25s
          integrity-shield-operator-controller-manager-65b7fb58f7-j25zd   2/2     Running   0          3h5m
          ```

          After installation, you can protect Kubernetes resources by following this [document](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).

          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:

          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)

          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/open-cluster-management/integrity-shield
        - name: Operator Documentation
          url: https://github.com/open-cluster-management/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/open-cluster-management/integrity-shield-operator:0.3.0
        version: 0.3.0
      entries:
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.0
    - currentCSV: integrity-shield-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/open-cluster-management/integrity-shield-observer",
                    "imagePullPolicy": "Always",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  resp.status_code == 200\n  result := json.unmarshal(resp.raw_body)\n  result.allow == false\n  not is_detect_mode\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"tls_insecure_skip_verify\": true\n  })\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: detect\n  admissionOnly: false\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/open-cluster-management/integrity-shield-api",
                    "imagePullPolicy": "Always",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/open-cluster-management/integrity-shield-operator:0.3.1
          createdAt: "2021-10-22T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.

          [K8s Integrity Shield](https://github.com/open-cluster-management/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.

          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Detect (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster

          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).

          ## Preparations before installation

          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).


          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.

          If you want to change the settings such as default run mode (detection/enforcement) or audit interval,  please check [here](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).

          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-7b7f768bf7-fhrpg                           1/1     Running   0          20s
          integrity-shield-observer-5bc66f75f7-tn8fw                      1/1     Running   0          25s
          integrity-shield-operator-controller-manager-65b7fb58f7-j25zd   2/2     Running   0          3h5m
          ```

          After installation, you can protect Kubernetes resources by following this [document](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).

          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:

          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)

          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/open-cluster-management/integrity-shield
        - name: Operator Documentation
          url: https://github.com/open-cluster-management/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/open-cluster-management/integrity-shield-operator:0.3.1
        version: 0.3.1
      entries:
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.1
    - currentCSV: integrity-shield-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/stolostron/integrity-shield-observer",
                    "imagePullPolicy": "IfNotPresent",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  should_be_denied(resp)\n  not is_detect_mode\n  result := get_result(resp)\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"timeout\": \"2500ms\",\n    \"raise_error\": false,\n    \"tls_ca_cert\": REPLACE_WITH_CA_CERT\n  })\n}\n\n# get decision result from API response\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  not is_null(raw_body)\n  result := json.unmarshal(raw_body)\n}\n# if error, return it as result\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  is_null(raw_body)\n  error_msg := object.get(resp, \"error\", \"\")\n  result := {\"http_error\":error_msg}\n}\n\n# if API response is null, deny\nshould_be_denied(resp) {\n  is_null(resp)\n}\n\n# if API response has status_code other than 200, deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  resp.status_code != 200\n}\n\n# if decision result is \"allow == false\" or not set, then deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  result := get_result(resp)\n  allow := object.get(result, \"allow\", false)\n  allow == false\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "reporter": {
                    "image": "quay.io/stolostron/integrity-shield-reporter",
                    "imagePullPolicy": "IfNotPresent",
                    "intervalSeconds": "10",
                    "name": "integrity-shield-reporter",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: inform\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\ndecisionReporterConfig:\n  enabled: true\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/stolostron/integrity-shield-api",
                    "imagePullPolicy": "IfNotPresent",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/stolostron/integrity-shield-operator:0.3.2
          createdAt: "2022-01-25T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.15.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.
          [K8s Integrity Shield](https://github.com/stolostron/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.
          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Inform (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster
          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).
          ## Preparations before installation
          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).
          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.
          If you want to change the settings such as default run mode (inform/enforce) or audit interval,  please check [here](https://github.com/stolostron/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).
          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-5799f69896-zbcjs                           2/2     Running   0          88s
          integrity-shield-observer-757f58f7d7-7pf5h                      1/1     Running   0          55s
          integrity-shield-operator-controller-manager-6f94d7d684-dg5bt   2/2     Running   0          4m27s
          ```
          After installation, you can protect Kubernetes resources by following this [document](https://github.com/stolostron/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).
          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:
          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)
          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/stolostron/integrity-shield
        - name: Operator Documentation
          url: https://github.com/stolostron/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/stolostron/integrity-shield-operator:0.3.2
        version: 0.3.2
      entries:
      - name: integrity-shield-operator.v0.3.2
        version: 0.3.2
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.2
    - currentCSV: integrity-shield-operator.v0.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": false,
                    "image": "quay.io/stolostron/integrity-shield-observer",
                    "imagePullPolicy": "IfNotPresent",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  should_be_denied(resp)\n  not is_detect_mode\n  result := get_result(resp)\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"timeout\": \"2500ms\",\n    \"raise_error\": false,\n    \"tls_ca_cert\": REPLACE_WITH_CA_CERT\n  })\n}\n\n# get decision result from API response\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  not is_null(raw_body)\n  result := json.unmarshal(raw_body)\n}\n# if error, return it as result\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  is_null(raw_body)\n  error_msg := object.get(resp, \"error\", \"\")\n  result := {\"http_error\":error_msg}\n}\n\n# if API response is null, deny\nshould_be_denied(resp) {\n  is_null(resp)\n}\n\n# if API response has status_code other than 200, deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  resp.status_code != 200\n}\n\n# if decision result is \"allow == false\" or not set, then deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  result := get_result(resp)\n  allow := object.get(result, \"allow\", false)\n  allow == false\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "reporter": {
                    "image": "quay.io/stolostron/integrity-shield-reporter",
                    "imagePullPolicy": "IfNotPresent",
                    "intervalSeconds": "10",
                    "name": "integrity-shield-reporter",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: inform\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\ndecisionReporterConfig:\n  enabled: true\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/stolostron/integrity-shield-api",
                    "imagePullPolicy": "IfNotPresent",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/stolostron/integrity-shield-operator:0.3.3
          createdAt: "2022-03-18T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.15.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.
          [K8s Integrity Shield](https://github.com/stolostron/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.
          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Inform (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster
          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).
          ## Preparations before installation
          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).
          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.
          If you want to change the settings such as default run mode (inform/enforce) or audit interval,  please check [here](https://github.com/stolostron/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).
          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-5799f69896-zbcjs                           2/2     Running   0          88s
          integrity-shield-observer-757f58f7d7-7pf5h                      1/1     Running   0          55s
          integrity-shield-operator-controller-manager-6f94d7d684-dg5bt   2/2     Running   0          4m27s
          ```
          After installation, you can protect Kubernetes resources by following this [document](https://github.com/stolostron/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).
          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:
          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)
          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/stolostron/integrity-shield
        - name: Operator Documentation
          url: https://github.com/stolostron/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/stolostron/integrity-shield-operator:0.3.3
        version: 0.3.3
      entries:
      - name: integrity-shield-operator.v0.3.3
        version: 0.3.3
      - name: integrity-shield-operator.v0.3.2
        version: 0.3.2
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.3
    defaultChannel: alpha-0.3.3
    packageName: integrity-shield-operator
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: gcp-filestore-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gcp-filestore-csi-driver-operator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel8-operator@sha256:1ec61d2988b95a10260c7932e9dc6a9d1d5b3eb4148b22afd438aa4ee75dc9b3
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure GCP Filestore CSI driver.
          olm.skipRange: '>=4.11.0-0 <4.13.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/gcp-filestore-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Operator that installs and configures the CSI driver for GCP Filestore. The GCP Filestore CSI Driver Operator is a Technology Preview feature. For more information on Red Hat Technology Preview features support scope, see https://access.redhat.com/support/offerings/techpreview/.
        displayName: GCP Filestore CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - filestore
        links:
        - name: Documentation
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: preview
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:700c0e50f86a70f506403aba75e020a17530ada284af7eee77d4b4ce3dbf4102
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:4f1e873080754ae9aebc361cd77f965e1ade369403346aa0cf4cf502082d18e4
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:ee45e11a6fe8cf60f3ac285cba1b23459ca95068bd9769042ae052c778c3e080
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:537fbc8fe2b212dc798361099d2c6fd981cc846c1cf5454b703ba08d488b61f8
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel8@sha256:a1a4e7b93d6609561a7a2cd732ce772bf0047448e5019298d66c4b985baecfb2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:767682dd3bd65651a8204c9fb732b9c48b99127189992b43abc6ecce027f4589
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel8-operator@sha256:1ec61d2988b95a10260c7932e9dc6a9d1d5b3eb4148b22afd438aa4ee75dc9b3
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:6f1b497772b953468f0b816face8810ace8ea44fa91dd4a3b3e583da2cac350c
        version: 4.13.0-202309181427
      entries:
      - name: gcp-filestore-csi-driver-operator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: preview
    - currentCSV: gcp-filestore-csi-driver-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9-operator@sha256:2dc13de58b391237dc48b9335d88b99f5e26d0e5edc5a664c7d39f16ca9b297d
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure GCP Filestore CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.11.0-0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/gcp-filestore-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Operator that installs and configures the CSI driver for GCP Filestore.
        displayName: GCP Filestore CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - filestore
        links:
        - name: Documentation
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:a5704d26d0b562dba9633d1c107f34e05c750c0e3a64dbcace4dab7b964fff8d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9@sha256:8e801d924071c8f2513555450e688bff60d4100b2ca56d2ba9517038a1d48d4d
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:9c3c3d2b857ce289324aee56df1ba4ab5e5ea1e76cd86ac988233463915834b4
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:78e02d937740d742eb33b08c7317f211c25020c375069d5a97344107716608d3
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:b28def976fc155dff57ff3003cc462a8e9923b325af0b319b65220c66688699e
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0627c09fe3f5a015f117fb30af1f4d17340aefa48b23994737d3901a804b63d3
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9-operator@sha256:2dc13de58b391237dc48b9335d88b99f5e26d0e5edc5a664c7d39f16ca9b297d
        version: 4.16.0-202509240123
      entries:
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202508270905
        version: 4.16.0-202508270905
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202504121405
        version: 4.16.0-202504121405
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: gcp-filestore-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-prometheusservice-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-prometheusservice-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "prometheusservice.services.k8s.aws/v1alpha1",
                "kind": "RuleGroupsNamespace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "prometheusservice.services.k8s.aws/v1alpha1",
                "kind": "Workspace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/prometheusservice-controller:1.3.1
          createdAt: "2025-09-19T21:11:38Z"
          description: AWS Prometheus controller is a service controller for managing
            Prometheus resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertManagerDefinition represents the state of an AWS prometheusservice
              AlertManagerDefinition resource.
            displayName: AlertManagerDefinition
            kind: AlertManagerDefinition
            name: alertmanagerdefinitions.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: LoggingConfiguration represents the state of an AWS prometheusservice
              LoggingConfiguration resource.
            displayName: LoggingConfiguration
            kind: LoggingConfiguration
            name: loggingconfigurations.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: RuleGroupsNamespace represents the state of an AWS prometheusservice
              RuleGroupsNamespace resource.
            displayName: RuleGroupsNamespace
            kind: RuleGroupsNamespace
            name: rulegroupsnamespaces.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: Workspace represents the state of an AWS prometheusservice
              Workspace resource.
            displayName: Workspace
            kind: Workspace
            name: workspaces.prometheusservice.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Prometheus resources in AWS from within your Kubernetes cluster.

          **About Amazon Managed Service for Prometheus**

          Amazon Managed Service for Prometheus is a serverless, Prometheus-compatible monitoring service for container metrics that makes it easier to securely monitor container environments at scale. With Amazon Managed Service for Prometheus, you can use the same open-source Prometheus data model and query language that you use today to monitor the performance of your containerized workloads, and also enjoy improved scalability, availability, and security without having to manage the underlying infrastructure.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Prometheus
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - prometheusservice
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Prometheus Developer Resources
          url: https://aws.amazon.com/prometheus/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: prometheus maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/prometheusservice-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-prometheusservice-controller.v1.3.1
        version: 1.3.1
      - name: ack-prometheusservice-controller.v1.3.0
        version: 1.3.0
      - name: ack-prometheusservice-controller.v1.2.27
        version: 1.2.27
      - name: ack-prometheusservice-controller.v1.2.26
        version: 1.2.26
      - name: ack-prometheusservice-controller.v1.2.24
        version: 1.2.24
      - name: ack-prometheusservice-controller.v1.2.23
        version: 1.2.23
      - name: ack-prometheusservice-controller.v1.2.22
        version: 1.2.22
      - name: ack-prometheusservice-controller.v1.2.20
        version: 1.2.20
      - name: ack-prometheusservice-controller.v1.2.19
        version: 1.2.19
      - name: ack-prometheusservice-controller.v1.2.17
        version: 1.2.17
      - name: ack-prometheusservice-controller.v1.2.16
        version: 1.2.16
      - name: ack-prometheusservice-controller.v1.2.15
        version: 1.2.15
      - name: ack-prometheusservice-controller.v1.2.14
        version: 1.2.14
      - name: ack-prometheusservice-controller.v1.2.13
        version: 1.2.13
      - name: ack-prometheusservice-controller.v1.2.12
        version: 1.2.12
      - name: ack-prometheusservice-controller.v1.2.11
        version: 1.2.11
      - name: ack-prometheusservice-controller.v1.2.10
        version: 1.2.10
      - name: ack-prometheusservice-controller.v1.2.9
        version: 1.2.9
      - name: ack-prometheusservice-controller.v1.2.6
        version: 1.2.6
      - name: ack-prometheusservice-controller.v1.2.5
        version: 1.2.5
      - name: ack-prometheusservice-controller.v1.2.4
        version: 1.2.4
      - name: ack-prometheusservice-controller.v1.2.3
        version: 1.2.3
      - name: ack-prometheusservice-controller.v1.2.2
        version: 1.2.2
      - name: ack-prometheusservice-controller.v1.2.1
        version: 1.2.1
      - name: ack-prometheusservice-controller.v1.2.0
        version: 1.2.0
      - name: ack-prometheusservice-controller.v1.1.0
        version: 1.1.0
      - name: ack-prometheusservice-controller.v1.0.0
        version: 1.0.0
      - name: ack-prometheusservice-controller.v0.1.3
        version: 0.1.3
      - name: ack-prometheusservice-controller.v0.1.2
        version: 0.1.2
      - name: ack-prometheusservice-controller.v0.1.1
        version: 0.1.1
      - name: ack-prometheusservice-controller.v0.1.0
        version: 0.1.0
      - name: ack-prometheusservice-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-prometheusservice-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: awss3-operator-registry
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: awss3operator.1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/screeley44/aws-s3-provisioner:v1.0.0
          createdAt: "2019-05-02 16:12:00"
          description: Manage the full lifecycle of installing, configuring and managing
            AWS S3 Provisioner.
          repository: https://github.com/yard-turkey/aws-s3-provisioner
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - description: instance of an AWS S3 Bucket
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
          - description: Request for an AWS S3 Bucket
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
        description: |
          AWS S3 Operator will deploy the AWS S3 Provisioner which will dynamically or statically
          provision AWS S3 Bucket storage and access. The operator deploys the ObjectBucket (OB) and ObjectBucketClaim (OBC)
          CustomResourceDefinitions. The OB/OBC model follows the traditional Kubernetes PV/PVC pattern, when an OBC is detected
          the operator will act on the OBC to either provision a brand new S3 Bucket in AWS or gain access to an existing
          S3 Bucket in AWS. The operator produces an OB and ConfigMap and Secret which can then be consumed by application pods.

          **Important Note**: Currently, while in preview, this operator does not
          support automatic upgrades. You must remove the old version of the operator
          manually before installing a new version.
          ## Using AWS S3 Operator

          ### Administrator Creates Secret
          This secret will contain the elevated/admin privileges needed by the provisioner
          to properly access and create S3 Buckets and IAM users and policies. The AWS Access ID
          and AWS Secret Key will be needed for this.
          1. Create the Kubernetes Secret for the Provisioner's Owner Access.
          ```yaml
          apiVersion: v1
          kind: Secret
          metadata:
            name: s3-bucket-owner [1]
            namespace: s3-provisioner [2]
          type: Opaque
          data:
            AWS_ACCESS_KEY_ID: *base64 encoded value* [3]
            AWS_SECRET_ACCESS_KEY: *base64 encoded value* [4]
          ```
          1. Name of the secret, this will be referenced in StorageClass.
          1. Namespace where the Secret will exist.
          1. Your AWS_ACCESS_KEY_ID base64 encoded.
          1. Your AWS_SECRET_ACCESS_KEY base64 encoded.
          ```
           # kubectl create -f creds.yaml
          secret/s3-bucket-owner created
          ```
          ### Administrator Creates StorageClass
          The StorageClass defines the name of the provisioner and holds other properties that are needed to provision a new bucket, including
          the Owner Secret and Namespace, and the AWS Region.
          #### Greenfield Example:
          For Greenfield, a new, dynamic bucket will be generated.
          1. Create the Kubernetes StorageClass for the Provisioner.
          ```yaml
          kind: StorageClass
          apiVersion: storage.k8s.io/v1
          metadata:
            name: s3-buckets [1]
          provisioner: aws-s3.io/bucket [2]
          parameters:
            region: us-west-1 [3]
            secretName: s3-bucket-owner [4]
            secretNamespace: s3-provisioner [5]
          reclaimPolicy: Delete [6]
          ```
          1. Name of the StorageClass, this will be referenced in the User ObjectBucketClaim.
          1. Provisioner name
          1. AWS Region that the StorageClass will serve
          1. Name of the bucket owner Secret created above
          1. Namespace where the Secret will exist
          1. reclaimPolicy (Delete or Retain) indicates if the bucket can be deleted when the OBC is deleted.
          **NOTE:** the absence of the `bucketName` Parameter key in the storage class indicates this is a new bucket and its name is based on the bucket name fields in the OBC.
          ```
           # kubectl create -f storageclass-greenfield.yaml
          storageclass.storage.k8s.io/s3-buckets created
          ```
          #### Brownfield Example:
          For brownfield, the StorageClass defines the name of the provisioner and the name of the existing bucket. It also includes other properties needed by the target
          provisioner, including: the Owner Secret and Namespace, and the AWS Region
          1. Create the Kubernetes StorageClass for the Provisioner.
          ```yaml
          kind: StorageClass
          apiVersion: storage.k8s.io/v1
          metadata:
            name: s3-existing-buckets [1]
          provisioner: aws-s3.io/bucket [2]
          parameters:
            bucketName: my-existing-bucket [3]
            region: us-west-1 [4]
            secretName: s3-bucket-owner [5]
            secretNamespace: s3-provisioner [6]
          ```
          1. Name of the StorageClass, this will be referenced in the User ObjectBucketClaim.
          1. Provisioner name
          1. Name of the existing bucket
          1. AWS Region that the StorageClass will serve
          1. Name of the bucket owner Secret created above
          1. Namespace for that bucket owner secret
          **NOTE:** the storage class's `reclaimPolicy` is ignored for existing buckets.
          ```
           # kubectl create -f storageclass-brownfield.yaml
          storageclass.storage.k8s.io/s3-buckets created
          ```
          ### User Creates ObjectBucketClaim
          An ObjectBucketClaim follows the same concept as a PVC, in that
          it is a request for Object Storage, the user doesn't need to
          concern him/herself with the underlying storage, just that
          they need access to it. The user will work with the cluster/storage
          administrator to get the proper StorageClass needed and will
          then request access via the OBC.
          #### Greenfield Request Example:
          1. Create the ObjectBucketClaim.
          ```yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucketClaim
          metadata:
            name: myobc [1]
            namespace: s3-provisioner [2]
          spec:
            generateBucketName: mybucket [3]
            bucketName: my-awesome-bucket [4]
            storageClassName: s3-buckets [5]
          ```
          1. Name of the OBC
          1. Namespace of the OBC
          1. Name prepended to a random string used to generate a bucket name. It is ignored if bucketName is defined
          1. Name of new bucket which must be unique across all AWS regions, otherwise an error occurs when creating the bucket. If present, this name overrides `generateName`
          1. StorageClass name
          **NOTE:** if both `generateBucketName` and `bucketName` are omitted, and the storage class does _not_ define a bucket name, then a new, random bucket name is generated with no prefix.
          ```
           # kubectl create -f obc-brownfield.yaml
          objectbucketclaim.objectbucket.io/myobc created
          ```
          #### Brownfield Request Example:
          1. Create the ObjectBucketClaim.
          ```yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucketClaim
          metadata:
            name: myobc [1]
            namespace: s3-provisioner [2]
          spec:
            storageClassName: s3-existing-buckets [3]
          ```
          1. Name of the OBC
          1. Namespace of the OBC
          1. StorageClass name
          **NOTE:** in the OBC here there is no reference to the bucket's name. This is defined in the storage class and is not a concern of the user creating the claim to this bucket.  An OBC does have fields for defining a bucket name for greenfield use only.
          ```
           # kubectl create -f obc-brownfield.yaml
          objectbucketclaim.objectbucket.io/myobc created
          ```
          ### Results and Recap
          Let's pause for a moment and digest what just happened.
          After creating the OBC, and assuming the S3 provisioner is running, we now have
          the following Kubernetes resources:
          .  a global ObjectBucket (OB) which contains: bucket endpoint info (including region and bucket name), a reference to the OBC, and a reference to the storage class. Unique to S3, the OB also contains the bucket Amazon Resource Name (ARN).Note: there is always a 1:1 relationship between an OBC and an OB.
          .  a ConfigMap in the same namespace as the OBC, which contains the same endpoint data found in the OB.
          .  a Secret in the same namespace as the OBC, which contains the AWS key-pairs needed to access the bucket.
          And of course, we have a *new* AWS S3 Bucket which you should be able to see via the AWS Console.
          *ObjectBucket*
          ```yaml
           # kubectl get ob obc-s3-provisioner-my-awesome-bucket -o yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucket
          metadata:
            creationTimestamp: "2019-04-03T15:42:22Z"
            generation: 1
            name: obc-s3-provisioner-my-awesome-bucket
            resourceVersion: "15057"
            selfLink: /apis/objectbucket.io/v1alpha1/objectbuckets/obc-s3-provisioner-my-awesome-bucket
            uid: 0bfe8e84-576d-4c4e-984b-f73c4460f736
          spec:
            Connection:
              additionalState:
                ARN: arn:aws:iam::<accountid>:policy/my-awesome-bucket-vSgD5 [1]
                UserName: my-awesome-bucket-vSgD5 [2]
              endpoint:
                additionalConfig: null
                bucketHost: s3-us-west-1.amazonaws.com
                bucketName: my-awesome-bucket [3]
                bucketPort: 443
                region: us-west-1
                ssl: true
                subRegion: ""
            claimRef: null [4]
            reclaimPolicy: null
            storageClassName: s3-buckets [5]
          ```
          1. The AWS Policy created for this user and bucket.
          1. The new user generated by the Provisioner to access this existing bucket.
          1. The bucket name.
          1. The reference to the OBC (not filled in yet).
          1. The reference to the StorageClass used.
          *ConfigMap*
          ```yaml
           # kubectl get cm myobc -n s3-provisioner -o yaml
          apiVersion: v1
          data:
            BUCKET_HOST: s3-us-west-1.amazonaws.com [1]
            BUCKET_NAME: my-awesome-bucket [2]
            BUCKET_PORT: "443"
            BUCKET_REGION: us-west-1
            BUCKET_SSL: "true"
            BUCKET_SUBREGION: ""
          kind: ConfigMap
          metadata:
            creationTimestamp: "2019-04-01T19:11:38Z"
            finalizers:
            - objectbucket.io/finalizer
            name: my-awesome-bucket
            namespace: s3-provisioner
            resourceVersion: "892"
            selfLink: /api/v1/namespaces/s3-provisioner/configmaps/my-awesome-bucket
            uid: 2edcc58a-aff8-4a29-814a-ffbb6439a9cd
          ```
          1. The AWS S3 host.
          1. The name of the new bucket we are gaining access to.
          *Secret*
          ```yaml
           # kubectl get secret my-awesome-bucket -n s3-provisioner -o yaml
          apiVersion: v1
          data:
            AWS_ACCESS_KEY_ID: *the_new_access_id* [1]
            AWS_SECRET_ACCESS_KEY: *the_new_access_key_value* [2]
          kind: Secret
          metadata:
            creationTimestamp: "2019-04-03T15:42:22Z"
            finalizers:
            - objectbucket.io/finalizer
            name: my-awesome-bucket
            namespace: s3-provisioner
            resourceVersion: "15058"
            selfLink: /api/v1/namespaces/s3-provisioner/secrets/screeley-provb-5
            uid: 225c71a5-9d75-4ccc-b41f-bfe91b272a13
          type: Opaque
          ```
          1. The new generated AWS Access Key ID.
          1. The new generated AWS Secret Access Key.
          What happened in AWS? The first thing we do on any OBC request is
          create a new IAM user and generate Access ID and Secret Keys.
          This allows us to also better control ACLs. We also create a policy
          in IAM which we then attach to the user and bucket. We also created a new bucket, called *my-awesome-bucket*.
          When the OBC is deleted all of its Kubernetes and AWS resources will also be deleted, which includes:
          the generated OB, Secret, ConfigMap, IAM user, and policy.
          If the _retainPolicy_ on the StorageClass for this bucket is *"Delete"*, then, in addition to the above cleanup, the physical bucket is also deleted.
          **NOTE:** The actual bucket is only deleted if the Storage Class's _reclaimPolicy_ is "Delete".
          ### User Creates Pod
          Now that we have our bucket and connection/access information, a pod
          can be used to access the bucket. This can be done in several different
          ways, but the key here is that the provisioner has provided the proper
          endpoints and keys to access the bucket. The user then simply references
          the keys.
          1. Create a Sample Pod to Access the Bucket.
          ```yaml
          apiVersion: v1
          kind: Pod
          metadata:
            name: photo1
            labels:
              name: photo1
          spec:
            containers:
            - name: photo1
              image: docker.io/screeley44/photo-gallery:latest
              imagePullPolicy: Always
              envFrom:
              - configMapRef:
                  name: my-awesome-bucket <1>
              - secretRef:
                  name: my-awesome-bucket <2>
              ports:
              - containerPort: 3000
                protocol: TCP
          ```
          1. Name of the generated configmap from the provisioning process
          1. Name of the generated secret from the provisioning process
          *[Note]* Generated ConfigMap and Secret are same name as the OBC!
          Lastly, expose the pod as a service so you can access the url from a browser. In this example,
          I exposed as a LoadBalancer
          ```
            # kubectl expose pod photo1 --type=LoadBalancer --name=photo1 -n your-namespace
          ```
          To access via a url use the EXTERNAL-IP
          ```
            # kubectl get svc photo1
            NAME                         TYPE           CLUSTER-IP       EXTERNAL-IP                                                               PORT(S)          AGE
            photo1                       LoadBalancer   100.66.124.105   a00c53ccb3c5411e9b6550a7c0e50a2a-2010797808.us-east-1.elb.amazonaws.com   3000:32344/TCP   6d
          ```
          **NOTE:** This is just one example of a Pod that can utilize the bucket information,
          there are several ways that these pod applications can be developed and therefore
          the method of getting the actual values needed from the Secrets and ConfigMaps
          will vary greatly, but the idea remains the same, that the pod consumes the generated
          ConfigMap and Secret created by the provisioner.
        displayName: AWS S3 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - awss3provisioner
        - aws
        - S3
        - provisioner
        - buckets
        links:
        - name: AWS S3 Provisioner
          url: https://github.com/yard-turkey/aws-s3-provisioner
        - name: Provisioner Examples
          url: https://github.com/yard-turkey/aws-s3-provisioner/tree/master/examples
        - name: Operator Documentation
          url: https://github.com/yard-turkey/awss3operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - quay.io/screeley44/aws-s3-provisioner:v1.0.0
        version: 1.0.1
      entries:
      - name: awss3operator.1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: awss3-operator-registry
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Managed OpenShift Black Belt
      provider-url: https://github.com/rh-mobb/ocm-operator
    name: ocm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocm-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "GitLabIdentityProvider",
                "metadata": {
                  "name": "gitlab-sample"
                },
                "spec": {
                  "clientID": "gitlab",
                  "clientSecret": {
                    "name": "gitlab"
                  },
                  "clusterName": "my-cluster",
                  "displayName": "gitlab-sample",
                  "mappingMethod": "claim",
                  "url": "https://gitlab.com"
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "LDAPIdentityProvider",
                "metadata": {
                  "name": "ldap-sample"
                },
                "spec": {
                  "bindDN": "CN=test,OU=Users,DC=example,DC=com",
                  "bindPassword": {
                    "name": "test"
                  },
                  "clusterName": "my-cluster",
                  "displayName": "ldap-sample",
                  "mappingMethod": "claim",
                  "url": "ldap://test.example.com:389"
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "MachinePool",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "aws": {
                    "spotInstances": {
                      "enabled": false
                    }
                  },
                  "clusterName": "my-cluster",
                  "instanceType": "m5.xlarge",
                  "labels": {
                    "owner": "my-user"
                  },
                  "maximumNodesPerZone": 1,
                  "minimumNodesPerZone": 1,
                  "wait": false
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "ROSACluster",
                "metadata": {
                  "name": "rosa-sample"
                },
                "spec": {
                  "accountID": "111111111111",
                  "defaultMachinePool": {
                    "instanceType": "m5.xlarge",
                    "labels": {
                      "owner": "my-owner"
                    },
                    "minimumNodesPerZone": 2
                  },
                  "displayName": "my-cluster",
                  "iam": {
                    "userRole": "arn:aws:iam::111111111111:role/ManagedOpenShift-User-my_user-Role"
                  },
                  "tags": {
                    "owner": "my-owner"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ocm-operator:v0.1.0
          createdAt: "2023-06-02T17:32:25Z"
          operatorframework.io/suggested-namespace: ocm-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ocm-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLabIdentityProvider is the Schema for the gitlabidentityproviders
              API.
            displayName: Git Lab Identity Provider
            kind: GitLabIdentityProvider
            name: gitlabidentityproviders.ocm.mobb.redhat.com
            version: v1alpha1
          - description: LDAPIdentityProvider is the Schema for the ldapidentityproviders
              API.
            displayName: LDAPIdentity Provider
            kind: LDAPIdentityProvider
            name: ldapidentityproviders.ocm.mobb.redhat.com
            version: v1alpha1
          - description: MachinePool is the Schema for the machinepools API.
            displayName: Machine Pool
            kind: MachinePool
            name: machinepools.ocm.mobb.redhat.com
            version: v1alpha1
          - description: ROSACluster is the Schema for the clusters API.
            displayName: ROSACluster
            kind: ROSACluster
            name: rosaclusters.ocm.mobb.redhat.com
            version: v1alpha1
        description: "# Quick Start\n\n## Prerequisites\n\n- [oc](https://docs.openshift.com/container-platform/4.8/cli_reference/openshift_cli/getting-started-cli.html)\n-
          [aws](https://docs.aws.amazon.com/cli/latest/userguide/getting-started-install.html)\n-
          [ROSA Cluster (4.12+)](https://mobb.ninja/docs/rosa/sts/)\n\n> **Warning:**\n>
          This is only tested on ROSA, but may work on other Kubernetes and OpenShift
          variants.  It requires\n> an AWS STS-enabled cluster.  Additionally, OpenShift
          versions 4.12+ are recommended, as they \n> enable [CRD CEL validation](https://kubernetes.io/blog/2022/09/23/crd-validation-rules-beta/).
          \ Versions\n> prior to 4.12 provide a lower version of Kubernetes that does
          not enable this feature.  They may \n> work, but provide no input validation
          when submitting custom resources to the cluster.\n\nBefore installing this
          operator, there are a couple secrets that must be created.\n\n1. OCM Token:
          used to authenticate against the OCM API.  The controller will not start
          without\nthis token as it used for all calls into OCM.\n2. AWS Credentials:
          uses an assume role to a role created in the instructions below in order\nto
          access AWS resources.  This is needed for cluster creation.  See \n[Create
          AWS IAM Policies and Roles](#create-aws-iam-policies-and-roles) for more
          details.\n\n> **Note:**\n> Certain custom resources have their own specific
          prerequisites.  Please use `oc explain` \n> or read the docs [here](https://github.com/rh-mobb/ocm-operator/tree/main/docs)
          for more details.\n\n### Create OCM Token Secret\n\n1. Create a namespace
          where you wish to install the operator:\n\n```bash\noc new-project ocm-operator\n```\n\n2.
          Create a secret containing the OCM_TOKEN.  This token can be obtained form
          \nhttps://console.redhat.com/openshift/token and is used by the operator
          to authenticate \nagainst the OCM API.  This token must exist in the same
          namespace that the operator \nis running and be named `ocm-token`.  It also
          expects the key to be called `OCM_TOKEN` \nas the operator is expecting
          this value as an environment variable.\n\n```bash\noc create secret generic
          ocm-token \\\n  --namespace=ocm-operator \\\n  --from-literal=OCM_TOKEN=${MY_OCM_TOKEN}\n```\n\n###
          Create AWS IAM Policies and Roles\n\nThe operator will need to elevate privileges
          in order to perform things like \ncreating the operator-roles for the clusters.
          \ Because of this, the operator \nmust have a specific role created to allow
          it these permissions.\n\n**NOTE:** please understand what you are doing
          if you deviate from the known good \npolicies.  If errors or more stringent
          security lockdowns are found, please submit a PR \nso that we can get this
          fixed.\n\n1. Set `ACCOUNT_ID` environment variable to define your AWS account
          ID:\n\n```bash\nexport ACCOUNT_ID=111111111111\n```\n\n2. Create the permissions
          boundary.  Because the operator needs to create policies and \nroles, the
          boundary ensures that the operator is not allowed to create additional\npermissions
          outside of the defined boundary.  The sample permission set is located \nat
          `test/aws/boundary.json` in this repository:\n\n```bash\nboundary=$(curl
          -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/boundary.json\")\naws
          iam create-policy \\\n  --policy-name \"OCMOperatorBoundary\" \\\n  --policy-document
          \"$boundary\"\n```\n\n3. Create the policy.  This policy sets what the operator
          is allowed to do.  For any \n`iam` permission, the boundary created above
          is used.  The sample permission set is \nlocated at `test/aws/iam_policy.json`
          in this repository:\n\n```bash\npolicy=$(curl -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/iam_policy.json\")\naws
          iam create-policy \\\n  --policy-name \"OCMOperator\" \\\n  --policy-document
          \"$policy\"\n```\n\n4. Create the role using a trust policy and attach the
          previously created role.  The trust \npolicy is located at `test/aws/trust_policy.json`
          in this repository.  Please note that \nthe trust policy requires an OIDC
          configuration.  The OIDC configuration refers to \n**where the operator
          is running, NOT what it is provisioning**:\n\n```bash\ntrust_policy=$(curl
          -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/trust_policy.json\")\naws
          iam create-role \\\n    --role-name OCMOperator \\\n    --assume-role-policy-document
          \"$trust_policy\"\naws iam attach-role-policy \\\n    --role-name OCMOperator
          \\\n    --policy-arn arn:aws:iam::$ACCOUNT_ID:policy/OCMOperator\n```\n\n5.
          Finally, create the secret containing the assume role credentials.  The
          previous steps allow \nthe operator to assume the role you created in the
          previous step with the permissions created \nvia the previous policies.\n\n```bash\ncat
          <<EOF > /tmp/credentials\n[default]\nrole_arn = arn:aws:iam::$ACCOUNT_ID:role/OCMOperator\nweb_identity_token_file
          = /var/run/secrets/openshift/serviceaccount/token\nEOF\n\noc create secret
          generic aws-credentials \\\n  --namespace=ocm-operator \\\n  --from-file=credentials=/tmp/credentials\n```\n\n##
          Install the Operator\n\n```bash\ncat <<EOF | oc apply -f -\napiVersion:
          operators.coreos.com/v1alpha1\nkind: Subscription\nmetadata:\n  name: ocm-operator\n
          \ namespace: ocm-operator\nspec:\n  channel: v0.1\n  installPlanApproval:
          Automatic\n  name: ocm-operator\n  source: community-operators\n  sourceNamespace:
          ocm-operator\n  startingCSV: ocm-operator.v0.1.0\nEOF\n```\n\n## Provision
          Resources\n\nOnce the operator is available and running, you can provision
          any of the \nresources that it manages.  Also note that documentation is
          always available \nonce the operator is installed as well by using the `oc
          explain` command.  For \nexample, `oc explain rosacluster.spec.clusterName`
          will give you detailed documentation \nabout what the field does.\n\nSee
          the following documentation for details:\n\n* [ROSA Clusters](https://github.com/rh-mobb/ocm-operator/blob/main/docs/clusters.md)\n*
          [Machine Pools](https://github.com/rh-mobb/ocm-operator/blob/main/docs/machinepools.md)\n*
          [Identity Providers](https://github.com/rh-mobb/ocm-operator/blob/main/docs/identityproviders.md)\n"
        displayName: OpenShift Cluster Manager Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ocm
        - rosa
        - redhat
        - machinepool
        - mobb
        links:
        - name: OCM Machine Pool Operator
          url: https://github.com/rh-mobb/ocm-operator
        maintainers:
        - email: dscott@redhat.com
          name: Dustin Scott
        maturity: alpha
        provider:
          name: Managed OpenShift Black Belt
          url: https://github.com/rh-mobb/ocm-operator
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/mobb/ocm-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: ocm-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ocm-operator
    provider:
      name: Managed OpenShift Black Belt
      url: https://github.com/rh-mobb/ocm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SAP LinuxLab - Red Hat
      provider-url: https://github.com/sap-linuxlab
    name: sap-hana-express-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sap-hana-express-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "db.sap-redhat.io/v1alpha1",
                "kind": "HanaExpress",
                "metadata": {
                  "name": "hanaexpress-sample"
                },
                "spec": {
                  "credential": {
                    "key": "hxepasswd.json",
                    "name": "hxepasswd"
                  },
                  "pvcSize": "1Gi"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/redhat-sap-cop/sap-hanaexpress-operator:0.1.4
          createdAt: "2023-07-16T20:53:02Z"
          description: Manage the lifecycle of SAP Hana Express
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HanaExpress is the Schema for the hanaexpresses API
            displayName: Hana Express
            kind: HanaExpress
            name: hanaexpresses.db.sap-redhat.io
            version: v1alpha1
        description: Operator for managing the lifecycle of SAP Hana Express on Kubernetes
          platform
        displayName: SAP Hana Express Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Database
        - SAP
        links:
        - name: Sap Hana Express Operator
          url: https://github.com/redhat-sap/sap-hana-express-operator
        maintainers:
        - email: mjiao@redhat.com
          name: RedHat@SAPLinuxLab
        maturity: alpha
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/redhat-sap-cop/sap-hanaexpress-operator:0.1.4
        version: 1.0.2
      entries:
      - name: sap-hana-express-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: sap-hana-express-operator
    provider:
      name: SAP LinuxLab - Red Hat
      url: https://github.com/sap-linuxlab
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-registry-3
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-3.v3.0.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "simple"
                },
                "spec": {
                  "app": {
                    "ingress": {
                      "host": "simple-app.apps.cluster.example"
                    }
                  },
                  "ui": {
                    "ingress": {
                      "host": "simple-ui.apps.cluster.example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-3-operator:3.0.15
          createdAt: "2025-09-22T13:04:40Z"
          description: Deploy and manage Apicurio Registry 3 on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Apicurio Registry
              version 3.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Apicurio Registry 3

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Allows you manage event schemas and API designs using the Apicurio Registry web console, REST API, Maven plug-in, or Java client
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, MySQL, SQL Server, in-memory
          - Compatible with Confluent APIs
          - Runs on a lightweight Quarkus platform

          ## Apicurio Registry 3 Operator

          Provides a quick and easy way to deploy and manage Apicurio Registry 3 on Kubernetes.

          **Features**
          - Supports installation and configuration of Apicurio Registry 3
          - Perform a rolling upgrade of Apicurio Registry 3

          ## Prerequisites

          This operator does not deploy storage for Apicurio Registry 3. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio Registry 3 and Apicurio Registry 3 Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Apicurio Registry 3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-3-operator:3.0.15
        version: 3.0.15
      entries:
      - name: apicurio-registry-3.v3.0.15
        version: 3.0.15
      - name: apicurio-registry-3.v3.0.14
        version: 3.0.14
      - name: apicurio-registry-3.v3.0.12
        version: 3.0.12
      - name: apicurio-registry-3.v3.0.9
        version: 3.0.9
      - name: apicurio-registry-3.v3.0.8
        version: 3.0.8
      - name: apicurio-registry-3.v3.0.7
        version: 3.0.7
      name: 3.x
    defaultChannel: 3.x
    packageName: apicurio-registry-3
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM Edge
      provider-url: ""
    name: infrastructure-asset-orchestrator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infrastructure-asset-orchestrator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infrastructure.asset.orchestrator.com/v1",
                "kind": "Generatekey",
                "metadata": {
                  "name": "modelbuilder-token"
                },
                "spec": {}
              },
              {
                "apiVersion": "infrastructure.asset.orchestrator.com/v1",
                "kind": "Mb-broker-service",
                "metadata": {
                  "name": "mb-broker-service-sample"
                },
                "spec": {
                  "env_type": "prod",
                  "cloud_type": "IBM",
                  "external_db": {
                    "enabled": false
                  },
                  "postgres": {
                    "enable_pg_tls": true,
                    "storage_class": "",
                    "storage_size": "20G",
                    "backup_type": "incremental",
                    "backup_frequency": "@daily"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/iao-operator@sha256:5cc395f22d86641e175339deb134af6c74de92b81470a65a5c736c8005d645fd
          createdAt: "2021-08-19T12:00:00Z"
          olm.skipRange: <1.0.2
          operators.operatorframework.io/builder: operator-sdk-v1.12.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/infrastructure-asset-orchestrator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Generate key CR
            displayName: GenerateKey
            kind: Generatekey
            name: generatekeys.infrastructure.asset.orchestrator.com
            version: v1
          - description: Deploy MB broker deployment
            displayName: MBBrokerService
            kind: Mb-broker-service
            name: mb-broker-services.infrastructure.asset.orchestrator.com
            version: v1
        description: Infrastructure Asset Orchestrator provides users access to cloud
          resources and services. Leverages the dynamic nature of the cloud platforms
          to provide significant cost savings and efficiencies. Using the Infrastructure
          Asset Orchestrator user can provision, configure, utilize and deprovision
          variety of services on the IBM Cloud Catalog.
        displayName: Infrastructure Asset Orchestrator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cloud-service
        - mbbroker-service
        links:
        - name: Infrastructure Asset Orchestrator
          url: https://infrastructure-asset-orchestrator.domain
        maintainers:
        - email: edge@us.ibm.com
          name: IBM-Edge
        maturity: alpha
        provider:
          name: IBM Edge
        relatedImages:
        - registry.connect.redhat.com/ibm/iao-operator@sha256:5cc395f22d86641e175339deb134af6c74de92b81470a65a5c736c8005d645fd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:609c19d1718947ee814b829334c37e1d720acf5fb642b0712b8ab1b14d57f53f
        - registry.connect.redhat.com/ibm/iao-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/iao-orchestrator-service@sha256:145928929fd0c2b7d9ed5fb374d7cc0e6b48269847e810c5472580a1538ad44c
        - registry.connect.redhat.com/ibm/iao-mb-broker-service@sha256:6531b4196c51246958027453c2d9c7dbcdb4a1f55c826d1da238d256cdbc8a10
        version: 1.0.2
      entries:
      - name: infrastructure-asset-orchestrator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: infrastructure-asset-orchestrator-certified
    provider:
      name: IBM Edge
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datasance
      provider-url: ""
    name: datasance-pot-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datasance-pot-operator.v3.4.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasance.com/v3",
                "kind": "ControlPlane",
                "metadata": {
                  "name": "pot"
                },
                "spec": {
                  "replicas": {
                    "controller": 2
                  },
                  "database": {
                    "provider": "postgres",
                    "user": "user",
                    "host": "postgre.host",
                    "port": 5432,
                    "password": "password",
                    "databaseName": "dbname"
                  },
                  "auth": {
                    "url": "auth.url",
                    "realm": "realm",
                    "realmKey": "realmKey",
                    "ssl": "none",
                    "controllerClient": "controllerClient",
                    "controllerSecret": "controllerClientSecret",
                    "viewerClient": "viewerClient"
                  },
                  "images": {
                    "pullSecret": "pullsecret",
                    "controller": "ghcr.io/datasance/controller@sha256:aa896bae6c3e6eda70938a03aaa649226356d9e4a1d03ba427d4e4ee3090ca6c",
                    "portManager": "ghcr.io/datasance/port-manager@sha256:4bf3793e9e3cd10df9b46c459e983631393825ca788ffbbb3ec3659158ad61fe",
                    "proxy": "ghcr.io/datasance/proxy@sha256:276dbed0c8ef4f5b03fdf29d5c863b6f3efcae50444ce8123220c5673272860f",
                    "router": "ghcr.io/datasance/router@sha256:82657fc8f39d33d8ef0161cf27d705cbdc0a97dca07b22a5fe78fd4213ef0d31"
                  },
                  "services": {
                    "controller": {
                      "type": "LoadBalancer",
                      "annotations": {
                        "service.beta.kubernetes.io/azure-load-balancer-internal": "false"
                      }
                    },
                    "proxy": {
                      "type": "LoadBalancer",
                      "annotations": {
                        "service.beta.kubernetes.io/azure-load-balancer-internal": "false"
                      }
                    },
                    "router": {
                      "type": "LoadBalancer",
                      "annotations": {
                        "service.beta.kubernetes.io/azure-load-balancer-internal": "false"
                      }
                    }
                  },
                  "controller": {
                    "ecnViewerUrl": "pot.demo.datasance.com",
                    "https": true,
                    "secretName": "secretName"
                  },
                  "ingresses": {
                    "controller": {
                      "annotations": {
                        "cert-manager.io/cluster-issuer": "letsencrypt"
                      },
                      "ingressClassName": "nginx",
                      "host": "pot.demo.datasance.com",
                      "secretName": "secretName"
                    },
                    "router": {
                      "address": "pot.demo.datasance.com",
                      "messagePort": 5672,
                      "interiorPort": 55672,
                      "edgePort": 45672
                    },
                    "httpProxy": {
                      "address": "pot.demo.datasance.com"
                    },
                    "tcpProxy": {
                      "address": "pot.demo.datasance.com"
                    }
                  }
                },
                "status": {
                  "conditions": [
                    {
                      "lastTransitionTime": "2022-04-10T22:44:09Z",
                      "message": "",
                      "reason": "initial_status",
                      "status": "True",
                      "type": "deploying"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          containerImage: ghcr.io/datasance/operator@sha256:143b9307c2520ef16b40c4c92e6faee8b24b7fb3a97682231e53236ddd532687
          createdAt: "2024-11-04T20:45:43Z"
          description: Datasance PoT is an Enteprise version of Eclipse ioFog Platform.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Datasance PoT Enterprise Open
            Source Subscriptions based on number of Edge device that ControlPlane
            manages"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/datasance/iofog-operator
          support: Datasance Teknoloji A.S.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Datasance PoT supported ioFog ControlPlane CRD for managing
              container workloads on Edge device
            kind: ControlPlane
            name: controlplanes.datasance.com
            version: v3
        description: ioFog Control Plane on Kubernetes
        displayName: iofog-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iofog
        - datasance
        - pot
        links:
        - name: Datasance PoT // Iofog Operator
          url: https://docs.datasance.com
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Datasance
        relatedImages:
        - ghcr.io/datasance/operator@sha256:143b9307c2520ef16b40c4c92e6faee8b24b7fb3a97682231e53236ddd532687
        - ghcr.io/datasance/controller@sha256:aa896bae6c3e6eda70938a03aaa649226356d9e4a1d03ba427d4e4ee3090ca6c
        - ghcr.io/datasance/router@sha256:82657fc8f39d33d8ef0161cf27d705cbdc0a97dca07b22a5fe78fd4213ef0d31
        - ghcr.io/datasance/proxy@sha256:276dbed0c8ef4f5b03fdf29d5c863b6f3efcae50444ce8123220c5673272860f
        - ghcr.io/datasance/port-manager@sha256:4bf3793e9e3cd10df9b46c459e983631393825ca788ffbbb3ec3659158ad61fe
        version: 3.4.9
      entries:
      - name: datasance-pot-operator.v3.4.9
        version: 3.4.9
      name: stable
    defaultChannel: stable
    packageName: datasance-pot-operator
    provider:
      name: Datasance
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sagemaker-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sagemaker-controller.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "App",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "DataQualityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Domain",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Endpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "EndpointConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "FeatureGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "HyperParameterTuningJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Model",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelBiasJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelExplainabilityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelPackage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelPackageGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelQualityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "MonitoringSchedule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "NotebookInstance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "NotebookInstanceLifecycleConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ProcessingJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "TrainingJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "TransformJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "UserProfile",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sagemaker-controller:1.4.1
          createdAt: "2025-09-22T19:07:36Z"
          description: ACK service controller for Amazon SageMaker manages SageMaker
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: App represents the state of an AWS sagemaker App resource.
            displayName: App
            kind: App
            name: apps.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: DataQualityJobDefinition represents the state of an AWS sagemaker
              DataQualityJobDefinition resource.
            displayName: DataQualityJobDefinition
            kind: DataQualityJobDefinition
            name: dataqualityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Domain represents the state of an AWS sagemaker Domain resource.
            displayName: Domain
            kind: Domain
            name: domains.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: EndpointConfig represents the state of an AWS sagemaker EndpointConfig
              resource.
            displayName: EndpointConfig
            kind: EndpointConfig
            name: endpointconfigs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Endpoint represents the state of an AWS sagemaker Endpoint
              resource.
            displayName: Endpoint
            kind: Endpoint
            name: endpoints.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: FeatureGroup represents the state of an AWS sagemaker FeatureGroup
              resource.
            displayName: FeatureGroup
            kind: FeatureGroup
            name: featuregroups.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: HyperParameterTuningJob represents the state of an AWS sagemaker
              HyperParameterTuningJob resource.
            displayName: HyperParameterTuningJob
            kind: HyperParameterTuningJob
            name: hyperparametertuningjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: InferenceComponent represents the state of an AWS sagemaker
              InferenceComponent resource.
            displayName: InferenceComponent
            kind: InferenceComponent
            name: inferencecomponents.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: LabelingJob represents the state of an AWS sagemaker LabelingJob
              resource.
            displayName: LabelingJob
            kind: LabelingJob
            name: labelingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelBiasJobDefinition represents the state of an AWS sagemaker
              ModelBiasJobDefinition resource.
            displayName: ModelBiasJobDefinition
            kind: ModelBiasJobDefinition
            name: modelbiasjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelExplainabilityJobDefinition represents the state of
              an AWS sagemaker ModelExplainabilityJobDefinition resource.
            displayName: ModelExplainabilityJobDefinition
            kind: ModelExplainabilityJobDefinition
            name: modelexplainabilityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelPackageGroup represents the state of an AWS sagemaker
              ModelPackageGroup resource.
            displayName: ModelPackageGroup
            kind: ModelPackageGroup
            name: modelpackagegroups.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelPackage represents the state of an AWS sagemaker ModelPackage
              resource.
            displayName: ModelPackage
            kind: ModelPackage
            name: modelpackages.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelQualityJobDefinition represents the state of an AWS
              sagemaker ModelQualityJobDefinition resource.
            displayName: ModelQualityJobDefinition
            kind: ModelQualityJobDefinition
            name: modelqualityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Model represents the state of an AWS sagemaker Model resource.
            displayName: Model
            kind: Model
            name: models.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: MonitoringSchedule represents the state of an AWS sagemaker
              MonitoringSchedule resource.
            displayName: MonitoringSchedule
            kind: MonitoringSchedule
            name: monitoringschedules.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: NotebookInstanceLifecycleConfig represents the state of an
              AWS sagemaker NotebookInstanceLifecycleConfig resource.
            displayName: NotebookInstanceLifecycleConfig
            kind: NotebookInstanceLifecycleConfig
            name: notebookinstancelifecycleconfigs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: NotebookInstance represents the state of an AWS sagemaker
              NotebookInstance resource.
            displayName: NotebookInstance
            kind: NotebookInstance
            name: notebookinstances.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: PipelineExecution represents the state of an AWS sagemaker
              PipelineExecution resource.
            displayName: PipelineExecution
            kind: PipelineExecution
            name: pipelineexecutions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Pipeline represents the state of an AWS sagemaker Pipeline
              resource.
            displayName: Pipeline
            kind: Pipeline
            name: pipelines.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ProcessingJob represents the state of an AWS sagemaker ProcessingJob
              resource.
            displayName: ProcessingJob
            kind: ProcessingJob
            name: processingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: TrainingJob represents the state of an AWS sagemaker TrainingJob
              resource.
            displayName: TrainingJob
            kind: TrainingJob
            name: trainingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: TransformJob represents the state of an AWS sagemaker TransformJob
              resource.
            displayName: TransformJob
            kind: TransformJob
            name: transformjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: UserProfile represents the state of an AWS sagemaker UserProfile
              resource.
            displayName: UserProfile
            kind: UserProfile
            name: userprofiles.sagemaker.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon SageMaker resources in AWS from within your Kubernetes cluster.

          **About Amazon SageMaker**

          Amazon SageMaker helps data scientists and developers to prepare, build, train, and deploy high-quality machine learning (ML) models quickly by bringing together a broad set of capabilities purpose-built for ML.

          For more information on Amazon SageMaker, visit the [product page](https://aws.amazon.com/sagemaker/).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SageMaker
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sagemaker
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SageMaker Developer Resources
          url: https://aws.amazon.com/sagemaker/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sagemaker maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sagemaker-controller:1.4.1
        version: 1.4.1
      entries:
      - name: ack-sagemaker-controller.v1.4.1
        version: 1.4.1
      - name: ack-sagemaker-controller.v1.4.0
        version: 1.4.0
      - name: ack-sagemaker-controller.v1.3.3
        version: 1.3.3
      - name: ack-sagemaker-controller.v1.3.2
        version: 1.3.2
      - name: ack-sagemaker-controller.v1.3.1
        version: 1.3.1
      - name: ack-sagemaker-controller.v1.3.0
        version: 1.3.0
      - name: ack-sagemaker-controller.v1.2.20
        version: 1.2.20
      - name: ack-sagemaker-controller.v1.2.19
        version: 1.2.19
      - name: ack-sagemaker-controller.v1.2.18
        version: 1.2.18
      - name: ack-sagemaker-controller.v1.2.17
        version: 1.2.17
      - name: ack-sagemaker-controller.v1.2.16
        version: 1.2.16
      - name: ack-sagemaker-controller.v1.2.13
        version: 1.2.13
      - name: ack-sagemaker-controller.v1.2.12
        version: 1.2.12
      - name: ack-sagemaker-controller.v1.2.11
        version: 1.2.11
      - name: ack-sagemaker-controller.v1.2.9
        version: 1.2.9
      - name: ack-sagemaker-controller.v1.2.8
        version: 1.2.8
      - name: ack-sagemaker-controller.v1.2.5
        version: 1.2.5
      - name: ack-sagemaker-controller.v1.2.4
        version: 1.2.4
      - name: ack-sagemaker-controller.v1.2.3
        version: 1.2.3
      - name: ack-sagemaker-controller.v1.2.2
        version: 1.2.2
      - name: ack-sagemaker-controller.v1.2.1
        version: 1.2.1
      - name: ack-sagemaker-controller.v1.2.0
        version: 1.2.0
      - name: ack-sagemaker-controller.v1.1.0
        version: 1.1.0
      - name: ack-sagemaker-controller.v1.0.0
        version: 1.0.0
      - name: ack-sagemaker-controller.v0.6.0
        version: 0.6.0
      - name: ack-sagemaker-controller.v0.5.1
        version: 0.5.1
      - name: ack-sagemaker-controller.v0.5.0
        version: 0.5.0
      - name: ack-sagemaker-controller.v0.4.5
        version: 0.4.5
      - name: ack-sagemaker-controller.v0.4.4
        version: 0.4.4
      - name: ack-sagemaker-controller.v0.4.3
        version: 0.4.3
      - name: ack-sagemaker-controller.v0.4.2
        version: 0.4.2
      - name: ack-sagemaker-controller.v0.4.0
        version: 0.4.0
      - name: ack-sagemaker-controller.v0.3.4
        version: 0.3.4
      - name: ack-sagemaker-controller.v0.3.3
        version: 0.3.3
      - name: ack-sagemaker-controller.v0.3.2
        version: 0.3.2
      - name: ack-sagemaker-controller.v0.3.1
        version: 0.3.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-sagemaker-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: networkservicemesh.io
      provider-url: networkservicemesh.io
    name: nsm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nsm-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nsm.networkservicemesh.io/v1alpha1",
                "kind": "NSM",
                "metadata": {
                  "name": "nsm-sample",
                  "namespace": "nsm"
                },
                "spec": {
                  "forwarders": [
                    {
                      "image": "ghcr.io/networkservicemesh/cmd-forwarder-vpp:v1.8.0",
                      "type": "vpp"
                    }
                  ],
                  "nsmLogLevel": "TRACE",
                  "nsmPullPolicy": "IfNotPresent",
                  "registry": {
                    "image": "ghcr.io/networkservicemesh/cmd-registry-memory:v1.8.0",
                    "type": "memory"
                  },
                  "version": "v1.8.0",
                  "webhook": {
                    "image": "ghcr.io/networkservicemesh/admission-webhook-k8s:v1.8.0"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: quay.io/acmenezes/nsm-operator:v1.8.0
          createdAt: "2023-03-02T19:36:46Z"
          operatorframework.io/suggested-namespace: nsm
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/networkservicemesh/nsm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkServiceEndpoint
            name: networkserviceendpoints.networkservicemesh.io
            version: v1
          - kind: NetworkService
            name: networkservices.networkservicemesh.io
            version: v1
          - kind: NSM
            name: nsms.nsm.networkservicemesh.io
            version: v1alpha1
        description: |-
          The Network Service Mesh Operator manages and install network service meshes.

          Check networkservicemesh.io

          Install requirements:

          Before installing the operator it's necessary to have installed [spire](https://github.com/spiffe/spire)

          We have a ready to install yaml file for spire [here](https://github.com/networkservicemesh/nsm-operator/blob/master/config/spire/spire.yaml)
        displayName: nsm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network
        - service mesh
        links:
        - name: Nsm Operator
          url: https://nsm-operator.domain
        maintainers:
        - email: adcmenezes@gmail.com
          name: Alexandre Menezes
        maturity: alpha
        provider:
          name: networkservicemesh.io
          url: networkservicemesh.io
        relatedImages:
        - quay.io/acmenezes/nsm-operator:v1.8.0
        version: 1.8.0
      entries:
      - name: nsm-operator.v1.8.0
        version: 1.8.0
      name: alpha
    defaultChannel: alpha
    packageName: nsm-operator
    provider:
      name: networkservicemesh.io
      url: networkservicemesh.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: observability-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      name: candidate
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      - name: observability-operator.v0.0.22
        version: 0.0.22
      name: development
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      name: stable
    defaultChannel: stable
    packageName: observability-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-atlas-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-atlas-kubernetes.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "id": "1",
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "WEEKLY",
                      "id": "2",
                      "retentionUnit": "DAYS",
                      "retentionValue": 6
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": true,
                  "policies": [
                    {
                      "name": "atlas-default-backuppolicy",
                      "namespace": "mongodb-atlas-system"
                    },
                    {
                      "name": "atlas-default-backuppolicy2",
                      "namespace": "mongodb-atlas-system"
                    }
                  ],
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "my-database-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "david"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-cluster"
                },
                "spec": {
                  "deploymentSpec": {
                    "name": "test-cluster",
                    "providerSettings": {
                      "instanceSizeName": "M10",
                      "providerName": "AWS",
                      "regionName": "US_EAST_1"
                    }
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          containerImage: registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator@sha256:3cfbd700e4771c2c7e75c2c41cf6de57c9ef71ecd286e96a21d2567e43424bc3
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasclusters API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB Atlas —
          the only multi-cloud document database service that gives you the versatility you need to build sophisticated and resilient applications that can adapt to changing customer demands and market trends.

          > Current Status: *Beta*. The Operator gives users the ability to provision
          > Atlas projects, clusters and database users using Kubernetes Specifications and bind connection information
          > into applications deployed to Kubernetes.

          ## Quick Start guide
          ### Step 1. Deploy Kubernetes operator by clicking Install button.

          ### Step 2. Create Atlas Cluster

          **1.** Create an Atlas API Key Secret
          In order to work with the Atlas Operator you need to provide [authentication information](https://docs.atlas.mongodb.com/configure-api-access)
           to allow the Atlas Operator to communicate with Atlas API. Once you have generated a Public and Private key in Atlas, you can create a Kuberentes Secret with:
          ```
          kubectl create secret generic mongodb-atlas-operator-api-key \
                   --from-literal="orgId=<the_atlas_organization_id>" \
                   --from-literal="publicApiKey=<the_atlas_api_public_key>" \
                   --from-literal="privateApiKey=<the_atlas_api_private_key>" \
                   -n openshift-operators
          ```
          (Note, that you should use the namespace where the Operator was installed - it's `openshift-operators` by default)

          **2.** Create an `AtlasProject` Custom Resource

          The `AtlasProject` CustomResource represents Atlas Projects in our Kubernetes cluster. You need to specify
          `projectIpAccessList` with the IP addresses or CIDR blocks of any hosts that will connect to the Atlas Cluster.
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasProject
          metadata:
            name: my-project
          spec:
            name: Test Atlas Operator Project
            projectIpAccessList:
              - ipAddress: "192.0.2.15"
                comment: "IP address for Application Server A"
              - ipAddress: "203.0.113.0/24"
                comment: "CIDR block for Application Server B - D"
          ```
          **3.** Create an `AtlasDeployment` Custom Resource.
          The example below is a minimal configuration to create an M10 Atlas cluster in the AWS US East region. For a full list of properties, check
          `atlasclusters.atlas.mongodb.com` [CRD specification](config/crd/bases/atlas.mongodb.com_atlasclusters.yaml)):
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasDeployment
          metadata:
            name: my-atlas-cluster
          spec:
            name: "Test-cluster"
            projectRef:
              name: my-project
            providerSettings:
              instanceSizeName: M10
              providerName: AWS
              regionName: US_EAST_1
          ```

          **4.** Create a database user password Kubernetes Secret
          The Secret must be created in the same namespace as the `AtlasDeployment` and `AtlasProject` were created.
          ```
          kubectl create secret generic the-user-password --from-literal="password=P@@sword%"
          ```

          **5.** Create an `AtlasDatabaseUser` Custom Resource

          In order to connect to an Atlas Cluster the database user needs to be created. `AtlasDatabaseUser` resource should reference
          the password Kubernetes Secret created in the previous step.
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasDatabaseUser
          metadata:
            name: my-database-user
          spec:
            roles:
              - roleName: "readWriteAnyDatabase"
                databaseName: "admin"
            projectRef:
              name: my-project
            username: theuser
            passwordSecretRef:
              name: the-user-password
          ```
          **6.** Wait for the `AtlasDatabaseUser` Custom Resource to be ready

          Wait until the AtlasDatabaseUser resource gets to "ready" status (it will wait until the cluster is created that may take around 10 minutes):
          ```
          kubectl get atlasdatabaseusers my-database-user -o=jsonpath='{.status.conditions[?(@.type=="Ready")].status}'
          True
          ```
          ### Step 3. Connect your application to the Atlas Cluster

          The Atlas Operator will create a Kubernetes Secret with the information necessary to connect to the Atlas Cluster created
          in the previous step. An application in the same Kubernetes Cluster can mount and use the Secret:

          ```
          ...
          containers:
                - name: test-app
                  env:
                   - name: "CONNECTION_STRING"
                     valueFrom:
                       secretKeyRef:
                         name: test-atlas-operator-project-test-cluster-theuser
                         key: connectionStringStandardSrv

          ```
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator@sha256:3cfbd700e4771c2c7e75c2c41cf6de57c9ef71ecd286e96a21d2567e43424bc3
        version: 1.0.0
      entries:
      - name: mongodb-atlas-kubernetes.v1.0.0
        version: 1.0.0
      name: beta
    - currentCSV: mongodb-atlas-kubernetes.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupCompliancePolicy",
                "metadata": {
                  "name": "my-backup-compliance-policy"
                },
                "spec": {
                  "authorizedEmail": "john.doe@example.com",
                  "authorizedUserFirstName": "John",
                  "authorizedUserLastName": "Doe",
                  "copyProtectionEnabled": false,
                  "encryptionAtRestEnabled": false,
                  "onDemandPolicy": {
                    "retentionUnit": "weeks",
                    "retentionValue": 3
                  },
                  "overwriteBackupPolicies": false,
                  "pointInTimeEnabled": true,
                  "restoreWindowDays": 42,
                  "scheduledPolicyItems": [
                    {
                      "frequencyInterval": 2,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "hourly",
                      "retentionUnit": "days",
                      "retentionValue": 2
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "weekly",
                      "retentionUnit": "weeks",
                      "retentionValue": 4
                    },
                    {
                      "frequencyInterval": 15,
                      "frequencyType": "monthly",
                      "retentionUnit": "months",
                      "retentionValue": 12
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": false,
                  "copySettings": [
                    {
                      "cloudProvider": "AWS",
                      "frequencies": [
                        "WEEKLY"
                      ],
                      "regionName": "US_WEST_1",
                      "shouldCopyOplogs": false
                    }
                  ],
                  "policy": {
                    "name": "atlasbackuppolicy-sample"
                  },
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2,
                  "useOrgAndGroupNamesInExportPrefix": true
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasCustomRole",
                "metadata": {
                  "name": "shard-operator-role"
                },
                "spec": {
                  "projectRef": {
                    "name": "my-project"
                  },
                  "role": {
                    "actions": [
                      {
                        "name": "getShardMap",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "shardingState",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "connPoolStats",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "getLog",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      }
                    ],
                    "inheritedRoles": [
                      {
                        "database": "admin",
                        "name": "operator-role-1"
                      }
                    ],
                    "name": "my-role"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDataFederation",
                "metadata": {
                  "name": "my-data-federation",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "name": "my-data-federation",
                  "privateEndpoints": [
                    {
                      "endpointId": "vpce-03f9eeaa764e32454",
                      "provider": "AWS",
                      "type": "DATA_LAKE"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project",
                    "namespace": "mongodb-atlas-system"
                  },
                  "storage": {
                    "databases": [
                      {
                        "collections": [
                          {
                            "dataSources": [
                              {
                                "storeName": "http-test",
                                "urls": [
                                  "https://data.cityofnewyork.us/api/views/vfnx-vebw/rows.csv"
                                ]
                              }
                            ],
                            "name": "test-collection-1"
                          }
                        ],
                        "name": "test-db-1"
                      }
                    ],
                    "stores": [
                      {
                        "name": "http-test",
                        "provider": "http"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "the-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "theuser"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-deployment"
                },
                "spec": {
                  "backupRef": {
                    "name": "atlasbackupschedule-sample"
                  },
                  "deploymentSpec": {
                    "backupEnabled": true,
                    "clusterType": "REPLICASET",
                    "mongoDBMajorVersion": "7.0",
                    "name": "test-deployment",
                    "replicationSpecs": [
                      {
                        "regionConfigs": [
                          {
                            "electableSpecs": {
                              "instanceSize": "M10",
                              "nodeCount": 3
                            },
                            "priority": 7,
                            "providerName": "AWS",
                            "regionName": "US_EAST_1"
                          }
                        ],
                        "zoneName": "Zone 1"
                      }
                    ],
                    "tags": [
                      {
                        "key": "team",
                        "value": "atlas-operator"
                      }
                    ]
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasFederatedAuth",
                "metadata": {
                  "name": "atlas-default-federated-auth",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "connectionSecretRef": {
                    "name": "my-org-secret",
                    "namespace": "mongodb-atlas-system"
                  },
                  "domainAllowList": [
                    "my-org-domain.com"
                  ],
                  "domainRestrictionEnabled": true,
                  "enabled": true,
                  "postAuthRoleGrants": [
                    "ORG_MEMBER"
                  ],
                  "roleMappings": [
                    {
                      "externalGroupName": "org-admin",
                      "roleAssignments": [
                        {
                          "role": "ORG_OWNER"
                        }
                      ]
                    },
                    {
                      "externalGroupName": "dev-team",
                      "roleAssignments": [
                        {
                          "role": "ORG_GROUP_CREATOR"
                        },
                        {
                          "projectName": "dev-project",
                          "role": "GROUP_OWNER"
                        }
                      ]
                    }
                  ],
                  "ssoDebugEnabled": false
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasIPAccessList",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasipaccesslist-sample"
                },
                "spec": {
                  "entries": [
                    {
                      "cidrBlock": "192.168.1.0/24"
                    },
                    {
                      "deleteAfterDate": "2025-01-20T23:59:59+02:00",
                      "ipAddress": "10.0.0.1"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkContainer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasnetworkcontainer-sample"
                },
                "spec": {
                  "cidrBlock": "10.11.0.0/16",
                  "connectionSecret": {
                    "name": "atlas-connection-secret"
                  },
                  "externalProjectRef": {
                    "projectId": "66e2f2b621571b7e69a89b66"
                  },
                  "provider": "AWS",
                  "region": "US_EAST_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkPeering",
                "metadata": {
                  "name": "atlasnetworkpeering-sample"
                },
                "spec": {
                  "atlasCidrBlock": "10.12.0.0/16",
                  "awsConfiguration": {
                    "accepterRegionName": "US_EAST_1",
                    "awsAccountId": "23214235145",
                    "routeTableCidrBlock": "10.11.0.0/16",
                    "vpcId": "vpcid-7123782134"
                  },
                  "containerId": "623412394512350",
                  "containerRegion": "US_EAST_1",
                  "projectRef": {
                    "name": "atlas-project",
                    "namespace": "namespace"
                  },
                  "provider": "AWS"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasPrivateEndpoint",
                "metadata": {
                  "name": "atlasprivateendpoint-sample"
                },
                "spec": {
                  "awsConfiguration": {
                    "id": "vpce-f4k34w51d"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "provider": "AWS",
                  "region": "EU_CENTRAL_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    },
                    {
                      "cidrBlock": "203.0.113.0/24",
                      "comment": "CIDR block for Application Server B - D"
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasSearchIndexConfig",
                "metadata": {
                  "name": "atlassearchindexconfig-sample"
                },
                "spec": {
                  "analyzer": "lucene.standard",
                  "analyzers": [
                    {
                      "name": "test-analyzer",
                      "tokenFilters": [
                        {
                          "asciiFolding": {
                            "originalTokens": "omit"
                          },
                          "type": "asciiFolding"
                        }
                      ],
                      "tokenizer": {
                        "type": "whiteSpace",
                        "whitespace": {
                          "maxTokenLenght": 200
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamConnection",
                "metadata": {
                  "name": "my-streamconnection-kafka"
                },
                "spec": {
                  "kafkaConfig": {
                    "authentication": {
                      "credentials": {
                        "name": "kafka-userpass"
                      },
                      "mechanism": "SCRAM-512"
                    },
                    "bootstrapServers": "kafka.server1:9001,kafka.server2:9002,kafka.server3:9003",
                    "security": {
                      "certificate": {
                        "name": "kafka-cert"
                      },
                      "protocol": "SSL"
                    }
                  },
                  "name": "kafka-config",
                  "type": "Kafka"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamInstance",
                "metadata": {
                  "name": "my-streaminstance-sample"
                },
                "spec": {
                  "clusterConfig": {
                    "provider": "AWS",
                    "region": "VIRGINIA_USA",
                    "tier": "SP30"
                  },
                  "connectionRegistry": [
                    {
                      "name": "my-streamconnection-sample"
                    }
                  ],
                  "name": "my-instance",
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasTeam",
                "metadata": {
                  "name": "atlasteam-sample"
                },
                "spec": {
                  "name": "my-team",
                  "usernames": [
                    "username@mongodb.com"
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasThirdPartyIntegration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasintegration-sample"
                },
                "spec": {
                  "datadog": {
                    "apiKeySecretRef": {
                      "name": "secret-ref"
                    },
                    "region": "string",
                    "sendCollectionLatencyMetrics": "string",
                    "sendDatabaseMetrics": "string"
                  },
                  "projectRef": {
                    "name": "my-atlas-project"
                  },
                  "type": "DATADOG"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:8ac3f6b1c0f31af5f32562ddeed37dd6586ac47e1e7e111947df3ed570c5bed9
          createdAt: "2025-09-25T14:34:34Z"
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/mongodb/mongodb-atlas-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The AtlasBackupCompliancePolicy is a configuration that enforces
              specific backup and retention requirements
            displayName: Atlas Backup Compliance Policy
            kind: AtlasBackupCompliancePolicy
            name: atlasbackupcompliancepolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasCustomRole is the Schema for the AtlasCustomRole API
            displayName: Atlas Custom Role
            kind: AtlasCustomRole
            name: atlascustomroles.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDataFederation is the Schema for the Atlas Data Federation
              API
            displayName: Atlas Data Federation
            kind: AtlasDataFederation
            name: atlasdatafederations.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasdeployments API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasFederatedAuth is the Schema for the Atlasfederatedauth
              API
            displayName: Atlas Federated Auth
            kind: AtlasFederatedAuth
            name: atlasfederatedauths.atlas.mongodb.com
            version: v1
          - description: AtlasIPAccessList is the Schema for the atlasipaccesslists
              API.
            displayName: Atlas IPAccess List
            kind: AtlasIPAccessList
            name: atlasipaccesslists.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkContainer is the Schema for the AtlasNetworkContainer
              API
            displayName: Atlas Network Container
            kind: AtlasNetworkContainer
            name: atlasnetworkcontainers.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkPeering is the Schema for the AtlasNetworkPeering
              API
            displayName: Atlas Network Peering
            kind: AtlasNetworkPeering
            name: atlasnetworkpeerings.atlas.mongodb.com
            version: v1
          - displayName: Atlas Org Settings
            kind: AtlasOrgSettings
            name: atlasorgsettings.atlas.mongodb.com
            version: v1
          - description: "The AtlasPrivateEndpoint custom resource definition (CRD)
              defines a desired [Private Endpoint](https://www.mongodb.com/docs/atlas/security-private-endpoint/#std-label-private-endpoint-overview)
              configuration for an Atlas project. It allows a private connection between
              your cloud provider and Atlas that doesn't send information through
              a public network. \n You can use private endpoints to create a unidirectional
              connection to Atlas clusters from your virtual network."
            displayName: Atlas Private Endpoint
            kind: AtlasPrivateEndpoint
            name: atlasprivateendpoints.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
          - description: Atlas Search Index Config is the Schema for the Atlas Search
              API
            displayName: Atlas Search Index Config
            kind: AtlasSearchIndexConfig
            name: atlassearchindexconfigs.atlas.mongodb.com
            version: v1
          - description: AtlasStreamConnection is the Schema for the atlasstreamconnections
              API
            displayName: Atlas Stream Connection
            kind: AtlasStreamConnection
            name: atlasstreamconnections.atlas.mongodb.com
            version: v1
          - description: AtlasStreamInstance is the Schema for the atlasstreaminstances
              API
            displayName: Atlas Stream Instance
            kind: AtlasStreamInstance
            name: atlasstreaminstances.atlas.mongodb.com
            version: v1
          - description: AtlasTeam is the Schema for the Atlas Teams API
            displayName: Atlas Team
            kind: AtlasTeam
            name: atlasteams.atlas.mongodb.com
            version: v1
          - description: AtlasThirdPartyIntegration is the Schema for the atlas 3rd
              party integrations API.
            displayName: Atlas Third Party Integration
            kind: AtlasThirdPartyIntegration
            name: atlasthirdpartyintegrations.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB
          Atlas — the only multi-cloud document database service that gives you the versatility you need to build sophisticated
          and resilient applications that can adapt to changing customer demands and market trends.

          The full documentation for the Operator can be found [here](https://docs.atlas.mongodb.com/atlas-operator/)

          ## Getting Started

          ### Supported features

          * Create and configure a project, or connect to an existing one.
          * Deploy, manage, scale, and tear down clusters.
          * Support for serverless instances.
          * Create and edit database users.
          * Manage IP Access Lists, network peering and private endpoints.
          * Configure and control Atlas’s fully managed cloud backup.
          * Configure federated authentication for your Atlas organization
          * Integrate Atlas monitoring with Prometheus.

          ... and more.

          To view the list of custom resources and each CR's schema, visit our [reference](https://www.mongodb.com/docs/atlas/operator/stable/custom-resources/)
          documentation. See the [Quickstart](https://www.mongodb.com/docs/atlas/operator/stable/ak8so-quick-start/) to get
          started with Atlas Kubernetes Operator.
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:8ac3f6b1c0f31af5f32562ddeed37dd6586ac47e1e7e111947df3ed570c5bed9
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:9eb6465e723ede43ac47eed6de22042d26c0efb276301a4424a6a3948a8f5b15
        version: 2.11.0
      entries:
      - name: mongodb-atlas-kubernetes.v2.11.0
        version: 2.11.0
      - name: mongodb-atlas-kubernetes.v2.10.0
        version: 2.10.0
      - name: mongodb-atlas-kubernetes.v2.9.1
        version: 2.9.1
      - name: mongodb-atlas-kubernetes.v2.9.0
        version: 2.9.0
      - name: mongodb-atlas-kubernetes.v2.8.2
        version: 2.8.2
      - name: mongodb-atlas-kubernetes.v2.8.1
        version: 2.8.1
      - name: mongodb-atlas-kubernetes.v2.8.0
        version: 2.8.0
      - name: mongodb-atlas-kubernetes.v2.7.1
        version: 2.7.1
      - name: mongodb-atlas-kubernetes.v2.7.0
        version: 2.7.0
      - name: mongodb-atlas-kubernetes.v2.6.1
        version: 2.6.1
      - name: mongodb-atlas-kubernetes.v2.6.0
        version: 2.6.0
      - name: mongodb-atlas-kubernetes.v2.5.0
        version: 2.5.0
      - name: mongodb-atlas-kubernetes.v2.4.1
        version: 2.4.1
      - name: mongodb-atlas-kubernetes.v2.4.0
        version: 2.4.0
      - name: mongodb-atlas-kubernetes.v2.3.1
        version: 2.3.1
      - name: mongodb-atlas-kubernetes.v2.3.0
        version: 2.3.0
      - name: mongodb-atlas-kubernetes.v2.2.2
        version: 2.2.2
      - name: mongodb-atlas-kubernetes.v2.2.1
        version: 2.2.1
      - name: mongodb-atlas-kubernetes.v2.2.0
        version: 2.2.0
      - name: mongodb-atlas-kubernetes.v2.1.0
        version: 2.1.0
      - name: mongodb-atlas-kubernetes.v2.0.1
        version: 2.0.1
      - name: mongodb-atlas-kubernetes.v2.0.0
        version: 2.0.0
      - name: mongodb-atlas-kubernetes.v1.9.1
        version: 1.9.1
      - name: mongodb-atlas-kubernetes.v1.9.0
        version: 1.9.0
      - name: mongodb-atlas-kubernetes.v1.8.2
        version: 1.8.2
      - name: mongodb-atlas-kubernetes.v1.8.1
        version: 1.8.1
      - name: mongodb-atlas-kubernetes.v1.8.0
        version: 1.8.0
      - name: mongodb-atlas-kubernetes.v1.7.3
        version: 1.7.3
      - name: mongodb-atlas-kubernetes.v1.7.2
        version: 1.7.2
      - name: mongodb-atlas-kubernetes.v1.7.1
        version: 1.7.1
      - name: mongodb-atlas-kubernetes.v1.7.0
        version: 1.7.0
      - name: mongodb-atlas-kubernetes.v1.6.1
        version: 1.6.1
      - name: mongodb-atlas-kubernetes.v1.6.0
        version: 1.6.0
      - name: mongodb-atlas-kubernetes.v1.5.1
        version: 1.5.1
      - name: mongodb-atlas-kubernetes.v1.5.0
        version: 1.5.0
      - name: mongodb-atlas-kubernetes.v1.4.1
        version: 1.4.1
      - name: mongodb-atlas-kubernetes.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-atlas-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco AppDynamics
      provider-url: https://www.appdynamics.com
    name: appdynamics-cloud-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appdynamics-cloud-operator.v24.7.0-133
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clustermon",
                "metadata": {
                  "name": "clustermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "clusterName": "test",
                  "image": "appdynamics/appdynamics-cloud-k8s-monitoring:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Containermon",
                "metadata": {
                  "name": "containermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "name": "ContainerMon"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "DbCollector",
                "metadata": {
                  "name": "dbcollector",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-db-collector:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "DbConfig",
                "metadata": {
                  "name": "dbconfig",
                  "namespace": "appdynamics"
                },
                "spec": null
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Inframon",
                "metadata": {
                  "name": "inframon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-k8s-monitoring:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "LogCollector",
                "metadata": {
                  "name": "logcollector-sample",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-log-collector-agent:latest",
                  "resources": {
                    "limits": {
                      "cpu": "400m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "150Mi"
                    }
                  }
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Servermon",
                "metadata": {
                  "name": "servermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "name": "ServerMon"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-08-06T08:33:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clustermon is the Schema for the clustermons API
            displayName: Clustermon
            kind: Clustermon
            name: clustermons.cluster.appdynamics.com
            version: v1alpha1
          - description: Containermon is the Schema for the containermons API
            displayName: Containermon
            kind: Containermon
            name: containermons.cluster.appdynamics.com
            version: v1alpha1
          - kind: DbCollector
            name: dbcollectors.cluster.appdynamics.com
            version: v1alpha1
          - kind: DbConfig
            name: dbconfigs.cluster.appdynamics.com
            version: v1alpha1
          - description: Inframon is the Schema for the inframons API
            displayName: Inframon
            kind: Inframon
            name: inframons.cluster.appdynamics.com
            version: v1alpha1
          - kind: LogCollector
            name: logcollectors.cluster.appdynamics.com
            version: v1alpha1
          - description: Servermon is the Schema for the servermons API
            displayName: Servermon
            kind: Servermon
            name: servermons.cluster.appdynamics.com
            version: v1alpha1
        description: The Appdynamics Cloud Operator is used to install Cluster Collector,
          Container Collector, Server Collector, Log Collector and Cluster Agent
        displayName: Cisco AppDynamics Cloud Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - appdynamics
        links:
        - name: Appdynamics Cloud Operator
          url: https://appdynamics-cloud-operator.domain
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: Cisco AppDynamics
          url: https://www.appdynamics.com
        relatedImages:
        - registry.connect.redhat.com/appdynamics/appdynamics-cloud-operator@sha256:2710d63425942f60af21f060a02eb3c40f5a5f0c952ea6593b3b9769c7ca1014
        version: 24.7.0-133
      entries:
      - name: appdynamics-cloud-operator.v24.7.0-133
        version: 24.7.0-133
      name: alpha
    defaultChannel: alpha
    packageName: appdynamics-cloud-operator
    provider:
      name: Cisco AppDynamics
      url: https://www.appdynamics.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: smb-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smb-csi-driver-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9-operator@sha256:1019812a3969f844a0f3fcd45091fccc7d0d816207e230a2f399556c41ebc705
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure CIFS/SMB CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.15.0-0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/csi-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          CIFS/SMB CSI Driver Operator provides Server Message Block (SMB) CSI Driver that enables you to create and mount CIFS/SMB PersistentVolumes.
        displayName: CIFS/SMB CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - smb
        - cifs
        links:
        - name: Documentation
          url: https://github.com/openshift/csi-operator
        - name: Source Repository
          url: https://github.com/openshift/csi-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: preview
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9-operator@sha256:1019812a3969f844a0f3fcd45091fccc7d0d816207e230a2f399556c41ebc705
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0627c09fe3f5a015f117fb30af1f4d17340aefa48b23994737d3901a804b63d3
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:a5704d26d0b562dba9633d1c107f34e05c750c0e3a64dbcace4dab7b964fff8d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9@sha256:9f516d06c279cf1f82e8948ce2290395eec5f6611abf5ea919e4d6548d81ec10
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:9c3c3d2b857ce289324aee56df1ba4ab5e5ea1e76cd86ac988233463915834b4
        version: 4.16.0-202509240123
      entries:
      - name: smb-csi-driver-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: smb-csi-driver-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: smb-csi-driver-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: smb-csi-driver-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: smb-csi-driver-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: smb-csi-driver-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: smb-csi-driver-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: smb-csi-driver-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: smb-csi-driver-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: smb-csi-driver-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: smb-csi-driver-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: smb-csi-driver-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: smb-csi-driver-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: smb-csi-driver-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: smb-csi-driver-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: smb-csi-driver-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: smb-csi-driver-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: smb-csi-driver-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: smb-csi-driver-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: smb-csi-driver-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: smb-csi-driver-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: smb-csi-driver-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: smb-csi-driver-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: smb-csi-driver-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: smb-csi-driver-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: smb-csi-driver-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: smb-csi-driver-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: smb-csi-driver-operator.v4.16.0-202406251906
        version: 4.16.0-202406251906
      - name: smb-csi-driver-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: smb-csi-driver-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: preview
    defaultChannel: preview
    packageName: smb-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: node-maintenance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-maintenance-operator.v5.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:58846ab62346b1004e11f36bd901c96f8d315958eb39fbb272d8b328faebcf04
          createdAt: "2023-10-20 15:36:51"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=4.10.0 <5.0.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-node-maintenance-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:58846ab62346b1004e11f36bd901c96f8d315958eb39fbb272d8b328faebcf04
        version: 5.0.1
      entries:
      - name: node-maintenance-operator.v5.0.1
        version: 5.0.1
      name: 4.12-eus
    - currentCSV: node-maintenance-operator.v5.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:2efd1976d2fb110391e0e5937b90476d3eeaacb4dd29cd52375f2b8566f17b41
          createdAt: "2023-10-20 15:50:49"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=5.0.0 <5.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:2efd1976d2fb110391e0e5937b90476d3eeaacb4dd29cd52375f2b8566f17b41
        version: 5.2.1
      entries:
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.14-eus
    - currentCSV: node-maintenance-operator.v5.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:7ca5853885534198544d23067087d80047b0e1421903085a2cb328bd00def147
          createdAt: "2024-08-10 19:55:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.2.0 <5.3.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:7ca5853885534198544d23067087d80047b0e1421903085a2cb328bd00def147
        version: 5.3.1
      entries:
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.16-eus
    - currentCSV: node-maintenance-operator.v5.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-10 10:59:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.3.0 <5.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.4/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel9-operator@sha256:752bff8bebcaf3df1f9d126dc4b07b622a44b02bfcf0f82b9100a7693627079a
        version: 5.4.1
      entries:
      - name: node-maintenance-operator.v5.4.1
        version: 5.4.1
      - name: node-maintenance-operator.v5.4.0
        version: 5.4.0
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.18-eus
    - currentCSV: node-maintenance-operator.v5.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-10 10:59:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.3.0 <5.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.4/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel9-operator@sha256:752bff8bebcaf3df1f9d126dc4b07b622a44b02bfcf0f82b9100a7693627079a
        version: 5.4.1
      entries:
      - name: node-maintenance-operator.v5.4.1
        version: 5.4.1
      - name: node-maintenance-operator.v5.4.0
        version: 5.4.0
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: stable
    defaultChannel: stable
    packageName: node-maintenance-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc
      provider-url: ""
    name: openshift-gitops-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-gitops-operator.v1.10.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:b0dc693854319ada341e2e3872b34d7db9823b470ed369edfde88d3ae73a40e1
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.8.19\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:a74e0a1fe3589ebe9f226a2bad20e9a5d5601a0e62c5f3290902f466f55b7160
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:2bec31b0c5407ffcecfd8b860d8afb1629eaab31a250f60191f94a824e0a52f4
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:c779ad87a7f71d851c2af818abf16e5151c43d3fd54e64eac2fc7117c00753f5
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:b0dc693854319ada341e2e3872b34d7db9823b470ed369edfde88d3ae73a40e1
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:b5c8a6c7438a45125fc7fd970946f715ab7193b7446d625fae80dde9d1f3f914
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:4c842f0ea8f0e9b53eb16d94488fc2f4298bd8e4be2f48713be9157da24549e6
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:ece2fee4227ad1becadca47028b78ceaf8717a98ec84e3f1b5188a135c4b093c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d35dff5d59e748ef0404e7e76b0237ea96b2bb19b86fa04a3a34df7a379277b2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5d5061dbc2ec5082cf14b6c600fb5400b83cf91d7ccebfa80680a238d275db
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:0905ba3f5cf87780a1e95b6a00ea7cd340d4ed1b5054e4b2c82406270baeaefa
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ec9f60018694dcc5d431ba47d5536b761b71cb3f66684978fe6bb74c157679ac
        version: 1.10.6
      entries:
      - name: openshift-gitops-operator.v1.10.6
        version: 1.10.6
      - name: openshift-gitops-operator.v1.10.5
        version: 1.10.5
      - name: openshift-gitops-operator.v1.10.4
        version: 1.10.4
      - name: openshift-gitops-operator.v1.10.3
        version: 1.10.3
      - name: openshift-gitops-operator.v1.10.2
        version: 1.10.2
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.10
    - currentCSV: openshift-gitops-operator.v1.11.7-0.1724840231.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d6ad2ee900bf99a3672049cc8ce9c94870d41176501ded1abbc1b229b9553401
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          olm.substitutesFor: openshift-gitops-operator.v1.11.7
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.9.21\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:4d22a7a6cc2f0442889512267ae3d9356f05ff0dfdbbe137c3d8ebebc24818c8
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:42f809f32f1d43186b849e3cc78063d3ebb2b54c55b2a1b77908af09a5ce044b
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:0b926fc9eae775425faa1ab9709a36fdac15afe4bfe523720e17ecf5df45bb28
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:0df37cf0a0e48efae02a0711c6da61c6165e7856db97097e98d56c8b6a137bb9
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:eade50b8321c1b2e0a8ac4dcb34596d43dea0a137b8b97b6eb68c6d469c8b200
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:20566173a8e907980e603d82fae4d96fcdfd5c43d9004c2e506cf0e09a47b91e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/rhel8/redis-6@sha256:c4a98010f7dd5faec2047bb9803061704233a8efc930be2a02386ac16d062d86
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c94c2167dc957856682a104b9a6187d3d321dec71ae794acc0dc37794d2cc373
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:7ea96000768b3cf7711150531c4c1c7edff2c45a2f29fe10bd9811ab6b281909
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d6ad2ee900bf99a3672049cc8ce9c94870d41176501ded1abbc1b229b9553401
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:959de6a95ef8978f66fdbe1033bcaa0ddeff9a376023e5d1d2698c5acb417305
        version: 1.11.7+0.1724840231.p
      entries:
      - name: openshift-gitops-operator.v1.11.7-0.1724840231.p
        version: 1.11.7+0.1724840231.p
      - name: openshift-gitops-operator.v1.11.7
        version: 1.11.7
      - name: openshift-gitops-operator.v1.11.6
        version: 1.11.6
      - name: openshift-gitops-operator.v1.11.5
        version: 1.11.5
      - name: openshift-gitops-operator.v1.11.4
        version: 1.11.4
      - name: openshift-gitops-operator.v1.11.3
        version: 1.11.3
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.11
    - currentCSV: openshift-gitops-operator.v1.12.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:48dd1254fd3a887aef2eae9239f68637eae18514a50f72ea504414d83dc727b8
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.10.18\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:dbf3038daf7c797123c511f9a609c3485cb1a5bcae53c614fad17e0a65fa7ee3
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a1d4e26ea755181818401697f18260b9e876aa441be14830f9142081ee22a24d
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:73918f3b3c507fc64006b7a500ecfa9d58151d694852f696c41a4412edffc78e
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:33bbb88343285a0510670384b87d0226dd6de8aa12c78b68921362c4d6f5a6aa
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:48dd1254fd3a887aef2eae9239f68637eae18514a50f72ea504414d83dc727b8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9e0fee30842a16d7b8a39d336123c80c5271cb2ceb50759fc7ee51a4ed09ceb3
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:06a6637c4776f4285bcb295d327ee60b1aed61b7adc5f7ab52898d79281ced70
        - registry.redhat.io/rhel8/redis-6@sha256:a848cc4edb4405846c673187dd68ce66c328dd1ad42f193f5737288c7971399d
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:b5736d1106db29fcdf42328ca20b96382d4339f90b97a07120ea73ee174ea1ad
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:7d82c719b1cef64607d30e8aaa5ab293955771c5b1dbd430e0befc988a43356d
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:1cf485a889c7586cb4944f000b53c0e97d08f97ba14e580bec57dad7cb13712e
        version: 1.12.6
      entries:
      - name: openshift-gitops-operator.v1.12.6
        version: 1.12.6
      - name: openshift-gitops-operator.v1.12.5-0.1727371523.p
        version: 1.12.5+0.1727371523.p
      - name: openshift-gitops-operator.v1.12.5
        version: 1.12.5
      - name: openshift-gitops-operator.v1.12.4
        version: 1.12.4
      - name: openshift-gitops-operator.v1.12.3
        version: 1.12.3
      - name: openshift-gitops-operator.v1.12.2
        version: 1.12.2
      - name: openshift-gitops-operator.v1.12.1
        version: 1.12.1
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.12
    - currentCSV: openshift-gitops-operator.v1.13.3-0.1741683398.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d41034bcf5c28bcb8afef08678793d0cc27a4032e2811a0c496c3bd0f10ecc17
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          olm.substitutesFor: openshift-gitops-operator.v1.13.3-0.1738142397.p
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.11.12\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_gitops/
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d41034bcf5c28bcb8afef08678793d0cc27a4032e2811a0c496c3bd0f10ecc17
        - registry.redhat.io/rhel8/redis-6@sha256:a848cc4edb4405846c673187dd68ce66c328dd1ad42f193f5737288c7971399d
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:b982df02ea7c45a7603d83fe82cac12e1faac303a1b72af6a727512bb97441df
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3f2804bc2af437ff1f06c214b48a1a780dfcfab995650d901c03270057eb9f3c
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:06a6637c4776f4285bcb295d327ee60b1aed61b7adc5f7ab52898d79281ced70
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a7c3e31ce414e6c6fecf1652343b22c52502902676bf4d7a75bf42bdabfca06
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:a366229e6ee26ebe2d7f765b2a2da6057a29e783835655b5d97545ac563b541a
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:d4d5724f67859d24c32f0f5735f5af3e325f77db97f58bc7f17ed4b831b7e534
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c02dc2c79ee24ce5450ea4e19493ac2a1997eeb5bfd359291642c529826d0cdd
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:e87515d016565bb69cb2c2ac40432e0127dd2cb20a490ac82019836032ba15e5
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:bde11fa83b7e708677b08ea995041f480b88edfb93e4a62c10380d101c0a349c
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        version: 1.13.3+0.1741683398.p
      entries:
      - name: openshift-gitops-operator.v1.13.3-0.1741683398.p
        version: 1.13.3+0.1741683398.p
      - name: openshift-gitops-operator.v1.13.3-0.1738142397.p
        version: 1.13.3+0.1738142397.p
      - name: openshift-gitops-operator.v1.13.3
        version: 1.13.3
      - name: openshift-gitops-operator.v1.13.2
        version: 1.13.2
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.13
    - currentCSV: openshift-gitops-operator.v1.14.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:7cf924af77d9c56e70710b99e080ed8aee162f05ec472f32085a76ef85cf9879
          createdAt: "2025-05-22 12:40:45"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.12.13-1\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_gitops/
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:2c45cfc01f9ec7577afc09cadc4c1a6f470836f6a0a7c24d66b791913f89bb05
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:42a79aad641e8607fc339e4ce7f9cf1d258be01f863dbc31a94d2c27edb645bd
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:97e32825f59c33cca7d8dba3580095205c20435ccd15b91cf0aff37db9841e0e
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:7cf924af77d9c56e70710b99e080ed8aee162f05ec472f32085a76ef85cf9879
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:b7d3c882fc25fbbea28951afb197d4d47ace253864889339396e20f0c6897a7f
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:bcf6b3136a5e71803cb518c8763c165bb46ee37ef7bf55503911d7fcb36b3a8f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c369df980874aae270ca7bc6455f1ca45467d75d25fe80d7bcb06a7826bdeaa3
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:cd4965d91c163160bcf43a02c54ca60b8704281c95de904f904df942bfd83bcd
        - registry.redhat.io/rhel8/redis-6@sha256:a9f30ed08abcab0122556a5485fbcf6e0fc03860f19954dcebdbf5e754876a45
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:7358787959af7e7e589edfc043a97d80ebab87536bde365d33be7a600071bf0e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1ed3b514bf9af6ac0d386e3739874d1d964e667d54e041c2d67edd769b0009fb
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:8096f824bd96e3148cf4b6936766c888d1fff0af9e9d0be440ae4feb6adc03f2
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:1112160243f227a5d393eeee165bd4fed0d3f42646a9280cbe9af60ef728aeb2
        version: 1.14.4
      entries:
      - name: openshift-gitops-operator.v1.14.4
        version: 1.14.4
      - name: openshift-gitops-operator.v1.14.3-0.1746016855.p
        version: 1.14.3+0.1746016855.p
      - name: openshift-gitops-operator.v1.14.3-0.1742836272.p
        version: 1.14.3+0.1742836272.p
      - name: openshift-gitops-operator.v1.14.3
        version: 1.14.3
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.14
    - currentCSV: openshift-gitops-operator.v1.15.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ad49a3d14f843bde0d2365093d00ffa6d3b9d59a75c81e0c56710ad37b53bd92
          createdAt: "2024-11-25T03:57:37Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v2.13.9\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:d41454824632475f0e88aa6e0031ec643bd37e1d575df36cd3ee63daf1403281
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:6cec76d7ea70d77923cf4285177e375d34e96e3aecbe77b68cd5ff769e6e540b
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:911733d78ae6f36f642b4d63f044cd102b7eee170508eaf518ac7acf0f339698
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ad49a3d14f843bde0d2365093d00ffa6d3b9d59a75c81e0c56710ad37b53bd92
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:41717c27915251bda8afeef651b609b8a9ed410e0fa09c4cf8fa7e91d6a94292
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:0c7f72546d0d918bfd21aff7157b983ff7c832b269f5efb84d2ed723a09bf612
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:5bda26e2c16beebf66dd8f3c3c55d48e2f00db870f2d9d4d44570d9d3b8bddb0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:a9e2360f00c5c9840a100a6748bd905b8f34169140a33fb0ca4af9cf6db71642
        version: 1.15.4
      entries:
      - name: openshift-gitops-operator.v1.15.4
        version: 1.15.4
      - name: openshift-gitops-operator.v1.15.3
        version: 1.15.3
      - name: openshift-gitops-operator.v1.15.2
        version: 1.15.2
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.15
    - currentCSV: openshift-gitops-operator.v1.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ead539d7cf54992cf1bfeae6a73a453059a7dffef44505991ff33319c8d40c50
          createdAt: "2025-03-07T13:13:07Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v2.14.16\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:58284a12174d19c2d04723fa1f551d96b490872e0fa2ad94af1b3e52dc6ad955
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a5e2a12ec5dea607a6651462d8e7920cc73f3272dba4041c397a14f2c73d7532
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:03d0d9093422396330aa8611adf77abdc9be759f3d2e23850e4abe778bd6002b
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:cab809fd1288d31c93f4f5c4facb7efaff9f844e21a859d1e4557831575a13c8
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c37aa141387c847eac825363dddda3eb59203d3694120dba439b00cf0c66f320
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:2598a3339bc920b91018d3f7acbe78ccbee684b5182c874e48296744d2373f9d
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ead539d7cf54992cf1bfeae6a73a453059a7dffef44505991ff33319c8d40c50
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:ceff1cde6d043daae599b578eec9759c6d49cd5cf770460b9d075fa879a16e81
        version: 1.16.3
      entries:
      - name: openshift-gitops-operator.v1.16.3
        version: 1.16.3
      - name: openshift-gitops-operator.v1.16.2
        version: 1.16.2
      - name: openshift-gitops-operator.v1.16.1
        version: 1.16.1
      - name: openshift-gitops-operator.v1.16.0-0.1746014725.p
        version: 1.16.0+0.1746014725.p
      - name: openshift-gitops-operator.v1.16.0
        version: 1.16.0
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.16
    - currentCSV: openshift-gitops-operator.v1.17.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:1b86eb4e071b1c526931b6d18c3d8f51522293699059c514488602c320271049
          createdAt: "2025-07-30T13:03:16Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.0.14\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:3fc4941fa4a6f2dbe9dd82622a0d3fb794ecd5a9dee286a38b60a7f6a947d685
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:cc45e212e7eb4e7e8daac6ea74071d5d98260de0b1f6d1cea0e7876111ccea8c
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:2f5f997bce924445de735ae0508dca1a7bba561bc4acdacf659928488233cb8a
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:5c9ea426cd60e7b8d1d8e4fe763909200612434c65596855334054e26cbfe3d0
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:70e5e3900d0c89999c69be1d58adf84c1b3664aacbffa9c5e0efc8ae1b6ef9c1
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:1b86eb4e071b1c526931b6d18c3d8f51522293699059c514488602c320271049
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:d73b4f5b355f556208751281923bb8b43edffc061a673c04d53b88f0e50ecb26
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:a2ae66772d90e734bb3385a59852e6016325a21b916a2580a33cf0da7cb0b527
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:c0a513ed54b631629efafe6e94b775d5bd0a943c44fbe96f02a0d5bb1f989099
        version: 1.17.1
      entries:
      - name: openshift-gitops-operator.v1.17.1
        version: 1.17.1
      - name: openshift-gitops-operator.v1.17.0
        version: 1.17.0
      name: gitops-1.17
    - currentCSV: openshift-gitops-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
          createdAt: "2025-08-21T01:20:45Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.18.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: NamespaceManagement
            name: namespacemanagements.argoproj.io
            version: v1beta1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.1.6\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/rhel9/redis-7@sha256:b4f58ad83ab8719a2e102f38ee526546cd1d56428901afc3add45c9a0e2a2f12
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:1eef9ecd58cf8391f29360ef74dd8aa5dcb48e1582d1500e445706d8302c951e
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a0544c14823492165550d83a6d8ba79dd632b46144d3fdcb543793726111d76
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:6284873bcf1593c210003e0292ac5583f03f03e80d937098d808d48fa0d01cf6
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:5a09fcba3d9e6fc462719e81a46af6022cf7f70a05bcb387ed959a49f6f37dbc
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:adca9c55b80d899d500bf0a7234cf61ae30ffed1b2c927ddea8d9fd07c9acb6a
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7f6ceae56c60fd432425c36f93a9e9bbbb96ef9422333ad6ea41ff3a2eaf8374
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:640e3298c7cd2faa84971b0d861446c4892ab8602108af05b8b5b049a4de624f
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c2283997d47298d49378e8200b149ea613c7735317ef12e74b7915a590f41ce2
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:08efd3e7a79c2945693c99ed31ec912223becff7238b1e28d1b1da5d63f93019
        version: 1.18.0
      entries:
      - name: openshift-gitops-operator.v1.18.0
        version: 1.18.0
      name: gitops-1.18
    - currentCSV: openshift-gitops-operator.v1.6.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:104cc7564fcc3261faa3fbf9c95720b4434e15103b876c080ed012b244ef1dae
          createdAt: "2021-09-15T22:31:21Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD v2.4.28\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces. \n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc.
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:104cc7564fcc3261faa3fbf9c95720b4434e15103b876c080ed012b244ef1dae
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:0cc4fa704e581cd5581c816668183f602fae6ef9ab3c04ce0d99d561287ef140
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:06f1150114d4c0bd236d2a5d4716c9c3e95a9733058297790389d6ba2a1bf264
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:e9e2f950cda8153cff112e43872b98e4977d3cbdf59713b8de92defef6d34b6d
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:dbcad68b70ab7918e944099aef4177c9bc3378e9ca49499dbb94a175927d3209
        - registry.redhat.io/rhel8/redis-6@sha256:53598a6effeb90e4f1b005b2521beffd2fa2b0c52d0e7f2347ee2abd2577cab3
        version: 1.6.6
      entries:
      - name: openshift-gitops-operator.v1.6.6
        version: 1.6.6
      - name: openshift-gitops-operator.v1.6.5
        version: 1.6.5
      - name: openshift-gitops-operator.v1.6.4
        version: 1.6.4
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.6
    - currentCSV: openshift-gitops-operator.v1.7.4-0.1690486082.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:e0967a4058fe5a9bfe7e998a3347529c99536cf3a44de9e40f9632035ff5a2fe
          createdAt: "2021-09-15T22:31:21Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.substitutesFor: openshift-gitops-operator.v1.7.4
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD v2.5.16\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficient privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces. \n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/rhel8/redis-6@sha256:53598a6effeb90e4f1b005b2521beffd2fa2b0c52d0e7f2347ee2abd2577cab3
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:e0967a4058fe5a9bfe7e998a3347529c99536cf3a44de9e40f9632035ff5a2fe
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c55aadf45397083bec4e42d6c0514ed79169a07fd3d9d75671904df2c0520e52
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:45789c49d9334e0ed062641313714c101bd0117dc5633316b1ec63e7a6c76c15
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:26050fb95978e90e8d00e895032045e77cb236008a0e1417459e1aeca8a48c69
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:5c11be8ec9db54331893f2f298aa27006e77b0eaded2d88d0531c77da91a421f
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:f1371d2012796e9aaca105112190ff89ec3d0be95968e05b19db101480f67975
        version: 1.7.4+0.1690486082.p
      entries:
      - name: openshift-gitops-operator.v1.7.4-0.1690486082.p
        version: 1.7.4+0.1690486082.p
      - name: openshift-gitops-operator.v1.7.4
        version: 1.7.4
      - name: openshift-gitops-operator.v1.7.3
        version: 1.7.3
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.7
    - currentCSV: openshift-gitops-operator.v1.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f32df77b3c23d6a833ff887386285ca7a5bd436641f0ec1ffc934c8247fa315f
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.skipRange: ""
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.6.153\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:8cc59901689858e06f5c2942f8c975449c17454fa8364da6153d9b5a3538a985
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:f77594bc053be144b33ff9603ee9675c7e82946ec0dbfb04d8f942c8d73155da
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f32df77b3c23d6a833ff887386285ca7a5bd436641f0ec1ffc934c8247fa315f
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:8aa68fddef88921aeeede24a849d3dabbceff01dc03185342c21e6ee58033eaf
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:16d9184345e6d5a19f326a7b52f76d8a59b96763f233c50546ff017a625e5ff5
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:0d13c03048b4435d9f5045a6fb7bd4b8808461b26aeca60c0047fa5937e1d45d
        version: 1.8.6
      entries:
      - name: openshift-gitops-operator.v1.8.6
        version: 1.8.6
      - name: openshift-gitops-operator.v1.8.5
        version: 1.8.5
      - name: openshift-gitops-operator.v1.8.4-0.1693214052.p
        version: 1.8.4+0.1693214052.p
      - name: openshift-gitops-operator.v1.8.4
        version: 1.8.4
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.8
    - currentCSV: openshift-gitops-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "argo-rollout"
                },
                "spec": null
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:11a6084e11aa32511a34bd6cfb4f4995aec690c1f5ed61861e99f6e42d62afb4
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.skipRange: ""
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.7.16\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:80f92842fe4cba008a873ad2ab6059a3a24e76e16671d8ece034dbbe666996bc
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:3d17786e2f4063154b9e3a40b96bf8ec0a14b86395cea5fe7ca18c5b43764478
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a568c8c647e2e2d6abdf12138eeef8f8bdde1fdcbf786e9ca8c8f93bc79acaf9
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:43bb961d847477902e66a92edd1539322cac8d5f74706b048a458ad037a0c7a3
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:11a6084e11aa32511a34bd6cfb4f4995aec690c1f5ed61861e99f6e42d62afb4
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ec9f60018694dcc5d431ba47d5536b761b71cb3f66684978fe6bb74c157679ac
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:3873d1f18b5643e1de2601419410270342650a21428bb26a4cbee20c0c63f98b
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:86012145d4b8798dc12f6f4f07a9316a193188aaac865e324b63f4e60e8ec584
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:136d3a936759e112d3ea2e78f7c02bd882b0fd3011f4c452f45c32c43822772d
        version: 1.9.4
      entries:
      - name: openshift-gitops-operator.v1.9.4
        version: 1.9.4
      - name: openshift-gitops-operator.v1.9.3
        version: 1.9.3
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.9
    - currentCSV: openshift-gitops-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
          createdAt: "2025-08-21T01:20:45Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.18.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: NamespaceManagement
            name: namespacemanagements.argoproj.io
            version: v1beta1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.1.6\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:adca9c55b80d899d500bf0a7234cf61ae30ffed1b2c927ddea8d9fd07c9acb6a
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7f6ceae56c60fd432425c36f93a9e9bbbb96ef9422333ad6ea41ff3a2eaf8374
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:6284873bcf1593c210003e0292ac5583f03f03e80d937098d808d48fa0d01cf6
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:08efd3e7a79c2945693c99ed31ec912223becff7238b1e28d1b1da5d63f93019
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:640e3298c7cd2faa84971b0d861446c4892ab8602108af05b8b5b049a4de624f
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c2283997d47298d49378e8200b149ea613c7735317ef12e74b7915a590f41ce2
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:5a09fcba3d9e6fc462719e81a46af6022cf7f70a05bcb387ed959a49f6f37dbc
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:1eef9ecd58cf8391f29360ef74dd8aa5dcb48e1582d1500e445706d8302c951e
        - registry.redhat.io/rhel9/redis-7@sha256:b4f58ad83ab8719a2e102f38ee526546cd1d56428901afc3add45c9a0e2a2f12
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a0544c14823492165550d83a6d8ba79dd632b46144d3fdcb543793726111d76
        version: 1.18.0
      entries:
      - name: openshift-gitops-operator.v1.18.0
        version: 1.18.0
      - name: openshift-gitops-operator.v1.17.1
        version: 1.17.1
      - name: openshift-gitops-operator.v1.17.0
        version: 1.17.0
      - name: openshift-gitops-operator.v1.16.2
        version: 1.16.2
      - name: openshift-gitops-operator.v1.16.1
        version: 1.16.1
      - name: openshift-gitops-operator.v1.16.0-0.1746014725.p
        version: 1.16.0+0.1746014725.p
      - name: openshift-gitops-operator.v1.16.0
        version: 1.16.0
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: latest
    defaultChannel: latest
    packageName: openshift-gitops-operator
    provider:
      name: Red Hat Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: node-maintenance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-maintenance-operator.v0.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-maintenance-operator:v0.17.0
          createdAt: "2024-03-13 18:56:47"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=0.12.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/medik8s/node-maintenance-operator:v0.17.0
        version: 0.17.0
      entries:
      - name: node-maintenance-operator.v0.17.0
        version: 0.17.0
      - name: node-maintenance-operator.v0.16.1
        version: 0.16.1
      - name: node-maintenance-operator.v0.16.0
        version: 0.16.0
      - name: node-maintenance-operator.v0.15.0
        version: 0.15.0
      - name: node-maintenance-operator.v0.14.0
        version: 0.14.0
      - name: node-maintenance-operator.v0.13.0
        version: 0.13.0
      - name: node-maintenance-operator.v0.12.1
        version: 0.12.1
      name: stable
    defaultChannel: stable
    packageName: node-maintenance-operator
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudwatch-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudwatch-controller.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudwatch-controller:1.3.0
          createdAt: "2025-10-02T23:05:17Z"
          description: AWS CloudWatch controller is a service controller for managing
            CloudWatch resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Dashboard represents the state of an AWS cloudwatch Dashboard
              resource.
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.cloudwatch.services.k8s.aws
            version: v1alpha1
          - description: MetricAlarm represents the state of an AWS cloudwatch MetricAlarm
              resource.
            displayName: MetricAlarm
            kind: MetricAlarm
            name: metricalarms.cloudwatch.services.k8s.aws
            version: v1alpha1
          - description: MetricStream represents the state of an AWS cloudwatch MetricStream
              resource.
            displayName: MetricStream
            kind: MetricStream
            name: metricstreams.cloudwatch.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudWatch resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudWatch**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudWatch
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudwatch
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudWatch Developer Resources
          url: https://aws.amazon.com/CloudWatch/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudwatch maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudwatch-controller:1.3.0
        version: 1.3.0
      entries:
      - name: ack-cloudwatch-controller.v1.3.0
        version: 1.3.0
      - name: ack-cloudwatch-controller.v1.2.1
        version: 1.2.1
      - name: ack-cloudwatch-controller.v1.2.0
        version: 1.2.0
      - name: ack-cloudwatch-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudwatch-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudwatch-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudwatch-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudwatch-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudwatch-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudwatch-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudwatch-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudwatch-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudwatch-controller.v1.0.5
        version: 1.0.5
      - name: ack-cloudwatch-controller.v1.0.4
        version: 1.0.4
      - name: ack-cloudwatch-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudwatch-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudwatch-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudwatch-controller.v0.0.12
        version: 0.0.12
      - name: ack-cloudwatch-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudwatch-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudwatch-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudwatch-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudwatch-controller.v0.0.3
        version: 0.0.3
      - name: ack-cloudwatch-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudwatch-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OpenShift Database Access
      provider-url: ""
    name: dbaas-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dbaas-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSConnection",
                "metadata": {
                  "name": "dbaasconnection-example"
                },
                "spec": {
                  "databaseServiceID": "instance-example",
                  "databaseServiceType": "instance",
                  "inventoryRef": {
                    "name": "inventory-example",
                    "namespace": "openshift-dbaas-operator"
                  }
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSInstance",
                "metadata": {
                  "name": "dbaasinstance-sample"
                },
                "spec": {
                  "cloudProvider": "AWS",
                  "cloudRegion": "US_EAST_1",
                  "inventoryRef": {
                    "name": "inventory-example",
                    "namespace": "openshift-dbaas-operator"
                  },
                  "name": "my-cluster-free"
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSInventory",
                "metadata": {
                  "name": "inventory-example"
                },
                "spec": {
                  "credentialsRef": {
                    "name": "dbaas-vendor-credentials"
                  },
                  "providerRef": {
                    "name": "provider-example-registration"
                  }
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSPolicy",
                "metadata": {
                  "name": "dbaaspolicy-sample"
                },
                "spec": {
                  "disableProvisions": false
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSProvider",
                "metadata": {
                  "labels": {
                    "related-to": "dbaas-operator",
                    "type": "dbaas-provider-registration"
                  },
                  "name": "provider-example-registration"
                },
                "spec": {
                  "allowsFreeTrial": true,
                  "connectionKind": "ProviderConnection",
                  "credentialFields": [
                    {
                      "displayName": "ProviderAPIKeyD",
                      "helpText": "This is the API Key for the example provider to show on UI",
                      "key": "APIKey",
                      "required": true,
                      "type": "maskedstring"
                    }
                  ],
                  "externalProvisionDescription": "Follow the guide to start a free Provider Serverless (beta) cluster",
                  "externalProvisionURL": "https://www.exmample.com/docs/provider/quickstart.html",
                  "groupVersion": "dbaas.redhat.com/v1beta1",
                  "instanceKind": "ProviderInstance",
                  "inventoryKind": "ProviderInventory",
                  "provider": {
                    "displayDescription": "This is an example for providers on how to implement their operator to integrate with DBaaS.",
                    "displayName": "DBaaS Provider Example ",
                    "icon": {
                      "base64data": "SGVsbG8sIHdvcmxkLg==",
                      "mediatype": "image/png"
                    },
                    "name": "Provider Example"
                  },
                  "provisioningParameters": {
                    "cloudProvider": {
                      "conditionalData": [
                        {
                          "defaultValue": "GCP",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "FREETRIAL"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        },
                        {
                          "defaultValue": "AWS",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "SERVERLESS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Amazon Web Services",
                              "value": "AWS"
                            },
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        },
                        {
                          "defaultValue": "AWS",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Amazon Web Services",
                              "value": "AWS"
                            },
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        }
                      ],
                      "displayName": "Cloud Provider"
                    },
                    "dedicatedLocationLabel": {
                      "displayName": "Select regions \u0026 nodes",
                      "helpText": "Select the geographical region where you want the database instance to run, and set the number of nodes you want running in this dedicated cluster."
                    },
                    "hardwareLabel": {
                      "displayName": "Hardware per node",
                      "helpText": "Select the compute and storage requirements for this database instance."
                    },
                    "machineType": {
                      "conditionalData": [
                        {
                          "defaultValue": "m5.large",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "AWS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "2 vCPU, 8 GiB RAM",
                              "value": "m5.large"
                            },
                            {
                              "displayValue": "4 vCPU, 16 GiB RAM",
                              "value": "m5.xlarge"
                            },
                            {
                              "displayValue": "8 vCPU, 32 GiB RAM",
                              "value": "m5.2xlarge"
                            },
                            {
                              "displayValue": "16 vCPU, 64 GiB RAM",
                              "value": "m5.4xlarge"
                            },
                            {
                              "displayValue": "32 vCPU, 128 GiB RAM",
                              "value": "m5.8xlarge"
                            }
                          ]
                        },
                        {
                          "defaultValue": "n1-standard-2",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "GCP"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "2 vCPU, 7.5 GiB RAM",
                              "value": "n1-standard-2"
                            },
                            {
                              "displayValue": "4 vCPU, 15 GiB RAM",
                              "value": "n1-standard-4"
                            },
                            {
                              "displayValue": "8 vCPU, 30 GiB RAM",
                              "value": "n1-standard-8"
                            },
                            {
                              "displayValue": "16 vCPU, 60 GiB RAM",
                              "value": "n1-standard-16"
                            },
                            {
                              "displayValue": "32 vCPU, 120 GiB RAM",
                              "value": "n1-standard-32"
                            }
                          ]
                        }
                      ],
                      "displayName": "Compute"
                    },
                    "name": {
                      "displayName": "Cluster name"
                    },
                    "nodes": {
                      "displayName": "nodes"
                    },
                    "plan": {
                      "conditionalData": [
                        {
                          "defaultValue": "SERVERLESS",
                          "options": [
                            {
                              "displayValue": "Free trial",
                              "value": "FREETRIAL"
                            },
                            {
                              "displayValue": "Serverless",
                              "value": "SERVERLESS"
                            },
                            {
                              "displayValue": "Dedicated",
                              "value": "DEDICATED"
                            }
                          ]
                        }
                      ],
                      "displayName": "Hosting plan"
                    },
                    "planLabel": {
                      "displayName": "Select a plan"
                    },
                    "regions": {
                      "displayName": "us-west-2"
                    },
                    "serverlessLocationLabel": {
                      "displayName": "Select regions",
                      "helpText": "Select the geographical region where you want the database instance to run."
                    },
                    "spendLimit": {
                      "conditionalData": [
                        {
                          "defaultValue": "0",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "SERVERLESS"
                            }
                          ]
                        }
                      ],
                      "displayName": "Spend limit"
                    },
                    "spendLimitLabel": {
                      "displayName": "Spend limit",
                      "helpText": "Set a spending limit on resources for this database instance."
                    },
                    "storageGib": {
                      "conditionalData": [
                        {
                          "defaultValue": "15",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "AWS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "15 GiB",
                              "value": "15"
                            },
                            {
                              "displayValue": "35 GiB",
                              "value": "35"
                            },
                            {
                              "displayValue": "75 GiB",
                              "value": "75"
                            },
                            {
                              "displayValue": "150 GiB",
                              "value": "150"
                            },
                            {
                              "displayValue": "300 GiB",
                              "value": "300"
                            },
                            {
                              "displayValue": "600 GiB",
                              "value": "600"
                            }
                          ]
                        },
                        {
                          "defaultValue": "15",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "GCP"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "15 GiB",
                              "value": "15"
                            },
                            {
                              "displayValue": "35 GiB",
                              "value": "35"
                            },
                            {
                              "displayValue": "75 GiB",
                              "value": "75"
                            },
                            {
                              "displayValue": "150 GiB",
                              "value": "150"
                            },
                            {
                              "displayValue": "300 GiB",
                              "value": "300"
                            },
                            {
                              "displayValue": "600 GiB",
                              "value": "600"
                            }
                          ]
                        }
                      ],
                      "displayName": "Storage"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: quay.io/ecosystem-appeng/dbaas-operator:v0.5.0
          description: Discover & connect with database instances hosted on 3rd-party
            ISV cloud platforms.
          operatorframework.io/suggested-namespace: openshift-dbaas-operator
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/internal-objects: '["dbaasconnections.dbaas.redhat.com","dbaasproviders.dbaas.redhat.com","dbaasplatforms.dbaas.redhat.com","dbaasinstances.dbaas.redhat.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/RHEcosystemAppEng/dbaas-operator
          support: rhoda-requests@redhat.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DBaaSConnection
            name: dbaasconnections.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSConnection defines the schema for the DBaaSConnection
              API.
            displayName: DBaaSConnection
            kind: DBaaSConnection
            name: dbaasconnections.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSInstance
            name: dbaasinstances.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSInstance defines the schema for the DBaaSInstance API.
            displayName: DBaaSInstance
            kind: DBaaSInstance
            name: dbaasinstances.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSInventory
            name: dbaasinventories.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSInventory defines the schema for the DBaaSInventory
              API. Inventory objects must be created in a valid namespace, determined
              by the existence of a DBaaSPolicy object.
            displayName: Provider Account
            kind: DBaaSInventory
            name: dbaasinventories.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSPlatform
            name: dbaasplatforms.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSPlatform defines the schema for the DBaaSPlatform API.
            displayName: DBaaSPlatform
            kind: DBaaSPlatform
            name: dbaasplatforms.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSPolicy
            name: dbaaspolicies.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSPolicy enables administrative capabilities within a
              namespace, and sets a default inventory policy. Policy defaults can
              be overridden on a per-inventory basis.
            displayName: Provider Account Policy
            kind: DBaaSPolicy
            name: dbaaspolicies.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSProvider
            name: dbaasproviders.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSProvider defines the schema for the DBaaSProvider API.
            displayName: DBaaSProvider
            kind: DBaaSProvider
            name: dbaasproviders.dbaas.redhat.com
            version: v1beta1
        description: |
          The OpenShift Database Access Operator enables OpenShift users to discover & connect with database instances
          hosted on 3rd-party ISV cloud platforms such as CrunchyData Bridge & CockroachCloud.

          ## Feedback & Support

          Please contact us at [rhoda-requests@redhat.com](mailto:rhoda-requests@redhat.com)

          For more information on how to use our service, please refer to our [Quick Starts](quickstart?keyword=database).
        displayName: OpenShift Database Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rhoda
        - oda
        - dbaas
        - database
        - db
        - data
        - access
        - crunchy
        - bridge
        - cockroach
        - cockroachdb
        - cloud
        - ack
        - amazon
        - rds
        - relational
        links:
        - name: GitHub
          url: https://github.com/RHEcosystemAppEng/dbaas-operator
        maintainers:
        - email: ecosystem-appeng@redhat.com
          name: Red Hat Ecosystem App Engineering
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: OpenShift Database Access
        relatedImages:
        - quay.io/openshift/origin-kube-rbac-proxy:4.10.0
        - quay.io/ecosystem-appeng/dbaas-dynamic-plugin:0.5.0
        - registry.developers.crunchydata.com/crunchydata/crunchy-bridge-operator-catalog:v0.0.6
        - gcr.io/cockroach-shared/ccapi-k8s-operator-catalog:v0.0.5
        - quay.io/ecosystem-appeng/rds-dbaas-operator-catalog:v0.4.0
        - quay.io/ecosystem-appeng/dbaas-operator:v0.5.0
        version: 0.5.0
      entries:
      - name: dbaas-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: dbaas-operator
    provider:
      name: OpenShift Database Access
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ControlPlane Ltd.
      provider-url: https://control-plane.io
    name: flux-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux-operator.v0.31.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "FluxInstance",
                "metadata": {
                  "name": "flux",
                  "namespace": "flux-system",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcileEvery": "1h",
                    "fluxcd.controlplane.io/reconcileTimeout": "5m"
                  }
                },
                "spec": {
                  "distribution": {
                    "version": "2.x",
                    "registry": "ghcr.io/fluxcd"
                  },
                  "components": [
                    "source-controller",
                    "kustomize-controller",
                    "helm-controller",
                    "notification-controller",
                    "image-reflector-controller",
                    "image-automation-controller"
                  ],
                  "cluster": {
                    "type": "openshift",
                    "size": "medium",
                    "domain": "cluster.local",
                    "networkPolicy": true,
                    "multitenant": true,
                    "tenantDefaultServiceAccount": "flux"
                  }
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "FluxReport",
                "metadata": {
                  "name": "flux",
                  "namespace": "flux-system"
                },
                "spec": {
                  "distribution": {
                    "entitlement": "Issued by controlplane",
                    "managedBy": "flux-operator",
                    "status": "Installed",
                    "version": "v2.3.0"
                  },
                "components": [
                    {
                      "image": "ghcr.io/fluxcd/kustomize-controller:v1.3.0@sha256:48a032574dd45c39750ba0f1488e6f1ae36756a38f40976a6b7a588d83acefc1",
                      "name": "kustomize-controller",
                      "ready": true,
                      "status": "Current Deployment is available. Replicas: 1"
                    },
                    {
                      "image": "ghcr.io/fluxcd/source-controller:v1.3.0@sha256:161da425b16b64dda4b3cec2ba0f8d7442973aba29bb446db3b340626181a0bc",
                      "name": "source-controller",
                      "ready": true,
                      "status": "Current Deployment is available. Replicas: 1"
                    }
                  ],
                "reconcilers": [
                    {
                      "apiVersion": "kustomize.toolkit.fluxcd.io/v1",
                      "kind": "Kustomization",
                      "stats": {
                        "failing": 0,
                        "running": 1,
                        "suspended": 0
                      }
                    },
                    {
                      "apiVersion": "source.toolkit.fluxcd.io/v1",
                      "kind": "GitRepository",
                      "stats": {
                        "failing": 0,
                        "running": 1,
                        "suspended": 0,
                        "totalSize": "3.7 MiB"
                      }
                    }
                  ],
                  "sync": {
                    "ready": true,
                    "id": "kustomization/flux-system",
                    "path": "clusters/production",
                    "source": "https://github.com/my-org/my-fleet.git",
                    "status": "Applied revision: refs/heads/main@sha1:a90cd1ac35de01c175f7199315d3f4cd60195911"
                  }
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "ResourceSet",
                "metadata": {
                  "name": "podinfo",
                  "namespace": "default",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcile": "enabled",
                    "fluxcd.controlplane.io/reconcileEvery": "30m",
                    "fluxcd.controlplane.io/reconcileTimeout": "5m"
                  }
                },
                "spec": {
                  "commonMetadata": {
                    "labels": {
                      "app.kubernetes.io/name": "podinfo"
                    }
                  },
                  "inputs": [
                    {
                      "tenant": "team1",
                      "version": "6.7.x",
                      "replicas": "2"
                    },
                    {
                      "tenant": "team2",
                      "version": "6.6.x",
                      "replicas": "3"
                    }
                  ],
                  "resources": [
                    {
                      "apiVersion": "source.toolkit.fluxcd.io/v1beta2",
                      "kind": "OCIRepository",
                      "metadata": {
                        "name": "podinfo-<< inputs.tenant >>",
                        "namespace": "default"
                      },
                      "spec": {
                        "interval": "10m",
                        "url": "oci://ghcr.io/stefanprodan/charts/podinfo",
                        "ref": {
                          "semver": "<< inputs.version | quote >>"
                        }
                      }
                    },
                    {
                      "apiVersion": "helm.toolkit.fluxcd.io/v2",
                      "kind": "HelmRelease",
                      "metadata": {
                        "name": "podinfo-<< inputs.tenant >>",
                        "namespace": "default"
                      },
                      "spec": {
                        "interval": "1h",
                        "releaseName": "podinfo-<< inputs.tenant >>",
                        "chartRef": {
                          "kind": "OCIRepository",
                          "name": "podinfo-<< inputs.tenant >>"
                        },
                        "values": {
                          "replicaCount": "<< inputs.replicas | int >>"
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "ResourceSetInputProvider",
                "metadata": {
                  "name": "demo",
                  "namespace": "default",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcile": "enabled",
                    "fluxcd.controlplane.io/reconcileEvery": "10m",
                    "fluxcd.controlplane.io/reconcileTimeout": "1m"
                  }
                },
                "spec": {
                  "type": "GitHubPullRequest",
                  "url": "https://github.com/fluxcd-testing/pr-testing",
                  "filter": {
                    "limit": 10,
                    "includeBranch": "^stefanprodan-patch-.*$",
                    "labels": [
                      "enhancement"
                    ]
                  },
                  "skip": {
                    "labels": [
                        "deploy/flux-preview-pause",
                        "!test-build-push/passed"
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/controlplaneio-fluxcd/flux-operator:v0.31.0-ubi
          createdAt: "2025-10-06T20:33:03Z"
          description: The Flux Operator manages the lifecycle of the CNCF Flux project.
          operatorframework.io/suggested-namespace: flux-system
          repository: https://github.com/controlplaneio-fluxcd/flux-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Flux Instance
            displayName: FluxInstance
            kind: FluxInstance
            name: fluxinstances.fluxcd.controlplane.io
            version: v1
          - description: Flux Report (Autogenerated)
            displayName: FluxReport
            kind: FluxReport
            name: fluxreports.fluxcd.controlplane.io
            version: v1
          - description: Resource Set
            displayName: ResourceSet
            kind: ResourceSet
            name: resourcesets.fluxcd.controlplane.io
            version: v1
          - description: ResourceSet Input Provider
            displayName: ResourceSetInputProvider
            kind: ResourceSetInputProvider
            name: resourcesetinputproviders.fluxcd.controlplane.io
            version: v1
        description: |-
          [Flux](https://fluxcd.io) is a leading GitOps Continuous Delivery tool used to streamline and automate application deployments on Kubernetes. Flux enables teams to achieve a reliable and auditable delivery process while promoting collaboration and traceability across different environments.
          Flux is powered by the GitOps Toolkit, a set of composable APIs and specialized tools that enable a wide range of continuous delivery use-cases, from simple Kubernetes deployment pipelines to multi-tenant and multi-cluster progressive delivery rollouts.
          For more information about the CNCF-graduated Flux project, please see the [Flux architecture overview](https://fluxcd.control-plane.io/guides/flux-architecture/).
          ### Operator Capabilities
          The [Flux Operator](https://github.com/controlplaneio-fluxcd/flux-operator) provides a declarative API for the installation, configuration and upgrade of CNCF Flux and the ControlPlane [enterprise distribution](https://fluxcd.control-plane.io/).
          The operator allows the configuration of Flux multi-tenancy lockdown, network policies, persistent storage, sharding, vertical scaling, custom patches, and the synchronization of the cluster state from Git repositories, OCI artifacts and S3-compatible storage.
          The Flux Operator supervises the Flux controllers and provides a unified view of all the Flux resources that define the GitOps workflows for the target cluster. The operator generates reports, emits events, and exports Prometheus metrics to help with monitoring and troubleshooting Flux.
          ### Operator Features
          The operator extends Flux with self-service capabilities, deployment windows, and preview environments for GitHub, GitLab and Azure DevOps pull requests testing.
          The operator ResourceSet API enables platform teams to define their own application standard as a group of Flux and Kubernetes resources that can be templated, parameterized and deployed as a single unit on self-service environments. For more information, please see the [ResourceSet documentation](https://fluxcd.control-plane.io/operator/resourcesets/introduction/).
          ### OpenShift Support
          The Flux Operator should be installed in a dedicated namespace, e.g. `flux-system`.
          To deploy Flux on OpenShift clusters, create a `FluxInstance` custom resource named `flux` in the same namespace as the operator and set the `.spec.cluster.type` field to `openshift`.
          For more information on how to configure the Flux instance, please see the [Flux Operator documentation](https://fluxcd.control-plane.io/operator/flux-config/).
        displayName: Flux Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - fluxcd
        - gitops
        - continuous delivery
        links:
        - name: Source Code
          url: https://github.com/controlplaneio-fluxcd/flux-operator
        - name: Documentation
          url: https://fluxcd.control-plane.io/operator/
        - name: Enterprise Support
          url: https://fluxcd.control-plane.io/pricing/
        maintainers:
        - email: flux-enterprise@control-plane.io
          name: ControlPlane Flux Team
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: ControlPlane Ltd.
          url: https://control-plane.io
        relatedImages:
        - ghcr.io/controlplaneio-fluxcd/flux-operator:v0.31.0-ubi
        version: 0.31.0
      entries:
      - name: flux-operator.v0.31.0
        version: 0.31.0
      - name: flux-operator.v0.30.0
        version: 0.30.0
      - name: flux-operator.v0.28.0
        version: 0.28.0
      - name: flux-operator.v0.27.0
        version: 0.27.0
      - name: flux-operator.v0.26.0
        version: 0.26.0
      - name: flux-operator.v0.24.1
        version: 0.24.1
      - name: flux-operator.v0.24.0
        version: 0.24.0
      - name: flux-operator.v0.23.0
        version: 0.23.0
      - name: flux-operator.v0.22.0
        version: 0.22.0
      - name: flux-operator.v0.21.0
        version: 0.21.0
      - name: flux-operator.v0.20.0
        version: 0.20.0
      - name: flux-operator.v0.19.0
        version: 0.19.0
      - name: flux-operator.v0.18.0
        version: 0.18.0
      name: stable
    defaultChannel: stable
    packageName: flux-operator
    provider:
      name: ControlPlane Ltd.
      url: https://control-plane.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-load-core
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-load-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Loadcore",
                "metadata": {
                  "name": "loadcore-sample"
                },
                "spec": {
                  "accessRoute": {
                    "enabled": true,
                    "host": "loadcore.apps.cluster.perf.keysight.lab"
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "reportTemplates": {
                    "enabled": false
                  },
                  "wirelessDataModel": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:42:44Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Loadcore
            name: loadcores.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight LoadCore Operator
        displayName: Keysight LoadCore Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Load Core
          url: https://keysight-load-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-load-core@sha256:245d9be60ff6e9a6e5e42a026a4ee7b8168f391a0491de23ae827d1906af88e5
        version: 0.1.21
      entries:
      - name: keysight-load-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-load-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CloudBees
      provider-url: https://www.cloudbees.com/products/continuous-integration
    name: cloudbees-ci-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudbees-ci.v3.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.cloudbees.com/v1alpha1",
                "kind": "CloudBeesCI",
                "metadata": {
                  "name": "cloudbeesci-sample"
                },
                "spec": {
                  "Agents": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-agent:2.289.3.2"
                    },
                    "ImagePullSecrets": null,
                    "SeparateNamespace": {
                      "Create": false,
                      "Enabled": false,
                      "Name": null
                    }
                  },
                  "Hibernation": {
                    "Enabled": false,
                    "Image": {
                      "dockerImage": "cloudbees/managed-master-hibernation-monitor:247.c5dfce00a179",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "NodeSelector": {},
                    "Tolerations": []
                  },
                  "Master": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-mm:2.289.3.2"
                    },
                    "JavaOpts": null,
                    "OperationsCenterNamespace": null
                  },
                  "NetworkPolicy": {
                    "Enabled": false,
                    "JMXSelectors": [],
                    "ingressControllerSelector": []
                  },
                  "OperationsCenter": {
                    "AgentListenerPort": 50000,
                    "Annotations": {},
                    "CSRF": {
                      "ProxyCompatibility": false
                    },
                    "CasC": {
                      "ConfigMapName": "oc-casc-bundle",
                      "Enabled": false
                    },
                    "ContainerPort": 8080,
                    "ContextPath": null,
                    "Enabled": true,
                    "ExtraConfigMaps": [],
                    "ExtraContainers": [],
                    "ExtraGroovyConfiguration": {},
                    "ExtraVolumeMounts": [],
                    "ExtraVolumes": [],
                    "HealthProbeLivenessFailureThreshold": 12,
                    "HealthProbes": true,
                    "HostName": null,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-cloud-core-oc:2.289.3.2",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "Ingress": {
                      "Annotations": {
                        "kubernetes.io/tls-acme": "false"
                      },
                      "Class": "nginx",
                      "tls": {
                        "Enable": false,
                        "SecretName": null
                      }
                    },
                    "JavaOpts": null,
                    "JenkinsOpts": null,
                    "LoadBalancerIP": null,
                    "LoadBalancerSourceRanges": [
                      "0.0.0.0/0"
                    ],
                    "Name": "cjoc",
                    "NodeSelector": {},
                    "Platform": "standard",
                    "Protocol": "http",
                    "Resources": {
                      "Limits": {
                        "Cpu": 1,
                        "Memory": "2G"
                      },
                      "Requests": {
                        "Cpu": 1,
                        "Memory": "2G"
                      }
                    },
                    "Route": {
                      "tls": {
                        "CACertificate": null,
                        "Certificate": null,
                        "DestinationCACertificate": null,
                        "Enable": false,
                        "InsecureEdgeTerminationPolicy": "Redirect",
                        "Key": null,
                        "Termination": "edge"
                      }
                    },
                    "ServiceAgentListenerPort": 50000,
                    "ServiceAnnotations": {},
                    "ServicePort": 80,
                    "ServiceType": "ClusterIP",
                    "Tolerations": []
                  },
                  "Persistence": {
                    "AccessMode": "ReadWriteOnce",
                    "Annotations": {},
                    "Size": "20Gi",
                    "StorageClass": null
                  },
                  "PodSecurityPolicy": {
                    "Annotations": {},
                    "Enabled": false
                  },
                  "Subdomain": false,
                  "ingress-nginx": {
                    "Enabled": false,
                    "controller": {
                      "admissionWebhooks": {
                        "patch": {
                          "nodeSelector": {
                            "kubernetes.io/os": "linux"
                          }
                        }
                      },
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "nginx-ingress": {
                    "Enabled": false,
                    "controller": {
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "rbac": {
                    "agentsServiceAccountName": "jenkins-agents",
                    "hibernationMonitorServiceAccountName": "managed-master-hibernation-monitor",
                    "install": true,
                    "masterServiceAccountName": "jenkins",
                    "serviceAccountName": "cjoc"
                  },
                  "sidecarinjector": {
                    "Enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cloudbees-ci-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cloudbees-ci-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudBeesCI
            name: cloudbeescis.charts.cloudbees.com
            version: v1alpha1
        description: CloudBees CI
        displayName: CloudBees CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudbees
        - ci
        - enterprise
        - continuous
        - integration
        - jenkins
        - automation
        - pipeline
        - devops
        links:
        - name: CloudBees CI
          url: https://www.cloudbees.com/products/continuous-integration
        maintainers:
        - email: support@cloudbees.com
          name: Support
        - email: sales@cloudbees.com
          name: Sales
        maturity: alpha
        provider:
          name: CloudBees
          url: https://www.cloudbees.com/products/continuous-integration
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.marketplace.redhat.com/rhm/cloudbees/cloudbees-ci-operator@sha256:bd0bd72e9e7807b0fa307de8a93a56967221fe2dd181bac02383f0f9be9120e4
        version: 3.34.1
      entries:
      - name: cloudbees-ci.v3.34.1
        version: 3.34.1
      name: alpha
    defaultChannel: alpha
    packageName: cloudbees-ci-rhmp
    provider:
      name: CloudBees
      url: https://www.cloudbees.com/products/continuous-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: bamoe-kogito-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bamoe-kogito-operator.8.0.5-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "ruleunit-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogitoinfra-sample"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "kafka.strimzi.io/v1beta2",
                    "kind": "Kafka",
                    "name": "kogito-kafka"
                  }
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator@sha256:8a99fcd217502b7e850af1e1097de58b6cb19d1e786394811a8e075c4d871279
          description: IBM BAMOE Kogito Operator for deployment and management of
            Kogito services.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["IBM Business Automation Manager
            Open Edition"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: https://www.ibm.com/mysupport
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.rhpam.kiegroup.org
            version: v1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.rhpam.kiegroup.org
            version: v1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.rhpam.kiegroup.org
            version: v1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.rhpam.kiegroup.org
            version: v1
        description: |-
          IBM BAMOE Kogito Operator is designed for deployment and management of Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Builds a IBM BAMOE Kogito service from a remote git repository and creates a custom image for the given runtime: Red Hat build of Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Deploy IBM BAMOE Kogito service from user provided image.
          * Setup Grafana Dashboard when Grafana Operator is available in cluster.
          * Create Prometheus monitor object when Prometheus operator is available in cluster.

          **IMPORTANT!** IBM BAMOE Kogito Operator may need Grafana and Prometheus Operator to configure monitoring in cases if you do not provide these services beforehand. If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the IBM BAMOE Kogito Operator.
        displayName: IBM BAMOE Kogito Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://ibm.com
        - name: Documentation
          url: https://www.ibm.com/docs/en/ibamoe
        - name: Source
          url: https://github.com/kiegroup/kogito-operator
        maintainers:
        - email: mysphelp@us.ibm.com
          name: IBM
        maturity: beta
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9492c67eaa7ab37df6f5fa623c4aac955b7d7ffaab9ae2e57dea745e82907e40
        - registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator@sha256:8a99fcd217502b7e850af1e1097de58b6cb19d1e786394811a8e075c4d871279
        version: 8.0.5-3
      entries:
      - name: bamoe-kogito-operator.8.0.5-3
        version: 8.0.5-3
      - name: bamoe-kogito-operator.8.0.5-2
        version: 8.0.5-2
      - name: bamoe-kogito-operator.8.0.5-1
        version: 8.0.5-1
      - name: bamoe-kogito-operator.8.0.4-2
        version: 8.0.4-2
      - name: bamoe-kogito-operator.8.0.4-1
        version: 8.0.4-1
      - name: bamoe-kogito-operator.8.0.3-1
        version: 8.0.3-1
      - name: bamoe-kogito-operator.8.0.2-3
        version: 8.0.2-3
      - name: bamoe-kogito-operator.8.0.2-2
        version: 8.0.2-2
      - name: bamoe-kogito-operator.8.0.2-1
        version: 8.0.2-1
      - name: bamoe-kogito-operator.8.0.1-2
        version: 8.0.1-2
      - name: bamoe-kogito-operator.8.0.1-1
        version: 8.0.1-1
      - name: bamoe-kogito-operator.8.0.0-1
        version: 8.0.0-1
      name: 8.x
    defaultChannel: 8.x
    packageName: bamoe-kogito-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dns-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dns-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSRecord",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dns-operator",
                    "app.kubernetes.io/instance": "dnsrecord-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dnsrecord",
                    "app.kubernetes.io/part-of": "dns-operator"
                  },
                  "name": "dnsrecord-sample"
                },
                "spec": {
                  "endpoints": [
                    {
                      "dnsName": "dnsrecord-simple.example.com",
                      "recordTTL": 60,
                      "recordType": "A",
                      "targets": [
                        "52.215.108.61",
                        "52.30.101.221"
                      ]
                    }
                  ],
                  "providerRef": {
                    "name": "dns-provider-creds"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/dns-operator:v0.6.0
          createdAt: "2024-09-13T09:55:39Z"
          description: A Kubernetes Operator to manage the lifecycle of DNS resources
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kuadrant/dns-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSRecord is the Schema for the dnsrecords API
            displayName: DNSRecord
            kind: DNSRecord
            name: dnsrecords.kuadrant.io
            version: v1alpha1
        description: A Kubernetes Operator to manage the lifecycle of DNS resources
        displayName: DNS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dns
        - kuadrant
        links:
        - name: DNS Operator
          url: https://github.com/Kuadrant/dns-operator
        maintainers:
        - email: mnairn@redhat.com
          name: Michael Nairn
        - email: pbrookes@redhat.com
          name: Phil Brookes
        - email: cbrookes@redhat.com
          name: Craig Brookes
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/dns-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: dns-operator.v0.6.0
        version: 0.6.0
      name: stable
    defaultChannel: stable
    packageName: dns-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://www.redhat.com
    name: sap-commerce-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sap-commerce-operator.v0.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hybris.hybris.org/v1alpha1",
                "kind": "HybrisApp",
                "metadata": {
                  "name": "hybrisapp-sample"
                },
                "spec": {
                  "aJPServicePort": 30000,
                  "apachejvmRouteName": "",
                  "baseImageName": "hybris-base",
                  "baseImageTag": "v0.2",
                  "hybrisANTTaskNames": "",
                  "podHealthyProbePath": "/platform/init",
                  "sourceRepoLocalPropertiesOverride": "",
                  "sourceRepoRef": "main",
                  "sourceRepoURL": "https://github.com/redhat-sap/sap-commerce-app-example.git",
                  "startupProbeFailureThreshold": 100,
                  "startupProbePeriodSecond": 60
                }
              },
              {
                "apiVersion": "hybris.hybris.org/v1alpha1",
                "kind": "HybrisBase",
                "metadata": {
                  "name": "hybrisbase-sample"
                },
                "spec": {
                  "URL": "http://nexus.redhat.com/nexus/service/local/repositories/releases/content/org/hybris/hybris/1.0/hybris-1.0.tar.gz",
                  "buildSourceRepo": "https://github.com/redhat-sap/sap-commerce-operator",
                  "buildSourceRepoBranch": "main",
                  "imageName": "hybris-base",
                  "imageTag": "v0.2",
                  "jdkURL": "https://github.com/SAP/SapMachine/releases/download/sapmachine-11.0.5/sapmachine-jdk-11.0.5-1.x86_64.rpm",
                  "password": "admin123",
                  "username": "admin"
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          containerImage: quay.io/redhat-sap-cop/hybris-operator:0.0.4
          description: Builds and runs SAP Commerce in OpenShift
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-sap/sap-commerce-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HybrisApp is the Schema for the hybrisapps API
            displayName: Hybris Application
            kind: HybrisApp
            name: hybrisapps.hybris.hybris.org
            version: v1alpha1
          - description: HybrisBase is the Schema for the Hybris Base API
            displayName: Hybris Base
            kind: HybrisBase
            name: hybrisbases.hybris.hybris.org
            version: v1alpha1
        description: |-
          Builds and runs SAP Commerce in OpenShift

          * The operator provides APIs to initialize SAP Commerce with the provided SAP Commerce Server. Though there are no restrictions from the operator on the base package to be used from SAP, the default SAP base image provides a server preconfigured with an application server based on Apache Tomcat. Both HSQLDB and Tomcat are included in the SAP Commerce package, and are designed to be used for testing, development, and demonstration purposes. The custom classes created for various business needs along with an external RDBMS configuration can be used in a HybrisApp creation to execute SAP Commerce in a production setup.

          * The operator also enables the use of SAP Commerce in a clustered model. A SAP Commerce Cluster is a number of individual SAP Commerce installations using a common set of data on one database. A cluster configuration lets you balance the load between the nodes to scale SAP Commerce in a way that is transparent to the user.

          * The clustering offers few communication protocols as options (UDP: both multicast or unicast, or JGroups). In order to support multicast for SAP Commerce clustering the corresponding OpenShift namespace has to be enabled for multicast usage.

          For more details on SAP Commerce consult the product documentation: [https://help.sap.com/viewer/product/SAP_COMMERCE](https://help.sap.com/viewer/product/SAP_COMMERCE)
        displayName: SAP Commerce Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sap-commerce-operator
        links:
        - name: SAP Commerce Operator
          url: https://github.com/redhat-sap/sap-commerce-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
          url: https://www.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/redhat-sap-cop/hybris-operator:0.0.4
        version: 0.0.4
      entries:
      - name: sap-commerce-operator.v0.0.4
        version: 0.0.4
      - name: sap-commerce-operator.v0.0.3
        version: 0.0.3
      - name: sap-commerce-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: sap-commerce-operator
    provider:
      name: Red Hat, Inc.
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cognitoidentityprovider-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cognitoidentityprovider-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cognitoidentityprovider-controller:1.1.1
          createdAt: "2025-09-19T20:58:49Z"
          description: AWS Cognito Identity Provider controller is a service controller
            for managing Cognito Identity Provider resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: UserPool represents the state of an AWS cognitoidentityprovider
              UserPool resource.
            displayName: UserPool
            kind: UserPool
            name: userpools.cognitoidentityprovider.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Cognito Identity Provider resources in AWS from within your Kubernetes cluster.

          **About Amazon Cognito Identity Provider**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Cognito Identity Provider
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cognitoidentityprovider
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Cognito Identity Provider Developer Resources
          url: https://aws.amazon.com/Cognito Identity Provider/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cognitoidentityprovider maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cognitoidentityprovider-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-cognitoidentityprovider-controller.v1.1.1
        version: 1.1.1
      - name: ack-cognitoidentityprovider-controller.v1.1.0
        version: 1.1.0
      - name: ack-cognitoidentityprovider-controller.v1.0.12
        version: 1.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-cognitoidentityprovider-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sumo Logic
      provider-url: https://www.sumologic.com/
    name: sumologic-kubernetes-collection-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "helm-operator.sumologic.com/v1alpha1",
                    "kind": "SumologicCollection",
                    "metadata": {
                        "name": "collection"
                    },
                    "spec": {
                        "sumologic": {
                            "accessId": "",
                            "accessKey": "",
                            "clusterName": "kubernetes",
                            "scc": {
                                "create": true
                            }
                        },
                        "kube-prometheus-stack": {
                            "prometheus-node-exporter": {
                                "service": {
                                    "port": 9200,
                                    "targetPort": 9200
                                }
                            }
                        },
                        "otellogs": {
                            "daemonset": {
                                "containers": {
                                    "otelcol": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                },
                                "initContainers": {
                                    "changeowner": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                }
                            }
                        }
                    }
                }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing,Monitoring
          containerImage: registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
          createdAt: "2024-07-19T10:48:00Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator.git
          support: Sumo Logic
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configuration for Sumo Logic Kubernetes Collection
            displayName: SumologicCollection
            kind: SumologicCollection
            name: sumologiccollections.helm-operator.sumologic.com
            version: v1alpha1
        description: Helm Operator for the Sumo Logic Kubernetes Collection Helm Chart
          v4.13.0-2
        displayName: Sumo Logic Kubernetes Collection Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - collection
        - logging
        - metrics
        - monitoring
        links:
        - name: Sumo Logic Kubernetes Collection Helm Operator
          url: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator
        maintainers:
        - email: collection@sumologic.com
          name: Sumo Logic
        maturity: alpha
        provider:
          name: Sumo Logic
          url: https://www.sumologic.com/
        relatedImages:
        - registry.connect.redhat.com/sumologic/autoinstrumentation-java@sha256:242958b7d8905f4ac02f6ad38897ba58b9badf2a6213a389e67dcc71fee3630c
        - registry.connect.redhat.com/sumologic/thanos@sha256:16faf7191668e86d8b576bba055dd0cbdeb009635976b9e07bd00f926032f11c
        - registry.connect.redhat.com/sumologic/tailing-sidecar@sha256:738d653a78dc415243ab3627db9135768175fb7f51d0803566f5d9e824c8744b
        - registry.connect.redhat.com/sumologic/opentelemetry-operator@sha256:5587fe5b7ebfa57ad320ad2521103f908317388d1741c40e5f7be5ed9e5f9ad7
        - registry.connect.redhat.com/sumologic/kube-state-metrics@sha256:3551ededa327a35db798ec8286cd843580e3ae395d2231e041e8c4b7621127ee
        - registry.connect.redhat.com/sumologic/tailing-sidecar-operator@sha256:c16b8b18e9b263fc3eb24059bba512bcfc3201b32105af6e5518392826ddd66c
        - registry.connect.redhat.com/sumologic/telegraf-operator@sha256:ff958222be01c0c3966474463e43bbd5bcf348790dd34bb761718cb1cea2d793
        - registry.connect.redhat.com/sumologic/kubernetes-tools-kubectl@sha256:d3155b864cc74f327adedf7d013a8f0fcb11c2e73a5655b7cafa848824e80172
        - registry.connect.redhat.com/sumologic/kubernetes-setup@sha256:f07eaf68fa11ffc708e4485a83bd3763fc5712f7d73d29edfe5b9c9e08f05df6
        - registry.connect.redhat.com/sumologic/sumologic-otel-collector@sha256:4554d43dba204e3ec8e91b36442a438f10234fa0cb4db7fb7880b257a6485112
        - registry.connect.redhat.com/sumologic/autoinstrumentation-nodejs@sha256:c206eef4fdecaa58b4f01e5e7f0dbd74e9606eab2b4debf47461d36191e248d1
        - registry.connect.redhat.com/sumologic/autoinstrumentation-dotnet@sha256:f4f385c4489c74b5c43f2fe6b6620a9eb8d9ecab470567236b3c45f636106640
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
        - registry.connect.redhat.com/sumologic/metrics-server@sha256:65d2a3ddc4b4183f438fa18b655079c93c905cb9e3b8c54977336205dc1fff07
        - registry.connect.redhat.com/sumologic/nginx-unprivileged@sha256:7260555d16c94c4fe1e016626b4cb8a5c0de6b700e2eea28744331fb298f71ae
        - registry.connect.redhat.com/sumologic/node-exporter@sha256:11e9e604432a66d9b388cb1bf70eac077c7f0372d1264342ef927d76b8ef0272
        - registry.connect.redhat.com/sumologic/prometheus-operator@sha256:58a329e6582456b32abebb470086191f8366b927af4703741613993465f116b8
        - registry.connect.redhat.com/sumologic/prometheus@sha256:0ef6eaf570a40499dd5de92a79593e950bc771b014e0ee6ff7419b4b82a36fcc
        - registry.connect.redhat.com/sumologic/autoinstrumentation-python@sha256:f0fa45fa45e64c7df88f21181ae18ee96fe83cd3988258663d0806ea6f39e6bb
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:e1f49fcf5b7e86d9eb49c16dbbf9484cb486f55f999fd51d7c738c3b1f2666cf
        - registry.connect.redhat.com/sumologic/telegraf@sha256:75efd2e509c6017ea0fc5c9103ce068cbc89087b16506a4ef48cb5ff00eab84e
        - registry.connect.redhat.com/sumologic/busybox@sha256:209bc8e0b1d958fc699f3bb75b1248544e372ad35cdc46c991bd1698f641e1f9
        - registry.connect.redhat.com/sumologic/prometheus-config-reloader@sha256:0b605d36782afd8f50fd631d0e9c7b19d39ce962af89b7a737b2f09a8ceff291
        version: 4.13.0-2
      entries:
      - name: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
        version: 4.13.0-2
      name: alpha
    defaultChannel: alpha
    packageName: sumologic-kubernetes-collection-helm-operator
    provider:
      name: Sumo Logic
      url: https://www.sumologic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: rh-service-binding-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-binding-operator.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ServiceBinding expresses intent to bind an operator-backed
              service with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding for Kubernetes. Describes the connection
              between a Provisioned Service and an Application Projection. Implements
              Service Binding Specifcation for Kubernetes (https://github.com/k8s-service-bindings/spec).
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: alpha
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:fe6629401bc8447da95ec55949b5a27e620364bb6e02f86230262699ac8a5d55
        version: 0.9.1
      entries:
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: beta
    - currentCSV: service-binding-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindablekinds-sample"
                }
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:e799bd06043028d214baeeffc683710ad2eed7b83272fc653e3b89f69711319b
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:e799bd06043028d214baeeffc683710ad2eed7b83272fc653e3b89f69711319b
        version: 0.11.0
      entries:
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: candidate
    - currentCSV: service-binding-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:3d7a3648e86848402577dd19c1f8fd130a5556ebdfbc0f663a72f9e141024812
        version: 0.10.0
      entries:
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: preview
    - currentCSV: service-binding-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindable-kinds"
                },
                "spec": {}
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  },
                  "type": "postgres",
                  "workload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: registry.redhat.io/ocp-tools-4/service-binding-rhel8-operator@sha256:9c8fb0c8f9d2f54685561af21e595aa88bf309eaebebdf23fb6997b75d026b18
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'.
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1alpha3
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1beta1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
          - description: Service Binding implementing community specification (https://github.com/k8s-service-bindings/spec)
              describing the connection between a backing service and an application
              workload.
            displayName: Service Binding (Spec API)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1beta1
        description: |
          The Service Binding Operator manages the data plane for applications and backing services. It enables developers to connect their
          applications to backing services (REST API, databases, event buses and others) with a consistent and predictable
          experience. The operator reads data made available by the control plane of backing services and projects the bindings
          data to applications either as accessible files mounted into the application’s container or environment variables.

          In Kubernetes, each service suggests a different way to access their secrets and developers will consume them in their
          application in a custom way. While this provides a good deal of flexibility, it also creates discrepancies and
          complexity dealing with each specific services, which leads large development teams to lose overall velocity.

          Service Binding Operator removes the error-prone manual configuration of binding information, provides service operators
          a low-touch administrative experience to provision and manage access to services and enriches development lifecycle.

          For additional details refer to the project [documentation](https://redhat-developer.github.io/service-binding-operator).

          Get Started with Service Binding Operator
          * Find out about SBO’s [prerequisites](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/installing-service-binding.html).
          * Refer to the [Quick Start Guide](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/quick-start.html) to see SBO in action in a simple scenario.

          Need Help
          * Raise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)

          Licence
          * Service Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE)

          Service Binding Operator is providing and supporting two different APIs.
          * `binding.operators.coreos.com/v1alpha1`: This API is compliant with the Service Binding specification. It is fully supported and considered as the mature API. It’s the API we recommend in production environments.
          * `servicebinding.io/v1beta1`: This API is the one defined in the Service Binding specification. It is available in Service Binding Operator 1.x.

          We’ll consider supporting only the API from the specification, once the Service Binding specification will be officially GA. Of course, that will be a breaking change and handle in a safe way for you.
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Documentation
          url: https://redhat-developer.github.io/service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4/service-binding-rhel8-operator@sha256:9c8fb0c8f9d2f54685561af21e595aa88bf309eaebebdf23fb6997b75d026b18
        version: 1.3.3
      entries:
      - name: service-binding-operator.v1.3.3
        version: 1.3.3
      - name: service-binding-operator.v1.3.1
        version: 1.3.1
      - name: service-binding-operator.v1.3.0
        version: 1.3.0
      - name: service-binding-operator.v1.2.0
        version: 1.2.0
      - name: service-binding-operator.v1.1.1
        version: 1.1.1
      - name: service-binding-operator.v1.1.0
        version: 1.1.0
      - name: service-binding-operator.v1.0.1
        version: 1.0.1
      - name: service-binding-operator.v1.0.0
        version: 1.0.0
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: rh-service-binding-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cluster-logging
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-logging.v5.9.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "name": "remote-elasticsearch",
                      "secret": {
                        "name": "elasticsearch"
                      },
                      "type": "elasticsearch",
                      "url": "remote.example.org:9200"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "audit",
                        "infrastructure"
                      ],
                      "name": "enable-default-log-store",
                      "outputRefs": [
                        "default"
                      ]
                    },
                    {
                      "inputRefs": [
                        "application"
                      ],
                      "name": "forward-to-remote",
                      "outputRefs": [
                        "remote-elasticsearch"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogging",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "collection": {
                    "type": "vector"
                  },
                  "logStore": {
                    "elasticsearch": {
                      "nodeCount": 3,
                      "redundancyPolicy": "SingleRedundancy",
                      "resources": {
                        "requests": {
                          "memory": "2Gi"
                        }
                      },
                      "storage": {
                        "size": "200G"
                      }
                    },
                    "retentionPolicy": {
                      "application": {
                        "maxAge": "7d"
                      }
                    },
                    "type": "elasticsearch"
                  },
                  "managementState": "Managed",
                  "visualization": {
                    "kibana": {
                      "replicas": 1
                    },
                    "type": "kibana"
                  }
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          console.openshift.io/plugins: '["logging-view-plugin"]'
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:54054faf0d2f375dd6d7fc7929a2dbb529508e1b3c004a56fc97c2a682d16be2
          createdAt: "2025-03-14T20:19:21Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing your aggregated logging stack.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.7.0-0 <5.9.15'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Logging (team-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.


              There are built-in input names for common log categories, and you can
              define custom inputs to do additional filtering.


              There is a built-in output name for the default openshift log store, but
              you can define your own outputs with a URL and other connection information
              to forward logs to other stores or processors, inside or outside the cluster.


              For more details see the documentation on the API fields.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.logging.openshift.io
            version: v1
          - description: A Red Hat OpenShift Logging instance. ClusterLogging is the
              Schema for the clusterloggings API
            displayName: Cluster Logging
            kind: ClusterLogging
            name: clusterloggings.logging.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates and manages the aggregated logging stack as a cluster-wide service.

          ##Features
          * **Create/Destroy**: Launch and create an aggregated logging stack to support the entire OCP cluster.
          * **Simplified Configuration**: Configure your aggregated logging cluster's structure like components and end points easily.
          * Configure log forwarding to an in-cluster, Red Hat managed log storage solution based on LokiStack or Elasticsearch

          ## Prerequisites
          ### Red Hat OpenShift Logging Namespace
          Cluster logging and the Red Hat OpenShift Logging Operator is only deployable to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - fluentd
        - logging
        - aggregated
        - efk
        - vector
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Fluentd
          url: https://www.fluentd.org/
        - name: Vector
          url: https://vector.dev/
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:54054faf0d2f375dd6d7fc7929a2dbb529508e1b3c004a56fc97c2a682d16be2
        - registry.redhat.io/openshift-logging/logging-view-plugin-rhel9@sha256:8023b072b45e9a21ddef83a56f4760d476053250d5a920d8eb5f03eecab8b1b3
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:54e3a5efaae5ecc2a08db03a421aa79e323b48c764115ea84d83a37c21d1d997
        - registry.redhat.io/openshift-logging/fluentd-rhel9@sha256:a676c9164ec2523b2002dc5eba234ce774926dbf7fe87de6d72b53f99cbf1db5
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:865572275b79462f14e84c69438a61afe9465cfe45ec164caf30965730b276f7
        version: 5.9.15
      entries:
      - name: cluster-logging.v5.9.15
        version: 5.9.15
      - name: cluster-logging.v5.9.14
        version: 5.9.14
      - name: cluster-logging.v5.9.13
        version: 5.9.13
      - name: cluster-logging.v5.9.12
        version: 5.9.12
      - name: cluster-logging.v5.9.11
        version: 5.9.11
      - name: cluster-logging.v5.9.10
        version: 5.9.10
      - name: cluster-logging.v5.9.9
        version: 5.9.9
      - name: cluster-logging.v5.9.8
        version: 5.9.8
      - name: cluster-logging.v5.9.7
        version: 5.9.7
      - name: cluster-logging.v5.9.6
        version: 5.9.6
      - name: cluster-logging.v5.9.5
        version: 5.9.5
      - name: cluster-logging.v5.9.4
        version: 5.9.4
      - name: cluster-logging.v5.9.3
        version: 5.9.3
      - name: cluster-logging.v5.9.2
        version: 5.9.2
      - name: cluster-logging.v5.9.1
        version: 5.9.1
      - name: cluster-logging.v5.9.0
        version: 5.9.0
      name: stable
    - currentCSV: cluster-logging.v5.8.22
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "name": "remote-elasticsearch",
                      "secret": {
                        "name": "elasticsearch"
                      },
                      "type": "elasticsearch",
                      "url": "remote.example.org:9200"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "audit",
                        "infrastructure"
                      ],
                      "name": "enable-default-log-store",
                      "outputRefs": [
                        "default"
                      ]
                    },
                    {
                      "inputRefs": [
                        "application"
                      ],
                      "name": "forward-to-remote",
                      "outputRefs": [
                        "remote-elasticsearch"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogging",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "collection": {
                    "type": "vector"
                  },
                  "logStore": {
                    "elasticsearch": {
                      "nodeCount": 3,
                      "redundancyPolicy": "SingleRedundancy",
                      "resources": {
                        "requests": {
                          "memory": "2Gi"
                        }
                      },
                      "storage": {
                        "size": "200G"
                      }
                    },
                    "retentionPolicy": {
                      "application": {
                        "maxAge": "7d"
                      }
                    },
                    "type": "elasticsearch"
                  },
                  "managementState": "Managed",
                  "visualization": {
                    "kibana": {
                      "replicas": 1
                    },
                    "type": "kibana"
                  }
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.12-v4.17
          console.openshift.io/plugins: '["logging-view-plugin"]'
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:8e13bd3e1654fe26c1d4d8d4a1bfb12e04cee7899cf16b5fa07b62b8f7494af5
          createdAt: "2024-10-17T17:45:03Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing your aggregated logging stack.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.22'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-5.8
          operators.operatorframework.io.bundle.channels.v1: stable-5.8
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Logging (team-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "ClusterLogForwarder is an API to configure forwarding logs.
              \n You configure forwarding by specifying a list of `pipelines`, which
              forward from a set of named inputs to a set of named outputs. \n There
              are built-in input names for common log categories, and you can define
              custom inputs to do additional filtering. \n There is a built-in output
              name for the default openshift log store, but you can define your own
              outputs with a URL and other connection information to forward logs
              to other stores or processors, inside or outside the cluster. \n For
              more details see the documentation on the API fields."
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.logging.openshift.io
            version: v1
          - description: A Red Hat OpenShift Logging instance. ClusterLogging is the
              Schema for the clusterloggings API
            displayName: Cluster Logging
            kind: ClusterLogging
            name: clusterloggings.logging.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          ### Notice for OCP 4.16 and 4.17 Only

          Logging 5.8 is supported on OCP 4.16 and 4.17 to help customers upgrade and migrate to Logging 6.0. **Logging 5.8 will be End of Life in November 2025.**  At that time, it will no longer be available or supported.

          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates and manages the aggregated logging stack as a cluster-wide service.

          ##Features
          * **Create/Destroy**: Launch and create an aggregated logging stack to support the entire OCP cluster.
          * **Simplified Configuration**: Configure your aggregated logging cluster's structure like components and end points easily.
          * Configure log forwarding to an in-cluster, Red Hat managed log storage solution based on LokiStack or Elasticsearch

          ## Prerequisites
          ### Red Hat OpenShift Logging Namespace
          Cluster logging and the Red Hat OpenShift Logging Operator is only deployable to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - fluentd
        - logging
        - aggregated
        - efk
        - vector
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Fluentd
          url: https://www.fluentd.org/
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/logging-view-plugin-rhel9@sha256:f00d4e7ce6b37f1a053c5dbe554359d844989c2243574000ada0f5bc678ffff8
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:8e13bd3e1654fe26c1d4d8d4a1bfb12e04cee7899cf16b5fa07b62b8f7494af5
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:f26518dd3dbf6b6a9e9e20e40741687996b1160c6ad8ad417989197800c2b3fe
        - registry.redhat.io/openshift-logging/fluentd-rhel9@sha256:9296e50fccae60dea580186bc5ae2a387172dc90ece35663d96c4553a2446a01
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:8f29d6eaeb4a11c9a61e3d763605e3c9c8655df7fbb4e39e5b8311d18449f04a
        version: 5.8.22
      entries:
      - name: cluster-logging.v5.8.22
        version: 5.8.22
      - name: cluster-logging.v5.8.21
        version: 5.8.21
      - name: cluster-logging.v5.8.20
        version: 5.8.20
      - name: cluster-logging.v5.8.19
        version: 5.8.19
      - name: cluster-logging.v5.8.18
        version: 5.8.18
      - name: cluster-logging.v5.8.17
        version: 5.8.17
      - name: cluster-logging.v5.8.16
        version: 5.8.16
      - name: cluster-logging.v5.8.15
        version: 5.8.15
      - name: cluster-logging.v5.8.14
        version: 5.8.14
      - name: cluster-logging.v5.8.13
        version: 5.8.13
      - name: cluster-logging.v5.8.12
        version: 5.8.12
      - name: cluster-logging.v5.8.11
        version: 5.8.11
      - name: cluster-logging.v5.8.10
        version: 5.8.10
      - name: cluster-logging.v5.8.9
        version: 5.8.9
      - name: cluster-logging.v5.8.8
        version: 5.8.8
      - name: cluster-logging.v5.8.7
        version: 5.8.7
      - name: cluster-logging.v5.8.6
        version: 5.8.6
      - name: cluster-logging.v5.8.5
        version: 5.8.5
      - name: cluster-logging.v5.8.4
        version: 5.8.4
      - name: cluster-logging.v5.8.3
        version: 5.8.3
      - name: cluster-logging.v5.8.2
        version: 5.8.2
      - name: cluster-logging.v5.8.1
        version: 5.8.1
      - name: cluster-logging.v5.8.0
        version: 5.8.0
      name: stable-5.8
    - currentCSV: cluster-logging.v5.9.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "name": "remote-elasticsearch",
                      "secret": {
                        "name": "elasticsearch"
                      },
                      "type": "elasticsearch",
                      "url": "remote.example.org:9200"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "audit",
                        "infrastructure"
                      ],
                      "name": "enable-default-log-store",
                      "outputRefs": [
                        "default"
                      ]
                    },
                    {
                      "inputRefs": [
                        "application"
                      ],
                      "name": "forward-to-remote",
                      "outputRefs": [
                        "remote-elasticsearch"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogging",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "collection": {
                    "type": "vector"
                  },
                  "logStore": {
                    "elasticsearch": {
                      "nodeCount": 3,
                      "redundancyPolicy": "SingleRedundancy",
                      "resources": {
                        "requests": {
                          "memory": "2Gi"
                        }
                      },
                      "storage": {
                        "size": "200G"
                      }
                    },
                    "retentionPolicy": {
                      "application": {
                        "maxAge": "7d"
                      }
                    },
                    "type": "elasticsearch"
                  },
                  "managementState": "Managed",
                  "visualization": {
                    "kibana": {
                      "replicas": 1
                    },
                    "type": "kibana"
                  }
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          console.openshift.io/plugins: '["logging-view-plugin"]'
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:54054faf0d2f375dd6d7fc7929a2dbb529508e1b3c004a56fc97c2a682d16be2
          createdAt: "2025-03-14T20:19:21Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing your aggregated logging stack.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.7.0-0 <5.9.15'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Logging (team-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.


              There are built-in input names for common log categories, and you can
              define custom inputs to do additional filtering.


              There is a built-in output name for the default openshift log store, but
              you can define your own outputs with a URL and other connection information
              to forward logs to other stores or processors, inside or outside the cluster.


              For more details see the documentation on the API fields.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.logging.openshift.io
            version: v1
          - description: A Red Hat OpenShift Logging instance. ClusterLogging is the
              Schema for the clusterloggings API
            displayName: Cluster Logging
            kind: ClusterLogging
            name: clusterloggings.logging.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates and manages the aggregated logging stack as a cluster-wide service.

          ##Features
          * **Create/Destroy**: Launch and create an aggregated logging stack to support the entire OCP cluster.
          * **Simplified Configuration**: Configure your aggregated logging cluster's structure like components and end points easily.
          * Configure log forwarding to an in-cluster, Red Hat managed log storage solution based on LokiStack or Elasticsearch

          ## Prerequisites
          ### Red Hat OpenShift Logging Namespace
          Cluster logging and the Red Hat OpenShift Logging Operator is only deployable to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - fluentd
        - logging
        - aggregated
        - efk
        - vector
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Fluentd
          url: https://www.fluentd.org/
        - name: Vector
          url: https://vector.dev/
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:54054faf0d2f375dd6d7fc7929a2dbb529508e1b3c004a56fc97c2a682d16be2
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:54e3a5efaae5ecc2a08db03a421aa79e323b48c764115ea84d83a37c21d1d997
        - registry.redhat.io/openshift-logging/fluentd-rhel9@sha256:a676c9164ec2523b2002dc5eba234ce774926dbf7fe87de6d72b53f99cbf1db5
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:865572275b79462f14e84c69438a61afe9465cfe45ec164caf30965730b276f7
        - registry.redhat.io/openshift-logging/logging-view-plugin-rhel9@sha256:8023b072b45e9a21ddef83a56f4760d476053250d5a920d8eb5f03eecab8b1b3
        version: 5.9.15
      entries:
      - name: cluster-logging.v5.9.15
        version: 5.9.15
      - name: cluster-logging.v5.9.14
        version: 5.9.14
      - name: cluster-logging.v5.9.13
        version: 5.9.13
      - name: cluster-logging.v5.9.12
        version: 5.9.12
      - name: cluster-logging.v5.9.11
        version: 5.9.11
      - name: cluster-logging.v5.9.10
        version: 5.9.10
      - name: cluster-logging.v5.9.9
        version: 5.9.9
      - name: cluster-logging.v5.9.8
        version: 5.9.8
      - name: cluster-logging.v5.9.7
        version: 5.9.7
      - name: cluster-logging.v5.9.6
        version: 5.9.6
      - name: cluster-logging.v5.9.5
        version: 5.9.5
      - name: cluster-logging.v5.9.4
        version: 5.9.4
      - name: cluster-logging.v5.9.3
        version: 5.9.3
      - name: cluster-logging.v5.9.2
        version: 5.9.2
      - name: cluster-logging.v5.9.1
        version: 5.9.1
      - name: cluster-logging.v5.9.0
        version: 5.9.0
      name: stable-5.9
    - currentCSV: cluster-logging.v6.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.14-v4.17
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:3e36ea9a808945d3817d169f736b1758875f2fbc6a7ae845266e1545a1b713c7
          createdAt: "2025-05-14T12:32:39Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.8.0-0 <6.0.10'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.0
          operators.operatorframework.io.bundle.channels.v1: stable-6.0
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:3e36ea9a808945d3817d169f736b1758875f2fbc6a7ae845266e1545a1b713c7
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:4e9936c090818d1d8f812487a9a8218fda5b11f8c06167fffe35bae1bafcadf9
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:15487d7b9d8f83d0df0b6fbf380f20df3518dfb1b831fac4c3209e811494b436
        version: 6.0.10
      entries:
      - name: cluster-logging.v6.0.10
        version: 6.0.10
      - name: cluster-logging.v6.0.9
        version: 6.0.9
      - name: cluster-logging.v6.0.8
        version: 6.0.8
      - name: cluster-logging.v6.0.7
        version: 6.0.7
      - name: cluster-logging.v6.0.6
        version: 6.0.6
      - name: cluster-logging.v6.0.5
        version: 6.0.5
      - name: cluster-logging.v6.0.4
        version: 6.0.4
      - name: cluster-logging.v6.0.3
        version: 6.0.3
      - name: cluster-logging.v6.0.2
        version: 6.0.2
      - name: cluster-logging.v6.0.1
        version: 6.0.1
      - name: cluster-logging.v6.0.0
        version: 6.0.0
      name: stable-6.0
    - currentCSV: cluster-logging.v6.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.15-v4.18
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f7101c962de4e5643bd7f760104a41cc48f020889159a97e2d6711703fbbec64
          createdAt: "2025-05-14T12:38:55Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.9.0-0 <6.1.8'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.1
          operators.operatorframework.io.bundle.channels.v1: stable-6.1
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f7101c962de4e5643bd7f760104a41cc48f020889159a97e2d6711703fbbec64
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:96e438dd5c080cfa47080410f375dc71da24f6d046cf609db41c8b873d85404f
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:76037c744b34b1c4fccf18f17888813d9670bb6c061c2c3de843aa43c0497170
        version: 6.1.8
      entries:
      - name: cluster-logging.v6.1.8
        version: 6.1.8
      - name: cluster-logging.v6.1.7
        version: 6.1.7
      - name: cluster-logging.v6.1.6
        version: 6.1.6
      - name: cluster-logging.v6.1.5
        version: 6.1.5
      - name: cluster-logging.v6.1.4
        version: 6.1.4
      - name: cluster-logging.v6.1.3
        version: 6.1.3
      - name: cluster-logging.v6.1.2
        version: 6.1.2
      - name: cluster-logging.v6.1.1
        version: 6.1.1
      - name: cluster-logging.v6.1.0
        version: 6.1.0
      name: stable-6.1
    - currentCSV: cluster-logging.v6.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.16-v4.19
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
          createdAt: "2025-07-29T15:25:15Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0-0 <6.2.5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.2
          operators.operatorframework.io.bundle.channels.v1: stable-6.2
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Installing logging](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html-single/logging/index#installing-logging-6-2)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:c76c853565fc93da88ba0116425839cfb86cab63cbde9856d7cca78278d7758a
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:502a5d2891049bbc32115fda28086841351cd9e6a3110b4c3c412f683f2b2491
        version: 6.2.5
      entries:
      - name: cluster-logging.v6.2.5
        version: 6.2.5
      - name: cluster-logging.v6.2.4
        version: 6.2.4
      - name: cluster-logging.v6.2.3
        version: 6.2.3
      - name: cluster-logging.v6.2.2
        version: 6.2.2
      - name: cluster-logging.v6.2.1
        version: 6.2.1
      - name: cluster-logging.v6.2.0
        version: 6.2.0
      name: stable-6.2
    defaultChannel: stable-6.2
    packageName: cluster-logging
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhacs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhacs-operator.v3.74.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
          createdAt: "2024-02-06T23:06:13.271846+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.73.0 < 3.74.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:2c0840f4573373a19eb8c9a26501138ef0e76efbd508fd69a6260e78194d86c8
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:0482d4464d6d39c01abab8df882604dd227c52b5420d36fe7414aaff8067cbec
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:71b9fdcaf4cc761b1499be8e6ccff6c463cc6b9856e9deb1f93baad45ea59eb0
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:74624389ac3261ed699b1f64e5717353da68895709d30a011bc8813b61a4a249
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:120b9891ee9db839ab0b2029a62e67ec4747c47c6ffe15af38e9035160f57b78
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:22bf00e1457b4a9e63901892aa4375a1d4a9d1b2c067c1f054d918a24b222b10
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:a108e9a7967001e0bdfed7124b01e9a3dbf1dec308444762973030aec57e1d5a
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:87eb62fe8aa7952d0d69dc13c4867ac2f51e821f56d149a9b588c826caec4c16
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d546f2d3f668ce2f20e594394ade0a86d808f6c8c88cadbba227083d4abac940
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:264102d4bca7d9bd384dc04e5bbff1a7b43343ea8e59cc79047cdda50a3f299f
        version: 3.74.9
      entries:
      - name: rhacs-operator.v3.74.9
        version: 3.74.9
      - name: rhacs-operator.v3.74.8
        version: 3.74.8
      - name: rhacs-operator.v3.74.7
        version: 3.74.7
      - name: rhacs-operator.v3.74.6
        version: 3.74.6
      - name: rhacs-operator.v3.74.5
        version: 3.74.5
      - name: rhacs-operator.v3.74.4
        version: 3.74.4
      - name: rhacs-operator.v3.74.3
        version: 3.74.3
      - name: rhacs-operator.v3.74.2
        version: 3.74.2
      - name: rhacs-operator.v3.74.1
        version: 3.74.1
      - name: rhacs-operator.v3.74.0
        version: 3.74.0
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: latest
    - currentCSV: rhacs-operator.v3.62.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:56905c2e7df0e29ca2b6301cd81b4f9bbe02b483374708f7f7da219a4b29428d
          createdAt: "2021-08-13T17:10:01.410173+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.61.0 < 3.62.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.5.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main page before deploying, or consult the RHACS
              documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create an cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead. If you want to deploy Secured Cluster Service in another namespace, additionally pass the `-n <namespace>` flag.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e30f7835cf52dbce9afdbaf55e5a93d8baf9c4b3a3ae488d1d42c1adba7f46af
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:56905c2e7df0e29ca2b6301cd81b4f9bbe02b483374708f7f7da219a4b29428d
        - registry.redhat.io/rh-acs/collector@sha256:98cfab15684506f1d0978eb701e8de478d91fc5bda3013ffd3b436aa95f07cc0
        - registry.redhat.io/rh-acs/main@sha256:9932304bb62b084ee1ae7e75bb21d6dfb737260c851f2d8daf05082f79230c66
        - registry.redhat.io/rh-acs/scanner@sha256:266a575c2547a3223890e8b73db0f14d69dc5a22217c7e4e3dc79644e6ab0258
        - registry.redhat.io/rh-acs/scanner-db@sha256:ff877f503e08ecca0ce7a7d8db760b2e4004e8e8ae2f27fb9b9c00a05b58c93a
        - registry.redhat.io/rh-acs/collector@sha256:558c0dabdc52efaee7bc889ef397ac2ac30f7c54562c2cc0070d0ef991a8b31b
        version: 3.62.1
      entries:
      - name: rhacs-operator.v3.62.1
        version: 3.62.1
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.62
    - currentCSV: rhacs-operator.v3.64.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fa478167f2f738f244bed6a06450679be73f8464c40f86acde4ffd390505692b
          createdAt: "2021-09-22T05:10:01.286657+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.63.0 < 3.64.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0cacd00c0a147cc8295138565ef2dc2ddbd9ad1fb25f6be817719a75ea4c6448
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fa478167f2f738f244bed6a06450679be73f8464c40f86acde4ffd390505692b
        - registry.redhat.io/rh-acs/scanner@sha256:8b7121bce8f9c7133d29245ebb265a4b8f9831a66a57f11b2d68b9b8b5cf79cc
        - registry.redhat.io/rh-acs/scanner-db@sha256:67e81bf896267b7acab8cb6755cde552173de4f0881c38eaca8e491bd926cc53
        - registry.redhat.io/rh-acs/collector@sha256:2b3a8bd3af62516a6714fe612e5fdc0433b91d873182e290d3ddc2b32ce291f0
        - registry.redhat.io/rh-acs/collector@sha256:3eb099da081cf208d46be5ccfa7b5e9ac0e2cad49f127fa4db91e49e3f26db42
        - registry.redhat.io/rh-acs/main@sha256:6aeb2223124540ac217c38b016dea99d770ab897dde4906a821d8776d91ea03d
        version: 3.64.2
      entries:
      - name: rhacs-operator.v3.64.2
        version: 3.64.2
      - name: rhacs-operator.v3.64.1
        version: 3.64.1
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.64
    - currentCSV: rhacs-operator.v3.65.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:229c7e5f967c70dfbc3cf638d521267ea6acf875b055499dfd30cb19c0b1f69c
          createdAt: "2021-09-21T13:26:28.711610+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.64.0 < 3.65.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/collector@sha256:3eb099da081cf208d46be5ccfa7b5e9ac0e2cad49f127fa4db91e49e3f26db42
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0cacd00c0a147cc8295138565ef2dc2ddbd9ad1fb25f6be817719a75ea4c6448
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:229c7e5f967c70dfbc3cf638d521267ea6acf875b055499dfd30cb19c0b1f69c
        - registry.redhat.io/rh-acs/main@sha256:e672ef9a1239065f8cf0f331757861fea082e971a9e9760d1e65795f399cbde1
        - registry.redhat.io/rh-acs/scanner@sha256:bfe9b2205bfca6aebed14827b4bda098bfad574e67ffa967ad339543062200fe
        - registry.redhat.io/rh-acs/scanner-db@sha256:20f258500ac2ba4e626df92dbcd193cf7ab8687ac251591be2442b8a46871835
        - registry.redhat.io/rh-acs/collector@sha256:2b3a8bd3af62516a6714fe612e5fdc0433b91d873182e290d3ddc2b32ce291f0
        version: 3.65.1
      entries:
      - name: rhacs-operator.v3.65.1
        version: 3.65.1
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.65
    - currentCSV: rhacs-operator.v3.66.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:10cdb0ff3e67b65737cde420536494bd84c2b7ccf65192bbed2eb43eaed66a41
          createdAt: "2021-10-20T16:38:38.091764+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.65.0 < 3.66.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bd74f0b95ff2881faf6eb4520feecd067b69a759f09bbfecd467447ce2a2a25e
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:10cdb0ff3e67b65737cde420536494bd84c2b7ccf65192bbed2eb43eaed66a41
        - registry.redhat.io/rh-acs/main@sha256:97230c01e76d90c7fe35ffeaa591be3d9175f350a8fd88599aded09a1edf930c
        - registry.redhat.io/rh-acs/scanner@sha256:323ee86d6a9cbeb32270c659c1154f3a62c36535c6ea894e4e29e93eb1f4a84f
        - registry.redhat.io/rh-acs/scanner-db@sha256:490b0bf49161e4bc6de2504726aaf360e0ca9a214aea6b886fff13153bdb263d
        - registry.redhat.io/rh-acs/collector@sha256:5673c27554a71502e06e08945a4424c6ce5da27f489854f3aa452e6d31240aee
        - registry.redhat.io/rh-acs/collector@sha256:d93e13ee1d1cc4b7f82e644a7d43e560aac42bd1e06d25b476cea9a4338bf9ce
        version: 3.66.1
      entries:
      - name: rhacs-operator.v3.66.1
        version: 3.66.1
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.66
    - currentCSV: rhacs-operator.v3.67.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:893ebcb2d0c96647e56bce3b16fe32851df59cd71f2c0fc8f6e66ca3a669bb70
          createdAt: "2021-12-16T12:54:10.117625+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.66.0 < 3.67.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/scanner@sha256:2860ed271da6064149bb53568c5ad60e129918d55a2f244f9ade4a0bea4c4933
        - registry.redhat.io/rh-acs/scanner-db@sha256:6f8d47277732eb56317d576c2d06fb1f539d46d16520e6ca34abec290392cef9
        - registry.redhat.io/rh-acs/collector@sha256:57c2dfa3dbc8368076291fbc627baa01dcb5ebc50f838e02ce7ac38a128d027b
        - registry.redhat.io/rh-acs/collector@sha256:e4648723ccbd43fa5acf005d9734bfc06d41e19e1be0bde839880e28fa361e16
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86e5fa1fa294987114be200890c2e516501e424aee0fb98ece25c95e7716295b
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:893ebcb2d0c96647e56bce3b16fe32851df59cd71f2c0fc8f6e66ca3a669bb70
        - registry.redhat.io/rh-acs/main@sha256:8db7630d094b661411ebb3f35248848d3147c735677d0922f7a04de9a41eacc9
        version: 3.67.2
      entries:
      - name: rhacs-operator.v3.67.2
        version: 3.67.2
      - name: rhacs-operator.v3.67.1
        version: 3.67.1
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.67
    - currentCSV: rhacs-operator.v3.68.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:195a694dc96096166fe82e81ecb3ceb493f96cfb5029112e576e44ab39b56abd
          createdAt: "2022-06-15T22:13:22.427644+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.67.0 < 3.68.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:195a694dc96096166fe82e81ecb3ceb493f96cfb5029112e576e44ab39b56abd
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:f0d2f5f11c2a96d34d6b4baf249cbf821fe064eade73c3f95c22f3c6981ac4d1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:105d10bd50a6062c998ed107dea388dd78f0caddb0b57713c8692b1d07042774
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:081f2198ad70bfb2ec66153b50b48b950b9a5550ba996de413c5a2aff4052783
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c9d0a5a2871c24c0b1058e419f2b090429363a319856397855412fd2c0037395
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:9b9b9146d9683d6da0ff2abbc8ea82b96d945b7a8c3740d72d8c018fb2b04fba
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:0bb76363a3a25916dd4493e177e9b2d36eac9633473b07fe0fffb853fd072da3
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:711293438bab35b17d2e1f162e7f74f667a6e16dd32e45122f917b6817fefb4e
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:b5fc3dcaa72bbc8dea005189c940e1cf2d00c94a6cf7676cf0e1790e1191d45c
        version: 3.68.2
      entries:
      - name: rhacs-operator.v3.68.2
        version: 3.68.2
      - name: rhacs-operator.v3.68.1
        version: 3.68.1
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.68
    - currentCSV: rhacs-operator.v3.69.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9b2beceb255124855ba1b352d16009b2f2cd422901a19c311f0783dab1fcc3a0
          createdAt: "2022-06-21T09:07:40.705944+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.68.0 < 3.69.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:08a7d8c892fad69c589b81ad8fbd21d36c8fa4078530dba361cfa9a8f9f05950
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:9ae03669cfdc0f4fe56b021d911edb17159c07757720ae30a1f22d7f7a897c07
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:f167e928740d1c614e26948ea58f49e569f6a4fddc438cd0abe52f60538432d0
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:867613178d491f416e57975843817251490c5fbca2fc69d8b55b169bea675279
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:8db5206532e95a7ad7cb23f389155d41757aaf5f445ddce92b008a06627d3fd5
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:77f5db63e17b140a8d6b84e3822fd77463035be9c0c832b55b4977696da13ccc
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:da6dd486a20fdba61ea12d2c77ea299d9795a9f22d9057d65ff992b1f5896e07
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:b254746ed3faa1f1cbcafdf1343a1957bf0214311b36f4e892be3000cc2aecf3
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9b2beceb255124855ba1b352d16009b2f2cd422901a19c311f0783dab1fcc3a0
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:2afb2823edc6f3434e18db60ee37381b5b79f51234c1a261b6d0f8ed74cc8b63
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:cf5629823f2dd583495bcbbcf3b83b4f29f855ace3c388f5cb7b24a6d34ed267
        version: 3.69.2
      entries:
      - name: rhacs-operator.v3.69.2
        version: 3.69.2
      - name: rhacs-operator.v3.69.1
        version: 3.69.1
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.69
    - currentCSV: rhacs-operator.v3.70.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fb903bf9e110ba03e9c63977ebaf707e6884dd8fd413bcd6b793a0ed795cdd0e
          createdAt: "2022-06-20T12:30:26.999648+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.69.0 < 3.70.1'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:dfeb239931f7a33ade5c829a102cabc01f5c348e1aa35a7ddc24757cfd549629
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:df8cfdaa40b50ff2f2670d97c2115b47ae4bea443ae2831caaa468dd62f3c75f
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fb903bf9e110ba03e9c63977ebaf707e6884dd8fd413bcd6b793a0ed795cdd0e
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:6c32b86ed8bc18e384706eb6188406e66910bf34c8869bd155a60c3525d0fea7
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:d2e827a209840b7197960c8115720972237532eb9cf9d6fee7a36107053408ce
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:3d7c8b77671740e324ee6c309a84bbaa4d925e6750979fe4e46c561bc79f1b4d
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:b47c96af1c9ffe4d8728da510a93efe21adc221225fe2b822cfcace6c6672ba2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:08a7d8c892fad69c589b81ad8fbd21d36c8fa4078530dba361cfa9a8f9f05950
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:ee55a233d2a8fd77d89ec2c69f357721c7debead6470a9f494b3cd6a001829c4
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:6b14e850cda9a54618847858292ea669abff599a40692e27fed0e337c4a60526
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:e71f2f2347dd46dec790405b1bb98d54a348a4583949f142f5703be9a790f96e
        version: 3.70.1
      entries:
      - name: rhacs-operator.v3.70.1
        version: 3.70.1
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.70
    - currentCSV: rhacs-operator.v3.71.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:469b6c1592226eb9a521131ef758a33e6a6ac7c96bc67fcc903cea30469c125e
          createdAt: "2022-12-07T08:31:36.777654+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.70.0 < 3.71.3'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4a0dc8c337299d7f6440ff75c2d5219d3ab68183f4f117f008ef065757dd5d4b
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:469b6c1592226eb9a521131ef758a33e6a6ac7c96bc67fcc903cea30469c125e
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:f16f92e074f7125b88e9e2033c7c12b44421145b93d040034fa4688626869af1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:c3ec88cd99908894d0e67153779062f90ac1b007980d66040a79b2be13798a10
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:4d8bac9b3857af6ba351c7768f70ef30ef7cb99ae4df60cb226b5a793353a052
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:7f279bb6e0f34a5208d8711ba4d7b62c78f08b683815ce33f6aa28c23c4246da
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:9bc25e81c21cda6d8b535d8ac67c1286fe57a1596001850e730f8f73bf371111
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:e282a8e5e606003ea344d44ac5ee7699563848f7160eeae183aa58ad740e0e94
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:cb2d0f470968c2b725b0d176d423cc1866fd56893344d077d28f2c16581d3614
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:df7b41999b542a477848823a1d0e7e5c29ae1b00fbac035007a5a1f33fac139f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:5fdfd3690af60bd5ecf04741cd1f73259491da501700b066120976960567f0db
        version: 3.71.3
      entries:
      - name: rhacs-operator.v3.71.3
        version: 3.71.3
      - name: rhacs-operator.v3.71.2
        version: 3.71.2
      - name: rhacs-operator.v3.71.1
        version: 3.71.1
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.71
    - currentCSV: rhacs-operator.v3.72.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:3ff61e1cc77b66a22c8a8c0ab306f0e122cd74e762005285aede23ebb6c2f6a2
          createdAt: "2023-02-28T17:02:08.443613+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.71.0 < 3.72.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:575b6c3e2914751df1668ffff83cbda187ce6e56f32214414cdf036b7a62bbd8
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d56a00b3e20e60b7169f5c35d03fa6a1bd3b9980dd3ebc9d89d759e1d6d20ad2
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:cd475a6c2356d63cfe736a11114de7fedfbd8fdcfecc56224a00b6c819103be6
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:42e5f6c4ba789f124ac3ebe7b6d1a607ede8e763a35888c97ccbd0259bec4a1c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:648b14e4158da2a85eedb46b3456fb502d680d52d375ad34bf45b4b94f8b2ef6
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:3ff61e1cc77b66a22c8a8c0ab306f0e122cd74e762005285aede23ebb6c2f6a2
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:cd0c205ee44eddcdf99b84e80fcf4403851d956efc50532b9d12013bafe44988
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:0b69595405e4e598183bbf4197ac5b7593724fa6162413f225aa01a542af18c9
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:622f30b7f591194d86a5bf95deede0225e66f7b624db197f6289a5fbbbe887a9
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:c07fae3616cbc681e4d31839abea6623b4bf4dade93273201c8db7ba0d635aad
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:cb5df858e0c5cae8f77b285e23daeb83062bd72570c90ed5ea3ce4a02d2f5f08
        version: 3.72.4
      entries:
      - name: rhacs-operator.v3.72.4
        version: 3.72.4
      - name: rhacs-operator.v3.72.3
        version: 3.72.3
      - name: rhacs-operator.v3.72.2
        version: 3.72.2
      - name: rhacs-operator.v3.72.1
        version: 3.72.1
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.72
    - currentCSV: rhacs-operator.v3.73.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a8486b66f977a639d77ef110492e5ecdde47be8acaefbaa1cbb5417085c56839
          createdAt: "2023-05-26T21:27:04.033276+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.72.0 < 3.73.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:7bf9415aa4b0746d11ca5a59d4179bc382a0fac4ded969f0c4294171fde27d58
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:e6687c12d070069f655c17832d94c619c18274a259ccc59715036f9908e162f7
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:ee2a9e4ab74f5a554d1e4ff148833a53bdaab70ac1e2668cc9ee2fceca290ba9
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:082bd6f4b40209e0591be001dea358f328cd86a08691cdceb7f71ea59d861d9d
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:4d0ead8969f120a629104a92a454842bb7103aec137deb6c8209b7db9c48a1b2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:648b14e4158da2a85eedb46b3456fb502d680d52d375ad34bf45b4b94f8b2ef6
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a8486b66f977a639d77ef110492e5ecdde47be8acaefbaa1cbb5417085c56839
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:e2c89f735969d438f75a89ff63c404a8ec4361a03fde48e880fee73006b9c669
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:e4a27d3ffb173c87728fb637ca7fde92a9e78ea9ee3edcecccc2a5e3d3748b8c
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:dce5db4f162dd40c09cad114cb802cd6d7c72d0c6986581f2e74c977493f75b0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:c020a80ff82bf4d9b5cec3eeac5de9990de76041e22c444e664a20e15c46ce3b
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:5df359156b01505ce43e5f14c3b08722856bdb86f3d0d9b4fba785a2e580f346
        version: 3.73.5
      entries:
      - name: rhacs-operator.v3.73.5
        version: 3.73.5
      - name: rhacs-operator.v3.73.4
        version: 3.73.4
      - name: rhacs-operator.v3.73.3
        version: 3.73.3
      - name: rhacs-operator.v3.73.2
        version: 3.73.2
      - name: rhacs-operator.v3.73.1
        version: 3.73.1
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.73
    - currentCSV: rhacs-operator.v3.74.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
          createdAt: "2024-02-06T23:06:13.271846+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.73.0 < 3.74.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d546f2d3f668ce2f20e594394ade0a86d808f6c8c88cadbba227083d4abac940
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:264102d4bca7d9bd384dc04e5bbff1a7b43343ea8e59cc79047cdda50a3f299f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:0482d4464d6d39c01abab8df882604dd227c52b5420d36fe7414aaff8067cbec
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:2c0840f4573373a19eb8c9a26501138ef0e76efbd508fd69a6260e78194d86c8
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:74624389ac3261ed699b1f64e5717353da68895709d30a011bc8813b61a4a249
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:87eb62fe8aa7952d0d69dc13c4867ac2f51e821f56d149a9b588c826caec4c16
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:120b9891ee9db839ab0b2029a62e67ec4747c47c6ffe15af38e9035160f57b78
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:22bf00e1457b4a9e63901892aa4375a1d4a9d1b2c067c1f054d918a24b222b10
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:a108e9a7967001e0bdfed7124b01e9a3dbf1dec308444762973030aec57e1d5a
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:71b9fdcaf4cc761b1499be8e6ccff6c463cc6b9856e9deb1f93baad45ea59eb0
        version: 3.74.9
      entries:
      - name: rhacs-operator.v3.74.9
        version: 3.74.9
      - name: rhacs-operator.v3.74.8
        version: 3.74.8
      - name: rhacs-operator.v3.74.7
        version: 3.74.7
      - name: rhacs-operator.v3.74.6
        version: 3.74.6
      - name: rhacs-operator.v3.74.5
        version: 3.74.5
      - name: rhacs-operator.v3.74.4
        version: 3.74.4
      - name: rhacs-operator.v3.74.3
        version: 3.74.3
      - name: rhacs-operator.v3.74.2
        version: 3.74.2
      - name: rhacs-operator.v3.74.1
        version: 3.74.1
      - name: rhacs-operator.v3.74.0
        version: 3.74.0
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.74
    - currentCSV: rhacs-operator.v4.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9bf3a3e0fd573f91f03794eaf5291364a90392e5e7d4b5d4ed550ef4600ab826
          createdAt: "2023-10-20T07:04:50.531740+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.74.0 < 4.0.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:08f537cba262605a0eb0a510b6faf78e3ba26465f636d03cc1f7a6ed6b40d355
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:eee1aa054d4d15b23be293f31d3ea659e0ffc1d05174ef89a6c977dd3d4377cc
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:4c848eb7465e928251e05d32ffbc93cd14e7bb77972d22b8721bd6eaa4a93d2f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:3250d89e78de852a2627593f29469f0b3ff639ea3644b4b7d6d057d0c1ad3659
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:5a492b856e0b6d9fdb020b556d13186ef5e7095b0c736a2031500c9c00dad0c1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:139d80d398e7e5cfc85839e3039a3630b48c7ae6650d755754430fccd94d6a27
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0aedd8e865f72ffa8b79fd2bab48974caec3eea3e3a6f77c89fe6dacb8947057
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9bf3a3e0fd573f91f03794eaf5291364a90392e5e7d4b5d4ed550ef4600ab826
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:cfa10bb6b461f302674f185b26c8d6efaaeba7f3f65d2cf05ae1f069aa2710d3
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:c39621837cbc8419d187a43a9575ec8cb4021c58ea1b56f215003d7264eb5d03
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:09e79b5db8977520f9951cc43c339c8e40b4659d7da513347d5a88b2500f809d
        version: 4.0.5
      entries:
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.0
    - currentCSV: rhacs-operator.v4.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:8ffb6f0ff1bf016c3919c1eefc6c3db3d4a8fbe0e79919f72b4c281afe6d4b1c
          createdAt: "2024-01-13T18:55:22.011709+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 4.0.0 < 4.1.6'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:364c107c34694ca33ac724de78db3e41ee55f9098d9ade0775a51b11f1533065
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:8ffb6f0ff1bf016c3919c1eefc6c3db3d4a8fbe0e79919f72b4c281afe6d4b1c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a8ad0e76ce4cae9f9b36aecf125c4ed56fc2d776fe3df113d41bfc1c2f8b3602
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:8547b76cf0533dfe6da2aafbbd063021bad4ff78b32b2dfed4a7607496e0a904
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:8760d6dc68c86f0c5b53d8f0bfada0df012f80bc8e0023ba3522fe6829a05a41
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:762ad216e79be0837e98873202b0072c7fd399925ee94bde57cc61c68cbd5f4e
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:01b9f9798d3ecbefe3da1ac2ac782e8e8b930e7eeceb82fe3473b5bfc1ccd40a
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:7ac504b7824d17bd7ac55b1e59c26be308506f315a0ab0d8c45d3b4934c0a16c
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:2ec62f3da93a1346ad215744257820377ccf4ead84e7471c52e348a31a431d3c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:bdbe08ef2d01b114702780fab76905e6eb1851684692ffa0a4c84604bda3ddb3
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:34694b9f4fe7fffc0c547be4e311579b5bb0c1fa92d8257bf6d1581ffd655c46
        version: 4.1.6
      entries:
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.1
    - currentCSV: rhacs-operator.v4.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4253eca84cbb0c5fc033507ddc2deef235b054c0ba944fa271ebbba21545b872
          createdAt: "2024-03-13T14:51:55.084122+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.1.0 < 4.2.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:b5a15f2985b114bf9bc24a0d9942db5d0cb818e60fd15a2f9094aa00904cce77
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4253eca84cbb0c5fc033507ddc2deef235b054c0ba944fa271ebbba21545b872
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:12db245af3b8475c9b5680bb0035c566109819e05cfe62bd1dfdfb5d6ed25bea
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:b82f7f654235688fc07fb037422a73e1106e27024d39a68ef7fac1422b69496d
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:ff27f03064633d6d08f8383649069e39c868def4aeca8f2622750397e049d0ec
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:2f326cabf03f780a3e6120995d5442641764641f62430d82e05edbe21b30c7b2
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:3858ab936d4626a11bbf895d41bad181364ee5df3312f6dd29ebeb9b7d31ec77
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:e2b07cb7d690477d78a2b8e1875545998d4bdb946f80dd0533a14187da99dbbf
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:59f1862c8ce188906499f08201ae05deeb47bd4fc5eab4be0386ca44ec065da8
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:e79a5e5681ebfe2d6349e2763ee0c1fa521b20319846958e402d23229b1f257b
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:4b35f5f39865238a8e95d3cba7ebb5f15ccfbf6a2563c34f57d91a0b5c927066
        version: 4.2.5
      entries:
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.2
    - currentCSV: rhacs-operator.v4.3.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:266cbbc270efafa26d5d15f3e333c96eaedee3b47854ff337ae1c572d34901ce
          createdAt: "2024-06-06T22:31:19.365878+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.2.0 < 4.3.8'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c36a557c9dc2ae4479c432933a9a4dbef38f2be53360a68a5b6befb614d01dac
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:b0c9b3858fc0547dcb4eb3065dedd5bbcf1e1f6707a8ab812ed06882d4f6d6ad
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:b7eb17fbcf6c4a6e4205ad66f1ffd9b10a4634f6ef5045d407c266566b1f255d
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:bdf9785bff51194636c6e8a518448c6d12a297bd31c4bb2e79d07084409c18b4
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:266cbbc270efafa26d5d15f3e333c96eaedee3b47854ff337ae1c572d34901ce
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:9fba889021a876ab627ff8c496e4b6c46c221c61142cd0c76daf33c89762e2c1
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:10d0cc82b6eddd089f3a7af37b872c33248513d0b4d3f094b6f62c01f4460825
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:6b55a25eb3a611c648d9be883828834592229c16f4166fd28ef40a25fc667061
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:2c35cb6f518393d002edd2bf5f714c1d053cd0fab55f6152a2f35676e4979202
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:2bb643651c8d51525223df3b2b675947174bbe9f1eaea8688e1fa83044cfa41f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:5092c3fbb4a38c06a2f42d125dc718ef0e1005f304ad8046afdd2a5e7ccbc580
        version: 4.3.8
      entries:
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.3
    - currentCSV: rhacs-operator.v4.4.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:2b87569d7f4ef5e06c989b273c378e007c6d87b26eec41774ac1fc67dcd9e4e4
          createdAt: "2025-02-10T19:50:13.597164+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.3.0 < 4.4.8'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:65707fdad55767ff5598a8d3679fd0779a37f59501e58971b81bebcb23f1f9fc
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:d6b4f8caf342b105c8df14c5c9180cb36123e08d34756ee8fe4f2f76c34f8064
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:664a5dc40dc1fa06c47a11e13b0ab5e0612a981d7b69bcfb29422abf3d1a13b2
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:39cfd01236e905034f6645a73063427436c14e351e0e0bc735e73e62aa5a1c3c
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:467d069ee957aea07882b83b60ae9fcc62ccca159f5d643aac13e531ccf7df88
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:2b87569d7f4ef5e06c989b273c378e007c6d87b26eec41774ac1fc67dcd9e4e4
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:0df4e95c7406aee7812d145cdd4d8c61cec32857fd2079dbc4244954a9d8288a
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:183c75920ab8e7cb947cb966f1543e39b6fdb5e1bb0caa1b0cf5f89671f0328f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:76f9dc2e2ee5595c36530d97dec216e2dbefada2fe0747e3f6a3a725b75722ab
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:763e7568704971cb4656727dc223374de13e82b4304dd3f5387ddc0f4e42b8b6
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:69c12230a785619c765e3749a7d56f76dce9ce4f530c20e8a9387b99eb65627b
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:ba6e25a413c79131b966a097f72c36478933ea52deb55a62acda7011a96b3983
        version: 4.4.8
      entries:
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.4
    - currentCSV: rhacs-operator.v4.5.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a7dfbe3590668f91fe6694335b5a35b2fca07b4735348cc5202921755d2f654d
          createdAt: "2025-04-09T22:51:15.157672+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.4.0 < 4.5.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:5914548f6ce2160e92c5d96580ffba23cdf1ae940c1f73f3f6424d15c6e97c37
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:dc7f23341c8b209f1846f29f889d00ef232fb58da9d5b4bd561f3933093b2786
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:d928600f75461465598a10de8df2495a275fd5232a92b078f609d1f5774eb778
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:1817cf8d2c08b04391792321702ef53d16a4b1ca06459c00c9c053128f534358
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:c2656b794ec4236e36f48f71bf26bdedd9de27fc4acef39deba2bdcfcddab03b
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:cc4b329cc795752ecbb178d2e4efc152bde54b04181334eb5bd70304c3354cad
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:0ff948b5517df3773cae2df02b708c348d8707409899b1de4108588e31601c0c
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a7dfbe3590668f91fe6694335b5a35b2fca07b4735348cc5202921755d2f654d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:5109b304eb9458f97af85c2ba63e08f51672fa3f9406226b7f357b86b5828dc0
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:fde78d34ecd2944c7c16b4e823c76b3667bf6c5b468e9991128ecc4272ec0e2e
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:ed33e7396357f1db03a065573c020846a5872b12a21284439770686689222434
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:f07cba0738897e0e7ae74971be0470ce2a855657f133a6e1cdd8a7092f6151b0
        version: 4.5.9
      entries:
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.5
    - currentCSV: rhacs-operator.v4.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:c4d434e1ff052508e5594d2b95dafa8726fd8adf5317754774c013a46de224d8
          createdAt: "2025-09-26T09:24:28.046251+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.5.0 < 4.6.10'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:b5cd08349931669055f03d21331499ef2ea04597f8a90fb201615d43000271bc
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:ffe41363dc00b05b0daa0c6d16efb30e534086aa4e3d7ccc599c86d33813138c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:713e61d3c8b55de1edc9d7834d3e8576718e1327665f94a6b58178fa7496c0cc
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:65793c05aec8085a53fab8bcf33a85192c17b46c91a7d86e0a457d1e3a113878
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:899da9c23000bb821bc9d544ebb56bcda4f3245ceee5b1b9c12d09f580b4b9a1
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:8fad8ab2efdcbc99d9efa736a90cca6f236f0a9d45fdb355ae0b6b4dc42d77f0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:90c47c6f95571cfcca04aa4e05a18befad75c1b44a6f3a055fcdfdab749d29c5
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:16b61ca1b9ee4e0821e7fc82fe4def75ed22950e502730f5704ec4fe63bb2e59
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:c4d434e1ff052508e5594d2b95dafa8726fd8adf5317754774c013a46de224d8
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:b630d53ad0b862151be7e14f0483ca5adac570216deae66f60e75af87a4dfd46
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:19db8c5d2425dcf18ae63b025f946fc46adaec0578e9505415cb585797069e95
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:6da59c5b82625894f6efa3fa0bc5ea0378edd4f95cf42dee772e9c8a565dba81
        version: 4.6.10
      entries:
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.6
    - currentCSV: rhacs-operator.v4.7.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:5b60d8271d6fa70df3c8f7e0ecd0d64feecfdf674dc84036500fe5f145c9415a
          createdAt: "2025-09-26T11:05:05.900682+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.6.0 < 4.7.7'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:def71e1ddea5b8c1e54c8582f39dc9ecca855b38aa9423690150aa6ef0ff189f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8882c00c9be9b579654a9ee8cd0c38537cf92d2f4bd8ebded1678aa0b6ae827
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:34ea97c6e9fc2d23427d0c6f2628075c0eaa0e151ca1c381539b039c249217f7
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:98553e6f18d3c03a29212b4f7ea1d30a42bcf8b52ca12aecbcbc47769fc35b53
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:bac79889d4717233e89bb6c8fae6732ed7bec5b9a55fd3b3de1f1311d870d2b1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:6bb454d64a9e339b1aeb81ca17b26f33db4a84f19ced4ad81858e40fd01878dc
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:5b60d8271d6fa70df3c8f7e0ecd0d64feecfdf674dc84036500fe5f145c9415a
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:7596529cbdce7118226b79d69e0667ab35ee692569e4014b74abdf0fa7415222
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:5ed9902b52830be93436d7977c10e73d346b6290532f6d7245cfe7cca9c93b31
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:0c32318ee4cbd547e2c1fd64433cd23a629d1b06523d7c0f07ddb98c2baaa708
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:d8fcfe1267495801d23a0b94101c7b4d06b48346a949c97946ebcbffe148f4b6
        version: 4.7.7
      entries:
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.7
    - currentCSV: rhacs-operator.v4.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
          createdAt: "2025-09-12T11:06:11.361461+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.7.0 < 4.8.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:adf47cb6615d344533fcca0aee41774f5e8df9b952b9b89a57c2be2b7e0fc03c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:92ff7691963e620dc5bf6da0a7694e6bcd68c2545afc1642a59a443ba2ab5611
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:db691065e46f604b980b0b202f6c2e14c8b204094d784c4a15db1e4fecb2965e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:97df59280c12727a99fdec0c9bf56eb874e6a1cacd12d849709797d4be635ce9
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:e5c200f28187c1f6971562e41cae969183ee15f3d0feb076bb202b22ff67cbcb
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8964b1441251607b23f92918e9158724174013dc4af423988f56d6098dfb2d5
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:e949830f3034c41aa5434a3654b2b1ccba7dac0373bf43d7c50a9ea43bb686fb
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:16912df2c26e5fc2e6051c0501170722856db274415063608de716cf53851038
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:d1758b76c32f03b5a1a44951d18fb2bb0c4fd0a99d24f8c42fe9e6b059574889
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0c92d5b5508032d77559606927e45686179eaafc59e16d5496e088f2f212d258
        version: 4.8.4
      entries:
      - name: rhacs-operator.v4.8.4
        version: 4.8.4
      - name: rhacs-operator.v4.8.3
        version: 4.8.3
      - name: rhacs-operator.v4.8.2
        version: 4.8.2
      - name: rhacs-operator.v4.8.1
        version: 4.8.1
      - name: rhacs-operator.v4.8.0
        version: 4.8.0
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.8
    - currentCSV: rhacs-operator.v4.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
          createdAt: "2025-09-12T11:06:11.361461+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.7.0 < 4.8.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:e5c200f28187c1f6971562e41cae969183ee15f3d0feb076bb202b22ff67cbcb
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:e949830f3034c41aa5434a3654b2b1ccba7dac0373bf43d7c50a9ea43bb686fb
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:16912df2c26e5fc2e6051c0501170722856db274415063608de716cf53851038
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:adf47cb6615d344533fcca0aee41774f5e8df9b952b9b89a57c2be2b7e0fc03c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:97df59280c12727a99fdec0c9bf56eb874e6a1cacd12d849709797d4be635ce9
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:d1758b76c32f03b5a1a44951d18fb2bb0c4fd0a99d24f8c42fe9e6b059574889
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8964b1441251607b23f92918e9158724174013dc4af423988f56d6098dfb2d5
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0c92d5b5508032d77559606927e45686179eaafc59e16d5496e088f2f212d258
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:db691065e46f604b980b0b202f6c2e14c8b204094d784c4a15db1e4fecb2965e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:92ff7691963e620dc5bf6da0a7694e6bcd68c2545afc1642a59a443ba2ab5611
        version: 4.8.4
      entries:
      - name: rhacs-operator.v4.8.4
        version: 4.8.4
      - name: rhacs-operator.v4.8.3
        version: 4.8.3
      - name: rhacs-operator.v4.8.2
        version: 4.8.2
      - name: rhacs-operator.v4.8.1
        version: 4.8.1
      - name: rhacs-operator.v4.8.0
        version: 4.8.0
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: stable
    defaultChannel: stable
    packageName: rhacs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-apigateway-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-apigateway-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Resource",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "RestAPI",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Stage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "VPCLink",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/apigateway-controller:1.3.1
          createdAt: "2025-09-19T21:13:12Z"
          description: AWS API Gateway controller is a service controller for managing
            API Gateway resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIIntegrationResponse represents the state of an AWS apigateway
              APIIntegrationResponse resource.
            displayName: APIIntegrationResponse
            kind: APIIntegrationResponse
            name: apiintegrationresponses.apigateway.services.k8s.aws
            version: v1alpha1
          - description: APIKey represents the state of an AWS apigateway APIKey resource.
            displayName: APIKey
            kind: APIKey
            name: apikeys.apigateway.services.k8s.aws
            version: v1alpha1
          - description: APIMethodResponse represents the state of an AWS apigateway
              APIMethodResponse resource.
            displayName: APIMethodResponse
            kind: APIMethodResponse
            name: apimethodresponses.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Authorizer represents the state of an AWS apigateway Authorizer
              resource.
            displayName: Authorizer
            kind: Authorizer
            name: authorizers.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Deployment represents the state of an AWS apigateway Deployment
              resource.
            displayName: Deployment
            kind: Deployment
            name: deployments.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Integration represents the state of an AWS apigateway Integration
              resource.
            displayName: Integration
            kind: Integration
            name: integrations.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Method represents the state of an AWS apigateway Method resource.
            displayName: Method
            kind: Method
            name: methods.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Resource represents the state of an AWS apigateway Resource
              resource.
            displayName: Resource
            kind: Resource
            name: resources.apigateway.services.k8s.aws
            version: v1alpha1
          - description: RestAPI represents the state of an AWS apigateway RestAPI
              resource.
            displayName: RestAPI
            kind: RestAPI
            name: restapis.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Stage represents the state of an AWS apigateway Stage resource.
            displayName: Stage
            kind: Stage
            name: stages.apigateway.services.k8s.aws
            version: v1alpha1
          - description: VPCLink represents the state of an AWS apigateway VPCLink
              resource.
            displayName: VPCLink
            kind: VPCLink
            name: vpclinks.apigateway.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon API Gateway resources in AWS from within your Kubernetes cluster.

          **About Amazon API Gateway**

          Amazon API Gateway is an AWS service for creating, publishing, maintaining, monitoring, and securing REST, HTTP, and WebSocket APIs at any scale. API developers can create APIs that access AWS or other web services, as well as data stored in the AWS Cloud. As an API Gateway API developer, you can create APIs for use in your own client applications. Or you can make your APIs available to third-party app developers.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon API Gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apigateway
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon API Gateway Developer Resources
          url: https://aws.amazon.com/api-gateway/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: apigateway maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/apigateway-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-apigateway-controller.v1.3.1
        version: 1.3.1
      - name: ack-apigateway-controller.v1.3.0
        version: 1.3.0
      - name: ack-apigateway-controller.v1.2.5
        version: 1.2.5
      - name: ack-apigateway-controller.v1.2.4
        version: 1.2.4
      - name: ack-apigateway-controller.v1.2.3
        version: 1.2.3
      - name: ack-apigateway-controller.v1.2.2
        version: 1.2.2
      - name: ack-apigateway-controller.v1.2.1
        version: 1.2.1
      - name: ack-apigateway-controller.v1.1.0
        version: 1.1.0
      - name: ack-apigateway-controller.v1.0.6
        version: 1.0.6
      - name: ack-apigateway-controller.v1.0.5
        version: 1.0.5
      - name: ack-apigateway-controller.v1.0.4
        version: 1.0.4
      - name: ack-apigateway-controller.v1.0.2
        version: 1.0.2
      - name: ack-apigateway-controller.v1.0.1
        version: 1.0.1
      - name: ack-apigateway-controller.v1.0.0
        version: 1.0.0
      - name: ack-apigateway-controller.v0.0.3
        version: 0.0.3
      - name: ack-apigateway-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-apigateway-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nexus-repository-ha-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-repository-ha-operator-certified.v3.84.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusRepo",
                "metadata": {
                  "name": "nexusrepo-sample"
                },
                "spec": {
                  "ingress": {
                    "additionalRules": null,
                    "annotations": null,
                    "defaultRule": false,
                    "dockerIngress": {
                      "annotations": null,
                      "enabled": false,
                      "host": "example.com",
                      "name": "nexus-docker-ingress"
                    },
                    "dockerSubdomain": false,
                    "enabled": false,
                    "host": "example.com",
                    "name": "nexus-ingress",
                    "tls": {
                      "enabled": false,
                      "secretName": "tlsSecretName"
                    }
                  },
                  "license": {
                    "fileContentsBase64": "your_license_file_contents_in_base_64",
                    "secretName": "nexus-repo-license"
                  },
                  "secret": {
                    "nexusSecret": {
                      "enabled" : false,
                      "name": "nexus-secret.json",
                      "nexusSecretsKeyId": "super_secret_key_id",
                      "secretKeyfileContentsBase64": "secretKeyfileContentsBase64",
                      "mountPath": "/var/nexus-repo-secrets"
                    }
                  },
                  "nexusData": {
                    "pvc": {
                      "accessMode": "ReadWriteOnce",
                      "size": "2Gi"
                    },
                    "storageClass": {
                      "enabled": false,
                      "name": "nexusrepo-storage",
                      "parameters": {},
                      "provisioner": "provisioner",
                      "reclaimPolicy": "Retain",
                      "volumeBindingMode": "WaitForFirstConsumer"
                    },
                    "volumeClaimTemplate": {
                      "enabled": false
                    }
                  },
                  "service": {
                    "docker": {
                      "enabled": false,
                      "name": "nexus-repo-docker-service",
                      "port": 9090,
                      "protocol": "TCP",
                      "targetPort": 9090,
                      "type": "NodePort"
                    },
                    "nexus": {
                      "enabled": false,
                      "name": "nexus-repo-service",
                      "port": 80,
                      "protocol": "TCP",
                      "targetPort": 8081,
                      "type": "NodePort"
                    }
                  },
                  "statefulset": {
                    "container": {
                      "containerPort": 8081,
                      "env": {
                        "clustered": false,
                        "install4jAddVmParams": "-Xms2703m -Xmx2703m",
                        "jdbcUrl": null,
                        "nexusInitialPassword": "nexusabc",
                        "password": "nexus",
                        "user": "nexus",
                        "zeroDowntimeEnabled": false
                      },
                      "imageName": "registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:fbad488ff2e368845473a11b4955992fe826a94811da3abf5cc3c6a33ff3974f",
                      "pullPolicy": "IfNotPresent",
                      "resources": {
                        "limits": {
                          "cpu": 4,
                          "memory": "8Gi"
                        },
                        "requests": {
                          "cpu": 4,
                          "memory": "8Gi"
                        }
                      },
                      "terminationGracePeriod": 120
                    },
                    "imagePullSecrets": {},
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 240,
                      "path": "/",
                      "periodSeconds": 60
                    },
                    "name": "nexusrepo-statefulset",
                    "readinessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 240,
                      "path": "/",
                      "periodSeconds": 60
                    },
                    "replicaCount": 1
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-certified@sha256:99f30aa5ba8469d96de3efe69d1b21e73ff022172fc79144bba496d1119ec8a2
          createdAt: "2023-08-14"
          description: |-
            Nexus Repository is the central source of control to efficiently manage all binaries
            and build artifacts across your DevOps pipeline.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.replaces: 3.80.0-2
          olm.skipRange: <3.84.0-1
          repository: https://github.com/sonatype/operator-nexus-repository
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus Repository
            displayName: NexusRepo
            kind: NexusRepo
            name: nexusrepos.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Repository is the central source of control to efficiently manage all binaries
          and build artifacts across your DevOps pipeline.
          The flow of open source and third-party components into and through an organization
          creates a complex software supply chain.
          Nexus Repository delivers speed, efficiency, and quality to the governance
          and management of all dependencies, libraries, and applications for your DevOps teams.

          ## Core Capabilities

          * **Dependency Management**:
            Improves reliability with repeatable, fast access to secure dependencies
          * **Developer Productivity**:
            Streamline developer workflows by enabling the sharing of components and applications across teams
          * **Supply Chain Performance**:
            Improve speed-to-market and reduced build times with release advanced staging and component tagging
          * **CI/CD Integrations**:
            Increase DevOps scalability with integrations to the most popular build and deployment tools

          Version control systems and package registries do not scale when managing proprietary,
          open source, and third-party components.
          Organizations need a central binary and build artifact repository to manage dependencies
          across the entire software supply chain.

          ## Limitations

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus Repository Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage
          By default, the example Custom Resource starts up a single instance of Nexus Repository in non-Pro mode using an H2 database.

          ### Connect to a PostgreSQL database
          We strongly recommend using Nexus Repository with a PostgreSQL database. To connect a Nexus Repository instance to a PostgreSQL database, do as follows:
          * Update the `license.fileContentsBase64` parameter with a base 64 representation of your Nexus Repository Pro license.
          * Set the `statefulset.container.env.jdbcUrl` parameter to your database url.
          * Set the `statefulset.container.env.user` parameter to your database user name.
          * Set the `statefulset.container.env.password` parameter to your database password.

          ### High Availability (HA) mode
          To start Nexus Repository in HA mode, you must edit the Custom Resource YAML as follows:

          * Update the `license.fileContentsBase64` parameter with a base 64 representation of your Nexus Repository Pro license.
          * Set the `statefulset.container.env.jdbcUrl` parameter to your database url.
          * Set the `statefulset.container.env.user` parameter to your database user name.
          * Set the `statefulset.container.env.password` parameter to your database password.
          * Set the `statefulset.container.env.clustered` parameter to true.
          * Set the `statefulset.replicaCount` parameter to the number of Nexus Repository pods you require in your HA cluster
          * You must use a persistent volume as described in the 'Storage' section below.

          ### Storage
          The default configuration uses an emptyDir volume for storing Nexus Repository logs. However, we strongly recommend that
          you configure dynamic provisioning of persistent storage based on your deployment environment as explained below.

          #### Cloud deployments (AWS/Azure)
          * Ensure the appropriate Container Storage Interface (CSI) driver(s) are installed for your chosen OpenShift cloud deployment.
            Please, refer to OpenShift documentation for details on configuring CSI drivers.
          Note: if you're using Red Hat OpenShift on AWS (ROSA), CSI drivers for dynamically provisioning EBS volumes are installed
          by default; you should see associated storage classes for them in your cluster web console.

          #### On premise deployments
          1. Attach separate disks (i.e. separate from the root disk) to your worker nodes.
          2. Install the Local Storage Operator. Please refer to Open Shift persistent storage documentation:
             https://docs.openshift.com/container-platform/4.13/storage/persistent_storage/persistent_storage_local/persistent-storage-local.html
          and the Local Storage Operator documentation: https://github.com/openshift/local-storage-operator.
          3. Use the Local Storage Operator to automatically create persistent volumes for your chosen storage class name as documented here:
             https://docs.openshift.com/container-platform/4.13/storage/persistent_storage/persistent_storage_local/persistent-storage-local.html
          4. Specify the storage class name from step 3 as the value of the `nexusData.storageClass.name` property and also set `nexusData.volumeClaimTemplate.enabled` to true.

          #### Configuring the operator for dynamic persistent volume provisioning
          * Set the `nexusData.storageClass.name` parameter to a storage class name. This could be one of the default storage classes
            automatically created in your OpenShift cluster (e.g., if you're using ROSA) or one that you would like the operator to create.
          * If you would like to create a dedicated storage class (i.e., you don't want to use the default), then in addition to
            specifying a value for `nexusData.storageClass.name`, you must also set `nexusData.storageClass.enabled` parameter to `true`.
          * Set the `nexusData.volumeClaimTemplate.enabled` parameter to `true`.

          ### Additional configuration
          * Once the operator has created the server instance and it is running, you can further configure Nexus Repository via the NexusRepo custom resource:

          | Parameter                                                   | Description                         | Default                                 |
          | ----------------------------------------------------------- | ----------------------------------  | ----------------------------------------|
          | `statefulset.name`                                          |  The suffix for the statefulset's service name    | nexusrepo-statefulset     |
          | `statefulset.replicaCount`                                  |  The desired number of Nexus Repository pods    | 3                           |
          | `statefulset.container.imageName`                           |  The Nexus repository image registry url    |
          | | | registry.connect.redhat.com/sonatype/nexus-repository-manager:<app version>-ubi-1|
          | `statefulset.container.resources.requests.cpu`              |  The minimum cpu the Nexus repository pod can request    | 4 |
          | `statefulset.container.resources.requests.memory`           |  The minimum memory the Nexus repository pod can request    | 8Gi|
          | `statefulset.container.resources.limits.cpu`                |  The maximum cpu the Nexus repository pod may get.    | 4|
          | `statefulset.container.resources.limits.memory`             |  The maximum memory the Nexus repository pod may get.  | 8Gi|
          | `statefulset.container.containerPort`                       |  The Nexus Repository container's HTTP port  | 8081|
          | `statefulset.container.pullPolicy`                          |  The Nexus Repository docker image pull policy  | IfNotPresent|
          | `statefulset.container.terminationGracePeriod`              |  The time given for the pod to gracefully shut down  | 120 seconds|
          | `statefulset.container.env.clustered`                       |  Controls whether Nexus Repository should start in clustered mode  | False|
          | `statefulset.container.env.install4jAddVmParams`            |  Xmx and Xms settings for JVM  | -Xms2703m -Xmx2703m|
          | `statefulset.container.env.user`                            |  The database user name | nexus (can be overridden via Kubernetes secret. See above)|
          | `statefulset.container.env.password`                        |  The database password  | nexus (can be overridden via Kubernetes secret. See above)|
          | `statefulset.container.env.jdbcUrl`                         |  The database url |
          | | | jdbc:postgresql://your_database_server:5432/nexus (must be overridden via Kubernetes |
          | | | secret object which feeds into env variable variables defined in the operator.yaml . See above)|
          | `statefulset.container.env.nexusInitialPassword`            |  The initial admin password for Nexus Repository UI. This is for first time |
          | | installs only. Subsequent installs/restarts on the same node ignores this setting | nexusabc|
          | `statefulset.livenessProbe.initialDelaySeconds`             |  The initial delay before the Nexus Repository application is pinged for a |
          | | response i.e. to determine whether the application is still alive | 240|
          | `statefulset.livenessProbe.path`                            |  The path to ping the application | |
          | `statefulset.livenessProbe.periodSeconds`                   |  The interval for checking that the application is still alive i.e. responsive, not deadlocked etc | 60|
          | `statefulset.livenessProbe.failureThreshold`                |  The maximum number of times the application ping can fail before the |
          | | application is considered unresponsive. Kubernetes will attempt to restart the application's pod in this case | 6|
          | `statefulset.readinessProbe.initialDelaySeconds`            |  The initial delay before the Nexus Repository application is checked |
          | | whether it is ready to receive traffic | 240|
          | `statefulset.readinessProbe.path`                           |  The path where the Nexus Repository application is checked for traffic readiness | |
          | `statefulset.readinessProbe.periodSeconds`                  |  The interval for checking the Nexus Repository application for traffic readiness | 60|
          | `statefulset.readinessProbe.failureThreshold`               |  The maximum number of times to check the Nexus Repository application for traffic readiness | 6|
          | `statefulset.imagePullSecrets`                              |  The pull secret for private image registries | {}|
          | `ingress.name`                                              |  The name of an Ingress object which listens for requests to Nexus Repository on port 80 | nexus-ingress|
          | `ingress.enabled`                                           |  Whether or not to create the Ingress | false|
          | `ingress.host`                                              |  Ingress host | "example.com"|
          | `ingress.defaultRule`                                       |  Whether or not to add a default rule for the Nexus Repository |
          | | Ingress which forwards traffic to a Service object  | false|
          | `ingress.additionalRules`                                   |  Additional rules to add to the ingress | null|
          | `ingress.tls.enabled`                                       |  Whether or not TLS should be enabled for the Ingress | false|
          | `ingress.tls.secretName`                                    | The name of a Secret object in which to store the TLS secret | tlsSecretName|
          | `ingress.dockerSubdomain`                                   |  Whether or not to add rules for docker subdomains | false|
          | `ingress.annotations`                                       |  Annotations for the Ingress object | null |
          | `license.fileContentsBase64`                                |  Your Base64 encoded Nexus Repository license | your_license_file_contents_in_base_64 |
          | `license.secretName`                                        |  The name of a secret object in which to store your |
          | | Base64-encoded Nexus Repository license | nexus-repo-license|
          | `ingress.dockerIngress.name`                                |  The name of an Ingress object which listens for requests to an |
          | | existing docker repository in Nexus Repository | nexus-docker-ingress|
          | `ingress.dockerIngress.enabled`                             |  Whether or not to create an Ingress for a docker repository | false|
          | `ingress.dockerIngress.host`                                |  The docker Ingress's host | "example.com"|
          | `ingress.dockerIngress.annotations`                         |  Annotations for the docker Ingress object | null |
          | `nexusData.pvc.size`                                        |  The volume size to request for storing Nexus logs | 2Gi |
          | `nexusData.pvc.accessMode`                                  |  The persistent volume claim access mode | ReadWriteOnce |
          | `nexusData.storageClass.name`                               |  The name of the storage class to use for dynamically requesting that |
          | | storage be provisioned when using dynamic provisioning or to randomly select a pre-created persisted volume when using the Local Storage Operator. | nexusrepo-storage |
          | `nexusData.storageClass.reclaimPolicy`                      |  The reclaim policy for persistent volume's created by the storage class | Retain |
          | `nexusData.storageClass.enabled`                            |  Whether or not a storage class object should be created. | false |
          | `nexusData.storageClass.volumeBindingMode`                  |  It controls when volume binding and dynamic provisioning should occur | WaitForFirstConsumer |
          | `nexusData.storageClass.provisioner`                        |  The name of the volume provisioner | provisioner |
          | `nexusData.storageClass.parameters`                         |  The additional parameters for the storage class | {} |
          | `nexusData.volumeClaimTemplate.enabled`                           |  Whether or not volume should be requested using a claim template. |
          | | You should set this property to true when using dynamic volume provisioning (cloud deployments) |
          | | or if you're using Local Storage Operator to automatically create persistent volumes for locally attached disks (i.e. not root disk) | false |
          | `license.secretName | A name for the Kubernetes secret object which stores the base64 representation of your |
          | | Nexus Repository license specified via `license.fileContentsBase64` is stored | nexus-repo-license |
          | `license.fileContentsBase64 | A base64 representation of your Nexus Repository license | your_license_file_contents_in_base_64 |
          | `service.nexus.name`                                              |  The name of a kubernetes Service Object which allows access to |
          | | Nexus Repository's main HTTP port | nexus-repo-service |
          | `service.nexus.enabled`                                           |  Whether or not to create the Service object | false|
          | `service.nexus.type`                                              |  The type of the Kubernetes Service | "NodePort"|
          | `service.nexus.protocol`                                       |  The protocol  | TCP |
          | `service.nexus.port`                                   |  The port to listen for incoming requests | 80|
          | `service.nexus.targetPort`                                   |  The port to forward requests to | 8081 |
          | `service.docker.name`                                              |  The name of a kubernetes Service Object which allows access to a |
          | | docker repository within Nexus Repository | nexus-repo-docker-service |
          | `service.docker.enabled`                                           |  Whether or not to create a Kubernetes Service object for a given |
          | | docker repository within Nexus Repository | false |
          | `service.docker.type`                                              |  The type of the Kubernetes Service | "NodePort"|
          | `service.docker.protocol`                                       |  The protocol  | TCP |
          | `service.docker.port`                                   |  The port to listen for incoming requests | 9090 |
          | `service.docker.targetPort`                                   |  The port to forward requests to | 9090 |
        displayName: Nexus Repository HA Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - repository
        - sonatype
        links:
        - name: Nexus Repository Manager
          url: https://www.sonatype.com/product-nexus-repository
        maintainers:
        - email: support@sonatype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-certified@sha256:99f30aa5ba8469d96de3efe69d1b21e73ff022172fc79144bba496d1119ec8a2
        - registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:fbad488ff2e368845473a11b4955992fe826a94811da3abf5cc3c6a33ff3974f
        version: 3.84.0-1
      entries:
      - name: nexus-repository-ha-operator-certified.v3.84.0-1
        version: 3.84.0-1
      name: stable
    defaultChannel: stable
    packageName: nexus-repository-ha-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-kubernetes.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "ClusterMongoDBRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-enterprise"
                  },
                  "name": "clustermongodbrole-sample"
                },
                "spec": {
                  "db": "admin",
                  "role": "rootMonitor",
                  "roles": [
                    {
                      "db": "admin",
                      "role": "root"
                    },
                    {
                      "db": "admin",
                      "role": "clusterMonitor"
                    }
                  ]
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "my-replica-set"
                },
                "spec": {
                  "credentials": "my-credentials",
                  "members": 3,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "sample-sharded-cluster"
                },
                "spec": {
                  "configServerCount": 3,
                  "credentials": "my-credentials",
                  "mongodsPerShardCount": 3,
                  "mongosCount": 2,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "shardCount": 1,
                  "type": "ShardedCluster",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBMultiCluster",
                "metadata": {
                  "name": "multi-replica-set"
                },
                "spec": {
                  "clusterSpecList": [
                    {
                      "clusterName": "e2e.cluster1.mongokubernetes.com",
                      "members": 2
                    },
                    {
                      "clusterName": "e2e.cluster2.mongokubernetes.com",
                      "members": 1
                    },
                    {
                      "clusterName": "e2e.cluster3.mongokubernetes.com",
                      "members": 2
                    }
                  ],
                  "credentials": "my-credentials",
                  "duplicateServiceObjects": false,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBOpsManager",
                "metadata": {
                  "name": "ops-manager"
                },
                "spec": {
                  "adminCredentials": "ops-manager-admin",
                  "applicationDatabase": {
                    "members": 3,
                    "podSpec": {
                      "cpu": 1
                    }
                  },
                  "configuration": {
                    "mms.fromEmailAddr": "admin@thecompany.com"
                  },
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "version": "6.0.3"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBUser",
                "metadata": {
                  "name": "my-replica-set-x509-user"
                },
                "spec": {
                  "db": "$external",
                  "mongodbResourceRef": {
                    "name": "my-replica-set"
                  },
                  "roles": [
                    {
                      "db": "admin",
                      "name": "dbOwner"
                    }
                  ],
                  "username": "CN=my-replica-set-x509-user,OU=cloud,O=MongoDB,L=New York,ST=New York,C=US"
                },
                "status": {
                  "phase": "Updated"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/mongodb/mongodb-kubernetes@sha256:745cc2c6bb893c1cb1e75d79b6f3bfc7b125a6a62f7431bdc43cf6cacd373998
          createdAt: "2025-09-16T14:56:17Z"
          description: The MongoDB Controllers for Kubernetes enable easy deploys
            of MongoDB into Kubernetes clusters, using our management, monitoring
            and backup platforms, Ops Manager and Cloud Manager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mongodb/mongodb-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterMongoDBRole
            name: clustermongodbroles.mongodb.com
            version: v1
          - description: MongoDB Deployment
            displayName: MongoDB Deployment
            kind: MongoDB
            name: mongodb.mongodb.com
            version: v1
          - kind: MongoDBCommunity
            name: mongodbcommunity.mongodbcommunity.mongodb.com
            version: v1
          - description: MongoDB Multi Deployment
            displayName: MongoDB Multi Deployment
            kind: MongoDBMultiCluster
            name: mongodbmulticluster.mongodb.com
            version: v1
          - kind: MongoDBSearch
            name: mongodbsearch.mongodb.com
            version: v1
          - description: MongoDB x509 User
            displayName: MongoDB User
            kind: MongoDBUser
            name: mongodbusers.mongodb.com
            version: v1
          - description: MongoDB Ops Manager
            displayName: MongoDB Ops Manager
            kind: MongoDBOpsManager
            name: opsmanagers.mongodb.com
            version: v1
        description: |
          MongoDB Controllers for Kubernetes (MCK) is an open-source project. It represents a merge of the previous
          MongoDB Community Operator (MCO) and MongoDB Enterprise Kubernetes Operator (MEKO).

          This release brings MongoDB Community and Enterprise editions together under a single, unified operator,
          making it easier to manage, scale, and upgrade your deployments. While the first version simply brings
          the capabilities of both into a single Operator, future changes will build on this to more closely align
          how Community and Enterprise are managed in Kubernetes, to offer an even more seamless and streamlined experience.
          As an open-source project, it now allows for community contributions, helping drive quicker bug fixes and ongoing innovation.
          The MongoDB Controllers for Kubernetes enable easy deploys of MongoDB
          into Kubernetes clusters, using our management, monitoring and backup
          platforms, Ops Manager and Cloud Manager.

          ## Before You Start

          To start using the operator you''ll need an account in MongoDB Cloud Manager or
          a MongoDB Ops Manager deployment.

          * [Create a Secret with your OpsManager API key](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-operator-credentials/#procedure)

          * [Create a ConfigMap with your OpsManager project ID and URL](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-project-using-configmap/)

          By installing this integration, you will be able to deploy MongoDB instances
          with a single simple command.

          ## Required Parameters

          * `opsManager` or `cloudManager` - Enter the name of the ConfigMap containing project information
          * `credentials` - Enter the name of the Secret containing your OpsManager credentials
          * `type` - Enter MongoDB Deployment Types ("Standalone", "ReplicaSet", "ShardedCluster"

          ## Supported MongoDB Deployment Types ##

          * Standalone: An instance of mongod that is running as a single server and
          not as part of a replica set, this is, it does not do any kind of
          replication.

          * Replica Set: A replica set in MongoDB is a group of mongod processes that
          maintain the same data set. Replica sets provide redundancy and high
          availability, and are the basis for all production deployments. This section
          introduces replication in MongoDB as well as the components and architecture
          of replica sets. The section also provides tutorials for common tasks
          related to replica sets.

          * Sharded Cluster: The set of nodes comprising a sharded MongoDB deployment.
          A sharded cluster consists of config servers, shards, and one or more mongos
          routing processes. Sharding is a A database architecture that partitions
          data by key ranges and distributes the data among two or more database
          instances. Sharding enables horizontal scaling.

          ## Requirements for deploying MongoDB OpsManager

          * In order to deploy resources of type MongoDB OpsManager, you will need to
          create a secret containing the [credentials](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/plan-om-resource/#om-rsrc-prereqs)
          for the Global Owner user

          ## Security ##

          The operator can enable TLS for all traffic between servers and also between
          clients and servers. Before enabling `security.tls.enabled` to `true` you
          should create your certificates.  or you can leave the operator to create all
          the certificates for you. The operator ability to create certs is been
          deprecated due to Kubernetes API changes.

          For more information, please read the official MongoDB
          Kubernetes Operator  [docs](https://docs.mongodb.com/kubernetes-operator/stable/).
        displayName: MongoDB Controllers for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mongodb
        - databases
        - nosql
        links:
        - name: Documentation
          url: https://docs.opsmanager.mongodb.com/current/tutorial/install-k8s-operator/index.html
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: stable
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-kubernetes-init-database@sha256:3ea9106c0e35c015bc8f15657c76d974bb1cadb9220dc58be2076c9699f26fb7
        - quay.io/mongodb/mongodb-agent@sha256:6e826bbe7dd812c101464e75295370c4b10e4e9c31477db89cfceb96759853e9
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:702f529cb9d8d9042e54263f8bc37b40c062ac2695efc3e7150819779886e323
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c48631742ad1fec8a9ee745fc3cc9ca1fe7d86ef7fa68ad8e8337c7a4566dc87
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7e5361c2a1604dec3fc671d629806fe6ddc6df524b69a441b86933a4b30093
        - quay.io/mongodb/mongodb-enterprise-server@sha256:85c1d669c073caf9cbaac815e4c8ea2ca18b71deefa03bf84cd6d8b8ebd53e12
        - quay.io/mongodb/mongodb-agent@sha256:fa30370a75e8e4ac648650ceb39a7a259a754aab8a48d7e2f2801cd1b777e91b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f1e72679b2468aed2f3c3106568327911e6594c9ee15f7348e953eaa0d8fd44a
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:85b4e954c9c4bbe1019774bad1e05e9c8dc175648b01d3badf0b64143f182019
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d323c1bcd14118cc537ac282996ea5a58756d20b6b73f1eec0efa8710d246077
        - quay.io/mongodb/mongodb-enterprise-server@sha256:67238129df09d33ee4746ff1acb5515d699a29d9f42350eae3f3f69ccb7fbf54
        - quay.io/mongodb/mongodb-kubernetes-database@sha256:f8c981a6223f70d7c6bbd30c503fc8b2497f351830b2b74c30e6f81863d78d2f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:a4590af027182305913f0de1a2e37ad8612bd73b8b31f4df11068b80089021ff
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3add8b68bfbaa42596fe5dbbee4623ba5ff1a826b18bb71301525adf574dcb94
        - quay.io/mongodb/mongodb-agent@sha256:ee9f3906d893004bb2aaaa6412646f58288b0684502b9e90f33cafadcfce4349
        - quay.io/mongodb/mongodb-enterprise-server@sha256:df6081325ea2fe01d1b27d0c70005dec8232aaf51a43d91661b722b8a1761263
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:146c4f5dc443f94f54d72808b4e4df93bb904e23589b222d17700166780776b4
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:09788976bdc1327acbf18cedf8efd5fca2159e43b506faa40e03ee42ad94928a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:655896bac9e82de8e9729d59a36c3a31f66235f22b93449260a2ccb7709de084
        - quay.io/mongodb/mongodb-enterprise-server@sha256:172e338a43df01ee5408d8d1e3cb038d2c55e77761c1a2ac6ed37959d1525140
        - quay.io/mongodb/mongodb-agent@sha256:3fa5bcfda3228e0ea880434a60ced829e3872e74763b347636c18475d866d0e1
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:33bbd29b74f95baad522b380d70bcc9f933a3fee57642c37b09376c86179c40f
        - quay.io/mongodb/mongodb-kubernetes@sha256:745cc2c6bb893c1cb1e75d79b6f3bfc7b125a6a62f7431bdc43cf6cacd373998
        - quay.io/mongodb/mongodb-enterprise-server@sha256:19925e1297ab652db59c35f53e7b7c8b2b277f8ecdc4215682961aaf2b3e924e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:f71b027fccf649f18f63f8df7a59b1bac421a2cdab6712e0c985a93c08ed970a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8445d59d491a6d3890ea9c8b421214bcd0fdcfbc420ed40497a790b79f0bc89e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:895dbdfb87ddc2e9693671c7108a31ab28e16fec2e5be6d046bdd2e30319eade
        - quay.io/mongodb/mongodb-enterprise-server@sha256:dfd4277f0c16ed476ac7de51894739e8c8d13a65ea5024b9a97b3dce0f6c6975
        - quay.io/mongodb/mongodb-enterprise-server@sha256:fb8397cbcdbf7c33f6ad142492792771127f0d9a8c46388f350b6927405ee590
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7885ab6e88214f9289d8babf5189136dde0dd4c84a15a5481ccd3ef512b2e66f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:5d299820bb903f0cdfdf2626a3c44099d2eaee16188cd8234d2ff3a57259b3a7
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:439d6fefb469613432e2d05149dfc96c32b16469d743835d5f402a2ada8be8cc
        - quay.io/mongodb/mongodb-agent@sha256:da77dcb60b927bf6059c2c9bf008f8c3c0eec45cf7d1b40c34a0c263e8cc18b8
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b8aa1edbc5fb02af312a863bc7db8535a282dfd692a07b1e321e5a1acbfeb6a1
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9a3a6496be0d2d446e6d654eb56386c7726e0cd2d7232f6dcec579923edf0014
        - quay.io/mongodb/mongodb-agent@sha256:89819704d70a3b6b4d2358745a22060ef740b33ad5c6debf8e6ac4d4a6fd63ba
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b7764c316eeacffd3f9ddae62d9fd2a74eb58484d2cfd5021e460095d604764f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:e7aeedb752dda77cf0e49df3430843b5a2bba64c8404628a983f46640d316808
        - quay.io/mongodb/mongodb-search@sha256:7149a7b6f1ee84646d3972837920f80fc5db49503c598406ce92e2f794bd5f0d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa8683500e8d3632c4588bbcdddb4e3186a01f48aefe1b6d95f8860dae8b6c76
        - quay.io/mongodb/mongodb-enterprise-server@sha256:15e31285c8949a4666f7702df53416b0905d76945a28575a6028a876cc8a6f0f
        - quay.io/mongodb/mongodb-kubernetes-init-ops-manager@sha256:93c4633cbc84e9bfa3563771ac2efbf5e6515f380dec2c085494cb8326546377
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5c8553e5ba115657fb878c0fe08704af9fd5decec572a87d89d2c8c156ffc9f9
        - quay.io/mongodb/mongodb-agent@sha256:5b064c470335df840b050b714b28b75b6061444843250f1eb9215cb2b8c1a149
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a86c4d8a0b9ca3c20cc524809676eb03a944ada43378ec98c107cd72cc64700f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:5e3982493e85ac1e40536758c3145b7c2fff9a7b4be9d053ee41201363a89920
        - quay.io/mongodb/mongodb-agent@sha256:5f5031e4c2a3729f618d026b0a7f34fd4509ea73d5a4e3068b551041dc291583
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:d81eac2129b6ab8346be1749c56ea67f6da64d4ee0fb9f3ff4171ba0f61918eb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3316015ac0699d66ac73f714e5dc0784738654e0349e80b1e1f4e953b3c88964
        - quay.io/mongodb/mongodb-agent@sha256:b387b7458f5640f907c9074fcb08ae4d4785be31d4bb65f2d7c459cd6762b2d6
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:adeab0dc766a27138db7df5f37cbb1da860c14179ef631b2e001e5bb875325e4
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ef732e3eea91cc86caf2eabedc0c9115ed235a8d204f33e74bbf090a0b0259fb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a48a4d5c275fd669a27469649fb39ea38c728dbcf9f0a9da4e7cf27a2ce912ed
        - quay.io/mongodb/mongodb-enterprise-server@sha256:720056b985c1a16d48b2aecb7a88a15cba3d8186e55b3c6c24729699a85c8232
        - quay.io/mongodb/mongodb-enterprise-server@sha256:523566f44cbb6bbf0cd12b1adae939a883f2254c78dc4b8c9d0098581337b1db
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f2d9a3264a8814f33143b73c56624572fb80afeab184133d3a901713a31c1a87
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5145c97d2281aabcd8538ff515c3833aa48acb66b0b739fb1881120c1985d88a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9bef77e905dadd1e61d0439dd53b21ed1f3606e49affd9ffc4ba8156578e8f0f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:cd869ff7d661059fad7593e30141d74e8cc751295b1579577f7a853396eb5d0b
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:fa4aa97767bd19cf3711a74bf21ecb9cbbc25fed6b6b09a2eb241ebd3fbc8f3a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d49417ea85ee1ddb84429552a5aac87a82c2843904d8c0300a43b657b7a49e0e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:33a30d11969eb21952681108092fe7f916e25c8e1de43e6a4e40e126025a2bec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:778f9e29e36d6687cd17dd95ec1f8127d2a732bde1a2dd5821212eb55c1c998a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1d418c65d82ccad205b7889be4ba161d3cbfa8b1fb75b21c8fcf02de38fad9ca
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b9ab7147b59b62e40a69c6400f87d0c7faecfa5462bed6e68ca653b00e50b887
        - quay.io/mongodb/mongodb-agent@sha256:6394391f78da5e3218bb2c694e7ec9f6682fd17a6dd7e7250b0c0b5e1729796d
        - quay.io/mongodb/mongodb-agent@sha256:a8358ce465adf463d8e73916d7944765d7aaf2b0ed10bd3b8119a87fc9080667
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:78f11c9cd5e63f2b0e10c67dc1491f4ce1279fa90178c108814725414defab9a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c752a7f0726d21d55f204a359b85fa29cd727a5fe571b37289ae3ea52d6206b0
        - quay.io/mongodb/mongodb-agent@sha256:b422829aac63f068aed4a6948fd14cbd3c3bfad6d47ab76523f6839bf20b0203
        - quay.io/mongodb/mongodb-agent@sha256:6c23febaa9dd52bc29ac1e52d087fc4b2b47c578f4d6ac3a1674bb750a38cb07
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:3bbd3d7fd45d107621e7770f2b4c2a758e148cd472e0ee21719442f71bf6ade0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0fde236fe26d5b20210d6f246aadacd7845874712def516cbe2eec11d79d5181
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d2d8b59c1f39c41951e7878963f8496eb494688720366f8edc7060887cdbe6a0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7a93a0276531ff9be4c90bb8fe8d104e0a9e930c29792aafe03cc6a76a9fa89c
        - quay.io/mongodb/mongodb-agent@sha256:6cc4e953df3f09d793f0dbc6e57e55a3c5277ef63177a6d8f691a41227facf64
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3fe62f3ad52707ccf0e9993c69891e44c569a4b4e1a3244cf60e3b5fdde57bd2
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1a77851485ad3b2566e765b4bb051ff0ddfa48703492cef6c6db021dde84d87f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:bfbcfcfc5f0de22e20520b2d96f826638e424adb028a909ccc192013d11f5833
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b00ac773623e1bf422b47e6d783263d2d08737865abe9abbe106d86d1ea25165
        - quay.io/mongodb/mongodb-enterprise-server@sha256:37af0913742b6de339c7e664b30f35dcef04506529e4a09a9bbbdd57643f7149
        - quay.io/mongodb/mongodb-agent@sha256:d8a2ef79d767f03162ec57fe028f9072294b292d376af8eb97d2967eb6d7047b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7266b405ebb2dd24a9f7c12d3921024f547c3d8c01831e6bcfc1d551d63be1
        - quay.io/mongodb/mongodb-agent@sha256:af3de4f631d0b00cae7f7e80c3900378ba7e449e01830bf9dc643b873d2e6d33
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa300abf739e26350a258f643fa8e57aedacd041055c7b923bd22b21ef554cdb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d21c65780cadc9d8668d2c28935d492625d6993f13ba562722200a819281cbcc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:376ad7e5fe42651d946cba5705c0eee44aa636c1b4ff258210e4ce2fe4205b4d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:454b1585c342375dea643b10ead688d0f62347841f5f30059fcd4f92e3297ee9
        - quay.io/mongodb/mongodb-enterprise-server@sha256:11ab324698f919f84a5dab5a964d93fe63f98a48146bb3b86974cba7606d30ec
        - quay.io/mongodb/mongodb-kubernetes-init-appdb@sha256:afb0d463304fc0bfb3d8dade91459414daac6bf6fcf742347a2012cbc9fea354
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c8aec83b76dfecc05b8f5d8f40fd61b24980c0287a07294f75f321c2d42c32b4
        - quay.io/mongodb/mongodb-agent@sha256:2d5652d0e263aa94554e7a068c93e4fd94b84ec223f077361d030d172bc93f67
        version: 1.4.0
      entries:
      - name: mongodb-kubernetes.v1.4.0
        version: 1.4.0
      - name: mongodb-kubernetes.v1.3.0
        version: 1.3.0
      - name: mongodb-kubernetes.v1.2.0
        version: 1.2.0
      - name: mongodb-kubernetes.v1.1.0
        version: 1.1.0
      - name: mongodb-kubernetes.v1.0.1
        version: 1.0.1
      - name: mongodb-kubernetes.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: namespace-configuration-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: namespace-configuration-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupConfig",
                "metadata": {
                  "name": "test-groupconfig"
                },
                "spec": {
                  "labelSelector": {
                    "matchLabels": {
                      "type": "team"
                    }
                  },
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-dev\n  labels:\n    group: {{ .Name }}  \n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-qa\n  labels:\n    group: {{ .Name }}          \n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-prod\n  labels:\n    group: {{ .Name }}                           \n"
                    },
                    {
                      "objectTemplate": "apiVersion: quota.openshift.io/v1\nkind: ClusterResourceQuota\nmetadata:\n  name: {{ .Name }}-quota\nspec:\n  quota:\n    hard:\n      pods: \"4\" \n      requests.cpu: \"1\" \n      requests.memory: 1Gi \n      requests.ephemeral-storage: 2Gi \n      limits.cpu: \"2\" \n      limits.memory: 2Gi \n      limits.ephemeral-storage: 4Gi\n  selector:\n    labels:\n      matchLabels:\n        group: {{ .Name }}\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NamespaceConfig",
                "metadata": {
                  "name": "test-namespaceconfig"
                },
                "spec": {
                  "labelSelector": {
                    "matchLabels": {
                      "type": "secure"
                    }
                  },
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: ResourceQuota\nmetadata:\n  name: standard-sandbox\n  namespace: {{ .Name }}\nspec:\n  hard:\n    pods: \"4\" \n    requests.cpu: \"1\" \n    requests.memory: 1Gi \n    requests.ephemeral-storage: 2Gi \n    limits.cpu: \"2\" \n    limits.memory: 2Gi \n    limits.ephemeral-storage: 4Gi\n"
                    },
                    {
                      "objectTemplate": "kind: EgressNetworkPolicy\napiVersion: network.openshift.io/v1\nmetadata:\n  name: air-gapped\n  namespace: {{ .Name }}  \nspec:\n  egress:\n  - type: Deny\n    to:\n      cidrSelector: 0.0.0.0/0\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "UserConfig",
                "metadata": {
                  "name": "test-user-config"
                },
                "spec": {
                  "providerName": "my-provider",
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-sandbox\n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: ResourceQuota\nmetadata:\n  name: standard-sandbox\n  namespace: {{ .Name }}-sandbox\nspec:\n  hard:\n    pods: \"4\" \n    requests.cpu: \"1\" \n    requests.memory: 1Gi \n    requests.ephemeral-storage: 2Gi \n    limits.cpu: \"2\" \n    limits.memory: 2Gi \n    limits.ephemeral-storage: 4Gi\n"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/namespace-configuration-operator@sha256:49ed7d6155342adaa2b12fd80c6761c3081d8e6149d187cb7ff91a247cdf2e7a
          createdAt: "2023-11-22T15:58:50Z"
          description: This operator provides a facility to define and enforce namespace
            configurations
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: namespace-configuration-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/namespace-configuration-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GroupConfig is the Schema for the groupconfigs API
            displayName: Group Config
            kind: GroupConfig
            name: groupconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: NamespaceConfig is the Schema for the namespaceconfigs API
            displayName: Namespace Config
            kind: NamespaceConfig
            name: namespaceconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: UserConfig is the Schema for the userconfigs API
            displayName: User Config
            kind: UserConfig
            name: userconfigs.redhatcop.redhat.io
            version: v1alpha1
        description: "The `namespace-configuration-operator` helps keeping configurations
          related to Users, Groups and Namespaces aligned with one of more policies
          specified as a CRs. The purpose is to provide the foundational building
          block to create an end-to-end onboarding process. \nBy onboarding process
          we mean all the provisioning steps needed to a developer team working on
          one or more applications to OpenShift.\n This usually involves configuring
          resources such as: Groups, RoleBindings, Namespaces, ResourceQuotas, NetworkPolicies,
          EgressNetworkPolicies, etc.... . Depending on the specific environment the
          list could continue.\nNaturally such a process should be as automatic and
          scalable as possible.\n\nWith the namespace-configuration-operator one can
          create rules that will react to the creation of Users, Groups and Namespace
          and will create and enforce a set of resources.\n\nHere are some examples
          of the type of onboarding processes that one could support:\n\n1. [developer
          sandbox](https://github.com/redhat-cop/namespace-configuration-operator/blob/master/examples/user-sandbox/readme.md)\n
          2. [team onboarding](https://github.com/redhat-cop/namespace-configuration-operator//blob/master/examples/team-onboarding/readme.md)
          with support of the entire SDLC in a multitentant environment.\n\nPolicies
          can be expressed with the following CRDs:\n\n| Watched Resource | CRD |\n|--|--|\n
          | Groups | [GroupConfig](#GroupConfig) |\n| Users | [UserConfig](#UserConfig)
          |\n| Namespace | [NamespaceConfig](#NamespaceConfig) |\n\nThese CRDs all
          share some commonalities:\n\n1. Templated Resources\n2. List of ignored
          jason path\n \n### Templated Resources\n\nEach has a parameter called `templatedResources`,
          which is an array. Each element of the array has two fields `objectTemplate`
          and `excludedPaths` (see below).\n\nThe `objectTemplate` field must contains
          a [go template](https://golang.org/pkg/text/template/) that resolves to
          a single API Resource expressed in `yaml`. The template is merged with the
          object selected by the CR. For example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: UserConfig\nmetadata:\n  name: test-user-config\nspec:\n  provider:
          corp-ldap\n  templates:\n  - objectTemplate: |\n      apiVersion: v1\n      kind:
          Namespace\n      metadata:\n        name: {{ .Name }}-sandbox\n```\n\nThis
          create a rule in which every time a user from the `corp-ldap` provider is
          created, a namespace called `<username>-sandbox` is also created.\n\n###
          Excluded Paths\n\nThe logic of the `namespace-configuration-operator` is
          to enforce that the resources resolved by processing the templates \"stays
          in place\". In other words if those resources are changed and/or deleted
          they will be reset by the operator.\nBut there are situations in which at
          least part of a resource is allowed to change. Common use cases are: annotations
          and in general the metadata section of a resource can be updated by the
          various operators watching that resources. The status field is often updated
          by the main operator managing that resources. Finally, when applicable the
          `spec.replicas` field should also be allowed to change.\n\nTo handle special
          use case, one can also specify additional *jsonpaths* that should be ignored
          when comparing the desired resource and the current resource and making
          a decision on whether that resource should be reset.\n\nThe following paths
          are always included:\n \n1. `.metadata`\n2. `.status`\n3. `.spec.replicas`\n\n##
          NamespaceConfig\n\n The `NamespaceConfig` CR allows specifying one or more
          objects that will be created in the selected namespaces.\n\nNamespaces can
          be selected by labels or annotations via a label selector for example:\n\n```yaml\napiVersion:
          redhatcop.redhat.io/v1alpha1\n kind: NamespaceConfig\nmetadata:\n  name:
          small-namespace\n  selector:\n    matchLabels:\n      size: small  \n  resources:\n
          \ - apiVersion: v1\n    kind: ResourceQuota\n    metadata:\n      name:
          small-size  \n    spec:\n      hard:\n        requests.cpu: \"4\"\n        requests.memory:
          \"2Gi\"\n```\n\nHere is a `NamespaceConfig` object using a `matchExpressions`
          selector:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n kind: NamespaceConfig\nmetadata:\n
          \ name: tier-config\nspec:\n  annotationSelector:\n    matchExpressions:\n
          \     - {key: tier, operator: In, values: [gold,silver]}\n ```\n\nAlthough
          not enforced by the operator the general expectation is that the NamespaceConfig
          CR will be used to create objects inside the selected namespace.\n \nExamples
          of NamespaceConfig usages can be found [here](https://github.com/redhat-cop/namespace-configuration-operator/blob/master/examples/namespace-config/readme.md)\n
          \n## GroupConfig\n\nThe `GroupConfig` CR allows specifying one or more objects
          that will be created in the selected Group.\nGroups can be selected by labels
          or annotations via a label selector, similarly to the `NamespaceConfig`.\n\nOften
          groups are created in OpenShift by a job that synchronizes an Identity Provider
          with OCP. So the idea is that when new groups are added or deleted the configuration
          in OpenShift will adapt automatically.\n\nAlthough not enforced by the operator,
          GroupConfig are expected to create cluster-scoped resources like Namespaces,
          ClusterResourceQuotas and potentially some namespaced resources like RoleBindings.\n\n##
          UserConfig\n \nIn OpenShift an external user is defined by two entities:
          Users and Identities. There is a relationship of on to many between Users
          and Identities. Given one user, there can be one Identity per authentication
          mechanism.\n\nThe `UserConfig` CR allows specifying one or more objects
          that will be created in the selected User.\nUsers can be selected by label
          or annotation like `NamespaceConfig` and `UserConfig`.\nUSers can also be
          selected by provider name (the name of the authentication mechanism) and
          identity extra field.\n\nHere is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          UserConfig\nmetadata:\n  name: test-user-config\nspec:\n  providerName:
          okta-provider\n  identityExtraFieldSelector:\n    matchLabels:\n      sandbox_enabled:
          \"true\"\n  templates:\n  - objectTemplate: |\n      apiVersion: v1\n      kind:
          Namespace\n      metadata:\n        name: {{ .Name }}-sandbox\n ```\n\nUser
          will be selected by this `UserConfig` only if they login via the *okta-provider*
          and if the extra field was populate with the label `sandbox_enabled: \"true\"
          `. Note that not all authentication provider allow populating the extra
          fields in the Identity object.\n"
        displayName: Namespace Configuration Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - namespace
        - user
        - group
        - configuration
        - policy
        - management
        links:
        - name: Namespace Configuration Operator
          url: https://github.com/redhat-cop/namespace-configuration-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/namespace-configuration-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/namespace-configuration-operator@sha256:49ed7d6155342adaa2b12fd80c6761c3081d8e6149d187cb7ff91a247cdf2e7a
        version: 1.2.6
      entries:
      - name: namespace-configuration-operator.v1.2.6
        version: 1.2.6
      - name: namespace-configuration-operator.v1.2.5
        version: 1.2.5
      - name: namespace-configuration-operator.v1.2.4
        version: 1.2.4
      - name: namespace-configuration-operator.v1.2.3
        version: 1.2.3
      - name: namespace-configuration-operator.v1.2.2
        version: 1.2.2
      - name: namespace-configuration-operator.v1.2.1
        version: 1.2.1
      - name: namespace-configuration-operator.v1.2.0
        version: 1.2.0
      - name: namespace-configuration-operator.v1.1.0
        version: 1.1.0
      - name: namespace-configuration-operator.v1.0.7
        version: 1.0.7
      - name: namespace-configuration-operator.v1.0.6
        version: 1.0.6
      - name: namespace-configuration-operator.v1.0.5
        version: 1.0.5
      - name: namespace-configuration-operator.v1.0.4
        version: 1.0.4
      - name: namespace-configuration-operator.v1.0.3
        version: 1.0.3
      - name: namespace-configuration-operator.v1.0.2
        version: 1.0.2
      - name: namespace-configuration-operator.v1.0.1
        version: 1.0.1
      - name: namespace-configuration-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: namespace-configuration-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: geunwookim
      provider-url: ""
    name: ncn-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ncn-operator.v0.3.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-1"
                },
                "spec": null
              },
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "label-1"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
          - description: Label is the Schema for the labels API
            displayName: Label
            kind: Label
            name: labels.cxl.antline.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: jwahn@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Antline
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/rasen708/ncn-operator:v0.4.0
        version: 0.3.18
      entries:
      - name: ncn-operator.v0.3.18
        version: 0.3.18
      - name: ncn-operator.v0.3.14
        version: 0.3.14
      name: alpha
    - currentCSV: ncn-operator.v0.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-1"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-19T01:21:08Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: jwahn@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        provider:
          name: Antline
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/rasen708/ncn-operator:v0.2.2
        version: 0.2.2
      entries:
      - name: ncn-operator.v0.2.2
        version: 0.2.2
      - name: ncn-operator.v0.2.1
        version: 0.2.1
      - name: ncn-operator.v0.2.0
        version: 0.2.0
      name: beta
    - currentCSV: ncn-operator.v0.9.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cmmd.gw.com/v1",
                "kind": "CMMD",
                "metadata": {
                  "name": "my-app",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "allocateMode": "auto",
                  "enable": true,
                  "payload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "metadata": {
                      "name": "my-app"
                    },
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "name": "my-app"
                        }
                      },
                      "template": {
                        "metadata": {
                          "labels": {
                            "name": "my-app"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "quay.io/my-account/my-app:latest",
                              "name": "my-app-01"
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "cmmd.gw.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "machineconfiguration.openshift.io/role": "cmmd"
                  },
                  "name": "cmmd-node-label",
                  "namespace": "socmmd"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          console.openshift.io/plugins: '["socmmd-console-plugin"]'
          containerImage: quay.io/rasen708/ncn-operator@sha256:a09963745bd9c1419532d42d91237b7496c7a78e6a1c460a8e714ea1a0ce4d45
          createdAt: "2024-08-01T05:45:38Z"
          description: Provides the ncn extend
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "cmmd.gw.com/v1",
              "kind": "Label",
              "metadata": {
                "name": "cmmd-node-label",
                "namespace": "socmmd",
                "labels": {}
              },
              "spec": {}
            }
          operatorframework.io/suggested-namespace: socmmd-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/geunwookim/ncn-operator
          support: geunwookim
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CMMD is the Schema for the cmmds API
            displayName: CMMD
            kind: CMMD
            name: cmmds.cmmd.gw.com
            version: v1
          - description: Label is the Schema for the labels API
            displayName: Node Label
            kind: Label
            name: labels.cmmd.gw.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: geunwookim@gmail.com
          name: geunwookim
        maturity: betav1
        provider:
          name: geunwookim
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/rasen708/ncn-operator@sha256:a09963745bd9c1419532d42d91237b7496c7a78e6a1c460a8e714ea1a0ce4d45
        version: 0.9.8
      entries:
      - name: ncn-operator.v0.9.8
        version: 0.9.8
      - name: ncn-operator.v0.9.2
        version: 0.9.2
      - name: ncn-operator.v0.8.1
        version: 0.8.1
      - name: ncn-operator.v0.5.6
        version: 0.5.6
      - name: ncn-operator.v0.4.3
        version: 0.4.3
      - name: ncn-operator.v0.4.2
        version: 0.4.2
      - name: ncn-operator.v0.3.18
        version: 0.3.18
      - name: ncn-operator.v0.3.14
        version: 0.3.14
      name: betav1
    defaultChannel: betav1
    packageName: ncn-operator
    provider:
      name: geunwookim
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CNCF
      provider-url: ""
    name: etcd
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: etcdoperator.v0.6.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          description: etcd is a distributed key value store providing a reliable
            way to store data across a cluster of machines.
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
        description: |
          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader. Your applications can read and write data into etcd.
          A simple use-case is to store database connection details or feature flags within etcd as key value pairs. These values can be watched, allowing your app to reconfigure itself when they change. Advanced uses take advantage of the consistency guarantees to implement database leader elections or do distributed locking across a cluster of workers.

          _The etcd Open Cloud Service is Public Alpha. The goal before Beta is to fully implement backup features._

          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` or with the API using the automatically generated Kubernetes Service.

          [Read the complete guide to using the etcd Open Cloud Service](https://coreos.com/tectonic/docs/latest/alm/etcd-ocs.html)

          ### Supported Features
          **High availability**
          Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.
          **Automated updates**
          Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.
          **Backups included**
          Coming soon, the ability to schedule backups to happen on or off cluster.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: support@coreos.com
          name: CoreOS, Inc
        maturity: alpha
        provider:
          name: CoreOS, Inc
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:bd944a211eaf8f31da5e6d69e8541e7cada8f16a9f7a5a570b22478997819943
        version: 0.6.1
      entries:
      - name: etcdoperator.v0.6.1
        version: 0.6.1
      name: alpha
    - currentCSV: etcdoperator.v0.9.4-clusterwide
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdCluster",
                "metadata": {
                  "name": "example",
                  "annotations": {
                    "etcd.database.coreos.com/scope": "clusterwide"
                  }
                },
                "spec": {
                  "size": 3,
                  "version": "3.2.13"
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdRestore",
                "metadata": {
                  "name": "example-etcd-cluster-restore"
                },
                "spec": {
                  "etcdCluster": {
                    "name": "example-etcd-cluster"
                  },
                  "backupStorageType": "S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdBackup",
                "metadata": {
                  "name": "example-etcd-cluster-backup"
                },
                "spec": {
                  "etcdEndpoints": ["<etcd-cluster-endpoints>"],
                  "storageType":"S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
          createdAt: "2019-02-28 01:03:00"
          description: Create and maintain highly-available etcd clusters on Kubernetes
          repository: https://github.com/coreos/etcd-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to backup an etcd cluster.
            displayName: etcd Backup
            kind: EtcdBackup
            name: etcdbackups.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to restore an etcd cluster from a backup.
            displayName: etcd Restore
            kind: EtcdRestore
            name: etcdrestores.etcd.database.coreos.com
            version: v1beta2
        description: |
          The etcd Operater creates and maintains highly-available etcd clusters on Kubernetes, allowing engineers to easily deploy and manage etcd clusters for their applications.

          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader.


          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` via port forwarding:

              $ kubectl --namespace default port-forward service/example-client 2379:2379
              $ etcdctl --endpoints http://127.0.0.1:2379 get /

          Or directly to the API using the automatically generated Kubernetes Service:

              $ etcdctl --endpoints http://example-client.default.svc:2379 get /

          Be sure to secure your etcd cluster (see Common Configurations) before exposing it outside of the namespace or cluster.


          ### Supported Features

          * **High availability** - Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.

          * **Automated updates** - Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.

          * **Backups included** - Create etcd backups and restore them through the etcd Operator.

          ### Common Configurations

          * **Configure TLS** - Specify [static TLS certs](https://github.com/coreos/etcd-operator/blob/master/doc/user/cluster_tls.md) as Kubernetes secrets.

          * **Set Node Selector and Affinity** - [Spread your etcd Pods](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-node-selector-and-anti-affinity-across-nodes) across Nodes and availability zones.

          * **Set Resource Limits** - [Set the Kubernetes limit and request](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-resource-requirement) values for your etcd Pods.

          * **Customize Storage** - [Set a custom StorageClass](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#custom-persistentvolumeclaim-definition) that you would like to use.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: etcd-dev@googlegroups.com
          name: etcd Community
        maturity: alpha
        provider:
          name: CNCF
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
        version: 0.9.4-clusterwide
      entries:
      - name: etcdoperator.v0.9.4-clusterwide
        version: 0.9.4-clusterwide
      - name: etcdoperator.v0.9.2-clusterwide
        version: 0.9.2-clusterwide
      - name: etcdoperator.v0.9.0
        version: 0.9.0
      name: clusterwide-alpha
    - currentCSV: etcdoperator.v0.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "size": 3,
                  "version": "3.2.13"
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdRestore",
                "metadata": {
                  "name": "example-etcd-cluster-restore"
                },
                "spec": {
                  "etcdCluster": {
                    "name": "example-etcd-cluster"
                  },
                  "backupStorageType": "S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdBackup",
                "metadata": {
                  "name": "example-etcd-cluster-backup"
                },
                "spec": {
                  "etcdEndpoints": ["<etcd-cluster-endpoints>"],
                  "storageType":"S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
          createdAt: "2019-02-28 01:03:00"
          description: Create and maintain highly-available etcd clusters on Kubernetes
          repository: https://github.com/coreos/etcd-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to backup an etcd cluster.
            displayName: etcd Backup
            kind: EtcdBackup
            name: etcdbackups.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to restore an etcd cluster from a backup.
            displayName: etcd Restore
            kind: EtcdRestore
            name: etcdrestores.etcd.database.coreos.com
            version: v1beta2
        description: |
          The etcd Operater creates and maintains highly-available etcd clusters on Kubernetes, allowing engineers to easily deploy and manage etcd clusters for their applications.

          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader.


          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` via port forwarding:

              $ kubectl --namespace default port-forward service/example-client 2379:2379
              $ etcdctl --endpoints http://127.0.0.1:2379 get /

          Or directly to the API using the automatically generated Kubernetes Service:

              $ etcdctl --endpoints http://example-client.default.svc:2379 get /

          Be sure to secure your etcd cluster (see Common Configurations) before exposing it outside of the namespace or cluster.


          ### Supported Features

          * **High availability** - Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.

          * **Automated updates** - Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.

          * **Backups included** - Create etcd backups and restore them through the etcd Operator.

          ### Common Configurations

          * **Configure TLS** - Specify [static TLS certs](https://github.com/coreos/etcd-operator/blob/master/doc/user/cluster_tls.md) as Kubernetes secrets.

          * **Set Node Selector and Affinity** - [Spread your etcd Pods](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-node-selector-and-anti-affinity-across-nodes) across Nodes and availability zones.

          * **Set Resource Limits** - [Set the Kubernetes limit and request](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-resource-requirement) values for your etcd Pods.

          * **Customize Storage** - [Set a custom StorageClass](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#custom-persistentvolumeclaim-definition) that you would like to use.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: etcd-dev@googlegroups.com
          name: etcd Community
        maturity: alpha
        provider:
          name: CNCF
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
        version: 0.9.4
      entries:
      - name: etcdoperator.v0.9.4
        version: 0.9.4
      - name: etcdoperator.v0.9.2
        version: 0.9.2
      - name: etcdoperator.v0.9.0
        version: 0.9.0
      name: singlenamespace-alpha
    defaultChannel: singlenamespace-alpha
    packageName: etcd
    provider:
      name: CNCF
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataStax
      provider-url: ""
    name: cass-operator-community
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cass-operator.v1.23.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cassandra.datastax.com/v1beta1",
                "kind": "CassandraDatacenter",
                "metadata": {
                  "name": "dc1"
                },
                "spec": {
                  "clusterName": "development",
                  "config": {
                    "cassandra-yaml": {
                      "authenticator": "PasswordAuthenticator",
                      "authorizer": "CassandraAuthorizer",
                      "num_tokens": 16,
                      "role_manager": "CassandraRoleManager"
                    },
                    "jvm-server-options": {
                      "initial_heap_size": "1G",
                      "max_heap_size": "1G"
                    }
                  },
                  "managementApiAuth": {
                    "insecure": {}
                  },
                  "podTemplateSpec": {
                    "spec": {
                      "containers": [
                        {
                          "name": "cassandra"
                        }
                      ],
                      "securityContext": {}
                    }
                  },
                  "racks": [
                    {
                      "name": "rack1"
                    },
                    {
                      "name": "rack2"
                    },
                    {
                      "name": "rack3"
                    }
                  ],
                  "resources": {
                    "requests": {
                      "cpu": "1000m",
                      "memory": "2Gi"
                    }
                  },
                  "serverType": "cassandra",
                  "serverVersion": "4.1.2",
                  "size": 3,
                  "storageConfig": {
                    "cassandraDataVolumeClaimSpec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "server-storage"
                    }
                  }
                }
              },
              {
                "apiVersion": "control.k8ssandra.io/v1alpha1",
                "kind": "CassandraTask",
                "metadata": {
                  "name": "example-cleanup"
                },
                "spec": {
                  "datacenter": {
                    "name": "dc2",
                    "namespace": "cass-operator"
                  },
                  "jobs": [
                    {
                      "command": "cleanup",
                      "name": "cleanup-run"
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database,Big Data,OpenShift Optional
          certified: "true"
          containerImage: cr.k8ssandra.io/k8ssandra/cass-operator:v1.23.2
          createdAt: "2025-03-20"
          description: |
            Simple provisioning, turn-key operations, and automated remediation of Apache Cassandra clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/k8ssandra/cass-operator
          support: DataStax
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CassandraDatacenter is the Schema for the cassandradatacenters
              API
            displayName: Cassandra Datacenter
            kind: CassandraDatacenter
            name: cassandradatacenters.cassandra.datastax.com
            version: v1beta1
          - description: CassandraTask is the Schema for the cassandrajobs API
            displayName: Cassandra Task
            kind: CassandraTask
            name: cassandratasks.control.k8ssandra.io
            version: v1alpha1
        description: "## Apache Cassandra\n\nThe Apache Cassandra database is the
          right choice when you need scalability and\nhigh availability without compromising
          performance. Linear scalability and\nproven fault-tolerance on commodity
          hardware or cloud infrastructure make it the\nperfect platform for mission-critical
          data. Cassandra's support for replicating\nacross multiple datacenters is
          best-in-class, providing lower latency for your\nusers and the peace of
          mind of knowing that you can survive regional outages.\n\n## DataStax Enterprise\n\nThe
          most advanced distribution of Apache Cassandra™ on the market, with the\nenterprise
          functionality needed for serious production systems and backed up and\nsupported
          by the best distributed-experts in the world. It's one platform for\nall
          types of applications anywhere, any cloud, any model: key-value, graph,\ntabular,
          JSON.\n\nDataStax Enterprise is a fully integrated and optimized database,
          with graph,\nanalytics, and search included, all with a unified security
          model. Simply put,\nit's the only database capable of meeting today's demanding
          requirements\n\n## Operator Details\n\n`cass-operator` is designed as a
          modular operator for Apache Cassandra and\nderived  distributions. Apache
          Cassandra is a distributed database consisting of\nmultiple nodes working
          in concert to store data and process queries along a\nnumber of fault domains.
          `cass-operator` has the deployment of a Cassandra\ncluster around the logical
          domain of a datacenter with the `CassandraDatacenter`\ncustom resource.\n
          \   \nUpon submission of one of these resources it handles provisioning
          the underlying\nstateful sets (analogous to C\\* logical racks), services,
          and configuration.\nAdditionally through monitoring pod state via Kubernetes
          callbacks it handles day to day\noperations such as restarting failed processes,
          scaling clusters up, and deploying\nconfiguration changes in a rolling,
          non-disruptive, fashion.\n    \nThis operator is designed to be `Namespace`
          scoped. A single Kubernetes cluster may\nbe running multiple instances of
          this operator, in separate namespaces, to support\na number of C\\* clusters
          and environments. Configuration is simple with the usage of\nYAML based
          overrides in the Custom Resource paired with an `init` container.\n    \nIn
          C\\* clusters ordering and timing of certain operations are important to
          keep the system\nevenly distributed. `cass-operator` takes advantage of
          a sidecar process within the\nmain container to handle the orchestration
          of starting our main C* process."
        displayName: DataStax Kubernetes Operator for Apache Cassandra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - k8ssandra
        - cass-operator
        - DataStax Enterprise
        - Apache Cassandra
        - DSE
        - C*
        - Cassandra
        links:
        - name: DataStax
          url: https://datastax.com
        - name: Operator GitHub Repo
          url: https://github.com/k8ssandra/cass-operator
        - name: Config Builder GitHub Repo
          url: https://github.com/datastax/cass-config-builder
        - name: Documentation [1]
          url: https://github.com/k8ssandra/cass-operator/tree/master/docs/user
        - name: Documentation [2]
          url: https://docs.datastax.com/en/cass-operator/doc/cass-operator/cassOperatorTOC.html
        maintainers:
        - email: cass-operator@datastax.com
          name: DataStax Operator Team
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: DataStax
        relatedImages:
        - cr.k8ssandra.io/k8ssandra/cass-operator:v1.23.2
        version: 1.23.2
      entries:
      - name: cass-operator.v1.23.2
        version: 1.23.2
      - name: cass-operator.v1.23.1
        version: 1.23.1
      - name: cass-operator.v1.23.0
        version: 1.23.0
      - name: cass-operator.v1.21.1
        version: 1.21.1
      - name: cass-operator.v1.19.1
        version: 1.19.1
      - name: cass-operator.v1.17.1
        version: 1.17.1
      - name: cass-operator.v1.16.0
        version: 1.16.0
      - name: cass-operator.v1.14.0
        version: 1.14.0
      - name: cass-operator.v1.13.0
        version: 1.13.0
      - name: cass-operator.v1.10.4
        version: 1.10.4
      - name: cass-operator.v1.10.1
        version: 1.10.1
      name: stable
    defaultChannel: stable
    packageName: cass-operator-community
    provider:
      name: DataStax
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: openstack-ansibleee-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openstack-ansibleee-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ansibleee.openstack.org/v1beta1",
                "kind": "OpenStackAnsibleEE",
                "metadata": {
                  "name": "openstackansibleee-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-09-23T20:57:34Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.3
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenStackAnsibleEE is the Schema for the openstackansibleees
              API
            displayName: OpenStack Ansible EE
            kind: OpenStackAnsibleEE
            name: openstackansibleees.ansibleee.openstack.org
            version: v1beta1
        description: The OpenStack Ansible Execution Environment Operator
        displayName: OpenStackAnsibleEE
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - execution
        links:
        - name: OpenStack Ansible EE Operator
          url: https://github.com/openstack-k8s-operators/openstack-ansibleee-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:24b468b09b43dd6cf77f5b08a68f2c6359d115722d28982fd8cdb1923e8122de
        - registry.redhat.io/rhoso-operators/openstack-ansibleee-rhel9-operator@sha256:bf096518547e92fb6fe0ecd526813bdc4a15f928a6d89eacadb71ffaad91db69
        - registry.redhat.io/rhoso-operators/ee-openstack-ansible-ee-rhel9@sha256:640efe595fbdb993ecc1c55c14457edc2b0e3e1c6dde8ad9ea1acf4b4503339f
        version: 1.0.3
      entries:
      - name: openstack-ansibleee-operator.v1.0.3
        version: 1.0.3
      - name: openstack-ansibleee-operator.v1.0.2
        version: 1.0.2
      - name: openstack-ansibleee-operator.v1.0.1
        version: 1.0.1
      - name: openstack-ansibleee-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: openstack-ansibleee-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: ""
    name: red-hat-hawtio-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: red-hat-hawtio-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              },
              {
                "apiVersion": "hawt.io/v2",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24,
                    "internalSSL": true
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
          createdAt: "2025-08-07T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v2
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
        version: 1.2.1
      entries:
      - name: red-hat-hawtio-operator.v1.2.1
        version: 1.2.1
      - name: red-hat-hawtio-operator.v1.2.0-0.1754874242.p
        version: 1.2.0+0.1754874242.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754338551.p
        version: 1.2.0+0.1754338551.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754317392.p
        version: 1.2.0+0.1754317392.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1753863723.p
        version: 1.2.0+0.1753863723.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752679472.p
        version: 1.2.0+0.1752679472.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752452299.p
        version: 1.2.0+0.1752452299.p
      - name: red-hat-hawtio-operator.v1.2.0
        version: 1.2.0
      - name: red-hat-hawtio-operator.v1.1.1
        version: 1.1.1
      - name: red-hat-hawtio-operator.v1.1.0
        version: 1.1.0
      - name: red-hat-hawtio-operator.v1.1.0-0.1742200932.p
        version: 1.1.0+0.1742200932.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1740358952.p
        version: 1.1.0+0.1740358952.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739994212.p
        version: 1.1.0+0.1739994212.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739992853.p
        version: 1.1.0+0.1739992853.p
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: latest
    - currentCSV: red-hat-hawtio-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              },
              {
                "apiVersion": "hawt.io/v2",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24,
                    "internalSSL": true
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
          createdAt: "2025-08-07T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v2
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
        version: 1.2.1
      entries:
      - name: red-hat-hawtio-operator.v1.2.1
        version: 1.2.1
      - name: red-hat-hawtio-operator.v1.2.0-0.1754874242.p
        version: 1.2.0+0.1754874242.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754338551.p
        version: 1.2.0+0.1754338551.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754317392.p
        version: 1.2.0+0.1754317392.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1753863723.p
        version: 1.2.0+0.1753863723.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752679472.p
        version: 1.2.0+0.1752679472.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752452299.p
        version: 1.2.0+0.1752452299.p
      - name: red-hat-hawtio-operator.v1.2.0
        version: 1.2.0
      - name: red-hat-hawtio-operator.v1.1.1
        version: 1.1.1
      - name: red-hat-hawtio-operator.v1.1.0
        version: 1.1.0
      - name: red-hat-hawtio-operator.v1.1.0-0.1742200932.p
        version: 1.1.0+0.1742200932.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1740358952.p
        version: 1.1.0+0.1740358952.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739994212.p
        version: 1.1.0+0.1739994212.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739992853.p
        version: 1.1.0+0.1739992853.p
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: stable-v1
    - currentCSV: red-hat-hawtio-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel8-operator@sha256:f04d81d24c757db7b678464382ad8c5182b1580e714123a9f4f5fc9fbd9b8c57
          createdAt: "2024-01-30T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-rhel8-operator@sha256:f04d81d24c757db7b678464382ad8c5182b1580e714123a9f4f5fc9fbd9b8c57
        version: 1.0.3
      entries:
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1
    packageName: red-hat-hawtio-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MinIO
      provider-url: https://min.io
    name: minio-key-manager-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-key-manager-operator.v2025.8.19175300
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aistor.min.io/v1alpha1",
                "kind": "KeyManager",
                "metadata": {
                  "name": "my-kms",
                  "namespace": "ns-1"
                },
                "spec": {
                  "apiKeySecret": {
                    "name": "my-api-key"
                  },
                  "configuration": {
                    "name": "my-kms-server-config"
                  },
                  "hsmSecret": {
                    "name": "my-kms-hsm"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "registry-creds"
                    }
                  ],
                  "replicas": 2,
                  "volumeClaimTemplate": {
                    "metadata": {
                      "name": "key-manager-volume"
                    },
                    "spec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "100Mi"
                        }
                      },
                      "storageClassName": "standard"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security, Storage
          createdAt: "2025-08-27T14:50:59Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KeyManager
            name: keymanagers.aistor.min.io
            version: v1alpha1
        description: |-
          Highly available, powerful and operationally simple, MinIO's Enterprise Key Management Server is optimized for large storage infrastructures where billions of cryptographic keys are required.

          MinIO's AIStor KMS establishes its foundational trust using the concept of an hardware security module (but given KES is software, this is only a concept). That module assumes a pivotal role in sealing and unsealing the KMS root encryption key. The module responsibility extends to safeguarding the integrity of KMS by allowing the unsealing of its encrypted on-disk state and facilitating communication among nodes within a KMS cluster.

          It solves the challenges associated with billions of cryptographic keys and hundreds of thousands of cryptographic operations per node per second - which are commonplace in larger deployments.

          <b>High Availability and Fault Tolerance</b>
          In the dynamic landscape of large-scale systems, network or node outages are inevitable. Taking down a cluster for maintenance is rarely feasible. MinIO's AIStor KMS ensures uninterrupted availability, even when faced with such disruptions, mitigating cascading effects that can take down the entire storage infrastructure. Specifically, you could lose all but one node of a cluster and still handle any encryption, decryption or data key generation requests.

          <b>Scalability</b>
          While the amount of data usually only increases, the load on a large-scale storage system may vary significantly from time to time. MinIO's AIStor KMS supports dynamic cluster resizing and nodes can be added or removed at any point without incurring any downtime.

          <b>Multi-Tenancy</b>
          Large-scale storage infrastructures are often used by many applications and teams across the entire organization. Isolating teams and groups into their own namespaces is a core requirement. MinIO's AIStor KMS supports namespacing in the form of enclaves. Each tenant can be assigned its own enclave which is completely independent and isolated from all other enclaves on the KMS cluster.

          <b>Predictable Behavior</b>
          MinIO's AIStor KMS is designed to be easily managed, providing operators with the ability to comprehend its state intuitively. Due to its simple design, MinIO's AIStor KMS is significantly easier to operate than similar solutions that rely on more complex consensus algorithms like Raft, or Paxos.

          <b>Consistent and Performant</b>
          The responsiveness of the KMS for GET/PUT operations directly influences the overall efficiency and speed of the storage system. MinIO's AIStor KMS nodes don’t have to coordinate when handling such requests from the storage system. Therefore, the performance of a MinIO's AIStor KMS cluster increases linearly with the number of nodes. Further, MinIO's AIStor KMS supports request pipelining to handle hundreds of thousands of cryptographic operations per node and second.

          <b>Simplicity</b>
          Operating a KMS cluster does not require expertise in cryptography or distributed systems. Everything can be done from the AIStor Console.

          <a href="https://resources.min.io/request-a-demo-marketplaces/">Request a demo today</a>
        displayName: Minio AIStor Key management System (KMS)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Storage
        - MinIO
        links:
        - name: Minio AIStor
          url: https://min.io
        - name: Support
          url: https://subnet.min.io
        - name: Github
          url: https://github.com/minio/aistor
        maintainers:
        - email: dev@min.io
          name: MinIO
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: MinIO
          url: https://min.io
        relatedImages:
        - quay.io/minio/aistor/operator@sha256:f80139efa062fdd83f78318b20f371a827b68e854b03dd9706b9eba4c493aeb2
        version: 2025.8.19175300
      entries:
      - name: minio-key-manager-operator.v2025.8.19175300
        version: 2025.8.19175300
      - name: minio-key-manager-operator.v2025.8.1184927
        version: 2025.8.1184927
      - name: minio-key-manager-operator.v2025.7.1011319
        version: 2025.7.1011319
      - name: minio-key-manager-operator.v2025.6.3065135
        version: 2025.6.3065135
      - name: minio-key-manager-operator.v2025.5.12190907
        version: 2025.5.12190907
      name: stable
    defaultChannel: stable
    packageName: minio-key-manager-operator
    provider:
      name: MinIO
      url: https://min.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator-certified.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a",
                  "imagePullPolicy": "Always",
                  "initImage": "registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:3c03fb18a5d73876576b80aca2b61e7c84ca052839ccb467d14e15f6d1f9d834
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a57926b53eef862b70a53818a7ccb0e9756f878f30d4128b608090b109bdc585
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator-certified.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator-certified.v1.20.0
        version: 1.20.0
      - name: percona-server-mongodb-operator-certified.v1.19.1
        version: 1.19.1
      - name: percona-server-mongodb-operator-certified.v1.19.0
        version: 1.19.0
      - name: percona-server-mongodb-operator-certified.v1.18.0
        version: 1.18.0
      - name: percona-server-mongodb-operator-certified.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator-certified
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Cloudera
      provider-url: https://cloudera.com
    name: cloudera-streams-messaging-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster"
                },
                "spec": {
                  "kafka": {
                    "version": "3.9.0.1.4",
                    "replicas": 3,
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2,
                      "inter.broker.protocol.version": "3.9"
                    },
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "zookeeper": {
                    "replicas": 3,
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "3.9.0.1.4",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
          createdAt: "2025-07-14 10:09:41"
          description: Run Apache Kafka with Cloudera Streams Messaging - Kubernetes
            Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: A valid license and current subscription
            for Cloudera Streaming - Kubernetes Operator is required. Contact Cloudera
            sales to obtain a license. https://www.cloudera.com/contact-sales.html
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |-
          Cloudera Streams Messaging - Kubernetes Operator brings enterprise-grade Kafka deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of Kafka for enterprise deployments. To learn more about Cloudera Streaming, [read this deep dive blog post](https://blog.cloudera.com/accelerating-deployments-of-streaming-pipelines-announcing-data-in-motion-on-kubernetes/).

          Cloudera Streams Messaging - Kubernetes Operator consists of and ships multiple components including Apache Kafka, Apache ZooKeeper, Cruise Control, Strimzi, and others.

          ## Installation prerequisites

          Installing Cloudera Streams Messaging - Kubernetes Operator requires a valid license as well as Cloudera credentials (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly. The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) where installation artifacts are pulled from.

          ### Obtaining a license

          Contact [Cloudera Sales](https://www.cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets

          Before installation, create  two `Secrets`. One containing your Cloudera credentials, and one containing your license. Cloudera credentials are provided to you together with your license.

          Create a `Secret` containing your Cloudera credentials.
          ```
          oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***] --docker-server=container.repository.cloudera.com
          ```
          Create a `Secret` containing your license.
          ```
          oc create secret generic csm-op-license --from-file=license.txt=[***PATH TO LICENSE FILE***]
          ```

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces where you plan to deploy Kafka or Kafka Connect clusters. Cloudera recommends that you create the `Secret` in all required namespaces before installation if you know what namespaces you will be using to deploy Kafka or Kafka Connect.

          ## Learn more
          To learn more, visit the [Cloudera Streams Messaging - Kubernetes Operator documentation library](https://docs.cloudera.com/csm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Streams Messaging - Kubernetes Operator, visit the following resources:

          * [Strimzi](https://strimzi.io/)
          * [Apache Kafka](https://kafka.apache.org/)
          * [Apache ZooKeeper](https://zookeeper.apache.org/)
          * [Cruise Control](https://github.com/linkedin/cruise-control)
        displayName: Cloudera Streams Messaging - Kubernetes Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Documentation
          url: https://docs.cloudera.com/csm-operator
        - name: Cloudera Community Forum
          url: https://community.cloudera.com/
        - name: Cloudera Stream Processing
          url: https://www.cloudera.com/products/stream-processing.html
        - name: Contact Cloudera
          url: https://www.cloudera.com/contact-sales.html
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
        - container.repository.cloudera.com/cloudera/kafka@sha256:aa64e161b9672153a6f6559f9ae9ef88d0e40d6040a80e21bd3dc8ee39df6523
        - container.repository.cloudera.com/cloudera/kafka@sha256:b675e7426a558c4681e53d8e03f278264d7ad555f7c02422017683a29708ac2b
        - container.repository.cloudera.com/cloudera/maven-builder@sha256:9f18203b175193e5079fdc0c2b71c14887641d997f92b4a1c918546f777166df
        version: 1.4.0-b199
      entries:
      - name: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
        version: 1.4.0-b199
      name: 1.2.x
    - currentCSV: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster"
                },
                "spec": {
                  "kafka": {
                    "version": "3.9.0.1.4",
                    "replicas": 3,
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2,
                      "inter.broker.protocol.version": "3.9"
                    },
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "zookeeper": {
                    "replicas": 3,
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "3.9.0.1.4",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
          createdAt: "2025-07-14 10:09:41"
          description: Run Apache Kafka with Cloudera Streams Messaging - Kubernetes
            Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: A valid license and current subscription
            for Cloudera Streaming - Kubernetes Operator is required. Contact Cloudera
            sales to obtain a license. https://www.cloudera.com/contact-sales.html
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |-
          Cloudera Streams Messaging - Kubernetes Operator brings enterprise-grade Kafka deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of Kafka for enterprise deployments. To learn more about Cloudera Streaming, [read this deep dive blog post](https://blog.cloudera.com/accelerating-deployments-of-streaming-pipelines-announcing-data-in-motion-on-kubernetes/).

          Cloudera Streams Messaging - Kubernetes Operator consists of and ships multiple components including Apache Kafka, Apache ZooKeeper, Cruise Control, Strimzi, and others.

          ## Installation prerequisites

          Installing Cloudera Streams Messaging - Kubernetes Operator requires a valid license as well as Cloudera credentials (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly. The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) where installation artifacts are pulled from.

          ### Obtaining a license

          Contact [Cloudera Sales](https://www.cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets

          Before installation, create  two `Secrets`. One containing your Cloudera credentials, and one containing your license. Cloudera credentials are provided to you together with your license.

          Create a `Secret` containing your Cloudera credentials.
          ```
          oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***] --docker-server=container.repository.cloudera.com
          ```
          Create a `Secret` containing your license.
          ```
          oc create secret generic csm-op-license --from-file=license.txt=[***PATH TO LICENSE FILE***]
          ```

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces where you plan to deploy Kafka or Kafka Connect clusters. Cloudera recommends that you create the `Secret` in all required namespaces before installation if you know what namespaces you will be using to deploy Kafka or Kafka Connect.

          ## Learn more
          To learn more, visit the [Cloudera Streams Messaging - Kubernetes Operator documentation library](https://docs.cloudera.com/csm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Streams Messaging - Kubernetes Operator, visit the following resources:

          * [Strimzi](https://strimzi.io/)
          * [Apache Kafka](https://kafka.apache.org/)
          * [Apache ZooKeeper](https://zookeeper.apache.org/)
          * [Cruise Control](https://github.com/linkedin/cruise-control)
        displayName: Cloudera Streams Messaging - Kubernetes Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Documentation
          url: https://docs.cloudera.com/csm-operator
        - name: Cloudera Community Forum
          url: https://community.cloudera.com/
        - name: Cloudera Stream Processing
          url: https://www.cloudera.com/products/stream-processing.html
        - name: Contact Cloudera
          url: https://www.cloudera.com/contact-sales.html
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/kafka@sha256:aa64e161b9672153a6f6559f9ae9ef88d0e40d6040a80e21bd3dc8ee39df6523
        - container.repository.cloudera.com/cloudera/kafka@sha256:b675e7426a558c4681e53d8e03f278264d7ad555f7c02422017683a29708ac2b
        - container.repository.cloudera.com/cloudera/maven-builder@sha256:9f18203b175193e5079fdc0c2b71c14887641d997f92b4a1c918546f777166df
        - container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
        version: 1.4.0-b199
      entries:
      - name: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
        version: 1.4.0-b199
      name: stable
    defaultChannel: stable
    packageName: cloudera-streams-messaging-kubernetes-operator
    provider:
      name: Cloudera
      url: https://cloudera.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: vertical-pod-autoscaler
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: verticalpodautoscaler.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "autoscaling.openshift.io/v1",
                "kind": "VerticalPodAutoscalerController",
                "metadata": {
                  "name": "default",
                  "namespace": "openshift-vertical-pod-autoscaler"
                },
                "spec": {
                  "podMinCPUMillicores": 25,
                  "podMinMemoryMb": 250,
                  "recommendationOnly": false,
                  "safetyMarginFraction": 0.15,
                  "minReplicas": 2
                }
              },
              {
                 "apiVersion": "autoscaling.k8s.io/v1",
                 "kind": "VerticalPodAutoscaler",
                 "metadata": {
                    "name": "myapp-vpa"
                 },
                 "spec": {
                    "targetRef": {
                       "apiVersion": "apps/v1",
                       "kind": "Deployment",
                       "name": "myapp-deployment"
                    },
                    "resourcePolicy": {
                       "containerPolicies": [
                          {
                             "containerName": "*",
                             "minAllowed": {
                                "cpu": "100m",
                                "memory": "50Mi"
                             },
                             "maxAllowed": {
                                "cpu": 1,
                                "memory": "500Mi"
                             },
                             "controlledResources": [
                                "cpu",
                                "memory"
                             ]
                          }
                       ]
                    }
                 }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          certifiedLevel: Primed
          containerImage: registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9-operator@sha256:edf76334695c9a4aee4a9c0297af40475a16dec7e35ec7113a35b8507724536e
          createdAt: 2019-10-14 12:17:13+00:00
          description: An operator to run the OpenShift Vertical Pod Autoscaler. Vertical
            Pod Autoscaler (VPA) can be configured to monitor a workload's resource
            utilization, and then adjust its CPU and memory limits by updating the
            pod (future) or restarting the pod with updated limits.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          healthIndex: B
          olm.skipRange: '>=4.5.0 <4.16.0-202509240123'
          operatorframework.io/suggested-namespace: openshift-vertical-pod-autoscaler
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/vertical-pod-autoscaler-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of the set of VPA controllers
            displayName: VPA Controller
            kind: VerticalPodAutoscalerController
            name: verticalpodautoscalercontrollers.autoscaling.openshift.io
            version: v1
          - description: An application whose resources are managed by the VPA
            displayName: VerticalPodAutoscaler
            kind: VerticalPodAutoscaler
            name: verticalpodautoscalers.autoscaling.k8s.io
            version: v1
          - description: Represents VPA checkpoints used by the VPA controllers to
              manage workloads
            displayName: VPA Checkpoint
            kind: VerticalPodAutoscalerCheckpoint
            name: verticalpodautoscalercheckpoints.autoscaling.k8s.io
            version: v1
        description: An operator to run the OpenShift Vertical Pod Autoscaler. Vertical
          Pod Autoscaler (VPA) can be configured to monitor a workload's resource
          utilization, and then adjust its CPU and memory limits by updating the pod
          (future) or restarting the pod with updated limits.
        displayName: VerticalPodAutoscaler
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Vertical Pod Autoscaler Documentation
          url: https://docs.openshift.com/container-platform/latest/nodes/pods/nodes-pods-vertical-autoscaler.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9-operator@sha256:edf76334695c9a4aee4a9c0297af40475a16dec7e35ec7113a35b8507724536e
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9@sha256:e6c8f34895184179bb87f1659c4ff30b96b131787fa7a64e7bc6b6adbb336c16
        version: 4.16.0-202509240123
      entries:
      - name: verticalpodautoscaler.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: verticalpodautoscaler.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: verticalpodautoscaler.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: verticalpodautoscaler.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: verticalpodautoscaler.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: verticalpodautoscaler.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: verticalpodautoscaler.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: verticalpodautoscaler.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: verticalpodautoscaler.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: verticalpodautoscaler.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: verticalpodautoscaler.v4.16.0-202501260334
        version: 4.16.0-202501260334
      - name: verticalpodautoscaler.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: verticalpodautoscaler.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: verticalpodautoscaler.v4.16.0-202412021735
        version: 4.16.0-202412021735
      - name: verticalpodautoscaler.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: verticalpodautoscaler.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: verticalpodautoscaler.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: verticalpodautoscaler.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: verticalpodautoscaler.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: verticalpodautoscaler.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: verticalpodautoscaler.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: verticalpodautoscaler.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: verticalpodautoscaler.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: verticalpodautoscaler.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: verticalpodautoscaler.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: verticalpodautoscaler.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: verticalpodautoscaler.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: verticalpodautoscaler.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: verticalpodautoscaler.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: verticalpodautoscaler.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: vertical-pod-autoscaler
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-businessautomation: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: businessautomation-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: businessautomation-operator.7.13.5-16
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.kiegroup.org/v2","kind":"KieApp","metadata":{"name":"rhpam-trial"},"spec":{"environment":"rhpam-trial"}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/rhpam-7/rhpam-rhel8-operator@sha256:3b4f420def9e5d2c768d5209264bd5f9af6d6a37b22ef9a988a167ff7d9959cf
          createdAt: "2025-07-30 19:54:27"
          description: Deploys and manages Red Hat Process Automation Manager and
            Red Hat Decision Manager environments.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Process Automation
            Manager", "Red Hat Decision Manager"]'
          repository: https://github.com/kiegroup/kie-cloud-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A project prescription running an RHPAM/RHDM environment.
            displayName: KieApp
            kind: KieApp
            name: kieapps.app.kiegroup.org
            version: v2
        description: |-
          Deploys and manages Red Hat Process Automation Manager and Red Hat Decision Manager environments.

          * **Red Hat Process Automation Manager** is a platform for developing containerized microservices and applications that automate business decisions and processes. It includes business process management (BPM), business rules management (BRM), and business resource optimization and complex event processing (CEP) technologies. It also includes a user experience platform to create engaging user interfaces for process and decision services with minimal coding.

          [See more](https://www.redhat.com/en/products/process-automation).
        displayName: Business Automation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kieapp
        - pam
        - decision
        - kie
        - cloud
        - bpm
        - process
        - automation
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-process-automation-manager
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_process_automation_manager/7.13/#category-deploying-red-hat-process-automation-manager-on-openshift
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-monitoring-rhel8@sha256:8d4481a11269565f1690ad007f152e24326bae4b2099cb26ddf7b21be67031e8
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:cb37360843e9b062c59755593d9d412e1bf2c1dc1d6902e8a65a2825812c0989
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhpam-7/rhpam-kieserver-rhel8@sha256:2526d12da63b67ae9ed9294c914e9406d833c090ffa777623303ea0be22755d8
        - registry.redhat.io/rhpam-7/rhpam-process-migration-rhel8@sha256:3d7bdbe38895e50549e97573bd40da47b11691932164db7dafff0155a051aab2
        - registry.redhat.io/rhpam-7/rhpam-controller-rhel8@sha256:3a901b54a9d35d38b5aca275df2160e3190a874e139fcb606b9969b38284b3dc
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        - registry.redhat.io/rhpam-7/rhpam-kieserver-rhel8@sha256:3fbef82d1c5da8da1606d1a2d5dcfd20000da574737eb32b772d4f0123b179cb
        - registry.redhat.io/rhpam-7/rhpam-smartrouter-rhel8@sha256:d1e5461db8983a13f98259152dc3cfa2aa677c29f799104190e5cefa83cf97b5
        - registry.redhat.io/rhpam-7/rhpam-dashbuilder-rhel8@sha256:46396dc8060f2f9cc6f87ad905ef55be5b8e5216e5162bdf54df98c9d70e29ef
        - registry.redhat.io/rhscl/mysql-80-rhel7@sha256:e4f535c78bc2fdb58914c52fb805185035ff31515452b4df714f9ab623ba383a
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-rhel8@sha256:228c6ab5b83e215a98f6eb76250feb2ae3096fd21398c44e5b2701257c9c3493
        - registry.redhat.io/rhpam-7/rhpam-smartrouter-rhel8@sha256:9d7efdd97f163770d8a84b1ce5236bd4f31508154dced1f11dc6efe75a7b5089
        - registry.redhat.io/rhscl/postgresql-13-rhel7@sha256:338dd09681dcf860e6a976a3bc48a1474c2376d698815e4eb611ae9c5aa68966
        - registry.redhat.io/rhpam-7/rhpam-rhel8-operator@sha256:3b4f420def9e5d2c768d5209264bd5f9af6d6a37b22ef9a988a167ff7d9959cf
        - registry.redhat.io/rhpam-7/rhpam-dashbuilder-rhel8@sha256:4edac282119898ab1395030eecaee584990064e88fc8e65fb999f19ea2ef3f8b
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-rhel8@sha256:1abbaba9d4b2ecaec38205bcccf5ef4c7b6388ed9725b4f4c6dc8dc664588eb1
        - registry.redhat.io/rhpam-7/rhpam-controller-rhel8@sha256:989dc6dc64a2d5b6723fb5c000b0411fbdc6f78c0056666d831ffeabeab4b612
        - registry.redhat.io/openshift4/ose-cli@sha256:c3eb8807caf73e30012ae53c6d4f5114471290a5596929c27acef5a699a5b813
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/rhpam-7/rhpam-process-migration-rhel8@sha256:bd5c6586b4859dec23ad9530ee98bab33497ab251d544f863fc0af01cbd45101
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:457fc73423fa94591bc7e35729a25c58d3465583afccf342fb338b6e11188575
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-monitoring-rhel8@sha256:32912270eba0f2464da88f315bde8bf1ef5bbded814db156811d7a415a458f86
        version: 7.13.5-16
      entries:
      - name: businessautomation-operator.7.13.5-16
        version: 7.13.5-16
      - name: businessautomation-operator.7.13.5-15
        version: 7.13.5-15
      - name: businessautomation-operator.7.13.5-14
        version: 7.13.5-14
      - name: businessautomation-operator.7.13.5-13
        version: 7.13.5-13
      - name: businessautomation-operator.7.13.5-12
        version: 7.13.5-12
      - name: businessautomation-operator.7.13.5-11
        version: 7.13.5-11
      - name: businessautomation-operator.7.13.5-10
        version: 7.13.5-10
      - name: businessautomation-operator.7.13.5-9
        version: 7.13.5-9
      - name: businessautomation-operator.7.13.5-8
        version: 7.13.5-8
      - name: businessautomation-operator.7.13.5-7
        version: 7.13.5-7
      - name: businessautomation-operator.7.13.5-6
        version: 7.13.5-6
      - name: businessautomation-operator.7.13.5-5
        version: 7.13.5-5
      - name: businessautomation-operator.7.13.5-4
        version: 7.13.5-4
      - name: businessautomation-operator.7.13.5-3
        version: 7.13.5-3
      - name: businessautomation-operator.7.13.5-2
        version: 7.13.5-2
      - name: businessautomation-operator.7.13.5-1
        version: 7.13.5-1
      - name: businessautomation-operator.7.13.4-4
        version: 7.13.4-4
      - name: businessautomation-operator.7.13.4-3
        version: 7.13.4-3
      - name: businessautomation-operator.7.13.4-2
        version: 7.13.4-2
      - name: businessautomation-operator.7.13.4-1
        version: 7.13.4-1
      - name: businessautomation-operator.7.13.3-5
        version: 7.13.3-5
      - name: businessautomation-operator.7.13.3-4
        version: 7.13.3-4
      - name: businessautomation-operator.7.13.3-3
        version: 7.13.3-3
      - name: businessautomation-operator.7.13.3-2
        version: 7.13.3-2
      - name: businessautomation-operator.7.13.3-1
        version: 7.13.3-1
      - name: businessautomation-operator.7.13.2-3
        version: 7.13.2-3
      - name: businessautomation-operator.7.13.2-2
        version: 7.13.2-2
      - name: businessautomation-operator.7.13.2-1
        version: 7.13.2-1
      - name: businessautomation-operator.7.13.1-2
        version: 7.13.1-2
      - name: businessautomation-operator.7.13.1-1
        version: 7.13.1-1
      - name: businessautomation-operator.7.13.0-1
        version: 7.13.0-1
      - name: businessautomation-operator.7.12.1-4
        version: 7.12.1-4
      - name: businessautomation-operator.7.12.1-3
        version: 7.12.1-3
      - name: businessautomation-operator.7.12.1-2
        version: 7.12.1-2
      - name: businessautomation-operator.7.12.1-1
        version: 7.12.1-1
      - name: businessautomation-operator.7.12.0-2
        version: 7.12.0-2
      - name: businessautomation-operator.7.12.0-1
        version: 7.12.0-1
      - name: businessautomation-operator.7.11.1-3
        version: 7.11.1-3
      - name: businessautomation-operator.7.11.1-2
        version: 7.11.1-2
      - name: businessautomation-operator.7.11.1-1
        version: 7.11.1-1
      - name: businessautomation-operator.7.11.0-2
        version: 7.11.0-2
      - name: businessautomation-operator.7.11.0-1
        version: 7.11.0-1
      - name: businessautomation-operator.7.10.1-2
        version: 7.10.1-2
      - name: businessautomation-operator.7.10.1-1
        version: 7.10.1-1
      - name: businessautomation-operator.7.10.0-2
        version: 7.10.0-2
      - name: businessautomation-operator.7.10.0-1
        version: 7.10.0-1
      - name: businessautomation-operator.7.9.1-1
        version: 7.9.1-1
      - name: businessautomation-operator.7.9.0
        version: 7.9.0
      - name: businessautomation-operator.7.9.0-2
        version: 7.9.0-2
      - name: businessautomation-operator.7.8.1
        version: 7.8.1
      - name: businessautomation-operator.7.8.0
        version: 7.8.0
      - name: businessautomation-operator.1.4.1
        version: 1.4.1
      - name: businessautomation-operator.1.4.0
        version: 1.4.0
      - name: businessautomation-operator.1.3.0
        version: 1.3.0
      - name: businessautomation-operator.1.2.1
        version: 1.2.1
      - name: businessautomation-operator.1.2.0
        version: 1.2.0
      name: stable
    defaultChannel: stable
    packageName: businessautomation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: quay-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:d73321dfa6dcced943405fae45c2ccf841663e0ab57e8be8cc1ddba251b7ce08
          createdAt: 2021-11-02 11:11 UTC
          description: Opinionated deployment of Red Hat Quay on Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:d73321dfa6dcced943405fae45c2ccf841663e0ab57e8be8cc1ddba251b7ce08
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:0e6bef51caf479074925115fa11621c41bcada375cc1925d0511a73084b93543
        - registry.redhat.io/rhel8/postgresql-10@sha256:b2695e9af379624621bef9ec785f0cb618bfdabae98c5f17eea82ff5f8d2d5cb
        - registry.redhat.io/rhel8/redis-5@sha256:8d7e66927ce4b3050ca15ab4df4dbaeb593cb855e16fb0a6f5dde950e9db7db9
        - registry.redhat.io/quay/quay-rhel8@sha256:ed8c4249d376b4226e35bf088ffe5bf688bf68b059718c5e218aa1184fe048d0
        - registry.redhat.io/quay/clair-rhel8@sha256:af5e2fd1543365b365b1a6cce00b86e44fe5e66cde6491e9b113324c21f629c8
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:2a471300f70d3995ab7e9fcc3df35c5513e73fcc1d140fb4d651a1c3eae949eb
        version: 3.4.7
      entries:
      - name: quay-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: quay-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:1f5e4cc9ec2d9b8b89d82f1a698acfa1ac2764ca4ebc5c85b3417da7730b5867
          createdAt: 2021-09-09 20:09 UTC
          description: Opinionated deployment of Red Hat Quay on Kubernetes.
          olm.skipRange: '>=3.4.x <3.5.7'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.5.7
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7fe408a1e428b2d0660cb578c1dcd4b961f838e37166fe7b21b2bed70d288594
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:5d72577ccfd63c6d9ad712b376b604c40d2d3ecb53e391e2ee0d18caa6e0225a
        - registry.redhat.io/rhel8/postgresql-10@sha256:a723bb3a7a5046527b9a5b628ddfb211c526e3814cc4242262c4f8a016f38edd
        - registry.redhat.io/rhel8/redis-5@sha256:7916760ff03af5ca230df5e96b2bf353fbf1829386de1c063c05f623837be7f7
        - registry.redhat.io/quay/quay-rhel8@sha256:c599892160ac20c744d833603d9375923af7aec1edfd86982b7513e02fb6d463
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:1f5e4cc9ec2d9b8b89d82f1a698acfa1ac2764ca4ebc5c85b3417da7730b5867
        - registry.redhat.io/quay/clair-rhel8@sha256:d242bd9d37f95b2a5a8267b3f1eba02c671d46121712baae43466c86d1da89f9
        version: 3.5.7
      entries:
      - name: quay-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: quay-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:3c0201e3cf88e711c921870324002a02cbd42c0d2eecafb4f242aa0601e24602
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.10.15
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:fc20b8bdac88c1bfce9aec993bdce6dd2cc132f8a9d4062d174b098819c74013
        - registry.redhat.io/quay/clair-rhel8@sha256:efa6e69500ea643042b544e6205f088214d79ab09a36864a612e73d07979d1c5
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:3c0201e3cf88e711c921870324002a02cbd42c0d2eecafb4f242aa0601e24602
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7e720c6a332ec4032e8576c31421af404f61d9baaf8e88f69f8333c52e5d1683
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:774cd1f181527787ba8631c7c14437eb2250c15ddb4a9c75fdfa35c5c1a19a2e
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        version: 3.10.15
      entries:
      - name: quay-operator.v3.10.15
        version: 3.10.15
      - name: quay-operator.v3.10.14
        version: 3.10.14
      - name: quay-operator.v3.10.13
        version: 3.10.13
      - name: quay-operator.v3.10.12
        version: 3.10.12
      - name: quay-operator.v3.10.11
        version: 3.10.11
      - name: quay-operator.v3.10.10
        version: 3.10.10
      - name: quay-operator.v3.10.9
        version: 3.10.9
      - name: quay-operator.v3.10.8
        version: 3.10.8
      - name: quay-operator.v3.10.7
        version: 3.10.7
      - name: quay-operator.v3.10.6
        version: 3.10.6
      - name: quay-operator.v3.10.5
        version: 3.10.5
      - name: quay-operator.v3.10.4
        version: 3.10.4
      - name: quay-operator.v3.10.3
        version: 3.10.3
      - name: quay-operator.v3.10.2
        version: 3.10.2
      - name: quay-operator.v3.10.1
        version: 3.10.1
      - name: quay-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:cb3a41caa3a572b7981d424be2a32b8e881127cc03b557ee0a36ab13f430b897
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.11.13
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:7230fa1d4970e877cb6de8716e99a9bd43302c0ff23b4501f2aa806a8d496d9b
        - registry.redhat.io/quay/clair-rhel8@sha256:3746677ef1cb0ddd4f237ab598b14cf0bc76dcce080ffde127b5bba87ffe6693
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:1a083cd72502609427428ed22fb60846613ddb41a4613496c5cddc797b382d3a
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:11b8c9d64c42f0f0c7b2df491bba859d0410666a8ef1d2fd41eb77e4ad9d2dc8
        - registry.redhat.io/rhel8/postgresql-13@sha256:35d86936c539007a78a89869a14b87b76788fe117df442a670f63936b978ed5b
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:cb3a41caa3a572b7981d424be2a32b8e881127cc03b557ee0a36ab13f430b897
        version: 3.11.13
      entries:
      - name: quay-operator.v3.11.13
        version: 3.11.13
      - name: quay-operator.v3.11.12
        version: 3.11.12
      - name: quay-operator.v3.11.11
        version: 3.11.11
      - name: quay-operator.v3.11.10
        version: 3.11.10
      - name: quay-operator.v3.11.9
        version: 3.11.9
      - name: quay-operator.v3.11.8
        version: 3.11.8
      - name: quay-operator.v3.11.7
        version: 3.11.7
      - name: quay-operator.v3.11.6
        version: 3.11.6
      - name: quay-operator.v3.11.5
        version: 3.11.5
      - name: quay-operator.v3.11.4
        version: 3.11.4
      - name: quay-operator.v3.11.3
        version: 3.11.3
      - name: quay-operator.v3.11.2
        version: 3.11.2
      - name: quay-operator.v3.11.1
        version: 3.11.1
      - name: quay-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:accb705b6327fb2dcc1ce657a99f61e343eca752c9873d63d89fea89c10c3187
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.12.12
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:accb705b6327fb2dcc1ce657a99f61e343eca752c9873d63d89fea89c10c3187
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:ccbd83e05aea0bb96bf099f63420175b36ed413bfc8a618f4c90b6a2abc53895
        - registry.redhat.io/quay/clair-rhel8@sha256:dc3d5fb35f2bd5d00480351dca705fe260a225d9a97958cf0276801af08d8307
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:8e5dd171e6e94de0137bb6a972533b1f93b56efc1731e19b80869d5e9b8a26a6
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:8cdd677449f3c723dad79bcd9c4191c8917d6eb0beeeeeda68fe19a0988bb2a8
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        version: 3.12.12
      entries:
      - name: quay-operator.v3.12.12
        version: 3.12.12
      - name: quay-operator.v3.12.11
        version: 3.12.11
      - name: quay-operator.v3.12.10
        version: 3.12.10
      - name: quay-operator.v3.12.9
        version: 3.12.9
      - name: quay-operator.v3.12.8
        version: 3.12.8
      - name: quay-operator.v3.12.7
        version: 3.12.7
      - name: quay-operator.v3.12.6
        version: 3.12.6
      - name: quay-operator.v3.12.5
        version: 3.12.5
      - name: quay-operator.v3.12.4
        version: 3.12.4
      - name: quay-operator.v3.12.3
        version: 3.12.3
      - name: quay-operator.v3.12.2
        version: 3.12.2
      - name: quay-operator.v3.12.1
        version: 3.12.1
      - name: quay-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:5481a4b914ce4aed6ba3ceee8432eddab6e264cad2caef528404795f904100eb
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.13.8
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/clair-rhel8@sha256:069ccf36346cb14f4e4ea54de1d59afb72548912b140cfef67a6bcb7534d25e9
        - registry.redhat.io/rhel8/postgresql-13@sha256:f17fd8dd4d56d8e916cb2ad3f403d05ad259f79fd6918d67a38a6e9e4a8896cf
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:966a3dafa2dd31ac1a39e826e08c1b988fe694189fe79449bbe14a48b2631213
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:5481a4b914ce4aed6ba3ceee8432eddab6e264cad2caef528404795f904100eb
        - registry.redhat.io/quay/quay-rhel8@sha256:8887c17547c7a4be61471523eade9820e5095cf966b1885a3758081da559871c
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:31a9095d8f291ce68e8e688877692e74a17143bdda8d998fe3a07bcee32575dd
        - registry.redhat.io/rhel8/postgresql-15@sha256:a5f36e7893a519c12752c6118233c92d4e83618670bc9ea7bd3a8a34a63bd50b
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:f62cbd6632454c227b7cbb01dadc8e501e8288b2d8e0abf8e7842b806fa3195c
        version: 3.13.8
      entries:
      - name: quay-operator.v3.13.8
        version: 3.13.8
      - name: quay-operator.v3.13.7
        version: 3.13.7
      - name: quay-operator.v3.13.6
        version: 3.13.6
      - name: quay-operator.v3.13.5
        version: 3.13.5
      - name: quay-operator.v3.13.4
        version: 3.13.4
      - name: quay-operator.v3.13.3
        version: 3.13.3
      - name: quay-operator.v3.13.2
        version: 3.13.2
      - name: quay-operator.v3.13.1
        version: 3.13.1
      - name: quay-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:8ebc1f0511f7d394ceaa39a4c993e955c560dc04588072cb3f8c8fab8c85e32b
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.14.5
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/clair-rhel8@sha256:e11e8a819d17bd14c67f093b5303b4958855e731f89ce67dcca70fcb7b77f45a
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:8ebc1f0511f7d394ceaa39a4c993e955c560dc04588072cb3f8c8fab8c85e32b
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-rhel8@sha256:33e984892650d315065313fcb5f8808e145aadec5a287219c7031d85b713235e
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:43ef8adddbf9b007b6e2396f53bd6a5475150d28c2f4254f031a0fe6c7de5826
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:a12a0918f6dea114b0aae231d7119a4acfd0cd297074d895f4ecf190d2aec43a
        version: 3.14.5
      entries:
      - name: quay-operator.v3.14.5
        version: 3.14.5
      - name: quay-operator.v3.14.4
        version: 3.14.4
      - name: quay-operator.v3.14.3
        version: 3.14.3
      - name: quay-operator.v3.14.2
        version: 3.14.2
      - name: quay-operator.v3.14.1
        version: 3.14.1
      - name: quay-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:05fb981d9848281b97b3683093f752b1c1dc5ced439a2b4ba18b26ff8e11231e
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.15.2
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:873a3daf1dadf2438eb96e151072326c31ba34ed48cac041614b604413d5f8e7
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:956f3110f944d880bc27bb6df9472d0804cd75f5dcf4ed11bcf721e9d1d4542b
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:aa8c9f7b6ce57441ca1fe9dbaaefaf2bb83fd44aa367647b663bb822bf890362
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:05fb981d9848281b97b3683093f752b1c1dc5ced439a2b4ba18b26ff8e11231e
        - registry.redhat.io/quay/clair-rhel8@sha256:861be31d30142bae125cd059bb716d79b28694e9c4336333c68adf4a3823b61d
        version: 3.15.2
      entries:
      - name: quay-operator.v3.15.2
        version: 3.15.2
      - name: quay-operator.v3.15.1
        version: 3.15.1
      - name: quay-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-operator.v3.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:7a6e592d7f5f760b4451f9c974dbb96f56b1d3de92f039adcbf284f78ba13a02
          createdAt: 2021-10-12 22:10 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.6.10
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:ec36241f0a9a2983b1894cd8501898e6b70379af51a81b1e09b0ba0e9b54b33f
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:75df6b8e6bef14844a08f7d994470fa6ec4ffa66d4b15faaa0a3ba373f612fbb
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:7a6e592d7f5f760b4451f9c974dbb96f56b1d3de92f039adcbf284f78ba13a02
        - registry.redhat.io/rhel8/postgresql-10@sha256:4b912c80085b88a03309aeb7907efcc29dd3342fa3952b6ea067afb1914bfe53
        - registry.redhat.io/rhel8/redis-6@sha256:d7c7852338717308cbb59e9303e1ea35cc8e5c01ceb2818569be20c15f3f943d
        - registry.redhat.io/quay/quay-rhel8@sha256:e75e4d084eba9bb65b9cc72e38294af314d49316b12a0beb75db808dc057830b
        - registry.redhat.io/quay/clair-rhel8@sha256:cbafddd1eb50c1c0f9159ffafd86f44a8c0b976194903c96c6f144bfe62e749d
        version: 3.6.10
      entries:
      - name: quay-operator.v3.6.10
        version: 3.6.10
      - name: quay-operator.v3.6.9
        version: 3.6.9
      - name: quay-operator.v3.6.8
        version: 3.6.8
      - name: quay-operator.v3.6.7
        version: 3.6.7
      - name: quay-operator.v3.6.6
        version: 3.6.6
      - name: quay-operator.v3.6.5
        version: 3.6.5
      - name: quay-operator.v3.6.4
        version: 3.6.4
      - name: quay-operator.v3.6.2
        version: 3.6.2
      - name: quay-operator.v3.6.1
        version: 3.6.1
      - name: quay-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: quay-operator.v3.7.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:a8283c2be661e07c5f2f2e5b683c520846bcb91d580d9a3a0aa27ce994ddfff6
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.7.14
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:b15335d88b240bd6bbedeb7fe26f0a1de2f6c330af9758b0fb2edab7ca69f6df
        - registry.redhat.io/quay/quay-rhel8@sha256:7bc9e115e9bace07505ffd225242cab83d5eb6d50358188d18fd9161bdf1a5a6
        - registry.redhat.io/quay/clair-rhel8@sha256:d676af83e96740220e6f18721c98817ba071fd023a62ebdc3815424e94838083
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:c54c77e09c95dc7d80d7b31e796857c8d06f2f733426629b53d3298206d23b20
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:04ac405c152bffb43b33fb2e7f45df8b3940eb21e055ebf6aa17bd81ebdeeab4
        - registry.redhat.io/rhel8/postgresql-10@sha256:a52ad402458ec8ef3f275972c6ebed05ad64398f884404b9bb8e3010c5c95291
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:a8283c2be661e07c5f2f2e5b683c520846bcb91d580d9a3a0aa27ce994ddfff6
        version: 3.7.14
      entries:
      - name: quay-operator.v3.7.14
        version: 3.7.14
      - name: quay-operator.v3.7.13
        version: 3.7.13
      - name: quay-operator.v3.7.12
        version: 3.7.12
      - name: quay-operator.v3.7.11
        version: 3.7.11
      - name: quay-operator.v3.7.10
        version: 3.7.10
      name: stable-3.7
    - currentCSV: quay-operator.v3.8.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:abec0fa7ee641c8e424e995b5fe07b98e522ad1eb92c3b90426f69941f33eb2e
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.8.15
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7e4e57071439d671971b0de46b597cf6cff8ef3879ab711125a6ae7ecb9c6d29
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:3c047c7b22fb42557bd3a9b9c9ac333f0df7d5e4a4d333a2b02a4313f57085be
        - registry.redhat.io/rhel8/postgresql-10@sha256:b7fa644277d5e406b642685ff2ace1b1de7aea2e2a1421b3cbcda7d60405f44d
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:abec0fa7ee641c8e424e995b5fe07b98e522ad1eb92c3b90426f69941f33eb2e
        - registry.redhat.io/rhel8/redis-6@sha256:444200241cb9186cd97d564d7acf077d83fc997a8ee5ee7fdae41b604895085c
        - registry.redhat.io/quay/quay-rhel8@sha256:de2b65506f99eebb5eaafca6bb42b4ebec08e9479dd0d50aa1286f197e42d6de
        - registry.redhat.io/quay/clair-rhel8@sha256:b327fcfd7472cdb9d83a564904976b26b285fa576dc0a76e5590990ec77fd871
        version: 3.8.15
      entries:
      - name: quay-operator.v3.8.15
        version: 3.8.15
      - name: quay-operator.v3.8.14
        version: 3.8.14
      - name: quay-operator.v3.8.13
        version: 3.8.13
      - name: quay-operator.v3.8.12
        version: 3.8.12
      - name: quay-operator.v3.8.11
        version: 3.8.11
      - name: quay-operator.v3.8.10
        version: 3.8.10
      - name: quay-operator.v3.8.9
        version: 3.8.9
      - name: quay-operator.v3.8.8
        version: 3.8.8
      - name: quay-operator.v3.8.7
        version: 3.8.7
      - name: quay-operator.v3.8.6
        version: 3.8.6
      - name: quay-operator.v3.8.5
        version: 3.8.5
      - name: quay-operator.v3.8.4
        version: 3.8.4
      - name: quay-operator.v3.8.3
        version: 3.8.3
      - name: quay-operator.v3.8.2
        version: 3.8.2
      - name: quay-operator.v3.8.1
        version: 3.8.1
      - name: quay-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:8a16803946272156b7aeb00eb2202ab6cb87d875fe4a7c327554f770f0f2ecb5
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.9.16
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/clair-rhel8@sha256:6c67163c783d7d9fa4d187963379885a8b76685347ecf1894c4f8ebd6656544b
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:af8ada0c88b6f19991e61ee86d04b49a148c19dacb57c282e4ab98630e7499b4
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:5a42689198701d845ff3f05502cbbebceb7e6ad7b12762ccc038c0131cf6479e
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:f87b7e7ee6c31f6ae80b08aef0e10729b21fe1b8656fe18f6dee2310bae28ec1
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:8a16803946272156b7aeb00eb2202ab6cb87d875fe4a7c327554f770f0f2ecb5
        version: 3.9.16
      entries:
      - name: quay-operator.v3.9.16
        version: 3.9.16
      - name: quay-operator.v3.9.15
        version: 3.9.15
      - name: quay-operator.v3.9.14
        version: 3.9.14
      - name: quay-operator.v3.9.13
        version: 3.9.13
      - name: quay-operator.v3.9.12
        version: 3.9.12
      - name: quay-operator.v3.9.11
        version: 3.9.11
      - name: quay-operator.v3.9.10
        version: 3.9.10
      - name: quay-operator.v3.9.9
        version: 3.9.9
      - name: quay-operator.v3.9.8
        version: 3.9.8
      - name: quay-operator.v3.9.7
        version: 3.9.7
      - name: quay-operator.v3.9.6
        version: 3.9.6
      - name: quay-operator.v3.9.5
        version: 3.9.5
      - name: quay-operator.v3.9.4
        version: 3.9.4
      - name: quay-operator.v3.9.3
        version: 3.9.3
      - name: quay-operator.v3.9.2
        version: 3.9.2
      - name: quay-operator.v3.9.1
        version: 3.9.1
      - name: quay-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: quay-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: glance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: glance-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "glance.openstack.org/v1beta1",
                "kind": "Glance",
                "metadata": {
                  "name": "glance"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\nenabled_backends=foo:bar,foo1:bar1\n",
                  "databaseAccount": "glance",
                  "databaseInstance": "openstack",
                  "secret": "osp-secret",
                  "serviceUser": "glance",
                  "storage": {
                    "storageClass": "local-storage",
                    "storageRequest": "10G"
                  }
                }
              },
              {
                "apiVersion": "glance.openstack.org/v1beta1",
                "kind": "GlanceAPI",
                "metadata": {
                  "annotation": {
                    "keystoneEndpoint": "true"
                  },
                  "name": "default"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\nenabled_backends=foo:bar,foo1:bar1\n[foo]\nhere-foo-config\n[foo1]\nhere-foo1-config\n",
                  "databaseAccount": "glance",
                  "databaseHostname": "glance",
                  "preserveJobs": false,
                  "replicas": 1,
                  "secret": "glance-secret",
                  "serviceAccount": "glance",
                  "serviceUser": "glance",
                  "storage": {
                    "storageRequest": "10G"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:45Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GlanceAPI is the Schema for the glanceapis API
            displayName: Glance API
            kind: GlanceAPI
            name: glanceapis.glance.openstack.org
            version: v1beta1
          - description: Glance is the Schema for the glances API
            displayName: Glance
            kind: Glance
            name: glances.glance.openstack.org
            version: v1beta1
        description: Glance Operator
        displayName: Glance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Image
        - Glance
        links:
        - name: Glance Operator
          url: https://github.com/openstack-k8s-operators/glance-operator
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/glance-rhel9-operator@sha256:74f9907b59493239fbe9c5f534399ff513c693a2bda11a037aa2cc5dc8cc3e91
        - registry.redhat.io/rhoso/openstack-glance-api-rhel9@sha256:b8f84f9b432a70f7ca7aebda60455a4890347529f6bed9a440a3b197e377e0a1
        version: 1.0.6
      entries:
      - name: glance-operator.v1.0.6
        version: 1.0.6
      - name: glance-operator.v1.0.5
        version: 1.0.5
      - name: glance-operator.v1.0.4
        version: 1.0.4
      - name: glance-operator.v1.0.3
        version: 1.0.3
      - name: glance-operator.v1.0.2
        version: 1.0.2
      - name: glance-operator.v1.0.1
        version: 1.0.1
      - name: glance-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: glance-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kafka-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kafka-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kafka-controller:1.2.1
          createdAt: "2025-09-20T21:41:07Z"
          description: AWS Kafka controller is a service controller for managing Kafka
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster represents the state of an AWS kafka Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.kafka.services.k8s.aws
            version: v1alpha1
          - description: Configuration represents the state of an AWS kafka Configuration
              resource.
            displayName: Configuration
            kind: Configuration
            name: configurations.kafka.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCluster represents the state of an AWS kafka ServerlessCluster
              resource.
            displayName: ServerlessCluster
            kind: ServerlessCluster
            name: serverlessclusters.kafka.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Kafka resources in AWS from within your Kubernetes cluster.

          **About Amazon Kafka**

          Amazon Managed Streaming for Apache Kafka (Amazon MSK) is a fully managed AWS streaming data service makes it easy for you to build and run applications that use Apache Kafka to process streaming data without needing to become an expert in operating Apache Kafka clusters. Amazon MSK manages the provisioning, configuration, and maintenance of Apache Kafka clusters and Apache ZooKeeper nodes for you. Amazon MSK also shows key Apache Kafka performance metrics in the AWS Console.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Kafka Developer Resources
          url: https://aws.amazon.com/msk/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kafka maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/kafka-controller:1.2.1
        version: 1.2.1
      entries:
      - name: ack-kafka-controller.v1.2.1
        version: 1.2.1
      - name: ack-kafka-controller.v1.2.0
        version: 1.2.0
      - name: ack-kafka-controller.v1.1.3
        version: 1.1.3
      - name: ack-kafka-controller.v1.1.2
        version: 1.1.2
      - name: ack-kafka-controller.v1.1.1
        version: 1.1.1
      - name: ack-kafka-controller.v1.1.0
        version: 1.1.0
      - name: ack-kafka-controller.v1.0.14
        version: 1.0.14
      - name: ack-kafka-controller.v1.0.13
        version: 1.0.13
      - name: ack-kafka-controller.v1.0.12
        version: 1.0.12
      - name: ack-kafka-controller.v1.0.11
        version: 1.0.11
      - name: ack-kafka-controller.v1.0.10
        version: 1.0.10
      - name: ack-kafka-controller.v1.0.8
        version: 1.0.8
      - name: ack-kafka-controller.v1.0.7
        version: 1.0.7
      - name: ack-kafka-controller.v1.0.6
        version: 1.0.6
      - name: ack-kafka-controller.v1.0.5
        version: 1.0.5
      - name: ack-kafka-controller.v1.0.4
        version: 1.0.4
      - name: ack-kafka-controller.v1.0.3
        version: 1.0.3
      - name: ack-kafka-controller.v1.0.2
        version: 1.0.2
      - name: ack-kafka-controller.v1.0.1
        version: 1.0.1
      - name: ack-kafka-controller.v1.0.0
        version: 1.0.0
      - name: ack-kafka-controller.v0.0.14
        version: 0.0.14
      - name: ack-kafka-controller.v0.0.12
        version: 0.0.12
      - name: ack-kafka-controller.v0.0.11
        version: 0.0.11
      - name: ack-kafka-controller.v0.0.10
        version: 0.0.10
      - name: ack-kafka-controller.v0.0.9
        version: 0.0.9
      - name: ack-kafka-controller.v0.0.8
        version: 0.0.8
      - name: ack-kafka-controller.v0.0.7
        version: 0.0.7
      - name: ack-kafka-controller.v0.0.6
        version: 0.0.6
      - name: ack-kafka-controller.v0.0.5
        version: 0.0.5
      - name: ack-kafka-controller.v0.0.4
        version: 0.0.4
      - name: ack-kafka-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-kafka-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-acmpca-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-acmpca-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acmpca.services.k8s.aws/v1alpha1",
                "kind": "CertificateAuthority",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/acmpca-controller:1.1.1
          createdAt: "2025-09-19T20:51:07Z"
          description: AWS ACM PCA controller is a service controller for managing
            ACM PCA resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CertificateAuthority represents the state of an AWS acmpca
              CertificateAuthority resource.
            displayName: CertificateAuthority
            kind: CertificateAuthority
            name: certificateauthorities.acmpca.services.k8s.aws
            version: v1alpha1
          - description: CertificateAuthorityActivation represents the state of an
              AWS acmpca CertificateAuthorityActivation resource.
            displayName: CertificateAuthorityActivation
            kind: CertificateAuthorityActivation
            name: certificateauthorityactivations.acmpca.services.k8s.aws
            version: v1alpha1
          - description: Certificate represents the state of an AWS acmpca Certificate
              resource.
            displayName: Certificate
            kind: Certificate
            name: certificates.acmpca.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ACM PCA resources in AWS from within your Kubernetes cluster.

          **About Amazon ACM PCA**

          AWS Private CA enables creation of private certificate authority (CA) hierarchies, including root and subordinate CAs, without the investment and maintenance costs of operating an on-premises CA. Your private CAs can issue end-entity X.509 certificates useful in scenarios including:
          - Creating encrypted TLS communication channels
          - Authenticating users, computers, API endpoints, and IoT devices
          - Cryptographically signing code
          - Implementing Online Certificate Status Protocol (OCSP) for obtaining certificate revocation status

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ACM PCA
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acmpca
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ACM PCA Developer Resources
          url: https://aws.amazon.com/private-ca/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: acmpca maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/acmpca-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-acmpca-controller.v1.1.1
        version: 1.1.1
      - name: ack-acmpca-controller.v1.1.0
        version: 1.1.0
      - name: ack-acmpca-controller.v1.0.10
        version: 1.0.10
      - name: ack-acmpca-controller.v1.0.9
        version: 1.0.9
      - name: ack-acmpca-controller.v1.0.8
        version: 1.0.8
      - name: ack-acmpca-controller.v1.0.7
        version: 1.0.7
      - name: ack-acmpca-controller.v1.0.6
        version: 1.0.6
      - name: ack-acmpca-controller.v1.0.5
        version: 1.0.5
      - name: ack-acmpca-controller.v1.0.4
        version: 1.0.4
      - name: ack-acmpca-controller.v1.0.3
        version: 1.0.3
      - name: ack-acmpca-controller.v1.0.2
        version: 1.0.2
      - name: ack-acmpca-controller.v1.0.1
        version: 1.0.1
      - name: ack-acmpca-controller.v1.0.0
        version: 1.0.0
      - name: ack-acmpca-controller.v0.0.20
        version: 0.0.20
      - name: ack-acmpca-controller.v0.0.16
        version: 0.0.16
      - name: ack-acmpca-controller.v0.0.15
        version: 0.0.15
      - name: ack-acmpca-controller.v0.0.14
        version: 0.0.14
      - name: ack-acmpca-controller.v0.0.13
        version: 0.0.13
      - name: ack-acmpca-controller.v0.0.12
        version: 0.0.12
      - name: ack-acmpca-controller.v0.0.11
        version: 0.0.11
      - name: ack-acmpca-controller.v0.0.10
        version: 0.0.10
      - name: ack-acmpca-controller.v0.0.9
        version: 0.0.9
      - name: ack-acmpca-controller.v0.0.8
        version: 0.0.8
      - name: ack-acmpca-controller.v0.0.5
        version: 0.0.5
      - name: ack-acmpca-controller.v0.0.4
        version: 0.0.4
      - name: ack-acmpca-controller.v0.0.3
        version: 0.0.3
      - name: ack-acmpca-controller.v0.0.2
        version: 0.0.2
      - name: ack-acmpca-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-acmpca-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MayaData
      provider-url: ""
    name: openebs
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openebsoperator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "openebs.io/v1alpha1",
                "kind": "OpenEBSInstallTemplate",
                "metadata": {
                  "name": "oebs",
                  "namespace": "openebs"
                },
                "spec": {
                  "rbac": {
                    "create": true,
                    "pspEnabled": false,
                    "kyvernoEnabled": false
                  },
                  "serviceAccount": {
                    "create": true,
                    "name": "openebs-maya-operator"
                  },
                  "imagePullSecrets": [],
                  "release": {
                    "version": "3.0.0"
                  },
                  "legacy": {
                    "enabled": false
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": ""
                  },
                  "apiserver": {
                    "enabled": true,
                    "image": "openebs/m-apiserver",
                    "imageTag": "2.12.2",
                    "replicas": 1,
                    "ports": {
                      "externalPort": 5656,
                      "internalPort": 5656
                    },
                    "sparse": {
                      "enabled": "false"
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "defaultStorageConfig": {
                    "enabled": "true"
                  },
                  "varDirectoryPath": {
                    "baseDir": "/var/openebs"
                  },
                  "provisioner": {
                    "enabled": true,
                    "image": "openebs/openebs-k8s-provisioner",
                    "imageTag": "2.12.2",
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "patchJivaNodeAffinity": "enabled",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "localprovisioner": {
                    "enabled": true,
                    "image": "openebs/provisioner-localpv",
                    "imageTag": "3.0.0",
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "enableDeviceClass": true,
                    "enableHostpathClass": true,
                    "basePath": "/var/openebs/local",
                    "waitForBDBindTimeoutRetryCount": "12",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "snapshotOperator": {
                    "enabled": true,
                    "controller": {
                      "image": "openebs/snapshot-controller",
                      "imageTag": "2.12.2",
                      "resources": {}
                    },
                    "provisioner": {
                      "image": "openebs/snapshot-provisioner",
                      "imageTag": "2.12.2",
                      "resources": {}
                    },
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "upgradeStrategy": "Recreate",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    }
                  },
                  "ndm": {
                    "enabled": true,
                    "image": "openebs/node-disk-manager",
                    "imageTag": "1.7.0",
                    "sparse": {
                      "path": "/var/openebs/sparse",
                      "size": "10737418240",
                      "count": "0"
                    },
                    "filters": {
                      "enableOsDiskExcludeFilter": true,
                      "osDiskExcludePaths": "/,/etc/hosts,/boot",
                      "enableVendorFilter": true,
                      "excludeVendors": "CLOUDBYT,OpenEBS",
                      "enablePathFilter": true,
                      "includePaths": "",
                      "excludePaths": "/dev/loop,/dev/fd0,/dev/sr0,/dev/ram,/dev/dm-,/dev/md,/dev/rbd,/dev/zd"
                    },
                    "probes": {
                      "enableSeachest": false
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "ndmOperator": {
                    "enabled": true,
                    "image": "openebs/node-disk-operator",
                    "imageTag": "1.7.0",
                    "replicas": 1,
                    "upgradeStrategy": "Recreate",
                    "nodeSelector": {},
                    "tolerations": [],
                    "healthCheck": {
                      "initialDelaySeconds": 15,
                      "periodSeconds": 20
                    },
                    "readinessCheck": {
                      "initialDelaySeconds": 5,
                      "periodSeconds": 10
                    },
                    "resources": {}
                  },
                  "ndmExporter": {
                    "enabled": false,
                    "image": {
                      "registry": null,
                      "repository": "openebs/node-disk-exporter",
                      "pullPolicy": "IfNotPresent",
                      "tag": "1.7.0"
                    },
                    "nodeExporter": {
                      "name": "ndm-node-exporter",
                      "podLabels": {
                        "name": "openebs-ndm-node-exporter"
                      },
                      "metricsPort": 9101
                    },
                    "clusterExporter": {
                      "name": "ndm-cluster-exporter",
                      "podLabels": {
                        "name": "openebs-ndm-cluster-exporter"
                      },
                      "metricsPort": 9100
                    }
                  },
                  "webhook": {
                    "enabled": true,
                    "image": "openebs/admission-server",
                    "imageTag": "2.12.2",
                    "failurePolicy": "Fail",
                    "replicas": 1,
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "hostNetwork": false,
                    "resources": {}
                  },
                  "helper": {
                    "image": "openebs/linux-utils",
                    "helperImageTag": "3.0.0"
                  },
                  "featureGates": {
                    "enabled": true,
                    "GPTBasedUUID": {
                      "enabled": true,
                      "featureGateFlag": "GPTBasedUUID"
                    },
                    "APIService": {
                      "enabled": false,
                      "featureGateFlag": "APIService",
                      "address": "0.0.0.0:9115"
                    },
                    "UseOSDisk": {
                      "enabled": false,
                      "featureGateFlag": "UseOSDisk"
                    },
                    "ChangeDetection": {
                      "enabled": false,
                      "featureGateFlag": "ChangeDetection"
                    }
                  },
                  "crd": {
                    "enableInstall": true
                  },
                  "policies": {
                    "monitoring": {
                      "enabled": true,
                      "image": "openebs/m-exporter",
                      "imageTag": "2.12.2"
                    }
                  },
                  "analytics": {
                    "enabled": true,
                    "pingInterval": "24h"
                  },
                  "jiva": {
                    "image": "openebs/jiva",
                    "imageTag": "2.12.2",
                    "replicas": 3,
                    "defaultStoragePath": "/var/openebs",
                    "enabled": false,
                    "openebsLocalpv": {
                      "enabled": false
                    },
                    "localpv-provisioner": {
                      "openebsNDM": {
                        "enabled": false
                      }
                    }
                  },
                  "cstor": {
                    "pool": {
                      "image": "openebs/cstor-pool",
                      "imageTag": "2.12.2"
                    },
                    "poolMgmt": {
                      "image": "openebs/cstor-pool-mgmt",
                      "imageTag": "2.12.2"
                    },
                    "target": {
                      "image": "openebs/cstor-istgt",
                      "imageTag": "2.12.2"
                    },
                    "volumeMgmt": {
                      "image": "openebs/cstor-volume-mgmt",
                      "imageTag": "2.12.2"
                    },
                    "enabled": false,
                    "openebsNDM": {
                      "enabled": false
                    }
                  },
                  "openebs-ndm": {
                    "enabled": false
                  },
                  "localpv-provisioner": {
                    "enabled": false,
                    "openebsNDM": {
                      "enabled": false
                    }
                  },
                  "zfs-localpv": {
                    "enabled": false
                  },
                  "lvm-localpv": {
                    "enabled": false
                  },
                  "nfs-provisioner": {
                    "enabled": false
                  },
                  "cleanup": {
                    "image": {
                      "registry": "",
                      "repository": "bitnami/kubectl",
                      "tag": "",
                      "imagePullSecrets": []
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: index.docker.io/openebs/helm-operator:v0.0.9
          createdAt: "2021-11-02T02:56:58Z"
          description: Creates and maintains OpenEBS Control Plane deployments
          repository: https://github.com/openebs/helm-operator
          support: https://slack.openebs.io/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a OpenEBS Install Operator
            displayName: OpenEBS Install Template
            kind: OpenEBSInstallTemplate
            name: openebsinstalltemplates.openebs.io
            version: v1alpha1
        description: |
          **OpenEBS** is a leading container attached storage solution that enables the use of containers for mission-critical, persistent workloads and for other stateful workloads such as logging or Prometheus for example.
          OpenEBS itself is deployed as just another container on your host and enables storage services that can be designated on a per pod, application, cluster or container level, including:
          * Data persistence across nodes
          * Synchronization of data across availability zones and cloud providers
          * A common layer whether you are running on the cloud, or your bare metal
          * Integration with Kubernetes, so developer and application intent flows into OpenEBS
          * Management of tiering to and from S3 and other targets.
          ## OpenEBS Operator
          OpenEBS primarily provides container attached block storage (iSCSI volumes) by leveraging/aggregating the storage on the nodes, with the storage controller itself running as a container. Different storage engines (Jiva & cStor) are supported, with tools available to dynamically provision Kubernetes Local PVs. The volumes are dynamically provisioned via PersistentVolumeClaims and are managed by a control plane component called "maya", which also runs as a deployment in the K8s cluster. In addition to maya, a typical OpenEBS installation comprises several other resources, which aid with various functionalities, ranging from snapshotting to disk management. All these components are described briefly below:
          * **Maya-API-Server** - A storage orchestrator which integrates into Kubernetes workflow to help provision
            and manage OpenEBS Jiva & cStor (storage engine) volumes

          * **Dynamic-OpenEBS-Provisioner** - A Kubernetes external storage provisioner that utilizes APIs exposed by maya-apiserver
            to perform provision & delete operations of Jiva & cStor volumes

          * **Dynamic-LocalPV-Provisioner** - A dynamic provisioner for Kubernetes Local PVs
          * **OpenEBS-Snapshot-Operator** - A Kubernetes snapshot controller that creates & restores OpenEBS volume snapshots
          * **Node-Disk-Manager** - A disk management controller which identifies available disks, maintains inventory, and
            dynamically attaches/detaches disks to backend storage pods

          The helm-based OpenEBS Operator eases the setup of all the above mentioned components, with a simple custom resource provided to define the install options, thereby enabling applications to start using the OpenEBS storageclasses in their PVCs. The OpenEBSInstallTemplate CR can be used to specify start-up parameters & also update/overwrite the definitions post install.
          ## Pre-Requisites
          Before installing OpenEBS control plane, perform the following steps to ensure successful deployment of the Node-Disk-Manager & the OpenEBS volume replicas, respectively.
          * Configure the OpenEBS service account on the openshift-operators namespace/project to use the privileged security context constraint.

            **Note**: The serviceaccount name is same as the one specified in the `spec.serviceAccount.name` field of the OpenEBSInstallTemplate CR.


            ```
            oc adm policy add-scc-to-user privileged system:serviceaccount:openshift-operators:openebs-maya-operator
            ```

          * Configure the default service account on the namespace/project in which the volume replicas are deployed to use privileged
            security context constraint.


            ```
            oc adm policy add-scc-to-user privileged system:serviceaccount:openshift-operators:default
            ```

          ## Getting Started
          * Try the quickstart [guide](https://github.com/openebs/helm-operator/blob/master/olm/README.md)
          * To learn how to contribute, please read the [contribution guide](https://github.com/openebs/helm-operator/blob/master/CONTRIBUTING.md)
          * OpenEBS welcomes your feedback and contributions in any form possible. [Join our Community](https://openebs.org/community)

          ## License
          OpenEBS is distributed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt)
        displayName: OpenEBS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenEBS
        - ContainerAttachedStorage
        links:
        - name: OpenEBS Website
          url: https://openebs.io
        - name: Operator Source Code
          url: https://github.com/openebs/helm-operator
        - name: Install Instructions
          url: https://github.com/openebs/helm-operator/blob/master/olm/README.md
        maintainers:
        - email: kiran.mova@openebs.io
          name: kmova
        maturity: alpha
        minKubeVersion: 1.12.0
        provider:
          name: MayaData
        relatedImages:
        - index.docker.io/openebs/helm-operator:v0.0.9
        version: 3.0.0
      entries:
      - name: openebsoperator.v3.0.0
        version: 3.0.0
      - name: openebsoperator.v2.12.2
        version: 2.12.2
      name: alpha
    defaultChannel: alpha
    packageName: openebs
    provider:
      name: MayaData
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-core
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Core",
                "metadata": {
                  "name": "core-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configService": {
                    "port": 8080
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "helper": {
                    "port": 8080
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "licenseService": {
                    "port": 8080
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "restStatsService": {
                    "port": 8080
                  },
                  "resultService": {
                    "port": 8080,
                    "storage": "64Gi"
                  },
                  "sessionManager": {
                    "port": 8080
                  },
                  "statsService": {
                    "port": 8080
                  },
                  "testResultsService": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:34:00Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Core
            name: cores.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Core Operator
        displayName: Keysight WAP Core Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Core
          url: https://keysight-wap-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-core@sha256:001b0d08cd84a9ebc13cd8fd35b68d431b96f84d71e933c1123ed6c0d244fd37
        version: 0.1.21
      entries:
      - name: keysight-wap-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-nats-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-nats-operator.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2023-09-19T12:44:10Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NatsCluster
            name: natsclusters.nats.io
            version: v1alpha2
          - kind: NatsServiceRole
            name: natsserviceroles.nats.io
            version: v1alpha2
        description: Keysight Nats Operator
        displayName: Keysight Nats Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Nats Operator
          url: https://keysight-nats-operator.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-nats-operator@sha256:2ee5dc8814274e83a5f82e89eabd0b643e8cde8e3d555a56535a61d05c72fd75
        version: 0.1.21
      entries:
      - name: keysight-nats-operator.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-nats-operator
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: github.com/stakater-ab/hibernation-operator
    name: hibernation-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hibernation-operator.v0.1.103
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Monitoring
          createdAt: "2025-10-02T05:44:42Z"
          description: Operator to engage and disengage hibernation on k8s resources
            using cron expressions
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.1.103
          operatorframework.io/suggested-namespace: hibernation-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Cluster Resource Supervisor
            kind: ClusterResourceSupervisor
            name: clusterresourcesupervisors.hibernation.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.hibernation.stakater.com
            version: v1beta1
        description: Schedule k8s resources using cron expressions via Hibernation
          Operator
        displayName: Hibernation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - redhat
        - k8s
        - stakater
        - hibernation
        - operator
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maintainers:
        - email: hello@stakater.com
          name: StakaterAB
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Stakater
          url: github.com/stakater-ab/hibernation-operator
        relatedImages:
        - registry.connect.redhat.com/stakater/hibernation-operator@sha256:3738f82834d53aec4f8fba5caad6258c7adb4bdc23a57bde47c7ec22030fae5c
        version: 0.1.103
      entries:
      - name: hibernation-operator.v0.1.103
        version: 0.1.103
      name: release-0.1
    defaultChannel: release-0.1
    packageName: hibernation-operator
    provider:
      name: Stakater
      url: github.com/stakater-ab/hibernation-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: runtime-component-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runtime-component.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/runtime-component-operator@sha256:d37b3f62575c9ad1ffcb22c89e2c3886275c326bf27aa49fb1808acba475853d
          createdAt: "2022-03-25T14:00:00Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          olm.skipRange: <0.8.1
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          This advanced Operator is capable of deploying any runtime component image with consistent, production-grade QoS. It enables enterprise architects to govern the way their applications get deployed & managed in the cluster, while dramatically reducing the learning curve for developers to deploy into Kubernetes - allowing them to focus on writing the code!
          Here are some key features:

          #### Application Lifecyle
          You can deploy your runtime component container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the Operator will watch for any updates and will re-deploy the modified image.

          #### Custom RBAC
          This Operator is capable of using a custom ServiceAccount from the caller, allowing it to follow RBAC restrictions. By default it creates a ServiceAccount if one is not specified, which can also be bound with specific roles.

          #### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as: labels, annotations, and environment variables from a ConfigMap, a Secret or a value.

          #### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configuration, such as TLS settings, are also easily enabled.

          #### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto scaling to create (and delete) instances based on resource consumption.

          #### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you.
          This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart.
          We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.

          #### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating kubernetes Secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to its required services without any intervention or interruption.

          #### Exposing metrics to Prometheus
          The Runtime Component Operator exposes the runtime container's metrics via the Prometheus Operator.
          You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `spec.monitoring` field to control configurations such as the poll interval and security credentials.

          #### Easily mount logs and transaction directories
          If you need to mount the logs and transaction data from your runtime component to an external volume such as NFS (or any storage supported in your cluster), simply add the following (to specify the volume size and the location to persist) to your RuntimeComponent CR:
          ``` storage: size: 2Gi mountPath: "/logs" ```

          #### Integration with OpenShift Serverless
          Deploy your serverless runtime component using a single toggle.  The Operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing your pod to automatically scale to 0 when it is idle.

          #### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and how they are connected.

          See our [**documentation**](https://github.com/application-stacks/runtime-component-operator/tree/main/doc/) for more information.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://github.com/application-stacks/runtime-component-operator/tree/main/doc
        - name: Repository
          url: https://github.com/application-stacks/runtime-component-operator
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: beta
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/runtime-component-operator@sha256:d37b3f62575c9ad1ffcb22c89e2c3886275c326bf27aa49fb1808acba475853d
        - registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4
        version: 0.8.1
      entries:
      - name: runtime-component.v0.8.1
        version: 0.8.1
      - name: runtime-component.v0.8.0
        version: 0.8.0
      name: beta2
    - currentCSV: runtime-component.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:1b818f1a9c27baab6a39ae42d86f78bac085d772cae5cd97b357d0ee5799cf85
          createdAt: "2023-08-28T16:46:50Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          olm.skipRange: '>=0.8.0 <1.2.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02
        - icr.io/appcafe/runtime-component-operator@sha256:1b818f1a9c27baab6a39ae42d86f78bac085d772cae5cd97b357d0ee5799cf85
        version: 1.2.2
      entries:
      - name: runtime-component.v1.2.2
        version: 1.2.2
      name: v1.2
    - currentCSV: runtime-component.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:682c5cb3e24566bf2933b1d4a4f6bf520d9b546edcedc1ee3214ec1ac4d48629
          createdAt: "2024-08-01T20:26:42Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.3.3'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/runtime-component-operator@sha256:682c5cb3e24566bf2933b1d4a4f6bf520d9b546edcedc1ee3214ec1ac4d48629
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91
        version: 1.3.3
      entries:
      - name: runtime-component.v1.3.3
        version: 1.3.3
      name: v1.3
    - currentCSV: runtime-component.v1.4.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:6f527c5173421c28d175b42d8c690eb63c4d1c5092c31d5350fd7a5e93661eaa
          createdAt: "2025-07-18T20:38:10Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.4.4'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/runtime-component-operator@sha256:6f527c5173421c28d175b42d8c690eb63c4d1c5092c31d5350fd7a5e93661eaa
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3
        version: 1.4.4
      entries:
      - name: runtime-component.v1.4.4
        version: 1.4.4
      name: v1.4
    - currentCSV: runtime-component.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:58bfceaed2fc828bce1c0e4e21a583d25f6d72f88abdc78df2510867a4f2020f
          createdAt: "2025-09-15T18:08:27Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.5.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8
        - icr.io/appcafe/runtime-component-operator@sha256:58bfceaed2fc828bce1c0e4e21a583d25f6d72f88abdc78df2510867a4f2020f
        version: 1.5.0
      entries:
      - name: runtime-component.v1.5.0
        version: 1.5.0
      name: v1.5
    defaultChannel: v1.5
    packageName: runtime-component-operator-certified
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio Project
      provider-url: ""
    name: apicurito
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: 1.0.x
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: candidate
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: latest
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: latest
    packageName: apicurito
    provider:
      name: Apicurio Project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Apache Software Foundation
      provider-url: ""
    name: camel-k
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: camel-k-operator.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.0.1
          createdAt: "2023-09-04T12:32:38Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
          - kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.0.1
        version: 2.0.1
      entries:
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: candidate
    - currentCSV: camel-k-operator.v2.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "setBody": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationProfile",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "debug"
                },
                "spec": {
                  "traits": {
                    "logging": {
                      "level": "DEBUG"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example-source"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "setBody": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "timer-to-log-example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "log-sink"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world",
                      "period": 2000
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "timer-source"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.8.0
          createdAt: "2025-09-21T09:07:43Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: IntegrationProfile is the Schema for the integrationprofiles
              API
            displayName: Integration Profile
            kind: IntegrationProfile
            name: integrationprofiles.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Pipe is the Schema for the pipes API
            displayName: Pipe
            kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers:
          the easiest way to build and manage your Camel applications on Kubernetes.

          ## Installation

          You may need to specify the container registry where to store the container images created for the Camel applications.
          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          spec:
            build:
              registry:
                address: docker.io/my-org
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step when using Openshift cluster.

          ## Running an Integration

          Run a Camel application on the cluster by creating a sample `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            annotations:
              camel.apache.org/operator.id: camel-k
            name: test
          spec:
            flows:
            - from:
                parameters:
                  period: "1000"
                steps:
                - setBody:
                    simple: Hello Camel from ${routeId}
                - log: ${body}
                uri: timer:yaml
          ```
          Log the result of the running application:
          ```
          kubectl wait --for=condition=ready integration test --timeout=180s
          kubectl logs deployment/test -f
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.8.0
        version: 2.8.0
      entries:
      - name: camel-k-operator.v2.8.0
        version: 2.8.0
      - name: camel-k-operator.v2.7.0
        version: 2.7.0
      - name: camel-k-operator.v2.6.0
        version: 2.6.0
      - name: camel-k-operator.v2.5.1
        version: 2.5.1
      - name: camel-k-operator.v2.5.0
        version: 2.5.0
      - name: camel-k-operator.v2.4.0
        version: 2.4.0
      - name: camel-k-operator.v2.3.3
        version: 2.3.3
      - name: camel-k-operator.v2.3.2
        version: 2.3.2
      - name: camel-k-operator.v2.3.1
        version: 2.3.1
      - name: camel-k-operator.v2.3.0
        version: 2.3.0
      - name: camel-k-operator.v2.2.0
        version: 2.2.0
      - name: camel-k-operator.v2.1.0
        version: 2.1.0
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: latest
    - currentCSV: camel-k-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.12.0
          createdAt: "2023-02-27T09:01:55Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.12.0
        version: 1.12.0
      entries:
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable
    - currentCSV: camel-k-operator.v1.10.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.10.4
          createdAt: "2023-01-28T13:49:05Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.10.4
        version: 1.10.4
      entries:
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.10
    - currentCSV: camel-k-operator.v1.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.11.1
          createdAt: "2023-01-30T08:45:04Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.11.1
        version: 1.11.1
      entries:
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.11
    - currentCSV: camel-k-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.12.1
          createdAt: "2023-06-12T08:46:58Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.12.1
        version: 1.12.1
      entries:
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.12
    - currentCSV: camel-k-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.5.1
          createdAt: "2021-08-31T09:03:41Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.5.1
        version: 1.5.1
      entries:
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.5
    - currentCSV: camel-k-operator.v1.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.6.1
          createdAt: "2021-10-27T21:06:20Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.6.1
        version: 1.6.1
      entries:
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.6
    - currentCSV: camel-k-operator.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.7.0
          createdAt: "2021-11-16T10:54:12Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.8.1+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.7.0
        version: 1.7.0
      entries:
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.7
    - currentCSV: camel-k-operator.v1.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.8.2
          createdAt: "2022-03-07T07:13:19Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.8.2
        version: 1.8.2
      entries:
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.8
    - currentCSV: camel-k-operator.v1.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.9.2
          createdAt: "2022-05-23T09:05:22Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.9.2
        version: 1.9.2
      entries:
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.9
    - currentCSV: camel-k-operator.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.0.1
          createdAt: "2023-09-04T12:32:38Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
          - kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.0.1
        version: 2.0.1
      entries:
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-2.0
    - currentCSV: camel-k-operator.v2.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "setBody": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationProfile",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "debug"
                },
                "spec": {
                  "traits": {
                    "logging": {
                      "level": "DEBUG"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example-source"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "setBody": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "timer-to-log-example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "log-sink"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world",
                      "period": 2000
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "timer-source"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.8.0
          createdAt: "2025-09-21T09:07:43Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: IntegrationProfile is the Schema for the integrationprofiles
              API
            displayName: Integration Profile
            kind: IntegrationProfile
            name: integrationprofiles.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Pipe is the Schema for the pipes API
            displayName: Pipe
            kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers:
          the easiest way to build and manage your Camel applications on Kubernetes.

          ## Installation

          You may need to specify the container registry where to store the container images created for the Camel applications.
          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          spec:
            build:
              registry:
                address: docker.io/my-org
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step when using Openshift cluster.

          ## Running an Integration

          Run a Camel application on the cluster by creating a sample `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            annotations:
              camel.apache.org/operator.id: camel-k
            name: test
          spec:
            flows:
            - from:
                parameters:
                  period: "1000"
                steps:
                - setBody:
                    simple: Hello Camel from ${routeId}
                - log: ${body}
                uri: timer:yaml
          ```
          Log the result of the running application:
          ```
          kubectl wait --for=condition=ready integration test --timeout=180s
          kubectl logs deployment/test -f
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.8.0
        version: 2.8.0
      entries:
      - name: camel-k-operator.v2.8.0
        version: 2.8.0
      - name: camel-k-operator.v2.7.0
        version: 2.7.0
      - name: camel-k-operator.v2.6.0
        version: 2.6.0
      - name: camel-k-operator.v2.5.1
        version: 2.5.1
      - name: camel-k-operator.v2.5.0
        version: 2.5.0
      - name: camel-k-operator.v2.4.0
        version: 2.4.0
      - name: camel-k-operator.v2.3.3
        version: 2.3.3
      - name: camel-k-operator.v2.3.2
        version: 2.3.2
      - name: camel-k-operator.v2.3.1
        version: 2.3.1
      - name: camel-k-operator.v2.3.0
        version: 2.3.0
      - name: camel-k-operator.v2.2.0
        version: 2.2.0
      - name: camel-k-operator.v2.1.0
        version: 2.1.0
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-v2
    defaultChannel: stable-v2
    packageName: camel-k
    provider:
      name: The Apache Software Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-cpd-ccs
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-ccs.v10.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ccs.cpd.ibm.com/v1beta1",
                "kind": "CCS",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ccs-cr",
                    "app.kubernetes.io/managed-by": "ibm-cpd-ccs-operator",
                    "app.kubernetes.io/name": "ibm-cpd-ccs-operator"
                  },
                  "name": "ccs-cr"
                },
                "spec": {
                  "scaleConfig": "small",
                  "license": {
                    "accept": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
          createdAt: "2025-04-18"
          description: Common Core Services
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-ccs
          olm.skipRange: '>=7.0.0 <10.3.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM watsonx.ai"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: |-
            [
              "runtimeassemblies.runtimes.ibm.com"
            ]
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation Runtime Manager CRD. This product is not deployable
              standalone and it is part of the LICCR already accepted.  For additional
              details regarding install parameters check: http://ibm.biz/-readme.
              License By installing this product you accept the license terms https://ibm.biz/BdaaLC
              or https://ibm.biz/BdaaL9.'
            kind: RuntimeAssembly
            name: runtimeassemblies.runtimes.ibm.com
            version: v1alpha1
          - description: 'Documentation This product is not deployable standalone
              and it is part of the LICCR already accepted.  For additional details
              regarding install parameters check: http://ibm.biz/-readme. License
              By installing this product you accept the license terms https://ibm.biz/BdaaLC
              or https://ibm.biz/BdaaL9.'
            kind: CCS
            name: ccs.ccs.cpd.ibm.com
            version: v1beta1
        description: |-
          # Introduction Common Core Services Operator
          Put stuff here

          # Details
          Various Details
          ## Prerequisites
          Various Prerequisites
          ## Resources Required
          Various Resources Required
          ## Installing
          Various Installation details
          # Configuration
          Various Configuration information
          ### Limitations
          Various Limitations
          ### SecurityContextConstraints Requirements
          All pods will use the restricted SCC, I think.
        displayName: Common Core Services
        installModes:
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: OwnNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commonc-core-services
        - cpd40
        links:
        - name: Ccs
          url: https://ccs.domain
        maintainers:
        - email: test@ibm.com
          name: test
        maturity: v1
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/jobs-api@sha256:173b88f90078e1817686ac949f48387af70a706e23e6fdab4a7c6217db5ad86e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-bundle@sha256:94c7a1dd290e2abd55c5b8998cf08ad174d3fbfc659ae6fa6fb6cbab61d5fe14
        - registry.marketplace.redhat.com/cp/cpd/dap-base-job@sha256:f208686a3f56d2980e7b8640b6ebdea9057109e7520c6817a731e68d08506a36
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:bcdf9b7eea8753b3e463973a2690140ae7aa8c02c3c7726511f116bd751ec2d8
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:c268d2852580286e5c6770857fcfc3b06a698521f5acd17ed83b8e2cb9ab55cf
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:523dcc671f1c05e690d5e578a90300d28d76570f605d4de7f80645b590f4a462
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:81d8b25221d827160dd5e88efd383453f532057ad02d7a2ee3dcd9c6eb16ba65
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:0e0ae96aff8540323da8a2379af014d3af1770e4b0043b84b9b415dda44501f7
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:6f2e7cc917adb54b3112dc4909f645ac08c33e8c2c33266fbc8f5133cbcd1973
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:21c81ff7c4b8809accaf62b242fd9b3d52f235489ae427556b7f71cc21bf8926
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:351375de35d7fc9d940b905b867ce8a00b7ff1416e70d3a100918797fe3d8c56
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:4fb2127ba7c525a959bbbcf024de364ef266ebb3ca7e696f74d990d2552cda59
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:56315ce0a0a70f5899191a7eab856105489143f0e889c8cdd1b6b8c225f02eb7
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:a3a0d58f895fca4d832e602d0dee80018df4ad7bad8800ad4058eea07fda4817
        - registry.marketplace.redhat.com/cp/cpd/ngp-projects-api@sha256:df420e9f616ad8a123d9eedd117fb6cb007b30a772138db534f47ad8801fa96d
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:257bf301338f7b123099e2ffe31c3f3b793bf28d6a0570a25b889271239579ba
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:6f96b8b6afb00cb26fcf903b8e4c30b0d6344bb86d8bce2f2026fc8cb0b575ec
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:0621eb9b83c9b2b8bfe5cd09797fe08fa37a854a5506914935ff6f0df599b631
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:0ae22914dbaf09d6679bd06bc9b6a5728e744c9ce2309cbd5f01e28308c403ea
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:562281dbb2bec5191a7c6b595ab853092e6fe3ff1432e77b2471f71ccdffcbc8
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:54c86367bfe0ab658611856cc0a6c5de24c14a336bd129ac9f16293945a55238
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:f5527ddbbd0b9878555fce852996b13020fd91b1604d8c03d34a61ca2fdb2d1d
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:fbc6eab07fcfc2d0c1895f18e17f751e9481ad111875959e9ff08e14f81b9a35
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:740a1d87a43e61e32cfdea032d852af33100d959c8b57ff2f165f4baf77fc011
        - registry.marketplace.redhat.com/cp/cpd/task-credentials@sha256:e162828d5ad07ceae4db0589f930979842714e3b766552102bee373a3a570af3
        - registry.marketplace.redhat.com/cp/cpd/environments-ui@sha256:2923664475c7a8ea5e63055c90748db37825fe76c7975b845f1185d4b32ee4ea
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/cpd/runtime-manager-api@sha256:a37773df07fe757dd698e3f09f8687902e7b3592d97855e703dd92ecc832aab0
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:8a6ce0dada706056efda73ad0cc1547566003e0a7aab8ed379b50e4776861b94
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e12599ff99db058fdc8f9c1eebfba31160c636ff092138deb1f6f6705f44fd62
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:3f2f97e6b2b18526255e88ff02d1039b24312eca51ee22b7cbf76d5680f7aee1
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:d023f35b65816f244b6d699b41362e6031553f7d4c416efcee4a7f2395a61355
        - registry.marketplace.redhat.com/cp/cpd/spaces-ui-job@sha256:95dade51444e30ec034c12207ecbef675b70889b81970f49f10aa75160b2f947
        - registry.marketplace.redhat.com/cp/cpd/environment-specifications-sync@sha256:894b4624d8ead7286cf9068c6477300992c32d96339644fe764bcb29b1c8ef5c
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:ad19b65042fb26cc1244a92f81a96e5ceb4313f8c823181efe28b66e9fecf9c8
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:044b663e4b7508c62dbb28b7fb2aa267c9771fc62ab509b08d46ea1499d95431
        - registry.marketplace.redhat.com/cp/cpd/environments-api@sha256:6dd60b530f81d7a2d55a534211f7f128814abf87b8cd3575d528dba53442c44e
        - registry.marketplace.redhat.com/cp/cpd/jobs-ui@sha256:49daf607918b378593485da16ac417056ee295bca94f1a5cac37b334381ae096
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:6cffef7631ee54612fba71efc0eff4e17bc08defc3e2137aee9881d685444298
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:722944bfd0eba280255f4c198f65f562423a26e2c4ad3580591ee71cb6e9c2ef
        - registry.marketplace.redhat.com/cp/cpd/job-trigger@sha256:f2ea487072393a24e4c6864b2107c1fce072ebba07f9dbc2c139bd37fbd062da
        - registry.marketplace.redhat.com/cp/cpd/environments-init-job@sha256:6af8f900af7c11eee6894e2c3e52f740b3c700dac0ee753fadef330230bd42dd
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:6fe90f051d9261ad224dccf4d9f6bc9f19ce0e8d22fa010a46b2752374bce2b3
        version: 10.3.0
      entries:
      - name: ibm-cpd-ccs.v10.3.0
        version: 10.3.0
      name: v10.3
    defaultChannel: v10.3
    packageName: ibm-cpd-ccs
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: swift-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: swift-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "swift.openstack.org/v1beta1",
                "kind": "Swift",
                "metadata": {
                  "name": "swift"
                },
                "spec": {
                  "swiftProxy": {
                    "passwordSelectors": {
                      "service": "SwiftPassword"
                    },
                    "replicas": 1
                  },
                  "swiftRing": {
                    "ringReplicas": 1
                  },
                  "swiftStorage": {
                    "replicas": 1,
                    "storageClass": "local-storage"
                  }
                }
              },
              {
                "apiVersion": "swift.openstack.org/v1beta1",
                "kind": "SwiftProxy",
                "metadata": {
                  "name": "swift-proxy"
                },
                "spec": {
                  "passwordSelectors": {
                    "service": "SwiftPassword"
                  },
                  "replicas": 1,
                  "secret": "osp-secret",
                  "serviceUser": "swift"
                }
              },
              {
                "apiVersion": "swift.openstack.org/v1beta1",
                "kind": "SwiftRing",
                "metadata": {
                  "name": "swift-ring"
                },
                "spec": {
                  "ringReplicas": 1
                }
              },
              {
                "apiVersion": "swift.openstack.org/v1beta1",
                "kind": "SwiftStorage",
                "metadata": {
                  "name": "swift-storage"
                },
                "spec": {
                  "replicas": 1,
                  "storageClass": "local-storage",
                  "storageRequest": "10G"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:34Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SwiftProxy is the Schema for the swiftproxies API
            displayName: Swift Proxy
            kind: SwiftProxy
            name: swiftproxies.swift.openstack.org
            version: v1beta1
          - description: SwiftRing is the Schema for the swiftrings API
            displayName: Swift Ring
            kind: SwiftRing
            name: swiftrings.swift.openstack.org
            version: v1beta1
          - description: Swift is the Schema for the swifts API
            displayName: Swift
            kind: Swift
            name: swifts.swift.openstack.org
            version: v1beta1
          - description: SwiftStorage is the Schema for the swiftstorages API
            displayName: Swift Storage
            kind: SwiftStorage
            name: swiftstorages.swift.openstack.org
            version: v1beta1
        description: OpenStack Swift operator
        displayName: Swift operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Swift
        - Object Storage
        links:
        - name: Swift Operator
          url: https://github.com/openstack-k8s-operators/swift-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-swift-object-rhel9@sha256:656f23b25ca0bd80d7b77ba717a21455366dfaf2049df1bd6d6536925a18b94b
        - registry.redhat.io/rhoso/openstack-swift-proxy-server-rhel9@sha256:6e654db83d570a3f8feedb912d5d9dff35ecba4d53169481a3b4982d3a832798
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/swift-rhel9-operator@sha256:18207f02c049447836211b326f09ce660782c384cda92e592db256692e3c8e73
        - registry.redhat.io/rhoso/openstack-swift-account-rhel9@sha256:ecebf3eb430188392f7a7b8cda3d1c0c0f658bcbbaff531d99eba88a7e0c9889
        - registry.redhat.io/rhoso/openstack-swift-container-rhel9@sha256:782866f8e9661de50d6a74bc52262f4546d7e0055d0c4da83ba6b891a50c35ea
        version: 1.0.6
      entries:
      - name: swift-operator.v1.0.6
        version: 1.0.6
      - name: swift-operator.v1.0.5
        version: 1.0.5
      - name: swift-operator.v1.0.4
        version: 1.0.4
      - name: swift-operator.v1.0.3
        version: 1.0.3
      - name: swift-operator.v1.0.2
        version: 1.0.2
      - name: swift-operator.v1.0.1
        version: 1.0.1
      - name: swift-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: swift-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/component: kmm
      app.kubernetes.io/name: kmm
      app.kubernetes.io/part-of: kmm
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: kernel-module-management
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kernel-module-management.v2.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:1022205d712a2ed9d6e76968961c180d4e66a2e0fb991221b0b02ed94ed830d2
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.0.2'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel-module-management
        - out-of-tree
        - kernel
        - kmm
        - module
        - modules
        - oot
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:1022205d712a2ed9d6e76968961c180d4e66a2e0fb991221b0b02ed94ed830d2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62c44057a27abd41cf3daf60b8736da59a82dadb65fa6c32bef65d19bd49ea49
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:25f39f20f325f64676bf8ef14cd8a60cfbdc77116753a89d4bb667ea2a86172c
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:192e08b121a04ae69aea60c35199e8f5e3c42305e1ebd024a605a371de4ab472
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:94b4ddc45cb34d4ffe794a51b99c868cb709f90cc4eb8b28284d70c526a0a226
        version: 2.0.2
      entries:
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.0
    - currentCSV: kernel-module-management.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a85037778c53c3766f39efe4c4d2523d01ae7743e5c5693b05bd8b2cbb50b1d8
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.1.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a85037778c53c3766f39efe4c4d2523d01ae7743e5c5693b05bd8b2cbb50b1d8
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:611556d999ea97f9dbb78fec5963baa02de7a26626d7096ba63b177e3bcf9707
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:14e100e5853444114029116255ebfb7a54e5ed51503ba969740ad48b5ca80f70
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:1a9a19df17b464d1708b2905d03b8726e4480ca1f11aa97a15ff4c049a24ea4b
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:ea3e5fbf7eaf6bd29031e2876d7e24859779fb4d3d7ef27c26a1d310f137ad17
        version: 2.1.1
      entries:
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.1
    - currentCSV: kernel-module-management.v2.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a224708950f8defb259ef0da49011d361c5c0378404350640600c6e8fbdbd606
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.2.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:b1d98c0f14448ae8eb80821cbec236150c338674aae2a2b37f8d0ab0d5f1430c
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:53fddf0b226eb579c71c86112a8d9b6b1d90a1498d6f6985932b16b8be532f35
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:c3dc9ad4ebc68cd85a074b520903a9310e72a90a73bdf4aa7c42b7ed66b7bcba
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:ce1d68fd96f5003285f9fbc0c7abd8aa23830dbdd9b76629d38d4d2238e30032
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a224708950f8defb259ef0da49011d361c5c0378404350640600c6e8fbdbd606
        version: 2.2.1
      entries:
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.2
    - currentCSV: kernel-module-management.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:750dcc75f908cdda86932386fa1a28edaf65d7b5379ef6d0447586a064db906f
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.3.0'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:750dcc75f908cdda86932386fa1a28edaf65d7b5379ef6d0447586a064db906f
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:c5005b74402f680d11c9ea682a9a0d3f74b6ab0d01d25298d232f910f1a29c34
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:055c2d6e80bd1ec225f0cb2041572ebb28cfd1a4ed82ba47ff0d638d52ded2a8
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:f9fde334af68220943017a7491130e6a82f15e63522f5bb3931aa7d9b00ccc40
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:06ac089cc89b303e525fd71d6439b958f34ab944a72209f1cc8cb562ecfd1112
        version: 2.3.0
      entries:
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.3
    - currentCSV: kernel-module-management.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:41Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:8e0a3cfb1cb92d6d3c9ab5a3030593d8df1378c3ce755d598fb77b08ea5793d3
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        version: 2.4.1
      entries:
      - name: kernel-module-management.v2.4.1
        version: 2.4.1
      - name: kernel-module-management.v2.4.0
        version: 2.4.0
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.4
    - currentCSV: kernel-module-management.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:41Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:8e0a3cfb1cb92d6d3c9ab5a3030593d8df1378c3ce755d598fb77b08ea5793d3
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        version: 2.4.1
      entries:
      - name: kernel-module-management.v2.4.1
        version: 2.4.1
      - name: kernel-module-management.v2.4.0
        version: 2.4.0
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: kernel-module-management
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: recipe
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: recipe.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "Recipe",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "recipe",
                    "app.kubernetes.io/instance": "recipe-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "recipe",
                    "app.kubernetes.io/part-of": "recipe"
                  },
                  "name": "recipe-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          createdAt: "2025-09-12T06:34:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Recipe is the Schema for the recipes API
            displayName: Recipe
            kind: Recipe
            name: recipes.ramendr.openshift.io
            version: v1alpha1
        description: |
          Recipe describes a workflow used for capturing or recovering Kubernetes resources. This can be referred to by a Ramen DRPlacementControl object when kubeObjectProtection is enabled.
        displayName: Recipe
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - recipe
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/recipe
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:af262b0890300183db1766bc392017b0bda723a9ba841dc25e3d928aa413db9a
        version: 4.16.16-rhodf
      entries:
      - name: recipe.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: recipe.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: recipe.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: recipe.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: recipe.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: recipe.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: recipe.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: recipe.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: recipe.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: recipe.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: recipe.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: recipe.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: recipe.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: recipe.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: recipe.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: recipe.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: recipe.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: recipe
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: New England Research Cloud
      provider-url: https://nerc.mghpcc.org/
    name: keycloak-permissions-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keycloak-permissions-operator.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.nerc.mghpcc.org/v1",
                "kind": "KeycloakAuthorization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keycloak-permissions-operator",
                    "app.kubernetes.io/instance": "keycloakauthorization-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "keycloakauthorization",
                    "app.kubernetes.io/part-of": "keycloak-permissions-operator"
                  },
                  "name": "example"
                },
                "spec": {
                  "cleanupUndefinedAuthorizations": false,
                  "description": "",
                  "keycloak": {
                    "adminSecret": {
                      "name": "keycloak-initial-admin",
                      "passwordKey": "password",
                      "usernameKey": "username"
                    }
                  },
                  "managementState": "Managed",
                  "realm": null,
                  "validateCerts": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          com.redhat.openshift.versions: v4.11
          containerImage: quay.io/nerc-images/keycloak-permissions-operator:1.4.1
          createdAt: "2024-05-07T21:09:02Z"
          description: Managing Keycloak resources, scopes, policies, and permissions
            for fine-grained resource permissions
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/nerc-images/keycloak-permissions-operator
          support: New England Research Cloud
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Keycloak resources, scopes, policies, and permissions
              for fine-grained resource permissions.
            displayName: Keycloak Authorization
            kind: KeycloakAuthorization
            name: keycloakauthorizations.keycloak.nerc.mghpcc.org
            version: v1
        description: "An OpenShift Operator for managing Keycloak resources, scopes,
          policies,\nand permissions for fine-grained resource permissions.\nAccording
          to the [Keycloak Authorization Services documentation](https://www.keycloak.org/docs/23.0.7/authorization_services/):\n\n-
          A resource can be a web page, a RESTFul resource, a file in your file system,
          or other entity you wish to authorize access to.\n- Scopes usually represent
          the actions that can be performed on a resource, but they are not limited
          to that. You can also use scopes to represent one or more attributes within
          a resource.\n- Policies define the conditions that must be satisfied to
          access or perform operations on something (resource or scope), but they
          are not tied to what they are protecting. They are generic and can be reused
          to build permissions or even more complex policies.\n- Permissions are coupled
          with the resource they are protecting. Here you specify what you want to
          protect (resource or scope) and the policies that must be satisfied to grant
          or deny permission.\n\n## Getting started with the Keycloak Permissions
          Operator\n\nStart by defining the name of your KeycloakAuthorization and
          the same namespace where Keycloak is deployed.\n\n```yaml\napiVersion: keycloak.nerc.mghpcc.org/v1\nkind:
          KeycloakAuthorization\nmetadata:\n  name: example\n  namespace: keycloak\n```\n\n##
          KeycloakAuthorization spec\n\nNext, define the spec. \n\n- `validateCerts:
          false` If your Keycloak does not have valid certificates, if you are deploying
          to OpenShift Local for example. You don't need this line if Keycloak has
          valid certificates.\n- `cleanupUndefinedAuthorizations: true` if you would
          like to remove any undefined authorizations that may have been added manually.
          This is useful for keeping in sync with GitOps. If you do not want the operator
          removing any resources, policies, or permissions, then remove this line.\n\n```yaml\nspec:\n
          \ validateCerts: false\n  cleanupUndefinedAuthorizations: true\n```\n\n##
          Keycloak admin credentials\n\nSpecify the details to connect to keycloak,
          including the `baseUrl`, and the Secret name containing the Keycloak admin
          usernameKey and passwordKey.\n\n```yaml\n  keycloak:\n    baseUrl: https://keycloak-service:8443\n
          \   adminSecret:\n      name: keycloak-initial-admin\n      usernameKey:
          username\n      passwordKey: password\n```\n\n## Realm\n\nSpecify the realm
          id you wish to add authorization configuration to. If you want to add authorization
          configuration to multiple realms, then create a KeycloakAuthorization resource
          for each realm.\n\n```yaml\n  realm:\n    id: NERC\n```\n\n## Clients\n\nFor
          each client in the realm you wish to configure, add authorizationSettings.\n\n```yaml\n
          \   clients:\n      - id: ai4cloudops\n        authorizationSettings:\n```\n\n##
          Authorization Scopes\n\nList the scopes you wish to define for the client.\n\n```yaml\n
          \         scopes:\n            - id: prod\n              name: prod\n              displayName:
          Production\n            - id: test\n              name: test\n              displayName:
          Test\n```\n\n## Authorization Resources\n\nList the resources and corresponding
          scopes you wish to define for the client.\n\n```yaml\n          resources:\n
          \           - name: cluster\n              displayName: cluster\n              scopes:\n
          \               - id: prod\n                  name: prod\n                  displayName:
          Production\n                - id: test\n                  name: test\n                  displayName:
          Test\n```\n\n## Authorization Policies\n\nDefine the policies you wish to
          use in your permissions. You can choose from all the different types of
          permissions, including: aggregated, client, client-scope, group, regex,
          role, time, and user.\n\n```yaml\n          policies:\n            - id:
          client-example\n              name: client-example\n              logic:
          POSITIVE\n              decisionStrategy: UNANIMOUS\n              type:\n
          \               client:\n                  clients:\n                    -
          example\n            - id: group-example\n              name: group-example\n
          \             logic: POSITIVE\n              decisionStrategy: UNANIMOUS\n
          \             type:\n                group:\n                  groups:\n
          \                   - id: example\n```\n\n## Authorization Permissions\n\nDefine
          the permissions, connecting the resources, scopes, and policies together.\n\n```yaml\n
          \         permissions:\n            - name: group-example-cluster-test\n
          \             client: test\n              policy: group-example\n              resource:
          cluster\n            - name: client-example-cluster-test\n              client:
          example\n              policy: client-test\n              resource: cluster\n
          \             scope: test\n```\n\n## Other details\n\nResources, scopes,
          policies, and permissions will all be applied to Keycloak if you have configured
          the admin credentials to Keycloak successfully.\nWhen the KeycloakAuthorization
          is deleted, then the configured resources, policies, and permissions will
          be deleted as well. "
        displayName: Keycloak Permissions Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - fine-grained resource permissions
        - keycloak
        - resources
        - scopes
        - policies
        - permissions
        links:
        - name: Keycloak Permissions Operator
          url: https://keycloak-permissions-operator.domain
        maintainers:
        - email: computate@computate.org
          name: Christopher Tate
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: New England Research Cloud
          url: https://nerc.mghpcc.org/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/nerc-images/keycloak-permissions-operator:1.4.1
        version: 1.4.1
      entries:
      - name: keycloak-permissions-operator.v1.4.1
        version: 1.4.1
      - name: keycloak-permissions-operator.v1.3.0
        version: 1.3.0
      - name: keycloak-permissions-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: keycloak-permissions-operator
    provider:
      name: New England Research Cloud
      url: https://nerc.mghpcc.org/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: mongodb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"opstreelabs.in/v1alpha1","kind":"MongoDB","metadata":{"name":"mongodb"},"spec":{"kubernetesConfig":{"image":"quay.io/opstree/mongo:v5.0.6","imagePullPolicy":"IfNotPresent","securityContext":{"fsGroup":1001}},"storage":{"accessModes":["ReadWriteOnce"],"storageSize":"1Gi","storageClass":"gp2"},"mongoDBSecurity":{"mongoDBAdminUser":"admin","secretRef":{"name":"mongodb-secret","key":"password"}}}},{"apiVersion":"opstreelabs.in/v1alpha1","kind":"MongoDBCluster","metadata":{"name":"mongodb"},"spec":{"clusterSize":3,"kubernetesConfig":{"image":"quay.io/opstree/mongo:v5.0.6","imagePullPolicy":"IfNotPresent","securityContext":{"fsGroup":1001}},"storage":{"accessModes":["ReadWriteOnce"],"storageSize":"1Gi","storageClass":"gp2"},"mongoDBSecurity":{"mongoDBAdminUser":"admin","secretRef":{"name":"mongodb-secret","key":"password"}}}}]'
          capabilities: Seamless Upgrades
          categories: Database,  OpenShift Optional, Storage
          certified: "false"
          containerImage: quay.io/opstree/mongodb-operator:v0.3.0
          createdAt: 05-04-2022
          description: A golang based operator to create and manage MongoDB standalone
            setup and cluster setup
          k8sMaxVersion: ""
          k8sMinVersion: 1.17.0
          repository: https://github.com/ot-container-kit/mongodb-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Mongo DB Cluster
            displayName: MongoDBCluster
            kind: MongoDBCluster
            name: mongodbclusters.opstreelabs.in
            version: v1alpha1
          - description: Mongo DB
            displayName: MongoDB
            kind: MongoDB
            name: mongodbs.opstreelabs.in
            version: v1alpha1
        description: |+
          MongoDB Operator is an operator created in Golang to create, update, and manage MongoDB standalone, replicated, and arbiter replicated setup on Kubernetes and Openshift clusters. This operator is capable of doing the setup for MongoDB with all the required best practices.

          ### Supported Features

          * MongoDB replicated cluster setup
          * MongoDB standalone setup
          * MongoDB replicated cluster failover and recovery
          * Monitoring support with MongoDB Exporter
          * Password based authentication for MongoDB
          * Kubernetes's resources for MongoDB standalone and cluster
          * Customizable configuration changes in MongoDB
          * Insightful Grafana dashboards

          ### Documentation

          **[Documentation](https://ot-mongodb-operator.netlify.app/)**
          ### Getting Help

          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.
          ### Contributing
          You can contribute to this project by:-

          * Raising Bugs and Feature Requests
          * Fixing issue by opening a Pull Request
          * Improving Documentation

          ### License

          MongoDB  Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/mongodb-operator/blob/master/LICENSE)


        displayName: MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - MongoDB
        - Opstree
        - Operatork
        - Cloud Native
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-mongodb-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: alpha
        minKubeVersion: 1.17.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/opstree/mongodb-operator:v0.3.0
        version: 0.3.0
      entries:
      - name: mongodb-operator.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: mongodb-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: https://konveyor.io
    name: move2kube-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: move2kube-operator.v0.3.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {
                  "deployment": {
                    "api": {
                      "cpu": "300m",
                      "imageTag": "",
                      "initContainer": false,
                      "memory": "1Gi",
                      "privilegedPods": false,
                      "replicas": 1
                    },
                    "authServer": {
                      "cpu": "300m",
                      "enable": false,
                      "memory": "1Gi",
                      "replicas": 1
                    },
                    "database": {
                      "cpu": "300m",
                      "enable": false,
                      "memory": "1Gi",
                      "replicas": 1,
                      "startEmpty": true
                    }
                  },
                  "ingress": {
                    "enable": true,
                    "host": "mydomain.com",
                    "preferRoute": false,
                    "tlsSecretName": ""
                  },
                  "persistentVolumeClaim": {
                    "createNew": true,
                    "enable": false,
                    "name": "",
                    "readWriteMany": true,
                    "storageClassName": "",
                    "storageSize": "1Gi"
                  },
                  "secret": {
                    "api": {
                      "configYAML": "",
                      "createNew": true,
                      "enable": false,
                      "name": ""
                    },
                    "authServer": {
                      "adminPassword": "password",
                      "adminUsername": "admin",
                      "createNew": true,
                      "databasePassword": "password",
                      "databaseUsername": "auth-server",
                      "name": "",
                      "realmJSON": "",
                      "standaloneHAXML": ""
                    }
                  },
                  "securityContextContraints": {
                    "enable": false
                  },
                  "serviceAccount": {
                    "createNew": true,
                    "enable": false,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.10
          createdAt: "2023-09-25T02:56:15Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/konveyor/move2kube-operator:v0.3.10
        version: 0.3.10
      entries:
      - name: move2kube-operator.v0.3.10
        version: 0.3.10
      - name: move2kube-operator.v0.3.10-rc.0
        version: 0.3.10-rc.0
      - name: move2kube-operator.v0.3.9
        version: 0.3.9
      - name: move2kube-operator.v0.3.8
        version: 0.3.8
      name: alpha
    - currentCSV: move2kube-operator.v0.3.15-rc.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.15-rc.0
          createdAt: "2024-10-18T16:10:56Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/konveyor/move2kube-operator:v0.3.15-rc.0
        version: 0.3.15-rc.0
      entries:
      - name: move2kube-operator.v0.3.15-rc.0
        version: 0.3.15-rc.0
      - name: move2kube-operator.v0.3.14-rc.0
        version: 0.3.14-rc.0
      - name: move2kube-operator.v0.3.13-rc.0
        version: 0.3.13-rc.0
      - name: move2kube-operator.v0.3.11-rc.0
        version: 0.3.11-rc.0
      name: prerelease
    - currentCSV: move2kube-operator.v0.3.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.15
          createdAt: "2025-03-07T06:09:48Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/konveyor/move2kube-operator:v0.3.15
        version: 0.3.15
      entries:
      - name: move2kube-operator.v0.3.15
        version: 0.3.15
      - name: move2kube-operator.v0.3.14
        version: 0.3.14
      - name: move2kube-operator.v0.3.13
        version: 0.3.13
      - name: move2kube-operator.v0.3.11
        version: 0.3.11
      name: stable
    defaultChannel: stable
    packageName: move2kube-operator
    provider:
      name: Konveyor
      url: https://konveyor.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Dell Technologies
      provider-url: https://github.com/dell/csm-operator
    name: dell-csm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dell-csm-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "authorization",
                  "namespace": "authorization"
                },
                "spec": {
                  "modules": [
                    {
                      "components": [
                        {
                          "enabled": true,
                          "name": "nginx"
                        },
                        {
                          "enabled": true,
                          "name": "cert-manager"
                        },
                        {
                          "authorizationController": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231",
                          "authorizationControllerReplicas": 1,
                          "certificate": "",
                          "controllerReconcileInterval": "5m",
                          "enabled": true,
                          "hostname": "csm-authorization.com",
                          "leaderElection": true,
                          "name": "proxy-server",
                          "opa": "docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095",
                          "opaKubeMgmt": "docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517",
                          "openTelemetryCollectorAddress": "",
                          "privateKey": "",
                          "proxyServerIngress": [
                            {
                              "annotations": {},
                              "hosts": [],
                              "ingressClassName": "nginx"
                            }
                          ],
                          "proxyService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682",
                          "proxyServiceReplicas": 1,
                          "roleService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c",
                          "roleServiceReplicas": 1,
                          "storageService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5",
                          "storageServiceReplicas": 1,
                          "tenantService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358",
                          "tenantServiceReplicas": 1
                        },
                        {
                          "commander": "docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a",
                          "name": "redis",
                          "redis": "docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f",
                          "redisCommander": "rediscommander",
                          "redisName": "redis-csm",
                          "redisReplicas": 5,
                          "redisSecretProviderClass": [
                            {
                              "redisPasswordKey": "",
                              "redisSecretName": "",
                              "redisUsernameKey": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "sentinel": "sentinel"
                        },
                        {
                          "configSecretProviderClass": [
                            {
                              "configSecretName": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "name": "config"
                        },
                        {
                          "name": "storage-system-credentials",
                          "secretProviderClasses": {
                            "vault": [
                              "secret-provider-class-1",
                              "secret-provider-class-2"
                            ]
                          }
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": true,
                      "forceRemoveModule": true,
                      "name": "authorization-proxy-server"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "isilon",
                  "namespace": "isilon"
                },
                "spec": {
                  "driver": {
                    "authSecret": "isilon-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VERBOSE",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_ISI_PORT",
                          "value": "8080"
                        },
                        {
                          "name": "X_CSI_ISI_PATH",
                          "value": "/ifs/data/csi"
                        },
                        {
                          "name": "X_CSI_ISI_NO_PROBE_ON_START",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_AUTOPROBE",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_AUTH_TYPE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_CUSTOM_TOPOLOGY_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOISILON_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "AZ_RECONCILE_INTERVAL",
                          "value": "1h"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_ISI_QUOTA_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_ACCESS_ZONE",
                          "value": "System"
                        },
                        {
                          "name": "X_CSI_ISI_VOLUME_PATH_PERMISSIONS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_IGNORE_UNRESOLVABLE_HOSTS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "isilon",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerscale"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSCALE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_PERFORMANCE_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSCALE_QUOTA_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "ISICLIENT_INSECURE",
                              "value": "true"
                            },
                            {
                              "name": "ISICLIENT_AUTH_TYPE",
                              "value": "1"
                            },
                            {
                              "name": "ISICLIENT_VERBOSE",
                              "value": "0"
                            },
                            {
                              "name": "POWERSCALE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSCALE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0",
                          "name": "metrics-powerscale"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-isilon/csi_sock",
                            "--mode=node",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powermax",
                  "namespace": "powermax"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powermax-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_MANAGED_ARRAYS",
                          "value": "000000000001,000000000002"
                        },
                        {
                          "name": "X_CSI_POWERMAX_PORTGROUPS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_TRANSPORT_PROTOCOL",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_K8S_CLUSTER_PREFIX",
                          "value": "CSM"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "X_CSI_VSPHERE_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VSPHERE_PORTGROUP",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VSPHERE_HOSTNAME",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VCENTER_HOST",
                          "value": ""
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "X_CSI_POWERMAX_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "X_CSI_POWERMAX_DEBUG",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_REVPROXY_USE_SECRET",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powermax",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERMAX_ISCSI_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_TOPOLOGY_CONTROL_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": [
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/memory-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/disk-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/network-unavailable",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REVPROXY_TLS_SECRET",
                              "value": "csirevproxy-tls-secret"
                            },
                            {
                              "name": "X_CSI_REVPROXY_PORT",
                              "value": "2222"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            },
                            {
                              "name": "DeployAsSidecar",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265",
                          "name": "csipowermax-reverseproxy"
                        }
                      ],
                      "configVersion": "v2.14.0",
                      "name": "csireverseproxy"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powermax"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERMAX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERMAX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414",
                          "name": "metrics-powermax"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/powermax.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powerstore",
                  "namespace": "powerstore"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powerstore-config",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_NODE_NAME_PREFIX",
                          "value": "csi-node"
                        },
                        {
                          "name": "X_CSI_FC_PORTS_FILTER_FILE_PATH",
                          "value": "/etc/fc-ports-filter"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOPOWERSTORE_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSM_AUTH_ENABLED",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_MULTI_NAS_FAILURE_THRESHOLD",
                          "value": "5"
                        },
                        {
                          "name": "X_CSI_MULTI_NAS_COOLDOWN_PERIOD",
                          "value": "5m"
                        },
                        {
                          "name": "X_CSI_NFS_ACLS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerstore",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-powerstore.dellemc.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerstore"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSTORE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_SPACE_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_ARRAY_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_FILE_SYSTEM_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSTORE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSTORE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "ZIPKIN_URI",
                              "value": ""
                            },
                            {
                              "name": "ZIPKIN_SERVICE_NAME",
                              "value": "metrics-powerstore"
                            },
                            {
                              "name": "ZIPKIN_PROBABILITY",
                              "value": "0.0"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60",
                          "name": "metrics-powerstore"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "unity",
                  "namespace": "unity"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_UNITY_ALLOW_MULTI_POD_ACCESS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_EPHEMERAL_STAGING_PATH",
                          "value": "/var/lib/kubelet/plugins/kubernetes.io/csi/pv/"
                        },
                        {
                          "name": "X_CSI_ISCSI_CHROOT",
                          "value": "/noderoot"
                        },
                        {
                          "name": "X_CSI_UNITY_SYNC_NODEINFO_INTERVAL",
                          "value": "15"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOUNITY_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOUNITY_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "TENANT_NAME",
                          "value": ""
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_UNITY_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "unity",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "vxflexos",
                  "namespace": "vxflexos"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLELISTVOLUMESNAPSHOT",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLESNAPSHOTCGDELETE",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "INFO"
                        },
                        {
                          "name": "GOSCALEIO_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOSCALEIO_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_QUOTA_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "INTERFACE_NAMES",
                          "value": null
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERFLEX_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "File",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerflex",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "initContainers": [
                      {
                        "envs": [
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          },
                          {
                            "name": "REPO_ADDRESS",
                            "value": "sftp://0.0.0.0"
                          },
                          {
                            "name": "REPO_USER",
                            "value": "sdcRepoUser"
                          },
                          {
                            "name": "MODULE_SIGCHECK",
                            "value": "0"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "imagePullPolicy": "IfNotPresent",
                        "name": "sdc"
                      }
                    ],
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_SDC_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_APPROVE_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_PREFIX",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_SDC_SFTP_REPO_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 1,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "enabled": false,
                        "envs": [
                          {
                            "name": "HOST_PID",
                            "value": "1"
                          },
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "name": "sdc-monitor"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "csi-external-health-monitor-controller"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERFLEX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_SDC_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_SDC_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERFLEX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d",
                          "name": "metrics-powerflex"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerflex"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--csisock=unix:/var/lib/kubelet/plugins/vxflexos.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
          createdAt: "2025-09-05T13:18:42Z"
          description: Easily install and manage Dell’s CSI Drivers and CSM
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/dell/csm-operator
          support: Dell Technologies
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ContainerStorageModule is the Schema for the containerstoragemodules
              API
            displayName: Container Storage Module
            kind: ContainerStorageModule
            name: containerstoragemodules.storage.dell.com
            version: v1
        description: "Dell Container Storage Modules (CSM) Operator is a Kubernetes
          Operator which can be \nused to install and manage Dell’s CSI drivers and
          CSM modules. \nBy using Dell CSM Operator, enterprises can quickly and easily
          deploy the \nCSM modules for CSI drivers making it easy for DevOps \nteams
          to build and optimize their Kubernetes workloads using enterprise grade
          storage \nwhile consistently aligning with the desired outcomes of the organization.\nThis
          operator is available as both community and certified operator for Openshift
          \nand as a community operator for upstream Kubernetes.\nOperator can be
          deployed via OperatorHub.io. using Operator Lifecycle Manager (OLM)\n\n###
          About Dell CSM\n\nDell CSM brings powerful enterprise storage features and
          functionality to Kubernetes for \neasier adoption of cloud native workloads,
          improved productivity and scalable operations. \n\nCSM delivers software
          modules for storage management that provide developers access to \nbuild
          automation for enhanced IT needs and other critical enterprise storage features.
          \n\nCurrently, the following modules are available as part of Dell CSM.
          \n*\tObservability - Delivers a high-level view of the storage capacity
          \n  and performance usage via Grafana dashboards to the Kubernetes users\n*\tReplication
          - Enables array replication capabilities for Kubernetes users\n*\tAuthorization
          - Provides both storage and Kubernetes administrators the ability \n  to
          apply RBAC for Dell CSI Drivers\n*\tResiliency - Designed to make Kubernetes
          applications more resilient to \n  node failures\n\nThese modules are available
          for various storage platforms in the Dell portfolio, \nsuch as Dell PowerStore,
          Dell PowerMax, Dell PowerFlex, Dell PowerScale, and Dell Unity XT.\n\nFor
          more information, visit [Dell’s Container Storage Module web page](https://www.dell.com/en-us/dt/solutions/kubernetes-containers/kubernetes-storage.htm)\n\nAdditional
          Materials\n*\t[Documentation](https://dell.github.io/csm-docs)\n*\t[CSM
          Solution Brief](https://www.delltechnologies.com/asset/en-us/products/storage/briefs-summaries/csm-solution-brief.pdf)\n"
        displayName: Dell Container Storage Modules
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dell Container Storage Modules
        - Dell CSI Driver
        - Dell CSM Modules
        - Powerflex
        - Powerscale
        - Powerstore
        - Unity
        - Authorization
        - Observability
        - Replication
        links:
        - name: Documentation
          url: https://dell.github.io/csm-docs
        maintainers:
        - email: container.storage.modules@dell.com
          name: Dell Container Storage Modules
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Dell Technologies
          url: https://github.com/dell/csm-operator
        relatedImages:
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5
        - registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334
        - docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517
        - registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
        - registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:59335acaca515fceed5f0350c46cfcfd6037475f0cf35ae8ab41d37333039e2e
        - registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d
        - registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60
        - ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0
        - registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265
        - registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414
        - docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f
        - quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540
        - docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231
        - registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16
        - registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        version: 1.10.0
      entries:
      - name: dell-csm-operator.v1.10.0
        version: 1.10.0
      - name: dell-csm-operator.v1.9.1
        version: 1.9.1
      - name: dell-csm-operator.v1.9.0
        version: 1.9.0
      - name: dell-csm-operator.v1.8.1
        version: 1.8.1
      - name: dell-csm-operator.v1.8.0
        version: 1.8.0
      - name: dell-csm-operator.v1.7.0
        version: 1.7.0
      - name: dell-csm-operator.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: dell-csm-operator
    provider:
      name: Dell Technologies
      url: https://github.com/dell/csm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: raftt.io
      provider-url: https:///www.raftt.io
    name: raftt-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: raftt-operator.v1.25.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "raftt.io/v1beta1",
                "kind": "Raftt",
                "metadata": {
                  "name": "raftt",
                  "namespace": "raftt"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/raftt/raftt-operator@sha256:73edc6d8416ee3815d7e59bb26c33e3d35d97cf1ceadd8fa457041be13e175ca
          createdAt: "2025-04-29T11:31:02Z"
          description: This operator installs and configures a Raftt instance
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/raftt-operator/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/raftt-operator/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: raftt
          operators.openshift.io/valid-subscription: Raftt commercial license
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Raftt systems
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Raftt represents an installation of Raftt cluster controller
              & supporting services
            displayName: Raftt
            kind: Raftt
            name: raftts.raftt.io
            version: v1beta1
        description: |
          Raftt allows fast code iterations and debugging over complex Kubernetes environments.
          Raftt provides a localhost development experience for applications running on Kubernetes.
          You can use Raftt to orchestrate the environments for you, so you get the optimal development experience -
          iterations that take under a second, OOTB debugging of your code as it runs in the cluster, easy introspection, and more.
        displayName: Raftt
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dx
        - development
        - debugging
        links:
        - name: Raftt Operator
          url: https://raftt-operator.domain
        - name: Documentation
          url: https://docs.raftt.io
        maintainers:
        - email: ops@raftt.io
          name: Raftt
        maturity: alpha
        provider:
          name: raftt.io
          url: https:///www.raftt.io
        relatedImages:
        - quay.io/raftt/raftt-operator@sha256:73edc6d8416ee3815d7e59bb26c33e3d35d97cf1ceadd8fa457041be13e175ca
        version: 1.25.2
      entries:
      - name: raftt-operator.v1.25.2
        version: 1.25.2
      name: alpha
    defaultChannel: alpha
    packageName: raftt-operator
    provider:
      name: raftt.io
      url: https:///www.raftt.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HashiCorp
      provider-url: https://www.hashicorp.com/
    name: vault-secrets-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vault-secrets-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultConnection",
                "metadata": {
                  "name": "vaultconnection-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "address": "http://vault.vault.svc.cluster.local:8200"
                }
              },
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultAuth",
                "metadata": {
                  "name": "vaultauth-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "vaultConnectionRef": "vaultconnection-sample",
                  "method": "kubernetes",
                  "mount": "kubernetes",
                  "kubernetes": {
                    "role": "sample",
                    "serviceAccount": "default"
                  }
                }
              },
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultStaticSecret",
                "metadata": {
                  "name": "vaultstaticsecret-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "vaultAuthRef": "vaultauth-sample",
                  "mount": "kvv2",
                  "type": "kv-v2",
                  "path": "secret",
                  "refreshAfter": "5s",
                  "destination": {
                    "create": true,
                    "name": "secret1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.connect.redhat.com/hashicorp/vault-secrets-operator@sha256:84ac7003c4cb90c35f1e296585bb58187a86d64f972133e2661839bdb532ddb8
          createdAt: "2025-09-26T21:50:51Z"
          description: The Vault Secrets Operator (VSO) allows Pods to consume Vault
            secrets natively from Kubernetes Secrets.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/hashicorp/vault-secrets-operator
          support: HashiCorp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CSISecrets
            name: csisecrets.secrets.hashicorp.com
            version: v1beta1
          - description: HCPAuth is the Schema for the hcpauths API
            displayName: HCPAuth
            kind: HCPAuth
            name: hcpauths.secrets.hashicorp.com
            version: v1beta1
          - description: HCPVaultSecretsApp is the Schema for the hcpvaultsecretsapps
              API
            displayName: HCPVault Secrets App
            kind: HCPVaultSecretsApp
            name: hcpvaultsecretsapps.secrets.hashicorp.com
            version: v1beta1
          - description: SecretTransformation is the Schema for the secrettransformations
              API
            displayName: Secret Transformation
            kind: SecretTransformation
            name: secrettransformations.secrets.hashicorp.com
            version: v1beta1
          - description: VaultAuthGlobal is the Schema for the vaultauthglobals API
            displayName: Vault Auth Global
            kind: VaultAuthGlobal
            name: vaultauthglobals.secrets.hashicorp.com
            version: v1beta1
          - description: VaultAuth is the Schema for the vaultauths API
            displayName: Vault Auth
            kind: VaultAuth
            name: vaultauths.secrets.hashicorp.com
            version: v1beta1
          - description: VaultConnection is the Schema for the vaultconnections API
            displayName: Vault Connection
            kind: VaultConnection
            name: vaultconnections.secrets.hashicorp.com
            version: v1beta1
          - description: VaultDynamicSecret is the Schema for the vaultdynamicsecrets
              API
            displayName: Vault Dynamic Secret
            kind: VaultDynamicSecret
            name: vaultdynamicsecrets.secrets.hashicorp.com
            version: v1beta1
          - description: VaultPKISecret is the Schema for the vaultpkisecrets API
            displayName: Vault PKISecret
            kind: VaultPKISecret
            name: vaultpkisecrets.secrets.hashicorp.com
            version: v1beta1
          - description: VaultStaticSecret is the Schema for the vaultstaticsecrets
              API
            displayName: Vault Static Secret
            kind: VaultStaticSecret
            name: vaultstaticsecrets.secrets.hashicorp.com
            version: v1beta1
        description: |-
          The Vault Secrets Operator (VSO) allows Pods to consume Vault secrets
          natively from Kubernetes Secrets.

          ## Overview

          The Vault Secrets Operator operates by watching for changes to its supported set of Custom Resource Definitions (CRD).
          Each CRD provides the specification required to allow the *Operator* to synchronize a Vault Secrets to a Kubernetes Secret.
          The *Operator* writes the *source* Vault secret data directly to the *destination* Kubernetes Secret, ensuring that any
          changes made to the *source* are replicated to the *destination* over its lifetime. In this way, an application only needs
          to have access to the *destination* secret in order to make use of the secret data contained within.

          See the developer docs for more info [here](https://developer.hashicorp.com/vault/docs/platform/k8s/vso), including
          [examples](https://developer.hashicorp.com/vault/docs/platform/k8s/vso/examples) and a
          [tutorial](https://developer.hashicorp.com/vault/tutorials/kubernetes/vault-secrets-operator)
          for getting started.
        displayName: Vault Secrets Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vault
        - security
        - secrets
        links:
        - name: Vault Secrets Operator
          url: https://github.com/hashicorp/vault-secrets-operator
        - name: Documentation
          url: https://developer.hashicorp.com/vault/docs/platform/k8s/vso
        maintainers:
        - email: support@hashicorp.com
          name: HashiCorp Support
        maturity: stable
        provider:
          name: HashiCorp
          url: https://www.hashicorp.com/
        relatedImages:
        - registry.connect.redhat.com/hashicorp/vault-secrets-operator@sha256:84ac7003c4cb90c35f1e296585bb58187a86d64f972133e2661839bdb532ddb8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:503d5ee0fbcd5bd3492785f4246e19cbdb92960be0d05f3c6f123a32fe0fe84d
        version: 1.0.1
      entries:
      - name: vault-secrets-operator.v1.0.1
        version: 1.0.1
      - name: vault-secrets-operator.v1.0.0
        version: 1.0.0
      - name: vault-secrets-operator.v0.10.0
        version: 0.10.0
      - name: vault-secrets-operator.v0.9.1
        version: 0.9.1
      - name: vault-secrets-operator.v0.9.0
        version: 0.9.0
      - name: vault-secrets-operator.v0.8.1
        version: 0.8.1
      - name: vault-secrets-operator.v0.8.0
        version: 0.8.0
      - name: vault-secrets-operator.v0.7.0
        version: 0.7.0
      - name: vault-secrets-operator.v0.6.0
        version: 0.6.0
      - name: vault-secrets-operator.v0.5.2
        version: 0.5.2
      - name: vault-secrets-operator.v0.5.1
        version: 0.5.1
      - name: vault-secrets-operator.v0.5.0
        version: 0.5.0
      - name: vault-secrets-operator.v0.4.3
        version: 0.4.3
      name: stable
    defaultChannel: stable
    packageName: vault-secrets-operator
    provider:
      name: HashiCorp
      url: https://www.hashicorp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ab Initio
      provider-url: ""
    name: abinitio-runtime-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: abinitio-runtime-operator.v4.3.3-1
      currentCSVDesc:
        annotations:
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: abinitio-system
          operators.operatorframework.io/builder: operator-sdk-v1.8.0-ocp
          operators.operatorframework.io/internal-objects: '["cooperatingsystemruntimepoolmetrics.cloud.abinitio.com",
            "cooperatingsystemruntimeprofiles.cloud.abinitio.com"]'
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Defines a *runtime claim*, an interface for reserving a runtime
              object (CoOperatingSystemRuntime) in a runtime pool (CoOperatingSystemRuntimePool)
              in order to launch a job.
            displayName: Co>Operating System Runtime Claim
            kind: CoOperatingSystemRuntimeClaim
            name: cooperatingsystemruntimeclaims.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Pool Metric
            kind: CoOperatingSystemRuntimePoolMetric
            name: cooperatingsystemruntimepoolmetrics.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime pool*, a set of templated runtime objects
              (CoOperatingSystemRuntime) that are available on demand for running
              jobs. A runtime pool's capacity is configurable and can be scheduled
              in response to usage patterns.
            displayName: Co>Operating System Runtime Pool
            kind: CoOperatingSystemRuntimePool
            name: cooperatingsystemruntimepools.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Profile
            kind: CoOperatingSystemRuntimeProfile
            name: cooperatingsystemruntimeprofiles.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime object*, a set of coordinated resources
              used to run a data processing job in parallel across a launcher pod
              and a dynamically scaled set of worker pods. Such a job leverages the
              full power of the Co>Operating System for processing complex enterprise
              data.
            displayName: Co>Operating System Runtime
            kind: CoOperatingSystemRuntime
            name: cooperatingsystemruntimes.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime template*, a reusable specification (CoOperatingSystemRuntimeSpec)
              that can be referenced and overridden by runtime objects as needed,
              making it possible to specify jobs concisely. The runtime objects in
              a runtime pool (CoOperatingSystemRuntimePool) are created from a runtime
              template.
            displayName: Co>Operating System Runtime Template
            kind: CoOperatingSystemRuntimeTemplate
            name: cooperatingsystemruntimetemplates.cloud.abinitio.com
            version: v1
        description: |-
          The Co>Operating System runtime operator brings the power of Ab Initio's Co>Operating
          System to Kubernetes by defining a set of custom API resources for provisioning and
          running data processing jobs.

          The operator is installed along with the following custom resource definitions (CRDs):

          - CoOperatingSystemRuntime
          - CoOperatingSystemRuntimeClaim
          - CoOperatingSystemRuntimePool
          - CoOperatingSystemRuntimeTemplate

          In combination, these resources give you significant flexibility in scheduling, running,
          and monitoring your jobs, which can run in parallel across a dynamically scaled set of
          containers. For reference documentation for these resources, see
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-FOUR-Reference/Co-Operating-System-runtime-operator-API-reference)
          in the Ab Initio Help.

          The following topics describe the basic requirements and procedures for creating custom
          resources for running jobs after you have installed the runtime operator.

          ## Requirements for using custom resources

          The runtime operator is fully configured when you install it in your OpenShift cluster
          through OperatorHub. Importantly, the operator is installed in unrestricted mode,
          which means it can manage the CoOperatingSystemRuntime family of CRDs in any namespace
          in your cluster.

          For creating custom resources in order to run jobs using the operator, you must meet
          the requirements and configure the settings described in the following sections.

          ### Container images for job agents

          To run jobs of kind CoOperatingSystemRuntime, you must have a way to build the following
          OCI container images:

          - A *processing* image with the Co>Operating System server software installed in it.
          - A *sandbox workspace* image with the graphs and other files needed to run your jobs.
          This image is based on the Co>Operating System image and functions as the job's launcher
          agent, which coordinates with a set of worker agents (processing containers) in order to
          run a job in parallel.

          For building and managing these container images, you need the following:

          - A Co>Operating System installation package.
          - Access to a technical repository containing the projects that make up your application.
          - A container-building tool.
          - Access to a private container registry for hosting the images you build. You need pull
          and push access to this registry from your container build context and pull access from
          your OpenShift cluster (for example, via a Kubernetes Secret).

          > WARNING! Never push container images with Ab Initio software to a public registry. Push
          them to a private container registry only.

          The Co>Operating System includes reference Dockerfiles and example scripts for building
          container images. Typically, you build these images as part of your CI/CD pipeline. For
          more information, contact your Ab Initio administrator or Ab Initio Support.

          ### Ab Initio key servers

          The key servers that you use to enable jobs must be reachable from your cluster. You
          specify these key servers via the `AB_HOSTNAME_KEYSERVER_URLS` configuration variable in
          the pod specifications for launcher and worker agents in the CoOperatingSystemRuntime
          resource.

          To discuss the software activation key requirements for jobs that run in Kubernetes,
          contact your key server administrator or Ab Initio Support.

          ### Namespace and role considerations

          The runtime operator is installed with the cluster-wide RBAC roles and role bindings that
          it needs in order to manage custom resources and run jobs.

          You typically use a separate namespace or namespaces for creating and managing custom
          resources to run jobs. Each of these namespaces must be equipped with a service
          account that can list pod limits and get and patch CoOperatingSystemRuntime resources.

          For each namespace in which you intend to create custom resources, create the needed RBAC
          role and role binding by applying the following Role and RoleBinding manifests:

          ```yaml
          apiVersion: rbac.authorization.k8s.io/v1
          kind: Role
          metadata:
            name: cooperatingsystemruntime-role
          rules:
          - apiGroups:
            - ""
            resources:
            - limitranges
            verbs:
            - list
          - apiGroups:
            - cloud.abinitio.com
            resources:
            - cooperatingsystemruntimes
            verbs:
            - get
            - patch
          ---
          apiVersion: rbac.authorization.k8s.io/v1
          kind: RoleBinding
          metadata:
            name: run-cooperatingsystemruntime-rolebinding
          subjects:
          - kind: ServiceAccount
            name: default
          roleRef:
            kind: Role
            name: cooperatingsystemruntime-role
            apiGroup: rbac.authorization.k8s.io
          ```

          One way to do this is to save these manifests to a file and specify a namespace when
          using `oc` to apply them to the cluster. In the following example, the manifests are
          saved to a file named `runtime-rbac.yaml` and the role and role binding are to be created
          for the `default` service account in the `hello-world` namespace:

          ```
          oc -n hello-world apply -f runtime-rbac.yaml
          ```

          To bind the role to a service account other than `default`, simply update the value of
          `subjects[0].name` in the provided RoleBinding manifest.

          ### Dynamic volume provisioning

          The launcher and worker template sections of the CoOperatingSystemRuntime resource
          require you to specify persistent volume claims (PVCs) for job recovery data. To ensure
          that these PVCs can be bound to approriate storage volumes, dynamic volume provisioning
          must be enabled in your cluster.

          Work with your cluster administrator to identify the storage needs of your jobs and to
          ensure that dynamic volume provisioning is enabled so that PVCs can be matched to
          persistent volumes of the right storage class.

          For more information about dynamic storage provisioning, see the
          [OpenShift documentation](https://docs.openshift.com/container-platform/latest/storage/dynamic-provisioning.html).

          ## Quick start and examples

          For an overview of the key concepts associated with running jobs in Kubernetes, see
          ["About running jobs in Kubernetes"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-TWO-Running-jobs-in-Kubernetes/About-running-jobs-in-Kubernetes)
          in the Ab Initio Help.

          As described in this topic, you might begin by creating a CoOperatingSystemRuntime
          manifest that defines the configuration and compute resources needed for a standalone test
          job. You could then use the specification for this job to create a CoOperatingSystemRuntimeTemplate
          resource, which would collect common settings and make it possible to concisely specify
          CoOperatingSystemRuntime jobs.

          You could also use a runtime pool (of kind CoOperatingSystemRuntime) to extend these basic
          resources: such a pool consists of an actively running group of templated runtime objects whose
          pods can be used to run jobs on demand. Finally, for reserving resources in a pool, you could
          create a CoOperatingSystemRuntimeClaim resource.

          For more information about these custom resources, refer to the
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-FOUR-Reference/Co-Operating-System-runtime-operator-API-reference).

          For a set of integrated examples (including complete YAML manifests) that show you how to
          create and use the custom resources, see ["Example manifests for running jobs"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-TWO-Running-jobs-in-Kubernetes/Configuring-and-running-Co-Operating-System-jobs-in-Kubernetes/Example-manifests-for-CoOperatingSystemRuntime-jobs).
        displayName: Co>Operating System Runtime Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - abinitio
        - runtime
        - operator
        links:
        - name: Ab Initio
          url: https://www.abinitio.com/
        maintainers:
        - email: support@abinitio.com
          name: Ab Initio Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Ab Initio
        relatedImages:
        - registry.connect.redhat.com/abinitio/cooperating-system-runtime-operator@sha256:665efa8cc2ba97ad31f219161cdd53e15c1aef16b6c05ac60506be9bbc9ffc9a
        version: 4.3.3-1
      entries:
      - name: abinitio-runtime-operator.v4.3.3-1
        version: 4.3.3-1
      - name: abinitio-runtime-operator.v4.3.1-0
        version: 4.3.1-0
      name: release-4.3
    - currentCSV: abinitio-runtime-operator.v4.2.3-0
      currentCSVDesc:
        annotations:
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: abinitio-system
          operators.operatorframework.io/builder: operator-sdk-v1.8.0-ocp
          operators.operatorframework.io/internal-objects: '["cooperatingsystemruntimepoolmetrics.cloud.abinitio.com",
            "cooperatingsystemruntimeprofiles.cloud.abinitio.com"]'
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Defines a *runtime claim*, an interface for reserving a runtime
              object (CoOperatingSystemRuntime) in a runtime pool (CoOperatingSystemRuntimePool)
              in order to launch a job.
            displayName: Co>Operating System Runtime Claim
            kind: CoOperatingSystemRuntimeClaim
            name: cooperatingsystemruntimeclaims.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Pool Metric
            kind: CoOperatingSystemRuntimePoolMetric
            name: cooperatingsystemruntimepoolmetrics.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime pool*, a set of templated runtime objects
              (CoOperatingSystemRuntime) that are available on demand for running
              jobs. A runtime pool's capacity is configurable and can be scheduled
              in response to usage patterns.
            displayName: Co>Operating System Runtime Pool
            kind: CoOperatingSystemRuntimePool
            name: cooperatingsystemruntimepools.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Profile
            kind: CoOperatingSystemRuntimeProfile
            name: cooperatingsystemruntimeprofiles.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime object*, a set of coordinated resources
              used to run a data processing job in parallel across a launcher pod
              and a dynamically scaled set of worker pods. Such a job leverages the
              full power of the Co>Operating System for processing complex enterprise
              data.
            displayName: Co>Operating System Runtime
            kind: CoOperatingSystemRuntime
            name: cooperatingsystemruntimes.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime template*, a reusable specification (CoOperatingSystemRuntimeSpec)
              that can be referenced and overridden by runtime objects as needed,
              making it possible to specify jobs concisely. The runtime objects in
              a runtime pool (CoOperatingSystemRuntimePool) are created from a runtime
              template.
            displayName: Co>Operating System Runtime Template
            kind: CoOperatingSystemRuntimeTemplate
            name: cooperatingsystemruntimetemplates.cloud.abinitio.com
            version: v1
        description: |-
          The Co>Operating System runtime operator brings the power of Ab Initio's Co>Operating
          System to Kubernetes by defining a set of custom API resources for provisioning and
          running data processing jobs.

          The operator is installed along with the following custom resource definitions (CRDs):

          - CoOperatingSystemRuntime
          - CoOperatingSystemRuntimeClaim
          - CoOperatingSystemRuntimePool
          - CoOperatingSystemRuntimeTemplate

          In combination, these resources give you significant flexibility in scheduling, running,
          and monitoring your jobs, which can run in parallel across a dynamically scaled set of
          containers. For reference documentation for these resources, see
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Co_Operating_System_runtime_operator_API_reference)
          in the Ab Initio Help.

          The following topics describe the basic requirements and procedures for creating custom
          resources for running jobs after you have installed the runtime operator.

          ## Requirements for using custom resources

          The runtime operator is fully configured when you install it in your OpenShift cluster
          through OperatorHub. Importantly, the operator is installed in unrestricted mode,
          which means it can manage the CoOperatingSystemRuntime family of CRDs in any namespace
          in your cluster.

          For creating custom resources in order to run jobs using the operator, you must meet
          the requirements and configure the settings described in the following sections.

          ### Container images for job agents

          To run jobs of kind CoOperatingSystemRuntime, you must have a way to build the following
          OCI container images:

          - A *processing* image with the Co>Operating System server software installed in it.
          - A *sandbox workspace* image with the graphs and other files needed to run your jobs.
          This image is based on the Co>Operating System image and functions as the job's launcher
          agent, which coordinates with a set of worker agents (processing containers) in order to
          run a job in parallel.

          For building and managing these container images, you need the following:

          - A Co>Operating System installation package.
          - Access to a technical repository containing the projects that make up your application.
          - A container-building tool.
          - Access to a private container registry for hosting the images you build. You need pull
          and push access to this registry from your container build context and pull access from
          your OpenShift cluster (for example, via a Kubernetes Secret).

          > WARNING! Never push container images with Ab Initio software to a public registry. Push
          them to a private container registry only.

          The Co>Operating System includes reference Dockerfiles and example scripts for building
          container images. Typically, you build these images as part of your CI/CD pipeline. For
          more information, contact your Ab Initio administrator or Ab Initio Support.

          ### Ab Initio key servers

          The key servers that you use to enable jobs must be reachable from your cluster. You
          specify these key servers via the `AB_HOSTNAME_KEYSERVER_URLS` configuration variable in
          the pod specifications for launcher and worker agents in the CoOperatingSystemRuntime
          resource.

          To discuss the software activation key requirements for jobs that run in Kubernetes,
          contact your key server administrator or Ab Initio Support.

          ### Namespace and role considerations

          The runtime operator is installed with the cluster-wide RBAC roles and role bindings that
          it needs in order to manage custom resources and run jobs.

          You typically use a separate namespace or namespaces for creating and managing custom
          resources to run jobs. Each of these namespaces must be equipped with a service
          account that can list pod limits and get and patch CoOperatingSystemRuntime resources.

          For each namespace in which you intend to create custom resources, create the needed RBAC
          role and role binding by applying the following Role and RoleBinding manifests:

          ```yaml
          apiVersion: rbac.authorization.k8s.io/v1
          kind: Role
          metadata:
            name: cooperatingsystemruntime-role
          rules:
          - apiGroups:
            - ""
            resources:
            - limitranges
            verbs:
            - list
          - apiGroups:
            - cloud.abinitio.com
            resources:
            - cooperatingsystemruntimes
            verbs:
            - get
            - patch
          ---
          apiVersion: rbac.authorization.k8s.io/v1
          kind: RoleBinding
          metadata:
            name: run-cooperatingsystemruntime-rolebinding
          subjects:
          - kind: ServiceAccount
            name: default
          roleRef:
            kind: Role
            name: cooperatingsystemruntime-role
            apiGroup: rbac.authorization.k8s.io
          ```

          One way to do this is to save these manifests to a file and specify a namespace when
          using `oc` to apply them to the cluster. In the following example, the manifests are
          saved to a file named `runtime-rbac.yaml` and the role and role binding are to be created
          for the `default` service account in the `hello-world` namespace:

          ```
          oc -n hello-world apply -f runtime-rbac.yaml
          ```

          To bind the role to a service account other than `default`, simply update the value of
          `subjects[0].name` in the provided RoleBinding manifest.

          ### Dynamic volume provisioning

          The launcher and worker template sections of the CoOperatingSystemRuntime resource
          require you to specify persistent volume claims (PVCs) for job recovery data. To ensure
          that these PVCs can be bound to approriate storage volumes, dynamic volume provisioning
          must be enabled in your cluster.

          Work with your cluster administrator to identify the storage needs of your jobs and to
          ensure that dynamic volume provisioning is enabled so that PVCs can be matched to
          persistent volumes of the right storage class.

          For more information about dynamic storage provisioning, see the
          [OpenShift documentation](https://docs.openshift.com/container-platform/latest/storage/dynamic-provisioning.html).

          ## Quick start and examples

          For an overview of the key concepts associated with running jobs in Kubernetes, see
          ["About running jobs in Kubernetes"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Configuring_and_running_Co_Operating_System_jobs_in_Kubernetes/About_running_jobs_in_Kubernetes)
          in the Ab Initio Help.

          As described in this topic, you might begin by creating a CoOperatingSystemRuntime
          manifest that defines the configuration and compute resources needed for a standalone test
          job. You could then use the specification for this job to create a CoOperatingSystemRuntimeTemplate
          resource, which would collect common settings and make it possible to concisely specify
          CoOperatingSystemRuntime jobs.

          You could also use a runtime pool (of kind CoOperatingSystemRuntime) to extend these basic
          resources: such a pool consists of an actively running group of templated runtime objects whose
          pods can be used to run jobs on demand. Finally, for reserving resources in a pool, you could
          create a CoOperatingSystemRuntimeClaim resource.

          For more information about these custom resources, refer to the
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Co_Operating_System_runtime_operator_API_reference).

          For a set of integrated examples (including complete YAML manifests) that show you how to
          create and use the custom resources, see ["Example manifests for running jobs"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Configuring_and_running_Co_Operating_System_jobs_in_Kubernetes/Creating_custom_resources_for_running_jobs/Example_manifests_for_running_jobs).
        displayName: Co>Operating System Runtime Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - abinitio
        - runtime
        - operator
        links:
        - name: Ab Initio
          url: https://www.abinitio.com/
        maintainers:
        - email: support@abinitio.com
          name: Ab Initio Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Ab Initio
        relatedImages:
        - registry.connect.redhat.com/abinitio/cooperating-system-runtime-operator@sha256:ee30210a153e8b3ba09080737216d881cc6a07e1a888e9684be0d17fd4ff791d
        version: 4.2.3-0
      entries:
      - name: abinitio-runtime-operator.v4.2.3-0
        version: 4.2.3-0
      - name: abinitio-runtime-operator.v4.2.2
        version: 4.2.2
      name: stable
    defaultChannel: release-4.3
    packageName: abinitio-runtime-operator
    provider:
      name: Ab Initio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microsoft
      provider-url: ""
    name: azure-service-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: azure-service-operator.v1.0.59040
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "APIMgmtAPI",
                "metadata": {
                  "name": "aso-apimgmtapi-sample"
                },
                "spec": {
                  "apiId": "apiId0",
                  "apiService": "apimservice-sample",
                  "location": "westus",
                  "properties": {
                    "apiVersionDescription": "API version description",
                    "description": "API description",
                    "displayName": "aso-apimgmtapi-sample",
                    "isCurrent": true,
                    "isOnline": true,
                    "path": "/api/test",
                    "protocols": [
                      "http"
                    ],
                    "subscriptionRequired": false
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ApimService",
                "metadata": {
                  "name": "apimservice-sample"
                },
                "spec": {
                  "appInsightsName": "appinsights-sample-wmm",
                  "appInsightsResourceGroup": "resourcegroup-azure-operators-wmm",
                  "location": "westus",
                  "publisherEmail": {
                    "Your email": ""
                  },
                  "publisherName": {
                    "Your name": ""
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "tier": "premium",
                  "vnetName": "virtualnetwork-sample",
                  "vnetResourceGroup": "resourcegroup-azure-operators",
                  "vnetSubnetName": "test1",
                  "vnetType": "none"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AppInsights",
                "metadata": {
                  "name": "appinsights-sample"
                },
                "spec": {
                  "applicationType": "other",
                  "kind": "other",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AppInsightsApiKey",
                "metadata": {
                  "name": "key-sample"
                },
                "spec": {
                  "appInsights": "appinsights-sample",
                  "authSDKControlChannel": true,
                  "readTelemetry": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "writeAnnotations": true
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureLoadBalancer",
                "metadata": {
                  "name": "lb-sample-01"
                },
                "spec": {
                  "backendAddressPoolName": "test",
                  "backendPort": 3000,
                  "frontendPortRangeEnd": 2000,
                  "frontendPortRangeStart": 1001,
                  "inboundNatPoolName": "test",
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "lb-pip-sample-01",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureNetworkInterface",
                "metadata": {
                  "name": "nic-sample-01"
                },
                "spec": {
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "nic-pip-sample-01",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnetName": "test1",
                  "vnetName": "vnet-sample-01"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureNetworkInterface",
                "metadata": {
                  "name": "nic-sample-02"
                },
                "spec": {
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "nic-pip-sample-02",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnetName": "test1",
                  "vnetName": "vnet-sample-02"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "lb-pip-sample-01"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "nic-pip-sample-01"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "nic-pip-sample-02"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLManagedUser",
                "metadata": {
                  "name": "azuresqlmanageduser-sample"
                },
                "spec": {
                  "dbName": "azuresqldatabase-sample",
                  "managedIdentityClientId": null,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "db_owner"
                  ],
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLUser",
                "metadata": {
                  "name": "sqluser-sample"
                },
                "spec": {
                  "dbName": "azuresqldatabase-sample1",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "db_owner"
                  ],
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLVNetRule",
                "metadata": {
                  "name": "azuresqlvnetrule-sample"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-vnet",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlAction",
                "metadata": {
                  "name": "azuresqlaction-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverName": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "azuresqldatabase-sample1"
                },
                "spec": {
                  "edition": 0,
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "valueA",
                    "tag2": "valueB"
                  },
                  "name": "azuresqldatabase-sample3"
                },
                "spec": {
                  "edition": 0,
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlFailoverGroup",
                "metadata": {
                  "name": "azuresqlfailovergroup-sample"
                },
                "spec": {
                  "databaselist": [
                    "azuresqldatabase-sample1",
                    "azuresqldatabase-sample2"
                  ],
                  "failovergraceperiod": 30,
                  "failoverpolicy": "Automatic",
                  "location": "eastus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "secondaryserver": "sqlserver-samplesec",
                  "secondaryserverresourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-samplepri"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlFirewallRule",
                "metadata": {
                  "name": "sqlf-allowazuresvcaccess"
                },
                "spec": {
                  "endipaddress": "0.0.0.0",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "startipaddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample-777"
                },
                "spec": {
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample777"
                },
                "spec": {
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVMScaleSet",
                "metadata": {
                  "name": "vmss-sample-01"
                },
                "spec": {
                  "adminUserName": "azureuser7",
                  "backendAddressPoolName": "test",
                  "capacity": 3,
                  "inboundNatPoolName": "test",
                  "loadBalancerName": "lb-sample-01",
                  "location": "SouthCentralUS",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": "{ssh public key}",
                  "subnetName": "test1",
                  "virtualNetworkName": "vnet-sample-01",
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachine",
                "metadata": {
                  "name": "hpfvm1"
                },
                "spec": {
                  "adminUserName": "azureuser",
                  "location": "SouthCentralUS",
                  "networkInterfaceName": "nic-sample-02",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": {
                    "ssh key": ""
                  },
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachine",
                "metadata": {
                  "name": "hpfvm20"
                },
                "spec": {
                  "adminUserName": "azureuser",
                  "location": "SouthCentralUS",
                  "networkInterfaceName": "nic-sample-01",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": "{ssh public key}",
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachineExtension",
                "metadata": {
                  "name": "vmext20a"
                },
                "spec": {
                  "autoUpgradeMinorVersion": true,
                  "forceUpdateTag": "Test2",
                  "location": "SouthCentralUS",
                  "protectedSettings": "{}",
                  "publisher": "Microsoft.Azure.Extensions",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "settings": "{\"commandToExecute\":\"echo 'hello, world'\"}",
                  "typeHandlerVersion": "2.1",
                  "typeName": "CustomScript",
                  "vmName": "vm20"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "BlobContainer",
                "metadata": {
                  "name": "blobcontainer-sample"
                },
                "spec": {
                  "accesslevel": "Container",
                  "accountname": "storageaccountsample777",
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ConsumerGroup",
                "metadata": {
                  "name": "consumergroup-sample-1"
                },
                "spec": {
                  "eventHub": "eventhub-sample-1",
                  "namespace": "eventhubnamespace-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "CosmosDB",
                "metadata": {
                  "name": "cosmosdb-sample-1"
                },
                "spec": {
                  "kind": "GlobalDocumentDB",
                  "location": "westus",
                  "properties": {
                    "databaseAccountOfferType": "Standard",
                    "isVirtualNetworkFilterEnabled": true
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "CosmosDB",
                "metadata": {
                  "labels": {
                    "CosmosAccountType": "Production"
                  },
                  "name": "cosmosdb-sample-2"
                },
                "spec": {
                  "ipRules": [
                    "104.42.195.92,40.76.54.131,52.176.6.30,52.169.50.45,52.187.184.26"
                  ],
                  "kind": "GlobalDocumentDB",
                  "location": "eastus",
                  "locations": [
                    {
                      "failoverPriority": 0,
                      "isZoneRedundant": true,
                      "locationName": "eastus"
                    },
                    {
                      "failoverPriority": 1,
                      "locationName": "westus"
                    }
                  ],
                  "properties": {
                    "databaseAccountOfferType": "Standard",
                    "enableMultipleWriteLocations": true
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "Eventhub",
                "metadata": {
                  "name": "eventhub-sample-1"
                },
                "spec": {
                  "authorizationRule": {
                    "name": "RootManageSharedAccessKey",
                    "rights": [
                      "Listen",
                      "Manage",
                      "Send"
                    ]
                  },
                  "location": "westus",
                  "namespace": "eventhubnamespace-sample-1",
                  "properties": {
                    "messageRetentionInDays": 7,
                    "partitionCount": 2
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "Eventhub",
                "metadata": {
                  "name": "eventhub-sample-2307-10-capture"
                },
                "spec": {
                  "authorizationRule": {
                    "name": "RootManageSharedAccessKey",
                    "rights": [
                      "Listen",
                      "Manage",
                      "Send"
                    ]
                  },
                  "location": "westus",
                  "namespace": "eventhubnamespace-sample-1",
                  "properties": {
                    "captureDescription": {
                      "destination": {
                        "archiveNameFormat": "{Namespace}/{EventHub}/{PartitionId}/{Year}/{Month}/{Day}/{Hour}/{Minute}/{Second}",
                        "blobContainer": "capturecontainer",
                        "name": "EventHubArchive.AzureBlockBlob",
                        "storageAccount": {
                          "accountName": "storageaccountauv1",
                          "resourceGroup": "my-resource-group"
                        }
                      },
                      "enabled": true,
                      "intervalInSeconds": 120,
                      "sizeLimitInBytes": 524288000
                    },
                    "messageRetentionInDays": 7,
                    "partitionCount": 2
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "EventhubNamespace",
                "metadata": {
                  "name": "eventhubnamespace-sample-1"
                },
                "spec": {
                  "location": "westus",
                  "properties": {
                    "isAutoInflateEnabled": false,
                    "kafkaEnabled": false,
                    "maximumThroughputUnits": 0
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "capacity": 1,
                    "name": "Standard",
                    "tier": "Standard"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVault",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "keyvaultsample123"
                },
                "spec": {
                  "accessPolicies": [
                    {
                      "objectID": "\u003cobjectID\u003e",
                      "permissions": {
                        "certificates": [
                          "list",
                          "get"
                        ],
                        "keys": [
                          "list",
                          "get"
                        ],
                        "secrets": [
                          "list",
                          "get"
                        ],
                        "storage": [
                          "list",
                          "get"
                        ]
                      },
                      "tenantID": "\u003ctenantID\u003e"
                    }
                  ],
                  "enableSoftDelete": false,
                  "location": "westus",
                  "networkPolicies": {
                    "bypass": "AzureServices",
                    "defaultAction": "Allow",
                    "ipRules": [
                      "172.16.0.0/24",
                      "172.16.1.0/24"
                    ],
                    "virtualNetworkRules": [
                      "/subscriptions/\u003csubid\u003e/resourceGroups/rg1/providers/Microsoft.Network/virtualNetworks/test-vnet/subnets/subnet1",
                      "/subscriptions/\u003csubid\u003e/resourceGroups/rg1/providers/Microsoft.Network/virtualNetworks/test-vnet/subnets/subnet2"
                    ]
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "name": "standard"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVault",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "keyvaultsample123-simple"
                },
                "spec": {
                  "enableSoftDelete": false,
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVaultKey",
                "metadata": {
                  "name": "keyvaultkey-sample"
                },
                "spec": {
                  "keySize": 4096,
                  "keyVault": "keyvaultsample123",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "type": "RSA"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLDatabase",
                "metadata": {
                  "name": "mysqldatabase-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLDatabase",
                "metadata": {
                  "name": "mysqldatabase-sample2"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLFirewallRule",
                "metadata": {
                  "name": "mysqlfirewallrule-sample"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLFirewallRule",
                "metadata": {
                  "name": "mysqlfirewallrule-sample2"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample2",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-replica"
                },
                "spec": {
                  "createMode": "Replica",
                  "location": "eastus2",
                  "replicaProperties": {
                    "sourceServerId": "/subscriptions/{SUBID}/resourceGroups/resourcegroup-azure-operators/providers/Microsoft.DBforMySQL/servers/mysqlserver-sample"
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-sample2"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLUser",
                "metadata": {
                  "name": "mysqluser-sample"
                },
                "spec": {
                  "dbName": "mysqldatabase-sample",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "SELECT"
                  ],
                  "server": "mysqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLVNetRule",
                "metadata": {
                  "name": "mysqlvnetrule-sample"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-vnet",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLDatabase",
                "metadata": {
                  "name": "postgresqldatabase-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLDatabase",
                "metadata": {
                  "name": "postgresqldatabase-sample2"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLFirewallRule",
                "metadata": {
                  "name": "postgresqlfirewallrule-sample"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLFirewallRule",
                "metadata": {
                  "name": "postgresqlfirewallrule-sample2"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample2",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-replica3"
                },
                "spec": {
                  "createMode": "Replica",
                  "location": "eastus",
                  "replicaProperties": {
                    "sourceServerId": "/subscriptions/{subscription ID}/resourceGroups/resourcegroup-azure-operators/providers/Microsoft.DBforPostgreSQL/servers/postgresqlserver-sample"
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-sample2"
                },
                "spec": {
                  "location": "southcentralus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 2,
                    "family": "Gen5",
                    "name": "B_Gen5_2",
                    "size": "51200",
                    "tier": "Basic"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLUser",
                "metadata": {
                  "name": "psqluser-sample"
                },
                "spec": {
                  "dbName": "postgresqldatabase-sample",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "azure_pg_admin"
                  ],
                  "server": "postgresqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLVNetRule",
                "metadata": {
                  "name": "postgresqlvnetrule-sample1"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-azure-operators",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCache",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "rediscache-sample-1"
                },
                "spec": {
                  "location": "westus",
                  "properties": {
                    "enableNonSslPort": true,
                    "sku": {
                      "capacity": 1,
                      "family": "C",
                      "name": "Basic"
                    }
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCacheAction",
                "metadata": {
                  "name": "rediscacheaction-sample-1"
                },
                "spec": {
                  "actionName": "rollallkeys",
                  "cacheName": "rediscache-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCacheFirewallRule",
                "metadata": {
                  "name": "rediscachefirewallrule"
                },
                "spec": {
                  "properties": {
                    "endIP": "0.0.0.0",
                    "startIP": "0.0.0.0"
                  },
                  "redisCache": "rediscache-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ResourceGroup",
                "metadata": {
                  "name": "resourcegroup-azure-operators"
                },
                "spec": {
                  "location": "westus"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "StorageAccount",
                "metadata": {
                  "name": "storageaccountsample777"
                },
                "spec": {
                  "accessTier": "Hot",
                  "kind": "StorageV2",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "name": "Standard_RAGRS"
                  },
                  "supportsHttpsTrafficOnly": true
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "VirtualNetwork",
                "metadata": {
                  "name": "vnet-sample-01"
                },
                "spec": {
                  "addressSpace": "10.0.0.0/8",
                  "location": "SouthCentralUS",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnets": [
                    {
                      "subnetAddressPrefix": "10.1.0.0/16",
                      "subnetName": "test1"
                    },
                    {
                      "subnetAddressPrefix": "10.2.0.0/16",
                      "subnetName": "test2"
                    }
                  ]
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "VirtualNetwork",
                "metadata": {
                  "name": "vnet-sample-02"
                },
                "spec": {
                  "addressSpace": "10.0.0.0/8",
                  "location": "SouthCentralUS",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnets": [
                    {
                      "subnetAddressPrefix": "10.1.0.0/16",
                      "subnetName": "test1"
                    },
                    {
                      "subnetAddressPrefix": "10.2.0.0/16",
                      "subnetName": "test2"
                    }
                  ]
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "BlobContainer",
                "metadata": {
                  "name": "blobcontainer-sample"
                },
                "spec": {
                  "accessLevel": "Container",
                  "accountName": "storageaccountsample777",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "MySQLServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "mysqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "postgresqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "azuresqldatabase-sample"
                },
                "spec": {
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "sku": {
                    "name": "Basic",
                    "tier": "Basic"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlFailoverGroup",
                "metadata": {
                  "name": "azuresqlfailovergroup-sample"
                },
                "spec": {
                  "databaseList": [
                    "azuresqldatabase-sample",
                    "azuresqldatabase-sample2"
                  ],
                  "failoverGracePeriod": 30,
                  "failoverPolicy": "automatic",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "secondaryServer": "sqlserver-samplesec",
                  "secondaryServerResourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-samplepri"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlFirewallRule",
                "metadata": {
                  "name": "sqlf-allowazuresvcaccess"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample-777"
                },
                "spec": {
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Cloud Provider
          certified: "false"
          containerImage: mcr.microsoft.com/k8s/azureserviceoperator:1.0.59040
          createdAt: Mon 12 Jun 2023 04:24:02 UTC
          description: The Azure Service Operator helps you provision Azure resources
            and connect your applications to them from within Kubernetes.
          displayName: Azure Service Operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Azure/azure-service-operator
          support: https://github.com/Azure/azure-service-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Creates an API with the specified properties in the specified
              API Management service.
            displayName: APIMgmtAPI
            kind: APIMgmtAPI
            name: apimgmtapis.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an API Management instance into a specified Resource
              Group at the specified location. It also provides the option to link
              to an Application Insights instance for logging, and to place the API
              Management instance in a specified Virtual Network.
            displayName: ApimService
            kind: ApimService
            name: apimservices.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Application Insights instance into a specified
              Resource Group at the specified location.
            displayName: AppInsights
            kind: AppInsights
            name: appinsights.azure.microsoft.com
            version: v1alpha1
          - description: Creates an Api Key to access a given Application Insights
              instance.
            displayName: AppInsightsApiKey
            kind: AppInsightsApiKey
            name: appinsightsapikeys.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Load Balancer (LB) into a specified Resource
              Group at the specified location.
            displayName: AzureLoadBalancer
            kind: AzureLoadBalancer
            name: azureloadbalancers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Network Interface (NIC) into a specified
              Resource Group at the specified location. Users can specify underlying
              public IP address and virtual network configurations in their NIC setup.
            displayName: AzureNetworkInterface
            kind: AzureNetworkInterface
            name: azurenetworkinterfaces.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Public IP Address (PIP) into a specified
              Resource Group at the specified location. Users can specify IP allocation
              method, idle timeout, IP address version, and SKU.
            displayName: AzurePublicIPAddress
            kind: AzurePublicIPAddress
            name: azurepublicipaddresses.azure.microsoft.com
            version: v1alpha1
          - description: Allows you to roll the password for the specified SQL server.
            displayName: AzureSqlAction
            kind: AzureSqlAction
            name: azuresqlactions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an SQL database given the SQL server.
            displayName: AzureSqlDatabase
            kind: AzureSqlDatabase
            name: azuresqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an SQL database given the SQL server
            displayName: AzureSqlDatabase
            kind: AzureSqlDatabase
            name: azuresqldatabases.azure.microsoft.com
            version: v1beta1
          - description: Deploys a failover group on a specified Azure SQL server
              given the secondary server and the databases to failover.
            displayName: AzureSqlFailoverGroup
            kind: AzureSqlFailoverGroup
            name: azuresqlfailovergroups.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a failover group on a specified Azure SQL server
              given the secondary server and the databases to failover.
            displayName: AzureSqlFailoverGroup
            kind: AzureSqlFailoverGroup
            name: azuresqlfailovergroups.azure.microsoft.com
            version: v1beta1
          - description: Deploys a firewall rule to allow access to the SQL server
              from specific IPs.
            displayName: AzureSqlFirewallRule
            kind: AzureSqlFirewallRule
            name: azuresqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the SQL server
              from specific IPs.
            displayName: AzureSqlFirewallRule
            kind: AzureSqlFirewallRule
            name: azuresqlfirewallrules.azure.microsoft.com
            version: v1beta1
          - description: Binds a specified Managed Identity as a user on the SQL database
            displayName: AzureSQLManagedUser
            kind: AzureSQLManagedUser
            name: azuresqlmanagedusers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure SQL server given the location and Resource
              Group.
            displayName: AzureSqlServer
            kind: AzureSqlServer
            name: azuresqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure SQL server given the location and Resource
              Group.
            displayName: AzureSqlServer
            kind: AzureSqlServer
            name: azuresqlservers.azure.microsoft.com
            version: v1beta1
          - description: Creates an user on the specified Azure SQL database and stores
              the username/password as secrets.
            displayName: AzureSQLUser
            kind: AzureSQLUser
            name: azuresqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a VirtualNetwork rule to allow access to the SQL
              server from specific VNets.
            displayName: AzureSQLVNetRule
            kind: AzureSQLVNetRule
            name: azuresqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys Azure Virtual Machine Extensions to provide post-deployment
              configuration and automation on Azure VMs.
            displayName: AzureVirtualMachineExtension
            kind: AzureVirtualMachineExtension
            name: azurevirtualmachineextensions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Machine (VM) into a specified Resource
              Group at the specified location. Users can specify platform image, size,
              user name and public SSH key, etc. for the VM.
            displayName: AzureVirtualMachine
            kind: AzureVirtualMachine
            name: azurevirtualmachines.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Machine Scale Set (VMSS) into a
              specified Resource Group at the specified location. Users can specify
              platform image, size, user name and public SSH key, etc. for the VMSS.
            displayName: AzureVMScaleSet
            kind: AzureVMScaleSet
            name: azurevmscalesets.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a Blob Container into a specified Resource Group
              at the specified location. A Storage Account must first be created prior
              to creating a Blob Container.
            displayName: BlobContainer
            kind: BlobContainer
            name: blobcontainers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a Blob Container into a specified Resource Group
              at the specified location. A Storage Account must first be created prior
              to creating a Blob Container.
            displayName: BlobContainer
            kind: BlobContainer
            name: blobcontainers.azure.microsoft.com
            version: v1alpha2
          - description: Deploys a consumer group given the Event Hub, Event Hubs
              namespace and Resource Group.
            displayName: ConsumerGroup
            kind: ConsumerGroup
            name: consumergroups.azure.microsoft.com
            version: v1alpha1
          - description: Provision a CosmosDB instance given the Cosmos DB instance
              type, location, and Resource Group.
            displayName: CosmosDB
            kind: CosmosDB
            name: cosmosdbs.azure.microsoft.com
            version: v1alpha1
          - kind: CosmosDBSQLDatabase
            name: cosmosdbsqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Event Hubs namespace given the Resource Group
              and location. Also has the ability to configure SKU, properties, and
              network rules.
            displayName: EventhubNamespace
            kind: EventhubNamespace
            name: eventhubnamespaces.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Event Hubs instance given the Event Hubs namespace,
              Resource Group and Location.
            displayName: Eventhub
            kind: Eventhub
            name: eventhubs.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Key Vault key given the location and Resource
              Group.
            displayName: KeyVaultKey
            kind: KeyVaultKey
            name: keyvaultkeys.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Key Vault given the location and Resource
              Group.
            displayName: KeyVault
            kind: KeyVault
            name: keyvaults.azure.microsoft.com
            version: v1alpha1
          - description: Add an AAD user to an existing MySQL database.
            kind: MySQLAADUser
            name: mysqlaadusers.azure.microsoft.com
            version: v1alpha1
          - description: Add an AAD user to an existing MySQL database.
            kind: MySQLAADUser
            name: mysqlaadusers.azure.microsoft.com
            version: v1alpha2
          - description: Deploys a database under the given Azure Database for MySQL
              server
            displayName: MySQLDatabase
            kind: MySQLDatabase
            name: mysqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the Azure Database
              for MySQL server from the specified IP range.
            displayName: MySQLFirewallRule
            kind: MySQLFirewallRule
            name: mysqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Create an AAD-based admininistrative user in a MySQL server.
            kind: MySQLServerAdministrator
            name: mysqlserveradministrators.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for MySQL server given the Location,
              Resource Group and other properties. Also helps creating read replicas
              for MySQL server.
            displayName: MySQLServer
            kind: MySQLServer
            name: mysqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for MySQL server given the Location,
              Resource Group and other properties. Also helps creating read replicas
              for MySQL server.
            displayName: MySQLServer
            kind: MySQLServer
            name: mysqlservers.azure.microsoft.com
            version: v1alpha2
          - description: Add a new user to an existing MySQL database.
            displayName: MySQLUser
            kind: MySQLUser
            name: mysqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Add a new user to an existing MySQL database.
            kind: MySQLUser
            name: mysqlusers.azure.microsoft.com
            version: v1alpha2
          - description: Add virtual network rules to the MySQL server.
            displayName: MySQLVNetRule
            kind: MySQLVNetRule
            name: mysqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a database under the given Azure Database for PostgreSQL
              server.
            displayName: PostgreSQLDatabase
            kind: PostgreSQLDatabase
            name: postgresqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the Azure Database
              for PostgreSQL server from the specified IP range.
            displayName: PostgreSQLFirewallRule
            kind: PostgreSQLFirewallRule
            name: postgresqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for PostgreSQL server given the
              Location and Resource Group.
            displayName: PostgreSQLServer
            kind: PostgreSQLServer
            name: postgresqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for PostgreSQL server given the
              Location and Resource Group.
            displayName: PostgreSQLServer
            kind: PostgreSQLServer
            name: postgresqlservers.azure.microsoft.com
            version: v1alpha2
          - description: Create and Manage Users for Azure Database for PostgreSQL
              server.
            displayName: PostgreSQLUser
            kind: PostgreSQLUser
            name: postgresqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Create and manage VNet service endpoints and VNet rules in
              Azure Database for PostgreSQL.
            displayName: PostgreSQLVNetRule
            kind: PostgreSQLVNetRule
            name: postgresqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Allows you to regenerate keys and reboot the RedisCache cluster.
            displayName: RedisCacheAction
            kind: RedisCacheAction
            name: rediscacheactions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the RedisCache
              from the specified IP range.
            displayName: RedisCacheFirewallRule
            kind: RedisCacheFirewallRule
            name: rediscachefirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Cache for Redis into a specified Resource
              Group at the specified location.
            displayName: RedisCache
            kind: RedisCache
            name: rediscaches.azure.microsoft.com
            version: v1alpha1
          - description: Provisions a Resource Group given the location and a name.
            displayName: ResourceGroup
            kind: ResourceGroup
            name: resourcegroups.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Storage Account into a specified Resource
              Group at the specified location.
            displayName: StorageAccount
            kind: StorageAccount
            name: storageaccounts.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Network into a specified Resource
              Group at the specified location, in the address space provided. Users
              are also able to add subnets to their virtual network through the operator.
            displayName: VirtualNetwork
            kind: VirtualNetwork
            name: virtualnetworks.azure.microsoft.com
            version: v1alpha1
        description: |
          ## Overview

          The Azure Service Operator comprises of:

          - The Custom Resource Definitions (CRDs) for each of the Azure services a Kubernetes user can provision.
          - The Kubernetes controller that watches for requests to create Custom Resources for each of these CRDs and creates them.

          The project was built using [Kubebuilder](https://book.kubebuilder.io/).

          Curious to see how it all works? Check out our [control flow diagrams](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/controlflow.md).

          **Note**: The current version of Azure Service Operator only supports watching all namespaces in the cluster.

          ## Supported Azure Services

          - [Resource Group](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/resourcegroup/resourcegroup.md)
          - [Event Hubs](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/eventhub/eventhub.md)
          - [Azure SQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/azuresql/azuresql.md)
          - [Azure Database for PostgreSQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/postgresql/postgresql.md)
          - [Azure Database for MySQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/mysql/mysql.md)
          - [Azure Key Vault](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/keyvault/keyvault.md)
          - [Azure Cache for Redis](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/rediscache/rediscache.md)
          - [Storage Account](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/storage/storageaccount.md)
          - [Blob Storage](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/storage/blobcontainer.md)
          - [Virtual Network](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/virtualnetwork/virtualnetwork.md)
          - [Application Insights](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/appinsights/appinsights.md)
          - [API Management](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/apimgmt/apimgmt.md)
          - [Cosmos DB](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/cosmosdb/cosmosdb.md)
          - [Virtual Machine](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/virtualmachine/virtualmachine.md)
          - [Virtual Machine Scale Set](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/vmscaleset/vmscaleset.md)

          ## Instructions - Please read before installing

          Before you begin, verify you're running the Azure CLI version 2.0.53 or later. Run `az --version` to find the version. If you need to install or upgrade, see [Install Azure CLI](https://docs.microsoft.com/en-us/cli/azure/install-azure-cli).

          1. Create an Azure Service Principal. You'll need this to grant Azure Service Operator permissions to create resources in your subscription.

              First, set the following environment variables to your Azure Tenant ID and Subscription ID with your values:
              ```yaml
              AZURE_TENANT_ID=<your-tenant-id-goes-here>
              AZURE_SUBSCRIPTION_ID=<your-subscription-id-goes-here>
              ```

              You can find these values by using the [Azure CLI](https://docs.microsoft.com/en-us/cli/azure/?view=azure-cli-latest):
              ```sh
              az account show
              ```

          2. Next, we'll create a Service Principal with Contributor permissions for your subscription, so ASO can create resources in your subscription on your behalf. Note that the [ServicePrincipal](https://docs.microsoft.com/en-us/cli/azure/create-an-azure-service-principal-azure-cli) you pass to the command below needs to have access to create resources in your subscription. If you'd like to use Managed Identity for authorization instead, check out instructions [here](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/managedidentity.md).

              ```sh
              az ad sp create-for-rbac -n "azure-service-operator" --role contributor \
                  --scopes /subscriptions/$AZURE_SUBSCRIPTION_ID
              ```

              This should give you output like the following:
              ```sh
              "appId": "xxxxxxxxxx",
              "displayName": "azure-service-operator",
              "name": "http://azure-service-operator",
              "password": "xxxxxxxxxxx",
              "tenant": "xxxxxxxxxxxxx"
              ```

          3.  Once you have created a Service Principal, gather the following values:

              `AZURE_TENANT_ID` is the Tenant ID from Step 1.

              `AZURE_SUBSCRIPTION_ID` is the Subscription ID from Step 1.

              `AZURE_CLIENT_ID` is the appID from the Service Principal created in Step 2.

              `AZURE_CLIENT_SECRET` is the password from the service Principal we created in Step 2.

              `AZURE_CLOUD_ENV` is the Azure Environment you'd like to use, i.e. AzurePublicCloud, AzureUSGovernmentCloud, AzureChinaCloud, AzureGermanCloud.

          4.  Set those values in a `Secret` called `azureoperatorsettings` within the `operators` Namespace:
              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name: azureoperatorsettings
                namespace: operators
              stringData:
                AZURE_TENANT_ID: <your-tenant-id-goes-here>
                AZURE_SUBSCRIPTION_ID: <your-subscription-id-goes-here>
                AZURE_CLIENT_ID: <your-client-id>
                AZURE_CLIENT_SECRET: <your-client-secret>
                AZURE_CLOUD_ENV: <your-azure-cloud-environment>
              ```

          5. Now you can proceed to install the Azure Service Operator to the `operators` Namespace via the Install button on the top right of this page. After the operator is installed, you will then see the Azure Service Operator pod running in your cluster:

              ```console
              $ kubectl get pods -n operators
              NAME                                                READY   STATUS    RESTARTS   AGE
              azureoperator-controller-manager-7dd75bbd97-mk4s9   2/2     Running   0          35s
              ```

          ## About the project

          This project maintains [releases of the Azure Service Operator](https://github.com/Azure/azure-service-operator/releases).

          Please see the [FAQ](https://github.com/Azure/azure-service-operator/blob/main/docs/faq.md) for answers to commonly asked questions about the Azure Service Operator.

          Have more questions? Feel free to consult our documentation [here](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/contents.md).

          ## Contributing

          The [contribution guide](https://github.com/Azure/azure-service-operator/blob/main/CONTRIBUTING.md) covers everything you need to know about how you can contribute to Azure Service Operators. The [developer guide](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/contents.md#developing-azure-service-operator) will help you onboard as a developer.

          ## Support

          Azure Service Operator is an open source project that is **not** covered by the [Microsoft Azure support policy](https://support.microsoft.com/en-us/help/2941892/support-for-linux-and-open-source-technology-in-azure). Please search open issues [here](https://github.com/Azure/azure-service-operator/issues). If your issue isn't already represented, please [open a new one](https://github.com/Azure/azure-service-operator/issues/new/choose). The Azure Service Operator project maintainers will respond to the best of their abilities.

          ## Code of conduct

          This project has adopted the [Microsoft Open Source Code of Conduct](https://opensource.microsoft.com/codeofconduct/). For more information, see the [Code of Conduct FAQ](https://opensource.microsoft.com/codeofconduct/faq) or contact [opencode@microsoft.com](mailto:opencode@microsoft.com) with any additional questions or comments.
        displayName: Azure Service Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - azure, microsoft, cloud service broker
        links:
        - name: Azure Service Operator
          url: https://github.com/Azure/azure-service-operator
        maintainers:
        - email: christian.muirhead@microsoft.com
          name: Christian Muirhead
        maturity: stable
        provider:
          name: Microsoft
        relatedImages:
        - mcr.microsoft.com/k8s/azureserviceoperator:1.0.59040
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 1.0.59040
      entries:
      - name: azure-service-operator.v1.0.59040
        version: 1.0.59040
      name: stable
    defaultChannel: stable
    packageName: azure-service-operator
    provider:
      name: Microsoft
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: service-registry-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-registry-operator.v2.0.7-0.1662549961.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-kafkasql"
                },
                "spec": {
                  "configuration": {
                    "kafkasql": {
                      "bootstrapServers": "\u003cservice name\u003e.\u003cnamespace\u003e.svc:9092"
                    },
                    "persistence": "kafkasql"
                  }
                }
              },
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-sql"
                },
                "spec": {
                  "configuration": {
                    "persistence": "sql",
                    "sql": {
                      "dataSource": {
                        "password": "\u003cpassword\u003e",
                        "url": "jdbc:postgresql://\u003cservice name\u003e.\u003cnamespace\u003e.svc:5432/\u003cdatabase name\u003e",
                        "userName": "postgres"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/integration/service-registry-rhel8-operator@sha256:6dfd539951a739ec06fb6620b79c42c6159d4ee148c0d9d336547e786cd4780c
          createdAt: "2022-06-24"
          description: Deploy and manage Red Hat Integration service registry on OpenShift.
          olm.substitutesFor: service-registry-operator.v2.0.7-0.1661784420.p
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents a Red Hat Integration Service
              Registry instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Service Registry
          Service Registry stores and retrieves API designs and event schemas, and gives you control of their evolution.
          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf - Provides a REST API and web UI to manage the artifacts and artifact meta-data - Includes Serializers and Deserializers for Kafka client integration - Configurable rules to control schema validity and evolution (compatibility) - Storage options: Kafka, PostgreSQL - Compatible with Confluent and IBM APIs - Runs on a lightweight Quarkus platform - Includes Maven plugin to integrate with Maven based builds
          ## Service Registry Operator
          Provides a quick and easy way to deploy and manage service registry on OpenShift.
          **Features**
          - Supports installation and configuration of the Registry for all storage options - Easily perform a rolling upgrade of the Registry
          ## Prerequisites
          This operator does not deploy storage for the Registry. Therefore, some storage options require that the chosen persistence service is already set up.
          ## License
          Red Hat Service Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - Service Registry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - service
        - service-registry
        links:
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/integration/service-registry-rhel8-operator@sha256:6dfd539951a739ec06fb6620b79c42c6159d4ee148c0d9d336547e786cd4780c
        - registry.redhat.io/integration/service-registry-kafkasql-rhel8@sha256:fc97a2a08b2f77d620ea86ec77bb4f750bfc0109ede3ad1080c2cb694019f8b8
        - registry.redhat.io/integration/service-registry-sql-rhel8@sha256:6c97fa8ead1f6f0323811f8f35a80ae33f0e160f6c58f2bdb31713dbeaa3f014
        version: 2.0.7+0.1662549961.p
      entries:
      - name: service-registry-operator.v2.0.7-0.1662549961.p
        version: 2.0.7+0.1662549961.p
      - name: service-registry-operator.v2.0.7-0.1661784420.p
        version: 2.0.7+0.1661784420.p
      - name: service-registry-operator.v2.0.7
        version: 2.0.7
      - name: service-registry-operator.v2.0.6
        version: 2.0.6
      - name: service-registry-operator.v2.0.5-0.1654503709.p
        version: 2.0.5+0.1654503709.p
      - name: service-registry-operator.v2.0.5-0.1652887585.p
        version: 2.0.5+0.1652887585.p
      - name: service-registry-operator.v2.0.5-0.1646309885.p
        version: 2.0.5+0.1646309885.p
      - name: service-registry-operator.v2.0.5-0.1647938417.p
        version: 2.0.5+0.1647938417.p
      - name: service-registry-operator.v2.0.5-0.1648813064.p
        version: 2.0.5+0.1648813064.p
      - name: service-registry-operator.v2.0.5-0.1651047781.p
        version: 2.0.5+0.1651047781.p
      - name: service-registry-operator.v2.0.5-0.1652429755.p
        version: 2.0.5+0.1652429755.p
      - name: service-registry-operator.v2.0.5
        version: 2.0.5
      - name: service-registry-operator.v2.0.5-0.1655396591.p
        version: 2.0.5+0.1655396591.p
      - name: service-registry-operator.v2.0.4
        version: 2.0.4
      - name: service-registry-operator.v2.0.3
        version: 2.0.3
      - name: service-registry-operator.v2.0.2
        version: 2.0.2
      - name: service-registry-operator.v2.0.1
        version: 2.0.1
      - name: service-registry-operator.v2.0.0
        version: 2.0.0
      name: 2.0.x
    - currentCSV: service-registry-operator.v2.6.13-r2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-kafkasql"
                },
                "spec": {
                  "configuration": {
                    "kafkasql": {
                      "bootstrapServers": "\u003cservice name\u003e.\u003cnamespace\u003e.svc:9092"
                    },
                    "persistence": "kafkasql"
                  }
                }
              },
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-sql"
                },
                "spec": {
                  "configuration": {
                    "persistence": "sql",
                    "sql": {
                      "dataSource": {
                        "password": "\u003cpassword\u003e",
                        "url": "jdbc:postgresql://\u003cservice name\u003e.\u003cnamespace\u003e.svc:5432/\u003cdatabase name\u003e",
                        "userName": "postgres"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/apicurio/apicurio-registry-rhel8-operator@sha256:61f00395b51819312dda7abcdaa47190a11a83b1e2b523101be0ba776b1c5b7c
          createdAt: "2025-09-15"
          description: Deploy and manage Red Hat build of Apicurio Registry on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations",
            "Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents a Red Hat build of Apicurio Registry
              instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Red Hat build of Apicurio Registry

          Red Hat build of Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Provides a REST API and web UI to manage the artifacts and artifact meta-data
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL
          - Compatible with Confluent and IBM APIs
          - Runs on a lightweight Quarkus platform
          - Includes Maven plugin to integrate with Maven based builds

          ## Red Hat build of Apicurio Registry Operator

          Provides a quick and easy way to deploy and manage Red Hat build of Apicurio Registry on OpenShift.

          **Features**
          - Supports installation and configuration of the Registry for all storage options
          - Easily perform a rolling upgrade of Red Hat build of Apicurio Registry

          ## Prerequisites

          This operator does not deploy storage for Red Hat build of Apicurio Registry. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Red Hat build of Apicurio Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/main/LICENSE)
        displayName: Red Hat build of Apicurio Registry
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - service
        - service-registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/apicurio/apicurio-registry-rhel8-operator@sha256:61f00395b51819312dda7abcdaa47190a11a83b1e2b523101be0ba776b1c5b7c
        - registry.redhat.io/apicurio/apicurio-registry-kafkasql-rhel8@sha256:f2da7cd92e126187d91043e2ae40521f8574586f8b3975be27d8370fe31b6732
        - registry.redhat.io/apicurio/apicurio-registry-sql-rhel8@sha256:a55425a9af1a16959ab5e528a5740d47ea06f6935593cc925e7727cded4e7d13
        version: 2.6.13-r2
      entries:
      - name: service-registry-operator.v2.6.13-r2
        version: 2.6.13-r2
      - name: service-registry-operator.v2.6.13-r1
        version: 2.6.13-r1
      - name: service-registry-operator.v2.6.12-r1
        version: 2.6.12-r1
      - name: service-registry-operator.v2.6.11-r1
        version: 2.6.11-r1
      - name: service-registry-operator.v2.6.10
        version: 2.6.10
      - name: service-registry-operator.v2.6.9
        version: 2.6.9
      - name: service-registry-operator.v2.6.8-0.1740477005.p
        version: 2.6.8+0.1740477005.p
      - name: service-registry-operator.v2.6.8-0.1739457262.p
        version: 2.6.8+0.1739457262.p
      - name: service-registry-operator.v2.6.8
        version: 2.6.8
      - name: service-registry-operator.v2.6.6-0.1736942038.p
        version: 2.6.6+0.1736942038.p
      - name: service-registry-operator.v2.6.6
        version: 2.6.6
      - name: service-registry-operator.v2.6.5-0.1731946870.p
        version: 2.6.5+0.1731946870.p
      - name: service-registry-operator.v2.6.5
        version: 2.6.5
      - name: service-registry-operator.v2.6.5-0.1729511936.p
        version: 2.6.5+0.1729511936.p
      - name: service-registry-operator.v2.6.5-0.1730982111.p
        version: 2.6.5+0.1730982111.p
      - name: service-registry-operator.v2.6.3
        version: 2.6.3
      - name: service-registry-operator.v2.6.1-0.1724927626.p
        version: 2.6.1+0.1724927626.p
      - name: service-registry-operator.v2.6.1
        version: 2.6.1
      - name: service-registry-operator.v2.3.5
        version: 2.3.5
      - name: service-registry-operator.v2.3.4-0.1716987426.p
        version: 2.3.4+0.1716987426.p
      - name: service-registry-operator.v2.3.4
        version: 2.3.4
      - name: service-registry-operator.v2.3.3
        version: 2.3.3
      - name: service-registry-operator.v2.3.2
        version: 2.3.2
      - name: service-registry-operator.v2.3.1-0.1706181988.p
        version: 2.3.1+0.1706181988.p
      - name: service-registry-operator.v2.3.1
        version: 2.3.1
      - name: service-registry-operator.v2.3.0
        version: 2.3.0
      - name: service-registry-operator.v2.2.3
        version: 2.2.3
      - name: service-registry-operator.v2.2.3-0.1698369839.p
        version: 2.2.3+0.1698369839.p
      - name: service-registry-operator.v2.2.2
        version: 2.2.2
      - name: service-registry-operator.v2.2.2-0.1697116886.p
        version: 2.2.2+0.1697116886.p
      - name: service-registry-operator.v2.2.1
        version: 2.2.1
      - name: service-registry-operator.v2.2.0
        version: 2.2.0
      - name: service-registry-operator.v2.1.5
        version: 2.1.5
      - name: service-registry-operator.v2.1.4
        version: 2.1.4
      - name: service-registry-operator.v2.1.3
        version: 2.1.3
      - name: service-registry-operator.v2.1.2-0.1676389596.p
        version: 2.1.2+0.1676389596.p
      - name: service-registry-operator.v2.1.2
        version: 2.1.2
      - name: service-registry-operator.v2.1.1
        version: 2.1.1
      - name: service-registry-operator.v2.1.0
        version: 2.1.0
      - name: service-registry-operator.v2.0.7
        version: 2.0.7
      - name: service-registry-operator.v2.0.7-0.1662549961.p
        version: 2.0.7+0.1662549961.p
      - name: service-registry-operator.v2.0.7-0.1661784420.p
        version: 2.0.7+0.1661784420.p
      - name: service-registry-operator.v2.0.6
        version: 2.0.6
      - name: service-registry-operator.v2.0.5-0.1648813064.p
        version: 2.0.5+0.1648813064.p
      - name: service-registry-operator.v2.0.5-0.1646309885.p
        version: 2.0.5+0.1646309885.p
      - name: service-registry-operator.v2.0.5-0.1654503709.p
        version: 2.0.5+0.1654503709.p
      - name: service-registry-operator.v2.0.5-0.1652887585.p
        version: 2.0.5+0.1652887585.p
      - name: service-registry-operator.v2.0.5-0.1652429755.p
        version: 2.0.5+0.1652429755.p
      - name: service-registry-operator.v2.0.5-0.1651047781.p
        version: 2.0.5+0.1651047781.p
      - name: service-registry-operator.v2.0.5-0.1655396591.p
        version: 2.0.5+0.1655396591.p
      - name: service-registry-operator.v2.0.5-0.1647938417.p
        version: 2.0.5+0.1647938417.p
      - name: service-registry-operator.v2.0.5
        version: 2.0.5
      - name: service-registry-operator.v2.0.4
        version: 2.0.4
      - name: service-registry-operator.v2.0.3
        version: 2.0.3
      - name: service-registry-operator.v2.0.2
        version: 2.0.2
      - name: service-registry-operator.v2.0.1
        version: 2.0.1
      - name: service-registry-operator.v2.0.0
        version: 2.0.0
      name: 2.x
    defaultChannel: 2.x
    packageName: service-registry-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-keyspaces-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-keyspaces-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keyspaces.services.k8s.aws/v1alpha1",
                "kind": "Keyspace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "keyspaces.services.k8s.aws/v1alpha1",
                "kind": "Table",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/keyspaces-controller:1.1.1
          createdAt: "2025-09-20T01:23:26Z"
          description: AWS Keyspaces controller is a service controller for managing
            Keyspaces resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Keyspace represents the state of an AWS keyspaces Keyspace
              resource.
            displayName: Keyspace
            kind: Keyspace
            name: keyspaces.keyspaces.services.k8s.aws
            version: v1alpha1
          - description: Table represents the state of an AWS keyspaces Table resource.
            displayName: Table
            kind: Table
            name: tables.keyspaces.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Keyspaces resources in AWS from within your Kubernetes cluster.

          **About Amazon Keyspaces**

          Amazon Keyspaces (for Apache Cassandra) is a scalable, highly available, and managed Apache Cassandra–compatible database service. With Amazon Keyspaces, you don’t have to provision, patch, or manage servers, and you don’t have to install, maintain, or operate software.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Keyspaces
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keyspaces
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Keyspaces Developer Resources
          url: https://aws.amazon.com/keyspaces/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: keyspaces maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/keyspaces-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-keyspaces-controller.v1.1.1
        version: 1.1.1
      - name: ack-keyspaces-controller.v1.1.0
        version: 1.1.0
      - name: ack-keyspaces-controller.v1.0.14
        version: 1.0.14
      - name: ack-keyspaces-controller.v1.0.12
        version: 1.0.12
      - name: ack-keyspaces-controller.v1.0.11
        version: 1.0.11
      - name: ack-keyspaces-controller.v1.0.10
        version: 1.0.10
      - name: ack-keyspaces-controller.v1.0.9
        version: 1.0.9
      - name: ack-keyspaces-controller.v1.0.7
        version: 1.0.7
      - name: ack-keyspaces-controller.v1.0.6
        version: 1.0.6
      - name: ack-keyspaces-controller.v1.0.5
        version: 1.0.5
      - name: ack-keyspaces-controller.v1.0.3
        version: 1.0.3
      - name: ack-keyspaces-controller.v1.0.2
        version: 1.0.2
      - name: ack-keyspaces-controller.v1.0.1
        version: 1.0.1
      - name: ack-keyspaces-controller.v1.0.0
        version: 1.0.0
      - name: ack-keyspaces-controller.v0.0.16
        version: 0.0.16
      - name: ack-keyspaces-controller.v0.0.15
        version: 0.0.15
      - name: ack-keyspaces-controller.v0.0.14
        version: 0.0.14
      - name: ack-keyspaces-controller.v0.0.13
        version: 0.0.13
      - name: ack-keyspaces-controller.v0.0.12
        version: 0.0.12
      - name: ack-keyspaces-controller.v0.0.11
        version: 0.0.11
      - name: ack-keyspaces-controller.v0.0.10
        version: 0.0.10
      - name: ack-keyspaces-controller.v0.0.9
        version: 0.0.9
      - name: ack-keyspaces-controller.v0.0.8
        version: 0.0.8
      - name: ack-keyspaces-controller.v0.0.7
        version: 0.0.7
      - name: ack-keyspaces-controller.v0.0.6
        version: 0.0.6
      - name: ack-keyspaces-controller.v0.0.5
        version: 0.0.5
      - name: ack-keyspaces-controller.v0.0.4
        version: 0.0.4
      - name: ack-keyspaces-controller.v0.0.3
        version: 0.0.3
      - name: ack-keyspaces-controller.v0.0.2
        version: 0.0.2
      - name: ack-keyspaces-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-keyspaces-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/kuadrant-operator
    name: kuadrant-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kuadrant-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "authScheme": {
                    "identity": [
                      {
                        "apiKey": {
                          "allNamespaces": true,
                          "selector": {
                            "matchLabels": {
                              "app": "toystore"
                            }
                          }
                        },
                        "credentials": {
                          "in": "authorization_header",
                          "keySelector": "APIKEY"
                        },
                        "name": "friends"
                      }
                    ],
                    "response": [
                      {
                        "json": {
                          "properties": [
                            {
                              "name": "userID",
                              "valueFrom": {
                                "authJSON": "auth.identity.metadata.annotations.secret\\.kuadrant\\.io/user-id"
                              }
                            }
                          ]
                        },
                        "name": "rate-limit-apikey",
                        "wrapper": "envoyDynamicMetadata",
                        "wrapperKey": "ext_auth_data"
                      }
                    ]
                  },
                  "rules": null,
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "rateLimits": [
                    {
                      "configurations": [
                        {
                          "actions": [
                            {
                              "metadata": {
                                "default_value": "no-user",
                                "descriptor_key": "userID",
                                "metadata_key": {
                                  "key": "envoy.filters.http.ext_authz",
                                  "path": [
                                    {
                                      "segment": {
                                        "key": "ext_auth_data"
                                      }
                                    },
                                    {
                                      "segment": {
                                        "key": "userID"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          ]
                        }
                      ],
                      "limits": [
                        {
                          "conditions": [],
                          "maxValue": 2,
                          "seconds": 10,
                          "variables": [
                            "userID"
                          ]
                        }
                      ]
                    }
                  ],
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/kuadrant-operator:v0.3.1
          createdAt: "2023-06-01T12:24:17Z"
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Enable AuthN and AuthZ based access control on workloads
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1beta1
          - description: Kuadrant is the Schema for the kuadrants API
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: Enable access control on workloads based on HTTP rate limiting
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1beta1
        description: A Kubernetes Operator to manage the lifecycle of the Kuadrant
          system
        displayName: Kuadrant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: Kuadrant Operator
          url: https://github.com/Kuadrant/kuadrant-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - quay.io/kuadrant/kuadrant-operator:v0.3.1
        - oci://quay.io/kuadrant/wasm-shim:v0.2.0
        version: 0.3.1
      entries:
      - name: kuadrant-operator.v0.3.1
        version: 0.3.1
      - name: kuadrant-operator.v0.3.0
        version: 0.3.0
      - name: kuadrant-operator.v0.2.1
        version: 0.2.1
      - name: kuadrant-operator.v0.2.0
        version: 0.2.0
      name: alpha
    - currentCSV: kuadrant-operator.v0.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSPolicy",
                "metadata": {
                  "name": "dnspolicy-sample"
                },
                "spec": {
                  "healthCheck": {
                    "endpoint": "/",
                    "protocol": "HTTP"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "TLSPolicy",
                "metadata": {
                  "name": "tlspolicy-sample"
                },
                "spec": {
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "self-signed-ca"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "kuadrant.io/v1beta2",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "rules": {
                    "authentication": {
                      "apikey": {
                        "apiKey": {
                          "selector": {}
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta2",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "limits": {
                    "toys": {
                      "rates": [
                        {
                          "duration": 1,
                          "limit": 50,
                          "unit": "minute"
                        }
                      ]
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/kuadrant-operator:v0.11.0
          createdAt: "2024-10-02T20:30:45Z"
          description: '[DEPRECATED] Follow https://docs.kuadrant.io instead'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthPolicy enables authentication and authorization for service
              workloads in a Gateway API network
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1beta2
          - description: DNSPolicy configures how North-South based traffic should
              be balanced and reach the gateways
            displayName: DNSPolicy
            kind: DNSPolicy
            name: dnspolicies.kuadrant.io
            version: v1alpha1
          - description: Kuadrant configures installations of Kuadrant Service Protection
              components
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: RateLimitPolicy enables rate limiting for service workloads
              in a Gateway API network
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1beta2
          - description: TLSPolicy provides tls for gateway listeners by managing
              the lifecycle of tls certificates
            displayName: TLSPolicy
            kind: TLSPolicy
            name: tlspolicies.kuadrant.io
            version: v1alpha1
        description: |-
          ## Deprecated
          This Kuadrant community operator is deprecated, Follow https://docs.kuadrant.io instead.
        displayName: '[DEPRECATED] Kuadrant Operator'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: Kuadrant Operator
          url: https://github.com/Kuadrant/kuadrant-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - quay.io/kuadrant/kuadrant-operator:v0.11.0
        - oci://quay.io/kuadrant/wasm-shim:v0.6.0
        version: 0.11.1
      entries:
      - name: kuadrant-operator.v0.11.1
        version: 0.11.1
      - name: kuadrant-operator.v0.11.0
        version: 0.11.0
      - name: kuadrant-operator.v0.10.0
        version: 0.10.0
      - name: kuadrant-operator.v0.7.2
        version: 0.7.2
      - name: kuadrant-operator.v0.4.1
        version: 0.4.1
      - name: kuadrant-operator.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: kuadrant-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/kuadrant-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Nutanix
      provider-url: ""
    name: nutanixcsioperator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nutanixcsioperator.v2.6.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crd.nutanix.com/v1alpha1",
                "kind": "NutanixCsiStorage",
                "metadata": {
                  "name": "nutanixcsistorage"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:859b6d1b2b36a48391e347c2a7431cafa08decafb56f801e21e544c3460a4b4b
          createdAt: "2024-10-30T17:10:05Z"
          description: An Operator for deploying and managing the Nutanix CSI Driver
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2.6.11
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi","disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/nutanix/csi-operator
          support: Nutanix Support Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a new instance of Nutanix CSI Driver
            displayName: Nutanix CSI Storage
            kind: NutanixCsiStorage
            name: nutanixcsistorages.crd.nutanix.com
            version: v1alpha1
        description: "# Overview \n\nThe Nutanix CSI Operator for Kubernetes packages,
          deploys, manages, upgrades the Nutanix CSI Driver on Kubernetes and OpenShift
          for dynamic provisioning of persistent volumes on Nutanix Enterprise Cloud
          platform.\n\nThe Nutanix Container Storage Interface (CSI) Driver for Kubernetes
          leverages Nutanix Volumes and Nutanix Files to provide scalable and persistent
          storage for stateful applications.\n\nWith the Nutanix CSI Provider you
          can:\n\n*  **Provide persistent storage to your containers**\n    * Leverage
          PVC resources to consume dynamically Nutanix storage\n    * With Files storage
          classes, applications on multiple pods can access the same storage, and
          also have the benefit of multi-pod read and write access. \n\n## Warning\n\nIf
          you upgrade from a previous version take care to validate your existing
          StorageClass parameters against the corresponding [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v2_6:CSI-Volume-Driver-v2_6).\n\n##
          Supported Access mode\n\n| Storage Class mode | ReadWriteOnce | ReadOnlyMany
          | ReadWriteMany |\n|--------------------|---------------|--------------|---------------|\n|
          Nutanix Volumes    | yes           | yes          | no            |\n| Nutanix
          Files      | yes           | yes          | yes           |\n\n\n## Supported
          orchestration platforms\nThe following table details orchestration platforms
          suitable for deployment of the Nutanix CSI driver.\n\n| Orchestration platform
          | Version | Architecture |\n|------------------------|---------|--------------|\n|
          Red Hat OpenShift      | 4.12    | x86          |\n| Red Hat OpenShift      |
          4.13    | x86          |\n| Red Hat OpenShift      | 4.14    | x86          |\n|
          Red Hat OpenShift      | 4.15    | x86          |\n| Red Hat OpenShift      |
          4.16    | x86          |\n| Red Hat OpenShift      | 4.17    | x86          |\n\n##
          Configuring k8s secret and storage class\nIn order to use this driver, deploy
          your NutanixCsiStorage instance and create the relevant storage classes
          and secrets, by following the [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v2_6:CSI-Volume-Driver-v2_6)
          or the example in below section:\n\n**1. Create a NutanixCsiStorage resource
          to deploy your driver**\n\nYou can do it directly inside the Operator UI
          with the `Create instance` button or with the following ressource\n```yaml\napiVersion:
          crd.nutanix.com/v1alpha1\nkind: NutanixCsiStorage\nmetadata:\n  name: nutanixcsistorage\n
          \ namespace: openshift-cluster-csi-drivers\nspec: {}\n```\n\n**2. Create
          a Nutanix secret**\n```yaml\napiVersion: v1\nkind: Secret\nmetadata:\n  name:
          ntnx-secret\n  namespace: openshift-cluster-csi-drivers\nstringData:\n  #
          prism-ip:prism-port:username:password.\n  key: 1.2.3.4:9440:admin:password\n```\n\n**3.
          Create storage classes**\n\n- volume mode example\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n  name: nutanix-volume\nprovisioner: csi.nutanix.com\nparameters:\n
          \ csi.storage.k8s.io/provisioner-secret-name: ntnx-secret\n  csi.storage.k8s.io/provisioner-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/node-publish-secret-name:
          ntnx-secret\n  csi.storage.k8s.io/node-publish-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/controller-expand-secret-name: ntnx-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/fstype: ext4\n  storageContainer:
          default-container\n  storageType: NutanixVolumes\n  #description: \"description
          added to each storage object created by the driver\"\n  #isSegmentedIscsiNetwork:
          \"false\"\n  #whitelistIPMode: ENABLED\n  #chapAuth: ENABLED\n  #isLVMVolume:
          \"false\"\n  #numLVMDisks: 4\nallowVolumeExpansion: true\nreclaimPolicy:
          Delete\n```\n\n- dynamic files mode example\n\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n    name: nutanix-dynfiles\nprovisioner: csi.nutanix.com\nparameters:\n
          \ dynamicProv: ENABLED\n  nfsServerName: fs\n  csi.storage.k8s.io/provisioner-secret-name:
          ntnx-secret\n  csi.storage.k8s.io/provisioner-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/node-publish-secret-name: ntnx-secret\n  csi.storage.k8s.io/node-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-expand-secret-name:
          ntnx-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace: openshift-cluster-csi-drivers\n
          \ storageType: NutanixFiles\n  squashType: \"none\"\n  #description: \"description
          added to each storage object created by the driver\"\nallowVolumeExpansion:
          true\n```"
        displayName: Nutanix CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Nutanix
        - CSI
        - Volumes
        - Files
        - Storage
        links:
        - name: Documentation
          url: https://portal.nutanix.com/page/documents/list?type=software&filterKey=product&filterVal=CSI
        maintainers:
        - email: cloudnative@nutanix.com
          name: Nutanix Cloud Native
        maturity: beta
        provider:
          name: Nutanix
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:9af9bf28430b00a0cedeb2ec29acadce45e6afcecd8bdf31c793c624cfa75fa7
        - registry.k8s.io/sig-storage/livenessprobe@sha256:82adbebdf5d5a1f40f246aef8ddbee7f89dea190652aefe83336008e69f9a89f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:e998f22243869416f9860fc6a1fb07d4202eac8846defc1b85ebd015c1207605
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:76a47d100de1f2afa23803bec503d9440605db03c7179b231f963e5663a3d12a
        - registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:859b6d1b2b36a48391e347c2a7431cafa08decafb56f801e21e544c3460a4b4b
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:2cddcc716c1930775228d56b0d2d339358647629701047edfdad5fcdfaf4ebcb
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:04c55b93a032ad8075cae540d60fc740a72c31006590dbdb61ad4ab9813959d0
        - registry.connect.redhat.com/nutanix/ntnx-csi-os@sha256:92962320ad682a7113aeadc8f9de1c7849996e06819acf00925aa8aa3bf51068
        version: 2.6.11
      entries:
      - name: nutanixcsioperator.v2.6.11
        version: 2.6.11
      name: stable
    - currentCSV: nutanixcsioperator.v3.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crd.nutanix.com/v1alpha1",
                "kind": "NutanixCsiStorage",
                "metadata": {
                  "name": "nutanixcsistorage"
                },
                "spec": {
                  "ntnxInitConfigMap": {
                    "usePC": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:9cc0ffc6e9c0a82320e207dad6b6b3ddc30ab8f83fa9e099a34450d5537ca435
          createdAt: "2025-06-18T11:04:03Z"
          description: An Operator for deploying and managing the Nutanix CSI Driver
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.3.4
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi","disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/nutanix/csi-operator
          support: Nutanix Support Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a new instance of Nutanix CSI Driver
            displayName: Nutanix CSI Storage
            kind: NutanixCsiStorage
            name: nutanixcsistorages.crd.nutanix.com
            version: v1alpha1
        description: "# Important\nPlease read the following instructions carefully
          before upgrading from 2.6.x to 3.3.4, for more information please refer
          to [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)
          \n- Please do not upgrade to the CSI 3.x operator if:\n    * You are using
          LVM volumes.\n    \n- To upgrade from the CSI 2.6.x to CSI 3.3 (interacting
          with Prism Central) operator\n    * Create a Nutanix Prism Central secret
          as explained below.\n    * Delete the csidriver object from the cluster:\n\n
          \       ```\n        oc delete csidriver csi.nutanix.com\n        ```\n\n
          \   * In the installed operators, go to Nutanix CSI Operator and change
          the subscription channel from stable to stable-3.x. \n    If you have installed
          the operator with automatic update approval, the operator will be automatically
          upgraded to CSI 3.3, and then the nutanixcsistorage resource will be upgraded.\n
          \   An update plan will be generated for manual updates. Upon approval,
          the operator will be successfully upgraded.\n    \n- Direct upgrades from
          CSI 2.6.x to CSI 3.3 interacting with Prism Element are not supported. \n
          \   The only solution is to recreate the nutanixcsistorage instance by following
          the below procedure:\n    - In the installed operators, go to Nutanix CSI
          Operator and delete the nutanixcsistorage instance.\n    - Next change the
          subscription channel from stable to stable-3.x.\n    - Verify the following
          points:\n        - Ensure a Nutanix Prism Element secret is present in the
          namespace.\n        - Ensure that all the storage classes with provisioner:
          csi.nutanix.com have a controller publish secret as explained below.\n\n
          \           ```\n            csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n            csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n            ```\n\n            If this secret
          is not present in the storage class please delete and recreate the storage
          classes with the required secrets.\n    - Create a new instance of nutanixcsistorage
          from this operator by specifying `usePC: false` in YAML spec section.\n
          \   - Caution: Moving from CSI driver interacting with Prism Central to
          CSI driver interacting with Prism Element is not supported.\n    \n- Troubleshooting:\n
          \   \n    If the upgrade was unsuccessful and you want to revert to version
          CSI 2.6.x, please delete the csidriver object as explained above, uninstall
          the operator (no need to delete the nutanixcsistorage custom resource),
          and install version CSI 2.6.x from the stable channel.\n# Overview \nThe
          Nutanix CSI Operator for Kubernetes packages, deploys, manages, and upgrades
          the Nutanix CSI Driver on Kubernetes and OpenShift for dynamic provisioning
          of persistent volumes on the Nutanix Enterprise Cloud platform.\nThe Nutanix
          Container Storage Interface (CSI) Driver for Kubernetes leverages Nutanix
          Volumes and Nutanix Files to provide scalable and persistent storage for
          stateful applications.\nWith the Nutanix CSI Provider, you can:\n*  **Provide
          persistent storage to your containers**\n    * Leverage PVC resources to
          consume dynamically Nutanix storage\n    * With Files storage classes, applications
          on multiple pods can access the same storage and have the benefit of multi-pod
          read and write access. \n## Warning\nIf you upgrade from a previous version,
          make sure to validate your existing StorageClass parameters against the
          corresponding [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)\n##
          Supported Access mode\n| Storage Class mode | ReadWriteOnce | ReadOnlyMany
          | ReadWriteMany |\n|--------------------|---------------|--------------|---------------|\n|
          Nutanix Volumes    | yes           | yes          | no            |\n| Nutanix
          Files      | yes           | yes          | yes           |\n## Supported
          orchestration platforms\nThe following table details orchestration platforms
          suitable for deployment of the Nutanix CSI driver.\n| Orchestration platform
          | Version | Architecture |\n|------------------------|---------|--------------|\n|
          Red Hat OpenShift      | 4.12    | x86          |\n| Red Hat OpenShift      |
          4.13    | x86          |\n| Red Hat OpenShift      | 4.14    | x86          |\n|
          Red Hat OpenShift      | 4.15    | x86          |\n| Red Hat OpenShift      |
          4.16    | x86          |\n| Red Hat OpenShift      | 4.17    | x86          |\n|
          Red Hat OpenShift      | 4.18    | x86          |\n| Red Hat OpenShift      |
          4.19    | x86          |\n## Configuring k8s secret and storage class\nTo
          use this driver, create the relevant storage classes and secrets and deploy
          your NutanixCsiStorage instance by following the [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)
          or the example in below section:\n**1. Create a Nutanix PE secret**\n```yaml\napiVersion:
          v1\nkind: Secret\nmetadata:\n  name: ntnx-pe-secret\n  namespace: openshift-cluster-csi-drivers\nstringData:\n
          \ # prism-ip:prism-port:username:password.\n  key: 1.2.3.4:9440:admin:password\n
          \ files-key: \"fileserver01.sample.com:csi:password1\"\n```\n**2. Create
          a Nutanix PC secret**\n```yaml\napiVersion: v1\nkind: Secret\nmetadata:\n
          \ name: ntnx-pc-secret\n  namespace: openshift-cluster-csi-drivers\nstringData:\n
          \ # prism-ip:prism-port:username:password.\n  key: 1.2.3.4:9440:admin:password\n```\n**3.
          Create a NutanixCsiStorage resource to deploy your driver**\nYou can do
          it directly inside the Operator UI with the `Create instance` button or
          with the following resource\n```yaml\napiVersion: crd.nutanix.com/v1alpha1\nkind:
          NutanixCsiStorage\nmetadata:\n  name: nutanixcsistorage\n  namespace: openshift-cluster-csi-drivers\nspec:
          \n  ntnxInitConfigMap:\n    usePC : true (false for PE based installations.
          Default is true.)\n```\n**4. Create storage classes**\n- volume mode example
          for 3.x installation\n```yaml\nkind: StorageClass\napiVersion: storage.k8s.io/v1\nmetadata:\n
          \ name: nutanix-volume\nprovisioner: csi.nutanix.com\nparameters:\n  csi.storage.k8s.io/provisioner-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/node-publish-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-expand-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/fstype: ext4\n  storageContainer:
          default-container\n  storageType: NutanixVolumes\n  #description: \"description
          added to each storage object created by the driver\"\n  #hypervisorAttached:
          ENABLED\nallowVolumeExpansion: true\nreclaimPolicy: Delete\n```\n- volume
          mode example for 2.x installation\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n  name: nutanix-volume\nprovisioner: csi.nutanix.com\nparameters:\n
          \ csi.storage.k8s.io/provisioner-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/node-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/controller-expand-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/fstype: ext4\n   storageContainer:
          default-container\n   storageType: NutanixVolumes\n   #description: \"description
          added to each storage object created by the driver\"\n   #isSegmentedIscsiNetwork:
          \"false\"\n   #whitelistIPMode: ENABLED\n   #chapAuth: ENABLED\n   #isLVMVolume:
          \"false\"\n   #numLVMDisks: 4\n allowVolumeExpansion: true\n reclaimPolicy:
          Delete\n```\n- dynamic files mode example\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n    name: nutanix-dynfiles\nprovisioner: csi.nutanix.com\nparameters:\n
          \ dynamicProv: ENABLED\n  nfsServerName: fs\n  csi.storage.k8s.io/provisioner-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/node-publish-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-expand-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  storageType: NutanixFiles\n  squashType:
          \"none\"\n  #description: \"description added to each storage object created
          by the driver\"\nallowVolumeExpansion: true\n```"
        displayName: Nutanix CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Nutanix
        - CSI
        - Volumes
        - Files
        - Storage
        links:
        - name: Documentation
          url: https://portal.nutanix.com/page/documents/list?type=software&filterKey=product&filterVal=CSI
        maintainers:
        - email: cloudnative@nutanix.com
          name: Nutanix Cloud Native
        maturity: beta
        provider:
          name: Nutanix
        relatedImages:
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:9af9bf28430b00a0cedeb2ec29acadce45e6afcecd8bdf31c793c624cfa75fa7
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:04c55b93a032ad8075cae540d60fc740a72c31006590dbdb61ad4ab9813959d0
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:9e581022a4e063b9d8b4f9466a36eabe4d0ced3badc8e353f13b700074ce4553
        - registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:9cc0ffc6e9c0a82320e207dad6b6b3ddc30ab8f83fa9e099a34450d5537ca435
        - registry.connect.redhat.com/nutanix/ntnx-csi-os@sha256:f0e191e7bf98829dbf718978fc971444436a1ecb9957ee22a5d8f9ba01a931b7
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0f64602ea791246712b51df334bbd701a0f31df9950a4cb9c28c059f367baa9e
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:f1bd6ee18c4021c1c94f29edfab89b49b6a4d1b800936c19dbef2d75f8202f2d
        - registry.connect.redhat.com/nutanix/ntnx-csi-os@sha256:9f121929227acf5581a813dc2f952105f517647c11a39cd7ea0863e768082586
        - registry.k8s.io/sig-storage/livenessprobe@sha256:82adbebdf5d5a1f40f246aef8ddbee7f89dea190652aefe83336008e69f9a89f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:3c116f543f0590aeff3299c8bb0683f250817d11a77d9e9071b15a0bffdabcd9
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:10624570c0aceb03f55f1eb07147b0c537e4676869cca2e9bd4bab113f810ac4
        - registry.k8s.io/sig-storage/csi-attacher@sha256:d7325367ab72b2d469a5091d87b4fc01142d2d13d1a28b2defbbe3e6fdbc4611
        version: 3.3.4
      entries:
      - name: nutanixcsioperator.v3.3.4
        version: 3.3.4
      name: stable-3.x
    defaultChannel: stable-3.x
    packageName: nutanixcsioperator
    provider:
      name: Nutanix
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/confidential-containers
    name: trustee-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trustee-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          containerImage: registry.redhat.io/build-of-trustee/trustee-rhel9-operator@sha256:fe8e23de14ad088c79b1bd5e7aab08a1b88a86679effcbf174be81e761bd9d6d
          createdAt: "2024-06-07T10:06:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trustee-operator-system
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/trustee-operator
          support: Confidential Containers Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KbsConfig is the Schema for the kbsconfigs API
            displayName: Kbs Config
            kind: KbsConfig
            name: kbsconfigs.confidentialcontainers.org
            version: v1alpha1
        description: Remote attestation operator for confidential containers
        displayName: Red Hat build of Trustee
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trustee
        - Confidential Computing
        - Attestation
        links:
        - name: Red Hat build of Trustee
          url: https://github.com/openshift/trustee-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
          url: https://github.com/confidential-containers
        relatedImages:
        - registry.redhat.io/build-of-trustee/trustee-rhel9-operator@sha256:fe8e23de14ad088c79b1bd5e7aab08a1b88a86679effcbf174be81e761bd9d6d
        version: 0.4.2
      entries:
      - name: trustee-operator.v0.4.2
        version: 0.4.2
      - name: trustee-operator.v0.4.1
        version: 0.4.1
      - name: trustee-operator.v0.4.0
        version: 0.4.0
      - name: trustee-operator.v0.3.0
        version: 0.3.0
      - name: trustee-operator.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: trustee-operator
    provider:
      name: Red Hat
      url: https://github.com/confidential-containers
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NooBaa
      provider-url: ""
    name: mcg-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mcg-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[{"kind":"NooBaa","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"noobaa","creationTimestamp":null},"spec":{"cleanupPolicy":{},"security":{"kms":{}},"loadBalancerSourceSubnets":{},"autoscaler":{}},"status":{"lastKeyRotateTime":null}},{"kind":"BackingStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"NamespaceStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"BucketClass","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{},"status":{}},{"kind":"NooBaaAccount","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"allow_bucket_creation":false},"status":{}}]'
          capabilities: Deep Insights
          categories: Storage,Big Data
          certified: "false"
          containerImage: registry.redhat.io/odf4/mcg-rhel9-operator@sha256:43ddea9d3c1cde5c8b38f74e8840a871dd0441ed9fd074528bacf78b439f024f
          createdAt: "2019-07-08T13:10:20.940Z"
          description: NooBaa is an object data service for hybrid and multi cloud
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operators.openshift.io/infrastructure-features: ֿ'["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/noobaa/noobaa-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A NooBaa system - Create this to start
            displayName: NooBaa
            kind: NooBaa
            name: noobaas.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos,
              PV's and more. Used in BucketClass to construct data placement policies.
            displayName: Backing Store
            kind: BackingStore
            name: backingstores.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos
              and more. Used in BucketClass to construct namespace policies.
            displayName: Namespace Store
            kind: NamespaceStore
            name: namespacestores.noobaa.io
            version: v1alpha1
          - description: Storage policy spec  tiering, mirroring, spreading, namespace
              policy. Combines BackingStores Or NamespaceStores. Referenced by ObjectBucketClaims.
            displayName: Bucket Class
            kind: BucketClass
            name: bucketclasses.noobaa.io
            version: v1alpha1
          - kind: NooBaaAccount
            name: noobaaaccounts.noobaa.io
            version: v1alpha1
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information.
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: "The noobaa operator creates and reconciles a NooBaa system in
          a Kubernetes/Openshift cluster.\n\nNooBaa provides an S3 object-store service
          abstraction and data placement policies to create hybrid and multi cloud
          data solutions.\n\nFor more information on using NooBaa refer to [Github](https://github.com/noobaa/noobaa-core)
          / [Website](https://www.noobaa.io) / [Articles](https://noobaa.desk.com).
          \n\n## How does it work?\n\n- The operator deploys the noobaa core pod and
          two services - Mgmt (UI/API) and S3 (object-store).\n- Both services require
          credentials which you will get from a secret that the operator creates -
          use describe noobaa to locate it.\n- The service addresses will also appear
          in the describe output - pick the one that is suitable for your client:\n
          \   - minikube - use the NodePort address.\n    - remote cluster - probably
          need one of the External addresses.\n    - connect an application on the
          same cluster - use Internal DNS (though any address should work)\n    \n-
          Feel free to email us or open github issues on any question.\n\n## Getting
          Started\n\n### Notes:\n- The following instructions are for **minikube**
          but it works on any Kubernetes/Openshift clusters.\n- This will setup noobaa
          in the **my-noobaa-operator** namespace.\n- You will need **jq**, **curl**,
          **kubectl** or **oc**, **aws-cli**.\n\n### 1. Install OLM (if you don't
          have it already):\n```\ncurl -sL https://github.com/operator-framework/operator-lifecycle-manager/releases/download/0.12.0/install.sh
          | bash -s 0.12.0\n```\n\n### 2. Install noobaa-operator:\n```\nkubectl create
          -f https://operatorhub.io/install/noobaa-operator.yaml\n```\nWait for it
          to be ready:\n```\nkubectl wait pod -n my-noobaa-operator -l noobaa-operator
          --for=condition=ready\n```\n\n### 3. Create noobaa system:\n```\ncurl -sL
          https://operatorhub.io/api/operator?packageName=noobaa-operator | \n    jq
          '.operator.customResourceDefinitions[0].yamlExample | .metadata.namespace=\"my-noobaa-operator\"'
          |\n    kubectl create -f -\n```\nWait for it to be ready:\n```\nkubectl
          wait pod -n my-noobaa-operator -l noobaa-core --for=condition=ready\nkubectl
          get noobaa -n my-noobaa-operator -w\n# NAME     PHASE   MGMT-ENDPOINTS                  S3-ENDPOINTS
          \                   IMAGE                    AGE\n# noobaa   **Ready**   [https://192.168.64.12:31121]
          \  [https://192.168.64.12:32557]   noobaa/noobaa-core:4.0   19m\n```\n\n###
          4. Get system information to your shell:\n```\nNOOBAA_SECRET=$(kubectl get
          noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.accounts.admin.secretRef.name'
          )\nNOOBAA_MGMT=$(kubectl get noobaa noobaa -n my-noobaa-operator -o json
          | jq -r '.status.services.serviceMgmt.nodePorts[0]' )\nNOOBAA_S3=$(kubectl
          get noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.services.serviceS3.nodePorts[0]'
          )\nNOOBAA_ACCESS_KEY=$(kubectl get secret $NOOBAA_SECRET -n my-noobaa-operator
          -o json | jq -r '.data.AWS_ACCESS_KEY_ID|@base64d')\nNOOBAA_SECRET_KEY=$(kubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq -r '.data.AWS_SECRET_ACCESS_KEY|@base64d')\n```\n\n###
          5. Connect to Mgmt UI:\n```\n# show email/password from the secret:\nkubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq '.data|map_values(@base64d)'\n\n#
          open mgmt UI login:\nopen $NOOBAA_MGMT\n```\n\n### 6. Connect to S3 with
          aws-cli:\n```\nalias s3='AWS_ACCESS_KEY_ID=$NOOBAA_ACCESS_KEY AWS_SECRET_ACCESS_KEY=$NOOBAA_SECRET_KEY
          aws --endpoint $NOOBAA_S3 --no-verify-ssl s3'\ns3 ls\ns3 sync /var/log/
          s3://first.bucket\ns3 ls s3://first.bucket\n```\n"
        displayName: NooBaa Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - noobaa
        - kubernetes
        - openshift
        - cloud
        - hybrid
        - multi
        - data
        - storage
        - s3
        - tiering
        - mirroring
        links:
        - name: Github
          url: https://github.com/noobaa/noobaa-core
        - name: Website
          url: https://www.noobaa.io
        - name: Articles
          url: https://noobaa.desk.com
        maintainers:
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: etamir@redhat.com
          name: Eran Tamir
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: NooBaa
        relatedImages:
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:f37e2c611f4cda32172d4e8938938e488076fd8a8f19c252c56a80c48044bc02
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhel8/postgresql-12@sha256:af1ab437b165fa4ee9dfd6ade637d7de5fa556d0b3851fb8edc78afe0e6fa95e
        - registry.redhat.io/odf4/mcg-rhel9-operator@sha256:43ddea9d3c1cde5c8b38f74e8840a871dd0441ed9fd074528bacf78b439f024f
        - registry.redhat.io/odf4/odf-cosi-sidecar-rhel9@sha256:bce36c7d33a85a11cd6b6b24f92887cffe6f38af70340e54d10e4f6f524aa782
        version: 4.15.20-rhodf
      entries:
      - name: mcg-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: mcg-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: mcg-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: mcg-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: mcg-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: mcg-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: mcg-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: mcg-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: mcg-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: mcg-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: mcg-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: mcg-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: mcg-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: mcg-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: mcg-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: mcg-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: mcg-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: mcg-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: mcg-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: mcg-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: mcg-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[{"kind":"NooBaa","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"noobaa","creationTimestamp":null},"spec":{"cleanupPolicy":{},"security":{"kms":{}},"loadBalancerSourceSubnets":{},"autoscaler":{}},"status":{"lastKeyRotateTime":null}},{"kind":"BackingStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"NamespaceStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"BucketClass","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{},"status":{}},{"kind":"NooBaaAccount","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"allow_bucket_creation":false},"status":{}}]'
          capabilities: Deep Insights
          categories: Storage,Big Data
          certified: "false"
          containerImage: registry.redhat.io/odf4/mcg-rhel9-operator@sha256:fdcdc8870b12d853973000877818463c3b4a5ed1d3316ff72a04a299577ed23d
          createdAt: "2019-07-08T13:10:20.940Z"
          description: NooBaa is an object data service for hybrid and multi cloud
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/infrastructure-features: ֿ'["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/noobaa/noobaa-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A NooBaa system - Create this to start
            displayName: NooBaa
            kind: NooBaa
            name: noobaas.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos,
              PV's and more. Used in BucketClass to construct data placement policies.
            displayName: Backing Store
            kind: BackingStore
            name: backingstores.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos
              and more. Used in BucketClass to construct namespace policies.
            displayName: Namespace Store
            kind: NamespaceStore
            name: namespacestores.noobaa.io
            version: v1alpha1
          - description: Storage policy spec  tiering, mirroring, spreading, namespace
              policy. Combines BackingStores Or NamespaceStores. Referenced by ObjectBucketClaims.
            displayName: Bucket Class
            kind: BucketClass
            name: bucketclasses.noobaa.io
            version: v1alpha1
          - kind: NooBaaAccount
            name: noobaaaccounts.noobaa.io
            version: v1alpha1
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information.
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: "The noobaa operator creates and reconciles a NooBaa system in
          a Kubernetes/Openshift cluster.\n\nNooBaa provides an S3 object-store service
          abstraction and data placement policies to create hybrid and multi cloud
          data solutions.\n\nFor more information on using NooBaa refer to [Github](https://github.com/noobaa/noobaa-core)
          / [Website](https://www.noobaa.io) / [Articles](https://noobaa.desk.com).
          \n\n## How does it work?\n\n- The operator deploys the noobaa core pod and
          two services - Mgmt (UI/API) and S3 (object-store).\n- Both services require
          credentials which you will get from a secret that the operator creates -
          use describe noobaa to locate it.\n- The service addresses will also appear
          in the describe output - pick the one that is suitable for your client:\n
          \   - minikube - use the NodePort address.\n    - remote cluster - probably
          need one of the External addresses.\n    - connect an application on the
          same cluster - use Internal DNS (though any address should work)\n    \n-
          Feel free to email us or open github issues on any question.\n\n## Getting
          Started\n\n### Notes:\n- The following instructions are for **minikube**
          but it works on any Kubernetes/Openshift clusters.\n- This will setup noobaa
          in the **my-noobaa-operator** namespace.\n- You will need **jq**, **curl**,
          **kubectl** or **oc**, **aws-cli**.\n\n### 1. Install OLM (if you don't
          have it already):\n```\ncurl -sL https://github.com/operator-framework/operator-lifecycle-manager/releases/download/0.12.0/install.sh
          | bash -s 0.12.0\n```\n\n### 2. Install noobaa-operator:\n```\nkubectl create
          -f https://operatorhub.io/install/noobaa-operator.yaml\n```\nWait for it
          to be ready:\n```\nkubectl wait pod -n my-noobaa-operator -l noobaa-operator
          --for=condition=ready\n```\n\n### 3. Create noobaa system:\n```\ncurl -sL
          https://operatorhub.io/api/operator?packageName=noobaa-operator | \n    jq
          '.operator.customResourceDefinitions[0].yamlExample | .metadata.namespace=\"my-noobaa-operator\"'
          |\n    kubectl create -f -\n```\nWait for it to be ready:\n```\nkubectl
          wait pod -n my-noobaa-operator -l noobaa-core --for=condition=ready\nkubectl
          get noobaa -n my-noobaa-operator -w\n# NAME     PHASE   MGMT-ENDPOINTS                  S3-ENDPOINTS
          \                   IMAGE                    AGE\n# noobaa   **Ready**   [https://192.168.64.12:31121]
          \  [https://192.168.64.12:32557]   noobaa/noobaa-core:4.0   19m\n```\n\n###
          4. Get system information to your shell:\n```\nNOOBAA_SECRET=$(kubectl get
          noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.accounts.admin.secretRef.name'
          )\nNOOBAA_MGMT=$(kubectl get noobaa noobaa -n my-noobaa-operator -o json
          | jq -r '.status.services.serviceMgmt.nodePorts[0]' )\nNOOBAA_S3=$(kubectl
          get noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.services.serviceS3.nodePorts[0]'
          )\nNOOBAA_ACCESS_KEY=$(kubectl get secret $NOOBAA_SECRET -n my-noobaa-operator
          -o json | jq -r '.data.AWS_ACCESS_KEY_ID|@base64d')\nNOOBAA_SECRET_KEY=$(kubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq -r '.data.AWS_SECRET_ACCESS_KEY|@base64d')\n```\n\n###
          5. Connect to Mgmt UI:\n```\n# show email/password from the secret:\nkubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq '.data|map_values(@base64d)'\n\n#
          open mgmt UI login:\nopen $NOOBAA_MGMT\n```\n\n### 6. Connect to S3 with
          aws-cli:\n```\nalias s3='AWS_ACCESS_KEY_ID=$NOOBAA_ACCESS_KEY AWS_SECRET_ACCESS_KEY=$NOOBAA_SECRET_KEY
          aws --endpoint $NOOBAA_S3 --no-verify-ssl s3'\ns3 ls\ns3 sync /var/log/
          s3://first.bucket\ns3 ls s3://first.bucket\n```\n"
        displayName: NooBaa Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - noobaa
        - kubernetes
        - openshift
        - cloud
        - hybrid
        - multi
        - data
        - storage
        - s3
        - tiering
        - mirroring
        links:
        - name: Github
          url: https://github.com/noobaa/noobaa-core
        - name: Website
          url: https://www.noobaa.io
        - name: Articles
          url: https://noobaa.desk.com
        maintainers:
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: etamir@redhat.com
          name: Eran Tamir
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: NooBaa
        relatedImages:
        - registry.redhat.io/odf4/mcg-rhel9-operator@sha256:fdcdc8870b12d853973000877818463c3b4a5ed1d3316ff72a04a299577ed23d
        - registry.redhat.io/odf4/odf-cosi-sidecar-rhel9@sha256:9fed589373d1586577d57f23b7a4f39ace5052fc55c99a03cb256da63a26c441
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:0084d9053ad2c74926ca02d2ede0461abd59227eee98d122c45e08a5ebaf629a
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhel8/postgresql-12@sha256:af1ab437b165fa4ee9dfd6ade637d7de5fa556d0b3851fb8edc78afe0e6fa95e
        version: 4.16.16-rhodf
      entries:
      - name: mcg-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: mcg-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: mcg-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: mcg-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: mcg-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: mcg-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: mcg-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: mcg-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: mcg-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: mcg-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: mcg-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: mcg-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: mcg-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: mcg-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: mcg-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: mcg-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: mcg-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: mcg-operator
    provider:
      name: NooBaa
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-eks-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-eks-controller.v1.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "FargateProfile",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Nodegroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/eks-controller:1.9.3
          createdAt: "2025-10-03T21:43:09Z"
          description: AWS EKS controller is a service controller for managing EKS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessEntry represents the state of an AWS eks AccessEntry
              resource.
            displayName: AccessEntry
            kind: AccessEntry
            name: accessentries.eks.services.k8s.aws
            version: v1alpha1
          - description: Addon represents the state of an AWS eks Addon resource.
            displayName: Addon
            kind: Addon
            name: addons.eks.services.k8s.aws
            version: v1alpha1
          - description: Cluster represents the state of an AWS eks Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.eks.services.k8s.aws
            version: v1alpha1
          - description: FargateProfile represents the state of an AWS eks FargateProfile
              resource.
            displayName: FargateProfile
            kind: FargateProfile
            name: fargateprofiles.eks.services.k8s.aws
            version: v1alpha1
          - description: IdentityProviderConfig represents the state of an AWS eks
              IdentityProviderConfig resource.
            displayName: IdentityProviderConfig
            kind: IdentityProviderConfig
            name: identityproviderconfigs.eks.services.k8s.aws
            version: v1alpha1
          - description: Nodegroup represents the state of an AWS eks Nodegroup resource.
            displayName: Nodegroup
            kind: Nodegroup
            name: nodegroups.eks.services.k8s.aws
            version: v1alpha1
          - description: PodIdentityAssociation represents the state of an AWS eks
              PodIdentityAssociation resource.
            displayName: PodIdentityAssociation
            kind: PodIdentityAssociation
            name: podidentityassociations.eks.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Kubernetes Service (EKS) resources in AWS from within your Kubernetes cluster.

          **About Amazon EKS**

          Amazon Elastic Kubernetes Service (Amazon EKS) is a managed service that you can use to run Kubernetes on AWS without needing to install, operate, and maintain your own Kubernetes control plane or nodes. Kubernetes is an open-source system for automating the deployment, scaling, and management of containerized applications.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EKS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eks
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EKS Developer Resources
          url: https://aws.amazon.com/eks/getting-started/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/eks-controller:1.9.3
        version: 1.9.3
      entries:
      - name: ack-eks-controller.v1.9.3
        version: 1.9.3
      - name: ack-eks-controller.v1.9.2
        version: 1.9.2
      - name: ack-eks-controller.v1.9.1
        version: 1.9.1
      - name: ack-eks-controller.v1.9.0
        version: 1.9.0
      - name: ack-eks-controller.v1.8.0
        version: 1.8.0
      - name: ack-eks-controller.v1.7.6
        version: 1.7.6
      - name: ack-eks-controller.v1.7.5
        version: 1.7.5
      - name: ack-eks-controller.v1.7.4
        version: 1.7.4
      - name: ack-eks-controller.v1.7.3
        version: 1.7.3
      - name: ack-eks-controller.v1.7.2
        version: 1.7.2
      - name: ack-eks-controller.v1.6.4
        version: 1.6.4
      - name: ack-eks-controller.v1.6.3
        version: 1.6.3
      - name: ack-eks-controller.v1.6.2
        version: 1.6.2
      - name: ack-eks-controller.v1.6.1
        version: 1.6.1
      - name: ack-eks-controller.v1.6.0
        version: 1.6.0
      - name: ack-eks-controller.v1.5.5
        version: 1.5.5
      - name: ack-eks-controller.v1.5.4
        version: 1.5.4
      - name: ack-eks-controller.v1.5.3
        version: 1.5.3
      - name: ack-eks-controller.v1.5.2
        version: 1.5.2
      - name: ack-eks-controller.v1.5.1
        version: 1.5.1
      - name: ack-eks-controller.v1.5.0
        version: 1.5.0
      - name: ack-eks-controller.v1.4.7
        version: 1.4.7
      - name: ack-eks-controller.v1.4.6
        version: 1.4.6
      - name: ack-eks-controller.v1.4.5
        version: 1.4.5
      - name: ack-eks-controller.v1.4.4
        version: 1.4.4
      - name: ack-eks-controller.v1.4.3
        version: 1.4.3
      - name: ack-eks-controller.v1.4.2
        version: 1.4.2
      - name: ack-eks-controller.v1.4.1
        version: 1.4.1
      - name: ack-eks-controller.v1.4.0
        version: 1.4.0
      - name: ack-eks-controller.v1.3.5
        version: 1.3.5
      - name: ack-eks-controller.v1.3.3
        version: 1.3.3
      - name: ack-eks-controller.v1.3.1
        version: 1.3.1
      - name: ack-eks-controller.v1.3.0
        version: 1.3.0
      - name: ack-eks-controller.v1.2.3
        version: 1.2.3
      - name: ack-eks-controller.v1.2.2
        version: 1.2.2
      - name: ack-eks-controller.v1.2.1
        version: 1.2.1
      - name: ack-eks-controller.v1.2.0
        version: 1.2.0
      - name: ack-eks-controller.v1.1.3
        version: 1.1.3
      - name: ack-eks-controller.v1.1.2
        version: 1.1.2
      - name: ack-eks-controller.v1.1.1
        version: 1.1.1
      - name: ack-eks-controller.v1.1.0
        version: 1.1.0
      - name: ack-eks-controller.v1.0.5
        version: 1.0.5
      - name: ack-eks-controller.v1.0.4
        version: 1.0.4
      - name: ack-eks-controller.v1.0.3
        version: 1.0.3
      - name: ack-eks-controller.v1.0.2
        version: 1.0.2
      - name: ack-eks-controller.v1.0.1
        version: 1.0.1
      - name: ack-eks-controller.v1.0.0
        version: 1.0.0
      - name: ack-eks-controller.v0.1.9
        version: 0.1.9
      - name: ack-eks-controller.v0.1.8
        version: 0.1.8
      - name: ack-eks-controller.v0.1.7
        version: 0.1.7
      - name: ack-eks-controller.v0.1.6
        version: 0.1.6
      - name: ack-eks-controller.v0.1.5
        version: 0.1.5
      - name: ack-eks-controller.v0.1.4
        version: 0.1.4
      - name: ack-eks-controller.v0.1.3
        version: 0.1.3
      - name: ack-eks-controller.v0.1.2
        version: 0.1.2
      - name: ack-eks-controller.v0.1.1
        version: 0.1.1
      - name: ack-eks-controller.v0.1.0
        version: 0.1.0
      - name: ack-eks-controller.v0.0.15
        version: 0.0.15
      - name: ack-eks-controller.v0.0.14
        version: 0.0.14
      - name: ack-eks-controller.v0.0.13
        version: 0.0.13
      - name: ack-eks-controller.v0.0.12
        version: 0.0.12
      - name: ack-eks-controller.v0.0.11
        version: 0.0.11
      - name: ack-eks-controller.v0.0.10
        version: 0.0.10
      - name: ack-eks-controller.v0.0.9
        version: 0.0.9
      - name: ack-eks-controller.v0.0.8
        version: 0.0.8
      - name: ack-eks-controller.v0.0.7
        version: 0.0.7
      - name: ack-eks-controller.v0.0.6
        version: 0.0.6
      - name: ack-eks-controller.v0.0.5
        version: 0.0.5
      - name: ack-eks-controller.v0.0.4
        version: 0.0.4
      - name: ack-eks-controller.v0.0.3
        version: 0.0.3
      - name: ack-eks-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-eks-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Carlos Eduardo de Paula
      provider-url: ""
    name: lbconfig-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lbconfig-operator.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-infra-sample",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "http",
                    "path": "/healthz",
                    "port": 1936
                  },
                  "ports": [
                    80,
                    443
                  ],
                  "provider": {
                    "creds": "f5-creds",
                    "host": "https://192.168.1.35",
                    "partition": "Common",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "F5_BigIP"
                  },
                  "type": "infra",
                  "vip": "192.168.1.45"
                }
              },
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-infra-sample-shard",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "http",
                    "path": "/healthz",
                    "port": 1936
                  },
                  "nodelabels": {
                    "kubernetes.io/region": "DC1",
                    "node-role.kubernetes.io/infra": ""
                  },
                  "ports": [
                    80,
                    443
                  ],
                  "provider": {
                    "creds": "f5-creds",
                    "host": "https://192.168.1.35",
                    "partition": "Common",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "F5_BigIP"
                  },
                  "type": "infra",
                  "vip": "192.168.1.46"
                }
              },
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-master-sample",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "https",
                    "path": "/healthz",
                    "port": 6443
                  },
                  "ports": [
                    6443
                  ],
                  "provider": {
                    "creds": "netscaler-creds",
                    "host": "https://192.168.1.36",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "Citrix_ADC"
                  },
                  "type": "master",
                  "vip": "192.168.1.40"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Networking
          certified: "false"
          containerImage: quay.io/carlosedp/lbconfig-operator:v0.5.1
          createdAt: "2025-01-09T15:45:28Z"
          description: Manage External Load Balancers allowing creation/update for
            VIPs and Servers dynamically via API.
          k8sMaxVersion: ""
          k8sMinVersion: ""
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/carlosedp/lbconfig-operator
          support: https://github.com/carlosedp/lbconfig-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalLoadBalancer is the Schema for the externalloadbalancers
              API
            displayName: ExternalLoadBalancer Instance
            kind: ExternalLoadBalancer
            name: externalloadbalancers.lb.lbconfig.carlosedp.com
            version: v1
          required:
          - description: ExternalLoadBalancer represents a configured instance of
              an external Load-Balancer for a specific group of nodes of the cluster.
              The Instance has a VIP and ports to be balanced to the cluster nodes
              based on a set of node labels.
            displayName: External Load-Balancer Configuration Instance
            kind: ExternalLoadBalancer
            name: externalloadbalancers.lb.lbconfig.carlosedp.com
            version: v1
        description: |
          ## About the Operator

          The LBConfig Operator, manages the configuration of External Load Balancer instances (on third-party equipment via it's API) and creates VIPs and IP Pools with Monitors for a set of OpenShift or Kubernetes nodes like Master-nodes (Control-Plane), Infra nodes (where the Routers or Ingress controllers are located) or based on it's roles and/or labels.

          The operator dynamically handles creating, updating or deleting the IPs of the pools in the Load Balancer based on the Node IPs for each role or label. On every change of the operator configuration (CRDs) or addition/change/removal or cluster Nodes, the operator updates the Load Balancer properly.

          ## Main Use Cases

          The main users for this operator is enterprise deployments or clusters composed of multiple nodes having an external load-balancer providing the balancing and high-availability to access the cluster in both API and Application levels.

          ### Create ExternalLoadBalancer Instances

          Create the instances for each Load Balancer instance you need (for example one for Master Nodes and another for the Infra Nodes). **If installing on OpenShift or Kubernetes with OLM (or in a different namespace), adjust the sample YAMLs to match the created namespace**.

          **The provider `vendor` field can be (case-insensitive):**

          * **`F5_BigIP`** - Tested on F5 BigIP version 15
          * **`Citrix_ADC`** - Tested on Citrix ADC (Netscaler) version 13
          * **`HAProxy`** - HAProxy with Dataplane API. ([Docs](./docs/haproxy/))
          * **`Dummy`** - Dummy backend used for testing to only print log messages on operations

          Create the secret holding the Load Balancer API user and password:

          ```sh
          oc create secret generic f5-creds --from-literal=username=admin --from-literal=password=admin123 --namespace lbconfig-operator-system
          ```

          #### Sample CRDs and Available Fields

          Master Nodes using a Citrix ADC LB:

          ```yaml
          apiVersion: lb.lbconfig.carlosedp.com/v1
          kind: ExternalLoadBalancer
          metadata:
            name: externalloadbalancer-master-sample
            namespace: lbconfig-operator-system
          spec:
            vip: "192.168.1.40"
            type: "master"
            ports:
              - 6443
            monitor:
              path: "/healthz"
              port: 6443
              monitortype: "https"
            provider:
              vendor: Citrix_ADC
              host: "https://192.168.1.36"
              port: 443
              creds: netscaler-creds
              validatecerts: false
          ```

          Infra Nodes using a F5 BigIP LB:

          ```yaml
          apiVersion: lb.lbconfig.carlosedp.com/v1
          kind: ExternalLoadBalancer
          metadata:
            name: externalloadbalancer-infra-sample
            namespace: lbconfig-operator-system
          spec:
            vip: "192.168.1.45"
            type: "infra"
            ports:
              - 80
              - 443
            monitor:
              path: "/healthz"
              port: 1936
              monitortype: http
            provider:
              vendor: F5_BigIP
              host: "https://192.168.1.35"
              port: 443
              creds: f5-creds
              partition: "Common"
              validatecerts: false
          ```

          To choose the nodes which will be part of the server pool, you can set either `type` or `nodelabels` fields. The yaml field `type: "master"` or `type: "infra"` selects nodes with the role label `"node-role.kubernetes.io/master"` and `"node-role.kubernetes.io/infra"` respectively. If the field `nodelabels` array is used instead, the operator will use nodes which match all labels.

          Clusters with sharded routers or using arbitrary labels to determine where the Ingress Controllers run can be configured like:

          ```yaml
          spec:
            vip: "10.0.0.6"
            ports:
              - 80
            nodelabels:
              "node.kubernetes.io/ingress-controller": "production"
              "kubernetes.io/region": "DC1"
            ...
          ```

          ## Getting Help

          This project fully open-source and is hosted at [GitHub](https://github.com/carlosedp/lbconfig-operator). Always check the project page and issues for the latest updates and/or solutions to problems encountered. If a problem is detected, please fill an issue. Feature request and improvement issues are welcome.

          Some fields inside `providers` are optional and depend on the used backend. Check the [API docs](https://pkg.go.dev/github.com/carlosedp/lbconfig-operator/api/lb.lbconfig.carlosedp.com/v1?utm_source=gopls#Provider) which fields are backend-specific.

          ## Disclaimers

          * The operator does not check if the requested configuration (names, IPs) already exists and/or conflicts with existing configuration in the Load Balancer. The user is responsible for these checks before deployment;
          * I am not responsible if the operator changes/deletes existing configuration on the Load Balancer if existing names are already configured.
          * The operator creates the entries(Pools, VIPs, Monitors) in the provided Load Balancer with the `name` of the instance configured in the CustomResource prefixed with the type. Eg. For a CR with name `externalloadbalancer-master-sample`, the operator creates a server pool named `Pool-externalloadbalancer-master-sample-6443` (suffixed with the port), a monitor named `Monitor-externalloadbalancer-master-sample` and a VIP named `VIP-externalloadbalancer-master-sample-6443` (suffixed with the port).
        displayName: External Load-Balancer Configuration Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 'load-balance '
        - infrastructure
        links:
        - name: GitHub
          url: https://github.com/carlosedp/lbconfig-operator
        - name: Documentation
          url: https://github.com/carlosedp/lbconfig-operator/docs
        - name: Issues
          url: https://github.com/carlosedp/lbconfig-operator/issues
        maintainers:
        - email: carlosedp@gmail.com
          name: Carlos Eduardo de Paula
        maturity: beta
        minKubeVersion: 1.18.0
        provider:
          name: Carlos Eduardo de Paula
        relatedImages:
        - quay.io/carlosedp/kube-rbac-proxy:v0.15.0
        - quay.io/carlosedp/lbconfig-operator:v0.5.1
        version: 0.5.1
      entries:
      - name: lbconfig-operator.v0.5.1
        version: 0.5.1
      - name: lbconfig-operator.v0.5.0
        version: 0.5.0
      - name: lbconfig-operator.v0.4.1
        version: 0.4.1
      - name: lbconfig-operator.v0.4.0
        version: 0.4.0
      - name: lbconfig-operator.v0.2.0
        version: 0.2.0
      name: beta
    defaultChannel: beta
    packageName: lbconfig-operator
    provider:
      name: Carlos Eduardo de Paula
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: group-sync-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: group-sync-operator.v0.0.36
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "azure-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "azure": {
                        "credentialsSecret": {
                          "name": "azure-group-sync",
                          "namespace": "group-sync-operator"
                        }
                      },
                      "name": "azure"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "github-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "github": {
                        "credentialsSecret": {
                          "name": "github-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "organization": "redhat-cop"
                      },
                      "name": "github"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "gitlab-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "gitlab": {
                        "credentialsSecret": {
                          "name": "gitlab-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "url": "https://gitlab.apps.openshift.com"
                      },
                      "name": "gitlab"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "keycloak-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "keycloak": {
                        "credentialsSecret": {
                          "name": "keycloak-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "insecure": true,
                        "realm": "ocp",
                        "url": "https://keycloak.apps.openshift.com"
                      },
                      "name": "keycloak"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "ldap-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "ldap": {
                        "credentialsSecret": {
                          "name": "ldap-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "insecure": true,
                        "rfc2307": {
                          "groupMembershipAttributes": [
                            "member"
                          ],
                          "groupNameAttributes": [
                            "cn"
                          ],
                          "groupUIDAttribute": "dn",
                          "groupsQuery": {
                            "baseDN": "ou=Groups,dc=example,dc=com",
                            "derefAliases": "never",
                            "filter": "(objectClass=groupofnames)",
                            "scope": "sub"
                          },
                          "tolerateMemberNotFoundErrors": true,
                          "tolerateMemberOutOfScopeErrors": true,
                          "userNameAttributes": [
                            "cn"
                          ],
                          "userUIDAttribute": "dn",
                          "usersQuery": {
                            "baseDN": "ou=Users,dc=example,dc=com",
                            "derefAliases": "never",
                            "scope": "sub"
                          }
                        },
                        "url": "ldap://localhost:389"
                      },
                      "name": "ldap"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "okta-sync"
                },
                "spec": {
                  "providers": [
                    {
                      "name": "okta",
                      "okta": {
                        "appId": "okta-sync-app-id",
                        "credentialsSecret": {
                          "name": "okta-api-token",
                          "namespace": "group-sync-operator"
                        },
                        "extractLoginUsername": true,
                        "profileKey": "login",
                        "url": "https://example.okta.com/"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/group-sync-operator@sha256:43ef79efc2a77ee5edd4e3123f9dab156a30d3e00fd1fba9f89600b96f0f68d6
          createdAt: "2025-08-12T19:19:21Z"
          description: Synchronize groups and users from external providers
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: group-sync-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/group-sync-operator
          support: Red Hat Community of Practice
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GroupSync is the Schema for the groupsyncs API
            displayName: Group Sync
            kind: GroupSync
            name: groupsyncs.redhatcop.redhat.io
            version: v1alpha1
        description: |
          Synchronizes groups from external providers into OpenShift

          ## Overview

          The OpenShift Container Platform contains functionality to synchronize groups found in external identity providers into the platform. Currently, the functionality that is included in OpenShift to limited to synchronizing LDAP only. This operator is designed to integrate with external providers in order to provide new solutions.

          Group Synchronization is facilitated by creating a `GroupSync` resource. The following describes the high level schema for this resource:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: example-groupsync
          spec:
            providers:
              - <One or more providers to synchronize>
          ```

          ## Authentication

          In most cases, authentication details must be provided in order to communicate with providers. Authentication details are provider specific with regards to the required values. In supported providers, the secret can be referenced in the `credentialsSecret` by name and namespace where it has been created as shown below:

          ```
          credentialsSecret:
            name: <secret_name>
            namespace: <secret_namespace>
          ```

          ## Providers

          Integration with external systems is made possible through a set of pluggable external providers. The following providers are currently supported:

          * [Azure](https://azure.microsoft.com/)
          * [GitHub](https://github.com)
          * [GitLab](https://gitlab.com)
          * [LDAP](https://en.wikipedia.org/wiki/Lightweight_Directory_Access_Protocol)
          * [Keycloak](https://www.keycloak.org/)/[Red Hat Single Sign On](https://access.redhat.com/products/red-hat-single-sign-on)
          * [Okta](https://www.okta.com/)
          * [IBM Security Verify](https://docs.verify.ibm.com/verify)

          The following sections describe the configuration options available for each provider


          ### Azure

          Groups contained within Azure Active Directory can be synchronized into OpenShift. The following table describes the set of configuration options for the Azure provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `authorityHost` | Azure Active Directory Endpoint | `https://login.microsoftonline.com` | No |
          | `baseGroups` | List of groups to start searching from instead of listing all groups in the directory | | No |
          | `credentialsSecret` | Name of the secret containing authentication details (See below) | | Yes |
          | `filter` | Graph API filter | | No |
          | `groups` | List of groups to filter against | | No |
          | `userNameAttributes` | Fields on a user record to use as the User Name | `userPrincipalName` | No |
          | `prune` | Prune Whether to prune groups that are no longer in Azure | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Azure provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: azure-groupsync
          spec:
            providers:
            - name: azure
              azure:
                credentialsSecret:
                  name: azure-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to Azure

          Authentication to Azure can be performed using Application Registration with access to query group information in Azure Active Directory.

          The App Registration must be granted access to the following Microsoft Graph API's:

          * Group.Read.All
          * GroupMember.Read.All
          * User.Read.All

          A secret must be created in the same namespace that contains the `GroupSync` resource:

          The following keys must be defined in the secret

          * `AZURE_TENANT_ID` - Tenant ID
          * `AZURE_CLIENT_ID` - Client ID
          * `AZURE_CLIENT_SECRET` - Client Secret

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic azure-group-sync --from-literal=AZURE_SUBSCRIPTION_ID=<AZURE_SUBSCRIPTION_ID> --from-literal=AZURE_TENANT_ID=<AZURE_TENANT_ID> --from-literal=AZURE_CLIENT_ID=<AZURE_CLIENT_ID> --from-literal=AZURE_CLIENT_SECRET=<AZURE_CLIENT_SECRET>
          ```

          ### GitHub

          Teams stored within a GitHub organization can be synchronized into OpenShift. The following table describes the set of configuration options for the GitHub provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `caSecret` | Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `organization` | Organization to synchronize against | | Yes |
          | `teams` | List of teams to filter against | | No |
          | `url` | Base URL for the GitHub or GitHub Enterprise host (Must contain a trailing slash) | | No |
          | `prune` | Prune Whether to prune groups that are no longer in GitHub | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Github provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: github-groupsync
          spec:
            providers:
            - name: github
              github:
                organization: ocp
                credentialsSecret:
                  name: github-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to GitHub

          Authentication to GitLab can be performed using a [Token](https://docs.gitlab.com/ee/security/token_overview.html) or a Username and Password (Note: 2FA not supported). A secret must be created in the same namespace that contains the `GroupSync` resource:

          When using an OAuth token, the following token types are supported:

          * Personal Access Token
          * OAuth Token
          * Job Token

          the following key is required:

          * `token` - OAuth token

          Optionally, the `tokenType` key can be specified to indicate the type of token being provided from the following values:

          * OAuth - `oauth`
          * Personal Access Token - `personal`
          * Job Token - `job`

          If no `tokenType` is provided, `oauth` is used by default

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token>
          ```

          To specify a token type, such as a Personal Access Token, the following command can be executed:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token> --from-literal=tokenType=personal
          ```

          The following keys are required for username and password:

          * `username` - Username for authenticating with GitHub
          * `password` - Password for authenticating with GitHub

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic github-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### GitLab

          Groups stored within a GitLab can be synchronized into OpenShift. The following table describes the set of configuration options for the GitLab provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `groups` | List of groups to filter against | | No |
          | `prune` | Prune Whether to prune groups that are no longer in GitLab | `false` | No |
          | `scope` | Scope for group synchronization. Options are `one` for one level or `sub` to include subgroups | `sub` | No |
          | `url` | Base URL for the GitLab instance | `https://gitlab.com` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Github provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: gitlab-groupsync
          spec:
            providers:
            - name: gitlab
              gitlab:
                credentialsSecret:
                  name: gitlab-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to GitLab

          Authentication to GitLab can be performed using an OAuth Personal Access Token or a Username and Password (Note: 2FA not supported). A secret must be created in the same namespace that contains the `GroupSync` resource:

          When using an OAuth token, the following key is required:

          * `token` - OAuth token

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token>
          ```

          The following keys are required for username and password:

          * `username` - Username for authenticating with GitLab
          * `password` - Password for authenticating with GitLab

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### LDAP

          Groups stored within an [LDAP](https://en.wikipedia.org/wiki/Lightweight_Directory_Access_Protocol) server can be synchronized into OpenShift. The LDAP provider implements the included features of the [Syncing LDAP groups](https://docs.openshift.com/container-platform/latest/authentication/ldap-syncing.html) feature and makes use of the libraries from the [OpenShift Command Line](https://docs.openshift.com/container-platform/latest/cli_reference/openshift_cli/getting-started-cli.html) tool to streamline the migration to this operator based implementation.

          The configurations of the three primary schemas (`rfc2307`, `activeDirectory` and `augmentedActiveDirectory`) can be directly migrated as is without any modification.

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | No |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `groupUIDNameMapping` | User defined name mapping | | No |
          | `rfc2307` | Configuration using the [rfc2307](https://docs.openshift.com/container-platform/latest/authentication/ldap-syncing.html#ldap-syncing-rfc2307_ldap-syncing-groups) schema | | No |
          | `activeDirectory` | Configuration using the [activeDirectory](https://docs.openshift.com/container-platform/4.5/authentication/ldap-syncing.html#ldap-syncing-activedir_ldap-syncing-groups) schema | | No |
          | `augmentedActiveDirectory` | Configuration using the [activeDirectory](https://docs.openshift.com/container-platform/4.5/authentication/ldap-syncing.html#ldap-syncing-augmented-activedir_ldap-syncing-groups) schema | | No |
          | `url` | Connection URL for the LDAP server | `ldap://ldapserver:389` | No |
          | `whitelist` | Explicit list of groups to synchronize |  | No |
          | `blacklist` | Explicit list of groups to not synchronize |  | No |
          | `prune` | Prune Whether to prune groups that are no longer in LDAP | `false` | No |

          The following is an example using the `rfc2307` schema:

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
                credentialsSecret:
                  name: ldap-group-sync
                  namespace: group-sync-operator
                insecure: true
                rfc2307:
                  groupMembershipAttributes:
                  - member
                  groupNameAttributes:
                  - cn
                  groupUIDAttribute: dn
                  groupsQuery:
                    baseDN: ou=Groups,dc=example,dc=com
                    derefAliases: never
                    filter: (objectClass=groupofnames)
                    scope: sub
                  tolerateMemberNotFoundErrors: true
                  tolerateMemberOutOfScopeErrors: true
                  userNameAttributes:
                  - cn
                  userUIDAttribute: dn
                  usersQuery:
                    baseDN: ou=Users,dc=example,dc=com
                    derefAliases: never
                    scope: sub
                url: ldap://ldapserver:389
              name: ldap
          ```

          The examples provided in the OpenShift documented referenced previously can be used to construct the schemas for the other LDAP synchronization types.

          #### Authenticating to LDAP

          If authentication is required in order to communicate with the LDAP server, a secret should be created in the same namespace that contains the `GroupSync` resource. The following keys can be defined:

          * `username` - Username (Bind DN) for authenticating with the LDAP server
          * `password` - Password for authenticating with the LDAP server

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic ldap-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          #### Whitelists and Blacklists

          Groups can be explicitly whitelisted or blacklisted in order to control the groups that are eligible to be synchronized into OpenShift. When running LDAP group synchronization using the command line, this configuration is referenced via separate files, but these are instead specified in the `blacklist` and `whitelist` properties as shown below:

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
          ...
                whitelist:
                - cn=Online Corporate Banking,ou=Groups,dc=example,dc=com
          ...
              name: ldap
          ```

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
          ...
                blacklist:
                - cn=Finance,ou=Groups,dc=example,dc=com
          ...
              name: ldap
          ```

          ### Keycloak

          Groups stored within Keycloak can be synchronized into OpenShift. The following table describes the set of configuration options for the Keycloak provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `groups` | List of groups to filter against | | No |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `loginRealm` | Realm to authenticate against | `master` | No |
          | `realm` | Realm to synchronize | | Yes |
          | `scope` | Scope for group synchronization. Options are `one` for one level or `sub` to include subgroups | `sub` | No |
          | `url` | Base URL for the Keycloak server. Older versions (<17.0.0) including Red Hat SSO should include the context path `/auth` appended to the hostname  | | Yes |
          | `prune` | Prune Whether to prune groups that are no longer in Keycloak | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Keycloak provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                url: https://keycloak-keycloak-operator.apps.openshift.com/auth
          ```

          #### Authenticating to Keycloak

          A user with rights to query for Keycloak groups must be available. The following permissions must be associated to the user:

          * Password must be set (Temporary option unselected) on the _Credentials_ tab
          * On the _Role Mappings_ tab, select _master-realm_ or _realm-management_ next to the _Client Roles_ dropdown and then select **Query Groups** and **Query Users**.

          A secret must be created in the same namespace that contains the `GroupSync` resource. It must contain the following keys for the user previously created:

          * `username` - Username for authenticating with Keycloak
          * `password` - Password for authenticating with Keycloak

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic keycloak-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### Okta

          [Okta Groups](https://help.okta.com/en/prod/Content/Topics/users-groups-profiles/usgp-main.htm) assigned to [Okta Applications](https://help.okta.com/en/prod/Content/Topics/Apps/Apps_Apps.htm) can be synchronized into OpenShift. The developer docs for the Okta API that the Okta Syncer uses can be found [here](https://developer.okta.com/docs/reference/api/apps/#list-groups-assigned-to-application).
          The following table describes the set of configuration options for the Okta provider:

            | Name | Description | Defaults | Required |
            | ----- | ---------- | -------- | ----- |
            | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
            | `groups` | List of groups to filter against | | No |
            | `url` | Okta organization URL (Must contain a trailing slash) | | Yes |
            | `appId` | Application ID of App Groups are assigned to | | Yes |
            | `extractLoginUsername` | Bool to determine if you should extract username from okta login | | No |
            | `profileKey` | Attribute field on Okta User Profile you would like to use as identity | `login` | No |
            | `groupLimit` | Integer to set the maximum number of groups to retrieve from OKTA per request. | `1000` | No |
            | `prune` | Prune Whether to prune groups that are no longer in OKTA | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with an Okta provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: okta-sync
          spec:
            providers:
              - name: okta
                okta:
                  credentialsSecret:
                    name: okta-api-token
                    namespace: group-sync-operator
                  url: "https://example.okta.com/"
                  appId: okta-sync-app-id
          ```

          #### Authenticating to Okta

          A secret must be created in the same namespace as the group-sync-operator pod. It must contain the following key:

            * `okta-api-token` - Okta API Token for interacting with Okta

          The secret can be created by executing the following command:

            ```shell
            oc create secret generic okta-api-token --from-literal=okta-api-token=<OKTA_API_TOKEN> -n group-sync-operator
            ```

          ### IBM Security Verify

          Groups defined in [IBM Security Verify](https://help.okta.com/en/prod/Content/Topics/users-groups-profiles/usgp-main.htm) (ISV) can be synchronized into OpenShift. Currently only the `userName` field from ISV will be synchronized. The developer docs for the ISV API can be found [here](https://docs.verify.ibm.com/verify/page/api-documentation).
          The following table describes the set of configuration options for the provider:

            | Name | Description | Defaults | Required |
            | ----- | ---------- | -------- | ----- |
            | `credentialsSecret` | Reference to a secret containing authentication details (see below) | `''`  | Yes |
            | `groups` | List of groups to synchronize (see below) | `nil`  | Yes |
            | `tenantUrl` | The ISV tenant URL, for example `https://my-isv.verify.ibm.com`) | `''`  | Yes |

          The following is an example of a minimal configuration that can be applied to integrate with an Okta provider:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ibmsecurityverify-sync
          spec:
              providers:
              - name: ibmsecurityverify
                ibmsecurityverify:
                  credentialsSecret:
                    name: isv-group-sync
                    namespace: group-sync-operator
                  tenantUrl: https://my-isv.verify.ibm.com
                  groups:
                    - name: 'application owners'
                      id: 645001V3V9
                    - name: developer
                      id: 645001V3VA
          ```

          #### Group Objects
          Each group object in the `groups` array must contain an `id` field. The group ID can be retrieved by pulling the group information from the ISV API. Optionally, the object may also contain a `name` which corresponds to the group's display name. When defined, the operator will confirm that the name defined in the YAML matches that received from the API when synchronization occurs; as the group IDs are not human-friendly, using the name can confirm the correct groups are configured. If the names do not match an error will be logged.

          #### Group Names
          The name of each groups created in OpenShift will match the group name in ISV. Any whitespace in the ISV group name will be replaced with a hyphen.

          #### Authenticating to IBM Security Verify

          A secret must be created in the same namespace as the group-sync-operator pod. It must contain the following keys:

          * `clientId` - The API client ID.
          * `clientSecret`- The API client secret.

          See the IBM Security Verify [API documentation](https://docs.verify.ibm.com/verify/docs/api-access) for setting up authentication.

          ### Support for Additional Metadata (Beta)

          Additional metadata based on Keycloak group are also added to the OpenShift groups as Annotations including:

          * Parent/child relationship between groups and their subgroups
          * Group attributes

          ## CA Certificates

          Several providers allow for certificates to be provided in either a _ConfigMap_ or _Secret_ to communicate securely to the target host through the use of a property called `ca`.

          The certificate can be added to a Secret called _keycloak-certs_ using the key `ca.crt` representing the certificate using the following command.

          ```
          oc create secret generic keycloak-certs --from-file=ca.crt=<file>
          ```

          An example of how the CA certificate can be added to the Keycloak provider is shown below:


          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                caSecret:
                  kind: Secret
                  name: keycloak-certs
                  namespace: group-sync-operator
                  key: ca.crt
                url: https://keycloak-keycloak-operator.apps.openshift.com
          ```

          Alteratively, a _ConfigMap_ can be used instead instead of a _Secret_. This is useful when using the [Certificate injection using Operators](https://docs.openshift.com/container-platform/latest/networking/configuring-a-custom-pki.html#certificate-injection-using-operators_configuring-a-custom-pki) feature.

          The following command can be used to create a _ConfigMap_ containing the certificate:

          ```
          oc create configmap keycloak-certs --from-file=ca.crt=<file>
          ````

          An example of how the CA certificate can be added to the Keycloak provider is shown below:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                caSecret:
                  kind: ConfigMap
                  name: keycloak-certs
                  namespace: group-sync-operator
                  key: ca.crt
                url: https://keycloak-keycloak-operator.apps.openshift.com
          ```

          ## Scheduled Execution

          A cron style expression can be specified for which a synchronization event will occur. The following specifies that a synchronization should occur nightly at 3AM

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            schedule: "0 3 * * *"
            providers:
            - ...
          ```

          If a schedule is not provided, synchronization will occur only when the object is reconciled by the platform.

          ## Metrics

          Prometheus compatible metrics are exposed by the Operator and can be integrated into OpenShift's default cluster monitoring. To enable OpenShift cluster monitoring, label the namespace the operator is deployed in with the label `openshift.io/cluster-monitoring="true"`.

          ```
          oc label namespace <namespace> openshift.io/cluster-monitoring="true"
          ```
        displayName: Group Sync Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - group
        - user
        - security
        - authorization
        links:
        - name: Group Sync Operator
          url: https://github.com/redhat-cop/group-sync-operator
        maintainers:
        - email: ablock@redhat.com
          name: Andrew Block
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:b8a174be91f3df75dff9656d4c72de28af51a731008e9e96db5eef3c7af58aac
        - quay.io/redhat-cop/group-sync-operator@sha256:43ef79efc2a77ee5edd4e3123f9dab156a30d3e00fd1fba9f89600b96f0f68d6
        version: 0.0.36
      entries:
      - name: group-sync-operator.v0.0.36
        version: 0.0.36
      - name: group-sync-operator.v0.0.35
        version: 0.0.35
      - name: group-sync-operator.v0.0.34
        version: 0.0.34
      - name: group-sync-operator.v0.0.33
        version: 0.0.33
      - name: group-sync-operator.v0.0.32
        version: 0.0.32
      - name: group-sync-operator.v0.0.31
        version: 0.0.31
      - name: group-sync-operator.v0.0.30
        version: 0.0.30
      - name: group-sync-operator.v0.0.29
        version: 0.0.29
      - name: group-sync-operator.v0.0.28
        version: 0.0.28
      - name: group-sync-operator.v0.0.27
        version: 0.0.27
      - name: group-sync-operator.v0.0.26
        version: 0.0.26
      - name: group-sync-operator.v0.0.25
        version: 0.0.25
      - name: group-sync-operator.v0.0.24
        version: 0.0.24
      - name: group-sync-operator.v0.0.23
        version: 0.0.23
      - name: group-sync-operator.v0.0.22
        version: 0.0.22
      - name: group-sync-operator.v0.0.21
        version: 0.0.21
      - name: group-sync-operator.v0.0.20
        version: 0.0.20
      - name: group-sync-operator.v0.0.19
        version: 0.0.19
      - name: group-sync-operator.v0.0.18
        version: 0.0.18
      - name: group-sync-operator.v0.0.17
        version: 0.0.17
      - name: group-sync-operator.v0.0.16
        version: 0.0.16
      - name: group-sync-operator.v0.0.15
        version: 0.0.15
      - name: group-sync-operator.v0.0.14
        version: 0.0.14
      - name: group-sync-operator.v0.0.13
        version: 0.0.13
      name: alpha
    defaultChannel: alpha
    packageName: group-sync-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubecost
      provider-url: https://www.kubecost.com
    name: kubecost-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubecost-operator.v2.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kubecost.com/v1alpha1",
                "kind": "CostAnalyzer",
                "metadata": {
                  "name": "costanalyzer-sample"
                },
                "spec": {
                  "clusterController": {
                    "image": {
                      "repository": "registry.connect.redhat.com/kubecost/kubecost-cluster-controller@sha256",
                      "tag": "1a6f11dcd7341ab3f24eba23ec06c7b75646a534a7ed960153601806425469d1"
                    }
                  },
                  "forecasting": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-modeling@sha256:8f59f8fd8696964c1b5322ca79fae78befe97323fa56c2b509bdcc67ee4c0edd"
                  },
                  "global": {
                    "platforms": {
                      "openshift": {
                        "createMonitoringClusterRoleBinding": true,
                        "createMonitoringResourceReaderRoleBinding": true,
                        "enabled": true,
                        "scc": {
                          "clusterController": true,
                          "networkCosts": true,
                          "nodeExporter": true
                        }
                      }
                    },
                    "prometheus": {
                      "enabled": false,
                      "fqdn": "https://prometheus-k8s.openshift-monitoring.svc.cluster.local:9091",
                      "kubeRBACProxy": true
                    }
                  },
                  "grafana": {
                    "image": {
                      "repository": "registry.redhat.io/rhel9/grafana@sha256",
                      "tag": "9f71de028c18034e07a8d15dac13ef68afb91cb899b9e8e7d39703547dd0eac1"
                    },
                    "sidecar": {
                      "image": {
                        "repository": "quay.io/kiwigrid/k8s-sidecar@sha256",
                        "tag": "d3749777d5bedf62549b452a7e2ccb205da53588dc7f04927efc75d0da5b6d23"
                      }
                    }
                  },
                  "kubecostAggregator": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef"
                  },
                  "kubecostFrontend": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-frontend@sha256:f9b2196dc6ac017b8074f8fe40f1a58e7a2630282c5b7a019ca9f3dbdb670dc8"
                  },
                  "kubecostModel": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef"
                  },
                  "networkCosts": {
                    "image": {
                      "repository": "registry.connect.redhat.com/kubecost/kubecost-network-costs@sha256",
                      "tag": "c444d475be4f1aa36680bd84358e1e07303dcfe5d13a95544ad9e59f80d4a5f0"
                    }
                  },
                  "prometheusRule": {
                    "enabled": true
                  },
                  "serviceMonitor": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          createdAt: "2025-08-08T19:34:34Z"
          description: Deploys Kubecost 2.8.2 via the Helm chart.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Contact Kubecost for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Kubecost.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The CostAnalyzer API describes a deployment of Kubecost and
              its dependencies.
            displayName: CostAnalyzer
            kind: CostAnalyzer
            name: costanalyzers.charts.kubecost.com
            version: v1alpha1
        description: |2-

          IMPORTANT - Before installation, you must run the following command to allow monitoring of the Kubecost namespace by the local OpenShift Promtheus. If you install in a namespace other than "Kubecost", adjust accordingly.

          `oc label namespace kubecost openshift.io/cluster-monitoring=true`

          Kubecost provides real-time cost visibility and insights for teams using Kubernetes, helping you continuously reduce your cloud costs.

          This OpenShift Operator deploys Kubecost from the official Helm chart.

          Kubecost allows you to:

          * See allocated spend across all native Kubernetes concepts and provide your teams with transparent, accurate cost data reconciled with your actual cloud bill.
          * Join in-cluster costs like CPU and memory with out-of-cluster spend from other cloud services.
          * Get complete cost visibility across OpenShift, On-Prem, AWS, GCP, Azure, Oracle, AliCloud + more.
          * Gain insights you can use to save 30-50%+ on your infrastructure spend, without exposing your private information — your data never gets shared externally, even with us.
          * Have real-time alerts and recurring reports, empowering teams to take control of their Kubernetes-enabled infrastructure and stay within budgeted limits.

          Visit the Kubecost website for more details including features, pricing, documentation, and more at www.kubecost.com.
        displayName: Kubecost
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kubernetes
        - cost
        - kubernetes costs
        - kubernetes optimization
        - k8s
        - cost monitoring
        - cloud costs
        - k8s cost
        links:
        - name: Kubecost
          url: https://www.kubecost.com
        maintainers:
        - email: support@kubecost.com
          name: Kubecost Support
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Kubecost
          url: https://www.kubecost.com
        relatedImages:
        - registry.connect.redhat.com/kubecost/kubecost-network-costs@sha256:c444d475be4f1aa36680bd84358e1e07303dcfe5d13a95544ad9e59f80d4a5f0
        - registry.redhat.io/rhel9/grafana@sha256:9f71de028c18034e07a8d15dac13ef68afb91cb899b9e8e7d39703547dd0eac1
        - quay.io/kiwigrid/k8s-sidecar@sha256:d3749777d5bedf62549b452a7e2ccb205da53588dc7f04927efc75d0da5b6d23
        - registry.connect.redhat.com/kubecost/kubecost-operator-manager@sha256:582223a66a77f35d1f7ca33460032cfca1581a8ff30f75ead289d8cd4cf62bb9
        - registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef
        - registry.connect.redhat.com/kubecost/kubecost-frontend@sha256:f9b2196dc6ac017b8074f8fe40f1a58e7a2630282c5b7a019ca9f3dbdb670dc8
        - registry.connect.redhat.com/kubecost/kubecost-modeling@sha256:8f59f8fd8696964c1b5322ca79fae78befe97323fa56c2b509bdcc67ee4c0edd
        - registry.connect.redhat.com/kubecost/kubecost-cluster-controller@sha256:1a6f11dcd7341ab3f24eba23ec06c7b75646a534a7ed960153601806425469d1
        version: 2.8.2
      entries:
      - name: kubecost-operator.v2.8.2
        version: 2.8.2
      - name: kubecost-operator.v2.8.1
        version: 2.8.1
      - name: kubecost-operator.v2.8.0
        version: 2.8.0
      - name: kubecost-operator.v2.7.2
        version: 2.7.2
      - name: kubecost-operator.v2.7.1
        version: 2.7.1
      - name: kubecost-operator.v2.5.4
        version: 2.5.4
      name: alpha
    defaultChannel: alpha
    packageName: kubecost-operator
    provider:
      name: Kubecost
      url: https://www.kubecost.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-access-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-access-operator.v25.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyAccess",
                "metadata": {
                  "name": "ivia-sample"
                },
                "spec": {
                  "image": "icr.io/ivia/ivia-wrp@sha256:c2dd5ec2b362136896e47ccbd0f3ee033af945fc2794ddaf075da603e4d4d7b2",
                  "instance": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isva/verify-access-operator@sha256:ca1d7be14b4dd313d4675f18541b2cc670ea0a0adf3ec87e347e5e20a86bc39c
          createdAt: "2025-05-01T01:27:17Z"
          description: The IBM Verify Identity Access Operator manages the lifecycle
            of IBM Verify Identity Access worker containers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM-Security/verify-access-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyAccess is the Schema for the ibmsecurityverifyaccesses
              API.
            displayName: IBMSecurity Verify Access
            kind: IBMSecurityVerifyAccess
            name: ibmsecurityverifyaccesses.ibm.com
            version: v1
        description: |+
          In a world of highly fragmented access management environments, [IBM Verify Identity Access](https://www.ibm.com/au-en/products/verify-access) helps you simplify your users' access while more securely adopting web, mobile and cloud technologies. This solution helps you strike a balance between usability and security through the use of risk-based access, single sign-on, integrated access management control, identity federation and its mobile multi-factor authentication capability, IBM Verify. Take back control of your access management with IBM Verify Identity Access.

          The IBM Verify Identity Access operator provides lifecycle management of the lightweight containers which are used to protect an environment, namely:
          * [Web Reverse Proxy](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-web-reverse-proxy)
          * [Runtime](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-runtime)
          * [Distributed Session Cache](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-distributed-session-cache)

          The operator will manage the deployment of these lightweight IBM Verify Identity Access worker containers, and also control the rolling restart of these containers when a configuration snapshot is updated.

          See the project [Readme](https://github.com/IBM-Security/verify-access-operator/blob/master/README.md) for further information and details.

        displayName: IBM Verify Identity Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Identity Access Product Information
          url: https://www.ibm.com/au-en/products/verify-access
        - name: Verify Identity Access Documentation
          url: https://www.ibm.com/docs/en/sva
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Identity Access Development Team
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - icr.io/isva/verify-access-operator@sha256:ca1d7be14b4dd313d4675f18541b2cc670ea0a0adf3ec87e347e5e20a86bc39c
        - icr.io/ivia/ivia-wrp@sha256:c2dd5ec2b362136896e47ccbd0f3ee033af945fc2794ddaf075da603e4d4d7b2
        version: 25.4.0
      entries:
      - name: ibm-security-verify-access-operator.v25.4.0
        version: 25.4.0
      - name: ibm-security-verify-access-operator.v24.12.0
        version: 24.12.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-access-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Redis
      provider-url: ""
    name: redis-enterprise-operator-cert
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-enterprise-operator.v6.2.10-45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
        - registry.connect.redhat.com/redislabs/services-manager@sha256:8b46f2aacfb3e1b243c0b64975accc55fd80b4de9c9ac4769c3c67a5676addc5
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:ffdf10a1b08115fd4de06f1cd751d36644a9bc9a8c11034cd432e9b66a5d2f73
        version: 6.2.10-45
      entries:
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: 6.2.10
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.22.0
    - currentCSV: redis-enterprise-operator.v7.22.0-11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:088e4b57c6e1ca1680b359078e5eb1bba67343cdbd9a539af26a0ea23a830593
        - registry.connect.redhat.com/redislabs/services-manager@sha256:3b96cdc65c3e42fd22eb165950be96b907725903ac74fe4f82cfa0073e28e95a
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:d56ba97e95e799210e76615095409143a5dddd5c5a1e1be1d551d7e3f9e4d3ef
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:0cfb07f774db131b5b74e5aa91291579c916c4d4cd1a693d7aaed554f0177a68
        version: 7.22.0-11.3
      entries:
      - name: redis-enterprise-operator.v7.22.0-11.3
        version: 7.22.0-11.3
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.22.0-11.3
    - currentCSV: redis-enterprise-operator.v7.4.6-7.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:473b5661049a6c8830ba113504975726154c7b172376ba3b63d891674d59c78f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5f6f32a63f4dd8be85b9d469f0fedd26d6aa76e32b7735bfa7ab7203cb272cdb
        - registry.connect.redhat.com/redislabs/services-manager@sha256:ba20ab644a473bc742af671171d5313537292d88942b16336d03a6386c061d8a
        version: 7.4.6-7.2
      entries:
      - name: redis-enterprise-operator.v7.4.6-7.2
        version: 7.4.6-7.2
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      name: 7.4.6
    - currentCSV: redis-enterprise-operator.v7.8.2-6.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:36c7a36dabb23a4c003e112085cda2873296bfb22e5d11b470d7c1aba55b7f0f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:c6e1f2109a0dac76b51c25c1cbf6030b4894b21b98147e2ad1db2dd8e7693576
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:acf817c90b5f03bc2796fbc305cf43ce5c0644dcbe1b685752090b3f600a0f7b
        version: 7.8.2-6.8
      entries:
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.2
    - currentCSV: redis-enterprise-operator.v7.8.4-9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:803be04534e59cdc7b8898b3bb336676aad1cb77cf266051e226be82679fcf6a
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:4ba0c414a74268bed9c6757e447b340db7c912afa864cfb096ddf88f11feeaf9
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:66ec124da16cc9d0abd66687e088be0b358e20cfd48f5774d34bf5874e6a4e65
        version: 7.8.4-9.0
      entries:
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.4
    - currentCSV: redis-enterprise-operator.v7.8.6-5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:1fff01daece34aeb7befcb0760d5112c8c7d261f2b86388ece416ac94076a573
        - registry.connect.redhat.com/redislabs/services-manager@sha256:88d483f5e37b5a901d1ae3d965304edc000c172d2e5e307297d5b7bcd77ae73b
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5125e83d3c93b224d1983c599994e43899a0ba0c6f2fc4eed31476f755b5fc0b
        version: 7.8.6-5.1
      entries:
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.6
    - currentCSV: redis-enterprise-operator.v6.2.10-41
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/outbound",
                      "versionTag": "6.2.10-41"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager",
                      "versionTag": "6.2.10-41"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise",
                    "versionTag": "6.2.14-9.rhel8-openshift"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/outbound@sha256:3c251754ae8650157465a56ef3a7b41d5a0ad4e9db09ef042e0014076f25e8a2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/outbound@sha256:3c251754ae8650157465a56ef3a7b41d5a0ad4e9db09ef042e0014076f25e8a2
        version: 6.2.10-41
      entries:
      - name: redis-enterprise-operator.v6.2.10-41
        version: 6.2.10-41
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: preview
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: production
    defaultChannel: production
    packageName: redis-enterprise-operator-cert
    provider:
      name: Redis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: candidate
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: fast
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: stable
    defaultChannel: stable
    packageName: stackgres
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-iam-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-iam-controller.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Role",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Group",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "OpenIDConnectProvider",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Policy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/iam-controller:1.5.2
          createdAt: "2025-10-03T22:22:57Z"
          description: AWS IAM controller is a service controller for managing IAM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Group represents the state of an AWS iam Group resource.
            displayName: Group
            kind: Group
            name: groups.iam.services.k8s.aws
            version: v1alpha1
          - description: InstanceProfile represents the state of an AWS iam InstanceProfile
              resource.
            displayName: InstanceProfile
            kind: InstanceProfile
            name: instanceprofiles.iam.services.k8s.aws
            version: v1alpha1
          - description: OpenIDConnectProvider represents the state of an AWS iam
              OpenIDConnectProvider resource.
            displayName: OpenIDConnectProvider
            kind: OpenIDConnectProvider
            name: openidconnectproviders.iam.services.k8s.aws
            version: v1alpha1
          - description: Policy represents the state of an AWS iam Policy resource.
            displayName: Policy
            kind: Policy
            name: policies.iam.services.k8s.aws
            version: v1alpha1
          - description: Role represents the state of an AWS iam Role resource.
            displayName: Role
            kind: Role
            name: roles.iam.services.k8s.aws
            version: v1alpha1
          - description: ServiceLinkedRole represents the state of an AWS iam ServiceLinkedRole
              resource.
            displayName: ServiceLinkedRole
            kind: ServiceLinkedRole
            name: servicelinkedroles.iam.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS iam User resource.
            displayName: User
            kind: User
            name: users.iam.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Identity and Access Management (IAM) resources in AWS from within your Kubernetes cluster.

          **About Amazon IAM**

          AWS Identity and Access Management (IAM) is a web service that helps you securely control access to AWS resources. You use IAM to control who is authenticated (signed in) and authorized (has permissions) to use resources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon IAM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iam
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon IAM Developer Resources
          url: https://aws.amazon.com/iam/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: iam maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/iam-controller:1.5.2
        version: 1.5.2
      entries:
      - name: ack-iam-controller.v1.5.2
        version: 1.5.2
      - name: ack-iam-controller.v1.5.1
        version: 1.5.1
      - name: ack-iam-controller.v1.5.0
        version: 1.5.0
      - name: ack-iam-controller.v1.4.4
        version: 1.4.4
      - name: ack-iam-controller.v1.4.3
        version: 1.4.3
      - name: ack-iam-controller.v1.4.2
        version: 1.4.2
      - name: ack-iam-controller.v1.4.1
        version: 1.4.1
      - name: ack-iam-controller.v1.4.0
        version: 1.4.0
      - name: ack-iam-controller.v1.3.22
        version: 1.3.22
      - name: ack-iam-controller.v1.3.21
        version: 1.3.21
      - name: ack-iam-controller.v1.3.20
        version: 1.3.20
      - name: ack-iam-controller.v1.3.19
        version: 1.3.19
      - name: ack-iam-controller.v1.3.18
        version: 1.3.18
      - name: ack-iam-controller.v1.3.17
        version: 1.3.17
      - name: ack-iam-controller.v1.3.15
        version: 1.3.15
      - name: ack-iam-controller.v1.3.14
        version: 1.3.14
      - name: ack-iam-controller.v1.3.13
        version: 1.3.13
      - name: ack-iam-controller.v1.3.12
        version: 1.3.12
      - name: ack-iam-controller.v1.3.11
        version: 1.3.11
      - name: ack-iam-controller.v1.3.10
        version: 1.3.10
      - name: ack-iam-controller.v1.3.9
        version: 1.3.9
      - name: ack-iam-controller.v1.3.8
        version: 1.3.8
      - name: ack-iam-controller.v1.3.6
        version: 1.3.6
      - name: ack-iam-controller.v1.3.2
        version: 1.3.2
      - name: ack-iam-controller.v1.3.1
        version: 1.3.1
      - name: ack-iam-controller.v1.3.0
        version: 1.3.0
      - name: ack-iam-controller.v1.2.6
        version: 1.2.6
      - name: ack-iam-controller.v1.2.5
        version: 1.2.5
      - name: ack-iam-controller.v1.2.3
        version: 1.2.3
      - name: ack-iam-controller.v1.2.2
        version: 1.2.2
      - name: ack-iam-controller.v1.2.1
        version: 1.2.1
      - name: ack-iam-controller.v1.2.0
        version: 1.2.0
      - name: ack-iam-controller.v1.1.3
        version: 1.1.3
      - name: ack-iam-controller.v1.1.2
        version: 1.1.2
      - name: ack-iam-controller.v1.1.1
        version: 1.1.1
      - name: ack-iam-controller.v1.1.0
        version: 1.1.0
      - name: ack-iam-controller.v1.0.0
        version: 1.0.0
      - name: ack-iam-controller.v0.1.2
        version: 0.1.2
      - name: ack-iam-controller.v0.1.1
        version: 0.1.1
      - name: ack-iam-controller.v0.1.0
        version: 0.1.0
      - name: ack-iam-controller.v0.0.22
        version: 0.0.22
      - name: ack-iam-controller.v0.0.21
        version: 0.0.21
      - name: ack-iam-controller.v0.0.20
        version: 0.0.20
      - name: ack-iam-controller.v0.0.19
        version: 0.0.19
      - name: ack-iam-controller.v0.0.18
        version: 0.0.18
      - name: ack-iam-controller.v0.0.17
        version: 0.0.17
      - name: ack-iam-controller.v0.0.16
        version: 0.0.16
      - name: ack-iam-controller.v0.0.15
        version: 0.0.15
      - name: ack-iam-controller.v0.0.14
        version: 0.0.14
      - name: ack-iam-controller.v0.0.13
        version: 0.0.13
      - name: ack-iam-controller.v0.0.12
        version: 0.0.12
      - name: ack-iam-controller.v0.0.11
        version: 0.0.11
      - name: ack-iam-controller.v0.0.10
        version: 0.0.10
      - name: ack-iam-controller.v0.0.9
        version: 0.0.9
      - name: ack-iam-controller.v0.0.8
        version: 0.0.8
      - name: ack-iam-controller.v0.0.7
        version: 0.0.7
      - name: ack-iam-controller.v0.0.6
        version: 0.0.6
      - name: ack-iam-controller.v0.0.5
        version: 0.0.5
      - name: ack-iam-controller.v0.0.4
        version: 0.0.4
      - name: ack-iam-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-iam-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ODH
      provider-url: ""
    name: opendatahub-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opendatahub-operator.v2.35.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Managed",
                      "registriesNamespace": "odh-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "opendatahub",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "opendatahub"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v2.35.0
          createdAt: "2025-10-01T13:12:27Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",
            "workbenches.components.platform.opendatahub.io", "monitorings.services.platform.opendatahub.io",
            "servicemeshes.services.platform.opendatahub.io", "modelcontrollers.components.platform.opendatahub.io",
            "feastoperators.components.platform.opendatahub.io", "llamastackoperators.components.platform.opendatahub.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/opendatahub-io/opendatahub-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: ServiceMesh is the Schema for the servicemesh API
            displayName: Service Mesh
            kind: ServiceMesh
            name: servicemeshes.services.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack - Unified open-source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odh
        - notebooks
        - serving
        - training
        - pipelines
        - modelmesh
        - workbenches
        - dashboard
        - kserve
        - distributed workloads
        - trustyai
        - modelregistry
        - feast
        - featurestore
        - llamastack
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ODH
        relatedImages:
        - quay.io/opendatahub/opendatahub-operator:v2.35.0
        version: 2.35.0
      entries:
      - name: opendatahub-operator.v2.35.0
        version: 2.35.0
      - name: opendatahub-operator.v2.34.0
        version: 2.34.0
      - name: opendatahub-operator.v2.33.0
        version: 2.33.0
      - name: opendatahub-operator.v2.32.0
        version: 2.32.0
      - name: opendatahub-operator.v2.31.0
        version: 2.31.0
      - name: opendatahub-operator.v2.30.0
        version: 2.30.0
      - name: opendatahub-operator.v2.29.0
        version: 2.29.0
      - name: opendatahub-operator.v2.28.0
        version: 2.28.0
      - name: opendatahub-operator.v2.27.0
        version: 2.27.0
      - name: opendatahub-operator.v2.26.0
        version: 2.26.0
      - name: opendatahub-operator.v2.25.0
        version: 2.25.0
      - name: opendatahub-operator.v2.24.0
        version: 2.24.0
      - name: opendatahub-operator.v2.23.0
        version: 2.23.0
      - name: opendatahub-operator.v2.22.0
        version: 2.22.0
      - name: opendatahub-operator.v2.21.0
        version: 2.21.0
      - name: opendatahub-operator.v2.20.0
        version: 2.20.0
      - name: opendatahub-operator.v2.19.0
        version: 2.19.0
      - name: opendatahub-operator.v2.18.2
        version: 2.18.2
      - name: opendatahub-operator.v2.18.1
        version: 2.18.1
      - name: opendatahub-operator.v2.18.0
        version: 2.18.0
      - name: opendatahub-operator.v2.17.0
        version: 2.17.0
      - name: opendatahub-operator.v2.16.0
        version: 2.16.0
      - name: opendatahub-operator.v2.15.0
        version: 2.15.0
      - name: opendatahub-operator.v2.14.0
        version: 2.14.0
      - name: opendatahub-operator.v2.13.0
        version: 2.13.0
      - name: opendatahub-operator.v2.12.1
        version: 2.12.1
      - name: opendatahub-operator.v2.12.0
        version: 2.12.0
      - name: opendatahub-operator.v2.11.1
        version: 2.11.1
      - name: opendatahub-operator.v2.11.0
        version: 2.11.0
      - name: opendatahub-operator.v2.10.0
        version: 2.10.0
      name: fast
    - currentCSV: opendatahub-operator.v2.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "opendatahub",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "opendatahub"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v2.8.0
          createdAt: "2024-2-20T00:00:00Z"
          operatorframework.io/initialization-resource: |-
            {
            "apiVersion": "dscinitialization.opendatahub.io/v1",
            "kind": "DSCInitialization",
            "metadata": {
            "labels": {
             "app.kubernetes.io/created-by": "opendatahub-operator",
             "app.kubernetes.io/instance": "default",
             "app.kubernetes.io/managed-by": "kustomize",
             "app.kubernetes.io/name": "dscinitialization",
             "app.kubernetes.io/part-of": "opendatahub-operator"
            },
            "name": "default-dsci"
            },
            "spec": {
            "applicationsNamespace": "opendatahub",
            "monitoring": {
             "managementState": "Managed",
             "namespace": "opendatahub"
            },
            "serviceMesh": {
             "controlPlane": {
               "metricsCollection": "Istio",
               "name": "data-science-smcp",
               "namespace": "istio-system"
             },
             "managementState": "Managed"
            },
            "trustedCABundle": {
             "customCABundle": "",
             "managementState": "Managed"
            }
            }
            }
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: "The Open Data Hub is a machine-learning-as-a-service platform
          built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open
          Data Hub integrates multiple AI/ML open source components into one operator
          that can easily be downloaded and installed by OpenShift users.\n\nOpen
          Data Hub operator allows users to install and manage components of the Open
          Data Hub. Users can mix and match tools from each project to fulfill the
          needs of their use case. Each of the projects share some components, but
          can be mostly seen as an extension of each other to provide a complete solution
          for both novice and skilled enterprise users.\n    \n### Components\n\n*
          Open Data Hub Dashboard - A web dashboard that displays installed Open Data
          Hub components with easy access to component UIs and documentation\n* ODH
          Notebook Controller - Secure management of Jupyter Notebook in Kubernetes
          environments built on top of Kubeflow Notebook Controller with support for
          OAuth\n* Jupyter Notebooks - JupyterLab notebook that provide Python support
          for GPU workloads\n* Data Science Pipelines - Pipeline solution for end
          to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton\n*
          Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh,
          a general-purpose model serving management/routing layer\n* Distributed
          Workloads(Incubation) - Stack built to make managing distributed compute
          infrastructure in the cloud easy and intuitive for Data Scientists. This
          stack consists of three components \n                                      Codeflare
          , KubeRay and Kueue.\n* Kserve - Kserve is the Controller for for serving
          machine learning (ML) models on arbitrary frameworks"
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odh
        - notebooks
        - serving
        - training
        - pipelines
        - modelmesh
        - workbenches
        - dashboard
        - kserve
        - distributed-workloads
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: ODH
        relatedImages:
        - quay.io/opendatahub/opendatahub-operator:v2.8.0
        version: 2.8.1
      entries:
      - name: opendatahub-operator.v2.8.1
        version: 2.8.1
      name: odh-2.8.z
    - currentCSV: opendatahub-operator.v1.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kfdef.apps.kubeflow.org/v1",
                "kind": "KfDef",
                "metadata": {
                  "name": "opendatahub",
                  "namespace": "opendatahub"
                },
                "spec": {
                  "applications": [
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-common"
                        }
                      },
                      "name": "odh-common"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-dashboard"
                        }
                      },
                      "name": "odh-dashboard"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "data-science-pipelines-operator/"
                        }
                      },
                      "name": "data-science-pipelines-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-notebook-controller"
                        }
                      },
                      "name": "odh-notebook-controller"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "notebook-images"
                        }
                      },
                      "name": "notebook-images"
                    },
                    {
                      "kustomizeConfig": {
                        "parameters": [
                          {
                            "name": "monitoring-namespace",
                            "value": "opendatahub"
                          }
                        ],
                        "repoRef": {
                          "name": "manifests",
                          "path": "model-mesh"
                        }
                      },
                      "name": "model-mesh"
                    }
                  ],
                  "repos": [
                    {
                      "name": "manifests",
                      "uri": "file:///opt/manifests/odh-manifests.tar.gz"
                    }
                  ]
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhApplication",
                "metadata": {
                  "name": "example-odhapplication",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started,Model development,Model training,Notebook environments,Package management,Data Management"
                  }
                },
                "spec": {
                  "displayName": "ODHAPPLICATION DISPLAY NAME",
                  "provider": "PROVIDER_NAME",
                  "description": "Description of your ODHApplication",
                  "kfdefApplications": [
                    "<NAME OF THE ODH APPLICATION IN KFDEF>"
                  ],
                  "route": "<APPLICATION_OCP_ROUTE_OBJECT_NAME>",
                  "img": "<svg xmlns=\"http://www.w3.org/2000/svg\" role=\"img\" viewBox=\"-1.13 -1.13 58.50 48.94\"><title>Open Data Hub logo</title><g><g><g fill=\"#4d5258\" stroke-width=\".006\"><path fill=\"#bbb\" d=\"M112.30231 22.797083c-.2034 0-.35737.143517-.35737.337407 0 .192939.15397.339307.35737.339307.20339 0 .35736-.147318.35736-.339307 0-.19294-.15397-.337407-.35736-.337407zm.002.131161c.1112 0 .20149.08934.20149.206246 0 .116904-.0903.208146-.20149.208146-.1112 0-.2053-.09124-.2053-.208146 0-.116904.0931-.206246.2053-.206246zM113.04141 22.801836h-.28228v.665308h.15017v-.190088h.13211c.16918 0 .26612-.08934.26612-.244263 0-.147319-.0969-.230957-.26612-.230957zm-.008.349762h-.12451v-.225255h.12451c.0836 0 .13116.03707.13116.111202 0 .07603-.0475.114053-.13116.114053zM113.40328 22.801836v.665308h.51703v-.124508h-.36687v-.147318h.32125V23.07081h-.32125v-.144467h.35642v-.124507zM114.03814 22.801836v.665308h.14352v-.426748l.3222.426748h.13686v-.665308h-.14257v.427698L114.175 22.801836z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/><path d=\"M115.29685 22.801836h-.28228v.665308h.27658c.21004 0 .35546-.136863.35546-.332654 0-.196741-.14351-.332654-.34976-.332654zm.004.538899h-.13591v-.412491h.12926c.11405 0 .19959.08459.19959.207196 0 .121657-.0817.205295-.19294.205295zM116.21878 23.467144h.16062l-.28418-.665308h-.15492l-.28894.665308h.15493l.0523-.12926h.30794zm-.3108-.249966l.1055-.25852.10455.25852zM116.31525 22.801836v.127359h.20244v.537949h.15017v-.537949h.2034v-.127359zM117.37527 23.467144h.16063l-.28419-.665308h-.15492l-.28893.665308h.15492l.0523-.12926h.30795zm-.31079-.249966l.1055-.25852.10454.25852zM117.83433 22.801836v.665308h.15017v-.262322h.30129v.262322h.15017v-.665308h-.15017v.277528h-.30129v-.277528zM118.72025 23.188665v-.386829h-.15017v.386829c0 .175831.11785.285132.30224.285132.18343 0 .29844-.109301.29844-.285132v-.386829h-.15017v.386829c0 .09599-.0589.153971-.14827.153971-.0893 0-.15207-.05798-.15207-.153971zM119.61059 22.801836h-.30604v.665308h.31649c.15492 0 .25092-.07033.25092-.184385 0-.08744-.0532-.152071-.13876-.168228.0713-.01806.115-.07413.115-.146368 0-.102648-.0912-.166327-.23761-.166327zm-.008.268024h-.14826v-.145418h.14826c.0561 0 .0913.02566.0913.07128 0 .04657-.0352.07413-.0913.07413zm0 .274677h-.14826v-.158723h.14826c.0713 0 .11406.02946.11406.07794 0 .05132-.0428.08079-.11406.08079z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g><g fill=\"#58595b\" stroke-width=\".01\"><path d=\"M112.19192 24.15027h.0381l-.13717-.306759h-.0364l-.1376.306759h.0372l.0342-.07844h.16696zm-.18712-.110871l.0688-.157763.0697.157763zM112.27858 23.843511v.306759h.0351v-.306759zM112.64205 23.843511h-.11525v.306759h.0351v-.09729h.0802c.0758 0 .11964-.039.11964-.10649 0-.0653-.0438-.102983-.11964-.102983zm-.001.177044h-.0789V23.87594h.0789c.0552 0 .0872.02454.0872.07143 0 .0482-.032.07318-.0872.07318zM112.81775 23.825105v.325165h.0342v-.325165zM113.1137 24.15027l-.00044-.152503c-.00044-.04908-.032-.07932-.089-.07932-.0394 0-.0644.01139-.0942.03111l.0145.0241c.0241-.01709.0478-.02673.0732-.02673.0408 0 .0618.0206.0618.05434v.01446h-.0758c-.0543.000439-.0863.02586-.0863.06617 0 .03944.0307.06968.0784.06968.0368 0 .0662-.01227.0846-.03725v.03593zm-.11131-.02717c-.0324 0-.053-.01709-.053-.04251 0-.02585.0202-.03769.0587-.03769h.0719v.02805c-.009.03287-.0386.05215-.0776.05215zM113.29915 24.108201c-.0127.0074-.025.01139-.0368.01183-.0197 0-.0307-.01139-.0307-.03944v-.127086h.0732v-.02673h-.0732v-.06048h-.0337v.06048h-.0333v.02673h.0333v.132783c0 .04558.025.06573.0587.0653.0184 0 .0359-.0061.0521-.01709zM113.44228 23.854466c.0118 0 .0245.0044.0351.01183l.014-.02629c-.014-.01052-.032-.01622-.0495-.01622-.0386-.000438-.0741.02717-.0741.07713v.02586h-.0298v.02673h.0298v.196764h.0337v-.196764h.0662v-.02673h-.0662v-.03024c0-.02717.0184-.04251.0408-.04207zM113.59236 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM113.80323 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM114.22747 23.918448c-.0443.000438-.0767.01972-.0911.06004-.01-.03813-.0394-.06004-.0828-.06004-.0416.000438-.0727.01753-.0881.05346v-.05215h-.0337v.230508h.0337v-.113501c0-.04996.0285-.08458.0758-.08589.0403 0 .064.02454.064.06617v.133221h.0346v-.113501c0-.04996.0276-.08458.0754-.08589.0399 0 .064.02454.064.06617v.133221h.0342V24.00828c0-.05609-.032-.08984-.0859-.08984zM114.63474 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0487.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11088-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM114.89934 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM115.02844 23.919762l.0863.230508h.0355l.0675-.188876.0679.188876h.0355l.0855-.230508h-.0355l-.0675.195888-.0679-.195888h-.0359l-.067.195888-.0684-.195888zM115.53009 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM115.73075 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM115.95034 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11482.116569.0386 0 .071-.01402.0929-.03813l-.0188-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18843c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM116.29474 23.825105v.144615c-.0171-.03287-.0491-.05171-.0903-.05171-.0653 0-.11087.04777-.11087.116569 0 .0688.0456.117445.11131.117445.0412 0 .0727-.01884.0898-.05127v.04952h.0342v-.325165zm-.0833.297557c-.0491 0-.0837-.03637-.0837-.08721 0-.05083.0346-.08765.0837-.08765.0486 0 .0833.03681.0833.08765 0 .05083-.0346.08721-.0833.08721zM116.65984 23.918448c-.0403 0-.0719.01884-.089.05083v-.144177h-.0342v.325165h.0342v-.04952c.0171.03199.0486.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM116.89657 24.183138l.10605-.263376h-.0351l-.0732.191944-.0767-.191944h-.0359l.0955.228756-.01.02191c-.009.02191-.0241.03331-.0421.03331-.0114 0-.0219-.0035-.0324-.01095l-.0145.02717c.014.01139.0294.01665.0486.01665.0307 0 .0548-.01841.0692-.05346zM117.29947 23.842634c-.0898 0-.16083.06793-.16083.154257 0 .08633.071.154694.16083.154694.0894 0 .16039-.06836.16039-.154694 0-.08633-.071-.154257-.16039-.154257zm0 .03243c.0684 0 .12446.0539.12446.121828 0 .06792-.0561.122703-.12446.122703-.0684 0-.12533-.05478-.12533-.122703 0-.06793.057-.121828.12533-.121828zM117.64414 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0486.05083.0898.05083.0653 0 .10999-.04777.10999-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM117.9061 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM118.19616 23.918448c-.0421.000438-.0732.01709-.0894.05215v-.05084h-.0342v.230508h.0342v-.12358c.004-.0447.032-.07494.0776-.07537.0408 0 .0653.02454.0653.06573v.133221h.0342v-.141986c0-.05609-.0324-.08984-.0876-.08984zM118.65836 23.901795l.0149-.03111c-.0254-.01709-.0618-.02848-.0973-.02848-.0635 0-.10605.03068-.10605.07888 0 .109995.17266.06661.17222.148559 0 .0298-.0276.04777-.0719.04777-.0346 0-.0736-.01578-.1021-.04295l-.0158.03024c.0289.02761.0727.04689.11745.04689.0657 0 .11087-.03287.11087-.0837.00044-.11131-.17223-.06968-.17223-.149436 0-.02629.0259-.04207.0653-.04207.0245 0 .057.0075.0846.02542zM118.84168 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .11701-.04777.11701-.117007 0-.06836-.0487-.11613-.11701-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM119.18446 23.919762v.113063c0 .04996-.0281.08502-.0758.08589-.0399 0-.064-.02454-.064-.06573v-.133222h-.0342v.141986c0 .05565.0324.08984.0863.08984.0416-.000438.0723-.01753.0876-.05346v.05215h.0342v-.230508zM119.33586 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM119.62557 23.976294l.0188-.02366c-.021-.02147-.0517-.03418-.089-.03418-.0666 0-.11438.0482-.11438.116568 0 .0688.0473.116569.11438.116569.0399 0 .0719-.0149.0925-.03988l-.0197-.02016c-.0162.01928-.0412.0298-.0714.0298-.0478 0-.082-.03506-.082-.08633 0-.05083.0342-.08589.082-.08589.028 0 .0513.01008.0688.02717zM119.78997 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g></g><g><g><path fill=\"#a586f3\" d=\"m 7808.8125,2097.1875 c -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 v 329.875 c -0.8781,10.2195 7.1804,19 17.4375,19 10.2571,0 19.0227,-8.7805 18.1446,-19 v -77.7324 l 0.4179,-6.1426 c 0,-55.1597 44.2153,-99.5625 99.375,-99.5625 55.1596,0 99.5625,44.4028 99.5625,99.5625 0.01,0.1837 0.015,0.3674 0.027,0.5508 -0.026,0.4409 -0.035,0.8826 -0.027,1.3242 v 110 c 0,9.6822 7.849,17.5312 17.5312,17.5312 9.6823,0 17.5313,-7.849 17.5313,-17.5312 v -110 c 0,-0.3245 -0.012,-0.6488 -0.031,-0.9727 0.018,-0.3004 0.029,-0.6013 0.031,-0.9023 1e-4,-74.075 -60.55,-134.4374 -134.625,-134.4375 -39.9177,1e-4 -75.8465,17.5354 -100.5,45.293 V 2115 c 0.1749,-9.8356 -7.7881,-17.8833 -17.625,-17.8125 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><path fill=\"#5ed3ee\" d=\"m 7765.8027,2098.123 c -0.038,5e-4 -0.077,0 -0.1152,0 -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 V 2361 c 0,41.9333 -26.0162,79.2671 -65.4375,93.5625 -39.4213,14.2954 -83.5594,2.3741 -110.4375,-29.8125 -26.878,-32.1865 -30.6579,-77.5719 -9.5625,-113.8125 21.0955,-36.2406 62.4262,-55.2897 103.6875,-47.8125 9.5173,1.7397 18.6463,-4.5547 20.4043,-14.0686 1.758,-9.514 -4.5187,-18.6551 -14.0293,-20.4314 -55.6354,-10.082 -111.8054,15.8219 -140.25,64.6875 -28.4446,48.8656 -23.3039,110.5381 12.9375,153.9375 36.2415,43.3993 95.908,59.4005 149.0625,40.125 44.7667,-16.2337 77.0286,-54.658 86.0977,-100.1562 0,0 1.9812,-4.7686 2.5898,-9.1563 0.6086,-4.3877 0,-17.0625 0,-17.0625 v -245.0625 c 0.1769,-9.865 -7.8313,-17.9263 -17.6973,-17.8145 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><g fill=\"none\" stroke-width=\".365\" transform=\"translate(-449.265 -83.342) scale(5.91262)\"><circle cx=\"79.677\" cy=\"16.823\" r=\".645\" stroke=\"#c6ea70\"/><path stroke=\"#dff3b1\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 80.808656,18.234303 a 1.8088735,1.8088735 0 0 1 -2.309183,-0.03848 1.8088735,1.8088735 0 0 1 -0.389241,-2.276466 1.8088735,1.8088735 0 0 1 2.165211,-0.803532\"/><path stroke=\"#f2fadf\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 82.216272,17.248644 a 2.455857,2.4425189 0 0 1 -2.468226,2.016391 2.455857,2.4425189 0 0 1 -2.383743,-2.114547 2.455857,2.4425189 0 0 1 1.733108,-2.668494\"/></g></g></g></g></svg>",
                  "category": "Open Data Hub",
                  "support": "open data hub",
                  "docsLink": "http://REPLACE.THIS.URL",
                  "quickStart": "example-odhquickstart",
                  "getStartedLink": "http://REPLACE.THIS.URL",
                  "getStartedMarkDown": "# Header for Getting Started MarkDown section\nHere is where you add information about your ODHApplication in markdown format\n\n1. List item\n1. List item\n1. List item"
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhDocument",
                "metadata": {
                  "name": "example-odhdocument",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHDocument",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "type": "how-to",
                  "description": "This is an example of an ODHDocument custom resource",
                  "url": "http://REPLACE.THIS.URL",
                  "durationMinutes": 5
                }
              },
              {
                "apiVersion": "console.openshift.io/v1",
                "kind": "OdhQuickStart",
                "metadata": {
                  "name": "example-odhquickstart",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHQuickstart Name",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "durationMinutes": 5,
                  "description": "Description of an example quickstart",
                  "introduction": "### Introduction Header\nThis is the body of the introduction for an ODHQuickstart",
                  "tasks": [
                    {
                      "title": "Title of the first task",
                      "description": "### Description header for the first task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the first task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    },
                    {
                      "title": "Title of the second task",
                      "description": "### Description header for the second task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the second task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    }
                  ],
                  "conclusion": "Message for the conclusion of the ODHQuickstart",
                  "nextQuickStart": null
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v1.11.0
          createdAt: "2023-10-31T00:00:00Z"
          description: Open Data Hub Operator for deployment and management of Open
            Data Hub
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
          support: Open Data Hub
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AcceleratorProfile is the Schema for the acceleratorprofiles
            displayName: AcceleratorProfile
            kind: AcceleratorProfile
            name: acceleratorprofiles.dashboard.opendatahub.io
            version: v1alpha
          - description: KfDef is the Schema for the kfdefs API
            displayName: Kf Def
            kind: KfDef
            name: kfdefs.kfdef.apps.kubeflow.org
            version: v1
          - description: OdhApplication is the Schema for the odhapplications
            displayName: Open Data Hub Dashboard Application
            kind: OdhApplication
            name: odhapplications.dashboard.opendatahub.io
            version: v1
          - description: OdhDashboardConfig is the Schema for the odhdashboardconfigs
            displayName: Open Data Hub Dashboard Configuration
            kind: OdhDashboardConfig
            name: odhdashboardconfigs.opendatahub.io
            version: v1alpha
          - description: OdhDocument is the Schema for the odhdocuments
            displayName: Open Data Hub Dashboard Documentation
            kind: OdhDocument
            name: odhdocuments.dashboard.opendatahub.io
            version: v1
          - description: Extension for guiding user through various workflows in the
              Open Data Hub dashboard.
            displayName: Open Data Hub Dashboard QuickStart
            kind: OdhQuickStart
            name: odhquickstarts.console.openshift.io
            version: v1
        description: |
          The Open Data Hub is a machine-learning-as-a-service platform built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open Data Hub integrates multiple AI/ML open source components into one operator that can easily be downloaded and installed by OpenShift users.

          Open Data Hub operator allows users to install and manage components of the Open Data Hub. Users can mix and match tools from each project to fulfill the needs of their use case. Each of the
          projects share some components, but can be mostly seen as an extension of each other to provide a complete solution for both novice and skilled enterprise users.

          ### Components
          * Open Data Hub Dashboard - A web dashboard that displays installed Open Data Hub components with easy access to component UIs and documentation
          * ODH Notebook Controller - Secure management of Jupyter Notebook in Kubernetes environments built on top of Kubeflow Notebook Controller with support for OAuth
          * Jupyter Notebooks - JupyterLab notebook that provide Python support for GPU workloads
          * Data Science Pipelines - Pipeline solution for end to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton
          * Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh, a general-purpose model serving management/routing layer.
          * TrustyAI - TrustyAI is an open source AI Explainability (XAI) Toolkit

          To install one or multiple of these components use the default KfDef provided with the operator.
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Open Data Hub
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/opendatahub/opendatahub-operator:v1.11.0
        version: 1.11.0
      entries:
      - name: opendatahub-operator.v1.11.0
        version: 1.11.0
      - name: opendatahub-operator.v1.10.1
        version: 1.10.1
      - name: opendatahub-operator.v1.10.0
        version: 1.10.0
      - name: opendatahub-operator.v1.9.0
        version: 1.9.0
      - name: opendatahub-operator.v1.8.0
        version: 1.8.0
      - name: opendatahub-operator.v1.7.0
        version: 1.7.0
      - name: opendatahub-operator.v1.6.0
        version: 1.6.0
      - name: opendatahub-operator.v1.5.0
        version: 1.5.0
      - name: opendatahub-operator.v1.4.2
        version: 1.4.2
      - name: opendatahub-operator.v1.4.1
        version: 1.4.1
      - name: opendatahub-operator.v1.4.0
        version: 1.4.0
      - name: opendatahub-operator.v1.3.0
        version: 1.3.0
      - name: opendatahub-operator.v1.2.0
        version: 1.2.0
      - name: opendatahub-operator.v1.1.2
        version: 1.1.2
      - name: opendatahub-operator.v1.1.1
        version: 1.1.1
      name: rolling
    - currentCSV: opendatahub-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kfdef.apps.kubeflow.org/v1",
                "kind": "KfDef",
                "metadata": {
                  "name": "opendatahub",
                  "namespace": "opendatahub"
                },
                "spec": {
                  "applications": [
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-common"
                        }
                      },
                      "name": "odh-common"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-dashboard"
                        }
                      },
                      "name": "odh-dashboard"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "prometheus/cluster"
                        }
                      },
                      "name": "prometheus-cluster"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "prometheus/operator"
                        }
                      },
                      "name": "prometheus-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "grafana/cluster"
                        }
                      },
                      "name": "grafana-cluster"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "grafana/grafana"
                        }
                      },
                      "name": "grafana-instance"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-notebook-controller"
                        }
                      },
                      "name": "odh-notebook-controller"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "notebook-images"
                        }
                      },
                      "name": "notebook-images"
                    },
                    {
                      "kustomizeConfig": {
                        "overlays": [
                          "odh-model-controller"
                        ],
                        "repoRef": {
                          "name": "manifests",
                          "path": "model-mesh"
                        }
                      },
                      "name": "model-mesh"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "data-science-pipelines-operator"
                        }
                      },
                      "name": "data-science-pipelines-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "trustyai-service"
                        }
                      },
                      "name": "trustyai"
                    }
                  ],
                  "repos": [
                    {
                      "name": "manifests",
                      "uri": "https://github.com/opendatahub-io/odh-manifests/tarball/v1.5"
                    }
                  ]
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhApplication",
                "metadata": {
                  "name": "example-odhapplication",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started,Model development,Model training,Notebook environments,Package management,Data Management"
                  }
                },
                "spec": {
                  "displayName": "ODHAPPLICATION DISPLAY NAME",
                  "provider": "PROVIDER_NAME",
                  "description": "Description of your ODHApplication",
                  "kfdefApplications": [
                    "<NAME OF THE ODH APPLICATION IN KFDEF>"
                  ],
                  "route": "<APPLICATION_OCP_ROUTE_OBJECT_NAME>",
                  "img": "<svg xmlns=\"http://www.w3.org/2000/svg\" role=\"img\" viewBox=\"-1.13 -1.13 58.50 48.94\"><title>Open Data Hub logo</title><g><g><g fill=\"#4d5258\" stroke-width=\".006\"><path fill=\"#bbb\" d=\"M112.30231 22.797083c-.2034 0-.35737.143517-.35737.337407 0 .192939.15397.339307.35737.339307.20339 0 .35736-.147318.35736-.339307 0-.19294-.15397-.337407-.35736-.337407zm.002.131161c.1112 0 .20149.08934.20149.206246 0 .116904-.0903.208146-.20149.208146-.1112 0-.2053-.09124-.2053-.208146 0-.116904.0931-.206246.2053-.206246zM113.04141 22.801836h-.28228v.665308h.15017v-.190088h.13211c.16918 0 .26612-.08934.26612-.244263 0-.147319-.0969-.230957-.26612-.230957zm-.008.349762h-.12451v-.225255h.12451c.0836 0 .13116.03707.13116.111202 0 .07603-.0475.114053-.13116.114053zM113.40328 22.801836v.665308h.51703v-.124508h-.36687v-.147318h.32125V23.07081h-.32125v-.144467h.35642v-.124507zM114.03814 22.801836v.665308h.14352v-.426748l.3222.426748h.13686v-.665308h-.14257v.427698L114.175 22.801836z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/><path d=\"M115.29685 22.801836h-.28228v.665308h.27658c.21004 0 .35546-.136863.35546-.332654 0-.196741-.14351-.332654-.34976-.332654zm.004.538899h-.13591v-.412491h.12926c.11405 0 .19959.08459.19959.207196 0 .121657-.0817.205295-.19294.205295zM116.21878 23.467144h.16062l-.28418-.665308h-.15492l-.28894.665308h.15493l.0523-.12926h.30794zm-.3108-.249966l.1055-.25852.10455.25852zM116.31525 22.801836v.127359h.20244v.537949h.15017v-.537949h.2034v-.127359zM117.37527 23.467144h.16063l-.28419-.665308h-.15492l-.28893.665308h.15492l.0523-.12926h.30795zm-.31079-.249966l.1055-.25852.10454.25852zM117.83433 22.801836v.665308h.15017v-.262322h.30129v.262322h.15017v-.665308h-.15017v.277528h-.30129v-.277528zM118.72025 23.188665v-.386829h-.15017v.386829c0 .175831.11785.285132.30224.285132.18343 0 .29844-.109301.29844-.285132v-.386829h-.15017v.386829c0 .09599-.0589.153971-.14827.153971-.0893 0-.15207-.05798-.15207-.153971zM119.61059 22.801836h-.30604v.665308h.31649c.15492 0 .25092-.07033.25092-.184385 0-.08744-.0532-.152071-.13876-.168228.0713-.01806.115-.07413.115-.146368 0-.102648-.0912-.166327-.23761-.166327zm-.008.268024h-.14826v-.145418h.14826c.0561 0 .0913.02566.0913.07128 0 .04657-.0352.07413-.0913.07413zm0 .274677h-.14826v-.158723h.14826c.0713 0 .11406.02946.11406.07794 0 .05132-.0428.08079-.11406.08079z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g><g fill=\"#58595b\" stroke-width=\".01\"><path d=\"M112.19192 24.15027h.0381l-.13717-.306759h-.0364l-.1376.306759h.0372l.0342-.07844h.16696zm-.18712-.110871l.0688-.157763.0697.157763zM112.27858 23.843511v.306759h.0351v-.306759zM112.64205 23.843511h-.11525v.306759h.0351v-.09729h.0802c.0758 0 .11964-.039.11964-.10649 0-.0653-.0438-.102983-.11964-.102983zm-.001.177044h-.0789V23.87594h.0789c.0552 0 .0872.02454.0872.07143 0 .0482-.032.07318-.0872.07318zM112.81775 23.825105v.325165h.0342v-.325165zM113.1137 24.15027l-.00044-.152503c-.00044-.04908-.032-.07932-.089-.07932-.0394 0-.0644.01139-.0942.03111l.0145.0241c.0241-.01709.0478-.02673.0732-.02673.0408 0 .0618.0206.0618.05434v.01446h-.0758c-.0543.000439-.0863.02586-.0863.06617 0 .03944.0307.06968.0784.06968.0368 0 .0662-.01227.0846-.03725v.03593zm-.11131-.02717c-.0324 0-.053-.01709-.053-.04251 0-.02585.0202-.03769.0587-.03769h.0719v.02805c-.009.03287-.0386.05215-.0776.05215zM113.29915 24.108201c-.0127.0074-.025.01139-.0368.01183-.0197 0-.0307-.01139-.0307-.03944v-.127086h.0732v-.02673h-.0732v-.06048h-.0337v.06048h-.0333v.02673h.0333v.132783c0 .04558.025.06573.0587.0653.0184 0 .0359-.0061.0521-.01709zM113.44228 23.854466c.0118 0 .0245.0044.0351.01183l.014-.02629c-.014-.01052-.032-.01622-.0495-.01622-.0386-.000438-.0741.02717-.0741.07713v.02586h-.0298v.02673h.0298v.196764h.0337v-.196764h.0662v-.02673h-.0662v-.03024c0-.02717.0184-.04251.0408-.04207zM113.59236 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM113.80323 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM114.22747 23.918448c-.0443.000438-.0767.01972-.0911.06004-.01-.03813-.0394-.06004-.0828-.06004-.0416.000438-.0727.01753-.0881.05346v-.05215h-.0337v.230508h.0337v-.113501c0-.04996.0285-.08458.0758-.08589.0403 0 .064.02454.064.06617v.133221h.0346v-.113501c0-.04996.0276-.08458.0754-.08589.0399 0 .064.02454.064.06617v.133221h.0342V24.00828c0-.05609-.032-.08984-.0859-.08984zM114.63474 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0487.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11088-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM114.89934 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM115.02844 23.919762l.0863.230508h.0355l.0675-.188876.0679.188876h.0355l.0855-.230508h-.0355l-.0675.195888-.0679-.195888h-.0359l-.067.195888-.0684-.195888zM115.53009 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM115.73075 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM115.95034 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11482.116569.0386 0 .071-.01402.0929-.03813l-.0188-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18843c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM116.29474 23.825105v.144615c-.0171-.03287-.0491-.05171-.0903-.05171-.0653 0-.11087.04777-.11087.116569 0 .0688.0456.117445.11131.117445.0412 0 .0727-.01884.0898-.05127v.04952h.0342v-.325165zm-.0833.297557c-.0491 0-.0837-.03637-.0837-.08721 0-.05083.0346-.08765.0837-.08765.0486 0 .0833.03681.0833.08765 0 .05083-.0346.08721-.0833.08721zM116.65984 23.918448c-.0403 0-.0719.01884-.089.05083v-.144177h-.0342v.325165h.0342v-.04952c.0171.03199.0486.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM116.89657 24.183138l.10605-.263376h-.0351l-.0732.191944-.0767-.191944h-.0359l.0955.228756-.01.02191c-.009.02191-.0241.03331-.0421.03331-.0114 0-.0219-.0035-.0324-.01095l-.0145.02717c.014.01139.0294.01665.0486.01665.0307 0 .0548-.01841.0692-.05346zM117.29947 23.842634c-.0898 0-.16083.06793-.16083.154257 0 .08633.071.154694.16083.154694.0894 0 .16039-.06836.16039-.154694 0-.08633-.071-.154257-.16039-.154257zm0 .03243c.0684 0 .12446.0539.12446.121828 0 .06792-.0561.122703-.12446.122703-.0684 0-.12533-.05478-.12533-.122703 0-.06793.057-.121828.12533-.121828zM117.64414 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0486.05083.0898.05083.0653 0 .10999-.04777.10999-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM117.9061 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM118.19616 23.918448c-.0421.000438-.0732.01709-.0894.05215v-.05084h-.0342v.230508h.0342v-.12358c.004-.0447.032-.07494.0776-.07537.0408 0 .0653.02454.0653.06573v.133221h.0342v-.141986c0-.05609-.0324-.08984-.0876-.08984zM118.65836 23.901795l.0149-.03111c-.0254-.01709-.0618-.02848-.0973-.02848-.0635 0-.10605.03068-.10605.07888 0 .109995.17266.06661.17222.148559 0 .0298-.0276.04777-.0719.04777-.0346 0-.0736-.01578-.1021-.04295l-.0158.03024c.0289.02761.0727.04689.11745.04689.0657 0 .11087-.03287.11087-.0837.00044-.11131-.17223-.06968-.17223-.149436 0-.02629.0259-.04207.0653-.04207.0245 0 .057.0075.0846.02542zM118.84168 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .11701-.04777.11701-.117007 0-.06836-.0487-.11613-.11701-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM119.18446 23.919762v.113063c0 .04996-.0281.08502-.0758.08589-.0399 0-.064-.02454-.064-.06573v-.133222h-.0342v.141986c0 .05565.0324.08984.0863.08984.0416-.000438.0723-.01753.0876-.05346v.05215h.0342v-.230508zM119.33586 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM119.62557 23.976294l.0188-.02366c-.021-.02147-.0517-.03418-.089-.03418-.0666 0-.11438.0482-.11438.116568 0 .0688.0473.116569.11438.116569.0399 0 .0719-.0149.0925-.03988l-.0197-.02016c-.0162.01928-.0412.0298-.0714.0298-.0478 0-.082-.03506-.082-.08633 0-.05083.0342-.08589.082-.08589.028 0 .0513.01008.0688.02717zM119.78997 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g></g><g><g><path fill=\"#a586f3\" d=\"m 7808.8125,2097.1875 c -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 v 329.875 c -0.8781,10.2195 7.1804,19 17.4375,19 10.2571,0 19.0227,-8.7805 18.1446,-19 v -77.7324 l 0.4179,-6.1426 c 0,-55.1597 44.2153,-99.5625 99.375,-99.5625 55.1596,0 99.5625,44.4028 99.5625,99.5625 0.01,0.1837 0.015,0.3674 0.027,0.5508 -0.026,0.4409 -0.035,0.8826 -0.027,1.3242 v 110 c 0,9.6822 7.849,17.5312 17.5312,17.5312 9.6823,0 17.5313,-7.849 17.5313,-17.5312 v -110 c 0,-0.3245 -0.012,-0.6488 -0.031,-0.9727 0.018,-0.3004 0.029,-0.6013 0.031,-0.9023 1e-4,-74.075 -60.55,-134.4374 -134.625,-134.4375 -39.9177,1e-4 -75.8465,17.5354 -100.5,45.293 V 2115 c 0.1749,-9.8356 -7.7881,-17.8833 -17.625,-17.8125 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><path fill=\"#5ed3ee\" d=\"m 7765.8027,2098.123 c -0.038,5e-4 -0.077,0 -0.1152,0 -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 V 2361 c 0,41.9333 -26.0162,79.2671 -65.4375,93.5625 -39.4213,14.2954 -83.5594,2.3741 -110.4375,-29.8125 -26.878,-32.1865 -30.6579,-77.5719 -9.5625,-113.8125 21.0955,-36.2406 62.4262,-55.2897 103.6875,-47.8125 9.5173,1.7397 18.6463,-4.5547 20.4043,-14.0686 1.758,-9.514 -4.5187,-18.6551 -14.0293,-20.4314 -55.6354,-10.082 -111.8054,15.8219 -140.25,64.6875 -28.4446,48.8656 -23.3039,110.5381 12.9375,153.9375 36.2415,43.3993 95.908,59.4005 149.0625,40.125 44.7667,-16.2337 77.0286,-54.658 86.0977,-100.1562 0,0 1.9812,-4.7686 2.5898,-9.1563 0.6086,-4.3877 0,-17.0625 0,-17.0625 v -245.0625 c 0.1769,-9.865 -7.8313,-17.9263 -17.6973,-17.8145 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><g fill=\"none\" stroke-width=\".365\" transform=\"translate(-449.265 -83.342) scale(5.91262)\"><circle cx=\"79.677\" cy=\"16.823\" r=\".645\" stroke=\"#c6ea70\"/><path stroke=\"#dff3b1\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 80.808656,18.234303 a 1.8088735,1.8088735 0 0 1 -2.309183,-0.03848 1.8088735,1.8088735 0 0 1 -0.389241,-2.276466 1.8088735,1.8088735 0 0 1 2.165211,-0.803532\"/><path stroke=\"#f2fadf\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 82.216272,17.248644 a 2.455857,2.4425189 0 0 1 -2.468226,2.016391 2.455857,2.4425189 0 0 1 -2.383743,-2.114547 2.455857,2.4425189 0 0 1 1.733108,-2.668494\"/></g></g></g></g></svg>",
                  "category": "Open Data Hub",
                  "support": "open data hub",
                  "docsLink": "http://REPLACE.THIS.URL",
                  "quickStart": "example-odhquickstart",
                  "getStartedLink": "http://REPLACE.THIS.URL",
                  "getStartedMarkDown": "# Header for Getting Started MarkDown section\nHere is where you add information about your ODHApplication in markdown format\n\n1. List item\n1. List item\n1. List item"
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhDocument",
                "metadata": {
                  "name": "example-odhdocument",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHDocument",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "type": "how-to",
                  "description": "This is an example of an ODHDocument custom resource",
                  "url": "http://REPLACE.THIS.URL",
                  "durationMinutes": 5
                }
              },
              {
                "apiVersion": "console.openshift.io/v1",
                "kind": "OdhQuickStart",
                "metadata": {
                  "name": "example-odhquickstart",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHQuickstart Name",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "durationMinutes": 5,
                  "description": "Description of an example quickstart",
                  "introduction": "### Introduction Header\nThis is the body of the introduction for an ODHQuickstart",
                  "tasks": [
                    {
                      "title": "Title of the first task",
                      "description": "### Description header for the first task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the first task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    },
                    {
                      "title": "Title of the second task",
                      "description": "### Description header for the second task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the second task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    }
                  ],
                  "conclusion": "Message for the conclusion of the ODHQuickstart",
                  "nextQuickStart": null
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v1.5.0
          createdAt: "2022-10-18T00:00:00Z"
          description: Open Data Hub Operator for deployment and management of Open
            Data Hub
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
          support: Open Data Hub
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KfDef is the Schema for the kfdefs API
            displayName: Kf Def
            kind: KfDef
            name: kfdefs.kfdef.apps.kubeflow.org
            version: v1
          - description: OdhApplication is the Schema for the odhapplications
            displayName: Open Data Hub Dashboard Application
            kind: OdhApplication
            name: odhapplications.dashboard.opendatahub.io
            version: v1
          - description: OdhDashboardConfig is the Schema for the odhdashboardconfigs
            displayName: Open Data Hub Dashboard Configuration
            kind: OdhDashboardConfig
            name: odhdashboardconfigs.opendatahub.io
            version: v1alpha
          - description: OdhDocument is the Schema for the odhdocuments
            displayName: Open Data Hub Dashboard Documentation
            kind: OdhDocument
            name: odhdocuments.dashboard.opendatahub.io
            version: v1
          - description: Extension for guiding user through various workflows in the
              Open Data Hub dashboard.
            displayName: Open Data Hub Dashboard QuickStart
            kind: OdhQuickStart
            name: odhquickstarts.console.openshift.io
            version: v1
        description: |
          The Open Data Hub is a machine-learning-as-a-service platform built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open Data Hub integrates multiple AI/ML open source components into one operator that can easily be downloaded and installed by OpenShift users.

          Open Data Hub operator allows users to install and manage components of the Open Data Hub. Users can mix and match tools from each project to fulfill the needs of their use case. Each of the
          projects share some components, but can be mostly seen as an extension of each other to provide a complete solution for both novice and skilled enterprise users.

          ### Components
          * Open Data Hub Dashboard - A web dashboard that displays installed Open Data Hub components with easy access to component UIs and documentation
          * ODH Notebook Controller - Secure management of Jupyter Notebook in Kubernetes environments built on top of Kubeflow Notebook Controller with support for OAuth
          * Jupyter Notebooks - JupyterLab notebook that provide Python support for GPU workloads
          * Data Science Pipelines - Pipeline solution for end to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton
          * Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh, a general-purpose model serving management/routing layer.

          To install one or multiple of these components use the default KfDef provided with the operator.

          ### Available Channels

          #### Stable

          Channel `stable` offers the major releases of Open Data Hub operator and ODH manifests. It is based on Kubeflow Operator and Kustomize for deployment configuration. This version is compatible with [Open Data Hub manifests](https://github.com/opendatahub-io/odh-manifests/) as well as [Kubeflow manifests](https://github.com/opendatahub-io/manifests). This channel will offer new components and features that have been thoroughly tested by consumers of the `rolling` channel

          NOTE: The `stable` channel will be deprecated starting with the ODH 1.6 release and we will only be publishing updates on the `rolling` channel for all future releases

          #### Rolling

          Channel `rolling` will offer the latest release of Open Data Hub operator and manifests. Subscribing to this branch will provide faster updates and access to components that will be in developement and testing for the next stable release
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io
        maintainers:
        - email: maintainers@lists.opendatahub.io
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Open Data Hub
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/opendatahub/opendatahub-operator:v1.5.0
        version: 1.5.0
      entries:
      - name: opendatahub-operator.v1.5.0
        version: 1.5.0
      - name: opendatahub-operator.v1.4.2
        version: 1.4.2
      - name: opendatahub-operator.v1.4.1
        version: 1.4.1
      - name: opendatahub-operator.v1.4.0
        version: 1.4.0
      - name: opendatahub-operator.v1.3.0
        version: 1.3.0
      - name: opendatahub-operator.v1.2.0
        version: 1.2.0
      - name: opendatahub-operator.v1.1.2
        version: 1.1.2
      - name: opendatahub-operator.v1.1.1
        version: 1.1.1
      name: stable
    defaultChannel: fast
    packageName: opendatahub-operator
    provider:
      name: ODH
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Thales Group
      provider-url: https://cpl.thalesgroup.com/data-protection
    name: cte-k8s-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cte-k8s-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cte-k8s-operator.csi.cte.cpl.thalesgroup.com/v1",
                "kind": "CteK8sOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cte-k8s-operator"
                  },
                  "name": "ctek8soperator-sample"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-06-05T08:24:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CteK8sOperator is the Schema for the ctek8soperators API
            displayName: Cte K8s Operator
            kind: CteK8sOperator
            name: ctek8soperators.cte-k8s-operator.csi.cte.cpl.thalesgroup.com
            version: v1
        description: |-
          # Kubernetes Operator to deploy CTE for Kubernetes on an OCP/K8s Cluster.
          This Operator deploys CTE for Kubernetes on the OCP/K8s Cluster.

          ## About this operator

          It is a GO based operator that manages the deployment, upgrade and
          deletion of the CTE for Kubernetes on all the nodes of the Cluster on
          which it is installed.

          ## Prerequisites

          Applicable only on OCP cluster
            For an OCP cluster the operator requires a SecurityContextConstraints (SCC)
            to be created before installation. To create the SecurityContextConstraints
            you can use a definition provided by the deployment scripts at
            https://github.com/thalescpl-io/ciphertrust-transparent-encryption-kubernetes

            Look for cte-csi-scc.yaml file in deploy/kubernetes/X.Y.Z/operator-deploy/
            Where X.Y.Z is the last released version of the deploy scripts.

          For more information refer to the guide at:
            https://thalesdocs.com/ctp/cte-con/cte-k8s/latest/cte-4-k8s/index.html
        displayName: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Encryption
        - Data Protection
        - Security
        - Kubernetes
        - Openshift
        links:
        - name: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          url: https://cpl.thalesgroup.com/encryption/ciphertrust-transparent-data-encryption-kubernetes
        maintainers:
        - email: technical.support.DIS@thalesgroup.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Thales Group
          url: https://cpl.thalesgroup.com/data-protection
        relatedImages:
        - docker.io/thalesciphertrust/ciphertrust-transparent-encryption-kubernetes-operator@sha256:d58cbb65d758fd0b9d4cfba4cb48a35bdaebe2e239fb25edb9b491cd8c2c3f73
        version: 1.6.5
      entries:
      - name: cte-k8s-operator.v1.6.5
        version: 1.6.5
      name: candidate
    - currentCSV: cte-k8s-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cte-k8s-operator.csi.cte.cpl.thalesgroup.com/v1",
                "kind": "CteK8sOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cte-k8s-operator"
                  },
                  "name": "ctek8soperator-sample"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-06-05T08:24:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CteK8sOperator is the Schema for the ctek8soperators API
            displayName: Cte K8s Operator
            kind: CteK8sOperator
            name: ctek8soperators.cte-k8s-operator.csi.cte.cpl.thalesgroup.com
            version: v1
        description: |-
          # Kubernetes Operator to deploy CTE for Kubernetes on an OCP/K8s Cluster.
          This Operator deploys CTE for Kubernetes on the OCP/K8s Cluster.

          ## About this operator

          It is a GO based operator that manages the deployment, upgrade and
          deletion of the CTE for Kubernetes on all the nodes of the Cluster on
          which it is installed.

          ## Prerequisites

          Applicable only on OCP cluster
            For an OCP cluster the operator requires a SecurityContextConstraints (SCC)
            to be created before installation. To create the SecurityContextConstraints
            you can use a definition provided by the deployment scripts at
            https://github.com/thalescpl-io/ciphertrust-transparent-encryption-kubernetes

            Look for cte-csi-scc.yaml file in deploy/kubernetes/X.Y.Z/operator-deploy/
            Where X.Y.Z is the last released version of the deploy scripts.

          For more information refer to the guide at:
            https://thalesdocs.com/ctp/cte-con/cte-k8s/latest/cte-4-k8s/index.html
        displayName: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Encryption
        - Data Protection
        - Security
        - Kubernetes
        - Openshift
        links:
        - name: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          url: https://cpl.thalesgroup.com/encryption/ciphertrust-transparent-data-encryption-kubernetes
        maintainers:
        - email: technical.support.DIS@thalesgroup.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Thales Group
          url: https://cpl.thalesgroup.com/data-protection
        relatedImages:
        - docker.io/thalesciphertrust/ciphertrust-transparent-encryption-kubernetes-operator@sha256:d58cbb65d758fd0b9d4cfba4cb48a35bdaebe2e239fb25edb9b491cd8c2c3f73
        version: 1.6.5
      entries:
      - name: cte-k8s-operator.v1.6.5
        version: 1.6.5
      name: stable
    defaultChannel: stable
    packageName: cte-k8s-operator
    provider:
      name: Thales Group
      url: https://cpl.thalesgroup.com/data-protection
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-team-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-team-operator.v2024.10.2-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": null,
                        "host": null,
                        "process": null
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master",
                        "operator": "",
                        "value": ""
                      }
                    ]
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": null
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:12c1d2e9347ec447196a618b481a0970dacac7a8a72a3b38ec22efa69668d180
          createdAt: "2024-10-25T01:47:52Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2024.10.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2023.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                   operator: ""
                   value: ""

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:7c4868db021beaffa762eb0de43739bbc865a4bf5900f1d17f78caed8612c073
        - registry.connect.redhat.com/ca/uma-operator@sha256:12c1d2e9347ec447196a618b481a0970dacac7a8a72a3b38ec22efa69668d180
        version: 2024.10.2-12
      entries:
      - name: uma-team-operator.v2024.10.2-12
        version: 2024.10.2-12
      name: "2024.10"
    - currentCSV: uma-team-operator.v2025.1.3-33
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": null,
                        "host": null,
                        "process": null
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.1.1.33,990033",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master",
                        "operator": "",
                        "value": ""
                      }
                    ]
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": null
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:78b66a291f58aa94123cdd59f2aa07dce4928162c481fe9de5ce524782bc93de
          createdAt: "2025-02-18T05:35:06Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.1.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2023.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.1.1.26,990026"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                   operator: ""
                   value: ""

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:78b66a291f58aa94123cdd59f2aa07dce4928162c481fe9de5ce524782bc93de
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:ae91a28d57c8a359f81757e4bf79472f2ac189ac0ae3ee0d87e8ff9c433f1ccc
        version: 2025.1.3-33
      entries:
      - name: uma-team-operator.v2025.1.3-33
        version: 2025.1.3-33
      name: "2025.1"
    - currentCSV: uma-team-operator.v2025.2.1-25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.2.1.25,990025",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master",
                        "operator": "",
                        "value": ""
                      }
                    ]
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:d0a0090a871ff7e23d3c97b620a1b0d004a8317efe1581ba0437d073c5656825
          createdAt: "2025-03-05T05:51:44Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.2.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2023.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.1.1.26,990026"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                   operator: ""
                   value: ""

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:d0a0090a871ff7e23d3c97b620a1b0d004a8317efe1581ba0437d073c5656825
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:8bcf0dac889f38a0bf77eb9babe3f726e3b2564d5988c05bdc5c2d91ae189c61
        version: 2025.2.1-25
      entries:
      - name: uma-team-operator.v2025.2.1-25
        version: 2025.2.1-25
      name: "2025.2"
    - currentCSV: uma-team-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:60fb1f76b4d4aab569eec5f2c20f7a06d813759f5fb6ac952156b41c4d295a52
          createdAt: "2025-04-16T06:57:55Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.4.5
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.4.2.5,990005"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:60fb1f76b4d4aab569eec5f2c20f7a06d813759f5fb6ac952156b41c4d295a52
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        version: 2025.4.5-5
      entries:
      - name: uma-team-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-team-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:052055a6280f60a69caccf89f50fd8bf92da61cd59f1962f61d21467c4191970
          createdAt: "2025-07-24T14:54:58Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.5.2.12,990012"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:052055a6280f60a69caccf89f50fd8bf92da61cd59f1962f61d21467c4191970
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        version: 2025.5.4-12
      entries:
      - name: uma-team-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-team-operator.v2025.7.3-21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:e3ae2240aeb618f36f41073a5f65b03a7ad5dfdccc7f9feb830506fbd2bfd604
          createdAt: "2025-08-23T09:01:14Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.7.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.7.1.23,990023"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:e3ae2240aeb618f36f41073a5f65b03a7ad5dfdccc7f9feb830506fbd2bfd604
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.3-21
      entries:
      - name: uma-team-operator.v2025.7.3-21
        version: 2025.7.3-21
      name: "2025.7"
    - currentCSV: uma-team-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "hostmonitoring": {
                        "linux": {
                          "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:b8422d0f3a9e576948ced8244a91abc3283b26217e4c7efc7e2669ca5508930e
          createdAt: "2025-08-15T03:20:57Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 hostmonitoring:
                   linux:
                   # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                   groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.8.1.9,990009"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:b8422d0f3a9e576948ced8244a91abc3283b26217e4c7efc7e2669ca5508930e
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        version: 2025.8.2-9
      entries:
      - name: uma-team-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-team-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
          createdAt: "2025-09-24T12:24:28Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.12,990012"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-team-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-team-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "hostmonitoring": {
                        "linux": {
                          "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:8c12315745ad3564035a4db923698bf63fe32dddbc7d246f5a3867945305713e
          createdAt: "2025-07-30T08:27:50Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 hostmonitoring:
                   linux:
                   # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                   groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "99.99.0.serp_F152813,19"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:8c12315745ad3564035a4db923698bf63fe32dddbc7d246f5a3867945305713e
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        version: 2025.7.2-19
      entries:
      - name: uma-team-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-team-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:89e0505b5e526beffda9b13a1ca7b49e5ba144fd0166a7e171e9d9b084915a56
          createdAt: "2025-09-11T11:16:37Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.8,990008"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:89e0505b5e526beffda9b13a1ca7b49e5ba144fd0166a7e171e9d9b084915a56
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-team-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-team-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
          createdAt: "2025-09-24T12:24:28Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.12,990012"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-team-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-team-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ocs-client-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocs-client-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          console.openshift.io/plugins: '["odf-client-console"]'
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageClassClaim is the Schema for the storageclassclaims
              API
            displayName: Storage Class Claim
            kind: StorageClassClaim
            name: storageclassclaims.ocs.openshift.io
            version: v1alpha1
          - description: StorageClient is the Schema for the storageclients API
            displayName: Storage Client
            kind: StorageClient
            name: storageclients.ocs.openshift.io
            version: v1alpha1
        description: This operator is not intended to be directly installed
        displayName: OpenShift Data Foundation Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: OCS Client Operator
          url: https://github.com/red-hat-storage/ocs-client-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:975529d3fbdcf6f78cc1602333832f543da2918ac5c018d1ee6bea2b315106be
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:ef4276b80cacae825d4831de88d3fe6ba7cc324464087b7b28734bd2b8de3315
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel8@sha256:d83b44a1257e8015409cd261e8eed5bd8c4e581b788e8452f19114ba696a30c1
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:bc637bbbed3457c45fbb5cb1e8940b2a1dacfa64959e5bd0bfb58c5d669427fd
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:1dd7c0ef17d39ebf3311aee326202e09038cece1b7554cea23c1efe582620d0b
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:c9e5b48d84febf10ff4c9b78a8210df36c90905fbb92b6a6b5b32e0fd9a6a41d
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:9a0008f662b6afa9e2677a8a83bd662d806ad6f4d03fb49bccb78ed487ce4c7a
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:028120f2212738917f49e0330b60be5f79a208ce2974bea22b71f533aa6ce3be
        - registry.redhat.io/odf4/ocs-client-console-rhel9@sha256:ebf3f6098a9a0a06706d20fe8fc86513be34b98604bc70fd82910f0fcd17f191
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:ffcc0806ec2afb8b40d75c70ef2a6d98bb887879f5b3f573e900861316ce9e41
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel8@sha256:05cf61192ee75dc11f5777bebeacdd209232042445553329e14f542ac0084c71
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:e206dcc33863f4d8d0aa6df173a86cd50ee1d52e6bcf9efc43b39831acc20573
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:0dc0733df67bb49886ac9715b81bd4ec4a36330fe1afdde657587e1883f5ca7a
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel8@sha256:647074ada8cb88f7ed99ca14621d3b754db6a60346f12b7561cd035c05f283f7
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:390b527d10e9c7a8a9191251b0a1ba501128057cefaff35e950f0efe04ef3122
        - registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:cb93b68adb857c82f23f292cdee47d52656dc4fc5bb83fac8581110869cbe841
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:645d4fd09628f7ddcfc765e4ddacaa79fa6a7c35c03292ca2f6e317d48765d6b
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel8@sha256:1f6a17909f1fe6e09a01cad38b6a50b41d850b1c5cffd24ef9a87c8798ff15e7
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:1711a0aabd28937aac1424adebe96619a48b91f7f888f0588117f80891001da4
        version: 4.15.20-rhodf
      entries:
      - name: ocs-client-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: ocs-client-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: ocs-client-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: ocs-client-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: ocs-client-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: ocs-client-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: ocs-client-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: ocs-client-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: ocs-client-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: ocs-client-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: ocs-client-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: ocs-client-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: ocs-client-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: ocs-client-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: ocs-client-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: ocs-client-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: ocs-client-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: ocs-client-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: ocs-client-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: ocs-client-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: ocs-client-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-client-console"]'
          containerImage: registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:bd8454b5a0b235669732650ae8502831811793d93fbfbd0d0f53278fcdfed459
          createdAt: "2025-09-12T06:29:47Z"
          description: OpenShift Data Foundation client operator enables consumption
            of storage services from a remote centralized OpenShift Data Foundation
            provider cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["storageclaims.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/ocs-client-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageClaim is the Schema for the storageclaims API
            displayName: Storage Claim
            kind: StorageClaim
            name: storageclaims.ocs.openshift.io
            version: v1alpha1
          - description: StorageClient is the Schema for the storageclients API
            displayName: Storage Client
            kind: StorageClient
            name: storageclients.ocs.openshift.io
            version: v1alpha1
        description: OpenShift Data Foundation client operator enables consumption
          of storage services from a remote centralized OpenShift Data Foundation
          provider cluster.
        displayName: OpenShift Data Foundation Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: OCS Client Operator
          url: https://github.com/red-hat-storage/ocs-client-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:80c01633a6bfca44df37ac38f764035be6998fce965c8709e4f6d54d17eeaa7b
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:82a4b481028195cfd75471656fe40c1cbbade4c7b2d170d6bf9422d9d3b84fa7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel8@sha256:d83b44a1257e8015409cd261e8eed5bd8c4e581b788e8452f19114ba696a30c1
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel8@sha256:1f6a17909f1fe6e09a01cad38b6a50b41d850b1c5cffd24ef9a87c8798ff15e7
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:975529d3fbdcf6f78cc1602333832f543da2918ac5c018d1ee6bea2b315106be
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c11a599bfd09b9b22888e03eabdbe1054d4049718a591d2a45bf0ed89800111e
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6fee56851da3a87b9f7f477fc59e3385e118dc0b398861a8f13459c998a5af77
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:0afcb516eb528857887c6ee1a9cba081dc8f6b7d47d9f26a0dbb2dca0cefe07c
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:8d8326c74795c9c6bd8f33a9f9e81d2304a70b1f9be3c2f8d50e76be4b0292ff
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/odf4/ocs-client-console-rhel9@sha256:beb4f82d3923171b80aacb90dfe345a22626453bf2f3921b03a16839b204d377
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:1711a0aabd28937aac1424adebe96619a48b91f7f888f0588117f80891001da4
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:ef4276b80cacae825d4831de88d3fe6ba7cc324464087b7b28734bd2b8de3315
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:028120f2212738917f49e0330b60be5f79a208ce2974bea22b71f533aa6ce3be
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:645d4fd09628f7ddcfc765e4ddacaa79fa6a7c35c03292ca2f6e317d48765d6b
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:bc637bbbed3457c45fbb5cb1e8940b2a1dacfa64959e5bd0bfb58c5d669427fd
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:ffcc0806ec2afb8b40d75c70ef2a6d98bb887879f5b3f573e900861316ce9e41
        - registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:bd8454b5a0b235669732650ae8502831811793d93fbfbd0d0f53278fcdfed459
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:693a60d6a1162659213ae554b1c6ee200b0300a7da72c01415044cf507acab49
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:c9e5b48d84febf10ff4c9b78a8210df36c90905fbb92b6a6b5b32e0fd9a6a41d
        version: 4.16.16-rhodf
      entries:
      - name: ocs-client-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: ocs-client-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: ocs-client-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: ocs-client-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: ocs-client-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: ocs-client-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: ocs-client-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: ocs-client-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: ocs-client-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: ocs-client-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: ocs-client-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: ocs-client-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: ocs-client-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: ocs-client-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: ocs-client-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: ocs-client-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: ocs-client-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: ocs-client-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://your.domain
    name: amq-streams-proxy
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: alpha
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: amq-streams-3.0.x
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: amq-streams-3.x
    defaultChannel: alpha
    packageName: amq-streams-proxy
    provider:
      name: Red Hat
      url: https://your.domain
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-secretsmanager-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-secretsmanager-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secretsmanager.services.k8s.aws/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/secretsmanager-controller:1.1.1
          createdAt: "2025-09-19T21:14:43Z"
          description: AWS Secrets Manager controller is a service controller for
            managing Secrets Manager resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret represents the state of an AWS secretsmanager Secret
              resource.
            displayName: Secret
            kind: Secret
            name: secrets.secretsmanager.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Secrets Manager resources in AWS from within your Kubernetes cluster.

          **About Amazon Secrets Manager**

          AWS Secrets Manager helps you manage, retrieve, and rotate database credentials, application credentials, OAuth tokens, API keys, and other secrets throughout their lifecycles. [Many AWS services that use secrets](https://docs.aws.amazon.com/secretsmanager/latest/userguide/integrating.html) store them in Secrets Manager.

          Secrets Manager helps you improve your security posture, because you no longer need hard-coded credentials in application source code. Storing the credentials in Secrets Manager helps avoid possible compromise by anyone who can inspect your application or the components. You replace hard-coded credentials with a runtime call to the Secrets Manager service to retrieve credentials dynamically when you need them.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Secrets Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secretsmanager
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Secrets Manager Developer Resources
          url: https://aws.amazon.com/secrets-manager/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: secretsmanager maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/secretsmanager-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-secretsmanager-controller.v1.1.1
        version: 1.1.1
      - name: ack-secretsmanager-controller.v1.1.0
        version: 1.1.0
      - name: ack-secretsmanager-controller.v1.0.14
        version: 1.0.14
      - name: ack-secretsmanager-controller.v1.0.13
        version: 1.0.13
      - name: ack-secretsmanager-controller.v1.0.12
        version: 1.0.12
      - name: ack-secretsmanager-controller.v1.0.11
        version: 1.0.11
      - name: ack-secretsmanager-controller.v1.0.10
        version: 1.0.10
      - name: ack-secretsmanager-controller.v1.0.9
        version: 1.0.9
      - name: ack-secretsmanager-controller.v1.0.8
        version: 1.0.8
      - name: ack-secretsmanager-controller.v1.0.7
        version: 1.0.7
      - name: ack-secretsmanager-controller.v1.0.6
        version: 1.0.6
      - name: ack-secretsmanager-controller.v1.0.5
        version: 1.0.5
      - name: ack-secretsmanager-controller.v1.0.4
        version: 1.0.4
      - name: ack-secretsmanager-controller.v1.0.3
        version: 1.0.3
      - name: ack-secretsmanager-controller.v1.0.1
        version: 1.0.1
      - name: ack-secretsmanager-controller.v1.0.0
        version: 1.0.0
      - name: ack-secretsmanager-controller.v0.0.11
        version: 0.0.11
      - name: ack-secretsmanager-controller.v0.0.10
        version: 0.0.10
      - name: ack-secretsmanager-controller.v0.0.9
        version: 0.0.9
      - name: ack-secretsmanager-controller.v0.0.8
        version: 0.0.8
      - name: ack-secretsmanager-controller.v0.0.7
        version: 0.0.7
      - name: ack-secretsmanager-controller.v0.0.2
        version: 0.0.2
      - name: ack-secretsmanager-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-secretsmanager-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: crunchy-postgres-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:c559c896e174674d151f93543ac92a1b5647eb4676313f7dc6cbb8faee786f18
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:4bdd0588ffbb3f628e7df1c72670e7a5087269438343b68779012bc3a13dea31
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-exporter@sha256:3b231902b2116cd38de05c7378202716c95813b9961fb8cee811b61d36cddf86
        - registry.connect.redhat.com/crunchydata/crunchy-upgrade@sha256:68bc93cce61d9b1e44822ae3e3ac8f7a9c2b712eddfaea090eabefacd60c67f6
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:0e283616c2d32e53258d62d02702bc4e179ef8c03933592154babed1a70bf666
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:56a5e3a5758982e086c9f78201a7137e0d934316f058a31cf30df44639a569a2
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:b2c3c74fdcc0e08f555e50c527f0d322292333c5d3dce061a0dee3680a1fbc56
        - registry.connect.redhat.com/crunchydata/crunchy-pgbackrest@sha256:b70da0035b74379f2507cfabdd9d1cc1e3c3eecb9066c22f292981c9a27916e1
        - registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
        - registry.connect.redhat.com/crunchydata/crunchy-pgadmin4@sha256:0d84acb1ce6a9b046e3f8b222a583ddcc91dda0a117c9f637ef173d0ac992f57
        - registry.connect.redhat.com/crunchydata/crunchy-pgbouncer@sha256:17cb58dec78447c2f70a414a2aeff6facbf921b42a0f366d0dbf7be95aa8987a
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:a7f62f367817fc08294a0b8b82ac9826a42c0ae668efdfacaa158a4287b8a688
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.4
        version: 5.4.4
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.4.0
        version: 5.4.0
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: crunchy-postgres-operator
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hbase-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hbase-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:e56051b66b0f419874d113c0320232dba7dd2547357256cf80eeaf18bbecb298
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hbase-operator@sha256:e56051b66b0f419874d113c0320232dba7dd2547357256cf80eeaf18bbecb298
        version: 24.11.1
      entries:
      - name: hbase-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: hbase-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:232ff30ffdf4bdc33f49d87950864e15436957510ba47f352dc0b133cb43279d
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hbase-operator@sha256:232ff30ffdf4bdc33f49d87950864e15436957510ba47f352dc0b133cb43279d
        version: 25.3.0
      entries:
      - name: hbase-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: hbase-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
        version: 25.7.0
      entries:
      - name: hbase-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hbase-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
        version: 25.7.0
      entries:
      - name: hbase-operator.v25.7.0
        version: 25.7.0
      - name: hbase-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hbase-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: node-healthcheck-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-healthcheck-operator.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:60d95ea8240ecf701ba974b6df3ee296a7fb3ef5393e0dcf43af15e782f2d153
          createdAt: "2023-11-02 11:44:50"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.3.0 <0.4.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://docs.openshift.com/container-platform/4.12/nodes/nodes/eco-node-health-check-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:60d95ea8240ecf701ba974b6df3ee296a7fb3ef5393e0dcf43af15e782f2d153
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:8d32dee5c72ec1c94e4e92f8d1da935a384e40e379f237c432086a2781f4ecb9
        version: 0.4.1
      entries:
      - name: node-healthcheck-operator.v0.4.1
        version: 0.4.1
      name: 4.12-eus
    - currentCSV: node-healthcheck-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:90cceb176fa903cc8e9aa4c0f7316560a8a354a3138dc26abbfc9884c1c576b5
          createdAt: "2023-10-30 14:21:41"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.4.0 <0.6.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/node-health-check-operator
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:90cceb176fa903cc8e9aa4c0f7316560a8a354a3138dc26abbfc9884c1c576b5
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:b8d5db068bfc81031c2eef86e8649e4fc008b7480e08c9cfab2ab4497098eecf
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel8@sha256:85054b34e119c16cc95cda9dc3f5d63f2d7bfc70634f70973410b2a1b1fa58d4
        version: 0.6.1
      entries:
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.14-eus
    - currentCSV: node-healthcheck-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:4811046e26846b1f12a3b20ddc6209980897c11b171ddd61d26b43630cc1d4b4
          createdAt: "2024-08-09T14:54:37Z"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.8.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          SNR is installed automatically when installing NHC.

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/node-health-check-operator
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:4811046e26846b1f12a3b20ddc6209980897c11b171ddd61d26b43630cc1d4b4
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:942f69efc7ae05d6c4d9c33335e4be71bfae2e6f9ef062dcd7887eebfe02120e
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel8@sha256:062cd9def39a087ae3802cd5208e2f0d5bfca29ba76ebd5139d88348e8c9e923
        version: 0.8.2
      entries:
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.16-eus
    - currentCSV: node-healthcheck-operator.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:f174ada653bb5fb7e58219510d0b10d868273a70db5c486e0b3ad05a1be50640
          createdAt: "2025-08-04 12:38:05"
          description: Detect failed Nodes and trigger remediation with a remediation
            operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <0.9.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.7/html/remediation_fencing_and_maintenance
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel9@sha256:8e9926558c3d8082f298e9e5f58bc5b473020acd44c2ce853c2916cdad71d649
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:f174ada653bb5fb7e58219510d0b10d868273a70db5c486e0b3ad05a1be50640
        - registry.redhat.io/workload-availability/node-remediation-console-rhel9@sha256:37f62407284fe574b973a1a685371d191b710ad70a6bfe8172f19a4568c11cf6
        version: 0.9.1
      entries:
      - name: node-healthcheck-operator.v0.9.1
        version: 0.9.1
      - name: node-healthcheck-operator.v0.9.0
        version: 0.9.0
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.18-eus
    - currentCSV: node-healthcheck-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:8e1dbc35126101e1025bd4c0eac4f4c614eb779ccafbb55050541a27ba7e9706
          createdAt: "2022-08-15 12:34:34"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.2.0 <0.3.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-node-health-check-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b5786bbbef725badf3dfcc2c2c7a86ead5ebb584c978c47aae8b9a62e241b80d
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:8e1dbc35126101e1025bd4c0eac4f4c614eb779ccafbb55050541a27ba7e9706
        version: 0.3.1
      entries:
      - name: node-healthcheck-operator.v0.3.1
        version: 0.3.1
      name: candidate
    - currentCSV: node-healthcheck-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:2160f2688e5b098c0bae7597cebcd380019b82ae02f4506d07c2ed85f0b3664d
          createdAt: "2025-09-01 18:07:56"
          description: Detect failed Nodes and trigger remediation with a remediation
            operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.8/html/remediation_fencing_and_maintenance
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:2160f2688e5b098c0bae7597cebcd380019b82ae02f4506d07c2ed85f0b3664d
        - registry.redhat.io/workload-availability/node-remediation-console-rhel9@sha256:cc0f671b126cde10476f8cc2061c6903932e7bf5e3dbf4e8bdda9add70d5deea
        version: 0.10.0
      entries:
      - name: node-healthcheck-operator.v0.10.0
        version: 0.10.0
      - name: node-healthcheck-operator.v0.9.1
        version: 0.9.1
      - name: node-healthcheck-operator.v0.9.0
        version: 0.9.0
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: stable
    defaultChannel: stable
    packageName: node-healthcheck-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-mq-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-mq-controller.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mq.services.k8s.aws/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/mq-controller:1.1.2
          createdAt: "2025-10-03T18:49:51Z"
          description: AWS MQ controller is a service controller for managing MQ resources
            in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker represents the state of an AWS mq Broker resource.
            displayName: Broker
            kind: Broker
            name: brokers.mq.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon MQ resources in AWS from within your Kubernetes cluster.

          **About Amazon MQ**

          Amazon MQ is a managed message broker service for Apache ActiveMQ and RabbitMQ that makes it easy to set up and operate message brokers on AWS. Amazon MQ reduces your operational responsibilities by managing the provisioning, setup, and maintenance of message brokers for you. Because Amazon MQ connects to your current applications with industry-standard APIs and protocols, you can easily migrate to AWS without having to rewrite code.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon MQ
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mq
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon MQ Developer Resources
          url: https://docs.aws.amazon.com/amazon-mq/latest/developer-guide/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: mq maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/mq-controller:1.1.2
        version: 1.1.2
      entries:
      - name: ack-mq-controller.v1.1.2
        version: 1.1.2
      - name: ack-mq-controller.v1.1.1
        version: 1.1.1
      - name: ack-mq-controller.v1.1.0
        version: 1.1.0
      - name: ack-mq-controller.v1.0.13
        version: 1.0.13
      - name: ack-mq-controller.v1.0.11
        version: 1.0.11
      - name: ack-mq-controller.v1.0.10
        version: 1.0.10
      - name: ack-mq-controller.v1.0.9
        version: 1.0.9
      - name: ack-mq-controller.v1.0.8
        version: 1.0.8
      - name: ack-mq-controller.v1.0.7
        version: 1.0.7
      - name: ack-mq-controller.v1.0.6
        version: 1.0.6
      - name: ack-mq-controller.v1.0.5
        version: 1.0.5
      - name: ack-mq-controller.v1.0.2
        version: 1.0.2
      - name: ack-mq-controller.v1.0.1
        version: 1.0.1
      - name: ack-mq-controller.v1.0.0
        version: 1.0.0
      - name: ack-mq-controller.v0.0.36
        version: 0.0.36
      - name: ack-mq-controller.v0.0.35
        version: 0.0.35
      - name: ack-mq-controller.v0.0.34
        version: 0.0.34
      - name: ack-mq-controller.v0.0.31
        version: 0.0.31
      - name: ack-mq-controller.v0.0.29
        version: 0.0.29
      - name: ack-mq-controller.v0.0.27
        version: 0.0.27
      - name: ack-mq-controller.v0.0.26
        version: 0.0.26
      - name: ack-mq-controller.v0.0.25
        version: 0.0.25
      - name: ack-mq-controller.v0.0.24
        version: 0.0.24
      - name: ack-mq-controller.v0.0.23
        version: 0.0.23
      - name: ack-mq-controller.v0.0.22
        version: 0.0.22
      - name: ack-mq-controller.v0.0.21
        version: 0.0.21
      - name: ack-mq-controller.v0.0.20
        version: 0.0.20
      - name: ack-mq-controller.v0.0.19
        version: 0.0.19
      - name: ack-mq-controller.v0.0.18
        version: 0.0.18
      - name: ack-mq-controller.v0.0.17
        version: 0.0.17
      - name: ack-mq-controller.v0.0.16
        version: 0.0.16
      - name: ack-mq-controller.v0.0.15
        version: 0.0.15
      - name: ack-mq-controller.v0.0.14
        version: 0.0.14
      - name: ack-mq-controller.v0.0.13
        version: 0.0.13
      - name: ack-mq-controller.v0.0.12
        version: 0.0.12
      - name: ack-mq-controller.v0.0.11
        version: 0.0.11
      - name: ack-mq-controller.v0.0.10
        version: 0.0.10
      - name: ack-mq-controller.v0.0.9
        version: 0.0.9
      - name: ack-mq-controller.v0.0.8
        version: 0.0.8
      - name: ack-mq-controller.v0.0.7
        version: 0.0.7
      name: alpha
    defaultChannel: alpha
    packageName: ack-mq-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Accuknox Inc.
      provider-url: https://accuknox.com
    name: accuknox-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: accuknox-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.accuknox.com/v1",
                "kind": "AccuknoxConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "accuknox-operator",
                    "app.kubernetes.io/instance": "accuknoxconfig",
                    "app.kubernetes.io/name": "accuknoxconfig-sample",
                    "app.kubernetes.io/part-of": "accuknox-operator"
                  },
                  "name": "accuknoxconfig-sample",
                  "namespace": "accuknox-agents"
                },
                "spec": {
                  "ppsHost": "PLEASE_FILL_IN",
                  "knoxGateway": "PLEASE_FILL_IN",
                  "spireHost": "PLEASE_FILL_IN"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Security
          containerImage: public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:b6fae25e229fb7fca757872fc0460642d52c4f9906f776afac50c54a69b05b68
          createdAt: "2024-02-09T17:06:30Z"
          operatorframework.io/suggested-namespace: accuknox-agents
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: public.ecr.aws/k9v9d5v2/accuknox-operator
          support: Accuknox Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccuknoxConfig is the Schema for the accuknoxconfigs API
            kind: AccuknoxConfig
            name: accuknoxconfigs.operator.accuknox.com
            version: v1
        description: |2-

          The AccuKnox Operator brings native K8s support for deploying and managing AccuKnox Components on a Kubernetes cluster.
          The AccuKnox Operator helps to easily onboard a k8s cluster to the AccuKnox SaaS.

          AccuKnox Operator deploys and manages the components:

          1. Discovery Engine

          2. Policy Enforcement Agent

          3. Shared Informer Agent

          4. Feeder Service

          5. Agents Operator

          ## Prerequisites for using this Operator

          1. Get AccuKnox licence - https://app.accuknox.com/sign-up

          2. After Successful Sign-up into AccuKnox Application.

             Please naviagate to Settings >> Manager Cluster >> Onboard Cluster >>

             After successful onboarding, Copy the join token created by AccuKnox

          3. Create a secret using jointoken by executing the below command:

          ```
          oc create secret generic accuknox-jointoken --from-literal=token=<joinToken> -n accuknox-agents
          ```

          ## Deploy AccuKnox Operator using AccuknoxConfig Custom Resource

          All the information to configure the AccuKnox Operator will be provided via Accuknox during Cluster Onboarding steps.

          ### AccuknoxConfig Spec

          | Spec | Description |
          | ---- | ----------- |
          | spireHost | "spire host url" |
          | ppsHost   |  "pps host url" |
          | knoxGateway | "knox gateway url" |
          | joinTokenSecretName | "join token secret name if it's different than accuknox-jointoken" |
          | joinTokenSecretKey | "key name used to store the jointoken with the secret" |

          *Example*

          ```
          apiVersion: operator.accuknox.com/v1
          kind: AccuknoxConfig
          metadata:
            name: accuknoxconfig-default
            namespace: accuknox-agents
          spec:
            ppsHost: <ppsHost url>
            knoxGateway: <knox gateway url>
            spireHost: <spire host url>
          ```
        displayName: Accuknox Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CNAPP
        - CSPM
        - CWPP
        - Cloud Native
        - Runtime Security
        - Zero-Trust
        maturity: alpha
        provider:
          name: Accuknox Inc.
          url: https://accuknox.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        - public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:b6fae25e229fb7fca757872fc0460642d52c4f9906f776afac50c54a69b05b68
        - public.ecr.aws/k9v9d5v2/feeder-service@sha256:f5bc340d9d3e34804dd54a03701d272e0ce5c04160d1aa41984cb3d877eb5b76
        - public.ecr.aws/k9v9d5v2/discovery-engine-discover@sha256:052efbe1064503853adab497415b9407f51578e1456931a8a0e60353aa53e0c2
        - public.ecr.aws/k9v9d5v2/discovery-engine-hardening@sha256:3eb28d44ef93e6ba331e9085e64475458e1c234017c77dc30b8bc802def434c3
        - public.ecr.aws/k9v9d5v2/agents-operator@sha256:656d6bac4371adef3e86658011ba68ae2f98514f6ddaea0667d4269eb88dd02e
        - public.ecr.aws/k9v9d5v2/shared-informer-agent@sha256:6d18d13312d4309b6e32b18c063d5b26f15fbf057fda2b9e6b7e8ebb1a5885e9
        - public.ecr.aws/k9v9d5v2/policy-enforcement-agent@sha256:f004fc3572a4d66da1bc0db9e05a279d761a42d6ed62f95de7a2bb42ce71e17b
        - public.ecr.aws/k9v9d5v2/discovery-engine-sumengine@sha256:4ca78f90b28b38badb0520f148ce702a3a6cd6b02bb62e8b1a6782a5b82e0868
        - public.ecr.aws/k9v9d5v2/discovery-engine-offloader@sha256:3c2291310e30e35a931d2af389f7ade656ce86f609d2e42ecdabf1000764695c
        version: 0.2.1
      entries:
      - name: accuknox-operator.v0.2.1
        version: 0.2.1
      name: alpha
    - currentCSV: accuknox-operator.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.accuknox.com/v1",
                "kind": "AccuknoxConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "accuknox-operator",
                    "app.kubernetes.io/instance": "accuknoxconfig",
                    "app.kubernetes.io/name": "accuknoxconfig-sample",
                    "app.kubernetes.io/part-of": "accuknox-operator"
                  },
                  "name": "accuknoxconfig-sample",
                  "namespace": "accuknox-agents"
                },
                "spec": {
                  "ppsHost": "PLEASE_FILL_IN",
                  "knoxGateway": "PLEASE_FILL_IN",
                  "spireHost": "PLEASE_FILL_IN"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Security
          containerImage: public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:a935ad806e82760bb838bada30d3ed1a54c8c61b95fe5c6109f48985c6faf4d7
          createdAt: "2024-08-23T04:25:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: accuknox-agents
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: docker.io/accuknox/accuknox-operator
          support: Accuknox Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccuknoxConfig is the Schema for the accuknoxconfigs API
            kind: AccuknoxConfig
            name: accuknoxconfigs.operator.accuknox.com
            version: v1
        description: |2-

          The AccuKnox Operator brings native K8s support for deploying and managing AccuKnox Components on a Kubernetes cluster.
          The AccuKnox Operator helps to easily onboard a k8s cluster to the AccuKnox SaaS.

          AccuKnox Operator deploys and manages the components:

          1. Discovery Engine

          2. Policy Enforcement Agent

          3. Shared Informer Agent

          4. Feeder Service

          5. Agents Operator

          ## Prerequisites for using this Operator

          1. Get AccuKnox licence - https://app.accuknox.com/sign-up

          2. After Successful Sign-up into AccuKnox Application.

             Please naviagate to Settings >> Manager Cluster >> Onboard Cluster >>

             After successful onboarding, Copy the join token created by AccuKnox

          3. Create a secret using jointoken by executing the below command:

          ```
          oc create secret generic accuknox-jointoken --from-literal=token=<joinToken> -n accuknox-agents
          ```

          ## Deploy AccuKnox Operator using AccuknoxConfig Custom Resource

          All the information to configure the AccuKnox Operator will be provided via Accuknox during Cluster Onboarding steps.

          ### AccuknoxConfig Spec

          | Spec | Description |
          | ---- | ----------- |
          | spireHost | "spire host url" |
          | ppsHost   |  "pps host url" |
          | knoxGateway | "knox gateway url" |
          | joinTokenSecretName | "join token secret name if it's different than accuknox-jointoken" |
          | joinTokenSecretKey | "key name used to store the jointoken with the secret" |

          *Example*

          ```
          apiVersion: operator.accuknox.com/v1
          kind: AccuknoxConfig
          metadata:
            name: accuknoxconfig-default
            namespace: accuknox-agents
          spec:
            ppsHost: <ppsHost url>
            knoxGateway: <knox gateway url>
            spireHost: <spire host url>
          ```
        displayName: Accuknox Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CNAPP
        - CSPM
        - CWPP
        - Cloud Native
        - Runtime Security
        - Zero-Trust
        maturity: stable
        provider:
          name: Accuknox Inc.
          url: https://accuknox.com
        relatedImages:
        - docker.io/accuknox/discovery-engine-discover@sha256:e2d37bfbd0e9d23c129c143eb7c9aab1c3ed55e9571b44888eea2b842f8a2361
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        - public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:a935ad806e82760bb838bada30d3ed1a54c8c61b95fe5c6109f48985c6faf4d7
        - docker.io/accuknox/discovery-engine-hardening@sha256:e596aaaabaec94e0abb4f92ad9266dc010c0e88a89d600e8750a360099cae3cc
        - docker.io/accuknox/feeder-service@sha256:295e897b6f27ce61c45fb0179b734ef6d4c9e91864bcc5d2884a2d1dd4db81b6
        - docker.io/accuknox/policy-enforcement-agent@sha256:a0364343afeac4537589bfc247f64fffd7aac16f8afe07ce050bc58b2af248fb
        - docker.io/accuknox/discovery-engine-offloader@sha256:59cc594009de89a574cd569e701b6765d59a6899a03ad07eee1c025331df4331
        - docker.io/accuknox/shared-informer-agent@sha256:9f53d7dba9569487c42c6cf9e88988a1adb4c0971742e8b521700dce2d22b8cc
        - docker.io/accuknox/discovery-engine-sumengine@sha256:f7927128b86fdfb635377bd599df06f0eb9b9f0d1b068a0047468e480855443e
        - docker.io/accuknox/agents-operator@sha256:b0ae2728f15ebae904efb778cd29215ccf07ffca945a59deb5484f6c19c1b21f
        version: 0.7.1
      entries:
      - name: accuknox-operator.v0.7.1
        version: 0.7.1
      name: stable
    defaultChannel: stable
    packageName: accuknox-operator-certified
    provider:
      name: Accuknox Inc.
      url: https://accuknox.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pachyderm, Inc.
      provider-url: https://pachyderm.com
    name: pachyderm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pachyderm-operator.v0.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "Pachyderm",
                "metadata": {
                  "name": "pachyderm-sample"
                },
                "spec": {
                  "console": {
                    "disable": true
                  },
                  "pachd": {
                    "metrics": {
                      "disable": false
                    },
                    "storage": {
                      "amazon": {
                        "credentialSecretName": "pachyderm-aws-secret"
                      },
                      "backend": "AMAZON"
                    }
                  },
                  "version": "v2.1.6"
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermExport",
                "metadata": {
                  "name": "pachydermexport-sample"
                },
                "spec": {
                  "backup": {
                    "target": "pachyderm-sample"
                  }
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermImport",
                "metadata": {
                  "name": "pachydermimport-sample"
                },
                "spec": {
                  "backup": "backups/pachyderm-backup-20220610.tar.gz",
                  "destination": {
                    "name": "pachyderm-restore",
                    "namespace": "pachyderm-workloads"
                  },
                  "storageSecret": "s3-credentials-secret"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
          createdAt: "2021-05-20T08:05:00Z"
          description: Pachyderm provides the data layer that allows data science
            teams to productionize and scale their machine learning lifecycle with
            data driven automation, petabyte scalability and end-to-end reproducibility.
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pachyderm/openshift-operator
          support: Pachyderm, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PachydermExport is the Schema for the pachydermexports API
            displayName: Pachyderm Export
            kind: PachydermExport
            name: pachydermexports.aiml.pachyderm.com
            version: v1beta1
          - description: PachydermImport is the Schema for the pachydermimports API
            displayName: Pachyderm Import
            kind: PachydermImport
            name: pachydermimports.aiml.pachyderm.com
            version: v1beta1
          - description: Pachyderm is the Schema for the pachyderms API
            displayName: Pachyderm
            kind: Pachyderm
            name: pachyderms.aiml.pachyderm.com
            version: v1beta1
        description: "The Pachyderm operator aims to make deploying and managing Pachyderm
          instances on Openshift container platform easier.\n\nPachyderm is the data
          foundation for machine learning.  Pachyderm provides industry leading data
          versioning, pipelines and lineage that allow data science teams to automate
          the machine learning lifecycle and optimize their machine learning operations
          (MLOps). \n\n## Getting started\n\n1. Pachyderm uses an object store to
          store your data. Set the pachd storage backend at `pachyderm.spec.pachd.storage.backend`
          to `AMAZON`, `MICROSOFT`, `GOOGLE` or `MINIO` depending on your storage
          provider. Additionally, create a user with read/write access to the object
          store.\n\n2. Setup the user credentials in the section, `pachyderm.spec.pachd.storage`.
          Required fields depend on the backend type set in step 1 above.\n\n3. Finally,
          deploy the pachyderm cluster in its own namespace.\n\nFor more information
          on getting started, reference the documentation.\n"
        displayName: Pachyderm
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pachyderm
        - versioning
        - machine learning
        - artificial intelligence
        links:
        - name: Documentation
          url: https://docs.pachyderm.com/latest/
        maintainers:
        - email: kyle@pachyderm.io
          name: Kyle White
        - email: eochieng@redhat.com
          name: Edmund Ochieng
        maturity: beta
        minKubeVersion: v1.19.0
        provider:
          name: Pachyderm, Inc.
          url: https://pachyderm.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - registry.connect.redhat.com/pachyderm/backup-handler@sha256:e2ef6bd8191e8ad58c765c0db8f5bb21ea7123caed0c0f58fe681a024201306d
        - registry.connect.redhat.com/pachyderm/etcd@sha256:7f330b51874379b1d4b71d94e37e0ecfa52b86d628296136970688cf204f53b5
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/pgbouncer@sha256:c3b6949227df24abb8934d111a34b44e87fb16194ac34a5d0fe284cff02e9ac0
        - registry.connect.redhat.com/pachyderm/init-utils@sha256:5c0258bfa0203079925648c0a7dda6d05a29b6f461854821a1eec541777ff359
        - registry.connect.redhat.com/pachyderm/worker@sha256:374f83d92899707bc87b0318fda6bffa78d2e9e523b827bb3d40693a2d1bdecd
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
        - registry.redhat.io/rhel8/postgresql-13@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/pachd@sha256:116477fefdf413ea02c6ac3d013f1f91d3f6e4b16aed2408c9890832cdb41430
        version: 0.1.4
      entries:
      - name: pachyderm-operator.v0.1.4
        version: 0.1.4
      name: stable
    defaultChannel: stable
    packageName: pachyderm-operator
    provider:
      name: Pachyderm, Inc.
      url: https://pachyderm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-listener-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: listener-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/tools@sha256:47204b2697dfbab5eaf9d6dcb4c160dd427a148d6b840727239b773810e74fc9
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/tools@sha256:47204b2697dfbab5eaf9d6dcb4c160dd427a148d6b840727239b773810e74fc9
        - quay.io/stackable/listener-operator@sha256:b9c7957e7e562889bddd08e679abff05bb677672b3f735932f68514b77eacfd5
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:a6d6631d35c32b874c995b6af889636ac8dd1f39dcb68810d9f9b10a4e2c1987
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:161e754437428c68490333337303fb30b695a9ac677c3a00a1dc27565b855ab8
        version: 24.11.1
      entries:
      - name: listener-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: listener-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/listener-operator@sha256:da53970954b889bf4d874ae3df7e411b9abf41950ef7d50d6289831826f045a0
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/listener-operator@sha256:da53970954b889bf4d874ae3df7e411b9abf41950ef7d50d6289831826f045a0
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:71f8465e35a67f4b96e4c4904e67ccac9bbb9ae347426afb8bbfaabd6709e983
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        version: 25.3.0
      entries:
      - name: listener-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: listener-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
        version: 25.7.0
      entries:
      - name: listener-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: listener-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        version: 25.7.0
      entries:
      - name: listener-operator.v25.7.0
        version: 25.7.0
      - name: listener-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: "25.7"
    packageName: stackable-listener-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      pod-security.kubernetes.io/audit: privileged
      pod-security.kubernetes.io/enforce: privileged
      pod-security.kubernetes.io/warn: privileged
      provider: NVIDIA Corporation
      provider-url: ""
    name: gpu-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gpu-operator-certified.v25.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "kernelModuleType": "auto",
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": []
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
          createdAt: Thu Sep 18 18:11:36 PDT 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <25.3.4'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:43aa99352e1dc47e9218f36340ae6328a6282ed1e7e9a61b4204475e980230ef
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:c22dc67465c33d2f17932931229e61a6e7d9e32c5ed2ecefe8d10e873ffedfeb
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:51c8f71d3b3c08ae4eb4853697e3f8e6f11e435e666e08210178e6a1faf8028f
        - nvcr.io/nvidia/driver@sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735
        - nvcr.io/nvidia/driver@sha256:44422f10e899d3419fbd611a02fc1c6127e071a2067bd36e3debb043bf251de0
        - nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:66c3d68ca2b35d47b5562300a10ff8711b953ba3a15df7dbc409d81613ebe5d0
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b689dfb7806ae51b0072620b47c26114f05477e8ad9725892ee10366e5d93d8d
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:119de9a331a47203858b99901f44d0c4a8052961b4e60327f4b100d0ab8c9df0
        - nvcr.io/nvidia/k8s-device-plugin@sha256:3c54348fe5a57e5700e7d8068e7531d2ef2d5f3ccb70c8f6bac0953432527abd
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:5c4e61f7ba83d7a64ff2523d447c209ce5bde1ddc79acaf1f32f19620b4912d6
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:6bb786c41fa7b1a6238c023566fdeaeeea3ba842a8daec015a7bd85a8037b249
        - nvcr.io/nvidia/cuda@sha256:d19fe621624c4eb6ac931b8558daa3ecc0c3f07f1e2a52e0267e083d22dceade
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3191cc0684e92d655e72fe9faf3da95fe6b8a6c2eb4ad396f75ba9b98c021aaf
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:619e7bf48168d76a7e087c3bf6190fda9b10449c9839309fa4f9ed5b8a9e8804
        version: 25.3.4
      entries:
      - name: gpu-operator-certified.v25.3.4
        version: 25.3.4
      - name: gpu-operator-certified.v25.3.3
        version: 25.3.3
      - name: gpu-operator-certified.v25.3.2
        version: 25.3.2
      - name: gpu-operator-certified.v25.3.1
        version: 25.3.1
      - name: gpu-operator-certified.v25.3.0
        version: 25.3.0
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: stable
    - currentCSV: gpu-operator-certified.v1.10.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "dcgmExporter": {
                    "config": {
                      "name": ""
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                  },
                  "devicePlugin": {
                  },
                  "driver": {
                    "enabled": true,
                    "use_ocp_driver_toolkit": true,
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "gfd": {
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "operator": {
                    "defaultRuntime": "crio",
                    "deployGFD": true,
                    "initContainer": {
                    }
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:c7f9074c1a7f58947c807f23f2eece3a8b04e11175127919156f8e864821d45a
          createdAt: Mon Apr 18 10:06:40 PDT 2022
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=v1.9.0 <v1.10.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cuda@sha256:7f0f7dc8c993a86865d4d9f9b0e36d113ea15603d5d5279d2879fa2bc21c8236
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:5b16056257acc51b517d9cdb1da3218693cefc214af93789e6e214fd2b4cacf1
        - nvcr.io/nvidia/k8s-device-plugin@sha256:aa95c16106280e36a5e32d2fe4c66e8b70f5a114860c6f4ed5b1a4085c63601b
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:e8f06ce53415e8572f633b2169289d973993cb1e49d4dcc686ac134194d88f33
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:24d804e8f005d7aeca8343aa13e5f92295d8642a4c47cb24e3ac86a22543bc37
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:67c48a8370c34f68b46eb57339e87d01c04198368a0423553fab7565c862c4df
        - nvcr.io/nvidia/driver@sha256:f99b66855e92451a59a722355dd52bf97eb985cc8a647172273de8178aeaa93c
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:c2e3d0fe41a0d227dbb70caec03c780cc76317515e5ab3875f31d50c63f41c66
        - nvcr.io/nvidia/gpu-operator@sha256:c7f9074c1a7f58947c807f23f2eece3a8b04e11175127919156f8e864821d45a
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:ac13256c295126b854a07421c45de0721e7a57974b38b3e66eb1fdc3a4e29f2e
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:9abb063e6b74085728fead5ed5bbe3bd7b3f57de60e87f6b964bba2d46f1bc67
        version: 1.10.1
      entries:
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v1.10
    - currentCSV: gpu-operator-certified.v1.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    },
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "dcgmExporter": {
                    "config": {
                      "name": ""
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                  },
                  "devicePlugin": {
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": true
                  },
                  "vgpuDeviceManager": {
                    "enabled": true,
                    "config": {
                      "name": "vgpu-devices-config",
                      "default": "default"
                      }
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:53180f81fb7e1e4bcd49a556bc6285c2262abbab8083b29d2688b02d4200dd2d
          createdAt: Thu Aug  4 11:57:56 PDT 2022
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=v1.9.0 <v1.11.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:6e2902d73c54cf57bec94edf166d206cfeb0b948b5aab70800ccbe025064b3a1
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:b90a519e9f30fbc013e1806bc081268724e9e80b4035c67c167ab3b43a37704a
        - nvcr.io/nvidia/gpu-operator@sha256:53180f81fb7e1e4bcd49a556bc6285c2262abbab8083b29d2688b02d4200dd2d
        - nvcr.io/nvidia/driver@sha256:cf8ce04907d0e8decf90a8dccfa6134af9c39509d2f8a21c355ed71223acacdb
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:39127d04c9faff3f22068dcf463ed48d02be684cd18e3f14db843bddbdd783a7
        - nvcr.io/nvidia/driver@sha256:b8799e7397fae5a27446737443091eed05b918816f53c400a48be71125ee22e8
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3804e2b03c44d25dd410e3882b7fb6d1fde41f55110d7702af47136b042a63ff
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:10d825ccbfcf87f8c74f841fa89ae345f2d381fb47a5b2b31f61b72ebbbb0915
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:f48e8b430248a9cadf716a9211bb0050d4a663d1f1386ea107e72917f82e5b5f
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:8490d555131a57a11311561667ea9ec8cf554a399d2492a6f9e6677c01ec4b9a
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7a0ccbcbe8d379bfc441e06504eb45d0d433a37e2a9e6dde83d13347296ebe81
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:e88e610d09f1a7528ab32cb54259aee340eba2b2bf66b92a28157626daadc82d
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:2d6e86db0e2b0db64c00566938e1851e25d4bb1019b1e7b696e2b199cba33397
        - nvcr.io/nvidia/driver@sha256:8c9fabc1cd773e9526eb7e8526084f065f6cca9d0fd8dc2a672a0bb717f1cc60
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:16ee5b1a8d9217f761c40a06c7ae598cbc893fbc35d99ff9b84fa886ea6b9b37
        - nvcr.io/nvidia/cuda@sha256:4400b56f81095b24fc9dde93ad7ad7d0ca6a7e9736c4c7f1d4e2bcbf0afbe45c
        version: 1.11.1
      entries:
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v1.11
    - currentCSV: gpu-operator-certified.v22.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:54d5846c0291db621931a8544c9cb6437512470a93278ca9130ba73f2e2f0055
          createdAt: Mon Jan 30 15:43:24 PST 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <22.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:c3cf59dd5d6160eba5d816ade2e81b35ebb10f4884df67971f6ace36f8e6efc1
        - nvcr.io/nvidia/driver@sha256:f250d1e4989fd4821e81bad9227eeeaa531ca8810f3a40e996f1ca2af7945f96
        - nvcr.io/nvidia/k8s-device-plugin@sha256:9c17d3a907eb77eb8f7b4f3faf52d8352e4252af92003f828083f80d629bd2c3
        - nvcr.io/nvidia/driver@sha256:b860ef04cc6cd8c481810fa19bef7221071f5a3862640b2ad6c2704728dc77c8
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:0d47dad29d2ef445b301c5c64717758eed43a606345b79f97bce2e64b40a91a8
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:64d757b4c80b910e64647a84a1d592fab2ea3313ff6dce30c25c3a08e180bd74
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:efb88937f73434994d1bbadc87b492a1df047aa9f8d6e9f5ec3b09536e6e7691
        - nvcr.io/nvidia/driver@sha256:f2ca4612f3418e0e3963be8c3771ffe463cdc1fc1b810437a3aad9d76bb6499d
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:bec9f026d9b3d9404c78d6091817a359015c6a7aa411735b34138c1518853b5d
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:00f1476548fbed9ee01961443a73bf65396c2e8bb2b84426f949dd56cb4d14cd
        - nvcr.io/nvidia/gpu-operator@sha256:54d5846c0291db621931a8544c9cb6437512470a93278ca9130ba73f2e2f0055
        - nvcr.io/nvidia/driver@sha256:c5a62bc2a6cce339798e4b7cf9cec47e289c2d30b40f451ba8c9e958a68e3e55
        - nvcr.io/nvidia/driver@sha256:268c26781c46b36fe691638e807eb5c3a2cd077029ddb095ac75264adac04925
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:9177a0ae30798b42d0387f6a20cd3ce3cd1799a91b7866bf812368764b05b1af
        - nvcr.io/nvidia/cuda@sha256:5f2a2d8977f2c59abe88394f05cc3c044194554f90994d8554b0e1117ab5900d
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:9a00cdfdddb73327ef8e8e0fa60e50926a388d380c551bba6a3d3012be40401d
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:50c2a2b998e467c82716efa79fe136aa6f7ea95fd23576cf384d251bb9628640
        version: 22.9.2
      entries:
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v22.9
    - currentCSV: gpu-operator-certified.v23.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:08ca33f491195b7ba867be1e6642b5ff8aff6a4b0b7cfc0ec87619252ccfddc7
          createdAt: Mon May 22 15:22:15 PDT 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <23.3.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:8ff0ee88fe6c4ff031ef642593f93625bfd9a48fe7be8a79a5c2736117e36941
        - nvcr.io/nvidia/k8s-device-plugin@sha256:80871275fbcf888d4683c58638c7a75b3c46091040cafc6210a8c93513adb146
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:03e43d6b1460cadb33509d5ee07a5e27603ab89d7ec35f67d8a51bb78745b463
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:cab21c93987a5c884075efe0fb4a8abaa1997e1696cbc773ba69889f42f8329b
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:84ce86490d0d313ed6517f2ac3a271e1179d7478d86c772da3846727d7feddc3
        - nvcr.io/nvidia/cuda@sha256:62d845b4bd6de4be65f0b6daaae6831c29d4529955f99db5d26f5f7179106285
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:5489125ceae5864280e4d6a9ab52ab0f650b3179349a7298c4a204feb60b661a
        - nvcr.io/nvidia/gpu-operator@sha256:08ca33f491195b7ba867be1e6642b5ff8aff6a4b0b7cfc0ec87619252ccfddc7
        - nvcr.io/nvidia/driver@sha256:73b8928a1b1fb650d2ca8d6451a7d7c551fd0509f7511c2fd5bfdcef83ae4f5d
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:2d7e32e3d30c2415b4eb0b48ff4ce5a4ccabaf69ede0486305ed51d26cab7713
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:21dfc9c56b5f8bce73e60361d6e83759c3fa14dc6afc2d5ebdf1b891a936daf6
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:0d47dad29d2ef445b301c5c64717758eed43a606345b79f97bce2e64b40a91a8
        - nvcr.io/nvidia/driver@sha256:917f92c589c3124dba54f66f89c44b93b0beaf418f01d0da6a546e1670fdf3a9
        - nvcr.io/nvidia/driver@sha256:3df118f2330e95db3c76173cca35e542b2da120ec0bb8f6aff83c4c96c0c2daf
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:8f03f74076b5aa6672a5d75a398fa7515fcde57fdb1cf8079e015b3366efd3b5
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:f10170b076badda8f669af8b0f7b47eb5470abaf35a67dda9c4e182d66138177
        version: 23.3.2
      entries:
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.3
    - currentCSV: gpu-operator-certified.v23.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:d2e0f8bd86e8e9f2ff3514e113657658584ca4a2d0b82dac2e47364cc0b30648
          createdAt: Thu Aug 31 12:22:44 PDT 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <23.6.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:878a6162a4fa0b61dae02c7cfc13ba4d28bb06f92fd62d0c521633ab507d675a
        - nvcr.io/nvidia/driver@sha256:bbdc4415fe2e71eeea1b8a6e50ee562e11704dabb46db1db4aea807102886a35
        - nvcr.io/nvidia/k8s-device-plugin@sha256:f1617c3fe54fd9f6d4a00358e5da1a9fa9e8963a4ef6b3ed06f47d4971ddff4b
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:b30f5c64791dc5406b57eaf51ac40195dc1051a18cd1f38fd0a43c65f1d6dba9
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:a229beba8540549c234a046e975bf4bf8c50d521dfe551af56f6e042b25ec017
        - nvcr.io/nvidia/gpu-operator@sha256:d2e0f8bd86e8e9f2ff3514e113657658584ca4a2d0b82dac2e47364cc0b30648
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:9484110986c80ab83bc404066ca4b7be115124ec04ca16bce775403e92bfd890
        - nvcr.io/nvidia/driver@sha256:b4517810f2884f345b4f2323e83e41d28f54cdc6e2caf0690bb7ed4f3947dbcc
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:bb845160b32fd12eb3fae3e830d2e6a7780bc7405e0d8c5b816242d48be9daa8
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:60086c33226bbe7f04b611c20d308fed5825226f8a0a5635bc58d8ed55cc08a5
        - nvcr.io/nvidia/cuda@sha256:64c0ff5fd4a27d4b1d46378634f7a7398c75f87c18c70c988dde18910d461560
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:00a91252be8e62da6f28cdf1ec3721f2e945ae296938b0b925ef9db2bf6dcc8b
        - nvcr.io/nvidia/driver@sha256:3382e254056f28831767bc6729bc2594353a5ff2a28fe9f2d94396c597bb23d8
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:0379c59b43e11861837787a0da22e70b528718373bb10de9e8b7752849bd4e7a
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:bf06c8ad380a36e5014607bd3ab05b80e8e5e7e77d23d0dd89fe53a5564152fa
        version: 23.6.1
      entries:
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.6
    - currentCSV: gpu-operator-certified.v23.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:7481a3697783dcdca9ae78e7b548a6900e86ea33ab49ec14f0ba55db2fdb1a2e",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:3988b3b570cf65b0d338cb29bd24b0b9bf1885e1563fbb398ef23cba0b59e6bb
          createdAt: Wed Mar 6 16:43:36 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <23.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:6a05d6a1923fda756aed0dddf7ed23a83c30cf1e6c519fc39dd70c0309ec8257
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:a67d8e92861a2dce5649105c07561e4422e9fe4ba81a6525dc0d70a7ef85f9c0
        - nvcr.io/nvidia/driver@sha256:56c79482582cdfbc58d3134e8672637c5bf05f328880f76898f526143d04c6af
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:8d1ec78f2b1ddb7f0c47453d0427231190747bda411733a7dd0c8f5196f09e9c
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:27c44f4720a4abf780217bd5e7903e4a008ebdbcf71238c4f106a0c22654776c
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:925f4a47710e4318ed457930f5406174c1f6d28b1bf6b1bc310687fec0fde712
        - nvcr.io/nvidia/driver@sha256:abda1ac56371d55917b96ff330109980f468e133c9d5705da0ef87429f14ccd7
        - nvcr.io/nvidia/k8s-device-plugin@sha256:bb81ae8d034d1e1c6a342091a5f1dda5884dd43ab47e96624487a133d1edd4b3
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:387021553dfb16aab633228d42f63f04fa932b4f46add07527f296dfe97e5148
        - nvcr.io/nvidia/driver@sha256:7481a3697783dcdca9ae78e7b548a6900e86ea33ab49ec14f0ba55db2fdb1a2e
        - nvcr.io/nvidia/gpu-operator@sha256:3988b3b570cf65b0d338cb29bd24b0b9bf1885e1563fbb398ef23cba0b59e6bb
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:011fb450af3fa2e8fe5d28d590e4c653631447bc23d149591ced3d89089c4f2c
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:9aefef081c3ab1123556374d2b15d0429f3990af2fbaccc3c9827801e1042703
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:59a3875e7a37eb370385e654184efa3a1b193c9ea352165818496b19cbe14aa4
        - nvcr.io/nvidia/cuda@sha256:714547d54e5fe4191019a1e5f1daffc7fab7481b619b79c378541dafc76c9e5d
        version: 23.9.2
      entries:
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.9
    - currentCSV: gpu-operator-certified.v24.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:96f25c67e5b1072d5981080e12d65ec37eb9ef2fc0494499416aa801b0a34da3",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:041e75a3df84039c2dbbd4b9d67763bd212138822dbb6dbc0008858c1c6eff8d
          createdAt: Tue Apr 30 14:18:26 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.3.0'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:ecb867c5787e44f1889b257e6c03b07748689a7b293c8f1affb1d4cb449b235c
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:f0e4e14a3081417be8e6a5d855fb0cb69e1b63bc54a74f17e6d7084abe275588
        - nvcr.io/nvidia/driver@sha256:56c79482582cdfbc58d3134e8672637c5bf05f328880f76898f526143d04c6af
        - nvcr.io/nvidia/driver@sha256:a836ccbe21da735aee1c39b81060ed5e2fdb4ffa339874baaf4634f1e9259f74
        - nvcr.io/nvidia/driver@sha256:96f25c67e5b1072d5981080e12d65ec37eb9ef2fc0494499416aa801b0a34da3
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:2edc1d4ed555830e70010c82558936198f5faa86fc29ecf5698219145102cfcc
        - nvcr.io/nvidia/k8s-device-plugin@sha256:80674b19898ecf9ed6949e39674da769d6feeeb01bea54b914ef9ff502834f49
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:d8774d6afd37110199512636172823ae8749e5ea2e13760f57c255a74f47018c
        - nvcr.io/nvidia/cuda@sha256:ae0623ec8634b6c88f815b88037763def160cbbac15013b77ddef257fc276c9a
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:1ae0b923bc57f47bab046b50c50110f6914bbaffbfef704df34b3fe332db2e31
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:0c77725de8c42d248ed825453efd2e005f4900f4be384fd23084f6b721ddd0e0
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:7798eb9b9424e09959fc808596f87bdb68ac5740174123c6646dee2166d3e7d2
        - nvcr.io/nvidia/gpu-operator@sha256:041e75a3df84039c2dbbd4b9d67763bd212138822dbb6dbc0008858c1c6eff8d
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:2cc2119d20515f8ca7a61a0f3932578d69a45bcdea49e6c320a89c56f105e7d9
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:2442fd2ec0ee29746abbced727c53741f017e8f3f615321c1ec42c1c3ffa0b12
        version: 24.3.0
      entries:
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.3
    - currentCSV: gpu-operator-certified.v24.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:858de27c152669f5a3cf4287406405b16dd5bb70c0373324eb735511997bb415",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:8e0969cffc030a89c4acd68e64d41dd54e3bce8a794106b178d4dbd636a07f1c
          createdAt: Wed Sep 25 08:25:16 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.6.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:21f4c8b88716e8e6f732f9fb4c2efaef937c227491a8631c5e55036f80f39a4d
        - nvcr.io/nvidia/driver@sha256:858de27c152669f5a3cf4287406405b16dd5bb70c0373324eb735511997bb415
        - nvcr.io/nvidia/gpu-operator@sha256:8e0969cffc030a89c4acd68e64d41dd54e3bce8a794106b178d4dbd636a07f1c
        - nvcr.io/nvidia/cuda@sha256:de5b598bca89850567c4c104411d66bb52f47c9179199e6a3be6829b7ac586a2
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:740abc3ff657545c10effd5354f09af525200ed9a1b7623f0c2e8c7bd9e4a4e2
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:96380b95396b7f29d2ed2ec8ef1ad5a8bcf9f55051db723295dc0a20db845331
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:ae63fac9a4057a7646f0cf0ee0566e8928529adde05c4c0a017cda0599e381b2
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:33de74eb590f071403c17b6c210c02963245851971168bc0c07c06c100a9f376
        - nvcr.io/nvidia/driver@sha256:07e11f85d54d49ec9648fb06e148b8d832ee1f9c3549a915eee853c9ef2949c2
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:781fb47e264d9e0fbc8da5bd046e5e678316c866bc36ddd4b56d4eb0de682d5b
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:969147c01d63be5d1fe458f32f1cc0c7408cf3062531db91408e2fc57b4d8a67
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:15dab1273345df4a5844c4c761d064dbc4b592101251dc39174e597137123027
        - nvcr.io/nvidia/driver@sha256:a6d12fb5753f267dda25dfd38910f972bc632c006a24107fa50e20bba3642d7c
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7ad2c9f71fe06f9f7745ac8635f46740fbdff4f11edd468addfab81afcdfa534
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:7bcc188703f2fac630f7a4ff8960e6733ac3a29adf6a946533b796d9a27b8acf
        version: 24.6.2
      entries:
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.6
    - currentCSV: gpu-operator-certified.v24.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:6fe74322562c726c8fade184d8c45ebae3da7b1ea0a21f0ff9dc42c66c65e692",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:92f5262bb4934e0ea29eb10ec0ad4fc293164f2f2f94d48a935922791abbfdac
          createdAt: Mon Jan 27 16:10:52 PST 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cuda@sha256:b3b340847db9abd0c39703278f5638d3e65e1369e9a526a19b3138d5a6563527
        - nvcr.io/nvidia/gpu-operator@sha256:92f5262bb4934e0ea29eb10ec0ad4fc293164f2f2f94d48a935922791abbfdac
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:34eab99d7992a57c35803ec5e0afd538d16e997341e951fef4c4019afb08793e
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:e8398b2451996e5c64fc3855e7a3c93b667a85f69c0e259049d3b0012c21545d
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:d934033f9ff89087d555850eb1da150eaaa27b11e5ad21db1b874b8bc0b8d35b
        - nvcr.io/nvidia/driver@sha256:6fe74322562c726c8fade184d8c45ebae3da7b1ea0a21f0ff9dc42c66c65e692
        - nvcr.io/nvidia/driver@sha256:38b66a8d44cab9e2c62da9e101f32cd9dbcb5e02d8e57b47671284d374ca3695
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:4ffa1cd2a6497eb647a89ed259dcfb007554737b9d80f69bc173a2c3cd72a1da
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:40830d3561c14743f484d45b498141f9e86b1308e16fae3978110783927264ab
        - nvcr.io/nvidia/driver@sha256:026f8f2d29b7058ecaaa5b98666e28e1b3646b73e7d7cda3dc8026d47a929152
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:cf39d78ced7fb5727a9668ee2cd44b14bb7a23a95b83d5464b7d755740e02121
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7089559ce6153018806857f5049085bae15b3bf6f1c8bd19d8b12f707d087dea
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b072c5793be65eee556eaff1b9cbbd115a1ef29982be95b2959adfcb4bc72382
        - nvcr.io/nvidia/driver@sha256:d55b57938866e538acc3a71ca32f8cf87e71c591abd4a34695ee428e7ec2fa73
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:7edd7a0413dcb39b6e3bcefaf06812f3293c8e480ca10783e821a561ed686200
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:857f7669ae5ce2a43a3b5691fa970085098321ef8e90a896fa8dda3dee5b1d2b
        version: 24.9.2
      entries:
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.9
    - currentCSV: gpu-operator-certified.v25.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "kernelModuleType": "auto",
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": []
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
          createdAt: Thu Sep 18 18:11:36 PDT 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <25.3.4'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cuda@sha256:d19fe621624c4eb6ac931b8558daa3ecc0c3f07f1e2a52e0267e083d22dceade
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:619e7bf48168d76a7e087c3bf6190fda9b10449c9839309fa4f9ed5b8a9e8804
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:66c3d68ca2b35d47b5562300a10ff8711b953ba3a15df7dbc409d81613ebe5d0
        - nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:5c4e61f7ba83d7a64ff2523d447c209ce5bde1ddc79acaf1f32f19620b4912d6
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:51c8f71d3b3c08ae4eb4853697e3f8e6f11e435e666e08210178e6a1faf8028f
        - nvcr.io/nvidia/driver@sha256:44422f10e899d3419fbd611a02fc1c6127e071a2067bd36e3debb043bf251de0
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3191cc0684e92d655e72fe9faf3da95fe6b8a6c2eb4ad396f75ba9b98c021aaf
        - nvcr.io/nvidia/driver@sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735
        - nvcr.io/nvidia/k8s-device-plugin@sha256:3c54348fe5a57e5700e7d8068e7531d2ef2d5f3ccb70c8f6bac0953432527abd
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:6bb786c41fa7b1a6238c023566fdeaeeea3ba842a8daec015a7bd85a8037b249
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:c22dc67465c33d2f17932931229e61a6e7d9e32c5ed2ecefe8d10e873ffedfeb
        - nvcr.io/nvidia/driver@sha256:43aa99352e1dc47e9218f36340ae6328a6282ed1e7e9a61b4204475e980230ef
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:119de9a331a47203858b99901f44d0c4a8052961b4e60327f4b100d0ab8c9df0
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b689dfb7806ae51b0072620b47c26114f05477e8ad9725892ee10366e5d93d8d
        version: 25.3.4
      entries:
      - name: gpu-operator-certified.v25.3.4
        version: 25.3.4
      - name: gpu-operator-certified.v25.3.3
        version: 25.3.3
      - name: gpu-operator-certified.v25.3.2
        version: 25.3.2
      - name: gpu-operator-certified.v25.3.1
        version: 25.3.1
      - name: gpu-operator-certified.v25.3.0
        version: 25.3.0
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v25.3
    defaultChannel: v25.3
    packageName: gpu-operator-certified
    provider:
      name: NVIDIA Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Yugabyte Inc
      provider-url: ""
    name: yugabyte-platform-operator-bundle-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: yugabyte-platform-operator.v0.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "yugabyte.com/v1alpha1",
                "kind": "YBPlatform",
                "metadata": {
                  "name": "ybplatform-sample"
                },
                "spec": {
                  "domainName": "cluster.local",
                  "helm": {
                    "package": "/opt/yugabyte/helm/yugabyte-latest.tgz",
                    "timeout": 900
                  },
                  "helm2Legacy": false,
                  "image": {
                    "commonRegistry": "",
                    "nginx": {
                      "name": "nginxinc/nginx-unprivileged",
                      "registry": "",
                      "tag": "1.17.4-amd64"
                    },
                    "postgres": {
                      "name": "postgres",
                      "registry": "",
                      "tag": 11.5
                    },
                    "prometheus": {
                      "name": "prom/prometheus",
                      "registry": "",
                      "tag": "v2.2.1"
                    },
                    "pullPolicy": "IfNotPresent",
                    "pullSecret": "yugabyte-k8s-pull-secret",
                    "repository": "registry.connect.redhat.com/yugabytedb/yugabyte-platform",
                    "tag": "latest",
                    "thirdparty-deps": {
                      "name": "yugabyte/thirdparty-deps",
                      "registry": "quay.io",
                      "tag": "latest"
                    }
                  },
                  "ip_version_support": "v4_only",
                  "ocpCompatibility": {
                    "enabled": true
                  },
                  "rbac": {
                    "create": true
                  },
                  "securityContext": {
                    "enabled": false,
                    "fsGroup": 1001,
                    "runAsUser": 1001
                  },
                  "tls": {
                    "certificate": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tDQpNSUlHTHpDQ0JSZWdBd0lCQWdJUUQ2R1ZmL0ViUDhSMFQvcWxjRGZyM3pBTkJna3Foa2lHOXcwQkFRc0ZBREJlDQpNUXN3Q1FZRFZRUUdFd0pWVXpFVk1CTUdBMVVFQ2hNTVJHbG5hVU5sY25RZ1NXNWpNUmt3RndZRFZRUUxFeEIzDQpkM2N1WkdsbmFXTmxjblF1WTI5dE1SMHdHd1lEVlFRREV4UlNZWEJwWkZOVFRDQlNVMEVnUTBFZ01qQXhPREFlDQpGdzB4T1RBNE1ERXdNREF3TURCYUZ3MHlNVEEzTXpFeE1qQXdNREJhTUI4eEhUQWJCZ05WQkFNTUZDb3VZMnh2DQpkV1F1ZVhWbllXSjVkR1V1WTI5dE1JSUJJakFOQmdrcWhraUc5dzBCQVFFRkFBT0NBUThBTUlJQkNnS0NBUUVBDQp3SmswU25ub1pRYW1hWm0rMVVJdjdRYjBLalhZeVEwckpJSGpsbE03elhFYUFoL0JqVjZ1M0h1TWxrTlZSekVFDQpGeEtGUm5nSlpXTVlPcy9sWDU1QmtBS0RocXNmaDl0ZVA4ZFVQU1RTVTN6bWRVc0dTcDJIMGpKbVhuRWNvMFpzDQpwYlVyYWhlcWlxSUUyZ1gwQlpvYldvWGRHc1EydXJiUGxyMGlVSnUxVnp4SGRoTS9nTnFQWEN5bE5YdENrbVI2DQo0eXpIa0Rob21pRTAySTMwTXBlOERxczl4d3lKd3F6T3J2ZnA0eUE0Ukhna2g1clovK1Erc0VQNkIyVzBHQXc2DQpzZDNTYytOTytYVUZoVnlFRDhqZkNsTndPVHZodi9keE5IWHZ4WHd0RWNjZTd1TVBBTFJzSFh1RW1kTkN5N2pvDQpibzJuUWFObzZBQ1c0U2s5RWJVZXB3SURBUUFCbzRJREpqQ0NBeUl3SHdZRFZSMGpCQmd3Rm9BVVU4b1hXZnhyDQp3QU1oTHhxdTVLcW9ISUpXMm5Vd0hRWURWUjBPQkJZRUZHemw3bjdOMy9GbFdDSDFvdWFCWldKMjI1UStNQjhHDQpBMVVkRVFRWU1CYUNGQ291WTJ4dmRXUXVlWFZuWVdKNWRHVXVZMjl0TUE0R0ExVWREd0VCL3dRRUF3SUZvREFkDQpCZ05WSFNVRUZqQVVCZ2dyQmdFRkJRY0RBUVlJS3dZQkJRVUhBd0l3UGdZRFZSMGZCRGN3TlRBem9ER2dMNFl0DQphSFIwY0RvdkwyTmtjQzV5WVhCcFpITnpiQzVqYjIwdlVtRndhV1JUVTB4U1UwRkRRVEl3TVRndVkzSnNNRXdHDQpBMVVkSUFSRk1FTXdOd1lKWUlaSUFZYjliQUVDTUNvd0tBWUlLd1lCQlFVSEFnRVdIR2gwZEhCek9pOHZkM2QzDQpMbVJwWjJsalpYSjBMbU52YlM5RFVGTXdDQVlHWjRFTUFRSUJNSFVHQ0NzR0FRVUZCd0VCQkdrd1p6QW1CZ2dyDQpCZ0VGQlFjd0FZWWFhSFIwY0RvdkwzTjBZWFIxY3k1eVlYQnBaSE56YkM1amIyMHdQUVlJS3dZQkJRVUhNQUtHDQpNV2gwZEhBNkx5OWpZV05sY25SekxuSmhjR2xrYzNOc0xtTnZiUzlTWVhCcFpGTlRURkpUUVVOQk1qQXhPQzVqDQpjblF3Q1FZRFZSMFRCQUl3QURDQ0FYNEdDaXNHQVFRQjFua0NCQUlFZ2dGdUJJSUJhZ0ZvQUhVQTdrdTl0M1hPDQpZTHJoUW1rZnErR2VacU1QZmwrd2N0aURBTVI3aVhxby9jc0FBQUZzVHEwL013QUFCQU1BUmpCRUFpQlg3NXJ0DQpwNkdQbUQxbVl1S1hvcXVRVytIcUtRNE8rV1RXbTNCbVBteWF1d0lnQUoya01WdmduaUx1dEh5Y3AyYzBwdDRsDQptL3M5SWN2Yi9aN09jMU82b2xJQWRnQ0hkYi9uV1h6NGpFT1pYNzN6YnY5V2pVZFdOdjlLdFdEQnRPci9YcUNEDQpEd0FBQVd4T3JUOXNBQUFFQXdCSE1FVUNJSHphY3NCcmpEN1FER1ZRSHA5YUZPMnJ4VEx4NU9GOVE2MTFYaUU2DQpZUEk3QWlFQW9xd3pTd1BVaGt4bVhYNzN0SDVEcVNwaUxuOHdrMXFWeWhZOTVmSmxZa0VBZHdCRWxHVXVzTzdPDQpyOFJBQjlpby9pakEydWFDdnRqTE1iVS8wek9XdGJhQnFBQUFBV3hPclQ1M0FBQUVBd0JJTUVZQ0lRQ1ZLaHQ5DQpWTWhlNTUvZEwrSGN6Qzk5a2xheG9DcW5zOStRNGVOTDZ2Smd4QUloQUsyZmp4a1RadGw4WFZ4WHEwc2NSTWNJDQpiSktHcGJYdEV0N3dFVzRkZDNGS01BMEdDU3FHU0liM0RRRUJDd1VBQTRJQkFRQm1GcDNCK1g2Y21Cb0FLWkhrDQpIWWxzanhIZk01bDQyZWJtSFdGdlQyTi9aVkZ3MEZ3YndBNExyNzlvd01QdU5FbVNNa2pmVVhxR011Q2VlVTdODQo0MlRwUFdPaGNzVHF6ZkxVa0hhMU9rK011L3RUZHBNcmhROEVqbkhMY3FKVE9MbmRSVy8zcE5IM2xrK2FJejNoDQpDVGFueElRVEE2cUxRNThRakszWmFMVzdqWVVNcVJFcnBGVDlBVXkxMDAxMjlOam5HVkhtYmU1bW1ZbEwvQlhMDQpmdHBJUUJJZnhCNXpHbHpTcjRHT3p5RDBES05YU040RU0yM2pIM1ljeUg4eFQrODU4ZkxSaXQ4SFlOQ3RQbkloDQoyRzR6UjV3emZFVDlnZUh6aWtuY0ZMWUpMMEpKMVIrZGVNRE91bmJWOUJqcmZJb3NNSk53TUV2ejFZczlBWXhHDQpxSFNYDQotLS0tLUVORCBDRVJUSUZJQ0FURS0tLS0tDQotLS0tLUJFR0lOIENFUlRJRklDQVRFLS0tLS0KTUlJRXNUQ0NBNW1nQXdJQkFnSVFDS1dpUnMxTFhJeUQxd0swdTZ0VFNUQU5CZ2txaGtpRzl3MEJBUXNGQURCaApNUXN3Q1FZRFZRUUdFd0pWVXpFVk1CTUdBMVVFQ2hNTVJHbG5hVU5sY25RZ1NXNWpNUmt3RndZRFZRUUxFeEIzCmQzY3VaR2xuYVdObGNuUXVZMjl0TVNBd0hnWURWUVFERXhkRWFXZHBRMlZ5ZENCSGJHOWlZV3dnVW05dmRDQkQKUVRBZUZ3MHhOekV4TURZeE1qSXpNek5hRncweU56RXhNRFl4TWpJek16TmFNRjR4Q3pBSkJnTlZCQVlUQWxWVApNUlV3RXdZRFZRUUtFd3hFYVdkcFEyVnlkQ0JKYm1NeEdUQVhCZ05WQkFzVEVIZDNkeTVrYVdkcFkyVnlkQzVqCmIyMHhIVEFiQmdOVkJBTVRGRkpoY0dsa1UxTk1JRkpUUVNCRFFTQXlNREU0TUlJQklqQU5CZ2txaGtpRzl3MEIKQVFFRkFBT0NBUThBTUlJQkNnS0NBUUVBNVMyb2loRW85bm5wZXpvemlEdHg0V1dMTENsbC9lMHQxRVllbUU1bgorTWdQNXZpYUhMeStWcEhQK25kWDVEMThJTkl1dUFWOHdGcTI2S0Y1VTBXTklaaVFwNm1MdElXalVlV0RQQTI4Ck9leWhUbGo5VExrMmJleXRidEZVNnlwYnBXVWx0bXZZNVY4bmdzcEM3bkZSTkNqcGZuREVEMmtSeUp6Tzh5b0sKTUZ6NEo0SkU4TjdOQTF1SndVRUZNVXZITHMwc2NMb1Baa0tjZXdJUm0xUlYyQXhtRlF4SmtkZjdZTjlQY2traQpmMlhnbTNiNDhCWm4wemYwcVhzU2VHdTg0dWE5Z3d6anpJN3RiVEJqYXlUcFQrL1hwV3VCVnY2ZnZhckk2YmlrCktCODU5T1NHUXV3NzNYWGdldUZ3RVBIVElSb1V0a3p1My9FUStMdHd6bmtrZFFJREFRQUJvNElCWmpDQ0FXSXcKSFFZRFZSME9CQllFRkZQS0YxbjhhOEFESVM4YXJ1U3FxQnlDVnRwMU1COEdBMVVkSXdRWU1CYUFGQVBlVURWVwowVXk3WnZDajRoc2J3NWV5UGRGVk1BNEdBMVVkRHdFQi93UUVBd0lCaGpBZEJnTlZIU1VFRmpBVUJnZ3JCZ0VGCkJRY0RBUVlJS3dZQkJRVUhBd0l3RWdZRFZSMFRBUUgvQkFnd0JnRUIvd0lCQURBMEJnZ3JCZ0VGQlFjQkFRUW8KTUNZd0pBWUlLd1lCQlFVSE1BR0dHR2gwZEhBNkx5OXZZM053TG1ScFoybGpaWEowTG1OdmJUQkNCZ05WSFI4RQpPekE1TURlZ05hQXpoakZvZEhSd09pOHZZM0pzTXk1a2FXZHBZMlZ5ZEM1amIyMHZSR2xuYVVObGNuUkhiRzlpCllXeFNiMjkwUTBFdVkzSnNNR01HQTFVZElBUmNNRm93TndZSllJWklBWWI5YkFFQ01Db3dLQVlJS3dZQkJRVUgKQWdFV0hHaDBkSEJ6T2k4dmQzZDNMbVJwWjJsalpYSjBMbU52YlM5RFVGTXdDd1lKWUlaSUFZYjliQUVCTUFnRwpCbWVCREFFQ0FUQUlCZ1puZ1F3QkFnSXdEUVlKS29aSWh2Y05BUUVMQlFBRGdnRUJBSDRqeC9MS05XNVprbEZjCllXczhFamJtMG55ektlWkMyS09WWVI3UDhnZXZLeXNsV200WG80QlN6S3IyMzVGc0o0YUZ0NnlBaXYxZVkwdFoKL1pOMThiT0dTR1N0b0VjL0pFNG9jSXpyOFA1TWcxMWtSWUhibWdZbnIxUnhla2k1bVNlYjM5REd4VHBKRDRrRwpoczVsWE5vbzRjb25VaWlKd0thcUg3dmgyYmFyeWQ4cE1JU2FnODNKVXF5VkdjMnRXUHBPMDMyOS9DV3Eya3J5CnF2NjZPU01qd3VsVXowZFhmNE9IUWFzUjdDTmZJcis0S1NjYzZBQmxRNVJERjg2UEdlRTZrZHdTUWtGaUIvY1EKeXNOeXEwakVEUVRrZmEycGptdVd0TUNOYkJuaEZYQlllamZ1YkloYVViRXYyRk9RQjNkQ2F2K0ZQZzVlRXZlWApUVnlNbkdvPQotLS0tLUVORCBDRVJUSUZJQ0FURS0tLS0tCg==",
                    "enabled": false,
                    "hostname": "localhost",
                    "key": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQpNSUlFb3dJQkFBS0NBUUVBd0prMFNubm9aUWFtYVptKzFVSXY3UWIwS2pYWXlRMHJKSUhqbGxNN3pYRWFBaC9CCmpWNnUzSHVNbGtOVlJ6RUVGeEtGUm5nSlpXTVlPcy9sWDU1QmtBS0RocXNmaDl0ZVA4ZFVQU1RTVTN6bWRVc0cKU3AySDBqSm1YbkVjbzBac3BiVXJhaGVxaXFJRTJnWDBCWm9iV29YZEdzUTJ1cmJQbHIwaVVKdTFWenhIZGhNLwpnTnFQWEN5bE5YdENrbVI2NHl6SGtEaG9taUUwMkkzME1wZThEcXM5eHd5Sndxek9ydmZwNHlBNFJIZ2toNXJaCi8rUStzRVA2QjJXMEdBdzZzZDNTYytOTytYVUZoVnlFRDhqZkNsTndPVHZodi9keE5IWHZ4WHd0RWNjZTd1TVAKQUxSc0hYdUVtZE5DeTdqb2JvMm5RYU5vNkFDVzRTazlFYlVlcHdJREFRQUJBb0lCQUJDbDdaZzJlM2tvSHdCbgowdEFhWFl3dW1ySi9YeExaK1kyNVVKUHc4KzFwUm1kQ21YZC81THVZVG1tNnZrZGVIRTFtUTNENDlVNzNSdVBtCmpwdzZVTTZDK1IyTjh6YktKcFo2dzNSWmwzaXlCalhEVzNMZHRSemRVRzMvZHd0Q0RpZ3d4ZVR3ME1RVkJ6aFYKWldlS2o2eStmNE5OekxGTkZocWdsTFBGRWVMTlk3WUgvYUdHdTZwaVpERkFMT0liMlFVUUVvVVRkUEJabDMxSwpKL0Y0cHE1RkV6ekQ0aHMwYXJzSlIxeUd4L0NCRWxqQWRSNWE4Y3FHdGhxYjFMdXJTMVcxT0dNUGkzNFdDblJKCnYwTC9zSU9ORCt5cjdjd3ZQaFFzcmI3Tk9NbGxRSDYwUGdyQ0JyOUVwalFiRm1ZN3NCTS9jYUZTWHVDTmZnTzMKT3hTQjF0RUNnWUVBNHpSMTh2aXl3Ykt5aE5xeEhkNzJ1UkxHS2k4SWZhVmhsNGxFeGdsR0pLNHQ0KzYreXhObAp2OFBkdGg3elJjL1pyQUlBUEtBRnZxa1RnVzkrc2llL3JPcE9PUzA2UnVIWVpJMjdpS3ZiL054N29YTm9TM0Y2CitXRlNJSVJqRFo0eHc2SU1iNUJFS0hOVGtKU1pBMXBkaXhYeFgzeDVTM2hCWTlmekl2NGdVVDhDZ1lFQTJRSDAKbnU5REZ0MDNmWmpvcjRMUkZCQS84NEV1dmdBL2RxMkZ4TzVvMjIrdk43Qk9LRW9IWTRuTzhYdFN2L2MzaVFZdApmTUdGZUR2eXdqRnFPU05SS2JxenMza3ZmY2gxNEhHdFltT29DVTF6Smd2VkF4SytWOFF4YWxMbmhlbjZsaTgwCnJKYmtWWUJXYnFHZHU1d2FPM1RyeGsvbHBMejFvSnliK0FQbmNKa0NnWUVBeHd3anRTeEtOc1hkL2VRMXZReWoKcTdvVjM4ajZzRkRkOGNzeGRDbVZ5UGJMZjErZXlQeUNiSUVrMExQR3ZzZDhWR1FnWHZ4MmEvVm1BRVIwZzVHdApLOGpOcE92Sm5XV29sVmRTRDBjNDRBc3RTR0Y1NG5aVm5QcTAzNFg3M1dkVkE2WUpMUWYvbmt5Q3ppeFE3UTRTCjI1SHF0NU91RS84VGttd1ROdm5aeURNQ2dZQlUxV0xmendFWEt4dktTY0NNL0VMbjZRSlhHeVUrdWJJVm52Z3cKYUZNdGxaZVYyYUowTlp4MTJTWEJXVWljS3ZUVzZFZFlBZjgvWlpBZFFsd1lRaW8wZWZ4T0tRM2J5a1hxbkFYLwo5SjE0SllMaVdMWEpJY2F2WU05QWxYQlU5WWMwdHZlTWtPQnFYR1Z0a2xpclF2Y3VUeUxzdjhjeWczd3ljWWEzCmlpOHI0UUtCZ0NaODZtcEZuUlN2aTdpTWw1MUdybDFQZER2YzZVMURRVkI5empTQlVzak1lamNTQXk4WmVJZlQKUDVoSGc3REFaUEhmUGpYLzZxUnpNQW8zWlFlZVBMdTNEYVgwbFBoekNhalF1bkpIbWZvYlA2QjFOUG45T056bQo0dE0wdEU1YTJVVmJuZW5haVZyaDRhSXlJTTFxVTdHT1ArcDNMNi9hNzVnSVQ2SVJRVlloCi0tLS0tRU5EIFJTQSBQUklWQVRFIEtFWS0tLS0tCg==",
                    "sslProtocols": ""
                  },
                  "yugaware": {
                    "health": {
                      "email": "",
                      "password": "",
                      "username": ""
                    },
                    "multiTenant": false,
                    "replicas": 1,
                    "service": {
                      "enabled": true,
                      "ip": "",
                      "type": "LoadBalancer"
                    },
                    "serviceAccount": "yugaware",
                    "storage": "100Gi",
                    "storageClass": "standard"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/yugabytedb/yugabyte-platform-operator@sha256:f227c9e2c625b04160d952349ddba552ac66dae8e7da35f56746f3447c270ce3
          description: Operator for Yugabyte Platform makes it easy to deploy and
            upgrade Yugabyte Platform on OpenShift environments.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/yugabyte-platform-operator-bundle-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/yugabyte-platform-operator-bundle-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/yugabyte/yugabyte-platform-operator
          support: YugabyteDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy and manage Yugabyte Platform as a Kubernetes native
              custom resource.
            displayName: Yugabyte Platform
            kind: YBPlatform
            name: ybplatforms.yugabyte.com
            version: v1alpha1
        description: |2

          Operator for Yugabyte Platform makes it easy to deploy and upgrade Yugabyte Platform on OpenShift environments.
          Use Yugabyte Platform's orchestration and monitoring to manage YugabyteDB universes. Yugabyte Platform can create a YugabyteDB universe with many instances (VMs, pods, machines, etc., provided by IaaS), logically grouped together to form one logical distributed database. Each universe includes one or more clusters. A universe is comprised of one primary cluster and, optionally, one or more read replica clusters. All instances belonging to a cluster run on the same type of cloud provider instance type.

          ### Yugabyte Platform Features
          * **YugabyteDB-as-a-Service on Kubernetes** - Deploy your own YugabyteDB-as-a-Service running on any Kubernetes infrastructure, Yugabyte Platform provides central management for multiple database clusters and Includes 24/7 production support.
          * **Monitoring & Alerting** - Comprehensive monitoring & alerting without any configuration.
          * **Scheduled Backups** - Use Yugabyte Platform to perform a regular scheduled backup of a full YugabyteDB universe or selected tables.

          ### Getting Started
          Here are a few resources for getting started with Yugabyte platform
          * [Quick start guide for Yugabyte Platform](https://docs.yugabyte.com/latest/yugabyte-platform/deploy/) - Install Yugabyte platform on OpenShift.
          * [Quick Start for YugabyteDB](http://docs.yugabyte.com/latest/quick-start/) - install, create a local cluster and read/write from YugabyteDB.
          * [Explore core features](https://docs.yugabyte.com/latest/explore/) - automatic sharding & re-balancing, linear scalability, fault tolerance, tunable reads etc.
          * [Ecosystem integrations](https://docs.yugabyte.com/latest/develop/ecosystem-integrations/) - integrations with Apache Kafka/KSQL, Apache Spark, JanusGraph, KairosDB, Presto and more.
          * [Real world apps](https://docs.yugabyte.com/latest/develop/realworld-apps/) - sample real-world, end-to-end applications built using YugabyteDB.

          ### Help
          Cannot find what you are looking for? Have a question? We love to hear from you - [please file a GitHub issue](https://github.com/YugaByte/yugabyte-db/issues) or talk to us live at [Slack](https://www.yugabyte.com/slack).
        displayName: Yugabyte Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - distributed sql
        links:
        - name: Yugabyte Platform Operator
          url: https://www.yugabyte.com/
        maintainers:
        - email: oss-maintainers@yugabyte.com
          name: Yugabyte
        maturity: alpha
        provider:
          name: Yugabyte Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9211b70edbb93d718428e1c9a992e5fc74a20cb213f8a0592ffe340c362890fe
        - registry.marketplace.redhat.com/rhm/yugabytedb/yugabyte-platform-operator@sha256:f227c9e2c625b04160d952349ddba552ac66dae8e7da35f56746f3447c270ce3
        version: 0.0.6
      entries:
      - name: yugabyte-platform-operator.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: yugabyte-platform-operator-bundle-rhmp
    provider:
      name: Yugabyte Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: cinder-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cinder-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cinder.openstack.org/v1beta1",
                "kind": "Cinder",
                "metadata": {
                  "name": "cinder",
                  "namespace": "openstack"
                },
                "spec": {
                  "cinderAPI": {},
                  "cinderBackup": {},
                  "cinderScheduler": {},
                  "cinderVolumes": {
                    "volume1": {}
                  },
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "cinder",
                  "databaseInstance": "openstack",
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "cinder-secret",
                  "serviceUser": "cinder"
                }
              },
              {
                "apiVersion": "cinder.openstack.org/v1beta1",
                "kind": "CinderAPI",
                "metadata": {
                  "name": "cinderapi-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cinder.openstack.org/v1beta1",
                "kind": "CinderBackup",
                "metadata": {
                  "name": "cinderbackup-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cinder.openstack.org/v1beta1",
                "kind": "CinderScheduler",
                "metadata": {
                  "name": "cinderscheduler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cinder.openstack.org/v1beta1",
                "kind": "CinderVolume",
                "metadata": {
                  "name": "cindervolume-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:43Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CinderAPI is the Schema for the cinderapis API
            displayName: Cinder API
            kind: CinderAPI
            name: cinderapis.cinder.openstack.org
            version: v1beta1
          - description: CinderBackup is the Schema for the cinderbackups API
            displayName: Cinder Backup
            kind: CinderBackup
            name: cinderbackups.cinder.openstack.org
            version: v1beta1
          - description: Cinder is the Schema for the cinders API
            displayName: Cinder
            kind: Cinder
            name: cinders.cinder.openstack.org
            version: v1beta1
          - description: CinderScheduler is the Schema for the cinderschedulers API
            displayName: Cinder Scheduler
            kind: CinderScheduler
            name: cinderschedulers.cinder.openstack.org
            version: v1beta1
          - description: CinderVolume is the Schema for the cindervolumes API
            displayName: Cinder Volume
            kind: CinderVolume
            name: cindervolumes.cinder.openstack.org
            version: v1beta1
        description: Cinder Operator
        displayName: Cinder Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Block Storage
        - Cinder
        links:
        - name: Cinder Operator
          url: https://github.com/openstack-k8s-operators/cinder-operator
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/cinder-rhel9-operator@sha256:e0e457500af9db0831c7080b10e010d11213cd800d89e40683974aaf7966a68d
        - registry.redhat.io/rhoso/openstack-cinder-api-rhel9@sha256:ec6835efb7ffc4fa7eed10bd029440263237e8f792e7e43d4cda337d00fecdb6
        - registry.redhat.io/rhoso/openstack-cinder-backup-rhel9@sha256:936e10d80300c325d0f58b4d6ca290fcf3c5cad774eb5cb696330a20c0151561
        - registry.redhat.io/rhoso/openstack-cinder-scheduler-rhel9@sha256:753775cc575321c727b9a6371daedc8774ba3a2287c96f3d44ad8f7025f63841
        - registry.redhat.io/rhoso/openstack-cinder-volume-rhel9@sha256:47b19d993bdfc790c508861dbcc87f0ec51302db8975698af8e1476a0f5209c7
        version: 1.0.6
      entries:
      - name: cinder-operator.v1.0.6
        version: 1.0.6
      - name: cinder-operator.v1.0.5
        version: 1.0.5
      - name: cinder-operator.v1.0.4
        version: 1.0.4
      - name: cinder-operator.v1.0.3
        version: 1.0.3
      - name: cinder-operator.v1.0.2
        version: 1.0.2
      - name: cinder-operator.v1.0.1
        version: 1.0.1
      - name: cinder-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: cinder-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: volsync-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: volsync-product.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
          createdAt: "2025-08-27T13:12:05Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.13.1'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
        version: 0.13.1
      entries:
      - name: volsync-product.v0.13.1
        version: 0.13.1
      - name: volsync-product.v0.13.0
        version: 0.13.0
      - name: volsync-product.v0.12.2
        version: 0.12.2
      - name: volsync-product.v0.12.1
        version: 0.12.1
      - name: volsync-product.v0.12.1-0.1749677898.p
        version: 0.12.1+0.1749677898.p
      - name: volsync-product.v0.12.0
        version: 0.12.0
      name: stable
    - currentCSV: volsync-product.v0.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:d345825e953c4c170db45a5d92639182b7ee18933287aaeea035f2a2bad13c37
          createdAt: 09 Jan 2025, 00:47
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.10.2'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:d345825e953c4c170db45a5d92639182b7ee18933287aaeea035f2a2bad13c37
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 0.10.2
      entries:
      - name: volsync-product.v0.10.2
        version: 0.10.2
      - name: volsync-product.v0.10.1
        version: 0.10.1
      - name: volsync-product.v0.10.1-0.1730236570.p
        version: 0.10.1+0.1730236570.p
      - name: volsync-product.v0.10.0
        version: 0.10.0
      name: stable-0.10
    - currentCSV: volsync-product.v0.11.2-0.1749674925.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:9da8572c79b75767952cf699e5b1699b9f46d83586aa2d6a17fb23c092713b62
          createdAt: 16 Apr 2025, 06:42
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.11.2'
          olm.substitutesFor: volsync-product.v0.11.2
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:9da8572c79b75767952cf699e5b1699b9f46d83586aa2d6a17fb23c092713b62
        version: 0.11.2+0.1749674925.p
      entries:
      - name: volsync-product.v0.11.2
        version: 0.11.2
      - name: volsync-product.v0.11.2-0.1749674925.p
        version: 0.11.2+0.1749674925.p
      - name: volsync-product.v0.11.1
        version: 0.11.1
      - name: volsync-product.v0.11.0
        version: 0.11.0
      name: stable-0.11
    - currentCSV: volsync-product.v0.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:7d1ac69fe3d0fddc66acdfce9361a4fc8dd3f549f66c98450f7a0e4c5cc893a0
          createdAt: "2025-07-07T13:46:18Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.12.2'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:7d1ac69fe3d0fddc66acdfce9361a4fc8dd3f549f66c98450f7a0e4c5cc893a0
        version: 0.12.2
      entries:
      - name: volsync-product.v0.12.2
        version: 0.12.2
      - name: volsync-product.v0.12.1
        version: 0.12.1
      - name: volsync-product.v0.12.1-0.1749677898.p
        version: 0.12.1+0.1749677898.p
      - name: volsync-product.v0.12.0
        version: 0.12.0
      name: stable-0.12
    - currentCSV: volsync-product.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
          createdAt: "2025-08-27T13:12:05Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.13.1'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
        version: 0.13.1
      entries:
      - name: volsync-product.v0.13.1
        version: 0.13.1
      - name: volsync-product.v0.13.0
        version: 0.13.0
      name: stable-0.13
    - currentCSV: volsync-product.v0.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:ef44675fe70ee9ec8109fb91c6559f0c5518b45219e19175a18a5004e40b65f7
          createdAt: 22 Aug 2024, 22:06
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.9.3'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ReplicationDestination defines the destination for a replicated
              volume
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: ReplicationSource defines the source for a replicated volume
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:ef44675fe70ee9ec8109fb91c6559f0c5518b45219e19175a18a5004e40b65f7
        version: 0.9.3
      entries:
      - name: volsync-product.v0.9.3
        version: 0.9.3
      - name: volsync-product.v0.9.2
        version: 0.9.2
      - name: volsync-product.v0.9.1
        version: 0.9.1
      - name: volsync-product.v0.9.0
        version: 0.9.0
      name: stable-0.9
    defaultChannel: stable
    packageName: volsync-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: deployment-validation-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: deployment-validation-operator.v0.7.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Application Runtime, Monitoring, Security
          certified: "false"
          containerImage: quay.io/deployment-validation-operator/dv-operator:0.7.12
          createdAt: "2025-09-11T00:00:00Z"
          description: The deployment validation operator
          ignore-check.kube-linter.io/minimum-three-replicas: This deployment uses
            1 pod as currently replicating does not replicate metric data causing
            installation issues
          repository: https://github.com/app-sre/deployment-validation-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Deployment Validation Operator (DVO) checks deployments and other resources against a curated collection of best practices.
          These best practices focus mainly on ensuring that the applications are fault-tolerant. DVO reports failed validations via Prometheus metrics.
          If the best-practice check has failed, the metrics will report `1`.
        displayName: Deployment Validation Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: false
          type: MultiNamespace
        keywords:
        - dvo
        links:
        - name: repository
          url: https://github.com/app-sre/deployment-validation-operator
        - name: containerImage
          url: https://quay.io/deployment-validation-operator/dv-operator:0.7.12
        maintainers:
        - email: dvo-owners@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/deployment-validation-operator/dv-operator:0.7.12
        version: 0.7.12
      entries:
      - name: deployment-validation-operator.v0.7.12
        version: 0.7.12
      - name: deployment-validation-operator.v0.7.9
        version: 0.7.9
      - name: deployment-validation-operator.v0.7.8
        version: 0.7.8
      - name: deployment-validation-operator.v0.7.7
        version: 0.7.7
      - name: deployment-validation-operator.v0.7.6
        version: 0.7.6
      - name: deployment-validation-operator.v0.7.5
        version: 0.7.5
      - name: deployment-validation-operator.v0.7.4
        version: 0.7.4
      - name: deployment-validation-operator.v0.7.3
        version: 0.7.3
      - name: deployment-validation-operator.v0.7.2
        version: 0.7.2
      - name: deployment-validation-operator.v0.7.1
        version: 0.7.1
      - name: deployment-validation-operator.v0.7.0
        version: 0.7.0
      - name: deployment-validation-operator.v0.6.0
        version: 0.6.0
      - name: deployment-validation-operator.v0.5.0
        version: 0.5.0
      - name: deployment-validation-operator.v0.4.0
        version: 0.4.0
      - name: deployment-validation-operator.v0.3.0
        version: 0.3.0
      - name: deployment-validation-operator.v0.2.2
        version: 0.2.2
      name: alpha
    defaultChannel: alpha
    packageName: deployment-validation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: ""
    name: gitops-primer
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitops-primer.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "primer.gitops.io/v1alpha1",
                "kind": "Export",
                "metadata": {
                  "name": "primer"
                },
                "spec": {
                  "method": "download"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/migtools/gitops-primer:v0.0.11
          operatorframework.io/suggested-namespace: gitops-primer-system
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Export is the Schema for the exports API
            displayName: Export
            kind: Export
            name: exports.primer.gitops.io
            version: v1alpha1
        description: Export and normalize Kubernetes YAML files for GitOps
        displayName: gitops-primer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - operator
        links:
        - name: Gitops Primer
          url: https://github.com/migtools/gitops-primer
        maturity: alpha
        provider:
          name: Konveyor
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6d57bfd91fac9b68eb72d27226bc297472ceb136c996628b845ecc54a48b31cb
        - quay.io/migtools/gitops-primer:v0.0.11
        - quay.io/migtools/gitops-primer-webhook:v0.0.11
        - quay.io/migtools/gitops-primer-downloader:v0.0.11
        - quay.io/migtools/gitops-primer-export:v0.0.11
        - quay.io/openshift/origin-oauth-proxy:4.9
        version: 0.0.11
      entries:
      - name: gitops-primer.v0.0.11
        version: 0.0.11
      - name: gitops-primer.v0.0.10
        version: 0.0.10
      - name: gitops-primer.v0.0.9
        version: 0.0.9
      - name: gitops-primer.v0.0.8
        version: 0.0.8
      - name: gitops-primer.v0.0.7
        version: 0.0.7
      - name: gitops-primer.v0.0.6
        version: 0.0.6
      - name: gitops-primer.v0.0.5
        version: 0.0.5
      - name: gitops-primer.v0.0.4
        version: 0.0.4
      - name: gitops-primer.v0.0.3
        version: 0.0.3
      - name: gitops-primer.v0.0.2
        version: 0.0.2
      - name: gitops-primer.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: gitops-primer
    provider:
      name: Konveyor
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Craig Trought
      provider-url: ""
    name: prometheus
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometheusoperator.0.56.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "Prometheus",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "replicas": 2,
                  "serviceAccountName": "prometheus-k8s",
                  "serviceMonitorSelector": {},
                  "ruleSelector": {},
                  "podMonitorSelector": {},
                  "probeSelector": {},
                  "alerting": {
                    "alertmanagers": [
                      {
                        "namespace": "monitoring",
                        "name": "alertmanager-main",
                        "port": "web"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "ServiceMonitor",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "operated-prometheus": "true"
                    }
                  },
                  "endpoints": [
                    {
                      "port": "web",
                      "interval": "30s"
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "PodMonitor",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "app": "prometheus"
                    }
                  },
                  "podMetricsEndpoints": [
                    {
                      "port": "web",
                      "interval": "30s"
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "Alertmanager",
                "metadata": {
                  "name": "alertmanager-main"
                },
                "spec": {
                  "replicas": 3,
                  "alertmanagerConfigSelector": {}
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "PrometheusRule",
                "metadata": {
                  "name": "prometheus-example-rules"
                },
                "spec": {
                  "groups": [
                    {
                      "name": "./example.rules",
                      "rules": [
                        {
                          "alert": "ExampleAlert",
                          "expr": "vector(1)"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1alpha1",
                "kind": "AlertmanagerConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "receivers": [
                    {
                      "name": "example"
                    }
                  ],
                  "route": {
                    "receiver": "example"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/prometheus-operator/prometheus-operator:v0.56.3
          createdAt: "2022-05-24T9:40:00Z"
          description: Manage the full lifecycle of configuring and managing Prometheus
            and Alertmanager servers.
          repository: https://github.com/prometheus-operator/prometheus-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A running Prometheus instance
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.coreos.com
            version: v1
          - description: A Prometheus Rule configures groups of sequentially evaluated
              recording and alerting rules.
            displayName: Prometheus Rule
            kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - description: Configures prometheus to monitor a particular k8s service
            displayName: Service Monitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - description: Configures prometheus to monitor a particular pod
            displayName: Pod Monitor
            kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
          - description: Configures an Alertmanager for the namespace
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.coreos.com
            version: v1
          - description: Configures subsections of Alertmanager configuration specific
              to each namespace
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.coreos.com
            version: v1alpha1
          - description: Defines how groups of ingresses and static targets should
              be monitored.
            displayName: Probe
            kind: Probe
            name: probes.monitoring.coreos.com
            version: v1
          - description: Allows recording and alerting rules to be processed across
              multiple Prometheus instances.
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.coreos.com
            version: v1
        description: |
          The Prometheus Operator for Kubernetes provides easy monitoring definitions for Kubernetes services and deployment and management of Prometheus instances.

          Once installed, the Prometheus Operator provides the following features:

          * **Create/Destroy**: Easily launch a Prometheus instance for your Kubernetes namespace, a specific application or team easily using the Operator.
          * **Simple Configuration**: Configure the fundamentals of Prometheus like versions, persistence, retention policies, and replicas from a native Kubernetes resource.
          * **Target Services via Labels**: Automatically generate monitoring target configurations based on familiar Kubernetes label queries; no need to learn a Prometheus specific configuration language.

          ### Other Supported Features

          **High availability**

          Multiple instances are run across failure zones and data is replicated. This keeps your monitoring available during an outage, when you need it most.

          **Updates via automated operations**

          New Prometheus versions are deployed using a rolling update with no downtime, making it easy to stay up to date.

          **Handles the dynamic nature of containers**

          Alerting rules are attached to groups of containers instead of individual instances, which is ideal for the highly dynamic nature of container deployment.
        displayName: Prometheus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - prometheus
        - monitoring
        - tsdb
        - alerting
        - observability
        links:
        - name: Prometheus
          url: https://www.prometheus.io/
        - name: Documentation
          url: https://github.com/prometheus-operator/prometheus-operator/tree/master/Documentation
        - name: Prometheus Operator
          url: https://github.com/prometheus-operator/prometheus-operator
        maintainers:
        - email: k8s@trought.ca
          name: Craig Trought
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Craig Trought
        relatedImages:
        - quay.io/prometheus-operator/prometheus-operator:v0.56.3
        version: 0.56.3
      entries:
      - name: prometheusoperator.0.56.3
        version: 0.56.3
      - name: prometheusoperator.0.47.0
        version: 0.47.0
      name: beta
    defaultChannel: beta
    packageName: prometheus
    provider:
      name: Craig Trought
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC
      provider-url: ""
    name: nspc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nspc-operator.v1.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.nec.com/v1",
                "kind": "NSPC",
                "metadata": {
                  "name": "nspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.connect.redhat.com/nec/nspc-operator@sha256:50fe10f05d8aa7db1b97280e5842bb0b67c6f9bfb01c0d17c5be3cd11b78e853
          createdAt: "2025-07-24T00:00:00Z"
          description: An operator for managing NEC Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: NEC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NSPC is the Schema for the nspcs API
            displayName: NSPC
            kind: NSPC
            name: nspcs.csi.nec.com
            version: v1
        description: |-
          ## About
          NEC Storage Plug-in for Containers (NSPC) is a plugin that integrates NEC Storage V Series into Kubernetes based clusters.
          NSPC provides persistent volume provisioning capabilities from NEC Storage V Series arrays.

          For full documentation, refer to the reference guide.

          ## Requirements

          ### Supported Driver Version

          * NSPC v3.16.0
        displayName: NEC Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - NEC
        - Storage
        - iStorage
        - V Series
        - Container Storage Interface
        - CSI
        maturity: stable
        provider:
          name: NEC
        relatedImages:
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:81d32545fb1b2c319bfeb36a70db4b4d3aea9c51c0574b209b8f64750b32d2f0
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.connect.redhat.com/nec/nspc-operator@sha256:50fe10f05d8aa7db1b97280e5842bb0b67c6f9bfb01c0d17c5be3cd11b78e853
        - registry.connect.redhat.com/nec/nspc-csi-driver@sha256:e50d60e025b3c8ae522a8c5ae78a07f8a624a1210ed7a078c8084efd76ccc646
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        version: 1.16.0
      entries:
      - name: nspc-operator.v1.16.0
        version: 1.16.0
      name: stable
    defaultChannel: stable
    packageName: nspc-operator
    provider:
      name: NEC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: tempo-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tempo-operator.v0.18.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoMonolithic",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "jaegerui": {
                    "enabled": true,
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    }
                  },
                  "storage": {
                    "traces": {
                      "backend": "memory"
                    }
                  }
                }
              },
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoStack",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "resources": {
                    "total": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2Gi"
                      }
                    }
                  },
                  "storage": {
                    "secret": {
                      "name": "my-storage-secret",
                      "type": "s3"
                    }
                  },
                  "storageSize": "1Gi",
                  "template": {
                    "queryFrontend": {
                      "jaegerQuery": {
                        "enabled": true,
                        "ingress": {
                          "type": "route"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/rhosdt/tempo-rhel8-operator@sha256:9bfa5a2ffb09b501f001a4275379210621f06620adcf6454a9327c65538872f5
          createdAt: 22 Sep 2025, 11:15
          description: Create and manage deployments of Tempo, a high-scale distributed
            tracing backend.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.18.0-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-tempo-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/tempo-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TempoMonolithic manages a Tempo deployment in monolithic
              mode.
            displayName: TempoMonolithic
            kind: TempoMonolithic
            name: tempomonolithics.tempo.grafana.com
            version: v1alpha1
          - description: TempoStack manages a Tempo deployment in microservices mode.
            displayName: TempoStack
            kind: TempoStack
            name: tempostacks.tempo.grafana.com
            version: v1alpha1
        description: |-
          Red Hat OpenShift distributed tracing platform based on Tempo. Tempo is an open-source, easy-to-use, and highly scalable distributed tracing backend. It provides observability for microservices architectures by allowing developers to track requests as they flow through distributed systems. Tempo is optimized to handle large volumes of trace data and is designed to be highly performant even under heavy loads.
          It can ingest common open source tracing protocols including Jaeger, Zipkin, and OpenTelemetry and requires only object storage to operate.

          The Tempo Operator supports Tempo deployments in Microservices mode (`TempoStack` CR) and Monolithic mode (`TempoMonolithic` CR).

          ### Core capabilities
          Tempo is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed transaction monitoring
          * Root cause analysis
          * Performance / latency optimization

          ### Operator features
          * **Resource Limits** - Specify overall resource requests and limits in the `TempoStack` CR; the operator assigns fractions of it to each component
          * **AuthN and AuthZ** - Supports OpenID Control (OIDC) and role-based access control (RBAC)
          * **Managed upgrades** - Updating the operator will automatically update all managed Tempo clusters
          * **Multitenancy** - Multiple tenants can send traces to the same Tempo cluster
          * **mTLS** - Communication between the Tempo components can be secured via mTLS
          * **Jaeger UI** - Traces can be visualized in Jaeger UI and exposed via Ingress or OpenShift Route
          * **Observability** - The operator and `TempoStack` operands expose telemetry (metrics, traces) and integrate with Prometheus `ServiceMonitor` and `PrometheusRule`

          ### Before you start
          `TempoStack` requires object storage to store its traces. `TempoMonolithic` can store traces in-memory, in a Persistent Volume and in object storage.
          Please ensure that your system has a compatible object storage solution that is supported, such as OpenShift Data Foundation, Minio, AWS S3, Azure Storage, or Google Cloud Storage.

          ### Examples
          There is a list of examples to help you deploy different configurations in [this GitHub repository](https://github.com/os-observability/redhat-rhosdt-samples).

          ### Support & Troubleshooting
          Tempo Operator is available and supported as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/distributed_tracing) is recommended when installing, configuring, and managing the Operator and instances.
        displayName: Tempo Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tempo
        - tracing
        - observability
        - monitoring
        - database
        links:
        - name: Tempo Operator
          url: https://github.com/grafana/tempo-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6bcfa981a7a75c242a16afa6dab36bf6e3b8a1442fa863875919ad551f48eceb
        - registry.redhat.io/rhosdt/tempo-rhel8@sha256:530a0ef3aaea3f07d6210a683a771c416cea8ce5accac8f5f751dcd7232cb341
        - registry.redhat.io/rhosdt/tempo-query-rhel8@sha256:5b89188bc587a81af3ecdca3caabd6e48356e9ba4f9070860f5160544ce30eb9
        - registry.redhat.io/rhosdt/tempo-rhel8-operator@sha256:9bfa5a2ffb09b501f001a4275379210621f06620adcf6454a9327c65538872f5
        - registry.redhat.io/rhosdt/tempo-jaeger-query-rhel8@sha256:53b62fea39ce211842026e8365d61f579b8e01e8587a1ce639423d72f3b44f45
        - registry.redhat.io/rhosdt/tempo-gateway-rhel8@sha256:095ec83219660bb1955758b94a9431b3add572da4dc2cba61b42b59fa6b4bba0
        - registry.redhat.io/rhosdt/tempo-gateway-opa-rhel8@sha256:22e876837d00371d24feac3f989ead3d4f034400b292f043b18f3839bfde564a
        version: 0.18.0-1
      entries:
      - name: tempo-operator.v0.18.0-1
        version: 0.18.0-1
      - name: tempo-operator.v0.16.0-2
        version: 0.16.0-2
      - name: tempo-operator.v0.16.0-1
        version: 0.16.0-1
      - name: tempo-operator.v0.15.4-1
        version: 0.15.4-1
      - name: tempo-operator.v0.15.3-1
        version: 0.15.3-1
      - name: tempo-operator.v0.14.1-2
        version: 0.14.1-2
      - name: tempo-operator.v0.14.1-1
        version: 0.14.1-1
      - name: tempo-operator.v0.13.0-1-0.1730817419.p
        version: 0.13.0-1+0.1730817419.p
      - name: tempo-operator.v0.13.0-1
        version: 0.13.0-1
      - name: tempo-operator.v0.10.0-8
        version: 0.10.0-8
      - name: tempo-operator.v0.10.0-7
        version: 0.10.0-7
      - name: tempo-operator.v0.10.0-6
        version: 0.10.0-6
      - name: tempo-operator.v0.8.0-3
        version: 0.8.0-3
      - name: tempo-operator.v0.8.0-2
        version: 0.8.0-2
      - name: tempo-operator.v0.6.0-2
        version: 0.6.0-2
      - name: tempo-operator.v0.3.1-3
        version: 0.3.1-3
      - name: tempo-operator.v0.1.0-6
        version: 0.1.0-6
      - name: tempo-operator.v0.1.0-6-0.1687199947.p
        version: 0.1.0-6+0.1687199947.p
      name: stable
    defaultChannel: stable
    packageName: tempo-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: logging-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: logging-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Elasticsearch","metadata":{"name":"elasticsearch"},"spec":{"esClusterName":"prod","esVersion":"7.17.0","esSecurity":{"autoGeneratePassword":true,"tlsEnabled":true}}},{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Fluentd","metadata":{"name":"fluentd"},"spec":{"esCluster":{"host":"elasticsearch-master"},"esSecurity":{"tlsEnabled":true,"existingSecret":"elasticsearch-password"},"indexNameStrategy":"namespace_name"}},{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Kibana","metadata":{"name":"kibana"},"spec":{"replicas":1,"esCluster":{"host":"https://elasticsearch-master:9200","esVersion":"7.16.0","clusterName":"elasticsearch"},"esSecurity":{"tlsEnabled":true,"existingSecret":"elasticsearch-password"}}}]'
          capabilities: Seamless Upgrades
          categories: Database,  Monitoring, Big Data
          certified: "false"
          containerImage: quay.io/opstree/logging-operator:v0.4.0
          createdAt: 04-07-2022
          description: A golang-based operator to create and manage EFK (Elasticsearch,
            Fluentd, and Kibana) stack on Kubernetes.
          k8sMaxVersion: ""
          k8sMinVersion: ""
          repository: https://github.com/ot-container-kit/logging-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Elasticsearch
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Index Life Cycle
            displayName: IndexLifeCycle
            kind: IndexLifeCycle
            name: indexlifecycles.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Index Template
            displayName: IndexTemplate
            kind: IndexTemplate
            name: indextemplates.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Fluentd
            displayName: Fluentd
            kind: Fluentd
            name: fluentds.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Kibana
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.logging.opstreelabs.in
            version: v1beta1
        description: |2+

          Logging Operator is an operator created in Golang to set up and manage EFK(Elasticsearch, Fluentd, and Kibana) cluster inside Kubernetes and Openshift environment. This operator is capable of setting up each individual component of EFK cluster separately.
          ### Supported Features
          - Setup of elasticsearch cluster with different nodes type:- master, data, ingestion, and client.
          - Customizable configuration for elasticsearch, fluentd stack.
          - Setup of fluentd as light-weight log-shipper as Daemonset.
          - Kibana will be set up as visualization tool for elastic stack.
          - Seamless upgrades of elasticsearch, fluentd, and kibana.
          - Security best practices support for complete stack such as TLS, elastic security.
          - Kubernetes resources objects support like:- resources, securityContext, affinity, tolerations, etc.
          - Elasticsearch plugins and keystore seamless support with operator
          ### Documentation
          **[Documentation](https://ot-logging-operator.netlify.app/)**
          ### Getting Help
          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.
          ### Contributing
          You can contribute to this project by:-
          * Raising Bugs and Feature Requests
          * Fixing issue by opening a Pull Request
          * Improving Documentation
          ### License
          MongoDB Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/logging-operator/blob/master/LICENSE)

        displayName: Logging Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Operator
        - Elasticsearch
        - Fluentd
        - Kibana
        - Opstree
        - Logging
        - Monitoring
        - Cloud Native
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-logging-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep
        maturity: beta
        minKubeVersion: 1.17.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/opstree/logging-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: logging-operator.v0.4.0
        version: 0.4.0
      - name: logging-operator.v0.3.0
        version: 0.3.0
      name: beta
    defaultChannel: beta
    packageName: logging-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Google
      provider-url: https://cloud.google.com/alloydb/omni
    name: alloydb-omni-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: alloydb-omni-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "backupPlanRef": "backupplan1",
                  "dbclusterRef": "dbcluster-sample",
                  "manual": true,
                  "physicalBackupSpec": {
                    "backupType": "full"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "BackupPlan",
                "metadata": {
                  "name": "backupplan1"
                },
                "spec": {
                  "backupRetainDays": 14,
                  "backupSchedules": {
                    "full": "0 0 * * 0",
                    "incremental": "0 21 * * *"
                  },
                  "dbclusterRef": "dbcluster-sample",
                  "paused": false
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "clone1"
                },
                "spec": {
                  "clonedDBClusterConfig": {
                    "dbclusterName": "new-dbcluster-sample"
                  },
                  "pointInTime": "2024-02-23T19:59:43Z",
                  "sourceDBCluster": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "DBCluster",
                "metadata": {
                  "name": "dbcluster-sample"
                },
                "spec": {
                  "databaseVersion": "16.3.0",
                  "primarySpec": {
                    "adminUser": {
                      "passwordRef": {
                        "name": "db-pw-dbcluster-sample"
                      }
                    },
                    "resources": {
                      "cpu": 1,
                      "disks": [
                        {
                          "name": "DataDisk",
                          "size": "10Gi"
                        }
                      ],
                      "memory": "5Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "DBInstance",
                "metadata": {
                  "name": "dbcluster-sample-rp-1"
                },
                "spec": {
                  "dbcParent": {
                    "name": "dbcluster-sample"
                  },
                  "instanceType": "ReadPool",
                  "nodeCount": 2,
                  "resources": {
                    "cpu": 2,
                    "disks": [
                      {
                        "name": "DataDisk",
                        "size": "15Gi"
                      }
                    ],
                    "memory": "6Gi"
                  },
                  "schedulingconfig": {
                    "nodeaffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "preference": {
                            "matchExpressions": [
                              {
                                "key": "another-node-label-key",
                                "operator": "In",
                                "values": [
                                  "another-node-label-value"
                                ]
                              }
                            ]
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "podAffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "podAffinityTerm": {
                            "labelSelector": {
                              "matchExpressions": [
                                {
                                  "key": "app",
                                  "operator": "In",
                                  "values": [
                                    "store"
                                  ]
                                }
                              ]
                            },
                            "topologyKey": "kubernetes.io/hostname"
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "podAntiAffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "podAffinityTerm": {
                            "labelSelector": {
                              "matchExpressions": [
                                {
                                  "key": "security",
                                  "operator": "In",
                                  "values": [
                                    "S1"
                                  ]
                                }
                              ]
                            },
                            "topologyKey": "topology.kubernetes.io/zone"
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane",
                        "operator": "Exists"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Failover",
                "metadata": {
                  "name": "failover-sample"
                },
                "spec": {
                  "dbclusterRef": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "PgBouncer",
                "metadata": {
                  "name": "mypgbouncer"
                },
                "spec": {
                  "allowSuperUserAccess": true,
                  "dbclusterRef": "dbcluster-sample",
                  "parameters": {
                    "default_pool_size": "15",
                    "ignore_startup_parameters": "extra_float_digits",
                    "max_client_conn": "800",
                    "max_db_connections": "160",
                    "pool_mode": "transaction"
                  },
                  "podSpec": {
                    "image": "gcr.io/alloydb-omni/operator/g-pgbouncer:1.5.0",
                    "resources": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "replicaCount": 1,
                  "serviceOptions": {
                    "type": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Replication",
                "metadata": {
                  "name": "replication-downstream-sample"
                },
                "spec": {
                  "dbcluster": {
                    "name": "dbcluster-sample"
                  },
                  "downstream": {
                    "control": "setup",
                    "host": "10.10.10.10",
                    "password": {
                      "name": "ha-rep-pw-dbcluster-sample"
                    },
                    "port": 5432,
                    "replicationSlotName": "dbcluster_sample_replication_upstream_sample",
                    "username": "alloydbreplica"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Replication",
                "metadata": {
                  "name": "replication-upstream-sample"
                },
                "spec": {
                  "dbcluster": {
                    "name": "dbcluster-sample"
                  },
                  "upstream": {}
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backup": "backup1",
                  "sourceDBCluster": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Sidecar",
                "metadata": {
                  "name": "sidecar-sample"
                },
                "spec": {
                  "sidecars": [
                    {
                      "args": [
                        "-c",
                        "while [ true ]\ndo\ndate\nset -x\nls -lh /logs/diagnostic\nset +x\ndone\n"
                      ],
                      "command": [
                        "/bin/sh"
                      ],
                      "image": "busybox",
                      "name": "sidecar-sample",
                      "volumeMounts": [
                        {
                          "mountPath": "/logs",
                          "name": "obsdisk"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Switchover",
                "metadata": {
                  "name": "switchover-sample"
                },
                "spec": {
                  "dbclusterRef": "dbcluster-sample",
                  "newPrimary": "aaaa-dbcluster-sample"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: gcr.io/alloydb-omni/operator/fleet-operator:1.5.0
          createdAt: "2025-07-25T04:12:04Z"
          operatorframework.io/suggested-namespace: alloydb-omni-system
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["backuprepositories.alloydbomni.internal.dbadmin.goog",
            "failovers.alloydbomni.internal.dbadmin.goog", "instancebackupplans.alloydbomni.internal.dbadmin.goog","instancebackups.alloydbomni.internal.dbadmin.goog","instancerestores.alloydbomni.internal.dbadmin.goog","instances.alloydbomni.internal.dbadmin.goog","instanceswitchovers.alloydbomni.internal.dbadmin.goog","instanceuserdefinedauthentications.alloydbomni.internal.dbadmin.goog","lrojobs.alloydbomni.internal.dbadmin.goog","replicationconfigs.alloydbomni.internal.dbadmin.goog","sidecars.alloydbomni.internal.dbadmin.goog","createstandbyjobs.alloydbomni.internal.dbadmin.goog","deletestandbyjobs.alloydbomni.internal.dbadmin.goog"]'
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BackupPlan is the Schema for the backupplans API.
            displayName: Backup Plan
            kind: BackupPlan
            name: backupplans.alloydbomni.dbadmin.goog
            version: v1
          - description: Backup is the Schema for the backups API.
            displayName: Backup
            kind: Backup
            name: backups.alloydbomni.dbadmin.goog
            version: v1
          - description: CreateStandbyJob is an internal workflow tracking object.
              Users should not directly interact with this.
            displayName: (Internal) Create Standby Job
            kind: CreateStandbyJob
            name: createstandbyjobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: DBCluster is the Schema for the DBClusters API.
            displayName: DB Cluster
            kind: DBCluster
            name: dbclusters.alloydbomni.dbadmin.goog
            version: v1
          - description: DBInstance is the Schema for the DBInstances API.
            displayName: DB Instance
            kind: DBInstance
            name: dbinstances.alloydbomni.dbadmin.goog
            version: v1
          - description: DeleteStandbyJob is an internal workflow tracking object.
              Users should not directly interact with this.
            displayName: (Internal) Delete Standby Job
            kind: DeleteStandbyJob
            name: deletestandbyjobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Failover represents the parameters and status of a single
              failover operation.
            displayName: Failover
            kind: Failover
            name: failovers.alloydbomni.dbadmin.goog
            version: v1
          - description: PgBouncer is the Schema for the PgBouncer API.
            displayName: PgBouncer
            kind: PgBouncer
            name: pgbouncers.alloydbomni.dbadmin.goog
            version: v1
          - description: Replication is the Schema for the Replications API.
            displayName: Replication
            kind: Replication
            name: replications.alloydbomni.dbadmin.goog
            version: v1
          - description: Restore is the Schema for the Restores API.
            displayName: Restore
            kind: Restore
            name: restores.alloydbomni.dbadmin.goog
            version: v1
          - description: Sidecar is the Schema for the Sidecar API.
            displayName: Sidecar
            kind: Sidecar
            name: sidecars.alloydbomni.dbadmin.goog
            version: v1
          - description: Switchover represents the parameters and status of a single
              switchover.
            displayName: Switchover
            kind: Switchover
            name: switchovers.alloydbomni.dbadmin.goog
            version: v1
          - description: UserDefinedAuthentication is the Schema for the UserDefinedAuthentications
              API.
            displayName: User Defined Authentication
            kind: UserDefinedAuthentication
            name: userdefinedauthentications.alloydbomni.dbadmin.goog
            version: v1
          - description: BackupRepository is the Schema for the backuprepositories
              API.
            displayName: (Internal) Backup Repository
            kind: BackupRepository
            name: backuprepositories.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Failover is the Schema for the failover API.
            displayName: (Internal) Failover
            kind: Failover
            name: failovers.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceBackupPlan is the Schema for the InstanceBackupPlan
              API.
            displayName: (Internal) Instance Backup Plan
            kind: InstanceBackupPlan
            name: instancebackupplans.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceBackup is the Schema for the InstanceBackup API.
            displayName: (Internal) Instance Backup
            kind: InstanceBackup
            name: instancebackups.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceRestore is the Schema for the InstanceRestore API.
            displayName: (Internal) Instance Restore
            kind: InstanceRestore
            name: instancerestores.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Instance is the Schema for the Instance API.
            displayName: (Internal) Instance
            kind: Instance
            name: instances.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceSwitchover is the Schema for the InstanceSwitchover
              API.
            displayName: (Internal) Instance Switchover
            kind: InstanceSwitchover
            name: instanceswitchovers.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceUserDefinedAuthentication is the Schema for the InstanceUserDefinedAuthentication
              API.
            displayName: (Internal) Instance User Defined Authentication
            kind: InstanceUserDefinedAuthentication
            name: instanceuserdefinedauthentications.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: LROJob is the Schema for the LROJob API.
            displayName: (Internal) LRO Job
            kind: LROJob
            name: lrojobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: ReplicationConfig is the Schema for the ReplicationConfig
              API.
            displayName: (Internal) Replication Config
            kind: ReplicationConfig
            name: replicationconfigs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Sidecar is the Schema for the Sidecar API.
            displayName: (Internal) Sidecar
            kind: Sidecar
            name: sidecars.alloydbomni.internal.dbadmin.goog
            version: v1
        description: |-
          AlloyDB Omni is a database that lets you deploy a streamlined edition of AlloyDB for PostgreSQL in your own computing environment. While being 100% compatible with PostgreSQL applications and tools, AlloyDB Omni delivers higher performance for both OLTP and OLAP workloads and makes administration easy with a number of autopilots. AlloyDB AI provides powerful tools to build generative AI applications with the ability to create, store, and index vector embeddings right in your database.

          ## Post-installation

          After installing the operator, you need to manually create the cert-manager resources in your cluster as this is a strict dependency for our operator. For this, follow these steps:

          ```
          kubectl create ns ${NAMESPACE:?}
          ```

          ```
          kubectl apply -f - <<EOF
          apiVersion: cert-manager.io/v1
          kind: ClusterIssuer
          metadata:
            name: alloydbomni-selfsigned-cluster-issuer
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Issuer
          metadata:
            name: fleet-selfsigned-issuer
            namespace: ${NAMESPACE:?}
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Certificate
          metadata:
            name: fleet-serving-cert
            namespace: ${NAMESPACE:?}
          spec:
            dnsNames:
            - fleet-webhook-service.alloydb-omni-system.svc
            - fleet-webhook-service.alloydb-omni-system.svc.cluster.local
            issuerRef:
              kind: Issuer
              name: fleet-selfsigned-issuer
            secretName: fleet-webhook-server-cert
          ---
          apiVersion: cert-manager.io/v1
          kind: Issuer
          metadata:
            name: local-selfsigned-issuer
            namespace: ${NAMESPACE:?}
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Certificate
          metadata:
            name: local-serving-cert
            namespace: ${NAMESPACE:?}
          spec:
            dnsNames:
            - local-webhook-service.alloydb-omni-system.svc
            - local-webhook-service.alloydb-omni-system.svc.cluster.local
            issuerRef:
              kind: Issuer
              name: local-selfsigned-issuer
            secretName: local-webhook-server-cert
          EOF
          ```

          Replace the following:
          * `NAMESPACE` : the namespace where you have installed your operator. Suggested namespace: `alloydb-omni-system`.

          ## Creating a database cluster

          After you install the AlloyDB Omni operator on your Kubernetes cluster, you can create an AlloyDB Omni database cluster. For each database cluster you create, you must first create a password secret using the following format:

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: db-pw-${DB_CLUSTER_NAME:s?}
            namespace: ${DB_CLUSTER_NAMESPACE:?}
          type: Opaque
          data:
            ${DB_CLUSTER_NAME:?}: "ENCODED_PASSWORD"
          ```

          Replace the following:
          * `DB_CLUSTER_NAME`: the name of this database cluster. For example, `my-db-cluster`.
          * `DB_CLUSTER_NAMESPACE` (Optional): the namespace where you want to create the database cluster. For example, `my-db-cluster-namespace`.
          * `ENCODED_PASSWORD`: the database login password for the default postgres user role, encoded as a base64 string. For example, `Q2hhbmdlTWUxMjM=` for password `ChangeMe123`.

          Follow [these][database-cluster] steps for creating a database cluster.

          ## Connect to AlloyDB Omni running on Kubernetes
          Follow [these][connect] steps to connect to AlloyDB Omni.

          [database-cluster]: https://cloud.google.com/alloydb/omni/docs/deploy-kubernetes#create
          [connect]: https://cloud.google.com/alloydb/omni/docs/run-connect#connect-kubernetes
        displayName: AlloyDB Omni Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - sql
        - postgres
        - vector
        - ai
        - analytics
        - postgresql
        - alloydb
        links:
        - name: Documentation
          url: https://cloud.google.com/alloydb/omni/docs
        - name: Release Notes
          url: https://cloud.google.com/alloydb/docs/release-notes
        maintainers:
        - email: alloydb-omni-ext@google.com
          name: Google Cloud
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Google
          url: https://cloud.google.com/alloydb/omni
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.1
        - gcr.io/alloydb-omni/operator/local-operator:1.5.0
        - gcr.io/alloydb-omni/operator/fleet-operator:1.5.0
        version: 1.5.0
      entries:
      - name: alloydb-omni-operator.v1.5.0
        version: 1.5.0
      - name: alloydb-omni-operator.v1.4.1
        version: 1.4.1
      - name: alloydb-omni-operator.v1.4.0
        version: 1.4.0
      - name: alloydb-omni-operator.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: alloydb-omni-operator
    provider:
      name: Google
      url: https://cloud.google.com/alloydb/omni
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Trilio
      provider-url: https://trilio.io/
    name: tvo-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tvo-operator.6.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOControlPlane",
                "metadata": {
                  "name": "tvocontrolplane-sample"
                },
                "spec": {
                  "crdType": "TVOControlPlane",
                  "common": {
                    "dmapi_workers": "16",
                    "memcached_servers": "",
                    "my_ip": {
                      "host_interface": null
                    },
                    "nova_group_id": "42436",
                    "nova_user_id": "42436",
                    "oslo_enable_proxy_headers_parsing": false,
                    "trustee_role": "creator"
                  },
                  "database": {
                    "common": {
                      "host": "EDIT_DB_HOSTNAME",
                      "root_password": "EDIT_PASSWORD",
                      "root_user_name": "root"
                    },
                    "datamover_api": {
                      "database": "dmapi",
                      "password": "EDIT_PASSWORD",
                      "user": "dmapi"
                    },
                    "wlm_api": {
                      "database": "workloadmgr",
                      "password": "EDIT_PASSWORD",
                      "user": "workloadmgr"
                    }
                  },
                  "helm3_hook": true,
                  "images": {
                    "triliovault_datamover_api": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_datamover_db_sync": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_cloud_trust": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_cron": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_db_sync": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_scheduler": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_workloads": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1"
                  },
                  "keystone": {
                    "common": {
                      "admin_role_name": "admin",
                      "auth_uri": "",
                      "auth_url": "",
                      "ca_cert": "CERT MULTIPLE LINES\n",
                      "cloud_admin_domain_name": "Default",
                      "cloud_admin_project_name": "admin",
                      "cloud_admin_user_name": "admin",
                      "cloud_admin_user_password": "",
                      "is_self_signed_ssl_cert": true,
                      "keystone_auth_host": "",
                      "keystone_auth_port": "",
                      "keystone_auth_protocol": "",
                      "region_name": "regionOne",
                      "service_project_domain_name": "Default",
                      "service_project_name": "service"
                    },
                    "datamover_api": {
                      "internal_endpoint": "ENDPOINT_PROTOCOL_INTERNAL://ENDPOINT_HOST_INTERNAL:8784/v2",
                      "password": "EDIT_PASSWORD",
                      "public_auth_host": "",
                      "public_endpoint": "ENDPOINT_PROTOCOL_PUBLIC://ENDPOINT_HOST_PUBLIC:8784/v2",
                      "service_desc": "TrilioVault Datamover Api Service",
                      "service_name": "dmapi",
                      "service_type": "datamover",
                      "user": "dmapi"
                    },
                    "identity": null,
                    "keystone_interface": "internal",
                    "wlm_api": {
                      "internal_endpoint": "ENDPOINT_PROTOCOL_INTERNAL://ENDPOINT_HOST_INTERNAL:8781/v1/$(tenant_id)s",
                      "password": "EDIT_PASSWORD",
                      "public_auth_host": "",
                      "public_endpoint": "ENDPOINT_PROTOCOL_PUBLIC://ENDPOINT_HOST_PUBLIC:8781/v1/$(tenant_id)s",
                      "service_desc": "TrilioVault Workloadmgr Api Service",
                      "service_name": "TrilioVaultWLM",
                      "service_type": "workloads",
                      "user": "triliovault"
                    }
                  },
                  "labels": {
                    "datamover_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_cron": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_scheduler": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_workloads": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    }
                  },
                  "pod": {
                    "replicas": {
                      "triliovault_datamover_api": 3,
                      "triliovault_object_store": 3,
                      "triliovault_wlm_api": 3,
                      "triliovault_wlm_cron": 1,
                      "triliovault_wlm_scheduler": 3,
                      "triliovault_wlm_workloads": 3
                    },
                    "security_context": {
                      "triliovault": {
                        "container": {
                          "triliovault_datamover_api": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 0,
                            "runAsUser": 63630
                          },
                          "triliovault_datamover_api_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 63630,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_api": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_cron": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_scheduler": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_workloads": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "wlm_cloud_trust_creation": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 42436
                          }
                        }
                      }
                    }
                  },
                  "rabbitmq": {
                    "common": {
                      "admin_password": "EDIT_PASSWORD",
                      "admin_user": "admin",
                      "driver": "messagingv2",
                      "host": "EDIT_HOST",
                      "port": "5672"
                    },
                    "datamover_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "",
                      "user": "dmapi",
                      "vhost": "dmapi"
                    },
                    "wlm_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "",
                      "user": "workloadmgr",
                      "vhost": "workloadmgr"
                    }
                  },
                  "triliovault_backup_targets": [
                    {
                      "backup_target_name": "S3_BT1",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY1",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET1",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL1",
                      "s3_region_name": "REGION1",
                      "s3_secret_key": "SECRETKEY1",
                      "s3_self_signed_cert": false,
                      "s3_signature_version": "default",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "S3_BT2",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY2",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET2",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL2",
                      "s3_region_name": "REGION2",
                      "s3_secret_key": "SECRETKEY2",
                      "s3_self_signed_cert": true,
                      "s3_signature_version": "default",
                      "s3_ssl_ca_cert": "SSL_CERT_LINE_1\n",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "NFS_BT3",
                      "backup_target_type": "nfs",
                      "is_default": false,
                      "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10",
                      "nfs_shares": "11.30.1.10:/mnt/share"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:3d80cad27a1043fb61735e7a6fb156c98b269cfcc68f8b9712713a7d4f6ad6bd
          createdAt: "2025-01-14T11:20:30Z"
          description: Cloud-Native Data Protection for OpenStack
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trilio-openstack
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trilioData/triliovault-cfg-scripts
          support: Trilio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trilio for OpenStack Control Plane. Includes all control
              plane components like Datamover Api, Trilio object store and Workloadmgr
              services.
            displayName: Trilio for OpenStack Control Plane
            kind: TVOControlPlane
            name: tvocontrolplanes.tvo.trilio.io
            version: v1
          - description: Trilio for OpenStack Backup Target. This CR is responsible
              to add a new backup target.
            displayName: Trilio for OpenStack Backup Target
            kind: TVOBackupTarget
            name: tvobackuptargets.tvo.trilio.io
            version: v1
        description: |-
          Trilio for OpenStack provides scalable, native, agentless, multi-tenant backup and recovery solutions, ensuring seamless protection of workloads with flexible storage options for dynamic cloud environments.

            * OpenStack Backup and Recovery: Recover any OpenStack application quickly in the event of data loss or misconfiguration. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Such as One-Click-Restore to entire project recovery on different clusters.
            * Disaster Recovery: Recover any OpenStack application quickly in the event of a disaster. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Trilio makes OpenStack disaster recovery easy.

            * Migration: A tenant or administrator can capture an application and its data and non-disruptively migrate that exact point-in-time to another OpenStack cloud. The need to migrate OpenStack workloads can be driven by economics, security or test/dev.

            * Network Functions Virtualization backup and recovery: Business critical services like NFV can quickly be configured to any state, assuring the optimal quality of service. Trilio provides recovery from data damage and allows restoration of entire VNFs, selected VNFs, or individual data items using a point-in-time working copy. The platform also allows recovery from a disaster by restoring the OpenStack network topology and populating it with a working copy of the entire NFV.

          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Backup
        - Recovery
        - RHOSO Backup
        - Disaster Recovery
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/openstack-backup-and-recovery/
        - name: Operator Documentation
          url: https://docs.trilio.io/openstack
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        provider:
          name: Trilio
          url: https://trilio.io/
        relatedImages:
        - registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82
        - registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1
        - registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:3d80cad27a1043fb61735e7a6fb156c98b269cfcc68f8b9712713a7d4f6ad6bd
        version: 6.1.1
      entries:
      - name: tvo-operator.6.1.1
        version: 6.1.1
      name: 6.0.x
    - currentCSV: tvo-operator.6.1.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOControlPlane",
                "metadata": {
                  "name": "tvocontrolplane-sample",
                  "namespace": "trilio-openstack"
                },
                "spec": {
                  "common": {
                    "dmapi_workers": "16",
                    "memcached_servers": "memcached-0.memcached.openstack.svc:11212",
                    "nova_group_id": "42436",
                    "nova_user_id": "42436",
                    "oslo_enable_proxy_headers_parsing": false,
                    "trustee_role": "member"
                  },
                  "crdType": "TVOControlPlane",
                  "database": {
                    "common": {
                      "host": "openstack.openstack.svc",
                      "port": 3306,
                      "root_password": "openstack",
                      "root_user_name": "root"
                    },
                    "datamover_api": {
                      "database": "dmapi",
                      "password": "dsw21aSyH5gHH6XE72k11Y3t3C308jY5rP02ndwJW",
                      "user": "dmapi"
                    },
                    "wlm_api": {
                      "database": "workloadmgr",
                      "password": "dfeeYb72OwUzX5YuRKcVjxXJaELV8YT0jZczIQwys0v",
                      "user": "workloadmgr"
                    }
                  },
                  "helm3_hook": true,
                  "images": {
                    "triliovault_datamover_api": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_datamover_db_sync": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_cloud_trust": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_cron": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_db_sync": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_scheduler": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_workloads": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd"
                  },
                  "keystone": {
                    "common": {
                      "admin_role_name": "admin",
                      "auth_uri": "https://keystone-internal.openstack.svc:5000/v3",
                      "auth_url": "https://keystone-internal.openstack.svc:5000",
                      "cloud_admin_domain_name": "Default",
                      "cloud_admin_project_name": "admin",
                      "cloud_admin_user_name": "admin",
                      "cloud_admin_user_password": "openstack",
                      "is_self_signed_ssl_cert": true,
                      "keystone_auth_host": "keystone-internal.openstack.svc",
                      "keystone_auth_port": "5000",
                      "keystone_auth_protocol": "https",
                      "region_name": "regionOne",
                      "service_project_domain_name": "Default",
                      "service_project_name": "service"
                    },
                    "datamover_api": {
                      "internal_endpoint": "https://triliovault-datamover-internal.trilio-openstack.svc:8784/v2",
                      "password": "defajgh48zgt88wFzzhwEt3xHGeCOdExYbLGQN2XRm",
                      "public_auth_host": "triliovault-datamover-public-trilio-openstack.apps.trilio.trilio.bos2",
                      "public_endpoint": "https://triliovault-datamover-public-trilio-openstack.apps.trilio.trilio.bos2/v2",
                      "service_desc": "TrilioVault Datamover Api Service",
                      "service_name": "dmapi",
                      "service_type": "datamover",
                      "user": "dmapi"
                    },
                    "keystone_interface": "internal",
                    "wlm_api": {
                      "internal_endpoint": "https://triliovault-wlm-internal.trilio-openstack.svc:8781/v1/$(tenant_id)s",
                      "password": "dsdeJl8axy0gv14fURwUUxy0qUOQbugUUOfXikz6j",
                      "public_auth_host": "triliovault-wlm-public-trilio-openstack.apps.trilio.trilio.bos2",
                      "public_endpoint": "https://triliovault-wlm-public-trilio-openstack.apps.trilio.trilio.bos2/v1/$(tenant_id)s",
                      "service_desc": "TrilioVault Workloadmgr Api Service",
                      "service_name": "TrilioVaultWLM",
                      "service_type": "workloads",
                      "user": "triliovault"
                    }
                  },
                  "labels": {
                    "datamover_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_cron": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_scheduler": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_workloads": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    }
                  },
                  "pod": {
                    "replicas": {
                      "triliovault_datamover_api": 3,
                      "triliovault_object_store": 3,
                      "triliovault_wlm_api": 3,
                      "triliovault_wlm_cron": 1,
                      "triliovault_wlm_scheduler": 3,
                      "triliovault_wlm_workloads": 3
                    },
                    "security_context": {
                      "triliovault": {
                        "container": {
                          "triliovault_datamover_api": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 0,
                            "runAsUser": 63630
                          },
                          "triliovault_datamover_api_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 63630,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_api": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_cron": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_scheduler": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_workloads": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "wlm_cloud_trust_creation": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 42436
                          }
                        }
                      }
                    }
                  },
                  "rabbitmq": {
                    "common": {
                      "admin_password": "EDIT_PASSWORD",
                      "admin_user": "admin",
                      "driver": "messagingv2",
                      "host": "EDIT_HOST",
                      "port": "15671",
                      "ssl": true
                    },
                    "datamover_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "rabbit://dmapi:${PASSWORD}@${RABBITMQ_HOST}:5671/dmapi?ssl=1",
                      "user": "dmapi",
                      "vhost": "dmapi"
                    },
                    "wlm_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "rabbit://workloadmgr:${PASSWORD}@${RABBITMQ_HOST}:5671/workloadmgr?ssl=1",
                      "user": "workloadmgr",
                      "vhost": "workloadmgr"
                    }
                  },
                  "triliovault_backup_targets": [
                    {
                      "backup_target_name": "BT1_S3",
                      "backup_target_type": "s3",
                      "is_default": false,
                      "s3_access_key": "",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "",
                      "s3_region_name": "",
                      "s3_secret_key": "",
                      "s3_self_signed_cert": false,
                      "s3_signature_version": "default",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "BT2_Other_S3",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY2",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET2",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL2",
                      "s3_region_name": "REGION2",
                      "s3_secret_key": "SECRETKEY2",
                      "s3_self_signed_cert": true,
                      "s3_signature_version": "default",
                      "s3_ssl_ca_cert": "SSL_CERT_LINE_1\nSSL_CERT_LINE_2\nSSL_CERT_LINE_N\n",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "other_s3"
                    },
                    {
                      "backup_target_name": "BT3_NFS",
                      "backup_target_type": "nfs",
                      "is_default": false,
                      "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10",
                      "nfs_shares": "192.168.2.3:/nfs/share"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOBackupTarget",
                "metadata": {
                  "name": "tvobackuptarget-sample",
                  "namespace": "trilio-openstack"
                },
                "spec": {
                  "crdType": "TVOBackupTarget",
                  "images": {
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd"
                  },
                  "triliovault_backup_target": {
                    "backup_target_name": "S3_BT4",
                    "backup_target_type": "s3",
                    "is_default": false,
                    "s3_type": "other_s3",
                    "s3_access_key": "",
                    "s3_secret_key": "",
                    "s3_region_name": "",
                    "s3_bucket": "",
                    "s3_endpoint_url": "",
                    "s3_signature_version": "default",
                    "s3_auth_version": "DEFAULT",
                    "s3_bucket_object_lock_enabled": false,
                    "s3_ssl_enabled": true,
                    "s3_ssl_verify": true,
                    "s3_self_signed_cert": true,
                    "s3_ssl_ca_cert": "SSL_CERT_LINE_1\nSSL_CERT_LINE_2\nSSL_CERT_LINE_N",
                    "nfs_shares": "192.168.2.30:/nfs/share",
                    "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:e282bb056a1780e5170db3e63efdf5099a5ac5240ad44162bc7af83cac893ac7
          createdAt: "2025-01-14T11:20:30Z"
          description: Cloud-Native Data Protection for OpenStack
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: tvo-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trilioData/triliovault-cfg-scripts
          support: Trilio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trilio for OpenStack Control Plane. Includes all control
              plane components like Datamover Api, Trilio object store and Workloadmgr
              services.
            displayName: Trilio for OpenStack Control Plane
            kind: TVOControlPlane
            name: tvocontrolplanes.tvo.trilio.io
            version: v1
          - description: Trilio for OpenStack Backup Target. This CR is responsible
              to add a new backup target.
            displayName: Trilio for OpenStack Backup Target
            kind: TVOBackupTarget
            name: tvobackuptargets.tvo.trilio.io
            version: v1
        description: |-
          Trilio for OpenStack provides scalable, native, agentless, multi-tenant backup and recovery solutions, ensuring seamless protection of workloads with flexible storage options for dynamic cloud environments.

            * OpenStack Backup and Recovery: Recover any OpenStack application quickly in the event of data loss or misconfiguration. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Such as One-Click-Restore to entire project recovery on different clusters.
            * Disaster Recovery: Recover any OpenStack application quickly in the event of a disaster. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Trilio makes OpenStack disaster recovery easy.

            * Migration: A tenant or administrator can capture an application and its data and non-disruptively migrate that exact point-in-time to another OpenStack cloud. The need to migrate OpenStack workloads can be driven by economics, security or test/dev.

            * Network Functions Virtualization backup and recovery: Business critical services like NFV can quickly be configured to any state, assuring the optimal quality of service. Trilio provides recovery from data damage and allows restoration of entire VNFs, selected VNFs, or individual data items using a point-in-time working copy. The platform also allows recovery from a disaster by restoring the OpenStack network topology and populating it with a working copy of the entire NFV.

          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Backup
        - Recovery
        - RHOSO Backup
        - Disaster Recovery
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/openstack-backup-and-recovery/
        - name: Operator Documentation
          url: https://docs.trilio.io/openstack
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        provider:
          name: Trilio
          url: https://trilio.io/
        relatedImages:
        - registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd
        - registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:e282bb056a1780e5170db3e63efdf5099a5ac5240ad44162bc7af83cac893ac7
        - registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82
        version: 6.1.9
      entries:
      - name: tvo-operator.6.1.9
        version: 6.1.9
      name: 6.1.x
    defaultChannel: 6.0.x
    packageName: tvo-operator
    provider:
      name: Trilio
      url: https://trilio.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: elasticsearch-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: elasticsearch-operator.v5.8.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Elasticsearch",
                "metadata": {
                  "name": "elasticsearch"
                },
                "spec": {
                  "indexManagement": {
                    "mappings": [
                      {
                        "aliases": [
                          "infra",
                          "logs.infra"
                        ],
                        "name": "infra",
                        "policyRef": "infra-policy"
                      }
                    ],
                    "policies": [
                      {
                        "name": "infra-policy",
                        "phases": {
                          "delete": {
                            "minAge": "2d",
                            "namespaceSpec": [
                              {
                                "minAge": "5h",
                                "namespace": "openshift-monitoring"
                              }
                            ],
                            "pruneNamespacesInterval": "24h"
                          },
                          "hot": {
                            "actions": {
                              "rollover": {
                                "maxAge": "8h"
                              }
                            }
                          }
                        },
                        "pollInterval": "30m"
                      }
                    ]
                  },
                  "managementState": "Managed",
                  "nodeSpec": {
                    "resources": {
                      "limits": {
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "512Mi"
                      }
                    }
                  },
                  "nodes": [
                    {
                      "nodeCount": 1,
                      "roles": [
                        "client",
                        "data",
                        "master"
                      ],
                      "storage": {
                        "size": "20G"
                      }
                    }
                  ],
                  "redundancyPolicy": "ZeroRedundancy"
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Kibana",
                "metadata": {
                  "name": "kibana"
                },
                "spec": {
                  "managementState": "Managed",
                  "nodeSelector": {},
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "memory": "512Mi"
                    },
                    "requests": {
                      "memory": "512Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
          createdAt: "2024-09-19T15:14:33Z"
          description: |
            The Elasticsearch Operator for OCP provides a means for configuring and managing an Elasticsearch cluster for tracing and cluster logging.
            ## Prerequisites and Requirements
            ### Elasticsearch Operator Namespace
            The Elasticsearch Operator must be deployed to the global operator group namespace
            ### Memory Considerations
            Elasticsearch is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Elasticsearch cluster.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory. Each ES node can operate with a
            lower memory setting though this is not recommended for production deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.21'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Cluster Logging, Jaeger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Elasticsearch cluster instance
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.openshift.io
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.openshift.io
            version: v1
        description: "### IN OCP 4.16 AND FORWARD \n\nELASTICSEARCH IS ONLY SUPPORTED
          FOR **SERVICEMESH, TRACING, AND KIALI**, AND THIS OPERATOR **WILL BE REMOVED
          FROM THE OPENSHIFT OPERATOR CATALOG IN NOVEMBER 2025**.  ELASTICSEARCH IS
          **NO LONGER SUPPORTED FOR LOG STORAGE**. **KIBANA IS NO LONGER SUPPORTED**
          IN OCP 4.16 AND FORWARD. \n\n### About \n\nThe Elasticsearch Operator for
          OCP provides a means for configuring and managing an Elasticsearch cluster
          for use in tracing \nand cluster logging as well as a Kibana instance to
          connect to it.\nThis operator only supports OCP Cluster Logging and Jaeger.
          \ It is tightly coupled to each and is not currently capable of\nbeing used
          as a general purpose manager of Elasticsearch clusters running on OCP.\n\nPlease
          note: For a general purpose Elasticsearch operator, please use Elastic's
          Elasticsearch (ECK) Operator [here](https://catalog.redhat.com/software/containers/elastic/eck-operator/5fabf6d1ecb52450895164be?container-tabs=gti)\n\nIt
          is recommended that this operator be installed in the `openshift-operators-redhat`
          namespace to \nproperly support the Cluster Logging and Jaeger use cases.\n\nOnce
          installed, the operator provides the following features for **Elasticsearch**:\n*
          **Create/Destroy**: Deploy an Elasticsearch cluster to the same namespace
          in which the elasticsearch CR is created.\n* **Update**: Changes to the
          elasticsearch CR will be scheduled and applied to the cluster in a controlled
          manner (most often as a rolling upgrade).\n* **Cluster health**: The operator
          will periodically poll the cluster to evaluate its current health (such
          as the number of active shards and if any cluster nodes have reached their
          storage watermark usage).\n* **Redeploys**: In the case where the provided
          secrets are updated, the Elasticsearch Operator will schedule and perform
          a full cluster restart.\n* **Index management**: The Elasticsearch Operator
          will create cronjobs to perform index management such as roll over and deletion.\n\nOnce
          installed, the operator provides the following features for **Kibana**:\n*
          **Create/Destroy**: Deploy a Kibana instance to the same namespace in which
          the kibana CR is created (this should be the same namespace as the elasticsearch
          CR).\n* **Update**: Changes to the kibana CR will be scheduled and applied
          to the cluster in a controlled manner.\n* **Redeploys**: In the case where
          the provided secrets are updated, the Elasticsearch Operator will perform
          a restart.\n\n### Additionally provided features\n* Out of the box multitenancy
          that is integrated with OCP user access control.\n* Document Level Security\n*
          mTLS communication between Elasticsearch, Kibana, Index Management cronjobs,
          and CLO's Fluentd\n* OCP prometheus dashboard for Elasticsearch clusters\n*
          Prometheus Alerting rules  \n"
        displayName: OpenShift Elasticsearch Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - jaeger
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Elasticsearch Operator
          url: https://github.com/openshift/elasticsearch-operator
        maintainers:
        - email: aos-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8d603100978c655020a8f4d81c527162348f468d93989edbc00463b8297fb8a5
        - registry.redhat.io/openshift-logging/kibana6-rhel8@sha256:a2c200667738fdb079ba11ed3b0df3c5901373428d5e560bc1efea9d5e1ddd67
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
        - registry.redhat.io/openshift-logging/logging-curator5-rhel9@sha256:394e0f46fb8be44c720b29aed716443747a636d2758226c9386f930f95758470
        - registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:b26d2218aa6b1efea83a1717d62f81009c97e8082434b918937c609f5cdab912
        version: 5.8.21
      entries:
      - name: elasticsearch-operator.v5.8.21
        version: 5.8.21
      - name: elasticsearch-operator.v5.8.20
        version: 5.8.20
      - name: elasticsearch-operator.v5.8.19
        version: 5.8.19
      - name: elasticsearch-operator.v5.8.18
        version: 5.8.18
      - name: elasticsearch-operator.v5.8.17
        version: 5.8.17
      - name: elasticsearch-operator.v5.8.16
        version: 5.8.16
      - name: elasticsearch-operator.v5.8.15
        version: 5.8.15
      - name: elasticsearch-operator.v5.8.14
        version: 5.8.14
      - name: elasticsearch-operator.v5.8.13
        version: 5.8.13
      - name: elasticsearch-operator.v5.8.12
        version: 5.8.12
      - name: elasticsearch-operator.v5.8.11
        version: 5.8.11
      - name: elasticsearch-operator.v5.8.10
        version: 5.8.10
      - name: elasticsearch-operator.v5.8.9
        version: 5.8.9
      - name: elasticsearch-operator.v5.8.8
        version: 5.8.8
      - name: elasticsearch-operator.v5.8.7
        version: 5.8.7
      - name: elasticsearch-operator.v5.8.6
        version: 5.8.6
      - name: elasticsearch-operator.v5.8.5
        version: 5.8.5
      - name: elasticsearch-operator.v5.8.4
        version: 5.8.4
      - name: elasticsearch-operator.v5.8.3
        version: 5.8.3
      - name: elasticsearch-operator.v5.8.2
        version: 5.8.2
      - name: elasticsearch-operator.v5.8.1
        version: 5.8.1
      - name: elasticsearch-operator.v5.8.0
        version: 5.8.0
      name: stable
    - currentCSV: elasticsearch-operator.v5.8.22
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Elasticsearch",
                "metadata": {
                  "name": "elasticsearch"
                },
                "spec": {
                  "indexManagement": {
                    "mappings": [
                      {
                        "aliases": [
                          "infra",
                          "logs.infra"
                        ],
                        "name": "infra",
                        "policyRef": "infra-policy"
                      }
                    ],
                    "policies": [
                      {
                        "name": "infra-policy",
                        "phases": {
                          "delete": {
                            "minAge": "2d",
                            "namespaceSpec": [
                              {
                                "minAge": "5h",
                                "namespace": "openshift-monitoring"
                              }
                            ],
                            "pruneNamespacesInterval": "24h"
                          },
                          "hot": {
                            "actions": {
                              "rollover": {
                                "maxAge": "8h"
                              }
                            }
                          }
                        },
                        "pollInterval": "30m"
                      }
                    ]
                  },
                  "managementState": "Managed",
                  "nodeSpec": {
                    "resources": {
                      "limits": {
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "512Mi"
                      }
                    }
                  },
                  "nodes": [
                    {
                      "nodeCount": 1,
                      "roles": [
                        "client",
                        "data",
                        "master"
                      ],
                      "storage": {
                        "size": "20G"
                      }
                    }
                  ],
                  "redundancyPolicy": "ZeroRedundancy"
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Kibana",
                "metadata": {
                  "name": "kibana"
                },
                "spec": {
                  "managementState": "Managed",
                  "nodeSelector": {},
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "memory": "512Mi"
                    },
                    "requests": {
                      "memory": "512Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.12-v4.17
          containerImage: registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:ce1f01f7a7fd8fa3d52f6bca62f63f65aebcb4c9a8ff8f6afaca7e5d121f0406
          createdAt: "2025-09-23T09:54:59Z"
          description: |
            The Elasticsearch Operator for OCP provides a means for configuring and managing an Elasticsearch cluster for tracing and cluster logging.
            ## Prerequisites and Requirements
            ### Elasticsearch Operator Namespace
            The Elasticsearch Operator must be deployed to the global operator group namespace
            ### Memory Considerations
            Elasticsearch is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Elasticsearch cluster.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory. Each ES node can operate with a
            lower memory setting though this is not recommended for production deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.22'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-5.8
          operators.operatorframework.io.bundle.channels.v1: stable-5.8
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: elasticsearch-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Cluster Logging, Jaeger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Elasticsearch cluster instance
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.openshift.io
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.openshift.io
            version: v1
        description: |
          ### IN OCP 4.16 AND FORWARD

          IN OCP 4.16 AND FORWARD ELASTICSEARCH and KIBANA ARE **NO LONGER SUPPORTED
          FOR LOG STORAGE** AND ELASTICSEARCH IS ONLY SUPPORTED FOR **SERVICEMESH,
          TRACING, AND KIALI**

          THIS OPERATOR **WILL REACH END OF LIFE ON NOVEMBER 2025**

          ### About

          The Elasticsearch Operator for OCP provides a means for configuring and
          managing an Elasticsearch cluster for use in tracing and cluster logging as
          well as a Kibana instance to connect to it. This operator only supports OCP
          Cluster Logging and Jaeger.  It is tightly coupled to each and is not
          currently capable of being used as a general purpose manager of
          Elasticsearch clusters running on OCP.

          Please note: For a general purpose Elasticsearch operator, please use
          Elastic's Elasticsearch (ECK) Operator
          [here](https://catalog.redhat.com/software/containers/elastic/eck-operator/5fabf6d1ecb52450895164be?container-tabs=gti)

          It is recommended that this operator be installed in the
          `openshift-operators-redhat` namespace to properly support the Cluster
          Logging and Jaeger use cases.

          Once installed, the operator provides the following features for
          **Elasticsearch**: * **Create/Destroy**: Deploy an Elasticsearch cluster to
          the same namespace in which the elasticsearch CR is created.
          * **Update**: Changes to the elasticsearch CR will be scheduled and applied
          to the cluster in a controlled manner (most often as a rolling upgrade).
          * **Cluster health**: The operator will periodically poll the cluster to
          evaluate its current health (such as the number of active shards and if any
          cluster nodes have reached their storage watermark usage).
          * **Redeploys**: In the case where the provided secrets are updated, the
          Elasticsearch Operator will schedule and perform a full cluster restart.
          * **Index management**: The Elasticsearch Operator will create cronjobs to
          perform index management such as roll over and deletion.

          Once installed, the operator provides the following features for **Kibana**:

          * **Create/Destroy**: Deploy a Kibana instance to the same namespace in
          which the kibana CR is created (this should be the same namespace as the
          elasticsearch CR).
          * **Update**: Changes to the kibana CR will be scheduled and applied to the
          cluster in a controlled manner.
          * **Redeploys**: In the case where the provided secrets are updated, the
          Elasticsearch Operator will perform a restart.

          ### Additionally provided features

          * Out of the box multitenancy that is integrated with OCP user access
          control.
          * Document Level Security
          * mTLS communication between Elasticsearch, Kibana, Index Management
          cronjobs, and CLO's Fluentd
          * OCP prometheus dashboard for Elasticsearch clusters
          * Prometheus Alerting rules
        displayName: OpenShift Elasticsearch Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - jaeger
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Elasticsearch Operator
          url: https://github.com/openshift/elasticsearch-operator
        maintainers:
        - email: aos-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:6388562dc496e524a297c733720dea23a2889468fa25394812b3247f5f08c907
        - registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:69878c9e40178530baf837a28d3f78e63eb6fba54865dbf05f30dda1a661d8d0
        - registry.redhat.io/openshift-logging/kibana6-rhel8@sha256:7cfb21861cbc3c6d8c6b6f0192ac3c687e4ac9b16fbccf56841e179a08553c84
        - registry.redhat.io/openshift-logging/logging-curator5-rhel9@sha256:dd1cef6a91ca25231af46e0417c7929663098a3b4ad1230dddf4a8171871ce0e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:ce1f01f7a7fd8fa3d52f6bca62f63f65aebcb4c9a8ff8f6afaca7e5d121f0406
        version: 5.8.22
      entries:
      - name: elasticsearch-operator.v5.8.22
        version: 5.8.22
      - name: elasticsearch-operator.v5.8.21
        version: 5.8.21
      - name: elasticsearch-operator.v5.8.20
        version: 5.8.20
      - name: elasticsearch-operator.v5.8.19
        version: 5.8.19
      - name: elasticsearch-operator.v5.8.18
        version: 5.8.18
      - name: elasticsearch-operator.v5.8.17
        version: 5.8.17
      - name: elasticsearch-operator.v5.8.16
        version: 5.8.16
      - name: elasticsearch-operator.v5.8.15
        version: 5.8.15
      - name: elasticsearch-operator.v5.8.14
        version: 5.8.14
      - name: elasticsearch-operator.v5.8.13
        version: 5.8.13
      - name: elasticsearch-operator.v5.8.12
        version: 5.8.12
      - name: elasticsearch-operator.v5.8.11
        version: 5.8.11
      - name: elasticsearch-operator.v5.8.10
        version: 5.8.10
      - name: elasticsearch-operator.v5.8.9
        version: 5.8.9
      - name: elasticsearch-operator.v5.8.8
        version: 5.8.8
      - name: elasticsearch-operator.v5.8.7
        version: 5.8.7
      - name: elasticsearch-operator.v5.8.6
        version: 5.8.6
      - name: elasticsearch-operator.v5.8.5
        version: 5.8.5
      - name: elasticsearch-operator.v5.8.4
        version: 5.8.4
      - name: elasticsearch-operator.v5.8.3
        version: 5.8.3
      - name: elasticsearch-operator.v5.8.2
        version: 5.8.2
      - name: elasticsearch-operator.v5.8.1
        version: 5.8.1
      - name: elasticsearch-operator.v5.8.0
        version: 5.8.0
      name: stable-5.8
    defaultChannel: stable-5.8
    packageName: elasticsearch-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: application-services-metering-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: application-services-metering-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "applicationservices.redhat.com/v1",
                "kind": "Meter",
                "metadata": {
                  "name": "app-services-metrics-example"
                },
                "spec": {
                  "includeInfrastructure": false,
                  "meterCollectionEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Application Runtime
          certified: "false"
          containerImage: quay.io/redhat-developer/application-services-metering-operator:0.6.1
          createdAt: "2021-09-28T13:00:00Z"
          description: Collect the core usage of products from the Application Services
            portfolio into a single metric labeled by Runtimes, Integration, and Automation.
          repository: https://github.com/redhat-developer/application-services-metering-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Meter
            displayName: Meter
            kind: Meter
            name: meters.applicationservices.redhat.com
            version: v1
        description: |-
          ## About the managed application

          There is no managed application created with this operator.

          ## About this Operator

          The metering operator watches pods for predetermined labels, across a cluster or namespace(s).
          The label schema for Red Hat Application Services is in the format:

          * `com.company=Red_Hat`
          * `rht.comp=PAM`
          * `rht.comp_ver=7.10`
          * `rht.prod_name=Red_Hat_Process_Automation`
          * `rht.prod_ver=7.11`
          * `rht.subcomp=rhpam-kogito-runtime`
          * `rht.subcomp_t=application`

          It's important not to remove labels as the operator will cease monitoring those pods.

          The operator exposes a metric named: `appsvcs_cpu_usage_cores`
          The metric has a label for product category, these are Runtimes, Integration and Automation.
          There is a prometheus rule using the metric to group by product name, `appsvcs_cores_by_product:sum`.

          The operator utilises the OpenShift Monitoring Stack to provide usage observability via cloud.redhat.com

          ## Prerequisites for enabling this Operator

          None.
        displayName: Application Services Metering Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - metering
        - operator
        links:
        - name: Red Hat Application Services portfolio
          url: https://www.redhat.com/en/products/middleware
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/redhat-developer/application-services-metering-operator:0.6.1
        version: 0.6.1
      entries:
      - name: application-services-metering-operator.v0.6.1
        version: 0.6.1
      - name: application-services-metering-operator.v0.6.0
        version: 0.6.0
      name: alpha
    defaultChannel: alpha
    packageName: application-services-metering-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-licensing-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-licensing-operator.v8.0.220
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Developer"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:feeabfe4ecd36205306e3f3bf6c11517cfee735d7ae7eb4350abd01e9d76610b
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          olm.skipRange: <8.0.220
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:2f3de2bf46920cfe8c251011fc99841b43999561c1d3fe44850b205e94585a68
        version: 8.0.220
      entries:
      - name: infoscale-licensing-operator.v8.0.220
        version: 8.0.220
      - name: infoscale-licensing-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-licensing-operator.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-licensing-operator.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:b357b2203466516684c8fbc209757a98a45ec2376981f8699faf7dd5e37486a3
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:b357b2203466516684c8fbc209757a98a45ec2376981f8699faf7dd5e37486a3
        version: 8.0.330
      entries:
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-licensing-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Arctera
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Arctera
        - InfoScale
        - Licensing
        - Arctera Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
        version: 8.0.410
      entries:
      - name: infoscale-licensing-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-licensing-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-licensing-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-licensing-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Arctera
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Arctera
        - InfoScale
        - Licensing
        - Arctera Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
        version: 8.0.410
      entries:
      - name: infoscale-licensing-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-licensing-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-licensing-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-licensing-operator.v8.0.110
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Developer"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:e7b8d3c44df6151e61ded91336273e35eeaec04bd19124cf6313dc16d1224660
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:e7b8d3c44df6151e61ded91336273e35eeaec04bd19124cf6313dc16d1224660
        version: 8.0.110
      entries:
      - name: infoscale-licensing-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-licensing-operator.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-licensing-operator
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kedify
      provider-url: kedify.io
    name: kedify-agent-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kedify-agent-helm-operator.v0.2.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kedify.io/v1alpha1",
                "kind": "KedifyAgent",
                "metadata": {
                  "name": "kedifyagent-sample"
                },
                "spec": {
                  "agent": {
                    "agentId": "",
                    "apiKey": "kfy_xxxxxxxxxxxxxxxxxxxxx",
                    "createCrds": true,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v0.2.16@sha256:bfcc0b24ff75de17464ec213406f48bc211a932308e888c450ed23ee1786d37a"
                    },
                    "kedifyProxy": {
                      "globalValues": {
                        "openshift": true
                      },
                      "image": {
                        "tag": "v1.34.1@sha256:683e2f32809657c8d4a1508648545481130ac5409cd1ce63ccc795ad8c71202e"
                      }
                    },
                    "kubectlImage": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3"
                    },
                    "orgId": "00000000-0000-0000-0000-000000000000",
                    "securityContext": {
                      "readOnlyRootFilesystem": true
                    }
                  },
                  "keda": {
                    "enabled": true,
                    "image": {
                      "keda": {
                        "tag": "v2.17.2-0@sha256:802592b691a7b81a249055fa117994927c2fd376b5023a3888bcc7973c65ba52"
                      },
                      "metricsApiServer": {
                        "tag": "v2.17.2-0@sha256:df44bc1de75324158017d623a9d0adf3e3e9f1d64db6afa7133f777adc7b6389"
                      },
                      "pullPolicy": "IfNotPresent",
                      "webhooks": {
                        "tag": "v2.17.2-0@sha256:73e180bd5793dec3785fcb9ecd7e7184e645f7c37d561a5728970e5303c19a41"
                      }
                    }
                  },
                  "keda-add-ons-http": {
                    "enabled": true,
                    "interceptor": {
                      "pullPolicy": "IfNotPresent",
                      "scaledObject": {
                        "waitForCrd": true
                      }
                    },
                    "kubectlImage": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3"
                    },
                    "pinnedImages": {
                      "interceptor": {
                        "tag": "v0.10.0-17@sha256:30ccf20e79c6a62c14781872a73ad0a811063add7136b77df986bf403d6ca6bb"
                      },
                      "scaler": {
                        "tag": "v0.10.0-17@sha256:5863ee0ca09efeca225966c2eb8781f90f2dfb4aed872d84a6075b711ab042b2"
                      }
                    },
                    "scaler": {
                      "pullPolicy": "IfNotPresent"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: '{"kedify-openshift-default":{"description":"Default
            Kedify Agent CR for OpenShift. Register at https://dashboard.kedify.io
            to get started."}}'
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:d61332d50587ce2c75aff29033fdd5cb1bf0ec49d199ab72323a27d9d19ff402
          createdAt: "2025-08-07T14:04:11Z"
          description: Kedify Autoscaler, powered by proven open source technologies
            and KEDA, is an enterprise-grade service that allows teams to autoscale
            any workload using various event sources including HTTP events
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: keda
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: http://github.com/kedify/kedify-agent-helm-operator
          support: support@kedify.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KedifyAgent
            name: kedifyagents.charts.kedify.io
            version: v1alpha1
        description: Kedify Autoscaler, powered by proven open source technologies
          and KEDA, is an enterprise-grade service that allows teams to autoscale
          any workload using various event sources including HTTP events
        displayName: kedify-agent-helm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - autoscaling
        - keda
        - kedify
        links:
        - name: Kedify Homepage
          url: https://kedify.io
        maintainers:
        - email: support@kedify.io
          name: Kedify
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Kedify
          url: kedify.io
        relatedImages:
        - ghcr.io/kedify/agent:v0.2.16@sha256:9a57d5eafe368aab1d94f0eec3101e02f0ff2e356d8499984030ab89161e42ae
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:d61332d50587ce2c75aff29033fdd5cb1bf0ec49d199ab72323a27d9d19ff402
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:8e3a458912ec6540cb414446425ad7b1546528f04a5d84fcd64d8ee5424392e6
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:fc91d1f7f8b632614f31dea14c0c8d4a5f8311a83fdef82303857aa8b43b5782
        - envoyproxy/envoy:v1.34.1@sha256:007da57c2c328a90bd4e6d99b70bd899132f1b4a9426ccafe25437cf84a60c14
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:4b7277ad53dfb01bf5f2201dd31d94f24519da00d008638c3c36026b96ea78b4
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:9300a5382373b179f7f541a663cd30e50284e54f352993e8cca34460acbd890d
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:c1c093a6fda7efc9bc30f110c55f8bb075250db929dbdb2534b9071efbfc3169
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:30ccf20e79c6a62c14781872a73ad0a811063add7136b77df986bf403d6ca6bb
        - envoyproxy/envoy:v1.34.1@sha256:683e2f32809657c8d4a1508648545481130ac5409cd1ce63ccc795ad8c71202e
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:a38fdaf3a49eb4c71c4f3ab1211befa22fa5de56391f88efc11335f0bb3b1d29
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:82b041027dc900fdab717e8db6a522139067801c04cd1ebfe950572e714938ab
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:720b90a4f112a91a94b1bdafe445df1681657ddaa4fb20aa6d2904a0eb664a2a
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:5863ee0ca09efeca225966c2eb8781f90f2dfb4aed872d84a6075b711ab042b2
        - envoyproxy/envoy:v1.34.1@sha256:75545b91d050aa51b52cb0673be172f2f263c971dae3d2a82a70d128d92e87c2
        - ghcr.io/kedify/agent:v0.2.16@sha256:bfcc0b24ff75de17464ec213406f48bc211a932308e888c450ed23ee1786d37a
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:df44bc1de75324158017d623a9d0adf3e3e9f1d64db6afa7133f777adc7b6389
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:73e180bd5793dec3785fcb9ecd7e7184e645f7c37d561a5728970e5303c19a41
        - ghcr.io/kedify/agent:v0.2.16@sha256:c59d765091a31659236b8d80f4033912f5cf47d86c2e2da96bcfa133b9f6fe25
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:33cd5a8c6dd5e90614341dc1863e1f041f2c2475bd61edf5176a2a0d98ea54f9
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:850922994153104a8e382416561c251b46a95ff2b2a216aacbd7fcb4be7f6382
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:802592b691a7b81a249055fa117994927c2fd376b5023a3888bcc7973c65ba52
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:4b63409c022ba137760e7a534b6308fc1e3639f06fdb53ee1f6df32787b68400
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:ea6b22bb12c8ff7cbab4633c2f4a0e49eaf3883b07ab1f47dbebff2a3e07f029
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:9fcc76f983035eb0ceff7beb941bf32fe7fc28b341ec1aac5191ffc1471f401b
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:cb40290842976fd09f2008d8e2e64ebf90a215f0584656d4886122d322006873
        version: 0.2.16
      entries:
      - name: kedify-agent-helm-operator.v0.2.16
        version: 0.2.16
      name: stable
    defaultChannel: stable
    packageName: kedify-agent-helm-operator
    provider:
      name: Kedify
      url: kedify.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-operator.v2024.10.2-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "24.10.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "false",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": null
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": null
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:f4b9eb4f209f5428b0940056c760efcd9939553f3baac29924d1554701bc2dce
          createdAt: "2024-10-25T01:46:32Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2024.10.2
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2024.3.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "99.99.0.nextgen_uma_beta,3"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:f4b9eb4f209f5428b0940056c760efcd9939553f3baac29924d1554701bc2dce
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:7c4868db021beaffa762eb0de43739bbc865a4bf5900f1d17f78caed8612c073
        version: 2024.10.2-12
      entries:
      - name: uma-operator.v2024.10.2-12
        version: 2024.10.2-12
      name: "2024.10"
    - currentCSV: uma-operator.v2025.1.3-33
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.1.1.33,990033",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "false",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": null
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": null
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperDataQueueSize": 300,
                            "scrapperProcessingRoutines": 50,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:4a1bc124f22cf05dc6b42edbcc971d6309862e7be9c39b2ab6b1e3d50de61f3f
          createdAt: "2025-02-18T05:33:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.1.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2024.3.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       scrapperProcessingRoutines: 50
                       scrapperDataQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.1.1.26,990026"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:4a1bc124f22cf05dc6b42edbcc971d6309862e7be9c39b2ab6b1e3d50de61f3f
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:ae91a28d57c8a359f81757e4bf79472f2ac189ac0ae3ee0d87e8ff9c433f1ccc
        version: 2025.1.3-33
      entries:
      - name: uma-operator.v2025.1.3-33
        version: 2025.1.3-33
      name: "2025.1"
    - currentCSV: uma-operator.v2025.2.1-25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.2.1.25,990025",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mixedMode": "false",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperDataQueueSize": 300,
                            "scrapperProcessingRoutines": 50,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:d8f4ca750252f1570a2dd89cabf018cc4f0b2c0cd6cadfedbf48507a34d4aee9
          createdAt: "2025-03-05T05:49:42Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.2.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2024.3.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       scrapperProcessingRoutines: 50
                       scrapperDataQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.1.1.26,990026"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:d8f4ca750252f1570a2dd89cabf018cc4f0b2c0cd6cadfedbf48507a34d4aee9
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:8bcf0dac889f38a0bf77eb9babe3f726e3b2564d5988c05bdc5c2d91ae189c61
        version: 2025.2.1-25
      entries:
      - name: uma-operator.v2025.2.1-25
        version: 2025.2.1-25
      name: "2025.2"
    - currentCSV: uma-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:8ff4aab6d0043ccd2542a86b7e9c24d60670a3c376021ef2d02fc1a289130275
          createdAt: "2025-04-16T06:54:48Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.4.5
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.4.2.5,990005"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:8ff4aab6d0043ccd2542a86b7e9c24d60670a3c376021ef2d02fc1a289130275
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        version: 2025.4.5-5
      entries:
      - name: uma-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:90b4b21d93bce41f2530699cec01f03e5daa73bfbf4f30cbcc8fbd6e75e97da6
          createdAt: "2025-07-24T15:07:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.5.2.12,990012""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:90b4b21d93bce41f2530699cec01f03e5daa73bfbf4f30cbcc8fbd6e75e97da6
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        version: 2025.5.4-12
      entries:
      - name: uma-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-operator.v2025.7.4-23
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:35f74109aa1916524c1fb483972caa31cb84381c3cff7c3b26fd9b676f9d748c
          createdAt: "2025-08-25T09:00:19Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.3
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.7.1.23,990023"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:35f74109aa1916524c1fb483972caa31cb84381c3cff7c3b26fd9b676f9d748c
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.4-23
      entries:
      - name: uma-operator.v2025.7.4-23
        version: 2025.7.4-23
      name: "2025.7"
    - currentCSV: uma-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:adea186632253865406e95a0062abbb312dee23c92ff108f5914e325dc0b8833
          createdAt: "2025-08-15T03:19:46Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.8.1.9,990009"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:adea186632253865406e95a0062abbb312dee23c92ff108f5914e325dc0b8833
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        version: 2025.8.2-9
      entries:
      - name: uma-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
          createdAt: "2025-09-24T12:22:41Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.12,990012"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:c892787785c8c5e0de1692709a9e3f54eef36f1f00c53f1b9feabb613769a48c
          createdAt: "2025-07-30T08:06:11Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "99.99.0.serp_F152813,19"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:c892787785c8c5e0de1692709a9e3f54eef36f1f00c53f1b9feabb613769a48c
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        version: 2025.7.2-19
      entries:
      - name: uma-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:dc15dfdc8b82ee159b52036cc9d057e6525b08c478ed759939cf5c3fc7daeeec
          createdAt: "2025-09-11T11:15:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.8,990008"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:dc15dfdc8b82ee159b52036cc9d057e6525b08c478ed759939cf5c3fc7daeeec
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
          createdAt: "2025-09-24T12:22:41Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.12,990012"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: EDB
      provider-url: ""
    name: cloud-native-postgresql
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloud-native-postgresql.v1.27.0-rc1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:5f5bebbf4d8021dafd0fb7485d629b30aaa362e899e69eb52fa5e74757af6e36
          createdAt: "2025-07-30T09:28:05Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:5f5bebbf4d8021dafd0fb7485d629b30aaa362e899e69eb52fa5e74757af6e36
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        version: 1.27.0-rc1
      entries:
      - name: cloud-native-postgresql.v1.27.0-rc1
        version: 1.27.0-rc1
      - name: cloud-native-postgresql.v1.26.0-rc3
        version: 1.26.0-rc3
      - name: cloud-native-postgresql.v1.26.0-rc2
        version: 1.26.0-rc2
      - name: cloud-native-postgresql.v1.26.0-rc1
        version: 1.26.0-rc1
      name: candidate
    - currentCSV: cloud-native-postgresql.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
          createdAt: "2025-08-19T12:00:36Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
        version: 1.27.0
      entries:
      - name: cloud-native-postgresql.v1.27.0
        version: 1.27.0
      - name: cloud-native-postgresql.v1.26.1
        version: 1.26.1
      - name: cloud-native-postgresql.v1.26.0
        version: 1.26.0
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      - name: cloud-native-postgresql.v1.24.1
        version: 1.24.1
      - name: cloud-native-postgresql.v1.24.0
        version: 1.24.0
      - name: cloud-native-postgresql.v1.23.3
        version: 1.23.3
      - name: cloud-native-postgresql.v1.23.2
        version: 1.23.2
      - name: cloud-native-postgresql.v1.23.1
        version: 1.23.1
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      - name: cloud-native-postgresql.v1.21.1
        version: 1.21.1
      - name: cloud-native-postgresql.v1.21.0
        version: 1.21.0
      - name: cloud-native-postgresql.v1.20.2
        version: 1.20.2
      - name: cloud-native-postgresql.v1.20.1
        version: 1.20.1
      - name: cloud-native-postgresql.v1.20.0
        version: 1.20.0
      - name: cloud-native-postgresql.v1.19.1
        version: 1.19.1
      - name: cloud-native-postgresql.v1.19.0
        version: 1.19.0
      name: fast
    - currentCSV: cloud-native-postgresql.v1.25.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
          createdAt: "2025-07-26T06:26:53Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.25.3
      entries:
      - name: cloud-native-postgresql.v1.25.3
        version: 1.25.3
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      - name: cloud-native-postgresql.v1.22.9
        version: 1.22.9
      - name: cloud-native-postgresql.v1.22.8
        version: 1.22.8
      - name: cloud-native-postgresql.v1.22.7
        version: 1.22.7
      - name: cloud-native-postgresql.v1.22.6
        version: 1.22.6
      - name: cloud-native-postgresql.v1.22.5
        version: 1.22.5
      - name: cloud-native-postgresql.v1.22.4
        version: 1.22.4
      - name: cloud-native-postgresql.v1.22.3
        version: 1.22.3
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      - name: cloud-native-postgresql.v1.18.13
        version: 1.18.13
      - name: cloud-native-postgresql.v1.18.12
        version: 1.18.12
      - name: cloud-native-postgresql.v1.18.11
        version: 1.18.11
      - name: cloud-native-postgresql.v1.18.10
        version: 1.18.10
      - name: cloud-native-postgresql.v1.18.9
        version: 1.18.9
      - name: cloud-native-postgresql.v1.18.8
        version: 1.18.8
      - name: cloud-native-postgresql.v1.18.7
        version: 1.18.7
      name: stable
    - currentCSV: cloud-native-postgresql.v1.18.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:69e439a2a961881309cc122174eaca273cf6d7873e9e871efbdb2d60804f40e4
          createdAt: "2024-06-13T16:08:55Z"
          description: EDB Postgres for Kubernetes is an open source operator designed
            to manage highly available PostgreSQL databases with a primary/standby
            architecture on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 1.18.0 < 1.18.13'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloud
        - native
        - edb
        - enterprisedb
        - cnp
        - pg4k
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: john.long@enterprisedb.com
          name: John Long
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: The EDB Postgres for Kubernetes Contributors
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:69e439a2a961881309cc122174eaca273cf6d7873e9e871efbdb2d60804f40e4
        version: 1.18.13
      entries:
      - name: cloud-native-postgresql.v1.18.13
        version: 1.18.13
      - name: cloud-native-postgresql.v1.18.12
        version: 1.18.12
      - name: cloud-native-postgresql.v1.18.11
        version: 1.18.11
      - name: cloud-native-postgresql.v1.18.10
        version: 1.18.10
      - name: cloud-native-postgresql.v1.18.9
        version: 1.18.9
      - name: cloud-native-postgresql.v1.18.8
        version: 1.18.8
      - name: cloud-native-postgresql.v1.18.7
        version: 1.18.7
      name: stable-v1.18
    - currentCSV: cloud-native-postgresql.v1.19.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:2a717edd0523798bd4d0b14c557b8d6b19ae3ad1f019c0cc4f61db49f10c3dad
          createdAt: "2023-11-08T19:02:47Z"
          description: Operator to manage Postgres high availability clusters with
            a primary/standby architecture.
          olm.skipRange: '>=1.15.3 < 1.19.6'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup manager
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster manager
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PgBouncer Pooler Manager
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL backups schedule
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: "EDB Postgres for Kubernetes is an operator designed, developed,
          and supported by EDB that covers the full \nlifecycle of a highly available
          Postgres database clusters with a primary/standby architecture, using native\nstreaming
          replication. It is based on the open source CloudNativePG operator, and
          provides additional value\nsuch as compatibility with Oracle using EDB Postgres
          Advanced Server and additional supported platforms such\nas IBM Power and
          OpenShift.\n\nKey features available include:\n\n* Kubernetes API integration
          for high availability\n* Self-healing through failover and automated recreation
          of replicas\n* Capacity management with scale up/down capabilities\n* Planned
          switchovers for scheduled maintenance\n* Read-only and read-write Kubernetes
          services definitions\n* Rolling updates for Postgres minor versions and
          operator upgrades\n* Continuous backup and point-in-time recovery\n* Connection
          Pooling with PgBouncer\n* Integrated metrics exporter out of the box\n*
          PostgreSQL replication across multiple Kubernetes clusters\n* Red Hat certified
          operator for OpenShift\n\nThe operator has been renamed from Cloud Native
          PostgreSQL. Existing users of Cloud Native PostgreSQL will not\nexperience
          any change, as the underlying components and resources have not changed.\n"
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloud
        - native
        - edb
        - enterprisedb
        - cnp
        - pg4k
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: sales@enterprisedb.com
          name: Sales
        - email: support@enterprisedb.com
          name: Support
        maturity: stable
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:2a717edd0523798bd4d0b14c557b8d6b19ae3ad1f019c0cc4f61db49f10c3dad
        version: 1.19.6
      entries:
      - name: cloud-native-postgresql.v1.19.6
        version: 1.19.6
      - name: cloud-native-postgresql.v1.19.5
        version: 1.19.5
      - name: cloud-native-postgresql.v1.19.4
        version: 1.19.4
      - name: cloud-native-postgresql.v1.19.3
        version: 1.19.3
      - name: cloud-native-postgresql.v1.19.2
        version: 1.19.2
      - name: cloud-native-postgresql.v1.19.1
        version: 1.19.1
      - name: cloud-native-postgresql.v1.19.0
        version: 1.19.0
      name: stable-v1.19
    - currentCSV: cloud-native-postgresql.v1.20.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:919c6669f8bfa333b19fdd1ca03aaeceebe0c5ca74d91cb66b2369680422e916
          createdAt: "2024-02-03T17:26:23Z"
          description: Operator to manage Postgres high availability clusters with
            a primary/standby architecture.
          olm.skipRange: '>=1.15.3 < 1.20.6'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup manager
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster manager
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PgBouncer Pooler Manager
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL backups schedule
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: "EDB Postgres for Kubernetes is an operator designed, developed,
          and supported by EDB that covers the full \nlifecycle of a highly available
          Postgres database clusters with a primary/standby architecture, using native\nstreaming
          replication. It is based on the open source CloudNativePG operator, and
          provides additional value\nsuch as compatibility with Oracle using EDB Postgres
          Advanced Server and additional supported platforms such\nas IBM Power and
          OpenShift.\n\nKey features available include:\n\n* Kubernetes API integration
          for high availability\n* Self-healing through failover and automated recreation
          of replicas\n* Capacity management with scale up/down capabilities\n* Planned
          switchovers for scheduled maintenance\n* Read-only and read-write Kubernetes
          services definitions\n* Rolling updates for Postgres minor versions and
          operator upgrades\n* Continuous backup and point-in-time recovery\n* Connection
          Pooling with PgBouncer\n* Integrated metrics exporter out of the box\n*
          PostgreSQL replication across multiple Kubernetes clusters\n* Red Hat certified
          operator for OpenShift\n\nThe operator has been renamed from Cloud Native
          PostgreSQL. Existing users of Cloud Native PostgreSQL will not\nexperience
          any change, as the underlying components and resources have not changed.\n"
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloud
        - native
        - edb
        - enterprisedb
        - cnp
        - pg4k
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: sales@enterprisedb.com
          name: Sales
        - email: support@enterprisedb.com
          name: Support
        maturity: stable
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:919c6669f8bfa333b19fdd1ca03aaeceebe0c5ca74d91cb66b2369680422e916
        version: 1.20.6
      entries:
      - name: cloud-native-postgresql.v1.20.6
        version: 1.20.6
      - name: cloud-native-postgresql.v1.20.5
        version: 1.20.5
      - name: cloud-native-postgresql.v1.20.4
        version: 1.20.4
      - name: cloud-native-postgresql.v1.20.3
        version: 1.20.3
      - name: cloud-native-postgresql.v1.20.2
        version: 1.20.2
      - name: cloud-native-postgresql.v1.20.1
        version: 1.20.1
      - name: cloud-native-postgresql.v1.20.0
        version: 1.20.0
      name: stable-v1.20
    - currentCSV: cloud-native-postgresql.v1.21.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:8cc328b1fc8efa3911d3dba689bc824ff789b305f7eb3284bf21ffcdc52a9683
          createdAt: "2024-06-13T16:41:26Z"
          description: EDB Postgres for Kubernetes is an open source operator designed
            to manage highly available PostgreSQL databases with a primary/standby
            architecture on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 1.18.0 < 1.21.6'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: john.long@enterprisedb.com
          name: John Long
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: The EDB Postgres for Kubernetes Contributors
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:8cc328b1fc8efa3911d3dba689bc824ff789b305f7eb3284bf21ffcdc52a9683
        version: 1.21.6
      entries:
      - name: cloud-native-postgresql.v1.21.6
        version: 1.21.6
      - name: cloud-native-postgresql.v1.21.5
        version: 1.21.5
      - name: cloud-native-postgresql.v1.21.4
        version: 1.21.4
      - name: cloud-native-postgresql.v1.21.3
        version: 1.21.3
      - name: cloud-native-postgresql.v1.21.2
        version: 1.21.2
      - name: cloud-native-postgresql.v1.21.1
        version: 1.21.1
      - name: cloud-native-postgresql.v1.21.0
        version: 1.21.0
      name: stable-v1.21
    - currentCSV: cloud-native-postgresql.v1.22.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:4d55ad54cc7e7807bc99f5a6f96bc92e88b4060f424ffd2cb6552be9505f38f4
          createdAt: "2025-07-25T19:02:56Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images. Only
              available in 1.25 or later
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster.
              Only available in 1.25 or later
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster. Only available in 1.25 or later
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication.
              Only available in 1.25 or later
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:4d55ad54cc7e7807bc99f5a6f96bc92e88b4060f424ffd2cb6552be9505f38f4
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.22.11
      entries:
      - name: cloud-native-postgresql.v1.22.11
        version: 1.22.11
      - name: cloud-native-postgresql.v1.22.10
        version: 1.22.10
      - name: cloud-native-postgresql.v1.22.9
        version: 1.22.9
      - name: cloud-native-postgresql.v1.22.8
        version: 1.22.8
      - name: cloud-native-postgresql.v1.22.7
        version: 1.22.7
      - name: cloud-native-postgresql.v1.22.6
        version: 1.22.6
      - name: cloud-native-postgresql.v1.22.5
        version: 1.22.5
      - name: cloud-native-postgresql.v1.22.4
        version: 1.22.4
      - name: cloud-native-postgresql.v1.22.3
        version: 1.22.3
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      name: stable-v1.22
    - currentCSV: cloud-native-postgresql.v1.23.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:934e01ed5cd35cf681414b9136a337514a1b73c0110871f1c78ba98447badb0f
          createdAt: "2024-12-23T16:37:54Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 1.18.0 < 1.23.6'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:934e01ed5cd35cf681414b9136a337514a1b73c0110871f1c78ba98447badb0f
        version: 1.23.6
      entries:
      - name: cloud-native-postgresql.v1.23.6
        version: 1.23.6
      - name: cloud-native-postgresql.v1.23.5
        version: 1.23.5
      - name: cloud-native-postgresql.v1.23.4
        version: 1.23.4
      - name: cloud-native-postgresql.v1.23.3
        version: 1.23.3
      - name: cloud-native-postgresql.v1.23.2
        version: 1.23.2
      - name: cloud-native-postgresql.v1.23.1
        version: 1.23.1
      name: stable-v1.23
    - currentCSV: cloud-native-postgresql.v1.24.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:a1acaaf5a413db0f33f3824a15aefe598974d1593172bcbccd0fc3d39d7b46d0
          createdAt: "2025-05-23T15:20:37Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:a1acaaf5a413db0f33f3824a15aefe598974d1593172bcbccd0fc3d39d7b46d0
        version: 1.24.4
      entries:
      - name: cloud-native-postgresql.v1.24.4
        version: 1.24.4
      - name: cloud-native-postgresql.v1.24.3
        version: 1.24.3
      - name: cloud-native-postgresql.v1.24.2
        version: 1.24.2
      - name: cloud-native-postgresql.v1.24.1
        version: 1.24.1
      - name: cloud-native-postgresql.v1.24.0
        version: 1.24.0
      name: stable-v1.24
    - currentCSV: cloud-native-postgresql.v1.25.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
          createdAt: "2025-07-26T06:26:53Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
        version: 1.25.3
      entries:
      - name: cloud-native-postgresql.v1.25.3
        version: 1.25.3
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      name: stable-v1.25
    - currentCSV: cloud-native-postgresql.v1.26.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:47d70233908ea855ddab3ff79c7a226719c3f2be5b8a5018c7934cd44bf0f546
          createdAt: "2025-07-26T07:59:14Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:47d70233908ea855ddab3ff79c7a226719c3f2be5b8a5018c7934cd44bf0f546
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.26.1
      entries:
      - name: cloud-native-postgresql.v1.26.1
        version: 1.26.1
      - name: cloud-native-postgresql.v1.26.0
        version: 1.26.0
      name: stable-v1.26
    - currentCSV: cloud-native-postgresql.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
          createdAt: "2025-08-19T12:00:36Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.27.0
      entries:
      - name: cloud-native-postgresql.v1.27.0
        version: 1.27.0
      name: stable-v1.27
    defaultChannel: fast
    packageName: cloud-native-postgresql
    provider:
      name: EDB
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: http://marketplace.redhat.com
    name: marketplace-games-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marketplace-games-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "game.marketplace.redhat.com/v1",
                "kind": "Arcade",
                "metadata": {
                  "name": "arcade-sample"
                },
                "spec": {
                  "size": 1
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/marketplace-games-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/marketplace-games-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Arcade is the Schema for the arcades API
            displayName: Arcade Instance
            kind: Arcade
            name: arcades.game.marketplace.redhat.com
            version: v1
        description: RHM Arcade Operator - a simple operator used to showcase how
          to build an operator around an existing web application that can can run
          on Openshift, Vanilla k8s, or on-boarded to Red Hat Marketplace.
        displayName: Arcade Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - rhm
        - red hat
        - marketplace
        - games
        - arcade
        links:
        - name: Marketplace Games Operator
          url: https://github.com/redhat-marketplace/marketplace-games-operator
        maturity: alpha
        provider:
          name: Red Hat Marketplace
          url: http://marketplace.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/rhm-labs/arcade-operator@sha256:d6959be6693728d2dc4cd38b69251551d23c5002cb52db2684df31a0a8f7b0ab
        version: 1.0.0
      entries:
      - name: marketplace-games-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: marketplace-games-operator-rhmp
    provider:
      name: Red Hat Marketplace
      url: http://marketplace.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/
    name: varnish-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: varnish-operator.v0.27.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "caching.ibm.com/v1alpha1",
                "kind": "VarnishCluster",
                "metadata": {
                  "name": "varnishcluster-sample"
                },
                "spec": {
                  "backend": {
                    "port": "web",
                    "selector": {
                      "app": "nginx"
                    }
                  },
                  "replicas": 1,
                  "service": {
                    "port": 80
                  },
                  "varnish": {
                    "args": [
                      "-p",
                      "default_ttl=3600",
                      "-p",
                      "default_grace=3600"
                    ]
                  },
                  "vcl": {
                    "configMapName": "vcl-files",
                    "entrypointFileName": "entrypoint.vcl"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: ibmcom/varnish-operator:0.27.2
          createdAt: 2021-02-15T11:15:31+0000
          description: Run and manage Varnish clusters on Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/varnish-operator
          support: https://ibm.github.io/varnish-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VarnishCluster is the Schema for the varnishclusters API
            displayName: Varnish Cluster
            kind: VarnishCluster
            name: varnishclusters.caching.ibm.com
            version: v1alpha1
        description: |
          Run and manage Varnish clusters on Kubernetes

          The operator manages the whole lifecycle of the cluster: creating, deleting and keeping the cluster configuration up to date.

          The operator is responsible for building the VCL configuration using templates defined by the users and keeping the configuration up to date when relevant events occur (backend pod failure, scaling of the deployment, VCL configuration change).

          Example of a simple `VarnishCluster`:

          ```yaml
          apiVersion: caching.ibm.com/v1alpha1
          kind: VarnishCluster
          metadata:
            name: varnishcluster-sample
            namespace: varnish-ns
          spec:
            vcl:
              configMapName: vcl-files
              entrypointFileName: entrypoint.vcl
            replicas: 3
            backend:
              selector:
                app: nginx
            service:
              port: 80
          ```

          Full documentation can be found [here](https://ibm.github.io/varnish-operator/)
        displayName: Varnish Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - varnish
        - cache
        links:
        - name: Varnish Operator Docs
          url: https://ibm.github.io/varnish-operator
        - name: Varnish Operator Source Code
          url: https://github.com/IBM/varnish-operator
        maintainers:
        - email: tomash.sidei@ibm.com
          name: Tomash Sidei
        - email: cingram@us.ibm.com
          name: Craig Ingram
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/
        relatedImages:
        - ibmcom/varnish-operator:0.27.2
        version: 0.27.2
      entries:
      - name: varnish-operator.v0.27.2
        version: 0.27.2
      name: alpha
    defaultChannel: alpha
    packageName: varnish-operator
    provider:
      name: IBM
      url: https://www.ibm.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CloudBees
      provider-url: https://www.cloudbees.com/products/continuous-integration
    name: cloudbees-ci
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudbees-ci.v3.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.cloudbees.com/v1alpha1",
                "kind": "CloudBeesCI",
                "metadata": {
                  "name": "cloudbeesci-sample"
                },
                "spec": {
                  "Agents": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-agent:2.289.3.2"
                    },
                    "ImagePullSecrets": null,
                    "SeparateNamespace": {
                      "Create": false,
                      "Enabled": false,
                      "Name": null
                    }
                  },
                  "Hibernation": {
                    "Enabled": false,
                    "Image": {
                      "dockerImage": "cloudbees/managed-master-hibernation-monitor:247.c5dfce00a179",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "NodeSelector": {},
                    "Tolerations": []
                  },
                  "Master": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-mm:2.289.3.2"
                    },
                    "JavaOpts": null,
                    "OperationsCenterNamespace": null
                  },
                  "NetworkPolicy": {
                    "Enabled": false,
                    "JMXSelectors": [],
                    "ingressControllerSelector": []
                  },
                  "OperationsCenter": {
                    "AgentListenerPort": 50000,
                    "Annotations": {},
                    "CSRF": {
                      "ProxyCompatibility": false
                    },
                    "CasC": {
                      "ConfigMapName": "oc-casc-bundle",
                      "Enabled": false
                    },
                    "ContainerPort": 8080,
                    "ContextPath": null,
                    "Enabled": true,
                    "ExtraConfigMaps": [],
                    "ExtraContainers": [],
                    "ExtraGroovyConfiguration": {},
                    "ExtraVolumeMounts": [],
                    "ExtraVolumes": [],
                    "HealthProbeLivenessFailureThreshold": 12,
                    "HealthProbes": true,
                    "HostName": null,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-cloud-core-oc:2.289.3.2",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "Ingress": {
                      "Annotations": {
                        "kubernetes.io/tls-acme": "false"
                      },
                      "Class": "nginx",
                      "tls": {
                        "Enable": false,
                        "SecretName": null
                      }
                    },
                    "JavaOpts": null,
                    "JenkinsOpts": null,
                    "LoadBalancerIP": null,
                    "LoadBalancerSourceRanges": [
                      "0.0.0.0/0"
                    ],
                    "Name": "cjoc",
                    "NodeSelector": {},
                    "Platform": "standard",
                    "Protocol": "http",
                    "Resources": {
                      "Limits": {
                        "Cpu": 1,
                        "Memory": "2G"
                      },
                      "Requests": {
                        "Cpu": 1,
                        "Memory": "2G"
                      }
                    },
                    "Route": {
                      "tls": {
                        "CACertificate": null,
                        "Certificate": null,
                        "DestinationCACertificate": null,
                        "Enable": false,
                        "InsecureEdgeTerminationPolicy": "Redirect",
                        "Key": null,
                        "Termination": "edge"
                      }
                    },
                    "ServiceAgentListenerPort": 50000,
                    "ServiceAnnotations": {},
                    "ServicePort": 80,
                    "ServiceType": "ClusterIP",
                    "Tolerations": []
                  },
                  "Persistence": {
                    "AccessMode": "ReadWriteOnce",
                    "Annotations": {},
                    "Size": "20Gi",
                    "StorageClass": null
                  },
                  "PodSecurityPolicy": {
                    "Annotations": {},
                    "Enabled": false
                  },
                  "Subdomain": false,
                  "ingress-nginx": {
                    "Enabled": false,
                    "controller": {
                      "admissionWebhooks": {
                        "patch": {
                          "nodeSelector": {
                            "kubernetes.io/os": "linux"
                          }
                        }
                      },
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "nginx-ingress": {
                    "Enabled": false,
                    "controller": {
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "rbac": {
                    "agentsServiceAccountName": "jenkins-agents",
                    "hibernationMonitorServiceAccountName": "managed-master-hibernation-monitor",
                    "install": true,
                    "masterServiceAccountName": "jenkins",
                    "serviceAccountName": "cjoc"
                  },
                  "sidecarinjector": {
                    "Enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudBeesCI
            name: cloudbeescis.charts.cloudbees.com
            version: v1alpha1
        description: CloudBees CI
        displayName: CloudBees CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudbees
        - ci
        - enterprise
        - continuous
        - integration
        - jenkins
        - automation
        - pipeline
        - devops
        links:
        - name: CloudBees CI
          url: https://www.cloudbees.com/products/continuous-integration
        maintainers:
        - email: support@cloudbees.com
          name: Support
        - email: sales@cloudbees.com
          name: Sales
        maturity: alpha
        provider:
          name: CloudBees
          url: https://www.cloudbees.com/products/continuous-integration
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.connect.redhat.com/cloudbees/cloudbees-ci-operator@sha256:bd0bd72e9e7807b0fa307de8a93a56967221fe2dd181bac02383f0f9be9120e4
        version: 3.34.1
      entries:
      - name: cloudbees-ci.v3.34.1
        version: 3.34.1
      name: alpha
    defaultChannel: alpha
    packageName: cloudbees-ci
    provider:
      name: CloudBees
      url: https://www.cloudbees.com/products/continuous-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: https://github.com/Mellanox/network-operator/
    name: nvidia-network-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nvidia-network-operator.v25.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "doca3.1.0-25.07-0.9.7.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "nvcr.io/nvidia/mellanox",
                    "version": "sha256:a87096761d155eeb6f470e042d2d167bb466d57e63b4aba957f57d745e15a9b2"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:51270965284cceeda659fb1098e70a888b39664ceade7e9d455fb95ac85914b7
          createdAt: "2025-08-21T19:12:20Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/mellanox/nvidia-k8s-ipam@sha256:3f692273c1598525b9c698c9d9cd463e5ddc55b638c6c031639f38786fc67475
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:9a831bfdf85f313b1f5749b7c9b2673bb8fff18b4ff768c9242dabaa4468e449
        - nvcr.io/nvidia/mellanox/k8s-rdma-shared-dev-plugin@sha256:a87096761d155eeb6f470e042d2d167bb466d57e63b4aba957f57d745e15a9b2
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:5ad37cd5ae96153e6f87670f437ff5f6e44febe7bb6eaeb1a0f219e28284e53d
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:d41eea65b102dc5bb5792871ae212b59a65dae492ca109c52138d5626eba1ffd
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:13c2a2d055e501c03c6451413b85476ceafbc2fbffc7443952294c196f3f89f3
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:30354799264480e1aee057397eef2c52b277bbc6872184e2b7b811f4152b3a67
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:423aea08b4a0f01e2e19b6a709e14347bd08f8274f40002e0224090dfa4904fc
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:f9d4b8358c68ba28f849ca710bf769c4327a639733c0ff30282ca33a744078f6
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:51270965284cceeda659fb1098e70a888b39664ceade7e9d455fb95ac85914b7
        - nvcr.io/nvidia/mellanox/nic-feature-discovery@sha256:1330e3a7ea2491bc310bfc4766769389ceb56d050831bf9c145b7bfb6aba36c4
        - nvcr.io/nvidia/mellanox/network-operator-init-container@sha256:02ee60ac6ee163acc97706bd4229d268f608a95bcbcb6fccee90ceb1d8a6c62b
        - nvcr.io/nvidia/mellanox/ipoib-cni@sha256:7564ad1142c863e61f1dd84f5714deba9a7c6f6d85efb383ac3a2566eb4f4b71
        - nvcr.io/nvidia/mellanox/sriov-network-device-plugin@sha256:7508a160f3f124e0b78d2b4d5a52deca6b9f56e226fee957c4d156e425eb101a
        - nvcr.io/nvidia/mellanox/nic-configuration-operator-daemon@sha256:b78c15904cecd343fa48374757b3619b5cbdc22059d5ee9731da87e35b94a2e2
        - nvcr.io/nvidia/mellanox/nic-configuration-operator@sha256:03f34f7fde3f65ca7b214b66c6fdbc1250d4f3863f9c02b0fba78e8aa84efca3
        - nvcr.io/nvidia/mellanox/ib-kubernetes@sha256:1fa251979b5f5201af50bc7aa6936829103d742f42bfbf752f20852f5fe44dfb
        version: 25.7.0
      entries:
      - name: nvidia-network-operator.v25.7.0
        version: 25.7.0
      - name: nvidia-network-operator.v25.4.0
        version: 25.4.0
      name: stable
    - currentCSV: nvidia-network-operator.v24.10.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "24.10-0.7.0.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "v1.5.2"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:23aeb481d662496210d1c13083c65cfa4ffdcf67f973e5f884b1b936f693c0f5
          createdAt: "2025-01-27T08:16:45Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - ghcr.io/mellanox/ib-kubernetes@sha256:100e6addf99d5b44e81cb3aa8d2fc892fced1de3183190931961b92ee924207b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:e36c46cf2c813c081bf7cc22b47fb4cf391a3d91804e47a6ce87f31d62f42535
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:4bb13a8b850c91011013509334d5ae4f4b7aa1d1b0c12958625e161cd50309d6
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:9f468fdc4449e65e4772575f83aa85840a00f97165f9a00ba34695c91d610fbd
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:5a87d04872d5d4530cab3cbb3db528722fa24435f3375debd8ed102b04e631c2
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:23aeb481d662496210d1c13083c65cfa4ffdcf67f973e5f884b1b936f693c0f5
        - ghcr.io/mellanox/ipoib-cni@sha256:81646c07e27ce1df2603050a75dd267f264685fdc64c285eca3925ae98ed5392
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:77e3b946c20f45728dfa6dd1d115356de5ea317f91cddfd1540b7121550f66dc
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:47e1bb84ac97f9af49f97b49c2de44b4ba82b890b487880edc5c9adf86a176f7
        - ghcr.io/mellanox/network-operator-init-container@sha256:1699d23027ea30c9fa59575a914114bdfd5a87a359caf8c0a9b16d409ec0d068
        - ghcr.io/mellanox/nic-feature-discovery@sha256:92dda9434519de39be0f94c80787766e3399e8b238d14839360aff4d6232a09b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:db5797121735eeb935d8f11e305b50c80ba07aa83c291d76d05dea7e303bc740
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:ecb2cfdc510b7ed09a59daed4817ee353e83a94d2d8581c3c052f4fa4b990574
        version: 24.10.1
      entries:
      - name: nvidia-network-operator.v24.10.1
        version: 24.10.1
      name: v24.10
    - currentCSV: nvidia-network-operator.v24.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "24.10-0.7.0.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "v1.5.2"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:f258e8665d094c09159654759bc9ae41fbe0f7fffe95139ec02964759c3d8a87
          createdAt: "2024-11-26T09:58:46Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:5a87d04872d5d4530cab3cbb3db528722fa24435f3375debd8ed102b04e631c2
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:ecb2cfdc510b7ed09a59daed4817ee353e83a94d2d8581c3c052f4fa4b990574
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:9f468fdc4449e65e4772575f83aa85840a00f97165f9a00ba34695c91d610fbd
        - ghcr.io/mellanox/nic-feature-discovery@sha256:92dda9434519de39be0f94c80787766e3399e8b238d14839360aff4d6232a09b
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:f258e8665d094c09159654759bc9ae41fbe0f7fffe95139ec02964759c3d8a87
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:db5797121735eeb935d8f11e305b50c80ba07aa83c291d76d05dea7e303bc740
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:e36c46cf2c813c081bf7cc22b47fb4cf391a3d91804e47a6ce87f31d62f42535
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:4bb13a8b850c91011013509334d5ae4f4b7aa1d1b0c12958625e161cd50309d6
        - ghcr.io/mellanox/ib-kubernetes@sha256:100e6addf99d5b44e81cb3aa8d2fc892fced1de3183190931961b92ee924207b
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:77e3b946c20f45728dfa6dd1d115356de5ea317f91cddfd1540b7121550f66dc
        - ghcr.io/mellanox/ipoib-cni@sha256:81646c07e27ce1df2603050a75dd267f264685fdc64c285eca3925ae98ed5392
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:47e1bb84ac97f9af49f97b49c2de44b4ba82b890b487880edc5c9adf86a176f7
        - ghcr.io/mellanox/network-operator-init-container@sha256:1699d23027ea30c9fa59575a914114bdfd5a87a359caf8c0a9b16d409ec0d068
        version: 24.10.0
      entries:
      - name: nvidia-network-operator.v24.10.0
        version: 24.10.0
      name: v24.10.0
    - currentCSV: nvidia-network-operator.v24.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "docaTelemetryService": {
                    "image": "doca_telemetry",
                    "repository": "nvcr.io/nvidia/doca",
                    "version": "1.16.5-doca2.6.0-host"
                  },
                  "nicFeatureDiscovery": {
                    "image": "nic-feature-discovery",
                    "repository": "ghcr.io/mellanox",
                    "version": "v0.0.1"
                  },
                  "nvIpam": {
                    "enableWebhook": false,
                    "image": "nvidia-k8s-ipam",
                    "repository": "ghcr.io/mellanox",
                    "version": "v0.2.0"
                  },
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "24.07-0.6.1.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "v1.5.1"
                  },
                  "secondaryNetwork": {
                    "ipoib": {
                      "image": "ipoib-cni",
                      "repository": "ghcr.io/mellanox",
                      "version": "v1.2.0"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:fcee44995d64058cb7bf14c02da604879443e191a382402b8a6ebfe10a3a2601
          createdAt: "2024-08-29T13:23:34Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - ghcr.io/mellanox/network-operator-init-container@sha256:1699d23027ea30c9fa59575a914114bdfd5a87a359caf8c0a9b16d409ec0d068
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:93e856dde88abb0aeb8e0b991e5f6d9f41884d8beea792f4d70cda06f18e783f
        - ghcr.io/mellanox/ib-kubernetes@sha256:b56bb88963dadaa2290a17cf0b57b86c55c09e7c075b7271c184959532555c91
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:c41eaf76c06186e1dabb0d870225d946ad3ecda2a072fa4be197dd131b7695f2
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:8324c009c3d694fe7702af36305ede78b903d839190b1e4506e82362f9aa5c02
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:fcee44995d64058cb7bf14c02da604879443e191a382402b8a6ebfe10a3a2601
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:47e1bb84ac97f9af49f97b49c2de44b4ba82b890b487880edc5c9adf86a176f7
        - ghcr.io/mellanox/nic-feature-discovery@sha256:92dda9434519de39be0f94c80787766e3399e8b238d14839360aff4d6232a09b
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - ghcr.io/mellanox/ipoib-cni@sha256:81646c07e27ce1df2603050a75dd267f264685fdc64c285eca3925ae98ed5392
        version: 24.7.0
      entries:
      - name: nvidia-network-operator.v24.7.0
        version: 24.7.0
      name: v24.7.0
    - currentCSV: nvidia-network-operator.v24.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "docaTelemetryService": {
                    "image": "doca_telemetry",
                    "repository": "nvcr.io/nvidia/doca",
                    "version": "1.16.5-doca2.6.0-host"
                  },
                  "nicFeatureDiscovery": {
                    "image": "nic-feature-discovery",
                    "repository": "ghcr.io/mellanox",
                    "version": "v0.0.1"
                  },
                  "nvIpam": {
                    "enableWebhook": false,
                    "image": "nvidia-k8s-ipam",
                    "repository": "ghcr.io/mellanox",
                    "version": "v0.2.0"
                  },
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "24.07-0.6.1.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "v1.5.1"
                  },
                  "secondaryNetwork": {
                    "ipoib": {
                      "image": "ipoib-cni",
                      "repository": "ghcr.io/mellanox",
                      "version": "v1.2.0"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:b8a07696d05fb8e5292991ba731758579313d6983ed3d75dd95e7356024d987c
          createdAt: "2024-09-04T06:38:33Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:d5a163d36ce60630e14ae8620ab437510ff0821924e043b88f0d7b67cee3bd77
        - ghcr.io/mellanox/network-operator-init-container@sha256:1699d23027ea30c9fa59575a914114bdfd5a87a359caf8c0a9b16d409ec0d068
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:403be1bcf708984f90b3ccfdf24f58594554336d69f878d3843535f4fa61ceca
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:7facd05f60e2be29925ffce9d0a6b2c712c0670c8889b815425582c46c82efdc
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:8324c009c3d694fe7702af36305ede78b903d839190b1e4506e82362f9aa5c02
        - ghcr.io/mellanox/ib-kubernetes@sha256:b56bb88963dadaa2290a17cf0b57b86c55c09e7c075b7271c184959532555c91
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:10684c0c36364b1d897a03a346abb2f691a18541ae1242b596d002f34f0d7390
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:7f87324adb351513d87fa27dfd0c9d9371ae42529d76cce8ecd68ce41c4d263f
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:676c74fcbe6d2c2341a092913b1afca65ba1a2f568ee6796b371542657e7a174
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:22e6ff1e9245f0cdf67e7904235d1a6181ee2bd3f39cedf76b39d4aeaa4a6c1d
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:b3f503b1ac39ebc55840a436973f142eb7cf67e48a070284129fd1e30390c032
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:93e856dde88abb0aeb8e0b991e5f6d9f41884d8beea792f4d70cda06f18e783f
        - ghcr.io/mellanox/ipoib-cni@sha256:81646c07e27ce1df2603050a75dd267f264685fdc64c285eca3925ae98ed5392
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:47e1bb84ac97f9af49f97b49c2de44b4ba82b890b487880edc5c9adf86a176f7
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:b8a07696d05fb8e5292991ba731758579313d6983ed3d75dd95e7356024d987c
        - ghcr.io/mellanox/nic-feature-discovery@sha256:92dda9434519de39be0f94c80787766e3399e8b238d14839360aff4d6232a09b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:c41eaf76c06186e1dabb0d870225d946ad3ecda2a072fa4be197dd131b7695f2
        version: 24.7.1
      entries:
      - name: nvidia-network-operator.v24.7.1
        version: 24.7.1
      name: v24.7.1
    - currentCSV: nvidia-network-operator.v25.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "25.01-0.6.0.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "sha256:9f468fdc4449e65e4772575f83aa85840a00f97165f9a00ba34695c91d610fbd"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:12f0576f372f5adc17b60795952ae47388e50d8fad563b9595f3091eadd8eb29
          createdAt: "2025-02-26T16:55:35Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:f5c1741cc6368fcec5b250285552d9c80ff7254414f68102cc89162e12ac0400
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:cabce074d10a0f1d62135e2cc5442d65b49094b95b8297fdd024a1a5f461319f
        - ghcr.io/mellanox/nic-feature-discovery@sha256:92dda9434519de39be0f94c80787766e3399e8b238d14839360aff4d6232a09b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:5b91974c74f08fc5b1ad2edd4ca97d81b2dc6cc897045c3eac777826fe5d40d8
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:3edf0cec91eea10df17c9f182b2db430cb383887ac144e27ac44043b308284d7
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:68b3735c4edca2d889c456beba80ac2077b0cc0edba11a32769dacd3ece64892
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:a257aae9c4e02b62f9f4ddeea66c4349c517519a544dd02fe4c8d3fee5229521
        - ghcr.io/mellanox/ib-kubernetes@sha256:100e6addf99d5b44e81cb3aa8d2fc892fced1de3183190931961b92ee924207b
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:47e1bb84ac97f9af49f97b49c2de44b4ba82b890b487880edc5c9adf86a176f7
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:7604a30cbd0927eba76ddc74ea1cfbe113563ccf068e6ab3fe78123e4deec28d
        - ghcr.io/mellanox/ipoib-cni@sha256:56f19687b0b956ba45bf079cf3a655b0d1bad0314da9d5ecfe0f4d98ff31fc86
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:12f0576f372f5adc17b60795952ae47388e50d8fad563b9595f3091eadd8eb29
        - ghcr.io/mellanox/network-operator-init-container@sha256:67e93ccf3ecb61f17597567faf0f72e1b8ddcf73c5d7440baeadcc1cb6bb811b
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:9f468fdc4449e65e4772575f83aa85840a00f97165f9a00ba34695c91d610fbd
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        version: 25.1.0
      entries:
      - name: nvidia-network-operator.v25.1.0
        version: 25.1.0
      name: v25.1
    - currentCSV: nvidia-network-operator.v25.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "25.04-0.6.1.0-2"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "sha256:1eb5ab7f4b6f0cf89c954cf7ac8eb2e6fb2290a4f5f57417136a5302f20b12c8"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:93c2063f4da6a93810784be95df0083f07618486699d93245c97a29a63861dee
          createdAt: "2025-05-28T19:40:27Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:1b20b78f889339834ed74e0da621fc5da582719b2537b36d8967ddc6a04679b8
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:56a41010b4dbee0dc31e8f9c1467a739241a2d248ed8b5054d4b418f7f79f04d
        - ghcr.io/mellanox/nic-configuration-operator-daemon@sha256:0e8478898e0b30aec13030b11081b205308666c2a9c5c5bf2c00121079901b41
        - ghcr.io/mellanox/ipoib-cni@sha256:513d88d639a01ef98b65365f711cee7fbaa166137a56245b7ab32818f745418e
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:9bc09051aca763f83cde99ee8909cf989b1d6c5d6ab683fbb1ca441d47f61fdf
        - ghcr.io/mellanox/network-operator-init-container@sha256:67e93ccf3ecb61f17597567faf0f72e1b8ddcf73c5d7440baeadcc1cb6bb811b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:4a620bd39bc70928cd6e249855176ced0f4feb585c8c7624c0b9f5a3e57664c9
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:69e6ab6a0249be918f60c1e04fc53a07f280b33563f8edc953806538a9cc3c30
        - ghcr.io/mellanox/nic-configuration-operator@sha256:353ab27f5f5fba08197cbe8968222e1c115431cc1b9b8d8445f1facb3c3c4d7b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:35ff3b173b2694198e57b61720545be67e944c1da18ac004007a5b4e68562adc
        - ghcr.io/mellanox/ib-kubernetes@sha256:9774f774df6ede9ddf780b9193ca9d919618056e682d985e6b41e477b13c3432
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:cabce074d10a0f1d62135e2cc5442d65b49094b95b8297fdd024a1a5f461319f
        - ghcr.io/mellanox/nic-feature-discovery@sha256:35b1820fe94a833a43613ef5c9db1f58b293ea0dbcf96ec0742b0987f23f133c
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:76529c24ea17e68b89a2c0d4823d66504fa6f53759760a449308fda396574097
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:93c2063f4da6a93810784be95df0083f07618486699d93245c97a29a63861dee
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:1eb5ab7f4b6f0cf89c954cf7ac8eb2e6fb2290a4f5f57417136a5302f20b12c8
        version: 25.4.0
      entries:
      - name: nvidia-network-operator.v25.4.0
        version: 25.4.0
      name: v25.4
    defaultChannel: stable
    packageName: nvidia-network-operator
    provider:
      name: NVIDIA
      url: https://github.com/Mellanox/network-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: node-observability-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-observability-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservability",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "labels": {
                    "kubernetes.io/hostname": "worker-a"
                  },
                  "type": "crio-kubelet"
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservabilityMachineConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "debug": {
                    "enableCrioProfiling": true
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservabilityRun",
                "metadata": {
                  "name": "nodeobservabilityrun-sample",
                  "namespace": "node-observability-operator"
                },
                "spec": {
                  "nodeObservabilityRef": {
                    "name": "cluster"
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservability",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "nodeSelector": {
                    "kubernetes.io/hostname": "worker-a"
                  },
                  "type": "crio-kubelet"
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservabilityMachineConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "debug": {
                    "enableCrioProfiling": true
                  },
                  "nodeSelector": {
                    "kubernetes.io/hostname": "worker-a"
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservabilityRun",
                "metadata": {
                  "name": "nodeobservabilityrun-sample",
                  "namespace": "node-observability-operator"
                },
                "spec": {
                  "nodeObservabilityRef": {
                    "name": "cluster"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/noo/node-observability-rhel8-operator@sha256:53ba77f648154ac8aca1ae6c83815078dfde211dba1e42f597e69844e4954c01
          createdAt: 2022-12-05T09:25:09
          olm.skipRange: <0.2.0
          operatorframework.io/suggested-namespace: node-observability-operator
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeObservability prepares a subset of worker nodes (identified
              by a label) for running node observability queries, such as profiling
            displayName: Node Observability
            kind: NodeObservability
            name: nodeobservabilities.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservability prepares a subset of worker nodes (identified
              by a label) for running node observability queries, such as profiling
            displayName: Node Observability
            kind: NodeObservability
            name: nodeobservabilities.nodeobservability.olm.openshift.io
            version: v1alpha2
          - description: NodeObservabilityMachineConfig is the Schema for the nodeobservabilitymachineconfigs
              API
            displayName: Node Observability Machine Config
            kind: NodeObservabilityMachineConfig
            name: nodeobservabilitymachineconfigs.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservabilityMachineConfig is the Schema for the nodeobservabilitymachineconfigs
              API
            displayName: Node Observability Machine Config
            kind: NodeObservabilityMachineConfig
            name: nodeobservabilitymachineconfigs.nodeobservability.olm.openshift.io
            version: v1alpha2
          - description: NodeObservabilityRun is a request to run observability actions
              on the nodes previously selected in NodeObservability resource
            displayName: Node Observability Run
            kind: NodeObservabilityRun
            name: nodeobservabilityruns.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservabilityRun is a request to run observability actions
              on the nodes previously selected in NodeObservability resource
            displayName: Node Observability Run
            kind: NodeObservabilityRun
            name: nodeobservabilityruns.nodeobservability.olm.openshift.io
            version: v1alpha2
        description: |-
          An Operator that will be able to gather debugging/profiling data over
          a custom period of time which would be helpful to troubleshoot and resolve issues
          for OpenShift customers.
        displayName: Node Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node-observability-operator
        links:
        - name: Source Code
          url: https://github.com/openshift/node-observability-operator
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/noo/node-observability-rhel8-operator@sha256:53ba77f648154ac8aca1ae6c83815078dfde211dba1e42f597e69844e4954c01
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/noo/node-observability-agent-rhel8@sha256:a8e60d9f63e2ed0cf1e27309aac54014e3f1efc947b51ca0a99376338d220875
        version: 0.2.0
      entries:
      - name: node-observability-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: node-observability-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lib-bucket-provisioner
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lib-bucket-provisioner.v2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage,Big Data
          certified: "false"
          containerImage: quay.io/noobaa/pause
          createdAt: "2020-07-01T06:28:48Z"
          description: Library for the dynamic provisioning of object store buckets
            to be used by object store providers.
          repository: https://github.com/kube-object-storage/lib-bucket-provisioner
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: ObjectBucketClaim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information
            displayName: ObjectBucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: |
          ### CRD-only Operator

          This operator package is **CRD-only** and the operator is a no-op operator.

          Instead, bucket provisioners using this library are using these CRD's and using CSV [required-crds](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/Documentation/design/building-your-csv.md#required-crds) them so that OLM can install it as a dependency.

          ### Generic Bucket Provisioning

          Kubernetes natively supports dynamic provisioning for many types of file and block storage, but lacks support for object bucket provisioning.

          This repo is a placeholder for an object store bucket provisioning library, very similar to the Kubernetes [sig-storage-lib-external-provisioner](https://github.com/kubernetes-sigs/sig-storage-lib-external-provisioner/blob/master/controller/controller.go) library.

          ### Known Provisioners
          - https://github.com/noobaa/noobaa-operator (NooBaa)
          - https://github.com/rook/rook (Rook-Ceph)
          - https://github.com/yard-turkey/aws-s3-provisioner (AWS-S3)
        displayName: lib-bucket-provisioner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - object
        - bucket
        - storage
        - cloud
        - s3
        links:
        - name: Github
          url: https://github.com/kube-object-storage/lib-bucket-provisioner
        maintainers:
        - email: jcope@redhat.com
          name: Jon Cope
        - email: jvance@redhat.com
          name: Jeff Vance
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: dzaken@redhat.com
          name: Danny Zaken
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/noobaa/pause
        version: 2.0.0
      entries:
      - name: lib-bucket-provisioner.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: lib-bucket-provisioner
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Odigos
      provider-url: https://odigos.io
    name: odigos-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odigos-operator.v1.0.212
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.odigos.io/v1alpha1",
                "kind": "Odigos",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "odigos-operator"
                  },
                  "name": "odigos-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing
          containerImage: registry.connect.redhat.com/odigos/odigos-certified-operator-ubi9@sha256:c221eb2be22c2fcabcee728429bbca1255e0e31642c572e4dc40096d1d35e0d1
          createdAt: "2025-07-18T13:23:10Z"
          description: Odigos enables automatic distributed tracing with OpenTelemetry
            and eBPF.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Odigos Enterprise subscription
            (for enterprise features)
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Odigos
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Odigos is the Schema for the odigos API
            displayName: Odigos
            kind: Odigos
            name: odigos.operator.odigos.io
            version: v1alpha1
        description: |-
          The Odigos Operator provides options for installing and configuring Odigos.

          Upon creation of an Odigos resource, the Odigos Operator installs the following Odigos components:

          - Instrumentor
          - Odiglet
          - Autoscaler
          - Scheduler
          - Frontend UI

          With Odigos installed, follow the Odigos docs at docs.odigos.io for instructions on instrumenting applications and configuring destinations.
        displayName: Odigos Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - OpenTelemetry
        - eBPF
        - tracing
        - observability
        - distributed tracing
        - otel
        - monitoring
        links:
        - name: Odigos
          url: https://odigos.io
        - name: Odigos Documentation
          url: https://docs.odigos.io
        - name: Odigos on Github
          url: https://github.com/odigos-io/odigos
        maintainers:
        - email: mike@odigos.io
          name: Mike Dame
        maturity: alpha
        minKubeVersion: 1.20.15
        provider:
          name: Odigos
          url: https://odigos.io
        relatedImages:
        - registry.connect.redhat.com/odigos/odigos-certified-operator-ubi9@sha256:c221eb2be22c2fcabcee728429bbca1255e0e31642c572e4dc40096d1d35e0d1
        - registry.connect.redhat.com/odigos/odigos-ui-ubi9@sha256:b2a27cdfc94f1968a06ae18de1977127bb7d4e7d553e875d09a212693302f9c6
        - registry.connect.redhat.com/odigos/odigos-enterprise-instrumentor-ubi9@sha256:647e526cdbe38023bb221a802b180f0b95cb25734a0b39fa11cbbe22a07482a3
        - registry.connect.redhat.com/odigos/odigos-scheduler-ubi9@sha256:9e613c9b22d2fbfa3d21cc24f3b1e2dbc2e7287790e1f4286f1817c92565c865
        - registry.connect.redhat.com/odigos/odigos-instrumentor-ubi9@sha256:16712373c8e927b7208e43e9fe6e21180289975dbf3a81df60578e8550cf7150
        - registry.connect.redhat.com/odigos/odigos-odiglet-ubi9@sha256:770193901b4e5abcec59833f56f1530b43d4ec9514f3b5a5cb0bbb4356e06c67
        - registry.connect.redhat.com/odigos/odigos-enterprise-odiglet-ubi9@sha256:aec9b66461375994dc9aa92e5b2cc7d4151cf0332150726840cec049a7cbe7d9
        - registry.connect.redhat.com/odigos/odigos-autoscaler-ubi9@sha256:692193851b40f72ca79f7c82903136686d721c7fed77ef599037269ac6e55fdb
        - registry.connect.redhat.com/odigos/odigos-collector-ubi9@sha256:b67138d3ffd9e721602c5bf83f0c03ae1dcee230eca5e15335dcb9abced281ba
        version: 1.0.212
      entries:
      - name: odigos-operator.v1.0.212
        version: 1.0.212
      name: alpha
    defaultChannel: alpha
    packageName: odigos-operator
    provider:
      name: Odigos
      url: https://odigos.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-superset-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: superset-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:e0c9be4ac01a3d02f908f71f47f96bb1423fee14ba454c1bc627b35654d9cae3
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:e0c9be4ac01a3d02f908f71f47f96bb1423fee14ba454c1bc627b35654d9cae3
        version: 24.11.1
      entries:
      - name: superset-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: superset-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:f12cc3ed35135c14a82d49fcafed55465f0e66d46c2a36d532e5e09b9dece31e
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:f12cc3ed35135c14a82d49fcafed55465f0e66d46c2a36d532e5e09b9dece31e
        version: 25.3.0
      entries:
      - name: superset-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: superset-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
        version: 25.7.0
      entries:
      - name: superset-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: superset-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
        version: 25.7.0
      entries:
      - name: superset-operator.v25.7.0
        version: 25.7.0
      - name: superset-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-superset-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: LINBIT
      provider-url: ""
    name: linstor-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: linstor-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorController",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerImage": "",
                  "dbConnectionURL": "etcd://linstor-etcd:2379",
                  "drbdRepoCred": "",
                  "priorityClassName": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorCSIDriver",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerReplicas": 1,
                  "controllerEndpoint": "http://linstor:3370",
                  "csiControllerServiceAccountName": "csi-controller",
                  "csiNodeServiceAccountName": "csi-node",
                  "linstorPluginImage": "",
                  "imagePullSecret": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorSatelliteSet",
                "metadata": {
                  "name": "linstor-satellites"
                },
                "spec": {
                  "automaticStorageType": "None",
                  "controllerEndpoint": "http://linstor:3370",
                  "drbdRepoCred": "",
                  "kernelModuleInjectionMode": "ShippedModules",
                  "priorityClassName": "",
                  "satelliteImage": ""
                },
                "status": {
                  "errors": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          createdAt: "2021-06-22T07:57:53+00:00"
          description: LINSTOR Kubernetes Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/piraeusdatastore/piraeus-operator
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LinstorController is the Schema for the linstorcontrollersets
              API
            kind: LinstorController
            name: linstorcontrollers.linstor.linbit.com
            version: v1
          - description: LinstorCSIDriver is the Schema for the linstorcsidrivers
              API
            kind: LinstorCSIDriver
            name: linstorcsidrivers.linstor.linbit.com
            version: v1
          - description: LinstorSatelliteSet is the Schema for the linstornodesets
              API
            kind: LinstorSatelliteSet
            name: linstorsatellitesets.linstor.linbit.com
            version: v1
        description: |-
          LINSTOR is a configuration management system for storage on Linux systems.
          It manages LVM logical volumes and/or ZFS ZVOLs on a cluster of nodes.
          It leverages DRBD for replication between different nodes and to provide block
          storage devices to users and applications. It manages snapshots, encryption and
          caching of HDD backed data in SSDs via bcache.

          LINBIT provides a certified LINSTOR operator to ease deployment of LINSTOR
          on Openshift by installing DRBD, managing Satellite and Controller pods,
          and integrating with Openshift to provision persistent storage for your workloads.

          For detailed instructions and more configuration options, see our [user guide].

          [user guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift

          ## Install

          Unlike deployment via the helm chart, the certified Openshift
          operator does not deploy the needed etcd cluster. You must deploy this
          yourself ahead of time. We do this via the etcd operator available in the
          OperatorHub.

          IMPORTANT: It is advised that the etcd deployment uses persistent
          storage of some type. Either use an existing storage provisioner with
          a default `StorageClass` or simply use `hostPath` volumes.

          ### Installing the operator

          Hit "Install", select the stable update channel and a namespace for the
          operator. Use of a new namespace is recommended.

          Hit "Install" again. At this point you should have just one pod, the
          operator pod, running. Next we needs to configure the remaining provided APIs.

          #### A note on operator namespaces
          The LINSTOR operator can only watch for events and manage
          custom resources that are within the same namespace it is deployed
          within (OwnNamsespace). This means the LINSTOR Controller, LINSTOR
          Satellites, and LINSTOR CSI Driver pods all need to be deployed in the
          same namsepace as the LINSTOR Operator pod.

          ### Deploying the LINSTOR Controller

          Navigate to the left-hand control pane of the Openshift Web
          Console. Expand the "Operators" section, selecting "Installed Operators".
          Find the entry for the "Linstor Operator", then select the "LinstorController"
          from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorController". Click the "Create LinstorController" button.

          Here you will be presented with options to configure the LINSTOR
          Controller. Either via the web-form view or the YAML View. Regardless
          of which view you select, make sure that the `dbConnectionURL` matches
          the endpoint provided from your etcd deployment. Otherwise, the
          defaults are usually fine for most purposes.

          Lastly hit "Create", you should now see a linstor-controller pod
          running.

          ### Deploying the LINSTOR Satellites

          Next we need to deploy the Satellites Set. Just as before navigate
          to the left-hand control pane of the Openshift Web Console. Expand the
          "Operators" section, but this time select "Installed Operators". Find
          the entry for the "Linstor Operator", then select the
          "LinstorSatelliteSet" from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorSatelliteSet". Click the "Create LinstorSatelliteSet" button.

          Here you will be presented with the options to configure the LINSTOR
          Satellites. The defaults should be enough to get you started.
          Make sure the `controllerEndpoint` matches what is available in the
          openshift endpoints. The default is usually correct here.

          You can edit the `storagePools` section to configure LINSTOR storage pools,
          including preparing the backing devices. See our [storage guide].

          [storage guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-storage

          Below is an example manifest:

          ```yaml
          apiVersion: linstor.linbit.com/v1
          kind: LinstorSatelliteSet
          metadata:
            name: linstor-satellites
          spec:
            satelliteImage: ''
            automaticStorageType: None
            storagePools:
              lvmThinPools:
              - name: openshift-pool
                volumeGroup: ""
                thinVolume: openshift
                devicePaths:
                - /dev/vdb
            drbdRepoCred: ''
            kernelModuleInjectionMode: ShippedModules
            controllerEndpoint: 'http://linstor:3370'
            priorityClassName: ''
          ```

          Lastly hit "Create", you should now see a linstor-node pod
          running on every worker node.

          ### Deploying the LINSTOR CSI driver

          Last bit left is the CSI pods to bridge the layer between the CSI and
          LINSTOR. Just as before navigate to the left-hand control pane of the
          Openshift Web Console. Expand the "Operators" section, but this time
          select "Installed Operators". Find the entry for the "Linstor Operator",
          then select the "LinstorCSIDriver" from the "Provided APIs" column on the
          right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorCSIDriver". Click the "Create LinstorCSIDriver" button.

          Again, you will be presented with the options. Make sure that the
          `controllerEndpoint` is correct. Otherwise the defaults are fine for
          most use cases.

          Lastly hit "Create". You will now see a single "linstor-csi-controller" pod,
          as well as a "linstor-csi-node" pod on all worker nodes.

          ## Interacting with LINSTOR in Openshift.

          The Controller pod includes a LINSTOR Client, making it easy to interact directly with LINSTOR.
          For instance:

          ```
          oc exec deployment/linstor-cs-controller -- linstor storage-pool list
          ```

          This should only be necessary for investigating problems and accessing advanced functionality.
          Regular operation such as creating volumes should be achieved via the [Openshift/Kubernetes integration].

          [Openshift/Kubernetes integration]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-basic-configuration-and-deployment
        displayName: Linstor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - DRBD
        links:
        - name: LINBIT
          url: https://www.linbit.com/
        - name: User Guide
          url: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift
        maintainers:
        - email: sales@linbit.com
          name: LINBIT
        maturity: beta
        provider:
          name: LINBIT
        relatedImages:
        - registry.marketplace.redhat.com/rhm/linbit/linstor-operator@sha256:75e3ec9a9331072dc88c13979f0cacdf6232f8b3a6bfaf55c95b2ee3a03a5d5b
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:395a5a4aa4cfe3a0093d2225ce2e67acdcec0fd894e4b61e30a750f22931448d
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:bba872052e42fe0d6cbb2eea6541bf57f035575cbc7d838a3260b8582b8fe572
        - registry.marketplace.redhat.com/rhm/linbit/linstor-controller@sha256:1084ba14eaec7511682dd87c562398ce5e6690ded700c964373188b73bfe3b17
        - registry.marketplace.redhat.com/rhm/linbit/drbd-9@sha256:6337b3adfa3de87e74adcc974edafa09b618cb8e0315658c1cbdf8dd1c6c4c32
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:ba217e7e51f0748b54f477486263ff4983b0029635fd1f7ccc67160a09f81608
        - registry.marketplace.redhat.com/rhm/linbit/linstor-csi@sha256:034e3641376aa7ad47ed309fb389083c3ff82974f246f4d71b8789be8f8f618d
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:53b0f48eeecaa3ff1e211a2a413cd49266b7be85a5aa3a0431ab72d32e3a4e5c
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:8e52a6b31f9f3834d709dbca3614c2110b8ebb4af382c76b7056921221c87600
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:91cfe63acc12c91c3053ba7cc12d72cce404d33aab4afcfac285311f5ddfc7fe
        - registry.marketplace.redhat.com/rhm/linbit/linstor-satellite@sha256:4a3503092c8cb9435ccfb4837ee6a9183b680eb18a35c7d40227fa5c8f4da996
        version: 1.5.1
      entries:
      - name: linstor-operator.v1.5.1
        version: 1.5.1
      name: alpha
    - currentCSV: linstor-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorController",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerImage": "",
                  "dbConnectionURL": "etcd://linstor-etcd:2379",
                  "drbdRepoCred": "",
                  "priorityClassName": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorCSIDriver",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerReplicas": 1,
                  "controllerEndpoint": "http://linstor:3370",
                  "csiControllerServiceAccountName": "csi-controller",
                  "csiNodeServiceAccountName": "csi-node",
                  "linstorPluginImage": "",
                  "imagePullSecret": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorSatelliteSet",
                "metadata": {
                  "name": "linstor-satellites"
                },
                "spec": {
                  "automaticStorageType": "None",
                  "controllerEndpoint": "http://linstor:3370",
                  "drbdRepoCred": "",
                  "kernelModuleInjectionMode": "ShippedModules",
                  "priorityClassName": "",
                  "satelliteImage": ""
                },
                "status": {
                  "errors": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          createdAt: "2021-06-22T07:57:53+00:00"
          description: LINSTOR Kubernetes Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/piraeusdatastore/piraeus-operator
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LinstorController is the Schema for the linstorcontrollersets
              API
            kind: LinstorController
            name: linstorcontrollers.linstor.linbit.com
            version: v1
          - description: LinstorCSIDriver is the Schema for the linstorcsidrivers
              API
            kind: LinstorCSIDriver
            name: linstorcsidrivers.linstor.linbit.com
            version: v1
          - description: LinstorSatelliteSet is the Schema for the linstornodesets
              API
            kind: LinstorSatelliteSet
            name: linstorsatellitesets.linstor.linbit.com
            version: v1
        description: |-
          LINSTOR is a configuration management system for storage on Linux systems.
          It manages LVM logical volumes and/or ZFS ZVOLs on a cluster of nodes.
          It leverages DRBD for replication between different nodes and to provide block
          storage devices to users and applications. It manages snapshots, encryption and
          caching of HDD backed data in SSDs via bcache.

          LINBIT provides a certified LINSTOR operator to ease deployment of LINSTOR
          on Openshift by installing DRBD, managing Satellite and Controller pods,
          and integrating with Openshift to provision persistent storage for your workloads.

          For detailed instructions and more configuration options, see our [user guide].

          [user guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift

          ## Install

          Unlike deployment via the helm chart, the certified Openshift
          operator does not deploy the needed etcd cluster. You must deploy this
          yourself ahead of time. We do this via the etcd operator available in the
          OperatorHub.

          IMPORTANT: It is advised that the etcd deployment uses persistent
          storage of some type. Either use an existing storage provisioner with
          a default `StorageClass` or simply use `hostPath` volumes.

          ### Installing the operator

          Hit "Install", select the stable update channel and a namespace for the
          operator. Use of a new namespace is recommended.

          Hit "Install" again. At this point you should have just one pod, the
          operator pod, running. Next we needs to configure the remaining provided APIs.

          #### A note on operator namespaces
          The LINSTOR operator can only watch for events and manage
          custom resources that are within the same namespace it is deployed
          within (OwnNamsespace). This means the LINSTOR Controller, LINSTOR
          Satellites, and LINSTOR CSI Driver pods all need to be deployed in the
          same namsepace as the LINSTOR Operator pod.

          ### Deploying the LINSTOR Controller

          Navigate to the left-hand control pane of the Openshift Web
          Console. Expand the "Operators" section, selecting "Installed Operators".
          Find the entry for the "Linstor Operator", then select the "LinstorController"
          from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorController". Click the "Create LinstorController" button.

          Here you will be presented with options to configure the LINSTOR
          Controller. Either via the web-form view or the YAML View. Regardless
          of which view you select, make sure that the `dbConnectionURL` matches
          the endpoint provided from your etcd deployment. Otherwise, the
          defaults are usually fine for most purposes.

          Lastly hit "Create", you should now see a linstor-controller pod
          running.

          ### Deploying the LINSTOR Satellites

          Next we need to deploy the Satellites Set. Just as before navigate
          to the left-hand control pane of the Openshift Web Console. Expand the
          "Operators" section, but this time select "Installed Operators". Find
          the entry for the "Linstor Operator", then select the
          "LinstorSatelliteSet" from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorSatelliteSet". Click the "Create LinstorSatelliteSet" button.

          Here you will be presented with the options to configure the LINSTOR
          Satellites. The defaults should be enough to get you started.
          Make sure the `controllerEndpoint` matches what is available in the
          openshift endpoints. The default is usually correct here.

          You can edit the `storagePools` section to configure LINSTOR storage pools,
          including preparing the backing devices. See our [storage guide].

          [storage guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-storage

          Below is an example manifest:

          ```yaml
          apiVersion: linstor.linbit.com/v1
          kind: LinstorSatelliteSet
          metadata:
            name: linstor-satellites
          spec:
            satelliteImage: ''
            automaticStorageType: None
            storagePools:
              lvmThinPools:
              - name: openshift-pool
                volumeGroup: ""
                thinVolume: openshift
                devicePaths:
                - /dev/vdb
            drbdRepoCred: ''
            kernelModuleInjectionMode: ShippedModules
            controllerEndpoint: 'http://linstor:3370'
            priorityClassName: ''
          ```

          Lastly hit "Create", you should now see a linstor-node pod
          running on every worker node.

          ### Deploying the LINSTOR CSI driver

          Last bit left is the CSI pods to bridge the layer between the CSI and
          LINSTOR. Just as before navigate to the left-hand control pane of the
          Openshift Web Console. Expand the "Operators" section, but this time
          select "Installed Operators". Find the entry for the "Linstor Operator",
          then select the "LinstorCSIDriver" from the "Provided APIs" column on the
          right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorCSIDriver". Click the "Create LinstorCSIDriver" button.

          Again, you will be presented with the options. Make sure that the
          `controllerEndpoint` is correct. Otherwise the defaults are fine for
          most use cases.

          Lastly hit "Create". You will now see a single "linstor-csi-controller" pod,
          as well as a "linstor-csi-node" pod on all worker nodes.

          ## Interacting with LINSTOR in Openshift.

          The Controller pod includes a LINSTOR Client, making it easy to interact directly with LINSTOR.
          For instance:

          ```
          oc exec deployment/linstor-cs-controller -- linstor storage-pool list
          ```

          This should only be necessary for investigating problems and accessing advanced functionality.
          Regular operation such as creating volumes should be achieved via the [Openshift/Kubernetes integration].

          [Openshift/Kubernetes integration]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-basic-configuration-and-deployment
        displayName: Linstor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - DRBD
        links:
        - name: LINBIT
          url: https://www.linbit.com/
        - name: User Guide
          url: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift
        maintainers:
        - email: sales@linbit.com
          name: LINBIT
        maturity: beta
        provider:
          name: LINBIT
        relatedImages:
        - registry.marketplace.redhat.com/rhm/linbit/linstor-csi@sha256:034e3641376aa7ad47ed309fb389083c3ff82974f246f4d71b8789be8f8f618d
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:8e52a6b31f9f3834d709dbca3614c2110b8ebb4af382c76b7056921221c87600
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:395a5a4aa4cfe3a0093d2225ce2e67acdcec0fd894e4b61e30a750f22931448d
        - registry.marketplace.redhat.com/rhm/linbit/linstor-controller@sha256:1084ba14eaec7511682dd87c562398ce5e6690ded700c964373188b73bfe3b17
        - registry.marketplace.redhat.com/rhm/linbit/linstor-satellite@sha256:4a3503092c8cb9435ccfb4837ee6a9183b680eb18a35c7d40227fa5c8f4da996
        - registry.marketplace.redhat.com/rhm/linbit/linstor-operator@sha256:75e3ec9a9331072dc88c13979f0cacdf6232f8b3a6bfaf55c95b2ee3a03a5d5b
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:53b0f48eeecaa3ff1e211a2a413cd49266b7be85a5aa3a0431ab72d32e3a4e5c
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:bba872052e42fe0d6cbb2eea6541bf57f035575cbc7d838a3260b8582b8fe572
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:ba217e7e51f0748b54f477486263ff4983b0029635fd1f7ccc67160a09f81608
        - registry.marketplace.redhat.com/rhm/linbit/drbd-9@sha256:6337b3adfa3de87e74adcc974edafa09b618cb8e0315658c1cbdf8dd1c6c4c32
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:91cfe63acc12c91c3053ba7cc12d72cce404d33aab4afcfac285311f5ddfc7fe
        version: 1.5.1
      entries:
      - name: linstor-operator.v1.5.1
        version: 1.5.1
      name: stable
    defaultChannel: stable
    packageName: linstor-operator-rhmp
    provider:
      name: LINBIT
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Mantisnet
      provider-url: https://www.mantisnet.com
    name: tawon-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tawon-operator.v2.29.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Directive",
                "metadata": {
                  "name": "amf-h2"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "amf"
                    }
                  },
                  "container": {
                    "resources": {
                      "limits": {
                        "cpu": "500m"
                      }
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  },
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 8080"
                      },
                      "task": "payload"
                    },
                    {
                      "task": "h2"
                    },
                    {
                      "config": {
                        "addr": "nats://nats:4222",
                        "topic": "hello"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "TopologyAggregator",
                "metadata": {
                  "name": "topology-aggregator"
                },
                "spec": {
                  "messaging": {
                    "nats": {
                      "url": "nats://nats:4222"
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-11-13T01:16:29Z"
          olm.skipRange: '>=2.23.0 <2.29.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Directive allows to define a Tawon task pipeline that will
              be applied to a Kubernetes cluster.
            displayName: Tawon Directive
            kind: Directive
            name: directives.tawon.mantisnet.com
            version: v1alpha1
          - description: TopologyAggregator is the Schema for the topologyaggregators
              API
            displayName: Tawon Topology Aggregator
            kind: TopologyAggregator
            name: topologyaggregators.tawon.mantisnet.com
            version: v1alpha1
        description: Tawon is a cloud-native observability fabric for deeply instrumenting
          microservice-based applications.
        displayName: Tawon
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - networking
        - metrics
        - ebpf
        links:
        - name: Tawon Docs
          url: https://docs.tawon.xmantis.net
        maintainers:
        - email: elliott@mantisnet.com
          name: Elliott Starin
        - email: nicolas@mantisnet.com
          name: Nicolas Paton
        - email: thiago@mantisnet.com
          name: Thiago Navarro
        maturity: alpha
        provider:
          name: Mantisnet
          url: https://www.mantisnet.com
        relatedImages:
        - quay.io/mantisnet/kernel-headers@sha256:242e837af831b6eed98d522b6bae7416928b9fbe451fb6e02ffa678686771068
        - quay.io/mantisnet/kernel-headers@sha256:2aa3bc7871fe0fc0a79ac812d941e55cd0d25a5f0d1d9b3e2435aab6f3d70c67
        - quay.io/mantisnet/tawonctl@sha256:d436ef902bb08d830f64f34fdb86b20d2bd18d86233c8240014f0722c60e5e94
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - quay.io/mantisnet/tawon-operator@sha256:d5669dde46bb30411964baa86844d97721655886769b96809626246a4123caad
        - quay.io/mantisnet/tawon@sha256:83575079fe00ae6e53bb13005b94ef0047cc1974d95015c19e64dd53e13bcb93
        - quay.io/mantisnet/nats@sha256:6afc7ced92181a7c12d709e9693740847297dc5d60ab78cbac7b9b6c859ecdd1
        version: 2.29.1
      entries:
      - name: tawon-operator.v2.29.1
        version: 2.29.1
      name: alpha
    - currentCSV: tawon-operator.v2.39.27-rc6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "ClusterDirective",
                "metadata": {
                  "name": "capture-coredns"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "coredns"
                    }
                  },
                  "debug": {
                    "logLevel": "debug"
                  },
                  "streams": [
                    {
                      "maxage": "6h0m0s",
                      "maxmsgs": 1000,
                      "name": "coredns",
                      "retentionPolicy": "Delete"
                    }
                  ],
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 53"
                      },
                      "task": "capture"
                    },
                    {
                      "config": {
                        "name": "coredns",
                        "type": "stream"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Dashboard",
                "metadata": {
                  "name": "tawon-dashboard"
                },
                "spec": null
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Directive",
                "metadata": {
                  "name": "capture-coredns"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "coredns"
                    }
                  },
                  "debug": {
                    "logLevel": "debug"
                  },
                  "streams": [
                    {
                      "maxage": "6h0m0s",
                      "maxmsgs": 1000,
                      "name": "coredns",
                      "retentionPolicy": "Delete"
                    }
                  ],
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 53"
                      },
                      "task": "capture"
                    },
                    {
                      "config": {
                        "name": "coredns",
                        "type": "stream"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Stream",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "tawon-operator",
                    "app.kubernetes.io/instance": "stream-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "stream",
                    "app.kubernetes.io/part-of": "tawon-operator"
                  },
                  "name": "stream-sample"
                },
                "spec": {
                  "maxage": "6h",
                  "maxbytes": "1G",
                  "maxmsgs": 100000,
                  "store": {
                    "name": "mystreamstore"
                  }
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "StreamStore",
                "metadata": {
                  "name": "mystreamstore"
                },
                "spec": {
                  "defaults": {
                    "maxage": "240h",
                    "maxbytes": "1Gi",
                    "maxsize": 1000000
                  },
                  "limits": {
                    "max": 10000000,
                    "maxage": "2400h",
                    "maxbytes": "2Gi"
                  },
                  "persistentVolumeClaimRetentionPolicy": {
                    "whenDeleted": "Delete"
                  },
                  "volumeClaimTemplate": {
                    "metadata": {
                      "name": "mypvclaim"
                    },
                    "spec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "ssd-csi"
                    }
                  }
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "TopologyAggregator",
                "metadata": {
                  "name": "topology-aggregator"
                },
                "spec": {
                  "messaging": {
                    "nats": {
                      "url": "nats://nats:4222"
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing, Monitoring, Networking
          containerImage: quay.io/mantisnet/tawon-operator@sha256:d8ce5c8966d3590ce5f286d8e778463b85582b970ad7730ddf7828bc68843a90
          createdAt: "2025-03-08T18:58:04Z"
          description: Tawon is a cloud-native observability fabric for deeply instrumenting
            microservice-based applications.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Mantisnet Tawon Operator Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: MantisNet
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterDirective is the Schema for the clusterdirectives
              API.
            displayName: Cluster Tawon Directive
            kind: ClusterDirective
            name: clusterdirectives.tawon.mantisnet.com
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.tawon.mantisnet.com
            version: v1alpha1
          - description: A Directive allows to define a Tawon task pipeline that will
              be applied to a Kubernetes cluster.
            displayName: Tawon Directive
            kind: Directive
            name: directives.tawon.mantisnet.com
            version: v1alpha1
          - description: Stream is a persistent Tawon data stream.
            displayName: Stream
            kind: Stream
            name: streams.tawon.mantisnet.com
            version: v1alpha1
          - description: StreamStore is the Schema for the streamstores API
            displayName: Stream Store
            kind: StreamStore
            name: streamstores.tawon.mantisnet.com
            version: v1alpha1
          - description: TopologyAggregator is the Schema for the topologyaggregators
              API
            displayName: Tawon Topology Aggregator
            kind: TopologyAggregator
            name: topologyaggregators.tawon.mantisnet.com
            version: v1alpha1
        description: Tawon is a cloud-native observability fabric for deeply instrumenting
          microservice-based applications.
        displayName: Tawon
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - networking
        - metrics
        - ebpf
        links:
        - name: Tawon Docs
          url: https://docs.tawon.xmantis.net
        maintainers:
        - email: elliott@mantisnet.com
          name: Elliott Starin
        - email: nicolas@mantisnet.com
          name: Nicolas Paton
        - email: thiago@mantisnet.com
          name: Thiago Navarro
        maturity: alpha
        provider:
          name: Mantisnet
          url: https://www.mantisnet.com
        relatedImages:
        - quay.io/mantisnet/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/mantisnet/kernel-headers@sha256:ca3baeba7ab8ce968c1ee129112532fc8b8d6fb6741e42831b6100f0a2451629
        - quay.io/mantisnet/tawon-operator@sha256:d8ce5c8966d3590ce5f286d8e778463b85582b970ad7730ddf7828bc68843a90
        - quay.io/mantisnet/diagnose@sha256:ab15f4eaeda32f510d81a3cc865b0fd77405b55f6a0c6f285131331dcc38f2ca
        - quay.io/mantisnet/tawonctl@sha256:8bf40ca0cc9984a252363e87d6135e43ac52839113b762399cc54a085af171d8
        - quay.io/mantisnet/sengat@sha256:00c3e1194c2c3e343d16c1543b7ce2a1ee9829f6ddd1c2870785624596b46ebf
        - quay.io/mantisnet/kernel-headers@sha256:9290eb56b186ee0edf65da615c1a8aabc66f7d290168e49ad8764ab8fc2d209a
        - quay.io/mantisnet/tawon@sha256:46e654c6f275be2be604254102aaee9d927859cfd05dfa16b7cb571d36b0f1a2
        - quay.io/mantisnet/tawon-tests@sha256:cd49c24deb7cc9f4b439cb9baa885208dfe91de385cd8c5d312fd08280615c2c
        - quay.io/openshift/origin-oauth-proxy@sha256:28856aae9e481fe763776df713b4b335632645687e6fd2e31619f76d11e0f47c
        - quay.io/mantisnet/tawon@sha256:5a8bf475ff18abacb96f631e4d3ea01c94bef8a38f84c00d1dc4047073b6cc14
        - quay.io/mantisnet/nats@sha256:5f9e7f8f685fd0c1279a3465b61c7215f4600c99d81785dda86a71c966cf829a
        - quay.io/mantisnet/kernel-headers@sha256:5d00bfa9f60c95afcfa363ede59a9625fef9925cd1cb4235327a1de94a0fc632
        version: 2.39.27-rc6
      entries:
      - name: tawon-operator.v2.39.27-rc6
        version: 2.39.27-rc6
      name: candidate-v2
    - currentCSV: tawon-operator.v2.39.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "ClusterDirective",
                "metadata": {
                  "name": "capture-coredns"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "coredns"
                    }
                  },
                  "debug": {
                    "logLevel": "debug"
                  },
                  "streams": [
                    {
                      "maxage": "6h0m0s",
                      "maxmsgs": 1000,
                      "name": "coredns",
                      "retentionPolicy": "Delete"
                    }
                  ],
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 53"
                      },
                      "task": "capture"
                    },
                    {
                      "config": {
                        "name": "coredns",
                        "type": "stream"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Dashboard",
                "metadata": {
                  "name": "tawon-dashboard"
                },
                "spec": null
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Directive",
                "metadata": {
                  "name": "capture-coredns"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "coredns"
                    }
                  },
                  "debug": {
                    "logLevel": "debug"
                  },
                  "streams": [
                    {
                      "maxage": "6h0m0s",
                      "maxmsgs": 1000,
                      "name": "coredns",
                      "retentionPolicy": "Delete"
                    }
                  ],
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 53"
                      },
                      "task": "capture"
                    },
                    {
                      "config": {
                        "name": "coredns",
                        "type": "stream"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Stream",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "tawon-operator",
                    "app.kubernetes.io/instance": "stream-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "stream",
                    "app.kubernetes.io/part-of": "tawon-operator"
                  },
                  "name": "stream-sample"
                },
                "spec": {
                  "maxage": "6h",
                  "maxbytes": "1G",
                  "maxmsgs": 100000,
                  "store": {
                    "name": "mystreamstore"
                  }
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "StreamStore",
                "metadata": {
                  "name": "mystreamstore"
                },
                "spec": {
                  "defaults": {
                    "maxage": "240h",
                    "maxbytes": "1Gi",
                    "maxsize": 1000000
                  },
                  "limits": {
                    "max": 10000000,
                    "maxage": "2400h",
                    "maxbytes": "2Gi"
                  },
                  "persistentVolumeClaimRetentionPolicy": {
                    "whenDeleted": "Delete"
                  },
                  "volumeClaimTemplate": {
                    "metadata": {
                      "name": "mypvclaim"
                    },
                    "spec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "ssd-csi"
                    }
                  }
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "TopologyAggregator",
                "metadata": {
                  "name": "topology-aggregator"
                },
                "spec": {
                  "messaging": {
                    "nats": {
                      "url": "nats://nats:4222"
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing, Monitoring, Networking
          containerImage: quay.io/mantisnet/tawon-operator@sha256:f831abf0f9fb3a4e096743d4ff7356984738da25790e305325e740ca97f0be5b
          createdAt: "2025-07-18T16:24:07Z"
          description: Tawon is a cloud-native observability fabric for deeply instrumenting
            microservice-based applications.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.23.0 <2.39.32'
          operators.openshift.io/valid-subscription: Mantisnet Tawon Operator Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: MantisNet
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterDirective is the Schema for the clusterdirectives
              API.
            displayName: Cluster Tawon Directive
            kind: ClusterDirective
            name: clusterdirectives.tawon.mantisnet.com
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.tawon.mantisnet.com
            version: v1alpha1
          - description: A Directive allows to define a Tawon task pipeline that will
              be applied to a Kubernetes cluster.
            displayName: Tawon Directive
            kind: Directive
            name: directives.tawon.mantisnet.com
            version: v1alpha1
          - description: Stream is a persistent Tawon data stream.
            displayName: Stream
            kind: Stream
            name: streams.tawon.mantisnet.com
            version: v1alpha1
          - description: StreamStore is the Schema for the streamstores API
            displayName: Stream Store
            kind: StreamStore
            name: streamstores.tawon.mantisnet.com
            version: v1alpha1
          - description: TopologyAggregator is the Schema for the topologyaggregators
              API
            displayName: Tawon Topology Aggregator
            kind: TopologyAggregator
            name: topologyaggregators.tawon.mantisnet.com
            version: v1alpha1
        description: Tawon is a cloud-native observability fabric for deeply instrumenting
          microservice-based applications.
        displayName: Tawon
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - networking
        - metrics
        - ebpf
        links:
        - name: Tawon Docs
          url: https://docs.tawon.xmantis.net
        maintainers:
        - email: elliott@mantisnet.com
          name: Elliott Starin
        - email: nicolas@mantisnet.com
          name: Nicolas Paton
        - email: thiago@mantisnet.com
          name: Thiago Navarro
        maturity: alpha
        provider:
          name: Mantisnet
          url: https://www.mantisnet.com
        relatedImages:
        - quay.io/mantisnet/nats@sha256:5f9e7f8f685fd0c1279a3465b61c7215f4600c99d81785dda86a71c966cf829a
        - quay.io/mantisnet/tawon-tests@sha256:294df1580816919502dc592c1c28ffff32b13ee892686f783eff181863fa9d18
        - quay.io/mantisnet/tawon@sha256:1640758505d9b4e0f684ab502ef3344135a270192054e256e7e711209fa51adf
        - quay.io/mantisnet/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/mantisnet/nginx@sha256:c9f91949187fa1c2b4615b88d3acf7902c7e2d4a2557f33ca0cf90164269a7ae
        - quay.io/mantisnet/tawonctl@sha256:88ac81b4fe00d360c5ff127f77456f9e36990e435442a9194d3dbf50b20e7206
        - quay.io/mantisnet/diagnose@sha256:ab15f4eaeda32f510d81a3cc865b0fd77405b55f6a0c6f285131331dcc38f2ca
        - quay.io/mantisnet/tawon@sha256:846d03ada7e6f6648fc74fd0ec9e68c3c6b17e4c155df22829a3db3ca556c395
        - quay.io/mantisnet/kernel-headers@sha256:9290eb56b186ee0edf65da615c1a8aabc66f7d290168e49ad8764ab8fc2d209a
        - quay.io/mantisnet/tawon-operator@sha256:f831abf0f9fb3a4e096743d4ff7356984738da25790e305325e740ca97f0be5b
        - quay.io/mantisnet/alpine@sha256:ec1b05d1eac264d9204a57f4ad9d4dc35e9e756e9fedaea0674aefc7edb1d6a4
        - quay.io/mantisnet/kernel-headers@sha256:ca3baeba7ab8ce968c1ee129112532fc8b8d6fb6741e42831b6100f0a2451629
        - quay.io/mantisnet/kernel-headers@sha256:5d00bfa9f60c95afcfa363ede59a9625fef9925cd1cb4235327a1de94a0fc632
        - quay.io/mantisnet/kernel-headers@sha256:b0975a44cc94bc1d75679f288a9e23a65c2c0b350ce1fb2f5eb421a76a9567c6
        - quay.io/mantisnet/sengat@sha256:483fbce2bacbd95b7458773e2278b936f06a19e65acb8987adffa8403d76e40a
        - quay.io/openshift/origin-oauth-proxy@sha256:28856aae9e481fe763776df713b4b335632645687e6fd2e31619f76d11e0f47c
        version: 2.39.32
      entries:
      - name: tawon-operator.v2.39.32
        version: 2.39.32
      - name: tawon-operator.v2.39.31
        version: 2.39.31
      - name: tawon-operator.v2.39.29
        version: 2.39.29
      - name: tawon-operator.v2.39.28
        version: 2.39.28
      - name: tawon-operator.v2.39.27
        version: 2.39.27
      name: fast-v2
    defaultChannel: fast-v2
    packageName: tawon-operator
    provider:
      name: Mantisnet
      url: https://www.mantisnet.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Perficient Inc
      provider-url: http://www.perficient.com
    name: devopsinabox
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devopsinabox.v0.0.45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "DevSecOpsManager",
                "metadata": {
                  "name": "devsecopsmanager-sample"
                },
                "spec": {
                  "config": {
                    "repoPath": "",
                    "repoUrl": ""
                  },
                  "forwarding": {
                    "fluentd": {
                      "affinity": {},
                      "caFile": "files/default-openshift-logging-fluentd.crt",
                      "image": "registry.redhat.io/openshift4/ose-logging-fluentd:v4.4",
                      "keyFile": "files/default-openshift-logging-fluentd.key",
                      "loglevel": "warn",
                      "nodeSelector": {},
                      "passphrase": "",
                      "persistence": {
                        "accessMode": "ReadWriteOnce",
                        "enabled": false,
                        "size": "10Gi",
                        "storageClass": ""
                      },
                      "port": 24224,
                      "replicas": 2,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "512Mi"
                        }
                      },
                      "sharedkey": "splunkforwarding",
                      "ssl": true,
                      "tolerations": [],
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      }
                    },
                    "preserveDefault": true
                  },
                  "fullnameOverride": "",
                  "git": {
                    "password": "",
                    "server": "https://gitlab.com",
                    "token": "",
                    "username": ""
                  },
                  "nameOverride": "",
                  "prometheus": {
                    "token": ""
                  }
                }
              },
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "ManagedNamespace",
                "metadata": {
                  "name": "managednamespace-sample"
                },
                "spec": {
                  "config": {
                    "repoBranch": "master",
                    "repoPath": "",
                    "repoUrl": ""
                  },
                  "container": {
                    "password": "",
                    "registry": "https://quay.io",
                    "username": ""
                  },
                  "fullnameOverride": "",
                  "git": {
                    "password": "",
                    "server": "https://gitlab.com",
                    "token": "",
                    "username": ""
                  },
                  "manager": {
                    "namespace": ""
                  },
                  "mirrors": {
                    "maven": ""
                  },
                  "nameOverride": ""
                }
              },
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "SpringBootApp",
                "metadata": {
                  "name": "springbootapp-sample"
                },
                "spec": {
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "config": {
                    "repo": ""
                  },
                  "container": {
                    "image": "placeholder",
                    "registry": ""
                  },
                  "fullnameOverride": "",
                  "git": {
                    "branch": "master",
                    "module": "",
                    "url": ""
                  },
                  "management": {
                    "port": 8081
                  },
                  "maven": {
                    "mirror": ""
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "replicaCount": 1,
                  "resources": {},
                  "service": {
                    "port": 8080
                  },
                  "serviceAccount": {
                    "annotations": {}
                  },
                  "version": {
                    "branch": "",
                    "image": "placeholder",
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration,Developer Tools,Application Runtime
          containerImage: quay.io/mrethers/devopsinabox-operator:v0.0.45
          createdAt: 2021-03-03 10:51:09Z
          description: Perficient's DevSecOps Manager For Kubernetes Applications
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/prft-rh/devopsinabox-operator
          support: Matthieu Rethers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DevSecOpsManager
            name: devsecopsmanagers.apimc.com.perficient
            version: v1alpha1
          - kind: ManagedNamespace
            name: managednamespaces.apimc.com.perficient
            version: v1alpha1
          - kind: SpringBootApp
            name: springbootapps.apimc.com.perficient
            version: v1alpha1
        description: "The devops-in-a-box Operator manages the entire lifecycle of
          your applications in Kubernetes. It encapsulates all the best devsecops
          practices gathered by Perficient over the years working on many large projects.
          Use it to migrate existing applications with minimum changes or deploy new
          ones with little to no knowledge of Kubernetes.\n\nOnce installed, the Devops-in-a-box
          operator provides the following features:\n\n* **Create/Destroy**: Easily
          launch a new DevSecOpsManager instance to bootstrap your cluster and start
          deploying new applications with total confidence.\n\n* **Gitops Simplified**:
          Configure your cluster and individual applications through Git and watch
          your changes propagate instantly. Just point to where your code is and let
          the operator do the rest with preconfigured cloud-native CI/CD pipelines.\n\n*
          **Production-Ready**: Automatically register your applications with Openshift
          logging and monitoring systems and enjoy special additions for Spring Boot
          like custom dashboards and searchable logs.\n\n*Contact us on Github for
          installation and usage instruction*\n\n### Other Supported Features\n\n**Blue/Green,
          Canary, A/B Deployments**\n\nRun multiple versions of your applications
          simultaneously and configure traffic rules to target specific groups of
          users using Service Mesh.\n\n**Updates via automated operations**\n\nGet
          the latest enhancements automatically through the Operator Lifecycle Manager
          and upgrade all your running applications at once. "
        displayName: Devops-in-a-box
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devops
        - devsecops
        - spring
        - boot
        - java
        - gitops
        - cicd
        - argo
        links:
        - name: Devops-in-a-box Operator
          url: https://github.com/prft-rh/devopsinabox-operator
        maintainers:
        - email: matthieu.rethers@perficient.com
          name: Matthieu Rethers
        maturity: alpha
        provider:
          name: Perficient Inc
          url: http://www.perficient.com
        relatedImages:
        - quay.io/mrethers/devopsinabox-operator:v0.0.45
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        version: 0.0.45
      entries:
      - name: devopsinabox.v0.0.45
        version: 0.0.45
      name: alpha
    defaultChannel: alpha
    packageName: devopsinabox
    provider:
      name: Perficient Inc
      url: http://www.perficient.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: dapr.io
      provider-url: https://dapr.io
    name: dapr-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dapr-kubernetes-operator.v0.0.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprControlPlane",
                "metadata": {
                  "name": "dapr-control-plane",
                  "namespace": "dapr-system"
                },
                "spec": {
                  "values": {}
                }
              },
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprCruiseControl",
                "metadata": {
                  "name": "dapr-cruise-control",
                  "namespace": "dapr-system"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprInstance",
                "metadata": {
                  "name": "dapr-instance",
                  "namespace": "dapr-system"
                },
                "spec": {
                  "values": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          containerImage: docker.io/daprio/kubernetes-operator@sha256:da599a273943bf28812314a11df1052411237dd1887edf0fe6104c5857f439b0
          createdAt: "2025-10-06T18:47:59Z"
          description: Dapr Control Plane Operator
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/dapr/kubernetes-operator
          support: contact@dapr.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DaprControlPlane is the Schema for the Dapr ControlPlane
              API.
            displayName: DaprControlPlane
            kind: DaprControlPlane
            name: daprcontrolplanes.operator.dapr.io
            version: v1alpha1
          - kind: DaprCruiseControl
            name: daprcruiscontrols.operator.dapr.io
            version: v1alpha1
          - description: DaprInstance is the Schema for the Dapr Instance API.
            displayName: DaprInstance
            kind: DaprInstance
            name: daprinstances.operator.dapr.io
            version: v1alpha1
        description: Dapr Control Plane Operator
        displayName: Dapr Control Plane Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dapr
        links:
        - name: Dapr
          url: https://dapr.io
        - name: Dapr Kubernetes Operator
          url: https://github.com/dapr/kubernetes-operator
        maintainers:
        - email: lburgazzoli@gmail.com
          name: Luca Burgazzoli
        - email: salaboy@gmail.com
          name: Mauricio Salatino
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: dapr.io
          url: https://dapr.io
        relatedImages:
        - docker.io/daprio/kubernetes-operator@sha256:da599a273943bf28812314a11df1052411237dd1887edf0fe6104c5857f439b0
        version: 0.0.12
      entries:
      - name: dapr-kubernetes-operator.v0.0.12
        version: 0.0.12
      - name: dapr-kubernetes-operator.v0.0.11
        version: 0.0.11
      - name: dapr-kubernetes-operator.v0.0.10
        version: 0.0.10
      - name: dapr-kubernetes-operator.v0.0.9
        version: 0.0.9
      - name: dapr-kubernetes-operator.v0.0.8
        version: 0.0.8
      - name: dapr-kubernetes-operator.v0.0.6
        version: 0.0.6
      - name: dapr-kubernetes-operator.v0.0.5
        version: 0.0.5
      - name: dapr-kubernetes-operator.v0.0.4
        version: 0.0.4
      - name: dapr-kubernetes-operator.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: dapr-kubernetes-operator
    provider:
      name: dapr.io
      url: https://dapr.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: gitwebhook-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitwebhook-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitWebhook",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "gitwebhook-operator",
                    "app.kubernetes.io/instance": "gitwebhook-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gitwebhook",
                    "app.kubernetes.io/part-of": "gitwebhook-operator"
                  },
                  "name": "gitwebhook-sample"
                },
                "spec": null
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/redhat-cop/gitwebhook-operator@sha256:2810d75eebb24069dbbedf6a82285136c6f16b4cc58b59f1b0fbdc38560cf8a8
          createdAt: "2024-05-10T18:17:13Z"
          description: An operator to create git webhooks declaratively
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: gitwebhook-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/gitwebhook-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitWebhook is the Schema for the gitwebhooks API
            displayName: Git Webhook
            kind: GitWebhook
            name: gitwebhooks.redhatcop.redhat.io
            version: v1alpha1
        description: "This operator automates the setting up of git webhooks. Its
          main purpose is to stream line the configuration of operators such as Tekton
          or ArgoCD which rely on git webhooks for some of their functionalities.\n\n##
          The GitWebhook CRD\n \nA CRD is provided by this operator to create a git
          web hook, here is an example for github\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GitWebhook\nmetadata:\n  name: gitwebhook-github\nspec:\n  gitHub:\n    gitServerCredentials:\n
          \     name: github-pat\n  repositoryOwner: ${repo_owner}\n  ownerType: user\n
          \ repositoryName: ${repo_name}\n  webhookURL: https://hellowebhook.com\n
          \ insecureSSL: false\n  webhookSecret:\n    name: webhook-secret\n  events:\n
          \   - push  \n  contentType: json\n  active: true  \n```    \n\nhere is
          an explanation of each field:\n\n- `gihub` specifies how to connect to the
          git api server. It also requires a local reference to a secret (in the same
          namespace) containing a key `token` with a valid github token to be used
          to authenticate. A similar `gitLab` section exists when connecting to gitlab.
          Only one of `gitLab` or `gitHub` can be defined. \n- `repositoryOwner` and
          `repositoryName` identify the repository for which we want to receive events.\n-
          `ownerType` can have two values: `user` and `organization` and identifies
          the kind of owner.\n- `webhookURL` is the URL for to be called.\n - `insecureSSL`
          defines whether the target URL certificate should be validated (default
          `false`). \n- `webhookSecret` defines a local reference to a secret containing
          the `secret` key. The value is a shared secret between the webhook caller
          and the received for farther validation or identification of the caller.\n-
          `events` is the list of the repo-level events that the webhook should generate.
          The list of valid events for github can be found [here](https://docs.github.com/en/developers/webhooks-and-events/webhooks/webhook-events-and-payloads).
          The list of valid events for gitlab can be found [here](https://docs.gitlab.com/ee/user/project/integrations/webhook_events.html).\n
          - `contentType` defines the format of the webhook payload (default `json`)
          (github).\n - `active` whether the webhook should be turned on (default
          `true`) (github).\n - `pushEventBranchFilter` a regular expression to filter
          from which branches push events should be generated (gitlab only).\n\n##
          Security Considerations\n\nThis operator does not own credentials for the
          git server, but instead always allocate a new connection based on the credentials
          referenced in the CR and every reconcile cycle. As a result there is no
          risk of security escalation or credential leaking between tenants of a cluster
          using this operator. On the other hand it is the responsibility of the namespace
          owners or the platform owner to ensure that valid git credentials are always
          available in the namespace where the GitWebhook CRs need to defined.\n\n##
          Current support\n\nCurrently this operator support creating repo-level webhooks
          for github and gitlab. Potentially this operator could be extended to support
          org-level webhook or other git systems. Contributions are welcome.  "
        displayName: GitWebhook operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - webhook
        - github
        - gitlab
        links:
        - name: GitWbehook Operator
          url: https://github.com/redhat-cop/gitwebhook-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/gitwebhook-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/gitwebhook-operator@sha256:2810d75eebb24069dbbedf6a82285136c6f16b4cc58b59f1b0fbdc38560cf8a8
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        version: 0.1.2
      entries:
      - name: gitwebhook-operator.v0.1.2
        version: 0.1.2
      - name: gitwebhook-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: gitwebhook-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ec2-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ec2-controller.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "DHCPOptions",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "ElasticIPAddress",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "Instance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "InternetGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "NATGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "RouteTable",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "SecurityGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "Subnet",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "TransitGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "VPC",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "VPCEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ec2-controller:1.7.0
          createdAt: "2025-09-27T00:04:55Z"
          description: AWS EC2 controller is a service controller for managing EC2
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CapacityReservation represents the state of an AWS ec2 CapacityReservation
              resource.
            displayName: CapacityReservation
            kind: CapacityReservation
            name: capacityreservations.ec2.services.k8s.aws
            version: v1alpha1
          - description: DHCPOptions represents the state of an AWS ec2 DHCPOptions
              resource.
            displayName: DHCPOptions
            kind: DHCPOptions
            name: dhcpoptions.ec2.services.k8s.aws
            version: v1alpha1
          - description: ElasticIPAddress represents the state of an AWS ec2 ElasticIPAddress
              resource.
            displayName: ElasticIPAddress
            kind: ElasticIPAddress
            name: elasticipaddresses.ec2.services.k8s.aws
            version: v1alpha1
          - description: FlowLog represents the state of an AWS ec2 FlowLog resource.
            displayName: FlowLog
            kind: FlowLog
            name: flowlogs.ec2.services.k8s.aws
            version: v1alpha1
          - description: Instance represents the state of an AWS ec2 Instance resource.
            displayName: Instance
            kind: Instance
            name: instances.ec2.services.k8s.aws
            version: v1alpha1
          - description: InternetGateway represents the state of an AWS ec2 InternetGateway
              resource.
            displayName: InternetGateway
            kind: InternetGateway
            name: internetgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: LaunchTemplate represents the state of an AWS ec2 LaunchTemplate
              resource.
            displayName: LaunchTemplate
            kind: LaunchTemplate
            name: launchtemplates.ec2.services.k8s.aws
            version: v1alpha1
          - description: NATGateway represents the state of an AWS ec2 NATGateway
              resource.
            displayName: NATGateway
            kind: NATGateway
            name: natgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: NetworkACL represents the state of an AWS ec2 NetworkACL
              resource.
            displayName: NetworkACL
            kind: NetworkACL
            name: networkacls.ec2.services.k8s.aws
            version: v1alpha1
          - description: RouteTable represents the state of an AWS ec2 RouteTable
              resource.
            displayName: RouteTable
            kind: RouteTable
            name: routetables.ec2.services.k8s.aws
            version: v1alpha1
          - description: SecurityGroup represents the state of an AWS ec2 SecurityGroup
              resource.
            displayName: SecurityGroup
            kind: SecurityGroup
            name: securitygroups.ec2.services.k8s.aws
            version: v1alpha1
          - description: Subnet represents the state of an AWS ec2 Subnet resource.
            displayName: Subnet
            kind: Subnet
            name: subnets.ec2.services.k8s.aws
            version: v1alpha1
          - description: TransitGateway represents the state of an AWS ec2 TransitGateway
              resource.
            displayName: TransitGateway
            kind: TransitGateway
            name: transitgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: TransitGatewayVPCAttachment represents the state of an AWS
              ec2 TransitGatewayVPCAttachment resource.
            displayName: TransitGatewayVPCAttachment
            kind: TransitGatewayVPCAttachment
            name: transitgatewayvpcattachments.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCEndpoint represents the state of an AWS ec2 VPCEndpoint
              resource.
            displayName: VPCEndpoint
            kind: VPCEndpoint
            name: vpcendpoints.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCEndpointServiceConfiguration represents the state of an
              AWS ec2 VPCEndpointServiceConfiguration resource.
            displayName: VPCEndpointServiceConfiguration
            kind: VPCEndpointServiceConfiguration
            name: vpcendpointserviceconfigurations.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCPeeringConnection represents the state of an AWS ec2 VPCPeeringConnection
              resource.
            displayName: VPCPeeringConnection
            kind: VPCPeeringConnection
            name: vpcpeeringconnections.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPC represents the state of an AWS ec2 VPC resource.
            displayName: VPC
            kind: VPC
            name: vpcs.ec2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Compute Cloud (EC2) resources in AWS from within your Kubernetes cluster.

          **About Amazon EC2**

          Amazon Elastic Compute Cloud (Amazon EC2) provides scalable computing capacity in the Amazon Web Services (AWS) Cloud. Using Amazon EC2 eliminates your need to invest in hardware up front, so you can develop and deploy applications faster. You can use Amazon EC2 to launch as many or as few virtual servers as you need, configure security and networking, and manage storage. Amazon EC2 enables you to scale up or down to handle changes in requirements or spikes in popularity, reducing your need to forecast traffic.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EC2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ec2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EC2 Developer Resources
          url: https://aws.amazon.com/ec2/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ec2 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ec2-controller:1.7.0
        version: 1.7.0
      entries:
      - name: ack-ec2-controller.v1.7.0
        version: 1.7.0
      - name: ack-ec2-controller.v1.6.2
        version: 1.6.2
      - name: ack-ec2-controller.v1.6.1
        version: 1.6.1
      - name: ack-ec2-controller.v1.6.0
        version: 1.6.0
      - name: ack-ec2-controller.v1.4.11
        version: 1.4.11
      - name: ack-ec2-controller.v1.4.10
        version: 1.4.10
      - name: ack-ec2-controller.v1.4.9
        version: 1.4.9
      - name: ack-ec2-controller.v1.4.8
        version: 1.4.8
      - name: ack-ec2-controller.v1.4.7
        version: 1.4.7
      - name: ack-ec2-controller.v1.4.6
        version: 1.4.6
      - name: ack-ec2-controller.v1.4.5
        version: 1.4.5
      - name: ack-ec2-controller.v1.4.4
        version: 1.4.4
      - name: ack-ec2-controller.v1.4.2
        version: 1.4.2
      - name: ack-ec2-controller.v1.4.1
        version: 1.4.1
      - name: ack-ec2-controller.v1.4.0
        version: 1.4.0
      - name: ack-ec2-controller.v1.3.7
        version: 1.3.7
      - name: ack-ec2-controller.v1.3.6
        version: 1.3.6
      - name: ack-ec2-controller.v1.3.5
        version: 1.3.5
      - name: ack-ec2-controller.v1.3.4
        version: 1.3.4
      - name: ack-ec2-controller.v1.3.3
        version: 1.3.3
      - name: ack-ec2-controller.v1.3.1
        version: 1.3.1
      - name: ack-ec2-controller.v1.3.0
        version: 1.3.0
      - name: ack-ec2-controller.v1.2.28
        version: 1.2.28
      - name: ack-ec2-controller.v1.2.27
        version: 1.2.27
      - name: ack-ec2-controller.v1.2.26
        version: 1.2.26
      - name: ack-ec2-controller.v1.2.25
        version: 1.2.25
      - name: ack-ec2-controller.v1.2.24
        version: 1.2.24
      - name: ack-ec2-controller.v1.2.23
        version: 1.2.23
      - name: ack-ec2-controller.v1.2.22
        version: 1.2.22
      - name: ack-ec2-controller.v1.2.21
        version: 1.2.21
      - name: ack-ec2-controller.v1.2.20
        version: 1.2.20
      - name: ack-ec2-controller.v1.2.17
        version: 1.2.17
      - name: ack-ec2-controller.v1.2.16
        version: 1.2.16
      - name: ack-ec2-controller.v1.2.15
        version: 1.2.15
      - name: ack-ec2-controller.v1.2.14
        version: 1.2.14
      - name: ack-ec2-controller.v1.2.13
        version: 1.2.13
      - name: ack-ec2-controller.v1.2.12
        version: 1.2.12
      - name: ack-ec2-controller.v1.2.11
        version: 1.2.11
      - name: ack-ec2-controller.v1.2.10
        version: 1.2.10
      - name: ack-ec2-controller.v1.2.9
        version: 1.2.9
      - name: ack-ec2-controller.v1.2.8
        version: 1.2.8
      - name: ack-ec2-controller.v1.2.7
        version: 1.2.7
      - name: ack-ec2-controller.v1.2.6
        version: 1.2.6
      - name: ack-ec2-controller.v1.2.4
        version: 1.2.4
      - name: ack-ec2-controller.v1.2.1
        version: 1.2.1
      - name: ack-ec2-controller.v1.2.0
        version: 1.2.0
      - name: ack-ec2-controller.v1.1.2
        version: 1.1.2
      - name: ack-ec2-controller.v1.1.1
        version: 1.1.1
      - name: ack-ec2-controller.v1.1.0
        version: 1.1.0
      - name: ack-ec2-controller.v1.0.7
        version: 1.0.7
      - name: ack-ec2-controller.v1.0.6
        version: 1.0.6
      - name: ack-ec2-controller.v1.0.4
        version: 1.0.4
      - name: ack-ec2-controller.v1.0.3
        version: 1.0.3
      - name: ack-ec2-controller.v1.0.2
        version: 1.0.2
      - name: ack-ec2-controller.v1.0.1
        version: 1.0.1
      - name: ack-ec2-controller.v1.0.0
        version: 1.0.0
      - name: ack-ec2-controller.v0.1.1
        version: 0.1.1
      - name: ack-ec2-controller.v0.1.0
        version: 0.1.0
      - name: ack-ec2-controller.v0.0.21
        version: 0.0.21
      - name: ack-ec2-controller.v0.0.20
        version: 0.0.20
      - name: ack-ec2-controller.v0.0.19
        version: 0.0.19
      - name: ack-ec2-controller.v0.0.18
        version: 0.0.18
      - name: ack-ec2-controller.v0.0.17
        version: 0.0.17
      - name: ack-ec2-controller.v0.0.16
        version: 0.0.16
      - name: ack-ec2-controller.v0.0.15
        version: 0.0.15
      - name: ack-ec2-controller.v0.0.14
        version: 0.0.14
      - name: ack-ec2-controller.v0.0.13
        version: 0.0.13
      - name: ack-ec2-controller.v0.0.12
        version: 0.0.12
      - name: ack-ec2-controller.v0.0.11
        version: 0.0.11
      - name: ack-ec2-controller.v0.0.10
        version: 0.0.10
      - name: ack-ec2-controller.v0.0.9
        version: 0.0.9
      - name: ack-ec2-controller.v0.0.8
        version: 0.0.8
      - name: ack-ec2-controller.v0.0.7
        version: 0.0.7
      - name: ack-ec2-controller.v0.0.6
        version: 0.0.6
      - name: ack-ec2-controller.v0.0.5
        version: 0.0.5
      - name: ack-ec2-controller.v0.0.4
        version: 0.0.4
      - name: ack-ec2-controller.v0.0.3
        version: 0.0.3
      - name: ack-ec2-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-ec2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: global-load-balancer-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: global-load-balancer-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSRecord",
                "metadata": {
                  "name": "hello-global-record"
                },
                "spec": {
                  "endpoints": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    }
                  ],
                  "globalZoneRef": {
                    "name": "external-dns-zone"
                  },
                  "loadBalancingPolicy": "Multivalue",
                  "name": "hello.my.base.domain",
                  "ttl": 60
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSRecord",
                "metadata": {
                  "name": "route53-geoproximity-global-record-with-healthcheck"
                },
                "spec": {
                  "endpoints": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    }
                  ],
                  "globalZoneRef": {
                    "name": "route53-global-dns-zone"
                  },
                  "healthCheck": {
                    "failureThreshold": 3,
                    "httpGet": {
                      "host": "geoproximity-hc.${global_base_domain}",
                      "path": "/healthz",
                      "port": 80,
                      "scheme": "HTTP"
                    },
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 2
                  },
                  "loadBalancingPolicy": "Geoproximity",
                  "name": "geoproximity-hc.my.base.domain",
                  "ttl": 60
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSZone",
                "metadata": {
                  "name": "external-dns-zone"
                },
                "spec": {
                  "domain": "myglobalzone.mybasedomain.io",
                  "provider": {
                    "externalDNS": {
                      "annotations": {
                        "type": "global"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSZone",
                "metadata": {
                  "name": "route53-global-dns-zone"
                },
                "spec": {
                  "domain": "myglobalzone.mybasedomain.io",
                  "provider": {
                    "route53": {
                      "credentialsSecretRef": {
                        "name": "route53-global-zone-credentials",
                        "namespace": "mynamespace"
                      },
                      "zoneID": "Z083892239ROKNEP70BDY"
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalRouteDiscovery",
                "metadata": {
                  "name": "external-dns-globalroutediscovery"
                },
                "spec": {
                  "clusters": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3"
                    }
                  ],
                  "defaultLoadBalancingPolicy": "Multivalue",
                  "defaultTTL": 30,
                  "globalZoneRef": {
                    "name": "external-dns-zone"
                  },
                  "routeSelector": {
                    "matchLabels": {
                      "route-type": "global"
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalRouteDiscovery",
                "metadata": {
                  "name": "route53-globalroutediscovery"
                },
                "spec": {
                  "clusters": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3"
                    }
                  ],
                  "defaultLoadBalancingPolicy": "Multivalue",
                  "defaultTTL": 30,
                  "globalZoneRef": {
                    "name": "route53-global-dns-zone"
                  },
                  "routeSelector": {
                    "matchLabels": {
                      "route-type": "global"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Cloud Provider
          certified: "false"
          containerImage: quay.io/redhat-cop/global-load-balancer-operator@sha256:c0b6ac9b03571a74a56b28302d720e8aa4442233f7867d05ed365e022f7cb8f5
          createdAt: "2021-12-02T17:33:23Z"
          description: This operator creates automation around a DNS to operate as
            a global load balancer for a set of OpenShift clusters.
          operatorframework.io/suggested-namespace: global-load-balancer-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/global-load-balancer-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DNSEndpoint
            name: dnsendpoints.externaldns.k8s.io
            version: v1alpha1
          - description: GlobalDNSRecord is the Schema for the globaldnsrecords API
            displayName: Global DNSRecord
            kind: GlobalDNSRecord
            name: globaldnsrecords.redhatcop.redhat.io
            version: v1alpha1
          - description: GlobalDNSZone is the Schema for the globaldnszones API
            displayName: Global DNSZone
            kind: GlobalDNSZone
            name: globaldnszones.redhatcop.redhat.io
            version: v1alpha1
          - description: GlobalRouteDiscovery is the Schema for the globalroutediscoveries
              API
            displayName: Global Route Discovery
            kind: GlobalRouteDiscovery
            name: globalroutediscoveries.redhatcop.redhat.io
            version: v1alpha1
        description: "The global-load-balancer-operator implements automation to program
          a DNS service to act as global load balancer for applications deployed to
          multiple OpenShift clusters.\nThis operator is designed to be deployed to
          a control cluster which will watch the load balanced clusters (controlled
          clusters).\nThere are two main concepts (APIs) provided by this operator:\n\n1.
          GlobalDNSZone\n2. GlobalDNSRecord\n\n## GlobalDNSZone\n\nThe `GlobalDNSZone`
          CR allows you to configure a zone which will contain global load balanced
          records and the provider used to populate it.\nHere is an example of GlobalDNSZone:\n\n```yaml\napiVersion:
          redhatcop.redhat.io/v1alpha1\nkind: GlobalDNSZone\nmetadata:\n  name: external-dns-zone\nspec:\n
          \ # Add fields here\n  domain: global.myzone.io\n  provider:\n    externalDNS:\n
          \     annotations:\n        type: global\n```\n\nHere is a table summarizing
          the supported providers and their capabilities:\n\n| Provider  | Supports
          Health Checks  | Supports Multivalue LB | Supports Latency LB  | Supports
          GeoProximity LB  |\n|:--:|:--:|:---:|:---:|:---:|\n| External-dns  | no
          \ | yes | no  | no  |\n| Route53  | yes | yes(*)  | yes(*)  | yes(*)  |\n\n(*)
          only if all controlled clusters run on AWS.\n\n## GlobalDNSRecord\n\nThe
          `GlobalDNSRecord` CR allows you to specify the intention to create a global
          dns record. Here is an example\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GlobalDNSRecord\nmetadata:\n  name: hello-global-record\nspec:\n  name:
          hello.global.myzone.io\n  endpoints:\n  - clusterName: cluster1\n    clusterCredentialRef:\n
          \     name: kubeconfig\n      namespace: cluster1\n    loadBalancerServiceRef:\n
          \     name: router-default\n      namespace: openshift-ingress\n  - clusterName:
          cluster2\n    clusterCredentialRef:\n      name: kubeconfig\n      namespace:
          cluster2\n    loadBalancerServiceRef:\n      name: router-default\n      namespace:
          openshift-ingress\n  - clusterName: cluster3\n    clusterCredentialRef:\n
          \     name: kubeconfig\n      namespace: cluster3\n    loadBalancerServiceRef:\n
          \     name: router-default\n      namespace: openshift-ingress\n  ttl: 60\n
          \ loadBalancingPolicy: Multivalue\n  globalZoneRef:\n    name: external-dns-zone\n
          \ healthCheck:\n    failureThreshold: 3\n    httpGet:\n      host: hello.global.myzone.io\n
          \     httpHeaders:\n        - name: ciao\n          value: hello\n      path:
          /readyz\n      port: 80\n      scheme: HTTP\n    periodSeconds: 10\n    successThreshold:
          1\n    timeoutSeconds: 1     \n```\n\nThe list of endpoints identifies the
          set of `LoadBalancer` type of services to watch on the remote clusters (at
          the moment only LoadBalancer services are supported). These are the `LoadBalancer`
          services created by the [ingress controller](https://docs.openshift.com/container-platform/4.5/networking/ingress-operator.html)
          on which the routers rely. Each entry of this list requires a reference
          to the loadbalancer service on the remote cluster and a reference to a local
          secret containing the credential to connect to the remote cluster.\n\nThis
          secret must contain one entry called `kubeconfig` with a kubeconfig file
          with a default context pointing to the remote cluster. The account used
          by the kubeconfig file (presumably a service account) will require at a
          minimum cluster-level permissions as described in this [cluster role](./docs/scripts/remote-account-cluster-role.yaml).\n\nThe
          `globalZoneRef` field refers to a local (same namespace) `GlobalZone` CR.
          The DNS record represented by this `GlobalDNSRecord`, will be created in
          the referenced zone.\n\n`ttl` is the TTL of the crated DNS record.\n\n`loadBalancingPolicy`
          is the load balancing policy for this global record. It must match one of
          the policy supported by the provider of the referenced `GlobalZone`.\n\nFinally,
          `healthcheck` represent a [probe](https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.18/#probe-v1-core)
          to be used to test the health of a record. This field will be ignored if
          the provider does not support health checks.\n\n## External DNS Provider\n\nThe
          [external-dns](https://github.com/kubernetes-sigs/external-dns) provider
          delegates to external-dns the creation of the actual DNS records by creating
          a `DNSEndpoint` object.\nThe `DNSEndpoint` object will be created in the
          same namespace as the `GlobalDNSRecord` and will be owned by it.\nThe `DNSEdnpoint`
          object will have the same labels as the `GlobalDNSRecord` and the annotations
          specified in the GlobalDNSZone configuration.\nExternal-dns should be configured
          to watch for DNSEnpoints at the cluster level and to point to the desired
          provider.\nDetails on configuration can be found at the external-dns git
          repository.\nThe External-dns should be used as a fall back option when
          other options are not available as it does not support health checks and
          advanced load balancing policies.\n\n## AWS Route53 provider\n\nAWS Route53
          provider uses the Route53 service as a global loadbalancer and offers advanced
          routing capabilities via [route53 traffic policies](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/traffic-flow.html)
          (note that traffic policies will trigger an expense).\nThe following routing
          polices are currently supported:\n\n1. [Multivalue](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-multivalue)\n2.
          [Geoproximity](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-geoproximity)\n3.
          [Latency](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-latency)\n\nAWS
          Route53 provider at the moment requires that all the controlled clusters
          run in AWS.\n\nIf health checks are defined, a route53 health check originating
          from any reason (you have to ensure connectivity) will be created for each
          of the endpoint. Because the endpoint represent s a shared ELB (shared with
          other apps, that is) and the health check is app specific, we cannot use
          the ELB health check, so the route53 endpoint is created with one of the
          two IP exposed by the ELB. This is suboptimal, but it works in most situations.\n\n##
          Global Route Auto Discovery\n\nThe aboev examples showed how to create global
          DNS records. This can be good in some situations, but most of the times
          in an openshift deployment global DNS records will point to routes that
          are intended to be global.\nThe global-load-balancer operator can auto-discover
          these routes and automatically create the corresponding `GloablDNSRecord`.
          The `GlobalRouteDiscovery` CRD is used to configure the discovery process,
          here is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GlobalRouteDiscovery\nmetadata:\n  name: route53-globalroutediscovery\nspec:\n
          \ clusters:\n  - clusterName: cluster1\n    clusterCredentialRef:\n      name:
          ${cluster1_secret_name}\n      namespace: cluster1\n...\n  routeSelector:\n
          \   matchLabels:\n      route-type: global\n  defaultLoadBalancingPolicy:
          Multivalue\n  defaultTTL: 30 \n  globalZoneRef:\n    name: route53-global-dns-zone\n```\n\nThis
          global discovery route will discover routes in the provided list of cluster.
          Only the routes that match the route selector will be considered global.
          \nThe default load balancing policy and default TTL can be expressed in
          the `GlobalRouteDiscovery` CR. However with the following annotations, it's
          possible to configure route-specific values:\n\n* `global-load-balancer-operator.redhat-cop.io/load-balancing-policy`
          to set the load balancing policy\n* `global-load-balancer-operator.redhat-cop.io/ttl`
          to set the TTL\n\nThe globalZoneRef refers to the global zone to be used
          for the created `GlobalDNSRecords`.\n\nHealth checks will also be automatically
          discovered. If the pods behind the route expose a readiness check of httpGet
          kind, that configuration will be used to create the GlobalDNSRecord health
          check.\nWhen more than one container is present in the pod, by default the
          first container will be examined for health check. This behavior can be
          overridden with the this annotation on the route: `global-load-balancer-operator.redhat-cop.io/container-probe`
          where the value will container the name of the container with teh correct
          readiness probe.\n\nIf routes with the same namespace and name exists in
          multiple cluster, the following conditions must be met:\n\n* all host names
          must be the same\n* all load balancing policy must be the same\n* all TTLs
          must be the same\n* all discovered readiness checks must be the same\n\n##
          Examples\n\nThese examples are intended to help you setting up working configuration
          with each of the providers\n\n### Cluster Set up\n\nTwo approaches for cluster
          setup are provided\n\n1. [One cluster, three ingress-gateways.](https://github.com/redhat-cop/global-load-balancer-operator/docs/one-cluster-three-ingress-gateways.md)
          This approach is intended for development purposes and has the objective
          to keep resource consumption at the minimum.\n2. [Control cluster and three
          controlled clusters in different regions.](https://github.com/redhat-cop/global-load-balancer-operator/docs/three-clusters.md).
          This approach represents a more realistic set-up albeit it consumes more
          resources.\n\nYou can also set up the cluster on your own, at the end the
          following conditions must be met:\n\nThree namespace `cluster1` `cluster2`
          `cluster3` are created.\nthe following environment variables are initialized
          for each cluster:\n\n1. <cluster>_secret_name. Pointing to a secret in each
          of the cluster namespaces containing a valid kubeconfig fot that cluster\n2.
          <cluster>_service_name.  Pointing to the name of the loadbalancer service
          to be used for that cluster.\n3. <cluster>_service_namespace. Pointing to
          the namespace of the loadbalancer service to be used for that cluster.\n\nHere
          are examples for the supported provider:\n\n1. [Setting up external-dns
          as provider](https://github.com/redhat-cop/global-load-balancer-operator/docs/external-dns-provider.md)\n2.
          [Setting up route53 as a provider](https://github.com/redhat-cop/global-load-balancer-operator/docs/aws-route53-provider.md)\n"
        displayName: Global Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Networking
        - Cloud Provider
        - Global Load Balancing
        - GLB
        - GTM
        - route53
        - DNS
        links:
        - name: GigHub Repository
          url: https://github.com/redhat-cop/global-load-balancer-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/global-load-balancer-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/global-load-balancer-operator@sha256:c0b6ac9b03571a74a56b28302d720e8aa4442233f7867d05ed365e022f7cb8f5
        version: 1.4.2
      entries:
      - name: global-load-balancer-operator.v1.4.2
        version: 1.4.2
      - name: global-load-balancer-operator.v1.4.1
        version: 1.4.1
      - name: global-load-balancer-operator.v1.4.0
        version: 1.4.0
      - name: global-load-balancer-operator.v1.3.0
        version: 1.3.0
      - name: global-load-balancer-operator.v1.2.0
        version: 1.2.0
      - name: global-load-balancer-operator.v1.1.0
        version: 1.1.0
      - name: global-load-balancer-operator.v1.0.7
        version: 1.0.7
      - name: global-load-balancer-operator.v1.0.6
        version: 1.0.6
      - name: global-load-balancer-operator.v1.0.5
        version: 1.0.5
      - name: global-load-balancer-operator.v1.0.4
        version: 1.0.4
      - name: global-load-balancer-operator.v1.0.3
        version: 1.0.3
      - name: global-load-balancer-operator.v1.0.2
        version: 1.0.2
      - name: global-load-balancer-operator.v1.0.1
        version: 1.0.1
      - name: global-load-balancer-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: global-load-balancer-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Netscaler
      provider-url: https://www.netscaler.com
    name: netscaler-adm-agent-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netscaler-adm-agent-operator.v141.37.44
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "AdmAgent",
                "metadata": {
                  "name": "admagent"
                },
                "spec": {
                  "accessSecret": "",
                  "admURL": "adm.cloud.com",
                  "affinity": {},
                  "clusterPrefix": "",
                  "core": {
                    "livenessProbe": {
                      "failureThreshold": 5,
                      "initialDelaySeconds": 600,
                      "periodSeconds": 60,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "readinessProbe": {
                      "failureThreshold": 30,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 30,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "100m",
                        "memory": "2000Mi"
                      }
                    }
                  },
                  "fullnameOverride": "",
                  "helper": {
                    "image": "{{ .Values.helper.imageRegistry }}/{{ .Values.helper.imageRepository }}:{{ .Values.helper.imageTag }}",
                    "imageRegistry": "quay.io",
                    "imageRepository": "citrix/adm-agent-helper",
                    "imageTag": "2.0.1",
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "200Mi"
                      }
                    }
                  },
                  "hpa": {
                    "maxReplicas": 10,
                    "minReplicas": 1,
                    "required": false,
                    "targetCPUUtilization": "1000m",
                    "targetMemoryUtilization": "4000Mi"
                  },
                  "image": "{{ .Values.imageRegistry }}/{{ .Values.imageRepository }}:{{ .Values.imageTag }}",
                  "imagePullSecrets": [],
                  "imageRegistry": "quay.io",
                  "imageRepository": "citrix/adm-agent",
                  "imageTag": "14.1-37.44",
                  "kad": {
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "50m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "lic": {
                    "livenessProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 600,
                      "periodSeconds": 20,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "100m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "loginSecret": "",
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "redis": {
                    "livenessProbe": {
                      "failureThreshold": 5,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "readinessProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 20,
                      "periodSeconds": 10,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true
                  },
                  "sharding": {
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-adm-agent-operator@sha256:1c462c2d6161877f87d5eed4888f8d94c733501e67a94757ab0873d983fb3a4a
          createdAt: "2024-11-18T06:40:58Z"
          description: This is an operator to install ADM Agent
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler ADM Agent Operator
            displayName: NetScaler ADM Agent Operator
            kind: AdmAgent
            name: admagents.netscaler.com
            version: v1
        description: Netscaler ADM Agent Operator
        displayName: Netscaler ADM Agent Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        - ADC
        - Kubernetes
        - OpenShift
        - Citrix
        - Controller
        - Observability
        - Metrices
        - adm
        - agent
        - monitoring
        - proxy
        links:
        - name: Netscaler Adm Agent Operator
          url: https://github.com/netscaler/netscaler-helm-charts/blob/master/adm-agent/README.md
        maintainers:
        - email: pavan.belani@cloud.com
          name: Pavan Belani
        - email: swapnil.khanderaokakade@cloud.com
          name: Swapnil Khanderao
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Netscaler
          url: https://www.netscaler.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6b8c101124579dd46086ec83715618dd9a6aebe66176d676242e6c0b0900a53c
        - registry.connect.redhat.com/citrix/netscaler-adm-agent-operator@sha256:1c462c2d6161877f87d5eed4888f8d94c733501e67a94757ab0873d983fb3a4a
        - quay.io/citrix/adm-agent@sha256:45cab90524687c76a6d53f014b55063fa25c8aaeec0261f0f4f0efa6bff42968
        - quay.io/citrix/adm-agent-helper@sha256:f567a0637e8024f9d1674977f6c013741cc0fd12cf78703f8dbe01126561c3c4
        version: 141.37.44
      entries:
      - name: netscaler-adm-agent-operator.v141.37.44
        version: 141.37.44
      - name: netscaler-adm-agent-operator.v141.31.38
        version: 141.31.38
      - name: netscaler-adm-agent-operator.v141.25.33
        version: 141.25.33
      - name: netscaler-adm-agent-operator.v141.23.31
        version: 141.23.31
      - name: netscaler-adm-agent-operator.v141.21.41
        version: 141.21.41
      - name: netscaler-adm-agent-operator.v141.16.39
        version: 141.16.39
      - name: netscaler-adm-agent-operator.v141.13.23
        version: 141.13.23
      name: stable
    defaultChannel: stable
    packageName: netscaler-adm-agent-operator
    provider:
      name: Netscaler
      url: https://www.netscaler.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: pg4k-pgd
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pg4k-pgd.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgd.k8s.enterprisedb.io/v1beta1",
                "kind": "PGDGroup",
                "metadata": {
                  "name": "region-a"
                },
                "spec": {
                  "backup": null,
                  "cnp": {
                    "storage": {
                      "size": "1Gi"
                    },
                    "walStorage": {
                      "size": "1Gi"
                    }
                  },
                  "connectivity": {
                    "dns": {
                      "additional": [
                        {
                          "domain": "alternate.domain"
                        },
                        {
                          "domain": "my.domain",
                          "hostSuffix": "-dc1"
                        }
                      ]
                    },
                    "tls": {
                      "clientCert": {
                        "caCertSecret": "client-ca-key-pair",
                        "certManager": {
                          "spec": {
                            "issuerRef": {
                              "group": "cert-manager.io",
                              "kind": "Issuer",
                              "name": "client-ca-issuer"
                            }
                          }
                        }
                      },
                      "mode": "verify-ca",
                      "serverCert": {
                        "caCertSecret": "server-ca-key-pair",
                        "certManager": {
                          "spec": {
                            "issuerRef": {
                              "group": "cert-manager.io",
                              "kind": "Issuer",
                              "name": "server-ca-issuer"
                            }
                          }
                        }
                      }
                    }
                  },
                  "instances": 2,
                  "pgd": {
                    "parentGroup": {
                      "create": true,
                      "name": "world"
                    }
                  },
                  "proxyInstances": 2,
                  "witness": {
                    "storage": {
                      "size": "1Gi"
                    },
                    "walStorage": {
                      "size": "1Gi"
                    }
                  },
                  "witnessInstances": 1
                }
              },
              {
                "apiVersion": "pgd.k8s.enterprisedb.io/v1beta1",
                "kind": "PGDGroupCleanup",
                "metadata": {
                  "name": "region-b-cleanup"
                },
                "spec": {
                  "executor": "region-a",
                  "target": "region-b"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: docker.enterprisedb.com/k8s_enterprise_pgd/pg4k-pgd-openshift@sha256:5127d913011619d61fd06b27be3c709df2236a4b5c0254fa58632698d58b50fb
          createdAt: "2025-09-26T01:49:29Z"
          description: EDB Postgres Distributed for Kubernetes is an operator designed
            to manage EDB Postgres Distributed (PGD) v5 workloads on Kubernetes both
            in a single Kubernetes cluster or across multiple ones.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterImageCatalog
            name: clusterimagecatalogs.pgd.k8s.enterprisedb.io
            version: v1beta1
          - kind: ImageCatalog
            name: imagecatalogs.pgd.k8s.enterprisedb.io
            version: v1beta1
          - description: Object to clean up a PGD Group
            displayName: PGD Group Cleanup
            kind: PGDGroupCleanup
            name: pgdgroupcleanups.pgd.k8s.enterprisedb.io
            version: v1beta1
          - description: A group of Postgres Clusters
            displayName: PGD Group
            kind: PGDGroup
            name: pgdgroups.pgd.k8s.enterprisedb.io
            version: v1beta1
        description: |
          EDB Postgres Distributed for Kubernetes is an operator designed to manage EDB Postgres Distributed (PGD) v5 workloads on Kubernetes both in a single Kubernetes cluster or across multiple ones.

          EDB Postgres Distributed is an extension for PostgreSQL that provides multi-master replication and data distribution with advanced conflict management, data-loss protection, and throughput up to 5X faster than native logical replication. It enables distributed PostgreSQL clusters with high availability up to five 9s.

          This operator requires access to the EDB private repositories.
        displayName: EDB Postgres Distributed for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - distributed
        - postgres
        - database
        - sql
        - cloud
        - native
        - edb
        - enterprisedb
        - cnp
        - pg4k
        - pgd
        links:
        - name: 'Documentation: EDB Postgres Distributed v5'
          url: https://www.enterprisedb.com/docs/pgd/latest/
        - name: 'Documentation: EDB Postgres Distributed for Kubernetes'
          url: https://www.enterprisedb.com/docs/postgres_distributed_for_kubernetes/latest/
        maintainers:
        - email: sales@enterprisedb.com
          name: Sales
        - email: support@enterprisedb.com
          name: Support
        maturity: alpha
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - docker.enterprisedb.com/k8s_enterprise_pgd/pg4k-pgd-openshift@sha256:5127d913011619d61fd06b27be3c709df2236a4b5c0254fa58632698d58b50fb
        version: 1.1.3
      entries:
      - name: pg4k-pgd.v1.1.3
        version: 1.1.3
      - name: pg4k-pgd.v1.1.2
        version: 1.1.2
      name: stable
    defaultChannel: stable
    packageName: pg4k-pgd
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Redis
      provider-url: ""
    name: redis-enterprise-operator-cert-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-enterprise-operator.v6.2.10-45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:ffdf10a1b08115fd4de06f1cd751d36644a9bc9a8c11034cd432e9b66a5d2f73
        - registry.connect.redhat.com/redislabs/services-manager@sha256:8b46f2aacfb3e1b243c0b64975accc55fd80b4de9c9ac4769c3c67a5676addc5
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
        version: 6.2.10-45
      entries:
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: 6.2.10
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.22.0
    - currentCSV: redis-enterprise-operator.v7.22.0-11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:088e4b57c6e1ca1680b359078e5eb1bba67343cdbd9a539af26a0ea23a830593
        - registry.connect.redhat.com/redislabs/services-manager@sha256:3b96cdc65c3e42fd22eb165950be96b907725903ac74fe4f82cfa0073e28e95a
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:d56ba97e95e799210e76615095409143a5dddd5c5a1e1be1d551d7e3f9e4d3ef
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:0cfb07f774db131b5b74e5aa91291579c916c4d4cd1a693d7aaed554f0177a68
        version: 7.22.0-11.3
      entries:
      - name: redis-enterprise-operator.v7.22.0-11.3
        version: 7.22.0-11.3
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.22.0-11.3
    - currentCSV: redis-enterprise-operator.v7.4.6-7.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:473b5661049a6c8830ba113504975726154c7b172376ba3b63d891674d59c78f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5f6f32a63f4dd8be85b9d469f0fedd26d6aa76e32b7735bfa7ab7203cb272cdb
        - registry.connect.redhat.com/redislabs/services-manager@sha256:ba20ab644a473bc742af671171d5313537292d88942b16336d03a6386c061d8a
        version: 7.4.6-7.2
      entries:
      - name: redis-enterprise-operator.v7.4.6-7.2
        version: 7.4.6-7.2
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      name: 7.4.6
    - currentCSV: redis-enterprise-operator.v7.8.2-6.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:36c7a36dabb23a4c003e112085cda2873296bfb22e5d11b470d7c1aba55b7f0f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:c6e1f2109a0dac76b51c25c1cbf6030b4894b21b98147e2ad1db2dd8e7693576
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:acf817c90b5f03bc2796fbc305cf43ce5c0644dcbe1b685752090b3f600a0f7b
        version: 7.8.2-6.8
      entries:
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.2
    - currentCSV: redis-enterprise-operator.v7.8.4-9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:803be04534e59cdc7b8898b3bb336676aad1cb77cf266051e226be82679fcf6a
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:4ba0c414a74268bed9c6757e447b340db7c912afa864cfb096ddf88f11feeaf9
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:66ec124da16cc9d0abd66687e088be0b358e20cfd48f5774d34bf5874e6a4e65
        version: 7.8.4-9.0
      entries:
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.4
    - currentCSV: redis-enterprise-operator.v7.8.6-5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:1fff01daece34aeb7befcb0760d5112c8c7d261f2b86388ece416ac94076a573
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5125e83d3c93b224d1983c599994e43899a0ba0c6f2fc4eed31476f755b5fc0b
        - registry.connect.redhat.com/redislabs/services-manager@sha256:88d483f5e37b5a901d1ae3d965304edc000c172d2e5e307297d5b7bcd77ae73b
        version: 7.8.6-5.1
      entries:
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.6
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: production
    defaultChannel: production
    packageName: redis-enterprise-operator-cert-rhmp
    provider:
      name: Redis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc
      provider-url: ""
    name: bare-metal-event-relay
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bare-metal-event-relay.v4.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "event.redhat-cne.org/v1alpha1",
                "kind": "HardwareEvent",
                "metadata": {
                  "name": "hardwareevent-sample"
                },
                "spec": {
                  "nodeSelector": {
                    "node-role.kubernetes.io/worker": ""
                  },
                  "storageType": "emptyDir",
                  "transportHost": "http://hw-event-publisher-service.openshift-bare-metal-events.svc.cluster.local:9043"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/openshift4/baremetal-hardware-event-proxy-rhel9@sha256:aaaea040164e1fd801e13c74024b2307997cdde954c1a5e35efd90023d2d7376
          createdAt: "2023-02-23"
          description: This software manages the lifecycle of the hw-event-proxy container.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.22.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/hw-event-proxy
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HardwareEvent is the Schema for the hardwareevents API
            displayName: Hardware Event
            kind: HardwareEvent
            name: hardwareevents.event.redhat-cne.org
            version: v1alpha1
        description: |
          ## Funtionality
          Install and manage a hardware event delivery framework.
          ## Installation
          To start using Hardware event proxy, install the operator and then create the following HardwareEvent:
          ```
          apiVersion: "event.redhat-cne.org/v1alpha1"
          kind: "HardwareEvent"
          metadata:
            name: "hardware-event"
          spec:
            nodeSelector:
              node-role.kubernetes.io/worker: ""
            transportHost: "http://hw-event-publisher-service.openshift-bare-metal-events.svc.cluster.local:9043"

          ```
          This operator needs a secret to be created to access Redfish
          Message Registry.
          The secret name and spec the operand expects are under the same namespace
          the operand is deployed on. For example:
          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: redfish-basic-auth
          type: Opaque
          stringData:
            username: <bmc_username>
            password: <bmc_password>
            # BMC host DNS or IP address
            hostaddr: <bmc_host_ip_address>
          ```
        displayName: Bare Metal Event Relay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hardware
        - event
        - cloudevents
        links:
        - name: Hardware Event Proxy Operator
          url: https://github.com/redhat-cne/hw-event-proxy-operator
        maturity: alpha
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:6b7beed3cba9093b3aeb3109f23ec38c20f5568915dff774ad5e9e6bb5db3a58
        - registry.redhat.io/openshift4/bare-metal-event-relay-rhel9-operator@sha256:2219f2f623ee27ef83518d28857995287de0a70cefa193792f28acb352ed8fc0
        - registry.redhat.io/openshift4/baremetal-hardware-event-proxy-rhel9@sha256:aaaea040164e1fd801e13c74024b2307997cdde954c1a5e35efd90023d2d7376
        - registry.redhat.io/openshift4/ose-cloud-event-proxy-rhel9@sha256:c99b448820cf03a33e9237030ac74c1d1ad6a4ce17f1b2840c69b039097c5733
        version: 4.16.1
      entries:
      - name: bare-metal-event-relay.v4.16.1
        version: 4.16.1
      name: "4.16"
    - currentCSV: bare-metal-event-relay.v4.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "event.redhat-cne.org/v1alpha1",
                "kind": "HardwareEvent",
                "metadata": {
                  "name": "hardwareevent-sample"
                },
                "spec": {
                  "nodeSelector": {
                    "node-role.kubernetes.io/worker": ""
                  },
                  "storageType": "emptyDir",
                  "transportHost": "http://hw-event-publisher-service.openshift-bare-metal-events.svc.cluster.local:9043"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/openshift4/baremetal-hardware-event-proxy-rhel9@sha256:aaaea040164e1fd801e13c74024b2307997cdde954c1a5e35efd90023d2d7376
          createdAt: "2023-02-23"
          description: This software manages the lifecycle of the hw-event-proxy container.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.22.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/hw-event-proxy
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HardwareEvent is the Schema for the hardwareevents API
            displayName: Hardware Event
            kind: HardwareEvent
            name: hardwareevents.event.redhat-cne.org
            version: v1alpha1
        description: |
          ## Funtionality
          Install and manage a hardware event delivery framework.
          ## Installation
          To start using Hardware event proxy, install the operator and then create the following HardwareEvent:
          ```
          apiVersion: "event.redhat-cne.org/v1alpha1"
          kind: "HardwareEvent"
          metadata:
            name: "hardware-event"
          spec:
            nodeSelector:
              node-role.kubernetes.io/worker: ""
            transportHost: "http://hw-event-publisher-service.openshift-bare-metal-events.svc.cluster.local:9043"

          ```
          This operator needs a secret to be created to access Redfish
          Message Registry.
          The secret name and spec the operand expects are under the same namespace
          the operand is deployed on. For example:
          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: redfish-basic-auth
          type: Opaque
          stringData:
            username: <bmc_username>
            password: <bmc_password>
            # BMC host DNS or IP address
            hostaddr: <bmc_host_ip_address>
          ```
        displayName: Bare Metal Event Relay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hardware
        - event
        - cloudevents
        links:
        - name: Hardware Event Proxy Operator
          url: https://github.com/redhat-cne/hw-event-proxy-operator
        maturity: alpha
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:6b7beed3cba9093b3aeb3109f23ec38c20f5568915dff774ad5e9e6bb5db3a58
        - registry.redhat.io/openshift4/bare-metal-event-relay-rhel9-operator@sha256:2219f2f623ee27ef83518d28857995287de0a70cefa193792f28acb352ed8fc0
        - registry.redhat.io/openshift4/ose-cloud-event-proxy-rhel9@sha256:c99b448820cf03a33e9237030ac74c1d1ad6a4ce17f1b2840c69b039097c5733
        - registry.redhat.io/openshift4/baremetal-hardware-event-proxy-rhel9@sha256:aaaea040164e1fd801e13c74024b2307997cdde954c1a5e35efd90023d2d7376
        version: 4.16.1
      entries:
      - name: bare-metal-event-relay.v4.16.1
        version: 4.16.1
      name: stable
    defaultChannel: stable
    packageName: bare-metal-event-relay
    provider:
      name: Red Hat Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ocs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocs-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |2-

            [
                {
                    "apiVersion": "ocs.openshift.io/v1",
                    "kind": "StorageCluster",
                    "metadata": {
                        "name": "example-storagecluster",
                        "namespace": "openshift-storage"
                    },
                    "spec": {
                        "manageNodes": false,
                        "monPVCTemplate": {
                            "spec": {
                                "accessModes": [
                                    "ReadWriteOnce"
                                ],
                                "resources": {
                                    "requests": {
                                        "storage": "10Gi"
                                    }
                                },
                                "storageClassName": "gp2-csi"
                            }
                        },
                        "storageDeviceSets": [
                            {
                                "count": 3,
                                "dataPVCTemplate": {
                                    "spec": {
                                        "accessModes": [
                                            "ReadWriteOnce"
                                        ],
                                        "resources": {
                                            "requests": {
                                                "storage": "1Ti"
                                            }
                                        },
                                        "storageClassName": "gp2-csi",
                                        "volumeMode": "Block"
                                    }
                                },
                                "name": "example-deviceset",
                                "placement": {},
                                "portable": true,
                                "resources": {}
                            }
                        ]
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Storage
          containerImage: registry.redhat.io/odf4/ocs-rhel9-operator@sha256:b1a9492bd9b314f86b4e192cf2c83144885ff863d9932e386a41691f50be6193
          description: Red Hat OpenShift Container Storage provides hyperconverged
            storage for applications within an OpenShift cluster.
          external.features.ocs.openshift.io/export-script: |-
            IiIiCkNvcHlyaWdodCAyMDIwIFRoZSBSb29rIEF1dGhvcnMuIEFsbCByaWdodHMgcmVzZXJ2ZWQu
            CgpMaWNlbnNlZCB1bmRlciB0aGUgQXBhY2hlIExpY2Vuc2UsIFZlcnNpb24gMi4wICh0aGUgIkxp
            Y2Vuc2UiKTsKeW91IG1heSBub3QgdXNlIHRoaXMgZmlsZSBleGNlcHQgaW4gY29tcGxpYW5jZSB3
            aXRoIHRoZSBMaWNlbnNlLgpZb3UgbWF5IG9idGFpbiBhIGNvcHkgb2YgdGhlIExpY2Vuc2UgYXQK
            CglodHRwOi8vd3d3LmFwYWNoZS5vcmcvbGljZW5zZXMvTElDRU5TRS0yLjAKClVubGVzcyByZXF1
            aXJlZCBieSBhcHBsaWNhYmxlIGxhdyBvciBhZ3JlZWQgdG8gaW4gd3JpdGluZywgc29mdHdhcmUK
            ZGlzdHJpYnV0ZWQgdW5kZXIgdGhlIExpY2Vuc2UgaXMgZGlzdHJpYnV0ZWQgb24gYW4gIkFTIElT
            IiBCQVNJUywKV0lUSE9VVCBXQVJSQU5USUVTIE9SIENPTkRJVElPTlMgT0YgQU5ZIEtJTkQsIGVp
            dGhlciBleHByZXNzIG9yIGltcGxpZWQuClNlZSB0aGUgTGljZW5zZSBmb3IgdGhlIHNwZWNpZmlj
            IGxhbmd1YWdlIGdvdmVybmluZyBwZXJtaXNzaW9ucyBhbmQKbGltaXRhdGlvbnMgdW5kZXIgdGhl
            IExpY2Vuc2UuCiIiIgoKaW1wb3J0IGVycm5vCmltcG9ydCBzeXMKaW1wb3J0IGpzb24KaW1wb3J0
            IGFyZ3BhcnNlCmltcG9ydCByZQppbXBvcnQgc3VicHJvY2VzcwppbXBvcnQgaG1hYwpmcm9tIGhh
            c2hsaWIgaW1wb3J0IHNoYTEgYXMgc2hhCmZyb20gb3MgaW1wb3J0IGxpbmVzZXAgYXMgTElORVNF
            UApmcm9tIG9zIGltcG9ydCBwYXRoCmZyb20gZW1haWwudXRpbHMgaW1wb3J0IGZvcm1hdGRhdGUK
            aW1wb3J0IHJlcXVlc3RzCmZyb20gcmVxdWVzdHMuYXV0aCBpbXBvcnQgQXV0aEJhc2UKCnB5M2sg
            PSBGYWxzZQppZiBzeXMudmVyc2lvbl9pbmZvLm1ham9yID49IDM6CiAgICBweTNrID0gVHJ1ZQog
            ICAgaW1wb3J0IHVybGxpYi5wYXJzZQogICAgZnJvbSBpcGFkZHJlc3MgaW1wb3J0IGlwX2FkZHJl
            c3MsIElQdjRBZGRyZXNzCgpNb2R1bGVOb3RGb3VuZEVycm9yID0gSW1wb3J0RXJyb3IKCnRyeToK
            ICAgIGltcG9ydCByYWRvcwpleGNlcHQgTW9kdWxlTm90Rm91bmRFcnJvciBhcyBub01vZEVycjoK
            ICAgIHByaW50KGYiRXJyb3I6IHtub01vZEVycn1cbkV4aXRpbmcgdGhlIHNjcmlwdC4uLiIpCiAg
            ICBzeXMuZXhpdCgxKQoKdHJ5OgogICAgaW1wb3J0IHJiZApleGNlcHQgTW9kdWxlTm90Rm91bmRF
            cnJvciBhcyBub01vZEVycjoKICAgIHByaW50KGYiRXJyb3I6IHtub01vZEVycn1cbkV4aXRpbmcg
            dGhlIHNjcmlwdC4uLiIpCiAgICBzeXMuZXhpdCgxKQoKdHJ5OgogICAgIyBmb3IgMi43LngKICAg
            IGZyb20gU3RyaW5nSU8gaW1wb3J0IFN0cmluZ0lPCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9y
            OgogICAgIyBmb3IgMy54CiAgICBmcm9tIGlvIGltcG9ydCBTdHJpbmdJTwoKdHJ5OgogICAgIyBm
            b3IgMi43LngKICAgIGZyb20gdXJscGFyc2UgaW1wb3J0IHVybHBhcnNlCiAgICBmcm9tIHVybGxp
            YiBpbXBvcnQgdXJsZW5jb2RlIGFzIHVybGVuY29kZQpleGNlcHQgTW9kdWxlTm90Rm91bmRFcnJv
            cjoKICAgICMgZm9yIDMueAogICAgZnJvbSB1cmxsaWIucGFyc2UgaW1wb3J0IHVybHBhcnNlCiAg
            ICBmcm9tIHVybGxpYi5wYXJzZSBpbXBvcnQgdXJsZW5jb2RlIGFzIHVybGVuY29kZQoKdHJ5Ogog
            ICAgZnJvbSBiYXNlNjQgaW1wb3J0IGVuY29kZXN0cmluZwpleGNlcHQ6CiAgICBmcm9tIGJhc2U2
            NCBpbXBvcnQgZW5jb2RlYnl0ZXMgYXMgZW5jb2Rlc3RyaW5nCgoKY2xhc3MgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbihFeGNlcHRpb24pOgogICAgcGFzcwoKCiMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMgRHVtbXlSYWRv
            cyAjIyMjIyMjIyMjIyMjIyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjCiMgdGhpcyBpcyBtYWlubHkgZm9yIHRlc3RpbmcgYW5kIGNvdWxkIGJlIHVz
            ZWQgd2hlcmUgJ3JhZG9zJyBpcyBub3QgYXZhaWxhYmxlCgoKY2xhc3MgRHVtbXlSYWRvcyhvYmpl
            Y3QpOgogICAgZGVmIF9faW5pdF9fKHNlbGYpOgogICAgICAgIHNlbGYucmV0dXJuX3ZhbCA9IDAK
            ICAgICAgICBzZWxmLmVycl9tZXNzYWdlID0gIiIKICAgICAgICBzZWxmLnN0YXRlID0gImNvbm5l
            Y3RlZCIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwID0ge30KICAgICAgICBzZWxmLmNtZF9u
            YW1lcyA9IHt9CiAgICAgICAgc2VsZi5faW5pdF9jbWRfb3V0cHV0X21hcCgpCiAgICAgICAgc2Vs
            Zi5kdW1teV9ob3N0X2lwX21hcCA9IHt9CgogICAgZGVmIF9pbml0X2NtZF9vdXRwdXRfbWFwKHNl
            bGYpOgogICAgICAgIGpzb25fZmlsZV9uYW1lID0gInRlc3QtZGF0YS9jZXBoLXN0YXR1cy1vdXQi
            CiAgICAgICAgc2NyaXB0X2RpciA9IHBhdGguYWJzcGF0aChwYXRoLmRpcm5hbWUoX19maWxlX18p
            KQogICAgICAgIGNlcGhfc3RhdHVzX3N0ciA9ICIiCiAgICAgICAgd2l0aCBvcGVuKAogICAgICAg
            ICAgICBwYXRoLmpvaW4oc2NyaXB0X2RpciwganNvbl9maWxlX25hbWUpLCBtb2RlPSJyIiwgZW5j
            b2Rpbmc9IlVURi04IgogICAgICAgICkgYXMganNvbl9maWxlOgogICAgICAgICAgICBjZXBoX3N0
            YXR1c19zdHIgPSBqc29uX2ZpbGUucmVhZCgpCiAgICAgICAgc2VsZi5jbWRfbmFtZXNbImZzIGxz
            Il0gPSAiIiJ7ImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJmcyBscyJ9IiIiCiAgICAgICAg
            c2VsZi5jbWRfbmFtZXNbInF1b3J1bV9zdGF0dXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3Jt
            YXQiOiAianNvbiIsICJwcmVmaXgiOiAicXVvcnVtX3N0YXR1cyJ9IiIiCiAgICAgICAgKQogICAg
            ICAgIHNlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXSA9ICgKICAgICAgICAgICAgIiIieyJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAibWdyIHNlcnZpY2VzIn0iIiIKICAgICAgICApCiAg
            ICAgICAgIyBhbGwgdGhlIGNvbW1hbmRzIGFuZCB0aGVpciBvdXRwdXQKICAgICAgICBzZWxmLmNt
            ZF9vdXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJmcyBscyJdXSA9ICgKICAgICAgICAgICAgIiIi
            W3sibmFtZSI6Im15ZnMiLCJtZXRhZGF0YV9wb29sIjoibXlmcy1tZXRhZGF0YSIsIm1ldGFkYXRh
            X3Bvb2xfaWQiOjIsImRhdGFfcG9vbF9pZHMiOlszXSwiZGF0YV9wb29scyI6WyJteWZzLXJlcGxp
            Y2F0ZWQiXX1dIiIiCiAgICAgICAgKQogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbc2VsZi5j
            bWRfbmFtZXNbInF1b3J1bV9zdGF0dXMiXV0gPSAoCiAgICAgICAgICAgICIiInsiZWxlY3Rpb25f
            ZXBvY2giOjMsInF1b3J1bSI6WzBdLCJxdW9ydW1fbmFtZXMiOlsiYSJdLCJxdW9ydW1fbGVhZGVy
            X25hbWUiOiJhIiwicXVvcnVtX2FnZSI6MTQzODUsImZlYXR1cmVzIjp7InF1b3J1bV9jb24iOiI0
            NTQwMTM4MjkyODM2Njk2MDYzIiwicXVvcnVtX21vbiI6WyJrcmFrZW4iLCJsdW1pbm91cyIsIm1p
            bWljIiwib3NkbWFwLXBydW5lIiwibmF1dGlsdXMiLCJvY3RvcHVzIl19LCJtb25tYXAiOnsiZXBv
            Y2giOjEsImZzaWQiOiJhZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5ZDBmMWMiLCJtb2Rp
            ZmllZCI6IjIwMjAtMDUtMDdUMDM6MzY6MzkuOTE4MDM1WiIsImNyZWF0ZWQiOiIyMDIwLTA1LTA3
            VDAzOjM2OjM5LjkxODAzNVoiLCJtaW5fbW9uX3JlbGVhc2UiOjE1LCJtaW5fbW9uX3JlbGVhc2Vf
            bmFtZSI6Im9jdG9wdXMiLCJmZWF0dXJlcyI6eyJwZXJzaXN0ZW50IjpbImtyYWtlbiIsImx1bWlu
            b3VzIiwibWltaWMiLCJvc2RtYXAtcHJ1bmUiLCJuYXV0aWx1cyIsIm9jdG9wdXMiXSwib3B0aW9u
            YWwiOltdfSwibW9ucyI6W3sicmFuayI6MCwibmFtZSI6ImEiLCJwdWJsaWNfYWRkcnMiOnsiYWRk
            cnZlYyI6W3sidHlwZSI6InYyIiwiYWRkciI6IjEwLjExMC4yMDUuMTc0OjMzMDAiLCJub25jZSI6
            MH0seyJ0eXBlIjoidjEiLCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OSIsIm5vbmNlIjowfV19
            LCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OS8wIiwicHVibGljX2FkZHIiOiIxMC4xMTAuMjA1
            LjE3NDo2Nzg5LzAiLCJwcmlvcml0eSI6MCwid2VpZ2h0IjowfV19fSIiIgogICAgICAgICkKICAg
            ICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXV0g
            PSAoCiAgICAgICAgICAgICIiInsiZGFzaGJvYXJkIjoiaHR0cHM6Ly9jZXBoLWRhc2hib2FyZDo4
            NDQzLyIsInByb21ldGhldXMiOiJodHRwOi8vY2VwaC1kYXNoYm9hcmQtZGI6OTI4My8ifSIiIgog
            ICAgICAgICkKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNh
            cHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMiLCAib3Nk
            IiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0
            YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250
            cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJdLCAiZW50aXR5Ijog
            ImNsaWVudC5oZWFsdGhjaGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRo
            IGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuaGVh
            bHRoY2hlY2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRB
            PT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cyIs
            Im9zZCI6InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5j
            b250cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJ9fV0iIiIKICAg
            ICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIs
            ICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAib3NkIiwgInBy
            b2ZpbGUgcmJkIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtbm9kZSIsICJmb3JtYXQiOiAi
            anNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIi
            W3siZW50aXR5IjoiY2xpZW50LmNzaS1yYmQtbm9kZSIsImtleSI6IkFRQk9nck5lSGJLMUF4QUF1
            YllCZVY4UzFVL0dQenE1U1ZlcTZnPT0iLCJjYXBzIjp7Im1vbiI6InByb2ZpbGUgcmJkLCBhbGxv
            dyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsIm9zZCI6InByb2ZpbGUgcmJkIn19XSIiIgogICAg
            ICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsibW9uIiwg
            InByb2ZpbGUgcmJkLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxs
            b3cgcnciLCAib3NkIiwgInByb2ZpbGUgcmJkIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQt
            cHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNy
            ZWF0ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktcmJkLXByb3Zp
            c2lvbmVyIiwia2V5IjoiQVFCTmdyTmUxZ2V5S3hBQThla1ZpUmRFK2hzczVPd2VZQmt3Tmc9PSIs
            ImNhcHMiOnsibWdyIjoiYWxsb3cgcnciLCJtb24iOiJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFu
            ZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJwcm9maWxlIHJiZCJ9fV0iIiIKICAgICAgICBzZWxm
            LmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyBy
            LCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3Nk
            IiwgImFsbG93IHJ3IHRhZyBjZXBoZnMgKj0qIiwgIm1kcyIsICJhbGxvdyBydyJdLCAiZW50aXR5
            IjogImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4Ijog
            ImF1dGggZ2V0LW9yLWNyZWF0ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVu
            dC5jc2ktY2VwaGZzLW5vZGUiLCJrZXkiOiJBUUJPZ3JOZUVOdW5LeEFBUENtZ0U3UjZHOERjWG5h
            SjFGMzJxZz09IiwiY2FwcyI6eyJtZHMiOiJhbGxvdyBydyIsIm1nciI6ImFsbG93IHJ3IiwibW9u
            IjoiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBy
            dyB0YWcgY2VwaGZzICo9KiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAg
            ICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kICdvc2Qg
            YmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3NkIiwgImFsbG93IHJ3IHRhZyBjZXBo
            ZnMgbWV0YWRhdGE9KiJdLCAiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVy
            IiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgog
            ICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIs
            ImtleSI6IkFRQk9nck5lQUZnY0dCQUF2R3FLT0FEMEQzeHhtVlkwUjkxMmRnPT0iLCJjYXBzIjp7
            Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2Nr
            bGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoifX1dIiIiCiAgICAg
            ICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAi
            YWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdyIiwgImFsbG93IHJ3
            IiwgIm9zZCIsICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiXSwgImVudGl0eSI6ICJj
            bGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVuc2hpZnQtc3RvcmFnZSIsICJmb3JtYXQi
            OiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0g
            IiIiW3siZW50aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0
            b3JhZ2UiLCJrZXkiOiJCUUJPZ3JOZUFGZ2NHQkFBdkdxS09BRDBEM3h4bVZZMFI5MTJkZz09Iiwi
            Y2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29z
            ZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIn19XSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsi
            bW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciIsICJh
            bGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT1teWZzIl0sICJl
            bnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0b3JhZ2Ut
            bXlmcyIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0i
            IiIKICAgICAgICBdID0gIiIiW3siZW50aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9u
            ZXItb3BlbnNoaWZ0LXN0b3JhZ2UtbXlmcyIsImtleSI6IkNRQk9nck5lQUZnY0dCQUF2R3FLT0FE
            MEQzeHhtVlkwUjkxMmRnPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cg
            ciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2Vw
            aGZzIG1ldGFkYXRhPW15ZnMifX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAg
            ICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCBxdW9y
            dW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIiwgImFsbG93IGNvbW1hbmQg
            Y29uZmlnIiwgIm9zZCIsICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFsbG93IHJ3eCBwb29sPWRl
            ZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9ZGVm
            YXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJndy5sb2csIGFsbG93IHgg
            cG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4Il0sICJlbnRpdHkiOiAiY2xpZW50LmhlYWx0
            aGNoZWNrZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0
            ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5oZWFsdGhjaGVja2VyIiwi
            a2V5IjoiQVFERmtiTmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsi
            bW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFu
            ZCB2ZXJzaW9uIiwgIm1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsICJvc2QiOiAicHJvZmls
            ZSByYmQtcmVhZC1vbmx5LCBhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhLCBhbGxvdyBy
            IHBvb2w9LnJndy5yb290LCBhbGxvdyBydyBwb29sPWRlZmF1bHQucmd3LmNvbnRyb2wsIGFsbG93
            IHJ4IHBvb2w9ZGVmYXVsdC5yZ3cubG9nLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0
            cy5pbmRleCJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAi
            IiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMs
            IGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cgY29tbWFuZCBjb25maWciLCAi
            b3NkIiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5j
            b250cm9sLCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1
            bHQucmd3LmJ1Y2tldHMuaW5kZXgiXSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIs
            ICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBjYXBzIn0iIiIKICAgICAgICBdID0g
            IiIiW3siZW50aXR5IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0NWJG
            UkFBVG5kTE5VU1JLcnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIjog
            ImFsbG93IGNvbW1hbmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFs
            bG93IHJ3eCBwb29sPWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFs
            bG93IHJ3IHBvb2w9ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJn
            dy5sb2csIGFsbG93IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4In19XSIiIgogICAg
            ICAgIHNlbGYuY21kX291dHB1dF9tYXBbIiIieyJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            bWdyIHNlcnZpY2VzIn0iIiJdID0gKAogICAgICAgICAgICAiIiJ7ImRhc2hib2FyZCI6ICJodHRw
            Oi8vcm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6NzAwMC8iLCAicHJvbWV0aGV1cyI6
            ICJodHRwOi8vcm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6OTI4My8ifSIiIgogICAg
            ICAgICkKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0
            eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            YXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJ7ImRhc2hib2FyZCI6ICJodHRwOi8vcm9vay1j
            ZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6NzAwMC8iLCAicHJvbWV0aGV1cyI6ICJodHRwOi8v
            cm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6OTI4My8ifSIiIgogICAgICAgIHNlbGYu
            Y21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNsaWVudC5oZWFsdGhj
            aGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldCJ9IiIiCiAgICAg
            ICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5oZWFsdGhjaGVja2VyIiwia2V5IjoiQVFERmti
            TmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjogImFsbG93
            IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFuZCB2ZXJzaW9uIiwg
            Im1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsICJvc2QiOiAicHJvZmlsZSByYmQtcmVhZC1v
            bmx5LCBhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhLCBhbGxvdyByIHBvb2w9LnJndy5y
            b290LCBhbGxvdyBydyBwb29sPWRlZmF1bHQucmd3LmNvbnRyb2wsIGFsbG93IHJ4IHBvb2w9ZGVm
            YXVsdC5yZ3cubG9nLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJ9fV0i
            IiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0eSI6
            ICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJh
            dXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0
            X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtbm9kZSIsICJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJb
            XSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5
            IjogImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRf
            b3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMt
            cHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIK
            ICAgICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAg
            ICAgICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVuc2hp
            ZnQtc3RvcmFnZSIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgog
            ICAgICAgIF0gPSAiIiJbXSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAg
            ICAgICIiInsiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlm
            dC1zdG9yYWdlLW15ZnMiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0i
            IiIKICAgICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAg
            ICAgICAgICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJb
            eyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsImtleSI6IkFRREZrYk5l
            ZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRBPT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIi
            LCAibWdyIjoiYWxsb3cgcnciLCAib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0q
            In19XSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2Fw
            cyI6IFsibW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1n
            ciIsICJhbGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIl0s
            ICJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpz
            b24iLCAicHJlZml4IjogImF1dGggY2FwcyJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmtiTmVmdDViRlJBQVRu
            ZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cgciwgIGFsbG93IGNv
            bW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciI6ImFsbG93IHJ3IiwgIm9zZCI6ImFsbG93IHJ3
            IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFw
            Wyd7ImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJzdGF0dXMifSddID0gY2VwaF9zdGF0dXNf
            c3RyCgogICAgZGVmIHNodXRkb3duKHNlbGYpOgogICAgICAgIHBhc3MKCiAgICBkZWYgZ2V0X2Zz
            aWQoc2VsZik6CiAgICAgICAgcmV0dXJuICJhZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5
            ZDBmMWMiCgogICAgZGVmIGNvbmZfcmVhZF9maWxlKHNlbGYpOgogICAgICAgIHBhc3MKCiAgICBk
            ZWYgY29ubmVjdChzZWxmKToKICAgICAgICBwYXNzCgogICAgZGVmIHBvb2xfZXhpc3RzKHNlbGYs
            IHBvb2xfbmFtZSk6CiAgICAgICAgcmV0dXJuIFRydWUKCiAgICBkZWYgbW9uX2NvbW1hbmQoc2Vs
            ZiwgY21kLCBvdXQpOgogICAgICAgIGpzb25fY21kID0ganNvbi5sb2FkcyhjbWQpCiAgICAgICAg
            anNvbl9jbWRfc3RyID0ganNvbi5kdW1wcyhqc29uX2NtZCwgc29ydF9rZXlzPVRydWUpCiAgICAg
            ICAgY21kX291dHB1dCA9IHNlbGYuY21kX291dHB1dF9tYXBbanNvbl9jbWRfc3RyXQogICAgICAg
            IHJldHVybiBzZWxmLnJldHVybl92YWwsIGNtZF9vdXRwdXQsIHN0cihzZWxmLmVycl9tZXNzYWdl
            LmVuY29kZSgidXRmLTgiKSkKCiAgICBkZWYgX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoc2VsZiwg
            aG9zdF9uYW1lKToKICAgICAgICBpcF9yZWdfeCA9IHJlLmNvbXBpbGUociJcZHsxLDN9LlxkezEs
            M30uXGR7MSwzfS5cZHsxLDN9IikKICAgICAgICAjIGlmIHByb3ZpZGVkIGhvc3QgaXMgZGlyZWN0
            bHkgYW4gSVAgYWRkcmVzcywgcmV0dXJuIHRoZSBzYW1lCiAgICAgICAgaWYgaXBfcmVnX3gubWF0
            Y2goaG9zdF9uYW1lKToKICAgICAgICAgICAgcmV0dXJuIGhvc3RfbmFtZQogICAgICAgIGltcG9y
            dCByYW5kb20KCiAgICAgICAgaG9zdF9pcCA9IHNlbGYuZHVtbXlfaG9zdF9pcF9tYXAuZ2V0KGhv
            c3RfbmFtZSwgIiIpCiAgICAgICAgaWYgbm90IGhvc3RfaXA6CiAgICAgICAgICAgIGhvc3RfaXAg
            PSBmIjE3Mi45LntyYW5kb20ucmFuZGludCgwLCAyNTQpfS57cmFuZG9tLnJhbmRpbnQoMCwgMjU0
            KX0iCiAgICAgICAgICAgIHNlbGYuZHVtbXlfaG9zdF9pcF9tYXBbaG9zdF9uYW1lXSA9IGhvc3Rf
            aXAKICAgICAgICBkZWwgcmFuZG9tCiAgICAgICAgcmV0dXJuIGhvc3RfaXAKCiAgICBAY2xhc3Nt
            ZXRob2QKICAgIGRlZiBSYWRvcyhjb25mZmlsZT1Ob25lKToKICAgICAgICByZXR1cm4gRHVtbXlS
            YWRvcygpCgoKY2xhc3MgUzNBdXRoKEF1dGhCYXNlKToKICAgICIiIkF0dGFjaGVzIEFXUyBBdXRo
            ZW50aWNhdGlvbiB0byB0aGUgZ2l2ZW4gUmVxdWVzdCBvYmplY3QuIiIiCgogICAgc2VydmljZV9i
            YXNlX3VybCA9ICJzMy5hbWF6b25hd3MuY29tIgoKICAgIGRlZiBfX2luaXRfXyhzZWxmLCBhY2Nl
            c3Nfa2V5LCBzZWNyZXRfa2V5LCBzZXJ2aWNlX3VybD1Ob25lKToKICAgICAgICBpZiBzZXJ2aWNl
            X3VybDoKICAgICAgICAgICAgc2VsZi5zZXJ2aWNlX2Jhc2VfdXJsID0gc2VydmljZV91cmwKICAg
            ICAgICBzZWxmLmFjY2Vzc19rZXkgPSBzdHIoYWNjZXNzX2tleSkKICAgICAgICBzZWxmLnNlY3Jl
            dF9rZXkgPSBzdHIoc2VjcmV0X2tleSkKCiAgICBkZWYgX19jYWxsX18oc2VsZiwgcik6CiAgICAg
            ICAgIyBDcmVhdGUgZGF0ZSBoZWFkZXIgaWYgaXQgaXMgbm90IGNyZWF0ZWQgeWV0LgogICAgICAg
            IGlmICJkYXRlIiBub3QgaW4gci5oZWFkZXJzIGFuZCAieC1hbXotZGF0ZSIgbm90IGluIHIuaGVh
            ZGVyczoKICAgICAgICAgICAgci5oZWFkZXJzWyJkYXRlIl0gPSBmb3JtYXRkYXRlKHRpbWV2YWw9
            Tm9uZSwgbG9jYWx0aW1lPUZhbHNlLCB1c2VnbXQ9VHJ1ZSkKICAgICAgICBzaWduYXR1cmUgPSBz
            ZWxmLmdldF9zaWduYXR1cmUocikKICAgICAgICBpZiBweTNrOgogICAgICAgICAgICBzaWduYXR1
            cmUgPSBzaWduYXR1cmUuZGVjb2RlKCJ1dGYtOCIpCiAgICAgICAgci5oZWFkZXJzWyJBdXRob3Jp
            emF0aW9uIl0gPSBmIkFXUyB7c2VsZi5hY2Nlc3Nfa2V5fTp7c2lnbmF0dXJlfSIKICAgICAgICBy
            ZXR1cm4gcgoKICAgIGRlZiBnZXRfc2lnbmF0dXJlKHNlbGYsIHIpOgogICAgICAgIGNhbm9uaWNh
            bF9zdHJpbmcgPSBzZWxmLmdldF9jYW5vbmljYWxfc3RyaW5nKHIudXJsLCByLmhlYWRlcnMsIHIu
            bWV0aG9kKQogICAgICAgIGlmIHB5M2s6CiAgICAgICAgICAgIGtleSA9IHNlbGYuc2VjcmV0X2tl
            eS5lbmNvZGUoInV0Zi04IikKICAgICAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZy5lbmNv
            ZGUoInV0Zi04IikKICAgICAgICBlbHNlOgogICAgICAgICAgICBrZXkgPSBzZWxmLnNlY3JldF9r
            ZXkKICAgICAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZwogICAgICAgIGggPSBobWFjLm5l
            dyhrZXksIG1zZywgZGlnZXN0bW9kPXNoYSkKICAgICAgICByZXR1cm4gZW5jb2Rlc3RyaW5nKGgu
            ZGlnZXN0KCkpLnN0cmlwKCkKCiAgICBkZWYgZ2V0X2Nhbm9uaWNhbF9zdHJpbmcoc2VsZiwgdXJs
            LCBoZWFkZXJzLCBtZXRob2QpOgogICAgICAgIHBhcnNlZHVybCA9IHVybHBhcnNlKHVybCkKICAg
            ICAgICBvYmplY3RrZXkgPSBwYXJzZWR1cmwucGF0aFsxOl0KCiAgICAgICAgYnVja2V0ID0gcGFy
            c2VkdXJsLm5ldGxvY1s6IC1sZW4oc2VsZi5zZXJ2aWNlX2Jhc2VfdXJsKV0KICAgICAgICBpZiBs
            ZW4oYnVja2V0KSA+IDE6CiAgICAgICAgICAgICMgcmVtb3ZlIGxhc3QgZG90CiAgICAgICAgICAg
            IGJ1Y2tldCA9IGJ1Y2tldFs6LTFdCgogICAgICAgIGludGVyZXN0aW5nX2hlYWRlcnMgPSB7ImNv
            bnRlbnQtbWQ1IjogIiIsICJjb250ZW50LXR5cGUiOiAiIiwgImRhdGUiOiAiIn0KICAgICAgICBm
            b3Iga2V5IGluIGhlYWRlcnM6CiAgICAgICAgICAgIGxrID0ga2V5Lmxvd2VyKCkKICAgICAgICAg
            ICAgdHJ5OgogICAgICAgICAgICAgICAgbGsgPSBsay5kZWNvZGUoInV0Zi04IikKICAgICAgICAg
            ICAgZXhjZXB0OgogICAgICAgICAgICAgICAgcGFzcwogICAgICAgICAgICBpZiBoZWFkZXJzW2tl
            eV0gYW5kICgKICAgICAgICAgICAgICAgIGxrIGluIGludGVyZXN0aW5nX2hlYWRlcnMua2V5cygp
            IG9yIGxrLnN0YXJ0c3dpdGgoIngtYW16LSIpCiAgICAgICAgICAgICk6CiAgICAgICAgICAgICAg
            ICBpbnRlcmVzdGluZ19oZWFkZXJzW2xrXSA9IGhlYWRlcnNba2V5XS5zdHJpcCgpCgogICAgICAg
            ICMgSWYgeC1hbXotZGF0ZSBpcyB1c2VkIGl0IHN1cGVyc2VkZXMgdGhlIGRhdGUgaGVhZGVyLgog
            ICAgICAgIGlmIG5vdCBweTNrOgogICAgICAgICAgICBpZiAieC1hbXotZGF0ZSIgaW4gaW50ZXJl
            c3RpbmdfaGVhZGVyczoKICAgICAgICAgICAgICAgIGludGVyZXN0aW5nX2hlYWRlcnNbImRhdGUi
            XSA9ICIiCiAgICAgICAgZWxzZToKICAgICAgICAgICAgaWYgIngtYW16LWRhdGUiIGluIGludGVy
            ZXN0aW5nX2hlYWRlcnM6CiAgICAgICAgICAgICAgICBpbnRlcmVzdGluZ19oZWFkZXJzWyJkYXRl
            Il0gPSAiIgoKICAgICAgICBidWYgPSBmInttZXRob2R9XG4iCiAgICAgICAgZm9yIGtleSBpbiBz
            b3J0ZWQoaW50ZXJlc3RpbmdfaGVhZGVycy5rZXlzKCkpOgogICAgICAgICAgICB2YWwgPSBpbnRl
            cmVzdGluZ19oZWFkZXJzW2tleV0KICAgICAgICAgICAgaWYga2V5LnN0YXJ0c3dpdGgoIngtYW16
            LSIpOgogICAgICAgICAgICAgICAgYnVmICs9IGYie2tleX06e3ZhbH1cbiIKICAgICAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgICAgIGJ1ZiArPSBmInt2YWx9XG4iCgogICAgICAgICMgYXBwZW5k
            IHRoZSBidWNrZXQgaWYgaXQgZXhpc3RzCiAgICAgICAgaWYgYnVja2V0ICE9ICIiOgogICAgICAg
            ICAgICBidWYgKz0gZiIve2J1Y2tldH0iCgogICAgICAgICMgYWRkIHRoZSBvYmplY3RrZXkuIGV2
            ZW4gaWYgaXQgZG9lc24ndCBleGlzdCwgYWRkIHRoZSBzbGFzaAogICAgICAgIGJ1ZiArPSBmIi97
            b2JqZWN0a2V5fSIKCiAgICAgICAgcmV0dXJuIGJ1ZgoKCmNsYXNzIFJhZG9zSlNPTjoKICAgIEVY
            VEVSTkFMX1VTRVJfTkFNRSA9ICJjbGllbnQuaGVhbHRoY2hlY2tlciIKICAgIEVYVEVSTkFMX1JH
            V19BRE1JTl9PUFNfVVNFUl9OQU1FID0gInJndy1hZG1pbi1vcHMtdXNlciIKICAgIEVNUFRZX09V
            VFBVVF9MSVNUID0gIkVtcHR5IG91dHB1dCBsaXN0IgogICAgREVGQVVMVF9SR1dfUE9PTF9QUkVG
            SVggPSAiZGVmYXVsdCIKICAgIERFRkFVTFRfTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUID0gIjky
            ODMiCgogICAgQGNsYXNzbWV0aG9kCiAgICBkZWYgZ2VuX2FyZ19wYXJzZXIoY2xzLCBhcmdzX3Rv
            X3BhcnNlPU5vbmUpOgogICAgICAgIGFyZ1AgPSBhcmdwYXJzZS5Bcmd1bWVudFBhcnNlcigpCgog
            ICAgICAgIGNvbW1vbl9ncm91cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJjb21tb24iKQog
            ICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoIi0tdmVyYm9zZSIsICItdiIsIGFjdGlv
            bj0ic3RvcmVfdHJ1ZSIsIGRlZmF1bHQ9RmFsc2UpCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9h
            cmd1bWVudCgKICAgICAgICAgICAgIi0tY2VwaC1jb25mIiwgIi1jIiwgaGVscD0iUHJvdmlkZSBh
            IGNlcGggY29uZiBmaWxlLiIsIHR5cGU9c3RyCiAgICAgICAgKQogICAgICAgIGNvbW1vbl9ncm91
            cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLWtleXJpbmciLCAiLWsiLCBoZWxwPSJQYXRo
            IHRvIGNlcGgga2V5cmluZyBmaWxlLiIsIHR5cGU9c3RyCiAgICAgICAgKQogICAgICAgIGNvbW1v
            bl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJ1bi1hcy11c2VyIiwKICAgICAg
            ICAgICAgIi11IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgdHlwZT1zdHIs
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIGEgdXNlciBuYW1lIHRvIGNoZWNrIHRoZSBjbHVz
            dGVyJ3MgaGVhbHRoIHN0YXR1cywgbXVzdCBiZSBwcmVmaXhlZCBieSAnY2xpZW50LiciLAogICAg
            ICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1j
            bHVzdGVyLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJL
            dWJlcm5ldGVzIGNsdXN0ZXIgbmFtZShsZWdhY3kgZmxhZyksIE5vdGU6IEVpdGhlciB1c2UgdGhp
            cyBvciAtLWs4cy1jbHVzdGVyLW5hbWUiLAogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAu
            YWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1rOHMtY2x1c3Rlci1uYW1lIiwgZGVmYXVsdD0i
            IiwgaGVscD0iS3ViZXJuZXRlcyBjbHVzdGVyIG5hbWUiCiAgICAgICAgKQogICAgICAgIGNvbW1v
            bl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLW5hbWVzcGFjZSIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIGhlbHA9Ik5hbWVzcGFjZSB3aGVyZSBDZXBoQ2x1
            c3RlciBpcyBydW5uaW5nIiwKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tcmd3LXBvb2wtcHJlZml4IiwgZGVmYXVsdD0iIiwgaGVscD0i
            UkdXIFBvb2wgcHJlZml4IgogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3Vt
            ZW50KAogICAgICAgICAgICAiLS1yZXN0cmljdGVkLWF1dGgtcGVybWlzc2lvbiIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlJlc3RyaWN0IGNlcGhDU0lLZXly
            aW5ncyBhdXRoIHBlcm1pc3Npb25zIHRvIHNwZWNpZmljIHBvb2xzLCBjbHVzdGVyLiIKICAgICAg
            ICAgICAgKyAiTWFuZGF0b3J5IGZsYWdzIHRoYXQgbmVlZCB0byBiZSBzZXQgYXJlIC0tcmJkLWRh
            dGEtcG9vbC1uYW1lLCBhbmQgLS1rOHMtY2x1c3Rlci1uYW1lLiIKICAgICAgICAgICAgKyAiLS1j
            ZXBoZnMtZmlsZXN5c3RlbS1uYW1lIGZsYWcgY2FuIGFsc28gYmUgcGFzc2VkIGluIGNhc2Ugb2Yg
            Y2VwaGZzIHVzZXIgcmVzdHJpY3Rpb24sIHNvIGl0IGNhbiByZXN0cmljdCB1c2VyIHRvIHBhcnRp
            Y3VsYXIgY2VwaGZzIGZpbGVzeXN0ZW0iCiAgICAgICAgICAgICsgInNhbXBsZSBydW46IGBweXRo
            b24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwtY2x1c3Rlci1yZXNvdXJjZXMucHkgLS1jZXBo
            ZnMtZmlsZXN5c3RlbS1uYW1lIG15ZnMgLS1yYmQtZGF0YS1wb29sLW5hbWUgcmVwbGljYXBvb2wg
            LS1rOHMtY2x1c3Rlci1uYW1lIHJvb2tzdG9yYWdlIC0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Np
            b24gdHJ1ZWAiCiAgICAgICAgICAgICsgIk5vdGU6IFJlc3RyaWN0aW5nIHRoZSBjc2ktdXNlcnMg
            cGVyIHBvb2wsIGFuZCBwZXIgY2x1c3RlciB3aWxsIHJlcXVpcmUgY3JlYXRpbmcgbmV3IGNzaS11
            c2VycyBhbmQgbmV3IHNlY3JldHMgZm9yIHRoYXQgY3NpLXVzZXJzLiIKICAgICAgICAgICAgKyAi
            U28gYXBwbHkgdGhlc2Ugc2VjcmV0cyBvbmx5IHRvIG5ldyBgQ29uc3VtZXIgY2x1c3RlcmAgZGVw
            bG95bWVudCB3aGlsZSB1c2luZyB0aGUgc2FtZSBgU291cmNlIGNsdXN0ZXJgLiIsCiAgICAgICAg
            KQogICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXYyLXBv
            cnQtZW5hYmxlIiwKICAgICAgICAgICAgYWN0aW9uPSJzdG9yZV90cnVlIiwKICAgICAgICAgICAg
            ZGVmYXVsdD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iRW5hYmxlIHYyIG1vbiBwb3J0KDMzMDAp
            IGZvciBtb25zIiwKICAgICAgICApCgogICAgICAgIG91dHB1dF9ncm91cCA9IGFyZ1AuYWRkX2Fy
            Z3VtZW50X2dyb3VwKCJvdXRwdXQiKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQo
            CiAgICAgICAgICAgICItLWZvcm1hdCIsCiAgICAgICAgICAgICItdCIsCiAgICAgICAgICAgIGNo
            b2ljZXM9WyJqc29uIiwgImJhc2giXSwKICAgICAgICAgICAgZGVmYXVsdD0ianNvbiIsCiAgICAg
            ICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBvdXRwdXQgZm9ybWF0IChqc29uIHwgYmFzaCkiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1vdXRwdXQiLAogICAgICAgICAgICAiLW8iLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAg
            ICAgICAgICBoZWxwPSJPdXRwdXQgd2lsbCBiZSBzdG9yZWQgaW50byB0aGUgcHJvdmlkZWQgZmls
            ZSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAog
            ICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgQ2VwaCBmaWxlc3lzdGVt
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tY2VwaGZzLW1ldGFkYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIs
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBjZXBoZnMgbWV0YWRh
            dGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAg
            ICAgICAgICAgICItLWNlcGhmcy1kYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9
            IiIsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBjZXBoZnMgZGF0
            YSBwb29sIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAg
            ICAgICAgICAgIi0tcmJkLWRhdGEtcG9vbC1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwK
            ICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRo
            ZSBuYW1lIG9mIHRoZSBSQkQgZGF0YXBvb2wiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3Jv
            dXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1hbGlhcy1yYmQtZGF0YS1wb29sLW5hbWUi
            LAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAg
            ICAgICAgICAgaGVscD0iUHJvdmlkZXMgYW4gYWxpYXMgZm9yIHRoZSAgUkJEIGRhdGEgcG9vbCBu
            YW1lLCBuZWNlc3NhcnkgaWYgYSBzcGVjaWFsIGNoYXJhY3RlciBpcyBwcmVzZW50IGluIHRoZSBw
            b29sIG5hbWUgc3VjaCBhcyBhIHBlcmlvZCBvciB1bmRlcnNjb3JlIiwKICAgICAgICApCiAgICAg
            ICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmd3LWVuZHBvaW50
            IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAg
            ICAgICAgICAgIGhlbHA9IlJBRE9TIEdhdGV3YXkgZW5kcG9pbnQgKGluIGA8SVB2ND46PFBPUlQ+
            YCBvciBgPFtJUHY2XT46PFBPUlQ+YCBvciBgPEZRRE4+OjxQT1JUPmAgZm9ybWF0KSIsCiAgICAg
            ICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJn
            dy10bHMtY2VydC1wYXRoIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVx
            dWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlJBRE9TIEdhdGV3YXkgZW5kcG9pbnQgVExT
            IGNlcnRpZmljYXRlIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVu
            dCgKICAgICAgICAgICAgIi0tcmd3LXNraXAtdGxzIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFs
            c2UsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9Iklnbm9yZSBU
            TFMgY2VydGlmaWNhdGlvbiB2YWxpZGF0aW9uIHdoZW4gYSBzZWxmLXNpZ25lZCBjZXJ0aWZpY2F0
            ZSBpcyBwcm92aWRlZCAoTk9UIFJFQ09NTUVOREVEIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tbW9uaXRvcmluZy1lbmRwb2ludCIs
            CiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAg
            ICAgICAgICBoZWxwPSJDZXBoIE1hbmFnZXIgcHJvbWV0aGV1cyBleHBvcnRlciBlbmRwb2ludHMg
            KGNvbW1hIHNlcGFyYXRlZCBsaXN0IG9mIChmb3JtYXQgYDxJUHY0PmAgb3IgYDxbSVB2Nl0+YCBv
            ciBgPEZRRE4+YCkgZW50cmllcyBvZiBhY3RpdmUgYW5kIHN0YW5kYnkgbWdycykiLAogICAgICAg
            ICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1tb25p
            dG9yaW5nLWVuZHBvaW50LXBvcnQiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAg
            ICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iQ2VwaCBNYW5hZ2VyIHByb21ldGhl
            dXMgZXhwb3J0ZXIgcG9ydCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJn
            dW1lbnQoCiAgICAgICAgICAgICItLXNraXAtbW9uaXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAg
            ICAgICAgIGhlbHA9IkRvIG5vdCBjaGVjayBmb3IgYSBtb25pdG9yaW5nIGVuZHBvaW50IGZvciB0
            aGUgQ2VwaCBjbHVzdGVyIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tcmJkLW1ldGFkYXRhLWVjLXBvb2wtbmFtZSIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBo
            ZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiBlcmFzdXJlIGNvZGVkIFJCRCBtZXRhZGF0YSBwb29s
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tZHJ5LXJ1biIsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFj
            dGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGhlbHA9IkRyeSBydW4gcHJpbnRzIHRoZSBl
            eGVjdXRlZCBjb21tYW5kcyB3aXRob3V0IHJ1bm5pbmcgdGhlbSIsCiAgICAgICAgKQogICAgICAg
            IG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJhZG9zLW5hbWVzcGFj
            ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAog
            ICAgICAgICAgICBoZWxwPSJEaXZpZGVzIGEgcG9vbCBpbnRvIHNlcGFyYXRlIGxvZ2ljYWwgbmFt
            ZXNwYWNlcywgdXNlZCBmb3IgY3JlYXRpbmcgUkJEIFBWQyBpbiBhIENlcGhCbG9ja1Bvb2xSYWRv
            c05hbWVzcGFjZSAoc2hvdWxkIGJlIGxvd2VyIGNhc2UpIiwKICAgICAgICApCiAgICAgICAgb3V0
            cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tc3Vidm9sdW1lLWdyb3VwIiwK
            ICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAg
            ICAgICAgIGhlbHA9InByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBzdWJ2b2x1bWUgZ3JvdXAiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1yZ3ctcmVhbG0tbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJl
            cXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUg
            cmd3LXJlYWxtIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgK
            ICAgICAgICAgICAgIi0tcmd3LXpvbmUtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAg
            ICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUg
            bmFtZSBvZiB0aGUgcmd3LXpvbmUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS1yZ3ctem9uZWdyb3VwLW5hbWUiLAogICAgICAgICAg
            ICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVs
            cD0icHJvdmlkZXMgdGhlIG5hbWUgb2YgdGhlIHJndy16b25lZ3JvdXAiLAogICAgICAgICkKCiAg
            ICAgICAgdXBncmFkZV9ncm91cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJ1cGdyYWRlIikK
            ICAgICAgICB1cGdyYWRlX2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tdXBncmFk
            ZSIsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9
            RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlVwZ3JhZGVzIHRoZSBjZXBoQ1NJS2V5cmluZ3MoRm9y
            IGV4YW1wbGU6IGNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyKSBhbmQgY2xpZW50LmhlYWx0
            aGNoZWNrZXIgY2VwaCB1c2VycyB3aXRoIG5ldyBwZXJtaXNzaW9ucyBuZWVkZWQgZm9yIHRoZSBu
            ZXcgY2x1c3RlciB2ZXJzaW9uIGFuZCBvbGRlciBwZXJtaXNzaW9uIHdpbGwgc3RpbGwgYmUgYXBw
            bGllZC4iCiAgICAgICAgICAgICsgIlNhbXBsZSBydW46IGBweXRob24zIC9ldGMvY2VwaC9jcmVh
            dGUtZXh0ZXJuYWwtY2x1c3Rlci1yZXNvdXJjZXMucHkgLS11cGdyYWRlYCwgdGhpcyB3aWxsIHVw
            Z3JhZGUgYWxsIHRoZSBkZWZhdWx0IGNzaSB1c2Vycyhub24tcmVzdHJpY3RlZCkiCiAgICAgICAg
            ICAgICsgIkZvciByZXN0cmljdGVkIHVzZXJzKEZvciBleGFtcGxlOiBjbGllbnQuY3NpLWNlcGhm
            cy1wcm92aXNpb25lci1vcGVuc2hpZnQtc3RvcmFnZS1teWZzKSwgdXNlcnMgY3JlYXRlZCB1c2lu
            ZyAtLXJlc3RyaWN0ZWQtYXV0aC1wZXJtaXNzaW9uIGZsYWcgbmVlZCB0byBwYXNzIG1hbmRhdG9y
            eSBmbGFncyIKICAgICAgICAgICAgKyAibWFuZGF0b3J5IGZsYWdzOiAnLS1yYmQtZGF0YS1wb29s
            LW5hbWUsIC0tazhzLWNsdXN0ZXItbmFtZSBhbmQgLS1ydW4tYXMtdXNlcicgZmxhZ3Mgd2hpbGUg
            dXBncmFkaW5nIgogICAgICAgICAgICArICJpbiBjYXNlIG9mIGNlcGhmcyB1c2VycyBpZiB5b3Ug
            aGF2ZSBwYXNzZWQgLS1jZXBoZnMtZmlsZXN5c3RlbS1uYW1lIGZsYWcgd2hpbGUgY3JlYXRpbmcg
            dXNlciB0aGVuIHdoaWxlIHVwZ3JhZGluZyBpdCB3aWxsIGJlIG1hbmRhdG9yeSB0b28iCiAgICAg
            ICAgICAgICsgIlNhbXBsZSBydW46IGBweXRob24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwt
            Y2x1c3Rlci1yZXNvdXJjZXMucHkgLS11cGdyYWRlIC0tcmJkLWRhdGEtcG9vbC1uYW1lIHJlcGxp
            Y2Fwb29sIC0tazhzLWNsdXN0ZXItbmFtZSByb29rc3RvcmFnZSAgLS1ydW4tYXMtdXNlciBjbGll
            bnQuY3NpLXJiZC1ub2RlLXJvb2tzdG9yYWdlLXJlcGxpY2Fwb29sYCIKICAgICAgICAgICAgKyAi
            UFM6IEFuIGV4aXN0aW5nIG5vbi1yZXN0cmljdGVkIHVzZXIgY2Fubm90IGJlIGNvbnZlcnRlZCB0
            byBhIHJlc3RyaWN0ZWQgdXNlciBieSB1cGdyYWRpbmcuIgogICAgICAgICAgICArICJVcGdyYWRl
            IGZsYWcgc2hvdWxkIG9ubHkgYmUgdXNlZCB0byBhcHBlbmQgbmV3IHBlcm1pc3Npb25zIHRvIHVz
            ZXJzLCBpdCBzaG91bGRuJ3QgYmUgdXNlZCBmb3IgY2hhbmdpbmcgdXNlciBhbHJlYWR5IGFwcGxp
            ZWQgcGVybWlzc2lvbiwgZm9yIGV4YW1wbGUgeW91IHNob3VsZG4ndCBjaGFuZ2UgaW4gd2hpY2gg
            cG9vbCB1c2VyIGhhcyBhY2Nlc3MiLAogICAgICAgICkKCiAgICAgICAgaWYgYXJnc190b19wYXJz
            ZToKICAgICAgICAgICAgYXNzZXJ0ICgKICAgICAgICAgICAgICAgIHR5cGUoYXJnc190b19wYXJz
            ZSkgPT0gbGlzdAogICAgICAgICAgICApLCAiQXJndW1lbnQgdG8gJ2dlbl9hcmdfcGFyc2VyJyBz
            aG91bGQgYmUgYSBsaXN0IgogICAgICAgIGVsc2U6CiAgICAgICAgICAgIGFyZ3NfdG9fcGFyc2Ug
            PSBzeXMuYXJndlsxOl0KICAgICAgICByZXR1cm4gYXJnUC5wYXJzZV9hcmdzKGFyZ3NfdG9fcGFy
            c2UpCgogICAgZGVmIHZhbGlkYXRlX3JiZF9tZXRhZGF0YV9lY19wb29sX25hbWUoc2VsZik6CiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lOgogICAg
            ICAgICAgICByYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5yYmRf
            bWV0YWRhdGFfZWNfcG9vbF9uYW1lCiAgICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9h
            cmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQoKICAgICAgICAgICAgaWYgcmJkX3Bvb2xfbmFt
            ZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24o
            CiAgICAgICAgICAgICAgICAgICAgIkZsYWcgJy0tcmJkLWRhdGEtcG9vbC1uYW1lJyBzaG91bGQg
            bm90IGJlIGVtcHR5IgogICAgICAgICAgICAgICAgKQoKICAgICAgICAgICAgaWYgcmJkX21ldGFk
            YXRhX2VjX3Bvb2xfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIkZsYWcgJy0tcmJkLW1ldGFkYXRh
            LWVjLXBvb2wtbmFtZScgc2hvdWxkIG5vdCBiZSBlbXB0eSIKICAgICAgICAgICAgICAgICkKCiAg
            ICAgICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAib3NkIGR1bXAiLCAiZm9ybWF0IjogImpz
            b24ifQogICAgICAgICAgICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1v
            bl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBs
            ZW4oanNvbl9vdXQpID09IDA6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJl
            RXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYie2NtZF9qc29uWydwcmVmaXgnXX0gY29t
            bWFuZCBmYWlsZWQuXG4iCiAgICAgICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYg
            cmV0X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAgICAgICAg
            ICApCiAgICAgICAgICAgIG1ldGFkYXRhX3Bvb2xfZXhpc3QsIHBvb2xfZXhpc3QgPSBGYWxzZSwg
            RmFsc2UKCiAgICAgICAgICAgIGZvciBrZXkgaW4ganNvbl9vdXRbInBvb2xzIl06CiAgICAgICAg
            ICAgICAgICAjIGlmIGVyYXN1cmVfY29kZV9wcm9maWxlIGlzIGVtcHR5IGFuZCBwb29sIG5hbWUg
            ZXhpc3RzIHRoZW4gaXQgcmVwbGljYSBwb29sCiAgICAgICAgICAgICAgICBpZiAoCiAgICAgICAg
            ICAgICAgICAgICAga2V5WyJlcmFzdXJlX2NvZGVfcHJvZmlsZSJdID09ICIiCiAgICAgICAgICAg
            ICAgICAgICAgYW5kIGtleVsicG9vbF9uYW1lIl0gPT0gcmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFt
            ZQogICAgICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgICAgICBtZXRhZGF0YV9wb29sX2V4
            aXN0ID0gVHJ1ZQogICAgICAgICAgICAgICAgIyBpZiBlcmFzdXJlX2NvZGVfcHJvZmlsZSBpcyBu
            b3QgZW1wdHkgYW5kIHBvb2wgbmFtZSBleGlzdHMgdGhlbiBpdCBpcyBlYyBwb29sCiAgICAgICAg
            ICAgICAgICBpZiBrZXlbImVyYXN1cmVfY29kZV9wcm9maWxlIl0gYW5kIGtleVsicG9vbF9uYW1l
            Il0gPT0gcmJkX3Bvb2xfbmFtZToKICAgICAgICAgICAgICAgICAgICBwb29sX2V4aXN0ID0gVHJ1
            ZQoKICAgICAgICAgICAgaWYgbm90IG1ldGFkYXRhX3Bvb2xfZXhpc3Q6CiAgICAgICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJQ
            cm92aWRlZCByYmRfZWNfbWV0YWRhdGFfcG9vbCBuYW1lLCIKICAgICAgICAgICAgICAgICAgICBm
            IiB7cmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZX0sIGRvZXMgbm90IGV4aXN0IgogICAgICAgICAg
            ICAgICAgKQogICAgICAgICAgICBpZiBub3QgcG9vbF9leGlzdDoKICAgICAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgZiJQcm92
            aWRlZCByYmRfZGF0YV9wb29sIG5hbWUsIHtyYmRfcG9vbF9uYW1lfSwgZG9lcyBub3QgZXhpc3Qi
            CiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIHJldHVybiByYmRfbWV0YWRhdGFfZWNfcG9v
            bF9uYW1lCgogICAgZGVmIGRyeV9ydW4oc2VsZiwgbXNnKToKICAgICAgICBpZiBzZWxmLl9hcmdf
            cGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHByaW50KCJFeGVjdXRlOiAiICsgIiciICsgbXNn
            ICsgIiciKQoKICAgIGRlZiB2YWxpZGF0ZV9yZ3dfZW5kcG9pbnRfdGxzX2NlcnQoc2VsZik6CiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZ3dfdGxzX2NlcnRfcGF0aDoKICAgICAgICAgICAg
            d2l0aCBvcGVuKHNlbGYuX2FyZ19wYXJzZXIucmd3X3Rsc19jZXJ0X3BhdGgsIGVuY29kaW5nPSJ1
            dGY4IikgYXMgZjoKICAgICAgICAgICAgICAgIGNvbnRlbnRzID0gZi5yZWFkKCkKICAgICAgICAg
            ICAgICAgIHJldHVybiBjb250ZW50cy5yc3RyaXAoKQoKICAgIGRlZiBfY2hlY2tfY29uZmxpY3Rp
            bmdfb3B0aW9ucyhzZWxmKToKICAgICAgICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNlci51cGdyYWRl
            IGFuZCBub3Qgc2VsZi5fYXJnX3BhcnNlci5yYmRfZGF0YV9wb29sX25hbWU6CiAgICAgICAgICAg
            IHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAiRWl0aGVy
            ICctLXVwZ3JhZGUnIG9yICctLXJiZC1kYXRhLXBvb2wtbmFtZSA8cG9vbF9uYW1lPicgc2hvdWxk
            IGJlIHNwZWNpZmllZCIKICAgICAgICAgICAgKQoKICAgIGRlZiBfaW52YWxpZF9lbmRwb2ludChz
            ZWxmLCBlbmRwb2ludF9zdHIpOgogICAgICAgICMgZXh0cmFjdCB0aGUgcG9ydCBieSBnZXR0aW5n
            IHRoZSBsYXN0IHNwbGl0IG9uIGA6YCBkZWxpbWl0ZXIKICAgICAgICB0cnk6CiAgICAgICAgICAg
            IGVuZHBvaW50X3N0cl9pcCwgcG9ydCA9IGVuZHBvaW50X3N0ci5yc3BsaXQoIjoiLCAxKQogICAg
            ICAgIGV4Y2VwdCBWYWx1ZUVycm9yOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJl
            RXhjZXB0aW9uKGYiTm90IGEgcHJvcGVyIGVuZHBvaW50OiB7ZW5kcG9pbnRfc3RyfSIpCgogICAg
            ICAgIHRyeToKICAgICAgICAgICAgaWYgZW5kcG9pbnRfc3RyX2lwWzBdID09ICJbIjoKICAgICAg
            ICAgICAgICAgIGVuZHBvaW50X3N0cl9pcCA9IGVuZHBvaW50X3N0cl9pcFsxIDogbGVuKGVuZHBv
            aW50X3N0cl9pcCkgLSAxXQogICAgICAgICAgICBpcF90eXBlID0gKAogICAgICAgICAgICAgICAg
            IklQdjQiIGlmIHR5cGUoaXBfYWRkcmVzcyhlbmRwb2ludF9zdHJfaXApKSBpcyBJUHY0QWRkcmVz
            cyBlbHNlICJJUHY2IgogICAgICAgICAgICApCiAgICAgICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAg
            ICAgICAgICAgIGlwX3R5cGUgPSAiRlFETiIKICAgICAgICBpZiBub3QgcG9ydC5pc2RpZ2l0KCk6
            CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oZiJQb3J0IG5vdCB2
            YWxpZDoge3BvcnR9IikKICAgICAgICBpbnRQb3J0ID0gaW50KHBvcnQpCiAgICAgICAgaWYgaW50
            UG9ydCA8IDEgb3IgaW50UG9ydCA+IDIqKjE2IC0gMToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0
            aW9uRmFpbHVyZUV4Y2VwdGlvbihmIk91dCBvZiByYW5nZSBwb3J0IG51bWJlcjoge3BvcnR9IikK
            CiAgICAgICAgcmV0dXJuIGlwX3R5cGUKCiAgICBkZWYgZW5kcG9pbnRfZGlhbChzZWxmLCBlbmRw
            b2ludF9zdHIsIGlwX3R5cGUsIHRpbWVvdXQ9MywgY2VydD1Ob25lKToKICAgICAgICAjIGlmIHRo
            ZSAnY2x1c3RlcicgaW5zdGFuY2UgaXMgYSBkdW1teSBvbmUsCiAgICAgICAgIyBkb24ndCB0cnkg
            dG8gcmVhY2ggb3V0IHRvIHRoZSBlbmRwb2ludAogICAgICAgIGlmIGlzaW5zdGFuY2Uoc2VsZi5j
            bHVzdGVyLCBEdW1teVJhZG9zKToKICAgICAgICAgICAgcmV0dXJuICIiLCAiIiwgIiIKICAgICAg
            ICBpZiBpcF90eXBlID09ICJJUHY2IjoKICAgICAgICAgICAgdHJ5OgogICAgICAgICAgICAgICAg
            ZW5kcG9pbnRfc3RyX2lwLCBlbmRwb2ludF9zdHJfcG9ydCA9IGVuZHBvaW50X3N0ci5yc3BsaXQo
            IjoiLCAxKQogICAgICAgICAgICBleGNlcHQgVmFsdWVFcnJvcjoKICAgICAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgZiJOb3Qg
            YSBwcm9wZXIgZW5kcG9pbnQ6IHtlbmRwb2ludF9zdHJ9IgogICAgICAgICAgICAgICAgKQogICAg
            ICAgICAgICBpZiBlbmRwb2ludF9zdHJfaXBbMF0gIT0gIlsiOgogICAgICAgICAgICAgICAgZW5k
            cG9pbnRfc3RyX2lwID0gIlsiICsgZW5kcG9pbnRfc3RyX2lwICsgIl0iCiAgICAgICAgICAgIGVu
            ZHBvaW50X3N0ciA9ICI6Ii5qb2luKFtlbmRwb2ludF9zdHJfaXAsIGVuZHBvaW50X3N0cl9wb3J0
            XSkKCiAgICAgICAgcHJvdG9jb2xzID0gWyJodHRwIiwgImh0dHBzIl0KICAgICAgICByZXNwb25z
            ZV9lcnJvciA9IE5vbmUKICAgICAgICBmb3IgcHJlZml4IGluIHByb3RvY29sczoKICAgICAgICAg
            ICAgdHJ5OgogICAgICAgICAgICAgICAgZXAgPSBmIntwcmVmaXh9Oi8ve2VuZHBvaW50X3N0cn0i
            CiAgICAgICAgICAgICAgICB2ZXJpZnkgPSBOb25lCiAgICAgICAgICAgICAgICAjIElmIHZlcmlm
            eSBpcyBzZXQgdG8gYSBwYXRoIHRvIGEgZGlyZWN0b3J5LAogICAgICAgICAgICAgICAgIyB0aGUg
            ZGlyZWN0b3J5IG11c3QgaGF2ZSBiZWVuIHByb2Nlc3NlZCB1c2luZyB0aGUgY19yZWhhc2ggdXRp
            bGl0eSBzdXBwbGllZCB3aXRoIE9wZW5TU0wuCiAgICAgICAgICAgICAgICBpZiBwcmVmaXggPT0g
            Imh0dHBzIiBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfc2tpcF90bHM6CiAgICAgICAgICAgICAg
            ICAgICAgdmVyaWZ5ID0gRmFsc2UKICAgICAgICAgICAgICAgICAgICByID0gcmVxdWVzdHMuaGVh
            ZChlcCwgdGltZW91dD10aW1lb3V0LCB2ZXJpZnk9RmFsc2UpCiAgICAgICAgICAgICAgICBlbGlm
            IHByZWZpeCA9PSAiaHR0cHMiIGFuZCBjZXJ0OgogICAgICAgICAgICAgICAgICAgIHZlcmlmeSA9
            IGNlcnQKICAgICAgICAgICAgICAgICAgICByID0gcmVxdWVzdHMuaGVhZChlcCwgdGltZW91dD10
            aW1lb3V0LCB2ZXJpZnk9Y2VydCkKICAgICAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAg
            ICAgICAgciA9IHJlcXVlc3RzLmhlYWQoZXAsIHRpbWVvdXQ9dGltZW91dCkKICAgICAgICAgICAg
            ICAgIGlmIHIuc3RhdHVzX2NvZGUgPT0gMjAwOgogICAgICAgICAgICAgICAgICAgIHJldHVybiBw
            cmVmaXgsIHZlcmlmeSwgIiIKICAgICAgICAgICAgZXhjZXB0IEV4Y2VwdGlvbiBhcyBlcnI6CiAg
            ICAgICAgICAgICAgICByZXNwb25zZV9lcnJvciA9IGVycgogICAgICAgICAgICAgICAgY29udGlu
            dWUKICAgICAgICBzeXMuc3RkZXJyLndyaXRlKAogICAgICAgICAgICBmInVuYWJsZSB0byBjb25u
            ZWN0IHRvIGVuZHBvaW50OiB7ZW5kcG9pbnRfc3RyfSwgZmFpbGVkIGVycm9yOiB7cmVzcG9uc2Vf
            ZXJyb3J9IgogICAgICAgICkKICAgICAgICByZXR1cm4gKAogICAgICAgICAgICAiIiwKICAgICAg
            ICAgICAgIiIsCiAgICAgICAgICAgICgiLTEiKSwKICAgICAgICApCgogICAgZGVmIF9faW5pdF9f
            KHNlbGYsIGFyZ19saXN0PU5vbmUpOgogICAgICAgIHNlbGYub3V0X21hcCA9IHt9CiAgICAgICAg
            c2VsZi5fZXhjbHVkZWRfa2V5cyA9IHNldCgpCiAgICAgICAgc2VsZi5fYXJnX3BhcnNlciA9IHNl
            bGYuZ2VuX2FyZ19wYXJzZXIoYXJnc190b19wYXJzZT1hcmdfbGlzdCkKICAgICAgICBzZWxmLl9j
            aGVja19jb25mbGljdGluZ19vcHRpb25zKCkKICAgICAgICBzZWxmLnJ1bl9hc191c2VyID0gc2Vs
            Zi5fYXJnX3BhcnNlci5ydW5fYXNfdXNlcgogICAgICAgIHNlbGYub3V0cHV0X2ZpbGUgPSBzZWxm
            Ll9hcmdfcGFyc2VyLm91dHB1dAogICAgICAgIHNlbGYuY2VwaF9jb25mID0gc2VsZi5fYXJnX3Bh
            cnNlci5jZXBoX2NvbmYKICAgICAgICBzZWxmLmNlcGhfa2V5cmluZyA9IHNlbGYuX2FyZ19wYXJz
            ZXIua2V5cmluZwogICAgICAgICMgaWYgdXNlciBub3QgcHJvdmlkZWQsIGdpdmUgYSBkZWZhdWx0
            IHVzZXIKICAgICAgICBpZiBub3Qgc2VsZi5ydW5fYXNfdXNlciBhbmQgbm90IHNlbGYuX2FyZ19w
            YXJzZXIudXBncmFkZToKICAgICAgICAgICAgc2VsZi5ydW5fYXNfdXNlciA9IHNlbGYuRVhURVJO
            QUxfVVNFUl9OQU1FCiAgICAgICAgaWYgbm90IHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJl
            Zml4IGFuZCBub3Qgc2VsZi5fYXJnX3BhcnNlci51cGdyYWRlOgogICAgICAgICAgICBzZWxmLl9h
            cmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeCA9IHNlbGYuREVGQVVMVF9SR1dfUE9PTF9QUkVGSVgK
            ICAgICAgICBpZiBzZWxmLmNlcGhfY29uZjoKICAgICAgICAgICAga3dhcmdzID0ge30KICAgICAg
            ICAgICAgaWYgc2VsZi5jZXBoX2tleXJpbmc6CiAgICAgICAgICAgICAgICBrd2FyZ3NbImNvbmYi
            XSA9IHsia2V5cmluZyI6IHNlbGYuY2VwaF9rZXlyaW5nfQogICAgICAgICAgICBzZWxmLmNsdXN0
            ZXIgPSByYWRvcy5SYWRvcyhjb25mZmlsZT1zZWxmLmNlcGhfY29uZiwgKiprd2FyZ3MpCiAgICAg
            ICAgZWxzZToKICAgICAgICAgICAgc2VsZi5jbHVzdGVyID0gcmFkb3MuUmFkb3MoKQogICAgICAg
            ICAgICBzZWxmLmNsdXN0ZXIuY29uZl9yZWFkX2ZpbGUoKQogICAgICAgIHNlbGYuY2x1c3Rlci5j
            b25uZWN0KCkKCiAgICBkZWYgc2h1dGRvd24oc2VsZik6CiAgICAgICAgaWYgc2VsZi5jbHVzdGVy
            LnN0YXRlID09ICJjb25uZWN0ZWQiOgogICAgICAgICAgICBzZWxmLmNsdXN0ZXIuc2h1dGRvd24o
            KQoKICAgIGRlZiBnZXRfZnNpZChzZWxmKToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRy
            eV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGggZnNpZCIpCiAgICAg
            ICAgcmV0dXJuIHN0cihzZWxmLmNsdXN0ZXIuZ2V0X2ZzaWQoKSkKCiAgICBkZWYgX2NvbW1vbl9j
            bWRfanNvbl9nZW4oc2VsZiwgY21kX2pzb24pOgogICAgICAgIGNtZCA9IGpzb24uZHVtcHMoY21k
            X2pzb24sIHNvcnRfa2V5cz1UcnVlKQogICAgICAgIHJldF92YWwsIGNtZF9vdXQsIGVycl9tc2cg
            PSBzZWxmLmNsdXN0ZXIubW9uX2NvbW1hbmQoY21kLCBiIiIpCiAgICAgICAgaWYgc2VsZi5fYXJn
            X3BhcnNlci52ZXJib3NlOgogICAgICAgICAgICBwcmludChmIkNvbW1hbmQgSW5wdXQ6IHtjbWR9
            IikKICAgICAgICAgICAgcHJpbnQoCiAgICAgICAgICAgICAgICBmIlJldHVybiBWYWw6IHtyZXRf
            dmFsfVxuQ29tbWFuZCBPdXRwdXQ6IHtjbWRfb3V0fVxuIgogICAgICAgICAgICAgICAgZiJFcnJv
            ciBNZXNzYWdlOiB7ZXJyX21zZ31cbi0tLS0tLS0tLS1cbiIKICAgICAgICAgICAgKQogICAgICAg
            IGpzb25fb3V0ID0ge30KICAgICAgICAjIGlmIHRoZXJlIGlzIG5vIGVycm9yIChpLmU7IHJldF92
            YWwgaXMgWkVSTykgYW5kICdjbWRfb3V0JyBpcyBub3QgZW1wdHkKICAgICAgICAjIHRoZW4gY29u
            dmVydCAnY21kX291dCcgdG8gYSBqc29uIG91dHB1dAogICAgICAgIGlmIHJldF92YWwgPT0gMCBh
            bmQgY21kX291dDoKICAgICAgICAgICAganNvbl9vdXQgPSBqc29uLmxvYWRzKGNtZF9vdXQpCiAg
            ICAgICAgcmV0dXJuIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnCgogICAgZGVmIGdldF9jZXBo
            X2V4dGVybmFsX21vbl9kYXRhKHNlbGYpOgogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAi
            cXVvcnVtX3N0YXR1cyIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgaWYgc2VsZi5fYXJnX3Bh
            cnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKCJjZXBoICIgKyBj
            bWRfanNvblsicHJlZml4Il0pCiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBz
            ZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICMgaWYgdGhlcmUgaXMg
            YW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihq
            c29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgICAgICIncXVvcnVtX3N0YXR1cycgY29tbWFuZCBmYWlsZWQuXG4iCiAg
            ICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxm
            LkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgKQogICAgICAgIHFfbGVhZGVyX25hbWUg
            PSBqc29uX291dFsicXVvcnVtX2xlYWRlcl9uYW1lIl0KICAgICAgICBxX2xlYWRlcl9kZXRhaWxz
            ID0ge30KICAgICAgICBxX2xlYWRlcl9tYXRjaGluZ19saXN0ID0gWwogICAgICAgICAgICBsIGZv
            ciBsIGluIGpzb25fb3V0WyJtb25tYXAiXVsibW9ucyJdIGlmIGxbIm5hbWUiXSA9PSBxX2xlYWRl
            cl9uYW1lCiAgICAgICAgXQogICAgICAgIGlmIGxlbihxX2xlYWRlcl9tYXRjaGluZ19saXN0KSA9
            PSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKCJObyBtYXRj
            aGluZyAnbW9uJyBkZXRhaWxzIGZvdW5kIikKICAgICAgICBxX2xlYWRlcl9kZXRhaWxzID0gcV9s
            ZWFkZXJfbWF0Y2hpbmdfbGlzdFswXQogICAgICAgICMgZ2V0IHRoZSBhZGRyZXNzIHZlY3RvciBv
            ZiB0aGUgcXVvcnVtLWxlYWRlcgogICAgICAgIHFfbGVhZGVyX2FkZHJ2ZWMgPSBxX2xlYWRlcl9k
            ZXRhaWxzLmdldCgicHVibGljX2FkZHJzIiwge30pLmdldCgiYWRkcnZlYyIsIFtdKQogICAgICAg
            ICMgaWYgdGhlIHF1b3J1bS1sZWFkZXIgaGFzIG9ubHkgb25lIGFkZHJlc3MgaW4gdGhlIGFkZHJl
            c3MtdmVjdG9yCiAgICAgICAgIyBhbmQgaXQgaXMgb2YgdHlwZSAndjInIChpZTsgd2l0aCA8SVA+
            OjMzMDApLAogICAgICAgICMgcmFpc2UgYW4gZXhjZXB0aW9uIHRvIG1ha2UgdXNlciBhd2FyZSB0
            aGF0CiAgICAgICAgIyB0aGV5IGhhdmUgdG8gZW5hYmxlICd2MScgKGllOyB3aXRoIDxJUD46Njc4
            OSkgdHlwZSBhcyB3ZWxsCiAgICAgICAgaWYgbGVuKHFfbGVhZGVyX2FkZHJ2ZWMpID09IDEgYW5k
            IHFfbGVhZGVyX2FkZHJ2ZWNbMF1bInR5cGUiXSA9PSAidjIiOgogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgIk9ubHkgJ3YyJyBhZGRy
            ZXNzIHR5cGUgaXMgZW5hYmxlZCwgdXNlciBzaG91bGQgYWxzbyBlbmFibGUgJ3YxJyB0eXBlIGFz
            IHdlbGwiCiAgICAgICAgICAgICkKICAgICAgICBpcF9hZGRyID0gc3RyKHFfbGVhZGVyX2RldGFp
            bHNbInB1YmxpY19hZGRyIl0uc3BsaXQoIi8iKVswXSkKCiAgICAgICAgaWYgc2VsZi5fYXJnX3Bh
            cnNlci52Ml9wb3J0X2VuYWJsZToKICAgICAgICAgICAgaWYgbGVuKHFfbGVhZGVyX2FkZHJ2ZWMp
            ID4gMToKICAgICAgICAgICAgICAgIGlmIHFfbGVhZGVyX2FkZHJ2ZWNbMF1bInR5cGUiXSA9PSAi
            djIiOgogICAgICAgICAgICAgICAgICAgIGlwX2FkZHIgPSBxX2xlYWRlcl9hZGRydmVjWzBdWyJh
            ZGRyIl0KICAgICAgICAgICAgICAgIGVsaWYgcV9sZWFkZXJfYWRkcnZlY1sxXVsidHlwZSJdID09
            ICJ2MiI6CiAgICAgICAgICAgICAgICAgICAgaXBfYWRkciA9IHFfbGVhZGVyX2FkZHJ2ZWNbMV1b
            ImFkZHIiXQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0
            ZSgKICAgICAgICAgICAgICAgICAgICAiJ3YyJyBhZGRyZXNzIHR5cGUgbm90IHByZXNlbnQsIGFu
            ZCAndjItcG9ydC1lbmFibGUnIGZsYWcgaXMgcHJvdmlkZWQiCiAgICAgICAgICAgICAgICApCgog
            ICAgICAgIHJldHVybiBmIntzdHIocV9sZWFkZXJfbmFtZSl9PXtpcF9hZGRyfSIKCiAgICBkZWYg
            X2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoc2VsZiwgaG9zdF9uYW1lLCBwb3J0LCBpcF90eXBlKToK
            ICAgICAgICAjIGlmICdjbHVzdGVyJyBpbnN0YW5jZSBpcyBhIGR1bW15IHR5cGUsCiAgICAgICAg
            IyBjYWxsIHRoZSBkdW1teSBpbnN0YW5jZSdzICJjb252ZXJ0IiBtZXRob2QKICAgICAgICBpZiBu
            b3QgaG9zdF9uYW1lOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KCJFbXB0eSBob3N0bmFtZSBwcm92aWRlZCIpCiAgICAgICAgaWYgaXNpbnN0YW5jZShzZWxmLmNs
            dXN0ZXIsIER1bW15UmFkb3MpOgogICAgICAgICAgICByZXR1cm4gc2VsZi5jbHVzdGVyLl9jb252
            ZXJ0X2hvc3RuYW1lX3RvX2lwKGhvc3RfbmFtZSkKCiAgICAgICAgaWYgaXBfdHlwZSA9PSAiRlFE
            TiI6CiAgICAgICAgICAgICMgY2hlY2sgd2hpY2ggaXAgRlFETiBzaG91bGQgYmUgY29udmVydGVk
            IHRvLCBJUHY0IG9yIElQdjYKICAgICAgICAgICAgIyBjaGVjayB0aGUgaG9zdCBpcCwgdGhlIGVu
            ZHBvaW50IGlwIHR5cGUgd291bGQgYmUgc2ltaWxhciB0byBob3N0IGlwCiAgICAgICAgICAgIGNt
            ZF9qc29uID0geyJwcmVmaXgiOiAib3JjaCBob3N0IGxzIiwgImZvcm1hdCI6ICJqc29uIn0KICAg
            ICAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pz
            b25fZ2VuKGNtZF9qc29uKQogICAgICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1
            bCBhdHRlbXB0LAogICAgICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9
            PSAwOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAg
            ICAgICAgICAgICAgICAgICAiJ29yY2ggaG9zdCBscycgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAg
            ICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYgcmV0X3ZhbCAhPSAwIGVsc2Ugc2Vs
            Zi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGhvc3Rf
            YWRkciA9IGpzb25fb3V0WzBdWyJhZGRyIl0KICAgICAgICAgICAgIyBhZGQgOjgwIHNhbXBsZSBw
            b3J0IGluIGlwX3R5cGUsIGFzIF9pbnZhbGlkX2VuZHBvaW50IGFsc28gdmVyaWZ5IHBvcnQKICAg
            ICAgICAgICAgaG9zdF9pcF90eXBlID0gc2VsZi5faW52YWxpZF9lbmRwb2ludChob3N0X2FkZHIg
            KyAiOjgwIikKICAgICAgICAgICAgaW1wb3J0IHNvY2tldAoKICAgICAgICAgICAgIyBleGFtcGxl
            IG91dHB1dCBbKDxBZGRyZXNzRmFtaWx5LkFGX0lORVQ6IDI+LCA8U29ja2V0S2luZC5TT0NLX1NU
            UkVBTTogMT4sIDYsICcnLCAoJzkzLjE4NC4yMTYuMzQnLCA4MCkpLCAuLi5dCiAgICAgICAgICAg
            ICMgd2UgbmVlZCB0byBnZXQgOTMuMTg0LjIxNi4zNCBzbyBpdCB3b3VsZCBiZSBpcFswXVs0XVsw
            XQogICAgICAgICAgICBpZiBob3N0X2lwX3R5cGUgPT0gIklQdjYiOgogICAgICAgICAgICAgICAg
            aXAgPSBzb2NrZXQuZ2V0YWRkcmluZm8oCiAgICAgICAgICAgICAgICAgICAgaG9zdF9uYW1lLCBw
            b3J0LCBmYW1pbHk9c29ja2V0LkFGX0lORVQ2LCBwcm90bz1zb2NrZXQuSVBQUk9UT19UQ1AKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxpZiBob3N0X2lwX3R5cGUgPT0gIklQdjQiOgog
            ICAgICAgICAgICAgICAgaXAgPSBzb2NrZXQuZ2V0YWRkcmluZm8oCiAgICAgICAgICAgICAgICAg
            ICAgaG9zdF9uYW1lLCBwb3J0LCBmYW1pbHk9c29ja2V0LkFGX0lORVQsIHByb3RvPXNvY2tldC5J
            UFBST1RPX1RDUAogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBkZWwgc29ja2V0CiAgICAg
            ICAgICAgIHJldHVybiBpcFswXVs0XVswXQogICAgICAgIHJldHVybiBob3N0X25hbWUKCiAgICBk
            ZWYgZ2V0X2FjdGl2ZV9hbmRfc3RhbmRieV9tZ3JzKHNlbGYpOgogICAgICAgIGlmIHNlbGYuX2Fy
            Z19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuICIiLCBzZWxmLmRyeV9ydW4oImNl
            cGggc3RhdHVzIikKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQgPSBzZWxmLl9hcmdf
            cGFyc2VyLm1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydAogICAgICAgIG1vbml0b3JpbmdfZW5kcG9p
            bnRfaXBfbGlzdCA9IHNlbGYuX2FyZ19wYXJzZXIubW9uaXRvcmluZ19lbmRwb2ludAogICAgICAg
            IHN0YW5kYnlfbWdycyA9IFtdCiAgICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBf
            bGlzdDoKICAgICAgICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJzdGF0dXMiLCAiZm9ybWF0
            IjogImpzb24ifQogICAgICAgICAgICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYu
            X2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgICAgICMgaWYgdGhlcmUgaXMg
            YW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBs
            ZW4oanNvbl9vdXQpID09IDA6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJl
            RXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICInbWdyIHNlcnZpY2VzJyBjb21tYW5kIGZh
            aWxlZC5cbiIKICAgICAgICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFs
            ICE9IDAgZWxzZSBzZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgICAgICkKICAg
            ICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludCA9ICgKICAgICAgICAgICAgICAgIGpzb25fb3V0
            LmdldCgibWdybWFwIiwge30pLmdldCgic2VydmljZXMiLCB7fSkuZ2V0KCJwcm9tZXRoZXVzIiwg
            IiIpCiAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnQ6
            CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAg
            ICAgICAgICAgICAgICJjYW4ndCBmaW5kIG1vbml0b3JpbmdfZW5kcG9pbnQsIHByb21ldGhldXMg
            bW9kdWxlIG1pZ2h0IG5vdCBiZSBlbmFibGVkLCAiCiAgICAgICAgICAgICAgICAgICAgImVuYWJs
            ZSB0aGUgbW9kdWxlIGJ5IHJ1bm5pbmcgJ2NlcGggbWdyIG1vZHVsZSBlbmFibGUgcHJvbWV0aGV1
            cyciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICMgbm93IGNoZWNrIHRoZSBzdGFuZC1i
            eSBtZ3ItcwogICAgICAgICAgICBzdGFuZGJ5X2FyciA9IGpzb25fb3V0LmdldCgibWdybWFwIiwg
            e30pLmdldCgic3RhbmRieXMiLCBbXSkKICAgICAgICAgICAgZm9yIGVhY2hfc3RhbmRieSBpbiBz
            dGFuZGJ5X2FycjoKICAgICAgICAgICAgICAgIGlmICJuYW1lIiBpbiBlYWNoX3N0YW5kYnkua2V5
            cygpOgogICAgICAgICAgICAgICAgICAgIHN0YW5kYnlfbWdycy5hcHBlbmQoZWFjaF9zdGFuZGJ5
            WyJuYW1lIl0pCiAgICAgICAgICAgIHRyeToKICAgICAgICAgICAgICAgIHBhcnNlZF9lbmRwb2lu
            dCA9IHVybHBhcnNlKG1vbml0b3JpbmdfZW5kcG9pbnQpCiAgICAgICAgICAgIGV4Y2VwdCBWYWx1
            ZUVycm9yOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigK
            ICAgICAgICAgICAgICAgICAgICBmImludmFsaWQgZW5kcG9pbnQ6IHttb25pdG9yaW5nX2VuZHBv
            aW50fSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9p
            cF9saXN0ID0gcGFyc2VkX2VuZHBvaW50Lmhvc3RuYW1lCiAgICAgICAgICAgIGlmIG5vdCBtb25p
            dG9yaW5nX2VuZHBvaW50X3BvcnQ6CiAgICAgICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50
            X3BvcnQgPSBzdHIocGFyc2VkX2VuZHBvaW50LnBvcnQpCgogICAgICAgICMgaWYgbW9uaXRvcmlu
            ZyBlbmRwb2ludCBwb3J0IGlzIG5vdCBzZXQsIHB1dCBhIGRlZmF1bHQgbW9uIHBvcnQKICAgICAg
            ICBpZiBub3QgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0OgogICAgICAgICAgICBtb25pdG9yaW5n
            X2VuZHBvaW50X3BvcnQgPSBzZWxmLkRFRkFVTFRfTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUCgog
            ICAgICAgICMgdXNlciBjb3VsZCBnaXZlIGNvbW1hIGFuZCBzcGFjZSBzZXBhcmF0ZWQgaW5wdXRz
            IChsaWtlIC0tbW9uaXRvcmluZy1lbmRwb2ludD0iPGlwMT4sIDxpcDI+IikKICAgICAgICBtb25p
            dG9yaW5nX2VuZHBvaW50X2lwX2xpc3QgPSBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3QucmVw
            bGFjZSgiLCIsICIgIikKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXQg
            PSBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Quc3BsaXQoKQogICAgICAgICMgaWYgbW9uaXRv
            cmluZy1lbmRwb2ludCBjb3VsZCBub3QgYmUgZm91bmQsIHJhaXNlIGFuIGVycm9yCiAgICAgICAg
            aWYgbGVuKG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdF9zcGxpdCkgPT0gMDoKICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigiTm8gJ21vbml0b3JpbmctZW5kcG9p
            bnQnIGZvdW5kIikKICAgICAgICAjIGZpcnN0IGlwIGlzIHRyZWF0ZWQgYXMgdGhlIG1haW4gbW9u
            aXRvcmluZy1lbmRwb2ludAogICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfaXAgPSBtb25pdG9y
            aW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXRbMF0KICAgICAgICAjIHJlc3Qgb2YgdGhlIGlwLXMg
            YXJlIGFkZGVkIHRvIHRoZSAnc3RhbmRieV9tZ3JzJyBsaXN0CiAgICAgICAgc3RhbmRieV9tZ3Jz
            LmV4dGVuZChtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXRbMTpdKQogICAgICAgIGZh
            aWxlZF9pcCA9IG1vbml0b3JpbmdfZW5kcG9pbnRfaXAKCiAgICAgICAgbW9uaXRvcmluZ19lbmRw
            b2ludCA9ICI6Ii5qb2luKAogICAgICAgICAgICBbbW9uaXRvcmluZ19lbmRwb2ludF9pcCwgbW9u
            aXRvcmluZ19lbmRwb2ludF9wb3J0XQogICAgICAgICkKICAgICAgICBpcF90eXBlID0gc2VsZi5f
            aW52YWxpZF9lbmRwb2ludChtb25pdG9yaW5nX2VuZHBvaW50KQogICAgICAgIHRyeToKICAgICAg
            ICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcCA9IHNlbGYuX2NvbnZlcnRfaG9zdG5hbWVfdG9f
            aXAoCiAgICAgICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X2lwLCBtb25pdG9yaW5nX2Vu
            ZHBvaW50X3BvcnQsIGlwX3R5cGUKICAgICAgICAgICAgKQogICAgICAgICAgICAjIGNvbGxlY3Qg
            YWxsIHRoZSAnc3RhbmQtYnknIG1nciBpcHMKICAgICAgICAgICAgbWdyX2lwcyA9IFtdCiAgICAg
            ICAgICAgIGZvciBlYWNoX3N0YW5kYnlfbWdyIGluIHN0YW5kYnlfbWdyczoKICAgICAgICAgICAg
            ICAgIGZhaWxlZF9pcCA9IGVhY2hfc3RhbmRieV9tZ3IKICAgICAgICAgICAgICAgIG1ncl9pcHMu
            YXBwZW5kKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAo
            CiAgICAgICAgICAgICAgICAgICAgICAgIGVhY2hfc3RhbmRieV9tZ3IsIG1vbml0b3JpbmdfZW5k
            cG9pbnRfcG9ydCwgaXBfdHlwZQogICAgICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAg
            ICkKICAgICAgICBleGNlcHQ6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNl
            cHRpb24oCiAgICAgICAgICAgICAgICBmIkNvbnZlcnNpb24gb2YgaG9zdDoge2ZhaWxlZF9pcH0g
            dG8gSVAgZmFpbGVkLiAiCiAgICAgICAgICAgICAgICAiUGxlYXNlIGVudGVyIHRoZSBJUCBhZGRy
            ZXNzZXMgb2YgYWxsIHRoZSBjZXBoLW1ncnMgd2l0aCB0aGUgJy0tbW9uaXRvcmluZy1lbmRwb2lu
            dCcgZmxhZyIKICAgICAgICAgICAgKQoKICAgICAgICBfLCBfLCBlcnIgPSBzZWxmLmVuZHBvaW50
            X2RpYWwobW9uaXRvcmluZ19lbmRwb2ludCwgaXBfdHlwZSkKICAgICAgICBpZiBlcnIgPT0gIi0x
            IjoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihlcnIpCiAgICAg
            ICAgIyBhZGQgdGhlIHZhbGlkYXRlZCBhY3RpdmUgbWdyIElQIGludG8gdGhlIGZpcnN0IGluZGV4
            CiAgICAgICAgbWdyX2lwcy5pbnNlcnQoMCwgbW9uaXRvcmluZ19lbmRwb2ludF9pcCkKICAgICAg
            ICBhbGxfbWdyX2lwc19zdHIgPSAiLCIuam9pbihtZ3JfaXBzKQogICAgICAgIHJldHVybiBhbGxf
            bWdyX2lwc19zdHIsIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydAoKICAgIGRlZiBjaGVja191c2Vy
            X2V4aXN0KHNlbGYsIHVzZXIpOgogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAiYXV0aCBn
            ZXQiLCAiZW50aXR5IjogZiJ7dXNlcn0iLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIHJldF92
            YWwsIGpzb25fb3V0LCBfID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAg
            ICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICBy
            ZXR1cm4gIiIKICAgICAgICByZXR1cm4gc3RyKGpzb25fb3V0WzBdWyJrZXkiXSkKCiAgICBkZWYg
            Z2V0X2NlcGhmc19wcm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50
            aXR5ID0gImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIgogICAgICAgIGNhcHMgPSB7CiAg
            ICAgICAgICAgICJtb24iOiAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCci
            LAogICAgICAgICAgICAibWdyIjogImFsbG93IHJ3IiwKICAgICAgICAgICAgIm9zZCI6ICJhbGxv
            dyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiLAogICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9h
            cmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uOgogICAgICAgICAgICBrOHNfY2x1
            c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1lCiAgICAgICAgICAg
            IGlmIGs4c19jbHVzdGVyX25hbWUgPT0gIiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRp
            b25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJrOHNfY2x1c3Rlcl9uYW1l
            IG5vdCBmb3VuZCwgcGxlYXNlIHNldCB0aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW0gPSBzZWxmLl9hcmdf
            cGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICAgICAgaWYgY2VwaGZzX2ZpbGVz
            eXN0ZW0gPT0gIiI6CiAgICAgICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1
            c3Rlcl9uYW1lfSIKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYi
            e2VudGl0eX0te2s4c19jbHVzdGVyX25hbWV9LXtjZXBoZnNfZmlsZXN5c3RlbX0iCiAgICAgICAg
            ICAgICAgICBjYXBzWyJvc2QiXSA9IGYiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT17Y2Vw
            aGZzX2ZpbGVzeXN0ZW19IgoKICAgICAgICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVmIGdl
            dF9jZXBoZnNfbm9kZV9jYXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gImNs
            aWVudC5jc2ktY2VwaGZzLW5vZGUiCiAgICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6
            ICJhbGxvdyByLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJt
            Z3IiOiAiYWxsb3cgcnciLAogICAgICAgICAgICAib3NkIjogImFsbG93IHJ3IHRhZyBjZXBoZnMg
            Kj0qIiwKICAgICAgICAgICAgIm1kcyI6ICJhbGxvdyBydyIsCiAgICAgICAgfQogICAgICAgIGlm
            IHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAg
            IGs4c19jbHVzdGVyX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAg
            ICAgICAgICAgaWYgazhzX2NsdXN0ZXJfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIms4c19jbHVz
            dGVyX25hbWUgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1rOHMtY2x1c3Rlci1uYW1lJyBm
            bGFnIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBjZXBoZnNfZmlsZXN5c3RlbSA9IHNl
            bGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgICAgICBpZiBjZXBo
            ZnNfZmlsZXN5c3RlbSA9PSAiIjoKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0t
            e2s4c19jbHVzdGVyX25hbWV9IgogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgZW50
            aXR5ID0gZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0te2NlcGhmc19maWxlc3lzdGVtfSIK
            ICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJdID0gZiJhbGxvdyBydyB0YWcgY2VwaGZzICo9e2Nl
            cGhmc19maWxlc3lzdGVtfSIKCiAgICAgICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBn
            ZXRfZW50aXR5KAogICAgICAgIHNlbGYsCiAgICAgICAgZW50aXR5LAogICAgICAgIHJiZF9wb29s
            X25hbWUsCiAgICAgICAgYWxpYXNfcmJkX3Bvb2xfbmFtZSwKICAgICAgICBrOHNfY2x1c3Rlcl9u
            YW1lLAogICAgICAgIHJhZG9zX25hbWVzcGFjZSwKICAgICk6CiAgICAgICAgaWYgKAogICAgICAg
            ICAgICByYmRfcG9vbF9uYW1lLmNvdW50KCIuIikgIT0gMAogICAgICAgICAgICBvciByYmRfcG9v
            bF9uYW1lLmNvdW50KCJfIikgIT0gMAogICAgICAgICAgICBvciBhbGlhc19yYmRfcG9vbF9uYW1l
            ICE9ICIiCiAgICAgICAgICAgICMgY2hlY2tpbmcgYWxpYXNfcmJkX3Bvb2xfbmFtZSBpcyBub3Qg
            ZW1wdHkgYXMgdGhlcmUgbWF5YmUgYSBzcGVjaWFsIGNoYXJhY3RlciB1c2VkIG90aGVyIHRoYW4g
            LiBvciBfCiAgICAgICAgKToKICAgICAgICAgICAgaWYgYWxpYXNfcmJkX3Bvb2xfbmFtZSA9PSAi
            IjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAg
            ICAgICAgICAgICAgICAgInBsZWFzZSBzZXQgdGhlICctLWFsaWFzLXJiZC1kYXRhLXBvb2wtbmFt
            ZScgZmxhZyBhcyB0aGUgcmJkIGRhdGEgcG9vbCBuYW1lIGNvbnRhaW5zICcuJyBvciAnXyciCiAg
            ICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmICgKICAgICAgICAgICAgICAgIGFsaWFzX3Ji
            ZF9wb29sX25hbWUuY291bnQoIi4iKSAhPSAwCiAgICAgICAgICAgICAgICBvciBhbGlhc19yYmRf
            cG9vbF9uYW1lLmNvdW50KCJfIikgIT0gMAogICAgICAgICAgICApOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJy0t
            YWxpYXMtcmJkLWRhdGEtcG9vbC1uYW1lJyBmbGFnIHZhbHVlIHNob3VsZCBub3QgY29udGFpbiAn
            Licgb3IgJ18nIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRp
            dHl9LXtrOHNfY2x1c3Rlcl9uYW1lfS17YWxpYXNfcmJkX3Bvb2xfbmFtZX0iCiAgICAgICAgZWxz
            ZToKICAgICAgICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0te3Ji
            ZF9wb29sX25hbWV9IgoKICAgICAgICBpZiByYWRvc19uYW1lc3BhY2U6CiAgICAgICAgICAgIGVu
            dGl0eSA9IGYie2VudGl0eX0te3JhZG9zX25hbWVzcGFjZX0iCiAgICAgICAgcmV0dXJuIGVudGl0
            eQoKICAgIGRlZiBnZXRfcmJkX3Byb3Zpc2lvbmVyX2NhcHNfYW5kX2VudGl0eShzZWxmKToKICAg
            ICAgICBlbnRpdHkgPSAiY2xpZW50LmNzaS1yYmQtcHJvdmlzaW9uZXIiCiAgICAgICAgY2FwcyA9
            IHsKICAgICAgICAgICAgIm1vbiI6ICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJs
            b2NrbGlzdCciLAogICAgICAgICAgICAibWdyIjogImFsbG93IHJ3IiwKICAgICAgICAgICAgIm9z
            ZCI6ICJwcm9maWxlIHJiZCIsCiAgICAgICAgfQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIu
            cmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBz
            ZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgICAgICBhbGlhc19yYmRf
            cG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5hbGlhc19yYmRfZGF0YV9wb29sX25hbWUKICAg
            ICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJf
            bmFtZQogICAgICAgICAgICByYWRvc19uYW1lc3BhY2UgPSBzZWxmLl9hcmdfcGFyc2VyLnJhZG9z
            X25hbWVzcGFjZQogICAgICAgICAgICBpZiByYmRfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAg
            ICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAg
            ICAibWFuZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1yYmQtZGF0YS1w
            b29sLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmIGs4c19jbHVz
            dGVyX25hbWUgPT0gIiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhj
            ZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBs
            ZWFzZSBzZXQgdGhlICctLWs4cy1jbHVzdGVyLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgIGVudGl0eSA9IHNlbGYuZ2V0X2VudGl0eSgKICAgICAgICAgICAgICAgIGVu
            dGl0eSwKICAgICAgICAgICAgICAgIHJiZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBhbGlh
            c19yYmRfcG9vbF9uYW1lLAogICAgICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAg
            ICAgICAgICAgIHJhZG9zX25hbWVzcGFjZSwKICAgICAgICAgICAgKQogICAgICAgICAgICBpZiBy
            YWRvc19uYW1lc3BhY2UgIT0gIiI6CiAgICAgICAgICAgICAgICBjYXBzWyJvc2QiXSA9ICgKICAg
            ICAgICAgICAgICAgICAgICBmInByb2ZpbGUgcmJkIHBvb2w9e3JiZF9wb29sX25hbWV9IG5hbWVz
            cGFjZT17cmFkb3NfbmFtZXNwYWNlfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxz
            ZToKICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJdID0gZiJwcm9maWxlIHJiZCBwb29sPXtyYmRf
            cG9vbF9uYW1lfSIKCiAgICAgICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfcmJk
            X25vZGVfY2Fwc19hbmRfZW50aXR5KHNlbGYpOgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3Np
            LXJiZC1ub2RlIgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAgICJtb24iOiAicHJvZmlsZSBy
            YmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwKICAgICAgICAgICAgIm9zZCI6ICJw
            cm9maWxlIHJiZCIsCiAgICAgICAgfQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJp
            Y3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9h
            cmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9u
            YW1lID0gc2VsZi5fYXJnX3BhcnNlci5hbGlhc19yYmRfZGF0YV9wb29sX25hbWUKICAgICAgICAg
            ICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQog
            ICAgICAgICAgICByYWRvc19uYW1lc3BhY2UgPSBzZWxmLl9hcmdfcGFyc2VyLnJhZG9zX25hbWVz
            cGFjZQogICAgICAgICAgICBpZiByYmRfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAibWFu
            ZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1yYmQtZGF0YS1wb29sLW5h
            bWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmIGs4c19jbHVzdGVyX25h
            bWUgPT0gIiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KAogICAgICAgICAgICAgICAgICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBz
            ZXQgdGhlICctLWs4cy1jbHVzdGVyLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAg
            ICAgICAgIGVudGl0eSA9IHNlbGYuZ2V0X2VudGl0eSgKICAgICAgICAgICAgICAgIGVudGl0eSwK
            ICAgICAgICAgICAgICAgIHJiZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBhbGlhc19yYmRf
            cG9vbF9uYW1lLAogICAgICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAgICAgICAg
            ICAgIHJhZG9zX25hbWVzcGFjZSwKICAgICAgICAgICAgKQogICAgICAgICAgICBpZiByYWRvc19u
            YW1lc3BhY2UgIT0gIiI6CiAgICAgICAgICAgICAgICBjYXBzWyJvc2QiXSA9ICgKICAgICAgICAg
            ICAgICAgICAgICBmInByb2ZpbGUgcmJkIHBvb2w9e3JiZF9wb29sX25hbWV9IG5hbWVzcGFjZT17
            cmFkb3NfbmFtZXNwYWNlfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxzZToKICAg
            ICAgICAgICAgICAgIGNhcHNbIm9zZCJdID0gZiJwcm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9u
            YW1lfSIKCiAgICAgICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfZGVmYXVsdFVz
            ZXJfY2Fwc19hbmRfZW50aXR5KHNlbGYpOgogICAgICAgIGVudGl0eSA9IHNlbGYucnVuX2FzX3Vz
            ZXIKICAgICAgICBjYXBzID0gewogICAgICAgICAgICAibW9uIjogImFsbG93IHIsIGFsbG93IGNv
            bW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFuZCB2ZXJzaW9uIiwKICAgICAgICAgICAg
            Im1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsCiAgICAgICAgICAgICJvc2QiOiBmInByb2Zp
            bGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9e3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bv
            b2xfcHJlZml4fS5yZ3cubWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9v
            bD17c2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5jb250cm9sLCBhbGxvdyBy
            eCBwb29sPXtzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmxvZywgYWxsb3cg
            eCBwb29sPXtzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmJ1Y2tldHMuaW5k
            ZXgiLAogICAgICAgIH0KCiAgICAgICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRf
            Y2Fwc19hbmRfZW50aXR5KHNlbGYsIHVzZXJfbmFtZSk6CiAgICAgICAgaWYgImNsaWVudC5jc2kt
            Y2VwaGZzLXByb3Zpc2lvbmVyIiBpbiB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQu
            Y3NpLWNlcGhmcy1wcm92aXNpb25lciIgIT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAg
            ICAgcmV0dXJuIHNlbGYuZ2V0X2NlcGhmc19wcm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoKQog
            ICAgICAgIGlmICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiBpbiB1c2VyX25hbWU6CiAgICAgICAg
            ICAgIGlmICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiAhPSB1c2VyX25hbWU6CiAgICAgICAgICAg
            ICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQog
            ICAgICAgICAgICByZXR1cm4gc2VsZi5nZXRfY2VwaGZzX25vZGVfY2Fwc19hbmRfZW50aXR5KCkK
            ICAgICAgICBpZiAiY2xpZW50LmNzaS1yYmQtcHJvdmlzaW9uZXIiIGluIHVzZXJfbmFtZToKICAg
            ICAgICAgICAgaWYgImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiAhPSB1c2VyX25hbWU6CiAg
            ICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9u
            ID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5nZXRfcmJkX3Byb3Zpc2lvbmVyX2NhcHNf
            YW5kX2VudGl0eSgpCiAgICAgICAgaWYgImNsaWVudC5jc2ktcmJkLW5vZGUiIGluIHVzZXJfbmFt
            ZToKICAgICAgICAgICAgaWYgImNsaWVudC5jc2ktcmJkLW5vZGUiICE9IHVzZXJfbmFtZToKICAg
            ICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24g
            PSBUcnVlCiAgICAgICAgICAgIHJldHVybiBzZWxmLmdldF9yYmRfbm9kZV9jYXBzX2FuZF9lbnRp
            dHkoKQogICAgICAgIGlmICJjbGllbnQuaGVhbHRoY2hlY2tlciIgaW4gdXNlcl9uYW1lOgogICAg
            ICAgICAgICBpZiAiY2xpZW50LmhlYWx0aGNoZWNrZXIiICE9IHVzZXJfbmFtZToKICAgICAgICAg
            ICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVl
            CiAgICAgICAgICAgIHJldHVybiBzZWxmLmdldF9kZWZhdWx0VXNlcl9jYXBzX2FuZF9lbnRpdHko
            KQoKICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICBm
            Im5vIHVzZXIgZm91bmQgd2l0aCB1c2VyX25hbWU6IHt1c2VyX25hbWV9LCAiCiAgICAgICAgICAg
            ICJnZXRfY2Fwc19hbmRfZW50aXR5IGNvbW1hbmQgZmFpbGVkLlxuIgogICAgICAgICkKCiAgICBk
            ZWYgY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3VzZXIoc2VsZiwgdXNlcik6CiAgICAgICAgIiIiCiAg
            ICAgICAgY29tbWFuZDogY2VwaCBhdXRoIGdldC1vci1jcmVhdGUgY2xpZW50LmNzaS1jZXBoZnMt
            cHJvdmlzaW9uZXIgbW9uICdhbGxvdyByJyBtZ3IgJ2FsbG93IHJ3JyBvc2QgJ2FsbG93IHJ3IHRh
            ZyBjZXBoZnMgbWV0YWRhdGE9KicKICAgICAgICAiIiIKICAgICAgICBjYXBzLCBlbnRpdHkgPSBz
            ZWxmLmdldF9jYXBzX2FuZF9lbnRpdHkodXNlcikKICAgICAgICBjbWRfanNvbiA9IHsKICAgICAg
            ICAgICAgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUiLAogICAgICAgICAgICAiZW50aXR5
            IjogZW50aXR5LAogICAgICAgICAgICAiY2FwcyI6IFtjYXAgZm9yIGNhcF9saXN0IGluIGxpc3Qo
            Y2Fwcy5pdGVtcygpKSBmb3IgY2FwIGluIGNhcF9saXN0XSwKICAgICAgICAgICAgImZvcm1hdCI6
            ICJqc29uIiwKICAgICAgICB9CgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoK
            ICAgICAgICAgICAgcmV0dXJuICgKICAgICAgICAgICAgICAgIHNlbGYuZHJ5X3J1bigKICAgICAg
            ICAgICAgICAgICAgICAiY2VwaCAiCiAgICAgICAgICAgICAgICAgICAgKyBjbWRfanNvblsicHJl
            Zml4Il0KICAgICAgICAgICAgICAgICAgICArICIgIgogICAgICAgICAgICAgICAgICAgICsgY21k
            X2pzb25bImVudGl0eSJdCiAgICAgICAgICAgICAgICAgICAgKyAiICIKICAgICAgICAgICAgICAg
            ICAgICArICIgIi5qb2luKGNtZF9qc29uWyJjYXBzIl0pCiAgICAgICAgICAgICAgICApLAogICAg
            ICAgICAgICAgICAgIiIsCiAgICAgICAgICAgICkKICAgICAgICAjIGNoZWNrIGlmIHVzZXIgYWxy
            ZWFkeSBleGlzdAogICAgICAgIHVzZXJfa2V5ID0gc2VsZi5jaGVja191c2VyX2V4aXN0KGVudGl0
            eSkKICAgICAgICBpZiB1c2VyX2tleSAhPSAiIjoKICAgICAgICAgICAgcmV0dXJuIHVzZXJfa2V5
            LCBmIntlbnRpdHkuc3BsaXQoJy4nLCAxKVsxXX0iCiAgICAgICAgICAgICMgZW50aXR5LnNwbGl0
            KCcuJywxKVsxXSB0byByZW5hbWUgZW50aXR5KGNsaWVudC5jc2ktcmJkLW5vZGUpIGFzIGNzaS1y
            YmQtbm9kZQoKICAgICAgICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1v
            bl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nl
            c3NmdWwgYXR0ZW1wdCwKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9
            PSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAg
            ICAgICAgICAgZiInYXV0aCBnZXQtb3ItY3JlYXRlIHt1c2VyfScgY29tbWFuZCBmYWlsZWQuXG4i
            CiAgICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBz
            ZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgKQogICAgICAgIHJldHVybiBzdHIo
            anNvbl9vdXRbMF1bImtleSJdKSwgZiJ7ZW50aXR5LnNwbGl0KCcuJywgMSlbMV19IgogICAgICAg
            ICMgZW50aXR5LnNwbGl0KCcuJywxKVsxXSB0byByZW5hbWUgZW50aXR5KGNsaWVudC5jc2ktcmJk
            LW5vZGUpIGFzIGNzaS1yYmQtbm9kZQoKICAgIGRlZiBnZXRfY2VwaGZzX2RhdGFfcG9vbF9kZXRh
            aWxzKHNlbGYpOgogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAiZnMgbHMiLCAiZm9ybWF0
            IjogImpzb24ifQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAg
            ICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigiY2VwaCAiICsgY21kX2pzb25bInByZWZpeCJdKQogICAg
            ICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dl
            bihjbWRfanNvbikKICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0
            LCByZXBvcnQgYW4gZXJyb3IKICAgICAgICBpZiByZXRfdmFsICE9IDA6CiAgICAgICAgICAgICMg
            aWYgZnMgYW5kIGRhdGFfcG9vbCBhcmd1bWVudHMgYXJlIG5vdCBzZXQsIHNpbGVudGx5IHJldHVy
            bgogICAgICAgICAgICBpZiAoCiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhm
            c19maWxlc3lzdGVtX25hbWUgPT0gIiIKICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFy
            c2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZSA9PSAiIgogICAgICAgICAgICApOgogICAgICAgICAg
            ICAgICAgcmV0dXJuCiAgICAgICAgICAgICMgaWYgdXNlciBoYXMgcHJvdmlkZWQgYW55IG9mIHRo
            ZQogICAgICAgICAgICAjICctLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUnIG9yICctLWNlcGhmcy1k
            YXRhLXBvb2wtbmFtZScgYXJndW1lbnRzLAogICAgICAgICAgICAjIHJhaXNlIGFuIGV4Y2VwdGlv
            biBhcyB3ZSBhcmUgdW5hYmxlIHRvIHZlcmlmeSB0aGUgYXJncwogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiInZnMgbHMnIGNlcGgg
            Y2FsbCBmYWlsZWQgd2l0aCBlcnJvcjoge2Vycl9tc2d9IgogICAgICAgICAgICApCgogICAgICAg
            IG1hdGNoaW5nX2pzb25fb3V0ID0ge30KICAgICAgICAjIGlmICctLWNlcGhmcy1maWxlc3lzdGVt
            LW5hbWUnIGFyZ3VtZW50IGlzIHByb3ZpZGVkLAogICAgICAgICMgY2hlY2sgd2hldGhlciB0aGUg
            cHJvdmlkZWQgZmlsZXN5c3RlbS1uYW1lIGV4aXN0cyBvciBub3QKICAgICAgICBpZiBzZWxmLl9h
            cmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWU6CiAgICAgICAgICAgICMgZ2V0IHRoZSBt
            YXRjaGluZyBsaXN0CiAgICAgICAgICAgIG1hdGNoaW5nX2pzb25fb3V0X2xpc3QgPSBbCiAgICAg
            ICAgICAgICAgICBtYXRjaGVkCiAgICAgICAgICAgICAgICBmb3IgbWF0Y2hlZCBpbiBqc29uX291
            dAogICAgICAgICAgICAgICAgaWYgc3RyKG1hdGNoZWRbIm5hbWUiXSkgPT0gc2VsZi5fYXJnX3Bh
            cnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lCiAgICAgICAgICAgIF0KICAgICAgICAgICAgIyB1
            bmFibGUgdG8gZmluZCBhIG1hdGNoaW5nIGZzLW5hbWUsIHJhaXNlIGFuIGVycm9yCiAgICAgICAg
            ICAgIGlmIGxlbihtYXRjaGluZ19qc29uX291dF9saXN0KSA9PSAwOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBmIkZp
            bGVzeXN0ZW0gcHJvdmlkZWQsICd7c2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9u
            YW1lfScsICIKICAgICAgICAgICAgICAgICAgICBmImlzIG5vdCBmb3VuZCBpbiB0aGUgZnMtbGlz
            dDoge1tzdHIoeFsnbmFtZSddKSBmb3IgeCBpbiBqc29uX291dF19IgogICAgICAgICAgICAgICAg
            KQogICAgICAgICAgICBtYXRjaGluZ19qc29uX291dCA9IG1hdGNoaW5nX2pzb25fb3V0X2xpc3Rb
            MF0KICAgICAgICAjIGlmIGNlcGhmcyBmaWxlc3lzdGVtIG5hbWUgaXMgbm90IHByb3ZpZGVkLAog
            ICAgICAgICMgdHJ5IHRvIGdldCBhIGRlZmF1bHQgZnMgbmFtZSBieSBkb2luZyB0aGUgZm9sbG93
            aW5nCiAgICAgICAgZWxzZToKICAgICAgICAgICAgIyBhLiBjaGVjayBpZiB0aGVyZSBpcyBvbmx5
            IG9uZSBmaWxlc3lzdGVtIGlzIHByZXNlbnQKICAgICAgICAgICAgaWYgbGVuKGpzb25fb3V0KSA9
            PSAxOgogICAgICAgICAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXQgPSBqc29uX291dFswXQogICAg
            ICAgICAgICAjIGIuIG9yIGVsc2UsIGNoZWNrIGlmIGRhdGFfcG9vbCBuYW1lIGlzIHByb3ZpZGVk
            CiAgICAgICAgICAgIGVsaWYgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWU6
            CiAgICAgICAgICAgICAgICAjIGFuZCBpZiBwcmVzZW50LCBjaGVjayB3aGV0aGVyIHRoZXJlIGV4
            aXN0cyBhIGZzIHdoaWNoIGhhcyB0aGUgZGF0YV9wb29sCiAgICAgICAgICAgICAgICBmb3IgZWFj
            aEogaW4ganNvbl9vdXQ6CiAgICAgICAgICAgICAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5j
            ZXBoZnNfZGF0YV9wb29sX25hbWUgaW4gZWFjaEpbImRhdGFfcG9vbHMiXToKICAgICAgICAgICAg
            ICAgICAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXQgPSBlYWNoSgogICAgICAgICAgICAgICAgICAg
            ICAgICBicmVhawogICAgICAgICAgICAgICAgIyBpZiB0aGVyZSBpcyBubyBtYXRjaGluZyBmcyBl
            eGlzdHMsIHRoYXQgbWVhbnMgcHJvdmlkZWQgZGF0YV9wb29sIG5hbWUgaXMgaW52YWxpZAogICAg
            ICAgICAgICAgICAgaWYgbm90IG1hdGNoaW5nX2pzb25fb3V0OgogICAgICAgICAgICAgICAgICAg
            IHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgICAg
            IGYiUHJvdmlkZWQgZGF0YV9wb29sIG5hbWUsIHtzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRh
            X3Bvb2xfbmFtZX0sIgogICAgICAgICAgICAgICAgICAgICAgICAiIGRvZXMgbm90IGV4aXN0cyIK
            ICAgICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICMgYy4gaWYgbm90aGluZyBpcyBzZXQg
            YW5kIGNvdWxkbid0IGZpbmQgYSBkZWZhdWx0LAogICAgICAgICAgICBlbHNlOgogICAgICAgICAg
            ICAgICAgIyBqdXN0IHJldHVybiBzaWxlbnRseQogICAgICAgICAgICAgICAgcmV0dXJuCgogICAg
            ICAgIGlmIG1hdGNoaW5nX2pzb25fb3V0OgogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNl
            cGhmc19maWxlc3lzdGVtX25hbWUgPSBzdHIobWF0Y2hpbmdfanNvbl9vdXRbIm5hbWUiXSkKICAg
            ICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfbWV0YWRhdGFfcG9vbF9uYW1lID0gc3Ry
            KAogICAgICAgICAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXRbIm1ldGFkYXRhX3Bvb2wiXQogICAg
            ICAgICAgICApCgogICAgICAgIGlmIGlzaW5zdGFuY2UobWF0Y2hpbmdfanNvbl9vdXRbImRhdGFf
            cG9vbHMiXSwgbGlzdCk6CiAgICAgICAgICAgICMgaWYgdGhlIHVzZXIgaGFzIGFscmVhZHkgcHJv
            dmlkZWQgZGF0YS1wb29sLW5hbWUsCiAgICAgICAgICAgICMgdGhyb3VnaCAtLWNlcGhmcy1kYXRh
            LXBvb2wtbmFtZQogICAgICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bv
            b2xfbmFtZToKICAgICAgICAgICAgICAgICMgaWYgdGhlIHByb3ZpZGVkIG5hbWUgaXMgbm90IG1h
            dGNoaW5nIHdpdGggdGhlIG9uZSBpbiB0aGUgbGlzdAogICAgICAgICAgICAgICAgaWYgKAogICAg
            ICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lCiAg
            ICAgICAgICAgICAgICAgICAgbm90IGluIG1hdGNoaW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0K
            ICAgICAgICAgICAgICAgICk6CiAgICAgICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAgICAgZiJQcm92aWRlZCBkYXRhLXBv
            b2wtbmFtZTogJ3tzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZX0nLCAiCiAg
            ICAgICAgICAgICAgICAgICAgICAgICJkb2Vzbid0IG1hdGNoIGZyb20gdGhlIGRhdGEtcG9vbHMg
            bGlzdDogIgogICAgICAgICAgICAgICAgICAgICAgICBmIntbc3RyKHgpIGZvciB4IGluIG1hdGNo
            aW5nX2pzb25fb3V0WydkYXRhX3Bvb2xzJ11dfSIKICAgICAgICAgICAgICAgICAgICApCiAgICAg
            ICAgICAgICMgaWYgZGF0YV9wb29sIG5hbWUgaXMgbm90IHByb3ZpZGVkLAogICAgICAgICAgICAj
            IHRoZW4gdHJ5IHRvIGZpbmQgYSBkZWZhdWx0IGRhdGEgcG9vbCBuYW1lCiAgICAgICAgICAgIGVs
            c2U6CiAgICAgICAgICAgICAgICAjIGlmIG5vIGRhdGFfcG9vbHMgZXhpc3QsIHNpbGVudGx5IHJl
            dHVybgogICAgICAgICAgICAgICAgaWYgbGVuKG1hdGNoaW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xz
            Il0pID09IDA6CiAgICAgICAgICAgICAgICAgICAgcmV0dXJuCiAgICAgICAgICAgICAgICBzZWxm
            Ll9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZSA9IHN0cigKICAgICAgICAgICAgICAg
            ICAgICBtYXRjaGluZ19qc29uX291dFsiZGF0YV9wb29scyJdWzBdCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgICMgaWYgdGhlcmUgYXJlIG1vcmUgdGhhbiBvbmUgJ2RhdGFfcG9vbHMnIGV4
            aXN0LAogICAgICAgICAgICAjIHRoZW4gd2FybiB0aGUgdXNlciB0aGF0IHdlIGFyZSB1c2luZyB0
            aGUgc2VsZWN0ZWQgbmFtZQogICAgICAgICAgICBpZiBsZW4obWF0Y2hpbmdfanNvbl9vdXRbImRh
            dGFfcG9vbHMiXSkgPiAxOgogICAgICAgICAgICAgICAgcHJpbnQoCiAgICAgICAgICAgICAgICAg
            ICAgIldBUk5JTkc6IE11bHRpcGxlIGRhdGEgcG9vbHMgZGV0ZWN0ZWQ6ICIKICAgICAgICAgICAg
            ICAgICAgICBmIntbc3RyKHgpIGZvciB4IGluIG1hdGNoaW5nX2pzb25fb3V0WydkYXRhX3Bvb2xz
            J11dfVxuIgogICAgICAgICAgICAgICAgICAgIGYiVXNpbmcgdGhlIGRhdGEtcG9vbDogJ3tzZWxm
            Ll9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZX0nXG4iCiAgICAgICAgICAgICAgICAp
            CgogICAgZGVmIGNyZWF0ZV9jaGVja2VyS2V5KHNlbGYsIHVzZXIpOgogICAgICAgIGNhcHMsIGVu
            dGl0eSA9IHNlbGYuZ2V0X2NhcHNfYW5kX2VudGl0eSh1c2VyKQogICAgICAgIGNtZF9qc29uID0g
            ewogICAgICAgICAgICAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0ZSIsCiAgICAgICAgICAg
            ICJlbnRpdHkiOiBlbnRpdHksCiAgICAgICAgICAgICJjYXBzIjogW2NhcCBmb3IgY2FwX2xpc3Qg
            aW4gbGlzdChjYXBzLml0ZW1zKCkpIGZvciBjYXAgaW4gY2FwX2xpc3RdLAogICAgICAgICAgICAi
            Zm9ybWF0IjogImpzb24iLAogICAgICAgIH0KCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5k
            cnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKAogICAgICAgICAgICAgICAg
            ImNlcGggIgogICAgICAgICAgICAgICAgKyBjbWRfanNvblsicHJlZml4Il0KICAgICAgICAgICAg
            ICAgICsgIiAiCiAgICAgICAgICAgICAgICArIGNtZF9qc29uWyJlbnRpdHkiXQogICAgICAgICAg
            ICAgICAgKyAiICIKICAgICAgICAgICAgICAgICsgIiAiLmpvaW4oY21kX2pzb25bImNhcHMiXSkK
            ICAgICAgICAgICAgKQogICAgICAgICMgY2hlY2sgaWYgdXNlciBhbHJlYWR5IGV4aXN0CiAgICAg
            ICAgdXNlcl9rZXkgPSBzZWxmLmNoZWNrX3VzZXJfZXhpc3QoZW50aXR5KQogICAgICAgIGlmIHVz
            ZXJfa2V5ICE9ICIiOgogICAgICAgICAgICByZXR1cm4gdXNlcl9rZXkKCiAgICAgICAgcmV0X3Zh
            bCwganNvbl9vdXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29u
            KQogICAgICAgICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAg
            aWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmFpc2Ug
            RXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgIGYiJ2F1dGggZ2V0LW9y
            LWNyZWF0ZSB7c2VsZi5ydW5fYXNfdXNlcn0nIGNvbW1hbmQgZmFpbGVkXG4iCiAgICAgICAgICAg
            ICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09V
            VFBVVF9MSVNUfSIKICAgICAgICAgICAgKQogICAgICAgIHJldHVybiBzdHIoanNvbl9vdXRbMF1b
            ImtleSJdKQoKICAgIGRlZiBnZXRfY2VwaF9kYXNoYm9hcmRfbGluayhzZWxmKToKICAgICAgICBj
            bWRfanNvbiA9IHsicHJlZml4IjogIm1nciBzZXJ2aWNlcyIsICJmb3JtYXQiOiAianNvbiJ9CiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2Vs
            Zi5kcnlfcnVuKCJjZXBoICIgKyBjbWRfanNvblsicHJlZml4Il0pCiAgICAgICAgcmV0X3ZhbCwg
            anNvbl9vdXQsIF8gPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAg
            ICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3Zh
            bCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmV0dXJuIE5vbmUKICAg
            ICAgICBpZiAiZGFzaGJvYXJkIiBub3QgaW4ganNvbl9vdXQ6CiAgICAgICAgICAgIHJldHVybiBO
            b25lCiAgICAgICAgcmV0dXJuIGpzb25fb3V0WyJkYXNoYm9hcmQiXQoKICAgIGRlZiBjcmVhdGVf
            cmd3X2FkbWluX29wc191c2VyKHNlbGYpOgogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgInJh
            ZG9zZ3ctYWRtaW4iLAogICAgICAgICAgICAidXNlciIsCiAgICAgICAgICAgICJjcmVhdGUiLAog
            ICAgICAgICAgICAiLS11aWQiLAogICAgICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9P
            UFNfVVNFUl9OQU1FLAogICAgICAgICAgICAiLS1kaXNwbGF5LW5hbWUiLAogICAgICAgICAgICAi
            Um9vayBSR1cgQWRtaW4gT3BzIHVzZXIiLAogICAgICAgICAgICAiLS1jYXBzIiwKICAgICAgICAg
            ICAgImJ1Y2tldHM9Kjt1c2Vycz0qO3VzYWdlPXJlYWQ7bWV0YWRhdGE9cmVhZDt6b25lPXJlYWQi
            LAogICAgICAgICAgICAiLS1yZ3ctcmVhbG0iLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd19yZWFsbV9uYW1lLAogICAgICAgICAgICAiLS1yZ3ctem9uZWdyb3VwIiwKICAgICAgICAg
            ICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZWdyb3VwX25hbWUsCiAgICAgICAgICAgICItLXJn
            dy16b25lIiwKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1lLAogICAg
            ICAgIF0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJl
            dHVybiBzZWxmLmRyeV9ydW4oImNlcGggIiArICIgIi5qb2luKGNtZCkpCiAgICAgICAgdHJ5Ogog
            ICAgICAgICAgICBvdXRwdXQgPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1z
            dWJwcm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vy
            cm9yIGFzIGV4ZWNFcnI6CiAgICAgICAgICAgICMgaWYgdGhlIHVzZXIgYWxyZWFkeSBleGlzdHMs
            IHdlIGp1c3QgcXVlcnkgaXQKICAgICAgICAgICAgaWYgZXhlY0Vyci5yZXR1cm5jb2RlID09IGVy
            cm5vLkVFWElTVDoKICAgICAgICAgICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgICAgICAgICAi
            cmFkb3Nndy1hZG1pbiIsCiAgICAgICAgICAgICAgICAgICAgInVzZXIiLAogICAgICAgICAgICAg
            ICAgICAgICJpbmZvIiwKICAgICAgICAgICAgICAgICAgICAiLS11aWQiLAogICAgICAgICAgICAg
            ICAgICAgIHNlbGYuRVhURVJOQUxfUkdXX0FETUlOX09QU19VU0VSX05BTUUsCiAgICAgICAgICAg
            ICAgICAgICAgIi0tcmd3LXJlYWxtIiwKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFy
            c2VyLnJnd19yZWFsbV9uYW1lLAogICAgICAgICAgICAgICAgICAgICItLXJndy16b25lZ3JvdXAi
            LAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1l
            LAogICAgICAgICAgICAgICAgICAgICItLXJndy16b25lIiwKICAgICAgICAgICAgICAgICAgICBz
            ZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUsCiAgICAgICAgICAgICAgICBdCiAgICAgICAg
            ICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVj
            a19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAgICAgICAgZXhj
            ZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yIGFzIGV4ZWNFcnI6CiAgICAgICAgICAg
            ICAgICAgICAgZXJyX21zZyA9ICgKICAgICAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8g
            ZXhlY3V0ZSBjb21tYW5kIHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAg
            ICAgICAgICAgICAgICAgICAgZiJDb2RlOiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtl
            eGVjRXJyLnN0ZGVycn0iCiAgICAgICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAg
            IHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykKICAgICAgICAgICAgICAgICAgICByZXR1cm4gTm9u
            ZSwgTm9uZSwgRmFsc2UsICItMSIKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVy
            cl9tc2cgPSAoCiAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5k
            IHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICBm
            IkNvZGU6IHtleGVjRXJyLnJldHVybmNvZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykK
            ICAgICAgICAgICAgICAgIHJldHVybiBOb25lLCBOb25lLCBGYWxzZSwgIi0xIgoKICAgICAgICAj
            IGlmIGl0IGlzIHB5dGhvbjIsIGRvbid0IGNoZWNrIGZvciBjZXBoIHZlcnNpb24gZm9yIGFkZGlu
            ZyBgaW5mbz1yZWFkYCBjYXAocmd3X3ZhbGlkYXRpb24pCiAgICAgICAgaWYgc3lzLnZlcnNpb25f
            aW5mby5tYWpvciA8IDM6CiAgICAgICAgICAgIGpzb25vdXRwdXQgPSBqc29uLmxvYWRzKG91dHB1
            dCkKICAgICAgICAgICAgcmV0dXJuICgKICAgICAgICAgICAgICAgIGpzb25vdXRwdXRbImtleXMi
            XVswXVsiYWNjZXNzX2tleSJdLAogICAgICAgICAgICAgICAganNvbm91dHB1dFsia2V5cyJdWzBd
            WyJzZWNyZXRfa2V5Il0sCiAgICAgICAgICAgICAgICBGYWxzZSwKICAgICAgICAgICAgICAgICIi
            LAogICAgICAgICAgICApCgogICAgICAgICMgc2VwYXJhdGVseSBhZGQgaW5mbz1yZWFkIGNhcHMg
            Zm9yIHJndy1lbmRwb2ludCBpcCB2YWxpZGF0aW9uCiAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVk
            ID0gVHJ1ZQogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAg
            ICAgICAgICAiY2FwcyIsCiAgICAgICAgICAgICJhZGQiLAogICAgICAgICAgICAiLS11aWQiLAog
            ICAgICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9PUFNfVVNFUl9OQU1FLAogICAgICAg
            ICAgICAiLS1jYXBzIiwKICAgICAgICAgICAgImluZm89cmVhZCIsCiAgICAgICAgICAgICItLXJn
            dy1yZWFsbSIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUsCiAg
            ICAgICAgICAgICItLXJndy16b25lZ3JvdXAiLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd196b25lZ3JvdXBfbmFtZSwKICAgICAgICAgICAgIi0tcmd3LXpvbmUiLAogICAgICAgICAg
            ICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUsCiAgICAgICAgXQogICAgICAgIHRyeToK
            ICAgICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9
            c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2VwdCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NF
            cnJvciBhcyBleGVjRXJyOgogICAgICAgICAgICAjIGlmIHRoZSBjZXBoIHZlcnNpb24gbm90IHN1
            cHBvcnRlZCBmb3IgYWRkaW5nIGBpbmZvPXJlYWRgIGNhcChyZ3dfdmFsaWRhdGlvbikKICAgICAg
            ICAgICAgaWYgKAogICAgICAgICAgICAgICAgImNvdWxkIG5vdCBhZGQgY2FwczogdW5hYmxlIHRv
            IGFkZCBjYXBzOiBpbmZvPXJlYWRcbiIKICAgICAgICAgICAgICAgIGluIGV4ZWNFcnIuc3RkZXJy
            LmRlY29kZSgidXRmLTgiKQogICAgICAgICAgICAgICAgYW5kIGV4ZWNFcnIucmV0dXJuY29kZSA9
            PSAyNDQKICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgIGluZm9fY2FwX3N1cHBvcnRlZCA9
            IEZhbHNlCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICBlcnJfbXNnID0gKAogICAg
            ICAgICAgICAgICAgICAgIGYiZmFpbGVkIHRvIGV4ZWN1dGUgY29tbWFuZCB7Y21kfS4gT3V0cHV0
            OiB7ZXhlY0Vyci5vdXRwdXR9LiAiCiAgICAgICAgICAgICAgICAgICAgZiJDb2RlOiB7ZXhlY0Vy
            ci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtleGVjRXJyLnN0ZGVycn0iCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKGVycl9tc2cpCiAgICAgICAgICAgICAg
            ICByZXR1cm4gTm9uZSwgTm9uZSwgRmFsc2UsICItMSIKCiAgICAgICAganNvbm91dHB1dCA9IGpz
            b24ubG9hZHMob3V0cHV0KQogICAgICAgIHJldHVybiAoCiAgICAgICAgICAgIGpzb25vdXRwdXRb
            ImtleXMiXVswXVsiYWNjZXNzX2tleSJdLAogICAgICAgICAgICBqc29ub3V0cHV0WyJrZXlzIl1b
            MF1bInNlY3JldF9rZXkiXSwKICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkLAogICAgICAg
            ICAgICAiIiwKICAgICAgICApCgogICAgZGVmIHZhbGlkYXRlX3JiZF9wb29sKHNlbGYpOgogICAg
            ICAgIGlmIG5vdCBzZWxmLmNsdXN0ZXIucG9vbF9leGlzdHMoc2VsZi5fYXJnX3BhcnNlci5yYmRf
            ZGF0YV9wb29sX25hbWUpOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0
            aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcG9vbCwgJ3tzZWxmLl9hcmdfcGFy
            c2VyLnJiZF9kYXRhX3Bvb2xfbmFtZX0nLCBkb2VzIG5vdCBleGlzdCIKICAgICAgICAgICAgKQoK
            ICAgIGRlZiBpbml0X3JiZF9wb29sKHNlbGYpOgogICAgICAgIGlmIGlzaW5zdGFuY2Uoc2VsZi5j
            bHVzdGVyLCBEdW1teVJhZG9zKToKICAgICAgICAgICAgcmV0dXJuCiAgICAgICAgcmJkX3Bvb2xf
            bmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgaW9jdHgg
            PSBzZWxmLmNsdXN0ZXIub3Blbl9pb2N0eChyYmRfcG9vbF9uYW1lKQogICAgICAgIHJiZF9pbnN0
            ID0gcmJkLlJCRCgpCiAgICAgICAgcmJkX2luc3QucG9vbF9pbml0KGlvY3R4LCBUcnVlKQoKICAg
            IGRlZiB2YWxpZGF0ZV9yYWRvc19uYW1lc3BhY2Uoc2VsZik6CiAgICAgICAgcmJkX3Bvb2xfbmFt
            ZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgcmFkb3NfbmFt
            ZXNwYWNlID0gc2VsZi5fYXJnX3BhcnNlci5yYWRvc19uYW1lc3BhY2UKICAgICAgICBpZiByYWRv
            c19uYW1lc3BhY2UgPT0gIiI6CiAgICAgICAgICAgIHJldHVybgogICAgICAgIGlmIHJhZG9zX25h
            bWVzcGFjZS5pc2xvd2VyKCkgPT0gRmFsc2U6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCByYWRvcyBOYW1l
            c3BhY2UsICd7cmFkb3NfbmFtZXNwYWNlfScsICIKICAgICAgICAgICAgICAgIGYiY29udGFpbnMg
            dXBwZXIgY2FzZSIKICAgICAgICAgICAgKQogICAgICAgIHJiZF9pbnN0ID0gcmJkLlJCRCgpCiAg
            ICAgICAgaW9jdHggPSBzZWxmLmNsdXN0ZXIub3Blbl9pb2N0eChyYmRfcG9vbF9uYW1lKQogICAg
            ICAgIGlmIHJiZF9pbnN0Lm5hbWVzcGFjZV9leGlzdHMoaW9jdHgsIHJhZG9zX25hbWVzcGFjZSkg
            aXMgRmFsc2U6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAg
            ICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCByYWRvcyBOYW1lc3BhY2UsICd7cmFkb3NfbmFt
            ZXNwYWNlfScsICIKICAgICAgICAgICAgICAgIGYiaXMgbm90IGZvdW5kIGluIHRoZSBwb29sICd7
            cmJkX3Bvb2xfbmFtZX0nIgogICAgICAgICAgICApCgogICAgZGVmIGdldF9vcl9jcmVhdGVfc3Vi
            dm9sdW1lX2dyb3VwKHNlbGYsIHN1YnZvbHVtZV9ncm91cCwgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFt
            ZSk6CiAgICAgICAgY21kID0gWwogICAgICAgICAgICAiY2VwaCIsCiAgICAgICAgICAgICJmcyIs
            CiAgICAgICAgICAgICJzdWJ2b2x1bWVncm91cCIsCiAgICAgICAgICAgICJnZXRwYXRoIiwKICAg
            ICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAgICAgICAgICAgc3Vidm9sdW1lX2dy
            b3VwLAogICAgICAgIF0KICAgICAgICB0cnk6CiAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNo
            ZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1
            YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICBjbWQgPSBbCiAgICAgICAg
            ICAgICAgICAiY2VwaCIsCiAgICAgICAgICAgICAgICAiZnMiLAogICAgICAgICAgICAgICAgInN1
            YnZvbHVtZWdyb3VwIiwKICAgICAgICAgICAgICAgICJjcmVhdGUiLAogICAgICAgICAgICAgICAg
            Y2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAgICAgICAgICAgICAgIHN1YnZvbHVtZV9ncm91cCwK
            ICAgICAgICAgICAgXQogICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBfID0gc3VicHJv
            Y2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAg
            ICBleGNlcHQgc3VicHJvY2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3I6CiAgICAgICAgICAgICAgICBy
            YWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYic3Vi
            dm9sdW1lIGdyb3VwIHtzdWJ2b2x1bWVfZ3JvdXB9IGlzIG5vdCBhYmxlIHRvIGdldCBjcmVhdGVk
            IgogICAgICAgICAgICAgICAgKQoKICAgIGRlZiBwaW5fc3Vidm9sdW1lKAogICAgICAgIHNlbGYs
            IHN1YnZvbHVtZV9ncm91cCwgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwgcGluX3R5cGUsIHBpbl9z
            ZXR0aW5nCiAgICApOgogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgImNlcGgiLAogICAgICAg
            ICAgICAiZnMiLAogICAgICAgICAgICAic3Vidm9sdW1lZ3JvdXAiLAogICAgICAgICAgICAicGlu
            IiwKICAgICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAgICAgICAgICAgc3Vidm9s
            dW1lX2dyb3VwLAogICAgICAgICAgICBwaW5fdHlwZSwKICAgICAgICAgICAgcGluX3NldHRpbmcs
            CiAgICAgICAgXQogICAgICAgIHRyeToKICAgICAgICAgICAgXyA9IHN1YnByb2Nlc3MuY2hlY2tf
            b3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAgICAgICBleGNlcHQgc3VicHJv
            Y2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3I6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1
            cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmInN1YnZvbHVtZSBncm91cCB7c3Vidm9sdW1l
            X2dyb3VwfSBpcyBub3QgYWJsZSB0byBnZXQgcGlubmVkIgogICAgICAgICAgICApCgogICAgZGVm
            IGdldF9yZ3dfZnNpZChzZWxmLCBiYXNlX3VybCwgdmVyaWZ5KToKICAgICAgICBhY2Nlc3Nfa2V5
            ID0gc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJfQUNDRVNTX0tFWSJdCiAgICAgICAg
            c2VjcmV0X2tleSA9IHNlbGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX1NFQ1JFVF9LRVki
            XQogICAgICAgIHJnd19lbmRwb2ludCA9IHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50CiAg
            ICAgICAgYmFzZV91cmwgPSBiYXNlX3VybCArICI6Ly8iICsgcmd3X2VuZHBvaW50ICsgIi9hZG1p
            bi9pbmZvPyIKICAgICAgICBwYXJhbXMgPSB7ImZvcm1hdCI6ICJqc29uIn0KICAgICAgICByZXF1
            ZXN0X3VybCA9IGJhc2VfdXJsICsgdXJsZW5jb2RlKHBhcmFtcykKCiAgICAgICAgdHJ5OgogICAg
            ICAgICAgICByID0gcmVxdWVzdHMuZ2V0KAogICAgICAgICAgICAgICAgcmVxdWVzdF91cmwsCiAg
            ICAgICAgICAgICAgICBhdXRoPVMzQXV0aChhY2Nlc3Nfa2V5LCBzZWNyZXRfa2V5LCByZ3dfZW5k
            cG9pbnQpLAogICAgICAgICAgICAgICAgdmVyaWZ5PXZlcmlmeSwKICAgICAgICAgICAgKQogICAg
            ICAgIGV4Y2VwdCByZXF1ZXN0cy5leGNlcHRpb25zLlRpbWVvdXQ6CiAgICAgICAgICAgIHN5cy5z
            dGRlcnIud3JpdGUoCiAgICAgICAgICAgICAgICBmImludmFsaWQgZW5kcG9pbnQ6LCBub3QgYWJs
            ZSB0byBjYWxsIGFkbWluLW9wcyBhcGl7cmd3X2VuZHBvaW50fSIKICAgICAgICAgICAgKQogICAg
            ICAgICAgICByZXR1cm4gIiIsICItMSIKICAgICAgICByMSA9IHIuanNvbigpCiAgICAgICAgaWYg
            cjEgaXMgTm9uZSBvciByMS5nZXQoImluZm8iKSBpcyBOb25lOgogICAgICAgICAgICBzeXMuc3Rk
            ZXJyLndyaXRlKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmd3IEVuZHBvaW50LCAn
            e3NlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50fScsIGlzIGludmFsaWQuIgogICAgICAgICAg
            ICApCiAgICAgICAgICAgIHJldHVybiAoCiAgICAgICAgICAgICAgICAiIiwKICAgICAgICAgICAg
            ICAgICItMSIsCiAgICAgICAgICAgICkKCiAgICAgICAgcmV0dXJuIHIxWyJpbmZvIl1bInN0b3Jh
            Z2VfYmFja2VuZHMiXVswXVsiY2x1c3Rlcl9pZCJdLCAiIgoKICAgIGRlZiB2YWxpZGF0ZV9yZ3df
            ZW5kcG9pbnQoc2VsZiwgaW5mb19jYXBfc3VwcG9ydGVkKToKICAgICAgICAjIGlmIHRoZSAnY2x1
            c3RlcicgaW5zdGFuY2UgaXMgYSBkdW1teSBvbmUsCiAgICAgICAgIyBkb24ndCB0cnkgdG8gcmVh
            Y2ggb3V0IHRvIHRoZSBlbmRwb2ludAogICAgICAgIGlmIGlzaW5zdGFuY2Uoc2VsZi5jbHVzdGVy
            LCBEdW1teVJhZG9zKToKICAgICAgICAgICAgcmV0dXJuCgogICAgICAgIHJnd19lbmRwb2ludCA9
            IHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50CgogICAgICAgICMgdmFsaWRhdGUgcmd3IGVu
            ZHBvaW50IG9ubHkgaWYgaXAgYWRkcmVzcyBpcyBwYXNzZWQKICAgICAgICBpcF90eXBlID0gc2Vs
            Zi5faW52YWxpZF9lbmRwb2ludChyZ3dfZW5kcG9pbnQpCgogICAgICAgICMgY2hlY2sgaWYgdGhl
            IHJndyBlbmRwb2ludCBpcyByZWFjaGFibGUKICAgICAgICBjZXJ0ID0gTm9uZQogICAgICAgIGlm
            IG5vdCBzZWxmLl9hcmdfcGFyc2VyLnJnd19za2lwX3RscyBhbmQgc2VsZi52YWxpZGF0ZV9yZ3df
            ZW5kcG9pbnRfdGxzX2NlcnQoKToKICAgICAgICAgICAgY2VydCA9IHNlbGYuX2FyZ19wYXJzZXIu
            cmd3X3Rsc19jZXJ0X3BhdGgKICAgICAgICBiYXNlX3VybCwgdmVyaWZ5LCBlcnIgPSBzZWxmLmVu
            ZHBvaW50X2RpYWwocmd3X2VuZHBvaW50LCBpcF90eXBlLCBjZXJ0PWNlcnQpCiAgICAgICAgaWYg
            ZXJyICE9ICIiOgogICAgICAgICAgICByZXR1cm4gIi0xIgoKICAgICAgICAjIGNoZWNrIGlmIHRo
            ZSByZ3cgZW5kcG9pbnQgYmVsb25ncyB0byB0aGUgc2FtZSBjbHVzdGVyCiAgICAgICAgIyBvbmx5
            IGNoZWNrIGlmIGBpbmZvYCBjYXAgaXMgc3VwcG9ydGVkCiAgICAgICAgaWYgaW5mb19jYXBfc3Vw
            cG9ydGVkOgogICAgICAgICAgICBmc2lkID0gc2VsZi5nZXRfZnNpZCgpCiAgICAgICAgICAgIHJn
            d19mc2lkLCBlcnIgPSBzZWxmLmdldF9yZ3dfZnNpZChiYXNlX3VybCwgdmVyaWZ5KQogICAgICAg
            ICAgICBpZiBlcnIgPT0gIi0xIjoKICAgICAgICAgICAgICAgIHJldHVybiAiLTEiCiAgICAgICAg
            ICAgIGlmIGZzaWQgIT0gcmd3X2ZzaWQ6CiAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRl
            KAogICAgICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHJndyBFbmRwb2ludCwgJ3tzZWxm
            Ll9hcmdfcGFyc2VyLnJnd19lbmRwb2ludH0nLCBpcyBpbnZhbGlkLiBXZSBhcmUgdmFsaWRhdGlu
            ZyBieSBjYWxsaW5nIHRoZSBhZG1pbm9wcyBhcGkgdGhyb3VnaCByZ3ctZW5kcG9pbnQgYW5kIHZh
            bGlkYXRpbmcgdGhlIGNsdXN0ZXJfaWQgJ3tyZ3dfZnNpZH0nIGlzIGVxdWFsIHRvIHRoZSBjZXBo
            IGNsdXN0ZXIgZnNpZCAne2ZzaWR9JyIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAg
            IHJldHVybiAiLTEiCgogICAgICAgICMgY2hlY2sgaWYgdGhlIHJndyBlbmRwb2ludCBwb29sIGV4
            aXN0CiAgICAgICAgIyBvbmx5IHZhbGlkYXRlIGlmIHJnd19wb29sX3ByZWZpeCBpcyBwYXNzZWQg
            ZWxzZSBpdCB3aWxsIHRha2UgZGVmYXVsdCB2YWx1ZSBhbmQgd2UgZG9uJ3QgY3JlYXRlIHRoZXNl
            IGRlZmF1bHQgcG9vbHMKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZp
            eCAhPSAiZGVmYXVsdCI6CiAgICAgICAgICAgIHJnd19wb29sc190b192YWxpZGF0ZSA9IFsKICAg
            ICAgICAgICAgICAgIGYie3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cubWV0
            YSIsCiAgICAgICAgICAgICAgICAiLnJndy5yb290IiwKICAgICAgICAgICAgICAgIGYie3NlbGYu
            X2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cuY29udHJvbCIsCiAgICAgICAgICAgICAg
            ICBmIntzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmxvZyIsCiAgICAgICAg
            ICAgIF0KICAgICAgICAgICAgZm9yIF9yZ3dfcG9vbF90b192YWxpZGF0ZSBpbiByZ3dfcG9vbHNf
            dG9fdmFsaWRhdGU6CiAgICAgICAgICAgICAgICBpZiBub3Qgc2VsZi5jbHVzdGVyLnBvb2xfZXhp
            c3RzKF9yZ3dfcG9vbF90b192YWxpZGF0ZSk6CiAgICAgICAgICAgICAgICAgICAgc3lzLnN0ZGVy
            ci53cml0ZSgKICAgICAgICAgICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcG9vbCwgJ3tf
            cmd3X3Bvb2xfdG9fdmFsaWRhdGV9JywgZG9lcyBub3QgZXhpc3QiCiAgICAgICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICAgICAgICAgIHJldHVybiAiLTEiCgogICAgICAgIHJldHVybiAiIgoK
            ICAgIGRlZiB2YWxpZGF0ZV9yZ3dfbXVsdGlzaXRlKHNlbGYsIHJnd19tdWx0aXNpdGVfY29uZmln
            X25hbWUsIHJnd19tdWx0aXNpdGVfY29uZmlnKToKICAgICAgICBpZiByZ3dfbXVsdGlzaXRlX2Nv
            bmZpZyAhPSAiIjoKICAgICAgICAgICAgY21kID0gWwogICAgICAgICAgICAgICAgInJhZG9zZ3ct
            YWRtaW4iLAogICAgICAgICAgICAgICAgcmd3X211bHRpc2l0ZV9jb25maWcsCiAgICAgICAgICAg
            ICAgICAiZ2V0IiwKICAgICAgICAgICAgICAgICItLXJndy0iICsgcmd3X211bHRpc2l0ZV9jb25m
            aWcsCiAgICAgICAgICAgICAgICByZ3dfbXVsdGlzaXRlX2NvbmZpZ19uYW1lLAogICAgICAgICAg
            ICBdCiAgICAgICAgICAgIHRyeToKICAgICAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNoZWNr
            X291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgICAgIGV4Y2VwdCBz
            dWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJvciBhcyBleGVjRXJyOgogICAgICAgICAgICAgICAg
            ZXJyX21zZyA9ICgKICAgICAgICAgICAgICAgICAgICBmImZhaWxlZCB0byBleGVjdXRlIGNvbW1h
            bmQge2NtZH0uIE91dHB1dDoge2V4ZWNFcnIub3V0cHV0fS4gIgogICAgICAgICAgICAgICAgICAg
            IGYiQ29kZToge2V4ZWNFcnIucmV0dXJuY29kZX0uIEVycm9yOiB7ZXhlY0Vyci5zdGRlcnJ9Igog
            ICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZShlcnJfbXNn
            KQogICAgICAgICAgICAgICAgcmV0dXJuICItMSIKICAgICAgICByZXR1cm4gIiIKCiAgICBkZWYg
            X2dlbl9vdXRwdXRfbWFwKHNlbGYpOgogICAgICAgIGlmIHNlbGYub3V0X21hcDoKICAgICAgICAg
            ICAgcmV0dXJuCiAgICAgICAgIyBzdXBwb3J0IGxlZ2FjeSBmbGFnIHdpdGggdXBncmFkZXMKICAg
            ICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNsdXN0ZXJfbmFtZToKICAgICAgICAgICAgc2VsZi5f
            YXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5jbHVzdGVyX25h
            bWUKICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUgPSAoCiAgICAgICAg
            ICAgIHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZS5sb3dlcigpCiAgICAgICAgKSAg
            IyBhbHdheXMgY29udmVydCBjbHVzdGVyIG5hbWUgdG8gbG93ZXJjYXNlIGNoYXJhY3RlcnMKICAg
            ICAgICBzZWxmLnZhbGlkYXRlX3JiZF9wb29sKCkKICAgICAgICBzZWxmLmluaXRfcmJkX3Bvb2wo
            KQogICAgICAgIHNlbGYudmFsaWRhdGVfcmFkb3NfbmFtZXNwYWNlKCkKICAgICAgICBzZWxmLl9l
            eGNsdWRlZF9rZXlzLmFkZCgiSzhTX0NMVVNURVJfTkFNRSIpCiAgICAgICAgc2VsZi5nZXRfY2Vw
            aGZzX2RhdGFfcG9vbF9kZXRhaWxzKCkKICAgICAgICBzZWxmLm91dF9tYXBbIk5BTUVTUEFDRSJd
            ID0gc2VsZi5fYXJnX3BhcnNlci5uYW1lc3BhY2UKICAgICAgICBzZWxmLm91dF9tYXBbIks4U19D
            TFVTVEVSX05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQogICAgICAg
            IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9GU0lEIl0gPSBzZWxmLmdldF9mc2lkKCkKICAg
            ICAgICBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUk5BTUUiXSA9IHNlbGYucnVuX2Fz
            X3VzZXIKICAgICAgICBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfQ0VQSF9NT05fREFUQSJd
            ID0gc2VsZi5nZXRfY2VwaF9leHRlcm5hbF9tb25fZGF0YSgpCiAgICAgICAgc2VsZi5vdXRfbWFw
            WyJST09LX0VYVEVSTkFMX1VTRVJfU0VDUkVUIl0gPSBzZWxmLmNyZWF0ZV9jaGVja2VyS2V5KAog
            ICAgICAgICAgICAiY2xpZW50LmhlYWx0aGNoZWNrZXIiCiAgICAgICAgKQogICAgICAgIHNlbGYu
            b3V0X21hcFsiUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSyJdID0gc2VsZi5nZXRfY2VwaF9k
            YXNoYm9hcmRfbGluaygpCiAgICAgICAgKAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9S
            QkRfTk9ERV9TRUNSRVQiXSwKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVf
            U0VDUkVUX05BTUUiXSwKICAgICAgICApID0gc2VsZi5jcmVhdGVfY2VwaENTSUtleXJpbmdfdXNl
            cigiY2xpZW50LmNzaS1yYmQtbm9kZSIpCiAgICAgICAgKAogICAgICAgICAgICBzZWxmLm91dF9t
            YXBbIkNTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUIl0sCiAgICAgICAgICAgIHNlbGYub3V0X21h
            cFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICkgPSBzZWxmLmNy
            ZWF0ZV9jZXBoQ1NJS2V5cmluZ191c2VyKCJjbGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIpCiAg
            ICAgICAgc2VsZi5vdXRfbWFwWyJDRVBIRlNfUE9PTF9OQU1FIl0gPSBzZWxmLl9hcmdfcGFyc2Vy
            LmNlcGhmc19kYXRhX3Bvb2xfbmFtZQogICAgICAgIHNlbGYub3V0X21hcFsiQ0VQSEZTX01FVEFE
            QVRBX1BPT0xfTkFNRSJdID0gKAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19t
            ZXRhZGF0YV9wb29sX25hbWUKICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJDRVBIRlNf
            RlNfTkFNRSJdID0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lCiAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJSRVNUUklDVEVEX0FVVEhfUEVSTUlTU0lPTiJdID0gKAogICAgICAg
            ICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uCiAgICAgICAg
            KQogICAgICAgIHNlbGYub3V0X21hcFsiUkFET1NfTkFNRVNQQUNFIl0gPSBzZWxmLl9hcmdfcGFy
            c2VyLnJhZG9zX25hbWVzcGFjZQogICAgICAgIHNlbGYub3V0X21hcFsiU1VCVk9MVU1FX0dST1VQ
            Il0gPSBzZWxmLl9hcmdfcGFyc2VyLnN1YnZvbHVtZV9ncm91cAogICAgICAgIHNlbGYub3V0X21h
            cFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVCJdID0gIiIKICAgICAgICBzZWxmLm91dF9tYXBbIkNT
            SV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUIl0gPSAiIgogICAgICAgICMgY3JlYXRlIENlcGhG
            UyBub2RlIGFuZCBwcm92aXNpb25lciBrZXlyaW5nIG9ubHkgd2hlbiBNRFMgZXhpc3RzCiAgICAg
            ICAgaWYgc2VsZi5vdXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdIGFuZCBzZWxmLm91dF9tYXBbIkNF
            UEhGU19QT09MX05BTUUiXToKICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgc2VsZi5vdXRf
            bWFwWyJDU0lfQ0VQSEZTX05PREVfU0VDUkVUIl0sCiAgICAgICAgICAgICAgICBzZWxmLm91dF9t
            YXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICApID0gc2VsZi5j
            cmVhdGVfY2VwaENTSUtleXJpbmdfdXNlcigiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIpCiAgICAg
            ICAgICAgICgKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9WSVNJ
            T05FUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19Q
            Uk9WSVNJT05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICApID0gc2VsZi5jcmVhdGVfY2Vw
            aENTSUtleXJpbmdfdXNlcigiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiKQogICAgICAg
            ICAgICAjIGNyZWF0ZSB0aGUgZGVmYXVsdCAiY3NpIiBzdWJ2b2x1bWVncm91cAogICAgICAgICAg
            ICBzZWxmLmdldF9vcl9jcmVhdGVfc3Vidm9sdW1lX2dyb3VwKAogICAgICAgICAgICAgICAgImNz
            aSIsIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgICAgICAp
            CiAgICAgICAgICAgICMgcGluIHRoZSBkZWZhdWx0ICJjc2kiIHN1YnZvbHVtZWdyb3VwCiAgICAg
            ICAgICAgIHNlbGYucGluX3N1YnZvbHVtZSgKICAgICAgICAgICAgICAgICJjc2kiLCBzZWxmLl9h
            cmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUsICJkaXN0cmlidXRlZCIsICIxIgogICAg
            ICAgICAgICApCiAgICAgICAgICAgIGlmIHNlbGYub3V0X21hcFsiU1VCVk9MVU1FX0dST1VQIl06
            CiAgICAgICAgICAgICAgICBzZWxmLmdldF9vcl9jcmVhdGVfc3Vidm9sdW1lX2dyb3VwKAogICAg
            ICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuc3Vidm9sdW1lX2dyb3VwLAogICAgICAg
            ICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgIHNlbGYucGluX3N1YnZvbHVtZSgKICAgICAg
            ICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnN1YnZvbHVtZV9ncm91cCwKICAgICAgICAg
            ICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAg
            ICAgICAgICAgICAgICAgImRpc3RyaWJ1dGVkIiwKICAgICAgICAgICAgICAgICAgICAiMSIsCiAg
            ICAgICAgICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJSR1dfVExTX0NFUlQiXSA9ICIi
            CiAgICAgICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UIl0gPSAiIgogICAgICAg
            IHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUIl0gPSAiIgogICAgICAgIGlm
            IG5vdCBzZWxmLl9hcmdfcGFyc2VyLnNraXBfbW9uaXRvcmluZ19lbmRwb2ludDoKICAgICAgICAg
            ICAgKAogICAgICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UIl0s
            CiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJd
            LAogICAgICAgICAgICApID0gc2VsZi5nZXRfYWN0aXZlX2FuZF9zdGFuZGJ5X21ncnMoKQogICAg
            ICAgIHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdID0gc2VsZi5fYXJnX3BhcnNlci5yYmRf
            ZGF0YV9wb29sX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBbIlJCRF9NRVRBREFUQV9FQ19QT09M
            X05BTUUiXSA9ICgKICAgICAgICAgICAgc2VsZi52YWxpZGF0ZV9yYmRfbWV0YWRhdGFfZWNfcG9v
            bF9uYW1lKCkKICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJSR1dfUE9PTF9QUkVGSVgi
            XSA9IHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4CiAgICAgICAgc2VsZi5vdXRfbWFw
            WyJSR1dfRU5EUE9JTlQiXSA9ICIiCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZ3dfZW5k
            cG9pbnQ6CiAgICAgICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAg
            ICAgICAgIHNlbGYuY3JlYXRlX3Jnd19hZG1pbl9vcHNfdXNlcigpCiAgICAgICAgICAgIGVsc2U6
            CiAgICAgICAgICAgICAgICBpZiAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfcmVhbG1fbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdf
            cGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBz
            ZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICk6CiAg
            ICAgICAgICAgICAgICAgICAgZXJyID0gc2VsZi52YWxpZGF0ZV9yZ3dfbXVsdGlzaXRlKAogICAg
            ICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lLCAicmVh
            bG0iCiAgICAgICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAgIGVyciA9IHNlbGYu
            dmFsaWRhdGVfcmd3X211bHRpc2l0ZSgKICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJn
            X3BhcnNlci5yZ3dfem9uZWdyb3VwX25hbWUsICJ6b25lZ3JvdXAiCiAgICAgICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICAgICAgICAgIGVyciA9IHNlbGYudmFsaWRhdGVfcmd3X211bHRpc2l0
            ZSgKICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1l
            LCAiem9uZSIKICAgICAgICAgICAgICAgICAgICApCgogICAgICAgICAgICAgICAgaWYgKAogICAg
            ICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUgPT0gIiIKICAg
            ICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZWdyb3VwX25hbWUg
            PT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZV9u
            YW1lID09ICIiCiAgICAgICAgICAgICAgICApIG9yICgKICAgICAgICAgICAgICAgICAgICBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lICE9ICIiCiAgICAgICAgICAgICAgICAgICAgYW5k
            IHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lICE9ICIiCiAgICAgICAgICAgICAg
            ICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVfbmFtZSAhPSAiIgogICAgICAgICAg
            ICAgICAgKToKICAgICAgICAgICAgICAgICAgICAoCiAgICAgICAgICAgICAgICAgICAgICAgIHNl
            bGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX0FDQ0VTU19LRVkiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJfU0VDUkVUX0tFWSJd
            LAogICAgICAgICAgICAgICAgICAgICAgICBpbmZvX2NhcF9zdXBwb3J0ZWQsCiAgICAgICAgICAg
            ICAgICAgICAgICAgIGVyciwKICAgICAgICAgICAgICAgICAgICApID0gc2VsZi5jcmVhdGVfcmd3
            X2FkbWluX29wc191c2VyKCkKICAgICAgICAgICAgICAgICAgICBlcnIgPSBzZWxmLnZhbGlkYXRl
            X3Jnd19lbmRwb2ludChpbmZvX2NhcF9zdXBwb3J0ZWQpCiAgICAgICAgICAgICAgICAgICAgaWYg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfdGxzX2NlcnRfcGF0aDoKICAgICAgICAgICAgICAgICAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJSR1dfVExTX0NFUlQiXSA9ICgKICAgICAgICAgICAgICAgICAgICAg
            ICAgICAgIHNlbGYudmFsaWRhdGVfcmd3X2VuZHBvaW50X3Rsc19jZXJ0KCkKICAgICAgICAgICAg
            ICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAgICMgaWYgdGhlcmUgaXMgbm8gZXJyb3Is
            IHNldCB0aGUgUkdXX0VORFBPSU5UCiAgICAgICAgICAgICAgICAgICAgaWYgZXJyICE9ICItMSI6
            CiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiUkdXX0VORFBPSU5UIl0gPSBz
            ZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludAogICAgICAgICAgICAgICAgZWxzZToKICAgICAg
            ICAgICAgICAgICAgICBlcnIgPSAiUGxlYXNlIHByb3ZpZGUgYWxsIHRoZSBSR1cgbXVsdGlzaXRl
            IHBhcmFtZXRlcnMgb3Igbm9uZSBvZiB0aGVtIgogICAgICAgICAgICAgICAgICAgIHN5cy5zdGRl
            cnIud3JpdGUoZXJyKQoKICAgIGRlZiBnZW5fc2hlbGxfb3V0KHNlbGYpOgogICAgICAgIHNlbGYu
            X2dlbl9vdXRwdXRfbWFwKCkKICAgICAgICBzaE91dElPID0gU3RyaW5nSU8oKQogICAgICAgIGZv
            ciBrLCB2IGluIHNlbGYub3V0X21hcC5pdGVtcygpOgogICAgICAgICAgICBpZiB2IGFuZCBrIG5v
            dCBpbiBzZWxmLl9leGNsdWRlZF9rZXlzOgogICAgICAgICAgICAgICAgc2hPdXRJTy53cml0ZShm
            ImV4cG9ydCB7a309e3Z9e0xJTkVTRVB9IikKICAgICAgICBzaE91dCA9IHNoT3V0SU8uZ2V0dmFs
            dWUoKQogICAgICAgIHNoT3V0SU8uY2xvc2UoKQogICAgICAgIHJldHVybiBzaE91dAoKICAgIGRl
            ZiBnZW5fanNvbl9vdXQoc2VsZik6CiAgICAgICAgc2VsZi5fZ2VuX291dHB1dF9tYXAoKQogICAg
            ICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuICIiCiAg
            ICAgICAganNvbl9vdXQgPSBbCiAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICJuYW1lIjog
            InJvb2stY2VwaC1tb24tZW5kcG9pbnRzIiwKICAgICAgICAgICAgICAgICJraW5kIjogIkNvbmZp
            Z01hcCIsCiAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAiZGF0
            YSI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9DRVBIX01PTl9EQVRBIl0sCiAgICAgICAg
            ICAgICAgICAgICAgIm1heE1vbklkIjogIjAiLAogICAgICAgICAgICAgICAgICAgICJtYXBwaW5n
            IjogInt9IiwKICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgIH0sCiAgICAgICAgICAgIHsK
            ICAgICAgICAgICAgICAgICJuYW1lIjogInJvb2stY2VwaC1tb24iLAogICAgICAgICAgICAgICAg
            ImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAg
            ICAgICAgICJhZG1pbi1zZWNyZXQiOiAiYWRtaW4tc2VjcmV0IiwKICAgICAgICAgICAgICAgICAg
            ICAiZnNpZCI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9GU0lEIl0sCiAgICAgICAgICAg
            ICAgICAgICAgIm1vbi1zZWNyZXQiOiAibW9uLXNlY3JldCIsCiAgICAgICAgICAgICAgICB9LAog
            ICAgICAgICAgICB9LAogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFtZSI6ICJyb29r
            LWNlcGgtb3BlcmF0b3ItY3JlZHMiLAogICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwK
            ICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICJ1c2VySUQiOiBz
            ZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUk5BTUUiXSwKICAgICAgICAgICAgICAgICAg
            ICAidXNlcktleSI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9VU0VSX1NFQ1JFVCJdLAog
            ICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgfSwKICAgICAgICBdCgogICAgICAgICMgaWYg
            J01PTklUT1JJTkdfRU5EUE9JTlQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAnbW9uaXRvcmluZy1l
            bmRwb2ludCcgdG8gQ2x1c3RlcgogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFw
            WyJNT05JVE9SSU5HX0VORFBPSU5UIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsiTU9O
            SVRPUklOR19FTkRQT0lOVF9QT1JUIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291dC5h
            cHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAibW9u
            aXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiQ2VwaENsdXN0
            ZXIiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAg
            ICAiTW9uaXRvcmluZ0VuZHBvaW50Ijogc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5U
            Il0sCiAgICAgICAgICAgICAgICAgICAgICAgICJNb25pdG9yaW5nUG9ydCI6IHNlbGYub3V0X21h
            cFsiTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUIl0sCiAgICAgICAgICAgICAgICAgICAgfSwKICAg
            ICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQoKICAgICAgICAjIGlmICdDU0lfUkJEX05PREVf
            U0VDUkVUJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ3Jvb2stY3NpLXJiZC1wcm92aXNpb25lcicg
            U2VjcmV0CiAgICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9E
            RV9TRUNSRVQiXQogICAgICAgICAgICBhbmQgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VD
            UkVUX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAg
            ICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9vay17c2VsZi5vdXRf
            bWFwWydDU0lfUkJEX05PREVfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICJr
            aW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAg
            ICAgICAgICAgICAgICJ1c2VySUQiOiBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9TRUNSRVRf
            TkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNlbGYub3V0X21hcFsi
            Q1NJX1JCRF9OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAg
            ICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdDU0lfUkJEX1BST1ZJU0lPTkVSX1NF
            Q1JFVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdyb29rLWNzaS1yYmQtcHJvdmlzaW9uZXInIFNl
            Y3JldAogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJ
            U0lPTkVSX1NFQ1JFVCJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9SQkRfUFJP
            VklTSU9ORVJfU0VDUkVUX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFw
            cGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9v
            ay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAg
            ICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJk
            YXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJD
            U0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAg
            ICJ1c2VyS2V5Ijogc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAog
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAg
            ICAgICAjIGlmICdDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCcgZXhpc3RzLCB0aGVuIG9u
            bHkgYWRkICdyb29rLWNzaS1jZXBoZnMtcHJvdmlzaW9uZXInIFNlY3JldAogICAgICAgIGlmICgK
            ICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJd
            CiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VD
            UkVUX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAg
            ICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9vay17c2VsZi5vdXRf
            bWFwWydDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAg
            ICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewog
            ICAgICAgICAgICAgICAgICAgICAgICAiYWRtaW5JRCI6IHNlbGYub3V0X21hcFsiQ1NJX0NFUEhG
            U19QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAiYWRt
            aW5LZXkiOiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUIl0sCiAg
            ICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAg
            ICAgICMgaWYgJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAn
            cm9vay1jc2ktY2VwaGZzLW5vZGUnIFNlY3JldAogICAgICAgIGlmICgKICAgICAgICAgICAgc2Vs
            Zi5vdXRfbWFwWyJDU0lfQ0VQSEZTX05PREVfU0VDUkVUIl0KICAgICAgICAgICAgYW5kIHNlbGYu
            b3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FIl0KICAgICAgICApOgogICAgICAg
            ICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAg
            ICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9O
            QU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAg
            ICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAiYWRtaW5JRCI6IHNl
            bGYub3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FIl0sCiAgICAgICAgICAgICAg
            ICAgICAgICAgICJhZG1pbktleSI6IHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JF
            VCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAg
            ICkKICAgICAgICAjIGlmICdST09LX0VYVEVSTkFMX0RBU0hCT0FSRF9MSU5LJyBleGlzdHMsIHRo
            ZW4gb25seSBhZGQgJ3Jvb2stY2VwaC1kYXNoYm9hcmQtbGluaycgU2VjcmV0CiAgICAgICAgaWYg
            c2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0RBU0hCT0FSRF9MSU5LIl06CiAgICAgICAgICAg
            IGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAi
            bmFtZSI6ICJyb29rLWNlcGgtZGFzaGJvYXJkLWxpbmsiLAogICAgICAgICAgICAgICAgICAgICJr
            aW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAg
            ICAgICAgICAgICAgICJ1c2VySUQiOiAiY2VwaC1kYXNoYm9hcmQtbGluayIsCiAgICAgICAgICAg
            ICAgICAgICAgICAgICJ1c2VyS2V5Ijogc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0RBU0hC
            T0FSRF9MSU5LIl0sCiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAg
            ICAgICAgICAgKQogICAgICAgICMgaWYgJ1JBRE9TX05BTUVTUEFDRScgZXhpc3RzLCB0aGVuIG9u
            bHkgYWRkIHRoZSAiUkFET1NfTkFNRVNQQUNFIiBuYW1lc3BhY2UKICAgICAgICBpZiAoCiAgICAg
            ICAgICAgIHNlbGYub3V0X21hcFsiUkFET1NfTkFNRVNQQUNFIl0KICAgICAgICAgICAgYW5kIHNl
            bGYub3V0X21hcFsiUkVTVFJJQ1RFRF9BVVRIX1BFUk1JU1NJT04iXQogICAgICAgICAgICBhbmQg
            bm90IHNlbGYub3V0X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdCiAgICAgICAgKToK
            ICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAg
            ICAgICAgICAgICJuYW1lIjogInJhZG9zLW5hbWVzcGFjZSIsCiAgICAgICAgICAgICAgICAgICAg
            ImtpbmQiOiAiQ2VwaEJsb2NrUG9vbFJhZG9zTmFtZXNwYWNlIiwKICAgICAgICAgICAgICAgICAg
            ICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgInJhZG9zTmFtZXNwYWNlTmFtZSI6
            IHNlbGYub3V0X21hcFsiUkFET1NfTkFNRVNQQUNFIl0sCiAgICAgICAgICAgICAgICAgICAgICAg
            ICJwb29sIjogc2VsZi5vdXRfbWFwWyJSQkRfUE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAg
            ICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICAgICBqc29uX291
            dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAi
            Y2VwaC1yYmQtcmFkb3MtbmFtZXNwYWNlIiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJT
            dG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAg
            ICAgICAgICAgICAicG9vbCI6IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAogICAgICAg
            ICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1u
            YW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05B
            TUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRy
            b2xsZXItZXhwYW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRf
            UFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3Np
            LnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0
            X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICB9
            LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgZWxzZToKICAgICAgICAg
            ICAgaWYgc2VsZi5vdXRfbWFwWyJSQkRfTUVUQURBVEFfRUNfUE9PTF9OQU1FIl06CiAgICAgICAg
            ICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICAgICAgewogICAgICAgICAg
            ICAgICAgICAgICAgICAibmFtZSI6ICJjZXBoLXJiZCIsCiAgICAgICAgICAgICAgICAgICAgICAg
            ICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAgICAgICAgICAgICJkYXRhIjog
            ewogICAgICAgICAgICAgICAgICAgICAgICAgICAgImRhdGFQb29sIjogc2VsZi5vdXRfbWFwWyJS
            QkRfUE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYu
            b3V0X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdLAogICAgICAgICAgICAgICAgICAg
            ICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNyZXQtbmFtZSI6IGYi
            cm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwK
            ICAgICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vY29udHJvbGxl
            ci1leHBhbmQtc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9QUk9W
            SVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiY3Np
            LnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0
            X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAg
            ICAgfSwKICAgICAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICAgICApCiAgICAgICAgICAg
            IGVsc2U6CiAgICAgICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICAg
            ICAgewogICAgICAgICAgICAgICAgICAgICAgICAibmFtZSI6ICJjZXBoLXJiZCIsCiAgICAgICAg
            ICAgICAgICAgICAgICAgICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAgICAg
            ICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAgICAgInBvb2wiOiBzZWxm
            Lm91dF9tYXBbIlJCRF9QT09MX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJj
            c2kuc3RvcmFnZS5rOHMuaW8vcHJvdmlzaW9uZXItc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYu
            b3V0X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAg
            ICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNl
            Y3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VD
            UkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4
            cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9S
            QkRfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgICAgIH0sCiAgICAg
            ICAgICAgICAgICAgICAgfQogICAgICAgICAgICAgICAgKQoKICAgICAgICAjIGlmICdDRVBIRlNf
            RlNfTkFNRScgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdjZXBoZnMnIFN0b3JhZ2VDbGFzcwogICAg
            ICAgIGlmIHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXToKICAgICAgICAgICAganNvbl9v
            dXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjog
            ImNlcGhmcyIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwKICAg
            ICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImZzTmFt
            ZSI6IHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAg
            ICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIkNFUEhGU19QT09MX05BTUUiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNyZXQtbmFtZSI6
            IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1F
            J119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250cm9s
            bGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfQ0VQSEZT
            X1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNz
            aS5zdG9yYWdlLms4cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91
            dF9tYXBbJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAg
            ICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICMgaWYgJ1JHV19F
            TkRQT0lOVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdjZXBoLXJndycgU3RvcmFnZUNsYXNzCiAg
            ICAgICAgaWYgc2VsZi5vdXRfbWFwWyJSR1dfRU5EUE9JTlQiXToKICAgICAgICAgICAganNvbl9v
            dXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjog
            ImNlcGgtcmd3IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAog
            ICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAiZW5k
            cG9pbnQiOiBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdLAogICAgICAgICAgICAgICAgICAg
            ICAgICAicG9vbFByZWZpeCI6IHNlbGYub3V0X21hcFsiUkdXX1BPT0xfUFJFRklYIl0sCiAgICAg
            ICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAg
            ICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAg
            ICAgIm5hbWUiOiAicmd3LWFkbWluLW9wcy11c2VyIiwKICAgICAgICAgICAgICAgICAgICAia2lu
            ZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAg
            ICAgICAgICAgICAiYWNjZXNzS2V5Ijogc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJf
            QUNDRVNTX0tFWSJdLAogICAgICAgICAgICAgICAgICAgICAgICAic2VjcmV0S2V5Ijogc2VsZi5v
            dXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJfU0VDUkVUX0tFWSJdLAogICAgICAgICAgICAgICAg
            ICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdSR1df
            VExTX0NFUlQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCB0aGUgImNlcGgtcmd3LXRscy1jZXJ0IiBz
            ZWNyZXQKICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VSVCJdOgogICAgICAgICAg
            ICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAg
            Im5hbWUiOiAiY2VwaC1yZ3ctdGxzLWNlcnQiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjog
            IlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAg
            ICAgICAgICJjZXJ0Ijogc2VsZi5vdXRfbWFwWyJSR1dfVExTX0NFUlQiXSwKICAgICAgICAgICAg
            ICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCgogICAgICAgIHJldHVy
            biBqc29uLmR1bXBzKGpzb25fb3V0KSArIExJTkVTRVAKCiAgICBkZWYgdXBncmFkZV91c2Vyc19w
            ZXJtaXNzaW9ucyhzZWxmKToKICAgICAgICB1c2VycyA9IFsKICAgICAgICAgICAgImNsaWVudC5j
            c2ktY2VwaGZzLW5vZGUiLAogICAgICAgICAgICAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9u
            ZXIiLAogICAgICAgICAgICAiY2xpZW50LmNzaS1yYmQtbm9kZSIsCiAgICAgICAgICAgICJjbGll
            bnQuY3NpLXJiZC1wcm92aXNpb25lciIsCiAgICAgICAgICAgICJjbGllbnQuaGVhbHRoY2hlY2tl
            ciIsCiAgICAgICAgXQogICAgICAgIGlmIHNlbGYucnVuX2FzX3VzZXIgIT0gIiIgYW5kIHNlbGYu
            cnVuX2FzX3VzZXIgbm90IGluIHVzZXJzOgogICAgICAgICAgICB1c2Vycy5hcHBlbmQoc2VsZi5y
            dW5fYXNfdXNlcikKICAgICAgICBmb3IgdXNlciBpbiB1c2VyczoKICAgICAgICAgICAgc2VsZi51
            cGdyYWRlX3VzZXJfcGVybWlzc2lvbnModXNlcikKCiAgICBkZWYgZ2V0X3Jnd19wb29sX25hbWVf
            ZHVyaW5nX3VwZ3JhZGUoc2VsZiwgdXNlciwgY2Fwcyk6CiAgICAgICAgaWYgdXNlciA9PSAiY2xp
            ZW50LmhlYWx0aGNoZWNrZXIiOgogICAgICAgICAgICAjIHdoZW4gYWRtaW4gaGFzIG5vdCBwcm92
            aWRlZCByZ3cgcG9vbCBuYW1lIGR1cmluZyB1cGdyYWRlLAogICAgICAgICAgICAjIGdldCB0aGUg
            cmd3IHBvb2wgbmFtZSBmcm9tIGNsaWVudC5oZWFsdGhjaGVja2VyIHVzZXIgd2hpY2ggd2FzIHVz
            ZWQgZHVyaW5nIGNvbm5lY3Rpb24KICAgICAgICAgICAgaWYgbm90IHNlbGYuX2FyZ19wYXJzZXIu
            cmd3X3Bvb2xfcHJlZml4OgogICAgICAgICAgICAgICAgIyBUbyBnZXQgdmFsdWUgJ2RlZmF1bHQn
            IHdoaWNoIGlzIHJndyBwb29sIG5hbWUgZnJvbSAnYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YScKICAgICAgICAgICAgICAgIHBhdHRlcm4gPSByInBvb2w9KC4qPylcLnJnd1wubWV0YSIK
            ICAgICAgICAgICAgICAgIG1hdGNoID0gcmUuc2VhcmNoKHBhdHRlcm4sIGNhcHMpCiAgICAgICAg
            ICAgICAgICBpZiBtYXRjaDoKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJn
            d19wb29sX3ByZWZpeCA9IG1hdGNoLmdyb3VwKDEpCiAgICAgICAgICAgICAgICBlbHNlOgogICAg
            ICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAg
            ICAgICAgICAgICAgICAgICJmYWlsZWQgdG8gZ2V0IHJndyBwb29sIG5hbWUgZm9yIHVwZ3JhZGUi
            CiAgICAgICAgICAgICAgICAgICAgKQoKICAgIGRlZiB1cGdyYWRlX3VzZXJfcGVybWlzc2lvbnMo
            c2VsZiwgdXNlcik6CiAgICAgICAgIyBjaGVjayB3aGV0aGVyIHRoZSBnaXZlbiB1c2VyIGV4aXN0
            cyBvciBub3QKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4IjogImF1dGggZ2V0IiwgImVudGl0
            eSI6IGYie3VzZXJ9IiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICByZXRfdmFsLCBqc29uX291
            dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAg
            aWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgcHJpbnQo
            ZiJ1c2VyIHt1c2VyfSBub3QgZm91bmQgZm9yIHVwZ3JhZGluZy4iKQogICAgICAgICAgICByZXR1
            cm4KICAgICAgICBleGlzdGluZ19jYXBzID0ganNvbl9vdXRbMF1bImNhcHMiXQogICAgICAgIHNl
            bGYuZ2V0X3Jnd19wb29sX25hbWVfZHVyaW5nX3VwZ3JhZGUodXNlciwgc3RyKGV4aXN0aW5nX2Nh
            cHMpKQogICAgICAgIG5ld19jYXAsIF8gPSBzZWxmLmdldF9jYXBzX2FuZF9lbnRpdHkodXNlcikK
            ICAgICAgICBjYXBfa2V5cyA9IFsibW9uIiwgIm1nciIsICJvc2QiLCAibWRzIl0KICAgICAgICBj
            YXBzID0gW10KICAgICAgICBmb3IgZWFjaENhcCBpbiBjYXBfa2V5czoKICAgICAgICAgICAgY3Vy
            X2NhcF92YWx1ZXMgPSBleGlzdGluZ19jYXBzLmdldChlYWNoQ2FwLCAiIikKICAgICAgICAgICAg
            bmV3X2NhcF92YWx1ZXMgPSBuZXdfY2FwLmdldChlYWNoQ2FwLCAiIikKICAgICAgICAgICAgY3Vy
            X2NhcF9wZXJtX2xpc3QgPSBbCiAgICAgICAgICAgICAgICB4LnN0cmlwKCkgZm9yIHggaW4gY3Vy
            X2NhcF92YWx1ZXMuc3BsaXQoIiwiKSBpZiB4LnN0cmlwKCkKICAgICAgICAgICAgXQogICAgICAg
            ICAgICBuZXdfY2FwX3Blcm1fbGlzdCA9IFsKICAgICAgICAgICAgICAgIHguc3RyaXAoKSBmb3Ig
            eCBpbiBuZXdfY2FwX3ZhbHVlcy5zcGxpdCgiLCIpIGlmIHguc3RyaXAoKQogICAgICAgICAgICBd
            CiAgICAgICAgICAgICMgYXBwZW5kIG5ld19jYXBfbGlzdCB0byBjdXJfY2FwX2xpc3QgdG8gbWFp
            bnRhaW4gdGhlIG9yZGVyIG9mIGNhcHMKICAgICAgICAgICAgY3VyX2NhcF9wZXJtX2xpc3QuZXh0
            ZW5kKG5ld19jYXBfcGVybV9saXN0KQogICAgICAgICAgICAjIGVsaW1pbmF0ZSBkdXBsaWNhdGVz
            IHdpdGhvdXQgdXNpbmcgJ3NldCcKICAgICAgICAgICAgIyBzZXQgcmUtb3JkZXJzIGl0ZW1zIGlu
            IHRoZSBsaXN0IGFuZCB3ZSBoYXZlIHRvIGtlZXAgdGhlIG9yZGVyCiAgICAgICAgICAgIG5ld19j
            YXBfbGlzdCA9IFtdCiAgICAgICAgICAgIFtuZXdfY2FwX2xpc3QuYXBwZW5kKHgpIGZvciB4IGlu
            IGN1cl9jYXBfcGVybV9saXN0IGlmIHggbm90IGluIG5ld19jYXBfbGlzdF0KICAgICAgICAgICAg
            ZXhpc3RpbmdfY2Fwc1tlYWNoQ2FwXSA9ICIsICIuam9pbihuZXdfY2FwX2xpc3QpCiAgICAgICAg
            ICAgIGlmIGV4aXN0aW5nX2NhcHNbZWFjaENhcF06CiAgICAgICAgICAgICAgICBjYXBzLmFwcGVu
            ZChlYWNoQ2FwKQogICAgICAgICAgICAgICAgY2Fwcy5hcHBlbmQoZXhpc3RpbmdfY2Fwc1tlYWNo
            Q2FwXSkKICAgICAgICBjbWRfanNvbiA9IHsKICAgICAgICAgICAgInByZWZpeCI6ICJhdXRoIGNh
            cHMiLAogICAgICAgICAgICAiZW50aXR5IjogdXNlciwKICAgICAgICAgICAgImNhcHMiOiBjYXBz
            LAogICAgICAgICAgICAiZm9ybWF0IjogImpzb24iLAogICAgICAgIH0KICAgICAgICByZXRfdmFs
            LCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24p
            CiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiInYXV0aCBjYXBzIHt1c2VyfScgY29tbWFu
            ZCBmYWlsZWQuXG4gRXJyb3I6IHtlcnJfbXNnfSIKICAgICAgICAgICAgKQogICAgICAgIHByaW50
            KGYiVXBkYXRlZCB1c2VyIHt1c2VyfSBzdWNjZXNzZnVsbHkuIikKCiAgICBkZWYgbWFpbihzZWxm
            KToKICAgICAgICBnZW5lcmF0ZWRfb3V0cHV0ID0gIiIKICAgICAgICBpZiBzZWxmLl9hcmdfcGFy
            c2VyLnVwZ3JhZGU6CiAgICAgICAgICAgIHNlbGYudXBncmFkZV91c2Vyc19wZXJtaXNzaW9ucygp
            CiAgICAgICAgZWxpZiBzZWxmLl9hcmdfcGFyc2VyLmZvcm1hdCA9PSAianNvbiI6CiAgICAgICAg
            ICAgIGdlbmVyYXRlZF9vdXRwdXQgPSBzZWxmLmdlbl9qc29uX291dCgpCiAgICAgICAgZWxpZiBz
            ZWxmLl9hcmdfcGFyc2VyLmZvcm1hdCA9PSAiYmFzaCI6CiAgICAgICAgICAgIGdlbmVyYXRlZF9v
            dXRwdXQgPSBzZWxmLmdlbl9zaGVsbF9vdXQoKQogICAgICAgIGVsc2U6CiAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlVuc3VwcG9y
            dGVkIGZvcm1hdDoge3NlbGYuX2FyZ19wYXJzZXIuZm9ybWF0fSIKICAgICAgICAgICAgKQogICAg
            ICAgIHByaW50KGdlbmVyYXRlZF9vdXRwdXQpCiAgICAgICAgaWYgc2VsZi5vdXRwdXRfZmlsZSBh
            bmQgZ2VuZXJhdGVkX291dHB1dDoKICAgICAgICAgICAgZk91dCA9IG9wZW4oc2VsZi5vdXRwdXRf
            ZmlsZSwgbW9kZT0idyIsIGVuY29kaW5nPSJVVEYtOCIpCiAgICAgICAgICAgIGZPdXQud3JpdGUo
            Z2VuZXJhdGVkX291dHB1dCkKICAgICAgICAgICAgZk91dC5jbG9zZSgpCgoKIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyBNQUlOICMjIyMjIyMjIyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMKaWYgX19uYW1lX18gPT0gIl9fbWFpbl9fIjoKICAgIHJqT2Jq
            ID0gUmFkb3NKU09OKCkKICAgIHRyeToKICAgICAgICByak9iai5tYWluKCkKICAgIGV4Y2VwdCBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uIGFzIGVycjoKICAgICAgICBwcmludChmIkV4ZWN1dGlv
            biBGYWlsZWQ6IHtlcnJ9IikKICAgICAgICByYWlzZSBlcnIKICAgIGV4Y2VwdCBLZXlFcnJvciBh
            cyBrRXJyOgogICAgICAgIHByaW50KGYiS2V5RXJyb3I6IHtrRXJyfSIpCiAgICBleGNlcHQgT1NF
            cnJvciBhcyBvc0VycjoKICAgICAgICBwcmludChmIkVycm9yIHdoaWxlIHRyeWluZyB0byBvdXRw
            dXQgdGhlIGRhdGE6IHtvc0Vycn0iKQogICAgZmluYWxseToKICAgICAgICByak9iai5zaHV0ZG93
            bigpCg==
          external.features.ocs.openshift.io/supported-platforms: '["BareMetal", "None",
            "VSphere", "OpenStack", "oVirt"]'
          external.features.ocs.openshift.io/validation: '{"secrets":["rook-ceph-operator-creds",
            "rook-csi-rbd-node", "rook-csi-rbd-provisioner"], "configMaps": ["rook-ceph-mon-endpoints",
            "rook-ceph-mon"], "storageClasses": ["ceph-rbd"], "cephClusters": ["monitoring-endpoint"]}'
          features.ocs.openshift.io/disabled: '["ss-list", "install-wizard", "block-pool",
            "mcg-resource", "odf-dashboard", "common", "storage-provider", "storage-provisioner",
            "dashboard-resources", "csv-actions", "inventory-item", "alert-actions"]'
          features.ocs.openshift.io/enabled: '["kms", "arbiter", "flexible-scaling",
            "multus", "pool-management", "namespace-store", "mcg-standalone", "taint-nodes",
            "vault-sa-kms", "hpcs-kms"]'
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operatorframework.io/initialization-resource: "\n    {\n        \"apiVersion\":
            \"ocs.openshift.io/v1\",\n        \"kind\": \"StorageCluster\",\n        \"metadata\":
            {\n            \"name\": \"example-storagecluster\",\n            \"namespace\":
            \"openshift-storage\"\n        },\n        \"spec\": {\n            \"manageNodes\":
            false,\n            \"monPVCTemplate\": {\n                \"spec\": {\n
            \                   \"accessModes\": [\n                        \"ReadWriteOnce\"\n
            \                   ],\n                    \"resources\": {\n                        \"requests\":
            {\n                            \"storage\": \"10Gi\"\n                        }\n
            \                   },\n                    \"storageClassName\": \"gp2-csi\"\n
            \               }\n            },\n            \"storageDeviceSets\":
            [\n                {\n                    \"count\": 3,\n                    \"dataPVCTemplate\":
            {\n                        \"spec\": {\n                            \"accessModes\":
            [\n                                \"ReadWriteOnce\"\n                            ],\n
            \                           \"resources\": {\n                                \"requests\":
            {\n                                    \"storage\": \"1Ti\"\n                                }\n
            \                           },\n                            \"storageClassName\":
            \"gp2-csi\",\n                            \"volumeMode\": \"Block\"\n
            \                       }\n                    },\n                    \"name\":
            \"example-deviceset\",\n                    \"placement\": {},\n                    \"portable\":
            true,\n                    \"resources\": {}\n                }\n            ]\n
            \       }\n    }\n\t"
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:03e50284f7fe090c57eddf90feea8ff48d78aff8ce95d1b297aa91e71c3ee577
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.4
          operators.operatorframework.io/internal-objects: '["ocsinitializations.ocs.openshift.io","storageclassrequests.ocs.openshift.io","storageconsumers.ocs.openshift.io","storageprofiles.ocs.openshift.io","cephclusters.ceph.rook.io","cephobjectstores.ceph.rook.io","cephobjectstoreusers.ceph.rook.io","cephnfses.ceph.rook.io","cephclients.ceph.rook.io","cephfilesystems.ceph.rook.io","cephfilesystemmirrors.ceph.rook.io","cephrbdmirrors.ceph.rook.io","cephobjectrealms.ceph.rook.io","cephobjectzonegroups.ceph.rook.io","cephobjectzones.ceph.rook.io","cephbucketnotifications.ceph.rook.io","cephbuckettopics.ceph.rook.io","cephfilesystemsubvolumegroups.ceph.rook.io","cephblockpoolradosnamespaces.ceph.rook.io","cephcosidrivers.ceph.rook.io"]'
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/ocs-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OCSInitialization represents the initial data to be created
              when the operator is installed.
            displayName: OCS Initialization
            kind: OCSInitialization
            name: ocsinitializations.ocs.openshift.io
            version: v1
          - description: StorageClassRequest is the Schema for the StorageClassRequests
              API
            displayName: Storage Class Request
            kind: StorageClassRequest
            name: storageclassrequests.ocs.openshift.io
            version: v1alpha1
          - description: StorageCluster represents a cluster including Ceph Cluster,
              NooBaa and all the storage and compute resources required.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.ocs.openshift.io
            version: v1
          - description: StorageConsumer is the Schema for the storageconsumers API
            displayName: Storage Consumer
            kind: StorageConsumer
            name: storageconsumers.ocs.openshift.io
            version: v1alpha1
          - kind: StorageProfile
            name: storageprofiles.ocs.openshift.io
            version: v1
          - description: Represents a Ceph cluster.
            displayName: Ceph Cluster
            kind: CephCluster
            name: cephclusters.ceph.rook.io
            version: v1
          - description: Represents a Ceph Block Pool.
            displayName: Block Pools
            kind: CephBlockPool
            name: cephblockpools.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store.
            displayName: Ceph Object Store
            kind: CephObjectStore
            name: cephobjectstores.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store User.
            displayName: Ceph Object Store User
            kind: CephObjectStoreUser
            name: cephobjectstoreusers.ceph.rook.io
            version: v1
          - description: Represents a cluster of Ceph NFS ganesha gateways.
            displayName: Ceph NFS
            kind: CephNFS
            name: cephnfses.ceph.rook.io
            version: v1
          - description: Represents a Ceph User.
            displayName: Ceph Client
            kind: CephClient
            name: cephclients.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem.
            displayName: Ceph Filesystem
            kind: CephFilesystem
            name: cephfilesystems.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem Mirror.
            displayName: Ceph Filesystem Mirror
            kind: CephFilesystemMirror
            name: cephfilesystemmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph RBD Mirror.
            displayName: Ceph RBD Mirror
            kind: CephRBDMirror
            name: cephrbdmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Realm.
            displayName: Ceph Object Store Realm
            kind: CephObjectRealm
            name: cephobjectrealms.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone Group.
            displayName: Ceph Object Store Zone Group
            kind: CephObjectZoneGroup
            name: cephobjectzonegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone.
            displayName: Ceph Object Store Zone
            kind: CephObjectZone
            name: cephobjectzones.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Notification.
            displayName: Ceph Bucket Notification
            kind: CephBucketNotification
            name: cephbucketnotifications.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Topic.
            displayName: Ceph Bucket Topic
            kind: CephBucketTopic
            name: cephbuckettopics.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem SubVolumeGroup.
            displayName: Ceph Filesystem SubVolumeGroup
            kind: CephFilesystemSubVolumeGroup
            name: cephfilesystemsubvolumegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph BlockPool Rados Namespace.
            displayName: Ceph BlockPool Rados Namespace
            kind: CephBlockPoolRadosNamespace
            name: cephblockpoolradosnamespaces.ceph.rook.io
            version: v1
          - description: Represents a Ceph COSI Driver.
            displayName: Ceph COSI Driver
            kind: CephCOSIDriver
            name: cephcosidrivers.ceph.rook.io
            version: v1
        description: |
          **Red Hat OpenShift Container Storage** deploys three operators.

          ### OpenShift Container Storage operator

          The OpenShift Container Storage operator is the primary operator for OpenShift Container Storage. It serves to facilitate the other operators in OpenShift Container Storage by performing administrative tasks outside their scope as well as watching and configuring their CustomResources.

          ### Rook

          [Rook][1] deploys and manages Ceph on OpenShift, which provides block and file storage.

          # Core Capabilities

          * **Self-managing service:** No matter which supported storage technologies you choose, OpenShift Container Storage ensures that resources can be deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Container Storage you can either build dedicated storage clusters or hyper-converged clusters where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Container Storage:** OpenShift Container Storage integrates Ceph with multiple storage presentations including object storage (compatible with S3), block storage, and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Container Storage efficiently distributes and replicates your data across your cluster to minimize the risk of data loss. With snapshots, cloning, and versioning, no more losing sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data storage for your workloads, using a single namespace.

          [1]: https://rook.io
        displayName: OpenShift Container Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - rook
        - ceph
        - block storage
        - shared filesystem
        - object storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ocs-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:1711a0aabd28937aac1424adebe96619a48b91f7f888f0588117f80891001da4
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:f37e2c611f4cda32172d4e8938938e488076fd8a8f19c252c56a80c48044bc02
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:028120f2212738917f49e0330b60be5f79a208ce2974bea22b71f533aa6ce3be
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:f75a12fc0c449008028d7fc8e18ca7c1afc37ba585885ac754fed860047794a0
        - registry.redhat.io/odf4/ocs-metrics-exporter-rhel9@sha256:c58c67263c08d84a68c8e45cfd77d59a35aee2151bf79b44377016ca0fb429ee
        - registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:9a0008f662b6afa9e2677a8a83bd662d806ad6f4d03fb49bccb78ed487ce4c7a
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:645d4fd09628f7ddcfc765e4ddacaa79fa6a7c35c03292ca2f6e317d48765d6b
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:ef4276b80cacae825d4831de88d3fe6ba7cc324464087b7b28734bd2b8de3315
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:0dc0733df67bb49886ac9715b81bd4ec4a36330fe1afdde657587e1883f5ca7a
        - registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:03e50284f7fe090c57eddf90feea8ff48d78aff8ce95d1b297aa91e71c3ee577
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:ffcc0806ec2afb8b40d75c70ef2a6d98bb887879f5b3f573e900861316ce9e41
        - registry.redhat.io/odf4/ocs-rhel9-operator@sha256:b1a9492bd9b314f86b4e192cf2c83144885ff863d9932e386a41691f50be6193
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:f8bfbf64b4bd030474e239aed5e03af9dfe920eeae444c5cc8ac3dcfa640fcc1
        version: 4.15.20-rhodf
      entries:
      - name: ocs-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: ocs-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: ocs-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: ocs-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: ocs-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: ocs-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: ocs-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: ocs-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: ocs-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: ocs-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: ocs-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: ocs-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: ocs-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: ocs-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: ocs-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: ocs-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: ocs-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: ocs-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: ocs-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: ocs-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: ocs-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |2-

            [
                {
                    "apiVersion": "ocs.openshift.io/v1",
                    "kind": "StorageCluster",
                    "metadata": {
                        "name": "example-storagecluster",
                        "namespace": "openshift-storage"
                    },
                    "spec": {
                        "manageNodes": false,
                        "monPVCTemplate": {
                            "spec": {
                                "accessModes": [
                                    "ReadWriteOnce"
                                ],
                                "resources": {
                                    "requests": {
                                        "storage": "10Gi"
                                    }
                                },
                                "storageClassName": "gp2-csi"
                            }
                        },
                        "storageDeviceSets": [
                            {
                                "count": 3,
                                "dataPVCTemplate": {
                                    "spec": {
                                        "accessModes": [
                                            "ReadWriteOnce"
                                        ],
                                        "resources": {
                                            "requests": {
                                                "storage": "1Ti"
                                            }
                                        },
                                        "storageClassName": "gp2-csi",
                                        "volumeMode": "Block"
                                    }
                                },
                                "name": "example-deviceset",
                                "placement": {},
                                "portable": true,
                                "resources": {}
                            }
                        ]
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Storage
          containerImage: registry.redhat.io/odf4/ocs-rhel9-operator@sha256:3ef1c2679e9229d51ebc04bd2f2b22ce9b11e8f7fdd7bddee69630fe5b1de47c
          description: Red Hat OpenShift Container Storage provides hyperconverged
            storage for applications within an OpenShift cluster.
          external.features.ocs.openshift.io/supported-platforms: '["BareMetal", "None",
            "VSphere", "OpenStack", "oVirt"]'
          external.features.ocs.openshift.io/validation: '{"secrets":["rook-ceph-operator-creds",
            "rook-csi-rbd-node", "rook-csi-rbd-provisioner"], "configMaps": ["rook-ceph-mon-endpoints",
            "rook-ceph-mon"], "storageClasses": ["ceph-rbd"], "cephClusters": ["monitoring-endpoint"]}'
          features.ocs.openshift.io/disabled: '["ss-list", "install-wizard", "block-pool",
            "mcg-resource", "odf-dashboard", "common", "storage-provider", "storage-provisioner",
            "dashboard-resources", "csv-actions", "inventory-item", "alert-actions"]'
          features.ocs.openshift.io/enabled: '["kms", "arbiter", "flexible-scaling",
            "multus", "pool-management", "namespace-store", "mcg-standalone", "taint-nodes",
            "vault-sa-kms", "hpcs-kms"]'
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operatorframework.io/initialization-resource: "\n    {\n        \"apiVersion\":
            \"ocs.openshift.io/v1\",\n        \"kind\": \"StorageCluster\",\n        \"metadata\":
            {\n            \"name\": \"example-storagecluster\",\n            \"namespace\":
            \"openshift-storage\"\n        },\n        \"spec\": {\n            \"manageNodes\":
            false,\n            \"monPVCTemplate\": {\n                \"spec\": {\n
            \                   \"accessModes\": [\n                        \"ReadWriteOnce\"\n
            \                   ],\n                    \"resources\": {\n                        \"requests\":
            {\n                            \"storage\": \"10Gi\"\n                        }\n
            \                   },\n                    \"storageClassName\": \"gp2-csi\"\n
            \               }\n            },\n            \"storageDeviceSets\":
            [\n                {\n                    \"count\": 3,\n                    \"dataPVCTemplate\":
            {\n                        \"spec\": {\n                            \"accessModes\":
            [\n                                \"ReadWriteOnce\"\n                            ],\n
            \                           \"resources\": {\n                                \"requests\":
            {\n                                    \"storage\": \"1Ti\"\n                                }\n
            \                           },\n                            \"storageClassName\":
            \"gp2-csi\",\n                            \"volumeMode\": \"Block\"\n
            \                       }\n                    },\n                    \"name\":
            \"example-deviceset\",\n                    \"placement\": {},\n                    \"portable\":
            true,\n                    \"resources\": {}\n                }\n            ]\n
            \       }\n    }\n\t"
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:b7758b197cf423573de96bdca0cb3045852b20f34fc9c17ad1632b9c3a3cf3d0
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.4
          operators.operatorframework.io/internal-objects: '["ocsinitializations.ocs.openshift.io","storageconsumers.ocs.openshift.io","storageprofiles.ocs.openshift.io","storagerequests.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/ocs-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OCSInitialization represents the initial data to be created
              when the operator is installed.
            displayName: OCS Initialization
            kind: OCSInitialization
            name: ocsinitializations.ocs.openshift.io
            version: v1
          - description: StorageCluster represents a cluster including Ceph Cluster,
              NooBaa and all the storage and compute resources required.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.ocs.openshift.io
            version: v1
          - description: StorageConsumer is the Schema for the storageconsumers API
            displayName: Storage Consumer
            kind: StorageConsumer
            name: storageconsumers.ocs.openshift.io
            version: v1alpha1
          - description: StorageProfile is the Schema for the storageprofiles API
            displayName: Storage Profile
            kind: StorageProfile
            name: storageprofiles.ocs.openshift.io
            version: v1
          - description: StorageRequest is the Schema for the StorageRequests API
            displayName: Storage Request
            kind: StorageRequest
            name: storagerequests.ocs.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Container Storage** deploys three operators.

          ### OpenShift Container Storage operator

          The OpenShift Container Storage operator is the primary operator for OpenShift Container Storage. It serves to facilitate the other operators in OpenShift Container Storage by performing administrative tasks outside their scope as well as watching and configuring their CustomResources.

          ### Rook

          [Rook][1] deploys and manages Ceph on OpenShift, which provides block and file storage.

          # Core Capabilities

          * **Self-managing service:** No matter which supported storage technologies you choose, OpenShift Container Storage ensures that resources can be deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Container Storage you can either build dedicated storage clusters or hyper-converged clusters where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Container Storage:** OpenShift Container Storage integrates Ceph with multiple storage presentations including object storage (compatible with S3), block storage, and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Container Storage efficiently distributes and replicates your data across your cluster to minimize the risk of data loss. With snapshots, cloning, and versioning, no more losing sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data storage for your workloads, using a single namespace.

          [1]: https://rook.io
        displayName: OpenShift Container Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - rook
        - ceph
        - block storage
        - shared filesystem
        - object storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ocs-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d
        - registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:b7758b197cf423573de96bdca0cb3045852b20f34fc9c17ad1632b9c3a3cf3d0
        - registry.redhat.io/odf4/ocs-rhel9-operator@sha256:3ef1c2679e9229d51ebc04bd2f2b22ce9b11e8f7fdd7bddee69630fe5b1de47c
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:70a3c6bf4bde84c7e43fd58c60abc5612d25c792042b902f6abee94bfaa63f80
        - registry.redhat.io/odf4/ocs-metrics-exporter-rhel9@sha256:3332a93e3ea2e83af15a7edab02a9e00d70e81e44338b63a0eacdea54142ad00
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:0084d9053ad2c74926ca02d2ede0461abd59227eee98d122c45e08a5ebaf629a
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:cf7f72469d25fb37301a8159406613673f8968c82af1035fd45a470e0ae8acfa
        version: 4.16.16-rhodf
      entries:
      - name: ocs-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: ocs-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: ocs-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: ocs-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: ocs-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: ocs-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: ocs-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: ocs-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: ocs-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: ocs-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: ocs-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: ocs-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: ocs-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: ocs-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: ocs-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: ocs-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: ocs-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: ocs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-pipes-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-pipes-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pipes.services.k8s.aws/v1alpha1",
                "kind": "Pipe",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/pipes-controller:1.1.1
          createdAt: "2025-09-19T21:15:33Z"
          description: AWS EventBridge Pipes controller is a service controller for
            managing Pipes resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pipe represents the state of an AWS pipes Pipe resource.
            displayName: Pipe
            kind: Pipe
            name: pipes.pipes.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EventBridge Pipes resources in AWS from within your Kubernetes cluster.

          **About Amazon EventBridge Pipes**

          Amazon EventBridge Pipes connects sources to targets. Pipes are intended for point-to-point integrations between supported [sources](https://docs.aws.amazon.com/eventbridge/latest/userguide/eb-pipes-event-source.html) and [targets](https://docs.aws.amazon.com/eventbridge/latest/userguide/eb-pipes-event-target.html), with support for advanced transformations and [enrichment](https://docs.aws.amazon.com/eventbridge/latest/userguide/pipes-enrichment.html). It reduces the need for specialized knowledge and integration code when developing event-driven architectures, fostering consistency across your company’s applications. To set up a pipe, you choose the source, add optional filtering, define optional enrichment, and choose the target for the event data.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EventBridge Pipes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pipes
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EventBridge Pipes Developer Resources
          url: https://aws.amazon.com/eventbridge/pipes/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: pipes maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/pipes-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-pipes-controller.v1.1.1
        version: 1.1.1
      - name: ack-pipes-controller.v1.1.0
        version: 1.1.0
      - name: ack-pipes-controller.v1.0.25
        version: 1.0.25
      - name: ack-pipes-controller.v1.0.24
        version: 1.0.24
      - name: ack-pipes-controller.v1.0.23
        version: 1.0.23
      - name: ack-pipes-controller.v1.0.22
        version: 1.0.22
      - name: ack-pipes-controller.v1.0.21
        version: 1.0.21
      - name: ack-pipes-controller.v1.0.20
        version: 1.0.20
      - name: ack-pipes-controller.v1.0.18
        version: 1.0.18
      - name: ack-pipes-controller.v1.0.17
        version: 1.0.17
      - name: ack-pipes-controller.v1.0.16
        version: 1.0.16
      - name: ack-pipes-controller.v1.0.15
        version: 1.0.15
      - name: ack-pipes-controller.v1.0.14
        version: 1.0.14
      - name: ack-pipes-controller.v1.0.13
        version: 1.0.13
      - name: ack-pipes-controller.v1.0.12
        version: 1.0.12
      - name: ack-pipes-controller.v1.0.11
        version: 1.0.11
      - name: ack-pipes-controller.v1.0.10
        version: 1.0.10
      - name: ack-pipes-controller.v1.0.9
        version: 1.0.9
      - name: ack-pipes-controller.v1.0.8
        version: 1.0.8
      - name: ack-pipes-controller.v1.0.6
        version: 1.0.6
      - name: ack-pipes-controller.v1.0.3
        version: 1.0.3
      - name: ack-pipes-controller.v1.0.2
        version: 1.0.2
      - name: ack-pipes-controller.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-pipes-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Sigstore Authors
      provider-url: https://sigstore.dev
    name: sigstore-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sigstore-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.sigstore.dev/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "name": "rekor"
                },
                "spec": {
                  "server": {
                    "ingress": {
                      "hostname": "rekor.example.com"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          containerImage: ghcr.io/sigstore/sigstore-helm-operator@sha256:d9c6eb162b97a1aa1ad3bf515db65d01a96e55b6f8bcd646864d3e64a3ee2b9e
          createdAt: "2021-09-01T21:20:28Z"
          description: Helm Charts for Sigstore Project.
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/sigstore/sigstore-helm-operator
          support: The Sigstore Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Rekor
            name: rekors.helm.sigstore.dev
            version: v1alpha1
        description: |-
          Helm Charts for the sigstore project

          ## Charts

          The following charts are available:

          * [Rekor](https://github.com/sigstore/helm-charts/tree/main/charts/rekor)
        displayName: sigstore
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        links:
        - name: Sigstore Helm Operator
          url: https://github.com/sabre1041/sigstore-helm-operator
        maintainers:
        - email: ablock@redhat.com
          name: Andrew Block
        maturity: alpha
        provider:
          name: The Sigstore Authors
          url: https://sigstore.dev
        relatedImages:
        - quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - ghcr.io/sigstore/sigstore-helm-operator@sha256:d9c6eb162b97a1aa1ad3bf515db65d01a96e55b6f8bcd646864d3e64a3ee2b9e
        version: 0.0.1
      entries:
      - name: sigstore-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: sigstore-helm-operator
    provider:
      name: The Sigstore Authors
      url: https://sigstore.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Loki SIG Operator
      provider-url: ""
    name: loki-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: docker.io/grafana/loki-operator:0.8.0
          createdAt: "2025-03-17T11:00:08Z"
          description: The Community Loki Operator provides Kubernetes native deployment
            and management of Loki and related logging components.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/loki/tree/main/operator
          support: Grafana Loki SIG Operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |-
          The Community Loki Operator provides Kubernetes native deployment and management of Loki and related logging components.
          The purpose of this project is to simplify and automate the configuration of a Loki based logging stack for Kubernetes clusters.

          ### Operator features

          The Loki Operator includes, but is not limited to, the following features:

          * Kubernetes Custom Resources: Use Kubernetes custom resources to deploy and manage Loki, Alerting rules, Recording rules, and related components.
          * Simplified Deployment Configuration: Configure the fundamentals of Loki like tenants, limits, replication factor and storage from a native Kubernetes resource.

          ### Feature Gates

          The Loki Operator Bundle provides a set of feature gates that enable/disable specific feature depending on the target Kubernetes distribution. The following feature gates are enabled by default:
          * `serviceMonitors`: Enables creating a Prometheus-Operator managed ServiceMonitor resource per LokiStack component.
          * `serviceMonitorTlsEndpoints`: Enables TLS for the ServiceMonitor endpoints.
          * `lokiStackAlerts`: Enables creating PrometheusRules for common Loki alerts.
          * `httpEncryption`: Enables TLS encryption for all HTTP LokiStack services.
          * `grpcEncryption`: Enables TLS encryption for all GRPC LokiStack services.
          * `builtInCertManagement`: Enables the built-in facility for generating and rotating TLS client and serving certificates for all LokiStack services and internal clients
          * `lokiStackGateway`: Enables reconciling the reverse-proxy lokistack-gateway component for multi-tenant authentication/authorization traffic control to Loki.
          * `runtimeSeccompProfile`: Enables the restricted seccomp profile on all Lokistack components.
          * `defaultNodeAffinity`: Enable the operator will set a default node affinity on all pods. This will limit scheduling of the pods to Nodes with Linux.
          * `lokiStackWebhook`: Enables the LokiStack CR validation and conversion webhooks.
          * `alertingRuleWebhook`: Enables the AlertingRule CR validation webhook.
          * `recordingRuleWebhook`: Enables the RecordingRule CR validation webhook.
          * `rulerConfigWebhook`: Enables the RulerConfig CR validation webhook.

          In addition it enables the following OpenShift-only related feature gates:
          * `servingCertsService`: Enables OpenShift ServiceCA annotations on the lokistack-gateway service only.
          * `ruleExtendedValidation`: Enables extended validation of AlertingRule and RecordingRule to enforce tenancy in an OpenShift context.
          * `clusterTLSPolicy`: Enables usage of TLS policies set in the API Server.
          * `clusterProxy`: Enables usage of the proxy variables set in the proxy resource.

          ### Before you start

          1. Ensure that the appropriate object storage solution, that will be used by Loki, is avaliable and configured.
        displayName: Community Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Documentation
          url: https://loki-operator.dev/
        maintainers:
        - email: loki-operator-team@googlegroups.com
          name: Grafana Loki SIG Operator
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Grafana Loki SIG Operator
        relatedImages:
        - docker.io/grafana/loki-operator:0.8.0
        - quay.io/openshift/origin-kube-rbac-proxy:latest
        - quay.io/observatorium/api:latest
        - quay.io/observatorium/opa-openshift:latest
        - docker.io/grafana/loki:3.4.2
        version: 0.8.0
      entries:
      - name: loki-operator.v0.8.0
        version: 0.8.0
      - name: loki-operator.v0.7.1
        version: 0.7.1
      - name: loki-operator.v0.7.0
        version: 0.7.0
      - name: loki-operator.v0.6.2
        version: 0.6.2
      - name: loki-operator.v0.6.1
        version: 0.6.1
      - name: loki-operator.v0.6.0
        version: 0.6.0
      - name: loki-operator.v0.5.0
        version: 0.5.0
      - name: loki-operator.v0.4.0
        version: 0.4.0
      - name: loki-operator.v0.3.0
        version: 0.3.0
      - name: loki-operator.v0.2.0
        version: 0.2.0
      - name: loki-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: loki-operator
    provider:
      name: Grafana Loki SIG Operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-license-service-reporter-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-license-service-reporter-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "IBMLicenseServiceReporter",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-license-service-reporter-operator",
                    "app.kubernetes.io/instance": "ibmlicenseservicereporter-instance",
                    "app.kubernetes.io/managed-by": "operator",
                    "app.kubernetes.io/name": "ibmlicenseservicereporter",
                    "app.kubernetes.io/part-of": "ibm-license-service-reporter-operator"
                  },
                  "name": "instance"
                },
                "spec": {
                  "authentication": {
                    "useradmin": {
                      "enabled": true
                    }
                  },
                  "license": {
                    "accept": false
                  },
                  "version": "4.2.14"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:277e2ab849907553e4992c2dccc11f03ae1e0302f20aae48c620968aa97d6afc
          createdAt: "2025-02-26T12:10:52Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <4.2.14'
          operatorframework.io/suggested-namespace: ibm-ls-reporter
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "IBM License Service Reporter is a singleton which may be
              used in a single or multi-cluster environment. It aggregates data pushed
              from IBM License Services, deployed on clusters and from ILMT. \n Documentation:
              https://ibm.biz/lsvc-rprtr. \n License: Please refer to the IBM Terms
              website (ibm.biz/lsvc-lic) to find the license terms for the particular
              IBM product for which you are deploying this component. \n IBM License
              Service Reporter is a free, optionally installed add-on – one of services
              of Cloud Pak Foundational Services. Thanks to the IBM License Service
              Reporter customer can: \n - see IBM software deployments on a single
              dashboard, verify and maintain license compliance and avoid audit infractions,
              \n - see historical data on IBM software deployments to support making
              informed decisions for future purchases, \n - see details of software
              deployments, including source (ILMT for VMs, License Service for containers),
              cluster and quantities of licenses deployed, \n - identify which software
              deployments are VMs vs. containers and can use this information to evaluate
              workloads for modernization."
            displayName: IBMLicense Service Reporter
            kind: IBMLicenseServiceReporter
            name: ibmlicenseservicereporters.operator.ibm.com
            version: v1alpha1
        description: |-
          IBM License Service Reporter is a free, optionally installed add-on – one of services of Cloud Pak Foundational Services. Thanks to the IBM License Service Reporter customer can:
           * see IBM software deployments on a single dashboard, verify and maintain license compliance and avoid audit infractions,
           * see historical data on IBM software deployments to support making informed decisions for future purchases,
           * see details of software deployments, including source (ILMT for VMs, License Service for containers), cluster and quantities of licenses deployed,
           * identify which software deployments are VMs vs. containers and can use this information to evaluate workloads for modernization.

          IBM License Service Reporter is a singleton which may be used in a single or multi-cluster environment. It aggregates data pushed from IBM License Services, deployed on clusters and from ILMT.

          **Supported platforms**

          Red Hat OpenShift Container Platform 4.10 or newer installed on Linux x86_64.

          **Prerequisites**

          Prerequisites depend on the integration of the License Service Reporter with an IBM Cloud Pak. For more information, see the applicable IBM Cloud Pak documentation.
        displayName: IBM License Service Reporter
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - licensing
        - usage
        - IBMLicenseServiceReporter
        - IBM
        - Cloud
        - License
        - License Service
        - Service
        - Reporter
        - License Service Reporter
        - License Reporter
        - Service Reporter
        maintainers:
        - email: talk2sam@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-ui@sha256:213889e3c4c1d7bb10c88926ac8276d7e78c558397993d098e36f6c5da617de1
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-oauth2-proxy@sha256:c403c24857e29ddc306ba5d9bc28c3d556b181c5265fee16879fd2b4d946f899
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator-bundle@sha256:648f92bacd52fb51655925e34da1d0bdd0da720da8489b79064e73a47dcdd54a
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:277e2ab849907553e4992c2dccc11f03ae1e0302f20aae48c620968aa97d6afc
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-postgresql@sha256:b6e11080774863f56dd16d5748d03c3f89f6f4872f662ee218d6bf600491a35f
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter@sha256:87e62695d733a60bfc2084cd751810df6801a5fbe1ab87b0a421f6814f1b887a
        version: 4.2.14
      entries:
      - name: ibm-license-service-reporter-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-license-service-reporter-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: orchestrator-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: orchestrator-operator.v1.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Orchestrator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "orchestrator-sample"
                  },
                  "name": "orchestrator-sample"
                },
                "spec": {
                  "argocd": {
                    "enabled": false,
                    "namespace": "orchestrator-gitops"
                  },
                  "platform": {
                    "eventing": {
                      "broker": {}
                    },
                    "monitoring": {
                      "enabled": false
                    },
                    "namespace": "sonataflow-infra",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "64Mi"
                      }
                    }
                  },
                  "postgres": {
                    "authSecret": {
                      "name": "sonataflow-psql-postgresql",
                      "passwordKey": "postgres-password",
                      "userKey": "postgres-username"
                    },
                    "database": "sonataflow",
                    "name": "sonataflow-psql-postgresql",
                    "namespace": "sonataflow-infra"
                  },
                  "rhdh": {
                    "devMode": true,
                    "installOperator": true,
                    "name": "my-rhdh",
                    "namespace": "rhdh",
                    "plugins": {
                      "notificationsEmail": {
                        "enabled": false,
                        "port": 587,
                        "replyTo": "",
                        "sender": ""
                      }
                    }
                  },
                  "serverless": {
                    "installOperator": true
                  },
                  "serverlessLogic": {
                    "installOperator": true
                  },
                  "tekton": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-08-15T00:51:04Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/rhdhorchestrator/orchestrator-go-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Orchestrator
            name: orchestrators.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub Orchestrator is a plugin that enables serverless asynchronous workflows to Backstage.

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://rhdhorchestrator.io)  displayName: Orchestrator Operator
        displayName: Orchestrator Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Orchestrator
        - RHDH
        links:
        - name: Orchestrator Operator
          url: https://github.com/rhdhorchestrator/orchestrator-go-operator
        maintainers:
        - email: jubah@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/rhdh-orchestrator-dev-preview-beta/controller-rhel9-operator@sha256:9f5423829ac55faf2395e2d0d3683b238339dc03285696e1b350c5c5a1c4df3a
        version: 1.6.1
      entries:
      - name: orchestrator-operator.v1.6.1
        version: 1.6.1
      - name: orchestrator-operator.v1.6.0
        version: 1.6.0
      - name: orchestrator-operator.v1.5.2
        version: 1.5.2
      - name: orchestrator-operator.v1.5.1
        version: 1.5.1
      - name: orchestrator-operator.v1.5.0
        version: 1.5.0
      - name: orchestrator-operator.v1.4.0
        version: 1.4.0
      - name: orchestrator-operator.v1.3.1
        version: 1.3.1
      - name: orchestrator-operator.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: orchestrator-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtv-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:d0158c37686bf378347dea5368c0c7838eec9674fbb0ff11f369e6fd99753652
          createdAt: "2024-09-23T16:42:11Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.7.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.5 is supported on OpenShift 4.12 and 4.13

          Migration Toolkit for Virtualization 2.6 is supported on OpenShift 4.14 and 4.15

          Migration Toolkit for Virtualization 2.7 is supported on OpenShift 4.15 and 4.16

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:d0158c37686bf378347dea5368c0c7838eec9674fbb0ff11f369e6fd99753652
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:f8b8d2ce23c1be9dd74d18c7a90324288213e91630da529ded561fbfe5d59a5d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:fffa2fd9f8c804d021264748a910ceb17cee388b4d55cc763a8598234db87f15
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:31e20712c82ae77807eb5dfba5e8616c045280ff164027443214ab8f259d964a
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:ac13419a8c9993bbe656563f478095526dedd34a93d6463979d26852d08ed217
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:19909db45872b2d665eac31f20dac2f0fa3bd9b9a343f2be839747106de6bc5b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:90742a924a1782dced8ae639e68aef93df3c14e4a7b65e2e6b1b73e22860e7a7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:3a07a1e7c74783b0de3c190717f17091d4c9d573776419b46c2ef0c55328d769
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:045cfaec032648cb1a734398fd625de7ac1cc25bfd621b7558e641da82316b58
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:8ce4bba6a69de762f221792c03f6a759afcd16dc576de4b9b1f341d660cb82b0
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:1b06a93c3458b92d40d8eb437043c9abbb9ec5290fe85b2f6ef12297bea0c838
        version: 2.7.0
      entries:
      - name: mtv-operator.v2.7.0
        version: 2.7.0
      name: candidate
    - currentCSV: mtv-operator.2.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui": "true",
                  "feature_validation": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "plan": {
                      "name": "example-plan",
                      "namespace": "openshift-mtv"
                   }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {}
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "type": "vmware",
                   "secret": {
                      "name": "",
                      "namespace": ""
                   },
                   "url": ""
                }
              },
              {
                "kind": "Host",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                     "namespace": "",
                     "name": ""
                   },
                   "id": "",
                   "ipAddress": ""
                }
              },
              {
                "kind": "NetworkMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-networkmap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "type": "pod",
                            "namespace": "",
                            "name": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "StorageMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-storagemap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "storageClass": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "Hook",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "kind": "Provisioner",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "cinder",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "name": "kubernetes.io/cinder",
                  "volumeModes": [
                    {
                      "name": "Block",
                      "priority": 0,
                      "accessModes": [
                        {
                          "name": "ReadWriteOnce",
                          "priority": 0
                        },
                        {
                          "name": "ReadWriteMany",
                          "priority": 1
                        }
                      ]
                    },
                    {
                      "name": "Filesystem",
                      "priority": 1,
                      "accessModes": [
                        {
                          "name": "ReadWriteOnce",
                          "priority": 0
                        },
                        {
                          "name": "ReadWriteMany",
                          "priority": 1
                        }
                      ]
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:74a740878707470ea79c7bf03a7f045c2ea77d924d0890042ba3554dc5613a33
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.1.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui": "true",
                "feature_validation": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          repository: https://github.com/konveyor/forklift-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: Storage Provisioner Capabilities
            displayName: Provisioner
            kind: Provisioner
            name: provisioners.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator fully manages the deployment and life cycle of Migration Toolkit for Virtualization on OpenShift.

          If you experience any issues or have feature requests, please [file an issue in our Github repository](https://github.com/konveyor/forklift-operator/issues)
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        links:
        - name: Konveyor Forklift Operator
          url: https://github.com/konveyor/forklift-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:584a696270554f4fb64755cdc5b525bc6852925e6780f776fb309b1bd8d81d0f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ui-rhel8@sha256:8089b0889f8956f2abf5a69f4ab8f4382919bc15a48663b98551fb0b865d8af7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel8@sha256:aead9b494562985e20fa466888f28119ab302d64e93d7f23bc7d8da1c60267e7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel8@sha256:8c6758c2d81f807c3875f41cec14d9e39f1fe7710584794de57a1449510110db
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:74a740878707470ea79c7bf03a7f045c2ea77d924d0890042ba3554dc5613a33
        version: 2.1.0
      entries:
      - name: mtv-operator.2.1.0
        version: 2.1.0
      name: release-v2.1.0
    - currentCSV: mtv-operator.v2.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui": "true",
                  "feature_validation": "true",
                  "feature_must_gather_api": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "plan": {
                      "name": "example-plan",
                      "namespace": "openshift-mtv"
                   }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {}
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "type": "vmware",
                   "secret": {
                      "name": "",
                      "namespace": ""
                   },
                   "url": ""
                }
              },
              {
                "kind": "Host",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                     "namespace": "",
                     "name": ""
                   },
                   "id": "",
                   "ipAddress": ""
                }
              },
              {
                "kind": "NetworkMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-networkmap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "type": "pod",
                            "namespace": "",
                            "name": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "StorageMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-storagemap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "storageClass": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "Hook",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "kind": "Provisioner",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "cinder",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "name": "kubernetes.io/cinder",
                  "volumeModes": [
                    {
                      "name": "Block",
                      "priority": 0,
                      "accessModes": [
                        {
                          "name": "ReadWriteOnce",
                          "priority": 0
                        },
                        {
                          "name": "ReadWriteMany",
                          "priority": 1
                        }
                      ]
                    },
                    {
                      "name": "Filesystem",
                      "priority": 1,
                      "accessModes": [
                        {
                          "name": "ReadWriteOnce",
                          "priority": 0
                        },
                        {
                          "name": "ReadWriteMany",
                          "priority": 1
                        }
                      ]
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:3d28138500d6b715e7d11983684db12ced28dfef9c97c619044f0b830534f4ed
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.2.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui": "true",
                "feature_validation": "true",
                "feature_must_gather_api": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          repository: https://github.com/konveyor/forklift-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: Storage Provisioner Capabilities
            displayName: Provisioner
            kind: Provisioner
            name: provisioners.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation
          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.
          * Must-gather-api, a service to generate targeted must-gather archives.

          ### Compatibility
          Migration Toolkit for Virtualization 2.0 is supported in OpenShift 4.7
          Migration Toolkit for Virtualization 2.1 is supported in OpenShift 4.8
          Migration Toolkit for Virtualization 2.2 is supported in OpenShift 4.9

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [Konveyor Community](https://github.com/konveyor/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/konveyor/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Konveyor Forklift Operator
          url: https://github.com/konveyor/forklift-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:3d28138500d6b715e7d11983684db12ced28dfef9c97c619044f0b830534f4ed
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel8@sha256:097a3c667e594060844cbe5d16adcc278d35402d1965ff3fca249c8f84a688dc
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:02ec1b0dbc12ba31fce85ee164c97560ba3a3184316324ab7c871989a319bf8d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-api-rhel8@sha256:d6bf2bb018c471f8f86f85c62ef52f4a99510b748ac8b6b7b177fc44fa53fb44
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ui-rhel8@sha256:697b20d6423ac25c0d2f41b2598f4909b6863e85265a4f52bc23b8ce77afc732
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel8@sha256:220d8f97173292dc8389bf4b1a350843757f01ea7861bbcebcf75749366b450b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel8@sha256:6f332be2991b026d5ce2b22e5bdef8d17da06c8b437fb3e5b26e8d35b75e3f95
        version: 2.2.0
      entries:
      - name: mtv-operator.v2.2.0
        version: 2.2.0
      name: release-v2.2.0
    - currentCSV: mtv-operator.v2.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui": "true",
                  "feature_validation": "true",
                  "feature_must_gather_api": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "plan": {
                      "name": "example-plan",
                      "namespace": "openshift-mtv"
                   }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {}
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "type": "vmware",
                   "secret": {
                      "name": "",
                      "namespace": ""
                   },
                   "url": ""
                }
              },
              {
                "kind": "Host",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                     "namespace": "",
                     "name": ""
                   },
                   "id": "",
                   "ipAddress": ""
                }
              },
              {
                "kind": "NetworkMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-networkmap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "type": "pod",
                            "namespace": "",
                            "name": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "StorageMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-storagemap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "storageClass": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "Hook",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:fc6a4503d00719a0d22736a78689b5828a8ebaf8c7895689a56df5cf7b43b97f
          createdAt: "2022-11-08T18:56:59Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.3.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui": "true",
                "feature_validation": "true",
                "feature_must_gather_api": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/forklift-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.
          * Must-gather-api, a service to generate targeted must-gather archives.

          ### Compatibility

          Migration Toolkit for Virtualization 2.2 is supported in OpenShift 4.9

          Migration Toolkit for Virtualization 2.3 is supported in OpenShift 4.10

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [Konveyor Community](https://github.com/konveyor/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/konveyor/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/konveyor/forklift-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ui-rhel8@sha256:e6c2337e610e0f634e57a200a92f492ebe1f111d9fccaa019f8756ba907b067b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel8@sha256:4423034aac33409721ce1b5344f6d5330c174c9b47fc6837b8de7b47018036d7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel8@sha256:872e2a803742071bf81b70d9a58aafd0ceb6ce1d80e651d876690820835854c0
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:fc6a4503d00719a0d22736a78689b5828a8ebaf8c7895689a56df5cf7b43b97f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel8@sha256:59ad5149ce8691eb7944fa291be75e5701c8e74160a0b9a74d098872be704a67
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:a0c201d3474996a08ce662b074ebb069e5082da36f3e7977cf54e3c4e1e54b1d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-api-rhel8@sha256:a7a2d6216ee87284a4f9e07d4c1973b9e001292c504c3b933515826e90073167
        version: 2.3.4
      entries:
      - name: mtv-operator.v2.3.4
        version: 2.3.4
      name: release-v2.3
    - currentCSV: mtv-operator.v2.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui": "true",
                  "feature_validation": "true",
                  "feature_must_gather_api": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "plan": {
                      "name": "example-plan",
                      "namespace": "openshift-mtv"
                   }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {}
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "type": "vmware",
                   "secret": {
                      "name": "",
                      "namespace": ""
                   },
                   "url": ""
                }
              },
              {
                "kind": "Host",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                     "namespace": "",
                     "name": ""
                   },
                   "id": "",
                   "ipAddress": ""
                }
              },
              {
                "kind": "NetworkMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-networkmap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "type": "pod",
                            "namespace": "",
                            "name": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "StorageMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-storagemap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "storageClass": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "Hook",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:fc6a4503d00719a0d22736a78689b5828a8ebaf8c7895689a56df5cf7b43b97f
          createdAt: "2022-11-08T18:56:59Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.3.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui": "true",
                "feature_validation": "true",
                "feature_must_gather_api": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/forklift-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.
          * Must-gather-api, a service to generate targeted must-gather archives.

          ### Compatibility

          Migration Toolkit for Virtualization 2.2 is supported in OpenShift 4.9

          Migration Toolkit for Virtualization 2.3 is supported in OpenShift 4.10

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [Konveyor Community](https://github.com/konveyor/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/konveyor/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/konveyor/forklift-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel8@sha256:4423034aac33409721ce1b5344f6d5330c174c9b47fc6837b8de7b47018036d7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel8@sha256:872e2a803742071bf81b70d9a58aafd0ceb6ce1d80e651d876690820835854c0
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel8@sha256:59ad5149ce8691eb7944fa291be75e5701c8e74160a0b9a74d098872be704a67
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:a0c201d3474996a08ce662b074ebb069e5082da36f3e7977cf54e3c4e1e54b1d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-api-rhel8@sha256:a7a2d6216ee87284a4f9e07d4c1973b9e001292c504c3b933515826e90073167
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ui-rhel8@sha256:e6c2337e610e0f634e57a200a92f492ebe1f111d9fccaa019f8756ba907b067b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:fc6a4503d00719a0d22736a78689b5828a8ebaf8c7895689a56df5cf7b43b97f
        version: 2.3.4
      entries:
      - name: mtv-operator.v2.3.4
        version: 2.3.4
      name: release-v2.3.0
    - currentCSV: mtv-operator.v2.4.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_must_gather_api": "true",
                  "feature_ui": "false",
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:f17fb7a2a840bda51551683e18d01ad1fae3ea73c293de1711f382ed309b8668
          createdAt: "2023-10-23T12:16:09Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.4.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui": "false",
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_must_gather_api": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.
          * Must-gather-api, a service to generate targeted must-gather archives.

          ### Compatibility

          Migration Toolkit for Virtualization 2.2 is supported in OpenShift 4.9

          Migration Toolkit for Virtualization 2.3 is supported in OpenShift 4.10 and 4.11

          Migration Toolkit for Virtualization 2.4 is supported in OpenShift 4.11 and 4.12

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-api-rhel8@sha256:cf22690c6bb420b9acc09ee5f90329ece6a35c18e5d3d6f6e40a320e2385ad85
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:7b4eb4de5dbc7e1af9c74d47fb6bba39f4973dc2d41531989c9fcf454b715b14
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:dc7ed113fae6ea3d5baaf165b08b5404a15f3a70f08089caca65b837623af4fb
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:8d3aeb8f6cfb17f5694f92f56a2361e08d220e2608183fb39ac94e92187b3fc7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-warm-rhel8@sha256:a6f91bc0ffc5e2f545a18b8888715281db7a751e1eff5f9eb89e8d319268a3ec
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:f17fb7a2a840bda51551683e18d01ad1fae3ea73c293de1711f382ed309b8668
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:8e7a7b3c2dc31b90e2102a82633f1eda95355496ff657964fedb26157f9ec229
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:ad9bacdfa4f10e020902d8e0381bdd84e48ce89ffab0f4144bd3c03fc91f7259
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:86716db780d3e43fa1e3029ab095ae6a557cf023797f4d79565a1aa5ebcb9b93
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:a7caa55c9d211a869b42e6f0101cd5099d64f5f3fd423f8a06091bad98d84aff
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ui-rhel8@sha256:54638e44269175e112a510fe2558f9a1dda4758bf3b3fb5e5b54c6617a5917bd
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:20a7b16f9f5d3dd58dbc7c061567590618a24bd595911ed089034095cc919d96
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:9aec00a487545eef76dcc4701680c5ec5bdc5e74302edbdfbc12c821b8ab4997
        version: 2.4.3
      entries:
      - name: mtv-operator.v2.4.3
        version: 2.4.3
      name: release-v2.4
    - currentCSV: mtv-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_must_gather_api": "true",
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:7dac245977fa0beeb7539b8ea5b3be68c457da132c979fd36405509fc879c7e1
          createdAt: "2023-08-31T15:03:42Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.5.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_must_gather_api": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.
          * Must-gather-api, a service to generate targeted must-gather archives.

          ### Compatibility

          Migration Toolkit for Virtualization 2.4 is supported on OpenShift 4.11 and 4.12

          Migration Toolkit for Virtualization 2.5 is supported on OpenShift 4.12 and 4.13

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:00fac23d8a514208c906ab2c911f188947368cb5e3076aacf3dc049e27b6c366
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:7dac245977fa0beeb7539b8ea5b3be68c457da132c979fd36405509fc879c7e1
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:8cc711f92258cad8f0747d3b98fef432d20661b90c4b67df8aa91650f2be3388
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-warm-rhel8@sha256:b5e0a88008c5c355844ed16b0889b0d73e63748fffafe57d0d1c52bc39c50f87
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:349cb43a620b2db9e724d874b231ef2ad896260009e3b0a93539121ac9cbe84b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:c157d6ebfb12f37a646f5f3bfbfdc90b7c8c669afe08ad342d34e474c39aa094
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:68cde15ed754189d74a528fc61cc2191861a27a4a9be112fb3ce1de3e836a77c
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-api-rhel8@sha256:c702dc948e42b9ba1e87b988f603be3df249a44dfdd6a11b0b9b49cb25f6de0c
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:55d773cad8fcf2919f5b0a5387dea62feb4ce676fb94eba310fc4d459f729d73
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:18591334cd99eb9e9f7e3d2609ff8666a9c0625e996d2f8751a14ebb9386457a
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:6097cc8e88905793bc61e48a62d77decffbbc59c1d4de83bb2eb5340ea1e8697
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:e2fca2e132bfc9a90a7c1f7e8e6b6d6c4fd8a63ec4177f5dea3f06c0c529e722
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:e665d8c5da798dc0dcd66b691db087de49564333e3c6e238d16a356c8c62d080
        version: 2.5.0
      entries:
      - name: mtv-operator.v2.5.0
        version: 2.5.0
      name: release-v2.5
    - currentCSV: mtv-operator.v2.6.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:1dfbb0993d3009608a6b272f178197f8a88edb047db4f4ff84aee8c023191d65
          createdAt: "2024-09-13T15:42:40Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.6.7'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.5 is supported on OpenShift 4.12 and 4.13

          Migration Toolkit for Virtualization 2.6 is supported on OpenShift 4.14 and 4.15

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:d34826b2a9370a1b6502d8e74a1254b2bd786684b25d90c2f0a960be0e99777d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-warm-rhel8@sha256:a46eb455027d248bcf591b612f34eb2ff4cd887cbbc3d116f78187b579ac31ab
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:3fef1c267f1f89d51ba7cd54bf39e17d7fe8ba5a4b6ecf9cdedcbe6d49d28bcc
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:1dfbb0993d3009608a6b272f178197f8a88edb047db4f4ff84aee8c023191d65
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:537f5fed16dcc46ee9cea05285a68c0e037c9b439289ff537590352865eced92
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:7dbf0f8d3d11cdb73033377552f999951331ed61d28c98cfd836903faf5601f9
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:a30fca5c5f260b94181c03a8c4c1cfe340048af87515e86571e9676e836d4b9f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:c98635ac256743873c4d74d63f4278cd26794dbaa1d56a82b15089dab510e905
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:a24acbfeca235270e75bd78046aa1dded9ff2a325a725cb46fc52ed3ff6880b8
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:403a5f8fd64926a9638342dbe568b557a7e3dc4b700cde79e64ecd9d948002a7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:6e0b0a7aca1a4afef62cdf511195e6a567feb644d3e5128c9899afc351807212
        version: 2.6.7
      entries:
      - name: mtv-operator.v2.6.7
        version: 2.6.7
      - name: mtv-operator.v2.6.6
        version: 2.6.6
      - name: mtv-operator.v2.6.5
        version: 2.6.5
      - name: mtv-operator.v2.6.4
        version: 2.6.4
      - name: mtv-operator.v2.6.3
        version: 2.6.3
      - name: mtv-operator.v2.6.2
        version: 2.6.2
      - name: mtv-operator.v2.6.1
        version: 2.6.1
      - name: mtv-operator.v2.6.0
        version: 2.6.0
      name: release-v2.6
    - currentCSV: mtv-operator.v2.7.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:db4cbc7214cdc7bf176ff16e562120111219ad8e663b7d4385dfd4c164253b7f
          createdAt: "2025-04-22T18:23:33Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.7.12'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.7 is supported on OpenShift 4.15, 4.16 and 4.17

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:a0360971c4cf448f99eb9ca79c9a153aec479e59e5985b91e019ee478382510b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:ed8cefd01f3b03df7faaf1ba0774cc18b7fc08b25f359d1e52c794b4ad60441e
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:33bae3e4c5b96df3319287d409d8d25db7db33704c7e9f8a58aec797bf20d12a
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:db4cbc7214cdc7bf176ff16e562120111219ad8e663b7d4385dfd4c164253b7f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:97ba4a4b479957e57eb296598d0c8797233f3fd0dbac20fb4686db02d3a8ed9f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:1622759bca392e2f7f66611df7988e7c63138807f917f8b34aefcc60467051ec
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:a32fba94e62f4c979e3b3596ab3a70750bd3bbf841e6004b5fd5305ac796027f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:51fdb9274911e93ae9dff9ab616a04a8a2c8fce4b6b21c05cf5f0145b1f93de0
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:a0b713409ad1d5018832cd50a2ffa3ce70ebf6cf1a4e532b3676b6f039316b52
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:ec2b67b3aff9b74f2246c310f1871198bbea8428b4e563d0e8c254a785b46a27
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:112a465da1695340286a7ff6e4dd66144209ce4389134f03813f786d30ba7508
        version: 2.7.12
      entries:
      - name: mtv-operator.v2.7.12
        version: 2.7.12
      - name: mtv-operator.v2.7.11
        version: 2.7.11
      - name: mtv-operator.v2.7.10
        version: 2.7.10
      - name: mtv-operator.v2.7.9
        version: 2.7.9
      - name: mtv-operator.v2.7.8
        version: 2.7.8
      - name: mtv-operator.v2.7.7
        version: 2.7.7
      - name: mtv-operator.v2.7.6
        version: 2.7.6
      - name: mtv-operator.v2.7.5
        version: 2.7.5
      - name: mtv-operator.v2.7.4
        version: 2.7.4
      - name: mtv-operator.v2.7.3
        version: 2.7.3
      - name: mtv-operator.v2.7.2
        version: 2.7.2
      - name: mtv-operator.v2.7.1
        version: 2.7.1
      - name: mtv-operator.v2.7.0
        version: 2.7.0
      name: release-v2.7
    - currentCSV: mtv-operator.v2.8.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "\u003cbase64-encoded-playbook\u003e"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "konveyor-forklift"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:e4145efb22dfc336d8b81f24f5b488db602d6d547479d615bf82b4fdce96c784
          createdAt: "2025-08-26T14:06:48Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: VSphere Xcopy Volume Populator
            displayName: VSphereXcopyVolumePopulator
            kind: VSphereXcopyVolumePopulator
            name: vspherexcopyvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.8 is supported on OpenShift 4.16, 4.17 and 4.18

          Migration Toolkit for Virtualization 2.9 is supported on OpenShift 4.17, 4.18 and 4.19

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/support/policy/updates/migration-toolkit-for-virtualization).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:e4145efb22dfc336d8b81f24f5b488db602d6d547479d615bf82b4fdce96c784
        version: 2.8.7
      entries:
      - name: mtv-operator.v2.8.7
        version: 2.8.7
      - name: mtv-operator.v2.8.6
        version: 2.8.6
      - name: mtv-operator.v2.8.5
        version: 2.8.5
      - name: mtv-operator.v2.8.4
        version: 2.8.4
      - name: mtv-operator.v2.8.3
        version: 2.8.3
      - name: mtv-operator.v2.8.2
        version: 2.8.2
      - name: mtv-operator.v2.8.1
        version: 2.8.1
      - name: mtv-operator.v2.8.0
        version: 2.8.0
      name: release-v2.8
    defaultChannel: release-v2.8
    packageName: mtv-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: serverless-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: serverless-operator.v1.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.36.0 <1.36.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:52ba6b1252b6f3d41e2a0b3b880cb768010ab9dd925ac34596f6adaedcbbe07c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-source-rhel8@sha256:88323da16360981f27b8e4686fbbb64053e4e0e3c79c65d695edb317b334e1e0
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:3fe326ad9a00b112fc888ffd345c8d5042f0a22cf32ea575931192c39ce1fece
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:259e5dab50d36cdbfc117a42325128c4e16f34c048ba64b36d4a89132ddb4c1a
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:4344716eb2940c6a3fa9a278250f936602bb1549fece6037912837719c178b24
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-transform-jsonata-rhel8@sha256:3af4c48030263eb5769e5f36a6dd13fd390d8566efaab37532f39e1df1c95c67
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sns-sink-rhel8@sha256:51b5e18e8d4c1ca81d7c5ddfea6352a3887f757acfcc53bd2f8bd3b01337ed42
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-ddb-streams-source-rhel8@sha256:a903480159ebf1d8652eeaa702d9885fb76bb106c02d75cec5c82c31d77d6410
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:e3ae1cc6012f19631c012a07b113fb015977d2ee4f9faab3207fbc2464c25aa7
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:596a7272d6de714d7dbc6e00e9e50a2872846c269547465252d953172ef0b009
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:f4d86b196457cdeebdc380b6f1b707c10f8dee789c7779c90ee1332e511af8c4
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:f2d112ba9fdfccea988589b99b463c8a7157475f7a3f03678e8999e65b678cc3
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-sink-rhel8@sha256:afa634e029bcf79a972f49372498d44b45bcfbdb6eb28ecaf4f0b0b2a0a1824b
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:dcfe5fbc57b9f8520adbff368c8bf29eb491c9c26d1a1737a1722cffea83eb1b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-source-rhel8@sha256:27e74287774cda208f667bc2a386debe6b6255e8991abfb37c959c8525f2af53
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:9b328fff0cc8545fb9bb1f24f11d78cfac3ee76dedc11fa4def4262d79a60e8c
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:82e69a1890042f5636ac7bb645b1aaeaba0a4894851eccc1fa1c63296918e31a
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:3d780ad1ab280b5092b90e277485d589da2a18e8b08090af7d95aecd17af545d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:631eb3a8dd9d83ceafa7d8a4da4b1e2d4f2740e6017c766828dcc7575cf1322f
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-log-sink-rhel8@sha256:30b8045ef5f925a36c58941386f1899bec125d005a08aeeeba5edd49b0859149
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:c7c2e738e0d9da1d2d1993f68938e72c4709b1f28a54fa92224ef6a0e456b0aa
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:a1834b620f875681388f0ae224dd0771773a8b61b1d8e5214a942f232854b87b
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:074ebfc4494581611647ba90049655a470ca43743610a27be77db66a8633bac6
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:429d5d9568cca31fbe83c663276a300286af4cdbf3d35685afdbe5f6f2f8c6e0
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:5352e4eb4e426a41521276decadf204b6d23b4762a298e4d3dc73969e3eddc68
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:bf82b37ab8fd29c7a7cd846008a16d14b4cfdc1c09df4f731d71be3957e90123
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:6b7316cce3432d391adc6ac38c92ff0ee86dd7ae0101957cc65cd338cc69e352
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:3b04e9d8cd4b5e43a07ece838d10478723a6c6529350f010550b4a0d1d0950d1
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:6210a946f373b9c36c9b31293debe6fed090008b8234c0ba17d37fa81646172e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:c39cb88afaddc295d93a3774b8a20d001b82392df540fb502ee02316d82460a9
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:0c2cdf917bbad1f5bca8a2a2ef12a96cfcd0c77c190cc35687eb57169b54ab6f
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:187fb9d64b53c55a029ffce3c0f9e7c3304683fa8a0e832cfb7134e8a3d0c237
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:d6f338e08fa937c8a4ccfec692424de4acded565f7ca3971ea507dffeb351503
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:f3d87a7fa5b5f2818effcac3304bd56eddcdb0db97a18ac0f35a06fc7e223e7d
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:4aeb68a22f63300e54391ec6b4cda8c4e665d904b5dcdd3c40155040525742a0
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:ed60e889a677fb139f2eb4bd9ab4fda6ecc39fd050ca2b6cc1c19717e428535a
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:77e4689b2ce7a713e905553e0e42b8e128045f40cb6c6d4ee082f126cbc46b00
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:4274004b6289fa9bea3fa6e56d64077a328a6b5e63a9f0e9c41f6d561f682bd7
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:a6f565d73375e827d830dca85838f4c4cb7f08b0c078acb0c0d567b25f0d256a
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-timer-source-rhel8@sha256:8fed5ef9a1f39a55077bb5026c8854d0ee7fc8d8ff5da277ce9c1bfc2ab2474b
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:78fee4c596f2848487ca704da3fa1074e3c9c99b3f96f66833aeba63836d7b93
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:a47b6b5237a87ccd90e5a49a53d06485fd55c87eebf12d01c335b87fc3a8c494
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-sink-rhel8@sha256:d9b944e64391a996104afbe8a21f66316a948e9fb7b91734e861847285be59c7
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:ee6533b71cf280877c9d7489dfd821c9209c6802bed7fa3fdb4f8f00fbeacb09
        version: 1.36.1
      entries:
      - name: serverless-operator.v1.36.1
        version: 1.36.1
      - name: serverless-operator.v1.36.0
        version: 1.36.0
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable
    - currentCSV: serverless-operator.v1.29.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2023-07-28T05:42:31Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.29.0 <1.29.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of a particular version of Knative
              Serving
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Eventing
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Kafka components
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:a3c5d5fd8014c432830be912aedebbc58f9c60a8cabd87ae6c42c7b8d27db8f9
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:a59ba5aeb8e66db3d8ba9074b8e553603c59e6b0e35b57b82c4ce3de589b94cc
        - registry.redhat.io/ocp-tools-4-tech-preview/source-to-image-rhel8@sha256:98d8cb3a255641ca6a1bce854e5e2460c20de9fb9b28e3cc67eb459f122873dd
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:56e48afb1dcbf4e266ace944105f3c45042550d8d4a0528c982225a793d26d56
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:31d45f10a8519addec9aac115c5edc8a2e14f9621c875d95f123d1291776ef4b
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:b780634fffe2257eb845f8c1d25498d9c654d0327c1f8841b40d90c176a2bfe0
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:04866e16881f64ff2b5edd379eab0731c9d66a47c25ebfe79c8d8ad33d8116e6
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:61fb98802490323bb03ff4a7060334d8fbda9e9f95cd6366b29b1b6494e4ef8b
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:3b7449287adf98f0cf0a1dc74e03b69218be3bcf5538327bc34ef776cc1c69c4
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:10788ca67809e5a2d76d6d72df7dd64ea4407773999707a7681e0cf1b5c4695f
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:216f58dc8c8166c0c69fc0312b667daad14cf39b7aeb7bc0fe1e313376f1a5fd
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:aea5f200a845e1d4dc718d76465fdbb1eeac2b20ecae224d5d20a98214ad0357
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:e5db11a0adb889f907cad797dad9cf2de09c433e5522fa333193cb35575802d8
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:c4cff5ad5b0ac79e71a16b88640a312ab320817041317f50caf575f7e12a864e
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:c2cb80bce34a74afa76151d2b44dc25ece73fb274cdfc6fdf243b1b0cb7d989f
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:3aab995872336e6ccb590ef6488c097cff38a91348055031145427a00aa6b732
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:b520305189c7567b2f8ac867b20468259fb22cd3edc5b4576af2bfcdb0c11855
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:8273842d8e1714fa03c7e0809924b1ffc5ef282f79b9606bd8ac88c03e7a16bd
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a10e76aa15707e708f82c8d9e8b5179bf67780d99f3c3e2e99b84ae2a0aa4ecb
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:6342358e6aed2339333ebf37fc223a41a0a3e199a0e4a840f3f66e5accf8f233
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:594c3ac24bf59802e2d3fa127e4051277bc07b6444f934f535fddeb668ccc144
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:4408d1fcf34c0ffe9327d4c9c0bc772afa969d831bc5de1097164d10b4e8d900
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:03a3d6f0314acf4da4f3d3d85f1e273a3ef25828d28f5b22f394441d55a645f3
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:7dd11d385b3d5a178566d901650ecd45e4af395faec2141d57b98af8fc1e5151
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:a4943c0563c45ceccfda0a01e26be3a01c1437c89d927714c577bf6060495958
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:fd86ea2a34de43014dea3ba2b8d705a3d94347d3a24a370a8bfd032ed239dc56
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:f37f494956adf0892ffca34d2a4c9b839ce28dc9f2779ca32e8801fb4f0e9105
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:557e73127326be051aa6257e7daeb9721e2af04d2f06cc26cfc7a1073666b989
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:fe916630003194dbcd3894dd880cad8b15a041a5e534cd8f76fde6296db00797
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:397eb425f383c1273c7976bea69a603530676a00a61f461ceb5f48d67901c4cd
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:bd134e64c156dc5c81d14522c9c9efcc1f56b2e972ac8ecb1bb1b130c76adc60
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:ab90ec9b3cd9203fdcdd7c43abf95f48c9d7479a9c7af479eb1e4b6599ba9642
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:c12c471902b8e2596588d9e553ad59b501d979dd04f301d45f7d45e794737a8c
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:de0634d14962bdb8701a8c29e80b7b70684824462bc5203e5fa7f139839cb73b
        - registry.redhat.io/rhel8/buildah@sha256:a1e5cc0fb334e333e5eab69689223e8bd1f0c060810d260603b26cf8c0da2023
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:0229a0a4f026f04f8fa0a1cd8e7afea10117fe3f12e21986ec6089769c8afcaf
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ad399867cbb712111db6affaf5c9e3573699ed0eb0def8ce6057de02054ced4f
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:3d7801f7ffe6420595779673a170235921a9090333f8ca51d5b7f664e19c82b2
        version: 1.29.1
      entries:
      - name: serverless-operator.v1.29.1
        version: 1.29.1
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.29
    - currentCSV: serverless-operator.v1.30.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2023-10-30T14:03:25Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.30.0 <1.30.2'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of a particular version of Knative
              Serving
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Eventing
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Kafka components
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:9a28cb1d4c4f4dd7d494829eeeb7122db2174a27747f915244672c95dc8f9d12
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:4800c85e8de031bf77f6fae0e4c8e6c1a9912ccdfb77799cc2ea574c37d870ac
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:66149df8549380673c24e1487d0c2c8eb638b7e502455671ca64ed6f32ff82c3
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:a0c5a91348ec71697747113990652feb024143fbbd0cdd67e8ae672730383a79
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:2d3e314a6584fc6974517c1a992fb59f5133f54c2f6cfaa42f579c0d332e0300
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:ce4b36c90955a356a5774410356aafdac2b60795526a3488b0b76ac2ec1a018c
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:1f7e65009168383cdcb6b2caa22fe4db49e03c71574973a8086f6c332dc5e26e
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:32445e4a73bc7c1e8007775adeb147175d934b17e7d3d4dcf422104dbe80a52b
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:c9353a022e6f883b7c0b2d87b1bc4653172feb13806ebf0aa075d6f9f43c264e
        - registry.redhat.io/rhel8/buildah@sha256:a1e5cc0fb334e333e5eab69689223e8bd1f0c060810d260603b26cf8c0da2023
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:92a39ca90984cebb1938447a31c9170b52994ae98f83caa860acdb2c3ebb75c7
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:06e9e073f092dce67d4227e282fd8eea6dc9f38b90a8d444cf448d80d027b616
        - registry.redhat.io/openshift-serverless-1-tech-preview/eventing-istio-controller-rhel8@sha256:b6ef281fca2b80d379b1e28fe3f23119df8ff765fb2cca291d514053fb512da6
        - registry.redhat.io/ubi8/python-39@sha256:0ff58549041d32b685cc850aaa85da32d1ef73f99cd4b6e0f6882dad5f0399bf
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:21a638b481639c3c58e8b026d875917c30a5874b08fa4ceac40d447076b2bfa4
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:124087e6402caa375b24b976c3b9fd7d53890e20ac3b0b57ba3ac646e352c7ce
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:16939b802f4c16a3d9d941917d0363325f40cdef28f6e1a982bc95401117674c
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:e9505aba7c44cffc3f261a5dc7617a0e54ed98e78f73ea3c9c7a508560b792ed
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:e8c756b659f6187790bc1a5115e0f71791bff96630859a905bc5b75d14596e0e
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:a9c95fc208cc03625cf49bcebdaac73da823d1fa3bfae4239d7db20d3a0b8723
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:3fb6e7626d2b0c92c2a2fef35e9b8171382b4bdf54393f2ab1281ae7ab441761
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/ocp-tools-4-tech-preview/source-to-image-rhel8@sha256:98d8cb3a255641ca6a1bce854e5e2460c20de9fb9b28e3cc67eb459f122873dd
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:12bc62246142a605f4efed0f6bb8843aa2df7699cc1c5ab53debb94b0682e9d1
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:3c8bf480dbb05e09e8d07cab12e4736643f61bf9b5d98efa6bc0557eb31887bc
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:1455606c8edf611d996a48e29ead53e0bb4078f55384b7689058bb25fc59fb2c
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:6057a2d18406e00b0b7b1731870f707844e39193d4cce6de50137b77f17d220b
        - registry.redhat.io/ubi8/nodejs-16@sha256:8abeddc6ccfffbe313f87e5e069e18cb7047f2c7880ad63cf15bcd0a79ba71ff
        - registry.redhat.io/ubi8/openjdk-17@sha256:af305e087061cbaa4107ce0b304a1ef987f6b54bae3f75bfd529cfd084cba9b5
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:2264d2cd4ee2a720667d625ff42b6f8be4f2cd65b57a91eb6044da086a110740
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:69ea22c646c09208ca68345d4b3508453d8c49f7f7fd9a4ff4c4cb78be7bff43
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:a5154f9d13a1c373b8ba47e2bcfb9d4474fd03e08db4134c92aa4f049536f1a8
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:96d6f52711ba433ab2f273be8080840a88e626e989027480424d19c2fe1e9454
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ef116868c9c60e5ee246060688da49832571ff6a7901cfe143a42d048743736e
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:01c8fee92889a84a19136b0843e4fed5db31c1f2a6d997669f103373fd995e36
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:5c933641efc6bf919b1945be5a01dfb7774ae865975b1bb1f44653b3e4ca9f1f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d03f2bf8f5e7cba47102ba26ea0149ae6719076c645d5e451058392d6044b738
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:3508434d5ab6b38c8a7466256f1158ebcd75e8b751afcc5255da8c21566804f0
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:cd1abeb2067afa44a0f1eca301cce37307582d56c1147820ff86b9b64a40588b
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:85a369d5696771d0932b61760336e52ae3f7c66f9a3def5312aefcac03ba516e
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:96c135ced2e3bff434addb012368d96d0778ef1d294ba507ae7e624811e20d46
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:e97eb2bb7eb529cbfe67c14e823ffa1269d614e2d35723896e89efa12a499c6a
        version: 1.30.2
      entries:
      - name: serverless-operator.v1.30.2
        version: 1.30.2
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.30
    - currentCSV: serverless-operator.v1.31.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-02-01T19:15:42Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.31.0 <1.31.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:9221a833ecc66fe2354aee7aaec2a2470b21e2d89cb6ed287ece32edab2a2edb
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:22ffe08d6a0d5461158613f2b2a2ac6936025357e282a09197a750c42b5e6922
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:3dab790fa845ed7eda088e5abdc87799da1354884a53ff1ff728739b8e278418
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:1ee675125f7141e91f4ee57ebce082967834bbdabe1a870c221f31aeb8750be4
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:486bf60d17a5d265ebd65d590a6ee1a7e13bf3953ea797b2f6c2b5910789858f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:bf693eaa845373e1a06041626a9314ec00519a6fd2437b12f7f4f3f104864039
        - registry.redhat.io/ubi8/openjdk-17@sha256:a8165bc2cd5051a96d6937e25ed03155bbd4b731da6e58cebfe2ea83209c16d8
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:565becc77dbe3bf03d89b1399d6e1853253276a71e80b6b14069285d03cb14b2
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:433821e6c7f43da5ca2e38c0550722feb179a0bab021decd9a97ad020cb78b76
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:e49220eaf3143a8958f0cb1f22115b2e834fb7177755433bdc592b753a54a8ec
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:1fcecc11b55007b9c6969a24940a391968378e4102eebf7ceabbe7efda138575
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:b37dd2cf4b459fecb6d681cac36918e486c3bd22501df50b90e20b15fe164d56
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:b2480154951109a159acfb88c5472a312a515d6a6b060d832b3f7bfc16f90ac5
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:887480662b1bfe09626fbca8ec7c5da879d097c46dca11764ea98474ec78ab5d
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:cda04e3b7c2c1f37164733ad162ea0c4e52f14fdc8eb8bdef8507f0ed5454640
        - registry.redhat.io/ubi8/nodejs-16@sha256:f90b28179d1d5de68d42ef1a9166d70a9dfea85a1c6a4bf12dce2c59eeb2fbb1
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:677be039353b2cee51d5faa5d5a2abd6a7bd46992ffe9521eb033eea9925c785
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:3181ad775ead55f317f96466c2719b2b73c5b35d31933733da21273955d160c4
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:9e120ecdbea54eecbc5e0412c1774f8a2ee70d7fca682ecd0757b8627f70680b
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:eacdad8db61b23634c3f89bffa3f1c30ac5c7bbb5e938c6208003ff00f9595f1
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:df1aee67a161fa19d7f49ba56d53e0682cc1996436e985a12a2068fa419b54ea
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:3b01ebdb0704dc6a1a993ac076819489495ab3fe32e3c14609e42d1358856be2
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:d8bdd96c6a9b31f48c961ca438c254511db785fd34bd10b98084b702baf0ebab
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:63e08279d2852d494d4f8836c54e2ce942a342f452a23ef437df9640bb2f5bc9
        - registry.redhat.io/openshift-serverless-1-tech-preview/eventing-istio-controller-rhel8@sha256:5b38e5c71e422161e40f9c37dd2f46212cb80e8e37b00247767276b957080397
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:80361e052d366ffffe9a99893cd83965bd69a9ff81d8c9262976de145bbc6184
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:d2a34100f6309c3cbac7c9ccd452b7bdab2a533afc99463bfbf91e3ba2706d54
        - registry.redhat.io/rhel8/buildah@sha256:b48f410efa0ff8ab0db6ead420a5d8d866d64af846fece5efb185230d7ecf591
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:0a2989f688b966277652957b312df736ada43fff78fd89b17f60b2137e38f115
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c078071cd51c7fd5584039b54a2004331c05a897ee8b0689bfcb9204a8ebae80
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:8353ccb4f0d70597d61e4fd1f799d4bb74e5bc038f6b9916059e712aa70b97c1
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:3be4fc173b489d0beb768739f60274e07d95ffee86d70a8423e1f59678c805ea
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:3a4e7412322755b4c278540c965a403f9e178f138bd4ad7929c1c24a8036dd14
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:de62762efa0df95ed23df56f5520490d8a231e47ce13c433c512254ceff81844
        - registry.redhat.io/ubi8/python-39@sha256:4f35cbcc6f4e108c69f7d2b70cce5afb138741b0c961506376d524e001b922f4
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:99f5dcb92a33c047f0b68236a3a42ffaec78037dd0a5eb79d1144d62d6bfcf43
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:2572d98363cca2dd94984d6cefa573c5054fc206ee70048f901ad4b846d8dd52
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:dbd9c0cac86f79b8d91ab65ae7543854144781d4cb17ef16183931ec9bc234a7
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:7941b421b3e18bda8c565305241621fcdf1212cbbe0dc02f38f4ab0565d2bfcb
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:af5c30fc9f631d9ed110641b815315d8b4aef539f687277a3821d9939ca5dd46
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:ef7d28580e75d746449a90cc0e5343ef3f4ddd75bd9fb866d97836bfbc3b314d
        version: 1.31.1
      entries:
      - name: serverless-operator.v1.31.1
        version: 1.31.1
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.31
    - currentCSV: serverless-operator.v1.32.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-08-01T10:56:58Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.32.0 <1.32.2'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:16c026f09b377c44253cdb902b9a5fa0ae766fd0185cca1863798a212e0daee1
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:2f0f84767f7d76dbbf4387ca2e272766267a2c46b5bc29f5d01f05dfa39cb944
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:20566173a8e907980e603d82fae4d96fcdfd5c43d9004c2e506cf0e09a47b91e
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:04e5b3882223d45a5f5383ce6481efb7b2cbedea57ebd894caafb1454ad8c09d
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:e1e2273fdf6384d95e1128e54c0794d7d620291a272bfb1f91e0c2723be7f1b7
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:22b4d47725ed8cdd663ee26a03f31198fff9e71dc703532e70c42bdfb4840811
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:e7e6cbb46720eeca3f0f4b9829dfeeb3db373a83c89cd808b667ded4d5e139a4
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:a5a7b2be50f211d4d9d555b2a2e2302096c139ff5011d984cc386f1530007f16
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:0729de0f3f03ed220216c6cb718fc34beece0ceacd926b5e2480bdfcf44df5cd
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:ab9a1b952e9603b4ecc5f197715a1d871abc25ebb0c5e6470d4f0b365ae94276
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:2a30ae808d5c32d812406ff4ac9b472e82b6bb0555e62eee4c2c6a972f8750c2
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:b008da825ee2429f04cf39258389740c68056f82cf4bde0cbd3db850b6256885
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:5930caa25ea875d29dff0ec5e235eaa13cb6dc98980d8045335eba1a09ae761d
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:da5f813dea324ffe9679524ea7f3801536fa48f7dbbd8da56a8492b916b15fc3
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:990962ad245a6d62cd11551bd4e1f8b14dc08abbebe47beddb2f4874f78d4f47
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:b0db9f0fc85fcdcd5da980633a3a44be2bbabae068f89d0965263de38789e812
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:9d36cde68bd822016e967148eebd2ef3e7ffab9dcdd37d17f33017e9b703d9b2
        - registry.redhat.io/ubi8/openjdk-17@sha256:a166bd2fedb99cb62ebf4695ab46e19c7e451f887940cdf0b8245c81e403aea2
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:9b6524f9c0e7f78e6ffb372a8587eaf5f4f1281eaf64f93dd00b63be84a4c21e
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:42f31eb8f12dd36541c8c0b6572d942c62e2064d5418428482d676d117a6c426
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:7c8cefd5823ad953f112b905c651f12a97491aafc008b2ba0111f0c9e83e50ae
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:3be20d63d7e82f35ac88274532d6ab7839f9a7306ca272e7d532ac03215faea9
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:46b8dc1c8ea6c40401d4d1a9868ab3edbe8b9e92364a43add6722489e2eefa04
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:ab09a1710aa2331e24df9e7283f3909fdb104c1b314dea88e919d0a396f334d3
        - registry.redhat.io/ubi8/openjdk-21@sha256:bb95ea866269ac8652f666ee1587d71f4f43da33493c2be7f6a9548bcfd2407c
        - registry.redhat.io/ubi8/nodejs-16@sha256:69439fe83fa957089cb6fd3c1042ab71d719f81dd457a5d1fbc56c1e3ea600da
        - registry.redhat.io/ubi8/python-39@sha256:8c96e8648cc6651d0cefbd20496a89f208621601f9664554d3fdeef8ced25453
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:e55c70591c0498b6b85f0d68e304e88b830ea0b35219ba8881143555757be4b7
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:25cbdf78b67e7f5a360313ccfbe3b6e2c98873b4e9ddee46c115a623e77c7abb
        - registry.redhat.io/rhel8/buildah@sha256:55f5c7d88c77e1ab5a2ab8e877172b2b66cb196e21f10dc45148470d0ee8bbae
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:1a1de677297525c0c5d317fe8a7a42684fe3f1b422a7d6bea1f70f0924653b08
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:0507ea8745267fdb53329a5274d59d23c387225d2594af51dd2603b4ad99db0f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:4c9faea6fd162db3340aee8c11af11a08c9d79bbed104967c25df26c4933c8ae
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:14ca32e999767776c3e585ff5eb1aa568f7d76fd3e5ae24cae05ce7564678080
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:8033b7c1e903ba313134645c626eb3bf7ab1a0d68dc2575bc8e26884ad4f805f
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:e70dd3a9e067075eb56e5f2772642545e308348cd49b89357c02de3979b1a5be
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:3ffcf746df914f0949477b46f90cb2992bb63f8198d73ffd54208e7a777a9cb1
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:1befc9969edd2062b3df69f45aee9771b33a2659c3a14db0aaaf5c3df985af58
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:0551f1657dc158bd41eb2338dd6514d55c39fc77300cdf1744d8f5ab14e9ed1b
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:4933b76d95d411d597d4dc460fc46c92875d28e107cda661946206a9d934b8b5
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:70c26c7ac48d1eaaf1f95ca5c5b8ba8ee6b047a3c9c4a2c2db9c49a506bb2205
        version: 1.32.2
      entries:
      - name: serverless-operator.v1.32.2
        version: 1.32.2
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.32
    - currentCSV: serverless-operator.v1.33.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-11-20T08:38:34Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.33.0 <1.33.3'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:e1cb1c24333605e52364456f9bd8d86d787d1bae17ef75dd04be61257fdb2256
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:6dbfa2573a2c78e450d3c1c04c6e7cbc50d9272233e828fe9b42e575ecf386c0
        - registry.redhat.io/ubi8/nodejs-20-minimal@sha256:acee3b00ccf8ff49bcebcf10e028e34a84d45f28f855363bafb3b93c75e84f11
        - registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:e1cb1c24333605e52364456f9bd8d86d787d1bae17ef75dd04be61257fdb2256
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:eb95deb2105f14c6a9f3bb2cdd7306b446109a9b142a7bfc726b6429f270d9e5
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:ea9e4de6228d5c3ed494c2ce5317ec290543ad40bcf7c43c2a3d6565e5d94e5d
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:359f8b9fbb4c3c2f79e8c70f3b50c994693b930f6f078c34c87112ebb93037e3
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:c3dee2598088bda36ef474f4748943a29169ed9589aeb0f77bc79a7f6bef9907
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:672022b6f2a080483d9d6f6f4a5f42e52e522a5797105f3465731809c6cf4947
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:28c9d05a3ba73b76f293e8f4902b574deb44f77a2bae8cf92473b8a8b10cee92
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:a0399a993155e755409355740b260d9b0019559a62be3b6253d8693a4fb59286
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:42d16e4d9f14a427e7a5b7e3a4ea07c258e0f2c74984baeee3b4926a8c5e2640
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:a8821b59d495e24598b27c6510b32ef19de2bb04ea1e568fc5afb9135d772036
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:9b850afa750a67120f68a49f4834ba7f105563df22cbcbc563317fe1545640d1
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:09060de091d617bd4606f7a289b4528265ae0b7a04737f383e79efd5b0d3eb79
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:dd1f7c850f7cf2a4d70403deedd8a57ecc6e0ee5455d68fdfc49ff175fe42dde
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:93b2f5211e79abc38bf92f0b8fb8adf046303efa2fe62c47a2fef9dfb5662f55
        - registry.redhat.io/ubi8/openjdk-21@sha256:095175ef8573de8055e67bd5c8a21098448ac73db4e0d6ccd5269dd36b4d263c
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:b871bd08b782c9168e9a54340efffc0b0effadde8079915b848e4eff72bb3e95
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:879fca768b46676f1df89de1cdc4ded8dac310209c146c2f381ed1deb19b06be
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:993597ff83a4d62a23f642fadc92f182944b953770716b73b7d19492b5a462c3
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:eab4aa93b9d175045b1f01bff09681107d839285e0d633cd5146d2a2efab373c
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:af66cc7dbed88907d96a390fe5c8807a2a60bbe3276b8cd13b4b061cbaf7b162
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:b12272e1a9cb4f4bb62c4badc62b0bab110abb99484612dc781c46bfeac13cc0
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:b2be693c71cfc4ce934d5b01efb863dc6e391f93a6b15fe7eeb15914063a8294
        - registry.redhat.io/openshift-serverless-1-tech-preview/backstage-plugins-eventmesh-rhel8@sha256:28ae309e4cf7feabaf78ac6067e1173f9f5b33a71c06d6c27cf25e01a0b1ed64
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:05e3dd6255294b691783523f7722cfaf13aca1a1e1dab344a25eeea72aaf9f9e
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:e9b38f1df225238836aa829f3a2016372859fb8709cdd2b30496498669a90be1
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:510e98837a549f489b2d74939f3b9c7ed90a8260ba914910f681d5c0dec7655c
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:d8ae7bc888f11121fbbba65074389b879bc22453bbabc551d79d0cee68fc33a3
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:5c7f29554d04eb9348932953344f2ea0806d14e7704355619381bf395d993f58
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:76a47d100de1f2afa23803bec503d9440605db03c7179b231f963e5663a3d12a
        - registry.redhat.io/ubi8/python-39@sha256:ac12e4706e0fdc464e791a3f274474ac1bb14d5af8f217c440fbb0c76d6a5d00
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:ace72435e3e63ab21dff2345f958cf6bd171537ef2846c5d5d91cf821877140d
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:9b94875e0e5d7abdb3b9ba7b04e217a2a9b7b4410d38d1608c86d1c21f929cc1
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:56c6ac59a68976741d2ba676a53e59fc494d04dc44b513d72fdf8a9ad894a54d
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:62bf4bee0ea17ecc60366ecc71431adc79d6f0a5b9f59489913b474128c4adb0
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ccbda564735555785815f27f8f25ad9c6c27dabe4e262586fc51609331692527
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:a9c84ac7e7c3fb1e5dedaf949d54a3aee8a62397d2bd6022094671de00b651da
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:2bd241898bd54dcb248ae86470430042dc545d45e7519ed470a1a9c8c63bdc37
        version: 1.33.3
      entries:
      - name: serverless-operator.v1.33.3
        version: 1.33.3
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.33
    - currentCSV: serverless-operator.v1.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-11-20T08:53:53Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.34.0 <1.34.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:a4b3d75903de1929326041e9885213074473f07fda91b299548d45680c295c00
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi8/python-39@sha256:ac12e4706e0fdc464e791a3f274474ac1bb14d5af8f217c440fbb0c76d6a5d00
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:633ea30ea3b6d5152bbdb61acb4271e5905d50fbe3c731fb4c8d8640a57f1970
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:586fc8ed88f91298dbff31bc2fc4c63e0d6d9356e3358b2f853f49511c3ff9bf
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:d8b94bbce98325fe8079c11c0b7a6c59bad8a7edd0532a7d27ef1632ff0de708
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:ef7f34a025e40f93fbd51f3a4596998a6dc508db824f137ffa1fbaf90e88f4a8
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:2adb0124f04d6c4a552ade79dd1499c8ca997edb45855bf438ca8043b2709e19
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:f94f0e8d21a6b6341b1db00513a1a2ed5cd84fcd898a376ecde942e940a518e7
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:01ea7a06332d60ac16ae0a27adcaaede3f680a98a22ff995259a6c42ac0e358d
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:b07010595201d7cfe0ec4f18e612d323eb38464c74eb187f275617a744d9df12
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:144883b09d8a999fd4d2a109d074e44d7d8b959987512379aa71c71e4060ca28
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:a9e8583c19ca3e01593f8c0e54731cb73df9b46b2ed47a2b1c711828c2295f82
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:76a47d100de1f2afa23803bec503d9440605db03c7179b231f963e5663a3d12a
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:6d042c614431aa9a9e77cf8ba0064423e903f7c418e367ecc23bc02a24a796c0
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:86ffb8be1762782d643c147cefcfbc2b4723b0f97d017314a99bd2275d05a1cf
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:30cb41f09067959cfe3792e070fca62820fcca4035683e5c43dd46d1cce97c01
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:45bc50a993d04b775a153e985a43d587d5455883fc4c6f31b772408297cf9040
        - registry.redhat.io/openshift-serverless-1-tech-preview/backstage-plugins-eventmesh-rhel8@sha256:09dd94090eea9cac7a1935a6d671cb97fd9fd5714c715c9d2aed461e468ca54a
        - registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:a4b3d75903de1929326041e9885213074473f07fda91b299548d45680c295c00
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:946daeff59b41a6f8c4842f8c2668f97ff1b49e2a8afae5d64ba3baa916b1d95
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:75228e70fb1383bbdbcc64d0cf24bf5d977c09112e3f32316824188264c3a2bc
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:33d62d13e16f47db9951ac826671f39fdc048fe7c3fb019eeb4bf9bd0f814222
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:d5ad3c9db7ce8e4fc950261f7c388476214a03d92e0043c7bbc4f0add2b1e69e
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:b9d85323e110d8eb6600223c38f9d4cbe89b50d59e03bf1e6b8808f7cef9a3a5
        - registry.redhat.io/ubi8/nodejs-20-minimal@sha256:acee3b00ccf8ff49bcebcf10e028e34a84d45f28f855363bafb3b93c75e84f11
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:512c4d93dc13cc53d923a6990770a7527d6fb657278e2c5d8f0ece3cd5020155
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:0f126893074cd0398654c4018df6e882b7610e148c82a41ac814f4bf3830bf9d
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:a5950f81cbc64d1edb64ae7f7f120c3435d36afcd14ffc34ae522af297d87e1e
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:3d17f12cc9da578549254ba1b8934b5afc66e453b1d5d667bb90281e45b8795c
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:c0376504edae12e7af6b6cce02ac53b6b374f5058ebefcbdbd32628b551bb0c4
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:1c82d20ccf5eb5f8e7725839f1060f4efdfd1c1b4773a99bd905e690a6e36f38
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:1252960c0c0a4dc7b39090cde0fd5742a19453b1a65a08b737bc41d044a0882a
        - registry.redhat.io/ubi8/openjdk-21@sha256:095175ef8573de8055e67bd5c8a21098448ac73db4e0d6ccd5269dd36b4d263c
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:d40daf67b36387faa9f55e70e50d10d02accfd19470c0ac11b78bf969947fed1
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:84994eabeb7342a0771c7cb1bdd70e7ee3928dd45874eafa64cde5bb385fe964
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:0f8d006d0e7468783739942a1c82c7aec376c9948f23841132c8ddece8ecba06
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:b801ed7b1cb32d4a1522a959e97f280f2eff6fc92e47fea2ea022f180d37831a
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:e520d1b6b87f85dfffab5e53cfb70562ab7edf450061c07609da39e3b2498fdd
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:4b6f161c14c68536ec2c7679d30681afa9a1fa2d512548f0f591791a10a8b894
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:76480201188dd142206040af845bd55b44155af92e0c5ee9bcdb8660bea0cdb8
        version: 1.34.1
      entries:
      - name: serverless-operator.v1.34.1
        version: 1.34.1
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.34
    - currentCSV: serverless-operator.v1.35.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.35.0 <1.35.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:f23d3ab058667c6fe9ee44dc249aa7e8449c1bb9581380ddbc66624511b286d2
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:b2de6b658a1595084a202203c90704c3f9de018a7a904ca276112ad606e374f1
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:0751b9ea7ca197480c4cbca17654d0653d1c0c760ca3c818c2a1018dc952cdc8
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:da8d975527ba100313b92e87d6d2690fddae9988aa8d82d728221295955b7224
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:ab8f85bb18311b0bfee00561ef6b5d9faf810358c954ac2c6f39cf2a7739e07c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:6b842b4eddfc71a2c8f6c1fcebb8f23a05f50c95be1ba2d20a1be660d3014c8e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:e743dd93b7ae855825716403979db3f34cd300caad130e7d3011172b1aa4eaf7
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:0e58edf5ee340fdfb282faeb6c9bee8cd6323e7f29a9b282dbb91c2622131a9c
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:477b1ac0db0abf1e16ae31bb22fdb73dd98f4dbd243fa0a7e6e94852b16a813c
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:b66bb9531f05cd46360f787955c849208496e3e45541cb50b23bfb6b68570aa8
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:9b176477be4c782ce60d81a72bacb566069a499579ce6df5a987a0841f05ad5c
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:86386b54cc10743023884193b7d6c7159b90052bc5e4a0608a6cdfd47865c966
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:cb2ebe98481808e788ca70e7f0a9b3402818b622158f327b99a1e87c23226d2d
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:d0650e4d7e72241e07a3796f20931adba231fbf5e401a45d794761f2c5afae48
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:2cbe84afebac23b5397d144b4d5660af33290df451141fa3625607b6033983e0
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:020264063c50f2a4590f3d7524b271f0f77350a1f8fa3c35ca6e28f0162aa9a5
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:d00a0059f6134b2519f9b1e034b81a943cae7ce391109703f24db21e56ee923d
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:d68fae6fc07efa9da9af0532aadb49f897a15afcdf9a98dc7517ff5e17735454
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:3130a5159adb9b4eaab75a0209d2ac639156d5e9f2346a8b2e6e0e22fad4b45f
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:80a155d2637f47e249fd69760b250c9b7a1eeb6781de9308b54162bb3e1bce71
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:ffaad2214fa897e89e92e574eb5cd1c671edab7f6eab87cf2ffd4ebf162b3cf4
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:b5db9a51bbb46640b23712c61e9c5a1c7bbdbcd9a53b0b92fa36b80dcaffffc0
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:f23d3ab058667c6fe9ee44dc249aa7e8449c1bb9581380ddbc66624511b286d2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:56e7a0158a713c70765ead847732db4e5f8439aacf31e9a4a22672b583028064
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:e7aa54986c3d388c1bb00828c9ccb294ca0b6c26376c7593ed6bfddda0ef4e5d
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:ab0342b325ee67cc2a7d321aa1eda9c6797dc91fe2b6f26e3cbab53fadf564c7
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:732827652b7a7a49181e86d561831001c28b72b7e12d5e3c1d71b2e496003229
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:018459cbba2e3169bc4be6564cf3747c9cd0677445912abf237f8fe6d58e46ac
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:6f4fa254d923f3260ab0f7d6a0b459554f83b1abfcb575a5ce2c0135fba32427
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:6722aecbbbefa80dfaf93143e033c562cea0982514e2eea27654dd4638e1d072
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:2980eca437e4c1cd22a2b6c2ccae9c4d6b55921b3d575fbd4b9241f1f0a8a3a4
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:dfcd11686f161d44786a205d12f9d628d9e847d40ffdb628ec9aa4a0da0b86f0
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:acb28d84cf03b8551efa32646864005dc3bc9bb341a696544892dfd0c87fbda6
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:818dc7bd69b453ba5882b4ccb82841f2d48c7552368417bac20bef1df64f52ff
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:4ca1bfaf2d7e9477583cddfa40f0b2ce59c20ff9ab20022038774dc3cc49662b
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:f56d21089566330ecbd9aa8acf14da332d3db7360aaf6da00c9a19df83f7334c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:8a4e5e0934197c32e0dfa1efc922d750940587fb61e0df49d476a2d035c53413
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:86074674b7926ad9bfc95556e50820792fcbc5e11ab02879378625ffabdd1889
        version: 1.35.1
      entries:
      - name: serverless-operator.v1.35.1
        version: 1.35.1
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.35
    - currentCSV: serverless-operator.v1.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.36.0 <1.36.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-ddb-streams-source-rhel8@sha256:a903480159ebf1d8652eeaa702d9885fb76bb106c02d75cec5c82c31d77d6410
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-timer-source-rhel8@sha256:8fed5ef9a1f39a55077bb5026c8854d0ee7fc8d8ff5da277ce9c1bfc2ab2474b
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:77e4689b2ce7a713e905553e0e42b8e128045f40cb6c6d4ee082f126cbc46b00
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:3d780ad1ab280b5092b90e277485d589da2a18e8b08090af7d95aecd17af545d
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:429d5d9568cca31fbe83c663276a300286af4cdbf3d35685afdbe5f6f2f8c6e0
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:4344716eb2940c6a3fa9a278250f936602bb1549fece6037912837719c178b24
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:f2d112ba9fdfccea988589b99b463c8a7157475f7a3f03678e8999e65b678cc3
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:5352e4eb4e426a41521276decadf204b6d23b4762a298e4d3dc73969e3eddc68
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:187fb9d64b53c55a029ffce3c0f9e7c3304683fa8a0e832cfb7134e8a3d0c237
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:6210a946f373b9c36c9b31293debe6fed090008b8234c0ba17d37fa81646172e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-transform-jsonata-rhel8@sha256:3af4c48030263eb5769e5f36a6dd13fd390d8566efaab37532f39e1df1c95c67
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-sink-rhel8@sha256:d9b944e64391a996104afbe8a21f66316a948e9fb7b91734e861847285be59c7
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:631eb3a8dd9d83ceafa7d8a4da4b1e2d4f2740e6017c766828dcc7575cf1322f
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:a6f565d73375e827d830dca85838f4c4cb7f08b0c078acb0c0d567b25f0d256a
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:a47b6b5237a87ccd90e5a49a53d06485fd55c87eebf12d01c335b87fc3a8c494
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:ed60e889a677fb139f2eb4bd9ab4fda6ecc39fd050ca2b6cc1c19717e428535a
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:074ebfc4494581611647ba90049655a470ca43743610a27be77db66a8633bac6
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:dcfe5fbc57b9f8520adbff368c8bf29eb491c9c26d1a1737a1722cffea83eb1b
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:f3d87a7fa5b5f2818effcac3304bd56eddcdb0db97a18ac0f35a06fc7e223e7d
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:4274004b6289fa9bea3fa6e56d64077a328a6b5e63a9f0e9c41f6d561f682bd7
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-source-rhel8@sha256:27e74287774cda208f667bc2a386debe6b6255e8991abfb37c959c8525f2af53
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:3b04e9d8cd4b5e43a07ece838d10478723a6c6529350f010550b4a0d1d0950d1
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:ee6533b71cf280877c9d7489dfd821c9209c6802bed7fa3fdb4f8f00fbeacb09
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:6b7316cce3432d391adc6ac38c92ff0ee86dd7ae0101957cc65cd338cc69e352
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:c39cb88afaddc295d93a3774b8a20d001b82392df540fb502ee02316d82460a9
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:82e69a1890042f5636ac7bb645b1aaeaba0a4894851eccc1fa1c63296918e31a
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:52ba6b1252b6f3d41e2a0b3b880cb768010ab9dd925ac34596f6adaedcbbe07c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-sink-rhel8@sha256:afa634e029bcf79a972f49372498d44b45bcfbdb6eb28ecaf4f0b0b2a0a1824b
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:f4d86b196457cdeebdc380b6f1b707c10f8dee789c7779c90ee1332e511af8c4
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:259e5dab50d36cdbfc117a42325128c4e16f34c048ba64b36d4a89132ddb4c1a
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-log-sink-rhel8@sha256:30b8045ef5f925a36c58941386f1899bec125d005a08aeeeba5edd49b0859149
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:3fe326ad9a00b112fc888ffd345c8d5042f0a22cf32ea575931192c39ce1fece
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:e3ae1cc6012f19631c012a07b113fb015977d2ee4f9faab3207fbc2464c25aa7
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:bf82b37ab8fd29c7a7cd846008a16d14b4cfdc1c09df4f731d71be3957e90123
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:d6f338e08fa937c8a4ccfec692424de4acded565f7ca3971ea507dffeb351503
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:4aeb68a22f63300e54391ec6b4cda8c4e665d904b5dcdd3c40155040525742a0
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:596a7272d6de714d7dbc6e00e9e50a2872846c269547465252d953172ef0b009
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:0c2cdf917bbad1f5bca8a2a2ef12a96cfcd0c77c190cc35687eb57169b54ab6f
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:c7c2e738e0d9da1d2d1993f68938e72c4709b1f28a54fa92224ef6a0e456b0aa
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:a1834b620f875681388f0ae224dd0771773a8b61b1d8e5214a942f232854b87b
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:9b328fff0cc8545fb9bb1f24f11d78cfac3ee76dedc11fa4def4262d79a60e8c
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:78fee4c596f2848487ca704da3fa1074e3c9c99b3f96f66833aeba63836d7b93
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sns-sink-rhel8@sha256:51b5e18e8d4c1ca81d7c5ddfea6352a3887f757acfcc53bd2f8bd3b01337ed42
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-source-rhel8@sha256:88323da16360981f27b8e4686fbbb64053e4e0e3c79c65d695edb317b334e1e0
        version: 1.36.1
      entries:
      - name: serverless-operator.v1.36.1
        version: 1.36.1
      - name: serverless-operator.v1.36.0
        version: 1.36.0
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.36
    defaultChannel: stable
    packageName: serverless-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Labs
      provider-url: https://grafana.com
    name: grafana-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: grafana-operator.v5.19.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "Grafana",
                "metadata": {
                  "labels": {
                    "dashboards": "grafana-a",
                    "folders": "grafana-a"
                  },
                  "name": "grafana-a"
                },
                "spec": {
                  "config": {
                    "auth": {
                      "disable_login_form": "false"
                    },
                    "log": {
                      "mode": "console"
                    },
                    "security": {
                      "admin_password": "start",
                      "admin_user": "root"
                    }
                  }
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaAlertRuleGroup",
                "metadata": {
                  "name": "grafanaalertrulegroup-sample"
                },
                "spec": {
                  "folderRef": "test-folder-from-operator",
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "interval": "5m",
                  "rules": [
                    {
                      "condition": "B",
                      "data": [
                        {
                          "datasourceUid": "grafanacloud-demoinfra-prom",
                          "model": {
                            "datasource": {
                              "type": "prometheus",
                              "uid": "grafanacloud-demoinfra-prom"
                            },
                            "editorMode": "code",
                            "expr": "weather_temp_c{}",
                            "instant": true,
                            "intervalMs": 1000,
                            "legendFormat": "__auto",
                            "maxDataPoints": 43200,
                            "range": false,
                            "refId": "A"
                          },
                          "refId": "A",
                          "relativeTimeRange": {
                            "from": 600
                          }
                        },
                        {
                          "datasourceUid": "__expr__",
                          "model": {
                            "conditions": [
                              {
                                "evaluator": {
                                  "params": [
                                    0
                                  ],
                                  "type": "gt"
                                },
                                "operator": {
                                  "type": "and"
                                },
                                "query": {
                                  "params": [
                                    "C"
                                  ]
                                },
                                "reducer": {
                                  "params": [],
                                  "type": "last"
                                },
                                "type": "query"
                              }
                            ],
                            "datasource": {
                              "type": "__expr__",
                              "uid": "__expr__"
                            },
                            "expression": "A",
                            "intervalMs": 1000,
                            "maxDataPoints": 43200,
                            "refId": "B",
                            "type": "threshold"
                          },
                          "refId": "B",
                          "relativeTimeRange": {
                            "from": 600
                          }
                        }
                      ],
                      "execErrState": "Error",
                      "for": "5m0s",
                      "noDataState": "NoData",
                      "title": "Temperature below freezing",
                      "uid": "4843de5c-4f8a-4af0-9509-23526a04faf8"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaContactPoint",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "grafana-operator",
                    "app.kubernetes.io/instance": "grafanacontactpoint-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "grafanacontactpoint",
                    "app.kubernetes.io/part-of": "grafana-operator"
                  },
                  "name": "grafanacontactpoint-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "name": "grafanacontactpoint-sample",
                  "settings": {
                    "email": null
                  },
                  "type": "email"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaDashboard",
                "metadata": {
                  "name": "grafanadashboard-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "json": "{\n\n  \"id\": null,\n  \"title\": \"Simple Dashboard\",\n  \"tags\": [],\n  \"style\": \"dark\",\n  \"timezone\": \"browser\",\n  \"editable\": true,\n  \"hideControls\": false,\n  \"graphTooltip\": 1,\n  \"panels\": [],\n  \"time\": {\n    \"from\": \"now-6h\",\n    \"to\": \"now\"\n  },\n  \"timepicker\": {\n    \"time_options\": [],\n    \"refresh_intervals\": []\n  },\n  \"templating\": {\n    \"list\": []\n  },\n  \"annotations\": {\n    \"list\": []\n  },\n  \"refresh\": \"5s\",\n  \"schemaVersion\": 17,\n  \"version\": 0,\n  \"links\": []\n}\n"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaDatasource",
                "metadata": {
                  "name": "grafanadatasource-sample"
                },
                "spec": {
                  "datasource": {
                    "access": "proxy",
                    "isDefault": true,
                    "jsonData": {
                      "timeInterval": "5s",
                      "tlsSkipVerify": true
                    },
                    "name": "prometheus",
                    "type": "prometheus",
                    "url": "http://prometheus-service:9090"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "plugins": [
                    {
                      "name": "grafana-clock-panel",
                      "version": "1.3.0"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaFolder",
                "metadata": {
                  "name": "grafanafolder-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "title": "Example Folder"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaLibraryPanel",
                "metadata": {
                  "name": "grafana-library-panel-inline-envs"
                },
                "spec": {
                  "envs": [
                    {
                      "name": "CUSTOM_RANGE_ENV",
                      "value": "now - 12h"
                    }
                  ],
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "jsonnet": "local myRange = std.extVar('CUSTOM_RANGE_ENV'); {\n\n  model: {}\n}\n",
                  "plugins": [
                    {
                      "name": "grafana-piechart-panel",
                      "version": "1.3.9"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaMuteTiming",
                "metadata": {
                  "name": "mutetiming-sample"
                },
                "spec": {
                  "editable": false,
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "name": "mutetiming-sample",
                  "time_intervals": [
                    {
                      "days_of_month": [
                        "1",
                        "15"
                      ],
                      "location": "Asia/Shanghai",
                      "times": [
                        {
                          "end_time": "06:00",
                          "start_time": "00:00"
                        }
                      ],
                      "weekdays": [
                        "saturday"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationPolicy",
                "metadata": {
                  "name": "grafananotificationpolicy-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "route": {
                    "group_by": [
                      "grafana_folder",
                      "alertname"
                    ],
                    "receiver": "Grafana Cloud OnCall",
                    "routes": [
                      {
                        "object_matchers": [
                          [
                            "foo",
                            "=",
                            "bar"
                          ]
                        ],
                        "receiver": "grafana-default-email",
                        "routes": [
                          {
                            "object_matchers": [
                              [
                                "severity",
                                "=",
                                "critical"
                              ]
                            ],
                            "receiver": "Grafana Cloud OnCall"
                          }
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationPolicyRoute",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "grafana-operator",
                    "app.kubernetes.io/instance": "grafananotificationpolicyroute-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "grafananotificationpolicyroute",
                    "app.kubernetes.io/part-of": "grafana-operator"
                  },
                  "name": "grafananotificationpolicyroute-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationTemplate",
                "metadata": {
                  "name": "test"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "name": "test",
                  "template": "{{ define \"SlackAlert\" }}\n  [{{.Status}}] {{ .Labels.alertname }}\n  {{ .Annotations.AlertValues }}\n{{ end }}\n\n{{ define \"SlackAlertMessage\" }}\n  {{ if gt (len .Alerts.Firing) 0 }}\n    {{ len .Alerts.Firing }} firing:\n    {{ range .Alerts.Firing }} {{ template \"SlackAlert\" . }} {{ end }}\n  {{ end }}\n  {{ if gt (len .Alerts.Resolved) 0 }}\n    {{ len .Alerts.Resolved }} resolved:\n    {{ range .Alerts.Resolved }} {{ template \"SlackAlert\" . }} {{ end }}\n  {{ end }}\n{{ end }}\n\n{{ template \"SlackAlertMessage\" . }}\n"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: ghcr.io/grafana/grafana-operator@sha256:b678bc3faad11f4ef53c240e2a650ce939b9e8920dd72a47681b2efcadc7b837
          createdAt: "2025-08-21T08:38:00Z"
          description: Deploys and manages Grafana instances, dashboards and data
            sources
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/grafana/grafana-operator
          support: Grafana Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GrafanaAlertRuleGroup is the Schema for the grafanaalertrulegroups
              API
            displayName: Grafana Alert Rule Group
            kind: GrafanaAlertRuleGroup
            name: grafanaalertrulegroups.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaContactPoint is the Schema for the grafanacontactpoints
              API
            displayName: Grafana Contact Point
            kind: GrafanaContactPoint
            name: grafanacontactpoints.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaDashboard is the Schema for the grafanadashboards
              API
            displayName: Grafana Dashboard
            kind: GrafanaDashboard
            name: grafanadashboards.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaDatasource is the Schema for the grafanadatasources
              API
            displayName: Grafana Datasource
            kind: GrafanaDatasource
            name: grafanadatasources.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaFolder is the Schema for the grafanafolders API
            displayName: Grafana Folder
            kind: GrafanaFolder
            name: grafanafolders.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaLibraryPanel
            name: grafanalibrarypanels.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaMuteTiming
            name: grafanamutetimings.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaNotificationPolicy is the Schema for the GrafanaNotificationPolicy
              API
            displayName: Grafana Notification policy
            kind: GrafanaNotificationPolicy
            name: grafananotificationpolicies.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaNotificationPolicyRoute is the Schema for the grafananotificationpolicyroutes
              API
            displayName: Grafana Notification Policy Route
            kind: GrafanaNotificationPolicyRoute
            name: grafananotificationpolicyroutes.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaNotificationTemplate
            name: grafananotificationtemplates.grafana.integreatly.org
            version: v1beta1
          - description: Grafana is the Schema for the grafanas API
            displayName: Grafana
            kind: Grafana
            name: grafanas.grafana.integreatly.org
            version: v1beta1
        description: Deploys and manages Grafana instances, dashboards and data sources
        displayName: Grafana Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Grafana
        - Metrics
        - Observability
        links:
        - name: Grafana Operator
          url: https://grafana.github.io/grafana-operator
        maintainers:
        - email: grafana-operator@grafana.com
          name: Grafana-operator maintainers
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Grafana Labs
          url: https://grafana.com
        relatedImages:
        - ghcr.io/grafana/grafana-operator@sha256:b678bc3faad11f4ef53c240e2a650ce939b9e8920dd72a47681b2efcadc7b837
        - docker.io/grafana/grafana@sha256:6ac590e7cabc2fbe8d7b8fc1ce9c9f0582177b334e0df9c927ebd9670469440f
        version: 5.19.4
      entries:
      - name: grafana-operator.v5.19.4
        version: 5.19.4
      - name: grafana-operator.v5.18.0
        version: 5.18.0
      - name: grafana-operator.v5.17.1
        version: 5.17.1
      - name: grafana-operator.v5.17.0
        version: 5.17.0
      - name: grafana-operator.v5.16.0
        version: 5.16.0
      - name: grafana-operator.v5.15.1
        version: 5.15.1
      - name: grafana-operator.v5.14.0
        version: 5.14.0
      name: v5
    defaultChannel: v5
    packageName: grafana-operator
    provider:
      name: Grafana Labs
      url: https://grafana.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: PingCAP
      provider-url: ""
    name: tidb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tidb-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pingcap.com/v1alpha1",
                "kind": "TidbCluster",
                "metadata": {
                  "name": "basic"
                },
                "spec": {
                  "version": "v5.4.0",
                  "timezone": "UTC",
                  "pvReclaimPolicy": "Delete",
                  "discovery": {},
                  "pd": {
                    "baseImage": "pingcap/pd",
                    "replicas": 1,
                    "requests": {
                      "storage": "1Gi"
                    },
                    "config": {}
                  },
                  "tikv": {
                    "baseImage": "pingcap/tikv",
                    "replicas": 1,
                    "requests": {
                      "storage": "1Gi"
                    },
                    "config": {
                      "storage": {
                        "reserve-space": "0MB"
                      }
                    }
                  },
                  "tidb": {
                    "baseImage": "pingcap/tidb",
                    "replicas": 1,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "config": {}
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: pingcap/tidb-operator:v1.3.1
          createdAt: 2021-2-15T07-18-39Z
          description: TiDB Operator manages TiDB clusters on Kubernetes and automates
            tasks related to operating a TiDB cluster. It makes TiDB a truly cloud-native
            database.
          repository: https://github.com/pingcap/tidb-operator
          support: KanShiori
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup
            displayName: Backup
            kind: Backup
            name: backups.pingcap.com
            version: v1alpha1
          - description: Backup Schedule
            displayName: BackupSchedule
            kind: BackupSchedule
            name: backupschedules.pingcap.com
            version: v1alpha1
          - description: DM Cluster
            displayName: DMCluster
            kind: DMCluster
            name: dmclusters.pingcap.com
            version: v1alpha1
          - description: Restore
            displayName: Restore
            kind: Restore
            name: restores.pingcap.com
            version: v1alpha1
          - description: Tidb Cluster Auto Scaler
            displayName: TidbClusterAutoScaler
            kind: TidbClusterAutoScaler
            name: tidbclusterautoscalers.pingcap.com
            version: v1alpha1
          - description: Tidb Cluster
            displayName: TidbCluster
            kind: TidbCluster
            name: tidbclusters.pingcap.com
            version: v1alpha1
          - description: Tidb Initializer
            displayName: TidbInitializer
            kind: TidbInitializer
            name: tidbinitializers.pingcap.com
            version: v1alpha1
          - description: Tidb Monitor
            displayName: TidbMonitor
            kind: TidbMonitor
            name: tidbmonitors.pingcap.com
            version: v1alpha1
          - description: Tidb NG Monitoring
            displayName: TidbNGMonitoring
            kind: TidbNGMonitoring
            name: tidbngmonitorings.pingcap.com
            version: v1alpha1
        description: |
          ## About the managed application
          TiDB ("Ti" stands for Titanium) is an open-source, distributed, NewSQL database that supports Hybrid Transactional and Analytical Processing (HTAP) workloads. It is MySQL compatible and features horizontal scalability, strong consistency, and high availability. TiDB can be deployed on-premise or in-cloud.
          ## About this Operator

          TiDB Operator manages TiDB clusters on Kubernetes and automates tasks related to operating a TiDB cluster. It makes TiDB a truly cloud-native database.

          ## Features

          * Safely scaling the TiDB cluster

          TiDB Operator empowers TiDB with horizontal scalability on the cloud.

          * Rolling update of the TiDB cluster

          Gracefully perform rolling updates for the TiDB cluster in order, achieving zero-downtime of the TiDB cluster.

          * Multi-tenant support

          Users can deploy and manage multiple TiDB clusters on a single Kubernetes cluster easily.

          * Automatic failover

          TiDB Operator automatically performs failover for your TiDB cluster when node failures occur.

          * Kubernetes package manager support

          By embracing Kubernetes package manager Helm, users can easily deploy TiDB clusters with only one command.

          * Automatically monitoring TiDB cluster at creating

          Automatically deploy Prometheus, Grafana for TiDB cluster monitoring.

          ## Quick Start
          You can follow our [Get Started](https://docs.pingcap.com/tidb-in-kubernetes/stable/get-started) guide to quickly start a testing Kubernetes cluster and play with TiDB Operator on your own machine.

          ## Documentation

          You can see our documentation at PingCAP website for more in-depth installation and instructions for production:

          - [English](https://docs.pingcap.com/tidb-in-kubernetes/stable)
          - [简体中文](https://docs.pingcap.com/zh/tidb-in-kubernetes/stable)

          All the TiDB Operator documentation is maintained in the [docs-tidb-operator repository](https://github.com/pingcap/docs-tidb-operator).

          ## Community

          Feel free to reach out if you have any questions. The maintainers of this project are reachable via:

          - [TiDB Community Slack](https://pingcap.com/tidbslack/) in the [#sig-k8s](https://tidbcommunity.slack.com/archives/CHD0HA3LZ) channel
          - [Filing an issue](https://github.com/pingcap/tidb-operator/issue) against this repo

          Pull Requests are welcome! Check the [issue tracker](https://github.com/pingcap/tidb-operator/issue) for `status/help-wanted` issues if you're unsure where to start.

          If you're planning a new feature, please file an issue or join [#sig-k8s](https://tidbcommunity.slack.com/archives/CHD0HA3LZ) channel to discuss first.
          ## Prerequisites for enabling this Operator
        displayName: TiDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TiDB
        - SQL
        - Database
        links:
        - name: GitHub
          url: https://github.com/pingcap/tidb-operator
        - name: Documentation
          url: https://docs.pingcap.com/tidb-in-kubernetes/stable
        maintainers:
        - email: zhanghailong@pingcap.com
          name: Daniel Zhang
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: PingCAP
        relatedImages:
        - pingcap/tidb-operator:v1.3.1
        version: 1.3.1
      entries:
      - name: tidb-operator.v1.3.1
        version: 1.3.1
      name: stable
    defaultChannel: stable
    packageName: tidb-operator
    provider:
      name: PingCAP
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v0.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: gcr.io/datadoghq/operator:0.8.4
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          olm.skipRange: <0.8.4
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - gcr.io/datadoghq/operator:0.8.4
        version: 0.8.4
      entries:
      - name: datadog-operator.v0.8.4
        version: 0.8.4
      - name: datadog-operator.v0.8.2
        version: 0.8.2
      - name: datadog-operator.v0.8.1
        version: 0.8.1
      - name: datadog-operator.v0.8.0
        version: 0.8.0
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: gcr.io/datadoghq/operator:1.18.0
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - gcr.io/datadoghq/operator:1.18.0
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.2
        version: 1.4.2
      - name: datadog-operator.v1.4.1
        version: 1.4.1
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      - name: datadog-operator.v1.2.0
        version: 1.2.0
      - name: datadog-operator.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: datadog-operator
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: servicemeshoperator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: servicemeshoperator.v2.6.10
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"maistra.io/v2\",\n    \"kind\":
            \"ServiceMeshControlPlane\",\n    \"metadata\": {\n      \"name\": \"basic\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"version\": \"v2.6\",\n      \"policy\": {\n         \"type\":
            \"Istiod\"\n      },\n      \"telemetry\": {\n         \"type\": \"Istiod\"\n
            \     },\n      \"addons\": {\n        \"prometheus\": {\n          \"enabled\":
            true\n        },\n        \"kiali\": {\n          \"enabled\": true\n
            \       },\n        \"grafana\": {\n          \"enabled\": true\n        }\n
            \     }\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n    \"kind\":
            \"ServiceMeshMemberRoll\",\n    \"metadata\": {\n      \"name\": \"default\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"members\": [\n        \"your-project\",\n        \"another-of-your-projects\"
            \n      ]\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n
            \   \"kind\": \"ServiceMeshMember\",\n    \"metadata\": {\n      \"name\":
            \"default\",\n      \"namespace\": \"application-namespace\"\n    },\n
            \   \"spec\": {\n      \"controlPlaneRef\": {\n        \"name\": \"basic\",\n
            \       \"namespace\": \"control-plane-namespace\"\n      }\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
          createdAt: 2025-08-11T09:32:44CEST
          description: The OpenShift Service Mesh 2 Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh 2.6
            and earlier releases. OpenShift Service Mesh 2 is based on the open source
            Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2.6.10-0'
          operators.openshift.io/infrastructure-features: '["Disconnected","fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/maistra/istio-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Istio control plane installation
            displayName: Istio Service Mesh Control Plane
            kind: ServiceMeshControlPlane
            name: servicemeshcontrolplanes.maistra.io
            version: v2
          - description: Marks the containing namespace as a member of the referenced
              Service Mesh
            displayName: Istio Service Mesh Member
            kind: ServiceMeshMember
            name: servicemeshmembers.maistra.io
            version: v1
          - description: A list of namespaces in Service Mesh
            displayName: Istio Service Mesh Member Roll
            kind: ServiceMeshMemberRoll
            name: servicemeshmemberrolls.maistra.io
            version: v1
        description: |-
          Red Hat OpenShift Service Mesh 2 is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh 2, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh 2
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh 2 provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh 2 supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### Joining Projects Into a Mesh

          Once an instance of Red Hat OpenShift Service Mesh 2 has been installed, it will only exercise control over services within its own
          project.  Other projects may be added into the mesh using one of two methods:

          1. A **ServiceMeshMember** resource may be created in other projects to add those projects into the mesh.  The
            **ServiceMeshMember** specifies a reference to the **ServiceMeshControlPlane** object that was used to install
            the control plane.  The user creating the **ServiceMeshMember** resource must have permission to *use* the
            **ServiceMeshControlPlane** object.  The adminstrator for the project containing the control plane can grant
            individual users or groups the *use* permissions.

          2. A **ServiceMeshMemberRoll** resource may be created in the project containing the control plane.  This resource
            contains a single *members* list of all the projects that should belong in the mesh.  The resource must be named
            *default*.  The user creating the resource must have *edit* or *admin* permissions for all projects in the
            *members* list.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 2
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - maistra
        - servicemesh
        links:
        - name: Red Hat OpenShift Service Mesh 2
          url: https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html
        - name: Istio
          url: https://istio.io/
        - name: Operator Source Code
          url: https://github.com/Maistra/istio-operator
        - name: Bugs
          url: https://issues.redhat.com/projects/OSSM
        maintainers:
        - email: istio-feedback@redhat.com
          name: Red Hat, OpenShift Service Mesh
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:4ed63a5440753b1b22d44191246c82e7671e3f545c1dbdae4cb4fad163ca3a46
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6b4623c416fc85bd8e38904c16494fa3c4aeae25c90dc0da9763a0a87141bc04
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:515c5d9e6700fae981b0ea3807436b250acb1534e9ecf50043e5df8474e66dce
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:b134578e0677963197e353d8acd0ac3330df1b6d2752aa41cb072e86b5b4b91b
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6fb7d50b188dc27ffb9a396edabbfb21767fe438344fd70d4754d1ec7cf27957
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:e734ca310e09f737b4f06f3ecfeb9ac12971c63884fdcc27e3bebef9d5b18722
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:a36d5db6097749c55e946bba3b8ddd13a84b1f6f68b62d0ccbad5a803e9b46fb
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:fc500a282c6668ff987a632efbbd19ecc3aebdc021f831a7b386e360a5c8324d
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:57eb7ba03f24781501a7ceed2e160ff95317869af6c2caf8f9a83990e6a6c7b9
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:1ca80927334c6af8e386de8d44704f568b4b35d3f7afbe785a999d0fa4efd0bf
        - registry.redhat.io/openshift4/ose-prometheus@sha256:423ae2e81cfe660b7217f7662754f366a91d457d6dd1589b4fc40733896eb9d6
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:d16a60204d266bf4cc266c005fabf175b533b827500f171013884a713c88857a
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel8@sha256:3d773d16fed72278949b9e3b54dc59e346e95341dc94b715bcfc328d3c9e83b2
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:6a420228e684596ce2b900614a874f3e1284a6f2bbbba4c5a71ade158867affe
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:c77183e50a34d33b5e2fede60350f54b18ab3536c3189e331569d5df862d9979
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:a3d65983d46dcf47b94f26e7265ed32082208ce08abcf3587a9a4e15c19b274a
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:b241452cce5c9af5aa0f12d1db3a7094cc004ce355988b8a5207091810a6f84b
        - registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:3dc1bddb070fd530ce6251666989b9a849de90591718a578da4a9900f98d93ec
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:c34b4cd9a82fef40708a25f0568f6dc632878affba35d5d631a48ec31da08cd3
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:c1cadf40bdb9511d612a0257c12eb3a9738823592ebb6e32a4f51b3a8dfe08db
        - registry.redhat.io/openshift-service-mesh/prometheus-rhel8@sha256:71e193a2d403bb06a7a76686b376985315aa0c32f495b8b9dacd73c7ccce2af3
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d95762c62087072cadee5948c706619f09202a349e1310f0ed81bfb42f11203f
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:4669724b430683bf2e895c5cf941c2d99536117815999d38131e5b6f162bbef0
        version: 2.6.10-0
      entries:
      - name: servicemeshoperator.v2.6.10
        version: 2.6.10-0
      - name: servicemeshoperator.v2.6.9
        version: 2.6.9-0
      - name: servicemeshoperator.v2.6.8
        version: 2.6.8-0
      - name: servicemeshoperator.v2.6.7
        version: 2.6.7-0
      - name: servicemeshoperator.v2.6.6
        version: 2.6.6-0
      - name: servicemeshoperator.v2.6.5
        version: 2.6.5-0
      - name: servicemeshoperator.v2.6.4
        version: 2.6.4-0
      - name: servicemeshoperator.v2.6.3
        version: 2.6.3-0
      - name: servicemeshoperator.v2.6.2
        version: 2.6.2-0
      - name: servicemeshoperator.v2.6.1
        version: 2.6.1-0
      - name: servicemeshoperator.v2.6.0
        version: 2.6.0-0
      - name: servicemeshoperator.v2.5.2
        version: 2.5.2-0
      - name: servicemeshoperator.v2.5.1
        version: 2.5.1-0
      - name: servicemeshoperator.v2.5.0
        version: 2.5.0-0
      - name: servicemeshoperator.v2.4.5
        version: 2.4.5-0
      - name: servicemeshoperator.v2.4.4
        version: 2.4.4-0
      - name: servicemeshoperator.v2.4.3
        version: 2.4.3-0
      - name: servicemeshoperator.v2.4.2
        version: 2.4.2-0
      - name: servicemeshoperator.v2.4.1
        version: 2.4.1-0
      - name: servicemeshoperator.v2.4.0
        version: 2.4.0-0
      - name: servicemeshoperator.v2.3.3
        version: 2.3.3-0
      - name: servicemeshoperator.v2.3.2
        version: 2.3.2-0
      - name: servicemeshoperator.v2.3.1
        version: 2.3.1-0
      - name: servicemeshoperator.v2.3.0
        version: 2.3.0-0
      - name: servicemeshoperator.v2.2.3
        version: 2.2.3-0
      - name: servicemeshoperator.v2.2.2
        version: 2.2.2-0
      - name: servicemeshoperator.v2.2.1
        version: 2.2.1-0
      - name: servicemeshoperator.v2.2.0
        version: 2.2.0-0
      - name: servicemeshoperator.v2.1.3
        version: 2.1.3-0
      - name: servicemeshoperator.v2.1.2
        version: 2.1.2-0
      - name: servicemeshoperator.v2.1.1
        version: 2.1.1-0
      - name: servicemeshoperator.v2.1.0
        version: 2.1.0-0
      - name: servicemeshoperator.v2.0.8
        version: 2.0.8-0
      name: "1.0"
    - currentCSV: servicemeshoperator.v2.6.10
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"maistra.io/v2\",\n    \"kind\":
            \"ServiceMeshControlPlane\",\n    \"metadata\": {\n      \"name\": \"basic\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"version\": \"v2.6\",\n      \"policy\": {\n         \"type\":
            \"Istiod\"\n      },\n      \"telemetry\": {\n         \"type\": \"Istiod\"\n
            \     },\n      \"addons\": {\n        \"prometheus\": {\n          \"enabled\":
            true\n        },\n        \"kiali\": {\n          \"enabled\": true\n
            \       },\n        \"grafana\": {\n          \"enabled\": true\n        }\n
            \     }\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n    \"kind\":
            \"ServiceMeshMemberRoll\",\n    \"metadata\": {\n      \"name\": \"default\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"members\": [\n        \"your-project\",\n        \"another-of-your-projects\"
            \n      ]\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n
            \   \"kind\": \"ServiceMeshMember\",\n    \"metadata\": {\n      \"name\":
            \"default\",\n      \"namespace\": \"application-namespace\"\n    },\n
            \   \"spec\": {\n      \"controlPlaneRef\": {\n        \"name\": \"basic\",\n
            \       \"namespace\": \"control-plane-namespace\"\n      }\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
          createdAt: 2025-08-11T09:32:44CEST
          description: The OpenShift Service Mesh 2 Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh 2.6
            and earlier releases. OpenShift Service Mesh 2 is based on the open source
            Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2.6.10-0'
          operators.openshift.io/infrastructure-features: '["Disconnected","fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/maistra/istio-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Istio control plane installation
            displayName: Istio Service Mesh Control Plane
            kind: ServiceMeshControlPlane
            name: servicemeshcontrolplanes.maistra.io
            version: v2
          - description: Marks the containing namespace as a member of the referenced
              Service Mesh
            displayName: Istio Service Mesh Member
            kind: ServiceMeshMember
            name: servicemeshmembers.maistra.io
            version: v1
          - description: A list of namespaces in Service Mesh
            displayName: Istio Service Mesh Member Roll
            kind: ServiceMeshMemberRoll
            name: servicemeshmemberrolls.maistra.io
            version: v1
        description: |-
          Red Hat OpenShift Service Mesh 2 is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh 2, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh 2
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh 2 provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh 2 supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### Joining Projects Into a Mesh

          Once an instance of Red Hat OpenShift Service Mesh 2 has been installed, it will only exercise control over services within its own
          project.  Other projects may be added into the mesh using one of two methods:

          1. A **ServiceMeshMember** resource may be created in other projects to add those projects into the mesh.  The
            **ServiceMeshMember** specifies a reference to the **ServiceMeshControlPlane** object that was used to install
            the control plane.  The user creating the **ServiceMeshMember** resource must have permission to *use* the
            **ServiceMeshControlPlane** object.  The adminstrator for the project containing the control plane can grant
            individual users or groups the *use* permissions.

          2. A **ServiceMeshMemberRoll** resource may be created in the project containing the control plane.  This resource
            contains a single *members* list of all the projects that should belong in the mesh.  The resource must be named
            *default*.  The user creating the resource must have *edit* or *admin* permissions for all projects in the
            *members* list.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 2
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - maistra
        - servicemesh
        links:
        - name: Red Hat OpenShift Service Mesh 2
          url: https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html
        - name: Istio
          url: https://istio.io/
        - name: Operator Source Code
          url: https://github.com/Maistra/istio-operator
        - name: Bugs
          url: https://issues.redhat.com/projects/OSSM
        maintainers:
        - email: istio-feedback@redhat.com
          name: Red Hat, OpenShift Service Mesh
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:c1cadf40bdb9511d612a0257c12eb3a9738823592ebb6e32a4f51b3a8dfe08db
        - registry.redhat.io/openshift4/ose-prometheus@sha256:423ae2e81cfe660b7217f7662754f366a91d457d6dd1589b4fc40733896eb9d6
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:4ed63a5440753b1b22d44191246c82e7671e3f545c1dbdae4cb4fad163ca3a46
        - registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:57eb7ba03f24781501a7ceed2e160ff95317869af6c2caf8f9a83990e6a6c7b9
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:3dc1bddb070fd530ce6251666989b9a849de90591718a578da4a9900f98d93ec
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel8@sha256:3d773d16fed72278949b9e3b54dc59e346e95341dc94b715bcfc328d3c9e83b2
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:a3d65983d46dcf47b94f26e7265ed32082208ce08abcf3587a9a4e15c19b274a
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:6a420228e684596ce2b900614a874f3e1284a6f2bbbba4c5a71ade158867affe
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:c77183e50a34d33b5e2fede60350f54b18ab3536c3189e331569d5df862d9979
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:d16a60204d266bf4cc266c005fabf175b533b827500f171013884a713c88857a
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:1ca80927334c6af8e386de8d44704f568b4b35d3f7afbe785a999d0fa4efd0bf
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6b4623c416fc85bd8e38904c16494fa3c4aeae25c90dc0da9763a0a87141bc04
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:515c5d9e6700fae981b0ea3807436b250acb1534e9ecf50043e5df8474e66dce
        - registry.redhat.io/openshift-service-mesh/prometheus-rhel8@sha256:71e193a2d403bb06a7a76686b376985315aa0c32f495b8b9dacd73c7ccce2af3
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d95762c62087072cadee5948c706619f09202a349e1310f0ed81bfb42f11203f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:e734ca310e09f737b4f06f3ecfeb9ac12971c63884fdcc27e3bebef9d5b18722
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:a36d5db6097749c55e946bba3b8ddd13a84b1f6f68b62d0ccbad5a803e9b46fb
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:b134578e0677963197e353d8acd0ac3330df1b6d2752aa41cb072e86b5b4b91b
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6fb7d50b188dc27ffb9a396edabbfb21767fe438344fd70d4754d1ec7cf27957
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:4669724b430683bf2e895c5cf941c2d99536117815999d38131e5b6f162bbef0
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:fc500a282c6668ff987a632efbbd19ecc3aebdc021f831a7b386e360a5c8324d
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:c34b4cd9a82fef40708a25f0568f6dc632878affba35d5d631a48ec31da08cd3
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:b241452cce5c9af5aa0f12d1db3a7094cc004ce355988b8a5207091810a6f84b
        version: 2.6.10-0
      entries:
      - name: servicemeshoperator.v2.6.10
        version: 2.6.10-0
      - name: servicemeshoperator.v2.6.9
        version: 2.6.9-0
      - name: servicemeshoperator.v2.6.8
        version: 2.6.8-0
      - name: servicemeshoperator.v2.6.7
        version: 2.6.7-0
      - name: servicemeshoperator.v2.6.6
        version: 2.6.6-0
      - name: servicemeshoperator.v2.6.5
        version: 2.6.5-0
      - name: servicemeshoperator.v2.6.4
        version: 2.6.4-0
      - name: servicemeshoperator.v2.6.3
        version: 2.6.3-0
      - name: servicemeshoperator.v2.6.2
        version: 2.6.2-0
      - name: servicemeshoperator.v2.6.1
        version: 2.6.1-0
      - name: servicemeshoperator.v2.6.0
        version: 2.6.0-0
      - name: servicemeshoperator.v2.5.2
        version: 2.5.2-0
      - name: servicemeshoperator.v2.5.1
        version: 2.5.1-0
      - name: servicemeshoperator.v2.5.0
        version: 2.5.0-0
      - name: servicemeshoperator.v2.4.5
        version: 2.4.5-0
      - name: servicemeshoperator.v2.4.4
        version: 2.4.4-0
      - name: servicemeshoperator.v2.4.3
        version: 2.4.3-0
      - name: servicemeshoperator.v2.4.2
        version: 2.4.2-0
      - name: servicemeshoperator.v2.4.1
        version: 2.4.1-0
      - name: servicemeshoperator.v2.4.0
        version: 2.4.0-0
      - name: servicemeshoperator.v2.3.3
        version: 2.3.3-0
      - name: servicemeshoperator.v2.3.2
        version: 2.3.2-0
      - name: servicemeshoperator.v2.3.1
        version: 2.3.1-0
      - name: servicemeshoperator.v2.3.0
        version: 2.3.0-0
      - name: servicemeshoperator.v2.2.3
        version: 2.2.3-0
      - name: servicemeshoperator.v2.2.2
        version: 2.2.2-0
      - name: servicemeshoperator.v2.2.1
        version: 2.2.1-0
      - name: servicemeshoperator.v2.2.0
        version: 2.2.0-0
      - name: servicemeshoperator.v2.1.3
        version: 2.1.3-0
      - name: servicemeshoperator.v2.1.2
        version: 2.1.2-0
      - name: servicemeshoperator.v2.1.1
        version: 2.1.1-0
      - name: servicemeshoperator.v2.1.0
        version: 2.1.0-0
      - name: servicemeshoperator.v2.0.8
        version: 2.0.8-0
      name: stable
    defaultChannel: stable
    packageName: servicemeshoperator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: service-binding-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-binding-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:b355d2370926063a3de8c7a09c0ca8c2348a8d158d6a7095d0e5e8381cc95be5
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/redhat-developer/servicebinding-operator@sha256:b355d2370926063a3de8c7a09c0ca8c2348a8d158d6a7095d0e5e8381cc95be5
        version: 0.10.0
      entries:
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: beta
    - currentCSV: service-binding-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindablekinds-sample"
                }
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:c9c9e45321db2d4d3cdc2d364337af844f69c9893d0cc74d2fa56c846ffcf3da
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/redhat-developer/servicebinding-operator@sha256:c9c9e45321db2d4d3cdc2d364337af844f69c9893d0cc74d2fa56c846ffcf3da
        version: 0.11.0
      entries:
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: candidate
    - currentCSV: service-binding-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindable-kinds"
                },
                "spec": {}
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  },
                  "type": "postgres",
                  "workload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:ac47f496fb7ecdcbc371f8c809fad2687ec0c35bbc8c522a7ab63b3e5ffd90ea
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'.
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1alpha3
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1beta1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
          - description: Service Binding implementing community specification (https://github.com/k8s-service-bindings/spec)
              describing the connection between a backing service and an application
              workload.
            displayName: Service Binding (Spec API)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1beta1
        description: |
          The Service Binding Operator manages the data plane for applications and backing services. It enables developers to connect their
          applications to backing services (REST API, databases, event buses and others) with a consistent and predictable
          experience. The operator reads data made available by the control plane of backing services and projects the bindings
          data to applications either as accessible files mounted into the application’s container or environment variables.

          In Kubernetes, each service suggests a different way to access their secrets and developers will consume them in their
          application in a custom way. While this provides a good deal of flexibility, it also creates discrepancies and
          complexity dealing with each specific services, which leads large development teams to lose overall velocity.

          Service Binding Operator removes the error-prone manual configuration of binding information, provides service operators
          a low-touch administrative experience to provision and manage access to services and enriches development lifecycle.

          For additional details refer to the project [documentation](https://redhat-developer.github.io/service-binding-operator).

          Get Started with Service Binding Operator
          * Find out about SBO’s [prerequisites](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/installing-service-binding.html).
          * Refer to the [Quick Start Guide](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/quick-start.html) to see SBO in action in a simple scenario.

          Need Help
          * Raise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)

          Licence
          * Service Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE)

          Service Binding Operator is providing and supporting two different APIs.
          * `binding.operators.coreos.com/v1alpha1`: This API is compliant with the Service Binding specification. It is fully supported and considered as the mature API. It’s the API we recommend in production environments.
          * `servicebinding.io/v1beta1`: This API is the one defined in the Service Binding specification. It is available in Service Binding Operator 1.x.

          We’ll consider supporting only the API from the specification, once the Service Binding specification will be officially GA. Of course, that will be a breaking change and handle in a safe way for you.
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Documentation
          url: https://redhat-developer.github.io/service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/redhat-developer/servicebinding-operator@sha256:ac47f496fb7ecdcbc371f8c809fad2687ec0c35bbc8c522a7ab63b3e5ffd90ea
        version: 1.4.0
      entries:
      - name: service-binding-operator.v1.4.0
        version: 1.4.0
      - name: service-binding-operator.v1.3.4
        version: 1.3.4
      - name: service-binding-operator.v1.3.3
        version: 1.3.3
      - name: service-binding-operator.v1.3.2
        version: 1.3.2
      - name: service-binding-operator.v1.3.1
        version: 1.3.1
      - name: service-binding-operator.v1.3.0
        version: 1.3.0
      - name: service-binding-operator.v1.2.0
        version: 1.2.0
      - name: service-binding-operator.v1.1.1
        version: 1.1.1
      - name: service-binding-operator.v1.1.0
        version: 1.1.0
      - name: service-binding-operator.v1.0.1
        version: 1.0.1
      - name: service-binding-operator.v1.0.0
        version: 1.0.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: service-binding-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: netobserv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netobserv-operator.v1.6.3-community
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowMetric",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netobserv-operator",
                    "app.kubernetes.io/instance": "flowmetric-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "flowmetric",
                    "app.kubernetes.io/part-of": "netobserv-operator"
                  },
                  "name": "flowmetric-sample"
                },
                "spec": {
                  "charts": [
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "",
                          "promQL": "sum(rate($METRIC[2m]))"
                        }
                      ],
                      "title": "External ingress traffic",
                      "type": "SingleStat",
                      "unit": "Bps"
                    },
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "{{DstK8S_Namespace}} / {{DstK8S_OwnerName}}",
                          "promQL": "sum(rate($METRIC{DstK8S_Namespace!=\"\"}[2m])) by (DstK8S_Namespace, DstK8S_OwnerName)"
                        }
                      ],
                      "sectionName": "External",
                      "title": "Top external ingress traffic per workload",
                      "type": "StackArea",
                      "unit": "Bps"
                    }
                  ],
                  "direction": "Ingress",
                  "filters": [
                    {
                      "field": "SrcSubnetLabel",
                      "matchType": "Absence"
                    }
                  ],
                  "labels": [
                    "DstK8S_HostName",
                    "DstK8S_Namespace",
                    "DstK8S_OwnerName",
                    "DstK8S_OwnerType"
                  ],
                  "metricName": "cluster_external_ingress_bytes_total",
                  "type": "Counter",
                  "valueField": "Bytes"
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "app"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "infra"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "enable": true,
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "includeList": [
                        "node_ingress_bytes_total",
                        "workload_ingress_bytes_total",
                        "namespace_flows_total"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta2",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "metrics": {
                        "server": {
                          "port": 9400
                        }
                      },
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "eBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "Disabled"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "\"app\""
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "\"infra\""
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "\"Pod\"",
                          "src_kind": "\"Pod\""
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "\"Service\""
                        },
                        "name": "Services network"
                      }
                    ],
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    }
                  },
                  "deploymentModel": "Direct",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "enable": true,
                    "lokiStack": {
                      "name": "loki"
                    },
                    "mode": "Monolithic",
                    "monolithic": {
                      "tenantID": "netobserv",
                      "tls": {
                        "caCert": {
                          "certFile": "service-ca.crt",
                          "name": "loki-gateway-ca-bundle",
                          "type": "configmap"
                        },
                        "enable": false
                      },
                      "url": "http://loki.netobserv.svc:3100/"
                    },
                    "readTimeout": "30s",
                    "writeBatchSize": 10485760,
                    "writeBatchWait": "1s",
                    "writeTimeout": "10s"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "Flows",
                    "metrics": {
                      "disableAlerts": [],
                      "server": {
                        "port": 9401
                      }
                    },
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "prometheus": {
                    "querier": {
                      "enable": true,
                      "mode": "Auto",
                      "timeout": "30s"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:1.6.2-community
          createdAt: "2024-08-22T18:26:56Z"
          description: Network flows collector and monitoring solution
          operatorframework.io/initialization-resource: '{"apiVersion":"flows.netobserv.io/v1beta2",
            "kind":"FlowCollector","metadata":{"name":"cluster"},"spec": {}}'
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
          support: NetObserv team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta2
          - description: '`FlowMetric` is the schema for the custom metrics API, which
              allows to generate more metrics out of flow logs. You can find examples
              here: https://github.com/netobserv/network-observability-operator/tree/1.6.2-community/config/samples/flowmetrics'
            displayName: Flow Metric
            kind: FlowMetric
            name: flowmetrics.flows.netobserv.io
            version: v1alpha1
        description: |-
          DEPRECATED: this distribution channel is deprecated and doesn't receive new updates. Please read https://github.com/orgs/netobserv/discussions/996 to learn more.

          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline that consists in:
          - an eBPF agent, that generates network flows from captured packets
          - flowlogs-pipeline, a component that collects, enriches and exports these flows
          - when used in OpenShift, a Console plugin for flows visualization with powerful filtering options, a topology representation and more

          Flow data is then available in multiple ways, each optional:

          - As Prometheus metrics
          - As raw flow logs stored in Grafana Loki
          - As raw flow logs exported to a collector

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, can optionally be used as the backend to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 10GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          If you prefer to not use Loki, you must set `spec.loki.enable` to `false` in `FlowCollector`.
          In that case, you can still get the Prometheus metrics or export raw flows to a custom collector. But be aware that some of the Console plugin features will be disabled. For instance, you will not be able to view raw flows there, and the metrics / topology will have a more limited level of details, missing information such as pods or IPs.

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for example [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/config/samples/flows_v1beta2_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide on every node, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you might have to configure differently if you used another installation method. Make sure to disable it (`spec.loki.enable`) if you don't want to use Loki.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`) an optional list of exporters to which to send enriched flows. KAFKA and IPFIX exporters are supported. This allows you to define any custom storage or processing that can read from Kafka or use the IPFIX standard.

          - To enable availability zones awareness, set `spec.processor.addZone` to `true`.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - More information on [configuring metrics](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/Metrics.md).
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: '[DEPRECATED] NetObserv Operator'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/network-observability-operator:1.6.2-community
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/netobserv/netobserv-ebpf-agent:v1.6.2-community
        - quay.io/netobserv/flowlogs-pipeline:v1.6.2-community
        - quay.io/netobserv/network-observability-console-plugin:v1.6.2-community
        version: 1.6.3-community
      entries:
      - name: netobserv-operator.v1.6.3-community
        version: 1.6.3-community
      - name: netobserv-operator.v1.6.2-community
        version: 1.6.2-community
      - name: netobserv-operator.v1.6.1-community
        version: 1.6.1-community
      - name: netobserv-operator.v1.6.0-community
        version: 1.6.0-community
      - name: netobserv-operator.v1.5.0-community
        version: 1.5.0-community
      name: community
    - currentCSV: netobserv-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "ipfix": {
                      "cacheActiveTimeout": "20s",
                      "cacheMaxFlows": 400,
                      "clusterNetworkOperator": {
                        "namespace": "openshift-network-operator"
                      },
                      "ovnKubernetes": {
                        "containerName": "ovnkube-node",
                        "daemonSetName": "ovnkube-node",
                        "namespace": "ovn-kubernetes"
                      },
                      "sampling": 400
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "staticLabels": {
                      "app": "netobserv-flowcollector"
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "enableKubeProbes": true,
                    "healthPort": 8080,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:0.2.1
          createdAt: "2022-12-07T09:00:48Z"
          description: Network flows collector and monitoring solution
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
        description: |-
          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline to collect and enrich network flows. These flows can be produced by the NetObserv eBPF agent, or by any device or CNI able to export flows in IPFIX format, such as OVN-Kubernetes.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When used in OpenShift, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For a normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionnaly be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/0.2.1/config/samples/flows_v1alpha1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Agent (`spec.agent.type`) can be `EBPF` (default) or `IPFIX`. eBPF is recommended, as it should work in more situations and offers better performances. If you can't, or don't want to use eBPF, note that the IPFIX option is fully functional only when using [OVN-Kubernetes](https://github.com/ovn-org/ovn-kubernetes/) CNI. Other CNIs are not officially supported, but you may still be able to configure them manually if they allow IPFIX exports.

          - Sampling (`spec.agent.ebpf.sampling` and `spec.agent.ipfix.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50) for eBPF and 400 (1:400) for IPFIX. Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub more more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - A [F.A.Q](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: NetObserv Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: mmaciasl@redhat.com
          name: Mario Macias
        - email: ocazade@redhat.com
          name: Olivier Cazade
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/netobserv-ebpf-agent:v0.2.2
        - quay.io/netobserv/flowlogs-pipeline:v0.1.5
        - quay.io/netobserv/network-observability-console-plugin:v0.1.6
        - quay.io/netobserv/network-observability-operator:0.2.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        version: 0.2.1
      entries:
      - name: netobserv-operator.v0.2.1
        version: 0.2.1
      - name: netobserv-operator.v0.2.0
        version: 0.2.0
      - name: netobserv-operator.v0.1.4
        version: 0.1.4
      - name: netobserv-operator.v0.1.3
        version: 0.1.3
      - name: netobserv-operator.v0.1.2
        version: 0.1.2
      - name: netobserv-operator.v0.1.1
        version: 0.1.1
      - name: netobserv-operator.v0.1.0
        version: 0.1.0
      name: latest
    - currentCSV: netobserv-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "ipfix": {
                      "cacheActiveTimeout": "20s",
                      "cacheMaxFlows": 400,
                      "clusterNetworkOperator": {
                        "namespace": "openshift-network-operator"
                      },
                      "ovnKubernetes": {
                        "containerName": "ovnkube-node",
                        "daemonSetName": "ovnkube-node",
                        "namespace": "ovn-kubernetes"
                      },
                      "sampling": 400
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "staticLabels": {
                      "app": "netobserv-flowcollector"
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "enableKubeProbes": true,
                    "healthPort": 8080,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:0.2.1
          createdAt: "2022-12-07T09:00:48Z"
          description: Network flows collector and monitoring solution
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
        description: |-
          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline to collect and enrich network flows. These flows can be produced by the NetObserv eBPF agent, or by any device or CNI able to export flows in IPFIX format, such as OVN-Kubernetes.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When used in OpenShift, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For a normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionnaly be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/0.2.1/config/samples/flows_v1alpha1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Agent (`spec.agent.type`) can be `EBPF` (default) or `IPFIX`. eBPF is recommended, as it should work in more situations and offers better performances. If you can't, or don't want to use eBPF, note that the IPFIX option is fully functional only when using [OVN-Kubernetes](https://github.com/ovn-org/ovn-kubernetes/) CNI. Other CNIs are not officially supported, but you may still be able to configure them manually if they allow IPFIX exports.

          - Sampling (`spec.agent.ebpf.sampling` and `spec.agent.ipfix.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50) for eBPF and 400 (1:400) for IPFIX. Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub more more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - A [F.A.Q](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: NetObserv Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: mmaciasl@redhat.com
          name: Mario Macias
        - email: ocazade@redhat.com
          name: Olivier Cazade
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/network-observability-operator:0.2.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/netobserv/network-observability-console-plugin:v0.1.6
        - quay.io/netobserv/netobserv-ebpf-agent:v0.2.2
        - quay.io/netobserv/flowlogs-pipeline:v0.1.5
        version: 0.2.1
      entries:
      - name: netobserv-operator.v0.2.1
        version: 0.2.1
      - name: netobserv-operator.v0.2.0
        version: 0.2.0
      - name: netobserv-operator.v0.1.4
        version: 0.1.4
      - name: netobserv-operator.v0.1.3
        version: 0.1.3
      - name: netobserv-operator.v0.1.2
        version: 0.1.2
      - name: netobserv-operator.v0.1.1
        version: 0.1.1
      - name: netobserv-operator.v0.1.0
        version: 0.1.0
      name: v0.2.x
    defaultChannel: community
    packageName: netobserv-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-license-service-scanner-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-license-service-scanner-operator.v4.2.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1",
                "kind": "IBMLicenseServiceScanner",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-licensing-scanner-operator",
                    "app.kubernetes.io/instance": "ibm-licensing-scanner-instance",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ibm-licensing-scanner-instance",
                    "app.kubernetes.io/part-of": "ibm-licensing-scanner-operator"
                  },
                  "name": "ibm-licensing-scanner-instance"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "license-service-upload-secret": "ibm-licensing-scanner-upload-secret",
                  "log-level": "INFO",
                  "registry-pull-secret": "ibm-licensing-scanner-registry-pull-secret",
                  "scan": {
                    "namespaces": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:77c9db26a13dba25438a6fc5a41290606418481e08ea2ecc0a9a397062af74fb
          createdAt: "2025-05-14T12:59:00Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <4.2.16'
          operatorframework.io/suggested-namespace: ibm-licensing-scanner
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMLicenseServiceScanner is the Schema for the ibmlicenseservicescanners
              API
            displayName: IBMLicense Service Scanner
            kind: IBMLicenseServiceScanner
            name: ibmlicenseservicescanners.operator.ibm.com
            version: v1
        description: IBM License Service Scanner
        displayName: IBM License Service Scanner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - licensing
        - usage
        - IBMLicenseServiceScanner
        - IBM
        - Cloud
        - License
        - License Service
        - Service
        - Scanner
        - License Service Scanner
        - License Scanner
        - Service Scanner
        links:
        - name: Ibm License Service Scanner Operator
          url: https://ibm-license-service-scanner-operator.domain
        maintainers:
        - email: talk2sam@us.ibm.com
          name: IBM
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-bundle@sha256:b6cb9e5976b749dd7818ac19d1cfd5e24673f3c052a41d6e2fe1d2d7094694b4
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-catalog@sha256:c016904e8c5cfcf8270241ad5e28b52f528859984ed894cdd5e1283fa90cd6e8
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:77c9db26a13dba25438a6fc5a41290606418481e08ea2ecc0a9a397062af74fb
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:f3503f72a90ff8abbadaef3e78a16db6e1fbff4c90b475c33ead17bf1e69990c
        version: 4.2.16
      entries:
      - name: ibm-license-service-scanner-operator.v4.2.16
        version: 4.2.16
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-license-service-scanner-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/component: operator
      app.kubernetes.io/managed-by: kustomize
      app.kubernetes.io/name: zabbix-operator-certified
      app.kubernetes.io/part-of: zabbix-operator-certified
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zabbix LLC
      provider-url: ""
    name: zabbix-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zabbix-operator-certified.v6.0.40
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent",
                "metadata": {
                  "labels": {
                    "app": "agent",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "log_remote_commands": true,
                  "max_lines_per_second": 20,
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent2",
                "metadata": {
                  "labels": {
                    "app": "agent2",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent2"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "enable_persistent_buffer": false,
                  "enable_status_port": false,
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "persistent_buffer_claim": "",
                  "persistent_buffer_period": "1h",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixFull",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-full"
                },
                "spec": {
                  "builtin_report_service": true,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysql_volumeclaim": "mysql-full-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxyMysql",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-mysql"
                },
                "spec": {
                  "db_server_port": 3306,
                  "enable_snmp_traps": false,
                  "internal_db": true,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix_proxy",
                  "proxy": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "zabbix_mysql_volumeclaim": "mysql-proxy-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxySqlite",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-sqlite"
                },
                "spec": {
                  "db_data_volumeclaim": "zabbix-proxy-sqlite",
                  "enable_snmp_traps": false,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "proxy": {
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "node_name_as_db_name": true,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "proxy_size": 1,
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  }
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixServer",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-server"
                },
                "spec": {
                  "builtin_report_service": true,
                  "db_server_host": "mysql-server",
                  "db_server_port": 3306,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix",
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              }
            ]
          ansible.operator-sdk/reconcile-period: 5s
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          certified: "true"
          containerImage: registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
          createdAt: "2025-05-26T07:41:22Z"
          description: Zabbix operator with multiple deployment variants and different
            components
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <6.0.40
          operatorframework.io/suggested-namespace: zabbix
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: ""
          support: Zabbix
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Zabbix agent 2 is a new generation of Zabbix agent and may
              be used in place of Zabbix agent
            displayName: Zabbix agent 2
            kind: ZabbixAgent2
            name: zabbixagent2s.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix agent is deployed on a monitoring nodes to actively
              monitor local resources and applications
            displayName: Zabbix agent
            kind: ZabbixAgent
            name: zabbixagents.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database server, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Full
            kind: ZabbixFull
            name: zabbixfulls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with MySQL database server
            displayName: Zabbix proxy (MySQL)
            kind: ZabbixProxyMysql
            name: zabbixproxymysqls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with SQLite database
            displayName: Zabbix proxy (SQLite3)
            kind: ZabbixProxySqlite
            name: zabbixproxysqlites.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database support, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Server
            kind: ZabbixServer
            name: zabbixservers.kubernetes.zabbix.com
            version: v1alpha1
        description: "## About this Operator\n\nZabbix helps you to real-time monitoring
          of millions of metrics collected from tens of thousands of servers, virtual
          machines and network devices.\nThe Zabbix Operator allows users to easily
          deploy, manage, and maintain Zabbix deployments on OpenShift. By installing
          this integration you will be able to deploy Zabbix server / proxies and
          other components with a single command.\n## Supported Features\n* **Zabbix
          Server** - Simple Zabbix installation with included Zabbix server, Zabbix
          web-interface and Zabbix Java Gateway with MySQL database support. The feature
          does not provide MySQL service and requires an external MySQL database.
          \n* **Zabbix Server (Full)** - Zabbix installation with included Zabbix
          server, Zabbix web-interface, Zabbix Java Gateway and MySQL server instance.\n*
          **Zabbix proxy (SQLite3)** - Very simple way to gain power of Zabbix proxy.
          The feature has  SQLite3 support for Zabbix proxies and allow to specify
          amount of proxies. \n* **Zabbix proxy (MySQL)** - Another option of Zabbix
          proxy. The option support and deliver MySQL database.\n* **Zabbix agent**
          - Zabbix agent can be deployed on each available node for stability and
          performance monitoring on remote nodes. It allows to gather metrics with
          full automation!* \n## Prerequisites\nAll deployment options are require
          additional information during deployment. Please, check the following instructions
          and provide required configuration:\n* **Zabbix Server** - MySQL database
          host information and MySQL database credentials in specially formatted *Secret*.
          Additionally it is possible to specify SSL certificates for HTTPS support
          in *Secret*.\n* **Zabbix Server (Full)** - MySQL database credentials in
          specially formatted *Secret*. MySQL database volume name information. Additionally
          it is possible to specify SSL certificates for HTTPS support in *Secret*.\n*
          **Zabbix proxy (SQLite3)** - Zabbix server host information only.\n* **Zabbix
          proxy (MySQL)** - MySQL database credentials in specially formatted *Secrets*
          and Zabbix server host information.\n* **Zabbix agent** - Zabbix server
          host information only for outgoing and incoming connections. Zabbix agent
          uses \"privileged\" mode to monitor node resources! For example, running
          processes.* \n### MySQL credentials\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-full-secrets\ndata:\n  mysql_root_password: emFiYml4X3N1cGVyX3Jvb3Q=
          [1]\n  mysql_zabbix_password: emFiYml4X3VzZXJfcGFzc3dvcmQ= [2]\n  mysql_zabbix_username:
          emFiYml4 [3]\ntype: Opaque\n```\nUsing MySQL root password (*mysqlrootpassword*
          [1]) Zabbix server / proxy will try to create MySQL database schema with
          grant permissions to *mysqlzabbixusername* [2] and *mysqlzabbixpassword*
          [3].\n### SSL certificates for HTTPS\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-web-sslsecret\ndata:\n  ssl.crt: >-\n   < ssl.crt data>\n
          \ ssl.key: >-\n\t < ssl.key data >\n  dhparam.pem: >-\n   <  dhparam.pem
          data >\n```\nFiles *ssl.crt*, *ssl.key* and *dhparam.perm* are required
          for Zabbix web-interface for SSL support.\n"
        displayName: Zabbix Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - zabbix
        - appliance
        - agent
        - metrics
        links:
        - name: Zabbix
          url: https://www.zabbix.com
        - name: Zabbix Official Documentation
          url: https://www.zabbix.com/documentation/6.0/en/manual/quickstart
        - name: Zabbix Operator Installation
          url: https://www.zabbix.com/documentation/6.0/en/manual/installation/containers/openshift
        - name: Downloads
          url: https://www.zabbix.com/download
        maintainers:
        - email: alexey.pustovalov@zabbix.com
          name: Alexey Pustovalov
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Zabbix LLC
        relatedImages:
        - registry.connect.redhat.com/zabbix/zabbix-agent2-60@sha256:424003d52f3cb4360adf2b3619d8b9f8de9196c630ee7e4ceb5d177c1c5a36e4
        - registry.connect.redhat.com/zabbix/zabbix-snmptraps-60@sha256:e61e6b1c551dec8488cce0998f94d83d6d10b5d89a465d7119b088313032cb81
        - registry.connect.redhat.com/zabbix/zabbix-server-mysql-60@sha256:637484faec6e9a55049c13dbd1f100c03d757c3a21749388c0605040f02d05ce
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
        - quay.io/zabbix/zabbix-operator@sha256:43d26488a19967bdb2a8aafde5f43820be95011811da9cc5ec623c9029801675
        - registry.connect.redhat.com/zabbix/zabbix-web-mysql-60@sha256:611d2670e1bf95b81226976c0a1f64434b2a7766f196dadbd646a7d8319f4cb1
        - registry.connect.redhat.com/zabbix/zabbix-java-gateway-60@sha256:a4feb725c12c4ac90eb9244c6fafdb6708679f2e0e7f6aa683c765127306412e
        - registry.connect.redhat.com/zabbix/zabbix-proxy-mysql-60@sha256:c2ae74d97c083ff348dc9c55932a33da4779cd9efdd9d7c5a79acfffae5b9ed0
        - registry.connect.redhat.com/zabbix/zabbix-proxy-sqlite-60@sha256:62e62ebc8c1bc13fb610f7a83412e65d852f74b3c2c212baa75ea27fc7fa03eb
        - registry.connect.redhat.com/zabbix/zabbix-web-service-60@sha256:4a333094e110713470a897b6406c5e340e962caf32a429dddf19865041fcc448
        - registry.redhat.io/rhel9/mysql-80@sha256:c4b5246d0c0d089a766ea07768c22d9d6be8049ac1379bd345db99390ffe2a36
        - registry.connect.redhat.com/zabbix/zabbix-agent-60@sha256:2895b5e45785aaf528d4a6f78c8af0a0d494dbe9362403fdd470673d911d5004
        version: 6.0.40
      entries:
      - name: zabbix-operator-certified.v6.0.40
        version: 6.0.40
      - name: zabbix-operator-certified.v6.0.38
        version: 6.0.38
      - name: zabbix-operator-certified.v6.0.36
        version: 6.0.36
      - name: zabbix-operator-certified.v6.0.33
        version: 6.0.33
      - name: zabbix-operator-certified.v6.0.29
        version: 6.0.29
      - name: zabbix-operator-certified.v6.0.27
        version: 6.0.27
      - name: zabbix-operator-certified.v6.0.26
        version: 6.0.26
      - name: zabbix-operator-certified.v6.0.17
        version: 6.0.17
      - name: zabbix-operator-certified.v6.0.14
        version: 6.0.14
      - name: zabbix-operator-certified.v6.0.13
        version: 6.0.13
      name: lts
    defaultChannel: lts
    packageName: zabbix-operator-certified
    provider:
      name: Zabbix LLC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/multiarch-tuning-operator
    name: multiarch-tuning-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multiarch-tuning-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multiarch.openshift.io/v1beta1",
                "kind": "ClusterPodPlacementConfig",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logVerbosity": "Normal",
                  "namespaceSelector": {
                    "matchExpressions": [
                      {
                        "key": "multiarch.openshift.io/exclude-pod-placement",
                        "operator": "DoesNotExist"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Other
          console.openshift.io/disable-operand-delete: "false"
          containerImage: registry.redhat.io/multiarch-tuning/multiarch-tuning-rhel9-operator@sha256:1a1d3f5258d4b91f1a0380b0fe2c809823e323d9ff74ae4171468168d255e7e0
          createdAt: "2025-03-05T12:27:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operator.openshift.io/uninstall-message: You must remove all Operands before
            uninstalling the operator.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-multiarch-tuning-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/multiarch-tuning-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPodPlacementConfig defines the configuration for the
              architecture aware pod placement operand. Users can only deploy a single
              object named "cluster". Creating the object enables the operand.
            displayName: Cluster Pod Placement Config
            kind: ClusterPodPlacementConfig
            name: clusterpodplacementconfigs.multiarch.openshift.io
            version: v1alpha1
          - description: ClusterPodPlacementConfig defines the configuration for the
              architecture aware pod placement operand. Users can only deploy a single
              object named "cluster". Creating the object enables the operand.
            displayName: Cluster Pod Placement Config
            kind: ClusterPodPlacementConfig
            name: clusterpodplacementconfigs.multiarch.openshift.io
            version: v1beta1
        description: |
          The Multiarch Tuning Operator optimizes workload management within multi-architecture clusters and in
          single-architecture clusters transitioning to multi-architecture environments.

          By default, the scheduler does not consider the architecture of a pod's container images
          when determining the placement of new pods onto nodes.

          The Multiarch Tuning Operator introduces the ClusterPodPlacementConfig custom resource to
          enable architecture-aware scheduling, ensuring that pods are assigned to nodes with
          a corresponding architecture for their container images.

          To enable this functionality, you must create a ClusterPodPlacementConfig resource.

          When the ClusterPodPlacementConfig object is created, the Multiarch Tuning Operator deploys the necessary operands
          to support architecture-aware workload scheduling.
        displayName: Multiarch Tuning Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multiarch-tuning-operator
        - pod-placement
        links:
        - name: Multiarch Tuning Operator Documentation
          url: https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/postinstallation_configuration/configuring-multi-architecture-compute-machines-on-an-openshift-cluster#multiarch-tuning-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/multiarch-tuning-operator
        relatedImages:
        - registry.redhat.io/multiarch-tuning/multiarch-tuning-rhel9-operator@sha256:1a1d3f5258d4b91f1a0380b0fe2c809823e323d9ff74ae4171468168d255e7e0
        version: 1.1.0
      entries:
      - name: multiarch-tuning-operator.v1.1.0
        version: 1.1.0
      - name: multiarch-tuning-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: multiarch-tuning-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/multiarch-tuning-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: infra-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infra-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instanceha.openstack.org/v1beta1",
                "kind": "InstanceHa",
                "metadata": {
                  "name": "instanceha"
                },
                "spec": {
                  "caBundleSecretName": "combined-ca-bundle"
                }
              },
              {
                "apiVersion": "memcached.openstack.org/v1beta1",
                "kind": "Memcached",
                "metadata": {
                  "name": "memcached"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "network.openstack.org/v1beta1",
                "kind": "DNSData",
                "metadata": {
                  "name": "allhosts"
                },
                "spec": {
                  "hosts": [
                    {
                      "hostnames": [
                        "edpm-compute-0.ctlplane"
                      ],
                      "ip": "192.168.122.100"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-0.internalapi"
                      ],
                      "ip": "172.17.0.100"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-0.storage"
                      ],
                      "ip": "172.18.0.100"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-0.tenant"
                      ],
                      "ip": "172.19.0.100"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-1.ctlplane"
                      ],
                      "ip": "192.168.122.101"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-1.internalapi"
                      ],
                      "ip": "172.17.0.101"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-1.storage"
                      ],
                      "ip": "172.18.0.101"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-1.tenant"
                      ],
                      "ip": "172.19.0.101"
                    },
                    {
                      "hostnames": [
                        "keystone-internal.openstack.svc",
                        "placement-internal.openstack.svc",
                        "cinder-internal.openstack.svc",
                        "glance-internal.openstack.svc",
                        "nova-internal.openstack.svc",
                        "nova-metadata-internal.openstack.svc",
                        "neutron-internal.openstack.svc"
                      ],
                      "ip": "172.17.0.80"
                    },
                    {
                      "hostnames": [
                        "rabbitmq.openstack.svc"
                      ],
                      "ip": "172.17.0.85"
                    },
                    {
                      "hostnames": [
                        "rabbitmq-cell1.openstack.svc"
                      ],
                      "ip": "172.17.0.86"
                    }
                  ]
                }
              },
              {
                "apiVersion": "network.openstack.org/v1beta1",
                "kind": "DNSMasq",
                "metadata": {
                  "name": "dnsmasq"
                },
                "spec": {
                  "debug": {
                    "service": false
                  },
                  "options": [
                    {
                      "key": "server",
                      "values": [
                        "192.168.122.1"
                      ]
                    },
                    {
                      "key": "local",
                      "values": [
                        "/example.com/"
                      ]
                    }
                  ],
                  "override": {
                    "service": {
                      "metadata": {
                        "annotations": {
                          "metallb.universe.tf/address-pool": "ctlplane",
                          "metallb.universe.tf/allow-shared-ip": "ctlplane",
                          "metallb.universe.tf/loadBalancerIPs": "192.168.122.80"
                        }
                      },
                      "spec": {
                        "type": "LoadBalancer"
                      }
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "network.openstack.org/v1beta1",
                "kind": "IPSet",
                "metadata": {
                  "name": "edpm-compute-0"
                },
                "spec": {
                  "networks": [
                    {
                      "defaultRoute": true,
                      "name": "CtlPlane",
                      "subnetName": "subnet1"
                    },
                    {
                      "name": "InternalApi",
                      "subnetName": "subnet1"
                    },
                    {
                      "name": "Storage",
                      "subnetName": "subnet1"
                    },
                    {
                      "name": "Tenant",
                      "subnetName": "subnet1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "network.openstack.org/v1beta1",
                "kind": "NetConfig",
                "metadata": {
                  "name": "netconfig"
                },
                "spec": {
                  "networks": [
                    {
                      "dnsDomain": "ctlplane.example.com",
                      "name": "ctlplane",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "192.168.122.120",
                              "start": "192.168.122.100"
                            },
                            {
                              "end": "192.168.122.200",
                              "start": "192.168.122.150"
                            }
                          ],
                          "cidr": "192.168.122.0/24",
                          "gateway": "192.168.122.1",
                          "name": "subnet1"
                        }
                      ]
                    },
                    {
                      "dnsDomain": "internalapi.example.com",
                      "name": "internalapi",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "172.17.0.250",
                              "start": "172.17.0.100"
                            }
                          ],
                          "cidr": "172.17.0.0/24",
                          "name": "subnet1",
                          "vlan": 20
                        }
                      ]
                    },
                    {
                      "dnsDomain": "external.example.com",
                      "name": "external",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "10.0.0.250",
                              "start": "10.0.0.100"
                            }
                          ],
                          "cidr": "10.0.0.0/24",
                          "gateway": "10.0.0.1",
                          "name": "subnet1"
                        }
                      ]
                    },
                    {
                      "dnsDomain": "storage.example.com",
                      "name": "storage",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "172.18.0.250",
                              "start": "172.18.0.100"
                            }
                          ],
                          "cidr": "172.18.0.0/24",
                          "name": "subnet1",
                          "vlan": 21
                        }
                      ]
                    },
                    {
                      "dnsDomain": "storagemgmt.example.com",
                      "name": "storagemgmt",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "172.20.0.250",
                              "start": "172.20.0.100"
                            }
                          ],
                          "cidr": "172.20.0.0/24",
                          "name": "subnet1",
                          "vlan": 23
                        }
                      ]
                    },
                    {
                      "dnsDomain": "tenant.example.com",
                      "name": "tenant",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "172.19.0.250",
                              "start": "172.19.0.100"
                            }
                          ],
                          "cidr": "172.19.0.0/24",
                          "name": "subnet1",
                          "vlan": 22
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "network.openstack.org/v1beta1",
                "kind": "Reservation",
                "metadata": {
                  "name": "edpm-compute-0"
                },
                "spec": {
                  "ipSetRef": {
                    "name": "edpm-compute-0",
                    "namespace": "openstack",
                    "uid": "56753ffb-e486-43ab-a79a-f85f27b2ac14"
                  },
                  "reservation": {
                    "CtlPlane": {
                      "address": "192.168.122.100",
                      "network": "CtlPlane",
                      "subnet": "subnet1"
                    },
                    "InternalApi": {
                      "address": "172.17.0.100",
                      "network": "InternalApi",
                      "subnet": "subnet1"
                    },
                    "Storage": {
                      "address": "172.18.0.100",
                      "network": "Storage",
                      "subnet": "subnet1"
                    },
                    "StorageMgmt": {
                      "address": "172.20.0.100",
                      "network": "StorageMgmt",
                      "subnet": "subnet1"
                    },
                    "Tenant": {
                      "address": "172.19.0.100",
                      "network": "Tenant",
                      "subnet": "subnet1"
                    }
                  }
                }
              },
              {
                "apiVersion": "rabbitmq.openstack.org/v1beta1",
                "kind": "TransportURL",
                "metadata": {
                  "name": "transporturl"
                },
                "spec": {
                  "rabbitmqClusterName": "rabbitmq"
                }
              },
              {
                "apiVersion": "redis.openstack.org/v1beta1",
                "kind": "Redis",
                "metadata": {
                  "name": "redis"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:45Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSData is the Schema for the dnsdata API
            displayName: DNSData
            kind: DNSData
            name: dnsdata.network.openstack.org
            version: v1beta1
          - description: DNSMasq is the Schema for the dnsmasqs API
            displayName: DNSMasq
            kind: DNSMasq
            name: dnsmasqs.network.openstack.org
            version: v1beta1
          - description: InstanceHa is the Schema for the instancehas API
            displayName: Instance Ha
            kind: InstanceHa
            name: instancehas.instanceha.openstack.org
            version: v1beta1
          - description: IPSet is the Schema for the ipsets API
            displayName: IPSet
            kind: IPSet
            name: ipsets.network.openstack.org
            version: v1beta1
          - description: Memcached is the Schema for the memcacheds API
            displayName: Memcached
            kind: Memcached
            name: memcacheds.memcached.openstack.org
            version: v1beta1
          - description: NetConfig is the Schema for the netconfigs API
            displayName: Net Config
            kind: NetConfig
            name: netconfigs.network.openstack.org
            version: v1beta1
          - description: Redis is the Schema for the redises API
            displayName: Redis
            kind: Redis
            name: redises.redis.openstack.org
            version: v1beta1
          - description: Reservation is the Schema for the reservations API
            displayName: Reservation
            kind: Reservation
            name: reservations.network.openstack.org
            version: v1beta1
          - description: TransportURL is the Schema for the transporturls API
            displayName: Transport URL
            kind: TransportURL
            name: transporturls.rabbitmq.openstack.org
            version: v1beta1
        description: OpenStack Infrastructure
        displayName: OpenStack Infra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Infrastructure
        links:
        - name: Infra Operator
          url: https://github.com/openstack-k8s-operators/infra-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/infra-rhel9-operator@sha256:177077029ad8baa959e41a785cd29ed63745e8f39be618aa97509e6867fbff7c
        - registry.redhat.io/rhoso/openstack-memcached-rhel9@sha256:962a314226009a8ca5c95fb66a75b15176506e02110847d0f42aeed6dd26a751
        - registry.redhat.io/rhoso/openstack-redis-rhel9@sha256:b83d23df9f2b367e604f404d97574fd6db9e0ed530eda23e9e0bbe01292c8db8
        - registry.redhat.io/rhoso/openstack-neutron-server-rhel9@sha256:806e5e009a77a29c82699487e539669e1ec2136514a0332d222b5cb3fed6f452
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: infra-operator.v1.0.6
        version: 1.0.6
      - name: infra-operator.v1.0.5
        version: 1.0.5
      - name: infra-operator.v1.0.4
        version: 1.0.4
      - name: infra-operator.v1.0.3
        version: 1.0.3
      - name: infra-operator.v1.0.2
        version: 1.0.2
      - name: infra-operator.v1.0.1
        version: 1.0.1
      - name: infra-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: infra-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: maximilianopizarro
      provider-url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
    name: jhipster-online-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jhipster-online-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maximilianopizarro.github.io/v1alpha1",
                "kind": "JhipsterOnline",
                "metadata": {
                  "name": "jhipsteronline-sample"
                },
                "spec": {
                  "affinity": {},
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "env": {
                    "APPLICATION_GITHUB_CLIENT-ID": "CLIENT-ID",
                    "APPLICATION_GITHUB_CLIENT-SECRET": "CLIENT-SECRET",
                    "APPLICATION_GITHUB_HOST": "https://github.com",
                    "APPLICATION_JHIPSTER-CMD_CMD": "jhipster-quarkus",
                    "JAVA_APP_JAR": "jhonline-2.33.0.war",
                    "SPRING_DATASOURCE_PASSWORD": "jhipster",
                    "SPRING_DATASOURCE_URL": "jdbc:mariadb://mariadb:3306/jhipsteronline",
                    "SPRING_DATASOURCE_USERNAME": "jhipster"
                  },
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "quay.io/maximilianopizarro/jhipster-online",
                    "tag": "quarkus"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "livenessProbe": {
                    "httpGet": {
                      "path": "/jdl-studio/",
                      "port": 8080
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "readinessProbe": {
                    "httpGet": {
                      "path": "/jdl-studio/",
                      "port": 8080
                    }
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "securityContext": {},
                  "service": {
                    "port": 8080,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automount": false,
                    "create": false,
                    "name": ""
                  },
                  "tolerations": [],
                  "volumeMounts": [],
                  "volumes": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/maximilianopizarro/jhipster-online-operator:v0.1.0
          createdAt: "2025-05-30T22:42:19Z"
          description: JHipster Online provides an intuitive web interface for generating
            JHipster projects, allowing you to define your application architecture
            using a JDL (JHipster Domain Language) model and generate the code directly
            into a GitHub repository.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/maximilianoPizarro/jhipster-online-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: JhipsterOnline
            name: jhipsteronlines.maximilianopizarro.github.io
            version: v1alpha1
        description: |
          JHipster Online Operator: This repository hosts the JHipster Online Operator, designed to seamlessly deploy JHipster Online on Red Hat OpenShift and Kubernetes environments.

          JHipster Online provides an intuitive web interface for generating JHipster projects, allowing you to define your application architecture using a JDL (JHipster Domain Language) model and generate the code directly into a GitHub repository.

          Overview: The JHipster Online Operator manages the deployment and lifecycle of the JHipster Online application. It leverages the power of Kubernetes Operators to provide a native Kubernetes experience for JHipster Online users.

          This artifact includes:
          - JHipster 8.8.0. for generate Spring Boot 3.4.1 projects.
          - generator-jhipster-quarkus 3.4.0 for generate Quarkus 3.11.1 projects.
          - JDL Studio for add JDL models files by PR on your repo.

          Usage: Once the operator is running, you can create instances of JHipster Online by defining `JhipsterOnline` custom resources. An example `JhipsterOnline` custom resource is provided in the `alm-examples`. You would apply this (or a similar) YAML file to your cluster to create an instance of JHipster Online. Feel free to customize the `JhipsterOnline` Custom Resource Definition (CRD) instance to suit your needs, such as configuring GitHub integration, database connections, and resource allocations.

          Configuration: The `JhipsterOnline` custom resource supports various configuration options including `replicaCount`, `image`, `imagePullSecrets`, `serviceAccount`, `podAnnotations`, `podLabels`, `podSecurityContext`, `securityContext`, `service` type and port, `ingress` settings (enabled, `className`, `annotations`, `hosts`, `tls`), OpenShift `Route` (enabled), `autoscaling` (enabled, `minReplicas`, `maxReplicas`, `targetCPUUtilizationPercentage`), `resource` requests and limits, `nodeSelector`, `tolerations`, `affinity`, `liveness` and `readiness` probes, `env` variables, `volumes`, and `volumeMounts`. Refer to the detailed `README.md` for a full table of parameters.
        displayName: JHipster Online Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - java
        - spring-boot
        - jhipster
        - quarkus
        links:
        - name: Jhipster Online Operator
          url: https://maximilianopizarro.github.io/jhipster-online-operator/
        maintainers:
        - email: maximiliano.pizarro.5@gmail.com
          name: Maximiliano Pizarro
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: maximilianopizarro
          url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/maximilianopizarro/jhipster-online-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: jhipster-online-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: jhipster-online-operator
    provider:
      name: maximilianopizarro
      url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Skupper Project
      provider-url: https://skupper.io
    name: skupper-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: alpha
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1
    - currentCSV: skupper-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:e5f8b2ddf378cf3e40f4c7b7ec8482eb3321947cd46481c868d727932c449d78
          createdAt: "2024-02-26T18:22:07Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/config-sync@sha256:659fe28d1df7c99bba3bb854b1e9991c715ae4de99d91c88d5db5e4f9ef4c852
        - quay.io/skupper/flow-collector@sha256:3870c5bc89e623b7cebe21856bfefc22de76313705efa0ec551a5963789fff3b
        - quay.io/skupper/site-controller@sha256:e5f8b2ddf378cf3e40f4c7b7ec8482eb3321947cd46481c868d727932c449d78
        - quay.io/prometheus/prometheus@sha256:d2ab0a27783fd4ad96a8853e2847b99a0be0043687b8a5d1ebfb2dd3fa4fd1b8
        - quay.io/openshift/origin-oauth-proxy@sha256:a7dff785d821acc4b972d94964947713fade45f007dedb5deee17a2869bf5d24
        - quay.io/skupper/skupper-router@sha256:bf8d5955ce4e406275a4934278efe51c1af9a1a269a70e8fe642b7d94d80cb81
        - quay.io/skupper/service-controller@sha256:345edf2d9d72b5a97bcce0e8cea8cda82d120f7bf83756d73f9ac114c2fcced8
        version: 1.6.0
      entries:
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.6
    - currentCSV: skupper-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:4809db7263d77a95797061798e910f1bdbb5d5fe2db2bb3c743b97f5d0c7b3af
          createdAt: "2024-06-12T14:36:59Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/prometheus/prometheus@sha256:d2ab0a27783fd4ad96a8853e2847b99a0be0043687b8a5d1ebfb2dd3fa4fd1b8
        - quay.io/openshift/origin-oauth-proxy@sha256:a7dff785d821acc4b972d94964947713fade45f007dedb5deee17a2869bf5d24
        - quay.io/skupper/skupper-router@sha256:5eb258a0927e7706e7c035f1def2eb5211c8c33dd371fafa7210517fd894f68b
        - quay.io/skupper/service-controller@sha256:624e88e695297df28e485db7e802945ad1978ac6b0c9e9edca8b5e6013af8991
        - quay.io/skupper/site-controller@sha256:4809db7263d77a95797061798e910f1bdbb5d5fe2db2bb3c743b97f5d0c7b3af
        - quay.io/skupper/config-sync@sha256:a884b26a32179af5a2bfa85118bf873852eec6e2930e92da778b57506c0a0880
        - quay.io/skupper/flow-collector@sha256:bf646b0cb75e1c29bbe0261aa9943e06d33118f5e6a5ccb7503c60c974167692
        version: 1.7.1
      entries:
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.7
    - currentCSV: skupper-operator.v1.8.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:ffe3ea5067d350b64b69120bdf6d9e3ee9a30e8a2559b77170a10d66635623b8
          createdAt: "2025-03-27T20:11:53Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/service-controller@sha256:a45de211b8f8a8294a3588932a0a72766cf0d5ce64b01f4f744832c8d09aaa46
        - quay.io/skupper/config-sync@sha256:9e96d6a585f037f60ca3d89b5efee27ec92444d353e8b289ca35b94aeed1e309
        - quay.io/skupper/flow-collector@sha256:252244ab6aaf6b80118bfc06b690d4f5e8d7829b999fca60d194a84452fee7f6
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/openshift/origin-oauth-proxy@sha256:c15baf57dbd17dcc517cd2fa2299462d62124f689d6f5a48a3c9fd154fc77ccf
        - quay.io/skupper/skupper-router@sha256:7633b143f1844ee559322e4642ca80e8248a08711c8e814d489c26c39ecf9bd0
        - quay.io/skupper/site-controller@sha256:ffe3ea5067d350b64b69120bdf6d9e3ee9a30e8a2559b77170a10d66635623b8
        version: 1.8.4
      entries:
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.8
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.9
    defaultChannel: stable
    packageName: skupper-operator
    provider:
      name: Skupper Project
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Harness Inc.
      provider-url: https://harness.io
    name: harness-gitops-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: harness-gitops-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gitops.harness.io/v1alpha1",
                "kind": "GitopsAgent",
                "metadata": {
                  "annotations": {
                    "containerImage": "harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde",
                    "createdAt": "2025-07-31T12:11:18Z"
                  },
                  "name": "gitops-agent-sample"
                },
                "spec": {
                  "identity": {
                    "accountIdentifier": "accountIdentifier",
                    "agentIdentifier": "agentIdentifier",
                    "agentName": "agentName"
                  },
                  "replicas": 1,
                  "secrets": {
                    "agentSecret": "agentSecret"
                  },
                  "target": {
                    "http": "https://app.harness.io/gitops/",
                    "protocol": "HTTP1"
                  },
                  "upgrade": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde
          createdAt: "2025-07-31T06:41:18Z"
          description: Harness operator to manage and provision GitOps Agent
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Harness Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Schema for the Gitops Agents API
            displayName: Gitops Agent
            kind: GitopsAgent
            name: gitopsagents.gitops.harness.io
            version: v1alpha1
        description: |-
          ## Overview
          Harness Gitops Operator manages the full lifecycle for [Gitops Agent](https://developer.harness.io/docs/continuous-delivery/gitops/connect-and-manage/install-a-harness-git-ops-agent/)

          The operator aims to provide the following:

            * Easy configuration and installation of Harness GitOps Agent.
            * Provide seamless upgrades for Harness GitOps Agent.
            * Monitoring and autorecovery for GitOps Agent.

          ## Usage
          Deploy gitops-agent in namespace where operator is installed with basic setup.

          ### Prerequisites:
            * Go to your Harness account and create a new gitops agent
            * Copy account, org and project identifiers
            * Retrieve the agent secret key

          ```
          apiVersion: gitops.harness.io/v1alpha1
          kind: GitopsAgent
          metadata:
            name: gitops-agent-sample
          spec:
            identity:
              accountIdentifier: accountIdentifier
              agentIdentifier: agentIdentifier
              agentName: agentName
            secrets:
              agentSecret: agentSecret
            target:
              http: https://app.harness.io/gitops/
              protocol: HTTP1
            replicas: 1
            upgrade:
              enabled: true
          ```
          **Note:** Gitops Agent require an active ArgoCD instance running in the same namespace where agent is installed.

          See [Harness Documentation](https://developer.harness.io/) for more information.
        displayName: Harness GitOps Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - harness
        - agent
        - cd
        - argocd
        - kubernetes
        links:
        - name: Harness Gitops Operator
          url: https://harness-gitops-operator.domain
        maintainers:
        - email: alexandru.bosii@harness.io
          name: Alexandru Bosii
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Harness Inc.
          url: https://harness.io
        relatedImages:
        - harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde
        - docker.io/harness/gitops-agent@sha256:c9f57338be294d32a73a4cbecf90dfff50f5d36a699f3ec32d962dcb64d8d085
        version: 0.3.0
      entries:
      - name: harness-gitops-operator.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: harness-gitops-operator
    provider:
      name: Harness Inc.
      url: https://harness.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataTrucker.IO
      provider-url: https://www.datatrucker.io/
    name: datatrucker-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datatrucker-operator.v2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datatrucker.datatrucker.io/v1",
                "kind": "DatatruckerConfig",
                "metadata": {
                  "name": "datatruckerconfig-sample"
                },
                "spec": {
                  "API": {
                    "crypto": {
                      "CredentailEncryption": {
                        "Algorithm": "aes-256-cbc",
                        "IV": "eThWmZq4t7w!z$C\u0026",
                        "Key": "JaNdRgUkXp2s5u8x/A?D(G+KbPeShVmY"
                      },
                      "JWT": {
                        "authorization_keys": {
                          "private_key": "privateJWT.key",
                          "public_key": "publicJWT.key"
                        },
                        "signOptions": {
                          "algorithm": "RS256",
                          "expiresIn": "60m",
                          "issuer": "datatrucker"
                        },
                        "verify": {
                          "issuer": "datatrucker"
                        }
                      },
                      "keycloak": {
                        "client_id": "nodejs",
                        "client_secret": "7c5542ae-7264-420d-bf5f-766f4ab39ad3",
                        "jwt_publickey_verifier": "keycloak.key",
                        "realm": "TestRealm",
                        "url": "http://localhost:30015"
                      },
                      "passwordHash": {
                        "Algorithm": "sha256",
                        "SaltLength": 16
                      }
                    },
                    "cryptokeys": {
                      "keycloakkey": "LS0tLS1CRUdJTiBQVUJMSUMgS0VZLS0tLS0KTUlJQklqQU5CZ2txaGtpRzl3MEJBUUVGQUFPQ0FROEFNSUlCQ2dLQ0FRRUF4ZnhGY1gxb3RmTUN6SWlkZ09tMndZakZWZXZHN1BQT1h5dTNrTEFuTWtTVkUxUXl4TzJ4SFFvSGcydDFVcUU4SVNXeTN3ZkZTbERzdU1VK0J1VVkvaWI5L04vSlpBV3k5MGRET0YxREdIeGVwM21HMVBTUllMQWFGQ0I3NTVsbWZjL1dna3VzZFdLdmxOTmZuNkd6S1J6OXBmV3RNUnhWWkFrejhmNlVUeG0zWE02Q0tOMUludThubEcxU3ZzaFc5cmt3Lyt6UmVvOFVCV2pPcTh5RjdPV0tnS0ZZZlhENFREa090OWFxN3RaaSsxaExYOE5iV2xxRnJZYmhJcFZINXlLSko3ZTA5cUVIRkVEelg5QXJzVVNHamNjbFFFYTl1dmJxU2d4Rzc3RUZVbk82aDZxUkNnU3lNNnVYdk9qcWlIMHp4RTNmRHB4R2NhY2pvTW1EMlFJREFRQUIKLS0tLS1FTkQgUFVCTElDIEtFWS0tLS0t",
                      "privateJWTkey": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQ0KTUlJRW9RSUJBQUtDQVFCVWd1ZDN1OFlXcUJCWWtYZ3M2K25pMS84bnM5YjRGNkZudzA4RDRKdGRlalg5L1ZaMg0KWFdOamJ1ekMxUkRHMnVseFM4QjZIdTZYeEhYWU9XU2ZzOEY0UHZ2eC9TUWdmRWtWUlNDL1pQN3pzM1ZETllaMA0Kc1JScDNSZzUza29qUG80T3o2azFaeGprbFRFZEpMWUtqYlZud3U1M3A3MTFsVTJkV3B4R3VEeGYxVTJuUWgrWA0KaUhKZkhyUmxTZVA1aVhYQkpxd21VTDFkK25lTU5BOXlGakR0MnNZMG5iSDRySDVxbk1VRk4xMEpsZ3V4NFdiUw0KaGFUNklNN3h1dUVHbDRDckVENmdTejBNNGF1UVk4ekZxQWxJNlFWUTlML28zQnBCRmxYUUMwWmowb3lwQVFKLw0KQ1VLaU9FcEhLRHgram02RU9KOE1aejlMQ01yTmdLME5KSWlaQWdNQkFBRUNnZ0VBRTBMMnNLSjZob1Z0bkZJag0KZXRHV01CNmN5QmJFSXkxM215ZTNwYUdXcDVxSUhjaVU0VXR0c0tjQ2pYRjBibTVNa0liOGFwa01Oa2tQdHVWMQ0KMnJSa254eHozZmFkT2s0NWRnQ1ljeWI0dU9DcEZGVnB2TVR6YXlGNk9RNUNCWUJuYmxiRU1TaTdVclh5U0xzcQ0KUm9OLzJoTml1S0RobktBWDNZcGVYZnYxdnIyaWIrMGtiRDVvMVZLcnB6SlJ5emlsWHFwQWJwNDQ4RExFZHVGeA0KTnJKc0tWb1dXMVhmS1FvR2VDMXZqNFJudzdhWkNyU1BjMEhvMUdKb29xSXlrNjRXZnhGWjhpVElsWDJMU1gzbQ0KQm1pN1ZXZnVJK0QvUXRtWEVpN05oTE1jUmExVXB4QkdiK1ZPZXlucEVyMEpxM3VZSmlva1VIZzloOGIzSDM1Yg0KK0lJaEFRS0JnUUNtbzUrRUNpOEJUcFRHdlpqQXBtdmxGZ055NkpWd0RvVHJLSGdCcHJHVHpsVGt6Q0tLYzJhTw0KTTMvQmVIVUQ1NVFqcTdUM094eGgzQUk3cG9kT0tCWS84eXgzenlYUExQUmVCMlR0eTNVQjE5YU5CclhNUGQ5RQ0KTG5hbUQ0SmZQeTV0WHpWWEloWVB0UzBiNjdGbjZnZFl4MlNBbGFYYWFjejU4eURLZnlZL3VRS0JnUUNCMUxNYg0KTFZNTU1Pd1U4R2xsVVB1V0Q5VnBsMnlURytCN2Nnd0ZiYTlrWVlVSFQ4TmlvaUdxQzlLZWxqeklKNnlUdyt3aA0KQisvNGNJU1o5Y0JDNHB0VnRlTWJwUkg3UTQwK20xU3M1SmQzU1VFcDBMOW5VbmdVNDhrMzlpSTV5aEgxU3JTRw0KREtBamVka2Rjb3hNVWFwUHB6cVprOVNKdGVHckVsTHY5WmcvNFFLQmdHcExQTmt3MWNMenFSc01uWDlscUxVOA0KVVg3ZXRWUUJMdkFCcEdzSk5ZU2V2TEFsTm5NeUlWemJwTVo0Vzh2ZDBEVGd0bEp0dGpFRjhiYXpCdVBuMWlsOA0KcUFUVWNiMnhZejhIWUczK3RlZ3pFWHhycDFOY3RDbjlxVmxMMDFHR0w4czhTLzRFbENYNDM0bEwwcUlDT2lWNw0KS3JTS3dIeU51RjVIbzhjSFRCS2hBb0dBVjhVRVpENEs0N1g3RTlqSFY3QWN0UUQxUkpJUzI0ekFBSlJFSGpOZA0KbkIzMVFkRWhoSCtaQkkxRGtPUitLTTN4ZEYvQ05UNVpjM29WbXNPeGUwWVl2eVBjYWViNU5nVWY2cEExK1QvOA0KU2E4ak4zS2d3dWZVdUVLWjJKUXo1UjkvYm5WWGZ6dnYwcXZnSkhFNCthdEgyTnR6MlRsQU9PelZwOElXZ1pSbA0KbmVFQ2dZQXpVS0RLS1ZKaXBYdC9TRC8xd2ZzWElCazFxYWpmdENHOFVTcFlGWGp5VzVYMlZ5bTlVTldmVEdJcA0KQlJkVldXamdFbWhyOU5Da1UrVEU4WEVjTXJNRnhWbXBEeTQ4YmxsQWltWFJtOEhNMlhKODQ4TlArTG5GdEtHdA0KTWswT3JuUHdPYitualZGcXdWUllNQ0QzeFN0Z2RnZUppUHFYcWd3SDhjR0M0OGVmZ0E9PQ0KLS0tLS1FTkQgUlNBIFBSSVZBVEUgS0VZLS0tLS0=",
                      "publicJWTkey": "LS0tLS1CRUdJTiBQVUJMSUMgS0VZLS0tLS0NCk1JSUJJVEFOQmdrcWhraUc5dzBCQVFFRkFBT0NBUTRBTUlJQkNRS0NBUUJVZ3VkM3U4WVdxQkJZa1hnczYrbmkNCjEvOG5zOWI0RjZGbncwOEQ0SnRkZWpYOS9WWjJYV05qYnV6QzFSREcydWx4UzhCNkh1Nlh4SFhZT1dTZnM4RjQNClB2dngvU1FnZkVrVlJTQy9aUDd6czNWRE5ZWjBzUlJwM1JnNTNrb2pQbzRPejZrMVp4amtsVEVkSkxZS2piVm4NCnd1NTNwNzExbFUyZFdweEd1RHhmMVUyblFoK1hpSEpmSHJSbFNlUDVpWFhCSnF3bVVMMWQrbmVNTkE5eUZqRHQNCjJzWTBuYkg0ckg1cW5NVUZOMTBKbGd1eDRXYlNoYVQ2SU03eHV1RUdsNENyRUQ2Z1N6ME00YXVRWTh6RnFBbEkNCjZRVlE5TC9vM0JwQkZsWFFDMFpqMG95cEFRSi9DVUtpT0VwSEtEeCtqbTZFT0o4TVp6OUxDTXJOZ0swTkpJaVoNCkFnTUJBQUU9DQotLS0tLUVORCBQVUJMSUMgS0VZLS0tLS0=",
                      "truckerapicrt": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSUQxVENDQXIyZ0F3SUJBZ0lVUFVzZ2NYVXZRYWtuQldjeUFDeEw2dmt3b204d0RRWUpLb1pJaHZjTkFRRUwKQlFBd2RqRUxNQWtHQTFVRUJoTUNRMEV4Q3pBSkJnTlZCQWdNQWs5T01ROHdEUVlEVlFRSERBWk5hV3gwYjI0eApGekFWQmdOVkJBb01Ea1JoZEdGVWNuVmphMlZ5TGtsUE1SUXdFZ1lEVlFRTERBdEZibWRwYm1WbGNtbHVaekVhCk1CZ0dBMVVFQXd3UlJHRjBZVlJ5ZFdOclpYSlNiMjkwUTBFd0hoY05NakF4TURFeU1UWXlOalV5V2hjTk1qTXcKTVRFMU1UWXlOalV5V2pCM01Rc3dDUVlEVlFRR0V3SkRRVEVMTUFrR0ExVUVDQXdDVDA0eER6QU5CZ05WQkFjTQpCazFwYkhSdmJqRVhNQlVHQTFVRUNnd09SR0YwWVZSeWRXTnJaWEl1U1U4eEZEQVNCZ05WQkFzTUMwVnVaMmx1ClpXVnlhVzVuTVJzd0dRWURWUVFEREJKa1lYUmhkSEoxWTJ0bGNpMWhjR2t1YVc4d2dnRWlNQTBHQ1NxR1NJYjMKRFFFQkFRVUFBNElCRHdBd2dnRUtBb0lCQVFEakM4ZmVCVEg4L3NnejFQcTJQYU9QQjc5YWtyeUZvWVpGV292bQpWOXNvMmJNRHp4NVBZSFJydVFGYmJQRzlVVlc2QWZpSHdEb1p2Z3NTVXpyS3ZyR1RlYjluNUhSVlVOMTl0NExTCjdtUFppNzFRbGE4bVR3QmdvTnE4TFppcnZUdzV0S2I4K3NmcFJxQUlUQkVsRmxVUEZISzllcDl2eFdvaHhwdGgKS1dPVkwvV2ZtWldhWXlRdjlLNDBkWUJpRWpGcTFKZVB4MDlnUFdHM2pOT1VYTTZQUVlMaHJDUXFFcy9VME9GWQpwVFRHUmNXZEx4TFVFZVNkcHJKK1ZEc3lEK1lVeUFRR2phYk14VmVTVlRUN1FGbFBjLzc3OVZsRkdXMC8xWVVaCmpqakNsSkFLUWZOTHg5bG85cWJERVBZL3grZmhnRnYyWUVaSTNTdGVyZWUvVDJoWEFnTUJBQUdqV2pCWU1COEcKQTFVZEl3UVlNQmFBRkNJVy9JYkZIUUNBeEdxZC9FVFpXbVhGeGMvQ01Ba0dBMVVkRXdRQ01BQXdDd1lEVlIwUApCQVFEQWdUd01CMEdBMVVkRVFRV01CU0NFbVJoZEdGMGNuVmphMlZ5TFdGd2FTNXBiekFOQmdrcWhraUc5dzBCCkFRc0ZBQU9DQVFFQUxrMlJDbnVweDA5Wmx0SmN3M1pUaFIveUZGNS9KVlVsWkdHTnc3MlAwQmR6SEloQncrVk4KRkNucGFPMjJ0WGF1NlJVeTVNUWhwc3JrbFNKYVdJT25jOUg0Uk0vbGh3YWsyRXN0MkFhdUFwNUdYSVFJMDVBeAo3NmtDZXEyMnUyalNXRWpFS1hGRnRiZXVDZnFpWlpqNWFLcGF1YlZWRG4ydEZTWjJRMFRib2RRN2dpTFdNUTBLCk5ycVgwQWNSU2J0ZnRSNjBhdGQ1YW41aFNXY0dyZUg5ZHdFUVRvVVNFblg2RFRZSURoeElsZStjQ2V5VjVSWk8KVS9hck5RZ0lFY3ZrZUM1d1lZaFFFL1FUN3lNb1lNajFnckxmOTZSZkRhbEVFNE1seUtDWm9wYWNJWUdzSmhuKwpLMWFISEQvbGZxTC9zbnNtY3pXdHlLUncxTDF1WkxPcjJnPT0KLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=",
                      "truckerapikey": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQpNSUlFb3dJQkFBS0NBUUVBNHd2SDNnVXgvUDdJTTlUNnRqMmpqd2UvV3BLOGhhR0dSVnFMNWxmYktObXpBODhlClQyQjBhN2tCVzJ6eHZWRlZ1Z0g0aDhBNkdiNExFbE02eXI2eGszbS9aK1IwVlZEZGZiZUMwdTVqMll1OVVKV3YKSms4QVlLRGF2QzJZcTcwOE9iU20vUHJINlVhZ0NFd1JKUlpWRHhSeXZYcWZiOFZxSWNhYllTbGpsUy8xbjVtVgptbU1rTC9TdU5IV0FZaEl4YXRTWGo4ZFBZRDFodDR6VGxGek9qMEdDNGF3a0toTFAxTkRoV0tVMHhrWEZuUzhTCjFCSGtuYWF5ZmxRN01nL21GTWdFQm8ybXpNVlhrbFUwKzBCWlQzUCsrL1ZaUlJsdFA5V0ZHWTQ0d3BTUUNrSHoKUzhmWmFQYW13eEQyUDhmbjRZQmI5bUJHU04wclhxM252MDlvVndJREFRQUJBb0lCQUZSbFo2aiswMk1RcmdpZQpUWWQ2aUJhOXNZS1FjdVZWY2x6TFdKK0ozdHhqZCtpZGo4Mi9zK2JyQVRjajJ3UWRmTGxFekVaaHpQY09ybnprClZVRlIreDR2cmxlSEV6czNjTzJ2VGtEci9rKzZZdWZXakNHNFVoS1pHRlhFdVRCMWlqS0ZGV0dLTXM5aXdrL3oKUHBoRG9lQ1dJRVNRbFQxdDY1S2NpZHRuRUM0K3oxaURxR21ZZEVEa2xEZFFIVHorRGpseS80VzV2MFpVQWN2Vwo1Unkvd0JldmNURlBXQy9uZGFrN3hGUnVrU2xNZUlhTWhvTlllOU5uVCt2Nmh5VVh1MXp6cG45UEFJRTVpQWpOCnBsZkFEZitwakwrZFJDbW9sWW1LR3NQVjEzSmtCY1p2cEVoUWgrUmJnaWFwczJLc2plQ0NmTmtST0hndHc0MU0KeGxOTU1Ha0NnWUVBOHM1akpuZ2FMblIvTzZtakZEb0wvd1p2UWdnT1h3Z1RBN2tsQ0trbjBYM0xJS0hhaTZESgpyOVRGQktMdkhUazVFK1Y4NkhIZ29VMUFQQnYyMWJpaHhuQXhUYm8wSHZMdG5QOXU3YnZJejFNOWMyd2ZlWUgzCmliMG1FNkV0Y2wvNnVIUHdMVlZtRGh5bllneEZpeTRRRXp1ZHZwaFIvSnlPazQvUytQUzh2ZXNDZ1lFQTcySW8KWDlkY0xvUmYzSk0vRTF1SXhrRGRLZm1VakN6Mk1NN3pyRkpNUERobTFSUlJhQjdidHNxUzl4Mm5KeTI5amtoSQpndTR0Y2s5STdncm9WRXY0cEtkU0lOVzVNVUg0MXJPckxTbFBobCtMaTM5QXNrTERFVXVQbkFmYko3YWZibFFHCjNFWE1Ka1d2WFQxRG90aE11MUpUMTFPMjhmYlBmb3R3YVRucXFFVUNnWUE1RnArblh0dkJDbkVUNVJuTDJxNFIKWTBEcmhKdGthckNWMG9WTG1RSWN0bmY1RHpVa1dVU3lpOU5hK0xoQjViOXUxeWxjT2RJMytzTGNuektWSTVrZAo0Q2tuZmwxYzNYOVN4NWdjVzJlbW40cTJSYmpNeCtzY3ZwVmlUaVl5V3pESEI3Tno1VDFlbWd3YytJZXRac0l1CjAzaU9aSi82UmhFYnp0Z3E0dGd4eXdLQmdCanhwSDFnSm9YZVcxTFRzUlhtU3RsRnR3WUJzTDN1alhUR2k5T2YKKzd1bXNPaHhKbzV2OExzYVhyZ0xTZmltSm9ObCt0QlIyMEIzRElWcGM0NGRJSHFmeU1vTkd6QUx6TThxaDZLOApGRUpQUUc0M3cxa1A1ZDJ1ZnZkbVJpcmNONzY2V01DcE1UMm5zejlHKy81REJiUGJQMHRIdU45N3RXRmRURWhhCnRaMzFBb0dCQU5xZnVycEV4ZHozdXRNUm91RnJjTmVkV2s4VlVpVVZIV1hZMjhNNWQvRjcxMExIOTZuOW12V08Kc1cwcDFjcXBJMEsyQko2WlYwNVpybnlyRHBsVVpVdVpSQWxGR3hYR1dLU2trQzlmWm90SkFOWDBkR29KSitqTwpEYUNLNGxlcXdOMHNPRE1YZFErMlFPaUtlSUxkNU5KcDNUSW1YV1Z0RFlXUkRwOW9iSUdlCi0tLS0tRU5EIFJTQSBQUklWQVRFIEtFWS0tLS0tCg=="
                    },
                    "jobsServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "debug"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": true,
                        "FileSFTP": true,
                        "IOTKafka": true,
                        "IOTProxy": false,
                        "IOTRedis": true,
                        "Login": false,
                        "ScriptJS": true,
                        "ScriptSSH": true,
                        "ScriptShell": true,
                        "UtilEcho": true,
                        "UtilFuzzy": true,
                        "UtilSentiment": true,
                        "credentials": false,
                        "jobs": true,
                        "resources": false,
                        "statuschecks": true,
                        "ui": false,
                        "usermanagement": false
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "loginServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "info"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": false,
                        "FileSFTP": false,
                        "IOTKafka": false,
                        "IOTProxy": false,
                        "IOTRedis": false,
                        "Login": true,
                        "ScriptJS": false,
                        "ScriptSSH": false,
                        "ScriptShell": false,
                        "UtilEcho": false,
                        "UtilFuzzy": false,
                        "UtilSentiment": false,
                        "credentials": false,
                        "jobs": false,
                        "resources": false,
                        "statuschecks": true,
                        "ui": false,
                        "usermanagement": false
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "managementServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "debug"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": false,
                        "FileSFTP": false,
                        "IOTKafka": false,
                        "IOTProxy": false,
                        "IOTRedis": false,
                        "Login": true,
                        "ScriptJS": false,
                        "ScriptSSH": false,
                        "ScriptShell": false,
                        "UtilEcho": false,
                        "UtilFuzzy": false,
                        "UtilSentiment": false,
                        "credentials": true,
                        "jobs": false,
                        "resources": false,
                        "statuschecks": true,
                        "ui": true,
                        "usermanagement": true
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "name": "API",
                    "resource": {
                      "Cache": {
                        "Creds": false,
                        "CredsFlush": 10,
                        "Resource": false,
                        "ResourceFlush": 10
                      },
                      "Templates": {
                        "Scripts": "scripts",
                        "Shell": "shell",
                        "keys": "keys",
                        "resourcedefinitions": "resourcedefinitions"
                      }
                    }
                  },
                  "DB": {
                    "databasename": "userdb",
                    "hostname": "db",
                    "password": "password",
                    "port": 5432,
                    "type": "pg",
                    "user": "testuser"
                  },
                  "TempDB": {
                    "enabled": false,
                    "pvc": "samplepvc"
                  }
                }
              },
              {
                "apiVersion": "datatrucker.datatrucker.io/v1",
                "kind": "DatatruckerFlow",
                "metadata": {
                  "name": "datatruckerflow-sample"
                },
                "spec": {
                  "API": {
                    "Image": {
                      "imageName": "datatrucker-server/datatrucker-api",
                      "imageNameUI": "datatruckerio/datatrucker-ui",
                      "repository": "registry.gitlab.com/datatruckerio",
                      "tagName": "latest"
                    },
                    "name": "api"
                  },
                  "DatatruckerConfig": "datatruckerconfig-sample",
                  "JobDefinitions": [
                    {
                      "name": "echoapi",
                      "restmethod": "GET",
                      "tenant": "Admin",
                      "type": "Util-Echo",
                      "validations": {
                        "properties": {
                          "voila": {
                            "enum": [
                              "true",
                              "false"
                            ],
                            "type": "string"
                          }
                        },
                        "type": "object"
                      }
                    }
                  ],
                  "Keys": {
                    "configmap": "placeholder"
                  },
                  "Replicas": 1,
                  "Resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "256Mi"
                    },
                    "requests": {
                      "cpu": "250m",
                      "memory": "256Mi"
                    }
                  },
                  "Scripts": {
                    "configmap": "placeholder"
                  },
                  "Type": "Job"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/datatruckerio/datatrucker-operator:2.0.0
          createdAt: "2021-09-08 01:01:01"
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: hhttps://gitlab.com/datatruckerio/datatrucker-server.git
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              - creates secrets to hold server configs
              - create tempDB for protoypting
              - create configs for seperating Login, management and Jobs endpoints
              - Create crypto configs
            displayName: Datatrucker Config
            kind: DatatruckerConfig
            name: datatruckerconfigs.datatrucker.datatrucker.io
            version: v1
          - description: '- creates an API as per Job Definitions'
            displayName: Datatrucker Flow
            kind: DatatruckerFlow
            name: datatruckerflows.datatrucker.datatrucker.io
            version: v1
        description: "## DataTrucker.IO\n\nDataTrucker.IO is a simple no-code / less-code
          API Backend\n\n### What does it do ?\n\nDatatrucker.IO is a product capable
          of reading simple json/yaml configs and building th code necessary for converting
          it into an API.\nAlong with building the code, it will also host the code
          base on a nodejs server , i.e. it will immediately make it available for
          consumption.\n\n### What does it offload from a developer ?\n\nDataTrucker
          is capable of removing the most common activities which a developer needs
          to do on every new project A few of such common activities are\n\n- Creating
          an API endpoint with a specified business logic (using simple plugins)\n-
          Applying standard RBAC\n- Applying Authorization logic\n- Applying hardening
          on endpoints\n- Log management\n- Connecting to variety of systems\n- Modularizing
          business logic\n- The best of doing it with little to no code\n\n### watch
          the quick elevator pitch to understand datatrucker.io\n\n https://youtu.be/DKLFDqhjs5M\n "
        displayName: DataTrucker.IO
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - datatrucker
        - operator
        - nocode
        - apibuilder
        links:
        - name: Website
          url: https://www.datatrucker.io/
        - name: Git Repo
          url: https://gitlab.com/datatruckerio/datatrucker-server.git
        maintainers:
        - email: gauravshankar.can@gmail.com
          name: Gaurav Shankar
        maturity: alpha
        provider:
          name: DataTrucker.IO
          url: https://www.datatrucker.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - docker.io/datatruckerio/datatrucker-operator:2.0.0
        version: 2.0.0
      entries:
      - name: datatrucker-operator.v2.0.0
        version: 2.0.0
      - name: datatrucker-operator.v1.3.2
        version: 1.3.2
      - name: datatrucker-operator.v1.3.1
        version: 1.3.1
      - name: datatrucker-operator.v1.3.0
        version: 1.3.0
      name: alpha
    defaultChannel: alpha
    packageName: datatrucker-operator
    provider:
      name: DataTrucker.IO
      url: https://www.datatrucker.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: postgresql
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: registry.developers.crunchydata.com/crunchydata/postgres-operator:ubi9-5.8.3-0
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.developers.crunchydata.com/crunchydata/postgres-operator:ubi9-5.8.3-0
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: postgresql
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Vertica
      provider-url: http://www.vertica.com
    name: verticadb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: verticadb-operator.v1.11.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "EventTrigger",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "verticadb-operator",
                    "app.kubernetes.io/instance": "eventtrigger-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eventtrigger",
                    "app.kubernetes.io/part-of": "verticadb-operator"
                  },
                  "name": "eventtrigger-sample"
                },
                "spec": {
                  "matches": [
                    {
                      "condition": {
                        "status": "True",
                        "type": "DBInitialized"
                      }
                    }
                  ],
                  "references": [
                    {
                      "object": {
                        "apiVersion": "vertica.com/v1beta1",
                        "kind": "VerticaDB",
                        "name": "verticadb-sample"
                      }
                    }
                  ],
                  "template": {
                    "metadata": {
                      "generateName": "create-user-table-"
                    },
                    "spec": {
                      "template": {
                        "spec": {
                          "containers": [
                            {
                              "command": [
                                "/opt/vertica/bin/vsql",
                                "-h",
                                "verticadb-sample-defaultsubcluster",
                                "-c",
                                "CREATE TABLE T1 (C1 INT);"
                              ],
                              "image": "vertica/vertica-k8s:latest",
                              "name": "main"
                            }
                          ],
                          "restartPolicy": "OnFailure"
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "serviceName": "defaultsubcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaDB",
                "metadata": {
                  "name": "verticadb-sample"
                },
                "spec": {
                  "communal": {
                    "endpoint": "https://s3.amazonaws.com"
                  },
                  "image": "vertica/vertica-k8s:23.3.0-0-minimal",
                  "subclusters": [
                    {
                      "name": "default_subcluster",
                      "size": 3
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: docker.io/vertica/verticadb-operator:1.11.2
          createdAt: "2023-08-02T11:29:39Z"
          description: Operator that manages a Vertica Eon Mode database.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/vertica/vertica-kubernetes
          support: Vertica
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventTrigger is the Schema for the eventtriggers API
            displayName: Event Trigger
            kind: EventTrigger
            name: eventtriggers.vertica.com
            version: v1beta1
          - description: VerticaAutoscaler is a CR that allows you to autoscale one
              or more subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1beta1
          - description: VerticaDB is the CR that defines a Vertica Eon mode cluster
              that is managed by the verticadb-operator.
            displayName: Vertica DB
            kind: VerticaDB
            name: verticadbs.vertica.com
            version: v1beta1
        description: |-
          ### What is Vertica?
          Vertica is a unified analytics platform, based on a massively scalable architecture with the broadest set of analytical functions spanning event and time series, pattern matching, geospatial and end-to-end in-database machine learning. Vertica enables you to easily apply these powerful functions to the largest and most demanding analytical workloads, arming you and your customers with predictive business insights faster than any analytics data warehouse in the market. Vertica provides a unified analytics platform across major public clouds and on-premises data centers and integrates data in cloud object storage and HDFS without forcing you to move any of your data.

          ### What does this operator do?
          This operator manages a Vertica Eon Mode database, and automates the following administrative tasks:

          * Installing Vertica
          * Creating and reviving a Vertica database
          * Restarting and rescheduling DOWN pods to maintain quorum
          * Upgrading Vertica to a new version while keeping the database online
          * Subcluster scaling, both manually and automatically with the Horizontal Pod Autoscaler
          * Service management and health monitoring for pods
          * Load balancing for internal and external traffic
          * Expose a metrics endpoint to integrate with Prometheus to monitor the operator's health

          For a brief overview on how to create a cluster, see the [Vertica GitHub repository](https://github.com/vertica/vertica-kubernetes). For an in-depth look at Vertica on Kubernetes, see the [Vertica documentation](https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm).
        displayName: VerticaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - database
        - vertica
        links:
        - name: Documentation
          url: https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm
        - name: Vertica Container Images
          url: https://hub.docker.com/u/vertica
        maintainers:
        - email: matt.spilchen@vertica.com
          name: Matt Spilchen
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Vertica
          url: http://www.vertica.com
        relatedImages:
        - docker.io/vertica/verticadb-operator:1.11.2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 1.11.2
      entries:
      - name: verticadb-operator.v1.11.2
        version: 1.11.2
      - name: verticadb-operator.v1.11.1
        version: 1.11.1
      - name: verticadb-operator.v1.11.0
        version: 1.11.0
      - name: verticadb-operator.v1.10.2
        version: 1.10.2
      - name: verticadb-operator.v1.10.1
        version: 1.10.1
      - name: verticadb-operator.v1.10.0
        version: 1.10.0
      - name: verticadb-operator.v1.9.0
        version: 1.9.0
      - name: verticadb-operator.v1.8.0
        version: 1.8.0
      - name: verticadb-operator.v1.7.0
        version: 1.7.0
      - name: verticadb-operator.v1.6.0
        version: 1.6.0
      - name: verticadb-operator.v1.5.0
        version: 1.5.0
      - name: verticadb-operator.v1.4.0
        version: 1.4.0
      - name: verticadb-operator.v1.3.1
        version: 1.3.1
      - name: verticadb-operator.v1.3.0
        version: 1.3.0
      - name: verticadb-operator.v1.2.0
        version: 1.2.0
      name: stable
    - currentCSV: verticadb-operator.v25.3.1-0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vertica.com/v1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "customAutoscaler": {},
                  "scalingGranularity": "Subcluster",
                  "serviceName": "default-subcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1",
                "kind": "VerticaDB",
                "metadata": {
                  "name": "verticadb-sample"
                },
                "spec": {
                  "communal": {
                    "endpoint": "https://s3.amazonaws.com"
                  },
                  "image": "opentext/vertica-k8s:24.1.0-0-minimal",
                  "subclusters": [
                    {
                      "name": "default_subcluster",
                      "size": 3
                    }
                  ]
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "EventTrigger",
                "metadata": {
                  "name": "eventtrigger-sample"
                },
                "spec": {
                  "matches": [
                    {
                      "condition": {
                        "status": "True",
                        "type": "DBInitialized"
                      }
                    }
                  ],
                  "references": [
                    {
                      "object": {
                        "apiVersion": "vertica.com/v1beta1",
                        "kind": "VerticaDB",
                        "name": "verticadb-sample"
                      }
                    }
                  ],
                  "template": {
                    "metadata": {
                      "generateName": "create-user-table-"
                    },
                    "spec": {
                      "template": {
                        "spec": {
                          "containers": [
                            {
                              "command": [
                                "/opt/vertica/bin/vsql",
                                "-h",
                                "verticadb-sample-defaultsubcluster",
                                "-c",
                                "CREATE TABLE T1 (C1 INT);"
                              ],
                              "image": "opentext/vertica-k8s:23.4.0-0-minimal",
                              "name": "main"
                            }
                          ],
                          "restartPolicy": "OnFailure"
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "customAutoscaler": {},
                  "scalingGranularity": "Subcluster",
                  "serviceName": "default-subcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaReplicator",
                "metadata": {
                  "name": "verticareplicator-sample"
                },
                "spec": {
                  "source": {
                    "verticaDB": "sourceDBName"
                  },
                  "target": {
                    "verticaDB": "targetDBName"
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaRestorePointsQuery",
                "metadata": {
                  "name": "verticarestorepointsquery-sample"
                },
                "spec": {
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaScrutinize",
                "metadata": {
                  "name": "verticascrutinize-sample"
                },
                "spec": {
                  "verticaDBName": "verticadb-sample"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: docker.io/opentext/verticadb-operator:25.3.1-0
          createdAt: "2025-09-24T00:58:06Z"
          description: Operator that manages a Vertica Eon Mode database.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/vertica/vertica-kubernetes
          support: Vertica
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventTrigger is the Schema for the eventtriggers API
            displayName: Event Trigger
            kind: EventTrigger
            name: eventtriggers.vertica.com
            version: v1beta1
          - description: |-
              VerticaAutoscaler is a CR that allows you to autoscale one or more
              subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1
          - description: |-
              VerticaAutoscaler is a CR that allows you to autoscale one or more
              subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1beta1
          - description: VerticaDB is the CR that defines a Vertica Eon mode cluster
              that is managed by the verticadb-operator.
            displayName: Vertica DB
            kind: VerticaDB
            name: verticadbs.vertica.com
            version: v1
          - description: VerticaReplicator is the Schema for the verticareplicators
              API
            displayName: Vertica Replicator
            kind: VerticaReplicator
            name: verticareplicators.vertica.com
            version: v1beta1
          - description: VerticaRestorePointsQuery is the Schema for the verticarestorepointsqueries
              API
            displayName: Vertica Restore Points Query
            kind: VerticaRestorePointsQuery
            name: verticarestorepointsqueries.vertica.com
            version: v1beta1
          - description: VerticaScrutinize is the schema for verticascrutinize API
            displayName: Vertica Scrutinize
            kind: VerticaScrutinize
            name: verticascrutinizers.vertica.com
            version: v1beta1
        description: |-
          ### What is Vertica?
          Vertica is a unified analytics platform, based on a massively scalable architecture with the broadest set of analytical functions spanning event and time series, pattern matching, geospatial and end-to-end in-database machine learning. Vertica enables you to easily apply these powerful functions to the largest and most demanding analytical workloads, arming you and your customers with predictive business insights faster than any analytics data warehouse in the market. Vertica provides a unified analytics platform across major public clouds and on-premises data centers and integrates data in cloud object storage and HDFS without forcing you to move any of your data.

          ### What does this operator do?
          This operator manages a Vertica Eon Mode database, and automates the following administrative tasks:

          * Installing Vertica
          * Creating and reviving a Vertica database
          * Restarting and rescheduling DOWN pods to maintain quorum
          * Upgrading Vertica to a new version while keeping the database online
          * Subcluster scaling, both manually and automatically with the Horizontal Pod Autoscaler
          * Service management and health monitoring for pods
          * Load balancing for internal and external traffic
          * Expose a metrics endpoint to integrate with Prometheus to monitor the operator's health

          For a brief overview on how to create a cluster, see the [Vertica GitHub repository](https://github.com/vertica/vertica-kubernetes). For an in-depth look at Vertica on Kubernetes, see the [Vertica documentation](https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm).
        displayName: VerticaDB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - vertica
        links:
        - name: Documentation
          url: https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm
        - name: Vertica Container Images
          url: https://hub.docker.com/u/vertica
        maintainers:
        - email: cchen8@opentext.com
          name: Cai Chen
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Vertica
          url: http://www.vertica.com
        relatedImages:
        - docker.io/opentext/verticadb-operator:25.3.1-0
        version: 25.3.1-0
      entries:
      - name: verticadb-operator.v25.3.1-0
        version: 25.3.1-0
      - name: verticadb-operator.v25.3.0-0
        version: 25.3.0-0
      - name: verticadb-operator.v25.2.1-0
        version: 25.2.1-0
      - name: verticadb-operator.v25.2.0-0
        version: 25.2.0-0
      - name: verticadb-operator.v25.1.0-0
        version: 25.1.0-0
      - name: verticadb-operator.v24.4.0-0
        version: 24.4.0-0
      - name: verticadb-operator.v2.2.0
        version: 2.2.0
      - name: verticadb-operator.v2.1.3
        version: 2.1.3
      - name: verticadb-operator.v2.1.2
        version: 2.1.2
      - name: verticadb-operator.v2.1.1
        version: 2.1.1
      - name: verticadb-operator.v2.1.0
        version: 2.1.0
      - name: verticadb-operator.v2.0.2
        version: 2.0.2
      - name: verticadb-operator.v2.0.1
        version: 2.0.1
      - name: verticadb-operator.v2.0.0
        version: 2.0.0
      name: v2-stable
    defaultChannel: v2-stable
    packageName: verticadb-operator
    provider:
      name: Vertica
      url: http://www.vertica.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: F5 Networks Inc.
      provider-url: ""
    name: f5-ipam-controller-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: f5-ipam-controller-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fic.f5.com/v1",
                "kind": "F5IpamCtlr",
                "metadata": {
                  "name": "f5ipamctlr-sample"
                },
                "spec": {
                  "args": {
                    "infoblox_grid_host": "IP_ADDRESS_OF_SERVER",
                    "infoblox_labels": "CIDR_Labels",
                    "infoblox_netview": "NetView",
                    "infoblox_password": "PASSWORD",
                    "infoblox_username": "USERNAME",
                    "infoblox_wapi_version": "VERSION",
                    "insecure": true,
                    "ip_range": "IP_RANGE_LABELS",
                    "log_level": "DEBUG_OR_INFO",
                    "orchestration": "kubernetes",
                    "provider": "INFOBLOX_OR_F5-IP-PROVIDER"
                  },
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "f5-ipam-controller",
                    "user": "f5networks",
                    "version": "latest"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "serviceAccount": {
                    "create": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
          createdAt: "2025-08-08T03:48:02Z"
          description: Operator to install F5 IPAM Controller.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/f5-ipam-controller
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This Operator installs F5 IPAM controller(FIC)
            kind: F5IpamCtlr
            name: f5ipamctlrs.fic.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 IPAM Controller in your OpenShift Cluster. This enables to configure and deploy FIC using Helm Charts.
          ## F5 IPAM Controller
          The F5 IPAM Controller is a Docker container that runs in an orchestration environment and interfaces with an IPAM system. It allocates IP addresses from an IPAM system’s address pool for hostnames in an orchestration environment. The F5 IPAM Controller watches orchestration-specific resources and consumes the hostnames within each resource.
          ## Documentation
          Refer to F5 IPAM Controller documentation
          - https://github.com/F5Networks/f5-ipam-controller/blob/main/README.md
          ## Prerequisites
          ### Configure Infoblox with network and netview [Provider: Infoblox]
          Refer Infoblox documentation
          ### Create persistent volume and persistent volume claim [Provider: f5-ip-provider]
          Create persistent volume and persistent volume claim
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/f5-ipam-controller/main/docs/config_examples/f5-ip-provider/localstorage-pv-pvc-example.yaml
          ```
        displayName: F5 Ipam controller
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipam
        - ip
        - ipaddress
        - controller
        - ipv4
        - ipv6
        - F5
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/userguide/ipam/
        - name: Github Repo
          url: https://github.com/F5Networks/f5-ipam-controller
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
        - registry.connect.redhat.com/f5networks/f5-ipam-controller@sha256:09a358db300d82e00fd3b4e255917df82c02e1bc58425f8cff21cc2c77eba1b6
        version: 0.0.8
      entries:
      - name: f5-ipam-controller-operator.v0.0.8
        version: 0.0.8
      name: alpha
    - currentCSV: f5-ipam-controller-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fic.f5.com/v1",
                "kind": "F5IpamCtlr",
                "metadata": {
                  "name": "f5ipamctlr-sample"
                },
                "spec": {
                  "args": {
                    "infoblox_grid_host": "IP_ADDRESS_OF_SERVER",
                    "infoblox_labels": "CIDR_Labels",
                    "infoblox_netview": "NetView",
                    "infoblox_password": "PASSWORD",
                    "infoblox_username": "USERNAME",
                    "infoblox_wapi_version": "VERSION",
                    "insecure": true,
                    "ip_range": "IP_RANGE_LABELS",
                    "log_level": "DEBUG_OR_INFO",
                    "orchestration": "kubernetes",
                    "provider": "INFOBLOX_OR_F5-IP-PROVIDER"
                  },
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "f5-ipam-controller",
                    "user": "f5networks",
                    "version": "latest"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "serviceAccount": {
                    "create": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
          createdAt: "2025-08-08T03:48:02Z"
          description: Operator to install F5 IPAM Controller.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/f5-ipam-controller
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This Operator installs F5 IPAM controller(FIC)
            kind: F5IpamCtlr
            name: f5ipamctlrs.fic.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 IPAM Controller in your OpenShift Cluster. This enables to configure and deploy FIC using Helm Charts.
          ## F5 IPAM Controller
          The F5 IPAM Controller is a Docker container that runs in an orchestration environment and interfaces with an IPAM system. It allocates IP addresses from an IPAM system’s address pool for hostnames in an orchestration environment. The F5 IPAM Controller watches orchestration-specific resources and consumes the hostnames within each resource.
          ## Documentation
          Refer to F5 IPAM Controller documentation
          - https://github.com/F5Networks/f5-ipam-controller/blob/main/README.md
          ## Prerequisites
          ### Configure Infoblox with network and netview [Provider: Infoblox]
          Refer Infoblox documentation
          ### Create persistent volume and persistent volume claim [Provider: f5-ip-provider]
          Create persistent volume and persistent volume claim
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/f5-ipam-controller/main/docs/config_examples/f5-ip-provider/localstorage-pv-pvc-example.yaml
          ```
        displayName: F5 Ipam controller
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipam
        - ip
        - ipaddress
        - controller
        - ipv4
        - ipv6
        - F5
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/userguide/ipam/
        - name: Github Repo
          url: https://github.com/F5Networks/f5-ipam-controller
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
        - registry.connect.redhat.com/f5networks/f5-ipam-controller@sha256:09a358db300d82e00fd3b4e255917df82c02e1bc58425f8cff21cc2c77eba1b6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        version: 0.0.8
      entries:
      - name: f5-ipam-controller-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: f5-ipam-controller-operator
    provider:
      name: F5 Networks Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: smart-gateway-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smart-gateway-operator.v5.0.1757650606
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "smartgateway.infra.watch/v2",
                "kind": "SmartGateway",
                "metadata": {
                  "name": "demo"
                },
                "spec": {
                  "applications": [
                    {
                      "config": "host: localhost\nport: 8081\nwithtimestamp: false\n",
                      "name": "prometheus"
                    }
                  ],
                  "blockEventBus": false,
                  "bridge": {
                    "amqpBlock": false,
                    "amqpUrl": "amqp://amq-interconnect:5672/collectd",
                    "enabled": true,
                    "ringBufferCount": 15000,
                    "ringBufferSize": 2048,
                    "socketBlock": true,
                    "statsPeriod": 60,
                    "stopCount": 0,
                    "unixSocketPath": "/tmp",
                    "verbose": true
                  },
                  "handleErrors": false,
                  "logLevel": "info",
                  "pluginDir": "/usr/lib64/sg-core",
                  "services": [
                    {
                      "name": "prometheus",
                      "ports": [
                        {
                          "name": "prom-http",
                          "port": 8081,
                          "protocol": "TCP",
                          "targetPort": 8081
                        }
                      ]
                    }
                  ],
                  "size": 1,
                  "transports": [
                    {
                      "config": "path: /tmp/smartgateway\n",
                      "handlers": [
                        {
                          "config": "",
                          "name": "collectd-metrics"
                        }
                      ],
                      "name": "socket"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/stf/smart-gateway-rhel9-operator@sha256:3363abcb91319f8642212b9b80ca49bb9f90cae69255b1f75f03929af3d3a8c3
          createdAt: "2025-09-12T04:17:59Z"
          description: Operator for managing the Smart Gateway Custom Resources, resulting
            in deployments of the Smart Gateway.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenStack Platform", "Cloud
            Infrastructure", "Cloud Suite"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/infrawatch/smart-gateway-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of the Smart Gateway
            displayName: SmartGateway
            kind: SmartGateway
            name: smartgateways.smartgateway.infra.watch
            version: v2
        description: |-
          # Smart Gateway for Service Telemetry Framework

          The Smart Gateway is middleware for OpenShift that takes metrics and events
          data streams from an AMQP 1.x message bus, and provides a method to having
          that data stream stored within supported data stores such as Prometheus and
          ElasticSearch.

          ## Core capabilities

          The Smart Gateway provides two modes:

          * metrics: provides an HTTP scrape endpoint for Prometheus
          * events: writes events directly to an ElasticSearch endpoint

          ## Operator features

          * **Two Service Types** - Supports the `metrics` and `events` service types which defines the type of Smart Gateway to be deployed

          * **Configuration** - Configuration of the Smart Gateway to allow for it to be connected to an AMQP 1.x bus and pointed at the appropriate storage mechanism based on service type.

          ## Before getting started

          1. Ensure an AMQP 1.x bus has been setup (such as AMQ Interconnect)
          2. If the Smart Gateway is running in events mode, be sure the ElasticSearch Operator and ElasticSearch instance have been pre-deployed.
        displayName: Smart Gateway Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - servicetelemetry
        - monitoring
        - telemetry
        - notifications
        - smartgateway
        links:
        - name: Source Code
          url: https://github.com/infrawatch/smart-gateway-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/stf/smart-gateway-rhel9-operator@sha256:3363abcb91319f8642212b9b80ca49bb9f90cae69255b1f75f03929af3d3a8c3
        - registry.redhat.io/stf/sg-core-rhel9@sha256:0e12903da0b15c64421102b5b914ca4c82d107295485c9b76ed19e91a45a4420
        - registry.redhat.io/stf/sg-bridge-rhel9@sha256:7fae342cf6cecc676eaa0a8fb64ee380ad101d667d6e8f0d89c943938bb6baf4
        - registry.redhat.io/openshift4/ose-oauth-proxy:latest@sha256:dc24aa382e02c4ac058d4b824a21f0e90a6259d6d817d0341b2a41809b5daa78
        version: 5.0.1757650606
      entries:
      - name: smart-gateway-operator.v5.0.1757650606
        version: 5.0.1757650606
      - name: smart-gateway-operator.v5.0.1743671602
        version: 5.0.1743671602
      - name: smart-gateway-operator.v5.0.1740051852
        version: 5.0.1740051852
      - name: smart-gateway-operator.v5.0.1731336910
        version: 5.0.1731336910
      - name: smart-gateway-operator.v5.0.1725953602
        version: 5.0.1725953602
      name: stable-1.5
    defaultChannel: stable-1.5
    packageName: smart-gateway-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-pipelines-operator-rh
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-pipelines-operator-rh.v1.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    },
                    {
                      "name": "legacyPipelineRbac",
                      "value": "true"
                    },
                    {
                      "name": "createCABundleConfigMaps",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.19.0 <1.20.0'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.3.2
          - Tekton Triggers: v0.33.0
          - Pipelines as Code: v0.37.0
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.22.3
          - Tekton Results (tech-preview): v0.16.0
          - Manual Approval Gate (tech-preview): v0.7.0
          - Tekton Pruner (tech-preview): v0.2.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.20/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:5021158a4752a26f8b33adea7da66695b22d2ad155d346d35dde5f971bc1a55b
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:5683321a5376e63c3ff3932270a1850d5e261f76212020ce867334166f90d2f7
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:a712bb411d2433b01793dff25785c1e265fa865a52f7c8732c9fb4418ab61757
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e7998e113f75e8f29602bae210c8b66ed06634b699f1d99c5875ef879220d733
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:96db721939b22572d7d8d8b95dcb5c92fdedb98e9cd0aabb7d17b703a915b07f
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:115eea1a5ca87f3b6b72db1df05b7386bf73dc70c9c2e4dc5e1a66700d960c37
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:980ed7e81ea55522e1681dc1e9b170b68d16f07061dd1f1ffd142367d59552f1
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:1b6aecfce45ed3163227b845b212e83385cdbe999edfc2e9387a0c76f3f8d8e9
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:161df929e2bd4c5acff594be31d5458365de5ae0d2189bb2f647c3fc6657fa28
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:bc10e7d6122ede5cf4f2882c200447683255465a1080e826761e90bfb983506e
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:df15eb1f1f5f237bfe481c16ab3d40bf7fb4123f3c930bcfb121e2cc8dc42882
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:d56d6b46331b4569077ab1e7a7b609ba01f0a1dd62b695d62f8aa0f32ea1b621
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:c0f438e6095aa81f3f7d820614e0414cc43374014c1505d7641bc1dba69a3903
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:818bfa4a3a874974b5f7f8bcf2faeb6c05e1795381ee1ceacf89f17b1ec0ef7f
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:9415e889ef98df6d2b2dca1e5d798fb765eeea0d9828e48424374d5267a6b0d1
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:441e1c7124c1674e3d4c38c3e82a4fe5568e76d13c89c04bf009d6b97435de09
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:7eafffa6e4e04f10c1d4e84e02d51c00eb8649c60f8f9a13da14492f24634891
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:bc403f93ef6c5526fa30152eab384c4a25bb0bc6ab4158a1e24f637b87a57259
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:2d496752b19f765f1ea36ec94b873ebd8fa604764a56aa67d3ed94b33e5fe75c
        - registry.redhat.io/rhel9/buildah@sha256:18453ab0a62154283aaf4c45efd7543d1c8ecca1aebae46dedc7ddc0f410b232
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:036250bcdc6fb95273bf07948dfdd4dce765ab84c803ceca0d0b4abf6ca19dec
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:8f0494341244a75546a4b80b065702be04e6d200421aa4e7ec90385c9dabc6ab
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:c4d05e546c466c57379fff502816012f615be073eb8cdce54677b4ab206a3f5c
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:bbd1918660be7b3c78777dcccc87d3b26bf5adfc077844ceffbbdd53696dc0a3
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:e3e64abef53151f6b15f15bcbff85668167a571e98badbd8a2ca252ff3a5b8dc
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:dd941acd357ae2c0fd7e15fb8d32d7ee6b2847435101dd1e1f107db48af18bf3
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:823e018e0329d764ac694f177ce88af9074f364c29a1c17929f9dd0b596d03dd
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:a7163a9e1fe891de021b1003558e6e64fe17ed17ab0820ba16bba09aa112401e
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:e63298d07ccec5804dfab4e42bcc8163c29cb3ac371c91248777423a4edd31ce
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:3cf990fb2fc422f3527a7c8b43d06ccd4d32bb5487ec8cb2d4a97f0f5d13f170
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:c3013c347bca8a520443a44ac2817fbba5744a9e91069ac5f09e91447ebf6b28
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:901b5108ceb669c985e75f5fb033c496e9f028863df99ab32651d1b08433a352
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:14dab8d87323fc7d112213f7f4db3ee7cf3841747de669ce710ae6a794053457
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:458f865f1b246c5f0682f5956795771a0515b89edfcffbbc0da1a3d3f532645e
        - registry.redhat.io/ubi9/ubi-minimal@sha256:2f06ae0e6d3d9c4f610d32c480338eef474867f435d8d28625f2985e8acde6e8
        - registry.redhat.io/ubi9/openjdk-17@sha256:555c8bcd2596637befd1a6da4aa949c5a2640cf66e09c2dfb932afe37e39095e
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:c31af237afb84214bf2a4c06a7636ca071fe97203b41da90d7878109ccd64784
        - registry.redhat.io/openshift-pipelines/pipelines-opc-rhel9@sha256:c1bee4a2bb495de6d4da962b6e7b2447caa51cc36c3005e4083bd31420ef5102
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:ec74a562c18fefe36ec7ae3ba083e4eb76c56b562f916578900f8767ae98e6c2
        - registry.redhat.io/rhel9/skopeo@sha256:a65a413f8a2864389a09dc750690d97afbcdc5c70821e0f85e99e8adba7954e0
        version: 1.20.0
      entries:
      - name: openshift-pipelines-operator-rh.v1.20.0
        version: 1.20.0
      - name: openshift-pipelines-operator-rh.v1.19.3
        version: 1.19.3
      - name: openshift-pipelines-operator-rh.v1.19.2
        version: 1.19.2
      - name: openshift-pipelines-operator-rh.v1.19.1
        version: 1.19.1
      - name: openshift-pipelines-operator-rh.v1.19.0
        version: 1.19.0
      - name: openshift-pipelines-operator-rh.v1.18.1
        version: 1.18.1
      - name: openshift-pipelines-operator-rh.v1.18.0
        version: 1.18.0
      - name: openshift-pipelines-operator-rh.v1.17.1
        version: 1.17.1
      - name: openshift-pipelines-operator-rh.v1.17.0
        version: 1.17.0
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      - name: openshift-pipelines-operator-rh.v1.14.1
        version: 1.14.1
      - name: openshift-pipelines-operator-rh.v1.13.0
        version: 1.13.0
      - name: openshift-pipelines-operator-rh.v1.12.2
        version: 1.12.2
      - name: openshift-pipelines-operator-rh.v1.12.1
        version: 1.12.1
      - name: openshift-pipelines-operator-rh.v1.12.0
        version: 1.12.0
      - name: openshift-pipelines-operator-rh.v1.11.1
        version: 1.11.1
      name: latest
    - currentCSV: openshift-pipelines-operator-rh.v1.14.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "db_user": "result",
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.14.6'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.56.4
          - Tekton Triggers: v0.26.2
          - Pipelines as Code: v0.24.6
          - Tekton Chains: v0.20.1
          - Tekton Hub (tech-preview): v1.16.0
          - Tekton Results (tech-preview): v0.9.2

          ## Getting Started
          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/container-platform/OPENSHIFT_DOCS_VERSION/cicd/pipelines/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)
          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi8/openjdk-17@sha256:a8165bc2cd5051a96d6937e25ed03155bbd4b731da6e58cebfe2ea83209c16d8
        - registry.redhat.io/ubi8/ubi-minimal@sha256:6825946f10e8c81586fa7815c4f06016af05cfc231d921ec9699402a6655d201
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:059c08316152c209de13017e322d6fe5eacfde81c3b0039e55b5de8f2f27183a
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:d01ba7ead670d5ab9c1355b90010657988e46b4d1aa0bb68b194eaf4001b2d7a
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:6327385e3d3c9df1ea50e803955d980b4cd35fb15593fcfec0d477d91a45021f
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel8@sha256:6e7c7a1ca458e08ad85e1146815ec00a30016f709b45e28b464043e875c2a3e8
        - registry.redhat.io/rhel8/skopeo@sha256:e297b84ba6f1d5eb2c37e2c3b31c9a9a4c39f9590e7d7c32388165245bde4835
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:ae0260abd2725c2f4bb25c0c07d86290978892a49050e0fd23f466401aba668c
        - registry.redhat.io/rhel8/buildah@sha256:b48f410efa0ff8ab0db6ead420a5d8d866d64af846fece5efb185230d7ecf591
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:2261f4f657f70759f4d0c898e44ffafceb8dbe2144063d50227140463bdbf2c5
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:7a63e7a762b9c7b15057ed9d5bcbea08acb4c8933a300c6ec336824d49a6e937
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:8d4450d68ac03d8275e337d2c55692ef35e954400d0c9f041a943d163a8b7680
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:bf6cf2e87fb19f7aa9a490b83c16af69834c0721220a643710a1b077959e91ca
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:aa6eb4c7ade9d6add1e4ab94197c2ce72a253fea0d7461698ca0e37854f1dcbe
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:71d91c7b80b6d196025f68749bd8518b8b0affd8d5064a7c5c70d125656fbc36
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:db3ff043410d1dd0621fef4483761727f7241d788b136e446fb861953350a2cb
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:5e7f5a2813a591e7b13fea9fc4e2f599b5ec855a015cd5a4c287c4e178358756
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:9c29f6eff1f7019993446d85d226f04c5614a18351dfb4042a8bcb62276810de
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:d9972bc64dfa7e260bd9b949bbb85ef5ae532e7c6df8d874d3d3963af1097726
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:bf0c38beefa24ca3d04ebfa6140a1cc5a9a56fba81ae05799279a559093b9033
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel8@sha256:b3bedee6084e9d9da549d3e637d2bf176aa365b5818e6b6377525577edd710dc
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:82c8a91d521d7e6d81fb462fa2b804f420f52ee062df5cda39545165a2f19e0c
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:ce0ea093c522af329acdc85148c2473ea45cd44270c15f8dd32965b0d56f304f
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:abf4442b6acc714c2392b86af046d5f32d247ee0c3763bb18c57d0473edfcf02
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:e2bd929246d62038a24a0685cb3e4873c131c9b7713d6a5268ca09a8614538be
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:83913744a1bb7ff23091cfa541b807cfb82b38dde0c37e0f1dd9c17cade99775
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:63fc1351191a54d12d4e463a1bb4cc732620cba3000af86ba7f7a579514f5fe4
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:e23c901868229e4f6a4016aa7798f1aba879ff35ffe594f5b0e4d39f1bff9e52
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:f68dc18a4dc6a1886507097551f4a6fce39062d03c6eb86c067d317812b0d34a
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:f9fed20cdfebbb226e035cc62f037174f5ecdb3139a5edc7e39f019f6d2b1e98
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:ddf0cb64ea57de872515e19de63b2dec5f56738c9d7eaa283a4a489545ef4902
        - registry.redhat.io/rhel8/postgresql-13@sha256:3903d562425dd564986123280dc076548fb3d24c74138dcd857652f53052e0a7
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:2e75096ffe20d1261e8a645fefd32e37295d727aa6a280f10041a44443917469
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:57fb86ba112d79a228cbacc96cd0525c123464bcac956354722e3f2aeff42508
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:ffa8c6d8fb5bacb59da27f4ebde87385359b3f97a6a023b4a4488ced2d46c574
        version: 1.14.6
      entries:
      - name: openshift-pipelines-operator-rh.v1.14.6
        version: 1.14.6
      - name: openshift-pipelines-operator-rh.v1.14.5
        version: 1.14.5
      - name: openshift-pipelines-operator-rh.v1.14.4
        version: 1.14.4
      - name: openshift-pipelines-operator-rh.v1.14.3
        version: 1.14.3
      - name: openshift-pipelines-operator-rh.v1.14.2
        version: 1.14.2
      - name: openshift-pipelines-operator-rh.v1.14.1
        version: 1.14.1
      - name: openshift-pipelines-operator-rh.v1.13.0
        version: 1.13.0
      - name: openshift-pipelines-operator-rh.v1.12.2
        version: 1.12.2
      - name: openshift-pipelines-operator-rh.v1.12.1
        version: 1.12.1
      - name: openshift-pipelines-operator-rh.v1.12.0
        version: 1.12.0
      - name: openshift-pipelines-operator-rh.v1.11.1
        version: 1.11.1
      name: pipelines-1.14
    - currentCSV: openshift-pipelines-operator-rh.v1.15.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "db_user": "result",
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.15.3'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.59.6
          - Tekton Triggers: v0.27.0
          - Pipelines as Code: v0.27.2
          - Tekton Chains: v0.20.1
          - Tekton Hub (tech-preview): v1.17.1
          - Tekton Results (tech-preview): v0.10.0
          - Manual Approval Gate (tech-preview): v0.2.2

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.15/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:322ef2385620661a7914f47c73a6d145ec0c9ce1ccde30645bb66148545ddc96
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:49c346bf7fc69a9796690236dc8597e35a874343b02fb74c5d2f1c05d356e41c
        - registry.redhat.io/rhel8/skopeo@sha256:29465586e92ed04d2878dcdeda2d508c7e8e7539762e10296b52d631929960aa
        - registry.redhat.io/rhel8/buildah@sha256:aac6629389db17e99894c5bee0da01d4c8065d11d8c6f6e1602f9484290baa70
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:bf94c350923da706c6e0a427e2d91af3e13e751a12388a4ed75d1c34dd7b3177
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:ac1a6e057a5599fda7e048ff77030e98c92b5d51cac2472c1ef3910964fc17e6
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:61c22119f37e620a3807056cde35bc07c90867765b149294056326b1d11494f1
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:abac279a0a5ecde57952b6bef634f8ba7527cd8c8956a2ea30e9449174ad52bd
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:166789a990df9f6ab04fd3df0a8cd301bc19b9bedde94215cb27306060eb1e2c
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:83eb08bb9579cc19132af282c455bca32a0fd4f5436a35e7ad23b5efa2faa3cf
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:479e3eed93e0aa13e3489514fe00375a519eb45969d099c490b4acd056ea5adf
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:d1f86458079449c86407ac716c3cbf608cad06de29edeecbde6586f76628daa9
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:e749db81c9d7e11e2802833a2a02e93afc357c83d1a2a827e537262df281a4f7
        - registry.redhat.io/ubi8/ubi-minimal@sha256:f729a7f5685ea823e87ffd68aff988f2b8ff8d52126ade4e6de7c68088f28ebd
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:7c81ae208a16e17d0102074867add37dfb1207ee95cbbc6c732d88e61bbc59fb
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:7152c3c174d25f2a5239456ba677ac5315bfb82bc3b3ac5ee0eb8f9f74fb6632
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:aca0024cac7a8c41b1ed151eb1e1188f09bc636866f4e03f1b973f90890b37cf
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:bf6cf2e87fb19f7aa9a490b83c16af69834c0721220a643710a1b077959e91ca
        - registry.redhat.io/rhel8/postgresql-13@sha256:a92a579f1aef66ac188d24fd489c456a1a3e311d95dcce652da6b81e28fbf725
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:7f00ff1b407d9ea107d3d969988a6279b28bde86943f490a007fa922dd40a1c3
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:eed464a3b4e23f0b2623aa1b923c10232d235f4272f562955926c1748e295d2b
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:7958925cecce1b4b02ac4649e3bfd7fa3b5361c16ab3593700b8da93fdf1c79a
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:312d1f90765d05a86845856c677e1ece4c57e81a9bfe3fd1e76868d624e3d72d
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:1300d171dbf09c63071ba5586e223491a13acfc6dbab27eff04648aeb788d455
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:538c63c8e7af3cba91526f2f9ac99fb8f6b0f4781db31771c959612c209dd090
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:601fcac1abc6c8efa1248c5ca229213d6660708470a1c416d41c6aff0ba4691a
        - registry.redhat.io/ubi8/openjdk-17@sha256:e8cc2e476282b75d89c73057bfa713db22d72bdb2808d62d981a84c33beb2575
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:e7b000a2735bbf98ee79bd093f7b703ed09d7c1e11ed2a5b22c4b8ac8d440386
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:827313c2691fb8833e42418b65b4a3360a11cabc2c92a7d8e7b20183f06cbf68
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:f02af8475c33f6f38a2f90ff1be1fb1b2094860c4d5e300aefdd7eb84b4a529a
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:acf8017029f4b2e7a1c3da9bc62c65239d313d24e3012712356a191bea9a2e6f
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:75d6228eb656b8778fb3bb8dc68c686ac1c9b787e5343ae3c2467f8e724af83e
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:1db14f23077785108660dde38c98e2e30c8301f8e2e1c594dd4636e0ba29f84f
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:70bb0ebcfe5da45e5c239aa011dc5f1ef48844116d7346bb05433b503107268a
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:f04ea12675e0eb2354635cd2f9e99d6e59619d1c611e0ea95c3ce8222d615322
        version: 1.15.3
      entries:
      - name: openshift-pipelines-operator-rh.v1.15.3
        version: 1.15.3
      - name: openshift-pipelines-operator-rh.v1.15.2
        version: 1.15.2
      - name: openshift-pipelines-operator-rh.v1.15.1
        version: 1.15.1
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      - name: openshift-pipelines-operator-rh.v1.14.1
        version: 1.14.1
      - name: openshift-pipelines-operator-rh.v1.13.0
        version: 1.13.0
      - name: openshift-pipelines-operator-rh.v1.12.2
        version: 1.12.2
      - name: openshift-pipelines-operator-rh.v1.12.1
        version: 1.12.1
      - name: openshift-pipelines-operator-rh.v1.12.0
        version: 1.12.0
      - name: openshift-pipelines-operator-rh.v1.11.1
        version: 1.11.1
      name: pipelines-1.15
    - currentCSV: openshift-pipelines-operator-rh.v1.16.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.16.4'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.62.9
          - Tekton Triggers: v0.29.2
          - Pipelines as Code: v0.28.2
          - Tekton Chains: v0.22.3
          - Tekton Hub (tech-preview): v1.18.2
          - Tekton Results (tech-preview): v0.12.3
          - Manual Approval Gate (tech-preview): v0.3.1

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.16/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:81a8b903d0fc4b81e6a23fac862d18da4cc18b14cca8213665df37dfaabbc2e9
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:5672f3a375c6a274200b423eb2643c818aba75488dec0b210fec2532814ba51e
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:47f094ccd12db7a43c8a91c1a3d2a190bc2b83f1745e7eb8ce097d0c8e1e2c88
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:80d46bc5ce8c931af22a97ccbce05b59740e313167c88ad910febcc2f2dc0320
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:90e2a7ea498da756e09c7e4d19c7fb91d429cc28c57f97c6958958e3466cb58c
        - registry.redhat.io/rhel8/skopeo@sha256:1dcb4189c92da21ee58b3ca5433615e0208194c8412f7f1fb45fa491179aaf9b
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:d2bb3f9f79ef719d8fdafef3a639e3eb8bc199ee56fcfb33e85975d536d2a6ac
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:7edbd11e5a449408d673cae1fc2b9b5bab492ecff2ac6e640c8cd69a9b90ff08
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:a5a7b2be50f211d4d9d555b2a2e2302096c139ff5011d984cc386f1530007f16
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:f9122d71c3862cd2afafc8eb7f1a39c83eee306e5baf38514e6e39215c70a72b
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel8@sha256:8cb87920e6fc8185ead49aafb7799fead6a4a510eafdae0ee29e85248dae3945
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:aacdabe243dc573f3161a3d04fbf014b77d9962207305d70fcafce475f7eea93
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:92042dcacd4ab0f2a78aab03fd91a9a39d7d7366b5398fd797ecb145acf5fce4
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:2fea7d8e29d70c5893ddb6caf2bf1d4144787f5f0de8739351d2097ed09ee751
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/rhel8/buildah@sha256:5c7cd7c9a3d49e8905fc98693f6da605aeafae36bde5622dc78e12f31db3cd59
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:474658c262b539dd1c83ac22bb090b3fdac5de9cffc06587d36dbc6fa4ce208c
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:e1fe4a86211cf22e5ad2b23c989d24b69f597db4c3604e84777dbd0c585ec186
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:6acc31f179b08658976cddcb9ef42f377edb29933d1d83a2ecebfd455bb68571
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:f3e6df07fd86acb91037d55a1970e5f8c6ab12a7517752692c224c7687022812
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:8129e62906c68f04075a44c939375a7bafdeaeac673a20a4fa74ffd6285187ac
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:7d92dca9da27c066cb530172e1811493d8c916e67d8068f5ddc471fa3f3f7856
        - registry.redhat.io/ubi8/ubi-minimal@sha256:a47c89f02b39a98290f88204ed3d162845db0a0c464b319c2596cfd1e94b444e
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:b94bd566ae3d9ce5f47273d867164486826b19fca0ce11d1a7d18312413e87ba
        - registry.redhat.io/ubi8/openjdk-17@sha256:632e78989471859ca4ed1148b951a911f7e3b6e6516482be20967c4171612c55
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel8@sha256:46635d636ea2e9a6ba0760bf8c528028700ee40b0ba8466797034371185b16d0
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:856229e9e37683c926e3449658d76f93d910384eff62c2ca192768a653b71533
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:c6d654c71aba35c3c7f5c1a2ca5b8401e95adf6f649c599a9aff25177c2b3323
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:549d74a18ec52db91fa949cac9b8e3d943d478d5dec9850d2ba0cb1bb526f578
        - registry.redhat.io/rhel8/postgresql-13@sha256:a308263f5d1071be8ac5c13fd84b40609988ca30ed56c22a8bc6769e107ce09b
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:b41b888fc1a3b7215e6b71ee97bf64a1816914a8f33e96ad3930a4e2d5b5ffa7
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:4e0e4c358faa9cafc3964d6d40ee4aa73f8dfe8b6338dfb8206613d7c5241842
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:dcab4a3d519975340ade59a0ca0222104e68acd052a61f2aeb8c736be4b672cb
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:e84c2d80e92815c6bf384e1504a194dec9b796a3aceac2eed555fff241249571
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:d36dbb60c7b5caca7f3be08705f8ab91d7f01abaf8bfee4aee73bdbbb3e7e71a
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:7da7f929d2711fc55dda01fe51a474bfa7adf5aaf78be6d65a6d65850958d6e7
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:19b3dd69c1d7e17890b7163aef322c1efbc76fed7f2b1faf26aaa05fd8161184
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:aeeed543412cde0bd3a9baec965cb99d80f2b6acb77ae6356ed51b4b1262d006
        version: 1.16.4
      entries:
      - name: openshift-pipelines-operator-rh.v1.16.4
        version: 1.16.4
      - name: openshift-pipelines-operator-rh.v1.16.3
        version: 1.16.3
      - name: openshift-pipelines-operator-rh.v1.16.2
        version: 1.16.2
      - name: openshift-pipelines-operator-rh.v1.16.1
        version: 1.16.1
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      - name: openshift-pipelines-operator-rh.v1.14.1
        version: 1.14.1
      - name: openshift-pipelines-operator-rh.v1.13.0
        version: 1.13.0
      - name: openshift-pipelines-operator-rh.v1.12.2
        version: 1.12.2
      - name: openshift-pipelines-operator-rh.v1.12.1
        version: 1.12.1
      - name: openshift-pipelines-operator-rh.v1.12.0
        version: 1.12.0
      - name: openshift-pipelines-operator-rh.v1.11.1
        version: 1.11.1
      name: pipelines-1.16
    - currentCSV: openshift-pipelines-operator-rh.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.16.0 <1.17.2'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.65.7
          - Tekton Triggers: v0.30.1
          - Pipelines as Code: v0.29.1
          - Tekton Chains: v0.23.1
          - Tekton Hub (tech-preview): v1.19.2
          - Tekton Results (tech-preview): v0.13.4
          - Manual Approval Gate (tech-preview): v0.4.1

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.17/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:a14270967460883f283ec1ce17ff8b82db1e0777dda4f6d690f15ac38dcbad85
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:650391ffd3e4c7b173bf9d718405b2375c6ff2027abc0b7ce4f67e79e7ef4b81
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:37109f6de75dda750496f6668164d0dceaf9db5278e8c034708436f39de3276e
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:a392d2c85baedd7428c579e91d9f8acf8232399b90523b1093dc3447aaf56a16
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:8e52b4b31b0d9ec772abe6e6ef9c6698e89893395f364968216bebc3367da4b2
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:4b87782b35e7284a365c49f1c094aae98531ac630db1958294009c041a2d8440
        - registry.redhat.io/ubi8/openjdk-17@sha256:e5fc9e34382da41b10d4ea566639d31707493abff9e8de81f5accfd66876409e
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:4224040b00cfd76e85f80aecc5095098564cc62e9c477fe2f7a428b30df7d4d9
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:41ec6db921812fd1e80b05d51c224ddbce63bf0c57f02ccf07461d2202e97b90
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:4fabae1312c1aaf8a57bd2de63bd040956faa0c728453f2a4b4002705fba0f0c
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:7abc88bea01dd6513f6e48334f0b98501f909a1dc1ab9e83efeb3c1a834f8f92
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:eab798472b76bd51235198fd3b95189740f8d650ebc6bb01fe653c025c1defaa
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:61c9e8b1660d8935ee316ecc113e8141daf05149d172685cabd861ae8aeed3eb
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:4240501c076605a01d969328bbb8ad885c598b44ec1dc1c066312491a0eb6715
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:7c710882752f12be3c22bedc8d32096c111e42c5b9b392163f61cf8d72be3ec0
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:66309ae437c4647f48efe3d6f2b696df83859323700427d0025b94795751a859
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:41e2f2c7397a8b9e093e18977666e93943734f51780c769c99958ae1d20af53e
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:4a96f1bb7e3a5c7ca62dff2199fa15ef964b129e4e36e9619837c1221538f496
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:3696cf724ce7865626ae408d82fd87727c741e4d3eb3475c271ec6d56f0882a4
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:44da76d75f567c2ba623bae646e1d552d2c0fded1accdde3aab4f4bcdf3adc28
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:26f7b695fc1430359ab2bfd2915cf0b7e88b06d8695e26761bdc2ce8935f0fb3
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:0b6ee6cd98cb7e835cfc9544996c6f85288d7252071c100bed853eb593d6befb
        - registry.redhat.io/rhel8/buildah@sha256:6d2dcb651ba680cf4ec74331f8349dec43d071d420625a1703370acc8d984e9e
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:d851a2dd8ec95bfd7ac6ec5515c8cdadfcbfc6e6ec84b40630b4fe595f978fa5
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:429bc3d6044ca7dab780ff4b5eb91d42d38a29eb7ef428a25c53b7dda49871bf
        - registry.redhat.io/rhel8/postgresql-13@sha256:757d8dd6f243717d820846da7144a49913f28f223b74ec305b2215f54720f8a0
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel8@sha256:b5cceead6b8e9c99655c9ba810679b9d6fb978e5ff18d020a46d475c4c26fa29
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:c92ebe828b6ced6c93a3df5f4ccdfba022471729ad08b184de2f7410ecf592cc
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:df11cdbab2a1c6b9b800ef124c73f2fb5e3981402b21bc822ccc13e95836197a
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:db27cb4f0c350d99546b0f1bdf02d1ba271a8053a4552d88ff302adca60a3e29
        - registry.redhat.io/rhel8/skopeo@sha256:7fd65200b4847fc84fba0466519b87f915e28c0e9bc90d3022beecfebffd4af8
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:99f5383d8c6bc046a0bf48fd6c2064a5d055645085248327305f3ee4ae71979c
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:e673c20fff2fc684f5098282dd6ac6cb265045945249e995b81289532ce8c856
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:9668d852123b160e49511242371edc6e941a8e8b8c2ffa60080d91cc6b97d3d7
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:84eed8ba1189448e02ffd284c7d913e7b3cb7efe7a8e5b0edf890f7eb17fc439
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:922b9e86a674a6f3eb82f510569ab9cde97a1ad833c263335cd4164bf4164b66
        - registry.redhat.io/ubi8/ubi-minimal@sha256:c12e67af6a7e15113d76bc72f10bef2045c026c71ec8b7124c8a075458188a83
        version: 1.17.2
      entries:
      - name: openshift-pipelines-operator-rh.v1.17.2
        version: 1.17.2
      - name: openshift-pipelines-operator-rh.v1.17.1
        version: 1.17.1
      - name: openshift-pipelines-operator-rh.v1.17.0
        version: 1.17.0
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      - name: openshift-pipelines-operator-rh.v1.14.1
        version: 1.14.1
      - name: openshift-pipelines-operator-rh.v1.13.0
        version: 1.13.0
      - name: openshift-pipelines-operator-rh.v1.12.2
        version: 1.12.2
      - name: openshift-pipelines-operator-rh.v1.12.1
        version: 1.12.1
      - name: openshift-pipelines-operator-rh.v1.12.0
        version: 1.12.0
      - name: openshift-pipelines-operator-rh.v1.11.1
        version: 1.11.1
      name: pipelines-1.17
    - currentCSV: openshift-pipelines-operator-rh.v1.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.17.0 <1.18.1'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.68.1
          - Tekton Triggers: v0.31.0
          - Pipelines as Code: v0.33.2
          - Tekton Chains: v0.24.0
          - Tekton Hub (tech-preview): v1.20.1
          - Tekton Results: v0.14.0
          - Manual Approval Gate (tech-preview): v0.5.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.18/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi9/openjdk-17@sha256:5d619e202695acf76a5f32c07ace6709f5ee5d3999bf977e1bf95ad34197f44a
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:86ea94b12e8923035e3e4f39b6a53b3ab32c5f77d7eea170aa9b34aecc6f8577
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:0a60ce46c05256c74f770901e0bb79ad89e88ba882b643c1ff7b2a1738463acc
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:3964c3248dfb1006ea9a46295baf7f8cfa71952d6b9863b52f87e8f3d7863b1e
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:36a4c6a3273f2371ab5e679f1955221c286ea753e29e6ab8da7c983681a68f02
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:b550a8bc2a87d07a5b22a3ec3fb83efd8eb26ca6e80d404f2e453cc9b8434da7
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:cd28c2c90dd340324b0d8709640cd7636f28286114b0023649b6e21be13e03c2
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:fa5c456fc91319c55fac9a011dc7bb32ef85db293c73d74436e903b76ceb4349
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:409c14649d8e7007d4b72b6feffb71dabd26322bf2a7c77f5307f2546fdfe81b
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:96f5e1922fb9d6dc17e271576dda9bd032bd6dcc04e7496fb15a2f6b5871cd67
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:16d799503674641d4690822edc9c399c61429ccd31621d295eee4a4d1ef1c076
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:cbd89c531953a43b8055bda72f7696be80f57fd1d782567a91d82e0fbcb3d26d
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:8fec8e7a8b3e5182c67104de4e3abe32593b682c4a8986d1e9ab835dafb95ff4
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:3696cf724ce7865626ae408d82fd87727c741e4d3eb3475c271ec6d56f0882a4
        - registry.redhat.io/rhel9/skopeo@sha256:d077e35100d6d7b3795f8e3594293dbf8b65c6bc18b909aa555fe56f048b356a
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:1b57b64657eca0002145100bae262a19f02eb40068e07994df467e28a1618f90
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:20e19c02efdecf94a7b16483047e5db81c50dc2a0aa59d4895067e6c7769efcc
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e2b4c3b505c4b5aa36f8a5e661389b0de4e799c74c2c7f16a065e5522b438e89
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:80d441ec5d07c5059277a3a9173a35dfc0b8372d6e4d7823514131a94bae3fd7
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:d72842a7c6a1e50a68576c93f60047afec801cf83db64d7d3e34b25bf3c753ec
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:e6a7738cdb0e8aa9e1af381273b1c18ec1414b08a66fc851f6e11e5aa7a6706f
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:fcd5fd10de7334bd59a8da4997f91008ced95c6cde40d56d50f3b3e32c0f19e1
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:c20ab6fa719506bca942669ea84b113335b05df3ed275257b88be234cade7c8d
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:2ba6782af0d7e5c4d623de31814f43f4f7c3c38d09b57705021957d334067f87
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:d055d2b35a663aef3e1aafdbed0b12957867c0670c946ebae66e9c44a768bda2
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:72901cf4e091fe5f6a577144fdffaea8570712e9887a8bd715d5ae5bc346ff1d
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:b9675b22bc8759903472c1eb0e03f90c75460f02799d894294c161641cad9414
        - registry.redhat.io/rhel9/postgresql-13@sha256:ecd5a5583db922c517c735797db3fca587c0ef0b410e7dd494ace2267d89cf05
        - registry.redhat.io/rhel9/buildah@sha256:6886997da8c92214e99e3e8da47a0f08239172d2544b4edae058d5cb018effab
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:c90a9c3a20f088241980b76986ba4f43b0243fd6d94a826fa7b5888e01c13093
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:776d34b0ea2e35c11ad4050e6b40cf4b40c84dbe0d5f3e6ad669c494a210ea8e
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:8b4bf043e0fb237553503d63b129c432600bb4e65b2096b246b257b404b42be9
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:eaad37453a5ebef51c28d4a068f56e9f394e710c72bb86342cd2189001420836
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:8ae0ca6411e4fcd5c0605344bc6930cf122ffd80e71199509d58423a8b41de1d
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:dbc462cb50385dec3a220b5d0caed0e0211c166182026f86e3c51a1351ba0de4
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:6abc903c784bb455e1f2fde45a48afe775239668e1d0dbc1b3cc6afca3c0b477
        - registry.redhat.io/ubi9/ubi-minimal@sha256:ac61c96b93894b9169221e87718733354dd3765dd4a62b275893c7ff0d876869
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:e27a30d471dda4983a5b8192709d283a9d082558f3caf31b30db2214e418b017
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:e63d956a0534f2e726301e00e877c4fe2f34ca4559e59d2df7567c04ded311a3
        version: 1.18.1
      entries:
      - name: openshift-pipelines-operator-rh.v1.18.1
        version: 1.18.1
      - name: openshift-pipelines-operator-rh.v1.18.0
        version: 1.18.0
      name: pipelines-1.18
    - currentCSV: openshift-pipelines-operator-rh.v1.19.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.18.0 <1.19.3'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.0.0
          - Tekton Triggers: v0.32.0
          - Pipelines as Code: v0.35.3
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.21.0
          - Tekton Results (tech-preview): v0.15.3
          - Manual Approval Gate (tech-preview): v0.6.0
          - Tekton Pruner (tech-preview): v0.1.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.19/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:22cfadc42e793553182acf05a34fb6a7ae0d7055431c2e393699ecfeb171b173
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:7a0c6794e9284209dc131f3087ffc01fd5946e20a9261422808c6dd3938407de
        - registry.redhat.io/rhel9/skopeo@sha256:993f799d58c48151ecae7a84e63860f6c0039e8e18efed5ad67e7752f4b7e408
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:fd6f4cd7801d7c20cf5a6e546022136404212da5d9997750e643f6e640b00e3d
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:2c3074ecc0c74d508020bf1e11667d868150915b55d436a43c2b4fc22815bf48
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:80af5fa2d87e7c405f16575519293ca7fccc49d635b4f3d0d334dea4fbd28721
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:25bcdb30f0fe34300c4f61f62513f4853f65fe11bafde58bcf19b7b3c3fb9af7
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:251c87e2c48e7f377bcb399ed64d50dff11de3459629a73a63bac17c353eb16b
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:363e28a82ed022e507c3976d3a42ee63c053eb7871712c213b12235d8234b7ef
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:b247e9f3a2db32b575999387ef4f91d685a67c1f3950a6e265e1fe14c1fac196
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:6aaabb9af37e46a856371e29f141de6b0fac623128099bb927fc30f8f61e9db4
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:d000daec36a6cdbe41a85ee4eac24a712f7628441a412f751baf8d7778626f62
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:8bcaeaa751c61eb927d6353984f81d6a92b4bd4f9a5be9dc620a11b06ff9e6ba
        - registry.redhat.io/rhel9/buildah@sha256:a6e04c061b8d261dba85d01dcb64d3f7dba8c0b9765e03f05ec79850a248d75a
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:41d669fbe8bb48481890fb9fb52de9d4cac98cb74e6a6b284d046a8d21b06e58
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:28258903330a0715141e4685b3760ff18b0d75da0694985bba7c3345d7808b5f
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:90a5eb2c0decb72234cbea5354bb7737cebc53c467229b2c16a1425b4bf408dc
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:3f7e1d83eb1e21860cc2ae0bfe3abe8c9058e92a69698dba9d5871069016e90d
        - registry.redhat.io/ubi9/openjdk-17@sha256:cdbd64ba1b7826d09bc180fa19aa6fbc09ddb5bff5e02b7e1351353bbabebb78
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:a90a6c269b3fc4c77652d4862878211eee5eca1d8047e561d07e3ef1292e52f5
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:31a40f8095a3f8cd04750810d964572339965ae0b2d357c57c0b671a95e1cdbe
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:1f870e70669ad584db1ee31b1dcd822d8b8d0cc03028c7cb09238fb7abf19a42
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:b8cdf2a1873a3df1fb8b8f6d58d06235cef28d228abd70d0bb83cdcfbe3bc230
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:3b1c6f8acdf17a5185413147e0248baa559378ed79e08cb098c724036020576e
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:c9fbb018a70df9a542236ace556fa3343ae350f43c0629f13e525854508f73f9
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:92b8c92eb0bec102215b97f1d60d76c7a57747aeec9594ce1e0e46f7c27391d6
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:5f39f9e61ddbce92629e6502bbef737365c8a1bb2dc2ecaf9f9d8092dea844c9
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:aa1b6f374643ed4e89ba194afde8a5138e401c8cdeb45a6bb5df5b608f58358b
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:22095c436562ff92771d9349f57d007d82415dae519e3da1de7d51e3f1042a93
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:198cf9e5c9641757a0453dec2747a24b018a9b85e10eb0dfdc7a2e75ed09cf6d
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:ba596c629fe8b5b70db97923ff11392dc12115c8e7afefee0d48c12440b8dd86
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:10c620aa3a9c4bfbca185fb460e2521473a5fe6d518342b89a3dada5cf14e65a
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:13608cfe2b87535776c65728aa40cfaa84ccf63afe64d1f2e799ea0e26819239
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:68d9007c5abb2ad3b7666249a61d9428c2cf07165bd8090a6e6def0b7cf70fa3
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:37e2f1fb8a2b4c7ac3d5390ba68a9bf98a0c83a38567b6a9778764061b3d2bab
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:c3ba6b9652205407cc4d4d47b4a928fe8fa86ce4427e6b44ee1d53faf3455f21
        - registry.redhat.io/ubi9/ubi-minimal@sha256:14f14e03d68f7fd5f2b18a13478b6b127c341b346c86b6e0b886ed2b7573b8e0
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:d21cc7e094aa46ba7f6ea717a3d7927da489024a46a6c1224c0b3c5834dcb7a6
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:2d2499b7b12aace07c4b5ab4121caad445aab5bc8dd73377171683b36d417d73
        version: 1.19.3
      entries:
      - name: openshift-pipelines-operator-rh.v1.19.3
        version: 1.19.3
      - name: openshift-pipelines-operator-rh.v1.19.2
        version: 1.19.2
      - name: openshift-pipelines-operator-rh.v1.19.1
        version: 1.19.1
      - name: openshift-pipelines-operator-rh.v1.19.0
        version: 1.19.0
      name: pipelines-1.19
    - currentCSV: openshift-pipelines-operator-rh.v1.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    },
                    {
                      "name": "legacyPipelineRbac",
                      "value": "true"
                    },
                    {
                      "name": "createCABundleConfigMaps",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.19.0 <1.20.0'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.3.2
          - Tekton Triggers: v0.33.0
          - Pipelines as Code: v0.37.0
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.22.3
          - Tekton Results (tech-preview): v0.16.0
          - Manual Approval Gate (tech-preview): v0.7.0
          - Tekton Pruner (tech-preview): v0.2.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.20/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:bc403f93ef6c5526fa30152eab384c4a25bb0bc6ab4158a1e24f637b87a57259
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:980ed7e81ea55522e1681dc1e9b170b68d16f07061dd1f1ffd142367d59552f1
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:bc10e7d6122ede5cf4f2882c200447683255465a1080e826761e90bfb983506e
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:c3013c347bca8a520443a44ac2817fbba5744a9e91069ac5f09e91447ebf6b28
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:1b6aecfce45ed3163227b845b212e83385cdbe999edfc2e9387a0c76f3f8d8e9
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:5683321a5376e63c3ff3932270a1850d5e261f76212020ce867334166f90d2f7
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/ubi9/ubi-minimal@sha256:2f06ae0e6d3d9c4f610d32c480338eef474867f435d8d28625f2985e8acde6e8
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:7eafffa6e4e04f10c1d4e84e02d51c00eb8649c60f8f9a13da14492f24634891
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:d56d6b46331b4569077ab1e7a7b609ba01f0a1dd62b695d62f8aa0f32ea1b621
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:e63298d07ccec5804dfab4e42bcc8163c29cb3ac371c91248777423a4edd31ce
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:8f0494341244a75546a4b80b065702be04e6d200421aa4e7ec90385c9dabc6ab
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:c31af237afb84214bf2a4c06a7636ca071fe97203b41da90d7878109ccd64784
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:115eea1a5ca87f3b6b72db1df05b7386bf73dc70c9c2e4dc5e1a66700d960c37
        - registry.redhat.io/rhel9/skopeo@sha256:a65a413f8a2864389a09dc750690d97afbcdc5c70821e0f85e99e8adba7954e0
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:dd941acd357ae2c0fd7e15fb8d32d7ee6b2847435101dd1e1f107db48af18bf3
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:5021158a4752a26f8b33adea7da66695b22d2ad155d346d35dde5f971bc1a55b
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:823e018e0329d764ac694f177ce88af9074f364c29a1c17929f9dd0b596d03dd
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:14dab8d87323fc7d112213f7f4db3ee7cf3841747de669ce710ae6a794053457
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:bbd1918660be7b3c78777dcccc87d3b26bf5adfc077844ceffbbdd53696dc0a3
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:a7163a9e1fe891de021b1003558e6e64fe17ed17ab0820ba16bba09aa112401e
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e7998e113f75e8f29602bae210c8b66ed06634b699f1d99c5875ef879220d733
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:ec74a562c18fefe36ec7ae3ba083e4eb76c56b562f916578900f8767ae98e6c2
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:c4d05e546c466c57379fff502816012f615be073eb8cdce54677b4ab206a3f5c
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:c0f438e6095aa81f3f7d820614e0414cc43374014c1505d7641bc1dba69a3903
        - registry.redhat.io/rhel9/buildah@sha256:18453ab0a62154283aaf4c45efd7543d1c8ecca1aebae46dedc7ddc0f410b232
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:818bfa4a3a874974b5f7f8bcf2faeb6c05e1795381ee1ceacf89f17b1ec0ef7f
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:9415e889ef98df6d2b2dca1e5d798fb765eeea0d9828e48424374d5267a6b0d1
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:a712bb411d2433b01793dff25785c1e265fa865a52f7c8732c9fb4418ab61757
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:3cf990fb2fc422f3527a7c8b43d06ccd4d32bb5487ec8cb2d4a97f0f5d13f170
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:458f865f1b246c5f0682f5956795771a0515b89edfcffbbc0da1a3d3f532645e
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:96db721939b22572d7d8d8b95dcb5c92fdedb98e9cd0aabb7d17b703a915b07f
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:036250bcdc6fb95273bf07948dfdd4dce765ab84c803ceca0d0b4abf6ca19dec
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:161df929e2bd4c5acff594be31d5458365de5ae0d2189bb2f647c3fc6657fa28
        - registry.redhat.io/ubi9/openjdk-17@sha256:555c8bcd2596637befd1a6da4aa949c5a2640cf66e09c2dfb932afe37e39095e
        - registry.redhat.io/openshift-pipelines/pipelines-opc-rhel9@sha256:c1bee4a2bb495de6d4da962b6e7b2447caa51cc36c3005e4083bd31420ef5102
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:901b5108ceb669c985e75f5fb033c496e9f028863df99ab32651d1b08433a352
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:df15eb1f1f5f237bfe481c16ab3d40bf7fb4123f3c930bcfb121e2cc8dc42882
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:441e1c7124c1674e3d4c38c3e82a4fe5568e76d13c89c04bf009d6b97435de09
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:e3e64abef53151f6b15f15bcbff85668167a571e98badbd8a2ca252ff3a5b8dc
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:2d496752b19f765f1ea36ec94b873ebd8fa604764a56aa67d3ed94b33e5fe75c
        version: 1.20.0
      entries:
      - name: openshift-pipelines-operator-rh.v1.20.0
        version: 1.20.0
      name: pipelines-1.20
    defaultChannel: latest
    packageName: openshift-pipelines-operator-rh
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: cpd-platform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cpd-platform-operator.v6.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cpd.ibm.com/v1",
                "kind": "Ibmcpd",
                "metadata": {
                  "name": "ibmcpd-sample"
                },
                "spec": {
                  "license": {
                    "accept": false,
                    "license": "EE"
                  },
                  "storageClass": "managed-nfs-storage"
                }
              }
            ]
          capabilities: Seamless Upgrades
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.0 <6.1.3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ibmcpd
            name: ibmcpds.cpd.ibm.com
            version: v1
          required:
          - kind: CommonService
            name: commonservices.operator.ibm.com
            version: v3
        description: "IBM Cloud Pak for Data is a cloud-native solution that enables
          you to put your data to work quickly and efficiently. \nCloud Pak for Data
          includes a suite of services that support you on your journey to AI.\n#
          Introduction\nWith Cloud Pak for Data, your users can collaborate from a
          single, unified interface that supports a large number of \nservices that
          are designed to work together. You can use the Cloud Pak for Data Operator
          to quickly provision services for \nimmediate use. These are just a few
          of the services that are available in Cloud Pak for Data:     \n- Watson
          Studio\n- Watson Machine Learning\n- Watson Knowledge Catalog\n- Watson
          OpenScale\n- Watson Machine Learning Accelerator\n- Data Virtualization\n-
          Db2 Data Management Console\n- Db2 Warehouse\n- Cognos® Dashboards\n- Analytics
          Engine Powered by Apache Spark\n- IBM Match 360 with Watson\n- Data Refinery\n-
          Db2® Big SQL\n- Decision Optimization\n- SPSS® Modeler\n- Db2 Data Gate\n\n#
          Details\n\n## Prerequisites\n- Red Hat OpenShift Container Platform 4.6
          or newer installed on x86_64\n- A user with cluster administrator role\n-
          No storage is required to install the operator, but you need storage and
          resources to install Cloud Pak for Data services. For more information,
          see [Storage considerations](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=planning-storage-considerations).\n-
          See [Pre-installation](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-pre-installation-tasks)\n-
          To install the Cloud Pak for Data control plane, review the [system\nrequirements
          for IBM Cloud Pak for Data](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=planning-system-requirements).\n-
          Before you install Cloud Pak for Data services, review the [system requirements\nfor
          services](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=requirements-system-services).\n-
          For air-gapped installations, see [Installing the Cloud Pak for Data Operator
          on air-gapped clusters](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=tasks-mirroring-images-your-container-registry).\n##
          Resources Required\n- The operator requires 100m CPU and 256Mi memory.\n-
          For the minimum requirements for services, see the [system requirements
          for services](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=requirements-system-services).\n##
          Installing\n- [Installing Cloud Pak for Data Control Plane] (https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-cloud-pak-data)\n-
          [Installing Cloud Pak for Data Service Subscriptions](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=tasks-creating-operator-subscriptions)\n-
          [Installing Cloud Pak for Data Services](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=integrations-services)\n##
          Configuration\n See [Cloud Pak for Data Post Installation](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-post-installation-tasks)\n##
          Limitations\n- `x86-64`, `PPCLE`, and `x390` platforms are supported by
          this operator.\n- See [limitations](https://www.ibm.com/docs/en/search/limitations?scope=SSQNUZ_4.0)
          in Cloud Pak for Data services\n## SecurityContextConstraints Requirements\nThe
          `restricted` security context constraint (SCC) is used for the operator.\n##
          Documentation\nSee the [Cloud Pak for Data documentation](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest)
          and [license information](https://ibm.biz/BdfVHT).\n"
        displayName: Cloud Pak for Data Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cloud Pak
        - Data
        - Analytics
        - IBM
        - AI
        - Machine Learning
        - Cloud Pak for Data
        - Auto AI
        - Db2
        - Watson
        - Data Governance
        links:
        - name: Cloud Pak for Data Operator
          url: https://www.ibm.com/products/cloud-pak-for-data/deployment-model-options
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:bbb457d4f6b7bab065e909756c2dae31af7162dce10bfb92f274aedbc98d0762
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:716b5823661ac8f1eacda11ffbe5f0ff08a0e9d2296d04acffb646b7526128a9
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:1f82599d91ab20d2024abe134d2f56caca1f9bca1f02d64352c81b5f1e0f1730
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:d6575de3156c7a9b0553887abaf8713a758d9c9916f9cdcef611923c50632867
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:113fdfb9395c96f7eb65f9a20e18ede8a2eae4aa6c91660be8fd3e517656e156
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:d0e0144eaf27e03aec1499e53cae5366e11a3da284536cd3b27ba5691b4f70d0
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:6a996123e69c7f2c3254f334e278ee0daf75fc53980183e1a72358323e41349c
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:1d6dc7a3039e6fa044053d4e89ef54ba2e83ceadd57bcc4613a6d447d027ae2f
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:7861185d55a5fd18d896dbc57501ae5cbc1de0bc02f1ae3fe558a36f0f80b4fc
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:46bbfdc1f5e418db95ba806228fa5696be22c4cf9eefc185ce27bfa68ec1d247
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:a95e42a4f9138f1f97937a74982c98d466c171d0e3e79c522ededba6c25b39a3
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:23ba1b277cf76bc28f952ce6f5bb02d6177f76eca8a8b4ff7925ef119a994a85
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:670a36f3400164cc9eec5d23422d184340e0c1edfb13a634e58f0345c325c21b
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:0dc719e58694fb501b8b36f3a375b3c0462624d27fe49afe6b835821e31393f3
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:257bf301338f7b123099e2ffe31c3f3b793bf28d6a0570a25b889271239579ba
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:d61b7b8e6a9785b59630d9c5776f72a29ffbf2e81eb753d7b11dcf2d9d77f1f0
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:30a5b2c4e3c231373fa18d8aaab16d3505cc11a8a5d4be36eaf99d7c0dee35f2
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:fbc6eab07fcfc2d0c1895f18e17f751e9481ad111875959e9ff08e14f81b9a35
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:aadebf0598c492639e7c1109ce681256731215b686fec808578bb09558a40315
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:6cd1825ba40071bca7f46c1af3e813ae7ea44a393dd3d10e3c83e62834563ffd
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:b3387bae415fe3a66db2264ee766cf43aa8f31e2b615b200baa9228216596a0a
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:508412a22fbd65074964150d6b270f38d6df898e17eb7ec5e28f279c4c2e1062
        - registry.marketplace.redhat.com/cpopen/cpd/cpd-toolbox@sha256:dcd74c128502dc301482eb302b34c21dc02173ae1cf522356d5b500ba42b4e25
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:f00158405adec00c7e992efa7d7a19a520df485cbbcf39972e579ed3c6e4d289
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-hook-utils@sha256:ec44be69b4c585e9a3d507e29bd287f5b464360d82c8e0969c5fb4ed3e95abc0
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:31470f873a0931d130484d5255692d445e3127398aae0e777a3c3cb702917743
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:f561f089c4e5037f7c04971d96fa024ba4df5e6d17834f0c44545d0e53c772d7
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:20863de342e07dddc9aa8eeb214fba5cb17478bb88613e4e1f167d77cd0af214
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:9460c98c7202ab368e732951c13bb6f5064cf18f69d36b3d26eedcfed233ec81
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:5aecbb4de6fca34f026de72b7480f35dbb2ee8d8ef1f22e524a85a71bea99a1d
        - registry.marketplace.redhat.com/cp/cpd/jobs-ui@sha256:49daf607918b378593485da16ac417056ee295bca94f1a5cac37b334381ae096
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e924abf025141f3b3437ed89830957c5915152d2610213090e33fde988bd879e
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:e6fc9d7a70877100b42e290cd091011f2e1c674a27985fce180ee245373391dc
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:8bd9fe1d35785fa0125b8fb3a60333219a50ee8be7a7ec07386e8e9d65316061
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-bundle@sha256:94c7a1dd290e2abd55c5b8998cf08ad174d3fbfc659ae6fa6fb6cbab61d5fe14
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:ad027c91c719fd1d931cbac8c76735474f676b228f9ab23d1f243201246b3238
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:2f40d8188dc8b41fb072256ce8d109167095469af416b322ac023071814cda8f
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:e4eb6a75e0d207232e479693d9c99ba863f8e62830f3191ca4d1e7ceaf3586d1
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:847751338f3b78cc06e29a4a26a18dc0f5d0bc24cbda23b765198a6a5c4503c1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:ca4da5a8a3c1e7add6ac630c3fe4713a08739f6b891b2290bb17026b94c31317
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:c268d2852580286e5c6770857fcfc3b06a698521f5acd17ed83b8e2cb9ab55cf
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:fb37fe521693bdb8b96400fd491a39c573545f35b4131fdfab501b7ba29713f6
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:4a5ed208b68613d5d88936687f6912b3aad0e93c75f0d311120a8e55200f5795
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:52ce08ac2bce7c02dc30e2c0aa74df4bde8f156debccc59960f02b4b0040ffea
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:d6cdb71edd7b7451a206d1620b35c321f88b3ba8028468329595cd538a45abf6
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:0f01a67524fa46829e455f654d3b9b00ceac44f3d870b21c022051dfcc8a00eb
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:b646ea3330f3fd9bb3174224de69f20ef58589a464d50d1c9399211141f67bd5
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:650e68c41252300836740ad495ad6776a7da94331e8070520e60a1c5e2c5f3ca
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:6f2e7cc917adb54b3112dc4909f645ac08c33e8c2c33266fbc8f5133cbcd1973
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:5e161ff695c0552d1333a1f993048bd055168a9be9e0a75adb7c4464c9c2b603
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:ff5a5be492347e75b309194661d7742d1c909c3765c9947c8b41f4c3af62b08c
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:7ac32270044a9ec08ddaa8dee62b61618d9363c8773fe44ba1dd86e1b53acbd5
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:c70d3c86d4c61bd87a427f7a7571f37fc4d0966ff4f420b73ac4a2af73ed1379
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:978cf4bf299cd3e8cde9d7ee7aa75b73fd9d9e6755567651d6c1c49036b8978b
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-usage@sha256:8438c2229b500fc9c40629280701856345e228fc645a5b468bee2135c61b2387
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:498ad7832e989e9356aeffbee96f310e3edaa6a7466320fda9b4f75034e73eec
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:7661b04bc59d64a4fafb2f660cffb99e3d326772749759ed0d2a1940ba147078
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:1f68dadaf2ab779dece300037f3abda1b0bb33603436b84336f8fb7d53d33644
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:0f46b288d060a15a48ecad8f328a8046bf232b8629c3d79da6a82ea712b363f4
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:cab91a231068a6bbef03ee8db32fadf11654317c4573c6714721a15d0e2a98dd
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:6c226a0ccde697b56c81427a970062872929eab73d99791c9dc670f4aa176ab8
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:67bfdf9115fbee782e35ac61dd93801bcb9e622d441806f1144208b4447f1430
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:2668a562eb8df1117c40acbc3600d6735469d210e5aae4aa03bf9f54d0ebfe46
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:e4888fadfadb3462e3a9ad0395a526a2631872f70caf291dc18b60269ded199f
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:f0c68715f78bd20c352c97dd48ac5a912cad58e19a00159b5e6878f6beeb00fb
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:b0735e396c09fbd08f328ce0686cd55fe9eafd5b759f36cf5677326b24cbb859
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:abd96661823553d623eb0ea1fac8271d64391243dd0f13eb4a4fcc858dfe10a5
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:48d237cd3d363da8a206ca5e092d0f91e8a5bc931ea81763498dcf6f28c0434b
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:c078bd65fb18b31ecfe03c832643bd6fd85f7330b94a1ca26f557d32d89903c1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:631a18f4b437e5092b9671fb4c86af05e99effe5540c4d6f964dc626466764a3
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:b63e2c10478735b51b6828bb11709569a2fce62a19b70157a8a5b1964820164d
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:47a3745c78280f98f2ee352bb9192cdcd2fcd6d4c1a3627e7aa42178a0f107fd
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:a28b71ef64c1967cdd36a16e01b611fff6a387b6a741dbcaf7556ed6be9b1f69
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:095aae63ced410e033d8f7f3d27bd14dc5b93a10068ef5a389f181115c5d07c2
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:00f8bcee4ba7e0c5cf5f0c52e9e7707df367041c9f4dd5508dbefe0dbdba8877
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:ec137115210dae9b69cb2899ed8a5437d94c8a658ae1ae5eb295232b34b0925e
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-velero-plugin@sha256:fe562192e86558a17d1c88a8fe21148bc000fa93c8928f7a3a897acce62e6798
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:e857323acd10f7b99c5b7fb4b0a5b89a808f68036a1d9a224eff55a7ae7744b6
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:1cf983f9b9cda18257066fc82e199bd3e22e2fcba13daf0f60c92b50a4a0e078
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:96d382b1b26a1e1e4074b7a1aa48b973cd4865bb5c3cf02795306e92c752a8b4
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:dadfa9ef372b8d165276fc5f1ee9c63c3f01367d156700810d63e1ba19f3da79
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:45194b56f9fcb3af83158eb0297248364db40e8164fb4fe4f5656c58aeac3fa5
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:0ceaaab514f65f9a1c40717e97ca652506f8b3595186222d470a99d653afcead
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:04f9ed67696fc193877f9da3d73ee3f6f4cadc7b8d09fd07fa12d0ee15f1152d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:9f7e6a1ca3faa5af1db2afe8693f6bf9182296be2a23e305ba6c990aee3fc301
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:562281dbb2bec5191a7c6b595ab853092e6fe3ff1432e77b2471f71ccdffcbc8
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-bundle@sha256:552fc53664722903eccaa7d3782a3e84413509fd357725fe0b0f4a077c48ab2c
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:cb9a33db3707c0c6045e6c823ecb21b0f229c17ff868da7d36602d0026eff351
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:d4738f64e36a8f0c88d40043eebaae1d561193895d438fb6d8bf1a042d0a1007
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:cb3620eb9aec24f7ee54c3ec6709fc223b685177500c503af8ffeb23642dcdb5
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:c330a321438943fdd264b292b68fdfb8be89c3bde93a7bd987bd92c9dd2512b0
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:9a970756cd76ec748b555d5bed97889b8c9f410db6765eedee84fad4976cd523
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:4b3ffae1ddb7a2c28f98199be6343ed826f8fd21b64fc9c584566341f44e44d9
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:48b7cfa31b9a185176c20afc119e37346a6f9380d74000edf6c729b4677ca44e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:a7c050ae29cb93754f59d8f83676275ee9a91ae9a9956395b97abf8c5e20e96e
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:65108559a5edc9a7c4f82962ab2293017e0873872730d60d8487963970761941
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:d34d561bbf8afbfba9362d2a9a58f2919faed1502abf72e11882dd9cd4fc9711
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:4f0cd4920e7cef64e9f49ae1abc5d87303102fc96819e0c3b699d86b8f8a3980
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-operator-bundle@sha256:086eaeaaaf440a1d4a254984afacd8d04583aafb9ec11ac072feb42a87b986f7
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:a2949c3868c2c91bf35e50c64b51fcc24372986744b55520e21e9ce31fde59f0
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:e21a346d15b286956f62bdd1fbb6d7b77a554ee10a475cb698a8796c42f3a267
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:6298c1561c066eae3d420f20212a3c7a8f9069a6996408685437588bf8335dec
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:47cd9776a46ae5fdac3a101d79a779baefc802ddf8e1f8fb02216f47feb55fa9
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:8444703b2a7822f82141afcf4f36bf255650d755321b9e9fe2cbd1bdea41d5c7
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:ea57c7cb1ccd5bce17f6cb35e326ab974718be8cd19b0cd38ea8112e91c51939
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:5d889be3dff96d1e888eff94bba94228aa46f4e94e5f30ba33077ef00692509b
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:ecbc65705de76804a3c7642339ecfb4cc0f5a61891058e07c20bf2b63f1b82b8
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:690e9b5218a200de93a7c8477fd2980d8d341f22be75002967201db09b5a5524
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:9c2093089ca0651bcabc1d01ee2b69d2fb0cf2a277a8b966cec9533950fdcdae
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:a1754e003811afb578dace78f9eaef91e7917e5468e590fcb41920fa71bcc22d
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:010bd5bb9943ffdffa8fd00cc8226e59d9583bb3baaeacfe70e5b00e805878f1
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:3d5c6573faf4ae00cbb21cbfe9b2ad5cd8ffb658df0c5ba09bd98b3bf54f9979
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:8cba5f256747002b62fcedb9066f7ea6dfdb1071eb359e9137fd413d20b29ff8
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:6f96b8b6afb00cb26fcf903b8e4c30b0d6344bb86d8bce2f2026fc8cb0b575ec
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:56315ce0a0a70f5899191a7eab856105489143f0e889c8cdd1b6b8c225f02eb7
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:d733c5a10a21cac2df2fff6378035151e5d0df57e3ed1f96bb3ac17c0c95b8eb
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:d06ef794046dac2c58163c201f55489b69081dd15dd91cab16387510a50e24b2
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:2c39ff1850b7a9d32a11e99393039ef74ab363f3fffec1e4fd542f2ae148a6b4
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:61ba2b05a6826f618853ba048496e2c9558e509a01648c74b50cce6d28d0d764
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:e0feb4d305c5ec9501b30ccdf8f5826cd8eb1ba5ba1d62d4d685101c785a708f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:c8786cffb4a27decb5dff3d2a9682d07d11b7096a689bda0d7cf105aa3156299
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:51e1275f6e8297418997b0ca9615642153371c29b3466f155e15d7fb9e80cda9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:f1aa6add76f59afeccdff84c9ea89a46b49c6696f62acd75ef8ff2f31a3080be
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:132670412a0ce6dd010869e569504d2e89aee0983ff221da371c9e30d5f8d8a9
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:b5a8fb14c3bb7c89ff970ca85f06dfcdca832e2b45ad86e8bb1b6c0a0e7e0764
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:ecac5e0327d939ca84d304b2fe95739ae43a06745ef775ee59e4e5378ced65c2
        - registry.marketplace.redhat.com/cp/cpd/environments-init-job@sha256:6af8f900af7c11eee6894e2c3e52f740b3c700dac0ee753fadef330230bd42dd
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:369f66883e040accb73c3deae984ed479e90eed242051b7035723f0ba21b51ac
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:7ae4dac24ea054ce527a507dea6c77a3233aca0047672ba5a367fd5742448517
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:8d7625a0ba656425f601b6ec9a398a6d30511fe1de119cf0e29f680c704ea225
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:7b2e9c373f9bfd91e16eff68ef2313171c69b94fe5a21d3d3158470c74f54a05
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:ad2215adfb8dd2dc600bf5097b7f88c257d5fbaf541409c01b36cdb67c130236
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:c41e4d3314fac0d2cf5f08e7db85801cb2098d37d0aa8104a347e9e82526e914
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:2e70036d7299168331c94af70a22e3b2063c56160f616362b1a2187c12d003a4
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter@sha256:be5fd8a8b43ae3792c58de1342cca91836287fb584bab003f7f9aeeeaca3f65f
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:54c86367bfe0ab658611856cc0a6c5de24c14a336bd129ac9f16293945a55238
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:603a6bf0e616f16b609bf157b9e21415802478aefc9c606e124008d0224a1188
        - registry.marketplace.redhat.com/cp/cpd/task-credentials@sha256:e162828d5ad07ceae4db0589f930979842714e3b766552102bee373a3a570af3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:4aa81bd0043d0bac60c9f1b0ab56e45a3035c88314614e60eaea48b2a29ffa16
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:587af68ec8b4974b29da46ede525c70054703fce8d5934f0cbe68739c6058fe4
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:fa01c7d11c23a47bb84375389e69619455c2879b5ed6bb5b9b753b945dcb02c3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:08dcf99453b4dc6dfb5620410b8f7e4eeb42fa824aedb2e480e2fb40ec8ce347
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:0164d8fbea591bbd9d67d84603162c3781b08551eddcc2a8a842aafc828af18c
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:c957787ae587a31e6a497b46649da9cd46102a2a3dff29243fd5cbac96e7e2d5
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:4e65e96603d5043e78a517a4c6e5debc5df7e3747fad4d8025ef1fdf1cd6a9da
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:e3d8d4d5285d058130345d6108ed4056ea9c871ed8c0575601a94629e86ec691
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr@sha256:43cf735b51e2ef45a911ac7032f03c74cf0125e2d2801258d4aa04831e1c6ce0
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:4a87dbcb7d2ca2d121512fbf57b645eaa9ab745439b79722f474b84efa9eb391
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:6fe90f051d9261ad224dccf4d9f6bc9f19ce0e8d22fa010a46b2752374bce2b3
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:9a537eae7e581322e2f0764ed38a797ba7cd260724a285f24d5095058b7b47a0
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:3be006596f8e1d36865a9000c77df3223c264e807bf0aae6049a23307e8079e8
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:d28e6a3c65b580051ffbe287492dfad28dbb1c669cb76aa3df23bd9428f4bbe9
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:8cb37bb0f0461efc603cab89b5420c56990aa104fe39c9b77f0393bc6b641172
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-bundle@sha256:98cf2a6c7b7af5d7cbfbc61ce140201272b30b12edabaec2b2d4287b0690d89e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:20ceaa70a3ab5ed8a2e451a1e2960d04c201f0087fba1679fb80c57c6be6117a
        - registry.marketplace.redhat.com/cp/cpd/job-trigger@sha256:f2ea487072393a24e4c6864b2107c1fce072ebba07f9dbc2c139bd37fbd062da
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:83038bb7e3f4b4637f0d5095b42dc906e7fc27bbffcbd66f6ad93dc9b1cd6bbb
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:75ca37701dea1bd079df8f8cf302cf95e8dd4ed817fe126b192ca368e92fb78c
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:b10ca594f31751faad4050b4f7717e0e98c7977fb7e26d4866e1c14125422225
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:b2034fae4f2287c8997104154064f2712b98d3c4382d1faea3df777338711f23
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:849ecfeaa36fe365c7dd8f1a76fa0d263a1196fd16af805dc66035643e8e9d9e
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-api@sha256:a76d3114418a3df1d4d3cfe03da587c37028ca1697795e2535cac9f6f190457d
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:481391c273cffdeca36f91543732ba1f082dc409ee851ae95d3c1d205a7a6a08
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:c730fd4a7baa3f46c03f6c73e2c6c7c0ff24a8900b79dfc0f2fd319909edce74
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:9d4ae40d34e38a48ad2ad6556e997a42f6308006d5f6b8638cb7bc4ce5b1e40a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:6e0a23c3434046195428792c45c95869d81fb4ac4ba75444e4343963fda6014b
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:deb66a64288a370c5da5c7cc0ba4e2ba4ab816d7c5a080a0c505b05f55fa6072
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:443b51f8b10acc85bdefde7193e0f45b1c423fa7fbdcaa28342805815c43db3d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:53d1fba68aa0227c2f4c8de26837f6c114c56ca4a397826e1f2e93bb857e5b7a
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:42ffa27ae47b577cdda214e0fbc46c10bd108211e7dc58777d8236c59cb78766
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:99589ee11b50af8ed6355a5f9c2272ce1fc7f9a13e80fda8eda5883a332782a3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:a0273c013c5943c9b9a223cf1aafbc96b3cdd2dbb147c73d2a9ed49befdbdfc2
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:9a0c76fd9ef9b2dd9af5b0f29e1813567ab7f7d35c1efdf71da1a315f2942b45
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:e3138406cfe81021370924992dca891cba22b3be2eef84546faa0e8f8aa57ad6
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:62e2b059f565f87beded59b36962c457c9662d1e401a78766fda328ea1aeb78f
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:166a3918b942a4ea7b6c371a1e30fd4c928586d7ffd5b6919c2fc728a086ccbb
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:d30397b74780cdc14930a3a02dfcf6b99264847702afe0513ce536983c32e2b1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:f662be48eeb6dc93a8a4b0c3374f59bc342b5b8f5360437978cbac8f74c3a084
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:12eece1fc0a2e14244a2f57dcf7c198a792ced9ad0d02c9606645bce1f295877
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:130872119947fe91cbfb8dd82b3baaf78f7adcaca5933ca394eca3f9ee84c8cc
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:f3f8c6c395ae860fed5761f67ce092a122e7073edbc205c9167c27a4ba85094f
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:aed24dcc7ec24707c978a17b86501b65c6f0a77bd3ea0d173c28684b9df9cd91
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:8b7941a3f518e952766a2e6f88acfa7c9914b6d2bcd4ad9ddd54df313fa30347
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:722944bfd0eba280255f4c198f65f562423a26e2c4ad3580591ee71cb6e9c2ef
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:0341c499ae144ee81dc2b92e8b99ee313ff6759374b7c1a6469156c20f522921
        - registry.marketplace.redhat.com/cp/cpd/runtime-manager-api@sha256:a37773df07fe757dd698e3f09f8687902e7b3592d97855e703dd92ecc832aab0
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bb074bdf89cbc7c703fadb17639608599ab4f66db9be5617e5f79a0550676790
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:e673f1d0c5a0e32e9abc7ac40eff905ce411f9f3d8ca0b03160e23ba5661e041
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:f2f256a2fb787419b25e99c765003d25d1edbbe60042151d91b46a5d029fc0e7
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:a16b822b41f5a745b475ceb353a459e998f9f43bb9bb867ca1640866ba51780e
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:8a6ce0dada706056efda73ad0cc1547566003e0a7aab8ed379b50e4776861b94
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:ec5e4a002cfc63c1ff1bec340111e8a1755c8624066445db36c23fdd1c3061f0
        - registry.marketplace.redhat.com/cp/cpd/environments-api@sha256:6dd60b530f81d7a2d55a534211f7f128814abf87b8cd3575d528dba53442c44e
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:b0c8bb70a01bb19c685d019e89f83d8ed327bfa670a13b78d3e803fed77499f3
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:2f2590734343c99183dba7aa9fa4debc73377ee707f9bb090680750e62acdd68
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:25420a21686a32c62e83b99fcb215df015dd13fec10f73f397f927c61226f964
        - registry.marketplace.redhat.com/cp/cpd/spaces-ui-job@sha256:95dade51444e30ec034c12207ecbef675b70889b81970f49f10aa75160b2f947
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:686700e561181ebe1431f71eb9aa11b550bc4361dab8a3668ce0c7ead217f79a
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:3d83ebdfddcbe6eca38705fc3d575f2e70580cf9f7e8735941afee0a33d3b5c5
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:99c863002ae1e93e97792fa606128d05063abbedad9e9404acf28c745ae1a5c4
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7d9b2d61a4eaac645490d1c9e6db97ba193d65015b6efa8d5a1f3ec6cccb1ad1
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:06403c4ad99655cfc10fde46b21e84f90c05ae143bb6d08464dbc8954486897a
        - registry.marketplace.redhat.com/cpopen/cpd/cpdtool@sha256:8ed09b7d65178aab93b77b9d39371bc86b1022f32c5e63a9192baef02c71016d
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:ded16bf3cc993bfa5af8907a1517cc34ecb511948e4d9ac73bab68888af5543b
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:107911ed6a60a024d151ec3724ddd6d2d1402aa70a519b9dbec1a519d88231fe
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:0157fe80f2c67786cacaaf1302c334a6e0ec0b941afd870e0c4ba163ce394216
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a24f4cb635e8b5759a5e2bda20ddefc6d5ea9b16fb1081592885c54aefb5480c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:0711e1c6001586c950e74fbdeef8f04d20ce7b93de2ad033e82d2d5421338936
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:ad892ce5e6a3f69ef484eed68e92d9c972597a86420e22ff80c247689ba02905
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:39700bf47b93c7f0db1894148c16c60eaf406196034351f54573f0a94c4c8aac
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:3112fa03dc9a530642d329666a0cc64c06a04308beb1ddca0f7472915719ef5d
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:66c1d2493e6721094e85a0c65011523004f9589f58b7f68cc6f33c5e5adb6984
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:68ae72930a346d0fa0871cff573c64285266311208f2bf64101697a540f60272
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:2712755961a1c1e23f30d49341c037858c0965dbfc065aebda88c84526291f21
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:4b6bc5640d9e68519a7f543ddbe3f16ca200f32a74ce9c07e996cf122da3b59f
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:39e6aee0bb64fe6dc18ce81496936b113de6d6746221655cea102e04c11dfa33
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:07142ce0d259412d0a9908b63dd78d845ef6b7c5f8a24442b93964b4c38372cd
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:525924934f74b28783dacf630a6401ccd4c88d253aafa2d312b7993d8b9a14c1
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:73c40c6c72b7f2399fb0ed7e462baa72c3e8b5aa246fd2fe1082904b4afc0796
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:fcb79d76f965e74f7e3459abb92e3561ba2f016b34e2ebb6f88298bb19f7bc11
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:46893c166cabdd8b3422a190e220ee2e34861e797aa068100ccf53da349e3fe7
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:a477d47eaf8f8e569601d3c23c1def66f31c098ac3f5a1ae350618c460917dd8
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d51033417b786dd2dfbc94e648db00ba68f316c23b06a0bc9e36a5c22f579124
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:86cdb5b6543e7184361cbf0b0fa556277232460737ad254e920756c6d365e47c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:e6ce3c998642b235080daa6c0b954733bd3fe973fc8d85a10ef0cdae2a73aeb0
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
        - registry.marketplace.redhat.com/cpopen/cpfs/cpdbr-hook-utils@sha256:63a5606df81013b8a7b5cb07bd1f29bf032c8998db843885184fe86a76fb93bb
        - registry.marketplace.redhat.com/cp/cpd/environment-specifications-sync@sha256:894b4624d8ead7286cf9068c6477300992c32d96339644fe764bcb29b1c8ef5c
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:3a872814d01ded278afc8dfa6b1569d9e933e8fbda28c55a991da010b4304c69
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:351375de35d7fc9d940b905b867ce8a00b7ff1416e70d3a100918797fe3d8c56
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:00c30fe372524fa82b5ed4a637509dd9d4373dc727a9abc706221f832858f9bd
        - registry.marketplace.redhat.com/cpopen/ds-operator-bundle@sha256:be2ddb821274938e5ab926627bd998279b6d19b7789d9683513e85a4b05bc2d8
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:edfb732843a53254ebf7d4a2f109af109623be8fb9bbae94366d30febd2d0494
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:ee15594f82d374da81dab59c559567af74153524a6d9bfeb28ee085bd4109b37
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:3ddc6bbed1ddae1be71043d177cbcb58830d0dad542b06567fcfa254dd11e626
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:75595bf2753fb43391e2a496fb6d59faa8785babebb4aeb1da34cd5e72c93699
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:bcfb9b7d3b379ee2bb0b449cdf5af3fbc88f99c77641441ca392ebed04403e00
        - registry.marketplace.redhat.com/cp/cpd/environments-ui@sha256:2923664475c7a8ea5e63055c90748db37825fe76c7975b845f1185d4b32ee4ea
        - registry.marketplace.redhat.com/cpopen/cpd/ibm-network-performance@sha256:30e0b58531f876a47174c14bc6f94a3c38a40f757403775a06745d4ae992060c
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:119be0bf4c60230c43bd64c0ea252108a0173adfffc9f05e5257ec260b335599
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:3c1e08c56430e9bfc0c831be2394fa28e395a3341ce5cb5201b08828298deb60
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:2698150edf65a99a1d24bf3e8336f02b57fd964fcad283e182794634b033af89
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:077107dd9cede2c49beec429decb66760372a229c476ca1089e42e4ed5ef40d9
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:9782a9de863ad10ed36341d07ea5087e40f6f213cd4453c3ac2ec100d8733ce2
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:894d67753dd0c5b8625369e53635a1ff53066d3910b73d91e5625dd8907bff43
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:6bec687b2bd5c4d04403a25f56d2245f822e914907771f26c25166462976258e
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:993f16d744a81c95e73339d3c8d5880b5cdbcf29beb914fcb2c6403af6d628a1
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:21ce62d80df2575c3aa7c8dfde706cf8af01b2fc86f72ea68f767bc1770fe733
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:2565a8ba9aa103956d6902f2bcb53ac13328b647a2d7a150bf80ea2749d5b972
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:17bac3307ad3843afea77ed213279603b5f775f24dcf4857095baf5029409c18
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:749dd365ce85a4c9ac8e6ea7f90a34b791a7bc5300319daa85d427b04f40af65
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:cff7466aa6063a2bca677815df01208822930f8c12784c1c406fa352d6bd6cb4
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:d5feb4aecf8f36531db4c52e672558c04f7e314a1a791b27b5638202f9477bf5
        - registry.marketplace.redhat.com/cpopen/cpd/ibm-service-functionality@sha256:ea4c32a19b1549698c8a68fa030d7036d3ea3cf2d000f3523b6e9fc7c6c708f1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:0e563d51d601adbd710833cbde26ab8f857092f85686213adfd6fb3fe5e2828f
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:f7e68137e1ca58a63924967a8e4c872912557fa68af8a7392e3e0d4aaaef1e78
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:009d98ff8e3cf353cf8ba71a973228e1e409e4de21373498534a6cce01c07ba5
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:fada47a5f4824d3cd5bdc262b92018ca0e697e18dd373f92fd93186fb8719a1e
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:f88a7294aa200d3c633c0a25c4e4ccda52c75de174ba508d847ffa3092d5b8d1
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:2d320dee224cc4818a84493723f2262d259891bef89dedef09e1f1e6ae5686cb
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:489d7f80d83c9890600e3c9c4cb9442a6f42b9c6c4f13dafdf0f614b72e67143
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:934bc42fb29cbe616edcd93d2a060dbffab5e5a8a2ef4caa7b2e538f26167ee3
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:96e6db93b2d16ba590252d6568b68f254548aaed73ec17e4bdd8d530ea668be6
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:5f0974565de8b045d3faf6acac7c8ebce6cd35327dd76290b39fe48b27ad5b62
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d4dff1a8968c18d984e16c4c9f1517c3a10e23a11bf166c64d734773324d4c9d
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:c411456aa41fe705b47dec60eb49c6d65d4b7bd1a64d1e0545817efc95afc67f
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:f5527ddbbd0b9878555fce852996b13020fd91b1604d8c03d34a61ca2fdb2d1d
        - registry.marketplace.redhat.com/cp/cpd/jobs-api@sha256:173b88f90078e1817686ac949f48387af70a706e23e6fdab4a7c6217db5ad86e
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:85669f2a35ecc4ebf6b46fd793ba6a6bdf996164c00662e65cd7ceece8d60cda
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ed81ba4ab2caa7ffd1cb90e11aa8915df1c5bc1b76aeb4c87532c84097bb102
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:23337786ca20ab624eb428b0d53490975e2b08a2db642ccc76d974dc0e455f55
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:b8c83fe5215b7412765181aaba84162b6cb188b5765e45972a6f29b2a9f19826
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:811d1cbda1d0287dab48cff90c7091dd1bc8aea385f562c896f2b09f075cd1f6
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:7a0e784b0722ab0a4c62754a4e817a1f25e5a2d5e960de76387971fa1ff7da41
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:36e3b8e88cbb895ca89288ed385fb800b888743844f50fb2233ebe5e6c7498c3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:de7094acd65a80daa67669d8eab809220dc715b9cd6aaa96c58e5fcbe34c460a
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90046eb6170e12439ccc5aa343f19a4d4fc99a10d98b383cf1dc0e020ebeed55
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:164f6063f3a50a36e28a240ae7c060c4f90266b7ed799bfdc075f90628bea7ff
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:78fffd8a74eb6495f104ef346bdd03a5636134e2d5e8da9ef5c2647a9ebc7624
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:19bb32e0389361e8314e627bcad2e9bd93984289485cf632466834353d9c3449
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-oadp@sha256:35594a16332ba3ac00226414be9b85243bb25dd0648e256d7ef9d42137dcf914
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:787e01e29abcbb3d1fe58e0b6051067b309085459983f38d6bd18d3811d608d9
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:2f45aabc99eb9646db4c74686569a7f2af3f27cc44362a30c92d171540bb274c
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:97b9ecef8caef156a0eb073b6eb2a34b63209b44511f5694a243f11285d88662
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:bcc579eac4be59609102836252ded82a17ee0088e5ce28c684fff41c80232a0e
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:da38999d66199df0eeabd7c9e934205fb858cf2c52be8bbaf6b34adb894dd38e
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:e29885c1ee083a73b952b0de8d1cd675b2b2830abf44d69a8577f56972a601fb
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:17f016c11039c34f90f67279b45ee488740cc3d033112a3bd589bb1a60ce6812
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:bcadf3d367f76d3454e119a9d39f6fdeeb6cb806c282f21a1bc3c38bed41b427
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:ce57bc323b95db548dcd2f90ec938e708988cf6103579e793ee39d36359d60cf
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:5f63e062489b8e1bf49b5943e7a836174fb47af8dbbc8092ace5c8b72baabb99
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:740a1d87a43e61e32cfdea032d852af33100d959c8b57ff2f165f4baf77fc011
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:d72cef9b19ea48f03f7658dd55cabe6e118be0209852025845d426b1fbf0e0ca
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:1a1bc54977b160b70cc00cc4eb1380250dcd639fa3d741363212b8081b35f72b
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:ac128758328325ae56273800ea010f9559902281cc60226c8f99de51b636b180
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:26500e941ccdddc08dea07d3a4b9aa9eae7840fd5b683191320060f6dfc88c97
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:19ee4cb8e3e60d9c1e8fa63b0bf3e218bd683b6c05b9e9553d9477c68c493a9d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:7dbff355db7739152961cb6a97887d2e43bc960ac58837c126e03ed1a4480a3a
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:075b6bc4866b812c4b5394c8902b012978449a710ce71a30f55231fead941105
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:697baab21258166a1897ce4e056e39bc4667ff837bb97cadf92dad5863242ceb
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:ef03bb6ed647d92b9c84d34810376da7049aad88589eb2e35798d82a22f5b994
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:6df671937bc55c410a9df13aba15687b2832c7c19c0eff4686c3f83648528068
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4dd54195a3283a820d260eaeaf83463722e49bee05d875333c1b99201a7e2b85
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0a773245bec514cb893f730086f39a3789d0204b82cd11f5d11a5c270b92f70f
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:fb15cbade10fb37b14afe063a93decefd9a40583895b7b6b436f36f08bd73aee
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:a6e98064effd12ffda0613db98e795e056b83ef77b5deee932a762b178ceec3f
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:be5679938413dfb7766f765fba7923d3079d2d9b6f464485a8309e8e306f19e6
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:0cc0283b768c2968849f24cbf7178213926e084da5d0c90c18d38e2260e19138
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:82f60c881048618473c07970d6463b3186518bc18e4b263e16ca61d15448ab30
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:e76b61f9f0efe16facbddd80808d80bc32d0d41b4c16aed566116f1366152987
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:b368c22956707b8e594441c0e9bac20981339a5713d4fdfbe3c64bbfdf677b35
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:9fa6eea66a3daff39b7c0c7839ef5349b483b29dc219e1db384aef6c4c31ecb4
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:dc257df716b270858ca5d07909bda9ab635a388c9887d30b015391f60cc7b5c1
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:bb45d0217b4bc4352b209e3771756cb82923b8d260846c75b33ec270b4a8b8dd
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:b797e00eb32ff9a6a56e71d355bea653ecdecfecc121ddb197a6e347b89dbf8c
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:0e0ae96aff8540323da8a2379af014d3af1770e4b0043b84b9b415dda44501f7
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:37853d2979d0ca5f88fed76fed5194b628507ef0480e668dfd6eb034fa1d2023
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:c0459e5d3acb3b2a6d9863efe6ca1a24e529b38a6ddd3cc1e13d143924e872e7
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:582ced3f60212f736f720558524225b0d4a1c79ad4890656c47dd58009ffe0de
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:351358f551dfcf817c5936c0b43c2e28db61adcde65d7d2c81cae666630426b8
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:07c13b18197ed0bdd8b97d65312a342f95d7a7fc43d50af746d788032a67f83a
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:67515618d631b0ec00ad95ecb9bcb48d7f199855af9a64e86c4b46e7ae3be69e
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:932825fc7af8b96c381d504cd0976c751661ce078257430e5c79d565f416bacb
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:5ec262d1ada13a7d5d519ad98cf82dfbbd915664bb7560408558001ea6bb8451
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:3be3ee728cbae6bb8cf29d47c37ed1a4cd5791b54e1262135089424e5bda7ea1
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-usage@sha256:fe5bc4a40472926ec0df3023777b2aa8105507128bd9ac02eac2c68ad08723c1
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:f6e8de9f11696ca22302b8576ffd5bc2c024df0c8dc3734c404cc799565f351a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:36f4c57c6594f997383c4731e1e62c4765b8642c0b5c7373ba22118220928d09
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:2d8f924a7e4ef373df9760b4f1f3ad8050facec7de0c0efccd8c27d46b756e91
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:719b0871f0d32da894ec80dc2a23483623b7802a5c6f935cf3672468f99faa40
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:b372540b630507f4cb74f5eab38d0335d7b5bd64197e3f83852b2c7c5ba9067a
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:f9c9202e106d5a159c8f6b8fd883da4d6ac0d81e75b558fa52e0f7776269404c
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:095c969026e191bc4fceb020b5760e7595e439cca91b88f0d8c3e422591d56aa
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:8d9e35c3f29f8f56d55d81a2e6cc4d1b0cc352ddc8d33d83dced83c4a111a0cb
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:88d3c1ee02acbffedbf6ddcc27b80be6727f85425c361d45cc5e5c4056cf6733
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:1279b116fbcebb1c052eb27a72c67011b44a12fe673a86c1ece28202afd911d4
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:13a9ea6ee24d9b0ac1d5fb09c3e42eff731779867e7e71ce40429b43ba2779f9
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:5e5176376a664755fa776efe623a1e11bc5d3c31a3d0af980b165e54cf06620d
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:6fc976bc622c2165dd8bf93c9244e1c29b7c973a4da7a5ce4d0a097a1c1ea4fb
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:18f44ab15289625fa2c6b72ece6e06afb51058dcba30b8f38b4fdeedf8c9e51b
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:4fb2127ba7c525a959bbbcf024de364ef266ebb3ca7e696f74d990d2552cda59
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:a99b1f5a4eb0306b83562de0184fc826c34835ffbf46d929ceb9e70e0d325361
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:2011f32eb1e68ed5b75a4e8a99c992eac546995e219701d17a511596a5a09361
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:99d0b3192e6e46167423505f48655e7d49af1a461ae00b95c9c7c9d4209e5b65
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:65d327900e3b7a874dfe4bfbd0f2c7a1ee402346aba0e8cbc628b406e5b5329b
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:bb85f2567aaf7b612343fb9819fe05e2e771ccca8383316d34282ec37fcd1383
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:b2ce782efc5349a49d9c7326e2f72b2c1d99d30eccfb6975f7cba10008f9b2d3
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:267b7da5936493b0daf8ce824e3e6b50ef3b2bcb6e3126ed18fa9e09cdb0da43
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:27195e9758afdf8521971866fd60ea3dafed91a146c3c4165876e5ea476a4622
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-bundle@sha256:cd92fd1c85a72560241fffcb3bb62cf965f96bf2a4e1f5cca84c0b52624ae962
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:a34169228887ff93161584753c990bb82a5520ad7945dfa1ff4078e70aede592
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:7dc67256c002e86ee416c60ebb2942b35d8dfe0a5f96b4db495eff7361d21466
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:7de2c2f7cd478baba2011461c8d28c1a4c70aa485404d4523d1166dcb94c0a79
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:427e431d5b8b56011a91bc4f1e1586cd5490944615011fa6ac20abc8f45f92a7
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:96451d4a9d1c4b759d82dff2eaa9aba2a6fd58518f6ac6306cc082ebb7766a64
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:eb157bfe50005f9ffd3560ce1ef7ee0fc283ddfe265c42b35a71f18409e3f561
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e1b10435f94bbb7165703720f58aa8c68acbdaa18006a2c36d1e24ffdd72cfc3
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:411a4c7164b59ab3eed86e67aa457e45bd028067c937d35d0a227de63870cf87
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:3743af183b7b801f72b279a658b0d32da0bdbfa526a00d191a9a44de6df2284f
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:2472053d912f693a5fcdd00b7d9f0a7ce98697c47eae955e31c0b602fd8a4a37
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:b2ca6bc1119571cce732f1ce2ccb9ce406b298550b55144f950bb46cf985df3a
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:7cec55c00827b747f23c2f1703d99e1ca5b30a0c445f8ad0471b25b0c9cca57c
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-ui@sha256:b9a255ea269a11c85f52fe556291b0e869efbff9918b488b65e4297b958c6059
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a4daf28771178b0af391965684337805f8de0019a677753112961710f6011991
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:b7468b824d519e46e3ea23f5a803a8ace030b74d71d4cd282819027665ff0e12
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:3f2f97e6b2b18526255e88ff02d1039b24312eca51ee22b7cbf76d5680f7aee1
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:183ee860bf48076ea155db08032eed7004c1204a363625f54218734b00175441
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:62102f79bec5a8f66c600d184478f2b42c8acfde059fc71bc262b16dd46975cb
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:3129d9c7b096f9fce9cc12002eafb5c3d731a162ddef61578455dcf8acd31fac
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:a72c2b70a9f0a93047d47247cddc1616138cbc87337289ddb208565af6e4369d
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:27210840cbff8a7c61acac2ccc328a415a405b4738422b4f47eb9a0110127d22
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:381170c9c497b1bda33eef3ddeb1bafefa07be1e6ca75e75c9eaef9dc87b8d4a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:6bbcb64cf9adb4d3b6a4c259eb910e7da5c3c58732643e60d416399b2083cc0d
        - registry.marketplace.redhat.com/cpopen/ds-operator-bundle@sha256:e30b09797e341a5fdcc3ec377657e0e65be5855fb850dfc31f35b200f43130d8
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:ef24cc79265473a8b6fea231b1930c5763293dca79eedd46c89ec63dd95812c5
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:a31e0fb7a05e969e31c10a74cc858c483d5d4da14ce3abc36f5318100e60e792
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:aba8f1d08d143a8901b4d2cd8790e067fc7b9eca9987696a1ec6f9f505a686ba
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e12599ff99db058fdc8f9c1eebfba31160c636ff092138deb1f6f6705f44fd62
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bcf193d6f30ce1c7f9337ba8efa8d9b853eb7961e0008aa82635aa383fbe9d76
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:ab31869bdab5180b816a89c45008328cb7013c26b09c531bdbe5b00382f0dd7d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8ccecc00f3a9656a3dd1b07af76ef065e301e4101bda4ae8dc5bb91097cad7ad
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:c2b9d370b6c3e3c8e30079465502ce376055261cbf0f95c48d8a2ae10ca36c76
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:a6d5bd253959be1f91b50f5f2c56ed77e1c82b06483e62094aa86ff57cd46eec
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:78206c6634f2be2bd36806eb717bb8e4db2c755263cefa6d026a425a19329324
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:673320deb0ae96c6132545032f6b3e569d40204d3a5565a568dd97c5c1a4a552
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:3f6e9eeb659857799e30011f222924cef35301c59eaa5327ab5e8d212af75ce4
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:be55d9ad6ae76bd061046af448d032e65926abaeef95c831c978e01e2b0355c6
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:f6f3758287dd31889933ed0461fcf02dfa85d872c3d193669608573ca5269524
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:14da175e95c91864142ae00e9087ac2fe0e2083fc1a3bcb20a6b29dbab99d8a3
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:0cdba17b95a959b6d04889179d105515ed3176ea3449bc104f18b86374876048
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:b19bbc36033c3e60f387a9924711aab1ed81f150ca9d8f435e5433fed7f6a5c8
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:a3a0d58f895fca4d832e602d0dee80018df4ad7bad8800ad4058eea07fda4817
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:8d744af7dfeb9892378b7cc67d0e85e29d12b6f79c8b6b2a53f7a53aeffe8231
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:fa3d8a54b1ba33465171240d66f62271286cb120c51c54267d2b4f8b731c6177
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:7a90c8e0180da12172b23f3b54536bd1afe35affe37d908ddd78996a81435751
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:d2680c5f05c886b44df644aef6cb2fdc9534bb6a238066f6ebbb97bbeaac431e
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:6ccbb20c9fc7ce00ddfbac19c8808417957481473eeb1ee62ecda165e0e275cc
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:ca175109562cd9140d98bea3be88d6b34490ccb8d79f0cbbde08c821db9c5fe9
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:4f0165926685f49cde22a93387fea401bc1799ee846041fe54bca148dfc2f62f
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:70d23c45cc636af2148fadc1c7a08802692fed14e1a31893214b8d2d82f06e1e
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:9ad526bd56829ef24834b8a790b29c50809866c1dfb60cf8e0600fb9a903d388
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:0ab7d6651cdeda78fd0416137eff72eb58a621c8cd55bbbcb16f269808dec7cd
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:545b41cc6c044d6807bfeb7453a045e074a5670cfff9f5c311f4918946901fcb
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:17f7039e49f28abb7239e0c4f3689de58c75f34c9bde9e7b33a6ffb93ba92677
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:0fb78efb9e92dd0c2e52ca2488e6f1e7431e197f87e6994e15f7d621b90319f2
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:4deb399f28d3385c50a256c6e927cf27115c311be5943021733e4b281c12a419
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:5f086cdc7a272384c8a9b680f8f90fd471799e9e96175b40c7f16126b73169e7
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:786b21db2a9ce6cb0eabd5cb7c93dc41a8b1db6c1f2d07725a5bdd75b3a1b7d5
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:cdfb097f2438b5739bb88905b8318d15d15c4c64d37cfa6fdb06f44dfe976e6d
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-velero-plugin@sha256:2dd02d8f15dca1dce7cecad098cd8791cf1cbeddd2abd8e8c6d8f066d85f9e2e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:9c635194a9c63fed440e7c0ab0be1fe6cf59b1c3f954216dfb9150fe9e0fc207
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:044b663e4b7508c62dbb28b7fb2aa267c9771fc62ab509b08d46ea1499d95431
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:9c24d02b1cbb3857d017453744bbb2e68fbe133210c35e7da5ce0341921d5b08
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:cf2b88be76a4aad436f45a79694f32d2e2175587c3630543a6d24d7725b60c5a
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:9b91aaeacb196806646a1f196d4adc83b5159de767fad1a2eab699f735938911
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:d023f35b65816f244b6d699b41362e6031553f7d4c416efcee4a7f2395a61355
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:1019e13e6f072c07da6d31277c8ac639e8ff2200599ece1666a3e34745e763d4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ee161954766211651d7cb13f3f5a3c025ac04094ba742be453ce789c6f0e24a
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:4d24fb274b5429ba61fc4bc6833e93b14b92b8d525556db64e2172b45433de31
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:772f98763527c5452c662b2ae516de59c1f6edcf339da46231f4be8aab6c580d
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:bb6492c452fa6b481551628da26bb85bad151d12afd34a89c1a272a9587ecd1c
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:00cc7d072ea6b6f5e9f3e5b95b9cb4ab8ad7bd942d8c259e70f8ba0e9c09a464
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:9c160303552a9b8a97b951311018de2ada410883468680dd61fb41519b9a9e7a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:80c0913e2c84530d9b7c6a2152de76ded5754c134f6768754d00944f46d0ed4f
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:777e090e0394ffdf39abdde285b5b2749d6835e61458f117dbdac108186c34c8
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:d7ca81372914a28c8298de5f9d3c35cdbd6d8d8582ac9b8fda5a85d22e3c0def
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:07c5e4b3dc18b921339a7a5c78b8fe9179205678eb662377e8124e46479d7f25
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:2a50cd6ee3cb352ba186a57ba73fb9df321cf98f0b79b3d81024cbe57e0c7f35
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:866ebafb730e660062cf7401023d82c5242a2b80cca0126d47a68193c84670e0
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:c2aa3566f8fce992f00f43867173e5f47593e8618357369b23044fdf3c67ba53
        - registry.marketplace.redhat.com/cpopen/cpd/cpd-toolbox@sha256:1b3469f6fcd037aacabde3e05605b908e22f012d331d7365d0450f6d933d3efd
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-api@sha256:dd9e4909317512a6fada5ecfc7cdb47c6807f266fa456a01b60d7b3486faeeb7
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:fb850a3d660729fe8281967c2cb466fbad3f01fdc302db17988c27682217ad29
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:6caf5203489492755e5cf6bffcba53be0bd32662e8cd331dada68a85b2c083ca
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:7693312b2bcee7b8502de0da8786be985607fbfe6616bc9b0da4af8a6fdedea8
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:4ca30e2017a0ff2d64100a03a84bee8d601c765f0e624e1f485cbb97748bb8a2
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:c35eaf4921bfd046985ea7a03fc325f79766bf73e3ddba69408f4467fbd7994f
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:4184f25155643f01850f286f15cc92294a92d8fdc8e5020b1b522977ba429c2c
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:61155acdb8982648ae5748be9f5944ba913db1590ec3e773206d6fbb1d8952cb
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:474187fe2837da4b86ed9482a96fbca0a7f83db149859f8ef23db84de9a6d7fb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:55d4c6a13452f76bbab485363288efcace23975623b8a4b6b83f3c549f844c58
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator-bundle@sha256:c743dd7de2edbd08c6c73e20cf15f3b147491ad77a81677339e88caffb93e84b
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:b8886b9b54e6ee4c6b2981ef092e54348a0286577d9924cf136cee881f5bf1f4
        - registry.marketplace.redhat.com/cpopen/operand-deployment-lifecycle-manager-bundle@sha256:68f71f2767a40ca9333f0a670e07493e8a8c13a5c7ca877ea5bd8a544b352d9a
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:2754384c6936b36ffd64f89df751cbe47993f73dd700a184af35f412b3408d72
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:887348b5035f5a70f1c31aede992a5c76397f5eddd2aebeb27074c3ea6248c00
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:1b29e04308bb8e3c990a8c41a5d3d1e854e197bbdf0cffb2833fea4866772ba9
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90d45b7a3c5ffcd30a1ca1447693ba4aedfb400983cf49779a8421a42dc0eecd
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:60dbe69693314c937a69bd213de438c3af0a8a290aa9b1cfaad8a57e9d04c448
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:ab686e9e4c38b7909994f9fb496943b745e1a07492fe61a61b9e8a78f756eb1c
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:abec4867d6b9294d674cc02b9f7f254e65e7de8eea0231fa94f50513f2b7c45d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:b6500764d4f6d281f1b42458eb223d21f6a413b5db55f2a2ba1a7973aad8465c
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:3c18707afb90a7dc1bd2ed1fff05c53cce89b7c85062e72261f538a12834ab37
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:245f81b08980bfd4790cc3661e609e2cf53f3dd4d4c1f988e5db4dcf316d90fb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:d76476d2ecd88ea7b4fd857e591cc9b71be9d31fe0fb0057baa1b21e6113f730
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:186e849f3cf782f0d78660eb4433853e1e61296234d230a9dc5921fa4bd41060
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:6e055b6e07cb1adaa652b8e7515c4ce91438593a56ab0ae051435f55373d96ab
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:24dff64a80b910f0b4c5442d4f611f9df7602a16c5231eff2c9d08f07b2a9f9a
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:0b640bda3afa672011d5da390af798bbd3e6dd9126029a9eef03846d41582743
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:62dda55cc870279d491679f6b43be177986991b6eb85d9dfeba3ed2e72b3f10e
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:25c98c67d09cfaa3a4d332c461fb6766700a65876a48a1aabbd73b58bc2653f3
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:bd237f42078822b038c13271f35e6710273c2a4f3fe2f3eb74429c93cd2a0e41
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:bc39ff644108a458d1b7e29fcd88bf53982ab4070ce2ea0e75b0908678b9205d
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:21dffe4500172f74345fdafb9beb476e0ccffa55f7bca59f372f312f8d28d247
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:8f9d3438f58604cf29d643691b125a4138275b2c09cca4afc8e4b29289febd77
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:e8e0ea5cb1a070443011794988f2f6c9d67758605dc2847bf0183e19bfde52eb
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:9f835bb2984329db9db7eb2ac5521371ce0ced216a7875e45ef3ec31d255b648
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:234a8dcd8a9c5a68533bce6e0f793324e3ad7a660fdedf95895c5ed73a4cee3d
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:48828f128c98e73f8e345d87bc228f0920a3ff2cdc325d4a4ffe73ad95c5c5e0
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:ceee171c5ce0a8e580dfeccfe0dbfb73474c109becd45a562411678f3ab08595
        - registry.marketplace.redhat.com/cp/cpd/dap-base-job@sha256:f208686a3f56d2980e7b8640b6ebdea9057109e7520c6817a731e68d08506a36
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:31edab3e5662489dc2fe9528d078966663c8d49836425e648b7e112f871808dc
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:bcdf9b7eea8753b3e463973a2690140ae7aa8c02c3c7726511f116bd751ec2d8
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:63d4a9fa16231cc0c01f1b8bb9363221ac1e93f1571ea132866f170b156731e8
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:8fe85b93b9dcab9c33306ae358e6910685e109f32662d02e2b6bcd8dae243968
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:da0190fee910cb8245cf7a61d6e7aa2bf599d41c679eeedc183950918ea2611f
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:ad19b65042fb26cc1244a92f81a96e5ceb4313f8c823181efe28b66e9fecf9c8
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:78e33fff4bfc2e33ea179eaed9902df9bbd11463b9152d0281d1e0924fe09563
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-bundle@sha256:842f2ac568db4db79cf0450b413b4b09b5f2180ef3541385b46fb650ac7066c0
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:c43770715b76e0aa3fd9b3c2b21998ee369f2142e3c5d04320c677c9e9b29add
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:4eab846d6e80d68ce488e2cda36329cec1a6bf2f81feadf666b1005fed294e48
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:aa8d647f007d921db565a505d52583511e25de4a6a1187fc6b4df15427944364
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:276d5d06661ed401f0b0672dbcfa3d9f07fa2cde662fbb532fcdffbe3e332227
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:c6862158a2a5ae0a4e4206bfdc7f73e1c2359544966c72cf4fc3a617acb13697
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:4794a0fdae9d060def1e0d75a98560e713ca18de2625e4d9e5b5fa4c4877fac2
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9669ed5ba26ef32ad36a6784b2658553c00ac2a89cf4e9caddadf7bf46c1592e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:8305f26c2ebc493b54abfea0757d65b679f6c96bc3a0840c89f08a034327ed9c
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:0ae22914dbaf09d6679bd06bc9b6a5728e744c9ce2309cbd5f01e28308c403ea
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:209ea4b77b3b0863c591bcdec6f53c4c39dafcb10ff13e27ae4e1cb986a59727
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:483bc8f74d02efcbfb40ceb6abb0ab0b6581523cdea576b34512709b98879233
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:f40a9375fe5722972f995a82b50c00f254e3fe37f202b0aa1bebf66b0504099f
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:e8eee5083cbe317b368c002f47ce09f98e0c5e3407e5a4712e1c5473c4090ba3
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:523dcc671f1c05e690d5e578a90300d28d76570f605d4de7f80645b590f4a462
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:7ca8d085dd9aeb568eec2c4da974c03aa72bbfe4648f5d66a62d0444a2d1c839
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:9cd2ba7ab3ecab7094c892086a5276bbde22ff96d26ce0d087df2c1791235b02
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:ab99b62d40ba05cef8e699ebc7cc65271ea6fcb5900a76fe17526e7d55b50b08
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:a64a93104dee64e0e7e3e381144340511f903155f72b7f619c82c1d440712227
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:aa0068aec14e37986620c982295c083df6fc6bbe95facba582ae3c6f769a8e3a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:a10a15f07b856a7803ce0dcd30a542741dcb591804dc74c2a5fd50c55de0a983
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:10a935e41cc2dc09f8f3939cd512d9f6bc71d68457f77793a6957c5420cf4bf5
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:b87fa8374976ccd3dad90ddbf6b65cd6ae5a028ba896562f284f089380ebe44a
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:81d8b25221d827160dd5e88efd383453f532057ad02d7a2ee3dcd9c6eb16ba65
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:973eda4612c905661ff5ebc28c7cac6d9fe16fe5edaa544548d2311496ed938e
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:b82a10ce8f75d4441e869bf62fbfb5da349a2f228b52178bb29723af62f13938
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:5436a25b6c5b419b77643a76eb073bbbf8f09517299dbf4a992ae14c1965972c
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:40064d2893ce54b358c02023f4cc3603d3d28202ba56a177b20636abd74cb559
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:ae5e6858ed698d67b4b3917c0d41d0b4f7e2564585694f834099584c21a3c012
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:268c763fec7defe810624bbd31c96407fa37fef47b904758bce4c58d95aac043
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-bundle@sha256:94662b9dba5d4bafeec53b8ad334e7a1235385b21ac0952e12975f4f0a7623a3
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:c59837981b9a2a645c9ffd84c40d4f9ddc3797911f7a785ce4e3a2bfdd0bbe5c
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:bcd879a7cc3442ef62bb9677cc5534a2bfc9e7ea248e566aa4f1f6a0242431a8
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:b87b6c32a2050f1d5ff311d163a7b45efa7d5de1d352984b2f70fa578301a952
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:b9c8a65c1742c72c6351b1c8f4a858940c2bfa00ac5ec271c57b2ee7e5e7a66b
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:78682d9b9f19eae363cec0b6d31e696b2a5fd1bd1f8429df6f65a197310b6778
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:96ebe954a21f697ebaecfe5aa2c497c60b2ddec17e9ec4b376f435e6c53b3c29
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:21c81ff7c4b8809accaf62b242fd9b3d52f235489ae427556b7f71cc21bf8926
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-oauth2-proxy@sha256:bf4052b099393b563332acd5dcba6ba37ac424aa431d2d4f21532838b5ed6d6f
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:f5248b19c4a9b605a9160ecdb5dbbe0906841b667ede6fd6cf8e1d6c4369fc8e
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator-bundle@sha256:e4340bd9481da998a6b89e94c884c46e17cfb50d9d0c1099f39a3c54e500d2b2
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-oadp@sha256:9aec2bab8c7a6941e97c18352cb9e838a99aec08b06269e8053ac764638a90b9
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:0a32e127dd305fc445808d0469147db6c9704cd15d924d57d791e035475ecb60
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:4e3ec61a34fc7c71d213999b68b1f5c3f866f97e19c12fb001d506d4166a50f1
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:99d3b5a88d20380bcb4161e3b182f305af2e4cb405460d0f2a91965dd6f99f80
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:173ef1b032594c4e070e2e013db469115ec58d970fa8b62f3b748163b46d6d3e
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:a65ed5ac55cd5be99cc43c5a8d33bf127e8694917fce5f83fb8401c04e31b1f1
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:83ca1235dc054bf9ead64932c85333ea0dbade076b4a0cd1c3fe2b6bc2e05896
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:4ccd786884e39641a74096d783742c219fb277caafe80b667a80396c3a9cf6c3
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:6cffef7631ee54612fba71efc0eff4e17bc08defc3e2137aee9881d685444298
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:a3dc5fcff6a2aeee7c7e562fe9f6e0cf0652fd7ad4ec61496021d366206e2544
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:e296fa15dc14ee827d5046e8b1434d7f6838059c41d355471e1275e09f087411
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:68a98f52b7eba7e41c78aa91602c4e88d21d4a2a2a11b9b7592f181ce30e41ea
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:dd76a8484628970b62238334e122768aeea820ab18e24d62d23a6328c83b868c
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:e9e408a13bd103fb46536f896c51e7a27cba38d9b224231b20328a28efe1e616
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:9ced60f18740063ab281b836d4caf21035242ac3f659c6bb022ddb546770bff9
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:ebcfc8941330da6991480dc18ac34d6df4fe879e562ee537e5c36df73f4bc536
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:bdbbd76c93f0fb8f838edd9f124b02cc646a6f735a68a430a00f025a06c5f407
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:c0b6ad3b88f216379fc8f14aa5f075f5adc06ad06cc46495d3a40cc765f3deba
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:bb2f19e1797a900d60fb9f05d9f115faf20d514000aa3acac55fdba701581d81
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:93f73afa23f9fe8b73fd389081281fe1b0fd74bba6aa9bf7a6df663053afd3b8
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:54d8ceaaa06196e1ff524b7be595cbe90811f3be1a57c767b1ba74f53edb2f0e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:20f05c662659be9b2b54407c02f1805b93f09a3c204011981b7f17a76caad8ee
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:758f9e49dccaec5ebd125fc0c383fb5a656ba9f2f301984f993d50b103208d12
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:7655f531c354cfcce438c61cc5cde2c7690bfee4e2eadf9f99e4b157f163cd73
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:66c11636fdf68631e37bda7467536093b9402e883999e7989dbdcc4b4d2ba7d7
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:0621eb9b83c9b2b8bfe5cd09797fe08fa37a854a5506914935ff6f0df599b631
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:57e3433f860ec2b65f5c5f86574614fa9273c09b6b37db96a9e1888f4f6932f4
        - registry.marketplace.redhat.com/cp/cpd/ngp-projects-api@sha256:df420e9f616ad8a123d9eedd117fb6cb007b30a772138db534f47ad8801fa96d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:6d903fd4bd0ef3ef361d22b209ba7b686601b49d6f8d4e29b5d030da8d949dd5
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:d904761e70d3969a4ccacd71a60ad4d77dda939c556615df8d65782ccc3273aa
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:74aadd4fe99e66b89471222f853e442fa8338a23d9c24b4e908d38536c66456a
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:c91fc0992cff4b674f5e3690398e7dfb70662c807b930b9b0e65ba507ff925d2
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:24ca69c2f80a4369e29aa382a94d64fd3546b110787df1142c45a20b8246e844
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:9021a04805bb7f21d68fc6211f844a21c1d06f700c332ec3238073065b7c224b
        - registry.marketplace.redhat.com/cpopen/cpd/cpdtool@sha256:714caed1042ee03461efc9256a3da1046bd8f43819caf3a91ec38b118d459869
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:2346f67ffdc1a802720b325e010018cdb908aa458f372fe50ac6d8a92d8fb733
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:ef38f12712c406f8206742dffae9378e344d66c774e5b37b4fcb0b657db456ca
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:7505eb74c6bf165d40f6a8f3511f8e891cc34c0201fee38240c693a08e002ab1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8da8b240176c00f8f6eadf32e5614c663cc1d56d546c73996cca614b715acf4a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:fd35c33c7361bf04a66e69112b05a078ad65fdbe8f8ca8b7f4ecc2143b1643e2
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:67a02a9264d9c6e82edc15ae36d4f99b3fa9099697c550088d27c76511189f37
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:25a4bb9cea1c3bdbeeeb205622512bcba98163f9629eba35625756346ce9d652
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:860c17d2b2c6b0e013d0f4c035a388e45d8ef1483f53aa7ad022913b025f73b0
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:7d8fb7d8e2551bf949edd33589393fd12f634871b486e26464cd089e3e4b429d
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9186c350fa54e8b792d72e47be6b54a67b610f726175c9641795010d45ad51f9
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:f289ac0d44803ff12d41c3659e6fdf04c404e0b3b233b6bc44eec671abb1b47e
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:c8d2825596b7e675c247158bdd06911e1cc0ad9af3364c9c175a5b4cf16d5d8e
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:f42395ea365edc2c1bbef914a55b091e6442d17bd87bf3730a8f9f6d401bf4f5
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:8160c688f6b66206f8428f78dcf5e21b65c3a1cadc38b765f22ba8403a87f4d6
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:aea8fcb72664a41c3881839086a47fef1b54dd672300d4aaf583fd0b575a64b4
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:b341d9af40ad6aceb4ecc792ec75295313128a1eea3a0974aac879aed24a89ba
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:838f7649df5a7c13d6155394e211ac565aa86233bd30a3a2aab9706ec4a74eec
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:8542b9e449b054e455abf6e7884dcfda708a6213015e932844568d1853b65eb2
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:3af0967b0fb0a28a83a3e007b2c78fe27ad56a5e9c153806c6a7b17d7d9b1df5
        version: 6.1.3
      entries:
      - name: cpd-platform-operator.v6.1.3
        version: 6.1.3
      name: v6.1
    defaultChannel: v6.1
    packageName: cpd-platform-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Arkmq-org
      provider-url: https://arkmq-org.io
    name: activemq-artemis-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: activemq-artemis-operator.v2.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/arkmq-org/activemq-artemis-operator:2.0.5
          createdAt: "2025-04-04 14:35:00"
          description: An operator for managing the Apache ActiveMQ Artemis message
            broker
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/arkmq-org/activemq-artemis-operator
          support: arkmq-org
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: ArkMQ Operator manages ActiveMQ Artemis messaging broker in the
          cloud
        displayName: ArkMQ Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - messaging
        - cloud
        - kubernetes
        - activemq
        - artemis
        - integration
        - operator
        links:
        - name: Documentation
          url: https://arkmq-org.io/docs
        - name: Repository
          url: https://github.com/arkmq-org/activemq-artemis-operator
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Arkmq-org
          url: https://arkmq-org.io
        relatedImages:
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:4bba8828e72fd7c9d38c22ba1c2a4f7130817a4e7eac8fb63c13e86465a8e1e7
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:b71d1a57c8b3820a8133cf7a25a6edde7b29e1c4d32d7c30325183415a8bacd1
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:49f34122ab46ab5d7ead094b6e40817b62c75b6a809f67db07d68bce750da51e
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:a4a0d0453f0761b9e114d0d4cc7c14091a2a9c30790f18ffa6bf1e365d5083fe
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:93168fd8cd4fb32f041795afb8fb0e21cf81081d6cc463a44356e153dc1ad032
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:e2438670831bad39bb08147bbebe67ef9465899fb7750f1a5a185569f99d5fb5
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:8fbdd1ba728058d5f437186b9055ca28596140068ebc69ecff4fa52717b19a82
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:81590ec33b2f00775d3c6596f610887b670de7c0b89148b09eab7584f5b0b0ec
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:ddaf00914e3a412d21ec3a814e24f3d4cc382b5c4d6c21e9b1b94dde16ddf194
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:2895b2cf6292c951a0cdde78e755514f5d7cb43ce9f7b6e7dc381d8dfeec31ab
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.7
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.0
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:3bbfb6177e4839087f615a01ee54f9e06a95acaddbb083e11851b6af9315dbf0
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.0
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.0
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:42230ce9d11a1949a5bb73b6ee218b1581f585bcd553485ee438e1c5b50fcf8b
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.10
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.7
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:8f592b2454d092c82b5f61b4e04c3aff173db246539b3c06e15a71960d3bd235
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:f10983c68aff358e285d89d827c27b3487a3ee3b8e121cdaa295bf5fa8501145
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:b293c1269166cc6181f68b8d803fdfcf20fdbd5df2787ed0453724b22e298fb2
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:f25db980c3229e66cf48690aa175f6f93e3dce3974c0e70ab23af6462e7d1893
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:9bbb915855f29ea16950a3fd50d0b9bac25f82cba7d197579e04d9c33a5b02c4
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:de82344daa0caeabc6559656da0af285436a687e8f371744c0c51bdbc309167d
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:e438bcacf099a92b143793e88c00c6ac0afb31effe4b0eed49a935c9265ad943
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:2d155bffd246f26f516a6075ec170babc1373c72bb664cba4e3c58521ffd3558
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:264b33a9aad90d32a41258904f35092a141d76dcef0c115d414475444100aace
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:56bae938a6b4cd87f9f5284931f440f9b9d116cda3b63b13f83a4b1c6c641959
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:da95bff69deb63545d9782ed20cbdd17c1a79c6b16165580bf96a1d5191e6a59
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:f2c853ff91809282c04d5b1381ceae366518bf07b16649a61aa6c60c47f10714
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.2
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:b6f61188379def8c796c738549c099bd40bcd9bb44f5c30682c5edd136347aca
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:fc9b9a4c7550d84fd90335dd4f9e0d8d39c09136c6034b28490a368368cb11f1
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.2
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:9e1fefc9a94d72a7b5a3f4e2d53ebd23e5ab221df17a5afde42ce973277838f6
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.1
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.7
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:634f3d8fd620607b56e0360a191d23a9cb8a32ba0ba1f24fe20791ec52343586
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:4be60cfa2a300b2eef9b5f10ea756bd3f55d75fbd8890232d907f8d64080449b
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:37ab7354273f4dcd19bad6936f10e92220d98b8c0266002221477a0e2d146c81
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:719178251893def955e444e706ff07042ea98d294671315d95da504c1f47de24
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.4
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:70d4c799c7d905fac27457d63e662ea72bca9e899a692e425609e4eaa45129e6
        - quay.io/arkmq-org/activemq-artemis-operator:2.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:5e768aefade6d18cd65cb1eb5ee233844f9f7e7b52631f93a0982b67d382239d
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:ffa0d7cb4068df84319776dbfb66210297da1b8466ccd5451318fc3de79736c3
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:6f2da9ecd8f5021964d0dfa85144830efcfbd0462d18f18f036971e5f17b3bb6
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:decb3932ebb5b848b90a4a23b083f0cb5237eec2844095d8c58a6798da8bb999
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:88a0098a628fe43a9b422f8ebb774086fb65ed389a195e2779482a4a63447e28
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:a4628c2591ec06caf21a3467f6e3faeaede7fa402b4f137d943b25b4dd4ad343
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.2
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:52f337cd5d7d5d199eebf4fc722f8c5279f652506f009efbc0f8f6de88e18718
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:1fd99a46758cd29fce846d0d590021758b2a7a3efedb817bb205359caa7d4f67
        version: 2.0.5
      entries:
      - name: activemq-artemis-operator.v2.0.5
        version: 2.0.5
      name: stable
    defaultChannel: stable
    packageName: activemq-artemis-operator
    provider:
      name: Arkmq-org
      url: https://arkmq-org.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Syndesis Team
      provider-url: https://github.com/syndesisio/syndesis
    name: syndesis
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: syndesis.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta2",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.12
          createdAt: "2021-08-11T12:00:00Z"
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta2
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.12
        - centos/postgresql-10-centos7
        version: 1.12.0
      entries:
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: 1.12.x
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.13
        - centos/postgresql-10-centos7
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: 1.13.x
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - centos/postgresql-10-centos7
        - quay.io/syndesis/syndesis-operator:1.13
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: candidate
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.13
        - centos/postgresql-10-centos7
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: latest
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.13
        - centos/postgresql-10-centos7
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: stable
    defaultChannel: latest
    packageName: syndesis
    provider:
      name: Syndesis Team
      url: https://github.com/syndesisio/syndesis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.22.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.5"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.22
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.22
        version: 42.22.0
      entries:
      - name: t8c-operator.v42.22.0
        version: 42.22.0
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: stable
    - currentCSV: t8c-operator.v42.10.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.10
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.10
        version: 42.10.0
      entries:
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.10.0
    - currentCSV: t8c-operator.v42.10.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.10
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.10
        version: 42.10.0
      entries:
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.12.0
    - currentCSV: t8c-operator.v42.13.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.7"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.13
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.13
        version: 42.13.0
      entries:
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.13.0
    - currentCSV: t8c-operator.v42.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.7"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.14
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.14
        version: 42.14.0
      entries:
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.14.0
    - currentCSV: t8c-operator.v42.15.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.6.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.15
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.15
        version: 42.15.0
      entries:
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.15.0
    - currentCSV: t8c-operator.v42.16.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.16
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.16
        version: 42.16.0
      entries:
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.16.0
    - currentCSV: t8c-operator.v42.17.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.6.6"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.17
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.17
        version: 42.17.0
      entries:
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.17.0
    - currentCSV: t8c-operator.v42.18.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.1"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.18
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.18
        version: 42.18.0
      entries:
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.18.0
    - currentCSV: t8c-operator.v42.19.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.2"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.19
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.19
        version: 42.19.0
      entries:
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.19.0
    - currentCSV: t8c-operator.v42.20.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.20
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.20
        version: 42.20.0
      entries:
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.20.0
    - currentCSV: t8c-operator.v42.21.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.21
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.21
        version: 42.21.0
      entries:
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.21.0
    - currentCSV: t8c-operator.v42.22.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.5"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.22
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.22
        version: 42.22.0
      entries:
      - name: t8c-operator.v42.22.0
        version: 42.22.0
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.22.0
    - currentCSV: t8c-operator.v42.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.4.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.6
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.6
        version: 42.6.0
      entries:
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.6.0
    - currentCSV: t8c-operator.v42.7.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.4.6"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.7
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.7
        version: 42.7.0
      entries:
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.7.0
    - currentCSV: t8c-operator.v42.8.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.0"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.8
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.8
        version: 42.8.0
      entries:
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.8.0
    - currentCSV: t8c-operator.v42.9.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.1"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.9
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.9
        version: 42.9.0
      entries:
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.9.0
    defaultChannel: stable
    packageName: t8c
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-db
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-db.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Db",
                "metadata": {
                  "name": "db-sample"
                },
                "spec": {
                  "clusterDomain": "cluster.local",
                  "databases": [
                    {
                      "name": "wapdb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "grafanadb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "brokerdb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "nimbusmosaic",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    }
                  ],
                  "image": {
                    "migrationJobImage": "wap-ose-cli",
                    "migrationJobRepository": "prod.harbor.keysight.digital/loadcore",
                    "migrationJobTag": "v4.13.0-3747-35-ocp",
                    "pgImage": "postgresql-ubi-image",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "pgTag": "14.2.0-3747-32"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "postgresql": {
                    "image": {
                      "pullSecrets": [
                        "load-core-secret"
                      ],
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "postgresql-ubi-image",
                      "tag": "14.2.0-3747-32"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "pgHbaConfiguration": "local all all trust\nhost all all 0.0.0.0/0 trust",
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "storageClass": "managed-loadcore-storage"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "volumePermissions": {
                        "containerSecurityContext": {
                          "runAsUser": "auto"
                        }
                      }
                    },
                    "serviceAccount": {
                      "automountServiceAccountToken": false,
                      "name": "load-core-service-account"
                    }
                  },
                  "storage": "8Gi"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:30:46Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Db
            name: dbs.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP DB Operator
        displayName: Keysight WAP DB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Db
          url: https://keysight-wap-db.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-db@sha256:5689d6a4da2fdd97223bc64aa61a8574bba810f93d09ab804cb45b4475f80ff5
        version: 0.1.21
      entries:
      - name: keysight-wap-db.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-db
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-certified-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "tag": "5.4.6",
                  "oem": "",
                  "imagePullSecrets": "",
                  "psp": false,
                  "rbac": true,
                  "serviceAccount": "default",
                  "leastPrivilege": false,
                  "global": {
                    "cattle": {
                      "url": ""
                    }
                  },
                  "bootstrapPassword": "",
                  "autoGenerateCert": true,
                  "defaultValidityPeriod": 365,
                  "internal": {
                    "certmanager": {
                      "enabled": false,
                      "secretname": "neuvector-internal"
                    },
                    "autoGenerateCert": true,
                    "autoRotateCert": true
                  },
                  "controller": {
                    "enabled": true,
                    "annotations": {},
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": "registry.connect.redhat.com/neuvector/controller@sha256:65ee3723c7c007db5ce3cd0493a785d4163d972cf957de6dec956b842807a551",
                    "imagePullPolicy": "IfNotPresent",
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "searchRegistries": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "nodePort": "",
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      }
                    },
                    "ranchersso": {
                      "enabled": false
                    },
                    "pvc": {
                      "enabled": false,
                      "existingClaim": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {
                        "userinitcfg.yaml": {
                          "users": [
                            {
                              "Fullname": "admin",
                              "Password": "",
                              "Role": "admin"
                            }
                          ]
                        }
                      }
                    },
                    "certupgrader": {
                      "env": [],
                      "schedule": "",
                      "imagePullPolicy": "IfNotPresent",
                      "timeout": 3600,
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "prime": {
                      "enabled": false,
                      "image": {
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "1.0.7"
                      }
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/enforcer@sha256:72a628a5c811c95bb267bbd0a5ffc9d74dbaca03a92aae615d43c7601b4d3b43",
                    "imagePullPolicy": "IfNotPresent",
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "env": [],
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/etcd"
                      }
                    ],
                    "resources": {},
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "manager": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/manager@sha256:89a42c9ee86f98ef56b5e695e23bdcc6334361ebd29cfb7729d7c86f04da98b0",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "",
                    "env": {
                      "ssl": true,
                      "envs": []
                    },
                    "svc": {
                      "type": "ClusterIP",
                      "nodePort": "",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": "",
                      "tls": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "topologySpreadConstraints": [],
                    "affinity": {},
                    "podLabels": {},
                    "podAnnotations": {},
                    "tolerations": [],
                    "nodeSelector": {},
                    "runAsUser": "",
                    "probes": {
                      "enabled": false,
                      "timeout": 1,
                      "periodSeconds": 10,
                      "startupFailureThreshold": 30
                    }
                  },
                  "cve": {
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "cacert": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                      "image": "registry.access.redhat.com/ubi8",
                      "imagePullPolicy": "IfNotPresent",
                      "schedule": "0 0 * * *",
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": "registry.connect.redhat.com/neuvector/scanner@sha256:b632f1d31998a4c8bad9f97216e8aa56cd7e6446bd70d2089d6c66845b964e8d",
                      "imagePullPolicy": "Always",
                      "priorityClassName": "",
                      "resources": {},
                      "topologySpreadConstraints": [],
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      },
                      "volumes": "",
                      "volumeMounts": ""
                    }
                  },
                  "resources": {},
                  "runtimePath": "",
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhooksvc": {
                    "enabled": true
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  },
                  "lease": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "true"
          containerImage: registry.connect.redhat.com/neuvector/neuvector-operator@sha256:e2f19c50ebc4127a83c943dbcfa1312ec46dec71bd36f6d99178b320d8da970b
          createdAt: "2020-12-02 14:19:21"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: Neuvector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm/tree/master/charts/core) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart.  NeuVector Operator versions are tied to NeuVector product versions. Version 1.5.2 of the NeuVector Certified Operator deploys version 5.4.6 of NeuVector.

          **Complete below steps to Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Privileged SCC is added to Service Account specified in the deploment yaml by Operator version 1.3.6 and above in new deployment. In case of upgrading NeuVector Operator from previous version to 1.3.6, please delete Privileged SCC before upgrading.

                   oc delete rolebinding -nneuvector system:openshift:scc:privileged

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s                    neuvector/default


          #Deploying the NeuVector Operator#


          Please refer to the instructions [here](https://github.com/neuvector/neuvector-operator/blob/master/README.md)


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Networking
        - Security
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - registry.connect.redhat.com/neuvector/manager@sha256:89a42c9ee86f98ef56b5e695e23bdcc6334361ebd29cfb7729d7c86f04da98b0
        - registry.connect.redhat.com/neuvector/scanner@sha256:b632f1d31998a4c8bad9f97216e8aa56cd7e6446bd70d2089d6c66845b964e8d
        - registry.access.redhat.com/ubi8@sha256:534c2c0efa4150ede18e3f9d7480d3b9ec2a52e62bc91cd54e08ee7336819619
        - registry.connect.redhat.com/neuvector/neuvector-operator@sha256:e2f19c50ebc4127a83c943dbcfa1312ec46dec71bd36f6d99178b320d8da970b
        - registry.connect.redhat.com/neuvector/controller@sha256:65ee3723c7c007db5ce3cd0493a785d4163d972cf957de6dec956b842807a551
        - registry.connect.redhat.com/neuvector/enforcer@sha256:72a628a5c811c95bb267bbd0a5ffc9d74dbaca03a92aae615d43c7601b4d3b43
        version: 1.5.2
      entries:
      - name: neuvector-operator.v1.5.2
        version: 1.5.2
      - name: neuvector-operator.v1.5.1
        version: 1.5.1
      - name: neuvector-operator.v1.5.0
        version: 1.5.0
      - name: neuvector-operator.v1.4.9
        version: 1.4.9
      - name: neuvector-operator.v1.4.8
        version: 1.4.8
      - name: neuvector-operator.v1.4.7
        version: 1.4.7
      - name: neuvector-operator.v1.4.6
        version: 1.4.6
      - name: neuvector-operator.v1.4.5
        version: 1.4.5
      - name: neuvector-operator.v1.4.4
        version: 1.4.4
      - name: neuvector-operator.v1.4.3
        version: 1.4.3
      - name: neuvector-operator.v1.4.2
        version: 1.4.2
      - name: neuvector-operator.v1.4.1
        version: 1.4.1
      - name: neuvector-operator.v1.4.0
        version: 1.4.0
      - name: neuvector-operator.v1.3.9
        version: 1.3.9
      - name: neuvector-operator.v1.3.8
        version: 1.3.8
      - name: neuvector-operator.v1.3.7
        version: 1.3.7
      - name: neuvector-operator.v1.3.6
        version: 1.3.6
      - name: neuvector-operator.v1.3.5
        version: 1.3.5
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-certified-operator
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: model-builder-for-vision-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-modelbuilder-for-vision.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "Modelbuilder",
                "metadata": {
                  "name": "modelbuilder-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "backup_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "metadata_storage": {
                    "storage_class": "ibmc-block-bronze",
                    "storage_size": "20G"
                  },
                  "photo_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "500G"
                  },
                  "in_memory_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              },
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "ModelbuilderAWS",
                "metadata": {
                  "name": "modelbuilder-aws-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "postgres": {
                    "storage_size": "20G",
                    "storage_class": "gp2",
                    "backup_storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
          createdAt: "2021-08-28T12:00:00Z"
          olm.skipRange: <1.0.3
          operators.operatorframework.io/builder: operator-sdk-v1.11.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/ibm-modelbuilder-for-vision/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Model Builder for Vision on AWS
            displayName: Deploy-Modelbuilder on AWS
            kind: ModelbuilderAWS
            name: modelbuildersaws.modelbuilder.com
            version: v1alpha1
          - description: Deploy Model Builder for Vision on IBM
            displayName: Deploy-Modelbuilder on IBM
            kind: Modelbuilder
            name: modelbuilders.modelbuilder.com
            version: v1alpha1
        description: IBM Services Software Model Builder for Vision is an AI training
          platform that can use the capabilities of IBM Cloud and Amazon Web Services
          and their GPUs to quickly train computer vision models for compatible mobile
          apps. IBM Services Software Inspection Workbench iOS and iPadOS app will
          be used as the exclusive UI for labeling and training. Specifically, IBM
          Model Builder for Vision will train computer vision models that can then
          be deployed for use on the IBM Services Software Inspector Portable , IBM
          Services Software Inspector Wearable and the IBM Maximo Visual Inspection
          Mobile apps.
        displayName: IBM Model Builder for Vision
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Model Builder
        - AI
        - Computer Vision
        - Inspection
        - Mobile
        - Core ML
        - TensorFlow Lite
        links:
        - name: IBM Model Builder for Vision
          url: http://ibm.biz/model_builder
        maintainers:
        - email: isshelp@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:2f65f46b25d84a3cb41a6a10f93786bcab47b7354ca9ac13a8e602ea833eba60
        - registry.connect.redhat.com/ibm/modelbuilder-image-server@sha256:ecdb665af651417c50335e7259b89ef8096521abec431976bd7abe4c1edc49d6
        - registry.connect.redhat.com/ibm/modelbuilder-cv-training-service@sha256:c3c9080b08b87022c784fd98ac3e2774ec89ce29a2dd19d91dee039e3fdb2b36
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d4c60531f59953f399c6a2b0f8239128cdc2640f72f2d1e9ac4d6add618f9f01
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.connect.redhat.com/ibm/modelbuilder-augmentor@sha256:ef5d0dde3eb334d919c2773af63007f1ef5e735ae040ef77fffe3440f854f068
        - registry.connect.redhat.com/ibm/modelbuilder-auth-service@sha256:f96b444fdcf7781775c9eba942bcc5e5ebdcf5f50820447f5eb26061c4c674cd
        - registry.connect.redhat.com/ibm/modelbuilder-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
        version: 1.0.3
      entries:
      - name: ibm-modelbuilder-for-vision.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: model-builder-for-vision-certified
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Coroot Inc.
      provider-url: https://coroot.com
    name: coroot-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: coroot-operator.1.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "coroot.com/v1",
                "kind": "Coroot",
                "metadata": {
                  "name": "coroot",
                  "namespace": "coroot"
                },
                "spec": {
                  "enterpriseEdition": {
                    "licenseKey": "<COROOT_LICENSE_KEY> from https://coroot.com/account"
                  },
                  "ingress": {
                    "host": "coroot.example.com"
                  },
                  "storage": {
                    "size": "50Gi"
                  },
                  "clickhouse": {
                    "replicas": 2,
                    "shards": 2,
                    "storage": {
                      "size": "100Gi"
                    }
                  },
                  "prometheus": {
                    "storage": {
                      "size": "100Gi"
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring, Logging & Tracing
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/coroot/coroot-operator
          support: support@coroot.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Coroot is the Schema for the coroots API
            displayName: Coroot
            kind: Coroot
            name: coroots.coroot.com
            version: v1
        description: An open-source observability platform using eBPF to gather telemetry
          data, providing actionable insights for quick issue detection and resolution
          in Kubernetes and VM environments.
        displayName: Coroot
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - coroot
        - observability
        links:
        - name: Documentation
          url: https://docs.coroot.com
        maintainers:
        - email: support@coroot.com
          name: Coroot Inc.
        maturity: stable
        provider:
          name: Coroot Inc.
          url: https://coroot.com
        relatedImages:
        - ghcr.io/coroot/kube-state-metrics@sha256:080a85e77b1e530df56f84c8e0ee2523c1f740adea152015367b87edbfac078d
        - ghcr.io/coroot/coroot-ee@sha256:09b9177ac07c2d26f4bd16050d07bb1e2132431055c32e047563083ee09194a8
        - ghcr.io/coroot/coroot-node-agent@sha256:3d552e98ca2ff8c6e125496857306172ec6f8e134ff13886c7f799abef29d1b8
        - ghcr.io/coroot/coroot-cluster-agent@sha256:674bcb147354873eaa0ee9c1328ddf758d43adbbe532c9fbbe41ddae7addc114
        - ghcr.io/coroot/clickhouse@sha256:afed1054626598371975c948e5b9e4dcaac54e8059b7ad9f0b905ef1c99f2260
        - ghcr.io/coroot/coroot-operator@sha256:13f1a4a07877c8fe8230a0bb7282ff47451007c9cf689a995a37d16697b9268a
        - ghcr.io/coroot/prometheus@sha256:e7117bd86e1088fc02ac12758bc94dba1029de54a5bc1ee04954569d8b79625d
        version: 1.1.5
      entries:
      - name: coroot-operator.1.1.5
        version: 1.1.5
      name: stable
    defaultChannel: stable
    packageName: coroot-operator
    provider:
      name: Coroot Inc.
      url: https://coroot.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cluster-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-manager.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "cluster-manager"
                },
                "spec": {
                  "addOnManagerImagePullSpec": "quay.io/open-cluster-management/addon-manager:v0.13.1",
                  "deployOption": {
                    "mode": "Default"
                  },
                  "placementImagePullSpec": "quay.io/open-cluster-management/placement:v0.13.1",
                  "registrationConfiguration": {
                    "featureGates": [
                      {
                        "feature": "DefaultClusterSet",
                        "mode": "Enable"
                      }
                    ]
                  },
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "cluster-manager"
                },
                "spec": {
                  "addOnManagerImagePullSpec": "quay.io/open-cluster-management/addon-manager:v0.13.1",
                  "deployOption": {
                    "hosted": {
                      "registrationWebhookConfiguration": {
                        "address": "management-control-plane",
                        "port": 30443
                      },
                      "workWebhookConfiguration": {
                        "address": "management-control-plane",
                        "port": 31443
                      }
                    },
                    "mode": "Hosted"
                  },
                  "placementImagePullSpec": "quay.io/open-cluster-management/placement:v0.13.1",
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/registration-operator:v0.13.1
          createdAt: "2024-04-10T15:27:49Z"
          description: Manages the installation and upgrade of the ClusterManager.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management-io/ocm
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterManager defines the configurations of controllers
              running hub that govern registration and work distribution for attached
              Klusterlets
            displayName: ClusterManager
            kind: ClusterManager
            name: clustermanagers.operator.open-cluster-management.io
            version: v1
        description: "Note: This operator will stop updating, please use the latest
          operator in OperatorHub.io https://operatorhub.io/operator/cluster-manager.
          \n\nThe Cluster Manager provides the multicluster hub, which can manage
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. This operator supports the installation and upgrade of ClusterManager.\n\n##
          How to Install\nInstall ClusterManager operator by following instructions
          in top right button `Install`. After installing the operator, create an
          instance of the ClusterManager resource to install the ClusterManager.\n\nThe
          pods of ClusterManager are created in the `open-cluster-management-hub`
          namespace:\n\n```\n$ kubectl get pods -n open-cluster-management-hub\nNAME
          \                                                      READY   STATUS    RESTARTS
          \  AGE\ncluster-manager-addon-manager-controller-7fcc67bfdf-x7hbz  1/1     Running
          \  0          61m\ncluster-manager-placement-controller-dbcfd4975-26jpw
          \      1/1     Running   0          61m\ncluster-manager-registration-controller-7949d97899-wwtg4
          \  1/1     Running   0          61m\ncluster-manager-registration-webhook-7f675bc4bb-v4rhj
          \     1/1     Running   0          61m\ncluster-manager-work-webhook-6c9d65fb95-jh97b
          \             1/1     Running   0          61m\n```\n\nThere is a ManagedCluster
          created in the Hub cluster when the managed cluster is registered to the
          Hub cluster.\n\n```\n$ kubectl get managedclusters.cluster.open-cluster-management.io\nNAME
          \                  HUB ACCEPTED   MANAGED CLUSTER URLS     JOINED   AVAILABLE
          \  AGE\ncluster-dev            true                                    True
          \    True        7h20m\ncluster-test           false          https://127.0.0.1:8443
          \                       7h25m\n```\n"
        displayName: Cluster Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open-cluster-management
        - ClusterManager
        links:
        - name: Registration Operator
          url: https://github.com/open-cluster-management-io/ocm
        maintainers:
        - email: zyin@redhat.com
          name: Zhiwei Yin
        - email: deads@redhat.com
          name: David Eads
        - email: jqiu@redhat.com
          name: Jian Qiu
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/open-cluster-management/registration-operator:v0.13.1
        version: 0.13.1
      entries:
      - name: cluster-manager.v0.13.1
        version: 0.13.1
      - name: cluster-manager.v0.13.0
        version: 0.13.0
      - name: cluster-manager.v0.12.0
        version: 0.12.0
      - name: cluster-manager.v0.11.0
        version: 0.11.0
      - name: cluster-manager.v0.10.0
        version: 0.10.0
      - name: cluster-manager.v0.9.1
        version: 0.9.1
      - name: cluster-manager.v0.8.0
        version: 0.8.0
      - name: cluster-manager.v0.7.0
        version: 0.7.0
      - name: cluster-manager.v0.6.0
        version: 0.6.0
      - name: cluster-manager.v0.5.0
        version: 0.5.0
      - name: cluster-manager.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: cluster-manager
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: ScaleOps
      provider-url: https://scaleops.com
    name: scaleops-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: scaleops-operator.v1.7.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1alpha1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops-sample"
                },
                "spec": {
                  "scaleopsToken": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: ghcr.io/scaleops-sh/operator-bundle@sha256:ee627042deefa4faceb63efd56aa6f6338c90f6c13877f29d7161efe603e02f4
          createdAt: "2024-09-19T07:04:10Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: ScaleOps
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1alpha1
        description: ScaleOps Cluster Optimization
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: vlad@scaleops.com
          name: Vladislav Shub
        - email: guy@scaleops.com
          name: Guy Baron
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - ghcr.io/scaleops-sh/operator@sha256:7de0651a76d6eb63fee81e9177df93f126906d50063d7891f12c4a72c817dc86
        - ghcr.io/scaleops-sh/kube-state-metrics/ubi@sha256:d8cb65795861ce257f08726b7a3aa60a869cc44ed152f559431191bc79e98908
        - ghcr.io/scaleops-sh/operator-bundle@sha256:ee627042deefa4faceb63efd56aa6f6338c90f6c13877f29d7161efe603e02f4
        - ghcr.io/scaleops-sh/scaleops-operator-bundle@sha256:128771d92c852f7ace78103493db079a474e06d1352f06cbde7a7d4666684610
        - ghcr.io/scaleops-sh/configmap-reload/ubi@sha256:fed19f1aad43228560912abc36372a6c124d9ddf5e3e39476269dc51604d26de
        - ghcr.io/scaleops-sh/prometheus/ubi@sha256:8aeeae2a33251493d4d6e2f02a2fff1d4c2f32ab54554e6a84619ecd8b4aa69a
        version: 0.0.1
      entries:
      - name: scaleops-operator.v1.7.5
        version: 0.0.1
      name: alpha
    - currentCSV: scaleops-operator.v1.21.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops",
                  "namespace": "scaleops-system"
                },
                "spec": {
                  "scaleopsToken": "${SCALEOPS_TOKEN}"
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: registry.connect.redhat.com/scaleops-sh/operator@sha256:528f2e09d2dad4f48e01a272b51baa1753c12069515831223e3f97c7dfe03e3c
          createdAt: "2025-09-29T13:35:24Z"
          description: |-
            ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

            ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

            Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

            Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

            Contact [us](mailto:team@scaleops.com) to start your free trial today
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.21.8
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://scaleops.com
          support: ScaleOps Support support@scaleops.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This will install an instance of the ScaleOps Platform on
              your cluster
            displayName: ScaleOps Platform
            kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1
        description: |-
          ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

          ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

          Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

          Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

          Contact [us](mailto:team@scaleops.com) to start your free trial today
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        - application runtime
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: support@scaleops.com
          name: ScaleOps
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - registry.connect.redhat.com/scaleops-sh/operator@sha256:528f2e09d2dad4f48e01a272b51baa1753c12069515831223e3f97c7dfe03e3c
        - registry.connect.redhat.com/scaleops-sh/scaleops-sh@sha256:57d5d06f4f4b5470562c4f24ea5d37203633e1f4b65cd497e71d9066fd2b9e41
        - registry.connect.redhat.com/scaleops-sh/scaleops-gpu@sha256:724aef7683ea4cb245ab0a464517a6d542efa4634e5469ff35b7a7b7f00b563e
        version: 1.21.8
      entries:
      - name: scaleops-operator.v1.21.8
        version: 1.21.8
      - name: scaleops-operator.v1.21.7
        version: 1.21.7
      - name: scaleops-operator.v1.21.6
        version: 1.21.6
      - name: scaleops-operator.v1.21.5
        version: 1.21.5
      - name: scaleops-operator.v1.21.3
        version: 1.21.3
      - name: scaleops-operator.v1.21.2
        version: 1.21.2
      - name: scaleops-operator.v1.21.1
        version: 1.21.1
      - name: scaleops-operator.v1.21.0
        version: 1.21.0
      - name: scaleops-operator.v1.20.9
        version: 1.20.9
      - name: scaleops-operator.v1.20.8
        version: 1.20.8
      - name: scaleops-operator.v1.20.7
        version: 1.20.7
      - name: scaleops-operator.v1.20.6
        version: 1.20.6
      - name: scaleops-operator.v1.20.5
        version: 1.20.5
      - name: scaleops-operator.v1.20.4
        version: 1.20.4
      - name: scaleops-operator.v1.20.2
        version: 1.20.2
      - name: scaleops-operator.v1.20.1
        version: 1.20.1
      - name: scaleops-operator.v1.19.8
        version: 1.19.8
      - name: scaleops-operator.v1.19.7
        version: 1.19.7
      - name: scaleops-operator.v1.19.6
        version: 1.19.6
      - name: scaleops-operator.v1.19.5
        version: 1.19.5
      - name: scaleops-operator.v1.19.4
        version: 1.19.4
      - name: scaleops-operator.v1.19.1
        version: 1.19.1
      - name: scaleops-operator.v1.19.0
        version: 1.19.0
      - name: scaleops-operator.v1.18.8
        version: 1.18.8
      - name: scaleops-operator.v1.18.7
        version: 1.18.7
      - name: scaleops-operator.v1.18.6
        version: 1.18.6
      - name: scaleops-operator.v1.18.5
        version: 1.18.5
      - name: scaleops-operator.v1.18.4
        version: 1.18.4
      - name: scaleops-operator.v1.18.3
        version: 1.18.3
      - name: scaleops-operator.v1.18.2
        version: 1.18.2
      name: stable
    - currentCSV: scaleops-operator.v1.10.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops",
                  "namespace": "scaleops-system"
                },
                "spec": {
                  "scaleopsToken": "${SCALEOPS_TOKEN}"
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: registry.connect.redhat.com/scaleops-sh/operator@sha256:8def662a546babd6740e5417401836d647d92d52b15d559b1b3e2cebd7f77f33
          createdAt: "2025-02-12T19:06:04Z"
          description: |-
            ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

            ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

            Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

            Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

            Contact [us](mailto:team@scaleops.com) to start your free trial today
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.10.32
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://scaleops.com
          support: ScaleOps Support support@scaleops.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This will install an instance of the ScaleOps Platform on
              your cluster
            displayName: ScaleOps Platform
            kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1
        description: |-
          ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

          ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

          Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

          Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

          Contact [us](mailto:team@scaleops.com) to start your free trial today
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        - application runtime
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: support@scaleops.com
          name: ScaleOps
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - registry.connect.redhat.com/scaleops-sh/operator@sha256:8def662a546babd6740e5417401836d647d92d52b15d559b1b3e2cebd7f77f33
        - registry.connect.redhat.com/scaleops-sh/scaleops-sh@sha256:5d77e81d7959483121c208072eb61b063eb3ad9abc49e7df3c7efb8e2b77a3ae
        version: 1.10.32
      entries:
      - name: scaleops-operator.v1.10.32
        version: 1.10.32
      name: stable-1.10
    - currentCSV: scaleops-operator.v1.12.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops",
                  "namespace": "scaleops-system"
                },
                "spec": {
                  "scaleopsToken": "${SCALEOPS_TOKEN}"
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: registry.connect.redhat.com/scaleops-sh/operator@sha256:4bd74a56a8680f47aa1da7a377f74529fe7ca1568995a4e74018af0cf67763ed
          createdAt: "2025-03-11T18:10:07Z"
          description: |-
            ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

            ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

            Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

            Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

            Contact [us](mailto:team@scaleops.com) to start your free trial today
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.12.6
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://scaleops.com
          support: ScaleOps Support support@scaleops.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This will install an instance of the ScaleOps Platform on
              your cluster
            displayName: ScaleOps Platform
            kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1
        description: |-
          ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

          ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

          Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

          Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

          Contact [us](mailto:team@scaleops.com) to start your free trial today
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        - application runtime
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: support@scaleops.com
          name: ScaleOps
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - registry.connect.redhat.com/scaleops-sh/operator@sha256:4bd74a56a8680f47aa1da7a377f74529fe7ca1568995a4e74018af0cf67763ed
        - registry.connect.redhat.com/scaleops-sh/scaleops-sh@sha256:5ea4e26f2fecee3570d63adc3e2588d34f9610345d703975d4febac9c94b4bd0
        version: 1.12.6
      entries:
      - name: scaleops-operator.v1.12.6
        version: 1.12.6
      name: stable-1.12
    - currentCSV: scaleops-operator.v1.14.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops",
                  "namespace": "scaleops-system"
                },
                "spec": {
                  "scaleopsToken": "${SCALEOPS_TOKEN}"
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: registry.connect.redhat.com/scaleops-sh/operator@sha256:9267f49ce774ee4d401458531599bbd6729c6313cecc4f2bcd3bf0b5eecd92e1
          createdAt: "2025-06-08T16:15:46Z"
          description: |-
            ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

            ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

            Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

            Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

            Contact [us](mailto:team@scaleops.com) to start your free trial today
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.14.14
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://scaleops.com
          support: ScaleOps Support support@scaleops.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This will install an instance of the ScaleOps Platform on
              your cluster
            displayName: ScaleOps Platform
            kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1
        description: |-
          ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

          ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

          Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

          Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

          Contact [us](mailto:team@scaleops.com) to start your free trial today
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        - application runtime
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: support@scaleops.com
          name: ScaleOps
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - registry.connect.redhat.com/scaleops-sh/operator@sha256:9267f49ce774ee4d401458531599bbd6729c6313cecc4f2bcd3bf0b5eecd92e1
        - registry.connect.redhat.com/scaleops-sh/scaleops-sh@sha256:9c6395202ccb70e8edae3a1b540de31991ec9232b7b235dd92755588b8491720
        version: 1.14.14
      entries:
      - name: scaleops-operator.v1.14.14
        version: 1.14.14
      name: stable-1.14
    defaultChannel: stable
    packageName: scaleops-operator
    provider:
      name: ScaleOps
      url: https://scaleops.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataStax
      provider-url: ""
    name: cass-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cass-operator.v1.23.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cassandra.datastax.com/v1beta1",
                "kind": "CassandraDatacenter",
                "metadata": {
                  "name": "dc1"
                },
                "spec": {
                  "clusterName": "development",
                  "config": {
                    "cassandra-yaml": {
                      "authenticator": "PasswordAuthenticator",
                      "authorizer": "CassandraAuthorizer",
                      "num_tokens": 16,
                      "role_manager": "CassandraRoleManager"
                    },
                    "jvm-server-options": {
                      "initial_heap_size": "1G",
                      "max_heap_size": "1G"
                    }
                  },
                  "managementApiAuth": {
                    "insecure": {}
                  },
                  "podTemplateSpec": {
                    "spec": {
                      "containers": [
                        {
                          "name": "cassandra"
                        }
                      ],
                      "securityContext": {}
                    }
                  },
                  "racks": [
                    {
                      "name": "rack1"
                    },
                    {
                      "name": "rack2"
                    },
                    {
                      "name": "rack3"
                    }
                  ],
                  "resources": {
                    "requests": {
                      "cpu": "1000m",
                      "memory": "2Gi"
                    }
                  },
                  "serverType": "cassandra",
                  "serverVersion": "4.1.2",
                  "size": 3,
                  "storageConfig": {
                    "cassandraDataVolumeClaimSpec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "server-storage"
                    }
                  }
                }
              },
              {
                "apiVersion": "control.k8ssandra.io/v1alpha1",
                "kind": "CassandraTask",
                "metadata": {
                  "name": "example-cleanup"
                },
                "spec": {
                  "datacenter": {
                    "name": "dc2",
                    "namespace": "cass-operator"
                  },
                  "jobs": [
                    {
                      "command": "cleanup",
                      "name": "cleanup-run"
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database,Big Data,OpenShift Optional
          certified: "true"
          containerImage: registry.connect.redhat.com/datastax/cass-operator@sha256:6abc6253b50a0d7f6bdd4b394d2bf96320a8845952aeb8f0bbed99863192ec52
          createdAt: "2025-03-20"
          description: |
            Simple provisioning, turn-key operations, and automated remediation of Apache Cassandra clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/k8ssandra/cass-operator
          support: DataStax
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CassandraDatacenter is the Schema for the cassandradatacenters
              API
            displayName: Cassandra Datacenter
            kind: CassandraDatacenter
            name: cassandradatacenters.cassandra.datastax.com
            version: v1beta1
          - description: CassandraTask is the Schema for the cassandrajobs API
            displayName: Cassandra Task
            kind: CassandraTask
            name: cassandratasks.control.k8ssandra.io
            version: v1alpha1
        description: "## Apache Cassandra\n\nThe Apache Cassandra database is the
          right choice when you need scalability and\nhigh availability without compromising
          performance. Linear scalability and\nproven fault-tolerance on commodity
          hardware or cloud infrastructure make it the\nperfect platform for mission-critical
          data. Cassandra's support for replicating\nacross multiple datacenters is
          best-in-class, providing lower latency for your\nusers and the peace of
          mind of knowing that you can survive regional outages.\n\n## DataStax Enterprise\n\nThe
          most advanced distribution of Apache Cassandra™ on the market, with the\nenterprise
          functionality needed for serious production systems and backed up and\nsupported
          by the best distributed-experts in the world. It's one platform for\nall
          types of applications anywhere, any cloud, any model: key-value, graph,\ntabular,
          JSON.\n\nDataStax Enterprise is a fully integrated and optimized database,
          with graph,\nanalytics, and search included, all with a unified security
          model. Simply put,\nit's the only database capable of meeting today's demanding
          requirements\n\n## Operator Details\n\n`cass-operator` is designed as a
          modular operator for Apache Cassandra and\nderived  distributions. Apache
          Cassandra is a distributed database consisting of\nmultiple nodes working
          in concert to store data and process queries along a\nnumber of fault domains.
          `cass-operator` has the deployment of a Cassandra\ncluster around the logical
          domain of a datacenter with the `CassandraDatacenter`\ncustom resource.\n
          \   \nUpon submission of one of these resources it handles provisioning
          the underlying\nstateful sets (analogous to C\\* logical racks), services,
          and configuration.\nAdditionally through monitoring pod state via Kubernetes
          callbacks it handles day to day\noperations such as restarting failed processes,
          scaling clusters up, and deploying\nconfiguration changes in a rolling,
          non-disruptive, fashion.\n    \nThis operator is designed to be `Namespace`
          scoped. A single Kubernetes cluster may\nbe running multiple instances of
          this operator, in separate namespaces, to support\na number of C\\* clusters
          and environments. Configuration is simple with the usage of\nYAML based
          overrides in the Custom Resource paired with an `init` container.\n    \nIn
          C\\* clusters ordering and timing of certain operations are important to
          keep the system\nevenly distributed. `cass-operator` takes advantage of
          a sidecar process within the\nmain container to handle the orchestration
          of starting our main C* process."
        displayName: DataStax Kubernetes Operator for Apache Cassandra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - k8ssandra
        - cass-operator
        - DataStax Enterprise
        - Apache Cassandra
        - DSE
        - C*
        - Cassandra
        links:
        - name: DataStax
          url: https://datastax.com
        - name: Operator GitHub Repo
          url: https://github.com/k8ssandra/cass-operator
        - name: Config Builder GitHub Repo
          url: https://github.com/datastax/cass-config-builder
        - name: Documentation [1]
          url: https://github.com/k8ssandra/cass-operator/tree/master/docs/user
        - name: Documentation [2]
          url: https://docs.datastax.com/en/cass-operator/doc/cass-operator/cassOperatorTOC.html
        maintainers:
        - email: cass-operator@datastax.com
          name: DataStax Operator Team
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: DataStax
        relatedImages:
        - registry.connect.redhat.com/datastax/cass-operator@sha256:6abc6253b50a0d7f6bdd4b394d2bf96320a8845952aeb8f0bbed99863192ec52
        - registry.connect.redhat.com/datastax/system-logger@sha256:3408620786e6265aa835be56f85aadc48b20060b2038dc21126f614685f77c6b
        version: 1.23.2
      entries:
      - name: cass-operator.v1.23.2
        version: 1.23.2
      name: stable
    defaultChannel: stable
    packageName: cass-operator
    provider:
      name: DataStax
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
    name: aikit-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aikit-operator.v2021.2.102120
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aikit.intel/v1alpha1",
                "kind": "AIKitContainer",
                "metadata": {
                  "name": "intel-aikit-container",
                  "namespace": "redhat-ods-applications"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagestream": {
                    "name": "oneapi-aikit",
                    "namespace": "redhat-ods-applications",
                    "registry": {
                      "root": "registry.connect.redhat.com",
                      "repo": "intel",
                      "name": "oneapi-aikit",
                      "version": "latest"
                    }
                  },
                  "nameOverride": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aikit-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aikit-operator-rhmp/support?utm_source=openshift_console
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AIKitContainer
            name: aikitcontainers.aikit.intel
            version: v1alpha1
        description: |
          The Intel® oneAPI AI Analytics Toolkit gives data scientists, AI developers, and researchers familiar Python* tools and frameworks to accelerate end-to-end data science and machine learning pipelines on Intel® architectures. The components are built using oneAPI libraries for low-level compute optimizations. This toolkit maximizes performance from preprocessing through machine and deep learning phases and provides interoperability for efficient model development and deployment across single and multinodes.
        displayName: Intel® oneAPI AI Analytics Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aikit-operator
        links:
        - name: Aikit Operator
          url: https://aikit-operator.domain
        maintainers:
        - email: kam.d.kasravi@intel.com
          name: Kam Kasravi
        - email: abolfazl.shahbazi@intel.com
          name: Abolfazl Shahbazi
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Intel
          url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
        relatedImages:
        - registry.marketplace.redhat.com/rhm/intel/aikit-operator@sha256:7abb5e6164ffb3e4b87c39061f45d09848218fcfb74b48bf5e1035d56b41ee0c
        version: 2021.2.102120
      entries:
      - name: aikit-operator.v2021.2.102120
        version: 2021.2.102120
      - name: aikit-operator.v2021.2.100410
        version: 2021.2.100410
      - name: aikit-operator.v2021.2.92911
        version: 2021.2.92911
      - name: aikit-operator.v2021.2.92814
        version: 2021.2.92814
      - name: aikit-operator.v2021.2.92808-dev-a60e5ba
        version: 2021.2.92808-dev-a60e5ba
      name: alpha
    defaultChannel: alpha
    packageName: aikit-operator-rhmp
    provider:
      name: Intel
      url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Apache Software Foundation
      provider-url: ""
    name: camel-karavan-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: camel-karavan-operator.v3.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Karavan",
                "metadata": {
                  "labels": {
                    "app": "karavan"
                  },
                  "name": "karavan"
                },
                "spec": {
                  "instances": 1,
                  "auth": "public",
                  "environment": "dev",
                  "runtimes": "quarkus,spring-boot",
                  "gitPullInterval": "10s"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/apache/camel-karavan-operator:3.20.1
          repository: https://github.com/apache/camel-karavan
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Karavan
            name: karavans.camel.apache.org
            version: v1
        description: |
          Apache Camel Karavan
          ====================
          Karavan is an Integration Toolkit for Apache Camel, which makes integration easy and fun through the visualization of pipelines,
          integration with runtimes and package, image build and deploy to kubernetes out-of-the-box.

          ## Installation

          1. Install Apache Camel Karavan Operator
          2. Install Tekton operator
          3. Create namespace, ex: `karavan`
          4. Create `Secret` in the namespace, ex:
          ```
          kind: Secret
          apiVersion: v1
          metadata:
            name: karavan
            namespace: karavan
          type: Opaque
          stringData:
            master-password: karavan
            oidc-secret: secret
            oidc-server-url: https://hostname/auth/realms/karavan
            oidc-frontend-url: https://hostname/auth
            git-repository: https://github.com/repository/projects.git
            git-password: password
            git-username: username
            git-branch: main
            image-registry: image-registry.openshift-image-registry.svc:5000
          ```
          5. Create `Karavan` instance in the namespace, ex:
          ```
          apiVersion: camel.apache.org/v1
          kind: Karavan
          metadata:
            name: karavan
            namespace: karavan
          spec:
            instances: 1
            auth: public
            environment: dev
            runtimes: quarkus,spring-boot
            gitPullInterval: 10s
          ```
        displayName: Camel Karavan Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - camel
        - karavan
        - integration
        - microservices
        - low-code
        links:
        - name: Apache Camel Karavan source code repository
          url: https://github.com/apache/camel-karavan
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - ghcr.io/apache/camel-karavan-operator:3.20.1
        version: 3.20.1
      entries:
      - name: camel-karavan-operator.v3.20.1
        version: 3.20.1
      - name: camel-karavan-operator.v3.20.0
        version: 3.20.0
      - name: camel-karavan-operator.v3.18.5
        version: 3.18.5
      name: alpha
    defaultChannel: alpha
    packageName: camel-karavan-operator
    provider:
      name: The Apache Software Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: cloudnative-pg
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudnative-pg.v1.27.0-rc1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: ghcr.io/cloudnative-pg/cloudnative-pg@sha256:c2212e806e950ed5b667455f476481aef179c8e467e435352b624dc670000443
          createdAt: "2025-07-31T12:14:29Z"
          description: CloudNativePG is an open source operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/cloudnative-pg/cloudnative-pg
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.cnpg.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.cnpg.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.cnpg.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.cnpg.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.cnpg.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.cnpg.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.cnpg.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnpg` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: CloudNativePG
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnpg
        links:
        - name: CloudNativePG
          url: https://cloudnative-pg.io/
        - name: Documentation
          url: https://cloudnative-pg.io/docs/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: The CloudNativePG Contributors
        relatedImages:
        - ghcr.io/cloudnative-pg/cloudnative-pg@sha256:c2212e806e950ed5b667455f476481aef179c8e467e435352b624dc670000443
        - ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b
        - ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891
        version: 1.27.0-rc1
      entries:
      - name: cloudnative-pg.v1.27.0-rc1
        version: 1.27.0-rc1
      - name: cloudnative-pg.v1.26.0-rc3
        version: 1.26.0-rc3
      - name: cloudnative-pg.v1.26.0-rc2
        version: 1.26.0-rc2
      - name: cloudnative-pg.v1.26.0-rc1
        version: 1.26.0-rc1
      name: candidate
    - currentCSV: cloudnative-pg.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: ghcr.io/cloudnative-pg/cloudnative-pg@sha256:53995e9d6a7760c77d77e5298f78b49029427494f4dd933d5fa7a74f7af3bb44
          createdAt: "2025-08-12T14:08:35Z"
          description: CloudNativePG is an open source operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/cloudnative-pg/cloudnative-pg
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.cnpg.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.cnpg.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.cnpg.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.cnpg.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.cnpg.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.cnpg.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.cnpg.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnpg` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: CloudNativePG
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnpg
        links:
        - name: CloudNativePG
          url: https://cloudnative-pg.io/
        - name: Documentation
          url: https://cloudnative-pg.io/docs/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - ghcr.io/cloudnative-pg/cloudnative-pg@sha256:53995e9d6a7760c77d77e5298f78b49029427494f4dd933d5fa7a74f7af3bb44
        - ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b
        - ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891
        version: 1.27.0
      entries:
      - name: cloudnative-pg.v1.27.0
        version: 1.27.0
      - name: cloudnative-pg.v1.26.1
        version: 1.26.1
      - name: cloudnative-pg.v1.26.0
        version: 1.26.0
      - name: cloudnative-pg.v1.25.1
        version: 1.25.1
      - name: cloudnative-pg.v1.25.0
        version: 1.25.0
      - name: cloudnative-pg.v1.24.1
        version: 1.24.1
      - name: cloudnative-pg.v1.24.0
        version: 1.24.0
      - name: cloudnative-pg.v1.23.3
        version: 1.23.3
      - name: cloudnative-pg.v1.23.2
        version: 1.23.2
      - name: cloudnative-pg.v1.23.1
        version: 1.23.1
      name: stable-v1
    defaultChannel: stable-v1
    packageName: cloudnative-pg
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cincinnati-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: update-service-operator.v5.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "updateservice.operator.openshift.io/v1",
                "kind": "UpdateService",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          description: Creates and maintains an OpenShift Update Service instance
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-update-service
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: UpdateService is the Schema for the updateservices API.
            displayName: Update Service
            kind: UpdateService
            name: updateservices.updateservice.operator.openshift.io
            version: v1
        description: |-
          # Use Case
          Running an Update Service instance in a cluster is appealing for offline OpenShift
          clusters or for admins that want to provide their own graph data instead of
          using [https://github.com/openshift/cincinnati-graph-data](https://github.com/openshift/cincinnati-graph-data/).

          # About Update Service
          Update Service uses **Cincinnati** which is an update protocol designed to facilitate
          automatic updates. It describes a particular method for representing transitions between
          releases of a project and allowing a client to perform automatic updates between these
          releases.

          ## Components
          A **Graph Builder** iterates over the release payloads hosted by the
          storage component and builds a DAG of the releases. It is responsible for
          verifying that the graph described by the releases is acyclic and connected.

          A **Policy Engine** is in charge of altering a client's view of the graph
          by applying a set of filters which are defined within the particular Policy
          Engine instance. Both the input to and the output from Policy Engines is a
          graph, allowing multiple Policy Engines to be chained together. The first
          Policy Engine in a chain will fetch its graph from the Graph Builder and the
          last Policy Engine in a chain will serve the modified graph to the client.

          An **Update Service client** is the end consumer of the release payloads. The
          client periodically queries the Policy Engine for updates and applys them if
          available.

          # Query OpenShift's Update Service Endpoint
          $ curl --silent --header 'Accept:application/json' 'https://api.openshift.com/api/upgrades_info/v1/graph?arch=amd64&channel=stable-4.2' | jq '. as $graph | $graph.nodes | map(.version == "4.2.13") | index(true) as $orig | $graph.edges | map(select(.[0] == $orig)[1]) | map($graph.nodes[.])'
        displayName: OpenShift Update Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Update Service
        - UpdateService
        - update-service
        - OSUS
        - Cincinnati
        - version
        - upgrade
        links:
        - name: Update Service Operator
          url: https://github.com/openshift/cincinnati-operator
        - name: OpenShift Update Service
          url: https://github.com/openshift/cincinnati
        - name: OpenShift Graph Data
          url: https://github.com/openshift/cincinnati-graph-data
        maintainers:
        - email: aos-team-ota@redhat.com
          name: OpenShift Update Service maintainers
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-update-service/openshift-update-service-rhel8-operator@sha256:759f4de107a5b429933c5e44046ffb4ed75c674aea75440513a6df8fa93a83a3
        - registry.redhat.io/openshift-update-service/openshift-update-service-rhel8@sha256:a9468a662babb272917df92bc9c827535b4b37883edb440b6d2d7c32ffb2145e
        version: 5.0.3
      entries:
      - name: update-service-operator.v5.0.3
        version: 5.0.3
      - name: update-service-operator.v5.0.2
        version: 5.0.2
      - name: update-service-operator.v5.0.1
        version: 5.0.1
      - name: update-service-operator.v5.0.0
        version: 5.0.0
      - name: update-service-operator.v4.9.1
        version: 4.9.1
      - name: update-service-operator.v4.9.0
        version: 4.9.0
      name: v1
    defaultChannel: v1
    packageName: cincinnati-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: www.redhat.com
    name: dpu-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dpu-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.openshift.io/v1",
                "kind": "DpuOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dpu-operator",
                    "app.kubernetes.io/instance": "dpuoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dpuoperatorconfig",
                    "app.kubernetes.io/part-of": "dpu-operator"
                  },
                  "name": "dpuoperatorconfig-sample"
                },
                "spec": {
                  "mode": "host"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-13T13:42:44Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: dpu-operator-system
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DpuOperatorConfig is the Schema for the dpuoperatorconfigs
              API
            displayName: Dpu Operator Config
            kind: DpuOperatorConfig
            name: dpuoperatorconfigs.config.openshift.io
            version: v1
        description: This operator will manage and configure data processing unit
          (DPUs) to be used in accelerating/offloading k8s networking functions
        displayName: DPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dpu
        - dpu-operator
        - ipu
        links:
        - name: Dpu Operator
          url: https://github.com/openshift/dpu-operator
        maintainers:
        - email: wizhao@redhat.com
          name: William Zhao
        - email: bnemeth@redhat.com
          name: Balazs Nemeth
        maturity: alpha
        provider:
          name: Red Hat
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-dpu-rhel9-operator@sha256:9d8283238af545c332781b2edef6b3a2857bcf4a264c9bae26720cc48e338a4b
        - registry.redhat.io/openshift4/ose-dpu-daemon-rhel9@sha256:a9a375f70dfb4a730aca0f7e58619fb05ef226261de6980d04eeffb8a043e6a4
        version: 4.16.0-202509240123
      entries:
      - name: dpu-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: dpu-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: dpu-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: dpu-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: dpu-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: dpu-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: dpu-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: dpu-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: dpu-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: dpu-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: dpu-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: dpu-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: dpu-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: dpu-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: dpu-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: dpu-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: dpu-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: dpu-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: dpu-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: dpu-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: dpu-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: dpu-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: dpu-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: dpu-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: dpu-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: dpu-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: dpu-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: dpu-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: dpu-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      name: stable
    defaultChannel: stable
    packageName: dpu-operator
    provider:
      name: Red Hat
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: odr-cluster-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odr-cluster-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "VolumeReplicationGroup",
                "metadata": {
                  "name": "volumereplicationgroup-sample"
                },
                "spec": {
                  "async": {
                    "schedulingInterval": "10m"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ],
                        "name": "cpd"
                      },
                      {
                        "includedResources": [
                          "Deployment"
                        ],
                        "name": "deployments"
                      },
                      {
                        "excludedResources": [
                          ""
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includeClusterResources": true,
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "cpd",
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      },
                      {
                        "backupName": "deployments",
                        "includedResources": [
                          "Deployment"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "excludedResources": [
                          "ConfigMap",
                          "Secret",
                          "Deployment",
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      }
                    ]
                  },
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  },
                  "replicationState": "primary",
                  "s3Profiles": [
                    "s3-profile-of-east",
                    "s3-profile-of-west"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: MaintenanceMode
            name: maintenancemodes.ramendr.openshift.io
            version: v1alpha1
          - kind: ProtectedVolumeReplicationGroupList
            name: protectedvolumereplicationgrouplists.ramendr.openshift.io
            version: v1alpha1
          - description: VolumeReplicationGroup is the Schema for the volumereplicationgroups
              API
            displayName: Volume Replication Group
            kind: VolumeReplicationGroup
            name: volumereplicationgroups.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Cluster is a disaster-recovery orchestrator for
          stateful applications, that operates from an Advanced Cluster Management
          (ACM) managed cluster and is controlled by Openshift DR Hub operator to
          orchestrate the life-cycle of an application, and its state on the managed
          cluster.
        displayName: Openshift DR Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:769acf15d016d4992793c9c527d7c268d2a67502dd7e0fadc240e4896398195b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 4.15.20-rhodf
      entries:
      - name: odr-cluster-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odr-cluster-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odr-cluster-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odr-cluster-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odr-cluster-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odr-cluster-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odr-cluster-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odr-cluster-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odr-cluster-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odr-cluster-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odr-cluster-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odr-cluster-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odr-cluster-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odr-cluster-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odr-cluster-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odr-cluster-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odr-cluster-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odr-cluster-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odr-cluster-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odr-cluster-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odr-cluster-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "VolumeReplicationGroup",
                "metadata": {
                  "name": "volumereplicationgroup-sample"
                },
                "spec": {
                  "async": {
                    "schedulingInterval": "10m"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ],
                        "name": "cpd"
                      },
                      {
                        "includedResources": [
                          "Deployment"
                        ],
                        "name": "deployments"
                      },
                      {
                        "excludedResources": [
                          ""
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includeClusterResources": true,
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "cpd",
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      },
                      {
                        "backupName": "deployments",
                        "includedResources": [
                          "Deployment"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "excludedResources": [
                          "ConfigMap",
                          "Secret",
                          "Deployment",
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      }
                    ]
                  },
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  },
                  "replicationState": "primary",
                  "s3Profiles": [
                    "s3-profile-of-east",
                    "s3-profile-of-west"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: MaintenanceMode
            name: maintenancemodes.ramendr.openshift.io
            version: v1alpha1
          - kind: ProtectedVolumeReplicationGroupList
            name: protectedvolumereplicationgrouplists.ramendr.openshift.io
            version: v1alpha1
          - description: VolumeReplicationGroup is the Schema for the volumereplicationgroups
              API
            displayName: Volume Replication Group
            kind: VolumeReplicationGroup
            name: volumereplicationgroups.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Cluster is a disaster-recovery orchestrator for
          stateful applications, that operates from an Advanced Cluster Management
          (ACM) managed cluster and is controlled by Openshift DR Hub operator to
          orchestrate the life-cycle of an application, and its state on the managed
          cluster.
        displayName: Openshift DR Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:af262b0890300183db1766bc392017b0bda723a9ba841dc25e3d928aa413db9a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        version: 4.16.16-rhodf
      entries:
      - name: odr-cluster-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odr-cluster-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odr-cluster-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odr-cluster-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odr-cluster-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odr-cluster-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odr-cluster-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odr-cluster-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odr-cluster-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odr-cluster-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odr-cluster-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odr-cluster-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odr-cluster-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odr-cluster-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odr-cluster-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odr-cluster-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odr-cluster-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odr-cluster-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: topology-aware-lifecycle-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: topology-aware-lifecycle-manager.v4.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ran.openshift.io/v1alpha1",
                "kind": "ClusterGroupUpgrade",
                "metadata": {
                  "name": "ClusterGroupUpgrade-sample"
                },
                "spec": {
                  "preCaching": true
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:bf24cad2940a227f5bd2f188be5d298ff19176c634fcf609c7225d7e0a6f2fd1
          description: The topology aware lifecycle manager is an operator that facilitates
            platform upgrades of group of clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-operators",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/cluster-group-upgrades-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterGroupUpgrade is the Schema for the ClusterGroupUpgrades
              API
            displayName: Cluster Group Upgrade
            kind: ClusterGroupUpgrade
            name: clustergroupupgrades.ran.openshift.io
            version: v1alpha1
          - description: ImageBasedGroupUpgrade is the schema for upgrading a group
              of clusters using IBU
            displayName: Image-Based Group Upgrade
            kind: ImageBasedGroupUpgrade
            name: imagebasedgroupupgrades.lcm.openshift.io
            version: v1alpha1
          - description: PreCachingConfig is the Schema for the precachingconfigs
              API
            displayName: Pre-caching Config
            kind: PreCachingConfig
            name: precachingconfigs.ran.openshift.io
            version: v1alpha1
        description: |-
          Topology Aware Lifecycle Manager is an operator that facilitates
          platform and operator upgrades of group of clusters
        displayName: Topology Aware Lifecycle Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm lifecyclemanagement upgrades cluster
        links:
        - name: Cluster Group Upgrades Operator
          url: https://cluster-group-upgrades-operator.domain
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:eed8234b0fe47083ab941f513709f778bfcb80e32747b71a65fd5f62328a5ca0
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:bf24cad2940a227f5bd2f188be5d298ff19176c634fcf609c7225d7e0a6f2fd1
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-aztp-rhel9@sha256:f6a37a0de10c847a53e22406990858dbaa4cdd3be075bc047e8dcf5bc6a2f2f7
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-precache-rhel9@sha256:7673a0d7f8944b245124376f2153bf987299f5bd7350b33f0d4bd1ac18543929
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-recovery-rhel9@sha256:37e234f247c841cd12c4617f92bf8bd5a5dbcbf46fa1be042d56b966994efaad
        version: 4.16.5
      entries:
      - name: topology-aware-lifecycle-manager.v4.16.5
        version: 4.16.5
      - name: topology-aware-lifecycle-manager.v4.16.4
        version: 4.16.4
      - name: topology-aware-lifecycle-manager.v4.16.3
        version: 4.16.3
      - name: topology-aware-lifecycle-manager.v4.16.2
        version: 4.16.2
      - name: topology-aware-lifecycle-manager.v4.16.1
        version: 4.16.1
      - name: topology-aware-lifecycle-manager.v4.16.0
        version: 4.16.0
      name: "4.16"
    - currentCSV: topology-aware-lifecycle-manager.v4.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ran.openshift.io/v1alpha1",
                "kind": "ClusterGroupUpgrade",
                "metadata": {
                  "name": "ClusterGroupUpgrade-sample"
                },
                "spec": {
                  "preCaching": true
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:bf24cad2940a227f5bd2f188be5d298ff19176c634fcf609c7225d7e0a6f2fd1
          description: The topology aware lifecycle manager is an operator that facilitates
            platform upgrades of group of clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-operators",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/cluster-group-upgrades-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterGroupUpgrade is the Schema for the ClusterGroupUpgrades
              API
            displayName: Cluster Group Upgrade
            kind: ClusterGroupUpgrade
            name: clustergroupupgrades.ran.openshift.io
            version: v1alpha1
          - description: ImageBasedGroupUpgrade is the schema for upgrading a group
              of clusters using IBU
            displayName: Image-Based Group Upgrade
            kind: ImageBasedGroupUpgrade
            name: imagebasedgroupupgrades.lcm.openshift.io
            version: v1alpha1
          - description: PreCachingConfig is the Schema for the precachingconfigs
              API
            displayName: Pre-caching Config
            kind: PreCachingConfig
            name: precachingconfigs.ran.openshift.io
            version: v1alpha1
        description: |-
          Topology Aware Lifecycle Manager is an operator that facilitates
          platform and operator upgrades of group of clusters
        displayName: Topology Aware Lifecycle Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm lifecyclemanagement upgrades cluster
        links:
        - name: Cluster Group Upgrades Operator
          url: https://cluster-group-upgrades-operator.domain
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:eed8234b0fe47083ab941f513709f778bfcb80e32747b71a65fd5f62328a5ca0
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:bf24cad2940a227f5bd2f188be5d298ff19176c634fcf609c7225d7e0a6f2fd1
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-aztp-rhel9@sha256:f6a37a0de10c847a53e22406990858dbaa4cdd3be075bc047e8dcf5bc6a2f2f7
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-precache-rhel9@sha256:7673a0d7f8944b245124376f2153bf987299f5bd7350b33f0d4bd1ac18543929
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-recovery-rhel9@sha256:37e234f247c841cd12c4617f92bf8bd5a5dbcbf46fa1be042d56b966994efaad
        version: 4.16.5
      entries:
      - name: topology-aware-lifecycle-manager.v4.16.5
        version: 4.16.5
      - name: topology-aware-lifecycle-manager.v4.16.4
        version: 4.16.4
      - name: topology-aware-lifecycle-manager.v4.16.3
        version: 4.16.3
      - name: topology-aware-lifecycle-manager.v4.16.2
        version: 4.16.2
      - name: topology-aware-lifecycle-manager.v4.16.1
        version: 4.16.1
      - name: topology-aware-lifecycle-manager.v4.16.0
        version: 4.16.0
      name: stable
    defaultChannel: stable
    packageName: topology-aware-lifecycle-manager
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Korrel8r
      provider-url: https://github.com/korrel8r/korrel8r
    name: korrel8r
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: korrel8r.v0.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "korrel8r.openshift.io/v1alpha1",
                "kind": "Korrel8r",
                "metadata": {
                  "name": "korrel8r",
                  "namespace": "korrel8r"
                },
                "spec": {
                  "debug": {
                    "verbose": 1
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional,Monitoring,Logging & Tracing
          containerImage: quay.io/korrel8r/operator:0.1.8
          createdAt: "2024-09-27T17:45:00Z"
          description: Correlate cluster resources and observability data.
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "korrel8r.openshift.io/v1alpha1",
              "kind": "Korrel8r",
              "metadata": {
                "name": "korrel8r",
                "namespace": "korrel8r"
              }
            }
          operatorframework.io/suggested-namespace: korrel8r
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/korrel8r/korrel8r
          support: Korrel8r Project
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Korrel8r is a service that correlates observabililty signals
              in the cluster.
            displayName: Korrel8r
            kind: Korrel8r
            name: korrel8rs.korrel8r.openshift.io
            version: v1alpha1
        description: |
          Korrel8r helps navigate from problem symptoms to related resources and signal data that can reveal the cause.
          It can follow relationships between disjointed observability "silos" (logs, metrics, alerts and more) to bring together all the data available to solve a problem.
          This operator allows Korrel8r to be deployed as a cluster service with a REST API.
          It can be used as a back-end for data correlation by the OpenShift console and other tools.
        displayName: Korrel8r
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - signal
        - correlation
        links:
        - name: Documentation
          url: https://korrel8r.github.io/korrel8r
        - name: Source Code and Issues
          url: https://github.com/korrel8r/korrel8r
        maintainers:
        - email: korrel8r@gmail.com
          name: Korrel8r Project
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Korrel8r
          url: https://github.com/korrel8r/korrel8r
        relatedImages:
        - quay.io/korrel8r/operator:0.1.8
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 0.1.8
      entries:
      - name: korrel8r.v0.1.8
        version: 0.1.8
      - name: korrel8r.v0.1.7
        version: 0.1.7
      - name: korrel8r.v0.1.5
        version: 0.1.5
      - name: korrel8r.v0.1.3
        version: 0.1.3
      - name: korrel8r.v0.1.2
        version: 0.1.2
      - name: korrel8r.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: korrel8r
    provider:
      name: Korrel8r
      url: https://github.com/korrel8r/korrel8r
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Upbound Inc.
      provider-url: https://upbound.io
    name: universal-crossplane
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: universal-crossplane.1.5.1-up.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"pkg.crossplane.io/v1",
                "kind":"Configuration",
                "metadata":{
                  "name":"xp-getting-started-with-aws"
                },
                "spec":{
                  "package":"registry.upbound.io/xp/getting-started-with-aws:v1.2.2"
                }
              },
              {
                "apiVersion":"apiextensions.crossplane.io/v1",
                "kind":"CompositeResourceDefinition",
                "metadata":{
                  "name":"compositepostgresqlinstances.database.example.org"
                },
                "spec":{
                  "group":"database.example.org",
                  "names":{
                    "kind":"CompositePostgreSQLInstance",
                    "plural":"compositepostgresqlinstances"
                  },
                  "claimNames":{
                    "kind":"PostgreSQLInstance",
                    "plural":"postgresqlinstances"
                  },
                  "connectionSecretKeys":[
                      "username",
                      "password",
                      "endpoint",
                      "port"
                  ],
                  "versions":[
                    {
                      "name":"v1alpha1",
                      "served":true,
                      "referenceable":true,
                      "schema":{
                        "openAPIV3Schema":{
                          "type":"object",
                          "properties":{
                            "spec":{
                              "type":"object",
                              "properties":{
                                "parameters":{
                                  "type":"object",
                                  "properties":{
                                    "storageGB":{
                                      "type":"integer"
                                    }
                                  },
                                  "required":[
                                      "storageGB"
                                  ]
                                }
                              },
                              "required":[
                                  "parameters"
                              ]
                            }
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion":"apiextensions.crossplane.io/v1",
                "kind":"Composition",
                "metadata":{
                  "name":"compositepostgresqlinstances.aws.database.example.org",
                  "labels":{
                    "provider":"aws",
                    "guide":"quickstart",
                    "vpc":"default"
                  }
                },
                "spec":{
                  "writeConnectionSecretsToNamespace":"crossplane-system",
                  "compositeTypeRef":{
                    "apiVersion":"database.example.org/v1alpha1",
                    "kind":"CompositePostgreSQLInstance"
                  },
                  "resources":[
                    {
                      "name":"rdsinstance",
                      "base":{
                        "apiVersion":"database.aws.crossplane.io/v1beta1",
                        "kind":"RDSInstance",
                        "spec":{
                          "forProvider":{
                            "region":"us-east-1",
                            "dbInstanceClass":"db.t2.small",
                            "masterUsername":"masteruser",
                            "engine":"postgres",
                            "engineVersion":"9.6",
                            "skipFinalSnapshotBeforeDeletion":true,
                            "publiclyAccessible":true
                          },
                          "writeConnectionSecretToRef":{
                            "namespace":"crossplane-system"
                          }
                        }
                      },
                      "patches":[
                        {
                          "fromFieldPath":"metadata.uid",
                          "toFieldPath":"spec.writeConnectionSecretToRef.name",
                          "transforms":[
                            {
                              "type":"string",
                              "string":{
                                "fmt":"%s-postgresql"
                              }
                            }
                          ]
                        },
                        {
                          "fromFieldPath":"spec.parameters.storageGB",
                          "toFieldPath":"spec.forProvider.allocatedStorage"
                        }
                      ],
                      "connectionDetails":[
                        {
                          "fromConnectionSecretKey":"username"
                        },
                        {
                          "fromConnectionSecretKey":"password"
                        },
                        {
                          "fromConnectionSecretKey":"endpoint"
                        },
                        {
                          "fromConnectionSecretKey":"port"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion":"pkg.crossplane.io/v1",
                "kind":"Provider",
                "metadata":{
                  "name":"provider-aws"
                },
                "spec":{
                  "package":"crossplane/provider-aws:master"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime, Cloud Provider, Database, Integration &
            Delivery, Networking, Storage
          com.redhat.component: universal-crossplane
          com.redhat.delivery.backport: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.6
          containerImage: upbound/uxp-bootstrapper
          createdAt: 2021-05-18T00:00:00+0000
          description: Upbound Universal Crossplane (UXP) is Upbound's official enterprise-grade
            distribution of Crossplane.
          io.k8s.display-name: universal-crossplane
          io.openshift.tags: uxp,crossplane,upbound
          maintainer: Upbound Inc. <info@upbound.io>
          name: universal-crossplane
          ocs.tags: v4.6
          operatorframework.io/initialization-resource: |
            apiVersion: v1
            kind: Secret
            metadata:
              name: upbound-control-plane-token
              namespace: upbound-system
            type: Opaque
            data:
              token: "Output of `up cloud controlplane attach | base64` command"
          operatorframework.io/suggested-namespace: upbound-system
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io.bundle.channels.v1: stable
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: universal-crossplane
          operators.operatorframework.io/internal-objects: '["locks.pkg.crossplane.io"]'
          repository: https://github.com/upbound/universal-crossplane
          summary: Upbound Universal Crossplane (UXP) is Upbound's official enterprise-grade
            distribution of Crossplane.
          support: Upbound Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: CompositeResourceDefinition
            kind: CompositeResourceDefinition
            name: compositeresourcedefinitions.apiextensions.crossplane.io
            version: v1
          - displayName: CompositionRevision
            kind: CompositionRevision
            name: compositionrevisions.apiextensions.crossplane.io
            version: v1alpha1
          - displayName: Composition
            kind: Composition
            name: compositions.apiextensions.crossplane.io
            version: v1
          - displayName: ConfigurationRevision
            kind: ConfigurationRevision
            name: configurationrevisions.pkg.crossplane.io
            version: v1
          - displayName: Configuration
            kind: Configuration
            name: configurations.pkg.crossplane.io
            version: v1
          - displayName: ControllerConfig
            kind: ControllerConfig
            name: controllerconfigs.pkg.crossplane.io
            version: v1alpha1
          - displayName: Lock
            kind: Lock
            name: locks.pkg.crossplane.io
            version: v1alpha1
          - displayName: ProviderRevision
            kind: ProviderRevision
            name: providerrevisions.pkg.crossplane.io
            version: v1
          - displayName: Provider
            kind: Provider
            name: providers.pkg.crossplane.io
            version: v1
        description: |
          ## Overview

          Upbound Universal Crossplane (UXP) is [Upbound's][upbound] official
          enterprise-grade distribution of [Crossplane][crossplane]. It's fully compatible
          with upstream Crossplane, open source, capable of connecting to Upbound Cloud
          for real-time dashboard visibility, and maintained by Upbound. It's the easiest
          way for both individual community members and enterprises to build their
          production control planes.

          ## Quick Start

          1. Install the [Upbound CLI][upbound-cli].
              ```console
              curl -sL https://cli.upbound.io | sh
              # Install with Homebrew
              brew install upbound/tap/up
              ```

          2. [Create an Upbound account][create-account] for a free dashboard for UXP.

          3. Connect UXP to Upbound Cloud.
              ```console
              # The name of your new UXP control plane.
              UXP_NAME=mycrossplane

              up cloud login
              up cloud controlplane attach ${UXP_NAME} | up uxp connect -
              ```

          4. Manage your UXP control plane by [signing in][login] to your Upbound account.

          ## Additional Resources

          - The [UXP Documentation][uxp-documentation] provides additional information
            about UXP and resources for developers, like examples.
          - The [developer guide][developer-guide] describes how to build and run UXP
            locally from source.
          - [UXP Slack][uxp-slack] is where you can go to get all of your UXP questions
            answered.

          [upbound]: https://upbound.io
          [crossplane]: https://crossplane.io/
          [upbound-cli]: https://github.com/upbound/up
          [create-account]: https://cloud.upbound.io/register
          [login]: https://cloud.upbound.io/login
          [uxp-documentation]: https://cloud.upbound.io/docs/uxp
          [developer-guide]: https://github.com/upbound/universal-crossplane/blob/main/docs/developer-guide.md
          [uxp-slack]: https://crossplane.slack.com/archives/upbound
        displayName: Upbound Universal Crossplane (UXP)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - infrastructure
        - services
        - application
        - database
        - cache
        - bucket
        - infra
        - app
        - ops
        - oam
        - gcp
        - azure
        - aws
        - alibaba
        - cloudsql
        - rds
        - s3
        - azuredatabase
        - asparadb
        - gke
        - aks
        - eks
        links:
        - name: GitHub
          url: https://github.com/upbound/universal-crossplane
        - name: Website
          url: https://upbound.io
        - name: Twitter
          url: https://twitter.com/upbound_io
        - name: Slack
          url: https://slack.crossplane.io/
        maintainers:
        - email: info@upbound.io
          name: Upbound Inc.
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Upbound Inc.
          url: https://upbound.io
        relatedImages:
        - upbound/uxp-bootstrapper:v1.5.1-up.1
        - upbound/crossplane:v1.5.1-up.1
        - upbound/xgql:v0.1.5
        version: 1.5.1-up.1
      entries:
      - name: universal-crossplane.1.5.1-up.1
        version: 1.5.1-up.1
      - name: universal-crossplane.1.4.3-up.1
        version: 1.4.3-up.1
      - name: universal-crossplane.1.3.3-up.1
        version: 1.3.3-up.1
      - name: universal-crossplane.1.3.1-up.1
        version: 1.3.1-up.1
      - name: universal-crossplane.1.2.3-up.1
        version: 1.2.3-up.1
      - name: universal-crossplane.1.2.2-up.1
        version: 1.2.2-up.1
      - name: universal-crossplane.1.2.1-up.4
        version: 1.2.1-up.4
      - name: universal-crossplane.1.2.1-up.3
        version: 1.2.1-up.3
      name: stable
    defaultChannel: stable
    packageName: universal-crossplane
    provider:
      name: Upbound Inc.
      url: https://upbound.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Palo Alto Networks
      provider-url: ""
    name: pcc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pcc-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "name": "pcc-console",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "ConsoleDefender",
                "metadata": {
                  "name": "pcc-consoledefender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "defenderConfig": {
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Defender",
                "metadata": {
                  "name": "pcc-defender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "defenderConfig": {
                    "clusterAddress": "twistlock-console.example.com",
                    "consoleAddress": "https://twistlock-console.example.com:8083",
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: quay.io/prismacloud/pcc-operator@sha256:51a343d5ea0115e3c43ccaa546492e85ab74cf2cf2658d725c1ea7099fc89191
          description: Deploy Prisma Cloud Compute for cloud-native security in your
            clusters
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator
          support: pcc-integrations@paloaltonetworks.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Install the Prisma Cloud Compute Console and Defenders
            displayName: Console and Defenders
            kind: ConsoleDefender
            name: consoledefenders.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install the Prisma Cloud Compute Console
            displayName: Console
            kind: Console
            name: consoles.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install Prisma Cloud Compute Defenders
            displayName: Defenders
            kind: Defender
            name: defenders.pcc.paloaltonetworks.com
            version: v1alpha1
        description: |
          ## Features
          - Deploy Console
              - Create initial local administrator user
              - Add license
          - Deploy Defenders
          - Support for offline deployments
          - Upgrade Console
          - Upgrade Defenders

          ## Installation
          See the [Prisma Cloud Compute Operator documentation](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator) for installation guides.

          ## Support
          This operator is available "as is," and should be seen as community-supported; however, Palo Alto Networks will continue contributing whenever possible.
          Palo Alto Networks does not provide technical support or help with using or troubleshooting the operator through normal support channels.
          We encourage you to open GitHub [issues](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/issues) and [pull requests](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/pulls) to track bugs and feature requests.

          ## Other links
          [Prisma Cloud Compute product documentation](https://docs.paloaltonetworks.com/prisma/prisma-cloud/prisma-cloud-admin-compute.html)

          [Prisma Cloud Compute API documentation](https://prisma.pan.dev/api/cloud/cwpp)
        displayName: Prisma Cloud Compute Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - palo
        - palo alto
        - palo alto networks
        - paloalto
        - twistlock
        - compute
        - prisma
        - prisma cloud
        - prisma cloud compute
        - security
        - image
        - images
        - container
        - containers
        - docker
        - podman
        - vulnerability
        - vulnerabilities
        - compliance
        - runtime
        - scan
        - scanning
        - monitor
        - monitoring
        - alert
        - alerting
        links:
        - name: Prisma Cloud
          url: https://www.paloaltonetworks.com/prisma/cloud
        maintainers:
        - email: pcc-integrations@paloaltonetworks.com
          name: Wyatt Gill
        maturity: alpha
        provider:
          name: Palo Alto Networks
        relatedImages:
        - quay.io/prismacloud/pcc-operator@sha256:51a343d5ea0115e3c43ccaa546492e85ab74cf2cf2658d725c1ea7099fc89191
        version: 0.2.3
      entries:
      - name: pcc-operator.v0.2.3
        version: 0.2.3
      name: stable
    defaultChannel: stable
    packageName: pcc-operator
    provider:
      name: Palo Alto Networks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-iam-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-iam-operator.v4.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "Authentication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-iam-operator",
                    "app.kubernetes.io/managed-by": "ibm-iam-operator",
                    "app.kubernetes.io/name": "ibm-iam-operator"
                  },
                  "name": "example-authentication"
                },
                "spec": {
                  "auditService": {
                    "imageName": "auditService no longer used - ignore",
                    "imageRegistry": "auditService no longer used - ignore",
                    "imageTag": "auditService no longer used - ignore"
                  },
                  "authService": {
                    "imageName": "icp-platform-auth",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "ldapsCACert": "platform-auth-ldaps-ca-cert",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "650Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "400Mi",
                        "memory": "350Mi"
                      }
                    },
                    "routerCertSecret": "saml-auth-secret"
                  },
                  "clientRegistration": {
                    "imageName": "im-initcontainer",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "158Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "config": {
                    "authUniqueHosts": "internal-ip1 internal-ip2 mycluster.icp",
                    "bootstrapUserId": "kubeadmin",
                    "claimsMap": "name=\"givenName\" family_name=\"givenName\" given_name=\"givenName\" preferred_username=\"displayName\" display_name=\"displayName\"",
                    "claimsSupported": "name,family_name,display_name,given_name,preferred_username",
                    "clusterCADomain": "mycluster.icp",
                    "clusterExternalAddress": "10.0.0.1",
                    "clusterInternalAddress": "10.0.0.1",
                    "clusterName": "mycluster",
                    "defaultAdminPassword": "password",
                    "defaultAdminUser": "cpadmin",
                    "defaultLogin": "",
                    "enableImpersonation": false,
                    "fipsEnabled": true,
                    "ibmCloudSaas": false,
                    "icpPort": 8443,
                    "installType": "fresh",
                    "isOpenshiftEnv": true,
                    "nonceEnabled": true,
                    "oidcIssuerURL": "https://127.0.0.1:443/idauth/oidc/endpoint/OP",
                    "openshiftPort": 443,
                    "preferredLogin": "",
                    "providerIssuerURL": "",
                    "roksEnabled": true,
                    "roksURL": "https://roks.domain.name:443",
                    "roksUserPrefix": "changeme",
                    "saasClientRedirectUrl": "",
                    "scimAdminPassword": "scimpassword",
                    "scimAdminUser": "scimadmin",
                    "scopeClaim": "profile=\"name,family_name,display_name,given_name,preferred_username\"",
                    "wlpClientID": "4444be3a738841016ab76d71b650e836",
                    "wlpClientRegistrationSecret": "f1362ca4d20b8389af2d1ea68042c9af",
                    "wlpClientSecret": "aa73bf39752053bf723d1143fb4cf8a2",
                    "xframeDomain": "",
                    "zenFrontDoor": false
                  },
                  "identityManager": {
                    "imageName": "icp-identity-manager",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "masterNodesList": "10.0.0.1",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "550Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "ephemeral-storage": "300Mi",
                        "memory": "150Mi"
                      }
                    }
                  },
                  "identityProvider": {
                    "imageName": "icp-identity-provider",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "550Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "ephemeral-storage": "300Mi",
                        "memory": "150Mi"
                      }
                    }
                  },
                  "initMongodb": {
                    "imageName": "im-initcontainer",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "178Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "operatorVersion": "0.14.1",
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
          createdAt: "2025-02-20T00:03:25Z"
          description: The IAM operator provides a simple Kubernetes CRD-Based API
            to manage the lifecycle of IAM services. With this operator, you can simply
            deploy and upgrade the IAM services
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.11.0
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-iam-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Authentication is the Schema for the authentications API
            displayName: Authentication
            kind: Authentication
            name: authentications.operator.ibm.com
            version: v1alpha1
          - description: Client is the Schema for the clients API
            displayName: Client
            kind: Client
            name: clients.oidc.security.ibm.com
            version: v1
        description: |-
          The `ibm-iam-operator` installs the IBM Cloud Platform Common Services Identity and access management (IAM) service.

          **Important:** Do not install this operator directly. Install this operator only by using the IBM Common Service Operator. For more information about installing this operator and other Common Services operators, see [Installer documentation](http://ibm.biz/cpcs_opinstall). Additionally, you can exit this panel and navigate to the IBM Common Services tile in OperatorHub to learn more about the operator. If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For more information about IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).

          You can use the `ibm-iam-operator` to install the authentication and authorization services for the IBM Cloud Platform Common Services.

          With these services, you can configure security for IBM Cloud Platform Common Services, IBM Certified Containers (IBM products), or IBM Cloud Paks that are installed.

          For more information about the available IBM Cloud Platform Common Services, see the [IBM Knowledge Center](http://ibm.biz/cpcsdocs).

          ## Supported platforms

           - Red Hat OpenShift Container Platform 4.3 or newer installed on one of the following platforms:

              - Linux x86_64
              - Linux on Power (ppc64le)
              - Linux on IBM Z and LinuxONE

          ## Prerequisites

          Before you install this operator, you need to first install the operator dependencies and prerequisites:

          - For the list of operator dependencies, see the IBM Knowledge Center [Common Services dependencies documentation](http://ibm.biz/cpcs_opdependencies).

          - For the list of prerequisites for installing the operator, see the IBM Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).

          ## Documentation

          To install the operator by using the IBM Common Services Operator, follow the installation and configuration instructions that are in the IBM Knowledge Center.

          - If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).
          - If you are using the operator with an IBM Containerized Software, see the IBM Cloud Platform Common Services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall).
        displayName: IBM IM Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:b8886b9b54e6ee4c6b2981ef092e54348a0286577d9924cf136cee881f5bf1f4
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-bundle@sha256:98cf2a6c7b7af5d7cbfbc61ce140201272b30b12edabaec2b2d4287b0690d89e
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:f289ac0d44803ff12d41c3659e6fdf04c404e0b3b233b6bc44eec671abb1b47e
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:52ce08ac2bce7c02dc30e2c0aa74df4bde8f156debccc59960f02b4b0040ffea
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:97b9ecef8caef156a0eb073b6eb2a34b63209b44511f5694a243f11285d88662
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:0f46b288d060a15a48ecad8f328a8046bf232b8629c3d79da6a82ea712b363f4
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:7cec55c00827b747f23c2f1703d99e1ca5b30a0c445f8ad0471b25b0c9cca57c
        version: 4.11.0
      entries:
      - name: ibm-iam-operator.v4.11.0
        version: 4.11.0
      name: v4.10
    defaultChannel: v4.10
    packageName: ibm-iam-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: vfunction-operator
      app.kubernetes.io/managed-by: operator
      app.kubernetes.io/name: vfunction
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      name: vfunction-server-operator
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: vFunction
      provider-url: ""
    name: vfunction-server-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vfunction-server-operator.v4.3.1762
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vfunction.com/v1",
                "kind": "VfunctionServer",
                "metadata": {
                  "name": "vfunction-server",
                  "namespace": "vfunction",
                  "labels": {
                      "name": "vfunction",
                      "app.kubernetes.io/name": "vfunction",
                      "app.kubernetes.io/instance": "vfunction-server",
                      "app.kubernetes.io/managed-by": "operator"
                  }
                },
                "spec": {
                  "admin": {
                    "email": "admin@mycompany.com",
                    "name": "Admin",
                    "password": "Password1!"
                  },
                  "host": "http://my.domain.com",
                  "org_name": "MyCompany",
                  "upgrade": "Never",
                  "measurements": {
                    "min_number_of_services": "2",
                    "max_number_of_services": "10",
                    "max_pod_memory_capacity": "16G",
                    "max_pod_cpu_capacity": "1",
                    "pod_ephemeral_storage": "1G",
                    "S3": {
                      "bucket": "",
                      "key": "",
                      "secret": "",
                      "region": ""
                    }
                  },
                  "smtp": {
                    "identity": "",
                    "password": "",
                    "url": "",
                    "user": ""
                  },
                  "authentication": {
                    "authority": "",
                    "client_id": "",
                    "client_secret": "",
                    "jwks_url": "",
                    "issuer": "",
                    "ca_root_crt": ""
                  },
                  "tls": {
                    "use_letsencrypt": "No",
                    "crt": "",
                    "key": ""
                  },
                  "nginx": {
                    "force_http": "No",
                    "ipv6_disabled": "No",
                    "service_type": ""
                  },
                  "route": {
                    "use_route": "Yes",
                    "tls_termination": "edge",
                    "use_tls_certificate": "No"
                  },
                  "mysql": {
                    "external_mysql_ip": "",
                    "external_mysql_user": "",
                    "external_mysql_password": "",
                    "max_pod_memory_capacity": "4G",
                    "max_pod_cpu_capacity": "1",
                    "pod_ephemeral_storage": "1G"
                  },
                  "storage": {
                    "storage_class": "",
                    "mysql_persistent_volume": "",
                    "mysql_persistent_volume_claim": "",
                    "mysql_persistent_volume_claim_access_mode": "ReadWriteOnce",
                    "storage_persistent_volume": "",
                    "storage_persistent_volume_claim": "",
                    "storage_persistent_volume_claim_access_mode": "ReadWriteOnce",
                    "lets_encrypt_persistent_volume": "",
                    "lets_encrypt_persistent_volume_claim": "",
                    "lets_encrypt_persistent_volume_claim_access_mode": "ReadWriteOnce"
                  },
                  "distributed": {
                    "otlp": {
                      "min_number_of_services": "1",
                      "max_number_of_services": "10",
                      "max_pod_memory_capacity": "8G",
                      "max_pod_cpu_capacity": "1",
                      "pod_ephemeral_storage": "250M",
                      "max_folder_size_in_mb": "200"
                    },
                    "dm": {
                      "min_number_of_services": "1",
                      "max_number_of_services": "10",
                      "max_pod_memory_capacity": "8G",
                      "max_pod_cpu_capacity": "1",
                      "pod_ephemeral_storage": "1G"
                    }
                  },
                  "backup": {
                    "mysql": {
                      "restore_during_installation": "No",
                      "S3": {
                        "bucket": "",
                        "key": "",
                        "secret": "",
                        "region": ""
                      }
                    }
                  },
                  "security": {
                    "image_pull_policy": "",
                    "force_hashed_images": "No",
                    "set_pod_resources": "Yes",
                    "proxy": {
                      "http_proxy": "",
                      "https_proxy": "",
                      "additional_no_proxy": ""
                    },
                    "use_rate_limiting": "No",
                    "requests_per_second": "10"
                  },
                  "affinity": {
                    "node_affinity": [
                      {
                        "key": "",
                        "operator": "",
                        "values": "",
                        "services": ""
                      }
                    ],
                    "node_selectors": [
                      {
                        "key": "",
                        "value": "",
                        "services": ""
                      }
                    ],
                    "additional_labels": [
                      {
                        "key": "",
                        "value": "",
                        "services": ""
                      }
                    ],
                    "additional_annotations": [
                      {
                        "key": "",
                        "value": "",
                        "services": ""
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, AI/Machine Learning
          certified: Red Hat certified
          containerImage: registry.connect.redhat.com/vfunction/vfunction-server-operator@sha256:a5a0f47f7fb9219139f8d749da14950a8c39698b340644011485dfbec54ebe8b
          createdAt: "2025-08-04T16:10:32Z"
          description: vFunction is a cutting-edge code analysis, machine learning,
            and automation to boost your Java modernization projects.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.3.1762
          support: vFunction
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of vFunction running on your cluster.
            displayName: vFunction
            kind: VfunctionServer
            name: vfunctionservers.vfunction.com
            version: v1
        description: |2-

          ## Introduction

          [vFunction](https://www.vfunction.com/) is a cloud-native modernization platform that combines dynamic and static code analysis, machine learning, and automation to automatically identify and extract services from existing applications. vFunction is the only platform purpose-built for modernization of Java applications.

          The vFunction modernization process starts by learning the running monolithic application, and surfacing the interdependencies within it. Using AI, the platform analyzes and identifies services that can be separated from the application. This decomposition can present a range of micro, mini, or even macro services, depending on your application environment, each being an independently deployable and scalable application component.

          vFunction automates the extraction of these services, enabling you to modernize your monolith, quickly and easily.

          ## The vFunction platform

          The platform consists of 3 basic components; the server, the controller package, and a tools package. The server runs as an operand on an OpenShift environment. The controller package is installed on the machine that runs the monolithic application which can be either a Linux or a Windows machine, and the tools are run on a development machine, with access to the code of the monolithic application.

          The controller package consists of three elements: the vFunction agent, that collects data during the dynamic analysis phase; the vFunction Viper application, that performs static analysis on the binaries of the application; and the vFunction controller that handles all the communication between the agent, Viper, and the vFunction server.

          The vFunction agent is a mix of a Java and native agent, and needs to run on the JVM that is currently running your application. [Refer to the vFunction OOB Support Matrix document](https://drive.google.com/file/d/1ccq8LFab1FrYAimDUxwgjiCOdk4QuVzs/view) for a list of supported application servers and JVMs.

          This operator installs a vFunction server instance to be connected later with one or more vFunction controllers installed on your application machine(s).

          ## Before You Start

          The vFunction operator requires a Red Hat OpenShift Kubernetes Application Platform 4.x.

          The cluster recommended configuration (which is the same as the minimal one) should consist of:

          - *1 master node*

          - *1 worker node*

          - *Storage capacity enough for dynamically provision 2 PVs of 50G each*

          - *A default StorageClass configured*

          vFunction doesn’t provide any encryption mechanism for data stored on any attached storage. If required, encryption can be achieved by the client by using its own external means on the storage itself.

          vFunction supports backup of all critical data as part of the application. Nevertheless, for complete backup of entire data it is recommended that the client apply its own storage based backup mechanism.

          ## Install Prerequisites

          ###Cluster:
          The operator was certified and tested on OCP 4.6 - 4.17.
          If you encounter any issue with other OCP versions or cloud providers please contact info@vfunction.com.

          ###Storage:
          The operator creates two new PersistentVolumeClaims (PVCs) during installation time. For their creation, it relies on the default StorageClass to dynamically provision two PersistentVolumes (PVs). The system must have a default StorageClass in place.
          Both PersistentVolumes are defaulted to access with ReadWriteOnce mode, but you can change this by specifying a different access mode in the YAML file.
          Notes:
          - There is an option to specify a specific StorageClass for the PersistentVolumeClaims.
          - There is an option to specify the use of a specific pre-created PersistentVolume for each PersistentVolumeClaim
          - There is an option to specify a pre-created PersistentVolumeClaim for each Deployment.
          - There is an option to change the access mode for each PersistentVolumeClaim.

          ###Project:
          The operator should be installed in a new and dedicated project (namespace). If you intend to install more than one vFunction operands in the same cluster, please use different projects for each one. Installation of more than one operand on the same cluster will allow scaling by load balancing multiple applications to different vFunction servers.

          ## Installation instructions

          1. Prepare the YAML file to use in the installation:

              a. Copy the following YAML template into a text editor:

              ```yaml
                  apiVersion: vfunction.com/v1
                  kind: VfunctionServer
                  metadata:
                    name: vfunction
                    namespace: vfunction
                    labels:
                      name: vfunction
                      app.kubernetes.io/name: vfunction
                      app.kubernetes.io/instance: vfunction-server
                      app.kubernetes.io/managed-by: operator
                  spec:
                    host: "http://my.domain.com"
                    org_name: "MyCompany"
                    upgrade: "Never"
                    admin:
                      email: "admin@mycompany.com"
                      name: "Admin"
                      password: "Password1!"
                    measurements:
                      min_number_of_services: "2"
                      max_number_of_services: "10"
                      max_pod_memory_capacity: "16G"
                      max_pod_cpu_capacity: "1"
                      S3:
                        bucket: ""
                        key: ""
                        secret: ""
                    smtp:
                      password: ""
                      url: ""
                      identity: ""
                      user: ""
                    authentication:
                      authority: ""
                      client_id: ""
                      client_secret: ""
                      jwks_url: ""
                      issuer: ""
                      ca_root_crt: |
                        -----BEGIN CERTIFICATE-----
                        ...
                        -----END CERTIFICATE-----
                    tls:
                      use_letsencrypt: "No"
                      crt: |
                        -----BEGIN CERTIFICATE-----
                        ...
                        -----END CERTIFICATE-----
                      key: |
                        -----BEGIN PRIVATE KEY-----
                        ...
                        -----END PRIVATE KEY-----
                    nginx:
                      force_http: "No"
                      ipv6_disabled: "No"
                      service_type: ""
                    route:
                      use_route: "Yes"
                      tls_termination: "edge"
                      use_tls_certificate: "No"
                    mysql:
                      external_mysql_ip: ""
                      external_mysql_user: ""
                      external_mysql_password: ""
                      max_pod_memory_capacity: "4G"
                      max_pod_cpu_capacity: "1"
                    storage:
                      storage_class: ""
                      mysql_persistent_volume: ""
                      mysql_persistent_volume_claim: ""
                      mysql_persistent_volume_claim_access_mode: "ReadWriteOnce"
                      storage_persistent_volume: ""
                      storage_persistent_volume_claim: ""
                      storage_persistent_volume_claim_access_mode: "ReadWriteOnce"
                      lets_encrypt_persistent_volume: ""
                      lets_encrypt_persistent_volume_claim: ""
                      lets_encrypt_persistent_volume_claim_access_mode: "ReadWriteOnce"
                    distributed:
                        otlp:
                          min_number_of_services: "1"
                          max_number_of_services: "10"
                          max_pod_memory_capacity: "8G"
                          max_pod_cpu_capacity: "1"
                          max_folder_size_in_mb: "200"
                        dm:
                          min_number_of_services: "1"
                          max_number_of_services: "10"
                          max_pod_memory_capacity: "8G"
                          max_pod_cpu_capacity: "1"
                    backup:
                      mysql:
                        restore_during_installation: "No"
                        S3:
                          bucket: ""
                          key: ""
                          secret: ""
                    security:
                      image_pull_policy: ""
                      force_hashed_images: "No"
                      set_pod_resources: "Yes"
                      proxy:
                        http_proxy: ""
                        https_proxy: ""
                        additional_no_proxy: ""
                      use_rate_limiting: "No"
                      requests_per_second: "10"
                    affinity:
                      node_affinity:
                        - key: ""
                          operator: ""
                          values: ""
                          services: ""
                      node_selectors:
                        - key: ""
                          value: ""
                          services: ""
                      additional_labels:
                        - key: ""
                          value: ""
                          services: ""
                      additional_annotations:
                        - key: ""
                          value: ""
                          services: ""
              ```
              b. Customize the template fields by replacing with your information:

              - **host**: Enter the FQDN for accessing the vFunction dashboard.
                          Make sure that you use a domain name and not an IP address. Writing "https://" will mandate TLS while writing "http://" will expose the server through HTTP.
              - **org_name**: Enter your organization name.
              - **upgrade**: Choose auto-upgrade mode.
                  - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically.
                  - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
                  - **Never**: The automatic upgrading mechanism is off (default).
                  - **Maintenance**: Put the operator in a maintenance mode.
              - **admin.email**: Enter the email address of your administrator.
              - **admin.name**: Enter the name of your administrator.
              - **admin.password**: Enter the password you want to use for the vFunction administrator.
                                    The password should be at least 8 characters long, and consist of at least one lowercase letter, at least one uppercase letter, at least one number, and at least one special character.
              - **measurements.min\_num\_of\_services**: Minimum Number of measurement service pods.
              - **measurements.max\_num\_of\_services**: Maximum Number of measurement service pods.
              - **measurements.max\_pod\_memory\_capacity**: Maximum memory capacity limit for measurement service pods. (Default is "8G").
              - **measurements.max\_pod\_cpu\_capacity**: Maximum CPU capacity limit for measurement service pods. (Default is "1").
              - **measurements.pod\_ephemeral\_storage**: Ephemeral Storage Capacity for measurement service pods. (Default is "1G").
              - **measurements.S3**: Save all measurements data to S3, instead of local PV (optional).
                  - **measurements.S3.bucket**: Measurements S3 bucket name.
                  - **measurements.S3.key**: Measurements S3 key.
                  - **measurements.S3.secret**: Measurements S3 secret.
                  - **measurements.S3.region**: Measurements S3 region.
              - **smtp.user**: Enter the email address for a designated user for the SMTP server (optional).
              - **smtp.password**: Enter this user’s password (optional).
              - **smtp.identity**: Enter this user’s password (optional).
              - **smtp.url**: Enter the SMTP server URL (optional).
              - **authentication.authority**: Authority for authentication (Optional).
              - **authentication.client_id**: Client ID for authentication (Optional).
              - **authentication.client_secret**: Client Secret for authentication (Optional).
              - **authentication.jwks_url**: JWKS URL for authentication (Optional).
              - **authentication.issuer**: Issuer for authentication (Optional).
              - **authentication.ca\_root\_crt**: CA root certificate for authentication (Optional).
              - **tls.use_letsencrypt**: If your host FQDN starts with “https”, choose whether to use Let’s Encrypt to obtain certificates.
              - **tls.crt**: If your host FQDN starts with “https”, paste in the certifications you have for using the TLS connection (Applicable only if you don’t use Let’s Encrypt).
              - **tls.key**: If your host FQDN starts with “https”, enter the key you have for using the TLS connection (Applicable only if you don’t use Let’s Encrypt).
              - **nginx.force_http**: Force internal NGINX to listen on http even if the host FQDN starts with “https”.
                                      This is useful if you want to use a TLS termination proxy in front of vFunction, other then the route provided by the operator.
              - **nginx.ipv6_disabled**: Disable IPv6 on internal NGINX.
              - **nginx.service_type**: Set the service type for the internal NGINX service - ClusterIP, NodePort or LoadBalancer.
              - **route.use_route**: Use OpenShift route to expose vFunction. (Default is "Yes").
              - **route.tls_termination**: Set the TLS termination type for the OpenShift route - edge, passthrough or reencrypt. (Default is "edge").
              - **route.use\_tls\_certificate**: Use TLS certificate for OpenShift route (as provided in tls.crt and tls.key) - instead of the default one defined for the route controller in your cluster. (Default is "No").
              - **mysql.external\_mysql\_ip**: If you want to use an external MySQL database, enter its IP address.
              - **mysql.external\_mysql\_user**: If you want to use an external MySQL database, enter its user name.
              - **mysql.external\_mysql\_password**: If you want to use an external MySQL database, enter its password.
              - **mysql.max\_pod\_memory\_capacity**: Maximum memory capacity limit for MySQL pod. (Default is "2G").
              - **mysql.max\_pod\_cpu\_capacity**: Maximum CPU capacity limit for MySQL pod. (Default is "1").
              - **mysql.pod\_ephemeral\_storage**: Ephemeral Storage Capacity for MySQL pod. (Default is "1G").
              - **storage.storage\_class**: If you want to use a specific StorageClass for all PVCs, other than the default one, enter its name.
              - **storage.mysql\_persistent\_volume**: If you want to use a specific PV for mysql PVC, other than the dynamically provisioned one, enter its name.
              - **storage.mysql\_persistent\_volume\_claim**: If you want to use a specific pre-created PVC for mysql deployment, other than the dynamically provisioned one, enter its name.
              - **storage.mysql\_persistent\_volume\_claim\access\_mode**: If you want to use a specific access mode for mysql PVC, other than the default one, enter its name. (Default is "ReadWriteOnce").
              - **storage.storage\_persistent\_volume**: If you want to use a specific PV for storage PVC, other than the dynamically provisioned one, enter its name.
              - **storage.storage\_persistent\_volume\_claim**: If you want to use a specific pre-created PVC for storage deployment, other than the dynamically provisioned one, enter its name.
              - **storage.storage\_persistent\_volume\_claim\access\_mode**: If you want to use a specific access mode for storage PVC, other than the default one, enter its name. (Default is "ReadWriteOnce").
              - **storage.lets\_encrypt\_persistent\_volume**: If you want to use a specific PV for lets encrypt PVC, other than the dynamically provisioned one, enter its name.
              - **storage.lets\_encrypt\_persistent\_volume\_claim**: If you want to use a specific pre-created PVC for lets encrypt deployment, other than the dynamically provisioned one, enter its name.
              - **storage.lets\_encrypt\_persistent\_volume\_claim\access\_mode**: If you want to use a specific access mode for lets encrypt PVC, other than the default one, enter its name. (Default is "ReadWriteOnce").
              - **distributed.otlp.min\_number\_of\_services**: Minimum Number of OTLP pods.
              - **distributed.otlp.max\_number\_of\_services**: Maximum Number of OTLP pods.
              - **distributed.otlp.max\_pod\_memory\_capacity**: Maximum memory capacity for OTLP pods. (Default is "1G").
              - **distributed.otlp.max\_pod\_cpu\_capacity**: Maximum CPU capacity for OTLP pods. (Default is "1").
              - **distributed.otlp.pod\_ephemeral\_storage**: Ephemeral Storage Capacity for OTLP service pods. (Default is "250M").
              - **distributed.otlp.max\_folder\_size\_in\_mb**: Maximum folder size in MB for OTLP pods. (Default is "200").
              - **distributed.dm.min\_number\_of\_services**: Minimum Number of dm pods.
              - **distributed.dm.max\_number\_of\_services**: Maximum Number of dm pods.
              - **distributed.dm.max\_pod\_memory\_capacity**: Maximum memory capacity for dm pods. (Default is "8G").
              - **distributed.dm.max\_pod\_cpu\_capacity**: Maximum CPU capacity for dm pods. (Default is "1").
              - **distributed.dm.pod\_ephemeral\_storage**: Ephemeral Storage Capacity for dm service pods. (Default is "1G").
              - **backup.mysql.restore\_during\_installation**: If you want, during installation, to restore the backup from the latest kept backup with the same major version (if available on attached DB storage or S3 bucket), choose "Yes".
              - **backup.mysql.S3**: Backup MySQL to S3 (optional).
                  - **backup.mysql.S3.bucket**: MySQL backup S3 bucket name.
                  - **backup.mysql.S3.key**: MySQL backup S3 key.
                  - **backup.mysql.S3.secret**: MySQL backup S3 secret.
                  - **backup.mysql.S3.region**: MySQL backup S3 region.
              - **security.image\_pull\_policy**: If you want to use a custom image pull policy for all deployments, enter it here (optional).
              - **security.force\_hashed\_images**: If you want to use hashed images for all deployments, choose "Yes".
              - **security.set\_pod\_resources**: If you want to not set resource requests and limits for all deployments, choose "No".
              - **security.proxy.http\_proxy**: Enter the address of an HTTP proxy (optional).
              - **security.proxy.https\_proxy**: Enter the address of an HTTPS proxy (optional).
              - **security.proxy.additional\_no\_proxy**: Enter additional addresses to exclude from proxying (optional).
              - **security.use\_rate\_limiting**: If you want to use rate limiting in nginx, choose "Yes".
              - **security.requests\_per\_second**: Enter the number of requests per second for rate limiting.
              - **affinity.node\_affinity**: Enter the node affinity for the deployments (optional). There is only one nodeSelectorTerm as requiredDuringSchedulingIgnoredDuringExecution, but it can contain multiple matchExpressions (which are ANDed together).
                  - **key**: Enter the key for the node affinity.
                  - **operator**: Enter the operator for the node affinity (In, NotIn, Exists, DoesNotExist, Gt, Lt).
                  - **values**: Enter the values for the node affinity. A comma-separated list of values.
                  - **services**: Enter the services for the node affinity. A comma-separated list of services to which the node affinity applies. If not specified, the node affinity will apply to all services. List of possible services: mysql, nginx, storage, vfapi-parser, vfapi-users, vfapi-organizations, vfapi-idp, vfapi-mcp, vfapi-measurements, vfapi-otlp, vfapi-dm. (optional).
              - **affinity.node\_selectors**: Enter the node selectors for the deployments (optional).
                  - **key**: Enter the key for the node selector.
                  - **value**: Enter the value for the node selector.
                  - **services**: Enter the services for the node selector. A comma-separated list of services to which the node selector applies. If not specified, the node selector will apply to all services. List of possible services: mysql, nginx, storage, vfapi-parser, vfapi-users, vfapi-organizations, vfapi-idp, vfapi-mcp, vfapi-measurements, vfapi-otlp, vfapi-dm. (optional).
              - **affinity.additional\_labels**: Enter additional labels for the deployments (optional).
                  - **key**: Enter the key for the additional label.
                  - **value**: Enter the value for the additional label.
                  - **services**: Enter the services for the additional label. See above at the node selectors section for more details. (optional).
              - **affinity.additional\_annotations**: Enter additional annotations for the deployments (optional).
                  - **key**: Enter the key for the additional annotation.
                  - **value**: Enter the value for the additional annotation.
                  - **services**: Enter the services for the additional annotation. See above at the node selectors section for more details. (optional).
          1. In the Installed Operators window choose and click on vFunction Operator and then Create Instance.
          1. In the Create vFunctionServer window, paste the YAML file that you created in step 1.
          1. Click Create.
          1. You can now click on your new created vFunction operand in order to see its details and installation progress.

          ## Verifying the installation

          Check that the operand installed successfully:

          1. Check that the "Operand State" property shows “Working”.
          1. Check that the "Successfully Installed" property shows "Yes".

          If "Operand State" shows "Failed", it indicates the operand wasn’t installed correctly. Check the events and logs for all pods (operator and image containers) for any issues. In the event that you cannot troubleshoot, contact vFunction support.

          ## After installation

          The vFunction site is now accessible via the newly created vFunction application custom address, for example, my.domain.com. There are two ways you can access the vFunction dashboard.

          ### Access using the router canonical hostname

          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address, and to the ”vfunction” subdomain of the OpenShift canonical router hostname as the alias.
          For example, ***my.domain.com.   CNAME   vfunction.apps.ocp4.my-openshift.com.***
          2. Find your cluster Router Canonical Hostname address in the newly created vfunction-route-xxx under your vFunction OpenShift project -> **Networking** -> **Routes** -> **vfunction-route-xxx** route -> **Router Canonical Hostname** field.
          You can now access the dashboard using your defined "host" spec property (as above).

          ### Access using the nginx service location

          You may use this access method if your OpenShift is installed on a provider that supports exposing LoadBalancer-type services.
          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address and to the vfunction-nginx-xxx service location.
          For example, ***my.domain.com.   CNAME   a05951ed7cdf-1394239323.us-east-1.elb.amazonaws.com.***
          1. Point your custom domain to the vfunction-nginx-xxx service's external IP location, which you can find in the OpenShift project > **Networking** -> **Services** -> **vfunction-nginx-xxx** service > **Service Address** > **Location** field.

          ## Upgrade and Rollback

          The vFunction operator includes a built-in auto-upgrade mechanism.

          You can choose one of 4 upgrading modes:
          - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically.
          - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
          - **Never**: The automatic upgrading mechanism is off (default).
          - **Maintenance**: Put the operator in a maintenance mode.

          For an on-demand upgrade,  you can change anytime the upgrading mode from “Never” to “Always”, wait for the operand to be upgraded, and change again to “Never”.

          During the upgrade, the "Operand State" property will change to “Upgrading...” and back again to “Working” after a successful upgrade.

          In case of a failure upgrading the operand, an automatic rollback to the last working version will occur. The failed version will be marked as defective and the operator will not try to upgrade to it again.

          Choosing the "Maintenance" option will tell the operator to take down all pods, so you can safely fix storage issues, return volumes from snapshots, etc.
          After finishing the maintenance, return the upgrade mode to your original desired policy.
        displayName: vFunction Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - vfunction
        - analyzing
        - services
        links:
        - name: vFunction
          url: https://vfunction.com
        - name: Terms
          url: https://www.vfunction.com/eula/
        maintainers:
        - email: raviv@vfunction.com
          name: Raviv Halivni
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: vFunction
        relatedImages:
        - registry.connect.redhat.com/vfunction/vfunction-server-operator@sha256:a5a0f47f7fb9219139f8d749da14950a8c39698b340644011485dfbec54ebe8b
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-users@sha256:c471d450d016aa0801ffb3de4648735245f2cf6901576106aed06cacc948d6f6
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-parser@sha256:b70fe4340d0de33c03a3b3458ea29e4cbff30cdd43f0e8d94265568e9a8d1ac7
        - registry.connect.redhat.com/vfunction/vfunction-storage@sha256:8c36770f1fa5034ba089ae81c36c587cb73a5176f63969023c8ab6f90d8d7abd
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-idp@sha256:17fe2e3ef706a4bbd8fdf0e0e08bd021c49104b331674aae7b8978b31730a03d
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-organizations@sha256:8dd4d043806da1e85a4565b2f14bb9e24671fc158f8c1c2adc3804fda5213571
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-dm@sha256:2f82294b15af873c012094a8984cb0c157bf322c28cbe58b38253adc917b6662
        - registry.connect.redhat.com/vfunction/vfunction-mysql@sha256:b4b6543eb02e396e1339877f61b74cc783977dd607cd3679edce8ac6b627e9e9
        - registry.connect.redhat.com/vfunction/vfunction-nginx@sha256:8370b871d05a24da9841a93fb255bf47945428b0a192b96b4eb95b9c0efb86a0
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-mcp@sha256:d50142c1b7f8a52944f9051a13bb821c0a790f137ebef09b0a84256bf7e5d794
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-measurements@sha256:81edae1c7344c13916530f20c8be09512ac8e20542770949783b0025c37079f8
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-otlp@sha256:80ca6d64b0d74f2d4f1b267002d5666f7c602956c656fd36a5c430f77eef0672
        version: 4.3.1762
      entries:
      - name: vfunction-server-operator.v4.3.1762
        version: 4.3.1762
      - name: vfunction-server-operator.v4.0.1576
        version: 4.0.1576
      name: stable
    defaultChannel: stable
    packageName: vfunction-server-operator
    provider:
      name: vFunction
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent
      provider-url: ""
    name: cilium
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cilium.v1.10.15-x084926b
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/hubble-ui@sha256:dec6e7e541cac91ab859245dd2a7fa02de5363b75caa3509a79cf76d4f622550
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/hubble-ui-backend@sha256:a07e13df3e9d7ec0dcc3b158f243cda5a08f6eca38a25a5aa3171f6b3ec491ed
        - quay.io/cilium/operator-generic@sha256:7f2f7b59b541e4f660d65ed0a21d25ab69e49022172529771dd1134315e95c82
        - quay.io/cilium/clustermesh-apiserver@sha256:e73151c3295589de00dccae75819aabdff8fd907260817020f9e8fe3738e78b2
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - registry.connect.redhat.com/isovalent/cilium-olm:970b1d74dbe2b87c28e8dc1c68936de33123a4df-v1.10.15@sha256:afa3ac140fae24b151ab6ff5a8cdef11566a58f7234d702a6a45a398822e6d34
        - quay.io/cilium/cilium@sha256:adbb60f9d19d641a5496fcfc4769ebb55c1d4cd7699fef752f7c94754194cf83
        - quay.io/cilium/hubble-relay@sha256:67a9183fd1febd9e51f4d02bd35092b81c8612a1716015988f69fffba79d999a
        version: 1.10.15+x084926b
      entries:
      - name: cilium.v1.10.15-x084926b
        version: 1.10.15+x084926b
      - name: cilium.v1.10.13-xa8bb1e5
        version: 1.10.13+xa8bb1e5
      - name: cilium.v1.10.12-x362ea11
        version: 1.10.12+x362ea11
      - name: cilium.v1.10.10-x6adc6ff
        version: 1.10.10+x6adc6ff
      name: "1.10"
    - currentCSV: cilium.v1.11.20-xdb5ed32
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/cilium/cilium@sha256:60df3cb7155886e0b62060c7a4a31e457933c6e35af79febad5fd6e43bab2a99
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:3677840d57e1d74e3ac14db0dfceb83088e6ebd70775d775855cdd3bfed68059
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/operator-generic@sha256:1439954acf620f048ef663524ae70b4a25693c58527a2f2cee51124496e29f90
        - quay.io/cilium/hubble-ui@sha256:bbf8b041152041d3ef000d6074a8dd5d40ed910b5c0067f365b00e52e449d467
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/clustermesh-apiserver@sha256:46760182f8c98227cfac27627275616987b71509227775350573d834133a6d49
        - quay.io/cilium/hubble-relay@sha256:e2f38b901fd8bd5adc9a765a5e68836364ebd1e7dfb85c2bcd8a5488b23c3470
        - quay.io/cilium/hubble-ui-backend@sha256:46070270b8ffe7ac3f1794721be0b5a09f906dff3492f8a199a5f995db362d2a
        version: 1.11.20+xdb5ed32
      entries:
      - name: cilium.v1.11.20-xdb5ed32
        version: 1.11.20+xdb5ed32
      name: "1.11"
    - currentCSV: cilium.v1.12.19-xef55de8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.12.0 <1.12.19+xef55de8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/hubble-ui-backend@sha256:f88a73a120521eeafbcdbd908f517117f6557bf61e115847853fac371f0d774c
        - quay.io/cilium/hubble-relay@sha256:19f2dee371523888985e6e34b30d07ef3f84f2a9206ad79bdfc8dede96e25d5c
        - quay.io/cilium/clustermesh-apiserver@sha256:61f8e871760dceaa0c0459bbf44a3049dd0f4ab15e57e063d9b3284434cb3dd8
        - quay.io/cilium/hubble-ui@sha256:e15af59a2ded739e420be82de6cbdd0ce22f8d3f00e3a10b3d2d2734e074a394
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/cilium/cilium@sha256:b997994ab759443d9fc9807eb39b4784bf03e13730a8132ffe5df3edef22a312
        - quay.io/cilium/operator-generic@sha256:da4d72facecde862c88d08bf99c3f476845beb407052a0cc4c06b68cc93b7570
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:d35994cf5cf319b88ff5a7985cc972edd23df5a282360f9ba5516d610390ebf8
        version: 1.12.19+xef55de8
      entries:
      - name: cilium.v1.12.19-xef55de8
        version: 1.12.19+xef55de8
      - name: cilium.v1.12.18-x63f4f4c
        version: 1.12.18+x63f4f4c
      - name: cilium.v1.12.17-x6cc0c18
        version: 1.12.17+x6cc0c18
      - name: cilium.v1.12.16-x14b1d1d
        version: 1.12.16+x14b1d1d
      - name: cilium.v1.12.15-x7d84617
        version: 1.12.15+x7d84617
      - name: cilium.v1.12.14-x17ed60a
        version: 1.12.14+x17ed60a
      - name: cilium.v1.12.13-xb414403
        version: 1.12.13+xb414403
      - name: cilium.v1.12.12-xd280cc3
        version: 1.12.12+xd280cc3
      - name: cilium.v1.12.11-xc8fb7eb
        version: 1.12.11+xc8fb7eb
      - name: cilium.v1.12.10-xe372776
        version: 1.12.10+xe372776
      - name: cilium.v1.12.9-x2ccb2b6
        version: 1.12.9+x2ccb2b6
      - name: cilium.v1.12.8-xa53bac1
        version: 1.12.8+xa53bac1
      - name: cilium.v1.12.7-xba75145
        version: 1.12.7+xba75145
      - name: cilium.v1.12.0-xb42b669
        version: 1.12.0+xb42b669
      name: "1.12"
    - currentCSV: cilium.v1.13.12-x6e56d47
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.13.12+x6e56d47'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:b0d22d4376b4c2b2b3b988955227f1ed2ad49a363c5367e53f64edae2cec0ff5
        - quay.io/cilium/cilium@sha256:d99204aa7b3b7bd2c9ab47fd398cc9f40290799bc0c7a4386c8dc5c1780cd3d3
        - quay.io/cilium/clustermesh-apiserver@sha256:f1b1d0a85bab65e7d6adc90d000513a56ac58bdb071aa391a8580d73d20e6b6a
        - quay.io/cilium/operator-generic@sha256:f83734bbe270f961d545c7929152785507ce04a05d818ebc9776941723736d02
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/hubble-relay@sha256:01b23ea40bcd81145dde6bfcbfc4d542749d08c2a1c6348954c85123a8d2b1fc
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        version: 1.13.12+x6e56d47
      entries:
      - name: cilium.v1.13.12-x6e56d47
        version: 1.13.12+x6e56d47
      - name: cilium.v1.13.11-xba1ab0b
        version: 1.13.11+xba1ab0b
      - name: cilium.v1.13.10-x412ae64
        version: 1.13.10+x412ae64
      - name: cilium.v1.13.9-x3ae7099
        version: 1.13.9+x3ae7099
      - name: cilium.v1.13.8-xe0355c7
        version: 1.13.8+xe0355c7
      - name: cilium.v1.13.7-x37f9269
        version: 1.13.7+x37f9269
      - name: cilium.v1.13.6-xdb5afcd
        version: 1.13.6+xdb5afcd
      - name: cilium.v1.13.5-xaf5caa6
        version: 1.13.5+xaf5caa6
      - name: cilium.v1.13.4-x129f2a7
        version: 1.13.4+x129f2a7
      - name: cilium.v1.13.3-x5854d64
        version: 1.13.3+x5854d64
      - name: cilium.v1.13.2-x82c79c4
        version: 1.13.2+x82c79c4
      - name: cilium.v1.13.1-x10a5de9
        version: 1.13.1+x10a5de9
      - name: cilium.v1.13.0-x32540df
        version: 1.13.0+x32540df
      name: "1.13"
    - currentCSV: cilium.v1.14.7-x3522df7
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.14.7+x3522df7'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:291a05c58de8d7daefeb0689c23cadd92c9368aea09751de70eaa817f7fcc358
        - quay.io/cilium/operator-generic@sha256:37ef0bd85c27c765c637cd58c3ff4a559f8734ae39f9d1839a3ac7803de7b952
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/cilium@sha256:45ce2b87696082ecf7d53ba1c64ceeb4217578033e5ef28ac479ec049a48bc32
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/hubble-relay@sha256:46762393daf4a0aaef76b106614c2615942f98f10aeacd435ea3fb1a0bdf69e4
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/clustermesh-apiserver@sha256:28f3ffe53365ca79831af600f09a95c0b3e9959f5f891b416dab8cedd90c263d
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        version: 1.14.7+x3522df7
      entries:
      - name: cilium.v1.14.7-x3522df7
        version: 1.14.7+x3522df7
      - name: cilium.v1.14.6-x11464e1
        version: 1.14.6+x11464e1
      - name: cilium.v1.14.5-x3583390
        version: 1.14.5+x3583390
      - name: cilium.v1.14.4-xb89471d
        version: 1.14.4+xb89471d
      - name: cilium.v1.14.3-x041dc10
        version: 1.14.3+x041dc10
      - name: cilium.v1.14.2-xedd5ce2
        version: 1.14.2+xedd5ce2
      - name: cilium.v1.14.1-xdd859b7
        version: 1.14.1+xdd859b7
      name: "1.14"
    - currentCSV: cilium.v1.15.1-x7095b76
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.15.1+x7095b76'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:9ab6be29447125e886300e9258b9a06bedf0a9d87405832aa8b6565ed1ba4215
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        - quay.io/cilium/cilium@sha256:351d6685dc6f6ffbcd5451043167cfa8842c6decf80d8c8e426a417c73fb56d4
        - quay.io/cilium/clustermesh-apiserver@sha256:b353badd255c2ce47eaa8f394ee4cbf70666773d7294bd887693e0c33503dc37
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/operator-generic@sha256:819c7281f5a4f25ee1ce2ec4c76b6fbc69a660c68b7825e9580b1813833fa743
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/hubble-relay@sha256:3254aaf85064bc1567e8ce01ad634b6dd269e91858c83be99e47e685d4bb8012
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        version: 1.15.1+x7095b76
      entries:
      - name: cilium.v1.15.1-x7095b76
        version: 1.15.1+x7095b76
      name: "1.15"
    - currentCSV: cilium.v1.9.14-x70dc065
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Basic Install
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/hubble-relay@sha256:fd6ab1aea260abc5f64eca26c1b1e7009983e4aaa8e5d098e8d442f7659603fb
        - quay.io/cilium/clustermesh-apiserver@sha256:a0da5edf0372899647da51de1b277f0bab8e676d694aee7f939cddfdd3172010
        - quay.io/cilium/hubble-ui-backend@sha256:2cf1c7b1fde6161f575f3147c2ef885739e60b8c59a1ed444fe74edbdf22c7c1
        - quay.io/cilium/cilium@sha256:2c6ce93fa7e625979043a387eb998c17ad57df8768d89facb9b715da42a4c51c
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:3a9ca9a4c2f78109f265c4758f5f693d01865f2d0e31bc5f1ecbc065e878628b
        - docker.io/envoyproxy/envoy@sha256:55a141c0b8e39497b32074cb758b4e5dc2384bd6d9926144969a4452eaa611fd
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/hubble-ui@sha256:c267219946b84f79a5fd2849f5cc71f39453e29d93039e57b9cef2b5802e9c50
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/cilium/operator-generic@sha256:bdcfd2eade99933f2fda55ef79ea697ddfad3512f65b15bcd0ba7702518c1ba3
        version: 1.9.14+x70dc065
      entries:
      - name: cilium.v1.9.14-x70dc065
        version: 1.9.14+x70dc065
      name: "1.9"
    - currentCSV: cilium.v1.10.4-x5bfd7b3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Basic Install
          categories: Networking,Security
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:b45ff6a6552d71cbc26e358e0b25759d33818fd92301f1656453c5bcf826ff13
        version: 1.10.4+x5bfd7b3
      entries:
      - name: cilium.v1.10.4-x5bfd7b3
        version: 1.10.4+x5bfd7b3
      name: stable
    defaultChannel: "1.15"
    packageName: cilium
    provider:
      name: Isovalent
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Couchbase
      provider-url: ""
    name: couchbase-enterprise-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: couchbase-operator.v2.6.4-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator@sha256:1f0a6865c360a14ba98706fd6fe8dff3bdbdede32e3622f93f19c82e377ca33f
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        version: 2.6.4-7
      entries:
      - name: couchbase-operator.v2.6.4-7
        version: 2.6.4-7
      - name: couchbase-operator.v2.6.4-6
        version: 2.6.4-6
      name: 2.6.4
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator@sha256:707571a7881bc12d5f1b0b1a94eb3c675d5c0ea0114c994a47ceb5aa79df4013
        version: 2.7.0-5
      entries:
      - name: couchbase-operator.v2.7.0-5
        version: 2.7.0-5
      name: 2.7.0
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:a79e1b6781d5f8badf664e7efea6e2889e43fae6e0d37169d766685dcd6ff394
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        version: 2.7.1-1
      entries:
      - name: couchbase-operator.v2.7.1-1
        version: 2.7.1-1
      name: 2.7.1
    - currentCSV: couchbase-operator.v2.8.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:df61928b064014d851c6e7dff6dcd63a433a27fa5623254b6e389c3eea170e36
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        version: 2.8.0-2
      entries:
      - name: couchbase-operator.v2.8.0-2
        version: 2.8.0-2
      - name: couchbase-operator.v2.8.0-1
        version: 2.8.0-1
      name: 2.8.0
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        version: 2.8.1-1
      entries:
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: 2.8.1
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        version: 2.8.1-1
      entries:
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: stable
    defaultChannel: stable
    packageName: couchbase-enterprise-certified-rhmp
    provider:
      name: Couchbase
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: metallb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: metallb-operator.v4.16.0-202509230119
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "basic",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64512,
                        "neighbors": [
                          {
                            "address": "172.30.0.3",
                            "asn": 64512,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            },
                            "toReceive": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "two-neighs",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64512,
                        "neighbors": [
                          {
                            "address": "172.30.0.3",
                            "asn": 4200000000,
                            "ebgpMultiHop": true,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          },
                          {
                            "address": "172.18.0.6",
                            "asn": 4200000000,
                            "port": 179,
                            "toAdvertise": {
                              "allowed": {
                                "prefixes": [
                                  "192.168.2.0/24"
                                ]
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "two-routers-vrf",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64513,
                        "neighbors": [
                          {
                            "address": "172.32.0.3",
                            "asn": 4200000000,
                            "ebgpMultiHop": true,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ]
                      },
                      {
                        "asn": 64515,
                        "neighbors": [
                          {
                            "address": "172.31.0.3",
                            "asn": 4200000000,
                            "port": 179,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ],
                        "vrf": "red"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRNodeState",
                "metadata": {
                  "name": "frr-k8s-worker"
                },
                "status": {
                  "lastConversionResult": "success",
                  "lastReloadResult": "success",
                  "runningConfig": "Building configuration...\n\nCurrent configuration:\n!\nfrr version 9.0.2_git\nfrr defaults traditional\nhostname frr-k8s-worker\nlog file /etc/frr/frr.log informational\nlog timestamp precision 3\nno ipv6 forwarding\nservice integrated-vtysh-config\n!\nip nht resolve-via-default\n!\nipv6 nht resolve-via-default\n!\nend\n"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BFDProfile",
                "metadata": {
                  "name": "bfd-profile-sample",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "receiveInterval": 380,
                  "transmitInterval": 270
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "peers": [
                    "ebgp-single-hop0"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPoolSelectors": [
                    {
                      "matchLabels": {
                        "test": "ipv6"
                      }
                    },
                    {
                      "matchLabels": {
                        "test": "ipv4"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "communities": [
                    "65535:65282"
                  ],
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "localPref": 50
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample4",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "communities": [
                    "NO_ADVERTISE"
                  ],
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "localPref": 50
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample5",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "nodeSelectors": [
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-control-plane"
                      }
                    },
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-worker"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-beta1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "Community",
                "metadata": {
                  "name": "community1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "communities": [
                    {
                      "name": "NO_ADVERTISE",
                      "value": "65535:65282"
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "name": "ip-addresspool-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "192.168.10.0/24",
                    "192.168.9.1-192.168.9.5",
                    "fc00:f853:0ccd:e799::/124"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "labels": {
                    "test": "ipv4"
                  },
                  "name": "ip-addresspool-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "172.20.0.100/24"
                  ],
                  "autoAssign": false
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "labels": {
                    "test": "ipv6"
                  },
                  "name": "ip-addresspool-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "2002:2:2::1-2002:2:2::100"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "L2Advertisement",
                "metadata": {
                  "name": "l2-adv-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "L2Advertisement",
                "metadata": {
                  "name": "l2-adv-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "nodeSelectors": [
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-control-plane"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "MetalLB",
                "metadata": {
                  "name": "metallb",
                  "namespace": "metallb-system"
                },
                "spec": {}
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "bfdProfile": "bfd-profile-sample",
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "holdTime": "9m0s",
                  "keepaliveTime": "0s",
                  "myASN": 64512,
                  "passwordSecret": {
                    "name": "secretname",
                    "namespace": "metallb-system"
                  },
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3",
                  "peerPort": 180
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:b07c046ba2ae71b93f27d5cdca7cb93f51f3198cb84bb6e65dd6320bcda268c8
          createdAt: "2023-06-06T15:25:00Z"
          description: An operator for deploying MetalLB on a kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.8.0 <4.16.0'
          operatorframework.io/suggested-namespace: metallb-system
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/metallb-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BFDProfile is the Schema for the bfdprofiles API
            kind: BFDProfile
            name: bfdprofiles.metallb.io
            version: v1beta1
          - description: BGPAdvertisement is the Schema for the bgpadvertisements
              API
            kind: BGPAdvertisement
            name: bgpadvertisements.metallb.io
            version: v1beta1
          - description: BGPPeer is the Schema for the bgppeers API
            kind: BGPPeer
            name: bgppeers.metallb.io
            version: v1beta1
          - description: BGPPeer is the Schema for the bgppeers API
            kind: BGPPeer
            name: bgppeers.metallb.io
            version: v1beta2
          - description: Community is the Schema for the communities API
            kind: Community
            name: communities.metallb.io
            version: v1beta1
          - description: FRRConfiguration is the Schema for the frrconfigurations
              API
            kind: FRRConfiguration
            name: frrconfigurations.frrk8s.metallb.io
            version: v1beta1
          - description: FRRNodeState is the Schema for the frrnodestates API
            kind: FRRNodeState
            name: frrnodestates.frrk8s.metallb.io
            version: v1beta1
          - description: IPAddressPool is the Schema for the ipaddresspools API
            kind: IPAddressPool
            name: ipaddresspools.metallb.io
            version: v1beta1
          - description: L2Advertisement is the Schema for the l2advertisements API
            kind: L2Advertisement
            name: l2advertisements.metallb.io
            version: v1beta1
          - description: MetalLB is the Schema for the metallbs API
            displayName: MetalLB
            kind: MetalLB
            name: metallbs.metallb.io
            version: v1beta1
        description: An operator for deploying MetalLB on a kubernetes cluster.
        displayName: MetalLB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - metallb-operator
        links:
        - name: MetalLB Operator
          url: https://github.com/openshift/metallb-operator
        maintainers:
        - email: fpaoline@redhat.com
          name: Federico Paolinelli
        - email: obraunsh@redhat.com
          name: Ori Braunshtein
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/metallb-rhel9@sha256:95abda97d689078bd124ef59c92cc12e661222cac8e8b0e526579470ab54d05e
        - registry.redhat.io/openshift4/frr-rhel9@sha256:48a361644b2fa8fc632f49117172db2e9c12022774e7e0fd409561d884a6c237
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:b07c046ba2ae71b93f27d5cdca7cb93f51f3198cb84bb6e65dd6320bcda268c8
        version: 4.16.0-202509230119
      entries:
      - name: metallb-operator.v4.16.0-202509230119
        version: 4.16.0-202509230119
      - name: metallb-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: metallb-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: metallb-operator.v4.16.0-202507221936
        version: 4.16.0-202507221936
      - name: metallb-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: metallb-operator.v4.16.0-202506090335
        version: 4.16.0-202506090335
      - name: metallb-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: metallb-operator.v4.16.0-202505051339
        version: 4.16.0-202505051339
      - name: metallb-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: metallb-operator.v4.16.0-202503252033
        version: 4.16.0-202503252033
      - name: metallb-operator.v4.16.0-202502260004
        version: 4.16.0-202502260004
      - name: metallb-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: metallb-operator.v4.16.0-202501232204
        version: 4.16.0-202501232204
      - name: metallb-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: metallb-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: metallb-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: metallb-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: metallb-operator.v4.16.0-202411110236
        version: 4.16.0-202411110236
      - name: metallb-operator.v4.16.0-202410292005
        version: 4.16.0-202410292005
      - name: metallb-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: metallb-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: metallb-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: metallb-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: metallb-operator.v4.16.0-202409161407
        version: 4.16.0-202409161407
      - name: metallb-operator.v4.16.0-202409062036
        version: 4.16.0-202409062036
      - name: metallb-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: metallb-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: metallb-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: metallb-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: metallb-operator.v4.16.0-202408051837
        version: 4.16.0-202408051837
      - name: metallb-operator.v4.16.0-202407221107
        version: 4.16.0-202407221107
      - name: metallb-operator.v4.16.0-202407180936
        version: 4.16.0-202407180936
      - name: metallb-operator.v4.16.0-202407091337
        version: 4.16.0-202407091337
      - name: metallb-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: metallb-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: metallb-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: metallb-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: metallb-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app: enmasse
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq-online
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-online.1.7.4-0.1657725339.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "StandardInfraConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "broker": {
                    "resources": {
                      "memory": "1Gi",
                      "storage": "5Gi"
                    },
                    "addressFullPolicy": "FAIL"
                  },
                  "router": {
                    "linkCapacity": 50,
                    "resources": {
                      "memory": "512Mi"
                    }
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "BrokeredInfraConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "broker": {
                    "resources": {
                      "memory": "4Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta2",
                "kind": "AddressPlan",
                "metadata": {
                  "name": "standard-small-queue"
                },
                "spec": {
                  "addressType": "queue",
                  "shortDescription": "Small Queue",
                  "resources": {
                    "router": 0.01,
                    "broker": 0.1
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta2",
                "kind": "AddressSpacePlan",
                "metadata": {
                  "name": "standard-small"
                },
                "spec": {
                  "addressSpaceType": "standard",
                  "infraConfigRef": "default",
                  "shortDescription": "Small Address Space Plan",
                  "resourceLimits": {
                    "router": 1.0,
                    "broker": 2.0,
                    "aggregate": 2.0
                  },
                  "addressPlans": [
                    "standard-small-queue"
                  ]
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "AuthenticationService",
                "metadata": {
                  "name": "standard-authservice"
                },
                "spec": {
                  "type": "standard",
                  "standard": {
                    "storage": {
                      "claimName": "standard-authservice",
                      "deleteClaim": true,
                      "size": "1Gi",
                      "type": "persistent-claim"
                    }
                  }
                }
              },
              {
                "apiVersion": "enmasse.io/v1beta1",
                "kind": "AddressSpace",
                "metadata": {
                  "name": "myspace"
                },
                "spec": {
                  "plan": "standard-small",
                  "type": "standard"
                }
              },
              {
                "apiVersion": "enmasse.io/v1beta1",
                "kind": "Address",
                "metadata": {
                  "name": "myspace.myqueue"
                },
                "spec": {
                  "address": "myqueue",
                  "plan": "standard-small-queue",
                  "type": "queue"
                }
              },
              {
                "apiVersion": "user.enmasse.io/v1beta1",
                "kind": "MessagingUser",
                "metadata": {
                  "name": "myspace.user"
                },
                "spec": {
                  "authentication": {
                    "password": "ZW5tYXNzZQ==",
                    "type": "password"
                  },
                  "authorization": [
                    {
                      "addresses": [
                        "myqueue"
                      ],
                      "operations": [
                        "send",
                        "recv"
                      ]
                    }
                  ],
                  "username": "user"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator@sha256:2aaf87224a96aa56dc595e96d8e7ec8d138aaf20240afa25a1e805b405941355
          createdAt: "2019-02-19T00:00:00Z"
          description: Red Hat Integration - AMQ Online provides messaging as a managed
            service on OpenShift
          olm.substitutesFor: amq-online.1.7.4
          operators.openshift.io/capability: '["license-required"]'
          operators.openshift.io/valid-licenses: '["Red Hat AMQ", "Red Hat Integration"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: ""
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A messaging user that can connect to an Address Space
            displayName: Messaging User
            kind: MessagingUser
            name: messagingusers.user.enmasse.io
            version: v1beta1
          - description: A resource representing the available schema of plans and
              authentication services
            displayName: AddressSpaceSchema
            kind: AddressSpaceSchema
            name: addressspaceschemas.enmasse.io
            version: v1beta1
          - description: A group of messaging addresses that can be accessed via the
              same endpoint
            displayName: Address Space
            kind: AddressSpace
            name: addressspaces.enmasse.io
            version: v1beta1
          - description: A messaging address that can be used to send/receive messages
              to/from
            displayName: Address
            kind: Address
            name: addresses.enmasse.io
            version: v1beta1
          - description: Infrastructure configuration template for the standard address
              space type
            displayName: Standard Infra Config
            kind: StandardInfraConfig
            name: standardinfraconfigs.admin.enmasse.io
            version: v1beta1
          - description: Infrastructure configuration template for the brokered address
              space type
            displayName: Brokered Infra Config
            kind: BrokeredInfraConfig
            name: brokeredinfraconfigs.admin.enmasse.io
            version: v1beta1
          - description: Plan describing the resource usage of a given address type
            displayName: Address Plan
            kind: AddressPlan
            name: addressplans.admin.enmasse.io
            version: v1beta2
          - description: Plan describing the capabilities and resource limits of a
              given address space type
            displayName: Address Space Plan
            kind: AddressSpacePlan
            name: addressspaceplans.admin.enmasse.io
            version: v1beta2
          - description: Authentication service configuration available to address
              spaces.
            displayName: Authentication Service
            kind: AuthenticationService
            name: authenticationservices.admin.enmasse.io
            version: v1beta1
          - description: Console Service Singleton for deploying global console.
            displayName: Console Service
            kind: ConsoleService
            name: consoleservices.admin.enmasse.io
            version: v1beta1
        description: "**Red Hat Integration - AMQ Online** provides messaging as a
          managed service on OpenShift.\nWith Red Hat Integration - AMQ Online, administrators
          can configure a cloud-native,\nmulti-tenant messaging service either in
          the cloud or on premise.\nDevelopers can provision messaging using the Red
          Hat Integration - AMQ Online Console.\nMultiple development teams can provision
          the brokers and queues from the\nconsole, without requiring each team to
          install, configure, deploy,\nmaintain, or patch any software. \n\n**The
          core capabilities include:**\n\n  * **Built-in authentication and authorization**
          - Use the built-in authentication service or\n    plug in your own authentication
          service for authentication and\n    authorization of messaging clients.\n\n
          \ * **Self-service messaging for applications** - The service administrator
          deploys\n    and manages the messaging infrastructure, while applications
          can request\n    messaging resources regardless of the messaging infrastructure.\n\n
          \ * **Support for a wide variety of messaging patterns** - Choose between\n
          \   JMS-style messaging with strict guarantees, or messaging that supports\n
          \   a larger number of connections and higher throughput.\n\n## Post-installation
          tasks\n\nTo fully use Red Hat Integration - AMQ Online, you need to create
          a few\ninfrastructure components after the installation, including:\n\n
          \ * An authentication service\n  * Infrastructure configuration for the
          standard and brokered address space\n  * Address and address space plans\n
          \ * (Optional) Create RBAC roles to allow users to discover available plans\n
          \ * (Optional) Create RBAC roles to allow users to self-manage addresses
          and\n    address spaces.\n\nFor a complete overview of how to configure
          Red Hat Integration - AMQ Online, see\n[Configuring Red Hat Integration
          - AMQ Online](https://access.redhat.com/documentation/en-us/red_hat_amq/7.7/html/installing_and_managing_amq_online_on_openshift/configuring-messaging).\n\n###
          Quickstart infrastructure configuration\n\nIf you simply want to get started
          quickly, you can import the following\nYAML by saving the content to a file
          and apply it by running the \n`oc apply -f <file>` command. You can also
          split the content (at the `---` marker)\nand import the single YAML document
          using the Web UI: \n\n~~~yaml\n---\napiVersion: admin.enmasse.io/v1beta1\nkind:
          StandardInfraConfig\nmetadata:\n  name: default\n---\napiVersion: admin.enmasse.io/v1beta2\nkind:
          AddressPlan\nmetadata:\n  name: standard-small-queue\nspec:\n  addressType:
          queue\n  resources:\n    router: 0.01\n    broker: 0.1\n---\napiVersion:
          admin.enmasse.io/v1beta2\nkind: AddressSpacePlan\nmetadata:\n  name: standard-small\nspec:\n
          \ addressSpaceType: standard\n  infraConfigRef: default\n  addressPlans:\n
          \ - standard-small-queue\n  resourceLimits:\n    router: 2.0\n    broker:
          3.0\n    aggregate: 4.0\n---\napiVersion: admin.enmasse.io/v1beta1\nkind:
          AuthenticationService\nmetadata:\n  name: standard-authservice\nspec:\n
          \ type: standard\n  standard:\n    storage:\n      claimName: standard-authservice\n
          \     deleteClaim: true\n      size: 1Gi\n      type: persistent-claim\n~~~\n\n###
          Create RBAC roles to allow users to discover available plans\n\nFor users
          to discover the available plans, cluster-wide roles to read the available\nschema
          can be created.  Import the following YAML by saving the content to a file
          and apply it by running the \n`oc apply -f <file>` command. You can also
          split the content (at the `---` marker)\nand import the single YAML document
          using the Web UI: \n\n~~~yaml\n---\napiVersion: rbac.authorization.k8s.io/v1\nkind:
          ClusterRole\nmetadata:\n  name: enmasse.io:schema\n  labels:\n    app: enmasse\nrules:\n
          \ - apiGroups: [ \"enmasse.io\" ]\n    resources: [ \"addressspaceschemas\"
          ]\n    verbs: [ \"get\", \"list\", \"watch\" ]\n---\napiVersion: rbac.authorization.k8s.io/v1\nkind:
          ClusterRoleBinding\nmetadata:\n  name: \"enmasse.io:schema\"\n  labels:\n
          \   app: enmasse\nroleRef:\n  apiGroup: rbac.authorization.k8s.io\n  kind:
          ClusterRole\n  name: enmasse.io:schema\nsubjects:\n- kind: Group\n  name:
          system:authenticated\n~~~\n\n### Creating infrastructure using the Web UI\n\nYou
          must create a new instance of each of the following custom resources. You
          can\nuse the example data directly, which is provided when using the\nWeb
          UI:\n\n  * *Authentication Service* – Create an authentication service.\n
          \ * *Brokered Infra Config* – Create the broker infrastructure configuration.\n
          \ * *Standard Infra Config* – Create the standard infrastructure\n    configuration.\n\nYou
          must also create at least one address space plan and one address plan.\n\n*Note*:
          The name of the address space plan and address plan is required\nlater when
          creating new addresses. Some examples use specific plan\nnames, which might
          not be available in your environment. You can\ncreate those plans, or edit
          the examples to use different plan names.\n\n## Configuration for messaging
          tenants\n\nWhile service administrators perform the infrastructure configuration,
          the following\nresources are for the actual users of the system, the messaging
          tenants.\n\nYou need to create those resources to satisfy your particular
          use case.\n\n  * *Address space* – A container for addresses\n  * *Address*
          – A messaging address (address, topic, queue, and so on)\n  * *Messaging
          user* – Manages access to an address\n"
        displayName: Red Hat Integration - AMQ Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amq
        - messaging
        - amqp
        - iot
        - mqtt
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/7.7/html/using_amq_online_on_openshift/index
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-online-1-topic-forwarder@sha256:de33a701339d91a4e9275f7cf65f917ff75368a59b8ab36dd66c90bc306875aa
        - registry.redhat.io/amq7/amq-online-1-broker-plugin@sha256:2466a336c0d78b196635de3fad636aeff2d9265839f7271b8ef94f3ba97d9bc0
        - registry.redhat.io/amq7/amq-interconnect@sha256:5ebd4c4a0ed0db2725ae466550c915531b32933c420344cc2083c616da3f326f
        - registry.redhat.io/redhat-sso-7/sso73-openshift@sha256:18d15980f28f0d0c6b8582131a1f062d84463577512875b5391a612c3d7d1096
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6d489a7897bd831ec273034f1ff0e223427a40afd01b9835ebb020a8bf9ddc66
        - registry.redhat.io/amq7/amq-broker@sha256:1ec0437bbd82722148fdbb02dc55682b114c311cd6f521a8bc621a8c1cb4b32b
        - registry.redhat.io/amq7/amq-online-1-address-space-controller@sha256:9acc555c8e869aec7d72f53b60320e1a96cc82d748638e0d50970837229fdbee
        - registry.redhat.io/amq7/amq-online-1-none-auth-service@sha256:a013fb1806c002b408a7753ff943cee24fbcdf8a657eff1e9582c42b5a7b0844
        - registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator@sha256:2aaf87224a96aa56dc595e96d8e7ec8d138aaf20240afa25a1e805b405941355
        - registry.redhat.io/amq7/amq-online-1-standard-controller@sha256:070349af30cec52fa8efc185fc940077a64bb40e3e36310d7cc6e39ce0971436
        - registry.redhat.io/amq7/amq-online-1-console-init@sha256:0970804909b82fdc3d4e7d5d88280e41157d743107de8bfa2d1baa1367543176
        - registry.redhat.io/openshift3/oauth-proxy@sha256:59b20f23c03d0012ec5c95c4725892ee64277c22dedc9d0a90a7b6b0bc80fc72
        - registry.redhat.io/amq7/amq-online-1-console-server-rhel7@sha256:1dcb3d0cc7bc6e84f4f96962ff88849511b7a7686a1f5dbd9300c4c2c8c05e2f
        - registry.redhat.io/amq7/amq-online-1-agent@sha256:f98ed8ee97ac59d8fdd40ec8b2c063a5afd360a3e183057c186e7a77a32801ee
        - registry.redhat.io/amq7/amq-online-1-auth-plugin@sha256:68926f434b2f6ce4d9698938bc3c53d514c21c3f76c73b6b4cdf7fcf51f7cb0d
        version: 1.7.4+0.1657725339.p
      entries:
      - name: amq-online.1.7.4
        version: 1.7.4
      - name: amq-online.1.7.4-0.1657725339.p
        version: 1.7.4+0.1657725339.p
      - name: amq-online.1.7.3
        version: 1.7.3
      - name: amq-online.1.7.2
        version: 1.7.2
      - name: amq-online.1.7.1
        version: 1.7.1
      - name: amq-online.1.7.1-0.1628610187.p
        version: 1.7.1+0.1628610187.p
      - name: amq-online.1.7.0
        version: 1.7.0
      name: stable
    defaultChannel: stable
    packageName: amq-online
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: sandboxed-containers-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sandboxed-containers-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:051bd7f1dad8cc3251430fee32184be8d64077aba78580184cef0255d267bdcf
          createdAt: 2021-06-07-13:38:06
          olm.skipRange: <1.0.2
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          OpenShift sandboxed containers based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift’s Defense-in-Depth strategy. For more information
          [see](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5).

          # Features & benefits
          - **Isolated Developer Environments & Priviliges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as CAP_ADMIN or CAP_BPF. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select preview-1.0 from the list of available Update Channel options.
            This ensures that you install the version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              openshift-sandboxed-containers-operator namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              openshift-sandboxed-containers-operator causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a KataConfig CRD instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.8/sandboxed_containers/understanding-sandboxed-containers.html).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kernel-isolated containers
        links:
        - name: OpenShift sandboxed containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        provider:
          name: Red Hat
          url: https://www.github.com/openshift/sandboxed-containers-operator
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:051bd7f1dad8cc3251430fee32184be8d64077aba78580184cef0255d267bdcf
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:c6c589d5e47ba9564c66c84fc2bc7e5e046dae1d56a3dc99d7343f01e42e4d31
        version: 1.0.2
      entries:
      - name: sandboxed-containers-operator.v1.0.2
        version: 1.0.2
      name: preview-1.0
    - currentCSV: sandboxed-containers-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:b1e824e126c579db0f56d04c3d1796d82ed033110c6bc923de66d95b67099611
          createdAt: "2021-10-08T13:20:20.250789Z"
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift’s Defense-in-Depth strategy. For more information
          [see](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5).

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `preview-1.1` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CRD instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.8/sandboxed_containers/understanding-sandboxed-containers.html).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kernel-isolated containers
        links:
        - name: OpenShift sandboxed containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        provider:
          name: Red Hat
          url: https://www.github.com/openshift/sandboxed-containers-operator
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:b1e824e126c579db0f56d04c3d1796d82ed033110c6bc923de66d95b67099611
        version: 1.1.0
      entries:
      - name: sandboxed-containers-operator.v1.1.0
        version: 1.1.0
      - name: sandboxed-containers-operator.v1.0.2
        version: 1.0.2
      name: preview-1.1
    - currentCSV: sandboxed-containers-operator.v1.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-12T14:30:28Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.10.3'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '["peerpods.confidentialcontainers.org","peerpodconfigs.confidentialcontainers.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/sandboxed-containers-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
          - kind: PeerPod
            name: peerpods.confidentialcontainers.org
            version: v1alpha1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `candidate` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the [OpenShift sandboxed containers documentation](https://docs.redhat.com/en/documentation/openshift_sandboxed_containers/).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-cloud-api-adaptor-rhel9@sha256:7d1fa1e85af0d34c23e75868b1388eff17abe633c04a8dc55fbefcc470195c05
        - registry.redhat.io/openshift-sandboxed-containers/osc-cloud-api-adaptor-webhook-rhel9@sha256:6802f00a939f24ecd6dad150433db5ea3fd72d171f0c410cb894d3af910fc7fd
        - registry.redhat.io/openshift-sandboxed-containers/osc-podvm-builder-rhel9@sha256:c15f36a639fa07fee11f74c6563696ec8ef4da3fb1a99b7bd3f308dd04881339
        - registry.redhat.io/openshift-sandboxed-containers/osc-podvm-payload-rhel9@sha256:8d6b1262c6cedd04e33441264539842d82ed057f64eda5c4bce59566bcb72558
        - registry.redhat.io/openshift-sandboxed-containers/osc-dm-verity-image@sha256:e633bc27853f811d919e64461d0c917e659f5430458abe2304a4c85b0d871c80
        - registry.redhat.io/openshift-sandboxed-containers/osc-monitor-rhel9@sha256:037cbc221ae8fda83cdf1093127795ea806ecda53893deb49cdf981e7863774c
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel9-operator@sha256:1ffb432d9a8833854fa5e5262dcbf42530dfbaadfddf6978f7bc7804b84d2e3f
        version: 1.10.3
      entries:
      - name: sandboxed-containers-operator.v1.10.3
        version: 1.10.3
      - name: sandboxed-containers-operator.v1.10.2
        version: 1.10.2
      - name: sandboxed-containers-operator.v1.10.1
        version: 1.10.1
      - name: sandboxed-containers-operator.v1.10.0
        version: 1.10.0
      - name: sandboxed-containers-operator.v1.9.0
        version: 1.9.0
      - name: sandboxed-containers-operator.v1.8.1
        version: 1.8.1
      - name: sandboxed-containers-operator.v1.7.0
        version: 1.7.0
      - name: sandboxed-containers-operator.v1.6.0
        version: 1.6.0
      - name: sandboxed-containers-operator.v1.5.3
        version: 1.5.3
      - name: sandboxed-containers-operator.v1.5.2
        version: 1.5.2
      - name: sandboxed-containers-operator.v1.5.1
        version: 1.5.1
      - name: sandboxed-containers-operator.v1.5.0
        version: 1.5.0
      name: stable
    - currentCSV: sandboxed-containers-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:8f30a9129d817c3f4e404d2c43fb47e196d8c8da3badba4c48f65d440a4d7584
          createdAt: "2021-11-22T14:44:50.817336Z"
          olm.skipRange: '>=1.1.0 <1.2.2'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `stable-1.2` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.9/sandboxed_containers/understanding-sandboxed-containers.html)
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:a660f0b54b9139bed9a3aeef3408001c0d50ba60648364a98a09059b466fbcc1
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:8f30a9129d817c3f4e404d2c43fb47e196d8c8da3badba4c48f65d440a4d7584
        version: 1.2.2
      entries:
      - name: sandboxed-containers-operator.v1.2.2
        version: 1.2.2
      name: stable-1.2
    - currentCSV: sandboxed-containers-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          olm.skipRange: '>=1.1.0 <1.3.3'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/sandboxed-containers-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `stable-1.3` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/latest/sandboxed_containers/understanding-sandboxed-containers.html)
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:5d2b03721043e5221dfb0cf164cf59eba396ba3aae40a56c53aa3496c625eea0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:cb836456974e510eb4bccbffadbc6d99d5f57c36caec54c767a158ffd8a025d5
        version: 1.3.3
      entries:
      - name: sandboxed-containers-operator.v1.3.3
        version: 1.3.3
      name: stable-1.3
    defaultChannel: stable
    packageName: sandboxed-containers-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/stolostron/multicluster-global-hub
    name: multicluster-global-hub-operator-rh
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-global-hub-operator-rh.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:01e8dc62d5dc48af84b92f5315fab22574f648f2f705ac94f0516f4a7dab9237
          createdAt: 27 Aug 2024, 22:21
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.1.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```


          ### Documentation

          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: release-1.1
        provider:
          name: Red Hat
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:01e8dc62d5dc48af84b92f5315fab22574f648f2f705ac94f0516f4a7dab9237
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:78628c4f7a0286417023410261cc3b36797b7cdf72b38d0f4a2533a0f3db89dc
        - registry.redhat.io/rhel9/postgresql-13@sha256:c67d00643295697d196dd70dea457948ebb9e9d9d61a42019490b350c8286b50
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:5c8cea421be39712503378e32c54c332a7a0bf77a1705781d10fc123224bd60e
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:4ea8d3a5317d3b9968fa93dd68416e07e12e00e492860ff20ed90f88ca6ba225
        version: 1.1.2
      entries:
      - name: multicluster-global-hub-operator-rh.v1.1.2
        version: 1.1.2
      - name: multicluster-global-hub-operator-rh.v1.1.1
        version: 1.1.1
      - name: multicluster-global-hub-operator-rh.v1.1.0
        version: 1.1.0
      name: release-1.1
    - currentCSV: multicluster-global-hub-operator-rh.v1.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:98d470c8683c346107ab943776991554e3173c9a6431202bdacc48a75c4cf18f
          createdAt: 15 May 2025, 21:22
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          kafka-zookeeper-0                                       1/1     Running   0          21m
          kafka-zookeeper-1                                       1/1     Running   0          21m
          kafka-zookeeper-2                                       1/1     Running   0          21m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgres-0                      2/2     Running   0          22m
          ```


          ### Documentation

          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: release-1.2
        provider:
          name: Red Hat
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:c83a7b01bfe5a39d6cdf585180a011448e935d109768cb51560308b29d9ffe14
        - registry.redhat.io/rhel9/postgresql-13@sha256:b1af637cbe0e4041cd2fe6fdbcf665c5183e2991d3cc3e5b7384235f801504d9
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:16bb76ad0c68c0d6b3149b2741126a96e154150c1059621b14c2e6c30a214f16
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:98d470c8683c346107ab943776991554e3173c9a6431202bdacc48a75c4cf18f
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:92f44cff160874efe57a985609e599b13c49c0c4cef846dccc80833af2ae74f3
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:64e5ade4fbcd4cb70d62a6a62d775e051ff49838b891b682574646d4e61d9751
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:9b98fc01b08c3262e5ab99096e74fb55982f2a9c0de61e8fe84e419522e725b0
        version: 1.2.3
      entries:
      - name: multicluster-global-hub-operator-rh.v1.2.3
        version: 1.2.3
      - name: multicluster-global-hub-operator-rh.v1.2.2
        version: 1.2.2
      - name: multicluster-global-hub-operator-rh.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator-rh.v1.2.0
        version: 1.2.0
      name: release-1.2
    - currentCSV: multicluster-global-hub-operator-rh.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:0ef24ded54588aa49e366ba1cafc0971162e9d32024045bbcd7ef16e9f38a2d5
          createdAt: 01 Apr 2025, 20:25
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          kafka-zookeeper-0                                       1/1     Running   0          21m
          kafka-zookeeper-1                                       1/1     Running   0          21m
          kafka-zookeeper-2                                       1/1     Running   0          21m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgres-0                      2/2     Running   0          22m
          ```


          ### Documentation

          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: release-1.3
        provider:
          name: Red Hat
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:a139da58c681de987d847b233dd5c1e6ced5d6e86f3aab0f1b4181a77fc583d4
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:a17e90abc5e23bbc3908613c56c4208ad554eb27ed9679c4ccd06c4a1e9fcbcc
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:0ef24ded54588aa49e366ba1cafc0971162e9d32024045bbcd7ef16e9f38a2d5
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:619d8cb6b3811169af183bd8bc26ed98cdb25486ec428d88b25fee0ef9e6c552
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:e848a90d17e4f00d8021cbde47e488b82c447d8a5f6588f85f3d7cfbd6877f0f
        - registry.redhat.io/rhel9/postgresql-13@sha256:76c49b7dbd7c4ed8b430c4931b32f546db6d07ddc2146232f5d17eb17d3243b4
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-kessel-inventory-api-rhel9@sha256:dbd3924eb24db7ef28301b17851b5b5cd20e044cf259f0c556c67a0721561803
        version: 1.3.3
      entries:
      - name: multicluster-global-hub-operator-rh.v1.3.3
        version: 1.3.3
      - name: multicluster-global-hub-operator-rh.v1.3.2
        version: 1.3.2
      - name: multicluster-global-hub-operator-rh.v1.3.1
        version: 1.3.1
      - name: multicluster-global-hub-operator-rh.v1.3.0
        version: 1.3.0
      name: release-1.3
    - currentCSV: multicluster-global-hub-operator-rh.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:c5b3b3907e8135d3eedba40e27ba18a232ea4f30854d652059fcb14f82a8bf15
          createdAt: "2025-07-23T09:04:08Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: release-1.4
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:047b6c378e2b37339dd498c9025060f20812014bde4d0c98183f1a1201c21e66
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:ea906dcf8d45956251df22e269fe0b8fa8d49734e96a61b10db7b25d626b2e6a
        - registry.redhat.io/rhel9/postgresql-16@sha256:2161182c88022f38d6d9b8b467c0d59156edcededc908d383b80caa4eedf57f9
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:2be8bc79667018386235823e322ae10788e4e7e889c10e63566ae27a65dcabf1
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:0f4683d18404b4a9412e42ed417f6d3ac0024b3c57aa72c9162ff72d6ea8d394
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:c5b3b3907e8135d3eedba40e27ba18a232ea4f30854d652059fcb14f82a8bf15
        version: 1.4.2
      entries:
      - name: multicluster-global-hub-operator-rh.v1.4.2
        version: 1.4.2
      - name: multicluster-global-hub-operator-rh.v1.4.1
        version: 1.4.1
      - name: multicluster-global-hub-operator-rh.v1.4.0
        version: 1.4.0
      name: release-1.4
    - currentCSV: multicluster-global-hub-operator-rh.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:8af7e57b8c6285ad232084e7776fef32d026e080eb0bf14b91986ac28d00bb6a
          createdAt: "2025-08-18T08:30:32Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: release-1.5
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:8af7e57b8c6285ad232084e7776fef32d026e080eb0bf14b91986ac28d00bb6a
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:c95052569312d805b6750477c8bc57f7c72e1efd26143f3899ea0eddc2aed37b
        - registry.redhat.io/rhel9/postgresql-16@sha256:2161182c88022f38d6d9b8b467c0d59156edcededc908d383b80caa4eedf57f9
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:921163d139bca9e47bf95df43d7b408ff6fe2bf9b0902fb6af9f4cfc9d4598cc
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:ab3ed25a4448a7d545648ae4cea52b493918cc4d3b0c5adc56bb2510c0838c80
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:eae390ec42c3582f7896a8346df8782b287d64440794983ae93755490639e21e
        version: 1.5.1
      entries:
      - name: multicluster-global-hub-operator-rh.v1.5.1
        version: 1.5.1
      - name: multicluster-global-hub-operator-rh.v1.5.0
        version: 1.5.0
      name: release-1.5
    defaultChannel: release-1.5
    packageName: multicluster-global-hub-operator-rh
    provider:
      name: Red Hat, Inc
      url: https://github.com/stolostron/multicluster-global-hub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-recyclebin-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-recyclebin-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "recyclebin.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/recyclebin-controller:1.1.1
          createdAt: "2025-09-20T01:20:03Z"
          description: AWS Recycle Bin controller is a service controller for managing
            Recycle Bin resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Rule represents the state of an AWS recyclebin Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.recyclebin.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Recycle Bin resources in AWS from within your Kubernetes cluster.

          **About Amazon Recycle Bin**

          Recycle Bin is a data recovery feature that enables you to restore accidentally deleted Amazon EBS snapshots and EBS-backed AMIs. When using Recycle Bin, if your resources are deleted, they are retained in the Recycle Bin for a time period that you specify before being permanently deleted.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Recycle Bin
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - recyclebin
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Recycle Bin Developer Resources
          url: https://docs.aws.amazon.com/ebs/latest/userguide/recycle-bin.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: recyclebin maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/recyclebin-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-recyclebin-controller.v1.1.1
        version: 1.1.1
      - name: ack-recyclebin-controller.v1.1.0
        version: 1.1.0
      - name: ack-recyclebin-controller.v1.0.13
        version: 1.0.13
      - name: ack-recyclebin-controller.v1.0.12
        version: 1.0.12
      - name: ack-recyclebin-controller.v1.0.11
        version: 1.0.11
      - name: ack-recyclebin-controller.v1.0.10
        version: 1.0.10
      - name: ack-recyclebin-controller.v1.0.9
        version: 1.0.9
      - name: ack-recyclebin-controller.v1.0.8
        version: 1.0.8
      - name: ack-recyclebin-controller.v1.0.7
        version: 1.0.7
      - name: ack-recyclebin-controller.v1.0.6
        version: 1.0.6
      - name: ack-recyclebin-controller.v1.0.5
        version: 1.0.5
      - name: ack-recyclebin-controller.v1.0.2
        version: 1.0.2
      - name: ack-recyclebin-controller.v1.0.1
        version: 1.0.1
      - name: ack-recyclebin-controller.v1.0.0
        version: 1.0.0
      - name: ack-recyclebin-controller.v0.0.3
        version: 0.0.3
      - name: ack-recyclebin-controller.v0.0.2
        version: 0.0.2
      - name: ack-recyclebin-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-recyclebin-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ian Boden
      provider-url: ""
    name: eclipse-amlen-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eclipse-amlen-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eclipse.org/v1",
                "kind": "Amlen",
                "metadata": {
                  "name": "amlen-sample"
                },
                "spec": {
                  "_amlen_image_tag": "latest",
                  "_amlen_monitor_image_tag": "latest",
                  "_amlen_name": "amlen-sample",
                  "_amlen_namespace": "osdk-test",
                  "_amlen_persistence": "false",
                  "config": "amlen-config",
                  "device_cert_issuer": {
                    "mode": "automatic",
                    "name": "amlen-selfsigned-issuer"
                  },
                  "ha": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "size": 2
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/amlen/operator:v1.0.1a
          createdAt: "2022-11-24T15:31:00Z"
          description: An operator to run the Eclipse Amlen MQTT Broker
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/eclipse/amlen
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Amlen
            name: amlens.eclipse.org
            version: v1
        description: Operator for the eclipse Amlen MQTT Broker\n\n### Supported Feature\n\nCurrently
          supports:\n* Install - simple install process\n* Upgrade - non-disruptive
          upgrade\n*High Availability - can be enabled/disabled via the CRD\n\n###
          Documentation\nThe main Amlen website has links to product documentation,
          forum, blog and the slack channel\nhttp://www.eclipse.org/amlen\n\n### Getting
          help\nThe quickest way to get help is to use the slack channel but the forum
          is monitored as well\n\nIssues can also be added to the github repository
          https://github.com/eclipse/amlen\n\n### License\nEclipse Amlen Operator
          is licensed under the [Eclipse Public License - v 2.0](https://www.eclipse.org/legal/epl-2.0/)
        displayName: eclipse-amlen-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amlen
        - mqtt
        - broker
        links:
        - name: Eclipse Amlen Operator
          url: https://eclipse.org/amlen
        maintainers:
        - email: ianboden@uk.ibm.com
          name: Ian Boden
        maturity: stable
        provider:
          name: Ian Boden
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/amlen/operator:v1.0.1a
        version: 1.0.0
      entries:
      - name: eclipse-amlen-operator.v1.0.0
        version: 1.0.0
      - name: eclipse-amlen-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: eclipse-amlen-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eclipse.org/v1",
                "kind": "Amlen",
                "metadata": {
                  "name": "amlen-sample"
                },
                "spec": {
                  "config": "amlen-config",
                  "device_cert_issuer": {
                    "mode": "automatic",
                    "name": "amlen-selfsigned-issuer"
                  },
                  "ha": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "ldap": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "size": 2
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/amlen/operator:v1.1.3
          createdAt: "2025-05-22T10:00:00Z"
          description: An operator to run the Eclipse Amlen MQTT Broker
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/eclipse-amlen/amlen
          support: Eclipse
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Amlen
            name: amlens.eclipse.org
            version: v1
        description: |-
          Operator for the eclipse Amlen MQTT Broker

          ### Supported Feature

          Currently supports:
          * Install - simple install process
          * Upgrade - non-disruptive upgrade
          * High Availability - can be enabled/disabled via the CRD

          ### Documentation
          The main Amlen website has links to product documentation, forum, blog and the slack channel
          http://www.eclipse.org/amlen

          ### Getting help
          The quickest way to get help is to use the slack channel but the forum is monitored as well
          Issues can also be added to the github repository https://github.com/eclipse-amlen/amlen

          ### License
          Eclipse Amlen Operator is licensed under [Eclipse Public License - v 2.0](https://www.eclipse.org/legal/epl-2.0/)
        displayName: eclipse-amlen-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amlen
        - mqtt
        - broker
        links:
        - name: Eclipse Amlen Operator
          url: https://eclipse.org/amlen
        maintainers:
        - email: ianboden@uk.ibm.com
          name: Ian Boden
        maturity: stable
        provider:
          name: Eclipse Amlen
          url: https://eclipse.org/amlen
        relatedImages:
        - controller:latest
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        version: 1.1.3
      entries:
      - name: eclipse-amlen-operator.v1.1.3
        version: 1.1.3
      - name: eclipse-amlen-operator.v1.1.0
        version: 1.1.0
      - name: eclipse-amlen-operator.v1.0.2
        version: 1.0.2
      - name: eclipse-amlen-operator.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: alpha
    packageName: eclipse-amlen-operator
    provider:
      name: Ian Boden
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: vault-config-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vault-config-operator.v0.8.34
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AuthEngineMount",
                "metadata": {
                  "name": "authenginemount-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kube-authengine-mount-sample",
                  "type": "kubernetes"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azureauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azureauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azureauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureCredentials": {
                    "passwordKey": "clientsecret",
                    "secret": {
                      "name": "aad-credentials"
                    },
                    "usernameKey": "clientid"
                  },
                  "clientID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "environment": "AzurePublicCloud",
                  "maxRetries": 3,
                  "maxRetryDelay": 60,
                  "path": "azure/test",
                  "resource": "https://management.azure.com/",
                  "retryDelay": 4,
                  "tenantID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azureauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azureauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azureauthenginerole-sample"
                },
                "spec": {
                  "BoundResourceGroups": [
                    "resourcegroup1",
                    "resourcegroup2"
                  ],
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "boundGroupIDs": [
                    "group1",
                    "group2"
                  ],
                  "boundLocations": [
                    "location1",
                    "location2"
                  ],
                  "boundScaleSets": [
                    "scaleset1",
                    "scaleset1"
                  ],
                  "boundServicePrincipalIDs": [
                    "sp1",
                    "sp2"
                  ],
                  "boundSubscriptionIDs": [
                    "subscription1",
                    "subscription2"
                  ],
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "dev-role",
                  "path": "azure",
                  "policies": [
                    "policy1",
                    "policy2"
                  ],
                  "tokenBoundCIDRs": [
                    "10.20.102.13",
                    "10.20.102.15"
                  ],
                  "tokenExplicitMaxTTL": "",
                  "tokenMaxTTL": "",
                  "tokenNoDefaultPolicy": false,
                  "tokenNumUses": 0,
                  "tokenPeriod": 0,
                  "tokenPolicies": [
                    "policy1",
                    "policy2"
                  ],
                  "tokenTTL": "",
                  "tokenType": ""
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureSecretEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azuresecretengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azuresecretengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azuresecretengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureCredentials": {
                    "passwordKey": "clientsecret",
                    "secret": {
                      "name": "aad-credentials"
                    },
                    "usernameKey": "clientid"
                  },
                  "clientID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "environment": "AzurePublicCloud",
                  "passwordPolicy": "",
                  "path": "azure",
                  "rootPasswordTTL": "128d",
                  "subscriptionID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "tenantID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureSecretEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azuresecretenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azuresecretenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azuresecretenginerole-sample"
                },
                "spec": {
                  "TTL": "",
                  "applicationObjectID": "",
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureGroups": "",
                  "azureRoles": "",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "maxTTL": "",
                  "name": "azure-role",
                  "path": "azure",
                  "permanentlyDelete": "",
                  "persistApp": "",
                  "signInAudience": "",
                  "tags": ""
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "CertAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "certauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "certauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "certauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "test",
                  "path": "cert"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "CertAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "certauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "certauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "certauthenginerole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "certificate": "-----BEGIN CERTIFICATE-----\nMIIDJTCCAg2gAwIBAgIUF119kmCNlnMwK6yOOpfRs4JGhAAwDQYJKoZIhvcNAQEL\nBQAwFjEUMBIGA1UEAwwLZXhhbXBsZS5jb20wHhcNMjUwMTIxMTM0MjExWhcNMjYw\nMTIxMTM0MjExWjAWMRQwEgYDVQQDDAtleGFtcGxlLmNvbTCCASIwDQYJKoZIhvcN\nAQEBBQADggEPADCCAQoCggEBALad4Sx97VyFSfOepJoXUGYljg7hL+z20pVL2/M8\n8HXAQl4qf/i45b//Qz8VogMxdg3Y9J7O/M0n9hxbL18i1gm270VjEFyzjJCyUq0R\n72WMM4Sr3OjQPnVtJoQkWnH7wR3zsWK2N4NuIPdE3Ofb48Olyachffv7xCI+MLIa\n/adcyRggM4Z5f2nBCMcXL8D/eY03qCgd/U1aaPJ0WyJTZHLL4euL0ewKlKtB1JnU\n4MrRa5mbPQ3B+OhxcaD/7BNV6hPXAe2zftEv8LcimhbVoj3XmLJ+om58YIaEMadk\nJj61xnXbf+F/iV41onzwvTPNWCSWiLowDl8P3zkabuS583sCAwEAAaNrMGkwHQYD\nVR0OBBYEFH971CwrqkTvUz2qrUncWMS5ULwKMB8GA1UdIwQYMBaAFH971CwrqkTv\nUz2qrUncWMS5ULwKMA8GA1UdEwEB/wQFMAMBAf8wFgYDVR0RBA8wDYILZXhhbXBs\nZS5jb20wDQYJKoZIhvcNAQELBQADggEBAIQzT113TQC9BqitOCOoDFCJqcM+BHmf\nITEKlEhlqubPvA40lD4wkz67RP2SdQbF1glzZW5sDTxMjpHL+8nhdmo65CJpsQGP\nyMPwnOvx7W5JWQJa1Yt+48JNhgIymDMun+xwDsjNnlBqp0Kr7HkOxPimMPWN4R6B\nydG96puDGBYqxB1ujx5seOGfGvxPBBr3h8kNg054rtIIDrUG2iUTfpNIILJrd8Q2\nshop9//3VcIpUJO7Z1brvK/Hp4DrcKBZVo7vd8n+5uCx3vBA0tjqNAXYtFnSRvDZ\n6XV+rCZecDBFkaLuJ5YgYEK15gGe6lQAez6DTVLXMAKhZ52kceefnwM=\n-----END CERTIFICATE-----\n",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "test-role",
                  "path": "cert/test"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineConfig",
                "metadata": {
                  "name": "my-postgresql-database"
                },
                "spec": {
                  "allowedRoles": [
                    "read-only"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "connectionURL": "postgresql://{{username}}:{{password}}@my-postgresql-database.test-vault-config-operator.svc:5432",
                  "path": "test-vault-config-operator/database",
                  "pluginName": "postgresql-database-plugin",
                  "rootCredentials": {
                    "passwordKey": "postgresql-password",
                    "secret": {
                      "name": "postgresql-admin-password"
                    }
                  },
                  "username": "postgres"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineRole",
                "metadata": {
                  "name": "read-only"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "creationStatements": [
                    "CREATE ROLE \"{{name}}\" WITH LOGIN PASSWORD '{{password}}' VALID UNTIL '{{expiration}}'; GRANT SELECT ON ALL TABLES IN SCHEMA public TO \"{{name}}\";"
                  ],
                  "dBName": "my-postgresql-database",
                  "defaultTTL": "1h",
                  "maxTTL": "24h",
                  "path": "test-vault-config-operator/database"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineStaticRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "databasesecretenginestaticrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "databasesecretenginestaticrole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "databasesecretenginestaticrole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "dBName": "db-name",
                  "path": "test-vault-config-operator/database",
                  "rotationPeriod": "24h",
                  "rotationStatements": [
                    "ALTER ROLE \"{{name}}\" WITH PASSWORD '{{password}}';git add"
                  ],
                  "username": "db-user"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GCPAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "gcpauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gcpauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "gcpauthengineconfig-sample"
                },
                "spec": {
                  "GCEalias": "role_id",
                  "GCEmetadata": "default",
                  "GCPCredentials": {
                    "passwordKey": "credentials",
                    "secret": {
                      "name": "gcp-serviceaccount-credentials"
                    },
                    "usernameKey": "serviceaccount"
                  },
                  "IAMalias": "default",
                  "IAMmetadata": "default",
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "customEndpoint": {},
                  "path": "azure",
                  "serviceAccount": "vault-sa"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GCPAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "gcpauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gcpauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "gcpauthenginerole-sample"
                },
                "spec": {
                  "addGroupAliases": false,
                  "allowGCEInference": false,
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "boundInstanceGroups": [],
                  "boundLabels": [],
                  "boundProjects": {},
                  "boundRegions": [],
                  "boundServiceAccounts": {},
                  "boundZones": [],
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "maxJWTExp": "",
                  "name": "gcp-iam-role",
                  "path": "gcp",
                  "policies": [],
                  "tokenBoundCIDRs": [],
                  "tokenExplicitMaxTTL": "",
                  "tokenMaxTTL": "",
                  "tokenNoDefaultPolicy": false,
                  "tokenNumUses": 0,
                  "tokenPeriod": 0,
                  "tokenPolicies": [],
                  "tokenTTL": "",
                  "tokenType": "",
                  "type": "iam"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitHubSecretEngineConfig",
                "metadata": {
                  "name": "githubsecretengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitHubSecretEngineRole",
                "metadata": {
                  "name": "githubsecretenginerole-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Group",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "group-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "group",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "group-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "metadata": {
                    "team": "team-abc"
                  },
                  "policies": [
                    "team-abc-access"
                  ],
                  "type": "external"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupAlias",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "groupalias-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "groupalias",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "groupalias-sample"
                },
                "spec": {
                  "authEngineMountPath": "kubernetes",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "groupName": "group-sample"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "JWTOIDCAuthEngineConfig",
                "metadata": {
                  "name": "azure-oidc-config"
                },
                "spec": {
                  "OIDCClientID": "000000000-0000-0000-0000-000000000",
                  "OIDCCredentials": {
                    "passwordKey": "oidc_client_secret",
                    "secret": {
                      "name": "oidccredentials"
                    },
                    "usernameKey": "oidc_client_id"
                  },
                  "OIDCDiscoveryURL": "https://login.microsoftonline.com/000000000-0000-0000-0000-000000000/v2.0",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "oidc/azuread-oidc/"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "JWTOIDCAuthEngineRole",
                "metadata": {
                  "name": "azure-oidc-dev-role"
                },
                "spec": {
                  "OIDCScopes": [
                    "https://graph.microsoft.com/.default"
                  ],
                  "allowedRedirectURIs": [
                    "http://localhost:8250/oidc/callback",
                    "http://localhost:8200/ui/vault/auth/oidc/azuread-oidc/oidc/callback"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "groupsClaim": "groups",
                  "name": "dev-role",
                  "path": "oidc-aad/azuread-oidc",
                  "roleType": "oidc",
                  "tokenPolicies": [
                    "dev",
                    "prod"
                  ],
                  "userClaim": "email"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesAuthEngineConfig",
                "metadata": {
                  "name": "kubernetesauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kube-authengine-mount-sample"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesAuthEngineRole",
                "metadata": {
                  "name": "database-engine-admin"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kubernetes",
                  "policies": [
                    "database-engine-admin"
                  ],
                  "targetNamespaces": {
                    "targetNamespaceSelector": {
                      "matchLabels": {
                        "database-engine-admin": "true"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesSecretEngineConfig",
                "metadata": {
                  "name": "kubernetessecretengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesSecretEngineRole",
                "metadata": {
                  "name": "kubernetessecretenginerole-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "LDAPAuthEngineConfig",
                "metadata": {
                  "name": "ldapauthengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "LDAPAuthEngineGroup",
                "metadata": {
                  "name": "ldapauthenginegroup-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin",
                    "serviceAccount": {
                      "name": "default"
                    }
                  },
                  "name": "test-group",
                  "path": "ldap/test"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PKISecretEngineConfig",
                "metadata": {
                  "name": "pkisecretengineconfig-sample"
                },
                "spec": {
                  "TTL": "8760h",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "commonName": "pki-vault-demo.internal.io",
                  "crlDistributionPoints": [
                    "https://vault-internal.vault.svc:8200/v1/test-vault-config-operator/pki/crl\""
                  ],
                  "issuingCertificates": [
                    "https://vault-internal.vault.svc:8200/v1/test-vault-config-operator/pki/ca"
                  ],
                  "path": "test-vault-config-operator/pki",
                  "privateKeyType": "internal",
                  "type": "root"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PKISecretEngineRole",
                "metadata": {
                  "name": "pkisecretenginerole-sample"
                },
                "spec": {
                  "allowGlobDomains": true,
                  "allowSubdomains": true,
                  "allowedDomains": [
                    "internal.io",
                    "pki-vault-demo.svc",
                    "example.com"
                  ],
                  "allowedOtherSans": "*",
                  "allowedURISans": [
                    "*-pki-vault-demo.apps.example.com"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "maxTTL": "8760h",
                  "path": "test-vault-config-operator/pki"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PasswordPolicy",
                "metadata": {
                  "name": "simple-password-policy"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "passwordPolicy": "length = 20\nrule \"charset\" {\n  charset = \"abcdefghijklmnopqrstuvwxyz\"\n}"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Policy",
                "metadata": {
                  "name": "database-engine-admin"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "policy": "# mount database secrets engines\npath \"/sys/mounts/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n  allowed_parameters = {\n    \"type\" = [\"database\"]\n    \"*\"   = []\n  }\n}\n\n# tune database secrets engines\npath \"/sys/mounts/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/tune\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n\n# Configure database secrets engines\npath \"/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/config/+\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n\n# Configure database roles\npath \"/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/roles/+\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineConfig",
                "metadata": {
                  "name": "quaysecretengineconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineRole",
                "metadata": {
                  "name": "quaysecretenginerole-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineStaticRole",
                "metadata": {
                  "name": "quaysecretenginestaticrole-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RabbitMQSecretEngineConfig",
                "metadata": {
                  "name": "rabbitmqsecretengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "rabbitMQ-engine-admin"
                  },
                  "connectionURI": "https://test.com",
                  "leaseMaxTTL": 86400,
                  "leaseTTL": 86400,
                  "path": "vault-config-operator/rabbitmq"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RabbitMQSecretEngineRole",
                "metadata": {
                  "name": "rabbitmqsecretenginerole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "rabbitmq-engine-admin"
                  },
                  "path": "vault-config-operator/rabbitmq",
                  "tags": "administrator",
                  "vhostTopics": [
                    {
                      "topics": [
                        {
                          "permissions": {
                            "configure": ".*",
                            "read": ".*",
                            "write": ".*"
                          },
                          "topicName": "my-topic"
                        },
                        {
                          "permissions": {
                            "read": ".*"
                          },
                          "topicName": "my-read-topic"
                        }
                      ],
                      "vhostName": "/"
                    }
                  ],
                  "vhosts": [
                    {
                      "permissions": {
                        "configure": ".*",
                        "read": ".*",
                        "write": ".*"
                      },
                      "vhostName": "/"
                    },
                    {
                      "permissions": {
                        "read": "my-queue",
                        "write": "my-exchange"
                      },
                      "vhostName": "my-vhost"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RandomSecret",
                "metadata": {
                  "name": "randomsecret-password"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "secret-writer"
                  },
                  "path": "test-vault-config-operator/kv",
                  "secretFormat": {
                    "passwordPolicyName": "simple-password-policy"
                  },
                  "secretKey": "password"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "SecretEngineMount",
                "metadata": {
                  "name": "database"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "path": "test-vault-config-operator",
                  "type": "database"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "VaultSecret",
                "metadata": {
                  "name": "vaultsecret-sample"
                },
                "spec": {
                  "output": {
                    "annotations": {
                      "refresh": "every-minute"
                    },
                    "labels": {
                      "app": "test-vault-config-operator"
                    },
                    "name": "randomsecret",
                    "stringData": {
                      "anotherpassword": "{{ .anotherrandomsecret.password }}",
                      "password": "{{ .randomsecret.password }}"
                    },
                    "type": "Opaque"
                  },
                  "refreshPeriod": "3m0s",
                  "vaultSecretDefinitions": [
                    {
                      "authentication": {
                        "path": "kubernetes",
                        "role": "secret-reader",
                        "serviceAccount": {
                          "name": "default"
                        }
                      },
                      "name": "randomsecret",
                      "path": "test-vault-config-operator/kv/randomsecret-password"
                    },
                    {
                      "authentication": {
                        "path": "kubernetes",
                        "role": "secret-reader",
                        "serviceAccount": {
                          "name": "default"
                        }
                      },
                      "name": "anotherrandomsecret",
                      "path": "test-vault-config-operator/kv/another-password"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/vault-config-operator@sha256:490852417fac37f3003ab7c1ee797f9035fd2905c354cc46eca5e33d2ad780ce
          createdAt: "2025-09-18T18:56:36Z"
          description: This operator provides primitives to declaratively configure
            Hashicorp Vault.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: vault-config-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/vault-config-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthEngineMount is the Schema for the authenginemounts API
            displayName: Auth Engine Mount
            kind: AuthEngineMount
            name: authenginemounts.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureAuthEngineConfig is the Schema for the azureauthengineconfigs
              API
            displayName: Azure Auth Engine Config
            kind: AzureAuthEngineConfig
            name: azureauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureAuthEngineRole is the Schema for the azureauthengineroles
              API
            displayName: Azure Auth Engine Role
            kind: AzureAuthEngineRole
            name: azureauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - kind: AzureSecretEngineConfig
            name: azuresecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureSecretEngineRole is the Schema for the azuresecretengineroles
              API
            displayName: Azure Secret Engine Role
            kind: AzureSecretEngineRole
            name: azuresecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: CertAuthEngineConfig is the Schema for the certauthengineconfigs
              API
            displayName: Cert Auth Engine Config
            kind: CertAuthEngineConfig
            name: certauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: CertAuthEngineRole is the Schema for the certauthengineroles
              API
            displayName: Cert Auth Engine Role
            kind: CertAuthEngineRole
            name: certauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineConfig is the Schema for the databasesecretengineconfigs
              API
            displayName: Database Secret Engine Config
            kind: DatabaseSecretEngineConfig
            name: databasesecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineRole is the Schema for the databasesecretengineroles
              API
            displayName: Database Secret Engine Role
            kind: DatabaseSecretEngineRole
            name: databasesecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineStaticRole is the Schema for the databasesecretenginestaticroles
              API
            displayName: Database Secret Engine Static Role
            kind: DatabaseSecretEngineStaticRole
            name: databasesecretenginestaticroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GCPAuthEngineConfig is the Schema for the gcpauthengineconfigs
              API
            displayName: GCPAuth Engine Config
            kind: GCPAuthEngineConfig
            name: gcpauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: GCPAuthEngineRole is the Schema for the gcpauthengineroles
              API
            displayName: GCPAuth Engine Role
            kind: GCPAuthEngineRole
            name: gcpauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GitHubSecretEngineConfig is the Schema for the githubsecretengineconfigs
              API
            displayName: Git Hub Secret Engine Config
            kind: GitHubSecretEngineConfig
            name: githubsecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: GitHubSecretEngineRole is the Schema for the githubsecretengineroles
              API
            displayName: Git Hub Secret Engine Role
            kind: GitHubSecretEngineRole
            name: githubsecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GroupAlias is the Schema for the groupaliases API
            displayName: Group Alias
            kind: GroupAlias
            name: groupaliases.redhatcop.redhat.io
            version: v1alpha1
          - description: Group is the Schema for the groups API
            displayName: Group
            kind: Group
            name: groups.redhatcop.redhat.io
            version: v1alpha1
          - description: JWTOIDCAuthEngineConfig is the Schema for the jwtoidcauthengineconfigs
              API
            displayName: JWTOIDCAuth Engine Config
            kind: JWTOIDCAuthEngineConfig
            name: jwtoidcauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: JWTOIDCAuthEngineRole is the Schema for the jwtoidcauthengineroles
              API
            displayName: JWTOIDCAuth Engine Role
            kind: JWTOIDCAuthEngineRole
            name: jwtoidcauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesAuthEngineConfig is the Schema for the kubernetesauthengineconfigs
              API
            displayName: Kubernetes Auth Engine Config
            kind: KubernetesAuthEngineConfig
            name: kubernetesauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesAuthEngineRole can be used to define a KubernetesAuthEngineRole
              for the kube-auth authentication method
            displayName: Kubernetes Auth Engine Role
            kind: KubernetesAuthEngineRole
            name: kubernetesauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesSecretEngineConfig is the Schema for the kubernetessecretengineconfigs
              API
            displayName: Kubernetes Secret Engine Config
            kind: KubernetesSecretEngineConfig
            name: kubernetessecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesSecretEngineRole is the Schema for the kubernetessecretengineroles
              API
            displayName: Kubernetes Secret Engine Role
            kind: KubernetesSecretEngineRole
            name: kubernetessecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: LDAPAuthEngineConfig is the Schema for the ldapauthengineconfigs
              API
            displayName: LDAPAuth Engine Config
            kind: LDAPAuthEngineConfig
            name: ldapauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: LDAPAuthEngineGroup is the Schema for the ldapauthenginegroups
              API
            displayName: LDAPAuth Engine Group
            kind: LDAPAuthEngineGroup
            name: ldapauthenginegroups.redhatcop.redhat.io
            version: v1alpha1
          - description: PasswordPolicy is the Schema for the passwordpolicies API
            displayName: Password Policy
            kind: PasswordPolicy
            name: passwordpolicies.redhatcop.redhat.io
            version: v1alpha1
          - description: PKISecretEngineConfig is the Schema for the pkisecretengineconfigs
              API
            displayName: PKISecret Engine Config
            kind: PKISecretEngineConfig
            name: pkisecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: PKISecretEngineRole is the Schema for the pkisecretengineroles
              API
            displayName: PKISecret Engine Role
            kind: PKISecretEngineRole
            name: pkisecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: Policy is the Schema for the policies API
            displayName: Policy
            kind: Policy
            name: policies.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineConfig is the Schema for the quaysecretengineconfigs
              API
            displayName: Quay Secret Engine Config
            kind: QuaySecretEngineConfig
            name: quaysecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineRole is the Schema for the quaysecretengineroles
              API
            displayName: Quay Secret Engine Role
            kind: QuaySecretEngineRole
            name: quaysecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineStaticRole is the Schema for the quaysecretenginestaticroles
              API
            displayName: Quay Secret Engine Static Role
            kind: QuaySecretEngineStaticRole
            name: quaysecretenginestaticroles.redhatcop.redhat.io
            version: v1alpha1
          - description: RabbitMQSecretEngineConfig is the Schema for the rabbitmqsecretengineconfigs
              API
            displayName: Rabbit MQSecret Engine Config
            kind: RabbitMQSecretEngineConfig
            name: rabbitmqsecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: RabbitMQSecretEngineRole is the Schema for the rabbitmqsecretengineroles
              API
            displayName: Rabbit MQSecret Engine Role
            kind: RabbitMQSecretEngineRole
            name: rabbitmqsecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: RandomSecret is the Schema for the randomsecrets API
            displayName: Random Secret
            kind: RandomSecret
            name: randomsecrets.redhatcop.redhat.io
            version: v1alpha1
          - description: SecretEngineMount is the Schema for the secretenginemounts
              API
            displayName: Secret Engine Mount
            kind: SecretEngineMount
            name: secretenginemounts.redhatcop.redhat.io
            version: v1alpha1
          - description: VaultSecret is the Schema for the vaultsecrets API
            displayName: Vault Secret
            kind: VaultSecret
            name: vaultsecrets.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator helps set up Vault Configurations. The main intent is to do so such that subsequently pods can consume the secrets made available.
          There are two main principles through all of the capabilities of this operator:

          1. high-fidelity API. The CRD exposed by this operator reflect field by field the Vault APIs. This is because we don't want to make any assumption on the kinds of configuration workflow that user will set up. That being said the Vault API is very extensive and we are starting with enough API coverage to support, we think, some simple and very common configuration workflows.
          2. attention to security (after all we are integrating with a security tool). To prevent credential leaks we give no permissions to the operator itself against Vault. All APIs exposed by this operator contains enough information to authenticate to Vault using a local service account (local to the namespace where the API exist). In other word for a namespace user to be abel to successfully configure Vault, a service account in that namespace must have been previously given the needed Vault permissions.

          Currently this operator supports the following CRDs:

          1. [Policy](https://github.com/redhat-cop/vault-config-operator#policy) Configures Vault [Policies](https://www.vaultproject.io/docs/concepts/policies)
          2. [KubernetesAuthEngineRole](https://github.com/redhat-cop/vault-config-operator#KubernetesAuthEngineRole) Configures a Vault [Kubernetes Authentication](https://www.vaultproject.io/docs/auth/kubernetes) Role
          3. [SecretEngineMount](https://github.com/redhat-cop/vault-config-operator#SecretEngineMount) Configures a Mount point for a [SecretEngine](https://www.vaultproject.io/docs/secrets)
          4. [DatabaseSecretEngineConfig](https://github.com/redhat-cop/vault-config-operator#DatabaseSecretEngineConfig) Configures a [Database Secret Engine](https://www.vaultproject.io/docs/secrets/databases) Connection
          5. [DatabaseSecretEngineRole](https://github.com/redhat-cop/vault-config-operator#DatabaseSecretEngineRole) Configures a [Database Secret Engine](https://www.vaultproject.io/docs/secrets/databases) Role
          6. [RandomSecret](https://github.com/redhat-cop/vault-config-operator#RandomSecret) Creates a random secret in a vault [kv Secret Engine](https://www.vaultproject.io/docs/secrets/kv) with one password field generated using a [PasswordPolicy](https://www.vaultproject.io/docs/concepts/password-policies)vault-config-operator
        displayName: Vault Config Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - secrets
        - gitops
        - policy
        - credentials
        links:
        - name: Vault Config Operator
          url: https://github.com/redhat-cop/vault-config-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/vault-config-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/vault-config-operator@sha256:490852417fac37f3003ab7c1ee797f9035fd2905c354cc46eca5e33d2ad780ce
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 0.8.34
      entries:
      - name: vault-config-operator.v0.8.34
        version: 0.8.34
      - name: vault-config-operator.v0.8.33
        version: 0.8.33
      - name: vault-config-operator.v0.8.32
        version: 0.8.32
      - name: vault-config-operator.v0.8.31
        version: 0.8.31
      - name: vault-config-operator.v0.8.30
        version: 0.8.30
      - name: vault-config-operator.v0.8.29
        version: 0.8.29
      - name: vault-config-operator.v0.8.28
        version: 0.8.28
      - name: vault-config-operator.v0.8.27
        version: 0.8.27
      - name: vault-config-operator.v0.8.25
        version: 0.8.25
      - name: vault-config-operator.v0.8.24
        version: 0.8.24
      - name: vault-config-operator.v0.8.23
        version: 0.8.23
      - name: vault-config-operator.v0.8.22
        version: 0.8.22
      - name: vault-config-operator.v0.8.21
        version: 0.8.21
      - name: vault-config-operator.v0.8.20
        version: 0.8.20
      - name: vault-config-operator.v0.8.19
        version: 0.8.19
      - name: vault-config-operator.v0.8.18
        version: 0.8.18
      - name: vault-config-operator.v0.8.17
        version: 0.8.17
      - name: vault-config-operator.v0.8.16
        version: 0.8.16
      - name: vault-config-operator.v0.8.15
        version: 0.8.15
      - name: vault-config-operator.v0.8.14
        version: 0.8.14
      - name: vault-config-operator.v0.8.13
        version: 0.8.13
      - name: vault-config-operator.v0.8.12
        version: 0.8.12
      - name: vault-config-operator.v0.8.11
        version: 0.8.11
      - name: vault-config-operator.v0.8.10
        version: 0.8.10
      - name: vault-config-operator.v0.8.9
        version: 0.8.9
      - name: vault-config-operator.v0.8.7
        version: 0.8.7
      - name: vault-config-operator.v0.8.5
        version: 0.8.5
      - name: vault-config-operator.v0.8.4
        version: 0.8.4
      - name: vault-config-operator.v0.8.3
        version: 0.8.3
      - name: vault-config-operator.v0.8.2
        version: 0.8.2
      - name: vault-config-operator.v0.8.0
        version: 0.8.0
      - name: vault-config-operator.v0.7.1
        version: 0.7.1
      - name: vault-config-operator.v0.6.5
        version: 0.6.5
      - name: vault-config-operator.v0.6.4
        version: 0.6.4
      - name: vault-config-operator.v0.6.3
        version: 0.6.3
      - name: vault-config-operator.v0.6.2
        version: 0.6.2
      - name: vault-config-operator.v0.6.1
        version: 0.6.1
      - name: vault-config-operator.v0.6.0
        version: 0.6.0
      - name: vault-config-operator.v0.5.1
        version: 0.5.1
      - name: vault-config-operator.v0.5.0
        version: 0.5.0
      - name: vault-config-operator.v0.4.1
        version: 0.4.1
      - name: vault-config-operator.v0.4.0
        version: 0.4.0
      - name: vault-config-operator.v0.3.1
        version: 0.3.1
      - name: vault-config-operator.v0.3.0
        version: 0.3.0
      - name: vault-config-operator.v0.2.0
        version: 0.2.0
      - name: vault-config-operator.v0.1.2
        version: 0.1.2
      - name: vault-config-operator.v0.1.1
        version: 0.1.1
      - name: vault-config-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: vault-config-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-rest-api-browser-v1
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-rest-api-browser-v1.v0.1.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rest.loadcore.keysight.com/v1alpha1",
                "kind": "Apibrowser",
                "metadata": {
                  "name": "apibrowser-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-06-29T14:08:05Z"
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Apibrowser
            name: apibrowsers.rest.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Rest API Browser Operator v1
        displayName: Keysight Rest API Browser Operator v1
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Rest Api Browser V1
          url: https://keysight-rest-api-browser-v1.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5d5061dbc2ec5082cf14b6c600fb5400b83cf91d7ccebfa80680a238d275db
        - prod.harbor.keysight.digital/loadcore/keysight-rest-api-browser-v1@sha256:3ada6a0fa630acb0fa962a11d2c272ff679dded1106c4693d2c8f5ec9ab414a4
        version: 0.1.11
      entries:
      - name: keysight-rest-api-browser-v1.v0.1.11
        version: 0.1.11
      name: alpha
    defaultChannel: alpha
    packageName: keysight-rest-api-browser-v1
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Qualcomm Technologies, Inc.
      provider-url: https://github.com/quic/aic-operator
    name: aic-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aic-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aic.quicinc.com/v1",
                "kind": "AIC",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "aic-operator",
                    "app.kubernetes.io/instance": "aic-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "aic",
                    "app.kubernetes.io/part-of": "aic-operator"
                  },
                  "name": "aic-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "false"
          containerImage: ghcr.io/quic/cloud_ai_openshift_operator@sha256:9cf581e9b5c9cd08819fab8f8b6313836a0a20098ad4c99b80d93dbcf5ab32bf
          description: Automate the management of OpenShift Operator for Qualcomm
            AI 100 accelerators.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: Qualcomm Technologies, Inc.
          repository: https://github.com/quic/aic-operator
          support: Qualcomm Technologies, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AIC is the Schema for the aics API
            displayName: AIC
            kind: AIC
            name: aics.aic.quicinc.com
            version: v1
        description: The Openshift Operator enables the Qualcomm® Cloud AI series
          of hardware on OpenShift clusters by automating the configuration and installation
          of their Linux device drivers and its Kubernetes Device Plugin.
        displayName: AIC Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AIC
        - AIC-Operator
        links:
        - name: AIC Operator
          url: https://github.com/quic/aic-operator
        maintainers:
        - email: quic_vmellemp@quicinc.com
          name: quic_vmellemp
        - email: quic_carls@quicinc.com
          name: quic_carls
        maturity: alpha
        provider:
          name: Qualcomm Technologies, Inc.
          url: https://github.com/quic/aic-operator
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:7a943db9bcb22f817d75341621472f4bb30abc5625c53b9a781d97d431bc4d09
        - ghcr.io/quic/cloud_ai_openshift_operator@sha256:9cf581e9b5c9cd08819fab8f8b6313836a0a20098ad4c99b80d93dbcf5ab32bf
        version: 0.3.1
      entries:
      - name: aic-operator.v0.3.1
        version: 0.3.1
      name: alpha
    defaultChannel: alpha
    packageName: aic-operator
    provider:
      name: Qualcomm Technologies, Inc.
      url: https://github.com/quic/aic-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kong Inc.
      provider-url: https://konghq.com
    name: kong-gateway-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kong-gateway-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "ControlPlane",
                "metadata": {
                  "name": "controlplane-example"
                },
                "spec": {
                  "dataplane": "dataplane-example",
                  "gatewayClass": "kong"
                }
              },
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "DataPlane",
                "metadata": {
                  "name": "dataplane-example"
                },
                "spec": {
                  "env": [
                    {
                      "name": "KONG_LOG_LEVEL",
                      "value": "debug"
                    }
                  ]
                }
              },
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "GatewayConfiguration",
                "metadata": {
                  "name": "kong",
                  "namespace": "default"
                },
                "spec": {
                  "controlPlaneDeploymentOptions": {
                    "env": [
                      {
                        "name": "CONTROLLER_LOG_LEVEL",
                        "value": "debug"
                      }
                    ]
                  },
                  "dataPlaneDeploymentOptions": {
                    "containerImage": "kong",
                    "version": "2.8"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking,Integration & Delivery
          operatorframework.io/suggested-namespace: kong-system
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ControlPlane is the Schema for the controlplanes API
            displayName: Control Plane
            kind: ControlPlane
            name: controlplanes.gateway-operator.konghq.com
            version: v1alpha1
          - description: DataPlane is the Schema for the dataplanes API
            displayName: Data Plane
            kind: DataPlane
            name: dataplanes.gateway-operator.konghq.com
            version: v1alpha1
          - description: GatewayConfiguration is the Schema for the gatewayconfigurations
              API
            displayName: Gateway Configuration
            kind: GatewayConfiguration
            name: gatewayconfigurations.gateway-operator.konghq.com
            version: v1alpha1
        description: Kong Operator for Kubernetes. Technical preview.
        displayName: Kong
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - gateway
        - ingress
        - kong
        - kubernetes
        - go
        links:
        - name: Github repository for the operator
          url: https://github.com/kong/gateway-operator
        maintainers:
        - email: support@konghq.com
          name: Kong Inc.
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Kong Inc.
          url: https://konghq.com
        relatedImages:
        - ghcr.io/kong/gateway-operator@sha256:76b227715a3e1915a28f8aaa1e2ff8495349d29d0d61e7b85a72d74b03fbcc7f
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.connect.redhat.com/kong/kong@sha256:54e7c7e57bf93810a7cc36327023b8699f754c4c40c5c289548d54170caa5db2
        - registry.connect.redhat.com/kong/kong-ingress-controller@sha256:f691db070e7c604da9e163800903d0942ec6fda879eb9dcd5d8e90618ae787da
        version: 0.0.2
      entries:
      - name: kong-gateway-operator.v0.0.2
        version: 0.0.2
      - name: kong-gateway-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: kong-gateway-operator
    provider:
      name: Kong Inc.
      url: https://konghq.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: https://www.turbonomic.com/
    name: t8c-tsc-client-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-tsc-client-certified.v1.4.2-candidate-20250806145822000
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "TurbonomicClient",
                "metadata": {
                  "name": "turbonomicclient-release"
                },
                "spec": {
                  "global": {
                    "version": "8.16.1"
                  },
                  "probes": {
                    "actionScript": {
                      "enabled": false
                    },
                    "actionStreamKafka": {
                      "enabled": false
                    },
                    "appDynamics": {
                      "enabled": false
                    },
                    "compellent": {
                      "enabled": false
                    },
                    "datadog": {
                      "enabled": false
                    },
                    "dynatrace": {
                      "enabled": false
                    },
                    "flexera": {
                      "enabled": false
                    },
                    "hds": {
                      "enabled": false
                    },
                    "hpe3par": {
                      "enabled": false
                    },
                    "hyperFlex": {
                      "enabled": false
                    },
                    "hyperV": {
                      "enabled": false
                    },
                    "ibmStorageFlashSystem": {
                      "enabled": false
                    },
                    "instana": {
                      "enabled": false
                    },
                    "jvm": {
                      "enabled": false
                    },
                    "mssql": {
                      "enabled": false
                    },
                    "mysql": {
                      "enabled": false
                    },
                    "netApp": {
                      "enabled": false
                    },
                    "newRelic": {
                      "enabled": false
                    },
                    "nutanix": {
                      "enabled": false
                    },
                    "oneView": {
                      "enabled": false
                    },
                    "oracle": {
                      "enabled": false
                    },
                    "powerVM": {
                      "enabled": false
                    },
                    "pure": {
                      "enabled": false
                    },
                    "scaleIO": {
                      "enabled": false
                    },
                    "serviceNow": {
                      "enabled": false
                    },
                    "snmp": {
                      "enabled": false
                    },
                    "terraform": {
                      "enabled": false
                    },
                    "tomcat": {
                      "enabled": false
                    },
                    "ucs": {
                      "enabled": false
                    },
                    "vCenter": {
                      "enabled": false
                    },
                    "vmax": {
                      "enabled": false
                    },
                    "vmm": {
                      "enabled": false
                    },
                    "vplex": {
                      "enabled": false
                    },
                    "webLogic": {
                      "enabled": false
                    },
                    "webSphere": {
                      "enabled": false
                    },
                    "wmi": {
                      "enabled": false
                    },
                    "xtremIO": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "VersionManager",
                "metadata": {
                  "name": "versionmanager-release"
                },
                "spec": {
                  "url": "http://remote-nginx-tunnel:9080/cluster-manager/clusterConfiguration"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-client-operator@sha256:8db7396c956e7a04aee92e1e90491cd3eb533cebe5f5616ca43ff16d9f3ed6e2
          createdAt: "2025-08-07T07:15:28Z"
          description: Turbonomic Secure Connect for Hybrid environment management
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.4.2-candidate-20250806145822000'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TurbonomicClient is the Schema for the turbonomicclients
              API
            displayName: Turbonomic Client
            kind: TurbonomicClient
            name: turbonomicclients.clients.turbonomic.ibm.com
            version: v1alpha1
          - description: VersionManager is the Schema for the versionmanagers API
            displayName: Version Manager
            kind: VersionManager
            name: versionmanagers.clients.turbonomic.ibm.com
            version: v1alpha1
        description: "### Turbonomic Secure Connect for Hybrid environment management\nTurbonomic
          Workload Automation for Multicloud simultaneously optimizes performance,
          compliance, and cost in real-time. Workloads are precisely resourced, automatically,
          to perform while satisfying business constraints:\n* Continuous placement
          of workload across multiple clouds both on-prem and public clouds providers.\n*
          Continuous scaling for applications and the underlying infrastructure.\n*
          Capability to connect both public cloud and on-prem targets on single saas
          instance\nIt assures application performance by giving workloads the resources
          they need when they need them.\n\n### How do I start?\nTurbonomic documentation
          can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm.\nTo
          setup Turbonomic Secure Connect client, follow the instructions below:\n1.
          Click Install on the landing page\n2. Upon moving to Install Operator page,
          do not modify the default selection for any of the configuration\n3. Select
          the Installed namespace from the drop down or add a new namespace and click
          on create project to create the entered namespace.\n4. Click Install on
          the Install Operator page.\nAt this point operator is installed.\n\nOn the
          Details tab, under Provided APIs: \n1. Click Create instance for the Turbonomic
          Client API. Enable/Disable the probes based on the requirement, and click
          Create.\n2. Click Create instance for the Version Manager API. Do not modify
          the JSON/YAML file, click on Create.\n\n### How does it work?\nTurbonomic
          uses a public APIs already exposed by application and infrastructure instrumentation
          to discover and monitor your environment.\nTurbonomic determines the right
          actions that drive continuous health, including continuous placement and
          continuous scaling for applications and the underlying cluster.\nTurbonomic
          leverages the built-on orchestration provided by the application and infrastructure
          deployment tools and automates the execution of these actions to continiously
          meet the respective service level objective of each application service."
        displayName: Turbonomic Secure Connect Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - turbonomic
        - turbo secure connect
        links:
        - name: T8c Tsc Client Certified
          url: https://t8c-tsc-certified.domain
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        relatedImages:
        - icr.io/cpopen/t8c-client-operator@sha256:8db7396c956e7a04aee92e1e90491cd3eb533cebe5f5616ca43ff16d9f3ed6e2
        version: 1.4.2-candidate-20250806145822000
      entries:
      - name: t8c-tsc-client-certified.v1.4.2-candidate-20250806145822000
        version: 1.4.2-candidate-20250806145822000
      name: candidate
    - currentCSV: t8c-tsc-client-certified.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "TurbonomicClient",
                "metadata": {
                  "name": "turbonomicclient-release"
                },
                "spec": {
                  "global": {
                    "version": "8.16.1"
                  },
                  "probes": {
                    "actionScript": {
                      "enabled": false
                    },
                    "actionStreamKafka": {
                      "enabled": false
                    },
                    "appDynamics": {
                      "enabled": false
                    },
                    "compellent": {
                      "enabled": false
                    },
                    "datadog": {
                      "enabled": false
                    },
                    "dynatrace": {
                      "enabled": false
                    },
                    "flexera": {
                      "enabled": false
                    },
                    "hds": {
                      "enabled": false
                    },
                    "hpe3par": {
                      "enabled": false
                    },
                    "hyperFlex": {
                      "enabled": false
                    },
                    "hyperV": {
                      "enabled": false
                    },
                    "ibmStorageFlashSystem": {
                      "enabled": false
                    },
                    "instana": {
                      "enabled": false
                    },
                    "jvm": {
                      "enabled": false
                    },
                    "mssql": {
                      "enabled": false
                    },
                    "mysql": {
                      "enabled": false
                    },
                    "netApp": {
                      "enabled": false
                    },
                    "newRelic": {
                      "enabled": false
                    },
                    "nutanix": {
                      "enabled": false
                    },
                    "oneView": {
                      "enabled": false
                    },
                    "oracle": {
                      "enabled": false
                    },
                    "powerVM": {
                      "enabled": false
                    },
                    "pure": {
                      "enabled": false
                    },
                    "scaleIO": {
                      "enabled": false
                    },
                    "serviceNow": {
                      "enabled": false
                    },
                    "snmp": {
                      "enabled": false
                    },
                    "terraform": {
                      "enabled": false
                    },
                    "tomcat": {
                      "enabled": false
                    },
                    "ucs": {
                      "enabled": false
                    },
                    "vCenter": {
                      "enabled": false
                    },
                    "vmax": {
                      "enabled": false
                    },
                    "vmm": {
                      "enabled": false
                    },
                    "vplex": {
                      "enabled": false
                    },
                    "webLogic": {
                      "enabled": false
                    },
                    "webSphere": {
                      "enabled": false
                    },
                    "wmi": {
                      "enabled": false
                    },
                    "xtremIO": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "VersionManager",
                "metadata": {
                  "name": "versionmanager-release"
                },
                "spec": {
                  "url": "http://remote-nginx-tunnel:9080/cluster-manager/clusterConfiguration"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-client-operator@sha256:626e024327388410f7a77bd9b8e5a5df29993fcafdb0176cd0bea08b05713c61
          createdAt: "2025-08-12T23:15:08Z"
          description: Turbonomic Secure Connect for Hybrid environment management
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.4.2'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TurbonomicClient is the Schema for the turbonomicclients
              API
            displayName: Turbonomic Client
            kind: TurbonomicClient
            name: turbonomicclients.clients.turbonomic.ibm.com
            version: v1alpha1
          - description: VersionManager is the Schema for the versionmanagers API
            displayName: Version Manager
            kind: VersionManager
            name: versionmanagers.clients.turbonomic.ibm.com
            version: v1alpha1
        description: "### Turbonomic Secure Connect for Hybrid environment management\nTurbonomic
          Workload Automation for Multicloud simultaneously optimizes performance,
          compliance, and cost in real-time. Workloads are precisely resourced, automatically,
          to perform while satisfying business constraints:\n* Continuous placement
          of workload across multiple clouds both on-prem and public clouds providers.\n*
          Continuous scaling for applications and the underlying infrastructure.\n*
          Capability to connect both public cloud and on-prem targets on single saas
          instance\nIt assures application performance by giving workloads the resources
          they need when they need them.\n\n### How do I start?\nTurbonomic documentation
          can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm.\nTo
          setup Turbonomic Secure Connect client, follow the instructions below:\n1.
          Click Install on the landing page\n2. Upon moving to Install Operator page,
          do not modify the default selection for any of the configuration\n3. Select
          the Installed namespace from the drop down or add a new namespace and click
          on create project to create the entered namespace.\n4. Click Install on
          the Install Operator page.\nAt this point operator is installed.\n\nOn the
          Details tab, under Provided APIs: \n1. Click Create instance for the Turbonomic
          Client API. Enable/Disable the probes based on the requirement, and click
          Create.\n2. Click Create instance for the Version Manager API. Do not modify
          the JSON/YAML file, click on Create.\n\n### How does it work?\nTurbonomic
          uses a public APIs already exposed by application and infrastructure instrumentation
          to discover and monitor your environment.\nTurbonomic determines the right
          actions that drive continuous health, including continuous placement and
          continuous scaling for applications and the underlying cluster.\nTurbonomic
          leverages the built-on orchestration provided by the application and infrastructure
          deployment tools and automates the execution of these actions to continiously
          meet the respective service level objective of each application service."
        displayName: Turbonomic Secure Connect Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - turbonomic
        - turbo secure connect
        links:
        - name: T8c Tsc Client Certified
          url: https://t8c-tsc-certified.domain
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        relatedImages:
        - icr.io/cpopen/t8c-client-operator@sha256:626e024327388410f7a77bd9b8e5a5df29993fcafdb0176cd0bea08b05713c61
        version: 1.4.2
      entries:
      - name: t8c-tsc-client-certified.v1.4.2
        version: 1.4.2
      name: stable
    defaultChannel: stable
    packageName: t8c-tsc-client-certified
    provider:
      name: Turbonomic, Inc.
      url: https://www.turbonomic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Aerospike
      provider-url: ""
    name: aerospike-kubernetes-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aerospike-kubernetes-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "memory-size": 3000000000,
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:6.3.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: quay.io/aerospike/aerospike-kubernetes-operator@sha256:a3033b50ae434b97de4de1ca03013dabcfac580d810a6ee8c0fe05b6a9301952
          createdAt: 2023-05-04T09:32:47UTC
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: <=2.4.0
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Aerospike
        relatedImages:
        - quay.io/aerospike/aerospike-kubernetes-operator@sha256:a3033b50ae434b97de4de1ca03013dabcfac580d810a6ee8c0fe05b6a9301952
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        version: 2.5.0
      entries:
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      name: alpha
    - currentCSV: aerospike-kubernetes-operator-rhmp.v4.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "data-size": 1073741824,
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:8.1.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackup",
                "metadata": {
                  "name": "aerospikebackup",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "aerospike-cluster": {
                      "aerospike-aerospikebackup-test-cluster": {
                        "credentials": {
                          "password": "admin123",
                          "user": "admin"
                        },
                        "seed-nodes": [
                          {
                            "host-name": "aerocluster.aerospike.svc.cluster.local",
                            "port": 3000
                          }
                        ]
                      }
                    },
                    "backup-routines": {
                      "aerospike-aerospikebackup-test-routine": {
                        "backup-policy": "test-policy",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "local"
                      },
                      "aerospike-aerospikebackup-test-routine1": {
                        "backup-policy": "test-policy1",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "s3Storage"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackupService",
                "metadata": {
                  "name": "aerospikebackupservice-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "config": {
                    "backup-policies": {
                      "test-policy": {
                        "parallel": 3
                      },
                      "test-policy1": {
                        "parallel": 3
                      }
                    },
                    "service": {
                      "http": {
                        "port": 8081
                      }
                    },
                    "storage": {
                      "local": {
                        "local-storage": {
                          "path": "/tmp/localStorage"
                        }
                      },
                      "s3Storage": {
                        "s3-storage": {
                          "bucket": "aerospike-kubernetes-operator-test",
                          "s3-profile": "default",
                          "s3-region": "us-east-1"
                        }
                      }
                    }
                  },
                  "image": "aerospike/aerospike-backup-service:3.2.0",
                  "secrets": [
                    {
                      "secretName": "aws-secret",
                      "volumeMount": {
                        "mountPath": "/root/.aws/credentials",
                        "name": "aws-secret",
                        "subPath": "credentials"
                      }
                    }
                  ],
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeRestore",
                "metadata": {
                  "name": "aerospikerestore-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "backup-data-path": "aerospike-aerospikebackup-test-routine/backup/1733506015827/data/test",
                    "destination": {
                      "credentials": {
                        "password": "admin123",
                        "user": "admin"
                      },
                      "label": "destinationCluster",
                      "seed-nodes": [
                        {
                          "host-name": "aerocluster.test.svc.cluster.local",
                          "port": 3000
                        }
                      ]
                    },
                    "policy": {
                      "no-generation": true,
                      "no-indexes": true,
                      "parallel": 3
                    },
                    "source": {
                      "local-storage": {
                        "path": "/tmp/localStorage"
                      }
                    }
                  },
                  "type": "Full"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/aerospike/aerospike-kubernetes-operator@sha256:e6d6bd7c58e036b3728e52475a9b764ca8c46b83cbc4751cac66294495bf2b31
          createdAt: "2025-09-15T10:14:05Z"
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
          support: Aerospike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeBackup is the Schema for the aerospikebackup API
            displayName: Aerospike Backup
            kind: AerospikeBackup
            name: aerospikebackups.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeBackupService is the Schema for the aerospikebackupservices
              API
            displayName: Aerospike Backup Service
            kind: AerospikeBackupService
            name: aerospikebackupservices.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1
          - description: AerospikeRestore is the Schema for the aerospikerestores
              API
            displayName: Aerospike Restore
            kind: AerospikeRestore
            name: aerospikerestores.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
           -   Aerospike cluster monitoring
           -   Backup and restore Aerospike clusters
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Aerospike
        relatedImages:
        - quay.io/aerospike/aerospike-kubernetes-operator@sha256:e6d6bd7c58e036b3728e52475a9b764ca8c46b83cbc4751cac66294495bf2b31
        version: 4.1.1
      entries:
      - name: aerospike-kubernetes-operator-rhmp.v4.1.1
        version: 4.1.1
      - name: aerospike-kubernetes-operator-rhmp.v4.1.0
        version: 4.1.0
      - name: aerospike-kubernetes-operator-rhmp.v4.0.2
        version: 4.0.2
      - name: aerospike-kubernetes-operator-rhmp.v4.0.1
        version: 4.0.1
      - name: aerospike-kubernetes-operator-rhmp.v4.0.0
        version: 4.0.0
      - name: aerospike-kubernetes-operator-rhmp.v3.4.1
        version: 3.4.1
      - name: aerospike-kubernetes-operator-rhmp.v3.4.0
        version: 3.4.0
      - name: aerospike-kubernetes-operator-rhmp.v3.3.1
        version: 3.3.1
      - name: aerospike-kubernetes-operator-rhmp.v3.3.0
        version: 3.3.0
      - name: aerospike-kubernetes-operator-rhmp.v3.2.2
        version: 3.2.2
      - name: aerospike-kubernetes-operator-rhmp.v3.2.1
        version: 3.2.1
      - name: aerospike-kubernetes-operator.v3.2.0
        version: 3.2.0
      - name: aerospike-kubernetes-operator.v3.1.0
        version: 3.1.0
      - name: aerospike-kubernetes-operator.v3.0.0
        version: 3.0.0
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      name: stable
    defaultChannel: stable
    packageName: aerospike-kubernetes-operator-rhmp
    provider:
      name: Aerospike
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: samsung
      provider-url: ""
    name: socmmd
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: socmmd.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cmmd.samsung.com/v1",
                "kind": "CMMD",
                "metadata": {
                  "name": "my-app",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "allocateMode": "auto",
                  "enable": true,
                  "payload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "metadata": {
                      "name": "my-app"
                    },
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "name": "my-app"
                        }
                      },
                      "template": {
                        "metadata": {
                          "labels": {
                            "name": "my-app"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "quay.io/my-account/my-app:latest",
                              "name": "my-app-01"
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "cmmd.samsung.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "machineconfiguration.openshift.io/role": "cmmd"
                  },
                  "name": "cmmd-node-label",
                  "namespace": "socmmd"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          console.openshift.io/plugins: '["socmmd-console-plugin"]'
          containerImage: quay.io/samsung_sag/socmmd@sha256:2ef361fb15f0fec99e0a8f1ef83ad4dd14f701d579c900cce084c4a62cf805f8
          createdAt: "2024-08-01T05:45:38Z"
          description: Samsung Operator For CMM-D
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "cmmd.samsung.com/v1",
              "kind": "Label",
              "metadata": {
                "name": "cmmd-node-label",
                "namespace": "socmmd",
                "labels": {}
              },
              "spec": {}
            }
          operatorframework.io/suggested-namespace: socmmd-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenMPDK/SMDK/tree/main/lib/redhat/socmmd
          support: samsung
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CMMD is the Schema for the cmmds API.
            displayName: CMMD
            kind: CMMD
            name: cmmds.cmmd.samsung.com
            version: v1
          - description: Label is the Schema for the labels API.
            displayName: Node Label
            kind: Label
            name: labels.cmmd.samsung.com
            version: v1
        description: Samsung Operator For CMM-D
        displayName: socmmd
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - samsung
        - memory
        - extend
        - cmm-d
        links:
        - name: Samsung Operator For CMM-D
          url: https://github.com/OpenMPDK/SMDK/tree/main/lib/redhat/socmmd
        maintainers:
        - email: junhyeok.im@samsung.com
          name: junhyeok-im
        maturity: stable
        provider:
          name: samsung
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/samsung_sag/socmmd@sha256:2ef361fb15f0fec99e0a8f1ef83ad4dd14f701d579c900cce084c4a62cf805f8
        version: 0.0.2
      entries:
      - name: socmmd.v0.0.2
        version: 0.0.2
      - name: socmmd.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: socmmd
    provider:
      name: samsung
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicast-community-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicast-community-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/apicast-operator:v0.8.0
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway built on top of NGINX. It is part
            of the Red Hat 3scale API Management Platform
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |
          The APIcast Operator creates and maintains the Red Hat 3scale API Gateway in several deployment configurations.

          APIcast is an API gateway built on top of [NGINX](https://www.nginx.com/). It is part of the [Red Hat 3scale API Management Platform](https://www.redhat.com/en/technologies/jboss-middleware/3scale).

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator/blob/v0.8.0/doc/operator-user-guide.md#upgrading-APIcast) for more information.

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/apicast-operator/tree/v0.8.0).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: APIcast
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://github.com/3scale/apicast-operator/tree/v0.8.0
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/apicast-operator:v0.8.0
        version: 0.8.0
      entries:
      - name: apicast-community-operator.v0.8.0
        version: 0.8.0
      - name: apicast-community-operator.v0.7.1
        version: 0.7.1
      - name: apicast-community-operator.v0.6.0
        version: 0.6.0
      - name: apicast-community-operator.v0.5.1
        version: 0.5.1
      name: alpha
    - currentCSV: apicast-community-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/apicast-operator:v0.8.0
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway built on top of NGINX. It is part
            of the Red Hat 3scale API Management Platform
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |
          The APIcast Operator creates and maintains the Red Hat 3scale API Gateway in several deployment configurations.

          APIcast is an API gateway built on top of [NGINX](https://www.nginx.com/). It is part of the [Red Hat 3scale API Management Platform](https://www.redhat.com/en/technologies/jboss-middleware/3scale).

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator/blob/v0.8.0/doc/operator-user-guide.md#upgrading-APIcast) for more information.

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/apicast-operator/tree/v0.8.0).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: APIcast
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://github.com/3scale/apicast-operator/tree/v0.8.0
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/apicast-operator:v0.8.0
        version: 0.8.0
      entries:
      - name: apicast-community-operator.v0.8.0
        version: 0.8.0
      - name: apicast-community-operator.v0.7.1
        version: 0.7.1
      - name: apicast-community-operator.v0.6.0
        version: 0.6.0
      - name: apicast-community-operator.v0.5.1
        version: 0.5.1
      name: stable
    defaultChannel: stable
    packageName: apicast-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-appsec
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-appsec.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Appsec",
                "metadata": {
                  "name": "appsec-sample"
                },
                "spec": {
                  "appsecDataModel": {
                    "port": 8080
                  },
                  "appsecResourceService": {
                    "port": 8080
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "remoteRepository": "docker.io",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "reportTemplates": {
                    "enabled": false
                  },
                  "systemMonitor": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:39:34Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Appsec
            name: appsecs.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP APPSEC Operator
        displayName: Keysight WAP APPSEC Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Appsec
          url: https://keysight-wap-appsec.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-appsec@sha256:e39ff225c26b750f86553693dda47d0dcc9834a4ab0975aea154361ee3527314
        version: 0.1.21
      entries:
      - name: keysight-wap-appsec.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-appsec
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-operator.v23.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerify",
                "metadata": {
                  "name": "ibmsecurityverify-sample"
                },
                "spec": {
                  "clientSecret": "--secret--",
                  "sessionLifetime": 3600,
                  "ssoPath": "/verify-sso"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isv-saas/verify-operator@sha256:c3bed2e5db48e54884d2278d24ca4864f389a56a667db481c627d01ab5ffaba7
          createdAt: "2023-03-22T02:49:11Z"
          description: The IBM Security Verify operator can consistently enforce policy-driven
            security, including authentication and authorization, using the Ingress
            networking capability of OpenShift.
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.9"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM-Security/verify-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerify is the Schema for the ibmsecurityverifies
              API.
            displayName: IBMSecurity Verify
            kind: IBMSecurityVerify
            name: ibmsecurityverifies.ibm.com
            version: v1
        description: "[IBM Security Verify](https://www.ibm.com/products/verify-saas)
          is an Identity-as-a-Service platform that allows IT, security and business
          leaders to protect their digital users, assets and data in a hybrid, multi-cloud
          world by enabling technical agility and operational efficiency. IBM Security
          Verify SaaS provides single sign-on (SSO), multi-factor authentication (MFA),
          AI-powered context for risk-based authentication for adaptive access decisions,
          user management, access recertification campaigns and identity analytics.\n\nAs
          part of using Red Hat OpenShift, you are entitled to Verify SaaS SSO for
          unlimited apps and users.  If you do not already have a Verify SaaS SSO
          tenant a new one can be [created](https://www.ibm.com/account/reg/signup?formid=urx-51255).\n\nFor
          a detailed description of IBM Security Verify refer to the [Official documentation](https://www.ibm.com/docs/en/security-verify).\n\nThe
          IBM Security Verify operator can consistently enforce policy-driven security
          by using the Ingress networking capability of OpenShift in conjunction with
          the [Nginx Ingress operator](https://www.nginx.com/blog/getting-started-nginx-ingress-operator-red-hat-openshift/).
          With this approach, you can enforce authentication and authorization policies
          for all of the applications in your cluster at the same time, without ever
          changing your application code. You can also dynamically register your application
          to start protecting them centrally from the cloud via Verify SaaS. \n\nSee
          the project [Readme](https://github.com/IBM-Security/verify-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Security Verify Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Product Information
          url: https://www.ibm.com/products/verify-saas
        - name: Verify Documentation
          url: https://www.ibm.com/docs/en/security-verify
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Development Team
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - icr.io/isv-saas/verify-operator@sha256:c3bed2e5db48e54884d2278d24ca4864f389a56a667db481c627d01ab5ffaba7
        version: 23.3.0
      entries:
      - name: ibm-security-verify-operator.v23.3.0
        version: 23.3.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hitachi
      provider-url: ""
    name: hspc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hspc-operator.v1.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.hitachi.com/v1",
                "kind": "HSPC",
                "metadata": {
                  "name": "hspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.hitachivantara.com/hitachicsi-oci-oss/hspc-operator@sha256:4dda4a43cefc19c8b1992756b499b1f9300920c5250dd77379f523997cfd5d54
          createdAt: "2025-06-13T10:30:39Z"
          description: An operator for managing Hitachi Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: Hitachi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HSPC is the Schema for the hspcs API
            displayName: HSPC
            kind: HSPC
            name: hspcs.csi.hitachi.com
            version: v1
        description: |-
          ## About
          Hitachi Storage Plug-in for Containers is a software plugin component for Kubernetes and OpenShift environments that is used to create and manage persistent block storage from Hitachi Virtual Storage Platform One storages.


          For full documentation, go to our [Product Documentation site](https://docs.hitachivantara.com/home) and refer to the reference guide for HSPC v3.16.0.

          ## Requirements

          ### Supported Driver Version

          * HSPC v3.16.0
        displayName: Hitachi Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hitachi
        - Storage
        - Container Storage Interface
        - CSI
        links:
        - name: Product Documentation
          url: https://docs.hitachivantara.com/home
        maintainers:
        - email: Container-info@hitachivantara.com
          name: Hitachi
        maturity: stable
        provider:
          name: Hitachi
        relatedImages:
        - registry.hitachivantara.com/hitachicsi-oci-oss/hspc-operator@sha256:4dda4a43cefc19c8b1992756b499b1f9300920c5250dd77379f523997cfd5d54
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - registry.hitachivantara.com/hitachicsi-oci-oss/hspc-csi-driver@sha256:2c580d98be55fffd8f6083379b6f23cbfe2aeb42fc4aec9e890072bc831c87c4
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:81d32545fb1b2c319bfeb36a70db4b4d3aea9c51c0574b209b8f64750b32d2f0
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        version: 1.16.0
      entries:
      - name: hspc-operator.v1.16.0
        version: 1.16.0
      name: stable
    defaultChannel: stable
    packageName: hspc-operator
    provider:
      name: Hitachi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CyberArk
      provider-url: https://www.cyberark.com/
    name: conjur-follower-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: conjur-follower-operator.v2.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:75fb5de2c111084ffcc3d973e6c63f1adfae78e29f6fd02152419fbca553ce7b",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:bee2cd31b3620a31e98f9d21326a2cdcecb36479099b06cc9660e4e83354d940",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:383776c23556a0a327a0bacc843d22c4b1516082ddd0cfb8a8275500f2eb841a",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:a9b109c7f3bfba5553aa4366338fa211896564f8e33856d75b4918acf8429f1e",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:b3096c17e7f3b9d723c906619dde293ea084e78c0da48c534335a8eee07b23c7",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:680a2044bc0e94fbc9376cd1eed9d067de91657f566c9752654480c939447841",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:ed53aea1866b1e2262386ac38cb9e1e47cc3f9047ae47f49070e2f506af6ca5a"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0 <2.5.3'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:cf46bdb4238cdb1dc51329822da55b7836958be2344d26f9e1819c0aa97fece6
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:383776c23556a0a327a0bacc843d22c4b1516082ddd0cfb8a8275500f2eb841a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:75fb5de2c111084ffcc3d973e6c63f1adfae78e29f6fd02152419fbca553ce7b
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:bee2cd31b3620a31e98f9d21326a2cdcecb36479099b06cc9660e4e83354d940
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:a9b109c7f3bfba5553aa4366338fa211896564f8e33856d75b4918acf8429f1e
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:b3096c17e7f3b9d723c906619dde293ea084e78c0da48c534335a8eee07b23c7
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:680a2044bc0e94fbc9376cd1eed9d067de91657f566c9752654480c939447841
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:ed53aea1866b1e2262386ac38cb9e1e47cc3f9047ae47f49070e2f506af6ca5a
        version: 2.5.3
      entries:
      - name: conjur-follower-operator.v2.5.3
        version: 2.5.3
      name: "2.5"
    - currentCSV: conjur-follower-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:c5337633eb3cc272f989e3e70c2e9bace9c819bba9e817537c7a691a8b6ee596
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813
        version: 2.6.3
      entries:
      - name: conjur-follower-operator.v2.6.3
        version: 2.6.3
      - name: conjur-follower-operator.v2.6.2
        version: 2.6.2
      - name: conjur-follower-operator.v2.6.1
        version: 2.6.1
      - name: conjur-follower-operator.v2.6.0
        version: 2.6.0
      - name: conjur-follower-operator.v2.5.0
        version: 2.5.0
      - name: conjur-follower-operator.v2.4.4
        version: 2.4.4
      - name: conjur-follower-operator.v2.4.3
        version: 2.4.3
      - name: conjur-follower-operator.v2.4.2
        version: 2.4.2
      - name: conjur-follower-operator.v2.3.5
        version: 2.3.5
      - name: conjur-follower-operator.v2.3.4
        version: 2.3.4
      - name: conjur-follower-operator.v2.3.1-290
        version: 2.3.1
      - name: conjur-follower-operator.v2.3.0
        version: 2.3.0
      - name: conjur-follower-operator.v2.2.4
        version: 2.2.4
      - name: conjur-follower-operator.v2.2.2
        version: 2.2.2
      - name: conjur-follower-operator.v2.2.1
        version: 2.2.1
      name: preview
    - currentCSV: conjur-follower-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:c5337633eb3cc272f989e3e70c2e9bace9c819bba9e817537c7a691a8b6ee596
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f
        version: 2.6.3
      entries:
      - name: conjur-follower-operator.v2.6.3
        version: 2.6.3
      - name: conjur-follower-operator.v2.6.2
        version: 2.6.2
      - name: conjur-follower-operator.v2.6.1
        version: 2.6.1
      - name: conjur-follower-operator.v2.6.0
        version: 2.6.0
      - name: conjur-follower-operator.v2.5.0
        version: 2.5.0
      - name: conjur-follower-operator.v2.4.4
        version: 2.4.4
      - name: conjur-follower-operator.v2.4.3
        version: 2.4.3
      - name: conjur-follower-operator.v2.4.2
        version: 2.4.2
      - name: conjur-follower-operator.v2.3.5
        version: 2.3.5
      - name: conjur-follower-operator.v2.3.4
        version: 2.3.4
      - name: conjur-follower-operator.v2.3.1-290
        version: 2.3.1
      - name: conjur-follower-operator.v2.3.0
        version: 2.3.0
      - name: conjur-follower-operator.v2.2.4
        version: 2.2.4
      - name: conjur-follower-operator.v2.2.2
        version: 2.2.2
      - name: conjur-follower-operator.v2.2.1
        version: 2.2.1
      name: stable
    defaultChannel: stable
    packageName: conjur-follower-operator
    provider:
      name: CyberArk
      url: https://www.cyberark.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: ovms-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openvino-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "ModelServer",
                "metadata": {
                  "name": "model-server-sample"
                },
                "spec": {
                  "image_name": "registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0",
                  "deployment_parameters": {
                    "replicas": 1
                  },
                  "service_parameters": {
                    "grpc_port": 8080,
                    "rest_port": 8081
                  },
                  "models_settings": {
                    "single_model_mode": true,
                    "config_configmap_name": "",
                    "config_path": "",
                    "model_name": "resnet",
                    "model_path": "gs://<bucket_name>/<model_dir>",
                    "nireq": 0,
                    "plugin_config": "{\"PERFORMANCE_HINT\":\"LATENCY\"}",
                    "batch_size": "",
                    "shape": "",
                    "model_version_policy": "{\"latest\": { \"num_versions\":1 }}",
                    "layout": "",
                    "target_device": "CPU",
                    "is_stateful": false,
                    "idle_sequence_cleanup": false,
                    "low_latency_transformation": true,
                    "max_sequence_number": 0
                  },
                  "server_settings": {
                    "file_system_poll_wait_seconds": 0,
                    "sequence_cleaner_poll_wait_minutes": 0,
                    "log_level": "INFO",
                    "grpc_workers": 1,
                    "rest_workers": 0
                  },
                  "models_repository": {
                    "https_proxy": "",
                    "http_proxy": "",
                    "storage_type": "google",
                    "models_host_path": "",
                    "models_volume_claim": "",
                    "aws_secret_access_key": "",
                    "aws_access_key_id": "",
                    "aws_region": "",
                    "s3_compat_api_endpoint": "",
                    "gcp_creds_secret_name": "",
                    "azure_storage_connection_string": ""
                  }
                }
              },
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "Notebook",
                "metadata": {
                  "name": "v2024.1"
                },
                "spec": {
                  "git_ref": "2024.1",
                  "auto_update_image": false
                }
              }
            ]
          alm-examples-metadata: '{ "model-server-sample":{"description":"Example
            of the model server configuration"}, "v2024":{"description":"Default notebook
            configuration"} }'
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
          createdAt: "2024-05-14T09:29:35Z"
          description: OpenVINO Toolkit Operator manages OpenVINO components in OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/ovms-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/ovms-operator-rhmp/support?utm_source=openshift_console
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.34.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Provides inference serving through gRPC & REST
            displayName: Model Server
            kind: ModelServer
            name: modelservers.intel.com
            version: v1alpha1
          - description: Adds a collection of ready-to-run Jupyter notebooks for learning
              and experimenting with the OpenVINO™ Toolkit to Red Hat® OpenShift®
              Data Science
            displayName: Notebook
            kind: Notebook
            name: notebooks.intel.com
            version: v1alpha1
        description: |
          OpenVINO Toolkit Operator manages OpenVINO components in OpenShift.

          Currently there available components are ModelServer and Notebook.

          # Model Server
          [OpenVINO™ Model Server](https://github.com/openvinotoolkit/model_server) (OVMS) is a scalable, high-performance solution for serving machine learning models optimized for Intel® architectures. The server provides an inference service via gRPC or REST API for any models trained in a framework that is supported by [OpenVINO](https://software.intel.com/en-us/openvino-toolki://docs.openvino.ai/latest/index.html).
          Model Server configuration parameters is explained [here](https://github.com/openvinotoolkit/operator/blob/main/docs/modelserver_params.md).
          ## Using the cluster
          OpenVINO Model Server can be consumed as a `Service`. It is called like with `ModelServer` resource with `-ovms` suffix.
          The suffix is omitted when `ovms` phrase is included in the name.
          The service exposes gRPC and REST API interfaces to run inference requests.
          ```
          oc get pods
          NAME                                        READY   STATUS    RESTARTS   AGE
          model-server-sample-ovms-586f6f76df-dpps4   1/1     Running   0          8h

          oc get services
          NAME                       TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)             AGE
          model-server-sample-ovms   ClusterIP   172.25.199.210   <none>        8080/TCP,8081/TCP   8h
          ```

          # Notebook
          The Notebook resource integrates JupyterHub from OpenShift Data Science or Open Data Hub with a container image that includes [developer
          tools](https://github.com/openvinotoolkit/openvino/blob/master/docs/install_guides/pypi-openvino-dev.md)
          from the OpenVINO toolkit and a set of Jupyter notebook tutorials. It enables selecting a defined image `OpenVINO™ Toolkit` from
          the Jupyter Spawner choice list.

          Create the `Notebook` resource in the same project with JupyterHub and RedHat OpenShift Data Science operator as seen in [documentation](https://github.com/openvinotoolkit/operator/blob/main/docs/notebook_in_rhods.md).
          It builds the image in the cluster based on Dockerfile from [openvino_notebooks](https://github.com/openvinotoolkit/openvino_notebooks).

          ## References
          OpenVINO Model Server on [Github](https://github.com/openvinotoolkit/model_server)

          OpenVINO Model Server Operator on [Github](https://github.com/openvinotoolkit/operator)
        displayName: OpenVINO Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Inference OpenVINO
        links:
        - name: OpenVINO Toolkit Operator
          url: https://github.com/openvinotoolkit/operator
        - name: OpenVINO Model Server
          url: https://github.com/openvinotoolkit/model_server
        - name: OpenVINO Notebooks
          url: https://github.com/openvinotoolkit/openvino_notebooks
        maintainers:
        - email: openvino-operator-support@intel.com
          name: Operator Support
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5714b93ac39277affdc10c3ba142d8c032ed0ddb2ba61c04cbddac710464c4
        - registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
        - registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0
        version: 1.2.0
      entries:
      - name: openvino-operator.v1.2.0
        version: 1.2.0
      - name: openvino-operator.v1.1.0
        version: 1.1.0
      - name: openvino-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: ovms-operator-rhmp
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/file-integrity-operator
    name: file-integrity-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: file-integrity-operator.v0.1.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <0.1.13'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator
          url: https://file-integrity-operator.domain
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/openshift4/file-integrity-rhel8-operator@sha256:a65f1244636649734763164a5d6155a6f6a7b79b3950ecc03fb3c58188583599
        version: 0.1.13
      entries:
      - name: file-integrity-operator.v0.1.13
        version: 0.1.13
      name: "4.7"
    - currentCSV: file-integrity-operator.v0.1.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <0.1.32'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:c3e34e6198eadb48c007cd949d8be38e26ef7d15d2f9047a8c3c89c03201e4ea
        version: 0.1.32
      entries:
      - name: file-integrity-operator.v0.1.32
        version: 0.1.32
      name: release-0.1
    - currentCSV: file-integrity-operator.v1.3.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          createdAt: "2024-11-21T18:02:08Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.3.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:f7039c6aaaa3e60dc4a3965cc1341c5217837e4b4030eb943771967e0b206578
        version: 1.3.6
      entries:
      - name: file-integrity-operator.v1.3.6
        version: 1.3.6
      - name: file-integrity-operator.v1.3.5
        version: 1.3.5
      - name: file-integrity-operator.v1.3.4
        version: 1.3.4
      - name: file-integrity-operator.v1.3.3
        version: 1.3.3
      name: stable
    - currentCSV: file-integrity-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <1.0.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:dd1dfeccf59214a0de653c3a4c4dbbc8c41640954318d13386cda73891660e4f
        version: 1.0.0
      entries:
      - name: file-integrity-operator.v1.0.0
        version: 1.0.0
      name: v1
    defaultChannel: stable
    packageName: file-integrity-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/file-integrity-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: resource-locker-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: resource-locker-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "ResourceLocker",
                "metadata": {
                  "name": "combined"
                },
                "spec": {
                  "patches": [
                    {
                      "id": "patch1",
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 0).metadata.name }}: {{ (index . 1).metadata.name }}\n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "kind": "Namespace",
                          "name": "resource-locker-test"
                        },
                        {
                          "apiVersion": "v1",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "resource-locker-test"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "resource-locker-test"
                      }
                    },
                    {
                      "id": "patch2",
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 0) }}: {{ (index . 1) }}\n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "fieldPath": "$.metadata.uid",
                          "kind": "Namespace",
                          "name": "resource-locker-test"
                        },
                        {
                          "apiVersion": "v1",
                          "fieldPath": "$.metadata.uid",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "resource-locker-test"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "resource-locker-test"
                      }
                    }
                  ],
                  "resources": [
                    {
                      "object": {
                        "apiVersion": "networking.k8s.io/v1",
                        "kind": "NetworkPolicy",
                        "metadata": {
                          "name": "allow-from-same-namespace",
                          "namespace": "resource-locker-test"
                        },
                        "spec": {
                          "ingress": [
                            {
                              "from": [
                                {
                                  "podSelector": {}
                                }
                              ]
                            }
                          ],
                          "podSelector": {}
                        }
                      }
                    },
                    {
                      "object": {
                        "apiVersion": "networking.k8s.io/v1",
                        "kind": "NetworkPolicy",
                        "metadata": {
                          "name": "allow-from-default-namespace",
                          "namespace": "resource-locker-test"
                        },
                        "spec": {
                          "ingress": [
                            {
                              "from": [
                                {
                                  "namespaceSelector": {
                                    "matchLabels": {
                                      "name": "default"
                                    }
                                  }
                                }
                              ]
                            }
                          ],
                          "podSelector": {}
                        }
                      }
                    }
                  ],
                  "serviceAccountRef": {
                    "name": "default"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/resource-locker-operator@sha256:2176db02f561c24012c5179495c89c3819500dff21e89a62052d465eb3314b44
          createdAt: "2021-11-30T19:38:40Z"
          description: This operator provides a facility to lock resources and / or
            patches into the state described by its custom resource preventing any
            drifts.
          operatorframework.io/suggested-namespace: resource-locker-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/resource-locker-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ResourceLocker Represents the intetion to create and enforce
              a set of resources and/or patches.
            displayName: Resource Locker
            kind: ResourceLocker
            name: resourcelockers.redhatcop.redhat.io
            version: v1alpha1
        description: |
          The resource locker operator allows you to specify a set of configurations that the operator will "keep in place" (lock) preventing any drifts.
          Two types of configurations may be specified:
          * Resources. This will instruct the operator to create and enforce the specified resource. In this case the operator "owns" the created resources.
          * Patches to resources. This will instruct the operator to patch- and enforce the change on- a pre-existing resource. In this case the operator does not "own" the resource.
          Locked resources are defined with the `ResourceLocker` CRD. Here is the high-level structure of this CRD:
          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: ResourceLocker
          metadata:
            name: test-simple-resource
          spec:
            resources:
            - object:
                apiVersion: v1
            ...
            patches:
            - targetObjectRef:
              ...
              patchTemplate: |
                metadata:
                  annotations:
                    ciao: hello
            ...
            serviceAccountRef:
              name: default
          ```
          It contains:
          * `resources`: representing an array of resources
          * `patches`: representing an array of patches
          * `serviceAccountRef`: a reference to a service account defined in the same namespace as the ResourceLocker CR, that will be used to create the resources and apply the patches. If not specified the service account will be defaulted to: `default`
          For each ResourceLocker a manager is dynamically allocated. For each resource and patch a controller with the needed watches is created and associated with the previously created manager.
          **Note**: creating a `ResourceLocker` CR will not work without due preparation (of the service account in particular). Follow these [instructions](https://github.com/redhat-cop/resource-locker-operator/blob/master/EXAMPLES.md) for a step by step walkthrough of the needed prerequisites to make this operator work.
        displayName: Resource Locker Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - resource_management
        - enforcing
        - patches
        - gitops
        links:
        - name: Resource Locker Operator
          url: https://github.com/redhat-cop/resource-locker-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/resource-locker-operator:latest
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/resource-locker-operator@sha256:2176db02f561c24012c5179495c89c3819500dff21e89a62052d465eb3314b44
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.1.3
      entries:
      - name: resource-locker-operator.v1.1.3
        version: 1.1.3
      - name: resource-locker-operator.v1.1.2
        version: 1.1.2
      - name: resource-locker-operator.v1.1.1
        version: 1.1.1
      - name: resource-locker-operator.v1.1.0
        version: 1.1.0
      - name: resource-locker-operator.v1.0.9
        version: 1.0.9
      name: alpha
    defaultChannel: alpha
    packageName: resource-locker-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Jason Skrzypek
      provider-url: https://github.com/shpwrck
    name: openshift-node-upgrade-mutex-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-node-upgrade-mutex-operator.v0.2.66
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexRule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutexrule-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutexrule",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutexrule-sample"
                },
                "spec": {
                  "jobName": "example",
                  "jobNamespace": "node-upgrade-mutex-operator",
                  "jobSpec": {
                    "selector": {},
                    "template": {
                      "metadata": {
                        "name": "mutex"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              "20"
                            ],
                            "image": "ubi9/toolbox",
                            "name": "mutex"
                          }
                        ],
                        "restartPolicy": "Never"
                      }
                    }
                  },
                  "type": "kubernetes"
                }
              },
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexRunConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutexrunconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutexrunconfig",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutexrunconfig-sample"
                },
                "spec": {
                  "canary": {
                    "name": "canary-example"
                  },
                  "mutexRules": [
                    {
                      "name": "mutexrule-sample",
                      "namespace": "node-upgrade-mutex-operator"
                    }
                  ],
                  "mutexTargets": [
                    {
                      "name": "mutextarget-sample",
                      "namespace": "node-upgrade-mutex-operator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexTarget",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutextarget-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutextarget",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutextarget-sample"
                },
                "spec": {
                  "machineConfigPools": [
                    "worker"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/jskrzypek/openshift-node-upgrade-mutex-operator
          createdAt: "2025-05-30T17:49:00Z"
          description: |
            This software injects required operations into the worker node upgrade process.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: node-upgrade-mutex-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          provider: Jason Skrzypek
          repository: https://github.com/shpwrck/openshift-upgrade-rollout-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A node mutex rule definition in the form of a kubernetes
              job spec
            displayName: Mutex Rule
            kind: MutexRule
            name: mutexrules.openshift-optional.skrzypek.io
            version: v1alpha1
          - description: Lists of Targets and Rules to apply to them, as well as an
              upgrade facilitating Machine Config Pool
            displayName: Mutex Run Config
            kind: MutexRunConfig
            name: mutexrunconfigs.openshift-optional.skrzypek.io
            version: v1alpha1
          - description: A list of Machine Config Pools to enforce upgrade procedure.
            displayName: Mutex Target
            kind: MutexTarget
            name: mutextargets.openshift-optional.skrzypek.io
            version: v1alpha1
        description: |
          ## Introduction

          This operator injects required operations in the form of a Kubernetes job into the worker node upgrade process using MachineConfigPools. In short, when provided a list of operations (`MutexRules`) and a list of MachineConfigPools (`MutexTargets`) , the OpenShift Node Upgrade Mutex Operator will execute each operation for a given target, enable updates for the given target by modifying it's MachineConfigPool membership, and wait until the node has been upgraded before proceeding onto subsequent operations and targets.

          ## Dependencies

          ### Release Version Label

          In order to confirm the successful upgrade of each target node, the NFD Operator is used to generate a version feature on every worker node. In order to proceed through given targets, each node must have this label
          `'machineconfiguration.openshift.io/release-image-version'` with a key providing the semantic version of this node (e.g. `'4.16.4'`)

          There are several ways to accomplish this, but an example MachineConfig has been provided [here](https://github.com/shpwrck/openshift-node-upgrade-mutex-operator/blob/main/machineconfig.yaml).

          ### Canary Machine Config Pool

          In addition to node labels, a MachineConfigPool that will be used to accomodate targets subject to upgrades must be provided matching the following criteria:
          * selects only worker nodes
          * is not paused

          Applying the example below will create a compatible canary pool:

          ```yaml
          ---
          apiVersion: machineconfiguration.openshift.io/v1
          kind: MachineConfigPool
          metadata:
            name: canary
          spec:
            machineConfigSelector:
              matchLabels:
                machineconfiguration.openshift.io/role: worker
            nodeSelector:
              matchLabels:
                node-role.kubernetes.io/worker: ''
              matchExpressions:
                - key: 'isCanary'
                  operator: Exists
            paused: false
          ```

          ## Usage / Custom Resources Definitions

          #### MutexRule:

          A resource storing a single rule referring to a template for a Kubernetes job to be created.
          This resource **does not** modify existing resources.

          ```yaml
          ---
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexRule
          metadata:
            name: k8s-example
          spec:
            jobName: "example" # Jobs will be run with this prefix name
            jobNamespace: "node-upgrade-mutex-operator" # Jobs will be run in this namespace
            jobSpec: # Accepts full 'batch/v1' job spec
              selector: {}
              template:
                metadata:
                  name: mutex
                spec:
                  containers:
                  - command:
                    - sleep
                    args:
                    - "30"
                    image: ubi9/toolbox
                    name: mutex
                  restartPolicy: Never
            cleanUpJobSpec: # Accepts full 'batch/v1' job spec (Optional)
              selector: {}
              template:
                metadata:
                  name: mutex
                spec:
                  containers:
                  - command:
                    - sleep
                    args:
                    - "30"
                    image: ubi9/toolbox
                    name: mutex
                  restartPolicy: Never
            type: kubernetes
          ```

          #### MutexTarget:

          A resource storing a set of MachineConfigPools to pause and incrementally upgrade.
          This resource **does not** modify existing resources.

          ```yaml
          ---
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexTarget
          metadata:
            name: worker-example
          spec:
            machineConfigPools:
              - worker # Name of existing machineconfigpool
          ```

          #### MutexRunConfig:

          A resource that references existing MutexRules and MutexTargets as well as an existing MachineConfigPool that will allow members to be upgraded. Every rule defined in the `mutexRules` list will be added as a mandatory task to every node that is a member of the each MachineConfigPool defined in the `mutexTargets` list.
          This resource **does** modify existing resources.

          ```yaml
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexRunConfig
          metadata:
            name: example
          spec:
            skipTLSVerify: false # Defaults to false if omitted
            canary:
              name: canary # Name of unpaused machineconfigpool to modify
            mutexRules: # List of existing mutexRules
              - name: k8s-example
                namespace: node-upgrade-mutex-operator
            mutexTargets: # List of existing mutexTargets
              - name: worker-example
                namespace: node-upgrade-mutex-operator
          ```

          ## Procedure

          #### MutexRule

          1. Upon creation/modification, the schema of each K8s rule is validated.
          2. Status is updated to indicate valid or invalid configuration.

          #### MutexTargets

          1. Upon creation/modification, the existence of each target is validated.
          2. Status is updated to indicate valid or invalid configuration.
          3. The nodes assigned to each target are added to the status field `nodes`.

          #### MutexRunConfigs

          1. Upon creation/modification, the existence as well as validation status of each rule, target, and canary pool is confirmed.
          2. If valid, the existence of an upgrade is queried. An upgrade is available if the following is true:
              * The `status.history` of the `cluster` ClusterVersion equals `Completed`
              * The `release-image-version` of any worker node is less than the `spec.desiredUpdate.version` of the `cluster` ClusterVersion.
          3. If an update exists, each target MachineConfigPool has `spec.paused` set to `true`.
          4. If each pool is paused, each node is updated to include a label for each rule with the following format:
              * `mutexrule.openshift-optional.skrzypek.io/{{MutexRule_Name}}.{{MutexRule_Namespace}}`
          5. If labels have been applied, the provided canary pool is modified to exclude nodes with labels provided by step (4).
          6. Assuming steps (2-5) have completed successfully, then the process for running jobs and updating nodes begins. The process per node is as follows:
              * Launch each K8S job sequentially
              * Remove the current node's rule labels
              * Wait until the `release-image-version` of the node matches the `spec.desiredUpdate.version` of the `cluster` ClusterVersion
              * Launch each K8S cleanup job sequentially
          7. If the update succeeds, all machine config pools are unpaused and all match expressions are removed.

          ## Permissions Granted

          In addition to the permissions required to operate the included CRDs, the following access is granted to the operator's service account.

          | Resource | Permissions |
          | :---: | :---: |
          | Jobs | Create & Read |
          | Nodes | Read & Update |
          | ClusterVersion | Read |
          | MachineConfigPools | Read & Update |

          ## Notes

          * This operator was built using the operator-sdk's ansible plugin and therefore has all of the inherited capabilities and restrictions.
          * This operator has undergone minimal testing and should not be run in production environments without acknowledging the risks.
          * This operator is currently alpha and subject to change based on requirements.

          ## Additional Information
          For additional information refer to the [GitHub Repository](https://github.com/shpwrck/openshift-node-upgrade-mutex-operator)
        displayName: OpenShift Node Upgrade Mutex Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - OpenShift Optional
        - Upgrade
        - Mutex
        links:
        - name: Openshift Node Upgrade Mutex Operator
          url: https://github.com/shpwrck/openshift-node-upgrade-mutex-operator
        maintainers:
        - email: jskrzype@redhat.com
          name: Jason Skrzypek
        maturity: alpha
        minKubeVersion: 1.27.0
        provider:
          name: Jason Skrzypek
          url: https://github.com/shpwrck
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - quay.io/jskrzypek/openshift-node-upgrade-mutex-operator@sha256:d4c28e8b256c00429d9999595c2386f1a25de597ee8e6c33871bb7b22b84c4af
        version: 0.2.66
      entries:
      - name: openshift-node-upgrade-mutex-operator.v0.2.66
        version: 0.2.66
      - name: openshift-node-upgrade-mutex-operator.v0.2.65
        version: 0.2.65
      - name: openshift-node-upgrade-mutex-operator.v0.2.60
        version: 0.2.60
      - name: openshift-node-upgrade-mutex-operator.v0.2.59
        version: 0.2.59
      - name: openshift-node-upgrade-mutex-operator.v0.2.53
        version: 0.2.53
      - name: openshift-node-upgrade-mutex-operator.v0.2.38
        version: 0.2.38
      - name: openshift-node-upgrade-mutex-operator.v0.2.34
        version: 0.2.34
      - name: openshift-node-upgrade-mutex-operator.v0.2.15
        version: 0.2.15
      - name: openshift-node-upgrade-mutex-operator.v0.2.14
        version: 0.2.14
      - name: openshift-node-upgrade-mutex-operator.v0.1.22
        version: 0.1.22
      name: alpha
    defaultChannel: alpha
    packageName: openshift-node-upgrade-mutex-operator
    provider:
      name: Jason Skrzypek
      url: https://github.com/shpwrck
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Digital.ai
      provider-url: https://digital.ai
    name: digitalai-deploy-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: digitalai-deploy-operator.v25.3.0-beta.813
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "quorum",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.0-debian-12-r4"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.0-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.5.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.2-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabsunsupported/deploy-operator@sha256:9ccca8ab999d4b6058c979bbb70568a63cf43a257ae65a787b2ff4d9bf470c39 '
          createdAt: "2025-09-02T05:43:37Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabsunsupported/deploy-operator@sha256:9ccca8ab999d4b6058c979bbb70568a63cf43a257ae65a787b2ff4d9bf470c39
        - docker.io/xebialabsunsupported/xl-deploy@sha256:41feb1d24d70a0c78a560b338d05c96c5383d570e3ef1ff37daff0eb8c92b386
        - docker.io/xebialabsunsupported/deploy-task-engine@sha256:15322f2263c2b9b70336177aec170e59497fa320281c2d73605918b30272b8f1
        - docker.io/xebialabsunsupported/central-configuration@sha256:3a35c2aada71a7f5ab8fe7d5c68441fe58b0b2c778b6ea85ac6f3f7b351f80ad
        version: 25.3.0-beta.813
      entries:
      - name: digitalai-deploy-operator.v25.3.0-beta.813
        version: 25.3.0-beta.813
      - name: digitalai-deploy-operator.v25.3.0-beta.711
        version: 25.3.0-beta.711
      - name: digitalai-deploy-operator.v25.3.0-beta.609
        version: 25.3.0-beta.609
      - name: digitalai-deploy-operator.v25.3.0-beta.512
        version: 25.3.0-beta.512
      name: beta
    - currentCSV: digitalai-deploy-operator.v24.1.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "ingressClass": "haproxy-daid",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "tag": "1.26.0-debian-12-r1"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "tag": "1.10.1-debian-12-r5"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "tag": "15.6.0-debian-12-r7"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "tag": "3.12.13-debian-12-r2"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator without SCC creation and with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:2b3413cd41f2b437e6027c8ae2cff12d1113e0ac1be69fc24840d68a05d3a665 '
          createdAt: "2024-10-16T05:39:21Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind:
          DigitalaiDeploy\nmetadata:\n  name: daid-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  securityContextConstraints:\n
          \   enabled: false\n  centralConfiguration:\n    replicaCount: 1\n    podSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  master:\n    replicaCount: 1\n    persistence:\n
          \     storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n      runAsUser:
          null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  worker:\n    replicaCount: 1\n    persistence:\n
          \     storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n      runAsUser:
          null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  route:\n    enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n        runAsUser: null\n        runAsGroup: null\n
          \       fsGroup: null\n      containerSecurityContext:\n        enabled:
          false\n        runAsUser: null\n        runAsGroup: null\n      securityContextConstraints:\n
          \       enabled: false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n
          \   install: true\n    persistence:\n      size: 1Gi\n      storageClass:
          ''\n    replicaCount: 1\n    podSecurityContext:\n      enabled: false\n
          \     runAsUser: null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     enabled: false\n      runAsUser: null\n      runAsGroup: null\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"daid-doc-digitalai-deploy-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xld-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Embedded PostgreSQL for
          DB management\n- Embedded RabbitMQ for message queue management\n\n### Security
          Configuration Details\n\n- `securityContextConstraints.enabled: false` -
          disables creation of SCCs;\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/deploy-operator@sha256:2b3413cd41f2b437e6027c8ae2cff12d1113e0ac1be69fc24840d68a05d3a665
        - docker.io/xebialabs/deploy-task-engine@sha256:8a381fd84483ea919fa7f8bee5cf19d9dec0d1d61cce73ec6b7c56fa504c6c93
        - docker.io/xebialabs/xl-deploy@sha256:77fdf6a3909367b060eba2ef698baddb77e4086d4d977f9c617d1e26a4427472
        - docker.io/xebialabs/central-configuration@sha256:83359e49c66e4fab502cc75663b456eb37d81d7cb512f2c48667f4d6a324e0df
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:771a9a173e033a3ad8b46f5c00a7036eaa88c8d8d1fbd89217325168998113ea
        version: 24.1.10
      entries:
      - name: digitalai-deploy-operator.v24.1.10
        version: 24.1.10
      name: stable
    - currentCSV: digitalai-deploy-operator.v24.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "image": {
                        "tag": "v0.14.7"
                      },
                      "ingressClass": "haproxy-daid",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.26.2-debian-12-r10"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "15.6.0-debian-12-r20"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.12.14-debian-12-r7"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator without SCC creation and with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:7d505a855728c761433ef7d46f0a64ffe2b4192f4819bad86c23582375099ac9 '
          createdAt: "2025-08-25T04:53:23Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind:
          DigitalaiDeploy\nmetadata:\n  name: daid-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  securityContextConstraints:\n
          \   enabled: false\n  centralConfiguration:\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  master:\n    replicaCount:
          1\n    persistence:\n      storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  worker:\n    replicaCount:
          1\n    persistence:\n      storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  route:\n    enabled: false\n
          \   annotations:\n      haproxy.router.openshift.io/cookie_name: SESSION_XLD\n
          \     haproxy.router.openshift.io/disable_cookies: \"false\"\n      haproxy.router.openshift.io/rewrite-target:
          /\n    hostname: '<mandatory-deploy-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n      securityContextConstraints:\n
          \       enabled: false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n
          \   install: true\n    persistence:\n      size: 1Gi\n      storageClass:
          ''\n    replicaCount: 1\n    podSecurityContext:\n      enabled: false\n
          \   containerSecurityContext:\n      enabled: false\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"daid-doc-digitalai-deploy-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xld-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Embedded PostgreSQL for
          DB management\n- Embedded RabbitMQ for message queue management\n\n### Security
          Configuration Details\n\n- `securityContextConstraints.enabled: false` -
          disables creation of SCCs;\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/xl-deploy@sha256:c2fe209f198fb2f42dc78a2e1544c026a6891d60682b4e47634234bbf7346540
        - docker.io/xebialabs/central-configuration@sha256:581a0bc71ae3652ceb643d81927fcf8cabea92f1d301f5d354076baff32fbcde
        - docker.io/xebialabs/deploy-operator@sha256:7d505a855728c761433ef7d46f0a64ffe2b4192f4819bad86c23582375099ac9
        - docker.io/xebialabs/deploy-task-engine@sha256:a97d67ea53e1b954ffb4d9514fd61426dcc0cab08922009eef02f32162026d6a
        version: 24.1.21
      entries:
      - name: digitalai-deploy-operator.v24.1.21
        version: 24.1.21
      - name: digitalai-deploy-operator.v24.1.20
        version: 24.1.20
      - name: digitalai-deploy-operator.v24.1.19
        version: 24.1.19
      - name: digitalai-deploy-operator.v24.1.18
        version: 24.1.18
      - name: digitalai-deploy-operator.v24.1.17
        version: 24.1.17
      - name: digitalai-deploy-operator.v24.1.14
        version: 24.1.14
      - name: digitalai-deploy-operator.v24.1.10
        version: 24.1.10
      name: stable-v24.1
    - currentCSV: digitalai-deploy-operator.v24.3.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "16.3.0-debian-12-r23"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n{{ if eq .Values.queueType \"quorum\" -}}\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n{{- end }}\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.13.7-debian-12-r5"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "queueType": "classic",
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:41ffc26c1d99a808e537f2366d92473ecf76dbffbac699dfa04785c3016f7470 '
          createdAt: "2025-09-26T18:18:41Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/central-configuration@sha256:4b83d4e9a523ae3615579907c65790478a3b22315ca5afce594793424d1d33b1
        - docker.io/xebialabs/deploy-operator@sha256:41ffc26c1d99a808e537f2366d92473ecf76dbffbac699dfa04785c3016f7470
        - docker.io/xebialabs/xl-deploy@sha256:8536efdf530dde45b84b14191ebfdc7ec5ebb1844f84871032fe2d335b13908d
        - docker.io/xebialabs/deploy-task-engine@sha256:75fe942305bf3bb403eb243b157b88844363eb702f03227a61e76ad4321bf0ee
        version: 24.3.11
      entries:
      - name: digitalai-deploy-operator.v24.3.11
        version: 24.3.11
      - name: digitalai-deploy-operator.v24.3.10
        version: 24.3.10
      - name: digitalai-deploy-operator.v24.3.9
        version: 24.3.9
      - name: digitalai-deploy-operator.v24.3.8
        version: 24.3.8
      - name: digitalai-deploy-operator.v24.3.7
        version: 24.3.7
      - name: digitalai-deploy-operator.v24.3.6
        version: 24.3.6
      - name: digitalai-deploy-operator.v24.3.5
        version: 24.3.5
      name: stable-v24.3
    - currentCSV: digitalai-deploy-operator.v25.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "quorum",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.1-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.4.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.3-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:d192ef9008d72c9792e1b7b988bdae8f1bbce45ed0e44afc48e3e268b657b38b '
          createdAt: "2025-09-26T16:30:53Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/deploy-task-engine@sha256:13bb779d6041836acb95c95f1eed75f4b5b2940d75687306183a9f8a2177cab3
        - docker.io/xebialabs/central-configuration@sha256:0c46275ffa3b251ccb76d2623e0e5a4cd1441bf7f94901845bda743f087bc6e4
        - docker.io/xebialabs/xl-deploy@sha256:e4212d20fad19f09cbe53796371bec7fc368cfe63316f1b7bccecf5faf6ea733
        - docker.io/xebialabs/deploy-operator@sha256:d192ef9008d72c9792e1b7b988bdae8f1bbce45ed0e44afc48e3e268b657b38b
        version: 25.1.5
      entries:
      - name: digitalai-deploy-operator.v25.1.5
        version: 25.1.5
      - name: digitalai-deploy-operator.v25.1.4
        version: 25.1.4
      - name: digitalai-deploy-operator.v25.1.3
        version: 25.1.3
      - name: digitalai-deploy-operator.v25.1.2
        version: 25.1.2
      - name: digitalai-deploy-operator.v25.1.1
        version: 25.1.1
      - name: digitalai-deploy-operator.v25.1.0
        version: 25.1.0
      name: stable-v25.1
    defaultChannel: stable-v24.3
    packageName: digitalai-deploy-operator
    provider:
      name: Digital.ai
      url: https://digital.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://access.redhat.com/products/red-hat-fuse
    name: fuse-online
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-online.v7.10.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:ed4faedcb97bb6a45f568586ff19da19da31ecbc304f12e0ae7978381247744a
          createdAt: 2022-01-31 15:55:46 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:ed4faedcb97bb6a45f568586ff19da19da31ecbc304f12e0ae7978381247744a
        - registry.redhat.io/rhel8/postgresql-12@sha256:44215c5b244b190c82d9d371a8ffd39b0c83576bd85c980cc72782a5cf9e6e1b
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:5048cd4fab2f98e6000bc60835c539cc40b70dbe265e927cfe185595347d8054
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:5dd1bc8f6b7e724736f4aeac86409bda3e0f64f62b736688e9572d14411232f1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:fcdd0292ebfc7e9edc36241f63b54a9d924a7605329c934ff2b10373af26dbbf
        - registry.redhat.io/openshift4/ose-prometheus@sha256:7f679d6c03535e300dac366f4bc6dd8d781d96dc47592836713f8bb76be36fa9
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:4f8849362e9ed6ff95d318a04e287b7ba33254a2cf7abdaf9b90cab7cbba6646
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:df2d55dd8092e9e94d034166890117cf5d1cfd5cf52ef0f90cfa2374cc9e2d06
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:3f0277e2c278740307c9b2d966f539ba12317e306198cb6c4c1c52a2ad9f4f40
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:96810c770f75057b8bba6b78a5ba414c8289315d8af1a387f681ae5e4fb7d6dd
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:ef229a4c8851ec495f28d0eeab83b137cd02698e283f76ea1b32271b7222621b
        version: 7.10.4
      entries:
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.10.x
    - currentCSV: fuse-online.v7.11.1-0.1687357787.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:658af3fc7819efb7cc39cd3a0504dcddc2ce7788c9b1e5178df869f7c9a742a8
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.11.1-0.1684718286.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:a2b1cd0256f30aff8fa37c332419b4483c0bfe34fb4f574ad8f18b0858969342
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:4de91e4527022e1b25df8d3835ddf778f773c20334314be9ba01a885c82287f5
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:8b9bd777bcbd20b1a1bbbc18274a5076a67bea1f928d772cb243c57fb1da7408
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:ac514973ec428d815d7a248d09d27f2a84f762aedfea63b3b48de9e633267ed5
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:658af3fc7819efb7cc39cd3a0504dcddc2ce7788c9b1e5178df869f7c9a742a8
        - registry.redhat.io/rhel8/postgresql-12@sha256:9248c4eaa8aeedacc1c06d7e3141ca1457147eef59e329273eb78e32fcd27e79
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:251c29e81f8e338f980b475b841647a977807908c388f4c86e18da7bff1b198a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:400a6ee03c3200c19b9b15c5d037a849633156a62e13922fb2e62838024ae69c
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:3d98d775bec85704c9a10c2f0d8e4f8116b7a41494f73c0b763e010262e5cd4d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        version: 7.11.1+0.1687357787.p
      entries:
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.11.x
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.12.x
    - currentCSV: fuse-online.v7.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta2",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:c70c5e76dc3db559c722bea2a767eb7dcba1eda786e1ef3a45c4cf32e1aca9ee
          createdAt: 2021-07-21 12:25:15 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta2
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to the product documentation at https://access.redhat.com/documentation/en-us/red_hat_fuse
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:d6df09b15599e776b49e1d46aa01728fd02d5629cee65e77dc277c0166554e0a
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:c668fe1f5d25baddf627acd79df4c8d2ab8b05309d3d41784ddc0a10dd5f34ad
        - registry.redhat.io/openshift4/ose-prometheus@sha256:13e9ea359cbcd3e0e94a6589b86b708dc84197f152701fd4092fffa234d11b58
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:60d5cf00411b03af3f1149a2bb3aeeb196c063d19625f876a094ae6baf46b54b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:f65600506db985689e5b7cd8b838e646d4268a0c4e272e19d7db3577a8dd7ac4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:75d5772c792a7c795281acec3d59e90021c109e9d7d1110868be0b882768aefb
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:c70c5e76dc3db559c722bea2a767eb7dcba1eda786e1ef3a45c4cf32e1aca9ee
        - registry.redhat.io/rhscl/postgresql-12-rhel7@sha256:c3e95e808eecba8ac57632452d3553509131ca783f62c68754a582cf5c55efab
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:eff9f8edf28ccbadcacab2022fff9e6bc93926af97521b49e3968eaf7b6fda65
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:d6f0c804c179fa559d3baed2e2d106c916a0439dc2215c4d34b878bc2e9bc281
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bafd85b0fd1256b8a60a2bb368a7e68f5dbad72718e5059a65660e01a425d8e2
        version: 7.9.3
      entries:
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.9.x
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: candidate
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: latest
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: stable
    defaultChannel: latest
    packageName: fuse-online
    provider:
      name: Red Hat
      url: https://access.redhat.com/products/red-hat-fuse
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: eG Innovations
      provider-url: https://www.eginnovations.com
    name: eginnovations-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eginnovations-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eginnovations.com/v1beta1",
                "kind": "EgUniversalAgent",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eginnovations-operator"
                  },
                  "name": "eguniversalagent",
                  "namespace": "egagent"
                },
                "spec": {
                  "agent": {
                    "env": [
                      {
                        "name": "EG_MANAGER",
                        "value": "Replace with eG Manager IP"
                      },
                      {
                        "name": "EG_MANAGER_PORT",
                        "value": "Replace with eG Manager Port"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "true"
          containerImage: docker.io/eginnovations/universal-agent-operator:0.0.11
          createdAt: "2025-09-11T20:08:00Z"
          description: eG Innovations provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with eG Universal Agent Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eginnovations/universal-agent-operator
          support: eG Innovations
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgUniversalAgent is the Schema for the eguniversalagents
              API
            displayName: Eg Universal Agent
            kind: EgUniversalAgent
            name: eguniversalagents.eginnovations.com
            version: v1beta1
        description: |
          eG Innovations’ converged application and infrastructure performance monitoring capabilities provide end-to-end visibility and correlated analytics for every layer, every tier of your IT landscape.

          The eG Enterprise Universal Agent Operator automatically configures a host agent on every Kubernetes worker node. The host agent auto-discovers the worker nodes and application containers running as Pods on each node and tracks their performance and utilization levels. In-depth monitoring of applications running on containers is also provided using the same host agent. No additional agents are required for the containers.

          ### Capabilities
            * Rolls out the egagent pod per node to monitor its pods and the node itself
            * Automatic BTM profiler injection using webhook based injection for deep application monitoring and transection tracing.

          ### Installation
            1. Create the namespace "egagent" for installing eG Agent on OpenShift Cluster.

               <br />``` $ oc create ns egagent ```
            2. Add "egagent" serviceaccount to scc( Security Context Constraints ) of OpenShift Cluster as privileged using below command.

               <br />``` $ oc adm policy add-scc-to-user privileged -z egagent -n egagent ```
            2. Install eG Innovations Universal Agent operator from [OpenShift Container Platform](https://www.openshift.com/) OperatorHub.
            3. Create the eG Universal Agent CRD (Custom Resource Definition)
               a. Click "View YAML Example" and copy the yaml contents and save it to (filename).yaml on your host.
               b. To apply the CRD run the following command

               <br />``` $ oc apply -f (filename).yaml ```

          ### Required Parameters
            * EG\_MANAGER, EG\_MANAGER\_PORT - provide the ip and port of your eG Manager Installation.

          ### Advanced Options
            * Disable Certificate Checking - disable any certificate validation that may interact poorly with proxies within your cluster
            * Image Override - use a copy of the eG Iniversal Agent container image from a registry other that Docker's or Red Hat's
            * Environment variables - define environment variables for the eG Universal Agent container of for APM profiler.
            * Namespace Selectors- select a subset of namesapces to enable instrumentation for applications.

          For complete list of supported parameters and the detailed documentation available in [eG Innovations Docs](https://docs.eginnovations.com)
        displayName: eG Innovations Universal Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Tracing
        links:
        - name: eG Innovations Kubernetes Monitoring Overview
          url: https://www.eginnovations.com/documentation/Kubernetes/Introduction-to-Monitoring-Kubernetes.htm
        maintainers:
        - email: devops@eginnovations.com
          name: eG Innovations
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: eG Innovations
          url: https://www.eginnovations.com
        relatedImages:
        - docker.io/eginnovations/universal-agent-operator:0.0.11
        version: 0.0.11
      entries:
      - name: eginnovations-operator.v0.0.11
        version: 0.0.11
      - name: eginnovations-operator.v0.0.10
        version: 0.0.10
      - name: eginnovations-operator.v0.0.9
        version: 0.0.9
      - name: eginnovations-operator.v0.0.8
        version: 0.0.8
      name: beta
    defaultChannel: beta
    packageName: eginnovations-operator
    provider:
      name: eG Innovations
      url: https://www.eginnovations.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: APIMatic.io
      provider-url: https://www.apimatic.io/
    name: apimatic-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apimatic-kubernetes-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicodegen.apimatic.io/v1beta1",
                "kind": "APIMatic",
                "metadata": {
                  "name": "apimatic-sample"
                },
                "spec": {
                  "licensespec": {
                    "licenseSourceName": "apimaticlicense",
                    "licenseSourceType": "ConfigMap"
                  },
                  "podspec": {
                    "apimaticContainerSpec": {
                      "image": "apimaticio/apimatic-codegen",
                      "imagePullSecret": "apimaticimagesecret"
                    }
                  },
                  "replicas": 3,
                  "servicespec": {
                    "apimaticserviceport": {
                      "nodePort": 32000,
                      "port": 8070
                    },
                    "servicetype": "NodePort"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: docker.io/apimaticio/apimatic-kubernetes-operator:v0.0.1
          createdAt: "2021-08-24 12:00:00"
          description: Generate client SDKs and interactive Documentation for your
            APIs in minutes
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/apimatic/apimatic-kubernetes-operator
          support: APIMatic.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIMatic is the Schema for the apimatics API
            displayName: APIMatic App
            kind: APIMatic
            name: apimatics.apicodegen.apimatic.io
            version: v1beta1
        description: "##About APIMatic \n\n APIMatic is a Developer Experience   platform
          for Web APIs that empowers users to create multi-language SDKs with API
          reference docs, dynamic code samples, assisted authentication and more -
          all auto-generated to provide a comprehensive developer experience beyond
          any API documentation tool.\n\n## Table of contents\n\n* [Introduction](#introduction)\n*
          [Features](#features)\n* [Running the Sample](#running-the-sample)\n  *
          [Prerequisites](#prerequisites)\n  * [Steps for Direct Deployment](#steps-for-direct-deployment)\n
          \ * [Steps for OLM Deployment](#steps-for-olm-deployment)\n* [Technical
          Support](#technical-support)\n* [Copyrights](#copyrights)\n\n## Introduction\n\nAPIMatic
          Operator simplifies the configuration and lifecycle management of the APIMatic
          code and docs generation solution on different Kubernetes distributions
          and OpenShift. The Operator encapsulates key operational knowledge on how
          to configure and upgrade the APIMatic CodeGen application, making it easy
          to get up and running.\n\n\nMore information about the underlying APIMatic
          CodeGen API that is exposed\nby this operator can be found [here](https://apimatic-core-v3-docs.netlify.app/#/http/getting-started/overview-apimatic-core).\n\n##
          Features\n\nAPIMatic Operator provides the following features:\n- Deploys
          the APIMatic CodeGen Web API service within the Kubernetes or OpenShift
          cluster.\n- Exposes the APIMatic CodeGen API external to the cluster, using
          Service type as [NodePort](https://kubernetes.io/docs/concepts/services-networking/service/#nodeport),
          [LoadBalancer](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          or [ExternalName](https://kubernetes.io/docs/concepts/services-networking/service/#externalname).\n-
          For exposing the service through an ingress resource, create an Ingress
          resource in the namespace of your APIMatic CR and set owned APIMatic service
          created by the operator as a backed service. More information can be found
          [here](https://kubernetes.io/docs/concepts/services-networking/ingress/).\n-
          Manual horizontal scaling of pods.\n  ```sh\n  kubectl scale apm apimatic-sample--replicas=2\n
          \ ```\n\n## Running the Sample \n\n### Prerequisites\n\nPlease contact APIMatic
          at [support@apimatic.io](mailto:support@apimatic.io) to register with the
          APIMatic CodeGen Docker Hub registry and acquire a valid license to run
          the APIMatic CodeGen API.\n\nFurther prerequisites for running the sample
          include:\n\n- [go](https://golang.org/) v1.16.*\n- [git](https://git-scm.com/)\n-
          [make](https://www.gnu.org/software/make/)\n- [Operator SDK](https://sdk.operatorframework.io/docs/overview/)\n-
          A running Kubernetes cluster with [kubectl](https://kubernetes.io/docs/reference/kubectl/overview/)
          on client. For testing purposes, you can use [Minikube](https://minikube.sigs.k8s.io/docs/)
          or [kind](https://kind.sigs.k8s.io/)\n- For checking the service created
          by the APIMatic operator on-prem, you can use [MetalLB](https://metallb.org/)\n\n###
          Steps for Direct Deployment\n\nTo run the sample for checking the APIMatic
          operator:\n\n- Clone the APIMatic repository into your working directory
          using the following command:\n  ```sh\n  git clone https://github.com/apimatic/apimatic-kubernetes-operator.git\n
          \ ```\n- Run `make deploy` to set up the APIMatic operator resources. This
          will deploy the '`apimatic-system`' namespace as well as the CRD and the
          RBAC manifests.\n\n- Create a secret named '`apimaticimagesecret`' to allow
          pulling the APIMatic CodeGen image using the Docker Hub username registered
          with APIMatic. If you have not done so, please contact support@apimatic.io
          for the steps required.\n  ```sh\n  kubectl create secret docker-registry
          apimaticimagesecret --docker-server=https://index.docker.io/v1/ --docker-username=<your-name>
          --docker-password=<your-pword> --docker-email=<your-email>\n  ```\n- Create
          a [configmap](https://kubernetes.io/docs/concepts/configuration/configmap/)
          resource named '`apimaticlicense`' that will serve as the volume storing
          the APIMatic license. An example of this is given below which will create
          the ConfigMap using the *License.lic* file located at */usr/local/apimatic/license/*\n
          \ ```sh\n  kubectl create configmap apimaticlicense --from-file /usr/local/apimatic/license/License.lic\n
          \ ```\n- This will deploy a ConfigMap resource with the following definition:\n
          \ ```sh\n  apiVersion: v1\n  data:\n    License.lic: \\\"<License file contents
          here>\\\"\n  kind: ConfigMap\n  metadata:\n    name: apimaticlicense\n    namespace:
          default\n  ```\n- Now run the sample using the following command:\n  ```sh\n
          \ kubectl apply -f config/samples/apicodegen_v1beta1_apimatic.yaml\n  ```\n-
          You will now see a new [Deployment](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/)
          with replica count of 3 and [Service](https://kubernetes.io/docs/concepts/services-networking/service/)
          of type `NodePort` created, both named ***apimatic-sample***. Accessing
          http://localhost:32000 from your browser (or using curl from within the
          Minikube or Kind cluster, if using that) should now show the APIMatic Titan
          page.\n- You can now use the exposed APIMatic CodeGen API to generate API
          SDKs and Docs using [curl](https://curl.se/), [Postman](https://www.postman.com/)
          or your own custom Web application that consumes the APIMatic CodeGen API
          service.\n\n- Once done, you can remove the APIMatic resources using the
          following command:\n  ```sh\n  make undeploy\n  ```\n\n### Steps for OLM
          Deployment\n\nThe following steps can be used to utilize [Operator LifeCycle
          Manager (OLM)](https://olm.operatorframework.io/docs/) to deploy the operator
          and run the sample. The steps are as follows:\n\n- If not already done so,
          clone the APIMatic repository into your working directory:\n  ```sh\n  git
          clone https://github.com/apimatic/apimatic-kubernetes-operator.git\n  ```
          \n- [Install OLM in your Kubernetes cluster](https://olm.operatorframework.io/docs/getting-started/#installing-olm-in-your-cluster).\n\n-
          Run the following script to install the resources required by OLM to deploy
          the APIMatic operator in the Kubernetes cluster within the `apimatic-system`
          namespace. Information about the different resources required can be found
          using the steps given [here](https://olm.operatorframework.io/docs/tasks/).\n
          \ ```sh\n  kubectl apply -f olm/manifests.yaml\n  ```\n\n- This should spin
          up the `ClusterServiceVersion` of the operator in the `apimatic-system`
          namespace, following which the operator pod will spin up. To ensure the
          that the operator installed successfully, check for the `ClusterServiceVersion`
          and the operator deployment in the namespace it was installed in.\n  ```sh\n
          \ kubectl get csv -n apimatic-system\n  \n  kubectl get deployment -n apimatic-system\n
          \ ```\n\n  - Create a secret named `apimaticimagesecret` to allow pulling
          the APIMatic CodeGen image using the Docker Hub username registered with
          APIMatic. If you have not done so, please contact support@apimatic.io for
          the steps required.\n  ```sh\n  kubectl create secret docker-registry apimaticimagesecret
          --docker-server=https://index.docker.io/v1/ --docker-username=<your-name>
          --docker-password=<your-pword> --docker-email=<your-email>\n  ```\n- Create
          a [configmap](https://kubernetes.io/docs/concepts/configuration/configmap/)
          resource named `apimaticlicense` that will serve as the volume storing the
          APIMatic license. An example of this is given below which will create the
          ConfigMap using the *License.lic* file located at */usr/local/apimatic/license/*\n
          \ ```sh\n  kubectl create configmap apimaticlicense --from-file /usr/local/apimatic/license/License.lic\n
          \ ```\n- This will deploy a ConfigMap resource with the following definition:\n
          \ ```sh\n  apiVersion: v1\n  data:\n    License.lic: \\\"<License file contents
          here>\\\"\n  kind: ConfigMap\n  metadata:\n    name: apimaticlicense\n    namespace:
          default\n  ```\n- Now run the sample using the following command:\n  ```sh\n
          \ kubectl apply -f config/samples/apicodegen_v1beta1_apimatic.yaml\n  ```\n\n-
          You will now see a new [Deployment](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/)
          with replica count of 3 and [Service](https://kubernetes.io/docs/concepts/services-networking/service/)
          of type `NodePort` created, both named ***apimatic-sample***. Accessing
          http://localhost:32000 from your browser (or using curl from within the
          Minikube or Kind cluster, if using that) should now show the APIMatic Titan
          page.\n- You can now use the exposed APIMatic CodeGen API to generate API
          SDKs and Docs using [curl](https://curl.se/), [Postman](https://www.postman.com/)
          or your own custom Web application that consumes the APIMatic CodeGen API
          service.\n- Once done, you can remove the APIMatic operator resources using
          the follow script:\n  ```sh\n  kubectl delete -f olm/manifests.yaml\n  ```\n\n##
          Technical Support\n\n- To request additional features in the future, or
          if you notice any discrepancy regarding this document, please drop an email
          to [support@apimatic.io](mailto:support@apimatic.io).\n\n### Copyrights\n\n&copy;
          2021 APIMatic.io\n"
        displayName: APIMatic Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Rest APIs
        - API SDK Generator
        - API Docs Generator
        - API Developer Portal
        - APIMatic
        - Developer Tools
        links:
        - name: Overview
          url: https://www.apimatic.io/
        - name: Developer Experience Portal
          url: https://www.apimatic.io/developer-experience-portal
        - name: Code-Gen as a Service
          url: https://www.apimatic.io/continuous-code-generation
        - name: API Transformer
          url: https://www.apimatic.io/transformer
        - name: Contact
          url: https://www.apimatic.io/contact
        - name: Documentation
          url: https://docs.apimatic.io/
        - name: Blogs
          url: https://www.apimatic.io/blog/
        maintainers:
        - email: support@apimatic.io
          name: Support
        maturity: alpha
        provider:
          name: APIMatic.io
          url: https://www.apimatic.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - docker.io/apimaticio/apimatic-kubernetes-operator:v0.0.1
        version: 0.0.1
      entries:
      - name: apimatic-kubernetes-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: apimatic-kubernetes-operator
    provider:
      name: APIMatic.io
      url: https://www.apimatic.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-openshift-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.14
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-06-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:154f85cc6721a70909dee755ddd88897ad51038afe2538b6dee4d26e7d8b9afb
        version: 0.0.14
      entries:
      - name: joget-dx8-openshift-operator.v0.0.14
        version: 0.0.14
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-openshift-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sumo Logic
      provider-url: https://www.sumologic.com/
    name: sumologic-kubernetes-collection-helm-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "helm-operator.sumologic.com/v1alpha1",
                    "kind": "SumologicCollection",
                    "metadata": {
                        "name": "collection"
                    },
                    "spec": {
                        "sumologic": {
                            "accessId": "",
                            "accessKey": "",
                            "clusterName": "kubernetes",
                            "scc": {
                                "create": true
                            }
                        },
                        "kube-prometheus-stack": {
                            "prometheus-node-exporter": {
                                "service": {
                                    "port": 9200,
                                    "targetPort": 9200
                                }
                            }
                        },
                        "otellogs": {
                            "daemonset": {
                                "containers": {
                                    "otelcol": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                },
                                "initContainers": {
                                    "changeowner": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                }
                            }
                        }
                    }
                }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing,Monitoring
          containerImage: registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
          createdAt: "2024-07-19T10:48:00Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/sumologic-kubernetes-collection-helm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/sumologic-kubernetes-collection-helm-operator-rhmp/support?utm_source=openshift_console
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator.git
          support: Sumo Logic
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configuration for Sumo Logic Kubernetes Collection
            displayName: SumologicCollection
            kind: SumologicCollection
            name: sumologiccollections.helm-operator.sumologic.com
            version: v1alpha1
        description: Helm Operator for the Sumo Logic Kubernetes Collection Helm Chart
          v4.13.0-2
        displayName: Sumo Logic Kubernetes Collection Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - collection
        - logging
        - metrics
        - monitoring
        links:
        - name: Sumo Logic Kubernetes Collection Helm Operator
          url: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator
        maintainers:
        - email: collection@sumologic.com
          name: Sumo Logic
        maturity: alpha
        provider:
          name: Sumo Logic
          url: https://www.sumologic.com/
        relatedImages:
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
        - registry.connect.redhat.com/sumologic/kube-state-metrics@sha256:3551ededa327a35db798ec8286cd843580e3ae395d2231e041e8c4b7621127ee
        - registry.connect.redhat.com/sumologic/metrics-server@sha256:65d2a3ddc4b4183f438fa18b655079c93c905cb9e3b8c54977336205dc1fff07
        - registry.connect.redhat.com/sumologic/prometheus@sha256:0ef6eaf570a40499dd5de92a79593e950bc771b014e0ee6ff7419b4b82a36fcc
        - registry.connect.redhat.com/sumologic/autoinstrumentation-python@sha256:f0fa45fa45e64c7df88f21181ae18ee96fe83cd3988258663d0806ea6f39e6bb
        - registry.connect.redhat.com/sumologic/kubernetes-setup@sha256:f07eaf68fa11ffc708e4485a83bd3763fc5712f7d73d29edfe5b9c9e08f05df6
        - registry.connect.redhat.com/sumologic/sumologic-otel-collector@sha256:4554d43dba204e3ec8e91b36442a438f10234fa0cb4db7fb7880b257a6485112
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:e1f49fcf5b7e86d9eb49c16dbbf9484cb486f55f999fd51d7c738c3b1f2666cf
        - registry.connect.redhat.com/sumologic/autoinstrumentation-dotnet@sha256:f4f385c4489c74b5c43f2fe6b6620a9eb8d9ecab470567236b3c45f636106640
        - registry.connect.redhat.com/sumologic/tailing-sidecar@sha256:738d653a78dc415243ab3627db9135768175fb7f51d0803566f5d9e824c8744b
        - registry.connect.redhat.com/sumologic/telegraf-operator@sha256:ff958222be01c0c3966474463e43bbd5bcf348790dd34bb761718cb1cea2d793
        - registry.connect.redhat.com/sumologic/autoinstrumentation-nodejs@sha256:c206eef4fdecaa58b4f01e5e7f0dbd74e9606eab2b4debf47461d36191e248d1
        - registry.connect.redhat.com/sumologic/node-exporter@sha256:11e9e604432a66d9b388cb1bf70eac077c7f0372d1264342ef927d76b8ef0272
        - registry.connect.redhat.com/sumologic/kubernetes-tools-kubectl@sha256:d3155b864cc74f327adedf7d013a8f0fcb11c2e73a5655b7cafa848824e80172
        - registry.connect.redhat.com/sumologic/thanos@sha256:16faf7191668e86d8b576bba055dd0cbdeb009635976b9e07bd00f926032f11c
        - registry.connect.redhat.com/sumologic/opentelemetry-operator@sha256:5587fe5b7ebfa57ad320ad2521103f908317388d1741c40e5f7be5ed9e5f9ad7
        - registry.connect.redhat.com/sumologic/nginx-unprivileged@sha256:7260555d16c94c4fe1e016626b4cb8a5c0de6b700e2eea28744331fb298f71ae
        - registry.connect.redhat.com/sumologic/prometheus-operator@sha256:58a329e6582456b32abebb470086191f8366b927af4703741613993465f116b8
        - registry.connect.redhat.com/sumologic/telegraf@sha256:75efd2e509c6017ea0fc5c9103ce068cbc89087b16506a4ef48cb5ff00eab84e
        - registry.connect.redhat.com/sumologic/tailing-sidecar-operator@sha256:c16b8b18e9b263fc3eb24059bba512bcfc3201b32105af6e5518392826ddd66c
        - registry.connect.redhat.com/sumologic/prometheus-config-reloader@sha256:0b605d36782afd8f50fd631d0e9c7b19d39ce962af89b7a737b2f09a8ceff291
        - registry.connect.redhat.com/sumologic/autoinstrumentation-java@sha256:242958b7d8905f4ac02f6ad38897ba58b9badf2a6213a389e67dcc71fee3630c
        - registry.connect.redhat.com/sumologic/busybox@sha256:209bc8e0b1d958fc699f3bb75b1248544e372ad35cdc46c991bd1698f641e1f9
        version: 4.13.0-2
      entries:
      - name: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
        version: 4.13.0-2
      name: alpha
    defaultChannel: alpha
    packageName: sumologic-kubernetes-collection-helm-operator-rhmp
    provider:
      name: Sumo Logic
      url: https://www.sumologic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-bamoe-businessautomation: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: bamoe-businessautomation-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bamoe-businessautomation-operator.8.0.8-1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.kiegroup.org/v2","kind":"KieApp","metadata":{"name":"rhpam-trial"},"spec":{"environment":"rhpam-trial"}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/ibm-bamoe/bamoe-rhel8-operator@sha256:0646da8707d10b3cc2e2344269abd8797db34127a3326bf76bbc1547aa4bbd73
          createdAt: "2025-07-22 17:40:17"
          description: Deploys and manages IBM Business Automation Manager Open Editions
            environment.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM Business Automation Manager
            Open Edition"]'
          repository: https://github.com/kiegroup/kie-cloud-operator
          support: https://www.ibm.com/mysupport
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A project prescription running an IBM BAMOE environment.
            displayName: KieApp
            kind: KieApp
            name: kieapps.app.kiegroup.org
            version: v2
        description: |-
          Deploys and manages IBM Business Automation Manager Open Editions environment.

          * **IBM Process Automation Manager** is a platform for developing containerized microservices and applications that automate business decisions and processes. It includes business process management (BPM), business rules management (BRM), and business resource optimization and complex event processing (CEP) technologies. It also includes a user experience platform to create engaging user interfaces for process and decision services with minimal coding.
        displayName: IBM Business Automation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kieapp
        - pam
        - decision
        - kie
        - cloud
        - bpm
        - process
        - automation
        - operator
        links:
        - name: Product Page
          url: https://ibm.com
        - name: Documentation
          url: https://www.ibm.com/docs/en/ibamoe
        maintainers:
        - email: mysphelp@us.ibm.com
          name: IBM
        maturity: 8.x-stable
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-monitoring-rhel8@sha256:ef1d77bb085172f927ee53d1674c85e2acaf1f580d7cad55bae8bf2c33c49c01
        - registry.redhat.io/openshift4/ose-cli@sha256:b2b93def4eb5edeccfa2d8ba22e493646325842674dff0446b45677bf641a342
        - registry.redhat.io/rhscl/postgresql-13-rhel7@sha256:338dd09681dcf860e6a976a3bc48a1474c2376d698815e4eb611ae9c5aa68966
        - registry.redhat.io/ibm-bamoe/bamoe-process-migration-rhel8@sha256:3fd45f2dabca7220772d823275a5697b42dd35f53c27671b093117933ef2d4f6
        - registry.redhat.io/rhscl/mysql-80-rhel7@sha256:e4f535c78bc2fdb58914c52fb805185035ff31515452b4df714f9ab623ba383a
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-rhel8@sha256:c3553707936cb9eee8a6a622643f72c91e1b0c11add5ec9ad5865242a9e133b4
        - registry.redhat.io/ibm-bamoe/bamoe-kieserver-rhel8@sha256:01bf97beffb3647968aae052cdd401d6d562c35f3e78ac483dbf55c44a71dc2f
        - registry.redhat.io/jboss-datagrid-7/datagrid73-openshift@sha256:4056b1cf8ff5995f56eecfdc087d9ece56a216dba6bd73acc5693f0778e8d573
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-rhel8@sha256:82d8ebde7eb3ecfbf72e5b71e7e392d3a3597fc084b6cf35e8fec23bfe41210d
        - registry.redhat.io/amq7/amq-broker@sha256:792c1cb19e62519cc47d834650fe5a6a6ca589a37d0b53d4f4549e80cabae3f9
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:98e9a6618558bc8a1b9cb91d1c9b3380968d335092f6e953d1abe6af96abf65c
        - registry.redhat.io/ibm-bamoe/bamoe-smartrouter-rhel8@sha256:2f1a7519617b5c0d082f943de0db295370649d968d699d76cfd829f4c9826a0f
        - registry.redhat.io/ibm-bamoe/bamoe-rhel8-operator@sha256:0646da8707d10b3cc2e2344269abd8797db34127a3326bf76bbc1547aa4bbd73
        - registry.redhat.io/ibm-bamoe/bamoe-process-migration-rhel8@sha256:a287a2fbaded371b8fcae5e4124b2d905bb228f94da0fc899ca2600b378dca0d
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-monitoring-rhel8@sha256:25025e489c15108edffe5270d257e1ed9be0a67aa067bcfc184647793aac2617
        - registry.redhat.io/ibm-bamoe/bamoe-dashbuilder-rhel8@sha256:efa55e17190b0ebe1618452f786876ef354b8fbec25c316aed7ed5bb1ce88116
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:457fc73423fa94591bc7e35729a25c58d3465583afccf342fb338b6e11188575
        - registry.redhat.io/ibm-bamoe/bamoe-smartrouter-rhel8@sha256:5ed8821328413ec11859ae317679f0692afff9945f134fc8909e07a3ac387794
        - registry.redhat.io/ibm-bamoe/bamoe-kieserver-rhel8@sha256:c9f08ddc6111cdd74faa0db5721d5730fc7716fccf4dc2a658595f7e0c184c10
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:f714b9c1555e7f27a3c9cf2cf687a74188f36b1283d4e09cec4db3c084d9fe64
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhscl/postgresql-12-rhel7@sha256:191237ccffc476e764714a69e35d9421dcee5ddd7184442d7a8306a06b97e6bb
        - registry.redhat.io/ibm-bamoe/bamoe-dashbuilder-rhel8@sha256:b3b7388b3b116ef37aabbb04b5d2c3e36e58d2fc95f1e59795b423010e071fc3
        - registry.redhat.io/ibm-bamoe/bamoe-controller-rhel8@sha256:3fcd9ae0bc194c198cbb5ee94272e80f302658733e467766bd1789357e9fd015
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:fb326be806b337b0ae1711f974d15e073aebcbad77aaa1123a63cf4f331da5f2
        - registry.redhat.io/ibm-bamoe/bamoe-controller-rhel8@sha256:6cd0f07856d612dca46b0a01a93b796f7a0f112b98cedfd6754f9fd770ee843d
        version: 8.0.8-1
      entries:
      - name: bamoe-businessautomation-operator.8.0.8-1
        version: 8.0.8-1
      - name: bamoe-businessautomation-operator.8.0.7-2
        version: 8.0.7-2
      - name: bamoe-businessautomation-operator.8.0.7-1
        version: 8.0.7-1
      - name: bamoe-businessautomation-operator.8.0.6-3
        version: 8.0.6-3
      - name: bamoe-businessautomation-operator.8.0.6-2
        version: 8.0.6-2
      - name: bamoe-businessautomation-operator.8.0.6-1
        version: 8.0.6-1
      - name: bamoe-businessautomation-operator.8.0.5-3
        version: 8.0.5-3
      - name: bamoe-businessautomation-operator.8.0.5-2
        version: 8.0.5-2
      - name: bamoe-businessautomation-operator.8.0.5-1
        version: 8.0.5-1
      - name: bamoe-businessautomation-operator.8.0.4-3
        version: 8.0.4-3
      - name: bamoe-businessautomation-operator.8.0.4-2
        version: 8.0.4-2
      - name: bamoe-businessautomation-operator.8.0.4-1
        version: 8.0.4-1
      - name: bamoe-businessautomation-operator.8.0.3-4
        version: 8.0.3-4
      - name: bamoe-businessautomation-operator.8.0.3-3
        version: 8.0.3-3
      - name: bamoe-businessautomation-operator.8.0.3-2
        version: 8.0.3-2
      - name: bamoe-businessautomation-operator.8.0.3-1
        version: 8.0.3-1
      - name: bamoe-businessautomation-operator.8.0.2-3
        version: 8.0.2-3
      - name: bamoe-businessautomation-operator.8.0.2-2
        version: 8.0.2-2
      - name: bamoe-businessautomation-operator.8.0.2-1
        version: 8.0.2-1
      - name: bamoe-businessautomation-operator.8.0.1-2
        version: 8.0.1-2
      - name: bamoe-businessautomation-operator.8.0.1-1
        version: 8.0.1-1
      - name: bamoe-businessautomation-operator.8.0.0-1
        version: 8.0.0-1
      name: 8.x-stable
    defaultChannel: 8.x-stable
    packageName: bamoe-businessautomation-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Odoo Community Association
      provider-url: ""
    name: odoo-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odoo-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"odoo-community.org/v1alpha1","kind":"Odoo","metadata":{"name":"example"},"spec":{"postgresqlVolumeSize":"4Gi","odooVolumeSize":"4Gi"}}]'
          capabilities: Full Lifecycle
          categories: Application Runtime
          certified: "false"
          containerImage: ursa/odoo-operator:0.0.2
          createdAt: "2021-04-26T12:00:00Z"
          description: Operator to manage Odoo instances within an OpenShift cluster
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          repository: https://github.com/ursais/odoo-operator
          support: Open Source Integrators
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Odoo
            displayName: Odoo
            kind: Odoo
            name: odoos.odoo-community.org
            version: v1alpha1
        description: |
          ## About Odoo

          [Odoo](https://odoo-community.org), is a suite of business management software tools including CRM, e-commerce, billing, accounting, manufacturing, warehouse, project management, and inventory management to name a few.
          ## About the Odoo Operator

          This is an Ansible-based operator created with operator-sdk v1.5.0.

          The upgrade is handled by the Odoo container itself using [Marabunta](https://pypi.org/project/marabunta/) and [Anthem](https://pypi.org/project/anthem/).

          ## Roadmap
          * Add horizontal pod autoscaler
          * Add cronjob for backup
          ## Prerequisites for enabling this Operator

          Please refer to the README file on the [GitHub project](https://github.com/ursais/odoo-operator).
        displayName: Odoo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ERP
        - Business Software
        links:
        - name: Odoo Community Association
          url: https://odoo-community.org
        - name: GitHub Repository
          url: https://github.com/ursais/odoo-operator
        - name: Open Source Integrators
          url: https://opensourceintegrators.com
        maintainers:
        - email: mchambreuil@opensourceintegrators.com
          name: Maxime Chambreuil
        maturity: alpha
        provider:
          name: Odoo Community Association
        relatedImages:
        - ursa/odoo-operator:0.0.2
        version: 0.0.2
      entries:
      - name: odoo-operator.v0.0.2
        version: 0.0.2
      - name: odoo-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: odoo-operator
    provider:
      name: Odoo Community Association
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fortanix Inc.
      provider-url: https://fortanix.com
    name: ccm-node-agent-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ccm-node-agent-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1",
                "kind": "CcmNodeAgent",
                "metadata": {
                  "name": "ccmnodeagent-sample"
                },
                "spec": {
                  "ccmJoinToken": {
                    "secretName": "ccm-join-token",
                    "token": ""
                  },
                  "fullnameOverride": "",
                  "image": {
                    "ccmNodeAgent": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/ccm-node-agent-ubi7",
                      "tag": "latest"
                    },
                    "ccmNodeAgentInit": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "busybox",
                      "tag": "latest"
                    },
                    "sgxSoftwareInstaller": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/sgx-software-installer-ubi7",
                      "tag": "latest"
                    }
                  },
                  "log": {
                    "debug": false
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "ports": {
                    "containerPort": 9092,
                    "hostPort": 9092
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CCM Node Agent software enables registration of the compute
              nodes to Fortanix CCM when installed in the Openshift Cluster Nodes<U+200B>.
            displayName: CCM Node Agent
            kind: CcmNodeAgent
            name: ccmnodeagents.charts.operatorhub.io
            version: v1
        description: "CCM Node Agent software enables registration of the compute
          nodes to Fortanix CCM when installed in the Openshift Cluster Nodes​. \nIt
          also assists with application attestation and visibility for Fortanix CCM.\n\n##
          About this Operator\nThis Operator is based on a Helm chart for CCM Node
          Agent. It enrolls all the nodes in the openshift cluster to a particular
          CCM account. You have to provide the base64 encode of join token of CCM
          account while creating CcmNodeAgent resource.\n\n## Prerequisites\nAll the
          worker nodes in the cluster must have SGX capabilities.\n\n## Installation\n*
          Choose a namespace to install the Operator. Operator will be available in
          all namespaces.\n* Select Approval Strategy: Available options are Automatic
          or Manual. Automatic approval will automatically upgrade the running instance
          of your Operator without human intervention when a new version is available.
          In case of Manual approval, when a new version is available, OLM will create
          an update request. You must manually approve the request to update the Operator
          to a new version.\n* Generate your CCM Account Join Token: Please log in
          to https://ccm.fortanix.com, and in the Management Console tab, click the
          + ENROLL NODE button. In the ENROLL NODE screen, a Join Token will be generated
          in the text box for \"Get a join token to register an SGX compute node\".
          This Join Token is used by the compute node to authenticate itself.\n* Base64
          encode the Join Token:\n```shell\necho -n <join-token> | openssl base64
          -A\n```\n* Create an instance of CcmNodeAgent by setting the `spec.ccmJoinToken.Token`
          to the base64 encoded CCM account join token.\n* After CcmNodeAgent resource
          is installed successfully you should be able to see all the worker nodes
          of the cluster in your CCM account under Compute Node Tab.\n\n## Getting
          Started\nNow to run your application inside SGX enclaves on your cluster
          follow the [documentation](https://support.fortanix.com/hc/en-us/articles/360043527431-User-s-Guide-Add-and-Edit-an-Application).\n"
        displayName: CCM Node Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dockerfile
        - Helm
        - CCM
        - Confidential Compute
        links:
        - name: CCM Node Agent Operator
          url: https://support.fortanix.com/hc/en-us/categories/360003107511-Confidential-Computing-Manager
        maintainers:
        - email: info@fortanix.com
          name: Fortanix Inc.
        maturity: alpha
        provider:
          name: Fortanix Inc.
          url: https://fortanix.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62575404000cb77fa8d2284c87b31e39e667a55b0fb47821263b23b3d1135a7e
        - docker.io/fortanix/ccm-node-agent-operator@sha256:e2a3d52551198102ddbe7f4ebd9621edc357c57ddeff47b7ed9f2cf56a1bed39
        version: 0.0.2
      entries:
      - name: ccm-node-agent-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ccm-node-agent-operator
    provider:
      name: Fortanix Inc.
      url: https://fortanix.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: https://intel.com
    name: intel-device-plugins-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: intel-device-plugins-operator.v0.29.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "GpuDevicePlugin",
                "metadata": {
                  "name": "gpudeviceplugin-sample"
                },
                "spec": {
                  "enableMonitoring": true,
                  "image": "registry.connect.redhat.com/intel/intel-gpu-plugin@sha256:963caf1e6e16b5e297671c536eec04dc57cf45f5cc5cf458bdcafa56bdb97972",
                  "preferredAllocationPolicy": "none",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/gpu": "true"
                  },
                  "sharedDevNum": 1
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "QatDevicePlugin",
                "metadata": {
                  "name": "qatdeviceplugin-sample"
                },
                "spec": {
                  "dpdkDriver": "vfio-pci",
                  "image": "registry.connect.redhat.com/intel/intel-qat-plugin@sha256:53617c3a72af169031cbaba1380cd6a25c9d2b50c9a2da822cf37b473c6e7a8f",
                  "initImage": "registry.connect.redhat.com/intel/intel-qat-initcontainer@sha256:14f282a0adae43262657f24ce99a620b671c9a2ce12a8216163e6b22ac6e252b",
                  "kernelVfDrivers": [
                    "c6xxvf",
                    "4xxxvf"
                  ],
                  "logLevel": 4,
                  "maxNumDevices": 128,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/qat": "true"
                  }
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "DsaDevicePlugin",
                "metadata": {
                  "name": "dsadeviceplugin-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/intel/intel-dsa-plugin@sha256:092575515739bf9a0f5982c3605b63382bc4d1250128269434c2bfe4a27ec99d",
                  "initImage": "registry.connect.redhat.com/intel/intel-idxd-config-initcontainer@sha256:649e4f4e7f362ba054fb85a67e3baa23ec9649c68df356d1e33721ca3eef2f6b",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/dsa": "true"
                  },
                  "sharedDevNum": 10
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "SgxDevicePlugin",
                "metadata": {
                  "name": "sgxdeviceplugin-sample"
                },
                "spec": {
                  "enclaveLimit": 110,
                  "image": "registry.connect.redhat.com/intel/intel-sgx-plugin@sha256:e54cdf61feb0b67e330ad24d92e7fa4000953c2ec4d4300d052e032d419d5e0a",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/sgx": "true"
                  },
                  "provisionLimit": 110
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Security, AI/Machine Learning
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/intel/intel-technology-enabling-for-openshift
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GpuDevicePlugin is the Schema for the gpudeviceplugins API.
              It represents the GPU device plugin responsible for advertising Intel
              GPU hardware resources to the kubelet.
            displayName: Intel GPU Device Plugin
            kind: GpuDevicePlugin
            name: gpudeviceplugins.deviceplugin.intel.com
            version: v1
          - description: DsaDevicePlugin is the Schema for the dsadeviceplugins API.
              It represents the DSA device plugin responsible for advertising Intel
              DSA hardware resources to the kubelet.
            displayName: Intel DSA Device Plugin
            kind: DsaDevicePlugin
            name: dsadeviceplugins.deviceplugin.intel.com
            version: v1
          - description: QatDevicePlugin is the Schema for the qatdeviceplugins API.
              It represents the QAT device plugin responsible for advertising Intel
              QuickAssist Technology hardware resources to the kubelet.
            displayName: Intel QuickAssist Technology Device Plugin
            kind: QatDevicePlugin
            name: qatdeviceplugins.deviceplugin.intel.com
            version: v1
          - description: SgxDevicePlugin is the Schema for the sgxdeviceplugins API.
              It represents the SGX device plugin responsible for advertising SGX
              device nodes to the kubelet.
            displayName: Intel Software Guard Extensions Device Plugin
            kind: SgxDevicePlugin
            name: sgxdeviceplugins.deviceplugin.intel.com
            version: v1
        description: |
          [Intel Device Plugins for Kubernetes](https://github.com/intel/intel-technology-enabling-for-openshift) is a collection of
          [device plugins](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/) advertising Intel specific hardware resources
          to the kubelet. Currently the operator has basic support for Intel SGX, Intel GPU, Intel QAT and Intel DSA. Intel IAA device plugin will be supported in a future release.
        displayName: Intel Device Plugins Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - qat
        - sgx
        - gpu
        - dsa
        - dlb
        links:
        - name: Intel Device Plugins Operator
          url: https://github.com/intel/intel-technology-enabling-for-openshift
        maintainers:
        - email: martin.xu@intel.com
          name: Martin Xu
        - email: chaitanya.kulkarni@intel.com
          name: Chaitanya Kulkarni
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Intel® Corporation
          url: https://intel.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - registry.connect.redhat.com/intel/intel-sgx-plugin@sha256:e54cdf61feb0b67e330ad24d92e7fa4000953c2ec4d4300d052e032d419d5e0a
        - registry.connect.redhat.com/intel/intel-gpu-plugin@sha256:963caf1e6e16b5e297671c536eec04dc57cf45f5cc5cf458bdcafa56bdb97972
        - registry.connect.redhat.com/intel/intel-qat-plugin@sha256:53617c3a72af169031cbaba1380cd6a25c9d2b50c9a2da822cf37b473c6e7a8f
        - registry.connect.redhat.com/intel/intel-qat-initcontainer@sha256:14f282a0adae43262657f24ce99a620b671c9a2ce12a8216163e6b22ac6e252b
        - registry.connect.redhat.com/intel/intel-dsa-plugin@sha256:092575515739bf9a0f5982c3605b63382bc4d1250128269434c2bfe4a27ec99d
        - registry.connect.redhat.com/intel/intel-idxd-initcontainer@sha256:649e4f4e7f362ba054fb85a67e3baa23ec9649c68df356d1e33721ca3eef2f6b
        - registry.connect.redhat.com/intel/intel-deviceplugin-operator@sha256:283e8b60d7ab8560f8ec8875b905e6446b59a91dd5e0d97da72e3a9aab235cc4
        version: 0.29.1
      entries:
      - name: intel-device-plugins-operator.v0.29.1
        version: 0.29.1
      - name: intel-device-plugins-operator.v0.29.0
        version: 0.29.0
      - name: intel-device-plugins-operator.v0.28.0
        version: 0.28.0
      name: alpha
    defaultChannel: alpha
    packageName: intel-device-plugins-operator
    provider:
      name: Intel® Corporation
      url: https://intel.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: designate-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: designate-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "designate.openstack.org/v1beta1",
                "kind": "Designate",
                "metadata": {
                  "name": "designate"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "designate",
                  "databaseInstance": "openstack",
                  "designateAPI": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate"
                  },
                  "designateBackendbind9": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "networkAttachments": [
                      "designate"
                    ],
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate",
                    "storageClass": "local-storage",
                    "storageRequest": "1G"
                  },
                  "designateCentral": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate"
                  },
                  "designateMdns": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "networkAttachments": [
                      "designate"
                    ],
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate"
                  },
                  "designateProducer": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate"
                  },
                  "designateUnbound": {
                    "replicas": 1
                  },
                  "designateWorker": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "networkAttachments": [
                      "designate"
                    ],
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate"
                  },
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "osp-secret",
                  "serviceUser": "designate"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:48Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DesignateAPI is the Schema for the designateapis API
            displayName: Designate API
            kind: DesignateAPI
            name: designateapis.designate.openstack.org
            version: v1beta1
          - kind: DesignateBackendbind9
            name: designatebackendbind9s.designate.openstack.org
            version: v1beta1
          - description: DesignateCentral is the Schema for the designatecentral API
            displayName: Designate Central
            kind: DesignateCentral
            name: designatecentrals.designate.openstack.org
            version: v1beta1
          - description: DesignateMdns is the Schema for the designatemdnses API
            displayName: Designate Mdns
            kind: DesignateMdns
            name: designatemdnses.designate.openstack.org
            version: v1beta1
          - description: DesignateProducer is the Schema for the designateproducer
              API
            displayName: Designate Producer
            kind: DesignateProducer
            name: designateproducers.designate.openstack.org
            version: v1beta1
          - description: Designate is the Schema for the designates API
            displayName: Designate
            kind: Designate
            name: designates.designate.openstack.org
            version: v1beta1
          - description: DesignateUnbound is the Schema for the designateunbound API
            displayName: Designate Unbound
            kind: DesignateUnbound
            name: designateunbounds.designate.openstack.org
            version: v1beta1
          - description: DesignateWorker is the Schema for the designateworker API
            displayName: Designate Worker
            kind: DesignateWorker
            name: designateworkers.designate.openstack.org
            version: v1beta1
        description: Designate Operator
        displayName: Designate Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Network
        - Designate
        links:
        - name: Designate Operator
          url: https://github.com/openstack-k8s-operators/designate-operator
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso/openstack-designate-api-rhel9@sha256:1a9baa4a0f99029eb8a367ddcce264581aa37fdf1834ba277c30a3f0ae33f400
        - registry.redhat.io/rhoso/openstack-designate-central-rhel9@sha256:ad610cdc29b31cd82a81cea22793176351574f43c0f7705273c8ffc03ad2fdac
        - registry.redhat.io/rhoso/openstack-designate-producer-rhel9@sha256:4beb652a073a65965f96f125d0fcb454255efa5e53a84bd0dd9120b57e59fcf3
        - registry.redhat.io/rhoso-operators/designate-rhel9-operator@sha256:e6c1ba511cfd0278fc8cd2836c6324d97ae3f6e9affc992600978062fa2d4549
        - registry.redhat.io/rhoso/openstack-designate-mdns-rhel9@sha256:a4109755360f340aadf6bc15d24599f882d82bd59739065f87275f9274d9eeb9
        - registry.redhat.io/rhoso/openstack-designate-worker-rhel9@sha256:94ef291ca85b06dffc93d149b09b412d9190c99f7a1b0b300d50bd70346401ad
        - registry.redhat.io/rhoso/openstack-designate-backend-bind9-rhel9@sha256:89ce30fc1eff339867ce422203e6e0c63296667e102ce85d823d44c30f7556ad
        - registry.redhat.io/rhoso/openstack-unbound-rhel9@sha256:ad95d40227a1e52f739f09a2ce8db647d6a2fa7ef63e58a2444688728fa5bdfd
        version: 1.0.6
      entries:
      - name: designate-operator.v1.0.6
        version: 1.0.6
      - name: designate-operator.v1.0.5
        version: 1.0.5
      - name: designate-operator.v1.0.4
        version: 1.0.4
      - name: designate-operator.v1.0.3
        version: 1.0.3
      - name: designate-operator.v1.0.2
        version: 1.0.2
      - name: designate-operator.v1.0.1
        version: 1.0.1
      - name: designate-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: designate-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-cert-manager-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-manager-operator.v1.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
          createdAt: 2025-08-06T08:21:40
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.16.1 <1.17.0'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.17.4](https://github.com/cert-manager/cert-manager/tree/v1.17.4), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:96d51e3a64bf30cbd92836c7cbd82f06edca16eef78ab1432757d34c16628659
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b
        version: 1.17.0
      entries:
      - name: cert-manager-operator.v1.17.0
        version: 1.17.0
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      - name: cert-manager-operator.v1.14.1
        version: 1.14.1
      - name: cert-manager-operator.v1.14.0
        version: 1.14.0
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1
    - currentCSV: cert-manager-operator.v1.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "\u003cACCESS_KEY_ID\u003e",
                            "hostedZoneID": "\u003cHOSTED_ZONE_ID\u003e",
                            "region": "\u003cAWS_REGION\u003e",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4e854c14fecbd27107b91b76164ce2e1575ce60202d835abba5962fb5d386560
          createdAt: 2024-04-30T13:29:33
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.13.0 <1.13.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/infrastructure-features: '["proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.13.6](https://github.com/cert-manager/cert-manager/tree/v1.13.6), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:afc82895c09d75e3d6a932aea2d3e3ef998f765306da92070de1dab26baf0593
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4e854c14fecbd27107b91b76164ce2e1575ce60202d835abba5962fb5d386560
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:578f74c3e96db7fc2f4e94dd9fbc7d6499d87b0bc55576930ee17de08a95022c
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:7915f8049d2451b425e3674eb60e596b9b11946ffda1497f5d5607799e2bf431
        version: 1.13.1
      entries:
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1.13
    - currentCSV: cert-manager-operator.v1.14.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:5c593bfd70309f87cd05716e233e65aa55394eafcdaf85f178698912cb29d79d
          createdAt: 2025-04-09T05:19:10
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.14.1 <1.14.2'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/infrastructure-features: '["proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.14.7](https://github.com/cert-manager/cert-manager/tree/v1.14.7), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:5c593bfd70309f87cd05716e233e65aa55394eafcdaf85f178698912cb29d79d
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:5df082f8e3787e6ac2edf395113806999ae192425ce18b4ddba9ab20c135fa5b
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:49fb8f710319d6f3479b98d2c5dd7d0c9df8c1678b2c00727314dea1a8e933a1
        version: 1.14.2
      entries:
      - name: cert-manager-operator.v1.14.2
        version: 1.14.2
      - name: cert-manager-operator.v1.14.1
        version: 1.14.1
      - name: cert-manager-operator.v1.14.0
        version: 1.14.0
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1.14
    - currentCSV: cert-manager-operator.v1.15.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:f328263e2d29e34ede65e4501f0447b2d9f84e9445a365c2fa2fbb253939e274
          createdAt: 2025-03-13T18:26:37
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.15.0 <1.15.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/infrastructure-features: '["proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.15.5](https://github.com/cert-manager/cert-manager/tree/v1.15.2), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:f328263e2d29e34ede65e4501f0447b2d9f84e9445a365c2fa2fbb253939e274
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9573d74bd2b926ec94af76f813e6358f14c5b2f4e0eedab7c1ff1070b7279a5c
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:6af3ee8b2a5a87042fb7158bda8d6cf2e6324d1e265974acf77214d4cd0ea0d3
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:20efff60a0caf5eafb38986fd21611697b5bc534c2e789da233983a9739938ed
        version: 1.15.1
      entries:
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      - name: cert-manager-operator.v1.14.1
        version: 1.14.1
      - name: cert-manager-operator.v1.14.0
        version: 1.14.0
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1.15
    - currentCSV: cert-manager-operator.v1.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:d30e755a266bb035d50e2b257e7a26c1fe6a5ccb66125576db99d537aa03ec27
          createdAt: 2025-07-10T08:52:43
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.15.1 <1.16.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.16.5](https://github.com/cert-manager/cert-manager/tree/v1.16.5), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:831e4b1e5aa6eed6fbb22f779ddbcb9c73e4bf6cc079d0c6652737a6a44e6320
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:d30e755a266bb035d50e2b257e7a26c1fe6a5ccb66125576db99d537aa03ec27
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:33308217d2226c2555eb18020710d16175295a8e349263a5435e4b158bbc735f
        version: 1.16.1
      entries:
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      - name: cert-manager-operator.v1.14.1
        version: 1.14.1
      - name: cert-manager-operator.v1.14.0
        version: 1.14.0
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1.16
    - currentCSV: cert-manager-operator.v1.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
          createdAt: 2025-08-06T08:21:40
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.16.1 <1.17.0'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.17.4](https://github.com/cert-manager/cert-manager/tree/v1.17.4), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:96d51e3a64bf30cbd92836c7cbd82f06edca16eef78ab1432757d34c16628659
        version: 1.17.0
      entries:
      - name: cert-manager-operator.v1.17.0
        version: 1.17.0
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      - name: cert-manager-operator.v1.14.1
        version: 1.14.1
      - name: cert-manager-operator.v1.14.0
        version: 1.14.0
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1.17
    defaultChannel: stable-v1
    packageName: openshift-cert-manager-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ram-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ram-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ram-controller:1.1.1
          createdAt: "2025-09-19T22:18:09Z"
          description: AWS RAM controller is a service controller for managing RAM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Permission represents the state of an AWS ram Permission
              resource.
            displayName: Permission
            kind: Permission
            name: permissions.ram.services.k8s.aws
            version: v1alpha1
          - description: ResourceShare represents the state of an AWS ram ResourceShare
              resource.
            displayName: ResourceShare
            kind: ResourceShare
            name: resourceshares.ram.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon RAM resources in AWS from within your Kubernetes cluster.

          **About Amazon RAM**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon RAM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ram
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon RAM Developer Resources
          url: https://aws.amazon.com/RAM/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ram maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ram-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ram-controller.v1.1.1
        version: 1.1.1
      - name: ack-ram-controller.v1.1.0
        version: 1.1.0
      - name: ack-ram-controller.v1.0.11
        version: 1.0.11
      - name: ack-ram-controller.v1.0.10
        version: 1.0.10
      name: alpha
    defaultChannel: alpha
    packageName: ack-ram-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operator-kogitocloud: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kogito-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kogito-operator.v1.44.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "process-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogito-infinispan-infra"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "infinispan.org/v1",
                    "kind": "Infinispan"
                  }
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/kiegroup/kogito-operator:1.44.1
          createdAt: "2019-08-22T13:12:22Z"
          description: Kogito Operator for deployment and management of Kogito services.
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.app.kiegroup.org
            version: v1beta1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.app.kiegroup.org
            version: v1beta1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.app.kiegroup.org
            version: v1beta1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.app.kiegroup.org
            version: v1beta1
        description: |-
          Kogito Operator is designed to deploy Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Creates two kinds of build configuration: Native or JVM. For more information refer to [Native build and deployment](https://docs.jboss.org/kogito/release/latest/html_single/#_native_build_and_deployment) section on our docs
          * Builds a Kogito service from a remote git repository and creates a custom image for the given runtime: Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Installs the [Kogito Data Index Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-data-index-service_kogito-deploying-on-openshift) to enable data index feature for Kogito services.
          * Installs the [Kogito Jobs Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-jobs-service_kogito-deploying-on-openshift) to enable jobs scheduling.

          For Kubernetes deployment of Kogito services, please refer to the section [Deploying Kogito services on Kubernetes](https://docs.jboss.org/kogito/release/latest/html_single/#proc-kogito-deploying-on-kubernetes_kogito-deploying-on-openshift) on our docs.

          **IMPORTANT!** Kogito Operator might need Infinispan Operator and Strimzi to implement persistence and messaging use cases if you do not provide these services beforehand.
          If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the Kogito Operator.
        displayName: Kogito
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://kogito.kie.org/
        - name: Documentation
          url: https://github.com/kiegroup/kogito-operator/blob/main/README.md
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/kiegroup/kogito-operator:1.44.1
        version: 1.44.1
      entries:
      - name: kogito-operator.v1.44.1
        version: 1.44.1
      - name: kogito-operator.v1.43.0
        version: 1.43.0
      - name: kogito-operator.v1.42.0
        version: 1.42.0
      - name: kogito-operator.v1.41.0
        version: 1.41.0
      - name: kogito-operator.v1.40.0
        version: 1.40.0
      - name: kogito-operator.v1.39.0
        version: 1.39.0
      - name: kogito-operator.v1.38.0
        version: 1.38.0
      - name: kogito-operator.v1.37.0
        version: 1.37.0
      - name: kogito-operator.v1.36.0
        version: 1.36.0
      - name: kogito-operator.v1.35.0
        version: 1.35.0
      - name: kogito-operator.v1.34.0
        version: 1.34.0
      - name: kogito-operator.v1.33.0
        version: 1.33.0
      - name: kogito-operator.v1.32.0
        version: 1.32.0
      - name: kogito-operator.v1.31.0
        version: 1.31.0
      - name: kogito-operator.v1.30.0
        version: 1.30.0
      - name: kogito-operator.v1.29.0
        version: 1.29.0
      - name: kogito-operator.v1.28.0
        version: 1.28.0
      - name: kogito-operator.v1.27.0
        version: 1.27.0
      - name: kogito-operator.v1.26.0
        version: 1.26.0
      - name: kogito-operator.v1.25.0
        version: 1.25.0
      - name: kogito-operator.v1.24.0
        version: 1.24.0
      - name: kogito-operator.v1.23.0
        version: 1.23.0
      - name: kogito-operator.v1.22.1
        version: 1.22.1
      - name: kogito-operator.v1.22.0
        version: 1.22.0
      - name: kogito-operator.v1.21.0
        version: 1.21.0
      - name: kogito-operator.v1.20.0
        version: 1.20.0
      - name: kogito-operator.v1.19.0
        version: 1.19.0
      - name: kogito-operator.v1.18.0
        version: 1.18.0
      - name: kogito-operator.v1.17.0
        version: 1.17.0
      - name: kogito-operator.v1.16.1
        version: 1.16.1
      - name: kogito-operator.v1.16.0
        version: 1.16.0
      - name: kogito-operator.v1.15.0
        version: 1.15.0
      - name: kogito-operator.v1.14.0
        version: 1.14.0
      - name: kogito-operator.v1.13.0
        version: 1.13.0
      - name: kogito-operator.v1.12.0
        version: 1.12.0
      - name: kogito-operator.v1.11.1
        version: 1.11.1
      - name: kogito-operator.v1.11.0
        version: 1.11.0
      - name: kogito-operator.v1.10.0
        version: 1.10.0
      - name: kogito-operator.v1.9.1
        version: 1.9.1
      - name: kogito-operator.v1.9.0
        version: 1.9.0
      - name: kogito-operator.v1.8.0
        version: 1.8.0
      - name: kogito-operator.v1.7.0
        version: 1.7.0
      - name: kogito-operator.v1.6.0
        version: 1.6.0
      - name: kogito-operator.v1.5.0
        version: 1.5.0
      - name: kogito-operator.v1.4.0
        version: 1.4.0
      name: 1.x
    - currentCSV: kogito-operator.v1.44.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "process-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogito-infinispan-infra"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "infinispan.org/v1",
                    "kind": "Infinispan"
                  }
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/kiegroup/kogito-operator:1.44.1
          createdAt: "2019-08-22T13:12:22Z"
          description: Kogito Operator for deployment and management of Kogito services.
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.app.kiegroup.org
            version: v1beta1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.app.kiegroup.org
            version: v1beta1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.app.kiegroup.org
            version: v1beta1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.app.kiegroup.org
            version: v1beta1
        description: |-
          Kogito Operator is designed to deploy Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Creates two kinds of build configuration: Native or JVM. For more information refer to [Native build and deployment](https://docs.jboss.org/kogito/release/latest/html_single/#_native_build_and_deployment) section on our docs
          * Builds a Kogito service from a remote git repository and creates a custom image for the given runtime: Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Installs the [Kogito Data Index Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-data-index-service_kogito-deploying-on-openshift) to enable data index feature for Kogito services.
          * Installs the [Kogito Jobs Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-jobs-service_kogito-deploying-on-openshift) to enable jobs scheduling.

          For Kubernetes deployment of Kogito services, please refer to the section [Deploying Kogito services on Kubernetes](https://docs.jboss.org/kogito/release/latest/html_single/#proc-kogito-deploying-on-kubernetes_kogito-deploying-on-openshift) on our docs.

          **IMPORTANT!** Kogito Operator might need Infinispan Operator and Strimzi to implement persistence and messaging use cases if you do not provide these services beforehand.
          If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the Kogito Operator.
        displayName: Kogito
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://kogito.kie.org/
        - name: Documentation
          url: https://github.com/kiegroup/kogito-operator/blob/main/README.md
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kiegroup/kogito-operator:1.44.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        version: 1.44.1
      entries:
      - name: kogito-operator.v1.44.1
        version: 1.44.1
      - name: kogito-operator.v1.43.0
        version: 1.43.0
      - name: kogito-operator.v1.42.0
        version: 1.42.0
      - name: kogito-operator.v1.41.0
        version: 1.41.0
      - name: kogito-operator.v1.40.0
        version: 1.40.0
      - name: kogito-operator.v1.39.0
        version: 1.39.0
      - name: kogito-operator.v1.38.0
        version: 1.38.0
      - name: kogito-operator.v1.37.0
        version: 1.37.0
      - name: kogito-operator.v1.36.0
        version: 1.36.0
      - name: kogito-operator.v1.35.0
        version: 1.35.0
      - name: kogito-operator.v1.34.0
        version: 1.34.0
      - name: kogito-operator.v1.33.0
        version: 1.33.0
      - name: kogito-operator.v1.32.0
        version: 1.32.0
      - name: kogito-operator.v1.31.0
        version: 1.31.0
      - name: kogito-operator.v1.30.0
        version: 1.30.0
      - name: kogito-operator.v1.29.0
        version: 1.29.0
      - name: kogito-operator.v1.28.0
        version: 1.28.0
      - name: kogito-operator.v1.27.0
        version: 1.27.0
      - name: kogito-operator.v1.26.0
        version: 1.26.0
      - name: kogito-operator.v1.25.0
        version: 1.25.0
      - name: kogito-operator.v1.24.0
        version: 1.24.0
      - name: kogito-operator.v1.23.0
        version: 1.23.0
      - name: kogito-operator.v1.22.1
        version: 1.22.1
      - name: kogito-operator.v1.22.0
        version: 1.22.0
      - name: kogito-operator.v1.21.0
        version: 1.21.0
      - name: kogito-operator.v1.20.0
        version: 1.20.0
      - name: kogito-operator.v1.19.0
        version: 1.19.0
      - name: kogito-operator.v1.18.0
        version: 1.18.0
      - name: kogito-operator.v1.17.0
        version: 1.17.0
      - name: kogito-operator.v1.16.1
        version: 1.16.1
      - name: kogito-operator.v1.16.0
        version: 1.16.0
      - name: kogito-operator.v1.15.0
        version: 1.15.0
      - name: kogito-operator.v1.14.0
        version: 1.14.0
      - name: kogito-operator.v1.13.0
        version: 1.13.0
      - name: kogito-operator.v1.12.0
        version: 1.12.0
      - name: kogito-operator.v1.11.1
        version: 1.11.1
      - name: kogito-operator.v1.11.0
        version: 1.11.0
      - name: kogito-operator.v1.10.0
        version: 1.10.0
      - name: kogito-operator.v1.9.1
        version: 1.9.1
      - name: kogito-operator.v1.9.0
        version: 1.9.0
      - name: kogito-operator.v1.8.0
        version: 1.8.0
      - name: kogito-operator.v1.7.0
        version: 1.7.0
      - name: kogito-operator.v1.6.0
        version: 1.6.0
      - name: kogito-operator.v1.5.0
        version: 1.5.0
      - name: kogito-operator.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: 1.x
    packageName: kogito-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: neutron-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neutron-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "neutron.openstack.org/v1beta1",
                "kind": "NeutronAPI",
                "metadata": {
                  "name": "neutron",
                  "namespace": "openstack"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "neutron",
                  "databaseInstance": "openstack",
                  "memcachedInstance": "memcached",
                  "preserveJobs": false,
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "neutron-secret",
                  "serviceUser": "neutron"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:57Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NeutronAPI is the Schema for the neutronapis API
            displayName: Neutron API
            kind: NeutronAPI
            name: neutronapis.neutron.openstack.org
            version: v1beta1
        description: Neutron Operator
        displayName: Neutron Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Network
        - Neutron
        links:
        - name: Neutron Operator
          url: https://github.com/openstack-k8s-operators/neutron-operator
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/neutron-rhel9-operator@sha256:16b1a5f83692f3d193754df664992e5b88cb357f158b69261eb802430f3a69f1
        - registry.redhat.io/rhoso/openstack-neutron-server-rhel9@sha256:806e5e009a77a29c82699487e539669e1ec2136514a0332d222b5cb3fed6f452
        version: 1.0.6
      entries:
      - name: neutron-operator.v1.0.6
        version: 1.0.6
      - name: neutron-operator.v1.0.5
        version: 1.0.5
      - name: neutron-operator.v1.0.4
        version: 1.0.4
      - name: neutron-operator.v1.0.3
        version: 1.0.3
      - name: neutron-operator.v1.0.2
        version: 1.0.2
      - name: neutron-operator.v1.0.1
        version: 1.0.1
      - name: neutron-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: neutron-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Tempo Operator SIG
      provider-url: ""
    name: tempo-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tempo-operator.v0.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoMonolithic",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "jaegerui": {
                    "enabled": true,
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    }
                  },
                  "storage": {
                    "traces": {
                      "backend": "memory"
                    }
                  }
                }
              },
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoStack",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "resources": {
                    "total": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2Gi"
                      }
                    }
                  },
                  "storage": {
                    "secret": {
                      "name": "my-storage-secret",
                      "type": "s3"
                    }
                  },
                  "storageSize": "1Gi",
                  "template": {
                    "queryFrontend": {
                      "jaegerQuery": {
                        "enabled": true,
                        "ingress": {
                          "type": "route"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          containerImage: ghcr.io/grafana/tempo-operator/tempo-operator:v0.18.0
          createdAt: "2025-09-17T14:23:28Z"
          description: Create and manage deployments of Tempo, a high-scale distributed
            tracing backend.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: tempo-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/tempo-operator
          support: Grafana Tempo Operator SIG
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TempoMonolithic manages a Tempo deployment in monolithic
              mode.
            displayName: TempoMonolithic
            kind: TempoMonolithic
            name: tempomonolithics.tempo.grafana.com
            version: v1alpha1
          - description: TempoStack manages a Tempo deployment in microservices mode.
            displayName: TempoStack
            kind: TempoStack
            name: tempostacks.tempo.grafana.com
            version: v1alpha1
        description: |-
          Tempo is an open source, easy-to-use, and high-scale distributed tracing backend.
          It can ingest common open source tracing protocols including Jaeger, Zipkin, and OpenTelemetry and requires only object storage to operate.
          Please refer to the [Tempo documentation](https://grafana.com/docs/tempo/latest/) for more information about Tempo.

          The Community Tempo Operator can manage Tempo deployments in Microservices mode (`TempoStack` CR) and Monolithic mode (`TempoMonolithic` CR).

          ### Operator features
          * **Resource Limits** - Specify overall resource requests and limits in the `TempoStack` CR; the operator assigns fractions of it to each component
          * **AuthN and AuthZ** - Supports OpenID Control (OIDC) and role-based access control (RBAC)
          * **Managed upgrades** - Updating the operator will automatically update all managed Tempo clusters
          * **Multitenancy** - Multiple tenants can send traces to the same Tempo cluster
          * **mTLS** - Communication between the Tempo components can be secured via mTLS
          * **Jaeger UI** - Traces can be visualized in Jaeger UI and exposed via Ingress or OpenShift Route
          * **Observability** - The operator and `TempoStack` operands expose telemetry (metrics, traces) and integrate with Prometheus `ServiceMonitor` and `PrometheusRule`

          ### Prerequisites
          `TempoStack` requires object storage to store its traces. `TempoMonolithic` can store traces in-memory, in a Persistent Volume and in object storage.
          Please ensure that an object storage solution is available and configured.
        displayName: Community Tempo Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tempo
        - tracing
        - observability
        - monitoring
        - database
        links:
        - name: Tempo Operator
          url: https://github.com/grafana/tempo-operator
        maintainers:
        - email: ruben.vp8510@gmail.com
          name: Ruben Vargas
        - email: p.loffay@gmail.com
          name: Pavol Loffay
        - email: bongartz@klimlive.de
          name: Benedikt Bongartz
        - email: andreas@gerstmayr.me
          name: Andreas Gerstmayr
        - email: iblancas@redhat.com
          name: Israel Blancas Alvarez
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Grafana Tempo Operator SIG
        relatedImages:
        - docker.io/jaegertracing/jaeger-query:1.68.0
        - docker.io/grafana/tempo-query:2.8.2
        - quay.io/observatorium/api:main-2025-08-25-7489eb0
        - quay.io/observatorium/opa-openshift:main-2025-06-16-ecdeca0
        - quay.io/openshift/origin-oauth-proxy:4.14
        - ghcr.io/grafana/tempo-operator/tempo-operator:v0.18.0
        - docker.io/grafana/tempo:2.8.2
        version: 0.18.0
      entries:
      - name: tempo-operator.v0.18.0
        version: 0.18.0
      - name: tempo-operator.v0.17.1
        version: 0.17.1
      - name: tempo-operator.v0.15.3
        version: 0.15.3
      - name: tempo-operator.v0.15.2
        version: 0.15.2
      - name: tempo-operator.v0.15.1
        version: 0.15.1
      - name: tempo-operator.v0.15.0
        version: 0.15.0
      - name: tempo-operator.v0.14.2
        version: 0.14.2
      - name: tempo-operator.v0.14.1
        version: 0.14.1
      - name: tempo-operator.v0.14.0
        version: 0.14.0
      - name: tempo-operator.v0.13.0
        version: 0.13.0
      - name: tempo-operator.v0.12.0
        version: 0.12.0
      - name: tempo-operator.v0.11.1
        version: 0.11.1
      - name: tempo-operator.v0.11.0
        version: 0.11.0
      - name: tempo-operator.v0.10.0
        version: 0.10.0
      - name: tempo-operator.v0.9.0
        version: 0.9.0
      - name: tempo-operator.v0.8.0
        version: 0.8.0
      - name: tempo-operator.v0.7.0
        version: 0.7.0
      - name: tempo-operator.v0.6.0
        version: 0.6.0
      - name: tempo-operator.v0.5.0
        version: 0.5.0
      - name: tempo-operator.v0.4.0
        version: 0.4.0
      - name: tempo-operator.v0.3.0
        version: 0.3.0
      - name: tempo-operator.v0.2.0
        version: 0.2.0
      - name: tempo-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: tempo-operator
    provider:
      name: Grafana Tempo Operator SIG
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: project-quay
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-operator.v3.10.0-nightly.20231014
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.10.0-nightly.20231014
          createdAt: 2023-10-14 06:26 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.10.0-nightly.20231014'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.10.0-nightly.20231014
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.10.0-nightly.20231014
        version: 3.10.0-nightly.20231014
      entries:
      - name: quay-operator.v3.10.0-nightly.20231014
        version: 3.10.0-nightly.20231014
      - name: quay-operator.v3.10.0-nightly.20231013
        version: 3.10.0-nightly.20231013
      - name: quay-operator.v3.10.0-nightly.20231012
        version: 3.10.0-nightly.20231012
      - name: quay-operator.v3.10.0-nightly.20231011
        version: 3.10.0-nightly.20231011
      - name: quay-operator.v3.10.0-nightly.20231010
        version: 3.10.0-nightly.20231010
      - name: quay-operator.v3.10.0-nightly.20231009
        version: 3.10.0-nightly.20231009
      - name: quay-operator.v3.10.0-nightly.20231008
        version: 3.10.0-nightly.20231008
      - name: quay-operator.v3.10.0-nightly.20231007
        version: 3.10.0-nightly.20231007
      - name: quay-operator.v3.10.0-nightly.20231006
        version: 3.10.0-nightly.20231006
      - name: quay-operator.v3.10.0-nightly.20231005
        version: 3.10.0-nightly.20231005
      - name: quay-operator.v3.10.0-nightly.20231002
        version: 3.10.0-nightly.20231002
      - name: quay-operator.v3.10.0-nightly.20230930
        version: 3.10.0-nightly.20230930
      - name: quay-operator.v3.10.0-nightly.20230929
        version: 3.10.0-nightly.20230929
      - name: quay-operator.v3.10.0-nightly.20230928
        version: 3.10.0-nightly.20230928
      - name: quay-operator.v3.10.0-nightly.20230926
        version: 3.10.0-nightly.20230926
      - name: quay-operator.v3.10.0-nightly.20230925
        version: 3.10.0-nightly.20230925
      - name: quay-operator.v3.10.0-nightly.20230923
        version: 3.10.0-nightly.20230923
      - name: quay-operator.v3.10.0-nightly.20230922
        version: 3.10.0-nightly.20230922
      - name: quay-operator.v3.10.0-nightly.20230921
        version: 3.10.0-nightly.20230921
      - name: quay-operator.v3.10.0-nightly.20230920
        version: 3.10.0-nightly.20230920
      - name: quay-operator.v3.10.0-nightly.20230919
        version: 3.10.0-nightly.20230919
      - name: quay-operator.v3.10.0-nightly.20230918
        version: 3.10.0-nightly.20230918
      - name: quay-operator.v3.10.0-nightly.20230917
        version: 3.10.0-nightly.20230917
      - name: quay-operator.v3.10.0-nightly.20230916
        version: 3.10.0-nightly.20230916
      - name: quay-operator.v3.10.0-nightly.20230915
        version: 3.10.0-nightly.20230915
      - name: quay-operator.v3.10.0-nightly.20230914
        version: 3.10.0-nightly.20230914
      - name: quay-operator.v3.10.0-nightly.20230913
        version: 3.10.0-nightly.20230913
      name: candidate-3.10
    - currentCSV: quay-operator.v3.9.0-nightly.20230812
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.9.0-nightly.20230812
          createdAt: 2023-08-12 06:06 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.9.0-nightly.20230812'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.9.0-nightly.20230812
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.9.0-nightly.20230812
        version: 3.9.0-nightly.20230812
      entries:
      - name: quay-operator.v3.9.0-nightly.20230812
        version: 3.9.0-nightly.20230812
      - name: quay-operator.v3.9.0-nightly.20230724
        version: 3.9.0-nightly.20230724
      - name: quay-operator.v3.9.0-nightly.20230723
        version: 3.9.0-nightly.20230723
      - name: quay-operator.v3.9.0-nightly.20230722
        version: 3.9.0-nightly.20230722
      - name: quay-operator.v3.9.0-nightly.20230721
        version: 3.9.0-nightly.20230721
      - name: quay-operator.v3.9.0-nightly.20230720
        version: 3.9.0-nightly.20230720
      - name: quay-operator.v3.9.0-nightly.20230717
        version: 3.9.0-nightly.20230717
      - name: quay-operator.v3.9.0-nightly.20230716
        version: 3.9.0-nightly.20230716
      - name: quay-operator.v3.9.0-nightly.20230715
        version: 3.9.0-nightly.20230715
      - name: quay-operator.v3.9.0-nightly.20230714
        version: 3.9.0-nightly.20230714
      - name: quay-operator.v3.9.0-nightly.20230712
        version: 3.9.0-nightly.20230712
      - name: quay-operator.v3.9.0-nightly.20230711
        version: 3.9.0-nightly.20230711
      name: candidate-3.9
    - currentCSV: quay-operator.v3.7.0-rc.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.7.0-rc.4
          createdAt: 2022-05-17 14:05 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.3.x <3.7.0-rc.4'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.7.0-rc.4
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.7.0-rc.4
        version: 3.7.0-rc.4
      entries:
      - name: quay-operator.v3.7.0-rc.4
        version: 3.7.0-rc.4
      - name: quay-operator.v3.7.0-rc.3
        version: 3.7.0-rc.3
      - name: quay-operator.v3.7.0-rc.2
        version: 3.7.0-rc.2
      - name: quay-operator.v3.7.0-rc.1
        version: 3.7.0-rc.1
      name: preview-3.7
    - currentCSV: quay-operator.v3.8.0-rc.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.8.0-rc.7
          createdAt: 2022-11-29 15:11 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.8.0-rc.7'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.8.0-rc.7
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.8.0-rc.7
        version: 3.8.0-rc.7
      entries:
      - name: quay-operator.v3.8.0-rc.7
        version: 3.8.0-rc.7
      - name: quay-operator.v3.8.0-rc.6
        version: 3.8.0-rc.6
      - name: quay-operator.v3.8.0-rc.5
        version: 3.8.0-rc.5
      - name: quay-operator.v3.8.0-rc.4
        version: 3.8.0-rc.4
      - name: quay-operator.v3.8.0-rc.3
        version: 3.8.0-rc.3
      - name: quay-operator.v3.8.0-rc.2
        version: 3.8.0-rc.2
      - name: quay-operator.v3.8.0-rc.1
        version: 3.8.0-rc.1
      name: preview-3.8
    - currentCSV: quay-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.10.15
          createdAt: "2025-09-25T14:06:05Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.10.15
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.10.15
        version: 3.10.15
      entries:
      - name: quay-operator.v3.10.15
        version: 3.10.15
      - name: quay-operator.v3.10.14
        version: 3.10.14
      - name: quay-operator.v3.10.13
        version: 3.10.13
      - name: quay-operator.v3.10.12
        version: 3.10.12
      - name: quay-operator.v3.10.11
        version: 3.10.11
      - name: quay-operator.v3.10.10
        version: 3.10.10
      - name: quay-operator.v3.10.9
        version: 3.10.9
      - name: quay-operator.v3.10.8
        version: 3.10.8
      - name: quay-operator.v3.10.7
        version: 3.10.7
      - name: quay-operator.v3.10.6
        version: 3.10.6
      - name: quay-operator.v3.10.5
        version: 3.10.5
      - name: quay-operator.v3.10.4
        version: 3.10.4
      - name: quay-operator.v3.10.3
        version: 3.10.3
      - name: quay-operator.v3.10.2
        version: 3.10.2
      - name: quay-operator.v3.10.1
        version: 3.10.1
      - name: quay-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.11.13
          createdAt: "2025-09-02T06:02:15Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.11.13
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.11.13
        version: 3.11.13
      entries:
      - name: quay-operator.v3.11.13
        version: 3.11.13
      - name: quay-operator.v3.11.12
        version: 3.11.12
      - name: quay-operator.v3.11.11
        version: 3.11.11
      - name: quay-operator.v3.11.10
        version: 3.11.10
      - name: quay-operator.v3.11.9
        version: 3.11.9
      - name: quay-operator.v3.11.8
        version: 3.11.8
      - name: quay-operator.v3.11.7
        version: 3.11.7
      - name: quay-operator.v3.11.6
        version: 3.11.6
      - name: quay-operator.v3.11.5
        version: 3.11.5
      - name: quay-operator.v3.11.4
        version: 3.11.4
      - name: quay-operator.v3.11.3
        version: 3.11.3
      - name: quay-operator.v3.11.2
        version: 3.11.2
      - name: quay-operator.v3.11.1
        version: 3.11.1
      - name: quay-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.12.12
          createdAt: "2025-09-23T09:59:30Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.12.12
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.12.12
        version: 3.12.12
      entries:
      - name: quay-operator.v3.12.12
        version: 3.12.12
      - name: quay-operator.v3.12.11
        version: 3.12.11
      - name: quay-operator.v3.12.10
        version: 3.12.10
      - name: quay-operator.v3.12.9
        version: 3.12.9
      - name: quay-operator.v3.12.8
        version: 3.12.8
      - name: quay-operator.v3.12.7
        version: 3.12.7
      - name: quay-operator.v3.12.6
        version: 3.12.6
      - name: quay-operator.v3.12.5
        version: 3.12.5
      - name: quay-operator.v3.12.4
        version: 3.12.4
      - name: quay-operator.v3.12.3
        version: 3.12.3
      - name: quay-operator.v3.12.2
        version: 3.12.2
      - name: quay-operator.v3.12.1
        version: 3.12.1
      - name: quay-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.13.8
          createdAt: "2025-08-28T03:28:40Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.13.8
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.13.8
        version: 3.13.8
      entries:
      - name: quay-operator.v3.13.8
        version: 3.13.8
      - name: quay-operator.v3.13.7
        version: 3.13.7
      - name: quay-operator.v3.13.6
        version: 3.13.6
      - name: quay-operator.v3.13.5
        version: 3.13.5
      - name: quay-operator.v3.13.4
        version: 3.13.4
      - name: quay-operator.v3.13.3
        version: 3.13.3
      - name: quay-operator.v3.13.2
        version: 3.13.2
      - name: quay-operator.v3.13.1
        version: 3.13.1
      - name: quay-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.14.5
          createdAt: "2025-09-22T14:19:49Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.14.5
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.14.5
        version: 3.14.5
      entries:
      - name: quay-operator.v3.14.5
        version: 3.14.5
      - name: quay-operator.v3.14.4
        version: 3.14.4
      - name: quay-operator.v3.14.3
        version: 3.14.3
      - name: quay-operator.v3.14.2
        version: 3.14.2
      - name: quay-operator.v3.14.1
        version: 3.14.1
      - name: quay-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.15.2
          createdAt: "2025-09-22T13:44:32Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.15.2
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.15.2
        version: 3.15.2
      entries:
      - name: quay-operator.v3.15.2
        version: 3.15.2
      - name: quay-operator.v3.15.1
        version: 3.15.1
      - name: quay-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-operator.v3.6.2
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"quay.redhat.com/v1\",\n    \"kind\":
            \"QuayRegistry\",\n    \"metadata\": {\n      \"name\": \"example-registry\"\n
            \   },\n    \"spec\": {\n      \"components\": [\n        {\"kind\": \"clair\",
            \"managed\": true},\n        {\"kind\": \"postgres\", \"managed\": true},\n
            \       {\"kind\": \"objectstorage\", \"managed\": true},\n        {\"kind\":
            \"redis\", \"managed\": true},\n        {\"kind\": \"horizontalpodautoscaler\",
            \"managed\": true},\n        {\"kind\": \"route\", \"managed\": true},\n
            \       {\"kind\": \"mirror\", \"managed\": true},\n        {\"kind\":
            \"monitoring\", \"managed\": true},\n        {\"kind\": \"tls\", \"managed\":
            true}\n      ]\n    }\n  }, \n  {\n    \"apiVersion\": \"redhatcop.redhat.io/v1alpha1\",\n
            \   \"kind\": \"QuayEcosystem\",\n    \"metadata\": {\n      \"name\":
            \"example-quayecosystem\"\n    },\n    \"spec\": {}\n  }\n]"
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:v3.6.2
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.5.x <3.6.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.6.2
          repository: https://github.com/quay/quay-operator
          support: Quay Team
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Project Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        - quay
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:v3.6.2
        version: 3.6.2
      entries:
      - name: quay-operator.v3.6.2
        version: 3.6.2
      name: stable-3.6
    - currentCSV: quay-operator.v3.7.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.7.11
          createdAt: 2023-02-22 10:02 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.3.x <3.7.11'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.7.11
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.7.11
        version: 3.7.11
      entries:
      - name: quay-operator.v3.7.11
        version: 3.7.11
      - name: quay-operator.v3.7.10
        version: 3.7.10
      - name: quay-operator.v3.7.9
        version: 3.7.9
      - name: quay-operator.v3.7.8
        version: 3.7.8
      - name: quay-operator.v3.7.7
        version: 3.7.7
      - name: quay-operator.v3.7.6
        version: 3.7.6
      - name: quay-operator.v3.7.5
        version: 3.7.5
      - name: quay-operator.v3.7.4
        version: 3.7.4
      - name: quay-operator.v3.7.3
        version: 3.7.3
      - name: quay-operator.v3.7.2
        version: 3.7.2
      - name: quay-operator.v3.7.1
        version: 3.7.1
      - name: quay-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: quay-operator.v3.8.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.8.13
          createdAt: 2023-10-17 16:13 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.8.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.8.13
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.8.13
        version: 3.8.13
      entries:
      - name: quay-operator.v3.8.13
        version: 3.8.13
      - name: quay-operator.v3.8.12
        version: 3.8.12
      - name: quay-operator.v3.8.11
        version: 3.8.11
      - name: quay-operator.v3.8.10
        version: 3.8.10
      - name: quay-operator.v3.8.8
        version: 3.8.8
      - name: quay-operator.v3.8.7
        version: 3.8.7
      - name: quay-operator.v3.8.6
        version: 3.8.6
      - name: quay-operator.v3.8.5
        version: 3.8.5
      - name: quay-operator.v3.8.4
        version: 3.8.4
      - name: quay-operator.v3.8.3
        version: 3.8.3
      - name: quay-operator.v3.8.2
        version: 3.8.2
      - name: quay-operator.v3.8.1
        version: 3.8.1
      - name: quay-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.9.16
          createdAt: "2025-09-25T14:31:58Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.9.16
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.9.16
        version: 3.9.16
      entries:
      - name: quay-operator.v3.9.16
        version: 3.9.16
      - name: quay-operator.v3.9.15
        version: 3.9.15
      - name: quay-operator.v3.9.14
        version: 3.9.14
      - name: quay-operator.v3.9.13
        version: 3.9.13
      - name: quay-operator.v3.9.12
        version: 3.9.12
      - name: quay-operator.v3.9.11
        version: 3.9.11
      - name: quay-operator.v3.9.10
        version: 3.9.10
      - name: quay-operator.v3.9.9
        version: 3.9.9
      - name: quay-operator.v3.9.6
        version: 3.9.6
      - name: quay-operator.v3.9.5
        version: 3.9.5
      - name: quay-operator.v3.9.4
        version: 3.9.4
      - name: quay-operator.v3.9.3
        version: 3.9.3
      - name: quay-operator.v3.9.2
        version: 3.9.2
      - name: quay-operator.v3.9.1
        version: 3.9.1
      - name: quay-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: project-quay
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-ui
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-ui.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Ui",
                "metadata": {
                  "name": "ui-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "volumeMounts": [
                    {
                      "mountPath": "/opt/app-root/src/assets/config.json",
                      "name": "config",
                      "subPath": "config.json"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:37:58Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ui
            name: uis.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP UI Operator
        displayName: Keysight WAP UI Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Ui
          url: https://keysight-wap-ui.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-ui@sha256:876a394f5cc2f745091ad2ee9ca0d08404cfb95284222e1f3a8607e193bd986a
        version: 0.1.21
      entries:
      - name: keysight-wap-ui.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-ui
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator-certified-rhmp.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator-certified-rhmp.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator-certified-rhmp.v2.6.1
        version: 2.6.1
      - name: percona-postgresql-operator-certified-rhmp.v2.6.0
        version: 2.6.0
      - name: percona-postgresql-operator-certified-rhmp.v2.5.0
        version: 2.5.0
      name: stable
    - currentCSV: percona-postgresql-operator-certified-rhmp.v2.7.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        version: 2.7.0-cw
      entries:
      - name: percona-postgresql-operator-certified-rhmp.v2.7.0-cw
        version: 2.7.0-cw
      - name: percona-postgresql-operator-certified-rhmp.v2.6.1-cw
        version: 2.6.1-cw
      - name: percona-postgresql-operator-certified-rhmp.v2.6.0-cw
        version: 2.6.0-cw
      - name: percona-postgresql-operator-certified-rhmp.v2.5.0-cw
        version: 2.5.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-postgresql-operator-certified-rhmp
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: aws-efs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-efs-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"aws-efs.managed.openshift.io/v1alpha1","kind":"SharedVolume","metadata":{"name":"sv1"},"spec":{"accessPointID":
            "fsap-0123456789abcdef", "fileSystemID": "fs-0123cdef"}}]'
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/app-sre/aws-efs-operator:9c87267
          createdAt: "2021-12-20T19:51:50Z"
          description: Manage read-write-many access to AWS EFS volumes in an OpenShift
            cluster.
          repository: https://github.com/openshift/aws-efs-operator
          support: https://github.com/openshift/aws-efs-operator/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Shared Volume
            displayName: SharedVolume
            kind: SharedVolume
            name: sharedvolumes.aws-efs.managed.openshift.io
            version: v1alpha1
        description: |-
          ## About the managed application
          AWS provides Elastic File System, an NFS-based shared storage solution. A couple of mechanisms exist to enable mounting such a file system in containers. Curating these options and making them work in a cluster is nontrivial.
          ## About this Operator
          The aws-efs-operator minimizes the barrier to entry for container use of EFS by:
          - Installing the [AWS EFS CSI Driver](https://github.com/kubernetes-sigs/aws-efs-csi-driver) and supporting infrastructure.
          - Reducing the identification of a shared storage resource to a simple Custom Resource Definition with only two fields.
          - Creating the necessary PersistentVolume and PersistentVolumeClaim resources that can simply be referenced from a Pod specification.

          More information can be found in the operator's [source repository](https://github.com/openshift/aws-efs-operator).
          ## Prerequisites for enabling this Operator
          An OpenShift cluster backed by AWS.
        displayName: AWS EFS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AWS
        - EFS
        - NFS
        - Shared storage
        links:
        - name: README
          url: https://github.com/openshift/aws-efs-operator/blob/master/README.md
        - name: Design
          url: https://github.com/openshift/aws-efs-operator/blob/master/DESIGN.md
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/app-sre/aws-efs-operator:9c87267
        version: 0.0.8
      entries:
      - name: aws-efs-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: aws-efs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-application-gateway-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-application-gateway-operator.v25.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMApplicationGateway",
                "metadata": {
                  "name": "iag-instance"
                },
                "spec": {
                  "configuration": [
                    {
                      "dataKey": "config",
                      "name": "test-config",
                      "type": "configmap"
                    },
                    {
                      "discoveryEndpoint": "https://isam.mmfa.ibm.com/mga/sps/oauth/oauth20/metadata/test",
                      "postData": [
                        {
                          "name": "redirect_uris",
                          "values": [
                            "https://isam.mmfa.ibm.com/pkmsoidc"
                          ]
                        },
                        {
                          "name": "client_name",
                          "value": "OperatorTest"
                        },
                        {
                          "name": "enforce_pkce",
                          "value": "false"
                        },
                        {
                          "name": "all_users_entitled",
                          "value": "true"
                        },
                        {
                          "name": "consent_action",
                          "value": "never_prompt"
                        }
                      ],
                      "secret": "oidc-client",
                      "type": "oidc_registration"
                    },
                    {
                      "headers": [
                        {
                          "name": "Authorization",
                          "secretKey": "value",
                          "type": "secret",
                          "value": "githubsecret"
                        }
                      ],
                      "type": "web",
                      "url": "https://raw.github.com/ibm-security/iag-config/master/test/sample1.yaml"
                    },
                    {
                      "type": "literal",
                      "value": "version: \"22.07\"\n\nserver:\n  local_applications:\n    cred_viewer:\n      path_segment: creds\n      enable_html: true\n\nidentity:\n  oidc:\n    discovery_endpoint: \"https://iag-dev.ibmcloudsecurity.com/oidc/endpoint/default/.well-known/openid-configuration\"\n    client_id: 1cbfe647-9e5f-4d99-8e05-8ed1c862eb47\n    client_secret: uPP8rM7N0e\n"
                    }
                  ],
                  "deployment": {
                    "image": "icr.io/ibmappgateway/ibm-application-gateway@sha256:1144b79a2930fe942ce627b0abcc3bed2f0b0770ae8313cb77ffe6772511c1f9",
                    "imagePullPolicy": "Always",
                    "imagePullSecrets": [
                      {
                        "name": "regcred"
                      }
                    ],
                    "lang": "C",
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 8,
                      "periodSeconds": 9,
                      "successThreshold": 7,
                      "timeoutSeconds": 1
                    },
                    "readinessProbe": {
                      "failureThreshold": 2,
                      "initialDelaySeconds": 7,
                      "periodSeconds": 8,
                      "successThreshold": 4,
                      "timeoutSeconds": 5
                    },
                    "serviceAccountName": "iag"
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/ibmappgateway/ibm-application-gateway-operator@sha256:6a16d4ef802a866942f504b11a948565245d303f322944200c1a1269fe7c371c
          createdAt: "2025-02-05T21:15:37Z"
          description: The IBM Application Gateway operator manages the lifecycle
            of IBM Application Gateway containers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM-Security/ibm-application-gateway-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMApplicationGateway is the Schema for the ibmapplicationgateways
              API
            displayName: IBMApplication Gateway
            kind: IBMApplicationGateway
            name: ibmapplicationgateways.ibm.com
            version: v1
        description: "The [IBM Application Gateway (IAG)](https://ibm.biz/ibm-app-gateway)
          image provides a containerized secure Web Reverse proxy which is designed
          to sit in front of your application, seamlessly adding authentication and
          authorization protection to your application.\n\nAn IAG instance deployed
          on Kubernetes can be a complex deployment.  In particular the configuration
          can be defined externally in one or more locations, and changes to this
          configuration may require all instances to be reloaded for the changes to
          take effect. \n\nThe internal Kubernetes deployment controller does not
          have any knowledge of how an IBM Application Gateway instance should behave
          when the configuration changes. As such this IBM Application Gateway specific
          Kubernetes operator is available to be deployed. Once deployed the operator
          can manage the lifecycle of the IBM Application Gateway instances.\n\nSee
          the project [Readme](https://github.com/IBM-Security/ibm-application-gateway-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Application Gateway Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Security
        - Identity and Access
        links:
        - name: IBM Application Gateway product documentation
          url: https://ibm.biz/ibm-app-gateway
        maintainers:
        - email: isamdev@au1.ibm.com
          name: IBM Application Gateway Development Team
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - icr.io/ibmappgateway/ibm-application-gateway-operator@sha256:6a16d4ef802a866942f504b11a948565245d303f322944200c1a1269fe7c371c
        - icr.io/ibmappgateway/ibm-application-gateway@sha256:1144b79a2930fe942ce627b0abcc3bed2f0b0770ae8313cb77ffe6772511c1f9
        version: 25.2.0
      entries:
      - name: ibm-application-gateway-operator.v25.2.0
        version: 25.2.0
      - name: ibm-application-gateway-operator.v24.4.0
        version: 24.4.0
      name: stable
    defaultChannel: stable
    packageName: ibm-application-gateway-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cockroach Labs
      provider-url: ""
    name: cockroachdb-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroach-operator.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v21.1.10",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cockroachdb/cockroachdb-operator@sha256:9fba7cd5fdf589317911925cb69338508bf9de928b1817c6f117c612635bbab3
          createdAt: "2021-10-19T15:47:03Z"
          description: CockroachDB Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:e67299fc0e79cff5b6caf9c0df154180dbaedc57b84f072acf68b7b2c958668d
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ec1672ce8091b677689a31cf09dbde5d1697f4caddae0151040b0fb156722e9d
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:e39bce1d9b08a7d62eb05d53e29fabc668bf6a4e4ae3eb9994eebf5b642ff6cf
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ab47ddd22328e0a1564814f049616c3e3569d24c0a290abee9308392c6a0de23
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:49421968126f9c98499bc0cf0c65729370ab566da3fbd1588c100b20052b972e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:c56f980b87a066d1e65ac8cff3d75040af9bd13093494cf4b4d30b5234d56edc
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:de94036e9cc8618e0493313512d880ae145e94d3bff54ed51b0de66a88da1319
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:3edda46eccce8350440a0ea78dad2e41a3eaa0838fe57945ad5318739e968516
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:0bd22d69db5ac570f30d3379afc67f6c2e46fc5b027c8aab42c3f562085d4672
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:bcf3f976437c4ee52e922aa5d7e382a86a031d780b711c08c6c158dbf2d26cfa
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:0ef0234b6fd42977b8a9eda2c59ca8194cc1c8cf6425b99d7bf8fde8feb826c5
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:87dcdbdc07904b217880e70484582c9d87dde7e4071e01cf9e2e6da43111190e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:2664d78fbe88c368927ddbf80704f430e6ca2ce1f72cb4b7e1906dfe72be7cd0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:fe0c615b4d1ef4ec6e089e85deae2b6bd85410be8ed525e2317c759669417f47
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1b56420753eac1126039b8ad4be82ddb59c89a4ec72fb7b6198cea88a10f3a91
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1ed18af1d3ffc65c24946c817d7d04d5f01025d3fb6c9eb78940e343b989bed0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:162a356cc8b423926634f83a1372b6321104d8de7a2c5b9b7b2681c769a88cc7
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:288ae92ebdfc848540ff80ef682b74e50809e9742cafce22b028112326d66b65
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:867c46bb4d8ca3f2fb8dc94063facc99758344c7b9d04f10e50932023ef5d262
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ba9da9aa9d662c0fa9fd16a40551eedf100869b008cf9c0c09cbb87a09fda697
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:25791b76b0d4b7220dadab3a405b3fdc330ba264ef6c9fbfdfb924c154c4bb5e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:cdab9fc62e07dd349703ae394adb48ab3d2281610df35e081f4b76e208b38c8e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:a5414e54c4c513c7877f925b09eadc4c28eb592d3e120bcf9c65ab10e07cd607
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:98b76cf63aabc39d026317905d6ba72eec3f0d35c0f770ebb8b789c3fa0709ee
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:9f5dfcde260804cd2d6976a6b3544ddabdd899365be2104210dc78ba13755ec3
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:27ef61d9f320a52ee9617f2ab8a1d1a1eacc576a0add2d85d99f9413912b651f
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:38e940144e34b69dfb7d1cf31456e7a0016c5dd40d5626baa346446c7678ceb0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1cea12847a0d236437f7fc28fba476fe512c343cd020d21bb50316e50da6ffd6
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:c1a07a8686bbadcb8c68ffba4e40e1285e581e24d43e5811f38a22d57fb9cc56
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:d290099496d070f135b5cc1e6bd59607512524d399df81698f11cfa09b8dca4a
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:b02549e5c3673b8154441990afa05a478d1ecc2bd0a70af64fd693cb6b469cb0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:5eb59093ee81f8115d82ec6a8a1c35eefad055cbb3a175978cef7afe9196e6a0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:fec6b0f56f2b94f141c8553a63b2e08b9494d4e6c5dc109d73268621ca445c56
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:d9a756df2ed2536b938d7c24fbf3de149cd1a0f44a3a1a0cc5d042fe8980362e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroachdb-operator@sha256:9fba7cd5fdf589317911925cb69338508bf9de928b1817c6f117c612635bbab3
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:6667919a41d304d5d4ade3ded4f11b42d722a995a4283e11e15320529f7f9abf
        version: 2.3.0
      entries:
      - name: cockroach-operator.v2.3.0
        version: 2.3.0
      name: beta
    - currentCSV: cockroach-operator.v2.18.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v25.2.2",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
          createdAt: "2025-07-22T11:30:56Z"
          description: CockroachDB Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9ab968e4ca62f23bf9848e5475ef4076df7318a8560f63d5b2090a32af6fe4e2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:90cf21c36d19d87b97d06edd1c0912a3bb288c769e6159651877db3206aa7355
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8f1f2fd1aca914a1ef007117c85ac944055bbfb873a468a7374d239ef20b065e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9b5a2e0c006eb4be8e1444734a98743551f278a24e81c20bebc96f0f3b631eb0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6371377718325235a2720dba0c97eb4321ee29b31a28cb893f76abd62baa339
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:68fe64fe3e112288ccfc2a7dffd69b9b2786a5da3d3bdedd6addf41659d460c9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5240f9f591263569fad7ee934e81fa8a1617de3c25da94ddd59ee5b2b3407f90
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f29c31bb267451b389608af3e8a4a28d6ca0691e67b77e8f57bb5e461b064893
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8737b67b91983817e3e4ff3f47f0b070bd36162d309b60993529426c6979cb24
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a3ac544d87a9d8e7c92897c4c23f803753c7ab31467db91b5c51244afc2c6be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf001173bbf00cdf443b8d64a3ad724389f1cf83e3628d0b609019c8cddad35a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce3d94a3fbdcde5858beb3c265987759971d991c397c6486e19904bd4706f44d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7d6320ac7fb0925f10cf6f524a13de629cbf13ccc371912e4fe8c777bc15d1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bd75aa918355e77cd671c3e11ee807861378e96245245f4fdf243cba5749f9a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d590946b2f7c7ec0c7ec057a338e81f864d6a9e3d3ac4e0769aa5756a8e13fc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:32fdca575c334822e4356aab36a7ed97b685c065925fe85f1b8ba8425c57159b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc4b75ed71845d5b18e1d9163d480fd9d0d3cb5ebbcfed00a2e4e1b174c0a5de
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f74c8a5dea4560ad59c6f6f6ab1e699f063334ee8db704460f870508dbeaeff0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4b3aa590623a38f435b075de4c09f8a343d8ece8b673e6bf7c6df5d5eb4bdf4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:46d9287ef11e3b21a36abfa31a26da3b93040bb6163b9223972a5dd1930fecff
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e0f21abea34825e55622c432a306f2e8aafd64d913eb78f2f601c51dc878a1da
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2b5c028f293c6f394e58a5495cb3dbfbedc215d0d03778d0217e045586d55a46
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08cec123813304ab03bd66d72a09949667a874a42320bedafda391ebcdb2e56c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6a2ee8d62c586eb74ebbef6bef256e06e1ba5cb51c1ae55b1654adafc2ca8859
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c2a62179e8ca293d7e30b9da599e48795345cf07a72abd3a83063107b53b03cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:72ea3d0ef72972dba93362dc31366c5946a9898f5c16780b877316a97068ef8f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1304d7f871a95c8b087bad1b548217562792edc66c8fe07c621e05242b8cd53e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9492894eaa5f43d00ac80b60029b32a0b0a85ae88ba609b46f98702c9349a5bb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bf36dab9b3257a4bba24d986f1451fc51a3a37023fbb988281f1941fd3faedd
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d467383af41aa80c26172964c34152abeba45121599804e502984655b72179f0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:022c6e1a0de875474bc953718a481a0af526c49dc910094dcdce0b5677aced18
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2d89398f03736f8ffaa82b10fed0f19fba166d3d2145b26f25cb71f3c39acf07
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e23af8f9d32c3f80a52d79ceb4b359dda1d75ba0dd4cb8ed4b66ac2f4e40c69b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:61e602b0e3582bf00460eb9328808ccf66e3ff1b1f6df07ee6c5c76a91846876
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8443c15eb1653dd9e5ec7aad54d4b8b092d41a0977e42b76e372e5e21280746e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0ec1dc184a2199736b4bd7de3241a357be7427caea5da1c931b08bc06f5c3dc0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c9a31bdf19175880d0981e21dd19a1458b6e8cde8d4acbcb554f3bea3578a64e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3f587e7b9f5a611013af2608e2151ed782859a52501b00adbc95739c56a8ff1e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7ad12df8e258b03c4099d375d02d272bd774bf188d108901d5cf6cb865e57ef3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4f753151af4c3d3f5a1df9014461316249a6d0621b8fba5c4430307b3355251
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:58de9d0b2ce005c7ad3e99152100e59b7677398557028a7a59e56e669fabb81c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1f6b1ec92251397dbcf7ed4a82f1f05619d64676aae2717f4eff1c33e4c2c27a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:266165ab87b19a05d6796e8b15ff0eba632d6ac96e6393400464c6136dd500ec
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:086d7b435993fdf21fd59b0093f52105a9028d6b769398b0033f98a65dfb7e79
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc1ab77a9537ccf236ce6dc370940dfedcfaa91b6e3db37bf2d5eade7ff84708
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4d8e804b35b73d7978c112748aea2d2bb68f65081b90bca62c39f2b6e2989378
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:744193037d6b64cb73fecab15ebe022abc2b79ab0335b76ec1415ac6379b7dfa
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5181196fff3bf8a5ec8fcfe4f1958cc69f285f0beb4f0b476b4b5e0912f3db42
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:40f00476b63cb9e370fe85f01407173693213a9242fc117d28b06378ca0d98e0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fd6c9b8ca82ff4f2d0fe29166d1fc9bca6439cb40907be60f4d8f1b9735de2f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:85a0e085901c2a73192556d5d70795d4b66cbcfe01d24c73b6fe6db8d5f2e5ca
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6bb15b36d64eebb6e4c1db5a5466e108b271d53383c58e0b6c78cec214756a9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b83eaff2e217c06098509709cead525b55b00ccd48019d8357e6231892008055
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:858dc82977b3c9e989a3a20626671b0c67c834c97132094a9212bb227ab55929
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da367cf0ac52045002e1c186f8e6964267ad87d5c25f9e72fd2c9b9a98a32702
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b77b80eed7ce149d2a3f228e5f188e76d73f32bc50ecde20bba9cfa7bdc4714c
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:53c7b960a2ed70a0998a4f8e28a4f78f7f676e01cffebe665e91e3a7629d88ed
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1558a1f5eab79f1a8c6c85d81740cc539dc23025ea457c0019d93a72dc05d57a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db4e5c2f27ba23e95a363c8174c01f0f9c6b6a27e0e581c86a38f3a1f68dc43c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:37bf9e1ec7e3eff47755fe7521223f31ec6aefc6e63901db3881b30898cf78f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:35ede65457d845e8c063ad7d9515d3a69f98e0751424ecbac484963678a0b1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:50393ad005fed19618d04f0716c6d55e2f1afce651e785c396638e852cac83b3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cad60044ad6573bd80b65e892a9ec0510dbe4fcbbfc6b51010ecf419f56f1024
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f42fa22aac3cc41ce72e144a72530ffbee765a392b24593dafd06cc098f8410b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:31b161f9bf9028a1a62cd2eef6abda01257d8889bdc3159822642549a75e3429
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9c85c25d5392817d17b68714dca6195ad53255d43e9c65b021d75b12d97e0579
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a6e2fe24ad52913e2160f85b9fe665a4e4c4d79b164b4b899f69dd9d92f999e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cc20270212f01fc231ff788e20b0476b8228c440bfc40430c0836d5eb702a042
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1374cbdc4ecf38e4b4226328f532e830b8932b34d9449bf2fbb36e60d0310521
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bd336715332f776a72aaf5c3ca52e62dd53a9e8521c1feb6e135cc868d694a39
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7481c0cf3b579c8d98d94d4aa0334ca94bf2ccfcd25ad70b57c42b22bbf4bcee
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4635550222775fc75737be5e5c0359303e3a23c4d1e1c296d7408cd15ad94aad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7daaf1368ab2c25694cdea0a598e1de59247b04a4b4b4b8348ed6edbaa194f9d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fe4a40006d1e8439423de1417cc9bc55716d790b882485074bb2a5ec68d7d1ad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5b3de46526d169fcd1318354545b3a18861843fa2022a23ed04d795493825287
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acc083f7dde1fbc557d858d03fc26b68b2056997fc1a457e4695650214d5b303
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0b0e0588c1aadf0c3826cd584bd2b3a7d6781882dcb5c5e037b47ff5cb84509b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3bf017b2b2506724500f5e29017d1b956825dd0f61a7c0049b0ff084c3554b5b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce1209ca6cc8724baee981c97c66e2852e512423bc0c0a16cdd80a6838b54031
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e82684b828256b3c2814930210031b3b70bc818514d4359a25f87ffdbf22be6e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:19f36f53f7da67755eb86da77098c5eeb84b381f9c18aa29ba573793dc498564
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f5a0ccc02dc9e938e484d5b5282ff650d1890d5f754c30a0c0d717989ed5d600
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9d0bfb642090ff55a3cc3da0f72663cae18063622a4ec785ba99bc71ef4db116
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f045fe5037d44aab1a0b2209241b3931ec7cb9897cb3c6a96a9f8d5280e55dd2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:44a0f40a6c46ba08297b482f3d96defe87b181c944803e5381c9acc42cabf97f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db07461824e73410456ef540b0bce3c7fe550aa67a73d6174300ca74b8f75276
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0754f7901feea9017bc5041cb91b7ac1cdbca10c2064f37c3648a2c395f24f83
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:28ae9661b38a72decf413c6a6c3758e61098debf5162197c27a12a070c7a096d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:743cda1eda6e273c798e3405a9a97e07dd6aa5ad4d5a52789be3b0f39a32a039
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:91f7e59f9e35a0395756eafed906963c535e886125294d671d930ce959a8f511
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d78c2e0709214f96794564f13386861f4b541e55b5bd316ad495cf104f7d40b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da3bd2e9ff2b08d5bb86b1f5faa95a865f85c5e8912df00b6b79190c1e518a2f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f3ffebd812881a830ff46c860984814dd61edcd51b478ebd3667759fb7710682
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5d6901b6b6fd6594d42bfc1842d050c7cffcb0867a5591a67469f0f860b1f21b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e8f26fd132affbaeb8ef6b1807a646c83b301e5dc334a592dfb2cf8dc3dbaf10
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c74cf927a2b2c5098ad74ad0880996f8b042186a9911d888a231ef5b9e2a0715
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5a54d564e08baefc2786f4b742b2507d31984664cd00333406aa7a3aa5b9c913
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:310310515625f099a928545865f7096997871ee4a16650a01c76c3799a18b684
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a0faa52a999289b308dc4caf050c6d9fe2f42e112aa30e4af282ed1f21c5c346
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a59cafff8d4ff2c62204ece15880117011c30978823c232902c7182d747ec827
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1e5d85f5db10fa79d09263db90daa44549bcc3b025e9f3f98cf0d4e51766394b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6ca927f137be28481e3a226e6c6bcb7c5ac27664984f17ffc6c1419cd7d08eb7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e50aab39722af22ccbf3d1db19e37972d52b0fc9a40998a64618dc966b2bac57
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d978678b3c254abd52f69f226b0aacc4b02c2aaca328c54ef10f6f9bb453582d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6408f89b0a88ec154f7ee98c314be574971c4c9aaf906b473d10124d822889be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:468b0a40a33be1dd5a2a89ce745d7e9012bc7371f2f01cc0f8f12c0ac887db43
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b06bd7f3802e8fa2af14bb0b07804cad4cc0fbbc97da18d09a9106bf074ed6d8
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f438fbbf9c87ca0d54c170e2c5105b1e265182e1c2ede48adea370e2be9dc895
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d8cfb221f67516675fdfde061f6ec5dfee1224ca7eb3d8be375c27d483f306af
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e151bded670f4cbd165a8ef864324805667d3e84ce3ace0cdf8f9aad087fccf1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cbc74889e9ca5edbd5aaa243d7649bca0c4da2719239f8d0ec9827868112a771
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a6f1c43486d044fed9f59406118f429c8037b389bd8c8d7a1c7aeb89ed2661f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2af60025ecb3bb933b61328be4c2b2bfd0e7d26f53b72430629208505e45c6d2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:79d688bebe0643ef0dd6ba7108cdb34c0442470c0c20ae0df88b5d02a3476cd0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9097e135159f46396ef4649f724d61bb5b9113c488f989d404586fbb7486fd85
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d14f77f70aea0422a906510c1ac2652ff76e131ed705cb002f80e13e887b0cd9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0fea6ed8a78ea624240ee6ffb3573d5cf1115186aa9180fcd7d1273351deaaa3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e18d73af3257c8cdf00245f2f718b56667524c1cc69d3db28e9b868a539ae16f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:dd521d7bad62da054eec54179be75c643a5492d981e11c0e66db41902cd2182c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:363d8feb560324a848c5771edd224dc19020013af7641979ee81ae8ee536d7e3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7b907d03b001b758d8a4b5e898725c73dda2ec2aa4b7425b685320b947154d11
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4b56c0311af14220a466aa11f25fd10fa38280825539bd02215c945bbafb861c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:42c9930d6861f6a2147ed6e01827eddd730279966763ec68868567a19cc164ab
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:740ba10f1ef0a6c5fa6ad31ab18855fc75f8e024ce4c23b0676c5a9c5b62cd75
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8e8c081c231b5b15eca483ceec16ea9190ba73aae6e421865162446eb584519a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:56109e57ee0379cf48644bcf8226a9238e01139cacc5499002c99f973f121911
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:742a6f69a3a209a7080a1e911fd96467d2a91aacb3f91d6be80a6a6e2e54b2ae
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0e2c7023f1f97acc20c46a4944635a9c5f57bb9bb3dd534cf41fea1a9cfdfdbe
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf7d3bea47638b763b312c7c11ca034a7ac62d993d75a13edceec74799b97e52
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:226d7f1c3f0d70ac6532ad9490bebcdef44a40959b6aa7c11b0f2d1383a46217
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6cd9f9b9fcb6e8f9a931a6ad01cb66329a74d40b22f9029fbc50e57248bad7cb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1ce71e80c0cd0e0063eba275ab91f77982f2e92a8aef8e89cab99b93feab0a88
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1b6d934a7b6c42a61e5ad51380e69af341a575b42205299fd58bc0c6ff306f6a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b046aa0a42dae992a03685ed82a20835532d6d4bf048e2366ff7382be4b3138
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acbfee4492327a720fe7da76929a86192d8af2442bbb76f0a089f6a81a890be3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:818f592c50d4ed5599faa053344f7ecd963003352f1faa3b72dc0f5e00a0e78b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4e5f7df1dc1e1db398c36d590431e7e5782897b209972d8e9e4671971c10d1b6
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5458a8fd89aa664b60e59c35bd933975cd37d18e19b11589e4ca0c2844ea0ed4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2a128b570352a5a7f82b27353f9b774caccce83e0e7bb2673e4802d96ad0bd51
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2171e1c75d0c665fe33e1ea334e605604c51de6dbc22d88a3a0ac604259c448f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ec6eb7c28c213cc83b2d7919cd87988f9a07f12276eb7351d0915f1567a5b095
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e5814748c5ab229a1dea5662a007063c0df06fbbfcfe041fd9a6df9ef67e07cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:82d8104a37349c513447e68ea3790efb256370632353176e5fc13c2e8df1bd44
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b78a4d3864e16d270979d9d7756012abf09e6cdb7f1eb4832f6497c26281099
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9f42e929f2cc7e9b6d24dd61e19c5d93dcfeabd34b9c8b935afbe78577efb18b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8fc8de4f3671a73f9aa747e5b23ec973d192fff6f2869c38985cd656063c1e3d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:02e1ef01a023da0eee653f63b314786fd66a78d184a97a314422f6f9c3ef8e7e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:65923c37fecda2d3174212fa9d9a6f34241d065389336ea69a7be99ba16cedc2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bed27f3f2c2d7627d1af4500d7ea0302da5c3fd3d875a6480fb6d98da481312e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:10b9df672260c023dfb37166e156c0d7e0e13144dc3e485af922e8151fd3ab05
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f05824bb8eaa5a93e54059449eaa236dd83d729049ad00065ef3242f5f74c91e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:87db6e079b96fb392d98ef5fc7063836bff10717efd75061016ce688be0ba38e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:29abb169ffefe36f6d69edd438a7135ee23059cedab8f87ad3b2d52361b96b35
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a4f932c3e9ff3aeb70ba1f7a92aa4642bd14cbb7d37d04ff750ed517b916cdb9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08496b84b3cc691909227be17e67bcd6f4d5185a93712f44d8a0985bd2c5fcd7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f53b07212320c6939d3bb4ddd4e6a0bce311ab5b63858eb1a8728531f0dbd510
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b0057de903654ea0315a2c96affcc8f50a92ede8ad121ed30edb3aaf0fb7b554
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d0014b1c19890b9e8ef1a4f3efe7eb5b269be5610ba6ababd4258b52d2f85e8a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:aeac9cd8e8acdbd2b0c339c48b85079c159b04f0ade24df8e3f9cae3f48b4c70
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6366ba29f301a82215c05e60d78e1de66bcee85440ce2fc3c692712d4da32685
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8a70ba12f610ac0e62066e417a0833bc1d01e91ac99d0dd794a2c0cd2a350164
        version: 2.18.2
      entries:
      - name: cockroach-operator.v2.18.2
        version: 2.18.2
      name: stable
    defaultChannel: stable
    packageName: cockroachdb-certified-rhmp
    provider:
      name: Cockroach Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      control-plane: odfmo-controller-manager
      full_version: 4.15.4-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-compose-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-compose-operator.v4.15.4-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "labels": {
                    "control-plane": "odfmo-controller-manager"
                  },
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              }
            ]
          capabilities: Basic Install
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.4-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          This is a CSV used for mirroring the entire product properly. Do not install it please.
        displayName: ODF Compose (DO NOT INSTALL)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: branto@redhat.com
          name: Boris Ranto
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-operator-bundle@sha256:3773158dbd141530cd851f36c3c8a5f54dc0892d003633118e123f5862923666
        - registry.redhat.io/odf4/odr-cluster-operator-bundle@sha256:4c7049b6137b3d49fa9b5e51e0cf7cd430d70c8790e0ed017e6bd811fcc506c8
        - registry.redhat.io/odf4/odf-csi-addons-operator-bundle@sha256:b2b8144c34df667b13987773b11df2fc8ac96d418999292303ef2c03cd3c79d6
        - registry.redhat.io/odf4/mcg-operator-bundle@sha256:9ce0ab75f56ebeeac63aefd0f48cba671da3377ac628830b3c03be9569f8c87a
        - registry.redhat.io/odf4/ocs-operator-bundle@sha256:1e91c5c6c122fdccab52c37ff2cef76317af80970dbdc06aec752115117b2fa0
        - registry.redhat.io/odf4/ocs-client-operator-bundle@sha256:8904c171ea317249e9a19c3a7638f688357a0b238582330b1bddc23a11a9925e
        - registry.redhat.io/odf4/odf-multicluster-operator-bundle@sha256:f7e1c75ee6b1a62fddbcc1e6a718a9cbd5ecf364e4ef10f29da7a15bc82d5b17
        - registry.redhat.io/odf4/odf-multicluster-rhel8-operator@sha256:861e56167698558109197b515aff93691f8d0fdba19311be18a32512f6896139
        - registry.redhat.io/odf4/odr-hub-operator-bundle@sha256:862a9d983013c16b7dfd5194ab5752c26a473ca5a4575bbcd80be59f97f8ac27
        version: 4.15.4-rhodf
      entries:
      - name: odf-compose-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      name: stable-4.15
    defaultChannel: stable-4.15
    packageName: odf-compose-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: unsupported
      operatorframework.io/arch.ppc64le: unsupported
      operatorframework.io/arch.s390x: unsupported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: mta-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mta-operator.v6.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:f5d10cd2528bebf5094c7744e47d9de6fe827fd09024c5ffc0e1b74f1a6cc3b4
          createdAt: "2023-04-26T14:04:20Z"
          description: MTA contains tools that support the modernization and migration
            of applications to Kubernetes and OpenShift
          olm.skipRange: '>=0.0.0 <6.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/tackle2-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is a collection of tools that supports large-scale application modernization and migration projects to Kubernetes and OpenShift.


          Migration Toolkit for Applications allows users to maintain their portfolio of applications with a full set of metadata and to assess their suitability for modernization leveraging a questionnaire based approach.


          ### Install

          By default, MTA requires a RWX capable storage class for caching purposes, this can be disabled via the rwx_supported parameter in Tackle CR.

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-windup-addon-rhel8@sha256:a4200fc0a39df67eae31c53ee9ae0c05769fb410d0312ebd0bbf8bf1567b2a48
        - registry.redhat.io/mta/mta-hub-rhel8@sha256:c75e1d9d3307d95be392a53440d9123601a6ada3693e3f2573e03022565e4339
        - registry.redhat.io/rhel8/postgresql-12@sha256:3a69efa85af6a6f95665eca69085b007fa1fcb645a37f356895aeb4c9527a621
        - registry.redhat.io/mta/mta-pathfinder-rhel8@sha256:a66dd030eb99207b6aaacc6b7c7cb5182901da836ebf50373c14f8e5ecb7b642
        - registry.redhat.io/mta/mta-ui-rhel8@sha256:de6cb95ce3663922ecc725b186876b043a9c7794e95510459ffbcba7d668fc26
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:f5d10cd2528bebf5094c7744e47d9de6fe827fd09024c5ffc0e1b74f1a6cc3b4
        - registry.redhat.io/mta/mta-admin-addon-rhel8@sha256:812b8dc621a06bb99cf2c59abb8257da3ceeb165c563e14d258e69803a557cc8
        version: 6.0.3
      entries:
      - name: mta-operator.v6.0.3
        version: 6.0.3
      - name: mta-operator.v6.0.2
        version: 6.0.2
      - name: mta-operator.v6.0.1
        version: 6.0.1
      - name: mta-operator.v6.0.0
        version: 6.0.0
      name: stable-v6.0
    - currentCSV: mta-operator.v6.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:6da71d4caecfbe049e63576d4325d4919002196da2e875f79703d8ff0d5a1705
          createdAt: "2023-10-26T15:29:26Z"
          description: MTA contains tools that support the modernization and migration
            of applications to Kubernetes and OpenShift
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0 <6.1.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/tackle2-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is a collection of tools that supports large-scale application modernization and migration projects to Kubernetes and OpenShift.


          Migration Toolkit for Applications allows users to maintain their portfolio of applications with a full set of metadata and to assess their suitability for modernization leveraging a questionnaire based approach.


          ### Install

          By default, MTA requires a RWX capable storage class for caching purposes, this can be disabled via the rwx_supported parameter in Tackle CR.

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:6da71d4caecfbe049e63576d4325d4919002196da2e875f79703d8ff0d5a1705
        - registry.redhat.io/mta/mta-hub-rhel8@sha256:32624f2bddfe8b6b9c6c1c3f2be541c8c61be29680d13b6d446e8874bdf8e219
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/mta/mta-pathfinder-rhel8@sha256:1df881da1e20c2501caf53e1cf59b6b366281d842f0948ca5f91694ec1db829d
        - registry.redhat.io/mta/mta-ui-rhel8@sha256:3845d6d246003ebeb49bcf1d95d6be2eb9cd05a280a5bf8320b201e878a41427
        - registry.redhat.io/mta/mta-windup-addon-rhel8@sha256:6bcaf8e8b63e169652d6568d77e91a289084e621a586105a2d429b5726569058
        version: 6.1.4
      entries:
      - name: mta-operator.v6.1.4
        version: 6.1.4
      - name: mta-operator.v6.1.3
        version: 6.1.3
      - name: mta-operator.v6.1.2
        version: 6.1.2
      - name: mta-operator.v6.1.1
        version: 6.1.1
      - name: mta-operator.v6.1.0
        version: 6.1.0
      name: stable-v6.1
    - currentCSV: mta-operator.v6.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:6938c3fc6dd8fb032e1765fc11d9586dbfd1aaba4b423b0d87da4120ee0bb60d
          createdAt: "2024-06-10T22:12:58Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0 <6.2.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-windup-addon-rhel9@sha256:d1f4939b92dd13c15cb7344ec47a571669f6d108dc2e7d795ce29350f2cb3838
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:6938c3fc6dd8fb032e1765fc11d9586dbfd1aaba4b423b0d87da4120ee0bb60d
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:e59725832547e80cab58077a1b6e42c313a4057621defa841f35c0a7627f149d
        - registry.redhat.io/rhel8/postgresql-12@sha256:c4716bee5e2cf2e43698bc99e5aa42f16a0f2d68c1fbdcd3602696bf88e408dc
        - registry.redhat.io/mta/mta-pathfinder-rhel9@sha256:a2a8b6d204d99494a6a005e8cb1fbd92afdc84dc01e4c1de4ad7d8798bd83741
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:c35f3ed429a7d351cb8c2d07600b4ad35615a5ba08e5cc9f8b9bd504aea3f7ee
        version: 6.2.3
      entries:
      - name: mta-operator.v6.2.3
        version: 6.2.3
      - name: mta-operator.v6.2.2
        version: 6.2.2
      - name: mta-operator.v6.2.1
        version: 6.2.1
      - name: mta-operator.v6.2.0
        version: 6.2.0
      name: stable-v6.2
    - currentCSV: mta-operator.v7.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:268ecf5efc51af71f9b2bc2d01b410f8cfcf81a948b82d2d476fef6c0cd3e033
          createdAt: "2024-05-10T19:37:30Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:268ecf5efc51af71f9b2bc2d01b410f8cfcf81a948b82d2d476fef6c0cd3e033
        - registry.redhat.io/mta/mta-windup-shim-rhel9@sha256:e2ea89055936d7ad229b5e890c6ba234c6cb22d8b688515ed4d6a34848efea83
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:d554cc36748fc05e30eb10e6fe816ff6aab2fe07b61cc60e9d93de5c3cf17748
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:0e1ec99af5666a060176adc52612f20209e3398a9a5a2aa1dcff11ef260b2006
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:1b5e016046d93b6a2674d63f08e0a46f8050e785bd58fd94ba7a70487efe7e2a
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:18c611415589a093e16e9c2f2fba9f3c83d9514f79c8e4fde9704ef0679ea4d5
        - registry.redhat.io/mta/mta-analyzer-lsp-rhel9@sha256:0c8f59f35a950ea8b45f16cf71362128ccc335b58a996a099384fd737dd6ff90
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:af879d8b9b5e396a5308fb675a1d6dba6d70151b9bdbe7829362e081a6adab1c
        - registry.redhat.io/rhel8/postgresql-12@sha256:da619f6617d99790b4fe00fdb23ab4aea0cdb0592590162d51b3ef94aca01a7b
        version: 7.0.3
      entries:
      - name: mta-operator.v7.0.3
        version: 7.0.3
      - name: mta-operator.v7.0.2
        version: 7.0.2
      - name: mta-operator.v7.0.1
        version: 7.0.1
      - name: mta-operator.v7.0.0
        version: 7.0.0
      name: stable-v7.0
    - currentCSV: mta-operator.v7.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:08b33b44140397785f8aa1b736796e09cf6e49ac30cf7c386a2fbd8b678d47b1
          createdAt: "2024-09-16T23:45:42Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.1.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:16cb28d683159ba71a5d7c233404ac1d6b7a07a374b46d24effe551a13a596c3
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:23c97ac122eb7cd1e953df8da968da420f6db8708baf772759517aca4929748b
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:e4e70201df23c9e1168646d3c6013a8ffa22481a85b07df491f240310c13ebe1
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:08b33b44140397785f8aa1b736796e09cf6e49ac30cf7c386a2fbd8b678d47b1
        - registry.redhat.io/rhel8/postgresql-15@sha256:16faf1ee82382263408f713ee4002f47ff13c82d4545e8a69d5b0ebd1d289330
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:30a3b1d5a70dec74a9253ca34f18bd8f2cdfaa68b85b0b25dc395158c2596330
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel8@sha256:582f28e609c1a0466d30c5586d6b6023f1b1de3c28f76cabb498322914ab24d4
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:ac38f1c3d34bc718928eb20743835c99783d846f63a3197d13839ea22f1634c8
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:4ad19b50cdd6be21e010155f2026abf0dc865e72aac1353e35bbb5eeb8320a4e
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:ee2c8ce2af0be097353758d9ee7eaa7e6b3fcd5f2bf5e04280e450e2b619d2d8
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        version: 7.1.1
      entries:
      - name: mta-operator.v7.1.1
        version: 7.1.1
      - name: mta-operator.v7.1.0
        version: 7.1.0
      name: stable-v7.1
    - currentCSV: mta-operator.v7.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:917d6cd534c158f6c4a7e9c83f0c66fa85ba81037c352d8a039e198581126a7f
          createdAt: "2025-06-17T16:04:15Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.2.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:917d6cd534c158f6c4a7e9c83f0c66fa85ba81037c352d8a039e198581126a7f
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel8@sha256:c0643c05833dd1b70ed361bf56c46f56ed6536aac3106b92f0d85b83da1b9d15
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:10ef4b218d603432508b5e1f3685fb32a80cbce6c327e35952cdcccb0465468b
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:9d6c02f6463771e747a30126de95fc0728fd95fff4ddfa6c70d828c55e0b6790
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:2dfc51d5398a90e73273b3e332f70d7ed94bac05e12b142fba22104ecc929436
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:b10fdc78c6f453a34a3d306313dc29fcb3152688a6a1d54cd236d00d9fab88af
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:4cde2c2df482293bd4d6b86a344336e13d2f9f81b29323698f603e656b14b190
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:70ff3b5e429a8fe0c68752ee05c47ead885b59bf9c8d7235337801270e7fc1de
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:c9447659195b9eebb8f8a265fd5052ca9b2611d57080017448297a33409863c6
        - registry.redhat.io/rhel8/postgresql-15@sha256:dd4aec2775da1cf187c0839384cb16d69268e9781d4330874bbc865f98153a7a
        version: 7.2.3
      entries:
      - name: mta-operator.v7.2.3
        version: 7.2.3
      - name: mta-operator.v7.2.2
        version: 7.2.2
      - name: mta-operator.v7.2.1
        version: 7.2.1
      - name: mta-operator.v7.2.0
        version: 7.2.0
      name: stable-v7.2
    - currentCSV: mta-operator.v7.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel9-operator@sha256:6abd6541a955543207c1ff4b903928debcc89dd6ea3d7aeaa3e47465849f7b37
          createdAt: "2025-08-15T16:32:16Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.3.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-15@sha256:854710b428c4c1132c448cbe9ad30087a8a4b3c33ae9c0c3a4de21a15c0ae642
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel9@sha256:9748d954ad8654ac203257286497ccf19cfdfa211fee40449cbe732ff48c1176
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:32328fc0459099ef1de0f92e064c15d29a46d88dfe2f54c88f822d44ab82e3f4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        - registry.redhat.io/mta/mta-rhel9-operator@sha256:6abd6541a955543207c1ff4b903928debcc89dd6ea3d7aeaa3e47465849f7b37
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:b2ff154a4b4ead22311005b7115c02346928b5ebf77d776ef909ad4af1f350bc
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:6118c1db15667851f26eb2acf6060e045ed9065459e3e649bbdd0e2625d825bd
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:072f676323e9abcb2ff22eb5e7a70f6681e9d885d8666ddd81060fbad54c443e
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:8ca00460b2fb64583a247f5b906c19743b7732b60ff8215f1baec12a766c61d1
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:80aa13813fe121da49e166d81d09d63749393f3375c3b96d762d0b09c6685ca9
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:ef57a58e609e244333d35681a164a1d3bf382c0eee37d33c993b018439814198
        version: 7.3.2
      entries:
      - name: mta-operator.v7.3.2
        version: 7.3.2
      - name: mta-operator.v7.3.1
        version: 7.3.1
      - name: mta-operator.v7.3.0
        version: 7.3.0
      name: stable-v7.3
    - currentCSV: mta-operator.v8.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel9-operator@sha256:83c2b804667d28f782d64132a0c393194083d10c820b88df49c2503d3351fd8b
          createdAt: "2025-09-26T03:03:31Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.3.0 <8.0.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Schema
            displayName: Schema
            kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-platform-addon-rhel9@sha256:dcb35178810b7eddc2f2713434d377003f8f4fb7022a40f6c1926b74b13fe23a
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:a77ee9b76225b6cca667fa3402492b0845587fe5e400fa8d141b5ab636c56811
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel9@sha256:2a6f0b392b86fe16244d5bf9fcb84ed9770a22329a96e9861f03872b7209fb04
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ef10b2ecbfa0c4ac8568af9ee64ee4a554e3fdc9089ee5683b6c2c3a74caafbe
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:616d04cfbab92341157c42eee0683f31867d81c5a289723da8da8208d1054774
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:aaf7bf10b1af107ad55289caab2cf9dd052e69211968d2e8d05c1d25f21ed2b3
        - registry.redhat.io/mta/mta-rhel9-operator@sha256:83c2b804667d28f782d64132a0c393194083d10c820b88df49c2503d3351fd8b
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:b1b38e218ea37b529f344eaa1f0f3b55b716ad7953e852c19f34f4a25ab6c5a7
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:af2c1790f2b98ebcfadb5d6bdd43cb9e7e96cdabd5b653d48d0fb67901781c49
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:460d1d45ed44ae9328ed93c5fed240e154549bbec662dc1fc55569ff7256aca9
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:8ae0887fa27d6f4281e5bb9b629bf6b64a9c460948a002839ba42da9b25b5521
        - registry.redhat.io/mta/mta-solution-server-rhel9@sha256:e9d48d11ae203108229d67bc96721daa1c3bb10160b9d6d957cd2e0ba68a2ef7
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        version: 8.0.0
      entries:
      - name: mta-operator.v8.0.0
        version: 8.0.0
      name: stable-v8.0
    defaultChannel: stable-v8.0
    packageName: mta-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: crunchy-postgres-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/crunchy-postgres-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/crunchy-postgres-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-exporter@sha256:3b231902b2116cd38de05c7378202716c95813b9961fb8cee811b61d36cddf86
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:0e283616c2d32e53258d62d02702bc4e179ef8c03933592154babed1a70bf666
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:a7f62f367817fc08294a0b8b82ac9826a42c0ae668efdfacaa158a4287b8a688
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:c559c896e174674d151f93543ac92a1b5647eb4676313f7dc6cbb8faee786f18
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:56a5e3a5758982e086c9f78201a7137e0d934316f058a31cf30df44639a569a2
        - registry.connect.redhat.com/crunchydata/crunchy-pgadmin4@sha256:0d84acb1ce6a9b046e3f8b222a583ddcc91dda0a117c9f637ef173d0ac992f57
        - registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:b2c3c74fdcc0e08f555e50c527f0d322292333c5d3dce061a0dee3680a1fbc56
        - registry.connect.redhat.com/crunchydata/crunchy-pgbackrest@sha256:b70da0035b74379f2507cfabdd9d1cc1e3c3eecb9066c22f292981c9a27916e1
        - registry.connect.redhat.com/crunchydata/crunchy-pgbouncer@sha256:17cb58dec78447c2f70a414a2aeff6facbf921b42a0f366d0dbf7be95aa8987a
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:4bdd0588ffbb3f628e7df1c72670e7a5087269438343b68779012bc3a13dea31
        - registry.connect.redhat.com/crunchydata/crunchy-upgrade@sha256:68bc93cce61d9b1e44822ae3e3ac8f7a9c2b712eddfaea090eabefacd60c67f6
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: crunchy-postgres-operator-rhmp
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-networkfirewall-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-networkfirewall-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "Firewall",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "FirewallPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "RuleGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/networkfirewall-controller:1.1.1
          createdAt: "2025-09-19T21:03:12Z"
          description: AWS Network Firewall controller is a service controller for
            managing Network Firewall resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FirewallPolicy represents the state of an AWS networkfirewall
              FirewallPolicy resource.
            displayName: FirewallPolicy
            kind: FirewallPolicy
            name: firewallpolicies.networkfirewall.services.k8s.aws
            version: v1alpha1
          - description: Firewall represents the state of an AWS networkfirewall Firewall
              resource.
            displayName: Firewall
            kind: Firewall
            name: firewalls.networkfirewall.services.k8s.aws
            version: v1alpha1
          - description: RuleGroup represents the state of an AWS networkfirewall
              RuleGroup resource.
            displayName: RuleGroup
            kind: RuleGroup
            name: rulegroups.networkfirewall.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Network Firewall resources in AWS from within your Kubernetes cluster.

          **About Amazon Network Firewall**

          AWS Network Firewall is a stateful, managed, network firewall and intrusion detection and prevention service for your virtual private cloud (VPC) that you create in Amazon Virtual Private Cloud (Amazon VPC).
          With Network Firewall, you can filter traffic at the perimeter of your VPC. This includes filtering traffic going to and coming from an internet gateway, NAT gateway, or over VPN or AWS Direct Connect. Network Firewall uses the open source intrusion prevention system (IPS), Suricata, for stateful inspection. Network Firewall supports Suricata compatible rules. For more information, see [Working with stateful rule groups in AWS Network Firewall](https://docs.aws.amazon.com/network-firewall/latest/developerguide/stateful-rule-groups-ips.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Network Firewall
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networkfirewall
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Network Firewall Developer Resources
          url: https://aws.amazon.com/network-firewall/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: network firewall maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/networkfirewall-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-networkfirewall-controller.v1.1.1
        version: 1.1.1
      - name: ack-networkfirewall-controller.v1.1.0
        version: 1.1.0
      - name: ack-networkfirewall-controller.v1.0.13
        version: 1.0.13
      - name: ack-networkfirewall-controller.v1.0.12
        version: 1.0.12
      - name: ack-networkfirewall-controller.v1.0.11
        version: 1.0.11
      - name: ack-networkfirewall-controller.v1.0.10
        version: 1.0.10
      - name: ack-networkfirewall-controller.v1.0.9
        version: 1.0.9
      - name: ack-networkfirewall-controller.v1.0.8
        version: 1.0.8
      - name: ack-networkfirewall-controller.v1.0.7
        version: 1.0.7
      - name: ack-networkfirewall-controller.v1.0.6
        version: 1.0.6
      - name: ack-networkfirewall-controller.v1.0.5
        version: 1.0.5
      - name: ack-networkfirewall-controller.v1.0.4
        version: 1.0.4
      - name: ack-networkfirewall-controller.v1.0.1
        version: 1.0.1
      - name: ack-networkfirewall-controller.v1.0.0
        version: 1.0.0
      - name: ack-networkfirewall-controller.v0.0.10
        version: 0.0.10
      - name: ack-networkfirewall-controller.v0.0.9
        version: 0.0.9
      - name: ack-networkfirewall-controller.v0.0.8
        version: 0.0.8
      - name: ack-networkfirewall-controller.v0.0.7
        version: 0.0.7
      - name: ack-networkfirewall-controller.v0.0.6
        version: 0.0.6
      - name: ack-networkfirewall-controller.v0.0.5
        version: 0.0.5
      - name: ack-networkfirewall-controller.v0.0.2
        version: 0.0.2
      - name: ack-networkfirewall-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-networkfirewall-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: opdev.github.io
    name: spark-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spark-helm-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apachespark.apache.org/v1alpha1",
                "kind": "Spark",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/install-disable-crds": "true"
                  },
                  "name": "spark-operator"
                },
                "spec": {
                  "affinity": {},
                  "batchScheduler": {
                    "enable": false
                  },
                  "controllerThreads": 10,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/googlecloudplatform/spark-operator",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "ingressUrlFormat": "",
                  "istio": {
                    "enabled": false
                  },
                  "labelSelectorFilter": "",
                  "leaderElection": {
                    "lockName": "spark-operator-lock",
                    "lockNamespace": ""
                  },
                  "logLevel": 2,
                  "metrics": {
                    "enable": true,
                    "endpoint": "/metrics",
                    "port": 10254,
                    "portName": "metrics",
                    "prefix": ""
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podMonitor": {
                    "enable": false,
                    "jobLabel": "spark-operator-podmonitor",
                    "labels": {},
                    "podMetricsEndpoint": {
                      "interval": "5s",
                      "scheme": "http"
                    }
                  },
                  "podSecurityContext": {},
                  "rbac": {
                    "create": false,
                    "createClusterRole": true,
                    "createRole": true
                  },
                  "replicaCount": 1,
                  "resourceQuotaEnforcement": {
                    "enable": false
                  },
                  "resources": {},
                  "resyncInterval": 30,
                  "securityContext": {},
                  "serviceAccounts": {
                    "spark": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "sparkoperator": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    }
                  },
                  "sparkJobNamespace": "",
                  "tolerations": [],
                  "uiService": {
                    "enable": true
                  },
                  "volumeMounts": [],
                  "volumes": [],
                  "webhook": {
                    "cleanupAnnotations": {
                      "helm.sh/hook": "pre-delete, pre-upgrade",
                      "helm.sh/hook-delete-policy": "hook-succeeded"
                    },
                    "enable": false,
                    "initAnnotations": {
                      "helm.sh/hook": "pre-install, pre-upgrade",
                      "helm.sh/hook-weight": "50"
                    },
                    "namespaceSelector": "",
                    "port": 8080,
                    "timeout": 30
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "ScheduledSparkApplication",
                "metadata": {
                  "name": "spark-pi-scheduled",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "concurrencyPolicy": "Allow",
                  "schedule": "@every 5m",
                  "template": {
                    "driver": {
                      "coreLimit": "1200m",
                      "cores": 1,
                      "labels": {
                        "version": "3.1.1"
                      },
                      "memory": "512m",
                      "serviceAccount": "spark-operator-spark",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "executor": {
                      "cores": 1,
                      "instances": 1,
                      "labels": {
                        "version": "3.1.1"
                      },
                      "memory": "512m",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "image": "gcr.io/spark-operator/spark:v3.1.1",
                    "imagePullPolicy": "Always",
                    "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.12-3.1.1.jar",
                    "mainClass": "org.apache.spark.examples.SparkPi",
                    "mode": "cluster",
                    "restartPolicy": {
                      "type": "Never"
                    },
                    "sparkVersion": "3.1.1",
                    "type": "Scala",
                    "volumes": [
                      {
                        "hostPath": {
                          "path": "/tmp",
                          "type": "Directory"
                        },
                        "name": "test-volume"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "pyspark-pi",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "driver": {
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "serviceAccount": "spark-operator-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "gcr.io/spark-operator/spark-py:v3.1.1",
                  "imagePullPolicy": "Always",
                  "mainApplicationFile": "local:///opt/spark/examples/src/main/python/pi.py",
                  "mode": "cluster",
                  "pythonVersion": "3",
                  "restartPolicy": {
                    "onFailureRetries": 3,
                    "onFailureRetryInterval": 10,
                    "onSubmissionFailureRetries": 5,
                    "onSubmissionFailureRetryInterval": 20,
                    "type": "OnFailure"
                  },
                  "sparkVersion": "3.1.1",
                  "type": "Python",
                  "volumes": [
                    {
                      "hostPath": {
                        "path": "/tmp",
                        "type": "Directory"
                      },
                      "name": "test-volume"
                    }
                  ]
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "spark-pi",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "driver": {
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "serviceAccount": "spark-operator-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "gcr.io/spark-operator/spark:v3.1.1",
                  "imagePullPolicy": "Always",
                  "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.12-3.1.1.jar",
                  "mainClass": "org.apache.spark.examples.SparkPi",
                  "mode": "cluster",
                  "restartPolicy": {
                    "type": "Never"
                  },
                  "sparkVersion": "3.1.1",
                  "type": "Scala",
                  "volumes": [
                    {
                      "hostPath": {
                        "path": "/tmp",
                        "type": "Directory"
                      },
                      "name": "test-volume"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/spark-helm-operator-controller:0.0.2
          createdAt: "2023-07-14T17:57:44Z"
          description: An experimental operator that installs the spark on k8s operator
            from its helm chart.
          operatorframework.io/initialization-resource: "{\n  \"apiVersion\": \"apachespark.apache.org/v1alpha1\",\n
            \ \"kind\": \"Spark\",\n  \"metadata\": {\n    \"name\": \"spark-operator\"\n
            \ },\n  \"spec\": {\n    \"affinity\": {},\n    \"batchScheduler\": {\n
            \     \"enable\": false\n    },\n    \"controllerThreads\": 10,\n    \"fullnameOverride\":
            \"\",\n    \"image\": {\n      \"pullPolicy\": \"IfNotPresent\",\n      \"repository\":
            \"ghcr.io/googlecloudplatform/spark-operator\",\n      \"tag\": \"\"\n
            \   },\n    \"imagePullSecrets\": [],\n    \"ingressUrlFormat\": \"\",\n
            \   \"istio\": {\n      \"enabled\": false\n    },\n    \"labelSelectorFilter\":
            \"\",\n    \"leaderElection\": {\n      \"lockName\": \"spark-operator-lock\",\n
            \     \"lockNamespace\": \"\"\n    },\n    \"logLevel\": 2,\n    \"metrics\":
            {\n      \"enable\": true,\n      \"endpoint\": \"/metrics\",\n      \"port\":
            10254,\n      \"portName\": \"metrics\",\n      \"prefix\": \"\"\n    },\n
            \   \"nameOverride\": \"\",\n    \"nodeSelector\": {},\n    \"podAnnotations\":
            {},\n    \"podLabels\": {},\n    \"podMonitor\": {\n      \"enable\":
            false,\n      \"jobLabel\": \"spark-operator-podmonitor\",\n      \"labels\":
            {},\n      \"podMetricsEndpoint\": {\n        \"interval\": \"5s\",\n
            \       \"scheme\": \"http\"\n      }\n    },\n    \"podSecurityContext\":
            {},\n    \"rbac\": {\n      \"create\": false,\n      \"createClusterRole\":
            true,\n      \"createRole\": true\n    },\n    \"replicaCount\": 1,\n
            \   \"resourceQuotaEnforcement\": {\n      \"enable\": false\n    },\n
            \   \"resources\": {},\n    \"resyncInterval\": 30,\n    \"securityContext\":
            {},\n    \"serviceAccounts\": {\n      \"spark\": {\n        \"annotations\":
            {},\n        \"create\": true,\n        \"name\": \"\"\n      },\n      \"sparkoperator\":
            {\n        \"annotations\": {},\n        \"create\": true,\n        \"name\":
            \"\"\n      }\n    },\n    \"sparkJobNamespace\": \"\",\n    \"tolerations\":
            [],\n    \"uiService\": {\n      \"enable\": true\n    },\n    \"volumeMounts\":
            [],\n    \"volumes\": [],\n    \"webhook\": {\n      \"cleanupAnnotations\":
            {\n        \"helm.sh/hook\": \"pre-delete, pre-upgrade\",\n        \"helm.sh/hook-delete-policy\":
            \"hook-succeeded\"\n      },\n      \"enable\": false,\n      \"initAnnotations\":
            {\n        \"helm.sh/hook\": \"pre-install, pre-upgrade\",\n        \"helm.sh/hook-weight\":
            \"50\"\n      },\n      \"namespaceSelector\": \"\",\n      \"port\":
            8080,\n      \"timeout\": 30\n    }\n  }\n}    "
          operatorframework.io/suggested-namespace: spark-operator
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys a scheduled spark application based on the spec provided
            displayName: Scheduled Spark Application
            kind: ScheduledSparkApplication
            name: scheduledsparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - description: Deploys a spark application based on the spec provided
            displayName: Spark Application
            kind: SparkApplication
            name: sparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - description: Deploys the spark operator based on the google cloud platform
              helm chart
            displayName: Spark Operator
            kind: Spark
            name: sparks.apachespark.apache.org
            version: v1alpha1
        description: An experimental helm based operator that wraps the spark on k8s
          helm chart from google cloud platform enabling installation and workload
          management of Apache Spark
        displayName: Spark Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - big data
        - map reduce
        - data science
        links:
        - name: Spark Helm Operator
          url: https://github.com/skattoju/spark-helm-operator
        maintainers:
        - email: skattoju@redhat.com
          name: Sid Kattoju
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: opdev.github.io
        relatedImages:
        - quay.io/opdev/spark-helm-operator-controller:0.0.2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 0.0.2
      entries:
      - name: spark-helm-operator.v0.0.2
        version: 0.0.2
      - name: spark-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: spark-helm-operator
    provider:
      name: opdev
      url: opdev.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-test-core
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-test-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Testcore",
                "metadata": {
                  "name": "testcore-sample"
                },
                "spec": {
                  "agentController": {
                    "port": 8080
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "dataseqService": {
                    "ingressEnabled": true,
                    "port": 8080,
                    "providers": {
                      "loadCore": {
                        "enabled": true
                      }
                    }
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "ingress": {
                    "backwardsCompatibilityEnabled": true
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "systemMonitor": {
                    "enabled": false
                  },
                  "trafficController": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:35:34Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Testcore
            name: testcores.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Test Core Operator
        displayName: Keysight WAP Test Core Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Test Core
          url: https://keysight-wap-test-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-test-core@sha256:9e2d81061abff49112f908e71dace03a748e283f32597aa22ad58471fff30caf
        version: 0.1.21
      entries:
      - name: keysight-wap-test-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-test-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp
      provider-url: https://www.netapp.com/
    name: netapp-spark-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netapp-spark-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.netapp.com/v1alpha1",
                "kind": "SparkOperator",
                "metadata": {
                  "name": "sparkoperator-sample"
                },
                "spec": {
                  "spark-operator": {
                    "affinity": {},
                    "controllerThreads": 20,
                    "disableExecutorReporting": false,
                    "hostNetwork": false,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "public.ecr.aws/f4k1p1n4/spark-operator",
                      "tag": "108810d-ubi9-rel20240916"
                    },
                    "imagePullSecrets": [],
                    "istio": {
                      "enabled": true
                    },
                    "jobAffinity": {},
                    "podLabels": {},
                    "pprof": {
                      "enable": true,
                      "port": 6060,
                      "portName": "pprof"
                    },
                    "rbac": {
                      "createClusterRole": true,
                      "createRole": true
                    },
                    "resources": {
                      "requests": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      }
                    },
                    "securityContext": {},
                    "serviceAccounts": {
                      "spark": {
                        "create": true
                      }
                    },
                    "sparkJobNamespace": "",
                    "tolerations": [],
                    "uiService": {
                      "enable": false
                    },
                    "webhook": {
                      "enable": true,
                      "jobTTLSecondsAfterFinished": 60,
                      "port": 9443
                    },
                    "workqueue": {
                      "maxDelay": {
                        "enabled": false,
                        "value": "6h"
                      },
                      "tokenBucketSize": 500,
                      "tokenRefillRate": 50
                    }
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "ScheduledSparkApplication",
                "metadata": {
                  "name": "spark-pi-scheduled"
                },
                "spec": {
                  "concurrencyPolicy": "Allow",
                  "schedule": "@every 5m",
                  "template": {
                    "arguments": [
                      "1000"
                    ],
                    "driver": {
                      "annotations": {
                        "openshift.io/required-scc": "sparkoperator-sample-spark"
                      },
                      "coreLimit": "1200m",
                      "cores": 1,
                      "labels": {
                        "version": "3.5.0"
                      },
                      "memory": "512m",
                      "serviceAccount": "sparkoperator-sample-spark",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "executor": {
                      "annotations": {
                        "openshift.io/required-scc": "sparkoperator-sample-spark"
                      },
                      "cores": 1,
                      "instances": 1,
                      "labels": {
                        "version": "3.5.0"
                      },
                      "memory": "512m",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "image": "public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce",
                    "imagePullPolicy": "IfNotPresent",
                    "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.13-3.5.0.jar",
                    "mainClass": "org.apache.spark.examples.SparkPi",
                    "mode": "cluster",
                    "restartPolicy": {
                      "type": "Never"
                    },
                    "sparkVersion": "3.5.0",
                    "type": "Scala",
                    "volumes": [
                      {
                        "emptyDir": {},
                        "name": "test-volume"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "spark-pi"
                },
                "spec": {
                  "arguments": [
                    "1000"
                  ],
                  "driver": {
                    "annotations": {
                      "openshift.io/required-scc": "sparkoperator-sample-spark"
                    },
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.5.0"
                    },
                    "memory": "512m",
                    "serviceAccount": "sparkoperator-sample-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "annotations": {
                      "openshift.io/required-scc": "sparkoperator-sample-spark"
                    },
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.5.0"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce",
                  "imagePullPolicy": "IfNotPresent",
                  "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.13-3.5.0.jar",
                  "mainClass": "org.apache.spark.examples.SparkPi",
                  "mode": "cluster",
                  "restartPolicy": {
                    "type": "Never"
                  },
                  "sparkVersion": "3.5.0",
                  "type": "Scala",
                  "volumes": [
                    {
                      "emptyDir": {},
                      "name": "test-volume"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Big Data, Integration & Delivery, AI/Machine Learning, Streaming
            & Messaging
          com.redhat.openshift.versions: v4.14
          containerImage: public.ecr.aws/f4k1p1n4/netapp-spark-operator-controller@sha256:9fe1d0793f11b1c525a9a92cbcde6f4bc1bf3c887617a8496bed3a5a0008a67e
          createdAt: "2024-10-21T13:47:47Z"
          description: Red Hat and NetApp have partnered to offer enterprise-grade
            supported Apache Spark on OpenShift.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ScheduledSparkApplication
            name: scheduledsparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - kind: SparkApplication
            name: sparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - kind: SparkOperator
            name: sparkoperators.charts.netapp.com
            version: v1alpha1
        description: |-
          # NetApp Supported Apache Spark on OpenShift

          Red Hat and NetApp have partnered to offer enterprise-grade supported Apache Spark on OpenShift.
          Use this certified operator to launch Apache Spark applications on OpenShift.

          ## Documentation

          Please find the official documentation at https://www.instaclustr.com/support/documentation/apache-spark/apache-spark-on-red-hat-openshift/getting-started/

          ## Support

          With questions about Apache Spark, contact NetApp's Instaclustr support team on the [web](https://support.instaclustr.com/) or by email at support@instaclustr.com

          With questions about Operators in the OpenShift OperatorHub, contact Red Hat's support team on the [web](https://www.redhat.com/en/services/support) or by email at customerservice@redhat.com

          With questions about pricing, email ng-netapp-apache-spark-support-redhat-openshift@netapp.com
        displayName: NetApp Supported Apache Spark on OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Spark
        - Big Data
        - Apache Spark
        - AI
        - ML
        - Data Engineering
        links:
        - name: NetApp Supported Apache Spark on OpenShift
          url: https://www.instaclustr.com/support-solutions
        maintainers:
        - email: thorsteinn.sigurdsson@netapp.com
          name: Thorsteinn Sigurdsson
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: NetApp
          url: https://www.netapp.com/
        relatedImages:
        - public.ecr.aws/f4k1p1n4/netapp-spark-operator-controller@sha256:9fe1d0793f11b1c525a9a92cbcde6f4bc1bf3c887617a8496bed3a5a0008a67e
        - public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce
        version: 1.1.3
      entries:
      - name: netapp-spark-operator.v1.1.3
        version: 1.1.3
      name: alpha
    defaultChannel: alpha
    packageName: netapp-spark-operator
    provider:
      name: NetApp
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Trilio
      provider-url: ""
    name: k8s-triliovault
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k8s-triliovault-stable.3.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/trilio/k8s-triliovault-operator@sha256:0e3094945427231ecbb884b903dc4d347dfaa73e5a04eb4c5cbaed6fce1cbbb8
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          olm.skipRange: '>=3.0.0 <3.1.3'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          TrilioVault for Kubernetes is an enterprise-grade, cloud-native platform purpose built for data protection and management of Kubernetes applications for IT managers, administrators and developers. TrilioVault supports upstream Kubernetes and OpenShift environments and offers the following features:
            * Application-Centric - protects both data and metadata for Helm, Operator or custom Label based applications.
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any Storage (CSI, NFS, S3), or any Cloud (Private or Public).
            * Application Deployment & Tooling: Helm, Operators, Labels, Prometheus, Fluentd.
          Tutorials
          ------------
          Please click the link below to access the TrilioVault for Kubernetes "How-To" series for deployment, best practice and use-case videos
          <a href="https://www.trilio.io/triliovault-for-kubernetes">TrilioVault for Kubernetes - Tutorials and How-To </a>”
          Licensing
          --------
          Customers can download a 30-day Free Trial or a 10-node Basic Edition at no cost.  You can also connect with the Trilio team for an Enterprise Edition license with no limitations and Premium Support.
          For more information on license plans please vist:
          <a href="www.trilio.io/plans"> Trilio Vault for Kubernetes licensing and plans </a>
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our TrilioVault technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. www.trilio.io and @triliodata on Twitter.
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://www.trilio.io/
        - name: Product Licensing
          url: https://www.trilio.io/plans
        - name: Tutorials
          url: https://www.trilio.io/triliovault-for-kubernetes/
        - name: Support for Free/Basic Users
          url: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Trilio
        relatedImages:
        - registry.connect.redhat.com/trilio/k8s-triliovault-operator@sha256:0e3094945427231ecbb884b903dc4d347dfaa73e5a04eb4c5cbaed6fce1cbbb8
        - registry.connect.redhat.com/trilio/trilio-webhook-init@sha256:c45e9b6b75862451d873d3eba29d6bd11cafc64823855f82ccb8c0b3d779ebb0
        - registry.connect.redhat.com/trilio/web@sha256:76865d8afa8059d40810cb94df29420950a7593a40e87d142d771c57b3319e1d
        - registry.connect.redhat.com/trilio/minio@sha256:cee0f976edc79ff9e7a3f3574a7115dd960042241f3ed5b4636240ad609f9bca
        - registry.connect.redhat.com/trilio/control-plane@sha256:01e36ed5b07104865f922d2996d2211b2b6000adc90266197d9bac74d908db97
        - registry.connect.redhat.com/trilio/datamover@sha256:c7bc15f8217899ed73a812d278eb118633edbee1bf284132054b6f9ca2afd801
        - registry.connect.redhat.com/trilio/dex@sha256:b6331cf3cb3a6a66d9b9ea639cda3c2559fbf1f92f7174032352c1c7bb5176d9
        - registry.connect.redhat.com/trilio/nats@sha256:47396bfcff3e65efb3410fa8369f11f20d794ab543f9ada87b451448b1813178
        - registry.connect.redhat.com/trilio/event-stack@sha256:37c33ae1adddfe739c98fd2b2183c49ec7cfdafc88dec2970ad7c670016ebaac
        version: 3.1.3
      entries:
      - name: k8s-triliovault-stable.3.1.3
        version: 3.1.3
      - name: k8s-triliovault-stable.3.1.2
        version: 3.1.2
      - name: k8s-triliovault-stable.3.1.1
        version: 3.1.1
      - name: k8s-triliovault-stable.3.1.0
        version: 3.1.0
      name: 3.1.x
    - currentCSV: k8s-triliovault-stable.4.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:69871613ceecefff7f691f21bb0b9147236af5a5dc2c7b36ad5e0cd8551f0e6a
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <4.0.5'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/event-stack@sha256:c793418df5a1fc1a49f629d335f4684d9cde1a5f2b462f7c9483ffba904456d2
        - quay.io/triliodata/control-plane@sha256:53368a72a60a38d3ce4293236a841346da6da2380ffd2525e5ffdb8dbcf11c7f
        - quay.io/triliodata/web@sha256:20de1a2f9477ad238bf24f48c8c16a38521c3a86164f2eb6b10d20f09839016e
        - quay.io/triliodata/datamover@sha256:e6848d94ba9f81843bde882a030b0f2556aae4bf16d79e2c5b199417cf626c85
        - quay.io/triliodata/k8s-triliovault-operator@sha256:69871613ceecefff7f691f21bb0b9147236af5a5dc2c7b36ad5e0cd8551f0e6a
        - quay.io/triliodata/operator-webhook-init@sha256:aa85b787f7f37a71e909f049d2abdda746cdd0627e2f74aaf7476d5e5a7f4401
        - quay.io/triliodata/dex@sha256:b4b2372fd03be586b9ff8cd269b9001797d813a34155ab37de31a44fe3cab1ea
        - quay.io/triliodata/nats@sha256:e6723595bcc5674b6201091d45d711d29a5cd1421c8b97c76cf00e468379fddf
        version: 4.0.5
      entries:
      - name: k8s-triliovault-stable.4.0.5
        version: 4.0.5
      - name: k8s-triliovault-stable.4.0.4
        version: 4.0.4
      - name: k8s-triliovault-stable.4.0.3
        version: 4.0.3
      - name: k8s-triliovault-stable.4.0.2
        version: 4.0.2
      - name: k8s-triliovault-stable.4.0.1
        version: 4.0.1
      - name: k8s-triliovault-stable.4.0.0
        version: 4.0.0
      name: 4.0.x
    - currentCSV: k8s-triliovault-stable.5.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:8d582842dbef1c64bbe3b0ce5eeb64790910bdc0416c5e5561557642dede7022
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <5.0.4'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/web@sha256:04c4baadae9c05f4add098ff62774d00102d56581e261626840d17f74effe438
        - quay.io/triliodata/datamover@sha256:104bdca9513c78f81c4340e9a3219df0a3bc8a5b026e4a252363a7e2c10977ba
        - quay.io/triliodata/dex@sha256:b4b2372fd03be586b9ff8cd269b9001797d813a34155ab37de31a44fe3cab1ea
        - quay.io/triliodata/nats@sha256:e6723595bcc5674b6201091d45d711d29a5cd1421c8b97c76cf00e468379fddf
        - quay.io/triliodata/event-stack@sha256:859535e19e0ebebc8261c649c2446f8ca5fe4b4e5b6236e69b780dcc0d0bb5b9
        - quay.io/triliodata/k8s-triliovault-operator@sha256:8d582842dbef1c64bbe3b0ce5eeb64790910bdc0416c5e5561557642dede7022
        - quay.io/triliodata/operator-webhook-init@sha256:ffcc55e85f884b884947b73fa7d2fb257704ff70ccd7f2dc37d07e02292abed6
        - quay.io/triliodata/control-plane@sha256:e60ef0a2c7a5b7879c9718e9a14c223f5290b4710b598fae19466a7558484824
        version: 5.0.4
      entries:
      - name: k8s-triliovault-stable.5.0.4
        version: 5.0.4
      - name: k8s-triliovault-stable.5.0.3
        version: 5.0.3
      - name: k8s-triliovault-stable.5.0.2
        version: 5.0.2
      - name: k8s-triliovault-stable.5.0.1
        version: 5.0.1
      - name: k8s-triliovault-stable.5.0.0
        version: 5.0.0
      name: 5.0.x
    - currentCSV: k8s-triliovault-stable.5.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "2Gi",
                      "cpu": "1"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:f0cf3f586561414671668210c59ac44071fde6abc954917e6ca04df5777a9a2f
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <5.1.1'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/k8s-triliovault-operator@sha256:f0cf3f586561414671668210c59ac44071fde6abc954917e6ca04df5777a9a2f
        - quay.io/triliodata/event-stack@sha256:548ae1a748919f518e811774ec87f128aa8857be234837e09308c295de165726
        - quay.io/triliodata/control-plane@sha256:2e7e44cf6bc70dd9b978dce2cb2d9a36b8886c0f55286d5a6f5eb0a5182c4350
        - quay.io/triliodata/datamover@sha256:8a1350c8975610f23dd0b40d604d218f59dca79db6beb81ffe78a2682e40864f
        - quay.io/triliodata/dex@sha256:6b9d1667916968fb1f187ad5945fcc08db702dcab4c947e309dd9d2881a7b9cf
        - quay.io/triliodata/nats@sha256:72283871b6b9343aeabc3ff090305979422f192d30a6cbc4b0f20b2de7533996
        - quay.io/triliodata/filerecovery-vm-ssh-server@sha256:09186be228bd0b59e83119b28f3239c6173e410a3cf36957fe2d9b56d25200f8
        - quay.io/triliodata/web@sha256:3908a39eed44d702c6b6804d392d08a570b3f5562590d050f17ddc7a5c87b37f
        - quay.io/triliodata/filerecovery@sha256:42559ac5d949d5ecf819cd4768a71fa688c72230502ff0b56539a8bb5e2f5838
        version: 5.1.1
      entries:
      - name: k8s-triliovault-stable.5.1.1
        version: 5.1.1
      - name: k8s-triliovault-stable.5.1.0
        version: 5.1.0
      name: 5.1.x
    defaultChannel: 5.1.x
    packageName: k8s-triliovault
    provider:
      name: Trilio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Couchbase
      provider-url: ""
    name: couchbase-enterprise-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: couchbase-operator.v2.6.4-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator@sha256:1f0a6865c360a14ba98706fd6fe8dff3bdbdede32e3622f93f19c82e377ca33f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        version: 2.6.4-7
      entries:
      - name: couchbase-operator.v2.6.4-7
        version: 2.6.4-7
      - name: couchbase-operator.v2.6.4-6
        version: 2.6.4-6
      name: 2.6.4
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:707571a7881bc12d5f1b0b1a94eb3c675d5c0ea0114c994a47ceb5aa79df4013
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        version: 2.7.0-5
      entries:
      - name: couchbase-operator.v2.7.0-5
        version: 2.7.0-5
      name: 2.7.0
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:a79e1b6781d5f8badf664e7efea6e2889e43fae6e0d37169d766685dcd6ff394
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        version: 2.7.1-1
      entries:
      - name: couchbase-operator.v2.7.1-1
        version: 2.7.1-1
      name: 2.7.1
    - currentCSV: couchbase-operator.v2.8.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator@sha256:df61928b064014d851c6e7dff6dcd63a433a27fa5623254b6e389c3eea170e36
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        version: 2.8.0-2
      entries:
      - name: couchbase-operator.v2.8.0-2
        version: 2.8.0-2
      - name: couchbase-operator.v2.8.0-1
        version: 2.8.0-1
      name: 2.8.0
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        version: 2.8.1-2
      entries:
      - name: couchbase-operator.v2.8.1-2
        version: 2.8.1-2
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: 2.8.1
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        version: 2.8.1-2
      entries:
      - name: couchbase-operator.v2.8.1-2
        version: 2.8.1-2
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: stable
    defaultChannel: stable
    packageName: couchbase-enterprise-certified
    provider:
      name: Couchbase
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubemq.io
      provider-url: ""
    name: kubemq-operator-marketplace-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubemq-operator.v1.7.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
              "apiVersion": "core.k8s.kubemq.io/v1beta1",
              "kind": "KubemqCluster",
              "metadata": {
                "name": "kubemq-cluster"
              },
              "spec": {
                "replicas": 3
              }
            },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqDashboard",
                "metadata": {
                  "name": "kubemq-dashboard"
                }
              },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqConnector",
                "metadata": {
                  "name": "kubemq-bridges"
                },
                "spec": {
                  "type": "bridges",
                  "config": "bindings: null"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/kubemq/kubemq-operator:v1.7.2
          createdAt: 2020-03-11 12:00:00+00:00
          description: Kubernetes Messaging Platform
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/kubemq-operator-marketplace-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/kubemq-operator-marketplace-rhmp/support?utm_source=openshift_console
          support: support@kubemq.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kubemq Cluster
            displayName: Kubemq Cluster
            kind: KubemqCluster
            name: kubemqclusters.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Cluster Dashboard
            displayName: Kubemq Dashboard
            kind: KubemqDashboard
            name: kubemqdashboards.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Connector
            displayName: Kubemq Connector
            kind: KubemqConnector
            name: kubemqconnectors.core.k8s.kubemq.io
            version: v1beta1
        description: |2-

          ## Kubernetes Messaging Platform
          A message-based architecture for microservices providing flexible connectivity for hybrid environments: multi-cloud, on-premises, and at the edge. Powered by enterprise-grade message broker and message queue with ready to use connectors, bridges, and control center.

          Messaging platform components:
            - **KubeMQ Cluster** - Enterprise-grade message broker and message queue, scalable, high available, and secured.
            - **KubeMQ Targets Connector** - connects a KubeMQ cluster with external systems and cloud services.
            - **KubeMQ Bridges Connector** - bridge, replicate, aggregate, and transform messages between KubeMQ clusters.
            - **KubeMQ Sources Connector** - transform external systems and cloud services data to KubeMQ cluster messages.
            - **KubeMQ Dashboard** - monitor KubeMQ cluster metrics.

          ### Kubernetes Native
          Innovative and modern message queue and message broker in a lightweight container developed to run in Kubernetes, certified in the CNCF landscape, and connect natively to the cloud-native ecosystem.

          ### Ease of use

          Simple deployment in Kubernetes in less than 1 minute. Developer friendly by simple to use SDKs and elimination of the many developers and DevOps-centered challenges to define exchanges, brokers, channels, routes, and predefined topics

          ### All messaging patterns

          KubeMQ is available with all messaging patterns such as Queue, Stream, Pub/Sub, and RPC, Saving the need to maintain multiple messaging platforms for different use cases.

          ### Hybrid Environments

          KubeMQ enables the microservices from multiple environments to communicate and build one hybrid infrastructure solution across clouds, on-prem, and at the edge.

          ### KubeMQ Cluster key features

          - All-batteries included Messaging Queue Broker for the Kubernetes environment
          - Deployed with Operator for full life cycle operation
          - Blazing fast (written in Go), small and lightweight Docker container
          - Asynchronous and Synchronous messaging with support for  `At Most Once Delivery` and `At Least Once Delivery` models
          - Supports durable FIFO based Queue, Publish-Subscribe Events, Publish-Subscribe with Persistence (Events Store), RPC Command and Query messaging patterns
          - Supports gRPC, Rest and WebSocket Transport protocols with TLS support (both RPC and Stream modes)
          - Supports Access control Authorization and Authentication
          - Supports message masticating and smart routing
          - No Message broker configuration needed (i.e., queues, exchanges)
          - .Net, Java, Python, Go, and NodeJS SDKs

          ### KubeMQ Targets Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major services types (databases, cache, messaging, serverless, HTTP, etc.)
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-targets)

          ### KubeMQ Bridges Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Build Any Topology - connects KubeMQ clusters in 1:1, 1:n , n:1, n:n
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-bridges)

          ### KubeMQ Sources Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major messaging services such as Kafka, ActiveMQ, SQS, and more.
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-sources)

          ## Getting Started

          [Start with OpenShift](https://docs.kubemq.io/getting-started/create-cluster/openshift)

          ## Get Enterprise Trial License

          [Register](https://account.kubemq.io/login/register)

          ## Request a Demo

          [See KubeMQ in Action](https://kubemq.io/request-a-demo/)
        displayName: Kubemq Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - messaging
        - queue
        - broker
        - pubsub
        - rpc
        links:
        - name: Learn more about Kubemq
          url: https://kubemq.io/
        - name: Register for free license
          url: https://account.kubemq.io/login/register
        - name: Contact Sales
          url: https://kubemq.io/contact-sales/
        - name: Documentation
          url: https://docs.kubemq.io/
        - name: Get Help
          url: https://kubemq.io/contact-support/
        maintainers:
        - email: support@kubemq.io
          name: kubemq
        maturity: alpha
        provider:
          name: Kubemq.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-grafana:v4.3
        - registry.marketplace.redhat.com/rhm/kubemq/kubemq-operator:v1.7.2
        - registry.marketplace.redhat.com/rhm/kubemq/kubemq-uni:v2.2.12
        - registry.redhat.io/openshift4/ose-prometheus:v4.3
        version: 1.7.8
      entries:
      - name: kubemq-operator.v1.7.8
        version: 1.7.8
      name: alpha
    defaultChannel: alpha
    packageName: kubemq-operator-marketplace-rhmp
    provider:
      name: Kubemq.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: aws-efs-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-efs-csi-driver-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-aws-efs-csi-driver-rhel9-operator@sha256:35ef24108df5f24d24fefe915f501965c98edea934fa6c2895e375ed006e5e7b
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure AWS EFS CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0-0 <4.16.0-202509240123'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/aws-efs-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          AWS EFS CSI Driver Operator provides AWS Elastic File Service (EFS) CSI Driver that enables you to create and mount AWS EFS PersistentVolumes.

          Please follow [AWS EFS CSI Driver Operator documentation](https://docs.openshift.com/container-platform/4.16/storage/container_storage_interface/persistent-storage-csi-aws-efs.html) to:

          1. [Install the operator](https://docs.openshift.com/container-platform/4.16/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#persistent-storage-csi-olm-operator-install_persistent-storage-csi-aws-efs). This section also covers details about configuring an AWS role in OCP cluster that uses Secure Token Services (STS).

          2. [Install the AWS EFS CSI driver using the operator](https://docs.openshift.com/container-platform/4.16/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#persistent-storage-csi-efs-driver-install_persistent-storage-csi-aws-efs)

          3. Finally, [create a StorageClass](https://docs.openshift.com/container-platform/4.16/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#storage-create-storage-class_persistent-storage-csi-aws-efs) that enables dynamic provisioning of PersistentVolumes.
        displayName: AWS EFS CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - efs
        links:
        - name: Documentation
          url: https://github.com/openshift/aws-efs-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/aws-efs-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-aws-efs-csi-driver-rhel9-operator@sha256:35ef24108df5f24d24fefe915f501965c98edea934fa6c2895e375ed006e5e7b
        - registry.redhat.io/openshift4/ose-aws-efs-csi-driver-container-rhel9@sha256:c857dda8f399fa104bd4f392a354b74365657acea6b8fc5b023fdb4bf55057c2
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:9c3c3d2b857ce289324aee56df1ba4ab5e5ea1e76cd86ac988233463915834b4
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0627c09fe3f5a015f117fb30af1f4d17340aefa48b23994737d3901a804b63d3
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:a5704d26d0b562dba9633d1c107f34e05c750c0e3a64dbcace4dab7b964fff8d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        version: 4.16.0-202509240123
      entries:
      - name: aws-efs-csi-driver-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: aws-efs-csi-driver-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: aws-efs-csi-driver-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: aws-efs-csi-driver-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: aws-efs-csi-driver-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: aws-efs-csi-driver-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: aws-efs-csi-driver-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: aws-efs-csi-driver-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: aws-efs-csi-driver-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: aws-efs-csi-driver-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: aws-efs-csi-driver-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: aws-efs-csi-driver-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: aws-efs-csi-driver-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: aws-efs-csi-driver-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: aws-efs-csi-driver-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: aws-efs-csi-driver-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: aws-efs-csi-driver-operator.v4.16.0-202410280205
        version: 4.16.0-202410280205
      - name: aws-efs-csi-driver-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: aws-efs-csi-driver-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: aws-efs-csi-driver-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: aws-efs-csi-driver-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: aws-efs-csi-driver-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: aws-efs-csi-driver-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: aws-efs-csi-driver-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: aws-efs-csi-driver-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: aws-efs-csi-driver-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: aws-efs-csi-driver-operator.v4.16.0-202408061508
        version: 4.16.0-202408061508
      - name: aws-efs-csi-driver-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: aws-efs-csi-driver-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: aws-efs-csi-driver-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: aws-efs-csi-driver-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: aws-efs-csi-driver-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: aws-efs-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: self-node-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: self-node-remediation.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:cb03286b9d9cda4a47e5b1c0058ca315bf082bcc8697506ea88d97ba77452bea
          createdAt: "2023-10-20 16:38:06"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          olm.skipRange: '>=0.4.0 <0.5.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the heath detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-self-node-remediation-operator.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:cb03286b9d9cda4a47e5b1c0058ca315bf082bcc8697506ea88d97ba77452bea
        version: 0.5.1
      entries:
      - name: self-node-remediation.v0.5.1
        version: 0.5.1
      name: 4.12-eus
    - currentCSV: self-node-remediation.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:e2048a8a2ec37e10b57dafeba2e131311e61982242cda4beb989e5e30c1a5537
          createdAt: "2023-10-20 15:32:20"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          olm.skipRange: '>=0.5.0 <0.7.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:e2048a8a2ec37e10b57dafeba2e131311e61982242cda4beb989e5e30c1a5537
        version: 0.7.1
      entries:
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.14-eus
    - currentCSV: self-node-remediation.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:6bd86a47d69f7662e24a93ce205a8179018152a97696a9223d41ed1887c35257
          createdAt: "2025-07-02 16:40:20"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.9.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:6bd86a47d69f7662e24a93ce205a8179018152a97696a9223d41ed1887c35257
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ef91935704bb9f0af4cdad07333e802587e3d458b2cd911e04d557b1defe3412
        version: 0.9.1
      entries:
      - name: self-node-remediation.v0.9.1
        version: 0.9.1
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.16-eus
    - currentCSV: self-node-remediation.v0.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
          createdAt: "2025-07-02 18:34:25"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        version: 0.10.2
      entries:
      - name: self-node-remediation.v0.10.2
        version: 0.10.2
      - name: self-node-remediation.v0.10.1
        version: 0.10.1
      - name: self-node-remediation.v0.10.0
        version: 0.10.0
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.18-eus
    - currentCSV: self-node-remediation.v0.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
          createdAt: "2025-07-02 18:34:25"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        version: 0.10.2
      entries:
      - name: self-node-remediation.v0.10.2
        version: 0.10.2
      - name: self-node-remediation.v0.10.1
        version: 0.10.1
      - name: self-node-remediation.v0.10.0
        version: 0.10.0
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: stable
    defaultChannel: stable
    packageName: self-node-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator-certified-rhmp.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a",
                  "imagePullPolicy": "Always",
                  "initImage": "registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-server-mongodb-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-server-mongodb-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:3c03fb18a5d73876576b80aca2b61e7c84ca052839ccb467d14e15f6d1f9d834
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a57926b53eef862b70a53818a7ccb0e9756f878f30d4128b608090b109bdc585
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator-certified-rhmp.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator-certified-rhmp.v1.20.0
        version: 1.20.0
      - name: percona-server-mongodb-operator-certified-rhmp.v1.19.1
        version: 1.19.1
      - name: percona-server-mongodb-operator-certified-rhmp.v1.19.0
        version: 1.19.0
      - name: percona-server-mongodb-operator-certified-rhmp.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator-certified-rhmp
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kms-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kms-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Alias",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Key",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kms-controller:1.1.1
          createdAt: "2025-09-19T22:02:02Z"
          description: AWS KMS controller is a service controller for managing KMS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alias represents the state of an AWS kms Alias resource.
            displayName: Alias
            kind: Alias
            name: aliases.kms.services.k8s.aws
            version: v1alpha1
          - description: Grant represents the state of an AWS kms Grant resource.
            displayName: Grant
            kind: Grant
            name: grants.kms.services.k8s.aws
            version: v1alpha1
          - description: Key represents the state of an AWS kms Key resource.
            displayName: Key
            kind: Key
            name: keys.kms.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Key Management Service (KMS) resources in AWS from within your Kubernetes cluster.

          **About Amazon KMS**

          AWS Key Management Service (AWS KMS) is a managed service that makes it easy for you to create and control the cryptographic keys that are used to protect your data. AWS KMS uses hardware security modules (HSM) to protect and validate your AWS KMS keys under the [FIPS 140-2 Cryptographic Module Validation Program](https://csrc.nist.gov/projects/cryptographic-module-validation-program/Certificate/3139), except in the China (Beijing) and China (Ningxia) Regions.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon KMS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kms
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon KMS Developer Resources
          url: https://aws.amazon.com/kms/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kms maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/kms-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-kms-controller.v1.1.1
        version: 1.1.1
      - name: ack-kms-controller.v1.1.0
        version: 1.1.0
      - name: ack-kms-controller.v1.0.31
        version: 1.0.31
      - name: ack-kms-controller.v1.0.30
        version: 1.0.30
      - name: ack-kms-controller.v1.0.29
        version: 1.0.29
      - name: ack-kms-controller.v1.0.28
        version: 1.0.28
      - name: ack-kms-controller.v1.0.27
        version: 1.0.27
      - name: ack-kms-controller.v1.0.26
        version: 1.0.26
      - name: ack-kms-controller.v1.0.25
        version: 1.0.25
      - name: ack-kms-controller.v1.0.24
        version: 1.0.24
      - name: ack-kms-controller.v1.0.23
        version: 1.0.23
      - name: ack-kms-controller.v1.0.22
        version: 1.0.22
      - name: ack-kms-controller.v1.0.19
        version: 1.0.19
      - name: ack-kms-controller.v1.0.18
        version: 1.0.18
      - name: ack-kms-controller.v1.0.17
        version: 1.0.17
      - name: ack-kms-controller.v1.0.16
        version: 1.0.16
      - name: ack-kms-controller.v1.0.15
        version: 1.0.15
      - name: ack-kms-controller.v1.0.14
        version: 1.0.14
      - name: ack-kms-controller.v1.0.13
        version: 1.0.13
      - name: ack-kms-controller.v1.0.12
        version: 1.0.12
      - name: ack-kms-controller.v1.0.8
        version: 1.0.8
      - name: ack-kms-controller.v1.0.7
        version: 1.0.7
      - name: ack-kms-controller.v1.0.5
        version: 1.0.5
      - name: ack-kms-controller.v1.0.4
        version: 1.0.4
      - name: ack-kms-controller.v1.0.3
        version: 1.0.3
      - name: ack-kms-controller.v1.0.2
        version: 1.0.2
      - name: ack-kms-controller.v1.0.0
        version: 1.0.0
      - name: ack-kms-controller.v0.1.5
        version: 0.1.5
      - name: ack-kms-controller.v0.1.4
        version: 0.1.4
      - name: ack-kms-controller.v0.1.3
        version: 0.1.3
      - name: ack-kms-controller.v0.1.2
        version: 0.1.2
      - name: ack-kms-controller.v0.1.1
        version: 0.1.1
      - name: ack-kms-controller.v0.1.0
        version: 0.1.0
      - name: ack-kms-controller.v0.0.20
        version: 0.0.20
      - name: ack-kms-controller.v0.0.19
        version: 0.0.19
      - name: ack-kms-controller.v0.0.18
        version: 0.0.18
      - name: ack-kms-controller.v0.0.17
        version: 0.0.17
      - name: ack-kms-controller.v0.0.16
        version: 0.0.16
      - name: ack-kms-controller.v0.0.15
        version: 0.0.15
      - name: ack-kms-controller.v0.0.14
        version: 0.0.14
      - name: ack-kms-controller.v0.0.13
        version: 0.0.13
      - name: ack-kms-controller.v0.0.12
        version: 0.0.12
      - name: ack-kms-controller.v0.0.11
        version: 0.0.11
      - name: ack-kms-controller.v0.0.10
        version: 0.0.10
      - name: ack-kms-controller.v0.0.9
        version: 0.0.9
      - name: ack-kms-controller.v0.0.8
        version: 0.0.8
      - name: ack-kms-controller.v0.0.7
        version: 0.0.7
      - name: ack-kms-controller.v0.0.6
        version: 0.0.6
      - name: ack-kms-controller.v0.0.5
        version: 0.0.5
      - name: ack-kms-controller.v0.0.4
        version: 0.0.4
      - name: ack-kms-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-kms-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: proactive-node-scaling-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: proactive-node-scaling-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NodeScalingWatermark",
                "metadata": {
                  "name": "ai-ml"
                },
                "spec": {
                  "nodeSelector": {
                    "machine.openshift.io/cluster-api-machine-type": "ai-ml"
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "workload",
                      "operator": "Equal",
                      "value": "ai-ml"
                    }
                  ],
                  "watermarkPercentage": 20
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NodeScalingWatermark",
                "metadata": {
                  "name": "us-west-2a"
                },
                "spec": {
                  "nodeSelector": {
                    "topology.kubernetes.io/zone": "us-west-2a"
                  },
                  "watermarkPercentage": 20
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "false"
          containerImage: quay.io/redhat-cop/proactive-node-scaling-operator@sha256:892303babb45090a6458cd4b1c61a01364012c43ea0bd6ed4d10bdcb8041e2ab
          createdAt: "2022-01-22T14:14:23Z"
          description: This operator provides automation proactively scaling nodes
            before pods get stuck waiting for the cluster autoscaler.
          operatorframework.io/suggested-namespace: proactive-node-scaling-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/proactive-node-scaling-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeScalingWatermark is the Schema for the nodescalingwatermarks
              API
            displayName: Node Scaling Watermark
            kind: NodeScalingWatermark
            name: nodescalingwatermarks.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator makes the [cluster autoscaler](https://github.com/kubernetes/autoscaler/tree/master/cluster-autoscaler) more proactive. As of now the cluster auto scaler will create new nodes only when a pod is pending because it cannot be allocated due to lack of capacity. This is not a goos user experience as the pending workload has to wait for several minutes as the new node is create and joins the cluster.

          The Proactive Node Scaling Operator improves the user experience by allocating low priority pods that don't do anything. When the cluster is full and a new user pod is created the following happens:

          1. some of the low priority pods are de-scheduled to make room for the user pod, which can then be scheduled. The user workload does not have to wait in this case.

          2. the de-scheduled low priority pods are rescheduled and in doing so the trigger the cluster autoscaler to add new nodes.

          Essentially this operator allows you to trade wasted resources for faster response time.

          In order for this operator to work correctly [pod priorities](https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/) must be defined. The default name for the priority class used by this operator is "proactive-node-autoscaling-pods" and it should have the lowest possible priority, 0. To ensure your regular workloads get a normal priority you should also define a PriorityClass for those with a higher priority than 0 and set globalDefault to true.

          For example:

          ```yaml
          apiVersion: scheduling.k8s.io/v1
          kind: PriorityClass
          metadata:
            name: proactive-node-autoscaling-pods
          value: 0
          globalDefault: false
          description: "This priority class is the priority class used for Proactive Node Scaling Pods."
          ---
          apiVersion: scheduling.k8s.io/v1
          kind: PriorityClass
          metadata:
            name: normal-workload
          value: 1000
          globalDefault: true
          description: "This priority classis the cluster default and should be used for normal workloads."
          ```

          Also for this operator to work the [cluster autoscaler](https://github.com/kubernetes/autoscaler/tree/master/cluster-autoscaler) must be active, see OpenShift instructions [here](https://docs.openshift.com/container-platform/4.6/machine_management/applying-autoscaling.html) on how to turn it on.

          To activate the proactive autoscaling, a CR must be defined, here is an example:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: NodeScalingWatermark
          metadata:
            name: us-west-2a
          spec:
            priorityClassName: proactive-node-autoscaling-pods
            watermarkPercentage: 20
            nodeSelector:
              topology.kubernetes.io/zone: us-west-2a
          ```

          The `nodeSelector` selects the nodes observed by this operator, which are also the nodes on which the low priority pods will be scheduled. The nodes observed by the cluster autoscaler should coincide with the nodes selected by this operator CR.

          The `watermarkPercentage` define the percentage of capacity of user workload that will be allocated to low priority pods. So in this example 20% of the user allocated capacity will be allocated via low priority pods. This also means that when the user workload reaches 80% capacity of the nodes selected by this CR (and the autoscaler), the cluster will start to scale.
        displayName: Proactive Node Scaling Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Capacity Management
        - Node Autoscaling
        - Proactive Autoscaling
        - Elastic Capacity
        links:
        - name: Proactive Node Scaling Operator
          url: https://github.com/redhat-cop/proactive-node-scaling-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/proactive-node-scaling-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/proactive-node-scaling-operator@sha256:892303babb45090a6458cd4b1c61a01364012c43ea0bd6ed4d10bdcb8041e2ab
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 0.4.2
      entries:
      - name: proactive-node-scaling-operator.v0.4.2
        version: 0.4.2
      - name: proactive-node-scaling-operator.v0.4.0
        version: 0.4.0
      - name: proactive-node-scaling-operator.v0.3.10
        version: 0.3.10
      - name: proactive-node-scaling-operator.v0.3.9
        version: 0.3.9
      - name: proactive-node-scaling-operator.v0.3.8
        version: 0.3.8
      - name: proactive-node-scaling-operator.v0.3.7
        version: 0.3.7
      - name: proactive-node-scaling-operator.v0.3.6
        version: 0.3.6
      - name: proactive-node-scaling-operator.v0.3.5
        version: 0.3.5
      - name: proactive-node-scaling-operator.v0.3.4
        version: 0.3.4
      - name: proactive-node-scaling-operator.v0.3.3
        version: 0.3.3
      - name: proactive-node-scaling-operator.v0.3.2
        version: 0.3.2
      - name: proactive-node-scaling-operator.v0.3.1
        version: 0.3.1
      - name: proactive-node-scaling-operator.v0.3.0
        version: 0.3.0
      - name: proactive-node-scaling-operator.v0.2.0
        version: 0.2.0
      - name: proactive-node-scaling-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: proactive-node-scaling-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Debezium Authors
      provider-url: https://debezium.io/
    name: debezium-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: debezium-operator.v2.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.4.2.Final
          createdAt: "2023-12-13T12:20:56Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.4.2.Final
        version: 2.4.2
      entries:
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.4.x
    - currentCSV: debezium-operator.v2.5.3-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.5.3.Final
          createdAt: "2024-03-19T08:17:17Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.5.3.Final
        version: 2.5.3-final
      entries:
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.5.x
    - currentCSV: debezium-operator.v2.6.1-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.6.1.Final
          createdAt: "2024-04-12T17:18:16Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.5.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.6.1.Final
        version: 2.6.1-final
      entries:
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.6.x
    - currentCSV: debezium-operator.v2.7.3-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.7.3.Final
          createdAt: "2024-09-20T10:31:16Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.7.3.Final
        version: 2.7.3-final
      entries:
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.7.x
    - currentCSV: debezium-operator.v3.0.0-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:3.0.0.Final
          createdAt: "2024-10-02T09:55:38Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.7.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:3.0.0.Final
        version: 3.0.0-final
      entries:
      - name: debezium-operator.v3.0.0-final
        version: 3.0.0-final
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-3.0.x
    - currentCSV: debezium-operator.v3.0.0-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:3.0.0.Final
          createdAt: "2024-10-02T09:55:38Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.7.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:3.0.0.Final
        version: 3.0.0-final
      entries:
      - name: debezium-operator.v3.0.0-final
        version: 3.0.0-final
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-latest
    defaultChannel: debezium-latest
    packageName: debezium-operator
    provider:
      name: Debezium Authors
      url: https://debezium.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-notifications
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-notifications.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Notifications",
                "metadata": {
                  "name": "notifications-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configuration": {
                    "port": 8080
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp",
                    "test": "wap-rhel8-toolbox:8.7.0-3747-36-ocp",
                    "testRepository": "prod.harbor.keysight.digital/loadcore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:34:47Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Notifications
            name: notifications.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Notifications Operator
        displayName: Keysight WAP Notifications Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Notifications
          url: https://keysight-wap-notifications.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-notifications@sha256:50c24180b3c16cf3eb82f6f4eeb61b91a62a198c122f7e59aa680a25bc488c2d
        version: 0.1.21
      entries:
      - name: keysight-wap-notifications.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-notifications
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-stats-dashboards
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-stats-dashboards.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Statsdashboard",
                "metadata": {
                  "name": "statsdashboard-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "statsDashboardService": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:41:08Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Statsdashboard
            name: statsdashboards.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Stats Dashboard Operator
        displayName: Keysight WAP Stats Dashboard Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Stats Dashboards
          url: https://keysight-wap-stats-dashboards.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-stats-dashboards@sha256:bcdc68bf0b9e9c22d775d704320829d258a9ebc77dfe41e2414061740f87f236
        version: 0.1.21
      entries:
      - name: keysight-wap-stats-dashboards.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-stats-dashboards
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Saison Technology Co., Ltd.
      provider-url: https://www.hulft.com/
    name: hulft
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hulft.v10.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hulft.saison-technology.com/v1",
                "kind": "Hulft",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "hulft",
                    "app.kubernetes.io/instance": "sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hulft",
                    "app.kubernetes.io/part-of": "hulft"
                  },
                  "name": "sample",
                  "namespace": "user-apps"
                },
                "spec": {
                  "applicationName": "hulft-apps",
                  "control": {
                    "allowExternalAccess": true,
                    "replicas": {
                      "max": 1,
                      "min": 1
                    }
                  },
                  "db": {
                    "authorizationSecret": "secret-hulft-db-authorization",
                    "connectionRetryCount": 30,
                    "hostname": "database.example.com",
                    "name": "hulft",
                    "port": 3306,
                    "tablePrefix": "hulft"
                  },
                  "licenseSecret": "secret-hulft-license",
                  "resourceIdentifier": "sample",
                  "transfer": {
                    "control": {
                      "retryCount": 10,
                      "retryInterval": 10000,
                      "timeout": 60000
                    },
                    "gracefulShutdownWait": 3600,
                    "observe": {
                      "websocket": {
                        "allowExternalAccess": false
                      }
                    },
                    "replicas": {
                      "max": 5,
                      "min": 1
                    },
                    "s3": {
                      "defaultRegion": "ap-northeast-1"
                    },
                    "websocket": {
                      "allowExternalAccess": false
                    }
                  }
                },
                "status": {
                  "applicationName": "hulft-apps",
                  "control": {
                    "allowExternalAccess": true,
                    "apiClientTlsSecret": "secret-hulft-sample-transfer-control-server-tls",
                    "consoleLoginSecret": "secret-hulft-sample-console-login",
                    "consoleTlsSecret": "secret-hulft-sample-control-console-tls",
                    "hostname": "",
                    "httpPort": 30080,
                    "httpsPort": 30443,
                    "replicas": {
                      "max": 1,
                      "min": 1
                    },
                    "samlSecret": ""
                  },
                  "db": {
                    "authorizationSecret": "secret-hulft-db-authorization",
                    "connectionRetryCount": 30,
                    "encryptionKeySecret": "secret-hulft-sample-db-encryption-key",
                    "hostname": "database.example.com",
                    "name": "hulft",
                    "port": 3306,
                    "tablePrefix": "hulft"
                  },
                  "licenseSecret": "secret-hulft-license",
                  "resourceIdentifier": "sample",
                  "transfer": {
                    "control": {
                      "clientTlsSecret": "secret-hulft-sample-control-console-tls",
                      "port": 30443,
                      "retryCount": 10,
                      "retryInterval": 10000,
                      "serverTlsSecret": "secret-hulft-sample-transfer-control-server-tls",
                      "timeout": 60000
                    },
                    "gracefulShutdownWait": 3600,
                    "hostname": "hulft-sample-transfer-wss-ci.apps.openshift.example.com",
                    "hulft": {
                      "port": 30000
                    },
                    "observe": {
                      "port": 31000,
                      "websocket": {
                        "allowExternalAccess": true,
                        "port": 33443,
                        "serverTlsSecret": "secret-hulft-sample-transfer-observe-wss-server-tls"
                      }
                    },
                    "replicas": {
                      "max": 5,
                      "min": 1
                    },
                    "s3": {
                      "authenticationSecret": "",
                      "defaultRegion": "ap-northeast-1"
                    },
                    "websocket": {
                      "allowExternalAccess": true,
                      "clientTlsSecret": "",
                      "port": 31443,
                      "serverTlsSecret": "secret-hulft-sample-transfer-wss-server-tls"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/saison-technology/hulft-operator@sha256:8ee0a29d0ce96f4ec230f387467b241dccbb2bd350cd642b275ef5f5d3b64972
          createdAt: "2025-09-30T07:54:26Z"
          description: HULFT10 for Container Platform is file integration middleware
            that provides transfer functions for the data accumulated in your IT system
            during the daily operations of your business, such as sales data, customer
            information, and various logs. File transfer from/to on-premises HULFT
            is also available.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: To use HULFT10 for Container
            Platform, you must purchase the product and have a key issued. Please
            refer to [Start Up Guide](https://www.hulft.com/help/ja-jp/HULFT-V10/CONP-SUGOS)
            for the deployment procedure.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Saison Technology Co., Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hulft is the Schema for the hulfts API
            displayName: Hulft
            kind: Hulft
            name: hulfts.hulft.saison-technology.com
            version: v1
        description: |
          HULFT10 for Container Platform is file integration middleware that provides transfer functions for the data accumulated in your IT system during the daily operations of your business, such as sales data, customer information, and various logs. File transfer from/to on-premises HULFT is also available.
          To use HULFT10 for Container Platform, you must purchase the product and have a key issued. Please refer to [Start Up Guide](https://www.hulft.com/help/ja-jp/HULFT-V10/CONP-SUGOS) for the deployment procedure.
        displayName: HULFT10 for Container Platform
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backup & Recovery
        - Storage
        - Migration
        links:
        - name: Product page
          url: https://www.hulft.com/software/hulft-10
        - name: Product manual
          url: https://www.hulft.com/help/ja-jp/HULFT-V10/index.htm
        maintainers:
        - email: salessupport_mp@hulft.com
          name: Saison Technology Co., Ltd.
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Saison Technology Co., Ltd.
          url: https://www.hulft.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fa0acefda43e309e7f9f31f5247c933d0b868ff1573ece2f38e0549d7e2ec942
        - registry.connect.redhat.com/saison-technology/hulft-operator@sha256:8ee0a29d0ce96f4ec230f387467b241dccbb2bd350cd642b275ef5f5d3b64972
        version: 10.5.0
      entries:
      - name: hulft.v10.5.0
        version: 10.5.0
      - name: hulft.v10.4.0
        version: 10.4.0
      - name: hulft.v10.3.0
        version: 10.3.0
      name: stable
    defaultChannel: stable
    packageName: hulft
    provider:
      name: Saison Technology Co., Ltd.
      url: https://www.hulft.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC Corporation
      provider-url: https://jpn.nec.com/webotx/
    name: webotx-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: webotx-operator.v11.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "webotx.nec.com/v3",
                "kind": "ApplicationServer",
                "metadata": {
                  "name": "webotx-express"
                },
                "spec": {
                  "deployment": {
                    "image": "registry.connect.redhat.com/nec/webotx-express:11.11.00.00@sha256:73a5dc48b4e3340951dbe207d744f5ac68474701dd134fa7aa05e65e3510c8ad",
                    "replicas": 1
                  },
                  "service": {
                    "externalTrafficPolicy": "Cluster",
                    "ports": [
                      {
                        "name": "http",
                        "nodePort": 30080,
                        "port": 8080,
                        "protocol": "TCP",
                        "targetPort": 8080
                      },
                      {
                        "name": "https",
                        "nodePort": 30443,
                        "port": 8443,
                        "protocol": "TCP",
                        "targetPort": 8443
                      }
                    ],
                    "sessionAffinity": "ClientIP",
                    "sessionAffinityConfig": {
                      "clientIP": {
                        "timeoutSeconds": 10800
                      }
                    },
                    "type": "NodePort"
                  },
                  "webotx": {
                    "metering": {
                      "enabled": false,
                      "port": 8443,
                      "scheme": "https"
                    },
                    "sessionReplication": {
                      "enabled": false,
                      "useAllProcessGroups": false,
                      "useJmxAgent": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApplicationServer is the Schema for the applicationservers
              API
            displayName: Application Server
            kind: ApplicationServer
            name: applicationservers.webotx.nec.com
            version: v3
        description: WebOTX Operator is the Operator for WebOTX Application Server.
          WebOTX Operator enables to deploy and administrate WebOTX Application Server
          automatically.
        displayName: WebOTX Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - webotx
        - WebOTX
        - nec
        - NEC
        links:
        - name: WebOTX Operator for Kubernetes
          url: https://github.com/webotx/operator
        maintainers:
        - email: info-webotx@isd.jp.nec.com
          name: NEC Corporation
        maturity: stable
        provider:
          name: NEC Corporation
          url: https://jpn.nec.com/webotx/
        relatedImages:
        - registry.connect.redhat.com/nec/webotx-operator:11.11.00.00@sha256:df0d0315ed88be875ca7a6474500f43f17855f763c5c7d0f4d3ec58a2e46f6fb
        version: 11.11.0
      entries:
      - name: webotx-operator.v11.11.0
        version: 11.11.0
      name: stable
    defaultChannel: stable
    packageName: webotx-operator
    provider:
      name: NEC Corporation
      url: https://jpn.nec.com/webotx/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: security-profiles-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: security-profiles-operator.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "AppArmorProfile",
                "metadata": {
                  "annotations": {
                    "description": "Block writing to any files in the disk."
                  },
                  "name": "test-profile"
                },
                "spec": {
                  "abstract": {
                    "filesystem": {
                      "readOnlyPaths": [],
                      "readWritePaths": [],
                      "writeOnlyPaths": []
                    }
                  }
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "SecurityProfilesOperatorDaemon",
                "metadata": {
                  "name": "spod",
                  "namespace": "security-profiles-operator"
                },
                "spec": {
                  "enableAppArmor": false,
                  "enableLogEnricher": false,
                  "enableSelinux": false
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "ProfileRecording",
                "metadata": {
                  "name": "test-recording"
                },
                "spec": {
                  "kind": "SeccompProfile",
                  "podSelector": {
                    "matchLabels": {
                      "app": "alpine"
                    }
                  },
                  "recorder": "bpf"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "ProfileBinding",
                "metadata": {
                  "name": "profile-binding"
                },
                "spec": {
                  "image": "nginx:1.19.1",
                  "profileRef": {
                    "kind": "SeccompProfile",
                    "name": "profile-allow-unsafe"
                  }
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha2",
                "kind": "RawSelinuxProfile",
                "metadata": {
                  "name": "errorlogger"
                },
                "spec": {
                  "policy": "(blockinherit container)\n(allow process var_log_t ( dir ( open read getattr lock search ioctl add_name remove_name write ))) \n(allow process var_log_t ( file ( getattr read write append ioctl lock map open create  ))) \n(allow process var_log_t ( sock_file ( getattr read write append open  ))) \n"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "Blocks all syscalls."
                  },
                  "name": "profile-block-all"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_ERRNO"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "UNSAFE: Allows all syscalls whilst logging their use. Similar to running as unconfined in terms of enforcement."
                  },
                  "name": "profile-complain-unsafe"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_LOG"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "UNSAFE: Allows all syscalls. Similar to running as unconfined as it provides no enforcement."
                  },
                  "name": "profile-allow-unsafe"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_ALLOW"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "Enables complain mode whilst blocking high-risk syscalls. Some essential syscalls are allowed to decrease log noise."
                  },
                  "name": "profile-complain-block-high-risk"
                },
                "spec": {
                  "architectures": [
                    "SCMP_ARCH_X86_64",
                    "SCMP_ARCH_PPC64LE"
                  ],
                  "defaultAction": "SCMP_ACT_LOG",
                  "syscalls": [
                    {
                      "action": "SCMP_ACT_ALLOW",
                      "names": [
                        "exit",
                        "exit_group",
                        "futex",
                        "nanosleep"
                      ]
                    },
                    {
                      "action": "SCMP_ACT_ERRNO",
                      "names": [
                        "acct",
                        "add_key",
                        "bpf",
                        "clock_adjtime",
                        "clock_settime",
                        "create_module",
                        "delete_module",
                        "finit_module",
                        "get_kernel_syms",
                        "get_mempolicy",
                        "init_module",
                        "ioperm",
                        "iopl",
                        "kcmp",
                        "kexec_file_load",
                        "kexec_load",
                        "keyctl",
                        "lookup_dcookie",
                        "mbind",
                        "mount",
                        "move_pages",
                        "name_to_handle_at",
                        "nfsservctl",
                        "open_by_handle_at",
                        "perf_event_open",
                        "personality",
                        "pivot_root",
                        "process_vm_readv",
                        "process_vm_writev",
                        "ptrace",
                        "query_module",
                        "quotactl",
                        "reboot",
                        "request_key",
                        "set_mempolicy",
                        "setns",
                        "settimeofday",
                        "stime",
                        "swapoff",
                        "swapon",
                        "_sysctl",
                        "sysfs",
                        "umount2",
                        "umount",
                        "unshare",
                        "uselib",
                        "userfaultfd",
                        "ustat",
                        "vm86old",
                        "vm86"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha2",
                "kind": "SelinuxProfile",
                "metadata": {
                  "name": "errorlogger"
                },
                "spec": {
                  "allow": {
                    "var_log_t": {
                      "dir": [
                        "open",
                        "read",
                        "getattr",
                        "lock",
                        "search",
                        "ioctl",
                        "add_name",
                        "remove_name",
                        "write"
                      ],
                      "file": [
                        "getattr",
                        "read",
                        "write",
                        "append",
                        "ioctl",
                        "lock",
                        "map",
                        "open",
                        "create"
                      ],
                      "sock_file": [
                        "getattr",
                        "read",
                        "write",
                        "append",
                        "open"
                      ]
                    }
                  },
                  "inherit": [
                    {
                      "name": "container"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          createdAt: "2025-02-25T13:34:54Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.1 <0.9.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-security-profiles
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ProfileBinding is the Schema for the profilebindings API.
            displayName: Profile Binding
            kind: ProfileBinding
            name: profilebindings.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: ProfileRecording is the Schema for the profilerecordings
              API.
            displayName: Profile Recording
            kind: ProfileRecording
            name: profilerecordings.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: SeccompProfile is a cluster level specification for a seccomp
              profile. See https://github.com/opencontainers/runtime-spec/blob/master/config-linux.md#seccomp
            displayName: Seccomp Profile
            kind: SeccompProfile
            name: seccompprofiles.security-profiles-operator.x-k8s.io
            version: v1beta1
          - description: SecurityProfileNodeStatus is a per-node status of a security
              profile
            displayName: Security Profile Node Status
            kind: SecurityProfileNodeStatus
            name: securityprofilenodestatuses.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: SecurityProfilesOperatorDaemon is the Schema to configure
              the spod deployment.
            displayName: Security Profiles Operator Daemon
            kind: SecurityProfilesOperatorDaemon
            name: securityprofilesoperatordaemons.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - kind: RawSelinuxProfile
            name: rawselinuxprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha2
          - description: SelinuxProfile is the Schema for the selinuxprofiles API.
            displayName: Selinux Profile
            kind: SelinuxProfile
            name: selinuxprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha2
          - description: AppArmorProfile is a cluster level specification for an AppArmor
              profile.
            displayName: App Armor Profile
            kind: AppArmorProfile
            name: apparmorprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha1
        description: SPO is an operator which aims to make it easier for users to
          use SELinux, seccomp and AppArmor in Kubernetes clusters
        displayName: Security Profiles Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - selinux
        - seccomp
        - apparmor
        - ebpf
        links:
        - name: Security Profiles Operator
          url: https://github.com/openshift/security-profiles-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-security-profiles-rhel8-operator@sha256:828274d71c65ef820e6a8b652e3f102624d09a0609fb00bae7b32dc62ff37fd1
        - registry.redhat.io/compliance/openshift-selinuxd-rhel9@sha256:3278d69814a8d0d1e31dcdef454c24828aadbb468a3d5d3e15910c76481765bd
        - registry.redhat.io/compliance/openshift-selinuxd-rhel8@sha256:81273ff149482109417064b40d333edb131481c05547e4ffa6485475460a48ba
        version: 0.9.0
      entries:
      - name: security-profiles-operator.v0.9.0
        version: 0.9.0
      - name: security-profiles-operator.v0.8.6
        version: 0.8.6
      - name: security-profiles-operator.v0.8.5
        version: 0.8.5
      - name: security-profiles-operator.v0.8.4
        version: 0.8.4
      - name: security-profiles-operator.v0.8.2
        version: 0.8.2
      name: release-alpha-rhel-8
    defaultChannel: release-alpha-rhel-8
    packageName: security-profiles-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: prometheus-exporter-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometheus-exporter-operator.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.3scale.net/v1alpha1",
                "kind": "PrometheusExporter",
                "metadata": {
                  "name": "example-memcached"
                },
                "spec": {
                  "dbHost": "your-memcached-host",
                  "dbPort": 11211,
                  "grafanaDashboard": {
                    "label": {
                      "key": "autodiscovery",
                      "value": "enabled"
                    }
                  },
                  "type": "memcached"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/3scale/prometheus-exporter-operator:v0.7.0
          createdAt: "2023-09-07 00:00:00"
          description: Operator to setup 3rd party prometheus exporters, with a collection
            of grafana dashboards
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/3scale-ops/prometheus-exporter-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configures a prometheus exporter to monitor a memcached instance
            displayName: PrometheusExporter
            kind: PrometheusExporter
            name: prometheusexporters.monitoring.3scale.net
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK to centralize the setup of 3rd party prometheus exporters on **Kubernetes/OpenShift**, with a collection of grafana dashboards.

          By just providing a few parameters like *dbHost* or *dbPort* (operator manage the container image, port, argument, command, volumes... and also prometheus `ServiceMonitor` and `GrafanaDashboard` k8s objects), you can setup different prometheus exporters to monitor:
          * The **internals from different databases**
          * **HTTP/TCP endpoints** (availability, latency, SSL/TLS certificate expiration...)
          * Any available **cloudwatch metric from any AWS Service**
          * Sendgrid email statistics (delivered, bounces, errors, spam...)

          Current prometheus exporters types supported, managed by same prometheus-exporter-operator:
          * memcached
          * redis
          * mysql
          * postgresql
          * sphinx
          * es (elasticsearch)
          * cloudwatch
          * probe (blackbox)
          * sendgrid

          The operator manages the lifecycle of the following objects:
          * Deployment (one per CR)
          * Service (one per CR)
          * ServiceMonitor (optional, one per CR)
          * GrafanaDashboard (optional, one per Namespace)

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale-ops/prometheus-exporter-operator#documentation).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale-ops/prometheus-exporter-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:
          * Raising any issues you find using Prometheus Exporter Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale-ops/prometheus-exporter-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving [documentation](https://github.com/3scale-ops/prometheus-exporter-operator)
          * Talking about Prometheus Exporter Operator
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale-ops/prometheus-exporter-operator/issues).

          ### License
          Prometheus Exporter Operator is licensed under the [Apache 2.0 license](https://github.com/3scale-ops/prometheus-exporter-operator/blob/main/LICENSE)
        displayName: Prometheus Exporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - metrics
        - observability
        - prometheus
        - exporter
        - grafana
        - mysql
        - postgresql
        - memcached
        - redis
        - sphinx
        - elasticsearch
        - cloudwatch
        - blackbox
        - sendgrid
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/prometheus-exporter-operator
        maintainers:
        - email: 3scale-operations@redhat.com
          name: 3scale Ops
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/3scale/prometheus-exporter-operator:v0.7.0
        version: 0.7.0
      entries:
      - name: prometheus-exporter-operator.v0.7.0
        version: 0.7.0
      - name: prometheus-exporter-operator.v0.3.4
        version: 0.3.4
      name: alpha
    defaultChannel: alpha
    packageName: prometheus-exporter-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: boxboat-github-practice
      provider-url: https://github.com/boxboat-github-practice
    name: github-arc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: github-arc-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "github-practice.boxboat.com/v1alpha1",
                "kind": "ActionsRunnerController",
                "metadata": {
                  "name": "arc-sample"
                },
                "spec": {
                  "authSecret": {
                    "name": "ghauth"
                  },
                  "createRunnerNamespaces": true,
                  "openshift": true,
                  "runnerNamespaces": [
                    "test1",
                    "test2"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery, OpenShift Optional
          containerImage: ghcr.io/boxboat-github-practice/github-arc-operator:1.1.0
          createdAt: "2023-03-03T20:49:16Z"
          description: Operator for deploying the github actions runner controller
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/boxboat-github-practice/github-arc-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActionsRunnerController is the schema for the ActionsRunnerController
              API
            displayName: Actions Runner Controller
            kind: ActionsRunnerController
            name: actionsrunnercontrollers.github-practice.boxboat.com
            version: v1alpha1
        description: helm based operator for github actions runner controller
        displayName: github-arc-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - github
        - actions
        - boxboat
        - ibm
        - self-hosted
        - runners
        links:
        - name: Github Arc Operator
          url: https://github.com/boxboat-github-practice/github-arc-operator
        - name: Actions Runner Controller
          url: https://github.com/actions/actions-runner-controller
        - name: Boxboat
          url: https://boxboat.com/
        maintainers:
        - email: Michael.McMullen@ibm.com
          name: mbmcmullen27
        maturity: alpha
        provider:
          name: boxboat-github-practice
          url: https://github.com/boxboat-github-practice
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - ghcr.io/boxboat-github-practice/github-arc-operator:1.1.0
        version: 1.1.0
      entries:
      - name: github-arc-operator.v1.1.0
        version: 1.1.0
      - name: github-arc-operator.v1.0.4
        version: 1.0.4
      name: alpha
    defaultChannel: alpha
    packageName: github-arc-operator
    provider:
      name: boxboat-github-practice
      url: https://github.com/boxboat-github-practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: special-resource-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: special-resource-operator.v4.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sro.openshift.io/v1beta1",
                "kind": "SpecialResource",
                "metadata": {
                  "name": "simple-kmod"
                },
                "spec": {
                  "chart": {
                    "name": "simple-kmod",
                    "repository": {
                      "name": "example",
                      "url": "file:///charts/example"
                    },
                    "version": "0.0.1"
                  },
                  "driverContainer": {
                    "source": {
                      "git": {
                        "ref": "master",
                        "uri": "https://github.com/openshift-psap/kvc-simple-kmod.git"
                      }
                    }
                  },
                  "namespace": "simple-kmod",
                  "set": {
                    "apiVersion": "sro.openshift.io/v1beta1",
                    "buildArgs": [
                      {
                        "name": "KMODVER",
                        "value": "SRO"
                      }
                    ],
                    "kind": "Values",
                    "kmodNames": [
                      "simple-kmod",
                      "simple-procfs-kmod"
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional, Storage
          containerImage: quay.io/openshift-psap/special-resource-operator:release-4.9
          description: The special resource operator is an orchestrator for resources
            in a cluster required to deploy special hardware and software. Examples
            include hardware accelerators or software defined storage, where extra
            management or kernel modules are needed.
          olm.skipRange: '>=4.6.0 <4.9.0'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpecialResource is the Schema for the specialresources API
            displayName: Special Resource
            kind: SpecialResource
            name: specialresources.sro.openshift.io
            version: v1beta1
          required:
          - description: Discovers special resources and labels the nodes with the
              PCI vendor ID.
            displayName: Node Feature Discovery
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
        description: "## About the managed application \nThe SRO will deploy the complete
          stack for enabling the accelerator in a cluster. \nIt manages the drivers,
          device-plugin, node-exporter, grafana and feature discovery.\n## About this
          Operator\n README: [https://github.com/openshift-psap/special-resource-operator](https://github.com/openshift-psap/special-resource-operator)\n\n##
          Prerequisites for enabling this Operator\nThe SRO has a dependency on NFD
          (Node Feature Discovery)\n"
        displayName: Special Resource Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GPI
        links:
        - name: README
          url: https://github.com/openshift-psap/special-resource-operator
        maintainers:
        - email: zkaiser@redhat.com
          name: Zvonko Kaiser
        - email: dagray@redhat.com
          name: David Gray
        maturity: alpha
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy
        - quay.io/openshift-psap/special-resource-operator:release-4.9
        version: 4.9.0
      entries:
      - name: special-resource-operator.v4.9.0
        version: 4.9.0
      - name: special-resource-operator.v4.8.2
        version: 4.8.2
      - name: special-resource-operator.v4.8.1
        version: 4.8.1
      - name: special-resource-operator.v4.8.0
        version: 4.8.0
      name: alpha
    defaultChannel: alpha
    packageName: special-resource-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Digital.ai
      provider-url: https://digital.ai
    name: digitalai-release-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: digitalai-release-operator.v25.3.0-beta.813
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 5,
                    "metrics": [],
                    "minReplicas": 3
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "cluster": {
                    "management": {
                      "ssl": {
                        "enabled": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    },
                    "manager": "xlr-legacy",
                    "mode": "full",
                    "name": "xl-release_cluster",
                    "node": {
                      "datacenter": "default"
                    },
                    "remote": {
                      "sslEngine": {
                        "enabled": false,
                        "enabledAlgorithms": [
                          "TLS_DHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_DHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_RSA_WITH_AES_128_CBC_SHA"
                        ],
                        "hostnameVerification": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "requireMutualAuthentication": false,
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"${XLR_CLUSTER_NAME}\"\n    manager = \"${XLR_CLUSTER_MANAGER}\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n      {{- if .Values.storage.reportingEngine.downingProviderClass }}\n      cluster {\n        downing-provider-class = {{.Values.storage.reportingEngine.downingProviderClass | quote}}\n      }\n      {{- end }}\n\n      {{- if .Values.cluster.remote.sslEngine.enabled }}\n      remote {\n        artery {  \n          ssl {  \n            config-ssl-engine {  \n              enabled-algorithms = [\n                {{- $last := sub (len .Values.cluster.remote.sslEngine.enabledAlgorithms) 1 }}\n                {{ range $index, $v := .Values.cluster.remote.sslEngine.enabledAlgorithms }}\n                {{ $v | quote }}{{ if ne $index $last }},{{ end }}\n                {{- end }}\n              ]  \n              key-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEY_PASSWORD}\"  \n              key-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.keystore \"default\" (printf \"cluster_remote_keystore.%s\" .Values.cluster.remote.sslEngine.keystoreType) \"context\" $) }}\"  \n              key-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEYSTORE_PASSWORD}\"  \n              trust-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.truststore \"default\" (printf \"cluster_remote_truststore.%s\" .Values.cluster.remote.sslEngine.truststoreType) \"context\" $) }}\"  \n              trust-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_TRUSTSTORE_PASSWORD}\"  \n              hostname-verification = {{ if .Values.cluster.remote.sslEngine.hostnameVerification }}on{{ else }}off{{ end }}\n              require-mutual-authentication = {{ if .Values.cluster.remote.sslEngine.requireMutualAuthentication }}on{{ else }}off{{ end }}  \n            }  \n          }  \n          transport = tls-tcp  \n        }  \n      }\n      {{- end }}\n    }\n    {{- if .Values.cluster.management.ssl.enabled }}\n    xlr-pekko-native {\n      pekko {\n        management {\n          http {  \n            route-providers-read-only = true  \n          }  \n          ssl {  \n            enabled = true\n            keyPassword = \"${?CLUSTER_MANAGEMENT_SSL_KEY_PASSWORD}\"  \n            keyStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.keystore \"default\" (printf \"cluster_management_keystore.%s\" .Values.cluster.management.ssl.keystoreType) \"context\" $) }}\"  \n            keyStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_KEYSTORE_PASSWORD}\"  \n            keystoreType = {{ .Values.cluster.management.ssl.keystoreType }}  \n            trustStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.truststore \"default\" (printf \"cluster_management_truststore.%s\" .Values.cluster.management.ssl.truststoreType) \"context\" $) }}\"  \n            trustStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_TRUSTSTORE_PASSWORD}\"  \n            trustStoreType = {{ .Values.cluster.management.ssl.truststoreType }}  \n          }  \n        }  \n      }\n    }\n    {{- end }}\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\nmanagement.endpoint.prometheus.enabled = {{ .Values.metrics.prometheus.enabled | default false }}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      datacenter = ${XLR_CLUSTER_DATACENTER}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "cd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "livenessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/liveness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "jmx": {
                      "enabled": false
                    },
                    "prometheus": {
                      "enabled": false,
                      "podAnnotations": {
                        "prometheus.io/port": "{{ .Values.service.ports.metrics }}",
                        "prometheus.io/scrape": "true"
                      },
                      "rule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "serviceMonitor": {
                        "annotations": {},
                        "basicAuth": {
                          "password": {
                            "key": "releasePassword",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          },
                          "username": {
                            "key": "releaseUsername",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          }
                        },
                        "enabled": false,
                        "honorLabels": false,
                        "interval": "30s",
                        "jobLabel": "",
                        "labels": {},
                        "metricRelabelings": [],
                        "namespace": "",
                        "params": {},
                        "path": "/s/actuator/prometheus",
                        "podTargetLabels": {},
                        "relabelings": [],
                        "scrapeTimeout": "",
                        "selector": {
                          "service-type": "normal"
                        },
                        "targetLabels": {}
                      }
                    }
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.0-debian-12-r4"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.0-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/extra/lib",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.5.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rbac": {
                    "create": true
                  },
                  "readinessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {
                      "service-type": "normal"
                    },
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "startupProbe": {
                    "enabled": true,
                    "failureThreshold": 12,
                    "initialDelaySeconds": 60,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 450,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnamilegacy/os-shell",
                      "tag": "12-debian-12-r49"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabsunsupported/release-operator@sha256:52914f17dc55dfaef55aed382e7269408fb676a213d12117eafa906d81506824 '
          createdAt: "2025-09-02T05:48:19Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql.\n\n```yaml\napiVersion:
          xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n
          \ global:\n    security:\n      allowInsecureImages: true\n  k8sSetup:\n
          \   platform: Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula:
          true\n  keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n```\n\n#### Configuration Details\n\nThe sample configuration
          uses:\n\n- Autogenerated limited license\n- Simple administrator initial
          password\n- Default keystore setup\n- Only one replica is set for all pods\n-
          All Pods are missing resource limit setup\n- Default storage class and minimal
          size for persistent storage\n- Release pods are using persistence access
          mode `ReadWriteOnce`, for Release pods the requirement is to have `ReadWriteMany`\n-
          Embedded PostgreSQL for DB management\n\n#### Security Configuration Details\n\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabsunsupported/release-operator@sha256:52914f17dc55dfaef55aed382e7269408fb676a213d12117eafa906d81506824
        - docker.io/xebialabsunsupported/xl-release@sha256:0548a1ba63da5ae09c59b67a30240bf48877ab8950e09b0f0f6fd7c88ba34c13
        version: 25.3.0-beta.813
      entries:
      - name: digitalai-release-operator.v25.3.0-beta.813
        version: 25.3.0-beta.813
      - name: digitalai-release-operator.v25.3.0-beta.711
        version: 25.3.0-beta.711
      - name: digitalai-release-operator.v25.3.0-beta.609
        version: 25.3.0-beta.609
      - name: digitalai-release-operator.v25.3.0-beta.512
        version: 25.3.0-beta.512
      name: beta
    - currentCSV: digitalai-release-operator.v24.1.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "fsGroup": null,
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "fsGroup": null,
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  jvm {\n      override-jvm-options = {{ .Values.jvm.overrideOptions }}\n      networkaddress.cache.ttl = {{ index .Values.jvm.options \"networkaddress.cache.ttl\" }}\n  }\n}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "ingressClass": "haproxy-dair",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-client/xl",
                        "kube",
                        "upgrade",
                        "--skip-context-check",
                        "--local-repo",
                        "/opt/xebialabs/xl-op-blueprints",
                        "--answers",
                        "/opt/xebialabs/xl-client/generated_answers.yaml"
                      ],
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    },
                    "overrideOptions": true
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "tag": "1.26.0-debian-12-r1"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "tag": "1.10.1-debian-12-r5"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "tag": "15.6.0-debian-12-r7"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "tag": "3.12.13-debian-12-r2"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnami/os-shell",
                      "tag": "12-debian-12-r16"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator without SCC creation and with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: docker.io/xebialabs/release-operator@sha256:47d6e31f66fce52c6566240b85fb028890409ad3addeea134120d1afa4814883
          createdAt: "2024-10-16T04:25:19Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind:
          DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    runAsUser: null\n    runAsGroup: null\n
          \   fsGroup: null\n  containerSecurityContext:\n    runAsUser: null\n    runAsGroup:
          null\n  securityContextConstraints:\n    enabled: false\n  volumePermissions:\n
          \   enabled: false\n  route:\n    enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n        runAsUser: null\n        runAsGroup: null\n        fsGroup:
          null\n      containerSecurityContext:\n        enabled: false\n        runAsUser:
          null\n        runAsGroup: null\n      securityContextConstraints:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n      runAsUser: null\n
          \     runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     enabled: false\n      runAsUser: null\n      runAsGroup: null\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"dair-doc-digitalai-release-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xlr-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Release pods are using
          persistence access mode `ReadWriteOnce`, for Release pods the requirement
          is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB management\n- Embedded
          RabbitMQ for message queue management\n\n#### Security Configuration Details\n\n-
          `securityContextConstraints.enabled: false` - disables creation of SCCs;\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:771a9a173e033a3ad8b46f5c00a7036eaa88c8d8d1fbd89217325168998113ea
        - docker.io/xebialabs/release-operator@sha256:47d6e31f66fce52c6566240b85fb028890409ad3addeea134120d1afa4814883
        - docker.io/xebialabs/xl-release@sha256:d1b65493ed524d37c60483f44b8bcc877ff398c81d36c7e767f2254dadec98cc
        version: 24.1.10
      entries:
      - name: digitalai-release-operator.v24.1.10
        version: 24.1.10
      name: stable
    - currentCSV: digitalai-release-operator.v24.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  jvm {\n      override-jvm-options = {{ .Values.jvm.overrideOptions }}\n      networkaddress.cache.ttl = {{ index .Values.jvm.options \"networkaddress.cache.ttl\" }}\n  }\n}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "image": {
                        "tag": "v0.14.7"
                      },
                      "ingressClass": "haproxy-dair",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-client/xl",
                        "kube",
                        "upgrade",
                        "--skip-context-check",
                        "--local-repo",
                        "/opt/xebialabs/xl-op-blueprints",
                        "--answers",
                        "/opt/xebialabs/xl-client/generated_answers.yaml"
                      ],
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    },
                    "overrideOptions": true
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.26.2-debian-12-r10"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "15.6.0-debian-12-r20"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.12.14-debian-12-r7"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnamilegacy/os-shell",
                      "tag": "12-debian-12-r49"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator without SCC creation and with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:aca76f0f5db170ddfc6ad692ea2874b9efb83fe732540946d88a0c078907ec8c '
          createdAt: "2025-08-22T08:54:25Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind:
          DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  securityContextConstraints:\n    enabled: false\n
          \ volumePermissions:\n    enabled: false\n  route:\n    enabled: false\n
          \   annotations:\n      haproxy.router.openshift.io/cookie_name: JSESSIONID\n
          \     haproxy.router.openshift.io/disable_cookies: 'false'\n      haproxy.router.openshift.io/rewrite-target:
          /\n      haproxy.router.openshift.io/timeout: 120s\n    hostname: '<mandatory-release-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n      securityContextConstraints:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   securityContextConstraints:\n      enabled: false\n    volumePermissions:\n
          \     enabled: false\n```\n\nIf SCC is required, it must be managed manually.
          The necessary permissions for accessing SecurityContextConstraints must
          be granted when SCC is enabled; otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"dair-doc-digitalai-release-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xlr-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Release pods are using
          persistence access mode `ReadWriteOnce`, for Release pods the requirement
          is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB management\n- Embedded
          RabbitMQ for message queue management\n\n#### Security Configuration Details\n\n-
          `securityContextConstraints.enabled: false` - disables creation of SCCs;\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/release-operator@sha256:aca76f0f5db170ddfc6ad692ea2874b9efb83fe732540946d88a0c078907ec8c
        - docker.io/xebialabs/xl-release@sha256:1cadd1d6ad7389ac9fe398ffba246b78238e54c180045b9dd44fdeefa3cf2af5
        version: 24.1.21
      entries:
      - name: digitalai-release-operator.v24.1.21
        version: 24.1.21
      - name: digitalai-release-operator.v24.1.20
        version: 24.1.20
      - name: digitalai-release-operator.v24.1.19
        version: 24.1.19
      - name: digitalai-release-operator.v24.1.18
        version: 24.1.18
      - name: digitalai-release-operator.v24.1.17
        version: 24.1.17
      - name: digitalai-release-operator.v24.1.14
        version: 24.1.14
      - name: digitalai-release-operator.v24.1.10
        version: 24.1.10
      name: stable-v24.1
    - currentCSV: digitalai-release-operator.v24.3.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "#!/bin/sh\n\ncd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "16.3.0-debian-12-r23"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n{{ if eq .Values.queueType \"quorum\" -}}\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n{{- end }}\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.13.7-debian-12-r5"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "queueType": "classic",
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-release",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:725fb5f7cfa4098c4110a40f54dc646f7663267bcba53603a7c1303c44051803 '
          createdAt: "2025-09-26T18:24:33Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n
          \ name: dair-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n```\n\n#### Configuration
          Details\n\nThe sample configuration uses:\n\n- Autogenerated limited license\n-
          Simple administrator initial password\n- Default keystore setup\n- Only
          one replica is set for all pods\n- All Pods are missing resource limit setup\n-
          Default storage class and minimal size for persistent storage\n- Release
          pods are using persistence access mode `ReadWriteOnce`, for Release pods
          the requirement is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB
          management\n- Embedded RabbitMQ for message queue management\n\n#### Security
          Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined ranges (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/release-operator@sha256:725fb5f7cfa4098c4110a40f54dc646f7663267bcba53603a7c1303c44051803
        - docker.io/xebialabs/xl-release@sha256:e9e87975888fcd4c8c164db90cb0a5c7abde88bf272c59d97e465af7b49bc6be
        version: 24.3.11
      entries:
      - name: digitalai-release-operator.v24.3.11
        version: 24.3.11
      - name: digitalai-release-operator.v24.3.10
        version: 24.3.10
      - name: digitalai-release-operator.v24.3.9
        version: 24.3.9
      - name: digitalai-release-operator.v24.3.8
        version: 24.3.8
      - name: digitalai-release-operator.v24.3.7
        version: 24.3.7
      - name: digitalai-release-operator.v24.3.6
        version: 24.3.6
      - name: digitalai-release-operator.v24.3.5
        version: 24.3.5
      name: stable-v24.3
    - currentCSV: digitalai-release-operator.v25.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "cluster": {
                    "management": {
                      "ssl": {
                        "enabled": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    },
                    "manager": "xlr-legacy",
                    "mode": "full",
                    "name": "xl-release_cluster",
                    "node": {
                      "datacenter": "default"
                    },
                    "remote": {
                      "sslEngine": {
                        "enabled": false,
                        "enabledAlgorithms": [
                          "TLS_DHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_DHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_RSA_WITH_AES_128_CBC_SHA"
                        ],
                        "hostnameVerification": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "requireMutualAuthentication": false,
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"${XLR_CLUSTER_NAME}\"\n    manager = \"${XLR_CLUSTER_MANAGER}\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n      {{- if .Values.storage.reportingEngine.downingProviderClass }}\n      cluster {\n        downing-provider-class = {{.Values.storage.reportingEngine.downingProviderClass | quote}}\n      }\n      {{- end }}\n\n      {{- if .Values.cluster.remote.sslEngine.enabled }}\n      remote {\n        artery {  \n          ssl {  \n            config-ssl-engine {  \n              enabled-algorithms = [\n                {{- $last := sub (len .Values.cluster.remote.sslEngine.enabledAlgorithms) 1 }}\n                {{ range $index, $v := .Values.cluster.remote.sslEngine.enabledAlgorithms }}\n                {{ $v | quote }}{{ if ne $index $last }},{{ end }}\n                {{- end }}\n              ]  \n              key-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEY_PASSWORD}\"  \n              key-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.keystore \"default\" (printf \"cluster_remote_keystore.%s\" .Values.cluster.remote.sslEngine.keystoreType) \"context\" $) }}\"  \n              key-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEYSTORE_PASSWORD}\"  \n              trust-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.truststore \"default\" (printf \"cluster_remote_truststore.%s\" .Values.cluster.remote.sslEngine.truststoreType) \"context\" $) }}\"  \n              trust-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_TRUSTSTORE_PASSWORD}\"  \n              hostname-verification = {{ if .Values.cluster.remote.sslEngine.hostnameVerification }}on{{ else }}off{{ end }}\n              require-mutual-authentication = {{ if .Values.cluster.remote.sslEngine.requireMutualAuthentication }}on{{ else }}off{{ end }}  \n            }  \n          }  \n          transport = tls-tcp  \n        }  \n      }\n      {{- end }}\n    }\n    {{- if .Values.cluster.management.ssl.enabled }}\n    xlr-pekko-native {\n      pekko {\n        management {\n          http {  \n            route-providers-read-only = true  \n          }  \n          ssl {  \n            enabled = true\n            keyPassword = \"${?CLUSTER_MANAGEMENT_SSL_KEY_PASSWORD}\"  \n            keyStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.keystore \"default\" (printf \"cluster_management_keystore.%s\" .Values.cluster.management.ssl.keystoreType) \"context\" $) }}\"  \n            keyStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_KEYSTORE_PASSWORD}\"  \n            keystoreType = {{ .Values.cluster.management.ssl.keystoreType }}  \n            trustStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.truststore \"default\" (printf \"cluster_management_truststore.%s\" .Values.cluster.management.ssl.truststoreType) \"context\" $) }}\"  \n            trustStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_TRUSTSTORE_PASSWORD}\"  \n            trustStoreType = {{ .Values.cluster.management.ssl.truststoreType }}  \n          }  \n        }  \n      }\n    }\n    {{- end }}\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\nmanagement.endpoint.prometheus.enabled = {{ .Values.metrics.prometheus.enabled | default false }}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      datacenter = ${XLR_CLUSTER_DATACENTER}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "#!/bin/sh\n\ncd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "livenessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/liveness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "jmx": {
                      "enabled": false
                    },
                    "prometheus": {
                      "enabled": false,
                      "podAnnotations": {
                        "prometheus.io/port": "{{ .Values.service.ports.metrics }}",
                        "prometheus.io/scrape": "true"
                      },
                      "rule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "serviceMonitor": {
                        "annotations": {},
                        "basicAuth": {
                          "password": {
                            "key": "releasePassword",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          },
                          "username": {
                            "key": "releaseUsername",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          }
                        },
                        "enabled": false,
                        "honorLabels": false,
                        "interval": "30s",
                        "jobLabel": "",
                        "labels": {},
                        "metricRelabelings": [],
                        "namespace": "",
                        "params": {},
                        "path": "/s/actuator/prometheus",
                        "podTargetLabels": {},
                        "relabelings": [],
                        "scrapeTimeout": "",
                        "selector": {
                          "service-type": "normal"
                        },
                        "targetLabels": {}
                      }
                    }
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.1-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.4.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.3-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "readinessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {
                      "service-type": "normal"
                    },
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "startupProbe": {
                    "enabled": true,
                    "failureThreshold": 12,
                    "initialDelaySeconds": 60,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 450,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-release",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:838443baf39514c5e0ac0a09cab411942a3cc4a13e261c8c60138788fe8d96e4 '
          createdAt: "2025-09-26T17:28:37Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n
          \ name: dair-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n```\n\n#### Configuration
          Details\n\nThe sample configuration uses:\n\n- Autogenerated limited license\n-
          Simple administrator initial password\n- Default keystore setup\n- Only
          one replica is set for all pods\n- All Pods are missing resource limit setup\n-
          Default storage class and minimal size for persistent storage\n- Release
          pods are using persistence access mode `ReadWriteOnce`, for Release pods
          the requirement is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB
          management\n- Embedded RabbitMQ for message queue management\n\n#### Security
          Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined ranges (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/release-operator@sha256:838443baf39514c5e0ac0a09cab411942a3cc4a13e261c8c60138788fe8d96e4
        - docker.io/xebialabs/xl-release@sha256:9b63789cdb87745832f1283d3f0e28ac5a4514d767895e0914a6be6672f2d14f
        version: 25.1.5
      entries:
      - name: digitalai-release-operator.v25.1.5
        version: 25.1.5
      - name: digitalai-release-operator.v25.1.4
        version: 25.1.4
      - name: digitalai-release-operator.v25.1.3
        version: 25.1.3
      - name: digitalai-release-operator.v25.1.2
        version: 25.1.2
      - name: digitalai-release-operator.v25.1.1
        version: 25.1.1
      - name: digitalai-release-operator.v25.1.0
        version: 25.1.0
      name: stable-v25.1
    defaultChannel: stable-v24.3
    packageName: digitalai-release-operator
    provider:
      name: Digital.ai
      url: https://digital.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Confluent, Inc.
      provider-url: https://confluent.io
    name: confluent-for-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: confluent-for-kubernetes.v2.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.8.4",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:0ca8769b04d6134ee231af84dfe98fb65161148d0d32e85f18dad6a3a345ed33
          createdAt: "2025-10-01T13:12:32Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0 <2.10.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - confluentinc/confluent-operator@sha256:0ca8769b04d6134ee231af84dfe98fb65161148d0d32e85f18dad6a3a345ed33
        - docker.io/confluentinc/confluent-operator@sha256:0ca8769b04d6134ee231af84dfe98fb65161148d0d32e85f18dad6a3a345ed33
        - docker.io/confluentinc/cp-server@sha256:45eb6e4f9c1201202d90216fbbdfcd8655eb378c937f09ac0129cd259956337d
        - docker.io/confluentinc/cp-server-connect@sha256:24aaa1cc8eb0df643f1efa37999b7ac33d844bfd1e4502da0d678996e4ac5ddd
        - docker.io/confluentinc/confluent-init-container@sha256:175ba68c876eca5804597b1591fdf551b2091d2f1c50e4daed2c32958af5d649
        - docker.io/confluentinc/cp-zookeeper@sha256:3e0af6ac6d34dcb8b9ed8f7a08109a48b2b2acd331c1dc4ea875222685c9afbd
        - docker.io/confluentinc/cp-schema-registry@sha256:bba639a3f9e2bf5f99b5d056ed520d2c268a125bee4f546d6cbca7a2c58f7dbb
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:2024d117dd413e207e49f37558d79dc882fd29f6b39e2d722708f18076813361
        - docker.io/confluentinc/cp-kafka-rest@sha256:8218e8db4731219ceaf9dbc378ead9c49aa86f98bc5106f62fbf70958dfd1ee4
        - docker.io/confluentinc/cp-ksqldb-server@sha256:093e87359c176d13d81317e972888a835cd01092d061bfc03fd60f0d1b4c8428
        version: 2.10.3
      entries:
      - name: confluent-for-kubernetes.v2.10.3
        version: 2.10.3
      name: "2.10"
    - currentCSV: confluent-for-kubernetes.v2.11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.9.3",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:85b512d3f20dc6e4a4da1759ee5ed306a4da0604592563a9af35be2cdf9ba1b6
          createdAt: "2025-10-01T13:10:39Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.10.0 <2.11.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/confluent-operator@sha256:85b512d3f20dc6e4a4da1759ee5ed306a4da0604592563a9af35be2cdf9ba1b6
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:351b68b062824c15db7ef999ba04062defdd4ea1ebe9665932147f99cf74879d
        - docker.io/confluentinc/cp-ksqldb-server@sha256:14a44cbbc13094974e7a5e14b33a829480ec3da75868130fdc7abaeda707dc5e
        - confluentinc/confluent-operator@sha256:85b512d3f20dc6e4a4da1759ee5ed306a4da0604592563a9af35be2cdf9ba1b6
        - docker.io/confluentinc/cp-server@sha256:e61e6c122b50ed345bc3e4e39cda6c93430e669535f8a70957e221563a7af1bd
        - docker.io/confluentinc/confluent-init-container@sha256:bccc7b723afdbe4e04c668720cdb1570835f8fd58ab7dd9a06ecf807f1176019
        - docker.io/confluentinc/cp-kafka-rest@sha256:cc72b3585ef9e7557910ab3efe3082444abe1fee191cdfb8747907fd2643971f
        - docker.io/confluentinc/cp-schema-registry@sha256:bd974dda5abf35fcad66c25fad92ee684e82305435b291b953d221f74d2bc119
        - docker.io/confluentinc/cp-zookeeper@sha256:03eadde2afa390f7339707e6baef2ecf7affb898aaa432378f61da54f5636ba3
        - docker.io/confluentinc/cp-server-connect@sha256:0d3b81e511cdf820daac4a2458b3cada40b4a7c3d142062a48880d38186bbcc4
        version: 2.11.3
      entries:
      - name: confluent-for-kubernetes.v2.11.3
        version: 2.11.3
      name: "2.11"
    - currentCSV: confluent-for-kubernetes.v2.8.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.6.4",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:551b14777b121b1e083524a0d5cce03cbe2fd2833537751bcd6be6cf9801f95e
          createdAt: "2024-12-02T06:51:43Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.0 <2.8.5'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-server@sha256:1b2ae325f87047b83d37d506066f073c27b0c42a0fd3c3409aa079ba736d1975
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:01bcde2aa35c0ad062ece49c63d41bf487618ea3c33e42ea4c23fde7194b26c9
        - docker.io/confluentinc/cp-schema-registry@sha256:dec88d30c8f21bc7fb38d28603c97c81c666ccaee63ff0136ecff48b3b3ec775
        - docker.io/confluentinc/confluent-init-container@sha256:b1ad1aa4daa81967a93b02323172f96db6a50c7135866f016344b2eb1a736a55
        - docker.io/confluentinc/cp-server-connect@sha256:255875bc57d81c322d47d5f6f11bc217e99ae709d1fdc5aec8e11829dd34279a
        - confluentinc/confluent-operator@sha256:551b14777b121b1e083524a0d5cce03cbe2fd2833537751bcd6be6cf9801f95e
        - docker.io/confluentinc/cp-zookeeper@sha256:7e728feebfc90a693d028216da68a5d79e5e8212855c63bac08e5b10b141d143
        - docker.io/confluentinc/cp-kafka-rest@sha256:2a5bc0d39693682e25d79f5036e18b71376df0941fedc2cb962a4b4d5a5eff82
        - docker.io/confluentinc/cp-ksqldb-server@sha256:50e4834c44759ab4a2a0bbab0f20545e124325b7ee12ba1a275af73b57084104
        - docker.io/confluentinc/confluent-operator@sha256:551b14777b121b1e083524a0d5cce03cbe2fd2833537751bcd6be6cf9801f95e
        version: 2.8.5
      entries:
      - name: confluent-for-kubernetes.v2.8.5
        version: 2.8.5
      name: "2.8"
    - currentCSV: confluent-for-kubernetes.v2.9.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.7.5",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:20648758fdb98d0f8913e75fb919ad0b622de8ecda9ab394912103e5e17d4007
          createdAt: "2025-10-03T06:52:55Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0 <2.9.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-server-connect@sha256:560833472dbdf23cf85b780e8306ae1bfe16ce6e2c0a9b326c64a513117525f1
        - docker.io/confluentinc/cp-zookeeper@sha256:2bd46d1f5d1eec0675bb5d1c08634d6458bbac7cacfb233aa61c68137fd4a7ce
        - docker.io/confluentinc/cp-kafka-rest@sha256:b6ac629a880474d81182eaf8459b4d83fe8126db796f1f285798beecda55a50a
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:257c38a75b3f4241780e20f16a7adc026dcca590f2e9c8da7c4e86b961e7611f
        - confluentinc/confluent-operator@sha256:20648758fdb98d0f8913e75fb919ad0b622de8ecda9ab394912103e5e17d4007
        - docker.io/confluentinc/confluent-operator@sha256:20648758fdb98d0f8913e75fb919ad0b622de8ecda9ab394912103e5e17d4007
        - docker.io/confluentinc/cp-ksqldb-server@sha256:6817a555ae4c041869f9b422a115c22894b2cd3988e82dc601694ce713b3ccd4
        - docker.io/confluentinc/confluent-init-container@sha256:b138d954318cc98767c10b4564a1412a22766317a83b9a452285bb0b3d893c63
        - docker.io/confluentinc/cp-server@sha256:79243e7e8b721f9fdf26fc57330ebb27d69c3b4686dbdf5319f6f39c2c7be836
        - docker.io/confluentinc/cp-schema-registry@sha256:6506a8c659d4154ae3e2d702f08b4ccac63513392ba454cbc9cf384d070c46eb
        version: 2.9.7
      entries:
      - name: confluent-for-kubernetes.v2.9.7
        version: 2.9.7
      name: "2.9"
    - currentCSV: confluent-for-kubernetes.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:8.0.1",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-ng-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "alertManagerClient": {
                      "url": "http://controlcenter-next-gen.confluent.svc.cluster.local:9093"
                    },
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    },
                    "prometheusClient": {
                      "url": "http://controlcenter-next-gen.confluent.svc.cluster.local:9090"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center-next-gen:2.2.0",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "replicas": 1,
                  "services": {
                    "alertmanager": {
                      "image": "confluentinc/cp-enterprise-alertmanager:2.2.0"
                    },
                    "prometheus": {
                      "image": "confluentinc/cp-enterprise-prometheus:2.2.0",
                      "pvc": {
                        "dataVolumeCapacity": "10Gi"
                      }
                    }
                  },
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:8.0.1",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:8.0.1",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:8.0.1",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:8.0.1",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:8.0.1",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:55b61717690fe334b5eebf5e7a55590537abddbd404f2eede81d4dd8cd29eda4
          createdAt: "2025-10-03T07:59:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <3.0.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-kafka-rest@sha256:f45f23e891ae073037ed7670659ecb6c3f72f9eea33e109f16243eec24a287a3
        - docker.io/confluentinc/cp-server-connect@sha256:4739f3a07d9f8eadd58f448a8e05b1392ef9b1dc9e9085d25a71f32b5b806a6e
        - docker.io/confluentinc/cp-ksqldb-server@sha256:df94bb805cff4f92315135a042f650c42f887bca70a40df8ecef2603532f11a7
        - docker.io/confluentinc/cp-enterprise-prometheus@sha256:db0989ef9859dbc88bc59f9b8ade86f2ba13e9dc36533e654566218328e5ff57
        - docker.io/confluentinc/cp-enterprise-control-center-next-gen@sha256:6378a0174433685ee52a7fe2e48424d3f598827ed2a5570ec74b916102485e0d
        - confluentinc/confluent-operator@sha256:55b61717690fe334b5eebf5e7a55590537abddbd404f2eede81d4dd8cd29eda4
        - docker.io/confluentinc/confluent-operator@sha256:55b61717690fe334b5eebf5e7a55590537abddbd404f2eede81d4dd8cd29eda4
        - docker.io/confluentinc/cp-schema-registry@sha256:2152975dfb62afbdf1d438dd0eb1f7fc3c929adf5804fe4790eccda0c96b9208
        - docker.io/confluentinc/cp-server@sha256:207cabc12e0d122199da875b2b0e39d1cf6f7fe435d5f184a9c8d7dc57ea81d9
        - docker.io/confluentinc/cp-enterprise-alertmanager@sha256:fcce6f5311c26cd58427a039530a9f8dfd009249ff5a519694aa945d34502252
        - docker.io/confluentinc/confluent-init-container@sha256:bf38b3bd9851bfd3ef0635e93b9e8b9ada570d393f650f60d8ae8943537730dc
        version: 3.0.1
      entries:
      - name: confluent-for-kubernetes.v3.0.1
        version: 3.0.1
      name: "3.0"
    defaultChannel: "3.0"
    packageName: confluent-for-kubernetes
    provider:
      name: Confluent, Inc.
      url: https://confluent.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pachyderm, Inc.
      provider-url: https://pachyderm.com
    name: pachyderm-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pachyderm-operator.v0.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "Pachyderm",
                "metadata": {
                  "name": "pachyderm-sample"
                },
                "spec": {
                  "console": {
                    "disable": true
                  },
                  "pachd": {
                    "metrics": {
                      "disable": false
                    },
                    "storage": {
                      "amazon": {
                        "credentialSecretName": "pachyderm-aws-secret"
                      },
                      "backend": "AMAZON"
                    }
                  },
                  "version": "v2.1.6"
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermExport",
                "metadata": {
                  "name": "pachydermexport-sample"
                },
                "spec": {
                  "backup": {
                    "target": "pachyderm-sample"
                  }
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermImport",
                "metadata": {
                  "name": "pachydermimport-sample"
                },
                "spec": {
                  "backup": "backups/pachyderm-backup-20220610.tar.gz",
                  "destination": {
                    "name": "pachyderm-restore",
                    "namespace": "pachyderm-workloads"
                  },
                  "storageSecret": "s3-credentials-secret"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
          createdAt: "2021-05-20T08:05:00Z"
          description: Pachyderm provides the data layer that allows data science
            teams to productionize and scale their machine learning lifecycle with
            data driven automation, petabyte scalability and end-to-end reproducibility.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/pachyderm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/pachyderm-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pachyderm/openshift-operator
          support: Pachyderm, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PachydermExport is the Schema for the pachydermexports API
            displayName: Pachyderm Export
            kind: PachydermExport
            name: pachydermexports.aiml.pachyderm.com
            version: v1beta1
          - description: PachydermImport is the Schema for the pachydermimports API
            displayName: Pachyderm Import
            kind: PachydermImport
            name: pachydermimports.aiml.pachyderm.com
            version: v1beta1
          - description: Pachyderm is the Schema for the pachyderms API
            displayName: Pachyderm
            kind: Pachyderm
            name: pachyderms.aiml.pachyderm.com
            version: v1beta1
        description: "The Pachyderm operator aims to make deploying and managing Pachyderm
          instances on Openshift container platform easier.\n\nPachyderm is the data
          foundation for machine learning.  Pachyderm provides industry leading data
          versioning, pipelines and lineage that allow data science teams to automate
          the machine learning lifecycle and optimize their machine learning operations
          (MLOps). \n\n## Getting started\n\n1. Pachyderm uses an object store to
          store your data. Set the pachd storage backend at `pachyderm.spec.pachd.storage.backend`
          to `AMAZON`, `MICROSOFT`, `GOOGLE` or `MINIO` depending on your storage
          provider. Additionally, create a user with read/write access to the object
          store.\n\n2. Setup the user credentials in the section, `pachyderm.spec.pachd.storage`.
          Required fields depend on the backend type set in step 1 above.\n\n3. Finally,
          deploy the pachyderm cluster in its own namespace.\n\nFor more information
          on getting started, reference the documentation.\n"
        displayName: Pachyderm
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pachyderm
        - versioning
        - machine learning
        - artificial intelligence
        links:
        - name: Documentation
          url: https://docs.pachyderm.com/latest/
        maintainers:
        - email: kyle@pachyderm.io
          name: Kyle White
        - email: eochieng@redhat.com
          name: Edmund Ochieng
        maturity: beta
        minKubeVersion: v1.19.0
        provider:
          name: Pachyderm, Inc.
          url: https://pachyderm.com
        relatedImages:
        - registry.connect.redhat.com/pachyderm/worker@sha256:374f83d92899707bc87b0318fda6bffa78d2e9e523b827bb3d40693a2d1bdecd
        - registry.connect.redhat.com/pachyderm/etcd@sha256:7f330b51874379b1d4b71d94e37e0ecfa52b86d628296136970688cf204f53b5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
        - registry.connect.redhat.com/pachyderm/backup-handler@sha256:e2ef6bd8191e8ad58c765c0db8f5bb21ea7123caed0c0f58fe681a024201306d
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/pachd@sha256:116477fefdf413ea02c6ac3d013f1f91d3f6e4b16aed2408c9890832cdb41430
        - registry.redhat.io/rhel8/postgresql-13@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/init-utils@sha256:5c0258bfa0203079925648c0a7dda6d05a29b6f461854821a1eec541777ff359
        - registry.connect.redhat.com/pachyderm/pgbouncer@sha256:c3b6949227df24abb8934d111a34b44e87fb16194ac34a5d0fe284cff02e9ac0
        version: 0.1.4
      entries:
      - name: pachyderm-operator.v0.1.4
        version: 0.1.4
      name: stable
    defaultChannel: stable
    packageName: pachyderm-operator-rhmp
    provider:
      name: Pachyderm, Inc.
      url: https://pachyderm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: local-storage-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: local-storage-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "local.storage.openshift.io/v1",
                "kind": "LocalVolume",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "storageClassDevices": [
                    {
                      "devicePaths": [
                        "/dev/disk/by-id/ata-crucial"
                      ],
                      "fsType": "ext4",
                      "storageClassName": "foobar",
                      "volumeMode": "Filesystem"
                    }
                  ]
                }
              },
              {
                "apiVersion": "local.storage.openshift.io/v1alpha1",
                "kind": "LocalVolumeSet",
                "metadata": {
                  "name": "example-localvolumeset"
                },
                "spec": {
                  "deviceInclusionSpec": {
                    "deviceMechanicalProperties": [
                      "Rotational",
                      "NonRotational"
                    ],
                    "deviceTypes": [
                      "disk"
                    ],
                    "maxSize": "100G",
                    "minSize": "10G"
                  },
                  "maxDeviceCount": 10,
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "storageClassName": "example-storageclass",
                  "volumeMode": "Block"
                }
              },
              {
                "apiVersion": "local.storage.openshift.io/v1alpha1",
                "kind": "LocalVolumeDiscovery",
                "metadata": {
                  "name": "auto-discover-devices"
                },
                "spec": {
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-local-storage-rhel9-operator@sha256:6f6491f023ffa21eff6a9b938cb1869f362d6b4a5b04c27dc5f13f49622ca178
          createdAt: "2019-08-14T00:00:00Z"
          description: Configure and use local storage volumes.
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.17"}]'
          olm.skipRange: '>=4.3.0 <4.16.0-202509240123'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-local-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["localvolumediscoveryresults.local.storage.openshift.io"]'
          repository: https://github.com/openshift/local-storage-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manage local storage volumes for OpenShift
            displayName: Local Volume
            kind: LocalVolume
            name: localvolumes.local.storage.openshift.io
            version: v1
          - description: A Local Volume set allows you to filter a set of storage
              volumes, group them and create a dedicated storage class to consume
              storage from the set of volumes.
            displayName: Local Volume Set
            kind: LocalVolumeSet
            name: localvolumesets.local.storage.openshift.io
            version: v1alpha1
          - description: Discover list of potentially usable disks on the chosen set
              of nodes
            displayName: Local Volume Discovery
            kind: LocalVolumeDiscovery
            name: localvolumediscoveries.local.storage.openshift.io
            version: v1alpha1
          - description: Disc inventory of available disks from selected nodes
            displayName: Local Volume Discovery Result
            kind: LocalVolumeDiscoveryResult
            name: localvolumediscoveryresults.local.storage.openshift.io
            version: v1alpha1
        description: |
          Operator that configures local storage volumes for use in Kubernetes and OpenShift. OpenShift 4.2 and above are the only supported OpenShift versions.
        displayName: Local Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - local storage
        links:
        - name: Documentation
          url: https://github.com/openshift/local-storage-operator/tree/master/docs
        - name: Source Repository
          url: https://github.com/openshift/local-storage-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-local-storage-rhel9-operator@sha256:6f6491f023ffa21eff6a9b938cb1869f362d6b4a5b04c27dc5f13f49622ca178
        - registry.redhat.io/openshift4/ose-local-storage-diskmaker-rhel9@sha256:d8d8b21fce5930a36c9755d82037fce1ab81f5ae2d5fb1bfd643d500c4ec7fba
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel9@sha256:ec5847d5006d66a014315decd579e10104b086a06bfefa33430545b4ffcece71
        version: 4.16.0-202509240123
      entries:
      - name: local-storage-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: local-storage-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: local-storage-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: local-storage-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: local-storage-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: local-storage-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: local-storage-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: local-storage-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: local-storage-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: local-storage-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: local-storage-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: local-storage-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: local-storage-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: local-storage-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: local-storage-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: local-storage-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: local-storage-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: local-storage-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: local-storage-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: local-storage-operator.v4.16.0-202410031007
        version: 4.16.0-202410031007
      - name: local-storage-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: local-storage-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: local-storage-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: local-storage-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: local-storage-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: local-storage-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: local-storage-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: local-storage-operator.v4.16.0-202408021139
        version: 4.16.0-202408021139
      - name: local-storage-operator.v4.16.0-202407190036
        version: 4.16.0-202407190036
      - name: local-storage-operator.v4.16.0-202407111006
        version: 4.16.0-202407111006
      - name: local-storage-operator.v4.16.0-202407081937
        version: 4.16.0-202407081937
      - name: local-storage-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: local-storage-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: local-storage-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: local-storage-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: local-storage-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-community-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.5.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "registry": "docker.io",
                  "tag": "5.4.6",
                  "oem": "",
                  "imagePullSecrets": "",
                  "psp": false,
                  "rbac": true,
                  "serviceAccount": "default",
                  "leastPrivilege": false,
                  "global": {
                    "cattle": {
                      "url": ""
                    },
                    "azure": {
                      "enabled": false,
                      "identity": {
                        "clientId": "DONOTMODIFY"
                      },
                      "marketplace": {
                        "planId": "DONOTMODIFY"
                      },
                      "extension": {
                        "resourceId": "DONOTMODIFY"
                      },
                      "serviceAccount": "csp",
                      "imagePullSecrets": "",
                      "images": {
                        "neuvector_csp_pod": {
                          "tag": "latest",
                          "image": "neuvector-billing-azure-by-suse-llc",
                          "registry": "registry.suse.de/suse/sle-15-sp5/update/pubclouds/images",
                          "imagePullPolicy": "IfNotPresent"
                        },
                        "controller": {
                          "tag": "5.2.4",
                          "image": "controller",
                          "registry": "docker.io/neuvector"
                        },
                        "manager": {
                          "tag": "5.2.4",
                          "image": "manager",
                          "registry": "docker.io/neuvector"
                        },
                        "enforcer": {
                          "tag": "5.2.4",
                          "image": "enforcer",
                          "registry": "docker.io/neuvector"
                        }
                      }
                    },
                    "aws": {
                      "enabled": false,
                      "accountNumber": "",
                      "roleName": "",
                      "serviceAccount": "csp",
                      "annotations": {},
                      "imagePullSecrets": "",
                      "image": {
                        "digest": "",
                        "repository": "neuvector/neuvector-csp-adapter",
                        "tag": "latest",
                        "imagePullPolicy": "IfNotPresent"
                      }
                    }
                  },
                  "bootstrapPassword": "",
                  "autoGenerateCert": true,
                  "defaultValidityPeriod": 365,
                  "internal": {
                    "certmanager": {
                      "enabled": false,
                      "secretname": "neuvector-internal"
                    },
                    "autoGenerateCert": true,
                    "autoRotateCert": true
                  },
                  "controller": {
                    "enabled": true,
                    "annotations": {},
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": {
                      "repository": "neuvector/controller",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "searchRegistries": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "nodePort": "",
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      }
                    },
                    "ranchersso": {
                      "enabled": false
                    },
                    "pvc": {
                      "enabled": false,
                      "existingClaim": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {
                        "userinitcfg.yaml": {
                          "users": [
                            {
                              "Fullname": "admin",
                              "Password": "",
                              "Role": "admin"
                            }
                          ]
                        }
                      }
                    },
                    "certupgrader": {
                      "env": [],
                      "schedule": "",
                      "imagePullPolicy": "IfNotPresent",
                      "timeout": 3600,
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "prime": {
                      "enabled": false,
                      "image": {
                        "repository": "neuvector/compliance-config",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "1.0.7",
                        "hash": ""
                      }
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": {
                      "repository": "neuvector/enforcer",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "env": [],
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/etcd"
                      }
                    ],
                    "resources": {},
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "manager": {
                    "enabled": true,
                    "image": {
                      "repository": "neuvector/manager",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "priorityClassName": "",
                    "env": {
                      "ssl": true,
                      "envs": []
                    },
                    "svc": {
                      "type": "ClusterIP",
                      "nodePort": "",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": "",
                      "tls": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "topologySpreadConstraints": [],
                    "affinity": {},
                    "podLabels": {},
                    "podAnnotations": {},
                    "tolerations": [],
                    "nodeSelector": {},
                    "runAsUser": "",
                    "probes": {
                      "enabled": false,
                      "timeout": 1,
                      "periodSeconds": 10,
                      "startupFailureThreshold": 30
                    }
                  },
                  "cve": {
                    "adapter": {
                      "enabled": false,
                      "image": {
                        "repository": "neuvector/registry-adapter",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "0.1.9",
                        "hash": ""
                      },
                      "priorityClassName": "",
                      "resources": {},
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt"
                      },
                      "harbor": {
                        "protocol": "https",
                        "secretName": ""
                      },
                      "svc": {
                        "type": "ClusterIP",
                        "loadBalancerIP": "",
                        "annotations": {}
                      },
                      "route": {
                        "enabled": true,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      },
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ingressClassName": "",
                        "path": "/",
                        "annotations": {
                          "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                        },
                        "tls": false,
                        "secretName": ""
                      },
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      }
                    },
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "cacert": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                      "image": {
                        "registry": "",
                        "repository": "neuvector/updater",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "0.0.5",
                        "hash": ""
                      },
                      "schedule": "0 0 * * *",
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": {
                        "registry": "",
                        "repository": "neuvector/scanner",
                        "imagePullPolicy": "Always",
                        "tag": "6",
                        "hash": ""
                      },
                      "priorityClassName": "",
                      "resources": {},
                      "topologySpreadConstraints": [],
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      },
                      "volumes": "",
                      "volumeMounts": ""
                    }
                  },
                  "resources": {},
                  "runtimePath": "",
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhooksvc": {
                    "enabled": true
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  },
                  "lease": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "false"
          containerImage: docker.io/neuvector/neuvector-operator@sha256:6815df6435eafaf7b2bc9d5e3f0c9b65f0a50ad4e05d2862e56b227e19aeca27
          createdAt: "2019-11-18T02:09:59Z"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: NeuVector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart. NeuVector Operator version 1.5.5 deploys NeuVector app version 5.4.6.

          **Complete below steps to create namespace and Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Privileged SCC is added to Service Account specified in the deploment yaml by Operator version 1.3.4 and above in new deployment. In case of upgrading NeuVector Operator from previous version to 1.3.4, please delete Privileged SCC before upgrading.

                   oc delete rolebinding -nneuvector system:openshift:scc:privileged

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s  neuvector/default


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - wavefront
        - wavefront collector
        - wavefront proxy
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - docker.io/neuvector/neuvector-operator@sha256:6815df6435eafaf7b2bc9d5e3f0c9b65f0a50ad4e05d2862e56b227e19aeca27
        version: 1.5.5
      entries:
      - name: neuvector-operator.v1.5.5
        version: 1.5.5
      - name: neuvector-operator.v1.5.4
        version: 1.5.4
      - name: neuvector-operator.v1.5.3
        version: 1.5.3
      - name: neuvector-operator.v1.5.2
        version: 1.5.2
      - name: neuvector-operator.v1.5.1
        version: 1.5.1
      - name: neuvector-operator.v1.5.0
        version: 1.5.0
      - name: neuvector-operator.v1.4.9
        version: 1.4.9
      - name: neuvector-operator.v1.4.8
        version: 1.4.8
      - name: neuvector-operator.v1.4.7
        version: 1.4.7
      - name: neuvector-operator.v1.4.6
        version: 1.4.6
      - name: neuvector-operator.v1.4.5
        version: 1.4.5
      - name: neuvector-operator.v1.4.4
        version: 1.4.4
      - name: neuvector-operator.v1.4.3
        version: 1.4.3
      - name: neuvector-operator.v1.4.2
        version: 1.4.2
      - name: neuvector-operator.v1.4.1
        version: 1.4.1
      - name: neuvector-operator.v1.4.0
        version: 1.4.0
      - name: neuvector-operator.v1.3.9
        version: 1.3.9
      - name: neuvector-operator.v1.3.8
        version: 1.3.8
      - name: neuvector-operator.v1.3.7
        version: 1.3.7
      - name: neuvector-operator.v1.3.6
        version: 1.3.6
      - name: neuvector-operator.v1.3.5
        version: 1.3.5
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-community-operator
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: mariadb-operator
      provider-url: https://github.com/mariadb-operator/mariadb-operator
    name: mariadb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-operator.v25.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.mariadb.mmontes.io/v1alpha1",
                "kind": "MariadbOperator",
                "metadata": {
                  "name": "mariadb-operator"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "caValidity": "35064h",
                    "certValidity": "8766h",
                    "enabled": true,
                    "extrArgs": [],
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "ha": {
                      "enabled": false,
                      "replicas": 3
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "lookaheadValidity": "2160h",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podSecurityContext": {},
                    "requeueDuration": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "automount": true,
                      "enabled": true,
                      "extraLabels": {},
                      "name": ""
                    },
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    },
                    "tolerations": []
                  },
                  "clusterName": "cluster.local",
                  "extrArgs": [],
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "fullnameOverride": "",
                  "ha": {
                    "enabled": false,
                    "replicas": 3
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "logLevel": "INFO",
                  "metrics": {
                    "enabled": false,
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "rbac": {
                    "enabled": true
                  },
                  "resources": {},
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "automount": true,
                    "enabled": true,
                    "extraLabels": {},
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "annotations": {},
                    "cert": {
                      "caPath": "/tmp/k8s-webhook-server/certificate-authority",
                      "certManager": {
                        "duration": "",
                        "enabled": false,
                        "issuerRef": {},
                        "renewBefore": ""
                      },
                      "path": "/tmp/k8s-webhook-server/serving-certs",
                      "secretAnnotations": {}
                    },
                    "extrArgs": [],
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "ha": {
                      "enabled": false,
                      "replicas": 3
                    },
                    "hostNetwork": false,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podSecurityContext": {},
                    "port": 10250,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "automount": true,
                      "enabled": true,
                      "extraLabels": {},
                      "name": ""
                    },
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    },
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup"
                },
                "spec": {
                  "args": [
                    "--single-transaction",
                    "--all-databases"
                  ],
                  "logLevel": "info",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "backups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Connection",
                "metadata": {
                  "name": "connection"
                },
                "spec": {
                  "database": "mariadb",
                  "healthCheck": {
                    "interval": "10s",
                    "retryInterval": "3s"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "params": {
                    "parseTime": "true"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "secretName": "connection",
                  "secretTemplate": {
                    "annotations": {
                      "k8s.mariadb.com/connection": "sample"
                    },
                    "databaseKey": "database",
                    "hostKey": "host",
                    "key": "dsn",
                    "labels": {
                      "k8s.mariadb.com/connection": "sample"
                    },
                    "passwordKey": "password",
                    "portKey": "port",
                    "usernameKey": "username"
                  },
                  "serviceName": "mariadb",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Database",
                "metadata": {
                  "name": "data-test"
                },
                "spec": {
                  "characterSet": "utf8",
                  "collate": "utf8_general_ci",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "retryInterval": "5s"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "ExternalMariaDB",
                "metadata": {
                  "name": "external-mariadb"
                },
                "spec": {
                  "connection": {
                    "healthCheck": {
                      "interval": "5s"
                    },
                    "secretName": "external-mariadb"
                  },
                  "host": "mariadb.default.svc.cluster.local",
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "port": 3306,
                  "username": "root"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "grant"
                },
                "spec": {
                  "database": "*",
                  "grantOption": true,
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "privileges": [
                    "SELECT",
                    "INSERT",
                    "UPDATE"
                  ],
                  "requeueInterval": "30s",
                  "retryInterval": "5s",
                  "table": "*",
                  "username": "user"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "MariaDB",
                "metadata": {
                  "name": "mariadb"
                },
                "spec": {
                  "connection": {
                    "secretName": "mariadb-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "database": "mariadb",
                  "galera": {
                    "enabled": true
                  },
                  "metrics": {
                    "enabled": true,
                    "passwordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "mariadb-metrics"
                    }
                  },
                  "myCnf": "[mariadb]\nbind-address=*\ndefault_storage_engine=InnoDB\nbinlog_format=row\ninnodb_autoinc_lock_mode=2\ninnodb_buffer_pool_size=1024M\nmax_allowed_packet=256M\n",
                  "passwordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-password"
                  },
                  "primaryConnection": {
                    "secretName": "mariadb-conn-primary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "primaryService": {
                    "type": "ClusterIP"
                  },
                  "replicas": 3,
                  "rootPasswordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-root"
                  },
                  "secondaryConnection": {
                    "secretName": "mariadb-conn-secondary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "secondaryService": {
                    "type": "ClusterIP"
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "storage": {
                    "size": "1Gi"
                  },
                  "updateStrategy": {
                    "type": "ReplicasFirstPrimaryLast"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "MaxScale",
                "metadata": {
                  "name": "maxscale-galera"
                },
                "spec": {
                  "admin": {
                    "guiEnabled": true,
                    "port": 8989
                  },
                  "auth": {
                    "generate": true
                  },
                  "config": {
                    "sync": {
                      "database": "mysql",
                      "interval": "5s",
                      "timeout": "10s"
                    }
                  },
                  "connection": {
                    "port": 3306,
                    "secretName": "mxs-galera-conn"
                  },
                  "guiKubernetesService": {
                    "metadata": {
                      "annotations": {
                        "metallb.universe.tf/loadBalancerIPs": "172.18.0.231"
                      }
                    },
                    "type": "LoadBalancer"
                  },
                  "kubernetesService": {
                    "annotations": {
                      "metallb.universe.tf/loadBalancerIPs": "172.18.0.224"
                    },
                    "type": "LoadBalancer"
                  },
                  "mariaDbRef": {
                    "name": "mariadb-galera"
                  },
                  "monitor": {
                    "cooperativeMonitoring": "majority_of_all",
                    "interval": "2s",
                    "params": {
                      "available_when_donor": "false",
                      "disable_master_failback": "false",
                      "disable_master_role_setting": "false"
                    }
                  },
                  "replicas": 3,
                  "requeueInterval": "10s",
                  "services": [
                    {
                      "listener": {
                        "params": {
                          "connection_metadata": "tx_isolation=auto"
                        },
                        "port": 3306,
                        "protocol": "MariaDBProtocol"
                      },
                      "name": "rw-router",
                      "params": {
                        "master_accept_reads": "true",
                        "max_replication_lag": "3s",
                        "max_slave_connections": "255",
                        "transaction_replay": "true",
                        "transaction_replay_attempts": "10",
                        "transaction_replay_timeout": "5s"
                      },
                      "router": "readwritesplit"
                    },
                    {
                      "listener": {
                        "port": 3307
                      },
                      "name": "rconn-master-router",
                      "params": {
                        "master_accept_reads": "true",
                        "max_replication_lag": "3s",
                        "router_options": "master"
                      },
                      "router": "readconnroute"
                    },
                    {
                      "listener": {
                        "port": 3308
                      },
                      "name": "rconn-slave-router",
                      "params": {
                        "max_replication_lag": "3s",
                        "router_options": "slave"
                      },
                      "router": "readconnroute"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "PhysicalBackup",
                "metadata": {
                  "name": "physicalbackup"
                },
                "spec": {
                  "compression": "bzip2",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "podAffinity": true,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "immediate": true,
                    "suspend": false
                  },
                  "serviceAccountName": "backup",
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "physicalbackups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "region": "us-east-1",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  },
                  "timeout": "1h"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore"
                },
                "spec": {
                  "backupRef": {
                    "name": "backup"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "targetRecoveryTime": "2023-12-19T09:00:00Z"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "SqlJob",
                "metadata": {
                  "name": "03-stars"
                },
                "spec": {
                  "database": "mariadb",
                  "dependsOn": [
                    {
                      "name": "01-users"
                    },
                    {
                      "name": "02-repos"
                    }
                  ],
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "sql": "CREATE TABLE IF NOT EXISTS stars (\n  id bigint PRIMARY KEY AUTO_INCREMENT,\n  user_id bigint NOT NULL,\n  repo_id bigint NOT NULL,\n  FOREIGN KEY (user_id) REFERENCES users(id) ON DELETE CASCADE,\n  FOREIGN KEY (repo_id) REFERENCES repos(id) ON DELETE CASCADE,\n  UNIQUE KEY (user_id, repo_id)\n);\nINSERT INTO stars(user_id, repo_id) \n  VALUES((SELECT id FROM users ORDER BY RAND() LIMIT 1), (SELECT id FROM repos ORDER BY RAND() LIMIT 1))\n  ON DUPLICATE KEY UPDATE id=id;\nDELETE FROM stars WHERE id = (SELECT id FROM stars ORDER BY RAND() LIMIT 1);\nSELECT r.name AS repo, COUNT(*) AS stars\nFROM stars s\nJOIN repos r\nON s.repo_id = r.id\nGROUP BY r.id\nORDER BY stars DESC;",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "user"
                },
                "spec": {
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxUserConnections": 20,
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "retryInterval": "5s"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: ghcr.io/mariadb-operator/mariadb-operator-helm:25.8.4
          createdAt: "2025-09-19T13:24:35Z"
          description: Run and operate MariaDB in a cloud native way
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mariadb-operator/mariadb-operator
          support: mariadb-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configures a backup
            displayName: Backup
            kind: Backup
            name: backups.k8s.mariadb.com
            version: v1alpha1
          - description: Configures a connection
            displayName: Connection
            kind: Connection
            name: connections.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a logical database
            displayName: Database
            kind: Database
            name: databases.k8s.mariadb.com
            version: v1alpha1
          - description: Configures an external MariaDB resources
            displayName: ExternalMariaDB
            kind: ExternalMariaDB
            name: externalmariadbs.k8s.mariadb.com
            version: v1alpha1
          - description: Grants permissions to an user in a database
            displayName: Grant
            kind: Grant
            name: grants.k8s.mariadb.com
            version: v1alpha1
          - description: Configures MariaDB helm chart based operator
            displayName: MariadbOperator
            kind: MariadbOperator
            name: mariadboperators.helm.mariadb.mmontes.io
            version: v1alpha1
          - description: Provisions a MariaDB instance
            displayName: MariaDB
            kind: MariaDB
            name: mariadbs.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a MaxScale database proxy
            displayName: MaxScale
            kind: MaxScale
            name: maxscales.k8s.mariadb.com
            version: v1alpha1
          - description: Configures a physical backup
            displayName: PhysicalBackup
            kind: PhysicalBackup
            name: physicalbackups.k8s.mariadb.com
            version: v1alpha1
          - description: Restores a backup
            displayName: Restore
            kind: Restore
            name: restores.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a SQL job
            displayName: SqlJob
            kind: SqlJob
            name: sqljobs.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a user
            displayName: User
            kind: User
            name: users.k8s.mariadb.com
            version: v1alpha1
        description: |
          Install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) via [OLM](https://olm.operatorframework.io/) using the [helm chart](https://artifacthub.io/packages/helm/mariadb-operator/mariadb-operator).

          This helm operator provides provides a 1:1 mapping between the official helm chart and the [`MariadbOperator`](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/config/samples/helm_v1alpha1_mariadboperator.yaml) CRD, allowing to install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) via OLM without having to do any change in the helm chart.

          Normally, you would install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) providing this `values.yaml` to the helm chart:
          ```yaml
          image:
            repository: ghcr.io/mariadb-operator/mariadb-operator
            pullPolicy: IfNotPresent
          logLevel: INFO
          ha:
            enabled: true
          metrics:
            enabled: true
            serviceMonitor:
              enabled: true
          webhook:
            cert:
              certManager:
                enabled: true
          ```

          This helm chart installation is abstracted in the [`MariadbOperator`](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/config/samples/helm_v1alpha1_mariadboperator.yaml) CRD, which will be reconciled by the helm operator:
          ```yaml
          apiVersion: helm.k8s.mariadb.com/v1alpha1
          kind: MariadbOperator
          metadata:
            name: mariadb-operator
          spec:
            image:
              repository: ghcr.io/mariadb-operator/mariadb-operator
              pullPolicy: IfNotPresent
            logLevel: INFO
            ha:
              enabled: true
            metrics:
              enabled: true
              serviceMonitor:
                enabled: true
            webhook:
              cert:
                certManager:
                  enabled: true
          ```

          Once you have installed the operator, you will able to install a [`MariaDB`](https://github.com/mariadb-operator/mariadb-operator/blob/main/examples/manifests/mariadb_v1alpha1_mariadb.yaml) instance. Refer to the main [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) documentation for getting started with the rest of CRDs.

          ## Documentation
          * [mariadb-operator](https://github.com/mariadb-operator/mariadb-operator/blob/main/README.md)
          * [mariadb-operator-helm](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/README.md)

          ## Releases
          This operator is automatically published in the following repositories whenever a new version of the [helm chart](https://artifacthub.io/packages/helm/mariadb-operator/mariadb-operator) is released:
          - [k8s-operatorhub/community-operators](https://github.com/k8s-operatorhub/community-operators)
          - [redhat-openshift-ecosystem/community-operators-prod](https://github.com/redhat-openshift-ecosystem/community-operators-prod)

          ## Roadmap
          Take a look at our [roadmap](https://github.com/mariadb-operator/mariadb-operator/blob/main/ROADMAP.md) and feel free to open an issue to suggest new features.

          ## Contributing
          We welcome and encourage contributions to this project! Please check our [contributing](https://github.com/mariadb-operator/mariadb-operator/blob/main/CONTRIBUTING.md) and [development](https://github.com/mariadb-operator/mariadb-operator/blob/main/docs/DEVELOPMENT.md) guides. PRs welcome!

          ## Get in touch
          Join us on Slack: **[MariaDB Community Slack](https://r.mariadb.com/join-community-slack)**.
        displayName: MariaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mariadb
        - mysql
        - operator
        - mariadb-operator
        - database
        - maxscale
        links:
        - name: GitHub
          url: https://github.com/mariadb-operator/mariadb-operator
        maintainers:
        - email: mariadb-operator@proton.me
          name: mmontes11
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: mariadb-operator
          url: https://github.com/mariadb-operator/mariadb-operator
        relatedImages:
        - ghcr.io/mariadb-operator/mariadb-operator-helm:25.8.4
        version: 25.8.4
      entries:
      - name: mariadb-operator.v25.8.4
        version: 25.8.4
      - name: mariadb-operator.v25.8.3
        version: 25.8.3
      - name: mariadb-operator.v25.8.2
        version: 25.8.2
      - name: mariadb-operator.v25.8.1
        version: 25.8.1
      - name: mariadb-operator.v25.8.0
        version: 25.8.0
      - name: mariadb-operator.v0.38.1
        version: 0.38.1
      - name: mariadb-operator.v0.38.0
        version: 0.38.0
      - name: mariadb-operator.v0.37.1
        version: 0.37.1
      - name: mariadb-operator.v0.37.0
        version: 0.37.0
      - name: mariadb-operator.v0.36.0
        version: 0.36.0
      - name: mariadb-operator.v0.35.1
        version: 0.35.1
      - name: mariadb-operator.v0.35.0
        version: 0.35.0
      - name: mariadb-operator.v0.34.0
        version: 0.34.0
      - name: mariadb-operator.v0.33.0
        version: 0.33.0
      - name: mariadb-operator.v0.32.0
        version: 0.32.0
      - name: mariadb-operator.v0.31.0
        version: 0.31.0
      - name: mariadb-operator.v0.30.0
        version: 0.30.0
      - name: mariadb-operator.v0.29.0
        version: 0.29.0
      - name: mariadb-operator.v0.28.1
        version: 0.28.1
      - name: mariadb-operator.v0.27.0
        version: 0.27.0
      - name: mariadb-operator.v0.26.0
        version: 0.26.0
      - name: mariadb-operator.v0.24.0
        version: 0.24.0
      - name: mariadb-operator.v0.23.1
        version: 0.23.1
      - name: mariadb-operator.v0.23.0
        version: 0.23.0
      - name: mariadb-operator.v0.22.0
        version: 0.22.0
      - name: mariadb-operator.v0.21.0
        version: 0.21.0
      - name: mariadb-operator.v0.20.0
        version: 0.20.0
      - name: mariadb-operator.v0.19.0
        version: 0.19.0
      - name: mariadb-operator.v0.18.0
        version: 0.18.0
      - name: mariadb-operator.v0.17.0
        version: 0.17.0
      - name: mariadb-operator.v0.16.0
        version: 0.16.0
      - name: mariadb-operator.v0.15.1
        version: 0.15.1
      - name: mariadb-operator.v0.15.0
        version: 0.15.0
      - name: mariadb-operator.v0.14.0
        version: 0.14.0
      - name: mariadb-operator.v0.13.0
        version: 0.13.0
      - name: mariadb-operator.v0.11.0
        version: 0.11.0
      - name: mariadb-operator.v0.10.0
        version: 0.10.0
      - name: mariadb-operator.v0.9.1
        version: 0.9.1
      - name: mariadb-operator.v0.9.0
        version: 0.9.0
      - name: mariadb-operator.v0.8.0
        version: 0.8.0
      - name: mariadb-operator.v0.7.1
        version: 0.7.1
      - name: mariadb-operator.v0.7.0
        version: 0.7.0
      - name: mariadb-operator.v0.6.7
        version: 0.6.7
      - name: mariadb-operator.v0.6.6
        version: 0.6.6
      - name: mariadb-operator.v0.6.5
        version: 0.6.5
      - name: mariadb-operator.v0.6.4
        version: 0.6.4
      name: alpha
    defaultChannel: alpha
    packageName: mariadb-operator
    provider:
      name: mariadb-operator
      url: https://github.com/mariadb-operator/mariadb-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Anchore Inc.
      provider-url: ""
    name: anchore-engine
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anchore-engine.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.anchore.io/v1alpha1",
                "kind": "AnchoreEngine",
                "metadata": {
                  "name": "anchoreengine-sample"
                },
                "spec": {
                  "anchoreGlobal": {
                    "defaultAdminEmail": "example@email.com"
                  },
                  "postgresql": {
                    "extraEnv": [
                      {
                        "name": "POSTGRESQL_USER",
                        "value": "anchoreengine"
                      },
                      {
                        "name": "POSTGRESQL_PASSWORD",
                        "value": "anchore-postgres,123"
                      },
                      {
                        "name": "POSTGRESQL_DATABASE",
                        "value": "anchore"
                      },
                      {
                        "name": "PGUSER",
                        "value": "postgres"
                      }
                    ],
                    "postgresDatabase": "anchore",
                    "postgresPassword": "anchore-postgres,123",
                    "postgresUser": "anchoreengine"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/anchore/engine-operator@sha256:baf0f5acd319bc4d0132c88922ee3f357fdcdc278d931f243d0cbec3f6d416c6
          createdAt: "2021-10-06T23:53:51Z"
          description: Anchore Engine - container image scanning service for policy-based
            security, best-practice and compliance enforcement.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/anchore/engine-operator
          support: anchore.slack.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys Anchore Engine
            displayName: Anchore Engine Operator
            kind: AnchoreEngine
            name: anchoreengines.charts.anchore.io
            version: v1alpha1
        description: "Anchore Engine is an open source software system that provides
          a centralized service for analyzing container images, \n scanning for security
          vulnerabilities, and enforcing deployment policies. Anchore Engine is provided
          as a Docker \n container image that can be run standalone or with an orchestration
          platform such as Kubernetes. Anchore Engine allows \n users to perform detailed
          analysis on their container images, run queries, produce reports, and define
          policies that \n can be used in CI/CD pipelines. Users can extend Anchore
          Engine with plugins that add new queries, image analysis, and \n policies.
          Anchore Engine can be accessed directly through a RESTful API or via the
          Anchore CLI.\n\n The Anchore Engine Operator is based on the official stable
          [Helm Chart](https://github.com/anchore/anchore-charts/tree/master/stable/anchore-engine)."
        displayName: Anchore Engine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - analysis
        - docker
        - anchore
        - anchore-engine
        - image
        - security
        - vulnerability
        - scanner
        links:
        - name: Anchore Inc.
          url: https://www.anchore.com
        - name: Anchore Engine
          url: https://github.com/anchore/anchore-engine
        maintainers:
        - email: bradyt@anchore.com
          name: Brady Todhunter
        - email: zach@anchore.com
          name: Zach Hill
        maturity: alpha
        provider:
          name: Anchore Inc.
        relatedImages:
        - registry.connect.redhat.com/anchore/engine-operator@sha256:baf0f5acd319bc4d0132c88922ee3f357fdcdc278d931f243d0cbec3f6d416c6
        - registry.connect.redhat.com/anchore/engine0@sha256:68e09d5ff695b46f24de3d7610f88dfc14b77b38758bbd68d03350dbae3deb0f
        - registry.redhat.io/rhel8/postgresql-96@sha256:e8177c5de05ccdd6d12e360b65c63c889b2de725d573ffdedd8914230b118639
        version: 1.0.1
      entries:
      - name: anchore-engine.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: anchore-engine
    provider:
      name: Anchore Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: https://github.com/Mellanox/maintenance-operator/
    name: nvidia-maintenance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nvidia-maintenance-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "MaintenanceOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "maintenanceoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "maintenanceoperatorconfig",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "maintenanceoperatorconfig-sample"
                },
                "spec": {
                  "logLevel": "info",
                  "maxNodeMaintenanceTimeSeconds": 1600,
                  "maxParallelOperations": 1,
                  "maxUnavailable": "30%"
                }
              },
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "nodemaintenance-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodemaintenance",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "cordon": true,
                  "drainSpec": {
                    "deleteEmptyDir": true,
                    "force": true,
                    "podEvictionFilters": [
                      {
                        "byResourceNameRegex": "nvidia.com/gpu-*"
                      },
                      {
                        "byResourceNameRegex": "nvidia.com/rdma*"
                      }
                    ],
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  },
                  "nodeName": "worker-node-01",
                  "requestorID": "some-cluster-component.vendor.com",
                  "waitForPodCompletion": {
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
          createdAt: "2025-07-13T12:33:06Z"
          description: Node maintenance in K8s cluster in a coordinated manner
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-maintenance-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: NVIDIA
          repository: https://github.com/Mellanox/maintenance-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MaintenanceOperatorConfig is the Schema for the maintenanceoperatorconfigs
              API
            displayName: Maintenance Operator Config
            kind: MaintenanceOperatorConfig
            name: maintenanceoperatorconfigs.maintenance.nvidia.com
            version: v1alpha1
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.maintenance.nvidia.com
            version: v1alpha1
        description: Coordinates And Performs Common Node Preparations For Maintenance
        displayName: NVIDIA Maintenance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node
        - maintenance
        - drain
        links:
        - name: NVIDIA Maintenance Operator
          url: https://github.com/Mellanox/maintenance-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/maintenance-operator/
        relatedImages:
        - ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
        version: 0.2.3
      entries:
      - name: nvidia-maintenance-operator.v0.2.3
        version: 0.2.3
      name: stable
    - currentCSV: nvidia-maintenance-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "MaintenanceOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "maintenanceoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "maintenanceoperatorconfig",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "maintenanceoperatorconfig-sample"
                },
                "spec": {
                  "logLevel": "info",
                  "maxNodeMaintenanceTimeSeconds": 1600,
                  "maxParallelOperations": 1,
                  "maxUnavailable": "30%"
                }
              },
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "nodemaintenance-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodemaintenance",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "cordon": true,
                  "drainSpec": {
                    "deleteEmptyDir": true,
                    "force": true,
                    "podEvictionFilters": [
                      {
                        "byResourceNameRegex": "nvidia.com/gpu-*"
                      },
                      {
                        "byResourceNameRegex": "nvidia.com/rdma*"
                      }
                    ],
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  },
                  "nodeName": "worker-node-01",
                  "requestorID": "some-cluster-component.vendor.com",
                  "waitForPodCompletion": {
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
          createdAt: "2025-07-13T12:33:06Z"
          description: Node maintenance in K8s cluster in a coordinated manner
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-maintenance-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: NVIDIA
          repository: https://github.com/Mellanox/maintenance-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MaintenanceOperatorConfig is the Schema for the maintenanceoperatorconfigs
              API
            displayName: Maintenance Operator Config
            kind: MaintenanceOperatorConfig
            name: maintenanceoperatorconfigs.maintenance.nvidia.com
            version: v1alpha1
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.maintenance.nvidia.com
            version: v1alpha1
        description: Coordinates And Performs Common Node Preparations For Maintenance
        displayName: NVIDIA Maintenance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node
        - maintenance
        - drain
        links:
        - name: NVIDIA Maintenance Operator
          url: https://github.com/Mellanox/maintenance-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/maintenance-operator/
        relatedImages:
        - ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
        version: 0.2.3
      entries:
      - name: nvidia-maintenance-operator.v0.2.3
        version: 0.2.3
      name: v0.2
    defaultChannel: v0.2
    packageName: nvidia-maintenance-operator
    provider:
      name: NVIDIA
      url: https://github.com/Mellanox/maintenance-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: mariadb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mariadb.openstack.org/v1beta1",
                "kind": "Galera",
                "metadata": {
                  "name": "openstack"
                },
                "spec": {
                  "replicas": 3,
                  "secret": "osp-secret",
                  "storageClass": "local-storage",
                  "storageRequest": "500M"
                }
              },
              {
                "apiVersion": "mariadb.openstack.org/v1beta1",
                "kind": "MariaDBAccount",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "mariadb-operator",
                    "app.kubernetes.io/instance": "mariadbaccount-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mariadbaccount",
                    "app.kubernetes.io/part-of": "mariadb-operator",
                    "mariaDBDatabaseName": "neutron"
                  },
                  "name": "neutron1"
                },
                "spec": {
                  "secret": "neutrondb-secret",
                  "userName": "neutron1"
                }
              },
              {
                "apiVersion": "mariadb.openstack.org/v1beta1",
                "kind": "MariaDBDatabase",
                "metadata": {
                  "labels": {
                    "dbName": "mariadb"
                  },
                  "name": "scott"
                },
                "spec": {
                  "name": "scott",
                  "secret": "tiger"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:39Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Galera is the Schema for the galeras API
            displayName: Galera
            kind: Galera
            name: galeras.mariadb.openstack.org
            version: v1beta1
          - description: MariaDBAccount is the Schema for the mariadbaccounts API
            displayName: Maria DBAccount
            kind: MariaDBAccount
            name: mariadbaccounts.mariadb.openstack.org
            version: v1beta1
          - description: MariaDBDatabase is the Schema for the mariadbdatabases API
            displayName: Maria DBDatabase
            kind: MariaDBDatabase
            name: mariadbdatabases.mariadb.openstack.org
            version: v1beta1
        description: MariaDB Operator
        displayName: MariaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Identity
        - MariaDB
        links:
        - name: MariaDB Operator
          url: https://github.com/openstack-k8s-operators/mariadb-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/mariadb-rhel9-operator@sha256:5ae64a3cf9bf5eaf244548c3bf2e8ef89c9436cb676721c1f1d1c2ff8285e7e5
        - registry.redhat.io/rhoso/openstack-mariadb-rhel9@sha256:13119fe0ec56263a2bff3fc9c3892ea6386b837955280d8488f05d8ea6c4e44d
        version: 1.0.6
      entries:
      - name: mariadb-operator.v1.0.6
        version: 1.0.6
      - name: mariadb-operator.v1.0.5
        version: 1.0.5
      - name: mariadb-operator.v1.0.4
        version: 1.0.4
      - name: mariadb-operator.v1.0.3
        version: 1.0.3
      - name: mariadb-operator.v1.0.2
        version: 1.0.2
      - name: mariadb-operator.v1.0.1
        version: 1.0.1
      - name: mariadb-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: mariadb-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: GitLab, Inc.
      provider-url: https://gitlab.com
    name: gitlab-runner-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-runner-operator.v1.41.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta2",
                "kind": "Runner",
                "metadata": {
                  "name": "gitlab-runner-example",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "gitlabUrl": "https://gitlab.example.com",
                  "token": "runner-token-secret",
                  "tags": "kubernetes,example",
                  "concurrent": 10,
                  "imagePullPolicy": "IfNotPresent"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery, Developer Tools
          containerImage: registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
          createdAt: "2025-09-25T15:08:40Z"
          description: GitLab Runner operator manages lifecycle of GitLab Runner instances
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.11.0 <1.41.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: none
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
          support: GitLab, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Runner is the open source project used to run your jobs and
              send the results back to GitLab
            displayName: GitLab Runner
            kind: Runner
            name: runners.apps.gitlab.com
            version: v1beta2
        description: |-
          GitLab Runner is the lightweight, highly-scalable agent that runs your build jobs and sends the results back to a GitLab instance. GitLab Runner works in conjunction with GitLab CI/CD, the open-source continuous integration service included with GitLab.

              The GitLab Runner operator manages the lifecycle of GitLab Runner in Kubernetes or Openshift clusters. The operator aims to automate the tasks needed to run your CI/CD jobs in your container orchestration platform.

              ## Prerequisites

                For Kubernetes cluster, install cert-manager:

                ```shell
                kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.7.1/cert-manager.yaml
                ```

              ## GitLab Runner version

              This version of **GitLab Runner Operator** ships with **GitLab Runner v18.4.0**.

              To use a different version of **GitLab Runner** change the [`runnerImage` and `helperImage` properties](https://docs.gitlab.com/runner/configuration/configuring_runner_operator.html#operator-properties).

              ## Usage

               To link a GitLab Runner instance to a self-hosted GitLab instance or the hosted [GitLab](https://gitlab.com), you first need to:

               1. Create a secret containing the `runner-token` from your GitLab project.

                ```
                cat > gitlab-runner-secret.yml << EOF
                apiVersion: v1
                kind: Secret
                metadata:
                  name: gitlab-runner-secret
                type: Opaque
                # Only one of the following fields must be set. The Operator will fail to register Runner if both are provided.
                # # NOTE: runner-registration-token is deprecated and will be removed in GitLab 18.0. You should use runner-token instead.
                stringData:
                  runner-token: REPLACE_ME # your project runner token
                  # runner-registration-token: "" # your project runner secret
                EOF
                ```

                ```
                oc apply -f gitlab-runner-secret.yml
                ```

               2. Create the Custom Resource Definition (CRD) file and include the following information. The tags value must be openshift for the job to run.

                 ```
                 cat > gitlab-runner.yml << EOF
                 apiVersion: apps.gitlab.com/v1beta2
                 kind: Runner
                 metadata:
                   name: gitlab-runner
                 spec:
                   gitlabUrl: https://gitlab.example.com
                   buildImage: alpine
                   token: gitlab-runner-secret
                   tags: openshift
                 EOF
                 ```

                ```
                oc apply -f gitlab-runner.yml
                ```

              ## Full documentation

              Visit [Install GitLab Runner Operator](https://docs.gitlab.com/runner/install/operator.html)
        displayName: GitLab Runner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GitLab
        - CI/CD
        - DevOps
        - SAST
        - DAST
        links:
        - name: Gitlab Runner Operator
          url: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
        - name: GitLab Docs
          url: https://docs.gitlab.com
        - name: GitLab Runner
          url: https://docs.gitlab.com/runner/
        - name: GitLab CI/CD
          url: https://docs.gitlab.com/ee/ci/quick_start/
        maintainers:
        - email: support@gitlab.com
          name: Georgi N. Georgiev
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: GitLab, Inc.
          url: https://gitlab.com
        relatedImages:
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-ocp@sha256:0f2a5a3b6779d719eecbf079b99cf06b652603d74b563f57d753e32966d253dd
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-helper-ocp@sha256:29727fae83a226bcf57607240d4046a0b8c208b5ff42ee1f8ca0f115bb188809
        version: 1.41.1
      entries:
      - name: gitlab-runner-operator.v1.41.1
        version: 1.41.1
      - name: gitlab-runner-operator.v1.41.0
        version: 1.41.0
      - name: gitlab-runner-operator.v1.40.1
        version: 1.40.1
      - name: gitlab-runner-operator.v1.40.0
        version: 1.40.0
      - name: gitlab-runner-operator.v1.39.0
        version: 1.39.0
      name: stable
    defaultChannel: stable
    packageName: gitlab-runner-operator
    provider:
      name: GitLab, Inc.
      url: https://gitlab.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "docker.io/percona/percona-postgresql-operator:2.7.0-ppg17.5.2-postgres",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "docker.io/percona/percona-pgbouncer:1.24.1",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "docker.io/percona/percona-pgbackrest:2.55.0",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:3.2.0",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: docker.io/percona/percona-postgresql-operator:2.7.0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          olm.skipRange: <v2.7.0
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native

          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - docker.io/percona/percona-postgresql-operator:2.7.0
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator.v2.6.1
        version: 2.6.1
      - name: percona-postgresql-operator.v2.6.0
        version: 2.6.0
      - name: percona-postgresql-operator.v2.5.0
        version: 2.5.0
      name: stable
    defaultChannel: stable
    packageName: percona-postgresql-operator
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: VMware Tanzu
      provider-url: ""
    name: rabbitmq-messaging-topology-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-messaging-topology-operator.v1.17.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [{"apiVersion":"rabbitmq.com/v1beta1","kind":"Binding","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Exchange","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Federation","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Permission","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Policy","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"OperatorPolicy","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Queue","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"SchemaReplication","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Shovel","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1alpha1","kind":"SuperStream","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"User","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Vhost","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"TopicPermission","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}}]
          capabilities: Full Lifecycle
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/rabbitmqoperator/messaging-topology-operator:1.17.4
          createdAt: 2025-08-21T11:15:42
          description: Kubernetes operator to allow developers to create and manage
            RabbitMQ messaging topologies within a RabbitMQ cluster using a declarative
            Kubernetes API.
          repository: https://github.com/rabbitmq/messaging-topology-operator
          support: VMware Tanzu
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Binding
            displayName: Binding
            kind: Binding
            name: bindings.rabbitmq.com
            version: v1beta1
          - description: Exchange
            displayName: Exchange
            kind: Exchange
            name: exchanges.rabbitmq.com
            version: v1beta1
          - description: Federation
            displayName: Federation
            kind: Federation
            name: federations.rabbitmq.com
            version: v1beta1
          - description: Permission
            displayName: Permission
            kind: Permission
            name: permissions.rabbitmq.com
            version: v1beta1
          - description: TopicPermission
            displayName: TopicPermission
            kind: TopicPermission
            name: topicpermissions.rabbitmq.com
            version: v1beta1
          - description: Policy
            displayName: Policy
            kind: Policy
            name: policies.rabbitmq.com
            version: v1beta1
          - description: OperatorPolicy
            displayName: OperatorPolicy
            kind: OperatorPolicy
            name: operatorpolicies.rabbitmq.com
            version: v1beta1
          - description: Queue
            displayName: Queue
            kind: Queue
            name: queues.rabbitmq.com
            version: v1beta1
          - description: Schema Replication
            displayName: SchemaReplication
            kind: SchemaReplication
            name: schemareplications.rabbitmq.com
            version: v1beta1
          - description: Shovel
            displayName: Shovel
            kind: Shovel
            name: shovels.rabbitmq.com
            version: v1beta1
          - description: Super Stream
            displayName: SuperStream
            kind: SuperStream
            name: superstreams.rabbitmq.com
            version: v1alpha1
          - description: User
            displayName: User
            kind: User
            name: users.rabbitmq.com
            version: v1beta1
          - description: Vhost
            displayName: Vhost
            kind: Vhost
            name: vhosts.rabbitmq.com
            version: v1beta1
          required:
          - description: Rabbitmq Cluster
            displayName: RabbitmqCluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: |
          ## About the managed application
          RabbitMQ is an open-source message-broker software that originally implemented the Advanced Message Queuing Protocol and has since been extended with a plug-in architecture to support Streaming Text Oriented Messaging Protocol, MQ Telemetry Transport, and other protocols.
          ## About this Operator
          Kubernetes operator to allow developers to create and manage RabbitMQ messaging topologies within a RabbitMQ cluster using a declarative Kubernetes API. A Messaging topology is the collection of objects such as exchanges, queues, bindings and policies that provides specific messaging or streaming scenario. This operator is used with RabbitMQ clusters deployed via the RabbitMQ Cluster Kubernetes Operator. This repository contains custom controllers and custom resource definitions (CRDs) enabling a declarative API for RabbitMQ messaging topologies.
          ## Prerequisites for enabling this Operator
          The operator works on top of the RabbitMQ cluster operator.
          Messaging Topology Operator is tested with the latest release of RabbitMQ Cluster Operator. It uses the generated default user secret from RabbitmqCluster (set in rabbitmqcluster.status.binding) to authenticate with RabbitMQ server.
        displayName: rabbitmq-messaging-topology-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        - messaging
        - amqp
        - amqp10
        - stomp
        - mqtt
        - mqtt5
        - queues
        - streaming
        links:
        - name: GitHub
          url: https://github.com/rabbitmq/messaging-topology-operator
        - name: Documentation
          url: https://www.rabbitmq.com/kubernetes/operator/install-topology-operator.html
        maintainers:
        - email: rabbitmq-users@googlegroups.com
          name: VMware Tanzu
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        provider:
          name: VMware Tanzu
        relatedImages:
        - quay.io/rabbitmqoperator/messaging-topology-operator:1.17.4
        version: 1.17.4
      entries:
      - name: rabbitmq-messaging-topology-operator.v1.17.4
        version: 1.17.4
      - name: rabbitmq-messaging-topology-operator.v1.16.0
        version: 1.16.0
      - name: rabbitmq-messaging-topology-operator.v1.15.0
        version: 1.15.0
      - name: rabbitmq-messaging-topology-operator.v1.14.2
        version: 1.14.2
      - name: rabbitmq-messaging-topology-operator.v1.14.1
        version: 1.14.1
      - name: rabbitmq-messaging-topology-operator.v1.13.0
        version: 1.13.0
      - name: rabbitmq-messaging-topology-operator.v1.12.2
        version: 1.12.2
      - name: rabbitmq-messaging-topology-operator.v1.12.1
        version: 1.12.1
      name: stable
    defaultChannel: stable
    packageName: rabbitmq-messaging-topology-operator
    provider:
      name: VMware Tanzu
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Omar Gaye IBM
      provider-url: omarg.net
    name: memcached-operator-ogaye
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: memcached-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.omarg.net/v1alpha1",
                "kind": "Memcached",
                "metadata": {
                  "name": "memcached-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.6.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Memcached is the Schema for the memcacheds API
            displayName: Memcached
            kind: Memcached
            name: memcacheds.cache.omarg.net
            version: v1alpha1
        description: Omar memcached operator
        displayName: memcached-operator-ogaye
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Omar
        - Memcached
        links:
        - name: Memcached Operator OGaye
          url: https://memcached-operator.domain
        maturity: alpha
        provider:
          name: Omar Gaye IBM
          url: omarg.net
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9211b70edbb93d718428e1c9a992e5fc74a20cb213f8a0592ffe340c362890fe
        - quay.io/ogaye/memcached-operator@sha256:5340ac11ce676dd3204bf3807a4c2333a8dddcbcec918db0aeafbc977163c1c6
        version: 0.0.1
      entries:
      - name: memcached-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: memcached-operator-ogaye
    provider:
      name: Omar Gaye IBM
      url: omarg.net
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Aqua Security, Inc.
      provider-url: ""
    name: aqua-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aqua-operator.2022.4.836
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaCsp",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "namespace": "aqua",
                    "version": "2022.4",
                    "requirements": true
                  },
                  "common": {
                    "dbDiskSize": 10,
                    "databaseSecret": {
                      "key": "db-password",
                      "name": "aqua-database-password"
                    }
                  },
                  "database": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "gateway": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "server": {
                    "replicas": 1,
                    "service": "LoadBalancer"
                  },
                  "route": true,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaDatabase",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "splitDB": false
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "diskSize": 10,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "runAsNonRoot": false,
                  "gateway": {
                    "host": "aqua-gateway",
                    "port": 8443
                  },
                  "token": "<<your-token>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaKubeEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "config": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "aqua-secure",
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "service": "ClusterIP"
                  },
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaGateway",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  }
                },
                "runAsNonRoot": false
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaScanner",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "deploy": {
                    "replicas": 1
                  },
                  "runAsNonRoot": false,
                  "login": {
                    "username": "<<YOUR AQUA USER NAME>>",
                    "password": "<<YOUR AQUA USER PASSWORD>>",
                    "host": "http://aqua-server:8080"
                  }
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaServer",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "LoadBalancer"
                  }
                },
                "runAsNonRoot": false
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaLightning",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "global": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "Default-cluster-name"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "kubeEnforcer": {
                    "infra": {
                      "serviceAccount": "aqua-kube-enforcer-sa"
                    },
                    "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                    "deploy": {
                      "service": "ClusterIP"
                    },
                    "starboard": {
                      "infra": {
                        "serviceAccount": "aqua-kube-enforcer-sa"
                      },
                      "config": {
                        "imagePullSecret": "aqua-registry"
                      },
                      "deploy": {
                        "replicas": 1
                      }
                    },
                    "env": []
                  },
                  "enforcer": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "deploy": null,
                    "token": "<<ENFORCER_GROUP_TOKEN>>",
                    "env": [
                    ],
                    "secret": {
                      "name": null,
                      "key": null
                    },
                    "runAsNonRoot": null,
                    "rhcosVersion": null
                  }
                }
              },
            {
                  "apiVersion": "operator.aquasec.com/v1alpha1",
                  "kind": "AquaCloudConnector",
                  "metadata": {
                    "name": "aqua"
                  },
                  "spec": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "common": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    },
                    "runAsNonRoot": false,
                    "gateway": {
                      "host": "aqua-gateway",
                      "port": "8443"
                    },
                    "login": {
                      "host": "aqua-gateway:8443",
                      "password": "<<YOUR AQUA USER PASSWORD>>",
                      "token": "<<YOUR AQUA CLOUD CONNECTOR TOKEN>>",
                      "username": "<<YOUR AQUA USER NAME>>"
                    },
                    "tunnels": [
                      {
                        "host": "<account>.azurecr.io",
                        "port": "443"
                      },
                      {
                        "host": "gke.gcr.io",
                        "port": "443"
                      },
                      {
                        "host": "<account>.jfrog.io",
                        "port": "443"
                      },
                      {
                        "host": "1.1.1.1",
                        "port": "443"
                      },
                      {
                        "region": "us-east1",
                        "type": "ecr"
                      }
                    ]
                  }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/aquasec/aquasec@sha256:c9f2c770a51e9664006c12d124ca18aab76bea8c32a6a253fe31d44c02633642
          createdAt: ""
          description: The Aqua Security Operator runs within a Openshift cluster
            and provides a means to deploy and manage Aqua Security cluster and components.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2022.4.836'
          operators.operatorframework.io/internal-objects: '["aquastarboards.aquasecurity.github.io","configauditreports.aquasecurity.github.io",
            "clusterconfigauditreports.aquasecurity.github.io"]'
          repository: https://github.com/aquasecurity/aqua-operator
          support: Aqua Security, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aqua Security CSP Deployment with Aqua Operator
            displayName: AquaCsp
            kind: AquaCsp
            name: aquacsps.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Lightning Enforcers Deployment with Aqua Operator
            displayName: AquaLightningEnforcers
            kind: AquaLightning
            name: aqualightnings.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Database Deployment with Aqua Operator
            displayName: AquaDatabase
            kind: AquaDatabase
            name: aquadatabases.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Enforcer Deployment with Aqua Operator
            displayName: AquaEnforcer
            kind: AquaEnforcer
            name: aquaenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security KubeEnforcer Deployment with Aqua Operator
            displayName: AquaKubeEnforcer
            kind: AquaKubeEnforcer
            name: aquakubeenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Gateway Deployment with Aqua Operator
            displayName: AquaGateway
            kind: AquaGateway
            name: aquagateways.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Scanner Deployment with Aqua Operator
            displayName: AquaScanner
            kind: AquaScanner
            name: aquascanners.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Server Deployment with Aqua Operator
            displayName: AquaServer
            kind: AquaServer
            name: aquaservers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Starboard for kube-enforcer.
            displayName: AquaStarboard
            kind: AquaStarboard
            name: aquastarboards.aquasecurity.github.io
            version: v1alpha1
          - description: Aqua Cloud Connector Deployment with Aqua Operator
            displayName: AquaCloudConnector
            kind: AquaCloudConnector
            name: aquacloudconnectors.operator.aquasec.com
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ClusterConfigAuditReport
            kind: ClusterConfigAuditReport
            name: clusterconfigauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ConfigAuditReport
            kind: ConfigAuditReport
            name: configauditreports.aquasecurity.github.io
            version: v1alpha1
        description: |-
          The Aqua Security Operator runs within an OpenShift cluster and provides a means to deploy and manage the Aqua Security cluster and components :
          * Server (aka “console”)
          * Database (for production environments we recommend to use an external database and not the Aqua default database)
          * Gateway
          * Enforcer (aka “agent”)
          * KubeEnforcer
          * Scanner
          * CSP (package that contains the Server, Database, and Gateway)
          * LightningEnforcer (package that contains Enforcer and Kube Enforcer) - Do not deploy using Certified Operator.
          * AquaCloudConnector.

          Use the Aqua-Operator to
          * Deploy Aqua Security components on OpenShift
          * Scale up Aqua Security components with extra replicas
          * Assign metadata tags to Aqua Security components

          The Aqua Operator provides a few [Custom Resources](https://github.com/aquasecurity/aqua-operator/tree/master/config/crd) for managing the Aqua CSP platform.

          ## Prerequisites
          There are only 4 prerequisites:
          1. Verify that the OpenShift version is officially supported [Verify Here](https://docs.aquasec.com/saas/workload-protection/product-information/system-requirements/)
          2. Make sure you have a license and to obtain one, please contact Aqua Security at [Contact Us](mailto:cloudsales@aquasec.com).
          3. Create a new project for aqua

              ```oc new-project aqua```

          4. Create the secret for Aqua Database password. If you are using an External Database, make sure you use those credentials.
          NOTE: This step is optional and you can specify the Database password when creating the CRs.

              ```oc create secret generic aqua-database-password --from-literal=db-password=<password> -n aqua```

          Please note that for the certified operator, the Docker registry secret is NOT needed. For the Red Hat certified operator, the Aqua application images are stored in the Red Hat Connect registry and will be automatically pulled by the Aqua operator.

          ## Choosing the right channel

          Channels are a flexible publishing mechanism that offers you different Aqua versions to deploy, at any point. These channels map to their respective Aqua versions, describing the maturity based on where they are in their lifecycle:
          * 4.6: This maps to the older version of Aqua 4.6.0
          * 5.0: This maps to the Aqua version 5.0.0
          * 5.3: This maps to the Aqua version 5.3.0
          * 6.0: This maps to the Aqua version 6.0.0
          * 6.2: This maps to the Aqua version 6.2.0
          * 6.5: This maps to the Aqua version 6.5.0
          * 2022.4: This is the latest and greatest version of Aqua (default)

          ## Deploying the Aqua Operator

          Aqua Operator follows an easy push button deployment that installs the operator in its own aqua namespace.
          Please refer to the instructions [here](https://github.com/aquasecurity/aqua-operator/blob/2022.4.0/docs/DeployOpenShiftOperator.md).

          ## Configuring the Aqua Operator

          Operators leverage CRDs as a configuration mechanism for the application. Aqua offers a number of CRDs that can be tweaked and configured as per your needs. Please refer to this [link](https://github.com/aquasecurity/aqua-operator/blob/6.2.0/docs/DeployOpenShiftOperator.md#aquacsp-crds) for Custom Resource examples for popular scenarios.

          ## Support

          For support please contact support@aquasec.com.
        displayName: Aqua Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aqua-security
        - scanning
        - security
        - runtime-security
        links:
        - name: Aqua Security
          url: https://www.aquasec.com/
        - name: Aqua Operator Github
          url: https://github.com/aquasecurity/aqua-operator
        maintainers:
        - email: support@aquasec.com
          name: Aqua Support, Aqua Security
        maturity: alpha
        provider:
          name: Aqua Security, Inc.
        relatedImages:
        - registry.connect.redhat.com/aquasec/console@sha256:ed371053e46e89ad68ed48282ddf28abc668a613a9c32e3aa45e5111512b8466
        - registry.connect.redhat.com/aquasec/kube-bench@sha256:bbb4e8a61033b303525af6764f685e62a2f441453c3f0eaaf4cbab6997bf93ee
        - registry.connect.redhat.com/aquasec/gateway@sha256:11aa01e681986f6b5bd53b41a07db4456023080a753970d8c872e000f68f3b9b
        - registry.connect.redhat.com/aquasec/kube-enforcer@sha256:c456b65a3f1dc2de7209f3d9cbcf1bb523130021433b8fd740c41d9bd6f7db9f
        - registry.connect.redhat.com/aquasec/aquasec@sha256:c9f2c770a51e9664006c12d124ca18aab76bea8c32a6a253fe31d44c02633642
        - registry.connect.redhat.com/aquasec/enforcer@sha256:50128b750e0a094de266c9a2cb42ce06a757f441a68d206a8a9c17f7b7119066
        - registry.connect.redhat.com/aquasec/scanner@sha256:c086d735c7e6ea842ad1aaeb1eb19d3a951e6e88d397a20f2cd559c6bc042d26
        - registry.connect.redhat.com/aquasec/cloud-connector@sha256:f9da1936ac5e4b8e4537dca69f920c682fb9fac98d091680e01d2122f2b0f85f
        - registry.connect.redhat.com/aquasec/database@sha256:c91ae82b0179c3ec4e156835a9f23b66767ba15514ff33fb268b35d8f04d6fde
        - registry.connect.redhat.com/aquasec/starboard@sha256:61673f7ba1d983ff4bb4bd76036ac55e5b81696c9a3e3aa4ab1a49a27a7a2106
        version: 2022.4.836
      entries:
      - name: aqua-operator.2022.4.836
        version: 2022.4.836
      - name: aqua-operator.2022.4.829
        version: 2022.4.829
      - name: aqua-operator.2022.4.818
        version: 2022.4.818
      - name: aqua-operator.2022.4.805
        version: 2022.4.805
      - name: aqua-operator.2022.4.760
        version: 2022.4.760
      - name: aqua-operator.2022.4.759
        version: 2022.4.759
      - name: aqua-operator.2022.4.720
        version: 2022.4.720
      - name: aqua-operator.2022.4.683
        version: 2022.4.683
      - name: aqua-operator.2022.4.662
        version: 2022.4.662
      - name: aqua-operator.2022.4.588
        version: 2022.4.588
      - name: aqua-operator.2022.4.587
        version: 2022.4.587
      - name: aqua-operator.2022.4.519
        version: 2022.4.519
      - name: aqua-operator.2022.4.518
        version: 2022.4.518
      name: "2022.4"
    defaultChannel: "2022.4"
    packageName: aqua-operator-certified
    provider:
      name: Aqua Security, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: netobserv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: network-observability-operator.v1.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowMetric",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netobserv-operator",
                    "app.kubernetes.io/instance": "flowmetric-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "flowmetric",
                    "app.kubernetes.io/part-of": "netobserv-operator"
                  },
                  "name": "flowmetric-sample"
                },
                "spec": {
                  "charts": [
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "",
                          "promQL": "sum(rate($METRIC[2m]))"
                        }
                      ],
                      "title": "External ingress traffic",
                      "type": "SingleStat",
                      "unit": "Bps"
                    },
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "{{DstK8S_Namespace}} / {{DstK8S_OwnerName}}",
                          "promQL": "sum(rate($METRIC{DstK8S_Namespace!=\"\"}[2m])) by (DstK8S_Namespace, DstK8S_OwnerName)"
                        }
                      ],
                      "sectionName": "External",
                      "title": "Top external ingress traffic per workload",
                      "type": "StackArea",
                      "unit": "Bps"
                    }
                  ],
                  "direction": "Ingress",
                  "filters": [
                    {
                      "field": "SrcSubnetLabel",
                      "matchType": "Absence"
                    }
                  ],
                  "labels": [
                    "DstK8S_HostName",
                    "DstK8S_Namespace",
                    "DstK8S_OwnerName",
                    "DstK8S_OwnerType"
                  ],
                  "metricName": "cluster_external_ingress_bytes_total",
                  "type": "Counter",
                  "valueField": "Bytes"
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "app"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "infra"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "enable": true,
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "includeList": [
                        "node_ingress_bytes_total",
                        "workload_ingress_bytes_total",
                        "namespace_flows_total"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta2",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "metrics": {
                        "server": {
                          "port": 9400
                        }
                      },
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "eBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "Disabled"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "\"app\""
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "\"infra\""
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "\"Pod\"",
                          "src_kind": "\"Pod\""
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "\"Service\""
                        },
                        "name": "Services network"
                      }
                    ],
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    }
                  },
                  "deploymentModel": "Direct",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "enable": true,
                    "lokiStack": {
                      "name": "loki"
                    },
                    "mode": "Monolithic",
                    "monolithic": {
                      "tenantID": "netobserv",
                      "tls": {
                        "caCert": {
                          "certFile": "service-ca.crt",
                          "name": "loki-gateway-ca-bundle",
                          "type": "configmap"
                        },
                        "enable": false
                      },
                      "url": "http://loki.netobserv.svc:3100/"
                    },
                    "readTimeout": "30s",
                    "writeBatchSize": 10485760,
                    "writeBatchWait": "1s",
                    "writeTimeout": "10s"
                  },
                  "namespace": "netobserv",
                  "networkPolicy": {
                    "additionalNamespaces": [],
                    "enable": false
                  },
                  "processor": {
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "Flows",
                    "metrics": {
                      "disableAlerts": [],
                      "server": {
                        "port": 9401
                      }
                    },
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "prometheus": {
                    "querier": {
                      "enable": true,
                      "mode": "Auto",
                      "timeout": "30s"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking, Observability, OpenShift Optional
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:13b6dd0ddf4bdd2836b0d3ec9c79119a902456dcee100cfac6898768e1f22aab
          createdAt: 2025-09-08T21:48:09
          description: Network flows collector and monitoring solution
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: '{"apiVersion":"flows.netobserv.io/v1beta2",
            "kind":"FlowCollector","metadata":{"name":"cluster"},"spec": {}}'
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
          support: NetObserv team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta2
          - description: '`FlowMetric` is the schema for the custom metrics API, which
              allows to generate more metrics out of flow logs. You can find examples
              here: https://github.com/netobserv/network-observability-operator/tree/main/config/samples/flowmetrics'
            displayName: Flow Metric
            kind: FlowMetric
            name: flowmetrics.flows.netobserv.io
            version: v1alpha1
        description: |
          Network Observability is an OpenShift operator that deploys a monitoring pipeline consisting in:
          - an eBPF agent, that generates network flows from captured packets
          - flowlogs-pipeline, a component that collects, enriches and exports these flows
          - a Console plugin for flows visualization with powerful filtering options, a topology representation and more

          Flow data is then available in multiple ways, each optional:

          - As Cluster Monitoring metrics
          - As raw flow logs stored in Grafana Loki
          - As raw flow logs exported to a collector

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, can optionally be used as the backend to store all collected flows. The Network Observability operator does not install Loki directly, however we provide some guidance to help you there.

          - For a production or production-like environment usage, refer to [the operator documentation](https://docs.openshift.com/container-platform/latest/network_observability/installing-operators.html).

          - For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 10GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          oc create namespace netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/1-storage.yaml) -n netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          If you prefer to not use Loki, you must set `spec.loki.enable` to `false` in `FlowCollector`.
          In that case, you can still get the Prometheus metrics or export raw flows to a custom collector. But be aware that some of the Console plugin features will be disabled. For instance, you will not be able to view raw flows there, and the metrics / topology will have a more limited level of details, missing information such as pods or IPs.

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for example [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/config/samples/flows_v1beta2_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          oc edit flowcollector cluster
          ```

          As it operates cluster-wide on every node, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you might have to configure differently if you used another installation method. Make sure to disable it (`spec.loki.enable`) if you don't want to use Loki.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`) an optional list of exporters to which to send enriched flows. KAFKA and IPFIX exporters are supported. This allows you to define any custom storage or processing that can read from Kafka or use the IPFIX standard.

          - To enable availability zones awareness, set `spec.processor.addZone` to `true`.

          ## Resource considerations

          The following table outlines examples of resource considerations for clusters with certain workload sizes.
          The examples outlined in the table demonstrate scenarios that are tailored to specific workloads. Consider each example only as a baseline from which adjustments can be made to accommodate your workload needs.


          | Resource recommendations                        | Extra small (10 nodes) | Small (25 nodes)       | Medium (65 nodes) **    | Large (120 nodes) **          |
          | ----------------------------------------------- | ---------------------- | ---------------------- | ----------------------- | ----------------------------- |
          | *Worker Node vCPU and memory*                   | 4 vCPUs\| 16GiB mem *  | 16 vCPUs\| 64GiB mem * | 16 vCPUs\| 64GiB mem  * |16 vCPUs\| 64GiB Mem *         |
          | *LokiStack size*                                | `1x.extra-small`       | `1x.small`             | `1x.small`              | `1x.medium`                   |
          | *Network Observability controller memory limit* | 400Mi (default)        | 400Mi (default)        | 400Mi (default)         | 800Mi                         |
          | *eBPF sampling rate*                            | 50 (default)           | 50 (default)           | 50 (default)            | 50 (default)                  |
          | *eBPF memory limit*                             | 800Mi (default)        | 800Mi (default)        | 2000Mi                  | 800Mi (default)               |
          | *FLP memory limit*                              | 800Mi (default)        | 800Mi (default)        | 800Mi (default)         | 800Mi (default)               |
          | *FLP Kafka partitions*                          | N/A                    | 48                     | 48                      | 48                            |
          | *Kafka consumer replicas*                       | N/A                    | 24                     | 24                      | 24                            |
          | *Kafka brokers*                                 | N/A                    | 3 (default)            | 3 (default)             | 3 (default)                   |

          *. Tested with AWS M6i instances.
          **. In addition to this worker and its controller, 3 infra nodes (size `M6i.12xlarge`) and 1 workload node (size `M6i.8xlarge`) were tested.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - More information on [configuring metrics](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/Metrics.md).
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: Network Observability
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/network-observability/network-observability-console-plugin-compat-rhel9@sha256:d130f94cdd3cc863a31c59f395a6c4f663e8e0da23747304353c6674b20d81be
        - registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:13b6dd0ddf4bdd2836b0d3ec9c79119a902456dcee100cfac6898768e1f22aab
        - registry.redhat.io/network-observability/network-observability-ebpf-agent-rhel9@sha256:0b4337f4db1e74a2bea1021a617cb42690251036f583bb771cc5ec888f38c97f
        - registry.redhat.io/network-observability/network-observability-flowlogs-pipeline-rhel9@sha256:09454618709a5c83beb4252f1c14c149716d1c09b6567c0a571d5390e8a639a2
        - registry.redhat.io/network-observability/network-observability-console-plugin-rhel9@sha256:7d92eadb3cd08a08672a65a196a5d3181e764c68e8c20cab2554933cdb10fd0c
        version: 1.9.3
      entries:
      - name: network-observability-operator.v1.9.3
        version: 1.9.3
      - name: network-observability-operator.v1.9.2
        version: 1.9.2
      - name: network-observability-operator.v1.9.1
        version: 1.9.1
      - name: network-observability-operator.v1.9.0
        version: 1.9.0
      - name: network-observability-operator.v1.8.1
        version: 1.8.1
      - name: network-observability-operator.v1.8.0
        version: 1.8.0
      - name: network-observability-operator.v1.7.0
        version: 1.7.0
      - name: network-observability-operator.v1.6.2
        version: 1.6.2
      - name: network-observability-operator.v1.6.1
        version: 1.6.1
      - name: network-observability-operator.v1.6.0
        version: 1.6.0
      - name: network-observability-operator.v1.5.0
        version: 1.5.0
      - name: network-observability-operator.v1.4.2
        version: 1.4.2
      - name: network-observability-operator.v1.4.1
        version: 1.4.1
      - name: network-observability-operator.v1.4.0
        version: 1.4.0
      - name: network-observability-operator.v1.3.0
        version: 1.3.0
      - name: network-observability-operator.v1.2.0
        version: 1.2.0
      - name: network-observability-operator.v1.1.0-202302110050
        version: 1.1.0
      - name: network-observability-operator.v1.0.0-202301070345
        version: 1.0.0
      name: stable
    - currentCSV: network-observability-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:fd943746840cba26996dabb47b1beb3445571fa43cb1b859fb073e4aa114b376
          createdAt: 2023-06-22T09:31:50
          description: Network flows collector and monitoring solution
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
          - description: FlowCollector is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
        description: |
          Network Observability is an OpenShift operator that deploys a monitoring pipeline to collect and enrich network flows that are produced by the Network Observability eBPF agent.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When a `FlowCollector` instance is created, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For normal usage, we recommend two options:

          - Installing the [Loki Operator](https://docs.openshift.com/container-platform/4.11//logging/cluster-logging-loki.html). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          oc create namespace netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.0.2/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.0.2/config/samples/flows_v1beta1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          oc edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.0.2/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: Network Observability
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:fd943746840cba26996dabb47b1beb3445571fa43cb1b859fb073e4aa114b376
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:37b3c9b2eed351d9d1a39c99045bbc7993ccb819f67f7b390deaa34930f45946
        - registry.redhat.io/network-observability/network-observability-console-plugin-rhel9@sha256:880c0812a2cd4b7194f417e31fbdcee951ce76776fa06e33c5c9cd479a854daa
        - registry.redhat.io/network-observability/network-observability-ebpf-agent-rhel9@sha256:e31acc12d8690c339e13b5dd22d33e707701362bdd31086c9b4c0f48cad62cc3
        - registry.redhat.io/network-observability/network-observability-flowlogs-pipeline-rhel9@sha256:7c7c42fff63e0599bb943a9e9617b43181360d5a5254b27f368010af275cf0c3
        version: 1.3.0
      entries:
      - name: network-observability-operator.v1.3.0
        version: 1.3.0
      - name: network-observability-operator.v1.2.0
        version: 1.2.0
      - name: network-observability-operator.v1.1.0-202302110050
        version: 1.1.0
      - name: network-observability-operator.v1.0.0-202301070345
        version: 1.0.0
      name: v1.0.x
    defaultChannel: stable
    packageName: netobserv-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: openshift-builds-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-builds-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "openshift-builds"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-rhel8-operator@sha256:7100a83aa053a976ceb77ace79f090ec8601bc9b2152ff6f7d00f0b1deea7f0b
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n\n## Usage\n\nTo deploy and manage [builds
          for Red Hat OpenShift](https://github.com/shipwright-io/build) in your cluster,\nfirst
          make sure this operator is installed and running on your cluster.\n\nNext,
          create the following:\n\n```yaml\n---\napiVersion: operator.shipwright.io/v1alpha1\nkind:
          ShipwrightBuild\nmetadata:\n  name: openshift-builds-operator\nspec:\n  targetNamespace:
          openshift-builds\n```\n\nThe operator will deploy Shipwright Build in the
          provided `targetNamespace`.\nWhen `.spec.targetNamespace` is not set, the
          namespace will default to ` openshift-builds`.\nRefer to the [ShipwrightBuild
          operator documentation](https://github.com/shipwright-io/operator/blob/main/docs/shipwrightbuild.md)
          for more information about this custom resource.\n"
        displayName: builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://shipwright.io
        - name: Shipwright Builds
          url: https://github.com/shipwright-io/build
        maintainers:
        - email: pipeline-integrations-team@redhat.com
          name: Red Hat Pipelines Integration Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-rhel8-operator@sha256:7100a83aa053a976ceb77ace79f090ec8601bc9b2152ff6f7d00f0b1deea7f0b
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel8@sha256:73efcb6ad82f698d7463718b06c86f2b2b3b5326689a62dc39c9f5d68f8943f2
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel8@sha256:0cf079a24b8b3a4b3e1f30710fea85039d9cf323cfedd06e7d1e33e6336fb848
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel8@sha256:22930726a25b11de2571da0395b53352939d142a16f4db6fd5460abd4cc37ba5
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel8@sha256:fd903804672ec08acf9a2b49aa7ecc5f9b543c12f7462637ef756733171b3817
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel8@sha256:f95348e5fc48676d93ebc919af2c27a041dfd632ec285296246d886daab11576
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel8@sha256:d01c1749ce76dd4bf5fefaa6c553725d33533d6e70e884e916f12217b674c3cf
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:97cade2c1ee468261aec5400728c8d44de387b459134aec7a4c3b5ec5a335d2c
        version: 1.0.2
      entries:
      - name: openshift-builds-operator.v1.0.2
        version: 1.0.2
      - name: openshift-builds-operator.v1.0.1
        version: 1.0.1
      - name: openshift-builds-operator.v1.0.0
        version: 1.0.0
      name: builds-1.0
    - currentCSV: openshift-builds-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          createdAt: "2025-01-23T14:48:43Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.openshift.com/builds/latest/about/overview-openshift-builds.html
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:778ad1946bdad68e1430a2fb366cee25cd98d87e44cf5ad34b87603c2f4171d0
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:e8582857566020c71653228355cbadc2c1cda41e36fc767c204a4edb05e7f42a
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:fa039b62ebd1fc6e86db8ec1c7ae5fe0dd3a6f6f642b63b4dc89af32a78cea6b
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:1ec9a1dcfc08b707df48175bba76f8a6cbfcc8396d9e9940cc3d7100e6731701
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:0e139db4370c37e91ef7eb3e0cf9a2587cb81f37709009a37f2724d4461ed2f3
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:949a1a5150642167b6c7d2b8f6d92a2885afe453319c02495f351b2e79eff3dd
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:7cfffdd43fdd33a20d115f35c2ba492da3c1d0e3ab970587bc888c68c47dfacf
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:97cade2c1ee468261aec5400728c8d44de387b459134aec7a4c3b5ec5a335d2c
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:e9f7b8741f265a332008424a2c5b3dae9ea154e623c469085c2abafb53fc62d4
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:5babdf590730b217deb59acbccf142c0500bc692542131810714cc00724ee02c
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:5ec474ab9e62c92e44c1a6635913032aead6246faec0895dcfbdb28ab0dc8a1a
        version: 1.1.1
      entries:
      - name: openshift-builds-operator.v1.1.1
        version: 1.1.1
      - name: openshift-builds-operator.v1.1.0
        version: 1.1.0
      name: builds-1.1
    - currentCSV: openshift-builds-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          createdAt: "2025-01-23T13:47:38Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.2
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:ee2792972dc52d81814603e26103e0d1117194db5ab3b5577736b0fc3285a8ab
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3fa22124916523b958c67af8ad652e73a2c3d68bb5579da1cba1ade537f3b7ae
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:8ace52981058ed51cec5b893e6d773de963b12ab715258ac1f2cbc6430ad52e7
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:775c15b93b19ba855791d34c4a95ef203dd9f8e51110c67ebf131e38eda49422
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:2b288e36e3606cf8a0d7723772edb8d4ac11772ec450539ef3448ad517155ecb
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:4bee8594d17172e5b9893a865a9922a1c7869027633ad64b3f684ec44755bac6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3fcd8e2bf0bcb8ff8c93a87af2c59a3bcae7be8792f9d3236c9b5bbd9b6db3b2
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:e8582857566020c71653228355cbadc2c1cda41e36fc767c204a4edb05e7f42a
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:78ee35ccc70b78638c667152b7b6247417c0aa1bb6d3f817d59b22eb179899ca
        - registry.redhat.io/ubi8/buildah@sha256:56c71ecd450f284325ba8434cbf5b3006078b11f0625c6258689699fa811c95d
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:ebfe4f8a97f6f0f9afe38953ae8cc2fdf2ec0b76401c819bb12c1df7cd671e9e
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:016c93578908f1dbdc3052333a1d5a9ab0c03104070ac36a6ddf99350d7510f4
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:87c4c0e9fe4d9baf8ae4f191c51d90c46729273e182fb5b3dfd951632e7bfad3
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:a91c2bb1239ef7fbce8c1ab9308da03b533df9cea75c7d978fd76f38087bc912
        version: 1.2.2
      entries:
      - name: openshift-builds-operator.v1.2.2
        version: 1.2.2
      - name: openshift-builds-operator.v1.2.1
        version: 1.2.1
      - name: openshift-builds-operator.v1.2.0
        version: 1.2.0
      name: builds-1.2
    - currentCSV: openshift-builds-operator.v1.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-01-29T08:58:54Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.2
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:6e44758fd4a7b34814ac7c71e52d93633781d39f7897ecab98a090859b487b4f
        - registry.redhat.io/ubi9/buildah@sha256:27d837f9bc69ad3c3651cf3315e2501b137c11baa553d9d46140e5cf7fa7873a
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:0783132b14ed8050664810f37e7ad6388fefbc586c45ecc7ba69dd5041a55c9f
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:e8582857566020c71653228355cbadc2c1cda41e36fc767c204a4edb05e7f42a
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:04b8bfc0dc6f260ef367ca1cce2bcc84032bc8be17b4a595be12a344813a2632
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:1e9a54212690f7930d2b681db87a6f79f5a2acc5c982734d4b283a2472b59db5
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:3523e7c2215e156081e5f05443884d65efad9b7b9ca81a35416695020231916a
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:d4b3f451becb7a42341c98dd5b9fc5d2b582c93cac2728070ef4bb83b64abb4c
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:a4ba591c57bbc815c93023f4a080213be32581ba6a1c2d94b0ecbc5c90b07a1e
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:0b1dd35f55181f34775188bc71b3ff8dff66c70acfcfabad19353a02e19351f7
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:6214708f3b517cb901af9f9bd0502ba54f8e5183d0d909a3b559deee6394026e
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:ed3fdd1e85f5c5434de6ec0feaec8c7fe4e680ca6c64258287f18ec2886bb71f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3fcd8e2bf0bcb8ff8c93a87af2c59a3bcae7be8792f9d3236c9b5bbd9b6db3b2
        version: 1.3.2
      entries:
      - name: openshift-builds-operator.v1.3.2
        version: 1.3.2
      - name: openshift-builds-operator.v1.3.1
        version: 1.3.1
      - name: openshift-builds-operator.v1.3.0
        version: 1.3.0
      name: builds-1.3
    - currentCSV: openshift-builds-operator.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2024-11-15T22:28:01Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:9d12e6cbf26e31a1e4458c99a972ed36125d759c5ec57773f203cd8dd0985de7
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:ed3fdd1e85f5c5434de6ec0feaec8c7fe4e680ca6c64258287f18ec2886bb71f
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:538416bc1384eb987e8a7e80b57eda3feddaf4260b196709a6e9a9125e3af39e
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:1a6ca51c2c68950af605ff367356ab6c0ef7d1b2fd6fb8f00aa665196ff7af84
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:f4b935c16dcbbc0393733437747497c1d24f2cc782c40c49738e5f37c037b3e7
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:c60ca44275297d2ee8c499d9d7fb4056d104f4bd5dec8a02e7388241ac466c44
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:1c7c24111f1709d7648156873ced09462638aa4b3cf7e7aee13297e51bcdfb3e
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:6251cb303774e29d137a37d6527875c371fe1d3f6a5fa0383e80f0e9ef5c02be
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:f87c2069105121927fb0c0a2e37f8f72bc6cb5e035fa8d22f7b41ce0e86c0631
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:4e9c152023b2db04ee5d55f102490294b393bf78baa3aa8917f85f8547fbe3c8
        - registry.redhat.io/ubi9/buildah@sha256:6886997da8c92214e99e3e8da47a0f08239172d2544b4edae058d5cb018effab
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:adc4a31d63b7044db7094fddcb1a4b235d848e842257f5baf22b6ead03fd6c91
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:5e87ba403f9e9fff3002a8275aba0d6c36aed54fb4754d90fa0516ad989f187a
        version: 1.4.1
      entries:
      - name: openshift-builds-operator.v1.4.1
        version: 1.4.1
      - name: openshift-builds-operator.v1.4.0
        version: 1.4.0
      name: builds-1.4
    - currentCSV: openshift-builds-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-08-14T12:12:15Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              OpenShiftBuild describes the desired state of Builds for OpenShift, and the status of
              all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I), Buildah, and Buildpacks. You can create and apply
          build resources, view logs of build runs, \nand manage builds in your OpenShift
          Container Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:38560dea7627632d8b24fa670c1d7ac1218cf86975329908786a460f76de558d
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ce7050192d9850ed4c35cb3295ccdeec1aa593e27e3a3da6d6aca97cde782abe
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c1a84feb88d53e93bdcf2cd5f76e2cbb18541c8b0e2979c132a888d6c280b664
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:ddcfcebd5d788d11f756939f07215b2e3507a1d8fe704a1052e15bb4ebe592e8
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:753e24a1cd57ca22552447363aed48f064b525b49e52f774f3e25f41d8dcb329
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:e2a7a42cd5916c05d95d3fb0decc34d0840420532270422aa7f6b7bdd94de1f2
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:bcae84267f582278591d93aecf94c8d7fb1376521d9b5208201c959a0fe56774
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:f4a75c115b5dc8853845f666f3bd063b2736e01e428a605fb5cd0f02736ecce8
        - registry.redhat.io/ubi9/buildah@sha256:8ea2b81b32f70b9502d88722a5f69de8c0cf5efa6f30df041873546e9f9438cb
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:4dc123fa703006662efcc5569a843a043bb31cf69b6868d6a9b6321c66bfca18
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:1d5a0bfe0013ab7602efe95be8a814be1e5586728fdf0e198c97739cc12f1d87
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:504179e8ee9f88a23f1d4c28c7683ab7395f4c1c4868765bf3c1406669fd52fc
        version: 1.5.1
      entries:
      - name: openshift-builds-operator.v1.5.1
        version: 1.5.1
      - name: openshift-builds-operator.v1.5.0
        version: 1.5.0
      name: builds-1.5
    - currentCSV: openshift-builds-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-09-23T13:38:59Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I), Buildah, and Buildpacks. You can create and apply
          build resources, view logs of build runs, \nand manage builds in your OpenShift
          Container Platform namespaces.\n\n## Prerequisites\n\n* OpenShift Pipelines
          operator must be installed before installing this operator\n\nRead more:
          [https://shipwright.io](https://shipwright.io)\n\n## Features\n\n* Standard
          Kubernetes-native API for building container images from source code and
          Dockerfile\n\n* Support for Source-to-Image (S2I) and Buildah build strategies\n\n*
          Extensibility with your own custom build strategies\n\n* Execution of builds
          from source code in a local directory\n\n* Shipwright CLI for creating and
          viewing logs, and managing builds on the cluster\n\n* Integrated user experience
          with the Developer perspective of the OpenShift Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:c6805e85ad4dc29f3a0582a1fbf99647a90c1af0b3924647839ad881cbb199fa
        - registry.redhat.io/ubi9/buildah@sha256:8ea2b81b32f70b9502d88722a5f69de8c0cf5efa6f30df041873546e9f9438cb
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:f9f43795c7ed1e416b92497cf2fe647beb5e45d755f6a6d89c3915b7a9c76570
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:1a61d4a8f69fa845ee678dd96f934685821828e7575852f97b6e7b1dd79a8297
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c1a84feb88d53e93bdcf2cd5f76e2cbb18541c8b0e2979c132a888d6c280b664
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:3d2e501abe1aa757e4e6ba03fe8e7a45bbfdb4303360ed7130c6344051887ed8
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:9f38cd42591b6a6b65e78a3cc5c165f8498ad3c79460a6c47acfb8744d2dd883
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:0ea2b0153ae07ee15972abbbe034ef7feb0dcaa403ac993afcb171906d9cc796
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:23e3f29ae63546db018701abb29b444b9b825fb3a2bda5084542649d9827d4ba
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:dd96e8c857bc52097d4d194db8fa2cb368bbdcbb6594722660d4e906d0bdc420
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ce7050192d9850ed4c35cb3295ccdeec1aa593e27e3a3da6d6aca97cde782abe
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:fd7c805c6d0638f2814503c6779719279119c7268b1eb0eda9b7b58ac775efa6
        version: 1.6.0
      entries:
      - name: openshift-builds-operator.v1.6.0
        version: 1.6.0
      name: builds-1.6
    - currentCSV: openshift-builds-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-09-23T13:38:59Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I), Buildah, and Buildpacks. You can create and apply
          build resources, view logs of build runs, \nand manage builds in your OpenShift
          Container Platform namespaces.\n\n## Prerequisites\n\n* OpenShift Pipelines
          operator must be installed before installing this operator\n\nRead more:
          [https://shipwright.io](https://shipwright.io)\n\n## Features\n\n* Standard
          Kubernetes-native API for building container images from source code and
          Dockerfile\n\n* Support for Source-to-Image (S2I) and Buildah build strategies\n\n*
          Extensibility with your own custom build strategies\n\n* Execution of builds
          from source code in a local directory\n\n* Shipwright CLI for creating and
          viewing logs, and managing builds on the cluster\n\n* Integrated user experience
          with the Developer perspective of the OpenShift Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:0ea2b0153ae07ee15972abbbe034ef7feb0dcaa403ac993afcb171906d9cc796
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:c6805e85ad4dc29f3a0582a1fbf99647a90c1af0b3924647839ad881cbb199fa
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:3d2e501abe1aa757e4e6ba03fe8e7a45bbfdb4303360ed7130c6344051887ed8
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:fd7c805c6d0638f2814503c6779719279119c7268b1eb0eda9b7b58ac775efa6
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:23e3f29ae63546db018701abb29b444b9b825fb3a2bda5084542649d9827d4ba
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ce7050192d9850ed4c35cb3295ccdeec1aa593e27e3a3da6d6aca97cde782abe
        - registry.redhat.io/ubi9/buildah@sha256:8ea2b81b32f70b9502d88722a5f69de8c0cf5efa6f30df041873546e9f9438cb
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:f9f43795c7ed1e416b92497cf2fe647beb5e45d755f6a6d89c3915b7a9c76570
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:1a61d4a8f69fa845ee678dd96f934685821828e7575852f97b6e7b1dd79a8297
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:9f38cd42591b6a6b65e78a3cc5c165f8498ad3c79460a6c47acfb8744d2dd883
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c1a84feb88d53e93bdcf2cd5f76e2cbb18541c8b0e2979c132a888d6c280b664
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:dd96e8c857bc52097d4d194db8fa2cb368bbdcbb6594722660d4e906d0bdc420
        version: 1.6.0
      entries:
      - name: openshift-builds-operator.v1.6.0
        version: 1.6.0
      - name: openshift-builds-operator.v1.5.1
        version: 1.5.1
      - name: openshift-builds-operator.v1.5.0
        version: 1.5.0
      - name: openshift-builds-operator.v1.4.1
        version: 1.4.1
      - name: openshift-builds-operator.v1.4.0
        version: 1.4.0
      - name: openshift-builds-operator.v1.3.2
        version: 1.3.2
      - name: openshift-builds-operator.v1.3.1
        version: 1.3.1
      - name: openshift-builds-operator.v1.3.0
        version: 1.3.0
      - name: openshift-builds-operator.v1.2.2
        version: 1.2.2
      - name: openshift-builds-operator.v1.2.1
        version: 1.2.1
      - name: openshift-builds-operator.v1.2.0
        version: 1.2.0
      - name: openshift-builds-operator.v1.1.1
        version: 1.1.1
      - name: openshift-builds-operator.v1.1.0
        version: 1.1.0
      - name: openshift-builds-operator.v1.0.2
        version: 1.0.2
      - name: openshift-builds-operator.v1.0.1
        version: 1.0.1
      - name: openshift-builds-operator.v1.0.0
        version: 1.0.0
      name: latest
    defaultChannel: latest
    packageName: openshift-builds-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: kaoto.io
      provider-url: https://kaoto.io
    name: kaoto-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kaoto-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "designer.kaoto.io/v1alpha1",
                "kind": "Kaoto",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "kaoto-operator",
                    "app.kubernetes.io/instance": "kaoto-demo",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kaoto",
                    "app.kubernetes.io/part-of": "kaoto-operator"
                  },
                  "name": "kaoto-demo"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          containerImage: quay.io/kaotoio/kaoto-operator@sha256:590d4958f148954ea1edb9825fb7fb4935d041ddcdb4a99302da14a605380071
          createdAt: "2025-09-03T12:48:58Z"
          description: Integration Editor for Apache Camel
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/KaotoIO/kaoto-operator
          support: contact@kaoto.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kaoto is the Schema for the kaotoes API.
            displayName: Kaoto
            kind: Kaoto
            name: kaotoes.designer.kaoto.io
            version: v1alpha1
        description: Kaoto is a low code and no code integration designer to edit
          and deploy based on Apache Camel . It is designed to be extendable and flexible
          and adaptable to different usecases.
        displayName: Kaoto Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - kaoto
        - camel-k
        - designer
        links:
        - name: Kaoto Operator
          url: https://kaoto-operator.domain
        maintainers:
        - email: info@kaoto.io
          name: kaoto
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: kaoto.io
          url: https://kaoto.io
        relatedImages:
        - quay.io/kaotoio/kaoto-operator@sha256:590d4958f148954ea1edb9825fb7fb4935d041ddcdb4a99302da14a605380071
        version: 2.7.0
      entries:
      - name: kaoto-operator.v2.7.0
        version: 2.7.0
      - name: kaoto-operator.v2.6.0
        version: 2.6.0
      - name: kaoto-operator.v2.5.0
        version: 2.5.0
      - name: kaoto-operator.v2.4.0
        version: 2.4.0
      - name: kaoto-operator.v2.3.0
        version: 2.3.0
      - name: kaoto-operator.v2.1.0
        version: 2.1.0
      - name: kaoto-operator.v2.0.1
        version: 2.0.1
      - name: kaoto-operator.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: kaoto-operator
    provider:
      name: kaoto.io
      url: https://kaoto.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-lambda-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-lambda-controller.v1.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "Alias",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "CodeSigningConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "EventSourceMapping",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/lambda-controller:1.9.1
          createdAt: "2025-09-19T19:23:07Z"
          description: AWS Lambda controller is a service controller for managing
            Lambda resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alias represents the state of an AWS lambda Alias resource.
            displayName: Alias
            kind: Alias
            name: aliases.lambda.services.k8s.aws
            version: v1alpha1
          - description: CodeSigningConfig represents the state of an AWS lambda CodeSigningConfig
              resource.
            displayName: CodeSigningConfig
            kind: CodeSigningConfig
            name: codesigningconfigs.lambda.services.k8s.aws
            version: v1alpha1
          - description: EventSourceMapping represents the state of an AWS lambda
              EventSourceMapping resource.
            displayName: EventSourceMapping
            kind: EventSourceMapping
            name: eventsourcemappings.lambda.services.k8s.aws
            version: v1alpha1
          - description: Function represents the state of an AWS lambda Function resource.
            displayName: Function
            kind: Function
            name: functions.lambda.services.k8s.aws
            version: v1alpha1
          - description: FunctionURLConfig represents the state of an AWS lambda FunctionURLConfig
              resource.
            displayName: FunctionURLConfig
            kind: FunctionURLConfig
            name: functionurlconfigs.lambda.services.k8s.aws
            version: v1alpha1
          - description: LayerVersion represents the state of an AWS lambda LayerVersion
              resource.
            displayName: LayerVersion
            kind: LayerVersion
            name: layerversions.lambda.services.k8s.aws
            version: v1alpha1
          - description: Version represents the state of an AWS lambda Version resource.
            displayName: Version
            kind: Version
            name: versions.lambda.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Lambda resources in AWS from within your Kubernetes cluster.

          **About Amazon Lambda**

          Lambda is a compute service that lets you run code without provisioning or managing servers. Lambda runs your code on a high-availability compute infrastructure and performs all of the administration of the compute resources, including server and operating system maintenance, capacity provisioning and automatic scaling, code monitoring and logging. With Lambda, you can run code for virtually any type of application or backend service. All you need to do is supply your code in one of the [languages that Lambda supports](https://docs.aws.amazon.com/lambda/latest/dg/lambda-runtimes.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Lambda
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lambda
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Lambda Developer Resources
          url: https://aws.amazon.com/lambda/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: lambda maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/lambda-controller:1.9.1
        version: 1.9.1
      entries:
      - name: ack-lambda-controller.v1.9.1
        version: 1.9.1
      - name: ack-lambda-controller.v1.9.0
        version: 1.9.0
      - name: ack-lambda-controller.v1.8.0
        version: 1.8.0
      - name: ack-lambda-controller.v1.7.7
        version: 1.7.7
      - name: ack-lambda-controller.v1.7.6
        version: 1.7.6
      - name: ack-lambda-controller.v1.7.5
        version: 1.7.5
      - name: ack-lambda-controller.v1.7.4
        version: 1.7.4
      - name: ack-lambda-controller.v1.7.2
        version: 1.7.2
      - name: ack-lambda-controller.v1.7.1
        version: 1.7.1
      - name: ack-lambda-controller.v1.7.0
        version: 1.7.0
      - name: ack-lambda-controller.v1.6.3
        version: 1.6.3
      - name: ack-lambda-controller.v1.6.2
        version: 1.6.2
      - name: ack-lambda-controller.v1.6.1
        version: 1.6.1
      - name: ack-lambda-controller.v1.6.0
        version: 1.6.0
      - name: ack-lambda-controller.v1.5.8
        version: 1.5.8
      - name: ack-lambda-controller.v1.5.7
        version: 1.5.7
      - name: ack-lambda-controller.v1.5.6
        version: 1.5.6
      - name: ack-lambda-controller.v1.5.5
        version: 1.5.5
      - name: ack-lambda-controller.v1.5.4
        version: 1.5.4
      - name: ack-lambda-controller.v1.5.3
        version: 1.5.3
      - name: ack-lambda-controller.v1.5.2
        version: 1.5.2
      - name: ack-lambda-controller.v1.5.1
        version: 1.5.1
      - name: ack-lambda-controller.v1.5.0
        version: 1.5.0
      - name: ack-lambda-controller.v1.4.2
        version: 1.4.2
      - name: ack-lambda-controller.v1.4.0
        version: 1.4.0
      - name: ack-lambda-controller.v1.3.4
        version: 1.3.4
      - name: ack-lambda-controller.v1.3.2
        version: 1.3.2
      - name: ack-lambda-controller.v1.3.1
        version: 1.3.1
      - name: ack-lambda-controller.v1.3.0
        version: 1.3.0
      - name: ack-lambda-controller.v1.2.1
        version: 1.2.1
      - name: ack-lambda-controller.v1.2.0
        version: 1.2.0
      - name: ack-lambda-controller.v1.1.1
        version: 1.1.1
      - name: ack-lambda-controller.v1.1.0
        version: 1.1.0
      - name: ack-lambda-controller.v1.0.1
        version: 1.0.1
      - name: ack-lambda-controller.v1.0.0
        version: 1.0.0
      - name: ack-lambda-controller.v0.1.7
        version: 0.1.7
      - name: ack-lambda-controller.v0.1.6
        version: 0.1.6
      - name: ack-lambda-controller.v0.1.5
        version: 0.1.5
      - name: ack-lambda-controller.v0.1.3
        version: 0.1.3
      - name: ack-lambda-controller.v0.1.2
        version: 0.1.2
      - name: ack-lambda-controller.v0.1.1
        version: 0.1.1
      - name: ack-lambda-controller.v0.1.0
        version: 0.1.0
      - name: ack-lambda-controller.v0.0.17
        version: 0.0.17
      - name: ack-lambda-controller.v0.0.16
        version: 0.0.16
      - name: ack-lambda-controller.v0.0.15
        version: 0.0.15
      - name: ack-lambda-controller.v0.0.14
        version: 0.0.14
      - name: ack-lambda-controller.v0.0.13
        version: 0.0.13
      - name: ack-lambda-controller.v0.0.12
        version: 0.0.12
      - name: ack-lambda-controller.v0.0.11
        version: 0.0.11
      - name: ack-lambda-controller.v0.0.10
        version: 0.0.10
      - name: ack-lambda-controller.v0.0.9
        version: 0.0.9
      - name: ack-lambda-controller.v0.0.8
        version: 0.0.8
      - name: ack-lambda-controller.v0.0.7
        version: 0.0.7
      - name: ack-lambda-controller.v0.0.6
        version: 0.0.6
      - name: ack-lambda-controller.v0.0.5
        version: 0.0.5
      - name: ack-lambda-controller.v0.0.4
        version: 0.0.4
      - name: ack-lambda-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-lambda-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: fence-agents-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fence-agents-remediation.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/fence-agents-remediation-operator:v0.4.0
          createdAt: "2024-05-23 19:57:04"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://www.medik8s.io/remediation/fence-agents-remediation/fence-agents-remediation/
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/medik8s/fence-agents-remediation-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      - name: fence-agents-remediation.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: fence-agents-remediation
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: hpe-filex-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-filex-csi-operator.v1.0.0-beta3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.hpe.com/v1",
                "kind": "HPEGreenLakeFileCSIDriver",
                "metadata": {
                  "name": "hpegreenlakefilecsidriver-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "disableNodeConformance": false,
                  "imagePullPolicy": "IfNotPresent",
                  "images": {
                    "csiAttacher": "registry.k8s.io/sig-storage/csi-attacher@sha256:6e54dae32284f60a2de1645c527715e4c934e5ce7899560c0d845bac957118dc",
                    "csiControllerDriver": "quay.io/hpestorage/filex-csi-driver@sha256:8e96df8f21db969dcd507c6fef8ecc82b7f4d9b33a9159f16345ebe7e02d8161",
                    "csiNodeDriver": "quay.io/hpestorage/filex-csi-driver@sha256:8e96df8f21db969dcd507c6fef8ecc82b7f4d9b33a9159f16345ebe7e02d8161",
                    "csiNodeDriverRegistrar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f",
                    "csiNodeInit": "quay.io/hpestorage/filex-csi-init@sha256:65a014b344d262f628100dcd96fecd228c42162bbfa67037952a5a67f16a1437",
                    "csiProvisioner": "registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f",
                    "csiResizer": "registry.k8s.io/sig-storage/csi-resizer@sha256:ab774734705a906561e15b67f6a96538f3ad520335d636f793aaafe87a3b5200",
                    "csiSnapshotter": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:dd788d79cf4c1b8edee6d9b80b8a1ebfc51a38a365c5be656986b129be9ac784"
                  },
                  "kubeletRootDir": "/var/lib/kubelet",
                  "node": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/hpestorage/filex-csi-driver-operator@sha256:e00066fccf2c673547ccafffb0a280130b6825ab6a82a52bd43d752535c63775
          createdAt: "2025-01-10T18:20:35Z"
          description: A Container Storage Interface (CSI) driver for HPE GreenLake
            File Storage.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.0.0-beta3'
          operatorhub.io/ui-metadata-max-k8s-version: "1.32"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/hpe-storage/co-deployments
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE GreenLake File Storage CSI Driver
            displayName: HPEGreenLakeFileCSIDriver
            kind: HPEGreenLakeFileCSIDriver
            name: hpegreenlakefilecsidrivers.storage.hpe.com
            version: v1
        description: |
          The HPE GreenLake File Storage CSI Operator deploys the  HPE GreenLake File Storage CSI Driver on OpenShift for dynamic  provisioning of persistent volumes.
          Please refer to our CSI driver [documentation](https://scod.hpedev.io/filex_csi_driver/index.html) for supported HPE platforms and full list of supported CSI features.
          ## Installation
          Refer to the HPE GreenLake File Storage CSI Operator [documentation](https://scod.hpedev.io/filex_csi_driver/deployment.html#operator).
        displayName: HPE GreenLake File Storage CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE CSI
        - HPE Storage
        - CSI
        links:
        - name: Documentation
          url: https://scod.hpedev.io/filex_csi_driver/index.html
        maintainers:
        - email: dev-hi-containers@hpe.com
          name: HPE Storage Containers Team
        maturity: beta
        minKubeVersion: 1.25.0
        provider:
          name: HPE
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - quay.io/hpestorage/filex-csi-init@sha256:65a014b344d262f628100dcd96fecd228c42162bbfa67037952a5a67f16a1437
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:dd788d79cf4c1b8edee6d9b80b8a1ebfc51a38a365c5be656986b129be9ac784
        - registry.k8s.io/sig-storage/csi-resizer@sha256:ab774734705a906561e15b67f6a96538f3ad520335d636f793aaafe87a3b5200
        - quay.io/hpestorage/filex-csi-driver@sha256:8e96df8f21db969dcd507c6fef8ecc82b7f4d9b33a9159f16345ebe7e02d8161
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3fcd8e2bf0bcb8ff8c93a87af2c59a3bcae7be8792f9d3236c9b5bbd9b6db3b2
        - registry.connect.redhat.com/hpestorage/filex-csi-driver-operator@sha256:e00066fccf2c673547ccafffb0a280130b6825ab6a82a52bd43d752535c63775
        - registry.k8s.io/sig-storage/csi-attacher@sha256:6e54dae32284f60a2de1645c527715e4c934e5ce7899560c0d845bac957118dc
        version: 1.0.0-beta3
      entries:
      - name: hpe-filex-csi-operator.v1.0.0-beta3
        version: 1.0.0-beta3
      name: beta
    defaultChannel: beta
    packageName: hpe-filex-csi-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SmartX
      provider-url: https://www.smartx.com
    name: iomesh-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: iomesh-operator.0.10.1-rc4
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
          createdAt: 2021-08-17 00:00:00+00:00
          repository: https://github.com/iomesh/docs
          support: SmartX, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: iomesh
            displayName: iomesh
            kind: IOMeshCluster
            name: iomeshclusters.iomesh.com
            version: v1alpha1
        description: "IOMesh (https://www.iomesh.com/) is a cloud-native storage product
          specifically designed and developed for Kubernetes. It takes ZBS as its
          core, a “production-ready” distributed storage system developed independently
          by SmartX, to provide production-level high performance and reliable persistent
          storage for business-critical applications such as MySQL, Cassandra, MongoDB,
          etc., contributing to the containerized transformation of stateful applications.
          \     \n## Pre-Requisites\n* According to https://docs.iomesh.com/deploy/prerequisites,
          make sure that your openshift cluster meets the \"Installation Requirements\"
          and set up the worker nodes correctly according to the \"Setup Worker Node\"
          chapter.\n* Run IOMesh Operator pre-installation script in an environment
          where oc or kubectl can be used to access the openshift cluster, the script
          will install the dependencies of IOMesh Operator and specific IOMesh settings
          for the openshift cluster:\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-pre-install-openshift.sh
          | sh -\n```\n## Install IOMesh Operator and IOMesh Cluster\n* Click install
          button on the current web page to install IOMesh Operator\n* On the **Installed
          Operators** -> **IOMesh Operator** -> **Create instance** -> **YAML view**,
          fill in an IOMesh Custom Resources according https://iomesh.run/iomesh.yaml,
          change the `spec.*.dataCIDR` to your own data network CIDR:\n## Post Install\n*
          Run IOMesh Operator post-installation script in an environment where oc
          or kubectl can be used to access the openshift cluster, the script will
          install IOMesh CSI Driver\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-post-install-openshift.sh
          | sh -\n```\n## Getting Started\n* Try to quickstart setup and use IOMesh:
          https://docs.iomesh.com/deploy/setup-iomesh\n"
        displayName: IOMesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iomesh-operator
        links:
        - name: IOMesh Operator
          url: https://iomesh.com
        maintainers:
        - email: na.wang@smartx.com
          name: Na Wang
        maturity: alpha
        provider:
          name: SmartX
          url: https://www.smartx.com
        relatedImages:
        - iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
        version: 0.10.1-rc4
      entries:
      - name: iomesh-operator.0.10.1-rc4
        version: 0.10.1-rc4
      name: stable
    defaultChannel: stable
    packageName: iomesh-operator
    provider:
      name: SmartX
      url: https://www.smartx.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-activemqartemis: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/
    name: amq-broker-rhel9
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-broker-operator.v7.13.1-opr-1-0.1758027574.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 9 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-broker-operator.v7.13.1-opr-1-0.1756381365.p
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 9 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel9-operator@sha256:f34672152017a94ed51056435e218341e22787788ceed52439c1ff066480c715
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9cc48eecf1442ae04b8543fa5d4381a13bc2831390850828834d387006d1342b
        - registry.redhat.io/amq7/amq-broker-rhel9@sha256:755d0f7421efba76f52818bb3a8667cd7d2dcff2c7054dac1a4637b74ecac51d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:47fd4ce2533496828aba37bd1f9715e2164d5c90bd0fc6b25e7e0786d723bf01
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b2402179b1d0ff9c950defe5d8808195ef9a4c3e13e3c469cae31ab51d5f5014
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:bb6fbd68475a7852b4d99eea6c4ab313f9267da7963162f0d75375d7063409e7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:40920687402fd747ca02d384e63692a5e1db625723a2bf3056a607c92ee8eee1
        - registry.redhat.io/amq7/amq-broker-rhel9@sha256:4304a5414911fe55ce4c0133723be234ff288463f4916bd18e743d6359489138
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:7f9b921d89b6d4d91b365659afff127d99f1455b94b48a10a6a7f4c38dd8b700
        - registry.redhat.io/amq7/amq-broker-init-rhel9@sha256:ff80e0a7accda6825bf834d9aec27f48205dd6329991890dd017b4d21d0da5c8
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:16253895bc113f395dc6ab5cd511901e43b6ee676369ec80d598a8741c3c9bf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d42d713da0ce6806fdc6492b6342586783e6865a82a8647d3c4288439b1751ee
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:dc13871dfed52402c40fd192900fc3cb5a49da4fa08e4ec4d0e23a879e83db25
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ffffa9875f0379e9373f89f05eb06e5a193273bb04bc3aa5f85b044357b79098
        - registry.redhat.io/amq7/amq-broker-init-rhel9@sha256:e2cb996e0bf131e71fb8e80c128af18ab2b9330c9fca7a5437842796158e2870
        version: 7.13.1-opr-1+0.1758027574.p
      entries:
      - name: amq-broker-operator.v7.13.1-opr-1
        version: 7.13.1-opr-1
      - name: amq-broker-operator.v7.13.1-opr-1-0.1754564924.p
        version: 7.13.1-opr-1+0.1754564924.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1756374447.p
        version: 7.13.1-opr-1+0.1756374447.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1756381365.p
        version: 7.13.1-opr-1+0.1756381365.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1758024989.p
        version: 7.13.1-opr-1+0.1758024989.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1758027574.p
        version: 7.13.1-opr-1+0.1758027574.p
      - name: amq-broker-operator.v7.13.0-opr-1
        version: 7.13.0-opr-1
      - name: amq-broker-operator.v7.13.0-opr-1-0.1747315188.p
        version: 7.13.0-opr-1+0.1747315188.p
      - name: amq-broker-operator.v7.13.0-opr-1-0.1747658456.p
        version: 7.13.0-opr-1+0.1747658456.p
      name: 7.13.x
    defaultChannel: 7.13.x
    packageName: amq-broker-rhel9
    provider:
      name: Red Hat
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: quay-bridge-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-bridge-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "my-quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecretName": "<NAMESPACE>/<SECRET>",
                  "quayHostname": "https://<QUAY_URL>",
                  "insecureRegistry": false,
                  "whitelistNamespaces": ["default"]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:501fe9b003b5066b3936def0ef4de05abd1a12e795e718d6c9a62d2c6c804c2e
          createdAt: 2021-11-02 11:11 UTC
          description: Enhance OCP using Red Hat Quay container registry
          repository: https://github.com/redhat-cop/quay-openshift-registry-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a connected Quay instance.
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.redhatcop.redhat.io
            version: v1alpha1
        description: "## Overview\n\nThe OpenShift Container Platform contains an
          in cluster image registry for pushing and pulling images that either produced
          within the cluster or are references from external sources. Quay is a highly
          available, enterprise grade image registry with enhanced role based access
          control (RBAC) features. The goal of this operator is to support replicating
          the features found within OpenShift intgrated image registry with Quay.\n\n##
          Functionality\n\nThe following capabilities are enabled through this operator:\n\n*
          Synchronization of OpenShift namespaces as Quay organizations\n    * Creation
          of Robot accounts for each default namespace service account\n    * Creation
          of Secrets for each created Robot Account\n        * Associate Robot Secret
          to Service Account as Mountable and Image Pull Secret\n    * Synchronization
          of OpenShift ImageStreams as Quay Repositories\n* Automatically rewrite
          new Builds making use of ImageStreams to output to Quay\n* Automatically
          import ImageStream tag once build completes\n\n## Prerequisites\n\nThe following
          requirements must be satisfied prior to setting and configuring the integration
          between OpenShift and Quay:\n\n* A Red Hat Quay environment with a user
          with superuser permissions\n* A Red Hat OpenShift Container Platform environment
          (at least version 4.0) with cluster-admin permissions on an OpenShift cluster
          \n*\tOpenShift Command Line tool (CLI)\n\n## Setup and Configuration\n\nThe
          setup process involves steps within both Quay and OpenShift as described
          below:\n\n### Quay Setup\n\nThe integration between OpenShift and Quay makes
          extensive use of the Quay RESTFul API. To allow for other methods aside
          from user credentials to be used for authenticating against the API, an
          application is required to be created. When an application is created, an
          OAuth token is generated which can be used by the operator in OpenShift.
          Applications are created within a Quay organization. While an existing organization
          can be utilized, it is recommended that a dedicated organization be used.\n\n1.\tLogin
          to Quay as a user with superuser access and select the organization for
          which the external application will be configured within.\n2.\tOn the left
          hand side of the page, select **Applications**\n3.\tCreate a new application
          by selecting the **Create New Application** button and entering a name for
          the application\n4.\tAfter the new application has been created, select
          the application that was created\n5.\tOn the lefthand navigation bar, select
          the **Generate Token** button to create a new OAuth2 token\n6.\tSelect all
          of the checkboxes to grant the necessary access that is needed for the integration\n7.\tReview
          the permissions that will be assigned and then select the **Authorize Application**
          button\n8.\tTake note of the generated Access Token as it will be needed
          in the following section\n\n### OpenShift Setup\n\nTo complete the process
          to begin to integrate OpenShift and Quay, several steps are required. Before
          beginning, ensure that you have the OpenShift Command Line tool installed
          and that you are logged into OpenShift as a cluster administrator. \n\n####
          Create A Secret for the Quay OAuth Token\n\nThe Operator will use the previously
          obtained Access Token to communicate with Quay. Store this token within
          OpenShift as a secret.\n\nExecute the following command to create a secret
          called `quay-integration` with a key called `token` containing the access
          token:\n\n```\n$ oc create secret generic quay-integration --from-literal=token=<access_token>\n```\n\nThis
          token will be referenced in the final section.\n\n#### MutatingWebhookConfiguration
          Support\n\nSupport for dynamic interception of API requests that are performed
          as part of OpenShift’s typical build process is facilitated through a [MutatingWebhookConfiguration](https://docs.openshift.com/container-platform/3.11/architecture/additional_concepts/dynamic_admission_controllers.html).
          A MutatingWebhookConfiguration allows for invoking an API running within
          a project on OpenShift when certain API requests are received. In particular,
          to support Quay’s integration with OpenShift, any new Build requests should
          be intercepted so that the output can be modified to target Quay instead
          of OpenShift’s integrated registry. \n\nKubernetes requires that the webhook
          endpoint is secured via SSL using a certificate that makes use of the certificate
          authority for the cluster. Fortunately, OpenShift provides support for generating
          a certificate signed by the cluster. A script is available in the documentation
          called `webhook-create-signed-cert.sh` that automates the process for requesting
          a new certificate.\n\nFirst, create a `Service` for the Operator's webhook
          endpoint:\n```yaml\napiVersion: v1\nkind: Service\nmetadata:\n  labels:\n
          \   name: quay-bridge-operator\n  name: quay-bridge-operator\n  namespace:
          openshift-operators\nspec:\n  ports:\n    - name: https\n      port: 443\n
          \     protocol: TCP\n      targetPort: 8443\n  selector:\n    name: quay-bridge-operator\n
          \ sessionAffinity: None\n  type: ClusterIP\n```\n\nExecute the following
          command to request the certificate:\n\n```\n./webhook-create-signed-cert.sh
          --namespace openshift-operators --secret quay-bridge-operator-webhook-certs
          --service quay-bridge-operator\n```\n\nThe result will place the newly created
          private key and certificate within a secret within the secret specified.
          The secret will be mounted into the appropriate located within the operator
          as declared in the Deployment of the Operator. \n\nOnce the secret has been
          created, focus can shift to the `MutatingWebhookConfiguration` (see below).
          The majority of the content should be in place to apply to the cluster.
          The one section that requires configuration is the caBundle property. This
          refers to the Certificate Authority (CA) for the OpenShift environment.
          The OpenShift CA is available as a _ConfigMap_ within the `kube-system`
          namespace.\n\nExecute the following command to retrieve the CA and format
          the result as a single line so that it can be entered into the MutatingWebhookConfiguration
          resource:\n\n```\noc get configmap -n kube-system extension-apiserver-authentication
          -o=jsonpath='{.data.client-ca-file}' | base64 | tr -d '\\n'\n```\n\nReplace
          the `${CA_BUNDLE}` variable in the following `MutatingWebhookConfiguration`
          YAML and create it:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\nkind:
          MutatingWebhookConfiguration \nmetadata:\n  name: quay-bridge-operator\nwebhooks:\n
          \ - name: quayintegration.redhatcop.redhat.io\n    clientConfig: \n      service:\n
          \       namespace: openshift-operators\n        name: quay-bridge-operator\n
          \       path: \"/admissionwebhook\"\n      caBundle: \"${CA_BUNDLE}\"\n
          \   rules: \n    - operations:  [ \"CREATE\" ]\n      apiGroups: [ \"build.openshift.io\"
          ]\n      apiVersions: [\"v1\" ]\n      resources: [ \"builds\" ]\n    failurePolicy:
          Fail\n```\n\n_Note: Until the operator is running, new requests for builds
          will fail since the webserver the MutatingWebhookConfiguration invokes is
          not available and a proper is response is required in order for the object
          to be persisted in etcd._\n\n#### Create the QuayIntegration Custom Resource\n\nFinally,
          to complete the integration between OpenShift and Quay, a `QuayIntegration`
          custom resource needs to be created.\n\nThe following is an example of a
          basic definition of a `QuayIntegration` resource associated from the associated
          CRD.\n\n```\napiVersion: redhatcop.redhat.io/v1alpha1\nkind: QuayIntegration\nmetadata:\n
          \ name: example-quayintegration\nspec:\n  clusterID: openshift\n  credentialsSecretName:
          <NAMESPACE>/<SECRET>\n  quayHostname: https://<QUAY_URL>\n```\n\nThe _clusterID_
          is a value which should be unique across the entire ecosystem. This value
          is optional and defaults to `openshift`.\n\nThe _credentialsSecretName_
          is a NamespacedName value of the secret containing the token that was previously
          created.\n\nNote: If Quay is using self signed certificates, the property
          `insecureRegistry: true`\n\nOrganizations within Quay should be created
          for the related namespaces in OpenShift."
        displayName: Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Source Code
          url: https://github.com/redhat-cop/quay-openshift-registry-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:501fe9b003b5066b3936def0ef4de05abd1a12e795e718d6c9a62d2c6c804c2e
        version: 3.4.7
      entries:
      - name: quay-bridge-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: quay-bridge-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "my-quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecretName": "<NAMESPACE>/<SECRET>",
                  "quayHostname": "https://<QUAY_URL>",
                  "insecureRegistry": false,
                  "whitelistNamespaces": ["default"]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:9149df5ffc83dce6593072b7f72805e07e700b84527e45b4b7486b9c38f70731
          createdAt: 2021-09-15 23:09 UTC
          description: Enhance OCP using Red Hat Quay container registry
          olm.skipRange: '>=3.4.x <3.5.7'
          repository: https://github.com/redhat-cop/quay-openshift-registry-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a connected Quay instance.
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.redhatcop.redhat.io
            version: v1alpha1
        description: "## Overview\n\nThe OpenShift Container Platform contains an
          in cluster image registry for pushing and pulling images that either produced
          within the cluster or are references from external sources. Quay is a highly
          available, enterprise grade image registry with enhanced role based access
          control (RBAC) features. The goal of this operator is to support replicating
          the features found within OpenShift intgrated image registry with Quay.\n\n##
          Functionality\n\nThe following capabilities are enabled through this operator:\n\n*
          Synchronization of OpenShift namespaces as Quay organizations\n    * Creation
          of Robot accounts for each default namespace service account\n    * Creation
          of Secrets for each created Robot Account\n        * Associate Robot Secret
          to Service Account as Mountable and Image Pull Secret\n    * Synchronization
          of OpenShift ImageStreams as Quay Repositories\n* Automatically rewrite
          new Builds making use of ImageStreams to output to Quay\n* Automatically
          import ImageStream tag once build completes\n\n## Prerequisites\n\nThe following
          requirements must be satisfied prior to setting and configuring the integration
          between OpenShift and Quay:\n\n* A Red Hat Quay environment with a user
          with superuser permissions\n* A Red Hat OpenShift Container Platform environment
          (at least version 4.0) with cluster-admin permissions on an OpenShift cluster
          \n*\tOpenShift Command Line tool (CLI)\n\n## Setup and Configuration\n\nThe
          setup process involves steps within both Quay and OpenShift as described
          below:\n\n### Quay Setup\n\nThe integration between OpenShift and Quay makes
          extensive use of the Quay RESTFul API. To allow for other methods aside
          from user credentials to be used for authenticating against the API, an
          application is required to be created. When an application is created, an
          OAuth token is generated which can be used by the operator in OpenShift.
          Applications are created within a Quay organization. While an existing organization
          can be utilized, it is recommended that a dedicated organization be used.\n\n1.\tLogin
          to Quay as a user with superuser access and select the organization for
          which the external application will be configured within.\n2.\tOn the left
          hand side of the page, select **Applications**\n3.\tCreate a new application
          by selecting the **Create New Application** button and entering a name for
          the application\n4.\tAfter the new application has been created, select
          the application that was created\n5.\tOn the lefthand navigation bar, select
          the **Generate Token** button to create a new OAuth2 token\n6.\tSelect all
          of the checkboxes to grant the necessary access that is needed for the integration\n7.\tReview
          the permissions that will be assigned and then select the **Authorize Application**
          button\n8.\tTake note of the generated Access Token as it will be needed
          in the following section\n\n### OpenShift Setup\n\nTo complete the process
          to begin to integrate OpenShift and Quay, several steps are required. Before
          beginning, ensure that you have the OpenShift Command Line tool installed
          and that you are logged into OpenShift as a cluster administrator. \n\n####
          Create A Secret for the Quay OAuth Token\n\nThe Operator will use the previously
          obtained Access Token to communicate with Quay. Store this token within
          OpenShift as a secret.\n\nExecute the following command to create a secret
          called `quay-integration` with a key called `token` containing the access
          token:\n\n```\n$ oc create secret generic quay-integration --from-literal=token=<access_token>\n```\n\nThis
          token will be referenced in the final section.\n\n#### MutatingWebhookConfiguration
          Support\n\nSupport for dynamic interception of API requests that are performed
          as part of OpenShift’s typical build process is facilitated through a [MutatingWebhookConfiguration](https://docs.openshift.com/container-platform/3.11/architecture/additional_concepts/dynamic_admission_controllers.html).
          A MutatingWebhookConfiguration allows for invoking an API running within
          a project on OpenShift when certain API requests are received. In particular,
          to support Quay’s integration with OpenShift, any new Build requests should
          be intercepted so that the output can be modified to target Quay instead
          of OpenShift’s integrated registry. \n\nKubernetes requires that the webhook
          endpoint is secured via SSL using a certificate that makes use of the certificate
          authority for the cluster. Fortunately, OpenShift provides support for generating
          a certificate signed by the cluster. A script is available in the documentation
          called `webhook-create-signed-cert.sh` that automates the process for requesting
          a new certificate.\n\nFirst, create a `Service` for the Operator's webhook
          endpoint:\n```yaml\napiVersion: v1\nkind: Service\nmetadata:\n  labels:\n
          \   name: quay-bridge-operator\n  name: quay-bridge-operator\n  namespace:
          openshift-operators\nspec:\n  ports:\n    - name: https\n      port: 443\n
          \     protocol: TCP\n      targetPort: 8443\n  selector:\n    name: quay-bridge-operator\n
          \ sessionAffinity: None\n  type: ClusterIP\n```\n\nExecute the following
          command to request the certificate:\n\n```\n./webhook-create-signed-cert.sh
          --namespace openshift-operators --secret quay-bridge-operator-webhook-certs
          --service quay-bridge-operator\n```\n\nThe result will place the newly created
          private key and certificate within a secret within the secret specified.
          The secret will be mounted into the appropriate located within the operator
          as declared in the Deployment of the Operator. \n\nOnce the secret has been
          created, focus can shift to the `MutatingWebhookConfiguration` (see below).
          The majority of the content should be in place to apply to the cluster.
          The one section that requires configuration is the caBundle property. This
          refers to the Certificate Authority (CA) for the OpenShift environment.
          The OpenShift CA is available as a _ConfigMap_ within the `kube-system`
          namespace.\n\nExecute the following command to retrieve the CA and format
          the result as a single line so that it can be entered into the MutatingWebhookConfiguration
          resource:\n\n```\noc get configmap -n kube-system extension-apiserver-authentication
          -o=jsonpath='{.data.client-ca-file}' | base64 | tr -d '\\n'\n```\n\nReplace
          the `${CA_BUNDLE}` variable in the following `MutatingWebhookConfiguration`
          YAML and create it:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\nkind:
          MutatingWebhookConfiguration \nmetadata:\n  name: quay-bridge-operator\nwebhooks:\n
          \ - name: quayintegration.redhatcop.redhat.io\n    clientConfig: \n      service:\n
          \       namespace: openshift-operators\n        name: quay-bridge-operator\n
          \       path: \"/admissionwebhook\"\n      caBundle: \"${CA_BUNDLE}\"\n
          \   rules: \n    - operations:  [ \"CREATE\" ]\n      apiGroups: [ \"build.openshift.io\"
          ]\n      apiVersions: [\"v1\" ]\n      resources: [ \"builds\" ]\n    failurePolicy:
          Fail\n    sideEffects: None\n    admissionReviewVersions: [v1beta1]\n    matchPolicy:
          Exact\n    timeoutSeconds: 30  \n```\n\n_Note: Until the operator is running,
          new requests for builds will fail since the webserver the MutatingWebhookConfiguration
          invokes is not available and a proper is response is required in order for
          the object to be persisted in etcd._\n\n#### Create the QuayIntegration
          Custom Resource\n\nFinally, to complete the integration between OpenShift
          and Quay, a `QuayIntegration` custom resource needs to be created.\n\nThe
          following is an example of a basic definition of a `QuayIntegration` resource
          associated from the associated CRD.\n\n```\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          QuayIntegration\nmetadata:\n  name: example-quayintegration\nspec:\n  clusterID:
          openshift\n  credentialsSecretName: <NAMESPACE>/<SECRET>\n  quayHostname:
          https://<QUAY_URL>\n```\n\nThe _clusterID_ is a value which should be unique
          across the entire ecosystem. This value is optional and defaults to `openshift`.\n\nThe
          _credentialsSecretName_ is a NamespacedName value of the secret containing
          the token that was previously created.\n\nNote: If Quay is using self signed
          certificates, the property `insecureRegistry: true`\n\nOrganizations within
          Quay should be created for the related namespaces in OpenShift."
        displayName: Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Source Code
          url: https://github.com/redhat-cop/quay-openshift-registry-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:9149df5ffc83dce6593072b7f72805e07e700b84527e45b4b7486b9c38f70731
        version: 3.5.7
      entries:
      - name: quay-bridge-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: quay-bridge-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:33dc7673155f01728225221768f6b90d87cc7609dcda46b394dba2421c136a23
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.10.15
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:33dc7673155f01728225221768f6b90d87cc7609dcda46b394dba2421c136a23
        version: 3.10.15
      entries:
      - name: quay-bridge-operator.v3.10.15
        version: 3.10.15
      - name: quay-bridge-operator.v3.10.14
        version: 3.10.14
      - name: quay-bridge-operator.v3.10.13
        version: 3.10.13
      - name: quay-bridge-operator.v3.10.12
        version: 3.10.12
      - name: quay-bridge-operator.v3.10.11
        version: 3.10.11
      - name: quay-bridge-operator.v3.10.10
        version: 3.10.10
      - name: quay-bridge-operator.v3.10.9
        version: 3.10.9
      - name: quay-bridge-operator.v3.10.8
        version: 3.10.8
      - name: quay-bridge-operator.v3.10.7
        version: 3.10.7
      - name: quay-bridge-operator.v3.10.6
        version: 3.10.6
      - name: quay-bridge-operator.v3.10.5
        version: 3.10.5
      - name: quay-bridge-operator.v3.10.4
        version: 3.10.4
      - name: quay-bridge-operator.v3.10.3
        version: 3.10.3
      - name: quay-bridge-operator.v3.10.2
        version: 3.10.2
      - name: quay-bridge-operator.v3.10.1
        version: 3.10.1
      - name: quay-bridge-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-bridge-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:511f8471db70e8d5fd9d0b547c73f1416bbac6f0fe6cdeb6c59a9ef8bd845ac2
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.11.13
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:511f8471db70e8d5fd9d0b547c73f1416bbac6f0fe6cdeb6c59a9ef8bd845ac2
        version: 3.11.13
      entries:
      - name: quay-bridge-operator.v3.11.13
        version: 3.11.13
      - name: quay-bridge-operator.v3.11.12
        version: 3.11.12
      - name: quay-bridge-operator.v3.11.11
        version: 3.11.11
      - name: quay-bridge-operator.v3.11.10
        version: 3.11.10
      - name: quay-bridge-operator.v3.11.9
        version: 3.11.9
      - name: quay-bridge-operator.v3.11.8
        version: 3.11.8
      - name: quay-bridge-operator.v3.11.7
        version: 3.11.7
      - name: quay-bridge-operator.v3.11.6
        version: 3.11.6
      - name: quay-bridge-operator.v3.11.5
        version: 3.11.5
      - name: quay-bridge-operator.v3.11.4
        version: 3.11.4
      - name: quay-bridge-operator.v3.11.3
        version: 3.11.3
      - name: quay-bridge-operator.v3.11.2
        version: 3.11.2
      - name: quay-bridge-operator.v3.11.1
        version: 3.11.1
      - name: quay-bridge-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-bridge-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:c6e04de3ad3b8c0ded95146d090e120bffed7f9dc5e74332a183fe1c2e45d16d
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.12.12
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:c6e04de3ad3b8c0ded95146d090e120bffed7f9dc5e74332a183fe1c2e45d16d
        version: 3.12.12
      entries:
      - name: quay-bridge-operator.v3.12.12
        version: 3.12.12
      - name: quay-bridge-operator.v3.12.11
        version: 3.12.11
      - name: quay-bridge-operator.v3.12.10
        version: 3.12.10
      - name: quay-bridge-operator.v3.12.9
        version: 3.12.9
      - name: quay-bridge-operator.v3.12.8
        version: 3.12.8
      - name: quay-bridge-operator.v3.12.7
        version: 3.12.7
      - name: quay-bridge-operator.v3.12.6
        version: 3.12.6
      - name: quay-bridge-operator.v3.12.5
        version: 3.12.5
      - name: quay-bridge-operator.v3.12.4
        version: 3.12.4
      - name: quay-bridge-operator.v3.12.3
        version: 3.12.3
      - name: quay-bridge-operator.v3.12.2
        version: 3.12.2
      - name: quay-bridge-operator.v3.12.1
        version: 3.12.1
      - name: quay-bridge-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-bridge-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:e8ee1944d8e0e9d515a6e63a3e39901c304abd421fcad2abbf9a3c43c7ca6243
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.13.8
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:e8ee1944d8e0e9d515a6e63a3e39901c304abd421fcad2abbf9a3c43c7ca6243
        version: 3.13.8
      entries:
      - name: quay-bridge-operator.v3.13.8
        version: 3.13.8
      - name: quay-bridge-operator.v3.13.7
        version: 3.13.7
      - name: quay-bridge-operator.v3.13.6
        version: 3.13.6
      - name: quay-bridge-operator.v3.13.5
        version: 3.13.5
      - name: quay-bridge-operator.v3.13.4
        version: 3.13.4
      - name: quay-bridge-operator.v3.13.3
        version: 3.13.3
      - name: quay-bridge-operator.v3.13.2
        version: 3.13.2
      - name: quay-bridge-operator.v3.13.1
        version: 3.13.1
      - name: quay-bridge-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-bridge-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b12f0dfb8fdeae3f521e41fcf60ef118bdfd08ad7f68bd87c5494b067b7c7c58
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.14.5
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b12f0dfb8fdeae3f521e41fcf60ef118bdfd08ad7f68bd87c5494b067b7c7c58
        version: 3.14.5
      entries:
      - name: quay-bridge-operator.v3.14.5
        version: 3.14.5
      - name: quay-bridge-operator.v3.14.4
        version: 3.14.4
      - name: quay-bridge-operator.v3.14.3
        version: 3.14.3
      - name: quay-bridge-operator.v3.14.2
        version: 3.14.2
      - name: quay-bridge-operator.v3.14.1
        version: 3.14.1
      - name: quay-bridge-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-bridge-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b2ac4a48c80209d7bebe8b1b2d066571f99a9f47b78604e47a8300d027941b66
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.15.2
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b2ac4a48c80209d7bebe8b1b2d066571f99a9f47b78604e47a8300d027941b66
        version: 3.15.2
      entries:
      - name: quay-bridge-operator.v3.15.2
        version: 3.15.2
      - name: quay-bridge-operator.v3.15.1
        version: 3.15.1
      - name: quay-bridge-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-bridge-operator.v3.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:0ad7af440c6e00fb578ef40e6d600964324fd94ce7835a28d0ec355b7415093f
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.6.10
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:99dc6b5af811bb33892c17334dd4e1a7eddc6982d49a793b1789f505e00f49b1
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:0ad7af440c6e00fb578ef40e6d600964324fd94ce7835a28d0ec355b7415093f
        version: 3.6.10
      entries:
      - name: quay-bridge-operator.v3.6.10
        version: 3.6.10
      - name: quay-bridge-operator.v3.6.9
        version: 3.6.9
      - name: quay-bridge-operator.v3.6.8
        version: 3.6.8
      - name: quay-bridge-operator.v3.6.7
        version: 3.6.7
      - name: quay-bridge-operator.v3.6.6
        version: 3.6.6
      - name: quay-bridge-operator.v3.6.5
        version: 3.6.5
      - name: quay-bridge-operator.v3.6.4
        version: 3.6.4
      - name: quay-bridge-operator.v3.6.2
        version: 3.6.2
      - name: quay-bridge-operator.v3.6.1
        version: 3.6.1
      - name: quay-bridge-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: quay-bridge-operator.v3.7.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:6a0dd80dc48ad0aaa576b3de78b2b265e29a3c10e4996aa6b8f129592fe20f09
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.7.14
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:975f287513039b2244894b9031e590f665d48f9af480dfdc4b85d4f96dd4bc37
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:6a0dd80dc48ad0aaa576b3de78b2b265e29a3c10e4996aa6b8f129592fe20f09
        version: 3.7.14
      entries:
      - name: quay-bridge-operator.v3.7.14
        version: 3.7.14
      - name: quay-bridge-operator.v3.7.13
        version: 3.7.13
      - name: quay-bridge-operator.v3.7.12
        version: 3.7.12
      - name: quay-bridge-operator.v3.7.11
        version: 3.7.11
      - name: quay-bridge-operator.v3.7.10
        version: 3.7.10
      - name: quay-bridge-operator.v3.7.9
        version: 3.7.9
      - name: quay-bridge-operator.v3.7.8
        version: 3.7.8
      - name: quay-bridge-operator.v3.7.7
        version: 3.7.7
      - name: quay-bridge-operator.v3.7.6
        version: 3.7.6
      - name: quay-bridge-operator.v3.7.5
        version: 3.7.5
      - name: quay-bridge-operator.v3.7.4
        version: 3.7.4
      - name: quay-bridge-operator.v3.7.3
        version: 3.7.3
      - name: quay-bridge-operator.v3.7.2
        version: 3.7.2
      - name: quay-bridge-operator.v3.7.1
        version: 3.7.1
      - name: quay-bridge-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: quay-bridge-operator.v3.8.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:f8544a546dd6d209355c3a1cbe35138f6910491fc4584b401634431da4414ec0
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.8.15
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:975f287513039b2244894b9031e590f665d48f9af480dfdc4b85d4f96dd4bc37
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:f8544a546dd6d209355c3a1cbe35138f6910491fc4584b401634431da4414ec0
        version: 3.8.15
      entries:
      - name: quay-bridge-operator.v3.8.15
        version: 3.8.15
      - name: quay-bridge-operator.v3.8.14
        version: 3.8.14
      - name: quay-bridge-operator.v3.8.13
        version: 3.8.13
      - name: quay-bridge-operator.v3.8.12
        version: 3.8.12
      - name: quay-bridge-operator.v3.8.11
        version: 3.8.11
      - name: quay-bridge-operator.v3.8.10
        version: 3.8.10
      - name: quay-bridge-operator.v3.8.9
        version: 3.8.9
      - name: quay-bridge-operator.v3.8.8
        version: 3.8.8
      - name: quay-bridge-operator.v3.8.7
        version: 3.8.7
      - name: quay-bridge-operator.v3.8.6
        version: 3.8.6
      - name: quay-bridge-operator.v3.8.5
        version: 3.8.5
      - name: quay-bridge-operator.v3.8.4
        version: 3.8.4
      - name: quay-bridge-operator.v3.8.3
        version: 3.8.3
      - name: quay-bridge-operator.v3.8.2
        version: 3.8.2
      - name: quay-bridge-operator.v3.8.1
        version: 3.8.1
      - name: quay-bridge-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-bridge-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:875e0b63fdb7d44879650eb2f788ea705d753520841a5d875428f0b3ca70a36c
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.9.16
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:875e0b63fdb7d44879650eb2f788ea705d753520841a5d875428f0b3ca70a36c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 3.9.16
      entries:
      - name: quay-bridge-operator.v3.9.16
        version: 3.9.16
      - name: quay-bridge-operator.v3.9.15
        version: 3.9.15
      - name: quay-bridge-operator.v3.9.14
        version: 3.9.14
      - name: quay-bridge-operator.v3.9.13
        version: 3.9.13
      - name: quay-bridge-operator.v3.9.12
        version: 3.9.12
      - name: quay-bridge-operator.v3.9.11
        version: 3.9.11
      - name: quay-bridge-operator.v3.9.10
        version: 3.9.10
      - name: quay-bridge-operator.v3.9.9
        version: 3.9.9
      - name: quay-bridge-operator.v3.9.8
        version: 3.9.8
      - name: quay-bridge-operator.v3.9.7
        version: 3.9.7
      - name: quay-bridge-operator.v3.9.6
        version: 3.9.6
      - name: quay-bridge-operator.v3.9.5
        version: 3.9.5
      - name: quay-bridge-operator.v3.9.4
        version: 3.9.4
      - name: quay-bridge-operator.v3.9.3
        version: 3.9.3
      - name: quay-bridge-operator.v3.9.2
        version: 3.9.2
      - name: quay-bridge-operator.v3.9.1
        version: 3.9.1
      - name: quay-bridge-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: quay-bridge-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/securesign/secure-sign-operator
    name: rhtas-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhtas-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": {
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname-intermediate",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname-leaf",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname-root",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname-intermediate",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname-leaf",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname-root",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": {
                  "database": {
                    "create": true
                  },
                  "server": {
                    "replicas": 1
                  },
                  "signer": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:0ebc6b3bd992309a914cc0da8ad8fc628224f583b66f1ca2cd10d2281b781cde
          createdAt: "2025-09-30T11:33:48Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:e7f5a9a2ea76fdfde34f6aac264b51c596cf4ab53f6922b58514c77a1977677e
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:f2798c21dbcb3334b256bf1370a42c013f8f3eefb0641e1c2f250428a82ddb82
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:2dd696771c0188a119d746711551b1643a14a579d0fc66d70918cca9de0c0107
        - registry.redhat.io/rhtas/createtree-rhel9@sha256:244758721aa1c63adfffbae31df58059ff8a06311f6cb60515eeae14b20e78ce
        - registry.redhat.io/rhtas/rekor-monitor-rhel9@sha256:da3aa5c441653f00c4b558ef6ebf21fef518731e70f152a1fd3a750b1145d901
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:b311fb891203fef071d53c9dab67a45961638af98bdfd8e02026dc2dbc25fd00
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:8f8a5d4abf7878ba34151f8632aa7db25218300dbb9f17b121c5f61e0e5c5e03
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:eaae082559c81c5e1253f1b5cf9e298b14235470c67e9111c18ece46078228d5
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:2d91dceb216d8cb8dacefcfe86f357f4ff4eb5e5e998f264d1ad231821a92fad
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:043103f7c888d150053b873276b92b2ee1907c8a7238652ad0770e36d44c10e2
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:0ebc6b3bd992309a914cc0da8ad8fc628224f583b66f1ca2cd10d2281b781cde
        - registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:08302a0af012c1a1f04ef15de86558c9eef55620ab7b179fbda170025f0584be
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:497203991dcbca71275d691ec970ba480cb8e80c9ada6af876f5559abe0d6155
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:110509f5f24240d79afff7cb0d8a4ead0ed038c6100df8e435b01f270b7d83b2
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:abaa247777eb05bf8c97ba1ffc30c00d0d84590fac2fdfbffa06ba374b924888
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:fcb225019f865dd4b7af1f5134a9aed9c5c8a425f3ee43b74e0517ce602e209c
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:e6a5bf7804f1c4371e4de06805d6e1ddfc98e356e355ac3b55a3be25c5636b4d
        version: 1.3.0
      entries:
      - name: rhtas-operator.v1.3.0
        version: 1.3.0
      - name: rhtas-operator.v1.2.1
        version: 1.2.1
      - name: rhtas-operator.v1.2.0
        version: 1.2.0
      - name: rhtas-operator.v1.1.2
        version: 1.1.2
      - name: rhtas-operator.v1.1.1
        version: 1.1.1
      - name: rhtas-operator.v1.1.0
        version: 1.1.0
      - name: rhtas-operator.v1.0.2
        version: 1.0.2
      - name: rhtas-operator.v1.0.1
        version: 1.0.1
      - name: rhtas-operator.v1.0.0
        version: 1.0.0
      name: stable
    - currentCSV: rhtas-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e1a78636cd7cc8e7227b6b0a5ddfd93d002db0464b15e966e88337d38989ea5a
          createdAt: "2024-06-27T09:45:39Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e1a78636cd7cc8e7227b6b0a5ddfd93d002db0464b15e966e88337d38989ea5a
        version: 1.0.2
      entries:
      - name: rhtas-operator.v1.0.2
        version: 1.0.2
      - name: rhtas-operator.v1.0.1
        version: 1.0.1
      - name: rhtas-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: rhtas-operator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e52d07e2cd8cbdc2ccd0336a4fd29cbb2e5b8a1be628350294a54579cc2636ed
          createdAt: "2025-04-08T09:29:59Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e52d07e2cd8cbdc2ccd0336a4fd29cbb2e5b8a1be628350294a54579cc2636ed
        version: 1.1.2
      entries:
      - name: rhtas-operator.v1.1.2
        version: 1.1.2
      - name: rhtas-operator.v1.1.1
        version: 1.1.1
      - name: rhtas-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: rhtas-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
          createdAt: "2025-08-17T21:52:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:02fbd4772c185edf38fcee4e5c5e1c9dad82b77c9da824addaebc445b67d35ac
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
        - registry.redhat.io/ubi9/httpd-24@sha256:badeee0aa3477e61de64f6a2b93e8a3673edd38fc860125f7536506f3d592132
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:c22cdb9a17f85e13c00412cba83aa644e90a0faea0236029d0f1c53fe212e227
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:12b438ea70f83ee6edb12c96b507f92510f4afba824eac34c8f50793c4199db6
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:fc018a45b7eb48690c46c83f52cac55fa40089e6b02686bc21265dc7c2205b8f
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:1d782a127d2f8bdab633f287cb2df2e400ae91bf5d4b913473539d8512dfaab6
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:09b4aeeb607c88c72f69e6f87cb840c82ef5752c64a58fa13551db749fac2530
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:eb38e98dbb9828fe033a5388609132f7246fe77d2f4d258d015a94ea30752b24
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:6270ba29adf779de5d3916d5e82f1cd9491ba89d4ae2042347b6ad6d1819dd26
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:1c2201d50469d70ec6e21546c7c74bd52251d420dc1dcfa5375c1cf61dd3a9fd
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:030986ea26f33db3a192c67a93cd2bde352f23c68a17b062fc062955675c3c51
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:531603563c69aac6905cd6b8157d30781040527404ac2313acba5a9310e0b713
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:fe31830df2d99359749cb86cd1ac2f32fcc3de07aae427d2f2cd3998d31d2833
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:a93df3206409de75613d398e7dcd8b30ff7cc36eb00349c435784f73f00a2335
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:640633823a54c11fe2a8b3d05e571399fda415b6f0b3adf1ba9806882a94bdd4
        version: 1.2.1
      entries:
      - name: rhtas-operator.v1.2.1
        version: 1.2.1
      - name: rhtas-operator.v1.2.0
        version: 1.2.0
      name: stable-v1.2
    - currentCSV: rhtas-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": {
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname-intermediate",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname-leaf",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname-root",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname-intermediate",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname-leaf",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname-root",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": {
                  "database": {
                    "create": true
                  },
                  "server": {
                    "replicas": 1
                  },
                  "signer": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:0ebc6b3bd992309a914cc0da8ad8fc628224f583b66f1ca2cd10d2281b781cde
          createdAt: "2025-09-30T11:33:48Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:abaa247777eb05bf8c97ba1ffc30c00d0d84590fac2fdfbffa06ba374b924888
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:eaae082559c81c5e1253f1b5cf9e298b14235470c67e9111c18ece46078228d5
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:2d91dceb216d8cb8dacefcfe86f357f4ff4eb5e5e998f264d1ad231821a92fad
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:e6a5bf7804f1c4371e4de06805d6e1ddfc98e356e355ac3b55a3be25c5636b4d
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:8f8a5d4abf7878ba34151f8632aa7db25218300dbb9f17b121c5f61e0e5c5e03
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:2dd696771c0188a119d746711551b1643a14a579d0fc66d70918cca9de0c0107
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:110509f5f24240d79afff7cb0d8a4ead0ed038c6100df8e435b01f270b7d83b2
        - registry.redhat.io/rhtas/createtree-rhel9@sha256:244758721aa1c63adfffbae31df58059ff8a06311f6cb60515eeae14b20e78ce
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:b311fb891203fef071d53c9dab67a45961638af98bdfd8e02026dc2dbc25fd00
        - registry.redhat.io/rhtas/rekor-monitor-rhel9@sha256:da3aa5c441653f00c4b558ef6ebf21fef518731e70f152a1fd3a750b1145d901
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:08302a0af012c1a1f04ef15de86558c9eef55620ab7b179fbda170025f0584be
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:043103f7c888d150053b873276b92b2ee1907c8a7238652ad0770e36d44c10e2
        - registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:e7f5a9a2ea76fdfde34f6aac264b51c596cf4ab53f6922b58514c77a1977677e
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:fcb225019f865dd4b7af1f5134a9aed9c5c8a425f3ee43b74e0517ce602e209c
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:0ebc6b3bd992309a914cc0da8ad8fc628224f583b66f1ca2cd10d2281b781cde
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:497203991dcbca71275d691ec970ba480cb8e80c9ada6af876f5559abe0d6155
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:f2798c21dbcb3334b256bf1370a42c013f8f3eefb0641e1c2f250428a82ddb82
        version: 1.3.0
      entries:
      - name: rhtas-operator.v1.3.0
        version: 1.3.0
      name: stable-v1.3
    defaultChannel: stable
    packageName: rhtas-operator
    provider:
      name: Red Hat
      url: https://github.com/securesign/secure-sign-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: sailoperator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sailoperator.v1.28.0-nightly-2025-10-06
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:07ea72de6ff9184f346ee2948228e73cfe4180c8d05fe7ee972c22378ad450ef
          createdAt: "2025-10-06T03:29:39Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0
          - master
          - v1.28-alpha.8de61b9a

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-testing/ztunnel@sha256:38b5902bf36f5163274ce0963dfb1e106000ac639d429f05846448de8454492d
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - quay.io/sail-dev/sail-operator@sha256:07ea72de6ff9184f346ee2948228e73cfe4180c8d05fe7ee972c22378ad450ef
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - gcr.io/istio-testing/install-cni@sha256:9525a3711201bcca04ce2ca9b1576d4094a03779d60db0a36b15a956020bbec1
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-testing/proxyv2@sha256:11cf6a3083a3c1ffa539a21cee30ddc0abe47fb44bb16520cdc2f4141dcfee03
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-testing/pilot@sha256:13e2125870f59f752dc5834b702c2dab28f36edc00dccdeeaf872023018e3df0
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        version: 1.28.0-nightly-2025-10-06
      entries:
      - name: sailoperator.v1.28.0-nightly-2025-10-06
        version: 1.28.0-nightly-2025-10-06
      - name: sailoperator.v1.28.0-nightly-2025-10-05
        version: 1.28.0-nightly-2025-10-05
      - name: sailoperator.v1.28.0-nightly-2025-10-04
        version: 1.28.0-nightly-2025-10-04
      - name: sailoperator.v1.28.0-nightly-2025-10-03
        version: 1.28.0-nightly-2025-10-03
      - name: sailoperator.v1.28.0-nightly-2025-10-02
        version: 1.28.0-nightly-2025-10-02
      - name: sailoperator.v1.28.0-nightly-2025-10-01
        version: 1.28.0-nightly-2025-10-01
      - name: sailoperator.v1.28.0-nightly-2025-09-30
        version: 1.28.0-nightly-2025-09-30
      - name: sailoperator.v1.28.0-nightly-2025-09-29
        version: 1.28.0-nightly-2025-09-29
      - name: sailoperator.v1.28.0-nightly-2025-09-28
        version: 1.28.0-nightly-2025-09-28
      - name: sailoperator.v1.28.0-nightly-2025-09-27
        version: 1.28.0-nightly-2025-09-27
      - name: sailoperator.v1.28.0-nightly-2025-09-26
        version: 1.28.0-nightly-2025-09-26
      - name: sailoperator.v1.28.0-nightly-2025-09-25
        version: 1.28.0-nightly-2025-09-25
      - name: sailoperator.v1.28.0-nightly-2025-09-23
        version: 1.28.0-nightly-2025-09-23
      - name: sailoperator.v1.28.0-nightly-2025-09-22
        version: 1.28.0-nightly-2025-09-22
      name: 1.28-nightly
    - currentCSV: sailoperator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.23.0"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.23.0"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/maistra-dev/sail-operator:0.1-latest
          createdAt: "2024-08-20T11:54:30Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1alpha1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1alpha1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1alpha1
          - description: |-
              RemoteIstio represents a remote Istio Service Mesh deployment consisting of one or more
              remote control plane instances (represented by one or more IstioRevision objects).
            displayName: Remote Istio
            kind: RemoteIstio
            name: remoteistios.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.23.0
          - v1.22.4

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - docker.io/istio/proxyv2:1.22.4
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.16.0
        - quay.io/maistra-dev/sail-operator:0.1-latest
        - docker.io/istio/pilot:1.23.0
        - docker.io/istio/proxyv2:1.23.0
        - docker.io/istio/install-cni:1.22.4
        - docker.io/istio/pilot:1.22.4
        - docker.io/istio/ztunnel:1.23.0
        - docker.io/istio/ztunnel:1.22.4
        - docker.io/istio/install-cni:1.23.0
        version: 0.1.0
      entries:
      - name: sailoperator.v0.1.0
        version: 0.1.0
      - name: sailoperator.v0.1.0-rc.1
        version: 0.1.0-rc.1
      name: candidates
    - currentCSV: sailoperator.v1.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
          createdAt: "2025-09-05T10:12:19Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        version: 1.27.1
      entries:
      - name: sailoperator.v1.27.1
        version: 1.27.1
      - name: sailoperator.v1.27.0
        version: 1.27.0
      - name: sailoperator.v1.26.3
        version: 1.26.3
      - name: sailoperator.v1.26.2
        version: 1.26.2
      - name: sailoperator.v1.26.1
        version: 1.26.1
      - name: sailoperator.v1.26.0
        version: 1.26.0
      - name: sailoperator.v1.25.2
        version: 1.25.2
      - name: sailoperator.v1.25.1
        version: 1.25.1
      - name: sailoperator.v1.25.0
        version: 1.25.0
      name: stable
    - currentCSV: sailoperator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.3"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.3"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator:1.0-latest
          createdAt: "2025-03-07T10:17:30Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.24-latest
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.23-latest
          - v1.23.5
          - v1.23.4
          - v1.23.3
          - v1.23.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/sail-dev/sail-operator:1.0-latest
        - gcr.io/istio-release/ztunnel:1.24.1
        - gcr.io/istio-release/proxyv2:1.24.2
        - gcr.io/istio-release/install-cni:1.24.3
        - gcr.io/istio-release/install-cni:1.23.0
        - gcr.io/istio-release/proxyv2:1.23.5
        - gcr.io/istio-release/proxyv2:1.23.4
        - gcr.io/istio-release/pilot:1.24.2
        - gcr.io/istio-release/pilot:1.23.0
        - gcr.io/istio-release/install-cni:1.24.2
        - gcr.io/istio-release/pilot:1.23.4
        - gcr.io/istio-release/proxyv2:1.23.3
        - gcr.io/istio-release/proxyv2:1.23.0
        - gcr.io/istio-release/pilot:1.23.3
        - gcr.io/istio-release/pilot:1.24.1
        - gcr.io/istio-release/pilot:1.23.5
        - gcr.io/istio-release/proxyv2:1.24.3
        - gcr.io/istio-release/ztunnel:1.24.3
        - gcr.io/istio-release/ztunnel:1.23.4
        - gcr.io/istio-release/install-cni:1.23.5
        - gcr.io/istio-release/ztunnel:1.23.3
        - gcr.io/istio-release/install-cni:1.24.1
        - gcr.io/istio-release/proxyv2:1.24.1
        - gcr.io/istio-release/ztunnel:1.23.0
        - gcr.io/istio-release/install-cni:1.23.4
        - gcr.io/istio-release/ztunnel:1.23.5
        - gcr.io/istio-release/ztunnel:1.24.2
        - gcr.io/istio-release/install-cni:1.23.3
        - gcr.io/istio-release/pilot:1.24.3
        version: 1.0.0
      entries:
      - name: sailoperator.v1.0.0
        version: 1.0.0
      name: stable-1.0
    - currentCSV: sailoperator.v1.25.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.25.2"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.25.2"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator:1.25.2
          createdAt: "2025-05-07T19:44:51Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.25.2'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.25-latest
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0
          - v1.23-latest
          - v1.23.5
          - v1.23.4
          - v1.23.3
          - v1.23.2
          - v1.22-latest
          - v1.22.8
          - v1.22.7
          - v1.22.6
          - v1.22.5
          - v1.21.6

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/ztunnel:1.22.6
        - gcr.io/istio-release/pilot:1.22.8
        - gcr.io/istio-release/install-cni:1.23.5
        - quay.io/sail-dev/sail-operator:1.25.2
        - gcr.io/istio-release/pilot:1.22.6
        - gcr.io/istio-release/install-cni:1.24.2
        - gcr.io/istio-release/install-cni:1.24.3
        - gcr.io/istio-release/install-cni:1.23.3
        - gcr.io/istio-release/ztunnel:1.22.5
        - gcr.io/istio-release/proxyv2:1.23.3
        - gcr.io/istio-release/pilot:1.23.2
        - gcr.io/istio-release/install-cni:1.25.1
        - gcr.io/istio-release/pilot:1.24.3
        - gcr.io/istio-release/ztunnel:1.25.1
        - gcr.io/istio-release/install-cni:1.22.8
        - gcr.io/istio-release/proxyv2:1.22.8
        - gcr.io/istio-release/ztunnel:1.21.6
        - gcr.io/istio-release/pilot:1.23.5
        - gcr.io/istio-release/pilot:1.24.1
        - gcr.io/istio-release/proxyv2:1.21.6
        - gcr.io/istio-release/proxyv2:1.25.1
        - gcr.io/istio-release/ztunnel:1.23.4
        - gcr.io/istio-release/proxyv2:1.23.5
        - gcr.io/istio-release/pilot:1.25.2
        - gcr.io/istio-release/install-cni:1.21.6
        - gcr.io/istio-release/ztunnel:1.23.3
        - gcr.io/istio-release/pilot:1.22.5
        - gcr.io/istio-release/proxyv2:1.25.2
        - gcr.io/istio-release/pilot:1.24.4
        - gcr.io/istio-release/pilot:1.23.3
        - gcr.io/istio-release/pilot:1.25.1
        - gcr.io/istio-release/proxyv2:1.23.4
        - gcr.io/istio-release/install-cni:1.22.7
        - gcr.io/istio-release/ztunnel:1.24.3
        - gcr.io/istio-release/ztunnel:1.24.4
        - gcr.io/istio-release/proxyv2:1.22.6
        - gcr.io/istio-release/proxyv2:1.24.2
        - gcr.io/istio-release/ztunnel:1.23.2
        - gcr.io/istio-release/ztunnel:1.24.2
        - gcr.io/istio-release/ztunnel:1.25.2
        - gcr.io/istio-release/install-cni:1.24.4
        - gcr.io/istio-release/pilot:1.22.7
        - gcr.io/istio-release/install-cni:1.25.2
        - gcr.io/istio-release/ztunnel:1.22.7
        - gcr.io/istio-release/proxyv2:1.22.7
        - gcr.io/istio-release/ztunnel:1.22.8
        - gcr.io/istio-release/pilot:1.24.0
        - gcr.io/istio-release/pilot:1.21.6
        - gcr.io/istio-release/install-cni:1.24.1
        - gcr.io/istio-release/install-cni:1.22.5
        - gcr.io/istio-release/install-cni:1.23.2
        - gcr.io/istio-release/install-cni:1.24.0
        - gcr.io/istio-release/ztunnel:1.23.5
        - gcr.io/istio-release/proxyv2:1.24.0
        - gcr.io/istio-release/install-cni:1.22.6
        - gcr.io/istio-release/proxyv2:1.23.2
        - gcr.io/istio-release/proxyv2:1.24.1
        - gcr.io/istio-release/pilot:1.24.2
        - gcr.io/istio-release/proxyv2:1.22.5
        - gcr.io/istio-release/pilot:1.23.4
        - gcr.io/istio-release/ztunnel:1.24.0
        - gcr.io/istio-release/install-cni:1.23.4
        - gcr.io/istio-release/ztunnel:1.24.1
        - gcr.io/istio-release/proxyv2:1.24.4
        - gcr.io/istio-release/proxyv2:1.24.3
        version: 1.25.2
      entries:
      - name: sailoperator.v1.25.2
        version: 1.25.2
      - name: sailoperator.v1.25.1
        version: 1.25.1
      - name: sailoperator.v1.25.0
        version: 1.25.0
      name: stable-1.25
    - currentCSV: sailoperator.v1.26.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.3"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.3"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:d37258646d76a9f708a69fbaa081a0a95908a2b0c8837a14d7ecbfcf505f9bd4
          createdAt: "2025-08-04T06:23:45Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.26-latest
          - v1.26.3
          - v1.26.2
          - v1.26.1
          - v1.26.0
          - v1.25-latest
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/proxyv2@sha256:fd734e6031566b4fb92be38f0f6bb02fdba6c199c45c2db5dc988bbc4fdee026
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/pilot@sha256:c4e33921e030bb22d3b8fabc693851d52c07a7af79a9534addc2f50005fa462a
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - quay.io/sail-dev/sail-operator@sha256:d37258646d76a9f708a69fbaa081a0a95908a2b0c8837a14d7ecbfcf505f9bd4
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/ztunnel@sha256:2d567893e5b452d5c0b9c43a43efd4f8e28e639a344e28d90498f13858f13d46
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/install-cni@sha256:8e782f5077f63c0a15fd4b9a4fa75f8543033c28ccf2457e25afbf1aefb6a72a
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        version: 1.26.3
      entries:
      - name: sailoperator.v1.26.3
        version: 1.26.3
      - name: sailoperator.v1.26.2
        version: 1.26.2
      - name: sailoperator.v1.26.1
        version: 1.26.1
      - name: sailoperator.v1.26.0
        version: 1.26.0
      name: stable-1.26
    - currentCSV: sailoperator.v1.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
          createdAt: "2025-09-05T10:12:19Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        version: 1.27.1
      entries:
      - name: sailoperator.v1.27.1
        version: 1.27.1
      - name: sailoperator.v1.27.0
        version: 1.27.0
      name: stable-1.27
    defaultChannel: stable
    packageName: sailoperator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-websocket-service
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-websocket-service.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "websocket.loadcore.keysight.com/v1alpha1",
                "kind": "Websocket",
                "metadata": {
                  "name": "websocket-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "websocket": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:26:39Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Websocket
            name: websockets.websocket.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Websocket Operator
        displayName: Keysight Websocket Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Websocket Service
          url: https://keysight-websocket-service.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-websocket-service@sha256:c91da3557ddcb8f052c9caa514d2dbf6c22b4963c080117a9d1859f1458fc404
        version: 0.1.21
      entries:
      - name: keysight-websocket-service.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-websocket-service
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/in-en/products/cloud-pak-for-data
    name: ibm-cpd-datastage-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-datastage-operator.v7.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ds.cpd.ibm.com/v1",
                "kind": "DataStage",
                "metadata": {
                  "name": "datastage"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "fileStorageClass": "managed-nfs-storage",
                  "scaleConfig": "small"
                }
              }
            ]
          capabilities: Seamless Upgrades
          cloudpakId: 49a42b864bb94569bef0188ead948f11
          cloudpakName: IBM DataStage Enterprise Plus Cartridge for IBM Cloud Pak
            for Data
          containerImage: registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
          description: The IBM DataStage Operator is used to deploy IBM DataStage
            services
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-datastage-operator
          olm.skipRange: '>=1.0.0 <7.3.0'
          operatorframework.io/arch.amd64: supported
          operatorframework.io/arch.s390x: supported
          operatorframework.io/os.linux: supported
          operatorframework.io/suggested-namespace: ibm-common-services
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation: [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)
              License: [IBM DataStage Enterprise Cartridge for Cloud Pak for Data](https://ibm.biz/Bdahic),
              [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)'
            kind: DataStage
            name: datastages.ds.cpd.ibm.com
            version: v1
          - description: Documentation [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)
              License [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)
            kind: PXRuntime
            name: pxruntimes.ds.cpd.ibm.com
            version: v1
        description: |-
          # Introduction
          Industry-leading batch and real-time data integration service for data engineers to build trusted pipelines (ETL/ELT) for AI and operational use cases across on-premises and hybrid cloud environments.
          With DataStage, you can design and run data flows that move and transform data anywhere, at any scale.
          No matter how complex your data landscape, DataStage can streamline your data movement costs and increase productivity. DataStage offers:
          <ul><li>A best-in-breed parallel processing engine that enables you to process your data where it resides</li><li>Automated job design</li><li>Simple integration with cloud data lakes, real-time data sources, relational databases, big data, and NoSQL data stores</li></ul>
          The DataStage service uses Cloud Pak for Data platform connections and integration points, with services like Data Virtualization, to simplify the process of connecting to and accessing your data.

          With DataStage, Data Engineers can use the simple user interface to build no-code/low-code data pipelines. The interface offers hundreds of functions and connectors that reduce development time and inconsistencies across pipelines. The interface also makes it easy to collaborate with your peers and control access to specific analytics projects.
          The service also provides automatic workload balancing to provide high performance pipelines that make efficient use of available compute resources.

          ## Details
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=SSQNUZ_4.0/svc-welcome/dsent.html)

          ## Documentation

          [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)

          ## Installing
          To install the operator follow the the installation and configuration instructions within the [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=datastage-installing).

          ## Prerequisites

          - OpenShift Version >= 4.6
          - IBM Cloud Pak for Data >= 4.0.2
          - IBM Cloud Pak for Data Common Core Services (CCS) >= 2.0.2 (The DataStage service automatically installs the CCS if they are not already installed.)

          ### Resources Required

          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=requirements-hardware)

          ## Configuration

          To scale DataStage services, you can specify spec.scaleConfig in the DataStage custom resource to small, medium, or large.

          ## License
          [IBM DataStage Enterprise Cartridge for Cloud Pak for Data](https://ibm.biz/Bdahic)
          [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)

          ## Limitations
          - Only the `x86-64` platform is supported by this operator.
          - See [Limitations and known issues in DataStage](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=issues-datastage)

          ## SecurityContextConstraints Requirements
          Custom SecurityContextConstraints definition:
          ```
          apiVersion: security.openshift.io/v1
          kind: SecurityContextConstraints
          allowHostDirVolumePlugin: false
          allowHostIPC: false
          allowHostNetwork: false
          allowHostPID: false
          allowHostPorts: false
          allowPrivilegeEscalation: true
          allowPrivilegedContainer: false
          allowedCapabilities: null
          defaultAddCapabilities: null
          fsGroup:
            type: MustRunAs
          metadata:
            annotations:
              include.release.openshift.io/self-managed-high-availability: "true"
              kubernetes.io/description: restrict denies access to all host features and requires pods to be run with a UID, and SELinux context that are allocated to the namespace.  This is the most restrictive SCC and it is used by default for authenticated users.
              release.openshift.io/create-only: "true"
            name: restrict
          priority: null
          readOnlyRootFilesystem: false
          requiredDropCapabilities:
          - KILL
          - MKNOD
          - SETUID
          - SETGID
          runAsUser:
            type: MustRunAsRange
          seLinuxContext:
            type: MustRunAs
          supplementalGroups:
            type: RunAsAny
          users: []
          volumes:
          - configMap
          - downwardAPI
          - emptyDir
          - persistentVolumeClaim
          - projected
          - secret
          ```
        displayName: IBM DataStage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DataStage
        - Analytics
        - Cloud Pak
        - Data
        - Cloud Pak for Data
        - IBM
        maintainers:
        - email: cp4d@ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/in-en/products/cloud-pak-for-data
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:046eb99c8f1f37dd4cffa1c6c9a93aea50bb5d10a549aa12c3c90134d03c6d1c
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:4604a076c837600cd4b3e29b31e22470915fe5b6b1191ae9401407f8620f2472
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:5f5faf7d15f46d79eea873c12ff631ed59bc55bc57f1f7aee9fd0a8f35fdd3cf
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:6c569c4687546a03a5f4e127acc13b219ce3efa057fbdfd55e96868e549c70cb
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:8b72cafe9de355634bc3218cc1dc701ffb376a29aecb5cd12a5719d43a1304bd
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:3f8133fcd904befa579b99d3a828fe737faf13ae14e842ed0408eeeee72de5fb
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:7015b410b947338dcafde98b68b95cb4199c08511ae46e733691e8fa835ad988
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:3eecc54f26cd87870c201b5ade1d037d2733ccfda40fc1aedd9a2164e582275d
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:001bf35df08cd397397c3fdd1ddbd7953cdf4bde06965427ae640e5608e60ddf
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:1b39e1d5d7c13451252250ceeb29b91c1e7ef3739fc17baf9958e1bf07c4c2fd
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:ef03bb6ed647d92b9c84d34810376da7049aad88589eb2e35798d82a22f5b994
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:10492010b28f57f8badbc7cadbe15d1447269052b387d878c6d9ecbd2f62923b
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:99fc418fb1a09ee6c45e90dfba47f73ea8e1da1b19f9f570d2a749b68118b7a3
        version: 7.3.0
      entries:
      - name: ibm-cpd-datastage-operator.v7.3.0
        version: 7.3.0
      name: v7.3
    defaultChannel: v7.3
    packageName: ibm-cpd-datastage-operator
    provider:
      name: IBM
      url: https://www.ibm.com/in-en/products/cloud-pak-for-data
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Advanced Micro Devices, Inc.
      provider-url: ""
    name: amd-gpu-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amd-gpu-operator.v1.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "amd.com/v1alpha1",
                "kind": "DeviceConfig",
                "metadata": {
                  "name": "test-deviceconfig",
                  "namespace": "openshift-amd-gpu"
                },
                "spec": {
                  "devicePlugin": {
                    "devicePluginImage": "rocm/k8s-device-plugin:rhubi-latest",
                    "nodeLabellerImage": "rocm/k8s-node-labeller:rhubi-latest"
                  },
                  "driver": {
                    "enable": "true",
                    "version": "6.4.2"
                  },
                  "selector": {
                    "feature.node.kubernetes.io/amd-gpu": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning,Monitoring
          containerImage: docker.io/rocm/gpu-operator@sha256:1583d4ec172d50d4416f4030eb85910ec54633b7e8e061d4bd79b14dbf98d28c
          createdAt: "2025-08-01T01:55:48Z"
          description: |-
            Operator responsible for deploying AMD GPU kernel drivers, device plugin, device test runner and device metrics exporter
            For more information, visit [documentation](https://instinct.docs.amd.com/projects/gpu-operator/en/latest/)
          deviceConfigManagerImage: docker.io/rocm/device-config-manager@sha256:d2896271eacfc664ac7651425f46a85af4f4ef98d713988c2a4fee4a846548ee
          devicePluginImage: docker.io/rocm/k8s-device-plugin@sha256:840c4ea1bf58df60ef3e80e3bf27bbf620a393854e42f58b2224b288767e3dd4
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          metricsExporterImage: docker.io/rocm/device-metrics-exporter@sha256:6c946c6ca10d5c0f40e67b5759759f12d51c45e3b934cb1af912821d5fd76e6e
          nodelabellerImage: docker.io/rocm/k8s-node-labeller@sha256:75f916c5a407e40e9ffdd26cdfc65b7da855710568937e740a969094e2991065
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-amd-gpu
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ROCm/gpu-operator
          support: Advanced Micro Devices, Inc.
          testRunnerImage: docker.io/rocm/test-runner@sha256:98e698263b4df26959341b14b3ddded9df51e504e7c6819d5c7c56c434855e9d
          utilsContainerImage: docker.io/rocm/gpu-operator-utils@sha256:9cbe60dce5b3a2ffd93f7148a363ed5f072a320377d5e5bcbbfb1e05d4959287
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DeviceConfig describes how to enable AMD GPU device
            displayName: DeviceConfig
            kind: DeviceConfig
            name: deviceconfigs.amd.com
            version: v1alpha1
        description: |-
          Operator responsible for deploying AMD GPU kernel drivers, device plugin, device test runner and device metrics exporter
          For more information, visit [documentation](https://instinct.docs.amd.com/projects/gpu-operator/en/latest/)
        displayName: amd-gpu-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AMD
        - GPU
        - AI
        - Deep Learning
        - Hardware
        - Driver
        - Monitoring
        links:
        - name: AMD GPU Operator
          url: https://github.com/ROCm/gpu-operator
        maintainers:
        - email: Yan.Sun3@amd.com
          name: Yan Sun
        - email: farshad.ghodsian@amd.com
          name: Farshad Ghodsian
        - email: shrey.ajmera@amd.com
          name: Shrey Ajmera
        maturity: stable
        provider:
          name: Advanced Micro Devices, Inc.
        relatedImages:
        - docker.io/rocm/gpu-operator@sha256:1583d4ec172d50d4416f4030eb85910ec54633b7e8e061d4bd79b14dbf98d28c
        - docker.io/rocm/device-metrics-exporter@sha256:6c946c6ca10d5c0f40e67b5759759f12d51c45e3b934cb1af912821d5fd76e6e
        - docker.io/rocm/k8s-device-plugin@sha256:840c4ea1bf58df60ef3e80e3bf27bbf620a393854e42f58b2224b288767e3dd4
        - docker.io/rocm/k8s-node-labeller@sha256:75f916c5a407e40e9ffdd26cdfc65b7da855710568937e740a969094e2991065
        - docker.io/rocm/gpu-operator-utils@sha256:9cbe60dce5b3a2ffd93f7148a363ed5f072a320377d5e5bcbbfb1e05d4959287
        - docker.io/rocm/test-runner@sha256:98e698263b4df26959341b14b3ddded9df51e504e7c6819d5c7c56c434855e9d
        - docker.io/rocm/device-config-manager@sha256:d2896271eacfc664ac7651425f46a85af4f4ef98d713988c2a4fee4a846548ee
        version: 1.3.2
      entries:
      - name: amd-gpu-operator.v1.3.2
        version: 1.3.2
      - name: amd-gpu-operator.v1.3.1
        version: 1.3.1
      - name: amd-gpu-operator.v1.2.1
        version: 1.2.1
      - name: amd-gpu-operator.v1.1.1
        version: 1.1.1
      name: alpha
    defaultChannel: alpha
    packageName: amd-gpu-operator
    provider:
      name: Advanced Micro Devices, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-zookeeper-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zookeeper-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:1d36865fe1cac88c9edc63c0e35e42a80456486668de5a575f10cc4b445d47db
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:1d36865fe1cac88c9edc63c0e35e42a80456486668de5a575f10cc4b445d47db
        version: 24.11.1
      entries:
      - name: zookeeper-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: zookeeper-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:6a4729289275e473907a3792c001dbe7dd2dc0498aabc719cf4175d956e3c87f
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:6a4729289275e473907a3792c001dbe7dd2dc0498aabc719cf4175d956e3c87f
        version: 25.3.0
      entries:
      - name: zookeeper-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: zookeeper-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
        version: 25.7.0
      entries:
      - name: zookeeper-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: zookeeper-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
        version: 25.7.0
      entries:
      - name: zookeeper-operator.v25.7.0
        version: 25.7.0
      - name: zookeeper-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-zookeeper-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cluster-kube-descheduler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clusterkubedescheduleroperator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "KubeDescheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-kube-descheduler-operator"
                },
                "spec": {
                  "deschedulingIntervalSeconds": 3600,
                  "managementState": "Managed",
                  "mode": "Predictive",
                  "profiles": ["AffinityAndTaints"]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-cluster-kube-descheduler-operator@sha256:ba0b71ff2a30a069b4a8a8f3c1e0898aaadc6db112e4cc12aff7c77ced7a0405
          createdAt: 2021/10/13
          description: An operator to run descheduler in Openshift cluster.
          olm.skipRange: '>=4.10.0 <4.13.0-202309181427'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-kube-descheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeDescheduler is the Schema for the deschedulers API
            displayName: Kube Descheduler
            kind: KubeDescheduler
            name: kubedeschedulers.operator.openshift.io
            version: v1
        description: |
          The Kube Descheduler Operator provides the ability to evict a running pod so that the pod can be rescheduled onto a more suitable node.

          There are several situations where descheduling can benefit your cluster:

          * Nodes are underutilized or overutilized.
          * Pod and node affinity requirements, such as taints or labels, have changed and the original scheduling decisions are no longer appropriate for certain nodes.
          * Node failure requires pods to be moved.
          * New nodes are added to clusters.

          ## Descheduler Profiles

          Once the operator is installed, you can configure one or more profiles to identify pods to evict. The scheduler will schedule the replacement of the evicted pods.

          The following profiles are available:

          * AffinityAndTaints
          * TopologyAndDuplicates
          * SoftTopologyAndDuplicates
          * LifecycleAndUtilization
          * EvictPodsWithPVC
          * EvictPodsWithLocalStorage

          These profiles are documented in detail in the [descheduler operator README](https://github.com/openshift/cluster-kube-descheduler-operator#profiles).

          ## Additional Parameters

          In addition to the profiles, the following parameters can be configured:

          * `deschedulingIntervalSeconds` - Set the number of seconds between descheduler runs. A value of `0` in this field runs the descheduler once and exits.
          * `profileCustomizations` - Allows certain profile parameters to be tweaked, such as `podLifetime` (see [README](https://github.com/openshift/cluster-kube-descheduler-operator#profile-customizations) for more info).
        displayName: Kube Descheduler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - scheduling
        - descheduler
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/cluster-kube-descheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-cluster-kube-descheduler-operator@sha256:ba0b71ff2a30a069b4a8a8f3c1e0898aaadc6db112e4cc12aff7c77ced7a0405
        - registry.redhat.io/openshift4/ose-descheduler@sha256:45dc69ad93ab50bdf9ce1bb79f6d98f849e320db68af30475b10b7f5497a1b13
        version: 4.13.0-202309181427
      entries:
      - name: clusterkubedescheduleroperator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: "4.13"
    - currentCSV: clusterkubedescheduleroperator.v5.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "KubeDescheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-kube-descheduler-operator"
                },
                "spec": {
                  "deschedulingIntervalSeconds": 3600,
                  "managementState": "Managed",
                  "mode": "Predictive",
                  "profiles": ["AffinityAndTaints"]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/kube-descheduler-operator/kube-descheduler-rhel9-operator@sha256:03404d74da227fdf5805f61657272af2c915ce0014d6a902ef384b7a70bd10e2
          createdAt: 2024/10/30
          description: An operator to run descheduler in Openshift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.12.0 <5.0.2'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-kube-descheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeDescheduler is the Schema for the deschedulers API
            displayName: Kube Descheduler
            kind: KubeDescheduler
            name: kubedeschedulers.operator.openshift.io
            version: v1
        description: |
          The Kube Descheduler Operator provides the ability to evict a running pod so that the pod can be rescheduled onto a more suitable node.

          There are several situations where descheduling can benefit your cluster:

          * Nodes are underutilized or overutilized.
          * Pod and node affinity requirements, such as taints or labels, have changed and the original scheduling decisions are no longer appropriate for certain nodes.
          * Node failure requires pods to be moved.
          * New nodes are added to clusters.

          ## Descheduler Profiles

          Once the operator is installed, you can configure one or more profiles to identify pods to evict. The scheduler will schedule the replacement of the evicted pods.

          The following profiles are available:

          * AffinityAndTaints
          * TopologyAndDuplicates
          * SoftTopologyAndDuplicates
          * LifecycleAndUtilization
          * EvictPodsWithPVC
          * EvictPodsWithLocalStorage

          These profiles are documented in detail in the [descheduler operator README](https://github.com/openshift/cluster-kube-descheduler-operator#profiles).

          ## Additional Parameters

          In addition to the profiles, the following parameters can be configured:

          * `deschedulingIntervalSeconds` - Set the number of seconds between descheduler runs. A value of `0` in this field runs the descheduler once and exits.
          * `profileCustomizations` - Allows certain profile parameters to be tweaked, such as `podLifetime` (see [README](https://github.com/openshift/cluster-kube-descheduler-operator#profile-customizations) for more info).
        displayName: Kube Descheduler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - scheduling
        - descheduler
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/cluster-kube-descheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/kube-descheduler-operator/kube-descheduler-rhel9-operator@sha256:03404d74da227fdf5805f61657272af2c915ce0014d6a902ef384b7a70bd10e2
        - registry.redhat.io/kube-descheduler-operator/descheduler-rhel9@sha256:7ba3f46cbbec2e47edcc255fae405bdbf2b9df3a232fc6809ce65fb1827a3b45
        version: 5.0.2
      entries:
      - name: clusterkubedescheduleroperator.v5.0.2
        version: 5.0.2
      - name: clusterkubedescheduleroperator.4.14.0-202311021650
        version: 4.14.0-202311021650
      - name: clusterkubedescheduleroperator.4.14.0-202310201027
        version: 4.14.0-202310201027
      name: stable
    defaultChannel: stable
    packageName: cluster-kube-descheduler-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: machine-deletion-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: machine-deletion-remediation.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:420a4d4a8610a3694fd3ef9809c55a4f140ec8cea7f09219647f82a84d1f411b
          createdAt: "2023-10-30 13:59:28"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          olm.skipRange: '>=0.1.0 <0.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:420a4d4a8610a3694fd3ef9809c55a4f140ec8cea7f09219647f82a84d1f411b
        version: 0.2.1
      entries:
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.14-eus
    - currentCSV: machine-deletion-remediation.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:1445e6df8f49672aa729806f0e257043ecc22e19e666bb5a799286c6b878b1dd
          createdAt: "2024-08-11 06:47:20"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.2.0 <0.3.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f4381aaf002b761a2ba614332a2af973f4de5f6a4c84dd96b8808725e561fa89
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:1445e6df8f49672aa729806f0e257043ecc22e19e666bb5a799286c6b878b1dd
        version: 0.3.1
      entries:
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.16-eus
    - currentCSV: machine-deletion-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-29 09:35:10"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.3.1 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.6/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel9-operator@sha256:dbe9095d494ebc7b56982cb1f74329d7a7fd0cbb7544e9f90207eb35de7997dc
        version: 0.4.1
      entries:
      - name: machine-deletion-remediation.v0.4.1
        version: 0.4.1
      - name: machine-deletion-remediation.v0.4.0
        version: 0.4.0
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.18-eus
    - currentCSV: machine-deletion-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:7e1cad02f1e74a1355c168fc53925bd0fbcc55f2535a46f734c8de082eb49400
          createdAt: "2023-04-24 08:22:17"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3af07bfe0fb73b1a7f00487eba302d6da442e742ab4007fc50e52efd4d45dae5
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:7e1cad02f1e74a1355c168fc53925bd0fbcc55f2535a46f734c8de082eb49400
        version: 0.1.0
      entries:
      - name: machine-deletion-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: machine-deletion-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-29 09:35:10"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.3.1 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.6/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel9-operator@sha256:dbe9095d494ebc7b56982cb1f74329d7a7fd0cbb7544e9f90207eb35de7997dc
        version: 0.4.1
      entries:
      - name: machine-deletion-remediation.v0.4.1
        version: 0.4.1
      - name: machine-deletion-remediation.v0.4.0
        version: 0.4.0
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: stable
    defaultChannel: stable
    packageName: machine-deletion-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Weights & Biases
      provider-url: https://wandb.ai
    name: wandb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: wandb-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.wandb.com/v1",
                "kind": "WeightsAndBiases",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "wandb-operator",
                    "app.kubernetes.io/instance": "weightsandbiases-sample",
                    "app.kubernetes.io/managed-by": "olm",
                    "app.kubernetes.io/name": "weightsandbiases",
                    "app.kubernetes.io/part-of": "wandb-operator"
                  },
                  "name": "weightsandbiases-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning,Developer Tools
          certified: "false"
          containerImage: quay.io/wandb_tools/wandb-k8s-operator:1.0.0
          createdAt: "2024-10-10T14:46:54Z"
          description: Operator for Weights and Biases, enabling seamless integration
            and management of machine learning experiments in Kubernetes environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: wandb-system
          repository: https://github.com/wandb/operator
          support: WeightsAndBiases
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: WeightsAndBiases is a custom resource for managing W&B configurations.
            displayName: Weights And Biases
            kind: WeightsAndBiases
            name: weightsandbiases.apps.wandb.com
            version: v1
        description: |-
          The Weights and Biases Operator simplifies the management of machine learning experiments, tracking, and infrastructure. It integrates with Kubernetes to deploy and manage Weights and Biases instances.
          Features:
          - Experiment Tracking
          - Hyperparameter Tuning
          - Model Management
          - Integration with popular ML frameworks
        displayName: Weights & Biases Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - machine learning
        - experiment tracking
        - ML infrastructure
        - weights and biases
        links:
        - name: Weights & Biases Documentation
          url: https://docs.wandb.ai/
        maintainers:
        - email: abhinav.garg@domain.com
          name: Abhinav Garg
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Weights & Biases
          url: https://wandb.ai
        relatedImages:
        - quay.io/wandb_tools/wandb-k8s-operator:1.0.0
        version: 1.0.0
      entries:
      - name: wandb-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: wandb-operator
    provider:
      name: Weights & Biases
      url: https://wandb.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: keepalived-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keepalived-operator.v1.5.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KeepalivedGroup",
                "metadata": {
                  "name": "keepalivedgroup-workers"
                },
                "spec": {
                  "blacklistRouterIDs": [
                    1,
                    2
                  ],
                  "image": "registry.redhat.io/openshift4/ose-keepalived-ipfailover",
                  "interface": "ens3",
                  "nodeSelector": {
                    "node-role.kubernetes.io/loadbalancer": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Networking, Cloud Provider
          certified: "false"
          containerImage: quay.io/redhat-cop/keepalived-operator@sha256:cd0c2f7df82df1178795e4a18a0942da33adb9670b192a7b7a2b1483a5b4ccf8
          createdAt: "2023-03-11T18:48:22Z"
          description: This operator provides a simple way to create self-hosted load
            balancers using keepalived.
          operatorframework.io/suggested-namespace: keepalived-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/keepalived-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KeepalivedGroup
            name: keepalivedgroups.redhatcop.redhat.io
            version: v1alpha1
        description: "The objective of the keepalived operator is to allow for a way
          to create self-hosted load balancers in an automated way. From a user experience
          point of view the behavior is the same as of when creating [`LoadBalancer`](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          services with a cloud provider able to manage them.\n\nThe keepalived operator
          can be used in all environments that allows nodes to advertise additional
          IPs on their NICs (and at least for now, in networks that allow multicast),
          however it's mainly aimed at supporting LoadBalancer services and ExternalIPs
          on bare metal installations (or other installation environments where a
          cloud provider is not available).\n\nOne possible use of the keepalived
          operator is also to support [OpenShift Ingresses](https://docs.openshift.com/container-platform/4.5/networking/configuring_ingress_cluster_traffic/overview-traffic.html)
          in environments where an external load balancer cannot be provisioned.\n\n##
          How it works\n\nThe keepalived operator will create one or more [VIPs](https://en.wikipedia.org/wiki/Virtual_IP_address)
          (an HA IP that floats between multiple nodes), based on the [`LoadBalancer`](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          services and/or services requesting [`ExternalIPs`](https://kubernetes.io/docs/concepts/services-networking/service/#external-ips).\n\nFor
          `LoadBalancer` services the IPs found at `.Status.LoadBalancer.Ingress[].IP`
          will become VIPs.\n\nFor services requesting a `ExternalIPs`, the IPs found
          at `.Spec.ExternalIPs[]` will become VIPs.\n\nNote that a service can be
          of `LoadBalancer` type and also request `ExternalIPs`, it this case both
          sets of IPs will become VIPs.\n\nDue to a [keepalived](https://www.keepalived.org/manpage.html)
          limitation a single keepalived cluster can manage up to 256 VIP configurations.
          Multiple keepalived clusters can coexists in the same network as long as
          they use different multicast ports [TODO verify this statement].\n\nTo address
          this limitation the `KeepalivedGroup` [CRD](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/)
          has been introduced. This CRD is supposed to be configured by an administrator
          and allows you to specify a node selector to pick on which nodes the keepalived
          pods should be deployed. Here is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          KeepalivedGroup\nmetadata:\n  name: keepalivedgroup-router\nspec:\n  image:
          registry.redhat.io/openshift4/ose-keepalived-ipfailover\n  interface: ens3\n
          \ nodeSelector:\n    node-role.kubernetes.io/loadbalancer: \"\"\n  blacklistRouterIDs:\n
          \ - 1\n  - 2  \n```\n\nThis KeepalivedGroup will be deployed on all the
          nodes with role `loadbalancer`. One must also specify the network device
          on which the VIPs will be exposed, it is assumed that all the nodes have
          the same network device configuration.\n\nServices must be annotated to
          opt-in to being observed by the keepalived operator and to specify which
          KeepalivedGroup they refer to. The annotation looks like this:\n\n`keepalived-operator.redhat-cop.io/keepalivedgroup:
          <keepalivedgroup namespace>/<keepalivedgroup-name>`\n\nThe image used for
          the keepalived containers can be specified with `.Spec.Image` it will default
          to `registry.redhat.io/openshift4/ose-keepalived-ipfailover` if undefined.
          \n\n## Requirements\n\n### Security Context Constraints\n\nEach KeepalivedGroup
          deploys a [daemonset](https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/)
          that requires the [privileged scc](https://docs.openshift.com/container-platform/4.5/authentication/managing-security-context-constraints.html),
          this permission must be given to the `default` service account in the namespace
          where the keepalived group is created by and administrator.\n\n```shell\noc
          adm policy add-scc-to-user privileged -z default -n keepalived-operator\n```\n\n###
          Cluster Network Operator\n\nIn Openshift, use of an external IP address
          is governed by the following fields in the `Network.config.openshift.io`
          CR named `cluster`\n\n* `spec.externalIP.autoAssignCIDRs` defines an IP
          address block used by the load balancer when choosing an external IP address
          for the service. OpenShift supports only a single IP address block for automatic
          assignment.\n\n* `spec.externalIP.policy` defines the permissible IP address
          blocks when manually specifying an IP address. OpenShift does not apply
          policy rules to IP address blocks defined by `spec.externalIP.autoAssignCIDRs`\n\nThe
          following patch can be used to configure the Cluster Network Operator:\n\n```yaml\nspec:\n
          \ externalIP:\n    policy:\n      allowedCIDRs:\n      - ${ALLOWED_CIDR}\n
          \   autoAssignCIDRs:\n      - \"${AUTOASSIGNED_CIDR}\"\n```\n\nHere is an
          example of how to apply the patch:\n\n```shell\nexport ALLOWED_CIDR=\"192.168.131.128/26\"\nexport
          AUTOASSIGNED_CIDR=\"192.168.131.192/26\"\noc patch network cluster -p \"$(envsubst
          < ./network-patch.yaml | yq r -j -)\" --type=merge\n```\nAdditionally, the
          fields can be edited manually via `oc edit Network.config.openshift.io cluster`\n\n##
          Blacklisting router IDs\n\nIf the Keepalived pods are deployed on nodes
          which are in the same network (same broadcast domain to be precise) with
          other keepalived the process, it's necessary to ensure that there is no
          collision between the used routers it.\nFor this purpose it is possible
          to provide a `blacklistRouterIDs` field with a list of black-listed IDs
          that will not be used.\n\n## OpenShift RHV, vSphere, OSP and bare metal
          IPI instructions\n\nWhen IPI is used for RHV, vSphere, OSP or bare metal
          platforms, three keepalived VIPs are deployed. To make sure that keepalived-operator
          can work in these environment we need to discover and blacklist the corresponding
          VRRP router IDs.\n\nTo discover the VRRP router IDs being used, run the
          following command, you can run this command from you laptop:\n\n```shell\npodman
          run quay.io/openshift/origin-baremetal-runtimecfg:4.5 vr-ids <cluster_name>\n```\n\nIf
          you don't know your cluster name, run this command:\n\n```shell\npodman
          run quay.io/openshift/origin-baremetal-runtimecfg:4.5 vr-ids $(oc get cm
          cluster-config-v1 -n kube-system -o jsonpath='{.data.install-config}'| yq
          -r .metadata.name)\n```\n\nThen use these [instructions](#Blacklisting-router-IDs)
          to blacklist those VRRP router IDs.\n\n## Verbatim Configurations\n\nKeepalived
          has dozens of [configurations](https://www.keepalived.org/manpage.html).
          At the early stage of this project it's difficult to tell which one should
          be modeled in the API. Yet, users of this project may still need to use
          them. To account for that there is a way to pass verbatim options both at
          the keepalived group level (which maps to the keepalived config `global_defs`
          section) and at the service level (which maps to the keepalived config `vrrp_instance`
          section).\n\nKeepalivedGroup-level verbatim configurations can be passed
          as in the following example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          KeepalivedGroup\nmetadata:\n  name: keepalivedgroup-router\nspec:\n  interface:
          ens3\n  nodeSelector:\n    node-role.kubernetes.io/loadbalancer: \"\"\n
          \ verbatimConfig:  \n    vrrp_iptables: my-keepalived\n```\n\nthis will
          map to the following `global_defs`:\n\n```\n    global_defs {\n        router_id
          keepalivedgroup-router\n        vrrp_iptables my-keepalived\n    }\n```\n\nService-level
          verbatim configurations can be passed as in the following example:\n\n```yaml\napiVersion:
          v1\nkind: Service\nmetadata:\n  annotations:\n    keepalived-operator.redhat-cop.io/keepalivedgroup:
          keepalived-operator/keepalivedgroup-router\n    keepalived-operator.redhat-cop.io/verbatimconfig:
          '{ \"track_src_ip\": \"\" }'\n```\n\nthis will map to the following `vrrp_instance`
          section\n\n```\n    vrrp_instance openshift-ingress/router-default {\n        interface
          ens3\n        virtual_router_id 1  \n        virtual_ipaddress {\n          192.168.131.129\n
          \       }\n        track_src_ip\n    }\n```\n\n## Metrics collection\n\nEach
          keepalived pod exposes a [Prometheus](https://prometheus.io/) metrics port
          at `9650`. Metrics are collected with [keepalived_exporter](github.com/gen2brain/keepalived_exporter),
          the available metrics are described in the project documentation.\n\nWhen
          a keepalived group is created a [`PodMonitor`](https://github.com/coreos/prometheus-operator/blob/master/Documentation/api.md#podmonitor)
          rule to collect those metrics. All PodMonitor resources created that way
          have the label: `metrics: keepalived`. It is up to you to make sure your
          Prometheus instance watches for those `PodMonitor` rules. Here is an example
          of a fragment of a `Prometheus` CR configured to collect the keepalived
          pod metrics:\n\n```yaml\n  podMonitorSelector:\n    matchLabels:\n      metrics:
          keepalived\n```\n"
        displayName: Keepalived Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadbalancer
        - keepalived
        - ipfailover
        - externalIP
        - self-hosted
        links:
        - name: Keepalived Operator
          url: https://github.com/redhat-cop/keepalived-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/keepalived-operator
        maintainers:
        - email: rspazzol@redhat.com'
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/keepalived-operator@sha256:cd0c2f7df82df1178795e4a18a0942da33adb9670b192a7b7a2b1483a5b4ccf8
        version: 1.5.4
      entries:
      - name: keepalived-operator.v1.5.4
        version: 1.5.4
      - name: keepalived-operator.v1.5.3
        version: 1.5.3
      - name: keepalived-operator.v1.5.2
        version: 1.5.2
      - name: keepalived-operator.v1.5.1
        version: 1.5.1
      - name: keepalived-operator.v1.5.0
        version: 1.5.0
      - name: keepalived-operator.v1.4.2
        version: 1.4.2
      - name: keepalived-operator.v1.4.1
        version: 1.4.1
      - name: keepalived-operator.v1.4.0
        version: 1.4.0
      - name: keepalived-operator.v1.3.4
        version: 1.3.4
      - name: keepalived-operator.v1.3.3
        version: 1.3.3
      - name: keepalived-operator.v1.3.2
        version: 1.3.2
      - name: keepalived-operator.v1.3.1
        version: 1.3.1
      - name: keepalived-operator.v1.3.0
        version: 1.3.0
      - name: keepalived-operator.v1.2.2
        version: 1.2.2
      - name: keepalived-operator.v1.2.1
        version: 1.2.1
      - name: keepalived-operator.v1.2.0
        version: 1.2.0
      - name: keepalived-operator.v1.1.0
        version: 1.1.0
      - name: keepalived-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: keepalived-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel Corporation
      provider-url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
    name: sriov-fec
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sriov-fec.v2.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecClusterConfig",
                "metadata": {
                  "name": "acc100SampleConfig",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "acceleratorSelector": {
                    "deviceID": "someDevice",
                    "driver": "someDriver",
                    "maxVirtualFunctions": 2,
                    "pciAddress": "somePciAddress",
                    "vendorID": "someVendor"
                  },
                  "drainSkip": false,
                  "nodeSelector": {
                    "expectedLabel1": "valueOfExpectedLabel1",
                    "expectedLabelN": "valueOfExpectedLabelN"
                  },
                  "physicalFunction": {
                    "bbDevConfig": {
                      "acc100": {
                        "downlink4G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 0
                        },
                        "downlink5G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 4
                        },
                        "maxQueueSize": 1024,
                        "numVfBundles": 16,
                        "pfMode": true,
                        "uplink4G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 0
                        },
                        "uplink5G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 4
                        }
                      }
                    },
                    "pfDriver": "pci-pf-stub",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  },
                  "priority": 100
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecClusterConfig",
                "metadata": {
                  "name": "n3000SampleConfig",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "acceleratorSelector": {
                    "deviceID": "someDevice",
                    "driver": "someDriver",
                    "maxVirtualFunctions": 2,
                    "pciAddress": "somePciAddress",
                    "vendorID": "someVendor"
                  },
                  "drainSkip": false,
                  "nodeSelector": {
                    "expectedLabel1": "valueOfExpectedLabel1",
                    "expectedLabelN": "valueOfExpectedLabelN"
                  },
                  "physicalFunction": {
                    "bbDevConfig": {
                      "n3000": {
                        "downlink": {
                          "bandwidth": 3,
                          "loadBalance": 128,
                          "queues": {
                            "vf0": 16,
                            "vf1": 16,
                            "vf2": 0,
                            "vf3": 0,
                            "vf4": 0,
                            "vf5": 0,
                            "vf6": 0,
                            "vf7": 0
                          }
                        },
                        "flrTimeout": 610,
                        "networkType": "FPGA_5GNR",
                        "pfMode": true,
                        "uplink": {
                          "bandwidth": 3,
                          "loadBalance": 128,
                          "queues": {
                            "vf0": 16,
                            "vf1": 16,
                            "vf2": 0,
                            "vf3": 0,
                            "vf4": 0,
                            "vf5": 0,
                            "vf6": 0,
                            "vf7": 0
                          }
                        }
                      }
                    },
                    "pfDriver": "pci-pf-stub",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  },
                  "priority": 100
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecNodeConfig",
                "metadata": {
                  "name": "acc100-worker",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "drainSkip": false,
                  "physicalFunctions": [
                    {
                      "bbDevConfig": {
                        "acc100": {
                          "downlink4G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 0
                          },
                          "downlink5G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 4
                          },
                          "maxQueueSize": 1024,
                          "numVfBundles": 16,
                          "pfMode": true,
                          "uplink4G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 0
                          },
                          "uplink5G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 4
                          }
                        }
                      },
                      "pci_addr": "somePciAddress",
                      "pf_driver": "pci-pf-stub",
                      "vf_amount": 2,
                      "vf_driver": "vfio-pci"
                    }
                  ]
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecNodeConfig",
                "metadata": {
                  "name": "n3000-worker",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "drainSkip": false,
                  "physicalFunctions": [
                    {
                      "bbDevConfig": {
                        "n3000": {
                          "downlink": {
                            "bandwidth": 3,
                            "loadBalance": 128,
                            "queues": {
                              "vf0": 16,
                              "vf1": 16,
                              "vf2": 0,
                              "vf3": 0,
                              "vf4": 0,
                              "vf5": 0,
                              "vf6": 0,
                              "vf7": 0
                            }
                          },
                          "flrTimeout": 610,
                          "networkType": "FPGA_5GNR",
                          "pfMode": true,
                          "uplink": {
                            "bandwidth": 3,
                            "loadBalance": 128,
                            "queues": {
                              "vf0": 16,
                              "vf1": 16,
                              "vf2": 0,
                              "vf3": 0,
                              "vf4": 0,
                              "vf5": 0,
                              "vf6": 0,
                              "vf7": 0
                            }
                          }
                        }
                      },
                      "pci_addr": "somePciAddress",
                      "pf_driver": "pci-pf-stub",
                      "vf_amount": 2,
                      "vf_driver": "vfio-pci"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/intel/sriov-fec-operator@sha256:adb6a4c2da6399c8ed40c906f790bce13bfcefae3326a047f6a50592b3f5e485
          createdAt: "2025-01-15"
          description: An operator for Intel Wireless FEC Accelerator to orchestrate
            and manage the resources/devices exposed by a range of Intel's vRAN FEC
            acceleration devices/hardware within the OpenShift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/intel/sriov-fec-operator
          support: Intel Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovFecClusterConfig is the Schema for the sriovfecclusterconfigs
              API
            displayName: SriovFecClusterConfig
            kind: SriovFecClusterConfig
            name: sriovfecclusterconfigs.sriovfec.intel.com
            version: v2
          - description: SriovFecNodeConfig is the Schema for the sriovfecnodeconfigs
              API
            displayName: SriovFecNodeConfig
            kind: SriovFecNodeConfig
            name: sriovfecnodeconfigs.sriovfec.intel.com
            version: v2
          - description: SriovVrbClusterConfig is the Schema for the sriovvrbclusterconfigs
              API
            displayName: SriovVrbClusterConfig
            kind: SriovVrbClusterConfig
            name: sriovvrbclusterconfigs.sriovvrb.intel.com
            version: v1
          - description: SriovVrbNodeConfig is the Schema for the sriovvrbnodeconfigs
              API
            displayName: SriovVrbNodeConfig
            kind: SriovVrbNodeConfig
            name: sriovvrbnodeconfigs.sriovvrb.intel.com
            version: v1
        description: The Intel vRAN Boost family of accelerators ACC100/VRB1/VRB2,
          to offload and accelerate the computing-intensive process of forward error
          correction (FEC) and more for 4G/LTE and 5G technology, freeing up processing
          power. It allows the optimization of data plane performance to reduce total
          cost of ownership while maintaining a high degree of flexibility.  The Intel
          vRAN Boost accelerators ACC100/VRB1/VRB2 plays a key role in accelerating
          5G and network functions virtualization (NFV) adoption for ecosystem partners
          such as telecommunications equipment manufacturers (TEMs) virtual network
          functions (VNF) vendors, system integrators and telcos, to bring scalable
          and high-performance solutions to market. The Intel ACC100 includes a variant
          that is design to be Network Equipment Building System (NEBS)-friendly,
          and features a Root-of-Trust device that helps protect systems from FPGA
          host security exploits.
        displayName: SRIOV-FEC Operator for Intel vRAN Boost Accelerators
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ACC100
        - vRAN
        - ORAN
        links:
        - name: SRIOV-FEC source code
          url: https://github.com/intel/sriov-fec-operator
        maturity: alpha
        provider:
          name: Intel Corporation
          url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
        relatedImages:
        - registry.connect.redhat.com/intel/sriov-fec-daemon@sha256:5b0c60b67bcf5d7b2797deab0d5e3fa079577179f0acc6b0062533584ea0bc23
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1ed3b514bf9af6ac0d386e3739874d1d964e667d54e041c2d67edd769b0009fb
        - registry.connect.redhat.com/intel/sriov-fec-operator@sha256:adb6a4c2da6399c8ed40c906f790bce13bfcefae3326a047f6a50592b3f5e485
        - registry.connect.redhat.com/intel/n3000-labeler@sha256:b417cfcedc998a35491784005cc7eee3d9c647b84eac5ca41cb459076870b8f6
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin@sha256:d0fcb6e9135d9a19ea26851bcdb4ac72c04bd747706fc5123020460e9dbfc21b
        version: 2.11.1
      entries:
      - name: sriov-fec.v2.11.1
        version: 2.11.1
      - name: sriov-fec.v2.11.0
        version: 2.11.0
      - name: sriov-fec.v2.10.0
        version: 2.10.0
      - name: sriov-fec.v2.9.0
        version: 2.9.0
      - name: sriov-fec.v2.8.0
        version: 2.8.0
      - name: sriov-fec.v2.7.2
        version: 2.7.2
      - name: sriov-fec.v2.7.1
        version: 2.7.1
      name: stable
    defaultChannel: stable
    packageName: sriov-fec
    provider:
      name: Intel Corporation
      url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: spyre-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spyre-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreClusterPolicy",
                "metadata": {
                  "name": "spyreclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2",
                      "spyreFilter": ".",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "latest"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "spyre-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "spyre-device-plugin-init",
                      "repository": "icr.io/ibmaiu_internal",
                      "version": "0.1.1"
                    },
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "spyre-exporter",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.2"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "spyre-webhook-validator",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  },
                  "scheduler": {
                    "image": "spyre-scheduler",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  }
                }
              },
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu_internal/spyre-operator@sha256:0f657c98668a8358c37b3fc2ed3f31ec78b7ef7f9a9678e7cd35d383d22ce2da
          createdAt: "2025-07-31T10:32:48Z"
          description: Operator for IBM Spyre devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM Spyre
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["spyrenodestates.spyre.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpyreClusterPolicy is the Schema for the Spyre API
            displayName: Spyre Cluster Policy
            kind: SpyreClusterPolicy
            name: spyreclusterpolicies.spyre.ibm.com
            version: v1alpha1
          - description: SpyreNodeState is the Schema for the SpyreNodeState API
            displayName: Spyre Node State
            kind: SpyreNodeState
            name: spyrenodestates.spyre.ibm.com
            version: v1alpha1
        description: Operator for IBM Spyre devices
        displayName: IBM Spyre Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: IBM Spyre Operator
          url: https://spyre-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: Satyam.Gupta.3@ibm.com
          name: Satyam Gupta
        - email: Rehan.Khan7@ibm.com
          name: Rehan Khan
        - email: Nishan.Acharya@ibm.com
          name: Nishan Acharya
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu_internal/spyre-operator@sha256:0f657c98668a8358c37b3fc2ed3f31ec78b7ef7f9a9678e7cd35d383d22ce2da
        - icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2
        version: 0.1.2
      entries:
      - name: spyre-operator.v0.1.2
        version: 0.1.2
      name: stable-v0.1
    - currentCSV: spyre-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreClusterPolicy",
                "metadata": {
                  "name": "spyreclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb",
                      "spyreFilter": ".",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "latest"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "spyre-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "spyre-device-plugin-init",
                      "repository": "quay.io/ibm-aiu",
                      "version": "0.2.0"
                    },
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "spyre-exporter",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "spyre-webhook-validator",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "scheduler": {
                    "image": "spyre-scheduler",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.1"
                  }
                }
              },
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/ibm-aiu/spyre-operator@sha256:d0cdd449f1d8f0dd9281c4d9a809333db6c4ed201bb14187b6c70ae686a30e99
          createdAt: "2025-09-02T13:25:34Z"
          description: Operator for IBM Spyre devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM Spyre
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["spyrenodestates.spyre.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpyreClusterPolicy is the Schema for the Spyre API
            displayName: Spyre Cluster Policy
            kind: SpyreClusterPolicy
            name: spyreclusterpolicies.spyre.ibm.com
            version: v1alpha1
          - description: SpyreNodeState is the Schema for the SpyreNodeState API
            displayName: Spyre Node State
            kind: SpyreNodeState
            name: spyrenodestates.spyre.ibm.com
            version: v1alpha1
        description: Operator for IBM Spyre devices
        displayName: IBM Spyre Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: IBM Spyre Operator
          url: https://spyre-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: Satyam.Gupta.3@ibm.com
          name: Satyam Gupta
        - email: Rehan.Khan7@ibm.com
          name: Rehan Khan
        - email: Nishan.Acharya@ibm.com
          name: Nishan Acharya
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - quay.io/ibm-aiu/spyre-operator@sha256:d0cdd449f1d8f0dd9281c4d9a809333db6c4ed201bb14187b6c70ae686a30e99
        version: 0.2.1
      entries:
      - name: spyre-operator.v0.2.1
        version: 0.2.1
      name: stable-v0.2
    - currentCSV: spyre-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreClusterPolicy",
                "metadata": {
                  "name": "spyreclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "config": {
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb",
                      "spyreFilter": ".",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "latest"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "spyre-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "spyre-device-plugin-init",
                      "repository": "quay.io/ibm-aiu",
                      "version": "1.0.0"
                    },
                    "repository": "quay.io/ibm-aiu",
                    "version": "1.0.0"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "spyre-exporter",
                    "repository": "quay.io/ibm-aiu",
                    "version": "1.0.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "spyre-webhook-validator",
                    "repository": "quay.io/ibm-aiu",
                    "version": "1.0.0"
                  },
                  "scheduler": {
                    "image": "spyre-scheduler",
                    "repository": "quay.io/ibm-aiu",
                    "version": "1.0.0"
                  }
                }
              },
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/ibm-aiu/spyre-operator@sha256:dd262e8c65f35b12c5d79571e8b5ddd54dd0ee2693f28bea05a6fc02ac1fe6f0
          createdAt: "2025-10-01T00:40:23Z"
          description: Operator for IBM Spyre devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: spyre-operator
          operators.openshift.io/valid-subscription: use with IBM Spyre
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["spyrenodestates.spyre.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpyreClusterPolicy is the Schema for the Spyre API
            displayName: Spyre Cluster Policy
            kind: SpyreClusterPolicy
            name: spyreclusterpolicies.spyre.ibm.com
            version: v1alpha1
          - description: SpyreNodeState is the Schema for the SpyreNodeState API
            displayName: Spyre Node State
            kind: SpyreNodeState
            name: spyrenodestates.spyre.ibm.com
            version: v1alpha1
        description: Operator for IBM Spyre devices
        displayName: IBM Spyre Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: IBM Spyre Operator
          url: https://spyre-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: Satyam.Gupta.3@ibm.com
          name: Satyam Gupta
        - email: Rehan.Khan7@ibm.com
          name: Rehan Khan
        - email: Nishan.Acharya@ibm.com
          name: Nishan Acharya
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - quay.io/ibm-aiu/spyre-operator@sha256:dd262e8c65f35b12c5d79571e8b5ddd54dd0ee2693f28bea05a6fc02ac1fe6f0
        version: 1.0.0
      entries:
      - name: spyre-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: spyre-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hive-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hive-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:e71fc728632622f7ea97357ec361dd2bb73e63c0707e8014ca8ecf8ec4ab93f1
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:e71fc728632622f7ea97357ec361dd2bb73e63c0707e8014ca8ecf8ec4ab93f1
        version: 24.11.1
      entries:
      - name: hive-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: hive-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:5530f5034cd83c2e8b860fce737ac491a2adc6d22246356bee0c79768ed64706
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:5530f5034cd83c2e8b860fce737ac491a2adc6d22246356bee0c79768ed64706
        version: 25.3.0
      entries:
      - name: hive-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: hive-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
        version: 25.7.0
      entries:
      - name: hive-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hive-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
        version: 25.7.0
      entries:
      - name: hive-operator.v25.7.0
        version: 25.7.0
      - name: hive-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hive-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CSI Addons Community
      provider-url: ""
    name: odf-csi-addons-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-csi-addons-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          createdAt: "2025-09-12T07:33:01Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIAddonsNode is the Schema for the csiaddonsnode API
            displayName: CSIAddons Node
            kind: CSIAddonsNode
            name: csiaddonsnodes.csiaddons.openshift.io
            version: v1alpha1
          - description: NetworkFence is the Schema for the networkfences API
            displayName: Network Fence
            kind: NetworkFence
            name: networkfences.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceCronJob is the Schema for the reclaimspacecronjobs
              API
            displayName: Reclaim Space CronJob
            kind: ReclaimSpaceCronJob
            name: reclaimspacecronjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceJob is the Schema for the reclaimspacejobs API
            displayName: Reclaim Space Job
            kind: ReclaimSpaceJob
            name: reclaimspacejobs.csiaddons.openshift.io
            version: v1alpha1
          - description: VolumeReplicationClass is the Schema for the volumereplicationclasses
              API
            displayName: Volume Replication Class
            kind: VolumeReplicationClass
            name: volumereplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplication is the Schema for the volumereplications
              API
            displayName: Volume Replication
            kind: VolumeReplication
            name: volumereplications.replication.storage.openshift.io
            version: v1alpha1
        description: CSI Addons provides the CSI Addons Controller that enables advanced
          storage operations for CSI-drivers.
        displayName: CSI Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: CSI Addons
          url: https://csi-addons.github.io
        maintainers:
        - email: csi-addons@redhat.com
          name: CSI Addons Community
        maturity: alpha
        provider:
          name: CSI Addons Community
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/odf4/odf-csi-addons-rhel9-operator@sha256:8b0c16c21fccca9b674c4d980a793b6cde482b86363b793e173126b030bdcc55
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:9a0008f662b6afa9e2677a8a83bd662d806ad6f4d03fb49bccb78ed487ce4c7a
        version: 4.15.20-rhodf
      entries:
      - name: odf-csi-addons-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odf-csi-addons-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odf-csi-addons-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odf-csi-addons-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odf-csi-addons-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odf-csi-addons-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odf-csi-addons-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odf-csi-addons-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odf-csi-addons-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odf-csi-addons-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odf-csi-addons-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odf-csi-addons-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odf-csi-addons-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odf-csi-addons-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odf-csi-addons-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odf-csi-addons-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odf-csi-addons-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odf-csi-addons-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odf-csi-addons-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odf-csi-addons-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odf-csi-addons-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          createdAt: "2025-09-12T06:32:07Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIAddonsNode is the Schema for the csiaddonsnode API
            displayName: CSIAddons Node
            kind: CSIAddonsNode
            name: csiaddonsnodes.csiaddons.openshift.io
            version: v1alpha1
          - description: NetworkFence is the Schema for the networkfences API
            displayName: Network Fence
            kind: NetworkFence
            name: networkfences.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceCronJob is the Schema for the reclaimspacecronjobs
              API
            displayName: Reclaim Space CronJob
            kind: ReclaimSpaceCronJob
            name: reclaimspacecronjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceJob is the Schema for the reclaimspacejobs API
            displayName: Reclaim Space Job
            kind: ReclaimSpaceJob
            name: reclaimspacejobs.csiaddons.openshift.io
            version: v1alpha1
          - description: VolumeReplicationClass is the Schema for the volumereplicationclasses
              API
            displayName: Volume Replication Class
            kind: VolumeReplicationClass
            name: volumereplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplication is the Schema for the volumereplications
              API
            displayName: Volume Replication
            kind: VolumeReplication
            name: volumereplications.replication.storage.openshift.io
            version: v1alpha1
        description: CSI Addons provides the CSI Addons Controller that enables advanced
          storage operations for CSI-drivers.
        displayName: CSI Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: CSI Addons
          url: https://csi-addons.github.io
        maintainers:
        - email: csi-addons@redhat.com
          name: CSI Addons Community
        maturity: alpha
        provider:
          name: CSI Addons Community
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/odf4/odf-csi-addons-rhel9-operator@sha256:d370f19578d1c466d39d00da82f1f3fe3d9f13d7021ed0c39f759b24b0ed7573
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:0afcb516eb528857887c6ee1a9cba081dc8f6b7d47d9f26a0dbb2dca0cefe07c
        version: 4.16.16-rhodf
      entries:
      - name: odf-csi-addons-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odf-csi-addons-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odf-csi-addons-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odf-csi-addons-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odf-csi-addons-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odf-csi-addons-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odf-csi-addons-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odf-csi-addons-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odf-csi-addons-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odf-csi-addons-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odf-csi-addons-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odf-csi-addons-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odf-csi-addons-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odf-csi-addons-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odf-csi-addons-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odf-csi-addons-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odf-csi-addons-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odf-csi-addons-operator
    provider:
      name: CSI Addons Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp, Inc.
      provider-url: https://www.netapp.com/
    name: trident-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trident-operator.v25.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "trident.netapp.io/v1",
                "kind": "TridentOrchestrator",
                "metadata": {
                  "name": "trident"
                },
                "spec": {
                  "IPv6": false,
                  "debug": true,
                  "nodePrep": [],
                  "imagePullSecrets": [],
                  "imageRegistry": "",
                  "namespace": "trident",
                  "silenceAutosupport": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: docker.io/netapp/trident-operator:25.06.2
          createdAt: "2025-08-03"
          description: Trident Operator, to manage NetApp Trident installations
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          repository: https://github.com/netapp/trident
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Used to deploy Astra Trident.
            displayName: Trident Orchestrator
            kind: TridentOrchestrator
            name: tridentorchestrators.trident.netapp.io
            version: v1
          - description: Automates AWS FSxN backend configuration
            displayName: Trident Configurator
            kind: TridentConfigurator
            name: tridentconfigurators.trident.netapp.io
            version: v1
        description: |
          NetApp Trident is an open source storage provisioner and orchestrator maintained by NetApp. It enables you to create storage volumes for containerized applications managed by Docker and Kubernetes. For full release information, including patch release changes, see https://docs.netapp.com/us-en/trident/trident-rn.html.
        displayName: NetApp Trident
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - NetApp
        - ONTAP
        - Trident
        - storage
        - performance
        - open source
        - Astra
        - Cloud Native
        - AWS
        - Azure
        - GCP
        - Cloud Volumes
        links:
        - name: GitHub Repository
          url: https://github.com/NetApp/trident
        - name: Trident documentation
          url: https://docs.netapp.com/us-en/trident/index.html
        - name: Support policy
          url: https://mysupport.netapp.com/site/info/version-support
        - name: Release Notes
          url: https://docs.netapp.com/us-en/trident/trident-rn.html
        maintainers:
        - email: ng-astra-product@netapp.com
          name: NetApp
        provider:
          name: NetApp, Inc.
          url: https://www.netapp.com/
        relatedImages:
        - docker.io/netapp/trident-operator:25.06.2
        version: 25.6.2
      entries:
      - name: trident-operator.v25.6.2
        version: 25.6.2
      - name: trident-operator.v25.6.1
        version: 25.6.1
      - name: trident-operator.v25.6.0
        version: 25.6.0
      - name: trident-operator.v25.2.1
        version: 25.2.1
      - name: trident-operator.v25.2.0
        version: 25.2.0
      - name: trident-operator.v24.10.1
        version: 24.10.1
      - name: trident-operator.v24.10.0
        version: 24.10.0
      - name: trident-operator.v24.6.0
        version: 24.6.0
      - name: trident-operator.v24.2.0
        version: 24.2.0
      name: stable
    defaultChannel: stable
    packageName: trident-operator
    provider:
      name: NetApp, Inc.
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: multicluster-operators-subscription
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-operators-subscription.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.3.0
          createdAt: "2021-07-19T18:00:00Z"
          description: An operator to subscribe resources from object Bucket, helm
            and github types of channels
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-application:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-channel:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-placementrule:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-deployable:2.3.0
        version: 0.3.0
      entries:
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.3
    - currentCSV: multicluster-operators-subscription.v0.3.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.4.0
          createdAt: "2021-09-29T10:23:00Z"
          description: An operator to subscribe resources from object Bucket, helm
            and github types of channels
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-placementrule:2.4.0
        - quay.io/open-cluster-management/multicloud-integrations:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-deployable:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-application:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-channel:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.4.0
        version: 0.3.6
      entries:
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.4
    - currentCSV: multicluster-operators-subscription.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.5.0
          createdAt: "2022-03-23T18:00:00Z"
          description: An operator for deploying resources from object Bucket, helm
            repo or github repo to multiple managed clusters
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management-io/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: subscription status per cluster per app
            displayName: App Subscription Report
            kind: SubscriptionReport
            name: subscriptionreports.apps.open-cluster-management.io
            version: v1alpha1
          - description: subscription status per package
            displayName: App Subscription Status
            kind: SubscriptionStatus
            name: subscriptionstatuses.apps.open-cluster-management.io
            version: v1alpha1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.5.0
        - quay.io/open-cluster-management/multicloud-integrations:2.5.0
        - quay.io/open-cluster-management/multicluster-operators-application:2.5.0
        - quay.io/open-cluster-management/multicluster-operators-channel:2.5.0
        version: 0.4.0
      entries:
      - name: multicluster-operators-subscription.v0.4.0
        version: 0.4.0
      - name: multicluster-operators-subscription.v0.3.9
        version: 0.3.9
      - name: multicluster-operators-subscription.v0.3.8
        version: 0.3.8
      - name: multicluster-operators-subscription.v0.3.7
        version: 0.3.7
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.5
    - currentCSV: multicluster-operators-subscription.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-operators-subscription:2.6.0
          createdAt: "2022-06-06T18:00:00Z"
          description: An operator for deploying resources from object Bucket, helm
            repo or github repo to multiple managed clusters
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management-io/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: subscription status per cluster per app
            displayName: App Subscription Report
            kind: SubscriptionReport
            name: subscriptionreports.apps.open-cluster-management.io
            version: v1alpha1
          - description: subscription status per package
            displayName: App Subscription Status
            kind: SubscriptionStatus
            name: subscriptionstatuses.apps.open-cluster-management.io
            version: v1alpha1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/multicluster-operators-subscription:2.6.0
        - quay.io/stolostron/multicloud-integrations:2.6.0
        - quay.io/stolostron/multicluster-operators-application:2.6.0
        - quay.io/stolostron/multicluster-operators-channel:2.6.0
        version: 0.4.1
      entries:
      - name: multicluster-operators-subscription.v0.4.1
        version: 0.4.1
      - name: multicluster-operators-subscription.v0.4.0
        version: 0.4.0
      - name: multicluster-operators-subscription.v0.3.9
        version: 0.3.9
      - name: multicluster-operators-subscription.v0.3.8
        version: 0.3.8
      - name: multicluster-operators-subscription.v0.3.7
        version: 0.3.7
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.6
    defaultChannel: release-2.5
    packageName: multicluster-operators-subscription
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.90.0-beta.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"externalDbIP":
            "10.0.2.15","customImageNames": false,"repository": "icr.io/cpopen/turbonomic","tag":"8.17.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
          createdAt: "2023-03-24T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=42.53.0-beta.1 <42.90.0-beta.2'
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints. Turbonomic platform install and configuration
              documentation can be found online here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How do I start?

          Turbonomic platform installation and configuration documentation can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
          Carefully read through the prerequisites before beginning the installation.  Once ready, on the `Details` tab for the Turbonomic Platform Operator, under `Provided APIs`, click `Create instance`.
          For samples of common configurations, refer to sample custom resource YAMLs provided here: https://github.com/turbonomic/t8c-install/tree/master/samples
          Looking to upgrade your current deployment? Refer to documentation here: https://www.ibm.com/docs/en/tarm/latest?topic=version-updating-using-red-hat-openshift-operatorhub

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
        version: 42.90.0-beta.2
      entries:
      - name: t8c-operator.v42.90.0-beta.2
        version: 42.90.0-beta.2
      name: beta
    - currentCSV: t8c-operator.v42.90.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"externalDbIP":
            "10.0.2.15","customImageNames": false,"repository": "icr.io/cpopen/turbonomic","tag":"8.17.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
          createdAt: "2023-03-24T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=42.53.0 <42.90.0'
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints. Turbonomic platform install and configuration
              documentation can be found online here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How do I start?

          Turbonomic platform installation and configuration documentation can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
          Carefully read through the prerequisites before beginning the installation.  Once ready, on the `Details` tab for the Turbonomic Platform Operator, under `Provided APIs`, click `Create instance`.
          For samples of common configurations, refer to sample custom resource YAMLs provided here: https://github.com/turbonomic/t8c-install/tree/master/samples
          Looking to upgrade your current deployment? Refer to documentation here: https://www.ibm.com/docs/en/tarm/latest?topic=version-updating-using-red-hat-openshift-operatorhub

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
        version: 42.90.0
      entries:
      - name: t8c-operator.v42.90.0
        version: 42.90.0
      name: stable
    defaultChannel: stable
    packageName: t8c-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: https://stakater.com
    name: tenant-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tenant-operator.v0.12.66
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: =0.12.65
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:9f806a065b25229741dac8437a32ec45a99acecf4e6bcbe6e7d1fa4192bc560d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a35b96eacb4e46114b39354f66429f8f05df09c55c98c9416b4ffa2768310ef7
        version: 0.12.66
      entries:
      - name: tenant-operator.v0.12.66
        version: 0.12.66
      name: release-0.12
    - currentCSV: tenant-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2024-10-16T08:15:54Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.0
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:feb71a5045e621f8a8cbfe7e41831100b6b85af9812138071b6217cc4396860a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a35b96eacb4e46114b39354f66429f8f05df09c55c98c9416b4ffa2768310ef7
        version: 1.0.0
      entries:
      - name: tenant-operator.v1.0.0
        version: 1.0.0
      name: release-1.0
    - currentCSV: tenant-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-01-27T13:23:27Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.0
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:f412b6d05038a185f5fe35f9c917388194a5ea66afb538a513423c24390983ce
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a35b96eacb4e46114b39354f66429f8f05df09c55c98c9416b4ffa2768310ef7
        version: 1.1.0
      entries:
      - name: tenant-operator.v1.1.0
        version: 1.1.0
      name: release-1.1
    - currentCSV: tenant-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-03-19T13:30:50Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.2.1
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a35b96eacb4e46114b39354f66429f8f05df09c55c98c9416b4ffa2768310ef7
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:a899d60b31efd5cf2358efcd48737b1c4f4a0356ab6c0abe7a5f62ea825a52cd
        version: 1.2.1
      entries:
      - name: tenant-operator.v1.2.1
        version: 1.2.1
      name: release-1.2
    - currentCSV: tenant-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-06-11T15:25:04Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.3.0
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:e88aa5b008ece406bca35ec22fdc87c842544049d613d4d81c8bc5dd2fb50812
        version: 1.3.0
      entries:
      - name: tenant-operator.v1.3.0
        version: 1.3.0
      name: release-1.3
    - currentCSV: tenant-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-09-08T08:14:28Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.4.2
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:7ee14c0974d1eda849d2c199a56786b0ae777ef6569085be917449f1998b1158
        version: 1.4.2
      entries:
      - name: tenant-operator.v1.4.2
        version: 1.4.2
      name: release-1.4
    defaultChannel: release-1.4
    packageName: tenant-operator
    provider:
      name: Stakater
      url: https://stakater.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.0.4
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
        version: 1.0.4
      entries:
      - name: datadog-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      name: stable
    defaultChannel: stable
    packageName: datadog-operator-certified-rhmp
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sfn-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sfn-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sfn.services.k8s.aws/v1alpha1",
                "kind": "Activity",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sfn.services.k8s.aws/v1alpha1",
                "kind": "StateMachine",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sfn-controller:1.1.1
          createdAt: "2025-09-19T18:41:33Z"
          description: AWS Step Functions controller is a service controller for managing
            Step Functions resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Activity represents the state of an AWS sfn Activity resource.
            displayName: Activity
            kind: Activity
            name: activities.sfn.services.k8s.aws
            version: v1alpha1
          - description: StateMachine represents the state of an AWS sfn StateMachine
              resource.
            displayName: StateMachine
            kind: StateMachine
            name: statemachines.sfn.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Step Functions resources in AWS from within your Kubernetes cluster.

          **About Amazon Step FUnctions**

          Step Functions is a serverless orchestration service that lets you combine AWS Lambda functions and other AWS services to build business-critical applications. Through Step Functions' graphical console, you see your application’s workflow as a series of event-driven steps.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Step Functions
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sfn
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Step Functions Developer Resources
          url: https://aws.amazon.com/step-functions/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: step functions maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sfn-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-sfn-controller.v1.1.1
        version: 1.1.1
      - name: ack-sfn-controller.v1.1.0
        version: 1.1.0
      - name: ack-sfn-controller.v1.0.29
        version: 1.0.29
      - name: ack-sfn-controller.v1.0.28
        version: 1.0.28
      - name: ack-sfn-controller.v1.0.27
        version: 1.0.27
      - name: ack-sfn-controller.v1.0.26
        version: 1.0.26
      - name: ack-sfn-controller.v1.0.25
        version: 1.0.25
      - name: ack-sfn-controller.v1.0.24
        version: 1.0.24
      - name: ack-sfn-controller.v1.0.23
        version: 1.0.23
      - name: ack-sfn-controller.v1.0.22
        version: 1.0.22
      - name: ack-sfn-controller.v1.0.21
        version: 1.0.21
      - name: ack-sfn-controller.v1.0.20
        version: 1.0.20
      - name: ack-sfn-controller.v1.0.19
        version: 1.0.19
      - name: ack-sfn-controller.v1.0.18
        version: 1.0.18
      - name: ack-sfn-controller.v1.0.17
        version: 1.0.17
      - name: ack-sfn-controller.v1.0.16
        version: 1.0.16
      - name: ack-sfn-controller.v1.0.15
        version: 1.0.15
      - name: ack-sfn-controller.v1.0.14
        version: 1.0.14
      - name: ack-sfn-controller.v1.0.13
        version: 1.0.13
      - name: ack-sfn-controller.v1.0.11
        version: 1.0.11
      - name: ack-sfn-controller.v1.0.10
        version: 1.0.10
      - name: ack-sfn-controller.v1.0.9
        version: 1.0.9
      - name: ack-sfn-controller.v1.0.8
        version: 1.0.8
      - name: ack-sfn-controller.v1.0.6
        version: 1.0.6
      - name: ack-sfn-controller.v1.0.5
        version: 1.0.5
      - name: ack-sfn-controller.v1.0.2
        version: 1.0.2
      - name: ack-sfn-controller.v1.0.1
        version: 1.0.1
      - name: ack-sfn-controller.v1.0.0
        version: 1.0.0
      - name: ack-sfn-controller.v0.1.3
        version: 0.1.3
      - name: ack-sfn-controller.v0.1.2
        version: 0.1.2
      - name: ack-sfn-controller.v0.1.1
        version: 0.1.1
      - name: ack-sfn-controller.v0.1.0
        version: 0.1.0
      - name: ack-sfn-controller.v0.0.19
        version: 0.0.19
      - name: ack-sfn-controller.v0.0.18
        version: 0.0.18
      - name: ack-sfn-controller.v0.0.17
        version: 0.0.17
      - name: ack-sfn-controller.v0.0.16
        version: 0.0.16
      - name: ack-sfn-controller.v0.0.15
        version: 0.0.15
      - name: ack-sfn-controller.v0.0.14
        version: 0.0.14
      - name: ack-sfn-controller.v0.0.13
        version: 0.0.13
      - name: ack-sfn-controller.v0.0.12
        version: 0.0.12
      - name: ack-sfn-controller.v0.0.11
        version: 0.0.11
      - name: ack-sfn-controller.v0.0.10
        version: 0.0.10
      - name: ack-sfn-controller.v0.0.9
        version: 0.0.9
      - name: ack-sfn-controller.v0.0.8
        version: 0.0.8
      - name: ack-sfn-controller.v0.0.7
        version: 0.0.7
      - name: ack-sfn-controller.v0.0.6
        version: 0.0.6
      - name: ack-sfn-controller.v0.0.5
        version: 0.0.5
      name: alpha
    defaultChannel: alpha
    packageName: ack-sfn-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeVirt project
      provider-url: ""
    name: community-kubevirt-hyperconverged
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubevirt-hyperconverged-operator.v1.10.7
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployTektonTaskResources":true,"deployVmConsoleProxy":true},"tektonPipelines":{"namespace":"kubevirt"},"tektonTasks":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:0ee30557d5d5a42dec19121e70ed77fbdacd8932f55b11e2c7c1a3d5ad5df4aa
          createdAt: "2024-02-09 14:34:30"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:38293a1050bce9c7cb786cbe8719be1edb91a922a6c583eedb61e56729ed4c72
        - quay.io/kubevirt/virt-controller@sha256:0789fafed2913b35a771e3db882748502b3250be04ece86d97f30201779b4e54
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:e79027973b09aac75860f267c7d7f830599978c38a081001f404b9a1c3d2990f
        - quay.io/kubevirt/mtq-lock-server@sha256:9614df0d5af7c998bcceaf73b548f2e3b29ff8ee4ca64c32613bfbdd6e452f6c
        - quay.io/kubevirt/cdi-uploadproxy@sha256:551221d79902a5053d1c734b81163d69f087217e2ac13c49bdf6900336ef0786
        - quay.io/kubevirt/virt-operator@sha256:cbca7f93afbb18a840f7e763e4c03aeeb1fd97967f381c12c443cc179b51da29
        - quay.io/kubevirt/hostpath-provisioner@sha256:eae30f37b3e837478a4500144b7bc8eb4f5ddf15f49fa7354bc9d5e240ea8aea
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5ff5392cb762a41ee9bc4767e3a693249ae08b46a4322f623f24ad6b63dd06b5
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:57573a24923e5588bca6bc337a8b2b08406c5b77583974365d2cf063c0dd5d06
        - quay.io/kubevirt/cni-default-plugins@sha256:825e3f9fec1996c54a52cec806154945b38f76476b160d554c36e38dfffe5e61
        - quay.io/kubevirt/kubevirt-template-validator@sha256:a148b6e812b70f326b48d0439b1e44a967a6e1f8cf2f98b1a08e9294a93674eb
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:f64ab5981d9ddef1662dd0a05888783f5dbadda8ede596585ba80bfc97a0f51e
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:f7567bea7f24184da06c3ed87b396c804342fcfd89f54989622780ce0b7b0724
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:3fbcc32bd4e4d15bd93c96def784a229cd84cca27942bf4858b581f31c97ee02
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4dc0b87ccd69f9865b89234d8555d3a614ab0a16ed94a3016ffd27f8106132ce
        - quay.io/kubevirt/virtio-container-disk@sha256:f8e71a3a930ce95d646d80862d5ff6f9d3ed7ef2541dd2e2acec0bce26274734
        - quay.io/kubevirt/virt-api@sha256:707003b221496b4432da2f507d1e36e528b45888b5d321e06d460f0678da44ae
        - quay.io/kubevirt/virt-artifacts-server:1.10.7-unstable
        - quay.io/kubevirt/virt-handler@sha256:138dfda5fea8622f3da0d6413fe214fef80c2fd6a6f9533592a0dbfa7e1865b5
        - quay.io/kubevirt/hostpath-csi-driver@sha256:0ca8b50b0adb20afd4e20da9974200f3290b6d51aa7c1129de6e1b953bf26594
        - quay.io/kubevirt/mtq-controller@sha256:96dd73ea0d7847d9f6322192559747bd85655e25ee25482d112e6e3c6fef21b2
        - quay.io/kubevirt/libguestfs-tools@sha256:b7e248be783691daad1326cf41641d5042be80c5c057ebc5315ca7a1619b0581
        - quay.io/kubevirt/cdi-apiserver@sha256:e9e39408413b1478d2e98eba68913f9e20c93000558b190b47de73bdfd1d9ac4
        - quay.io/kubevirt/ovs-cni-plugin@sha256:08f72edf2bef876bba0b0f5513d30225304ad5e7ad6912a61c083664acdb99ff
        - quay.io/kubevirt/macvtap-cni@sha256:434420511e09b2b5ede785a2c9062b6658ffbc26fbdd4629ce06110f9039c600
        - quay.io/kubevirt/cdi-importer@sha256:3143bbc67cdc6267eb48b7eaac664b8551ac4c11401dfbf4921efd3f233e6ce9
        - ghcr.io/kubevirt/kubesecondarydns@sha256:e87e829380a1e576384145f78ccaa885ba1d5690d5de7d0b73d40cfb804ea24d
        - quay.io/kubevirt/virt-exportproxy@sha256:f14444b0200a85efb4b3c176caefe70aabba7d295e33d2af14027a59ce297e24
        - quay.io/kubevirt/cdi-operator@sha256:6c63521d835578b0fbb77d3145b648f95ab0932d5d1b36ff2068ed8fcd91bc5a
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:16ecabb1d733aa17084cd16d88963c5255d04f389e0f4d41941d67e8be973594
        - quay.io/kubevirt/virt-launcher@sha256:4c5fce3de2e2589197de72fb0c9436490ea318aca952c05a622c43e067023f35
        - quay.io/kubevirt/bridge-marker@sha256:bba066e3b5ff3fb8c5e20861fe8abe51e3c9b50ad6ce3b2616af9cb5479a06d0
        - quay.io/kubevirt/cdi-uploadserver@sha256:30f1827d3696cf996b081c22c3267ca78e7219c872fdb54950198fa54359f6ee
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - quay.io/kubevirt/cdi-controller@sha256:27c47883a08226f83757971d3adafb0cd9bcb26e58fbcf7208236070e0adf37e
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:0ee30557d5d5a42dec19121e70ed77fbdacd8932f55b11e2c7c1a3d5ad5df4aa
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:86c10c122767661e31ee26a9e9de7ea606c2c1168b32793b4ea567ede2ae3b21
        - quay.io/kubevirt/virt-exportserver@sha256:73311f79a9c71007f8572b3cc40cd6f6da404c7ef0a9c6509fb717d979546582
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f6717ce72a2615c7fbc746b4068f788e78579c54c43b8716e5ce650d97af2df1
        - quay.io/kubevirt/kubemacpool@sha256:3214020c21204c189cdf2d767beb8efb61ee3ef65dd8de7f0f1edc61e13e8adf
        - quay.io/kubevirt/ssp-operator@sha256:49cac00844f091d6e80301573e00235d685415185ec29c89b4bd7361f938711e
        - quay.io/kubevirt/mtq-operator@sha256:ef5c4bb1fa72f9e6e75d5585ca352f7563720389e6c9876aa8b07d12d1be6194
        - quay.io/kubevirt/cdi-cloner@sha256:9d31b14f23259398c5bac636f5ead13ad0afd6fe8eeab4499e8e047b4d85074f
        version: 1.10.7
      entries:
      - name: kubevirt-hyperconverged-operator.v1.10.7
        version: 1.10.7
      name: 1.10.7
    - currentCSV: kubevirt-hyperconverged-operator.v1.11.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployTektonTaskResources":true,"deployVmConsoleProxy":true},"tektonPipelines":{"namespace":"kubevirt"},"tektonTasks":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:dd6538e720f18b723b0807037ca7cc1014cba09e57b20706e10e6eb641ac6acc
          createdAt: "2024-03-06 15:11:00"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: quay.io/kubevirt/must-gather
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - quay.io/kubevirt/aaq-server@sha256:0c6b01c48030e03e99b3ff2b319658c2bf2b2001f4e3e030ba70d1143b25717a
        - quay.io/kubevirt/virt-exportserver@sha256:f8473fc824f8bac5f21c64b71cdb4227de28f9bb25c049932f1259216b874a86
        - quay.io/kubevirt/cdi-importer@sha256:7320dce01f4628363992997b592200b4804daf15a9f237cd50a4a6acab7798e8
        - quay.io/kubevirt/virt-artifacts-server@sha256:22fa3218deca38dc98fa0e905df16173a47ccec85707eb91404a367da6480817
        - quay.io/kubevirt/virt-operator@sha256:0e6d56ea0ebdf9bb832f29ab0dc4603f0f619474bb8bc02a01e0a06e9d8d6f4a
        - quay.io/kubevirt/mtq-operator@sha256:90c7608b0ba085582a65ccdc11d089049c5e6b0438d018dd9dc76eb3490383c7
        - quay.io/kubevirt/cni-default-plugins@sha256:c884d6d08f8c0db98964f1eb3877b44ade41fa106083802a9914775df17d5291
        - quay.io/kubevirt/bridge-marker@sha256:bba066e3b5ff3fb8c5e20861fe8abe51e3c9b50ad6ce3b2616af9cb5479a06d0
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:dd6538e720f18b723b0807037ca7cc1014cba09e57b20706e10e6eb641ac6acc
        - quay.io/kubevirt/virt-handler@sha256:238550bb263608ba334b44a15647cbf875482ccfb46049713c14af5d5f0cc788
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:f1bd6ee18c4021c1c94f29edfab89b49b6a4d1b800936c19dbef2d75f8202f2d
        - quay.io/kubevirt/kubemacpool@sha256:cf8daa57ae6603b776d3af512331b143fa03bc2f4b72f28420fddcf5e4156d0a
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:4ed1f37116cbaf883edd8783201d091fbbe121fcb84e7184ee531d3b1abd77ab
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - registry.k8s.io/sig-storage/livenessprobe@sha256:5baeb4a6d7d517434292758928bb33efc6397368cbb48c8a4cf29496abf4e987
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:bf17da674ed6dac95aa4fd7637173fd4e026ebb4523b66f529c0eff394a25bcc
        - quay.io/kubevirt/cdi-uploadserver@sha256:16d257e575b2bd504f148f63b57dc181fa6058d33a11ada64888183c3e4d4863
        - quay.io/kubevirt/virt-exportproxy@sha256:942c548d4848a87945eba51683002797eb4e1f5c463a1873115b2ec0e1e6e169
        - quay.io/kubevirt/ovs-cni-plugin@sha256:e16ac74343da21abb8fb668ce71e728053d00503a992dae2164b9e94a280113e
        - quay.io/kubevirt/hostpath-csi-driver@sha256:8d84b9a33e0ec3749aece7c5f9879d1e0d248e8028ab3f776528a044d387033c
        - ghcr.io/kubevirt/kubesecondarydns@sha256:e87e829380a1e576384145f78ccaa885ba1d5690d5de7d0b73d40cfb804ea24d
        - quay.io/kubevirt/mtq-controller@sha256:ae7dd07e4018c63414b021eba34aa4776a6c7d6ff4fcdf9fd2d7ca30aae98a29
        - quay.io/kubevirt/cdi-apiserver@sha256:95f6e71b36883abcb9ec463cfea2332f5490ec9eb3e4868bdf0a4da769279094
        - quay.io/kubevirt/cdi-controller@sha256:606de89e68be2ac911466e6713345932bda496c1e4f20e34ef5efa270456c9fd
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:dd2948ca8e62d17eafba40bdcaad8367a97f092fd3b8d9849b2d62aec3152a72
        - quay.io/kubevirt/aaq-operator@sha256:590d24742401eeff3fb6a9c8ed977c63442f8f0a0505509fae69aa1bf8aa84ae
        - quay.io/kubevirt/macvtap-cni@sha256:850b89343ace7c7ea6b18dd8e11964613974e9d1f7377af03854d407fb15230a
        - quay.io/kubevirt/virtio-container-disk@sha256:f8e71a3a930ce95d646d80862d5ff6f9d3ed7ef2541dd2e2acec0bce26274734
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:3fbcc32bd4e4d15bd93c96def784a229cd84cca27942bf4858b581f31c97ee02
        - quay.io/kubevirt/virt-controller@sha256:4f576e30b83db7b9f96dde77896898bde87db3517963c4227634d0c3e96bb86d
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:70efc16ab2bfec04f73ccc292064d080870659cfd47c2e0c9017ff7b6fd78863
        - quay.io/kubevirt/cdi-operator@sha256:8a27e1450c3bf8c1d8924e74470a62419d02ed2202924b082a74e2414b1fc21a
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:c53535af8a7f7e3164609838c4b191b42b2d81238d75c1b2a2b582ada62a9780
        - quay.io/kubevirt/libguestfs-tools@sha256:4971af1ee42959d366a3db6129ab684d11064e00bbd3c2b36324ad14e90cb903
        - quay.io/kubevirt/cdi-cloner@sha256:b22985af0ee7b1ce4e1dedded08c853d90d9e7209f98c5d470897bbefce2a063
        - quay.io/kubevirt/ssp-operator@sha256:d57ef05fea09c4690dbe728ebe6a39d9df1a3574dfa7d18346bf3fe66afed702
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:57573a24923e5588bca6bc337a8b2b08406c5b77583974365d2cf063c0dd5d06
        - quay.io/kubevirt/virt-launcher@sha256:4fbfd76561667e98fafc75c0b804f77df251f96471d1121d69f149cf8fb9a4d8
        - quay.io/kubevirt/kubevirt-template-validator@sha256:f6db539fe4d4b752644c6bc8a308f5a80ae09a71da5c5d535114dc4c3cf18e0a
        - quay.io/kubevirt/aaq-controller@sha256:cb2fa52ae3f6a1a99b2de77b58c5db54c43ba8ee86c6540b4844c901180ce611
        - quay.io/kubevirt/virt-api@sha256:735662c824da3d6069624d40674163d758c2256f8043cf6f02f505c3a724afa3
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:e2def4213ec0657e72eb790ae8a115511d5b8f164a62d3568d2f1bff189917e8
        - quay.io/kubevirt/mtq-lock-server@sha256:8e686107dd9990dab1a3d90ee1068541938f673bf54ea9e1308b2ed7218546b0
        - quay.io/kubevirt/hostpath-provisioner@sha256:adb1d33e95e8991dc47107926172469b5ecbeace509391ce021ca42d305a41d3
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:38293a1050bce9c7cb786cbe8719be1edb91a922a6c583eedb61e56729ed4c72
        - quay.io/kubevirt/cdi-uploadproxy@sha256:b89104ef5513d63cab43b34ac52f752c3d935cb986d5527d920c2c31d3b6f239
        version: 1.11.0
      entries:
      - name: kubevirt-hyperconverged-operator.v1.11.0
        version: 1.11.0
      name: 1.11.0
    - currentCSV: kubevirt-hyperconverged-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:7afeef450521fb7bbc918e874d9297dc3762cdeb28db0bef96d66d48b8752a9d
          createdAt: "2024-08-07 15:33:27"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: quay.io/kubevirt/must-gather
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:7afeef450521fb7bbc918e874d9297dc3762cdeb28db0bef96d66d48b8752a9d
        - quay.io/kubevirt/virt-artifacts-server@sha256:cc3e1e17bc3fa2a895f934210c9106bee01c3f9bf001f988d920eb58628e347b
        - quay.io/kubevirt/virt-operator@sha256:1ecc62c25f869e17027c930150d5c9b52b72d7e1fb434376bc0be0d886952b2e
        - quay.io/kubevirt/aaq-operator@sha256:002513b78dfd795f5195534b651eace136ead3d73c8ac8cae145607e79c8a8ba
        - quay.io/kubevirt/hostpath-csi-driver@sha256:b7ccc469ad7e640fed2e98b50eae57a64e81c7b9862a799b7a9892d14ade8937
        - quay.io/kubevirt/virt-controller@sha256:3471b00eb9f4774ae33867c7d0b83e1d26a065a2bd19d305b5d3946ab665b5e9
        - quay.io/kubevirt/aaq-controller@sha256:170abd7a75397ddce07ba0d64886ead25183c49cacbccf0d191a4dd37ca013fd
        - quay.io/kubevirt/ssp-operator@sha256:a15a53d0f3ef530fe469acf3b7673486011605d0021d4dbe29af94096b3b65ee
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:413a6c229ebb608868e8b2c7c06b251c6b1d43c07af12a76b893292cbed8bef9
        - quay.io/kubevirt/ovs-cni-plugin@sha256:e16ac74343da21abb8fb668ce71e728053d00503a992dae2164b9e94a280113e
        - quay.io/kubevirt/kubevirt-template-validator@sha256:8f7c389a7cd7ef34f2edc84245bc1d0ebcf13e86da5e0b8893f56733c482b10d
        - quay.io/kubevirt/hostpath-provisioner@sha256:2b93b6c4ce0307a8f0dbdae2226da4acb1bae781788fbde7f0c32d6ad6421aa0
        - quay.io/kubevirt/mtq-lock-server@sha256:8b5189518ac6f68161d6a259e2a6f70e7902855ea1510aa260e79093520a7803
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac
        - quay.io/kubevirt/libguestfs-tools@sha256:0de3de89720d4e5433a33944f00dac590daed6ab12e4d26a657ef082d0573747
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:e2def4213ec0657e72eb790ae8a115511d5b8f164a62d3568d2f1bff189917e8
        - quay.io/kubevirt/mtq-operator@sha256:859524721ae1a3b1e5dd6e29036b4c80acff13f74a4ee6769d867283e9ccbba5
        - registry.k8s.io/sig-storage/livenessprobe@sha256:5baeb4a6d7d517434292758928bb33efc6397368cbb48c8a4cf29496abf4e987
        - quay.io/kubevirt/virtio-container-disk@sha256:f8e71a3a930ce95d646d80862d5ff6f9d3ed7ef2541dd2e2acec0bce26274734
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:935475c2850466aa5ac57e4de627fb177515cb2c402a95842ead095d82b6df5f
        - quay.io/kubevirt/mtq-controller@sha256:04b38720b65479046270c8132cff5ab0f2dabc9cae91de1cc95fecf167efc356
        - quay.io/kubevirt/virt-exportserver@sha256:acac33a19528592cc3ef0cd3a25fa4c72bcf992e4c2c247437007fcb9de0aa6d
        - quay.io/kubevirt/cdi-uploadproxy@sha256:86285df5cf52533732ddf4fdb74674cbce877ba69833cb8ed058c2a157cb5573
        - quay.io/kubevirt/cdi-cloner@sha256:05da8df0b64024f9e2a48a1a54f06c821d79a3efe7828fad2c4fda6a01d0cd86
        - quay.io/kubevirt/cni-default-plugins@sha256:0c354fa9d695b8cab97b459e8afea2f7662407a987e83f6f6f1a8af4b45726be
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:3c70cb833819d8346e5531b10ac516283bab7098e782708768a6fe00ae0ddfac
        - quay.io/kubevirt/cdi-operator@sha256:f68e4f2a082fc72d92bc1df0340285af03cd12d8c7721b38c619d1de0f4c175a
        - ghcr.io/kubevirt/kubesecondarydns@sha256:6268d84154e2483fbce8c1adacbdaf6f0839117b2d48d9fa4687cc8f76bd5130
        - quay.io/kubevirt/virt-api@sha256:aeeae94c9dd3b352ff2fbcdba77a6b084fb4ebe0fed35fe7af9481e9e2b1f24c
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:83b460502671fb4f34116363a1a39b2ddfc9d14a920ee0a6413bfc3bd0580404
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - quay.io/kubevirt/cdi-apiserver@sha256:4ba004f57e7d589ddb5372a9729360460a97effecc6e5ecaccd7fbf16c6fcf13
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:d51821f001c959a79943cb22f5ea47589d7cf5c032e11ec1d571e8db7bf29b78
        - quay.io/kubevirt/virt-exportproxy@sha256:dd3b5c12314fa76b88d55d978210de81562fb888351748ea3997551c8c96a877
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:c4b6b02737bc24906fcce57fe6626d1a36cb2b91baa971af2a5e5a919093c34e
        - quay.io/kubevirt/cdi-importer@sha256:de5b89c879a531165482d540deffdc2dec0c188fd0f7b5fbd49c0298cf9a5b1d
        - quay.io/kubevirt/kubemacpool@sha256:20e156be33e6d3692c456081acbb91a4349f94de448f7d1f1cddd0228931b31f
        - quay.io/kubevirt/virt-launcher@sha256:cd37543ed2bc6baadbcbffa6951d740e4c65dc0ea366dec0a8593193c384e6de
        - quay.io/kubevirt/virt-handler@sha256:2a7d96da4de3ea51339b3a047abb4b05e088e3d1080110b42a85ea5baff4a0cc
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:246b3819cfa3833f9a8cb01e8cb859aa61a1924a85ea4a112e43e8414f32c429
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:3fbcc32bd4e4d15bd93c96def784a229cd84cca27942bf4858b581f31c97ee02
        - quay.io/kubevirt/cdi-controller@sha256:392ead2fdc5050c584765ebf99706cc5c0ab9c2b424b692ad35005e77d110b58
        - quay.io/kubevirt/cdi-uploadserver@sha256:d7fb653bb33dade84583fff6048092346639532d6675aaf9f88070171cb0fe7c
        - quay.io/kubevirt/bridge-marker@sha256:bba066e3b5ff3fb8c5e20861fe8abe51e3c9b50ad6ce3b2616af9cb5479a06d0
        - quay.io/kubevirt/aaq-server@sha256:9db6ca9147b1d75ba4d6bb1326c4ec289327eb9f8aadbcd951fcc689e584cd0d
        - quay.io/kubevirt/macvtap-cni@sha256:850b89343ace7c7ea6b18dd8e11964613974e9d1f7377af03854d407fb15230a
        version: 1.12.0
      entries:
      - name: kubevirt-hyperconverged-operator.v1.12.0
        version: 1.12.0
      - name: kubevirt-hyperconverged-operator.v1.11.1
        version: 1.11.1
      - name: kubevirt-hyperconverged-operator.v1.10.0
        version: 1.10.0
      - name: kubevirt-hyperconverged-operator.v1.9.0
        version: 1.9.0
      - name: kubevirt-hyperconverged-operator.v1.8.2
        version: 1.8.2
      - name: kubevirt-hyperconverged-operator.v1.8.1
        version: 1.8.1
      - name: kubevirt-hyperconverged-operator.v1.8.0
        version: 1.8.0
      - name: kubevirt-hyperconverged-operator.v1.7.0
        version: 1.7.0
      - name: kubevirt-hyperconverged-operator.v1.6.0
        version: 1.6.0
      - name: kubevirt-hyperconverged-operator.v1.5.2
        version: 1.5.2
      - name: kubevirt-hyperconverged-operator.v1.5.1
        version: 1.5.1
      - name: kubevirt-hyperconverged-operator.v1.5.0
        version: 1.5.0
      - name: kubevirt-hyperconverged-operator.v1.4.4
        version: 1.4.4
      - name: kubevirt-hyperconverged-operator.v1.4.3
        version: 1.4.3
      - name: kubevirt-hyperconverged-operator.v1.4.2
        version: 1.4.2
      - name: kubevirt-hyperconverged-operator.v1.4.1
        version: 1.4.1
      - name: kubevirt-hyperconverged-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: community-kubevirt-hyperconverged
    provider:
      name: KubeVirt project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: StreamNative
      provider-url: https://streamnative.io
    name: sn-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
        - docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21
        - docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      name: alpha
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
        - docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21
        - docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      name: beta
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21
        - quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
        - docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      name: stable
    defaultChannel: alpha
    packageName: sn-operator
    provider:
      name: StreamNative
      url: https://streamnative.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzo-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzo-operator.v2.0.101
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzo.cambridgesemantics.com/v1",
                "kind": "Anzo",
                "metadata": {
                  "name": "agent01"
                },
                "spec": {
                  "nodeConfig": {
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "app": "anzo"
                        }
                      },
                      "serviceName": "anzo-agent01",
                      "template": {
                        "metadata": {
                          "labels": {
                            "app": "anzo"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo@sha256:d1c2d54ca205d669c5ea1bb6aae105afe2385cc5695cecd78423ffa1d98929f8",
                              "name": "anzo",
                              "resources": {
                                "limits": {
                                  "cpu": "4000m",
                                  "memory": "12Gi"
                                },
                                "requests": {
                                  "cpu": "4000m",
                                  "memory": "12Gi"
                                }
                              }
                            }
                          ],
                          "serviceAccountName": "anzo-operator"
                        }
                      }
                    }
                  },
                  "role": "AnzoAgent"
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-operator@sha256:563e964f5771b9bfd860ea993b6e0936a54504908f5b2d1dff6f99ade74bb7a0
          createdAt: "2021-08-02"
          description: kubernetes operator for Anzo
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzo-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzo-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Anzo is the Schema for the anzos API
            displayName: Anzo Operator
            kind: Anzo
            name: anzos.anzo.cambridgesemantics.com
            version: v1
        description: |-
          The Anzo Operator provides the way to install and configure an anzo agent setup on Red Hat K8S environment.
          Currently, this is supported only through when deployed as an Anzo Agent as part of an Anzo Unstructured deployment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v2.0.0/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers through [ Cambridge Semantics Customer Center ]( https://customercenter.cambridgesemantics.com/ ).
        displayName: Anzo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        links:
        - name: Anzo Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v2.0.0/README_openshift.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-operator@sha256:563e964f5771b9bfd860ea993b6e0936a54504908f5b2d1dff6f99ade74bb7a0
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo@sha256:d1c2d54ca205d669c5ea1bb6aae105afe2385cc5695cecd78423ffa1d98929f8
        version: 2.0.0
      entries:
      - name: anzo-operator.v2.0.101
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anzo-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: odr-hub-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odr-hub-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east"
                },
                "spec": {
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east1"
                },
                "spec": {
                  "cidrs": [
                    "192.168.39.45/32"
                  ],
                  "clusterFence": "ManuallyFenced",
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east1"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPlacementControl",
                "metadata": {
                  "name": "drplacementcontrol-sample",
                  "namespace": "application-namespace"
                },
                "spec": {
                  "drPolicyRef": {
                    "name": "drpolicy-sample"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "excludedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "includedResources": [
                          "*"
                        ]
                      }
                    ]
                  },
                  "placementRef": {
                    "kind": "PlacementRule",
                    "name": "application-placement-rule"
                  },
                  "preferredCluster": "east",
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east",
                    "west"
                  ],
                  "replicationClassSelector": {
                    "matchLabels": {
                      "class": "ramen"
                    }
                  },
                  "schedulingInterval": "1h"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "metrodr-drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east1",
                    "east2"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DRCluster is the Schema for the drclusters API
            displayName: DRCluster
            kind: DRCluster
            name: drclusters.ramendr.openshift.io
            version: v1alpha1
          - description: DRPlacementControl is the Schema for the drplacementcontrols
              API
            displayName: DRPlacement Control
            kind: DRPlacementControl
            name: drplacementcontrols.ramendr.openshift.io
            version: v1alpha1
          - description: DRPolicy is the Schema for the drpolicies API
            displayName: DRPolicy
            kind: DRPolicy
            name: drpolicies.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Hub is a disaster-recovery orchestrator for stateful
          applications. It operates from an Advanced Cluster Management (ACM) hub
          cluster to orchestrate the recovery of application state, and scheduling
          of ACM PlacementRule for disaster recovery operations.
        displayName: Openshift DR Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:769acf15d016d4992793c9c527d7c268d2a67502dd7e0fadc240e4896398195b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 4.15.20-rhodf
      entries:
      - name: odr-hub-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odr-hub-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odr-hub-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odr-hub-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odr-hub-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odr-hub-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odr-hub-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odr-hub-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odr-hub-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odr-hub-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odr-hub-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odr-hub-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odr-hub-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odr-hub-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odr-hub-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odr-hub-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odr-hub-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odr-hub-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odr-hub-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odr-hub-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odr-hub-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east"
                },
                "spec": {
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east1"
                },
                "spec": {
                  "cidrs": [
                    "192.168.39.45/32"
                  ],
                  "clusterFence": "ManuallyFenced",
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east1"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPlacementControl",
                "metadata": {
                  "name": "drplacementcontrol-sample",
                  "namespace": "application-namespace"
                },
                "spec": {
                  "drPolicyRef": {
                    "name": "drpolicy-sample"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "excludedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "includedResources": [
                          "*"
                        ]
                      }
                    ]
                  },
                  "placementRef": {
                    "kind": "PlacementRule",
                    "name": "application-placement-rule"
                  },
                  "preferredCluster": "east",
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east",
                    "west"
                  ],
                  "replicationClassSelector": {
                    "matchLabels": {
                      "class": "ramen-replication-class"
                    }
                  },
                  "schedulingInterval": "1h",
                  "volumeSnapshotClassSelector": {
                    "matchLabels": {
                      "class": "ramen-snapshot-class"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "metrodr-drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east1",
                    "east2"
                  ],
                  "replicationClassSelector": {},
                  "schedulingInterval": {},
                  "volumeSnapshotClassSelector": {}
                }
              }
            ]
          capabilities: Deep Insights
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DRCluster is the Schema for the drclusters API
            displayName: DRCluster
            kind: DRCluster
            name: drclusters.ramendr.openshift.io
            version: v1alpha1
          - description: DRPlacementControl is the Schema for the drplacementcontrols
              API
            displayName: DRPlacement Control
            kind: DRPlacementControl
            name: drplacementcontrols.ramendr.openshift.io
            version: v1alpha1
          - description: DRPolicy is the Schema for the drpolicies API
            displayName: DRPolicy
            kind: DRPolicy
            name: drpolicies.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Hub is a disaster-recovery orchestrator for stateful
          applications. It operates from an Advanced Cluster Management (ACM) hub
          cluster to orchestrate the recovery of application state, and scheduling
          of ACM PlacementRule for disaster recovery operations.
        displayName: Openshift DR Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:af262b0890300183db1766bc392017b0bda723a9ba841dc25e3d928aa413db9a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        version: 4.16.16-rhodf
      entries:
      - name: odr-hub-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odr-hub-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odr-hub-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odr-hub-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odr-hub-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odr-hub-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odr-hub-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odr-hub-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odr-hub-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odr-hub-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odr-hub-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odr-hub-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odr-hub-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odr-hub-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odr-hub-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odr-hub-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odr-hub-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odr-hub-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: container-security-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: container-security-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secscan.quay.redhat.com/v1alpha1",
                "kind": "ImageManifestVuln",
                "metadata": {
                  "selfLink": "/apis/secscan.quay.redhat.com/v1alpha1/namespaces/openshift-cluster-version/imagemanifestvulns/sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18",
                  "resourceVersion": "14565395",
                  "name": "sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18",
                  "uid": "3f16a188-e69a-406c-aa06-df340195409c",
                  "creationTimestamp": "2021-03-17T12:07:16Z",
                  "generation": 2,
                  "managedFields": [
                    {
                      "apiVersion": "secscan.quay.redhat.com/v1alpha1",
                      "fieldsType": "FieldsV1",
                      "fieldsV1": {
                        "f:metadata": {
                          "f:labels": {
                            ".": {},
                            "f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": {}
                          }
                        },
                        "f:spec": {
                          ".": {},
                          "f:features": {},
                          "f:image": {},
                          "f:manifest": {}
                        },
                        "f:status": {
                          ".": {},
                          "f:affectedPods": {
                            ".": {},
                            "f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": {}
                          },
                          "f:fixableCount": {},
                          "f:highCount": {},
                          "f:highestSeverity": {},
                          "f:lastUpdate": {},
                          "f:lowCount": {},
                          "f:mediumCount": {}
                        }
                      },
                      "manager": "security-labeller",
                      "operation": "Update",
                      "time": "2021-03-17T13:07:20Z"
                    }
                  ],
                  "namespace": "openshift-cluster-version",
                  "labels": {
                    "openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": "true"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:447e05859134a0455ca55f8aae16c8a27ce06af4c18c191b41f6339df553f1bf
          createdAt: 2021-11-02 11:11 UTC
          description: Identify image vulnerabilities in Kubernetes pods
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Quay Container Security Operator (CSO) brings Quay and Clair
          metadata to Kubernetes / OpenShift. Starting with vulnerability information
          the scope will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Quay Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Quay Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:447e05859134a0455ca55f8aae16c8a27ce06af4c18c191b41f6339df553f1bf
        version: 3.4.7
      entries:
      - name: container-security-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: container-security-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"secscan.quay.redhat.com/v1alpha1\",\n
            \   \"kind\": \"ImageManifestVuln\",\n    \"metadata\": {\n      \"selfLink\":
            \"/apis/secscan.quay.redhat.com/v1alpha1/namespaces/openshift-cluster-version/imagemanifestvulns/sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18\",\n
            \     \"resourceVersion\": \"14565395\",\n      \"name\": \"sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18\",\n
            \     \"uid\": \"3f16a188-e69a-406c-aa06-df340195409c\",\n      \"creationTimestamp\":
            \"2021-03-17T12:07:16Z\",\n      \"generation\": 2,\n      \"managedFields\":
            [\n        {\n          \"apiVersion\": \"secscan.quay.redhat.com/v1alpha1\",\n
            \         \"fieldsType\": \"FieldsV1\",\n          \"fieldsV1\": {\n            \"f:metadata\":
            {\n              \"f:labels\": {\n                \".\": {},\n                \"f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            {}\n              }\n            },\n            \"f:spec\": {\n              \".\":
            {},\n              \"f:features\": {},\n              \"f:image\": {},\n
            \             \"f:manifest\": {}\n            },\n            \"f:status\":
            {\n              \".\": {},\n              \"f:affectedPods\": {\n                \".\":
            {},\n                \"f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            {}\n              },\n              \"f:fixableCount\": {},\n              \"f:highCount\":
            {},\n              \"f:highestSeverity\": {},\n              \"f:lastUpdate\":
            {},\n              \"f:lowCount\": {},\n              \"f:mediumCount\":
            {}\n            }\n          },\n          \"manager\": \"security-labeller\",\n
            \         \"operation\": \"Update\",\n          \"time\": \"2021-03-17T13:07:20Z\"\n
            \       }\n      ],\n      \"namespace\": \"openshift-cluster-version\",\n
            \     \"labels\": {\n        \"openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            \"true\"\n      }\n    },\n    \"spec\": {\n      \"features\": [\n        {\n
            \         \"name\": \"musl\",\n          \"namespaceName\": \"alpine:v3.14\",\n
            \         \"version\": \"1.2.2-r3\",\n          \"versionformat\": \"dpkg\",\n
            \         \"vulnerabilities\": [\n            {\n              \"fixedby\":
            \"1.2.2_pre2-r0\",\n              \"link\": \"https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28928\",\n
            \             \"metadata\": \"{\\\"NVD\\\": {\\\"CVSSv3\\\": {\\\"Vectors\\\":
            \\\"CVSS:3.0/AV:L/AC:L/PR:L/UI:N/S:U/C:N/I:N/A:H\\\", \\\"ImpactScore\\\":
            3.6, \\\"Score\\\": 5.5, \\\"ExploitabilityScore\\\": 1.8}, \\\"CVSSv2\\\":
            {\\\"Score\\\": 2.1, \\\"PublishedDateTime\\\": \\\"2020-11-24T18:15Z\\\",
            \\\"Vectors\\\": \\\"AV:L/AC:L/Au:N/C:N/I:N/A:P\\\"}}}\",\n              \"name\":
            \"CVE-2020-28928\",\n              \"namespaceName\": \"alpine:v3.14\",\n
            \             \"severity\": \"Low\"\n            }\n          ]\n        }\n
            \     ],\n      \"image\": \"quay.io/harishg/pruned-redhat-operators-index\",\n
            \     \"manifest\": \"sha256:19631c86642609254777ff59fc9e128aa5d0a94a7dbb504ef1fcfb2bbc6e486e\"\n
            \   }\n  }\n]    "
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:6b3581e62550d6a615d36eee6eaf4c89b9c35ba53cab8433bd93c68393a9d115
          createdAt: 2021-09-10 23:09 UTC
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.4.x <3.5.7'
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Quay Container Security Operator (CSO) brings Quay and Clair
          metadata to Kubernetes / OpenShift. Starting with vulnerability information
          the scope will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Quay Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Quay Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:6b3581e62550d6a615d36eee6eaf4c89b9c35ba53cab8433bd93c68393a9d115
        version: 3.5.7
      entries:
      - name: container-security-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: container-security-operator.v3.10.15
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:de7c66a0004f3c0643c09d4c3bbb4a0f1f8f2cbf264250802cdc58028fc72b50
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.10.15
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:de7c66a0004f3c0643c09d4c3bbb4a0f1f8f2cbf264250802cdc58028fc72b50
        version: 3.10.15
      entries:
      - name: container-security-operator.v3.10.15
        version: 3.10.15
      - name: container-security-operator.v3.10.14
        version: 3.10.14
      - name: container-security-operator.v3.10.13
        version: 3.10.13
      - name: container-security-operator.v3.10.12
        version: 3.10.12
      - name: container-security-operator.v3.10.11
        version: 3.10.11
      - name: container-security-operator.v3.10.10
        version: 3.10.10
      - name: container-security-operator.v3.10.9
        version: 3.10.9
      - name: container-security-operator.v3.10.8
        version: 3.10.8
      - name: container-security-operator.v3.10.7
        version: 3.10.7
      - name: container-security-operator.v3.10.6
        version: 3.10.6
      - name: container-security-operator.v3.10.5
        version: 3.10.5
      - name: container-security-operator.v3.10.4
        version: 3.10.4
      - name: container-security-operator.v3.10.3
        version: 3.10.3
      - name: container-security-operator.v3.10.2
        version: 3.10.2
      - name: container-security-operator.v3.10.1
        version: 3.10.1
      - name: container-security-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: container-security-operator.v3.11.13
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:3914389fd53c371627921298c5dc3564999d92bc720f27a240f74c034ebbf5ec
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.11.13
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:3914389fd53c371627921298c5dc3564999d92bc720f27a240f74c034ebbf5ec
        version: 3.11.13
      entries:
      - name: container-security-operator.v3.11.13
        version: 3.11.13
      - name: container-security-operator.v3.11.12
        version: 3.11.12
      - name: container-security-operator.v3.11.11
        version: 3.11.11
      - name: container-security-operator.v3.11.10
        version: 3.11.10
      - name: container-security-operator.v3.11.9
        version: 3.11.9
      - name: container-security-operator.v3.11.8
        version: 3.11.8
      - name: container-security-operator.v3.11.7
        version: 3.11.7
      - name: container-security-operator.v3.11.6
        version: 3.11.6
      - name: container-security-operator.v3.11.5
        version: 3.11.5
      - name: container-security-operator.v3.11.4
        version: 3.11.4
      - name: container-security-operator.v3.11.3
        version: 3.11.3
      - name: container-security-operator.v3.11.2
        version: 3.11.2
      - name: container-security-operator.v3.11.1
        version: 3.11.1
      - name: container-security-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: container-security-operator.v3.12.12
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:1a15328f0b2c64d3a857f2f155ad4def095027e809610de98545e303eec63225
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.12.12
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:1a15328f0b2c64d3a857f2f155ad4def095027e809610de98545e303eec63225
        version: 3.12.12
      entries:
      - name: container-security-operator.v3.12.12
        version: 3.12.12
      - name: container-security-operator.v3.12.11
        version: 3.12.11
      - name: container-security-operator.v3.12.10
        version: 3.12.10
      - name: container-security-operator.v3.12.9
        version: 3.12.9
      - name: container-security-operator.v3.12.8
        version: 3.12.8
      - name: container-security-operator.v3.12.7
        version: 3.12.7
      - name: container-security-operator.v3.12.6
        version: 3.12.6
      - name: container-security-operator.v3.12.5
        version: 3.12.5
      - name: container-security-operator.v3.12.4
        version: 3.12.4
      - name: container-security-operator.v3.12.3
        version: 3.12.3
      - name: container-security-operator.v3.12.2
        version: 3.12.2
      - name: container-security-operator.v3.12.1
        version: 3.12.1
      - name: container-security-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: container-security-operator.v3.13.8
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:ba50621739581da9a661c2c2e355849c328cfa33b7311b06abb148e1183a6321
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.13.8
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:ba50621739581da9a661c2c2e355849c328cfa33b7311b06abb148e1183a6321
        version: 3.13.8
      entries:
      - name: container-security-operator.v3.13.8
        version: 3.13.8
      - name: container-security-operator.v3.13.7
        version: 3.13.7
      - name: container-security-operator.v3.13.6
        version: 3.13.6
      - name: container-security-operator.v3.13.5
        version: 3.13.5
      - name: container-security-operator.v3.13.4
        version: 3.13.4
      - name: container-security-operator.v3.13.3
        version: 3.13.3
      - name: container-security-operator.v3.13.2
        version: 3.13.2
      - name: container-security-operator.v3.13.1
        version: 3.13.1
      - name: container-security-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: container-security-operator.v3.14.5
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:c42487ccf098df3ec6dbeb5d409346bdd30e92781cf3ffea5fdb6dd0b196d0cf
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.14.5
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:c42487ccf098df3ec6dbeb5d409346bdd30e92781cf3ffea5fdb6dd0b196d0cf
        version: 3.14.5
      entries:
      - name: container-security-operator.v3.14.5
        version: 3.14.5
      - name: container-security-operator.v3.14.4
        version: 3.14.4
      - name: container-security-operator.v3.14.3
        version: 3.14.3
      - name: container-security-operator.v3.14.2
        version: 3.14.2
      - name: container-security-operator.v3.14.1
        version: 3.14.1
      - name: container-security-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: container-security-operator.v3.15.2
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:81e929ec095f791a83310ee7fcda239d5c348ba69a3cf5adcf1b3a77c349ce77
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.15.2
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:81e929ec095f791a83310ee7fcda239d5c348ba69a3cf5adcf1b3a77c349ce77
        version: 3.15.2
      entries:
      - name: container-security-operator.v3.15.2
        version: 3.15.2
      - name: container-security-operator.v3.15.1
        version: 3.15.1
      - name: container-security-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: container-security-operator.v3.6.10
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:641b80df8d7102d49a5dfd5b99245669f60393880a6f1bc8ba8450b88a640d96
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.6.10
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:641b80df8d7102d49a5dfd5b99245669f60393880a6f1bc8ba8450b88a640d96
        version: 3.6.10
      entries:
      - name: container-security-operator.v3.6.10
        version: 3.6.10
      - name: container-security-operator.v3.6.9
        version: 3.6.9
      - name: container-security-operator.v3.6.8
        version: 3.6.8
      - name: container-security-operator.v3.6.7
        version: 3.6.7
      - name: container-security-operator.v3.6.6
        version: 3.6.6
      - name: container-security-operator.v3.6.5
        version: 3.6.5
      - name: container-security-operator.v3.6.4
        version: 3.6.4
      - name: container-security-operator.v3.6.2
        version: 3.6.2
      - name: container-security-operator.v3.6.1
        version: 3.6.1
      - name: container-security-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: container-security-operator.v3.7.14
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:10736a06beb503b4eb73260a87a2cde9bd83353374ee0ac9e8dd0a6260cf2665
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.7.14
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:10736a06beb503b4eb73260a87a2cde9bd83353374ee0ac9e8dd0a6260cf2665
        version: 3.7.14
      entries:
      - name: container-security-operator.v3.7.14
        version: 3.7.14
      - name: container-security-operator.v3.7.13
        version: 3.7.13
      - name: container-security-operator.v3.7.12
        version: 3.7.12
      - name: container-security-operator.v3.7.11
        version: 3.7.11
      - name: container-security-operator.v3.7.10
        version: 3.7.10
      - name: container-security-operator.v3.7.9
        version: 3.7.9
      - name: container-security-operator.v3.7.8
        version: 3.7.8
      - name: container-security-operator.v3.7.7
        version: 3.7.7
      - name: container-security-operator.v3.7.6
        version: 3.7.6
      - name: container-security-operator.v3.7.5
        version: 3.7.5
      - name: container-security-operator.v3.7.4
        version: 3.7.4
      - name: container-security-operator.v3.7.3
        version: 3.7.3
      - name: container-security-operator.v3.7.2
        version: 3.7.2
      - name: container-security-operator.v3.7.1
        version: 3.7.1
      - name: container-security-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: container-security-operator.v3.8.15
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:108bd033543beb7cce077130c96f992c202f7a258229c239795f8f19e5fae5e5
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.8.15
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:108bd033543beb7cce077130c96f992c202f7a258229c239795f8f19e5fae5e5
        version: 3.8.15
      entries:
      - name: container-security-operator.v3.8.15
        version: 3.8.15
      - name: container-security-operator.v3.8.14
        version: 3.8.14
      - name: container-security-operator.v3.8.13
        version: 3.8.13
      - name: container-security-operator.v3.8.12
        version: 3.8.12
      - name: container-security-operator.v3.8.11
        version: 3.8.11
      - name: container-security-operator.v3.8.10
        version: 3.8.10
      - name: container-security-operator.v3.8.9
        version: 3.8.9
      - name: container-security-operator.v3.8.8
        version: 3.8.8
      - name: container-security-operator.v3.8.7
        version: 3.8.7
      - name: container-security-operator.v3.8.6
        version: 3.8.6
      - name: container-security-operator.v3.8.5
        version: 3.8.5
      - name: container-security-operator.v3.8.4
        version: 3.8.4
      - name: container-security-operator.v3.8.3
        version: 3.8.3
      - name: container-security-operator.v3.8.2
        version: 3.8.2
      - name: container-security-operator.v3.8.1
        version: 3.8.1
      - name: container-security-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: container-security-operator.v3.9.16
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:df8e1d246273d21049c6bb8169a2e5f3fc8d69c85c564359abea7364969c6acc
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.9.16
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:df8e1d246273d21049c6bb8169a2e5f3fc8d69c85c564359abea7364969c6acc
        version: 3.9.16
      entries:
      - name: container-security-operator.v3.9.16
        version: 3.9.16
      - name: container-security-operator.v3.9.15
        version: 3.9.15
      - name: container-security-operator.v3.9.14
        version: 3.9.14
      - name: container-security-operator.v3.9.13
        version: 3.9.13
      - name: container-security-operator.v3.9.12
        version: 3.9.12
      - name: container-security-operator.v3.9.11
        version: 3.9.11
      - name: container-security-operator.v3.9.10
        version: 3.9.10
      - name: container-security-operator.v3.9.9
        version: 3.9.9
      - name: container-security-operator.v3.9.8
        version: 3.9.8
      - name: container-security-operator.v3.9.7
        version: 3.9.7
      - name: container-security-operator.v3.9.6
        version: 3.9.6
      - name: container-security-operator.v3.9.5
        version: 3.9.5
      - name: container-security-operator.v3.9.4
        version: 3.9.4
      - name: container-security-operator.v3.9.3
        version: 3.9.3
      - name: container-security-operator.v3.9.2
        version: 3.9.2
      - name: container-security-operator.v3.9.1
        version: 3.9.1
      - name: container-security-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: container-security-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: AppDynamics LLC
      provider-url: https://www.appdynamics.com/
    name: appdynamics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appdynamics-operator.v24.7.0-861
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clusteragent",
                "metadata": {
                  "name": "k8s-cluster-agent",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "appName": "<app-name>",
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/cluster-agent:tag"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "InfraViz",
                "metadata": {
                  "name": "appdynamics-infraviz",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/machine-agent:latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3c420cce8b626d8f5cf424410527c6a7a17ca3348e903dd8cdfcd74819b8de41
          createdAt: "2023-02-06T14:26:07Z"
          description: AppDynamics Operator to install AppDynamics Cluster Agent &
            AppDynamics Machine Agent for K8s Monitoring for various Cloud Providers
            like EKS, AKS etc.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: support@appdynamics.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clusteragent is the Schema for the clusteragents API
            displayName: Clusteragent
            kind: Clusteragent
            name: clusteragents.cluster.appdynamics.com
            version: v1alpha1
          - description: InfraViz is the Schema for the infravizs API
            displayName: Infra Viz
            kind: InfraViz
            name: infravizs.cluster.appdynamics.com
            version: v1alpha1
        description: |-
          AppDynamics Operator to install AppDynamics Cluster Agent & Machine Agent for K8s Monitoring on various Cloud Providers like EKS, AKS, GCP, Openshift etc.
          **About AppDynamics Operator**
          AppDynamics Operator simplifies the installation and lifecycle management of the AppDynamics ClusterAgent and the AppDynamics Machine Agent for K8s Monitoring on different Cloud Providers like EKS, AKS, GCP, OpenShift etc.
          **Installation Steps**
          Please refer below links for installation guide:
          [Install AppDynamics Cluster Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli)
          [Install AppDynamics Machine Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli)
          [Install AppDynamics Cluster Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts)
          [Install Appdynamics Machine Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts)
        displayName: AppDynamics Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AppDynamics
        - Operator
        - AppDynamics Cluster Agent
        - AppDynamics Machine Agent
        - AppDynamics K8s Monitoring
        links:
        - name: AppDynamics Kubernetes Monitoring Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent
        - name: AppDynamics Cluster Agent Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/overview-of-cluster-monitoring
        - name: AppDynamics Cluster Agent Requirements
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/cluster-agent-requirements-and-supported-environments
        - name: Install AppDynamics Cluster Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli
        - name: Install AppDynamics Cluster Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts
        - name: Install AppDynamics Machine Agent & NetViz Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli
        - name: Install AppDynamics Machine Agent & NetViz Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: AppDynamics LLC
          url: https://www.appdynamics.com/
        relatedImages:
        - registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3c420cce8b626d8f5cf424410527c6a7a17ca3348e903dd8cdfcd74819b8de41
        version: 24.7.0-861
      entries:
      - name: appdynamics-operator.v24.7.0-861
        version: 24.7.0-861
      name: alpha
    - currentCSV: appdynamics-operator.v25.9.0-1218
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clusteragent",
                "metadata": {
                  "name": "k8s-cluster-agent",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "appName": "<app-name>",
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/cluster-agent:tag"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "InfraViz",
                "metadata": {
                  "name": "appdynamics-infraviz",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/machine-agent:latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3320c2c8d5f60adc541befd099dac1ef96257b16e0df28b727b0e589f78d981a
          createdAt: "2023-02-06T14:26:07Z"
          description: AppDynamics Operator to install AppDynamics Cluster Agent &
            AppDynamics Machine Agent for K8s Monitoring for various Cloud Providers
            like EKS, AKS etc.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: support@appdynamics.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clusteragent is the Schema for the clusteragents API
            displayName: Clusteragent
            kind: Clusteragent
            name: clusteragents.cluster.appdynamics.com
            version: v1alpha1
          - description: InfraViz is the Schema for the infravizs API
            displayName: Infra Viz
            kind: InfraViz
            name: infravizs.cluster.appdynamics.com
            version: v1alpha1
        description: |-
          AppDynamics Operator to install AppDynamics Cluster Agent & Machine Agent for K8s Monitoring on various Cloud Providers like EKS, AKS, GCP, Openshift etc.
          **About AppDynamics Operator**
          AppDynamics Operator simplifies the installation and lifecycle management of the AppDynamics ClusterAgent and the AppDynamics Machine Agent for K8s Monitoring on different Cloud Providers like EKS, AKS, GCP, OpenShift etc.
          **Installation Steps**
          Please refer below links for installation guide:
          [Install AppDynamics Cluster Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli)
          [Install AppDynamics Machine Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli)
          [Install AppDynamics Cluster Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts)
          [Install Appdynamics Machine Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts)
        displayName: AppDynamics Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - AppDynamics
        - Operator
        - AppDynamics Cluster Agent
        - AppDynamics Machine Agent
        - AppDynamics K8s Monitoring
        links:
        - name: AppDynamics Kubernetes Monitoring Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent
        - name: AppDynamics Cluster Agent Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/overview-of-cluster-monitoring
        - name: AppDynamics Cluster Agent Requirements
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/cluster-agent-requirements-and-supported-environments
        - name: Install AppDynamics Cluster Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli
        - name: Install AppDynamics Cluster Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts
        - name: Install AppDynamics Machine Agent & NetViz Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli
        - name: Install AppDynamics Machine Agent & NetViz Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: AppDynamics LLC
          url: https://www.appdynamics.com/
        relatedImages:
        - registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3320c2c8d5f60adc541befd099dac1ef96257b16e0df28b727b0e589f78d981a
        version: 25.9.0-1218
      entries:
      - name: appdynamics-operator.v25.9.0-1218
        version: 25.9.0-1218
      name: stable
    defaultChannel: alpha
    packageName: appdynamics-operator
    provider:
      name: AppDynamics LLC
      url: https://www.appdynamics.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: www.redhat.com
    name: odf-node-recovery-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-node-recovery-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "NodeRecovery",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "odf-node-recovery-operator",
                    "app.kubernetes.io/instance": "noderecovery-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "noderecovery",
                    "app.kubernetes.io/part-of": "odf-node-recovery-operator"
                  },
                  "name": "noderecovery-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Storage,OpenShift Optional
          description: ODF Node Recovery is an operator that assist in the recovery
            of and ODF cluster that has a device or a node replaced
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/jordigilh/odf-node-recovery-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeRecovery is the Schema for the noderecoveries API
            displayName: Node Recovery
            kind: NodeRecovery
            name: noderecoveries.odf.openshift.io
            version: v1alpha1
        description: |
          ODF Node Recovery is an operator that assist in the recovery of an
          ODF cluster in these 2 scenarios:
          * <a href="https://docs.redhat.com/en/documentation/red_hat_openshift_data_foundation/4.14/html-single/replacing_devices/index"> Device replacement<a/>
          * <a href="https://docs.redhat.com/en/documentation/red_hat_openshift_data_foundation/4.14/html-single/replacing_nodes/index"> Node replacement</a>
        displayName: ODF Node Recovery Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - node recovery
        links:
        - name: GitHub Repository
          url: https://github.com/jordigilh/odf-node-recovery-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
          url: www.redhat.com
        relatedImages:
        - quay.io/jordigilh/odf-node-recovery-controller@sha256:4164fafa9bb1c4d6121f4c5fc700fa2827c3994ca98a65d641471a2a46d4f59c
        version: 1.1.0
      entries:
      - name: odf-node-recovery-operator.v1.1.0
        version: 1.1.0
      - name: odf-node-recovery-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: odf-node-recovery-operator
    provider:
      name: Red Hat, Inc.
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: authorino-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: authorino-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/authorino-operator:v0.8.0
          createdAt: 2021-12-08T10-00-00Z
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta1
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: The operator to manage instances of Authorino
        displayName: Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: Authorino Operator
          url: https://github.com/Kuadrant/authorino-operator
        - name: Authorino
          url: https://github.com/Kuadrant/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/authorino-operator:v0.8.0
        version: 0.8.0
      entries:
      - name: authorino-operator.v0.8.0
        version: 0.8.0
      - name: authorino-operator.v0.7.0
        version: 0.7.0
      - name: authorino-operator.v0.6.0
        version: 0.6.0
      - name: authorino-operator.v0.5.0
        version: 0.5.0
      - name: authorino-operator.v0.4.1
        version: 0.4.1
      - name: authorino-operator.v0.4.0
        version: 0.4.0
      - name: authorino-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: authorino-operator.v0.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta3",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/authorino-operator:v0.16.0
          createdAt: "2024-11-25T13:22:24Z"
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta3
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: The operator to manage instances of Authorino
        displayName: Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: Authorino Operator
          url: https://github.com/Kuadrant/authorino-operator
        - name: Authorino
          url: https://github.com/Kuadrant/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/authorino-operator:v0.16.0
        - quay.io/kuadrant/authorino:v0.20.0
        version: 0.16.0
      entries:
      - name: authorino-operator.v0.16.0
        version: 0.16.0
      - name: authorino-operator.v0.15.1
        version: 0.15.1
      - name: authorino-operator.v0.15.0
        version: 0.15.0
      - name: authorino-operator.v0.14.0
        version: 0.14.0
      - name: authorino-operator.v0.13.1
        version: 0.13.1
      - name: authorino-operator.v0.13.0
        version: 0.13.0
      - name: authorino-operator.v0.12.0
        version: 0.12.0
      - name: authorino-operator.v0.11.1
        version: 0.11.1
      - name: authorino-operator.v0.10.0
        version: 0.10.0
      - name: authorino-operator.v0.9.0
        version: 0.9.0
      - name: authorino-operator.v0.8.0
        version: 0.8.0
      - name: authorino-operator.v0.7.0
        version: 0.7.0
      - name: authorino-operator.v0.6.0
        version: 0.6.0
      - name: authorino-operator.v0.5.0
        version: 0.5.0
      - name: authorino-operator.v0.4.1
        version: 0.4.1
      - name: authorino-operator.v0.4.0
        version: 0.4.0
      - name: authorino-operator.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: authorino-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Eclipse Foundation
      provider-url: ""
    name: eclipse-che
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eclipse-che.v7.110.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "org.eclipse.che/v1",
                "kind": "CheCluster",
                "metadata": {
                  "name": "eclipse-che",
                  "namespace": "eclipse-che"
                },
                "spec": {
                  "auth": {
                    "identityProviderURL": "",
                    "oAuthClientName": "",
                    "oAuthSecret": ""
                  },
                  "database": {
                    "externalDb": false
                  },
                  "k8s": {
                    "ingressDomain": null,
                    "tlsSecretName": null
                  },
                  "metrics": {
                    "enable": true
                  },
                  "server": {
                    "externalDevfileRegistries": [
                      {
                        "url": "https://registry.devfile.io"
                      }
                    ],
                    "externalDevfileRegistry": true,
                    "externalPluginRegistry": true,
                    "workspaceNamespaceDefault": "<username>-che"
                  },
                  "storage": {
                    "pvcStrategy": "common"
                  }
                }
              },
              {
                "apiVersion": "org.eclipse.che/v2",
                "kind": "CheCluster",
                "metadata": {
                  "name": "eclipse-che",
                  "namespace": "eclipse-che"
                },
                "spec": {
                  "components": {
                    "devfileRegistry": {
                      "disableInternalRegistry": true,
                      "externalDevfileRegistries": [
                        {
                          "url": "https://registry.devfile.io"
                        }
                      ]
                    },
                    "pluginRegistry": {
                      "disableInternalRegistry": true
                    }
                  },
                  "containerRegistry": {},
                  "devEnvironments": {},
                  "gitServices": {},
                  "networking": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/eclipse/che-operator@sha256:10c4d05c2cb848d04b97caf752c7f4d181e6973683263e45adec417342164e51
          createdAt: "2025-10-02T14:44:20Z"
          description: A Kube-native development solution that delivers portable and
            collaborative developer workspaces.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eclipse-che/che-operator
          support: Eclipse Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              The `CheCluster` custom resource allows defining and managing Eclipse Che server installation.
              Based on these settings, the  Operator automatically creates and maintains several ConfigMaps:
              `che`, `plugin-registry` that will contain the appropriate environment variables
              of the various components of the installation. These generated ConfigMaps must NOT be updated manually.
            displayName: Eclipse Che instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v2
          - description: The `CheCluster` custom resource allows defining and managing
              a Che server installation
            displayName: Eclipse Che instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v1
        description: |
          A collaborative Kubernetes-native development solution that delivers OpenShift workspaces and in-browser IDE for rapid cloud application development.
          This operator installs the Plugin and Devfile registries, Dashboard, Gateway and the Eclipse Che server, and configures these services.
          OpenShift OAuth is used directly for authentication. TLS mode is on.

          ## How to Install
          Press the **Install** button, choose the channel and the upgrade strategy, and wait for the **Installed** Operator status.
          When the operator is installed, create a new Custom Resource (CR) of Kind CheCluster (click the **Create New** button).
          The CR spec contains all defaults. You can start using Eclipse Che when the CR status is set to **Available**, and you see a URL to Eclipse Che.

          ## Defaults
          By default, the operator deploys Eclipse Che with:
          * 10Gi storage
          * Auto-generated passwords
          * Bundled Plugin and Devfile registries

          Use `oc edit checluster/eclipse-che -n eclipse-che` to update Eclipse Che default installation options.
          See more in the [Installation guide](https://www.eclipse.org/che/docs/stable/administration-guide/configuring-che/).

          ### Certificates
          Operator uses a default router certificate to secure Eclipse Che routes.
          Follow the [guide](https://www.eclipse.org/che/docs/stable/administration-guide/importing-untrusted-tls-certificates/)
          to import certificates into Eclipse Che.
        displayName: Eclipse Che
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eclipse che
        - workspaces
        - devtools
        - developer
        - ide
        - java
        links:
        - name: Product Page
          url: https://www.eclipse.org/che
        - name: Documentation
          url: https://www.eclipse.org/che/docs
        - name: Operator GitHub Repo
          url: https://github.com/eclipse-che/che-operator
        maintainers:
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Eclipse Foundation
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy@sha256:738c854322f56d63ebab75de5210abcdd5e0782ce2d30c0ecd4620f63b24694d
        - quay.io/devfile/universal-developer-image@sha256:3cf53fb9eab60ee3156e9edd4ec5af72b70e55d5ffb28f41ecb5b26e01dc9838
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:354fc75eb7a21a934381e93d03ef9d42bc2c8ae8989fdcacecfb39b863b96ced
        - quay.io/eclipse/che-operator@sha256:10c4d05c2cb848d04b97caf752c7f4d181e6973683263e45adec417342164e51
        - quay.io/eclipse/che--traefik@sha256:4cf907247939b5d20bf4eff73abd21cb413c339600dde76dbc94a874b2578a27
        - quay.io/che-incubator/che-code@sha256:cfcfbe4d665b18b11401c3c35b23b04d292c31e7f4be6d78a5e6d60fce776ed0
        - quay.io/che-incubator/che-code-server@sha256:357c95297ce2182ef0a792038ea122075fdeb6c8e835099be09e8d4f105eac2a
        - quay.io/eclipse/che-tls-secret-creator@sha256:cbb2fd8ce903d822377f2093af1aa2543672ed8c516d607a7a9dd857379dac5d
        - quay.io/che-incubator/configbump@sha256:396714521307ae9f79453c492fb8064cf940cd6b4e6183a825f1056045623874
        - quay.io/che-incubator/che-idea-dev-server@sha256:87c8b53cc295e3d7eb35b326e221a949553be3345135a4b6c1999d66f345a8b2
        - quay.io/che-incubator/che-code-sshd@sha256:7eb5381fba8878b8f252523caf70c9957b9f9e699db4564061017fb0a023cc1a
        - quay.io/che-incubator/che-idea@sha256:8aae69dc4b0c122491a75400639af0fe92b5e214c6e68ac97cda29fb58b44151
        - quay.io/eclipse/che-plugin-registry@sha256:a70a0d98505ba5aeed691fc70c3c59aed3304155fba0ab3ad70a6d0453b22093
        - quay.io/che-incubator/che-code-server@sha256:2093baab1a73e9682f5892f80582481abfbef99d744472ee6098e2d82fb2899f
        - quay.io/che-incubator/header-rewrite-proxy@sha256:bd7873b8feef35f218f54c6251ea224bea2c8bf202a328230019a0ba2941245d
        - quay.io/oauth2-proxy/oauth2-proxy@sha256:dcb6ff8dd21bf3058f6a22c6fa385fa5b897a9cd3914c88a2cc2bb0a85f8065d
        - quay.io/che-incubator/che-code@sha256:ac62a2f42ae6d6f6628620a00133afece954fcbed2df309fa3de79e4c28cc029
        - quay.io/openshift/origin-oauth-proxy@sha256:870bfe92a4663720775c0dfe5728ecbb10a17f0644eef5f57276ec135034c6a1
        - quay.io/eclipse/che-dashboard@sha256:2db3255752aebd75dceea578cd27b7e8ecb41c2290827ff7104ed77e1b61feb9
        - quay.io/eclipse/che-server@sha256:b83d536cdd3cb2b86def5599495c40870dde20c194a003736a2b47a5e5e17a06
        version: 7.110.0
      entries:
      - name: eclipse-che.v7.110.0
        version: 7.110.0
      - name: eclipse-che.v7.109.0
        version: 7.109.0
      - name: eclipse-che.v7.108.0
        version: 7.108.0
      - name: eclipse-che.v7.107.0
        version: 7.107.0
      - name: eclipse-che.v7.106.0
        version: 7.106.0
      - name: eclipse-che.v7.105.0
        version: 7.105.0
      - name: eclipse-che.v7.104.0
        version: 7.104.0
      - name: eclipse-che.v7.103.0
        version: 7.103.0
      - name: eclipse-che.v7.102.0
        version: 7.102.0
      - name: eclipse-che.v7.101.0
        version: 7.101.0
      - name: eclipse-che.v7.100.0
        version: 7.100.0
      - name: eclipse-che.v7.99.0
        version: 7.99.0
      - name: eclipse-che.v7.98.0
        version: 7.98.0
      - name: eclipse-che.v7.97.0
        version: 7.97.0
      - name: eclipse-che.v7.96.0
        version: 7.96.0
      - name: eclipse-che.v7.95.0
        version: 7.95.0
      - name: eclipse-che.v7.94.0
        version: 7.94.0
      - name: eclipse-che.v7.93.0
        version: 7.93.0
      - name: eclipse-che.v7.92.0
        version: 7.92.0
      - name: eclipse-che.v7.91.0
        version: 7.91.0
      - name: eclipse-che.v7.90.0
        version: 7.90.0
      - name: eclipse-che.v7.89.0
        version: 7.89.0
      - name: eclipse-che.v7.88.0
        version: 7.88.0
      - name: eclipse-che.v7.87.0
        version: 7.87.0
      - name: eclipse-che.v7.86.0
        version: 7.86.0
      - name: eclipse-che.v7.85.0
        version: 7.85.0
      - name: eclipse-che.v7.84.0
        version: 7.84.0
      - name: eclipse-che.v7.83.0
        version: 7.83.0
      - name: eclipse-che.v7.82.0
        version: 7.82.0
      - name: eclipse-che.v7.81.0
        version: 7.81.0
      - name: eclipse-che.v7.80.0
        version: 7.80.0
      - name: eclipse-che.v7.79.0
        version: 7.79.0
      - name: eclipse-che.v7.78.0
        version: 7.78.0
      - name: eclipse-che.v7.77.0
        version: 7.77.0
      - name: eclipse-che.v7.76.0
        version: 7.76.0
      - name: eclipse-che.v7.75.0
        version: 7.75.0
      - name: eclipse-che.v7.74.0
        version: 7.74.0
      - name: eclipse-che.v7.73.0
        version: 7.73.0
      - name: eclipse-che.v7.72.0
        version: 7.72.0
      - name: eclipse-che.v7.71.0
        version: 7.71.0
      - name: eclipse-che.v7.70.0
        version: 7.70.0
      - name: eclipse-che.v7.69.0
        version: 7.69.0
      - name: eclipse-che.v7.68.0
        version: 7.68.0
      - name: eclipse-che.v7.67.0
        version: 7.67.0
      - name: eclipse-che.v7.66.0
        version: 7.66.0
      - name: eclipse-che.v7.65.0
        version: 7.65.0
      - name: eclipse-che.v7.64.0
        version: 7.64.0
      - name: eclipse-che.v7.63.0
        version: 7.63.0
      - name: eclipse-che.v7.62.0
        version: 7.62.0
      - name: eclipse-che.v7.61.0
        version: 7.61.0
      - name: eclipse-che.v7.60.1
        version: 7.60.1
      - name: eclipse-che.v7.60.0
        version: 7.60.0
      - name: eclipse-che.v7.59.0
        version: 7.59.0
      - name: eclipse-che.v7.58.0
        version: 7.58.0
      - name: eclipse-che.v7.57.0
        version: 7.57.0
      - name: eclipse-che.v7.56.0
        version: 7.56.0
      - name: eclipse-che.v7.55.0
        version: 7.55.0
      - name: eclipse-che.v7.54.0
        version: 7.54.0
      - name: eclipse-che.v7.53.0
        version: 7.53.0
      - name: eclipse-che.v7.52.0
        version: 7.52.0
      - name: eclipse-che.v7.51.0
        version: 7.51.0
      - name: eclipse-che.v7.50.0
        version: 7.50.0
      - name: eclipse-che.v7.49.0
        version: 7.49.0
      - name: eclipse-che.v7.48.1
        version: 7.48.1
      - name: eclipse-che.v7.47.0
        version: 7.47.0
      - name: eclipse-che.v7.46.0
        version: 7.46.0
      - name: eclipse-che.v7.45.0
        version: 7.45.0
      - name: eclipse-che.v7.44.0
        version: 7.44.0
      - name: eclipse-che.v7.42.0
        version: 7.42.0
      - name: eclipse-che.v7.41.2
        version: 7.41.2
      - name: eclipse-che.v7.41.1
        version: 7.41.1
      - name: eclipse-che.v7.41.0
        version: 7.41.0
      - name: eclipse-che.v7.40.2
        version: 7.40.2
      - name: eclipse-che.v7.40.1
        version: 7.40.1
      name: stable
    defaultChannel: stable
    packageName: eclipse-che
    provider:
      name: Eclipse Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solution-Soft
      provider-url: www.solution-soft.com
    name: timemachine-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: timemachine-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tm.solution-soft.com/v1alpha1",
                "kind": "TimeMachine",
                "metadata": {
                  "name": "timemachine-sample"
                },
                "spec": {
                  "cluster_provider": "ibm",
                  "aws": {
                    "shared_volume_name": ""
                  },
                  "ibm": {
                    "persistentvolume": {
                      "storageclass": "ibmc-file-silver-gid"
                    }
                  },
                  "tm_lichost": "52.152.148.40",
                  "tm_licpass": "qwaszx",
                  "tm_licport": "57777",
                  "imagePullSecrets": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
          createdAt: "2024-01-30T10:20:30Z"
          description: Time Machine Operator enables time travel of OpenShift pods,
            without the need to change the system clock on any of the cluster nodes
          olm.skipRange: '>=0.0.4 <1.0.17'
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: SolutionSoft Systems, Inc. - support@solution-soft.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Time Machine
            displayName: Time Machine
            kind: TimeMachine
            name: timemachines.tm.solution-soft.com
            version: v1alpha1
        description: |-
          Time Machine operator enables time travel of OpenShift pods/deployments.

          To time travel a pod where your application is, you just need to install the Time Machine operator and create a Time Machine custom resource in the same namespace, without the need of changing the container images of the application you’re using, or adding containers to the pod with your application.

          Time travelling is done on the namespace level where all the (configured) target pods/deployments see the same virtual time.
        displayName: Time Machine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - time shift
        - time travel
        - virtual clocks
        - time machine
        - virtual time
        - solution-soft
        links:
        - name: Timemachine Operator
          url: https://catalog.redhat.com/software/operators/detail/5e98727a6c5dcb34dfbb19e3
        maintainers:
        - email: dtrajkovic@solution-soft.com
          name: Dragan Trajkovic
        maturity: alpha
        provider:
          name: Solution-Soft
          url: www.solution-soft.com
        relatedImages:
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
        - registry.connect.redhat.com/rhel-timemachine-container/virtual-clock-testing-tool@sha256:a81a823328b0ee795c40668c89a08bcf545c5192d9eed0146071e3721a0a8372
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0c304c11986790f446604319dfc29db494b15f5f9fa13f2c436688ec29c92dbe
        version: 1.0.17
      entries:
      - name: timemachine-operator.v1.0.17
        version: 1.0.17
      - name: timemachine-operator.v1.0.16
        version: 1.0.16
      name: alpha
    defaultChannel: alpha
    packageName: timemachine-operator
    provider:
      name: Solution-Soft
      url: www.solution-soft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: antrea.io
      provider-url: ""
    name: antrea-by-broadcom-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: antrea-by-broadcom-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.antrea.vmware.com/v1",
                "kind": "AntreaInstall",
                "metadata": {
                  "name": "antrea-install",
                  "namespace": "antrea-operator"
                },
                "spec": {
                  "antreaAgentConfig": "# FeatureGates is a map of feature names to bools that enable or disable experimental features.\nfeatureGates:\n# AllAlpha is a global toggle for alpha features. Per-feature key values override the default set by AllAlpha.\n#  AllAlpha: false\n\n# AllBeta is a global toggle for beta features. Per-feature key values override the default set by AllBeta.\n#  AllBeta: false\n\n# Enable AntreaProxy which provides ServiceLB for in-cluster Services in antrea-agent.\n# It should be enabled on Windows, otherwise NetworkPolicy will not take effect on\n# Service traffic.\n#  AntreaProxy: true\n\n# Enable EndpointSlice support in AntreaProxy. Don't enable this feature unless that EndpointSlice\n# API version v1 is supported and set as enabled in Kubernetes. If AntreaProxy is not enabled,\n# this flag will not take effect.\n#  EndpointSlice: false\n\n# Enable TopologyAwareHints in AntreaProxy. This requires AntreaProxy and EndpointSlice to be\n# enabled, otherwise this flag will not take effect.\n#  TopologyAwareHints: false\n\n# Enable traceflow which provides packet tracing feature to diagnose network issue.\n#  Traceflow: true\n\n# Enable NodePortLocal feature to make the Pods reachable externally through NodePort\n#  NodePortLocal: true\n\n# Enable Antrea ClusterNetworkPolicy feature to complement K8s NetworkPolicy for cluster admins\n# to define security policies which apply to the entire cluster, and Antrea NetworkPolicy\n# feature that supports priorities, rule actions and externalEntities in the future.\n#  AntreaPolicy: true\n\n# Enable flowexporter which exports polled conntrack connections as IPFIX flow records from each\n# agent to a configured collector.\n#  FlowExporter: false\n\n# Enable collecting and exposing NetworkPolicy statistics.\n#  NetworkPolicyStats: true\n\n# Enable controlling SNAT IPs of Pod egress traffic.\n#  Egress: true\n\n# Enable AntreaIPAM, which can allocate IP addresses from IPPools. AntreaIPAM is required by the\n# bridging mode and allocates IPs to Pods in bridging mode. It is also required to use Antrea for\n# IPAM when configuring secondary network interfaces with Multus.\n#  AntreaIPAM: false\n\n# Enable multicast traffic.\n#  Multicast: false\n\n# Enable Antrea Multi-cluster Gateway to support cross-cluster traffic.\n# This feature is supported only with encap mode.\n#  Multicluster: false\n\n# Enable support for provisioning secondary network interfaces for Pods (using\n# Pod annotations). At the moment, Antrea can only create secondary network\n# interfaces using SR-IOV VFs on baremetal Nodes.\n#  SecondaryNetwork: false\n\n# Enable managing external IPs of Services of LoadBalancer type.\n#  ServiceExternalIP: false\n\n# Enable mirroring or redirecting the traffic Pods send or receive.\n#  TrafficControl: false\n\n# Enable certificated-based authentication for IPsec.\n#  IPsecCertAuth: false\n\n# Enable collecting support bundle files with SupportBundleCollection CRD.\n#  SupportBundleCollection: false\n\n# Enable users to protect their applications by specifying how they are allowed to communicate with others, taking\n# into account application context.\n#  L7NetworkPolicy: false\n\n# Name of the OpenVSwitch bridge antrea-agent will create and use.\n# Make sure it doesn't conflict with your existing OpenVSwitch bridges.\novsBridge: \"br-int\"\n\n# Datapath type to use for the OpenVSwitch bridge created by Antrea. At the moment, the only\n# supported value is 'system', which corresponds to the kernel datapath.\n#ovsDatapathType: system\n\n# Name of the interface antrea-agent will create and use for host \u003c--\u003e pod communication.\n# Make sure it doesn't conflict with your existing interfaces.\nhostGateway: \"antrea-gw0\"\n\n# Determines how traffic is encapsulated. It has the following options:\n# encap(default):    Inter-node Pod traffic is always encapsulated and Pod to external network\n#                    traffic is SNAT'd.\n# noEncap:           Inter-node Pod traffic is not encapsulated; Pod to external network traffic is\n#                    SNAT'd if noSNAT is not set to true. Underlying network must be capable of\n#                    supporting Pod traffic across IP subnets.\n# hybrid:            noEncap if source and destination Nodes are on the same subnet, otherwise encap.\n# networkPolicyOnly: Antrea enforces NetworkPolicy only, and utilizes CNI chaining and delegates Pod\n#                    IPAM and connectivity to the primary CNI.\n#\ntrafficEncapMode: \"encap\"\n\n# Whether or not to SNAT (using the Node IP) the egress traffic from a Pod to the external network.\n# This option is for the noEncap traffic mode only, and the default value is false. In the noEncap\n# mode, if the cluster's Pod CIDR is reachable from the external network, then the Pod traffic to\n# the external network needs not be SNAT'd. In the networkPolicyOnly mode, antrea-agent never\n# performs SNAT and this option will be ignored; for other modes it must be set to false.\nnoSNAT: false\n\n# Tunnel protocols used for encapsulating traffic across Nodes. If WireGuard is enabled in trafficEncryptionMode,\n# this option will not take effect. Supported values:\n# - geneve (default)\n# - vxlan\n# - gre\n# - stt\n# Note that \"gre\" is not supported for IPv6 clusters (IPv6-only or dual-stack clusters).\ntunnelType: \"geneve\"\n\n# TunnelPort is the destination port for UDP and TCP based tunnel protocols (Geneve, VXLAN, and STT).\n# If zero, it will use the assigned IANA port for the protocol, i.e. 6081 for Geneve, 4789 for VXLAN,\n# and 7471 for STT.\ntunnelPort: 0\n\n# TunnelCsum determines whether to compute UDP encapsulation header (Geneve or VXLAN) checksums on outgoing\n# packets. For Linux kernel before Mar 2021, UDP checksum must be present to trigger GRO on the receiver for better\n# performance of Geneve and VXLAN tunnels. The issue has been fixed by\n# https://github.com/torvalds/linux/commit/89e5c58fc1e2857ccdaae506fb8bc5fed57ee063, thus computing UDP checksum is\n# no longer necessary.\n# It should only be set to true when you are using an unpatched Linux kernel and observing poor transfer performance.\ntunnelCsum: false\n\n# Determines how tunnel traffic is encrypted. Currently encryption only works with encap mode.\n# It has the following options:\n# - none (default):  Inter-node Pod traffic will not be encrypted.\n# - ipsec:           Enable IPsec (ESP) encryption for Pod traffic across Nodes. Antrea uses\n#                    Preshared Key (PSK) for IKE authentication. When IPsec tunnel is enabled,\n#                    the PSK value must be passed to Antrea Agent through an environment\n#                    variable: ANTREA_IPSEC_PSK.\n# - wireGuard:       Enable WireGuard for tunnel traffic encryption.\ntrafficEncryptionMode: \"none\"\n\n# Enable bridging mode of Pod network on Nodes, in which the Node's transport interface is connected\n# to the OVS bridge, and cross-Node/VLAN traffic of AntreaIPAM Pods (Pods whose IP addresses are\n# allocated by AntreaIPAM from IPPools) is sent to the underlay network, and forwarded/routed by the\n# underlay network.\n# This option requires the `AntreaIPAM` feature gate to be enabled. At this moment, it supports only\n# IPv4 and Linux Nodes, and can be enabled only when `ovsDatapathType` is `system`,\n# `trafficEncapMode` is `noEncap`, and `noSNAT` is true.\nenableBridgingMode: false\n\n# Disable TX checksum offloading for container network interfaces. It's supposed to be set to true when the\n# datapath doesn't support TX checksum offloading, which causes packets to be dropped due to bad checksum.\n# It affects Pods running on Linux Nodes only.\ndisableTXChecksumOffload: false\n\n# Default MTU to use for the host gateway interface and the network interface of each Pod.\n# If omitted, antrea-agent will discover the MTU of the Node's primary interface and\n# also adjust MTU to accommodate for tunnel encapsulation overhead (if applicable).\ndefaultMTU: 0\n\n# wireGuard specifies WireGuard related configurations.\nwireGuard:\n  # The port for WireGuard to receive traffic.\n  port: 51820\n\negress:\n  # exceptCIDRs is the CIDR ranges to which outbound Pod traffic will not be SNAT'd by Egresses.\n  exceptCIDRs:\n\n# ClusterIP CIDR range for Services. It's required when AntreaProxy is not enabled, and should be\n# set to the same value as the one specified by --service-cluster-ip-range for kube-apiserver. When\n# AntreaProxy is enabled, this parameter is not needed and will be ignored if provided.\nserviceCIDR: \"\"\n\n# ClusterIP CIDR range for IPv6 Services. It's required when using kube-proxy to provide IPv6 Service in a Dual-Stack\n# cluster or an IPv6 only cluster. The value should be the same as the configuration for kube-apiserver specified by\n# --service-cluster-ip-range. When AntreaProxy is enabled, this parameter is not needed.\n# No default value for this field.\nserviceCIDRv6: \"\"\n\n# The port for the antrea-agent APIServer to serve on.\n# Note that if it's set to another value, the `containerPort` of the `api` port of the\n# `antrea-agent` container must be set to the same value.\napiPort: 10350\n\n# Enable metrics exposure via Prometheus. Initializes Prometheus metrics listener.\nenablePrometheusMetrics: true\n\n# Provide the IPFIX collector address as a string with format \u003cHOST\u003e:[\u003cPORT\u003e][:\u003cPROTO\u003e].\n# HOST can either be the DNS name or the IP of the Flow Collector. For example,\n# \"flow-aggregator.flow-aggregator.svc\" can be provided as DNS name to connect\n# to the Antrea Flow Aggregator service. If IP, it can be either IPv4 or IPv6.\n# However, IPv6 address should be wrapped with [].\n# If PORT is empty, we default to 4739, the standard IPFIX port.\n# If no PROTO is given, we consider \"tls\" as default. We support \"tls\", \"tcp\" and\n# \"udp\" protocols. \"tls\" is used for securing communication between flow exporter and\n# flow aggregator.\nflowCollectorAddr: \"flow-aggregator.flow-aggregator.svc:4739:tls\"\n\n# Provide flow poll interval as a duration string. This determines how often the\n# flow exporter dumps connections from the conntrack module. Flow poll interval\n# should be greater than or equal to 1s (one second).\n# Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\", \"m\", \"h\".\nflowPollInterval: \"5s\"\n\n# Provide the active flow export timeout, which is the timeout after which a flow\n# record is sent to the collector for active flows. Thus, for flows with a continuous\n# stream of packets, a flow record will be exported to the collector once the elapsed\n# time since the last export event is equal to the value of this timeout.\n# Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\", \"m\", \"h\".\nactiveFlowExportTimeout: \"5s\"\n\n# Provide the idle flow export timeout, which is the timeout after which a flow\n# record is sent to the collector for idle flows. A flow is considered idle if no\n# packet matching this flow has been observed since the last export event.\n# Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\", \"m\", \"h\".\nidleFlowExportTimeout: \"15s\"\n\nnodePortLocal:\n# Enable NodePortLocal, a feature used to make Pods reachable using port forwarding on the host. To\n# enable this feature, you need to set \"enable\" to true, and ensure that the NodePortLocal feature\n# gate is also enabled (which is the default).\n  enable: false\n# Provide the port range used by NodePortLocal. When the NodePortLocal feature is enabled, a port\n# from that range will be assigned whenever a Pod's container defines a specific port to be exposed\n# (each container can define a list of ports as pod.spec.containers[].ports), and all Node traffic\n# directed to that port will be forwarded to the Pod.\n  portRange: \"61000-62000\"\n\n# Provide the address of Kubernetes apiserver, to override any value provided in kubeconfig or InClusterConfig.\n# Defaults to \"\". It must be a host string, a host:port pair, or a URL to the base of the apiserver.\nkubeAPIServerOverride: \"\"\n\n# Provide the address of DNS server, to override the kube-dns service. It's used to resolve hostname in FQDN policy.\n# Defaults to \"\". It must be a host string or a host:port pair of the DNS server (e.g. 10.96.0.10, 10.96.0.10:53,\n# [fd00:10:96::a]:53).\ndnsServerOverride: \"\"\n\n# Comma-separated list of Cipher Suites. If omitted, the default Go Cipher Suites will be used.\n# https://golang.org/pkg/crypto/tls/#pkg-constants\n# Note that TLS1.3 Cipher Suites cannot be added to the list. But the apiserver will always\n# prefer TLS1.3 Cipher Suites whenever possible.\ntlsCipherSuites: \"\"\n\n# TLS min version from: VersionTLS10, VersionTLS11, VersionTLS12, VersionTLS13.\ntlsMinVersion: \"\"\n\n# The name of the interface on Node which is used for tunneling or routing the traffic across Nodes.\n# If there are multiple IP addresses configured on the interface, the first one is used. The IP\n# address used for tunneling or routing traffic to remote Nodes is decided in the following order of\n# preference (from highest to lowest):\n# 1. transportInterface\n# 2. transportInterfaceCIDRs\n# 3. The Node IP\ntransportInterface: \"\"\n\nmulticast:\n# The names of the interfaces on Nodes that are used to forward multicast traffic.\n# Defaults to transport interface if not set.\n  multicastInterfaces:\n\n# The interval at which the antrea-agent sends IGMP queries to Pods.\n# Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\", \"m\", \"h\".\n  igmpQueryInterval: \"125s\"\n\n# The network CIDRs of the interface on Node which is used for tunneling or routing the traffic across\n# Nodes. If there are multiple interfaces configured the same network CIDR, the first one is used. The\n# IP address used for tunneling or routing traffic to remote Nodes is decided in the following order of\n# preference (from highest to lowest):\n# 1. transportInterface\n# 2. transportInterfaceCIDRs\n# 3. The Node IP\ntransportInterfaceCIDRs:\n\n# Option antreaProxy contains AntreaProxy related configuration options.\nantreaProxy:\n  # ProxyAll tells antrea-agent to proxy all Service traffic, including NodePort, LoadBalancer, and ClusterIP traffic,\n  # regardless of where they come from. Therefore, running kube-proxy is no longer required. This requires the AntreaProxy\n  # feature to be enabled.\n  # Note that this option is experimental. If kube-proxy is removed, option kubeAPIServerOverride must be used to access\n  # apiserver directly.\n  proxyAll: false\n  # A string array of values which specifies the host IPv4/IPv6 addresses for NodePort. Values can be valid IP blocks.\n  # (e.g. 1.2.3.0/24, 1.2.3.4/32). An empty string slice is meant to select all host IPv4/IPv6 addresses.\n  # Note that the option is only valid when proxyAll is true.\n  nodePortAddresses:\n  # An array of string values to specify a list of Services which should be ignored by AntreaProxy (traffic to these\n  # Services will not be load-balanced). Values can be a valid ClusterIP (e.g. 10.11.1.2) or a Service name\n  # with Namespace (e.g. kube-system/kube-dns)\n  skipServices:\n  # When ProxyLoadBalancerIPs is set to false, AntreaProxy no longer load-balances traffic destined to the\n  # External IPs of LoadBalancer Services. This is useful when the external LoadBalancer provides additional\n  # capabilities (e.g. TLS termination) and it is desirable for Pod-to-ExternalIP traffic to be sent to the\n  # external LoadBalancer instead of being load-balanced to an Endpoint directly by AntreaProxy.\n  # Note that setting ProxyLoadBalancerIPs to false usually only makes sense when ProxyAll is set to true and\n  # kube-proxy is removed from the cluser, otherwise kube-proxy will still load-balance this traffic.\n  proxyLoadBalancerIPs: true\n\n# IPsec tunnel related configurations.\nipsec:\n  # The authentication mode of IPsec tunnel. It has the following options:\n  # - psk (default): Use pre-shared key (PSK) for IKE authentication.\n  # - cert:          Use CA-signed certificates for IKE authentication. This option requires the `IPsecCertAuth`\n  #                  feature gate to be enabled.\n  authenticationMode: \"psk\"\n\nmulticluster:\n# Enable Antrea Multi-cluster Gateway to support cross-cluster traffic.\n# This feature is supported only with encap mode.\n  enable: false\n# The Namespace where Antrea Multi-cluster Controller is running.\n# The default is antrea-agent's Namespace.\n  namespace: \"\"\n# Enable StretchedNetworkPolicy which could be enforced on cross-cluster traffic.\n# This feature is supported only with encap mode.\n  enableStretchedNetworkPolicy: false\n",
                  "antreaCNIConfig": "{\n    \"cniVersion\":\"0.3.0\",\n    \"name\": \"antrea\",\n    \"plugins\": [\n        {\n            \"type\": \"antrea\",\n            \"ipam\": {\n                \"type\": \"host-local\"\n            }\n        }\n        ,\n        {\n            \"type\": \"portmap\",\n            \"capabilities\": {\"portMappings\": true}\n        }\n        ,\n        {\n            \"type\": \"bandwidth\",\n            \"capabilities\": {\"bandwidth\": true}\n        }\n    ]\n}\n",
                  "antreaControllerConfig": "# FeatureGates is a map of feature names to bools that enable or disable experimental features.\nfeatureGates:\n# AllAlpha is a global toggle for alpha features. Per-feature key values override the default set by AllAlpha.\n#  AllAlpha: false\n\n# AllBeta is a global toggle for beta features. Per-feature key values override the default set by AllBeta.\n#  AllBeta: false\n\n# Enable traceflow which provides packet tracing feature to diagnose network issue.\n#  Traceflow: true\n\n# Enable Antrea ClusterNetworkPolicy feature to complement K8s NetworkPolicy for cluster admins\n# to define security policies which apply to the entire cluster, and Antrea NetworkPolicy\n# feature that supports priorities, rule actions and externalEntities in the future.\n#  AntreaPolicy: true\n\n# Enable collecting and exposing NetworkPolicy statistics.\n#  NetworkPolicyStats: true\n\n# Enable multicast traffic.\n#  Multicast: false\n\n# Enable controlling SNAT IPs of Pod egress traffic.\n#  Egress: true\n\n# Run Kubernetes NodeIPAMController with Antrea.\n#  NodeIPAM: false\n\n# Enable AntreaIPAM, which can allocate IP addresses from IPPools. AntreaIPAM is required by the\n# bridging mode and allocates IPs to Pods in bridging mode. It is also required to use Antrea for\n# IPAM when configuring secondary network interfaces with Multus.\n#  AntreaIPAM: false\n\n# Enable managing external IPs of Services of LoadBalancer type.\n#  ServiceExternalIP: false\n\n# Enable certificated-based authentication for IPsec.\n#  IPsecCertAuth: false\n\n# Enable managing ExternalNode for unmanaged VM/BM.\n#  ExternalNode: false\n\n# Enable collecting support bundle files with SupportBundleCollection CRD.\n#  SupportBundleCollection: false\n\n# Enable multi-cluster features.\n#  Multicluster: false\n\n# Enable users to protect their applications by specifying how they are allowed to communicate with others, taking\n# into account application context.\n#  L7NetworkPolicy: false\n\n# The port for the antrea-controller APIServer to serve on.\n# Note that if it's set to another value, the `containerPort` of the `api` port of the\n# `antrea-controller` container must be set to the same value.\napiPort: 10349\n\n# Enable metrics exposure via Prometheus. Initializes Prometheus metrics listener.\nenablePrometheusMetrics: true\n\n# Indicates whether to use auto-generated self-signed TLS certificate.\n# If false, a Secret named \"antrea-controller-tls\" must be provided with the following keys:\n#   ca.crt: \u003cCA certificate\u003e\n#   tls.crt: \u003cTLS certificate\u003e\n#   tls.key: \u003cTLS private key\u003e\nselfSignedCert: true\n\n# Comma-separated list of Cipher Suites. If omitted, the default Go Cipher Suites will be used.\n# https://golang.org/pkg/crypto/tls/#pkg-constants\n# Note that TLS1.3 Cipher Suites cannot be added to the list. But the apiserver will always\n# prefer TLS1.3 Cipher Suites whenever possible.\ntlsCipherSuites: \"\"\n\n# TLS min version from: VersionTLS10, VersionTLS11, VersionTLS12, VersionTLS13.\ntlsMinVersion: \"\"\n\nnodeIPAM:\n  # Enable the integrated Node IPAM controller within the Antrea controller.\n  enableNodeIPAM: false\n  # CIDR ranges for Pods in cluster. String array containing single CIDR range, or multiple ranges.\n  # The CIDRs could be either IPv4 or IPv6. At most one CIDR may be specified for each IP family.\n  # Value ignored when enableNodeIPAM is false.\n  clusterCIDRs:\n  # CIDR ranges for Services in cluster. It is not necessary to specify it when there is no overlap with clusterCIDRs.\n  # Value ignored when enableNodeIPAM is false.\n  serviceCIDR: \"\"\n  serviceCIDRv6: \"\"\n  # Mask size for IPv4 Node CIDR in IPv4 or dual-stack cluster. Value ignored when enableNodeIPAM is false\n  # or when IPv4 Pod CIDR is not configured. Valid range is 16 to 30.\n  nodeCIDRMaskSizeIPv4: 24\n  # Mask size for IPv6 Node CIDR in IPv6 or dual-stack cluster. Value ignored when enableNodeIPAM is false\n  # or when IPv6 Pod CIDR is not configured. Valid range is 64 to 126.\n  nodeCIDRMaskSizeIPv6: 64\n\nipsecCSRSigner:\n  # Determines the auto-approve policy of Antrea CSR signer for IPsec certificates management.\n  # If enabled, Antrea will auto-approve the CertificateSingingRequest (CSR) if its subject and x509 extensions\n  # are permitted, and the requestor can be validated. If K8s `BoundServiceAccountTokenVolume` feature is enabled,\n  # the Pod identity will also be validated to provide maximum security.\n  # If set to false, Antrea will not auto-approve CertificateSingingRequests and they need to be approved\n  # manually by `kubectl certificate approve`.\n  autoApprove: true\n  # Indicates whether to use auto-generated self-signed CA certificate.\n  # If false, a Secret named \"antrea-ipsec-ca\" must be provided with the following keys:\n  #   tls.crt: \u003cCA certificate\u003e\n  #   tls.key: \u003cCA private key\u003e\n  selfSignedCA: true\n\nmulticluster:\n  # Enable StretchedNetworkPolicy which allow Antrea-native policies to select peers\n  # from other clusters in a ClusterSet.\n  enableStretchedNetworkPolicy: false\n",
                  "antreaImage": "antrea/antrea-ubi:v2.1.0",
                  "antreaPlatform": "openshift"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: projects.packages.broadcom.com/antreainterworking/antrea-operator:v2.1.0_vmware.3@sha256:70f129499c801f164adf8772e7ccd5d1fc659df5f298d7851e6bf02ee760f2a9
          createdAt: "2025-04-16T12:00:00Z"
          description: An operator which installs Antrea network CNI plugin on the
            Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          support: projectantrea-maintainers@googlegroups.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AntreaInstall is the Schema for the antreainstalls API
            displayName: Antrea Install
            kind: AntreaInstall
            name: antreainstalls.operator.antrea.vmware.com
            version: v1
        description: An operator which installs Antrea network CNI plugin on the Kubernetes
          cluster.
        displayName: Antrea-by-Broadcom Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        links:
        - name: Antrea Operator For Kubernetes
          url: https://github.com/vmware/antrea-operator-for-kubernetes
        maintainers:
        - email: projectantrea-maintainers@googlegroups.com
          name: Project Antrea Maintainers
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: antrea.io
        relatedImages:
        - projects.packages.broadcom.com/antreainterworking/antrea-operator:v2.1.0_vmware.3@sha256:70f129499c801f164adf8772e7ccd5d1fc659df5f298d7851e6bf02ee760f2a9
        version: 1.10.0
      entries:
      - name: antrea-by-broadcom-operator.v1.10.0
        version: 1.10.0
      name: alpha
    defaultChannel: alpha
    packageName: antrea-by-broadcom-operator
    provider:
      name: antrea.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Nokia
      provider-url: https://docs.eda.dev
    name: eda-k8s-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eda-openshift-operator.v4.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.eda.nokia.com/v1",
                "kind": "ConnectControllerConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "operator"
                  },
                  "name": "connectcontrollerconfig-sample",
                  "namespace": "eda-connect-k8s"
                },
                "spec": {
                  "connectPluginName": "",
                  "controllerImagePullSecret": "",
                  "edaNamespace": "eda",
                  "heartBeat": "",
                  "logLevel": "",
                  "skipTLSVerify": false,
                  "tlsCertificateData": "",
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:c43a7fa7475db83011cc08b728982511fc0279a259e1aacc8ddde9e15f947725
          createdAt: "2025-06-10T15:47:27Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: eda-connect-k8s
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["connectnetworkattachmentstatuses.connect.eda.nokia.com",
            connectnetworkdefinitions.connect.eda.nokia.com]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConnectNetworkAttachmentStatus is the Schema for the connectnetworkattachmentstatuses
              API
            displayName: Connect NetworkAttachment Status
            kind: ConnectNetworkAttachmentStatus
            name: connectnetworkattachmentstatuses.connect.eda.nokia.com
            version: v1
          - description: ConnectNetworkDefinition is the Schema for the connectnetworkdefinitions
              API
            displayName: Connect Network Definition
            kind: ConnectNetworkDefinition
            name: connectnetworkdefinitions.connect.eda.nokia.com
            version: v1
          - description: ConnectControllerConfig is the Schema for the connectcontrollerconfigs
              API
            displayName: Connect Controller Config
            kind: ConnectControllerConfig
            name: connectcontrollerconfigs.config.eda.nokia.com
            version: v1
        description: "K8s operator to deploy eda connect controller to configure EDA
          Connect.\eb"
        displayName: eda-openshift-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - EDA Connect
        - K8s Controller
        - EDA K8s Controller Operator
        links:
        - name: EDA Connect K8s Controller Operator
          url: https://docs.eda.dev/connect/openshift-plugin/
        maintainers:
        - email: support@nokia.com
          name: nokia
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Nokia
          url: https://docs.eda.dev
        relatedImages:
        - ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:c43a7fa7475db83011cc08b728982511fc0279a259e1aacc8ddde9e15f947725
        version: 4.0.0
      entries:
      - name: eda-openshift-operator.v4.0.0
        version: 4.0.0
      name: alpha
    - currentCSV: eda-openshift-operator.v4.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.eda.nokia.com/v1",
                "kind": "ConnectControllerConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "operator"
                  },
                  "name": "connectcontrollerconfig-sample",
                  "namespace": "eda-connect-k8s"
                },
                "spec": {
                  "connectPluginName": "",
                  "controllerImagePullSecret": "",
                  "edaNamespace": "eda",
                  "heartBeat": "",
                  "logLevel": "",
                  "skipTLSVerify": false,
                  "tlsCertificateData": "",
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:ec264b52d581b8fd76e737b9811ab894957a4edf2f627f90bd9fe707890e6bf5
          createdAt: "2025-06-10T15:47:27Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: eda-connect-k8s
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["connectnetworkattachmentstatuses.connect.eda.nokia.com",
            connectnetworkdefinitions.connect.eda.nokia.com]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConnectNetworkAttachmentStatus is the Schema for the connectnetworkattachmentstatuses
              API
            displayName: Connect NetworkAttachment Status
            kind: ConnectNetworkAttachmentStatus
            name: connectnetworkattachmentstatuses.connect.eda.nokia.com
            version: v1
          - description: ConnectNetworkDefinition is the Schema for the connectnetworkdefinitions
              API
            displayName: Connect Network Definition
            kind: ConnectNetworkDefinition
            name: connectnetworkdefinitions.connect.eda.nokia.com
            version: v1
          - description: ConnectControllerConfig is the Schema for the connectcontrollerconfigs
              API
            displayName: Connect Controller Config
            kind: ConnectControllerConfig
            name: connectcontrollerconfigs.config.eda.nokia.com
            version: v1
        description: "K8s operator to deploy eda connect controller to configure EDA
          Connect.\eb"
        displayName: eda-openshift-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - EDA Connect
        - K8s Controller
        - EDA K8s Controller Operator
        links:
        - name: EDA Connect K8s Controller Operator
          url: https://docs.eda.dev/connect/openshift-plugin/
        maintainers:
        - email: support@nokia.com
          name: nokia
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Nokia
          url: https://docs.eda.dev
        relatedImages:
        - ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:ec264b52d581b8fd76e737b9811ab894957a4edf2f627f90bd9fe707890e6bf5
        version: 4.0.1
      entries:
      - name: eda-openshift-operator.v4.0.1
        version: 4.0.1
      name: stable
    defaultChannel: alpha
    packageName: eda-k8s-operator
    provider:
      name: Nokia
      url: https://docs.eda.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: prometurbo-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometurbo-operator.v8.17.40-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5-beta.1 <8.17.40-beta.1'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
          - kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.

          ### Get Started
          Please follow [this](https://www.ibm.com/docs/en/tarm/8.15.x?topic=configuration-prometheus) for an overview of Prometurbo and deployment instructions.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/prometurbo
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/prometurbo-operator@sha256:a1605506884f6bce7471d79c6243728bce991ceb381d7fe0c51f1cdcbeb96e22
        version: 8.17.40-beta.1
      entries:
      - name: prometurbo-operator.v8.17.40-beta.1
        version: 8.17.40-beta.1
      name: beta
    - currentCSV: prometurbo-operator.v8.17.40
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5 <8.17.40'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
          - kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.

          ### Get Started
          Please follow [this](https://www.ibm.com/docs/en/tarm/8.15.x?topic=configuration-prometheus) for an overview of Prometurbo and deployment instructions.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/prometurbo
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/prometurbo-operator@sha256:cbcf90e44ffcedec8e3f161eff4c95b581924a1db195bf55d13da3714346b26f
        version: 8.17.40
      entries:
      - name: prometurbo-operator.v8.17.40
        version: 8.17.40
      name: stable
    defaultChannel: stable
    packageName: prometurbo-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-prometheus-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-prometheus-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:b6071d6a1ec7cdd5f65b6d7296c1e9cd5660972a8ed614da4277e3a4ab5d1df1
          createdAt: "2024-03-01T23:43:00Z"
          description: Manage the full lifecycle of configuring and managing Prometheus
            and Alertmanager servers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/prometheus-operator/prometheus-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Prometheus Operator for Kubernetes provides easy monitoring definitions for Kubernetes services and deployment and management of Prometheus instances.

          Once installed, the Prometheus Operator provides the following features:

          * **Create/Destroy**: Easily launch a Prometheus instance for your Kubernetes namespace, a specific application or team easily using the Operator.
          * **Simple Configuration**: Configure the fundamentals of Prometheus like versions, persistence, retention policies, and replicas from a native Kubernetes resource.
          * **Target Services via Labels**: Automatically generate monitoring target configurations based on familiar Kubernetes label queries; no need to learn a Prometheus specific configuration language.

          ### Other Supported Features

          **High availability**

          Multiple instances are run across failure zones and data is replicated. This keeps your monitoring available during an outage, when you need it most.

          **Updates via automated operations**

          New Prometheus versions are deployed using a rolling update with no downtime, making it easy to stay up to date.

          **Handles the dynamic nature of containers**

          Alerting rules are attached to groups of containers instead of individual instances, which is ideal for the highly dynamic nature of container deployment.
        displayName: Prometheus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - prometheus
        - monitoring
        - tsdb
        - alerting
        - observability
        links:
        - name: Prometheus
          url: https://www.prometheus.io/
        - name: Documentation
          url: https://github.com/prometheus-operator/prometheus-operator/tree/master/Documentation
        - name: Prometheus Operator
          url: https://github.com/prometheus-operator/prometheus-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:b6071d6a1ec7cdd5f65b6d7296c1e9cd5660972a8ed614da4277e3a4ab5d1df1
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:413f7da5cb2e78abf33eb4e43f5fd483c19e791416eec7f5d4c3dbeb54db33a0
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader-rhel9@sha256:eb7614528884f2eefe1ae5e33a09624590c23e0e75510b4d88a274880cf090bc
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager-rhel9@sha256:7d1a3392872fd3990ae3acce79da659ba2238b30421824b7be533dcf8b93e89c
        version: 4.16.16-rhodf
      entries:
      - name: odf-prometheus-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odf-prometheus-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odf-prometheus-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odf-prometheus-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odf-prometheus-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odf-prometheus-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odf-prometheus-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odf-prometheus-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odf-prometheus-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odf-prometheus-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odf-prometheus-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odf-prometheus-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odf-prometheus-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odf-prometheus-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odf-prometheus-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odf-prometheus-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odf-prometheus-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odf-prometheus-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Alvearie
      provider-url: https://alvearie.github.io
    name: alvearie-imaging-ingestion
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: imaging-ingestion-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomEventBridge",
                "metadata": {
                  "name": "events"
                },
                "spec": {
                  "dicomEventDrivenIngestionName": "core",
                  "eventBridge": {},
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "events",
                  "natsTokenSecret": "nats-events-secure-bound-token",
                  "natsUrl": "jetstream.imaging-ingestion.svc.cluster.local:4222",
                  "role": "hub"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomEventDrivenIngestion",
                "metadata": {
                  "name": "core"
                },
                "spec": {
                  "databaseConfigName": "db-config",
                  "databaseSecretName": "db-secret",
                  "eventProcessor": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  },
                  "imagePullPolicy": "Always"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomInstanceBinding",
                "metadata": {
                  "name": "stow"
                },
                "spec": {
                  "bindingConfigName": "instance-binding-config",
                  "bindingSecretName": "instance-binding-secret",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "instanceBinding": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomStudyBinding",
                "metadata": {
                  "name": "fhir"
                },
                "spec": {
                  "bindingConfigName": "study-binding-config",
                  "bindingSecretName": "study-binding-secret",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "studyBinding": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomwebIngestionService",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "bucketConfigName": "imaging-ingestion",
                  "bucketSecretName": "imaging-ingestion",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "providerName": "provider",
                  "stowService": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  },
                  "wadoService": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DimseIngestionService",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "applicationEntityTitle": "DICOM-INGEST",
                  "bucketConfigName": "imaging-ingestion",
                  "bucketSecretName": "imaging-ingestion",
                  "dicomEventDrivenIngestionName": "core",
                  "dimseService": {},
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "DIMSE",
                  "natsTokenSecret": "ingestion-nats-secure-bound-token",
                  "natsUrl": "nats-secure.imaging-ingestion.svc.cluster.local:4222",
                  "providerName": "provider"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DimseProxy",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "applicationEntityTitle": "DCM4CHEE",
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "DIMSE",
                  "natsTokenSecret": "ingestion-nats-secure-bound-token",
                  "natsUrl": "nats-secure.imaging-ingestion.svc.cluster.local:4222",
                  "proxy": {},
                  "targetDimseHost": "arc.dcm4chee.svc.cluster.local",
                  "targetDimsePort": 11112
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: alvearie/imaging-ingestion-operator:0.0.3
          createdAt: "2021-04-08T15:29:55Z"
          description: The Alvearie Imaging Ingestion provides a collection of components
            for extending enterprise medical imaging informatics systems to Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Alvearie/imaging-ingestion
          support: alvearie.slack.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DicomEventBridge is the Schema for the dicomeventbridges
              API
            displayName: Dicom Event Bridge
            kind: DicomEventBridge
            name: dicomeventbridges.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Event driven manifest of all DICOM data across all associated
              S3 buckets
            displayName: Dicom Event Driven Ingestion
            kind: DicomEventDrivenIngestion
            name: dicomeventdriveningestions.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Fan-out notification of DICOM instance data (STOW-RS)
            displayName: Dicom Instance Binding
            kind: DicomInstanceBinding
            name: dicominstancebindings.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Fan-out notification of DICOM studies (FHIRv4 ImagingStudy)
            displayName: Dicom Study Binding
            kind: DicomStudyBinding
            name: dicomstudybindings.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides DICOMweb WADO-RS and STOW-RS access to a S3 bucket
            displayName: Dicomweb Ingestion Service
            kind: DicomwebIngestionService
            name: dicomwebingestionservices.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides a proxied DIMSE Application Entity (AE) in the cluster
              for C-STORE operations to a storage space
            displayName: Dimse Ingestion Service
            kind: DimseIngestionService
            name: dimseingestionservices.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides a bidirectional proxied DIMSE Application Entity
              (AE) in the cluster
            displayName: Dimse Proxy
            kind: DimseProxy
            name: dimseproxies.imaging-ingestion.alvearie.org
            version: v1alpha1
        description: |
          The Alvearie Imaging Ingestion provides a collection of components for extending enterprise medical imaging informatics systems to Kubernetes.

          ## About this Operator
          This operator enables standards-based medical imaging application interfaces such as DICOM Message Service Element (DIMSE) and DICOMweb.
          The API's provided allow for extending enterprise imaging systems to the cloud with custom, fit for purpose deployments.
          This is possible at speed, low cost, and little or zero integration effort.or extending enterprise medical imaging informatics systems to a modern cloud.


          Some potential usages:

          - Perform lightweight secure bridging of DICOM from the enterprise imaging zone to a cloud availability zone.
          - Utilize S3 buckets for segregating DICOM data to be used for clinical or scientific purposes.
          - Fan-out DICOM data to a number of different DICOM subscribers
          - Generate FHIRv4 *ImagingStudy* resources from ingested content and publish to a FHIRv4 based patient logitudinal record.
          - Distribute imaging study arrival and update notifications to imaging study subscribers


          ## Prerequisites

          - [Knative](https://operatorhub.io/operator/knative-operator)
          - ProgresSQL such as [Crunchy](https://operatorhub.io/operator/postgresql)
          - S3 storage such as [MinIO](https://operatorhub.io/operator/minio-operator)
          - [NATS operator](https://github.com/nats-io/nats-operator/) (Optional:  only required for DIMSE Proxy component)



          ## Documentation

          Visit the [official github repository](https://github.com/Alvearie/imaging-ingestion) for documentation and examples.
        displayName: Alvearie Imaging Ingestion Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - dicom
        - medical imaging
        - fhir
        - dimse-c
        - dimse
        - imaging study
        links:
        - name: Alvearie Imaging Ingestion Operator
          url: https://github.com/Alvearie/imaging-ingestion
        maintainers:
        - email: rduggan@ca.ibm.com
          name: Richard Duggan
        - email: jjacob@us.ibm.com
          name: Jeesmon Jacob
        maturity: alpha
        provider:
          name: Alvearie
          url: https://alvearie.github.io
        relatedImages:
        - alvearie/imaging-ingestion-operator:0.0.3
        version: 0.0.3
      entries:
      - name: imaging-ingestion-operator.v0.0.3
        version: 0.0.3
      - name: imaging-ingestion-operator.v0.0.2
        version: 0.0.2
      - name: imaging-ingestion-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: alvearie-imaging-ingestion
    provider:
      name: Alvearie
      url: https://alvearie.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solace Corporation
      provider-url: www.solace.com
    name: pubsubplus-eventbroker-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pubsubplus-eventbroker-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pubsubplus.solace.com/v1beta1",
                "kind": "PubSubPlusEventBroker",
                "metadata": {
                  "name": "non-ha-standalone-example"
                },
                "spec": {
                  "redundancy": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "true"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.10
          containerImage: registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator@sha256:d68db92bcb217d94dc0ad0a1cea625a0a9c9299d886d325a903163f9f887dcd5
          createdAt: "2025-07-31T19:48:36Z"
          description: The Solace PubSub+ Event Broker Operator deploys and manages
            the lifecycle of PubSub+ Event Brokers
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
          support: Solace Products
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PubSub+ Event Broker
            displayName: PubSub+ Event Broker
            kind: PubSubPlusEventBroker
            name: pubsubpluseventbrokers.pubsubplus.solace.com
            version: v1beta1
        description: |
          ## Solace PubSub+ Platform
          Solace [PubSub+ Platform](https://solace.com/products/platform/) is a complete event streaming and management platform
          for real-time enterprises. The [PubSub+ Software Event Broker](https://solace.com/products/event-broker/software/) efficiently streams
          event-driven information between applications, IoT devices, and user interfaces running in the cloud, on-premises,
          and in hybrid environments using open APIs and protocols like AMQP, JMS, MQTT, REST and WebSocket.
          It can be installed into a variety of public and private clouds, PaaS, and on-premises environments.
          Event brokers in multiple locations can be linked together in an [Event Mesh](https://solace.com/what-is-an-event-mesh/)
          to dynamically share events across the distributed enterprise.

          ## Solace PubSub+ Event Broker Operator
          Solace Pubsub+ Event Broker Operator automatically deploys and manages PubSub+ Software Event Brokers on Kubernetes and OpenShift environments.

          Features include:
          * Initial deployment configuration: Production-ready HA, non-HA or for developers
          * Rolling update for configuration changes or broker upgrades
          * Prometheus monitoring support

          ## Getting started
          Refer to the [Solace PubSub+ Event Broker Operator Quick Start](https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/README.md)
        displayName: Solace PubSub+ Event Broker Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - solace
        - pubsubplus
        - pubsub+
        - pubsub
        - messaging
        - advanced event broker
        - event broker
        - event mesh
        - event streaming
        - data streaming
        - event integration
        - middleware
        links:
        - name: Operator QuickStart
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
        - name: Operator Documentation
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/docs/EventBrokerOperatorUserGuide.md
        - name: Solace.com
          url: https://solace.com
        - name: Understanding Solace Technology
          url: https://solace.com/products/platform/
        - name: Solace PubSub+ Technical Documentation
          url: https://docs.solace.com/
        - name: Solace Developer Portal
          url: https://www.solace.dev/
        - name: Solace Community
          url: https://solace.community/
        maintainers:
        - email: support@solace.com
          name: Solace R&D
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Solace Corporation
          url: www.solace.com
        relatedImages:
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator@sha256:d68db92bcb217d94dc0ad0a1cea625a0a9c9299d886d325a903163f9f887dcd5
        version: 1.4.0
      entries:
      - name: pubsubplus-eventbroker-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: pubsubplus-eventbroker-operator
    provider:
      name: Solace Corporation
      url: www.solace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: StreamNative
      provider-url: https://streamnative.io
    name: function-mesh
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: function-mesh.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "java-function-sample",
                  "namespace": "default"
                },
                "spec": {
                  "autoAck": true,
                  "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                  "clusterName": "test-pulsar",
                  "forwardSourceMessageProperty": true,
                  "input": {
                    "topics": [
                      "persistent://public/default/java-function-input-topic"
                    ],
                    "typeClassName": "java.lang.String"
                  },
                  "java": {
                    "extraDependenciesDir": "random-dir/",
                    "jar": "pulsar-functions-api-examples.jar",
                    "jarLocation": "public/default/nlu-test-java-function"
                  },
                  "logTopic": "persistent://public/default/logging-function-logs",
                  "maxPendingAsyncRequests": 1000,
                  "maxReplicas": 5,
                  "output": {
                    "topic": "persistent://public/default/java-function-output-topic",
                    "typeClassName": "java.lang.String"
                  },
                  "pod": {
                    "annotations": {
                      "managed-function": "true"
                    },
                    "env": [
                      {
                        "name": "EXAMPLE_VARIABLE",
                        "value": "exampleValue"
                      }
                    ],
                    "imagePullSecrets": [
                      {
                        "name": "regcred"
                      }
                    ],
                    "initContainers": [
                      {
                        "command": [
                          "sh",
                          "-c",
                          "echo The app is running! \u0026\u0026 sleep 30"
                        ],
                        "image": "busybox:1.28",
                        "name": "init-function"
                      }
                    ],
                    "labels": {
                      "locaction": "mtv"
                    },
                    "sidecars": [
                      {
                        "command": [
                          "sh",
                          "-c",
                          "echo The app is running! \u0026\u0026 sleep 30000"
                        ],
                        "image": "busybox:1.28",
                        "name": "sidecar-function"
                      }
                    ],
                    "volumes": [
                      {
                        "emptyDir": {},
                        "name": "cache-volume"
                      }
                    ]
                  },
                  "pulsar": {
                    "authSecret": "test-auth",
                    "pulsarConfig": "test-pulsar",
                    "tlsSecret": "test-tls"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "secretsMap": {
                    "name": {
                      "key": "username",
                      "path": "test-secret"
                    },
                    "pwd": {
                      "key": "password",
                      "path": "test-secret"
                    }
                  },
                  "volumeMounts": [
                    {
                      "mountPath": "/cache",
                      "name": "cache-volume"
                    }
                  ]
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "FunctionMesh",
                "metadata": {
                  "name": "functionmesh-sample"
                },
                "spec": {
                  "functions": [
                    {
                      "autoAck": true,
                      "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                      "clusterName": "test-pulsar",
                      "forwardSourceMessageProperty": true,
                      "input": {
                        "topics": [
                          "persistent://public/default/functionmesh-input-topic"
                        ],
                        "typeClassName": "java.lang.String"
                      },
                      "java": {
                        "jar": "pulsar-functions-api-examples.jar",
                        "jarLocation": "public/default/nlu-test-functionmesh-ex1"
                      },
                      "logTopic": "persistent://public/default/logging-function-log",
                      "maxReplicas": 1,
                      "name": "ex1",
                      "output": {
                        "topic": "persistent://public/default/mid-topic",
                        "typeClassName": "java.lang.String"
                      },
                      "pulsar": {
                        "pulsarConfig": "mesh-test-pulsar"
                      },
                      "replicas": 1,
                      "resources": {
                        "limits": {
                          "cpu": "0.2",
                          "memory": "1.1G"
                        },
                        "requests": {
                          "cpu": "0.1",
                          "memory": "1G"
                        }
                      }
                    },
                    {
                      "autoAck": true,
                      "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                      "clusterName": "test-pulsar",
                      "forwardSourceMessageProperty": true,
                      "input": {
                        "topics": [
                          "persistent://public/default/mid-topic"
                        ],
                        "typeClassName": "java.lang.String"
                      },
                      "java": {
                        "jar": "pulsar-functions-api-examples.jar",
                        "jarLocation": "public/default/nlu-test-functionmesh-ex2"
                      },
                      "logTopic": "persistent://public/default/logging-function-logs",
                      "maxReplicas": 1,
                      "name": "ex2",
                      "output": {
                        "topic": "persistent://public/default/functionmesh-output-topic",
                        "typeClassName": "java.lang.String"
                      },
                      "pulsar": {
                        "pulsarConfig": "mesh-test-pulsar"
                      },
                      "replicas": 1,
                      "resources": {
                        "limits": {
                          "cpu": "0.2",
                          "memory": "1.1G"
                        },
                        "requests": {
                          "cpu": "0.1",
                          "memory": "1G"
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Sink",
                "metadata": {
                  "name": "sink-sample"
                },
                "spec": {
                  "autoAck": true,
                  "className": "org.apache.pulsar.io.elasticsearch.ElasticSearchSink",
                  "clusterName": "test-pulsar",
                  "image": "streamnative/pulsar-io-elastic-search:2.10.0.0-rc10",
                  "input": {
                    "topics": [
                      "persistent://public/default/input"
                    ],
                    "typeClassName": "[B"
                  },
                  "java": {
                    "jar": "connectors/pulsar-io-elastic-search-2.10.0.0-rc10.nar",
                    "jarLocation": ""
                  },
                  "maxReplicas": 1,
                  "pulsar": {
                    "pulsarConfig": "test-sink"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "sinkConfig": {
                    "elasticSearchUrl": "http://quickstart-es-http.default.svc.cluster.local:9200",
                    "indexName": "my_index",
                    "password": "wJ757TmoXEd941kXm07Z2GW3",
                    "typeName": "doc",
                    "username": "elastic"
                  }
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Source",
                "metadata": {
                  "name": "source-sample"
                },
                "spec": {
                  "className": "org.apache.pulsar.io.debezium.mongodb.DebeziumMongoDbSource",
                  "clusterName": "test-pulsar",
                  "forwardSourceMessageProperty": true,
                  "image": "streamnative/pulsar-io-debezium-mongodb:2.10.0.0-rc10",
                  "java": {
                    "jar": "connectors/pulsar-io-debezium-mongodb-2.10.0.0-rc10.nar",
                    "jarLocation": ""
                  },
                  "maxReplicas": 1,
                  "output": {
                    "producerConf": {
                      "maxPendingMessages": 1000,
                      "maxPendingMessagesAcrossPartitions": 50000,
                      "useThreadLocalProducers": true
                    },
                    "topic": "persistent://public/default/destination",
                    "typeClassName": "org.apache.pulsar.common.schema.KeyValue"
                  },
                  "pulsar": {
                    "pulsarConfig": "test-source"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "sourceConfig": {
                    "database.whitelist": "inventory",
                    "mongodb.hosts": "rs0/mongo-dbz-0.mongo.default.svc.cluster.local:27017,rs0/mongo-dbz-1.mongo.default.svc.cluster.local:27017,rs0/mongo-dbz-2.mongo.default.svc.cluster.local:27017",
                    "mongodb.name": "dbserver1",
                    "mongodb.password": "dbz",
                    "mongodb.task.id": "1",
                    "mongodb.user": "debezium",
                    "pulsar.service.url": "pulsar://test-pulsar-broker.default.svc.cluster.local:6650"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/function-mesh@sha256:3e13e5cc3ac4d7387ee0adc02cb5183aad6cf2301f76ab230482609c4da1c0b7
          createdAt: "2022-10-27T07:49:28Z"
          description: The Function Mesh Operator manages the Pulsar Functions and
            Connectors deployed on a Kubernetes cluster.
          olm.properties: '[{"type":"olm.maxOpenShiftVersion","value":"4.8"}]'
          olm.skipRange: <0.7.0
          operatorhub.io/ui-metadata-max-k8s-version: "1.21"
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/streamnative/function-mesh
          support: StreamNative
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FunctionMesh is the Schema for the functionmeshes API
            displayName: Mesh
            kind: FunctionMesh
            name: functionmeshes.compute.functionmesh.io
            version: v1alpha1
          - description: Function is the Schema for the functions API
            displayName: Function
            kind: Function
            name: functions.compute.functionmesh.io
            version: v1alpha1
          - description: Sink is the Schema for the sinks API
            displayName: Sink
            kind: Sink
            name: sinks.compute.functionmesh.io
            version: v1alpha1
          - description: Source is the Schema for the sources API
            displayName: Source
            kind: Source
            name: sources.compute.functionmesh.io
            version: v1alpha1
        description: |
          [Function Mesh](https://functionmesh.io/) is a serverless framework purpose-built for stream processing applications. It brings powerful event-streaming capabilities to your applications by orchestrating multiple [Pulsar Functions](http://pulsar.apache.org/docs/en/next/functions-overview/) and [Pulsar IO connectors](http://pulsar.apache.org/docs/en/next/io-overview/) for complex stream processing jobs on Kubernetes.
          Prerequisites:
          - Install cert-manager operator first.
        displayName: Function Mesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pulsar
        - function-mesh
        - serverless
        - pulsar-functions
        - pulsar-io-connectors
        - StreamNative
        links:
        - name: Function Mesh
          url: https://function-mesh.io
        maintainers:
        - email: function-mesh@streamnative.io
          name: Function Mesh
        maturity: alpha
        minKubeVersion: v1.17.0
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/function-mesh@sha256:3e13e5cc3ac4d7387ee0adc02cb5183aad6cf2301f76ab230482609c4da1c0b7
        - docker.cloudsmith.io/streamnative/mirrors/gcr.io/kubebuilder/kube-rbac-proxy@sha256:67ecb332573384515406ebd71816781366b70adb0eb66345e5980e92603373e1
        version: 0.7.0
      entries:
      - name: function-mesh.v0.7.0
        version: 0.7.0
      name: alpha
    defaultChannel: alpha
    packageName: function-mesh
    provider:
      name: StreamNative
      url: https://streamnative.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.3.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo-operator@sha256:32178a16dd2115e0a11a2e3dfbfc130ae1d0afd8e293131e76b6fc1a79b36c12
          createdAt: "2019-05-01T00:00:00Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/kubeturbo-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/kubeturbo-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo-operator@sha256:32178a16dd2115e0a11a2e3dfbfc130ae1d0afd8e293131e76b6fc1a79b36c12
        - registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo@sha256:671be694176d9cc31b8a22c11bbfefc2504ae72d27fd010b315b6839f9ab59d5
        version: 8.3.3
      entries:
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: stable
    defaultChannel: stable
    packageName: kubeturbo-certified-rhmp
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: devspaces
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devspacesoperator.v3.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "org.eclipse.che/v1",
                "kind": "CheCluster",
                "metadata": {
                  "name": "devspaces",
                  "namespace": "openshift-devspaces"
                },
                "spec": {
                  "auth": {
                    "identityProviderURL": "",
                    "oAuthClientName": "",
                    "oAuthSecret": ""
                  },
                  "database": {
                    "externalDb": false
                  },
                  "k8s": {
                    "ingressDomain": null,
                    "tlsSecretName": null
                  },
                  "metrics": {
                    "enable": true
                  },
                  "server": {
                    "workspaceNamespaceDefault": "<username>-devspaces"
                  },
                  "storage": {
                    "pvcStrategy": "common"
                  }
                }
              },
              {
                "apiVersion": "org.eclipse.che/v2",
                "kind": "CheCluster",
                "metadata": {
                  "name": "devspaces",
                  "namespace": "openshift-devspaces"
                },
                "spec": {
                  "components": {},
                  "containerRegistry": {},
                  "devEnvironments": {},
                  "gitServices": {},
                  "networking": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/devspaces/devspaces-rhel9-operator@sha256:39c95279bced68928d028ef026f82eafe53eddb3dbe57703c8007cb5a170a70a
          createdAt: "2025-09-11T22:30:49+00:00"
          description: Devfile v2 and v1 development solution, 1 instance per cluster,
            for portable, collaborative k8s workspaces.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/devspaces-images/
          skipRange: '>=3.0.0 <3.23.0'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              The `CheCluster` custom resource allows defining and managing Red Hat OpenShift Dev Spaces server installation.
              Based on these settings, the  Operator automatically creates and maintains several ConfigMaps:
              `che`, `plugin-registry` that will contain the appropriate environment variables
              of the various components of the installation. These generated ConfigMaps must NOT be updated manually.
            displayName: Red Hat OpenShift Dev Spaces instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v2
          - description: The `CheCluster` custom resource allows defining and managing
              a Che server installation
            displayName: Red Hat OpenShift Dev Spaces instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v1
        description: |
          A collaborative Kubernetes-native development solution that delivers OpenShift workspaces and in-browser IDE for rapid cloud application development.
          This operator installs the Plugin and Devfile registries, Dashboard, Gateway and the Red Hat OpenShift Dev Spaces server, and configures these services.
          OpenShift OAuth is used directly for authentication. TLS mode is on.

          ## How to Install
          Press the **Install** button, choose the channel and the upgrade strategy, and wait for the **Installed** Operator status.
          When the operator is installed, create a new Custom Resource (CR) of Kind CheCluster (click the **Create New** button).
          The CR spec contains all defaults. You can start using Red Hat OpenShift Dev Spaces when the CR status is set to **Available**, and you see a URL to Red Hat OpenShift Dev Spaces.

          ## Defaults
          By default, the operator deploys Red Hat OpenShift Dev Spaces with:
          * 10Gi storage
          * Auto-generated passwords
          * Bundled Plugin and Devfile registries

          Use `oc edit checluster/devspaces -n openshift-devspaces` to update Red Hat OpenShift Dev Spaces default installation options.
          See more in the [Installation guide](https://docs.redhat.com/en/documentation/red_hat_openshift_dev_spaces/3.23/html/administration_guide/configuring-devspaces).

          ### Certificates
          Operator uses a default router certificate to secure Red Hat OpenShift Dev Spaces routes.
          Follow the [guide](https://docs.redhat.com/en/documentation/red_hat_openshift_dev_spaces/3.23/html/administration_guide/configuring-devspaces#importing-untrusted-tls-certificates)
          to import certificates into Red Hat OpenShift Dev Spaces.
        displayName: Red Hat OpenShift Dev Spaces
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eclipse che
        - workspaces
        - devtools
        - developer
        - ide
        - java
        - devspaces
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/openshift-dev-spaces/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_openshift_dev_spaces
        - name: Operator GitHub Repo
          url: https://github.com/redhat-developer/devspaces-images/
        maintainers:
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/devspaces/pluginregistry-rhel9@sha256:3ee131041761f5250b29d497043ecf7820e02e03e73ed0f929c17e2247c0d7a3
        - registry.redhat.io/devspaces/traefik-rhel9@sha256:95cf6a28028ec7a546a3d8d78377f204bb559c3d1022a73f08c646e76b85ce85
        - registry.redhat.io/devspaces-tech-preview/jetbrains-ide-rhel9@sha256:46bf7bea716df74f14d2629fb7e0e2ab692564cec348ffb8a6f4fea1739b23c2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5a6edfecf1636805e6f0bcad02df10bceec764f0580f3ac8fe62aba978d50006
        - registry.redhat.io/devspaces/server-rhel9@sha256:c0ec8c792f88aacf6ee31a4b6d2d847d864707a084148b48e4e402388989f868
        - registry.redhat.io/devspaces/devspaces-rhel9-operator@sha256:39c95279bced68928d028ef026f82eafe53eddb3dbe57703c8007cb5a170a70a
        - registry.redhat.io/devspaces/code-rhel9@sha256:abc9b12131c8cdca0719e0404d354310f53769990b089f1a3adea31193798a8e
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6bcfa981a7a75c242a16afa6dab36bf6e3b8a1442fa863875919ad551f48eceb
        - registry.redhat.io/devspaces/configbump-rhel9@sha256:cecbc3796058618da9ad0473a6f937aae5aca25776bb434894286f22baa4677c
        - registry.redhat.io/devspaces/dashboard-rhel9@sha256:c34cf5ac9bf95c1c5082afd5a2941b5e412be8ca43aef6a72e8fe5a279a43271
        - registry.redhat.io/devspaces/udi-rhel9@sha256:f7325ee7259a0585e8017d4442d778a37f4c8a67d81fe70c241a014001e37539
        - registry.redhat.io/devspaces-tech-preview/idea-rhel9@sha256:df24ebb239a2cc0d4543309f51a7268c7d524d428a536e0b3443c51089c3f87e
        version: 3.23.0
      entries:
      - name: devspacesoperator.v3.23.0
        version: 3.23.0
      - name: devspacesoperator.v3.22.0
        version: 3.22.0
      - name: devspacesoperator.v3.21.0
        version: 3.21.0
      - name: devspacesoperator.v3.20.0
        version: 3.20.0
      - name: devspacesoperator.v3.19.0
        version: 3.19.0
      - name: devspacesoperator.v3.18.1-0.1739475915.p
        version: 3.18.1+0.1739475915.p
      - name: devspacesoperator.v3.18.1
        version: 3.18.1
      - name: devspacesoperator.v3.18.0
        version: 3.18.0
      - name: devspacesoperator.v3.17.0
        version: 3.17.0
      - name: devspacesoperator.v3.16.1
        version: 3.16.1
      - name: devspacesoperator.v3.16.0
        version: 3.16.0
      - name: devspacesoperator.v3.15.0
        version: 3.15.0
      - name: devspacesoperator.v3.14.0
        version: 3.14.0
      - name: devspacesoperator.v3.13.0
        version: 3.13.0
      - name: devspacesoperator.v3.12.0-0.1712003563.p
        version: 3.12.0+0.1712003563.p
      - name: devspacesoperator.v3.12.0
        version: 3.12.0
      - name: devspacesoperator.v3.11.0
        version: 3.11.0
      - name: devspacesoperator.v3.10.1
        version: 3.10.1
      - name: devspacesoperator.v3.10.0
        version: 3.10.0
      - name: devspacesoperator.v3.9.1
        version: 3.9.1
      - name: devspacesoperator.v3.9.0
        version: 3.9.0
      - name: devspacesoperator.v3.8.0
        version: 3.8.0
      - name: devspacesoperator.v3.7.2
        version: 3.7.2
      - name: devspacesoperator.v3.7.1-0.1691407091.p
        version: 3.7.1+0.1691407091.p
      - name: devspacesoperator.v3.7.1
        version: 3.7.1
      - name: devspacesoperator.v3.7.1-0.1691548593.p
        version: 3.7.1+0.1691548593.p
      - name: devspacesoperator.v3.7.0
        version: 3.7.0
      - name: devspacesoperator.v3.6.0
        version: 3.6.0
      - name: devspacesoperator.v3.6.0-0.1685320856.p
        version: 3.6.0+0.1685320856.p
      - name: devspacesoperator.v3.6.0-0.1685322634.p
        version: 3.6.0+0.1685322634.p
      - name: devspacesoperator.v3.5.0
        version: 3.5.0
      - name: devspacesoperator.v3.4.0-0.1675914643.p
        version: 3.4.0+0.1675914643.p
      - name: devspacesoperator.v3.4.0
        version: 3.4.0
      - name: devspacesoperator.v3.3.0-0.1673528130.p
        version: 3.3.0+0.1673528130.p
      - name: devspacesoperator.v3.3.0
        version: 3.3.0
      name: stable
    defaultChannel: stable
    packageName: devspaces
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Jooho Lee
      provider-url: https://github.com/jooho/nfs-provisioner-operator
    name: nfs-provisioner-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfs-provisioner-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.jhouse.com/v1alpha1",
                "kind": "NFSProvisioner",
                "metadata": {
                  "name": "nfsprovisioner-sample"
                },
                "spec": {
                  "nfsImageConfiguration": {
                    "image": "k8s.gcr.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439",
                    "imagePullPolicy": "IfNotPresent"
                  },
                  "scForNFS": "nfs",
                  "scForNFSPvc": "local-sc",
                  "storageSize": "1G"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: quay.io/jooholee/nfs-provisioner-operator@sha256:d9c013967421ec72644a588a155975bf856a856f1ef38ba71dfea306fdb47acd
          createdAt: "2025-08-08T20:23:50Z"
          description: Create and manage NFS Server and Provisioner
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/jooho/nfs-provisioner-operator
          support: NFS
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NFSProvisioner is the Schema for the nfsprovisioners API
            displayName: NFS Provisioner App
            kind: NFSProvisioner
            name: nfsprovisioners.cache.jhouse.com
            version: v1alpha1
        description: "This operator deploy NFS server with local storage and also
          provide provisioner for storageClass.\n### Core Capabilities\n* **NFS Server:**
          Deployed\n* **NFS Provisioner:** Help customers to create PV using StorageClass\n*
          **StorageClass:** Dynamically create PV for requested PVC\n### NFS Provisioner
          Operator Features\n* NFS Server can use localStorage PVC or HostPath on
          the node\n### Must-Gather \n> isv-cli must-gather --image=quay.io/jooholee/nfs-provisioner-operator-must-gather:0.0.1\n"
        displayName: NFS Provisioner Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nfs
        - storage
        links:
        - name: Nfs Provisioner Operator
          url: https://github.com/jooho/nfs-provisioner-operator
        maintainers:
        - email: ljhiyh@gmail.com
          name: Jooho
        maturity: alpha
        provider:
          name: Jooho Lee
          url: https://github.com/jooho/nfs-provisioner-operator
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - quay.io/jooholee/nfs-provisioner-operator@sha256:d9c013967421ec72644a588a155975bf856a856f1ef38ba71dfea306fdb47acd
        version: 0.0.8
      entries:
      - name: nfs-provisioner-operator.v0.0.8
        version: 0.0.8
      - name: nfs-provisioner-operator.v0.0.7
        version: 0.0.7
      - name: nfs-provisioner-operator.v0.0.6
        version: 0.0.6
      - name: nfs-provisioner-operator.v0.0.5
        version: 0.0.5
      - name: nfs-provisioner-operator.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: nfs-provisioner-operator
    provider:
      name: Jooho Lee
      url: https://github.com/jooho/nfs-provisioner-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-licensing
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-licensing.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Licensing",
                "metadata": {
                  "name": "licensing-sample"
                },
                "spec": {
                  "environment": {
                    "deploymentType": "openshift"
                  },
                  "hostid": {
                    "enabled": true,
                    "environment": {},
                    "hostMounts": {},
                    "image": "prod.harbor.keysight.digital/loadcore/kcos-licensing-hostid-vmuuid@sha256:076a2c6aeec875470414af7905613a25b56bcb2aca5edcb1c4c53037cd01b332",
                    "imagePullPolicy": "IfNotPresent"
                  },
                  "image": {
                    "dbimage": "prod.harbor.keysight.digital/loadcore/postgresql-ubi-image@sha256:17f6a8f775d06650939d2c73e82b45d73d50190cc735ca70695d082f7790834b",
                    "initOwnership": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf",
                    "repository": "prod.harbor.keysight.digital/loadcore"
                  },
                  "ingress": {
                    "annotations": {
                      "nginx.ingress.kubernetes.io/auth-method": "POST",
                      "nginx.ingress.kubernetes.io/auth-response-headers": "X-AUTH-USER",
                      "nginx.ingress.kubernetes.io/auth-url": "https://127.0.0.1/token/validate"
                    },
                    "enabled": true
                  },
                  "nodeSelector": {
                    "node-role.kubernetes.io/kcos-licensing": "kcos-licensing"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:29:04Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Licensing
            name: licensings.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Licensing Operator
        displayName: Keysight KCOS Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Licensing
          url: https://keysight-kcos-licensing.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-licensing@sha256:a09291923ad8d7533b33e40e1a3c4e735c2a4ae3b2b87be7c9d1a1a9244b3dad
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - prod.harbor.keysight.digital/loadcore/postgresql-ubi-image@sha256:17f6a8f775d06650939d2c73e82b45d73d50190cc735ca70695d082f7790834b
        - prod.harbor.keysight.digital/loadcore/kcos-licensing-hostid-vmuuid@sha256:076a2c6aeec875470414af7905613a25b56bcb2aca5edcb1c4c53037cd01b332
        version: 0.1.21
      entries:
      - name: keysight-kcos-licensing.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-licensing
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: multicluster-engine
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-engine.v2.5.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-08-06T12:21:10Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0 <2.5.10'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '[]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine is the Schema for the multiclusterengines
              API
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:80f33a762535ff69f57ecb04de043b8fe249a63767712f4942ac6685450e8f28
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:1af88c3703b668020137f90657e6262b73e407aa858bbc6d9c12ba3a1cab8043
        - registry.redhat.io/multicluster-engine/agent-service-rhel8@sha256:92464c57700dcb13ccf5431c104a3fba6845f651776aa5beb1ce7171087dc1ea
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:cb2c57c51edc8a2320a10ac217cca6a401db09de55c0c9d0c9a7b4545e2636e9
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:7f32ed997e03dba1746b3627d5a33fc69282b2e831edd21ec74020fe4fa1ca5d
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:90e792c92ef8c75ffb97545394206c9197997d474493d93f09fca273f65133e4
        - registry.redhat.io/multicluster-engine/hive-rhel8@sha256:27eefe6ad7d93e752f4d76b776884129615c2e87ff24116c401a886e98abfcc5
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:1e9eff48ea837a1d19c595986d765d74810d6d1397602ad79e2e820b1cbb5c6e
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:297d8d0d5855db90ad40af702d6ad5f3f32a4aab486686919d86bbd21c367ca9
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:b0a70eee404ba1267c0ae1a8eb140251d054e722305d680685223cfda360aa37
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:53b8d48e540c3d9bb5d948af0375c26377271e595fffdb3d5f921e3a358fa298
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:36abf06619cd3971da8f42f4193621a1384970365fce0f48e86a65e2840bca38
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel8@sha256:360d3dfb823d8e199574cf1c3fbf890fa74c71fc48644e22b47339363e941b3e
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:c8d48e7ccdce43ce72154e464e74c32477efb0094aab327ed0eaab6c1244774a
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:515eda1e4e1ced222709103789af14f7b4e99d4da037deddd8499aad3b45451b
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:3e3f7ee898e0f6b0f74026399753d7fde41013ade76f7963d5c25e578eb1906e
        - registry.redhat.io/rhel8/postgresql-12@sha256:da619f6617d99790b4fe00fdb23ab4aea0cdb0592590162d51b3ef94aca01a7b
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel8@sha256:6bf850d8fb250740113bfb004b19edbd5cf560c8175b81e2f3766280b3482e27
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:43cc171404571bb3c87275a86e9ab4409861c1c01b5f1441c455565d56e1921d
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:d71c1441577b1cd360d074581ad19c2b0cfa99cdfe1869af1777ff7de6821735
        - registry.redhat.io/multicluster-engine/cluster-api-provider-aws-rhel9@sha256:7f65119dc5d81b8ba485b651efe1d9013b62ec57e4725dbf8bdc2e3b7748f677
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:1a5ff852d96bd15c631a6f89e2b0047cd8dde6264646f51106f2b523decc64ba
        - registry.redhat.io/multicluster-engine/aws-encryption-provider-rhel9@sha256:22cefc43e205e4bdd7874c7e1fccba60fe556df4b2bab055d01c834d62f194bd
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:53ad9402a0a13c894b485f418dd1353c8870792893a5ce33f9d6c3cc9839ebc4
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:2497e077a00dfa94b913359be0c696126e932db1a45e066ead65dee76a1db8fd
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel8@sha256:1e755b64358c10309804a3bf44ca640a7d5bb6e0474d33774a8d76095f2b8c3d
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:25b30d4f4ccc23fda707db88969c0ae036ecb4a2baf15b82221e3351c23292f9
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:a4f58b94d45c1ae1c4ba69b12030b9945b37f3ce6e96c27919f256ae887ec28b
        - registry.redhat.io/multicluster-engine/cluster-api-rhel9@sha256:0e329ca39b1f18520da093a5194ef6e1d5527270c74bf260517620577a21b6c7
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:19dba8be2295f54a93853dbe566132011f0120b2c4b48de2816509f5d201e420
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:bd6d0189ea2bbc87b5745f9d7f9c82118f8dcc9ee63fe0a36bb5e7dfc0e56021
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:424ca238d9c8b20dc434a07aa6a27fe6e7c63b3e2480eead284570864d682175
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:3d420ba8bc5e1dbe629c0d4aa65c0a79a7914ac349828046994e245f4f2adb2b
        - registry.redhat.io/multicluster-engine/assisted-installer-reporter-rhel8@sha256:5d41ca13f4ae102760b65375578c2f632aa6ce41092732300d6727f195da3d55
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:460d37a0032813aa85a3b18bd6f152e4ed6b89f802d17d787c3365082420d379
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:d7ccc8ddbeb11ac8a3a0d4946920a0ab87d9e9595d7c61740bd57f3f4a19a711
        version: 2.5.10
      entries:
      - name: multicluster-engine.v2.5.10
        version: 2.5.10
      - name: multicluster-engine.v2.5.9
        version: 2.5.9
      - name: multicluster-engine.v2.5.8
        version: 2.5.8
      - name: multicluster-engine.v2.5.7
        version: 2.5.7
      - name: multicluster-engine.v2.5.6
        version: 2.5.6
      - name: multicluster-engine.v2.5.5
        version: 2.5.5
      - name: multicluster-engine.v2.5.4
        version: 2.5.4
      - name: multicluster-engine.v2.5.3
        version: 2.5.3
      - name: multicluster-engine.v2.5.2
        version: 2.5.2
      - name: multicluster-engine.v2.5.1
        version: 2.5.1
      - name: multicluster-engine.v2.5.0-103
        version: 2.5.0-103
      name: stable-2.5
    - currentCSV: multicluster-engine.v2.6.8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-08-04T11:16:16Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0 <2.6.8'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:69a40339bf3b3dd3c3b6359721895313147a5f682b62faf5151288ba63b65812
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:f6b15f827aeabc33074de1ff9ed20edaf04271b4e4d2eb16abbe786c78e033ca
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:9e1ea96b998b0c7854ca3c7c8a505fc96cbc1b35e6ce9f49b1dbec703fdc8070
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:7dde47f7fe16a26676a6e5b1739954d76cad1cddbefce2d6869ffb2af399e403
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:f563ff0acb4ea9d1bb112765b67bee535063304a98a2aa2cb7f21b0155d49262
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:54b8e17cacae2d355d82677ed334b4981b50e825579ff30406039933833e828b
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:a8117c99b33eedc702baeab32f0347892b587dfc1860f467cb1900565c3fdb3b
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:00ea8fe4945e93d2cbeceda4440acd6a53a48b4dd420dfe2cbd5287e20f0362f
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel8@sha256:42fee1066ff00066a172f82754ad9f08e8c29029e75f76712cb283d9fe5d1d9a
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:1d098eab26421f99ef71ffce82e4fa59ef0f9043de04ee74ca43486cbdb82006
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:b39819aa916974f4a408d23177e99cfb71e21af8ff9b0eacf6c37159c354ba70
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:0737e57ae8af6a62ac580710276d2c1620ac1b1283d3b84369db43e76e518610
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:5f9d5bdfd45f931989d069dcb1dea91a69020d83011fe0846ddfe22bba14b001
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel8@sha256:a630a20a633d057e475b4c39b342512119a71021e199aa0becf80899d8840ad7
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:c3c654be0b20cac501af1e31631464c32aecdbd16cbfb88f865900c70ccf77c2
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:7c5f339e762db1714b3f6cf1fc1bc8f08892cac433ab5a46e73b530fa08486fb
        - registry.redhat.io/multicluster-engine/cluster-api-rhel9@sha256:bd4eb05427ba2ea35f01227600cebd0fe8598d4de456348c24bb5e88d9ed8720
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:aff06db527e4ddc713a2fb8885824f0482647cbdcd844b311ccc3880a0dd2c3f
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:5ef8acc612817d478cede3b1bda55abafac00d0debebd68d948d157ee5b09286
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:294928c58b33ae8e776f86df211dac3b63a2cd9fc00aa995a0de037a29cb2c7d
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:2d807348774b8c6c0255c4bc931a956fc987c961fae6be689835410e8e301fd0
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel8@sha256:4d96379ad99d299d2a89eb1b0eb8a22db8bdada9a16248503cd8a9a5900eab35
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:d2c1d90a4e829b181fdf3d2838764c8f2b61c22b761279f53cb22316ad46ca03
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:93d65bfa0102c7ad7ec6ebfaad9ab062a31098b7ad7026313bcd8c6a495d5ca7
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:cd23f237c91dd8585946b31e17e38193f2e5ac8993d65459bb703a68f6fa8ca0
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:80a6615500949cc8c7a5b2a1e8c8c5f4564e9cdb9f995787a44f462e5bbd62cb
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel8@sha256:a703f2b1f4eb538aa02dbc0f4ff793506076480e94905c367b0d4ae352b87b65
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:5d488c8a644f2bfce97444d770516de10ceecc988cb3e9fbf0072d0b0a2eeb83
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:be1aa4cf5e09914774dcf69ab7ccdaf67504309d40c7e8cf4d91e9c1080e706f
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:82c0af2d285ba943dfa1da9f0856b26ec7e4c2c431577fcd4ab031ae040cfaea
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:87f5db53ad18f6d84501a5c2294cff73a5425276cdbde92225909826091b5ffa
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:46c4ac0cacef587b6e678b6cb0f68c29f5ac63bd59209d42a3ae0b7807e99972
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:d7c8bef4c3dec710ccb299482a10bfdf6efacd4e2ed2e6e323d448ed505000b3
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:d71ef80913d4e34389bce56c4a282cf7d2be705f003575dab87c07ca71f8c2d6
        version: 2.6.8
      entries:
      - name: multicluster-engine.v2.6.8
        version: 2.6.8
      - name: multicluster-engine.v2.6.7
        version: 2.6.7
      - name: multicluster-engine.v2.6.6
        version: 2.6.6
      - name: multicluster-engine.v2.6.5
        version: 2.6.5
      - name: multicluster-engine.v2.6.4
        version: 2.6.4
      - name: multicluster-engine.v2.6.3
        version: 2.6.3
      - name: multicluster-engine.v2.6.2
        version: 2.6.2
      - name: multicluster-engine.v2.6.1
        version: 2.6.1
      - name: multicluster-engine.v2.6.0
        version: 2.6.0
      name: stable-2.6
    - currentCSV: multicluster-engine.v2.7.5
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-21T22:32:21Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0 <2.7.5'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:44b07a18116ad827c23ca5a98b9fe8ef74a99abe4787cd01d848de42df44178d
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:15c2873703afcc74e41ae27dbb4761ad19c4bd2594ca6de82d3b6c02752e0ce3
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:816712bb3ccb58b705a85e5f4a55e519316fd8b9d7d9295d1042420ce44a4bad
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:e6a5e71f0aee8a682179cfd8f85b45a2ee04a1de680fc03dabdb659a2d2c0396
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:2a78ab192a554e78f3fc367c47e571f96b650ee9fb5ba0d06674f340f4320d94
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:9d1be5eb4fa357fe6bd97a52a53b3eeb537c431daf7af8a30efc3e0b6c5d5d66
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:2e3fe5e3b39721cc77b7d32b5682300eb60b1bc6940b8e83662b458d88489140
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:a24e9855a11d879d5308e29af6e3b2947365b7e466e7fa12dc57997140753604
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:b124242a55b1b2a2258cb12094617a0e009cf1cf58a60ab1b42d48d4ea2514ac
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:b66f2f4b5e0d77d501f9cc9da483f976f3c0eaf0ca72a5e9c9435c92d58e1388
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:00d415071f4c10575f9e5068868a7f80b0c0fb9c8614da86bc515923bed136a3
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:e2e16ff6213c0274126c9d64d495c0abbe17379e841ab730de66520908f1a190
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:4a0f1d3e696310f7a5bccbb358c538023a28f5bbc23b5743d10fd61b452d45d0
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:041044c08a1d526b349b5ad5fc4c4457f3632b1353530ed925c65ef2cce8e9c5
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:2025e424c11c6fe4a4395945c2e7d124d07cde43bb8d8cbe70286fe6ae345754
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:8c420720a48ea7c3aa096ae1929dcaa979c39e7865c5d43cba899e294745fc94
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:081151ca906b0f1afbcd3ee9c6d9e5bfe234494223c660ef376aa4004b0508f4
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:3defab1abc7395c9f9c22912ec6825be11b120a1416404eb0599e91f7f9c679a
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:c0fb7985620d5bd0dfadd57cb75664ee6777b8dac5a3463967e9b649e9eb836c
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:9e5c66ffdf39ae923ecccf79be538b883f2bdc8019169d55057565f077997d67
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:4fa34ab0aa8854eb5e942f2ed315041bf2026ce0bed86f57125878af0c73dc8b
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:7369bc79a283ff8b34c3189ea83e55f88a00d619cc78f8dad03d302dcb34a550
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:fedc699c0f786fb303d3956769f9c5fca1473ec4bf014ec87c5a4d40e34dcdb3
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:b82b1837dfc24a367f68741dece6d19c89c576706f006ee8d5123a0c6782c492
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:90c19d2aaa50e7cc9184f2f220b568c34c9fdcfdcba314da555885c999078057
        - registry.redhat.io/multicluster-engine/cluster-api-rhel9@sha256:cd8d6e0056fb0436275c12cc6b42f11c0ccd06729c7108e964b03d6c48d063bd
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:0684bcde2c018ae893bd71be8df3448afa09d199415e83285d2c740cff66192d
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:2216b015943eb8d9f25bf4ac37fbad2b7215a5c0003bd7c3b3d97b3b0a5ba708
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:555fcfce226e3b059f8f911d44a48578b898e085b6ad2575479feee71ec8baaf
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:a6736c7cac00f78c0e8877549ae2b74b7207cc348f24f175133db49fd38f240a
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:8e41d489e1def9b540ddd6387a1cc151b24e91a5fa44a74fbbaa0d1978588d57
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:45571dbb55bd213400eec6d2065b40615185bf3e52efdf495ab73201ec94faff
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:1c176688f13052783662115d5cabe8d614b8d9f79507dc2bdd4be1604e2eecc0
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:2cbd8c470dc4b263fd11952aa792094a7550f4b99e10f22186d51621bc8d3b3f
        version: 2.7.5
      entries:
      - name: multicluster-engine.v2.7.5
        version: 2.7.5
      - name: multicluster-engine.v2.7.4
        version: 2.7.4
      - name: multicluster-engine.v2.7.3
        version: 2.7.3
      - name: multicluster-engine.v2.7.2
        version: 2.7.2
      - name: multicluster-engine.v2.7.1
        version: 2.7.1
      - name: multicluster-engine.v2.7.0
        version: 2.7.0
      name: stable-2.7
    - currentCSV: multicluster-engine.v2.8.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-09-08T13:08:30Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0 <2.8.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:f869555beec0875c74f30a88428ac9641029dfed0365d45b23d5cae469cf7d0f
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:7ca7e87615e12b9d7fa75befd05f4f45bdc4bf8ef1b576c69f36a23c66c90cd6
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:988eb4aa01ba0d7e6728b23d043e4c0babe872ac23b6520f4daac714619a360b
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:f809a8774b0e3f891af46931dd4d2c11b95da4e7f5a8aee1814152d9a468d357
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:c5ac848e3b7eb88cdbdbf0b5f4aa96698d852b3a8118cf7146eb7e1948afaeba
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:a7f63077799a62fe6ece76a12346a87b5cbae966ab69df6ec6d951259dbf0e4a
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:c3b62ab1234257d30f41fa1ab33cb319321777a18de8ed751be1fcd48b204f2e
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:318ca0c7e9cd9a7336178519cadbf21eaeb474cd4a8eb5cf411a640d1c77372e
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:5ea86c20383a6932c593e020fd3b7b1a2a8e666bc374bff8b4bca0ee2d4ecc33
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:aa2754d35c8460a2879d4d523ebf32351d36e8fd7886335b5c0d2ca4befb8e8e
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:bee8f0c047cf8dbfa3f28712ce91bea55e926b2edb26c394f18a2d780e142083
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:7c86c89ff57091d4112783d82d1add47f2d56dba584418d5664c198ba048e0c2
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:f708f3e1007ed2b0e420e51d8b7ff9dd9bccaea9234755f4375f5514018211b4
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:3788b2674dda0b2cb4e9ca21bdce0b9e44702d72458b2396ef0acdad37fa866b
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:18b0e86305afe2eae3c26e637755fa9e8ab2abf30bc15e5428467a9a81ec99b0
        - registry.redhat.io/openshift4/ose-aws-cluster-api-controllers-rhel9@sha256:ada8b480a86d0a9e3ecc58f8b1cb53dbd7452b2b9183b3ba2b15a6a58c4436d1
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:f694b6e9e76941b11cc8a15153d9005cffe9bbbf58d0962f8ff546e90a63d93c
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:806b6c956f7bad824f0fe9acc3b534964e8f5f3399a1813f75511bc0f3904414
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:cf457b5a13cdd54c45a7b88f0f150a04e756bfac727c4365cb12f113d904937d
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:5461643a61850d95409d989e81c304053372f1dc5dbba291f3f3386de9ecab1c
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:261b99b93d98825ade0a0d3b9c4e465a785f35d2a310dedd4bbd6e4aa1f33ef9
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:0392c81a404be343d658dbc64153cbc46a64d0adf69aa8024c3f727351102d75
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:8379827635dfc4b51e04145d8048f7dcf9c6956ea1a22371bd22f948ae01b183
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:ee014c576263e19798df98460ebb5f1a21cb6249f09fe8a267b2dfa98b07d1eb
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:d4d11d0aeb3e1c9ff2a80e769d92dce509a0202a2ef1c9d2430030acdbf85bce
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:83a82f9558c406899e573308947c15b1f36de454006bb0e2ac44709f0c2c28d7
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:ea2ba066abd170e1087558c7b701faeb1be2e1667943dc14440f19c2ee5b0c30
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:fb2b8cef9916f387d26e65d9b53c7567ae1433d43d16e879ddb582ddc248eaa3
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:5ad96dad74c6f4967bd5fd397d6697e01b3b25bb3b8bae123a38e96964ddfa17
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:b9afa4d5328f17c7c3ce1dc105d177ecacf3447092d51a50f250a4b0ea692e2c
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:c6cd578123c1c86fcb3305ecca117543106f943cbcc4baf9065683988b2e10ff
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:e5e459e665afc14a1540e300015222b7c1ac896e2bee889283d67c3e19ca01ac
        - registry.redhat.io/openshift4/ose-cluster-api-rhel9@sha256:ec58463239eaa9bd15b31cbcfea0a34fd650201bac6120aa7ddaf72734451e06
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:ab0898d513b9b5ca2f5ca5879adc4e73da3232d515c0647a2a0ec28792264db1
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:0d1dfcb06eaa26de74994891371a5657aded7a59972f006f8c4a3859369877f3
        version: 2.8.3
      entries:
      - name: multicluster-engine.v2.8.3
        version: 2.8.3
      - name: multicluster-engine.v2.8.2
        version: 2.8.2
      - name: multicluster-engine.v2.8.1
        version: 2.8.1
      - name: multicluster-engine.v2.8.0
        version: 2.8.0
      name: stable-2.8
    - currentCSV: multicluster-engine.v2.9.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-30T19:22:43Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0 <2.9.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/open-cluster-management/backplane-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational components for managing multiple Kubernetes-based clusters. The deployment of the multicluster engine components is
              determined based on the configuration that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:8901e984b3b47f0650cddc1c36bdd89af22390f3d4a562d562fde7774d8ba6db
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:f3d65f68538e94260fad54f145433a36366f8c88f9c9f80b7fc5545ba62e1773
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:0a9c0ccb6a75273474fb04b785aa8fc58aec2f9ffbff8c232d359e14892bbae2
        - registry.redhat.io/multicluster-engine/capoa-control-plane-rhel9@sha256:b510a25a6370ca6b8d2f74165814ebfdae0ce9898cc343e73fab6344eebf2bed
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:8db1386d12d40dfbc80054c023897e5bfc0a987f6d07792b0f9f50f59f1a68d8
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:57699ecb529cb46843de847deef3e3757ca252910249d8deac16fcb00c01b636
        - registry.redhat.io/multicluster-engine/capoa-bootstrap-rhel9@sha256:a8b1278a2df9b17cee66813715995385bb3539af05e4df9b5bf7d9ccd508c313
        - registry.redhat.io/multicluster-engine/mce-capi-webhook-config-rhel9@sha256:7f1f70813e6d753b4acb04c7e9d7acb0047ad91f8a9a10e0fe2fff6c4411c8b4
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:a91276fcf5a589ef11f01b39df7d33ceb58ebcb47f95dfab772114bce7c1b4a8
        - registry.redhat.io/openshift4/ose-cluster-api-rhel9@sha256:9a72b8081a57846d942981ced07da86ad8e197309a1bc3c9d53876528b58f8ec
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:6a75df76fb0d9516b13fa6b079bc6c4d49dc6774f0d4f31343abfb8d70edcc23
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:ce0fd7bdc73cd2c84d1a9c5af528916488ca8116dd05d83dacf72f2af4248e65
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:ac43feaf079fa0d250d53cd9685fb68677d888dd091730faa32757dcdf2c95e4
        - registry.redhat.io/rhel8/postgresql-12@sha256:4773a29755ea826c0d3b58a8b8297c1d7388019516195cae9bd815f28ec4bca0
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:5e6594464bf7ddc2a66f127f645d93ad35d8cd95dd70544fa7fcec791f67f45c
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:36a6000b6482966f8bbf86e45c25335c066736ae1f8695a8b68e6b76b7c491ac
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:1b218f0aa0ea1f494094e3bc783ef40968cbff780e22de186d578a1d64c33bf8
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:3c450cb6a3ba6f2e3028856f14967803c1a691d95b8f54bd9d29d93691d6e795
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:8076db8ccd5578cb5aba2d66f53790563d3f42b93a6cfd14c7094d3a25707982
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:ad382d5eb347c448c46495a0a17503d789db4eb1096ad71a588b316b235e437f
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:6b95560a6a6dc1ddd41c76f275f83baec52ec9d79de04bfe54feaed02db123a3
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:ee75e805f0f9d73f073ce870f5d46962c0033d63158017841d17410febfc2c57
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:25ec9a0cc4ac92d4bed75c38fa37ef8d5ff9478cc475049b303d74eb58687a81
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:f8c3898e29f8c0a20c3be92bc4ee5a9443b9fc8218db95ba541fe3e57a89c40d
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:bf343592bf2f4399b5ac61709f56528435990e903812e264cbef845ff4433231
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:9555fd5f577f32a1d8ebe25f0d03e453c8ea925e4a12dbbe1015c58d3c8528b0
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:1cc6fddaf9e55637e90d5c3e598467df8967b8a362f8b8d46286b256bad6dd3f
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:a41cafc65e30deb073e051542702e258901b552ad00e42eacab05d149d6f0565
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:ea57247cf11f6ab966893ce4e6d2948232a4314588af2ca33c1d9689da7482dd
        - registry.redhat.io/openshift4/ose-baremetal-cluster-api-controllers-rhel9@sha256:f855d5ce17d3ac4560afc73c35d03f045f84551fdccc00312c890288a889b9ce
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:72a4ac73bb5e7797329e1856f785751d48815a64e82350a5d694bb5a8e121956
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:cd9dc89b933b394b08fe05ab5f0bbecea727bfb2c29770f11ec1db4aa5740b3c
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:37087ad8678f3e8ec8e2b2d035d3a12087ce9e0c993194a04a9874fd07e80d91
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:cecec1f7d7f0d37cf17fb25ccf34602728f2a09c61af7e212967b4ee28d2dec3
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:d068d73f41d8f8ecbec36752d0a432913bfbe2580f489abefe513c06233e80c9
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:edbb174cb69dd31339a53542d24c4fd00552e41fa09d7e2553baa7a1ad8dd810
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:1bbd134d44839258172e013fe84bfd174d36d6b7232e2cc8a3d8556998181d9d
        - registry.redhat.io/openshift4/ose-aws-cluster-api-controllers-rhel9@sha256:cf13b0658271cd0dbd1ce6dcc350e64c5459ef02ba4dc88a0031de9d6f26ffd1
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:8d189b565e4c3c434988aa4ff0ecad4c5f86b2a94120325c142ff111b25facde
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:902c5b3b4a4e8cb0a14fa5aba8451ec02830d5167ff1b91323ded62255983cd9
        version: 2.9.0
      entries:
      - name: multicluster-engine.v2.9.0
        version: 2.9.0
      name: stable-2.9
    defaultChannel: stable-2.9
    packageName: multicluster-engine
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JohnSnowLabs
      provider-url: ""
    name: nlp-server
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nlp-server.v0.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.johnsnowlabs.com/v1",
                "kind": "Nlpserver",
                "metadata": {
                  "name": "nlpserver-sample"
                },
                "spec": {
                  "ExtraEnv": {
                    "johnsnowlabs_client_id": "sI4MKSmLHOX2Pg7XhM3McJS2oyKG5PHcp0BlANEW",
                    "nb_uid": "1000",
                    "nb_user": "johnsnowlabs",
                    "pyspark_version": "3.0.2",
                    "spark_nlp_version": "3.4.4"
                  },
                  "affinity": {},
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "johnsnowlabs/nlp-server",
                    "tag": "0.8.0rc1"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {
                      "kubernetes.io/ingress.class": "nginx",
                      "nginx.ingress.kubernetes.io/ssl-redirect": "true"
                    },
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "nlp-server.jsl.aiplatform.eks.test.johnsnowlabs.com",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": [
                      {
                        "hosts": [
                          "nlp-server.jsl.aiplatform.eks.test.johnsnowlabs.com"
                        ],
                        "secretName": "annotationlab-tls-secret"
                      }
                    ]
                  },
                  "livenessProbe": "httpGet:\n  path: /\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 3\n",
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "readinessProbe": "httpGet:\n  path: /\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 3\n",
                  "replicaCount": 1,
                  "resources": {},
                  "securityContext": {},
                  "service": {
                    "port": 5000,
                    "portName": "http",
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": "nlp-server-sa"
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.johnsnowlabs.com/nlp-server/nlp-server-operator:0.8.6
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nlpserver
            name: nlpservers.apps.johnsnowlabs.com
            version: v1
        description: NLPServer from JohnSnow labs
        displayName: nlp-server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nlp-server
        links:
        - name: Nlp Server
          url: https://johnsnowlabs.com
        maturity: alpha
        provider:
          name: JohnSnowLabs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - registry.johnsnowlabs.com/nlp-server/nlp-server-operator:0.8.6
        version: 0.8.6
      entries:
      - name: nlp-server.v0.8.6
        version: 0.8.6
      name: alpha
    defaultChannel: alpha
    packageName: nlp-server
    provider:
      name: JohnSnowLabs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator-certified.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator-certified.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator-certified.v2.6.1
        version: 2.6.1
      - name: percona-postgresql-operator-certified.v2.6.0
        version: 2.6.0
      - name: percona-postgresql-operator-certified.v2.5.0
        version: 2.5.0
      name: stable
    - currentCSV: percona-postgresql-operator-certified.v2.7.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        version: 2.7.0-cw
      entries:
      - name: percona-postgresql-operator-certified.v2.7.0-cw
        version: 2.7.0-cw
      - name: percona-postgresql-operator-certified.v2.6.1-cw
        version: 2.6.1-cw
      - name: percona-postgresql-operator-certified.v2.6.0-cw
        version: 2.6.0-cw
      - name: percona-postgresql-operator-certified.v2.5.0-cw
        version: 2.5.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-postgresql-operator-certified
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Starburst Data, Inc.
      provider-url: https://www.starburstdata.com
    name: starburst-enterprise-helm-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: starburst-enterprise-helm-operator.v402.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstEnterprise",
                "metadata": {
                  "name": "starburstenterprise-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "cache": {
                    "diskUsagePercentage": 80,
                    "enabled": false,
                    "ttl": "7d",
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "catalogs": {
                    "tpch": "connector.name=tpch\ntpch.splits-per-node=4"
                  },
                  "commonLabels": {},
                  "coordinator": {
                    "additionalProperties": "",
                    "affinity": {},
                    "deploymentAnnotations": {},
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=true\nnode-scheduler.include-coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://localhost:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "resources": {
                      "cpu": 16,
                      "memory": "60Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "tolerations": []
                  },
                  "environment": null,
                  "expose": {
                    "clusterIp": {
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      }
                    },
                    "ingress": {
                      "annotations": {},
                      "host": null,
                      "ingressClassName": null,
                      "ingressName": "coordinator-ingress",
                      "path": "/",
                      "pathType": "ImplementationSpecific",
                      "serviceName": "starburst",
                      "servicePort": 8080,
                      "tls": {
                        "enabled": true,
                        "secretName": null
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "nodePort": 30080,
                          "port": 8080
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "externalSecrets": {
                    "enabled": false,
                    "goDaddy": {
                      "backendType": "secretsManager"
                    },
                    "secretPrefix": "external0",
                    "type": "goDaddy"
                  },
                  "extraArguments": [],
                  "extraSecret": {
                    "file": null,
                    "name": null
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise",
                    "tag": "402-e.1"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "initImage": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise-init",
                    "tag": "1.3.2"
                  },
                  "internal": {
                    "ports": {
                      "http": {
                        "port": 8080
                      },
                      "https": {
                        "port": 8443
                      }
                    }
                  },
                  "internalTls": false,
                  "licenseManager": {
                    "cronjob": {
                      "affinity": {},
                      "nodeSelector": {},
                      "priorityClassName": null,
                      "securityContext": {},
                      "tolerations": []
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "registry.connect.redhat.com/starburst/license-manager",
                      "tag": "0.0.2"
                    },
                    "url": "https://license-api.dogfood.eng.starburstdata.net/test/license/checkout"
                  },
                  "livenessProbe": null,
                  "prometheus": {
                    "agent": {
                      "config": "/etc/starburst/telemetry/prometheus.yaml",
                      "port": 8081,
                      "version": "0.16.1"
                    },
                    "enabled": true,
                    "rules": [
                      {
                        "attrNameSnakeCase": true,
                        "name": "$1",
                        "pattern": "trino.execution<name=QueryManager><>(running_queries|queued_queries)",
                        "type": "GAUGE"
                      },
                      {
                        "name": "failed_queries",
                        "pattern": "trino.execution<name=QueryManager><>FailedQueries\\.TotalCount",
                        "type": "COUNTER"
                      }
                    ]
                  },
                  "query": {
                    "maxConcurrentQueries": 3
                  },
                  "readinessProbe": null,
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "securityContext": {},
                  "sharedSecret": null,
                  "siac": {
                    "additionalHeapSizePercentage": 15,
                    "enabled": false
                  },
                  "spilling": {
                    "enabled": false,
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "startupProbe": null,
                  "usageMetrics": {
                    "enabled": false,
                    "usageClient": {
                      "initialDelay": "1m",
                      "interval": "1m"
                    }
                  },
                  "userDatabase": {
                    "enabled": false,
                    "users": [
                      {
                        "password": "46991b33f7a75ff79213c0dc0e610610",
                        "username": "admin"
                      }
                    ]
                  },
                  "worker": {
                    "additionalProperties": "",
                    "affinity": {},
                    "autoscaling": {
                      "enabled": false,
                      "maxReplicas": 100,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "deploymentAnnotations": {},
                    "deploymentTerminationGracePeriodSeconds": 300,
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://coordinator:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "kedaScaler": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.connect.redhat.com/starburst/keda-trino-scaler",
                        "tag": "0.1.7"
                      },
                      "port": 8021,
                      "scaledObjectSpec": {
                        "advanced": {},
                        "cooldownPeriod": 300,
                        "idleReplicaCount": 0,
                        "maxReplicaCount": 100,
                        "minReplicaCount": 1,
                        "pollingInterval": 30,
                        "scaleTargetRef": {
                          "name": "worker"
                        },
                        "triggers": [
                          {
                            "metadata": {
                              "numberOfQueriesPerWorker": "10",
                              "scaleInToIdleReplicaCountIfNoQueuedQueriesLeft": "true",
                              "scaleMethod": "query_queue"
                            },
                            "type": "external"
                          }
                        ]
                      }
                    },
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "replicas": 2,
                    "resources": {
                      "cpu": 16,
                      "memory": "100Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "starburstWorkerShutdownGracePeriodSeconds": 120,
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstHive",
                "metadata": {
                  "name": "starbursthive-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "affinity": {},
                  "commonLabels": {},
                  "database": {
                    "external": {
                      "driver": null,
                      "jdbcUrl": null,
                      "password": null,
                      "user": null
                    },
                    "internal": {
                      "databaseName": "hive",
                      "databasePassword": "HivePass1234",
                      "databaseUser": "hive",
                      "driver": "org.postgresql.Driver",
                      "env": [],
                      "envFrom": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.redhat.io/rhel8/postgresql-10",
                        "tag": "1"
                      },
                      "port": 5432,
                      "resources": {
                        "limits": {
                          "cpu": 2,
                          "memory": "1Gi"
                        },
                        "requests": {
                          "cpu": 2,
                          "memory": "1Gi"
                        }
                      },
                      "volume": {
                        "emptyDir": {}
                      }
                    },
                    "type": "internal"
                  },
                  "deploymentAnnotations": {},
                  "env": [],
                  "envFrom": [],
                  "expose": {
                    "clusterIp": {
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "nodePort": 30083,
                          "port": 9083
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "extraArguments": [],
                  "hdfs": {
                    "hadoopUserName": null
                  },
                  "heapSizePercentage": 85,
                  "hiveMetastoreStorageAuthorization": true,
                  "hiveMetastoreWarehouseDir": null,
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/hive",
                    "tag": "3.1.2-e.22"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "nodeSelector": {},
                  "objectStorage": {
                    "adl": {
                      "oauth2": {
                        "clientId": null,
                        "credential": null,
                        "refreshUrl": null
                      }
                    },
                    "awsS3": {
                      "accessKey": null,
                      "endpoint": null,
                      "pathStyleAccess": false,
                      "region": null,
                      "secretKey": null
                    },
                    "azure": {
                      "abfs": {
                        "accessKey": {
                          "accessKey": null,
                          "storageAccount": null
                        },
                        "authType": "accessKey",
                        "oauth": {
                          "clientId": null,
                          "endpoint": null,
                          "secret": null
                        }
                      },
                      "wasb": {
                        "accessKey": null,
                        "storageAccount": null
                      }
                    },
                    "gs": {
                      "cloudKeyFileSecret": null
                    }
                  },
                  "podAnnotations": {},
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "resources": {
                    "limits": {
                      "cpu": 1,
                      "memory": "1Gi"
                    },
                    "requests": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "securityContext": {},
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
          description: Deploys StarburstEnterprise Helm Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Starburst Data
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: StarburstEnterprise
            name: starburstenterprises.charts.starburstdata.com
            version: v1alpha1
          - kind: StarburstHive
            name: starbursthives.charts.starburstdata.com
            version: v1alpha1
        description: CRs for deploying starburst enterprise and hive
        displayName: Starburst Enterprise Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - starburst
        - trino
        - sep
        - bigdata
        - data
        links:
        - name: Starburst Enterprise Helm Operator
          url: https://www.starburstdata.com
        maintainers:
        - email: hello@starburstdata.com
          name: Starburst Data, Inc.
        maturity: alpha
        provider:
          name: Starburst Data, Inc.
          url: https://www.starburstdata.com
        relatedImages:
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
        version: 402.1.0
      entries:
      - name: starburst-enterprise-helm-operator.v402.1.0
        version: 402.1.0
      name: alpha
    defaultChannel: alpha
    packageName: starburst-enterprise-helm-operator-rhmp
    provider:
      name: Starburst Data, Inc.
      url: https://www.starburstdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: keystone-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keystone-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keystone.openstack.org/v1beta1",
                "kind": "KeystoneAPI",
                "metadata": {
                  "name": "keystone"
                },
                "spec": {
                  "adminProject": "admin",
                  "adminUser": "admin",
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "keystone",
                  "databaseInstance": "openstack",
                  "preserveJobs": false,
                  "region": "regionOne",
                  "resources": {
                    "requests": {
                      "cpu": "1.0",
                      "memory": "500Mi"
                    }
                  },
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "keystone.openstack.org/v1beta1",
                "kind": "KeystoneEndpoint",
                "metadata": {
                  "name": "placement"
                },
                "spec": {
                  "endpoints": {
                    "admin": "http://placement-admin-openstack.apps-crc.testing",
                    "internal": "http://placement-internal-openstack.apps-crc.testing",
                    "public": "http://placement-public-openstack.apps-crc.testing"
                  },
                  "serviceName": "placement"
                }
              },
              {
                "apiVersion": "keystone.openstack.org/v1beta1",
                "kind": "KeystoneService",
                "metadata": {
                  "name": "placement"
                },
                "spec": {
                  "enabled": true,
                  "passwordSelector": "PlacementPassword",
                  "secret": "osp-secret",
                  "serviceDescription": "Placement service",
                  "serviceName": "placement",
                  "serviceType": "placement",
                  "serviceUser": "placement"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:01Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KeystoneAPI is the Schema for the keystoneapis API
            displayName: Keystone API
            kind: KeystoneAPI
            name: keystoneapis.keystone.openstack.org
            version: v1beta1
          - description: KeystoneEndpoint is the Schema for the keystoneendpoints
              API
            displayName: Keystone Endpoint
            kind: KeystoneEndpoint
            name: keystoneendpoints.keystone.openstack.org
            version: v1beta1
          - description: KeystoneService is the Schema for the keystoneservices API
            displayName: Keystone Service
            kind: KeystoneService
            name: keystoneservices.keystone.openstack.org
            version: v1beta1
        description: Keystone Operator
        displayName: Keystone Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Identity
        - Keystone
        links:
        - name: Keystone Operator
          url: https://github.com/openstack-k8s-operators/keystone-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-keystone-rhel9@sha256:740e92a78d1e3874d2a40e37c970c6163d9e295dbc236914f1204baed070c595
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/keystone-rhel9-operator@sha256:8b055945a3808f7397265a10e6a6463af0ee3ef80fdff9d66402b488e5b70523
        version: 1.0.6
      entries:
      - name: keystone-operator.v1.0.6
        version: 1.0.6
      - name: keystone-operator.v1.0.5
        version: 1.0.5
      - name: keystone-operator.v1.0.4
        version: 1.0.4
      - name: keystone-operator.v1.0.3
        version: 1.0.3
      - name: keystone-operator.v1.0.2
        version: 1.0.2
      - name: keystone-operator.v1.0.1
        version: 1.0.1
      - name: keystone-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: keystone-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: ""
    name: ingress-node-firewall
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ingress-node-firewall.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ingressnodefirewall.openshift.io/v1alpha1",
                "kind": "IngressNodeFirewall",
                "metadata": {
                  "name": "ingressnodefirewall-demo2"
                },
                "spec": {
                  "ingress": [
                    {
                      "rules": [
                        {
                          "action": "Deny",
                          "order": 10,
                          "protocolConfig": {
                            "icmp": {
                              "icmpType": 8
                            },
                            "protocol": "ICMP"
                          }
                        },
                        {
                          "action": "Deny",
                          "order": 20,
                          "protocolConfig": {
                            "protocol": "TCP",
                            "tcp": {
                              "ports": "8000-9000"
                            }
                          }
                        }
                      ],
                      "sourceCIDRs": [
                        "172.16.0.0/12"
                      ]
                    },
                    {
                      "rules": [
                        {
                          "action": "Deny",
                          "order": 10,
                          "protocolConfig": {
                            "icmpv6": {
                              "icmpType": 128
                            },
                            "protocol": "ICMPv6"
                          }
                        }
                      ],
                      "sourceCIDRs": [
                        "fc00:f853:ccd:e793::0/64"
                      ]
                    }
                  ],
                  "interfaces": [
                    "eth0"
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "do-node-ingress-firewall": "true"
                    }
                  }
                }
              },
              {
                "apiVersion": "ingressnodefirewall.openshift.io/v1alpha1",
                "kind": "IngressNodeFirewallConfig",
                "metadata": {
                  "name": "ingressnodefirewallconfig",
                  "namespace": "ingress-node-firewall-system"
                },
                "spec": {
                  "nodeSelector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ingress-node-firewall-rhel9-operator@sha256:144403b009d1b5f2b2ee25ed3cebc525f3588f21bfa671b25c662cf800798839
          createdAt: "2024-03-13T12:36:01Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.11.0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-ingress-node-firewall
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/ingress-node-firewall
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: IngressNodeFirewallConfig
            name: ingressnodefirewallconfigs.ingressnodefirewall.openshift.io
            version: v1alpha1
          - kind: IngressNodeFirewallNodeState
            name: ingressnodefirewallnodestates.ingressnodefirewall.openshift.io
            version: v1alpha1
          - kind: IngressNodeFirewall
            name: ingressnodefirewalls.ingressnodefirewall.openshift.io
            version: v1alpha1
        description: The Ingress node firewall operator, manages the configuration
          of cluster scoped Ingress Node firewall rules for Kubernetes.
        displayName: Ingress Node Firewall Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ingressnodefirewall
        links:
        - name: Ingress Node Firewall
          url: https://github.com/openshift/ingress-node-firewall
        maintainers:
        - email: mmahmoud@redhat.com
          name: msherif1234
        maturity: alpha
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ingress-node-firewall-rhel9@sha256:4e3d69793f9280f6910a6749e80e8f6fe5d20c7a2c160d5682f310309c497e64
        - registry.redhat.io/openshift4/ingress-node-firewall-rhel9-operator@sha256:144403b009d1b5f2b2ee25ed3cebc525f3588f21bfa671b25c662cf800798839
        version: 4.16.0-202509240123
      entries:
      - name: ingress-node-firewall.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: ingress-node-firewall.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: ingress-node-firewall.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: ingress-node-firewall.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: ingress-node-firewall.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: ingress-node-firewall.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: ingress-node-firewall.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: ingress-node-firewall.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: ingress-node-firewall.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: ingress-node-firewall.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: ingress-node-firewall.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: ingress-node-firewall.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: ingress-node-firewall.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: ingress-node-firewall.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: ingress-node-firewall.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: ingress-node-firewall.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: ingress-node-firewall.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: ingress-node-firewall.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: ingress-node-firewall.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: ingress-node-firewall.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: ingress-node-firewall.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: ingress-node-firewall.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: ingress-node-firewall.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: ingress-node-firewall.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: ingress-node-firewall.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: ingress-node-firewall.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: ingress-node-firewall.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: ingress-node-firewall.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: ingress-node-firewall.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: ingress-node-firewall.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: ingress-node-firewall.v4.16.0-202405281506
        version: 4.16.0-202405281506
      name: stable
    defaultChannel: stable
    packageName: ingress-node-firewall
    provider:
      name: Red Hat, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cluster-observability-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-observability-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e
          createdAt: "2025-06-24T11:30:26Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=..0 <1.2.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-cluster-observability-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs",
              "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs",
              "prometheusagents.monitoring.rhobs",
              "perses.perses.dev"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
          support: Cluster Observability (https://issues.redhat.com/projects/COO/)
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: Perses is the Schema for the perses API
            displayName: Perses
            kind: Perses
            name: perses.perses.dev
            version: v1alpha1
          - description: A Perses Dashboard
            displayName: Perses Dashboard
            kind: PersesDashboard
            name: persesdashboards.perses.dev
            version: v1alpha1
          - description: A Perses Datasource
            displayName: Perses Datasource
            kind: PersesDatasource
            name: persesdatasources.perses.dev
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
          - description: UIPlugin defines a console plugin for observability
            displayName: UIPlugin
            kind: UIPlugin
            name: uiplugins.observability.openshift.io
            version: v1alpha1
        description: |-
          Cluster Observability Operator is a Go based Kubernetes operator to easily setup and manage various observability tools.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          - Setup console plugins
          - Setup korrel8r
          - Setup Perses
          - Setup Cluster Health Analyzer
          ### Documentation
          - **[Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/cluster_observability_operator/index)**
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)
        displayName: Cluster Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad
        - registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf
        - registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd
        - registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a
        - registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552
        - registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:e2681bce57dc9c15701f5591532c2dfe8f19778606661339553a28dc003dbca5
        - registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e
        - registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:e54c1e1301be66933f3ecb01d5a0ca27f58aabfd905b18b7d057bbf23bdb7b0d
        - registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:8597c48fc71fc6ec8e87dbe40dace4dbb7b817c1039db608af76a0d90f7ac2d0
        - registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404
        - registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223
        - registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00
        - registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1
        - registry.redhat.io/cluster-observability-operator/perses-0-1-rhel9-operator@sha256:bfed9f442aea6e8165644f1dc615beea06ec7fd84ea3f8ca393a63d3627c6a7c
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed
        version: 1.2.2
      entries:
      - name: cluster-observability-operator.v1.2.2
        version: 1.2.2
      - name: cluster-observability-operator.v1.2.1
        version: 1.2.1
      - name: cluster-observability-operator.v1.2.0
        version: 1.2.0
      - name: cluster-observability-operator.v1.1.1
        version: 1.1.1
      - name: cluster-observability-operator.v1.1.0
        version: 1.1.0
      - name: cluster-observability-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: cluster-observability-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kinesis-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kinesis-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kinesis.services.k8s.aws/v1alpha1",
                "kind": "Stream",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kinesis-controller:1.1.1
          createdAt: "2025-09-19T20:56:43Z"
          description: AWS Kinesis controller is a service controller for managing
            Kinesis resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Stream represents the state of an AWS kinesis Stream resource.
            displayName: Stream
            kind: Stream
            name: streams.kinesis.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Kinesis resources in AWS from within your Kubernetes cluster.

          **About Amazon Kinesis**

          You can use Amazon Kinesis Data Streams to collect and process large streams of data records in real time. You can create data-processing applications, known as Kinesis Data Streams applications. A typical Kinesis Data Streams application reads data from a data stream as data records. These applications can use the Kinesis Client Library, and they can run on Amazon EC2 instances.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Kinesis
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kinesis
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Kinesis Developer Resources
          url: https://aws.amazon.com/kinesis/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kinesis maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/kinesis-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-kinesis-controller.v1.1.1
        version: 1.1.1
      - name: ack-kinesis-controller.v1.1.0
        version: 1.1.0
      - name: ack-kinesis-controller.v1.0.16
        version: 1.0.16
      - name: ack-kinesis-controller.v1.0.15
        version: 1.0.15
      - name: ack-kinesis-controller.v1.0.14
        version: 1.0.14
      - name: ack-kinesis-controller.v1.0.13
        version: 1.0.13
      - name: ack-kinesis-controller.v1.0.12
        version: 1.0.12
      - name: ack-kinesis-controller.v1.0.11
        version: 1.0.11
      - name: ack-kinesis-controller.v1.0.10
        version: 1.0.10
      - name: ack-kinesis-controller.v1.0.9
        version: 1.0.9
      - name: ack-kinesis-controller.v1.0.8
        version: 1.0.8
      - name: ack-kinesis-controller.v1.0.7
        version: 1.0.7
      - name: ack-kinesis-controller.v1.0.6
        version: 1.0.6
      - name: ack-kinesis-controller.v1.0.5
        version: 1.0.5
      - name: ack-kinesis-controller.v1.0.4
        version: 1.0.4
      - name: ack-kinesis-controller.v1.0.3
        version: 1.0.3
      - name: ack-kinesis-controller.v1.0.2
        version: 1.0.2
      - name: ack-kinesis-controller.v1.0.1
        version: 1.0.1
      - name: ack-kinesis-controller.v1.0.0
        version: 1.0.0
      - name: ack-kinesis-controller.v0.0.19
        version: 0.0.19
      - name: ack-kinesis-controller.v0.0.18
        version: 0.0.18
      - name: ack-kinesis-controller.v0.0.17
        version: 0.0.17
      - name: ack-kinesis-controller.v0.0.16
        version: 0.0.16
      - name: ack-kinesis-controller.v0.0.15
        version: 0.0.15
      - name: ack-kinesis-controller.v0.0.10
        version: 0.0.10
      - name: ack-kinesis-controller.v0.0.9
        version: 0.0.9
      - name: ack-kinesis-controller.v0.0.8
        version: 0.0.8
      - name: ack-kinesis-controller.v0.0.6
        version: 0.0.6
      - name: ack-kinesis-controller.v0.0.5
        version: 0.0.5
      - name: ack-kinesis-controller.v0.0.4
        version: 0.0.4
      - name: ack-kinesis-controller.v0.0.3
        version: 0.0.3
      - name: ack-kinesis-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-kinesis-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Seldon Technologies
      provider-url: ""
    name: seldon-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: seldon-operator.v1.14.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1",
                "kind": "SeldonDeployment",
                "metadata": {
                  "labels": {
                    "app": "seldon",
                    "app.kubernetes.io/instance": "seldon1",
                    "app.kubernetes.io/name": "seldon",
                    "app.kubernetes.io/version": "v0.5"
                  },
                  "name": "seldon-model"
                },
                "spec": {
                  "name": "test-deployment",
                  "predictors": [
                    {
                      "componentSpecs": [
                        {
                          "spec": {
                            "containers": [
                              {
                                "image": "seldonio/mock_classifier:1.6.0",
                                "name": "classifier"
                              }
                            ]
                          }
                        }
                      ],
                      "graph": {
                        "children": [],
                        "name": "classifier",
                        "type": "MODEL"
                      },
                      "name": "example",
                      "replicas": 1
                    }
                  ]
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: docker.io/seldonio/seldon-core-operator:1.14.1
          createdAt: "2019-05-21 15:00:00"
          description: The Seldon operator for management, monitoring and operations
            of machine learning systems through the Seldon Engine. Once installed,
            the Seldon Operator provides multiple functions which facilitate the productisation,
            monitoring and maintenance of machine learning systems at scale.
          olm.skipRange: <1.13.1
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/SeldonIO/seldon-core
          support: Clive Cox
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A seldon engine deployment
            displayName: Seldon Deployment
            kind: SeldonDeployment
            name: seldondeployments.machinelearning.seldon.io
            version: v1
        description: "The Seldon operator enables for native operation of production
          machine learning workloads, including monitoring and operations of language-agnostic
          models with the benefits of real-time metrics and log analysis.\n   \n##
          Overview\nSeldon Core is an open source platform for deploying machine learning
          models on a Kubernetes cluster.\n\n* Deploy machine learning models in the
          cloud or on-premise.\n* Get metrics and ensure proper governance and compliance
          for your running machine learning models.\n* Create powerful inference graphs
          made up of multiple components.\n* Provide a consistent serving layer for
          models built using heterogeneous ML toolkits.\n\nYou can get started by
          following the guides in our documentation at https://docs.seldon.io/projects/seldon-core/en/latest/workflow/README.html\n"
        displayName: Seldon Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - aiops
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://docs.seldon.io/projects/seldon-core/en/latest/
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Seldon Technologies
        relatedImages:
        - docker.io/seldonio/seldon-core-operator:1.14.1
        - ""
        version: 1.14.1
      entries:
      - name: seldon-operator.v1.14.1
        version: 1.14.1
      name: stable
    defaultChannel: stable
    packageName: seldon-operator
    provider:
      name: Seldon Technologies
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: keycloak-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keycloak-operator.v19.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "False"
          containerImage: quay.io/keycloak/keycloak-operator:19.0.3-legacy
          createdAt: "2022-10-07 17:00:00"
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
          * Install Extensions
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/docs/latest/server_installation/index.html#_operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@lists.jboss.org
          name: Keycloak DEV mailing list
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/keycloak/keycloak-operator:19.0.3-legacy
        version: 19.0.3
      entries:
      - name: keycloak-operator.v19.0.3
        version: 19.0.3
      - name: keycloak-operator.v19.0.2
        version: 19.0.2
      - name: keycloak-operator.v19.0.1
        version: 19.0.1
      - name: keycloak-operator.v19.0.0
        version: 19.0.0
      - name: keycloak-operator.v18.0.2
        version: 18.0.2
      - name: keycloak-operator.v18.0.1
        version: 18.0.1
      - name: keycloak-operator.v18.0.0
        version: 18.0.0
      - name: keycloak-operator.v17.0.1
        version: 17.0.1
      - name: keycloak-operator.v17.0.0
        version: 17.0.0
      - name: keycloak-operator.v16.1.0
        version: 16.1.0
      - name: keycloak-operator.v16.0.0
        version: 16.0.0
      - name: keycloak-operator.v15.1.1
        version: 15.1.1
      - name: keycloak-operator.v15.1.0
        version: 15.1.0
      - name: keycloak-operator.v15.0.2
        version: 15.0.2
      - name: keycloak-operator.v15.0.1
        version: 15.0.1
      - name: keycloak-operator.v15.0.0
        version: 15.0.0
      - name: keycloak-operator.v14.0.0
        version: 14.0.0
      name: alpha
    - currentCSV: keycloak-operator.v20.0.0-alpha.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "False"
          containerImage: quay.io/keycloak/keycloak-operator:19.0.3
          createdAt: 10/07/22 17:00:00
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/docs/latest/server_installation/index.html#_operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@lists.jboss.org
          name: Keycloak DEV mailing list
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/keycloak/keycloak-operator:19.0.3
        version: 20.0.0-alpha.7
      entries:
      - name: keycloak-operator.v20.0.0-alpha.7
        version: 20.0.0-alpha.7
      - name: keycloak-operator.v20.0.0-alpha.6
        version: 20.0.0-alpha.6
      - name: keycloak-operator.v20.0.0-alpha.5
        version: 20.0.0-alpha.5
      - name: keycloak-operator.v20.0.0-alpha.4
        version: 20.0.0-alpha.4
      - name: keycloak-operator.v20.0.0-alpha.3
        version: 20.0.0-alpha.3
      - name: keycloak-operator.v20.0.0-alpha.2
        version: 20.0.0-alpha.2
      - name: keycloak-operator.v20.0.0-alpha.1
        version: 20.0.0-alpha.1
      name: candidate
    - currentCSV: keycloak-operator.v26.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/keycloak/keycloak-operator:26.4.0
          createdAt: "2025-09-30T11:58:06Z"
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/keycloak/keycloak-operator:26.4.0
        version: 26.4.0
      entries:
      - name: keycloak-operator.v26.4.0
        version: 26.4.0
      - name: keycloak-operator.v26.3.5
        version: 26.3.5
      - name: keycloak-operator.v26.3.4
        version: 26.3.4
      - name: keycloak-operator.v26.3.3
        version: 26.3.3
      - name: keycloak-operator.v26.3.2
        version: 26.3.2
      - name: keycloak-operator.v26.3.1
        version: 26.3.1
      - name: keycloak-operator.v26.3.0
        version: 26.3.0
      - name: keycloak-operator.v26.2.5
        version: 26.2.5
      - name: keycloak-operator.v26.2.4
        version: 26.2.4
      - name: keycloak-operator.v26.2.3
        version: 26.2.3
      - name: keycloak-operator.v26.2.2
        version: 26.2.2
      - name: keycloak-operator.v26.2.1
        version: 26.2.1
      - name: keycloak-operator.v26.2.0
        version: 26.2.0
      - name: keycloak-operator.v26.1.5
        version: 26.1.5
      - name: keycloak-operator.v26.1.4
        version: 26.1.4
      - name: keycloak-operator.v26.1.3
        version: 26.1.3
      - name: keycloak-operator.v26.1.2
        version: 26.1.2
      - name: keycloak-operator.v26.1.1
        version: 26.1.1
      - name: keycloak-operator.v26.1.0
        version: 26.1.0
      - name: keycloak-operator.v26.0.8
        version: 26.0.8
      - name: keycloak-operator.v26.0.7
        version: 26.0.7
      - name: keycloak-operator.v26.0.6
        version: 26.0.6
      - name: keycloak-operator.v26.0.5
        version: 26.0.5
      - name: keycloak-operator.v26.0.4
        version: 26.0.4
      - name: keycloak-operator.v26.0.2
        version: 26.0.2
      - name: keycloak-operator.v26.0.1
        version: 26.0.1
      - name: keycloak-operator.v26.0.0
        version: 26.0.0
      - name: keycloak-operator.v25.0.6
        version: 25.0.6
      - name: keycloak-operator.v25.0.5
        version: 25.0.5
      - name: keycloak-operator.v25.0.4
        version: 25.0.4
      - name: keycloak-operator.v25.0.2
        version: 25.0.2
      - name: keycloak-operator.v25.0.1
        version: 25.0.1
      - name: keycloak-operator.v25.0.0
        version: 25.0.0
      - name: keycloak-operator.v24.0.5
        version: 24.0.5
      - name: keycloak-operator.v24.0.4
        version: 24.0.4
      - name: keycloak-operator.v24.0.3
        version: 24.0.3
      - name: keycloak-operator.v24.0.2
        version: 24.0.2
      - name: keycloak-operator.v24.0.1
        version: 24.0.1
      - name: keycloak-operator.v24.0.0
        version: 24.0.0
      - name: keycloak-operator.v23.0.7
        version: 23.0.7
      - name: keycloak-operator.v23.0.6
        version: 23.0.6
      - name: keycloak-operator.v23.0.5
        version: 23.0.5
      - name: keycloak-operator.v23.0.4
        version: 23.0.4
      - name: keycloak-operator.v23.0.3
        version: 23.0.3
      - name: keycloak-operator.v23.0.2
        version: 23.0.2
      - name: keycloak-operator.v23.0.1
        version: 23.0.1
      - name: keycloak-operator.v23.0.0
        version: 23.0.0
      - name: keycloak-operator.v22.0.5
        version: 22.0.5
      - name: keycloak-operator.v22.0.4
        version: 22.0.4
      - name: keycloak-operator.v22.0.3
        version: 22.0.3
      - name: keycloak-operator.v22.0.2
        version: 22.0.2
      - name: keycloak-operator.v22.0.1
        version: 22.0.1
      - name: keycloak-operator.v22.0.0
        version: 22.0.0
      - name: keycloak-operator.v21.1.2
        version: 21.1.2
      - name: keycloak-operator.v21.1.1
        version: 21.1.1
      - name: keycloak-operator.v21.1.0
        version: 21.1.0
      - name: keycloak-operator.v21.0.2
        version: 21.0.2
      - name: keycloak-operator.v21.0.1
        version: 21.0.1
      - name: keycloak-operator.v21.0.0
        version: 21.0.0
      - name: keycloak-operator.v20.0.3
        version: 20.0.3
      - name: keycloak-operator.v20.0.2
        version: 20.0.2
      - name: keycloak-operator.v20.0.1
        version: 20.0.1
      - name: keycloak-operator.v20.0.0
        version: 20.0.0
      name: fast
    defaultChannel: fast
    packageName: keycloak-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: redis-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"Redis","metadata":{"name":"redis"},"spec":{"mode":"cluster","size":3,"global":{"image":"quay.io/opstree/redis:v6.2","imagePullPolicy":"IfNotPresent","password":"Opstree@1234","resources":{"requests":{"cpu":"100m","memory":"128Mi"},"limits":{"cpu":"100m","memory":"128Mi"}}},"master":{"service":{"type":"ClusterIP"},"redisConfig":{}},"slave":{"service":{"type":"ClusterIP"},"redisConfig":{}},"service":{"type":"ClusterIP"},"redisConfig":{},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:1.0","imagePullPolicy":"Always","resources":{"requests":{"cpu":"100m","memory":"128Mi"},"limits":{"cpu":"100m","memory":"128Mi"}}},"storage":{"volumeClaimTemplate":{"spec":{"storageClassName":"buildpiper-storage","accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}}}}]'
          capabilities: Seamless Upgrades
          categories: Database, Storage
          certified: "false"
          containerImage: quay.io/opstree/redis-operator:v0.4.0
          createdAt: 06-02-2021
          description: A Golang based redis operator that will make/oversee Redis
            standalone/cluster mode setup on top of the Kubernetes.
          repository: https://github.com/ot-container-kit/redis-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Redis
            displayName: Redis
            kind: Redis
            name: redis.redis.redis.opstreelabs.in
            version: v1beta1
        description: |+
          A Golang based redis operator that will make/oversee Redis standalone/cluster mode setup on top of the Kubernetes. It can create a redis cluster setup with best practices on Cloud as well as the Bare metal environment. Also, it provides an in-built monitoring capability using redis-exporter.

          ### Supported Features

          Here the features which are supported by this operator:-
          - Redis cluster/standalone mode setup
          - Inbuilt monitoring with prometheus exporter
          - Dynamic storage provisioning with pvc template
          - Resources restrictions with k8s requests and limits
          - Password/Password-less setup
          - Node selector and affinity
          - Priority class to manage setup priority
          - SecurityContext to manipulate kernel parameters
          ### Documentation

          [Documentation](https://docs.opstreelabs.in/redis-operator/)

          ### Getting Help

          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.

          ### Contributing

          You can contribute to this project by:-
          - Raising Bugs and Feature Requests
          - Fixing issue by opening a Pull Request
          - Improving Documentation

          ### License

          Logging Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/redis-operator/blob/master/LICENSE)


        displayName: Redis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - Redis
        - Opstree
        - CRD
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://docs.opstreelabs.in/redis-operator/
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: beta
        minKubeVersion: 1.11.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/opstree/redis-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: redis-operator.v0.4.0
        version: 0.4.0
      - name: redis-operator.v0.3.0
        version: 0.3.0
      name: preview
    - currentCSV: redis-operator.v0.15.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisCluster","metadata":{"name":"redis-cluster"},"spec":{"clusterSize":3,"clusterVersion":"v7","securityContext":{"runAsUser":1000,"fsGroup":1000},"persistenceEnabled":true,"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"Redis","metadata":{"name":"redis-standalone"},"spec":{"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"IfNotPresent"},"securityContext":{"runAsUser":1000,"fsGroup":1000}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisSentinel","metadata":{"name":"redis-sentinel"},"spec":{"clusterSize":3,"podSecurityContext":{"runAsUser":1000,"fsGroup":1000},"redisSentinelConfig":{"redisReplicationName":"redis-replication"},"kubernetesConfig":{"image":"quay.io/opstree/redis-sentinel:v7.0.7","imagePullPolicy":"IfNotPresent","resources":{"requests":{"cpu":"101m","memory":"128Mi"},"limits":{"cpu":"101m","memory":"128Mi"}}}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisReplication","metadata":{"name":"redis-replication"},"spec":{"clusterSize":3,"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"Always"},"podSecurityContext":{"runAsUser":1000,"fsGroup":1000}}}]'
          capabilities: Seamless Upgrades
          categories: Database, Storage
          certified: "false"
          containerImage: quay.io/opstree/redis-operator:v0.15.1
          createdAt: 24-09-2023
          description: A Golang based redis operator that will make/oversee Redis
            standalone/cluster mode setup on top of the Kubernetes.
          repository: https://github.com/ot-container-kit/redis-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Redis Cluster
            displayName: RedisCluster
            kind: RedisCluster
            name: redisclusters.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis Replication
            displayName: RedisReplication
            kind: RedisReplication
            name: redisreplications.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis Sentinel
            displayName: RedisSentinel
            kind: RedisSentinel
            name: redissentinels.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis
            displayName: Redis
            kind: Redis
            name: redis.redis.redis.opstreelabs.in
            version: v1beta1
        description: "A Golang based redis operator that will make/oversee Redis\nstandalone/cluster
          mode setup on top of the Kubernetes. It can create a\nredis cluster setup
          with best practices on Cloud as well as the Bare metal\nenvironment. Also,
          it provides an in-built monitoring capability using\nredis-exporter.\n\nOrganizations
          that are using Redis Operator to manage their redis workload \ncan be found
          here. If your organization is also using Redis Operator, please \nfree to
          add by creating a PR https://github.com/OT-CONTAINER-KIT/redis-operator/pulls.\n\nThis
          operator only supports versions of redis =>6.\n\n### Supported Features\n\nHere
          the features which are supported by this operator:-\n- Redis cluster and
          standalone mode setup\n- Redis cluster failover and recovery\n- Redis leader
          and follower setup with Sentinel for High Availability\n- Inbuilt monitoring
          with redis exporter\n- Password and password-less setup of redis\n- TLS
          support for additional security layer\n- Ipv4 and Ipv6 support for redis
          setup\n- External configuration support for leader and follower\n- Detailed
          monitoring grafana dashboard\n\n### Documentation\n\n[Documentation](https://ot-redis-operator.netlify.app/)\n\n###
          Getting Help\nThis project is managed by OpsTree Solutions. If you have
          any queries or\nsuggestions, mail us at opensource@opstree.com.\n\n### Contributing\n\nYou
          can contribute to this project by:-\n- Raising Bugs and Feature Requests\n-
          Fixing issue by opening a Pull Request\n- Improving Documentation\n\n###
          License\n\nLogging Operator is licensed under [Apache License, Version\n2.0](https://github.com/OT-CONTAINER-KIT/redis-operator/blob/master/LICENSE)\n"
        displayName: Redis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - Redis
        - Opstree
        - CRD
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-redis-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/opstree/redis-operator:v0.15.1
        version: 0.15.1
      entries:
      - name: redis-operator.v0.15.1
        version: 0.15.1
      - name: redis-operator.v0.15.0
        version: 0.15.0
      - name: redis-operator.v0.14.0
        version: 0.14.0
      - name: redis-operator.v0.13.0
        version: 0.13.0
      - name: redis-operator.v0.8.0
        version: 0.8.0
      - name: redis-operator.v0.6.0
        version: 0.6.0
      - name: redis-operator.v0.5.0
        version: 0.5.0
      - name: redis-operator.v0.4.0
        version: 0.4.0
      - name: redis-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: redis-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Veeam Kasten
      provider-url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
    name: k10-kasten-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k10-kasten-operator-rhmp.v8.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apik10.kasten.io/v1alpha1",
                "kind": "K10",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/reconcile-period": "2m",
                    "helm.sdk.operatorframework.io/rollback-force": "false"
                  },
                  "name": "k10"
                },
                "spec": {
                  "auth": {
                    "basicAuth": {
                      "enabled": false,
                      "htpasswd": "",
                      "secretName": ""
                    },
                    "tokenAuth": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "persistence": {
                      "catalog": {
                        "size": ""
                      },
                      "storageClass": ""
                    }
                  },
                  "metering": {
                    "mode": ""
                  },
                  "route": {
                    "enabled": false,
                    "host": "",
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "apik10.kasten.io/v1alpha1",
                "kind": "K10restore",
                "metadata": {
                  "name": "k10restore"
                },
                "spec": {
                  "pointInTime": null,
                  "profileName": "",
                  "skipResource": null,
                  "sourceClusterID": ""
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          console.openshift.io/plugins: '["veeam-kasten-console-plugin"]'
          containerImage: registry.connect.redhat.com/kasten/k10-operator@sha256:4e8ccf38e47f92d2a308587591f85fb4f15aea1e7fd5c248803764c94eaecd4a
          createdAt: Wed Sep 24 02:11:59 PM UTC 2025
          description: Kasten data management platform for OpenShift, The Market Leader
            for Kubernetes Backup and DR
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/k10-kasten-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/k10-kasten-operator-rhmp/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: kasten-io
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: contact@kasten.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              k10restore can recover the K10 platform in case of a variety of disasters such as
              the accidental deletion of K10, failure of underlying storage that K10 uses for its catalog,
              or even the accidental destruction of the Kubernetes cluster on which K10 is deployed.

              **Before creating k10restore**: [K10 Disaster Recovery documentation](https://docs.kasten.io/latest/operating/dr.html#recovering-with-the-operator)
            kind: K10restore
            name: k10restores.apik10.kasten.io
            version: v1alpha1
          - description: |-
              [Kasten's k10](https://docs.kasten.io/) is a data lifecycle management system for all your persistence enabled container-based applications.

              **Note: The complete list of values supported by the K10 spec field in the "YAML View" are documented [here](https://docs.kasten.io/latest/install/advanced.html#complete-list-of-k10-helm-options).**
            kind: K10
            name: k10s.apik10.kasten.io
            version: v1alpha1
        description: |-
          Kasten data management platform for OpenShift, The Market Leader for Kubernetes Backup and DR.

          The supported Kubernetes versions are listed here - https://docs.kasten.io/latest/operating/support.html
          Please review release notes before an upgrade - https://docs.kasten.io/latest/releasenotes.html
        displayName: Veeam Kasten (Free)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kubernetes
        - storage
        - data-management
        links:
        - name: Kasten Documentation
          url: https://docs.kasten.io
        - name: Veeam Kasten
          url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
        maintainers:
        - email: contact@kasten.io
          name: Veeam Kasten
        maturity: stable
        provider:
          name: Veeam Kasten
          url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:503d5ee0fbcd5bd3492785f4246e19cbdb92960be0d05f3c6f123a32fe0fe84d
        - registry.connect.redhat.com/kasten/bloblifecyclemanager@sha256:7b209314b96bc8ebaa77e97426bb34176d05844cc3e8d4b462cfba5c4a8d3798
        - registry.connect.redhat.com/kasten/kanister-tools@sha256:7af2e37c9027f46c565901d35a3a7da4d9c7ddd2812628b67275d4acf006b80c
        - registry.connect.redhat.com/kasten/kanister@sha256:1d69ff91cec8e7c131dfbe1ad3e282aab854d3412f5f6e3b808d25eb7b8c38bc
        - registry.connect.redhat.com/kasten/metric-sidecar@sha256:20c142258de56cf8c60d4b0afa5e72b9317023ebf4ab109fb84e0d57fae854a4
        - registry.connect.redhat.com/kasten/controllermanager@sha256:ab3b815bb12cd3c263cbfdba3939a95097450f1864431cca7380827986d704ed
        - registry.connect.redhat.com/kasten/configmap-reload@sha256:01010f2bb580eb131d63d3271efe40eadd8269b84cdb02bdea36750d1359a43f
        - registry.connect.redhat.com/kasten/aggregatedapis@sha256:db868afd5c6d4a8d2e7c8376ed6943963c4be299fca6b26931f907430763d244
        - registry.connect.redhat.com/kasten/jobs@sha256:c6358878e981baee4ac72abcc1269c206ee014af2c16584d1b41fe18e0f0e57f
        - registry.connect.redhat.com/kasten/upgrade@sha256:bf2200697d38f3ee95ba2863dfbae4d982273d2fd8450c9ffbc3a85be7d6d87d
        - registry.connect.redhat.com/kasten/prometheus@sha256:29d913bb06c3245cca5798772dba7d7d5dc9b76ed2e0700f5eb4d20416e24fa2
        - registry.connect.redhat.com/kasten/k10tools@sha256:59caa1920c76a1bc86e7733731602e3f0fc08026d9a16000bb748503a4d9cd9b
        - registry.connect.redhat.com/kasten/garbagecollector@sha256:6f2972bb255f5a2de33ace5ce805b15d5e04d3c259c9c8d018c1257c74562878
        - registry.connect.redhat.com/kasten/events@sha256:7cae1c0909ea7a468e96484d99300677bdad49f34cc57c412ebaf342cb176264
        - registry.connect.redhat.com/kasten/logging@sha256:014bfdfc3620cc2fe6540b9993b9b9e7d70c9b969714477db4de91a90bdb4a38
        - registry.connect.redhat.com/kasten/k10-operator@sha256:4e8ccf38e47f92d2a308587591f85fb4f15aea1e7fd5c248803764c94eaecd4a
        - registry.connect.redhat.com/kasten/auth@sha256:f8543e67de3d5276c597c26f16053bf9bbcbd9cbdad4652214935a1e40a26817
        - registry.connect.redhat.com/kasten/frontend@sha256:a021964ed85133897e4956014b5c5013da3332a36a7e96a0de2e2e1b67888491
        - registry.connect.redhat.com/kasten/init@sha256:cc7fbbda1b6946e93dd083fb7092f38789f626c26ac68c226b61fe57064a8f11
        - registry.connect.redhat.com/kasten/dex@sha256:4c8522a27c0d4f0977baa074800bba96a9bd9aefc57473de283b1d95586f7517
        - registry.connect.redhat.com/kasten/catalog@sha256:11a77ac968634850c72cd10f2bde55afc632c68cdc0881aeae83c100288c6aec
        - registry.connect.redhat.com/kasten/dashboardbff@sha256:e1208c60146faa8c3c7ef73b2290663840276ed8d3b3c990f495cee2967104ee
        - registry.connect.redhat.com/kasten/state@sha256:191383c76c4ff22805c06217e9f32a2f41a1206afab5f16acec4a5dd56e4a6b4
        - registry.connect.redhat.com/kasten/repositories@sha256:924a607874e1807a877cb454e00539d64f4487e7435d4bfb524f3e569ad1f942
        - registry.connect.redhat.com/kasten/metering@sha256:f809696ac1309fc54fde87e5d2030267d266386fb73a56287868bbac433756d8
        - registry.connect.redhat.com/kasten/restorectl@sha256:9dc3c834bd3d4543115ae1ad586ee1248fc2b637dbc2d1e2c66ecc2a459db67e
        - registry.connect.redhat.com/kasten/datamover@sha256:c9e2dc5b646aa0c465a7b2daf011462273a6f5b53fdff772d879293208e2d57c
        - registry.connect.redhat.com/kasten/vbrintegrationapi@sha256:1a20b0da4efb7e802516224675b37751bd56d406e57b16216388d2b5167c648d
        - registry.connect.redhat.com/kasten/executor@sha256:32c4ac1aa2ca802ff77e913991949e6841149963863ad527a67f6bab6c94a5af
        - registry.connect.redhat.com/kasten/gateway@sha256:dc2d7bdeb8abaf9454470afd53af2cd5430505351a97b6501ff65a60dd2a6db1
        - registry.connect.redhat.com/kasten/ocpconsoleplugin@sha256:69f6bfb9421f4264dd7d066d809311022c0db57c41319a6242532f6acfbd2145
        - registry.connect.redhat.com/kasten/crypto@sha256:86bfbbf4c4f06bab1e4b091f9f312d536813e2c11aa899690b3277ff73110a09
        version: 8.0.9
      entries:
      - name: k10-kasten-operator-rhmp.v8.0.9
        version: 8.0.9
      - name: k10-kasten-operator-rhmp.v8.0.8
        version: 8.0.8
      - name: k10-kasten-operator-rhmp.v8.0.7
        version: 8.0.7
      - name: k10-kasten-operator-rhmp.v8.0.6
        version: 8.0.6
      - name: k10-kasten-operator-rhmp.v8.0.5
        version: 8.0.5
      - name: k10-kasten-operator-rhmp.v8.0.4
        version: 8.0.4
      - name: k10-kasten-operator-rhmp.v8.0.3
        version: 8.0.3
      - name: k10-kasten-operator-rhmp.v8.0.2
        version: 8.0.2
      - name: k10-kasten-operator-rhmp.v8.0.1
        version: 8.0.1
      - name: k10-kasten-operator-rhmp.v8.0.0
        version: 8.0.0
      - name: k10-kasten-operator-rhmp.v7.5.10
        version: 7.5.10
      - name: k10-kasten-operator-rhmp.v7.5.9
        version: 7.5.9
      - name: k10-kasten-operator-rhmp.v7.5.8
        version: 7.5.8
      - name: k10-kasten-operator-rhmp.v7.5.7
        version: 7.5.7
      - name: k10-kasten-operator-rhmp.v7.5.4
        version: 7.5.4
      - name: k10-kasten-operator-rhmp.v7.5.3
        version: 7.5.3
      - name: k10-kasten-operator-rhmp.v7.5.2
        version: 7.5.2
      - name: k10-kasten-operator-rhmp.v7.5.1
        version: 7.5.1
      - name: k10-kasten-operator-rhmp.v7.5.0
        version: 7.5.0
      - name: k10-kasten-operator-rhmp.v7.0.14
        version: 7.0.14
      - name: k10-kasten-operator-rhmp.v7.0.13
        version: 7.0.13
      - name: k10-kasten-operator-rhmp.v7.0.12
        version: 7.0.12
      - name: k10-kasten-operator-rhmp.v7.0.11
        version: 7.0.11
      - name: k10-kasten-operator-rhmp.v7.0.9
        version: 7.0.9
      - name: k10-kasten-operator-rhmp.v7.0.8
        version: 7.0.8
      - name: k10-kasten-operator-rhmp.v7.0.7
        version: 7.0.7
      - name: k10-kasten-operator-rhmp.v7.0.6
        version: 7.0.6
      - name: k10-kasten-operator-rhmp.v7.0.5
        version: 7.0.5
      - name: k10-kasten-operator-rhmp.v7.0.4
        version: 7.0.4
      - name: k10-kasten-operator-rhmp.v7.0.3
        version: 7.0.3
      - name: k10-kasten-operator-rhmp.v7.0.1
        version: 7.0.1
      - name: k10-kasten-operator-rhmp.v6.5.14
        version: 6.5.14
      - name: k10-kasten-operator-rhmp.v6.5.13
        version: 6.5.13
      - name: k10-kasten-operator-rhmp.v6.5.12
        version: 6.5.12
      - name: k10-kasten-operator-rhmp.v6.5.11
        version: 6.5.11
      - name: k10-kasten-operator-rhmp.v6.5.10
        version: 6.5.10
      - name: k10-kasten-operator-rhmp.v6.5.8
        version: 6.5.8
      - name: k10-kasten-operator-rhmp.v6.5.6
        version: 6.5.6
      - name: k10-kasten-operator-rhmp.v6.5.5
        version: 6.5.5
      - name: k10-kasten-operator-rhmp.v6.5.4
        version: 6.5.4
      - name: k10-kasten-operator-rhmp.v6.5.3
        version: 6.5.3
      - name: k10-kasten-operator-rhmp.v6.5.1
        version: 6.5.1
      - name: k10-kasten-operator-rhmp.v6.5.0
        version: 6.5.0
      - name: k10-kasten-operator-rhmp.v6.0.12
        version: 6.0.12
      - name: k10-kasten-operator-rhmp.v6.0.11
        version: 6.0.11
      - name: k10-kasten-operator-rhmp.v6.0.9
        version: 6.0.9
      - name: k10-kasten-operator-rhmp.v6.0.8
        version: 6.0.8
      - name: k10-kasten-operator-rhmp.v6.0.7
        version: 6.0.7
      - name: k10-kasten-operator-rhmp.v6.0.6
        version: 6.0.6
      - name: k10-kasten-operator-rhmp.v6.0.5
        version: 6.0.5
      - name: k10-kasten-operator-rhmp.v6.0.4
        version: 6.0.4
      - name: k10-kasten-operator-rhmp.v6.0.3
        version: 6.0.3
      - name: k10-kasten-operator-rhmp.v6.0.2
        version: 6.0.2
      - name: k10-kasten-operator-rhmp.v6.0.1
        version: 6.0.1
      - name: k10-kasten-operator-rhmp.v6.0.0
        version: 6.0.0
      name: stable
    defaultChannel: stable
    packageName: k10-kasten-operator-rhmp
    provider:
      name: Veeam Kasten
      url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SusQL Operator Contributors
      provider-url: https://github.com/sustainable-computing-io/susql-operator
    name: susql-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: susql-operator.v0.0.35
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "susql.ibm.com/v1",
                "kind": "LabelGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "susql-operator",
                    "app.kubernetes.io/instance": "labelgroup-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "susql-operator",
                    "app.kubernetes.io/part-of": "susql-operator"
                  },
                  "name": "labelgroup-sample"
                },
                "spec": {
                  "labels": [
                    "labelgroup-sample-1",
                    "labelgroup-sample-2"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: quay.io/sustainable_computing_io/susql_operator:0.0.34
          createdAt: "2025-07-10T01:20:37Z"
          description: 'Aggregates energy and CO2 emission data for pods tagged with
            SusQL labels '
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/sustainable-computing-io/susql-operator
          support: https://github.com/sustainable-computing-io/susql-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LabelGroup is the Schema for the LabelGroups API
            displayName: Label Group
            kind: LabelGroup
            name: labelgroups.susql.ibm.com
            version: v1
        description: |-
          ### About this Operator

          SusQL is a Kubernetes operator that aggregates energy and estimated
          carbon dioxide emission data for pods tagged with SusQL specific
          labels. The energy measurements are taken from Kepler which should
          be deployed in the cluster before using SusQL.

          ### Prerequisites

          1. Deployment of [Kepler](https://sustainable-computing.io/) on the cluster
          2. Ensure that [User Project Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          is enabled to monitor energy consumed in user projects.
          3. Creation of a SusQL Service Monitor:
                `oc apply -n <NAMESPACE> -f https://raw.githubusercontent.com/sustainable-computing-io/susql-operator/main/hack/susql-servicemonitor.yaml`

          ### API Backward Compatibility

          Until Operator supports **seamless upgrades**, a new version of the operator may
          introduce a change that is **NOT** backwards compatible with the previous version.
          Thus, to upgrade the operator, uninstall the already installed version completely
          (including CRDs) and install the new version. For example to upgrade from 0.1.0
          to 0.2.0, you must first uninstall 0.1.0.

          ### LabelGroup spec Example

          ```
          apiVersion: susql.ibm.com/v1
          kind: LabelGroup
          metadata:
            name: labelgroup-name
            namespace: susql-operator
          spec:
            labels:
              - my-label-1
              - my-label-2
          ```

          ### Documentation

          Documentation can be found at [SusQL Operator](https://github.com/sustainable-computing-io/susql-operator#readme).

          ### Getting help

          If you encounter any issues while using the SusQL Operator, you can create an
          [issue](https://github.com/sustainable-computing-io/susql-operator/issues)
          on our GitHub repo for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by raising [issues](https://github.com/sustainable-computing-io/susql-operator/issues)
          related to susql-operator.

          ### License
          susql-operator is licensed under the Apache 2.0 license
        displayName: SusQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - aggregation
        - sustainability
        - energy
        - kepler
        - susql
        - carbon dioxide (CO2) emissions
        - green computing
        links:
        - name: SusQL Operator
          url: https://github.com/sustainable-computing-io/susql-operator
        - name: Sustainable Computing
          url: https://github.com/sustainable-computing-io
        maintainers:
        - email: trent@jp.ibm.com
          name: Scott Trent
        - email: mohammed.munir.abdi@ibm.com
          name: Mohammed Abdi
        maturity: alpha
        provider:
          name: SusQL Operator Contributors
          url: https://github.com/sustainable-computing-io/susql-operator
        relatedImages:
        - quay.io/sustainable_computing_io/susql_operator:0.0.34
        version: 0.0.35
      entries:
      - name: susql-operator.v0.0.35
        version: 0.0.35
      - name: susql-operator.v0.0.34
        version: 0.0.34
      - name: susql-operator.v0.0.33
        version: 0.0.33
      - name: susql-operator.v0.0.32
        version: 0.0.32
      - name: susql-operator.v0.0.30
        version: 0.0.30
      - name: susql-operator.v0.0.29
        version: 0.0.29
      - name: susql-operator.v0.0.28
        version: 0.0.28
      - name: susql-operator.v0.0.27
        version: 0.0.27
      name: alpha
    defaultChannel: alpha
    packageName: susql-operator
    provider:
      name: SusQL Operator Contributors
      url: https://github.com/sustainable-computing-io/susql-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cockroach Labs
      provider-url: ""
    name: cockroachdb-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroach-operator.v2.18.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v25.2.2",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
          createdAt: "2025-07-22T11:30:55Z"
          description: CockroachDB Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1304d7f871a95c8b087bad1b548217562792edc66c8fe07c621e05242b8cd53e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6408f89b0a88ec154f7ee98c314be574971c4c9aaf906b473d10124d822889be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:65923c37fecda2d3174212fa9d9a6f34241d065389336ea69a7be99ba16cedc2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3bf017b2b2506724500f5e29017d1b956825dd0f61a7c0049b0ff084c3554b5b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:85a0e085901c2a73192556d5d70795d4b66cbcfe01d24c73b6fe6db8d5f2e5ca
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bd75aa918355e77cd671c3e11ee807861378e96245245f4fdf243cba5749f9a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e8f26fd132affbaeb8ef6b1807a646c83b301e5dc334a592dfb2cf8dc3dbaf10
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6ca927f137be28481e3a226e6c6bcb7c5ac27664984f17ffc6c1419cd7d08eb7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da367cf0ac52045002e1c186f8e6964267ad87d5c25f9e72fd2c9b9a98a32702
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc1ab77a9537ccf236ce6dc370940dfedcfaa91b6e3db37bf2d5eade7ff84708
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e0f21abea34825e55622c432a306f2e8aafd64d913eb78f2f601c51dc878a1da
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1b6d934a7b6c42a61e5ad51380e69af341a575b42205299fd58bc0c6ff306f6a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a6f1c43486d044fed9f59406118f429c8037b389bd8c8d7a1c7aeb89ed2661f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b06bd7f3802e8fa2af14bb0b07804cad4cc0fbbc97da18d09a9106bf074ed6d8
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:740ba10f1ef0a6c5fa6ad31ab18855fc75f8e024ce4c23b0676c5a9c5b62cd75
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c9a31bdf19175880d0981e21dd19a1458b6e8cde8d4acbcb554f3bea3578a64e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf7d3bea47638b763b312c7c11ca034a7ac62d993d75a13edceec74799b97e52
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f438fbbf9c87ca0d54c170e2c5105b1e265182e1c2ede48adea370e2be9dc895
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9ab968e4ca62f23bf9848e5475ef4076df7318a8560f63d5b2090a32af6fe4e2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d978678b3c254abd52f69f226b0aacc4b02c2aaca328c54ef10f6f9bb453582d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:dd521d7bad62da054eec54179be75c643a5492d981e11c0e66db41902cd2182c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d0014b1c19890b9e8ef1a4f3efe7eb5b269be5610ba6ababd4258b52d2f85e8a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5a54d564e08baefc2786f4b742b2507d31984664cd00333406aa7a3aa5b9c913
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5181196fff3bf8a5ec8fcfe4f1958cc69f285f0beb4f0b476b4b5e0912f3db42
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9f42e929f2cc7e9b6d24dd61e19c5d93dcfeabd34b9c8b935afbe78577efb18b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:266165ab87b19a05d6796e8b15ff0eba632d6ac96e6393400464c6136dd500ec
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:46d9287ef11e3b21a36abfa31a26da3b93040bb6163b9223972a5dd1930fecff
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acbfee4492327a720fe7da76929a86192d8af2442bbb76f0a089f6a81a890be3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6371377718325235a2720dba0c97eb4321ee29b31a28cb893f76abd62baa339
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:79d688bebe0643ef0dd6ba7108cdb34c0442470c0c20ae0df88b5d02a3476cd0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e23af8f9d32c3f80a52d79ceb4b359dda1d75ba0dd4cb8ed4b66ac2f4e40c69b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:58de9d0b2ce005c7ad3e99152100e59b7677398557028a7a59e56e669fabb81c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7daaf1368ab2c25694cdea0a598e1de59247b04a4b4b4b8348ed6edbaa194f9d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:35ede65457d845e8c063ad7d9515d3a69f98e0751424ecbac484963678a0b1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f42fa22aac3cc41ce72e144a72530ffbee765a392b24593dafd06cc098f8410b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a4f932c3e9ff3aeb70ba1f7a92aa4642bd14cbb7d37d04ff750ed517b916cdb9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:19f36f53f7da67755eb86da77098c5eeb84b381f9c18aa29ba573793dc498564
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4635550222775fc75737be5e5c0359303e3a23c4d1e1c296d7408cd15ad94aad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f3ffebd812881a830ff46c860984814dd61edcd51b478ebd3667759fb7710682
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2a128b570352a5a7f82b27353f9b774caccce83e0e7bb2673e4802d96ad0bd51
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f74c8a5dea4560ad59c6f6f6ab1e699f063334ee8db704460f870508dbeaeff0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bd336715332f776a72aaf5c3ca52e62dd53a9e8521c1feb6e135cc868d694a39
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5240f9f591263569fad7ee934e81fa8a1617de3c25da94ddd59ee5b2b3407f90
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d14f77f70aea0422a906510c1ac2652ff76e131ed705cb002f80e13e887b0cd9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:10b9df672260c023dfb37166e156c0d7e0e13144dc3e485af922e8151fd3ab05
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2af60025ecb3bb933b61328be4c2b2bfd0e7d26f53b72430629208505e45c6d2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:32fdca575c334822e4356aab36a7ed97b685c065925fe85f1b8ba8425c57159b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:42c9930d6861f6a2147ed6e01827eddd730279966763ec68868567a19cc164ab
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4b3aa590623a38f435b075de4c09f8a343d8ece8b673e6bf7c6df5d5eb4bdf4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:02e1ef01a023da0eee653f63b314786fd66a78d184a97a314422f6f9c3ef8e7e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:310310515625f099a928545865f7096997871ee4a16650a01c76c3799a18b684
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6a2ee8d62c586eb74ebbef6bef256e06e1ba5cb51c1ae55b1654adafc2ca8859
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acc083f7dde1fbc557d858d03fc26b68b2056997fc1a457e4695650214d5b303
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b78a4d3864e16d270979d9d7756012abf09e6cdb7f1eb4832f6497c26281099
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:61e602b0e3582bf00460eb9328808ccf66e3ff1b1f6df07ee6c5c76a91846876
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2d89398f03736f8ffaa82b10fed0f19fba166d3d2145b26f25cb71f3c39acf07
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5458a8fd89aa664b60e59c35bd933975cd37d18e19b11589e4ca0c2844ea0ed4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:50393ad005fed19618d04f0716c6d55e2f1afce651e785c396638e852cac83b3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ec6eb7c28c213cc83b2d7919cd87988f9a07f12276eb7351d0915f1567a5b095
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6366ba29f301a82215c05e60d78e1de66bcee85440ce2fc3c692712d4da32685
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5d6901b6b6fd6594d42bfc1842d050c7cffcb0867a5591a67469f0f860b1f21b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4d8e804b35b73d7978c112748aea2d2bb68f65081b90bca62c39f2b6e2989378
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:743cda1eda6e273c798e3405a9a97e07dd6aa5ad4d5a52789be3b0f39a32a039
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4f753151af4c3d3f5a1df9014461316249a6d0621b8fba5c4430307b3355251
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce1209ca6cc8724baee981c97c66e2852e512423bc0c0a16cdd80a6838b54031
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f5a0ccc02dc9e938e484d5b5282ff650d1890d5f754c30a0c0d717989ed5d600
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f045fe5037d44aab1a0b2209241b3931ec7cb9897cb3c6a96a9f8d5280e55dd2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08cec123813304ab03bd66d72a09949667a874a42320bedafda391ebcdb2e56c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5b3de46526d169fcd1318354545b3a18861843fa2022a23ed04d795493825287
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d590946b2f7c7ec0c7ec057a338e81f864d6a9e3d3ac4e0769aa5756a8e13fc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0754f7901feea9017bc5041cb91b7ac1cdbca10c2064f37c3648a2c395f24f83
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db07461824e73410456ef540b0bce3c7fe550aa67a73d6174300ca74b8f75276
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c74cf927a2b2c5098ad74ad0880996f8b042186a9911d888a231ef5b9e2a0715
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7d6320ac7fb0925f10cf6f524a13de629cbf13ccc371912e4fe8c777bc15d1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:28ae9661b38a72decf413c6a6c3758e61098debf5162197c27a12a070c7a096d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:226d7f1c3f0d70ac6532ad9490bebcdef44a40959b6aa7c11b0f2d1383a46217
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:44a0f40a6c46ba08297b482f3d96defe87b181c944803e5381c9acc42cabf97f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b046aa0a42dae992a03685ed82a20835532d6d4bf048e2366ff7382be4b3138
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1ce71e80c0cd0e0063eba275ab91f77982f2e92a8aef8e89cab99b93feab0a88
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:29abb169ffefe36f6d69edd438a7135ee23059cedab8f87ad3b2d52361b96b35
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:68fe64fe3e112288ccfc2a7dffd69b9b2786a5da3d3bdedd6addf41659d460c9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8443c15eb1653dd9e5ec7aad54d4b8b092d41a0977e42b76e372e5e21280746e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b83eaff2e217c06098509709cead525b55b00ccd48019d8357e6231892008055
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bed27f3f2c2d7627d1af4500d7ea0302da5c3fd3d875a6480fb6d98da481312e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b0057de903654ea0315a2c96affcc8f50a92ede8ad121ed30edb3aaf0fb7b554
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cbc74889e9ca5edbd5aaa243d7649bca0c4da2719239f8d0ec9827868112a771
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1558a1f5eab79f1a8c6c85d81740cc539dc23025ea457c0019d93a72dc05d57a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf001173bbf00cdf443b8d64a3ad724389f1cf83e3628d0b609019c8cddad35a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:87db6e079b96fb392d98ef5fc7063836bff10717efd75061016ce688be0ba38e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:72ea3d0ef72972dba93362dc31366c5946a9898f5c16780b877316a97068ef8f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1374cbdc4ecf38e4b4226328f532e830b8932b34d9449bf2fbb36e60d0310521
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2171e1c75d0c665fe33e1ea334e605604c51de6dbc22d88a3a0ac604259c448f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db4e5c2f27ba23e95a363c8174c01f0f9c6b6a27e0e581c86a38f3a1f68dc43c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d8cfb221f67516675fdfde061f6ec5dfee1224ca7eb3d8be375c27d483f306af
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:aeac9cd8e8acdbd2b0c339c48b85079c159b04f0ade24df8e3f9cae3f48b4c70
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:086d7b435993fdf21fd59b0093f52105a9028d6b769398b0033f98a65dfb7e79
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9d0bfb642090ff55a3cc3da0f72663cae18063622a4ec785ba99bc71ef4db116
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3f587e7b9f5a611013af2608e2151ed782859a52501b00adbc95739c56a8ff1e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:82d8104a37349c513447e68ea3790efb256370632353176e5fc13c2e8df1bd44
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c2a62179e8ca293d7e30b9da599e48795345cf07a72abd3a83063107b53b03cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d467383af41aa80c26172964c34152abeba45121599804e502984655b72179f0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e5814748c5ab229a1dea5662a007063c0df06fbbfcfe041fd9a6df9ef67e07cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a6e2fe24ad52913e2160f85b9fe665a4e4c4d79b164b4b899f69dd9d92f999e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6cd9f9b9fcb6e8f9a931a6ad01cb66329a74d40b22f9029fbc50e57248bad7cb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:56109e57ee0379cf48644bcf8226a9238e01139cacc5499002c99f973f121911
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f53b07212320c6939d3bb4ddd4e6a0bce311ab5b63858eb1a8728531f0dbd510
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce3d94a3fbdcde5858beb3c265987759971d991c397c6486e19904bd4706f44d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e151bded670f4cbd165a8ef864324805667d3e84ce3ace0cdf8f9aad087fccf1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6bb15b36d64eebb6e4c1db5a5466e108b271d53383c58e0b6c78cec214756a9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0e2c7023f1f97acc20c46a4944635a9c5f57bb9bb3dd534cf41fea1a9cfdfdbe
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:91f7e59f9e35a0395756eafed906963c535e886125294d671d930ce959a8f511
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f05824bb8eaa5a93e54059449eaa236dd83d729049ad00065ef3242f5f74c91e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b77b80eed7ce149d2a3f228e5f188e76d73f32bc50ecde20bba9cfa7bdc4714c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cad60044ad6573bd80b65e892a9ec0510dbe4fcbbfc6b51010ecf419f56f1024
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8fc8de4f3671a73f9aa747e5b23ec973d192fff6f2869c38985cd656063c1e3d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:022c6e1a0de875474bc953718a481a0af526c49dc910094dcdce0b5677aced18
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:468b0a40a33be1dd5a2a89ce745d7e9012bc7371f2f01cc0f8f12c0ac887db43
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4e5f7df1dc1e1db398c36d590431e7e5782897b209972d8e9e4671971c10d1b6
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7481c0cf3b579c8d98d94d4aa0334ca94bf2ccfcd25ad70b57c42b22bbf4bcee
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08496b84b3cc691909227be17e67bcd6f4d5185a93712f44d8a0985bd2c5fcd7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fd6c9b8ca82ff4f2d0fe29166d1fc9bca6439cb40907be60f4d8f1b9735de2f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:742a6f69a3a209a7080a1e911fd96467d2a91aacb3f91d6be80a6a6e2e54b2ae
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e82684b828256b3c2814930210031b3b70bc818514d4359a25f87ffdbf22be6e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8737b67b91983817e3e4ff3f47f0b070bd36162d309b60993529426c6979cb24
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4b56c0311af14220a466aa11f25fd10fa38280825539bd02215c945bbafb861c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9097e135159f46396ef4649f724d61bb5b9113c488f989d404586fbb7486fd85
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1e5d85f5db10fa79d09263db90daa44549bcc3b025e9f3f98cf0d4e51766394b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a0faa52a999289b308dc4caf050c6d9fe2f42e112aa30e4af282ed1f21c5c346
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1f6b1ec92251397dbcf7ed4a82f1f05619d64676aae2717f4eff1c33e4c2c27a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a59cafff8d4ff2c62204ece15880117011c30978823c232902c7182d747ec827
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:858dc82977b3c9e989a3a20626671b0c67c834c97132094a9212bb227ab55929
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:744193037d6b64cb73fecab15ebe022abc2b79ab0335b76ec1415ac6379b7dfa
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d78c2e0709214f96794564f13386861f4b541e55b5bd316ad495cf104f7d40b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9b5a2e0c006eb4be8e1444734a98743551f278a24e81c20bebc96f0f3b631eb0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e50aab39722af22ccbf3d1db19e37972d52b0fc9a40998a64618dc966b2bac57
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:818f592c50d4ed5599faa053344f7ecd963003352f1faa3b72dc0f5e00a0e78b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0b0e0588c1aadf0c3826cd584bd2b3a7d6781882dcb5c5e037b47ff5cb84509b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:53c7b960a2ed70a0998a4f8e28a4f78f7f676e01cffebe665e91e3a7629d88ed
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f29c31bb267451b389608af3e8a4a28d6ca0691e67b77e8f57bb5e461b064893
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bf36dab9b3257a4bba24d986f1451fc51a3a37023fbb988281f1941fd3faedd
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7b907d03b001b758d8a4b5e898725c73dda2ec2aa4b7425b685320b947154d11
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8f1f2fd1aca914a1ef007117c85ac944055bbfb873a468a7374d239ef20b065e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:37bf9e1ec7e3eff47755fe7521223f31ec6aefc6e63901db3881b30898cf78f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:40f00476b63cb9e370fe85f01407173693213a9242fc117d28b06378ca0d98e0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8a70ba12f610ac0e62066e417a0833bc1d01e91ac99d0dd794a2c0cd2a350164
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2b5c028f293c6f394e58a5495cb3dbfbedc215d0d03778d0217e045586d55a46
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da3bd2e9ff2b08d5bb86b1f5faa95a865f85c5e8912df00b6b79190c1e518a2f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7ad12df8e258b03c4099d375d02d272bd774bf188d108901d5cf6cb865e57ef3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cc20270212f01fc231ff788e20b0476b8228c440bfc40430c0836d5eb702a042
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:31b161f9bf9028a1a62cd2eef6abda01257d8889bdc3159822642549a75e3429
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0fea6ed8a78ea624240ee6ffb3573d5cf1115186aa9180fcd7d1273351deaaa3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9c85c25d5392817d17b68714dca6195ad53255d43e9c65b021d75b12d97e0579
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a3ac544d87a9d8e7c92897c4c23f803753c7ab31467db91b5c51244afc2c6be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e18d73af3257c8cdf00245f2f718b56667524c1cc69d3db28e9b868a539ae16f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:363d8feb560324a848c5771edd224dc19020013af7641979ee81ae8ee536d7e3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fe4a40006d1e8439423de1417cc9bc55716d790b882485074bb2a5ec68d7d1ad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:90cf21c36d19d87b97d06edd1c0912a3bb288c769e6159651877db3206aa7355
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0ec1dc184a2199736b4bd7de3241a357be7427caea5da1c931b08bc06f5c3dc0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8e8c081c231b5b15eca483ceec16ea9190ba73aae6e421865162446eb584519a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc4b75ed71845d5b18e1d9163d480fd9d0d3cb5ebbcfed00a2e4e1b174c0a5de
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9492894eaa5f43d00ac80b60029b32a0b0a85ae88ba609b46f98702c9349a5bb
        version: 2.18.2
      entries:
      - name: cockroach-operator.v2.18.2
        version: 2.18.2
      name: stable
    defaultChannel: stable
    packageName: cockroachdb-certified
    provider:
      name: Cockroach Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Elastic
      provider-url: ""
    name: elasticsearch-eck-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: elasticsearch-eck-operator-certified.v3.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "elasticsearch.k8s.elastic.co/v1",
                    "kind": "Elasticsearch",
                    "metadata": {
                        "name": "elasticsearch-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "nodeSets": [
                            {
                                "name": "default",
                                "config": {
                                    "node.roles": ["master", "data"],
                                    "node.attr.attr_name": "attr_value",
                                    "node.store.allow_mmap": false
                                },
                                "podTemplate": {
                                    "metadata": {
                                        "labels": {
                                            "foo": "bar"
                                        }
                                    },
                                    "spec": {
                                        "containers": [
                                            {
                                                "name": "elasticsearch",
                                                "resources": {
                                                    "requests": {
                                                        "memory": "4Gi",
                                                        "cpu": 1
                                                    },
                                                    "limits": {
                                                        "memory": "4Gi",
                                                        "cpu": 2
                                                    }
                                                }
                                            }
                                        ]
                                    }
                                },
                                "count": 3
                            }
                        ]
                    }
                },
                {
                    "apiVersion": "kibana.k8s.elastic.co/v1",
                    "kind": "Kibana",
                    "metadata": {
                        "name": "kibana-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        },
                        "podTemplate": {
                            "metadata": {
                                "labels": {
                                    "foo": "bar"
                                }
                            },
                            "spec": {
                                "containers": [
                                    {
                                        "name": "kibana",
                                        "resources": {
                                            "requests": {
                                                "memory": "1Gi",
                                                "cpu": 0.5
                                            },
                                            "limits": {
                                                "memory": "2Gi",
                                                "cpu": 2
                                            }
                                        }
                                    }
                                ]
                            }
                        }
                    }
                },
                {
                    "apiVersion": "apm.k8s.elastic.co/v1",
                    "kind": "ApmServer",
                    "metadata": {
                        "name": "apmserver-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                    "apiVersion": "enterprisesearch.k8s.elastic.co/v1",
                    "kind": "EnterpriseSearch",
                    "metadata": {
                        "name": "ent-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "config": {
                            "ent_search.external_url": "https://localhost:3002"
                        },
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                  "apiVersion": "beat.k8s.elastic.co/v1beta1",
                  "kind": "Beat",
                  "metadata": {
                    "name": "heartbeat-sample"
                  },
                  "spec": {
                    "type": "heartbeat",
                    "version": "9.1.0",
                    "elasticsearchRef": {
                      "name": "elasticsearch-sample"
                    },
                    "config": {
                      "heartbeat.monitors": [
                        {
                          "type": "tcp",
                          "schedule": "@every 5s",
                          "hosts": [
                            "elasticsearch-sample-es-http.default.svc:9200"
                          ]
                        }
                      ]
                    },
                    "deployment": {
                      "replicas": 1,
                      "podTemplate": {
                        "spec": {
                          "securityContext": {
                            "runAsUser": 0
                          }
                        }
                      }
                    }
                  }
                },
                {
                  "apiVersion": "agent.k8s.elastic.co/v1alpha1",
                  "kind": "Agent",
                  "metadata": {
                    "name": "agent-sample"
                  },
                  "spec": {
                    "version": "9.1.0",
                    "elasticsearchRefs": [
                      {
                        "name": "elasticsearch-sample"
                      }
                    ],
                    "daemonSet": {},
                    "config": {
                      "inputs": [
                        {
                          "name": "system-1",
                          "revision": 1,
                          "type": "system/metrics",
                          "use_output": "default",
                          "meta": {
                            "package": {
                              "name": "system",
                              "version": "0.9.1"
                            }
                          },
                          "data_stream": {
                            "namespace": "default"
                          },
                          "streams": [
                            {
                              "id": "system/metrics-system.cpu",
                              "data_stream": {
                                "dataset": "system.cpu",
                                "type": "metrics"
                              },
                              "metricsets": [
                                "cpu"
                              ],
                              "cpu.metrics": [
                                "percentages",
                                "normalized_percentages"
                              ],
                              "period": "10s"
                            }
                          ]
                        }
                      ]
                    }
                  }
                },
                {
                    "apiVersion": "maps.k8s.elastic.co/v1alpha1",
                    "kind": "ElasticMapsServer",
                    "metadata": {
                        "name": "ems-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                    "apiVersion": "logstash.k8s.elastic.co/v1alpha1",
                    "kind": "Logstash",
                    "metadata" : {
                        "name": "logstash-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.connect.redhat.com/elastic/eck-operator@sha256:1750178c5dd716c031155f2e7058e19d4b7f1d3fdbb3e8a806b177a71be88500
          createdAt: "2025-07-30 11:10:34"
          description: Run Elasticsearch, Kibana, APM Server, Beats, Enterprise Search,
            Elastic Agent, Elastic Maps Server and Logstash on Kubernetes and OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Elastic Basic license
          repository: https://github.com/elastic/cloud-on-k8s
          support: elastic.co
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Elastic Agent instance
            displayName: Elastic Agent
            kind: Agent
            name: agents.agent.k8s.elastic.co
            version: v1alpha1
          - description: APM Server instance
            displayName: APM Server
            kind: ApmServer
            name: apmservers.apm.k8s.elastic.co
            version: v1
          - description: Beats instance
            displayName: Beats
            kind: Beat
            name: beats.beat.k8s.elastic.co
            version: v1beta1
          - description: Elastic Maps Server instance
            displayName: Elastic Maps Server
            kind: ElasticMapsServer
            name: elasticmapsservers.maps.k8s.elastic.co
            version: v1alpha1
          - description: Instance of an Elasticsearch autoscaler
            displayName: Elasticsearch Autoscaler
            kind: ElasticsearchAutoscaler
            name: elasticsearchautoscalers.autoscaling.k8s.elastic.co
            version: v1alpha1
          - description: Instance of an Elasticsearch cluster
            displayName: Elasticsearch Cluster
            kind: Elasticsearch
            name: elasticsearches.elasticsearch.k8s.elastic.co
            version: v1
          - description: Enterprise Search instance
            displayName: Enterprise Search
            kind: EnterpriseSearch
            name: enterprisesearches.enterprisesearch.k8s.elastic.co
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.kibana.k8s.elastic.co
            version: v1
          - description: Logstash instance
            displayName: Logstash
            kind: Logstash
            name: logstashes.logstash.k8s.elastic.co
            version: v1alpha1
          - description: Elastic Stack Config Policy
            displayName: Elastic Stack Config Policy
            kind: StackConfigPolicy
            name: stackconfigpolicies.stackconfigpolicy.k8s.elastic.co
            version: v1alpha1
        description: |-
          Elastic Cloud on Kubernetes (ECK) is the official operator by Elastic for automating the deployment, provisioning, management, and orchestration of Elasticsearch, Kibana, APM Server, Beats, Enterprise Search, Elastic Agent, Elastic Maps Server, and Logstash on Kubernetes.

          Current features:

          *  Elasticsearch, Kibana, APM Server, Enterprise Search, Beats, Elastic Agent, Elastic Maps Server, and Logstash deployments
          *  TLS Certificates management
          *  Safe Elasticsearch cluster configuration and topology changes
          *  Persistent volumes usage
          *  Custom node configuration and attributes
          *  Secure settings keystore updates

          Supported versions:

          * Kubernetes 1.29-1.33
          * OpenShift 4.15-4.19
          * Google Kubernetes Engine (GKE), Azure Kubernetes Service (AKS), and Amazon Elastic Kubernetes Service (EKS)
          * Elasticsearch, Kibana, APM Server: 7.17+, 8+, 9+
          * Enterprise Search: 7.17+, 8+
          * Beats: 7.17+, 8+, 9+
          * Elastic Agent: 7.17+, 8+, 9+
          * Elastic Maps Server: 7.17+, 8+, 9+
          * Logstash 8.12+, 9+

          ECK should work with all conformant installers as listed in these [FAQs](https://github.com/cncf/k8s-conformance/blob/master/faq.md#what-is-a-distribution-hosted-platform-and-an-installer). Distributions include source patches and so may not work as-is with ECK.
          Alpha, beta, and stable API versions follow the same [conventions used by Kubernetes](https://kubernetes.io/docs/concepts/overview/kubernetes-api/#api-versioning).
          See the full [Elastic support matrix](https://www.elastic.co/support/matrix#matrix_kubernetes) for more information.
          See the [Quickstart](https://www.elastic.co/guide/en/cloud-on-k8s/3.1/k8s-quickstart.html) to get started with ECK.
        displayName: Elasticsearch (ECK) Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - analytics
        - search
        - database
        - apm
        - logstash
        links:
        - name: Documentation
          url: https://www.elastic.co/guide/en/cloud-on-k8s/3.1/index.html
        maintainers:
        - email: eck@elastic.co
          name: Elastic
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Elastic
        relatedImages:
        - registry.connect.redhat.com/elastic/eck-operator@sha256:1750178c5dd716c031155f2e7058e19d4b7f1d3fdbb3e8a806b177a71be88500
        version: 3.1.0
      entries:
      - name: elasticsearch-eck-operator-certified.v3.1.0
        version: 3.1.0
      - name: elasticsearch-eck-operator-certified.v3.0.0
        version: 3.0.0
      - name: elasticsearch-eck-operator-certified.v2.16.1
        version: 2.16.1
      - name: elasticsearch-eck-operator-certified.v2.16.0
        version: 2.16.0
      - name: elasticsearch-eck-operator-certified.v2.15.0
        version: 2.15.0
      - name: elasticsearch-eck-operator-certified.v2.14.0
        version: 2.14.0
      - name: elasticsearch-eck-operator-certified.v2.13.0
        version: 2.13.0
      - name: elasticsearch-eck-operator-certified.v2.12.1
        version: 2.12.1
      - name: elasticsearch-eck-operator-certified.v2.12.0
        version: 2.12.0
      - name: elasticsearch-eck-operator-certified.v2.11.1
        version: 2.11.1
      - name: elasticsearch-eck-operator-certified.v2.11.0
        version: 2.11.0
      - name: elasticsearch-eck-operator-certified.v2.10.0
        version: 2.10.0
      - name: elasticsearch-eck-operator-certified.v2.9.0
        version: 2.9.0
      - name: elasticsearch-eck-operator-certified.v2.8.0
        version: 2.8.0
      - name: elasticsearch-eck-operator-certified.v2.7.0
        version: 2.7.0
      - name: elasticsearch-eck-operator-certified.v2.6.2
        version: 2.6.2
      - name: elasticsearch-eck-operator-certified.v2.6.1
        version: 2.6.1
      - name: elasticsearch-eck-operator-certified.v2.5.0
        version: 2.5.0
      - name: elasticsearch-eck-operator-certified.v2.4.0
        version: 2.4.0
      - name: elasticsearch-eck-operator-certified.v2.3.0
        version: 2.3.0
      - name: elasticsearch-eck-operator-certified.v2.2.0
        version: 2.2.0
      - name: elasticsearch-eck-operator-certified.v2.1.0
        version: 2.1.0
      - name: elasticsearch-eck-operator-certified.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: elasticsearch-eck-operator-certified
    provider:
      name: Elastic
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Portworx by PureStorage
      provider-url: ""
    name: portworx-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: portworx-operator.v25.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageCluster",
                "metadata": {
                  "name": "portworx",
                  "namespace": "test-operator",
                  "annotations": {
                    "portworx.io/is-openshift": "true"
                  }
                },
                "spec": {}
              },
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageNode",
                "metadata": {
                  "name": "example",
                  "namespace": "test-operator"
                },
                "spec": {}
              }
            ]
          capabilities: Auto Pilot
          categories: Storage
          certified: "true"
          console.openshift.io/plugins: '["portworx"]'
          containerImage: registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
          createdAt: "2025-09-06T19:38:43Z"
          description: Cloud native storage solution for production workloads
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "core.libopenstorage.org/v1",
              "kind": "StorageCluster",
              "metadata": {
                "name": "portworx",
                "annotations": {
                  "portworx.io/is-openshift": "true"
                }
              }
            }
          support: Portworx, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Storage Cluster installs Portworx in the cluster. It has
              all the necessary configurations to setup and update a Portworx cluster.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.core.libopenstorage.org
            version: v1
          - description: Do not create Storage Node as it is internally created by
              the operator. It represents the status of a Portworx node.
            displayName: Storage Node
            kind: StorageNode
            name: storagenodes.core.libopenstorage.org
            version: v1
        description: |
          Portworx-Enterprise is the most widely-used and reliable cloud-native
          storage solution for production workloads and provides high-availability,
          data protection and security for containerized applications.

          Portworx Enterprise enables you to migrate entire applications, including
          data, between clusters in a single data center or cloud, or between clouds,
          with a single kubectl command.

          The cloud native storage and data management platform that enterprises trust
          to manage data in containers now has an operator which simplifies the install,
          configuration, upgrades and manages the Portworx Enterprise cluster lifecycle.

          Learn more about the Portworx Enterprise
          [the data platform for Kubernetes](https://portworx.com/products/introduction)

          To learn more about the platform features, please visit our
          [product features page](https://portworx.com/products/features)

          ### About Portworx

          Portworx is the solution for running stateful containers in production,
          designed with DevOps in mind. With Portworx, users can manage any database
          or stateful service on any infrastructure using any container scheduler,
          including Kubernetes, Mesosphere DC/OS, and Docker Swarm. Portworx solves
          the five most common problems DevOps teams encounter when running stateful
          services in production: persistence, high availability, data automation,
          security, and support for multiple data stores and infrastructure.

          ### How to install StorageCluster

          To customize your cluster's configuration (specification), use the
          [Spec Generator](https://central.portworx.com/) from PX-Central.

          ### Prerequisite

          Ensure ports 17001-17020 on worker nodes are reachable from master and other worker nodes.

          ### Tutorials

          * [Portworx Enterprise on Openshift](https://portworx.com/openshift)

          * [Stateful applications on Kubernetes](https://docs.portworx.com/portworx-enterprise/concepts/kubernetes-storage-101/applications)

          * [Portworx Enterprise on Kubernetes](https://docs.portworx.com/portworx-install-with-kubernetes)

          * [Kafka on Kubernetes](https://portworx.com/kafka-kubernetes)

          * [Elastisearch on Kubernetes](https://portworx.com/elasticsearch-kubernetes)

          * [PostgreSQL on Kubernetes](https://portworx.com/postgres-kubernetes/)

          * [MongoDB on Kubernetes](https://portworx.com/mongodb-kubernetes/)

          * [Cassandra on Kubernetes](https://portworx.com/cassandra-kubernetes/)

          * [Kubernetes backup and recovery](https://portworx.com/kubernetes-backup/)

          * [Disaster Recovery for Kubernetes](https://portworx.com/kubernetes-disaster-recovery/)

          ### Uninstall

          Deleting the StorageCluster object for Portworx cluster does not stop Portworx
          service running on the nodes, to avoid application downtime.

          To uninstall Portworx completely without wiping the data, you should add the
          following delete strategy to the StorageCluster spec:
          ```
          spec:
            deleteStrategy:
              type: Uninstall
          ```
          **Caution:** To uninstall Portworx and **wipe all the data**, you should use the following
          delete strategy:
          ```
          spec:
            deleteStrategy:
              type: UninstallAndWipe
          ```
        displayName: Portworx Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - portworx
        - persistent storage
        - storage
        - cloud native
        - open source
        links:
        - name: Product Features
          url: https://portworx.com/products/features
        - name: Documentation
          url: https://docs.portworx.com/portworx-enterprise/platform/openshift
        - name: Support
          url: https://docs.portworx.com/portworx-enterprise/support/contact-support
        - name: Source Code
          url: https://github.com/libopenstorage/operator
        maintainers:
        - email: support@portworx.com
          name: Portworx
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Portworx by PureStorage
        relatedImages:
        - registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
        version: 25.3.1
      entries:
      - name: portworx-operator.v25.3.1
        version: 25.3.1
      - name: portworx-operator.v25.3.0
        version: 25.3.0
      - name: portworx-operator.v25.2.2
        version: 25.2.2
      - name: portworx-operator.v25.2.1
        version: 25.2.1
      - name: portworx-operator.v25.2.0
        version: 25.2.0
      - name: portworx-operator.v24.2.4
        version: 24.2.4
      - name: portworx-operator.v24.2.3
        version: 24.2.3
      - name: portworx-operator.v24.2.2
        version: 24.2.2
      - name: portworx-operator.v24.2.1
        version: 24.2.1
      - name: portworx-operator.v24.2.0
        version: 24.2.0
      - name: portworx-operator.v24.1.3
        version: 24.1.3
      - name: portworx-operator.v24.1.2
        version: 24.1.2
      - name: portworx-operator.v24.1.1
        version: 24.1.1
      - name: portworx-operator.v24.1.0
        version: 24.1.0
      - name: portworx-operator.v23.10.5
        version: 23.10.5
      - name: portworx-operator.v23.10.4
        version: 23.10.4
      - name: portworx-operator.v23.10.3
        version: 23.10.3
      - name: portworx-operator.v23.10.2
        version: 23.10.2
      - name: portworx-operator.v23.10.1
        version: 23.10.1
      - name: portworx-operator.v23.7.0
        version: 23.7.0
      - name: portworx-operator.v23.5.1
        version: 23.5.1
      - name: portworx-operator.v23.5.0
        version: 23.5.0
      - name: portworx-operator.v23.4.0
        version: 23.4.0
      - name: portworx-operator.v23.3.1
        version: 23.3.1
      - name: portworx-operator.v23.3.0
        version: 23.3.0
      - name: portworx-operator.v1.10.5
        version: 1.10.5
      - name: portworx-operator.v1.10.4
        version: 1.10.4
      - name: portworx-operator.v1.10.3
        version: 1.10.3
      - name: portworx-operator.v1.10.2
        version: 1.10.2
      - name: portworx-operator.v1.10.1
        version: 1.10.1
      - name: portworx-operator.v1.10.0
        version: 1.10.0
      - name: portworx-operator.v1.9.1
        version: 1.9.1
      - name: portworx-operator.v1.9.0
        version: 1.9.0
      - name: portworx-operator.v1.8.1
        version: 1.8.1
      - name: portworx-operator.v1.8.0
        version: 1.8.0
      - name: portworx-operator.v1.7.1
        version: 1.7.1
      - name: portworx-operator.v1.6.1
        version: 1.6.1
      name: alpha
    - currentCSV: portworx-operator.v25.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageCluster",
                "metadata": {
                  "name": "portworx",
                  "namespace": "test-operator",
                  "annotations": {
                    "portworx.io/is-openshift": "true"
                  }
                },
                "spec": {}
              },
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageNode",
                "metadata": {
                  "name": "example",
                  "namespace": "test-operator"
                },
                "spec": {}
              }
            ]
          capabilities: Auto Pilot
          categories: Storage
          certified: "true"
          console.openshift.io/plugins: '["portworx"]'
          containerImage: registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
          createdAt: "2025-09-06T19:38:43Z"
          description: Cloud native storage solution for production workloads
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "core.libopenstorage.org/v1",
              "kind": "StorageCluster",
              "metadata": {
                "name": "portworx",
                "annotations": {
                  "portworx.io/is-openshift": "true"
                }
              }
            }
          support: Portworx, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Storage Cluster installs Portworx in the cluster. It has
              all the necessary configurations to setup and update a Portworx cluster.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.core.libopenstorage.org
            version: v1
          - description: Do not create Storage Node as it is internally created by
              the operator. It represents the status of a Portworx node.
            displayName: Storage Node
            kind: StorageNode
            name: storagenodes.core.libopenstorage.org
            version: v1
        description: |
          Portworx-Enterprise is the most widely-used and reliable cloud-native
          storage solution for production workloads and provides high-availability,
          data protection and security for containerized applications.

          Portworx Enterprise enables you to migrate entire applications, including
          data, between clusters in a single data center or cloud, or between clouds,
          with a single kubectl command.

          The cloud native storage and data management platform that enterprises trust
          to manage data in containers now has an operator which simplifies the install,
          configuration, upgrades and manages the Portworx Enterprise cluster lifecycle.

          Learn more about the Portworx Enterprise
          [the data platform for Kubernetes](https://portworx.com/products/introduction)

          To learn more about the platform features, please visit our
          [product features page](https://portworx.com/products/features)

          ### About Portworx

          Portworx is the solution for running stateful containers in production,
          designed with DevOps in mind. With Portworx, users can manage any database
          or stateful service on any infrastructure using any container scheduler,
          including Kubernetes, Mesosphere DC/OS, and Docker Swarm. Portworx solves
          the five most common problems DevOps teams encounter when running stateful
          services in production: persistence, high availability, data automation,
          security, and support for multiple data stores and infrastructure.

          ### How to install StorageCluster

          To customize your cluster's configuration (specification), use the
          [Spec Generator](https://central.portworx.com/) from PX-Central.

          ### Prerequisite

          Ensure ports 17001-17020 on worker nodes are reachable from master and other worker nodes.

          ### Tutorials

          * [Portworx Enterprise on Openshift](https://portworx.com/openshift)

          * [Stateful applications on Kubernetes](https://docs.portworx.com/portworx-enterprise/concepts/kubernetes-storage-101/applications)

          * [Portworx Enterprise on Kubernetes](https://docs.portworx.com/portworx-install-with-kubernetes)

          * [Kafka on Kubernetes](https://portworx.com/kafka-kubernetes)

          * [Elastisearch on Kubernetes](https://portworx.com/elasticsearch-kubernetes)

          * [PostgreSQL on Kubernetes](https://portworx.com/postgres-kubernetes/)

          * [MongoDB on Kubernetes](https://portworx.com/mongodb-kubernetes/)

          * [Cassandra on Kubernetes](https://portworx.com/cassandra-kubernetes/)

          * [Kubernetes backup and recovery](https://portworx.com/kubernetes-backup/)

          * [Disaster Recovery for Kubernetes](https://portworx.com/kubernetes-disaster-recovery/)

          ### Uninstall

          Deleting the StorageCluster object for Portworx cluster does not stop Portworx
          service running on the nodes, to avoid application downtime.

          To uninstall Portworx completely without wiping the data, you should add the
          following delete strategy to the StorageCluster spec:
          ```
          spec:
            deleteStrategy:
              type: Uninstall
          ```
          **Caution:** To uninstall Portworx and **wipe all the data**, you should use the following
          delete strategy:
          ```
          spec:
            deleteStrategy:
              type: UninstallAndWipe
          ```
        displayName: Portworx Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - portworx
        - persistent storage
        - storage
        - cloud native
        - open source
        links:
        - name: Product Features
          url: https://portworx.com/products/features
        - name: Documentation
          url: https://docs.portworx.com/portworx-enterprise/platform/openshift
        - name: Support
          url: https://docs.portworx.com/portworx-enterprise/support/contact-support
        - name: Source Code
          url: https://github.com/libopenstorage/operator
        maintainers:
        - email: support@portworx.com
          name: Portworx
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Portworx by PureStorage
        relatedImages:
        - registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
        version: 25.3.1
      entries:
      - name: portworx-operator.v25.3.1
        version: 25.3.1
      - name: portworx-operator.v25.3.0
        version: 25.3.0
      - name: portworx-operator.v25.2.2
        version: 25.2.2
      - name: portworx-operator.v25.2.1
        version: 25.2.1
      - name: portworx-operator.v25.2.0
        version: 25.2.0
      - name: portworx-operator.v24.2.4
        version: 24.2.4
      - name: portworx-operator.v24.2.3
        version: 24.2.3
      - name: portworx-operator.v24.2.2
        version: 24.2.2
      - name: portworx-operator.v24.2.1
        version: 24.2.1
      - name: portworx-operator.v24.2.0
        version: 24.2.0
      - name: portworx-operator.v24.1.3
        version: 24.1.3
      - name: portworx-operator.v24.1.2
        version: 24.1.2
      - name: portworx-operator.v24.1.1
        version: 24.1.1
      - name: portworx-operator.v24.1.0
        version: 24.1.0
      - name: portworx-operator.v23.10.5
        version: 23.10.5
      - name: portworx-operator.v23.10.4
        version: 23.10.4
      - name: portworx-operator.v23.10.3
        version: 23.10.3
      - name: portworx-operator.v23.10.2
        version: 23.10.2
      - name: portworx-operator.v23.10.1
        version: 23.10.1
      - name: portworx-operator.v23.7.0
        version: 23.7.0
      - name: portworx-operator.v23.5.1
        version: 23.5.1
      - name: portworx-operator.v23.5.0
        version: 23.5.0
      - name: portworx-operator.v23.4.0
        version: 23.4.0
      - name: portworx-operator.v23.3.1
        version: 23.3.1
      - name: portworx-operator.v23.3.0
        version: 23.3.0
      - name: portworx-operator.v1.10.5
        version: 1.10.5
      - name: portworx-operator.v1.10.4
        version: 1.10.4
      - name: portworx-operator.v1.10.3
        version: 1.10.3
      - name: portworx-operator.v1.10.2
        version: 1.10.2
      - name: portworx-operator.v1.10.1
        version: 1.10.1
      - name: portworx-operator.v1.10.0
        version: 1.10.0
      - name: portworx-operator.v1.9.1
        version: 1.9.1
      - name: portworx-operator.v1.9.0
        version: 1.9.0
      - name: portworx-operator.v1.8.1
        version: 1.8.1
      - name: portworx-operator.v1.8.0
        version: 1.8.0
      - name: portworx-operator.v1.7.1
        version: 1.7.1
      - name: portworx-operator.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: portworx-certified
    provider:
      name: Portworx by PureStorage
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: clusterresourceoverride
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clusterresourceoverride-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.autoscaling.openshift.io/v1",
                "kind": "ClusterResourceOverride",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "podResourceOverride": {
                    "spec": {
                      "memoryRequestToLimitPercent": 50,
                      "cpuRequestToLimitPercent": 25,
                      "limitCPUToMemoryPercent": 200
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certifiedLevel: Primed
          containerImage: registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9-operator@sha256:40b19c94f688c3a58189fcb404df35a40a8acbbdbf73fadb8818a86fd14a9ab9
          createdAt: 2019/11/15
          description: An operator to manage the OpenShift ClusterResourceOverride
            Mutating Admission Webhook Server
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          healthIndex: B
          olm.skipRange: '>=4.3.0 <4.16.0-202509240123'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-resource-override-admission-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of ClusterResourceOverride Admission
              Webhook
            displayName: ClusterResourceOverride
            kind: ClusterResourceOverride
            name: clusterresourceoverrides.operator.autoscaling.openshift.io
            version: v1
        description: "ClusterResourceOverride\n==============\n\nContainers can specify
          compute resource requests and limits. Requests are used for scheduling your
          container and provide a minimum service guarantee. Limits constrain the
          amount of compute resource that may be consumed on your node.\n\nThe scheduler
          attempts to optimize the compute resource use across all nodes in your cluster.
          It places pods onto specific nodes, taking the pods' compute resource requests
          and nodes' available capacity into consideration.\n\nRequests and limits
          enable administrators to allow and manage the overcommitment of resources
          on a node, which may be desirable in development environments where a trade
          off of guaranteed performance for capacity is acceptable.\n\n### Requests
          and Limits\n\nFor each compute resource, a container may specify a resource
          request and limit. Scheduling decisions are made based on the request to
          ensure that a node has enough capacity available to meet the requested value.
          If a container specifies limits, but omits requests, the requests are defaulted
          to the limits. A container is not able to exceed the specified limit on
          the node.\n\nThe enforcement of limits is dependent upon the compute resource
          type. If a container makes no request or limit, the container is scheduled
          to a node with no resource guarantees. In practice, the container is able
          to consume as much of the specified resource as is available with the lowest
          local priority. In low resource situations, containers that specify no resource
          requests are given the lowest quality of service.\n\n### Compute Resources\n\nThe
          node-enforced behavior for compute resources is specific to the resource
          type.\n\n#### CPU\n\nA container is guaranteed the amount of CPU it requests
          and is additionally able to consume excess CPU available on the node, up
          to any limit specified by the container. If multiple containers are attempting
          to use excess CPU, CPU time is distributed based on the amount of CPU requested
          by each container.\n\nFor example, if one container requested 500m of CPU
          time and another container requested 250m of CPU time, then any extra CPU
          time available on the node is distributed among the containers in a 2:1
          ratio. If a container specified a limit, it will be throttled not to use
          more CPU than the specified limit.\n\nCPU requests are enforced using the
          CFS shares support in the Linux kernel. By default, CPU limits are enforced
          using the CFS quota support in the Linux kernel over a 100ms measuring interval,
          though this can be disabled.\n\n#### Memory\n\nA container is guaranteed
          the amount of memory it requests. A container may use more memory than requested,
          but once it exceeds its requested amount, it could be killed in a low memory
          situation on the node.\n\nIf a container uses less memory than requested,
          it will not be killed unless system tasks or daemons need more memory than
          was accounted for in the node’s resource reservation. If a container specifies
          a limit on memory, it is immediately killed if it exceeds the limit amount.\n\n###
          Configuring the Cluster for Overcommitment\n\nScheduling is based on resources
          requested, while quota and hard limits refer to resource limits, which can
          be set higher than requested resources. The difference between request and
          limit determines the level of overcommit; for instance, if a container is
          given a memory request of 1Gi and a memory limit of 2Gi, it is scheduled
          based on the 1Gi request being available on the node, but could use up to
          2Gi; so it is 200% overcommitted. If OpenShift Container Platform administrators
          would like to control the level of overcommit and manage container density
          on nodes, ClusterResourceOverride Admission Webhook can be configured to
          override the ratio between request and limit set on developer containers.
          In conjunction with a per-project LimitRange specifying limits and defaults,
          this adjusts the container limit and request to achieve the desired level
          of overcommit.\n\nThis requires creating a custom resource of `ClusterResourceOverride`
          type as in the following example:\n\n    \n    - apiVersion: operator.autoscaling.openshift.io/v1\n
          \   - kind: ClusterResourceOverride\n    - metadata:\n    -   name: cluster\n
          \   - spec:\n    -   podResourceOverride:\n    -     spec:\n    -       memoryRequestToLimitPercent:
          25\n    -       cpuRequestToLimitPercent: 25\n    -       limitCPUToMemoryPercent:
          200\n    \n **memoryRequestToLimitPercent**: (optional, 1-100) If a container
          memory limit has been specified or defaulted, the memory request is overridden
          to this percentage of the limit.\n\n **cpuRequestToLimitPercent**: (optional,
          1-100) If a container CPU limit has been specified or defaulted, the CPU
          request is overridden to this percentage of the limit.\n\n **limitCPUToMemoryPercent**:
          (optional, positive integer) If a container memory limit has been specified
          or defaulted, the CPU limit is overridden to a percentage of the memory
          limit, with a 100 percentage scaling 1Gi of RAM to equal 1 CPU core. This
          is processed prior to overriding CPU request (if configured).\n\n Note that
          these overrides have no effect if no limits have been set on containers.
          [Create a LimitRange object] (https://docs.openshift.com/container-platform/3.3/admin_guide/limits.html#admin-guide-limits)
          with default limits (per individual project, or in the [project template](https://docs.openshift.com/container-platform/3.3/admin_guide/managing_projects.html#modifying-the-template-for-new-projects))
          in order to ensure that the overrides apply.\n\nWhen configured, overrides
          can be enabled per-project by applying the following label.\n    ```\n    clusterresourceoverrides.admission.autoscaling.openshift.io/enabled:
          \"true\"\n    ```\n\n"
        displayName: ClusterResourceOverride Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - deschedule
        - scale
        - binpack
        - efficiency
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9-operator@sha256:40b19c94f688c3a58189fcb404df35a40a8acbbdbf73fadb8818a86fd14a9ab9
        - registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9@sha256:e4992c556fe6be50fa30a45b05cffd6029ebd8914eca7c1c683564ff869e0e7d
        version: 4.16.0-202509240123
      entries:
      - name: clusterresourceoverride-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: clusterresourceoverride-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: clusterresourceoverride-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: clusterresourceoverride-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: clusterresourceoverride-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: clusterresourceoverride-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: clusterresourceoverride-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: clusterresourceoverride-operator.v4.16.0-202505092110
        version: 4.16.0-202505092110
      - name: clusterresourceoverride-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: clusterresourceoverride-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: clusterresourceoverride-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: clusterresourceoverride-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: clusterresourceoverride-operator.v4.16.0-202501241104
        version: 4.16.0-202501241104
      - name: clusterresourceoverride-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: clusterresourceoverride-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: clusterresourceoverride-operator.v4.16.0-202412021735
        version: 4.16.0-202412021735
      - name: clusterresourceoverride-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: clusterresourceoverride-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: clusterresourceoverride-operator.v4.16.0-202411081636
        version: 4.16.0-202411081636
      - name: clusterresourceoverride-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: clusterresourceoverride-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: clusterresourceoverride-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: clusterresourceoverride-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: clusterresourceoverride-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: clusterresourceoverride-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: clusterresourceoverride-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: clusterresourceoverride-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: clusterresourceoverride-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: clusterresourceoverride-operator.v4.16.0-202407040206
        version: 4.16.0-202407040206
      - name: clusterresourceoverride-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: clusterresourceoverride-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: clusterresourceoverride-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: clusterresourceoverride-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: clusterresourceoverride
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicurio-registry-3
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-3.v3.0.15-r1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "simple"
                },
                "spec": {
                  "app": {
                    "ingress": {
                      "host": "simple-app.apps.cluster.example"
                    }
                  },
                  "ui": {
                    "ingress": {
                      "host": "simple-ui.apps.cluster.example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/apicurio/apicurio-registry-rhel9-operator@sha256:e6d2ceab8accdb6e1b9766733039e24df620d7128eb1e13e60a61339130d52a7
          createdAt: "2025-09-29"
          description: Deploy and manage Red Hat build of Apicurio Registry 3 on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations",
            "Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Red Hat build
              of Apicurio Registry version 3.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Red Hat build of Apicurio Registry 3

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Allows you manage event schemas and API designs using the Apicurio Registry web console, REST API, Maven plug-in, or Java client
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, MySQL\*, SQL Server\*, in-memory\*
          - Compatible with Confluent APIs
          - Runs on a lightweight Quarkus platform

          _Note: Storage options marked with * have not yet been tested by Red Hat for production deployments._

          ## Red Hat build of Apicurio Registry 3 Operator

          Provides a quick and easy way to deploy and manage Red Hat build of Apicurio Registry 3 on OpenShift.

          **Features**
          - Supports installation and configuration of Red Hat build of Apicurio Registry 3
          - Perform a rolling upgrade of Red Hat build of Apicurio Registry 3

          ## Prerequisites

          This operator does not deploy storage for Red Hat build of Apicurio Registry 3. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Red Hat build of Apicurio Registry 3 and Apicurio Registry 3 Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Red Hat build of Apicurio Registry 3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/apicurio/apicurio-registry-ui-rhel9@sha256:e4791546fd371b34c7580c6aca2faed8527c985c8029fa8087fb46fd94ab8feb
        - registry.redhat.io/apicurio/apicurio-registry-rhel9-operator@sha256:e6d2ceab8accdb6e1b9766733039e24df620d7128eb1e13e60a61339130d52a7
        - registry.redhat.io/apicurio/apicurio-registry-rhel9@sha256:c86d73d5e2b4146f016ce7ec67265fe80bd596f472247eb9f5590cc0c89dba75
        version: 3.0.15-r1
      entries:
      - name: apicurio-registry-3.v3.0.15-r1
        version: 3.0.15-r1
      - name: apicurio-registry-3.v3.0.9-r3
        version: 3.0.9-r3
      - name: apicurio-registry-3.v3.0.9-r2
        version: 3.0.9-r2
      - name: apicurio-registry-3.v3.0.9-r1
        version: 3.0.9-r1
      - name: apicurio-registry-3.v3.0.7-r2
        version: 3.0.7-r2
      - name: apicurio-registry-3.v3.0.7-r1
        version: 3.0.7-r1
      - name: apicurio-registry-3.v3.0.6-beta.r1
        version: 3.0.6-beta.r1
      - name: apicurio-registry-3.v3.0.6-beta.r1-0.1740495889.p
        version: 3.0.6-beta.r1+0.1740495889.p
      - name: apicurio-registry-3.v3.0.6-beta
        version: 3.0.6-beta
      name: 3.x
    defaultChannel: 3.x
    packageName: apicurio-registry-3
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Instana
      provider-url: ""
    name: instana-agent-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: instana-agent-operator.v2.1.36
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instana.io/v1",
                "kind": "InstanaAgent",
                "metadata": {
                  "name": "instana-agent",
                  "namespace": "instana-agent"
                },
                "spec": {
                  "agent": {
                    "configuration_yaml": "# You can leave this empty, or use this to configure your instana agent.\n# See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.\n",
                    "endpointHost": "ingress-red-saas.instana.io",
                    "endpointPort": "443",
                    "env": {},
                    "key": "replace-key"
                  },
                  "cluster": {
                    "name": "my-cluster"
                  },
                  "opentelemetry": {
                    "grpc": {
                      "enabled": true
                    },
                    "http": {
                      "enabled": true
                    }
                  },
                  "zone": {
                    "name": "edited-zone"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring,Logging & Tracing,OpenShift Optional
          certified: "false"
          containerImage: icr.io/instana/instana-agent-operator@sha256:ab8b35f0b45c68f69210d1b17c5dd609799f6bb0bc9f12d0d4e5302c910607a1
          description: Fully automated Application Performance Monitoring (APM) for
            microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/instana-agent-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/instana-agent-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/instana/instana-agent-operator
          support: Instana
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InstanaAgent is the Schema for the agents API
            displayName: Instana Agent
            kind: InstanaAgent
            name: agents.instana.io
            version: v1
          - description: InstanaAgentRemote is the Schema for the agents API
            displayName: Instana Agent Remote
            kind: InstanaAgentRemote
            name: agentsremote.instana.io
            version: v1
        description: |
          # Instana

          Instana is an [APM solution](https://www.instana.com/product-overview/) built for microservices that enables IT Ops to build applications faster and deliver higher quality services by automating monitoring, tracing and root cause analysis. The solution is optimized for [Kubernetes](https://www.instana.com/automatic-kubernetes-monitoring/) and [OpenShift](https://www.instana.com/blog/automatic-root-cause-analysis-for-openshift-applications/).

          ## Instana Agent Operator

          This is the Kubernetes Operator for installing the Instana Agent on Kubernetes or OpenShift.

          ## Prerequisites for OpenShift

          Before the agent will be able to run in OpenShift, you need to perform a couple of extra configuration steps.

          You need to set up a project for the Instana Agent and configure it's permissions.

          The project you create here needs to be the namespace where you create the Instana Agent custom resource that the operator will use to deploy the agent.

          For example, create the `instana-agent` project:

              oc new-project instana-agent

          Then, ensure the `instana-agent` service account is in the privileged security context:

              oc adm policy add-scc-to-user privileged -z instana-agent

          The `instana-agent-remote` service account must be allowed to use any userid:

              oc adm policy add-scc-to-user anyuid -z instana-agent-remote -n instana-agent

          This service account will be created by the operator.

          Now you can proceed with installing the operator for the Instana agent.

          ## Installation and Configuration

          First, install this operator from [OperatorHub.io](https://operatorhub.io/), [OpenShift Container Platform](https://www.openshift.com/), or [OKD](https://www.okd.io/).

          Second, create a custom resource with the agent configuration in the target namespace (for now, this MUST always be the `instana-agent` namespace). The operator will pick up the custom resource and install the Instana agent accordingly.

          The following is a minimal template of the custom resource:

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgent
          metadata:
            name: instana-agent
            namespace: instana-agent
          spec:
            zone:
              name: my-zone # (optional) name of the zone of the host
            cluster:
              name: replace-me # replace with the name of your Kubernetes cluster
            agent:
              key: replace-me # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io # the monitoring ingress endpoint
              endpointPort: "443" # the monitoring ingress endpoint port, wrapped in quotes
              env:
                INSTANA_AGENT_TAGS: example
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://ibm.biz/monitoring-k8s
          ```

          Save the template in a file `instana-agent.yaml` and edit the following values:

          * If your target namespace is not `instana-agent`, replace the `namespace:` accordingly.
          * `agent.key` must be set with your Instana agent key.
          * `agent.endpointHost` must be set with the monitoring ingress endpoint, generally either `saas-us-west-2.instana.io` or `saas-eu-west-1.instana.io`.
          * `agent.endpointPort` must be set with the monitoring ingress port, generally "443" (wrapped in quotes).
          * `zone.name` should be set with the name of the Kubernetes cluster that is be displayed in Instana.

          For advanced configuration, you can edit the contents of the `configuration.yaml` file. View documentation [here](https://ibm.biz/monitoring-k8s).

          Apply the custom resource with `kubectl apply -f instana-agent.yaml`. After some time, you should see `instana-agent` Pods being created on each node of your cluster, and your cluster should show on the infrastructure map on your Instana Web interface.

          In addition to the InstanaAgent custom resource, a remote agent can be deployed. This agent will not monitor workload running inside the cluster, but can deploy a single replica of an agent which can get an independent configuration for running remote monitoring for certain sensors.

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgentRemote
          metadata:
            name: remote-agent
            namespace: "instana-agent"
          spec:
            zone:
              name: edited-zone # name of the zone of the host
            hostname:
              name: use-custom-hostname-for-remote-agent # (optional) give your remote agent pod a custom hostname
            agent:
              key: replace-key # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io
              endpointPort: "443"
              env: {}
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.
          ```

          ## Uninstalling

          In order to uninstall the Instana agent, simply remove the custom resource with `kubectl delete -f instana-agent.yaml`.

          ## Source Code

          The Instana agent operator is an open source project hosted on [https://github.com/instana/instana-agent-operator](https://github.com/instana/instana-agent-operator/).
        displayName: Instana Agent Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - APM
        - Instana
        links:
        - name: GitHub Repository
          url: https://github.com/instana/instana-agent-operator
        maintainers:
        - email: support@instana.com
          name: Instana
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: Instana
        relatedImages:
        - icr.io/instana/instana-agent-operator@sha256:ab8b35f0b45c68f69210d1b17c5dd609799f6bb0bc9f12d0d4e5302c910607a1
        - icr.io/instana/agent@sha256:ed71e6c3e5bdb155853a2db8c916460e6a867fb02627e3a332186188cbbd98dd
        version: 2.1.36
      entries:
      - name: instana-agent-operator.v2.1.36
        version: 2.1.36
      - name: instana-agent-operator.v2.1.35
        version: 2.1.35
      - name: instana-agent-operator.v2.1.34
        version: 2.1.34
      - name: instana-agent-operator.v2.1.33
        version: 2.1.33
      - name: instana-agent-operator.v2.1.32
        version: 2.1.32
      - name: instana-agent-operator.v2.1.31
        version: 2.1.31
      - name: instana-agent-operator.v2.1.30
        version: 2.1.30
      - name: instana-agent-operator.v2.1.28
        version: 2.1.28
      - name: instana-agent-operator.v2.1.27
        version: 2.1.27
      - name: instana-agent-operator.v2.1.24
        version: 2.1.24
      - name: instana-agent-operator.v2.1.23
        version: 2.1.23
      - name: instana-agent-operator.v2.1.22
        version: 2.1.22
      - name: instana-agent-operator.v2.1.21
        version: 2.1.21
      - name: instana-agent-operator.v2.1.20
        version: 2.1.20
      - name: instana-agent-operator.v2.1.19
        version: 2.1.19
      - name: instana-agent-operator.v2.1.18
        version: 2.1.18
      - name: instana-agent-operator.v2.1.17
        version: 2.1.17
      - name: instana-agent-operator.v2.1.16
        version: 2.1.16
      - name: instana-agent-operator.v2.1.15
        version: 2.1.15
      - name: instana-agent-operator.v2.1.14
        version: 2.1.14
      - name: instana-agent-operator.v2.1.13
        version: 2.1.13
      - name: instana-agent-operator.v2.1.12
        version: 2.1.12
      - name: instana-agent-operator.v2.1.11
        version: 2.1.11
      - name: instana-agent-operator.v2.1.10
        version: 2.1.10
      - name: instana-agent-operator.v2.1.9
        version: 2.1.9
      - name: instana-agent-operator.v2.1.8
        version: 2.1.8
      - name: instana-agent-operator.v2.1.7
        version: 2.1.7
      - name: instana-agent-operator.v2.1.6
        version: 2.1.6
      - name: instana-agent-operator.v2.1.5
        version: 2.1.5
      - name: instana-agent-operator.v2.1.4
        version: 2.1.4
      - name: instana-agent-operator.v2.1.3
        version: 2.1.3
      - name: instana-agent-operator.v2.1.2
        version: 2.1.2
      - name: instana-agent-operator.v2.1.1
        version: 2.1.1
      - name: instana-agent-operator.v2.1.0
        version: 2.1.0
      - name: instana-agent-operator.v2.0.32
        version: 2.0.32
      - name: instana-agent-operator.v2.0.31
        version: 2.0.31
      - name: instana-agent-operator-rhmp.v2.0.30
        version: 2.0.30
      - name: instana-agent-operator.v2.0.27
        version: 2.0.27
      - name: instana-agent-operator.v2.0.19
        version: 2.0.19
      - name: instana-agent-operator.v2.0.18
        version: 2.0.18
      - name: instana-agent-operator.v2.0.17
        version: 2.0.17
      - name: instana-agent-operator.v2.0.13
        version: 2.0.13
      - name: instana-agent-operator.v2.0.12
        version: 2.0.12
      - name: instana-agent-operator.v2.0.11
        version: 2.0.11
      - name: instana-agent-operator.v2.0.10
        version: 2.0.10
      - name: instana-agent-operator.v2.0.9
        version: 2.0.9
      - name: instana-agent-operator.v2.0.8
        version: 2.0.8
      - name: instana-agent-operator.v2.0.7
        version: 2.0.7
      - name: instana-agent-operator.v2.0.5
        version: 2.0.5
      - name: instana-agent-operator.v2.0.4
        version: 2.0.4
      name: stable
    defaultChannel: stable
    packageName: instana-agent-operator-rhmp
    provider:
      name: Instana
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: pelorus-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pelorus-operator.v0.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.pelorus.dora-metrics.io/v1alpha1",
                "kind": "Pelorus",
                "metadata": {
                  "name": "pelorus-sample"
                },
                "spec": {
                  "exporters": {
                    "global": {},
                    "instances": [
                      {
                        "app_name": "deploytime-exporter",
                        "exporter_type": "deploytime"
                      },
                      {
                        "app_name": "failuretime-exporter",
                        "enabled": false,
                        "exporter_type": "failure"
                      },
                      {
                        "app_name": "committime-exporter",
                        "exporter_type": "committime"
                      },
                      {
                        "app_name": "webhook-exporter",
                        "exporter_type": "webhook"
                      }
                    ]
                  },
                  "federate_openshift_monitoring": {
                    "enabled": false
                  },
                  "openshift_prometheus_basic_auth_pass": "changeme",
                  "openshift_prometheus_htpasswd_auth": "internal:{SHA}+pvrmeQCmtWmYVOZ57uuITVghrM=",
                  "prometheus_retention": "1y",
                  "prometheus_retention_size": "1GB",
                  "prometheus_storage": false,
                  "prometheus_storage_pvc_capacity": "2Gi",
                  "prometheus_storage_pvc_storageclass": "gp2"
                }
              }
            ]
          capabilities: Basic Install
          categories: |
            Modernization & Migration,Developer Tools,Monitoring,Integration & Delivery
          containerImage: quay.io/pelorus/pelorus-operator:0.0.9
          createdAt: "2024-05-27T20:04:01Z"
          description: |
            Tool that helps IT organizations measure their impact on the overall performance of their organization
          operatorframework.io/suggested-namespace: pelorus
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/dora-metrics/pelorus/
          support: Pelorus Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pelorus is the Schema for pelorus API to adopt instance to
              the requested workflow
            displayName: Pelorus
            kind: Pelorus
            name: peloruses.charts.pelorus.dora-metrics.io
            version: v1alpha1
        description: |
          Pelorus is a tool that helps IT organizations measure their impact on the overall performance of their organization. It does this by gathering metrics about team and organizational behaviors over time in some key areas of IT that have been shown to impact the value they deliver to the organization as a whole. Some of the key outcomes Pelorus can focus on are:

          - Software Delivery Performance
          - Product Quality and Sustainability
          - Customer experience

          For more background on the project you can read [@trevorquinn](https://github.com/trevorquinn)'s blog post on [Metrics Driven Transformation](https://www.openshift.com/blog/exploring-a-metrics-driven-approach-to-transformation).

          ## Software Delivery Performance as an outcome

          Currently, Pelorus functionality can capture proven metrics that measure Software Delivery Performance -- a significant outcome that IT organizations aim to deliver.

          Pelorus is a Grafana dashboard that can easily be deployed to an OpenShift cluster, and provides an organizational-level view of the [four critical measures of software delivery performance](https://blog.openshift.com/exploring-a-metrics-driven-approach-to-transformation/).

          ## Software Delivery Metrics Dashboard

          A short video describing each of metrics provided by Pelorus is available [here](https://www.youtube.com/watch?v=7-iB_KhUaQg).

          ## Demo

          [YouTube Video](https://www.youtube.com/watch?v=VPCOIfDcgso) with the Pelorus in action recorded during online Konveyor Community event.

          ## Documentation

          Pelorus documentation is available at [pelorus.readthedocs.io](https://pelorus.readthedocs.io/en/latest/).

          ## Contributing to Pelorus

          If you are interested in contributing to the Pelorus project, please review our Contribution guide which can be found in the [contribution guide](https://github.com/dora-metrics/pelorus/blob/master/CONTRIBUTING.md)

          ## Statement of Support

          Our support policy can be found in the [Upstream Support statement](https://github.com/dora-metrics/pelorus/blob/master/docs/UpstreamSupport.md)

          ## Code of Conduct
          Refer to dora-metrics's Code of Conduct [here](https://github.com/dora-metrics/pelorus/blob/master/CODE_OF_CONDUCT.md).

          ## License

          This repository is licensed under the terms of [Apache-2.0 License](https://raw.githubusercontent.com/dora-metrics/pelorus/master/LICENSE).
        displayName: Pelorus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pelorus-operator
        - dora
        - dora metrics
        - pelorus
        - metrics
        - transformation
        - devops
        links:
        - name: Pelorus Docs
          url: https://pelorus.readthedocs.io/en/latest/
        - name: Pelorus Getting Started
          url: https://pelorus.readthedocs.io/en/latest/GettingStarted/Overview/
        - name: Pelorus GIT repository
          url: https://github.com/dora-metrics/pelorus/
        maintainers:
        - email: kgranger@redhat.com
          name: Kevin Granger
        - email: whayutin@redhat.com
          name: Wesley Hayutin
        - email: mpryc@redhat.com
          name: Michal Pryc
        - email: msouzaol@redhat.com
          name: Mateus Souza Oliveira
        - email: esauer@redhat.com
          name: Eric Sauer
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - quay.io/pelorus/pelorus-operator:0.0.9
        version: 0.0.9
      entries:
      - name: pelorus-operator.v0.0.9
        version: 0.0.9
      - name: pelorus-operator.v0.0.8
        version: 0.0.8
      - name: pelorus-operator.v0.0.7
        version: 0.0.7
      - name: pelorus-operator.v0.0.6
        version: 0.0.6
      - name: pelorus-operator.v0.0.5
        version: 0.0.5
      - name: pelorus-operator.v0.0.4
        version: 0.0.4
      - name: pelorus-operator.v0.0.3
        version: 0.0.3
      - name: pelorus-operator.v0.0.2
        version: 0.0.2
      - name: pelorus-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: pelorus-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CodeFlare
      provider-url: https://github.com/project-codeflare
    name: codeflare-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: codeflare-operator.v1.16.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: quay.io/project-codeflare/codeflare-operator:v1.16.0
          createdAt: "2025-06-16T10:53:51Z"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/project-codeflare/codeflare-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AppWrapper
            name: appwrappers.workload.codeflare.dev
            version: v1beta2
        description: CodeFlare allows you to scale complex pipelines anywhere
        displayName: CodeFlare Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Pipelines
        - Scaling
        - App
        - Jobs
        links:
        - name: Codeflare Operator
          url: https://codeflare.dev/
        maintainers:
        - email: aasthana@redhat.com
          name: Anish Asthana
        - email: kpostlet@redhat.com
          name: Kevin Postlethwait
        - email: meyceoz@redhat.com
          name: Mustafa Eyceoz
        maturity: alpha
        provider:
          name: CodeFlare
          url: https://github.com/project-codeflare
        relatedImages:
        - quay.io/project-codeflare/codeflare-operator:v1.16.0
        version: 1.16.0
      entries:
      - name: codeflare-operator.v1.16.0
        version: 1.16.0
      - name: codeflare-operator.v1.15.0
        version: 1.15.0
      - name: codeflare-operator.v1.14.0
        version: 1.14.0
      - name: codeflare-operator.v1.13.0
        version: 1.13.0
      - name: codeflare-operator.v1.12.0
        version: 1.12.0
      - name: codeflare-operator.v1.11.0
        version: 1.11.0
      - name: codeflare-operator.v1.10.0
        version: 1.10.0
      - name: codeflare-operator.v1.9.0
        version: 1.9.0
      - name: codeflare-operator.v1.8.0
        version: 1.8.0
      - name: codeflare-operator.v1.7.0
        version: 1.7.0
      - name: codeflare-operator.v1.6.0
        version: 1.6.0
      - name: codeflare-operator.v1.5.0
        version: 1.5.0
      - name: codeflare-operator.v1.4.4
        version: 1.4.4
      - name: codeflare-operator.v1.4.3
        version: 1.4.3
      - name: codeflare-operator.v1.4.2
        version: 1.4.2
      - name: codeflare-operator.v1.4.1
        version: 1.4.1
      - name: codeflare-operator.v1.4.0
        version: 1.4.0
      - name: codeflare-operator.v1.3.1
        version: 1.3.1
      - name: codeflare-operator.v1.3.0
        version: 1.3.0
      - name: codeflare-operator.v1.2.0
        version: 1.2.0
      - name: codeflare-operator.v1.1.0
        version: 1.1.0
      - name: codeflare-operator.v1.0.1
        version: 1.0.1
      - name: codeflare-operator.v1.0.0
        version: 1.0.0
      - name: codeflare-operator.v1.0.0-rc.4
        version: 1.0.0-rc.4
      - name: codeflare-operator.v1.0.0-rc.3
        version: 1.0.0-rc.3
      - name: codeflare-operator.v1.0.0-rc.2
        version: 1.0.0-rc.2
      - name: codeflare-operator.v1.0.0-rc.1
        version: 1.0.0-rc.1
      - name: codeflare-operator.v0.2.3
        version: 0.2.3
      - name: codeflare-operator.v0.2.2
        version: 0.2.2
      - name: codeflare-operator.v0.2.1
        version: 0.2.1
      - name: codeflare-operator.v0.2.0
        version: 0.2.0
      - name: codeflare-operator.v0.1.0
        version: 0.1.0
      - name: codeflare-operator.v0.0.6
        version: 0.0.6
      - name: codeflare-operator.v0.0.5
        version: 0.0.5
      - name: codeflare-operator.v0.0.4
        version: 0.0.4
      - name: codeflare-operator.v0.0.3
        version: 0.0.3
      - name: codeflare-operator.v0.0.2
        version: 0.0.2
      - name: codeflare-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: codeflare-operator
    provider:
      name: CodeFlare
      url: https://github.com/project-codeflare
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: opdev.github.io
    name: airflow-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airflow-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "workflow.apache.org/v1alpha1",
                "kind": "AirFlow",
                "metadata": {
                  "name": "airflow-helm"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/airflow-helm-operator:0.0.1
          createdAt: "2023-06-28T16:57:21Z"
          description: An experimental operator that installs Apache Airflow.
          operatorframework.io/suggested-namespace: airflow-helm
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys Apache Airflow based on the official helm chart in
              openshift
            kind: AirFlow
            name: airflows.workflow.apache.org
            version: v1alpha1
        description: Deploys Apache Airflow based on the official helm chart with
          openshift specific modifications applied by default
        displayName: Airflow Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        - data
        - batch
        - scheduler
        - data workflow
        links:
        - name: Airflow Helm Operator
          url: https://airflow-helm-operator.domain
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: opdev.github.io
        relatedImages:
        - quay.io/opdev/airflow-helm-operator:0.0.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 0.0.1
      entries:
      - name: airflow-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: airflow-helm-operator
    provider:
      name: opdev
      url: opdev.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-sso
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-sso.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Sso",
                "metadata": {
                  "name": "sso-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "kcos-keycloak": {
                    "auth": {
                      "baseURL": "https://keycloak.kcos-sso.svc.cluster.local:8443",
                      "enabled": true,
                      "sessionCookie": "AT"
                    },
                    "authproxy": {
                      "image": "prod.harbor.keysight.digital/loadcore/kcos-auth-proxy@sha256:a777ee5116280fbeb1e1342445fef5135f9ca61e983a3e2821f9ca4b2038b044",
                      "initContainer": {
                        "image": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf"
                      }
                    },
                    "crsCleanupJob": {
                      "clientExec": "oc",
                      "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "realm": {
                      "accessTokenLifespan": 266400,
                      "accessTokenLifespanForImplicitFlow": 266400,
                      "admin": {
                        "password": "admin"
                      },
                      "offlineSessionIdleTimeout": 2592000,
                      "offlineSessionMaxLifespan": 5184000,
                      "registrationAllowed": true,
                      "ssoSessionIdleTimeout": 266400,
                      "ssoSessionMaxLifespan": 266400
                    }
                  },
                  "subcharts": {
                    "enable": {
                      "kcos-keycloak": true,
                      "keycloak-operator": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:29:57Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Sso
            name: ssoes.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS SSO Operator
        displayName: Keysight KCOS SSO Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Sso
          url: https://keysight-kcos-sso.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-sso@sha256:298417fac2fc053bdc352f77b59b14e39eadd66c871cfb547304a517489f294e
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - prod.harbor.keysight.digital/loadcore/kcos-auth-proxy@sha256:a777ee5116280fbeb1e1342445fef5135f9ca61e983a3e2821f9ca4b2038b044
        version: 0.1.21
      entries:
      - name: keysight-kcos-sso.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-sso
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-cpd-wspipelines
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-wspipelines.v10.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wspipelines.cpd.ibm.com/v1beta1",
                "kind": "WSPipelines",
                "metadata": {
                  "name": "wspipelines-cr",
                  "labels": {
                    "app.kubernetes.io/name": "wspipelines",
                    "app.kubernetes.io/instance": "wspipelines.wspipelines.cpd.ibm.com",
                    "app.kubernetes.io/managed-by": "wspipelines-operator",
                    "wspipelinesBuildNumber": "2017"
                }
                },
                "spec": {
                  "version": "5.1.3",
                  "scaleConfig": "small",
                  "license": {
                    "accept": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:f9f616d17f816468957b69bbb09e6c92d678acb4ba7a40f1cb493d08fb98deb8
          createdAt: "2025-02-11"
          description: WS Pipelines CSV file
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-wspipelines
          olm.skipRange: '>=1.0.0 <10.3.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM watsonx.ai"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation For additional details regarding install parameters
              check: https://ibm.biz/BdaaL9 or https://ibm.biz/BdaaLC. License By
              installing this product you accept the license terms https://ibm.biz/BdaaL9
              or https://ibm.biz/BdaaLC.'
            kind: WSPipelines
            name: wspipelines.wspipelines.cpd.ibm.com
            version: v1beta1
        description: |-
          # Name
          IBM&reg; Watson Pipelines
          # Introduction
          ## Summary
          The Watson Pipelines editor provides a graphical interface for orchestrating an end-to-end flow of assets from creation through deployment. It enables to automate an end-to-end flow to prepare data, then create, train, deploy, monitor and update machine learning models and Python scripts. To design a pipeline you drag nodes onto the canvas, specify objects and parameters, then run and monitor the pipeline.
          ## Features
          Watson Pipelines are built off of Kubeflow pipelines on Tekton runtime and are fully integrated into the Watson Studio platform, allowing users to combine tools including:
          - Notebooks
          - Data refinery flows
          - AutoAI experiments
          - Web service / online deployments
          - Batch deployments
          - Import and export of Project / Space assets
          into repeatable and automated pipelines.
          # Details
          ## Prerequisites
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-pre-installation-tasks)
          ### Resources Required
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=planning-system-requirements)
          # Installing
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing)
          ## Configuration
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing)
          ## Storage
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=planning-storage-considerations)
          ## Limitations
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=issues-watson-studio)
          ## Documentation
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=overview)
          ## PodSecurityPolicy Requirements
          Custom PodSecurityPolicy definition:
          ```
          Not Applicable
          ```
          ## SecurityContextConstraints Requirements
          Custom SecurityContextConstraints definition:
          ```
          apiVersion: security.openshift.io/v1
          kind: SecurityContextConstraints
          allowHostDirVolumePlugin: false
          allowHostIPC: false
          allowHostNetwork: false
          allowHostPID: false
          allowHostPorts: false
          allowPrivilegeEscalation: false
          allowPrivilegedContainer: false
          allowedCapabilities: null
          defaultAddCapabilities: null
          fsGroup:
            type: MustRunAs
          metadata:
            annotations:
              include.release.openshift.io/self-managed-high-availability: "true"
              kubernetes.io/description: restrict denies access to all host features and requires pods to be run with a UID, and SELinux context that are allocated to the namespace.  This is the most restrictive SCC and it is used by default for authenticated users.
              release.openshift.io/create-only: "true"
            name: restrict
          priority: null
          readOnlyRootFilesystem: false
          requiredDropCapabilities:
          - ALL
          runAsUser:
            type: MustRunAsRange
          seLinuxContext:
            type: MustRunAs
          supplementalGroups:
            type: RunAsAny
          users: []
          volumes:
          - configMap
          - downwardAPI
          - emptyDir
          - persistentVolumeClaim
          - projected
          - secret
          ```
        displayName: Watson Pipelines
        installModes:
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: OwnNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - wspipelines
        - WSPipelines
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: v1
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:a477d47eaf8f8e569601d3c23c1def66f31c098ac3f5a1ae350618c460917dd8
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:0164d8fbea591bbd9d67d84603162c3781b08551eddcc2a8a842aafc828af18c
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:164f6063f3a50a36e28a240ae7c060c4f90266b7ed799bfdc075f90628bea7ff
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:ebcfc8941330da6991480dc18ac34d6df4fe879e562ee537e5c36df73f4bc536
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a24f4cb635e8b5759a5e2bda20ddefc6d5ea9b16fb1081592885c54aefb5480c
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:29cd5c9cd60bbcfe211337786616622b911e2319db9870eb9d38d4f630c03e75
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:f0c68715f78bd20c352c97dd48ac5a912cad58e19a00159b5e6878f6beeb00fb
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:a2949c3868c2c91bf35e50c64b51fcc24372986744b55520e21e9ce31fde59f0
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-bundle@sha256:842f2ac568db4db79cf0450b413b4b09b5f2180ef3541385b46fb650ac7066c0
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:f9f616d17f816468957b69bbb09e6c92d678acb4ba7a40f1cb493d08fb98deb8
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:61155acdb8982648ae5748be9f5944ba913db1590ec3e773206d6fbb1d8952cb
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:8d7625a0ba656425f601b6ec9a398a6d30511fe1de119cf0e29f680c704ea225
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:e4888fadfadb3462e3a9ad0395a526a2631872f70caf291dc18b60269ded199f
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:48b7cfa31b9a185176c20afc119e37346a6f9380d74000edf6c729b4677ca44e
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:19bb32e0389361e8314e627bcad2e9bd93984289485cf632466834353d9c3449
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:e29885c1ee083a73b952b0de8d1cd675b2b2830abf44d69a8577f56972a601fb
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:00f8bcee4ba7e0c5cf5f0c52e9e7707df367041c9f4dd5508dbefe0dbdba8877
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:99d0b3192e6e46167423505f48655e7d49af1a461ae00b95c9c7c9d4209e5b65
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:25a4bb9cea1c3bdbeeeb205622512bcba98163f9629eba35625756346ce9d652
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:6fc976bc622c2165dd8bf93c9244e1c29b7c973a4da7a5ce4d0a097a1c1ea4fb
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:ad2215adfb8dd2dc600bf5097b7f88c257d5fbaf541409c01b36cdb67c130236
        version: 10.3.0
      entries:
      - name: ibm-cpd-wspipelines.v10.3.0
        version: 10.3.0
      name: v10.3
    defaultChannel: v10.3
    packageName: ibm-cpd-wspipelines
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-enterprise
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-enterprise.v1.33.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "my-replica-set"
                },
                "spec": {
                  "credentials": "my-credentials",
                  "members": 3,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "sample-sharded-cluster"
                },
                "spec": {
                  "configServerCount": 3,
                  "credentials": "my-credentials",
                  "mongodsPerShardCount": 3,
                  "mongosCount": 2,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "shardCount": 1,
                  "type": "ShardedCluster",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBMultiCluster",
                "metadata": {
                  "name": "multi-replica-set"
                },
                "spec": {
                  "clusterSpecList": [
                    {
                      "clusterName": "e2e.cluster1.mongokubernetes.com",
                      "members": 2
                    },
                    {
                      "clusterName": "e2e.cluster2.mongokubernetes.com",
                      "members": 1
                    },
                    {
                      "clusterName": "e2e.cluster3.mongokubernetes.com",
                      "members": 2
                    }
                  ],
                  "credentials": "my-credentials",
                  "duplicateServiceObjects": false,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBOpsManager",
                "metadata": {
                  "name": "ops-manager"
                },
                "spec": {
                  "adminCredentials": "ops-manager-admin",
                  "applicationDatabase": {
                    "members": 3,
                    "podSpec": {
                      "cpu": 1
                    }
                  },
                  "configuration": {
                    "mms.fromEmailAddr": "admin@thecompany.com"
                  },
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "version": "6.0.3"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBUser",
                "metadata": {
                  "name": "my-replica-set-x509-user"
                },
                "spec": {
                  "db": "$external",
                  "mongodbResourceRef": {
                    "name": "my-replica-set"
                  },
                  "roles": [
                    {
                      "db": "admin",
                      "name": "dbOwner"
                    }
                  ],
                  "username": "CN=my-replica-set-x509-user,OU=cloud,O=MongoDB,L=New York,ST=New York,C=US"
                },
                "status": {
                  "phase": "Updated"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/mongodb/mongodb-enterprise-operator-ubi@sha256:6d394a5247cc346c4421bd9b37948c0378970dd3f4fc5d6eaac681f4c9655957
          createdAt: "2025-04-15T11:32:47Z"
          description: The MongoDB Enterprise Kubernetes Operator enables easy deploys
            of MongoDB into Kubernetes clusters, using our management, monitoring
            and backup platforms, Ops Manager and Cloud Manager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mongodb/mongodb-enterprise-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MongoDB Deployment
            displayName: MongoDB Deployment
            kind: MongoDB
            name: mongodb.mongodb.com
            version: v1
          - description: MongoDB Multi Deployment
            displayName: MongoDB Multi Deployment
            kind: MongoDBMultiCluster
            name: mongodbmulticluster.mongodb.com
            version: v1
          - description: MongoDB x509 User
            displayName: MongoDB User
            kind: MongoDBUser
            name: mongodbusers.mongodb.com
            version: v1
          - description: MongoDB Ops Manager
            displayName: MongoDB Ops Manager
            kind: MongoDBOpsManager
            name: opsmanagers.mongodb.com
            version: v1
        description: |
          The MongoDB Enterprise Kubernetes Operator enables easy deploys of MongoDB
          into Kubernetes clusters, using our management, monitoring and backup
          platforms, Ops Manager and Cloud Manager.

          ## Before You Start

          To start using the operator you''ll need an account in MongoDB Cloud Manager or
          a MongoDB Ops Manager deployment.

          * [Create a Secret with your OpsManager API key](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-operator-credentials/#procedure)

          * [Create a ConfigMap with your OpsManager project ID and URL](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-project-using-configmap/)

          By installing this integration, you will be able to deploy MongoDB instances
          with a single simple command.

          ## Required Parameters

          * `opsManager` or `cloudManager` - Enter the name of the ConfigMap containing project information
          * `credentials` - Enter the name of the Secret containing your OpsManager credentials
          * `type` - Enter MongoDB Deployment Types ("Standalone", "ReplicaSet", "ShardedCluster"

          ## Supported MongoDB Deployment Types ##

          * Standalone: An instance of mongod that is running as a single server and
          not as part of a replica set, this is, it does not do any kind of
          replication.

          * Replica Set: A replica set in MongoDB is a group of mongod processes that
          maintain the same data set. Replica sets provide redundancy and high
          availability, and are the basis for all production deployments. This section
          introduces replication in MongoDB as well as the components and architecture
          of replica sets. The section also provides tutorials for common tasks
          related to replica sets.

          * Sharded Cluster: The set of nodes comprising a sharded MongoDB deployment.
          A sharded cluster consists of config servers, shards, and one or more mongos
          routing processes. Sharding is a A database architecture that partitions
          data by key ranges and distributes the data among two or more database
          instances. Sharding enables horizontal scaling.

          ## Requirements for deploying MongoDB OpsManager

          * In order to deploy resources of type MongoDB OpsManager, you will need to
          create a secret containing the [credentials](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/plan-om-resource/#om-rsrc-prereqs)
          for the Global Owner user

          ## Security ##

          The operator can enable TLS for all traffic between servers and also between
          clients and servers. Before enabling `security.tls.enabled` to `true` you
          should create your certificates.  or you can leave the operator to create all
          the certificates for you. The operator ability to create certs is been
          deprecated due to Kubernetes API changes.

          For more information, please read the official MongoDB
          Kubernetes Operator  [docs](https://docs.mongodb.com/kubernetes-operator/stable/).
        displayName: MongoDB Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mongodb
        - databases
        - nosql
        links:
        - name: Documentation
          url: https://docs.opsmanager.mongodb.com/current/tutorial/install-k8s-operator/index.html
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: stable
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-enterprise-server@sha256:fb8397cbcdbf7c33f6ad142492792771127f0d9a8c46388f350b6927405ee590
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa8683500e8d3632c4588bbcdddb4e3186a01f48aefe1b6d95f8860dae8b6c76
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:7d93aee462c60cacc4e5e66729f5131c5c0fa434e6e6400c8e2ccac4bd5e20ba
        - quay.io/mongodb/mongodb-agent-ubi@sha256:85f883076bc8edc516800704e2290c54b1bee14a9a533df84928c7ce1ae807d2
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0b2d01d543d38feddcc9d351e711df33a0a5187b22d95cdb926f6015ee1093bd
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ca2aaf57d70a0b9d743a6533c4e1b88324056197453de540d3352e679a3c9c92
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7e5361c2a1604dec3fc671d629806fe6ddc6df524b69a441b86933a4b30093
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b2bd8eb7f8b96adb874468ff1b0ebae2db4ebe490b6e1da4c9a6c5108f813ff5
        - quay.io/mongodb/mongodb-agent-ubi@sha256:766e6cbeec27f995211aa509ff7f118b6945f3e728141f28f9c4ea0dd2297af5
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3add8b68bfbaa42596fe5dbbee4623ba5ff1a826b18bb71301525adf574dcb94
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0fde236fe26d5b20210d6f246aadacd7845874712def516cbe2eec11d79d5181
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:dfec259209995bb040a1eff1fcafb1d6c3fdd145129f12c04a56ec74834b80a6
        - quay.io/mongodb/mongodb-enterprise-server@sha256:110a560f5a0ebe1fde1b2c20ae5ff5638682229d1c8656b993bdaf7b71eb94bc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:19925e1297ab652db59c35f53e7b7c8b2b277f8ecdc4215682961aaf2b3e924e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa300abf739e26350a258f643fa8e57aedacd041055c7b923bd22b21ef554cdb
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:83af86f8098e3c81b6e0879eabd582592ef6cbb1d6592f5e175c0794b9df2a0e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5421e00c0935b7a9ced51a11e935773ef5df4d259f3e1b1930fbf2f1b9d46586
        - quay.io/mongodb/mongodb-agent-ubi@sha256:38175af108178d89e4d3359ee2754534e5e9b24e843f12711fc14fe1dead516a
        - quay.io/mongodb/mongodb-agent-ubi@sha256:6ebe232ed0ea87045792e578ecb11dd46411333ad0ea61383c247704b4a0efe9
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:d3248506b6bd6b01c5973ab7bdac51ee41efa0ec8dd6211be4cc0b76a53a40c1
        - quay.io/mongodb/mongodb-enterprise-init-database-ubi@sha256:ab12a330e6b40735ee516168073887a4f37b58694bd10026a451eb9972e55500
        - quay.io/mongodb/mongodb-agent-ubi@sha256:5119a1ceeae6440dbe88b89550a7ae18de8e85c9a2059f467ad66adf7ec5d75f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:11ab324698f919f84a5dab5a964d93fe63f98a48146bb3b86974cba7606d30ec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c752a7f0726d21d55f204a359b85fa29cd727a5fe571b37289ae3ea52d6206b0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8f8735d377d214f64961381b5b1ed7cd3c306e6354dbc6b45c51134a25444b93
        - quay.io/mongodb/mongodb-enterprise-server@sha256:db63f6d93e03a3e644ef6ef4846ed7d24576c688ffb584226c78e1d1fba38878
        - quay.io/mongodb/mongodb-agent-ubi@sha256:575a43be4c667f3249f46f1fdcc06dd2e68cacc9eed989da0f6edd7e214d51b7
        - quay.io/mongodb/mongodb-agent-ubi@sha256:23d962d038888c6aba5540c8f5d81d7e86c4ccfaa07b9e159825d98676609f8e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0ee80032abe873b330537d4b64511ee76c5efb512af95287f28ddeffab13bb6d
        - quay.io/mongodb/mongodb-agent-ubi@sha256:8c11348eac180f5faa113db2804817302ec78847472e1891b063744adba4aa60
        - quay.io/mongodb/mongodb-agent-ubi@sha256:6d007271faed24bdf06eb64f2b54e5cdc3e11ce71840a7f586dabe2ee02c640e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:83db493ac6158d654d3ce55619f312d53cfbac258e0faf22e9110ca004e174bd
        - quay.io/mongodb/mongodb-agent-ubi@sha256:2395fe9d38a1a9160f7cb2f400b776ea3fc2e72187d23b6dee7fd1e4fb9ca721
        - quay.io/mongodb/mongodb-agent-ubi@sha256:66b49d9dce414dcd657c3550b26f71964b82edfb67a41a3ce8d2386b04ef07e8
        - quay.io/mongodb/mongodb-agent-ubi@sha256:d927afa9d3d569d943518f0a684e19afd4adb0875f55df7c904b9dd019589e5e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:c96ad084edbdc6a6d907f57795fa38f39199b730328dbc64997d56ea50d55dfc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:67238129df09d33ee4746ff1acb5515d699a29d9f42350eae3f3f69ccb7fbf54
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:f64d99a77d6a6b9ed10de9a28b1b8b8155414eca2e561f3fdb04bd66914fb3c5
        - quay.io/mongodb/mongodb-enterprise-server@sha256:15e31285c8949a4666f7702df53416b0905d76945a28575a6028a876cc8a6f0f
        - quay.io/mongodb/mongodb-agent-ubi@sha256:31cfefd2660da3efe1b6d63e39090464c28ae006d6ed1dffb2189a09183b9c61
        - quay.io/mongodb/mongodb-agent-ubi@sha256:97bb6bb5ecd5a802e8a91e69aa10d8bde0f648b106dc5591d49ee94209822daa
        - quay.io/mongodb/mongodb-agent-ubi@sha256:f5f3b14ee6533b21d89d9ed6d4af47dbdee9e071f909dc686dfcdf6742179e46
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:985f770667e6c4c083e0c4bb79a948b83efa0cdc4ed3012b1dbf72fe48b38986
        - quay.io/mongodb/mongodb-agent-ubi@sha256:9d1f21df008b14ca721c9ed0ff905af23e1a56b64cd4eddab4df1e8e9b63df8c
        - quay.io/mongodb/mongodb-enterprise-server@sha256:523566f44cbb6bbf0cd12b1adae939a883f2254c78dc4b8c9d0098581337b1db
        - quay.io/mongodb/mongodb-agent-ubi@sha256:2dbaaece4dd48087370e43d9bcf6b0a0312f86cb5dfe916e8725323bd5401992
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9a3a6496be0d2d446e6d654eb56386c7726e0cd2d7232f6dcec579923edf0014
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1d418c65d82ccad205b7889be4ba161d3cbfa8b1fb75b21c8fcf02de38fad9ca
        - quay.io/mongodb/mongodb-enterprise-server@sha256:172e338a43df01ee5408d8d1e3cb038d2c55e77761c1a2ac6ed37959d1525140
        - quay.io/mongodb/mongodb-enterprise-server@sha256:655896bac9e82de8e9729d59a36c3a31f66235f22b93449260a2ccb7709de084
        - quay.io/mongodb/mongodb-agent-ubi@sha256:695074fbbfcadf454230b80fbf1094260e631f6108cefb2050473edaee2a217a
        - quay.io/mongodb/mongodb-agent-ubi@sha256:b175972792aa1be268d5fb3cf3ad0156f4aecffa323e1b473360fc2a3a8d8891
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d876f57ea45582e84748b02459748e440af443a7eb0882585638973409c08f6a
        - quay.io/mongodb/mongodb-agent-ubi@sha256:411f85772e7a4b2c76e21d26d320b7743612080bf0cb4c7900d9a626cfeea33b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a48a4d5c275fd669a27469649fb39ea38c728dbcf9f0a9da4e7cf27a2ce912ed
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:2be0c6c89898b18618932b6a6303aeefdde60f48ec37d5c1fca5b8d04a00fd78
        - quay.io/mongodb/mongodb-agent-ubi@sha256:bd7edb8d0931dd952524f1f0ef99f61b0f69385d9e61efd034a4f45fd315a40b
        - quay.io/mongodb/mongodb-agent-ubi@sha256:ee150e969c0bb786969c7e0a45fb1812766b752aee539072e28648817826433b
        - quay.io/mongodb/mongodb-agent-ubi@sha256:7144ac5e5038e92a6d142bf39eba46554c1b5fafeb8ea59f0390896a0d35ed17
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:2127d7f2684ae36b123422f55237055aad02fdb71c4c21a5b8f0b0f6d8baedfc
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:c9a211a2c31b97eed3a25a814198547f2aa3470df93d661ca3700b828100d8de
        - quay.io/mongodb/mongodb-agent-ubi@sha256:c03e4a7663206f54c5ef39974502816f66d8d64936c19849f0337ea4547e614e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:35b263861d6ea66b541abc190ce1aaf4d4e36f128e5d8ba2a5e9fa8e25d84446
        - quay.io/mongodb/mongodb-enterprise-server@sha256:85c1d669c073caf9cbaac815e4c8ea2ca18b71deefa03bf84cd6d8b8ebd53e12
        - quay.io/mongodb/mongodb-enterprise-server@sha256:37af0913742b6de339c7e664b30f35dcef04506529e4a09a9bbbdd57643f7149
        - quay.io/mongodb/mongodb-agent-ubi@sha256:442f5d99d02674099525abaa69daa42cec758ce5554198f978927aa5c49006c8
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f1bae6375f846af6fca7faeeff4aedf9fab41a9d45b515ba74e9fd7336408617
        - quay.io/mongodb/mongodb-enterprise-server@sha256:605dcca06a45e5ea127edb9ff262a1c3ddafdab2f30b6d3a18f73bfb8a6acdb0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ef732e3eea91cc86caf2eabedc0c9115ed235a8d204f33e74bbf090a0b0259fb
        - quay.io/mongodb/mongodb-agent-ubi@sha256:0f8d102780a9d2ce54b7453f9575d0e24b15f014cd9b74816b1e6a4090b979b2
        - quay.io/mongodb/mongodb-enterprise-init-appdb-ubi@sha256:be4295843d92ce0f325b8daaeb3a5490c11772d0348cbd9ea35a6126d038b7fb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:33a30d11969eb21952681108092fe7f916e25c8e1de43e6a4e40e126025a2bec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7aa00dca44559318356627e88440765576dd6706187044edd1f21d0d0dcb0bf6
        - quay.io/mongodb/mongodb-enterprise-server@sha256:df6081325ea2fe01d1b27d0c70005dec8232aaf51a43d91661b722b8a1761263
        - quay.io/mongodb/mongodb-enterprise-server@sha256:dfd4277f0c16ed476ac7de51894739e8c8d13a65ea5024b9a97b3dce0f6c6975
        - quay.io/mongodb/mongodb-enterprise-server@sha256:584255582db31d0bb1f9593feb3b675d934d8750766fc98d3a1b3340feae4995
        - quay.io/mongodb/mongodb-enterprise-database-ubi@sha256:e2ca129f04ad6bf4e0dbb988badb083187725f28bac01bb7080cc80cad9ac316
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7885ab6e88214f9289d8babf5189136dde0dd4c84a15a5481ccd3ef512b2e66f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f2d9a3264a8814f33143b73c56624572fb80afeab184133d3a901713a31c1a87
        - quay.io/mongodb/mongodb-agent-ubi@sha256:03763de5018fd5fa14843444d24adfc5e387605970b531801ef2d0916233a010
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d21c65780cadc9d8668d2c28935d492625d6993f13ba562722200a819281cbcc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b00ac773623e1bf422b47e6d783263d2d08737865abe9abbe106d86d1ea25165
        - quay.io/mongodb/mongodb-enterprise-operator-ubi@sha256:6d394a5247cc346c4421bd9b37948c0378970dd3f4fc5d6eaac681f4c9655957
        - quay.io/mongodb/mongodb-agent-ubi@sha256:b23cfa639c5a4d20381245772dfe396ec915af3a3046df5d9b10016b900431f5
        - quay.io/mongodb/mongodb-enterprise-server@sha256:cd0cdf1b2a58f4c3b514e602cc1d96543410f631825045bbf65bd3fb84113f1f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8445d59d491a6d3890ea9c8b421214bcd0fdcfbc420ed40497a790b79f0bc89e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c54eeb910855594ed4bf5dcbbbcc14aa6f733e97313896a8381fb4be6aeac583
        - quay.io/mongodb/mongodb-agent-ubi@sha256:36c84115c6c645fd7d65d1004d2507ef3592aa1f07176645b1ee8e8b2a4f5f73
        - quay.io/mongodb/mongodb-enterprise-init-ops-manager-ubi@sha256:6ccc8148c266a9037ea744e8d19875146b1c231909f29a8f01c7ce415ca15936
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d49417ea85ee1ddb84429552a5aac87a82c2843904d8c0300a43b657b7a49e0e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:bd868284d9dfcb7c5a20e1160218d0172358384229b556c37e0789af0030eb24
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c07f20f8e7a5479e43cac7c74fc9f18cb4d4a1bd8b6b46f7dcc6f7ec7525fe28
        - quay.io/mongodb/mongodb-enterprise-server@sha256:13186ad45a2ba3ab55ed5c2a008eca494ee3937c8519475379476cc8e6f1d055
        - quay.io/mongodb/mongodb-agent-ubi@sha256:3b57a03dc17a301fda86c8afed1917c10d9b98947680116399c975ece944f5de
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d323c1bcd14118cc537ac282996ea5a58756d20b6b73f1eec0efa8710d246077
        - quay.io/mongodb/mongodb-agent-ubi@sha256:8467b1c2924a19929fc101ef2ff8d7e53cb0115a0cdc2e371dc0733bd0262395
        version: 1.33.0
      entries:
      - name: mongodb-enterprise.v1.33.0
        version: 1.33.0
      - name: mongodb-enterprise.v1.32.0
        version: 1.32.0
      - name: mongodb-enterprise.v1.31.0
        version: 1.31.0
      - name: mongodb-enterprise.v1.30.0
        version: 1.30.0
      - name: mongodb-enterprise.v1.29.0
        version: 1.29.0
      - name: mongodb-enterprise.v1.28.0
        version: 1.28.0
      - name: mongodb-enterprise.v1.27.0
        version: 1.27.0
      - name: mongodb-enterprise.v1.26.0
        version: 1.26.0
      - name: mongodb-enterprise.v1.25.0
        version: 1.25.0
      - name: mongodb-enterprise.v1.24.0
        version: 1.24.0
      - name: mongodb-enterprise.v1.23.0
        version: 1.23.0
      - name: mongodb-enterprise.v1.22.0
        version: 1.22.0
      - name: mongodb-enterprise.v1.21.0
        version: 1.21.0
      - name: mongodb-enterprise.v1.20.1
        version: 1.20.1
      - name: mongodb-enterprise.v1.19.1
        version: 1.19.1
      - name: mongodb-enterprise.v1.18.0
        version: 1.18.0
      - name: mongodb-enterprise.v1.17.2
        version: 1.17.2
      - name: mongodb-enterprise.v1.17.1
        version: 1.17.1
      - name: mongodb-enterprise.v1.17.0
        version: 1.17.0
      - name: mongodb-enterprise.v1.16.4
        version: 1.16.4
      - name: mongodb-enterprise.v1.16.3
        version: 1.16.3
      - name: mongodb-enterprise.v1.16.2
        version: 1.16.2
      - name: mongodb-enterprise.v1.16.1
        version: 1.16.1
      name: stable
    defaultChannel: stable
    packageName: mongodb-enterprise
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: MariaDB Corporation
      provider-url: ""
    name: mariadb-enterprise-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-enterprise-operator.v25.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup"
                },
                "spec": {
                  "args": [
                    "--single-transaction",
                    "--all-databases"
                  ],
                  "logLevel": "info",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "0 0 * * *",
                    "suspend": false
                  },
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "backups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Connection",
                "metadata": {
                  "name": "connection"
                },
                "spec": {
                  "database": "mariadb",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "secretName": "connection",
                  "secretTemplate": {
                    "key": "dsn"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Database",
                "metadata": {
                  "name": "database"
                },
                "spec": {
                  "characterSet": "utf8",
                  "collate": "utf8_general_ci",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "name": "my-logical-database"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "grant"
                },
                "spec": {
                  "database": "*",
                  "grantOption": true,
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "privileges": [
                    "SELECT",
                    "INSERT",
                    "UPDATE"
                  ],
                  "table": "*",
                  "username": "my-user"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "MariaDB",
                "metadata": {
                  "name": "mariadb"
                },
                "spec": {
                  "connection": {
                    "secretName": "mariadb-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "database": "mariadb",
                  "galera": {
                    "enabled": true
                  },
                  "metrics": {
                    "enabled": true,
                    "passwordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "mariadb-metrics"
                    }
                  },
                  "myCnf": "[mariadb]\nbind-address=*\ndefault_storage_engine=InnoDB\nbinlog_format=row\ninnodb_autoinc_lock_mode=2\ninnodb_buffer_pool_size=1024M\nmax_allowed_packet=256M\n",
                  "passwordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-password"
                  },
                  "primaryConnection": {
                    "secretName": "mariadb-conn-primary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "primaryService": {
                    "type": "ClusterIP"
                  },
                  "replicas": 3,
                  "rootPasswordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-root"
                  },
                  "secondaryConnection": {
                    "secretName": "mariadb-conn-secondary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "secondaryService": {
                    "type": "ClusterIP"
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "storage": {
                    "size": "1Gi"
                  },
                  "updateStrategy": {
                    "type": "ReplicasFirstPrimaryLast"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "MaxScale",
                "metadata": {
                  "name": "maxscale"
                },
                "spec": {
                  "admin": {
                    "guiEnabled": true,
                    "port": 8989
                  },
                  "auth": {
                    "adminPasswordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "maxscale-admin"
                    },
                    "adminUsername": "mariadb-enterprise-operator",
                    "generate": true
                  },
                  "config": {
                    "sync": {
                      "database": "mysql",
                      "interval": "5s",
                      "timeout": "10s"
                    }
                  },
                  "connection": {
                    "port": 3306,
                    "secretName": "maxscale-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "guiKubernetesService": {
                    "type": "ClusterIP"
                  },
                  "kubernetesService": {
                    "type": "ClusterIP"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "metrics": {
                    "enabled": true
                  },
                  "monitor": {
                    "cooperativeMonitoring": "majority_of_all",
                    "interval": "2s",
                    "params": {
                      "available_when_donor": "false",
                      "disable_master_failback": "false",
                      "disable_master_role_setting": "false"
                    }
                  },
                  "replicas": 2,
                  "requeueInterval": "10s",
                  "services": [
                    {
                      "listener": {
                        "port": 3306,
                        "protocol": "MariaDBProtocol"
                      },
                      "name": "rw-router",
                      "params": {
                        "transaction_replay": "true",
                        "transaction_replay_attempts": "10",
                        "transaction_replay_timeout": "5s"
                      },
                      "router": "readwritesplit"
                    }
                  ]
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "PhysicalBackup",
                "metadata": {
                  "name": "physicalbackup"
                },
                "spec": {
                  "compression": "bzip2",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "podAffinity": true,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "0 0 * * *",
                    "immediate": true,
                    "suspend": false
                  },
                  "serviceAccountName": "backup",
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "physicalbackups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "region": "us-east-1",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  },
                  "timeout": "1h"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore"
                },
                "spec": {
                  "backupRef": {
                    "name": "backup"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "targetRecoveryTime": "2023-12-19T09:00:00Z"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "SqlJob",
                "metadata": {
                  "name": "sqljob"
                },
                "spec": {
                  "database": "mariadb",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "sql": "CREATE TABLE IF NOT EXISTS orders (\n  id bigint PRIMARY KEY AUTO_INCREMENT,\n  customer_id INT,\n  order_date DATE,\n  total_amount_cents INT\n);\nINSERT INTO orders (customer_id, order_date, total_amount_cents) VALUES\n  (101, '2024-05-15', 5000),\n  (102, '2024-05-14', 7525),\n  (103, '2024-05-13', 10050),\n  (104, '2024-05-12', 2575);",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "user"
                },
                "spec": {
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxUserConnections": 20,
                  "name": "my-user",
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.mariadb.com/mariadb-enterprise-operator@sha256:1effbbfa7632be92c17ef61c18a25f010b187840beec42c31d2410d0209e6985
          createdAt: "2025-08-28T11:35:03Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["MariaDB Enterprise"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: MariaDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API. It is used to define
              backup jobs and its storage.
            displayName: Backup
            kind: Backup
            name: backups.enterprise.mariadb.com
            version: v1alpha1
          - description: Connection is the Schema for the connections API. It is used
              to configure connection strings for the applications connecting to MariaDB.
            displayName: Connection
            kind: Connection
            name: connections.enterprise.mariadb.com
            version: v1alpha1
          - description: Database is the Schema for the databases API. It is used
              to define a logical database as if you were running a 'CREATE DATABASE'
              statement.
            displayName: Database
            kind: Database
            name: databases.enterprise.mariadb.com
            version: v1alpha1
          - description: Grant is the Schema for the grants API. It is used to define
              grants as if you were running a 'GRANT' statement.
            displayName: Grant
            kind: Grant
            name: grants.enterprise.mariadb.com
            version: v1alpha1
          - description: MariaDB is the Schema for the mariadbs API. It is used to
              define MariaDB clusters.
            displayName: Maria DB
            kind: MariaDB
            name: mariadbs.enterprise.mariadb.com
            version: v1alpha1
          - description: MaxScale is the Schema for the maxscales API. It is used
              to define MaxScale clusters.
            displayName: Max Scale
            kind: MaxScale
            name: maxscales.enterprise.mariadb.com
            version: v1alpha1
          - description: PhysicalBackup is the Schema for the physicalbackups API.
              It is used to define physical backup jobs and its storage.
            displayName: Physical Backup
            kind: PhysicalBackup
            name: physicalbackups.enterprise.mariadb.com
            version: v1alpha1
          - description: Restore is the Schema for the restores API. It is used to
              define restore jobs and its restoration source.
            displayName: Restore
            kind: Restore
            name: restores.enterprise.mariadb.com
            version: v1alpha1
          - description: SqlJob is the Schema for the sqljobs API. It is used to run
              sql scripts as jobs.
            displayName: Sql Job
            kind: SqlJob
            name: sqljobs.enterprise.mariadb.com
            version: v1alpha1
          - description: User is the Schema for the users API.  It is used to define
              grants as if you were running a 'CREATE USER' statement.
            displayName: User
            kind: User
            name: users.enterprise.mariadb.com
            version: v1alpha1
        description: |
          MariaDB Enterprise Operator for Openshift is an operator that manages MariaDB Enterprise operands declaratively via Custom Resources, with support for the following features:

          * Provision and Configure MariaDB and MaxScale Declaratively: Define MariaDB Enterprise Server and MaxScale clusters in YAML manifests and deploy them with ease in Kubernetes.
          * High Availability with Galera: Ensure availability with MariaDB Enterprise Cluster, providing synchronous multi-master replication.
          * Query and Connection-Based Routing with MaxScale: MaxScale provides query routing and connection load balancing for improved application performance.
          * Cluster-Aware Rolling Updates: Perform rolling updates on MariaDB and MaxScale clusters, ensuring zero-downtime upgrades with no disruptions to your applications.
          * Flexible Storage Configuration and Volume Expansion: Easily configure storage for MariaDB instances, including the ability to expand volumes as needed.
          * Physical Backups based on [mariadb-backup](https://mariadb.com/docs/server/server-usage/backup-and-restore/mariadb-backup/full-backup-and-restore-with-mariadb-backup) and [Kubernetes VolumeSnapshots](https://kubernetes.io/docs/concepts/storage/volume-snapshots/). By leveraging the [BACKUP STAGE](https://mariadb.com/docs/server/reference/sql-statements/administrative-sql-statements/backup-commands/backup-stage) feature, backups are taken without long read locks or service interruptions.
          * Logical Backups based on [mariadb-dump](https://mariadb.com/docs/server/clients-and-utilities/backup-restore-and-import-clients/mariadb-dump).
          * Backup Management: Take, restore, and schedule backups with multiple storage types supported: S3, PVCs, Kubernetes volumes and VolumeSnapshots..
          * Policy-Driven Backup Retention: Implement backup retention policies with bzip2 and gzip compression.
          * Target Recovery Time: Restore your database to the closest available backup based on a specified recovery time.
          * Bootstrap New Instances: Initialize new MariaDB instances from backups, S3, PVCs or VolumeSnapshots to quickly spin up new clusters.
          * TLS Certificate Management: Issue, configure, and rotate TLS certificates and Certificate Authorities (CAs) for secure connections.
          * Advanced TLS Support: customize certificate lifetime, private key algorithm and TLS version.
          * Native Integration with cert-manager: Leverage [cert-manager](https://cert-manager.io/docs/), the de-facto standard for managing certificates in Kubernetes, to enable issuance with private CAs, public CAs and HashiCorp Vault.
          * Prometheus Metrics: Expose metrics using the MariaDB and MaxScale Prometheus exporters.
          * Native Integration with prometheus-operator: Leverage [prometheus-operator](https://github.com/prometheus-operator/prometheus-operator) to scrape metrics from MariaDB and MaxScale instances.
          * Declarative User and Database Management: Manage users, grants, and logical databases in a declarative manner using Kubernetes resources.
          * Secure, immutable and lightweight images based on Red Hat UBI, available for multiple architectires (amd64, arm64 and ppc64le).
          * [Operator certified ](https://catalog.redhat.com/en/software/container-stacks/detail/65789bcbe17f1b31944acb1d#overview) by Red Hat.

          ## Installation

          In order to use this product, you will need a [MariaDB Enterprise subscription](https://mariadb.com/products/enterprise/).

          To be able to pull MariaDB Enterprise images, you will need to configure the MariaDB registry credentials in Openshift using the [MariaDB customer download token](https://customers.mariadb.com/downloads/token/) as described below:

          - Extract your [Openshift global pull secret](https://docs.redhat.com/en/documentation/openshift_container_platform/4.18/html/images/managing-images#images-update-global-pull-secret_using-image-pull-secrets):
            ```bash
            oc extract secret/pull-secret -n openshift-config --confirm
            ```

          - Login in the MariaDB registry providing the customer download token as password:
            ```bash
            oc registry login \
              --registry="docker.mariadb.com" \
              --auth-basic="<email>:<customer-download-token>" \
              --to=.dockerconfigjson
            ```

          - Update the global pull secret:
            ```bash
            oc set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson
            ```

          Alternatively, you can also create a `Secret` named `mariadb-enterprise` in the namespace where are you installing the operator by using the following command:
            ```bash
            oc create secret docker-registry mariadb-enterprise \
              --docker-server=docker.mariadb.com \
              --docker-username=<email> \
              --docker-password=<customer-download-token> \
              -n <operator-namespace>
            ```

          At this point, you can proceed to install the MariaDB Enterprise Operator, either via the Openshift console or by creating a `Subscription` object as described in the [Openshift installation docs](https://mariadb.com/kb/en/openshift/).

          Refer to the [quickstart guide](https://mariadb.com/kb/en/quickstart/), the [examples catalog](https://mariadb.com/kb/en/examples-catalog/) and the [API reference](https://mariadb.com/kb/en/api-reference/) for getting started with the Custom Resources offered by this operator.
        displayName: MariaDB Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mariadb
        - mysql
        - database
        - operator
        - mariadb-enterprise-operator
        - database
        links:
        - name: MariaDB Enterprise
          url: https://mariadb.com/products/enterprise/
        - name: Documentation
          url: https://mariadb.com/kb/en/mariadb-enterprise-operator/
        maintainers:
        - email: operator@mariadb.com
          name: MariaDB Corporation
        - email: support@mariadb.com
          name: Support
        - email: sales-EMEA@mariadb.com
          name: Sales EMEA
        - email: sales-APAC@mariadb.com
          name: Sales APAC
        - email: sales-nam@mariadb.com
          name: Sales Americas
        maturity: beta
        minKubeVersion: 1.29.0
        provider:
          name: MariaDB Corporation
        relatedImages:
        - docker.mariadb.com/enterprise-server@sha256:9872721d35ebe8d820b1aa9c6dd83b363a7b2df9af11a65bff6b80ff4144460a
        - docker.mariadb.com/maxscale@sha256:d9e04a7a406353dab2051e190be0c32ed20b780ed71e56e55cedd4ccb61af416
        - mariadb/mariadb-prometheus-exporter-ubi@sha256:f072175198cad050e80617d13e63f415bacdfe47e2dd177fde7283a96ee56c07
        - mariadb/maxscale-prometheus-exporter-ubi@sha256:cc17c1554795ad01b4d8db799f6aa90e00e41ede431f09d30eec64d9dd4479d9
        - docker.mariadb.com/mariadb-enterprise-operator@sha256:1effbbfa7632be92c17ef61c18a25f010b187840beec42c31d2410d0209e6985
        version: 25.8.1-1
      entries:
      - name: mariadb-enterprise-operator.v25.8.1-1
        version: 25.8.1-1
      - name: mariadb-enterprise-operator.v25.8.0
        version: 25.8.0
      - name: mariadb-enterprise-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: mariadb-enterprise-operator
    provider:
      name: MariaDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: manila-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: manila-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "manila.openstack.org/v1beta1",
                "kind": "Manila",
                "metadata": {
                  "name": "manila",
                  "namespace": "openstack"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "manila",
                  "databaseInstance": "openstack",
                  "manilaAPI": {},
                  "manilaScheduler": {},
                  "manilaShares": {
                    "share1": {}
                  },
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "osp-secret",
                  "serviceUser": "manila"
                }
              },
              {
                "apiVersion": "manila.openstack.org/v1beta1",
                "kind": "ManilaAPI",
                "metadata": {
                  "name": "manilaapi-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "manila.openstack.org/v1beta1",
                "kind": "ManilaScheduler",
                "metadata": {
                  "name": "manilascheduler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "manila.openstack.org/v1beta1",
                "kind": "ManilaShare",
                "metadata": {
                  "name": "manilashare-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:57Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManilaAPI is the Schema for the manilaapis API
            displayName: Manila API
            kind: ManilaAPI
            name: manilaapis.manila.openstack.org
            version: v1beta1
          - description: Manila is the Schema for the manilas API
            displayName: Manila
            kind: Manila
            name: manilas.manila.openstack.org
            version: v1beta1
          - description: ManilaScheduler is the Schema for the manilaschedulers API
            displayName: Manila Scheduler
            kind: ManilaScheduler
            name: manilaschedulers.manila.openstack.org
            version: v1beta1
          - description: ManilaShare is the Schema for the manilashares API
            displayName: Manila Share
            kind: ManilaShare
            name: manilashares.manila.openstack.org
            version: v1beta1
        description: Manila Operator
        displayName: Manila Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Share
        - Manila
        links:
        - name: Manila Operator
          url: https://github.com/openstack-k8s-operators/manila-operator
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/manila-rhel9-operator@sha256:1208676b94d8a726c277088591d898e08a3a8331a0125ba7bff101152afba934
        - registry.redhat.io/rhoso/openstack-manila-scheduler-rhel9@sha256:7a8f645126c073213ce066cdcac2164df51276f8aa4a3f89954664d3010e6fae
        - registry.redhat.io/rhoso/openstack-manila-share-rhel9@sha256:83a6d292785805a0ac2895810ea466cfb6e1cb7aa1a53ae8f4e22671381490df
        - registry.redhat.io/rhoso/openstack-manila-api-rhel9@sha256:4cbf725169614c2974d6d1a683cfc0179bfe06ca29cf30f7dca88caeb368c51a
        version: 1.0.6
      entries:
      - name: manila-operator.v1.0.6
        version: 1.0.6
      - name: manila-operator.v1.0.5
        version: 1.0.5
      - name: manila-operator.v1.0.4
        version: 1.0.4
      - name: manila-operator.v1.0.3
        version: 1.0.3
      - name: manila-operator.v1.0.2
        version: 1.0.2
      - name: manila-operator.v1.0.1
        version: 1.0.1
      - name: manila-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: manila-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: horizon-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: horizon-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "horizon.openstack.org/v1beta1",
                "kind": "Horizon",
                "metadata": {
                  "name": "horizon"
                },
                "spec": {
                  "customServiceConfig": "DEBUG = True\nLOGGING['handlers']['console']['level'] = 'DEBUG'\n",
                  "replicas": 1,
                  "secret": "osp-secret"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:00Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Horizon is the Schema for the horizons API
            displayName: Horizon
            kind: Horizon
            name: horizons.horizon.openstack.org
            version: v1beta1
        description: Horizon Operator
        displayName: Horizon Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - UI
        - Horizon
        links:
        - name: Horizon Operator
          url: https://github.com/openstack-k8s-operators/horizon-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-horizon-rhel9@sha256:b02c115d40aacc30e6763e295da90d69681b24fba20ebbc53133be852393b399
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/horizon-rhel9-operator@sha256:b76038b96a064c275721c1d45b95d7b0dfdb9fc626a2b3250d0adb1f97f89319
        version: 1.0.6
      entries:
      - name: horizon-operator.v1.0.6
        version: 1.0.6
      - name: horizon-operator.v1.0.5
        version: 1.0.5
      - name: horizon-operator.v1.0.4
        version: 1.0.4
      - name: horizon-operator.v1.0.3
        version: 1.0.3
      - name: horizon-operator.v1.0.2
        version: 1.0.2
      - name: horizon-operator.v1.0.1
        version: 1.0.1
      - name: horizon-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: horizon-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq-streams
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amqstreams.v2.2.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.3",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.3",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:f908d848b5e67de4b162597b190ef18d449ffd4b6540457b24b491445d5c5ea3
          createdAt: "2022-07-01 10:33:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.1.0-0 <2.2.2-2'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ", "Red Hat Application Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from AMQ Streams 1.7, migration to `v1beta2` needs to be completed for all AMQ Streams CRDs and CRs before the upgrade to 2.2 is done! If upgrading from AMQ Streams version earlier than 1.7, you need to first install the CRDs from AMQ Streams 1.7 and complete the migration to `v1beta2` for all AMQ Streams CRDs and CRs before the upgrade to 2.2 is done! For more details about the CRD upgrades, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.2/html/deploying_and_upgrading_amq_streams_on_openshift/assembly-upgrade-str#assembly-upgrade-resources-str).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.2/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: Red Hat Integration - AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        - registry.redhat.io/amq7/amq-streams-kafka-31-rhel8@sha256:d0607285fae95365b9d34b9474466824fb5ad012871204e8b13e52e438b66bf5
        - registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:f908d848b5e67de4b162597b190ef18d449ffd4b6540457b24b491445d5c5ea3
        - registry.redhat.io/amq7/amq-streams-kafka-32-rhel8@sha256:e3bbfff505f12c0360e175f4640c2dd9a7adcf3dce1ae2f586f2bde805ed8b1f
        - registry.redhat.io/amq7/amq-streams-bridge-rhel8@sha256:5d49bd23da40590e74422424b9457398bad89da8276718ff6a8a088e901c8345
        version: 2.2.2-2
      entries:
      - name: amqstreams.v2.2.2-2
        version: 2.2.2-2
      - name: amqstreams.v2.2.2-1
        version: 2.2.2-1
      - name: amqstreams.v2.2.2-0
        version: 2.2.2-0
      - name: amqstreams.v2.2.1-7
        version: 2.2.1-7
      - name: amqstreams.v2.2.1-6
        version: 2.2.1-6
      - name: amqstreams.v2.2.1-5
        version: 2.2.1-5
      - name: amqstreams.v2.2.1-4
        version: 2.2.1-4
      - name: amqstreams.v2.2.1-3
        version: 2.2.1-3
      - name: amqstreams.v2.2.1-2
        version: 2.2.1-2
      - name: amqstreams.v2.2.1-1
        version: 2.2.1-1
      - name: amqstreams.v2.2.1-0
        version: 2.2.1-0
      - name: amqstreams.v2.2.0-5
        version: 2.2.0-5
      name: amq-streams-2.2.x
    - currentCSV: amqstreams.v2.3.0-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.3.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.3.1",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:2e46a25144d697761f0119deaf528c6f1c2ccd04e2acd95c661b0df4f32e1a9f
          createdAt: "2022-10-31 17:05:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.2.0-0 <2.3.0-4'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.3.0 release, AMQ Streams supports only OCP 4.8 to 4.12!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 2.2 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more details, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.3/html/deploying_and_upgrading_amq_streams_on_openshift/assembly-upgrade-str#assembly-upgrade-resources-str

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.3/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-streams-kafka-32-rhel8@sha256:0152a85f2d0bc85881dacaee62c6e8b0b8b04f57a542b9b82dc93ad5303d6bf9
        - registry.redhat.io/amq7/amq-streams-kafka-33-rhel8@sha256:046c2539fef32bd928fc1f4f4c41e63f9a143b371fcc201d78c66f48fc34ea32
        - registry.redhat.io/amq7/amq-streams-bridge-rhel8@sha256:c822b3d87f7dea233fc523238d8d0516e42f9a3d252f50f4d427c934995f9b22
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        - registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:2e46a25144d697761f0119deaf528c6f1c2ccd04e2acd95c661b0df4f32e1a9f
        version: 2.3.0-4
      entries:
      - name: amqstreams.v2.3.0-4
        version: 2.3.0-4
      name: amq-streams-2.3.x
    - currentCSV: amqstreams.v2.4.0-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:cd1fefcd199400a504b4b9942c24ba6becafcecc952a9e39fa2729228be79e77
          createdAt: "2022-10-31 17:05:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.3.0-0 <2.4.0-3'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.4.0 release, AMQ Streams supports only OCP 4.8 to 4.12!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more details, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.4/html/deploying_and_managing_amq_streams_on_openshift/assembly-upgrade-str)

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.4/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:8d467fc746b78440df06f6d88d266f6ac9b5a4574a2b17fd870c173a107f8d65
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:cd1fefcd199400a504b4b9942c24ba6becafcecc952a9e39fa2729228be79e77
        - registry.redhat.io/amq-streams/kafka-33-rhel8@sha256:aa780a250444c2426be79e95366050a33f4a4c9d68c08b58743413ea67a2d6d7
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:50b3be789993cfeedc4923a8e760c1cf52498b2178be3abfa3719fc2d15ce723
        version: 2.4.0-3
      entries:
      - name: amqstreams.v2.4.0-3
        version: 2.4.0-3
      - name: amqstreams.v2.4.0-2
        version: 2.4.0-2
      - name: amqstreams.v2.4.0-1
        version: 2.4.0-1
      name: amq-streams-2.4.x
    - currentCSV: amqstreams.v2.5.2-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
          createdAt: "2023-07-10 21:47:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.2.0-0 <2.5.2-8'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.5.2 release, AMQ Streams supports only OCP 4.12 to 4.16!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more information, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:8c7e8ccb6c2d5da32b70d9d0c3919c303ad73a25447f6993d964f7b19cca1072
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:4ede5eeed159bddacfe830a768bab4996003d4c57a496fc8ec9b41df916d4acc
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:95bc35951eb545b918114af1e92a7139e2e3323b7d054dcb2a060843c7326d28
        version: 2.5.2-8
      entries:
      - name: amqstreams.v2.5.2-8
        version: 2.5.2-8
      - name: amqstreams.v2.5.2-7
        version: 2.5.2-7
      - name: amqstreams.v2.5.2-6
        version: 2.5.2-6
      - name: amqstreams.v2.5.2-5
        version: 2.5.2-5
      - name: amqstreams.v2.5.2-4
        version: 2.5.2-4
      - name: amqstreams.v2.5.2-3
        version: 2.5.2-3
      - name: amqstreams.v2.5.2-2
        version: 2.5.2-2
      - name: amqstreams.v2.5.2-1
        version: 2.5.2-1
      - name: amqstreams.v2.5.2-0
        version: 2.5.2-0
      - name: amqstreams.v2.5.1-5
        version: 2.5.1-5
      - name: amqstreams.v2.5.1-4
        version: 2.5.1-4
      - name: amqstreams.v2.5.1-3
        version: 2.5.1-3
      - name: amqstreams.v2.5.1-2
        version: 2.5.1-2
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.5.x
    - currentCSV: amqstreams.v2.6.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.6.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:dacf7ddffe7cbe2dc35e7d3290637b3334117e09dde986e6eaeaceded7069fbf
          createdAt: "2023-10-14 18:18:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.6.0-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: "**Red Hat AMQ Streams** is a massively scalable, distributed,
          and high performance data streaming platform based on the Apache Kafka®
          project. AMQ Streams provides an event streaming backbone that allows microservices
          and other application components to exchange data with extremely high throughput
          and low latency.\n**From the 2.6.0 release, AMQ Streams supports only OCP
          4.11 to 4.14! \n**The core capabilities include:** * A pub/sub messaging
          model, similar to a traditional enterprise messaging system, in which application
          components publish and consume events to/from an ordered stream\n* The long
          term, fault-tolerant storage of events\n* The ability for a consumer to
          replay streams of events\n* The ability to partition topics for horizontal
          scalability\n### Upgrades\n**!!! IMPORTANT !!!** **Direct upgrade from AMQ
          Stream 1.7 or earlier is not supported!** You have to upgrade first to one
          of the previous versions of AMQ Streams. You will also need to convert the
          CRD resources. For more information, see the [documentation for AMQ Streams
          on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).\n###
          Supported Features\n* **Manages the Kafka Cluster** - Deploys and manages
          all of the components of this complex application, including dependencies
          like Apache ZooKeeper® that are traditionally hard to administer.\n* **Includes
          Kafka Connect** - Allows for configuration of common data sources and sinks
          to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for mirroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashboards\n### Upgrading your Clusters\nThe
          AMQ Streams Operator understands how to run and upgrade between a set of
          Kafka versions. When specifying a new version in your config, check to make
          sure you aren't using any features that may have been removed. For more
          information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).\n###
          Storage\nAn efficient data storage infrastructure is essential to the optimal
          performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires
          block storage. The use of file storage (for example, NFS) is not recommended.\nThe
          AMQ Streams Operator supports three types of data storage:\n* Ephemeral
          (Recommended for development only!)\n* Persistent\n* JBOD (Just a Bunch
          of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nAMQ Streams
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation for the current release
          can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).\n"
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:dacf7ddffe7cbe2dc35e7d3290637b3334117e09dde986e6eaeaceded7069fbf
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:be7466ffd7f624f769eedca2e4106c597d2bf1d95cc9679f235e6dcdec66cfa9
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:fa8361e5089ab4449261f5a0ad57bd0bd059dc254b94bd9ab68340faae386947
        - registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:8b10bfb697b48ba3ed246fb846384f4cc67f05f670e9521edf0f47f829869404
        version: 2.6.0-2
      entries:
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.6.x
    - currentCSV: amqstreams.v2.7.0-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:671c1a2fcb0856cb208c0ead539fff9571fa2281a1e3d45b3df9ebd3df8c787e
          createdAt: "2024-03-09 21:27:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0-0 <2.7.0-7'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.6.0 release, Streams supports only OCP 4.12 to 4.14!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:671c1a2fcb0856cb208c0ead539fff9571fa2281a1e3d45b3df9ebd3df8c787e
        - registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:66362c800b78003d26df9471b7f21f5c9ad7fca8e8fd2511428dfc549cdf46c2
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:be9b82af13cbc120bb415abf4fcc97420e543172d6450b836c47344e4dcacd00
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:a7620559f3e5a0a4d5ecee44538619f698158847d9d0a06c7168404ed6f7cd36
        - registry.redhat.io/amq-streams/kafka-36-rhel9@sha256:8890c3217343d9aa3dd2ab9c7fa1742f960b5a26e09906581ce6075e61dc2888
        version: 2.7.0-7
      entries:
      - name: amqstreams.v2.7.0-7
        version: 2.7.0-7
      - name: amqstreams.v2.7.0-6
        version: 2.7.0-6
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.7.x
    - currentCSV: amqstreams.v2.8.0-0-0.1738265624.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:675d4acc375fb20e1a3fd5755184a318341c15f9337a11ea41e902fcda6d0bec
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.0-0 <2.8.0-0'
          olm.substitutesFor: amqstreams.v2.8.0-0
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.8.0 release, Streams supports only OCP 4.12 to 4.14!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:675d4acc375fb20e1a3fd5755184a318341c15f9337a11ea41e902fcda6d0bec
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:b46457a0a7f63518bc12b2789f75afb0e44496e62de918b9b584129c7078de86
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:4df906cf1490922a4587d96250f3be6ff605fe2caf87dfba7b812c144c5c5774
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:e771a84a458388c07ecf970ec625e4203c18e5d2bd91e7571d9c44ebb09313de
        - registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:17b2d255439f38c6ef54c0a56929ae636743b12c688d42223ac3fb58729e51ad
        version: 2.8.0-0+0.1738265624.p
      entries:
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.8.x
    - currentCSV: amqstreams.v2.9.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.9.2-1'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.9.0 release, Streams supports only OCP 4.14 to 4.18!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:f5985592b7f05c9b3c64caa0ac90f34f26b9b4277b342b6bd4a6415d2c563f9a
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:86ea2e823cd83c476521f2a1e9da5619c4810661c60d28295d0f5482c8aa519d
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:ddfb50ec2f1cf45ccccc632182bf0ff549c14969da35b7248802eab67d9c2e9d
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:67647bff1a76adb5b51f43c99d5f88a8aa38055c1e87190add8972bb621ddc15
        version: 2.9.2-1
      entries:
      - name: amqstreams.v2.9.2-1
        version: 2.9.2-1
      - name: amqstreams.v2.9.2-0
        version: 2.9.2-0
      - name: amqstreams.v2.9.1-2
        version: 2.9.1-2
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.9.x
    - currentCSV: amqstreams.v2.9.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.9.2-1'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.9.0 release, Streams supports only OCP 4.14 to 4.18!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:86ea2e823cd83c476521f2a1e9da5619c4810661c60d28295d0f5482c8aa519d
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:ddfb50ec2f1cf45ccccc632182bf0ff549c14969da35b7248802eab67d9c2e9d
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:67647bff1a76adb5b51f43c99d5f88a8aa38055c1e87190add8972bb621ddc15
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:f5985592b7f05c9b3c64caa0ac90f34f26b9b4277b342b6bd4a6415d2c563f9a
        version: 2.9.2-1
      entries:
      - name: amqstreams.v2.9.2-1
        version: 2.9.2-1
      - name: amqstreams.v2.9.2-0
        version: 2.9.2-0
      - name: amqstreams.v2.9.1-2
        version: 2.9.1-2
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.x
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-3.0.x
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-3.x
    - currentCSV: amqstreams.v2.5.2-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
          createdAt: "2023-07-10 21:47:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.2.0-0 <2.5.2-8'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.5.2 release, AMQ Streams supports only OCP 4.12 to 4.16!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more information, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:4ede5eeed159bddacfe830a768bab4996003d4c57a496fc8ec9b41df916d4acc
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:95bc35951eb545b918114af1e92a7139e2e3323b7d054dcb2a060843c7326d28
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:8c7e8ccb6c2d5da32b70d9d0c3919c303ad73a25447f6993d964f7b19cca1072
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
        version: 2.5.2-8
      entries:
      - name: amqstreams.v2.5.2-8
        version: 2.5.2-8
      - name: amqstreams.v2.5.2-7
        version: 2.5.2-7
      - name: amqstreams.v2.5.2-6
        version: 2.5.2-6
      - name: amqstreams.v2.5.2-5
        version: 2.5.2-5
      - name: amqstreams.v2.5.2-4
        version: 2.5.2-4
      - name: amqstreams.v2.5.2-3
        version: 2.5.2-3
      - name: amqstreams.v2.5.2-2
        version: 2.5.2-2
      - name: amqstreams.v2.5.2-1
        version: 2.5.2-1
      - name: amqstreams.v2.5.2-0
        version: 2.5.2-0
      - name: amqstreams.v2.5.1-5
        version: 2.5.1-5
      - name: amqstreams.v2.5.1-4
        version: 2.5.1-4
      - name: amqstreams.v2.5.1-3
        version: 2.5.1-3
      - name: amqstreams.v2.5.1-2
        version: 2.5.1-2
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-lts
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: stable
    defaultChannel: stable
    packageName: amq-streams
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JFrog
      provider-url: ""
    name: openshiftxray-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: xray-operator.v2.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftXray",
                "metadata": {
                  "name": "openshiftxray"
                },
                "spec": {
                  "xray": {
                    "common": {
                      "xrayUserId": "1000721035",
                      "xrayGroupId": "1000721035"
                    },
                    "analysis": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-analysis"
                      },
                      "name": "xray-analysis",
                      "podManagementPolicy": "Parallel",
                      "preStartCommand": null,
                      "updateStrategy": "RollingUpdate"
                    },
                    "database": {
                      "password": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "global": {
                      "postgresqlTlsSecret": null
                    },
                    "indexer": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-indexer"
                      },
                      "name": "xray-indexer",
                      "podManagementPolicy": "Parallel",
                      "updateStrategy": "RollingUpdate"
                    },
                    "persist": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-persist"
                      },
                      "name": "xray-persist",
                      "persistence": {
                        "size": "10Gi"
                      },
                      "podManagementPolicy": "Parallel",
                      "preStartCommand": null,
                      "updateStrategy": "RollingUpdate"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "rabbitmq": {
                      "enabled": true,
                      "replicaCount": 1,
                      "rbac": {
                        "create": true
                      },
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-rabbitmq"
                      },
                      "podSecurityContext": {
                        "fsGroup": 1000721001,
                        "runAsUser": 1000721001
                      },
                      "auth": {
                        "username": "xray",
                        "password": "xray",
                        "erlangCookie": "XRAYRABBITMQCLUSTER",
                        "tls": {
                          "enabled": false,
                          "failIfNoPeerCert": true,
                          "sslOptionsVerify": "verify_peer",
                          "caCertificate": "",
                          "serverCertificate": "",
                          "serverKey": ""
                        }
                      }
                    },
                    "rabbitmq-ha": {
                      "enabled": false,
                      "image": {
                        "repository": "registry.connect.redhat.com/jfrog/xray-rabbitmq"
                      },
                      "rabbitmqUsername": "guest",
                      "rabbitmqPassword": "guest",
                      "managementUsername": "guest",
                      "managementPassword": "guest",
                      "initContainer": {
                        "enabled": false
                      },
                      "securityContext": {
                        "fsGroup": 1000721035,
                        "runAsUser": 1000721035,
                        "runAsGroup": 1000721035
                      },
                      "replicaCount": 1
                    },
                    "replicaCount": 1,
                    "unifiedUpgradeAllowed":  "true",
                    "router": {
                      "image": {
                        "imagePullPolicy": "IfNotPresent",
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-router"
                      },
                      "name": "router"
                    },
                    "server": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-server"
                      },
                      "name": "xray-server",
                      "podManagementPolicy": "Parallel",
                      "replicaCount": 1,
                      "updateStrategy": "RollingUpdate"
                    },
                    "xray": {
                      "consoleLog": false,
                      "jfrogUrl": "OVERRIDE",
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/xray-operator@sha256:da7f6977170876b335875f4074ddfc0266b665351249295ce3c53f0e95891cc2
          createdAt: 2020-05-22 00:00:00+00:00
          description: JFrog Xray Enterprise deploys Xray continuous security scanner
            into Openshift (Requires Jfrog Artifactory)
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Xray Instances
            displayName: Xray
            kind: OpenshiftXray
            name: openshiftxrays.charts.helm.k8s.io
            version: v1
        description: "## Overview Openshift Operator to deploy JFrog Xray Continuous
          Security scanner into your Openshift cluster. NOTE: Artifactory is required
          for JFrog Xray to connect with and scan artifacts against.\n### Major version
          v2.0.4 Breaking change notification!!\nThis release has major bug fix for
          Rabbitmq Pod. Please update to version 2.0.4 of the operator.\n### Major
          version v2.0.0 Breaking change notification!! Xray is dependent upon Rabbitmq.
          The version of Rabbitmq that this chart uses has significantly changed from
          version 1.1.8 of the operator. For full details on how to migrate please
          visit: `https://github.com/jfrog/charts/blob/master/stable/xray/RABBITMQ_MIGRATION_NOTES.md`\nIf
          you are using operator version < 2.1.0 there is a issue with upgrade where
          xray version is not updated if you update operator version.As workaround
          you can update the Custom Resource OpenshiftXray which has hardcored image
          tag value.After upgrading to version 2.1.0 you can remove the image tags
          parameters from your CR so that future upgrade works without workaround.\n###
          Security Context Constraints To deploy this helm chart you will need to
          be a cluster admin w/ access to the anyuid scc and add the operator service
          account to the anyuid scc. ``` oc adm policy add-scc-to-user anyuid -z xray-operator
          ``` ``` oc adm policy add-scc-to-user anyuid -z openshiftxray ``` ``` oc
          adm policy add-scc-to-user anyuid -z openshiftxray-rabbitmq-ha ``` ## Usage
          An external DB is required. The operator will not deploy a DB but will require
          you to specify the configuration values to connect to it. Search for JFrog
          and click JFrog Xray Operator to install. Go to the Installed Operators.
          Wait for the JFrog Xray Operator to complete the installation. Open the
          Operator and click on the provided API: Xray Click Create New Instance and
          provide the following parameters for your DB configuration: ``` JFROG_URL
          DATABASE_URL DATABASE_USER DATABASE_PASSWORD ``` JFROG_URL is the external
          ip or DNS of your Artifactory to connect Xray to. Artifactory is required
          to use this operator. You can set your JFROG_URL to the service name of
          your Artifactory Nginx: ``` oc get svc -n my_namespace | grep nginx ```
          DATABASE_URL must be a Postgresql URL in the format: `` postgres://postgres-postgresql:5432/xraydb?sslmode=disable
          `` DATABASE_USER and DATABASE_PASSWORD must supply a valid user on Postgresql.
          Click Create for Xray to deploy into OpenShift. Open Artifactory in a web
          browser to complete the onboarding wizard for Xray! ### Air gap environments
          To use Xray Operator in an air gap environment you will need to download
          the images as image streams into your Openshift air gap cluster manually.
          Use the image overrides to then specify the image stream locations that
          are local to your cluster. Next you will need to setup the database of index
          data for xray to use when scanning artifacts. Follow the link below for
          instructions on setup: https://www.jfrog.com/confluence/display/JFROG/Configuring+Xray#ConfiguringXray-SynchronizingtheDatabase "
        displayName: JFrog Xray Continuous Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Xray Continuous Security
          url: https://jfrog.com/xray/
        - name: JFrog Xray Continuous Security Video
          url: https://youtu.be/wGzyKlQklKQ
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: alpha
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/xray-indexer@sha256:7f40d2dfe4a35b25d16dfcbfa04b4a58f227c132d4c3c86345e7b2bd19a97311
        - registry.connect.redhat.com/jfrog/xray-router@sha256:019235289e17d0e22dcd23a1505f71a78710827dfab338d8fa980f64c14f162f
        - registry.connect.redhat.com/jfrog/xray-rabbitmq@sha256:459d7d5e78192a6f45a5527e8fbaf1d8b30299b10ad4c42515d3d2bb5fcc75f4
        - registry.connect.redhat.com/jfrog/init@sha256:3f27e44b44c3556c50318110dbd542e34ef1061e28d7f0e0e7b8baa73c5f3e87
        - registry.connect.redhat.com/jfrog/xray-server@sha256:c5584c4e9528f0199e840b2b1ad1144220850ca9595ebf3e270bc6f86de4a71a
        - registry.connect.redhat.com/jfrog/xray-analysis@sha256:bcd8211f1bc753bfd7735165ce39b476e4efff198bffa877883620b41fcb33b5
        - registry.connect.redhat.com/jfrog/xray-persist@sha256:93fae0c124919a57d5162e32e88d3456ef6254889e454f89c21c89e69e6ae323
        - registry.connect.redhat.com/jfrog/xray-operator@sha256:da7f6977170876b335875f4074ddfc0266b665351249295ce3c53f0e95891cc2
        version: 2.1.6
      entries:
      - name: xray-operator.v2.1.6
        version: 2.1.6
      name: alpha
    defaultChannel: alpha
    packageName: openshiftxray-operator
    provider:
      name: JFrog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: m88i Labs
      provider-url: ""
    name: nexus-operator-m88i
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.m88i.io/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "name": "nexus3"
                },
                "spec": {
                  "networking": {
                    "expose": false
                  },
                  "persistence": {
                    "persistent": false
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    },
                    "requests": {
                      "cpu": "1",
                      "memory": "2Gi"
                    }
                  },
                  "useRedHatImage": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/m88i/nexus-operator:0.6.0
          createdAt: "2019-11-16T13:12:22Z"
          description: Nexus Operator to deploy and manage Nexus 3.x servers
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/m88i/nexus-operator
          support: m88i Labs
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: Nexus
            kind: Nexus
            name: nexus.apps.m88i.io
            version: v1alpha1
        description: |-
          Creates a new Nexus 3.x deployment in a Kubernetes cluster. Will help DevOps to have a quick Nexus application exposed to the world that can be used in a CI/CD process:

          * Deploys a new Nexus 3.x server based on either Community or Red Hat images
          * Creates an [Ingress](https://kubernetes.io/docs/concepts/services-networking/ingress/) in Kubernetes (1.14+) environments to expose the application to the world
          * On OpenShift, creates a Route to expose the service outside the cluster
          * Automatically creates Apache, Red Hat and JBoss Maven repositories
          * Automatically updates Nexus within the same minor version

          [See our documentation](https://github.com/m88i/nexus-operator/blob/main/README.md) for more installation and usage scenarios.

          If you experience any issues or have any ideas for new features, please [file an issue in our Github repository](https://github.com/m88i/nexus-operator/issues) or send an email to our maillist: [nexus-operator@googlegroups.com](mailto:nexus-operator@googlegroups.com)

          *Please note that the operator is an individual work and it's not provided nor supported by Sonatype.*
        displayName: Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nexus
        - sonatype
        - maven
        - docker
        - ci
        - continuous integration
        - continuous delivery
        - repository
        - repository manager
        - dev tools
        links:
        - name: Documentation
          url: https://github.com/m88i/nexus-operator/blob/main/README.md
        - name: Nexus Operator source code repository
          url: https://github.com/m88i/nexus-operator
        maintainers:
        - email: nexus-operator@googlegroups.com
          name: m88i Labs
        maturity: alpha
        provider:
          name: m88i Labs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/m88i/nexus-operator:0.6.0
        version: 0.6.0
      entries:
      - name: nexus-operator.v0.6.0
        version: 0.6.0
      name: alpha
    defaultChannel: alpha
    packageName: nexus-operator-m88i
    provider:
      name: m88i Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Aerospike
      provider-url: ""
    name: aerospike-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aerospike-kubernetes-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "memory-size": 3000000000,
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:6.3.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: docker.io/aerospike/aerospike-kubernetes-operator:2.5.0
          createdAt: 2023-05-04T09:32:47UTC
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Aerospike
        relatedImages:
        - docker.io/aerospike/aerospike-kubernetes-operator:2.5.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 2.5.0
      entries:
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      - name: aerospike-kubernetes-operator.v2.4.0
        version: 2.4.0
      - name: aerospike-kubernetes-operator.v2.3.0
        version: 2.3.0
      - name: aerospike-kubernetes-operator.v2.2.1
        version: 2.2.1
      - name: aerospike-kubernetes-operator.v2.2.0
        version: 2.2.0
      - name: aerospike-kubernetes-operator.v2.1.0
        version: 2.1.0
      name: alpha
    - currentCSV: aerospike-kubernetes-operator.v4.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "data-size": 1073741824,
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:8.1.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackup",
                "metadata": {
                  "name": "aerospikebackup",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "aerospike-cluster": {
                      "aerospike-aerospikebackup-test-cluster": {
                        "credentials": {
                          "password": "admin123",
                          "user": "admin"
                        },
                        "seed-nodes": [
                          {
                            "host-name": "aerocluster.aerospike.svc.cluster.local",
                            "port": 3000
                          }
                        ]
                      }
                    },
                    "backup-routines": {
                      "aerospike-aerospikebackup-test-routine": {
                        "backup-policy": "test-policy",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "local"
                      },
                      "aerospike-aerospikebackup-test-routine1": {
                        "backup-policy": "test-policy1",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "s3Storage"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackupService",
                "metadata": {
                  "name": "aerospikebackupservice-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "config": {
                    "backup-policies": {
                      "test-policy": {
                        "parallel": 3
                      },
                      "test-policy1": {
                        "parallel": 3
                      }
                    },
                    "service": {
                      "http": {
                        "port": 8081
                      }
                    },
                    "storage": {
                      "local": {
                        "local-storage": {
                          "path": "/tmp/localStorage"
                        }
                      },
                      "s3Storage": {
                        "s3-storage": {
                          "bucket": "aerospike-kubernetes-operator-test",
                          "s3-profile": "default",
                          "s3-region": "us-east-1"
                        }
                      }
                    }
                  },
                  "image": "aerospike/aerospike-backup-service:3.2.0",
                  "secrets": [
                    {
                      "secretName": "aws-secret",
                      "volumeMount": {
                        "mountPath": "/root/.aws/credentials",
                        "name": "aws-secret",
                        "subPath": "credentials"
                      }
                    }
                  ],
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeRestore",
                "metadata": {
                  "name": "aerospikerestore-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "backup-data-path": "aerospike-aerospikebackup-test-routine/backup/1733506015827/data/test",
                    "destination": {
                      "credentials": {
                        "password": "admin123",
                        "user": "admin"
                      },
                      "label": "destinationCluster",
                      "seed-nodes": [
                        {
                          "host-name": "aerocluster.test.svc.cluster.local",
                          "port": 3000
                        }
                      ]
                    },
                    "policy": {
                      "no-generation": true,
                      "no-indexes": true,
                      "parallel": 3
                    },
                    "source": {
                      "local-storage": {
                        "path": "/tmp/localStorage"
                      }
                    }
                  },
                  "type": "Full"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: docker.io/aerospike/aerospike-kubernetes-operator:4.1.1
          createdAt: "2025-09-15T10:14:05Z"
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          olm.skipRange: '>=4.0.0 <4.1.1'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
          support: Aerospike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeBackup is the Schema for the aerospikebackup API
            displayName: Aerospike Backup
            kind: AerospikeBackup
            name: aerospikebackups.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeBackupService is the Schema for the aerospikebackupservices
              API
            displayName: Aerospike Backup Service
            kind: AerospikeBackupService
            name: aerospikebackupservices.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1
          - description: AerospikeRestore is the Schema for the aerospikerestores
              API
            displayName: Aerospike Restore
            kind: AerospikeRestore
            name: aerospikerestores.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
           -   Aerospike cluster monitoring
           -   Backup and restore Aerospike clusters
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Aerospike
        relatedImages:
        - docker.io/aerospike/aerospike-kubernetes-operator:4.1.1
        version: 4.1.1
      entries:
      - name: aerospike-kubernetes-operator.v4.1.1
        version: 4.1.1
      - name: aerospike-kubernetes-operator.v4.1.0
        version: 4.1.0
      - name: aerospike-kubernetes-operator.v4.0.2
        version: 4.0.2
      - name: aerospike-kubernetes-operator.v4.0.1
        version: 4.0.1
      - name: aerospike-kubernetes-operator.v4.0.0
        version: 4.0.0
      - name: aerospike-kubernetes-operator.v3.4.1
        version: 3.4.1
      - name: aerospike-kubernetes-operator.v3.4.0
        version: 3.4.0
      - name: aerospike-kubernetes-operator.v3.3.1
        version: 3.3.1
      - name: aerospike-kubernetes-operator.v3.3.0
        version: 3.3.0
      - name: aerospike-kubernetes-operator.v3.2.2
        version: 3.2.2
      - name: aerospike-kubernetes-operator.v3.2.1
        version: 3.2.1
      - name: aerospike-kubernetes-operator.v3.2.0
        version: 3.2.0
      - name: aerospike-kubernetes-operator.v3.1.0
        version: 3.1.0
      - name: aerospike-kubernetes-operator.v3.0.0
        version: 3.0.0
      name: stable
    defaultChannel: stable
    packageName: aerospike-kubernetes-operator
    provider:
      name: Aerospike
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: 3scale-community-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: 3scale-community-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.8.2
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Improving [documentation](https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.8.2
        version: 0.8.2
      entries:
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.11
    - currentCSV: 3scale-community-operator.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.9.0
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.9.0/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.9.0/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.9.0
        version: 0.9.0
      entries:
      - name: 3scale-community-operator.v0.9.0
        version: 0.9.0
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.12
    - currentCSV: 3scale-community-operator.v0.10.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "ProductCRName": "product1-sample",
                  "Production": true
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.10.1
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.10.1/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Improving [documentation](https://github.com/3scale/3scale-operator/blob/master/doc/operator-user-guide.md)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.10.1/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.10.1
        version: 0.10.1
      entries:
      - name: 3scale-community-operator.v0.10.1
        version: 0.10.1
      - name: 3scale-community-operator.v0.9.0
        version: 0.9.0
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.13
    defaultChannel: threescale-2.13
    packageName: 3scale-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: https://www.intel.com
    name: ipm-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ipm-controller.v25.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ipm.intel.com/v1",
                "kind": "CPUPowerProv",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ipm-controller",
                    "node-label": "ipm-provisioning"
                  },
                  "name": "cpupowerprov-sample"
                },
                "spec": {
                  "cstates": {
                    "cores": {
                      "5": {
                        "C1": true,
                        "C1E": true,
                        "C6": false
                      },
                      "9,12": {
                        "C1": true,
                        "C6": true
                      }
                    },
                    "idleGovernor": "ladder"
                  },
                  "pstates": {
                    "cores": {
                      "5": {
                        "energyPerformanceBias": "normal",
                        "energyPerformancePreference": "performance",
                        "freqGovernor": "performance",
                        "maxFreqMHz": "3500",
                        "minFreqMHz": "800"
                      },
                      "9,12": {
                        "freqGovernor": "powersave",
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    },
                    "global": {
                      "conservative": {
                        "downThreshold": 20,
                        "freqStep": 5,
                        "ignoreNiceLoad": 0,
                        "samplingDownFactor": 1,
                        "samplingRate": 8888,
                        "upThreshold": 88
                      },
                      "maxPerfPct": 95,
                      "minPerfPct": 45,
                      "ondemand": {
                        "ignoreNiceLoad": 0,
                        "ioIsBusy": 1,
                        "powersaveBias": 0,
                        "samplingDownFactor": 1,
                        "samplingRate": 8888,
                        "upThreshold": 88
                      },
                      "schedutil": {
                        "rateLimitUs": 5000
                      },
                      "turbo": true
                    }
                  },
                  "uncoreFreqCPU": {
                    "cpu0": [
                      {
                        "maxFreqMHz": "1600",
                        "minFreqMHz": "800"
                      }
                    ],
                    "cpu1": [
                      {
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    ]
                  },
                  "uncoreFreqIO": {
                    "cpu0": [
                      {
                        "maxFreqMHz": "1600",
                        "minFreqMHz": "800"
                      }
                    ],
                    "cpu1": [
                      {
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/ipm-controller@sha256:32135c888f86a2c725d093ada9a59d222229bee5640d37f6e19a9e9812b4b67a
          createdAt: "2025-08-07T13:13:25Z"
          description: IPM Controller manages resources via REST API interactions.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.operatorframework.io/required-sccs: '["privileged"]'
          olm.securityContextConstraints.required: privileged
          operators.openshift.io/infrastructure-features: RestrictedNetwork
          operators.openshift.io/valid-subscription: Intel OBL Limited Distribution
            (Commercial Use) License Agreement [v2022.12.20]
          operators.operatorframework.io.bundle.package.v1: ipm-controller
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Intel® Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CPUPowerProv is the Schema for the cpupowerprovs API
            displayName: CPUPower Prov
            kind: CPUPowerProv
            name: cpupowerprovs.ipm.intel.com
            version: v1
        description: Infrastructure Power Manager (IPM) governs platform power technologies
          based on input from application workloads, the platform itself and customer
          input
        displayName: Intel® Infrastructure Power Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - intel
        - infrastructure
        - power
        - manager
        - management
        links:
        - name: IPM
          url: https://builders.intel.com/intel-technologies/power-management
        maintainers:
        - email: infrastructure.power.manager.operator@intel.com
          name: Infrastructure Power Manager Operator
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Intel® Corporation
          url: https://www.intel.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.13.0@sha256:471549d9cc07fcec0f285a05745b3d39a48e9a503b01b4047156a86ec1a30c00
        - registry.connect.redhat.com/intel/ipm-controller@sha256:32135c888f86a2c725d093ada9a59d222229bee5640d37f6e19a9e9812b4b67a
        - registry.connect.redhat.com/intel/ipm-agent@sha256:54de45cdc2e34768f6cdbf1cdee219406684143dddc4d236877709dc2536fe9c
        - registry.connect.redhat.com/intel/ipm-watcher@sha256:51eb773d5a4d1f401d4cbf2b42b8df4d0359beb85d5e274b8392dd7c5165e6ff
        version: 25.7.0
      entries:
      - name: ipm-controller.v25.7.0
        version: 25.7.0
      name: stable
    defaultChannel: stable
    packageName: ipm-controller
    provider:
      name: Intel® Corporation
      url: https://www.intel.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: ""
    name: nim-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nim-operator-certified.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMCache",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "source": {
                    "ngc": {
                      "modelPuller": "nvcr.io/nim/meta/llama-3.2-1b-instruct@sha256:1ff21ad32124d2382f8c733879394e7849cc3df5172651a24fcb671ac4fe77c6",
                      "pullSecret": "ngc-secret",
                      "authSecret": "ngc-api-secret",
                      "model": {
                        "lora": false,
                        "precision": "fp16",
                        "engine": "tensorrt_llm",
                        "qosProfile": "throughput",
                        "gpus": [
                          {
                            "product": "a100",
                            "ids": []
                          }
                        ]
                      }
                    }
                  },
                  "storage": {
                    "pvc": {
                      "create": true,
                      "storageClass": "local-path",
                      "size": "50Gi",
                      "volumeAccessMode": "ReadWriteOnce"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMService",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "image": {
                    "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                    "tag": "1.12.0",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [
                      "ngc-secret"
                    ]
                  },
                  "authSecret": "ngc-api-secret",
                  "storage": {
                    "nimCache": {
                      "name": "meta-llama3-1b-instruct",
                      "profile": ""
                    }
                  },
                  "scale": {
                    "enabled": false,
                    "hpa": {
                      "minReplicas": 1,
                      "maxReplicas": 10
                    }
                  },
                  "resources": {
                    "limits": {
                      "nvidia.com/gpu": 1
                    }
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMPipeline",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "k8s-nim-operator",
                    "app.kubernetes.io/managed-by": "k8s-nim-operator"
                  },
                  "name": "nimpipeline-sample"
                },
                "spec": {
                  "services": [
                    {
                      "name": "meta-llama3-1b-instruct",
                      "enabled": true,
                      "spec": {
                        "image": {
                          "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                          "tag": "1.12.0",
                          "pullPolicy": "IfNotPresent",
                          "pullSecrets": [
                            "ngc-secret"
                          ]
                        },
                        "authSecret": "ngc-api-secret",
                        "storage": {
                          "nimCache": {
                            "name": "meta-llama3-1b-instruct",
                            "profile": ""
                          }
                        },
                        "replicas": 1,
                        "resources": {
                          "limits": {
                            "nvidia.com/gpu": 1
                          }
                        },
                        "expose": {
                          "service": {
                            "type": "ClusterIP",
                            "port": 8000
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMBuild",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "nimCache": {
                    "name": "meta-llama3-1b-instruct",
                    "profile": "7cc8597690a35aba19a3636f35e7f1c7e7dbc005fe88ce9394cad4a4adeed414"
                  },
                  "image": {
                    "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                    "tag": "1.12.0",
                    "pullSecrets": [
                     "ngc-secret"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoCustomizer",
                "metadata": {
                  "name": "nemocustomizer-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "scheduler": {
                    "type": "volcano"
                  },
                  "wandb": {
                    "secretName": "wandb-secret",
                    "apiKeyKey": "apiKey",
                    "encryptionKey": "encryptionKey"
                  },
                  "otel": {
                    "enabled": true,
                    "exporterOtlpEndpoint": "http://customizer-otel-opentelemetry-collector.nemo.svc.cluster.local:4317"
                  },
                  "databaseConfig": {
                    "credentials": {
                      "user": "ncsuser",
                      "secretName": "customizer-pg-existing-secret",
                      "passwordKey": "password"
                    },
                    "host": "customizer-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "ncsdb"
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/customizer-api",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "entitystore": {
                    "endpoint": "http://nemoentitystore-sample.nemo.svc.cluster.local:8000"
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000"
                  },
                  "mlflow": {
                    "endpoint": "http://mlflow-tracking.nemo.svc.cluster.local:80"
                  },
                  "nemoDatastoreTools": {
                    "image": "nvcr.io/nvidia/nemo-microservices/nds-v2-huggingface-cli@sha256:7958e5244217de8af0ed723dfe7ea8e291b3afbb2a0aa73f5c27b218a1149b11"
                  },
                  "modelDownloadJobs": {
                    "image": "nvcr.io/nvidia/nemo-microservices/customizer-api@sha256:c590fa19121af2aac0379933deace3991f27c3ef88a8043560d6e3a172273b0b",
                    "ngcAPISecret": {
                      "name": "ngc-api-secret",
                      "key": "NGC_API_KEY"
                    },
                    "securityContext": {
                      "fsGroup": 1000,
                      "runAsNonRoot": true,
                      "runAsUser": 1000,
                      "runAsGroup": 1000
                    },
                    "ttlSecondsAfterFinished": 600,
                    "pollIntervalSeconds": 15
                  },
                  "modelConfig": {
                    "name": "nemo-model-config"
                  },
                  "trainingConfig": {
                    "configMap": {
                      "name": "nemo-training-config"
                    },
                    "modelPVC": {
                      "create": true,
                      "name": "finetuning-ms-models-pvc",
                      "storageClass": "",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": 50
                    },
                    "workspacePVC": {
                      "storageClass": "local-path",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": 10,
                      "mountPath": "/pvc/workspace"
                    },
                    "image": {
                      "repository": "nvcr.io/nvidia/nemo-microservices/customizer",
                      "tag": "25.08"
                    },
                    "env": [
                      { "name": "LOG_LEVEL", "value": "INFO" }
                    ],
                    "networkConfig": [
                      { "name": "NCCL_IB_SL", "value": "0" },
                      { "name": "NCCL_IB_TC", "value": "41" },
                      { "name": "NCCL_IB_QPS_PER_CONNECTION", "value": "4" },
                      { "name": "UCX_TLS", "value": "TCP" },
                      { "name": "UCX_NET_DEVICES", "value": "eth0" },
                      { "name": "HCOLL_ENABLE_MCAST_ALL", "value": "0" },
                      { "name": "NCCL_IB_GID_INDEX", "value": "3" }
                    ],
                    "ttlSecondsAfterFinished": 3600,
                    "timeout": 3600,
                    "tolerations": [
                      {
                        "key": "nvidia.com/gpu",
                        "operator": "Exists",
                        "effect": "NoSchedule"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoGuardrail",
                "metadata": {
                  "name": "nemoguardrails-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "configStore": {
                    "pvc": {
                      "name": "pvc-guardrail-config",
                      "create": true,
                      "storageClass": "",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": "1Gi"
                    }
                  },
                  "nimEndpoint": {
                    "baseURL": "http://meta-llama3-1b-instruct.nemo.svc.cluster.local:8000/v1"
                  },
                  "databaseConfig": {
                    "host": "guardrail-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "guardraildb",
                    "credentials": {
                      "user": "guardrailuser",
                      "secretName": "guardrail-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/guardrails",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [
                      "ngc-secret"
                    ]
                  },
                  "metrics": {
                    "serviceMonitor": {}
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "ephemeral-storage": "10Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoEvaluator",
                "metadata": {
                  "name": "nemoevaluator-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "evaluationImages": {
                    "bigcodeEvalHarness": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-bigcode@sha256:62f74c32d1ed07925ddb906de8975025bf63d73e82e7faefb840339aaf21a9dc",
                    "lmEvalHarness": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-lm-eval-harness@sha256:58e7c3da7f1edfed2ff508fa8abab7660366c2c17bb55d17aa68159d6e16be6d",
                    "similarityMetrics": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-custom-eval@sha256:eb51edf668481d1f5d2b100c850958dadf184fd8779433e45da40fa2a35ec4b9",
                    "llmAsJudge": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055",
                    "mtBench": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055",
                    "retriever": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-retriever@sha256:cb99fb3ea5b4e9989832679d4c6028573c28b99e43744e008ac5c9748b7dd250",
                    "rag": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-rag@sha256:b5e35cf277185e0901a4bbba35780b3673be71ae2c5e8efcdbab7710a8a6f783",
                    "bfcl": "nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-bfcl@sha256:95f45f66140d481d544b36b16bc3de7ad5e090d76225fd73ed70461b9bd9907c",
                    "agenticEval": "nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-agentic-eval@sha256:b38bbc5e61df0930204efa096cf4f40db4655b92d03741190513b019d67fc7dc"
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/evaluator",
                    "tag": "25.06",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "argoWorkflows": {
                    "endpoint": "https://argo-workflows-server.nemo.svc.cluster.local:2746",
                    "serviceAccount": "argo-workflows-executor"
                  },
                  "vectorDB": {
                    "endpoint": "http://milvus.nemo.svc.cluster.local:19530"
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000/v1/hf"
                  },
                  "entitystore": {
                    "endpoint": "http://nemoentitystore-sample.nemo.svc.cluster.local:8000"
                  },
                  "databaseConfig": {
                    "host": "evaluator-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "evaldb",
                    "credentials": {
                      "user": "evaluser",
                      "secretName": "evaluator-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "otel": {
                    "enabled": true,
                    "exporterOtlpEndpoint": "http://evaluator-otel-opentelemetry-collector.nemo.svc.cluster.local:4317"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoEntitystore",
                "metadata": {
                  "name": "nemoentitystore-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/entity-store",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "databaseConfig": {
                    "databaseName": "nesdb",
                    "host": "entity-store-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "credentials": {
                      "user": "nesuser",
                      "secretName": "entity-store-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000"
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoDatastore",
                "metadata": {
                  "name": "nemodatastore-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "secrets": {
                    "datastoreConfigSecret": "nemo-ms-nemo-datastore",
                    "datastoreInitSecret": "nemo-ms-nemo-datastore-init",
                    "datastoreInlineConfigSecret": "nemo-ms-nemo-datastore-inline-config",
                    "giteaAdminSecret": "gitea-admin-credentials",
                    "lfsJwtSecret": "nemo-ms-nemo-datastore--lfs-jwt"
                  },
                  "databaseConfig": {
                    "credentials": {
                      "user": "ndsuser",
                      "secretName": "datastore-pg-existing-secret",
                      "passwordKey": "password"
                    },
                    "host": "datastore-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "ndsdb"
                  },
                  "pvc": {
                    "name": "pvc-shared-data",
                    "create": true,
                    "storageClass": "",
                    "volumeAccessMode": "ReadWriteOnce",
                    "size": "10Gi"
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/datastore",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "replicas": 1,
                  "resources": {
                    "requests": {
                      "memory": "256Mi",
                      "cpu": "500m"
                    },
                    "limits": {
                      "memory": "512Mi",
                      "cpu": "1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: nvcr.io/nvidia/cloud-native/k8s-nim-operator@sha256:f42dd5c3dc269553ee46725ed4b1ab2d2f253ef06117f1d9d52faaa22d8fe60d
          createdAt: Wed Sep 10 05:49:48 PDT 2025
          description: An Operator for deployment and maintenance of NVIDIA NIMs and
            NeMo microservices in a Kubernetes environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          k8sMinVersion: 1.26.0
          repository: github.com/NVIDIA/k8s-nim-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NIM Cache
            displayName: NIMCache
            kind: NIMCache
            name: nimcaches.apps.nvidia.com
            version: v1alpha1
          - description: NIM Service
            displayName: NIMService
            kind: NIMService
            name: nimservices.apps.nvidia.com
            version: v1alpha1
          - description: NIM Pipeline
            displayName: NIMPipeline
            kind: NIMPipeline
            name: nimpipelines.apps.nvidia.com
            version: v1alpha1
          - description: NIM Build
            displayName: NIMBuild
            kind: NIMBuild
            name: nimbuilds.apps.nvidia.com
            version: v1alpha1
          - description: NEMO guardrail Service
            displayName: NemoGuardrail
            kind: NemoGuardrail
            name: nemoguardrails.apps.nvidia.com
            version: v1alpha1
          - description: NEMO datastore Service
            displayName: NemoDatastore
            kind: NemoDatastore
            name: nemodatastores.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Eval Service
            displayName: NemoEvaluator
            kind: NemoEvaluator
            name: nemoevaluators.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Customizer Service
            displayName: NemoCustomizer
            kind: NemoCustomizer
            name: nemocustomizers.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Entitystore Service
            displayName: NemoEntitystore
            kind: NemoEntitystore
            name: nemoentitystores.apps.nvidia.com
            version: v1alpha1
        description: |-
          ## NVIDIA NIM Operator
          The NVIDIA NIM Operator enables Kubernetes cluster administrators to operate the software components and services necessary to run NVIDIA NIMs in various domains such as reasoning, retrieval, speech, and biology. Additionally, it allows the use of NeMo Microservices to fine-tune, evaluate, or apply guardrails to your models.
          The Operator manages the life cycle of the following microservices and the models they use:
          ### NVIDIA NIM models, such as:
          * Reasoning LLMs
          * Retrieval - Embedding, Reranking, etc.
          * Speech
          * Biology
          ### NeMo core microservices:
          * NeMo Customizer
          * NeMo Evaluator
          * NeMo Guardrails
          ### NeMo platform component microservices:
          * NeMo Data Store
          * NeMo Entity Store
        displayName: The NVIDIA NIM Operator for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        links:
        - name: NVIDIA NIM
          url: https://docs.nvidia.com/nim/index.html
        - name: Documentation
          url: https://docs.nvidia.com/nim-operator/latest
        maintainers:
        - email: operator_feedback@nvidia.com
          name: shivamerla
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: NVIDIA
        relatedImages:
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-retriever@sha256:cb99fb3ea5b4e9989832679d4c6028573c28b99e43744e008ac5c9748b7dd250
        - nvcr.io/nvidia/nemo-microservices/customizer-api@sha256:c590fa19121af2aac0379933deace3991f27c3ef88a8043560d6e3a172273b0b
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-lm-eval-harness@sha256:58e7c3da7f1edfed2ff508fa8abab7660366c2c17bb55d17aa68159d6e16be6d
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-custom-eval@sha256:eb51edf668481d1f5d2b100c850958dadf184fd8779433e45da40fa2a35ec4b9
        - nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-bfcl@sha256:95f45f66140d481d544b36b16bc3de7ad5e090d76225fd73ed70461b9bd9907c
        - nvcr.io/nvidia/cloud-native/k8s-nim-operator@sha256:f42dd5c3dc269553ee46725ed4b1ab2d2f253ef06117f1d9d52faaa22d8fe60d
        - nvcr.io/nvidia/nemo-microservices/nds-v2-huggingface-cli@sha256:7958e5244217de8af0ed723dfe7ea8e291b3afbb2a0aa73f5c27b218a1149b11
        - nvcr.io/nim/meta/llama-3.2-1b-instruct@sha256:1ff21ad32124d2382f8c733879394e7849cc3df5172651a24fcb671ac4fe77c6
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-bigcode@sha256:62f74c32d1ed07925ddb906de8975025bf63d73e82e7faefb840339aaf21a9dc
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-rag@sha256:b5e35cf277185e0901a4bbba35780b3673be71ae2c5e8efcdbab7710a8a6f783
        - nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-agentic-eval@sha256:b38bbc5e61df0930204efa096cf4f40db4655b92d03741190513b019d67fc7dc
        version: 3.0.0
      entries:
      - name: nim-operator-certified.v3.0.0
        version: 3.0.0
      - name: nim-operator-certified.v2.0.2
        version: 2.0.2
      - name: nim-operator-certified.v2.0.1
        version: 2.0.1
      - name: nim-operator-certified.v2.0.0
        version: 2.0.0
      - name: nim-operator-certified.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: stable
    packageName: nim-operator-certified
    provider:
      name: NVIDIA
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-druid-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: druid-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:909a03d95f8173c418f6547092421a5531ff8957aa4319ce5fc890f32fe1f7be
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:909a03d95f8173c418f6547092421a5531ff8957aa4319ce5fc890f32fe1f7be
        version: 24.11.1
      entries:
      - name: druid-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: druid-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:253239268b8d32dbea10233fe6ab202bbe8a074c369a929630681dea37935383
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:253239268b8d32dbea10233fe6ab202bbe8a074c369a929630681dea37935383
        version: 25.3.0
      entries:
      - name: druid-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: druid-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
        version: 25.7.0
      entries:
      - name: druid-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: druid-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
        version: 25.7.0
      entries:
      - name: druid-operator.v25.7.0
        version: 25.7.0
      - name: druid-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-druid-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: eap
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eap-operator.v3.2.2
      currentCSVDesc:
        annotations:
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "false"
          containerImage: registry.redhat.io/jboss-eap/eap-rhel9-operator@sha256:2361e73b8c517177e80f60d204996ebf65b166e4e8b60b329323a2b409c54f6c
          createdAt: "2025-07-10T14:38:00Z"
          description: Operator that creates and manages Java applications running
            on JBoss EAP.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat JBoss Enterprise Application
            Platform", "Red Hat Runtimes", "Red Hat Integration", "Red Hat Process
            Automation"]'
          repository: https://github.com/wildfly/wildfly-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An application running on EAP application runtime.
            displayName: WildFlyServer
            kind: WildFlyServer
            name: wildflyservers.wildfly.org
            version: v1alpha1
        description: "# Red Hat JBoss Enterprise Application Platform (JBoss EAP)\n\nRed
          Hat® JBoss® Enterprise Application Platform (JBoss EAP) delivers enterprise-grade
          security, performance, and scalability in any environment.  Some features
          of JBoss EAP are:\n\n**Optimized for cloud and containers**: with simplified
          deployment and full Jakarta EE performance for applications in any environment.\n\n**Lightweight,
          flexible architecture**: which is well-suited for microservices and traditional
          applications.\n\n**Support and standardize microservices development**:
          With the JBoss Enterprise Application Platform expansion pack that supports
          Eclipse MicroProfile application programming interfaces (APIs).\n\n# JBoss
          EAP Operator\nOperators are software components that extend the Kubernetes
          API to customize how to manage applications and their components.\n\nThe
          JBoss EAP Operator facilitates the deployment and management of JBoss EAP
          on container orchestration platforms such as Red Hat OpenShift. The operator
          acts as an automated management tool, simplifying the process of provisioning,
          scaling, and updating JBoss EAP applications.\n\n# JBoss EAP Operator capabilities\nThe
          JBoss EAP Operator deploys and manages multiple instances of the JBoss EAP
          Java application in a Red Hat OpenShift cluster. \n\n**Simplifies management**:
          Creates, configures, manages, and seamlessly upgrades instances of complex
          stateful applications. No more manual configuration, making management a
          breeze.\n\n**Efficient scaling**: Effortlessly scales the instances of your
          JBoss EAP Java application based on demand.\n\n**Faster updates**: Seamlessly
          updates, enabling you to roll out new operator features and bug fixes without
          downtime\n\n**Safe transaction recovery**: Provides safe transaction recovery
          in your application cluster by verifying all transactions are completed
          before scaling down the replicas and marking a pod as clean for termination.\n\n#
          Operator installation and configuration\nLet's walk through the steps to
          install the JBoss EAP Operator on your OpenShift Cluster and deploy your
          first example application:\n\n**Step 1: Install JBoss EAP from OperatorHub**\n1.
          Click Install in the JBoss EAP operator window.\n2. In the Install Operator
          window, for Installation mode, select All namespaces on the cluster The
          operator is available in openshift-operators namespace.\n\n**Step 2: Build
          an application image**\n\nThe JBoss EAP Operator deploys an application
          image. If you do not have your JBoss EAP application image built yet, you
          can use the following example to build the image using Helm Charts.\n\n1.
          In the main navigation, click the dropdown menu (perspective switcher) and
          select Developer.\n2. Click Add > Helm Chart.\n3. Search for \"JBoss EAP
          8\" (Provided by Red Hat) and create a new Helm Chart release.\n4. On the
          Release name text box, type “eap-application”. \n5. Switch to the YAML view
          and replace the existing code with the following:\n\n    ```\n    build:\n
          \     uri: https://github.com/jboss-eap-up-and-running/eap8-getting-started\n
          \   deploy:\n      enabled: false\n    ```\n\n6. Click Create to build the
          application image.\n\n    The application image is created with the name
          eap-application:latest.\n\n**Step 3: Deploy your application with JBoss
          EAP Operator**\n\n1. In the main navigation, click the dropdown menu (perspective
          switcher) and select Developer.\n2. Click Add > Operator Backed > WildFlyServer
          > Create.\n3. Switch to the YAML view and replace the existing code with
          the following:\n\n    ```\n    apiVersion: wildfly.org/v1alpha1\n    kind:
          WildFlyServer\n    metadata:\n      name: eap-application\n    spec:\n      applicationImage:
          eap-application:latest\n      replicas: 1\n    ```\n\n 4. Click Create to
          deploy the application with JBoss EAP Operator.\n"
        displayName: JBoss EAP
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eap
        - java
        - open source
        - application runtime
        links:
        - name: JBoss EAP Homepage
          url: https://www.redhat.com/en/technologies/jboss-middleware/application-platform
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_jboss_enterprise_application_platform/
        - name: Blog
          url: https://developers.redhat.com/blog/
        - name: Quickstart
          url: https://github.com/jboss-developer/jboss-eap-quickstarts
        - name: Red Hat Customer Portal
          url: https://access.redhat.com/support
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat, Inc.
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/jboss-eap/eap-rhel9-operator@sha256:2361e73b8c517177e80f60d204996ebf65b166e4e8b60b329323a2b409c54f6c
        version: 3.2.2
      entries:
      - name: eap-operator.v3.2.2
        version: 3.2.2
      - name: eap-operator.v3.2.1
        version: 3.2.1
      - name: eap-operator.v3.2.0
        version: 3.2.0
      - name: eap-operator.v3.1.9
        version: 3.1.9
      - name: eap-operator.v3.1.8
        version: 3.1.8
      - name: eap-operator.v3.1.7
        version: 3.1.7
      - name: eap-operator.v3.1.6
        version: 3.1.6
      - name: eap-operator.v3.1.5
        version: 3.1.5
      - name: eap-operator.v3.1.4
        version: 3.1.4
      - name: eap-operator.v3.1.3
        version: 3.1.3
      - name: eap-operator.v3.1.2
        version: 3.1.2
      - name: eap-operator.v3.1.1
        version: 3.1.1
      - name: eap-operator.v3.1.0
        version: 3.1.0
      - name: eap-operator.v3.0.2
        version: 3.0.2
      - name: eap-operator.v3.0.1
        version: 3.0.1
      - name: eap-operator.v3.0.0
        version: 3.0.0
      - name: eap-operator.v2.4.3
        version: 2.4.3
      - name: eap-operator.v2.4.2
        version: 2.4.2
      - name: eap-operator.v2.4.1
        version: 2.4.1
      - name: eap-operator.v2.4.0
        version: 2.4.0
      - name: eap-operator.v2.3.10
        version: 2.3.10
      - name: eap-operator.v2.3.9
        version: 2.3.9
      - name: eap-operator.v2.3.8
        version: 2.3.8
      - name: eap-operator.v2.3.7
        version: 2.3.7
      - name: eap-operator.v2.3.6
        version: 2.3.6
      - name: eap-operator.v2.3.5
        version: 2.3.5
      - name: eap-operator.v2.3.4
        version: 2.3.4
      - name: eap-operator.v2.3.3
        version: 2.3.3
      - name: eap-operator.v2.3.2
        version: 2.3.2
      - name: eap-operator.v2.3.1
        version: 2.3.1
      - name: eap-operator.v2.3.0
        version: 2.3.0
      - name: eap-operator.v2.2.2
        version: 2.2.2
      - name: eap-operator.v2.2.1
        version: 2.2.1
      - name: eap-operator.v2.2.0
        version: 2.2.0
      name: stable
    defaultChannel: stable
    packageName: eap
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lifecycle-agent
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lifecycle-agent.v4.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "ImageBasedUpgrade",
                "metadata": {
                  "name": "upgrade"
                },
                "spec": {
                  "autoRollbackOnFailure": {},
                  "extraManifests": [
                    {
                      "name": "sno-extramanifests",
                      "namespace": "openshift-lifecycle-agent"
                    }
                  ],
                  "oadpContent": [
                    {
                      "name": "oadp-cm-sno-backup",
                      "namespace": "openshift-adp"
                    }
                  ],
                  "seedImageRef": {
                    "image": "quay.io/xyz",
                    "version": "4.16.0"
                  },
                  "stage": "Idle"
                }
              },
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "SeedGenerator",
                "metadata": {
                  "name": "seedimage"
                },
                "spec": {
                  "seedImage": "quay.io/xyz"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:6c603c89401823086d7a02d7ad221da44d8c84000ba63dd30675ae0cbcacc95f
          description: The Lifecycle Agent for OpenShift provides local lifecycle
            management services for Single Node Openshift (SNO) clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.3'
          operatorframework.io/suggested-namespace: openshift-lifecycle-agent
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-lifecycle-agent",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/lifecycle-agent
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ImageBasedUpgrade is the Schema for the ImageBasedUpgrades
              API
            displayName: Image-based Cluster Upgrade
            kind: ImageBasedUpgrade
            name: imagebasedupgrades.lca.openshift.io
            version: v1
          - description: SeedGenerator is the Schema for the seedgenerators API
            displayName: Seed Generator
            kind: SeedGenerator
            name: seedgenerators.lca.openshift.io
            version: v1
        description: "# Lifecycle Agent for OpenShift\nThe Lifecycle Agent for OpenShift
          provides local lifecycle management services \nfor Single Node Openshift
          (SNO) clusters.\n\n## Where to find more information\nYou can find additional
          guidance in the [agent repository](https://github.com/openshift-kni/lifecycle-agent).\n"
        displayName: Lifecycle Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - upgrade
        links:
        - name: Lifecycle Agent
          url: https://lifecycle-agent.domain
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift4/recert-rhel9@sha256:c91def5d3da4e777d2555c354e39bf021d9f499993f86614969dee5ecc00fc49
        - registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:6c603c89401823086d7a02d7ad221da44d8c84000ba63dd30675ae0cbcacc95f
        version: 4.16.3
      entries:
      - name: lifecycle-agent.v4.16.3
        version: 4.16.3
      - name: lifecycle-agent.v4.16.2
        version: 4.16.2
      - name: lifecycle-agent.v4.16.1
        version: 4.16.1
      - name: lifecycle-agent.v4.16.0
        version: 4.16.0
      name: "4.16"
    - currentCSV: lifecycle-agent.v4.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "ImageBasedUpgrade",
                "metadata": {
                  "name": "upgrade"
                },
                "spec": {
                  "autoRollbackOnFailure": {},
                  "extraManifests": [
                    {
                      "name": "sno-extramanifests",
                      "namespace": "openshift-lifecycle-agent"
                    }
                  ],
                  "oadpContent": [
                    {
                      "name": "oadp-cm-sno-backup",
                      "namespace": "openshift-adp"
                    }
                  ],
                  "seedImageRef": {
                    "image": "quay.io/xyz",
                    "version": "4.16.0"
                  },
                  "stage": "Idle"
                }
              },
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "SeedGenerator",
                "metadata": {
                  "name": "seedimage"
                },
                "spec": {
                  "seedImage": "quay.io/xyz"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:6c603c89401823086d7a02d7ad221da44d8c84000ba63dd30675ae0cbcacc95f
          description: The Lifecycle Agent for OpenShift provides local lifecycle
            management services for Single Node Openshift (SNO) clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.3'
          operatorframework.io/suggested-namespace: openshift-lifecycle-agent
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-lifecycle-agent",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/lifecycle-agent
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ImageBasedUpgrade is the Schema for the ImageBasedUpgrades
              API
            displayName: Image-based Cluster Upgrade
            kind: ImageBasedUpgrade
            name: imagebasedupgrades.lca.openshift.io
            version: v1
          - description: SeedGenerator is the Schema for the seedgenerators API
            displayName: Seed Generator
            kind: SeedGenerator
            name: seedgenerators.lca.openshift.io
            version: v1
        description: "# Lifecycle Agent for OpenShift\nThe Lifecycle Agent for OpenShift
          provides local lifecycle management services \nfor Single Node Openshift
          (SNO) clusters.\n\n## Where to find more information\nYou can find additional
          guidance in the [agent repository](https://github.com/openshift-kni/lifecycle-agent).\n"
        displayName: Lifecycle Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - upgrade
        links:
        - name: Lifecycle Agent
          url: https://lifecycle-agent.domain
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:6c603c89401823086d7a02d7ad221da44d8c84000ba63dd30675ae0cbcacc95f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift4/recert-rhel9@sha256:c91def5d3da4e777d2555c354e39bf021d9f499993f86614969dee5ecc00fc49
        version: 4.16.3
      entries:
      - name: lifecycle-agent.v4.16.3
        version: 4.16.3
      - name: lifecycle-agent.v4.16.2
        version: 4.16.2
      - name: lifecycle-agent.v4.16.1
        version: 4.16.1
      - name: lifecycle-agent.v4.16.0
        version: 4.16.0
      name: stable
    defaultChannel: stable
    packageName: lifecycle-agent
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-opensearchservice-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-opensearchservice-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opensearchservice.services.k8s.aws/v1alpha1",
                "kind": "Domain",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/opensearchservice-controller:1.1.1
          createdAt: "2025-09-19T18:05:50Z"
          description: AWS OpenSearch Service controller is a service controller for
            managing OpenSearch Service resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Domain represents the state of an AWS opensearchservice Domain
              resource.
            displayName: Domain
            kind: Domain
            name: domains.opensearchservice.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage OpenSearch Service resources in AWS from within your Kubernetes cluster.

          **About Amazon OpenSearch Service**

          Amazon OpenSearch Service (successor to Amazon Elasticsearch Service) is a managed service that makes it easy to deploy, operate, and scale OpenSearch clusters in the AWS Cloud. Amazon OpenSearch Service supports OpenSearch and legacy Elasticsearch OSS. When you create a cluster, you have the option of which search engine to use. OpenSearch Service offers broad compatibility with Elasticsearch OSS 7.10, the final open source version of the software. For information about what changed with the recent service renaming, and the actions you might need to take, see [Amazon OpenSearch Service - Summary of changes](https://docs.aws.amazon.com/opensearch-service/latest/developerguide/rename.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon OpenSearch Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opensearchservice
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon OpenSearch Service Developer Resources
          url: https://aws.amazon.com/opensearch-service/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: opensearch service maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/opensearchservice-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-opensearchservice-controller.v1.1.1
        version: 1.1.1
      - name: ack-opensearchservice-controller.v1.1.0
        version: 1.1.0
      - name: ack-opensearchservice-controller.v1.0.15
        version: 1.0.15
      - name: ack-opensearchservice-controller.v1.0.12
        version: 1.0.12
      - name: ack-opensearchservice-controller.v1.0.10
        version: 1.0.10
      - name: ack-opensearchservice-controller.v1.0.9
        version: 1.0.9
      - name: ack-opensearchservice-controller.v1.0.8
        version: 1.0.8
      - name: ack-opensearchservice-controller.v1.0.7
        version: 1.0.7
      - name: ack-opensearchservice-controller.v1.0.6
        version: 1.0.6
      - name: ack-opensearchservice-controller.v1.0.5
        version: 1.0.5
      - name: ack-opensearchservice-controller.v1.0.4
        version: 1.0.4
      - name: ack-opensearchservice-controller.v1.0.2
        version: 1.0.2
      - name: ack-opensearchservice-controller.v1.0.1
        version: 1.0.1
      - name: ack-opensearchservice-controller.v1.0.0
        version: 1.0.0
      - name: ack-opensearchservice-controller.v0.0.28
        version: 0.0.28
      - name: ack-opensearchservice-controller.v0.0.27
        version: 0.0.27
      - name: ack-opensearchservice-controller.v0.0.26
        version: 0.0.26
      - name: ack-opensearchservice-controller.v0.0.24
        version: 0.0.24
      - name: ack-opensearchservice-controller.v0.0.23
        version: 0.0.23
      - name: ack-opensearchservice-controller.v0.0.19
        version: 0.0.19
      - name: ack-opensearchservice-controller.v0.0.17
        version: 0.0.17
      - name: ack-opensearchservice-controller.v0.0.15
        version: 0.0.15
      - name: ack-opensearchservice-controller.v0.0.14
        version: 0.0.14
      - name: ack-opensearchservice-controller.v0.0.13
        version: 0.0.13
      - name: ack-opensearchservice-controller.v0.0.12
        version: 0.0.12
      - name: ack-opensearchservice-controller.v0.0.11
        version: 0.0.11
      - name: ack-opensearchservice-controller.v0.0.10
        version: 0.0.10
      - name: ack-opensearchservice-controller.v0.0.9
        version: 0.0.9
      - name: ack-opensearchservice-controller.v0.0.8
        version: 0.0.8
      - name: ack-opensearchservice-controller.v0.0.7
        version: 0.0.7
      - name: ack-opensearchservice-controller.v0.0.6
        version: 0.0.6
      - name: ack-opensearchservice-controller.v0.0.5
        version: 0.0.5
      - name: ack-opensearchservice-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-opensearchservice-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
    name: flink-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flink-kubernetes-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"flink.apache.org/v1beta1",
                "kind":"FlinkDeployment",
                "metadata":{
                    "name":"basic-example"
                },
                "spec":{
                    "image":"flink:1.16",
                    "flinkVersion":"v1_16",
                    "flinkConfiguration":{
                        "taskmanager.numberOfTaskSlots":"2"
                    },
                    "serviceAccount":"flink",
                    "jobManager":{
                        "resource":{
                            "memory":"2048m",
                            "cpu":1
                        }
                    },
                    "taskManager":{
                        "resource":{
                            "memory":"2048m",
                            "cpu":1
                        }
                    },
                    "job":{
                        "jarURI":"local:///opt/flink/examples/streaming/StateMachineExample.jar",
                        "parallelism":2,
                        "upgradeMode":"stateless"
                    }
                }
              },
              {
                "apiVersion":"flink.apache.org/v1beta1",
                "kind":"FlinkSessionJob",
                "metadata":{
                  "name":"basic-session-job-example2"
                },
                "spec":{
                  "deploymentName":"basic-session-cluster",
                  "job":{
                      "jarURI":"https://repo1.maven.org/maven2/org/apache/flink/flink-examples-streaming_2.12/1.16.1/flink-examples-streaming_2.12-1.16.1.jar",
                      "parallelism":2,
                      "upgradeMode":"stateless",
                      "entryClass":"org.apache.flink.streaming.examples.statemachine.StateMachineExample"
                  }
                }
              },
              {
                "apiVersion": "flink.apache.org/v1beta1",
                "kind": "FlinkStateSnapshot",
                "metadata": {
                  "name": "example-savepoint"
                },
                "spec": {
                  "savepoint": {
                    "alreadyExists": true,
                    "disposeOnDelete": false,
                    "path": "/flink-data/savepoints/savepoint-98682a-21931bf10c39"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "basic-example": {
                "description": "Example flink deployment"
              },
              "basic-session-job-example2": {
                "description": "Example flink session job"
              }
            }
          capabilities: Deep Insights
          categories: Big Data, Streaming & Messaging
          containerImage: ghcr.io/apache/flink-kubernetes-operator:1.12.0
          createdAt: "2025-06-23 18:42:54"
          description: Flink Kubernetes Operator manages the complete deployment lifecycle
            of Apache Flink applications.
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/apache/flink-kubernetes-operator.git
          support: Apache Software Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The FlinkDeployment CR defines Flink Application and Session
              cluster deployments
            displayName: Flink Deployment
            kind: FlinkDeployment
            name: flinkdeployments.flink.apache.org
            version: v1beta1
          - description: The FlinkSessionJob CR defines the session job on the Session
              cluster and each Session cluster can run multiple FlinkSessionJob.
            displayName: Flink Session Job
            kind: FlinkSessionJob
            name: flinksessionjobs.flink.apache.org
            version: v1beta1
          - description: The FlinkStateSnapshots CR is used to describe savepoint
              or checkpoint for a Flink job.
            displayName: Flink State Snapshot
            kind: FlinkStateSnapshot
            name: flinkstatesnapshots.flink.apache.org
            version: v1beta1
        description: |-
          Flink Kubernetes Operator acts as a control plane to manage the complete deployment lifecycle of Apache Flink applications.
          ## Installation
          This operator can be installed from [OperatorHub.io](https://operatorhub.io/operator/flink-kubernetes-operator).

          By default, the Flink operator monitors and run Flink applications in the same namespace as the operator.
          To run Flink jobs in another namespace, users are responsible for setting up the RBAC in that namespace.

          For example, users need to do the following for each additional namespace that runs the Flink applications:

          1. Switch to the namespace by running:

              ```sh
              kubectl config set-context --current --namespace=CHANGEIT
              ```

          2. Create the service account, role, and role binding in the namespace using the commands below:

              ```sh
              kubectl apply -f - <<EOF
              apiVersion: v1
              kind: ServiceAccount
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink
              ---
              apiVersion: rbac.authorization.k8s.io/v1
              kind: Role
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink
              rules:
              - apiGroups:
                - ""
                resources:
                - pods
                - configmaps
                verbs:
                - '*'
              - apiGroups:
                - apps
                resources:
                - deployments
                - deployments/finalizers
                verbs:
                - '*'
              ---
              apiVersion: rbac.authorization.k8s.io/v1
              kind: RoleBinding
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink-role-binding
              roleRef:
                apiGroup: rbac.authorization.k8s.io
                kind: Role
                name: flink
              subjects:
              - kind: ServiceAccount
                name: flink
              EOF
              ```

          3. Proceed deploying Flink applications using the the Custom Resource Definition below.

              ```sh
              kubectl create -f https://raw.githubusercontent.com/apache/flink-kubernetes-operator/release-1.12.0/examples/basic.yaml
              ```

          See [Flink custom resources](https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/custom-resource/overview/#overview) for more detail.
        displayName: Flink Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flink
        - streaming
        links:
        - name: Website
          url: https://flink.apache.org/
        - name: Documentation
          url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/
        - name: Mailing list
          url: https://lists.apache.org/list.html?dev@flink.apache.org
        - name: Slack
          url: https://apache-flink.slack.com/join/shared_invite/zt-1llkzbgyt-K2nNGGg88rfsDGLkT09Qzg#/shared-invite/email
        - name: GitHub
          url: https://github.com/apache/flink-kubernetes-operator
        maintainers:
        - email: mbalassi@apache.org
          name: Marton Balassi
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: jbusche@us.ibm.com
          name: James Busche
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Community
          url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
        relatedImages:
        - ghcr.io/apache/flink-kubernetes-operator:1.12.0
        version: 1.12.0
      entries:
      - name: flink-kubernetes-operator.v1.12.0
        version: 1.12.0
      - name: flink-kubernetes-operator.v1.11.0
        version: 1.11.0
      - name: flink-kubernetes-operator.v1.10.0
        version: 1.10.0
      - name: flink-kubernetes-operator.v1.6.0
        version: 1.6.0
      - name: flink-kubernetes-operator.v1.5.0
        version: 1.5.0
      - name: flink-kubernetes-operator.v1.4.0
        version: 1.4.0
      - name: flink-kubernetes-operator.v1.3.1
        version: 1.3.1
      - name: flink-kubernetes-operator.v1.3.0
        version: 1.3.0
      name: alpha
    defaultChannel: alpha
    packageName: flink-kubernetes-operator
    provider:
      name: Community
      url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeRocketCI
      provider-url: https://docs.kuberocketci.io/
    name: edp-keycloak-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: edp-keycloak-operator.v1.19.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "keycloak-sample"
                },
                "spec": {
                  "secret": "my-keycloak-secret",
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakAuthFlow",
                "metadata": {
                  "name": "keycloakauthflow-sample"
                },
                "spec": {
                  "alias": "MyBrowser",
                  "authenticationExecutions": [
                    {
                      "authenticator": "auth-cookie",
                      "priority": 0,
                      "requirement": "ALTERNATIVE"
                    },
                    {
                      "authenticator": "identity-provider-redirector",
                      "authenticatorConfig": {
                        "alias": "my-alias",
                        "config": {
                          "defaultProvider": "my-alias"
                        }
                      },
                      "priority": 1,
                      "requirement": "REQUIRED"
                    }
                  ],
                  "builtIn": false,
                  "description": "browser with idp",
                  "providerId": "basic-flow",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "topLevel": true
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-sample"
                },
                "spec": {
                  "advancedProtocolMappers": true,
                  "clientId": "agocd",
                  "defaultClientScopes": [
                    "argocd_groups"
                  ],
                  "directAccess": true,
                  "public": false,
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-name:client-secret-key",
                  "webUrl": "https://argocd.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClientScope",
                "metadata": {
                  "name": "keycloakclientscope-sample"
                },
                "spec": {
                  "description": "Group Membership",
                  "name": "groups",
                  "protocol": "openid-connect",
                  "protocolMappers": [
                    {
                      "config": {
                        "access.token.claim": "true",
                        "claim.name": "groups",
                        "full.path": "false",
                        "id.token.claim": "true",
                        "userinfo.token.claim": "true"
                      },
                      "name": "groups",
                      "protocol": "openid-connect",
                      "protocolMapper": "oidc-group-membership-mapper"
                    }
                  ],
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "keycloakrealm-sample"
                },
                "spec": {
                  "id": "d1-id-kc-realm-name",
                  "keycloakRef": {
                    "kind": "Keycloak",
                    "name": "keycloak-sample"
                  },
                  "passwordPolicy": [
                    {
                      "type": "forceExpiredPasswordChange",
                      "value": "365"
                    },
                    {
                      "type": "length",
                      "value": "8"
                    }
                  ],
                  "realmEventConfig": {
                    "adminEventsDetailsEnabled": false,
                    "adminEventsEnabled": true,
                    "enabledEventTypes": [
                      "UPDATE_CONSENT_ERROR",
                      "CLIENT_LOGIN"
                    ],
                    "eventsEnabled": true,
                    "eventsExpiration": 15000,
                    "eventsListeners": [
                      "jboss-logging"
                    ]
                  },
                  "realmName": "d2-id-kc-realm-name",
                  "ssoAutoRedirectEnabled": false,
                  "ssoRealmEnabled": false,
                  "ssoRealmName": "openshift"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmComponent",
                "metadata": {
                  "name": "keycloakrealmcomponent-sample"
                },
                "spec": {
                  "config": {
                    "allowPasswordAuthentication": [
                      "true"
                    ],
                    "cachePolicy": [
                      "EVICT_WEEKLY"
                    ],
                    "debug": [
                      "true"
                    ],
                    "editMode": [
                      "READ_ONLY"
                    ],
                    "enabled": [
                      "true"
                    ],
                    "evictionDay": [
                      "3"
                    ],
                    "evictionHour": [
                      "5"
                    ],
                    "evictionMinute": [
                      "7"
                    ],
                    "kerberosRealm": [
                      "test-realm"
                    ],
                    "keyTab": [
                      "test-key-tab"
                    ],
                    "priority": [
                      "0"
                    ],
                    "serverPrincipal": [
                      "srv-principal-test"
                    ],
                    "updateProfileFirstLogin": [
                      "true"
                    ]
                  },
                  "name": "cr-kerb-test",
                  "providerId": "kerberos",
                  "providerType": "org.keycloak.storage.UserStorageProvider",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmGroup",
                "metadata": {
                  "name": "keycloakrealmgroup-sample"
                },
                "spec": {
                  "name": "ArgoCDAdmins",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "keycloakrealmidentityprovider-sample"
                },
                "spec": {
                  "alias": "instagram",
                  "authenticateByDefault": false,
                  "config": {
                    "clientId": "foo",
                    "clientSecret": "$secretName:secretKey",
                    "hideOnLoginPage": "true",
                    "syncMode": "IMPORT",
                    "useJwksUrl": "true"
                  },
                  "enabled": true,
                  "firstBrokerLoginFlowAlias": "first broker login",
                  "mappers": [
                    {
                      "config": {
                        "role": "role-tr",
                        "syncMode": "INHERIT"
                      },
                      "identityProviderAlias": "instagram",
                      "identityProviderMapper": "oidc-hardcoded-role-idp-mapper",
                      "name": "test3212"
                    },
                    {
                      "config": {
                        "attribute": "foo",
                        "attribute.value": "bar",
                        "syncMode": "IMPORT"
                      },
                      "identityProviderAlias": "instagram",
                      "identityProviderMapper": "hardcoded-attribute-idp-mapper",
                      "name": "test-33221"
                    }
                  ],
                  "providerId": "instagram",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRole",
                "metadata": {
                  "name": "keycloakrealmrole-sample"
                },
                "spec": {
                  "composite": true,
                  "description": "default developer role",
                  "name": "developer",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRoleBatch",
                "metadata": {
                  "name": "keycloakrealmrolebatch-sample"
                },
                "spec": {
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "roles": [
                    {
                      "composite": true,
                      "description": "default developer role",
                      "isDefault": false,
                      "name": "developer"
                    },
                    {
                      "composite": true,
                      "description": "default administrator role",
                      "isDefault": false,
                      "name": "administrator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmUser",
                "metadata": {
                  "name": "keycloakrealmuser-sample"
                },
                "spec": {
                  "attributes": {
                    "baz": "jazz",
                    "foo": "bar"
                  },
                  "email": "john.snow13@example.com",
                  "emailVerified": true,
                  "enabled": true,
                  "firstName": "John",
                  "keepResource": true,
                  "lastName": "Snow",
                  "password": "12345678",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "requiredUserActions": [
                    "UPDATE_PASSWORD"
                  ],
                  "username": "john.snow13"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloak",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloak-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloak",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloak-sample"
                },
                "spec": {
                  "secret": "keycloak-access",
                  "url": "https://keycloak.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloakRealm",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloakrealm-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloakrealm",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloakrealm-sample"
                },
                "spec": {
                  "clusterKeycloakRef": "clusterkeycloak-sample",
                  "realmName": "realm-sample"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          containerImage: docker.io/epamedp/keycloak-operator:1.19.0
          createdAt: "2023-11-15T16:04:45Z"
          description: An Operator for managing Keycloak
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-keycloak-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterKeycloakRealm is the Schema for the clusterkeycloakrealms
              API.
            displayName: Cluster Keycloak Realm
            kind: ClusterKeycloakRealm
            name: clusterkeycloakrealms.v1.edp.epam.com
            version: v1alpha1
          - description: ClusterKeycloak is the Schema for the clusterkeycloaks API.
            displayName: Cluster Keycloak
            kind: ClusterKeycloak
            name: clusterkeycloaks.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakAuthFlow is the Schema for the keycloak authentication
              flow API.
            displayName: Keycloak Auth Flow
            kind: KeycloakAuthFlow
            name: keycloakauthflows.v1.edp.epam.com
            version: v1
          - description: KeycloakClient is the Schema for the keycloak clients API.
            displayName: Keycloak Client
            kind: KeycloakClient
            name: keycloakclients.v1.edp.epam.com
            version: v1
          - description: KeycloakClientScope is the Schema for the keycloakclientscopes
              API.
            displayName: Keycloak Client Scope
            kind: KeycloakClientScope
            name: keycloakclientscopes.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmComponent is the Schema for the keycloak component
              API.
            displayName: Keycloak Realm Component
            kind: KeycloakRealmComponent
            name: keycloakrealmcomponents.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmGroup is the Schema for the keycloak group API.
            displayName: Keycloak Realm Group
            kind: KeycloakRealmGroup
            name: keycloakrealmgroups.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmIdentityProvider is the Schema for the keycloak
              realm identity provider API.
            displayName: Keycloak Realm Identity Provider
            kind: KeycloakRealmIdentityProvider
            name: keycloakrealmidentityproviders.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRoleBatch is the Schema for the keycloak roles
              API.
            displayName: Keycloak Realm Role Batch
            kind: KeycloakRealmRoleBatch
            name: keycloakrealmrolebatches.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRole is the Schema for the keycloak group API.
            displayName: Keycloak Realm Role
            kind: KeycloakRealmRole
            name: keycloakrealmroles.v1.edp.epam.com
            version: v1
          - description: KeycloakRealm is the Schema for the keycloak realms API.
            displayName: Keycloak Realm
            kind: KeycloakRealm
            name: keycloakrealms.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmUser is the Schema for the keycloak user API.
            displayName: Keycloak Realm User
            kind: KeycloakRealmUser
            name: keycloakrealmusers.v1.edp.epam.com
            version: v1
          - description: Keycloak is the Schema for the keycloaks API.
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.v1.edp.epam.com
            version: v1
        description: |
          Keycloak Operator is an operator that is responsible for establishing
          a connection to provided Keycloak Server, reconciling Keycloak entities (realms,
          roles, groups, users, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Keycloak `Master` realm, and assign a `create-realm` role.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  keycloak-access
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Keycloak` with Keycloak instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: Keycloak
              metadata:
                name: keycloak-sample
              spec:
                secret: keycloak-access             # Secret name
                url: https://keycloak.example.com   # Keycloak URL
              ```

              Wait for the `.status` field with  `status.connected: true`

          4. Create Keycloak realm and group using Custom Resources:

            ```yaml
            apiVersion: v1.edp.epam.com/v1
            kind: KeycloakRealm
            metadata:
              name: keycloakrealm-sample
            spec:
              realmName: realm-sample
              keycloakOwner: keycloak-sample   # the name of `kind: Keycloak`
              ```

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: KeycloakRealmGroup
              metadata:
                name: argocd-admins
              spec:
                name: ArgoCDAdmins
                realm: keycloakrealm-sample   # the name of `kind: KeycloakRealm`
              ```
        displayName: EDP Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - authentication
        - authorization
        - edp
        - idp
        - keycloak
        - oauth
        - oidc
        - operator
        - saml
        - sso
        links:
        - name: Edp Keycloak Operator
          url: https://github.com/epam/edp-keycloak-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/keycloak-operator:1.19.0
        version: 1.19.0
      entries:
      - name: edp-keycloak-operator.v1.19.0
        version: 1.19.0
      - name: edp-keycloak-operator.v1.18.2
        version: 1.18.2
      - name: edp-keycloak-operator.v1.17.1
        version: 1.17.1
      - name: edp-keycloak-operator.v1.13.0
        version: 1.13.0
      name: alpha
    - currentCSV: edp-keycloak-operator.v1.29.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "keycloak-sample"
                },
                "spec": {
                  "secret": "my-keycloak-secret",
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakAuthFlow",
                "metadata": {
                  "name": "keycloakauthflow-sample"
                },
                "spec": {
                  "alias": "MyBrowser",
                  "authenticationExecutions": [
                    {
                      "authenticator": "auth-cookie",
                      "priority": 0,
                      "requirement": "ALTERNATIVE"
                    },
                    {
                      "authenticator": "identity-provider-redirector",
                      "authenticatorConfig": {
                        "alias": "my-alias",
                        "config": {
                          "defaultProvider": "my-alias"
                        }
                      },
                      "priority": 1,
                      "requirement": "REQUIRED"
                    }
                  ],
                  "builtIn": false,
                  "description": "browser with idp",
                  "providerId": "basic-flow",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "topLevel": true
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-authorization-sample"
                },
                "spec": {
                  "adminFineGrainedPermissionsEnabled": true,
                  "authorization": {
                    "permissions": [
                      {
                        "decisionStrategy": "AFFIRMATIVE",
                        "description": "Resource permission",
                        "logic": "POSITIVE",
                        "name": "resource-permission",
                        "policies": [
                          "role-policy"
                        ],
                        "resources": [
                          "resource1"
                        ],
                        "type": "resource"
                      },
                      {
                        "decisionStrategy": "CONSENSUS",
                        "description": "Scope permission",
                        "logic": "POSITIVE",
                        "name": "scope-permission",
                        "policies": [
                          "role-policy"
                        ],
                        "scopes": [
                          "scope1"
                        ],
                        "type": "scope"
                      }
                    ],
                    "policies": [
                      {
                        "decisionStrategy": "AFFIRMATIVE",
                        "description": "Role policy",
                        "logic": "POSITIVE",
                        "name": "role-policy",
                        "rolePolicy": {
                          "roles": [
                            {
                              "name": "developer",
                              "required": true
                            }
                          ]
                        },
                        "type": "role"
                      },
                      {
                        "aggregatedPolicy": {
                          "policies": [
                            "policy1",
                            "policy2"
                          ]
                        },
                        "description": "Aggregate policy",
                        "name": "aggregate-policy",
                        "type": "aggregate"
                      },
                      {
                        "clientPolicy": {
                          "clients": [
                            "client1",
                            "client2"
                          ]
                        },
                        "description": "Client policy",
                        "name": "client-policy",
                        "type": "client"
                      },
                      {
                        "description": "Group policy",
                        "groupPolicy": {
                          "groups": [
                            {
                              "extendChildren": true,
                              "name": "group1"
                            }
                          ]
                        },
                        "name": "group-policy",
                        "type": "group"
                      },
                      {
                        "description": "Role policy",
                        "name": "role-policy",
                        "rolePolicy": {
                          "roles": [
                            {
                              "name": "developer",
                              "required": true
                            }
                          ]
                        },
                        "type": "role"
                      },
                      {
                        "description": "Time policy",
                        "name": "time-policy",
                        "timePolicy": {
                          "notBefore": "2021-01-01T00:00:00Z",
                          "notOnOrAfter": "2021-12-31T23:59:59Z"
                        },
                        "type": "time"
                      },
                      {
                        "description": "User policy",
                        "name": "user-policy",
                        "type": "user",
                        "userPolicy": {
                          "users": [
                            "user1",
                            "user2"
                          ]
                        }
                      }
                    ],
                    "resources": [
                      {
                        "displayName": "Resource 1",
                        "iconUri": "https://example.com/icon.png",
                        "name": "resource1",
                        "scopes": [
                          "scope1"
                        ],
                        "type": "test"
                      }
                    ],
                    "scopes": [
                      "scope1"
                    ]
                  },
                  "authorizationServicesEnabled": true,
                  "clientId": "authorization-sample",
                  "directAccess": true,
                  "permission": {
                    "scopePermissions": [
                      {
                        "name": "token-exchange",
                        "policies": [
                          "policy1"
                        ]
                      }
                    ]
                  },
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-authorization-sample:client-secret-key",
                  "serviceAccount": {
                    "enabled": true
                  },
                  "webUrl": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-sample"
                },
                "spec": {
                  "advancedProtocolMappers": true,
                  "authenticationFlowBindingOverrides": {
                    "browser": "browser",
                    "directGrant": "direct grant"
                  },
                  "clientId": "agocd",
                  "defaultClientScopes": [
                    "argocd_groups"
                  ],
                  "directAccess": true,
                  "public": false,
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-name:client-secret-key",
                  "webUrl": "https://argocd.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClientScope",
                "metadata": {
                  "name": "keycloakclientscope-sample"
                },
                "spec": {
                  "description": "Group Membership",
                  "name": "groups",
                  "protocol": "openid-connect",
                  "protocolMappers": [
                    {
                      "config": {
                        "access.token.claim": "true",
                        "claim.name": "groups",
                        "full.path": "false",
                        "id.token.claim": "true",
                        "userinfo.token.claim": "true"
                      },
                      "name": "groups",
                      "protocol": "openid-connect",
                      "protocolMapper": "oidc-group-membership-mapper"
                    }
                  ],
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "keycloakrealm-sample"
                },
                "spec": {
                  "id": "d1-id-kc-realm-name",
                  "keycloakRef": {
                    "kind": "Keycloak",
                    "name": "keycloak-sample"
                  },
                  "organizationsEnabled": true,
                  "passwordPolicy": [
                    {
                      "type": "forceExpiredPasswordChange",
                      "value": "365"
                    },
                    {
                      "type": "length",
                      "value": "8"
                    }
                  ],
                  "realmEventConfig": {
                    "adminEventsDetailsEnabled": false,
                    "adminEventsEnabled": true,
                    "enabledEventTypes": [
                      "UPDATE_CONSENT_ERROR",
                      "CLIENT_LOGIN"
                    ],
                    "eventsEnabled": true,
                    "eventsExpiration": 15000,
                    "eventsListeners": [
                      "jboss-logging"
                    ]
                  },
                  "realmName": "d2-id-kc-realm-name",
                  "userProfileConfig": {
                    "attributes": [
                      {
                        "annotations": {
                          "inputType": "text"
                        },
                        "displayName": "Test Attribute",
                        "group": "test-group",
                        "multivalued": true,
                        "name": "test-attribute",
                        "permissions": {
                          "edit": [
                            "admin"
                          ],
                          "view": [
                            "admin",
                            "user"
                          ]
                        },
                        "required": {
                          "roles": [
                            "admin"
                          ],
                          "scopes": [
                            "profile"
                          ]
                        },
                        "selector": {
                          "scopes": [
                            "profile"
                          ]
                        },
                        "validations": {
                          "email": {
                            "max-local-length": {
                              "intVal": 64
                            }
                          },
                          "local-date": {},
                          "multivalued": {
                            "max": {
                              "stringVal": "10"
                            },
                            "min": {
                              "stringVal": "1"
                            }
                          },
                          "options": {
                            "options": {
                              "sliceVal": [
                                "option1",
                                "option2"
                              ]
                            }
                          }
                        }
                      }
                    ],
                    "groups": [
                      {
                        "annotations": {
                          "groupAnnotation": "groupAnnotation"
                        },
                        "displayDescription": "Test Group",
                        "displayHeader": "Test Group",
                        "name": "test-group"
                      }
                    ],
                    "unmanagedAttributePolicy": "ENABLED"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmComponent",
                "metadata": {
                  "name": "keycloakrealmcomponent-sample"
                },
                "spec": {
                  "config": {
                    "allowPasswordAuthentication": [
                      "true"
                    ],
                    "cachePolicy": [
                      "EVICT_WEEKLY"
                    ],
                    "debug": [
                      "true"
                    ],
                    "editMode": [
                      "READ_ONLY"
                    ],
                    "enabled": [
                      "true"
                    ],
                    "evictionDay": [
                      "3"
                    ],
                    "evictionHour": [
                      "5"
                    ],
                    "evictionMinute": [
                      "7"
                    ],
                    "kerberosRealm": [
                      "test-realm"
                    ],
                    "keyTab": [
                      "test-key-tab"
                    ],
                    "priority": [
                      "0"
                    ],
                    "serverPrincipal": [
                      "srv-principal-test"
                    ],
                    "updateProfileFirstLogin": [
                      "true"
                    ]
                  },
                  "name": "cr-kerb-test",
                  "providerId": "kerberos",
                  "providerType": "org.keycloak.storage.UserStorageProvider",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmGroup",
                "metadata": {
                  "name": "keycloakrealmgroup-sample"
                },
                "spec": {
                  "name": "ArgoCDAdmins",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "test-org-idp",
                  "namespace": "default"
                },
                "spec": {
                  "alias": "test-org-idp",
                  "config": {
                    "clientId": "test-org-client-id",
                    "clientSecret": "test-org-client-secret",
                    "kc.org.broker.redirect.mode.email-matches": "true",
                    "kc.org.domain": "example.com"
                  },
                  "enabled": true,
                  "providerId": "github",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "test-org-realm"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "keycloakrealmidentityprovider-sample"
                },
                "spec": {
                  "alias": "github",
                  "authenticateByDefault": false,
                  "config": {
                    "clientId": "foo",
                    "clientSecret": "$secretName:secretKey",
                    "syncMode": "IMPORT",
                    "useJwksUrl": "true"
                  },
                  "enabled": true,
                  "firstBrokerLoginFlowAlias": "first broker login",
                  "mappers": [
                    {
                      "config": {
                        "role": "role-tr",
                        "syncMode": "INHERIT"
                      },
                      "identityProviderAlias": "github",
                      "identityProviderMapper": "oidc-hardcoded-role-idp-mapper",
                      "name": "test3212"
                    },
                    {
                      "config": {
                        "attribute": "foo",
                        "attribute.value": "bar",
                        "syncMode": "IMPORT"
                      },
                      "identityProviderAlias": "github",
                      "identityProviderMapper": "hardcoded-attribute-idp-mapper",
                      "name": "test-33221"
                    }
                  ],
                  "providerId": "github",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRole",
                "metadata": {
                  "name": "keycloakrealmrole-sample"
                },
                "spec": {
                  "composite": true,
                  "description": "default developer role",
                  "name": "developer",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRoleBatch",
                "metadata": {
                  "name": "keycloakrealmrolebatch-sample"
                },
                "spec": {
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "roles": [
                    {
                      "composite": true,
                      "description": "default developer role",
                      "isDefault": false,
                      "name": "developer"
                    },
                    {
                      "composite": true,
                      "description": "default administrator role",
                      "isDefault": false,
                      "name": "administrator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmUser",
                "metadata": {
                  "name": "keycloakrealmuser-sample"
                },
                "spec": {
                  "attributes": {
                    "baz": "jazz",
                    "foo": "bar"
                  },
                  "email": "john.snow13@example.com",
                  "emailVerified": true,
                  "enabled": true,
                  "firstName": "John",
                  "keepResource": true,
                  "lastName": "Snow",
                  "password": "12345678",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "requiredUserActions": [
                    "UPDATE_PASSWORD"
                  ],
                  "username": "john.snow13"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloak",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloak-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloak",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloak-sample"
                },
                "spec": {
                  "secret": "keycloak-access",
                  "url": "https://keycloak.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloakRealm",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloakrealm-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloakrealm",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloakrealm-sample"
                },
                "spec": {
                  "authenticationFlows": {
                    "browserFlow": "browserFlow-sample"
                  },
                  "clusterKeycloakRef": "clusterkeycloak-sample",
                  "organizationsEnabled": false,
                  "realmName": "realm-sample"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "KeycloakOrganization",
                "metadata": {
                  "name": "test-keycloak-organization",
                  "namespace": "default"
                },
                "spec": {
                  "alias": "test-org",
                  "attributes": {
                    "department": [
                      "engineering",
                      "qa"
                    ],
                    "location": [
                      "us-east"
                    ]
                  },
                  "description": "Test organization",
                  "domains": [
                    "example.com",
                    "test.com"
                  ],
                  "identityProviders": [
                    {
                      "alias": "test-org-idp"
                    }
                  ],
                  "name": "Test Organization",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "test-org-realm"
                  },
                  "redirectUrl": "https://example.com/redirect"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          containerImage: docker.io/epamedp/keycloak-operator:1.29.0
          createdAt: "2025-09-26T12:16:55Z"
          description: An Operator for managing Keycloak
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/epam/edp-keycloak-operator
          support: KubeRocketCI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterKeycloakRealm is the Schema for the clusterkeycloakrealms
              API.
            displayName: Cluster Keycloak Realm
            kind: ClusterKeycloakRealm
            name: clusterkeycloakrealms.v1.edp.epam.com
            version: v1alpha1
          - description: ClusterKeycloak is the Schema for the clusterkeycloaks API.
            displayName: Cluster Keycloak
            kind: ClusterKeycloak
            name: clusterkeycloaks.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakAuthFlow is the Schema for the keycloak authentication
              flow API.
            displayName: Keycloak Auth Flow
            kind: KeycloakAuthFlow
            name: keycloakauthflows.v1.edp.epam.com
            version: v1
          - description: KeycloakClient is the Schema for the keycloak clients API.
            displayName: Keycloak Client
            kind: KeycloakClient
            name: keycloakclients.v1.edp.epam.com
            version: v1
          - description: KeycloakClientScope is the Schema for the keycloakclientscopes
              API.
            displayName: Keycloak Client Scope
            kind: KeycloakClientScope
            name: keycloakclientscopes.v1.edp.epam.com
            version: v1
          - description: KeycloakOrganization is the Schema for the organizations
              API.
            displayName: Keycloak Organization
            kind: KeycloakOrganization
            name: keycloakorganizations.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakRealmComponent is the Schema for the keycloak component
              API.
            displayName: Keycloak Realm Component
            kind: KeycloakRealmComponent
            name: keycloakrealmcomponents.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmGroup is the Schema for the keycloak group API.
            displayName: Keycloak Realm Group
            kind: KeycloakRealmGroup
            name: keycloakrealmgroups.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmIdentityProvider is the Schema for the keycloak
              realm identity provider API.
            displayName: Keycloak Realm Identity Provider
            kind: KeycloakRealmIdentityProvider
            name: keycloakrealmidentityproviders.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRoleBatch is the Schema for the keycloak roles
              API.
            displayName: Keycloak Realm Role Batch
            kind: KeycloakRealmRoleBatch
            name: keycloakrealmrolebatches.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRole is the Schema for the keycloak group API.
            displayName: Keycloak Realm Role
            kind: KeycloakRealmRole
            name: keycloakrealmroles.v1.edp.epam.com
            version: v1
          - description: KeycloakRealm is the Schema for the keycloak realms API.
            displayName: Keycloak Realm
            kind: KeycloakRealm
            name: keycloakrealms.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmUser is the Schema for the keycloak user API.
            displayName: Keycloak Realm User
            kind: KeycloakRealmUser
            name: keycloakrealmusers.v1.edp.epam.com
            version: v1
          - description: Keycloak is the Schema for the keycloaks API.
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.v1.edp.epam.com
            version: v1
        description: |
          Keycloak Operator is an operator that is responsible for establishing
          a connection to provided Keycloak Server, reconciling Keycloak entities (realms,
          roles, groups, users, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Keycloak `Master` realm, and assign a `create-realm` role.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  keycloak-access
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Keycloak` with Keycloak instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: Keycloak
              metadata:
                name: keycloak-sample
              spec:
                secret: keycloak-access             # Secret name
                url: https://keycloak.example.com   # Keycloak URL
              ```

              Wait for the `.status` field with  `status.connected: true`

          4. Create Keycloak realm and group using Custom Resources:

            ```yaml
            apiVersion: v1.edp.epam.com/v1
            kind: KeycloakRealm
            metadata:
              name: keycloakrealm-sample
            spec:
              realmName: realm-sample
              keycloakOwner: keycloak-sample   # the name of `kind: Keycloak`
              ```

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: KeycloakRealmGroup
              metadata:
                name: argocd-admins
              spec:
                name: ArgoCDAdmins
                realm: keycloakrealm-sample   # the name of `kind: KeycloakRealm`
              ```
        displayName: EDP Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - authentication
        - authorization
        - edp
        - idp
        - keycloak
        - kuberocketci
        - oauth
        - oidc
        - operator
        - saml
        - sso
        links:
        - name: Edp Keycloak Operator
          url: https://github.com/epam/edp-keycloak-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: KubeRocketCI
          url: https://docs.kuberocketci.io/
        relatedImages:
        - docker.io/epamedp/keycloak-operator:1.29.0
        version: 1.29.0
      entries:
      - name: edp-keycloak-operator.v1.29.0
        version: 1.29.0
      - name: edp-keycloak-operator.v1.28.0
        version: 1.28.0
      - name: edp-keycloak-operator.v1.27.1
        version: 1.27.1
      - name: edp-keycloak-operator.v1.26.0
        version: 1.26.0
      - name: edp-keycloak-operator.v1.25.0
        version: 1.25.0
      - name: edp-keycloak-operator.v1.24.0
        version: 1.24.0
      - name: edp-keycloak-operator.v1.23.0
        version: 1.23.0
      name: stable
    defaultChannel: stable
    packageName: edp-keycloak-operator
    provider:
      name: KubeRocketCI
      url: https://docs.kuberocketci.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: minhokim
      provider-url: ""
    name: node-discovery-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-discovery-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.wunicorns.com/v1alpha1",
                "kind": "NodeDiscovery",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "node-discovery-operator",
                    "app.kubernetes.io/instance": "nodediscovery-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodediscovery",
                    "app.kubernetes.io/part-of": "node-discovery-operator"
                  },
                  "name": "nodediscovery-sample"
                },
                "spec": {
                  "containers": [
                    {
                      "args": [
                        "tail",
                        "-f",
                        "/dev/null"
                      ],
                      "image": "alicek106/stress",
                      "imagePullPolicy": "IfNotPresent",
                      "name": "stress",
                      "resources": {
                        "limits": {
                          "cpu": "100m",
                          "memory": "1000Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "1000Mi"
                        }
                      },
                      "terminationGracePeriodSeconds": 10
                    }
                  ],
                  "cpus": "0",
                  "enabled": "true",
                  "mems": "0",
                  "nodeName": "worker3"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-18T01:32:51Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeDiscovery is the Schema for the nodediscoveries API
            displayName: Node Discovery
            kind: NodeDiscovery
            name: nodediscoveries.cache.wunicorns.com
            version: v1alpha1
        description: find cgroup infor and set value.
        displayName: node-discovery-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cgroup
        - cpulist
        - mems
        - cpus
        - numa
        links:
        - name: Node Discovery Operator
          url: https://node-discovery-operator.domain
        maintainers:
        - email: wunicorns1436@gmail.com
          name: minhokim
        maturity: alpha
        provider:
          name: minhokim
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/caspell/cgroup-discovery@sha256:b437dcde0107124e43a4e2fb8026ec191f44b5e761bedc2c295f47ff489cf343
        version: 0.0.1
      entries:
      - name: node-discovery-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: node-discovery-operator
    provider:
      name: minhokim
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HashiCorp
      provider-url: https://www.hashicorp.com/
    name: hcp-terraform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hcp-terraform-operator.v2.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "AgentPool",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "agentTokens": [
                    {
                      "name": "token"
                    }
                  ],
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Module",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "module": {
                    "source": "redeux/terraform-cloud-agent/kubernetes",
                    "version": "1.0.1"
                  },
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  },
                  "workspace": {
                    "id": "WORKSPACE_ID"
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Project",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Workspace",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.connect.redhat.com/hashicorp/hcp-terraform-operator@sha256:8885a6a5eae092117a670ec5541a3e4a24c6b23807fd17b0446d91b01eb6e421
          createdAt: "2025-05-28T13:58:52Z"
          description: HCP Terraform Operator for Kubernetes allows managing HCP Terraform
            / Terraform Enterprise resources via Kubernetes Custom Resources.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Valid HCP Terraform / Terraform
            Enterprise subscription is required
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hashicorp/hcp-terraform-operator
          support: HashiCorp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              AgentPool manages HCP Terraform Agent Pools, HCP Terraform Agent Tokens and can perform HCP Terraform Agent scaling.
              More infromation:
                - https://developer.hashicorp.com/terraform/cloud-docs/agents/agent-pools
                - https://developer.hashicorp.com/terraform/cloud-docs/users-teams-organizations/api-tokens#agent-api-tokens
                - https://developer.hashicorp.com/terraform/cloud-docs/agents
            displayName: Agent Pool
            kind: AgentPool
            name: agentpools.app.terraform.io
            version: v1alpha2
          - description: |-
              Module implements API-driven Run Workflows.
              More information:
                - https://developer.hashicorp.com/terraform/cloud-docs/run/api
            displayName: Module
            kind: Module
            name: modules.app.terraform.io
            version: v1alpha2
          - description: |-
              Project manages HCP Terraform Projects.
              More information:
              - https://developer.hashicorp.com/terraform/cloud-docs/projects/manage
            displayName: Project
            kind: Project
            name: projects.app.terraform.io
            version: v1alpha2
          - description: |-
              Workspace manages HCP Terraform Workspaces.
              More information:
                - https://developer.hashicorp.com/terraform/cloud-docs/workspaces
            displayName: Workspace
            kind: Workspace
            name: workspaces.app.terraform.io
            version: v1alpha2
        description: HCP Terraform Operator for Kubernetes allows managing HCP Terraform
          / Terraform Enterprise resources via Kubernetes Custom Resources.
        displayName: HCP Terraform Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - terraform
        links:
        - name: HCP Terraform Operator
          url: https://github.com/hashicorp/hcp-terraform-operator
        maintainers:
        - email: hcp-tf-operator@hashicorp.com
          name: HashiCorp
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: HashiCorp
          url: https://www.hashicorp.com/
        relatedImages:
        - registry.connect.redhat.com/hashicorp/hcp-terraform-operator@sha256:8885a6a5eae092117a670ec5541a3e4a24c6b23807fd17b0446d91b01eb6e421
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        version: 2.9.2
      entries:
      - name: hcp-terraform-operator.v2.9.2
        version: 2.9.2
      - name: hcp-terraform-operator.v2.9.1
        version: 2.9.1
      - name: hcp-terraform-operator.v2.9.0
        version: 2.9.0
      - name: hcp-terraform-operator.v2.8.1
        version: 2.8.1
      - name: hcp-terraform-operator.v2.8.0
        version: 2.8.0
      name: alpha
    defaultChannel: alpha
    packageName: hcp-terraform-operator
    provider:
      name: HashiCorp
      url: https://www.hashicorp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Appian Corporation
      provider-url: https://appian.com/
    name: appian-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appian-operator.v0.184.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: AI/Machine Learning, Application Runtime, Big Data, Integration
            & Delivery, Modernization & Migration
          createdAt: "2025-08-07T15:50:26Z"
          description: The Appian operator facilitates managing instances of Appian
            on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appian-operator
          operators.openshift.io/valid-subscription: '["Appian License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Appian Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Appian
            name: appians.crd.k8s.appian.com
            version: v1beta1
        description: The Appian operator facilitates managing instances of Appian
          on Kubernetes.
        displayName: Appian operator
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - Appian
        links:
        - name: Appian
          url: https://appian.com/
        - name: Documentation
          url: https://docs.appian.com/suite/help/latest/k8s
        maintainers:
        - email: support@appian.com
          name: Appian Corporation
        maturity: planning
        provider:
          name: Appian Corporation
          url: https://appian.com/
        relatedImages:
        - 937620572175.dkr.ecr.us-east-1.amazonaws.com/appian/appian-operator@sha256:26c2ba27f9360226f34f289109cb0d10e9bf44777146ccfd5f367adf2367a400
        version: 0.184.0
      entries:
      - name: appian-operator.v0.184.0
        version: 0.184.0
      - name: appian-operator.v0.171.0
        version: 0.171.0
      name: alpha
    defaultChannel: alpha
    packageName: appian-operator
    provider:
      name: Appian Corporation
      url: https://appian.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: aiu-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aiu-operator.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "labels": {
                    "operator-version": "2.1.0-dev"
                  },
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "aiu": {
                    "strategy": "single"
                  },
                  "devicePlugin": {
                    "aiuSelector": {
                      "deviceID": "06a7",
                      "vendor": "1014"
                    },
                    "image": "aiu-device-plugin",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "deviceType": "vfio-pci",
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation"
                  ],
                  "featureDiscovery": {
                    "image": "aiu-feature-discovery",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "isRdma": false,
                  "loglevel": "info",
                  "metricsExporter": {
                    "enabled": true,
                    "image": "aiu-exporter",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "numVfs": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
          createdAt: "2024-10-07T13:07:51Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 2.1.1
      entries:
      - name: aiu-operator.v2.1.1
        version: 2.1.1
      name: candidate-2.1
    - currentCSV: aiu-operator.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "labels": {
                    "operator-version": "2.1.0-dev"
                  },
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "aiu": {
                    "strategy": "single"
                  },
                  "devicePlugin": {
                    "aiuSelector": {
                      "deviceID": "06a7",
                      "vendor": "1014"
                    },
                    "image": "aiu-device-plugin",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "deviceType": "vfio-pci",
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation"
                  ],
                  "featureDiscovery": {
                    "image": "aiu-feature-discovery",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "isRdma": false,
                  "loglevel": "info",
                  "metricsExporter": {
                    "enabled": true,
                    "image": "aiu-exporter",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "numVfs": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
          createdAt: "2024-10-07T13:07:51Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
        version: 2.1.1
      entries:
      - name: aiu-operator.v2.1.1
        version: 2.1.1
      name: fast-2.1
    - currentCSV: aiu-operator.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "labels": {
                    "operator-version": "2.1.0-dev"
                  },
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "aiu": {
                    "strategy": "single"
                  },
                  "devicePlugin": {
                    "aiuSelector": {
                      "deviceID": "06a7",
                      "vendor": "1014"
                    },
                    "image": "aiu-device-plugin",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "deviceType": "vfio-pci",
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation"
                  ],
                  "featureDiscovery": {
                    "image": "aiu-feature-discovery",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "isRdma": false,
                  "loglevel": "info",
                  "metricsExporter": {
                    "enabled": true,
                    "image": "aiu-exporter",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "numVfs": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
          createdAt: "2024-10-07T13:07:51Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 2.1.1
      entries:
      - name: aiu-operator.v2.1.1
        version: 2.1.1
      name: stable-2.1
    - currentCSV: aiu-operator.v2.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "labels": {
                    "operator-version": "2.3.3"
                  },
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "aiu-device-plugin",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.1"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "featureDiscovery": {
                    "image": "aiu-feature-discovery",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "aiu-exporter",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "aiu-webhook-validator",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "scheduler": {
                    "image": "aiu-kube-scheduler-ocp-4.16",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.2"
                  }
                }
              },
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuNodeState",
                "metadata": {
                  "labels": {
                    "operator-version": "2.3.3"
                  },
                  "name": "aiunodestate"
                },
                "spec": {
                  "nodeName": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2025_03/aiu-operator@sha256:2ea8268e94d44581b3490cd4bbbd5525359073129bae42c7f007c05727afe1c3
          createdAt: "2025-04-03T01:19:38Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["aiunodestates.aiu.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu/release_2025_03/aiu-operator@sha256:2ea8268e94d44581b3490cd4bbbd5525359073129bae42c7f007c05727afe1c3
        version: 2.3.3
      entries:
      - name: aiu-operator.v2.3.3
        version: 2.3.3
      name: stable-v2.3
    - currentCSV: aiu-operator.v2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "aiuFilter": ".",
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "release_2025_05"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "aiu-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "e2e_stable",
                      "repository": "icr.io/ibmaiu_internal/x86_64/dd2",
                      "version": "release_2025_05"
                    },
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "aiu-exporter",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "aiu-webhook-validator",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "scheduler": {
                    "image": "aiu-kube-scheduler-ocp-4.16",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  }
                }
              },
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2025_05/aiu-operator@sha256:9796ab1c39d1dd55f8aa79c9f585be0f988b5026745184bdc890dfd1846909df
          createdAt: "2025-05-28T05:53:45Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["aiunodestates.aiu.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu/release_2025_05/aiu-operator@sha256:9796ab1c39d1dd55f8aa79c9f585be0f988b5026745184bdc890dfd1846909df
        - icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453
        version: 2.4.0
      entries:
      - name: aiu-operator.v2.4.0
        version: 2.4.0
      name: stable-v2.4
    defaultChannel: stable-v2.4
    packageName: aiu-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-ingress
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-ingress.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Ingress",
                "metadata": {
                  "name": "ingress-sample"
                },
                "spec": {
                  "admissionWebhooks": {
                    "patch": {
                      "image": {
                        "repository": "docker.io/jettech/kube-webhook-certgen"
                      }
                    }
                  },
                  "configMapMerger": {
                    "registry": "prod.harbor.keysight.digital/loadcore",
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "environment": "openshift",
                  "ingress-nginx": {
                    "controller": {
                      "config": {
                        "global-auth-signin": "https://$host/eula/static/",
                        "global-auth-url": "http://kcos-framework-v1-kcos-eula.kcos-framework.svc.cluster.local/v1/check"
                      },
                      "daemonset": {
                        "useHostPort": false
                      },
                      "hostNetwork": false,
                      "image": {
                        "registry": "prod.harbor.keysight.digital/loadcore"
                      },
                      "ingressClassResource": {
                        "default": true
                      },
                      "nodeSelector": "",
                      "service": {
                        "nodePorts": {
                          "http": 80,
                          "https": 443
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "defaultBackend": {
                      "image": {
                        "repository": "k8s.gcr.io/defaultbackend-amd64"
                      }
                    }
                  },
                  "jobDelWebhook": {
                    "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:28:18Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ingress
            name: ingresses.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Ingress Operator
        displayName: Keysight KCOS Ingress Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Ingress
          url: https://keysight-kcos-ingress.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-ingress@sha256:c6f6f8b44276efd8f5ffaa099b393e7e97d9cfa627952c0d049b98d92dc3ab59
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        version: 0.1.21
      entries:
      - name: keysight-kcos-ingress.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-ingress
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-migration
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-migration.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Migration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {
                  "auth": {
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configuration": {
                    "port": 8080
                  },
                  "database": {
                    "image": {
                      "name": "postgresql-ubi-image",
                      "tag": "14.2.0-3747-32"
                    }
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "esinfo": {
                    "batchSize": 10000,
                    "inputIndex": "wireless_stats"
                  },
                  "image": {
                    "remoteRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "systemMonitor": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:36:21Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Migration
            name: migrations.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Migration Operator
        displayName: Keysight WAP Migration Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Migration
          url: https://keysight-wap-migration.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-migration@sha256:118b466f1c59832a3620fa9b153507cc66ba38ee08b337b3a1a4225b158bebc5
        version: 0.1.21
      entries:
      - name: keysight-wap-migration.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-migration
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-xtradb-cluster-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator-certified.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator-certified.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator-certified.v1.17.0
        version: 1.17.0
      - name: percona-xtradb-cluster-operator-certified.v1.16.1
        version: 1.16.1
      - name: percona-xtradb-cluster-operator-certified.v1.16.0
        version: 1.16.0
      - name: percona-xtradb-cluster-operator-certified.v1.15.1
        version: 1.15.1
      - name: percona-xtradb-cluster-operator-certified.v1.15.0
        version: 1.15.0
      name: stable
    - currentCSV: percona-xtradb-cluster-operator-certified.v1.18.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        version: 1.18.0-cw
      entries:
      - name: percona-xtradb-cluster-operator-certified.v1.18.0-cw
        version: 1.18.0-cw
      - name: percona-xtradb-cluster-operator-certified.v1.17.0-cw
        version: 1.17.0-cw
      - name: percona-xtradb-cluster-operator-certified.v1.16.1-cw
        version: 1.16.1-cw
      - name: percona-xtradb-cluster-operator-certified.v1.16.0-cw
        version: 1.16.0-cw
      - name: percona-xtradb-cluster-operator-certified.v1.15.0-cw
        version: 1.15.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator-certified
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-storage-odf-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-storage-odf-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:962ca15393dd2c48f8a7189c1801b43e9fd1cd073a1971286d485f4b96060426
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.6.0'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:962ca15393dd2c48f8a7189c1801b43e9fd1cd073a1971286d485f4b96060426
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:2f785096aaa63b042e415aa0b64ace2a085cb7c219853f089d803bc37d8295ce
        version: 1.6.0
      entries:
      - name: ibm-storage-odf-operator.v1.6.0
        version: 1.6.0
      name: stable-v1.6
    - currentCSV: ibm-storage-odf-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:2dd8608645b13f81330466e25664b132d46af480a3ab2b0dd48fb1feba38870a
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.7.1'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:2dd8608645b13f81330466e25664b132d46af480a3ab2b0dd48fb1feba38870a
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:24eae9d1502ce98bb024d3eba154ddec07fdac0d78ecad0c3a64b55f5b3600f7
        version: 1.7.1
      entries:
      - name: ibm-storage-odf-operator.v1.7.1
        version: 1.7.1
      name: stable-v1.7
    - currentCSV: ibm-storage-odf-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:7432340ac156e620240eea73a19321a871fa89cc513c340a91d97d3f3fd4ded2
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.8.0'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:7432340ac156e620240eea73a19321a871fa89cc513c340a91d97d3f3fd4ded2
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:3344c7e987674fb98173bbc6ae1b189613e3ccefb468f1af9866734ddbf0408f
        version: 1.8.0
      entries:
      - name: ibm-storage-odf-operator.v1.8.0
        version: 1.8.0
      name: stable-v1.8
    defaultChannel: stable-v1.8
    packageName: ibm-storage-odf-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-nifi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nifi-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:05946a5e8f040e8ee46d56f2d4cca3d97e02d086e6d8224200204de4d76851fa
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:05946a5e8f040e8ee46d56f2d4cca3d97e02d086e6d8224200204de4d76851fa
        version: 24.11.1
      entries:
      - name: nifi-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: nifi-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:e5c3c4e2ad1c455fbb04a60658936bfb44c2bc9d2b616737ce1a0eba201b7a3f
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiClusterSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:e5c3c4e2ad1c455fbb04a60658936bfb44c2bc9d2b616737ce1a0eba201b7a3f
        version: 25.3.0
      entries:
      - name: nifi-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: nifi-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiClusterSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
        version: 25.7.0
      entries:
      - name: nifi-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: nifi-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiClusterSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
        version: 25.7.0
      entries:
      - name: nifi-operator.v25.7.0
        version: 25.7.0
      - name: nifi-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-nifi-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/en/products/edge/validated-patterns
    name: patterns-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: patterns-operator.v0.0.62
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gitops.hybrid-cloud-patterns.io/v1alpha1",
                "kind": "Pattern",
                "metadata": {
                  "name": "pattern-sample"
                },
                "spec": {
                  "clusterGroupName": "hub",
                  "gitSpec": {
                    "targetRepo": "https://github.com/validatedpatterns/multicloud-gitops",
                    "targetRevision": "main"
                  },
                  "multiSourceConfig": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/hybridcloudpatterns/patterns-operator:0.0.62
          createdAt: "2025-09-02 16:14:40"
          description: "An operator to deploy and manage architecture patterns from
            https://validatedpatterns.io. \nThis operator collects some Analytics
            like cluster version, cloud type, etc. To disable this\nadd \"ANALYTICS:
            false\" to the environment variables in the subscription."
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/validatedpatterns/patterns-operator
          support: validatedpatterns@googlegroups.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pattern is the Schema for the patterns API
            displayName: Pattern
            kind: Pattern
            name: patterns.gitops.hybrid-cloud-patterns.io
            version: v1alpha1
        description: Deploys and manages architecture patterns from https://validatedpatterns.io
        displayName: Validated Patterns Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Validated
        - Pattern
        - Blueprint
        - Architecture
        links:
        - name: Hybrid Cloud Patterns
          url: https://validatedpatterns.io
        maintainers:
        - email: validatedpatterns@googlegroups.com
          name: patterns-team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/en/products/edge/validated-patterns
        relatedImages:
        - quay.io/hybridcloudpatterns/kube-rbac-proxy@sha256:409a464692c7db3cf68893376479dca1b53802412237637725927f7e78aa0844
        - quay.io/hybridcloudpatterns/patterns-operator@sha256:15ed44938b112bcf9b3a9a566749b39ba2b9053a6a8b9b499ef247977c18496c
        version: 0.0.62
      entries:
      - name: patterns-operator.v0.0.62
        version: 0.0.62
      - name: patterns-operator.v0.0.61
        version: 0.0.61
      - name: patterns-operator.v0.0.60
        version: 0.0.60
      - name: patterns-operator.v0.0.59
        version: 0.0.59
      - name: patterns-operator.v0.0.58
        version: 0.0.58
      - name: patterns-operator.v0.0.57
        version: 0.0.57
      - name: patterns-operator.v0.0.56
        version: 0.0.56
      - name: patterns-operator.v0.0.55
        version: 0.0.55
      - name: patterns-operator.v0.0.54
        version: 0.0.54
      - name: patterns-operator.v0.0.53
        version: 0.0.53
      - name: patterns-operator.v0.0.52
        version: 0.0.52
      - name: patterns-operator.v0.0.51
        version: 0.0.51
      - name: patterns-operator.v0.0.50
        version: 0.0.50
      - name: patterns-operator.v0.0.49
        version: 0.0.49
      - name: patterns-operator.v0.0.48
        version: 0.0.48
      - name: patterns-operator.v0.0.47
        version: 0.0.47
      - name: patterns-operator.v0.0.46
        version: 0.0.46
      - name: patterns-operator.v0.0.45
        version: 0.0.45
      - name: patterns-operator.v0.0.44
        version: 0.0.44
      - name: patterns-operator.v0.0.43
        version: 0.0.43
      - name: patterns-operator.v0.0.42
        version: 0.0.42
      - name: patterns-operator.v0.0.41
        version: 0.0.41
      - name: patterns-operator.v0.0.40
        version: 0.0.40
      - name: patterns-operator.v0.0.39
        version: 0.0.39
      - name: patterns-operator.v0.0.38
        version: 0.0.38
      - name: patterns-operator.v0.0.37
        version: 0.0.37
      - name: patterns-operator.v0.0.36
        version: 0.0.36
      - name: patterns-operator.v0.0.35
        version: 0.0.35
      - name: patterns-operator.v0.0.34
        version: 0.0.34
      - name: patterns-operator.v0.0.33
        version: 0.0.33
      - name: patterns-operator.v0.0.32
        version: 0.0.32
      - name: patterns-operator.v0.0.31
        version: 0.0.31
      - name: patterns-operator.v0.0.30
        version: 0.0.30
      - name: patterns-operator.v0.0.29
        version: 0.0.29
      - name: patterns-operator.v0.0.28
        version: 0.0.28
      - name: patterns-operator.v0.0.27
        version: 0.0.27
      - name: patterns-operator.v0.0.26
        version: 0.0.26
      - name: patterns-operator.v0.0.25
        version: 0.0.25
      - name: patterns-operator.v0.0.24
        version: 0.0.24
      - name: patterns-operator.v0.0.23
        version: 0.0.23
      - name: patterns-operator.v0.0.22
        version: 0.0.22
      - name: patterns-operator.v0.0.21
        version: 0.0.21
      - name: patterns-operator.v0.0.20
        version: 0.0.20
      - name: patterns-operator.v0.0.19
        version: 0.0.19
      - name: patterns-operator.v0.0.18
        version: 0.0.18
      - name: patterns-operator.v0.0.17
        version: 0.0.17
      - name: patterns-operator.v0.0.16
        version: 0.0.16
      - name: patterns-operator.v0.0.15
        version: 0.0.15
      - name: patterns-operator.v0.0.14
        version: 0.0.14
      - name: patterns-operator.v0.0.13
        version: 0.0.13
      - name: patterns-operator.v0.0.12
        version: 0.0.12
      - name: patterns-operator.v0.0.11
        version: 0.0.11
      - name: patterns-operator.v0.0.10
        version: 0.0.10
      - name: patterns-operator.v0.0.9
        version: 0.0.9
      - name: patterns-operator.v0.0.8
        version: 0.0.8
      - name: patterns-operator.v0.0.7
        version: 0.0.7
      - name: patterns-operator.v0.0.6
        version: 0.0.6
      - name: patterns-operator.v0.0.5
        version: 0.0.5
      - name: patterns-operator.v0.0.4
        version: 0.0.4
      - name: patterns-operator.v0.0.3
        version: 0.0.3
      - name: patterns-operator.v0.0.2
        version: 0.0.2
      - name: patterns-operator.v0.0.1
        version: 0.0.1
      name: fast
    defaultChannel: fast
    packageName: patterns-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com/en/products/edge/validated-patterns
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sns-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sns-controller.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "PlatformApplication",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "PlatformEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sns-controller:1.2.2
          createdAt: "2025-09-30T21:19:34Z"
          description: AWS SNS controller is a service controller for managing SNS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PlatformApplication represents the state of an AWS sns PlatformApplication
              resource.
            displayName: PlatformApplication
            kind: PlatformApplication
            name: platformapplications.sns.services.k8s.aws
            version: v1alpha1
          - description: PlatformEndpoint represents the state of an AWS sns PlatformEndpoint
              resource.
            displayName: PlatformEndpoint
            kind: PlatformEndpoint
            name: platformendpoints.sns.services.k8s.aws
            version: v1alpha1
          - description: Subscription represents the state of an AWS sns Subscription
              resource.
            displayName: Subscription
            kind: Subscription
            name: subscriptions.sns.services.k8s.aws
            version: v1alpha1
          - description: Topic represents the state of an AWS sns Topic resource.
            displayName: Topic
            kind: Topic
            name: topics.sns.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Simple Notification Service (SNS) resources in AWS from within your Kubernetes cluster.

          **About Amazon SNS**

          Amazon Simple Notification Service (Amazon SNS) is a managed service that provides message delivery from publishers to subscribers (also known as producers and consumers). Publishers communicate asynchronously with subscribers by sending messages to a topic, which is a logical access point and communication channel. Clients can subscribe to the SNS topic and receive published messages using a supported endpoint type, such as Amazon Kinesis Data Firehose, Amazon SQS, AWS Lambda, HTTP, email, mobile push notifications, and mobile text messages (SMS).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SNS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sns
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SNS Developer Resources
          url: https://aws.amazon.com/sns/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sns maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sns-controller:1.2.2
        version: 1.2.2
      entries:
      - name: ack-sns-controller.v1.2.2
        version: 1.2.2
      - name: ack-sns-controller.v1.2.1
        version: 1.2.1
      - name: ack-sns-controller.v1.2.0
        version: 1.2.0
      - name: ack-sns-controller.v1.1.16
        version: 1.1.16
      - name: ack-sns-controller.v1.1.15
        version: 1.1.15
      - name: ack-sns-controller.v1.1.14
        version: 1.1.14
      - name: ack-sns-controller.v1.1.13
        version: 1.1.13
      - name: ack-sns-controller.v1.1.12
        version: 1.1.12
      - name: ack-sns-controller.v1.1.11
        version: 1.1.11
      - name: ack-sns-controller.v1.1.10
        version: 1.1.10
      - name: ack-sns-controller.v1.1.9
        version: 1.1.9
      - name: ack-sns-controller.v1.1.8
        version: 1.1.8
      - name: ack-sns-controller.v1.1.7
        version: 1.1.7
      - name: ack-sns-controller.v1.1.6
        version: 1.1.6
      - name: ack-sns-controller.v1.1.5
        version: 1.1.5
      - name: ack-sns-controller.v1.1.4
        version: 1.1.4
      - name: ack-sns-controller.v1.1.3
        version: 1.1.3
      - name: ack-sns-controller.v1.1.2
        version: 1.1.2
      - name: ack-sns-controller.v1.1.1
        version: 1.1.1
      - name: ack-sns-controller.v1.1.0
        version: 1.1.0
      - name: ack-sns-controller.v1.0.14
        version: 1.0.14
      - name: ack-sns-controller.v1.0.13
        version: 1.0.13
      - name: ack-sns-controller.v1.0.12
        version: 1.0.12
      - name: ack-sns-controller.v1.0.11
        version: 1.0.11
      - name: ack-sns-controller.v1.0.9
        version: 1.0.9
      - name: ack-sns-controller.v1.0.6
        version: 1.0.6
      - name: ack-sns-controller.v1.0.5
        version: 1.0.5
      - name: ack-sns-controller.v1.0.4
        version: 1.0.4
      - name: ack-sns-controller.v1.0.3
        version: 1.0.3
      - name: ack-sns-controller.v1.0.2
        version: 1.0.2
      - name: ack-sns-controller.v1.0.1
        version: 1.0.1
      - name: ack-sns-controller.v1.0.0
        version: 1.0.0
      - name: ack-sns-controller.v0.0.12
        version: 0.0.12
      - name: ack-sns-controller.v0.0.11
        version: 0.0.11
      - name: ack-sns-controller.v0.0.10
        version: 0.0.10
      - name: ack-sns-controller.v0.0.9
        version: 0.0.9
      - name: ack-sns-controller.v0.0.8
        version: 0.0.8
      - name: ack-sns-controller.v0.0.7
        version: 0.0.7
      - name: ack-sns-controller.v0.0.5
        version: 0.0.5
      - name: ack-sns-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-sns-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Helm Community
      provider-url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
    name: cockroachdb
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroachdb.v5.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1alpha1",
                "kind": "Cockroachdb",
                "metadata": {
                  "name": "cockroachdb-sample"
                },
                "spec": {
                  "clusterDomain": "cluster.local",
                  "conf": {
                    "attrs": [],
                    "cache": "25%",
                    "cluster-name": "",
                    "disable-cluster-name-verification": false,
                    "http-port": 8080,
                    "join": [],
                    "locality": "",
                    "logtostderr": "INFO",
                    "max-sql-memory": "25%",
                    "port": 26257,
                    "single-node": false,
                    "sql-audit-dir": ""
                  },
                  "image": {
                    "credentials": {},
                    "pullPolicy": "IfNotPresent",
                    "repository": "cockroachdb/cockroach",
                    "tag": "v20.2.4"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [],
                    "labels": {},
                    "paths": [
                      "/"
                    ],
                    "tls": []
                  },
                  "init": {
                    "affinity": {},
                    "annotations": {},
                    "labels": {
                      "app.kubernetes.io/component": "init"
                    },
                    "nodeSelector": {},
                    "resources": {},
                    "tolerations": []
                  },
                  "labels": {},
                  "networkPolicy": {
                    "enabled": false,
                    "ingress": {
                      "grpc": [],
                      "http": []
                    }
                  },
                  "service": {
                    "discovery": {
                      "annotations": {},
                      "labels": {
                        "app.kubernetes.io/component": "cockroachdb"
                      }
                    },
                    "ports": {
                      "grpc": {
                        "external": {
                          "name": "grpc",
                          "port": 26257
                        },
                        "internal": {
                          "name": "grpc-internal",
                          "port": 26257
                        }
                      },
                      "http": {
                        "name": "http",
                        "port": 8080
                      }
                    },
                    "public": {
                      "annotations": {},
                      "labels": {
                        "app.kubernetes.io/component": "cockroachdb"
                      },
                      "type": "ClusterIP"
                    }
                  },
                  "serviceMonitor": {
                    "annotations": {},
                    "enabled": false,
                    "interval": "10s",
                    "labels": {}
                  },
                  "statefulset": {
                    "annotations": {},
                    "args": [],
                    "budget": {
                      "maxUnavailable": 1
                    },
                    "env": [],
                    "labels": {
                      "app.kubernetes.io/component": "cockroachdb"
                    },
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "podAffinity": {},
                    "podAntiAffinity": {
                      "topologyKey": "kubernetes.io/hostname",
                      "type": "soft",
                      "weight": 100
                    },
                    "podManagementPolicy": "Parallel",
                    "priorityClassName": "",
                    "replicas": 3,
                    "resources": {},
                    "secretMounts": [],
                    "tolerations": [],
                    "topologySpreadConstraints": {
                      "maxSkew": 1,
                      "topologyKey": "topology.kubernetes.io/zone",
                      "whenUnsatisfiable": "ScheduleAnyway"
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    }
                  },
                  "storage": {
                    "hostPath": "",
                    "persistentVolume": {
                      "annotations": {},
                      "enabled": true,
                      "labels": {},
                      "size": "100Gi",
                      "storageClass": ""
                    }
                  },
                  "tls": {
                    "certs": {
                      "clientRootSecret": "cockroachdb-root",
                      "nodeSecret": "cockroachdb-node",
                      "provided": false,
                      "tlsSecret": false
                    },
                    "enabled": false,
                    "init": {
                      "image": {
                        "credentials": {},
                        "pullPolicy": "IfNotPresent",
                        "repository": "cockroachdb/cockroach-k8s-request-cert",
                        "tag": "0.4"
                      }
                    },
                    "serviceAccount": {
                      "create": true,
                      "name": ""
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          certified: "false"
          containerImage: quay.io/helmoperators/cockroachdb:v5.0.4
          createdAt: 2021-01-23T10-00-00Z
          description: CockroachDB Operator based on the CockroachDB helm chart
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/dmesser/cockroachdb-operator
          support: dmesser
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a CockroachDB cluster
            displayName: CockroachDB
            kind: Cockroachdb
            name: cockroachdbs.charts.operatorhub.io
            version: v1alpha1
        description: |
          CockroachDB is a scalable, survivable, strongly-consistent SQL database.
          ## About this Operator
          This Operator is based on a Helm chart for CockroachDB. It supports the same parameters like the official [CockroachDB Helm Chart](https://github.com/cockroachdb/helm-charts/tree/master/cockroachdb).

          ## Core capabilities
          * **StatefulSet** - Sets up a dynamically scalable CockroachDB cluster using a Kubernetes StatefulSet
          * **Expand Replicas** - Supports expanding the set of replicas by simply editing your object
          * **Dashboard** - Installs the CockroachDB user interface to administer your cluster. Easily expose it via an Ingress rule.
          Review all of the [confiuguration options](https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration) to best run your database instance. The example configuration is derived from the chart's [`values.yaml`](https://github.com/helm/charts/blob/master/stable/cockroachdb/values.yaml).
          ## Using the cluster
          The resulting cluster endpoint can be consumed from a `Service` that follows the pattern: `<StatefulSet-name>-public`. For example to connect using the command line client, use something like the following to obtain the name of the service:
          ```
          kubectl get service -l chart=cockroachdb-5.0.4
          NAME                                           TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)              AGE
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb          ClusterIP   None             <none>        26257/TCP,8080/TCP   24m
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public   ClusterIP   10.106.249.134   <none>        26257/TCP,8080/TCP   24m
          ```
          Then you can use the CockroachDB command line client to connect to the database cluster:
          ```
          kubectl run -it --rm cockroach-client --image=cockroachdb/cockroach --restart=Never --command -- ./cockroach sql --insecure --host example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public
          ```
          ## Before you start
          This Operator requires a cluster with PV support in order to run correctly.

          Note that for a production cluster, you will likely want to override the following parameters in [`values.yaml`](values.yaml) with your own values.

          - `statefulset.resources.requests.memory` and `statefulset.resources.limits.memory` allocate memory resources to CockroachDB pods in your cluster.
          - `conf.cache` and `conf.max-sql-memory` are memory limits that we recommend setting to 1/4 of the above resource allocation. When running CockroachDB, you must set these limits explicitly to avoid running out of memory.
          - `storage.persistentVolume.size` defaults to `100Gi` of disk space per pod, which you may increase or decrease for your use case.
          - `storage.persistentVolume.storageClass` uses the default storage class for your environment. We strongly recommend that you specify a storage class which uses an SSD.
          - `tls.enabled` must be set to `yes`/`true` to deploy in secure mode.

          For more information on overriding the `values.yaml` parameters, please see:
          > <https://www.cockroachlabs.com/docs/stable/orchestrate-cockroachdb-with-kubernetes.html#step-2-start-cockroachdb>

          If you are running in secure mode (with configuration parameter `tls.enabled` set to `yes`/`true`) and `tls.certs.provided` set to `no`/`false`), then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          When `tls.certs.provided` is set to `yes`/`true`, this chart will use certificates created outside of Kubernetes. You may want to use this if you want to use a different certificate authority from the one being used by Kubernetes or if your Kubernetes cluster doesn't fully support certificate-signing requests. To use this, first set up your certificates and load them into your Kubernetes cluster as Secrets using the commands below:

          ```
          mkdir certs
          mkdir my-safe-directory
          cockroach cert create-ca --certs-dir=certs --ca-key=my-safe-directory/ca.key
          cockroach cert create-client root --certs-dir=certs --ca-key=my-safe-directory/ca.key
          kubectl create secret generic cockroachdb-root --from-file=certs
          cockroach cert create-node --certs-dir=certs --ca-key=my-safe-directory/ca.key localhost 127.0.0.1 eerie-horse-cockroachdb-public eerie-horse-cockroachdb-public.default eerie-horse-cockroachdb-public.default.svc.cluster.local *.eerie-horse-cockroachdb *.eerie-horse-cockroachdb.default *.eerie-horse-cockroachdb.default.svc.cluster.local
          kubectl create secret generic cockroachdb-node --from-file=certs
          ```

          Set `tls.certs.tlsSecret` to `yes/true` if you make use of [cert-manager][3] in your cluster.

          [cert-manager][3] stores generated certificates in dedicated TLS secrets. Thus, they are always named:

          * `ca.crt`
          * `tls.crt`
          * `tls.key`

          On the other hand, CockroachDB also demands dedicated certificate filenames:

          * `ca.crt`
          * `node.crt`
          * `node.key`
          * `client.root.crt`
          * `client.root.key`

          By activating `tls.certs.tlsSecret` we benefit from projected secrets and convert the TLS secret filenames to their according CockroachDB filenames.

          If you are running in secure mode, then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          Confirm that all pods are `Running` successfully and init has been completed:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS      RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running     0          1m
          my-release-cockroachdb-1            1/1       Running     0          1m
          my-release-cockroachdb-2            1/1       Running     0          1m
          my-release-cockroachdb-init-k6jcr   0/1       Completed   0          1m
          ```

          Confirm that persistent volumes are created and claimed for each pod:

          ```shell
          kubectl get pv
          ```

          ```
          NAME                                       CAPACITY   ACCESS MODES   RECLAIM POLICY   STATUS    CLAIM                                      STORAGECLASS   REASON    AGE
          pvc-64878ebf-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-0   standard                 51s
          pvc-64945b4f-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-1   standard                 51s
          pvc-649d920d-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-2   standard                 51s
          ```

          ## Upgrading the cluster

          ### Chart version 3.0.0 and after

          Launch a temporary interactive pod and start the built-in SQL client:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          Set `cluster.preserve_downgrade_option`, where `$current_version` is the CockroachDB version currently running (e.g., `19.2`):

          ```sql
          > SET CLUSTER SETTING cluster.preserve_downgrade_option = '$current_version';
          ```

          Exit the shell and delete the temporary pod:

          ```sql
          > \q
          ```

          Kick off the upgrade process by changing the new Docker image, where `$new_version` is the CockroachDB version to which you are upgrading:

          ```shell
          kubectl delete job my-release-cockroachdb-init
          ```

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"image":{"tag":"<new version tag here>"}}}' --type merge
          ```

          Kubernetes will carry out a safe [rolling upgrade](https://kubernetes.io/docs/tutorials/stateful-application/basic-stateful-set/#updating-statefulsets) of your CockroachDB nodes one-by-one. Monitor the cluster's pods until all have been successfully restarted:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS              RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running             0          2m
          my-release-cockroachdb-1            1/1       Running             0          3m
          my-release-cockroachdb-2            1/1       Running             0          3m
          my-release-cockroachdb-3            0/1       ContainerCreating   0          25s
          my-release-cockroachdb-init-nwjkh   0/1       ContainerCreating   0          6s
          ```

          ```shell
          kubectl get pods \
          -o jsonpath='{range .items[*]}{.metadata.name}{"\t"}{.spec.containers[0].image}{"\n"}'
          ```

          ```
          my-release-cockroachdb-0    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-1    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-2    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-3    cockroachdb/cockroach:v20.2.4
          ```

          Resume normal operations. Once you are comfortable that the stability and performance of the cluster is what you'd expect post-upgrade, finalize the upgrade:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          ```sql
          > RESET CLUSTER SETTING cluster.preserve_downgrade_option;
          > \q
          ```

          ## Configuration

          The following table lists the configurable parameters of the CockroachDB chart and their default values.
          For details see the [`values.yaml`](values.yaml) file.

          | Parameter                                                 | Description                                                     | Default                                          |
          | ---------                                                 | -----------                                                     | -------                                          |
          | `clusterDomain`                                           | Cluster's default DNS domain                                    | `cluster.local`                                  |
          | `conf.attrs`                                              | CockroachDB node attributes                                     | `[]`                                             |
          | `conf.cache`                                              | Size of CockroachDB's in-memory cache                           | `25%`                                            |
          | `conf.cluster-name`                                       | Name of CockroachDB cluster                                     | `""`                                             |
          | `conf.disable-cluster-name-verification`                  | Disable CockroachDB cluster name verification                   | `no`                                             |
          | `conf.join`                                               | List of already-existing CockroachDB instances                  | `[]`                                             |
          | `conf.max-disk-temp-storage`                              | Max storage capacity for temp data                              | `0`                                              |
          | `conf.max-offset`                                         | Max allowed clock offset for CockroachDB cluster                | `500ms`                                          |
          | `conf.max-sql-memory`                                     | Max memory to use processing SQL querie                         | `25%`                                            |
          | `conf.locality`                                           | Locality attribute for this deployment                          | `""`                                             |
          | `conf.single-node`                                        | Disable CockroachDB clustering (standalone mode)                | `no`                                             |
          | `conf.sql-audit-dir`                                      | Directory for SQL audit log                                     | `""`                                             |
          | `conf.port`                                               | CockroachDB primary serving port in Pods                        | `26257`                                          |
          | `conf.http-port`                                          | CockroachDB HTTP port in Pods                                   | `8080`                                           |
          | `image.repository`                                        | Container image name                                            | `cockroachdb/cockroach`                          |
          | `image.tag`                                               | Container image tag                                             | `v20.2.4`                                        |
          | `image.pullPolicy`                                        | Container pull policy                                           | `IfNotPresent`                                   |
          | `image.credentials`                                       | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `statefulset.replicas`                                    | StatefulSet replicas number                                     | `3`                                              |
          | `statefulset.updateStrategy`                              | Update strategy for StatefulSet Pods                            | `{"type": "RollingUpdate"}`                      |
          | `statefulset.podManagementPolicy`                         | `OrderedReady`/`Parallel` Pods creation/deletion order          | `Parallel`                                       |
          | `statefulset.budget.maxUnavailable`                       | k8s PodDisruptionBudget parameter                               | `1`                                              |
          | `statefulset.args`                                        | Extra command-line arguments                                    | `[]`                                             |
          | `statefulset.env`                                         | Extra env vars                                                  | `[]`                                             |
          | `statefulset.secretMounts`                                | Additional Secrets to mount at cluster members                  | `[]`                                             |
          | `statefulset.labels`                                      | Additional labels of StatefulSet and its Pods                   | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `statefulset.annotations`                                 | Additional annotations of StatefulSet Pods                      | `{}`                                             |
          | `statefulset.nodeAffinity`                                | [Node affinity rules][2] of StatefulSet Pods                    | `{}`                                             |
          | `statefulset.podAffinity`                                 | [Inter-Pod affinity rules][1] of StatefulSet Pods               | `{}`                                             |
          | `statefulset.podAntiAffinity`                             | [Anti-affinity rules][1] of StatefulSet Pods                    | auto                                             |
          | `statefulset.podAntiAffinity.topologyKey`                 | The topologyKey for auto [anti-affinity rules][1]               | `kubernetes.io/hostname`                         |
          | `statefulset.podAntiAffinity.type`                        | Type of auto [anti-affinity rules][1]                           | `soft`                                           |
          | `statefulset.podAntiAffinity.weight`                      | Weight for `soft` auto [anti-affinity rules][1]                 | `100`                                            |
          | `statefulset.nodeSelector`                                | Node labels for StatefulSet Pods assignment                     | `{}`                                             |
          | `statefulset.priorityClassName`                           | [PriorityClassName][4] for StatefulSet Pods                     | `""`                                             |
          | `statefulset.tolerations`                                 | Node taints to tolerate by StatefulSet Pods                     | `[]`                                             |
          | `statefulset.topologySpreadConstraints`                   | [Topology Spread Constraints rules][5] of StatefulSet Pods      | auto                                             |
          | `statefulset.topologySpreadConstraints.maxSkew`           | Degree to which Pods may be unevenly distributed                | `1`                                              |
          | `statefulset.topologySpreadConstraints.topologyKey`       | The key of node labels                                          | `topology.kubernetes.io/zone`                    |
          | `statefulset.topologySpreadConstraints.whenUnsatisfiable` | `ScheduleAnyway`/`DoNotSchedule` for unsatisfiable constraints  | `ScheduleAnyway`                                 |
          | `statefulset.resources`                                   | Resource requests and limits for StatefulSet Pods               | `{}`                                             |
          | `service.ports.grpc.external.port`                        | CockroachDB primary serving port in Services                    | `26257`                                          |
          | `service.ports.grpc.external.name`                        | CockroachDB primary serving port name in Services               | `grpc`                                           |
          | `service.ports.grpc.internal.port`                        | CockroachDB inter-communication port in Services                | `26257`                                          |
          | `service.ports.grpc.internal.name`                        | CockroachDB inter-communication port name in Services           | `grpc-internal`                                  |
          | `service.ports.http.port`                                 | CockroachDB HTTP port in Services                               | `8080`                                           |
          | `service.ports.http.name`                                 | CockroachDB HTTP port name in Services                          | `http`                                           |
          | `service.public.type`                                     | Public Service type                                             | `ClusterIP`                                      |
          | `service.public.labels`                                   | Additional labels of public Service                             | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.public.annotations`                              | Additional annotations of public Service                        | `{}`                                             |
          | `service.discovery.labels`                                | Additional labels of discovery Service                          | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.discovery.annotations`                           | Additional annotations of discovery Service                     | `{}`                                             |
          | `ingress.enabled`                                         | Enable ingress resource for CockroachDB                         | `false`                                          |
          | `ingress.labels`                                          | Additional labels of Ingress                                    | `{}`                                             |
          | `ingress.annotations`                                     | Additional annotations of Ingress                               | `{}`                                             |
          | `ingress.paths`                                           | Paths for the default host                                      | `[/]`                                            |
          | `ingress.hosts`                                           | CockroachDB Ingress hostnames                                   | `[]`                                             |
          | `ingress.tls[0].hosts`                                    | CockroachDB Ingress tls hostnames                               | `nil`                                            |
          | `ingress.tls[0].secretName`                               | CockroachDB Ingress tls secret name                             | `nil`                                            |
          | `serviceMonitor.enabled`                                  | Create [ServiceMonitor](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/design.md#servicemonitor) Resource for scraping metrics using [PrometheusOperator](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/user-guides/getting-started.md#prometheus-operator)                     | `false`                                             |
          | `serviceMonitor.labels`                                   | Additional labels of ServiceMonitor                             | `{}`                                             |
          | `serviceMonitor.annotations`                              | Additional annotations of ServiceMonitor                        | `{}`                                             |
          | `serviceMonitor.interval`                                 | ServiceMonitor scrape metrics interval                          | `10s`                                            |
          | `serviceMonitor.scrapeTimeout`                            | ServiceMonitor scrape timeout                                   | `nil`                                            |
          | `storage.hostPath`                                        | Absolute path on host to store data                             | `""`                                             |
          | `storage.persistentVolume.enabled`                        | Whether to use PersistentVolume to store data                   | `yes`                                            |
          | `storage.persistentVolume.size`                           | PersistentVolume size                                           | `100Gi`                                          |
          | `storage.persistentVolume.storageClass`                   | PersistentVolume class                                          | `""`                                             |
          | `storage.persistentVolume.labels`                         | Additional labels of PersistentVolumeClaim                      | `{}`                                             |
          | `storage.persistentVolume.annotations`                    | Additional annotations of PersistentVolumeClaim                 | `{}`                                             |
          | `init.labels`                                             | Additional labels of init Job and its Pod                       | `{"app.kubernetes.io/component": "init"}`        |
          | `init.annotations`                                        | Additional labels of the Pod of init Job                        | `{}`                                             |
          | `init.affinity`                                           | [Affinity rules][2] of init Job Pod                             | `{}`                                             |
          | `init.nodeSelector`                                       | Node labels for init Job Pod assignment                         | `{}`                                             |
          | `init.tolerations`                                        | Node taints to tolerate by init Job Pod                         | `[]`                                             |
          | `init.resources`                                          | Resource requests and limits for the Pod of init Job            | `{}`                                             |
          | `tls.enabled`                                             | Whether to run securely using TLS certificates                  | `no`                                             |
          | `tls.serviceAccount.create`                               | Whether to create a new RBAC service account                    | `yes`                                            |
          | `tls.serviceAccount.name`                                 | Name of RBAC service account to use                             | `""`                                             |
          | `tls.certs.provided`                                      | Bring your own certs scenario, i.e certificates are provided    | `no`                                             |
          | `tls.certs.clientRootSecret`                              | If certs are provided, secret name for client root cert         | `cockroachdb-root`                               |
          | `tls.certs.nodeSecret`                                    | If certs are provided, secret name for node cert                | `cockroachdb-node`                               |
          | `tls.certs.tlsSecret`                                     | Own certs are stored in TLS secret                              | `no`                                             |
          | `tls.init.image.repository`                               | Image to use for requesting TLS certificates                    | `cockroachdb/cockroach-k8s-request-cert`         |
          | `tls.init.image.tag`                                      | Image tag to use for requesting TLS certificates                | `0.4`                                            |
          | `tls.init.image.pullPolicy`                               | Requesting TLS certificates container pull policy               | `IfNotPresent`                                   |
          | `tls.init.image.credentials`                              | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `networkPolicy.enabled`                                   | Enable NetworkPolicy for CockroachDB's Pods                     | `no`                                             |
          | `networkPolicy.ingress.grpc`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |
          | `networkPolicy.ingress.http`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |


          Override the default parameters using the `CockroachDB` custom resource.

          ## Deep dive

          ### Connecting to the CockroachDB cluster

          Once you've created the cluster, you can start talking to it by connecting to its `-public` Service. CockroachDB is PostgreSQL wire protocol compatible, so there's a [wide variety of supported clients](https://www.cockroachlabs.com/docs/install-client-drivers.html). As an example, we'll open up a SQL shell using CockroachDB's built-in shell and play around with it a bit, like this (likely needing to replace `my-release-cockroachdb-public` with the name of the `-public` Service that was created with your installed chart):

          ```shell
          kubectl run cockroach-client --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host my-release-cockroachdb-public
          ```

          ```
          Waiting for pod default/cockroach-client to be running, status is Pending,
          pod ready: false
          If you don't see a command prompt, try pressing enter.
          root@my-release-cockroachdb-public:26257> SHOW DATABASES;
          +--------------------+
          |      Database      |
          +--------------------+
          | information_schema |
          | pg_catalog         |
          | system             |
          +--------------------+
          (3 rows)
          root@my-release-cockroachdb-public:26257> CREATE DATABASE bank;
          CREATE DATABASE
          root@my-release-cockroachdb-public:26257> CREATE TABLE bank.accounts (id INT
          PRIMARY KEY, balance DECIMAL);
          CREATE TABLE
          root@my-release-cockroachdb-public:26257> INSERT INTO bank.accounts VALUES
          (1234, 10000.50);
          INSERT 1
          root@my-release-cockroachdb-public:26257> SELECT * FROM bank.accounts;
          +------+---------+
          |  id  | balance |
          +------+---------+
          | 1234 | 10000.5 |
          +------+---------+
          (1 row)
          root@my-release-cockroachdb-public:26257> \q
          Waiting for pod default/cockroach-client to terminate, status is Running
          pod "cockroach-client" deleted
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          ### Cluster health

          Because our pod spec includes regular health checks of the CockroachDB processes, simply running `kubectl get pods` and looking at the `STATUS` column is sufficient to determine the health of each instance in the cluster.

          If you want more detailed information about the cluster, the best place to look is the Admin UI.

          ### Accessing the Admin UI

          If you want to see information about how the cluster is doing, you can try pulling up the CockroachDB Admin UI by port-forwarding from your local machine to one of the pods (replacing `my-release-cockroachdb-0` with the name of one of your pods:

          ```shell
          kubectl port-forward my-release-cockroachdb-0 8080
          ```

          You should then be able to access the Admin UI by visiting <http://localhost:8080/> in your web browser.

          ### Failover

          If any CockroachDB member fails, it is restarted or recreated automatically by the Kubernetes infrastructure, and will re-join the cluster automatically when it comes back up. You can test this scenario by killing any of the CockroachDB pods:

          ```shell
          kubectl delete pod my-release-cockroachdb-1
          ```

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          After a while:

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-1  1/1       Running       0          20s
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          You can check the state of re-joining from the new pod's logs:

          ```shell
          kubectl logs my-release-cockroachdb-1
          ```

          ```
          [...]
          I161028 19:32:09.754026 1 server/node.go:586  [n1] node connected via gossip and
          verified as part of cluster {"35ecbc27-3f67-4e7d-9b8f-27c31aae17d6"}
          [...]
          cockroachdb-0.my-release-cockroachdb.default.svc.cluster.local:26257
          build:      beta-20161027-55-gd2d3c7f @ 2016/10/28 19:27:25 (go1.7.3)
          admin:      http://0.0.0.0:8080
          sql:
          postgresql://root@my-release-cockroachdb-1.my-release-cockroachdb.default.svc.cluster.local:26257?sslmode=disable
          logs:       cockroach-data/logs
          store[0]:   path=cockroach-data
          status:     restarted pre-existing node
          clusterID:  {35ecbc27-3f67-4e7d-9b8f-27c31aae17d6}
          nodeID:     2
          [...]
          ```

          ### NetworkPolicy

          To enable NetworkPolicy for CockroachDB, install [a networking plugin that implements the Kubernetes NetworkPolicy spec](https://kubernetes.io/docs/tasks/administer-cluster/declare-network-policy#before-you-begin), and set `networkPolicy.enabled` to `yes`/`true`.

          For Kubernetes v1.5 & v1.6, you must also turn on NetworkPolicy by setting the `DefaultDeny` Namespace annotation. Note: this will enforce policy for _all_ pods in the Namespace:

          ```shell
          kubectl annotate namespace default "net.beta.kubernetes.io/network-policy={\"ingress\":{\"isolation\":\"DefaultDeny\"}}"
          ```

          For more precise policy, set `networkPolicy.ingress.grpc` and `networkPolicy.ingress.http` rules. This will only allow pods that match the provided rules to connect to CockroachDB.

          ### Scaling

          Scaling should be managed via the `helm upgrade` command. After resizing your cluster on your cloud environment (e.g., GKE or EKS), run the following command to add a pod. This assumes you scaled from 3 to 4 nodes:

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"statefulset":{"replicas":"4"}}}' --type merge
          ```

          Note, that if you are running in secure mode (`tls.enabled` is `yes`/`true`) and increase the size of your cluster, you will also have to approve the CSR (certificate-signing request) of each new node (using `kubectl get csr` and `kubectl certificate approve`).

          [1]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [2]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#node-affinity
          [3]: https://cert-manager.io/
          [4]: https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass
          [5]: https://kubernetes.io/docs/concepts/workloads/pods/pod-topology-spread-constraints/
        displayName: CockroachDB Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cockroach
        - cockroachdb
        links:
        - name: Helm Chart Source
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb
        - name: Configuration Options
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration
        - name: CockroachDB Source
          url: https://github.com/cockroachdb/cockroach
        maintainers:
        - email: dmesser@redhat.com
          name: Daniel Messer
        maturity: stable
        minKubeVersion: 1.8.0
        provider:
          name: Helm Community
          url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/helmoperators/cockroachdb:v5.0.4
        version: 5.0.4
      entries:
      - name: cockroachdb.v5.0.4
        version: 5.0.4
      - name: cockroachdb.v5.0.3
        version: 5.0.3
      name: stable-5.x
    - currentCSV: cockroachdb.v6.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1alpha1",
                "kind": "Cockroachdb",
                "metadata": {
                  "name": "cockroachdb-sample"
                },
                "spec": {"image":{"repository":"cockroachdb/cockroach","tag":"v22.2.4","pullPolicy":"IfNotPresent","credentials":{}},"labels":{},"clusterDomain":"cluster.local","conf":{"attrs":[],"cache":"25%","cluster-name":"","disable-cluster-name-verification":false,"join":[],"log":{"enabled":false,"config":{}},"logtostderr":"INFO","max-sql-memory":"25%","locality":"","single-node":false,"sql-audit-dir":"","port":26257,"http-port":8080,"path":"cockroach-data","store":{"enabled":false,"type":null,"size":null,"attrs":null}},"statefulset":{"replicas":3,"updateStrategy":{"type":"RollingUpdate"},"podManagementPolicy":"Parallel","budget":{"maxUnavailable":1},"args":[],"env":[],"secretMounts":[],"labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{},"nodeAffinity":{},"podAffinity":{},"podAntiAffinity":{"topologyKey":"kubernetes.io/hostname","type":"soft","weight":100},"nodeSelector":{},"priorityClassName":"","tolerations":[],"topologySpreadConstraints":{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"},"resources":{},"customLivenessProbe":{},"customReadinessProbe":{}},"service":{"ports":{"grpc":{"external":{"port":26257,"name":"grpc"},"internal":{"port":26257,"name":"grpc-internal"}},"http":{"port":8080,"name":"http"}},"public":{"type":"ClusterIP","labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{}},"discovery":{"labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{}}},"ingress":{"enabled":false,"labels":{},"annotations":{},"paths":["/"],"hosts":[],"tls":[]},"prometheus":{"enabled":true},"serviceMonitor":{"enabled":false,"labels":{},"annotations":{},"interval":"10s","namespaced":false},"storage":{"hostPath":"","persistentVolume":{"enabled":true,"size":"100Gi","storageClass":"","labels":{},"annotations":{}}},"init":{"labels":{"app.kubernetes.io/component":"init"},"jobAnnotations":{},"annotations":{},"affinity":{},"nodeSelector":{},"tolerations":[],"resources":{},"provisioning":{"enabled":false,"clusterSettings":null,"users":[],"databases":[]}},"tls":{"enabled":true,"serviceAccount":{"create":true,"name":""},"copyCerts":{"image":"busybox"},"certs":{"provided":false,"clientRootSecret":"cockroachdb-root","nodeSecret":"cockroachdb-node","tlsSecret":false,"selfSigner":{"enabled":true,"caProvided":false,"caSecret":"","minimumCertDuration":"624h","caCertDuration":"43800h","caCertExpiryWindow":"648h","clientCertDuration":"672h","clientCertExpiryWindow":"48h","nodeCertDuration":"8760h","nodeCertExpiryWindow":"168h","rotateCerts":true,"readinessWait":"30s","podUpdateTimeout":"2m"},"certManager":false,"certManagerIssuer":{"group":"cert-manager.io","kind":"Issuer","name":"cockroachdb","clientCertDuration":"672h","clientCertExpiryWindow":"48h","nodeCertDuration":"8760h","nodeCertExpiryWindow":"168h"},"useCertManagerV1CRDs":false},"selfSigner":{"image":{"repository":"cockroachlabs-helm-charts/cockroach-self-signer-cert","tag":"1.3","pullPolicy":"IfNotPresent","credentials":{},"registry":"gcr.io"}}},"networkPolicy":{"enabled":false,"ingress":{"grpc":[],"http":[]}},"iap":{"enabled":false}}
              }
            ]
          capabilities: Basic Install
          categories: Database
          certified: "false"
          containerImage: quay.io/cockroachdb/cockroach-helm-operator:6.0.0
          createdAt: 2021-01-23T10-00-00Z
          description: CockroachDB Operator based on the CockroachDB helm chart
          olm.skipRange: <6.0.0
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/cockroachdb/helm-charts
          support: CockroachDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a CockroachDB cluster
            displayName: CockroachDB
            kind: Cockroachdb
            name: cockroachdbs.charts.operatorhub.io
            version: v1alpha1
        description: |
          CockroachDB is a scalable, survivable, strongly-consistent SQL database.
          ## About this Operator
          This Operator is based on a Helm chart for CockroachDB. It supports reconfiguration for some parameters, but notably does not handle scale down of the replica count in a seamless manner. Scale up works great.

          ## Core capabilities
          * **StatefulSet** - Sets up a dynamically scalable CockroachDB cluster using a Kubernetes StatefulSet
          * **Expand Replicas** - Supports expanding the set of replicas by simply editing your object
          * **Dashboard** - Installs the CockroachDB user interface to administer your cluster. Easily expose it via an Ingress rule.
          Review all of the [confiuguration options](https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration) to best run your database instance. The example configuration is derived from the chart's [`values.yaml`](https://github.com/helm/charts/blob/master/stable/cockroachdb/values.yaml).
          ## Using the cluster
          The resulting cluster endpoint can be consumed from a `Service` that follows the pattern: `<StatefulSet-name>-public`. For example to connect using the command line client, use something like the following to obtain the name of the service:
          ```
          kubectl get service -l chart=cockroachdb-6.0.0
          NAME                                           TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)              AGE
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb          ClusterIP   None             <none>        26257/TCP,8080/TCP   24m
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public   ClusterIP   10.106.249.134   <none>        26257/TCP,8080/TCP   24m
          ```
          Then you can use the CockroachDB command line client to connect to the database cluster:
          ```
          kubectl run -it --rm cockroach-client --image=cockroachdb/cockroach --restart=Never --command -- ./cockroach sql --insecure --host example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public
          ```
          ## Before you start
          This Operator requires a cluster with PV support in order to run correctly.

          Note that for a production cluster, you will likely want to override the following parameters in [`values.yaml`](values.yaml) with your own values.

          - `statefulset.resources.requests.memory` and `statefulset.resources.limits.memory` allocate memory resources to CockroachDB pods in your cluster.
          - `conf.cache` and `conf.max-sql-memory` are memory limits that we recommend setting to 1/4 of the above resource allocation. When running CockroachDB, you must set these limits explicitly to avoid running out of memory.
          - `storage.persistentVolume.size` defaults to `100Gi` of disk space per pod, which you may increase or decrease for your use case.
          - `storage.persistentVolume.storageClass` uses the default storage class for your environment. We strongly recommend that you specify a storage class which uses an SSD.
          - `tls.enabled` must be set to `yes`/`true` to deploy in secure mode.

          For more information on overriding the `values.yaml` parameters, please see:
          > <https://www.cockroachlabs.com/docs/stable/orchestrate-cockroachdb-with-kubernetes.html#step-2-start-cockroachdb>

          If you are running in secure mode (with configuration parameter `tls.enabled` set to `yes`/`true`) and `tls.certs.provided` set to `no`/`false`), then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          When `tls.certs.provided` is set to `yes`/`true`, this chart will use certificates created outside of Kubernetes. You may want to use this if you want to use a different certificate authority from the one being used by Kubernetes or if your Kubernetes cluster doesn't fully support certificate-signing requests. To use this, first set up your certificates and load them into your Kubernetes cluster as Secrets using the commands below:

          ```
          mkdir certs
          mkdir my-safe-directory
          cockroach cert create-ca --certs-dir=certs --ca-key=my-safe-directory/ca.key
          cockroach cert create-client root --certs-dir=certs --ca-key=my-safe-directory/ca.key
          kubectl create secret generic cockroachdb-root --from-file=certs
          cockroach cert create-node --certs-dir=certs --ca-key=my-safe-directory/ca.key localhost 127.0.0.1 eerie-horse-cockroachdb-public eerie-horse-cockroachdb-public.default eerie-horse-cockroachdb-public.default.svc.cluster.local *.eerie-horse-cockroachdb *.eerie-horse-cockroachdb.default *.eerie-horse-cockroachdb.default.svc.cluster.local
          kubectl create secret generic cockroachdb-node --from-file=certs
          ```

          Set `tls.certs.tlsSecret` to `yes/true` if you make use of [cert-manager][3] in your cluster.

          [cert-manager][3] stores generated certificates in dedicated TLS secrets. Thus, they are always named:

          * `ca.crt`
          * `tls.crt`
          * `tls.key`

          On the other hand, CockroachDB also demands dedicated certificate filenames:

          * `ca.crt`
          * `node.crt`
          * `node.key`
          * `client.root.crt`
          * `client.root.key`

          By activating `tls.certs.tlsSecret` we benefit from projected secrets and convert the TLS secret filenames to their according CockroachDB filenames.

          If you are running in secure mode, then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          Confirm that all pods are `Running` successfully and init has been completed:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS      RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running     0          1m
          my-release-cockroachdb-1            1/1       Running     0          1m
          my-release-cockroachdb-2            1/1       Running     0          1m
          my-release-cockroachdb-init-k6jcr   0/1       Completed   0          1m
          ```

          Confirm that persistent volumes are created and claimed for each pod:

          ```shell
          kubectl get pv
          ```

          ```
          NAME                                       CAPACITY   ACCESS MODES   RECLAIM POLICY   STATUS    CLAIM                                      STORAGECLASS   REASON    AGE
          pvc-64878ebf-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-0   standard                 51s
          pvc-64945b4f-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-1   standard                 51s
          pvc-649d920d-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-2   standard                 51s
          ```

          ## Upgrading the cluster

          ### Chart version 3.0.0 and after

          Launch a temporary interactive pod and start the built-in SQL client:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          Set `cluster.preserve_downgrade_option`, where `$current_version` is the CockroachDB version currently running (e.g., `19.2`):

          ```sql
          > SET CLUSTER SETTING cluster.preserve_downgrade_option = '$current_version';
          ```

          Exit the shell and delete the temporary pod:

          ```sql
          > \q
          ```

          Kick off the upgrade process by changing the new Docker image, where `$new_version` is the CockroachDB version to which you are upgrading:

          ```shell
          kubectl delete job my-release-cockroachdb-init
          ```

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"image":{"tag":"<new version tag here>"}}}' --type merge
          ```

          Kubernetes will carry out a safe [rolling upgrade](https://kubernetes.io/docs/tutorials/stateful-application/basic-stateful-set/#updating-statefulsets) of your CockroachDB nodes one-by-one. Monitor the cluster's pods until all have been successfully restarted:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS              RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running             0          2m
          my-release-cockroachdb-1            1/1       Running             0          3m
          my-release-cockroachdb-2            1/1       Running             0          3m
          my-release-cockroachdb-3            0/1       ContainerCreating   0          25s
          my-release-cockroachdb-init-nwjkh   0/1       ContainerCreating   0          6s
          ```

          ```shell
          kubectl get pods \
          -o jsonpath='{range .items[*]}{.metadata.name}{"\t"}{.spec.containers[0].image}{"\n"}'
          ```

          ```
          my-release-cockroachdb-0    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-1    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-2    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-3    cockroachdb/cockroach:v20.2.4
          ```

          Resume normal operations. Once you are comfortable that the stability and performance of the cluster is what you'd expect post-upgrade, finalize the upgrade:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          ```sql
          > RESET CLUSTER SETTING cluster.preserve_downgrade_option;
          > \q
          ```

          ## Configuration

          The following table lists the configurable parameters of the CockroachDB chart and their default values.
          For details see the [`values.yaml`](values.yaml) file.

          | Parameter                                                 | Description                                                     | Default                                          |
          | ---------                                                 | -----------                                                     | -------                                          |
          | `clusterDomain`                                           | Cluster's default DNS domain                                    | `cluster.local`                                  |
          | `conf.attrs`                                              | CockroachDB node attributes                                     | `[]`                                             |
          | `conf.cache`                                              | Size of CockroachDB's in-memory cache                           | `25%`                                            |
          | `conf.cluster-name`                                       | Name of CockroachDB cluster                                     | `""`                                             |
          | `conf.disable-cluster-name-verification`                  | Disable CockroachDB cluster name verification                   | `no`                                             |
          | `conf.join`                                               | List of already-existing CockroachDB instances                  | `[]`                                             |
          | `conf.max-disk-temp-storage`                              | Max storage capacity for temp data                              | `0`                                              |
          | `conf.max-offset`                                         | Max allowed clock offset for CockroachDB cluster                | `500ms`                                          |
          | `conf.max-sql-memory`                                     | Max memory to use processing SQL querie                         | `25%`                                            |
          | `conf.locality`                                           | Locality attribute for this deployment                          | `""`                                             |
          | `conf.single-node`                                        | Disable CockroachDB clustering (standalone mode)                | `no`                                             |
          | `conf.sql-audit-dir`                                      | Directory for SQL audit log                                     | `""`                                             |
          | `conf.port`                                               | CockroachDB primary serving port in Pods                        | `26257`                                          |
          | `conf.http-port`                                          | CockroachDB HTTP port in Pods                                   | `8080`                                           |
          | `image.repository`                                        | Container image name                                            | `cockroachdb/cockroach`                          |
          | `image.tag`                                               | Container image tag                                             | `v20.2.4`                                        |
          | `image.pullPolicy`                                        | Container pull policy                                           | `IfNotPresent`                                   |
          | `image.credentials`                                       | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `statefulset.replicas`                                    | StatefulSet replicas number                                     | `3`                                              |
          | `statefulset.updateStrategy`                              | Update strategy for StatefulSet Pods                            | `{"type": "RollingUpdate"}`                      |
          | `statefulset.podManagementPolicy`                         | `OrderedReady`/`Parallel` Pods creation/deletion order          | `Parallel`                                       |
          | `statefulset.budget.maxUnavailable`                       | k8s PodDisruptionBudget parameter                               | `1`                                              |
          | `statefulset.args`                                        | Extra command-line arguments                                    | `[]`                                             |
          | `statefulset.env`                                         | Extra env vars                                                  | `[]`                                             |
          | `statefulset.secretMounts`                                | Additional Secrets to mount at cluster members                  | `[]`                                             |
          | `statefulset.labels`                                      | Additional labels of StatefulSet and its Pods                   | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `statefulset.annotations`                                 | Additional annotations of StatefulSet Pods                      | `{}`                                             |
          | `statefulset.nodeAffinity`                                | [Node affinity rules][2] of StatefulSet Pods                    | `{}`                                             |
          | `statefulset.podAffinity`                                 | [Inter-Pod affinity rules][1] of StatefulSet Pods               | `{}`                                             |
          | `statefulset.podAntiAffinity`                             | [Anti-affinity rules][1] of StatefulSet Pods                    | auto                                             |
          | `statefulset.podAntiAffinity.topologyKey`                 | The topologyKey for auto [anti-affinity rules][1]               | `kubernetes.io/hostname`                         |
          | `statefulset.podAntiAffinity.type`                        | Type of auto [anti-affinity rules][1]                           | `soft`                                           |
          | `statefulset.podAntiAffinity.weight`                      | Weight for `soft` auto [anti-affinity rules][1]                 | `100`                                            |
          | `statefulset.nodeSelector`                                | Node labels for StatefulSet Pods assignment                     | `{}`                                             |
          | `statefulset.priorityClassName`                           | [PriorityClassName][4] for StatefulSet Pods                     | `""`                                             |
          | `statefulset.tolerations`                                 | Node taints to tolerate by StatefulSet Pods                     | `[]`                                             |
          | `statefulset.topologySpreadConstraints`                   | [Topology Spread Constraints rules][5] of StatefulSet Pods      | auto                                             |
          | `statefulset.topologySpreadConstraints.maxSkew`           | Degree to which Pods may be unevenly distributed                | `1`                                              |
          | `statefulset.topologySpreadConstraints.topologyKey`       | The key of node labels                                          | `topology.kubernetes.io/zone`                    |
          | `statefulset.topologySpreadConstraints.whenUnsatisfiable` | `ScheduleAnyway`/`DoNotSchedule` for unsatisfiable constraints  | `ScheduleAnyway`                                 |
          | `statefulset.resources`                                   | Resource requests and limits for StatefulSet Pods               | `{}`                                             |
          | `service.ports.grpc.external.port`                        | CockroachDB primary serving port in Services                    | `26257`                                          |
          | `service.ports.grpc.external.name`                        | CockroachDB primary serving port name in Services               | `grpc`                                           |
          | `service.ports.grpc.internal.port`                        | CockroachDB inter-communication port in Services                | `26257`                                          |
          | `service.ports.grpc.internal.name`                        | CockroachDB inter-communication port name in Services           | `grpc-internal`                                  |
          | `service.ports.http.port`                                 | CockroachDB HTTP port in Services                               | `8080`                                           |
          | `service.ports.http.name`                                 | CockroachDB HTTP port name in Services                          | `http`                                           |
          | `service.public.type`                                     | Public Service type                                             | `ClusterIP`                                      |
          | `service.public.labels`                                   | Additional labels of public Service                             | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.public.annotations`                              | Additional annotations of public Service                        | `{}`                                             |
          | `service.discovery.labels`                                | Additional labels of discovery Service                          | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.discovery.annotations`                           | Additional annotations of discovery Service                     | `{}`                                             |
          | `ingress.enabled`                                         | Enable ingress resource for CockroachDB                         | `false`                                          |
          | `ingress.labels`                                          | Additional labels of Ingress                                    | `{}`                                             |
          | `ingress.annotations`                                     | Additional annotations of Ingress                               | `{}`                                             |
          | `ingress.paths`                                           | Paths for the default host                                      | `[/]`                                            |
          | `ingress.hosts`                                           | CockroachDB Ingress hostnames                                   | `[]`                                             |
          | `ingress.tls[0].hosts`                                    | CockroachDB Ingress tls hostnames                               | `nil`                                            |
          | `ingress.tls[0].secretName`                               | CockroachDB Ingress tls secret name                             | `nil`                                            |
          | `serviceMonitor.enabled`                                  | Create [ServiceMonitor](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/design.md#servicemonitor) Resource for scraping metrics using [PrometheusOperator](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/user-guides/getting-started.md#prometheus-operator)                     | `false`                                             |
          | `serviceMonitor.labels`                                   | Additional labels of ServiceMonitor                             | `{}`                                             |
          | `serviceMonitor.annotations`                              | Additional annotations of ServiceMonitor                        | `{}`                                             |
          | `serviceMonitor.interval`                                 | ServiceMonitor scrape metrics interval                          | `10s`                                            |
          | `serviceMonitor.scrapeTimeout`                            | ServiceMonitor scrape timeout                                   | `nil`                                            |
          | `storage.hostPath`                                        | Absolute path on host to store data                             | `""`                                             |
          | `storage.persistentVolume.enabled`                        | Whether to use PersistentVolume to store data                   | `yes`                                            |
          | `storage.persistentVolume.size`                           | PersistentVolume size                                           | `100Gi`                                          |
          | `storage.persistentVolume.storageClass`                   | PersistentVolume class                                          | `""`                                             |
          | `storage.persistentVolume.labels`                         | Additional labels of PersistentVolumeClaim                      | `{}`                                             |
          | `storage.persistentVolume.annotations`                    | Additional annotations of PersistentVolumeClaim                 | `{}`                                             |
          | `init.labels`                                             | Additional labels of init Job and its Pod                       | `{"app.kubernetes.io/component": "init"}`        |
          | `init.annotations`                                        | Additional labels of the Pod of init Job                        | `{}`                                             |
          | `init.affinity`                                           | [Affinity rules][2] of init Job Pod                             | `{}`                                             |
          | `init.nodeSelector`                                       | Node labels for init Job Pod assignment                         | `{}`                                             |
          | `init.tolerations`                                        | Node taints to tolerate by init Job Pod                         | `[]`                                             |
          | `init.resources`                                          | Resource requests and limits for the Pod of init Job            | `{}`                                             |
          | `tls.enabled`                                             | Whether to run securely using TLS certificates                  | `no`                                             |
          | `tls.serviceAccount.create`                               | Whether to create a new RBAC service account                    | `yes`                                            |
          | `tls.serviceAccount.name`                                 | Name of RBAC service account to use                             | `""`                                             |
          | `tls.certs.provided`                                      | Bring your own certs scenario, i.e certificates are provided    | `no`                                             |
          | `tls.certs.clientRootSecret`                              | If certs are provided, secret name for client root cert         | `cockroachdb-root`                               |
          | `tls.certs.nodeSecret`                                    | If certs are provided, secret name for node cert                | `cockroachdb-node`                               |
          | `tls.certs.tlsSecret`                                     | Own certs are stored in TLS secret                              | `no`                                             |
          | `tls.init.image.repository`                               | Image to use for requesting TLS certificates                    | `cockroachdb/cockroach-k8s-request-cert`         |
          | `tls.init.image.tag`                                      | Image tag to use for requesting TLS certificates                | `0.4`                                            |
          | `tls.init.image.pullPolicy`                               | Requesting TLS certificates container pull policy               | `IfNotPresent`                                   |
          | `tls.init.image.credentials`                              | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `networkPolicy.enabled`                                   | Enable NetworkPolicy for CockroachDB's Pods                     | `no`                                             |
          | `networkPolicy.ingress.grpc`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |
          | `networkPolicy.ingress.http`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |


          Override the default parameters using the `CockroachDB` custom resource.

          ## Deep dive

          ### Connecting to the CockroachDB cluster

          Once you've created the cluster, you can start talking to it by connecting to its `-public` Service. CockroachDB is PostgreSQL wire protocol compatible, so there's a [wide variety of supported clients](https://www.cockroachlabs.com/docs/install-client-drivers.html). As an example, we'll open up a SQL shell using CockroachDB's built-in shell and play around with it a bit, like this (likely needing to replace `my-release-cockroachdb-public` with the name of the `-public` Service that was created with your installed chart):

          ```shell
          kubectl run cockroach-client --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host my-release-cockroachdb-public
          ```

          ```
          Waiting for pod default/cockroach-client to be running, status is Pending,
          pod ready: false
          If you don't see a command prompt, try pressing enter.
          root@my-release-cockroachdb-public:26257> SHOW DATABASES;
          +--------------------+
          |      Database      |
          +--------------------+
          | information_schema |
          | pg_catalog         |
          | system             |
          +--------------------+
          (3 rows)
          root@my-release-cockroachdb-public:26257> CREATE DATABASE bank;
          CREATE DATABASE
          root@my-release-cockroachdb-public:26257> CREATE TABLE bank.accounts (id INT
          PRIMARY KEY, balance DECIMAL);
          CREATE TABLE
          root@my-release-cockroachdb-public:26257> INSERT INTO bank.accounts VALUES
          (1234, 10000.50);
          INSERT 1
          root@my-release-cockroachdb-public:26257> SELECT * FROM bank.accounts;
          +------+---------+
          |  id  | balance |
          +------+---------+
          | 1234 | 10000.5 |
          +------+---------+
          (1 row)
          root@my-release-cockroachdb-public:26257> \q
          Waiting for pod default/cockroach-client to terminate, status is Running
          pod "cockroach-client" deleted
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          ### Cluster health

          Because our pod spec includes regular health checks of the CockroachDB processes, simply running `kubectl get pods` and looking at the `STATUS` column is sufficient to determine the health of each instance in the cluster.

          If you want more detailed information about the cluster, the best place to look is the Admin UI.

          ### Accessing the Admin UI

          If you want to see information about how the cluster is doing, you can try pulling up the CockroachDB Admin UI by port-forwarding from your local machine to one of the pods (replacing `my-release-cockroachdb-0` with the name of one of your pods:

          ```shell
          kubectl port-forward my-release-cockroachdb-0 8080
          ```

          You should then be able to access the Admin UI by visiting <http://localhost:8080/> in your web browser.

          ### Failover

          If any CockroachDB member fails, it is restarted or recreated automatically by the Kubernetes infrastructure, and will re-join the cluster automatically when it comes back up. You can test this scenario by killing any of the CockroachDB pods:

          ```shell
          kubectl delete pod my-release-cockroachdb-1
          ```

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          After a while:

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-1  1/1       Running       0          20s
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          You can check the state of re-joining from the new pod's logs:

          ```shell
          kubectl logs my-release-cockroachdb-1
          ```

          ```
          [...]
          I161028 19:32:09.754026 1 server/node.go:586  [n1] node connected via gossip and
          verified as part of cluster {"35ecbc27-3f67-4e7d-9b8f-27c31aae17d6"}
          [...]
          cockroachdb-0.my-release-cockroachdb.default.svc.cluster.local:26257
          build:      beta-20161027-55-gd2d3c7f @ 2016/10/28 19:27:25 (go1.7.3)
          admin:      http://0.0.0.0:8080
          sql:
          postgresql://root@my-release-cockroachdb-1.my-release-cockroachdb.default.svc.cluster.local:26257?sslmode=disable
          logs:       cockroach-data/logs
          store[0]:   path=cockroach-data
          status:     restarted pre-existing node
          clusterID:  {35ecbc27-3f67-4e7d-9b8f-27c31aae17d6}
          nodeID:     2
          [...]
          ```

          ### NetworkPolicy

          To enable NetworkPolicy for CockroachDB, install [a networking plugin that implements the Kubernetes NetworkPolicy spec](https://kubernetes.io/docs/tasks/administer-cluster/declare-network-policy#before-you-begin), and set `networkPolicy.enabled` to `yes`/`true`.

          For Kubernetes v1.5 & v1.6, you must also turn on NetworkPolicy by setting the `DefaultDeny` Namespace annotation. Note: this will enforce policy for _all_ pods in the Namespace:

          ```shell
          kubectl annotate namespace default "net.beta.kubernetes.io/network-policy={\"ingress\":{\"isolation\":\"DefaultDeny\"}}"
          ```

          For more precise policy, set `networkPolicy.ingress.grpc` and `networkPolicy.ingress.http` rules. This will only allow pods that match the provided rules to connect to CockroachDB.

          ### Scaling

          Scaling should be managed via the `helm upgrade` command. After resizing your cluster on your cloud environment (e.g., GKE or EKS), run the following command to add a pod. This assumes you scaled from 3 to 4 nodes:

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"statefulset":{"replicas":"4"}}}' --type merge
          ```

          Note, that if you are running in secure mode (`tls.enabled` is `yes`/`true`) and increase the size of your cluster, you will also have to approve the CSR (certificate-signing request) of each new node (using `kubectl get csr` and `kubectl certificate approve`).

          [1]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [2]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#node-affinity
          [3]: https://cert-manager.io/
          [4]: https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass
          [5]: https://kubernetes.io/docs/concepts/workloads/pods/pod-topology-spread-constraints/
        displayName: CockroachDB Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cockroach
        - cockroachdb
        links:
        - name: Helm Chart Source
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb
        - name: Configuration Options
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration
        - name: CockroachDB Source
          url: https://github.com/cockroachdb/cockroach
        maintainers:
        - email: yandu@cockroachlabs.com
          name: Yandu Oppacher
        maturity: stable
        minKubeVersion: 1.8.0
        provider:
          name: Helm Community
          url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/cockroachdb/cockroach-helm-operator:6.0.0
        version: 6.0.0
      entries:
      - name: cockroachdb.v6.0.0
        version: 6.0.0
      name: stable-v6.x
    defaultChannel: stable-v6.x
    packageName: cockroachdb
    provider:
      name: Helm Community
      url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CrowdStrike
      provider-url: https://crowdStrike.com
    name: falcon-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: falcon-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconContainer",
                "metadata": {
                  "name": "falcon-sidecar-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "sidecar"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconNodeSensor",
                "metadata": {
                  "name": "falcon-node-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "daemonset"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security,Monitoring
          containerImage: quay.io/crowdstrike/falcon-operator:0.8.1
          createdAt: "2024-06-06T10:00:00Z"
          description: '[DEPRECATED] Use the CrowdStrike Falcon Operator from the
            certified channel instead'
          operatorframework.io/suggested-namespace: falcon-operator
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/CrowdStrike/falcon-operator
          support: Community Only
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FalconContainer is the Schema for the falconcontainers API
            displayName: Falcon Container
            kind: FalconContainer
            name: falconcontainers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconNodeSensor is the Schema for the falconnodesensors
              API
            displayName: Falcon Node Sensor
            kind: FalconNodeSensor
            name: falconnodesensors.falcon.crowdstrike.com
            version: v1alpha1
        description: |-
          ## Deprecated
          This CrowdStrike community operator is deprecated, use the CrowdStrike Falcon Operator from the certified channel instead.

          ## About the CrowdStrike Falcon Operator
          The CrowdStrike Falcon Operator deploys CrowdStrike Falcon Workload Protection to the cluster. The operator exposes 2 custom resources that allows you to deploy either the Falcon Container Sensor or Falcon Node Sensor.

          ## About Custom Resources

          | Custom Resource                                       | Description                                                      |
          | :--------                                             | :------------                                                    |
          | [FalconContainer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/container/README.md) | Manages installation of Falcon Container Sensor on the cluster   |
          | [FalconNodeSensor](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/node/README.md)     | Manages installation of Falcon Linux Sensor on the cluster nodes |

          ## Installation and Deployment

          For installation and deployment of the CrowdStrike Falcon Operator and its Custom Resources, please read the [Installation and Deployment Guide](./docs/install_guide.md) and choose the deployment method that is right for your target environment.

          ## Getting Help
          If you encounter any issues while using the Falcon Operator, you can create an issue on our [Github repo](https://github.com/CrowdStrike/falcon-operator) for bugs, enhancements, or other requests.

          ## Contributing
          You can contribute by:

          * Raising any issues you find using Falcon Operator
          * Fixing issues by opening [Pull Requests](https://github.com/CrowdStrike/falcon-operator/pulls)
          * Improving documentation
          * Talking about the CrowdStrike Falcon Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/CrowdStrike/falcon-operator/issues).

          ## Additional Resources
           - CrowdStrike Container Security: [Product Page](https://www.crowdstrike.com/products/cloud-security/falcon-cloud-workload-protection/container-security/)
           - So You Think Your Containers Are Secure? Four Steps to Ensure a Secure Container Deployment: [Blog Post](https://www.crowdstrike.com/blog/four-steps-to-ensure-a-secure-containter-deployment/)
           - Container Security With CrowdStrike: [Blog Post](https://www.crowdstrike.com/blog/tech-center/container-security/)
           - To learn more about Falcon Container Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/support/documentation/146/falcon-container-sensor-for-linux), [Release Notes](https://falcon.crowdstrike.com/support/news/release-notes-falcon-container-sensor-for-linux)
           - To learn more about Falcon Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/documentation/20/falcon-sensor-for-linux#kubernetes-support)
           - [Developer Documentation](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/developer_guide.md)
        displayName: '[DEPRECATED] CrowdStrike Operator'
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - runtime-protection
        links:
        - name: Falcon Operator
          url: https://github.com/CrowdStrike/falcon-operator
        maintainers:
        - email: integrations@crowdstrike.com
          name: CrowdStrike Solutions Architecture
        maturity: alpha
        minKubeVersion: 1.17.0
        provider:
          name: CrowdStrike
          url: https://crowdStrike.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/crowdstrike/falcon-operator:0.8.1
        version: 0.8.2
      entries:
      - name: falcon-operator.v0.8.2
        version: 0.8.2
      - name: falcon-operator.v0.8.1
        version: 0.8.1
      name: alpha
    defaultChannel: alpha
    packageName: falcon-operator
    provider:
      name: CrowdStrike
      url: https://crowdStrike.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Foundation Model Stack
      provider-url: ""
    name: multi-nic-cni-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multi-nic-cni-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.3.0",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.3.0",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.3.0
          createdAt: "2024-04-17T04:51:40Z"
          description: Automate container multi-network configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: http://github.com/foundation-model-stack/multi-nic-cni
          support: Foundation Model Stack Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Foundation Model Stack
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.3.0
        version: 1.3.0
      entries:
      - name: multi-nic-cni-operator.v1.3.0
        version: 1.3.0
      - name: multi-nic-cni-operator.v1.2.9
        version: 1.2.9
      - name: multi-nic-cni-operator.v1.2.8
        version: 1.2.8
      - name: multi-nic-cni-operator.v1.2.2
        version: 1.2.2
      - name: multi-nic-cni-operator.v1.2.1
        version: 1.2.1
      - name: multi-nic-cni-operator.v1.0.2
        version: 1.0.2
      name: alpha
    - currentCSV: multi-nic-cni-operator.v1.2.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.7",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.7",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.7
          createdAt: "2025-05-09T12:45:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: minzhang@redhat.com
          name: Min Zhang
        - email: ebendror@redhat.com
          name: Eran Ben Dror
        maturity: alpha
        provider:
          name: Foundation Model Stack
        relatedImages:
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.7
        version: 1.2.7
      entries:
      - name: multi-nic-cni-operator.v1.2.7
        version: 1.2.7
      - name: multi-nic-cni-operator.v1.2.3
        version: 1.2.3
      - name: multi-nic-cni-operator.v1.2.0
        version: 1.2.0
      - name: multi-nic-cni-operator.v1.1.0
        version: 1.1.0
      name: beta
    - currentCSV: multi-nic-cni-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.6",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.6",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.6
          createdAt: "2025-03-24T14:49:54Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: minzhang@redhat.com
          name: Min Zhang
        - email: ebendror@redhat.com
          name: Eran Ben Dror
        maturity: alpha
        provider:
          name: Foundation Model Stack
        relatedImages:
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.6
        version: 1.2.6
      entries:
      - name: multi-nic-cni-operator.v1.2.6
        version: 1.2.6
      - name: multi-nic-cni-operator.v1.2.5
        version: 1.2.5
      - name: multi-nic-cni-operator.v1.2.4
        version: 1.2.4
      - name: multi-nic-cni-operator.v1.0.5
        version: 1.0.5
      - name: multi-nic-cni-operator.v1.0.4
        version: 1.0.4
      - name: multi-nic-cni-operator.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: multi-nic-cni-operator
    provider:
      name: Foundation Model Stack
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: assisted-service-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: assisted-service-operator.v0.7.74
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-x86_64-live.x86_64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-aarch64-live.aarch64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-s390x-live.s390x.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-x86_64-live.x86_64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-aarch64-live.aarch64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-s390x-live.s390x.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-x86_64-live.x86_64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-aarch64-live.aarch64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-ppc64le-live.ppc64le.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-s390x-live.s390x.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-x86_64-live.x86_64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-aarch64-live.aarch64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-ppc64le-live.ppc64le.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-s390x-live.s390x.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-s390x-live.s390x.iso",
                      "version": "418.94.202409162337-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service:latest
          createdAt: "2024-10-16T17:43:34Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.74'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AgentClassification is the Schema for the AgentClassifications
              API
            displayName: Agent Classification
            kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - description: Agent is the Schema for the hosts API
            displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: Infra Env
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service:latest
        version: 0.7.74
      entries:
      - name: assisted-service-operator.v0.7.74
        version: 0.7.74
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: assisted-service-operator.v0.7.71
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-x86_64-live.x86_64.iso",
                      "version": "414.92.202310170514-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-aarch64-live.aarch64.iso",
                      "version": "414.92.202310170514-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202309201615-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-s390x-live.s390x.iso",
                      "version": "414.92.202309201615-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-s390x-live.s390x.iso",
                      "version": "415.92.202310310037-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:cffbdc5fd51c08f16ec2aa484cf226e3b29510919a8aa9668d766007de8f3d79
          createdAt: "2023-12-15T15:26:08Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.71'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service@sha256:cffbdc5fd51c08f16ec2aa484cf226e3b29510919a8aa9668d766007de8f3d79
        - quay.io/edge-infrastructure/assisted-image-service@sha256:027fa9c5a881848786745aafeac5e48575eefe22417abe29d9e8eb17940b8f80
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/edge-infrastructure/assisted-installer@sha256:0f4ecf9a75b6755d02a45f9b70357ac45b6b4db66f24f0e83f888269d64f4293
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:40bafed94547a5fe96deb5492f10225501b965198227c9634869931692bda492
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:98740646fbfc0520f9d85ac5b483d1a898f22e164c5234901be51970bef7fcf3
        version: 0.7.71
      entries:
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.10
    - currentCSV: assisted-service-operator.v0.7.73
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-x86_64-live.x86_64.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-aarch64-live.aarch64.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-s390x-live.s390x.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-x86_64-live.x86_64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-aarch64-live.aarch64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-s390x-live.s390x.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.16.0-ec.4/rhcos-4.16.0-ec.4-x86_64-live.x86_64.iso",
                      "version": "416.94.202403071059-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-aarch64-live.aarch64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-s390x-live.s390x.iso",
                      "version": "415.92.202402130021-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:641fb65d8f0117cc6007567337ae38d5e912add454a76905617b60db99b60a64
          createdAt: "2024-03-20T10:08:49Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.73'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:2cc14304ba27c51d61b86db05445f3f0acc68593a4d76081da8eb9d6bae41757
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:d0da07ec81877302fc376e8049dca69c1acdbf56a37aa3e320cd75f2b85a2ed0
        - quay.io/edge-infrastructure/assisted-service@sha256:641fb65d8f0117cc6007567337ae38d5e912add454a76905617b60db99b60a64
        - quay.io/edge-infrastructure/assisted-image-service@sha256:b33068bbe88e38c0444fb158e233e5644fb127416198d52fdc809a3f2e684fde
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/edge-infrastructure/assisted-installer@sha256:234e4680b0ff4376f2c11d6f94a75a5dd6ae16c3218d6b63d2003cf43c65e185
        version: 0.7.73
      entries:
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.11
    - currentCSV: assisted-service-operator.v0.7.74
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-x86_64-live.x86_64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-aarch64-live.aarch64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-s390x-live.s390x.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-x86_64-live.x86_64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-aarch64-live.aarch64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-s390x-live.s390x.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-x86_64-live.x86_64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-aarch64-live.aarch64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-ppc64le-live.ppc64le.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-s390x-live.s390x.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-x86_64-live.x86_64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-aarch64-live.aarch64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-ppc64le-live.ppc64le.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-s390x-live.s390x.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-s390x-live.s390x.iso",
                      "version": "418.94.202409162337-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service:latest
          createdAt: "2024-10-16T17:43:34Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.74'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AgentClassification is the Schema for the AgentClassifications
              API
            displayName: Agent Classification
            kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - description: Agent is the Schema for the hosts API
            displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: Infra Env
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service:latest
        version: 0.7.74
      entries:
      - name: assisted-service-operator.v0.7.74
        version: 0.7.74
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.13
    - currentCSV: assisted-service-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "osImages": [
                    {
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202105062123-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/ocpmetal/assisted-service@sha256:64e88c5aa41c606bc2db2b0991acdbb1166559306474e13f5e2810f325616d64
          createdAt: "2021-09-23T11:11:20Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.0 <0.0.8'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled, for
              all other names a "Warning" event will be emitted + message logged and
              it will be ignored forever.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent".
          No other name will be accepted.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/ocpmetal/assisted-service@sha256:64e88c5aa41c606bc2db2b0991acdbb1166559306474e13f5e2810f325616d64
        - quay.io/ocpmetal/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/ocpmetal/assisted-installer@sha256:db97f7a565a8d5af61f3876d6465996e90c7f79dad6ec71c5bb5cce837444974
        - quay.io/ocpmetal/assisted-installer-agent@sha256:a391892766d1ce81300bf82943318e7c3be0d2945fc19065527ab9f64fc24cf8
        - quay.io/ocpmetal/assisted-installer-controller@sha256:43b65c310eee985ee0fc20ad8ee684f7557ffe704fd1a2fa53831e3dbc41e849
        version: 0.0.8
      entries:
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.3
    - currentCSV: assisted-service-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.2/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.2/rhcos-4.8.2-x86_64-live.x86_64.iso",
                      "version": "48.84.202107202156-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-x86_64-live.x86_64.iso",
                      "version": "49.84.202110081407-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:c3a852842b88f47c51cf1c62d3a19b102729fb355df19ced0e03e8265739c39c
          createdAt: "2022-04-07T11:23:06Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.1.3'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled, for
              all other names a "Warning" event will be emitted + message logged and
              it will be ignored forever.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent".
          No other name will be accepted.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-image-service@sha256:add19e251cafdfc4f34264d0aec5b660f0ea2dcc13326f4a2c8d50a2e465c0a2
        - quay.io/edge-infrastructure/assisted-service@sha256:c3a852842b88f47c51cf1c62d3a19b102729fb355df19ced0e03e8265739c39c
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/edge-infrastructure/assisted-installer@sha256:461d60b21d1d3ccb15a170f937f4871088cd0f83108b2271e84bdf2883a66870
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:979650f2e03e457d71958e8813285fcb7185fe0035f21a6ba956b4277cb29b70
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:658787d1e6c1f733c5f03863655aff315258c0b38e1972573c3878df46e4e463
        version: 0.1.3
      entries:
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.4
    - currentCSV: assisted-service-operator.v0.3.30
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-x86_64-live.x86_64.iso",
                      "version": "49.84.202110081407-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-aarch64-live-rootfs.aarch64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-aarch64-live.aarch64.iso",
                      "version": "49.84.202110080947-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.3/rhcos-4.10.3-x86_64-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.3/rhcos-4.10.3-x86_64-live.x86_64.iso",
                      "version": "410.84.202201251210-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:7c13ce52d82f761aec47bc5a6639d7baabe99d0f4f2e2106c4341018c606b458
          createdAt: "2022-06-13T23:16:28Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.3.30'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service@sha256:7c13ce52d82f761aec47bc5a6639d7baabe99d0f4f2e2106c4341018c606b458
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/edge-infrastructure/assisted-installer@sha256:57cf8d5fe49067a8c1345bedaf90f2b63aededdd7d615878496182c35a13892e
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:6271d639e4c0647bd28fbc85cea54f31143a876c3237376b9c670ec254a9c2fc
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:2267d2d7be38f81d530fed977b0c46f6c976369cd89c123c8e036c3b1dd988d5
        - quay.io/edge-infrastructure/assisted-image-service@sha256:d6396f9360b877a4f57139ba5b4969f243278db709be4de3573f2fba75e50018
        version: 0.3.30
      entries:
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.5
    - currentCSV: assisted-service-operator.v0.4.41
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.40/rhcos-4.9.40-x86_64-live.x86_64.iso",
                      "version": "49.84.202206171736-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.16/rhcos-4.10.16-x86_64-live.x86_64.iso",
                      "version": "410.84.202205191234-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.16/rhcos-4.10.16-aarch64-live.aarch64.iso",
                      "version": "410.84.202205191023-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.11.0-rc.2/rhcos-4.11.0-rc.2-x86_64-live.x86_64.iso",
                      "version": "411.86.202207090519-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.11.0-rc.2/rhcos-4.11.0-rc.2-aarch64-live.aarch64.iso",
                      "version": "411.86.202207090159-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:0fa869095185759c26c391f0e353f293c903ea739576f160d6668f5f20404bdb
          createdAt: "2022-09-11T14:56:07Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.4.41'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:56484f99799d22f345215bb65c6b81b6ee14e07d26b625c3f283a2ddfdc37238
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:951b9d625bc300b5b72a3916cb382015527ea786f10b7acbeebf83ae854ca51a
        - quay.io/edge-infrastructure/assisted-service@sha256:0fa869095185759c26c391f0e353f293c903ea739576f160d6668f5f20404bdb
        - quay.io/edge-infrastructure/assisted-image-service@sha256:104d2aae313c724fb10d3ef3bba78b2370fc0597de31335453894b679a450ab8
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:4ca00219efbc0d9dcce415dadc1039e9495f3a3b00ab3bd575b41554d7934899
        - quay.io/edge-infrastructure/assisted-installer@sha256:803168aeb7e01db77e3b2ff7418ceea3d93daccaceeb95f8fc02027144ebac75
        version: 0.4.41
      entries:
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.6
    - currentCSV: assisted-service-operator.v0.4.50
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-x86_64-live.x86_64.iso",
                      "version": "411.86.202210041459-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-aarch64-live.aarch64.iso",
                      "version": "411.86.202210032347-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.12.0-ec.2/rhcos-4.12.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "412.86.202208101039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.12.0-ec.2/rhcos-4.12.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "412.86.202208101040-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:0ca1cfe04f088cd4c7318cb91978754d6472ada29e3436d05e22335740bfff49
          createdAt: "2022-11-23T19:17:47Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.4.50'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service@sha256:0ca1cfe04f088cd4c7318cb91978754d6472ada29e3436d05e22335740bfff49
        - quay.io/edge-infrastructure/assisted-image-service@sha256:71e8c9c76caa8fe01255fdb4a1172cda372969f7fe1399a3a54fb93e967e68cf
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:54b39c6cafefb70d258ec1d491b6a0f1dd888cd85bd442c8f40a4805cf4721ca
        - quay.io/edge-infrastructure/assisted-installer@sha256:f225c86242b0bfbf58753123ef33b7c07f367a5099641d19cfda5a7121382586
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:34e0341c0471b1772db11535107db06e511c8d5d9816edc3971e745928164ecd
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:c1c7281b9f06e56330cc55ac8d74185d42ee82dff105ae747db677b8eb059aea
        version: 0.4.50
      entries:
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.7
    - currentCSV: assisted-service-operator.v0.6.65
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-x86_64-live.x86_64.iso",
                      "version": "411.86.202210041459-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-aarch64-live.aarch64.iso",
                      "version": "411.86.202210032347-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-s390x-live.s390x.iso",
                      "version": "411.86.202210171842-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202210171844-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-x86_64-live.x86_64.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-aarch64-live.aarch64.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-s390x-live.s390x.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-x86_64-live.x86_64.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-aarch64-live.aarch64.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-s390x-live.s390x.iso",
                      "version": "413.92.202303281804-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:2dfaf2c3ac58b2987b0088df302c7cf717d0b6640cdb8b2327370aa110120784
          createdAt: "2023-05-24T11:15:57Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.6.65'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-image-service@sha256:b1ba1076fe16b2f868ed31a660b16583a957a83634c7ec3cf03574b56e02144c
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:1287d947464fbf5275feb31d106d2feb856c6117f0eaec8d947f2e03bd8618c5
        - quay.io/edge-infrastructure/assisted-installer@sha256:f8a7e7cb6ff1a4a7daa0cb26f0fab8da17f109653af5979c9d53c35b7ca2c1bb
        - quay.io/edge-infrastructure/assisted-service@sha256:2dfaf2c3ac58b2987b0088df302c7cf717d0b6640cdb8b2327370aa110120784
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:1ad91b13c763437863480df68f5b65576e95c558d4537886432a81c0b8d3cedc
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:a6f99547564cc2e047df74003019817b2dbd3dad992a0c8e53aaaba4f5b02ab6
        version: 0.6.65
      entries:
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.8
    - currentCSV: assisted-service-operator.v0.7.67
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-s390x-live.s390x.iso",
                      "version": "414.92.202305090606-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:fb0397a1402d42a64e3128e66a0817aa9466e67eb407cb792dda0d380bd5a189
          createdAt: "2023-10-08T07:14:20Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.67'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service@sha256:fb0397a1402d42a64e3128e66a0817aa9466e67eb407cb792dda0d380bd5a189
        - quay.io/edge-infrastructure/assisted-image-service@sha256:69ff087be8e5d8458a97f49d88b8b2025ee341d362dcf5618988c011b66688e0
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/edge-infrastructure/assisted-installer@sha256:d85be59bc379927f46d0a88ef04172c5c6733e31e1e4a8188de2d434d5bedb66
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:f2aeb4f7d193ea2c5702cae0f6039c02991b9f417d091468d46974408c97efe4
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:35c118d8c07dd72fe3307424cc051d9e0426e916412171449a05e41445b9f599
        version: 0.7.67
      entries:
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.9
    defaultChannel: alpha
    packageName: assisted-service-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Strimzi
      provider-url: ""
    name: strimzi-kafka-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: strimzi-cluster-operator.v0.48.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.1.0",
                      "metadataVersion": "4.1-IV1",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.1.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
          createdAt: "2025-09-25 10:32:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.48.0 supports only Kubernetes 1.27 and newer.**
            Kubernetes 1.25 and 1.26 are not supported anymore

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.48.0
          * Support for Kafka 4.1.0
          * Add monitoring of custom resources using kubernetes-state-metrics (KSM)
          * Support for type: custom client authentication
          * Added support for Strimzi Metrics Reporter to Kafka Connect, Mirror Maker 2 and Kafka Bridge.
          * Added support for KIP-745 in Kafka Connector and Mirror Maker 2, allowing the usage of includeTasks and onlyFailed arguments in Kafka connectors restart.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
        - quay.io/strimzi/kafka@sha256:34450afc1a3399a9026fc3c6ec90e9f6adfec9faf6520545b4d84e90cda34964
        - quay.io/strimzi/kafka-bridge@sha256:53034f64f0b672f10b5bacea1c7a25132f118df7fd5c9032c4dbf702ed93796a
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:64380efcb8da87fdc0e0de238a20456be9abb830174c2f2331882b7a84657e57
        - quay.io/strimzi/kafka@sha256:48db05ad2e9fa39eb388089e7e25654bcbf4da1ddc3a5635d675056c28bbcb96
        version: 0.48.0
      entries:
      - name: strimzi-cluster-operator.v0.48.0
        version: 0.48.0
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: stable
    - currentCSV: strimzi-cluster-operator.v0.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnectS2I",
                "metadata":{
                  "name":"my-connect-s2i-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "whitelist":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 2,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": 1,
                        "offset.storage.replication.factor": 1,
                        "status.storage.replication.factor": 1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "NetworkOutboundCapacityGoal",
                    "CpuCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:59d939f342f96d0e8a54c96fd187832c4b19cd4137bf55ace53d486728ae6c5f
          createdAt: "2021-05-14 11:37:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster with Source 2 Image support
            displayName: Kafka Connect Source to Image
            kind: KafkaConnectS2I
            name: kafkaconnects2is.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  Starting with Strimzi 0.23, we support only the `v1beta2` API version of all Strimzi custom resources.  This is a required for migration to `apiextensions/v1` which is needed because Kubernetes 1.22 will remove support for `apiextensions/v1beta1`.  Migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs **before upgrading to Strimzi 0.23**. For more details about the CRD upgrades, see the [documentation](https://strimzi.io/docs/operators/0.23.0/deploying.html#assembly-upgrade-resources-str).
          ### New in 0.23
          * Migrate to CRD v1 (required by Kubernetes 1.22+)
          * Strimzi API versions `v1alpha1` and `v1beta1` were removed from all Strimzi custom resources apart from `KafkaTopic` and `KafkaUser` (use `v1beta2` versions instead)
          * Add support for Kafka 2.8.0 and 2.6.2, remove support for Kafka 2.5.x
          * Make it possible to configure maximum number of connections and maximum connection creation rate in listener configuration
          * Add support for configuring finalizers for `loadbalancer` type listeners
          * Use dedicated Service Account for Kafka Connect Build on Kubernetes
          * Remove direct ZooKeeper access for handling user quotas in the User Operator. Add usage of Admin Client API instead.
          * Support for configuring custom Authorizer implementation
          * Changed Reconciliation interval for Topic Operator from 90 to 120 seconds (to keep it the same as for other operators)
          * Changed Zookeeper session timeout default value to 18 seconds for Topic and User Operators (for improved resiliency)
          * Removed requirement for replicas and partitions KafkaTopic spec making these parameters optional
          * Support to configure a custom filter for parent CR's labels propagation into subresources
          * Allow disabling service links (environment variables describing Kubernetes services) in Pod template
          * Add support for separate control plane listener (disabled by default, available via the `ControlPlaneListener` feature gate)
          * Support for Dual Stack networking
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _master_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/master/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/master/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:59d939f342f96d0e8a54c96fd187832c4b19cd4137bf55ace53d486728ae6c5f
        - quay.io/strimzi/kafka@sha256:699bff48ee445d8ba381a791cd0340015e3ecede8d169f41bc151095ab7fb5cc
        - quay.io/strimzi/kafka@sha256:d2745f48ed262290ce66c3362ffb9b9a8cfeb9c76beac40c995f72f3f06ad10e
        - quay.io/strimzi/kafka@sha256:dbc53bb374d2cb9513a5bb8f80fea751c91b249cabd36d8b3c04af5e5c961738
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/kafka@sha256:e90695b35d6dd5db20c9900dd0dab95d1ef908f3145b4f9f436b903303c88549
        - quay.io/strimzi/kafka-bridge@sha256:bed07186355a4c04ee50e90048e3bd3a023264093a9ef28f0278cce44640e93f
        - quay.io/strimzi/kafka@sha256:641f46114c59d61577bcb3f4ef5af16aea3ed42c2a575f8d3641c82848df85e8
        - quay.io/strimzi/jmxtrans@sha256:b4ec1844f4cc5eb74cd0aac83d74b81b977e178bf670d1136d1af0b68663d0a3
        version: 0.23.0
      entries:
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.23.x
    - currentCSV: strimzi-cluster-operator.v0.24.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnectS2I",
                "metadata":{
                  "name":"my-connect-s2i-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 2,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:786ea69c77e0651f009a73a8b3ce0b3d41b3cd71b917dc0d9bdc61c725f494a6
          createdAt: "2021-06-24 13:13:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster with Source 2 Image support
            displayName: Kafka Connect Source to Image
            kind: KafkaConnectS2I
            name: kafkaconnects2is.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: "Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org)
          cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/)
          in various deployment configurations. See our [website](https://strimzi.io)
          for more details about the project.\n### CRD Upgrades\n**!!! IMPORTANT !!!**
          \ This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.
          \ If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed
          for all Strimzi CRDs and CRs before the upgrade to 0.24 is done!  If upgrading
          from Strimzi version earlier than 0.22, you need to first install the CRDs
          from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi
          CRDs and CRs before the upgrade to 0.24 is done!  For more details about
          the CRD upgrades, see the documentation.\n### New in 0.24\n* Add support
          for Kafka 2.7.1 and remove support for 2.6.0, 2.6.1, and 2.6.2\n* Add support
          for [Kubernetes Configuration Provider for Apache Kafka](https://github.com/strimzi/kafka-kubernetes-config-provider)\n*
          Use Red Hat UBI8 base image\n* Support for patching of service accounts
          and configuring their labels and annotations. The feature is disabled by
          default and enabled using the new `ServiceAccountPatching` feature gate.\n*
          Added support for configuring cluster-operator's worker thread pool size
          that is used for various sync and async tasks\n* Add Kafka Quotas plugin
          with produce, consume, and storage quotas\n* Support pausing reconciliation
          of KafkaTopic CR with annotation `strimzi.io/pause-reconciliation`\n* Update
          cruise control to 2.5.57\n* Update to Strimzi Kafka Bridge to 0.20.1\n*
          Support for broker load information added to the rebalance optimization
          proposal. Information on the load difference, before and after a rebalance
          is stored in a ConfigMap\n* Add support for selectively changing the verbosity
          of logging for individual CRs, using markers.\n* Added support for `controller_mutation_rate'
          quota. Creation/Deletion of topics and creation of partitions can be configured
          through this.\n* Use newer version of Kafka Exporter with different bugfixes
          \n### Supported Features\n* **Manages the Kafka Cluster** - Deploys and
          manages all of the components of this complex application, including dependencies
          like Apache ZooKeeper® that are traditionally hard to administer.\n* **Includes
          Kafka Connect** - Allows for configuration of common data sources and sinks
          to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for morroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashabords\n### Upgrading your Clusters\nThe
          Strimzi Operator understands how to run and upgrade between a set of Kafka
          versions. When specifying a new version in your config, check to make sure
          you aren't using any features that may have been removed. See [the upgrade
          guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str)
          for more information.\n### Storage\nAn efficient data storage infrastructure
          is essential to the optimal performance of Apache Kafka®. Apache Kafka®
          deployed via Strimzi requires block storage. The use of file storage (for
          example, NFS) is not recommended.\nThe Strimzi Operator supports three types
          of data storage:\n* Ephemeral (Recommended for development only!)\n* Persistent\n*
          JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nStrimzi
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation to the current _main_
          branch as well as all releases can be found on our [website](https://strimzi.io/documentation).\n###
          Getting help\nIf you encounter any issues while using Strimzi, you can get
          help using:\n* [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          Contributing\nYou can contribute by:\n* Raising any issues you find using
          Strimzi\n* Fixing issues by opening Pull Requests\n* Improving documentation\n*
          Talking about Strimzi\n\nAll bugs, tasks or enhancements are tracked as
          [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues).
          Issues which  might be a good start for new contributors are marked with
          [\"good-start\"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start)
          label.\n\nThe [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md)
          describes how to build Strimzi and how to  test your changes before submitting
          a patch or opening a PR.\n\nThe [Documentation Contributor Guide](https://strimzi.io/contributing/guide/)
          describes how to contribute to Strimzi documentation.\n\nIf you want to
          get in touch with us first before contributing, you can use:\n* [Strimzi
          mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          License\nStrimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).\n"
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/kafka@sha256:95cfe9000afda2f7def269ca46472d3803ee85146c521884884d8505a7187daf
        - quay.io/strimzi/kafka@sha256:8959b7968ab8b3306906cdbff2ebb8d63329af37e58124a601843795c4ef5bd6
        - quay.io/strimzi/kafka@sha256:fbb08410d9595029bc4a02ed859971264e6ce2dc85dd6a9855eaa7bb58b52a25
        - quay.io/strimzi/kafka-bridge@sha256:537ea10e87619251204747ebc1005c632a6c6ed270f5ef06f86c0c3ae4a7553b
        - quay.io/strimzi/operator@sha256:786ea69c77e0651f009a73a8b3ce0b3d41b3cd71b917dc0d9bdc61c725f494a6
        - quay.io/strimzi/jmxtrans@sha256:2c54adabf6f31c087dbb898057f9c5e200aaf4c5b39fb3a47bbaf7f6587df72d
        version: 0.24.0
      entries:
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.24.x
    - currentCSV: strimzi-cluster-operator.v0.25.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:afaf1dd1f2f696ae80674d74159661ea239ae609f78a73a1423c7a17a39e5044
          createdAt: "2021-06-24 13:13:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.25 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.25 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.25
          * Move from Scala 2.12 to Scala 2.13
          * Open Policy Agent authorizer updated to a new version supporting Scala 2.13.
            The new version of the authroizer uses different format of the input data sent to the OPA server and might require you to update your policies.
            See [release notes](https://github.com/strimzi/strimzi-kafka-operator/releases/tag/0.25.0) for more details.

          * Allow a custom password to be set for SCRAM-SHA-512 users by referencing a secret in the `KafkaUser` resource
          * Add support for [EnvVar Configuration Provider for Apache Kafka](https://github.com/strimzi/kafka-env-var-config-provider)
          * Add support for `tls-external` authentication to User Operator to allow management of ACLs and Quotas for TLS users with user certificates generated externally
          * Support for disabling the automatic generation of network policies by the Cluster Operator
          * Configure fixed size limit for `emptyDir` volumes used for temporary files
          * Update Strimzi Kafka Bridge to 0.20.2
          * The `KafkaConnectS2I` resource has been removed and is no longer supported by the operator.
            Please use the [migration guide](https://strimzi.io/docs/operators/0.24.0/full/using.html#proc-migrating-kafka-connect-s2i-str) to migrate your `KafkaConnectS2I` deployments to [`KafkaConnect` Build](https://strimzi.io/docs/operators/latest/full/deploying.html#creating-new-image-using-kafka-connect-build-str) instead.

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:d8991608ec71f17e850c33714ee02ca87ba94bc0f077be62ad1ed127ca244823
        - quay.io/strimzi/kafka@sha256:ee17eb3611019b0cfa5e9e196610edc1dbd7084471fb20990f055c89da8eef45
        - quay.io/strimzi/kafka@sha256:56a33815f7d50a97ea9bb4cce1e3a6d04d918b841b5aa285f16b792d8930059e
        - quay.io/strimzi/kafka-bridge@sha256:a005e20d7fbe04d6f587f2093e21a00556ec62b850df757b9f7d9d57e3141c40
        - quay.io/strimzi/jmxtrans@sha256:af5eb8459ac9a0acbca51e67c6e0217f62ba404cba9f2251968c927dcef75252
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/operator@sha256:afaf1dd1f2f696ae80674d74159661ea239ae609f78a73a1423c7a17a39e5044
        version: 0.25.0
      entries:
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.25.x
    - currentCSV: strimzi-cluster-operator.v0.26.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.0.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"3.0",
                        "inter.broker.protocol.version":"3.0"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.0.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:5d90caa66f57939ba62a9a50efa00fce319ba596e719a3f8a42170d7635ee3f4
          createdAt: "2021-12-14 00:16:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.26 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.26 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.26.1
          * **Updated Log4j2 to fix CVE-2021-44228**
          ### New in 0.26.0
          * Add support for Kafka 2.8.1 and 3.0.0; remove Kafka 2.7.0 and 2.7.1
          * Add support for exposing JMX in Zookeeper
          * Allow configuring labels and annotations for JMX authentication secrets
          * Enable Cruise Control anomaly.detection configurations
          * Add support for building connector images from the Maven coordinates
          * Allow Kafka Connect Build artifacts to be downloaded from insecure servers
          * Add option to specify pull secret in Kafka Connect Build on OpenShift
          * Configurable authentication, authorization, and SSL for Cruise Control API
          * Allow to configure `/tmp` volume size via Pod template. By default `1Mi` is used.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:d3899c79747e4cb3a6df5a9fcd80b81f038ca0634a97329c1d160499469876dd
        - quay.io/strimzi/jmxtrans@sha256:bff5d3bab575d0b9ed95dd79d2d22518cc00c4d1c9dbb5614ed16149d4920a1b
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/maven-builder@sha256:663eb81388ae8f824e7920c272f6d2e2274cf6c140d61416607261cdce9d50e2
        - quay.io/strimzi/kafka@sha256:8c022a8535742b2eea691737f863af5ddef1e8df08b415b8139ecd02a79052a5
        - quay.io/strimzi/kafka@sha256:bd22a134b0c351ddfc78a3263850c2f16733c52f57ac92984cbe0ffff477c724
        - quay.io/strimzi/operator@sha256:5d90caa66f57939ba62a9a50efa00fce319ba596e719a3f8a42170d7635ee3f4
        - quay.io/strimzi/kafka@sha256:6367b0fbba9593e3ac736040ea2d7ef71296b1ec144d3b54d9370352141b90e4
        version: 0.26.1
      entries:
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.26.x
    - currentCSV: strimzi-cluster-operator.v0.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.0.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"3.0",
                        "inter.broker.protocol.version":"3.0"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.0.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:9ea49407a04006d0bddfaaee822adfb95f5857af32ffcee81ba60ce891b07ba8
          createdAt: "2022-01-14 20:00:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.27 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.27 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.27.1
          * Updated Log4j2 to version 2.17.1 and Fabric8 Kubernetes Client to 5.10.2
          ### New in 0.27.0
          * Updated Log4j2 to version 2.17.0 (adresses CVE-2021-44228, CVE-2021-45046, CVE-2021-45105)
          * (Experimental) support for AArch64 / ARM64 architecture
          * `ControlPlaneListener` and `ServiceAccountPatching` feature gates are now in the beta phase and are enabled by default
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a07ebe6600d6baba09980abdde2931b3b6abd2f75aeb2cac5c6a2d22a37aa807
        - quay.io/strimzi/maven-builder@sha256:9acd178cc504be8acaa70a39f3c17830aae1f0e93a656110790f9476b09260d5
        - quay.io/strimzi/kafka@sha256:f4d68bb94447c6612f70de7f6587e9e7ef712c83769ea1f11a8dbda0d241a059
        - quay.io/strimzi/kafka@sha256:a3ec567ca529df2fa41f8986d21e86dde66be94aa4baa56faa66db4dcb3b258e
        - quay.io/strimzi/kafka@sha256:5682a6bfcc4ee13d53a30d587c8f96bc40abe3618880989336c5af274ea3cefa
        - quay.io/strimzi/kafka-bridge@sha256:55a52316aead02efa2004174e11ad4728bb8b6f1c1415f725f597f1d2f2da188
        - quay.io/strimzi/operator@sha256:9ea49407a04006d0bddfaaee822adfb95f5857af32ffcee81ba60ce891b07ba8
        - quay.io/strimzi/jmxtrans@sha256:d2d2477517eb1ac98b51ad6d64d8c1c578141637afa8d46037076f0013183c0b
        version: 0.27.1
      entries:
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.27.x
    - currentCSV: strimzi-cluster-operator.v0.28.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.1.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.1"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.1.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.1.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:8b010c5fb8244b742a37f264feefc116d75bb2abd2d6895375cbeea250727b39
          createdAt: "2022-02-12 23:02:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.28 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.28 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.28.0
          * Support for Kafka 3.1.0
          * Support for `StrimziPodSet` resources (disabled by default through the `UseStrimziPodSets` feature gate)
          * Support custom authentication mechanisms in Kafka listeners
          * Intra-broker disk balancing using Cruise Control
          * Add support for disabling the FIPS mode in OpenJDK
          * Fix renewing your own CA certificates
          * Added the option `createBootstrapService` to disable the creation of the bootstrap service for the Load Balancer Type Listener
          * Dependency updates (Strimzi Kafka Bridge, Strimzi OAuth, Open Policy Agent Authroizer plugin and others)
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a07ebe6600d6baba09980abdde2931b3b6abd2f75aeb2cac5c6a2d22a37aa807
        - quay.io/strimzi/maven-builder@sha256:3d5d664e089651944712361a6d8ba42b578dbd9bb3dfd9972f614520c9cfab75
        - quay.io/strimzi/kafka@sha256:7c04f65e216bb941458aeadffb85e9eeee43d5ca67da566017c2c93a0664c98e
        - quay.io/strimzi/kafka@sha256:06e9bda15495feb2b92d9950e834cdeed6ef20064f021a879610f1b59dd46950
        - quay.io/strimzi/kafka-bridge@sha256:c345c783d8fbbee4edeca2ebc47e7c454fcc2e26686820b8037de97369280ba0
        - quay.io/strimzi/jmxtrans@sha256:3a55a81790dc88b90d1bf858e2cfd7e591ad0e79f763c6d1667b2d3eb0d0c79d
        - quay.io/strimzi/operator@sha256:8b010c5fb8244b742a37f264feefc116d75bb2abd2d6895375cbeea250727b39
        version: 0.28.0
      entries:
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.28.x
    - currentCSV: strimzi-cluster-operator.v0.29.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:afde94c6d2544207cb7f3c94b54455643558a9d5daaa13b9f736cc8da20e319c
          createdAt: "2022-05-15 00:01:07"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.29 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.29 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.29.0
          * Add support for Apache Kafka 3.0.1, 3.1.1 and 3.2.0
          * Renew user certificates in User Operator only during maintenance windows
          * Remove TLS sidecar from Cruise Control pod. Cruise Control is now configured to not using ZooKeeper, so the TLS sidecar is not needed anymore.
          * Allow Cruise Control topic names to be configured
          * Add support for `spec.rack.topologyKey` property in Mirror Maker 2 to enable "fetch from the closest replica" feature.
          * Experiemntal support for the s390x platform
          * Added new rebalancing modes on the `KafkaRebalance` custom resource to add or remove brokers
          * Experimental KRaft mode (ZooKeeper-less Kafka). Use it for development and testing only!
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:544c4987f7f3ca558b04e9ef4357be204f088212b7b5b9ed0f909401a8bc5238
        - quay.io/strimzi/maven-builder@sha256:0c096a59faf8fd224298c01415a895043db4647534c060cb7e53207e9448bf65
        - quay.io/strimzi/operator@sha256:afde94c6d2544207cb7f3c94b54455643558a9d5daaa13b9f736cc8da20e319c
        - quay.io/strimzi/kafka@sha256:95678e843bc87cbe571f3f299045dd075725816b59e1b87821cd6025876ec559
        - quay.io/strimzi/kafka@sha256:935caab03f1d0365b47ddefec0c1d8fa449fc3de715ccdd298aedef5ce968c73
        - quay.io/strimzi/kafka@sha256:ac190141120c1800cc7131d3a1f701e64efdcd4a5d87f0b2e0e2cadfdef8b438
        - quay.io/strimzi/kafka@sha256:9dfc09273ad8fc4bdc75cbe93b4a061a05a5defb5884a221ae1e178ba602787f
        - quay.io/strimzi/kafka@sha256:284cde40784b6ff7bfb81b6dd28928e1e8c598fe6a72504f025592e5b43fa94c
        - quay.io/strimzi/kafka-bridge@sha256:b7bc3dd90e87223766f608a0087b521f0bba79d1c21eaa833bc2708c71148b7f
        - quay.io/strimzi/jmxtrans@sha256:f89e3c9618430908a8b390f2490f5aa43687b12889b434472a962ce0170c6463
        version: 0.29.0
      entries:
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.29.x
    - currentCSV: strimzi-cluster-operator.v0.30.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:587db6f4389f56209f48853beedbfe249f6d590a4cb0000be2c42b57e43c2adb
          createdAt: "2022-07-16 00:00:10"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: "Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org)
          cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/)
          in various deployment configurations. See our [website](https://strimzi.io)
          for more details about the project.\n### CRD Upgrades\n**!!! IMPORTANT !!!**
          \ This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.
          \ If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed
          for all Strimzi CRDs and CRs before the upgrade to 0.30 is done!  If upgrading
          from Strimzi version earlier than 0.22, you need to first install the CRDs
          from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi
          CRDs and CRs before the upgrade to 0.30 is done!  For more details about
          the CRD upgrades, see the documentation.\n### New in 0.30.0\n* The `UseStrimziPodSets`
          feature gate moves to beta stage. By default, StrimziPodSets are used instead
          of StatefulSets. If needed, `UseStrimziPodSets` can be disabled in the feature
          gates configuration in the Cluster Operator.\n* The `ServiceAccountPatching`
          feature gate moves to GA. It cannot be disabled anymore and will be permanently
          enabled.\n* Remove support for Kafka 3.0.0 and 3.0.1\n* Add support for
          `simple` authorization and for the User Operator to the experimental `UseKRaft`
          feature gate. _(Note: Due to [KAFKA-13909](https://issues.apache.org/jira/browse/KAFKA-13909),
          broker restarts currently don't work when authorization is enabled.)_\n*
          Add CPU and network capacity overrides for Cruise Control capacity config\n*
          Use better encryption and digest algorithms when creating the PKCS12 stores.
          For existing clusters, the certificates will not be updated during upgrade
          but only next time the PKCS12 store is created. \n* Operator emits Kafka
          events to explain why it restarted a Kafka broker\n* Better configurability
          of the Kafka Admin client in the User Operator\n* Update Strimzi Kafka Bridge
          to 0.21.6\n### Supported Features\n* **Manages the Kafka Cluster** - Deploys
          and manages all of the components of this complex application, including
          dependencies like Apache ZooKeeper® that are traditionally hard to administer.\n*
          **Includes Kafka Connect** - Allows for configuration of common data sources
          and sinks to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for mirroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashboards\n### Upgrading your Clusters\nThe
          Strimzi Operator understands how to run and upgrade between a set of Kafka
          versions. When specifying a new version in your config, check to make sure
          you aren't using any features that may have been removed. See [the upgrade
          guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str)
          for more information.\n### Storage\nAn efficient data storage infrastructure
          is essential to the optimal performance of Apache Kafka®. Apache Kafka®
          deployed via Strimzi requires block storage. The use of file storage (for
          example, NFS) is not recommended.\nThe Strimzi Operator supports three types
          of data storage:\n* Ephemeral (Recommended for development only!)\n* Persistent\n*
          JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nStrimzi
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation to the current _main_
          branch as well as all releases can be found on our [website](https://strimzi.io/documentation).\n###
          Getting help\nIf you encounter any issues while using Strimzi, you can get
          help using:\n* [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          Contributing\nYou can contribute by:\n* Raising any issues you find using
          Strimzi\n* Fixing issues by opening Pull Requests\n* Improving documentation\n*
          Talking about Strimzi\nAll bugs, tasks or enhancements are tracked as [GitHub
          issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues
          which  might be a good start for new contributors are marked with [\"good-start\"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start)
          label.\nThe [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md)
          describes how to build Strimzi and how to  test your changes before submitting
          a patch or opening a PR.\nThe [Documentation Contributor Guide](https://strimzi.io/contributing/guide/)
          describes how to contribute to Strimzi documentation.\nIf you want to get
          in touch with us first before contributing, you can use:\n* [Strimzi mailing
          list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n* [Strimzi
          Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          License\nStrimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).\n"
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:544c4987f7f3ca558b04e9ef4357be204f088212b7b5b9ed0f909401a8bc5238
        - quay.io/strimzi/maven-builder@sha256:f2cb45234f3a7a010824a370d03032b706271cd0d1d260da929f046f9259d7e0
        - quay.io/strimzi/kafka@sha256:68d6f4cb5cc5ccad44de799bd9cf417c469a09e1024dbe42298942f25a09e004
        - quay.io/strimzi/kafka@sha256:d9258ddd453c3a53ca8e737b989d27725818da95f7a4fc879c9f07fe743c0cef
        - quay.io/strimzi/kafka@sha256:5d881605425ab0b37d0b9cac6957d00ba904df0db185ca6181c810151bdab8e7
        - quay.io/strimzi/operator@sha256:587db6f4389f56209f48853beedbfe249f6d590a4cb0000be2c42b57e43c2adb
        - quay.io/strimzi/kafka-bridge@sha256:77fc68da171c74d3a4b75b1030eaaa5eae12e200002f7aa2ee1969385eba1d66
        - quay.io/strimzi/jmxtrans@sha256:de1ca88b713b926b72df535aca2cfad24fb1b2bf40481065d6fa0c421b872320
        version: 0.30.0
      entries:
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.30.x
    - currentCSV: strimzi-cluster-operator.v0.31.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.3",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.3",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:6e5e207c70c25c595d3ce60268c0214c7518e95836c9b1212ef1eb2f008c7a76
          createdAt: "2022-09-20 21:07:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          **This is the last version of Strimzi with support for Kubernetes 1.16, 1.17 and 1.18!**
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.31 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.31 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.31.1
          * Kafka 3.1.2 and 3.2.3 (fixes CVE-2022-34917)
          * Make `sasl.server.max.receive.size` broker option user configurable
          * Update Strimzi Kafka Bridge to 0.22.1
          ### New in 0.31.0
          * Support for Apache Kafka 3.2.1
          * Pluggable Pod Security Profiles with built-in support for _restricted_ Kubernetes Security Profile
          * Support for leader election and running multiple operator replicas (1 active leader replicas and one or more stand-by replicas)
          * Support for IPv6 addresses being used in Strimzi issued certificates
          * `StrimziPodSet` reconciliation metrics
          * Updated Strimzi Kafka Bridge to 0.22.0 and Kafka Exporter to 1.6.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:03cc48bc5cf367f898338665b253855bed7d6096fe577adfe344da2f2cddcfe4
        - quay.io/strimzi/operator@sha256:6e5e207c70c25c595d3ce60268c0214c7518e95836c9b1212ef1eb2f008c7a76
        - quay.io/strimzi/kaniko-executor@sha256:06581fb615f0b13ddfe27fae26ba5836b12c6e28afd55a81cba7685c9ca4bf79
        - quay.io/strimzi/maven-builder@sha256:670ca70900f2229ef140a592b77f2107934601f8b050b79250eab17ff62099b6
        - quay.io/strimzi/kafka-bridge@sha256:3ddac62a4f8c8a3ebb72d5841d3826f7fd3fbb0e00eb75fb40bced0a576b86a9
        - quay.io/strimzi/kafka@sha256:b7c059bd72d8b063a96dde788d91d1ad87498bad3e73acba600929d595aeb7d6
        - quay.io/strimzi/kafka@sha256:430b3eb70e47a91d7b5b6f230d5be024c2de853099cc0437e219b443e0ead9f5
        - quay.io/strimzi/kafka@sha256:9553aae22d35448681cbbf7a831a28149fb159f3df58ccf8c8ec4ca7eb9c1fdf
        - quay.io/strimzi/kafka@sha256:83acc1e2cc1e64a88bdbb54b9818a6ad3658c3da60e72c4be365f824c5fcf57c
        - quay.io/strimzi/jmxtrans@sha256:52dd107062d78c29bb58c905ac654c39df72bbe3c25613023aa33e834930606d
        - quay.io/strimzi/kafka@sha256:652924a1dba46aa716caba8381f1e791c09035e31ce62fdcc81f9b69af36cd02
        version: 0.31.1
      entries:
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.31.x
    - currentCSV: strimzi-cluster-operator.v0.32.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.3.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.3.1",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:c5e0e5dca7503a8371b8d366db0f49e3fc68f62ea480262e39b6cc277991c97c
          createdAt: "2022-10-31 17:05:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **From the 0.32.0 release, Strimzi supports only Kubernetes 1.19 and newer! Kubernetes 1.16, 1.17 and 1.18 are not supported anymore.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.32.0/deploying.html#assembly-upgrade-str).
          ### New in 0.32.0
          * Add support for Kafka 3.3.1 and remove support for Kafka 3.1.0, 3.1.1, and 3.1.2
          * Update KafkaConnector CR status so the 'NotReady' condition is added if the connector or any tasks are reporting a 'FAILED' state.
          * Add auto-approval mechanism on KafkaRebalance resource when an optimization proposal is ready
          * The ControlPlaneListener feature gate moves to GA
          * Add client rack-awareness support to Strimzi Bridge pods
          * Add support for OpenTelemetry for distributed tracing
          * ZookeeperRoller considers unready pods
          * Support multiple operations per ACLRule
          * Add new listener type "cluster-ip" which is using per-broker services instead of the pod DNS names
          * Update Strimzi OAuth library to 0.11.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:8c9f38b45dd814e617c9c51cc90fd51e23d7dd23e8efe7e9eff013522bd6a5a0
        - quay.io/strimzi/kafka@sha256:bb97008b7f928ed98939b2008a7baf8909c88e726a59aaf607f2eeba55375cc0
        - quay.io/strimzi/kafka@sha256:680ae1958dbcb9da8ee4128a67c1163a6ee2744221f7d29f73d8bcc237fd0173
        - quay.io/strimzi/maven-builder@sha256:593e9980b786ca8ad3223151c3e17f16c7652ee0e86d722af4c34b023038ecc2
        - quay.io/strimzi/operator@sha256:c5e0e5dca7503a8371b8d366db0f49e3fc68f62ea480262e39b6cc277991c97c
        - quay.io/strimzi/jmxtrans@sha256:208928133b2854f60bb34912e07561a3511de895b815960e73d565be2d553918
        - quay.io/strimzi/kafka@sha256:4bacf5b54041bd0ab9199c8c5d51e8058758139db4b7abc380cf0400737716ef
        - quay.io/strimzi/kafka-bridge@sha256:aad0cee4381cc1c8c6638ab5732b135acd271b5889486ca9a33312cce67e4a02
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        version: 0.32.0
      entries:
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.32.x
    - currentCSV: strimzi-cluster-operator.v0.33.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:cf0d41fdb99b20e91063e23eb1d664817c18866c9a2f5c4430eac897e78d5d53
          createdAt: "2023-02-18 21:21:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.33 supports only Kubernetes 1.19 and newer! Kubernetes versions 1.16, 1.17 and 1.18 are not supported anymore since Strimzi 0.32.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.33.2/deploying.html#assembly-upgrade-str).
          ### New in 0.33
          _Note: Strimzi 0.33.2 add support for Kafka 3.4.0 which fixes CVE-2023-25194._
          * Add support for Kafka 3.3.2
          * Automatic restarts of failed Connect or Mirror Maker 2 connectors
          * Redesign of Strimzi User Operator to improve its scalability
          * Move to Java 17
          * Improved FIPS support
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/maven-builder@sha256:de9e51365e7acc63feae745618804e6286fd730f41c3b31582cbb7319a93b483
        - quay.io/strimzi/kafka@sha256:116cf543fbadadd0b45007ce24d44a0c79a62280e8faa1fc80df40291fbdf05d
        - quay.io/strimzi/kafka@sha256:9eca149cfc436580e3cdd6cee0ca318366b8d0ec53c2cf007e4e2c614b988921
        - quay.io/strimzi/kafka@sha256:301dc52ed5891ed4fa33e79ca3c344b1ac0eec7469dd46a5a7ac22064e053fc2
        - quay.io/strimzi/kafka-bridge@sha256:3a380d5f3ea509f90cdf95280d83b42f5de4f86ee3a512e02234e7639f50c9eb
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/kafka@sha256:30945aecb9d63cf8863c672222f130469c89aefa07078c95aaed52c40fd16146
        - quay.io/strimzi/jmxtrans@sha256:5ebdafa7a5e3718ed3274d6254c9e655613c272340094b997b1c97fde2598b4a
        - quay.io/strimzi/operator@sha256:cf0d41fdb99b20e91063e23eb1d664817c18866c9a2f5c4430eac897e78d5d53
        - quay.io/strimzi/kafka@sha256:fe9d3f8b029ff32d5d725e57384592914957314d63e92fda3abf4028fe4e66f6
        - quay.io/strimzi/kafka@sha256:40a847f8455ec6ea9ff9fe1fff20b1ed01f848326dc13977d13ac911f0cb40b7
        version: 0.33.2
      entries:
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.33.x
    - currentCSV: strimzi-cluster-operator.v0.34.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:f4ee3ca482a2dd19785f59634578a0850273444c1173a5699a99cf02f3a018fe
          createdAt: "2023-03-17 13:02:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.34 supports only Kubernetes 1.19 and newer! Kubernetes versions 1.16, 1.17 and 1.18 are not supported anymore since Strimzi 0.32.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.34.0/deploying.html#assembly-upgrade-str).
          ### New in 0.34
          * Add support for Apache Kafka 3.4.0 (fixes CVE-2023-25194) and remove support for Apache Kafka 3.2.x
          * Stable Pod identities for Kafka Connect and MirrorMaker 2 (disabled by default)
          * Use JDK HTTP client in the Kubernetes client instead of the OkHttp client
          * Add truststore configuration for HTTPS connections to Open Policy Agent server
          * Update Strimzi HTTP Bridge to 0.25.0 and strimzi Oauth to 0.12.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:37cf8d0fea1d2078511920190a72b49977ff4047b6ecbfed6b394c6f9e1e2ff0
        - quay.io/strimzi/kafka-bridge@sha256:51babca0f7a328f4db005c6a16daa8ac4a4306c46fe3ca89382919134aa09179
        - quay.io/strimzi/jmxtrans@sha256:62272e8a03da2144340bd7267f7a824d1ed70f4f3e30dc7aadfba9ee1b9baf76
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/maven-builder@sha256:f781b6b857d437c2de62585462e97b9fae357912c9f6711c1df39f431278878d
        - quay.io/strimzi/operator@sha256:f4ee3ca482a2dd19785f59634578a0850273444c1173a5699a99cf02f3a018fe
        - quay.io/strimzi/kafka@sha256:1da35fd7b36b113d1343e254fdc6efd97ca8660cc7977d8985aa3a3547d5ebb1
        - quay.io/strimzi/kafka@sha256:1ace2cc0db9653ac1b388477fa56c3c1cd97c7b3846eba3c7824dc58ea179476
        version: 0.34.0
      entries:
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.34.x
    - currentCSV: strimzi-cluster-operator.v0.35.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:06a94a3021cf028ccc1a49271f35f79216029e344536e664f196c1725ff2c663
          createdAt: "2023-06-07 16:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.35 is the last version with support for Kubernetes 1.19 and 1.20. From Strimzi 0.36, only Kubernetes 1.21 and newer will be supported.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.35.1/deploying.html#assembly-upgrade-str).
          ### New in 0.35
          * Move feature gate `UseStrimziPodSets` to GA and remove support for StatefulSets
          * Remove support for JMX Trans
          * Redesigned the Cluster and User Operator configuration to make it more efficient and flexible
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:39778b90c2b2afc30261e4ad5135805e1a10a2b60e2e53108fb9f80487f1208a
        - quay.io/strimzi/maven-builder@sha256:88a79eff3b3a386880a630658964b7754caed9e99dd6e645a4c0d23d0fdb47ee
        - quay.io/strimzi/operator@sha256:06a94a3021cf028ccc1a49271f35f79216029e344536e664f196c1725ff2c663
        - quay.io/strimzi/kafka@sha256:4de4874a7b722ad813f4dcc58acf509527bca0609999b81e70d81e3b38534d9d
        - quay.io/strimzi/kafka@sha256:0d910e7138cb49e1cd8cd84cef88bce35698b93ddd683a3398f1d485a3162693
        - quay.io/strimzi/kafka@sha256:54c6b25b31f51ef401c1b6e2a1b27432911e819bf3e502e05186f01be3f798e5
        - quay.io/strimzi/kafka-bridge@sha256:d6be183e492f8f88157ab9fe0af53950df8b6711a8a8c33da465de6064f6f86e
        version: 0.35.1
      entries:
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.35.x
    - currentCSV: strimzi-cluster-operator.v0.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:e9e03b31007c0f9b67a6a33e6b644e8c1ba35bd9976804ee3d22a04e4bf511b1
          createdAt: "2023-07-10 21:47:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!! Strimzi 0.36.1 supports only Kubernetes 1.21 and newer! Kubernetes versions 1.19 and 1.20 are not supported anymore since Strimzi 0.36.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.35.1/deploying.html#assembly-upgrade-str).
          ### New in 0.36
          * Add support for Apache Kafka 3.4.1, 3.5.0, and 3.5.1
          * Remove support for 3.3.1 and 3.3.2
          * Add support for _Kafka node pools_
          * Add support for _Unidirectional Topic Operator_
          * Enable SCRAM-SHA authentication in KRaft mode (supported in Apache Kafka 3.5.0 and newer)
          * Add support for insecure flag in Maven artifacts in Kafka Connect Build
          * Improve Kafka rolling update to avoid rolling broker in log recovery
          * Added support for Kafka Exporter topic exclude and consumer group exclude parameters
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:9f35f40be3d47b0cb1f6628f7ffd9c6a687ab9f564829f341dd66fd0759c064f
        - quay.io/strimzi/kafka@sha256:67c2b563a65c85c9393608f9af7eaf12ea48dcc0ef5f726d51f27273a3845186
        - quay.io/strimzi/operator@sha256:e9e03b31007c0f9b67a6a33e6b644e8c1ba35bd9976804ee3d22a04e4bf511b1
        - quay.io/strimzi/kafka@sha256:ea35817ff1ef295c682d70562f209ef658fcd87c61f467a4549d7a2f39e71700
        - quay.io/strimzi/kafka-bridge@sha256:fefee252895e9f94756ed5c9ea71749e3fa89281df01bc55878a244ca7d17697
        - quay.io/strimzi/kaniko-executor@sha256:5ff9fc7d0b7b694875ad0150fc787c582d42da3b4864d807ee8a6524e47b0e4e
        - quay.io/strimzi/maven-builder@sha256:2a469b40122027b981914db5b5c102be48280e039c3089963422c181174af226
        - quay.io/strimzi/kafka@sha256:a0da6a48b5bcbff355365f07df2d99ef6443d22c5690495f10a6417e761b3799
        version: 0.36.1
      entries:
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.36.x
    - currentCSV: strimzi-cluster-operator.v0.37.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:52f376e64b9b31d74162e26248c278e9bd301d2d8b4240ee5bcf780dfe2f8ef8
          createdAt: "2023-09-03 17:45:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.37.0/deploying.html#assembly-upgrade-str).
          ### New in 0.37
          * The StableConnectIdentites feature gate moves to a beta stage.
          * Support for the ppc64le platform
          * Version fields to the Kafka custom resource status to track installation and upgrade state
          * Infinite auto-restarts of Kafka Connect and Kafka Mirror Maker 2 connectors
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:1de6e073e8625a193fcfd11dd1a6717cc17596911c10bcebd792c00b08a7ebaf
        - quay.io/strimzi/kafka@sha256:2997ba1c169eb792a6004d7a41b674560515f3ea6e62ab8b9a1fdb747eab40c1
        - quay.io/strimzi/kafka@sha256:dfeb1eb7b94107c443b6bc82fcfee7fcf782516c83c6fc64451fb6a7e2de74a6
        - quay.io/strimzi/kafka-bridge@sha256:5b821f6f3396002ffa7c5804391b627ff9da3b494b2777dde518c55c2b73cefc
        - quay.io/strimzi/kaniko-executor@sha256:5729f1f0511da95e64e358cfabb3980ad9a660af2968118a87d6755f5d652a67
        - quay.io/strimzi/maven-builder@sha256:995be06a50b86c992fcee9f00e44f6a21c3406975e4e35802f10e88b72ffb9df
        - quay.io/strimzi/operator@sha256:52f376e64b9b31d74162e26248c278e9bd301d2d8b4240ee5bcf780dfe2f8ef8
        - quay.io/strimzi/kafka@sha256:38bbadcdb196a373e1698be890caa1dc01134ce91763d8f95c71768a117e1932
        version: 0.37.0
      entries:
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.37.x
    - currentCSV: strimzi-cluster-operator.v0.38.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:aee45bbe439335e45a0423e6565bae0d0727d85c7b807812949bfa127d467fe6
          createdAt: "2023-10-14 18:18:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.38.0/deploying.html#assembly-upgrade-str).
          ### New in 0.38
          * Add support for Apache Kafka 3.6.0 and remove support for 3.4.0 and 3.4.1
          * Container signing and SBOMs
          * Stopping Connect and MirrorMaker 2 connectors
          * Manual rolling updates of Kafka Connect and Kafka Mirror Maker 2 pods using annotation
          * Prevent scale-down of brokers that still contain any replicas
          * Run ZooKeeper- and KRaft-based clusters in parallel
          * Unidirectional Topic Operator improvements
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:0cc01438e8596643407b6b8ff05fafcea7ef54fa9410d4ae2492a9dbf41118b1
        - quay.io/strimzi/kafka@sha256:20d451fb32ec82eab3c16891198ace46bb9b9193f1f5511bb60ce249080dd7d4
        - quay.io/strimzi/kafka-bridge@sha256:13838889cc20a3a382ea64e749cad94f997c0b928840588f2d1befcf19247bea
        - quay.io/strimzi/kaniko-executor@sha256:4fe7a3f3f78e93369284a5ab456873c911230926ca071d1d88167e63d8ce376c
        - quay.io/strimzi/maven-builder@sha256:922d91dab9e2b8e14b6db25d8b48d7f4e3521bedd435468cc22039599ac4f888
        - quay.io/strimzi/operator@sha256:aee45bbe439335e45a0423e6565bae0d0727d85c7b807812949bfa127d467fe6
        - quay.io/strimzi/kafka@sha256:50e24c797455f9cdbb3cdde8e4f26bb8e1f0503469456da5d179017fd3e03ffe
        version: 0.38.0
      entries:
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.38.x
    - currentCSV: strimzi-cluster-operator.v0.39.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.6.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:50b49f9b5e554cca517fc2794e11319d5aaba1676494b285baa3a74fc79eb501
          createdAt: "2023-12-14 20:28:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.39.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.39 is the last minor release with support for Kubernetes 1.21 and 1.22.** From Strimzi 0.40 on, only Kubernetes 1.23 and newer will be supported.
          ### New in 0.39
          * Add support for Apache Kafka 3.5.2 and 3.6.1
          * The `StableConnectIdentities` feature gate moves to GA stage and is now permanently enabled without the possibility to disable it.
            All Connect and Mirror Maker 2 operands will now use StrimziPodSets.

          * The `KafkaNodePools` feature gate moves to the beta stage and is enabled by default.
            If needed, `KafkaNodePools` can be disabled in the feature gates configuration in the Cluster Operator.

          * The `UnidirectionalTopicOperator` feature gate moves to the beta stage and is enabled by default.
            If needed, `UnidirectionalTopicOperator` can be disabled in the feature gates configuration in the Cluster Operator.

          * Improved Kafka Connect metrics and dashboard example files
          * Allow specifying and managing KRaft metadata version
          * Add support for KRaft to KRaft upgrades (Apache Kafka upgrades for the KRaft-based clusters)
          * Improved Kafka Mirror Maker 2 dashboard example file
          * Support for rolling updates of KRaft controller nodes
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:dd1706f36031c96dfcd0da7b3f804168e5f076220caa302973455fdcd76bd03e
        - quay.io/strimzi/kaniko-executor@sha256:723b7add2473e73c45c81361e1708af4971c321b0391c1298a0557fb8ca4f7f6
        - quay.io/strimzi/maven-builder@sha256:6ce3c16284379d5db9e42a6f3e2b765c981091164a28fdd6bf927b6c5d4ab557
        - quay.io/strimzi/kafka@sha256:49b7530ada1175750ad5a706e545ba3594d85e1e3a90f98b0200a73fa9875057
        - quay.io/strimzi/kafka@sha256:84691b3afb67b312400e01572ff08b64c1e1b1d32a4276e0c21bba10c71c36fe
        - quay.io/strimzi/kafka-bridge@sha256:13838889cc20a3a382ea64e749cad94f997c0b928840588f2d1befcf19247bea
        - quay.io/strimzi/kafka@sha256:53b960411e852f5cb182ecc5d395a3562865f83e15e6251f3a9af0eb413feeed
        - quay.io/strimzi/kafka@sha256:2badc65532b483bba6c1f6a8b0c9895e993ce24e9ebd3cd953d0d17fbe36d27a
        - quay.io/strimzi/operator@sha256:50b49f9b5e554cca517fc2794e11319d5aaba1676494b285baa3a74fc79eb501
        version: 0.39.0
      entries:
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.39.x
    - currentCSV: strimzi-cluster-operator.v0.40.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:2228d89b53e3524e69d1bcf5c2cd373884ce696d9005389a6c78dad23310e837
          createdAt: "2024-03-09 21:27:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.40.0/deploying.html#assembly-upgrade-str).

          **From Strimzi 0.40 on, we support only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.40
          * Add support for Apache Kafka 3.7.0. Remove support for Apache Kafka 3.5.0, 3.5.1, and 3.5.2.
          * The `UseKRaft` feature gate moves to beta stage and is enabled by default. If needed, `UseKRaft` can be disabled in the feature gates configuration in the Cluster Operator.
          * Add support for ZooKeeper to KRaft migration by enabling the users to move from using ZooKeeper to store metadata to KRaft.
          * Add support for moving from dedicated controller-only KRaft nodes to mixed KRaft nodes
          * Added support for Kafka Exporter `offset.show-all` parameter
          * Prevent removal of the `broker` process role from KRaft mixed-nodes that have assigned partition-replicas
          * Improve broker scale-down prevention to continue in reconciliation when scale-down cannot be executed
          * Added support for Tiered Storage by enabling the configuration of custom storage plugins through the Kafka custom resource.
          * Update HTTP bridge to the latest 0.28.0 release
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:73deb1af0d93e9fa8955a5a412d0e7138384bdc80e7574e815d21796e446b636
        - quay.io/strimzi/kaniko-executor@sha256:3fb4ace22589fcf8d3f01b74c37b10b32d87e15d103ea59e0933b61c4e46ff4a
        - quay.io/strimzi/maven-builder@sha256:d02a9b48c77a44eeab6426c9af4adf75405217994bcb01ce0196e616041350d8
        - quay.io/strimzi/operator@sha256:2228d89b53e3524e69d1bcf5c2cd373884ce696d9005389a6c78dad23310e837
        - quay.io/strimzi/kafka@sha256:2c143eb10b66037ef877311eb767457e4b877d5ff2c47d393c76babec155caff
        - quay.io/strimzi/kafka@sha256:e2b9f2835498831bbf30ac2d651de143d57088bcb23943101b066e941db92aca
        - quay.io/strimzi/kafka@sha256:c491a95f9b9b58f406461faf222ef612ad8d2f4d7b53f7f3645b20c99d7d1583
        version: 0.40.0
      entries:
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.40.x
    - currentCSV: strimzi-cluster-operator.v0.41.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:4c687a4fac96ffc22433ca6d00fb1bd31722767b44c9d87c465f95b57fb70089
          createdAt: "2024-05-09 22:23:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.41.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.41 supports only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.41
          * Support for Apache Kafka 3.6.2
          * Metrics to monitor CA expiration
          * Support for JBOD storage in KRaft mode. (Note: JBOD support in KRaft mode is considered early access in Apache Kafka 3.7.x)
          * Support for changing topic replication factor by changing the `KafkaTopic` resources.
          * The `KafkaNodePools` feature gate moves to GA stage and is permanently enabled without the possibility to disable it. To use the Kafka Node Pool resources, you still need to use the `strimzi.io/node-pools: enabled` annotation on the `Kafka` custom resources.
          * Support for configuring the `externalIPs` field in node port type services.
          * The `UnidirectionalTopicOperator` feature gate moves to GA stage and is permanently enabled without the possibility to disable it. If the topics whose names start with `strimzi-store-topic` and `strimzi-topic-operator` still exist, you can delete them.
          * Improve validation of `KafkaMirrorMaker2` resource to make sure the Connect cluster is set to the target cluster
          * Continue reconciliation after failed manual rolling update using the `strimzi.io/manual-rolling-update` annotation (when the `ContinueReconciliationOnManualRollingUpdateFailure` feature gate is enabled)
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/maven-builder@sha256:17f09b07fc41a059e6a048cb766046d43a072cfd3a8266c97388a37543dadee8
        - quay.io/strimzi/kafka@sha256:c3a346235ce4dea42986648ca07f2482d8fc0bf654c1d441b1301a23999304db
        - quay.io/strimzi/kafka@sha256:bf49499980bbe5d4dd399de0dd396089b0fd9dd29bc46e14d39b20f1e826aec1
        - quay.io/strimzi/kafka@sha256:41811bb106d6182978a889eaedffeadffaa5ab1599c4cd8e4f5015b131c7ac30
        - quay.io/strimzi/kafka@sha256:6243d98890c665ef65b5f5fa2ffd046412281ffef31bc7a07cac37bd77b2579d
        - quay.io/strimzi/kafka-bridge@sha256:c833f930ead2d1e9054912a20689ee979d335f5e5ca8769e3c6b8a5514a0df96
        - quay.io/strimzi/operator@sha256:4c687a4fac96ffc22433ca6d00fb1bd31722767b44c9d87c465f95b57fb70089
        - quay.io/strimzi/kaniko-executor@sha256:9404037bd5431b012c8c633ec3345907667453c5532e5bfaaedc78286de22c62
        version: 0.41.0
      entries:
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.41.x
    - currentCSV: strimzi-cluster-operator.v0.42.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:ca5c1b70e1196deec4bf519921bfa55a967aa1ea15bd4c182bc56bea08d2e114
          createdAt: "2024-07-03 12:34:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.42.0/deploying.html#assembly-upgrade-str).

          **From version 0.40.0, Strimzi supports only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.42
          * Support for Apache Kafka 3.7.1
          * The UseKRaft feature gate moves to GA stage and is permanently enabled without the possibility of disabling it. To use KRaft (ZooKeeper-less Apache Kafka), you still need to use the strimzi.io/kraft: enabled annotation on the Kafka custom resources or migrate from an existing ZooKeeper-based cluster.
          * Update the base image used by Strimzi containers from UBI8 to UBI9
          * Add support for filename patterns when configuring trusted certificates
          * Enhance KafkaBridge resource with consumer inactivity timeout and HTTP consumer/producer enablement.
          * Add support for setting publishNotReadyAddresses on services for listener types other than internal.
          * Added support for configuring the quotas plugin with type strimzi or kafka in the Kafka custom resource. The Strimzi Quotas plugin version was updated to 0.3.1.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:6527c426f753521c07f3a3df2ee432273a41b109594096b4c6cae2e8e609d669
        - quay.io/strimzi/kafka@sha256:6c519531cdd5f40f326b295d8827344140efe2607a1826f7f1cab8c3427f43f7
        - quay.io/strimzi/kafka-bridge@sha256:fdddffb3de59ba48c05964bb0e760593088d24d1ee35b090112fe6cc6b553cd8
        - quay.io/strimzi/kafka@sha256:bc49c10b3e839e161e515c3a1e411d1f8df9de5e2c8d71aa7b63ccdbbae9e8d2
        - quay.io/strimzi/kaniko-executor@sha256:11821b0078a270612f1bcc37f5dc716ab273c974648c40f95fa4804352e647a7
        - quay.io/strimzi/operator@sha256:ca5c1b70e1196deec4bf519921bfa55a967aa1ea15bd4c182bc56bea08d2e114
        - quay.io/strimzi/maven-builder@sha256:342e8d0086c53cf7334127798e5bd3764b28f5e4909ba67d65f6d937dc3dba05
        - quay.io/strimzi/kafka@sha256:841a46e9c10481df3ef702412caf362b55ae07c30a3a5927073356cd36dc049f
        - quay.io/strimzi/kafka@sha256:c56c5abf94897f20f976603cba0678a33cdfbbcd6f9030a193f7684cb9894a6c
        version: 0.42.0
      entries:
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.42.x
    - currentCSV: strimzi-cluster-operator.v0.43.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:d3732245f34afaa6b571f1f912e1cd017891b22003d6367ae8ca059b2a5cb13c
          createdAt: "2024-08-22 15:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.43.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.43 is the last version with support for Kubernetes 1.23 and 1.24.** From Strimzi 0.44, only Kubernetes versions 1.25 and newer will be supported.
          ### New in 0.43
          * Support for Apache Kafka 3.8.0
          * Removed support for Apache Kafka 3.6.0, 3.6.1, and 3.6.2
          * Specifying additional volumes and volume mounts in Strimzi custom resources
          * Support for custom Cruise Control API users
          * Unregistration of KRaft nodes after scale-down
          * Kafka Exporter updated to 1.8.0 (and the Grafana dashboard updated to work with it)
          * Service account based authentication
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:80f329ecd3ba96fac12e75be874dce0debc8395337737581de34238ce11c8fb1
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:6414f8f3985f221b0eda53867c00494140108d17579c4dcbbc93934bb00e5d49
        - quay.io/strimzi/kafka@sha256:d27cf5bf3624e18ba705246b14f92d12b8a15eb8551b13abf027cf71b549baf0
        - quay.io/strimzi/kafka@sha256:ad06ead5125bce87498b1582a76e78143a6d21408ac4b12bb7e6007aa4cabd84
        - quay.io/strimzi/operator@sha256:d3732245f34afaa6b571f1f912e1cd017891b22003d6367ae8ca059b2a5cb13c
        - quay.io/strimzi/kafka@sha256:b2f4e8895e9ed66228512ce14aec685101fa75dc329128d47cb1c15e3aef3c9c
        version: 0.43.0
      entries:
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.43.x
    - currentCSV: strimzi-cluster-operator.v0.44.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:b07b81f7e282dea2e4e29a7c93cfdd911d4715a5c32fe4c4e3d7c71cba3091e8
          createdAt: "2024-08-22 15:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.44.0/deploying.html#assembly-upgrade-str).

          **From version 0.44, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**

          **ZooKeeper support will be soon removed from Apache Kafka and Strimzi.** Currently, the last Strimzi version with ZooKeeper support is expected to be Strimzi 0.45. Please plan your migration to KRaft (ZooKeeper-less Apache Kafka) accordingly. Follow the [documentation](https://strimzi.io/docs/operators/latest/full/deploying.html#assembly-kraft-mode-str) for more details.

          **Kafka Mirror Maker 1 support will be soon removed from Apache Kafka and Strimzi.** Currently, the last Strimzi version with Mirror Maker 1 support is expected to be Strimzi 0.45. Please plan your migration to Mirror Maker 2 or another mirroring tool.

          ### New in 0.44
          * Managing offsets in Connect and Mirror Maker 2 connectors through custom resources
          * Templating `advertisedHost` and `host` fields in listener configuration
          * Configuring environment variables in operands based on Secrets and ConfigMaps
          * Disabling Pod disruption Budget management
          * Auto-rebalancing at scale-up/scale-down
          * `ContinueReconciliationOnManualRollingUpdateFailure` feature gate moves to the beta stage and is enabled by default
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:9a4224635e09ece2c1d3ffe9d715813cf63037461c9f06c1b5f53554aed5ecdc
        - quay.io/strimzi/kafka-bridge@sha256:cc95ce9f496057f3b18d5e8297b40a410cc66f9fd4a2227615971458a26bac25
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:7b55f971a3b62923a9a61a61c82f542e07deeb6ffdbaede68244ba358a0853f2
        - quay.io/strimzi/operator@sha256:b07b81f7e282dea2e4e29a7c93cfdd911d4715a5c32fe4c4e3d7c71cba3091e8
        - quay.io/strimzi/kafka@sha256:238bb080825b7ecebaa819be09788bc3d056ae8162ebcd41444076dc22f85740
        - quay.io/strimzi/kafka@sha256:98d332b7fa5c9e48f60b516e00ed5812bd9268e014fd2337284e7cc3e664f1eb
        version: 0.44.0
      entries:
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.44.x
    - currentCSV: strimzi-cluster-operator.v0.45.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:58e8722216b0b0e8b1dd79109725238fa4016a16cc6a9fca0e327bccb1506810
          createdAt: "2025-06-27 01:38:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.45.1/deploying.html#assembly-upgrade-str).

          **From version 0.45, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**

          **ZooKeeper and Kafka Mirror Maker 1 support will be removed from Apache Kafka and Strimzi.** Strimzi 0.45 is the last minor Strimzi version with support for ZooKeeper-based Apache Kafka clusters and MirrorMaker 1 deployments. Please make sure to [migrate to KRaft](https://strimzi.io/docs/operators/latest/full/deploying.html#assembly-kraft-mode-str) and MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          **Strimzi EnvVar Configuration Provider and Strimzi MirrorMaker 2 Extensions will be removed from Strimzi.** Strimzi 0.45 is the last Strimzi version to include the Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0). Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.
          ### New in 0.45.1
          * Add support for Kafka 3.9.1
          * Update HTTP bridge to 0.31.2
          * Update Strimzi Kafka OAuth library to 0.15.1
          * Fixed bug which may lead to metadata loss within the cluster when restarting a KRaft migration after a previous rollback due to missing `/migration` znode deletion.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:e98fbe49cadbc5d3afbd8f17abea218f216c95d52f4fa1fe514a7a988624da24
        - quay.io/strimzi/kafka@sha256:ba52ed046b1dccdbd96f4e68057ce014d862a7c9c1fc670760c023b9aa09f23f
        - quay.io/strimzi/kafka-bridge@sha256:c462441a966f0732855e684b4840563bfad45391345c9511fa49e1630ba9c3e3
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:770e410f1e7e0e92c31bae6c19831ba36d59dd11b44e90c159aef16a965aa998
        - quay.io/strimzi/kafka@sha256:173cb5793756f75739c51a83bbe06d7f0094fd003a2982160a5cc6e216cd6d28
        - quay.io/strimzi/kafka@sha256:6a05924839a2c7da0431408e6d6223a055e75d88b3f057ac5caa7f9097227f56
        - quay.io/strimzi/operator@sha256:58e8722216b0b0e8b1dd79109725238fa4016a16cc6a9fca0e327bccb1506810
        version: 0.45.1
      entries:
      - name: strimzi-cluster-operator.v0.45.1
        version: 0.45.1
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.45.x
    - currentCSV: strimzi-cluster-operator.v0.46.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.0.0",
                      "metadataVersion": "4.0-IV3",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:6a22e4968e9d409405423d12c04a0e3fecaea2ee2794bd27de7cf21f59453265
          createdAt: "2025-06-26 11:58:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.46.0/deploying.html#assembly-upgrade-str).

          **From version 0.46, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**
          **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed. Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!
          **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images. Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.
          **Support for MirrorMaker 1 has been removed. Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.
          ### New in 0.46.1
          * Update CSV file for OLM installation to add missing permission for broker Pods to "get" nodes Kubernetes resource.
          * Allow the injection of annotations on the deployment object in Helm charts.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:3e1815aed3f74b8573cebfced66b2c78ca0cf7c746c26db2002380cd39d09ae6
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:b532234c234fc30fecec992ad1a6154c5ef08ebf9a20ec8bc62e0c88d791ff36
        - quay.io/strimzi/kafka@sha256:f3837f8cb2a60e906e84486128c41c398424f4085df2348e9e0486bfc12089fd
        - quay.io/strimzi/kafka@sha256:82273d873ce21a73f8f3118bf61627cb15734b43ea7ca92d43b290340b44e485
        - quay.io/strimzi/operator@sha256:6a22e4968e9d409405423d12c04a0e3fecaea2ee2794bd27de7cf21f59453265
        version: 0.46.1
      entries:
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.46.x
    - currentCSV: strimzi-cluster-operator.v0.47.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.0.0",
                      "metadataVersion": "4.0-IV3",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:589c4d63641d9a944462dd682f6e5febe8b38ff55b9253949b061aca16feb154
          createdAt: "2025-07-12 17:21:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.47 is the last Strimzi version with support for Kubernetes 1.25 and 1.26.**
            From Strimzi 0.48.0 on, we will support only Kubernetes 1.27 and newer.

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.47.0
          * Support for Kafka 3.9.1
          * Progress tracking for Cruise Control rebalances
          * Support for Kubernetes Image Volumes to mount custom plugins
          * Support for [JsonTemplateLayout](https://logging.apache.org/log4j/2.x/manual/json-template-layout.html) in Operators, Cruise Control, and Kafka 4.0.0
          * Support for Strimzi Metrics Reporter for the Kafka brokers / controllers
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/operator@sha256:589c4d63641d9a944462dd682f6e5febe8b38ff55b9253949b061aca16feb154
        - quay.io/strimzi/maven-builder@sha256:c67a3ad3d92420ac8b7f8435eeec53ebffaea4e894e65414a7c47c20e3faba27
        - quay.io/strimzi/kafka@sha256:978e2ed5f0d315e8bf0ab0503ee839acbd9da01ca427e6581b59ac25d36bc9c9
        - quay.io/strimzi/kafka@sha256:8a793cdf5084cdb72c6663419530985158b16330a518720e9e9d1d96babbaf7e
        - quay.io/strimzi/kafka@sha256:551882f47a6d2dc8300bcfca251fffffc3379acb2724a7efe16f17c0c4f529fb
        - quay.io/strimzi/kafka-bridge@sha256:428c5587f547dabfc995a73745569d113380c7e6b8bf0b2504c6a9a00d6c4608
        version: 0.47.0
      entries:
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.47.x
    - currentCSV: strimzi-cluster-operator.v0.48.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.1.0",
                      "metadataVersion": "4.1-IV1",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.1.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
          createdAt: "2025-09-25 10:32:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.48.0 supports only Kubernetes 1.27 and newer.**
            Kubernetes 1.25 and 1.26 are not supported anymore

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.48.0
          * Support for Kafka 4.1.0
          * Add monitoring of custom resources using kubernetes-state-metrics (KSM)
          * Support for type: custom client authentication
          * Added support for Strimzi Metrics Reporter to Kafka Connect, Mirror Maker 2 and Kafka Bridge.
          * Added support for KIP-745 in Kafka Connector and Mirror Maker 2, allowing the usage of includeTasks and onlyFailed arguments in Kafka connectors restart.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:64380efcb8da87fdc0e0de238a20456be9abb830174c2f2331882b7a84657e57
        - quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
        - quay.io/strimzi/kafka@sha256:48db05ad2e9fa39eb388089e7e25654bcbf4da1ddc3a5635d675056c28bbcb96
        - quay.io/strimzi/kafka@sha256:34450afc1a3399a9026fc3c6ec90e9f6adfec9faf6520545b4d84e90cda34964
        - quay.io/strimzi/kafka-bridge@sha256:53034f64f0b672f10b5bacea1c7a25132f118df7fd5c9032c4dbf702ed93796a
        version: 0.48.0
      entries:
      - name: strimzi-cluster-operator.v0.48.0
        version: 0.48.0
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.48.x
    defaultChannel: stable
    packageName: strimzi-kafka-operator
    provider:
      name: Strimzi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: web-terminal
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: web-terminal.v1.11.1-0.1747215995.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: registry.redhat.io/web-terminal/web-terminal-rhel9-operator@sha256:0478d14e92df84fdfbf4584384d34dc9a71427a6487c2564d2eb7815ba1ac12b
          createdAt: "2021-10-26T07:24:32Z"
          description: Start a web terminal in your browser with common CLI tools
            for interacting with the cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.substitutesFor: web-terminal.v1.11.1
          operatorframework.io/suggested-namespace: openshift-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/redhat-developer/web-terminal-operator/
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - kind: DevWorkspaceRouting
            name: devworkspaceroutings.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha1
        description: |
          Start a web terminal in your browser with common CLI tools for interacting with
          the cluster.

          **Note:** The Web Terminal Operator integrates with the OpenShift Console in
          OpenShift 4.5.3 and higher to simplify web terminal instance creation and
          automate OpenShift login. In earlier versions of OpenShift, the operator can
          be installed but web terminals will have to be created and accessed manually.

          ## Description
          The Web Terminal Operator leverages the
          [DevWorkspace Operator](https://github.com/devfile/devworkspace-operator)
          to provision enviroments which support common cloud CLI tools. When this
          operator is installed, the DevWorkspace Operator will be installed as a
          dependency.

          ## How to Install
          Press the **Install** button, choose the upgrade strategy, and wait for the
          **Installed** Operator status.

          When the operator is installed, you will see a terminal button appear on the
          top right of the console after refreshing the OpenShift console window.

          ## How to Uninstall
          The Web Terminal Operator requires manual steps to fully uninstall the operator.
          As the Web Terminal Operator is designed as a way to access the OpenShift
          cluster, web terminal instances store user credentials. To avoid exposing these
          credentials to unwanted parties, the operator deploys webhooks and finalizers
          that aren't removed when the operator is uninstalled. See the
          [uninstall guide](https://docs.openshift.com/container-platform/latest/web_console/web_terminal/uninstalling-web-terminal.html)
          for more details.

          ## Documentation
          Documentation for this Operator is available at https://docs.openshift.com/container-platform/latest/web\_console/web\_terminal/installing-web-terminal.html
        displayName: Web Terminal
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - workspace
        - devtools
        - developer
        - ide
        - terminal
        links:
        - name: Web Terminal Repo
          url: https://github.com/redhat-developer/web-terminal-operator/
        maintainers:
        - email: aobuchow@redhat.com
          name: Andrew Obuchowicz
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/web-terminal/web-terminal-tooling-rhel9@sha256:5c24220f884dcdf1b1e5ac1e20dc6b7c8c4300bb89e8f118c0b11c331a56ab3f
        - registry.redhat.io/web-terminal/web-terminal-exec-rhel9@sha256:cfc8200340655a045f45d02fa327538f87e98d6369bdef2b46cf447053c44426
        - registry.redhat.io/web-terminal/web-terminal-rhel9-operator@sha256:0478d14e92df84fdfbf4584384d34dc9a71427a6487c2564d2eb7815ba1ac12b
        version: 1.11.1+0.1747215995.p
      entries:
      - name: web-terminal.v1.11.1-0.1747215995.p
        version: 1.11.1+0.1747215995.p
      - name: web-terminal.v1.11.1
        version: 1.11.1
      - name: web-terminal.v1.11.0
        version: 1.11.0
      - name: web-terminal.v1.10.1
        version: 1.10.1
      - name: web-terminal.v1.10.1-0.1740684238.p
        version: 1.10.1+0.1740684238.p
      - name: web-terminal.v1.10.0-0.1731481377.p
        version: 1.10.0+0.1731481377.p
      - name: web-terminal.v1.10.0-0.1732652667.p
        version: 1.10.0+0.1732652667.p
      - name: web-terminal.v1.10.0
        version: 1.10.0
      - name: web-terminal.v1.10.0-0.1727169028.p
        version: 1.10.0+0.1727169028.p
      - name: web-terminal.v1.10.0-0.1720435222.p
        version: 1.10.0+0.1720435222.p
      - name: web-terminal.v1.10.0-0.1720402943.p
        version: 1.10.0+0.1720402943.p
      - name: web-terminal.v1.9.0-0.1708477317.p
        version: 1.9.0+0.1708477317.p
      - name: web-terminal.v1.9.0
        version: 1.9.0
      - name: web-terminal.v1.8.0-0.1708477299.p
        version: 1.8.0+0.1708477299.p
      - name: web-terminal.v1.8.0-0.1701199376.p
        version: 1.8.0+0.1701199376.p
      - name: web-terminal.v1.8.0-0.1692219820.p
        version: 1.8.0+0.1692219820.p
      - name: web-terminal.v1.8.0
        version: 1.8.0
      - name: web-terminal.v1.7.0-0.1682321121.p
        version: 1.7.0+0.1682321121.p
      - name: web-terminal.v1.7.0-0.1692219820.p
        version: 1.7.0+0.1692219820.p
      - name: web-terminal.v1.7.0
        version: 1.7.0
      - name: web-terminal.v1.7.0-0.1681197295.p
        version: 1.7.0+0.1681197295.p
      - name: web-terminal.v1.7.0-0.1708477265.p
        version: 1.7.0+0.1708477265.p
      - name: web-terminal.v1.7.0-0.1684429884.p
        version: 1.7.0+0.1684429884.p
      - name: web-terminal.v1.6.0
        version: 1.6.0
      - name: web-terminal.v1.6.0-0.1692219820.p
        version: 1.6.0+0.1692219820.p
      - name: web-terminal.v1.5.1-0.1661829403.p
        version: 1.5.1+0.1661829403.p
      - name: web-terminal.v1.5.1
        version: 1.5.1
      - name: web-terminal.v1.5.0-0.1657220207.p
        version: 1.5.0+0.1657220207.p
      - name: web-terminal.v1.5.0
        version: 1.5.0
      - name: web-terminal.v1.4.0
        version: 1.4.0
      - name: web-terminal.v1.3.0
        version: 1.3.0
      name: fast
    defaultChannel: fast
    packageName: web-terminal
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM Quantum
      provider-url: ""
    name: openshift-qiskit-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-qiskit-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "qiskit.ibm.com/v1alpha1",
                "kind": "QiskitPlayground",
                "metadata": {
                  "name": "qiskitplayground-sample"
                },
                "spec": {
                  "image": "qiskit/jupyter:0.1",
                  "imagePullPolicy": "Always",
                  "resources": {
                    "requests": {
                      "memory": "2G",
                      "cpu": "1"
                    },
                    "limits": {
                      "memory": "3G",
                      "cpu": "1.5"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: docker.io/qiskit/operator:0.1
          createdAt: 2021-004-23 22:25:53
          description: |-
            Launches a developement environment Jupyter notebok development environment for
            developing quantum applications using Qiskit.
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator
          support: https://github.com/qiskit-community/openshift-quantum-operators/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QiskitPlayground is the Schema for the qiskitplaygrounds
              API
            displayName: Qiskit Playground
            kind: QiskitPlayground
            name: qiskitplaygrounds.qiskit.ibm.com
            version: v1alpha1
        description: |-
          ### Introduction
          Launches a developement Jupiter lab environment with Qiskit.

          QiskitPlayground is a Jupyter notebook with the pre installed qiskit libraries for implementing quantum application development.
          It comes with pre installed python packages for visualising results. QiskitPlayground can run quantum
          applications on simulator or on real quantum machines using IBM Q account.

          ### Documentation
          For detailed installation guide [follow](https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator)

          ### Contributing
          You can contribute by
            * Raising any [issues](https://github.com/qiskit-community/openshift-quantum-operators/issues) you find using ibm-quantum-operator community operator
            * Fixing issues by opening [Pull Requests](https://github.com/qiskit-community/openshift-quantum-operators/pulls)
            * Talking about openshift-qiskit-operator

          ### License
          kubernetes-qiskit-operator is licensed under the [Apache 2.0 license](https://github.com/qiskit-community/openshift-quantum-operators/blob/master/LICENSE)

          #### Note
          Any api could be changed any time without any notice. That said, your feedback is very important and appreciated to make this project more stable and useful.
        displayName: QiskitPlayground
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Qiskit
        - Open Source
        - Quantum
        links:
        - name: GitHub
          url: https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator
        maintainers:
        - email: blublinsky@ibm.com
          name: boris
        maturity: alpha
        provider:
          name: IBM Quantum
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - docker.io/qiskit/operator:0.1
        version: 0.2.1
      entries:
      - name: openshift-qiskit-operator.v0.2.1
        version: 0.2.1
      - name: openshift-qiskit-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: openshift-qiskit-operator
    provider:
      name: IBM Quantum
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: ""
    name: hawtio-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hawtio-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.14.0",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/hawtio/operator:0.5.0
          createdAt: "2022-04-08T00:00:00Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Hawtio Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - docker.io/hawtio/operator:0.5.0
        version: 0.5.0
      entries:
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: alpha
    - currentCSV: hawtio-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/hawtio/operator:1.2.0
          createdAt: "2024-12-02T17:22:48Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          olm.skipRange: '>=1.0.0 <1.0.2'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Hawtio console that eases the discovery and management
              of hawtio-enabled applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#jvm-agent).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/hawtio/operator:1.2.0
        version: 1.2.0
      entries:
      - name: hawtio-operator.v1.2.0
        version: 1.2.0
      - name: hawtio-operator.v1.1.1
        version: 1.1.1
      - name: hawtio-operator.v1.1.0
        version: 1.1.0
      - name: hawtio-operator.v1.0.1
        version: 1.0.1
      - name: hawtio-operator.v1.0.0
        version: 1.0.0
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: latest
    - currentCSV: hawtio-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/hawtio/operator:1.2.0
          createdAt: "2024-12-02T17:22:48Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          olm.skipRange: '>=1.0.0 <1.0.2'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Hawtio console that eases the discovery and management
              of hawtio-enabled applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#jvm-agent).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/hawtio/operator:1.2.0
        version: 1.2.0
      entries:
      - name: hawtio-operator.v1.2.0
        version: 1.2.0
      - name: hawtio-operator.v1.1.1
        version: 1.1.1
      - name: hawtio-operator.v1.1.0
        version: 1.1.0
      - name: hawtio-operator.v1.0.1
        version: 1.0.1
      - name: hawtio-operator.v1.0.0
        version: 1.0.0
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: stable-v1
    defaultChannel: stable-v1
    packageName: hawtio-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "docker.io/percona/percona-server-mongodb:7.0.18-11",
                  "imagePullPolicy": "Always",
                  "initImage": "docker.io/percona/percona-server-mongodb-operator:1.20.1",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:2.44.1",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "docker.io/percona/percona-backup-mongodb:2.9.1",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.io/percona/percona-server-mongodb-operator:1.20.1
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.20.1
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - docker.io/percona/percona-server-mongodb-operator:1.20.1
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator.v1.20.0
        version: 1.20.0
      - name: percona-server-mongodb-operator.v1.19.1
        version: 1.19.1
      - name: percona-server-mongodb-operator.v1.19.0
        version: 1.19.0
      - name: percona-server-mongodb-operator.v1.18.0
        version: 1.18.0
      - name: percona-server-mongodb-operator.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: directpv
      provider-url: ""
    name: minio-volume-manager-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-volume-manager-operator.v5.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.quay.io/v1alpha1",
                "kind": "AistorVolumemanager",
                "metadata": {
                  "name": "aistorvolumemanager-sample"
                },
                "spec": {
                  "annotations": {},
                  "controller": {
                    "podAnnotations": {},
                    "podLabels": {},
                    "replicas": 3
                  },
                  "images": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": []
                  },
                  "labels": {},
                  "license": "",
                  "nodeserver": {
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-10-05T23:18:24Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AistorVolumemanager
            name: aistorvolumemanagers.charts.quay.io
            version: v1alpha1
        description: directpv
        displayName: directpv
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - directpv
        links:
        - name: Private Dpv Operator
          url: https://private-dpv-operator.domain
        maintainers:
        - email: cesarcelis@minio.io
          name: celis
        maturity: alpha
        provider:
          name: directpv
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - quay.io/cniackz4/directpv-operator@sha256:af06e57661a7e05809e44c3fac571ed949cc04fd656fdb0c8324b533778eb604
        version: 5.0.3
      entries:
      - name: minio-volume-manager-operator.v5.0.3
        version: 5.0.3
      - name: private-dpv-operator.v5.0.0
        version: 5.0.0
      name: stable
    defaultChannel: stable
    packageName: minio-volume-manager-operator
    provider:
      name: directpv
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-commonui-operator-app
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-commonui-operator.v4.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operators.ibm.com/v1alpha1",
                "kind": "CommonWebUI",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "example-commonwebui",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator"
                  },
                  "name": "example-commonwebui"
                },
                "spec": {
                  "commonWebUIConfig": {
                    "cpuLimits": "300",
                    "cpuMemory": "256",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.8.0",
                    "ingressPath": "/common-nav",
                    "landingPage": "",
                    "requestLimits": "300",
                    "requestMemory": "256",
                    "serviceName": "common-web-ui"
                  },
                  "globalUIConfig": {
                    "cloudPakVersion": "3.6.0",
                    "defaultAdminUser": "admin",
                    "defaultAuth": "",
                    "enterpriseLDAP": "",
                    "enterpriseSAML": "",
                    "osAuth": "",
                    "sessionPollingInterval": 5000
                  },
                  "operatorVersion": "4.8.0",
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "256Mi"
                    },
                    "requests": {
                      "cpu": "300m",
                      "memory": "256Mi",
                      "ephemeral-storage": "251Mi"
                    }
                  },
                  "version": "4.8.0"
                }
              },
              {
                "apiVersion": "foundation.ibm.com/v1",
                "kind": "NavConfiguration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "common-web-ui-config",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator",
                    "default": "true",
                    "name": "common-web-ui-config"
                  },
                  "name": "common-web-ui-config"
                },
                "spec": {
                  "about": {
                    "copyright": "© 2018, 2020 IBM. All rights reserved.",
                    "licenses": [
                      "yq, version 3.3.0, MIT+GPL",
                      "MongoDB, version 4.0.16 Community Edition, SSPL",
                      "Ansible: © 2017 Red Hat, Inc., http://www.redhat.com; © Henry Graham (hzgraham) \u003cHenry.Graham@mail.wvu.edu\u003e",
                      "calico-bird: © 1998–2008, Martin Mares \u003cmj@ucw.cz\u003e; © 1998–2000 Pavel Machek \u003cpavel@ucw.cz\u003e;",
                      "© 1998–2008, Ondrej Filip \u003cfeela@network.cz\u003e; © 2009–2013,  CZ.NIC z.s.p.o.",
                      "chrony: © Richard P. Curnow  1997-2003, GPL v2",
                      "collectd, © 2017-2018, version 5.7.2, GPL v2, \u003chttps://github.com/collectd/collectd/tree/collectd-5.7.2\u003e",
                      "crudini: © Pádraig Brady \u003cP@draigBrady.com\u003e",
                      "Galera-3: © 2007–2014 Codership Oy \u003cinfo@codership.com\u003e",
                      "glusterfs: © 2010–2013+ James Shubin \u003chttps://ttboj.wordpress.com/\u003e",
                      "haproxy: © 2000–2013  Willy Tarreau \u003cw@1wt.eu\u003e",
                      "heketi v6.0.0: © 2015 The heketi Authors, GPL v2",
                      "heketi v8.0.0: © 2015 The heketi Authors, GPL v2",
                      "heketi-master/apps.app.go: © 2015 The heketi Authors",
                      "heketi-master/client/api/go-client/backup.go: © 2016 The heketi Authors",
                      "heketi-master/doc/man/heketi-cli.8: © 2016 The heketi Authors",
                      "heketi-master/extras/docker/gluster/gluster-setup.sh: © 2016 Red Hat, Inc. \u003chttp://www.redhat.com\u003e",
                      "ieee-data: © 2013 Luciano Bello \u003cluciano@debian.org\u003e",
                      "javax.mail: © 2017 Oracle and/or its affiliates. All rights reserved.",
                      "keepalived: © 2001-2017 Alexandre Cassen \u003cacassen@gmail.com\u003e",
                      "libonig2: © 2006–2008 Max Kellermann \u003cmax@duempel.org\u003e; © 2014–2015 Jörg Frings-Fürst \u003cdebian@jff-webhosting.net\u003e",
                      "libtomcrypt: © 2004 Sam Hocevar \u003csam@hocevar.net\u003e, GPL v2",
                      "mariadb-common: © 2018 MariaDB. All rights reserved.",
                      "mariaDB: © 2018 MariaDB. All rights reserved. \u003chttps://mariadb.com/\u003e",
                      "mariadb-server: © 2018 MariaDB. All rights reserved.",
                      "minitar: © 2004 Mauricio Julio Fernandez Pradier and Austin Ziegler",
                      "MongoDB: © 2007 Free Software Foundation, Inc. \u003chttp://fsf.org/\u003e",
                      "nvmi-cli: © 1989, 1991 Free Software Foundation, Inc., GPL v2",
                      "OpenJDK: © 2018 Oracle Corporation and/or its affiliates",
                      "openshift-mariadb-galera: © 2007 Free Software Foundation, Inc. \u003chttp://fsf.org/\u003e",
                      "percona-xtrabackup: © 2006–2018 Percona LLC.",
                      "pwgen: © Christian Thöing \u003cc.thoeing@web.de\u003e",
                      "rdoc: © 2001–2003 Dave Thomas, The Pragmatic Programmers",
                      "readline: © Chet Ramey \u003cchet.ramey@case.edu\u003e",
                      "John the Ripper password cracker: © 1996–2013 by Solar Designer \u003csolar@openwall.com\u003e",
                      "spdx-exceptions: © 2018 SPDX Workgroup a Linux Foundation Project. All rights reserved.",
                      "socat: © 2001–2010 Gerhard Rieger",
                      "sshpass: © 2006, 2008 Lingnu Open Source Consulting Ltd. \u003chttp://www.lingnu.com\u003e",
                      "timelimit: © 2001, 2007 - 2010  Peter Pentchev, GPL v2",
                      "ua-parser-js: © 2012-2018 Faisal Salman \u003cf@faisalman.com\u003e, GPL v2",
                      "ubuntu-cloud-keyring: © 2010 Michael Vogt \u003cmichael.vogt@canonical.com\u003e",
                      "unboundid-ldapsdk: © 2015 UnboundID. The LDAP SDK for Java is developed by UnboundID. \u003cinfo@unboundid.com\u003e",
                      "xmpp4r: © Lucas Nussbaum \u003clucas@lucas-nussbaum.net\u003e, Stephan Maka \u003cstephan@spaceboyz.net\u003e, and others.",
                      "module-assistant: © 2003-2008 Eduard Bloch \u003cblade@debian.org\u003e, version 0.11.8, GPL v2;",
                      "© 2009 Cyril Brulebois \u003ckibi@debian.org\u003e, version 0.11.8, GPL v2; © 2013-2018 Andreas Beckmann \u003canbe@debian.org\u003e, version 0.11.8, GPL v2",
                      "module-init-tools: © 2011 ProFUSION embedded systems, version 22, GPL v2",
                      "thin: © 2017 Marc-Andre Cournoyer \u003cmacournoyer@gmail.com\u003e, version 1.7.2, GPL v2",
                      "gosu, © 1999-2014, version 1.1, GPL v3",
                      "mercurial (Python), © 2006-2018 ,version v4.5.3, GPL v2",
                      "garden-runc, © 2015-Present CloudFoundry.org Foundation, Inc. All Rights Reserved, version 1.17.0, GPLv2",
                      "libtomcrypt0, © 2003-2007 Tom St Denis \u003ctomstdenis@gmail.com\u003e, version 1.17-7, GPLv2",
                      "console-setup-min, © 1999,2000,2001,2002,2003,2006,2007,2008,2009,2010,2011 Anton Zinoviev, \u003canton@lml.bas.bg\u003e,version 1.108, GPLv2",
                      "dracut, © 2009 Harald Hoyer \u003charald@redhat.com\u003e, version 044+3-3, GPLv2",
                      "dracut-core, © 2009 Harald Hoyer \u003charald@redhat.com\u003e, version 044+3-3, GPLv2",
                      "g++, version 5.4.0-6ubuntu, GPL v2",
                      "libstdc++6, version 5.4.0-6ubuntu, GPL v3",
                      "libstdc++-5-dev, version 5.4.0-6ubuntu, GPL v3",
                      "docker-engine-selinux, version 3b5fac4, GPLv2",
                      "unorm, version 1.5.0, GPL v2",
                      "psmisc, version 22.20, GPL v2",
                      "lvm2-devel, version 2.0.2, GPL v2",
                      "nfs-utils, version 1.3, GPL v2",
                      "popt-static, version 1.13, GPL v2",
                      "sysvinit-tools, version 2.88, GPL v2",
                      "stunnel, version 5.53, GPL v2",
                      "stunnel, version 5.39, GPL v2",
                      "LVM2, version 2.02.180-10.el7_6.2, GPL v2",
                      "sysdig, version 2c43237, GPL",
                      "chisels, version 9722dbc, GPL",
                      "MongoDB, version 4.0.12, SSPL",
                      "ffi (Ruby Gem), 1.11.1, GPL",
                      "inotify-tools, v3.14, GPL v2",
                      "logrotate, v3.8.6, GPL v2",
                      "checker-qual, version 2.0.0, GPLv2",
                      "ocrad-bower, v1.0.0, GPL v3",
                      "Grafana, v7.5.12, AGPL",
                      "IBM Semeru Runtime Open Edition binaries, GPL v2",
                      "checker-compat-qual 2.5.5, GPL-2.0",
                      "javax.annotation-api 1.3.1, GPL-2.0"
                    ],
                    "logoUrl": "IBM Cloud Pak | Administration Hub",
                    "logoAltText": "IBM Cloud Pak | Administration Hub"
                  },
                  "header": {
                    "disabledItems": [
                      "createResource",
                      "catalog"
                    ],
                    "docUrlMapping": "http://ibm.biz/cpcs_adminui",
                    "logoAltText": "IBM Cloud Pak Administration Hub",
                    "logoHeight": "47px",
                    "logoUrl": "/common-nav/graphics/ibm-cloudpack-logo.svg",
                    "logoWidth": "190px"
                  },
                  "login": {
                    "loginDialog": {
                      "acceptText": "Your acceptance text here",
                      "dialogText": "You must set your dialog for this environment",
                      "enable": false,
                      "headerText": "Header text here"
                    },
                    "logoAltText": "Cloud Pak",
                    "logoHeight": "47px",
                    "logoUrl": "/common-nav/api/graphics/logincloudpak.svg",
                    "logoWidth": "190px"
                  },
                  "navItems": [
                    {
                      "id": "providers",
                      "label": "Identity providers",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/realms"
                    }
                  ]
                }
              },
              {
                "apiVersion": "foundation.ibm.com/v1",
                "kind": "NavConfiguration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "icp4i",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator",
                    "name": "icp4i"
                  },
                  "name": "icp4i"
                },
                "spec": {
                  "header": {
                    "disabledItems": [
                      "createResource",
                      "catalog",
                      "bookmark"
                    ],
                    "logoAltText": "Cloud Pak for Integration",
                    "logoUrl": "/common-nav/graphics/ibm-cloudpak-integration.svg"
                  },
                  "navItems": [
                    {
                      "detectionServiceName": true,
                      "id": "metering",
                      "label": "Metering",
                      "serviceId": "metering-ui",
                      "serviceName": "metering-ui",
                      "url": "/metering/dashboard?ace_config={ 'showClusterData': false }\u0026dashboard=cpi.icp.main"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "monitoring",
                      "isAuthorized": [
                        "Administrator",
                        "ClusterAdministrator",
                        "Operator"
                      ],
                      "label": "Monitoring",
                      "serviceId": "monitoring-ui",
                      "serviceName": "ibm-monitoring-grafana",
                      "target": "_blank",
                      "url": "/grafana"
                    },
                    {
                      "id": "id-access",
                      "label": "Identity and access",
                      "serviceId": "webui-nav"
                    },
                    {
                      "id": "providers",
                      "parentId": "id-access",
                      "label": "Identity providers",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/realms?useNav=icp4i"
                    },
                    {
                      "id": "teams-ids",
                      "parentId": "id-access",
                      "label": "Teams and service IDs",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/teams?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "logging",
                      "label": "Logging",
                      "serviceId": "kibana",
                      "serviceName": "kibana",
                      "target": "_blank",
                      "url": "/kibana"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "releases",
                      "label": "Helm Releases",
                      "serviceId": "catalog-ui",
                      "serviceName": "catalog-ui",
                      "url": "/catalog/instances?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "repos",
                      "label": "Helm Repositories",
                      "serviceId": "catalog-ui",
                      "serviceName": "catalog-ui",
                      "url": "/catalog/repositories?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "licensing",
                      "label": "Licensing",
                      "serviceId": "ibm-license-service-reporter",
                      "url": "/license-service-reporter",
                      "isAuthorized": [
                        "ClusterAdministrator",
                        "CloudPakAdministrator"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operators.ibm.com/v1alpha1",
                "kind": "SwitcherItem",
                "metadata": {
                  "name": "example-switcheritem"
                },
                "spec": {
                  "cloudPakInfo": {
                    "display": "Administration",
                    "label": "administration-hub",
                    "landingPage": "/common-nav/dashboard",
                    "logoURL": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
          description: The IBM Common Web UI delivers the common header API and the
            identity and access pages for IBM Cloud Platform Common Services.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.8.0
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-commonui-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            displayName: CommonWebUI service
            kind: CommonWebUI
            name: commonwebuis.operators.ibm.com
            version: v1alpha1
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            kind: NavConfiguration
            name: navconfigurations.foundation.ibm.com
            version: v1
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            kind: SwitcherItem
            name: switcheritems.operators.ibm.com
            version: v1alpha1
        description: "**Important:** Do not install this operator directly. Only install
          this operator using the IBM Common Services Operator. For more information
          about installing this operator and other Common Services operators, see
          [Installer documentation](http://ibm.biz/cpcs_opinstall). Additionally,
          you can exit this panel and navigate to the IBM Common Services tile in
          OperatorHub to learn more about the operator.\n\n If you are using this
          operator as part of an IBM Cloud Pak, see the documentation for that IBM
          Cloud Pak to learn more about how to install and use the operator service.
          For more information about IBM Cloud Paks, see [IBM Cloud Paks that use
          Common Services](http://ibm.biz/cpcs_cloudpaks).\n\n You can use the ibm-commonui-operator
          to install the Common Web UI service for the IBM Cloud Platform Common Services
          and access the Common Web UI console. You can use the Common Web UI console
          to access information and features from other IBM Cloud Platform Common
          Services or IBM Cloud Paks that you install. \n\nFor more information about
          the available IBM Cloud Platform Common Services, see the [IBM Knowledge
          Center](http://ibm.biz/cpcsdocs). \n## Supported platforms \n\n Red Hat
          OpenShift Container Platform 4.2 or newer installed on one of the following
          platforms: \n\n- Linux x86_64 \n- Linux on Power (ppc64le) \n- Linux on
          IBM Z and LinuxONE \n## Prerequisites\n\n The Common Web UI service has
          dependencies on other IBM Cloud Platform Common Services. Before you install
          this operator, you need to first install the operator dependencies and prerequisites:
          \n For the list of operator dependencies, see the IBM Knowledge Center [Common
          Services dependencies documentation](http://ibm.biz/cpcs_opdependencies).
          \n For the list of prerequisites for installing the operator, see the IBM
          Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).
          \n## Documentation \n\n To install the operator with the IBM Common Services
          Operator follow the the installation and configuration instructions within
          the IBM Knowledge Center. \n- If you are using the operator as part of an
          IBM Cloud Pak, see the documentation for that IBM Cloud Pak, for a list
          of IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).
          \n- If you are using the operator with an IBM Containerized Software, see
          the IBM Cloud Platform Common Services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall)."
        displayName: Ibm Common UI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        - Web Console
        - Common Services
        links:
        - name: GitHub
          url: https://github.com/IBM/ibm-commonui-operator
        maintainers:
        - email: siddarth.patel@ibm.com
          name: Siddarth Patel
        - email: sgrube@us.ibm.com
          name: Steve Grube
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:e4eb6a75e0d207232e479693d9c99ba863f8e62830f3191ca4d1e7ceaf3586d1
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator-bundle@sha256:e4340bd9481da998a6b89e94c884c46e17cfb50d9d0c1099f39a3c54e500d2b2
        version: 4.8.0
      entries:
      - name: ibm-commonui-operator.v4.8.0
        version: 4.8.0
      name: v4.7
    defaultChannel: v4.7
    packageName: ibm-commonui-operator-app
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent
      provider-url: ""
    name: cilium-enterprise
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cilium.v1.10.13-xed86eae
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"prometheus":{"serviceMonitor":{"enabled":false}},"mode":"cluster-pool"},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium Enterprise - eBPF-based Networking, Security, and Observability
        displayName: Cilium Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Enterprise Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Cilium Enterprise
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-ui-enterprise@sha256:c9921c65892a5293a4f1bef787636d6ae58cec1af65da06862a47e3450afdc6e
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/isovalent/hubble-relay@sha256:9145b03acc954efde54e1ab9da878462526c5e820b83a92e32cc24291f784d33
        - quay.io/isovalent/clustermesh-apiserver@sha256:eb3ee3d2cb07e6c5966e4bc2e50615949303be5d5128e156d5c0dbe8404dc093
        - quay.io/isovalent/hubble-enterprise-metadata@sha256:72c7aa3961f14df413569cd156820a45553bb24d393b11246a18bf8b6553e022
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        - quay.io/isovalent/cilium@sha256:6887537789603946543aaa370a3ec592287a235231bf16aa386e091823285515
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:cc1f0e6b27ccda8a843dfb7fa4a8fef052ba29716646557815677c5d1bea2a1a
        - quay.io/isovalent/hubble-export-fluentd@sha256:05cb13f533f2c89b98a47cd58badda07add54077cd22283ca5028cb39f2e4486
        - quay.io/isovalent/hubble-export-s3@sha256:ce4f71f7f13301133b7d471b87fb81108b2c2da7fc29865ef96bc8431b2be72b
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/oauth2-proxy/oauth2-proxy@sha256:ecd26b74a01f2b547ddaed4d32d35f8f5e09c378d5c1fc6cfa63f0adf659ac2b
        - quay.io/isovalent/hubble-enterprise-operator@sha256:0d5e6a5d9c07780c18e5aa81d8f1c95c399bff2d36a44fa5f61ac89a788ca83b
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/hubble-ui-enterprise-backend@sha256:e892cba6dafcdf6704df3135a0b38d11894a5f276ace40683e74cb62b9c311d8
        - quay.io/isovalent/operator-generic@sha256:1cc0b0d2e71564ad8759be326c66ea13b4955b9441fb67e906da534e1b38aefa
        - quay.io/cilium/hubble-export-stdout@sha256:f1b1475a3c89950481264f5f8ca63eea3b911a45dc83fdf15122d695a0fba9b3
        - quay.io/isovalent/hubble-enterprise@sha256:1e1d3e6c199546b2398f4166ff64210c04075bb407c8f07c5be8d3db08560d4a
        - quay.io/isovalent/hubble-rbac@sha256:03f3304c157024356a6de13c3d91b4506b8e54d587920d80ebcc76cbcfeaeda6
        version: 1.10.13+xed86eae
      entries:
      - name: cilium.v1.10.13-xed86eae
        version: 1.10.13+xed86eae
      - name: cilium.v1.10.12-xb203fae
        version: 1.10.12+xb203fae
      - name: cilium.v1.10.11-x10ae342
        version: 1.10.11+x10ae342
      name: "1.10"
    - currentCSV: cilium.v1.11.21-xd0d07b1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium Enterprise - eBPF-based Networking, Security, and Observability
        displayName: Cilium Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Enterprise Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Cilium Enterprise
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - cgr.dev/chainguard/oauth2-proxy:7.4.0@sha256:54857c70321004d3e0b8b8d7208ce5422c1f71c18db5be02d2543f187ea8e8e2
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:372511f6fa3523dadfe9128caf77520cf74ce24df4a19e1583ee197375e06014
        - quay.io/isovalent/hubble-ui-enterprise:v0.23.0@sha256:02635d1392cca7e09ab7797f69147880068ba7e3b0d865be97eeab33edd646a0
        - quay.io/cilium/certgen:v0.1.5@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/startup-script:62093c5c233ea914bfa26a10ba41f8780d9b737f@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/isovalent/clustermesh-apiserver:v1.11.21-cee.1@sha256:e16d8375ed8a1c3a872047b3aacf3469bde43552f1a51ee3e159ef2e9fea3e97
        - quay.io/coreos/etcd:v3.4.13@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/isovalent/hubble-relay:v1.11.21-cee.1@sha256:5933dd1f8111d6bca0c68db9247385b5d215d5b613095bafab8244098f2c5a26
        - quay.io/isovalent/hubble-ui-enterprise-backend:v0.23.0@sha256:2fbdad86bb5244075f31487262d790f6fa75cdd993f868b952252905c118621c
        - quay.io/isovalent/hubble-rbac:v1.3.0@sha256:4c106dd133861a678eb24cf98b35d5023cccca0834c954c49b68514c366aedf0
        - quay.io/isovalent/hubble-enterprise-operator:v1.10.5@sha256:31b18a1ea7002936f04dd070549e0c9a02f8a33394666fc5b0523ab1350c6120
        - quay.io/isovalent/hubble-export-stdout:v1.0.3@sha256:bef600089c1671fce11bb4bcb98b81f25e65152a1a2f13b64ea637cfc5779c11
        - quay.io/isovalent/hubble-export-s3:v1.0.0@sha256:89c178c0a010166c68819ab5de32ab59b64a88aa23a05d8e4f78adb1a4b6c3d3
        - quay.io/isovalent/hubble-export-fluentd:v1.5.1@sha256:eaf42a84821509b69e2ef2b658796062e79a90e63045a98689da3da99a12e492
        - quay.io/isovalent/hubble-enterprise:v1.10.5@sha256:340700d4e28ded2aa9c50ff9f938996bb43faa84e0b2ba38ec79452648d7e5d7
        - quay.io/isovalent/operator-generic:v1.11.21-cee.1@sha256:25b668c1febe1d1a70b63dfe926be4e55fbf77fd186e9df7581e41d7e9813469
        - quay.io/isovalent/cilium:v1.11.21-cee.1@sha256:f41241a15cd96d26727aa1ec413ace643728ddbe42f7fb7b234dbda22bf3950b
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:d3148bf3b9765467936ee067b25b5ef3683d3e55260f517c7fe271ac0e19cb47
        version: 1.11.21+xd0d07b1
      entries:
      - name: cilium.v1.11.21-xd0d07b1
        version: 1.11.21+xd0d07b1
      - name: cilium.v1.11.20-x0f6d966
        version: 1.11.20+x0f6d966
      name: "1.11"
    - currentCSV: cilium.v1.12.21-x3861686
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.12.14 <1.12.21'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/cilium:v1.12.21-cee.1@sha256:3a462308cf12b572db987c03d190ee33691ccb04775d01beb192703a52f8d47b
        - quay.io/isovalent/hubble-relay:v1.12.21-cee.1@sha256:9d8dadc3f330acf0fa169c428d5fdf1d138d0294065eec5c628f6647106fe8b6
        - quay.io/coreos/etcd:v3.5.4@sha256:795d8660c48c439a7c3764c2330ed9222ab5db5bb524d8d0607cac76f7ba82a3
        - quay.io/isovalent/hubble-export-s3:v1.0.0@sha256:89c178c0a010166c68819ab5de32ab59b64a88aa23a05d8e4f78adb1a4b6c3d3
        - quay.io/isovalent/hubble-enterprise-operator:v1.10.5@sha256:31b18a1ea7002936f04dd070549e0c9a02f8a33394666fc5b0523ab1350c6120
        - quay.io/isovalent/clustermesh-apiserver:v1.12.21-cee.1@sha256:3f6549409966086a49e338c7829b9141a73ef2ebeebaac91db430aeb3ccd2e22
        - quay.io/isovalent/hubble-ui-enterprise:v0.28.1@sha256:11dfba0e4fcf43415aa6e8b3b6e98a5c4f468ec67b377328ad48b92037bbfbb0
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:f95c9e8002d8e381d604752b1d59b75595badbdb0b11237ac1cac4ecd04f4800
        - quay.io/isovalent/hubble-ui-enterprise-backend:v0.28.1@sha256:515aeb8166dd6995b04d4cd3333c94a740182943ba4e90a38871ee89c0b64e92
        - quay.io/isovalent/hubble-export-fluentd:v1.5.1@sha256:eaf42a84821509b69e2ef2b658796062e79a90e63045a98689da3da99a12e492
        - quay.io/cilium/certgen:v0.1.8@sha256:4a456552a5f192992a6edcec2febb1c54870d665173a33dc7d876129b199ddbd
        - quay.io/isovalent/hubble-enterprise:v1.10.5@sha256:340700d4e28ded2aa9c50ff9f938996bb43faa84e0b2ba38ec79452648d7e5d7
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/isovalent/hubble-rbac:v1.3.0@sha256:4c106dd133861a678eb24cf98b35d5023cccca0834c954c49b68514c366aedf0
        - quay.io/isovalent/operator-generic:v1.12.21-cee.1@sha256:4e19cda799a747b69f53b5e5a860ba256399d18221d5a4bbf243142ba1c69bb8
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/hubble-export-stdout:v1.0.3@sha256:bef600089c1671fce11bb4bcb98b81f25e65152a1a2f13b64ea637cfc5779c11
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:b6e40d76f0925e76f0a5b181be3f6c62728fa3e4aaef979caf9ea245f4ed03da
        - quay.io/cilium/startup-script:62093c5c233ea914bfa26a10ba41f8780d9b737f@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        version: 1.12.21+x3861686
      entries:
      - name: cilium.v1.12.21-x3861686
        version: 1.12.21+x3861686
      - name: cilium.v1.12.19-xca684e0
        version: 1.12.19+xca684e0
      - name: cilium.v1.12.18-xcb1adcf
        version: 1.12.18+xcb1adcf
      - name: cilium.v1.12.17-xae21c14
        version: 1.12.17+xae21c14
      - name: cilium.v1.12.16-x38e0b62
        version: 1.12.16+x38e0b62
      - name: cilium.v1.12.15-x8ca64a3
        version: 1.12.15+x8ca64a3
      - name: cilium.v1.12.14-xc86b260
        version: 1.12.14+xc86b260
      - name: cilium.v1.12.13-x2b5cec9
        version: 1.12.13+x2b5cec9
      - name: cilium.v1.12.12-xbd7b835
        version: 1.12.12+xbd7b835
      - name: cilium.v1.12.11-xffdea76
        version: 1.12.11+xffdea76
      - name: cilium.v1.12.10-x4a0e9e7
        version: 1.12.10+x4a0e9e7
      - name: cilium.v1.12.9-x4be66e4
        version: 1.12.9+x4be66e4
      - name: cilium.v1.12.8-x87b90a9
        version: 1.12.8+x87b90a9
      - name: cilium.v1.12.7-x53a2aab
        version: 1.12.7+x53a2aab
      name: "1.12"
    - currentCSV: cilium-enterprise.v1.13.21-x96449fb
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.7 <1.13.21+x96449fb'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/startup-script:c54c7edeab7fde4da68e59acd319ab24af242c3f@sha256:a9d0f451353f1b2f29ed3868e230ecfae10150eab9fd521de67adb3c8ca9c33a
        - quay.io/cilium/certgen:v0.1.13@sha256:01802e6a153a9473b06ebade7ee5730f8f2c6cc8db8768508161da3cdd778641
        - quay.io/isovalent/hubble-export-fluentd:v1.11.0@sha256:50cd2e4c2d04e8befcc2ae1a1463f07cba01edb9b546fb455fe7976311165385
        - quay.io/isovalent/hubble-enterprise:v1.11.6@sha256:eb60d59d1fb1b136855829274b31805c4206f8d050c3bf068e63a552576c1b38
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:b04e75456359b50b18e75464da5d7b467fdfeb058fc08ce3a10f057ba5f60107
        - quay.io/isovalent/cilium:v1.13.21-cee.2@sha256:f760cd85fa64eb8a197ec282644a8db630745a8a2854ee08aa2662cf67188db5
        - quay.io/isovalent/hubble-export-s3:v1.0.2@sha256:518daea1074d260d824c3e03a93ab6b2a7a5f4c22b8e07938770add8257f4a52
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:fc7046dee1a204163e68a6d9a42fec200b02129493a7d166f899dd12e2763364
        - quay.io/isovalent/hubble-rbac:v1.5.6@sha256:739ca45ea09829b0fa8a6198901b34eb5254df55be071bc854bbd37cae12a557
        - quay.io/isovalent/hubble-relay:v1.13.21-cee.2@sha256:aaf5f959a76d6b517e627d2d0b09dab86ae942616819a9a95f057ed4f45ea480
        - quay.io/isovalent/operator-generic:v1.13.21-cee.2@sha256:6f2d0d575e3f74395c959833992cd5fbb6971f4e292df1b7cf45be8b07be5e7a
        - quay.io/isovalent/clustermesh-apiserver:v1.13.21-cee.2@sha256:980b0235fd16817ecc4457ad4317a4601993b76df2e81b07a53f6f437c2705af
        - quay.io/isovalent/hubble-export-stdout:v1.0.4@sha256:a0ff77efa890478f18f5deeba2cd1e2dfd6c5c77981adac57b14d7734de71ff1
        - quay.io/isovalent/hubble-ui-enterprise:v1.0.3@sha256:55bf787c96561f4a460e2bfe5966421e2d85f3f9a38ac2487c6f6486c97cd5af
        - quay.io/isovalent/hubble-ui-enterprise-backend:v1.0.3@sha256:6756d2b808c38efce4765f90641d26aa2e0e79329ecc378a123cd3dc2c439640
        - quay.io/isovalent/hubble-enterprise-operator:v1.11.6@sha256:38253c2e17a8b2e88a226f0976c9cf4a00445212e2886e6f07cbb3a52757324f
        - quay.io/coreos/etcd:v3.5.4@sha256:795d8660c48c439a7c3764c2330ed9222ab5db5bb524d8d0607cac76f7ba82a3
        version: 1.13.21+x96449fb
      entries:
      - name: cilium-enterprise.v1.13.21-x96449fb
        version: 1.13.21+x96449fb
      - name: cilium-enterprise.v1.13.20-x5b70650
        version: 1.13.20+x5b70650
      - name: cilium-enterprise.v1.13.19-xdb27142
        version: 1.13.19+xdb27142
      - name: cilium-enterprise.v1.13.18-x0c21a94
        version: 1.13.18+x0c21a94
      - name: cilium-enterprise.v1.13.17-xc8f905f
        version: 1.13.17+xc8f905f
      - name: cilium-enterprise.v1.13.16-xcfe62ed
        version: 1.13.16+xcfe62ed
      - name: cilium-enterprise.v1.13.15-x1dfd26f
        version: 1.13.15+x1dfd26f
      - name: cilium.v1.13.14-xb70933f
        version: 1.13.14+xb70933f
      - name: cilium.v1.13.12-x453f6ad
        version: 1.13.12+x453f6ad
      - name: cilium.v1.13.11-xf6d6935
        version: 1.13.11+xf6d6935
      - name: cilium.v1.13.10-x9d57fc7
        version: 1.13.10+x9d57fc7
      - name: cilium.v1.13.9-xe8a359b
        version: 1.13.9+xe8a359b
      - name: cilium.v1.13.8-x96e1ca9
        version: 1.13.8+x96e1ca9
      - name: cilium.v1.13.7-x7320e7d
        version: 1.13.7+x7320e7d
      - name: cilium.v1.13.6-xd3714ef
        version: 1.13.6+xd3714ef
      - name: cilium.v1.13.5-x8386cae
        version: 1.13.5+x8386cae
      - name: cilium.v1.13.4-xc8b22ea
        version: 1.13.4+xc8b22ea
      - name: cilium.v1.13.3-x69e9561
        version: 1.13.3+x69e9561
      - name: cilium.v1.13.2-x4b1c6c8
        version: 1.13.2+x4b1c6c8
      name: "1.13"
    - currentCSV: cilium-enterprise.v1.14.22-x6d86c88
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.14.22+x6d86c88'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-enterprise:v1.11.6@sha256:eb60d59d1fb1b136855829274b31805c4206f8d050c3bf068e63a552576c1b38
        - quay.io/isovalent/hubble-rbac:v1.5.2@sha256:29309837e2246a1b6849d7a09d01ceebfdb43fa0ade973d14584f0e52b856b95
        - quay.io/coreos/etcd:v3.5.4@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:148b1029eac14049e2ba84d4d75be1ca43ef2d89951adbdeddec392ad7c9fe1e
        - quay.io/isovalent/hubble-ui-enterprise:v1.2.2@sha256:3e03a58f6eea5d2df195f42ee2817f7b28fa20bbe140f5086937eee8da1b8d56
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/isovalent/hubble-enterprise-operator:v1.11.6@sha256:38253c2e17a8b2e88a226f0976c9cf4a00445212e2886e6f07cbb3a52757324f
        - quay.io/cilium/startup-script:c54c7edeab7fde4da68e59acd319ab24af242c3f@sha256:a9d0f451353f1b2f29ed3868e230ecfae10150eab9fd521de67adb3c8ca9c33a
        - quay.io/isovalent/operator-generic:v1.14.22-cee.1@sha256:a997fe24ff0b7f7df3ee12748413ae53d6e07c803162b0db4d812b916b0cfc74
        - quay.io/isovalent/hubble-export-fluentd:v1.8.1@sha256:ce0cce63a8473d5ae73872fc65d82c54f2875511bb03f0bc8ee8480148c43ee1
        - quay.io/cilium/certgen:v0.1.17@sha256:07959f3125dae659ab9036c14df046e2536f7a4f6c92dabda837388b0344dc74
        - quay.io/isovalent/hubble-relay:v1.14.22-cee.1@sha256:d7ffc880d2e42e0876417ec26115b0e9155ad9a2e7f47c02c434746afa42b29a
        - quay.io/isovalent/clustermesh-apiserver:v1.14.22-cee.1@sha256:588e2b173ab409288eda2d404571184562d3d9fd12e2d9afe7282a5a7d79113a
        - quay.io/cilium/cilium-envoy:v1.32.5-1743506100-0821ef0acdf9f824d47d34e02932be522b3e7233@sha256:305c8b7da56ef1b64e4deb0642acf93a3944f2712e40c99a40bb005e6dc9761b
        - quay.io/isovalent/hubble-ui-enterprise-backend:v1.2.2@sha256:4258f7e0bff24ff1e0b214f62b37097032b6426ba28edd3dfc4f451a34de0110
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:fc7046dee1a204163e68a6d9a42fec200b02129493a7d166f899dd12e2763364
        - quay.io/isovalent/hubble-export-stdout:v1.0.4@sha256:a0ff77efa890478f18f5deeba2cd1e2dfd6c5c77981adac57b14d7734de71ff1
        - quay.io/isovalent/cilium:v1.14.22-cee.1@sha256:0e664a5a9e0edba38ab1a25eb7fa50861bb15f69ed0bbe3f3c792bd909240da3
        - quay.io/isovalent/hubble-export-s3:v1.0.2@sha256:518daea1074d260d824c3e03a93ab6b2a7a5f4c22b8e07938770add8257f4a52
        version: 1.14.22+x6d86c88
      entries:
      - name: cilium-enterprise.v1.14.22-x6d86c88
        version: 1.14.22+x6d86c88
      - name: cilium-enterprise.v1.14.21-xe0c9553
        version: 1.14.21+xe0c9553
      - name: cilium-enterprise.v1.14.20-xe01e743
        version: 1.14.20+xe01e743
      - name: cilium-enterprise.v1.14.19-xc1b9aff
        version: 1.14.19+xc1b9aff
      name: "1.14"
    - currentCSV: cilium-enterprise.v1.15.21-x2661a01
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.15.21+x2661a01'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:bd1a4d47ef48a6eb84c49d979e8f5ce867af5932828dfd3233fa9f78c2eac1f4
        - quay.io/isovalent/cilium-envoy-ubi:v1.33.4-1752151664-7c2edb0b44cf95f326d628b837fcdd845102ba68@sha256:e81fdde48c3d5cd93111e1d51d9f8554adb32ee937601a9124354e0894933b99
        - quay.io/isovalent/operator-generic-ubi:v1.15.21-cee.1@sha256:97daac94735e497e1e897beaa01eb50866abace1aa9a3919fc57971e45f538fe
        - quay.io/isovalent/hubble-ui-frontend-ubi:v1.3.7@sha256:f2b4e935abe3eb94f82dab41c9d3d929fca71d82fa39a6c8002683375b5303a4
        - quay.io/isovalent/oauth2-proxy:7.12.0-20250819@sha256:fe8f76728c36536d498b2a5c11172cb18dc30bbb555df2a6d14c036b29cb9f10
        - quay.io/isovalent/cilium:v1.15.21-cee.1@sha256:39ef22d467783481eccb376fcb5d6bc254ed8f4b3b25032121498db40278f0b8
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.15.21-cee.1@sha256:36d00eb9f8ed45a3d8892942cfae097782dd6509889ef7d1a451f9b0334d17e1
        - quay.io/isovalent/certgen-ubi:v0.1.19@sha256:c32e96b0972c9bcd3896945c46ea23e330c9dea407afee3a637daae637296f4c
        - quay.io/isovalent/hubble-ui-backend-ubi:v1.3.7@sha256:a4b5e7a032c1bbcd969a6873bc12935f8b6aa4c2d3a2ed87f5494fcfcdfd02d4
        - quay.io/isovalent/hubble-relay-ubi:v1.15.21-cee.1@sha256:8461f720ca95a9c543a298bee1c1e56241f56d5426030c13afed8262f2f5c5ff
        - quay.io/isovalent/startup-script-ubi:1755531540-60ee83e@sha256:2e1d5048cbfcd3c77f732da8203a6c792cbf19897b5cf137b9ad23a6a267f6b2
        version: 1.15.21+x2661a01
      entries:
      - name: cilium-enterprise.v1.15.21-x2661a01
        version: 1.15.21+x2661a01
      - name: cilium-enterprise.v1.15.20-x6df37bb
        version: 1.15.20+x6df37bb
      - name: cilium-enterprise.v1.15.19-x0be49ab
        version: 1.15.19+x0be49ab
      - name: cilium-enterprise.v1.15.18-xeeaa70b
        version: 1.15.18+xeeaa70b
      - name: cilium-enterprise.v1.15.17-xdfd9482
        version: 1.15.17+xdfd9482
      - name: cilium-enterprise.v1.15.16-xc7e31ca
        version: 1.15.16+xc7e31ca
      - name: cilium-enterprise.v1.15.15-xf37f22c
        version: 1.15.15+xf37f22c
      - name: cilium-enterprise.v1.15.14-x0e512e3
        version: 1.15.14+x0e512e3
      - name: cilium-enterprise.v1.15.13-x134fa29
        version: 1.15.13+x134fa29
      name: "1.15"
    - currentCSV: cilium-enterprise.v1.16.15-x6419e26
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.16.0 <1.16.15+x6419e26'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/startup-script-ubi:v1.16.15-cee.1@sha256:5d59a46be44c4d899576e69e0f20e148245e17d905a0769a73fd01f2432d1bab
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:ec8a17956e75e91b60e2999d10e67ef76f9b4cfd754db415aaf387d21690da87
        - quay.io/isovalent/cilium-envoy-ubi:v1.16.15-cee.1@sha256:843e6fdf6845f3d5f841dbd2dce75224b16c9d34c2b0ce7f329d9666466bfc6b
        - quay.io/isovalent/cilium-ubi:v1.16.15-cee.1@sha256:1d7bbbcb613ef20cd1719bd77473e71c198960ee5d8e563aec9c65cfd3c2efe2
        - quay.io/isovalent/hubble-relay-ubi:v1.16.15-cee.1@sha256:c992d549627e8f9180c272e7d96ff3783de184796633ed5d0b6fe2cf7e74055d
        - quay.io/isovalent/operator-generic-ubi:v1.16.15-cee.1@sha256:970aa156b6f5c390b6835a626190e18656f02db6a5cf27aaa1ac9c7872f4b15a
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.16.15-cee.1@sha256:f7573ab4145d306051b5d6fa7ff1af395fc71935f5204839ab1c9f574aee17ad
        - quay.io/isovalent/certgen-ubi:v1.16.15-cee.1@sha256:e3f1fa5ee59887d3a15c43947a853c8a3f4fc650ed72bcd891b277037ea16a9d
        version: 1.16.15+x6419e26
      entries:
      - name: cilium-enterprise.v1.16.15-x6419e26
        version: 1.16.15+x6419e26
      - name: cilium-enterprise.v1.16.13-xb2de075
        version: 1.16.13+xb2de075
      - name: cilium-enterprise.v1.16.12-xd373871
        version: 1.16.12+xd373871
      - name: cilium-enterprise.v1.16.11-xbd1cd23
        version: 1.16.11+xbd1cd23
      - name: cilium-enterprise.v1.16.10-x89ab4b0
        version: 1.16.10+x89ab4b0
      - name: cilium-enterprise.v1.16.9-x8ec4137
        version: 1.16.9+x8ec4137
      - name: cilium-enterprise.v1.16.8-yd2b50b4
        version: 1.16.8+yd2b50b4
      - name: cilium-enterprise.v1.16.7-x4867421
        version: 1.16.7+x4867421
      - name: cilium-enterprise.v1.16.6-x7e6eb1c
        version: 1.16.6+x7e6eb1c
      name: "1.16"
    defaultChannel: "1.16"
    packageName: cilium-enterprise
    provider:
      name: Isovalent
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-commons-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: commons-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:6768e4189fa8deebf337e4732e5f12bb2ad4acdc112d9d05f5dc1a4ed81f08de
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for S3ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for S3BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/commons-operator@sha256:6768e4189fa8deebf337e4732e5f12bb2ad4acdc112d9d05f5dc1a4ed81f08de
        version: 24.11.1
      entries:
      - name: commons-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: commons-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:adc2b25a676f7357ca3971c13ecf5c9ed3cb1af24ffc902b03733cc814e8946f
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for S3ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for S3BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/commons-operator@sha256:adc2b25a676f7357ca3971c13ecf5c9ed3cb1af24ffc902b03733cc814e8946f
        version: 25.3.0
      entries:
      - name: commons-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: commons-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
        version: 25.7.0
      entries:
      - name: commons-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: commons-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
        version: 25.7.0
      entries:
      - name: commons-operator.v25.7.0
        version: 25.7.0
      - name: commons-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-commons-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: heat-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: heat-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "heat.openstack.org/v1beta1",
                "kind": "Heat",
                "metadata": {
                  "name": "heat"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = True\n",
                  "databaseInstance": "openstack",
                  "heatAPI": {},
                  "heatCfnAPI": {},
                  "heatEngine": {},
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "heat.openstack.org/v1beta1",
                "kind": "HeatAPI",
                "metadata": {
                  "name": "heat-api"
                },
                "spec": {
                  "databaseAccount": "heat",
                  "databaseHostname": "openstack",
                  "passwordSelectors": {
                    "authEncryptionKey": "HeatAuthEncryptionKey",
                    "service": "HeatPassword"
                  },
                  "replicas": 1,
                  "resources": {},
                  "secret": "osp-secret",
                  "serviceAccount": "heat-heat",
                  "serviceUser": "heat",
                  "transportURLSecret": "rabbitmq-transport-url-heat-heat-transport"
                }
              },
              {
                "apiVersion": "heat.openstack.org/v1beta1",
                "kind": "HeatCfnAPI",
                "metadata": {
                  "name": "heat-cfnapi"
                },
                "spec": {
                  "databaseAccount": "heat",
                  "databaseHostname": "openstack",
                  "passwordSelectors": {
                    "authEncryptionKey": "HeatAuthEncryptionKey",
                    "service": "HeatPassword"
                  },
                  "replicas": 1,
                  "resources": {},
                  "secret": "osp-secret",
                  "serviceAccount": "heat-heat",
                  "serviceUser": "heat",
                  "transportURLSecret": "rabbitmq-transport-url-heat-heat-transport"
                }
              },
              {
                "apiVersion": "heat.openstack.org/v1beta1",
                "kind": "HeatEngine",
                "metadata": {
                  "name": "heat-engine"
                },
                "spec": {
                  "databaseAccount": "heat",
                  "databaseHostname": "openstack",
                  "passwordSelectors": {
                    "authEncryptionKey": "HeatAuthEncryptionKey",
                    "service": "HeatPassword"
                  },
                  "replicas": 1,
                  "resources": {},
                  "secret": "osp-secret",
                  "serviceAccount": "heat-heat",
                  "serviceUser": "heat",
                  "transportURLSecret": "rabbitmq-transport-url-heat-heat-transport"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:16Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HeatAPI is the Schema for the heatapis API
            displayName: Heat API
            kind: HeatAPI
            name: heatapis.heat.openstack.org
            version: v1beta1
          - description: HeatCfnAPI is the Schema for the heatcfnapis API
            displayName: Heat Cfn API
            kind: HeatCfnAPI
            name: heatcfnapis.heat.openstack.org
            version: v1beta1
          - description: HeatEngine is the Schema for the heatengines API
            displayName: Heat Engine
            kind: HeatEngine
            name: heatengines.heat.openstack.org
            version: v1beta1
          - description: Heat is the Schema for the heats API
            displayName: Heat
            kind: Heat
            name: heats.heat.openstack.org
            version: v1beta1
        description: Heat Operator
        displayName: Heat Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Orchestration
        - Heat
        links:
        - name: Heat Operator
          url: https://github.com/openstack-k8s-operators/heat-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/heat-rhel9-operator@sha256:fa61aeee8d63087d150a7641d1e8d4b67dd067bcfed6a140401fe396f1e32403
        - registry.redhat.io/rhoso/openstack-heat-api-rhel9@sha256:dba236e5fed89fbe46397b6ab01de5795e66e9194f5cd99fdf137c37a7b96135
        - registry.redhat.io/rhoso/openstack-heat-api-cfn-rhel9@sha256:95548184e68325a5fe1604d5c4d35d1e9cca7f39b84303fce50479e0e35f8280
        - registry.redhat.io/rhoso/openstack-heat-engine-rhel9@sha256:0b6fc782986bf11b26c0839fd81c52257f4a31aa5e5a6728ce251f3670537aba
        version: 1.0.6
      entries:
      - name: heat-operator.v1.0.6
        version: 1.0.6
      - name: heat-operator.v1.0.5
        version: 1.0.5
      - name: heat-operator.v1.0.4
        version: 1.0.4
      - name: heat-operator.v1.0.3
        version: 1.0.3
      - name: heat-operator.v1.0.2
        version: 1.0.2
      - name: heat-operator.v1.0.1
        version: 1.0.1
      - name: heat-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: heat-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: servicemeshoperator3
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: servicemeshoperator3.v3.0.0-tp.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh-tech-preview/istio-rhel9-operator@sha256:f5dfe495ba92ecb880485b66cfafbf6451f8142b8e0d5fd7f29db94171c51336
          createdAt: "2024-12-12T10:28:38Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1alpha1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1alpha1
          - description: IstioRevisionTag references a Istio or IstioRevision object
              and serves as an alias for sidecar injection.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1alpha1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-rhel9-operator@sha256:f5dfe495ba92ecb880485b66cfafbf6451f8142b8e0d5fd7f29db94171c51336
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-pilot-rhel9@sha256:376f64eab0ebeb06035983a60d1e7647c5d41354f31a49cb34f4deb7b6708e3e
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-pilot-rhel9@sha256:ea9372fabe06cd8e65ea7b4567ce3c8fbddd9d673dc34a12743fd95c548059d1
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-proxyv2-rhel9@sha256:c21481d6dae58b82dac93b1b47b9d9f53a30617fa9d930a0b5d075620cffa74b
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-must-gather-rhel9@sha256:2bf3704ba8d5945973ccf6bd13418457e929836352acf0ea6a47976ffa246616
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-proxyv2-rhel9@sha256:cbd6bc47341ab8618609f89d063552e17232541c727429855c01deab73e4d06c
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-must-gather-rhel9@sha256:e3978aecaafe2678355e21a7303f8620f53d8b774f6592f10527c15bae957224
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-cni-rhel9@sha256:d0ab36246c556249b6d67346eb0c72ebd22483a6502caa0fc75112a8d327b053
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-cni-rhel9@sha256:f16cd8a29ddcbb3de77cbfcf7e1da78e42cc0c1c502132d477a01c52689cdf23
        version: 3.0.0-tp.2
      entries:
      - name: servicemeshoperator3.v3.0.0-tp.2
        version: 3.0.0-tp.2
      - name: servicemeshoperator3.v3.0.0-tp.1
        version: 3.0.0-tp.1
      name: candidates
    - currentCSV: servicemeshoperator3.v3.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.4"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.4"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
          createdAt: "2025-09-10T06:22:42Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:366266f10e658c4cea86bddf6ee847e1908ea4dd780cb5f7fb0e466bac8995f1
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:0d6930075b8906bee0ab1aeac82f6e2431c16163cf87f7f0c68c10b2c40552a1
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:26747627ae22bbdffcf9de58077454fc0c890cda83659d7834b15dea2b5aaaf2
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2ea112ab90b8540f11e9949d77c3e7e3b3ef57ac3bf23f6cf1e883a88430e1f9
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:028e10651db0d1ddb769a27c9483c6d41be6ac597f253afd9d599f395d9c82d8
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:ecc6a22a471f4f4b836648f3a72ea7f57a4ea960ebcbc7dbe3fe729500902d0b
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d518f3d1539f45e1253c5c9fa22062802804601d4998cd50344e476a3cc388fe
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:33feac0c9c888b375e207e86b3f9a1a289126ca46b64632a7f693e2191cfbbfd
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:2ca96716812d339775d3dfd396c863834dd3810720e98bfcf87bc67c5fbd29b5
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:6bb44c84363a4113c5bce2ae0d6528b345fddd12bc559bbfaacf6222060ddcb0
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:6881fb9f5a4b447be4c468b8585fff86171bd3d82671a48559f3ceabcb9dad24
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:14c5a52faf20267baa43d9a72ee6416f56fbc41dd640adc2aba3c4043802a0e9
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:b757b81d34711cb8abfcb0e8896c023d39d4778f237dd80b637ba4bf613df893
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:f7ab868455b5e887e2d89b7078678531290365f9166cae99ce56119f31fa2ba4
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d757e700a124d7583b9456ec9e08e5e7e249c1de202bc4e973102922b9f69191
        version: 3.1.2
      entries:
      - name: servicemeshoperator3.v3.1.2
        version: 3.1.2
      - name: servicemeshoperator3.v3.1.1
        version: 3.1.1
      - name: servicemeshoperator3.v3.1.0
        version: 3.1.0
      - name: servicemeshoperator3.v3.0.5
        version: 3.0.5
      - name: servicemeshoperator3.v3.0.4
        version: 3.0.4
      - name: servicemeshoperator3.v3.0.3
        version: 3.0.3
      - name: servicemeshoperator3.v3.0.2
        version: 3.0.2
      - name: servicemeshoperator3.v3.0.1
        version: 3.0.1
      - name: servicemeshoperator3.v3.0.0
        version: 3.0.0
      name: stable
    - currentCSV: servicemeshoperator3.v3.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.6"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.6"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:d716f3558fbdf8514f225de708135ba87fdcac4c718ef101f394daf546514afb
          createdAt: "2025-09-03T07:53:34Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:d716f3558fbdf8514f225de708135ba87fdcac4c718ef101f394daf546514afb
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        version: 3.0.5
      entries:
      - name: servicemeshoperator3.v3.0.5
        version: 3.0.5
      - name: servicemeshoperator3.v3.0.4
        version: 3.0.4
      - name: servicemeshoperator3.v3.0.3
        version: 3.0.3
      - name: servicemeshoperator3.v3.0.2
        version: 3.0.2
      - name: servicemeshoperator3.v3.0.1
        version: 3.0.1
      - name: servicemeshoperator3.v3.0.0
        version: 3.0.0
      name: stable-3.0
    - currentCSV: servicemeshoperator3.v3.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.4"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.4"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
          createdAt: "2025-09-10T06:22:42Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:028e10651db0d1ddb769a27c9483c6d41be6ac597f253afd9d599f395d9c82d8
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d518f3d1539f45e1253c5c9fa22062802804601d4998cd50344e476a3cc388fe
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:366266f10e658c4cea86bddf6ee847e1908ea4dd780cb5f7fb0e466bac8995f1
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:33feac0c9c888b375e207e86b3f9a1a289126ca46b64632a7f693e2191cfbbfd
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:ecc6a22a471f4f4b836648f3a72ea7f57a4ea960ebcbc7dbe3fe729500902d0b
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:b757b81d34711cb8abfcb0e8896c023d39d4778f237dd80b637ba4bf613df893
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:14c5a52faf20267baa43d9a72ee6416f56fbc41dd640adc2aba3c4043802a0e9
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:6bb44c84363a4113c5bce2ae0d6528b345fddd12bc559bbfaacf6222060ddcb0
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:0d6930075b8906bee0ab1aeac82f6e2431c16163cf87f7f0c68c10b2c40552a1
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:6881fb9f5a4b447be4c468b8585fff86171bd3d82671a48559f3ceabcb9dad24
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d757e700a124d7583b9456ec9e08e5e7e249c1de202bc4e973102922b9f69191
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2ea112ab90b8540f11e9949d77c3e7e3b3ef57ac3bf23f6cf1e883a88430e1f9
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:26747627ae22bbdffcf9de58077454fc0c890cda83659d7834b15dea2b5aaaf2
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:f7ab868455b5e887e2d89b7078678531290365f9166cae99ce56119f31fa2ba4
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:2ca96716812d339775d3dfd396c863834dd3810720e98bfcf87bc67c5fbd29b5
        version: 3.1.2
      entries:
      - name: servicemeshoperator3.v3.1.2
        version: 3.1.2
      - name: servicemeshoperator3.v3.1.1
        version: 3.1.1
      - name: servicemeshoperator3.v3.1.0
        version: 3.1.0
      name: stable-3.1
    defaultChannel: stable
    packageName: servicemeshoperator3
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: must-gather-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: must-gather-operator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "MustGather",
                "metadata": {
                  "name": "example-mustgather"
                },
                "spec": {
                  "caseID": "02527285",
                  "caseManagementAccountSecretRef": {
                    "name": "case-management-creds"
                  },
                  "serviceAccountRef": {
                    "name": "must-gather-admin"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security, Monitoring
          certified: "false"
          containerImage: quay.io/redhat-cop/must-gather-operator@sha256:aebef122cdb8409f5a5d70d3cdc585e02272704d57e103553a84bc6fc238e646
          createdAt: "2021-12-02T17:13:53Z"
          description: This operator provides a facility to easily upload must-gather
            reports to a Red Hat case.
          operatorframework.io/suggested-namespace: global-load-balancer-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/must-gather-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: MustGather
            name: mustgathers.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The must gather operator helps collecting must gather information on a cluster and uploading it to a case.
          To use the operator a cluster administrator can create the following must gather CR:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: MustGather
          metadata:
            name: example-mustgather
          spec:
            caseID: '02527285'
            caseManagementAccountSecretRef:
              name: case-management-creds
          ```

          this request will collect the standard must gather info and upload it to case `#02527285` using the credentials found in the `caseManagementCreds` secret.

          A more complex example:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: MustGather
          metadata:
            name: full-mustgather
          spec:
            caseID: '02527285'
            caseManagementAccountSecretRef:
              name: case-management-creds
            serviceAccountRef:
              name: must-gather-admin
            mustGatherImages:
            - quay.io/kubevirt/must-gather:latest
            - quay.io/ocs-dev/ocs-must-gather
          ```

          in this example we are using a specific service account (which must have cluster admin permissions as per must-gather requirements) and we are specifying a couple of additional must gather images to be run for the `kubevirt` and `ocs` subsystem. If not specified serviceAccountRef.Name will default to `default`. Also the standard must gather image: `quay.io/openshift/origin-must-gather:latest` is always added by default.
        displayName: Must Gather Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - case_management
        - support
        - must_gather
        links:
        - name: Must Gather Operator
          url: https://github.com/redhat-cop/must-gather-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/must-gather-operator
        maintainers:
        - email: rspazzol@redhat.com'
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/must-gather-operator@sha256:aebef122cdb8409f5a5d70d3cdc585e02272704d57e103553a84bc6fc238e646
        version: 1.1.2
      entries:
      - name: must-gather-operator.v1.1.2
        version: 1.1.2
      - name: must-gather-operator.v1.1.1
        version: 1.1.1
      - name: must-gather-operator.v1.1.0
        version: 1.1.0
      - name: must-gather-operator.v1.0.6
        version: 1.0.6
      - name: must-gather-operator.v1.0.5
        version: 1.0.5
      - name: must-gather-operator.v1.0.4
        version: 1.0.4
      - name: must-gather-operator.v1.0.3
        version: 1.0.3
      - name: must-gather-operator.v1.0.2
        version: 1.0.2
      - name: must-gather-operator.v1.0.1
        version: 1.0.1
      - name: must-gather-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: must-gather-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CrowdStrike
      provider-url: https://crowdStrike.com
    name: falcon-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: falcon-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconAdmission",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-admission",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconadmission",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-kac"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "admission_controller"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconContainer",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-sidecar-sensor",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconcontainer",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-container-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "sidecar"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconDeployment",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcondeployment-sample",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falcon-deployment",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-deployment"
                },
                "spec": {
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconImageAnalyzer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "falcon-operator",
                    "app.kubernetes.io/instance": "falconimageanalyzer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "falconimageanalyzer",
                    "app.kubernetes.io/part-of": "falcon-operator"
                  },
                  "name": "falcon-image-analyzer"
                },
                "spec": {
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconNodeSensor",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-node-sensor",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconnodesensor",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-node-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "daemonset"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security,Monitoring
          containerImage: registry.connect.redhat.com/crowdstrike/falcon-operator@sha256:ffe00e9e22e209d1e08f29c3d257bbd51937fcbde68a1ed93b542a26dd436dee
          createdAt: "2025-06-03T22:34:20Z"
          description: Falcon Operator installs CrowdStrike Falcon Sensors on the
            cluster
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: falcon-operator
          operators.openshift.io/valid-subscription: Not required to use this operator
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: registry.connect.redhat.com/crowdstrike/falcon-operator
          support: CrowdStrike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FalconAdmission is the Schema for the falconadmissions API
            displayName: Falcon Admission
            kind: FalconAdmission
            name: falconadmissions.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconContainer is the Schema for the falconcontainers API
            displayName: Falcon Container
            kind: FalconContainer
            name: falconcontainers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconDeployment is the Schema for the falcondeployments
              API
            displayName: Falcon Deployment
            kind: FalconDeployment
            name: falcondeployments.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconImageAnalyzer is the Schema for the falconImageAnalyzers
              API
            displayName: Falcon Image Analyzer
            kind: FalconImageAnalyzer
            name: falconimageanalyzers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconNodeSensor is the Schema for the falconnodesensors
              API
            displayName: Falcon Node Sensor
            kind: FalconNodeSensor
            name: falconnodesensors.falcon.crowdstrike.com
            version: v1alpha1
        description: |-
          The CrowdStrike Falcon Operator installs CrowdStrike Falcon custom resources on a Kubernetes cluster.

          ## About the CrowdStrike Falcon Operator
          The CrowdStrike Falcon Operator deploys CrowdStrike Falcon to the cluster. The operator exposes custom resources that allow you to protect your Kubernetes clusters when deployed.

          ## About Custom Resources

          | Custom Resource                                       | Description                                                      |
          | :--                                             | :--                                                    |
          | [FalconAdmission](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/admission/README.md) | Manages installation of Falcon Admission Controller on the cluster |
          | [FalconImageAnalyzer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/imageanalyzer/README.md) | Manages installation of Falcon Image Assessment at Runtime on the cluster |
          | [FalconContainer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/container/README.md) | Manages installation of Falcon Container Sensor on the cluster   |
          | [FalconNodeSensor](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/node/README.md)     | Manages installation of Falcon Linux Sensor on the cluster nodes |
          | [FalconDeployment](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/falcondeployment/README.md)       | Deploys FalconAdmission, FalconImageAnalyzer, FalconContainer, and FalconNodeSensor CRs from a single manifest |

          ## Installation and Deployment

          For installation and deployment of the CrowdStrike Falcon Operator and its Custom Resources, please read the [Installation and Deployment Guide](./docs/install_guide.md) and choose the deployment method that is right for your target environment.

          ## Getting Help
          If you encounter any issues while using the Falcon Operator, you can create an issue on our [Github repo](https://github.com/CrowdStrike/falcon-operator) for bugs, enhancements, or other requests.

          ## Contributing
          You can contribute by:

          * Raising any issues you find using Falcon Operator
          * Fixing issues by opening [Pull Requests](https://github.com/CrowdStrike/falcon-operator/pulls)
          * Improving documentation
          * Talking about the CrowdStrike Falcon Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/CrowdStrike/falcon-operator/issues).

          ## Additional Resources
           - CrowdStrike Container Security: [Product Page](https://www.crowdstrike.com/products/cloud-security/falcon-cloud-workload-protection/container-security/)
           - So You Think Your Containers Are Secure? Four Steps to Ensure a Secure Container Deployment: [Blog Post](https://www.crowdstrike.com/blog/four-steps-to-ensure-a-secure-containter-deployment/)
           - Container Security With CrowdStrike: [Blog Post](https://www.crowdstrike.com/blog/tech-center/container-security/)
           - To learn more about Falcon Container Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/support/documentation/146/falcon-container-sensor-for-linux), [Release Notes](https://falcon.crowdstrike.com/support/news/release-notes-falcon-container-sensor-for-linux)
           - To learn more about Falcon Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/documentation/20/falcon-sensor-for-linux#kubernetes-support)
           - [Developer Documentation](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/developer_guide.md)
        displayName: CrowdStrike Falcon Platform - Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - runtime-protection
        links:
        - name: Falcon Operator
          url: https://github.com/CrowdStrike/falcon-operator
        maintainers:
        - email: integrations@crowdstrike.com
          name: CrowdStrike Solutions Architecture
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: CrowdStrike
          url: https://crowdStrike.com
        relatedImages:
        - registry.connect.redhat.com/crowdstrike/falcon-operator@sha256:ffe00e9e22e209d1e08f29c3d257bbd51937fcbde68a1ed93b542a26dd436dee
        - registry.crowdstrike.com/falcon-container/us-1/release/falcon-sensor@sha256:92d222cd604c9f2c9e9f2c9b402b9fb0858fcb04dac38c15e25125978f43c99b
        - registry.crowdstrike.com/falcon-kac/us-1/release/falcon-kac@sha256:5049128dcfe9565f39fc6e28fb94f5bf05fb7c51cf787fc6fe5c63a2b04f800d
        - registry.crowdstrike.com/falcon-imageanalyzer/us-1/release/falcon-imageanalyzer@sha256:f08603648658041ad52ccf0cfd058da40b385904830030aa82dc50609aed7ceb
        - registry.crowdstrike.com/falcon-sensor/us-1/release/falcon-sensor@sha256:ed4f3fe8d9f5f491600f18d4440dc1d8e3a07d4b9c08a1f17382a23a3a5cf523
        version: 1.8.0
      entries:
      - name: falcon-operator.v1.8.0
        version: 1.8.0
      - name: falcon-operator.v1.7.0
        version: 1.7.0
      - name: falcon-operator.v1.6.0
        version: 1.6.0
      - name: falcon-operator.v1.5.0
        version: 1.5.0
      - name: falcon-operator.v1.4.0
        version: 1.4.0
      - name: falcon-operator.v1.3.1
        version: 1.3.1
      - name: falcon-operator.v1.3.0
        version: 1.3.0
      - name: falcon-operator.v1.2.0
        version: 1.2.0
      - name: falcon-operator.v1.1.0
        version: 1.1.0
      - name: falcon-operator.v1.0.1
        version: 1.0.1
      - name: falcon-operator.v1.0.0
        version: 1.0.0
      name: certified-1.0
    defaultChannel: certified-1.0
    packageName: falcon-operator
    provider:
      name: CrowdStrike
      url: https://crowdStrike.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Airlock
      provider-url: ""
    name: airlock-microgateway
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airlock-microgateway.v4.5.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControl","metadata":{"name":"access-control-example"},"spec":{"policies":[{"requestConditions":{"path":{"matcher":{"exact":"/secure"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}},{"oidc":{"claim":{"name":"roles","value":{"matcher":{"contains":"\"admin\""}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"oidc-example"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"identity-propagation-example"}}],"onFailure":"Pass"}},{"authorization":{}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurity","metadata":{"name":"content-security-example"},"spec":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1"},"spec":{"value":{"name":"ext-authz","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}}}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"lifetime":"16h"},"prefix":"session_prefix_example"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SidecarGateway","metadata":{"name":"sidecar-gateway-example"},"spec":{"podSelector":{"matchLabels":{"app":"example-app"}},"sessionHandlingRef":{"name":"session-handling-example"},"applications":[{"containerPort":8443,"routes":[{"pathPrefix":"/metrics","unsecured":{}},{"pathPrefix":"/","secured":{"accessControlRef":{"name":"access-control-example"},"contentSecurityRef":{"name":"content-security-example"}}}],"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}},"restrictions":{"http":{"headersLength":"80Ki"}},"protocol":{"auto":{}},"tls":{"enable":true,"secretRef":{"name":"example-tls-secret"},"protocol":{"minimum":"TLSv1_3"},"clientCertificate":{"ignored":{}},"xfcc":"AlwaysForwardOnly"}},"upstream":{"tls":{"enable":true,"protocol":{"minimum":"TLSv1_2"}}},"envoyHTTPFilterRefs":{"prepend":[{"name":"envoy-http-filter-example-1"}]},"telemetryRef":{"name":"telemetry-example"}}],"envoyClusterRefs":[{"name":"envoy-cluster-example-1"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-06-13T17:33:31Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: AccessControl
            name: accesscontrols.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurity
            name: contentsecurities.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: SidecarGateway
            name: sidecargateways.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056)
          * Gateway API
          * (Optional) Airlock Microgateway CNI (Required for [data plane mode sidecar](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000137)).

          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ### (Optional) Install Airlock Microgateway CNI
          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          Refer to [our documentation](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000140) for installation instructions

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD_MONITOR\_CREATE | "false" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.5/
        - name: Airlock Microgateway Labs
          url: https://play.instruqt.com/airlock/invite/hyi9fy4b4jzc?icp_referrer=openshift
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:806366a0b3f78844fe14775fc290fbfbde8d2a701d34a6478b019a0e7140b699
        version: 4.5.4
      entries:
      - name: airlock-microgateway.v4.5.4
        version: 4.5.4
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.5"
    - currentCSV: airlock-microgateway.v4.6.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControl","metadata":{"name":"access-control-example"},"spec":{"policies":[{"requestConditions":{"path":{"matcher":{"exact":"/secure"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}},{"oidc":{"claim":{"name":"roles","value":{"matcher":{"contains":"\"admin\""}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"oidc-example"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"token-exchange-example"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"identity-propagation-example"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"jwt-example"}}}}},{"authorization":{}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurity","metadata":{"name":"content-security-example"},"spec":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1"},"spec":{"value":{"name":"ext-authz","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}}}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"lifetime":"16h"},"prefix":"session_prefix_example"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SidecarGateway","metadata":{"name":"sidecar-gateway-example"},"spec":{"podSelector":{"matchLabels":{"app":"example-app"}},"sessionHandlingRef":{"name":"session-handling-example"},"applications":[{"containerPort":8443,"routes":[{"pathPrefix":"/metrics","unsecured":{}},{"pathPrefix":"/","secured":{"accessControlRef":{"name":"access-control-example"},"contentSecurityRef":{"name":"content-security-example"}}}],"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}},"restrictions":{"http":{"headersLength":"80Ki"}},"protocol":{"auto":{}},"tls":{"enable":true,"secretRef":{"name":"example-tls-secret"},"protocol":{"minimum":"TLSv1_3"},"clientCertificate":{"ignored":{}},"xfcc":"AlwaysForwardOnly"}},"upstream":{"tls":{"enable":true,"protocol":{"minimum":"TLSv1_2"}}},"envoyHTTPFilterRefs":{"prepend":[{"name":"envoy-http-filter-example-1"}]},"telemetryRef":{"name":"telemetry-example"}}],"envoyClusterRefs":[{"name":"envoy-cluster-example-1"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-23T06:27:21Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: AccessControl
            name: accesscontrols.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurity
            name: contentsecurities.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: SidecarGateway
            name: sidecargateways.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056)
          * Gateway API
          * (Optional) Airlock Microgateway CNI (Required for [data plane mode sidecar](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000137)).

          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.3.0/standard-install.yaml
          ```

          ### (Optional) Install Airlock Microgateway CNI
          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          Refer to [our documentation](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000140) for installation instructions

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD_MONITOR\_CREATE | "false" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.6/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:8fa300ced5b69acbad824c037d2b344fd1926bcfb39e264e9f0f055ceb9bd68d
        version: 4.6.4
      entries:
      - name: airlock-microgateway.v4.6.4
        version: 4.6.4
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.6"
    - currentCSV: airlock-microgateway.v4.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1","namespace":"default"},"spec":{"value":{"name":"envoycluster/default/envoy-cluster-example-1","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyExtensionPolicy","metadata":{"name":"envoy-extension-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"envoyHTTPFilterRefs":{"prepend":[{"name":"geoip-blocker"}],"append":[{"name":"lua-transform-request"}]},"envoyClusterRefs":[{"name":"geoip-server"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"volumes":[{"name":"vol-1","emptyDir":{}}],"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}},"volumeMounts":[{"name":"vol-1","mountPath":"/tmp/data"}]}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"telemetryRef":{"name":"telemetry-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"mode":"Enforce","persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"idle":"30s","lifetime":"16h"},"prefix":"session_prefix_example","cookie":{"name":"SESS","attributes":{"domain":"example.org","path":"/","sameSite":"Strict","secure":true}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"},"session":{"id":"%SESSION_ID%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-03T13:02:23Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyExtensionPolicy
            name: envoyextensionpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ***Note:*** Sidecar-based data plane mode is *not* supported by this OLM Bundle.
          If you want to use the Sidecar-based data plane mode, use Helm instead. Please note that the Airlock Microgateway CNI and the Helm charts are not certified for RedHat OpenShift.
          Refer to https://github.com/airlock/microgateway for installation instruction with Helm.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056)
          * [Gateway API](https://gateway-api.sigs.k8s.io/)

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ***Note:*** *This step is only required for OpenShift versions below 4.19, as Gateway API is included by default in OpenShift 4.19+.*

          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD\_MONITOR\_CREATE | "true" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gateway API
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.7/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:d4ce35184922953c209a9ebb234ca3370dbfc5054de411b5503c26ac2730b7ae
        version: 4.7.1
      entries:
      - name: airlock-microgateway.v4.7.1
        version: 4.7.1
      - name: airlock-microgateway.v4.7.0
        version: 4.7.0
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.7"
    - currentCSV: airlock-microgateway.v4.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1","namespace":"default"},"spec":{"value":{"name":"envoycluster/default/envoy-cluster-example-1","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyExtensionPolicy","metadata":{"name":"envoy-extension-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"envoyHTTPFilterRefs":{"prepend":[{"name":"geoip-blocker"}],"append":[{"name":"lua-transform-request"}]},"envoyClusterRefs":[{"name":"geoip-server"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"volumes":[{"name":"vol-1","emptyDir":{}}],"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}},"volumeMounts":[{"name":"vol-1","mountPath":"/tmp/data"}]}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"telemetryRef":{"name":"telemetry-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"mode":"Enforce","persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"idle":"30s","lifetime":"16h"},"prefix":"session_prefix_example","cookie":{"name":"SESS","attributes":{"domain":"example.org","path":"/","sameSite":"Strict","secure":true}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"},"session":{"id":"%SESSION_ID%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-03T13:02:23Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyExtensionPolicy
            name: envoyextensionpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ***Note:*** Sidecar-based data plane mode is *not* supported by this OLM Bundle.
          If you want to use the Sidecar-based data plane mode, use Helm instead. Please note that the Airlock Microgateway CNI and the Helm charts are not certified for RedHat OpenShift.
          Refer to https://github.com/airlock/microgateway for installation instruction with Helm.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056)
          * [Gateway API](https://gateway-api.sigs.k8s.io/)

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ***Note:*** *This step is only required for OpenShift versions below 4.19, as Gateway API is included by default in OpenShift 4.19+.*

          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD\_MONITOR\_CREATE | "true" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gateway API
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.7/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:d4ce35184922953c209a9ebb234ca3370dbfc5054de411b5503c26ac2730b7ae
        version: 4.7.1
      entries:
      - name: airlock-microgateway.v4.7.1
        version: 4.7.1
      - name: airlock-microgateway.v4.7.0
        version: 4.7.0
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: stable
    defaultChannel: "4.7"
    packageName: airlock-microgateway
    provider:
      name: Airlock
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: marin3r
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marin3r.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyConfig",
                "metadata": {
                  "name": "envoyconfig-example",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "nodeID": "example",
                  "resources": [
                    {
                      "type": "cluster",
                      "value": {
                        "connect_timeout": "0.01s",
                        "dns_lookup_family": "V4_ONLY",
                        "lb_policy": "ROUND_ROBIN",
                        "load_assignment": {
                          "cluster_name": "example",
                          "endpoints": [
                            {
                              "lb_endpoints": [
                                {
                                  "endpoint": {
                                    "address": {
                                      "socket_address": {
                                        "address": "example",
                                        "port_value": 8080
                                      }
                                    }
                                  }
                                }
                              ]
                            }
                          ]
                        },
                        "name": "example",
                        "type": "STRICT_DNS"
                      }
                    },
                    {
                      "type": "route",
                      "value": {
                        "name": "local",
                        "virtual_hosts": [
                          {
                            "domains": [
                              "*"
                            ],
                            "name": "all",
                            "routes": [
                              {
                                "match": {
                                  "prefix": "/"
                                },
                                "route": {
                                  "cluster": "example"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    {
                      "type": "listener",
                      "value": {
                        "address": {
                          "socket_address": {
                            "address": "0.0.0.0",
                            "port_value": 8443
                          }
                        },
                        "filter_chains": [
                          {
                            "filters": [
                              {
                                "name": "envoy.filters.network.http_connection_manager",
                                "typed_config": {
                                  "@type": "type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager",
                                  "access_log": [
                                    {
                                      "name": "envoy.access_loggers.file",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.access_loggers.file.v3.FileAccessLog",
                                        "path": "/dev/stdout"
                                      }
                                    }
                                  ],
                                  "http_filters": [
                                    {
                                      "name": "envoy.filters.http.router",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.filters.http.router.v3.Router"
                                      }
                                    }
                                  ],
                                  "rds": {
                                    "config_source": {
                                      "ads": {},
                                      "resource_api_version": "V3"
                                    },
                                    "route_config_name": "local"
                                  },
                                  "stat_prefix": "https"
                                }
                              }
                            ],
                            "transport_socket": {
                              "name": "envoy.transport_sockets.tls",
                              "typed_config": {
                                "@type": "type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext",
                                "common_tls_context": {
                                  "tls_certificate_sds_secret_configs": [
                                    {
                                      "name": "example.default.svc",
                                      "sds_config": {
                                        "ads": {},
                                        "resource_api_version": "V3"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          }
                        ],
                        "name": "https"
                      }
                    },
                    {
                      "blueprint": "tlsCertificate",
                      "generateFromTlsSecret": "example.default.svc",
                      "type": "secret"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "DiscoveryService",
                "metadata": {
                  "name": "discoveryservice-example",
                  "namespace": "my-namespace"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyDeployment",
                "metadata": {
                  "name": "envoydeployment-sample",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "discoveryServiceRef": "discoveryservice-example",
                  "envoyConfigRef": "envoyconfig-example",
                  "ports": [
                    {
                      "name": "https",
                      "port": 8443
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: quay.io/3scale/marin3r
          createdAt: "2024-08-05T16:08:25Z"
          description: Lighweight, CRD based Envoy control plane for Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["envoyconfigrevisions.marin3r.3scale.net","discoveryservicecertificates.operator.marin3r.3scale.net"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/3scale-ops/marin3r
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DiscoveryServiceCertificate is an internal resource used
              to create certificates. This resource is used by the DiscoveryService
              controller to create the required certificates for the different components.
              Direct use of DiscoveryServiceCertificate objects is discouraged.
            displayName: DiscoveryServiceCertificate
            kind: DiscoveryServiceCertificate
            name: discoveryservicecertificates.operator.marin3r.3scale.net
            version: v1alpha1
          - description: DiscoveryService represents an envoy discovery service server.
              Only one instance per namespace is currently supported.
            displayName: DiscoveryService
            kind: DiscoveryService
            name: discoveryservices.operator.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfigRevision is an internal resource that stores a
              specific version of an EnvoyConfig resource. EnvoyConfigRevisions are
              automatically created and deleted by the EnvoyConfig controller and
              are not intended to be directly used. Use EnvoyConfig objects instead.
            displayName: EnvoyConfigRevision
            kind: EnvoyConfigRevision
            name: envoyconfigrevisions.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfig holds the configuration for a given envoy nodeID.
              The spec of an EnvoyConfig object holds the Envoy resources that conform
              the desired configuration for the given nodeID and that the discovery
              service will send to any envoy client that identifies itself with that
              nodeID.
            displayName: EnvoyConfig
            kind: EnvoyConfig
            name: envoyconfigs.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyDeployment is a resource to deploy and manage a Kubernetes
              Deployment of Envoy Pods.
            displayName: EnvoyDeployment
            kind: EnvoyDeployment
            name: envoydeployments.operator.marin3r.3scale.net
            version: v1alpha1
        description: |
          MARIN3R implements a control plane to deploy, configure and operate a fleet of envoy instances within a Kubernetes cluster.
          It operates a group of gateways and Pod sidecar containers to provide general purpose functionality required to operate production systems.

          Features:

          * Deploy and manage an Envoy xDS server using the DiscoveryService custom resource.
          * Inject Envoy sidecar containers based on Pod annotations.
          * Deploy Envoy as a Kubernetes Deployment using the EnvoyDeployment custom resource.
          * Dynamic Envoy configuration and re-configuration using the EnvoyConfig custom resource.
          * Use Kubernetes Secrets as certificate sources.
          * Syntactic validation of Envoy configurations.
          * Lifecycle management of Envoy containers (graceful shutdown and connection draining).
          * Self-healing.

          Check the [docs](https://github.com/3scale-ops/marin3r#readme) for more information.

          ## License
          MARIN3R is licensed under the [Apache 2.0 license](https://github.com/3scale/prometheus-exporter-operator/blob/master/LICENSE)
        displayName: MARIN3R
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - envoy
        - discovery
        - controlplane
        - proxy
        - sidecar
        - xds
        - network
        - gateway
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/marin3r
        maintainers:
        - email: 3scale-operations+marin3r@redhat.com
          name: 3scale Operations Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/marin3r:v0.13.1
        version: 0.13.1
      entries:
      - name: marin3r.v0.13.1
        version: 0.13.1
      - name: marin3r.v0.12.3
        version: 0.12.3
      - name: marin3r.v0.11.1
        version: 0.11.1
      - name: marin3r.v0.10.0
        version: 0.10.0
      name: alpha
    - currentCSV: marin3r.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyConfig",
                "metadata": {
                  "name": "envoyconfig-example",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "nodeID": "example",
                  "resources": [
                    {
                      "type": "cluster",
                      "value": {
                        "connect_timeout": "0.01s",
                        "dns_lookup_family": "V4_ONLY",
                        "lb_policy": "ROUND_ROBIN",
                        "load_assignment": {
                          "cluster_name": "example",
                          "endpoints": [
                            {
                              "lb_endpoints": [
                                {
                                  "endpoint": {
                                    "address": {
                                      "socket_address": {
                                        "address": "example",
                                        "port_value": 8080
                                      }
                                    }
                                  }
                                }
                              ]
                            }
                          ]
                        },
                        "name": "example",
                        "type": "STRICT_DNS"
                      }
                    },
                    {
                      "type": "route",
                      "value": {
                        "name": "local",
                        "virtual_hosts": [
                          {
                            "domains": [
                              "*"
                            ],
                            "name": "all",
                            "routes": [
                              {
                                "match": {
                                  "prefix": "/"
                                },
                                "route": {
                                  "cluster": "example"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    {
                      "type": "listener",
                      "value": {
                        "address": {
                          "socket_address": {
                            "address": "0.0.0.0",
                            "port_value": 8443
                          }
                        },
                        "filter_chains": [
                          {
                            "filters": [
                              {
                                "name": "envoy.filters.network.http_connection_manager",
                                "typed_config": {
                                  "@type": "type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager",
                                  "access_log": [
                                    {
                                      "name": "envoy.access_loggers.file",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.access_loggers.file.v3.FileAccessLog",
                                        "path": "/dev/stdout"
                                      }
                                    }
                                  ],
                                  "http_filters": [
                                    {
                                      "name": "envoy.filters.http.router",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.filters.http.router.v3.Router"
                                      }
                                    }
                                  ],
                                  "rds": {
                                    "config_source": {
                                      "ads": {},
                                      "resource_api_version": "V3"
                                    },
                                    "route_config_name": "local"
                                  },
                                  "stat_prefix": "https"
                                }
                              }
                            ],
                            "transport_socket": {
                              "name": "envoy.transport_sockets.tls",
                              "typed_config": {
                                "@type": "type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext",
                                "common_tls_context": {
                                  "tls_certificate_sds_secret_configs": [
                                    {
                                      "name": "example.default.svc",
                                      "sds_config": {
                                        "ads": {},
                                        "resource_api_version": "V3"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          }
                        ],
                        "name": "https"
                      }
                    },
                    {
                      "blueprint": "tlsCertificate",
                      "generateFromTlsSecret": "example.default.svc",
                      "type": "secret"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "DiscoveryService",
                "metadata": {
                  "name": "discoveryservice-example",
                  "namespace": "my-namespace"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyDeployment",
                "metadata": {
                  "name": "envoydeployment-sample",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "discoveryServiceRef": "discoveryservice-example",
                  "envoyConfigRef": "envoyconfig-example",
                  "ports": [
                    {
                      "name": "https",
                      "port": 8443
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: quay.io/3scale/marin3r
          createdAt: "2024-08-05T16:08:25Z"
          description: Lighweight, CRD based Envoy control plane for Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["envoyconfigrevisions.marin3r.3scale.net","discoveryservicecertificates.operator.marin3r.3scale.net"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/3scale-ops/marin3r
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DiscoveryServiceCertificate is an internal resource used
              to create certificates. This resource is used by the DiscoveryService
              controller to create the required certificates for the different components.
              Direct use of DiscoveryServiceCertificate objects is discouraged.
            displayName: DiscoveryServiceCertificate
            kind: DiscoveryServiceCertificate
            name: discoveryservicecertificates.operator.marin3r.3scale.net
            version: v1alpha1
          - description: DiscoveryService represents an envoy discovery service server.
              Only one instance per namespace is currently supported.
            displayName: DiscoveryService
            kind: DiscoveryService
            name: discoveryservices.operator.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfigRevision is an internal resource that stores a
              specific version of an EnvoyConfig resource. EnvoyConfigRevisions are
              automatically created and deleted by the EnvoyConfig controller and
              are not intended to be directly used. Use EnvoyConfig objects instead.
            displayName: EnvoyConfigRevision
            kind: EnvoyConfigRevision
            name: envoyconfigrevisions.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfig holds the configuration for a given envoy nodeID.
              The spec of an EnvoyConfig object holds the Envoy resources that conform
              the desired configuration for the given nodeID and that the discovery
              service will send to any envoy client that identifies itself with that
              nodeID.
            displayName: EnvoyConfig
            kind: EnvoyConfig
            name: envoyconfigs.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyDeployment is a resource to deploy and manage a Kubernetes
              Deployment of Envoy Pods.
            displayName: EnvoyDeployment
            kind: EnvoyDeployment
            name: envoydeployments.operator.marin3r.3scale.net
            version: v1alpha1
        description: |
          MARIN3R implements a control plane to deploy, configure and operate a fleet of envoy instances within a Kubernetes cluster.
          It operates a group of gateways and Pod sidecar containers to provide general purpose functionality required to operate production systems.

          Features:

          * Deploy and manage an Envoy xDS server using the DiscoveryService custom resource.
          * Inject Envoy sidecar containers based on Pod annotations.
          * Deploy Envoy as a Kubernetes Deployment using the EnvoyDeployment custom resource.
          * Dynamic Envoy configuration and re-configuration using the EnvoyConfig custom resource.
          * Use Kubernetes Secrets as certificate sources.
          * Syntactic validation of Envoy configurations.
          * Lifecycle management of Envoy containers (graceful shutdown and connection draining).
          * Self-healing.

          Check the [docs](https://github.com/3scale-ops/marin3r#readme) for more information.

          ## License
          MARIN3R is licensed under the [Apache 2.0 license](https://github.com/3scale/prometheus-exporter-operator/blob/master/LICENSE)
        displayName: MARIN3R
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - envoy
        - discovery
        - controlplane
        - proxy
        - sidecar
        - xds
        - network
        - gateway
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/marin3r
        maintainers:
        - email: 3scale-operations+marin3r@redhat.com
          name: 3scale Operations Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/marin3r:v0.13.1
        version: 0.13.1
      entries:
      - name: marin3r.v0.13.1
        version: 0.13.1
      - name: marin3r.v0.12.3
        version: 0.12.3
      - name: marin3r.v0.11.1
        version: 0.11.1
      - name: marin3r.v0.10.0
        version: 0.10.0
      name: stable
    defaultChannel: stable
    packageName: marin3r
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: machine-deletion-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: machine-deletion-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/machine-deletion-remediation-operator:v0.1.0
          createdAt: "2023-05-11 10:46:12"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/machine-deletion-remediation-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: machine-deletion-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: machine-deletion-remediation.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/machine-deletion-remediation-operator:v0.3.0
          createdAt: "2024-03-05 18:45:16"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/medik8s/machine-deletion-remediation-operator:v0.3.0
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        version: 0.3.0
      entries:
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      - name: machine-deletion-remediation.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: machine-deletion-remediation
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-nimbusmosaic-dusim
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-nimbusmosaic-dusim.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Nimbusmosaicdusim",
                "metadata": {
                  "name": "nimbusmosaicdusim-sample"
                },
                "spec": {
                  "development": false,
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "microservices": {
                    "jvm": "-Xmx300M -Xms64M -Xss256k"
                  },
                  "service": {
                    "type": "ClusterIP"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:41:56Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nimbusmosaicdusim
            name: nimbusmosaicdusims.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Nimbus Mosaic DUSim Operator
        displayName: Keysight Nimbus Mosaic DUSim Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Nimbusmosaic Dusim
          url: https://keysight-nimbusmosaic-dusim.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-nimbusmosaic-dusim@sha256:655dd6595a9c8e82947bed24991ac52f8fa55be99ca410e7d9f0026c51ef4520
        version: 0.1.21
      entries:
      - name: keysight-nimbusmosaic-dusim.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-nimbusmosaic-dusim
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-namespace-scope-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-namespace-scope-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1",
                "kind": "NamespaceScope",
                "metadata": {
                  "name": "namespacescope"
                },
                "spec": {
                  "namespaceMembers": [
                    "ibm-common-services",
                    "default"
                  ],
                  "restartLabels": {
                    "intent": "projected"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
          createdAt: "2025-02-19T00:54:24Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.2.14
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/IBM/ibm-namespace-scope-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NamespaceScope is the Schema for the namespacescopes API
            displayName: Namespace Scope
            kind: NamespaceScope
            name: namespacescopes.operator.ibm.com
            version: v1
        description: This operator automates the extension of operator watch and service
          account permission scope to other namespaces in an openshift cluster.
        displayName: IBM NamespaceScope Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Namespace Scope Operator
          url: https://github.com/IBM/ibm-namespace-scope-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:0774b1d916ebc354050612474800dda3a501c66a7a5aab6d18698c8c4e81ac4e
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-bundle@sha256:b6cb9e5976b749dd7818ac19d1cfd5e24673f3c052a41d6e2fe1d2d7094694b4
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:7a2cd72736bf9d30341763fd1652ff4b9ce3c785da5123645ee7664ee38c7283
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-catalog@sha256:c016904e8c5cfcf8270241ad5e28b52f528859984ed894cdd5e1283fa90cd6e8
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator-bundle@sha256:c743dd7de2edbd08c6c73e20cf15f3b147491ad77a81677339e88caffb93e84b
        version: 4.2.14
      entries:
      - name: ibm-namespace-scope-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-namespace-scope-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hitachi
      provider-url: ""
    name: hnas-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hnas-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.csi.hitachi.com/v1",
                "kind": "HNAS",
                "metadata": {
                  "name": "hnas",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-operator@sha256:9d3f9f10fc7172ee9cebbd9ee318906756d7e96909f2e66e0534e98bb53e0a7b
          createdAt: "2025-07-30T15:12:23Z"
          description: An operator for managing Hitachi NAS CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HNAS is the Schema for the hnas API
            displayName: HNAS
            kind: HNAS
            name: hnas.charts.csi.hitachi.com
            version: v1
        description: |-
          The Hitachi NAS CSI driver operator deploys the Hitachi NAS CSI driver on Kubernetes and OpenShift environments.  The driver can be used to create and manage persistent storage hosted on Hitachi NAS products, including those in the Hitachi VSP One File 3x range and the HNAS range of products.

          For the complete documentation, go to our [Product Documentation site](https://docs.hitachivantara.com/home) and refer to the Hitachi NAS CSI Driver User Guide - MK-92HNAS091.  The guide covers prerequisites for the NAS system and covers the driver functionality and all configuration options available for the driver.
        displayName: Hitachi NAS CSI Driver Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hitachi
        - Storage
        - Container Storage Interface
        - CSI
        - NAS
        links:
        - name: Product Documentation
          url: https://docs.hitachivantara.com/home
        maturity: stable
        minKubeVersion: 1.28.0
        provider:
          name: Hitachi
        relatedImages:
        - registry.k8s.io/sig-storage/csi-resizer@sha256:ab774734705a906561e15b67f6a96538f3ad520335d636f793aaafe87a3b5200
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:dd788d79cf4c1b8edee6d9b80b8a1ebfc51a38a365c5be656986b129be9ac784
        - registry.k8s.io/sig-storage/livenessprobe@sha256:2c5f9dc4ea5ac5509d93c664ae7982d4ecdec40ca7b0638c24e5b16243b8360f
        - registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-operator@sha256:9d3f9f10fc7172ee9cebbd9ee318906756d7e96909f2e66e0534e98bb53e0a7b
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:d7138bcc3aa5f267403d45ad4292c95397e421ea17a0035888850f424c7de25d
        - registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-driver@sha256:6b6f6816c9f4e608e93e7fbc0347255415426a0ff0c9a53b7750e5ba54d91262
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d5e46da8aff7d73d6f00c761dae94472bcda6e78f4f17b3802dc89d44de0111b
        - registry.k8s.io/sig-storage/csi-attacher@sha256:a399393ff5bd156277c56bae0c08389b1a1b95b7fd6ea44a316ce55e0dd559d7
        version: 1.0.0
      entries:
      - name: hnas-csi-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: hnas-csi-operator
    provider:
      name: Hitachi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-memorydb-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-memorydb-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "ACL",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "ParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "Snapshot",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "SubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/memorydb-controller:1.1.1
          createdAt: "2025-09-19T19:36:49Z"
          description: Amazon MemoryDB for Redis controller is a service controller
            for managing MemoryDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ACL represents the state of an AWS memorydb ACL resource.
            displayName: ACL
            kind: ACL
            name: acls.memorydb.services.k8s.aws
            version: v1alpha1
          - description: Cluster represents the state of an AWS memorydb Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.memorydb.services.k8s.aws
            version: v1alpha1
          - description: ParameterGroup represents the state of an AWS memorydb ParameterGroup
              resource.
            displayName: ParameterGroup
            kind: ParameterGroup
            name: parametergroups.memorydb.services.k8s.aws
            version: v1alpha1
          - description: Snapshot represents the state of an AWS memorydb Snapshot
              resource.
            displayName: Snapshot
            kind: Snapshot
            name: snapshots.memorydb.services.k8s.aws
            version: v1alpha1
          - description: SubnetGroup represents the state of an AWS memorydb SubnetGroup
              resource.
            displayName: SubnetGroup
            kind: SubnetGroup
            name: subnetgroups.memorydb.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS memorydb User resource.
            displayName: User
            kind: User
            name: users.memorydb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon MemoryDB for Redis resources in AWS from within your Kubernetes cluster.

          **About Amazon MemoryDB for Redis**

          MemoryDB for Redis is a durable, in-memory database service that delivers ultra-fast performance. It is purpose-built for modern applications with microservices architectures.
          MemoryDB is compatible with Redis, a popular open source data store, enabling you to quickly build applications using the same flexible and friendly Redis data structures, APIs, and commands that they already use today. With MemoryDB, all of your data is stored in memory, which enables you to achieve microsecond read and single-digit millisecond write latency and high throughput. MemoryDB also stores data durably across multiple Availability Zones (AZs) using a Multi-AZ transactional log to enable fast failover, database recovery, and node restarts.
          Delivering both in-memory performance and Multi-AZ durability, MemoryDB can be used as a high-performance primary database for your microservices applications, eliminating the need to separately manage both a cache and durable database.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - MemoryDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - memorydb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon MemoryDB for Redis Developer Resources
          url: https://docs.aws.amazon.com/memorydb/?id=docs_gateway
        maintainers:
        - email: ack-maintainers@amazon.com
          name: memorydb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/memorydb-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-memorydb-controller.v1.1.1
        version: 1.1.1
      - name: ack-memorydb-controller.v1.1.0
        version: 1.1.0
      - name: ack-memorydb-controller.v1.0.15
        version: 1.0.15
      - name: ack-memorydb-controller.v1.0.13
        version: 1.0.13
      - name: ack-memorydb-controller.v1.0.12
        version: 1.0.12
      - name: ack-memorydb-controller.v1.0.11
        version: 1.0.11
      - name: ack-memorydb-controller.v1.0.10
        version: 1.0.10
      - name: ack-memorydb-controller.v1.0.9
        version: 1.0.9
      - name: ack-memorydb-controller.v1.0.8
        version: 1.0.8
      - name: ack-memorydb-controller.v1.0.7
        version: 1.0.7
      - name: ack-memorydb-controller.v1.0.6
        version: 1.0.6
      - name: ack-memorydb-controller.v1.0.5
        version: 1.0.5
      - name: ack-memorydb-controller.v1.0.4
        version: 1.0.4
      - name: ack-memorydb-controller.v1.0.3
        version: 1.0.3
      - name: ack-memorydb-controller.v1.0.2
        version: 1.0.2
      - name: ack-memorydb-controller.v1.0.1
        version: 1.0.1
      - name: ack-memorydb-controller.v1.0.0
        version: 1.0.0
      - name: ack-memorydb-controller.v0.0.12
        version: 0.0.12
      - name: ack-memorydb-controller.v0.0.11
        version: 0.0.11
      - name: ack-memorydb-controller.v0.0.10
        version: 0.0.10
      - name: ack-memorydb-controller.v0.0.9
        version: 0.0.9
      - name: ack-memorydb-controller.v0.0.8
        version: 0.0.8
      - name: ack-memorydb-controller.v0.0.7
        version: 0.0.7
      - name: ack-memorydb-controller.v0.0.6
        version: 0.0.6
      - name: ack-memorydb-controller.v0.0.5
        version: 0.0.5
      - name: ack-memorydb-controller.v0.0.4
        version: 0.0.4
      - name: ack-memorydb-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-memorydb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: GitLab, Inc.
      provider-url: https://gitlab.com
    name: gitlab-runner-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-runner-operator.v1.41.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta2",
                "kind": "Runner",
                "metadata": {
                  "name": "gitlab-runner-example",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "gitlabUrl": "https://gitlab.example.com",
                  "token": "runner-token-secret",
                  "tags": "kubernetes,example",
                  "concurrent": 10,
                  "imagePullPolicy": "IfNotPresent"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery, Developer Tools
          containerImage: registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
          createdAt: "2025-09-25T15:08:40Z"
          description: GitLab Runner operator manages lifecycle of GitLab Runner instances
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.11.0 <1.41.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: none
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
          support: GitLab, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Runner is the open source project used to run your jobs and
              send the results back to GitLab
            displayName: GitLab Runner
            kind: Runner
            name: runners.apps.gitlab.com
            version: v1beta2
        description: |-
          GitLab Runner is the lightweight, highly-scalable agent that runs your build jobs and sends the results back to a GitLab instance. GitLab Runner works in conjunction with GitLab CI/CD, the open-source continuous integration service included with GitLab.

              The GitLab Runner operator manages the lifecycle of GitLab Runner in Kubernetes or Openshift clusters. The operator aims to automate the tasks needed to run your CI/CD jobs in your container orchestration platform.

              ## Prerequisites

                For Kubernetes cluster, install cert-manager:

                ```shell
                kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.7.1/cert-manager.yaml
                ```

              ## GitLab Runner version

              This version of **GitLab Runner Operator** ships with **GitLab Runner v18.4.0**.

              To use a different version of **GitLab Runner** change the [`runnerImage` and `helperImage` properties](https://docs.gitlab.com/runner/configuration/configuring_runner_operator.html#operator-properties).

              ## Usage

               To link a GitLab Runner instance to a self-hosted GitLab instance or the hosted [GitLab](https://gitlab.com), you first need to:

               1. Create a secret containing the `runner-token` from your GitLab project.

                ```
                cat > gitlab-runner-secret.yml << EOF
                apiVersion: v1
                kind: Secret
                metadata:
                  name: gitlab-runner-secret
                type: Opaque
                # Only one of the following fields must be set. The Operator will fail to register Runner if both are provided.
                # # NOTE: runner-registration-token is deprecated and will be removed in GitLab 18.0. You should use runner-token instead.
                stringData:
                  runner-token: REPLACE_ME # your project runner token
                  # runner-registration-token: "" # your project runner secret
                EOF
                ```

                ```
                oc apply -f gitlab-runner-secret.yml
                ```

               2. Create the Custom Resource Definition (CRD) file and include the following information. The tags value must be openshift for the job to run.

                 ```
                 cat > gitlab-runner.yml << EOF
                 apiVersion: apps.gitlab.com/v1beta2
                 kind: Runner
                 metadata:
                   name: gitlab-runner
                 spec:
                   gitlabUrl: https://gitlab.example.com
                   buildImage: alpine
                   token: gitlab-runner-secret
                   tags: openshift
                 EOF
                 ```

                ```
                oc apply -f gitlab-runner.yml
                ```

              ## Full documentation

              Visit [Install GitLab Runner Operator](https://docs.gitlab.com/runner/install/operator.html)
        displayName: GitLab Runner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GitLab
        - CI/CD
        - DevOps
        - SAST
        - DAST
        links:
        - name: Gitlab Runner Operator
          url: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
        - name: GitLab Docs
          url: https://docs.gitlab.com
        - name: GitLab Runner
          url: https://docs.gitlab.com/runner/
        - name: GitLab CI/CD
          url: https://docs.gitlab.com/ee/ci/quick_start/
        maintainers:
        - email: support@gitlab.com
          name: Georgi N. Georgiev
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: GitLab, Inc.
          url: https://gitlab.com
        relatedImages:
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-ocp@sha256:0f2a5a3b6779d719eecbf079b99cf06b652603d74b563f57d753e32966d253dd
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-helper-ocp@sha256:29727fae83a226bcf57607240d4046a0b8c208b5ff42ee1f8ca0f115bb188809
        version: 1.41.1
      entries:
      - name: gitlab-runner-operator.v1.41.1
        version: 1.41.1
      - name: gitlab-runner-operator.v1.41.0
        version: 1.41.0
      - name: gitlab-runner-operator.v1.40.1
        version: 1.40.1
      - name: gitlab-runner-operator.v1.40.0
        version: 1.40.0
      - name: gitlab-runner-operator.v1.39.0
        version: 1.39.0
      name: stable
    defaultChannel: stable
    packageName: gitlab-runner-operator
    provider:
      name: GitLab, Inc.
      url: https://gitlab.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Anjuna Security
      provider-url: https://www.anjuna.io
    name: anjuna-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anjuna-operator.v2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anjuna.io/v1",
                "kind": "AnjunaConfig",
                "metadata": {
                  "name": "anjuna-config-remote"
                },
                "spec": {
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "runtimeImage": "registry.connect.redhat.com/anjuna/anjuna-k8s-sev-runtime@sha256:1531b1ba955b2d3ea5d9238d2e20617a534bb1bab73f5fadadf8c9d3aee410b4"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-02-19T01:11:18Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          operatorframework.io/suggested-namespace: anjuna-operator
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnjunaConfig operand instructs the Anjuna Seaglass Operator
              to know (1) what needs to be installed into the cluster and (2) what
              nodes will be affected.
            displayName: Anjuna Config
            kind: AnjunaConfig
            name: anjunaconfigs.anjuna.io
            version: v1
        description: "The Anjuna Seaglass Operator for OpenShift allows running any
          Pod in\nOpenShift in a Confidential Pod leveraging the full power of the
          Confidential\nComputing technology.\n\n# Anjuna Seaglass\nAnjuna Seaglass™
          is the Universal Confidential Computing Platform to run applications in
          any cloud, and on-premises, with complete data security and privacy. \nIt
          isolates workloads in a hardware-backed environment that creates a zero-trust
          model for code and data.\n\nThe Anjuna Seaglass platform allows users to
          leverage the power of Confidential Computing:\n\nFor any Confidential Computing
          technology\nIn any Cloud Service Provider, as well as on-premises\nWith
          no changes to the application\n\nRunning a Confidential Pod with Anjuna
          Seaglass protects the Pod's memory with hardware-enforced encryption and
          integrity protection, and enables you to authenticate and trust the workload
          using Remote Attestation.\n\n# Data-in-use protection\nWhen protecting sensitive
          data managed by a Pod in OpenShift, one can protect the data at rest by
          encrypting the data written to the persistent storage, and protecting the
          data in transit when sending data over the network.\n\nHowever, when the
          data is used in the Pod's memory, it cannot be both encrypted and used at
          the same time.\nThis means that if a malicious insider (either human or
          software) had access to the OpenShift Node, they could dump the Pod's memory
          and access all of its sensitive data. This includes accessing the secrets
          used to encrypt the data at rest and in transit.\n\nRunning the Pod in a
          Confidential Pod with the Anjuna Seaglass platform solves this problem.
          Anjuna Seaglass runs the Pod inside a Trusted Execution Environment, where
          the Pod memory is encrypted by the hardware so that no privileged human
          or software on the Pod can access this memory in the clear.\n\n# Enabling
          zero trust for the Pod\nWhen running a Pod in OpenShift, the Pod may need
          to create trust with other Pods or with software running outside of the
          OpenShift cluster.\nThe standard way to create this trust is to identify
          with a token or credentials that could be easily read and stolen from the
          Pod disk or by providing a machine identity, a process name, or an IP that
          could be spoofed or reused.\n\nConfidential Computing technology provides
          a mechanism called Remote Attestation, which provides an identity to the
          software via the Hardware—the Software Measurement.\n\nWhen running a Pod
          as a Confidential Pod with the Anjuna Seaglass platform, a user can leverage
          the Remote Attestation to perform trusted authentication with the Confidential
          Pod, in a way that cannot be spoofed.\nThe Anjuna Policy Manager allows
          you to protect the first secret of a Confidential Pod and remove any secret
          from the Pod image or Kubernetes control plane that could be read by malicious
          insiders.\n\n# Quick start and configuration instructions\nFollow the directions
          in Installing the [Anjuna Kubernetes Toolset to OpenShift on GCP](https://docs.anjuna.io/k8s-sev/latest/installation/openshift_gcp.html)
          \nto learn how to deploy the Operator into your OpenShift cluster.\n\nFollow
          the directions in [Quickstart guides for OpenShift on GCP](https://docs.anjuna.io/k8s-sev/latest/applications/gcp.html)
          \nto learn how to deploy your first Confidential Pod in your OpenShift cluster.\n\nFollow
          the directions in [Attestation with Anjuna Policy Manager](https://docs.anjuna.io/k8s-sev/latest/attestation.html)
          \nto learn how to leverage Remote Attestation to protect your Pod's first
          secret."
        displayName: Anjuna Seaglass™ Operator for OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Security
        links:
        - name: Anjuna Operator Documentation
          url: https://docs.anjuna.io/k8s-sev/latest/index.html
        - name: Anjuna Seaglass
          url: https://www.anjuna.io/essence-of-anjuna-seaglass
        - name: Anjuna Security
          url: https://www.anjuna.io
        maintainers:
        - email: support@anjuna.io
          name: Anjuna Security
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Anjuna Security
          url: https://www.anjuna.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - registry.connect.redhat.com/anjuna/anjuna-k8s-sev-tools@sha256:e5ba31c240aaff8ca8f6addf12577fd6045bf6619a1d2cb2987b2ff5d75f4028
        - registry.connect.redhat.com/anjuna/anjuna-k8s-sev-runtime@sha256:1531b1ba955b2d3ea5d9238d2e20617a534bb1bab73f5fadadf8c9d3aee410b4
        version: 2.0.0
      entries:
      - name: anjuna-operator.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anjuna-operator
    provider:
      name: Anjuna Security
      url: https://www.anjuna.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzograph-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzograph-operator.v3.2.500
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzograph.clusters.cambridgesemantics.com/v2",
                "kind": "AnzoGraph",
                "metadata": {
                  "name": "azg01"
                },
                "spec": {
                  "db": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_data": "anzograph-data-grpc",
                            "app_mgmt": "anzograph-mgmt-grpc"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_data": "anzograph-data-grpc",
                              "app_mgmt": "anzograph-mgmt-grpc"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzograph-db@sha256:30ed2ec337135701ab8b54e75bb48ad8928a2c48818643c23b9dba4484255a2e",
                                "name": "db",
                                "resources": {
                                  "limits": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  },
                                  "requests": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  }
                                }
                              }
                            ],
                            "securityContext": {
                              "fsGroup": 1000,
                              "runAsNonRoot": true,
                              "runAsUser": 1000,
                              "seccompProfile": {
                                "type": "RuntimeDefault"
                              }
                            },
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  },
                  "deployFrontend": false,
                  "frontend": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_mgmt": "anzograph-frontend"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_mgmt": "anzograph-frontend"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzograph-frontend@sha256:e5446b3c04889f2189e955702c4165eb243eb89f47ac797eb417b60b1459647f",
                                "name": "frontend",
                                "resources": {
                                  "limits": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "securityContext": {
                              "fsGroup": 1000,
                              "runAsNonRoot": true,
                              "runAsUser": 1000,
                              "seccompProfile": {
                                "type": "RuntimeDefault"
                              }
                            },
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/anzograph-operator@sha256:2333154a14ab7f3a03a601f00c017feeafb9d9814aff5e59c95cef615ab2abd6
          createdAt: "2025-09-22T12:03:18Z"
          description: kubernetes operator for Altair Graph Lakehouse DB
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Altair Engineering Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoGraph is the Schema for the anzographs API
            displayName: Altair Graph Lakehouse Operator
            kind: AnzoGraph
            name: anzographs.anzograph.clusters.cambridgesemantics.com
            version: v2
        description: |-
          The Altair Graph Lakehouse Operator provides the way to install and configure an Altair Graph Lakehouse
          cluster on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v3.2.5/README_openshift_marketplace.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzograph/userdoc/ )

          ### Support

          We offer Support to our customers with the Altair Graph Lakehouse db Enterprise Edition License [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Lakehouse Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        - Database
        links:
        - name: Anzograph Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v3.2.5/README_openshift_marketplace.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzograph-operator@sha256:2333154a14ab7f3a03a601f00c017feeafb9d9814aff5e59c95cef615ab2abd6
        - registry.connect.redhat.com/cambridgesemantics/anzograph-db@sha256:30ed2ec337135701ab8b54e75bb48ad8928a2c48818643c23b9dba4484255a2e
        - registry.connect.redhat.com/cambridgesemantics/anzograph-frontend@sha256:e5446b3c04889f2189e955702c4165eb243eb89f47ac797eb417b60b1459647f
        version: 3.2.500
      entries:
      - name: anzograph-operator.v3.2.500
        version: 3.2.500
      - name: anzograph-operator.v3.2.300
        version: 3.2.300
      - name: anzograph-operator.v3.2.200
        version: 3.2.200
      - name: anzograph-operator.v3.2.104
        version: 3.2.104
      name: stable
    defaultChannel: stable
    packageName: anzograph-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: istio-workspace-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: istio-workspace-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "workspace.maistra.io/v1alpha1",
                "kind": "Session",
                "metadata": {
                  "name": "session-sample"
                },
                "spec": {
                  "ref": [
                    {
                      "args": {
                        "image": "quay.io/maistra-dev/istio-workspace-test:latest"
                      },
                      "name": "ratings-v1",
                      "strategy": "prepared-image"
                    }
                  ],
                  "route": {
                    "name": "x-workspace-route",
                    "type": "header",
                    "value": "feature-x"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/maistra/istio-workspace:v0.5.3
          createdAt: "2022-11-16T13:50:35Z"
          description: Safely develop and test on any Kubernetes cluster without affecting
            others.
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/maistra/istio-workspace
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Session controls the creation of the specialized hidden routes.
            displayName: Session
            kind: Session
            name: sessions.workspace.maistra.io
            version: v1alpha1
        description: "Istio Workspace enables developers to:\n\n * run one or more
          services locally during development but make it appear like it would be
          developed in the actual cluster\n   * this way you can use your favourite
          development tools\n   * have instant feedback loop\n   * don't make your
          machine overheating trying to run the whole cluster locally\n * safely preview
          incoming changes in pull requests without affecting regular users \n * have
          confidence in testing or troubleshooting your services directly in the cluster\n\nIstio
          Workspace supports testing on multi-user environment in an unobtrusive way.\nIt
          doesn’t really matter if it is QE cluster or actual production. We give
          you confidence that your changes won’t blow up the cluster, and your users
          won’t even notice a glitch.\n\nYou will need [Kubernetes](https://k8s.io)
          or [Openshift](https://openshift.com) cluster with  [Istio](https://istio.io/)
          installed. \n\nYou can read more about our vision on [Red Hat Developer’s
          blog](https://developers.redhat.com/blog/2020/07/14/developing-and-testing-on-production-with-kubernetes-and-istio-workspace/)
          or ...\n\n## See it in action!\n\n[![https://youtu.be/XTNVadUzMCc](https://img.youtube.com/vi/XTNVadUzMCc/hqdefault.jpg)](https://youtu.be/XTNVadUzMCc)\n\nWatch
          demo: [\"How to develop on production: An introduction to Istio-Workspaces\"](https://youtu.be/XTNVadUzMCc).\n\n##
          Documentation\n\nMore details can be found on our [documentation page](https://istio-workspace-docs.netlify.com/).\n"
        displayName: Istio Workspace
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - telepresence
        links:
        - name: Istio Workspace Documentation
          url: https://istio-workspace-docs.netlify.com/
        - name: How to install CLI tool
          url: https://github.com/maistra/istio-workspace#install-in-two-easy-steps
        maintainers:
        - email: aslak@redhat.com
          name: Aslak Knutsen
        - email: bartosz@redhat.com
          name: Bartosz Majsak
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/maistra/istio-workspace:v0.5.3
        version: 0.5.3
      entries:
      - name: istio-workspace-operator.v0.5.3
        version: 0.5.3
      - name: istio-workspace-operator.v0.5.2
        version: 0.5.2
      - name: istio-workspace-operator.v0.5.1
        version: 0.5.1
      - name: istio-workspace-operator.v0.5.0
        version: 0.5.0
      - name: istio-workspace-operator.v0.4.0
        version: 0.4.0
      - name: istio-workspace-operator.v0.3.0
        version: 0.3.0
      - name: istio-workspace-operator.v0.2.0
        version: 0.2.0
      - name: istio-workspace-operator.v0.1.0
        version: 0.1.0
      - name: istio-workspace-operator.v0.0.10
        version: 0.0.10
      - name: istio-workspace-operator.v0.0.9
        version: 0.0.9
      - name: istio-workspace-operator.v0.0.8
        version: 0.0.8
      - name: istio-workspace-operator.v0.0.7
        version: 0.0.7
      - name: istio-workspace-operator.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: istio-workspace-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibmcloud-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibmcloud-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"ibmcloud.ibm.com/v1","kind":"Service","metadata":{"annotations":{"ibmcloud.ibm.com/self-healing":"enabled"},"name":"mytranslator"},"spec":{"plan":"lite","serviceClass":"language-translator"}},{"apiVersion":"ibmcloud.ibm.com/v1","kind":"Binding","metadata":{"name":"binding-translator"},"spec":{"parameters":[{"name":"serviceid_crn","value":"crn:v1:bluemix:public:iam-identity::a/ACCOUNT_ID::serviceid:SERVICE_ID"}],"secretName":"translator-secret","serviceName":"mytranslator"}}]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: cloudoperators/ibmcloud-operator:1.1.0
          createdAt: "2022-03-17T04:17:40Z"
          description: The IBM Cloud Operator provides a Kubernetes CRD-Based API
            to manage the lifecycle of IBM public cloud services.
          repository: https://github.com/IBM/cloud-operators
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Binding is an instance of a service binding resource on IBM
              Cloud. A Binding creates a secret with the service instance credentials.
            displayName: Binding
            kind: Binding
            name: bindings.ibmcloud.ibm.com
            version: v1
          - description: Service is an instance of a Service resource on IBM Cloud.
            displayName: Service
            kind: Service
            name: services.ibmcloud.ibm.com
            version: v1
        description: |-
          With the IBM Cloud Operator, you can provision and bind [IBM public cloud services](https://cloud.ibm.com/catalog#services) to your Kubernetes cluster in a Kubernetes-native way. The IBM Cloud Operator is based on the [Kubernetes custom resource definition (CRD) API](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/) so that your applications can create, update, and delete IBM Cloud services from within the cluster by calling Kubnernetes APIs, instead of needing to use several IBM Cloud APIs in addition to configuring your app for Kubernetes.


          ## Features

          * **Service provisioning**: Create any service with any plan that is available in the [IBM Cloud catalog](https://cloud.ibm.com/catalog#services).

          * **Bindings management**: Automatically create Kubernetes secrets in your Kubernetes cluster with the credentials to bind to any provisioned service to the cluster.


          ## Upgrading the operator


          To upgrade an OperatorHub installation, use the OpenShift web console to perform the upgrade. If upgrading from the `alpha` channel subscription to `stable`, follow the additional upgrade instructions below.

          ### Upgrading to version v0.3 or v1.0
          **IMPORTANT NOTICE:** v0.1 and v0.2 used a different naming scheme for secrets and configmaps. Before you update the IBM Cloud Operator, create secret and configmap resources with these names and copy the contents of your previous resources to the new resources. Then, the upgraded operator recognizes and continues to update the resources.

          Existing `Service` and `Binding` resources **do not** need modification for the upgrade.

          | Previous names (v0.1 and v0.2)         | **Current names (v0.3 and v1.0)**             | Description                                                                           |
          |:---------------------------------------|:--------------------------------------------|:--------------------------------------------------------------------------------------|
          | secret-ibm-cloud-operator              | **ibmcloud-operator-secret**                | Secret with the API key, scoped to the namespace.                                     |
          | config-ibm-cloud-operator              | **ibmcloud-operator-defaults**              | ConfigMap with the default values for new resources.                                  |
          | ibm-cloud-operator                     | **ibmcloud-operator-config**                | ConfigMap with the management namespace configuration.                                |
          | ${namespace}-secret-ibm-cloud-operator | **${namespace}-ibmcloud-operator-secret**   | Management namespace Secret with the API key for ${namespace}.                        |
          | ${namespace}-config-ibm-cloud-operator | **${namespace}-ibmcloud-operator-defaults** | Management namespace ConfigMap with default values for new resources in ${namespace}. |

          **Tip:** Forgot to update the secret and configmap names before upgrading? The operator will not take action on Service and Binding resources until the new secrets and configmaps have been created. Creating these after the upgrade will also work.


          ### OperatorHub stable channel
          Continuous updates and bug fixes are provided in the latest stable channel. Subscribing to the `stable` channel in OperatorHub means you automatically get the latest updates without breaking backward compatibility.

          ### Changelog

          For a list of recent changes, see the [releases page](https://github.com/IBM/cloud-operators/releases).


          ## Prerequisites

          1.  Have an [IBM Cloud account](https://cloud.ibm.com/registration).
          2.  Have a cluster that runs Kubernetes version 1.11 or later (OpenShift 3.11 or later).
          3.  Install the required command line tools.
              *   [IBM Cloud CLI](https://cloud.ibm.com/docs/cli?topic=cloud-cli-getting-started) (`ibmcloud`)
              *   [Kubernetes CLI](https://kubernetes.io/docs/tasks/tools/install-kubectl/) (`kubectl`)
          4.  Log in to your IBM Cloud account from the CLI.

              ```bash
              ibmcloud login
              ```

          5.  Target the appropriate resource group (`-g`) and default region (`-r`) for provisioning services:

              ```bash
              ibmcloud target -g default -r us-south
              ```

          6.  If you also want to provision Cloud Foundry services, target the appropriate `org` and `space` (`--cf`).

              ```bash
              ibmcloud target --cf
              ```

          7.  Set the Kubernetes context of your CLI to your cluster so that you can run `kubectl` commands. For example, if your cluster runs OpenShift, use the `oc login` command.


          ## Setting up the operator

          You can use an installation script to set up the IBM Cloud Operator.
          By default, the installation script stores an API key in a Kubernetes secret in your cluster that can be accessed by the IBM Cloud Operator.
          Next, the script sets default values that are used to provision IBM Cloud services, like the resource group and region to provision the services in. Later, you can override any default value in the `Service` custom resource. Finally, the script deploys the operator in your cluster.

          To use your own API key, set the `IBMCLOUD_API_KEY` environment variable to the key before running the installation script:

              export IBMCLOUD_API_KEY="CUSTOM_API_KEY"

          If installed with Operator Hub, the operator will run in the `openshift-operators` namespace. Otherwise, it will run in the `ibmcloud-operator-system` namespace.

          Prefer to create the secrets and defaults yourself? See the [IBM Cloud Operator installation guide](https://github.com/IBM/cloud-operators/blob/master/docs/install.md).

          ### Using a service ID

          By default, the installation script creates an IBM Cloud API key that impersonates your user credentials, to use to set up the IBM Cloud Operator. However, you might want to create a service ID in IBM Cloud Identity and Access Managment (IAM). By using a service ID, you can control access for the IBM Cloud Operator without having the permissions tied to a particular user, such as if that user leaves the company. For more information, see the [IBM Cloud docs](https://cloud.ibm.com/docs/account?topic=account-serviceids).

          1.  Create a service ID in IBM Cloud IAM. If possible, do not use spaces in the names for your IAM credentials. When you use the operator binding feature, any spaces are replaced with underscores.

              ```bash
              ibmcloud iam service-id-create serviceid-ico -d service-ID-for-ibm-cloud-operator
              ```

          2.  Assign the service ID access to the required permissions to work with the IBM Cloud services that you want the operator to manage. The required permissions vary with each IBM Cloud service. You can also scope an access policy to different regions or resource groups. For example, the following command grants the service ID the **Administrator** platform role in the default resource group in the US South (Dallas) region. For more information, see the [IBM Cloud docs](https://cloud.ibm.com/docs/account?topic=account-userroles).

              ```bash
              ibmcloud iam service-policy-create serviceid-ico --roles Administrator --resource-group-name default --region us-south
              ```

          3.  Create an API key for the service ID.

              ```bash
              ibmcloud iam service-api-key-create apikey-ico serviceid-ico -d api-key-for-ibm-cloud-operator
              ```

          4.  Set the API key of the service ID as your CLI environment variable. Now, when you run the installation script, the script uses the service ID's API key. The following command is an example for macOS.

              ```bash
              export IBMCLOUD_API_KEY=<apikey-ico-value>
              ```

          5.  Confirm that the API key environment variable is set in your CLI.

              ```bash
              echo $IBMCLOUD_API_KEY
              ```

          6.  Follow the [prerequisite steps](https://github.com/IBM/cloud-operators/blob/master/README.md#prerequisites) to log in to the IBM Cloud account that owns the service ID.


          ### Installing the operator for OpenShift clusters

          Before you begin, complete the [prerequisite steps](https://github.com/IBM/cloud-operators/blob/master/README.md#prerequisites) to log in to IBM Cloud and your cluster, and optionally set up a [service ID API key](https://github.com/IBM/cloud-operators/blob/master/README.md#using-a-service-id).

          To configure the latest release for OpenShift before installing via the OperatorHub, run the following script:

          *   **Latest release**:

              ```bash
              curl -sL https://raw.githubusercontent.com/IBM/cloud-operators/master/hack/configure-operator.sh | bash
              ```

          *   **Specific release**: Replace `-v 0.0.0` with the specific version that you want to install.

              ```bash
              curl -sL https://raw.githubusercontent.com/IBM/cloud-operators/master/hack/configure-operator.sh | bash -s -- -v 0.0.0 store-creds
              ```


          ## Using the IBM Cloud Operator

          To use the IBM Cloud Operator, create a service instance and then bind the service to your cluster. For more information, see the [sample configuration files](https://github.com/IBM/cloud-operators/blob/master/config/samples), [user guide](https://github.com/IBM/cloud-operators/blob/master/docs/user-guide.md), and [reference documentation](https://github.com/IBM/cloud-operators/blob/master/README.md#reference).

          #### Step 1: Creating a service instance

          1.  To create an instance of an IBM public cloud service, first create a `Service` custom resource file. For more options, see the [Service properties](https://github.com/IBM/cloud-operators/blob/master/README.md#service-properties) reference doc.
              *   `<SERVICE_CLASS>` is the IBM Cloud service that you want to create. To list IBM Cloud services, run `ibmcloud catalog service-marketplace` and use the **Name** value from the output.
              *   `<PLAN>` is the plan for the IBM Cloud service that you want to create, such as `free` or `standard`. To list available plans, run `ibmcloud catalog service <SERVICE_CLASS> | grep plan`.

              ```yaml
              apiVersion: ibmcloud.ibm.com/v1
              kind: Service
              metadata:
                  name: myservice
              spec:
                  plan: <PLAN>
                  serviceClass: <SERVICE_CLASS>
              ```

          2.  Create the service instance in your cluster.

              ```bash
              kubectl apply -f filepath/myservice.yaml
              ```

          3.  Check that your service status is **Online** in your cluster.

              ```bash
              kubectl get services.ibmcloud
              NAME           STATUS   AGE
              myservice      Online   12s
              ```

          4.  Verify that your service instance is created in IBM Cloud.

              ```bash
              ibmcloud resource service-instances | grep myservice
              ```

          #### Step 2: Binding the service instance

          1.  To bind your service to the cluster so that your apps can use the service, create a `Binding` custom resource, where the `serviceName` field is the name of the `Service` custom resource that you previously created. For more options, see [Binding properties](https://github.com/IBM/cloud-operators/blob/master/README.md#binding-properties).

              ```yaml
              apiVersion: ibmcloud.ibm.com/v1
              kind: Binding
              metadata:
                  name: mybinding
              spec:
                  serviceName: myservice
              ```

          2.  Create the binding in your cluster.

              ```bash
              kubectl apply -f filepath/mybinding.yaml
              ```

          3.  Check that your service status is **Online**.

              ```bash
              kubectl get bindings.ibmcloud
              NAME         STATUS   AGE
              mybinding    Online   25s
              ```

          4.  Check that a secret of the same name as your binding is created. The secret contains the service credentials that apps in your cluster can use to access the service.

              ```bash
              kubectl get secrets
              NAME         TYPE      DATA   AGE
              mybinding    Opaque    6      102s
              ```
        displayName: IBM Cloud Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: Source on GitHub
          url: https://github.com/ibm/cloud-operators
        maintainers:
        - email: johnstarich@johnstarich.com
          name: John Starich
        - email: mvaziri@us.ibm.com
          name: vazirim
        - email: greg.lanthier@ca.ibm.com
          name: Greg Lanthier
        - email: halacs@users.noreply.github.com
          name: Halacs
        - email: artberger@users.noreply.github.com
          name: Art
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: IBM
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - cloudoperators/ibmcloud-operator:1.1.0
        version: 1.1.0
      entries:
      - name: ibmcloud-operator.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: ibmcloud-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Infinispan
      provider-url: ""
    name: infinispan
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infinispan-operator.v2.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.2.5.Final
          description: Create and manage Infinispan clusters.
          olm.skipRange: '>=2.1.x <2.2.1'
          operators.operatorframework.io/builder: operator-sdk-v1.3.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/infinispan/operator:2.2.5.Final
        version: 2.2.5
      entries:
      - name: infinispan-operator.v2.2.5
        version: 2.2.5
      - name: infinispan-operator.v2.2.4
        version: 2.2.4
      - name: infinispan-operator.v2.2.3
        version: 2.2.3
      - name: infinispan-operator.v2.2.2
        version: 2.2.2
      - name: infinispan-operator.v2.2.1
        version: 2.2.1
      - name: infinispan-operator.v2.2.0
        version: 2.2.0
      name: 2.2.x
    - currentCSV: infinispan-operator.v2.3.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.3.7.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/infinispan/operator:2.3.7.Final
        version: 2.3.8
      entries:
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: 2.3.x
    - currentCSV: infinispan-operator.v2.4.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.4.17.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/infinispan/operator:2.4.17.Final
        version: 2.4.17
      entries:
      - name: infinispan-operator.v2.4.17
        version: 2.4.17
      - name: infinispan-operator.v2.4.16
        version: 2.4.16
      - name: infinispan-operator.v2.4.15
        version: 2.4.15
      - name: infinispan-operator.v2.4.14
        version: 2.4.14
      - name: infinispan-operator.v2.4.13
        version: 2.4.13
      - name: infinispan-operator.v2.4.12
        version: 2.4.12
      - name: infinispan-operator.v2.4.11
        version: 2.4.11
      - name: infinispan-operator.v2.4.10
        version: 2.4.10
      - name: infinispan-operator.v2.4.9
        version: 2.4.9
      - name: infinispan-operator.v2.4.8
        version: 2.4.8
      - name: infinispan-operator.v2.4.7
        version: 2.4.7
      - name: infinispan-operator.v2.4.6
        version: 2.4.6
      - name: infinispan-operator.v2.4.5
        version: 2.4.5
      - name: infinispan-operator.v2.4.4
        version: 2.4.4
      - name: infinispan-operator.v2.4.3
        version: 2.4.3
      - name: infinispan-operator.v2.4.2
        version: 2.4.2
      - name: infinispan-operator.v2.4.1
        version: 2.4.1
      - name: infinispan-operator.v2.4.0
        version: 2.4.0
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: 2.4.x
    - currentCSV: infinispan-operator.v2.4.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.4.18.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/infinispan/operator:2.4.18.Final
        version: 2.4.18
      entries:
      - name: infinispan-operator.v2.4.18
        version: 2.4.18
      - name: infinispan-operator.v2.4.16
        version: 2.4.16
      - name: infinispan-operator.v2.4.15
        version: 2.4.15
      - name: infinispan-operator.v2.4.14
        version: 2.4.14
      - name: infinispan-operator.v2.4.13
        version: 2.4.13
      - name: infinispan-operator.v2.4.12
        version: 2.4.12
      - name: infinispan-operator.v2.4.11
        version: 2.4.11
      - name: infinispan-operator.v2.4.10
        version: 2.4.10
      - name: infinispan-operator.v2.4.9
        version: 2.4.9
      - name: infinispan-operator.v2.4.8
        version: 2.4.8
      - name: infinispan-operator.v2.4.7
        version: 2.4.7
      - name: infinispan-operator.v2.4.6
        version: 2.4.6
      - name: infinispan-operator.v2.4.5
        version: 2.4.5
      - name: infinispan-operator.v2.4.4
        version: 2.4.4
      - name: infinispan-operator.v2.4.3
        version: 2.4.3
      - name: infinispan-operator.v2.4.2
        version: 2.4.2
      - name: infinispan-operator.v2.4.1
        version: 2.4.1
      - name: infinispan-operator.v2.4.0
        version: 2.4.0
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: stable
    defaultChannel: stable
    packageName: infinispan
    provider:
      name: Infinispan
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent, part of Cisco
      provider-url: https://isovalent.com/
    name: clife
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clife.v1.17.8-cee.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cilium.io/v1alpha1",
                "kind": "CiliumConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "clife"
                  },
                  "name": "ciliumconfig"
                },
                "spec": {
                  "cluster": {
                    "name": "default"
                  },
                  "clusterHealthPort": 9940,
                  "cni": {
                    "binPath": "/var/lib/cni/bin",
                    "chainingMode": "portmap",
                    "confPath": "/var/run/multus/cni/net.d",
                    "exclusive": false
                  },
                  "enterprise": {
                    "featureGate": {
                      "approved": [
                        "CNIChainingMode"
                      ]
                    }
                  },
                  "hubble": {
                    "enabled": true
                  },
                  "ipam": {
                    "mode": "cluster-pool"
                  },
                  "kubeProxyReplacement": false,
                  "operator": {
                    "prometheus": {
                      "enabled": true,
                      "serviceMonitor": {
                        "enabled": true
                      }
                    }
                  },
                  "prometheus": {
                    "enabled": true,
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "securityContext": {
                    "privileged": true
                  },
                  "sessionAffinity": true,
                  "tunnelPort": 4789
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Networking,Security
          createdAt: "2025-09-23T07:55:19Z"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CiliumConfig defines the configuration of Isovalent Networking
              for Kubernetes and all its components
            displayName: CiliumConfig
            kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Networking for Kubernetes, built on Cilium, provides
          eBPF-based networking, observability, and security with optimal scale and
          performance for platform teams operating Kubernetes environments across
          cloud and on-prem infrastructure.
        displayName: Isovalent Networking for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cilium
        - networking
        - observability
        - security
        - eBPF
        - CNI
        links:
        - name: Isovalent Networking for Kubernetes
          url: https://isovalent.com/product/
        maintainers:
        - email: support@isovalent.com
          name: Isovalent, part of Cisco
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Isovalent, part of Cisco
          url: https://isovalent.com/
        relatedImages:
        - quay.io/isovalent/startup-script-ubi:v1.17.8-cee.1@sha256:9a6a413262ec59026345aefc743de847891d80e31ef2f41db954acdcee11c2ff
        - quay.io/isovalent/certgen-ubi:v1.17.8-cee.1@sha256:fb63bed641a279cc09114e93821c7b7555c61393ab594ea2b823ce34fa73335e
        - quay.io/isovalent/cilium-envoy-ubi:v1.17.8-cee.1@sha256:c3de0334f90f9b2540cdec7e2356d813f2df93f0d7d8701005eab7161c7663b5
        - quay.io/isovalent/clife@sha256:ba9bd4a5c35e1d26e395afbc2a4de89f42df1e5e61c9b23959c23194368b7b0e
        - quay.io/isovalent/operator-generic-ubi:v1.17.8-cee.1@sha256:37e4b9f65b2c67d24afeeebebe5d0cb59c09280179de57cd77da2b68e19c3c1b
        - quay.io/isovalent/cilium-ubi:v1.17.8-cee.1@sha256:e9816096e481fad5d95209f902e4a7018aeb65d1e3be1574e6b289e21229f87a
        - quay.io/isovalent/hubble-relay-ubi:v1.17.8-cee.1@sha256:88964d4836d59420dc7e0abb758654ceaf2abcfb53e19bbe2c7962aa68874e37
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.17.8-cee.1@sha256:b097c823638b4fdbe5bdd2b6040a17fef0c70a020e67a18f8231d60eeac02dd5
        version: 1.17.8-cee.1
      entries:
      - name: clife.v1.17.8-cee.1
        version: 1.17.8-cee.1
      - name: clife.v1.17.7-cee.1
        version: 1.17.7-cee.1
      - name: clife.v1.17.5-cee.1
        version: 1.17.5-cee.1
      - name: clife.v1.17.4-cee.1
        version: 1.17.4-cee.1
      - name: clife.v0.0.1
        version: 0.0.1
      name: "1.17"
    defaultChannel: "1.17"
    packageName: clife
    provider:
      name: Isovalent, part of Cisco
      url: https://isovalent.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microsafe
      provider-url: https://microsafe.com.mx/
    name: operator-aims-mobile
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operator-aims-mobile.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.microsafe.com.mx/v1",
                "kind": "AimsMobile",
                "metadata": {
                  "name": "aimsmobile-sample"
                },
                "spec": {
                  "adminApp": {
                    "autoscaling": {
                      "enabled": false
                    },
                    "config": {
                      "NODE_ENV": "production",
                      "VITE_EXPRESS_HOST": "aims-mob-service",
                      "VITE_EXPRESS_PORT": "8080",
                      "VITE_EXPRESS_PROTOCOL": "http",
                      "VITE_KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "VITE_KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "VITE_KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "VITE_VERSION": "1.0.0"
                    },
                    "image": "quay.io/microsafe_app_ban/aims-mob-admin-private@sha256:8e333f900885b0d3f6442c8ffcb43bd93540b8c38ad5bf17cbbd78929e0b1be0",
                    "ingress": {
                      "enabled": true,
                      "isOpenshiftRoute": true
                    },
                    "name": "aims-mob-admin",
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "secretName": "aims-mob-admin-secrets",
                    "service": {
                      "enabled": true,
                      "port": 8083,
                      "targetPort": 8080
                    }
                  },
                  "dbService": {
                    "autoscaling": {
                      "enabled": false
                    },
                    "config": {
                      "MSSQL_HOST": "mssql-database.prod.host",
                      "MSSQL_PORT": "mssql-database.prod.port"
                    },
                    "image": "quay.io/microsafe_app_ban/aims-db-service-private@sha256:bcb4d8624626436526549427f9f169cff714b39cbcc530dd2b3470c6712abaf9",
                    "ingress": {
                      "enabled": false
                    },
                    "name": "aims-db-service",
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "300m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "128Mi"
                      }
                    },
                    "secretName": "aims-db-service-secrets",
                    "service": {
                      "enabled": true,
                      "port": 3000,
                      "targetPort": 3000
                    }
                  },
                  "mainService": {
                    "annotations": {
                      "instrumentation.opentelemetry.io/inject-nodejs": "true"
                    },
                    "autoscaling": {
                      "enabled": true,
                      "maxReplicas": 2,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "config": {
                      "AIMS1_API_URL": "http://aims-db-service:3000",
                      "EXPRESS_PORT": "4000",
                      "FRONTEND_EXPRESS_HOST": "aims-mob-admin",
                      "FRONTEND_EXPRESS_PORT": "4000",
                      "FRONTEND_EXPRESS_PROTOCOL": "http",
                      "FRONTEND_IP": "aims-mob-admin",
                      "FRONTEND_KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "FRONTEND_KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "FRONTEND_KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "FRONTEND_PORT": "80",
                      "FRONTEND_PROTOCOL": "http",
                      "FRONTEND_VERSION": "1.0.0",
                      "KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "MSSQL_HOST": "mssql-database.prod.host",
                      "MSSQL_PORT": "mssql-database.prod.port",
                      "NODE_ENV": "production",
                      "OTLP_LOG_EXPORTER_URL": "opentelemetry-logs.prod.url",
                      "OTLP_METRIC_EXPORTER_URL": "opentelemetry-metrics.prod.url",
                      "OTLP_TRACE_EXPORTER_URL": "opentelemetry-traces.prod.url",
                      "REMEDY_API_URL": "remedy-aims.prod.svc",
                      "REMEDY_USER": "RMDYUSR"
                    },
                    "ingress": {
                      "enabled": true,
                      "isOpenshiftRoute": true
                    },
                    "name": "aims-mob-service",
                    "proxyContainer": {
                      "image": "quay.io/microsafe_app_ban/aims-mob-proxy-private@sha256:29b82de987b28c23f17d5f2176877c5aecea1a0ef182bb3a63612290ff469949",
                      "name": "aims-mob-service-proxy",
                      "resources": {
                        "limits": {
                          "cpu": "2000m",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "replicas": 2,
                    "secretName": "aims-mob-service-secrets",
                    "service": {
                      "enabled": true,
                      "port": 8085,
                      "targetPort": 8081
                    },
                    "serviceContainer": {
                      "image": "quay.io/microsafe_app_ban/aims-mob-service-private@sha256:f9586200106aa4de259f5cae2d22d13dc147a76dfd06e4f4bea09797071dad17",
                      "name": "aims-mob-service-main",
                      "resources": {
                        "limits": {
                          "cpu": "2500m",
                          "memory": "2048Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "128Mi"
                        }
                      }
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": false,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          createdAt: "2025-09-25T19:23:26Z"
          description: Este operador despliega los servicios AIMS Mobile.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/network-aware: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          instrumentation.opentelemetry.io/inject-nodejs: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: Valid Subscription Microsafe
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://quay.io/repository/microsafe_app_ban/operador-aims-mobile-private
          support: Microsafe
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AimsMobile
            name: aimsmobiles.cache.microsafe.com.mx
            version: v1
        description: Este operador despliega la aplicación AIMS Mobile.
        displayName: AIMS Mobile Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aims mobile
        - aims
        links:
        - name: AIMS Mobile Operator
          url: https://microsafe.com.mx/
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Microsafe
          url: https://microsafe.com.mx/
        relatedImages:
        - quay.io/microsafe_app_ban/aims-db-service-private@sha256:bcb4d8624626436526549427f9f169cff714b39cbcc530dd2b3470c6712abaf9
        - quay.io/microsafe_app_ban/aims-mob-admin-private@sha256:8e333f900885b0d3f6442c8ffcb43bd93540b8c38ad5bf17cbbd78929e0b1be0
        - quay.io/microsafe_app_ban/operador-aims-mobile-private@sha256:731f2d731c79ade02f4b6c39b380590833cacddfd795b703cec10ad55f730fad
        - quay.io/microsafe_app_ban/aims-mob-proxy-private@sha256:29b82de987b28c23f17d5f2176877c5aecea1a0ef182bb3a63612290ff469949
        - quay.io/microsafe_app_ban/aims-mob-service-private@sha256:f9586200106aa4de259f5cae2d22d13dc147a76dfd06e4f4bea09797071dad17
        version: 0.0.10
      entries:
      - name: operator-aims-mobile.v0.0.10
        version: 0.0.10
      - name: operator-aims-mobile.v0.0.9
        version: 0.0.9
      - name: operator-aims-mobile.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: operator-aims-mobile
    provider:
      name: Microsafe
      url: https://microsafe.com.mx/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzograph-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzograph-operator.v2.0.102
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzograph.clusters.cambridgesemantics.com/v2",
                "kind": "AnzoGraph",
                "metadata": {
                  "name": "azg01"
                },
                "spec": {
                  "db": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_data": "anzograph-data-grpc",
                            "app_mgmt": "anzograph-mgmt-grpc"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_data": "anzograph-data-grpc",
                              "app_mgmt": "anzograph-mgmt-grpc"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-db@sha256:8a5bd26551990929fdb776d6b421a9713cb84e03b19c18f81b625582f7f41af5",
                                "name": "db",
                                "resources": {
                                  "limits": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  },
                                  "requests": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  },
                  "deployFrontend": false,
                  "frontend": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_mgmt": "anzograph-frontend"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_mgmt": "anzograph-frontend"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-frontend@sha256:c1443bcdc8e998fc699f61980a665f1a083350adf2d3d1d55a7cb55e31ca50b3",
                                "name": "frontend",
                                "resources": {
                                  "limits": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-operator@sha256:a5b212558f748ddc8156541b546290320b656506e039a18903ad9fa6e638e732
          createdAt: "2021-08-11"
          description: kubernetes operator for AnzoGraph DB
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzograph-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzograph-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoGraph is the Schema for the anzographs API
            displayName: AnzoGraph Operator
            kind: AnzoGraph
            name: anzographs.anzograph.clusters.cambridgesemantics.com
            version: v2
        description: |-
          The AnzoGraph Operator provides the way to install and configure an AnzoGraph
          cluster on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v2.0.1/README_openshift_marketplace.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzograph/userdoc/ )

          ### Support

          We offer Support to our customers with the AnzoGraph db Enterprise Edition License [here]( https://customercenter.cambridgesemantics.com/ ). For AnzoGraph db Free Edition questions, get help from our Anzograph User Community at Stack Overflow. When submitting a question, include the tag 'anzograph'.
        displayName: AnzoGraph Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        - Database
        links:
        - name: Anzograph Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v2.0.1/README_openshift_marketplace.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-operator@sha256:a5b212558f748ddc8156541b546290320b656506e039a18903ad9fa6e638e732
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph@sha256:e3bdddf811b5ba69f388d023aea1ba538edd00e9f415c33ec555321d39129a36
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-db@sha256:8a5bd26551990929fdb776d6b421a9713cb84e03b19c18f81b625582f7f41af5
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-frontend@sha256:c1443bcdc8e998fc699f61980a665f1a083350adf2d3d1d55a7cb55e31ca50b3
        version: 2.0.1
      entries:
      - name: anzograph-operator.v2.0.102
        version: 2.0.1
      name: stable
    defaultChannel: stable
    packageName: anzograph-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: kubernetes-nmstate-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubernetes-nmstate-operator.4.13.0-202309261702
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nmstate.io/v1",
                "kind": "NMState",
                "metadata": {
                  "name": "nmstate"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift4/kubernetes-nmstate-rhel8-operator@sha256:bdcb80591b839ec6b4b94efa838ca41585f3044826534413b91efd2acdca4b4d
          createdAt: "2022-02-21 08:46:16"
          description: |
            Kubernetes NMState is a declaritive means of configuring NetworkManager.
          olm.skipRange: '>=4.3.0 <4.13.0'
          operatorframework.io/suggested-namespace: openshift-nmstate
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/kubernetes-nmstate
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an NMState deployment.
            displayName: NMState
            kind: NMState
            name: nmstates.nmstate.io
            version: v1
        description: A Kubernetes Operator to install Kubernetes NMState
        displayName: Kubernetes NMState Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - nmstate
        - networking
        - NetworkManager
        links:
        - name: Kubernetes Nmstate Operator
          url: https://github.com/nmstate/kubernetes-nmstate
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: GA
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/kubernetes-nmstate-rhel8-operator@sha256:bdcb80591b839ec6b4b94efa838ca41585f3044826534413b91efd2acdca4b4d
        - registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel8@sha256:a96434b959c1f61c1cfb4352284aaae453d2e41ca0824d016975a1216451b665
        version: 4.13.0-202309261702
      entries:
      - name: kubernetes-nmstate-operator.4.13.0-202309261702
        version: 4.13.0-202309261702
      name: "4.13"
    - currentCSV: kubernetes-nmstate-operator.4.16.0-202509241326
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nmstate.io/v1",
                "kind": "NMState",
                "metadata": {
                  "name": "nmstate"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:14750ac40ba2a9a92e2b37351eed94c22f2663747ea09f241a3a80607a9fca09
          createdAt: "2022-02-21 08:46:16"
          description: |
            Kubernetes NMState is a declaritive means of configuring NetworkManager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-nmstate
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/kubernetes-nmstate
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an NMState deployment.
            displayName: NMState
            kind: NMState
            name: nmstates.nmstate.io
            version: v1
        description: A Kubernetes Operator to install Kubernetes NMState
        displayName: Kubernetes NMState Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - nmstate
        - networking
        - NetworkManager
        links:
        - name: Kubernetes Nmstate Operator
          url: https://github.com/nmstate/kubernetes-nmstate
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: GA
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:14750ac40ba2a9a92e2b37351eed94c22f2663747ea09f241a3a80607a9fca09
        - registry.redhat.io/openshift4/nmstate-console-plugin-rhel9@sha256:8d03a6c414da30521e0794d3d61336bfbb1d9d4c3fec5ab5dac4bd24c897a540
        - registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel9@sha256:c592fdb65d99de2c3d274d7e72621408e578bdba39753bb425d0765ac016b11a
        version: 4.16.0-202509241326
      entries:
      - name: kubernetes-nmstate-operator.4.16.0-202509241326
        version: 4.16.0-202509241326
      - name: kubernetes-nmstate-operator.4.16.0-202509101329
        version: 4.16.0-202509101329
      - name: kubernetes-nmstate-operator.4.16.0-202508251030
        version: 4.16.0-202508251030
      - name: kubernetes-nmstate-operator.4.16.0-202507221936
        version: 4.16.0-202507221936
      - name: kubernetes-nmstate-operator.4.16.0-202507100305
        version: 4.16.0-202507100305
      - name: kubernetes-nmstate-operator.4.16.0-202506250035
        version: 4.16.0-202506250035
      - name: kubernetes-nmstate-operator.4.16.0-202506041738
        version: 4.16.0-202506041738
      - name: kubernetes-nmstate-operator.4.16.0-202505150035
        version: 4.16.0-202505150035
      - name: kubernetes-nmstate-operator.4.16.0-202505060805
        version: 4.16.0-202505060805
      - name: kubernetes-nmstate-operator.4.16.0-202504141106
        version: 4.16.0-202504141106
      - name: kubernetes-nmstate-operator.4.16.0-202503260136
        version: 4.16.0-202503260136
      - name: kubernetes-nmstate-operator.4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: kubernetes-nmstate-operator.4.16.0-202502190034
        version: 4.16.0-202502190034
      - name: kubernetes-nmstate-operator.4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: kubernetes-nmstate-operator.4.16.0-202501281135
        version: 4.16.0-202501281135
      - name: kubernetes-nmstate-operator.4.16.0-202501211505
        version: 4.16.0-202501211505
      - name: kubernetes-nmstate-operator.4.16.0-202501080105
        version: 4.16.0-202501080105
      - name: kubernetes-nmstate-operator.4.16.0-202412180105
        version: 4.16.0-202412180105
      - name: kubernetes-nmstate-operator.4.16.0-202412110604
        version: 4.16.0-202412110604
      - name: kubernetes-nmstate-operator.4.16.0-202412040032
        version: 4.16.0-202412040032
      - name: kubernetes-nmstate-operator.4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: kubernetes-nmstate-operator.4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: kubernetes-nmstate-operator.4.16.0-202410300036
        version: 4.16.0-202410300036
      - name: kubernetes-nmstate-operator.4.16.0-202410231737
        version: 4.16.0-202410231737
      - name: kubernetes-nmstate-operator.4.16.0-202410180404
        version: 4.16.0-202410180404
      - name: kubernetes-nmstate-operator.4.16.0-202410091137
        version: 4.16.0-202410091137
      - name: kubernetes-nmstate-operator.4.16.0-202410021935
        version: 4.16.0-202410021935
      - name: kubernetes-nmstate-operator.4.16.0-202409251605
        version: 4.16.0-202409251605
      - name: kubernetes-nmstate-operator.4.16.0-202409111235
        version: 4.16.0-202409111235
      - name: kubernetes-nmstate-operator.4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: kubernetes-nmstate-operator.4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: kubernetes-nmstate-operator.4.16.0-202408211407
        version: 4.16.0-202408211407
      - name: kubernetes-nmstate-operator.4.16.0-202408140143
        version: 4.16.0-202408140143
      - name: kubernetes-nmstate-operator.4.16.0-202408070540
        version: 4.16.0-202408070540
      - name: kubernetes-nmstate-operator.4.16.0-202407251436
        version: 4.16.0-202407251436
      - name: kubernetes-nmstate-operator.4.16.0-202407240806
        version: 4.16.0-202407240806
      - name: kubernetes-nmstate-operator.4.16.0-202407181806
        version: 4.16.0-202407181806
      - name: kubernetes-nmstate-operator.4.16.0-202407101306
        version: 4.16.0-202407101306
      - name: kubernetes-nmstate-operator.4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: kubernetes-nmstate-operator.4.16.0-202406260236
        version: 4.16.0-202406260236
      - name: kubernetes-nmstate-operator.4.16.0-202406140306
        version: 4.16.0-202406140306
      - name: kubernetes-nmstate-operator.4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: kubernetes-nmstate-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Infinidat
      provider-url: https://infinidat.com
    name: infinibox-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infinibox-operator-certified.v2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csidriver.infinidat.com/v1alpha1",
                "kind": "InfiniboxCsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infinibox-operator-certified",
                    "app.kubernetes.io/instance": "infiniboxcsidriver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infiniboxcsidriver",
                    "app.kubernetes.io/part-of": "infinibox-operator-certified"
                  },
                  "name": "infiniboxcsidriver-sample"
                },
                "spec": {
                  "Ibox_Creds": [
                    {
                      "SecretName": "infinibox-creds",
                      "hostname": "ibox0000",
                      "inbound_secret": "0.0000000000000",
                      "inbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonout",
                      "outbound_secret": "0.0000000000001",
                      "outbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonin",
                      "password": "my-pool-admin-passwd",
                      "username": "my-pool-admin"
                    }
                  ],
                  "Infinibox_Protocol_Secrets": [
                    {
                      "SecretName": "infinibox-protocol-secret",
                      "iscsi": {
                        "network_space": "default_iscsi_space",
                        "useCHAP": "none"
                      },
                      "nfs": {
                        "network_space": "default_nas_space",
                        "nfs_export_permissions": ""
                      },
                      "nvme": {
                        "network_space": "default_nvme_space"
                      },
                      "preferred_auto_order": "fc,nvme,iscsi",
                      "storage_protocol": "fc"
                    }
                  ],
                  "autoUpdate": false,
                  "cleanupNFSPerms": false,
                  "controllerserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "controllerservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "createEvents": true,
                  "csiDriverName": "infinibox-csi-driver",
                  "csiDriverVersion": "v2.23.0",
                  "e2etesting": false,
                  "images": {
                    "attachersidecar": "registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901",
                    "attachersidecar_pull_policy": "IfNotPresent",
                    "csidriver": "registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61",
                    "csidriver_pull_policy": "Always",
                    "livenesssidecar": "registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc",
                    "livenesssidecar_pull_policy": "IfNotPresent",
                    "provisionersidecar": "registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2",
                    "provisionersidecar_pull_policy": "IfNotPresent",
                    "registrarsidecar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24",
                    "registrarsidecar_pull_policy": "IfNotPresent",
                    "resizersidecar": "registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d",
                    "resizersidecar_pull_policy": "IfNotPresent",
                    "snapshottersidecar": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a",
                    "snapshottersidecar_pull_policy": "IfNotPresent"
                  },
                  "instanceCount": 1,
                  "logLevel": "debug",
                  "nodeSelector": {
                    "kubernetes.io/os": "linux"
                  },
                  "nodeserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "nodeservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "protocolSecret": "",
                  "removeDomainName": false,
                  "replicaCount": 1,
                  "skipCredentialsCreation": false,
                  "volumeNamePrefix": "ibox"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
          createdAt: "2025-09-27T16:12:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2.23.0
          operatorframework.io/suggested-namespace: infinidat-csi
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Infinidat/infinibox-csi-driver
          support: Infinidat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Iboxreplica
            name: iboxreplicas.csidriver.infinidat.com
            version: v1
          - kind: InfiniboxCsiDriver
            name: infiniboxcsidrivers.csidriver.infinidat.com
            version: v1alpha1
        description: "Infinidat InfiniBox Container Storage Interface (CSI) Driver
          is a\n    CNCF-compliant Kubernetes integration for InfiniBox storage systems,
          offering\n    advanced enterprise functionality for petabyte-scale Kubernetes
          deployments including\n    Red Hat OpenShift.\n \n## Features and Benefits\n
          \ \n* **Multi-protocol flexibility** - manage\n    Kubernetes Persistent
          Volumes attached via block and file protocols,\n    including Fibre Channel,
          iSCSI, and NFS, with all Kubernetes PV access modes\n    \n* **Multi-petabyte
          scalability** - support hundreds of thousands of PVs per InfiniBox\n    system
          and control multiple InfiniBox arrays within a single Kubernetes cluster\n
          \   \\ \n* **Advanced enterprise features** - manage native InfiniBox snapshots
          and\n    clones, including restoring from snapshots, and import PVs created
          outside of\n    InfiniBox CSI Driver\n  \n## Required Parameters\n  \n*
          `hostname` - IP address\n    or hostname of the InfiniBox management interface\n*
          `username` / `password` -\n    InfiniBox credentials\n* `SecretName` - secret
          name, to be used in the StorageClass\n    to define a specific InfiniBox
          for persistent volumes\n\n## Optional Parameters\n*\n    `inbound_user`
          / `inbound_secret` / `outbound_user` / `outbound_secret` - credentials \n
          \   for iSCSI CHAP authentication\n\n    \n## Installation Instructions\n
          \n1. Create *infinidat-csi* namespace to install operator into.\n    \n2.
          Delete the following *clusterrolebindings*, if they exist, before installing
          operator\n     * infinidat-csi-operator-infinidat-csi-attacher\n     * infinidat-csi-operator-infinidat-csi-controller
          \                    \n     * infinidat-csi-operator-infinidat-csi-driver
          \                    \n     * infinidat-csi-operator-infinidat-csi-node
          \                              \n     * infinidat-csi-operator-infinidat-csi-provisioner
          \                       \n     * infinidat-csi-operator-infinidat-csi-resizer
          \          \n     * infinidat-csi-operator-infinidat-csi-snapshotter \n
          \   \n3. Install Infinidat Infinibox CSI Driver - Operator (this) into the
          *infinidat-csi* namespace.\n    \n4. Apply the following Security Context
          Constraints to the created service accounts in the *infinidat-csi* \n    namespace\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    \n5. Recreate or edit any existing storage classes, pointing to the
          new namespace\n\n     \n6. If you choose a different namespace other than
          *infinidat-csi* you will need to edit the clusterrolebindings listed above
          to specify your custom namespace\n\n     \n7. Install an instance of Infinidat
          Driver into the *infinidat-csi* namespace"
        displayName: InfiniBox CSI Driver - Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - csi
        - nas
        - san
        - storage
        - infinidat
        - infinibox
        links:
        - name: Infinidat
          url: https://infinidat.com
        maintainers:
        - email: info@infinidat.com
          name: Infinidat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Infinidat
          url: https://infinidat.com
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified@sha256:42ff34a019bc67076515e89735b4bb9921fbd4610c127ac8b61e1fa6a6a08194
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24
        version: 2.23.0
      entries:
      - name: infinibox-operator-certified.v2.23.0
        version: 2.23.0
      - name: infinibox-operator-certified.v2.22.0
        version: 2.22.0
      - name: infinibox-operator-certified.v2.21.0
        version: 2.21.0
      name: alpha
    - currentCSV: infinibox-operator-certified.v2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csidriver.infinidat.com/v1alpha1",
                "kind": "InfiniboxCsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infinibox-operator-certified",
                    "app.kubernetes.io/instance": "infiniboxcsidriver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infiniboxcsidriver",
                    "app.kubernetes.io/part-of": "infinibox-operator-certified"
                  },
                  "name": "infiniboxcsidriver-sample"
                },
                "spec": {
                  "Ibox_Creds": [
                    {
                      "SecretName": "infinibox-creds",
                      "hostname": "ibox0000",
                      "inbound_secret": "0.0000000000000",
                      "inbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonout",
                      "outbound_secret": "0.0000000000001",
                      "outbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonin",
                      "password": "my-pool-admin-passwd",
                      "username": "my-pool-admin"
                    }
                  ],
                  "Infinibox_Protocol_Secrets": [
                    {
                      "SecretName": "infinibox-protocol-secret",
                      "iscsi": {
                        "network_space": "default_iscsi_space",
                        "useCHAP": "none"
                      },
                      "nfs": {
                        "network_space": "default_nas_space",
                        "nfs_export_permissions": ""
                      },
                      "nvme": {
                        "network_space": "default_nvme_space"
                      },
                      "preferred_auto_order": "fc,nvme,iscsi",
                      "storage_protocol": "fc"
                    }
                  ],
                  "autoUpdate": false,
                  "cleanupNFSPerms": false,
                  "controllerserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "controllerservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "createEvents": true,
                  "csiDriverName": "infinibox-csi-driver",
                  "csiDriverVersion": "v2.23.0",
                  "e2etesting": false,
                  "images": {
                    "attachersidecar": "registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901",
                    "attachersidecar_pull_policy": "IfNotPresent",
                    "csidriver": "registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61",
                    "csidriver_pull_policy": "Always",
                    "livenesssidecar": "registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc",
                    "livenesssidecar_pull_policy": "IfNotPresent",
                    "provisionersidecar": "registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2",
                    "provisionersidecar_pull_policy": "IfNotPresent",
                    "registrarsidecar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24",
                    "registrarsidecar_pull_policy": "IfNotPresent",
                    "resizersidecar": "registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d",
                    "resizersidecar_pull_policy": "IfNotPresent",
                    "snapshottersidecar": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a",
                    "snapshottersidecar_pull_policy": "IfNotPresent"
                  },
                  "instanceCount": 1,
                  "logLevel": "debug",
                  "nodeSelector": {
                    "kubernetes.io/os": "linux"
                  },
                  "nodeserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "nodeservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "protocolSecret": "",
                  "removeDomainName": false,
                  "replicaCount": 1,
                  "skipCredentialsCreation": false,
                  "volumeNamePrefix": "ibox"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
          createdAt: "2025-09-27T16:12:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2.23.0
          operatorframework.io/suggested-namespace: infinidat-csi
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Infinidat/infinibox-csi-driver
          support: Infinidat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Iboxreplica
            name: iboxreplicas.csidriver.infinidat.com
            version: v1
          - kind: InfiniboxCsiDriver
            name: infiniboxcsidrivers.csidriver.infinidat.com
            version: v1alpha1
        description: "Infinidat InfiniBox Container Storage Interface (CSI) Driver
          is a\n    CNCF-compliant Kubernetes integration for InfiniBox storage systems,
          offering\n    advanced enterprise functionality for petabyte-scale Kubernetes
          deployments including\n    Red Hat OpenShift.\n \n## Features and Benefits\n
          \ \n* **Multi-protocol flexibility** - manage\n    Kubernetes Persistent
          Volumes attached via block and file protocols,\n    including Fibre Channel,
          iSCSI, and NFS, with all Kubernetes PV access modes\n    \n* **Multi-petabyte
          scalability** - support hundreds of thousands of PVs per InfiniBox\n    system
          and control multiple InfiniBox arrays within a single Kubernetes cluster\n
          \   \\ \n* **Advanced enterprise features** - manage native InfiniBox snapshots
          and\n    clones, including restoring from snapshots, and import PVs created
          outside of\n    InfiniBox CSI Driver\n  \n## Required Parameters\n  \n*
          `hostname` - IP address\n    or hostname of the InfiniBox management interface\n*
          `username` / `password` -\n    InfiniBox credentials\n* `SecretName` - secret
          name, to be used in the StorageClass\n    to define a specific InfiniBox
          for persistent volumes\n\n## Optional Parameters\n*\n    `inbound_user`
          / `inbound_secret` / `outbound_user` / `outbound_secret` - credentials \n
          \   for iSCSI CHAP authentication\n\n    \n## Installation Instructions\n
          \n1. Create *infinidat-csi* namespace to install operator into.\n    \n2.
          Delete the following *clusterrolebindings*, if they exist, before installing
          operator\n     * infinidat-csi-operator-infinidat-csi-attacher\n     * infinidat-csi-operator-infinidat-csi-controller
          \                    \n     * infinidat-csi-operator-infinidat-csi-driver
          \                    \n     * infinidat-csi-operator-infinidat-csi-node
          \                              \n     * infinidat-csi-operator-infinidat-csi-provisioner
          \                       \n     * infinidat-csi-operator-infinidat-csi-resizer
          \          \n     * infinidat-csi-operator-infinidat-csi-snapshotter \n
          \   \n3. Install Infinidat Infinibox CSI Driver - Operator (this) into the
          *infinidat-csi* namespace.\n    \n4. Apply the following Security Context
          Constraints to the created service accounts in the *infinidat-csi* \n    namespace\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    \n5. Recreate or edit any existing storage classes, pointing to the
          new namespace\n\n     \n6. If you choose a different namespace other than
          *infinidat-csi* you will need to edit the clusterrolebindings listed above
          to specify your custom namespace\n\n     \n7. Install an instance of Infinidat
          Driver into the *infinidat-csi* namespace"
        displayName: InfiniBox CSI Driver - Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - csi
        - nas
        - san
        - storage
        - infinidat
        - infinibox
        links:
        - name: Infinidat
          url: https://infinidat.com
        maintainers:
        - email: info@infinidat.com
          name: Infinidat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Infinidat
          url: https://infinidat.com
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified@sha256:42ff34a019bc67076515e89735b4bb9921fbd4610c127ac8b61e1fa6a6a08194
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24
        version: 2.23.0
      entries:
      - name: infinibox-operator-certified.v2.23.0
        version: 2.23.0
      - name: infinibox-operator-certified.v2.22.0
        version: 2.22.0
      - name: infinibox-operator-certified.v2.21.0
        version: 2.21.0
      name: stable
    defaultChannel: stable
    packageName: infinibox-operator-certified
    provider:
      name: Infinidat
      url: https://infinidat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Shipwright Contributors
      provider-url: https://shipwright.io
    name: shipwright-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: shipwright-operator.v0.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "shipwright-build"
                },
                "spec": {
                  "targetNamespace": "shipwright-build"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: ghcr.io/shipwright-io/operator/operator:0.13.0@sha256:12cd5f936043c96f31cf571d7a4f8b6f2e240852cd9ad1a5423dbab0c208b542
          description: Shipwright is a framework for building container images on
            Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/shipwright-io/operator
          support: The Shipwright Contributors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
        description: |
          Shipwright is a framework for building container images on Kubernetes.

          Read more: [https://shipwright.io](https://shipwright.io)

          ## Usage

          To deploy and manage [Shipwright Builds](https://github.com/shipwright-io/build) in your cluster,
          first make sure this operator is installed and running on your cluster.

          Next, create the following:

          ```yaml
          ---
          apiVersion: operator.shipwright.io/v1alpha1
          kind: ShipwrightBuild
          metadata:
            name: shipwright-operator
          spec:
            targetNamespace: shipwright-build
          ```

          The operator will deploy Shipwright Builds in the provided `targetNamespace`.
          When `.spec.targetNamespace` is not set, the namespace will default to `shipwright-build`.
          Refer to the [ShipwrightBuild documentation](https://github.com/shipwright-io/operator/blob/main/docs/shipwrightbuild.md) for more information about this custom resource.
        displayName: Shipwright Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://shipwright.io
        - name: Shipwright Builds
          url: https://github.com/shipwright-io/build
        maintainers:
        - email: shipwright-dev@lists.shipwright.io
          name: The Shipwright Contributors
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Shipwright Contributors
          url: https://shipwright.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - ghcr.io/shipwright-io/operator/operator:0.13.0@sha256:12cd5f936043c96f31cf571d7a4f8b6f2e240852cd9ad1a5423dbab0c208b542
        version: 0.13.0
      entries:
      - name: shipwright-operator.v0.13.0
        version: 0.13.0
      - name: shipwright-operator.v0.9.0
        version: 0.9.0
      name: alpha
    defaultChannel: alpha
    packageName: shipwright-operator
    provider:
      name: The Shipwright Contributors
      url: https://shipwright.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ThinkParQ
      provider-url: https://www.thinkparq.com
    name: beegfs-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: beegfs-csi-driver-operator.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "beegfs.csi.netapp.com/v1",
                "kind": "BeegfsDriver",
                "metadata": {
                  "name": "csi-beegfs-cr"
                },
                "spec": {
                  "nodeAffinityControllerService": {
                    "preferredDuringSchedulingIgnoredDuringExecution": [
                      {
                        "preference": {
                          "matchExpressions": [
                            {
                              "key": "node-role.kubernetes.io/master",
                              "operator": "Exists"
                            }
                          ]
                        },
                        "weight": 50
                      }
                    ],
                    "requiredDuringSchedulingIgnoredDuringExecution": {
                      "nodeSelectorTerms": [
                        {
                          "matchExpressions": [
                            {
                              "key": "node.openshift.io/os_id",
                              "operator": "NotIn",
                              "values": [
                                "rhcos"
                              ]
                            }
                          ]
                        }
                      ]
                    }
                  },
                  "nodeAffinityNodeService": {
                    "requiredDuringSchedulingIgnoredDuringExecution": {
                      "nodeSelectorTerms": [
                        {
                          "matchExpressions": [
                            {
                              "key": "node.openshift.io/os_id",
                              "operator": "NotIn",
                              "values": [
                                "rhcos"
                              ]
                            }
                          ]
                        }
                      ]
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: ghcr.io/thinkparq/beegfs-csi-driver-operator:v1.7.0
          createdAt: 2021-06-30:03:11:00
          description: Install and manage the BeeGFS CSI driver
          operatorframework.io/suggested-namespace: beegfs-csi
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ThinkParQ/beegfs-csi-driver
          support: ThinkParQ
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys the BeeGFS CSI driver
            displayName: BeeGFS Driver
            kind: BeegfsDriver
            name: beegfsdrivers.beegfs.csi.netapp.com
            version: v1
        description: |
          The BeeGFS Container Storage Interface (CSI) driver provides high performing and scalable storage for workloads
          running in container orchestrators like Kubernetes. It allows containers to access existing datasets or request
          on-demand ephemeral or persistent high speed storage backed by BeeGFS parallel file systems.

          This operator eases the installation and management of the BeeGFS CSI driver in an OpenShift or other Operator
          Lifecycle Manager enabled cluster.

          ### Documentation

          Compatibility information and a quick start guide for the BeeGFS CSI driver itself are available on our [main GitHub
          page](https://www.github.com/ThinkParQ/beegfs-csi-driver). Additional detailed documentation for the driver is
          available in the [docs directory](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/docs) and for the
          operator is available in the [operator
          README](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/operator/README.md).

          ### Important Notes

          The BeeGFS client packages and kernel module must be pre-installed on all nodes that will run the driver. See the
          driver ["Getting Started" guide](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/README.md#getting-started)
          for additional information.

          The BeeGFS CSI driver is only supported on nodes running a [BeeGFS supported Linux
          distribution](https://doc.beegfs.io/latest/release_notes.html#supported-linux-distributions-and-kernels). Red Hat
          CoreOS (RHCOS), the default distribution in OpenShift environments, is NOT supported. However, Red Hat Enterprise
          Linux (RHEL) nodes can be added to an OpenShift cluster and RHEL is supported for running BeeGFS. By default, the
          operator will install the driver with a node affinity that ensures it does not run on RHCOS nodes. See the driver
          [compatibility
          matrix](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/README.md#interoperability-and-csi-feature-matrix)
          for additional information.

          We have recently published experimental manifests that enable the BeeGFS CSI driver on OpenShift nodes running
          RHCOS. These manifests are not intended for production use. See [this
          README](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/deploy/openshift-beegfs-client/README.md) for more
          details.

          All driver configuration can be specified through the BeegfsDrivers API (in a BeegfsDriver Custom Resource)
          except for BeeGFS connAuth information, which must be placed separately into a Kubernetes Secret. See the [ConnAuth
          Configuration
          section](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/operator/README.md#connauth-configuration) of the
          operator documentation for additional information.

          Starting with version 1.4.0 of the BeeGFS CSI driver the driver and operator images will be signed using the Cosign
          tool. Refer to the [BeeGFS CSI driver operator
          README](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/operator/README.md)
          for details on how to verify the operator and driver image signatures.
        displayName: BeeGFS CSI Driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CSI
        - BeeGFS
        - ThinkParQ
        - HPC
        - AI
        - High-Performance Computing
        links:
        - name: BeeGFS introduction
          url: https://www.beegfs.io/c/
        - name: BeeGFS documentation
          url: https://doc.beegfs.io/latest/index.html
        - name: BeeGFS CSI driver source code
          url: https://github.com/ThinkParQ/beegfs-csi-driver
        - name: BeeGFS blog posts
          url: https://www.beegfs.io/c/blog/
        maintainers:
        - email: csi-driver@beegfs.io
          name: ThinkParQ
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: ThinkParQ
          url: https://www.thinkparq.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - ghcr.io/thinkparq/beegfs-csi-driver-operator:v1.7.0
        version: 1.7.0
      entries:
      - name: beegfs-csi-driver-operator.v1.7.0
        version: 1.7.0
      - name: beegfs-csi-driver-operator.v1.6.0
        version: 1.6.0
      - name: beegfs-csi-driver-operator.v1.5.0
        version: 1.5.0
      - name: beegfs-csi-driver-operator.v1.4.0
        version: 1.4.0
      - name: beegfs-csi-driver-operator.v1.3.0
        version: 1.3.0
      - name: beegfs-csi-driver-operator.v1.2.2
        version: 1.2.2
      - name: beegfs-csi-driver-operator.v1.2.1
        version: 1.2.1
      - name: beegfs-csi-driver-operator.v1.2.0
        version: 1.2.0
      name: stable
    defaultChannel: stable
    packageName: beegfs-csi-driver-operator
    provider:
      name: ThinkParQ
      url: https://www.thinkparq.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stackwatch
      provider-url: www.kubecost.com
    name: kubecost
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubecost.v2.2.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kubecost.com/v1alpha1",
                "kind": "CostAnalyzer",
                "metadata": {
                  "name": "costanalyzer-sample"
                },
                "spec": {
                  "global": {
                    "prometheus": {
                      "enabled": true,
                      "fqdn": "http://cost-analyzer-prometheus-server.default.svc"
                    },
                    "grafana": {
                      "enabled": false,
                      "domainName": "cost-analyzer-grafana.default.svc",
                      "scheme": "http",
                      "proxy": false
                    },
                    "gcpstore": {
                      "enabled": false
                    },
                    "gmp": {
                      "enabled": false,
                      "prometheusServerEndpoint": "http://localhost:8085/",
                      "gmpProxy": {
                        "enabled": false,
                        "image": "gke.gcr.io/prometheus-engine/frontend:v0.4.1-gke.0",
                        "imagePullPolicy": "Always",
                        "name": "gmp-proxy",
                        "port": 8085,
                        "projectId": "YOUR_PROJECT_ID"
                      }
                    },
                    "amp": {
                      "enabled": false,
                      "prometheusServerEndpoint": "http://localhost:8005/workspaces/<workspaceId>/",
                      "remoteWriteService": "https://aps-workspaces.us-west-2.amazonaws.com/workspaces/<workspaceId>/api/v1/remote_write",
                      "sigv4": {
                        "region": "us-west-2"
                      }
                    },
                    "mimirProxy": {
                      "enabled": false,
                      "name": "mimir-proxy",
                      "image": "nginxinc/nginx-unprivileged",
                      "port": 8085,
                      "mimirEndpoint": "$mimir_endpoint",
                      "orgIdentifier": "$your_tenant_ID"
                    },
                    "notifications": {
                      "alertmanager": {
                        "enabled": false,
                        "fqdn": "http://cost-analyzer-prometheus-server.default.svc"
                      }
                    },
                    "savedReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Saved Report 0",
                          "window": "today",
                          "aggregateBy": "namespace",
                          "chartDisplay": "category",
                          "idle": "separate",
                          "rate": "cumulative",
                          "accumulate": false,
                          "filters": [
                            {
                              "key": "cluster",
                              "operator": ":",
                              "value": "dev"
                            }
                          ]
                        },
                        {
                          "title": "Example Saved Report 1",
                          "window": "month",
                          "aggregateBy": "controllerKind",
                          "chartDisplay": "category",
                          "idle": "share",
                          "rate": "monthly",
                          "accumulate": false,
                          "filters": [
                            {
                              "key": "namespace",
                              "operator": "!:",
                              "value": "kubecost"
                            }
                          ]
                        },
                        {
                          "title": "Example Saved Report 2",
                          "window": "2020-11-11T00:00:00Z,2020-12-09T23:59:59Z",
                          "aggregateBy": "service",
                          "chartDisplay": "category",
                          "idle": "hide",
                          "rate": "daily",
                          "accumulate": true,
                          "filters": []
                        }
                      ]
                    },
                    "assetReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Asset Report 0",
                          "window": "today",
                          "aggregateBy": "type",
                          "accumulate": false,
                          "filters": [
                            {
                              "property": "cluster",
                              "value": "cluster-one"
                            }
                          ]
                        }
                      ]
                    },
                    "advancedReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Advanced Report 0",
                          "window": "7d",
                          "aggregateBy": "namespace",
                          "filters": [
                            {
                              "key": "cluster",
                              "operator": ":",
                              "value": "dev"
                            }
                          ],
                          "cloudBreakdown": "service",
                          "cloudJoin": "label:kubernetes_namespace"
                        }
                      ]
                    },
                    "cloudCostReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Cloud Cost Report 0",
                          "window": "today",
                          "aggregateBy": "service",
                          "accumulate": false
                        }
                      ]
                    },
                    "podAnnotations": {},
                    "additionalLabels": {},
                    "securityContext": {
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      },
                      "fsGroup": 1001,
                      "runAsGroup": 1001,
                      "runAsUser": 1001,
                      "fsGroupChangePolicy": "OnRootMismatch"
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "privileged": false,
                      "readOnlyRootFilesystem": true,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "platforms": {
                      "openshift": {
                        "enabled": true,
                        "route": {
                          "enabled": false,
                          "annotations": {}
                        },
                        "scc": {
                          "nodeExporter": false,
                          "networkCosts": false
                        },
                        "securityContext": {
                          "runAsNonRoot": true,
                          "seccompProfile": {
                            "type": "RuntimeDefault"
                          }
                        }
                      },
                      "cicd": {
                        "enabled": true,
                        "skipSanityChecks": true
                      }
                    }
                  },
                  "upgrade": {
                    "toV2": false
                  },
                  "kubecostToken": null,
                  "pricingCsv": {
                    "enabled": false,
                    "location": {
                      "provider": "AWS",
                      "region": "us-east-1",
                      "URI": "s3://kc-csv-test/pricing_schema.csv",
                      "csvAccessCredentials": "pricing-schema-access-secret"
                    }
                  },
                  "saml": {
                    "enabled": false,
                    "rbac": {
                      "enabled": false
                    }
                  },
                  "oidc": {
                    "enabled": false,
                    "clientID": "",
                    "clientSecret": "",
                    "existingCustomSecret": {
                      "enabled": false,
                      "name": ""
                    },
                    "skipOnlineTokenValidation": false,
                    "rbac": {
                      "enabled": false
                    }
                  },
                  "systemProxy": {
                    "enabled": false,
                    "httpProxyUrl": "",
                    "httpsProxyUrl": "",
                    "noProxy": ""
                  },
                  "kubecostFrontend": {
                    "enabled": true,
                    "deployMethod": "singlepod",
                    "haReplicas": 2,
                    "image": "gcr.io/kubecost1/frontend",
                    "imagePullPolicy": "Always",
                    "resources": {
                      "requests": {
                        "cpu": "10m",
                        "memory": "55Mi"
                      }
                    },
                    "deploymentStrategy": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 1,
                      "periodSeconds": 5,
                      "failureThreshold": 6
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 1,
                      "periodSeconds": 5,
                      "failureThreshold": 6
                    },
                    "ipv6": {
                      "enabled": true
                    },
                    "useDefaultFqdn": false
                  },
                  "kubecostMetrics": {
                    "exporter": {
                      "enabled": false,
                      "port": 9005,
                      "prometheusScrape": true,
                      "resources": {},
                      "tolerations": [],
                      "affinity": {},
                      "service": {
                        "annotations": {}
                      },
                      "serviceMonitor": {
                        "enabled": false,
                        "additionalLabels": {},
                        "metricRelabelings": [],
                        "relabelings": []
                      },
                      "priorityClassName": "",
                      "additionalLabels": {},
                      "nodeSelector": {},
                      "extraArgs": []
                    }
                  },
                  "sigV4Proxy": {
                    "image": "public.ecr.aws/aws-observability/aws-sigv4-proxy:latest",
                    "imagePullPolicy": "Always",
                    "name": "aps",
                    "port": 8005,
                    "region": "us-west-2",
                    "host": "aps-workspaces.us-west-2.amazonaws.com",
                    "extraEnv": null,
                    "resources": {}
                  },
                  "kubecostModel": {
                    "image": "gcr.io/kubecost1/cost-model",
                    "imagePullPolicy": "Always",
                    "outOfClusterPromMetricsEnabled": false,
                    "warmCache": false,
                    "etl": true,
                    "etlFileStoreEnabled": true,
                    "etlDailyStoreDurationDays": 91,
                    "etlHourlyStoreDurationHours": 49,
                    "etlWeeklyStoreDurationWeeks": 53,
                    "etlReadOnlyMode": false,
                    "plugins": {
                      "enabled": false,
                      "install": {
                        "enabled": false,
                        "fullImageName": "curlimages/curl:latest",
                        "securityContext": {
                          "allowPrivilegeEscalation": false,
                          "seccompProfile": {
                            "type": "RuntimeDefault"
                          },
                          "capabilities": {
                            "drop": [
                              "ALL"
                            ]
                          },
                          "readOnlyRootFilesystem": true,
                          "runAsNonRoot": true,
                          "runAsUser": 1001
                        }
                      },
                      "folder": "/opt/opencost/plugin",
                      "enabledPlugins": [],
                      "configSecret": "kubecost-plugin-secret"
                    },
                    "cloudCost": {
                      "labelList": {
                        "IsIncludeList": false,
                        "labels": ""
                      },
                      "topNItems": 1000
                    },
                    "allocation": null,
                    "containerStatsEnabled": true,
                    "maxQueryConcurrency": 5,
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "55Mi"
                      }
                    },
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "extraArgs": [],
                    "ingress": {
                      "enabled": false,
                      "labels": null,
                      "annotations": null,
                      "paths": [
                        "/"
                      ],
                      "pathType": "ImplementationSpecific",
                      "hosts": [
                        "cost-analyzer-model.local"
                      ],
                      "tls": []
                    },
                    "utcOffset": "+00:00",
                    "extraPorts": []
                  },
                  "etlUtils": {
                    "enabled": false,
                    "fullImageName": null,
                    "resources": {},
                    "env": {},
                    "nodeSelector": {},
                    "tolerations": {},
                    "affinity": {}
                  },
                  "ingress": {
                    "enabled": false,
                    "labels": null,
                    "annotations": null,
                    "paths": [
                      "/"
                    ],
                    "pathType": "ImplementationSpecific",
                    "hosts": [
                      "cost-analyzer.local"
                    ],
                    "tls": []
                  },
                  "nodeSelector": {},
                  "tolerations": [],
                  "affinity": {},
                  "topologySpreadConstraints": [],
                  "priority": {
                    "enabled": false,
                    "name": ""
                  },
                  "networkPolicy": {
                    "enabled": false,
                    "denyEgress": true,
                    "sameNamespace": true,
                    "costAnalyzer": {
                      "enabled": false,
                      "annotations": {},
                      "additionalLabels": {}
                    }
                  },
                  "extraVolumes": [],
                  "extraVolumeMounts": [],
                  "persistentVolume": {
                    "size": "32Gi",
                    "dbSize": "32.0Gi",
                    "enabled": true,
                    "labels": {},
                    "annotations": {},
                    "dbPVEnabled": false
                  },
                  "service": {
                    "type": "ClusterIP",
                    "port": 9090,
                    "targetPort": 9090,
                    "nodePort": {},
                    "labels": {},
                    "annotations": {},
                    "sessionAffinity": {
                      "enabled": false,
                      "timeoutSeconds": 10800
                    }
                  },
                  "prometheus": {
                    "rbac": {
                      "create": true
                    },
                    "serviceAccounts": {
                      "alertmanager": {
                        "create": true,
                        "name": null
                      },
                      "nodeExporter": {
                        "create": true,
                        "name": null
                      },
                      "pushgateway": {
                        "create": true,
                        "name": null
                      },
                      "server": {
                        "create": true,
                        "name": null,
                        "annotations": {}
                      }
                    },
                    "imagePullSecrets": null,
                    "extraScrapeConfigs": "- job_name: kubecost\n  honor_labels: true\n  scrape_interval: 1m\n  scrape_timeout: 60s\n  metrics_path: /metrics\n  scheme: http\n  dns_sd_configs:\n  - names:\n    - {{ template \"cost-analyzer.serviceName\" . }}\n    type: 'A'\n    port: 9003\n- job_name: kubecost-networking\n  kubernetes_sd_configs:\n    - role: pod\n  relabel_configs:\n  # Scrape only the the targets matching the following metadata\n    - source_labels: [__meta_kubernetes_pod_label_app_kubernetes_io_instance]\n      action: keep\n      regex:  kubecost\n    - source_labels: [__meta_kubernetes_pod_label_app_kubernetes_io_name]\n      action: keep\n      regex:  network-costs\n",
                    "server": {
                      "enabled": true,
                      "name": "server",
                      "sidecarContainers": null,
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "image": {
                        "repository": "quay.io/prometheus/prometheus",
                        "tag": "v2.50.1",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "prefixURL": "",
                      "baseURL": "",
                      "env": [],
                      "extraFlags": [
                        "web.enable-lifecycle"
                      ],
                      "configPath": "/etc/config/prometheus.yml",
                      "global": {
                        "scrape_interval": "1m",
                        "scrape_timeout": "60s",
                        "evaluation_interval": "1m",
                        "external_labels": {
                          "cluster_id": "cluster-one"
                        }
                      },
                      "remoteWrite": {},
                      "remoteRead": {},
                      "extraArgs": {
                        "query.max-concurrency": 1,
                        "query.max-samples": 100000000
                      },
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "extraHostPathMounts": [],
                      "extraConfigmapMounts": [],
                      "extraSecretMounts": [],
                      "configMapOverrideName": "",
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "extraLabels": {},
                        "hosts": [],
                        "pathType": "Prefix",
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "affinity": {},
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "32Gi",
                        "subPath": ""
                      },
                      "emptyDir": {
                        "sizeLimit": ""
                      },
                      "podAnnotations": {},
                      "deploymentAnnotations": {},
                      "podLabels": {},
                      "alertmanagers": [],
                      "replicaCount": 1,
                      "statefulSet": {
                        "enabled": false,
                        "annotations": {},
                        "labels": {},
                        "podManagementPolicy": "OrderedReady",
                        "headless": {
                          "annotations": {},
                          "labels": {},
                          "servicePort": 80
                        }
                      },
                      "readinessProbeInitialDelay": 5,
                      "readinessProbeTimeout": 3,
                      "readinessProbeFailureThreshold": 3,
                      "readinessProbeSuccessThreshold": 1,
                      "livenessProbeInitialDelay": 5,
                      "livenessProbeTimeout": 3,
                      "livenessProbeFailureThreshold": 3,
                      "livenessProbeSuccessThreshold": 1,
                      "resources": {},
                      "verticalAutoscaler": {
                        "enabled": false
                      },
                      "securityContext": {},
                      "containerSecurityContext": {},
                      "service": {
                        "annotations": {},
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 80,
                        "sessionAffinity": "None",
                        "type": "ClusterIP",
                        "gRPC": {
                          "enabled": false,
                          "servicePort": 10901
                        },
                        "statefulsetReplica": {
                          "enabled": false,
                          "replica": 0
                        }
                      },
                      "terminationGracePeriodSeconds": 300,
                      "retention": "97h"
                    },
                    "alertmanager": {
                      "enabled": false,
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "name": "alertmanager",
                      "image": {
                        "repository": "quay.io/prometheus/alertmanager",
                        "tag": "v0.27.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "extraArgs": {},
                      "prefixURL": "",
                      "baseURL": "http://localhost:9093",
                      "extraEnv": {},
                      "extraSecretMounts": [],
                      "configMapOverrideName": "",
                      "configFromSecret": "",
                      "configFileName": "alertmanager.yml",
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "extraLabels": {},
                        "hosts": [],
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "affinity": {},
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "2Gi",
                        "subPath": ""
                      },
                      "podAnnotations": {},
                      "podLabels": {},
                      "replicaCount": 1,
                      "statefulSet": {
                        "enabled": false,
                        "podManagementPolicy": "OrderedReady",
                        "headless": {
                          "annotations": {},
                          "labels": {},
                          "servicePort": 80
                        }
                      },
                      "resources": {},
                      "securityContext": {
                        "runAsUser": 1001,
                        "runAsNonRoot": true,
                        "runAsGroup": 1001,
                        "fsGroup": 1001
                      },
                      "service": {
                        "annotations": {},
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 80,
                        "sessionAffinity": "None",
                        "type": "ClusterIP"
                      }
                    },
                    "alertmanagerFiles": {
                      "alertmanager.yml": {
                        "global": {},
                        "receivers": [
                          {
                            "name": "default-receiver"
                          }
                        ],
                        "route": {
                          "group_wait": "10s",
                          "group_interval": "5m",
                          "receiver": "default-receiver",
                          "repeat_interval": "3h"
                        }
                      }
                    },
                    "configmapReload": {
                      "prometheus": {
                        "enabled": false,
                        "name": "configmap-reload",
                        "image": {
                          "repository": "quay.io/prometheus-operator/prometheus-config-reloader",
                          "tag": "v0.72.0",
                          "pullPolicy": "IfNotPresent"
                        },
                        "extraArgs": {},
                        "extraVolumeDirs": [],
                        "extraConfigmapMounts": [],
                        "resources": {},
                        "containerSecurityContext": {}
                      },
                      "alertmanager": {
                        "enabled": false,
                        "name": "configmap-reload",
                        "image": {
                          "repository": "quay.io/prometheus-operator/prometheus-config-reloader",
                          "tag": "v0.72.0",
                          "pullPolicy": "IfNotPresent"
                        },
                        "extraArgs": {},
                        "extraVolumeDirs": [],
                        "extraConfigmapMounts": [],
                        "resources": {}
                      }
                    },
                    "nodeExporter": {
                      "enabled": false,
                      "hostNetwork": true,
                      "hostPID": true,
                      "dnsPolicy": "ClusterFirstWithHostNet",
                      "name": "node-exporter",
                      "image": {
                        "repository": "prom/node-exporter",
                        "tag": "v1.7.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "extraArgs": {},
                      "extraHostPathMounts": [],
                      "extraConfigmapMounts": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "deploymentAnnotations": {},
                      "pod": {
                        "labels": {}
                      },
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "resources": {},
                      "securityContext": {},
                      "service": {
                        "annotations": {
                          "prometheus.io/scrape": "true"
                        },
                        "labels": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "hostPort": 9100,
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 9100,
                        "type": "ClusterIP"
                      }
                    },
                    "pushgateway": {
                      "enabled": false,
                      "name": "pushgateway",
                      "image": {
                        "repository": "prom/pushgateway",
                        "tag": "v1.6.2",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "extraArgs": {},
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "hosts": [],
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "replicaCount": 1,
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "resources": {},
                      "securityContext": {
                        "runAsUser": 1001,
                        "runAsNonRoot": true
                      },
                      "service": {
                        "annotations": {
                          "prometheus.io/probe": "pushgateway"
                        },
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 9091,
                        "type": "ClusterIP"
                      },
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "2Gi",
                        "subPath": ""
                      }
                    },
                    "serverFiles": {
                      "alerting_rules.yml": {},
                      "alerts": {},
                      "recording_rules.yml": {},
                      "prometheus.yml": {
                        "rule_files": [
                          "/etc/config/recording_rules.yml",
                          "/etc/config/alerting_rules.yml",
                          "/etc/config/rules",
                          "/etc/config/alerts"
                        ],
                        "scrape_configs": [
                          {
                            "job_name": "prometheus",
                            "static_configs": [
                              {
                                "targets": [
                                  "localhost:9090"
                                ]
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-nodes-cadvisor",
                            "scheme": "https",
                            "tls_config": {
                              "ca_file": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                              "insecure_skip_verify": true
                            },
                            "bearer_token_file": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                            "kubernetes_sd_configs": [
                              {
                                "role": "node"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_node_label_(.+)"
                              },
                              {
                                "target_label": "__address__",
                                "replacement": "kubernetes.default.svc:443"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_node_name"
                                ],
                                "regex": "(.+)",
                                "target_label": "__metrics_path__",
                                "replacement": "/api/v1/nodes/$1/proxy/metrics/cadvisor"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(container_cpu_usage_seconds_total|container_memory_working_set_bytes|container_network_receive_errors_total|container_network_transmit_errors_total|container_network_receive_packets_dropped_total|container_network_transmit_packets_dropped_total|container_memory_usage_bytes|container_cpu_cfs_throttled_periods_total|container_cpu_cfs_periods_total|container_fs_usage_bytes|container_fs_limit_bytes|container_cpu_cfs_periods_total|container_fs_inodes_free|container_fs_inodes_total|container_fs_usage_bytes|container_fs_limit_bytes|container_cpu_cfs_throttled_periods_total|container_cpu_cfs_periods_total|container_network_receive_bytes_total|container_network_transmit_bytes_total|container_fs_inodes_free|container_fs_inodes_total|container_fs_usage_bytes|container_fs_limit_bytes|container_spec_cpu_shares|container_spec_memory_limit_bytes|container_network_receive_bytes_total|container_network_transmit_bytes_total|container_fs_reads_bytes_total|container_network_receive_bytes_total|container_fs_writes_bytes_total|container_fs_reads_bytes_total|cadvisor_version_info|kubecost_pv_info)",
                                "action": "keep"
                              },
                              {
                                "source_labels": [
                                  "container"
                                ],
                                "target_label": "container_name",
                                "regex": "(.+)",
                                "action": "replace"
                              },
                              {
                                "source_labels": [
                                  "pod"
                                ],
                                "target_label": "pod_name",
                                "regex": "(.+)",
                                "action": "replace"
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-nodes",
                            "scheme": "https",
                            "tls_config": {
                              "ca_file": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                              "insecure_skip_verify": true
                            },
                            "bearer_token_file": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                            "kubernetes_sd_configs": [
                              {
                                "role": "node"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_node_label_(.+)"
                              },
                              {
                                "target_label": "__address__",
                                "replacement": "kubernetes.default.svc:443"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_node_name"
                                ],
                                "regex": "(.+)",
                                "target_label": "__metrics_path__",
                                "replacement": "/api/v1/nodes/$1/proxy/metrics"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(kubelet_volume_stats_used_bytes)",
                                "action": "keep"
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-service-endpoints",
                            "kubernetes_sd_configs": [
                              {
                                "role": "endpoints"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_scrape"
                                ],
                                "action": "keep",
                                "regex": true
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_endpoints_name"
                                ],
                                "action": "keep",
                                "regex": "(.*node-exporter|kubecost-network-costs)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_scheme"
                                ],
                                "action": "replace",
                                "target_label": "__scheme__",
                                "regex": "(https?)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_path"
                                ],
                                "action": "replace",
                                "target_label": "__metrics_path__",
                                "regex": "(.+)"
                              },
                              {
                                "source_labels": [
                                  "__address__",
                                  "__meta_kubernetes_service_annotation_prometheus_io_port"
                                ],
                                "action": "replace",
                                "target_label": "__address__",
                                "regex": "([^:]+)(?::\\d+)?;(\\d+)",
                                "replacement": "$1:$2"
                              },
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_service_label_(.+)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_namespace"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_namespace"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_name"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_name"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_pod_node_name"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_node"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(container_cpu_allocation|container_cpu_usage_seconds_total|container_fs_limit_bytes|container_fs_writes_bytes_total|container_gpu_allocation|container_memory_allocation_bytes|container_memory_usage_bytes|container_memory_working_set_bytes|container_network_receive_bytes_total|container_network_transmit_bytes_total|DCGM_FI_DEV_GPU_UTIL|deployment_match_labels|kube_daemonset_status_desired_number_scheduled|kube_daemonset_status_number_ready|kube_deployment_spec_replicas|kube_deployment_status_replicas|kube_deployment_status_replicas_available|kube_job_status_failed|kube_namespace_annotations|kube_namespace_labels|kube_node_info|kube_node_labels|kube_node_status_allocatable|kube_node_status_allocatable_cpu_cores|kube_node_status_allocatable_memory_bytes|kube_node_status_capacity|kube_node_status_capacity_cpu_cores|kube_node_status_capacity_memory_bytes|kube_node_status_condition|kube_persistentvolume_capacity_bytes|kube_persistentvolume_status_phase|kube_persistentvolumeclaim_info|kube_persistentvolumeclaim_resource_requests_storage_bytes|kube_pod_container_info|kube_pod_container_resource_limits|kube_pod_container_resource_limits_cpu_cores|kube_pod_container_resource_limits_memory_bytes|kube_pod_container_resource_requests|kube_pod_container_resource_requests_cpu_cores|kube_pod_container_resource_requests_memory_bytes|kube_pod_container_status_restarts_total|kube_pod_container_status_running|kube_pod_container_status_terminated_reason|kube_pod_labels|kube_pod_owner|kube_pod_status_phase|kube_replicaset_owner|kube_statefulset_replicas|kube_statefulset_status_replicas|kubecost_cluster_info|kubecost_cluster_management_cost|kubecost_cluster_memory_working_set_bytes|kubecost_load_balancer_cost|kubecost_network_internet_egress_cost|kubecost_network_region_egress_cost|kubecost_network_zone_egress_cost|kubecost_node_is_spot|kubecost_pod_network_egress_bytes_total|node_cpu_hourly_cost|node_cpu_seconds_total|node_disk_reads_completed|node_disk_reads_completed_total|node_disk_writes_completed|node_disk_writes_completed_total|node_filesystem_device_error|node_gpu_count|node_gpu_hourly_cost|node_memory_Buffers_bytes|node_memory_Cached_bytes|node_memory_MemAvailable_bytes|node_memory_MemFree_bytes|node_memory_MemTotal_bytes|node_network_transmit_bytes_total|node_ram_hourly_cost|node_total_hourly_cost|pod_pvc_allocation|pv_hourly_cost|service_selector_labels|statefulSet_match_labels|kubecost_pv_info|up)",
                                "action": "keep"
                              }
                            ]
                          }
                        ]
                      },
                      "rules": {
                        "groups": [
                          {
                            "name": "CPU",
                            "rules": [
                              {
                                "expr": "sum(rate(container_cpu_usage_seconds_total{container!=\"\"}[5m]))",
                                "record": "cluster:cpu_usage:rate5m"
                              },
                              {
                                "expr": "rate(container_cpu_usage_seconds_total{container!=\"\"}[5m])",
                                "record": "cluster:cpu_usage_nosum:rate5m"
                              },
                              {
                                "expr": "avg(irate(container_cpu_usage_seconds_total{container!=\"POD\", container!=\"\"}[5m])) by (container,pod,namespace)",
                                "record": "kubecost_container_cpu_usage_irate"
                              },
                              {
                                "expr": "sum(container_memory_working_set_bytes{container!=\"POD\",container!=\"\"}) by (container,pod,namespace)",
                                "record": "kubecost_container_memory_working_set_bytes"
                              },
                              {
                                "expr": "sum(container_memory_working_set_bytes{container!=\"POD\",container!=\"\"})",
                                "record": "kubecost_cluster_memory_working_set_bytes"
                              }
                            ]
                          },
                          {
                            "name": "Savings",
                            "rules": [
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind!=\"DaemonSet\"}) by (pod) * sum(container_cpu_allocation) by (pod))",
                                "record": "kubecost_savings_cpu_allocation",
                                "labels": {
                                  "daemonset": "false"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind=\"DaemonSet\"}) by (pod) * sum(container_cpu_allocation) by (pod)) / sum(kube_node_info)",
                                "record": "kubecost_savings_cpu_allocation",
                                "labels": {
                                  "daemonset": "true"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind!=\"DaemonSet\"}) by (pod) * sum(container_memory_allocation_bytes) by (pod))",
                                "record": "kubecost_savings_memory_allocation_bytes",
                                "labels": {
                                  "daemonset": "false"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind=\"DaemonSet\"}) by (pod) * sum(container_memory_allocation_bytes) by (pod)) / sum(kube_node_info)",
                                "record": "kubecost_savings_memory_allocation_bytes",
                                "labels": {
                                  "daemonset": "true"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    "alertRelabelConfigs": null,
                    "networkPolicy": {
                      "enabled": false
                    }
                  },
                  "networkCosts": {
                    "enabled": false,
                    "image": {
                      "repository": "gcr.io/kubecost1/kubecost-network-costs",
                      "tag": "v0.17.3"
                    },
                    "imagePullPolicy": "Always",
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "prometheusScrape": false,
                    "trafficLogging": true,
                    "logLevel": "info",
                    "port": 3001,
                    "resources": {
                      "limits": {
                        "cpu": "500m"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "20Mi"
                      }
                    },
                    "extraArgs": [],
                    "config": {
                      "destinations": {
                        "in-zone": [
                          "127.0.0.0/8",
                          "169.254.0.0/16",
                          "10.0.0.0/8",
                          "172.16.0.0/12",
                          "192.168.0.0/16"
                        ],
                        "in-region": [],
                        "cross-region": [],
                        "internet": [],
                        "direct-classification": []
                      },
                      "services": {
                        "google-cloud-services": false,
                        "amazon-web-services": false,
                        "azure-cloud-services": false
                      }
                    },
                    "tolerations": [],
                    "affinity": {},
                    "service": {
                      "annotations": {},
                      "labels": {}
                    },
                    "priorityClassName": "",
                    "podMonitor": {
                      "enabled": false,
                      "additionalLabels": {}
                    },
                    "additionalLabels": {},
                    "nodeSelector": {},
                    "annotations": {},
                    "healthCheckProbes": {},
                    "additionalSecurityContext": {}
                  },
                  "kubecostDeployment": {
                    "replicas": 1,
                    "labels": {},
                    "annotations": {}
                  },
                  "forecasting": {
                    "enabled": true,
                    "fullImageName": "gcr.io/kubecost1/kubecost-modeling:v0.1.6",
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "limits": {
                        "cpu": "1500m",
                        "memory": "1Gi"
                      }
                    },
                    "env": {
                      "GUNICORN_CMD_ARGS": "--log-level info -t 1200"
                    },
                    "priority": {
                      "enabled": false,
                      "name": ""
                    },
                    "nodeSelector": {},
                    "tolerations": {},
                    "affinity": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    }
                  },
                  "kubecostAggregator": {
                    "deployMethod": "singlepod",
                    "enabled": false,
                    "replicas": 1,
                    "stagingEmptyDirSizeLimit": "2Gi",
                    "numDBCopyPartitions": 25,
                    "env": {
                      "LOG_LEVEL": "info",
                      "DB_READ_THREADS": "1",
                      "DB_WRITE_THREADS": "1",
                      "DB_CONCURRENT_INGESTION_COUNT": "3"
                    },
                    "persistentConfigsStorage": {
                      "storageClass": "",
                      "storageRequest": "1Gi"
                    },
                    "aggregatorDbStorage": {
                      "storageClass": "",
                      "storageRequest": "128Gi"
                    },
                    "resources": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "cloudCost": {
                      "enabled": false,
                      "resources": {},
                      "readinessProbe": {
                        "enabled": true,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 10,
                        "failureThreshold": 200
                      }
                    },
                    "jaeger": {
                      "enabled": false,
                      "image": "jaegertracing/all-in-one",
                      "imageVersion": "latest"
                    }
                  },
                  "diagnostics": {
                    "enabled": true,
                    "primary": {
                      "enabled": false,
                      "retention": "7d",
                      "readonly": false
                    },
                    "pollingInterval": "300s",
                    "keepDiagnosticHistory": false,
                    "collectHelmValues": false,
                    "deployment": {
                      "enabled": false,
                      "resources": {
                        "requests": {
                          "cpu": "10m",
                          "memory": "20Mi"
                        }
                      },
                      "env": {},
                      "labels": {},
                      "securityContext": {},
                      "containerSecurityContext": {},
                      "nodeSelector": {},
                      "tolerations": {},
                      "affinity": {}
                    }
                  },
                  "clusterController": {
                    "enabled": false,
                    "image": {
                      "repository": "gcr.io/kubecost1/cluster-controller",
                      "tag": "v0.16.0"
                    },
                    "imagePullPolicy": "Always",
                    "priorityClassName": "",
                    "tolerations": [],
                    "actionConfigs": {
                      "clusterTurndown": [],
                      "namespaceTurndown": null,
                      "clusterRightsize": null,
                      "containerRightsize": null
                    },
                    "kubescaler": {
                      "defaultResizeAll": false
                    },
                    "namespaceTurndown": {
                      "rbac": {
                        "enabled": true
                      }
                    }
                  },
                  "reporting": {
                    "logCollection": true,
                    "productAnalytics": true,
                    "errorReporting": true,
                    "valuesReporting": true
                  },
                  "serviceMonitor": {
                    "enabled": false,
                    "additionalLabels": {},
                    "metricRelabelings": [],
                    "relabelings": [],
                    "networkCosts": {
                      "enabled": false,
                      "scrapeTimeout": "10s",
                      "additionalLabels": {},
                      "metricRelabelings": [],
                      "relabelings": []
                    }
                  },
                  "prometheusRule": {
                    "enabled": false,
                    "additionalLabels": {}
                  },
                  "supportNFS": false,
                  "initChownDataImage": "busybox",
                  "initChownData": {
                    "resources": {}
                  },
                  "grafana": {
                    "rbac": {
                      "create": true
                    },
                    "serviceAccount": {
                      "create": true,
                      "name": ""
                    },
                    "replicas": 1,
                    "deploymentStrategy": "RollingUpdate",
                    "readinessProbe": {
                      "httpGet": {
                        "path": "/api/health",
                        "port": 3000
                      }
                    },
                    "livenessProbe": {
                      "httpGet": {
                        "path": "/api/health",
                        "port": 3000
                      },
                      "initialDelaySeconds": 60,
                      "timeoutSeconds": 30,
                      "failureThreshold": 10
                    },
                    "image": {
                      "repository": "grafana/grafana",
                      "tag": "10.3.4",
                      "pullPolicy": "IfNotPresent"
                    },
                    "securityContext": {},
                    "priorityClassName": "",
                    "downloadDashboardsImage": {
                      "repository": "curlimages/curl",
                      "tag": "latest",
                      "pullPolicy": "IfNotPresent"
                    },
                    "podAnnotations": {},
                    "annotations": {},
                    "service": {
                      "type": "ClusterIP",
                      "port": 80,
                      "annotations": {},
                      "labels": {}
                    },
                    "ingress": {
                      "enabled": false,
                      "annotations": {},
                      "labels": {},
                      "path": "/",
                      "pathType": "Prefix",
                      "hosts": [
                        "chart-example.local"
                      ],
                      "tls": []
                    },
                    "resources": {},
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "persistence": {
                      "enabled": false
                    },
                    "adminUser": "admin",
                    "adminPassword": "strongpassword",
                    "env": {},
                    "envFromSecret": "",
                    "extraSecretMounts": [],
                    "plugins": [],
                    "dashboardProviders": {},
                    "dashboards": {},
                    "dashboardsConfigMaps": {},
                    "ldap": {
                      "existingSecret": "",
                      "config": ""
                    },
                    "smtp": {
                      "existingSecret": ""
                    },
                    "sidecar": {
                      "image": {
                        "repository": "kiwigrid/k8s-sidecar",
                        "tag": "1.26.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "resources": {},
                      "dashboards": {
                        "enabled": true,
                        "label": "grafana_dashboard",
                        "labelValue": "1",
                        "annotations": {},
                        "error_throttle_sleep": 0,
                        "folder": "/tmp/dashboards"
                      },
                      "datasources": {
                        "enabled": false,
                        "error_throttle_sleep": 0,
                        "label": "grafana_datasource"
                      }
                    },
                    "grafana.ini": {
                      "server": {
                        "serve_from_sub_path": false,
                        "root_url": "%(protocol)s://%(domain)s:%(http_port)s/grafana"
                      },
                      "paths": {
                        "data": "/var/lib/grafana/data",
                        "logs": "/var/log/grafana",
                        "plugins": "/var/lib/grafana/plugins",
                        "provisioning": "/etc/grafana/provisioning"
                      },
                      "analytics": {
                        "check_for_updates": true
                      },
                      "log": {
                        "mode": "console"
                      },
                      "grafana_net": {
                        "url": "https://grafana.net"
                      },
                      "auth.anonymous": {
                        "enabled": true,
                        "org_role": "Editor",
                        "org_name": "Main Org."
                      }
                    }
                  },
                  "serviceAccount": {
                    "create": true,
                    "annotations": {}
                  },
                  "awsstore": {
                    "useAwsStore": false,
                    "imageNameAndVersion": "gcr.io/kubecost1/awsstore:latest",
                    "createServiceAccount": false,
                    "priorityClassName": "",
                    "nodeSelector": {},
                    "annotations": {}
                  },
                  "federatedETL": {
                    "agentOnly": false,
                    "federatedCluster": false,
                    "redirectS3Backup": false,
                    "useMultiClusterDB": false
                  },
                  "kubecostAdmissionController": {
                    "enabled": false,
                    "secretName": "webhook-server-tls",
                    "caBundle": "${CA_BUNDLE}"
                  },
                  "costEventsAudit": {
                    "enabled": false
                  },
                  "extraObjects": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: gcr.io/kubecost1/operator/kubecost-operator-controller:v2.2.4
          createdAt: "2024-05-19T20:46:35Z"
          description: An operator that deploys Kubecost 2.2.4 via the Helm chart.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Stackwatch
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The CostAnalyzer API describes a deployment of Kubecost and
              its dependencies.
            displayName: CostAnalyzer
            kind: CostAnalyzer
            name: costanalyzers.charts.kubecost.com
            version: v1alpha1
        description: |-
          Kubecost provides real-time cost visibility and insights for teams using Kubernetes, helping you continuously reduce your cloud costs.

          This OpenShift operator deploys Kubecost from the official Helm chart.

          Kubecost allows you to:

          * See allocated spend across all native Kubernetes concepts, so you can provide your teams with transparent, accurate cost data reconciled with your actual cloud bill.
          * Join in-cluster costs like CPU and memory with out-of-cluster spend from cloud infrastructure services for complete cost visibility—all services from AWS, GCP, and Microsoft Azure are supported.
          * Gain insights you can use to save 30-50% or more on your infrastructure spend, without exposing your private information—your data never gets shared externally, even with us.
          * Have real-time alerts and recurring reports, empowering teams to take control of their Kubernetes-enabled infrastructure, stay within budgeted limits, and address monitoring interruptions immediately.

          Visit the Kubecost website for more details including features, pricing, documentation, and more at www.kubecost.com.
        displayName: Kubecost
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubecost
        - cost
        - kubernetes
        links:
        - name: Kubecost
          url: https://www.kubecost.com
        maintainers:
        - email: czoller@kubecost.com
          name: Chip Zoller
        maturity: alpha
        provider:
          name: Stackwatch
          url: www.kubecost.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - gcr.io/kubecost1/operator/kubecost-operator-controller:v2.2.4
        version: 2.2.4
      entries:
      - name: kubecost.v2.2.4
        version: 2.2.4
      - name: kubecost.v2.1.0
        version: 2.1.0
      - name: kubecost-operator.v1.107.1
        version: 1.107.1
      - name: kubecost-operator.v1.106.2
        version: 1.106.2
      - name: kubecost-operator.v1.106.0
        version: 1.106.0
      - name: kubecost.v1.99.0
        version: 1.99.0
      - name: kubecost.v1.97.0
        version: 1.97.0
      name: alpha
    defaultChannel: alpha
    packageName: kubecost
    provider:
      name: Stackwatch
      url: www.kubecost.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: prometurbo
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometurbo-operator.v8.7.1-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.7.1-SNAPSHOT
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.7.1-SNAPSHOT
        version: 8.7.1-beta.1
      entries:
      - name: prometurbo-operator.v8.7.1-beta.1
        version: 8.7.1-beta.1
      name: beta
    - currentCSV: prometurbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.6
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.6
        version: 8.6.0
      entries:
      - name: prometurbo-operator.v8.6.0
        version: 8.6.0
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: stable
    - currentCSV: prometurbo-operator.v8.5.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.5
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.5
        version: 8.5.0
      entries:
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: v8.5.0
    - currentCSV: prometurbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.6
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.6
        version: 8.6.0
      entries:
      - name: prometurbo-operator.v8.6.0
        version: 8.6.0
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: v8.6.0
    defaultChannel: stable
    packageName: prometurbo
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apache KIE
      provider-url: ""
    name: sonataflow-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sonataflow-operator.v10.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "docker.io/apache",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: docker.io/apache/incubator-kie-sonataflow-operator:10.0.0
          description: SonataFlow Kubernetes Operator for deploying workflow applications
            based on the CNCF Serverless Workflow specification
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/apache/incubator-kie-tools/packages/sonataflow-operator
          support: Apache Software Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: Sonata Flow Build
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: Sonata Flow Cluster Platform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: Sonata Flow Platform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: Sonata Flow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          SonataFlow Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification](https://serverlessworkflow.io/):

          * Deploy workflow applications using the [dev profile](https://sonataflow.org/serverlessworkflow/latest/cloud/operator/developing-workflows.html), suited for the your development cycle
          * Build workflow applications based on the platform you're currently working on.
        displayName: SonataFlow Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        links:
        - name: Product Page
          url: https://github.com/apache/incubator-kie-tools
        maintainers:
        - email: dev@kie.apache.org
          name: Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Apache Software Foundation
        relatedImages:
        - docker.io/apache/incubator-kie-sonataflow-operator:10.0.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 10.0.0
      entries:
      - name: sonataflow-operator.v10.0.0
        version: 10.0.0
      - name: sonataflow-operator.v1.40.1
        version: 1.40.1
      - name: sonataflow-operator.v1.40.0
        version: 1.40.0
      name: alpha
    - currentCSV: sonataflow-operator.v10.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "host/namespace",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: docker.io/apache/incubator-kie-sonataflow-operator:10.1.0
          createdAt: "2025-09-11T14:44:56Z"
          description: SonataFlow Kubernetes Operator for deploying workflow applications
            based on the CNCF Serverless Workflow specification
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/apache/incubator-kie-tools/packages/sonataflow-operator
          support: Apache KIE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: SonataFlowBuild
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: SonataFlowClusterPlatform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: SonataFlowPlatform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: SonataFlow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          SonataFlow Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification](https://serverlessworkflow.io/):

          * Deploy workflow applications using the [dev profile](https://sonataflow.org/serverlessworkflow/latest/cloud/operator/developing-workflows.html), suited for the your development cycle
          * Build workflow applications based on the platform you're currently working on.
        displayName: SonataFlow Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        links:
        - name: Product Page
          url: https://sonataflow.org/serverlessworkflow/latest/index.html
        maintainers:
        - email: dev@kie.apache.org
          name: Apache KIE
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Apache KIE
        relatedImages:
        - docker.io/apache/incubator-kie-sonataflow-operator:10.1.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 10.1.0
      entries:
      - name: sonataflow-operator.v10.1.0
        version: 10.1.0
      - name: sonataflow-operator.v10.0.0
        version: 10.0.0
      - name: sonataflow-operator.v1.40.1
        version: 1.40.1
      - name: sonataflow-operator.v1.40.0
        version: 1.40.0
      name: stable
    defaultChannel: stable
    packageName: sonataflow-operator
    provider:
      name: Apache KIE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: directpv
      provider-url: ""
    name: private-dpv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: private-dpv-operator.v5.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.quay.io/v1alpha1",
                "kind": "AistorVolumemanager",
                "metadata": {
                  "name": "aistorvolumemanager-sample"
                },
                "spec": {
                  "annotations": {},
                  "controller": {
                    "podAnnotations": {},
                    "podLabels": {},
                    "replicas": 3
                  },
                  "images": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": []
                  },
                  "labels": {},
                  "license": "",
                  "nodeserver": {
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-10-05T15:47:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AistorVolumemanager
            name: aistorvolumemanagers.charts.quay.io
            version: v1alpha1
        description: directpv
        displayName: directpv
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - directpv
        links:
        - name: Private Dpv Operator
          url: https://private-dpv-operator.domain
        maintainers:
        - email: cesarcelis@minio.io
          name: celis
        maturity: alpha
        provider:
          name: directpv
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - quay.io/cniackz4/directpv-operator@sha256:af06e57661a7e05809e44c3fac571ed949cc04fd656fdb0c8324b533778eb604
        version: 5.0.3
      entries:
      - name: private-dpv-operator.v5.0.3
        version: 5.0.3
      - name: private-dpv-operator.v5.0.2
        version: 5.0.2
      name: alpha
    defaultChannel: alpha
    packageName: private-dpv-operator
    provider:
      name: directpv
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fortanix Inc.
      provider-url: https://fortanix.com
    name: ccm-node-agent-dcap-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ccm-node-agent-dcap-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1",
                "kind": "CcmNodeAgent",
                "metadata": {
                  "name": "ccmnodeagent-sample"
                },
                "spec": {
                  "ccmJoinToken": {
                    "secretName": "ccm-join-token",
                    "token": ""
                  },
                  "fullnameOverride": "",
                  "image": {
                    "aesmd": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/aesmd-dcap",
                      "tag": "latest"
                    },
                    "ccmNodeAgent": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/ccm-node-agent-ubi7",
                      "tag": "latest"
                    },
                    "ccmNodeAgentInit": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "busybox",
                      "tag": "latest"
                    },
                    "multiPlatformRegistration": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/multi-platform-registration-ubi8",
                      "tag": "latest"
                    }
                  },
                  "log": {
                    "debug": false
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "ports": {
                    "containerPort": 9092,
                    "hostPort": 9092
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CCM Node Agent software enables registration of the compute
              nodes to Fortanix CCM when installed in the Openshift Cluster Nodes<U+200B>.
            displayName: CCM Node Agent
            kind: CcmNodeAgent
            name: ccmnodeagents.charts.operatorhub.io
            version: v1
        description: "CCM Node Agent software enables registration of the compute
          nodes to Fortanix CCM when installed in the Openshift Cluster Nodes​. \nIt
          also assists with application attestation and visibility for Fortanix CCM.\n\n##
          About this Operator\nThis Operator is based on a Helm chart for CCM Node
          Agent. It enrolls all the nodes in the openshift cluster to a particular
          CCM account. You have to provide the base64 encode of join token of CCM
          account while creating CcmNodeAgent resource.\n\n## Prerequisites\nAll the
          worker nodes in the cluster must have SGX capabilities, in-kernel driver
          installed and fsgsbase support.\n\n## Installation\n * Choose a namespace
          to install the Operator. \n Operator will be available in all namespaces.\n*
          Select Approval Strategy: Available options are Automatic or Manual. Automatic
          approval will automatically upgrade the running instance of your Operator
          without human intervention when a new version is available. In case of Manual
          approval, when a new version is available, OLM will create an update request.
          You must manually approve the request to update the Operator to a new version.\n*
          Generate your CCM Account Join Token: Please log in to https://ccm.fortanix.com,
          and in the Management Console tab, click the + ENROLL NODE button. In the
          ENROLL NODE screen, a Join Token will be generated in the text box for \"Get
          a join token to register an SGX compute node\". This Join Token is used
          by the compute node to authenticate itself.\n* Base64 encode the Join Token:\n```shell\necho
          -n <join-token> | openssl base64 -A\n```\n* Create an instance of CcmNodeAgent
          by setting the `spec.ccmJoinToken.Token` to the base64 encoded CCM account
          join token.\n* After CcmNodeAgent resource is installed successfully you
          should be able to see all the worker nodes of the cluster in your CCM account
          under Compute Node Tab.\n\n## Getting Started\nNow to run your application
          inside SGX enclaves on your cluster follow the [documentation](https://support.fortanix.com/hc/en-us/articles/360043527431-User-s-Guide-Add-and-Edit-an-Application).\n"
        displayName: CCM Node Agent DCAP Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dockerfile
        - Helm
        - CCM
        - Confidential Compute
        links:
        - name: CCM Node Agent DCAP Operator
          url: https://support.fortanix.com/hc/en-us/categories/360003107511-Confidential-Computing-Manager
        maintainers:
        - email: info@fortanix.com
          name: Fortanix Inc.
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Fortanix Inc.
          url: https://fortanix.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62575404000cb77fa8d2284c87b31e39e667a55b0fb47821263b23b3d1135a7e
        - docker.io/fortanix/ccm-node-agent-dcap-operator@sha256:4fd28bea26aee461eb9c1800cdf5f041b64ae3ea6652203f8ad9feb680aaefee
        version: 0.0.2
      entries:
      - name: ccm-node-agent-dcap-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ccm-node-agent-dcap-operator
    provider:
      name: Fortanix Inc.
      url: https://fortanix.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: service-telemetry-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-telemetry-operator.v1.5.1757650583
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infra.watch/v1beta1",
                "kind": "ServiceTelemetry",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "alerting": {
                    "alertmanager": {
                      "receivers": {
                        "snmpTraps": {
                          "alertOidLabel": "oid",
                          "community": "public",
                          "enabled": false,
                          "port": 162,
                          "retries": 5,
                          "target": "192.168.24.254",
                          "timeout": 1,
                          "trapDefaultOid": "1.3.6.1.4.1.50495.15.1.2.1",
                          "trapDefaultSeverity": "",
                          "trapOidPrefix": "1.3.6.1.4.1.50495.15"
                        }
                      },
                      "storage": {
                        "persistent": {
                          "pvcStorageRequest": "20G"
                        },
                        "strategy": "persistent"
                      }
                    },
                    "enabled": true
                  },
                  "backends": {
                    "events": {
                      "elasticsearch": {
                        "certificates": {
                          "caCertDuration": "70080h",
                          "endpointCertDuration": "70080h"
                        },
                        "enabled": false,
                        "forwarding": {
                          "hostUrl": "https://elasticsearch-es-http:9200",
                          "tlsSecretName": "elasticsearch-es-cert",
                          "tlsServerName": "",
                          "useBasicAuth": true,
                          "useTls": true,
                          "userSecretName": "elasticsearch-es-elastic-user"
                        },
                        "storage": {
                          "persistent": {
                            "pvcStorageRequest": "20Gi"
                          },
                          "strategy": "persistent"
                        },
                        "version": "7.16.1"
                      }
                    },
                    "metrics": {
                      "prometheus": {
                        "enabled": true,
                        "scrapeInterval": "30s",
                        "storage": {
                          "persistent": {
                            "pvcStorageRequest": "20G"
                          },
                          "retention": "24h",
                          "strategy": "persistent"
                        }
                      }
                    }
                  },
                  "clouds": [
                    {
                      "metrics": {
                        "collectors": [
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 16384,
                              "verbose": false
                            },
                            "collectorType": "collectd",
                            "debugEnabled": false,
                            "subscriptionAddress": "collectd/cloud1-telemetry"
                          },
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 16384,
                              "verbose": false
                            },
                            "collectorType": "ceilometer",
                            "debugEnabled": false,
                            "subscriptionAddress": "anycast/ceilometer/cloud1-metering.sample"
                          },
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 65535,
                              "verbose": false
                            },
                            "collectorType": "sensubility",
                            "debugEnabled": false,
                            "subscriptionAddress": "sensubility/cloud1-telemetry"
                          }
                        ]
                      },
                      "name": "cloud1"
                    }
                  ],
                  "graphing": {
                    "enabled": false,
                    "grafana": {
                      "baseImage": "registry.redhat.io/rhel8/grafana:9",
                      "dashboards": {
                        "enabled": true
                      },
                      "disableSignoutMenu": false,
                      "ingressEnabled": true
                    }
                  },
                  "highAvailability": {
                    "enabled": false
                  },
                  "observabilityStrategy": "use_redhat",
                  "transports": {
                    "qdr": {
                      "auth": "basic",
                      "certificates": {
                        "caCertDuration": "70080h",
                        "endpointCertDuration": "70080h"
                      },
                      "enabled": true,
                      "web": {
                        "enabled": false
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/stf/service-telemetry-rhel9-operator@sha256:822becd77fa8835e753c88b3e307f62e42e090bdedc6e1fbcbe9fd6110310846
          createdAt: "2025-09-12T04:17:40Z"
          description: Service Telemetry Framework. Umbrella Operator for instantiating
            the required dependencies and configuration of various components to build
            a Service Telemetry platform for telco grade monitoring.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: service-telemetry
          operators.openshift.io/valid-subscription: '["OpenStack Platform", "Cloud
            Infrastructure", "Cloud Suite"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/infrawatch/service-telemetry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of the Service Telemetry Framework
            displayName: STF Cluster
            kind: ServiceTelemetry
            name: servicetelemetrys.infra.watch
            version: v1beta1
        description: Service Telemetry Operator for monitoring clouds
        displayName: Service Telemetry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - servicetelemetry
        - monitoring
        - telemetry
        - notifications
        - telecommunications
        links:
        - name: Source Code
          url: https://github.com/infrawatch/service-telemetry-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_openstack_platform/17.1/html/service_telemetry_framework_1.5/index
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy:latest@sha256:dc24aa382e02c4ac058d4b824a21f0e90a6259d6d817d0341b2a41809b5daa78
        - registry.redhat.io/cluster-observability-operator/coo-prometheus-rhel8:0.2.0@sha256:6bbb1aebb1cfdc3310e398ec7720ca1d4a9a89cecf7406412071bfaceb0d3cb3
        - registry.redhat.io/cluster-observability-operator/coo-prometheus-alertmanager-rhel8:0.2.0@sha256:cfd18e83d64a55ac310937eefe83fc5b9e98660962a127cd4f8d9a382764c171
        - registry.redhat.io/stf/service-telemetry-rhel9-operator@sha256:822becd77fa8835e753c88b3e307f62e42e090bdedc6e1fbcbe9fd6110310846
        - registry.redhat.io/stf/prometheus-webhook-snmp-rhel9@sha256:03199f8a3b6b22c82e01ede3adab6e98b276369d642029ae03641aac3b25581a
        version: 1.5.1757650583
      entries:
      - name: service-telemetry-operator.v1.5.1757650583
        version: 1.5.1757650583
      - name: service-telemetry-operator.v1.5.1743671601
        version: 1.5.1743671601
      - name: service-telemetry-operator.v1.5.1740051851
        version: 1.5.1740051851
      - name: service-telemetry-operator.v1.5.1731336911
        version: 1.5.1731336911
      - name: service-telemetry-operator.v1.5.1725952694
        version: 1.5.1725952694
      name: stable-1.5
    defaultChannel: stable-1.5
    packageName: service-telemetry-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: barbican-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: barbican-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "barbican.openstack.org/v1beta1",
                "kind": "Barbican",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "barbican-operator",
                    "app.kubernetes.io/instance": "barbican",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "barbican",
                    "app.kubernetes.io/part-of": "barbican-operator"
                  },
                  "name": "barbican"
                },
                "spec": {
                  "barbicanAPI": {
                    "customServiceConfig": "[optional]\noverrides = True\n",
                    "defautlConfigOverwrite": {
                      "optional_policy.json": "{\"some\": \"custom policy\"}\n"
                    },
                    "passwordSelectors": {
                      "database": "BarbicanDatabasePassword",
                      "service": "BarbicanPassword"
                    }
                  },
                  "barbicanKeystoneListener": {
                    "customServiceConfig": "[optional]\noverrides = True\n",
                    "defautlConfigOverwrite": {
                      "optional_policy.json": "{\"some\": \"custom policy\"}\n"
                    }
                  },
                  "barbicanWorker": {
                    "customServiceConfig": "[optional]\noverrides = True\n",
                    "defautlConfigOverwrite": {
                      "optional_policy.json": "{\"some\": \"custom policy\"}\n"
                    }
                  },
                  "customServiceConfig": "[DEFAULT]\ndebug = True\n",
                  "databaseAccount": "barbican",
                  "databaseInstance": "openstack",
                  "defaultConfigOverwrite": {
                    "policy.json": "{\"some\": \"custom policy\"}\n"
                  },
                  "passwordSelectors": {
                    "database": "BarbicanDatabasePassword",
                    "service": "BarbicanPassword",
                    "simplecryptokek": "BarbicanSimpleCryptoKEK"
                  },
                  "preserveJobs": true,
                  "rabbitMqCusterName": "barbican_rabbit",
                  "secret": "osp-secret",
                  "serviceAccount": "barbican",
                  "serviceUser": "barbican"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:32Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BarbicanAPI is the Schema for the barbicanapis API
            displayName: Barbican API
            kind: BarbicanAPI
            name: barbicanapis.barbican.openstack.org
            version: v1beta1
          - kind: BarbicanKeystoneListener
            name: barbicankeystonelisteners.barbican.openstack.org
            version: v1beta1
          - description: Barbican is the Schema for the barbicans API
            displayName: Barbican
            kind: Barbican
            name: barbicans.barbican.openstack.org
            version: v1beta1
          - description: BarbicanWorker is the Schema for the barbicanworkers API
            displayName: Barbican Worker
            kind: BarbicanWorker
            name: barbicanworkers.barbican.openstack.org
            version: v1beta1
        description: Barbican Operator
        displayName: Barbican Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - KeyManager
        - Barbican
        links:
        - name: Barbican Operator
          url: https://github.com/openstack-k8s-operators/barbican-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-barbican-worker-rhel9@sha256:b5133f539658ed5d532e0c3cf63d90bc0f76c7f0368963a8ebec49123f27f7a7
        - registry.redhat.io/rhoso/openstack-barbican-keystone-listener-rhel9@sha256:ac112444c0a12664e8a7900e72cd97fa795df167c0d95437a029644f0819f30d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/barbican-rhel9-operator@sha256:1275a3f7f09999e9dc25bc38f917a3b5d780d9d309442cddd0e7adf313d6f576
        - registry.redhat.io/rhoso/openstack-barbican-api-rhel9@sha256:0db6c1ab5b4ca302e38338a3ab9cee9bd4ae7c8e1e7bd3419e47b65df9ef1f4f
        version: 1.0.6
      entries:
      - name: barbican-operator.v1.0.6
        version: 1.0.6
      - name: barbican-operator.v1.0.5
        version: 1.0.5
      - name: barbican-operator.v1.0.4
        version: 1.0.4
      - name: barbican-operator.v1.0.3
        version: 1.0.3
      - name: barbican-operator.v1.0.2
        version: 1.0.2
      - name: barbican-operator.v1.0.1
        version: 1.0.1
      - name: barbican-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: barbican-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: ansible-cloud-addons-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aca-operator.v2.4.0-0.1737660894
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Cloud Addons Operator manages everything Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.1.1-0.1645500829 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy an AAP AWS Billing instance
            displayName: AnsibleAutomationPlatformAwsBilling
            kind: AnsibleAutomationPlatformAwsBilling
            name: ansibleautomationplatformawsbillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Azure Billing instance
            displayName: AnsibleAutomationPlatformAzureBilling
            kind: AnsibleAutomationPlatformAzureBilling
            name: ansibleautomationplatformazurebillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Metrics instance
            displayName: AnsibleAutomationPlatformMetrics
            kind: AnsibleAutomationPlatformMetrics
            name: ansibleautomationplatformmetrics.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP UI instance
            displayName: AutomationPlatform
            kind: AutomationPlatform
            name: automationplatforms.platform.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Cloud Addons is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Cloud Addons Operator provides for push-button deployments of Cloud Add-ons into your OpenShift environment. Ansible Cloud Addons deployments include:
            * Cloud Billing - Cloud billing connector for Ansible Automation Platform
            * Cloud UI - Unified welcome page UI for Ansible Automation Platform

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Cloud Addons operator by following instructions presented when you click the Install button.
        displayName: Ansible Cloud Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-platform
        - ansible-automation-platform
        - aap
        - cloud
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:38ee3484ba60bb474e7231e9d0718157386716cea7ffd6da6f32ddd551a5243d
        - registry.redhat.io/ansible-automation-platform-24/aap-cloud-billing-rhel8-operator@sha256:c461b0c4781e95d2b2e31e7bbd6d6d192806e8a8dd2e8b9225b5735da93dd6a2
        - registry.redhat.io/ansible-automation-platform-24/aap-cloud-ui-rhel8-operator@sha256:06bf313d4daf080ce6f2b1ba859fefd84431c7a2a0618feae55b92c9f5fff594
        - registry.redhat.io/ansible-automation-platform-24/aap-cloud-billing-rhel8@sha256:84fa86d312663a74502516f2ba838ac80e500c7bc17dc518940984a603da2548
        - registry.redhat.io/ansible-automation-platform-24/aap-cloud-metrics-collector-rhel8@sha256:6bd27ad06d0a65c7fa1cd2419bfc30976d53a5b89698d309cabbbd65b531a1ab
        - registry.redhat.io/ansible-automation-platform-24/aap-cloud-ui-rhel8@sha256:8489c253072186db064289f5ecc0a08ed3654a73be3c3f2130be479098e59047
        version: 2.4.0+0.1737660894
      entries:
      - name: aca-operator.v2.4.0-0.1708974547
        version: 2.4.0+0.1708974547
      - name: aca-operator.v2.4.0-0.1711587287
        version: 2.4.0+0.1711587287
      - name: aca-operator.v2.4.0-0.1713463331
        version: 2.4.0+0.1713463331
      - name: aca-operator.v2.4.0-0.1715891874
        version: 2.4.0+0.1715891874
      - name: aca-operator.v2.4.0-0.1717108266
        version: 2.4.0+0.1717108266
      - name: aca-operator.v2.4.0-0.1717558193
        version: 2.4.0+0.1717558193
      - name: aca-operator.v2.4.0-0.1718149348
        version: 2.4.0+0.1718149348
      - name: aca-operator.v2.4.0-0.1720478479
        version: 2.4.0+0.1720478479
      - name: aca-operator.v2.4.0-0.1725253934
        version: 2.4.0+0.1725253934
      - name: aca-operator.v2.4.0-0.1726108417
        version: 2.4.0+0.1726108417
      - name: aca-operator.v2.4.0-0.1727255735
        version: 2.4.0+0.1727255735
      - name: aca-operator.v2.4.0-0.1730152463
        version: 2.4.0+0.1730152463
      - name: aca-operator.v2.4.0-0.1733181569
        version: 2.4.0+0.1733181569
      - name: aca-operator.v2.4.0-0.1733937635
        version: 2.4.0+0.1733937635
      - name: aca-operator.v2.4.0-0.1737660894
        version: 2.4.0+0.1737660894
      name: stable-2.4-cluster-scoped
    - currentCSV: aca-operator.v2.5.0-0.1758143647
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Cloud Addons Operator manages everything Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.1.1-0.1645500829 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy an AAP AWS Billing instance
            displayName: AnsibleAutomationPlatformAwsBilling
            kind: AnsibleAutomationPlatformAwsBilling
            name: ansibleautomationplatformawsbillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Azure Billing instance
            displayName: AnsibleAutomationPlatformAzureBilling
            kind: AnsibleAutomationPlatformAzureBilling
            name: ansibleautomationplatformazurebillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Metrics instance
            displayName: AnsibleAutomationPlatformMetrics
            kind: AnsibleAutomationPlatformMetrics
            name: ansibleautomationplatformmetrics.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP UI instance
            displayName: AutomationPlatform
            kind: AutomationPlatform
            name: automationplatforms.platform.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Cloud Addons is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Cloud Addons Operator provides for push-button deployments of Cloud Add-ons into your OpenShift environment. Ansible Cloud Addons deployments include:
            * Cloud Billing - Cloud billing connector for Ansible Automation Platform
            * Cloud UI - Unified welcome page UI for Ansible Automation Platform

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Cloud Addons operator by following instructions presented when you click the Install button.
        displayName: Ansible Cloud Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-platform
        - ansible-automation-platform
        - aap
        - cloud
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-billing-rhel8-operator@sha256:f4b29ec59498bb6c9fdb2790ffc9d8c878efa3fd9da3416f7fe9f2a253171508
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-ui-rhel8-operator@sha256:47e2c40cdcfe218d5bf7abbf6fc35340bacdc2d02dac7f81deee7876b5570c53
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-billing-rhel8@sha256:0e50f539bfeaec91f0b18409cbc1888297a26af5114252a8ca07e1b97436001d
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-metrics-collector-rhel8@sha256:c6ab25b3839f8c3134e7d5e1f0b7716ea5fd66fe8d131261ff08aabd97c3cd2d
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-ui-rhel8@sha256:869efe830cdf23f9344263f0e9b5a209cb13b84d9d7588a1744fd3ea4cd3f83b
        version: 2.5.0+0.1758143647
      entries:
      - name: aca-operator.v2.5.0-0.1727240231
        version: 2.5.0+0.1727240231
      - name: aca-operator.v2.5.0-0.1727874815
        version: 2.5.0+0.1727874815
      - name: aca-operator.v2.5.0-0.1728516369
        version: 2.5.0+0.1728516369
      - name: aca-operator.v2.5.0-0.1729737962
        version: 2.5.0+0.1729737962
      - name: aca-operator.v2.5.0-0.1731452082
        version: 2.5.0+0.1731452082
      - name: aca-operator.v2.5.0-0.1733188778
        version: 2.5.0+0.1733188778
      - name: aca-operator.v2.5.0-0.1733964232
        version: 2.5.0+0.1733964232
      - name: aca-operator.v2.5.0-0.1736892777
        version: 2.5.0+0.1736892777
      - name: aca-operator.v2.5.0-0.1737095650
        version: 2.5.0+0.1737095650
      - name: aca-operator.v2.5.0-0.1737671171
        version: 2.5.0+0.1737671171
      - name: aca-operator.v2.5.0-0.1739244024
        version: 2.5.0+0.1739244024
      - name: aca-operator.v2.5.0-0.1740088272
        version: 2.5.0+0.1740088272
      - name: aca-operator.v2.5.0-0.1740769078
        version: 2.5.0+0.1740769078
      - name: aca-operator.v2.5.0-0.1741367998
        version: 2.5.0+0.1741367998
      - name: aca-operator.v2.5.0-0.1742427652
        version: 2.5.0+0.1742427652
      - name: aca-operator.v2.5.0-0.1743655743
        version: 2.5.0+0.1743655743
      - name: aca-operator.v2.5.0-0.1746134365
        version: 2.5.0+0.1746134365
      - name: aca-operator.v2.5.0-0.1747339736
        version: 2.5.0+0.1747339736
      - name: aca-operator.v2.5.0-0.1749072064
        version: 2.5.0+0.1749072064
      - name: aca-operator.v2.5.0-0.1750895274
        version: 2.5.0+0.1750895274
      - name: aca-operator.v2.5.0-0.1753899647
        version: 2.5.0+0.1753899647
      - name: aca-operator.v2.5.0-0.1755834478
        version: 2.5.0+0.1755834478
      - name: aca-operator.v2.5.0-0.1758143647
        version: 2.5.0+0.1758143647
      name: stable-2.5-cluster-scoped
    defaultChannel: stable-2.5-cluster-scoped
    packageName: ansible-cloud-addons-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SAP LinuxLab - Red Hat
      provider-url: https://github.com/sap-linuxlab
    name: sap-data-intelligence-observer-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: observer-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sdi.sap-redhat.io/v1alpha1",
                "kind": "SDIObserver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "observer-operator",
                    "app.kubernetes.io/instance": "sdiobserver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sdiobserver",
                    "app.kubernetes.io/part-of": "observer-operator"
                  },
                  "name": "sdiobserver-sample"
                },
                "spec": {
                  "manageSDINodeConfig": true,
                  "sdiNamespace": "sdi",
                  "sdiVSystemRoute": {
                    "managementState": "Managed"
                  },
                  "slcbNamespace": "sap-slcbridge",
                  "slcbRoute": {
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/redhat-sap-cop/sdi-observer-operator:0.1.5
          description: Configure and enable SAP Data Intelligence to run on OpenShift
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SDIObserver is the Schema for the sdiobservers API
            displayName: SDIObserver
            kind: SDIObserver
            name: sdiobservers.sdi.sap-redhat.io
            version: v1alpha1
        description: Operator for monitoring SAP Data Intelligence (SDI) namespace
          and modifying objects in there that enable running of SDI on top of OpenShift.
          The observer shall be run in a dedicated namespace. It must be deployed
          before the SDI installation is started. This is the prerequisite for the
          successful installation of the SDI on OpenShift. Please refer to https://access.redhat.com/articles/5100521
          for detailed information.
        displayName: SAP Data Intelligence 3 - Observer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sdi
        - sap data intelligence
        links:
        - name: SAP Data Intelligence 3 - Observer Operator
          url: https://github.com/redhat-sap/sap-data-intelligence/tree/master/observer-operator
        maintainers:
        - email: mjiao@redhat.com
          name: Manjun Jiao
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/redhat-sap-cop/sdi-observer-operator:0.1.5
        version: 0.0.3
      entries:
      - name: observer-operator.v0.0.3
        version: 0.0.3
      name: alpha
    - currentCSV: observer-operator.v1.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sdi.sap-redhat.io/v1alpha1",
                "kind": "SDIObserver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/part-of": "observer-operator"
                  },
                  "name": "sdiobserver-sample",
                  "namespace": "sdi-observer"
                },
                "spec": {
                  "manageSDINodeConfig": true,
                  "sdiNamespace": "sdi",
                  "sdiVSystemRoute": {
                    "managementState": "Managed"
                  },
                  "slcbNamespace": "sap-slcbridge",
                  "slcbRoute": {
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/redhat-sap-cop/sdi-observer-operator:0.1.32
          createdAt: "20240726"
          description: Configure and enable SAP Data Intelligence 3 to run on OpenShift
            4
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mjiao/sap-data-intelligence/tree/master/observer-operator
          support: RedHat-SAPLinuxLab
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SDIObserver is the Schema for the sdiobservers API
            displayName: SDIObserver
            kind: SDIObserver
            name: sdiobservers.sdi.sap-redhat.io
            version: v1alpha1
        description: Operator for monitoring SAP Data Intelligence (SDI) namespace
          and modifying objects in there that enable running of SDI on top of OpenShift.
          It must be deployed before the SDI installation is started. This is the
          prerequisite for the successful installation of the SDI on OpenShift. Please
          refer to https://access.redhat.com/articles/7042265#ocp-install-sdi-observer-operator
          for detailed information.
        displayName: SAP Data Intelligence 3 - Observer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sdi
        - sap data intelligence
        links:
        - name: SAP Data Intelligence 3 - Observer Operator
          url: https://github.com/redhat-sap/sap-data-intelligence/tree/master/observer-operator
        maintainers:
        - email: mjiao@redhat.com
          name: Manjun Jiao
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/redhat-sap-cop/sdi-observer-operator:0.1.32
        version: 1.0.9
      entries:
      - name: observer-operator.v1.0.9
        version: 1.0.9
      - name: observer-operator.v1.0.8
        version: 1.0.8
      - name: observer-operator.v1.0.7
        version: 1.0.7
      - name: observer-operator.v1.0.6
        version: 1.0.6
      - name: observer-operator.v1.0.5
        version: 1.0.5
      - name: observer-operator.v1.0.4
        version: 1.0.4
      - name: observer-operator.v1.0.3
        version: 1.0.3
      - name: observer-operator.v1.0.2
        version: 1.0.2
      - name: observer-operator.v1.0.1
        version: 1.0.1
      - name: observer-operator.v0.0.3
        version: 0.0.3
      name: stable
    defaultChannel: stable
    packageName: sap-data-intelligence-observer-operator
    provider:
      name: SAP LinuxLab - Red Hat
      url: https://github.com/sap-linuxlab
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: node-healthcheck-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-healthcheck-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "maxUnhealthy": "49%",
                  "remediationTemplate": {
                    "apiVersion": "poison-pill.medik8s.io/v1alpha1",
                    "kind": "PoisonPillRemediationTemplate",
                    "name": "group-x",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-healthcheck-operator:0.1.0-latest
          createdAt: "2021-04-20 12:00:00"
          description: An operator to monitor node healthcheck and remdiate via 3rd
            party providers like poison-pill.
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: This operator deploys the Node Health Check controller and its
          artifacts
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://node-healthcheck-operator.domain
        maintainers:
        - email: medik8s@googlegroups.com
          name: medik8s team
        maturity: alpha
        provider:
          name: medik8s
          url: https://github.com/medik8s
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/node-healthcheck-operator:0.1.0-latest
        version: 0.1.1
      entries:
      - name: node-healthcheck-operator.v0.1.1
        version: 0.1.1
      name: alpha
    - currentCSV: node-healthcheck-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-healthcheck-operator:v0.3.0
          createdAt: "2022-08-24 11:40:38"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.2.0 <0.3.0'
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: medik8s team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: medik8s
          url: https://github.com/medik8s
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/node-healthcheck-operator:v0.3.0
        version: 0.3.0
      entries:
      - name: node-healthcheck-operator.v0.3.0
        version: 0.3.0
      name: candidate
    - currentCSV: node-healthcheck-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: quay.io/medik8s/node-healthcheck-operator:v0.6.1
          createdAt: "2023-11-14T15:45:34Z"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.1.0 <0.6.1'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/medik8s/node-healthcheck-operator:v0.6.1
        - quay.io/medik8s/node-remediation-console:v0.6.1
        version: 0.6.1
      entries:
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      - name: node-healthcheck-operator.v0.6.0
        version: 0.6.0
      - name: node-healthcheck-operator.v0.5.0
        version: 0.5.0
      - name: node-healthcheck-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: node-healthcheck-operator
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: GitLab Inc
      provider-url: https://about.gitlab.com/
    name: gitlab-operator-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "gitlab": {
                        "gitlab-shell": {
                          "maxReplicas": 2,
                          "minReplicas": 1
                        },
                        "sidekiq": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1000M"
                            }
                          }
                        },
                        "webservice": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1500M"
                            }
                          }
                        }
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "installCertmanager": false,
                      "minio": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      },
                      "nginx-ingress": {
                        "controller": {
                          "electionID": "ingress-controller-leader",
                          "replicaCount": 1
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      },
                      "redis": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-29T12:42:23Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      name: stable
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "gitlab": {
                        "gitlab-shell": {
                          "maxReplicas": 2,
                          "minReplicas": 1
                        },
                        "sidekiq": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1000M"
                            }
                          }
                        },
                        "webservice": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1500M"
                            }
                          }
                        }
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "installCertmanager": false,
                      "minio": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      },
                      "nginx-ingress": {
                        "controller": {
                          "electionID": "ingress-controller-leader",
                          "replicaCount": 1
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      },
                      "redis": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-29T12:42:23Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      name: unstable
    defaultChannel: stable
    packageName: gitlab-operator-kubernetes
    provider:
      name: GitLab Inc
      url: https://about.gitlab.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: edb-hcp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: edb-hcp-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edbpgai.edb.com/v1alpha1",
                "kind": "HybridControlPlane",
                "metadata": {
                  "name": "placeholder"
                },
                "spec": {
                  "beaconAgent": {
                    "applianceProviderEnabled": true,
                    "beaconDBConnected": true,
                    "beaconServerAddress": "upm-beacon-server-ingestion.upm-beacon.svc.cluster.local:8080",
                    "inCluster": true,
                    "location": "devspatcher",
                    "logLevel": "debug",
                    "plaintext": true,
                    "providers": [
                      "appliance"
                    ],
                    "provisioning": {
                      "imageDiscovery": true,
                      "imagesetDiscoveryAuthenticationType": "token",
                      "imagesetDiscoveryContainerRegistryURL": "placeholder",
                      "loadBalancersEnabled": true,
                      "nodePortDomain": "nodeportdomain.com",
                      "openshift": true,
                      "provider": "aws"
                    },
                    "tlsInsecure": false,
                    "transparentDataEncryptionMethods": [
                      "passphrase",
                      "hashicorp_vault"
                    ]
                  },
                  "beaconServer": {
                    "logLevel": "debug"
                  },
                  "componentsParameters": {
                    "edb-migration-portal": {
                      "enable_pdb": "false"
                    },
                    "transporter-data-operator": {
                      "fips_enabled": "false"
                    },
                    "transporter-dp-agent": {
                      "rw_service_url": "placeholder"
                    },
                    "transporter-rw-service": {
                      "domain_name": "placeholder"
                    },
                    "upm-api-pnxt-metric": {
                      "log_level": "info",
                      "tsdb_host": "thanos-query.monitoring.svc.cluster.local",
                      "tsdb_port": "9090"
                    },
                    "upm-api-pnxt-pd": {
                      "log_level": "info"
                    },
                    "upm-beaco-ff-base": {
                      "usage_generator_interval": "24h"
                    },
                    "upm-beaco-usages": {
                      "usage_generator_interval": "24h"
                    },
                    "upm-beacon": {
                      "beacon_db_hibernation": "\"off\"",
                      "beacon_server_replicas": "1",
                      "external_metrics_replicas": "1",
                      "image_discovery_credentials_secret_name": "placeholder",
                      "server_host": "placeholder"
                    },
                    "upm-istio": {
                      "ingress_grpc_tls_node_port": "30290",
                      "ingress_http_node_port": "32542",
                      "ingress_https_node_port": "30288"
                    },
                    "upm-istio-gateway": {
                      "cookie_aeskey": "rzkutHl8NJNztPMEJYykZouHslNiA7xmIXH+58ISUVo=",
                      "ingress_grpc_tls_node_port": "30290",
                      "ingress_http_node_port": "32542",
                      "ingress_https_node_port": "30288",
                      "openshift_console_domain_name": "placeholder"
                    },
                    "upm-prometheus": {
                      "cpu": "200m",
                      "memory": "6Gi",
                      "prometheus_retention_size": "15GB",
                      "prometheus_storage_size": "20Gi"
                    },
                    "upm-thanos": {
                      "memcached_cpu": "500m",
                      "memcached_memory": "2Gi",
                      "query_frontend_cpu": "200m",
                      "query_frontend_memory": "128Mi"
                    },
                    "upm-trust-manager": {
                      "cert_manager_namespace": "cert-manager"
                    }
                  },
                  "flavour": "placeholder",
                  "globalParameters": {
                    "default_ingress_domain": "placeholder",
                    "internal_backup_folder": "a7462dbc7106",
                    "portal_certificate_secret": "beaco-istio-gw-cert",
                    "portal_domain_name": "placeholder",
                    "portal_port": "443",
                    "storage_class": "placeholder"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "placeholder",
                      "namespace": "default"
                    }
                  ],
                  "imageRegistry": "placeholder",
                  "pgai": {
                    "portal": {
                      "authentication": {
                        "clientSecret": "edb-postgres-ai-secret",
                        "idpConnectors": [],
                        "staticPasswords": [
                          {
                            "email": "owner@mycompany.com",
                            "hash": "$2y$10/JLTqiuGblH.RemIutWxkztN5m",
                            "userID": "c5998173-a605-449a-a9a5-4a9c33e26df7",
                            "username": "Owner MyCompany"
                          }
                        ]
                      }
                    }
                  },
                  "source": {
                    "useLocalKustomizations": true
                  },
                  "version": "placeholder"
                }
              }
            ]
          capabilities: Basic Install
          categories: Database,AI/Machine Learning
          containerImage: docker.enterprisedb.com/pgai-platform/edb-hcp-operator/manager@sha256:82610c624e956a67df4858929e1aabd561a7537ba68d40e6512b68f55e2649e0
          createdAt: "2025-06-17T11:00:42Z"
          description: Unify transactional, analytical, and AI workloads and bring
            cloud agility to your hybrid data infrastructure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/EnterpriseDB/edb-hcp-operator
          support: techsupport@enterprisedb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HybridControlPlane is the Schema for the hybridcontrolplanes
              API
            displayName: Hybrid Control Plane
            kind: HybridControlPlane
            name: hybridcontrolplanes.edbpgai.edb.com
            version: v1alpha1
        description: |
          EDB Postgres AI is a sovereign data and AI platform that brings analytics and AI closer to core operational data and enables cloud agility in private environments.
          Unify your structured and unstructured data in one multi-model data platform with centralized management and observability to:
          - Accelerate enterprise-grade transactional databases
          - Enable rapid analytics and powerful lakehouse integrations
          - Automate AI pipelines and semantic search for faster GenAI development
          Using EDB’s unmatched Postgres expertise, EDB Postgres AI automates database operations, ensuring expert-level execution across deployments – boosting operational efficiency by up to 30% and accelerating innovation.
        displayName: EDB Postgres AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Postgres
        - Database
        - DBaaS
        - EDB Postgres AI
        links:
        - name: EDB Postgres AI Software Deployment
          url: https://www.enterprisedb.com/products/software-deployment
        maintainers:
        - email: techsupport@enterprisedb.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - docker.enterprisedb.com/pgai-platform/edb-hcp-operator/manager@sha256:82610c624e956a67df4858929e1aabd561a7537ba68d40e6512b68f55e2649e0
        version: 1.2.0
      entries:
      - name: edb-hcp-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: edb-hcp-operator
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-documentdb-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-documentdb-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBInstance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/documentdb-controller:1.1.1
          createdAt: "2025-09-19T22:13:37Z"
          description: AWS DocumentDB controller is a service controller for managing
            DocumentDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DBCluster represents the state of an AWS documentdb DBCluster
              resource.
            displayName: DBCluster
            kind: DBCluster
            name: dbclusters.documentdb.services.k8s.aws
            version: v1alpha1
          - description: DBInstance represents the state of an AWS documentdb DBInstance
              resource.
            displayName: DBInstance
            kind: DBInstance
            name: dbinstances.documentdb.services.k8s.aws
            version: v1alpha1
          - description: DBSubnetGroup represents the state of an AWS documentdb DBSubnetGroup
              resource.
            displayName: DBSubnetGroup
            kind: DBSubnetGroup
            name: dbsubnetgroups.documentdb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon DocumentDB resources in AWS from within your Kubernetes cluster.

          **About Amazon DocumentDB**

          Amazon DocumentDB (with MongoDB compatibility) is a fast, reliable, and fully managed database service. Amazon DocumentDB makes it easy to set up, operate, and scale MongoDB-compatible databases in the cloud. With Amazon DocumentDB, you can run the same application code and use the same drivers and tools that you use with MongoDB.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon DocumentDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - documentdb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon DocumentDB Developer Resources
          url: https://aws.amazon.com/documentdb/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: documentdb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/documentdb-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-documentdb-controller.v1.1.1
        version: 1.1.1
      - name: ack-documentdb-controller.v1.1.0
        version: 1.1.0
      - name: ack-documentdb-controller.v1.0.14
        version: 1.0.14
      - name: ack-documentdb-controller.v1.0.13
        version: 1.0.13
      - name: ack-documentdb-controller.v1.0.12
        version: 1.0.12
      - name: ack-documentdb-controller.v1.0.11
        version: 1.0.11
      - name: ack-documentdb-controller.v1.0.10
        version: 1.0.10
      - name: ack-documentdb-controller.v1.0.9
        version: 1.0.9
      - name: ack-documentdb-controller.v1.0.7
        version: 1.0.7
      - name: ack-documentdb-controller.v1.0.6
        version: 1.0.6
      - name: ack-documentdb-controller.v1.0.5
        version: 1.0.5
      - name: ack-documentdb-controller.v1.0.2
        version: 1.0.2
      - name: ack-documentdb-controller.v1.0.1
        version: 1.0.1
      - name: ack-documentdb-controller.v1.0.0
        version: 1.0.0
      - name: ack-documentdb-controller.v0.0.9
        version: 0.0.9
      - name: ack-documentdb-controller.v0.0.8
        version: 0.0.8
      - name: ack-documentdb-controller.v0.0.7
        version: 0.0.7
      - name: ack-documentdb-controller.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: ack-documentdb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: k8gb.io
      provider-url: https://github.com/k8gb-io/k8gb
    name: k8gb
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k8gb.v0.11.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"k8gb.absa.oss/v1beta1","kind":"Gslb","metadata":{"name":"test-gslb-failover","namespace":"test-gslb"},"spec":{"ingress":{"rules":[{"host":"failover.test.k8gb.io","http":{"paths":[{"backend":{"service":{"name":"frontend-podinfo","port":{"name":"http"}}},"path":"/"}]}}]},"strategy":{"primaryGeoTag":"eu-west-1","type":"failover"}}}]'
          capabilities: Seamless Upgrades
          categories: Networking
          certified: "false"
          containerImage: docker.io/absaoss/k8gb:v0.11.4
          createdAt: "2021-09-24 12:00:00"
          description: A cloud native Kubernetes Global Balancer
          operatorframework.io/suggested-namespace: k8gb
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: k8gb
          operators.operatorframework.io/builder: operator-sdk-v1.12.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/k8gb-io/k8gb
          support: cncf-k8gb-maintainers@lists.cncf.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gslb resource for global load balancing strategy configuration
            displayName: Gslb
            kind: Gslb
            name: gslbs.k8gb.absa.oss
            version: v1beta1
          - description: Using ExternalDNS it synchronizes exposed Kubernetes Services
              and Ingresses with DNS providers
            displayName: DNSEndpoint
            kind: DNSEndpoint
            name: dnsendpoints.externaldns.k8s.io
            version: v1alpha1
        description: |
          **A cloud native Kubernetes Global Balancer**

          A Global Service Load Balancing solution with a focus on having cloud native qualities and work natively in a Kubernetes context.

          Key Differentiators:
           - Load balancing is based on timeproof DNS protocol which is perfect for global scope and extremely reliable
           - No dedicated management cluster and no single point of failure
           - Kubernetes native application health checks utilizing status of Liveness and Readiness probes for load balancing decisions
           - Configuration with a single Kubernetes CRD of `Gslb` kind

          Operator needs to be configured by setting couple of environment variables. This can be done by modifying
          the subscription yaml in the web ui or via kubectl. Here is an example of changing the `EDGE_DNS_ZONE`:

          ```yaml
          kind: Subscription
            ...
          spec:
            ...
            config:
              env:
              - name: CLUSTER_GEO_TAG
                value: us
              - name: EDGE_DNS_ZONE
                value: mycloud.example.com
                ...
           ```
           For more deployment variables see [helm code](https://github.com/k8gb-io/k8gb/blob/v0.8.3/chart/k8gb/templates/operator.yaml#L53:L68).

           For more information see [k8gb.io](https://k8gb.io).
        displayName: k8gb
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gslb
        - dns-lb
        - kuberneters-global-balancer
        - kubernetes-operator
        - balancer
        - multi-cluster
        links:
        - name: Home
          url: https://www.k8gb.io/
        - name: Source
          url: https://github.com/k8gb-io/k8gb
        maintainers:
        - email: dinar.valeev@absa.africa
          name: Dinar Valeev
        - email: jiri.kremser@gmail.com
          name: Jiri Kremser
        - email: kuritka@gmail.com
          name: Michal Kuritka
        - email: timofey.ilinykh@absa.africa
          name: Timofey Ilinykh
        - email: yury.tsarev@absa.africa
          name: Yury Tsarev
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: k8gb.io
          url: https://github.com/k8gb-io/k8gb
        relatedImages:
        - absaoss/k8s_crd:v0.0.11
        - docker.io/absaoss/k8gb:v0.11.4
        version: 0.11.4
      entries:
      - name: k8gb.v0.11.4
        version: 0.11.4
      name: alpha
    defaultChannel: alpha
    packageName: k8gb
    provider:
      name: k8gb.io
      url: https://github.com/k8gb-io/k8gb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Internal
      provider-url: redhat.com
    name: l5-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: l5-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pets.bestie.com/v1",
                "kind": "Bestie",
                "metadata": {
                  "name": "bestie"
                },
                "spec": {
                  "image": "quay.io/opdev/bestie",
                  "maxReplicas": 10,
                  "size": 3,
                  "version": "1.3"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Application Runtime
          certified: "false"
          containerImage: quay.io/opdev/l5-operator:v0.0.3
          createdAt: "2023-06-27T16:56:10Z"
          description: L5 Operator is a demo operator that aims to demonstrate level
            5 capabilities for web applications.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          relatedImages: quay.io/opdev/bestie@sha256:0e99ada2b1de19043845dd9076789584aeddc2646402e5f0987dd00424be1ee7
          repository: https://github.com/opdev/l5-operator-demo
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Bestie is the Schema for the besties API.
            displayName: Bestie
            kind: Bestie
            name: besties.pets.bestie.com
            version: v1
        description: |
          ## Usage
          The L5 operator is meant to demonstrate a minimum viable prototype of all capability levels outilned in the operator framework. The L5 operator is compatible with both Kubernetes and OpenShift clusters. When running on vanilla kubernetes a few prerequisites need to be kept in mind which may not be available out of the box. These include an ingress controller and a prometheus operator for level 4 capability. The L5 operator also depends on the Crunchy Data PostgreSQL operator. On openshift when installing via the Operator Lifecycle Manager, this will automatically be installed when installing the operator via the embdeded operator hub. The Operator Lifecycle Manager is also available for vanilla kubernetes however it may not be installed out of the box. In this case the Crunchy Data PostgreSQL operator must be installed manually. The Crunchy Data PostgreSQL operator can be installed on vanilla kuberentes using the [instructions provided by Crunchy Data](https://access.crunchydata.com/documentation/postgres-operator/4.7.5/installation/postgres-operator/) More documentation and examples can be found on [github](https://github.com/opdev/l5-operator-demo).
        displayName: L5 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - l5-operator
        links:
        - name: L5 Operator
          url: https://opdev.github.io/l5-operator-demo/
        maintainers:
        - email: partnereng@redhat.com
          name: OpDev Team
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat Internal
          url: redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/opdev/l5-operator:v0.0.3
        version: 0.0.3
      entries:
      - name: l5-operator.v0.0.3
        version: 0.0.3
      - name: l5-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: l5-operator
    provider:
      name: Red Hat Internal
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
    name: aikit-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aikit-operator.v2021.2.102120
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aikit.intel/v1alpha1",
                "kind": "AIKitContainer",
                "metadata": {
                  "name": "intel-aikit-container",
                  "namespace": "redhat-ods-applications"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagestream": {
                    "name": "oneapi-aikit",
                    "namespace": "redhat-ods-applications",
                    "registry": {
                      "root": "registry.connect.redhat.com",
                      "repo": "intel",
                      "name": "oneapi-aikit",
                      "version": "latest"
                    }
                  },
                  "nameOverride": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AIKitContainer
            name: aikitcontainers.aikit.intel
            version: v1alpha1
        description: |
          The Intel® oneAPI AI Analytics Toolkit gives data scientists, AI developers, and researchers familiar Python* tools and frameworks to accelerate end-to-end data science and machine learning pipelines on Intel® architectures. The components are built using oneAPI libraries for low-level compute optimizations. This toolkit maximizes performance from preprocessing through machine and deep learning phases and provides interoperability for efficient model development and deployment across single and multinodes.
        displayName: Intel® oneAPI AI Analytics Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aikit-operator
        links:
        - name: Aikit Operator
          url: https://aikit-operator.domain
        maintainers:
        - email: kam.d.kasravi@intel.com
          name: Kam Kasravi
        - email: abolfazl.shahbazi@intel.com
          name: Abolfazl Shahbazi
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Intel
          url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
        relatedImages:
        - registry.connect.redhat.com/intel/aikit-operator@sha256:7abb5e6164ffb3e4b87c39061f45d09848218fcfb74b48bf5e1035d56b41ee0c
        version: 2021.2.102120
      entries:
      - name: aikit-operator.v2021.2.102120
        version: 2021.2.102120
      - name: aikit-operator.v2021.2.100410
        version: 2021.2.100410
      - name: aikit-operator.v2021.2.92911
        version: 2021.2.92911
      - name: aikit-operator.v2021.2.92814
        version: 2021.2.92814
      - name: aikit-operator.v2021.2.92808-dev-a60e5ba
        version: 2021.2.92808-dev-a60e5ba
      name: alpha
    defaultChannel: alpha
    packageName: aikit-operator
    provider:
      name: Intel
      url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-airflow-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airflow-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:d8f8c439f33e5a80e16c7867a008d71693fd4928357ac6919c77bd0ff0bb1339
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:d8f8c439f33e5a80e16c7867a008d71693fd4928357ac6919c77bd0ff0bb1339
        version: 24.11.1
      entries:
      - name: airflow-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: airflow-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:173363db3cb0997aff09fe4b7f3a54c04f22bc56db83d2b30757e3cc745c4587
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:173363db3cb0997aff09fe4b7f3a54c04f22bc56db83d2b30757e3cc745c4587
        version: 25.3.0
      entries:
      - name: airflow-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: airflow-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
        version: 25.7.0
      entries:
      - name: airflow-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: airflow-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
        version: 25.7.0
      entries:
      - name: airflow-operator.v25.7.0
        version: 25.7.0
      - name: airflow-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-airflow-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: windows-machine-config-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: windows-machine-config-operator.v10.16.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          createdAt: "2025-06-06T22:08:07Z"
          description: An operator that enables Windows container workloads on OCP
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=10.15.0 <10.16.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.openshift.io/valid-subscription: '["Red Hat OpenShift support
            for Windows Containers"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ### Introduction
          The Windows Machine Config Operator configures Windows Machines into nodes, enabling Windows container workloads to
          be run on OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/4.16/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://docs.openshift.com/container-platform/4.16/windows_containers/byoh-windows-instance.html)
          The operator completes all the necessary steps to configure the Windows instance so that it can join the cluster as a worker node.

          Usage of this operator requires an OpenShift Container Platform for Windows Containers subscription. Users looking to
          deploy Windows containers workloads in production clusters should acquire a subscription before attempting to
          install this operator. Users without a subscription can try the community operator, a distribution which lacks
          official support.

          ### Pre-requisites
          * A Red Hat OpenShift Container Platform for Windows Containers [subscription](https://access.redhat.com/support/policy/updates/openshift#windows)
          * OCP 4.16 cluster running on Azure, AWS, GCP or vSphere configured with hybrid OVN Kubernetes networking
          * [WMCO prerequisites](https://docs.openshift.com/container-platform/4.16/windows_containers/understanding-windows-container-workloads.html#wmco-prerequisites__understanding-windows-container-workloads)

          ### Usage
          Once the `openshift-windows-machine-config-operator` namespace has been created, a secret must be created containing
          the private key that will be used to access the Windows VMs:
          ```
          # Create secret containing the private key in the openshift-windows-machine-config-operator namespace
          oc create secret generic cloud-private-key --from-file=private-key.pem=/path/to/key -n openshift-windows-machine-config-operator
          ```
          We strongly recommend not using the same
          [private key](https://docs.openshift.com/container-platform/4.16/installing/installing_azure/installing-azure-default.html#ssh-agent-using_installing-azure-default)
          used when installing the cluster

          Below is an example of a vSphere Windows MachineSet which can create Windows Machines that the WMCO can react upon.
          Please note that the windows-user-data secret will be created by the WMCO lazily when it is configuring the first
          Windows Machine. After that, the windows-user-data will be available for the subsequent MachineSets to be consumed.
          It might take around 10 minutes for the Windows VM to be configured so that it joins the cluster. Please note that
          the MachineSet should have following labels:
          * `machine.openshift.io/os-id: Windows`
          * `machine.openshift.io/cluster-api-machine-role: worker`
          * `machine.openshift.io/cluster-api-machine-type: worker`

          The following label must be added to the Machine spec within the MachineSet spec:
          * `node-role.kubernetes.io/worker: ""`

          Not having these labels will result in the Windows node not being marked as a worker.

          `<infrastructureID>` should be replaced with the output of:
          ```
          oc get -o jsonpath='{.status.infrastructureName}{"\n"}' infrastructure cluster
          ```

          The following template variables need to be replaced as follows with values from your vSphere environment:
          * `<Windows_VM_template>`: template name
          * `<VM Network Name>`: network name
          * `<vCenter DataCenter Name>`: datacenter name
          * `<Path to VM Folder in vCenter>`: path where your OpenShift cluster is running
          * `<vCenter Datastore Name>`: datastore name
          * `<vCenter Server FQDN/IP>`: IP address or FQDN of the vCenter server

          Please note that on vSphere, Windows Machine names cannot be more than 15 characters long. The MachineSet name therefore
          cannot be more than 9 characters long, due to the way Machine names are generated from it.
          ```
          apiVersion: machine.openshift.io/v1beta1
          kind: MachineSet
          metadata:
            labels:
              machine.openshift.io/cluster-api-cluster: <infrastructureID>
            name: winworker
            namespace: openshift-machine-api
          spec:
            replicas: 1
            selector:
              matchLabels:
                machine.openshift.io/cluster-api-cluster: <infrastructureID>
                machine.openshift.io/cluster-api-machineset: winworker
            template:
              metadata:
                labels:
                  machine.openshift.io/cluster-api-cluster: <infrastructureID>
                  machine.openshift.io/cluster-api-machine-role: worker
                  machine.openshift.io/cluster-api-machine-type: worker
                  machine.openshift.io/cluster-api-machineset: winworker
                  machine.openshift.io/os-id: Windows
              spec:
                metadata:
                  labels:
                    node-role.kubernetes.io/worker: ""
                providerSpec:
                  value:
                    apiVersion: vsphereprovider.openshift.io/v1beta1
                    credentialsSecret:
                      name: vsphere-cloud-credentials
                      namespace: openshift-machine-api
                    diskGiB: 128
                    kind: VSphereMachineProviderSpec
                    memoryMiB: 16384
                    metadata:
                      creationTimestamp: null
                    network:
                      devices:
                      - networkName: "<VM Network Name>"
                    numCPUs: 4
                    numCoresPerSocket: 1
                    snapshot: ""
                    template: <Windows_VM_template
                    userDataSecret:
                      name: windows-user-data
                    workspace:
                      datacenter: <vCenter DataCenter Name>
                      datastore: <vCenter Datastore Name>
                      folder: <Path to VM Folder in vCenter> # e.g. /DC/vm/ocp45-2tdrm
                      server: <vCenter Server FQDN/IP>
          ```
          Example MachineSet for other cloud providers:
          - [AWS](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-aws.html)
          - [Azure](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-azure.html)
          - [GCP](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-gcp.html)

          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.

          ### Reporting issues
          Support for this distribution of WMCO requires a Red Hat OpenShfit subscription. Support should be requested through the Red Hat Customer Portal.
          Please read through the [troubleshooting document](https://docs.openshift.com/container-platform/4.16/support/troubleshooting/troubleshooting-windows-container-workload-issues.html)
          before opening a support case.
        displayName: Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4-wincw/windows-machine-config-rhel9-operator@sha256:c9a6166158e1ccb1764007d79c6e0f879505c3546eb45cbbe79ebeefcc78a728
        version: 10.16.2
      entries:
      - name: windows-machine-config-operator.v10.16.2
        version: 10.16.2
      - name: windows-machine-config-operator.v10.16.1
        version: 10.16.1
      - name: windows-machine-config-operator.v10.16.0
        version: 10.16.0
      name: preview
    - currentCSV: windows-machine-config-operator.v10.16.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          createdAt: "2025-06-06T22:08:07Z"
          description: An operator that enables Windows container workloads on OCP
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=10.15.0 <10.16.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.openshift.io/valid-subscription: '["Red Hat OpenShift support
            for Windows Containers"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ### Introduction
          The Windows Machine Config Operator configures Windows Machines into nodes, enabling Windows container workloads to
          be run on OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/4.16/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://docs.openshift.com/container-platform/4.16/windows_containers/byoh-windows-instance.html)
          The operator completes all the necessary steps to configure the Windows instance so that it can join the cluster as a worker node.

          Usage of this operator requires an OpenShift Container Platform for Windows Containers subscription. Users looking to
          deploy Windows containers workloads in production clusters should acquire a subscription before attempting to
          install this operator. Users without a subscription can try the community operator, a distribution which lacks
          official support.

          ### Pre-requisites
          * A Red Hat OpenShift Container Platform for Windows Containers [subscription](https://access.redhat.com/support/policy/updates/openshift#windows)
          * OCP 4.16 cluster running on Azure, AWS, GCP or vSphere configured with hybrid OVN Kubernetes networking
          * [WMCO prerequisites](https://docs.openshift.com/container-platform/4.16/windows_containers/understanding-windows-container-workloads.html#wmco-prerequisites__understanding-windows-container-workloads)

          ### Usage
          Once the `openshift-windows-machine-config-operator` namespace has been created, a secret must be created containing
          the private key that will be used to access the Windows VMs:
          ```
          # Create secret containing the private key in the openshift-windows-machine-config-operator namespace
          oc create secret generic cloud-private-key --from-file=private-key.pem=/path/to/key -n openshift-windows-machine-config-operator
          ```
          We strongly recommend not using the same
          [private key](https://docs.openshift.com/container-platform/4.16/installing/installing_azure/installing-azure-default.html#ssh-agent-using_installing-azure-default)
          used when installing the cluster

          Below is an example of a vSphere Windows MachineSet which can create Windows Machines that the WMCO can react upon.
          Please note that the windows-user-data secret will be created by the WMCO lazily when it is configuring the first
          Windows Machine. After that, the windows-user-data will be available for the subsequent MachineSets to be consumed.
          It might take around 10 minutes for the Windows VM to be configured so that it joins the cluster. Please note that
          the MachineSet should have following labels:
          * `machine.openshift.io/os-id: Windows`
          * `machine.openshift.io/cluster-api-machine-role: worker`
          * `machine.openshift.io/cluster-api-machine-type: worker`

          The following label must be added to the Machine spec within the MachineSet spec:
          * `node-role.kubernetes.io/worker: ""`

          Not having these labels will result in the Windows node not being marked as a worker.

          `<infrastructureID>` should be replaced with the output of:
          ```
          oc get -o jsonpath='{.status.infrastructureName}{"\n"}' infrastructure cluster
          ```

          The following template variables need to be replaced as follows with values from your vSphere environment:
          * `<Windows_VM_template>`: template name
          * `<VM Network Name>`: network name
          * `<vCenter DataCenter Name>`: datacenter name
          * `<Path to VM Folder in vCenter>`: path where your OpenShift cluster is running
          * `<vCenter Datastore Name>`: datastore name
          * `<vCenter Server FQDN/IP>`: IP address or FQDN of the vCenter server

          Please note that on vSphere, Windows Machine names cannot be more than 15 characters long. The MachineSet name therefore
          cannot be more than 9 characters long, due to the way Machine names are generated from it.
          ```
          apiVersion: machine.openshift.io/v1beta1
          kind: MachineSet
          metadata:
            labels:
              machine.openshift.io/cluster-api-cluster: <infrastructureID>
            name: winworker
            namespace: openshift-machine-api
          spec:
            replicas: 1
            selector:
              matchLabels:
                machine.openshift.io/cluster-api-cluster: <infrastructureID>
                machine.openshift.io/cluster-api-machineset: winworker
            template:
              metadata:
                labels:
                  machine.openshift.io/cluster-api-cluster: <infrastructureID>
                  machine.openshift.io/cluster-api-machine-role: worker
                  machine.openshift.io/cluster-api-machine-type: worker
                  machine.openshift.io/cluster-api-machineset: winworker
                  machine.openshift.io/os-id: Windows
              spec:
                metadata:
                  labels:
                    node-role.kubernetes.io/worker: ""
                providerSpec:
                  value:
                    apiVersion: vsphereprovider.openshift.io/v1beta1
                    credentialsSecret:
                      name: vsphere-cloud-credentials
                      namespace: openshift-machine-api
                    diskGiB: 128
                    kind: VSphereMachineProviderSpec
                    memoryMiB: 16384
                    metadata:
                      creationTimestamp: null
                    network:
                      devices:
                      - networkName: "<VM Network Name>"
                    numCPUs: 4
                    numCoresPerSocket: 1
                    snapshot: ""
                    template: <Windows_VM_template
                    userDataSecret:
                      name: windows-user-data
                    workspace:
                      datacenter: <vCenter DataCenter Name>
                      datastore: <vCenter Datastore Name>
                      folder: <Path to VM Folder in vCenter> # e.g. /DC/vm/ocp45-2tdrm
                      server: <vCenter Server FQDN/IP>
          ```
          Example MachineSet for other cloud providers:
          - [AWS](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-aws.html)
          - [Azure](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-azure.html)
          - [GCP](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-gcp.html)

          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.

          ### Reporting issues
          Support for this distribution of WMCO requires a Red Hat OpenShfit subscription. Support should be requested through the Red Hat Customer Portal.
          Please read through the [troubleshooting document](https://docs.openshift.com/container-platform/4.16/support/troubleshooting/troubleshooting-windows-container-workload-issues.html)
          before opening a support case.
        displayName: Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4-wincw/windows-machine-config-rhel9-operator@sha256:c9a6166158e1ccb1764007d79c6e0f879505c3546eb45cbbe79ebeefcc78a728
        version: 10.16.2
      entries:
      - name: windows-machine-config-operator.v10.16.2
        version: 10.16.2
      - name: windows-machine-config-operator.v10.16.1
        version: 10.16.1
      - name: windows-machine-config-operator.v10.16.0
        version: 10.16.0
      name: stable
    defaultChannel: stable
    packageName: windows-machine-config-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JohnSnowLABS
      provider-url: nlp.johnsnowlabs.com
    name: annotationlab
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: annotationlab.v5.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.johnsnowlabs.com/v1",
                "kind": "Annotationlab",
                "metadata": {
                  "name": "annotationlab"
                },
                "spec": {
                  "activelearning": {
                    "extraEnv": "",
                    "extraVolumeMounts": "",
                    "extraVolumes": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/johnsnowlabs/annotationlab",
                      "tag": "active-learning-5.7.0"
                    },
                    "lifecycleHooks": ""
                  },
                  "affinity": "podAntiAffinity:\n  requiredDuringSchedulingIgnoredDuringExecution:\n    - labelSelector:\n        matchLabels:\n          {{- include \"annotationlab.selectorLabels\" . | nindent 10 }}\n        matchExpressions:\n          - key: role\n            operator: NotIn\n            values:\n              - test\n      topologyKey: kubernetes.io/hostname\n  preferredDuringSchedulingIgnoredDuringExecution:\n    - weight: 100\n      podAffinityTerm:\n        labelSelector:\n          matchLabels:\n            {{- include \"annotationlab.selectorLabels\" . | nindent 12 }}\n          matchExpressions:\n            - key: role\n              operator: NotIn\n              values:\n                - test\n        topologyKey: failure-domain.beta.kubernetes.io/zone\n",
                  "airflow": {
                    "activelearning": {
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/johnsnowlabs/annotationlab",
                        "tag": "active-learning-5.7.0"
                      }
                    },
                    "affinity": {},
                    "airflowConfigAnnotations": {},
                    "airflowHome": "/opt/airflow",
                    "airflowLocalSettings": "{{- if semverCompare \"\u003e=2.2.0\" .Values.airflowVersion }}\n{{- if not (or .Values.webserverSecretKey .Values.webserverSecretKeySecretName) }}\nfrom airflow.www.utils import UIAlert\n\nDASHBOARD_UIALERTS = [\n  UIAlert(\n    'Usage of a dynamic webserver secret key detected. We recommend a static webserver secret key instead.'\n    ' See the \u003ca href='\n    '\"https://airflow.apache.org/docs/helm-chart/stable/production-guide.html#webserver-secret-key\"\u003e'\n    'Helm Chart Production Guide\u003c/a\u003e for more details.',\n    category=\"warning\",\n    roles=[\"Admin\"],\n    html=True,\n  )\n]\n{{- end }}\n{{- end }}",
                    "airflowPodAnnotations": {},
                    "airflowVersion": "2.5.1",
                    "allowPodLaunching": true,
                    "cleanup": {
                      "affinity": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow kubernetes cleanup-pods --namespace={{ .Release.Namespace }}"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "failedJobsHistoryLimit": null,
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "resources": {},
                      "schedule": "*/15 * * * *",
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "successfulJobsHistoryLimit": null,
                      "tolerations": [],
                      "topologySpreadConstraints": []
                    },
                    "config": {
                      "api": {
                        "auth_backend": "airflow.api.auth.backend.default"
                      },
                      "celery": {
                        "flower_url_prefix": "{{ .Values.ingress.flower.path }}",
                        "worker_concurrency": 16
                      },
                      "celery_kubernetes_executor": {
                        "kubernetes_queue": "kubernetes"
                      },
                      "core": {
                        "colored_console_log": "False",
                        "dags_are_paused_at_creation": "False",
                        "dags_folder": "{{ include \"airflow_dags\" . }}",
                        "executor": "{{ .Values.executor }}",
                        "load_examples": "False",
                        "remote_logging": "{{- ternary \"True\" \"False\" .Values.elasticsearch.enabled }}"
                      },
                      "elasticsearch": {
                        "json_format": "True",
                        "log_id_template": "{dag_id}_{task_id}_{execution_date}_{try_number}"
                      },
                      "elasticsearch_configs": {
                        "max_retries": 3,
                        "retry_timeout": "True",
                        "timeout": 30
                      },
                      "kerberos": {
                        "ccache": "{{ .Values.kerberos.ccacheMountPath }}/{{ .Values.kerberos.ccacheFileName }}",
                        "keytab": "{{ .Values.kerberos.keytabPath }}",
                        "principal": "{{ .Values.kerberos.principal }}",
                        "reinit_frequency": "{{ .Values.kerberos.reinitFrequency }}"
                      },
                      "kubernetes": {
                        "airflow_configmap": "{{ include \"airflow_config\" . }}",
                        "airflow_local_settings_configmap": "{{ include \"airflow_config\" . }}",
                        "multi_namespace_mode": "{{ ternary \"True\" \"False\" .Values.multiNamespaceMode }}",
                        "namespace": "{{ .Release.Namespace }}",
                        "pod_template_file": "{{ include \"airflow_pod_template_file\" . }}/pod_template_file.yaml",
                        "worker_container_repository": "{{ .Values.images.airflow.repository | default .Values.defaultAirflowRepository }}",
                        "worker_container_tag": "{{ .Values.images.airflow.tag | default .Values.defaultAirflowTag }}"
                      },
                      "logging": {
                        "colored_console_log": "False",
                        "remote_logging": "{{- ternary \"True\" \"False\" .Values.elasticsearch.enabled }}"
                      },
                      "metrics": {
                        "statsd_host": "{{ printf \"%s-statsd\" .Release.Name }}",
                        "statsd_on": "{{ ternary \"True\" \"False\" .Values.statsd.enabled }}",
                        "statsd_port": 9125,
                        "statsd_prefix": "airflow"
                      },
                      "scheduler": {
                        "run_duration": 41460,
                        "standalone_dag_processor": "{{ ternary \"True\" \"False\" .Values.dagProcessor.enabled }}",
                        "statsd_host": "{{ printf \"%s-statsd\" .Release.Name }}",
                        "statsd_on": "{{ ternary \"True\" \"False\" .Values.statsd.enabled }}",
                        "statsd_port": 9125,
                        "statsd_prefix": "airflow"
                      },
                      "webserver": {
                        "authenticate": "False",
                        "base_url": "http://0.0.0.0:8080",
                        "enable_proxy_fix": "True",
                        "rbac": "False"
                      }
                    },
                    "createUserJob": {
                      "affinity": {},
                      "annotations": {},
                      "applyCustomEnv": true,
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"users create\" \"create_user\" }} \"$@\"",
                        "--",
                        "-r",
                        "{{ .Values.webserver.defaultUser.role }}",
                        "-u",
                        "{{ .Values.webserver.defaultUser.username }}",
                        "-e",
                        "{{ .Values.webserver.defaultUser.email }}",
                        "-f",
                        "{{ .Values.webserver.defaultUser.firstName }}",
                        "-l",
                        "{{ .Values.webserver.defaultUser.lastName }}",
                        "-p",
                        "{{ .Values.webserver.defaultUser.password }}"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "jobAnnotations": {},
                      "labels": {},
                      "nodeSelector": {},
                      "resources": {},
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-create-user-job"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "useHelmHooks": true
                    },
                    "dagProcessor": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow dag-processor"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 60,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "dags": {
                      "gitSync": {
                        "branch": "v2-2-stable",
                        "containerName": "git-sync",
                        "depth": 1,
                        "enabled": false,
                        "env": [],
                        "extraVolumeMounts": [],
                        "maxFailures": 0,
                        "repo": "https://github.com/apache/airflow.git",
                        "resources": {},
                        "rev": "HEAD",
                        "securityContext": {},
                        "subPath": "tests/dags",
                        "uid": 65533,
                        "wait": 5
                      },
                      "persistence": {
                        "accessMode": "ReadWriteOnce",
                        "enabled": false,
                        "existingClaim": null,
                        "size": "1Gi",
                        "storageClassName": null,
                        "subPath": null
                      }
                    },
                    "data": {
                      "brokerUrl": null,
                      "brokerUrlSecretName": null,
                      "metadataConnection": {
                        "db": "airflow",
                        "pass": "airflow",
                        "port": 5432,
                        "protocol": "postgresql",
                        "sslmode": "disable",
                        "user": "airflow"
                      },
                      "metadataSecretName": null,
                      "resultBackendConnection": null,
                      "resultBackendSecretName": null
                    },
                    "defaultAirflowRepository": "apache/airflow",
                    "defaultAirflowTag": "2.5.1",
                    "elasticsearch": {
                      "connection": {},
                      "enabled": false,
                      "secretName": null
                    },
                    "enableBuiltInSecretEnvVars": {
                      "AIRFLOW_CONN_AIRFLOW_DB": true,
                      "AIRFLOW__CELERY__BROKER_URL": true,
                      "AIRFLOW__CELERY__CELERY_RESULT_BACKEND": true,
                      "AIRFLOW__CELERY__RESULT_BACKEND": true,
                      "AIRFLOW__CORE__FERNET_KEY": true,
                      "AIRFLOW__CORE__SQL_ALCHEMY_CONN": true,
                      "AIRFLOW__DATABASE__SQL_ALCHEMY_CONN": true,
                      "AIRFLOW__ELASTICSEARCH__ELASTICSEARCH_HOST": true,
                      "AIRFLOW__ELASTICSEARCH__HOST": true,
                      "AIRFLOW__WEBSERVER__SECRET_KEY": true
                    },
                    "env": [],
                    "executor": "CeleryExecutor",
                    "extraConfigMaps": {},
                    "extraEnv": "- name: ACTIVE_LEARNING_IMAGE_PULL_POLICY\n  value: \"{{ .Values.activelearning.image.pullPolicy }}\"\n- name: ACTIVE_LEARNING_IMAGE\n  value: \"{{ .Values.activelearning.image.repository }}:{{ .Values.activelearning.image.tag }}\"\n- name: NUMBER_OF_PREANNOTATOR_DAG\n  value: {{ .Values.preannotator_queues.count | quote }}\n- name: AIRFLOW__LOGGING__LOGGING_LEVEL\n  value: INFO\n- name: ANNOTATIONLAB_FULLNAME\n  value: {{ include \"annotationlab.annotationlab.fullname\" . }}\n- name: ANNOTATIONLAB_URL\n  value: http://{{ include \"annotationlab.annotationlab.fullname\" . }}:8200\n- name: AIRFLOW_NAMESPACE\n  value: {{ .Release.Namespace }}\n- name: DATABASE_CONNECTION_STRING\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"annotationlab.annotationlab.fullname\" . }}-secret\n      key: DATABASE_CONNECTION_STRING\n",
                    "extraEnvFrom": null,
                    "extraSecrets": {},
                    "fernetKey": null,
                    "fernetKeySecretName": null,
                    "flower": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"celery flower\" \"flower\" }}"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "networkPolicy": {
                        "ingress": {
                          "from": [],
                          "ports": [
                            {
                              "port": "{{ .Values.ports.flowerUI }}"
                            }
                          ]
                        }
                      },
                      "nodeSelector": {},
                      "password": null,
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "secretName": null,
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "loadBalancerIP": null,
                        "loadBalancerSourceRanges": [],
                        "ports": [
                          {
                            "name": "flower-ui",
                            "port": "{{ .Values.ports.flowerUI }}"
                          }
                        ],
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "username": null
                    },
                    "fullnameOverride": "",
                    "gid": 0,
                    "global": {},
                    "image": {
                      "repository": "docker.io/johnsnowlabs/annotationlab"
                    },
                    "images": {
                      "airflow": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": "dataflows-5.7.0"
                      },
                      "flower": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": null
                      },
                      "gitSync": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "k8s.gcr.io/git-sync/git-sync",
                        "tag": "v3.6.3"
                      },
                      "migrationsWaitTimeout": 60,
                      "pgbouncer": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "apache/airflow",
                        "tag": "airflow-pgbouncer-2021.04.28-1.14.0"
                      },
                      "pgbouncerExporter": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "apache/airflow",
                        "tag": "airflow-pgbouncer-exporter-2021.09.22-0.12.0"
                      },
                      "pod_template": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": null
                      },
                      "redis": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/redis",
                        "tag": "alpine3.16"
                      },
                      "statsd": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "quay.io/prometheus/statsd-exporter",
                        "tag": "v0.22.8"
                      },
                      "useDefaultImageForMigration": false
                    },
                    "ingress": {
                      "enabled": false,
                      "flower": {
                        "annotations": {},
                        "enabled": false,
                        "host": "",
                        "hosts": [],
                        "ingressClassName": "",
                        "path": "/",
                        "pathType": "ImplementationSpecific",
                        "tls": {
                          "enabled": false,
                          "secretName": ""
                        }
                      },
                      "web": {
                        "annotations": {},
                        "enabled": false,
                        "host": "",
                        "hosts": [],
                        "ingressClassName": "",
                        "path": "/",
                        "pathType": "ImplementationSpecific",
                        "precedingPaths": [],
                        "succeedingPaths": [],
                        "tls": {
                          "enabled": false,
                          "secretName": ""
                        }
                      }
                    },
                    "kerberos": {
                      "ccacheFileName": "cache",
                      "ccacheMountPath": "/var/kerberos-ccache",
                      "config": "# This is an example config showing how you can use templating and how \"example\" config\n# might look like. It works with the test kerberos server that we are using during integration\n# testing at Apache Airflow (see `scripts/ci/docker-compose/integration-kerberos.yml` but in\n# order to make it production-ready you must replace it with your own configuration that\n# Matches your kerberos deployment. Administrators of your Kerberos instance should\n# provide the right configuration.\n\n[logging]\ndefault = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kerberos_libs.log\"\nkdc = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kerberos_kdc.log\"\nadmin_server = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kadmind.log\"\n\n[libdefaults]\ndefault_realm = FOO.COM\nticket_lifetime = 10h\nrenew_lifetime = 7d\nforwardable = true\n\n[realms]\nFOO.COM = {\n  kdc = kdc-server.foo.com\n  admin_server = admin_server.foo.com\n}\n",
                      "configPath": "/etc/krb5.conf",
                      "enabled": false,
                      "keytabBase64Content": null,
                      "keytabPath": "/etc/airflow.keytab",
                      "principal": "airflow@FOO.COM",
                      "reinitFrequency": 3600
                    },
                    "kubeVersionOverride": "",
                    "labels": {},
                    "limits": [],
                    "logs": {
                      "persistence": {
                        "enabled": false,
                        "existingClaim": null,
                        "size": "100Gi",
                        "storageClassName": null
                      }
                    },
                    "migrateDatabaseJob": {
                      "affinity": {},
                      "annotations": {},
                      "applyCustomEnv": true,
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"db upgrade\" \"upgradedb\" }}"
                      ],
                      "command": null,
                      "enabled": true,
                      "extraContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "jobAnnotations": {},
                      "nodeSelector": {},
                      "resources": {},
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-migrate-database-job"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "useHelmHooks": true
                    },
                    "multiNamespaceMode": false,
                    "nameOverride": "",
                    "networkPolicies": {
                      "enabled": false
                    },
                    "nodeSelector": {},
                    "pgbouncer": {
                      "affinity": {},
                      "annotations": {},
                      "args": null,
                      "auth_file": "/etc/pgbouncer/users.txt",
                      "auth_type": "md5",
                      "ciphers": "normal",
                      "command": [
                        "pgbouncer",
                        "-u",
                        "nobody",
                        "/etc/pgbouncer/pgbouncer.ini"
                      ],
                      "configSecretName": null,
                      "enabled": false,
                      "extraIni": null,
                      "extraIniMetadata": null,
                      "extraIniResultBackend": null,
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "logConnections": 0,
                      "logDisconnections": 0,
                      "maxClientConn": 100,
                      "metadataPoolSize": 10,
                      "metricsExporterSidecar": {
                        "resources": {},
                        "sslmode": "disable"
                      },
                      "nodeSelector": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "resultBackendPoolSize": 5,
                      "revisionHistoryLimit": null,
                      "service": {
                        "extraAnnotations": {}
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "ssl": {
                        "ca": null,
                        "cert": null,
                        "key": null
                      },
                      "sslmode": "prefer",
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 65534,
                      "verbose": 0
                    },
                    "podTemplate": null,
                    "ports": {
                      "airflowUI": 8080,
                      "flowerUI": 5555,
                      "pgbouncer": 6543,
                      "pgbouncerScrape": 9127,
                      "redisDB": 6379,
                      "statsdIngest": 9125,
                      "statsdScrape": 9102,
                      "workerLogs": 8793
                    },
                    "postgresql": {
                      "auth": {
                        "enablePostgresUser": true,
                        "password": null,
                        "postgresPassword": "postgres",
                        "username": null
                      },
                      "common": {
                        "exampleValue": "common-chart",
                        "global": {
                          "postgresql": {}
                        }
                      },
                      "commonAnnotations": {},
                      "enabled": true,
                      "extraEnv": [],
                      "global": {
                        "postgresql": {}
                      },
                      "image": {
                        "debug": false,
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgresql",
                        "tag": "11"
                      },
                      "ldap": {
                        "baseDN": "",
                        "bindDN": "",
                        "enabled": false,
                        "port": "",
                        "prefix": "",
                        "scheme": "",
                        "search_attr": "",
                        "search_filter": "",
                        "server": "",
                        "suffix": "",
                        "tls": false,
                        "url": ""
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 30,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "master": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": [],
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "metrics": {
                        "enabled": false,
                        "extraEnvVars": {},
                        "image": {
                          "pullPolicy": "IfNotPresent",
                          "registry": "docker.io",
                          "repository": "bitnami/postgres-exporter",
                          "tag": "0.8.0-debian-10-r166"
                        },
                        "livenessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "prometheusRule": {
                          "additionalLabels": {},
                          "enabled": false,
                          "namespace": "",
                          "rules": []
                        },
                        "readinessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "securityContext": {},
                        "service": {
                          "annotations": {
                            "prometheus.io/port": "9187",
                            "prometheus.io/scrape": "true"
                          },
                          "type": "ClusterIP"
                        },
                        "serviceMonitor": {
                          "additionalLabels": {},
                          "enabled": false
                        }
                      },
                      "nameOverride": "airflow-postgresql",
                      "networkPolicy": {
                        "allowExternal": true,
                        "enabled": false,
                        "explicitNamespacesSelector": {}
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "enabled": true,
                        "mountPath": "/bitnami/postgresql",
                        "size": "8Gi",
                        "subPath": ""
                      },
                      "postgresqlDataDir": "/bitnami/postgresql/data",
                      "postgresqlDatabase": "airflow",
                      "postgresqlPassword": "airflow",
                      "postgresqlUsername": "airflow",
                      "psp": {
                        "create": false
                      },
                      "rbac": {
                        "create": false
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "replication": {
                        "applicationName": "my_application",
                        "enabled": false,
                        "numSynchronousReplicas": 0,
                        "password": "repl_password",
                        "slaveReplicas": 1,
                        "synchronousCommit": "off",
                        "user": "repl_user"
                      },
                      "resources": {
                        "requests": {
                          "cpu": "250m",
                          "memory": "256Mi"
                        }
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "port": 5432,
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "enabled": true,
                        "name": "annotationlab-airflow-postgresql"
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": true
                        },
                        "enabled": true
                      },
                      "slave": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "tls": {
                        "certFilename": "",
                        "certKeyFilename": "",
                        "certificatesSecret": "",
                        "enabled": false,
                        "preferServerCiphers": true
                      },
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "volumePermissions": {
                        "enabled": false,
                        "image": {
                          "pullPolicy": "Always",
                          "registry": "docker.io",
                          "repository": "bitnami/minideb",
                          "tag": "buster"
                        },
                        "securityContext": {}
                      }
                    },
                    "preannotator_queues": {
                      "count": 5
                    },
                    "quotas": {},
                    "rbac": {
                      "create": true,
                      "createSCCRoleBinding": false
                    },
                    "redis": {
                      "affinity": {},
                      "enabled": true,
                      "nodeSelector": {},
                      "password": "DefaulTRediS-P0ss",
                      "passwordSecretName": null,
                      "persistence": {
                        "annotations": {},
                        "enabled": true,
                        "size": "1Gi",
                        "storageClassName": null
                      },
                      "podAnnotations": {},
                      "resources": {},
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "annotationlab-airflow-redis"
                      },
                      "terminationGracePeriodSeconds": 600,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 0
                    },
                    "registry": {
                      "connection": {},
                      "secretName": "kubernetesdockerhub"
                    },
                    "registryCredentials": {
                      "pullPolicy": "IfNotPresent",
                      "registryUrl": "docker.io/annotationlab"
                    },
                    "revisionHistoryLimit": null,
                    "scheduler": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow scheduler"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "logGroomerSidecar": {
                        "args": [
                          "bash",
                          "/clean-logs"
                        ],
                        "command": null,
                        "enabled": true,
                        "resources": {},
                        "retentionDays": 15
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-scheduler"
                      },
                      "strategy": null,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "updateStrategy": null,
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "secret": [],
                    "securityContext": {},
                    "statsd": {
                      "affinity": {},
                      "args": [
                        "--statsd.mapping-config=/etc/statsd-exporter/mappings.yml"
                      ],
                      "enabled": false,
                      "extraMappings": [],
                      "extraNetworkPolicies": [],
                      "nodeSelector": {},
                      "overrideMappings": [],
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "securityContext": {},
                      "service": {
                        "extraAnnotations": {}
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-statsd"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 65534
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "triggerer": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow triggerer"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "annotationlab-airflow-triggerer"
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 60,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "uid": 50000,
                    "volumeMounts": [],
                    "volumes": [],
                    "webserver": {
                      "affinity": {},
                      "allowPodLogReading": true,
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow webserver"
                      ],
                      "command": null,
                      "defaultUser": {
                        "email": "admin@example.com",
                        "enabled": false,
                        "firstName": "admin",
                        "lastName": "user",
                        "password": "admin",
                        "role": "Admin",
                        "username": "admin"
                      },
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "failureThreshold": 20,
                        "initialDelaySeconds": 15,
                        "periodSeconds": 5,
                        "scheme": "HTTP",
                        "timeoutSeconds": 30
                      },
                      "networkPolicy": {
                        "ingress": {
                          "from": [],
                          "ports": [
                            {
                              "port": "{{ .Values.ports.airflowUI }}"
                            }
                          ]
                        }
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "readinessProbe": {
                        "failureThreshold": 20,
                        "initialDelaySeconds": 15,
                        "periodSeconds": 5,
                        "scheme": "HTTP",
                        "timeoutSeconds": 30
                      },
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "loadBalancerIP": null,
                        "loadBalancerSourceRanges": [],
                        "ports": [
                          {
                            "name": "airflow-ui",
                            "port": "{{ .Values.ports.airflowUI }}"
                          }
                        ],
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-webserver"
                      },
                      "strategy": null,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      },
                      "webserverConfig": null,
                      "webserverConfigConfigMapName": null
                    },
                    "webserverSecretKey": "f565cd64401775f93c820dc63fc3d4f6",
                    "webserverSecretKeySecretName": null,
                    "workers": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"celery worker\" \"worker\" }}"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "hostAliases": [],
                      "keda": {
                        "advanced": {},
                        "cooldownPeriod": 30,
                        "enabled": false,
                        "maxReplicaCount": 10,
                        "minReplicaCount": 0,
                        "namespaceLabels": {},
                        "pollingInterval": 5
                      },
                      "kerberosSidecar": {
                        "enabled": false,
                        "resources": {}
                      },
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "enabled": true,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "logGroomerSidecar": {
                        "args": [
                          "bash",
                          "/clean-logs"
                        ],
                        "command": null,
                        "enabled": true,
                        "resources": {},
                        "retentionDays": 15
                      },
                      "nodeSelector": {},
                      "persistence": {
                        "annotations": {},
                        "enabled": true,
                        "fixPermissions": false,
                        "size": "5Gi",
                        "storageClassName": null
                      },
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-worker"
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 600,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "updateStrategy": null,
                      "waitForMigrations": {
                        "env": []
                      }
                    }
                  },
                  "backup": {
                    "enable": false,
                    "files": true,
                    "image": {
                      "repository": null,
                      "tag": null
                    },
                    "s3_access_key": "",
                    "s3_bucket_fullpath": "",
                    "s3_secret_key": ""
                  },
                  "configuration": {
                    "FLASK_SECRET_KEY": "rpaQghA-9lmpLwB8qxW0FsW1TJyZZYuB60ppzswpTzM=",
                    "KEYCLOAK_CLIENT_ID": "annotationlab",
                    "KEYCLOAK_CLIENT_SECRET_KEY": "09a71c59-0351-4ce6-bc8f-8fd3feb9d2ff",
                    "KEYCLOAK_REALM_NAME": "master",
                    "PERSISTENT_STORAGE": "/projects"
                  },
                  "extraContainers": "",
                  "extraEnv": "",
                  "extraNetworkPolicies": "- namespaceSelector: {}\n",
                  "extraVolumeMounts": "",
                  "extraVolumes": "",
                  "github_oauth_origin": "https://github.oauth.johnsnowlabs.com",
                  "google_analytics": {
                    "api_secret": "WENfd00yeUxRRWl4N0Utbml5SWt3Zw==",
                    "measurement_id": "Ry1XODBIMDhTTFBZ"
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "docker.io/johnsnowlabs/annotationlab",
                    "tag": "5.7.0"
                  },
                  "ingress": {
                    "annotations": {},
                    "defaultBackend": false,
                    "enabled": true,
                    "hosts": [
                      {
                        "host": "annotationlab",
                        "path": "/"
                      }
                    ],
                    "ingressroute": {
                      "enabled": false
                    },
                    "servicePort": "http",
                    "tls": [],
                    "uploadLimitInMegabytes": 16
                  },
                  "installAirflow": true,
                  "installKeycloak": true,
                  "installPostgresql": true,
                  "keycloak": {
                    "affinity": "podAntiAffinity:\n  requiredDuringSchedulingIgnoredDuringExecution:\n    - labelSelector:\n        matchLabels:\n          {{- include \"keycloak.selectorLabels\" . | nindent 10 }}\n        matchExpressions:\n          - key: app.kubernetes.io/component\n            operator: NotIn\n            values:\n              - test\n      topologyKey: kubernetes.io/hostname\n  preferredDuringSchedulingIgnoredDuringExecution:\n    - weight: 100\n      podAffinityTerm:\n        labelSelector:\n          matchLabels:\n            {{- include \"keycloak.selectorLabels\" . | nindent 12 }}\n          matchExpressions:\n            - key: app.kubernetes.io/component\n              operator: NotIn\n              values:\n                - test\n        topologyKey: topology.kubernetes.io/zone\n",
                    "args": [],
                    "auththeme": {
                      "image": {
                        "repository": "docker.io/johnsnowlabs/annotationlab",
                        "tag": "auth-theme-5.7.0"
                      }
                    },
                    "autoscaling": {
                      "behavior": {
                        "scaleDown": {
                          "policies": [
                            {
                              "periodSeconds": 300,
                              "type": "Pods",
                              "value": 1
                            }
                          ],
                          "stabilizationWindowSeconds": 300
                        }
                      },
                      "enabled": false,
                      "labels": {},
                      "maxReplicas": 10,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 80,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 3
                    },
                    "cache": {
                      "stack": "default"
                    },
                    "clusterDomain": "cluster.local",
                    "command": [
                      "/opt/keycloak/bin/kc.sh",
                      "--verbose",
                      "start",
                      "--http-enabled=true",
                      "--http-port=8080",
                      "--hostname-strict=false",
                      "--hostname-strict-https=false",
                      "--spi-events-listener-jboss-logging-success-level=info",
                      "--spi-events-listener-jboss-logging-error-level=warn"
                    ],
                    "database": {
                      "enabled": true,
                      "postgresqlDatabase": "keycloak",
                      "postgresqlPassword": "keycloak",
                      "postgresqlUsername": "keycloak"
                    },
                    "dbchecker": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/busybox",
                        "tag": 1.32
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "32Mi"
                        },
                        "requests": {
                          "cpu": "20m",
                          "memory": "32Mi"
                        }
                      },
                      "securityContext": {}
                    },
                    "enableServiceLinks": true,
                    "extraContainers": "",
                    "extraEnv": "- name: KC_FEATURES\n  value: \"authorization,account-api,admin-api,admin-fine-grained-authz,ciba,impersonation,kerberos,preview,step-up-authentication,token-exchange,web-authn\"\n- name: KEYCLOAK_ADMIN\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"keycloak.fullname\" . }}-admincreds\n      key: user\n- name: KEYCLOAK_ADMIN_PASSWORD\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"keycloak.fullname\" . }}-admincreds\n      key: password\n- name: JAVA_OPTS_APPEND\n  value: \u003e-\n    -XX:+UseContainerSupport\n    -XX:MaxRAMPercentage=50.0\n    -Djava.awt.headless=true\n    -Djgroups.dns.query={{ include \"keycloak.fullname\" . }}-headless\n",
                    "extraEnvFrom": "",
                    "extraInitContainers": "- name: theme-provider\n  image: {{ .Values.auththeme.image.repository }}:{{ .Values.auththeme.image.tag }}\n  imagePullPolicy: IfNotPresent\n  command:\n    - sh\n  args:\n    - -c\n    - |\n      echo \"Copying theme...\"\n      cp -R /johnsnowlabs/* /theme\n  volumeMounts:\n    - name: theme\n      mountPath: /theme\n",
                    "extraPorts": [],
                    "extraServiceMonitor": {
                      "annotations": {},
                      "enabled": false,
                      "interval": "10s",
                      "labels": {},
                      "namespace": "",
                      "namespaceSelector": {},
                      "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/realms/master/metrics",
                      "port": "http",
                      "scrapeTimeout": "10s"
                    },
                    "extraVolumeMounts": "- name: admincreds\n  mountPath: /secrets/admincreds\n  readOnly: true\n- name: theme\n  mountPath: /opt/keycloak/themes/johnsnowlabs\n",
                    "extraVolumes": "- name: admincreds\n  secret:\n    secretName: '{{ include \"annotationlab.keycloak.fullname\" . }}-admincreds'\n- name: theme\n  emptyDir: {}\n",
                    "fullnameOverride": "",
                    "global": {},
                    "health": {
                      "enabled": true
                    },
                    "hostAliases": [],
                    "http": {
                      "relativePath": "/auth"
                    },
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/keycloak/keycloak",
                      "tag": "20.0.3"
                    },
                    "imagePullSecrets": [
                      {
                        "name": "kubernetesdockerhub"
                      }
                    ],
                    "ingress": {
                      "annotations": {
                        "nginx.ingress.kubernetes.io/proxy-buffer-size": "16k"
                      },
                      "console": {
                        "annotations": {},
                        "enabled": false,
                        "ingressClassName": "",
                        "rules": [
                          {
                            "host": "{{ .Release.Name }}.keycloak.example.com",
                            "paths": [
                              {
                                "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/admin",
                                "pathType": "Prefix"
                              }
                            ]
                          }
                        ],
                        "tls": []
                      },
                      "enabled": false,
                      "ingressClassName": "",
                      "labels": {},
                      "rules": [
                        {
                          "host": "keycloak.example.com",
                          "paths": [
                            {
                              "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/",
                              "pathType": "Prefix"
                            }
                          ]
                        }
                      ],
                      "servicePort": "http",
                      "tls": []
                    },
                    "lifecycleHooks": "",
                    "livenessProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health/live'\n  port: http\ninitialDelaySeconds: 0\ntimeoutSeconds: 5\n",
                    "metrics": {
                      "enabled": true
                    },
                    "nameOverride": "",
                    "networkPolicy": {
                      "egress": [],
                      "enabled": false,
                      "extraFrom": [],
                      "labels": {}
                    },
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podDisruptionBudget": {},
                    "podLabels": {},
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {},
                    "postgresql": {
                      "common": {
                        "exampleValue": "common-chart",
                        "global": {
                          "postgresql": {}
                        }
                      },
                      "commonAnnotations": {},
                      "enabled": true,
                      "extraEnv": [],
                      "global": {
                        "postgresql": {}
                      },
                      "image": {
                        "debug": false,
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgresql",
                        "tag": "11.20.0-debian-11-r12"
                      },
                      "ldap": {
                        "baseDN": "",
                        "bindDN": "",
                        "bind_password": null,
                        "enabled": false,
                        "port": "",
                        "prefix": "",
                        "scheme": "",
                        "search_attr": "",
                        "search_filter": "",
                        "server": "",
                        "suffix": "",
                        "tls": false,
                        "url": ""
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 30,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "master": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": [],
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "metrics": {
                        "enabled": false,
                        "extraEnvVars": {},
                        "image": {
                          "pullPolicy": "IfNotPresent",
                          "registry": "docker.io",
                          "repository": "bitnami/postgres-exporter",
                          "tag": "0.8.0-debian-10-r166"
                        },
                        "livenessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "prometheusRule": {
                          "additionalLabels": {},
                          "enabled": false,
                          "namespace": "",
                          "rules": []
                        },
                        "readinessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "securityContext": {},
                        "service": {
                          "annotations": {
                            "prometheus.io/port": "9187",
                            "prometheus.io/scrape": "true"
                          },
                          "loadBalancerIP": null,
                          "type": "ClusterIP"
                        },
                        "serviceMonitor": {
                          "additionalLabels": {},
                          "enabled": false
                        }
                      },
                      "nameOverride": "keycloak-postgres",
                      "networkPolicy": {
                        "allowExternal": true,
                        "enabled": false,
                        "explicitNamespacesSelector": {}
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "enabled": true,
                        "mountPath": "/bitnami/postgresql",
                        "size": "8Gi",
                        "subPath": ""
                      },
                      "postgresqlDataDir": "/bitnami/postgresql/data",
                      "postgresqlDatabase": "keycloak",
                      "postgresqlPassword": "keycloak",
                      "postgresqlUsername": "keycloak",
                      "psp": {
                        "create": false
                      },
                      "rbac": {
                        "create": false
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "replication": {
                        "applicationName": "my_application",
                        "enabled": false,
                        "numSynchronousReplicas": 0,
                        "password": "repl_password",
                        "slaveReplicas": 1,
                        "synchronousCommit": "off",
                        "user": "repl_user"
                      },
                      "resources": {
                        "requests": {
                          "cpu": "250m",
                          "memory": "256Mi"
                        }
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "port": 5432,
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "enabled": true,
                        "name": "annotationlab-keycloak-postgresql"
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": true
                        },
                        "enabled": true
                      },
                      "slave": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "tls": {
                        "certCAFilename": null,
                        "certFilename": "",
                        "certKeyFilename": "",
                        "certificatesSecret": "",
                        "crlFilename": null,
                        "enabled": false,
                        "preferServerCiphers": true
                      },
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "volumePermissions": {
                        "enabled": false,
                        "image": {
                          "pullPolicy": "Always",
                          "registry": "docker.io",
                          "repository": "bitnami/minideb",
                          "tag": "buster"
                        },
                        "securityContext": {}
                      }
                    },
                    "priorityClassName": "",
                    "prometheusRule": {
                      "annotations": {},
                      "enabled": false,
                      "labels": {},
                      "namespace": "",
                      "rules": []
                    },
                    "proxy": {
                      "enabled": true,
                      "mode": "edge"
                    },
                    "rbac": {
                      "create": false,
                      "rules": []
                    },
                    "readinessProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health/ready'\n  port: http\ninitialDelaySeconds: 10\ntimeoutSeconds: 1\n",
                    "replicas": 1,
                    "resources": {},
                    "restartPolicy": "Always",
                    "route": {
                      "annotations": {},
                      "enabled": false,
                      "host": "",
                      "labels": {},
                      "path": "/",
                      "tls": {
                        "enabled": true,
                        "insecureEdgeTerminationPolicy": "Redirect",
                        "termination": "edge"
                      }
                    },
                    "secrets": {
                      "admincreds": {
                        "stringData": {
                          "password": "secret",
                          "user": "admin"
                        }
                      }
                    },
                    "securityContext": {},
                    "service": {
                      "annotations": {},
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "httpNodePort": null,
                      "httpPort": 80,
                      "httpsNodePort": null,
                      "httpsPort": 8443,
                      "labels": {},
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "sessionAffinity": "",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "serviceAccount": {
                      "allowReadPods": false,
                      "annotations": {},
                      "automountServiceAccountToken": true,
                      "create": false,
                      "imagePullSecrets": [],
                      "labels": {},
                      "name": "annotationlab-keyclo"
                    },
                    "serviceHeadless": {
                      "annotations": {}
                    },
                    "serviceMonitor": {
                      "annotations": {},
                      "enabled": false,
                      "interval": "10s",
                      "labels": {},
                      "namespace": "",
                      "namespaceSelector": {},
                      "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/metrics",
                      "port": "http",
                      "scrapeTimeout": "10s"
                    },
                    "skipInitContainers": false,
                    "startupProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health'\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 1\nfailureThreshold: 60\nperiodSeconds: 5\n",
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {},
                    "terminationGracePeriodSeconds": 60,
                    "test": {
                      "deletionPolicy": "before-hook-creation",
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/seleniarm/standalone-chromium",
                        "tag": "111.0"
                      },
                      "podSecurityContext": {},
                      "securityContext": {}
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": null,
                    "updateStrategy": "RollingUpdate"
                  },
                  "licenseserver": {
                    "origin": "https://license.johnsnowlabs.com"
                  },
                  "lifecycleHooks": "# postStart:\n#   exec:\n#     command: [\"/bin/sh\", \"-c\", \"ls\"]\n",
                  "modelshub_origin": "https://modelshub.johnsnowlabs.com",
                  "myjohnsnowlabs": {
                    "client_id": "UIGgoiIL9szXW8Nnik0O7tPT4to87oVtWvYWfc6r",
                    "origin": "https://my.johnsnowlabs.com"
                  },
                  "networkPolicy": {
                    "enabled": false
                  },
                  "nodeSelector": {},
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "annotations": {},
                    "enabled": true,
                    "size": "8Gi"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "postgresql": {
                    "common": {
                      "exampleValue": "common-chart",
                      "global": {
                        "postgresql": {}
                      }
                    },
                    "commonAnnotations": {},
                    "extraEnv": [],
                    "global": {
                      "postgresql": {}
                    },
                    "image": {
                      "debug": false,
                      "pullPolicy": "IfNotPresent",
                      "registry": "docker.io",
                      "repository": "bitnami/postgresql",
                      "tag": "11.20.0-debian-11-r12"
                    },
                    "ldap": {
                      "baseDN": "",
                      "bindDN": "",
                      "bind_password": null,
                      "enabled": false,
                      "port": "",
                      "prefix": "",
                      "scheme": "",
                      "search_attr": "",
                      "search_filter": "",
                      "server": "",
                      "suffix": "",
                      "tls": false,
                      "url": ""
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "master": {
                      "affinity": {},
                      "annotations": {},
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podLabels": {},
                      "priorityClassName": "",
                      "service": {},
                      "sidecars": [],
                      "tolerations": []
                    },
                    "metrics": {
                      "enabled": false,
                      "extraEnvVars": {},
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgres-exporter",
                        "tag": "0.8.0-debian-10-r166"
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "prometheusRule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {
                          "prometheus.io/port": "9187",
                          "prometheus.io/scrape": "true"
                        },
                        "loadBalancerIP": null,
                        "type": "ClusterIP"
                      },
                      "serviceMonitor": {
                        "additionalLabels": {},
                        "enabled": false
                      }
                    },
                    "networkPolicy": {
                      "allowExternal": true,
                      "enabled": false,
                      "explicitNamespacesSelector": {}
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {},
                      "enabled": true,
                      "mountPath": "/bitnami/postgresql",
                      "size": "8Gi",
                      "subPath": ""
                    },
                    "postgresqlDataDir": "/bitnami/postgresql/data",
                    "postgresqlDatabase": "annotationlab",
                    "postgresqlPassword": "annotationlab",
                    "postgresqlUsername": "annotationlab",
                    "psp": {
                      "create": false
                    },
                    "rbac": {
                      "create": false
                    },
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 6,
                      "initialDelaySeconds": 5,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "replication": {
                      "applicationName": "my_application",
                      "enabled": false,
                      "numSynchronousReplicas": 0,
                      "password": "repl_password",
                      "slaveReplicas": 1,
                      "synchronousCommit": "off",
                      "user": "repl_user"
                    },
                    "resources": {
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {},
                    "service": {
                      "annotations": {},
                      "port": 5432,
                      "type": "ClusterIP"
                    },
                    "serviceAccount": {
                      "enabled": false
                    },
                    "shmVolume": {
                      "chmod": {
                        "enabled": true
                      },
                      "enabled": true
                    },
                    "slave": {
                      "affinity": {},
                      "annotations": {},
                      "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podLabels": {},
                      "priorityClassName": "",
                      "service": {},
                      "sidecars": [],
                      "tolerations": []
                    },
                    "tls": {
                      "certCAFilename": null,
                      "certFilename": "",
                      "certKeyFilename": "",
                      "certificatesSecret": "",
                      "crlFilename": null,
                      "enabled": false,
                      "preferServerCiphers": true
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "volumePermissions": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "Always",
                        "registry": "docker.io",
                        "repository": "bitnami/minideb",
                        "tag": "buster"
                      },
                      "securityContext": {}
                    }
                  },
                  "preannotator_queues": {
                    "count": 5,
                    "service": {
                      "port": 5000,
                      "portName": "http",
                      "type": "ClusterIP"
                    }
                  },
                  "priorityClassName": "",
                  "readinessProbe": "httpGet:\n  path: /healthz\n  port: http\ninitialDelaySeconds: 10\ntimeoutSeconds: 1\n",
                  "registryCredentials": {
                    "create": false,
                    "password": "XXXXX",
                    "secretName": "kubernetesdockerhub",
                    "username": "XXXXX"
                  },
                  "replicas": 1,
                  "resources": {},
                  "restartPolicy": "Always",
                  "securityContext": {
                    "enabled": false,
                    "fsGroup": 65534,
                    "fsGroupChangePolicy": "OnRootMismatch",
                    "runAsUser": 65534
                  },
                  "service": {
                    "port": 8200,
                    "portName": "http",
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": false,
                    "name": "annotationlab"
                  },
                  "sharedData": {
                    "enabled": true,
                    "images": {
                      "storageSize": "5Gi"
                    },
                    "storageSize": "100Gi"
                  },
                  "sqlalchemy_pool_size": 10,
                  "telemetry": {
                    "enabled": true
                  },
                  "terminationGracePeriodSeconds": 60,
                  "tolerations": [],
                  "updateStrategy": {
                    "type": "Recreate"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.johnsnowlabs.com/annotationlab/annotationlab-operator:v5.7.0
          createdAt: "2023-12-19T10:26:54Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/annotationlab/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/annotationlab/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: johnsnowlabs-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: NLP Lab
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Annotationlab
            name: annotationlabs.apps.johnsnowlabs.com
            version: v1
        description: "NLP Lab is a data labeling and annotation tool that helps you\nlabel
          different types of data using a simple interface with a standardized output\nformat.\n\nIt
          contains enterprise features like Project Management and sharing, SSO\nintegration,
          OpenAPI Specification 3.0 compliant API, task management, and active\nlearning/preannotation
          using Spark NLP.\n\nFurthermore, the production deployment of the application
          is done using a helm chart.\nThe same helm chart is also used to create
          an AWS AMI for an EC2 based single server \ndeployment using packer.\n\nMake
          sure you have defined shared storage such as efs/nfs/cephfs prior to installing
          the NLP Lab Operator.\nUpdate the host property in the YAML configuration
          to define the required domain name to use instead of the \ndefault hostname
          annotationlab.\n\nNLP Lab Operator Installtion guide [NLP Lab Operator][]:\n[NLP
          Lab Operator]: https://nlp.johnsnowlabs.com/docs/en/alab/install    \n"
        displayName: NLPLab
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nlplab
        links:
        - name: Annotationlab
          url: https://annotationlab.domain
        maintainers:
        - email: bhandari@johnsnowlabs.com
          name: bikash
        maturity: alpha
        provider:
          name: JohnSnowLABS
          url: nlp.johnsnowlabs.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - registry.johnsnowlabs.com/annotationlab/annotationlab-operator:v5.7.0
        version: 5.7.0
      entries:
      - name: annotationlab.v5.7.0
        version: 5.7.0
      - name: annotationlab.v5.4.1
        version: 5.4.1
      - name: annotationlab.v4.9.2
        version: 4.9.2
      name: alpha
    defaultChannel: alpha
    packageName: annotationlab
    provider:
      name: JohnSnowLABS
      url: nlp.johnsnowlabs.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NGINX Inc
      provider-url: ""
    name: nginx-ingress-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nginx-ingress-operator.v3.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.nginx.org/v1alpha1",
                "kind": "NginxIngress",
                "metadata": {
                  "name": "nginxingress-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "annotations": {},
                    "appprotect": {
                      "enable": false
                    },
                    "appprotectdos": {
                      "debug": false,
                      "enable": false,
                      "maxDaemons": 0,
                      "maxWorkers": 0,
                      "memory": 0
                    },
                    "autoscaling": {
                      "annotations": {},
                      "behavior": {},
                      "enabled": false,
                      "maxReplicas": 3,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 50,
                      "targetMemoryUtilizationPercentage": 50
                    },
                    "config": {
                      "annotations": {},
                      "entries": {}
                    },
                    "containerPort": {
                      "http": 80,
                      "https": 443
                    },
                    "customConfigMap": "",
                    "customPorts": [],
                    "defaultHTTPListenerPort": 80,
                    "defaultHTTPSListenerPort": 443,
                    "defaultTLS": {
                      "cert": "",
                      "key": "",
                      "secret": ""
                    },
                    "disableIPV6": false,
                    "dnsPolicy": "ClusterFirst",
                    "enableCertManager": false,
                    "enableCustomResources": true,
                    "enableExternalDNS": false,
                    "enableLatencyMetrics": false,
                    "enableOIDC": false,
                    "enableSSLDynamicReload": true,
                    "enableSnippets": false,
                    "enableTLSPassthrough": false,
                    "env": [],
                    "extraContainers": [],
                    "globalConfiguration": {
                      "create": false,
                      "spec": {}
                    },
                    "healthStatus": false,
                    "healthStatusURI": "/nginx-health",
                    "hostNetwork": false,
                    "hostPort": {
                      "enable": false,
                      "http": 80,
                      "https": 443
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "nginx/nginx-ingress",
                      "tag": "5.2.0-ubi"
                    },
                    "ingressClass": {
                      "create": true,
                      "name": "nginx",
                      "setAsDefaultIngress": false
                    },
                    "initContainerResources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "initContainers": [],
                    "kind": "deployment",
                    "lifecycle": {},
                    "logFormat": "glog",
                    "logLevel": "info",
                    "mgmt": {
                      "licenseTokenSecretName": "license-token"
                    },
                    "minReadySeconds": 0,
                    "name": "controller",
                    "nginxDebug": false,
                    "nginxReloadTimeout": 60000,
                    "nginxStatus": {
                      "allowCidrs": "127.0.0.1",
                      "enable": true,
                      "port": 8080
                    },
                    "nginxplus": false,
                    "pod": {
                      "annotations": {},
                      "extraLabels": {}
                    },
                    "podDisruptionBudget": {
                      "annotations": {},
                      "enabled": false
                    },
                    "readOnlyRootFilesystem": false,
                    "readyStatus": {
                      "enable": true,
                      "initialDelaySeconds": 0,
                      "port": 8081
                    },
                    "replicaCount": 1,
                    "reportIngressStatus": {
                      "annotations": {},
                      "enable": true,
                      "enableLeaderElection": true,
                      "ingressLink": "",
                      "leaderElectionLockName": "nginx-ingress-leader"
                    },
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "selectorLabels": {},
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "create": true,
                      "customPorts": [],
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraLabels": {},
                      "httpPort": {
                        "enable": true,
                        "port": 80,
                        "targetPort": 80
                      },
                      "httpsPort": {
                        "enable": true,
                        "port": 443,
                        "targetPort": 443
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "type": "LoadBalancer"
                    },
                    "serviceAccount": {
                      "annotations": {},
                      "imagePullSecretName": "",
                      "imagePullSecretsNames": []
                    },
                    "shareProcessNamespace": false,
                    "strategy": {},
                    "terminationGracePeriodSeconds": 30,
                    "tlsPassthroughPort": 443,
                    "tolerations": [],
                    "volumeMounts": [],
                    "volumes": [],
                    "watchNamespace": "",
                    "watchNamespaceLabel": "",
                    "watchSecretNamespace": "",
                    "wildcardTLS": {
                      "cert": "",
                      "key": "",
                      "secret": ""
                    }
                  },
                  "nginxServiceMesh": {
                    "enable": false,
                    "enableEgress": false
                  },
                  "prometheus": {
                    "create": true,
                    "port": 9113,
                    "scheme": "http",
                    "secret": "",
                    "service": {
                      "create": false,
                      "labels": {
                        "service": "nginx-ingress-prometheus-service"
                      }
                    },
                    "serviceMonitor": {
                      "create": false,
                      "endpoints": [
                        {
                          "port": "prometheus"
                        }
                      ],
                      "labels": {},
                      "selectorMatchLabels": {
                        "service": "nginx-ingress-prometheus-service"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "serviceInsight": {
                    "create": false,
                    "port": 9114,
                    "scheme": "http",
                    "secret": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking
          certified: "true"
          containerImage: quay.io/nginx/nginx-ingress-operator@sha256:fb372438634b4a667218e725091bd448a253e7a9c40e7dda1044a318780d995f
          createdAt: "2025-09-15T16:08:18Z"
          description: The NGINX Ingress Operator is a Kubernetes/OpenShift component
            which deploys and manages one or more NGINX/NGINX Plus Ingress Controllers
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nginx-ingress
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/nginx/nginx-ingress-helm-operator
          support: NGINX Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The `NginxIngress` Custom Resource is the definition of a
              deployment of the Ingress Controller.
            displayName: Nginx Ingress Controller
            kind: NginxIngress
            name: nginxingresses.charts.nginx.org
            version: v1alpha1
        description: The NGINX Ingress Operator is a Kubernetes/OpenShift component
          which deploys and manages one or more NGINX/NGINX Plus Ingress Controllers
        displayName: Nginx Ingress Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nginx
        - ingress-controller
        - ingress
        - controller
        - kubernetes
        - openshift
        links:
        - name: Nginx Ingress Operator
          url: https://github.com/nginx/nginx-ingress-helm-operator
        maintainers:
        - email: kubernetes@nginx.com
          name: NGINX Inc
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: NGINX Inc
        relatedImages:
        - quay.io/nginx/nginx-ingress-operator@sha256:fb372438634b4a667218e725091bd448a253e7a9c40e7dda1044a318780d995f
        - quay.io/brancz/kube-rbac-proxy@sha256:754ab2a723c8fc7ea8e09ae1224f8af5be91849b8b2f2a3735c6237f1c534187
        version: 3.3.0
      entries:
      - name: nginx-ingress-operator.v3.3.0
        version: 3.3.0
      name: alpha
    defaultChannel: alpha
    packageName: nginx-ingress-operator
    provider:
      name: NGINX Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: QuestLabs
      provider-url: https://www.questlabs.in
    name: alloy
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: alloy.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.alloys.com/v1alpha1",
                "kind": "Alloy",
                "metadata": {
                  "name": "alloy"
                },
                "spec": {
                  "acmeEmail": "",
                  "dnsFQDN": "",
                  "sslTermination": "true"
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Alloy
            name: alloys.charts.alloys.com
            version: v1alpha1
        description: |-
          Alloy is platform to build trusted data in OSDU, available on-premise or in the cloud. It is equipped with built-in data quality rules derived from PPDM and OSDU standards that control the ingestion of data.

              • Applicable for conventional and new energy streams
              • Highly Scalable

              Benefits
              • Helps to build robust applications/solutions on stable data foundation using full data lifecycle APIs & built-in data/business rules
              • Help data governance team to monitor/oversea data across organisation and intervene when required
              • Substantially low implementation time for any new data entity or new domain (solar, wind etc.) because of modular rules packages
              • Capability to deploy workflows for complex corporate specific data quality and enrichment requirements
              • Spend less time & resources to deploy fit-for-purpose data
              • Schema based checks out of the box for conventional and new energy streams.
              • Sustainable and future proof


              Prerequisites
              1. Please create alloy project before installing alloy operator

          Getting Help

          Once Alloy is successfully installed you can also browse documentation at https://your.domain.com/alloy/docs/introduction
          Please visit https://www.questlabs.in/alloy for more details and if you have any queries or suggestions, mail us at info@questlabs.in
        displayName: Alloy
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - alloy
        links:
        - name: alloy
          url: https://www.questlabs.in
        maintainers:
        - email: info@questlabs.in
          name: QuestLabs
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: QuestLabs
          url: https://www.questlabs.in
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d291e0d38bd95eb2731a10bd5ff06bfa07c6621528d3c2bbe888397d87b3feaa
        - alloy.azurecr.io/alloy@sha256:bf29f69889d408507cf1a9a940beb80d95ba9faff1da2d70a4aea82cc235d96d
        version: 1.0.4
      entries:
      - name: alloy.v1.0.4
        version: 1.0.4
      - name: alloy.v1.0.3
        version: 1.0.3
      - name: alloy.v1.0.2
        version: 1.0.2
      - name: alloy.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: alloy
    provider:
      name: QuestLabs
      url: https://www.questlabs.in
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: SynaXG Engineering Team
      provider-url: ""
    name: synaxg-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: synaxg-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "card.synaxg.app.com/v1",
                "kind": "SG3Reset",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "synaxg-operator",
                    "app.kubernetes.io/instance": "sg3reset-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sg3reset",
                    "app.kubernetes.io/part-of": "synaxg-operator"
                  },
                  "name": "sg3reset-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/synaxg/synaxg-reset-utility@sha256:4050d5731a21bbec6e26165b8dc2b7632577ffa376e7eb056aa3245cf9574753",
                  "sg3ResetList": [
                    {
                      "nodeName": "test",
                      "pciList": [
                        {
                          "enable": "false",
                          "pciAddr": "000d:01:00.0"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "card.synaxg.app.com/v1",
                "kind": "SG3Update",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "synaxg-operator",
                    "app.kubernetes.io/instance": "sg3update-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sg3update",
                    "app.kubernetes.io/part-of": "synaxg-operator"
                  },
                  "name": "sg3update-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/synaxg/synaxg-upgrade-utility@sha256:7ffec3249edd2467f2414c2ce4232e8e436278fd1b43a51b66f54b6ea10a60d1",
                  "sg3UpdateList": [
                    {
                      "nodeName": "test",
                      "pciList": [
                        {
                          "enable": "false",
                          "pciAddr": "000d:01:00.0",
                          "sdkImageAddr": "quay.io/synaxgcom/sdk-img@sha256:1ed57013f278e5a858f71f343649652dbd52b6f3cf451374e45b6ac427fa76f9"
                        }
                      ]
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/synaxg/synaxg-operator-utility@sha256:3b22a8022a97cb1ffdb47ffcc9d9d562601f48669aa14fd58cd8ba6128b39bd7
          createdAt: "2025-07-01T03:42:07Z"
          description: SynaXG Operator will work with the SuperMicro Server + SG3
            card  and Openshift Container Platform.The Marvell kernel module “octeon_ep”
            and “octeon_ep_vf” are necessary to the Operator. The Operator can send
            command of “reset” and “update” to SG3 board via the VF channel.As a result,
            users can realize the process of “reset card” and “update firmware” on
            the web GUI, and monitor the executing result.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: free text
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: SynaXG
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SG3Reset is the Schema for the sg3resets API
            displayName: SG3 Reset
            kind: SG3Reset
            name: sg3resets.card.synaxg.app.com
            version: v1
          - description: SG3Update is the Schema for the sg3updates API
            displayName: SG3 Update
            kind: SG3Update
            name: sg3updates.card.synaxg.app.com
            version: v1
        description: SynaXG Operator will work with the SuperMicro Server + SG3 card  and
          Openshift Container Platform.The Marvell kernel module “octeon_ep” and “octeon_ep_vf”
          are necessary to the Operator. The Operator can send command of “reset”
          and “update” to SG3 board via the VF channel.As a result, users can realize
          the process of “reset card” and “update firmware” on the web GUI, and monitor
          the executing result.
        displayName: SynaXG vDU card 102X Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vDU
        - 102X
        - SynaXG
        links:
        - name: Synaxg Operator
          url: https://synaxg-operator.domain
        maintainers:
        - email: notice@synaxg.com
          name: Mantosh Malhotra
        maturity: alpha
        provider:
          name: SynaXG Engineering Team
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:38ee3484ba60bb474e7231e9d0718157386716cea7ffd6da6f32ddd551a5243d
        - registry.connect.redhat.com/synaxg/synaxg-operator-utility@sha256:3b22a8022a97cb1ffdb47ffcc9d9d562601f48669aa14fd58cd8ba6128b39bd7
        - registry.connect.redhat.com/synaxg/synaxg-reset-utility@sha256:4050d5731a21bbec6e26165b8dc2b7632577ffa376e7eb056aa3245cf9574753
        - registry.connect.redhat.com/synaxg/synaxg-upgrade-utility@sha256:7ffec3249edd2467f2414c2ce4232e8e436278fd1b43a51b66f54b6ea10a60d1
        - quay.io/synaxgcom/sdk-img@sha256:1ed57013f278e5a858f71f343649652dbd52b6f3cf451374e45b6ac427fa76f9
        version: 1.2.0
      entries:
      - name: synaxg-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: synaxg-operator
    provider:
      name: SynaXG Engineering Team
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-opa-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opa-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:08f995a39db5c2c6a17008cffa5216bb9bebd784d13d2b94154474d727fc796f
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/opa-operator@sha256:08f995a39db5c2c6a17008cffa5216bb9bebd784d13d2b94154474d727fc796f
        version: 24.11.1
      entries:
      - name: opa-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: opa-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:826aa12df4531d93a337ba70bd0c0d1e2e8062efe6954fc22797f6474c7732ca
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/opa-operator@sha256:826aa12df4531d93a337ba70bd0c0d1e2e8062efe6954fc22797f6474c7732ca
        version: 25.3.0
      entries:
      - name: opa-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: opa-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaClusterSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
        version: 25.7.0
      entries:
      - name: opa-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: opa-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaClusterSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
        version: 25.7.0
      entries:
      - name: opa-operator.v25.7.0
        version: 25.7.0
      - name: opa-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-opa-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: datagrid
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datagrid-operator.v8.2.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "adminAuth": {
                    "secretName": "basic-auth"
                  },
                  "clusterName": "example-infinispan",
                  "name": "mycache"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC batch-cache"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:1435aa60e8ab4d25c86daae68dcb93d14999f50d7c1e11a3c5726a86d3d30453
          createdAt: "2021-08-24T17:39:44Z"
          description: Create and manage Red Hat Data Grid clusters.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Restore API schema.
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Interoperability:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Data Grid Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
          Red Hat Data Grid is a distributed, in-memory data store that increases application performance and delivers open-source capabilities to handle demanding use cases.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.2/html/data_grid_operator_8.2_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.2
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:1435aa60e8ab4d25c86daae68dcb93d14999f50d7c1e11a3c5726a86d3d30453
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4d8057709958ef60bcabdcec52f05eae8df121fd3f4762c3218abca38130f638
        - registry.redhat.io/datagrid/datagrid-8-openj9-11-rhel8@sha256:c8607d6538a7528e27e6ecb138e0a93a03b3c9fa1f8a9a5aad3fd9f4a6814448
        version: 8.2.8
      entries:
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.2.x
    - currentCSV: datagrid-operator.v8.3.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:4d9d7221b029866837add5f3cb52dd0b7956f57996ce2d8b76344ce2de678716
          createdAt: "2022-11-03T08:40:57Z"
          description: Create and manage Red Hat Data Grid clusters.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.3/html/data_grid_operator_8.3_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.3
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:4d9d7221b029866837add5f3cb52dd0b7956f57996ce2d8b76344ce2de678716
        version: 8.3.9
      entries:
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.3.x
    - currentCSV: datagrid-operator.v8.4.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:e5315682869de72514e3e027bfb02bf0bb64919453ca738c7ecfa1e34634ac2b
          createdAt: 2024-05-30T08:05:19 EDT
          description: Create and manage Red Hat Data Grid clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.4/html/data_grid_operator_8.4_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.4
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:985193caf9bc3698c4d85aeedd1cfea09bccda881636d67885387b2a83c65ebd
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9d9e466b7c367d0406406550c74bc790aa68f3f875ea4f652eb28f2e25a05458
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:b56269cb849afca34e7f82cc8e01a44bb32af6126f24db28019a58e556d4cb3c
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:d522a20ca6909bd464f85967d8b86fe42b83f773478a80785f9a85a0cb87dfed
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bf7760520f5bea151a531d538d525c4591c47166995aefb4d77fef138de11544
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bcb040bf640691568a93a4c71c8072c37ffb76b4296ade83186f279862b7f197
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:24cf79efa89bbfa495665b1711a27531b7ab0813695a89b2c38c96c2e792eb24
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:532dca6facb584c1261c0480c499af9f857c70dc390e0d46e46afec2561468af
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:78a6bed6fcb6962e5689b92806789b72a66db21b7c607b392a573c98593f2c09
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:ae852e4679978725bb7d24c1bdfe25b40bd883130ed04a36d4fca49dbea024ef
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:424b30dccd773f69c079b35a42e8c3e339246c5d490e14e3fcfb8bd8144a9160
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9223b4b5c1c866a91033dac40f30a4caae0f6946cdb64935a3fdfffdb9ede660
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:c15d4fd827f554291c14851f187947f4109c0e20257fa2f1e141b32669569358
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:54bc26a0fddc160a4e0d1a813aec78abb3c9cb77db5039b74c23de759faa67f5
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:554b831a0803d55a4e3b7022011b5a5f6d2b4b65009065a9b28bc8e55fc3ad46
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:e5315682869de72514e3e027bfb02bf0bb64919453ca738c7ecfa1e34634ac2b
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4dbe1b7a1a36990c0a31d5a66f3cc4ca2954c9e1ec3b87431113a4111530277d
        version: 8.4.16
      entries:
      - name: datagrid-operator.v8.4.16
        version: 8.4.16
      - name: datagrid-operator.v8.4.15
        version: 8.4.15
      - name: datagrid-operator.v8.4.14
        version: 8.4.14
      - name: datagrid-operator.v8.4.13
        version: 8.4.13
      - name: datagrid-operator.v8.4.12
        version: 8.4.12
      - name: datagrid-operator.v8.4.11
        version: 8.4.11
      - name: datagrid-operator.v8.4.10
        version: 8.4.9
      - name: datagrid-operator.v8.4.8
        version: 8.4.8
      - name: datagrid-operator.v8.4.7
        version: 8.4.7
      - name: datagrid-operator.v8.4.6
        version: 8.4.6
      - name: datagrid-operator.v8.4.5
        version: 8.4.5
      - name: datagrid-operator.v8.4.4
        version: 8.4.4
      - name: datagrid-operator.v8.4.3
        version: 8.4.3
      - name: datagrid-operator.v8.4.2
        version: 8.4.2
      - name: datagrid-operator.v8.4.1
        version: 8.4.1
      - name: datagrid-operator.v8.4.0
        version: 8.4.0
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.4.x
    - currentCSV: datagrid-operator.v8.5.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel9-operator@sha256:d719d778b0848588a7e61d3c3fa689763f51a6533af4cf9e1c738c4ed8fac760
          createdAt: 2025-08-29T05:44:37 UTC
          description: Create and manage Red Hat Data Grid clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.5/html/data_grid_operator_8.5_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.5
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:a711c50e489995fb60e5dec23ce286360743feda5066e4a0861d448aa4f2e903
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bf7760520f5bea151a531d538d525c4591c47166995aefb4d77fef138de11544
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bcb040bf640691568a93a4c71c8072c37ffb76b4296ade83186f279862b7f197
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:4b9bf39d2ff14bcfbd4628a399093703d87a2b2b9c8e74c3d7cfdff4b5167fa4
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:ae852e4679978725bb7d24c1bdfe25b40bd883130ed04a36d4fca49dbea024ef
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:d522a20ca6909bd464f85967d8b86fe42b83f773478a80785f9a85a0cb87dfed
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9223b4b5c1c866a91033dac40f30a4caae0f6946cdb64935a3fdfffdb9ede660
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:11bfe5c8109d26164cff428232b2001ab8df464f14a382661b936543336e2960
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:ad4f91f984b4b898012963d52e6ebcd166ec03e76e26eba3191ecb87f51fb18d
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:c15d4fd827f554291c14851f187947f4109c0e20257fa2f1e141b32669569358
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:5656cd188d054551489d45be996266d03e3faa8814b8bfc0f360cf0dceb83cbe
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:424b30dccd773f69c079b35a42e8c3e339246c5d490e14e3fcfb8bd8144a9160
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:879f3dcc24ffe795472779f777d955666ba4041c17e717d5e165cd7f12d7955a
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:985193caf9bc3698c4d85aeedd1cfea09bccda881636d67885387b2a83c65ebd
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:b56269cb849afca34e7f82cc8e01a44bb32af6126f24db28019a58e556d4cb3c
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:9bc0b730570ab55fbeca83075949f4caa96ed5c37637420a59cd4a7db01ea840
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:532dca6facb584c1261c0480c499af9f857c70dc390e0d46e46afec2561468af
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9d9e466b7c367d0406406550c74bc790aa68f3f875ea4f652eb28f2e25a05458
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:54bc26a0fddc160a4e0d1a813aec78abb3c9cb77db5039b74c23de759faa67f5
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:78a6bed6fcb6962e5689b92806789b72a66db21b7c607b392a573c98593f2c09
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4dbe1b7a1a36990c0a31d5a66f3cc4ca2954c9e1ec3b87431113a4111530277d
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:48cf13e71953cea6f62354289f0d8ece7a8a481ee0bb2c76eb9352d438db9ca3
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:0d368459c98c5788dd55d2add5b4b7d438b9a8c814ea8aee2222ea956699cbb6
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:c25e3152d6a6b6d5cd86d311bf567e594b9948440785209cd8d7986076a5a10d
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:f59ae21773dfc207454a709ac5f7aad2be6a6a20ac88da26c379edeed54537b8
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:24cf79efa89bbfa495665b1711a27531b7ab0813695a89b2c38c96c2e792eb24
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:6e1766784b684561666134d41dca5331b80d50e63695fe7b1a1d6d4cdb73f0bd
        - registry.redhat.io/datagrid/datagrid-8-rhel9-operator@sha256:d719d778b0848588a7e61d3c3fa689763f51a6533af4cf9e1c738c4ed8fac760
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:554b831a0803d55a4e3b7022011b5a5f6d2b4b65009065a9b28bc8e55fc3ad46
        version: 8.5.11
      entries:
      - name: datagrid-operator.v8.5.11
        version: 8.5.11
      - name: datagrid-operator.v8.5.10
        version: 8.5.10
      - name: datagrid-operator.v8.5.9
        version: 8.5.9
      - name: datagrid-operator.v8.5.8
        version: 8.5.8
      - name: datagrid-operator.v8.5.7
        version: 8.5.7
      - name: datagrid-operator.v8.5.6
        version: 8.5.6
      - name: datagrid-operator.v8.5.5
        version: 8.5.5
      - name: datagrid-operator.v8.5.4
        version: 8.5.4
      - name: datagrid-operator.v8.5.3
        version: 8.5.3
      - name: datagrid-operator.v8.5.2
        version: 8.5.2
      - name: datagrid-operator.v8.5.1
        version: 8.5.1
      - name: datagrid-operator.v8.5.0
        version: 8.5.0
      - name: datagrid-operator.v8.4.16
        version: 8.4.16
      - name: datagrid-operator.v8.4.15
        version: 8.4.15
      - name: datagrid-operator.v8.4.14
        version: 8.4.14
      - name: datagrid-operator.v8.4.13
        version: 8.4.13
      - name: datagrid-operator.v8.4.12
        version: 8.4.12
      - name: datagrid-operator.v8.4.11
        version: 8.4.11
      - name: datagrid-operator.v8.4.10
        version: 8.4.9
      - name: datagrid-operator.v8.4.8
        version: 8.4.8
      - name: datagrid-operator.v8.4.7
        version: 8.4.7
      - name: datagrid-operator.v8.4.6
        version: 8.4.6
      - name: datagrid-operator.v8.4.5
        version: 8.4.5
      - name: datagrid-operator.v8.4.4
        version: 8.4.4
      - name: datagrid-operator.v8.4.3
        version: 8.4.3
      - name: datagrid-operator.v8.4.2
        version: 8.4.2
      - name: datagrid-operator.v8.4.1
        version: 8.4.1
      - name: datagrid-operator.v8.4.0
        version: 8.4.0
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: stable
    defaultChannel: stable
    packageName: datagrid
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeRocketCI
      provider-url: https://docs.kuberocketci.io/
    name: sonar-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sonar-operator.v3.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Sonar",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonar-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonar",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonar-sample"
                },
                "spec": {
                  "defaultPermissionTemplate": "edp-default",
                  "secret": "sonar-secret",
                  "settings": [
                    {
                      "key": "sonar.global.exclusions",
                      "values": [
                        "**/*.jsx",
                        "**/*.tsx"
                      ]
                    },
                    {
                      "fieldValues": {
                        "beginBlockRegexp": ".*",
                        "endBlockRegexp": ".*"
                      },
                      "key": "sonar.issue.ignore.block"
                    },
                    {
                      "key": "sonar.dbcleaner.hoursBeforeKeepingOnlyOneSnapshotByDay",
                      "value": "20"
                    }
                  ],
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonargroup-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonargroup",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonargroup-sample"
                },
                "spec": {
                  "description": "Developers",
                  "name": "sonar-developers",
                  "permissions": [
                    "admin"
                  ],
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarPermissionTemplate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarpermissiontemplate-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarpermissiontemplate",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarpermissiontemplate-sample"
                },
                "spec": {
                  "default": true,
                  "description": "EDP permission templates",
                  "groupsPermissions": {
                    "sonar-developers": [
                      "scan"
                    ],
                    "sonar-users": [
                      "issueadmin"
                    ]
                  },
                  "name": "edp-default",
                  "projectKeyPattern": ".+",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityGate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarqualitygate-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualitygate",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarqualitygate-sample"
                },
                "spec": {
                  "conditions": {
                    "new_duplicated_lines_density": {
                      "error": "1",
                      "op": "GT"
                    }
                  },
                  "default": true,
                  "name": "sonarqualitygate-sample",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityProfile",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarqualityprofile-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualityprofile",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarqualityprofile-sample"
                },
                "spec": {
                  "default": true,
                  "language": "go",
                  "name": "sonarqualityprofile-sample",
                  "rules": {
                    "go:S1151": {
                      "params": "max=\"6\"",
                      "severity": "MAJOR"
                    }
                  },
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonaruser-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonaruser",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonaruser-sample"
                },
                "spec": {
                  "email": "sonaruser@mail.com",
                  "groups": [
                    "sonar-administrators"
                  ],
                  "login": "sonaruser",
                  "name": "sonar user",
                  "permissions": [
                    "scan"
                  ],
                  "secret": "sonaruser-secret",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/sonar-operator:3.1.0
          createdAt: "2023-09-08T12:53:57Z"
          description: An Operator for managing Sonar entities.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-sonar-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonarGroup is the Schema for the sonar group API.
            displayName: Sonar Group
            kind: SonarGroup
            name: sonargroups.edp.epam.com
            version: v1alpha1
          - description: SonarPermissionTemplate is the Schema for the sonar permission
              template API.
            displayName: Sonar Permission Template
            kind: SonarPermissionTemplate
            name: sonarpermissiontemplates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityGate is the Schema for the sonarqualitygates
              API
            displayName: Sonar Quality Gate
            kind: SonarQualityGate
            name: sonarqualitygates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityProfile is the Schema for the sonarqualityprofiles
              API
            displayName: Sonar Quality Profile
            kind: SonarQualityProfile
            name: sonarqualityprofiles.edp.epam.com
            version: v1alpha1
          - description: Sonar is the Schema for the sonars API.
            displayName: Sonar
            kind: Sonar
            name: sonars.edp.epam.com
            version: v1alpha1
          - description: SonarUser is the Schema for the sonarusers API
            displayName: Sonar User
            kind: SonarUser
            name: sonarusers.edp.epam.com
            version: v1alpha1
        description: |
          Sonar Operator is an operator that is responsible for establishing
          a connection to provided SonarQube server, reconciling Sonar entities (users,
          groups, permission templates, etc) according to the created CRs.
        displayName: EDP Sonar Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - analysis
        - automation
        - code
        - continuous
        - devops
        - edp
        - inspection
        - metrics
        - metrics
        - monitoring
        - operator
        - quality
        - security
        - software
        - static
        - technical
        - vulnerabilities
        links:
        - name: Sonar Operator
          url: https://github.com/epam/edp-sonar-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/sonar-operator:3.1.0
        version: 3.1.0
      entries:
      - name: sonar-operator.v3.1.0
        version: 3.1.0
      name: alpha
    - currentCSV: sonar-operator.v3.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Sonar",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonar"
                  },
                  "name": "sonar-sample"
                },
                "spec": {
                  "defaultPermissionTemplate": "edp-default",
                  "secret": "sonar-secret",
                  "settings": [
                    {
                      "key": "sonar.global.exclusions",
                      "values": [
                        "**/*.jsx",
                        "**/*.tsx"
                      ]
                    },
                    {
                      "fieldValues": {
                        "beginBlockRegexp": ".*",
                        "endBlockRegexp": ".*"
                      },
                      "key": "sonar.issue.ignore.block"
                    },
                    {
                      "key": "sonar.dbcleaner.hoursBeforeKeepingOnlyOneSnapshotByDay",
                      "value": "20"
                    }
                  ],
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonargroup"
                  },
                  "name": "sonargroup-sample"
                },
                "spec": {
                  "description": "Developers",
                  "name": "sonar-developers",
                  "permissions": [
                    "admin"
                  ],
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarPermissionTemplate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarpermissiontemplate"
                  },
                  "name": "sonarpermissiontemplate-sample"
                },
                "spec": {
                  "default": true,
                  "description": "EDP permission templates",
                  "groupsPermissions": {
                    "sonar-developers": [
                      "scan"
                    ],
                    "sonar-users": [
                      "issueadmin"
                    ]
                  },
                  "name": "edp-default",
                  "projectKeyPattern": ".+",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityGate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualitygate"
                  },
                  "name": "sonarqualitygate-sample"
                },
                "spec": {
                  "conditions": {
                    "new_duplicated_lines_density": {
                      "error": "1",
                      "op": "GT"
                    }
                  },
                  "default": true,
                  "name": "sonarqualitygate-sample",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityProfile",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualityprofile"
                  },
                  "name": "sonarqualityprofile-sample"
                },
                "spec": {
                  "default": true,
                  "language": "go",
                  "name": "sonarqualityprofile-sample",
                  "rules": {
                    "go:S1151": {
                      "params": "max=\"6\"",
                      "severity": "MAJOR"
                    }
                  },
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonaruser"
                  },
                  "name": "sonaruser-sample"
                },
                "spec": {
                  "email": "sonaruser@mail.com",
                  "groups": [
                    "sonar-administrators"
                  ],
                  "login": "sonaruser",
                  "name": "sonar user",
                  "permissions": [
                    "scan"
                  ],
                  "secret": "sonaruser-secret",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/sonar-operator:3.3.0
          createdAt: "2025-04-22T11:39:32Z"
          description: An Operator for managing Sonar entities.
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/epam/edp-sonar-operator
          support: KubeRocketCI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonarGroup is the Schema for the sonar group API.
            displayName: Sonar Group
            kind: SonarGroup
            name: sonargroups.edp.epam.com
            version: v1alpha1
          - description: SonarPermissionTemplate is the Schema for the sonar permission
              template API.
            displayName: Sonar Permission Template
            kind: SonarPermissionTemplate
            name: sonarpermissiontemplates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityGate is the Schema for the sonarqualitygates
              API
            displayName: Sonar Quality Gate
            kind: SonarQualityGate
            name: sonarqualitygates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityProfile is the Schema for the sonarqualityprofiles
              API
            displayName: Sonar Quality Profile
            kind: SonarQualityProfile
            name: sonarqualityprofiles.edp.epam.com
            version: v1alpha1
          - description: Sonar is the Schema for the sonars API.
            displayName: Sonar
            kind: Sonar
            name: sonars.edp.epam.com
            version: v1alpha1
          - description: SonarUser is the Schema for the sonarusers API.
            displayName: Sonar User
            kind: SonarUser
            name: sonarusers.edp.epam.com
            version: v1alpha1
        description: |
          Sonar Operator is an operator that is responsible for establishing
          a connection to provided SonarQube server, reconciling Sonar entities (users,
          groups, permission templates, etc) according to the created CRs.
        displayName: EDP Sonar Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - analysis
        - automation
        - code
        - continuous
        - devops
        - edp
        - inspection
        - metrics
        - metrics
        - monitoring
        - operator
        - quality
        - security
        - software
        - static
        - technical
        - vulnerabilities
        links:
        - name: Sonar Operator
          url: https://github.com/epam/edp-sonar-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: KubeRocketCI
          url: https://docs.kuberocketci.io/
        relatedImages:
        - docker.io/epamedp/sonar-operator:3.3.0
        version: 3.3.0
      entries:
      - name: sonar-operator.v3.3.0
        version: 3.3.0
      - name: sonar-operator.v3.2.0
        version: 3.2.0
      name: stable
    defaultChannel: stable
    packageName: sonar-operator
    provider:
      name: KubeRocketCI
      url: https://docs.kuberocketci.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.6.5-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6.5
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/kubeturbo-operator:8.6.5
        version: 8.6.5-beta.1
      entries:
      - name: kubeturbo-operator.v8.6.5-beta.1
        version: 8.6.5-beta.1
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: beta
    - currentCSV: kubeturbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/kubeturbo-operator:8.6
        version: 8.6.0
      entries:
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: stable
    - currentCSV: kubeturbo-operator.v8.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.4
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/kubeturbo-operator:8.4
        version: 8.4.0
      entries:
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.4.0
    - currentCSV: kubeturbo-operator.v8.5.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.5
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/kubeturbo-operator:8.5
        version: 8.5.0
      entries:
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.5.0
    - currentCSV: kubeturbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/kubeturbo-operator:8.6
        version: 8.6.0
      entries:
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.6.0
    defaultChannel: stable
    packageName: kubeturbo
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Silicom Ltd.
      provider-url: http://www.silicom-usa.com
    name: silicom-sts-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: silicom-sts-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 3,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-BC-8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/mode": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "hp-ocp",
                  "namespase": "openshift-operators"
                },
                "spec": {
                  "dummy": 0
                },
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicom-sts/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicom-sts/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/rmr/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - quay.io/silicom/gpsd@sha256:e479c542e061b5422865f2617ed0563794f265c6e6660e5a9d204c9e1cb2f346
        - quay.io/silicom/tsyncd@sha256:3f1cbd4a8b5d62d7d3e525af5f26d9f27b8c08e886b0639a1e368163e166b537
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - quay.io/silicom/sts-operator@sha256:7d474d0aa91ac57d8f5817b033b67b511c8601e0dbd83d808a69004bd24b4e60
        - quay.io/silicom/grpc-tsyncd@sha256:631bb754742e1a24376913baed60e480a6e6ae988ad9a3892254ff96996c15d1
        - quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
        - quay.io/silicom/phc2sys@sha256:b067167a1a5d93e9fe715c7774f6cdbbcd93cac27780f022e44dcb8a4d4d7bb9
        - quay.io/silicom/sts-plugin@sha256:233bc95f8617a06bd452548ec116d91551dd4933970860d8d5ccb81f75ad2220
        - quay.io/silicom/ice-driver-src@sha256:569c8333e21c7c022ea71fa5b418404899040d5b3a3a4128eeac21f62aacaf3e
        version: 0.1.3
      entries:
      - name: silicom-sts-operator.v0.1.3
        version: 0.1.3
      - name: silicom-sts-operator.v0.0.9
        version: 0.0.9
      - name: silicom-sts-operator.v0.0.8
        version: 0.0.8
      name: alpha
    - currentCSV: silicom-sts-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 3,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-BC-8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/mode": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "hp-ocp",
                  "namespase": "openshift-operators"
                },
                "spec": {
                  "dummy": 0
                },
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicom-sts/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicom-sts/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/rmr/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - quay.io/silicom/sts-plugin@sha256:233bc95f8617a06bd452548ec116d91551dd4933970860d8d5ccb81f75ad2220
        - quay.io/silicom/sts-operator@sha256:7d474d0aa91ac57d8f5817b033b67b511c8601e0dbd83d808a69004bd24b4e60
        - quay.io/silicom/tsyncd@sha256:3f1cbd4a8b5d62d7d3e525af5f26d9f27b8c08e886b0639a1e368163e166b537
        - quay.io/silicom/grpc-tsyncd@sha256:631bb754742e1a24376913baed60e480a6e6ae988ad9a3892254ff96996c15d1
        - quay.io/silicom/ice-driver-src@sha256:569c8333e21c7c022ea71fa5b418404899040d5b3a3a4128eeac21f62aacaf3e
        - quay.io/silicom/gpsd@sha256:e479c542e061b5422865f2617ed0563794f265c6e6660e5a9d204c9e1cb2f346
        - quay.io/silicom/phc2sys@sha256:b067167a1a5d93e9fe715c7774f6cdbbcd93cac27780f022e44dcb8a4d4d7bb9
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
        version: 0.1.3
      entries:
      - name: silicom-sts-operator.v0.1.3
        version: 0.1.3
      - name: silicom-sts-operator.v0.0.9
        version: 0.0.9
      - name: silicom-sts-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: silicom-sts-operator
    provider:
      name: Silicom Ltd.
      url: http://www.silicom-usa.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-flashsystemcluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "flashsystemcluster.odf.ibm.com/v1alpha1",
                  "name": "my-flashsystemcluster",
                  "namespace": "openshift-storage"
                }
              },
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-storagecluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "storagecluster.ocs.openshift.io/v1",
                  "name": "my-storagecluster",
                  "namespace": "openshift-storage"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-console"]'
          containerImage: registry.redhat.io/odf4/odf-rhel9-operator@sha256:8d377d2d8a94e461a6eca8d1c3b4f5cf1f9ec1ba6a03044304403b58bbbabaa4
          createdAt: "2025-09-12T07:34:03Z"
          description: OpenShift Data Foundation provides a common control plane for
            storage solutions on OpenShift Container Platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.16"}]'
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operator.openshift.io/uninstall-message: 'Warning: deleting the operand
            (StorageSystem) will result in data loss'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "odf.openshift.io/v1alpha1",
              "kind": "StorageSystem",
              "metadata": {
                "name": "ocs-storagecluster-storagesystem",
                "namespace": "openshift-storage"
              },
              "spec": {
                "kind": "storagecluster.ocs.openshift.io/v1",
                "name": "ocs-storagecluster",
                "namespace": "openshift-storage"
              }
            }
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/odf-operator
          support: Red Hat
          vendors.odf.openshift.io/kind: '["storagecluster.ocs.openshift.io/v1", "flashsystemcluster.odf.ibm.com/v1alpha1"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageSystem is the Schema for the storagesystems API
            displayName: Storage System
            kind: StorageSystem
            name: storagesystems.odf.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Data Foundation** deploys three operators.

          ### OpenShift Data Foundation
          The OpenShift Data Foundation operator is the primary operator for
          OpenShift Data Foundation. It serves to facilitate the other operators in
          OpenShift Data Foundation by performing administrative tasks outside their
          scope as well as watching and configuring their CustomResources.

          ### Rook
          [Rook][1] deploys and manages Ceph on OpenShift, which provides block,
          file and object storage.

          ### NooBaa operator
          The NooBaa operator deploys and manages the [NooBaa][2] Multi-Cloud Gateway
          on OpenShift, which provides object storage.

          # Core Capabilities
          * **Self-managing service:** No matter which supported storage technologies
          you choose, OpenShift Data Foundation ensures that resources can be
          deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Data Foundation you
          can either build dedicated storage clusters or hyper-converged clusters
          where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Data Foundation:**
          OpenShift Data Foundation integrates Ceph with multiple storage
          presentations including object storage (compatible with S3), block storage,
          and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Data Foundation efficiently
          distributes and replicates your data across your cluster to minimize the
          risk of data loss. With snapshots, cloning, and versioning, no more losing
          sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your
          datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data
          storage for your workloads, using a single namespace.

          # Subscriptions
          OpenShift Data Foundation requires a matching 1:1 ratio with the OpenShift
          subscription used. The Essential subscription is included in the OpenShift
          Platform Plus subscription. Disaster Recovery features, external mode, and
          KMS integration for encryption at rest require a bundle of OpenShift
          Platform Plus with OpenShift Data Foundation Advanced subscription.

          [1]: https://rook.io
          [2]: https://www.noobaa.io
        displayName: OpenShift Data Foundation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/odf4/odf-rhel9-operator@sha256:8d377d2d8a94e461a6eca8d1c3b4f5cf1f9ec1ba6a03044304403b58bbbabaa4
        - registry.redhat.io/odf4/odf-console-rhel9@sha256:5d0d27a091a74e65985d2f047d3f6322043f2ae80fa5ff3192ec84cce92f143d
        version: 4.15.20-rhodf
      entries:
      - name: odf-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odf-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odf-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odf-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odf-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odf-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odf-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odf-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odf-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odf-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odf-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odf-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odf-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odf-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odf-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odf-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odf-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odf-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odf-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odf-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odf-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-flashsystemcluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "flashsystemcluster.odf.ibm.com/v1alpha1",
                  "name": "my-flashsystemcluster",
                  "namespace": "openshift-storage"
                }
              },
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-storagecluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "storagecluster.ocs.openshift.io/v1",
                  "name": "my-storagecluster",
                  "namespace": "openshift-storage"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-console"]'
          containerImage: registry.redhat.io/odf4/odf-rhel9-operator@sha256:20cb12259667539a0244cfde37297b975e2682833fbe537caecb5107ebecf15f
          createdAt: "2025-09-12T06:33:04Z"
          description: OpenShift Data Foundation provides a common control plane for
            storage solutions on OpenShift Container Platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.17"}]'
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operator.openshift.io/uninstall-message: 'Warning: deleting the operand
            (StorageSystem) will result in data loss'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "odf.openshift.io/v1alpha1",
              "kind": "StorageSystem",
              "metadata": {
                "name": "ocs-storagecluster-storagesystem",
                "namespace": "openshift-storage"
              },
              "spec": {
                "kind": "storagecluster.ocs.openshift.io/v1",
                "name": "ocs-storagecluster",
                "namespace": "openshift-storage"
              }
            }
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/odf-operator
          support: Red Hat
          vendors.odf.openshift.io/kind: '["storagecluster.ocs.openshift.io/v1", "flashsystemcluster.odf.ibm.com/v1alpha1"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageSystem is the Schema for the storagesystems API
            displayName: Storage System
            kind: StorageSystem
            name: storagesystems.odf.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Data Foundation** deploys three operators.

          ### OpenShift Data Foundation
          The OpenShift Data Foundation operator is the primary operator for
          OpenShift Data Foundation. It serves to facilitate the other operators in
          OpenShift Data Foundation by performing administrative tasks outside their
          scope as well as watching and configuring their CustomResources.

          ### Rook
          [Rook][1] deploys and manages Ceph on OpenShift, which provides block,
          file and object storage.

          ### NooBaa operator
          The NooBaa operator deploys and manages the [NooBaa][2] Multi-Cloud Gateway
          on OpenShift, which provides object storage.

          # Core Capabilities
          * **Self-managing service:** No matter which supported storage technologies
          you choose, OpenShift Data Foundation ensures that resources can be
          deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Data Foundation you
          can either build dedicated storage clusters or hyper-converged clusters
          where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Data Foundation:**
          OpenShift Data Foundation integrates Ceph with multiple storage
          presentations including object storage (compatible with S3), block storage,
          and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Data Foundation efficiently
          distributes and replicates your data across your cluster to minimize the
          risk of data loss. With snapshots, cloning, and versioning, no more losing
          sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your
          datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data
          storage for your workloads, using a single namespace.

          # Subscriptions
          OpenShift Data Foundation requires a matching 1:1 ratio with the OpenShift
          subscription used. The Essential subscription is included in the OpenShift
          Platform Plus subscription. Disaster Recovery features, external mode, and
          KMS integration for encryption at rest require a bundle of OpenShift
          Platform Plus with OpenShift Data Foundation Advanced subscription.

          [1]: https://rook.io
          [2]: https://www.noobaa.io
        displayName: OpenShift Data Foundation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/odf4/odf-rhel9-operator@sha256:20cb12259667539a0244cfde37297b975e2682833fbe537caecb5107ebecf15f
        - registry.redhat.io/odf4/odf-console-rhel9@sha256:8bd822a3410e520fc57ac69c7599831173003e2c34b980c517f79aca97210257
        version: 4.16.16-rhodf
      entries:
      - name: odf-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odf-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odf-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odf-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odf-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odf-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odf-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odf-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odf-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odf-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odf-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odf-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odf-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odf-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odf-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odf-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odf-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odf-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: jaeger-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jaeger-operator.v1.65.0-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "jaegertracing.io/v1",
                "kind": "Jaeger",
                "metadata": {
                  "name": "my-jaeger"
                },
                "spec": {
                  "strategy": "allInOne"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/rhosdt/jaeger-rhel8-operator@sha256:51a8685e4dc70888c991996f149cd42233049d18f702ec72b066d6dad77a07c9
          createdAt: 25 Aug 2025, 13:30
          description: Provides tracing, monitoring and troubleshooting for microservices-based
            distributed systems
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.65.0-4'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-distributed-tracing
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/os-observability/konflux-jaeger
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Jaeger is the Schema for the jaegers API
            displayName: Jaeger
            kind: Jaeger
            name: jaegers.jaegertracing.io
            version: v1
        description: |2-

          ## Operator deprecation warning

          This Operator is deprecated and not recommended for new deployments. This Operator will be removed from the `redhat-operators` catalog in a future release.

          If this is not the first time you are installing it, migrate to the Red Hat build of OpenTelemetry Operator and the Tempo Operator for distributed tracing collection and storage. For more information, see "Migrating" in the Red Hat build of OpenTelemetry documentation, "Installing" in the Red Hat build of OpenTelemetry documentation, and "Installing" in the distributed tracing platform (Tempo) documentation.

          ## About

          Red Hat OpenShift distributed tracing platform based on Jaeger. Jaeger is project inspired by [Dapper](https://research.google.com/pubs/pub36356.html) and [OpenZipkin](http://zipkin.io/). It is a distributed tracing system released as open source by Uber Technologies. It is used for monitoring and troubleshooting microservices-based distributed systems.

          ### Core capabilities
          Jaeger is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed context propagation
          * Distributed transaction monitoring
          * Root cause analysis
          * Service dependency analysis
          * Performance / latency optimization
          * OpenTracing compatible data model
          * Multiple storage backends: Elasticsearch, Memory.

          ### Operator features

          * **Multiple modes** - Provides `allInOne`, `production` and `streaming` [modes of deployment](https://www.jaegertracing.io/docs/latest/operator/#deployment-strategies).

          * **Configuration** - The Operator manages [configuration information](https://www.jaegertracing.io/docs/latest/operator/#configuring-the-custom-resource) when installing Jaeger instances.

          * **Storage** - [Configure storage](https://www.jaegertracing.io/docs/latest/operator/#storage-options) used by Jaeger. By default, `memory` is used. Other options include `elasticsearch`. The operator can delegate creation of an Elasticsearch cluster to the Elasticsearch Operator if deployed.

          * **Agent** - can be deployed as [sidecar](https://www.jaegertracing.io/docs/latest/operator/#auto-injecting-jaeger-agent-sidecars) (default) and/or [daemonset](https://www.jaegertracing.io/docs/latest/operator/#installing-the-agent-as-daemonset).

          * **UI** - Optionally setup secure route to provide [access to the Jaeger UI](https://www.jaegertracing.io/docs/latest/operator/#accessing-the-jaeger-console-ui).

          ### Before you start
          1. Ensure that the appropriate storage solution, that will be used by the Jaeger instance, is available and configured.
          2. If intending to deploy an Elasticsearch cluster via the Jaeger custom resource, then the Elasticsearch Operator must first be installed.

          ### Support & Troubleshooting

          Red Hat OpenShift distributed tracing Jaeger is available and supported as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/4.17/html/distributed_tracing/distributed-tracing-platform-jaeger) is recommended when installing, configuring, and managing the Operator and instances.
          Support is provided to Red Hat OpenShift entitled customers subject to the [Production Scope for Coverage](https://access.redhat.com/support/offerings/production/soc) and the [Red Hat OpenShift distributed tracing Life Cycle](https://access.redhat.com/support/policy/updates/openshift#jaeger).
        displayName: Red Hat OpenShift distributed tracing platform
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tracing
        - monitoring
        - troubleshooting
        links:
        - name: Jaeger Operator Source Code
          url: https://github.com/jaegertracing/jaeger-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:667d6545b92aa8381f65369de3b5a6781d1de5b4bcfa58ccde4721b6b0406f3d
        - registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:d92003d6cf0b4f3dae11f849355c9179e78e768f3b7dbeaf3224a4a019ec5d5d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e77efd450e757772d0753a73062753dabd08e53f21315e9b87f4cc1efa332661
        - registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:1d586f9aafd1cf847a14f509692ce54956b0e754a9a04beaa7a75c480f86717c
        - registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:7b021ed467d35b2bcae87765dfc64f639b03bf71f7e34bbea1dfad16626847b9
        - registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:1f4a9daf152420d9ca7cf89c38ae5550b12dd274bada389c7b85ea4cbcbc4fb7
        - registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:fd8110529daa8cf3e7f9fcf71346adc5f48cb05717ebbad1c25bc61c94f31656
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        - registry.redhat.io/rhosdt/jaeger-rhel8-operator@sha256:51a8685e4dc70888c991996f149cd42233049d18f702ec72b066d6dad77a07c9
        version: 1.65.0-4
      entries:
      - name: jaeger-operator.v1.65.0-4
        version: 1.65.0-4
      - name: jaeger-operator.v1.65.0-3
        version: 1.65.0-3
      - name: jaeger-operator.v1.65.0-2
        version: 1.65.0-2
      - name: jaeger-operator.v1.65.0-1
        version: 1.65.0-1
      - name: jaeger-operator.v1.62.0-2
        version: 1.62.0-2
      - name: jaeger-operator.v1.62.0-1
        version: 1.62.0-1
      - name: jaeger-operator.v1.57.0-10
        version: 1.57.0-10
      - name: jaeger-operator.v1.57.0-10-0.1730817420.p
        version: 1.57.0-10+0.1730817420.p
      - name: jaeger-operator.v1.57.0-7
        version: 1.57.0-7
      - name: jaeger-operator.v1.57.0-6
        version: 1.57.0-6
      - name: jaeger-operator.v1.57.0-5
        version: 1.57.0-5
      - name: jaeger-operator.v1.53.0-4
        version: 1.53.0-4
      - name: jaeger-operator.v1.53.0-3
        version: 1.53.0-3
      - name: jaeger-operator.v1.51.0-1
        version: 1.51.0-1
      - name: jaeger-operator.v1.47.1-5
        version: 1.47.1-5
      - name: jaeger-operator.v1.42.0-5-0.1687199951.p
        version: 1.42.0-5+0.1687199951.p
      - name: jaeger-operator.v1.42.0-5
        version: 1.42.0-5
      - name: jaeger-operator.v1.34.1-5
        version: 1.34.1-5
      - name: jaeger-operator.v1.30.2
        version: 1.30.2
      name: stable
    defaultChannel: stable
    packageName: jaeger-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator.v1.6.3
        version: 1.6.3
      - name: dynatrace-operator.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator.v1.4.0
        version: 1.4.0
      - name: dynatrace-operator.v1.3.2
        version: 1.3.2
      - name: dynatrace-operator.v1.3.1
        version: 1.3.1
      - name: dynatrace-operator.v1.3.0
        version: 1.3.0
      - name: dynatrace-operator.v0.6.0
        version: 0.6.0
      - name: dynatrace-operator.v0.5.1
        version: 0.5.1
      - name: dynatrace-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco
      provider-url: ""
    name: aci-containers-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aci-containers-operator.v1.0.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"aci.ctrl/v1alpha1","kind":"AciContainersOperator","metadata":{"name":"acicnioperator","namespace":"aci-containers-system"},"spec":{}}]'
          capabilities: Seamless Upgrades
          categories: Networking
          certified: "false"
          containerImage: quay.io/noiro/aci-containers-operator@sha256:2b2d685e254c206be5371285852a8f95d0318af69cb85b786427969e035c125c
          createdAt: "2025-04-10T12:00:00Z"
          description: The Cisco Application Centric Infrastructure (ACI) Containers
            Operator brings the ACI CNI Plugin and Policy model to Kubernetes clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://github.com/noironetworks/aci-containers/blob/master/README.md
          repository: https://github.com/noironetworks/aci-containers
          support: acisupport@cisco.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aci Containers Operator
            displayName: AciContainersOperator
            kind: AciContainersOperator
            name: acicontainersoperators.aci.ctrl
            version: v1alpha1
        description: |
          ## ACI CNI Plugin

          The Cisco Application Centric Infrastructure (ACI) CNI plugin brings the ACI Networking and Policy model to Kubernetes clusters that reside on-prem or in the cloud. It is fully open source and relies on the Opflex Protocol to program Open vSwitch instances running on the Kubernetes nodes. It provides IP address management, L2/L3 networking, load balancing, and security functions for container workloads.

          The Cisco ACI CNI plugin for the Red Hat OpenShift Container Platform provides a single, programmable network infrastructure that eliminates bottlenecks between OpenShift workloads and the rest of the infrastructure while providing enterprise-grade security and offers flexible micro segmentation possibilities.

          Unified Networking - By extending the integrated overlay of the ACI fabric to OpenShift workloads, Pod and Service endpoints are managed as native ACI network resources at the same level as Bare Metal or Virtual Machines.

          No Egress Router Required - Applications running in OpenShift Pods that need high-bandwidth, low-latency traffic for data transfers external to the cluster suffer the bottleneck imposed by the egress router implementation. Pods can easily connect with other fabric or non-fabric endpoints leveraging the combined hardware and software distributed routing and switching provided by ACI and the ACI CNI plugin.

          Simplified Operations and Enhanced Visibility - Whether policies and networking are defined by the devops or fabric or security admins, the APIC becomes the single source of truth for operations and security governance. Kubernetes workload visibility is provided to Network Operators within their familiar Cisco tools.

          Hardware-assisted Load Balancing - Ingress connections to LoadBalancer-type services using ACI’s Policy Based Redirect technology moves load balancing into the fabric's hardware.

          Granular Security - Security can be implemented by using native NetworkPolicy or by using ACI EPGs and contracts, or both models complementing each other. Network policy of container workloads becomes a shared responsibility while ensuring global compliance and tenant constraints are enforced.
        displayName: Cisco ACI Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cisco
        - aci
        links:
        - name: Cisco ACI Homepage
          url: https://www.cisco.com/c/en/us/td/docs/switches/datacenter/aci/apic/sw/kb/b_Kubernetes_Integration_with_ACI.html
        maintainers:
        - email: abhishs9@cisco.com
          name: 'Abhishek Sharma '
        - email: sumitnaiksatam@gmail.com
          name: Sumit Naiksatam
        - email: vlella@cisco.com
          name: Vishal Lella
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Cisco
        relatedImages:
        - quay.io/noiro/aci-containers-operator@sha256:2b2d685e254c206be5371285852a8f95d0318af69cb85b786427969e035c125c
        - quay.io/noiro/acc-provision-operator@sha256:9c8fa6d705602af13d539fb919e72e7ee5064c6998fc9ca95d10108e986c68fe
        version: 1.0.10
      entries:
      - name: aci-containers-operator.v1.0.10
        version: 1.0.10
      name: stable
    defaultChannel: stable
    packageName: aci-containers-operator
    provider:
      name: Cisco
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Accuknox
      provider-url: https://accuknox.com
    name: kubearmor-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubearmor-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          com.redhat.openshift.versions: v4.12
          createdAt: "2023-05-01T05:17:07Z"
          description: The KubeArmor Operator deloys and manages KubeArmor daemonset
            and its componenets within a cluster.
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.28.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/kubearmor/KubeArmor
          support: Accuknox
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          The KubeArmor Operator deploys and manages KubeArmor and its components:
          - KubeArmor Daemonset
          - KubeArmor Relay Server
          - KubeArmor Controller to manage CRDs (KubeArmorPolicy, KubeArmorHostPolicy)

          KubeArmor is a cloud-native runtime security enforcement system that restricts the behavior (such as process
          execution, file access, and networking operations) of pods, containers, and nodes (VMs) at the system level.

          KubeArmor leverages Linux security modules (LSMs) such as AppArmor, SELinux, or BPF-LSM to enforce the user-specified policies. KubeArmor generates rich alerts/telemetry events with container/pod/namespace identities by leveraging eBPF.

          ## Documentation

          github: [https://github.com/kubearmor/KubeArmor](https://github.com/kubearmor/KubeArmor)
          website: [https://kubearmor.io/](https://kubearmor.io/)
        displayName: Kubearmor Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubearmor-operator
        - runtime-security
        - security
        - cloud-native
        links:
        - name: Kubearmor
          url: https://kubearmor.io
        maintainers:
        - email: ramakant@accuknox.com
          name: Ramakant Sharma
        maturity: alpha
        provider:
          name: Accuknox
          url: https://www.accuknox.com/
        relatedImages:
        - docker.io/kubearmor/kubearmor-init@sha256:44fa534587972c5f85e3e6f362dd8268e9088a33c5657a21b72a7e8563164fb5
        - docker.io/kubearmor/kubearmor-relay-server@sha256:695f9ade3382bcf858077e2a7b9b01e0713071517603778d72f7aca5a6d71323
        - docker.io/kubearmor/kubearmor-controller@sha256:ccec0fb8bb1e91bd48631cc89baf752b842f19c33b50edab08f8b33e57af64d4
        - docker.io/kubearmor/kubearmor-operator@sha256:8c2d56781a2adf78ee58f956e01417f38b6ad2f319feada6470e93f4f9bf0808
        - docker.io/kubearmor/kubearmor@sha256:52f7049ecdbfa779663d2cd44ddd7f36aecd8ccd2ca91e69709b628600029f13
        version: 1.0.0
      entries:
      - name: kubearmor-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: kubearmor-operator.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.kubearmor.com/v1",
                "kind": "KubeArmorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kubearmorconfig",
                    "app.kubernetes.io/part-of": "kubearmoroperator"
                  },
                    "name": "kubearmorconfig-default",
                    "namespace": "placeholder"
                },
                "spec": {
                  "adapters": {
                    "elasticsearch": {
                      "url": "",
                      "enabled": false,
                      "auth": {
                        "secretName": "elastic-secret",
                        "usernameKey": "username",
                        "passwordKey": "password",
                        "allowInsecureTLS": false,
                        "caCertSecretName": "",
                        "caCertKey": "ca.crt"
                      },
                      "alertsIndex": "kubearmor-alerts"
                    }
                  },
                  "recommendedPolicies": {
                      "enable": true,
                      "matchExpressions": [
                          {
                              "key": "namespace",
                              "operator": "NotIn",
                              "values": [
                                  "kube-system",
                                  "kubearmor",
                                  "openshift-infra",
                                  "openshift-ingress",
                                  "openshift-ingress-canary",
                                  "openshift-ingress-operator",
                                  "openshift-insights",
                                  "openshift-kni-infra",
                                  "openshift-kube-apiserver",
                                  "openshift-kube-apiserver-operator",
                                  "openshift-kube-controller-manager",
                                  "openshift-kube-controller-manager-operator",
                                  "openshift-kube-scheduler",
                                  "openshift-kube-scheduler-operator",
                                  "openshift-kube-storage-version-migrator",
                                  "openshift-kube-storage-version-migrator-operator",
                                  "openshift-logging",
                                  "openshift-machine-api",
                                  "openshift-machine-config-operator",
                                  "openshift-managed-node-metadata-operator",
                                  "openshift-managed-upgrade-operator",
                                  "openshift-marketplace",
                                  "openshift-monitoring",
                                  "openshift-multus",
                                  "openshift-must-gather-operator",
                                  "openshift-network-diagnostics",
                                  "openshift-network-node-identity",
                                  "openshift-network-operator",
                                  "openshift-node",
                                  "openshift-nutanix-infra",
                                  "openshift-oauth-apiserver",
                                  "openshift-observability-operator",
                                  "openshift-ocm-agent-operator",
                                  "openshift-openstack-infra",
                                  "openshift-operator-lifecycle-manager",
                                  "openshift-operators",
                                  "openshift-operators-redhat",
                                  "openshift-osd-metrics",
                                  "openshift-ovirt-infra",
                                  "openshift-ovn-kubernetes",
                                  "openshift-package-operator",
                                  "openshift-rbac-permissions",
                                  "openshift-route-controller-manager",
                                  "openshift-route-monitor-operator",
                                  "openshift-security",
                                  "openshift-service-ca",
                                  "openshift-service-ca-operator",
                                  "openshift-splunk-forwarder-operator",
                                  "openshift-sre-pruning",
                                  "openshift-user-workload-monitoring",
                                  "openshift-validation-webhook",
                                  "openshift-velero",
                                  "openshift-vsphere-infra",
                                  "openshift",
                                  "openshift-addon-operator",
                                  "openshift-apiserver",
                                  "openshift-apiserver-operator",
                                  "openshift-aqua",
                                  "openshift-authentication",
                                  "openshift-authentication-operator",
                                  "openshift-backplane",
                                  "openshift-backplane-cee",
                                  "openshift-backplane-csa",
                                  "openshift-backplane-cse",
                                  "openshift-backplane-csm",
                                  "openshift-backplane-managed-scripts",
                                  "openshift-backplane-mcs-tier-two",
                                  "openshift-backplane-mobb",
                                  "openshift-backplane-srep",
                                  "openshift-backplane-tam",
                                  "openshift-cloud-controller-manager",
                                  "openshift-cloud-controller-manager-operator",
                                  "openshift-cloud-credential-operator",
                                  "openshift-cloud-ingress-operator",
                                  "openshift-cloud-network-config-controller",
                                  "openshift-cloud-platform-infra",
                                  "openshift-cluster-csi-drivers",
                                  "openshift-cluster-machine-approver",
                                  "openshift-cluster-node-tuning-operator",
                                  "openshift-cluster-samples-operator",
                                  "openshift-cluster-storage-operator",
                                  "openshift-cluster-version",
                                  "openshift-codeready-workspaces",
                                  "openshift-config",
                                  "openshift-config-managed",
                                  "openshift-config-operator",
                                  "openshift-console",
                                  "openshift-console-operator",
                                  "openshift-console-user-settings",
                                  "openshift-controller-manager",
                                  "openshift-controller-manager-operator",
                                  "openshift-custom-domains-operator",
                                  "openshift-customer-monitoring",
                                  "openshift-deployment-validation-operator",
                                  "openshift-dns",
                                  "openshift-dns-operator",
                                  "openshift-etcd",
                                  "openshift-etcd-operator",
                                  "openshift-host-network",
                                  "openshift-image-registry"
                              ]
                          }
                      ]
                  },
                  "defaultCapabilitiesPosture": "audit",
                  "defaultFilePosture": "audit",
                  "defaultNetworkPosture": "audit",
                  "defaultVisibility": "process,network",
                  "enableStdOutLogs": false,
                  "enableStdOutAlerts": false,
                  "enableStdOutMsgs": false,
                  "seccompEnabled": false
                }
              }
            ]
          alm-examples-metadata: '{"kubearmorconfig-default":{"description":"Default
            KubeArmorConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Security, Monitoring
          certified: "true"
          com.redhat.openshift.versions: v4.9
          containerImage: docker.io/kubearmor/kubearmor-operator@sha256:d6c6b16e636474568835ded97cecb5ed17bc6004501227f6b6d63522ed57b01c
          createdAt: "2024-03-20T17:43:15Z"
          description: KubeArmor Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: kubearmor
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kubearmor/KubeArmor/blob/main/pkg/KubeArmorOperator/
          support: Accuknox
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeArmorConfig is the Schema for the kubearmorconfigs API
            kind: KubeArmorConfig
            name: kubearmorconfigs.operator.kubearmor.com
            version: v1
        description: |2-

          ## Prerequisites

          1. Create a new project for KubeArmor (Recommended)

             ```
             oc new-project kubearmor
             ```

           2. Create [KubeArmor SCC](https://github.com/kubearmor/KubeArmor/blob/main/pkg/KubeArmorOperator/config/rbac/kubearmor-scc.yaml)

             ```
             oc create -f https://raw.githubusercontent.com/kubearmor/KubeArmor/main/pkg/KubeArmorOperator/config/rbac/kubearmor-scc.yaml
             ```

             Note: please change the serviceaccount namespace if you are using some other namespace than the kubearmor.

          KubeArmor is a cloud-native runtime security enforcement system that restricts the behavior
          (such as process execution, file access, and networking operations) of pods, containers,
          and nodes (VMs) at the system level.
          KubeArmor leverages Linux security modules (LSMs) such as AppArmor, SELinux, or BPF-LSM to enforce the
          user-specified policies. KubeArmor generates rich alerts/telemetry events with container/pod/namespace
          identities by leveraging eBPF.

          ## Features and Benefits

          ### Harden Infrastructure

          safeguards vital paths like certificate bundles, enforces security rules based on MITRE, STIGs, and CIS guidelines,
          and restricts access to raw database tables, enhancing overall cluster security.

          ### Least Permissive Access

          controls sensitive asset access through process and network whitelisting,
          reducing security risks and maintaining system integrity.

          ### Application Behavior

          in-depth monitoring of process executions, file system access, service bindings,
          ingress and egress connections, and sensitive system call profiling

          ## Documentation

          github: [https://github.com/kubearmor/KubeArmor](https://github.com/kubearmor/KubeArmor)

          website: [https://kubearmor.io/](https://kubearmor.io/)

          ## Installing the KubeArmor Operator

          The KubeArmor Operator, available through Red Hat's Operator catalog, offers stable update channel.
          **Stable Channel**: This channel provides the latest stable version of KubeArmor along with its patches.

          The KubeArmor operator is responsible for deploying various components of the KubeArmor application, including the KubeArmor
          Daemonset, KubeArmor Controller, KubeArmor Relay Server, and KubeArmor Snitch. The role of KubeArmor Snitch is to gather information
          about the enabled Linux security modules on the host, which is utilized by KubeArmor to enforce runtime security policies.
          Additionally, KubeArmor Snitch collects details about the container runtime socket and storage volume cluster configurations.
          This deployment is executed as a short-term job and necessitates root user privileges. KubeArmor Snitch is certified by RedHat
          as a container with root privileges.

          KubeArmor Operator comes with a custom resource:

          - **KubeArmorConfig** - User will provide the initial configurations of the KubeArmor using KubeArmorConfig CR.

          ### KubeArmorConfig Customization Options:

          | Parameter                                 | Description                                          | Default                                 |
          | ---                                       | ---                                                  | ---                                     |
          | `alertThrottling`                         | Enable throttling for the KubeArmor generated alerts | False                                   |
          | `defaultFilePosture`                      | Default file posture for KubeArmor                   | audit                                   |
          | `defaultNetworkPosture`                   | Default network posture for KubeArmor                | audit                                   |
          | `defaultCapabilitiesPosture`              | Default capabilities posture for KubeArmor           | audit                                   |
          | `defaultVisibility`                       | Default visibility for KubeArmor                     | process,file,network                    |
          | `enableStdOutLogs` |  show stdout logs for relay server | false |
          | `enableStdOutAlerts` |  show stdout alerts for relay server | false |
          | `enableStdOutMsgs` |  show stdout messages for relay server | false |
          | `kubearmorImage.image`                    | Default image for KubeArmor Daemonset                | kubearmor/kubearmor-ubi:stable          |
          | `kubearmorImage.imagePullPolicy`          | Default imagePullPolicy for KubeArmor Daemonset      | Always                                  |
          | `kubearmorInitImage.image`                | Default image for KubeArmor Init Container           | kubearmor/kubearmor-init:stable         |
          | `kubearmorInitImage.imagePullPolicy`      | Default imagePullPolicy for KubeArmor Init Container | Always                                  |
          | `kubearmorRelayImage.image`               | Default image for KubeArmor Relay Server             | kubearmor/kubearmor-relay-server:latest |
          | `kubearmorRelayImage.imagePullPolicy`     | Default imagePullPolicy for KubeArmor Relay Server   | Always                                  |
          | `kubearmorControllerImage.image`          | Default image for KubeArmor Controller               | kubearmor/kubearmor-controller:latest   |
          | `kubearmorControllerImage.imagePullPolicy`| Default imagePullPolicy for KubeArmor Controller     | Always                                  |
          | `maxAlertPerSec`                          | Maximum no of alerts per second to trigger throttling| 10                                      |
          | `recommendedPolicies`                     | Recommended Policies Spec                            | -                                       |
          | `recommendedPolicies.matchExpressions`    | Target Namespaces for recommended policies           | -                                       |
          | `recommendedPolicies.excludePolicy`       | Exclude policies from recommended policies           | -                                       |
          | `seccompEnabled`                          | Deploy KubeArmor Daemon with default seccomp profile | True                                    |
          | `throttleSec`                             | Thottling duration in seconds                        | 30                                      |

          Users are not required to specify the images in the KubearmorConfig custom resource (CR). By default, the operator will utilize the images published with the operator release.

          For detailed information on Default Posture refer [this guide](https://github.com/kubearmor/KubeArmor/blob/main/getting-started/default_posture.md)

          similarly for default visibility refer [this guide](https://github.com/kubearmor/KubeArmor/blob/main/getting-started/kubearmor_visibility.md)

          ## Support

          For Support please join KubeArmor Slack [Here](https://join.slack.com/t/kubearmor/shared_invite/zt-1ltmqdbc6-rSHw~LM6MesZZasmP2hAcA)
        displayName: KubeArmor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - security
        - alerting
        - runtime-security
        - cloud native
        - container security
        links:
        - name: Website
          url: https://kubearmor.io
        - name: Github
          url: https://github.com/kubearmor/KubeArmor.git
        - name: Slack
          url: https://join.slack.com/t/kubearmor/shared_invite/zt-1ltmqdbc6-rSHw~LM6MesZZasmP2hAcA
        maintainers:
        - email: ramakant@accuknox.com
          name: Ramakant Sharma
        maturity: stable
        provider:
          name: Accuknox
          url: https://accuknox.com
        relatedImages:
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:1153a4592746b05e019bde4d818d176ff9350c013f84d49829032540de882841
        - docker.io/kubearmor/kubearmor-operator@sha256:d6c6b16e636474568835ded97cecb5ed17bc6004501227f6b6d63522ed57b01c
        - docker.io/kubearmor/kubearmor-snitch@sha256:c3d0f45590f9a0ebb50254047f12a7468844bcae7961eccb11f5ef578832c0eb
        - docker.io/kubearmor/kubearmor-ubi@sha256:629e295a2e64953bf95f3f3e842c42ff665ed6e4db32db8e43dcde4befdd3fc2
        - docker.io/kubearmor/kubearmor-init@sha256:e30ee508aa95d273f74a543e1b412b98c1714d04f093b59991c0ec2abaa450be
        - docker.io/kubearmor/kubearmor-relay-server@sha256:0a1a20ef02bd05eaa4a900948cadee6ae7650b0dde70abe1935608eaa5f931b7
        - docker.io/kubearmor/kubearmor-controller@sha256:2d369138cdbb31468c41e9ce70eda986a16bbeb9dd1192cd8a011c9df8a83857
        version: 1.5.2
      entries:
      - name: kubearmor-operator.v1.5.2
        version: 1.5.2
      name: stable
    defaultChannel: stable
    packageName: kubearmor-operator-certified
    provider:
      name: Accuknox
      url: https://accuknox.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Tigera
      provider-url: https://tigera.io/
    name: tigera-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tigera-operator.v1.20.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:bd93e66be5cfbc123fe4e8c3f6e8ddf7654217bf26136f06dcf23f28b53063d2
          createdAt: "2021-10-05T20:02:14.564507Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.20.4
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:bd93e66be5cfbc123fe4e8c3f6e8ddf7654217bf26136f06dcf23f28b53063d2
        version: 1.20.4
      entries:
      - name: tigera-operator.v1.20.4
        version: 1.20.4
      - name: tigera-operator.v1.20.3
        version: 1.20.3
      - name: tigera-operator.v1.20.2
        version: 1.20.2
      - name: tigera-operator.v1.20.1
        version: 1.20.1
      name: release-v1.20
    - currentCSV: tigera-operator.v1.23.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:f258275fe3d48c2ae16346a57f628523a6f2fb7b854b6d43b169056e158537f1
          createdAt: "2022-01-12T22:55:57.409244Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.23.4
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:f258275fe3d48c2ae16346a57f628523a6f2fb7b854b6d43b169056e158537f1
        version: 1.23.4
      entries:
      - name: tigera-operator.v1.23.4
        version: 1.23.4
      - name: tigera-operator.v1.23.3
        version: 1.23.3
      - name: tigera-operator.v1.23.2
        version: 1.23.2
      - name: tigera-operator.v1.23.1
        version: 1.23.1
      - name: tigera-operator.v1.23.0
        version: 1.23.0
      name: release-v1.23
    - currentCSV: tigera-operator.v1.25.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:e862708c700df55f99a9c9bb3c453e31f22aef1993d6079d80433eb8da831c94
          createdAt: "2022-02-17T05:39:13.720885Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.25.1
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:e862708c700df55f99a9c9bb3c453e31f22aef1993d6079d80433eb8da831c94
        version: 1.25.1
      entries:
      - name: tigera-operator.v1.25.1
        version: 1.25.1
      name: release-v1.25
    - currentCSV: tigera-operator.v1.28.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:9e8e47045fc67a2b00f9f2fd575fa526d7ddcac50af0b8b6d583b7518ccce120
          createdAt: "2023-05-17T16:13:52.148441156Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.28.13
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:9e8e47045fc67a2b00f9f2fd575fa526d7ddcac50af0b8b6d583b7518ccce120
        version: 1.28.13
      entries:
      - name: tigera-operator.v1.28.13
        version: 1.28.13
      name: release-v1.28
    - currentCSV: tigera-operator.v1.29.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:2a5acd29d8dca6adb318af04dbb15d93d899ed2b28b2ee9aaac09147395efb58
          createdAt: "2023-03-31T20:49:52.518230079Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.29.3
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:2a5acd29d8dca6adb318af04dbb15d93d899ed2b28b2ee9aaac09147395efb58
        version: 1.29.3
      entries:
      - name: tigera-operator.v1.29.3
        version: 1.29.3
      name: release-v1.29
    - currentCSV: tigera-operator.v1.30.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:431f037ff18b5c867d01312e42671effc55602421aeed25dd3f6109f70596b4a
          createdAt: "2023-11-17T00:52:31.263564294Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.30.9
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: APIServer
            name: apiservers.operator.tigera.io
            version: v1
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:431f037ff18b5c867d01312e42671effc55602421aeed25dd3f6109f70596b4a
        version: 1.30.9
      entries:
      - name: tigera-operator.v1.30.9
        version: 1.30.9
      name: release-v1.30
    - currentCSV: tigera-operator.v1.32.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:715ac9a30f8a9579e44258af20de354715429e11836b493918e9e1a696e9b028
          createdAt: "2023-12-15T23:45:03.964812504Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.32.3
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:715ac9a30f8a9579e44258af20de354715429e11836b493918e9e1a696e9b028
        version: 1.32.3
      entries:
      - name: tigera-operator.v1.32.3
        version: 1.32.3
      name: release-v1.32
    - currentCSV: tigera-operator.v1.36.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:3341fa9475c0325b86228c8726389f9bae9fd6c430c66fe5cd5dc39d7bb6ad4b
          createdAt: "2025-02-06T02:39:01.007218336Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.36.5
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:3341fa9475c0325b86228c8726389f9bae9fd6c430c66fe5cd5dc39d7bb6ad4b
        version: 1.36.5
      entries:
      - name: tigera-operator.v1.36.5
        version: 1.36.5
      name: release-v1.36
    - currentCSV: tigera-operator.v1.38.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "calicoNetwork": {
                    "linuxDataplane": "BPF"
                  },
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing, Monitoring, Networking, Security
          containerImage: quay.io/tigera/operator@sha256:e0a34b265aebce1a2db906d8dad99190706e8bf3910cae626b9c2eb6bbb21775
          createdAt: "2025-05-06T13:27:07.802897329Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          features.operators.openshift.io/valid-subscription: No subscription required
          olm.skipRange: <1.38.0
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift.
        displayName: Tigera Operator v1.38
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tigera-operator
        - networking
        - security
        - monitoring
        links:
        - name: Tigera
          url: https://tigera.io/
        - name: Calico Introduction
          url: https://docs.tigera.io/calico/latest/about/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.tigera.io/calico/latest/getting-started/kubernetes/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
          url: https://tigera.io/
        relatedImages:
        - quay.io/tigera/operator@sha256:e0a34b265aebce1a2db906d8dad99190706e8bf3910cae626b9c2eb6bbb21775
        version: 1.38.0
      entries:
      - name: tigera-operator.v1.38.0
        version: 1.38.0
      name: release-v1.38
    defaultChannel: release-v1.38
    packageName: tigera-operator
    provider:
      name: Tigera
      url: https://tigera.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solution-Soft
      provider-url: www.solution-soft.com
    name: timemachine-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: timemachine-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tm.solution-soft.com/v1alpha1",
                "kind": "TimeMachine",
                "metadata": {
                  "name": "timemachine-sample"
                },
                "spec": {
                  "cluster_provider": "ibm",
                  "aws": {
                    "shared_volume_name": ""
                  },
                  "ibm": {
                    "persistentvolume": {
                      "storageclass": "ibmc-file-silver-gid"
                    }
                  },
                  "tm_lichost": "52.152.148.40",
                  "tm_licpass": "qwaszx",
                  "tm_licport": "57777",
                  "imagePullSecrets": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
          createdAt: "2024-02-05T09:46:17Z"
          description: Time Machine Operator enables time travel of OpenShift pods,
            without the need to change the system clock on any of the cluster nodes
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/timemachine-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/timemachine-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=0.0.4 <1.0.17'
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: SolutionSoft Systems, Inc. - support@solution-soft.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Time Machine
            displayName: Time Machine
            kind: TimeMachine
            name: timemachines.tm.solution-soft.com
            version: v1alpha1
        description: |-
          Time Machine operator enables time travel of OpenShift pods/deployments.

          To time travel a pod where your application is, you just need to install Time Machine operator and create a Time Machine custom resource in the same namespace, without the need of changing the container images of the application you’re using, or adding containers to the pod with your application.

          Time travelling is done on the namespace level where all the (configured) target pods/deployments see the same virtual time.
        displayName: Time Machine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - time shift
        - time travel
        - virtual clocks
        - time machine
        - virtual time
        - solution-soft
        links:
        - name: Timemachine Operator
          url: https://catalog.redhat.com/software/operators/detail/5e98727a6c5dcb34dfbb19e3
        maintainers:
        - email: dtrajkovic@solution-soft.com
          name: Dragan Trajkovic
        maturity: alpha
        provider:
          name: Solution-Soft
          url: www.solution-soft.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0c304c11986790f446604319dfc29db494b15f5f9fa13f2c436688ec29c92dbe
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
        - registry.connect.redhat.com/rhel-timemachine-container/virtual-clock-testing-tool@sha256:a81a823328b0ee795c40668c89a08bcf545c5192d9eed0146071e3721a0a8372
        version: 1.0.17
      entries:
      - name: timemachine-operator.v1.0.17
        version: 1.0.17
      - name: timemachine-operator.v1.0.16
        version: 1.0.16
      name: alpha
    defaultChannel: alpha
    packageName: timemachine-operator-rhmp
    provider:
      name: Solution-Soft
      url: www.solution-soft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: ironic-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ironic-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "Ironic",
                "metadata": {
                  "name": "ironic"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseInstance": "openstack",
                  "ironicAPI": {},
                  "ironicConductors": [
                    {
                      "storageRequest": "10G"
                    }
                  ],
                  "ironicInspector": {},
                  "ironicNeutronAgent": {},
                  "secret": "osp-secret",
                  "serviceUser": "ironic",
                  "storageClass": "local-storage"
                }
              },
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "IronicAPI",
                "metadata": {
                  "name": "ironicapi-sample"
                },
                "spec": {
                  "databaseHostname": "openstack.ironic-kuttl-tests.svc"
                }
              },
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "IronicConductor",
                "metadata": {
                  "name": "ironicconductor-sample"
                },
                "spec": {
                  "databaseHostname": "openstack.ironic-kuttl-tests.svc",
                  "storageRequest": "10G"
                }
              },
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "IronicInspector",
                "metadata": {
                  "name": "ironicinspector-sample"
                },
                "spec": {
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "IronicNeutronAgent",
                "metadata": {
                  "name": "ironic-neutron-agent-sample",
                  "namespace": "openstack"
                },
                "spec": {
                  "secret": "osp-secret"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:02Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IronicAPI is the Schema for the ironicapis API
            displayName: Ironic API
            kind: IronicAPI
            name: ironicapis.ironic.openstack.org
            version: v1beta1
          - description: IronicConductor is the Schema for the ironicconductors Conductor
            displayName: Ironic Conductor
            kind: IronicConductor
            name: ironicconductors.ironic.openstack.org
            version: v1beta1
          - description: IronicInspector is the Schema for the ironicinspectors
            displayName: Ironic Inspector
            kind: IronicInspector
            name: ironicinspectors.ironic.openstack.org
            version: v1beta1
          - description: IronicNeutronAgent is the Schema for the ML2 networking-baremetal's
              agent
            displayName: Ironic Neutron Agent
            kind: IronicNeutronAgent
            name: ironicneutronagents.ironic.openstack.org
            version: v1beta1
          - description: Ironic is the Schema for the ironics API
            displayName: Ironic
            kind: Ironic
            name: ironics.ironic.openstack.org
            version: v1beta1
        description: Ironic Operator
        displayName: Ironic Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Identity
        - Ironic
        links:
        - name: Ironic Operator
          url: https://github.com/openstack-k8s-operators/ironic-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/ironic-rhel9-operator@sha256:184a1783a3b2a38bdb23f9811243414a2b21c16d78ca79ed83e2bcaa720d199b
        - registry.redhat.io/rhoso/openstack-ironic-api-rhel9@sha256:86c5330a3065a11f251e8c333bd855d37294c4e9d36cc68208fd16b3f95ef15a
        - registry.redhat.io/rhoso/openstack-ironic-conductor-rhel9@sha256:4950d74335eb48ed04f1a4f602730f1dfd42f139233a4edfebef0d8403c00b2d
        - registry.redhat.io/rhoso/openstack-ironic-inspector-rhel9@sha256:34ff13efc254d4390fbef0a5c088c569241d44d063ec16aed42e36d8179e0c85
        - registry.redhat.io/rhoso/openstack-ironic-pxe-rhel9@sha256:d49af96baa8d90dad3b244ed8c006a9945c41d05e5fa197cc35cc662b49cdd7c
        - registry.redhat.io/rhoso/openstack-ironic-neutron-agent-rhel9@sha256:132627eaf483353c5daefbbc84c6b06f51c50d0813a229ddc558e9df18bb5578
        - registry.redhat.io/rhoso/ironic-python-agent-rhel9@sha256:f41166809d2e7d76446c01f81d325a320df21e64f4b776d0f83cadcb9387eb7f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: ironic-operator.v1.0.6
        version: 1.0.6
      - name: ironic-operator.v1.0.5
        version: 1.0.5
      - name: ironic-operator.v1.0.4
        version: 1.0.4
      - name: ironic-operator.v1.0.3
        version: 1.0.3
      - name: ironic-operator.v1.0.2
        version: 1.0.2
      - name: ironic-operator.v1.0.1
        version: 1.0.1
      - name: ironic-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: ironic-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kove
      provider-url: https://kove.com
    name: kove-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kove-operator.v3.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "containers.kove.com/v1",
                "kind": "KoveMemoryConfig",
                "metadata": {
                  "labels": {
                    "KoveMemoryConfig": ""
                  },
                  "name": "example-memory-config"
                },
                "spec": {
                  "container": "",
                  "namespace": "",
                  "size": ""
                }
              },
              {
                "apiVersion": "containers.kove.com/v1",
                "kind": "KoveSystem",
                "metadata": {
                  "labels": {
                    "KoveSystem": ""
                  },
                  "name": "kove-system"
                },
                "spec": {
                  "autogrowInitialIncrement": "1 GiB",
                  "autogrowMaxIncrement": "1 TiB",
                  "l4MaxMemoryPercent": 75,
                  "managementConsoleAddress": "",
                  "memoryPlusLocalMemoryMaxSize": "0 MiB"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: quay.io/kove/kove-operator@sha256:0fb5fa2e6702d5e5bf692c25f190f65d306fc72cdceebf0814ff64adcb5f6994
          createdAt: "2025-09-25T16:31:36Z"
          description: Configure containers with Kove® pooled memory
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: kove-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Kove
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configure containers with Kove® pooled memory
            displayName: Kove Memory Configuration
            kind: KoveMemoryConfig
            name: kovememoryconfigs.containers.kove.com
            version: v1
          - description: Configure cluster with Kove:SDM™
            displayName: Kove System Configuration
            kind: KoveSystem
            name: kovesystems.containers.kove.com
            version: v1
        description: |
          ## About the Operator
          The Kove:SDM™ operator integrates Kove's software-defined memory
          solution with Kubernetes and OpenShift clusters. The operator enables
          containers to transparently provision any amount of memory, enabling total
          container memory capacity to exceed physical hardware limitations.

          Use the Kove:SDM™ operator to:

          * Create custom resources to automatically allocate memory to specific
            containers
          * Increase container flexibility with dynamic memory capacity that
            auto-grows as needed
          * Scale up container density by removing memory as an orchestration
            bottleneck

          ## Using the Operator
          Using the operator is as simple as creating a custom resource which
          specifies an initial Kove memory allocation size, a namespace, and a
          container name. Containers in the specified namespace with a matching
          container name will automatically be provided with a Kove memory allocation
          which will auto-grow as needed.

          ## About Kove
          Kove makes memory a provisionable resource, just like storage, CPUs, and
          networking. Kove:SDM™ provides a transparent, linearly scalable
          memory resource accessible to any application without code changes, from
          bare-metal apps to virtual machines and containers.
        displayName: Kove:SDM™ Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kove
        - Pooled Memory
        - Memory Pool
        - Memory Target
        - External Memory
        - Memory
        links:
        - name: Kove
          url: https://kove.com
        maintainers:
        - email: contact-kove@kove.com
          name: Kove
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Kove
          url: https://kove.com
        relatedImages:
        - registry.access.redhat.com/ubi8/pause@sha256:eacbaaadc932633330e799eda24e9b731b5a4ba7ed823ee70791d21029724a9a
        - quay.io/kove/kove-agent@sha256:9d026b925e1d414a531530a30ed8b5f10b421ee12c7c13949ea0ee9907d36b75
        - quay.io/kove/kove-client-drivers@sha256:1c1786e1bbf16609218154126b4340b2f7c4b71cbd9045589402f75c032ea38d
        - quay.io/kove/kove-compatibility-detection@sha256:0303830a1f785b5e4ce6824db7c50f8feee2678fd5e87ded983686ab3f60b232
        - quay.io/kove/kove-operator@sha256:0fb5fa2e6702d5e5bf692c25f190f65d306fc72cdceebf0814ff64adcb5f6994
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fa0acefda43e309e7f9f31f5247c933d0b868ff1573ece2f38e0549d7e2ec942
        - quay.io/kove/kove-container-hook@sha256:66ee33d59ced63406d9b40f3a3226fab7f52071f8bed9aa1ba7d859837038528
        - quay.io/kove/kove-xmem@sha256:41bfc536a278153b0a4460d2836b56762b56cf50b94d17e123a63be066c28fe7
        - quay.io/kove/kove-client-drivers@sha256:716ccda2d8f082a954f2042fd189fec257bd2dcb2c50e497bad42b9a9e5f0e3d
        version: 3.0.2
      entries:
      - name: kove-operator.v3.0.2
        version: 3.0.2
      - name: kove-operator.v3.0.1
        version: 3.0.1
      - name: kove-operator.v3.0.0
        version: 3.0.0
      name: stable
    defaultChannel: stable
    packageName: kove-operator
    provider:
      name: Kove
      url: https://kove.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Fujitsu
      provider-url: https://www.postgresql.fastware.com/
    name: fujitsu-enterprise-postgres-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fujitsu-enterprise-postgres-operator.v7.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPAction",
                "metadata": {
                  "name": "new-fep-action"
                },
                "spec": {
                  "fepAction": {
                    "args": [
                      "new-fep-sts-0",
                      "new-fep-sts-1"
                    ],
                    "type": "reload"
                  },
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "targetClusterName": "new-fep"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPExporter",
                "metadata": {
                  "name": "new-fep-exporter"
                },
                "spec": {
                  "fepExporter": {
                    "exporterLogLevel": "error",
                    "fepClusterList": [
                      "new-fep1"
                    ],
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "userCustomQueries": "usr_example:\n  query: \"SELECT EXTRACT(EPOCH FROM (now() - pg_last_xact_replay_timestamp())) as lag\"\n  master: true\n  metrics:\n    - lag:\n        usage: \"GAUGE\"\n        description: \"Replication lag behind master in seconds\""
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPLogging",
                "metadata": {
                  "name": "new-fep-logging"
                },
                "spec": {
                  "fepLogging": {
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPPgpool2",
                "metadata": {
                  "name": "new-fep-pgpool2"
                },
                "spec": {
                  "clientAuthMethod": "scram",
                  "count": 2,
                  "customhba": "local   all         all                               trust\nhost    all         all         127.0.0.1/32          trust\nhost    all         all         ::1/128               trust\n",
                  "customlogsize": "128Mi",
                  "customparams": "listen_addresses = '*'\npcp_listen_addresses = '*'\nnum_init_children = 32\nreserved_connections = 0\nenable_pool_hba = off\nallow_clear_text_frontend_auth = off\nauthentication_timeout = 80\nbackend_weight0 = 1\nbackend_weight1 = 1\nbackend_flag0 = 'ALWAYS_PRIMARY'\nbackend_flag1 = 'DISALLOW_TO_FAILOVER'\nconnection_cache = on\nmax_pool = 4\nlisten_backlog_multiplier = 2\nserialize_accept = off\nchild_life_time = 300\nclient_idle_limit = 0\nchild_max_connections = 0\nconnection_life_time = 0\nreset_query_list = 'ABORT; DISCARD ALL'\nclient_min_messages = info\nlog_min_messages = debug1\nlog_statement = on\nlog_per_node_statement = on\nlog_client_messages = on\nlog_hostname = on\nlog_connections = on\nlog_line_prefix = '%t: pid %p: '\nload_balance_mode = on\nignore_leading_white_space = on\nwhite_function_list = ''\nblack_function_list = 'currval,lastval,nextval,setval'\nblack_query_pattern_list = ''\ndatabase_redirect_preference_list = ''\napp_name_redirect_preference_list = ''\nallow_sql_comments = off\ndisable_load_balance_on_write = 'transaction'\nstatement_level_load_balance = on\nsr_check_period = 0\nsr_check_user = 'postgres'\ndelay_threshold = 0\nlog_standby_delay = 'none'\nssl = on\nssl_ciphers = 'HIGH:MEDIUM:+3DES:!aNULL'\nssl_prefer_server_ciphers = off\nssl_ecdh_curve = 'prime256v1'\nssl_dh_params_file = ''\nrelcache_expire = 0\nrelcache_size = 256\ncheck_temp_table = catalog\ncheck_unlogged_table = on\nenable_shared_relcache = off\nrelcache_query_target = primary\nwd_port0 = 9000\nfailover_on_backend_error = off\n",
                  "custompcp": "none",
                  "customsslcacert": "none",
                  "customsslcert": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                  "customsslkey": "none",
                  "fepclustername": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "limits": {
                    "cpu": "400m",
                    "memory": "512Mi"
                  },
                  "requests": {
                    "cpu": "200m",
                    "memory": "256Mi"
                  },
                  "serviceport": 9999,
                  "statusport": 9898
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPRestore",
                "metadata": {
                  "name": "new-fep-restore"
                },
                "spec": {
                  "fromFEPcluster": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "mcSpec": {
                    "limits": {
                      "cpu": "200m",
                      "memory": "300Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "restoretype": "latest",
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "toFEPcluster": "new-fep-2"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v2",
                "kind": "FEPCluster",
                "metadata": {
                  "name": "new-fep"
                },
                "spec": {
                  "fep": {
                    "customAnnotations": {
                      "allDeployments": {}
                    },
                    "forceSsl": true,
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "instances": 1,
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "podAntiAffinity": false,
                    "podDisruptionBudget": false,
                    "servicePort": 27500,
                    "syncMode": "off",
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "usePodName": true
                  },
                  "fepChildCrVal": {
                    "backup": {
                      "image": {
                        "pullPolicy": "IfNotPresent"
                      },
                      "mcSpec": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      },
                      "pgbackrestParams": "# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[global]\nrepo1-retention-full=7\nrepo1-retention-full-type=time\nlog-path=/database/log/backup\n",
                      "postScript": " ",
                      "preScript": " ",
                      "schedule": {
                        "num": 2
                      },
                      "schedule1": {
                        "schedule": "15 0 * * 0",
                        "type": "full"
                      },
                      "schedule2": {
                        "schedule": "15 0 * * 1-6",
                        "type": "incr"
                      },
                      "schedule3": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule4": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule5": {
                        "schedule": " ",
                        "type": " "
                      }
                    },
                    "customPgAudit": "# define pg audit custom params here to override defaults.\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[output]\nlogger = 'auditlog'\nlog_directory = '/database/log/audit'\nlog_truncate_on_rotation = on\nlog_filename = 'pgaudit-%a.log'\nlog_rotation_age = 1d\nlog_rotation_size = 0\n[rule]\n",
                    "customPgHba": "# define pg_hba custom rules here to be merged with default rules.\n# TYPE     DATABASE        USER        ADDRESS        METHOD\n",
                    "customPgParams": "# define custom postgresql.conf parameters below to override defaults.\n# Current values are as per default FEP deployment\n# If you add a library to shared_preload_libraries, add it after the default value.\n# fsep_operator_security is available in FEP15 and later container images.\nshared_preload_libraries='pgx_datamasking,pg_prewarm,pg_stat_statements,fsep_operator_security'\nsession_preload_libraries='pg_prewarm'\nmax_prepared_transactions = 100\nmax_worker_processes = 30\nmax_connections = 100\nwork_mem = 1MB\nmaintenance_work_mem = 12MB\nshared_buffers = 128MB\neffective_cache_size = 384MB\ncheckpoint_completion_target = 0.8\n# tcp parameters\ntcp_keepalives_idle = 30\ntcp_keepalives_interval = 10\ntcp_keepalives_count = 3\n# logging parameters in default fep installation\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\nlog_directory = '/database/log'\nlog_filename = 'logfile-%a.log'\nlog_file_mode = 0600\nlog_truncate_on_rotation = on\nlog_rotation_age = 1d\nlog_rotation_size = 0\nlog_checkpoints = on\nlog_line_prefix = '%e %t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h'\nlog_lock_waits = on\nlog_autovacuum_min_duration = 60s\nlogging_collector = on\npgaudit.config_file='/opt/app-root/src/pgaudit-cfg/pgaudit.conf'\nlog_replication_commands = on\nlog_min_messages = WARNING\nlog_destination = csvlog\n# wal_archive parameters in default fep installation\narchive_mode = on\narchive_command = 'pgbackrest --stanza=backupstanza --config=/database/userdata/pgbackrest.conf archive-push %p'\nwal_level = replica\nmax_wal_senders = 12\nwal_keep_segments = 64\ntrack_activities = on\ntrack_counts = on\npassword_encryption = 'scram-sha-256'\n",
                    "storage": {
                      "archivewalVol": {
                        "size": "1Gi"
                      },
                      "backupVol": {
                        "size": "2Gi"
                      },
                      "dataVol": {
                        "size": "2Gi"
                      },
                      "logVol": {
                        "size": "1Gi"
                      },
                      "tablespaceVol": {
                        "size": "512Mi"
                      },
                      "walVol": {
                        "size": "1200Mi"
                      }
                    },
                    "sysUsers": {
                      "pgRewindPassword": "rewind_password",
                      "pgRewindUser": "rewind_user",
                      "pgdb": "mydb",
                      "pgpassword": "mydbpassword",
                      "pgreplpassword": "repluserpwd",
                      "pgrepluser": "repluser",
                      "pguser": "mydbuser",
                      "tdepassphrase": "tde-passphrase"
                    },
                    "systemCertificates": {
                      "cacrt": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                      "crt": "-----BEGIN CERTIFICATE-----\nMIIDUTCCAjmgAwIBAgIRAMocW3qMoHrD6qRvMPppMkMwDQYJKoZIhvcNAQELBQAw\nNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9yIEt1\nYmVybmV0ZXMwHhcNMjEwMjA2MDQzMjM2WhcNMjYwMjA1MDQzMjM2WjA/MRAwDgYD\nVQQKEwdGdWppdHN1MSswKQYDVQQDEyJGVUpJVFNVIEVudGVycHJpc2UgUG9zdGdy\nZXMgU2VydmVyMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA4AI33yvH\nZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I2e4SceTKi6O3C/I1XuvWlpng\n5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4ANB5JWWqDOjrRT3o7nRPGXfila\nbP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYpmjdbfxabTz07ig0+6/cwKoRR\nxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTFYvmAH7gcdssSFBt8NPlUATHE\nsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6Wdgmu5H2pDml8CDNLDv98Aj7i\n+I5SRKKcVPlnuQIDAQABo1AwTjAdBgNVHSUEFjAUBggrBgEFBQcDAQYIKwYBBQUH\nAwIwDAYDVR0TAQH/BAIwADAfBgNVHSMEGDAWgBQcwrrUO0u+FhIUuVdrDRCQRsi6\nZjANBgkqhkiG9w0BAQsFAAOCAQEAm5dxBoI9pScOCvRAchg4CprdRDSJb9K6yB3O\nnCAxnM47iHeXnY3WlnI388kHu8DU7O4ba1tJbGs3KY9KzioPk43pU12jWkO1onoF\n+mTDjx/Ef1cYWA9r5q/LtgTa6Q2sxV4O2x67QW82aAnaxO34dV5zWCPIvAoovZBV\nHRT+BgCg3r2vD1RGKK2nl1aYJtWhO1SZubam+VttdZ/vbM9oOJctxmImsEtBXjkY\nKteePdQtLL5o03JhyXWyRshCq+HMmKf2KgyY8gvydGcP4eLQdBWcW40LcnVq6UjT\n0kJycJEKngMVademq1ZWHGaiYB7hyT6GhgIcHUJ2cKrPgbEh1Q==\n-----END CERTIFICATE-----",
                      "key": "-----BEGIN RSA PRIVATE KEY-----\nMIIEowIBAAKCAQEA4AI33yvHZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I\n2e4SceTKi6O3C/I1XuvWlpng5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4AN\nB5JWWqDOjrRT3o7nRPGXfilabP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYp\nmjdbfxabTz07ig0+6/cwKoRRxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTF\nYvmAH7gcdssSFBt8NPlUATHEsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6W\ndgmu5H2pDml8CDNLDv98Aj7i+I5SRKKcVPlnuQIDAQABAoIBAFPQYKlOzw/+BA0b\nyMIUpdctIMb/54CR/xR0mVw1DbSjigNVPjHUQvB8Y1B2FAITQObgJO06bAv0QdWN\nRb0/v/yYiNJDFjaLjaIAHlO/2+oWrXbFaZqgpVDJhB+e1xaZr2x7XGxm+p925k30\nl6pvIRY+I8JRKvZiV1VZHwL/R3JOtPr++xMZtLVjVOI+f+ySqJ+TZHuAjm49EKxj\ncEmmJ28b7QcziXsvKy00f+zbqLIBKXQdZAFU5eEr1BsDRXdRW+Kf0XIvftuy4BJZ\nvoKT+VGhEvF/qysswL4+6IAO6tpuYnnM0Y2d3sOGoWPkTcQK0MekYKzL/WmtCjNs\n9hodJtECgYEA5EWyhEOf4uOKe5TDp697UCUvXLoOR58FDe/S8XNvScn29jjOkqIg\nOMoqo9xAkJTNTzqn5UUdt1x/pgM2NxlPLFijrc0zQlX3SoOO2ryDd9WNi7YKtN16\nKJqa536WeZu2OEbuAZ+S3GALVy1RPeTNPnUOmKnF06DjDUGzLNCZy10CgYEA+zfw\n952DWuz1U0Z4wvAEqqcgUKXPKrkTXV/iUnjkDkrLYVr0ZofDNTXrdHl+UedFmaOC\ncieZn6DNhcdz5tKtyysGMH3g/qs9PfoGUngvcXsy0Egk04l3x1jc8TTCLqXZXYaQ\nHMsx51n+R58oncPtzYSUOr9qQ6PbC2CstTbFJA0CgYEAjGEsUliAB/jknfEzjXjG\nPdhQUxb8VyE864Az2lah9t/kJzFyIAziAeqZ5GE7t247AGFTBRTHHI8e1Qoemi3P\nWbc9GVIbFs1lIYbcIDpUIyrKPEP8O5QEXtoNLxXTFgAjRGKiVY87spjCAJ+W2ZhO\ne/1it5GYXfgQCYQA2yuBmOUCgYANRkR2YR1axaCk+NlSu6oTdmdPu6M5x7PNQE7O\nOtMaKjua9lppvIzFGAdMDUtueoEEAE7ZR1xnwfB6PDLUpJdIYAqgr1YfPt8qkjaZ\nTv56yZ7CwL0pbF8m6nwqRrZoDp1wwraEvvvxFKFKGY/k3kCHlpTakdjEoDjn3gDi\nRnWeVQKBgCEneMSzucei5LRppRtRaJw/Btll8qlPMlX3W7dxQ3cLwpmLOn0m51Fp\nPIZ44zYK8R6fu4+/sSrlfaIg86Ugeufp6YNxyNROKxUGza5vDIu5OftwWtBeg+UK\nZ8lLWNdX6pp7WMujmF3H1DrkBbauYMUKZ4UxUYtelgHERMePIxwb\n-----END RSA PRIVATE KEY-----"
                    }
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "false"
          containerImage: quay.io/fujitsu/fujitsu-enterprise-postgres-operator@sha256:b7dd55c0bfda84e398cbfbb87345fcec218cb866e286e18a111330321ab6ad54
          createdAt: "2025-09-10T11:50:37Z"
          description: OpenShift Operator for Fujitsu Enterprise Postgres
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.1.0 <7.3.0'
          operators.openshift.io/valid-subscription: '["Fujitsu Enterprise Postgres
            Advanced Edition Annual Subscription License per Core 17 Operator for
            Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/internal-objects: '["fepautoscales.fep.fujitsu.io","fepconfigs.fep.fujitsu.io","fepusers.fep.fujitsu.io","fepcerts.fep.fujitsu.io","fepvolumes.fep.fujitsu.io","fepbackups.fep.fujitsu.io"]'
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: fj-dbaas-tec@dl.jp.fujitsu.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: FEPAction
            name: fepactions.fep.fujitsu.io
            version: v1
          - kind: FEPAutoscale
            name: fepautoscales.fep.fujitsu.io
            version: v1
          - kind: FEPBackup
            name: fepbackups.fep.fujitsu.io
            version: v1
          - kind: FEPCert
            name: fepcerts.fep.fujitsu.io
            version: v1
          - kind: FEPCluster
            name: fepclusters.fep.fujitsu.io
            version: v2
          - kind: FEPConfig
            name: fepconfigs.fep.fujitsu.io
            version: v1
          - kind: FEPExporter
            name: fepexporters.fep.fujitsu.io
            version: v1
          - kind: FEPLogging
            name: feploggings.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2Cert
            name: feppgpool2certs.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2
            name: feppgpool2s.fep.fujitsu.io
            version: v1
          - kind: FEPRestore
            name: feprestores.fep.fujitsu.io
            version: v1
          - kind: FEPUpgrade
            name: fepupgrades.fep.fujitsu.io
            version: v1
          - kind: FEPUser
            name: fepusers.fep.fujitsu.io
            version: v1
          - kind: FEPVolume
            name: fepvolumes.fep.fujitsu.io
            version: v1
        description: "Fujitsu Enterprise Postgres 17 delivers an enterprise-grade
          PostgreSQL on OpenShift Container Platform.\n\nThis solution provides the
          flexibility of a hybrid cloud solution while delivering an enhanced distribution
          of PostgreSQL to support enterprise-level workloads and provide improved
          deployment and management, availability, performance, data governance and
          security.  \n\nAvailable as a multi-architecture container built for both
          amd64, s390x and ppc64le.\n\nThe download and Use of the Product is strictly
          subject to the terms of the End User License Agreement with Fujitsu Limited
          found at https://www.fast.fujitsu.com/fujitsu-enterprise-postgres-license-agreements.
          Where the Product that has been embedded as a whole or part into a third
          party program, only Authorised Customers may download and use the Product.\n"
        displayName: Fujitsu Enterprise Postgres Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - Fujitsu
        - fep
        links:
        - name: Fujitsu Enterprise Postgres
          url: https://www.postgresql.fastware.com/
        maintainers:
        - email: fj-dbaas-tec@dl.jp.fujitsu.com
          name: Fujitsu
        maturity: stable
        provider:
          name: Fujitsu
          url: https://www.postgresql.fastware.com/
        relatedImages:
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-upgrade@sha256:bf3b069a71202ea78c7c0c7b36315c8bef8e6a1c11cd0668dcd888329e26ecfb
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-backup@sha256:dd8b84e200c9d0fb32cd567a33968e15e15efc115ff120f54fdd8862c7ea3111
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentbit@sha256:9622d2191525da39e5100e1882dd2cfa3eda740d8094f3fbb71ca8a5734d2843
        - quay.io/fujitsu/fujitsu-enterprise-postgres-utils@sha256:57a3fb792480a0d3a8f075c48dee2d02f1d79cf459761e3305cf62799575a885
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-backup@sha256:42252444e2dcfe5435532f318cce041990d2b71efa26007924717924d30a46ca
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-utils@sha256:edeec4d00fd8282aaa7ec6fe3ec4cd84070a6402611002a18addcac87fb8ee94
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-server@sha256:fa4e8849d75daf331ba7d488bcb52c083cd92bb5f288e4b62c980754e299f618
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-pgpool2@sha256:185af2e4fd1701aded1ff5e44d4336562f5a219c30ba535c2a3097b914c24efe
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-upgrade@sha256:04af997b4f548ffec711b63a06c35715b4e7863537c83697b3ee0c1d95a6aec3
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-pgpool2@sha256:8eaace2f9c8e5e844992fe5e008cd1061af33cc79b4d442c4a83a65610bcc502
        - quay.io/fujitsu/fujitsu-enterprise-postgres-exporter@sha256:1c0e22aeab65ff3848772424fc3ea6baa6b58b0d88db5413bda160a08435bcff
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-upgrade@sha256:8a0f7ca1f59ba162424a308e8a70c4717135e2847e61210753b71876c9b0a75f
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-restore@sha256:09be02274427e3022849fb635a4c9abdc34696838cae39f3afafc8009567db9d
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-backup@sha256:78a3ad639d0c63858b9db0a5de36ce11a9ea121f4535e9ca07ac7defb689bc0d
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-server@sha256:3e93a368a6198818e1b7c5525324c8130c5ed8d7b4cb499a47d4bec45ef0347b
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-pgpool2@sha256:c878935ffc796c405d811a6e8a9194aaaaa8652ea7d7b879c82919ce09794b35
        - quay.io/fujitsu/fujitsu-enterprise-postgres-cronjob@sha256:7582964eef14f89f2ca9046298357afd0132fb236f30f44f84cee14692227b0c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-restore@sha256:a81a0ce926572b75aa3151e2d2cc45a0ca411cc0158d69b89247ef70418f3740
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-server@sha256:1c5823afc7e77d46940214693c12c0ad6ffd6c7d9a156e26399ab55cd5ea1dfa
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-restore@sha256:17301c2191c0d75be82d12c5a35c0f74b8d65304bcfb6d4ab312a10ffa286eda
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-restore@sha256:a6e90dd77634033f2009942b635078cceb0f4c6b15b031624c5dce820713b14c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-server@sha256:5c65f8e76f70341c512100760de6792a880ad0d6eebc9e68154b5bbd1332a896
        - quay.io/fujitsu/fujitsu-enterprise-postgres-operator@sha256:b7dd55c0bfda84e398cbfbb87345fcec218cb866e286e18a111330321ab6ad54
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-utils@sha256:b7ae535f5e5fa5eb51c6f4cdb406c7b7a9dabe8573939b0431dada6f7d27b87e
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentd@sha256:3c5534b62695867c5973c9106fc46895ed73a4b9136ad83b2586bf7e71c427d0
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-pgpool2@sha256:38aac8108dd3590abcfd854a6ec2c025a68d2170a1ba68d359b45fea6c83cec1
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-backup@sha256:ac268e2af652bdabb8ee5d764a2b335dc73c48b9022eee8c9f3cd422227a4d98
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-upgrade@sha256:c0e44447349f02eb9f6105001136402f849d876a288d61997c8c05625facff45
        version: 7.3.0
      entries:
      - name: fujitsu-enterprise-postgres-operator.v7.3.0
        version: 7.3.0
      - name: fujitsu-enterprise-postgres-operator.v7.2.7
        version: 7.2.7
      name: stable
    defaultChannel: stable
    packageName: fujitsu-enterprise-postgres-operator
    provider:
      name: Fujitsu
      url: https://www.postgresql.fastware.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-block-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12
    - currentCSV: ibm-block-csi-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "k8s.gcr.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
        version: 1.12.0
      entries:
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.0
    - currentCSV: ibm-block-csi-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
        version: 1.12.1
      entries:
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.1
    - currentCSV: ibm-block-csi-operator.v1.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:af9215670675fe9c4fa2dac7a14d547c249f3aa0c0e72f1e5390d237e411735b
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:af9215670675fe9c4fa2dac7a14d547c249f3aa0c0e72f1e5390d237e411735b
        version: 1.12.2
      entries:
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.2
    - currentCSV: ibm-block-csi-operator.v1.12.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:6e7e13d259f4d743524ebcd546af5c73d1a3ba90edaf4b5edd2005a3f8adcdfe
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:6e7e13d259f4d743524ebcd546af5c73d1a3ba90edaf4b5edd2005a3f8adcdfe
        version: 1.12.3
      entries:
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.3
    - currentCSV: ibm-block-csi-operator.v1.12.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:42728e7fd1d2302511ab08f51335ab188eef125870fa27f322df83ebb8e1f447
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:42728e7fd1d2302511ab08f51335ab188eef125870fa27f322df83ebb8e1f447
        version: 1.12.4
      entries:
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.4
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.5
    defaultChannel: stable
    packageName: ibm-block-csi-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cli-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cli-manager-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "CliManager",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-cli-manager-operator"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed",
                  "operatorLogLevel": "Normal"
                }
              }
            ]
          capabilities: Basic Install
          categories: Openshift Optional
          createdAt: "2024-06-04"
          description: "The CLI Manager is a comprehensive tool designed to simplify
            the management of \nOpenShift CLI plugins within the OpenShift environment.
            \nModeled after the popular krew plugin manager, it offers seamless integration,
            \neasy installation, and efficient handling of a wide array of plugins,
            \nenhancing your OpenShift command-line experience.\n"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          repository: https://github.com/openshift/cli-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: CLI Manager
            kind: CliManager
            name: climanagers.operator.openshift.io
            version: v1
          - displayName: CLI Plugin
            kind: Plugin
            name: plugins.config.openshift.io
            version: v1alpha1
        description: "The CLI Manager is a comprehensive tool designed to simplify
          the management of \nOpenShift CLI plugins within the OpenShift environment.
          \nModeled after the popular krew plugin manager, it offers seamless integration,
          \neasy installation, and efficient handling of a wide array of plugins,
          \nenhancing your OpenShift command-line experience.\n"
        displayName: CLI Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cli-manager-operator
        - cli-manager
        links:
        - name: Source Code
          url: https://github.com/openshift/cli-manager
        - name: Source Code
          url: https://github.com/openshift/cli-manager-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/cli-manager-operator/cli-manager-rhel9-operator@sha256:31973d7f43d052ddfb9554dd623e37fb541a8c1f08ea981558eaf4bb5ab22d5a
        - registry.redhat.io/cli-manager-operator/cli-manager-rhel9@sha256:07853bc33a0ce783d8d45c08c161e454a3abe09a8a8dadeca70c3a199466b58b
        version: 0.1.0
      entries:
      - name: cli-manager-operator.v0.1.0
        version: 0.1.0
      name: tech-preview
    defaultChannel: tech-preview
    packageName: cli-manager
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cryostat-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cryostat-operator.v4.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.cryostat.io/v1beta2",
                "kind": "Cryostat",
                "metadata": {
                  "name": "cryostat-sample"
                },
                "spec": {
                  "enableCertManager": true,
                  "eventTemplates": [],
                  "networkPolicies": {},
                  "reportOptions": {
                    "replicas": 0
                  },
                  "trustedCertSecrets": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Developer Tools
          containerImage: registry.redhat.io/cryostat/cryostat-rhel9-operator@sha256:867a649b54ad144295c7aceda6b90631841aa31a62062df62d4980313723f359
          createdAt: "2025-09-30T15:10:08Z"
          description: JVM monitoring and profiling tool
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.0 <4.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "operator.cryostat.io/v1beta2",
              "kind": "Cryostat",
              "metadata": {
                "name": "cryostat-sample"
              },
              "spec": {
                "enableCertManager": true,
                "reportOptions": {
                  "replicas": 0
                }
              }
            }
          operators.openshift.io/valid-subscription: '["Red Hat build of OpenJDK"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/cryostatio/cryostat-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cryostat allows you to install Cryostat for a single namespace,
              or multiple namespaces. It contains configuration options for controlling
              the Deployment of the Cryostat application and its related components.
              A Cryostat instance must be created to instruct the operator to deploy
              the Cryostat application.
            displayName: Cryostat
            kind: Cryostat
            name: cryostats.operator.cryostat.io
            version: v1beta2
          - description: Cryostat allows you to install Cryostat for a single namespace.
              It contains configuration options for controlling the Deployment of
              the Cryostat application and its related components. A Cryostat instance
              must be created to instruct the operator to deploy the Cryostat application.
            displayName: Cryostat
            kind: Cryostat
            name: cryostats.operator.cryostat.io
            version: v1beta1
        description: |
          The Red Hat build of Cryostat provides a cloud-based solution for interacting with the JDK Flight Recorder already present in OpenJDK 11+ JVMs. With Cryostat, users can remotely start, stop, retrieve, and even analyze JFR event data, providing the capability to easily take advantage of Flight Recorder's extremely low runtime cost and overhead and the flexibility to monitor applications and analyze recording data without transferring data outside of the cluster the application runs within.
          ##Prerequisites
          Cryostat requires [cert-manager](https://cert-manager.io/) to run. If not already installed in your cluster, please [install](https://cert-manager.io/docs/installation/) it using your preferred method.
          When running on Kubernetes, the operator also requires [Ingress configurations](https://github.com/cryostatio/cryostat-operator/blob/v4.0.0/docs/config.md#network-options) for each of its services to make them available outside of the cluster. The user is responsible for providing the hostnames for each Ingress.
        displayName: Red Hat build of Cryostat
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flightrecorder
        - java
        - jdk
        - jfr
        - jmc
        - missioncontrol
        - monitoring
        - profiling
        - diagnostic
        links:
        - name: Upstream Project
          url: https://github.com/cryostatio/cryostat
        - name: Website
          url: https://cryostat.io/
        maintainers:
        - email: cryostat-development@googlegroups.com
          name: The Cryostat Authors
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi9/nginx-124@sha256:7c1d7172880a3649591a0dc1ceb670c133ef88eed3d8192d8a2e80e3878d04c3
        - registry.redhat.io/cryostat/cryostat-reports-rhel9@sha256:5535a539b185fa2781621741ccd7f10a6b9947adc6170a95e930f2d9d7a9f608
        - registry.redhat.io/insights-runtimes-tech-preview/runtimes-inventory-rhel8-operator@sha256:58e01d11c3d30abc5b0ee3795511bd04e39aea21e2853b4f4cc6d1c8a2bd7bb0
        - registry.redhat.io/cryostat/cryostat-rhel9@sha256:fadae8713daf82c329573191381f8f9e4ebf3fc852ac5f1a68ce0ea1da376469
        - registry.redhat.io/cryostat/jfr-datasource-rhel9@sha256:1e5f29aa096d783fd0a18b3ba31c391dc203761ba3f4af8e5aeeef1f7cc86cee
        - registry.redhat.io/cryostat/cryostat-db-rhel9@sha256:23d61204ecaf037cf17bf9a5cc4428461a60fecce077884d9458802fb071daab
        - registry.redhat.io/cryostat/cryostat-grafana-dashboard-rhel9@sha256:5715e897f9a8969aedd0066b2ba597264495b4c85e4ae8ae7d8871f93d13f4c1
        - registry.redhat.io/cryostat/cryostat-agent-init-rhel9@sha256:273d6c3d3194a4e9ec50c1d8e147cd9bc6d277d8d66878d885a18528b9ba7dc6
        - registry.redhat.io/cryostat/cryostat-rhel9-operator@sha256:867a649b54ad144295c7aceda6b90631841aa31a62062df62d4980313723f359
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        - registry.redhat.io/cryostat/cryostat-ose-oauth-proxy-rhel9@sha256:d38ce4dcac5b073b001d886f0494e0fc9b750f8bd68b855fc103767a1ed8b7df
        - registry.redhat.io/cryostat/cryostat-storage-rhel9@sha256:f642ca8a0ca0aa130462c0f423ce4c775cb296f8a40c9270e5369176e87c7ee6
        - registry.redhat.io/cryostat/cryostat-openshift-console-plugin-rhel9@sha256:ca748f96414ad648eb1e3953aeeeb828b187f3c5c30e014513adc3619a698683
        version: 4.0.3
      entries:
      - name: cryostat-operator.v4.0.3
        version: 4.0.3
      - name: cryostat-operator.v4.0.2-4
        version: 4.0.2+4
      - name: cryostat-operator.v4.0.2
        version: 4.0.2
      - name: cryostat-operator.v4.0.1
        version: 4.0.1
      - name: cryostat-operator.v4.0.1-5
        version: 4.0.1+5
      - name: cryostat-operator.v4.0.1-4
        version: 4.0.1+4
      - name: cryostat-operator.v4.0.1-3
        version: 4.0.1+3
      - name: cryostat-operator.v4.0.1-2
        version: 4.0.1+2
      - name: cryostat-operator.v4.0.0-11
        version: 4.0.0+11
      - name: cryostat-operator.v4.0.0
        version: 4.0.0
      - name: cryostat-operator.v4.0.0-9
        version: 4.0.0+9
      - name: cryostat-operator.v4.0.0-10
        version: 4.0.0+10
      - name: cryostat-operator.v3.0.1-7
        version: 3.0.1+7
      - name: cryostat-operator.v3.0.1-9
        version: 3.0.1+9
      - name: cryostat-operator.v3.0.1-8
        version: 3.0.1+8
      - name: cryostat-operator.v3.0.1
        version: 3.0.1
      - name: cryostat-operator.v3.0.1-6
        version: 3.0.1+6
      - name: cryostat-operator.v3.0.1-5
        version: 3.0.1+5
      - name: cryostat-operator.v3.0.1-3
        version: 3.0.1+3
      - name: cryostat-operator.v3.0.1-13
        version: 3.0.1+13
      - name: cryostat-operator.v3.0.1-12
        version: 3.0.1+12
      - name: cryostat-operator.v3.0.1-11
        version: 3.0.1+11
      - name: cryostat-operator.v3.0.1-10
        version: 3.0.1+10
      - name: cryostat-operator.v3.0.0-12
        version: 3.0.0+12
      - name: cryostat-operator.v3.0.0
        version: 3.0.0
      - name: cryostat-operator.v3.0.0-6
        version: 3.0.0+6
      - name: cryostat-operator.v3.0.0-8
        version: 3.0.0+8
      - name: cryostat-operator.v2.4.0
        version: 2.4.0
      - name: cryostat-operator.v2.4.0-8
        version: 2.4.0+8
      - name: cryostat-operator.v2.4.0-7
        version: 2.4.0+7
      - name: cryostat-operator.v2.4.0-6
        version: 2.4.0+6
      - name: cryostat-operator.v2.4.0-4
        version: 2.4.0+4
      - name: cryostat-operator.v2.4.0-3
        version: 2.4.0+3
      - name: cryostat-operator.v2.3.1-6
        version: 2.3.1+6
      - name: cryostat-operator.v2.3.1-8
        version: 2.3.1+8
      - name: cryostat-operator.v2.3.1-9
        version: 2.3.1+9
      - name: cryostat-operator.v2.3.1
        version: 2.3.1
      - name: cryostat-operator.v2.3.0
        version: 2.3.0
      - name: cryostat-operator.v2.3.0-6
        version: 2.3.0+6
      - name: cryostat-operator.v2.2.1-14
        version: 2.2.1+14
      - name: cryostat-operator.v2.2.1-8
        version: 2.2.1+8
      - name: cryostat-operator.v2.2.1-5
        version: 2.2.1+5
      - name: cryostat-operator.v2.2.1-13
        version: 2.2.1+13
      - name: cryostat-operator.v2.2.1-12
        version: 2.2.1+12
      - name: cryostat-operator.v2.2.1-10
        version: 2.2.1+10
      - name: cryostat-operator.v2.2.1
        version: 2.2.1
      - name: cryostat-operator.v2.2.0
        version: 2.2.0
      - name: cryostat-operator.v2.1.1-2.1659469835.p
        version: 2.1.1+2.1659469835.p
      - name: cryostat-operator.v2.1.1-2.1655334614.p
        version: 2.1.1+2.1655334614.p
      - name: cryostat-operator.v2.1.1
        version: 2.1.1
      - name: cryostat-operator.v2.1.0
        version: 2.1.0
      name: stable
    defaultChannel: stable
    packageName: cryostat-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: community-windows-machine-config-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: community-windows-machine-config-operator.v6.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-windows/community-windows-machine-config-operator:community-4.11-6c3601f
          createdAt: "2022-08-31T18:57:22Z"
          description: An operator that enables Windows container workloads on OCP
          olm.skipRange: '>=5.0.0 <6.0.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ## WARNING
          Community distribution of the Windows Machine Config Operator.
          This is a preview build, and is not meant for production workloads.
          Issues with this distribution of WMCO can be opened against the [WMCO repository](https://github.com/openshift/windows-machine-config-operator).
          Please read through the [troubleshooting doc](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/docs/TROUBLESHOOTING.md)
          before opening an issue.
          Please ensure that when installing this operator the starting CSV you subscribe to is supported on the
          version of OKD/OCP you are using. This CSV is meant for OKD/OCP community-4.11.
          ## Documentation
          ### Introduction
          The Windows Machine Config Operator configures Windows instances into nodes, enabling Windows container workloads
          to be ran within OKD/OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/latest/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/README.md#configuring-byoh-bring-your-own-host-windows-instances).
          The operator will do all the necessary steps to configure the instance so that it can join the cluster as a worker node.
          ### Pre-requisites
          - [Cluster and OS pre-requisites](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/docs/wmco-prerequisites.md)
          ### Usage
          Please see the usage section of [README.md](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/README.md#usage).
          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.
        displayName: Community Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: preview
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-windows/community-windows-machine-config-operator:community-4.11-6c3601f
        version: 6.0.0
      entries:
      - name: community-windows-machine-config-operator.v6.0.0
        version: 6.0.0
      name: preview
    defaultChannel: preview
    packageName: community-windows-machine-config-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom
      provider-url: https://www.broadcom.com/
    name: layer7-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: layer7-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "security.brcmlabs.com/v1",
                "kind": "Gateway",
                "metadata": {
                  "name": "ssg"
                },
                "spec": {
                  "app": {
                    "image": "docker.io/caapim/gateway:11.1.2",
                    "management": {
                      "cluster": {
                        "hostname": "gateway.brcmlabs.com",
                        "password": "7layer"
                      },
                      "password": "7layer",
                      "username": "admin"
                    },
                    "replicas": 1,
                    "service": {
                      "ports": [
                        {
                          "name": "https",
                          "port": 8443,
                          "protocol": "TCP",
                          "targetPort": 8443
                        },
                        {
                          "name": "management",
                          "port": 9443,
                          "protocol": "TCP",
                          "targetPort": 9443
                        }
                      ],
                      "type": "LoadBalancer"
                    }
                  },
                  "license": {
                    "accept": false,
                    "secretName": "gateway-license"
                  },
                  "version": "11.1.2"
                },
                "status": {}
              },
              {
                "apiVersion": "security.brcmlabs.com/v1",
                "kind": "Repository",
                "metadata": {
                  "name": "my-repository"
                },
                "spec": {
                  "auth": {},
                  "branch": "main",
                  "enabled": true,
                  "endpoint": "https://github.com/\u003cusername\u003e/\u003crepository\u003e",
                  "type": "git"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7Api",
                "metadata": {
                  "name": "l7api-sample"
                },
                "spec": {
                  "deploymentTags": [
                    "ssg",
                    "ssg1"
                  ],
                  "graphmanBundle": "jsongz",
                  "portalPublished": false,
                  "serviceUrl": "/v1/api"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7Portal",
                "metadata": {
                  "name": "dev-portal"
                },
                "spec": {
                  "auth": {},
                  "deploymentTags": [
                    "ssg"
                  ],
                  "enabled": true,
                  "endpoint": "dev-ssg.brcmlabs.com",
                  "enrollmentBundle": "",
                  "portalTenant": "portal"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7StateStore",
                "metadata": {
                  "name": "state-store-example"
                },
                "spec": {
                  "redis": {
                    "database": 0,
                    "groupName": "l7gw",
                    "masterPassword": "7layer",
                    "standalone": {
                      "host": "redis-standalone",
                      "port": 6379
                    },
                    "storeId": "test",
                    "type": "standalone"
                  },
                  "type": "redis"
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: docker.io/caapim/layer7-operator:v1.2.1
          createdAt: "2025-07-04T17:22:57Z"
          description: The Layer7 Operator covers all aspects of deploying, maintaining
            and upgrading Layer7 API Gateways in Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: layer7-operator-system
          operators.openshift.io/valid-subscription: Layer7 (Broadcom) API Gateway
            license
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/caapim/layer7-operator
          support: Broadcom Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gateway is the Schema for the Gateway Custom Resource
            displayName: Gateway
            kind: Gateway
            name: gateways.security.brcmlabs.com
            version: v1
          - description: L7Api is the Schema for the l7apis API
            displayName: L7 Api
            kind: L7Api
            name: l7apis.security.brcmlabs.com
            version: v1alpha1
          - description: L7Portal is the Schema for the l7portals API
            displayName: L7 Portal
            kind: L7Portal
            name: l7portals.security.brcmlabs.com
            version: v1alpha1
          - description: L7StateStore is the Schema for the l7statestores API
            displayName: L7 State Store
            kind: L7StateStore
            name: l7statestores.security.brcmlabs.com
            version: v1alpha1
          - description: Repository is the Schema for the repositories API
            displayName: Repository
            kind: Repository
            name: repositories.security.brcmlabs.com
            version: v1
        description: |
          The [Layer7 Operator](https://github.com/CAAPIM/layer7-operator) covers all aspects of deploying, maintaining and upgrading Layer7 API Gateways in Kubernetes. The Layer7 Operator is exclusive to Kubernetes and utilizes the Container Gateway with specific emphasis on running in Ephemeral mode (without an external MySQL policy database).

          ## Getting Started
          This [Getting Started](https://github.com/CAAPIM/layer7-operator/wiki/Getting-Started) guide covers a simple introduction to the Layer7 Operator. You can find more documentation [here](https://github.com/CAAPIM/layer7-operator/wiki).
        displayName: Layer7 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - layer7-operator
        links:
        - name: Layer7 Operator
          url: https://github.com/caapim/layer7-operator
        maintainers:
        - email: gary.vermeulen@broadcom.com
          name: Gary Vermeulen
        maturity: alpha
        provider:
          name: Broadcom
          url: https://www.broadcom.com/
        relatedImages:
        - docker.io/caapim/graphman-static-init:1.0.3
        - docker.io/caapim/gateway:11.1.2
        - docker.io/caapim/layer7-operator:v1.2.1
        version: 1.2.1
      entries:
      - name: layer7-operator.v1.2.1
        version: 1.2.1
      - name: layer7-operator.v1.2.0
        version: 1.2.0
      name: preview
    defaultChannel: preview
    packageName: layer7-operator
    provider:
      name: Broadcom
      url: https://www.broadcom.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: eG Innovations
      provider-url: https://www.eginnovations.com
    name: eginnovations-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eginnovations-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eginnovations.com/v1beta1",
                "kind": "EgUniversalAgent",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eginnovations-operator"
                  },
                  "name": "eguniversalagent",
                  "namespace": "egagent"
                },
                "spec": {
                  "agent": {
                    "env": [
                      {
                        "name": "EG_MANAGER",
                        "value": "Replace with eG Manager IP"
                      },
                      {
                        "name": "EG_MANAGER_PORT",
                        "value": "Replace with eG Manager Port"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "true"
          containerImage: docker.io/eginnovations/universal-agent-operator@sha256:b8e3e26dca1b94fabfd79496feb36f3f2de4434331ef9ca810125892a936dca7
          createdAt: "2025-09-11T20:08:00Z"
          description: eG Innovations provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with eG Universal Agent Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eginnovations/universal-agent-operator
          support: eG Innovations
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgUniversalAgent is the Schema for the eguniversalagents
              API
            displayName: Eg Universal Agent
            kind: EgUniversalAgent
            name: eguniversalagents.eginnovations.com
            version: v1beta1
        description: |
          eG Innovations’ converged application and infrastructure performance monitoring capabilities provide end-to-end visibility and correlated analytics for every layer, every tier of your IT landscape.

          The eG Enterprise Universal Agent Operator automatically configures a host agent on every Kubernetes worker node. The host agent auto-discovers the worker nodes and application containers running as Pods on each node and tracks their performance and utilization levels. In-depth monitoring of applications running on containers is also provided using the same host agent. No additional agents are required for the containers.

          ### Capabilities
            * Rolls out the egagent pod per node to monitor its pods and the node itself
            * Automatic BTM profiler injection using webhook based injection for deep application monitoring and transection tracing.

          ### Installation
            1. Create the namespace "egagent" for installing eG Agent on OpenShift Cluster.

               <br />``` $ oc create ns egagent ```
            2. Add "egagent" serviceaccount to scc( Security Context Constraints ) of OpenShift Cluster as privileged using below command.

               <br />``` $ oc adm policy add-scc-to-user privileged -z egagent -n egagent ```
            2. Install eG Innovations Universal Agent operator from [OpenShift Container Platform](https://www.openshift.com/) OperatorHub.
            3. Create the eG Universal Agent CRD (Custom Resource Definition)
               a. Click "View YAML Example" and copy the yaml contents and save it to (filename).yaml on your host.
               b. To apply the CRD run the following command

               <br />``` $ oc apply -f (filename).yaml ```

          ### Required Parameters
            * EG\_MANAGER, EG\_MANAGER\_PORT - provide the ip and port of your eG Manager Installation.

          ### Advanced Options
            * Disable Certificate Checking - disable any certificate validation that may interact poorly with proxies within your cluster
            * Image Override - use a copy of the eG Iniversal Agent container image from a registry other that Docker's or Red Hat's
            * Environment variables - define environment variables for the eG Universal Agent container of for APM profiler.
            * Namespace Selectors- select a subset of namesapces to enable instrumentation for applications.

          For complete list of supported parameters and the detailed documentation available in [eG Innovations Docs](https://docs.eginnovations.com)
        displayName: eG Innovations Universal Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Tracing
        links:
        - name: eG Innovations Kubernetes Monitoring Overview
          url: https://www.eginnovations.com/documentation/Kubernetes/Introduction-to-Monitoring-Kubernetes.htm
        maintainers:
        - email: devops@eginnovations.com
          name: eG Innovations
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: eG Innovations
          url: https://www.eginnovations.com
        relatedImages:
        - docker.io/eginnovations/universal-agent-operator@sha256:b8e3e26dca1b94fabfd79496feb36f3f2de4434331ef9ca810125892a936dca7
        version: 0.0.11
      entries:
      - name: eginnovations-operator.v0.0.11
        version: 0.0.11
      name: beta
    defaultChannel: beta
    packageName: eginnovations-operator
    provider:
      name: eG Innovations
      url: https://www.eginnovations.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-trino-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trino-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:904eb8ac055171c75c8c86e4923df85461b81e6d5d9269d9cddfb03fc615f9f4
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:904eb8ac055171c75c8c86e4923df85461b81e6d5d9269d9cddfb03fc615f9f4
        version: 24.11.1
      entries:
      - name: trino-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: trino-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b50b247ed8f0ad262ed59e8aafd099919e98156472e5663957f9922616140bd9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b50b247ed8f0ad262ed59e8aafd099919e98156472e5663957f9922616140bd9
        version: 25.3.0
      entries:
      - name: trino-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: trino-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
        version: 25.7.0
      entries:
      - name: trino-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: trino-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
        version: 25.7.0
      entries:
      - name: trino-operator.v25.7.0
        version: 25.7.0
      - name: trino-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-trino-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-kafka-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kafka-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:f76b5558088bead829be8018a89e3efc1bb7a84d40429a0a1df1c88552db169b
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:f76b5558088bead829be8018a89e3efc1bb7a84d40429a0a1df1c88552db169b
        version: 24.11.1
      entries:
      - name: kafka-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: kafka-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:72b390c2d382af0a9cbe2d9ae394fc61c62b9a3f6c7310cf5f325d0f17257f1f
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:72b390c2d382af0a9cbe2d9ae394fc61c62b9a3f6c7310cf5f325d0f17257f1f
        version: 25.3.0
      entries:
      - name: kafka-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: kafka-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
        version: 25.7.0
      entries:
      - name: kafka-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: kafka-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
        version: 25.7.0
      entries:
      - name: kafka-operator.v25.7.0
        version: 25.7.0
      - name: kafka-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-kafka-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: vfunction-operator
      app.kubernetes.io/managed-by: operator
      app.kubernetes.io/name: vfunction
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      name: vfunction-server-operator
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: vFunction
      provider-url: ""
    name: vfunction-server-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vfunction-server-operator.v2.2.469
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vfunction.com/v1",
                "kind": "VfunctionServer",
                "metadata": {
                  "name": "example-vfunction-server",
                  "namespace": "vfunction",
                  "labels": {
                      "name": "vfunction",
                      "app.kubernetes.io/name": "vfunction",
                      "app.kubernetes.io/instance": "vfunction-server",
                      "app.kubernetes.io/managed-by": "operator"
                  }
                },
                "spec": {
                  "admin": {
                    "email": "admin@mycompany.com",
                    "name": "Admin",
                    "password": "Password1!"
                  },
                  "host": "http://my.domain.com",
                  "org_name": "MyCompany",
                  "upgrade": "Daily",
                  "measurements": {
                    "auto_scaling": "No",
                    "min_number_of_services": "2",
                    "max_number_of_services": "10"
                  },
                  "smtp": {
                    "identity": "",
                    "password": "",
                    "url": "",
                    "user": ""
                  },
                  "tls": {
                    "crt": "",
                    "key": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, AI/Machine Learning
          certified: Red Hat certified
          containerImage: registry.marketplace.redhat.com/rhm/vfunction/vfunction-server-operator@sha256:60d1965460101d0fd90ecc9499d0a6d4c9684a54ada61a43a253d30661fab22f
          createdAt: "2021-08-03T19:08:54Z"
          description: vFunction is a cutting-edge code analysis, machine learning,
            and automation to boost your Java modernization projects.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/vfunction-server-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/vfunction-server-operator-rhmp/support?utm_source=openshift_console
          support: vFunction
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of vFunction running on your cluster.
            displayName: vFunction
            kind: VfunctionServer
            name: vfunctionservers.vfunction.com
            version: v1
        description: |2-

          ## Introduction

          [vFunction](https://www.vfunction.com/) is a cloud-native modernization platform that combines dynamic and static code analysis, machine learning, and automation to automatically identify and extract services from existing applications. vFunction is the only platform purpose-built for modernization of Java applications.

          The vFunction modernization process starts by learning the running monolithic application, and surfacing the interdependencies within it. Using AI, the platform analyzes and identifies services that can be separated from the application. This decomposition can present a range of micro, mini, or even macro services, depending on your application environment, each being an independently deployable and scalable application component.

          vFunction automates the extraction of these services, enabling you to modernize your monolith, quickly and easily.

          ## The vFunction platform

          The platform consists of 3 basic components; the server, the controller package, and a tools package. The server runs as an operand on an OpenShift environment. The controller package is installed on the machine that runs the monolithic application which can be either a Linux or a Windows machine, and the tools are run on a development machine, with access to the code of the monolithic application.

          The controller package consists of three elements: the vFunction agent, that collects data during the dynamic analysis phase; the vFunction Viper application, that performs static analysis on the binaries of the application; and the vFunction controller that handles all the communication between the agent, Viper, and the vFunction server.

          The vFunction agent is a mix of a Java and native agent, and needs to run on the JVM that is currently running your application. [Refer to the vFunction OOB Support Matrix document](https://drive.google.com/file/d/1ccq8LFab1FrYAimDUxwgjiCOdk4QuVzs/view) for a list of supported application servers and JVMs.

          This operator installs a vFunction server instance to be connected later with one or more vFunction controllers installed on your application machine(s).

          ## Before You Start

          The vFunction operator requires a Red Hat OpenShift Kubernetes Application Platform 4.x.

          The cluster recommended configuration (which is the same as the minimal one) should consist of:

          - *1 master node*

          - *1 worker node*

          - *Storage capacity enough for dynamically provision 2 PVs of 50G each*

          - *A default StorageClass configured*

          vFunction doesn’t provide any encryption mechanism for data stored on any attached storage. If required, encryption can be achieved by the client by using its own external means on the storage itself.

          vFunction supports backup of all critical data as part of the application. Nevertheless, for complete backup of entire data it is recommended that the client apply its own storage based backup mechanism.

          ## Install Prerequisites

          ###Cluster:
          The operator was certified and tested on OCP 4.6 and 4.7.
          If you encounter any issue with other OCP versions or cloud providers please contact info@vfunction.com.

          ###Storage:
          The operator creates two new PersistentVolumeClaims (PVCs) during installation time. For their creation, it relies on the default StorageClass to dynamically provision two PersistentVolumes (PVs). The system must have a default StorageClass in place.

          Both PersistentVolumes are accessed with ReadWriteOnce mode.

          ###Project:
          The operator should be installed in a new and dedicated project (namespace). If you intend to install more than one vFunction operands in the same cluster, please use different projects for each one. Installation of more than one operand on the same cluster will allow scaling by load balancing multiple applications to different vFunction servers.

          ## Installation instructions

          1. Prepare the YAML file to use in the installation:

              a. Copy the following YAML template into a text editor:

              ```yaml
                  apiVersion: vfunction.com/v1
                  kind: VfunctionServer
                  metadata:
                    name: vfunction
                    namespace: vfunction
                    labels:
                      name: vfunction
                      app.kubernetes.io/name: vfunction
                      app.kubernetes.io/instance: vfunction-server
                      app.kubernetes.io/managed-by: operator
                  spec:
                    host: "http://my.domain.com"
                    org_name: "MyCompany"
                    upgrade: "Daily"
                    admin:
                      email: "admin@mycompany.com"
                      name: "Admin"
                      password: "Password1!"
                    smtp:
                      password: ""
                      url: ""
                      identity: ""
                      user: ""
                    measurements:
                      auto_scaling: "No"
                      min_number_of_services: "2"
                      max_number_of_services: "10"
                    tls:
                      crt: |
                        -----BEGIN CERTIFICATE-----
                        ...
                        -----END CERTIFICATE-----
                      key: |
                        -----BEGIN PRIVATE KEY-----
                        ...
                        -----END PRIVATE KEY-----
              ```
              b. Customize the template fields by replacing with your information:

              - **host**: Enter the FQDN for accessing the vFunction dashboard.
                          Make sure that you use a domain name and not an IP address. Writing "https://" will mandate TLS while writing "http://" will expose the server through HTTP.
              - **org_name**: Enter your organization name.
              - **upgrade**: Choose auto-upgrade mode
                  - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically (default).
                  - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
                  - **Never**: The automatic upgrading mechanism is off.
                  - **Maintenance**: Put the operator in a maintenance mode.

              - **admin.email**: Enter the email address of your administrator.
              - **admin.name**: Enter the name of your administrator.
              - **admin.password**: Enter the password you want to use for the vFunction administrator.
                                    The password should be at least 8 characters long, and consist of at least one lowercase letter, at least one uppercase letter, at least one number, and at least one special character.
              - **smtp.user**: Enter the email address for a designated user for the SMTP server (optional).
              - **smtp.password**: Enter this user’s password (optional).
              - **smtp.identity**: Enter this user’s password (optional).
              - **smtp.url**: Enter the SMTP server URL (optional).
              - **tls.crt**: If your host FQDN starts with “https”, paste in the certifications you have for using the TLS connection.
              - **tls.key**: If your host FQDN starts with “https”, enter the key you have for using the TLS connection.
              - **measurements.auto_scaling**: Indicates if the measurement services auto scalling is active.
              - **measurements.min_num_of_services**: Minimum Number of measurement service pods.
              - **measurements.max_num_of_services**: Maximum Number of measurement service pods.
              - **measurements.S3**: Save all measurements data to S3, instead of local PV (optional).
                  - **measurements.S3.bucket**: Measurements S3 bucket name.
                  - **measurements.S3.key**: Measurements S3 key.
                  - **measurements.S3.secret**: Measurements S3 secret.
                  - **measurements.S3.region**: Measurements S3 region.
              - **backup.mysql.S3**: Backup MySQL to S3 (optional).
                  - **backup.mysql.S3.bucket**: MySQL backup S3 bucket name.
                  - **backup.mysql.S3.key**: MySQL backup S3 key.
                  - **backup.mysql.S3.secret**: MySQL backup S3 secret.
                  - **backup.mysql.S3.region**: MySQL backup S3 region.

          1. In the Installed Operators window choose and click on vFunction Operator and then Create Instance.
          1. In the Create vFunctionServer window, paste the YAML file that you created in step 1.
          1. Click Create.
          1. You can now click on your new created vFunction operand in order to see its details and installation progress.

          ## Verifying the installation

          Check that the operand installed successfully:

          1. Check that the "Operand State" property shows “Working”.
          1. Check that the "Successfully Installed" property shows "Yes".

          If "Operand State" shows "Failed", it indicates the operand wasn’t installed correctly. Check the events and logs for all pods (operator and image containers) for any issues. In the event that you cannot troubleshoot, contact vFunction support.

          ## After installation

          The vFunction site is now accessible via the newly created vFunction application custom address, for example, my.domain.com. There are two ways you can access the vFunction dashboard.

          ### Access using the router canonical hostname

          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address, and to the ”vfunction” subdomain of the  OpenShift canonical router hostname as the alias.
          For example, ***my.domain.com.   CNAME   vfunction.apps.ocp4.my-openshift.com.***
          2. Find your cluster Router Canonical Hostname address in the newly created vfunction-route-xxx under your vFunction OpenShift project -> **Networking** -> **Routes** -> **vfunction-route-xxx** route -> **Router Canonical Hostname** field.
          You can now access the dashboard using your defined "host" spec property (as above).

          ### Access using the nginx service location

          You may use this access method if your OpenShift is installed on a provider that supports exposing LoadBalancer-type services.
          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address and to the vfunction-nginx-xxx service location.
          For example, ***my.domain.com.   CNAME   a05951ed7cdf-1394239323.us-east-1.elb.amazonaws.com.***
          1. Point your custom domain to the vfunction-nginx-xxx service's external IP location, which you can find in the OpenShift project > **Networking** -> **Services** -> **vfunction-nginx-xxx** service > **Service Address** > **Location** field.

          ## Upgrade and Rollback

          The vFunction operator includes a built-in auto-upgrade mechanism.

          You can choose one of 4 upgrading modes:
          - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically (default).
          - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
          - **Never**: The automatic upgrading mechanism is off.
          - **Maintenance**: Put the operator in a maintenance mode.

          For an on-demand upgrade,  you can change anytime the upgrading mode from “Never” to “Always”, wait for the operand to be upgraded, and change again to “Never”.

          During the upgrade, the "Operand State" property will change to “Upgrading...” and back again to “Working” after a successful upgrade.

          In case of a failure upgrading the operand, an automatic rollback to the last working version will occur. The failed version will be marked as defective and the operator will not try to upgrade to it again.

          Choosing the "Maintenance" option will tell the operator to take down all pods, so you can safely fix storage issues, return volumes from snapshots, etc.
          After finishing the maintenance, return the upgrade mode to your original desired policy.
        displayName: vFunction Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - vfunction
        - analyzing
        - services
        links:
        - name: vFunction
          url: https://vfunction.com
        - name: Terms
          url: https://www.vfunction.com/eula/
        maintainers:
        - email: raviv@vfunction.com
          name: Raviv Halivni
        maturity: alpha
        provider:
          name: vFunction
        relatedImages:
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-organizations@sha256:bcc2c3d6222af95e614318daea3108fffc6bc77aa17b23031cce296a3811b79f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-users@sha256:e47cc051953976a0e075e8cfb457cd4749e9b9e4394f2fbecab2458c2ade92ca
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-mysql@sha256:a6e06ea08fffa33b709d4d1240587bb3ab5b4cc6a28f493a227555e88e645798
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-nginx@sha256:435d1489aaf29a811081ddb51f40ae692e5833f42c31dbced6bc189c0ddc40ab
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-storage@sha256:b778cb9c724a59e51b4cb9088047be471cdf338f1138c31b81f79e50a491203f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-idp@sha256:21ca95a0209f3c557a45658452d154777f928539a93129148470562c79e68d70
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-measurements@sha256:35ffbee4b0eead58706b23cf5a702b1d55d2a7913cd0019cefbff94957d4e5d6
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-server-operator@sha256:60d1965460101d0fd90ecc9499d0a6d4c9684a54ada61a43a253d30661fab22f
        version: 2.2.469
      entries:
      - name: vfunction-server-operator.v2.2.469
        version: 2.2.469
      name: stable
    defaultChannel: stable
    packageName: vfunction-server-operator-rhmp
    provider:
      name: vFunction
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-xtradb-cluster-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.17.0
        version: 1.17.0
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.16.1
        version: 1.16.1
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.16.0
        version: 1.16.0
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.15.1
        version: 1.15.1
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.15.0
        version: 1.15.0
      name: stable
    - currentCSV: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        version: 1.18.0-cw
      entries:
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0-cw
        version: 1.18.0-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.17.0-cw
        version: 1.17.0-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.16.1-cw
        version: 1.16.1-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.16.0-cw
        version: 1.16.0-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.15.0-cw
        version: 1.15.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator-certified-rhmp
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: JWS team
      provider-url: ""
    name: jws-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jws-operator.v2.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "web.servers.org/v1alpha1",
                "kind": "WebServer",
                "metadata": {
                  "name": "webserver-example"
                },
                "spec": {
                  "applicationName": "webapp",
                  "replicas": 1,
                  "useInsightsClient": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/jboss-webserver-5/jws5-rhel8-operator@sha256:ea042277bdcdb70b2dc32083ea868efd4bf474d5a8ec0f9b47f5bea35aaeae1e
          createdAt: "2025-09-12T08:31:52.558719Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat JBoss Web Server",
            "Red Hat Runtimes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/web-servers/jws-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Web Server is the schema for the webservers API
            displayName: Web Server
            kind: WebServer
            name: webservers.web.servers.org
            version: v1alpha1
        description: WebServer operator can manage JBoss Web Server images.
        displayName: WebServer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WebServer
        - JBoss
        links:
        - name: Jws Operator
          url: https://jws-operator.domain
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: JWS team
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:021ff671b0fb72eed635b05489ec86342fcb7ada3ba0353591fdb9328871d067
        - registry.redhat.io/jboss-webserver-5/jws5-rhel8-operator@sha256:ea042277bdcdb70b2dc32083ea868efd4bf474d5a8ec0f9b47f5bea35aaeae1e
        - registry.redhat.io/insights-runtimes-tech-preview/runtimes-inventory-rhel8-operator@sha256:58e01d11c3d30abc5b0ee3795511bd04e39aea21e2853b4f4cc6d1c8a2bd7bb0
        version: 2.2.0
      entries:
      - name: jws-operator.v2.2.0
        version: 2.2.0
      - name: jws-operator.v2.1.3
        version: 2.1.3
      - name: jws-operator.v2.1.2
        version: 2.1.2
      - name: jws-operator.v2.1.1
        version: 2.1.1
      - name: jws-operator.v2.1.0
        version: 2.1.0
      - name: jws-operator.v2.0.16
        version: 2.0.16
      - name: jws-operator.v2.0.15
        version: 2.0.15
      - name: jws-operator.v2.0.14
        version: 2.0.14
      - name: jws-operator.v2.0.13
        version: 2.0.13
      - name: jws-operator.v2.0.12
        version: 2.0.12
      - name: jws-operator.v2.0.11
        version: 2.0.11
      - name: jws-operator.v2.0.10
        version: 2.0.10
      - name: jws-operator.v2.0.9
        version: 2.0.9
      - name: jws-operator.v2.0.8
        version: 2.0.8
      - name: jws-operator.v2.0.7
        version: 2.0.7
      - name: jws-operator.v2.0.6
        version: 2.0.6
      - name: jws-operator.v2.0.5
        version: 2.0.5
      - name: jws-operator.v2.0.4
        version: 2.0.4
      - name: jws-operator.v2.0.3
        version: 2.0.3
      - name: jws-operator.v2.0.2
        version: 2.0.2
      - name: jws-operator.v2.0.1
        version: 2.0.1
      name: alpha
    defaultChannel: alpha
    packageName: jws-operator
    provider:
      name: JWS team
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Windup
      provider-url: ""
    name: windup-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: windup-operator.v0.0.25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
             {
               "apiVersion": "windup.jboss.org/v1",
               "kind": "Windup",
               "metadata": {
                 "name": "windup",
                 "labels": {
                   "application": "windup"
                 }
               },
               "spec": {}
             }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/windupeng/windup-operator-native:0.0.25
          createdAt: 01/22/24
          description: Windup is an analysis tool that supports the modernization
            and migration of Java applications.
          imagePullPolicy: Always
          operatorframework.io/suggested-namespace: windup
          repository: https://github.com/windup/windup-operator/
          support: https://issues.redhat.com/projects/WINDUP
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "Create a Windup Web Console application.  \n\nTo access
              the Windup Web Console application from the \"Developer\" perspective,
              please go to Topology and click on the \"Open URL\" icon on the web-console
              pod.  \n\nAlternatively from the \"Administrator\" perspective, please
              go to Networking->Routes and click on the Location hyperlink.  \n\nKnown
              issue  \nIf you want to customize the Windup Web Console instance's
              parameters and you can not see any in the `Form view`, please switch
              to the `YAML view` and change them as needed.\n"
            displayName: Windup
            kind: Windup
            name: windups.windup.jboss.org
            version: v1
        description: |
          Windup is a web console application that supports large-scale Java application modernization and migration projects across a broad range of transformations and use cases.

          It analyzes application code, supports effort estimation, accelerates code migration, and enables users to move applications to containers.

          For more information please refer to the [https://windup.github.io/](https://windup.github.io/) page.
        displayName: Windup Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - modernization
        - windup
        links:
        - name: Website
          url: https://windup.github.io
        - name: GitHub
          url: https://github.com/windup/windup-operator
        maintainers:
        - email: migrate@redhat.com
          name: Windup
        maturity: alpha
        provider:
          name: Windup
        relatedImages:
        - quay.io/windupeng/windup-operator-native:0.0.25
        version: 0.0.25
      entries:
      - name: windup-operator.v0.0.25
        version: 0.0.25
      - name: windup-operator.v0.0.24
        version: 0.0.24
      - name: windup-operator.v0.0.23
        version: 0.0.23
      - name: windup-operator.v0.0.22
        version: 0.0.22
      - name: windup-operator.v0.0.21
        version: 0.0.21
      - name: windup-operator.v0.0.20
        version: 0.0.20
      - name: windup-operator.v0.0.19
        version: 0.0.19
      - name: windup-operator.v0.0.18
        version: 0.0.18
      - name: windup-operator.v0.0.17
        version: 0.0.17
      - name: windup-operator.v0.0.16
        version: 0.0.16
      - name: windup-operator.v0.0.15
        version: 0.0.15
      - name: windup-operator.v0.0.14
        version: 0.0.14
      - name: windup-operator.v0.0.13
        version: 0.0.13
      - name: windup-operator.v0.0.12
        version: 0.0.12
      - name: windup-operator.v0.0.11
        version: 0.0.11
      name: alpha
    defaultChannel: alpha
    packageName: windup-operator
    provider:
      name: Windup
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stolostron
      provider-url: ""
    name: cluster-aas-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-aas-operator.v0.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplate",
                "metadata": {
                  "name": "clustertemplate-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplateInstance",
                "metadata": {
                  "name": "clustertemplateinstance-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplateQuota",
                "metadata": {
                  "name": "clustertemplatequota-sample"
                },
                "spec": {}
              }
            ]
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/cluster-templates-operator:2.11.0-0c3131154065397d37e3b66ae635d7ea88650bf9
          createdAt: "2024-06-06T19:34:44Z"
          description: Easily install fully configured clusters with guard-rails.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/cluster-templates-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents instance of a cluster
            displayName: Cluster template instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.clustertemplate.openshift.io
            version: v1alpha1
          - description: Defines which ClusterTemplates can be used in a given namespace
            displayName: Cluster template quota
            kind: ClusterTemplateQuota
            name: clustertemplatequotas.clustertemplate.openshift.io
            version: v1alpha1
          - description: Template of a cluster - both installation and post-install
              setup are defined as ArgoCD application spec. Any application source
              is supported - typically a Helm chart
            displayName: Cluster template
            kind: ClusterTemplate
            name: clustertemplates.clustertemplate.openshift.io
            version: v1alpha1
          - kind: ClusterTemplateSetup
            name: clustertemplatesetup.clustertemplate.openshift.io
            version: v1alpha1
          - kind: Config
            name: config.clustertemplate.openshift.io
            version: v1alpha1
        description: |
          **Self-service clusters with guardrails!**
          Cluster as a service operator provides an easy way to define clusters as templates and allows creating instances of those templates even for non-privileged developer/devops engineers. Cluster templates operator also allows specifing quotas for the developer/devops engineers.

          ### User Interface
          The User interface is installed but is not enabled by default. To enable go to `Home` -> `Overview` -> in the status card click `Dynamic Plugins` -> click `View all` -> click the `pencil` icon -> pick `Enable` and hit `Save`. You can access the UI in the `All Clusters` perspective under `Infrastructure` -> `Cluster Templates`.

          ### Documentation
          Documentation can be found on our [website](https://github.com/stolostron/cluster-templates-operator).

          ### Getting help
          To report any issues or ask questions, create an issue on our [Github repo](https://github.com/stolostron/cluster-templates-operator/issues)
        displayName: Cluster as a service operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cluster
        - templates
        - hive
        - hypershift
        - selfservice
        - openshift
        - multicluster
        - quota
        - helm
        - argocd
        links:
        - name: Cluster as a service operator
          url: https://github.com/stolostron/cluster-templates-operator
        maintainers:
        - email: rawagner@redhat.com
          name: Rastislav Wagner
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Stolostron
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/stolostron/cluster-templates-operator:2.11.0-0c3131154065397d37e3b66ae635d7ea88650bf9
        version: 0.1.5
      entries:
      - name: cluster-aas-operator.v0.1.5
        version: 0.1.5
      - name: cluster-aas-operator.v0.1.4
        version: 0.1.4
      - name: cluster-aas-operator.v0.1.3
        version: 0.1.3
      - name: cluster-aas-operator.v0.1.2
        version: 0.1.2
      - name: cluster-aas-operator.v0.1.1
        version: 0.1.1
      - name: cluster-aas-operator.v0.1.0
        version: 0.1.0
      - name: cluster-aas-operator.v0.0.6
        version: 0.0.6
      - name: cluster-aas-operator.v0.0.5
        version: 0.0.5
      - name: cluster-aas-operator.v0.0.4
        version: 0.0.4
      - name: cluster-aas-operator.v0.0.3
        version: 0.0.3
      - name: cluster-aas-operator.v0.0.2
        version: 0.0.2
      - name: cluster-aas-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cluster-aas-operator
    provider:
      name: Stolostron
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: ibm-spectrum-scale-csi-operator
      app.kubernetes.io/managed-by: ibm-spectrum-scale-csi-operator
      app.kubernetes.io/name: ibm-spectrum-scale-csi-operator
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-spectrum-scale-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-spectrum-scale-csi-operator.v2.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "CSIScaleOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-spectrum-scale-csi-operator",
                    "app.kubernetes.io/managed-by": "ibm-spectrum-scale-csi-operator",
                    "app.kubernetes.io/name": "ibm-spectrum-scale-csi-operator",
                    "release": "ibm-spectrum-scale-csi-operator"
                  },
                  "name": "ibm-spectrum-scale-csi",
                  "namespace": "ibm-spectrum-scale-csi-driver"
                },
                "spec": {
                  "attacherNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "clusters": [
                    {
                      "id": "\u003cPrimary Cluster ID\u003e",
                      "primary": {
                        "primaryFs": "\u003c Primary Filesystem \u003e"
                      },
                      "restApi": [
                        {
                          "guiHost": "\u003c Primary cluster GUI IP/Hostname \u003e"
                        }
                      ],
                      "secrets": "secret1",
                      "secureSslMode": false
                    }
                  ],
                  "pluginNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "provisionerNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "resizerNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "snapshotterNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ]
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/ibm-spectrum-scale/ibm-spectrum-scale-csi-operator@sha256:b62ad588fbc338090f405740fd8bceaf59b3267cdb532c60ae965643ccae4d5d
          createdAt: Thu May 5 04:42:15 IST 2022
          description: An operator for deploying and managing the IBM Spectrum Scale
            CSI Driver.
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-spectrum-scale-csi/
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIScaleOperator is the Schema for the csiscaleoperators
              API
            displayName: IBM CSI Spectrum Scale Driver
            kind: CSIScaleOperator
            name: csiscaleoperators.csi.ibm.com
            version: v1
        description: |
          The IBM Spectrum Scale CSI Operator for Kubernetes installs, manages,
          upgrades the IBM Spectrum Scale CSI Driver on OpenShift and Kubernetes
          for provisioning of persistent volumes on IBM Spectrum Scale.

          ### Documentation Documentation for IBM Spectrum Scale CSI Driver and Operator can be found at [IBM Documentation](https://www.ibm.com/docs/en/spectrum-scale-csi).

          ### Getting help
          IBM Spectrum Scale CSI driver is part of the IBM Spectrum Scale offering.
          Please follow the [IBM support procedure](https://www.ibm.com/mysupport/s/?language=en_US) for any issues with the driver.

          For non-urgent issues, suggestions, recommendations feel free to open an
          issue in [github](https://github.com/IBM/ibm-spectrum-scale-csi). Issues will be addressed as team availability permits.

          **Note:** Custom Resource example represents the minimum settings needed to run
          the operator. If your environment needs more advanced settings (e.g. remote clusters,
          node mapping, image pull secret, Taint and Toleration etc.) please refer to  the sample [Custom Resource](https://raw.githubusercontent.com/IBM/ibm-spectrum-scale-csi/v2.6.0/operator/config/samples/csiscaleoperators.csi.ibm.com_cr.yaml).
        displayName: IBM Spectrum Scale CSI Plugin Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - GPFS
        - Spectrum Scale
        - CSI Driver
        links:
        - name: IBM CSI Spectrum Scale Documentation
          url: https://www.ibm.com/docs/en/spectrum-scale-csi
        - name: CSI Developer Documentation
          url: https://kubernetes-csi.github.io/docs/
        maintainers:
        - email: madhu.punjabi@in.ibm.com
          name: Madhu Thorat
        - email: amdabhad@in.ibm.com
          name: Amol Dabhade
        - email: Nitish.Kumar4@ibm.com
          name: Nitish Kumar
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibm-spectrum-scale/ibm-spectrum-scale-csi-operator@sha256:b62ad588fbc338090f405740fd8bceaf59b3267cdb532c60ae965643ccae4d5d
        version: 2.6.0
      entries:
      - name: ibm-spectrum-scale-csi-operator.v2.6.0
        version: 2.6.0
      - name: ibm-spectrum-scale-csi-operator.v2.5.1
        version: 2.5.1
      - name: ibm-spectrum-scale-csi-operator.v2.5.0
        version: 2.5.0
      - name: ibm-spectrum-scale-csi-operator.v2.4.0
        version: 2.4.0
      - name: ibm-spectrum-scale-csi-operator.v2.3.1
        version: 2.3.1
      - name: ibm-spectrum-scale-csi-operator.v2.3.0
        version: 2.3.0
      name: stable
    defaultChannel: stable
    packageName: ibm-spectrum-scale-csi-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Managed Openshift Black Belt
      provider-url: ""
    name: ecr-secret-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ecr-secret-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "ecr-secret-sample",
                  "namespace": "test-ecr-secret-operator"
                },
                "spec": {
                  "ecr_registry": "ACCOUNT_ID.dkr.ecr.us-east-2.amazonaws.com",
                  "frequency": "10h",
                  "generated_secret_name": "ecr-docker-secret",
                  "region": "us-east-2"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ecr-secret-operator:v0.3.2
          operatorframework.io/suggested-namespace: ecr-secret-operator
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ecr-secret-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret is the Schema for the secrets API
            displayName: Secret
            kind: Secret
            name: secrets.ecr.mobb.redhat.com
            version: v1alpha1
        description: "# ECR Secret Operator\n\nAmazon Elastic Container Registry [Private
          Registry Authentication](https://docs.aws.amazon.com/AmazonECR/latest/userguide/registry_auth.html)
          provides a temporary token that is valid only for 12 hours. It is a challenge
          for automatic container image build process to refresh the token or secret
          in a timely manner.\n\nThis operators frequently talks with AWS ECR GetAuthroization
          Token and create/update the secret, so that the service account can perform
          docker image build.\n\n\n## How to use this operator\n\n### Prerequisites\n\n*
          [Create an ECR private repository](https://docs.aws.amazon.com/AmazonECR/latest/userguide/repository-create.html)\n*
          Create An Openshift Cluster\n* Provide AWS Authentication to the operator.
          Two Options:\n  **Note: You have to provide a secret aws-ecr-cloud-credentials
          in the installed namespace**\n  * [IAM User](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_user.md)\n
          \ * [STS Assume Role](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_assume_role.md)\n\n###
          Install the operator\n\n### Create the ECR Secret CRD\n\n```\napiVersion:
          ecr.mobb.redhat.com/v1alpha1\nkind: Secret\nmetadata:\n  name: ecr-secret\n
          \ namespace: test-ecr-secret-operator\nspec:\n  generated_secret_name: ecr-docker-secret\n
          \ ecr_registry: [ACCOUNT_ID].dkr.ecr.us-east-2.amazonaws.com\n  frequency:
          10h\n  region: us-east-2\n```\n\n```\noc create -f CRD_FILE\n```\n\nA docker
          registry secret is created by the operator momentally and the token is patched
          every 10 hours\n\n```\noc get secret ecr-docker-secret   \nNAME                TYPE
          \                            DATA   AGE\necr-docker-secret   kubernetes.io/dockerconfigjson
          \  1      16h\n```\n\n### A sample build process with generated secret\n\n\nLink
          the secret to builder\n\n```\noc secrets link builder ecr-docker-secret
          \n```\n\nConfigure [build config](https://github.com/rh-mobb/ecr-secret-operator/tree/main/samples/build-config.yaml)
          to point to your ECR Container repository\n\n```\noc create imagestream
          ruby\noc tag openshift/ruby:2.5-ubi8 ruby:2.5\noc create -f samples/build-config.yaml\noc
          start-build ruby-sample-build --wait\n```\n\nBuild should succeed and push
          the image to the the private ECR Container repository\n\n"
        displayName: ECR Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - redhat
        links:
        - name: Ecr Secret Operator
          url: https://github.com/rh-mobb/ecr-secret-operator
        maintainers:
        - email: shading@redhat.com
          name: Shaozhen Ding
        maturity: alpha
        provider:
          name: Managed Openshift Black Belt
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/mobb/ecr-secret-operator:v0.3.2
        version: 0.3.2
      entries:
      - name: ecr-secret-operator.v0.3.2
        version: 0.3.2
      name: v0.3
    - currentCSV: ecr-secret-operator.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "ecr-secret-sample",
                  "namespace": "test-ecr-secret-operator"
                },
                "spec": {
                  "ecr_registry": "ACCOUNT_ID.dkr.ecr.us-east-2.amazonaws.com",
                  "frequency": "10h",
                  "generated_secret_name": "ecr-docker-secret",
                  "region": "us-east-2"
                }
              },
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "ArgoHelmRepoSecret",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ecr-secret-operator",
                    "app.kubernetes.io/instance": "argohelmreposecret-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "argohelmreposecret",
                    "app.kubernetes.io/part-of": "ecr-secret-operator"
                  },
                  "name": "argohelmreposecret-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ecr-secret-operator:v0.4.1
          createdAt: "2023-04-14T14:52:13Z"
          operatorframework.io/suggested-namespace: ecr-secret-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ecr-secret-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ArgoHelmRepoSecret is the Schema for the argohelmreposecrets
              API
            displayName: Argo Helm Repo Secret
            kind: ArgoHelmRepoSecret
            name: argohelmreposecrets.ecr.mobb.redhat.com
            version: v1alpha1
          - description: Secret is the Schema for the secrets API
            displayName: Secret
            kind: Secret
            name: secrets.ecr.mobb.redhat.com
            version: v1alpha1
        description: "# ECR Secret Operator\n\nAmazon Elastic Container Registry [Private
          Registry Authentication](https://docs.aws.amazon.com/AmazonECR/latest/userguide/registry_auth.html)
          provides a temporary token that is valid only for 12 hours. It is a challenge
          for automatic container image build process to refresh the token or secret
          in a timely manner.\n\nThis operators frequently talks with AWS ECR GetAuthroization
          Token and create/update the secret, so that the service account can perform
          docker image build.\n\n\n## How to use this operator\n\n### Prerequisites\n\n*
          [Create an ECR private repository](https://docs.aws.amazon.com/AmazonECR/latest/userguide/repository-create.html)\n*
          Create An Openshift Cluster\n* Provide AWS Authentication to the operator.
          Two Options:\n  **Note: You have to provide a secret aws-ecr-cloud-credentials
          in the installed namespace**\n  * [IAM User](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_user.md)\n
          \ * [STS Assume Role](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_assume_role.md)\n\n###
          Install the operator\n\n### Create the ECR Secret CRD\n\n```\napiVersion:
          ecr.mobb.redhat.com/v1alpha1\nkind: Secret\nmetadata:\n  name: ecr-secret\n
          \ namespace: test-ecr-secret-operator\nspec:\n  generated_secret_name: ecr-docker-secret\n
          \ ecr_registry: [ACCOUNT_ID].dkr.ecr.us-east-2.amazonaws.com\n  frequency:
          10h\n  region: us-east-2\n```\n\n```\noc create -f CRD_FILE\n```\n\nA docker
          registry secret is created by the operator momentally and the token is patched
          every 10 hours\n\n```\noc get secret ecr-docker-secret   \nNAME                TYPE
          \                            DATA   AGE\necr-docker-secret   kubernetes.io/dockerconfigjson
          \  1      16h\n```\n\n### A sample build process with generated secret\n\n\nLink
          the secret to builder\n\n```\noc secrets link builder ecr-docker-secret
          \n```\n\nConfigure [build config](https://github.com/rh-mobb/ecr-secret-operator/tree/main/samples/build-config.yaml)
          to point to your ECR Container repository\n\n```\noc create imagestream
          ruby\noc tag openshift/ruby:2.5-ubi8 ruby:2.5\noc create -f samples/build-config.yaml\noc
          start-build ruby-sample-build --wait\n```\n\nBuild should succeed and push
          the image to the the private ECR Container repository\n\n"
        displayName: ECR Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - redhat
        links:
        - name: Ecr Secret Operator
          url: https://github.com/rh-mobb/ecr-secret-operator
        maintainers:
        - email: shading@redhat.com
          name: Shaozhen Ding
        maturity: alpha
        provider:
          name: Managed Openshift Black Belt
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/mobb/ecr-secret-operator:v0.4.1
        version: 0.4.1
      entries:
      - name: ecr-secret-operator.v0.4.1
        version: 0.4.1
      - name: ecr-secret-operator.v0.4.0
        version: 0.4.0
      name: v0.4
    defaultChannel: v0.4
    packageName: ecr-secret-operator
    provider:
      name: Managed Openshift Black Belt
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-apigatewayv2-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-apigatewayv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Deployment",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "API",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Authorizer",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Route",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Stage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "VPCLink",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/apigatewayv2-controller:1.1.1
          createdAt: "2025-09-19T21:58:07Z"
          description: AWS API Gateway v2 controller is a service controller for managing
            API Gateway v2 resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIMapping represents the state of an AWS apigatewayv2 APIMapping
              resource.
            displayName: APIMapping
            kind: APIMapping
            name: apimappings.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: API represents the state of an AWS apigatewayv2 API resource.
            displayName: API
            kind: API
            name: apis.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Authorizer represents the state of an AWS apigatewayv2 Authorizer
              resource.
            displayName: Authorizer
            kind: Authorizer
            name: authorizers.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Deployment represents the state of an AWS apigatewayv2 Deployment
              resource.
            displayName: Deployment
            kind: Deployment
            name: deployments.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: DomainName represents the state of an AWS apigatewayv2 DomainName
              resource.
            displayName: DomainName
            kind: DomainName
            name: domainnames.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Integration represents the state of an AWS apigatewayv2 Integration
              resource.
            displayName: Integration
            kind: Integration
            name: integrations.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Route represents the state of an AWS apigatewayv2 Route resource.
            displayName: Route
            kind: Route
            name: routes.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Stage represents the state of an AWS apigatewayv2 Stage resource.
            displayName: Stage
            kind: Stage
            name: stages.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: VPCLink represents the state of an AWS apigatewayv2 VPCLink
              resource.
            displayName: VPCLink
            kind: VPCLink
            name: vpclinks.apigatewayv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon API Gateway v2 resources in AWS from within your Kubernetes cluster. Currently only HTTP APIs are supported. WebSocket API support will be added soon. For HTTP APIs, DomainName and APIMapping resource support will be added soon.

          **About Amazon API Gateway v2**

          Amazon API Gateway is a fully managed service that makes it easy for developers to create, publish, maintain, monitor, and secure APIs at any scale. APIs act as the "front door" for applications to access data, business logic, or functionality from your backend services. Using API Gateway, you can create RESTful APIs and WebSocket APIs that enable real-time two-way communication applications. API Gateway supports containerized and serverless workloads, as well as web applications.

          API Gateway handles all the tasks involved in accepting and processing up to hundreds of thousands of concurrent API calls, including traffic management, CORS support, authorization and access control, throttling, monitoring, and API version management. API Gateway has no minimum fees or startup costs. You pay for the API calls you receive and the amount of data transferred out and, with the API Gateway tiered pricing model, you can reduce your cost as your API usage scales.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon API Gateway v2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apigatewayv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon API Gateway v2 Developer Resources
          url: https://aws.amazon.com/api-gateway/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: api gateway maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/apigatewayv2-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-apigatewayv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-apigatewayv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-apigatewayv2-controller.v1.0.30
        version: 1.0.30
      - name: ack-apigatewayv2-controller.v1.0.29
        version: 1.0.29
      - name: ack-apigatewayv2-controller.v1.0.28
        version: 1.0.28
      - name: ack-apigatewayv2-controller.v1.0.26
        version: 1.0.26
      - name: ack-apigatewayv2-controller.v1.0.25
        version: 1.0.25
      - name: ack-apigatewayv2-controller.v1.0.23
        version: 1.0.23
      - name: ack-apigatewayv2-controller.v1.0.21
        version: 1.0.21
      - name: ack-apigatewayv2-controller.v1.0.17
        version: 1.0.17
      - name: ack-apigatewayv2-controller.v1.0.16
        version: 1.0.16
      - name: ack-apigatewayv2-controller.v1.0.15
        version: 1.0.15
      - name: ack-apigatewayv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-apigatewayv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-apigatewayv2-controller.v1.0.7
        version: 1.0.7
      - name: ack-apigatewayv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-apigatewayv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-apigatewayv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-apigatewayv2-controller.v1.0.3
        version: 1.0.3
      - name: ack-apigatewayv2-controller.v1.0.2
        version: 1.0.2
      - name: ack-apigatewayv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-apigatewayv2-controller.v0.1.6
        version: 0.1.6
      - name: ack-apigatewayv2-controller.v0.1.4
        version: 0.1.4
      - name: ack-apigatewayv2-controller.v0.1.3
        version: 0.1.3
      - name: ack-apigatewayv2-controller.v0.1.2
        version: 0.1.2
      - name: ack-apigatewayv2-controller.v0.1.1
        version: 0.1.1
      - name: ack-apigatewayv2-controller.v0.1.0
        version: 0.1.0
      - name: ack-apigatewayv2-controller.v0.0.22
        version: 0.0.22
      - name: ack-apigatewayv2-controller.v0.0.21
        version: 0.0.21
      - name: ack-apigatewayv2-controller.v0.0.20
        version: 0.0.20
      - name: ack-apigatewayv2-controller.v0.0.19
        version: 0.0.19
      - name: ack-apigatewayv2-controller.v0.0.18
        version: 0.0.18
      - name: ack-apigatewayv2-controller.v0.0.17
        version: 0.0.17
      - name: ack-apigatewayv2-controller.v0.0.16
        version: 0.0.16
      - name: ack-apigatewayv2-controller.v0.0.15
        version: 0.0.15
      - name: ack-apigatewayv2-controller.v0.0.14
        version: 0.0.14
      - name: ack-apigatewayv2-controller.v0.0.13
        version: 0.0.13
      - name: ack-apigatewayv2-controller.v0.0.12
        version: 0.0.12
      - name: ack-apigatewayv2-controller.v0.0.11
        version: 0.0.11
      - name: ack-apigatewayv2-controller.v0.0.10
        version: 0.0.10
      - name: ack-apigatewayv2-controller.v0.0.9
        version: 0.0.9
      name: alpha
    defaultChannel: alpha
    packageName: ack-apigatewayv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Backube
      provider-url: https://github.com/backube
    name: snapscheduler
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: snapscheduler.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "snapscheduler.backube/v1",
                "kind": "SnapshotSchedule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "snapscheduler",
                    "app.kubernetes.io/instance": "daily",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "snapshotschedule",
                    "app.kubernetes.io/part-of": "snapscheduler"
                  },
                  "name": "daily"
                },
                "spec": {
                  "retention": {
                    "maxCount": 7
                  },
                  "schedule": "0 0 * * *"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/backube/snapscheduler:3.5.0
          createdAt: "2025-05-14T14:46:43Z"
          description: The SnapScheduler operator takes snapshots of Kubernetes CSI-based
            persistent volumes according to user-supplied schedules.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/snapscheduler
          support: https://github.com/backube/snapscheduler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SnapshotSchedule defines a schedule for taking automated
              snapshots of PVC(s)
            displayName: Snapshot Schedule
            kind: SnapshotSchedule
            name: snapshotschedules.snapscheduler.backube
            version: v1
        description: |-
          ## About this operator

          The SnapScheduler operator takes snapshots of CSI-based PersistentVolumes
          according to a configurable
          [Cron-like](https://en.wikipedia.org/wiki/Cron#Overview) schedule. The
          schedules include configurable retention policies for snapshots as well as
          selectors to limit the volumes that are snapshotted. An example schedule
          could be:

          > *Snapshot **all volumes** in a namespace **daily at midnight**, retaining
          > the most recent **7** snapshots for each volume.*

          Multiple schedules can be combined to provide more elaborate protection
          schemes. For example, a given volume (or collection of volumes) could be
          protected with:

          - 6 hourly snapshots
          - 7 daily snapshots
          - 4 weekly snapshots
          - 12 monthly snapshots

          ### How it works

          The operator watches for `SnapshotSchedule` CRs in each namespace. When the
          current time matches the schedule's cronspec, the operator creates a
          `VolumeSnapshot` object for each `PersistentVolumeClaim` in the namespace
          (or subset thereof if a label selector is provided). The `VolumeSnapshot`
          objects are named according to the template:
          `<pvcname>-<schedulename>-<timestamp>`. After creating the new snapshots,
          the oldest snapshots are removed if necessary, according to the retention
          policy of the schedule.

          Please see the [full
          documentation](https://backube.github.io/snapscheduler/) for more
          information.

          ## Requirements

          - OpenShift >= 4.7 (Kubernetes >= 1.20)
          - CSI-based storage driver that supports snapshots (i.e. has the
            `CREATE_DELETE_SNAPSHOT` capability)

          ## Examples

          The schedule for snapshotting is controlled by the
          `snapshotschedules.snapscheduler.backube` Custom Resource. This is a
          namespaced resource that applies only to the PersistentVolumeClaims in its
          namespace.

          Below is a simple example that keeps 7 daily (taken at midnight) snapshots
          of all PVCs in a given namespace:

          ```yaml
          ---
          apiVersion: snapscheduler.backube/v1
          kind: SnapshotSchedule
          metadata:
            name: daily
          spec:
            retention:
              maxCount: 7
            schedule: "0 0 * * *"
          ```

          See the [usage
          documentation](https://backube.github.io/snapscheduler/usage.html) for full
          details, including how to:

          - add label selectors to restrict which PVCs this schedule applies to
          - set the VolumeSnapshotClass used by the schedule
          - apply custom labels to the automatically created VolumeSnapshot objects
        displayName: SnapScheduler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - snapshot
        - storage
        - scheduler
        links:
        - name: Documentation
          url: https://backube.github.io/snapscheduler/
        - name: Source
          url: https://github.com/backube/snapscheduler
        - name: Blog
          url: https://backube.github.io/
        maintainers:
        - email: jstrunk@redhat.com
          name: John Strunk
        maturity: stable
        minKubeVersion: 1.20.0-0
        provider:
          name: Backube
          url: https://github.com/backube
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.19.1
        - quay.io/backube/snapscheduler:3.5.0
        version: 3.5.0
      entries:
      - name: snapscheduler.v3.5.0
        version: 3.5.0
      - name: snapscheduler.v3.4.0
        version: 3.4.0
      - name: snapscheduler.v3.3.0
        version: 3.3.0
      - name: snapscheduler.v3.2.0
        version: 3.2.0
      - name: snapscheduler.v3.1.0
        version: 3.1.0
      - name: snapscheduler.v3.0.0
        version: 3.0.0
      - name: snapscheduler.v2.1.0
        version: 2.1.0
      - name: snapscheduler.v2.0.0
        version: 2.0.0
      name: candidate
    - currentCSV: snapscheduler.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "snapscheduler.backube/v1",
                "kind": "SnapshotSchedule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "snapscheduler",
                    "app.kubernetes.io/instance": "daily",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "snapshotschedule",
                    "app.kubernetes.io/part-of": "snapscheduler"
                  },
                  "name": "daily"
                },
                "spec": {
                  "retention": {
                    "maxCount": 7
                  },
                  "schedule": "0 0 * * *"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/backube/snapscheduler:3.5.0
          createdAt: "2025-05-14T14:46:43Z"
          description: The SnapScheduler operator takes snapshots of Kubernetes CSI-based
            persistent volumes according to user-supplied schedules.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/snapscheduler
          support: https://github.com/backube/snapscheduler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SnapshotSchedule defines a schedule for taking automated
              snapshots of PVC(s)
            displayName: Snapshot Schedule
            kind: SnapshotSchedule
            name: snapshotschedules.snapscheduler.backube
            version: v1
        description: |-
          ## About this operator

          The SnapScheduler operator takes snapshots of CSI-based PersistentVolumes
          according to a configurable
          [Cron-like](https://en.wikipedia.org/wiki/Cron#Overview) schedule. The
          schedules include configurable retention policies for snapshots as well as
          selectors to limit the volumes that are snapshotted. An example schedule
          could be:

          > *Snapshot **all volumes** in a namespace **daily at midnight**, retaining
          > the most recent **7** snapshots for each volume.*

          Multiple schedules can be combined to provide more elaborate protection
          schemes. For example, a given volume (or collection of volumes) could be
          protected with:

          - 6 hourly snapshots
          - 7 daily snapshots
          - 4 weekly snapshots
          - 12 monthly snapshots

          ### How it works

          The operator watches for `SnapshotSchedule` CRs in each namespace. When the
          current time matches the schedule's cronspec, the operator creates a
          `VolumeSnapshot` object for each `PersistentVolumeClaim` in the namespace
          (or subset thereof if a label selector is provided). The `VolumeSnapshot`
          objects are named according to the template:
          `<pvcname>-<schedulename>-<timestamp>`. After creating the new snapshots,
          the oldest snapshots are removed if necessary, according to the retention
          policy of the schedule.

          Please see the [full
          documentation](https://backube.github.io/snapscheduler/) for more
          information.

          ## Requirements

          - OpenShift >= 4.7 (Kubernetes >= 1.20)
          - CSI-based storage driver that supports snapshots (i.e. has the
            `CREATE_DELETE_SNAPSHOT` capability)

          ## Examples

          The schedule for snapshotting is controlled by the
          `snapshotschedules.snapscheduler.backube` Custom Resource. This is a
          namespaced resource that applies only to the PersistentVolumeClaims in its
          namespace.

          Below is a simple example that keeps 7 daily (taken at midnight) snapshots
          of all PVCs in a given namespace:

          ```yaml
          ---
          apiVersion: snapscheduler.backube/v1
          kind: SnapshotSchedule
          metadata:
            name: daily
          spec:
            retention:
              maxCount: 7
            schedule: "0 0 * * *"
          ```

          See the [usage
          documentation](https://backube.github.io/snapscheduler/usage.html) for full
          details, including how to:

          - add label selectors to restrict which PVCs this schedule applies to
          - set the VolumeSnapshotClass used by the schedule
          - apply custom labels to the automatically created VolumeSnapshot objects
        displayName: SnapScheduler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - snapshot
        - storage
        - scheduler
        links:
        - name: Documentation
          url: https://backube.github.io/snapscheduler/
        - name: Source
          url: https://github.com/backube/snapscheduler
        - name: Blog
          url: https://backube.github.io/
        maintainers:
        - email: jstrunk@redhat.com
          name: John Strunk
        maturity: stable
        minKubeVersion: 1.20.0-0
        provider:
          name: Backube
          url: https://github.com/backube
        relatedImages:
        - quay.io/backube/snapscheduler:3.5.0
        - quay.io/brancz/kube-rbac-proxy:v0.19.1
        version: 3.5.0
      entries:
      - name: snapscheduler.v3.5.0
        version: 3.5.0
      - name: snapscheduler.v3.4.0
        version: 3.4.0
      - name: snapscheduler.v3.3.0
        version: 3.3.0
      - name: snapscheduler.v3.2.0
        version: 3.2.0
      - name: snapscheduler.v3.1.0
        version: 3.1.0
      - name: snapscheduler.v3.0.0
        version: 3.0.0
      - name: snapscheduler.v2.1.0
        version: 2.1.0
      - name: snapscheduler.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: snapscheduler
    provider:
      name: Backube
      url: https://github.com/backube
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CNCF
      provider-url: ""
    name: jaeger
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jaeger-operator.v1.65.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "jaegertracing.io/v1",
                "kind": "Jaeger",
                "metadata": {
                  "name": "my-jaeger"
                },
                "spec": {
                  "strategy": "allInOne"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing
          certified: "false"
          containerImage: quay.io/jaegertracing/jaeger-operator:1.65.0
          createdAt: "2025-01-22T20:40:19Z"
          description: Provides tracing, monitoring and troubleshooting for microservices-based
            distributed systems
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/jaegertracing/jaeger-operator
          support: Jaeger Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Jaeger is the Schema for the jaegers API
            displayName: Jaeger
            kind: Jaeger
            name: jaegers.jaegertracing.io
            version: v1
        description: |-
          Jaeger, inspired by [Dapper](https://research.google.com/pubs/pub36356.html) and [OpenZipkin](http://zipkin.io/), is a distributed tracing system released as open source by Uber Technologies. It is used for monitoring and troubleshooting microservices-based distributed systems.
          ### Core capabilities
          Jaeger is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed context propagation
          * Distributed transaction monitoring
          * Root cause analysis
          * Service dependency analysis
          * Performance / latency optimization
          * OpenTracing compatible data model
          * Multiple storage backends: Badger, Cassandra, Elasticsearch, Memory.
          ### Operator features
          * **Multiple modes** - Supports `allInOne`, `production` and `streaming` [modes of deployment](https://www.jaegertracing.io/docs/latest/operator/#deployment-strategies).
          * **Configuration** - The Operator manages [configuration information](https://www.jaegertracing.io/docs/latest/operator/#configuring-the-custom-resource) when installing Jaeger instances.
          * **Storage** - [Configure storage](https://www.jaegertracing.io/docs/latest/operator/#storage-options) used by Jaeger. By default, `memory` is used. Other options include `badger`, `cassandra` or `elasticsearch`. On OpenShift, the operator can delegate creation of an Elasticsearch cluster to the Elasticsearch Operator if deployed.
          * **Agent** - can be deployed as [sidecar](https://www.jaegertracing.io/docs/latest/operator/#auto-injecting-jaeger-agent-sidecars) (default) and/or [daemonset](https://www.jaegertracing.io/docs/latest/operator/#installing-the-agent-as-daemonset).
          * **UI** - Optionally setup ingress (Kubernetes) or secure route (OpenShift) to provide [access to the Jaeger UI](https://www.jaegertracing.io/docs/latest/operator/#accessing-the-jaeger-console-ui).
          ### Before you start
          1. Ensure that the appropriate storage solution, that will be used by the Jaeger instance, is available and configured.
          2. If intending to deploy an Elasticsearch cluster via the Jaeger custom resource, then the Elasticsearch Operator must first be installed.
          ### Troubleshooting
          * https://www.jaegertracing.io/docs/latest/troubleshooting/
        displayName: Community Jaeger Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tracing
        - monitoring
        - troubleshooting
        links:
        - name: Jaeger Operator Source Code
          url: https://github.com/jaegertracing/jaeger-operator
        maintainers:
        - email: jaeger-tracing@googlegroups.com
          name: Jaeger Google Group
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: CNCF
        relatedImages:
        - quay.io/jaegertracing/jaeger-operator:1.65.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 1.65.0
      entries:
      - name: jaeger-operator.v1.65.0
        version: 1.65.0
      - name: jaeger-operator.v1.62.0
        version: 1.62.0
      - name: jaeger-operator.v1.61.0
        version: 1.61.0
      - name: jaeger-operator.v1.60.0
        version: 1.60.0
      - name: jaeger-operator.v1.59.0
        version: 1.59.0
      - name: jaeger-operator.v1.57.0
        version: 1.57.0
      - name: jaeger-operator.v1.56.0
        version: 1.56.0
      - name: jaeger-operator.v1.55.0
        version: 1.55.0
      - name: jaeger-operator.v1.53.0
        version: 1.53.0
      - name: jaeger-operator.v1.52.0
        version: 1.52.0
      - name: jaeger-operator.v1.51.0
        version: 1.51.0
      - name: jaeger-operator.v1.49.0
        version: 1.49.0
      - name: jaeger-operator.v1.48.1
        version: 1.48.1
      - name: jaeger-operator.v1.47.0
        version: 1.47.0
      - name: jaeger-operator.v1.46.0
        version: 1.46.0
      - name: jaeger-operator.v1.45.0
        version: 1.45.0
      - name: jaeger-operator.v1.44.0
        version: 1.44.0
      - name: jaeger-operator.v1.43.0
        version: 1.43.0
      - name: jaeger-operator.v1.42.0
        version: 1.42.0
      - name: jaeger-operator.v1.41.1
        version: 1.41.1
      - name: jaeger-operator.v1.41.0
        version: 1.41.0
      - name: jaeger-operator.v1.40.0
        version: 1.40.0
      - name: jaeger-operator.v1.39.0
        version: 1.39.0
      - name: jaeger-operator.v1.38.1
        version: 1.38.1
      - name: jaeger-operator.v1.38.0
        version: 1.38.0
      - name: jaeger-operator.v1.37.0
        version: 1.37.0
      - name: jaeger-operator.v1.36.0
        version: 1.36.0
      - name: jaeger-operator.v1.35.0
        version: 1.35.0
      - name: jaeger-operator.v1.34.1
        version: 1.34.1
      - name: jaeger-operator.v1.34.0
        version: 1.34.0
      - name: jaeger-operator.v1.33.0
        version: 1.33.0
      - name: jaeger-operator.v1.32.0
        version: 1.32.0
      - name: jaeger-operator.v1.31.0
        version: 1.31.0
      - name: jaeger-operator.v1.30.0
        version: 1.30.0
      - name: jaeger-operator.v1.29.1
        version: 1.29.1
      - name: jaeger-operator.v1.25.0
        version: 1.25.0
      - name: jaeger-operator.v1.24.0
        version: 1.24.0
      name: stable
    defaultChannel: stable
    packageName: jaeger
    provider:
      name: CNCF
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: watcher-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: watcher-operator.v1.0.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "Watcher",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcher",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcher",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcher"
                },
                "spec": {
                  "apiServiceTemplate": {
                    "tls": {
                      "caBundleSecretName": "combined-ca-bundle"
                    }
                  },
                  "databaseInstance": "openstack"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherAPI",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcher-operator"
                  },
                  "name": "watcherapi-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherApplier",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcherapplier-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcherapplier",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcherapplier-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherDecisionEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcherdecisionengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcherdecisionengine",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcherdecisionengine-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-09-02T13:32:31Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.14
          operatorframework.io/suggested-namespace: openstack-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: WatcherAPI is the Schema for the watcherapis API
            displayName: Watcher API
            kind: WatcherAPI
            name: watcherapis.watcher.openstack.org
            version: v1beta1
          - description: WatcherApplier is the Schema for the watcherappliers API
            displayName: Watcher Applier
            kind: WatcherApplier
            name: watcherappliers.watcher.openstack.org
            version: v1beta1
          - description: WatcherDecisionEngine is the Schema for the watcherdecisionengines
              API
            displayName: Watcher Decision Engine
            kind: WatcherDecisionEngine
            name: watcherdecisionengines.watcher.openstack.org
            version: v1beta1
          - description: Watcher is the Schema for the watchers API
            displayName: Watcher
            kind: Watcher
            name: watchers.watcher.openstack.org
            version: v1beta1
        description: The Watcher Operator project
        displayName: Watcher Operator (Technology Preview)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        links:
        - name: Watcher Operator
          url: https://github.com/openstack-k8s-operators/watcher-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/watcher-rhel9-operator@sha256:63c341a3e20a522c601098502a8fd8f1164d23c773ae8f55782c43ff4b6c8c18
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:eed8234b0fe47083ab941f513709f778bfcb80e32747b71a65fd5f62328a5ca0
        - registry.redhat.io/rhoso/openstack-watcher-api-rhel9@sha256:e97369e3a47bb882f1920c6af3f6345a842af45c58f0f5c5df9ac2933cf3098a
        - registry.redhat.io/rhoso/openstack-watcher-decision-engine-rhel9@sha256:c759d66cf0986c0036be91c008cbd059b4412c94160ed82dd12cd4e00d0685a9
        - registry.redhat.io/rhoso/openstack-watcher-applier-rhel9@sha256:c92a82f901613991926367bb8aab4f875c27a3e4a5282d92847a8463d98cd9e3
        version: 1.0.14
      entries:
      - name: watcher-operator.v1.0.14
        version: 1.0.14
      - name: watcher-operator.v1.0.13
        version: 1.0.13
      - name: watcher-operator.v1.0.12
        version: 1.0.12
      - name: watcher-operator.v1.0.11
        version: 1.0.11
      - name: watcher-operator.v1.0.10
        version: 1.0.10
      - name: watcher-operator.v1.0.9
        version: 1.0.9
      - name: watcher-operator.v1.0.8
        version: 1.0.8
      - name: watcher-operator.v1.0.7
        version: 1.0.7
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: watcher-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NdmSpc project
      provider-url: ""
    name: ndmspc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ndmspc-operator.v0.20250209.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.ndmspc.io/v1alpha1",
                "kind": "NdmSpcConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ndmspc-operator",
                    "app.kubernetes.io/instance": "ndmspcconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ndmspcconfig",
                    "app.kubernetes.io/part-of": "ndmspc-operator"
                  },
                  "name": "ndmspc-sample"
                },
                "spec": {
                  "executor": {
                    "image": "registry.gitlab.com/ndmspc/api:v0.20240215.2",
                    "name": "executor"
                  },
                  "executors": 1,
                  "ingressclasstype": "nginx",
                  "ingressenabled": true,
                  "ingresshostdomain": "127.0.0.1.sslip.io",
                  "ingressnostname": "ndmspc",
                  "ingressssltraefik": false,
                  "naf": {
                    "image": "registry.gitlab.com/ndmspc/naf:v0.20250129.0",
                    "name": "naf"
                  },
                  "ndmvr": {
                    "image": "registry.gitlab.com/ndmspc/ndmvr/shared:v0.20230320.0",
                    "name": "ndmvr"
                  },
                  "salsa": {
                    "image": "registry.gitlab.com/ndmspc/user/ndmspc:v0.20240412.0"
                  },
                  "storage": {
                    "image": "registry.gitlab.com/ndmspc/storage-server:v0.20250205.0",
                    "mongodb": {
                      "image": "docker.io/library/mongo:latest",
                      "password": "admin",
                      "port": 27017,
                      "username": "admin"
                    },
                    "name": "storage",
                    "port": 8080
                  },
                  "webapp": {
                    "image": "registry.gitlab.com/ndmspc/react-ndmspc:v0.20240412.0",
                    "name": ""
                  },
                  "webappcontainerport": 8080,
                  "webapps": 1,
                  "worker": {
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "1Gi"
                      }
                    }
                  },
                  "workers": 2,
                  "workerslots": 1,
                  "zmq2ws": {
                    "image": "registry.gitlab.com/ndmspc/zmq2ws:v1.20231012.0",
                    "name": "zmq2ws"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: registry.gitlab.com/ndmspc/ndmspc-operator:0.20250209.0
          createdAt: "2025-02-09T12:49:06Z"
          description: The NdmSpc operator automates the deployment and management
            of Ndmspc suite on Kubernetes cluster.
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://gitlab.com/ndmspc/ndmspc-operator
          support: NDMSPC Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NdmSpc configuration
            displayName: NdmSpc Config
            kind: NdmSpcConfig
            name: ndmspcconfigs.apps.ndmspc.io
            version: v1alpha1
        description: "The NdmSpc operator automates the deployment and management
          of Ndmspc suite on Kubernetes cluster. This operator allows user to deploy
          SALSA cluster with ZMQ2WS and job executor service and specify custom webapp
          page.\n\n### Capabilities\n\n- SALSA cluster\n- ZMQ2WS api service\n- Job
          executor service \n- Custom WebApp page\n"
        displayName: NdmSpc operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ndmspc
        links:
        - name: Ndmspc Operator
          url: https://ndmspc.gitlab.io/docs
        maintainers:
        - email: ndmspc@gmail.com
          name: NdmSpc Support
        maturity: alpha
        provider:
          name: NdmSpc project
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - registry.gitlab.com/ndmspc/ndmspc-operator:0.20250209.0
        version: 0.20250209.0
      entries:
      - name: ndmspc-operator.v0.20250209.0
        version: 0.20250209.0
      - name: ndmspc-operator.v0.20250131.0
        version: 0.20250131.0
      - name: ndmspc-operator.v0.20240412.0
        version: 0.20240412.0
      - name: ndmspc-operator.v0.11.6
        version: 0.11.6
      - name: ndmspc-operator.v0.11.4
        version: 0.11.4
      - name: ndmspc-operator.v0.11.3
        version: 0.11.3
      - name: ndmspc-operator.v0.11.2
        version: 0.11.2
      - name: ndmspc-operator.v0.11.1
        version: 0.11.1
      - name: ndmspc-operator.v0.10.2
        version: 0.10.2
      - name: ndmspc-operator.v0.10.1
        version: 0.10.1
      - name: ndmspc-operator.v0.10.0
        version: 0.10.0
      - name: ndmspc-operator.v0.9.0
        version: 0.9.0
      - name: ndmspc-operator.v0.7.4
        version: 0.7.4
      - name: ndmspc-operator.v0.7.1
        version: 0.7.1
      - name: ndmspc-operator.v0.6.0
        version: 0.6.0
      - name: ndmspc-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: ndmspc-operator
    provider:
      name: NdmSpc project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: VMware Tanzu
      provider-url: ""
    name: rabbitmq-cluster-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-cluster-operator.v2.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "rabbitmq.com/v1beta1",
                "kind": "RabbitmqCluster",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "override": {
                    "statefulset": {
                      "spec": {
                        "template": {
                          "spec": {
                            "containers": [],
                            "securityContext": {}
                          }
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/rabbitmqoperator/cluster-operator:2.16.0
          createdAt: 2025-08-12T09:34:33
          repository: https://github.com/rabbitmq/cluster-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Rabbitmq Cluster
            displayName: RabbitmqCluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: "## About the managed application\nRabbitMQ is an open-source
          message broker that originally implemented the\nAdvanced Message Queuing
          Protocol (AMQP). Nowadays features a plug-in\narchitecture supporting protocols
          such as Streaming Text Oriented Messaging\nProtocol (STOMP), MQ Telemetry
          Transport (MQTT), and others. RabbitMQ\nsupports native streaming and AMQP
          1.0.\n\nVisit [rabbitmq.com](rabbitmq.com) for more information.\n\n## About
          this Operator\nRabbitMQ Cluster Kubernetes Operator is developed and maintened
          by the VMware \nRabbitMQ team and provides a consistent and easy way to
          deploy RabbitMQ\nclusters to Kubernetes and run them, including \"day two\"
          (continuous)\noperations. RabbitMQ clusters deployed using the Operator
          can be used by\napplications running on Kubernetes or outside of Kubernetes.\n\nRead
          [RabbitMQ operator docs](https://www.rabbitmq.com/kubernetes/operator/operator-overview)\nfor
          more information.\n"
        displayName: RabbitMQ-cluster-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        - messaging
        - amqp
        - amqp10
        - stomp
        - mqtt
        - mqtt5
        - queues
        - streaming
        links:
        - name: Documentation
          url: https://www.rabbitmq.com/kubernetes/operator/operator-overview.html
        - name: GitHub
          url: https://github.com/rabbitmq/cluster-operator
        maintainers:
        - email: rabbitmq-users@googlegroups.com
          name: VMware Tanzu
        maturity: stable
        minKubeVersion: 1.26.0
        nativeApis:
        - group: apps
          kind: Deployment
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: ReplicaSet
          version: v1
        - group: ""
          kind: Pod
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: ConfigMap
          version: v1
        provider:
          name: VMware Tanzu
        relatedImages:
        - quay.io/rabbitmqoperator/cluster-operator:2.16.0
        version: 2.16.0
      entries:
      - name: rabbitmq-cluster-operator.v2.16.0
        version: 2.16.0
      - name: rabbitmq-cluster-operator.v2.15.0
        version: 2.15.0
      - name: rabbitmq-cluster-operator.v2.14.0
        version: 2.14.0
      - name: rabbitmq-cluster-operator.v2.13.0
        version: 2.13.0
      - name: rabbitmq-cluster-operator.v2.12.1
        version: 2.12.1
      - name: rabbitmq-cluster-operator.v2.10.0
        version: 2.10.0
      - name: rabbitmq-cluster-operator.v2.9.0
        version: 2.9.0
      - name: rabbitmq-cluster-operator.v2.8.0
        version: 2.8.0
      - name: rabbitmq-cluster-operator.v2.7.0
        version: 2.7.0
      - name: rabbitmq-cluster-operator.v2.6.0
        version: 2.6.0
      - name: rabbitmq-cluster-operator.v2.5.0
        version: 2.5.0
      - name: rabbitmq-cluster-operator.v2.4.0
        version: 2.4.0
      - name: rabbitmq-cluster-operator.v2.3.0
        version: 2.3.0
      - name: rabbitmq-cluster-operator.v2.2.0
        version: 2.2.0
      - name: rabbitmq-cluster-operator.v2.1.0
        version: 2.1.0
      - name: rabbitmq-cluster-operator.v2.0.0
        version: 2.0.0
      - name: rabbitmq-cluster-operator.v1.14.0
        version: 1.14.0
      name: stable
    defaultChannel: stable
    packageName: rabbitmq-cluster-operator
    provider:
      name: VMware Tanzu
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: ""
    name: ibm-metrics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-metrics-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MarketplaceConfig",
                "metadata": {
                  "name": "marketplaceconfig"
                },
                "spec": {
                  "features": {
                    "deployment": true,
                    "meterDefinitionCatalogServer": false,
                    "registration": true
                  },
                  "installIBMCatalogSource": true,
                  "isDisconnected": false,
                  "license": {
                    "accept": false
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterBase",
                "metadata": {
                  "name": "meterbase"
                },
                "spec": {
                  "enabled": true,
                  "prometheus": {
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "1",
                        "memory": "1G"
                      }
                    },
                    "storage": {
                      "size": "20Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterReport",
                "metadata": {
                  "name": "example-meterreport"
                },
                "spec": {
                  "endTime": "2020-08-05T00:00:00Z",
                  "prometheusService": {
                    "name": "rhm-prometheus-meterbase",
                    "namespace": "openshift-redhat-marketplace",
                    "targetPort": "rbac"
                  },
                  "startTime": "2020-08-04T00:00:00Z"
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "example-meterdefinition-4"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "meterdefinition-sample"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
          createdAt: "2025-08-28T14:18:55Z"
          description: The IBM Metrics Operator provides workload metering and reporting
            for IBM and Red Hat Marketplace customers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
          - description: MeterBase is the resource that sets up Metering for Red Hat
              Marketplace. This is an internal resource not meant to be modified directly.
            displayName: Meter Base
            kind: MeterBase
            name: meterbases.marketplace.redhat.com
            version: v1alpha1
          - description: MeterDefinition defines the meter workloads used to enable
              pay for use billing.
            displayName: Meter Definition
            kind: MeterDefinition
            name: meterdefinitions.marketplace.redhat.com
            version: v1beta1
          - kind: MeterReport
            name: meterreports.marketplace.redhat.com
            version: v1alpha1
        description: "The IBM Metrics Operator provides workload metering and reporting
          for IBM Software Central and Red Hat Marketplace customers.\n### **Important
          Note**\nA set of instructions for onboarding is provided here. For more
          detailed onboarding instructions or information about what is installed
          please visit [swc.saas.ibm.com](https://swc.saas.ibm.com).\n\nUsage metrics
          may be monitored through [https://swc.saas.ibm.com](https://swc.saas.ibm.com)
          with only IBM Metrics Operator and a Red Hat Marketplace account.\n\n###
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n### Prerequisites\n1. Installations are required
          to [enable monitoring for user-defined projects](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          as the Prometheus provider.\n2. Edit the cluster-monitoring-config ConfigMap
          object:\n\n   ```sh\n   $ oc -n openshift-monitoring edit configmap cluster-monitoring-config\n
          \   ```\n\n3. Add enableUserWorkload: true under data/config.yaml:\n  \n
          \   ```sh\n    apiVersion: v1\n    kind: ConfigMap\n    metadata:\n      name:
          cluster-monitoring-config\n        namespace: openshift-monitoring\n    data:\n
          \     config.yaml: |\n        enableUserWorkload: true\n    ```\n\n4. Configure
          the user-workload-monitoring-config ConfigMap object:\n\n    ```sh\n    $
          oc -n openshift-user-workload-monitoring edit configmap user-workload-monitoring-config\n
          \   ```\n\n5. Configure a minimum retention time of 168h and minimum storage
          capacity of 40Gi\n  \n    ```sh\n    apiVersion: v1\n    kind: ConfigMap\n
          \   metadata:\n      name: user-workload-monitoring-config\n      namespace:
          openshift-user-workload-monitoring\n\n    data:\n      config.yaml: |\n
          \       prometheus:\n          retention: 168h\n          volumeClaimTemplate:\n
          \           spec:\n              resources:\n                requests:\n
          \                 storage: 40Gi\n    ```\n\n### Resources Required\n\nMinimum
          system resources required:\n\n| Operator                | Memory (MB) |
          CPU (cores) | Disk (GB) | Nodes |\n| ----------------------- | -----------
          | ----------- | --------- | ----- |\n| **Metrics**   |        750  |     0.25
          \   | 3x1       |    3  |\n| **Deployment** |        250  |     0.25    |
          -         |    1  |\n\n| Prometheus Provider  | Memory (GB) | CPU (cores)
          | Disk (GB) | Nodes |\n| --------- | ----------- | ----------- | ---------
          | ----- |\n| **[Openshift User Workload Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)**
          |          1  |     0.1       | 2x40        |   2    |\n\nMultiple nodes
          are required to provide pod scheduling for high availability for IBM Metrics
          Operator Data Service and Prometheus.\n\nThe IBM Metrics Operator automatically
          creates 3 x 1Gi PersistentVolumeClaims to store reports as part of the data
          service, with _ReadWriteOnce_ access mode. Te PersistentVolumeClaims are
          automatically created by the ibm-metrics-operator after creating a `redhat-marketplace-pull-secret`
          and accepting the license in `marketplaceconfig`.\n\n| NAME                                |
          CAPACITY | ACCESS MODES |\n| ----------------------------------- | --------
          | ------------ |\n| rhm-data-service-rhm-data-service-0 | 1Gi | RWO |\n|
          rhm-data-service-rhm-data-service-1 | 1Gi | RWO |\n| rhm-data-service-rhm-data-service-2
          | 1Gi | RWO |\n\n### Supported Storage Providers\n\n- OpenShift Container
          Storage / OpenShift Data Foundation version 4.x, from version 4.2 or higher\n-
          IBM Cloud Block storage and IBM Cloud File storage\n- IBM Storage Suite
          for IBM Cloud Paks:\n  - File storage from IBM Spectrum Fusion/Scale \n
          \ - Block storage from IBM Spectrum Virtualize, FlashSystem or DS8K\n- Portworx
          Storage, version 2.5.5 or above\n- Amazon Elastic File Storage\n\n### Access
          Modes required\n\n - ReadWriteOnce (RWO)\n\n### Provisioning Options supported\n\nChoose
          one of the following options to provision storage for the ibm-metrics-operator
          data-service\n\n#### Dynamic provisioning using a default StorageClass\n
          \  - A StorageClass is defined with a `metadata.annotations: storageclass.kubernetes.io/is-default-class:
          \"true\"`\n   - PersistentVolumes will be provisioned automatically for
          the generated PersistentVolumeClaims\n--- \n#### Manually create each PersistentVolumeClaim
          with a specific StorageClass\n   - Must be performed before creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`. Otherwise, the automatically
          generated PersistentVolumeClaims are immutable.\n```\napiVersion: v1\nkind:
          PersistentVolumeClaim\nmetadata:\n  labels:\n    app: rhm-data-service\n
          \ name: rhm-data-service-rhm-data-service-0\n  namespace: ibm-software-central\nspec:\n
          \ storageClassName: rook-cephfs\n  accessModes:\n  - ReadWriteOnce\n  resources:\n
          \   requests:\n      storage: 1Gi\n```\n---\n#### Manually provision each
          PersistentVolume for the generated PersistentVolumeClaims with a specific
          StorageClass\n  - May be performed before or after creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`.  \n```\napiVersion: v1\nkind:
          PersistentVolume\nmetadata:\n  name: rhm-data-service-rhm-data-service-0\nspec:\n
          \ csi:\n    driver: rook-ceph.cephfs.csi.ceph.com\n    volumeHandle: rhm-data-service-rhm-data-service-0\n
          \ capacity:\n    storage: 1Gi\n  accessModes:\n    - ReadWriteOnce\n  persistentVolumeReclaimPolicy:
          Delete\n  storageClassName: rook-cephfs\n  volumeMode: Filesystem\n  claimRef:\n
          \   kind: PersistentVolumeClaim\n    namespace: ibm-software-central\n    name:
          rhm-data-service-rhm-data-service-0\n```\n\n### Installation\n1. Create
          or get your [pull secret](https://swc.saas.ibm.com/en-us/documentation/clusters#get-pull-secret).\n2.
          Install the IBM Metrics Operator\n3. Create a Kubernetes secret in the installed
          namespace with the name `redhat-marketplace-pull-secret` and key `PULL_SECRET`
          with the value of the pull secret.\n    ```sh\n    # Replace ${PULL_SECRET}
          with your pull secret\n    oc create secret generic redhat-marketplace-pull-secret
          -n ibm-software-central --from-literal=PULL_SECRET=${PULL_SECRET}\n    ```\n4.
          Use of the platform is governed by the:\n\n    [IBM Cloud Services Agreement](https://www.ibm.com/support/customer/csol/terms/?id=Z126-6304_WS&_ga=2.116312197.2046730452.1684328846-812467790.1684328846)
          (or other base agreement between you and IBM such as a [Passport Advantage
          Agreement](https://www.ibm.com/software/passportadvantage/pa_agreements.html?_ga=2.116312197.2046730452.1684328846-812467790.1684328846))
          and the [Service Description for the Red Hat Marketplace](https://www.ibm.com/support/customer/csol/terms/?id=i126-8719&_ga=2.83289621.2046730452.1684328846-812467790.1684328846).\n
          \   \n5. Update MarketplaceConfig to accept the license.\n    ```\n    oc
          patch marketplaceconfig marketplaceconfig -n ibm-software-central --type='merge'
          -p '{\"spec\": {\"license\": {\"accept\": true}}}'\n    ```\n6. Install
          the pull secret as a global pull secret on the cluster.\n\n    These steps
          require `oc`, `jq`, and `base64` to be available on your machine.\n\n    ```sh\n
          \   # Create the docker pull secret file using your pull secret.\n    #
          Store it in a file called entitledregistryconfigjson.\n    oc create secret
          docker-registry entitled-registry --docker-server=registry.marketplace.redhat.com
          --docker-username \"cp\" --docker-password \"${PULL_SECRET}\" --dry-run=client
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > entitledregistryconfigjson\n
          \   # Get the current global secrets on the cluster and store it as a file
          named dockerconfigjson\n    oc get secret pull-secret -n openshift-config
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > dockerconfigjson\n
          \   # Merge the two dockerconfigs together into a file called dockerconfigjson-merged.\n
          \   jq -s '.[0] * .[1]' dockerconfigjson entitledregistryconfigjson > dockerconfigjson-merged\n
          \   # Set the cluster's dockerconfig file to the new merged version.\n    oc
          set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson=dockerconfigjson-merged\n
          \   ```\n  7. (Optional) Your IBM product may have installed IBM Metrics
          Operator using an `ibm-entitlement-key` for authentication to IBM Software
          Central. To authorize the automatic creation of an IBM Software Central
          account for the IBM account owning the `ibm-entitlement-key`\n      ```\n
          \     oc patch marketplaceconfig marketplaceconfig -n ibm-software-central
          --type='merge' -p '{\"spec\": {\"authorizeAccountCreation\": true}}'\n      ```\n\n###
          Why is a global pull secret required?\nIn order to successfully install
          the products hosted by the container image registry, you will need to make
          the pull secret available across the cluster. This can be achieved by applying
          the pull as a [global pull secret](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html#images-update-global-pull-secret_using-image-pull-secrets).
          For alternative approaches, please see the official OpenShift [documentation](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html).\n\n\n###
          SecurityContextConstraints requirements\n\nThe Operators and their components
          support running under the OpenShift Container Platform default restricted
          and restricted-v2 security context constraints.\n\n### Installation Namespace
          and ClusterRoleBinding requirements\n\nThe IBM Metrics Operator components
          require specific ClusterRoleBindings.\n- The metric-state component requires
          a ClusterRoleBinding for the the `view` ClusterRole.\n  - Ability to read
          non-sensitive CustomResources. \n  - The `view` ClusterRole dynamically
          adds new CustomResourceDefinitions.\n- The operator & reporter component
          requires a ClusterRoleBinding for the the `cluster-monitoring-view` ClusterRole.\n
          \ - Ability to view Prometheus metrics. \n  - The underlying ClusterRole
          RBAC often updated between OpenShift versions.\n\nDue to limitations of
          Operator Lifecycle Manager (OLM), this ClusterRoleBinding can not be provided
          dynamically for arbitrary installation target namespaces.\n\nA static ClusterRoleBinding
          is included for installation to the default namespace of `ibm-software-central`,
          and namespaces `openshift-redhat-marketplace`,  `redhat-marketplace`, `ibm-common-services`.\n\nTo
          create the ClusterRoleBindings for installation to an alternate namespace\n```\noc
          project INSTALL-NAMESPACE\noc adm policy add-cluster-role-to-user view -z
          ibm-metrics-operator-metric-state\noc adm policy add-cluster-role-to-user
          cluster-monitoring-view -z ibm-metrics-operator-controller-manager,ibm-metrics-operator-reporter\n```\n\n###
          Metric State scoping requirements\nThe metric-state Deployment obtains `get/list/watch`
          access to metered resources via the `view` ClusterRole. For operators deployed
          using Operator Lifecycle Manager (OLM), permissions are added to `clusterrole/view`
          dynamically via a generated and annotated `-view` ClusterRole. If you wish
          to meter an operator, and its Custom Resource Definitions (CRDs) are not
          deployed through OLM, one of two options are required\n1. Add the following
          label to a clusterrole that has get/list/watch access to your CRD: `rbac.authorization.k8s.io/aggregate-to-view:
          \"true\"`, thereby dynamically adding it to `clusterrole/view`\n2. Create
          a ClusterRole that has get/list/watch access to your CRD, and create a ClusterRoleBinding
          for the metric-state ServiceAccount\n\nAttempting to meter a resource with
          a MeterDefinition without the required permissions will log an `AccessDeniedError`
          in metric-state.\n\n### Disaster Recovery\n\nTo plan for disaster recovery,
          note the PhysicalVolumeClaims `rhm-data-service-rhm-data-service-N`. \n-
          In connected environments, MeterReport data upload attempts occur hourly,
          and are then removed from data-service. There is a low risk of losing much
          unreported data.\n- In an airgap environment, MeterReport data must be pulled
          from data-service and uploaded manually using [datactl](https://github.com/redhat-marketplace/datactl/).
          To prevent data loss in a disaster scenario, the data-service volumes should
          be considered in a recovery plan.\n\n### Subscription Config\n\nIt is possible
          to [configure](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md)
          how OLM deploys an Operator via the `config` field in the [Subscription](https://github.com/operator-framework/olm-book/blob/master/docs/subscriptions.md)
          object.\n\nThe IBM Metrics Operator will also read the `config` and append
          it to the operands. The primary use case is to control scheduling using
          [Tolerations](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/)
          and [NodeSelectors](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/).\n\nA
          limitation is that the `config` elements are only appended to the operands.
          The elements in the operands are not removed if the `config` is removed
          from the `Subscripion`. The operand must be modified manually, or deleted
          and recreated by the controller.\n\n### Cluster permission requirements\n\n|API
          group             |Resources              |Verbs                                     |\n|----------------------|-----------------------|------------------------------------------|\n|apps
          \                 |statefulsets           |get;list;watch                            |\n|authentication.k8s.io
          |tokenreviews           |create                                    |\n|authorization.k8s.io
          \ |subjectaccessreviews   |create                                    |\n|config.openshift.io
          \  |clusterversions        |get;list;watch                            |\n|marketplace.redhat.com|meterdefinitions
          \      |get;list;watch;create;update;patch;delete |\n|operators.coreos.com
          \ |clusterserviceversions |get;list;watch                            |\n|operators.coreos.com
          \ |subscriptions          |get;list;watch                            |\n|operators.coreos.com
          \ |operatorgroups         |get;list                                  |\n|storage.k8s.io
          \       |storageclasses         |get;list;watch                            |\n|
          \                     |namespaces             |get;list;watch                            |\n|
          \                     |configmaps             |get;list;watch                            |\n|
          \                     |services               |get;list;watch                            |\n|
          \                     |**clusterrole/view**   |                                          |\n\n\n###
          Documentation\nYou can find our documentation [here.](https://swc.saas.ibm.com/en-us/documentation/)\n\n###
          Getting help\nIf you encounter any issues while using the operators, you
          can create an issue on our [Github\nrepo](https://github.com/redhat-marketplace/redhat-marketplace-operator)
          for bugs, enhancements, or other requests. You can also visit our main page
          and\nreview our [support](https://swc.saas.ibm.com/en-us/support) and [documentation](https://swc.saas.ibm.com/en-us/documentation/).\n\n###
          Readme\nYou can find our readme [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/README.md)\n\n###
          License information\nYou can find our license information [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/LICENSE)\n"
        displayName: IBM Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        - buy
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/redhat-marketplace-authcheck@sha256:13a81b8aca79ae8c89ee24dcc30f3322a4cfdedf929e7bb984295c1e7e631fa9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/redhat-marketplace-data-service@sha256:30fd8d2996b921cfce51f697ee4490a6d9a836021952429b877336f5edf2749d
        - quay.io/rh-marketplace/redhat-marketplace-reporter@sha256:2bd098f057ca07f2b14be9b60ea6ad8ca6b0cae1d226b1ca3b27e23d04ccdf20
        - quay.io/rh-marketplace/redhat-marketplace-metric-state@sha256:e4014c9a546c10e84b13a09db6a18431ee9489ad1296808635e48d9ef4963cf1
        - quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
        version: 2.20.0
      entries:
      - name: ibm-metrics-operator.v2.20.0
        version: 2.20.0
      name: beta
    - currentCSV: ibm-metrics-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MarketplaceConfig",
                "metadata": {
                  "name": "marketplaceconfig"
                },
                "spec": {
                  "features": {
                    "deployment": true,
                    "meterDefinitionCatalogServer": false,
                    "registration": true
                  },
                  "installIBMCatalogSource": true,
                  "isDisconnected": false,
                  "license": {
                    "accept": false
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterBase",
                "metadata": {
                  "name": "meterbase"
                },
                "spec": {
                  "enabled": true,
                  "prometheus": {
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "1",
                        "memory": "1G"
                      }
                    },
                    "storage": {
                      "size": "20Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterReport",
                "metadata": {
                  "name": "example-meterreport"
                },
                "spec": {
                  "endTime": "2020-08-05T00:00:00Z",
                  "prometheusService": {
                    "name": "rhm-prometheus-meterbase",
                    "namespace": "openshift-redhat-marketplace",
                    "targetPort": "rbac"
                  },
                  "startTime": "2020-08-04T00:00:00Z"
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "example-meterdefinition-4"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "meterdefinition-sample"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
          createdAt: "2025-08-28T14:18:55Z"
          description: The IBM Metrics Operator provides workload metering and reporting
            for IBM and Red Hat Marketplace customers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
          - description: MeterBase is the resource that sets up Metering for Red Hat
              Marketplace. This is an internal resource not meant to be modified directly.
            displayName: Meter Base
            kind: MeterBase
            name: meterbases.marketplace.redhat.com
            version: v1alpha1
          - description: MeterDefinition defines the meter workloads used to enable
              pay for use billing.
            displayName: Meter Definition
            kind: MeterDefinition
            name: meterdefinitions.marketplace.redhat.com
            version: v1beta1
          - kind: MeterReport
            name: meterreports.marketplace.redhat.com
            version: v1alpha1
        description: "The IBM Metrics Operator provides workload metering and reporting
          for IBM Software Central and Red Hat Marketplace customers.\n### **Important
          Note**\nA set of instructions for onboarding is provided here. For more
          detailed onboarding instructions or information about what is installed
          please visit [swc.saas.ibm.com](https://swc.saas.ibm.com).\n\nUsage metrics
          may be monitored through [https://swc.saas.ibm.com](https://swc.saas.ibm.com)
          with only IBM Metrics Operator and a Red Hat Marketplace account.\n\n###
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n### Prerequisites\n1. Installations are required
          to [enable monitoring for user-defined projects](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          as the Prometheus provider.\n2. Edit the cluster-monitoring-config ConfigMap
          object:\n\n   ```sh\n   $ oc -n openshift-monitoring edit configmap cluster-monitoring-config\n
          \   ```\n\n3. Add enableUserWorkload: true under data/config.yaml:\n  \n
          \   ```sh\n    apiVersion: v1\n    kind: ConfigMap\n    metadata:\n      name:
          cluster-monitoring-config\n        namespace: openshift-monitoring\n    data:\n
          \     config.yaml: |\n        enableUserWorkload: true\n    ```\n\n4. Configure
          the user-workload-monitoring-config ConfigMap object:\n\n    ```sh\n    $
          oc -n openshift-user-workload-monitoring edit configmap user-workload-monitoring-config\n
          \   ```\n\n5. Configure a minimum retention time of 168h and minimum storage
          capacity of 40Gi\n  \n    ```sh\n    apiVersion: v1\n    kind: ConfigMap\n
          \   metadata:\n      name: user-workload-monitoring-config\n      namespace:
          openshift-user-workload-monitoring\n\n    data:\n      config.yaml: |\n
          \       prometheus:\n          retention: 168h\n          volumeClaimTemplate:\n
          \           spec:\n              resources:\n                requests:\n
          \                 storage: 40Gi\n    ```\n\n### Resources Required\n\nMinimum
          system resources required:\n\n| Operator                | Memory (MB) |
          CPU (cores) | Disk (GB) | Nodes |\n| ----------------------- | -----------
          | ----------- | --------- | ----- |\n| **Metrics**   |        750  |     0.25
          \   | 3x1       |    3  |\n| **Deployment** |        250  |     0.25    |
          -         |    1  |\n\n| Prometheus Provider  | Memory (GB) | CPU (cores)
          | Disk (GB) | Nodes |\n| --------- | ----------- | ----------- | ---------
          | ----- |\n| **[Openshift User Workload Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)**
          |          1  |     0.1       | 2x40        |   2    |\n\nMultiple nodes
          are required to provide pod scheduling for high availability for IBM Metrics
          Operator Data Service and Prometheus.\n\nThe IBM Metrics Operator automatically
          creates 3 x 1Gi PersistentVolumeClaims to store reports as part of the data
          service, with _ReadWriteOnce_ access mode. Te PersistentVolumeClaims are
          automatically created by the ibm-metrics-operator after creating a `redhat-marketplace-pull-secret`
          and accepting the license in `marketplaceconfig`.\n\n| NAME                                |
          CAPACITY | ACCESS MODES |\n| ----------------------------------- | --------
          | ------------ |\n| rhm-data-service-rhm-data-service-0 | 1Gi | RWO |\n|
          rhm-data-service-rhm-data-service-1 | 1Gi | RWO |\n| rhm-data-service-rhm-data-service-2
          | 1Gi | RWO |\n\n### Supported Storage Providers\n\n- OpenShift Container
          Storage / OpenShift Data Foundation version 4.x, from version 4.2 or higher\n-
          IBM Cloud Block storage and IBM Cloud File storage\n- IBM Storage Suite
          for IBM Cloud Paks:\n  - File storage from IBM Spectrum Fusion/Scale \n
          \ - Block storage from IBM Spectrum Virtualize, FlashSystem or DS8K\n- Portworx
          Storage, version 2.5.5 or above\n- Amazon Elastic File Storage\n\n### Access
          Modes required\n\n - ReadWriteOnce (RWO)\n\n### Provisioning Options supported\n\nChoose
          one of the following options to provision storage for the ibm-metrics-operator
          data-service\n\n#### Dynamic provisioning using a default StorageClass\n
          \  - A StorageClass is defined with a `metadata.annotations: storageclass.kubernetes.io/is-default-class:
          \"true\"`\n   - PersistentVolumes will be provisioned automatically for
          the generated PersistentVolumeClaims\n--- \n#### Manually create each PersistentVolumeClaim
          with a specific StorageClass\n   - Must be performed before creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`. Otherwise, the automatically
          generated PersistentVolumeClaims are immutable.\n```\napiVersion: v1\nkind:
          PersistentVolumeClaim\nmetadata:\n  labels:\n    app: rhm-data-service\n
          \ name: rhm-data-service-rhm-data-service-0\n  namespace: ibm-software-central\nspec:\n
          \ storageClassName: rook-cephfs\n  accessModes:\n  - ReadWriteOnce\n  resources:\n
          \   requests:\n      storage: 1Gi\n```\n---\n#### Manually provision each
          PersistentVolume for the generated PersistentVolumeClaims with a specific
          StorageClass\n  - May be performed before or after creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`.  \n```\napiVersion: v1\nkind:
          PersistentVolume\nmetadata:\n  name: rhm-data-service-rhm-data-service-0\nspec:\n
          \ csi:\n    driver: rook-ceph.cephfs.csi.ceph.com\n    volumeHandle: rhm-data-service-rhm-data-service-0\n
          \ capacity:\n    storage: 1Gi\n  accessModes:\n    - ReadWriteOnce\n  persistentVolumeReclaimPolicy:
          Delete\n  storageClassName: rook-cephfs\n  volumeMode: Filesystem\n  claimRef:\n
          \   kind: PersistentVolumeClaim\n    namespace: ibm-software-central\n    name:
          rhm-data-service-rhm-data-service-0\n```\n\n### Installation\n1. Create
          or get your [pull secret](https://swc.saas.ibm.com/en-us/documentation/clusters#get-pull-secret).\n2.
          Install the IBM Metrics Operator\n3. Create a Kubernetes secret in the installed
          namespace with the name `redhat-marketplace-pull-secret` and key `PULL_SECRET`
          with the value of the pull secret.\n    ```sh\n    # Replace ${PULL_SECRET}
          with your pull secret\n    oc create secret generic redhat-marketplace-pull-secret
          -n ibm-software-central --from-literal=PULL_SECRET=${PULL_SECRET}\n    ```\n4.
          Use of the platform is governed by the:\n\n    [IBM Cloud Services Agreement](https://www.ibm.com/support/customer/csol/terms/?id=Z126-6304_WS&_ga=2.116312197.2046730452.1684328846-812467790.1684328846)
          (or other base agreement between you and IBM such as a [Passport Advantage
          Agreement](https://www.ibm.com/software/passportadvantage/pa_agreements.html?_ga=2.116312197.2046730452.1684328846-812467790.1684328846))
          and the [Service Description for the Red Hat Marketplace](https://www.ibm.com/support/customer/csol/terms/?id=i126-8719&_ga=2.83289621.2046730452.1684328846-812467790.1684328846).\n
          \   \n5. Update MarketplaceConfig to accept the license.\n    ```\n    oc
          patch marketplaceconfig marketplaceconfig -n ibm-software-central --type='merge'
          -p '{\"spec\": {\"license\": {\"accept\": true}}}'\n    ```\n6. Install
          the pull secret as a global pull secret on the cluster.\n\n    These steps
          require `oc`, `jq`, and `base64` to be available on your machine.\n\n    ```sh\n
          \   # Create the docker pull secret file using your pull secret.\n    #
          Store it in a file called entitledregistryconfigjson.\n    oc create secret
          docker-registry entitled-registry --docker-server=registry.marketplace.redhat.com
          --docker-username \"cp\" --docker-password \"${PULL_SECRET}\" --dry-run=client
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > entitledregistryconfigjson\n
          \   # Get the current global secrets on the cluster and store it as a file
          named dockerconfigjson\n    oc get secret pull-secret -n openshift-config
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > dockerconfigjson\n
          \   # Merge the two dockerconfigs together into a file called dockerconfigjson-merged.\n
          \   jq -s '.[0] * .[1]' dockerconfigjson entitledregistryconfigjson > dockerconfigjson-merged\n
          \   # Set the cluster's dockerconfig file to the new merged version.\n    oc
          set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson=dockerconfigjson-merged\n
          \   ```\n  7. (Optional) Your IBM product may have installed IBM Metrics
          Operator using an `ibm-entitlement-key` for authentication to IBM Software
          Central. To authorize the automatic creation of an IBM Software Central
          account for the IBM account owning the `ibm-entitlement-key`\n      ```\n
          \     oc patch marketplaceconfig marketplaceconfig -n ibm-software-central
          --type='merge' -p '{\"spec\": {\"authorizeAccountCreation\": true}}'\n      ```\n\n###
          Why is a global pull secret required?\nIn order to successfully install
          the products hosted by the container image registry, you will need to make
          the pull secret available across the cluster. This can be achieved by applying
          the pull as a [global pull secret](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html#images-update-global-pull-secret_using-image-pull-secrets).
          For alternative approaches, please see the official OpenShift [documentation](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html).\n\n\n###
          SecurityContextConstraints requirements\n\nThe Operators and their components
          support running under the OpenShift Container Platform default restricted
          and restricted-v2 security context constraints.\n\n### Installation Namespace
          and ClusterRoleBinding requirements\n\nThe IBM Metrics Operator components
          require specific ClusterRoleBindings.\n- The metric-state component requires
          a ClusterRoleBinding for the the `view` ClusterRole.\n  - Ability to read
          non-sensitive CustomResources. \n  - The `view` ClusterRole dynamically
          adds new CustomResourceDefinitions.\n- The operator & reporter component
          requires a ClusterRoleBinding for the the `cluster-monitoring-view` ClusterRole.\n
          \ - Ability to view Prometheus metrics. \n  - The underlying ClusterRole
          RBAC often updated between OpenShift versions.\n\nDue to limitations of
          Operator Lifecycle Manager (OLM), this ClusterRoleBinding can not be provided
          dynamically for arbitrary installation target namespaces.\n\nA static ClusterRoleBinding
          is included for installation to the default namespace of `ibm-software-central`,
          and namespaces `openshift-redhat-marketplace`,  `redhat-marketplace`, `ibm-common-services`.\n\nTo
          create the ClusterRoleBindings for installation to an alternate namespace\n```\noc
          project INSTALL-NAMESPACE\noc adm policy add-cluster-role-to-user view -z
          ibm-metrics-operator-metric-state\noc adm policy add-cluster-role-to-user
          cluster-monitoring-view -z ibm-metrics-operator-controller-manager,ibm-metrics-operator-reporter\n```\n\n###
          Metric State scoping requirements\nThe metric-state Deployment obtains `get/list/watch`
          access to metered resources via the `view` ClusterRole. For operators deployed
          using Operator Lifecycle Manager (OLM), permissions are added to `clusterrole/view`
          dynamically via a generated and annotated `-view` ClusterRole. If you wish
          to meter an operator, and its Custom Resource Definitions (CRDs) are not
          deployed through OLM, one of two options are required\n1. Add the following
          label to a clusterrole that has get/list/watch access to your CRD: `rbac.authorization.k8s.io/aggregate-to-view:
          \"true\"`, thereby dynamically adding it to `clusterrole/view`\n2. Create
          a ClusterRole that has get/list/watch access to your CRD, and create a ClusterRoleBinding
          for the metric-state ServiceAccount\n\nAttempting to meter a resource with
          a MeterDefinition without the required permissions will log an `AccessDeniedError`
          in metric-state.\n\n### Disaster Recovery\n\nTo plan for disaster recovery,
          note the PhysicalVolumeClaims `rhm-data-service-rhm-data-service-N`. \n-
          In connected environments, MeterReport data upload attempts occur hourly,
          and are then removed from data-service. There is a low risk of losing much
          unreported data.\n- In an airgap environment, MeterReport data must be pulled
          from data-service and uploaded manually using [datactl](https://github.com/redhat-marketplace/datactl/).
          To prevent data loss in a disaster scenario, the data-service volumes should
          be considered in a recovery plan.\n\n### Subscription Config\n\nIt is possible
          to [configure](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md)
          how OLM deploys an Operator via the `config` field in the [Subscription](https://github.com/operator-framework/olm-book/blob/master/docs/subscriptions.md)
          object.\n\nThe IBM Metrics Operator will also read the `config` and append
          it to the operands. The primary use case is to control scheduling using
          [Tolerations](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/)
          and [NodeSelectors](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/).\n\nA
          limitation is that the `config` elements are only appended to the operands.
          The elements in the operands are not removed if the `config` is removed
          from the `Subscripion`. The operand must be modified manually, or deleted
          and recreated by the controller.\n\n### Cluster permission requirements\n\n|API
          group             |Resources              |Verbs                                     |\n|----------------------|-----------------------|------------------------------------------|\n|apps
          \                 |statefulsets           |get;list;watch                            |\n|authentication.k8s.io
          |tokenreviews           |create                                    |\n|authorization.k8s.io
          \ |subjectaccessreviews   |create                                    |\n|config.openshift.io
          \  |clusterversions        |get;list;watch                            |\n|marketplace.redhat.com|meterdefinitions
          \      |get;list;watch;create;update;patch;delete |\n|operators.coreos.com
          \ |clusterserviceversions |get;list;watch                            |\n|operators.coreos.com
          \ |subscriptions          |get;list;watch                            |\n|operators.coreos.com
          \ |operatorgroups         |get;list                                  |\n|storage.k8s.io
          \       |storageclasses         |get;list;watch                            |\n|
          \                     |namespaces             |get;list;watch                            |\n|
          \                     |configmaps             |get;list;watch                            |\n|
          \                     |services               |get;list;watch                            |\n|
          \                     |**clusterrole/view**   |                                          |\n\n\n###
          Documentation\nYou can find our documentation [here.](https://swc.saas.ibm.com/en-us/documentation/)\n\n###
          Getting help\nIf you encounter any issues while using the operators, you
          can create an issue on our [Github\nrepo](https://github.com/redhat-marketplace/redhat-marketplace-operator)
          for bugs, enhancements, or other requests. You can also visit our main page
          and\nreview our [support](https://swc.saas.ibm.com/en-us/support) and [documentation](https://swc.saas.ibm.com/en-us/documentation/).\n\n###
          Readme\nYou can find our readme [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/README.md)\n\n###
          License information\nYou can find our license information [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/LICENSE)\n"
        displayName: IBM Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        - buy
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/redhat-marketplace-metric-state@sha256:e4014c9a546c10e84b13a09db6a18431ee9489ad1296808635e48d9ef4963cf1
        - quay.io/rh-marketplace/redhat-marketplace-authcheck@sha256:13a81b8aca79ae8c89ee24dcc30f3322a4cfdedf929e7bb984295c1e7e631fa9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/redhat-marketplace-data-service@sha256:30fd8d2996b921cfce51f697ee4490a6d9a836021952429b877336f5edf2749d
        - quay.io/rh-marketplace/redhat-marketplace-reporter@sha256:2bd098f057ca07f2b14be9b60ea6ad8ca6b0cae1d226b1ca3b27e23d04ccdf20
        - quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
        version: 2.20.0
      entries:
      - name: ibm-metrics-operator.v2.20.0
        version: 2.20.0
      name: stable
    defaultChannel: stable
    packageName: ibm-metrics-operator
    provider:
      name: Red Hat Marketplace
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-dr-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-dr-manager.v8.0.220
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com",
                  "resRecord": {
                    "HostName_1": "100.200.30.41",
                    "HostName_2": "100.200.30.43",
                    "abc": "HostName_2",
                    "www": "HostName_1",
                    "xyz": "HostName_3"
                  },
                  "stealthMasters": [
                    "1.2.3.4"
                  ],
                  "tsigKeyFile": "/Kvcslnxdns.com.+157+10641.private"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "dataReplicationPointer": "test-datareplication",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 10,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          olm.skipRange: <8.0.220
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
            - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:a98f9cc51f2c7baeec6dbb4933a42ae8ae20ff1a42c9d48a37174cf5211bbcf8
        version: 8.0.220
      entries:
      - name: infoscale-dr-manager.v8.0.220
        version: 8.0.220
      - name: infoscale-dr-manager.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-dr-manager.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          createdAt: "2024-06-17T10:14:13Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:c46cbf78631fc790c77b6c13c1d1d4b5ab35f49164f7c2e36c7ec59220cbdce2
        version: 8.0.330
      entries:
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-dr-manager.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
          createdAt: "2024-02-22T14:14:42Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721648-168743109)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721654-168743109)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
        version: 8.0.410
      entries:
      - name: infoscale-dr-manager.v8.0.410
        version: 8.0.410
      - name: infoscale-dr-manager.v8.0.409
        version: 8.0.409
      - name: infoscale-dr-manager.v8.0.400
        version: 8.0.400
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-dr-manager.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
          createdAt: "2024-02-22T14:14:42Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721648-168743109)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721654-168743109)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
        version: 8.0.410
      entries:
      - name: infoscale-dr-manager.v8.0.410
        version: 8.0.410
      - name: infoscale-dr-manager.v8.0.409
        version: 8.0.409
      - name: infoscale-dr-manager.v8.0.400
        version: 8.0.400
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-dr-manager.v8.0.100
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com",
                  "resRecord": {
                    "abc": "r7515-054-vm9",
                    "r7515-054-vm10": "10.221.85.83",
                    "r7515-054-vm8": "10.221.85.81",
                    "r7515-054-vm9": "10.221.85.82",
                    "www": "r7515-054-vm8",
                    "xyz": "r7515-054-vm10"
                  },
                  "stealthMasters": [
                    "1.2.3.4"
                  ],
                  "tsigKeyFile": "/Kvcslnxdns.com.+157+10641.private"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "currentPrimary": "Clus1",
                  "force": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "fail",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "dataReplicationPointer": "test-datareplication",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 10,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:eb999187a1f9a9ce5566089e24adf598b1c229d31c5e57ddcfe4439ba25e97dc
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
            - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membershrip
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membershrip
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponsding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:eb999187a1f9a9ce5566089e24adf598b1c229d31c5e57ddcfe4439ba25e97dc
        version: 8.0.100
      entries:
      - name: infoscale-dr-manager.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-dr-manager
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-s3control-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-s3control-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/s3control-controller:1.1.1
          createdAt: "2025-09-19T20:52:02Z"
          description: AWS S3 Control controller is a service controller for managing
            S3 Control resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessPoint represents the state of an AWS s3control AccessPoint
              resource.
            displayName: AccessPoint
            kind: AccessPoint
            name: accesspoints.s3control.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon S3 Control resources in AWS from within your Kubernetes cluster.

          **About Amazon S3 Control**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon S3 Control
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - s3control
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon S3 Control Developer Resources
          url: https://aws.amazon.com/S3 Control/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: s3control maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/s3control-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-s3control-controller.v1.1.1
        version: 1.1.1
      - name: ack-s3control-controller.v1.1.0
        version: 1.1.0
      - name: ack-s3control-controller.v1.0.11
        version: 1.0.11
      - name: ack-s3control-controller.v1.0.10
        version: 1.0.10
      name: alpha
    defaultChannel: alpha
    packageName: ack-s3control-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: sosivio
      provider-url: https://www.sosiv.io
    name: sosivio
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sosivio-operator.v1.5.1-97
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sosivio.sosiv.io/v1alpha1",
                "kind": "Sosivio",
                "metadata": {
                  "name": "sosivio"
                },
                "spec": {
                  "platform": "ocp4",
                  "useSha": false
                }
              },
              {
                "apiVersion": "sosivio.sosiv.io/v1alpha1",
                "kind": "Sosivio",
                "metadata": {
                  "name": "sosivio"
                },
                "spec": {
                  "platform": "ocp4",
                  "useSha": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Monitoring,Integration & Delivery,Logging & Tracing,Developer
            Tools
          com.redhat.openshift.versions: v4.1-v4.1
          containerImage: quay.io/sosivio/sosivio-operator:v1.5.1-97
          createdAt: "2021-10-28 01:01:01"
          description: Sosivio | Predictive Troubleshooting for Kubernetes
          operatorhub.io/ui-metadata-max-k8s-version: "1.25"
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Sosivio/operator
          support: Sosivio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Sosivio is the Schema for the sosivios API
            displayName: Sosivio
            kind: Sosivio
            name: sosivios.sosivio.sosiv.io
            version: v1alpha1
        description: "# Sosivio Operator\nThe Sosivio Operator for RedHat’s Openshift
          Container Platform allows users to quickly and easily deploy Sosivio’s Kubernetes
          Predictive Troubleshooting Platform on your cluster.\n\n# Sosivio Overview\n\nSosivio
          is the first and only non-intrusive and predictive autonomous cloud-native
          platform built specifically for Kubernetes from the ground up. Sosivio provides
          predictive troubleshooting tools, root cause determination and instant remediation
          for any failures in Kubernetes environments.\n\nPredict and Prevent!\n\nThe
          Operator will deploy and manage Sosivio on Kubernetes and OpenShift. Included
          are the following features;\n-   **Root Cause Analysis and Remediation**
          - predictive troubleshooting to prevent failures before they occur. Instantly
          understand where the root cause is and how to fix it.\n-   **Application
          Profiling** - analyze container resource consumption and recommendations
          for requests and limits.\n-   **Real-time Metrics** - in-depth, granular,
          and accurate real-time data from the process level inside containers (and
          pods) with failure correlation.\n-   **Cluster Health Checks** - proactive
          health checks for cluster infrastructure and applications running on the
          cluster.\n-   **Multi-Cluster View** - management, monitoring, and maintenance
          capabilities for all clusters from one location. \n-   **Kubernetes Components
          Alerting** - notifications sent through Slack, email and other 3rd party
          platforms.\n\n# Installation\nSosivio is completely **non-intrusive** and
          **does not require any persistent storage**. There is **no data offloading**
          outside of your Kubernetes cluster. Applying our Data Swirling methodology
          enables the product to perform at a **remarkably low resource footprint**
          and requires **no persistent volume configurations**.\n\n\nSosivio is installed
          under one namespace (labeled: \"sosivio\") which creates all the required
          components for the product.\n\nAll users will start with 4 weeks of Sosivio
          Premium after which Sosivio will automatically convert to the Community
          Version (free forever).\n\n# Help\nMore information about Sosivio and our
          product usage instructions can be found in our documentation [here](https://docs.sosiv.io/)
          \ \n"
        displayName: Sosivio | Predictive Troubleshooting for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sosivio
        - kubernetes
        - observanility
        - monitoring
        - root cause
        links:
        - name: Sosivio
          url: https://www.sosiv.io
        - name: Documentation
          url: https://docs.sosiv.io
        maintainers:
        - email: info@sosiv.io
          name: Sosivio
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: sosivio
          url: https://www.sosiv.io
        relatedImages:
        - quay.io/sosivio/discovery-engine@sha256:b87b3f7c9ed3aed80d0a77788fd8925789450d9dc05d25017e89dff611cd82da
        - quay.io/sosivio/nsq@sha256:781aac23ef0a1ed0af2b8e93d625ae49bd11dbdd98787d3c79dde8d128f3810d
        - quay.io/sosivio/actuator@sha256:a9abe9730800ca2b0a1bbd7a361b2ab3a818cbf127798314e5477fb1cb16ba4c
        - quay.io/sosivio/sosivio-node-collector@sha256:61ed8dd52de5edc09d16a31b18f2f65ad9344c0b6c9cfab323c3eb3d7e7ed060
        - quay.io/sosivio/crud-manager@sha256:38289a28829fa737917d76dd4810ba6140e123061ead0a4b79fa79b2e246a293
        - quay.io/sosivio/sosivio-operator:v1.5.1-97
        - quay.io/sosivio/sosivio-dashboard@sha256:fdfca5bb2123d1008cec2b656cde73c7e9364f84c5ff4f4a6fee5751b1bfb2f8
        - quay.io/sosivio/sosivio-node-netpacket@sha256:84dd159b1dfdb52dee3eb1c4c93f323bae49ce517ff44663b69978e14bdd8390
        - quay.io/sosivio/classifier@sha256:dd8bf145524db23ff333efa1bcdff258dab116d3534f95f73775d8dd339e6963
        - quay.io/sosivio/communicator@sha256:fffec9a6c66594fde5f05e2dfb84361544ca9d53a345188c197a0ea6f53bc4b9
        - quay.io/sosivio/authentication@sha256:0ea6025f333576792f5b1fc4d96d9ab1a404749f185830410e71248683004da2
        - quay.io/sosivio/sosiviodb@sha256:a3c6c3975bc2b0316135ee5c547f1eb215f3a1f9c8b0a9c8eceff23415f93597
        - quay.io/sosivio/kube-rbac-proxy:v0.8.0
        - quay.io/sosivio/sequence-recognition@sha256:7335cf8f3e40bc46e8ca0de525d99f0655833b7c52d36dc273aa7b096eb70fa2
        - quay.io/sosivio/network-aggregator@sha256:e59d937f4d9c52acad0adadab39c6849d20b4ee72a83e239a33ea137f08323f4
        - quay.io/sosivio/correlation-entities@sha256:c00af09b2589bad4740fd44d7a81c69b2e29138c360939b2a45f65edd44b6901
        - quay.io/sosivio/contract-testing@sha256:cad6fac6ed3fdf2597ceb09ba5dbb7d81f70cf45f483a7064e3968215dd982e3
        - quay.io/sosivio/kube-rbac-proxy@sha256:34e8724e0f47e31eb2ec3279ac398b657db5f60f167426ee73138e2e84af6486
        - quay.io/sosivio/analyzer@sha256:3825608fe7c43ef059108e76665acc99fe72bea77d4d9e3852f6ad45e74379b1
        - quay.io/sosivio/notifier@sha256:35f5f3e8e5ccad914fcfe34fb4b4a1b044ee87f5a6141284cd5273f0b2b3bf22
        - quay.io/sosivio/sosivio-node-pinger@sha256:937de52ad7b76e6cff4ad5197de9705988d61f285ebadac5ebadb5856f59ee58
        - quay.io/sosivio/sosivio-operator@sha256:11bc627773bb5e08b629e0b013e7f600fcb4aaefbd97f2de337d339475616763
        version: 1.5.1
      entries:
      - name: sosivio-operator.v1.5.1-97
        version: 1.5.1
      - name: sosivio-operator.v1.5.0-95
        version: 1.5.0
      - name: sosivio-operator.v1.4.4-85
        version: 1.4.4
      - name: sosivio-operator.v1.4.3-83
        version: 1.4.3
      - name: sosivio-operator.v1.4.2-82
        version: 1.4.2
      - name: sosivio-operator.v1.4.1-81
        version: 1.4.1-81
      - name: sosivio-operator.v1.4.1.1-80
        version: 1.4.1-80
      - name: sosivio-operator.v1.4.1-78
        version: 1.4.1-79
      - name: sosivio-operator.v1.3.1-76
        version: 1.3.1-76
      - name: sosivio-operator.v1.3.0-75
        version: 1.3.0-75
      - name: sosivio-operator.v1.2.1-1
        version: 1.2.1-1
      - name: sosivio-operator.v1.2.0-1
        version: 1.2.0-1
      name: stable
    defaultChannel: stable
    packageName: sosivio
    provider:
      name: sosivio
      url: https://www.sosiv.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-keycloak-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-keycloak-operator.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2023-09-19T12:43:29Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KCOSClientRoleMapping
            name: kcosclientrolemappings.keycloak.org
            version: v1alpha1
          - kind: KCOSRealmRole
            name: kcosrealmroles.keycloak.org
            version: v1alpha1
          - kind: KCOSUserRoleMapping
            name: kcosuserrolemappings.keycloak.org
            version: v1alpha1
          - kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: Keysight Keycloak Operator
        displayName: Keysight Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Keycloak Operator
          url: https://keysight-keycloak-operator.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak@sha256:772f73cce6a77b9c46f8035faac99a28602219657d18b75b19b4d905d91942c9
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak-postgresql@sha256:b636955199065e18f9d4679835ae78214843258e5abb4ae758288d6f00df2f8e
        - prod.harbor.keysight.digital/loadcore/keysight-keycloak-operator@sha256:d74e7e98a114768a6f3327edabce9336743a078844fffee985053b6a90100da8
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak-init-container@sha256:90965beaf41c760cb8c8bf906acf08f8fe4770d313de8dfd92d18d7c25348a38
        version: 0.1.21
      entries:
      - name: keysight-keycloak-operator.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-keycloak-operator
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: fence-agents-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fence-agents-remediation.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  }
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:00cc420250d8e25d9da7c43d9c4cff35c7edeb7d390af754c48c0ce00e9e7770
          createdAt: "2023-10-30 14:06:57"
          description: Fence Agents Remediation Operator for remediating nodes using
            upstream fence-agents.
          olm.skipRange: '>=0.2.0 <0.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: "Fence Agents Remediation Operator (FAR)\n\nThe operator will
          remediate/fence a node when a FenceAgentsRemediation (far) custom resource
          exists with the node's name.\nFAR uses an existing set of upstream fencing
          agents for environments with a \ntraditional API end-point (e.g., IPMI)
          for power cycling cluster nodes.\n\nThe operator is recommended to be used
          with Node Healthcheck Operator for easier and smoother experience,\nbut
          it can also work as a standalonde remediator for the more advanced user.\n"
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:00cc420250d8e25d9da7c43d9c4cff35c7edeb7d390af754c48c0ce00e9e7770
        version: 0.2.1
      entries:
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.14-eus
    - currentCSV: fence-agents-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:a367e8fdd80938e97ac195631a996e0c537bf3666aef0199147534762dcf739f
          createdAt: "2024-08-10 22:49:23"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.2.0 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:a367e8fdd80938e97ac195631a996e0c537bf3666aef0199147534762dcf739f
        version: 0.4.1
      entries:
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.16-eus
    - currentCSV: fence-agents-remediation.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-16 10:25:57"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.5.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.5/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel9-operator@sha256:da5975557078fbfa6882df87cff5bda5671b9838b5fb0a738fea724469d03576
        version: 0.5.1
      entries:
      - name: fence-agents-remediation.v0.5.1
        version: 0.5.1
      - name: fence-agents-remediation.v0.5.0
        version: 0.5.0
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.18-eus
    - currentCSV: fence-agents-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  }
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:5c3778c7aae45f54e949fbae3d133777f7aa78fe1251ac3d5510aa28b71e2ba4
          createdAt: "2023-04-20 15:36:05"
          description: Fence Agents Remediation Operator for remediating nodes using
            upstream fence-agents.
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: "Fence Agents Remediation Operator (FAR)\n\nThe operator will
          remediate/fence a node when a FenceAgentsRemediation (far) custom resource
          exists with the node's name.\nFAR uses an existing set of upstream fencing
          agents for environments with a \ntraditional API end-point (e.g., IPMI)
          for power cycling cluster nodes.\n\nThe operator is recommended to be used
          with Node Healthcheck Operator for easier and smoother experience,\nbut
          it can also work as a standalonde remediator for the more advanced user.\n"
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3af07bfe0fb73b1a7f00487eba302d6da442e742ab4007fc50e52efd4d45dae5
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:5c3778c7aae45f54e949fbae3d133777f7aa78fe1251ac3d5510aa28b71e2ba4
        version: 0.1.0
      entries:
      - name: fence-agents-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: fence-agents-remediation.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeSecretNames": {
                    "master-0": "fence-agents-credentials-master0",
                    "master-1": "fence-agents-credentials-master1",
                    "master-2": "fence-agents-credentials-master2",
                    "worker-0": "fence-agents-credentials-worker0",
                    "worker-1": "fence-agents-credentials-worker1",
                    "worker-2": "fence-agents-credentials-worker2"
                  },
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedSecretName": "fence-agents-credentials-shared",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {
                    "spec": {
                      "agent": "fence_ipmilan",
                      "nodeSecretNames": {
                        "master-0": "fence-agents-credentials-master0",
                        "master-1": "fence-agents-credentials-master1",
                        "master-2": "fence-agents-credentials-master2",
                        "worker-0": "fence-agents-credentials-worker0",
                        "worker-1": "fence-agents-credentials-worker1",
                        "worker-2": "fence-agents-credentials-worker2"
                      },
                      "nodeparameters": {
                        "--ipport": {
                          "master-0": "6230",
                          "master-1": "6231",
                          "master-2": "6232",
                          "worker-0": "6233",
                          "worker-1": "6234",
                          "worker-2": "6235"
                        }
                      },
                      "remediationStrategy": "ResourceDeletion",
                      "retrycount": 5,
                      "retryinterval": "5s",
                      "sharedSecretName": "fence-agents-credentials-shared",
                      "sharedparameters": {
                        "--action": "reboot",
                        "--ip": "192.168.111.1",
                        "--lanplus": "",
                        "--username": "admin"
                      },
                      "timeout": "60s"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-09-08 09:19:00"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.5.0 <0.6.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.8/html/remediation_fencing_and_maintenance/about-remediation-fencing-maintenance
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel9-operator@sha256:eef1a298718650a22a3da4a07c140ce21b6f287d87af79801610317fa8f63d52
        version: 0.6.0
      entries:
      - name: fence-agents-remediation.v0.6.0
        version: 0.6.0
      - name: fence-agents-remediation.v0.5.1
        version: 0.5.1
      - name: fence-agents-remediation.v0.5.0
        version: 0.5.0
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: stable
    defaultChannel: stable
    packageName: fence-agents-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ssm-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ssm-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ssm-controller:1.1.1
          createdAt: "2025-09-19T20:57:07Z"
          description: AWS SSM controller is a service controller for managing SSM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Document represents the state of an AWS ssm Document resource.
            displayName: Document
            kind: Document
            name: documents.ssm.services.k8s.aws
            version: v1alpha1
          - description: PatchBaseline represents the state of an AWS ssm PatchBaseline
              resource.
            displayName: PatchBaseline
            kind: PatchBaseline
            name: patchbaselines.ssm.services.k8s.aws
            version: v1alpha1
          - description: ResourceDataSync represents the state of an AWS ssm ResourceDataSync
              resource.
            displayName: ResourceDataSync
            kind: ResourceDataSync
            name: resourcedatasyncs.ssm.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon SSM resources in AWS from within your Kubernetes cluster.

          **About Amazon SSM**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon SSM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ssm
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SSM Developer Resources
          url: https://aws.amazon.com/SSM/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ssm maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ssm-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ssm-controller.v1.1.1
        version: 1.1.1
      - name: ack-ssm-controller.v1.1.0
        version: 1.1.0
      - name: ack-ssm-controller.v1.0.12
        version: 1.0.12
      - name: ack-ssm-controller.v1.0.10
        version: 1.0.10
      - name: ack-ssm-controller.v1.0.9
        version: 1.0.9
      - name: ack-ssm-controller.v1.0.8
        version: 1.0.8
      - name: ack-ssm-controller.v1.0.7
        version: 1.0.7
      - name: ack-ssm-controller.v1.0.6
        version: 1.0.6
      - name: ack-ssm-controller.v1.0.5
        version: 1.0.5
      - name: ack-ssm-controller.v1.0.4
        version: 1.0.4
      - name: ack-ssm-controller.v1.0.3
        version: 1.0.3
      - name: ack-ssm-controller.v1.0.2
        version: 1.0.2
      - name: ack-ssm-controller.v1.0.1
        version: 1.0.1
      - name: ack-ssm-controller.v1.0.0
        version: 1.0.0
      - name: ack-ssm-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-ssm-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-efs-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-efs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "AccessPoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "FileSystem",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "MountTarget",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/efs-controller:1.1.1
          createdAt: "2025-09-19T18:38:13Z"
          description: AWS EFS controller is a service controller for managing EFS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessPoint represents the state of an AWS efs AccessPoint
              resource.
            displayName: AccessPoint
            kind: AccessPoint
            name: accesspoints.efs.services.k8s.aws
            version: v1alpha1
          - description: FileSystem represents the state of an AWS efs FileSystem
              resource.
            displayName: FileSystem
            kind: FileSystem
            name: filesystems.efs.services.k8s.aws
            version: v1alpha1
          - description: MountTarget represents the state of an AWS efs MountTarget
              resource.
            displayName: MountTarget
            kind: MountTarget
            name: mounttargets.efs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EFS resources in AWS from within your Kubernetes cluster.

          **About Amazon Elastic File System (Amazon EFS)**

          Amazon Elastic File System (Amazon EFS) provides serverless, fully elastic file storage so that you can share file data without provisioning or managing storage capacity and performance. Amazon EFS is built to scale on demand to petabytes without disrupting applications, growing and shrinking automatically as you add and remove files. Because Amazon EFS has a simple web services interface, you can create and configure file systems quickly and easily. The service manages all the file storage infrastructure for you, meaning that you can avoid the complexity of deploying, patching, and maintaining complex file system configurations.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Elastic File System (Amazon
          EFS)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - efs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EFS Developer Resources
          url: https://aws.amazon.com/efs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: efs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/efs-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-efs-controller.v1.1.1
        version: 1.1.1
      - name: ack-efs-controller.v1.1.0
        version: 1.1.0
      - name: ack-efs-controller.v1.0.19
        version: 1.0.19
      - name: ack-efs-controller.v1.0.18
        version: 1.0.18
      - name: ack-efs-controller.v1.0.17
        version: 1.0.17
      - name: ack-efs-controller.v1.0.16
        version: 1.0.16
      - name: ack-efs-controller.v1.0.15
        version: 1.0.15
      - name: ack-efs-controller.v1.0.14
        version: 1.0.14
      - name: ack-efs-controller.v1.0.13
        version: 1.0.13
      - name: ack-efs-controller.v1.0.12
        version: 1.0.12
      - name: ack-efs-controller.v1.0.11
        version: 1.0.11
      - name: ack-efs-controller.v1.0.10
        version: 1.0.10
      - name: ack-efs-controller.v1.0.8
        version: 1.0.8
      - name: ack-efs-controller.v1.0.6
        version: 1.0.6
      - name: ack-efs-controller.v1.0.5
        version: 1.0.5
      - name: ack-efs-controller.v1.0.3
        version: 1.0.3
      - name: ack-efs-controller.v1.0.2
        version: 1.0.2
      - name: ack-efs-controller.v1.0.0
        version: 1.0.0
      - name: ack-efs-controller.v0.0.9
        version: 0.0.9
      - name: ack-efs-controller.v0.0.8
        version: 0.0.8
      - name: ack-efs-controller.v0.0.7
        version: 0.0.7
      - name: ack-efs-controller.v0.0.6
        version: 0.0.6
      - name: ack-efs-controller.v0.0.5
        version: 0.0.5
      - name: ack-efs-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-efs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solace Corporation
      provider-url: www.solace.com
    name: pubsubplus-eventbroker-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pubsubplus-eventbroker-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pubsubplus.solace.com/v1beta1",
                "kind": "PubSubPlusEventBroker",
                "metadata": {
                  "name": "non-ha-standalone-example"
                },
                "spec": {
                  "redundancy": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "true"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.10
          containerImage: registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator:1.4.0
          createdAt: "2025-07-31T19:48:36Z"
          description: The Solace PubSub+ Event Broker Operator deploys and manages
            the lifecycle of PubSub+ Event Brokers
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
          support: Solace Products
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PubSub+ Event Broker
            displayName: PubSub+ Event Broker
            kind: PubSubPlusEventBroker
            name: pubsubpluseventbrokers.pubsubplus.solace.com
            version: v1beta1
        description: |
          ## Solace PubSub+ Platform
          Solace [PubSub+ Platform](https://solace.com/products/platform/) is a complete event streaming and management platform
          for real-time enterprises. The [PubSub+ Software Event Broker](https://solace.com/products/event-broker/software/) efficiently streams
          event-driven information between applications, IoT devices, and user interfaces running in the cloud, on-premises,
          and in hybrid environments using open APIs and protocols like AMQP, JMS, MQTT, REST and WebSocket.
          It can be installed into a variety of public and private clouds, PaaS, and on-premises environments.
          Event brokers in multiple locations can be linked together in an [Event Mesh](https://solace.com/what-is-an-event-mesh/)
          to dynamically share events across the distributed enterprise.

          ## Solace PubSub+ Event Broker Operator
          Solace Pubsub+ Event Broker Operator automatically deploys and manages PubSub+ Software Event Brokers on Kubernetes and OpenShift environments.

          Features include:
          * Initial deployment configuration: Production-ready HA, non-HA or for developers
          * Rolling update for configuration changes or broker upgrades
          * Prometheus monitoring support

          ## Getting started
          Refer to the [Solace PubSub+ Event Broker Operator Quick Start](https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/README.md)
        displayName: Solace PubSub+ Event Broker Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - solace
        - pubsubplus
        - pubsub+
        - pubsub
        - messaging
        - advanced event broker
        - event broker
        - event mesh
        - event streaming
        - data streaming
        - event integration
        - middleware
        links:
        - name: Operator QuickStart
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
        - name: Operator Documentation
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/docs/EventBrokerOperatorUserGuide.md
        - name: Solace.com
          url: https://solace.com
        - name: Understanding Solace Technology
          url: https://solace.com/products/platform/
        - name: Solace PubSub+ Technical Documentation
          url: https://docs.solace.com/
        - name: Solace Developer Portal
          url: https://www.solace.dev/
        - name: Solace Community
          url: https://solace.community/
        maintainers:
        - email: support@solace.com
          name: Solace R&D
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Solace Corporation
          url: www.solace.com
        relatedImages:
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator:1.4.0
        version: 1.4.0
      entries:
      - name: pubsubplus-eventbroker-operator.v1.4.0
        version: 1.4.0
      - name: pubsubplus-eventbroker-operator.v1.3.0
        version: 1.3.0
      - name: pubsubplus-eventbroker-operator.v1.2.0
        version: 1.2.0
      - name: pubsubplus-eventbroker-operator.v1.1.2
        version: 1.1.2
      - name: pubsubplus-eventbroker-operator.v1.1.1
        version: 1.1.1
      - name: pubsubplus-eventbroker-operator.v1.1.0
        version: 1.1.0
      - name: pubsubplus-eventbroker-operator.v1.0.3
        version: 1.0.3
      - name: pubsubplus-eventbroker-operator.v1.0.2
        version: 1.0.2
      - name: pubsubplus-eventbroker-operator.v1.0.1
        version: 1.0.1
      - name: pubsubplus-eventbroker-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pubsubplus-eventbroker-operator
    provider:
      name: Solace Corporation
      url: www.solace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: ibm-data-reporter-operator
      app.kubernetes.io/managed-by: ibm-data-reporter-operator
      app.kubernetes.io/name: ibm-data-reporter-operator
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: ""
    name: ibm-data-reporter-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-data-reporter-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "DataReporterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-data-reporter-operator",
                    "app.kubernetes.io/instance": "datareporterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datareporterconfig",
                    "app.kubernetes.io/part-of": "ibm-data-reporter-operator"
                  },
                  "name": "datareporterconfig"
                },
                "spec": {
                  "userConfig": [
                    {
                      "metadata": {
                        "ameta1": "ametadata1",
                        "bmeta1": "bmetadata1",
                        "cmeta1": "cmetadata1",
                        "dmeta1": "dmetadata1"
                      },
                      "userName": "system:serviceaccount:redhat-marketplace:ibm-data-reporter-operator-api"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
          createdAt: ""
          description: The IBM Data Reporter Operator accepts events and transforms
            them into reports submitted to the Data Service of the IBM Metrics Operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReporterConfig is the Schema for the datareporterconfigs
              API
            displayName: Data Reporter Config
            kind: DataReporterConfig
            name: datareporterconfigs.marketplace.redhat.com
            version: v1alpha1
          required:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
        description: "# IBM Data Reporter Operator\n\n# Introduction\n\nThe IBM Data
          Reporter Operator accepts events and transforms them into reports submitted
          to the IBM Metrics Operator Data Service.\n\n# Details\n\nThe IBM Data Reporter
          Operator deploys a service that provides an endpoint to which callers can
          send raw json event data. The event data is transformed into a report and
          is sent to the IBM Metrics Operator's Data Service. The IBM Metrics Operator
          Data Service periodically uploads the reports to IBM Software Central. It
          can also filter, transform and forward events to alternate endpoints.\n\n##
          Prerequisites\n\n- OpenShift Container Platform, major version 4 with any
          available supported minor version\n- Install IBM Metrics Operator\n  - IBM
          Data Reporter Operator prerequisties the IBM Metrics Operator's Data Service\n
          \ - Register the Cluster by creating a `redhat-marketplace-pull-secret`,
          as per the instructions\n  - Once `rhm-data-service` has started IBM Data
          Reporter Operator will accept requests\n\n## SecurityContextConstraints
          Requirements\n\n- The operator runs under Red Hat restricted SCC\n\n## Resources
          Required\n\n- The operator requires at least 85 Mi memory.\n\n## Limitations\n\n-
          Only runs on amd64, s390x and ppc64le architectures\n\n## Installing\n\n-
          A user with the cluster administrator role.\n- Install this operator in
          the same namespace as the IBM Metrics Operator\n  - default namespace: `ibm-software-central`\n\n##
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n# Configuration\n\nOptional:\n- Configure the
          Custom Resource Definition `DataReporterConfig` named `datareporterconfig`
          as per the following sample\n  - Configuration that can not be reconciled
          successfully will be reported on the DataReporterConfig Status\n  - jsonPath
          expressions are handled by: https://github.com/ohler55/ojg\n    - jsonPath
          comparison: https://cburgmer.github.io/json-path-comparison/\n  - Transformer
          type currently supported is `kazaam`\n    - https://github.com/willie68/kazaam\n\nSample
          DataReporterConfig:\n```YAML\napiVersion: marketplace.redhat.com/v1alpha1\nkind:
          DataReporterConfig\nmetadata:\n  name: datareporterconfig\nspec:\n  # true:
          report generated for 1 event and provide return code from data-service,
          confirming delivery\n  # false: events will be accumulated before building
          report, and immediately return 200, delivery unfconfirmed\n  confirmDelivery:
          false\n  dataFilters:   # dataFilters matches the first dataFilter selector,
          applies transforms and forwards to data-service and destinations\n  - altDestinations:
          \  # List of desintations to transform and foward events to\n    - authorization:
          \  # optional endpoint if an authorization token must first be requested
          \n        authDestHeader: Authorization   # http header key to be appended
          to destination request\n        authDestHeaderPrefix: 'Bearer '   # optional
          prefix for constructing destination request http header value\n        bodyData:
          \  # body data to send to authorization request, often an apiKey\n          secretKeyRef:\n
          \           name: auth-body-data-secret\n            key: bodydata\n        header:
          \  # secret map to use for authorization request headers\n          secret:\n
          \           name: auth-header-map-secret\n        tokenExpr: $.token   #
          optionally extract the returned body data using a jsonPath expression\n
          \       url: https://127.0.0.1:8443/api/2.0/accounts/my_account_id/apikeys/token
          \  # the url to make an authorization request to\n      header:   # secret
          map to use for destination request headers\n        secret:\n          name:
          dest-header-map-secret\n      transformer:   # The transformer to apply
          to event data for this destination\n        configMapKeyRef:\n          key:
          kazaam.json\n          name: kazaam-configmap\n        type: kazaam\n      url:
          https://127.0.0.1:8443/metering/api/1.0/usage   # the url to send the event
          to\n      urlSuffixExpr: $.properties.productId   # optional jsonPath expression
          on the event to extract and set a suffix to the url path\n    manifestType:
          dataReporter   # override manifestType of report sent to data-service\n
          \   selector:   # Matches against event data and a user. Empty selector
          matches all.\n      matchExpressions:   # jsonPath expression must return
          a result for all expressions to match (AND)\n      - $[?($.event == \"Account
          Contractual Usage\")]\n      - $.properties.productId\n      - $[?($.properties.source
          != null)]\n      - $[?($.properties.unit == \"AppPoints\")]\n      - $[?($.properties.quantity
          >= 0)]\n      - $[?($.timestamp != null)]\n      matchUsers:   # must match
          one of these users (OR). Omitting users matches any user\n      - system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n
          \   transformer:   # The transformer to apply to the event for data-service\n
          \     configMapKeyRef:\n        key: kazaam.json\n        name: kazaam-configmap\n
          \     type: kazaam\n  tlsConfig:   # TLS configuration for requests outbound
          to destinations\n    caCerts:\n    - name: tls-ca-certificates\n      key:
          ca.crt\n    certificates:\n    - clientCert:\n        secretKeyRef:\n          name:
          tls-ca-certificates\n          key: tls.crt\n      clientKey:\n        secretKeyRef:\n
          \         name: tls-ca-certificates\n          key: tls.key\n    cipherSuites:
          \n    - TLS_AES_128_GCM_SHA256\n    - TLS_AES_256_GCM_SHA384\n    - TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\n    - TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\n    insecureSkipVerify: false\n
          \   minVersion: VersionTLS12\n  userConfig:   # Optional metadata to apply
          to report sent to data-service for a specific user\n  - metadata:\n      ameta1:
          ametadata1\n      bmeta1: bmetadata1\n      cmeta1: cmetadata1\n      dmeta1:
          dmetadata1\n    userName: system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n```\n\nSample
          Header Secrets:\n```YAML\napiVersion: v1\nstringData:\n  accept: | \n    */*\nkind:
          Secret\nmetadata:\n  name: dest-header-map-secret\ntype: Opaque\n---\napiVersion:
          v1\nstringData:\n  accept: application/json\n  Content-Type: application/json\nkind:
          Secret\nmetadata:\n  name: auth-header-map-secret\ntype: Opaque\n```\n\nSample
          Authentication Body Secret:\n```YAML\napiVersion: v1\nstringData:\n  bodydata:
          |\n    {\"apikey\": \"myapikey\"}\nkind: Secret\nmetadata:\n  name: auth-body-data-secret\ntype:
          Opaque\n```\n\nSample TLS Secret:\n```YAML\napiVersion: v1\ndata:\n  ca.crt:
          LS0t...\n  tls.key: LS0t...\n  tls.crt: LS0t...\nkind: Secret\nmetadata:\n
          \ name: tls-ca-certificates\n```\n\nSample Kazaam ConfigMap:\n```YAML\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: kazaam-configmap\ndata:\n  kazaam.json:
          |\n    [\n        {\n            \"operation\": \"timestamp\",\n            \"spec\":
          {\n                \"timestamp\": {\n                \"inputFormat\": \"2006-01-02T15:04:05.999999+00:00\",\n
          \               \"outputFormat\": \"$unixmilli\"\n                }\n            }\n
          \       },\n        {\n            \"operation\": \"shift\",\n            \"spec\":
          {\n                \"instances[0].instanceId\": \"properties.source\",\n
          \               \"instances[0].startTime\": \"timestamp\",\n                \"instances[0].endTime\":
          \"timestamp\",\n                \"instances[0].metricUsage[0].metricId\":
          \"properties.unit\",\n                \"instances[0].metricUsage[0].quantity\":
          \"properties.quantity\"\n            }\n        },\n        {\n            \"operation\":
          \"default\",\n            \"spec\": {\n                \"meteringModel\":
          \"point-in-time\",\n                \"meteringPlan\": \"contract\"\n            }\n
          \       }\n    ]\n```\n\n### API Service User Configuration\n\n- The ClusterRole
          for api access is `clusterrole/ibm-data-reporter-operator-api`\n- The default
          ServiceAccount provided as an api user is `system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api`\n
          \ - The default ClusterRoleBinding for this user is `clusterrolebinding/ibm-data-reporter-operator-api`\n\nOptional:\n\n-
          To create an additional ServiceAccount and ClusterRoleBinding for api access\n\n```SHELL\nNAMESPACE=$(oc
          config view --minify -o jsonpath='{..namespace}') && \\\noc create serviceaccount
          my-api-service-account && \\\noc create clusterrolebinding ibm-data-reporter-operator-api
          --clusterrole=ibm-data-reporter-operator-api --serviceaccount=${NAMESPACE}:my-api-service-account\n```\n\n-
          Update datareporterconfig to attach metadata to reports associated with
          this user\n\n## API Service Usage\n\n- Get Token & Host\n\n```SHELL\noc
          project ibm-software-central\nexport DRTOKEN=$(oc create token ibm-data-reporter-operator-api
          --namespace ibm-software-central --duration 1h)\nexport DRHOST=$(oc get
          route ibm-data-reporter --template='{{ .spec.host }}')\n```\n\n- Get the
          Status\n\n```SHELL\ncurl -k -H \"Authorization: Bearer ${DRTOKEN}\" https://${DRHOST}/v1/status
          \n```\n\n- Post an Event\n\n```SHELL\ncurl -k -H \"Authorization: Bearer
          ${DRTOKEN}\" -X POST -d '{\"event\":\"myevent\"}' https://${DRHOST}/v1/event\n```\n\n##
          Storage\n\n- The operator temporarily stores events in pod memory, and writes
          to the IBM Metrics Operator data-service, which requires a PersistentVolume\n\n##
          License\n\nCopyright IBM Corporation 2023. All Rights Reserved.\n\nLicensed
          under the Apache License, Version 2.0 (the \"License\");\nyou may not use
          this file except in compliance with the License.\nYou may obtain a copy
          of the License at\n\n    http://www.apache.org/licenses/LICENSE-2.0\n\nUnless
          required by applicable law or agreed to in writing, software\ndistributed
          under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES
          OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for
          the specific language governing permissions and\nlimitations under the License.\n"
        displayName: IBM Data Reporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 2.20.0
      entries:
      - name: ibm-data-reporter-operator.v2.20.0
        version: 2.20.0
      name: beta
    - currentCSV: ibm-data-reporter-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "DataReporterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-data-reporter-operator",
                    "app.kubernetes.io/instance": "datareporterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datareporterconfig",
                    "app.kubernetes.io/part-of": "ibm-data-reporter-operator"
                  },
                  "name": "datareporterconfig"
                },
                "spec": {
                  "userConfig": [
                    {
                      "metadata": {
                        "ameta1": "ametadata1",
                        "bmeta1": "bmetadata1",
                        "cmeta1": "cmetadata1",
                        "dmeta1": "dmetadata1"
                      },
                      "userName": "system:serviceaccount:redhat-marketplace:ibm-data-reporter-operator-api"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
          createdAt: ""
          description: The IBM Data Reporter Operator accepts events and transforms
            them into reports submitted to the Data Service of the IBM Metrics Operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReporterConfig is the Schema for the datareporterconfigs
              API
            displayName: Data Reporter Config
            kind: DataReporterConfig
            name: datareporterconfigs.marketplace.redhat.com
            version: v1alpha1
          required:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
        description: "# IBM Data Reporter Operator\n\n# Introduction\n\nThe IBM Data
          Reporter Operator accepts events and transforms them into reports submitted
          to the IBM Metrics Operator Data Service.\n\n# Details\n\nThe IBM Data Reporter
          Operator deploys a service that provides an endpoint to which callers can
          send raw json event data. The event data is transformed into a report and
          is sent to the IBM Metrics Operator's Data Service. The IBM Metrics Operator
          Data Service periodically uploads the reports to IBM Software Central. It
          can also filter, transform and forward events to alternate endpoints.\n\n##
          Prerequisites\n\n- OpenShift Container Platform, major version 4 with any
          available supported minor version\n- Install IBM Metrics Operator\n  - IBM
          Data Reporter Operator prerequisties the IBM Metrics Operator's Data Service\n
          \ - Register the Cluster by creating a `redhat-marketplace-pull-secret`,
          as per the instructions\n  - Once `rhm-data-service` has started IBM Data
          Reporter Operator will accept requests\n\n## SecurityContextConstraints
          Requirements\n\n- The operator runs under Red Hat restricted SCC\n\n## Resources
          Required\n\n- The operator requires at least 85 Mi memory.\n\n## Limitations\n\n-
          Only runs on amd64, s390x and ppc64le architectures\n\n## Installing\n\n-
          A user with the cluster administrator role.\n- Install this operator in
          the same namespace as the IBM Metrics Operator\n  - default namespace: `ibm-software-central`\n\n##
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n# Configuration\n\nOptional:\n- Configure the
          Custom Resource Definition `DataReporterConfig` named `datareporterconfig`
          as per the following sample\n  - Configuration that can not be reconciled
          successfully will be reported on the DataReporterConfig Status\n  - jsonPath
          expressions are handled by: https://github.com/ohler55/ojg\n    - jsonPath
          comparison: https://cburgmer.github.io/json-path-comparison/\n  - Transformer
          type currently supported is `kazaam`\n    - https://github.com/willie68/kazaam\n\nSample
          DataReporterConfig:\n```YAML\napiVersion: marketplace.redhat.com/v1alpha1\nkind:
          DataReporterConfig\nmetadata:\n  name: datareporterconfig\nspec:\n  # true:
          report generated for 1 event and provide return code from data-service,
          confirming delivery\n  # false: events will be accumulated before building
          report, and immediately return 200, delivery unfconfirmed\n  confirmDelivery:
          false\n  dataFilters:   # dataFilters matches the first dataFilter selector,
          applies transforms and forwards to data-service and destinations\n  - altDestinations:
          \  # List of desintations to transform and foward events to\n    - authorization:
          \  # optional endpoint if an authorization token must first be requested
          \n        authDestHeader: Authorization   # http header key to be appended
          to destination request\n        authDestHeaderPrefix: 'Bearer '   # optional
          prefix for constructing destination request http header value\n        bodyData:
          \  # body data to send to authorization request, often an apiKey\n          secretKeyRef:\n
          \           name: auth-body-data-secret\n            key: bodydata\n        header:
          \  # secret map to use for authorization request headers\n          secret:\n
          \           name: auth-header-map-secret\n        tokenExpr: $.token   #
          optionally extract the returned body data using a jsonPath expression\n
          \       url: https://127.0.0.1:8443/api/2.0/accounts/my_account_id/apikeys/token
          \  # the url to make an authorization request to\n      header:   # secret
          map to use for destination request headers\n        secret:\n          name:
          dest-header-map-secret\n      transformer:   # The transformer to apply
          to event data for this destination\n        configMapKeyRef:\n          key:
          kazaam.json\n          name: kazaam-configmap\n        type: kazaam\n      url:
          https://127.0.0.1:8443/metering/api/1.0/usage   # the url to send the event
          to\n      urlSuffixExpr: $.properties.productId   # optional jsonPath expression
          on the event to extract and set a suffix to the url path\n    manifestType:
          dataReporter   # override manifestType of report sent to data-service\n
          \   selector:   # Matches against event data and a user. Empty selector
          matches all.\n      matchExpressions:   # jsonPath expression must return
          a result for all expressions to match (AND)\n      - $[?($.event == \"Account
          Contractual Usage\")]\n      - $.properties.productId\n      - $[?($.properties.source
          != null)]\n      - $[?($.properties.unit == \"AppPoints\")]\n      - $[?($.properties.quantity
          >= 0)]\n      - $[?($.timestamp != null)]\n      matchUsers:   # must match
          one of these users (OR). Omitting users matches any user\n      - system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n
          \   transformer:   # The transformer to apply to the event for data-service\n
          \     configMapKeyRef:\n        key: kazaam.json\n        name: kazaam-configmap\n
          \     type: kazaam\n  tlsConfig:   # TLS configuration for requests outbound
          to destinations\n    caCerts:\n    - name: tls-ca-certificates\n      key:
          ca.crt\n    certificates:\n    - clientCert:\n        secretKeyRef:\n          name:
          tls-ca-certificates\n          key: tls.crt\n      clientKey:\n        secretKeyRef:\n
          \         name: tls-ca-certificates\n          key: tls.key\n    cipherSuites:
          \n    - TLS_AES_128_GCM_SHA256\n    - TLS_AES_256_GCM_SHA384\n    - TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\n    - TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\n    insecureSkipVerify: false\n
          \   minVersion: VersionTLS12\n  userConfig:   # Optional metadata to apply
          to report sent to data-service for a specific user\n  - metadata:\n      ameta1:
          ametadata1\n      bmeta1: bmetadata1\n      cmeta1: cmetadata1\n      dmeta1:
          dmetadata1\n    userName: system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n```\n\nSample
          Header Secrets:\n```YAML\napiVersion: v1\nstringData:\n  accept: | \n    */*\nkind:
          Secret\nmetadata:\n  name: dest-header-map-secret\ntype: Opaque\n---\napiVersion:
          v1\nstringData:\n  accept: application/json\n  Content-Type: application/json\nkind:
          Secret\nmetadata:\n  name: auth-header-map-secret\ntype: Opaque\n```\n\nSample
          Authentication Body Secret:\n```YAML\napiVersion: v1\nstringData:\n  bodydata:
          |\n    {\"apikey\": \"myapikey\"}\nkind: Secret\nmetadata:\n  name: auth-body-data-secret\ntype:
          Opaque\n```\n\nSample TLS Secret:\n```YAML\napiVersion: v1\ndata:\n  ca.crt:
          LS0t...\n  tls.key: LS0t...\n  tls.crt: LS0t...\nkind: Secret\nmetadata:\n
          \ name: tls-ca-certificates\n```\n\nSample Kazaam ConfigMap:\n```YAML\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: kazaam-configmap\ndata:\n  kazaam.json:
          |\n    [\n        {\n            \"operation\": \"timestamp\",\n            \"spec\":
          {\n                \"timestamp\": {\n                \"inputFormat\": \"2006-01-02T15:04:05.999999+00:00\",\n
          \               \"outputFormat\": \"$unixmilli\"\n                }\n            }\n
          \       },\n        {\n            \"operation\": \"shift\",\n            \"spec\":
          {\n                \"instances[0].instanceId\": \"properties.source\",\n
          \               \"instances[0].startTime\": \"timestamp\",\n                \"instances[0].endTime\":
          \"timestamp\",\n                \"instances[0].metricUsage[0].metricId\":
          \"properties.unit\",\n                \"instances[0].metricUsage[0].quantity\":
          \"properties.quantity\"\n            }\n        },\n        {\n            \"operation\":
          \"default\",\n            \"spec\": {\n                \"meteringModel\":
          \"point-in-time\",\n                \"meteringPlan\": \"contract\"\n            }\n
          \       }\n    ]\n```\n\n### API Service User Configuration\n\n- The ClusterRole
          for api access is `clusterrole/ibm-data-reporter-operator-api`\n- The default
          ServiceAccount provided as an api user is `system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api`\n
          \ - The default ClusterRoleBinding for this user is `clusterrolebinding/ibm-data-reporter-operator-api`\n\nOptional:\n\n-
          To create an additional ServiceAccount and ClusterRoleBinding for api access\n\n```SHELL\nNAMESPACE=$(oc
          config view --minify -o jsonpath='{..namespace}') && \\\noc create serviceaccount
          my-api-service-account && \\\noc create clusterrolebinding ibm-data-reporter-operator-api
          --clusterrole=ibm-data-reporter-operator-api --serviceaccount=${NAMESPACE}:my-api-service-account\n```\n\n-
          Update datareporterconfig to attach metadata to reports associated with
          this user\n\n## API Service Usage\n\n- Get Token & Host\n\n```SHELL\noc
          project ibm-software-central\nexport DRTOKEN=$(oc create token ibm-data-reporter-operator-api
          --namespace ibm-software-central --duration 1h)\nexport DRHOST=$(oc get
          route ibm-data-reporter --template='{{ .spec.host }}')\n```\n\n- Get the
          Status\n\n```SHELL\ncurl -k -H \"Authorization: Bearer ${DRTOKEN}\" https://${DRHOST}/v1/status
          \n```\n\n- Post an Event\n\n```SHELL\ncurl -k -H \"Authorization: Bearer
          ${DRTOKEN}\" -X POST -d '{\"event\":\"myevent\"}' https://${DRHOST}/v1/event\n```\n\n##
          Storage\n\n- The operator temporarily stores events in pod memory, and writes
          to the IBM Metrics Operator data-service, which requires a PersistentVolume\n\n##
          License\n\nCopyright IBM Corporation 2023. All Rights Reserved.\n\nLicensed
          under the Apache License, Version 2.0 (the \"License\");\nyou may not use
          this file except in compliance with the License.\nYou may obtain a copy
          of the License at\n\n    http://www.apache.org/licenses/LICENSE-2.0\n\nUnless
          required by applicable law or agreed to in writing, software\ndistributed
          under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES
          OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for
          the specific language governing permissions and\nlimitations under the License.\n"
        displayName: IBM Data Reporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
        version: 2.20.0
      entries:
      - name: ibm-data-reporter-operator.v2.20.0
        version: 2.20.0
      name: stable
    defaultChannel: stable
    packageName: ibm-data-reporter-operator
    provider:
      name: Red Hat Marketplace
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-xtradb-cluster-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "docker.io/percona/percona-xtradb-cluster:8.0.42-33.1",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "docker.io/percona/haproxy:2.8.15",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "docker.io/percona/proxysql2:2.7.3",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "docker.io/percona/fluentbit:4.0.1",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:3.3.1",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "docker.io/percona/percona-xtrabackup:8.0.35-34.1",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.io/percona/percona-xtradb-cluster-operator:1.18.0
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <v1.18.0
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            clustercheck: Y2x1c3RlcmNoZWNrcGFzc3dvcmQ=
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            clustercheck: Y2x1c3RlcmNoZWNrcGFzc3dvcmQ=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ```
          ###  Release Highlights
          * **PMM3 support**

          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**

          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**

          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**

          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.

        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - database
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - docker.io/percona/percona-xtradb-cluster-operator:1.18.0
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator.v1.17.0
        version: 1.17.0
      - name: percona-xtradb-cluster-operator.v1.16.1
        version: 1.16.1
      - name: percona-xtradb-cluster-operator.v1.16.0
        version: 1.16.0
      - name: percona-xtradb-cluster-operator.v1.15.1
        version: 1.15.1
      - name: percona-xtradb-cluster-operator.v1.15.0
        version: 1.15.0
      - name: percona-xtradb-cluster-operator.v1.9.0
        version: 1.9.0
      name: stable
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: authorino-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: authorino-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta3",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/authorino-rhel9-operator@sha256:2d81613856df9725ad20928a7a9895d14482c42effce0a3a0de88f40ec35d411
          createdAt: "2025-03-25T22:39:23Z"
          description: Enables authentication and authorization for Gateways and applications
            in a Gateway API network.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta3
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: |-
          [Authorino](https://docs.kuadrant.io/authorino/) is Red Hat's Kubernetes-native lightweight external authorization service for tailor-made Zero Trust API security.

          Install this Red Hat official distribution of [Authorino Operator](https://docs.kuadrant.io/authorino-operator/) to manage instances of Authorino in this cluster.

          The current state of this distribution of the operator is: **Generally Available**.

          The Community version of Authorino Operator, based on upstream public images, is available in [OperatorHub.io](https://operatorhub.io/operator/authorino-operator).

          ## Getting started

          After installing the Operator, request an instance of the external authorization service by creating an `Authorino` custom resource.

          **Minimal example (namespaced)**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            listener:
              tls:
                enabled: false
            oidcServer:
              tls:
                enabled: false
          ```

          **Extended example**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            clusterWide: true
            authConfigLabelSelectors: environment=production
            secretLabelSelectors: authorino.kuadrant.io/component=authorino,environment=production

            replicas: 2

            evaluatorCacheSize: 2 # mb

            logLevel: info
            logMode: production

            listener:
              ports:
                grpc: 50001
                http: 5001
              tls:
                certSecretRef:
                  name: authorino-server-cert # secret must contain `tls.crt` and `tls.key` entries
              timeout: 2

            oidcServer:
              port: 8083
              tls:
                certSecretRef:
                  name: authorino-oidc-server-cert # secret must contain `tls.crt` and `tls.key` entries

            metrics:
              port: 8080
              deep: true

            healthz:
              port: 8081

            tracing:
              endpoint: rpc://otel-collector.observability.svc.cluster.local:4317
              insecure: true

            volumes:
              items:
                - name: keycloak-tls-cert
                  mountPath: /etc/ssl/certs
                  configMaps:
                    - keycloak-tls-cert
                  items: # details to mount the k8s configmap in the authorino pods
                    - key: keycloak.crt
                      path: keycloak.crt
              defaultMode: 420
          ```

          ### Cluster-wide vs Namespaced

          Namespaced instances only watch auth resources (`AuthConfig` and `Secrets`) created in the same namespace as the Authorino service. Use this mode for dedicated instances that do not require elevated privileges.

          Cluster-wide instances watch resources across the entire cluster (all namespaces.) Deploying and running Authorino in this mode requires elevated privileges.

          ### Multi-tenancy

          Use the `authConfigLabelSelectors` field of the `Authorino` custom resource to narrow the scope of the Authorino instance.

          Only `AuthConfig` custom resources whose labels match the value of this field will be handled by the Authorino instance.

          ## Protect a host

          To protect a host, create an `AuthConfig` custom resource for the host. E.g.:

          ```yaml
          apiVersion: authorino.kuadrant.io/v1beta3
          kind: AuthConfig
          metadata:
            name: my-api-protection
          spec:
            hosts:
            - my-api.io

            authentication:
              "keycloak":
                jwt:
                  issuerUrl: https://keycloak.keycloak.svc.cluster.local:8080/realms/my-realm

            authorization:
              "k8s-rbac":
                kubernetesSubjectAccessReview:
                  user:
                    selector: auth.identity.user.username
                  resourceAttributes:
                    resource:
                      value: my-api
                    verb:
                      selector: request.method
                cache:
                  key:
                    selector: auth.identity.user.username
                  ttl: 30
              "after-2am-only":
                rego: |
                  allow {
                    [hour, _, _] := time.clock(time.now_ns())
                    hour >= 2
                  }
          ```

          Make sure all requests to the host are fisrt checked with the Authorino instance, by configuring an Envoy proxy for external authz:

          ```yaml
          clusters:
          - name: my-api
            …
          - name: authorino
            connect_timeout: 0.25s
            type: STRICT_DNS
            lb_policy: ROUND_ROBIN
            http2_protocol_options: {}
            load_assignment:
              cluster_name: authorino
              endpoints:
              - lb_endpoints:
                - endpoint:
                    address:
                      socket_address:
                        address: authorino-authorino-authorization
                        port_value: 50051
          listeners:
          - filter_chains:
            - filters:
              name: envoy.http_connection_manager
              typed_config:
                "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
                use_remote_address: true
                route_config:
                  name: my-api-config
                  virtual_hosts:
                  - name: my-api-vs
                    domains:
                    - my-api.io
                    routes:
                    - match:
                        prefix: /
                      route:
                        cluster: my-api
                http_filters:
                - name: envoy.filters.http.ext_authz
                  typed_config:
                    "@type": type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz
                    transport_api_version: V3
                    failure_mode_allow: false
                    include_peer_certificate: true
                    grpc_service:
                      envoy_grpc:
                        cluster_name: authorino
                      timeout: 1s
          ```

          ...or, if using Istio, by creating an [`AuthorizationPolicy`](https://istio.io/latest/docs/reference/config/security/authorization-policy/#AuthorizationPolicy) custom resource. Use `action: CUSTOM` in the resource and the Authorino authorization service configured in the mesh extension provider settings.

          ## Features

          **Authentication**

          * JWT validation (with OpenID Connect Discovery)
          * OAuth 2.0 Token Introspection (opaque tokens)
          * Kubernetes TokenReview (ServiceAccount tokens)
          * API key authentication
          * X.509 client certificate authentication
          * Anonymous access
          * Proxy-handled (authentication performed by the proxy)

          **Authorization**

          * Built-in simple pattern matching (e.g. JWT claims, request attributes checking)
          * OPA policies (inline Rego and fetch from external registry)
          * Kubernetes SubjectAccessReview (resource and non-resource attributes)
          * Authzed SpiceDB

          **External metadata**

          * HTTP request
          * OpenID Connect User Info
          * UMA-protected resource attributes

          **Custom responses**

          * Header injection (Festival Wristbands tokens, JSON, plain text)
          * Envoy Dynamic Metadata
          * Custom HTTP response (status code, headers, messages, body, etc)

          **Callbacks**

          * HTTP webhooks

          **Caching**

          * OpenID Connect and User-Managed Access configs
          * JSON Web Keys (JWKs) and JSON Web Key Sets (JWKS)
          * Access tokens
          * External metadata
          * Precompiled Rego policies
          * Policy evaluation

          Check out the full [Feature Specification](https://docs.kuadrant.io/authorino/docs/features/) and how-to guides in the [Kuadrant Docs](https://docs.kuadrant.io) website.
        displayName: Red Hat - Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: GitHub
          url: https://github.com/kuadrant/authorino-operator
        - name: Documentation
          url: https://docs.kuadrant.io/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhcl-1/authorino-rhel9-operator@sha256:2d81613856df9725ad20928a7a9895d14482c42effce0a3a0de88f40ec35d411
        - registry.redhat.io/rhcl-1/authorino-rhel9@sha256:413d3796ad0101bf14d3341d31d9367e67fea1e73ebb3fb79124ef4c5dde0c09
        version: 1.2.2
      entries:
      - name: authorino-operator.v1.2.2
        version: 1.2.2
      - name: authorino-operator.v1.2.1
        version: 1.2.1
      - name: authorino-operator.v1.2.0
        version: 1.2.0
      - name: authorino-operator.v1.1.3
        version: 1.1.3
      - name: authorino-operator.v1.1.2
        version: 1.1.2
      - name: authorino-operator.v1.1.1
        version: 1.1.1
      - name: authorino-operator.v1.1.0
        version: 1.1.0
      - name: authorino-operator.v1.0.2
        version: 1.0.2
      - name: authorino-operator.v0.16.1
        version: 0.16.1
      - name: authorino-operator.v0.16.0
        version: 0.16.0
      name: stable
    - currentCSV: authorino-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta1",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "hosts": [
                    "my-api.io"
                  ],
                  "identity": [
                    {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "in": "authorization_header",
                        "keySelector": "APIKEY"
                      },
                      "name": "api-key-users"
                    }
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/3scale-tech-preview/authorino-rhel9-operator@sha256:eb574781cf39917f8f5d9e5c806f520534601f819ee1ebcbb751832da8938008
          createdAt: "2024-11-08T11:06:33Z"
          description: Authorino
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta1
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: |-
          [Authorino](https://docs.kuadrant.io/authorino/) is Red Hat's Kubernetes-native lightweight external authorization service for tailor-made Zero Trust API security.

          Install this Red Hat official distribution of [Authorino Operator](https://docs.kuadrant.io/authorino-operator/) to manage instances of Authorino in this cluster.

          The current state of this distribution of the operator is: **_Technical Preview_**.

          The Community version of Authorino Operator, based on upstream public images, is available in [OperatorHub.io](https://operatorhub.io/operator/authorino-operator).

          ## Getting started

          After installing the Operator, request an instance of the external authorization service by creating an `Authorino` custom resource.

          **Minimal example (namespaced)**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            listener:
              tls:
                enabled: false
            oidcServer:
              tls:
                enabled: false
          ```

          **Extended example**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            clusterWide: true
            authConfigLabelSelectors: environment=production
            secretLabelSelectors: authorino.kuadrant.io/component=authorino,environment=production

            replicas: 2

            evaluatorCacheSize: 2 # mb

            logLevel: info
            logMode: production

            listener:
              ports:
                grpc: 50001
                http: 5001
              tls:
                certSecretRef:
                  name: authorino-server-cert # secret must contain `tls.crt` and `tls.key` entries
              timeout: 2

            oidcServer:
              port: 8083
              tls:
                certSecretRef:
                  name: authorino-oidc-server-cert # secret must contain `tls.crt` and `tls.key` entries

            metrics:
              port: 8080
              deep: true

            healthz:
              port: 8081

            tracing:
              endpoint: rpc://otel-collector.observability.svc.cluster.local:4317
              insecure: true

            volumes:
              items:
                - name: keycloak-tls-cert
                  mountPath: /etc/ssl/certs
                  configMaps:
                    - keycloak-tls-cert
                  items: # details to mount the k8s configmap in the authorino pods
                    - key: keycloak.crt
                      path: keycloak.crt
              defaultMode: 420
          ```

          ### Cluster-wide vs Namespaced

          Namespaced instances only watch auth resources (`AuthConfig` and `Secrets`) created in the same namespace as the Authorino service. Use this mode for dedicated instances that do not require elevated privileges.

          Cluster-wide instances watch resources across the entire cluster (all namespaces.) Deploying and running Authorino in this mode requires elevated privileges.

          ### Multi-tenancy

          Use the `authConfigLabelSelectors` field of the `Authorino` custom resource to narrow the scope of the Authorino instance.

          Only `AuthConfig` custom resources whose labels match the value of this field will be handled by the Authorino instance.

          ## Protect a host

          To protect a host, create an `AuthConfig` custom resource for the host. E.g.:

          ```yaml
          apiVersion: authorino.kuadrant.io/v1beta2
          kind: AuthConfig
          metadata:
            name: my-api-protection
          spec:
            hosts:
            - my-api.io

            authentication:
              "keycloak":
                jwt:
                  issuerUrl: https://keycloak.keycloak.svc.cluster.local:8080/realms/my-realm

            authorization:
              "k8s-rbac":
                kubernetesSubjectAccessReview:
                  user:
                    selector: auth.identity.user.username
                  resourceAttributes:
                    resource:
                      value: my-api
                    verb:
                      selector: request.method
                cache:
                  key:
                    selector: auth.identity.user.username
                  ttl: 30
              "after-2am-only":
                rego: |
                  allow {
                    [hour, _, _] := time.clock(time.now_ns())
                    hour >= 2
                  }
          ```

          Make sure all requests to the host are fisrt checked with the Authorino instance, by configuring an Envoy proxy for external authz:

          ```yaml
          clusters:
          - name: my-api
            …
          - name: authorino
            connect_timeout: 0.25s
            type: STRICT_DNS
            lb_policy: ROUND_ROBIN
            http2_protocol_options: {}
            load_assignment:
              cluster_name: authorino
              endpoints:
              - lb_endpoints:
                - endpoint:
                    address:
                      socket_address:
                        address: authorino-authorino-authorization
                        port_value: 50051
          listeners:
          - filter_chains:
            - filters:
              name: envoy.http_connection_manager
              typed_config:
                "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
                use_remote_address: true
                route_config:
                  name: my-api-config
                  virtual_hosts:
                  - name: my-api-vs
                    domains:
                    - my-api.io
                    routes:
                    - match:
                        prefix: /
                      route:
                        cluster: my-api
                http_filters:
                - name: envoy.filters.http.ext_authz
                  typed_config:
                    "@type": type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz
                    transport_api_version: V3
                    failure_mode_allow: false
                    include_peer_certificate: true
                    grpc_service:
                      envoy_grpc:
                        cluster_name: authorino
                      timeout: 1s
          ```

          ...or, if using Istio, by creating an [`AuthorizationPolicy`](https://istio.io/latest/docs/reference/config/security/authorization-policy/#AuthorizationPolicy) custom resource. Use `action: CUSTOM` in the resource and the Authorino authorization service configured in the mesh extension provider settings.

          ## Features

          **Authentication**

          * JWT validation (with OpenID Connect Discovery)
          * OAuth 2.0 Token Introspection (opaque tokens)
          * Kubernetes TokenReview (ServiceAccount tokens)
          * API key authentication
          * X.509 client certificate authentication
          * Anonymous access
          * Proxy-handled (authentication performed by the proxy)

          **Authorization**

          * Built-in simple pattern matching (e.g. JWT claims, request attributes checking)
          * OPA policies (inline Rego and fetch from external registry)
          * Kubernetes SubjectAccessReview (resource and non-resource attributes)
          * Authzed SpiceDB

          **External metadata**

          * HTTP request
          * OpenID Connect User Info
          * UMA-protected resource attributes

          **Custom responses**

          * Header injection (Festival Wristbands tokens, JSON, plain text)
          * Envoy Dynamic Metadata
          * Custom HTTP response (status code, headers, messages, body, etc)

          **Callbacks**

          * HTTP webhooks

          **Caching**

          * OpenID Connect and User-Managed Access configs
          * JSON Web Keys (JWKs) and JSON Web Key Sets (JWKS)
          * Access tokens
          * External metadata
          * Precompiled Rego policies
          * Policy evaluation

          Check out the full [Feature Specification](https://docs.kuadrant.io/authorino/docs/features/) and how-to guides in the [Kuadrant Docs](https://docs.kuadrant.io) website.
        displayName: Red Hat - Authorino (Technical Preview)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: GitHub
          url: https://github.com/kuadrant/authorino-operator
        - name: Documentation
          url: https://docs.kuadrant.io/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-tech-preview/authorino-rhel9@sha256:a473dae20e71e3e813ac30ba978f2ab3c5e19d7d39b501ae9103dca892107c87
        - registry.redhat.io/3scale-tech-preview/authorino-rhel9-operator@sha256:eb574781cf39917f8f5d9e5c806f520534601f819ee1ebcbb751832da8938008
        - registry.redhat.io/openshift4/ose-cli-rhel9@sha256:3a398158d10018bfac5df291d6998b513cf9a30e5477bcaefddcf66d229627c7
        version: 1.1.3
      entries:
      - name: authorino-operator.v1.1.3
        version: 1.1.3
      - name: authorino-operator.v1.1.2
        version: 1.1.2
      - name: authorino-operator.v1.1.1
        version: 1.1.1
      - name: authorino-operator.v1.1.0
        version: 1.1.0
      - name: authorino-operator.v1.0.2
        version: 1.0.2
      name: tech-preview-v1
    defaultChannel: stable
    packageName: authorino-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-athena-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-athena-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "athena.services.k8s.aws/v1alpha1",
                "kind": "WorkGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/athena-controller:1.1.1
          createdAt: "2025-09-19T21:06:02Z"
          description: AWS Athena controller is a service controller for managing
            Athena resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PreparedStatement represents the state of an AWS athena PreparedStatement
              resource.
            displayName: PreparedStatement
            kind: PreparedStatement
            name: preparedstatements.athena.services.k8s.aws
            version: v1alpha1
          - description: WorkGroup represents the state of an AWS athena WorkGroup
              resource.
            displayName: WorkGroup
            kind: WorkGroup
            name: workgroups.athena.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Athena resources in AWS from within your Kubernetes cluster.

          **About Amazon Athena**

          Amazon Athena is an interactive query service that makes it easy to analyze data directly in Amazon Simple Storage Service (Amazon S3) using standard SQL. With a few actions in the AWS Management Console, you can point Athena at your data stored in Amazon S3 and begin using standard SQL to run ad-hoc queries and get results in seconds.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Athena
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - athena
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Athena Developer Resources
          url: https://aws.amazon.com/athena/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: athena maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/athena-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-athena-controller.v1.1.1
        version: 1.1.1
      - name: ack-athena-controller.v1.1.0
        version: 1.1.0
      - name: ack-athena-controller.v1.0.14
        version: 1.0.14
      - name: ack-athena-controller.v1.0.13
        version: 1.0.13
      - name: ack-athena-controller.v1.0.12
        version: 1.0.12
      - name: ack-athena-controller.v1.0.11
        version: 1.0.11
      - name: ack-athena-controller.v1.0.10
        version: 1.0.10
      - name: ack-athena-controller.v1.0.9
        version: 1.0.9
      - name: ack-athena-controller.v1.0.8
        version: 1.0.8
      - name: ack-athena-controller.v1.0.7
        version: 1.0.7
      - name: ack-athena-controller.v1.0.6
        version: 1.0.6
      - name: ack-athena-controller.v1.0.5
        version: 1.0.5
      - name: ack-athena-controller.v1.0.4
        version: 1.0.4
      - name: ack-athena-controller.v1.0.1
        version: 1.0.1
      - name: ack-athena-controller.v1.0.0
        version: 1.0.0
      - name: ack-athena-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-athena-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Flux project
      provider-url: ""
    name: flux
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.toolkit.fluxcd.io/v2beta1",
                "kind": "HelmRelease",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "5m",
                    "chart": {
                      "spec": {
                          "chart": "podinfo",
                          "version": "4.0.x",
                          "sourceRef": {
                            "kind": "HelmRepository",
                            "name": "podinfo",
                            "namespace": "flux-system"
                          },
                          "interval": "1m"
                      }
                    },
                    "values": {
                      "replicaCount": 2
                    }
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "GitRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "1m",
                    "ref": {
                      "branch": "master"
                    },
                    "url": "https://github.com/stefanprodan/podinfo"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmRepository",
                "metadata": {
                    "name": "stable",
                    "namespace": "default"
                },
                "spec": {
                    "url": "https://kubernetes-charts.storage.googleapis.com/",
                    "interval": "10m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmChart",
                "metadata": {
                    "name": "redis",
                    "namespace": "default"
                },
                "spec": {
                    "chart": "redis",
                    "version": "10.5.7",
                    "sourceRef": {
                      "name": "stable",
                      "kind": "HelmRepository"
                    },
                    "interval": "5m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "Bucket",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "1m",
                    "provider": "generic",
                    "bucketName": "podinfo",
                    "endpoint": "minio.minio.svc.cluster.local:9000",
                    "insecure": true,
                    "secretRef": {
                      "name": "minio-credentials"
                    }
                }
              },
              {
                "apiVersion": "kustomize.toolkit.fluxcd.io/v1beta2",
                "kind": "Kustomization",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "5m",
                    "path": "./deploy/overlays/dev",
                    "prune": true,
                    "sourceRef": {
                      "kind": "GitRepository",
                      "name": "podinfo"
                    },
                    "timeout": "2m",
                    "wait": true
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "address": "https://github.com/stefanprodan/podinfo",
                    "secretRef": {
                      "name": "api-token"
                    }
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Alert",
                "metadata": {
                    "name": "webapp",
                    "namespace": "default"
                },
                "spec": {
                    "providerRef": {
                      "name": "on-call-slack"
                    },
                    "eventSeverity": "info",
                    "eventSources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "Bucket",
                          "name": "secrets"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-backend"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-frontend"
                      }
                    ]
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Receiver",
                "metadata": {
                    "name": "github-receiver",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "events": [
                      "ping",
                      "push"
                    ],
                    "secretRef": {
                      "name": "webhook-token"
                    },
                    "resources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "HelmRepository",
                          "name": "webapp"
                      }
                    ]
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "image": "ghcr.io/stefanprodan/podinfo",
                    "interval": "1m0s"
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImagePolicy",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "imageRepositoryRef": {
                      "name": "podinfo"
                    },
                    "policy": {
                      "semver": {
                          "range": "5.0.x"
                      }
                    }
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageUpdateAutomation",
                "metadata": {
                  "name": "flux-system",
                  "namespace": "flux-system"
                },
                "spec": {
                  "interval": "1m0s",
                  "sourceRef": {
                    "kind": "GitRepository",
                    "name": "flux-system"
                  },
                  "git": {
                    "checkout": {
                      "ref": {
                        "branch": "main"
                      }
                    },
                    "commit": {
                      "author": {
                        "email": "fluxcdbot@users.noreply.github.com",
                        "name": "fluxcdbot"
                      },
                      "messageTemplate": "{{range .Updated.Images}}{{println .}}{{end}}"
                    },
                    "push": {
                      "branch": "main"
                    }
                  },
                  "update": {
                    "path": "./clusters/my-cluster",
                    "strategy": "Setters"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/fluxcd/source-controller:v1.3.0
          description: Flux is a Continuous Delivery solution for Kubernetes.
          operatorframework.io/suggested-namespace: flux-system
          support: The Flux community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta3
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta1
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta2
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta1
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta2
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta2
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta2
          - description: OCIRepository
            displayName: OCIRepository
            kind: OCIRepository
            name: ocirepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta3
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta2
        description: |-
          [Flux](https://fluxcd.io/) is a Continuous Delivery solution for Kubernetes.
          Flux is constructed with the [GitOps Toolkit](https://toolkit.fluxcd.io/components/), a set of composable APIs and specialized tools for keeping Kubernetes clusters in sync with sources of configuration (like Git & Helm repositories), and automating updates to configuration when there is new code to deploy.

          Flux is a CNCF incubation project made for:

          - **cluster operators** who automate provision and configuration of clusters
          - **platform engineers** who build continuous delivery for developer teams
          - **app developers** who rely on continuous delivery to get their code live

          Features:

          - Source configuration from Git and Helm repositories, and S3-compatible buckets (e.g., Minio)
          - Kustomize and Helm support
          - Event-triggered and periodic reconciliation
          - Integration with Kubernetes RBAC
          - Health assessment (clusters and workloads)
          - Dependency management (infrastructure and workloads)
          - Secrets management (builtin support for Mozilla SOPS and Cloud KMS)
          - Alerting to external systems (webhook senders)
          - External events handling (webhook receivers)
          - Automated container image updates to Git (image scanning and patching)
          - Policy-driven validation (OPA, Kyverno, admission controllers)
          - Seamless integration with Git providers (GitHub, GitLab, Bitbucket)
          - Interoperability with Cluster API (CAPI) providers

          ### Steps to support bootstraping with Flux CLI

          SecComp Profile should be removed as it requires a special privilege to get the access.
          Also, it requires to enforce `runUserAs` to the same UID provided by the images to prevent OpenShift
          to alter the user id expected by our controllers.

          To do so, please specifiy a patch using the following Kustomization file, for example.

          ```yaml
          apiVersion: kustomize.config.k8s.io/v1beta1
          kind: Kustomization
          resources:
            - gotk-components.yaml
            - gotk-sync.yaml
          patches:
            - patch: |
                apiVersion: apps/v1
                kind: Deployment
                metadata:
                  name: all
                spec:
                  template:
                    spec:
                      containers:
                        - name: manager
                          securityContext:
                            runAsUser: 65534
                            seccompProfile:
                              $patch: delete
              target:
                kind: Deployment
                labelSelector: app.kubernetes.io/part-of=flux
          ```

          Then before installing Flux, please also perform the following steps to bind Flux to the `nonroot` SecurityContextConstraint (SCC). This applies to installation via OC's OLM or via Flux CLI.

          ```sh
          # replace default "flux-system" with the project/namespace name for your flux installation
          namespace="flux-system"
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:source-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:kustomize-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:helm-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:notification-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-automation-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-reflector-controller

          ```

          To see complete steps please refer to this [documentation](https://github.com/fluxcd/website/blob/main/content/en/docs/use-cases/openshift.md).

          If you've already installed Flux, you might need to uninstall it after applying the above pre-requisite steps.

          ### Network Policies

          This version of Flux is shipped without the normally provided network Policies.
          By default all traffic is enabled, and anyone can access the source-controller
          artifacts.

          It is not recommended to run it in production as is.

          Please visit [Flux's policies](https://github.com/fluxcd/flux2/tree/main/manifests/policies)
          to deploy the needed network policies and secure your Flux setup.

          ### Breaking Changes

          - Flux 0.18.x ships with v1beta2 of the kustomize.toolkit.fluxcd.io/Kustomization CRD.
          - Flux 0.25.x is the last Flux release where Kubernetes 1.19 is supported.

          ### Flux on OpenShift Use Cases

          - [The GitOps with Flux + OpenShift](https://www.youtube.com/watch?v=8ybCtNVP5To) talk by Andrew Block (Red Hat) showing how to setup multi-tenacy on OpenShift using Flux.
          - [Flux OpenShift Use Case](https://fluxcd.io/docs/use-cases/openshift/) on FluxCD.io
        displayName: Flux
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - gitops
        - cd
        - continuous delivery
        - fluxcd
        - helm
        - kustomize
        links:
        - name: Homepage
          url: https://fluxcd.io/
        - name: Documentation
          url: https://fluxcd.io/docs/
        - name: Slack
          url: https://cloud-native.slack.com/messages/flux
        - name: Mailing List
          url: https://lists.cncf.io/g/cncf-flux-dev
        - name: Source Code
          url: https://github.com/fluxcd/flux2
        maintainers:
        - email: support@weave.works
          name: The Flux maintainers
        maturity: stable
        provider:
          name: The Flux project
        relatedImages:
        - ghcr.io/fluxcd/helm-controller:v1.0.1
        - ghcr.io/fluxcd/image-automation-controller:v0.38.0
        - ghcr.io/fluxcd/image-reflector-controller:v0.32.0
        - ghcr.io/fluxcd/kustomize-controller:v1.3.0
        - ghcr.io/fluxcd/notification-controller:v1.3.0
        - ghcr.io/fluxcd/source-controller:v1.3.0
        version: 2.3.0
      entries:
      - name: flux.v2.3.0
        version: 2.3.0
      - name: flux.v2.2.2
        version: 2.2.2
      - name: flux.v2.2.0
        version: 2.2.0
      - name: flux.v2.0.1
        version: 2.0.1
      - name: flux.v0.41.2
        version: 0.41.2
      - name: flux.v0.40.0
        version: 0.40.0
      - name: flux.v0.39.0
        version: 0.39.0
      - name: flux.v0.38.3
        version: 0.38.3
      - name: flux.v0.37.0
        version: 0.37.0
      - name: flux.v0.36.0
        version: 0.36.0
      - name: flux.v0.35.0
        version: 0.35.0
      - name: flux.v0.31.4
        version: 0.31.4
      - name: flux.v0.31.3
        version: 0.31.3
      - name: flux.v0.30.2
        version: 0.30.2
      - name: flux.v0.28.5
        version: 0.28.5
      - name: flux.v0.28.4
        version: 0.28.4
      - name: flux.v0.28.3
        version: 0.28.3
      - name: flux.v0.27.4
        version: 0.27.4
      - name: flux.v0.27.3
        version: 0.27.3
      - name: flux.v0.27.2
        version: 0.27.2
      - name: flux.v0.27.0
        version: 0.27.0
      - name: flux.v0.26.1
        version: 0.26.1
      - name: flux.v0.26.0
        version: 0.26.0
      - name: flux.v0.25.3
        version: 0.25.3
      - name: flux.v0.25.2
        version: 0.25.2
      - name: flux.v0.24.1
        version: 0.24.1
      - name: flux.v0.24.0
        version: 0.24.0
      - name: flux.v0.23.0
        version: 0.23.0
      - name: flux.v0.22.1
        version: 0.22.1
      - name: flux.v0.22.0
        version: 0.22.0
      - name: flux.v0.21.1
        version: 0.21.1
      - name: flux.v0.21.0
        version: 0.21.0
      - name: flux.v0.20.1
        version: 0.20.1
      - name: flux.v0.20.0
        version: 0.20.0
      - name: flux.v0.19.1
        version: 0.19.1
      - name: flux.v0.18.3
        version: 0.18.3
      - name: flux.v0.18.1
        version: 0.18.1
      - name: flux.v0.17.2
        version: 0.17.2
      - name: flux.v0.17.1
        version: 0.17.1
      - name: flux.v0.17.0
        version: 0.17.0
      - name: flux.v0.16.2
        version: 0.16.2
      - name: flux.v0.16.1
        version: 0.16.1
      - name: flux.v0.16.0
        version: 0.16.0
      - name: flux.v0.15.3
        version: 0.15.3
      - name: flux.v0.15.2
        version: 0.15.2
      - name: flux.v0.14.2
        version: 0.14.2
      - name: flux.v0.14.1
        version: 0.14.1
      - name: flux.v0.14.0
        version: 0.14.0
      - name: flux.v0.13.4
        version: 0.13.4
      name: stable
    defaultChannel: stable
    packageName: flux
    provider:
      name: The Flux project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          olm.skipRange: <1.0.4
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
        version: 1.0.4
      entries:
      - name: datadog-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      name: stable
    defaultChannel: stable
    packageName: datadog-operator-certified
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kiali-ossm
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kiali-operator.v2.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "Kiali",
                "metadata": {
                  "name": "kiali"
                },
                "spec": {
                  "version": "default",
                  "istio_namespace": "istio-system",
                  "deployment": {
                    "logger": {
                      "log_level": "info"
                    },
                    "view_only_mode": false
                  }
                }
              },
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "OSSMConsole",
                "metadata": {
                  "name": "ossmconsole"
                },
                "spec": {
                  "version": "default",
                  "kiali": {
                    "serviceName": "",
                    "serviceNamespace": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9-operator@sha256:0612dfa52f2bf39a1d87e640409313585fedc7cfcff6d00e77bb14809d27d646
          createdAt: "2024-12-12T10:44:15Z"
          description: This productized operator provides Kiali and OSSMC. Kiali is
            the Istio observability and management Console. OSSMC is the OpenShift
            Service Mesh Console plugin, powered by Kiali.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.1.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kiali/kiali
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A configuration file for a Kiali installation.
            displayName: Kiali
            kind: Kiali
            name: kialis.kiali.io
            version: v1alpha1
          - description: A configuration file for a OpenShift Service Mesh Console
              installation.
            displayName: OpenShift Service Mesh Console
            kind: OSSMConsole
            name: ossmconsoles.kiali.io
            version: v1alpha1
        description: |-
          ## About the managed application

          A Microservice Architecture breaks up the monolith into many smaller pieces
          that are composed together. Patterns to secure the communication between
          services like fault tolerance (via timeout, retry, circuit breaking, etc.)
          have come up as well as distributed tracing to be able to see where calls
          are going.

          A service mesh can now provide these services on a platform level and frees
          the application writers from those tasks. Routing decisions are done at the
          mesh level.

          Kiali works with OpenShift Service Mesh to visualize the service
          mesh topology, to provide visibility into features like circuit breakers,
          request rates and more. It offers insights about the mesh components at
          different levels, from abstract Applications to Services and Workloads.

          See [https://www.kiali.io](https://www.kiali.io) to read more.

          ### Accessing the UI

          By default, the Kiali operator exposes the Kiali UI as an OpenShift Route.

          If on OpenShift, you can create an OSSMConsole CR to have the operator
          install the OpenShift ServiceMesh Console plugin to the OpenShift Console
          thus providing an interface directly integrated with the OpenShift Console.

          ## About this Operator

          ### Kiali Custom Resource Configuration Settings

          For quick descriptions of all the settings you can configure in the Kiali
          Custom Resource (CR), see
          [the kiali.io docs](https://kiali.io/docs/configuration/kialis.kiali.io/).

          ## Prerequisites for enabling this Operator

          Kiali is a companion tool for OpenShift Service Mesh. So before you install Kiali, you must have
          already installed Service Mesh.
        displayName: Kiali Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - service-mesh
        - observability
        - monitoring
        - istio
        links:
        - name: Installation Guide
          url: https://www.kiali.io/docs/installation
        - name: Features
          url: https://www.kiali.io/docs/features
        - name: Documentation Home
          url: https://www.kiali.io
        - name: Blogs and Articles
          url: https://medium.com/kialiproject
        - name: Server and UI Source Code
          url: https://github.com/kiali/kiali
        - name: Operator Source Code
          url: https://github.com/kiali/kiali-operator
        - name: OpenShift Service Mesh Console Code
          url: https://github.com/kiali/openshift-servicemesh-plugin
        maintainers:
        - email: kiali-dev@googlegroups.com
          name: Kiali Developers Google Group
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel8@sha256:e37d76dd7c7a2db79ff1e96afef19b07c3beaa8ebf0c7bd85b808908b8f0ce77
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9-operator@sha256:0612dfa52f2bf39a1d87e640409313585fedc7cfcff6d00e77bb14809d27d646
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9@sha256:06fde6c696a33d17b0d2f7620ec46422c82277fd17a3fe6d7319de05c732aace
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:976a31ca8cf17ac0887b8a586edfcc29eb30175802461b3114758af525e7b800
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:66e4d933a9d69158bf87dd45835183bf2308b77c48908ab64496c5d22c816293
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:27a6aebffa3d049ee48050126e6107db2e07c49e9ce9297175facb6368e1e792
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-ossmc-rhel9@sha256:99643f294c3f92a6fc82f321cab7c756e4879fefb48fa824f1bf49ff7182658d
        version: 2.1.2
      entries:
      - name: kiali-operator.v2.1.2
        version: 2.1.2
      - name: kiali-operator.v1.89.8
        version: 1.89.8
      - name: kiali-operator.v1.89.7
        version: 1.89.7
      - name: kiali-operator.v1.89.6
        version: 1.89.6
      - name: kiali-operator.v1.89.1
        version: 1.89.1
      - name: kiali-operator.v1.73.10
        version: 1.73.10
      - name: kiali-operator.v1.73.8
        version: 1.73.8
      - name: kiali-operator.v1.73.7
        version: 1.73.7
      - name: kiali-operator.v1.73.4
        version: 1.73.4
      - name: kiali-operator.v1.65.11
        version: 1.65.11
      - name: kiali-operator.v1.65.10
        version: 1.65.10
      - name: kiali-operator.v1.65.9
        version: 1.65.9
      - name: kiali-operator.v1.65.8
        version: 1.65.8
      - name: kiali-operator.v1.65.7
        version: 1.65.7
      - name: kiali-operator.v1.65.6
        version: 1.65.6
      - name: kiali-operator.v1.57.7
        version: 1.57.7
      name: candidate
    - currentCSV: kiali-operator.v2.11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "Kiali",
                "metadata": {
                  "name": "kiali"
                },
                "spec": {
                  "version": "default",
                  "istio_namespace": "istio-system",
                  "deployment": {
                    "logger": {
                      "log_level": "info"
                    },
                    "view_only_mode": false
                  }
                }
              },
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "OSSMConsole",
                "metadata": {
                  "name": "ossmconsole"
                },
                "spec": {
                  "version": "default",
                  "kiali": {
                    "serviceName": "",
                    "serviceNamespace": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/kiali-rhel9-operator@sha256:2970861bd8e65e36d5080dc4d56340d08a52dbb1217be352998d135eb854732f
          createdAt: "2025-09-10T00:28:49Z"
          description: This productized operator provides Kiali and OSSMC. Kiali is
            the Istio observability and management Console. OSSMC is the OpenShift
            Service Mesh Console plugin, powered by Kiali.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kiali/kiali
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A configuration file for a Kiali installation.
            displayName: Kiali
            kind: Kiali
            name: kialis.kiali.io
            version: v1alpha1
          - description: A configuration file for a OpenShift Service Mesh Console
              installation.
            displayName: OpenShift Service Mesh Console
            kind: OSSMConsole
            name: ossmconsoles.kiali.io
            version: v1alpha1
        description: |-
          ## About the managed application

          A Microservice Architecture breaks up the monolith into many smaller pieces
          that are composed together. Patterns to secure the communication between
          services like fault tolerance (via timeout, retry, circuit breaking, etc.)
          have come up as well as distributed tracing to be able to see where calls
          are going.

          A service mesh can now provide these services on a platform level and frees
          the application writers from those tasks. Routing decisions are done at the
          mesh level.

          Kiali works with OpenShift Service Mesh to visualize the service
          mesh topology, to provide visibility into features like circuit breakers,
          request rates and more. It offers insights about the mesh components at
          different levels, from abstract Applications to Services and Workloads.

          See [https://www.kiali.io](https://www.kiali.io) to read more.

          ### Accessing the UI

          By default, the Kiali operator exposes the Kiali UI as an OpenShift Route.

          If on OpenShift, you can create an OSSMConsole CR to have the operator
          install the OpenShift ServiceMesh Console plugin to the OpenShift Console
          thus providing an interface directly integrated with the OpenShift Console.

          ## About this Operator

          ### Kiali Custom Resource Configuration Settings

          For quick descriptions of all the settings you can configure in the Kiali
          Custom Resource (CR), see
          [the kiali.io docs](https://kiali.io/docs/configuration/kialis.kiali.io/).

          ## Prerequisites for enabling this Operator

          Kiali is a companion tool for OpenShift Service Mesh. So before you install Kiali, you must have
          already installed Service Mesh.
        displayName: Kiali Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - service-mesh
        - observability
        - monitoring
        - istio
        links:
        - name: Installation Guide
          url: https://www.kiali.io/docs/installation
        - name: Features
          url: https://www.kiali.io/docs/features
        - name: Documentation Home
          url: https://www.kiali.io
        - name: Blogs and Articles
          url: https://medium.com/kialiproject
        - name: Server and UI Source Code
          url: https://github.com/kiali/kiali
        - name: Operator Source Code
          url: https://github.com/kiali/kiali-operator
        - name: OpenShift Service Mesh Console Code
          url: https://github.com/kiali/openshift-servicemesh-plugin
        maintainers:
        - email: kiali-dev@googlegroups.com
          name: Kiali Developers Google Group
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9-operator@sha256:2970861bd8e65e36d5080dc4d56340d08a52dbb1217be352998d135eb854732f
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:7487fbb4b776dfca28ad09311ba857e07ad98bc91ca5ee6fcaaa5b7783d04fbd
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel9@sha256:9a0d4862a25c2999a3f85b9a7a0332dab4ec1744548191d7a00c838b06ba26d2
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel9@sha256:fe25be563d50e5ad4e5475f153e03d4b24c7db094d7e824c51ebaae79418b1f7
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel8@sha256:81db94df331799f4087596378914b712d85d117dde3df2d4902079a82d4a9c55
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9@sha256:a00ec00e466e080592dca5c1cc153698b1412a54e1979647f74281f79b5691cf
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9@sha256:85eeb972e46f26ff30e04991e76ec3050d98b334e7b7f201a094f8d50b06854e
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:d729ef8d74c2f6e02d8c68ca761b1fb3c7c3706f1290b684468f0acb0ef313e0
        version: 2.11.3
      entries:
      - name: kiali-operator.v2.11.3
        version: 2.11.3
      - name: kiali-operator.v2.11.2
        version: 2.11.2
      - name: kiali-operator.v2.11.1
        version: 2.11.1
      - name: kiali-operator.v2.4.7
        version: 2.4.7
      - name: kiali-operator.v2.4.6
        version: 2.4.6
      - name: kiali-operator.v2.4.5
        version: 2.4.5
      - name: kiali-operator.v2.4.4
        version: 2.4.4
      - name: kiali-operator.v2.4.3
        version: 2.4.3
      - name: kiali-operator.v1.89.10
        version: 1.89.10
      - name: kiali-operator.v1.89.9
        version: 1.89.9
      - name: kiali-operator.v1.89.8
        version: 1.89.8
      - name: kiali-operator.v1.89.7
        version: 1.89.7
      - name: kiali-operator.v1.89.6
        version: 1.89.6
      - name: kiali-operator.v1.89.1
        version: 1.89.1
      - name: kiali-operator.v1.73.10
        version: 1.73.10
      - name: kiali-operator.v1.73.8
        version: 1.73.8
      - name: kiali-operator.v1.73.7
        version: 1.73.7
      - name: kiali-operator.v1.73.4
        version: 1.73.4
      - name: kiali-operator.v1.65.11
        version: 1.65.11
      - name: kiali-operator.v1.65.10
        version: 1.65.10
      - name: kiali-operator.v1.65.9
        version: 1.65.9
      - name: kiali-operator.v1.65.8
        version: 1.65.8
      - name: kiali-operator.v1.65.7
        version: 1.65.7
      - name: kiali-operator.v1.65.6
        version: 1.65.6
      - name: kiali-operator.v1.57.7
        version: 1.57.7
      name: stable
    defaultChannel: stable
    packageName: kiali-ossm
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: test-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: test-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "test.openstack.org/v1beta1",
                "kind": "HorizonTest",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "test-operator",
                    "app.kubernetes.io/instance": "horizontest-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "horizontest",
                    "app.kubernetes.io/part-of": "test-operator"
                  },
                  "name": "horizontest-sample"
                },
                "spec": {
                  "adminPassword": "12345678",
                  "adminUsername": "admin",
                  "authUrl": "https://keystone-public-openstack.apps.ocp.openstack.lab",
                  "backoffLimit": 0,
                  "containerImage": "",
                  "dashboardUrl": "https://horizon-openstack.apps.ocp.openstack.lab/",
                  "flavorName": "m1.tiny",
                  "horizonRepoBranch": "master",
                  "horizonTestDir": "/var/lib/horizontest",
                  "imageUrl": "http://download.cirros-cloud.net/0.6.2/cirros-0.6.2-x86_64-disk.img",
                  "logsDirectoryName": "horizon",
                  "parallel": false,
                  "password": "horizontest",
                  "projectName": "horizontest",
                  "repoUrl": "https://review.opendev.org/openstack/horizon",
                  "storageClass": "local-storage",
                  "user": "horizontest"
                }
              },
              {
                "apiVersion": "test.openstack.org/v1beta1",
                "kind": "Tempest",
                "metadata": {
                  "name": "tempest-tests",
                  "namespace": "openstack"
                },
                "spec": {
                  "containerImage": "",
                  "tempestRun": {
                    "concurrency": 8,
                    "includeList": "tempest.api.identity.v3.*\n"
                  },
                  "tempestconfRun": null
                }
              },
              {
                "apiVersion": "test.openstack.org/v1beta1",
                "kind": "Tobiko",
                "metadata": {
                  "name": "tobiko-tests",
                  "namespace": "openstack"
                },
                "spec": {
                  "config": "[DEFAULT]\nlog_file = value\n\n[testcase]\ntimeout = value\ntest_runner_timeout = value\n\n[ubuntu]\ninterface_name = value\n\n[keystone]\ninterface_name = value\n",
                  "containerImage": "",
                  "testenv": "py3",
                  "version": "master",
                  "workflow": [
                    {
                      "stepName": "unit-tests",
                      "testenv": "py3"
                    },
                    {
                      "stepName": "functional-tests",
                      "testenv": "functional"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:43Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: Ansible Test
            kind: AnsibleTest
            name: ansibletests.test.openstack.org
            version: v1beta1
          - displayName: Horizon Test
            kind: HorizonTest
            name: horizontests.test.openstack.org
            version: v1beta1
          - displayName: Tempest
            kind: Tempest
            name: tempests.test.openstack.org
            version: v1beta1
          - kind: Tobiko
            name: tobikoes.test.openstack.org
            version: v1beta1
        description: Test Operator
        displayName: OpenStack Test Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Tempest
        - Tobiko
        - Tests
        links:
        - name: Test Operator
          url: https://github.com/openstack-k8s-operators/test-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/test-rhel9-operator@sha256:17949ad6a69a14ea38870d471fd4f5a51122927f795976ac7e2b0189e2d0da5a
        - registry.redhat.io/rhoso/openstack-tempest-all-rhel9@sha256:2f07239d932a67a6589844f4380f613cc16614ff4de7699f64c17359bbdc5534
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: test-operator.v1.0.6
        version: 1.0.6
      - name: test-operator.v1.0.5
        version: 1.0.5
      - name: test-operator.v1.0.4
        version: 1.0.4
      - name: test-operator.v1.0.3
        version: 1.0.3
      - name: test-operator.v1.0.2
        version: 1.0.2
      - name: test-operator.v1.0.1
        version: 1.0.1
      - name: test-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: test-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhsso-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhsso-operator.7.5.1-opr-005
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm"
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "False"
          containerImage: registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:66577ec32a9a975f304e52b8ab49b9f7602158dcab1ef5d8c28a5d3455491336
          createdAt: "2021-08-09 00:00:00"
          description: An Operator for installing and managing Red Hat Single Sign-On
          olm.skipRange: <7.5.0
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Red Hat Single Sign-On Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Red Hat Single Sign-On.

          Red Hat Single Sign-On lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
        displayName: Red Hat Single Sign-On Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://access.redhat.com/products/red-hat-single-sign-on
        - name: Keycloak
          url: https://www.keycloak.org/
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat Customer Service
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso7-rhel8-init-container@sha256:614a9b1d7975769e8559e0b8c1578d16f6e3f7fee0778283797f4d62f5fc2100
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:5f0840a69642d13016430151892ec987fd67e711563b1dfb1a57353037c3bf79
        - registry.redhat.io/rh-sso-7/sso75-openj9-openshift-rhel8@sha256:2d4ea7226f43e7e8e8d3fcef6d425eb5104a15acf07791355033300b80db2901
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:66577ec32a9a975f304e52b8ab49b9f7602158dcab1ef5d8c28a5d3455491336
        version: 7.5.1-opr-005
      entries:
      - name: rhsso-operator.7.5.1-opr-005
        version: 7.5.1-opr-005
      - name: rhsso-operator.7.5.1-opr-004
        version: 7.5.1-opr-004
      - name: rhsso-operator.7.5.1-opr-003
        version: 7.5.1-opr-003
      - name: rhsso-operator.7.5.1-opr-002
        version: 7.5.1-opr-002
      - name: rhsso-operator.7.5.1-opr-001
        version: 7.5.1-opr-001
      - name: rhsso-operator.7.5.0
        version: 7.5.0
      name: alpha
    - currentCSV: rhsso-operator.7.6.12-opr-004
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm"
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "False"
          containerImage: registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:229265579ec77f2bd18c936ba19e1f9dbc6b2f0a13d963d5eae5b52c260010d1
          createdAt: "2021-08-09 00:00:00"
          description: An Operator for installing and managing Red Hat Single Sign-On
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Red Hat Single Sign-On Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Red Hat Single Sign-On.

          Red Hat Single Sign-On lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
        displayName: Red Hat Single Sign-On Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://access.redhat.com/products/red-hat-single-sign-on
        - name: Keycloak
          url: https://www.keycloak.org/
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat Customer Service
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:229265579ec77f2bd18c936ba19e1f9dbc6b2f0a13d963d5eae5b52c260010d1
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:923206112ed0873ee58edd0817dbce7bb4c9a89dc75d9b3ba2affe9aed4811de
        - registry.redhat.io/rh-sso-7/sso7-rhel8-init-container@sha256:282daf78934ca495ad4051371a58e3af0917bb9241d043a66e23fc8b12fbd1ce
        version: 7.6.12-opr-004
      entries:
      - name: rhsso-operator.7.6.12-opr-004
        version: 7.6.12-opr-004
      - name: rhsso-operator.7.6.12-opr-003
        version: 7.6.12-opr-003
      - name: rhsso-operator.7.6.12-opr-002
        version: 7.6.12-opr-002
      - name: rhsso-operator.7.6.12-opr-001
        version: 7.6.12-opr-001
      - name: rhsso-operator.7.6.11-opr-006
        version: 7.6.11-opr-006
      - name: rhsso-operator.7.6.11-opr-005
        version: 7.6.11-opr-005
      - name: rhsso-operator.7.6.11-opr-004
        version: 7.6.11-opr-004
      - name: rhsso-operator.7.6.11-opr-003
        version: 7.6.11-opr-003
      - name: rhsso-operator.7.6.11-opr-002
        version: 7.6.11-opr-002
      - name: rhsso-operator.7.6.11-opr-001
        version: 7.6.11-opr-001
      - name: rhsso-operator.7.6.10-opr-001
        version: 7.6.10-opr-001
      - name: rhsso-operator.7.6.9-opr-003
        version: 7.6.9-opr-003
      - name: rhsso-operator.7.6.9-opr-002
        version: 7.6.9-opr-002
      - name: rhsso-operator.7.6.9-opr-001
        version: 7.6.9-opr-001
      - name: rhsso-operator.7.6.8-opr-002
        version: 7.6.8-opr-002
      - name: rhsso-operator.7.6.8-opr-001
        version: 7.6.8-opr-001
      - name: rhsso-operator.7.6.7-opr-002
        version: 7.6.7-opr-002
      - name: rhsso-operator.7.6.7-opr-001
        version: 7.6.7-opr-001
      - name: rhsso-operator.7.6.6-opr-004
        version: 7.6.6-opr-004
      - name: rhsso-operator.7.6.6-opr-003
        version: 7.6.6-opr-003
      - name: rhsso-operator.7.6.6-opr-002
        version: 7.6.6-opr-002
      - name: rhsso-operator.7.6.6-opr-001
        version: 7.6.6-opr-001
      - name: rhsso-operator.7.6.5-opr-004
        version: 7.6.5-opr-004
      - name: rhsso-operator.7.6.5-opr-003
        version: 7.6.5-opr-003
      - name: rhsso-operator.7.6.5-opr-002
        version: 7.6.5-opr-002
      - name: rhsso-operator.7.6.5-opr-001
        version: 7.6.5-opr-001
      - name: rhsso-operator.7.6.4-opr-003
        version: 7.6.4-opr-003
      - name: rhsso-operator.7.6.4-opr-002
        version: 7.6.4-opr-002
      - name: rhsso-operator.7.6.4-opr-001
        version: 7.6.4-opr-001
      - name: rhsso-operator.7.6.3-opr-002
        version: 7.6.3-opr-002
      - name: rhsso-operator.7.6.3-opr-001
        version: 7.6.3-opr-001
      - name: rhsso-operator.7.6.2-opr-003
        version: 7.6.2-opr-003
      - name: rhsso-operator.7.6.2-opr-002
        version: 7.6.2-opr-002
      - name: rhsso-operator.7.6.2-opr-001
        version: 7.6.2-opr-001
      - name: rhsso-operator.7.6.1-opr-005
        version: 7.6.1-opr-005
      - name: rhsso-operator.7.6.1-opr-004
        version: 7.6.1-opr-004
      - name: rhsso-operator.7.6.1-opr-003
        version: 7.6.1-opr-003
      - name: rhsso-operator.7.6.1-opr-002
        version: 7.6.1-opr-002
      - name: rhsso-operator.7.6.1-opr-001
        version: 7.6.1-opr-001
      - name: rhsso-operator.7.6.0-opr-005
        version: 7.6.0-opr-005
      - name: rhsso-operator.7.6.0-opr-004
        version: 7.6.0-opr-004
      - name: rhsso-operator.7.6.0-opr-003
        version: 7.6.0-opr-003
      - name: rhsso-operator.7.6.0-opr-002
        version: 7.6.0-opr-002
      - name: rhsso-operator.7.6.0-opr-001
        version: 7.6.0-opr-001
      - name: rhsso-operator.7.5.2-opr-003
        version: 7.5.2-opr-003
      - name: rhsso-operator.7.5.2-opr-002
        version: 7.5.2-opr-002
      - name: rhsso-operator.7.5.2-opr-001
        version: 7.5.2-opr-001
      - name: rhsso-operator.7.5.1-opr-011
        version: 7.5.1-opr-011
      - name: rhsso-operator.7.5.1-opr-010
        version: 7.5.1-opr-010
      - name: rhsso-operator.7.5.1-opr-009
        version: 7.5.1-opr-009
      - name: rhsso-operator.7.5.1-opr-009-0.1646836600.p
        version: 7.5.1-opr-009+0.1646836600.p
      - name: rhsso-operator.7.5.1-opr-008
        version: 7.5.1-opr-008
      - name: rhsso-operator.7.5.1-opr-007
        version: 7.5.1-opr-007
      - name: rhsso-operator.7.5.1-opr-006
        version: 7.5.1-opr-006
      - name: rhsso-operator.7.5.1-opr-005
        version: 7.5.1-opr-005
      - name: rhsso-operator.7.5.1-opr-004
        version: 7.5.1-opr-004
      - name: rhsso-operator.7.5.1-opr-003
        version: 7.5.1-opr-003
      - name: rhsso-operator.7.5.1-opr-002
        version: 7.5.1-opr-002
      - name: rhsso-operator.7.5.1-opr-001
        version: 7.5.1-opr-001
      - name: rhsso-operator.7.5.0
        version: 7.5.0
      name: stable
    defaultChannel: stable
    packageName: rhsso-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-glue-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-glue-controller.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/glue-controller:0.2.1
          createdAt: "2025-09-19T20:55:49Z"
          description: AWS Glue controller is a service controller for managing Glue
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Job represents the state of an AWS glue Job resource.
            displayName: Job
            kind: Job
            name: jobs.glue.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Glue resources in AWS from within your Kubernetes cluster.

          **About Amazon Glue**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Glue
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - glue
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Glue Developer Resources
          url: https://aws.amazon.com/Glue/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: glue maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/glue-controller:0.2.1
        version: 0.2.1
      entries:
      - name: ack-glue-controller.v0.2.1
        version: 0.2.1
      - name: ack-glue-controller.v0.2.0
        version: 0.2.0
      - name: ack-glue-controller.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ack-glue-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Protocol Labs
      provider-url: ""
    name: ipfs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ipfs-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.ipfs.io/v1alpha1",
                "kind": "CircuitRelay",
                "metadata": {
                  "name": "circuitrelay-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cluster.ipfs.io/v1alpha1",
                "kind": "IpfsCluster",
                "metadata": {
                  "name": "ipfscluster-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Storage, Networking, Big Data
          containerImage: quay.io/redhat-et-ipfs/ipfs-operator:v0.0.1
          createdAt: 2022-04-05T15:50:37-0400
          operatorframework.io/suggested-namespace: ipfs-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ipfs-cluster/ipfs-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CircuitRelay is the Schema for the circuitrelays API.
            displayName: Circuit Relay
            kind: CircuitRelay
            name: circuitrelays.cluster.ipfs.io
            version: v1alpha1
          - description: IpfsCluster is the Schema for the ipfs API.
            displayName: Ipfs Cluster
            kind: IpfsCluster
            name: ipfsclusters.cluster.ipfs.io
            version: v1alpha1
        description: "The IPFS Operator enables users to easily create IPFS Cluster
          instances\nwithin their Kubernetes or OpenShift clusters. \n\n### Supported
          Features\n* Horizontal scaling\n* Joining collaborative clusters\n* Controlling
          IPFS node resource usage\n* Specifying reprovider settings  \n\n### Documentation
          \nDocumentation can be found on our [read-the-docs page](https://ipfs-operator.readthedocs.io/en/latest/),
          as well as our [GitHub repository](https://github.com/ipfs-cluster/ipfs-operator).\n\n###
          Getting Help\nIn case you run into issues while using the operator, please
          create an issue in our [GitHub\nrepository](https://github.com/ipfs-cluster/ipfs-operator).\n\n###
          Contributing\nWe welcome contributions of all types and sizes! You can find
          open issues\nin our [GitHub Repository](https://github.com/ipfs-cluster/ipfs-operator/issues),\nas
          well as make your own issue if you feel that something is missing.\n\n###
          License\nIPFS Operator is licensed under the [Apache 2.0 License](https://github.com/ipfs-cluster/ipfs-operator/blob/main/LICENSE)"
        displayName: IPFS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipfs
        - ipfs cluster
        - decentralized storage
        - peer-to-peer
        - distributed hash table
        - content-addressed
        links:
        - name: IPFS Operator GitHub page
          url: https://github.com/ipfs-cluster/ipfs-operator
        - name: IPFS Cluster
          url: https://ipfscluster.io/
        - name: IPFS
          url: https://ipfs.io/
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Protocol Labs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/redhat-et-ipfs/ipfs-operator:v0.0.1
        version: 0.0.2
      entries:
      - name: ipfs-operator.v0.0.2
        version: 0.0.2
      - name: ipfs-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ipfs-operator
    provider:
      name: Protocol Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: gaudi-ai-sw-tools-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gaudi-ai-sw-tools-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aitools.intel/v1",
                "kind": "GaudiAIToolsContainer",
                "metadata": {
                  "name": "intel-gaudi-1.17.0-495"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagespec": {
                    "registry": "registry.connect.redhat.com",
                    "repo": "intel/gaudi-notebooks",
                    "tags": [
                      {
                        "gaudi_software": "1.17.0-495",
                        "rhel_os": "9.2"
                      }
                    ]
                  },
                  "nameOverride": "",
                  "namespace": "redhat-ods-applications"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: registry.connect.redhat.com/intel/gaudi-ai-sw-tools-operator@sha256:0c286f1e4cec57a189d749501f82345afbe0a1dbbbf92b906ad3c549e47b89cd
          createdAt: "2024-09-11T21:27:02Z"
          description: Operator to enable Gaudi AI SW Tools into OpenShift AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: GaudiAIToolsContainer
            name: gaudiaitoolscontainers.aitools.intel
            version: v1
        description: Operator to enable Gaudi AI SW Tools into OpenShift AI
        displayName: Intel® Gaudi AI SW Tools Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gaudi
        - opendatahub
        - openshiftai
        - notebooks
        links:
        - name: Intel® Gaudi AI SW Tools Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/66d7aa630eb66a02febc8103
        maintainers:
        - email: sharvil.shah@intel.com
          name: sharvil
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.connect.redhat.com/intel/gaudi-ai-sw-tools-operator@sha256:0c286f1e4cec57a189d749501f82345afbe0a1dbbbf92b906ad3c549e47b89cd
        version: 0.0.2
      entries:
      - name: gaudi-ai-sw-tools-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: gaudi-ai-sw-tools-operator
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-certified-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "psp": false,
                  "serviceAccount": "default",
                  "controller": {
                    "enabled": true,
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": "registry.connect.redhat.com/neuvector/controller",
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": ""
                      }
                    },
                    "pvc": {
                      "enabled": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "path": "/",
                          "annotations": {
                            "ingress.kubernetes.io/protocol": "https"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "path": "/",
                          "annotations": {
                            "ingress.kubernetes.io/protocol": "https"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "path": "/",
                      "annotations": {
                        "ingress.kubernetes.io/protocol": "https"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {}
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/enforcer",
                    "priorityClassName": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      }
                    ],
                    "resources": {}
                  },
                  "manager": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/manager",
                    "priorityClassName": "",
                    "env": {
                      "ssl": true
                    },
                    "svc": {
                      "type": "NodePort",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "path": "/",
                      "annotations": {},
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "affinity": {},
                    "tolerations": [],
                    "nodeSelector": {}
                  },
                  "cve": {
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "image": "registry.access.redhat.com/ubi8@sha256:228824aa581f3b31bf79411f8448b798291c667a37155bdea61cfa128b2833f2",
                      "schedule": "0 0 * * *",
                      "priorityClassName": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": "registry.connect.redhat.com/neuvector/scanner@sha256:42004e13b348bb43b58ae8538c983a70b237e3aca870758ac629fb4fc43e8130",
                      "priorityClassName": "",
                      "resources": {},
                      "affinity": {},
                      "tolerations": [],
                      "nodeSelector": {}
                    }
                  },
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "resources": {},
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "true"
          containerImage: registry.connect.redhat.com/neuvector/neuvector-operator@sha256:6155ac3c95aa6471c9857e0a6f5e5d19e6fc1d8cc7d014108d7d4a10d14f96bb
          createdAt: "2020-12-02 14:19:21"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/neuvector-certified-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/neuvector-certified-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: Neuvector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm/tree/master/charts/core) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart.  NeuVector Operator versions are tied to NeuVector product versions. Version 1.3.4 of the NeuVector Certified Operator deploys version 4.4.3 of NeuVector.

          **Complete below steps to Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Grant Service Account Access to the Privileged SCC

                   oc -n neuvector adm policy add-scc-to-user privileged -z default

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s                    neuvector/default


          **Add NeuVector license from NeuVector WebUI->setting**


          #Deploying the NeuVector Operator#


          Please refer to the instructions [here](https://github.com/neuvector/neuvector-operator/blob/master/README.md)


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Networking
        - Security
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - registry.connect.redhat.com/neuvector/scanner@sha256:42004e13b348bb43b58ae8538c983a70b237e3aca870758ac629fb4fc43e8130
        - registry.access.redhat.com/ubi8@sha256:228824aa581f3b31bf79411f8448b798291c667a37155bdea61cfa128b2833f2
        - registry.connect.redhat.com/neuvector/neuvector-operator@sha256:6155ac3c95aa6471c9857e0a6f5e5d19e6fc1d8cc7d014108d7d4a10d14f96bb
        - registry.connect.redhat.com/neuvector/controller@sha256:acc943568520282061547b310c50080dc32924a531bbb543d959b281eb6664f4
        - registry.connect.redhat.com/neuvector/enforcer@sha256:be1d299ec3532d55444681a882e9fd1238949da6598987308e2800ed26291b7a
        - registry.connect.redhat.com/neuvector/manager@sha256:8245cd179d11fff279690c5a422f46b4ff252ed72b9b998f0c9b5453bc7d8fc0
        version: 1.3.4
      entries:
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-certified-operator-rhmp
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dpu-network-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dpu-network-operator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dpu.openshift.io/v1alpha1",
                "kind": "OVNKubeConfig",
                "metadata": {
                  "name": "ovnkubeconfig-sample"
                },
                "spec": {
                  "kubeConfigFile": "tenant-cluster-1-kubeconf",
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/dpu-worker": ""
                    }
                  },
                  "poolName": "dpu"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional, Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/dpu-network-rhel8-operator@sha256:d07b8ecb5d1625599878b5c1e34b0de94188b81d90c829787a2d13969f6d5971
          createdAt: "2023-07-13T18:03:11Z"
          description: The operator is responsible for the life-cycle management of
            the ovn-kube components and the necessary host network initialization
            on DPU cards.
          olm.skipRange: '>=4.10.0-0 <4.13.0-202309181427'
          operatorframework.io/suggested-namespace: openshift-dpu-network-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/dpu-network-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OVNKubeConfig is the Schema for the ovnkubeconfigs API
            displayName: OVNKube Config
            kind: OVNKubeConfig
            name: ovnkubeconfigs.dpu.openshift.io
            version: v1alpha1
        description: The operator to be responsible for the life-cycle management
          of the ovn-kube components and the necessary host network initialization
          on DPU cards.
        displayName: DPU Network Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dpu
        - network
        links:
        - name: Dpu Network Operator
          url: https://dpu-network-operator.domain
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:767682dd3bd65651a8204c9fb732b9c48b99127189992b43abc6ecce027f4589
        - registry.redhat.io/openshift4/dpu-network-rhel8-operator@sha256:d07b8ecb5d1625599878b5c1e34b0de94188b81d90c829787a2d13969f6d5971
        version: 4.13.0-202309181427
      entries:
      - name: dpu-network-operator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: stable
    defaultChannel: stable
    packageName: dpu-network-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-odlm
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operand-deployment-lifecycle-manager.v4.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandBindInfo",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "bindings": {
                    "public": {
                      "configmap": "mongodb-configmap",
                      "secret": "mongodb-secret"
                    }
                  },
                  "description": "Binding information that should be accessible to MongoDB adopters",
                  "operand": "mongodb-atlas-kubernetes",
                  "registry": "example-service"
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "services": [
                    {
                      "name": "jaeger",
                      "spec": {
                        "jaeger": {
                          "strategy": "allinone"
                        }
                      }
                    },
                    {
                      "name": "mongodb-atlas-kubernetes",
                      "spec": {
                        "atlasDeployment": {
                          "deploymentSpec": {
                            "name": "test-deployment"
                          },
                          "projectRef": {
                            "name": "my-project"
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "operators": [
                    {
                      "channel": "stable",
                      "installMode": "cluster",
                      "name": "jaeger",
                      "namespace": "default",
                      "packageName": "jaeger",
                      "sourceName": "community-operators",
                      "sourceNamespace": "openshift-marketplace"
                    },
                    {
                      "channel": "stable",
                      "name": "mongodb-atlas-kubernetes",
                      "namespace": "default",
                      "packageName": "mongodb-atlas-kubernetes",
                      "sourceName": "community-operators",
                      "sourceNamespace": "openshift-marketplace"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandRequest",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "requests": [
                    {
                      "operands": [
                        {
                          "name": "jaeger"
                        },
                        {
                          "name": "mongodb-atlas-kubernetes"
                        }
                      ],
                      "registry": "example-service"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, Monitoring, Logging & Tracing, Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
          createdAt: "2025-02-14T21:16:49Z"
          description: The Operand Deployment Lifecycle Manager provides a Kubernetes
            CRD-based API to manage the lifecycle of operands.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-odlm
          olm.skipRange: '>=1.2.0 <4.5.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/operand-deployment-lifecycle-manager
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OperandBindInfo is the Schema for the operandbindinfoes API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandBindInfo
            kind: OperandBindInfo
            name: operandbindinfos.operator.ibm.com
            version: v1alpha1
          - description: OperandConfig is the Schema for the operandconfigs API. Documentation
              For additional details regarding install parameters check https://ibm.biz/icpfs39install.
              License By installing this product you accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandConfig
            kind: OperandConfig
            name: operandconfigs.operator.ibm.com
            version: v1alpha1
          - description: OperandRegistry is the Schema for the operandregistries API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandRegistry
            kind: OperandRegistry
            name: operandregistries.operator.ibm.com
            version: v1alpha1
          - description: OperandRequest is the Schema for the operandrequests API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandRequest
            kind: OperandRequest
            name: operandrequests.operator.ibm.com
            version: v1alpha1
          - description: OperatorConfig is the Schema for the operatorconfigs API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.ibm.com
            version: v1alpha1
        description: |-
          # Introduction

          The operand is the instance managed by the operator. ODLM is used to manage the lifecycle of for a group of operands, compared with operator lifecycle manager, ODLM focus on the management of operands but not operators.

          # Details

          ### What are operands?

          Operator is a method of packaging, deploying and managing a Kubernetes application.
          Operands are the services and applications that Operator manage.

          For example, cert-manager operator deploys a cert-manager deployment, then the cert-manager-operator is an operator and the cert-manager deployment is an operand.

          The ODLM will have four CRDs:

          | Resource                 | Short Name | Description                                                                                |
          |--------------------------|------------|--------------------------------------------------------------------------------------------|
          | OperandRequest | opreq | It defines which operator/operand want to be installed in the cluster |
          | OperandRegistry | opreg | It defines the OLM information, like channel and catalog source, for each operator |
          | OperandConfig | opcon | It defines the parameters that should be used to install the operator's operand |
          | OperandBindInfo | opbi | It identifies secrets and/or configmaps that should be shared with requests |

          ### Goal

          1. A single entry point to manage a group of operands
          1. User can select a set of operands to install
          1. The install can be invoked through either OCP UI or CLI

          ## Prerequisites
          - [operator-sdk] version v1.3.0.
          - [go] version 1.15.7+
          - [oc] version v3.11+ or [kubectl] v1.11.3+
          - Access to an Openshift v4.3+ cluster

          ### Resources Required

          #### Minimum scheduling capacity

          | Software                   | Memory (MB) | CPU (cores) |
          | -------------------------- | ----------- | ----------- |
          | Operand Deployment Lifecycle Manager | 200          | 0.2        |
          | **Total**                  | 200         | 0.2         |

          ## Installing

          ### Create Operator NS, Group, Subscription

          ```yaml
          apiVersion: v1
          kind: Namespace
          metadata:
            name: odlm

          ---
          apiVersion: operators.coreos.com/v1alpha2
          kind: OperatorGroup
          metadata:
            name: operatorgroup
            namespace: odlm
          spec:
            targetNamespaces:
            - odlm

          ---
          apiVersion: v1
          data:
            namespaces: odlm
          kind: ConfigMap
          metadata:
            name: namespace-scope
            namespace: odlm

          ---
          apiVersion: operators.coreos.com/v1alpha1
          kind: Subscription
          metadata:
            name: operand-deployment-lifecycle-manager
            namespace: odlm
          spec:
            channel: v3.20
            name: ibm-odlm
            source: opencloud-operators
            sourceNamespace: olm
            config:
              env:
              - name: INSTALL_SCOPE
                value: namespaced
          END
          ```

          ## Configuration

          OperandRegistry defines the OLM information used for installation, like package name and catalog source, for each operator.

          Following is an example of the OperandRegistry CR:

          **NOTE:** The "spec.operators[*].name" parameter must be unique for each entry.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRegistry
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            operators:
            - name: jenkins [3]
              namespace: default [4]
              channel: alpha [5]
              packageName: jenkins-operator [6]
              scope: public [7]
              sourceName: community-operators [8]
              sourceNamespace: openshift-marketplace [9]
              installMode: cluster [10]
              installPlanApproval: Manual [11]
          ```

          The OperandRegistry Custom Resource (CR) lists OLM Operator information for operands that may be requested for installation and/or access by an application that runs in a namespace. The registry CR specifies:

          1. `name` of the OperandRegistry
          2. `namespace` of the OperandRegistry
          3. `name` is the name of the operator, which should be the same as the services name in the OperandConfig and OperandRequest.
          4. `namespace` defines the namespace where the operator and its CR will be deployed. (1) When InstallMode is `cluster`, the operator will be deployed into the `openshift-operators` namespace and the operator CRs will be deployed into the namespace this parameter defines. (2) When InstallMode is empty or set to `namespace`, it is the namespace where both operator and operator CR will be deployed.
          5. `channel` is the name of OLM channel that is subscribed for the operator.
          6. `packageName` is the name of the package in CatalogSource that is subscribed for the operator.
          7. (optional) `scope` is an indicator, either public or private, that dictates whether deployment can be requested from other namespaces (public) or only from the namespace of this OperandRegistry (private). The default value is private.
          8. `sourceName` is the name of the CatalogSource.
          9. `sourceNamespace` is the namespace of the CatalogSource.
          10. (optional) `installMode` is the install mode of the operator, can be either `namespace` (OLM one namespace) or `cluster` (OLM all namespaces). The default value is `namespace`. Operator is deployed in `openshift-operators` namespace when InstallMode is set to `cluster`.
          11. (optional) `installPlanApproval` is the approval mode for emitted installplan. The default value is `Automatic`.

          ### OperandConfig Spec

          OperandConfig defines the individual operand configuration. The OperandConfig Custom Resource (CR) defines the parameters for each operator that is listed in the OperandRegistry that should be used to install the operator instance by specifying an installation CR.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          Kind: OperandConfigs
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            services:
            - name: jenkins [3]
              spec: [4]
                jenkins:
                  port: 8081
          ```

          OperandConfig defines the individual operand deployment config:

          1. `name` of the OperandConfig
          2. `namespace` of the OperandConfig
          3. `name` is the name of the operator, which should be the same as the services name in the OperandRegistry and OperandRequest.
          4. `spec` defines a map. Its key is the kind name of the custom resource. Its value is merged to the spec field of custom resource. For more details, you can check the following topic **How does ODLM create the individual operator CR?**

          #### How does Operator create the individual operator CR

          Jenkins Operator has one CRD: Jenkins:

          The OperandConfig CR has

          ```yaml
          - name: jenkins
            spec:
              jenkins:
                service:
                  port: 8081
          ```

          The Jenkins Operator CSV has

          ```yaml
          apiVersion: operators.coreos.com/v1alpha1
          kind: ClusterServiceVersion
          metadata:
            annotations:
            alm-examples: |-
              [
              {
                "apiVersion":"jenkins.io/v1alpha2",
                "kind":"Jenkins",
                "metadata": {
                  "name":"example"
                },
                "spec":{
                  ...
                  "service":{"port":8080,"type":"ClusterIP"},
                  ...
                }
              }
            ]
          ```

          The ODLM will deep merge the OperandConfig CR spec and Jenkins Operator CSV alm-examples to create the Jenkins CR.

          ```yaml
          apiVersion: jenkins.io/v1alpha2
          kind: Jenkins
          metadata:
            name: example
          spec:
            ...
            service:
              port: 8081
              type: ClusterIP
            ...
          ```

          For day2 operations, the ODLM will patch the OperandConfigs CR spec to the existing Jenkins CR.

          ### OperandRequest Spec

          OperandRequest defines which operator/operand you want to install in the cluster.

          **NOTE:** OperandRequest custom resource is used to trigger a deployment for Operators and Operands.

          #### OperandRequest sample to create custom resource via OperandConfig

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRequest
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            requests:
            - registry: example-service [3]
              registryNamespace: example-service-ns [4]
              operands: [5]
              - name: jenkins [6]
                bindings: [7]
                  public:
                    secret: jenkins-operator-credential [8]
                    configmap: jenkins-operator-base-configuration [9]
          ```

          1. `name` of the OperandRequest
          2. `namespace` of the OperandRequest
          3. `registry` identifies the name of the OperandRegistry CR from which this operand deployment is being requested.
          4. (optional) `registryNamespace` identifies the namespace in which the OperandRegistry CR is defined. **Note:** If the `registryNamespace` is not specified then it is assumed that the OperandRegistry CR is in the current (OperandRequest's) namespace.
          5. `operands` in the CR is a list of operands.
          6. `name` of operands in the CR must match a name specification in an OperandRegistry's CR.
          7. (optional) The `bindings` of the operands is a map to get and rename the secret and/or configmap from the provider and create them in the requester's namespace. If the requester wants to rename the secret and/or configmap, they need to know the key of the binding in the OperandBindInfo. If the key of the bindings map is prefixed with public, it means the secret and/or configmap can be shared with the requester in the other namespace. If the key of the bindings map is prefixed with private, it means the secret and/or configmap can only be shared within its own namespace.
          8. (optional) `secret` names a secret that should be created in the requester's namespace with formatted data that can be used to interact with the service.
          9. (optional) `configmap` names a configmap that should be created in the requester's namespace with formatted data that can be used to interact with the service.

          #### OperandRequest sample to create custom resource via OperandRequest

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRequest
          metadata:
            name: example-service
            namespace: example-service-ns
          spec:
            requests:
            - registry: example-service
              registryNamespace: example-service-ns
              operands:
              - name: jenkins
                kind: Jenkins [1]
                apiVersion: "jenkins.io/v1alpha2" [2]
                instanceName: "example" [3]
                spec: [4]
                  service:
                    port: 8081
          ```

          1. `kind` of the target custom resource. If it is set in the operand item, ODLM will create CR according to OperandRequest only and get rid of the alm-example and OperandConfig.
          2. `apiVersion` of the target custom resource.
          3. `instanceName` is the name of the custom resource. If `instanceName` is not set, the name of the custom resource will be created with the name of the OperandRequest as a prefix.
          4. `spec` is the spec field of the target CR.

          ### OperandBindInfo Spec

          The ODLM will use the OperandBindInfo to copy the generated secret and/or configmap to a requester's namespace when a service is requested with the OperandRequest CR. An example specification for an OperandBindInfo CR is shown below.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandBindInfo
          metadata:
            name: publicjenkinsbinding [1]
            namespace: example-service-ns [2]
          spec:
            operand: jenkins [3]
            registry: example-service [4]
            description: "Binding information that should be accessible to jenkins adopters" [5]
            bindings: [6]
              public:
                secret: jenkins-operator-credentials-example [7]
                configmap: jenkins-operator-base-configuration-example [8]
          ```

          Fields in this CR are described below.

          1. `name` of the OperandBindInfo
          2. `namespace` of the OperandBindInfo
          3. The `operand` should be the the individual operator name.
          4. The `registry` section must match the name in the OperandRegistry CR in the current namespace.
          5. `description` is used to add a detailed description of a service.
          6. The `bindings` section is used to specify information about the access/configuration data that is to be shared. If the key of the bindings map is prefixed with public, it means the secret and/or configmap can be shared with the requester in the other namespace. If the key of the bindings map is prefixed with private, it means the secret and/or configmap can only be shared within its own namespace. If the key of the bindings map is prefixed with protected, it means the secret and/or configmap can only be shared if it is explicitly declared in the OperandRequest.
          7. The `secret` field names an existing secret, if any, that has been created and holds information that is to be shared with the requester.
          8. The `configmap` field identifies a configmap object, if any, that should be shared with the requester

          ODLM will use the OperandBindInfo CR to pass information to an adopter when they create a OperandRequest to access the service, assuming that both have compatible scopes. ODLM will copy the information from the shared service's "OperandBindInfo.bindinfo[].secret" and/or "OperandBindInfo.bindinfo[].configmap" to the requester namespace.

          **NOTE:** If in the OperandRequest, there is no secret and/or configmap name specified in the bindings or no bindings field in the element of operands, ODLM will copy the secret and/or configmap to the requester's namespace and rename them to the name of the OperandBindInfo + secret/configmap name.

          ### E2E Use Case

          1. User installs ODLM from OLM

              The ODLM will automatically generate two default CRD CRs, since OperandRegistry and OperandConfigs CRs don't define the state, so it should be fine.

          2. **Optional:** Users can update the OperandConfigs CR with their own parameter values

              Users can skip this step if the default configurations match the requirement.

          3. User creates the OperandRequest CR from OLM

              This tells ODLM that users want to install some of the individual operator/operand.

          4. The rest works will be done by OLM and ODLM

              Finally, users will get what they want.

          ### Operator/Operand Upgrade

          - For operator/operand upgrade, you only need to publish your operator OLM to your operator channel, and OLM will handle the upgrade automatically.
          - If there are major version, then you may want to update `channel` in `OperandRegistry` to trigger upgrade.

          ### Limitations

          ODLM depends on NamespaceScope Operator

          ### SecurityContextConstraints Requirements

          The Operand Deployment Lifecycle Manager supports running under the OpenShift Container Platform default restricted security context constraints.
        displayName: Operand Deployment Lifecycle Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Operand Deployment Lifecycle Manager Project
          url: https://github.com/IBM/operand-deployment-lifecycle-manager
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d4dff1a8968c18d984e16c4c9f1517c3a10e23a11bf166c64d734773324d4c9d
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
        - registry.marketplace.redhat.com/cpopen/operand-deployment-lifecycle-manager-bundle@sha256:68f71f2767a40ca9333f0a670e07493e8a8c13a5c7ca877ea5bd8a544b352d9a
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:27195e9758afdf8521971866fd60ea3dafed91a146c3c4165876e5ea476a4622
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bb074bdf89cbc7c703fadb17639608599ab4f66db9be5617e5f79a0550676790
        version: 4.5.0
      entries:
      - name: operand-deployment-lifecycle-manager.v4.5.0
        version: 4.5.0
      name: v4.5
    defaultChannel: v4.5
    packageName: ibm-odlm
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fujitsu
      provider-url: https://www.postgresql.fastware.com/
    name: fujitsu-enterprise-postgres-operator-with-cg-module
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPAction",
                "metadata": {
                  "name": "new-fep-action"
                },
                "spec": {
                  "fepAction": {
                    "args": [
                      "new-fep-sts-0",
                      "new-fep-sts-1"
                    ],
                    "type": "reload"
                  },
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "targetClusterName": "new-fep"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPExporter",
                "metadata": {
                  "name": "new-fep-exporter"
                },
                "spec": {
                  "fepExporter": {
                    "exporterLogLevel": "error",
                    "fepClusterList": [
                      "new-fep1"
                    ],
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "userCustomQueries": "usr_example:\n  query: \"SELECT EXTRACT(EPOCH FROM (now() - pg_last_xact_replay_timestamp())) as lag\"\n  master: true\n  metrics:\n    - lag:\n        usage: \"GAUGE\"\n        description: \"Replication lag behind master in seconds\""
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPLogging",
                "metadata": {
                  "name": "new-fep-logging"
                },
                "spec": {
                  "fepLogging": {
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPPgpool2",
                "metadata": {
                  "name": "new-fep-pgpool2"
                },
                "spec": {
                  "clientAuthMethod": "scram",
                  "count": 2,
                  "customhba": "local   all         all                               trust\nhost    all         all         127.0.0.1/32          trust\nhost    all         all         ::1/128               trust\n",
                  "customlogsize": "128Mi",
                  "customparams": "listen_addresses = '*'\npcp_listen_addresses = '*'\nnum_init_children = 32\nreserved_connections = 0\nenable_pool_hba = off\nallow_clear_text_frontend_auth = off\nauthentication_timeout = 80\nbackend_weight0 = 1\nbackend_weight1 = 1\nbackend_flag0 = 'ALWAYS_PRIMARY'\nbackend_flag1 = 'DISALLOW_TO_FAILOVER'\nconnection_cache = on\nmax_pool = 4\nlisten_backlog_multiplier = 2\nserialize_accept = off\nchild_life_time = 300\nclient_idle_limit = 0\nchild_max_connections = 0\nconnection_life_time = 0\nreset_query_list = 'ABORT; DISCARD ALL'\nclient_min_messages = info\nlog_min_messages = debug1\nlog_statement = on\nlog_per_node_statement = on\nlog_client_messages = on\nlog_hostname = on\nlog_connections = on\nlog_line_prefix = '%t: pid %p: '\nload_balance_mode = on\nignore_leading_white_space = on\nwhite_function_list = ''\nblack_function_list = 'currval,lastval,nextval,setval'\nblack_query_pattern_list = ''\ndatabase_redirect_preference_list = ''\napp_name_redirect_preference_list = ''\nallow_sql_comments = off\ndisable_load_balance_on_write = 'transaction'\nstatement_level_load_balance = on\nsr_check_period = 0\nsr_check_user = 'postgres'\ndelay_threshold = 0\nlog_standby_delay = 'none'\nssl = on\nssl_ciphers = 'HIGH:MEDIUM:+3DES:!aNULL'\nssl_prefer_server_ciphers = off\nssl_ecdh_curve = 'prime256v1'\nssl_dh_params_file = ''\nrelcache_expire = 0\nrelcache_size = 256\ncheck_temp_table = catalog\ncheck_unlogged_table = on\nenable_shared_relcache = off\nrelcache_query_target = primary\nwd_port0 = 9000\nfailover_on_backend_error = off\n",
                  "custompcp": "none",
                  "customsslcacert": "none",
                  "customsslcert": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                  "customsslkey": "none",
                  "fepclustername": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "limits": {
                    "cpu": "400m",
                    "memory": "512Mi"
                  },
                  "requests": {
                    "cpu": "200m",
                    "memory": "256Mi"
                  },
                  "serviceport": 9999,
                  "statusport": 9898
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPRestore",
                "metadata": {
                  "name": "new-fep-restore"
                },
                "spec": {
                  "fromFEPcluster": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "mcSpec": {
                    "limits": {
                      "cpu": "200m",
                      "memory": "300Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "restoretype": "latest",
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "toFEPcluster": "new-fep-2"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v2",
                "kind": "FEPCluster",
                "metadata": {
                  "name": "new-fep"
                },
                "spec": {
                  "fep": {
                    "customAnnotations": {
                      "allDeployments": {}
                    },
                    "forceSsl": true,
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "instances": 1,
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "podAntiAffinity": false,
                    "podDisruptionBudget": false,
                    "servicePort": 27500,
                    "syncMode": "off",
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "usePodName": true
                  },
                  "fepChildCrVal": {
                    "backup": {
                      "image": {
                        "pullPolicy": "IfNotPresent"
                      },
                      "mcSpec": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      },
                      "pgbackrestParams": "# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[global]\nrepo1-retention-full=7\nrepo1-retention-full-type=time\nlog-path=/database/log/backup\n",
                      "postScript": " ",
                      "preScript": " ",
                      "schedule": {
                        "num": 2
                      },
                      "schedule1": {
                        "schedule": "15 0 * * 0",
                        "type": "full"
                      },
                      "schedule2": {
                        "schedule": "15 0 * * 1-6",
                        "type": "incr"
                      },
                      "schedule3": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule4": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule5": {
                        "schedule": " ",
                        "type": " "
                      }
                    },
                    "customPgAudit": "# define pg audit custom params here to override defaults.\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[output]\nlogger = 'auditlog'\nlog_directory = '/database/log/audit'\nlog_truncate_on_rotation = on\nlog_filename = 'pgaudit-%a.log'\nlog_rotation_age = 1d\nlog_rotation_size = 0\n[rule]\n",
                    "customPgHba": "# define pg_hba custom rules here to be merged with default rules.\n# TYPE     DATABASE        USER        ADDRESS        METHOD\n",
                    "customPgParams": "# define custom postgresql.conf parameters below to override defaults.\n# Current values are as per default FEP deployment\n# If you add a library to shared_preload_libraries, add it after the default value.\n# fsep_operator_security is available in FEP15 and later container images.\nshared_preload_libraries='pgx_datamasking,pg_prewarm,pg_stat_statements,fsep_operator_security'\nsession_preload_libraries='pg_prewarm'\nmax_prepared_transactions = 100\nmax_worker_processes = 30\nmax_connections = 100\nwork_mem = 1MB\nmaintenance_work_mem = 12MB\nshared_buffers = 128MB\neffective_cache_size = 384MB\ncheckpoint_completion_target = 0.8\n# tcp parameters\ntcp_keepalives_idle = 30\ntcp_keepalives_interval = 10\ntcp_keepalives_count = 3\n# logging parameters in default fep installation\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\nlog_directory = '/database/log'\nlog_filename = 'logfile-%a.log'\nlog_file_mode = 0600\nlog_truncate_on_rotation = on\nlog_rotation_age = 1d\nlog_rotation_size = 0\nlog_checkpoints = on\nlog_line_prefix = '%e %t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h'\nlog_lock_waits = on\nlog_autovacuum_min_duration = 60s\nlogging_collector = on\npgaudit.config_file='/opt/app-root/src/pgaudit-cfg/pgaudit.conf'\nlog_replication_commands = on\nlog_min_messages = WARNING\nlog_destination = csvlog\n# wal_archive parameters in default fep installation\narchive_mode = on\narchive_command = 'pgbackrest --stanza=backupstanza --config=/database/userdata/pgbackrest.conf archive-push %p'\nwal_level = replica\nmax_wal_senders = 12\nwal_keep_segments = 64\ntrack_activities = on\ntrack_counts = on\npassword_encryption = 'scram-sha-256'\n",
                    "storage": {
                      "archivewalVol": {
                        "size": "1Gi"
                      },
                      "backupVol": {
                        "size": "2Gi"
                      },
                      "dataVol": {
                        "size": "2Gi"
                      },
                      "logVol": {
                        "size": "1Gi"
                      },
                      "tablespaceVol": {
                        "size": "512Mi"
                      },
                      "walVol": {
                        "size": "1200Mi"
                      }
                    },
                    "sysUsers": {
                      "pgRewindPassword": "rewind_password",
                      "pgRewindUser": "rewind_user",
                      "pgdb": "mydb",
                      "pgpassword": "mydbpassword",
                      "pgreplpassword": "repluserpwd",
                      "pgrepluser": "repluser",
                      "pguser": "mydbuser",
                      "tdepassphrase": "tde-passphrase"
                    },
                    "systemCertificates": {
                      "cacrt": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                      "crt": "-----BEGIN CERTIFICATE-----\nMIIDUTCCAjmgAwIBAgIRAMocW3qMoHrD6qRvMPppMkMwDQYJKoZIhvcNAQELBQAw\nNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9yIEt1\nYmVybmV0ZXMwHhcNMjEwMjA2MDQzMjM2WhcNMjYwMjA1MDQzMjM2WjA/MRAwDgYD\nVQQKEwdGdWppdHN1MSswKQYDVQQDEyJGVUpJVFNVIEVudGVycHJpc2UgUG9zdGdy\nZXMgU2VydmVyMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA4AI33yvH\nZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I2e4SceTKi6O3C/I1XuvWlpng\n5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4ANB5JWWqDOjrRT3o7nRPGXfila\nbP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYpmjdbfxabTz07ig0+6/cwKoRR\nxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTFYvmAH7gcdssSFBt8NPlUATHE\nsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6Wdgmu5H2pDml8CDNLDv98Aj7i\n+I5SRKKcVPlnuQIDAQABo1AwTjAdBgNVHSUEFjAUBggrBgEFBQcDAQYIKwYBBQUH\nAwIwDAYDVR0TAQH/BAIwADAfBgNVHSMEGDAWgBQcwrrUO0u+FhIUuVdrDRCQRsi6\nZjANBgkqhkiG9w0BAQsFAAOCAQEAm5dxBoI9pScOCvRAchg4CprdRDSJb9K6yB3O\nnCAxnM47iHeXnY3WlnI388kHu8DU7O4ba1tJbGs3KY9KzioPk43pU12jWkO1onoF\n+mTDjx/Ef1cYWA9r5q/LtgTa6Q2sxV4O2x67QW82aAnaxO34dV5zWCPIvAoovZBV\nHRT+BgCg3r2vD1RGKK2nl1aYJtWhO1SZubam+VttdZ/vbM9oOJctxmImsEtBXjkY\nKteePdQtLL5o03JhyXWyRshCq+HMmKf2KgyY8gvydGcP4eLQdBWcW40LcnVq6UjT\n0kJycJEKngMVademq1ZWHGaiYB7hyT6GhgIcHUJ2cKrPgbEh1Q==\n-----END CERTIFICATE-----",
                      "key": "-----BEGIN RSA PRIVATE KEY-----\nMIIEowIBAAKCAQEA4AI33yvHZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I\n2e4SceTKi6O3C/I1XuvWlpng5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4AN\nB5JWWqDOjrRT3o7nRPGXfilabP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYp\nmjdbfxabTz07ig0+6/cwKoRRxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTF\nYvmAH7gcdssSFBt8NPlUATHEsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6W\ndgmu5H2pDml8CDNLDv98Aj7i+I5SRKKcVPlnuQIDAQABAoIBAFPQYKlOzw/+BA0b\nyMIUpdctIMb/54CR/xR0mVw1DbSjigNVPjHUQvB8Y1B2FAITQObgJO06bAv0QdWN\nRb0/v/yYiNJDFjaLjaIAHlO/2+oWrXbFaZqgpVDJhB+e1xaZr2x7XGxm+p925k30\nl6pvIRY+I8JRKvZiV1VZHwL/R3JOtPr++xMZtLVjVOI+f+ySqJ+TZHuAjm49EKxj\ncEmmJ28b7QcziXsvKy00f+zbqLIBKXQdZAFU5eEr1BsDRXdRW+Kf0XIvftuy4BJZ\nvoKT+VGhEvF/qysswL4+6IAO6tpuYnnM0Y2d3sOGoWPkTcQK0MekYKzL/WmtCjNs\n9hodJtECgYEA5EWyhEOf4uOKe5TDp697UCUvXLoOR58FDe/S8XNvScn29jjOkqIg\nOMoqo9xAkJTNTzqn5UUdt1x/pgM2NxlPLFijrc0zQlX3SoOO2ryDd9WNi7YKtN16\nKJqa536WeZu2OEbuAZ+S3GALVy1RPeTNPnUOmKnF06DjDUGzLNCZy10CgYEA+zfw\n952DWuz1U0Z4wvAEqqcgUKXPKrkTXV/iUnjkDkrLYVr0ZofDNTXrdHl+UedFmaOC\ncieZn6DNhcdz5tKtyysGMH3g/qs9PfoGUngvcXsy0Egk04l3x1jc8TTCLqXZXYaQ\nHMsx51n+R58oncPtzYSUOr9qQ6PbC2CstTbFJA0CgYEAjGEsUliAB/jknfEzjXjG\nPdhQUxb8VyE864Az2lah9t/kJzFyIAziAeqZ5GE7t247AGFTBRTHHI8e1Qoemi3P\nWbc9GVIbFs1lIYbcIDpUIyrKPEP8O5QEXtoNLxXTFgAjRGKiVY87spjCAJ+W2ZhO\ne/1it5GYXfgQCYQA2yuBmOUCgYANRkR2YR1axaCk+NlSu6oTdmdPu6M5x7PNQE7O\nOtMaKjua9lppvIzFGAdMDUtueoEEAE7ZR1xnwfB6PDLUpJdIYAqgr1YfPt8qkjaZ\nTv56yZ7CwL0pbF8m6nwqRrZoDp1wwraEvvvxFKFKGY/k3kCHlpTakdjEoDjn3gDi\nRnWeVQKBgCEneMSzucei5LRppRtRaJw/Btll8qlPMlX3W7dxQ3cLwpmLOn0m51Fp\nPIZ44zYK8R6fu4+/sSrlfaIg86Ugeufp6YNxyNROKxUGza5vDIu5OftwWtBeg+UK\nZ8lLWNdX6pp7WMujmF3H1DrkBbauYMUKZ4UxUYtelgHERMePIxwb\n-----END RSA PRIVATE KEY-----"
                    }
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "false"
          containerImage: quay.io/fujitsu/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:eaf601e7aef525ebcd4b769c859ed74c7a8077f843b3a75d76713eabf0295745
          createdAt: "2025-09-11T09:32:52Z"
          description: OpenShift Operator for Fujitsu Enterprise Postgres
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.1.0 <7.1.6'
          operators.openshift.io/valid-subscription: '["Fujitsu Enterprise Postgres
            Advanced Edition with Cryptographic Module Annual Subscription License
            per Core 17 Operator for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/internal-objects: '["fepautoscales.fep.fujitsu.io","fepconfigs.fep.fujitsu.io","fepusers.fep.fujitsu.io","fepcerts.fep.fujitsu.io","fepvolumes.fep.fujitsu.io","fepbackups.fep.fujitsu.io"]'
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: fj-dbaas-tec@dl.jp.fujitsu.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: FEPAction
            name: fepactions.fep.fujitsu.io
            version: v1
          - kind: FEPAutoscale
            name: fepautoscales.fep.fujitsu.io
            version: v1
          - kind: FEPBackup
            name: fepbackups.fep.fujitsu.io
            version: v1
          - kind: FEPCert
            name: fepcerts.fep.fujitsu.io
            version: v1
          - kind: FEPCluster
            name: fepclusters.fep.fujitsu.io
            version: v2
          - kind: FEPConfig
            name: fepconfigs.fep.fujitsu.io
            version: v1
          - kind: FEPExporter
            name: fepexporters.fep.fujitsu.io
            version: v1
          - kind: FEPLogging
            name: feploggings.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2Cert
            name: feppgpool2certs.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2
            name: feppgpool2s.fep.fujitsu.io
            version: v1
          - kind: FEPRestore
            name: feprestores.fep.fujitsu.io
            version: v1
          - kind: FEPUpgrade
            name: fepupgrades.fep.fujitsu.io
            version: v1
          - kind: FEPUser
            name: fepusers.fep.fujitsu.io
            version: v1
          - kind: FEPVolume
            name: fepvolumes.fep.fujitsu.io
            version: v1
        description: "Fujitsu Enterprise Postgres 17 delivers an enterprise-grade
          PostgreSQL on OpenShift Container Platform.\n\nThis solution provides the
          flexibility of a hybrid cloud solution while delivering an enhanced distribution
          of PostgreSQL to support enterprise-level workloads and provide improved
          deployment and management, availability, performance, data governance and
          security.  \n\nAvailable as a multi-architecture container built for both
          amd64.\n\nThe download and Use of the Product is strictly subject to the
          terms of the End User License Agreement with Fujitsu Limited found at https://www.fast.fujitsu.com/fujitsu-enterprise-postgres-license-agreements.
          Where the Product that has been embedded as a whole or part into a third
          party program, only Authorised Customers may download and use the Product.\n"
        displayName: Fujitsu Enterprise Postgres Operator with Cryptographic Module
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - Fujitsu
        - fep
        links:
        - name: Fujitsu Enterprise Postgres
          url: https://www.postgresql.fastware.com/
        maintainers:
        - email: fj-dbaas-tec@dl.jp.fujitsu.com
          name: Fujitsu
        maturity: stable
        provider:
          name: Fujitsu
          url: https://www.postgresql.fastware.com/
        relatedImages:
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-restore-with-cryptographic-module@sha256:edeb7b72a89e05995661e6a17e75b5b477a3ca351e0db146a939a33e88e16c97
        - quay.io/fujitsu/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:eaf601e7aef525ebcd4b769c859ed74c7a8077f843b3a75d76713eabf0295745
        - quay.io/fujitsu/fujitsu-enterprise-postgres-utils-with-cryptographic-module@sha256:7d169df0b8fe47000891afde79be8dd7978da3f893d1e92f01f57fd8c0b3b4c6
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-pgpool2-with-cryptographic-module@sha256:e510bc2d30df33e4c9d3babb898ff9cbd13704fa3fa69a284bf672765fb03083
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-server-with-cryptographic-module@sha256:7778f4b03861b56371d18f7726b68ad21d0faa7c1a25cb38fe7cac483d81c095
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-restore-with-cryptographic-module@sha256:6c0d6e10cfe1868be2986fdde14c6f55f86b12cea01a82199b1e2e0c3154ef75
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-upgrade-with-cryptographic-module@sha256:b8be0344b6cf99b2c5de2b16f04f6f2b05bde861d9b38afd319bbbc4dd72a620
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-upgrade-with-cryptographic-module@sha256:caf6e1d2cc31d01958e4adcb4c3c307d761f350d5b0963ce19214f041c14184a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-restore-with-cryptographic-module@sha256:98a3b0bfa6ec7a9a4a08fb524c331c3de6a9b580d6386a7b5430955ed84b2fc5
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-backup-with-cryptographic-module@sha256:dd15cf3b6b4a5c959efef8b602abe5d64b4f5a424fb6a3e55079a2040e67cd7a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-backup-with-cryptographic-module@sha256:03b7f3c76e3eb775e4ec7f66f7506ee1feeab56dca96e868ac3f38b9c1ca5567
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentd-with-cryptographic-module@sha256:f743f35012e6fd3d54a6b54baeeee98bd7c65fe52923acd87546066fa9b54d7e
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-upgrade-with-cryptographic-module@sha256:1fe690a9c04a55213062ad3bef8c0cff73849f8777c3945b9a8ae2dcff6330fa
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-utils-with-cryptographic-module@sha256:8b2b6a981c863f7075ced1235c614317b6526626b42405295a9c664736f0a80f
        - quay.io/fujitsu/fujitsu-enterprise-postgres-cronjob-with-cryptographic-module@sha256:809a64600396378ff8c825595d40f72aedbc2b57f394420548262b085a3f2ea5
        - quay.io/fujitsu/fujitsu-enterprise-postgres-exporter-with-cryptographic-module@sha256:71e1d1e715e2d89757e80d90b8d5b11baa729110cd5f7898df920a4c85f2bf8a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentbit-with-cryptographic-module@sha256:bb661b3aad0b825d56a9696d6d2d9f5c04add2a83e793fc3d4a73980f4a7a0ac
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-utils-with-cryptographic-module@sha256:4caf1dc5211ac03893ce9a62b417f412ffd411f13569f3ac75e12ca5128af9e4
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-pgpool2-with-cryptographic-module@sha256:a4c5a59ca18f6fa295b78d570f03991f1a4b4bd6df52f8526a1254ac1a567990
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-pgpool2-with-cryptographic-module@sha256:e8aa27cbb57503ec5f4f37beeb2e833ace08d4c583d4361b7ce945c7461a904c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-server-with-cryptographic-module@sha256:cad4e36dc2b1e5ca6bf9308d9f93958d4c8f2bbdc14e000ca0b5700fa335df44
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-backup-with-cryptographic-module@sha256:5b49aa46386ca954ebcbea28fbabd15929cb64fc2a4d2dd2796c9d2e15b212c6
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-server-with-cryptographic-module@sha256:a012f21f63729ba667af097375e08874156e2cf1b94ae203985d45f11031fea9
        version: 7.1.6
      entries:
      - name: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.6
        version: 7.1.6
      - name: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.5
        version: 7.1.5
      name: stable
    defaultChannel: stable
    packageName: fujitsu-enterprise-postgres-operator-with-cg-module
    provider:
      name: Fujitsu
      url: https://www.postgresql.fastware.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-05-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx8-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx8-operator-rhmp/support?utm_source=openshift_console
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:154f85cc6721a70909dee755ddd88897ad51038afe2538b6dee4d26e7d8b9afb
        version: 0.0.10
      entries:
      - name: joget-dx8-openshift-operator.v0.0.10
        version: 0.0.10
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-operator-rhmp
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://www.redhat.com/
    name: rhdh
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhdh-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
          createdAt: "2025-09-12T06:25:11Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.7.1'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.7/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:62f32e50727c5006766a34daa5b6472b649cd9894f3fe9543b8ecc67e6760e8e
        version: 1.7.1
      entries:
      - name: rhdh-operator.v1.7.1
        version: 1.7.1
      - name: rhdh-operator.v1.7.0
        version: 1.7.0
      - name: rhdh-operator.v1.6.5
        version: 1.6.5
      - name: rhdh-operator.v1.6.4
        version: 1.6.4
      - name: rhdh-operator.v1.6.3
        version: 1.6.3
      - name: rhdh-operator.v1.6.2
        version: 1.6.2
      - name: rhdh-operator.v1.6.1
        version: 1.6.1
      - name: rhdh-operator.v1.6.0
        version: 1.6.0
      - name: rhdh-operator.v1.5.3
        version: 1.5.3
      - name: rhdh-operator.v1.5.2
        version: 1.5.2
      - name: rhdh-operator.v1.5.1
        version: 1.5.1
      - name: rhdh-operator.v1.4.3
        version: 1.4.3
      - name: rhdh-operator.v1.4.2
        version: 1.4.2
      - name: rhdh-operator.v1.4.1
        version: 1.4.1
      - name: rhdh-operator.v1.4.0
        version: 1.4.0
      - name: rhdh-operator.v1.3.5
        version: 1.3.5
      - name: rhdh-operator.v1.3.4
        version: 1.3.4
      - name: rhdh-operator.v1.3.3
        version: 1.3.3
      - name: rhdh-operator.v1.3.1
        version: 1.3.1
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast
    - currentCSV: rhdh-operator.v1.1.2-0.1714688890.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:9bea7eabdea44342248dfd6091a8f5c3b6e65884c916b3c2a073a4a64481aa6f
          createdAt: "2023-03-20T16:11:34Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins, configuration settings, and deployment
            mechanisms, which can help streamline the process of setting up a self-managed
            internal developer portal for adopters who are just starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: rhdh-operator.v1.1.2
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.1.0 <1.2.0'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/janus-idp/operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: asoro@redhat.com
          name: Armel Soro
        - email: gazarenk@redhat.com
          name: Gennady Azarenkov
        - email: jianrzha@redhat.com
          name: Jianrong Zhang
        - email: nboldt@redhat.com
          name: Nick Boldt
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:95ed239eb427c7b1e422ece28c5277b5018efb751c2b472bf502b8216219c047
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:9bea7eabdea44342248dfd6091a8f5c3b6e65884c916b3c2a073a4a64481aa6f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:d7ae2456857cab09a9e9acf69c9990ef18deb7f920046b8c2483fbfd0d42bbfc
        - registry.redhat.io/rhel9/postgresql-15@sha256:d5f3d496ecd55688e384cc0627700a97af6924582123bcf00749c375724af3de
        version: 1.1.2+0.1714688890.p
      entries:
      - name: rhdh-operator.v1.1.2
        version: 1.1.2
      - name: rhdh-operator.v1.1.2-0.1714608881.p
        version: 1.1.2+0.1714608881.p
      - name: rhdh-operator.v1.1.2-0.1714688618.p
        version: 1.1.2+0.1714688618.p
      - name: rhdh-operator.v1.1.2-0.1714688890.p
        version: 1.1.2+0.1714688890.p
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.1
    - currentCSV: rhdh-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:c9051a6c339cd40cbc7474a33b32a6b94913af964e12ca5da0c0b7e1c6f444d5
          createdAt: "2024-11-28T17:55:19Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.2.6'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.2/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.2/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: asoro@redhat.com
          name: Armel Soro
        - email: cdaley@redhat.com
          name: Corey Daley
        - email: gazarenk@redhat.com
          name: Gennady Azarenkov
        - email: nboldt@redhat.com
          name: Nick Boldt
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:ef29f2b340a6dc52c5014239876d54a54f7c0f0bf4ea917391438a69c94a8532
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:c9051a6c339cd40cbc7474a33b32a6b94913af964e12ca5da0c0b7e1c6f444d5
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:025eb58c69b9cf5992ecf2fccb29a3d36061cbcf62652a1c6878f0c4d767c1fb
        - registry.redhat.io/rhel9/postgresql-15@sha256:24fb4e7914a6e1464d015be9e5582cc4b9da224137408bd429e7ea4f391aa198
        version: 1.2.6
      entries:
      - name: rhdh-operator.v1.2.6
        version: 1.2.6
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.2
    - currentCSV: rhdh-operator.v1.3.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:5ce579edc10bde5c1c400104f7e0d9f86ca277cd565aaf3dd84b8636f5b949a8
          createdAt: "2025-02-10T18:12:52Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.3.5'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.3/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:5ce579edc10bde5c1c400104f7e0d9f86ca277cd565aaf3dd84b8636f5b949a8
        - registry.redhat.io/rhel9/postgresql-15@sha256:24fb4e7914a6e1464d015be9e5582cc4b9da224137408bd429e7ea4f391aa198
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:c044fed8485a62a5051ac726525da546936be22f268db05491bf81582e0e6194
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b4934c45ba595b36e65c73b734c58d1517e7f43b7cfc309ca8bc853733582f76
        version: 1.3.5
      entries:
      - name: rhdh-operator.v1.3.5
        version: 1.3.5
      - name: rhdh-operator.v1.3.4
        version: 1.3.4
      - name: rhdh-operator.v1.3.3
        version: 1.3.3
      - name: rhdh-operator.v1.3.1
        version: 1.3.1
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.3
    - currentCSV: rhdh-operator.v1.4.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:ae5b52c1b35e82f3c564402e367a2a2028e16806867bba0868643089ba7ed223
          createdAt: "2025-04-03T05:42:32Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.4.3'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.4/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:ae5b52c1b35e82f3c564402e367a2a2028e16806867bba0868643089ba7ed223
        version: 1.4.3
      entries:
      - name: rhdh-operator.v1.4.3
        version: 1.4.3
      - name: rhdh-operator.v1.4.2
        version: 1.4.2
      - name: rhdh-operator.v1.4.1
        version: 1.4.1
      - name: rhdh-operator.v1.4.0
        version: 1.4.0
      name: fast-1.4
    - currentCSV: rhdh-operator.v1.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:8fe167cc649a1c53994c376f3cf3eed8c6e318741a0c08e5aaf07cb0b74e7f8a
          createdAt: "2025-06-05T14:31:21Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.5.3'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.5/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:8fe167cc649a1c53994c376f3cf3eed8c6e318741a0c08e5aaf07cb0b74e7f8a
        - registry.redhat.io/rhel9/postgresql-15@sha256:74377f5a61dd7e3e0ba5d2fa880b7107025433a026396567c79eac927b4212bc
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:2ad8c87e82df27c2b6ceba94179d1f0a40fe775368663c1b5572611888e0e02b
        version: 1.5.3
      entries:
      - name: rhdh-operator.v1.5.3
        version: 1.5.3
      - name: rhdh-operator.v1.5.2
        version: 1.5.2
      - name: rhdh-operator.v1.5.1
        version: 1.5.1
      name: fast-1.5
    - currentCSV: rhdh-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:50f20300c81621221260d03d22b1442750e3f1d2e62f3e511d63925d84c4d1f9
          createdAt: "2025-08-27T18:43:12Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.6.5'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.6/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:50f20300c81621221260d03d22b1442750e3f1d2e62f3e511d63925d84c4d1f9
        - registry.redhat.io/rhel9/postgresql-15@sha256:5e225c1646251d2509b3a6e45fad23abfb11dd5c45d3224b59178f90deb38a70
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:177a7308f61a6b1ba21bc0f7bf550a93d4478e52479e2c94a2437f2152e3d8ff
        version: 1.6.5
      entries:
      - name: rhdh-operator.v1.6.5
        version: 1.6.5
      - name: rhdh-operator.v1.6.4
        version: 1.6.4
      - name: rhdh-operator.v1.6.3
        version: 1.6.3
      - name: rhdh-operator.v1.6.2
        version: 1.6.2
      - name: rhdh-operator.v1.6.1
        version: 1.6.1
      - name: rhdh-operator.v1.6.0
        version: 1.6.0
      name: fast-1.6
    - currentCSV: rhdh-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
          createdAt: "2025-09-12T06:25:11Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.7.1'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.7/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:62f32e50727c5006766a34daa5b6472b649cd9894f3fe9543b8ecc67e6760e8e
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
        version: 1.7.1
      entries:
      - name: rhdh-operator.v1.7.1
        version: 1.7.1
      - name: rhdh-operator.v1.7.0
        version: 1.7.0
      name: fast-1.7
    defaultChannel: fast
    packageName: rhdh
    provider:
      name: Red Hat Inc.
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecrpublic-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecrpublic-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecrpublic-controller:1.1.1
          createdAt: "2025-09-20T01:21:52Z"
          description: AWS ECR PUBLIC controller is a service controller for managing
            ECR PUBLIC resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Repository represents the state of an AWS ecrpublic Repository
              resource.
            displayName: Repository
            kind: Repository
            name: repositories.ecrpublic.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ECR PUBLIC resources in AWS from within your Kubernetes cluster.

          **About Amazon ECR PUBLIC**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECR PUBLIC
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecrpublic
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECR PUBLIC Developer Resources
          url: https://aws.amazon.com/ECR PUBLIC/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecrpublic maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecrpublic-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ecrpublic-controller.v1.1.1
        version: 1.1.1
      - name: ack-ecrpublic-controller.v1.1.0
        version: 1.1.0
      - name: ack-ecrpublic-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecrpublic-controller.v1.0.12
        version: 1.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecrpublic-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: egressip-ipam-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: egressip-ipam-operator.v1.2.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "EgressIPAM",
                "metadata": {
                  "name": "egressipam-aws"
                },
                "spec": {
                  "cidrAssignments": [
                    {
                      "CIDR": "10.0.128.0/19",
                      "labelValue": "us-west-2a"
                    },
                    {
                      "CIDR": "10.0.160.0/19",
                      "labelValue": "us-west-2b",
                      "reservedIPs": [
                        "10.0.160.126"
                      ]
                    },
                    {
                      "CIDR": "10.0.192.0/19",
                      "labelValue": "us-west-2c"
                    }
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "topologyLabel": "failure-domain.beta.kubernetes.io/zone"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "EgressIPAM",
                "metadata": {
                  "name": "egressipam-baremetal"
                },
                "spec": {
                  "cidrAssignments": [
                    {
                      "CIDR": "192.169.0.0/24",
                      "labelValue": "true"
                    }
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "topologyLabel": "egressGateway"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Networking, Cloud Provider, Security
          certified: "false"
          containerImage: quay.io/redhat-cop/egressip-ipam-operator@sha256:d6129bd287b87696a23e0da64dcd598958f779dc88ae33b46b3c909dbab80adf
          createdAt: "2022-07-25T15:07:22Z"
          description: This operator provides automation around managing egressIPs.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: egressip-ipam-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/egressip-ipam-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgressIPAM is the Schema for the egressipams API
            displayName: Egress IPAM
            kind: EgressIPAM
            name: egressipams.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator automates the assignment of egressIPs to namespaces.
          Namespaces can opt in to receiving one or more egressIPs with the following annotation `egressip-ipam-operator.redhat-cop.io/egressipam:<egressIPAM>`, where `egressIPAM` is the CRD that controls how egressIPs are assigned.
          IPs assigned to the namespace can be looked up in the following annotation: `egressip-ipam-operator.redhat-cop.io/egressips`.

          EgressIP assignments is managed by the EgressIPAM CRD. Here is an example of it:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: EgressIPAM
          metadata:
            name: example-egressipam
          spec:
            cidrAssignments:
              - labelValue: "true"
                CIDR: 192.169.0.0/24
                reservedIPs:
                - "192.159.0.5"
            topologyLabel: egressGateway
            nodeSelector:
              node-role.kubernetes.io/worker: ""
          ```

          This EgressIPAM specifies that all nodes that comply with the specified node selector and that also have labels `egressGateway: "true"` will be assigned egressIP from the specified CIDR.

          Note that the `cidrAssigments` field is an array and therefore, multiple groups of nodes can be identified with the labelValue and different CIDRs can be assigned to them. This is usually not necessary on a bare metal deployment.

          In the bare metal scenario, when this egressCRD is created, all the `hostsubnet` relative to the nodes selected by this EgressIPAM will be update to have the EgressCIDRs field equal to the specified CIDR (see below for cloud provider scenarios).

          When a namespace is created with the opt-in annotation: `egressip-ipam-operator.redhat-cop.io/egressipam=<egressIPAM>`, an available egressIP is selected from the CIDR and assigned to the namespace.The `netnamespace` associated with this namespace is updated to use that egressIP.

          It is possible to specify a set of reserved IPs. These IPs must belong to the CIDR and will never be assigned.

          ## Passing EgressIPs as input

          The normal mode of operation of this operator is to pick a random IP from the configured CIDR. However, it also supports a scenario where egressIPs are picked by an external process and passed as input.

          In this case IPs must me passed as an annotation to the namespace, like this: `egressip-ipam-operator.redhat-cop.io/egressips=IP1,IP2...`. The value of the annotation is a comma separated array of ip with no spaces.

          There must be exactly one IP per CIDR defined in the referenced egressIPAM. Moreover, each IP must belong to the corresponding CIDR. Because this situation can lead to inconsistencies, failing to have correspondence between IPs in the namespace annotation and CIDRs in the egressIPAM CR will cause the operator to error out and stop processing all the namespaces associated with the given EgressIPAM CR.

          It also also a responsibility of the progress picking the IPs to ensure that those IPs are actually available.

          ## Assumptions

          The following assumptions apply when using this operator:

          1. If multiple EgressIPAMs are defined, their selected nodes MUST NOT overlap.
          2. When using a cloud provider the topology label MUST be `failure-domain.beta.kubernetes.io/zone`.

          ## Support for AWS

          In AWS as well as other cloud providers, one cannot freely assign IPs to machines. Additional steps need to be performed in this case. Considering this EgressIPAM

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: EgressIPAM
          metadata:
            name: egressipam-aws
          spec:
            cidrAssignments:
              - labelValue: "eu-central-1a"
                CIDR: 10.0.128.0/20
              - labelValue: "eu-central-1b"
                CIDR: 10.0.144.0/20
              - labelValue: "eu-central-1c"
                CIDR: 10.0.160.0/20
            topologyLabel: failure-domain.beta.kubernetes.io/zone
            nodeSelector:
              matchLabels:
                node-role.kubernetes.io/worker: ""
          ```

          When a namespace with the opt-in annotation is created, the following happens:

          1. for each of the CIDRs, an available IP is selected and assigned to the namespace
          2. the relative `netnamespace` is update to reflect the assignment (multiple IPs will be assigned in this case).
          3. one node per zone is selected to carry the egressIP
          4. the relative aws machines are assigned the additional IP on the main interface (support for secondary interfaces in not available)
          5. the relative `hostsubnets` are updated to reflect the assigned IP, the `egressIP` field is updated.
        displayName: Egressip Ipam Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - egressIPs
        - high availability
        - firewall support
        - traffic management
        - ipam
        links:
        - name: Egressip Ipam Operator
          url: https://github.com/redhat-cop/egressip-ipam-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/egressip-ipam-operator:v0.1.2
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/egressip-ipam-operator@sha256:d6129bd287b87696a23e0da64dcd598958f779dc88ae33b46b3c909dbab80adf
        version: 1.2.4
      entries:
      - name: egressip-ipam-operator.v1.2.4
        version: 1.2.4
      - name: egressip-ipam-operator.v1.2.3
        version: 1.2.3
      - name: egressip-ipam-operator.v1.2.2
        version: 1.2.2
      - name: egressip-ipam-operator.v1.2.1
        version: 1.2.1
      - name: egressip-ipam-operator.v1.2.0
        version: 1.2.0
      - name: egressip-ipam-operator.v1.1.0
        version: 1.1.0
      - name: egressip-ipam-operator.v1.0.8
        version: 1.0.8
      - name: egressip-ipam-operator.v1.0.7
        version: 1.0.7
      - name: egressip-ipam-operator.v1.0.6
        version: 1.0.6
      - name: egressip-ipam-operator.v1.0.5
        version: 1.0.5
      - name: egressip-ipam-operator.v1.0.4
        version: 1.0.4
      - name: egressip-ipam-operator.v1.0.3
        version: 1.0.3
      - name: egressip-ipam-operator.v1.0.2
        version: 1.0.2
      - name: egressip-ipam-operator.v1.0.1
        version: 1.0.1
      - name: egressip-ipam-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: egressip-ipam-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: hive-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hive-operator.v1.2.4953-c248a51
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:c248a51ae1
          createdAt: "2025-09-22T13:41:19Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeploymentCustomization is the Schema for clusterdeploymentcustomizations
              API.
            displayName: ClusterDeploymentCustomization
            kind: ClusterDeploymentCustomization
            name: clusterdeploymentcustomizations.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: |-
              ClusterPool represents a pool of clusters that should be kept ready to be given out to users. Clusters are removed
              from the pool once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: |-
              MachinePoolNameLease is the Schema for the MachinePoolNameLeases API. This resource is mostly empty
              as we're primarily relying on the name to determine if a lease is available.
              Note that not all cloud providers require the use of a lease for naming, at present this
              is only required for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * IBM Cloud
          * Red Hat OpenStack
          * vSphere

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Troubleshooting](https://github.com/openshift/hive/blob/master/docs/troubleshooting.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
              logLevel: debug
              targetNamespace: hive
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:c248a51ae1
        version: 1.2.4953-c248a51
      entries:
      - name: hive-operator.v1.2.4953-c248a51
        version: 1.2.4953-c248a51
      - name: hive-operator.v1.2.4928-76a65b6
        version: 1.2.4928-76a65b6
      - name: hive-operator.v1.2.4902-4929f16
        version: 1.2.4902-4929f16
      - name: hive-operator.v1.2.4896-a29a238
        version: 1.2.4896-a29a238
      - name: hive-operator.v1.2.4868-45ae912
        version: 1.2.4868-45ae912
      - name: hive-operator.v1.2.4863-e9ad0fa
        version: 1.2.4863-e9ad0fa
      - name: hive-operator.v1.2.4774-87bff59
        version: 1.2.4774-87bff59
      - name: hive-operator.v1.2.4766-6a82857
        version: 1.2.4766-6a82857
      - name: hive-operator.v1.2.4746-d25e0db
        version: 1.2.4746-d25e0db
      - name: hive-operator.v1.2.4738-b19773e
        version: 1.2.4738-b19773e
      - name: hive-operator.v1.2.4731-41f153f
        version: 1.2.4731-41f153f
      - name: hive-operator.v1.2.4703-02f3f00
        version: 1.2.4703-02f3f00
      - name: hive-operator.v1.2.4693-ec3bc72
        version: 1.2.4693-ec3bc72
      - name: hive-operator.v1.2.4686-553c6c2
        version: 1.2.4686-553c6c2
      - name: hive-operator.v1.2.4682-406d6ca
        version: 1.2.4682-406d6ca
      - name: hive-operator.v1.2.4672-af54e2f
        version: 1.2.4672-af54e2f
      - name: hive-operator.v1.2.4668-3ccd44d
        version: 1.2.4668-3ccd44d
      - name: hive-operator.v1.2.4664-00f84cc
        version: 1.2.4664-00f84cc
      - name: hive-operator.v1.2.4658-c13e207
        version: 1.2.4658-c13e207
      - name: hive-operator.v1.2.4654-3268a18
        version: 1.2.4654-3268a18
      - name: hive-operator.v1.2.4650-531d549
        version: 1.2.4650-531d549
      - name: hive-operator.v1.2.4644-365f074
        version: 1.2.4644-365f074
      - name: hive-operator.v1.2.4638-56d6d64
        version: 1.2.4638-56d6d64
      - name: hive-operator.v1.2.4625-389d717
        version: 1.2.4625-389d717
      - name: hive-operator.v1.2.4615-52100dd
        version: 1.2.4615-52100dd
      - name: hive-operator.v1.2.4605-b41ca3f
        version: 1.2.4605-b41ca3f
      - name: hive-operator.v1.2.4584-95efcb7
        version: 1.2.4584-95efcb7
      - name: hive-operator.v1.2.4574-1ec27ad
        version: 1.2.4574-1ec27ad
      - name: hive-operator.v1.2.4568-b3eca52
        version: 1.2.4568-b3eca52
      - name: hive-operator.v1.2.4564-bdf9d08
        version: 1.2.4564-bdf9d08
      - name: hive-operator.v1.2.4552-bf2d689
        version: 1.2.4552-bf2d689
      - name: hive-operator.v1.2.4532-cddd4f9
        version: 1.2.4532-cddd4f9
      - name: hive-operator.v1.2.4511-1fce770
        version: 1.2.4511-1fce770
      - name: hive-operator.v1.2.4500-b1ac27b
        version: 1.2.4500-b1ac27b
      - name: hive-operator.v1.2.4485-47df0d9
        version: 1.2.4485-47df0d9
      - name: hive-operator.v1.2.4469-7be6433
        version: 1.2.4469-7be6433
      - name: hive-operator.v1.2.4463-d7ead60
        version: 1.2.4463-d7ead60
      - name: hive-operator.v1.2.4461-97686bb
        version: 1.2.4461-97686bb
      - name: hive-operator.v1.2.4459-f910c9b
        version: 1.2.4459-f910c9b
      - name: hive-operator.v1.2.4451-83aedb9
        version: 1.2.4451-83aedb9
      - name: hive-operator.v1.2.4446-e11a23b
        version: 1.2.4446-e11a23b
      - name: hive-operator.v1.2.4441-fdbfaf0
        version: 1.2.4441-fdbfaf0
      - name: hive-operator.v1.2.4436-44323bc
        version: 1.2.4436-44323bc
      - name: hive-operator.v1.2.4428-ceaafa8
        version: 1.2.4428-ceaafa8
      - name: hive-operator.v1.2.4426-fb27114
        version: 1.2.4426-fb27114
      - name: hive-operator.v1.2.4420-fe0e302
        version: 1.2.4420-fe0e302
      - name: hive-operator.v1.2.4418-eb5403f
        version: 1.2.4418-eb5403f
      - name: hive-operator.v1.2.4416-fba8d52
        version: 1.2.4416-fba8d52
      - name: hive-operator.v1.2.4408-9c5806a
        version: 1.2.4408-9c5806a
      - name: hive-operator.v1.2.4404-410fed9
        version: 1.2.4404-410fed9
      - name: hive-operator.v1.2.4400-8c54fc9
        version: 1.2.4400-8c54fc9
      - name: hive-operator.v1.2.4386-6f1fd3e
        version: 1.2.4386-6f1fd3e
      - name: hive-operator.v1.2.4382-0e95d5b
        version: 1.2.4382-0e95d5b
      - name: hive-operator.v1.2.4366-df65e7e
        version: 1.2.4366-df65e7e
      - name: hive-operator.v1.2.4360-de169dd
        version: 1.2.4360-de169dd
      - name: hive-operator.v1.2.4355-e5f809f
        version: 1.2.4355-e5f809f
      name: alpha
    - currentCSV: hive-operator.v2.5.3516-a2ed9b3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.5.3516-a2ed9b3
          createdAt: "2022-06-03T13:24:12Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:v2.5.3516-a2ed9b3
        version: 2.5.3516-a2ed9b3
      entries:
      - name: hive-operator.v2.5.3516-a2ed9b3
        version: 2.5.3516-a2ed9b3
      - name: hive-operator.v2.5.3514-4052426
        version: 2.5.3514-4052426
      - name: hive-operator.v2.5.3508-6cb94c6
        version: 2.5.3508-6cb94c6
      - name: hive-operator.v2.5.3506-f42617c
        version: 2.5.3506-f42617c
      - name: hive-operator.v2.5.3495-5cfdd44
        version: 2.5.3495-5cfdd44
      - name: hive-operator.v2.5.3491-6feded1
        version: 2.5.3491-6feded1
      - name: hive-operator.v2.5.3489-6dec9c3
        version: 2.5.3489-6dec9c3
      name: mce-2.0
    - currentCSV: hive-operator.v2.3.3039-60b8a9a
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.3.3039-60b8a9a
          createdAt: "2022-08-15T17:23:39Z"
          description: OpenShift cluster provisioning and management at scale.
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:v2.3.3039-60b8a9a
        version: 2.3.3039-60b8a9a
      entries:
      - name: hive-operator.v2.3.3039-60b8a9a
        version: 2.3.3039-60b8a9a
      name: ocm-2.3
    - currentCSV: hive-operator.v2.4.3231-ba51985
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.4.3231-ba51985
          createdAt: "2022-08-15T11:53:28Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:v2.4.3231-ba51985
        version: 2.4.3231-ba51985
      entries:
      - name: hive-operator.v2.4.3231-ba51985
        version: 2.4.3231-ba51985
      - name: hive-operator.v2.4.3227-74831db
        version: 2.4.3227-74831db
      - name: hive-operator.v2.4.3222-d68346c
        version: 2.4.3222-d68346c
      - name: hive-operator.v2.4.3206-b0714a7
        version: 2.4.3206-b0714a7
      - name: hive-operator.v2.4.3203-672cfe7
        version: 2.4.3203-672cfe7
      - name: hive-operator.v2.4.3201-8c50440
        version: 2.4.3201-8c50440
      - name: hive-operator.v2.4.3199-cce4cb9
        version: 2.4.3199-cce4cb9
      - name: hive-operator.v2.4.3197-6c9c8e4
        version: 2.4.3197-6c9c8e4
      - name: hive-operator.v2.4.3194-f82b7d8
        version: 2.4.3194-f82b7d8
      - name: hive-operator.v2.4.3192-fae0111
        version: 2.4.3192-fae0111
      - name: hive-operator.v2.4.3189-026033b
        version: 2.4.3189-026033b
      - name: hive-operator.v2.4.3180-7299056
        version: 2.4.3180-7299056
      - name: hive-operator.v1.2.3179-0756b70e6
        version: 1.2.3179-0756b70e6
      - name: hive-operator.v1.1.16
        version: 1.1.16
      - name: hive-operator.v1.1.15
        version: 1.1.15
      - name: hive-operator.v1.1.14
        version: 1.1.14
      - name: hive-operator.v1.1.13
        version: 1.1.13
      - name: hive-operator.v1.1.12
        version: 1.1.12
      - name: hive-operator.v1.1.11
        version: 1.1.11
      - name: hive-operator.v1.1.10
        version: 1.1.10
      - name: hive-operator.v1.1.9
        version: 1.1.9
      - name: hive-operator.v1.1.8
        version: 1.1.8
      - name: hive-operator.v1.1.7
        version: 1.1.7
      - name: hive-operator.v1.1.6
        version: 1.1.6
      name: ocm-2.4
    defaultChannel: alpha
    packageName: hive-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-admin-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-admin-operator.v2024.10.2-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": null
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:ed279ff5416cce855118f66a73d55385aba26af19735eed052010ba653f64b75
          createdAt: "2024-10-25T01:47:18Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2024.10.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2023.9.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:7c4868db021beaffa762eb0de43739bbc865a4bf5900f1d17f78caed8612c073
        - registry.connect.redhat.com/ca/uma-operator@sha256:ed279ff5416cce855118f66a73d55385aba26af19735eed052010ba653f64b75
        version: 2024.10.2-12
      entries:
      - name: uma-admin-operator.v2024.10.2-12
        version: 2024.10.2-12
      name: "2024.10"
    - currentCSV: uma-admin-operator.v2025.1.3-33
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.1.1.33,990033",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": null
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:6a932a065add9f36f9615a93037315603f2d9cd8a2e3c64b1b1fc390bb4f1dcf
          createdAt: "2025-02-18T05:34:34Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.1.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2023.9.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.1.1.26,990026"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:6a932a065add9f36f9615a93037315603f2d9cd8a2e3c64b1b1fc390bb4f1dcf
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:ae91a28d57c8a359f81757e4bf79472f2ac189ac0ae3ee0d87e8ff9c433f1ccc
        version: 2025.1.3-33
      entries:
      - name: uma-admin-operator.v2025.1.3-33
        version: 2025.1.3-33
      name: "2025.1"
    - currentCSV: uma-admin-operator.v2025.2.1-25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.2.1.25,990025",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverCollectorCert": "",
                    "serverCollectorKey": "",
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:d201790c2d9ded6bc7a1c5f68a267574e7b5d39c39ce2f98e20ef630b0f7365f
          createdAt: "2025-03-05T05:51:02Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.2.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2023.9.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.1.1.26,990026"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:d201790c2d9ded6bc7a1c5f68a267574e7b5d39c39ce2f98e20ef630b0f7365f
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:8bcf0dac889f38a0bf77eb9babe3f726e3b2564d5988c05bdc5c2d91ae189c61
        version: 2025.2.1-25
      entries:
      - name: uma-admin-operator.v2025.2.1-25
        version: 2025.2.1-25
      name: "2025.2"
    - currentCSV: uma-admin-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:4c2a7f68a47428f3e386107254f74fa8c34e477eb40fc17c173df92f3c66c559
          createdAt: "2025-04-16T06:55:39Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.4.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.4.2.5,990005"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        - registry.connect.redhat.com/ca/uma-operator@sha256:4c2a7f68a47428f3e386107254f74fa8c34e477eb40fc17c173df92f3c66c559
        version: 2025.4.5-5
      entries:
      - name: uma-admin-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-admin-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:514b953eaf0959bc120f80bdc0a2e19d8f059d3fe2883a4f6843a398470a56b1
          createdAt: "2025-07-24T14:54:21Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.5.2.12,990012"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:514b953eaf0959bc120f80bdc0a2e19d8f059d3fe2883a4f6843a398470a56b1
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        version: 2025.5.4-12
      entries:
      - name: uma-admin-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-admin-operator.v2025.7.3-21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:188a2b8d2255ded6141a4b795f1b6d7bd98017b4fd57c9c1a1d38fc701358057
          createdAt: "2025-08-23T09:00:37Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.3
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.7.1.23,990023"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:188a2b8d2255ded6141a4b795f1b6d7bd98017b4fd57c9c1a1d38fc701358057
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.3-21
      entries:
      - name: uma-admin-operator.v2025.7.3-21
        version: 2025.7.3-21
      name: "2025.7"
    - currentCSV: uma-admin-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:0cfb63c62258e54d67e7baca874003a04477f30bcf7f56861f565ad068070580
          createdAt: "2025-08-15T03:20:28Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-admin-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.8.1.9,990009"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:0cfb63c62258e54d67e7baca874003a04477f30bcf7f56861f565ad068070580
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        version: 2025.8.2-9
      entries:
      - name: uma-admin-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-admin-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
          createdAt: "2025-09-24T12:23:56Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.12,990012"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-admin-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-admin-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:7b93de818fa5160e72f878b0350d13fc1d5943eac95b0794d3d4f930cf21b59a
          createdAt: "2025-07-30T08:07:39Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-admin-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "99.99.0.serp_F152813,19"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:7b93de818fa5160e72f878b0350d13fc1d5943eac95b0794d3d4f930cf21b59a
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        version: 2025.7.2-19
      entries:
      - name: uma-admin-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-admin-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1cf3909dd2caeabc53209d6f8565dcb6c3562b9c6eb715dda8ba1953f264668e
          createdAt: "2025-09-11T11:16:06Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.8,990008"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:1cf3909dd2caeabc53209d6f8565dcb6c3562b9c6eb715dda8ba1953f264668e
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-admin-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-admin-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
          createdAt: "2025-09-24T12:23:56Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.12,990012"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-admin-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-admin-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: octavia-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: octavia-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "octavia.openstack.org/v1beta1",
                "kind": "Octavia",
                "metadata": {
                  "name": "octavia"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "octavia",
                  "databaseInstance": "openstack",
                  "lbMgmtNetwork": {
                    "availabilityZones": [
                      "zone-1"
                    ]
                  },
                  "octaviaAPI": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "octavia",
                    "databaseInstance": "openstack",
                    "persistenceDatabaseAccount": "octavia-persistence",
                    "preserveJobs": false,
                    "secret": "osp-secret",
                    "serviceAccount": "octavia",
                    "serviceUser": "octavia"
                  },
                  "octaviaHealthManager": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "octavia",
                    "databaseInstance": "openstack",
                    "networkAttachments": [
                      "octavia"
                    ],
                    "persistenceDatabaseAccount": "octavia-persistence",
                    "preserveJobs": false,
                    "role": "healthmanager",
                    "secret": "osp-secret",
                    "serviceAccount": "octavia",
                    "serviceUser": "octavia"
                  },
                  "octaviaHousekeeping": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "octavia",
                    "databaseInstance": "openstack",
                    "networkAttachments": [
                      "octavia"
                    ],
                    "persistenceDatabaseAccount": "octavia-persistence",
                    "preserveJobs": false,
                    "role": "housekeeping",
                    "secret": "osp-secret",
                    "serviceAccount": "octavia",
                    "serviceUser": "octavia"
                  },
                  "octaviaWorker": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "octavia",
                    "databaseInstance": "openstack",
                    "networkAttachments": [
                      "octavia"
                    ],
                    "persistenceDatabaseAccount": "octavia-persistence",
                    "preserveJobs": false,
                    "role": "worker",
                    "secret": "osp-secret",
                    "serviceAccount": "octavia",
                    "serviceUser": "octavia"
                  },
                  "persistenceDatabaseAccount": "octavia-persistence",
                  "preserveJobs": false,
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "osp-secret",
                  "serviceUser": "octavia"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:01Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: Octavia Amphora Controller
            kind: OctaviaAmphoraController
            name: octaviaamphoracontrollers.octavia.openstack.org
            version: v1beta1
          - description: OctaviaAPI is the Schema for the octaviaapis API
            displayName: Octavia API
            kind: OctaviaAPI
            name: octaviaapis.octavia.openstack.org
            version: v1beta1
          - kind: OctaviaRsyslog
            name: octaviarsyslogs.octavia.openstack.org
            version: v1beta1
          - description: Octavia is the Schema for the octavia API
            displayName: Octavia
            kind: Octavia
            name: octavias.octavia.openstack.org
            version: v1beta1
        description: Octavia Operator
        displayName: Octavia Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Load-Balancer
        - Octavia
        links:
        - name: Octavia Operator
          url: https://github.com/openstack-k8s-operators/octavia-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/octavia-rhel9-operator@sha256:ddd280ec099cde63e38bd964d7f0608c1c34e5c9f6b140f86dde2126b0357ef5
        - registry.redhat.io/rhoso/openstack-octavia-health-manager-rhel9@sha256:7d305af6e9e6c636510186dd80037849d19bb1e6d2b2507c025d1d0d9254874a
        - registry.redhat.io/rhoso/openstack-octavia-worker-rhel9@sha256:0919ddff6421430f8ad24cb21a89c65f02f98df424267f7e436c97d555112e2a
        - registry.redhat.io/ubi9/httpd-24@sha256:dbfbe3383e12d63df7178e4cdd5b1e6e51a84c781041b9192026ea6edb72fbf3
        - registry.redhat.io/rhoso/openstack-octavia-api-rhel9@sha256:ae9aae94c2523933a346bdade16158b4302b3a0bee93dec89bae5d15606c8dfb
        - registry.redhat.io/rhoso/openstack-octavia-housekeeping-rhel9@sha256:d54868077feb205199b76e83fc9220d9b4215dbeba037cf13fb68dc6577a07dc
        version: 1.0.6
      entries:
      - name: octavia-operator.v1.0.6
        version: 1.0.6
      - name: octavia-operator.v1.0.5
        version: 1.0.5
      - name: octavia-operator.v1.0.4
        version: 1.0.4
      - name: octavia-operator.v1.0.3
        version: 1.0.3
      - name: octavia-operator.v1.0.2
        version: 1.0.2
      - name: octavia-operator.v1.0.1
        version: 1.0.1
      - name: octavia-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: octavia-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: gatekeeper-operator-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gatekeeper-operator-product.v3.11.2-0.1725401426.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel8-operator@sha256:c8894e4d02fb2c822ab2e02e489dd9b67c953d96bd3b96a8305219a2245dfd99
          createdAt: 21 Mar 2024, 01:59
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.11.0
          olm.substitutesFor: gatekeeper-operator-product.v3.11.2-0.1721233953.p
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/gatekeeper/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the Schema for the gatekeepers API
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |-
          Gatekeeper allows administrators to detect and reject non-compliant commits to an infrastructure-as-code system's source-of-truth. This strengthens compliance efforts and prevents a bad state from slowing down the organization.

          ## Support & Troubleshooting
          Gatekeeper is available and supported as part of a valid Red Hat software subscription when it is used with Red Hat Advanced Cluster Management for Kubernetes. Troubleshooting information is available in the Gatekeeper documentation.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/gatekeeper/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - registry.redhat.io/gatekeeper/gatekeeper-rhel8-operator@sha256:c8894e4d02fb2c822ab2e02e489dd9b67c953d96bd3b96a8305219a2245dfd99
        - registry.redhat.io/gatekeeper/gatekeeper-rhel8@sha256:701b165def9e4b92d91b8eac4917721e76e634d2159b8efc69bef23e10b51b9d
        version: 3.11.2+0.1725401426.p
      entries:
      - name: gatekeeper-operator-product.v3.11.2-0.1725401426.p
        version: 3.11.2+0.1725401426.p
      - name: gatekeeper-operator-product.v3.11.2-0.1721233953.p
        version: 3.11.2+0.1721233953.p
      - name: gatekeeper-operator-product.v3.11.2-0.1718224960.p
        version: 3.11.2+0.1718224960.p
      - name: gatekeeper-operator-product.v3.11.2
        version: 3.11.2
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.11"
    - currentCSV: gatekeeper-operator-product.v3.14.3-0.1746550072.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e47f0e9da800eb0cf4a0da6af98986508a35798a41581158ee7dbc977cb651dc
          createdAt: 16 Jan 2025, 06:57
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.14.3
          olm.substitutesFor: gatekeeper-operator-product.v3.14.3-0.1744033158.p
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the Schema for the gatekeepers API
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |-
          Gatekeeper allows administrators to detect and reject non-compliant commits to an infrastructure-as-code system's source-of-truth. This strengthens compliance efforts and prevents a bad state from slowing down the organization.

          ## Support & Troubleshooting
          Gatekeeper is available and supported as part of a valid Red Hat software subscription when it is used with Red Hat Advanced Cluster Management for Kubernetes. Troubleshooting information is available in the Gatekeeper documentation.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e47f0e9da800eb0cf4a0da6af98986508a35798a41581158ee7dbc977cb651dc
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:6255f7a673e6989617176842d8f0cc897e4630b5283e812a7a2d52dbb2a1f335
        version: 3.14.3+0.1746550072.p
      entries:
      - name: gatekeeper-operator-product.v3.14.3-0.1746550072.p
        version: 3.14.3+0.1746550072.p
      - name: gatekeeper-operator-product.v3.14.3-0.1744033158.p
        version: 3.14.3+0.1744033158.p
      - name: gatekeeper-operator-product.v3.14.3-0.1742934403.p
        version: 3.14.3+0.1742934403.p
      - name: gatekeeper-operator-product.v3.14.3-0.1740676608.p
        version: 3.14.3+0.1740676608.p
      - name: gatekeeper-operator-product.v3.14.3
        version: 3.14.3
      - name: gatekeeper-operator-product.v3.14.2
        version: 3.14.2
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.14"
    - currentCSV: gatekeeper-operator-product.v3.15.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:8b43248aed06563e06f7a0347ff1c45fa5892e3a9adb2b5c363962a0b152eff0
          createdAt: 13 Mar 2025, 17:07
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.15.4
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: Gatekeeper allows administrators to detect and reject non-compliant
          commits to an infrastructure-as-code system's source-of-truth. This strengthens
          compliance efforts and prevents a bad state from slowing down the organization.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:8b43248aed06563e06f7a0347ff1c45fa5892e3a9adb2b5c363962a0b152eff0
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:29690144667b3753a7e76095e136d6fc42325270e7b9f22f97dfd411439e9f54
        version: 3.15.4
      entries:
      - name: gatekeeper-operator-product.v3.15.4
        version: 3.15.4
      - name: gatekeeper-operator-product.v3.15.3
        version: 3.15.3
      - name: gatekeeper-operator-product.v3.15.2
        version: 3.15.2
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.15"
    - currentCSV: gatekeeper-operator-product.v3.17.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:3c027ec78398df0a8fc09eb79e222a9508b4f18154d1e1b477d4f156181ae51a
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.17.3
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e44078fc6073a79168ed52922b370773b451d69bf335aeb91d26a1462df21a9b
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:3c027ec78398df0a8fc09eb79e222a9508b4f18154d1e1b477d4f156181ae51a
        version: 3.17.3
      entries:
      - name: gatekeeper-operator-product.v3.17.3
        version: 3.17.3
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.17"
    - currentCSV: gatekeeper-operator-product.v3.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:0803b7219ceff7aa676488eeba58607a19c9c18a2bb8eae8b3b16cfec411cb6e
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.18.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:49726cb2702441305e8047b24054575170fb595cfe5ebf815cdbfdaa957b02f4
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:0803b7219ceff7aa676488eeba58607a19c9c18a2bb8eae8b3b16cfec411cb6e
        version: 3.18.1
      entries:
      - name: gatekeeper-operator-product.v3.18.1
        version: 3.18.1
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.18"
    - currentCSV: gatekeeper-operator-product.v3.19.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.19.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:60ee347425c6cfb36f0e1e834b33cbb3d77aa2d6c89ac513216d0b9fc55d4b52
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
        version: 3.19.1
      entries:
      - name: gatekeeper-operator-product.v3.19.1
        version: 3.19.1
      - name: gatekeeper-operator-product.v3.19.0
        version: 3.19.0
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.19"
    - currentCSV: gatekeeper-operator-product.v3.19.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.19.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:60ee347425c6cfb36f0e1e834b33cbb3d77aa2d6c89ac513216d0b9fc55d4b52
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
        version: 3.19.1
      entries:
      - name: gatekeeper-operator-product.v3.19.1
        version: 3.19.1
      - name: gatekeeper-operator-product.v3.19.0
        version: 3.19.0
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: stable
    defaultChannel: stable
    packageName: gatekeeper-operator-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-activemqartemis: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/
    name: amq-broker-rhel8
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-broker-operator.v7.10.7-opr-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                    "name": "ex-aao",
                    "application": "ex-aao-app"
                },
                "spec": {
                    "deploymentPlan": {
                        "image": "placeholder",
                        "size": 2,
                        "requireLogin": false,
                        "persistenceEnabled": false,
                        "journalType": "nio",
                        "messageMigration": false,
                        "jolokiaAgentEnabled": false,
                        "managementRBACEnabled": true
                    }
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                    "name": "ex-aaoaddress"
                },
                "spec": {
                    "addressName": "myAddress0",
                    "queueName": "myQueue0",
                    "routingType": "anycast"
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                    "name": "ex-aaoscaledown"
                },
                "spec": {
                    "localOnly": "true"
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                    "name": "ex-keycloak"
                },
                "spec": {
                    "loginModules": {
                        "propertiesLoginModules": [
                            {
                                "name": "prop-module",
                                "users": [
                                    {
                                        "name": "superman",
                                        "roles": [
                                            "root"
                                        ]
                                    }
                                ]
                            }
                        ]
                    },
                    "securityDomains": {
                        "brokerDomain": {
                            "name": "activemq",
                            "loginModules": [
                                {
                                    "name": "prop-module",
                                    "flag": "sufficient"
                                }
                            ]
                        }
                    }
                }
            }]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:1a7aa54d2799d238eb5f49f7a95a78a896f6bf8d222567e9118e0e3963cc9aad
          createdAt: "2021-06-18T00:02:04Z"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A stateful deployment of one or more brokers
            displayName: AMQ Broker
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: Adding and removing addresses via custom resource definitions
            displayName: AMQ Broker Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: Provides message migration on clustered broker scaledown
            displayName: AMQ Broker Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: Security configuration for the broker
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
          to deploy and manage stateful AMQ Broker broker clusters
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: rkieley@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:1a7aa54d2799d238eb5f49f7a95a78a896f6bf8d222567e9118e0e3963cc9aad
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ba21779c29b431a22c824e91820705c9501fb0b7d6f7e1ae22fcd3becc8c8a2f
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:d4010af53429502f41d722c39d69a7bb147bb06d19e3d4a69de70dbe88e3990e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b575d6eebf44dc9049445e357488efb17a690953852419b534df30e0105b9ca4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ff6fefad518a6c997d4c5a6e475ba89640260167f0bc27715daf3cc30116fad1
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:56282c6a461e945a83bffb13d3f3d8d9621800f0fc9cd755a75d21eacd8f5fbf
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:979b59325aa0f34eb05625201beba53fccbb83bd5eb80a89dcb5261ae358138f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:83faefa67b33d93836a1be3cb438bb147242d602565dcbe59b4ed42b26072191
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:967d840b831b310abd3a468a8652d1fa48b2ce28b342bfab53999245deb004f1
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:2cb14c917246ffcbd5d861ca6f3d58cb1146a4e55715ee08160e115d3dfc1c3e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f37f98c809c6f29a83e3d5a3ac4494e28efe9b25d33c54f533c6a08662244622
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:4e09ee3bd8ff3479fdd00e221b7b1505ae04ca6c64fd6bc9e2f1a34b8e98d137
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0d857b95084e39c011651be89482d4dcc77c2d1c3ac3c6a819f3c9d5a57a425f
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:9465e398a55c89c121b3ae50374414190a991839fa5e1018fc02551f6aa3b9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3c00b524722b500ff0922fe7a8722868557512183a8f69248f37a68c45c6a770
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a94fa8135d3765363e442bda45df6e02a3e22bee7ea198bcb0d3584331a4320c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c5f4c08e068b9721967cf7c7cbd9a9e93fb5e39b264dd13b653e99d8f3fa9e0e
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:aa7af7e038a12b5202108b362c8af243954603e84034a16c0192684c86d5d28b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b74d03ed852a3731467ffda95266ce49f2065972f1c37bf254f3d52b34c11991
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:7abfdcfd7b2d3d12b9234f7b057bd199bceefa47bdf02f62236f82a13fdbf5b7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:869cda365447b0b7b53ed21d2496884c15dc9c07295a9dfa0337c2c9974fe018
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:5ff28e3062d3eabb9f52d8271b837fdab06cafbd8b8f732b1bc00de49ad69aad
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f58678f03a250dbb296ebeb112c9150340155799ba4dad9e90bf347b9ed0ca7b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3219f90789690a3c5cfcc86e5f602c443e75f1fdb3a19660f43ee3996bfa8f62
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:f2651cb097eb631fc6af90e4a8404d36516aa2205c546125bf0c37739ab525fa
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:35b7d96d5065f280a528626abd2c3df84e79ccea93ab6bd8b8837af66c330f30
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:45a5e054546a2c1ee338518f8477f728f5f7108c9dd8b516b4c79d4a7d6f5ba7
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:1a9f13fd565113055b2aa2022bd0709a63c1dc058a8570e0fd7eba5c7b101334
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:97c3aa5673610875a53571449f7d5210ae8d675072d044f17b416725ed0dba88
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:ba0f5df7da08bc030ff5e0e68ecb52a931c25f6b67c6556aa8a9661c82d69d42
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:83ed89fa942ccbf4451954956db42138c83f9b015e7de07fc159ddc4bb6dc843
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9a02decbc460ddfc501d8a4a9b41cd084776090951906b027838c0829e1fc15a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ff819e913406911735da39c51699405f747bd697116a273a00897ff176bf05ff
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:5d39f26f06769fe5c989ea0e0ee8f9166120e152ce2c9bc3336fc8d7cab2ff92
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:e6f71001dccfbbe1cd7cf4fded69e181c10f46cc5c8bd3a8d3045a3f3ac41e2f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:61e5a8c1e46f1c18648e0406e2a57c52f911979079020cc57c7a4b06a56940bb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:e03945756d86bd189e8c33027fa668d2065c168510085dfd06defd9f40d3e0ab
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:34739366153b825dcb113e0ed6115ef5656085cf8d47492789ff17464042b576
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:34ab1f66d8daa544e3823464383e9788e307665c875499838fbed95f3a74a4dd
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0b2126cfb6054fdf428c1f43b69e36e93a09a49ce15350e9273c98cc08c6598b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a8b2a9364fd06c8e1c29b8f8f28d3e4844bf7840a62b67137b663c558982d4a3
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:6e9e0e3fddc44fbf87591d8904176407f1ec409d578a7d108c13cdd48d85131b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:67ae74e72927557c55611d0248347b2099b117ac1aa245dac05856fe782fb59c
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:c3ac4602b48867f3d4f7cc0efd5c9f133098e409626cfb402f80386ff460dbf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c9eba0d1f9fba588711774f4db75221f057578407065127ed60c33eef564f499
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:86ff145bd68e797a0507495570063b54cbe4186fcb4acd83c900406619ef9524
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:d4a9176f49ffd9db6051fbda15d47a1aa32c282fb5010d0caf5b9243e80bf055
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:4440b20349032bb5c0bb2a8deb89832b37767af6f0b3208b067867a85641913e
        version: 7.10.7-opr-1
      entries:
      - name: amq-broker-operator.v7.10.7-opr-1
        version: 7.10.7-opr-1
      - name: amq-broker-operator.v7.10.6-opr-1
        version: 7.10.6-opr-1
      - name: amq-broker-operator.v7.10.5-opr-1
        version: 7.10.5-opr-1
      - name: amq-broker-operator.v7.10.4-opr-1
        version: 7.10.4-opr-1
      - name: amq-broker-operator.v7.10.3-opr-3
        version: 7.10.3-opr-3
      - name: amq-broker-operator.v7.10.3-opr-2
        version: 7.10.3-opr-2
      - name: amq-broker-operator.v7.10.3-opr-1
        version: 7.10.3-opr-1
      - name: amq-broker-operator.v7.10.2-opr-2-0.1676475747.p
        version: 7.10.2-opr-2+0.1676475747.p
      - name: amq-broker-operator.v7.10.2-opr-2-0.1680622941.p
        version: 7.10.2-opr-2+0.1680622941.p
      - name: amq-broker-operator.v7.10.2-opr-2
        version: 7.10.2-opr-2
      - name: amq-broker-operator.v7.10.2-opr-1
        version: 7.10.2-opr-1
      - name: amq-broker-operator.v7.10.1-opr-2
        version: 7.10.1-opr-2
      - name: amq-broker-operator.v7.10.1-opr-1
        version: 7.10.1-opr-1
      - name: amq-broker-operator.v7.10.0-opr-4
        version: 7.10.0-opr-4
      - name: amq-broker-operator.v7.10.0-opr-3
        version: 7.10.0-opr-3
      - name: amq-broker-operator.v7.10.0-opr-2
        version: 7.10.0-opr-2
      - name: amq-broker-operator.v7.10.0-opr-1
        version: 7.10.0-opr-1
      name: 7.10.x
    - currentCSV: amq-broker-operator.v7.11.8-opr-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisSecurity is the Schema for the activemqartemissecurities
              API
            displayName: Active MQArtemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:b9aecae6b0ce0b241c7a5b9e6e096395940bb6a384373a4d121822cf10c26d99
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b74d03ed852a3731467ffda95266ce49f2065972f1c37bf254f3d52b34c11991
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:442b6638133d42fa7faa41d86e47be95cdca56c2e16efcd6e7e22991c3812a36
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f37f98c809c6f29a83e3d5a3ac4494e28efe9b25d33c54f533c6a08662244622
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0b2126cfb6054fdf428c1f43b69e36e93a09a49ce15350e9273c98cc08c6598b
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:4440b20349032bb5c0bb2a8deb89832b37767af6f0b3208b067867a85641913e
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:7abfdcfd7b2d3d12b9234f7b057bd199bceefa47bdf02f62236f82a13fdbf5b7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b9f93c243ead118651083880786a81c624f390fcc3596c7ec5ebd55f34f95ea3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:35b7d96d5065f280a528626abd2c3df84e79ccea93ab6bd8b8837af66c330f30
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b575d6eebf44dc9049445e357488efb17a690953852419b534df30e0105b9ca4
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:f2651cb097eb631fc6af90e4a8404d36516aa2205c546125bf0c37739ab525fa
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:83ed89fa942ccbf4451954956db42138c83f9b015e7de07fc159ddc4bb6dc843
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:967d840b831b310abd3a468a8652d1fa48b2ce28b342bfab53999245deb004f1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0d857b95084e39c011651be89482d4dcc77c2d1c3ac3c6a819f3c9d5a57a425f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:e03945756d86bd189e8c33027fa668d2065c168510085dfd06defd9f40d3e0ab
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:56282c6a461e945a83bffb13d3f3d8d9621800f0fc9cd755a75d21eacd8f5fbf
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5e5d8b551ab8ca377e0bc939037970a2f31ee14e9f555853ebea2759d673895a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9a02decbc460ddfc501d8a4a9b41cd084776090951906b027838c0829e1fc15a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0f443f5aa818dd85bfa93e51e37bdac3909628a283aa52a89125b527dd75d9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:5d39f26f06769fe5c989ea0e0ee8f9166120e152ce2c9bc3336fc8d7cab2ff92
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:9465e398a55c89c121b3ae50374414190a991839fa5e1018fc02551f6aa3b9aa
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:1a9f13fd565113055b2aa2022bd0709a63c1dc058a8570e0fd7eba5c7b101334
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:45a5e054546a2c1ee338518f8477f728f5f7108c9dd8b516b4c79d4a7d6f5ba7
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c9eba0d1f9fba588711774f4db75221f057578407065127ed60c33eef564f499
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a94fa8135d3765363e442bda45df6e02a3e22bee7ea198bcb0d3584331a4320c
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:34ab1f66d8daa544e3823464383e9788e307665c875499838fbed95f3a74a4dd
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:86ff145bd68e797a0507495570063b54cbe4186fcb4acd83c900406619ef9524
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c5f4c08e068b9721967cf7c7cbd9a9e93fb5e39b264dd13b653e99d8f3fa9e0e
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:d4a9176f49ffd9db6051fbda15d47a1aa32c282fb5010d0caf5b9243e80bf055
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:29f07006ad62023c20d34971cb3ae900349efae5ab2918aaa8e5647d788032be
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ff819e913406911735da39c51699405f747bd697116a273a00897ff176bf05ff
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b0b3684ef1ebf8b44d3a4c46391984173792dde1bd7363ea06d43a8f95a717d0
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:cc438b4bc2b14f6bff4e5d32e7dc0fad6d2ca391c4e84b5ae925b7b40af76729
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ba21779c29b431a22c824e91820705c9501fb0b7d6f7e1ae22fcd3becc8c8a2f
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:ba0f5df7da08bc030ff5e0e68ecb52a931c25f6b67c6556aa8a9661c82d69d42
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:e6f71001dccfbbe1cd7cf4fded69e181c10f46cc5c8bd3a8d3045a3f3ac41e2f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:67ae74e72927557c55611d0248347b2099b117ac1aa245dac05856fe782fb59c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:979b59325aa0f34eb05625201beba53fccbb83bd5eb80a89dcb5261ae358138f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ade37271426ab8f543d72f8da3e9f663d665f9427379a2539fc9323f6897cc44
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:34739366153b825dcb113e0ed6115ef5656085cf8d47492789ff17464042b576
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:869cda365447b0b7b53ed21d2496884c15dc9c07295a9dfa0337c2c9974fe018
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ff6fefad518a6c997d4c5a6e475ba89640260167f0bc27715daf3cc30116fad1
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:61e5a8c1e46f1c18648e0406e2a57c52f911979079020cc57c7a4b06a56940bb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0dde839c2dce7cb684094bf26523c8e16677de03149a0fff468b8c3f106e1f4f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5f842ebe9a6a4294a056ad10189a5f277cfd3fb4e6e0b3225e194bd5c131151d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3c00b524722b500ff0922fe7a8722868557512183a8f69248f37a68c45c6a770
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:4e09ee3bd8ff3479fdd00e221b7b1505ae04ca6c64fd6bc9e2f1a34b8e98d137
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:d4010af53429502f41d722c39d69a7bb147bb06d19e3d4a69de70dbe88e3990e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a8b2a9364fd06c8e1c29b8f8f28d3e4844bf7840a62b67137b663c558982d4a3
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:5ff28e3062d3eabb9f52d8271b837fdab06cafbd8b8f732b1bc00de49ad69aad
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:d2ae3defac36f00112c6b773e13626336ba15d6f50f905ad1c33f122bfe1a6a9
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:97c3aa5673610875a53571449f7d5210ae8d675072d044f17b416725ed0dba88
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f58678f03a250dbb296ebeb112c9150340155799ba4dad9e90bf347b9ed0ca7b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:40ccaf2bb2e0f8132f2541b39cb835f17b38050b3ca9232f0861fd5fcaf4324a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:6e9e0e3fddc44fbf87591d8904176407f1ec409d578a7d108c13cdd48d85131b
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:c3ac4602b48867f3d4f7cc0efd5c9f133098e409626cfb402f80386ff460dbf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:137b893ad15e0cd5353bae134c8fae833257a12641b12d0f774f1565ce66f63c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:89bb769b4c33eb931cc7866eeabec9a315b913851cfa84bc6e11ae8d1e8b2cbb
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:2cb14c917246ffcbd5d861ca6f3d58cb1146a4e55715ee08160e115d3dfc1c3e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:25c3732e945b4bc3e9aca721ce709d0217339b065d9dac9afdb6202e9fd8a799
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:8b4c4400a02c6cacfd978147b158e2ccbe5da955578b4b970a5ee7edac865326
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3219f90789690a3c5cfcc86e5f602c443e75f1fdb3a19660f43ee3996bfa8f62
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:7d53744ae8c77aea09c59165486c9ed0eaa131140f6b346042293bce0bd8a745
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:aa7af7e038a12b5202108b362c8af243954603e84034a16c0192684c86d5d28b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9c1ac38bf8eec67a3da57c3cc61105513776a749408cb287fe6448a17889de4e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:83faefa67b33d93836a1be3cb438bb147242d602565dcbe59b4ed42b26072191
        version: 7.11.8-opr-1
      entries:
      - name: amq-broker-operator.v7.11.8-opr-1
        version: 7.11.8-opr-1
      - name: amq-broker-operator.v7.11.7-opr-2-1742200979
        version: 7.11.7-opr-2-1742200979
      - name: amq-broker-operator.v7.11.7-opr-2-1740583644
        version: 7.11.7-opr-2-1740583644
      - name: amq-broker-operator.v7.11.7-opr-2-1736956309
        version: 7.11.7-opr-2-1736956309
      - name: amq-broker-operator.v7.11.7-opr-2-1731330819
        version: 7.11.7-opr-2-1731330819
      - name: amq-broker-operator.v7.11.7-opr-2
        version: 7.11.7-opr-2
      - name: amq-broker-operator.v7.11.7-opr-1-1726157430
        version: 7.11.7-opr-1-1726157430
      - name: amq-broker-operator.v7.11.7-opr-1-1723204800
        version: 7.11.7-opr-1-1723204800
      - name: amq-broker-operator.v7.11.7-opr-1
        version: 7.11.7-opr-1
      - name: amq-broker-operator.v7.11.6-opr-2
        version: 7.11.6-opr-2
      - name: amq-broker-operator.v7.11.5-opr-1
        version: 7.11.5-opr-1
      - name: amq-broker-operator.v7.11.4-opr-1
        version: 7.11.4-opr-1
      - name: amq-broker-operator.v7.11.3-opr-1
        version: 7.11.3-opr-1
      - name: amq-broker-operator.v7.11.2-opr-1
        version: 7.11.2-opr-1
      - name: amq-broker-operator.v7.11.1-opr-2
        version: 7.11.1-opr-2
      - name: amq-broker-operator.v7.11.1-opr-1
        version: 7.11.1-opr-1
      - name: amq-broker-operator.v7.11.0-opr-3
        version: 7.11.0-opr-3
      - name: amq-broker-operator.v7.11.0-opr-2
        version: 7.11.0-opr-2
      name: 7.11.x
    - currentCSV: amq-broker-operator.v7.12.5-opr-1-0.1757928550.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-broker-operator.v7.12.5-opr-1-0.1756216162.p
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:7d53744ae8c77aea09c59165486c9ed0eaa131140f6b346042293bce0bd8a745
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9c1ac38bf8eec67a3da57c3cc61105513776a749408cb287fe6448a17889de4e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:29f07006ad62023c20d34971cb3ae900349efae5ab2918aaa8e5647d788032be
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:dc13871dfed52402c40fd192900fc3cb5a49da4fa08e4ec4d0e23a879e83db25
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5e5d8b551ab8ca377e0bc939037970a2f31ee14e9f555853ebea2759d673895a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:cc438b4bc2b14f6bff4e5d32e7dc0fad6d2ca391c4e84b5ae925b7b40af76729
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d71506472636d940b5dbd3e708d3e0448d777eeecfd38de61ba49e281d2979c6
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ffffa9875f0379e9373f89f05eb06e5a193273bb04bc3aa5f85b044357b79098
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:b9aecae6b0ce0b241c7a5b9e6e096395940bb6a384373a4d121822cf10c26d99
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:442b6638133d42fa7faa41d86e47be95cdca56c2e16efcd6e7e22991c3812a36
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:25c3732e945b4bc3e9aca721ce709d0217339b065d9dac9afdb6202e9fd8a799
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ade37271426ab8f543d72f8da3e9f663d665f9427379a2539fc9323f6897cc44
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:16253895bc113f395dc6ab5cd511901e43b6ee676369ec80d598a8741c3c9bf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:89bb769b4c33eb931cc7866eeabec9a315b913851cfa84bc6e11ae8d1e8b2cbb
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:bb6fbd68475a7852b4d99eea6c4ab313f9267da7963162f0d75375d7063409e7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b9f93c243ead118651083880786a81c624f390fcc3596c7ec5ebd55f34f95ea3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9cc48eecf1442ae04b8543fa5d4381a13bc2831390850828834d387006d1342b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:47fd4ce2533496828aba37bd1f9715e2164d5c90bd0fc6b25e7e0786d723bf01
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:023478b773a0c42739b86e822a4d435e01ef341b8136566ea52463f1b7e447ed
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:eacc4fb55e8c4c30894a51944960d536508001a49fda62782c6cb2ca8767961d
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0dde839c2dce7cb684094bf26523c8e16677de03149a0fff468b8c3f106e1f4f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:8b4c4400a02c6cacfd978147b158e2ccbe5da955578b4b970a5ee7edac865326
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d42d713da0ce6806fdc6492b6342586783e6865a82a8647d3c4288439b1751ee
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:7f9b921d89b6d4d91b365659afff127d99f1455b94b48a10a6a7f4c38dd8b700
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:90f1678c19c74454b77aed14fa6b3074a1c6f80fe884f1c6136e7e907f402aec
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b2402179b1d0ff9c950defe5d8808195ef9a4c3e13e3c469cae31ab51d5f5014
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:8329acdff170c5a8ae6ae16f6c37d3f7ab68d106a6e8be50c3fe131ec5ecceeb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5f842ebe9a6a4294a056ad10189a5f277cfd3fb4e6e0b3225e194bd5c131151d
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:40920687402fd747ca02d384e63692a5e1db625723a2bf3056a607c92ee8eee1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0f443f5aa818dd85bfa93e51e37bdac3909628a283aa52a89125b527dd75d9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:40ccaf2bb2e0f8132f2541b39cb835f17b38050b3ca9232f0861fd5fcaf4324a
        version: 7.12.5-opr-1+0.1757928550.p
      entries:
      - name: amq-broker-operator.v7.12.5-opr-1-0.1757928550.p
        version: 7.12.5-opr-1+0.1757928550.p
      - name: amq-broker-operator.v7.12.5-opr-1-0.1756216162.p
        version: 7.12.5-opr-1+0.1756216162.p
      - name: amq-broker-operator.v7.12.5-opr-1-0.1754272113.p
        version: 7.12.5-opr-1+0.1754272113.p
      - name: amq-broker-operator.v7.12.5-opr-1
        version: 7.12.5-opr-1
      - name: amq-broker-operator.v7.12.4-opr-1
        version: 7.12.4-opr-1
      - name: amq-broker-operator.v7.12.4-opr-1-0.1747217191.p
        version: 7.12.4-opr-1+0.1747217191.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741881887.p
        version: 7.12.3-opr-1+0.1741881887.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739991235.p
        version: 7.12.3-opr-1+0.1739991235.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1730142550.p
        version: 7.12.3-opr-1+0.1730142550.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1730813578.p
        version: 7.12.3-opr-1+0.1730813578.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1731483798.p
        version: 7.12.3-opr-1+0.1731483798.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1732654016.p
        version: 7.12.3-opr-1+0.1732654016.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1733315503.p
        version: 7.12.3-opr-1+0.1733315503.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1737643631.p
        version: 7.12.3-opr-1+0.1737643631.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1738596517.p
        version: 7.12.3-opr-1+0.1738596517.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739390944.p
        version: 7.12.3-opr-1+0.1739390944.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739889884.p
        version: 7.12.3-opr-1+0.1739889884.p
      - name: amq-broker-operator.v7.12.3-opr-1
        version: 7.12.3-opr-1
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741884937.p
        version: 7.12.3-opr-1+0.1741884937.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741883943.p
        version: 7.12.3-opr-1+0.1741883943.p
      - name: amq-broker-operator.v7.12.2-opr-1
        version: 7.12.2-opr-1
      - name: amq-broker-operator.v7.12.1-opr-1-0.1724054375.p
        version: 7.12.1-opr-1+0.1724054375.p
      - name: amq-broker-operator.v7.12.1-opr-1-0.1723669176.p
        version: 7.12.1-opr-1+0.1723669176.p
      - name: amq-broker-operator.v7.12.1-opr-1-0.1721247907.p
        version: 7.12.1-opr-1+0.1721247907.p
      - name: amq-broker-operator.v7.12.1-opr-1
        version: 7.12.1-opr-1
      - name: amq-broker-operator.v7.12.0-opr-2
        version: 7.12.0-opr-2
      - name: amq-broker-operator.v7.12.0-opr-1
        version: 7.12.0-opr-1
      name: 7.12.x
    defaultChannel: 7.12.x
    packageName: amq-broker-rhel8
    provider:
      name: Red Hat
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://github.com/redhat-developer/app-services-operator/
    name: rhoas-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhoas-operator.0.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "ServiceRegistryConnection",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace",
                  "labels": {
                    "app.kubernetes.io/component": "external-service",
                    "app.kubernetes.io/managed-by": "rhoas"
                  }
                },
                "spec": {
                  "accessTokenSecretName": "rh-managed-services-api-accesstoken",
                  "serviceRegistryId": "exampleId",
                  "credentials": {
                    "serviceAccountSecretName": "service-account-secret"
                  }
                }
              },
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "CloudServiceAccountRequest",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace"
                },
                "spec": {
                  "serviceAccountName": "rhoas-sa",
                  "serviceAccountDescription": "Operator created service account",
                  "serviceAccountSecretName": "service-account-credentials",
                  "accessTokenSecretName": "rh-managed-services-api-accesstoken"
                }
              },
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "CloudServicesRequest",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace",
                  "labels": {
                    "app.kubernetes.io/component": "external-service",
                    "app.kubernetes.io/managed-by": "rhoas"
                  }
                },
                "spec": {
                  "accessTokenSecretName": "rh-cloud-services-api-accesstoken"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: quay.io/rhoas/service-operator:0.13.0
          createdAt: "2023-01-20 16:09:37"
          description: Red Hat OpenShift Application Services Operator
          olm.skipRange: '>=0.0.1 <0.13.0'
          repository: https://github.com/redhat-developer/app-services-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Creates RHOAS Service Accounts
            displayName: Cloud Service Account Request
            kind: CloudServiceAccountRequest
            name: cloudserviceaccountrequests.rhoas.redhat.com
            version: v1alpha1
          - description: Cloud Services Request
            displayName: Cloud Services Request
            kind: CloudServicesRequest
            name: cloudservicesrequests.rhoas.redhat.com
            version: v1alpha1
          - description: Binds Kafka Instance
            displayName: Kafka Connection
            kind: KafkaConnection
            name: kafkaconnections.rhoas.redhat.com
            version: v1alpha1
          - description: Binds Service Registry Instance
            displayName: Service Registry Connection
            kind: ServiceRegistryConnection
            name: serviceregistryconnections.rhoas.redhat.com
            version: v1alpha1
        description: "**OpenShift Application Services Operator (RHOAS-Operator)**
          creates instances of Red Hat OpenShift Application Services in your cluster.
          Operator is currently available as limited Beta and it cannot be consumed
          without invitation.\n\nRed Hat OpenShift Application Services, when added
          to Red Hat OpenShift Dedicated,  provide a streamlined developer experience
          for building, deploying,  and scaling cloud-native applications in open
          hybrid-cloud environments.   \n\nRHOAS-Operator enables developers to connect
          with their Application Services directly in their own cluster by using [RHOAS
          cli](https://github.com/redhat-developer/app-services-cli) and OpenShift
          Console UI.  \n\nRHOAS-Operator requires developers to connect with their
          Red Hat OpenShift Streams for Apache Kafka (RHOSAK) instances.  "
        displayName: OpenShift Application Services (RHOAS)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - rhoas
        - kafka
        - serviceregistry
        - app-services
        - application
        - services
        links:
        - name: RHOAS Operator Documentation
          url: https://github.com/redhat-developer/app-services-operator/tree/main/docs
        maintainers:
        - email: rhosak-eval-support@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://github.com/redhat-developer/app-services-operator/
        relatedImages:
        - quay.io/rhoas/service-operator:0.13.0
        version: 0.13.0
      entries:
      - name: rhoas-operator.0.13.0
        version: 0.13.0
      - name: rhoas-operator.0.12.0
        version: 0.12.0
      - name: rhoas-operator.0.11.0
        version: 0.11.0
      - name: rhoas-operator.0.9.6
        version: 0.9.6
      - name: rhoas-operator.0.9.5
        version: 0.9.5
      - name: rhoas-operator.0.9.3
        version: 0.9.3
      - name: rhoas-operator.0.9.2
        version: 0.9.2
      - name: rhoas-operator.0.9.1
        version: 0.9.1
      - name: rhoas-operator.0.9.0
        version: 0.9.0
      - name: rhoas-operator.0.8.2
        version: 0.8.2
      - name: rhoas-operator.0.7.10
        version: 0.7.10
      - name: rhoas-operator.0.7.9
        version: 0.7.9
      - name: rhoas-operator.0.7.8
        version: 0.7.8
      - name: rhoas-operator.0.7.6
        version: 0.7.6
      - name: rhoas-operator.0.7.2
        version: 0.7.2
      - name: rhoas-operator.0.7.1
        version: 0.7.1
      - name: rhoas-operator.0.6.8
        version: 0.6.8
      name: beta
    defaultChannel: beta
    packageName: rhoas-operator
    provider:
      name: Red Hat Inc.
      url: https://github.com/redhat-developer/app-services-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-s3-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-s3-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "s3.services.k8s.aws/v1alpha1",
                "kind": "Bucket",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/s3-controller:1.1.1
          createdAt: "2025-09-19T19:23:45Z"
          description: AWS S3 controller is a service controller for managing S3 resources
            in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Bucket represents the state of an AWS s3 Bucket resource.
            displayName: Bucket
            kind: Bucket
            name: buckets.s3.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Simple Storage Service (S3) resources in AWS from within your Kubernetes cluster.

          **About Amazon S3**

          Amazon Simple Storage Service (Amazon S3) is an object storage service that offers industry-leading scalability, data availability, security, and performance. This means customers of all sizes and industries can use it to store and protect any amount of data for a range of use cases, such as data lakes, websites, mobile applications, backup and restore, archive, enterprise applications, IoT devices, and big data analytics. Amazon S3 provides easy-to-use management features so you can organize your data and configure finely-tuned access controls to meet your specific business, organizational, and compliance requirements. Amazon S3 is designed for 99.999999999% (11 9s) of durability, and stores data for millions of applications for companies all around the world.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon S3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - s3
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon S3 Developer Resources
          url: https://aws.amazon.com/s3/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: s3 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/s3-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-s3-controller.v1.1.1
        version: 1.1.1
      - name: ack-s3-controller.v1.1.0
        version: 1.1.0
      - name: ack-s3-controller.v1.0.35
        version: 1.0.35
      - name: ack-s3-controller.v1.0.33
        version: 1.0.33
      - name: ack-s3-controller.v1.0.32
        version: 1.0.32
      - name: ack-s3-controller.v1.0.31
        version: 1.0.31
      - name: ack-s3-controller.v1.0.30
        version: 1.0.30
      - name: ack-s3-controller.v1.0.29
        version: 1.0.29
      - name: ack-s3-controller.v1.0.28
        version: 1.0.28
      - name: ack-s3-controller.v1.0.27
        version: 1.0.27
      - name: ack-s3-controller.v1.0.26
        version: 1.0.26
      - name: ack-s3-controller.v1.0.25
        version: 1.0.25
      - name: ack-s3-controller.v1.0.24
        version: 1.0.24
      - name: ack-s3-controller.v1.0.23
        version: 1.0.23
      - name: ack-s3-controller.v1.0.21
        version: 1.0.21
      - name: ack-s3-controller.v1.0.20
        version: 1.0.20
      - name: ack-s3-controller.v1.0.19
        version: 1.0.19
      - name: ack-s3-controller.v1.0.18
        version: 1.0.18
      - name: ack-s3-controller.v1.0.17
        version: 1.0.17
      - name: ack-s3-controller.v1.0.16
        version: 1.0.16
      - name: ack-s3-controller.v1.0.15
        version: 1.0.15
      - name: ack-s3-controller.v1.0.14
        version: 1.0.14
      - name: ack-s3-controller.v1.0.13
        version: 1.0.13
      - name: ack-s3-controller.v1.0.12
        version: 1.0.12
      - name: ack-s3-controller.v1.0.11
        version: 1.0.11
      - name: ack-s3-controller.v1.0.9
        version: 1.0.9
      - name: ack-s3-controller.v1.0.7
        version: 1.0.7
      - name: ack-s3-controller.v1.0.6
        version: 1.0.6
      - name: ack-s3-controller.v1.0.5
        version: 1.0.5
      - name: ack-s3-controller.v1.0.4
        version: 1.0.4
      - name: ack-s3-controller.v1.0.3
        version: 1.0.3
      - name: ack-s3-controller.v1.0.2
        version: 1.0.2
      - name: ack-s3-controller.v1.0.1
        version: 1.0.1
      - name: ack-s3-controller.v1.0.0
        version: 1.0.0
      - name: ack-s3-controller.v0.1.8
        version: 0.1.8
      - name: ack-s3-controller.v0.1.7
        version: 0.1.7
      - name: ack-s3-controller.v0.1.6
        version: 0.1.6
      - name: ack-s3-controller.v0.1.5
        version: 0.1.5
      - name: ack-s3-controller.v0.1.4
        version: 0.1.4
      - name: ack-s3-controller.v0.1.3
        version: 0.1.3
      - name: ack-s3-controller.v0.1.2
        version: 0.1.2
      - name: ack-s3-controller.v0.1.1
        version: 0.1.1
      - name: ack-s3-controller.v0.1.0
        version: 0.1.0
      - name: ack-s3-controller.v0.0.20
        version: 0.0.20
      - name: ack-s3-controller.v0.0.19
        version: 0.0.19
      - name: ack-s3-controller.v0.0.18
        version: 0.0.18
      - name: ack-s3-controller.v0.0.17
        version: 0.0.17
      - name: ack-s3-controller.v0.0.16
        version: 0.0.16
      - name: ack-s3-controller.v0.0.15
        version: 0.0.15
      - name: ack-s3-controller.v0.0.14
        version: 0.0.14
      - name: ack-s3-controller.v0.0.13
        version: 0.0.13
      - name: ack-s3-controller.v0.0.12
        version: 0.0.12
      - name: ack-s3-controller.v0.0.11
        version: 0.0.11
      - name: ack-s3-controller.v0.0.10
        version: 0.0.10
      - name: ack-s3-controller.v0.0.9
        version: 0.0.9
      - name: ack-s3-controller.v0.0.8
        version: 0.0.8
      - name: ack-s3-controller.v0.0.7
        version: 0.0.7
      name: alpha
    defaultChannel: alpha
    packageName: ack-s3-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: numaresources-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: numaresources-operator.v4.16.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodetopology.openshift.io/v1alpha1",
                "kind": "NUMAResourcesOperator",
                "metadata": {
                  "name": "numaresourcesoperator"
                },
                "spec": {
                  "nodeGroups": [
                    {
                      "machineConfigPoolSelector": {
                        "matchLabels": {
                          "pools.operator.machineconfiguration.openshift.io/worker": ""
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "nodetopology.openshift.io/v1alpha1",
                "kind": "NUMAResourcesScheduler",
                "metadata": {
                  "name": "numaresourcesscheduler"
                },
                "spec": {
                  "imageSpec": "URL_OF_SCHEDULER_IMAGE_FROM_REDHAT_REGISTRY",
                  "logLevel": "Normal",
                  "schedulerName": "topo-aware-scheduler"
                }
              }
            ]
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "true"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.15.0 <4.16.6'
          operators.openshift.io/valid-subscription: "[\n\t\"OpenShift Kubernetes
            Engine\",\n\t\"OpenShift Container Platform\",\n\t\"OpenShift Platform
            Plus\"\n]"
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NUMAResourcesOperator is the Schema for the numaresourcesoperators
              API
            displayName: NUMA Resources Operator
            kind: NUMAResourcesOperator
            name: numaresourcesoperators.nodetopology.openshift.io
            version: v1
          - description: NUMAResourcesOperator is the Schema for the numaresourcesoperators
              API
            displayName: NUMA Resources Operator
            kind: NUMAResourcesOperator
            name: numaresourcesoperators.nodetopology.openshift.io
            version: v1alpha1
          - description: NUMAResourcesScheduler is the Schema for the numaresourcesschedulers
              API
            displayName: NUMA Aware Scheduler
            kind: NUMAResourcesScheduler
            name: numaresourcesschedulers.nodetopology.openshift.io
            version: v1
          - description: NUMAResourcesScheduler is the Schema for the numaresourcesschedulers
              API
            displayName: NUMA Aware Scheduler
            kind: NUMAResourcesScheduler
            name: numaresourcesschedulers.nodetopology.openshift.io
            version: v1alpha1
        description: NUMA resources exporter operator
        displayName: numaresources-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - RTE
        - NUMA
        links:
        - name: Numaresources Operator
          url: https://github.com/openshift-kni/numaresources-operator
        maintainers:
        - email: fromani@redhat.com
          name: fromani
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/numaresources-rhel9-operator@sha256:aca8232e89172ca9fd9ce4a66cae9e8bb027b6d16482ce6b8e27fb7779c398e2
        version: 4.16.6
      entries:
      - name: numaresources-operator.v4.16.6
        version: 4.16.6
      - name: numaresources-operator.v4.16.5
        version: 4.16.5
      - name: numaresources-operator.v4.16.4
        version: 4.16.4
      - name: numaresources-operator.v4.16.3
        version: 4.16.3
      - name: numaresources-operator.v4.16.2
        version: 4.16.2
      - name: numaresources-operator.v4.16.1
        version: 4.16.1
      - name: numaresources-operator.v4.16.0
        version: 4.16.0
      name: "4.16"
    defaultChannel: "4.16"
    packageName: numaresources-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.9-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lvms-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lvms-operator.v4.15.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "my-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "default": true,
                        "fstype": "xfs",
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:696edeb4c168646b49b8b0ca5b3d06071ab66fdb010e7cd46888d8b2f3a0c7a8
          description: Logical volume manager storage provides dynamically provisioned
            local storage for container workloads
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.11'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "test-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
          operatorframework.io/suggested-namespace: openshift-storage
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-storage",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/internal-objects: '["logicalvolumes.topolvm.io",
            "lvmvolumegroups.lvm.topolvm.io", "lvmvolumegroupnodestatuses.lvm.topolvm.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/lvm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LogicalVolume
            name: logicalvolumes.topolvm.io
            version: v1
          - description: LVMCluster is the Schema for the lvmclusters API
            displayName: LVMCluster
            kind: LVMCluster
            name: lvmclusters.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroupNodeStatus
            name: lvmvolumegroupnodestatuses.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroup
            name: lvmvolumegroups.lvm.topolvm.io
            version: v1alpha1
        description: Logical volume manager storage provides dynamically provisioned
          local storage.
        displayName: LVM Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - local storage
        - operator
        - LVM
        links:
        - name: Source Repository
          url: https://github.com/openshift/lvm-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:771d513c4d4e5d77d797b4c18f86e23162a20c6d39955e93d24f62b94e62e785
        - registry.redhat.io/lvms4/lvms-must-gather-rhel9@sha256:382ceea1151da56d0931403207c9f85721d58a66b39b1cd79790eca6219807d0
        - registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:696edeb4c168646b49b8b0ca5b3d06071ab66fdb010e7cd46888d8b2f3a0c7a8
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:82c01635b32821f04a7cfcf80da8b3f7e097786f220e0c85aab47bda4536a1d6
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:2db4f32824a8f9f6780dc3d236f73684cb20db41466ffa8f9a89fd3f5eb62fdd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:7ce50398cd29df6e07ccd89790a53a711a7e5b484698c48f6581f5d4e9efb05a
        - registry.redhat.io/lvms4/topolvm-rhel9@sha256:074a206d36c9e4198d90baed86f3046abd2ec12f9b36a715b60900011366b3d4
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:05dceaff7f0b2b769ddfd9261cc66d42aab9158ce99ed3dbd6c229e784072635
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:479454d0d4de6609d8a4696b9c7f0097115afb45db3bdd46631d58afd77dcc58
        version: 4.15.11
      entries:
      - name: lvms-operator.v4.15.11
        version: 4.15.11
      - name: lvms-operator.v4.15.10
        version: 4.15.10
      - name: lvms-operator.v4.15.9
        version: 4.15.9
      - name: lvms-operator.v4.15.8
        version: 4.15.8
      - name: lvms-operator.v4.15.7
        version: 4.15.7
      - name: lvms-operator.v4.15.6
        version: 4.15.6
      - name: lvms-operator.v4.15.5
        version: 4.15.5
      - name: lvms-operator.v4.15.4
        version: 4.15.4
      - name: lvms-operator.v4.15.3
        version: 4.15.3
      - name: lvms-operator.v4.15.2
        version: 4.15.2
      - name: lvms-operator.v4.15.1
        version: 4.15.1
      name: stable-4.15
    - currentCSV: lvms-operator.v4.16.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "my-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "default": true,
                        "fstype": "xfs",
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:b9d80bb263a54192501d5ba9366a4f0c19b34875b51bad4fa75ef22c60bb3ee7
          description: Logical volume manager storage provides dynamically provisioned
            local storage for container workloads
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.9'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "test-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
          operatorframework.io/suggested-namespace: openshift-storage
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-storage",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["logicalvolumes.topolvm.io",
            "lvmvolumegroups.lvm.topolvm.io", "lvmvolumegroupnodestatuses.lvm.topolvm.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lvm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LogicalVolume
            name: logicalvolumes.topolvm.io
            version: v1
          - description: LVMCluster is the Schema for the lvmclusters API
            displayName: LVMCluster
            kind: LVMCluster
            name: lvmclusters.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroupNodeStatus
            name: lvmvolumegroupnodestatuses.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroup
            name: lvmvolumegroups.lvm.topolvm.io
            version: v1alpha1
        description: Logical volume manager storage provides dynamically provisioned
          local storage.
        displayName: LVM Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - local storage
        - operator
        - LVM
        links:
        - name: Source Repository
          url: https://github.com/openshift/lvm-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/lvms4/topolvm-rhel9@sha256:227748103049d2cced921ba9fc4e2d3671f08c897cc2a5627d755fe5f9d2767d
        - registry.redhat.io/lvms4/lvms-must-gather-rhel9@sha256:20b28b87b3fe6d6bf739181376231d6f8b7454c34b59eaf2e06c04a9fca8f41d
        - registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:b9d80bb263a54192501d5ba9366a4f0c19b34875b51bad4fa75ef22c60bb3ee7
        version: 4.16.9
      entries:
      - name: lvms-operator.v4.16.9
        version: 4.16.9
      - name: lvms-operator.v4.16.8
        version: 4.16.8
      - name: lvms-operator.v4.16.7
        version: 4.16.7
      - name: lvms-operator.v4.16.6
        version: 4.16.6
      - name: lvms-operator.v4.16.5
        version: 4.16.5
      - name: lvms-operator.v4.16.4
        version: 4.16.4
      - name: lvms-operator.v4.16.3
        version: 4.16.3
      - name: lvms-operator.v4.16.2
        version: 4.16.2
      - name: lvms-operator.v4.16.1
        version: 4.16.1
      - name: lvms-operator.v4.16.0
        version: 4.16.0
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: lvms-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-custom-metrics-autoscaler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: custom-metrics-autoscaler.v2.17.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eventing.keda.sh/v1alpha1",
                "kind": "CloudEventSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keda",
                    "app.kubernetes.io/instance": "cloudeventsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cloudeventsource",
                    "app.kubernetes.io/part-of": "keda"
                  },
                  "name": "cloudeventsource-sample"
                },
                "spec": {
                  "clusterName": "cluster-sample",
                  "destination": {
                    "http": {
                      "uri": "http://foo.bar"
                    }
                  }
                }
              },
              {
                "apiVersion": "eventing.keda.sh/v1alpha1",
                "kind": "ClusterCloudEventSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keda",
                    "app.kubernetes.io/instance": "clustercloudeventsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clustercloudeventsource",
                    "app.kubernetes.io/part-of": "keda"
                  },
                  "name": "clustercloudeventsource-sample"
                },
                "spec": {
                  "clusterName": "clustercluster-sample",
                  "destination": {
                    "http": {
                      "uri": "http://foo.bar"
                    }
                  }
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ClusterTriggerAuthentication",
                "metadata": {
                  "name": "example-clustertriggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "example-secret-parameter"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "KedaController",
                "metadata": {
                  "name": "keda",
                  "namespace": "openshift-keda"
                },
                "spec": {
                  "admissionWebhooks": {
                    "logEncoder": "console",
                    "logLevel": "info"
                  },
                  "metricsServer": {
                    "logLevel": "0"
                  },
                  "operator": {
                    "logEncoder": "console",
                    "logLevel": "info"
                  },
                  "serviceAccount": null,
                  "watchNamespace": ""
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledJob",
                "metadata": {
                  "name": "scaledjob-sample"
                },
                "spec": {
                  "failedJobsHistoryLimit": 5,
                  "jobTargetRef": {
                    "activeDeadlineSeconds": 600,
                    "backoffLimit": 6,
                    "completions": 1,
                    "parallelism": 1,
                    "template": null
                  },
                  "maxReplicaCount": 100,
                  "pollingInterval": 30,
                  "successfulJobsHistoryLimit": 5,
                  "triggers": [
                    {
                      "metadata": {
                        "property": "example-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledObject",
                "metadata": {
                  "name": "example-scaledobject"
                },
                "spec": {
                  "cooldownPeriod": 300,
                  "maxReplicaCount": 100,
                  "minReplicaCount": 0,
                  "pollingInterval": 30,
                  "scaleTargetRef": {
                    "name": "example-deployment"
                  },
                  "triggers": [
                    {
                      "metadata": {
                        "property": "example-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "TriggerAuthentication",
                "metadata": {
                  "name": "example-triggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "example-secret-parameter"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Cloud Provider
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9-operator@sha256:9a359ae6612d2f22fe983653418143873cb9939f83dca6f008ec3cb8f3d53990
          createdAt: 20 Aug 2025, 21:31
          description: Custom Metrics Autoscaler Operator, an event-driven autoscaler
            based upon KEDA
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.1 <2.17.2-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-keda
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/custom-metrics-autoscaler-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudEventSource
            name: cloudeventsources.eventing.keda.sh
            version: v1alpha1
          - kind: ClusterCloudEventSource
            name: clustercloudeventsources.eventing.keda.sh
            version: v1alpha1
          - description: Describes cluster wide authentication parameters
            displayName: ClusterTriggerAuthentication
            kind: ClusterTriggerAuthentication
            name: clustertriggerauthentications.keda.sh
            version: v1alpha1
          - description: |
              Represents an installation of a particular version of KEDA Controller.
            displayName: KedaController
            kind: KedaController
            name: kedacontrollers.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your jobs and what the triggers
            displayName: ScaledJob
            kind: ScaledJob
            name: scaledjobs.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your application and what the
              triggers are
            displayName: ScaledObject
            kind: ScaledObject
            name: scaledobjects.keda.sh
            version: v1alpha1
          - description: Describes authentication parameters
            displayName: TriggerAuthentication
            kind: TriggerAuthentication
            name: triggerauthentications.keda.sh
            version: v1alpha1
        description: |
          ## About the managed application
          Custom Metrics Autoscaler for OpenShift is an event driven autoscaler based upon KEDA.  Custom Metrics Autoscaler can monitor event sources like Kafka, RabbitMQ, or cloud event sources and feed the metrics from those sources into the Kubernetes horizontal pod autoscaler.  With Custom Metrics Autoscaler, you can have event driven and serverless scale of deployments within any Kubernetes cluster.
          ## About this Operator
          The Custom Metrics Autoscaler Operator deploys and manages installation of KEDA Controller in the cluster. Install this operator and follow installation instructions on how to install Custom Metrics Autoscaler in your cluster.

          ## Prerequisites for enabling this Operator
          ## How to install Custom Metrics Autoscaler in the cluster
          The installation of Custom Metrics Autoscaler is triggered by the creation of `KedaController` resource. Please refer to the [KedaController Spec](https://github.com/openshift/custom-metrics-autoscaler-operator/blob/main/README.md#the-kedacontroller-custom-resource) for more details on available options.

          Only a resource named `keda` in the namespace where the operator is installed will trigger the installation, reconfiguration or removal of the KEDA Controller resource.

          There should be only one KEDA Controller in the cluster.

          The operator will behave in this manner whether it is installed with the `AllNamespaces` or `OwnNamespace` install mode. While the operator more closely matches the `OwnNamespace` semantics, `AllNamespaces` is a supported installation mode to allow it to be installed to namespaces with existing `OperatorGroups` which require that installation mode, such as `openshift-operators`.
        displayName: Custom Metrics Autoscaler
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keda
        - scaling
        - event
        links:
        - name: Custom Metrics Autoscaler Documentation
          url: https://docs.openshift.com/container-platform/latest/nodes/cma/nodes-cma-autoscaling-custom.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9-operator@sha256:9a359ae6612d2f22fe983653418143873cb9939f83dca6f008ec3cb8f3d53990
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9@sha256:ace9358fbd2d81a6e16bfa049c763ca16acf6cc2a59cdc0b6282e5e0dfc902cc
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-adapter-rhel9@sha256:0c8ae9323f927346e168ba829f6c7679af6b2dd359c8d0a020dd0ae3954481ce
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-admission-webhooks-rhel9@sha256:75631a21de057844e9ae45ae2496a22e16f3a32a6fd5f6cb5f90e173e3be2180
        version: 2.17.2-1
      entries:
      - name: custom-metrics-autoscaler.v2.17.2-1
        version: 2.17.2-1
      - name: custom-metrics-autoscaler.v2.15.1-6
        version: 2.15.1-6
      - name: custom-metrics-autoscaler.v2.15.1-4
        version: 2.15.1-4
      - name: custom-metrics-autoscaler.v2.14.1-467
        version: 2.14.1-467
      - name: custom-metrics-autoscaler.v2.14.1-454
        version: 2.14.1-454
      - name: custom-metrics-autoscaler.v2.13.1-427
        version: 2.13.1-427
      - name: custom-metrics-autoscaler.v2.12.1-394
        version: 2.12.1-394
      - name: custom-metrics-autoscaler.v2.11.2-322
        version: 2.11.2-322
      - name: custom-metrics-autoscaler.v2.11.2-311
        version: 2.11.2-311
      - name: custom-metrics-autoscaler.v2.10.1-267
        version: 2.10.1-267
      name: stable
    defaultChannel: stable
    packageName: openshift-custom-metrics-autoscaler-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com
    name: aws-load-balancer-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-load-balancer-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
          createdAt: 2022-06-28T17:56:41
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:d7bc364512178c36671d8a4b5a76cf7cb10f8e56997106187b0fe1f032670ece
        version: 0.0.1
      entries:
      - name: aws-load-balancer-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: aws-load-balancer-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
          createdAt: 2022-12-09T15:32:32
          olm.skipRange: <0.2.0
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:cfd210a2cb705156883fda126d052e05eb84345cc27ba04688067f8ed9a802ea
        version: 0.2.0
      entries:
      - name: aws-load-balancer-operator.v0.2.0
        version: 0.2.0
      name: stable-v0
    - currentCSV: aws-load-balancer-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
          createdAt: 2022-06-28T17:56:41
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:d7bc364512178c36671d8a4b5a76cf7cb10f8e56997106187b0fe1f032670ece
        version: 0.0.1
      entries:
      - name: aws-load-balancer-operator.v0.0.1
        version: 0.0.1
      name: stable-v0.1
    - currentCSV: aws-load-balancer-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
          createdAt: 2022-12-09T15:32:32
          olm.skipRange: <0.2.0
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:cfd210a2cb705156883fda126d052e05eb84345cc27ba04688067f8ed9a802ea
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        version: 0.2.0
      entries:
      - name: aws-load-balancer-operator.v0.2.0
        version: 0.2.0
      name: stable-v0.2
    - currentCSV: aws-load-balancer-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
          createdAt: 2024-01-23T11:21:00
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:2e0b9332a44d8d9c23e19c7accab0813a651f39210257820db508cac28876595
        version: 1.1.1
      entries:
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1
    - currentCSV: aws-load-balancer-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:6e15df39a3d25755a79888f739e1af8b052c3fba57bf8f6b3434ef58fd4288cf
          createdAt: 2024-01-24T09:13:21
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat <<EOF | oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
          EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note about UPI:__ For UPI based installs, additional documentation can be found in
          [VPC and subnets](https://github.com/openshift/aws-load-balancer-operator/blob/release-1.0/docs/prerequisites.md#vpc-and-subnets)

          __Note about STS:__ For clusters using STS, use the [STS documentation](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#for-sts-clusters) instead.
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9f9f8f69ab98c1f25cc301572339e0cc5c5b76f2e29ed26e10ac91ad20e0090c
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:6e15df39a3d25755a79888f739e1af8b052c3fba57bf8f6b3434ef58fd4288cf
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:73db97d20ee9415e9a8941cc6a36ed81d901d760d9949e95e10c90866ef8bf64
        version: 1.0.1
      entries:
      - name: aws-load-balancer-operator.v1.0.1
        version: 1.0.1
      - name: aws-load-balancer-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: aws-load-balancer-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
          createdAt: 2024-01-23T11:21:00
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:2e0b9332a44d8d9c23e19c7accab0813a651f39210257820db508cac28876595
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
        version: 1.1.1
      entries:
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    defaultChannel: stable-v1
    packageName: aws-load-balancer-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres-community
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.17.0-rc1
        version: 1.17.0-rc1
      - name: stackgres-community.v1.16.3
        version: 1.16.3
      - name: stackgres-community.v1.16.2
        version: 1.16.2
      - name: stackgres-community.v1.16.1
        version: 1.16.1
      - name: stackgres-community.v1.16.0
        version: 1.16.0
      - name: stackgres-community.v1.16.0-rc3
        version: 1.16.0-rc3
      - name: stackgres-community.v1.16.0-rc2
        version: 1.16.0-rc2
      - name: stackgres-community.v1.16.0-rc1
        version: 1.16.0-rc1
      - name: stackgres-community.v1.15.2
        version: 1.15.2
      - name: stackgres-community.v1.15.1
        version: 1.15.1
      - name: stackgres-community.v1.15.0
        version: 1.15.0
      - name: stackgres-community.v1.15.0-rc4
        version: 1.15.0-rc4
      - name: stackgres-community.v1.15.0-rc3
        version: 1.15.0-rc3
      - name: stackgres-community.v1.15.0-rc2
        version: 1.15.0-rc2
      - name: stackgres-community.v1.15.0-rc1
        version: 1.15.0-rc1
      - name: stackgres-community.v1.14.3
        version: 1.14.3
      - name: stackgres-community.v1.14.2
        version: 1.14.2
      - name: stackgres-community.v1.14.1
        version: 1.14.1
      - name: stackgres-community.v1.14.0
        version: 1.14.0
      - name: stackgres-community.v1.14.0-rc2
        version: 1.14.0-rc2
      - name: stackgres-community.v1.14.0-rc1
        version: 1.14.0-rc1
      - name: stackgres-community.v1.13.0
        version: 1.13.0
      - name: stackgres-community.v1.13.0-rc1
        version: 1.13.0-rc1
      - name: stackgres-community.v1.12.0
        version: 1.12.0
      - name: stackgres-community.v1.12.0-rc1
        version: 1.12.0-rc1
      name: candidate
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.17.0-rc1
        version: 1.17.0-rc1
      - name: stackgres-community.v1.16.3
        version: 1.16.3
      - name: stackgres-community.v1.16.2
        version: 1.16.2
      - name: stackgres-community.v1.16.1
        version: 1.16.1
      - name: stackgres-community.v1.16.0
        version: 1.16.0
      - name: stackgres-community.v1.16.0-rc3
        version: 1.16.0-rc3
      - name: stackgres-community.v1.16.0-rc2
        version: 1.16.0-rc2
      - name: stackgres-community.v1.16.0-rc1
        version: 1.16.0-rc1
      - name: stackgres-community.v1.15.2
        version: 1.15.2
      - name: stackgres-community.v1.15.1
        version: 1.15.1
      - name: stackgres-community.v1.15.0
        version: 1.15.0
      - name: stackgres-community.v1.15.0-rc4
        version: 1.15.0-rc4
      - name: stackgres-community.v1.15.0-rc3
        version: 1.15.0-rc3
      - name: stackgres-community.v1.15.0-rc2
        version: 1.15.0-rc2
      - name: stackgres-community.v1.15.0-rc1
        version: 1.15.0-rc1
      - name: stackgres-community.v1.14.3
        version: 1.14.3
      - name: stackgres-community.v1.14.2
        version: 1.14.2
      - name: stackgres-community.v1.14.1
        version: 1.14.1
      - name: stackgres-community.v1.14.0
        version: 1.14.0
      - name: stackgres-community.v1.14.0-rc2
        version: 1.14.0-rc2
      - name: stackgres-community.v1.14.0-rc1
        version: 1.14.0-rc1
      - name: stackgres-community.v1.13.0
        version: 1.13.0
      - name: stackgres-community.v1.13.0-rc1
        version: 1.13.0-rc1
      - name: stackgres-community.v1.12.0
        version: 1.12.0
      - name: stackgres-community.v1.12.0-rc1
        version: 1.12.0-rc1
      name: fast
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.16.3
        version: 1.16.3
      - name: stackgres-community.v1.16.2
        version: 1.16.2
      - name: stackgres-community.v1.16.1
        version: 1.16.1
      - name: stackgres-community.v1.16.0
        version: 1.16.0
      - name: stackgres-community.v1.15.2
        version: 1.15.2
      - name: stackgres-community.v1.15.1
        version: 1.15.1
      - name: stackgres-community.v1.15.0
        version: 1.15.0
      - name: stackgres-community.v1.14.3
        version: 1.14.3
      - name: stackgres-community.v1.14.2
        version: 1.14.2
      - name: stackgres-community.v1.14.1
        version: 1.14.1
      - name: stackgres-community.v1.14.0
        version: 1.14.0
      - name: stackgres-community.v1.13.0
        version: 1.13.0
      - name: stackgres-community.v1.12.0
        version: 1.12.0
      name: stable
    defaultChannel: stable
    packageName: stackgres-community
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Rebaca Technologies Pvt Ltd
      provider-url: https://www.rebaca.com/telecom/solutions
    name: abot-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: abot-operator-rhmp.3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "abot.com/v3",
                "kind": "Abot",
                "metadata": {
                  "name": "abot"
                },
                "spec": {
                  "abotanalytics": {
                    "image": "registry.connect.redhat.com/rebaca/abot-analytics@sha256:c854d9eafbe26f834e10f5a97a587b2a2bf05558554897817bb8e2556124b75d",
                    "resources": {
                      "limits": {
                        "cpu": "4000m",
                        "ephemeral-storage": "8Gi",
                        "memory": "6000Mi"
                      },
                      "requests": {
                        "cpu": "2000m",
                        "ephemeral-storage": "4Gi",
                        "memory": "4000Mi"
                      }
                    }
                  },
                  "abotanalyticsbase": {
                    "image": "registry.connect.redhat.com/rebaca/abot-analyticsbase@sha256:dcc2a417a889acd6ec63538a6dfa393d4c1942dac178a04e5d65189a1ae054dc",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotbaseimage": "registry.connect.redhat.com/rebaca/abot-baseimage@sha256:2cbe38298d26be5896e0dc071aae111dd921d0dbd05c3525290f6ff31c73e926",
                  "abotbdd": {
                    "image": "registry.connect.redhat.com/rebaca/abot-bdd@sha256:d5c6bfe12a548cbdb4e0d83d5f18e5ab0c17c3af5340e437b2d0c23740c9c9e6",
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "ephemeral-storage": "8Gi",
                        "memory": "2000Mi"
                      },
                      "requests": {
                        "cpu": "1000m",
                        "ephemeral-storage": "4Gi",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotgrafana": {
                    "image": "registry.redhat.io/rhel8/grafana@sha256:d894cd34f2d8cac30e0fab69bd869f4b0f258b5086a8fd7ead0c5886ce2c5f7a",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "1500Mi"
                      },
                      "requests": {
                        "cpu": "300m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotgui": {
                    "image": "registry.connect.redhat.com/rebaca/abot-gui@sha256:78b512d6bd223a1f5d35f4cbd221f6637e2c9267726ce8899ee5d86aaac82b6f",
                    "resources": {
                      "limits": {
                        "cpu": "4000m",
                        "ephemeral-storage": "8Gi",
                        "memory": "6000Mi"
                      },
                      "requests": {
                        "cpu": "2000m",
                        "ephemeral-storage": "4Gi",
                        "memory": "4000Mi"
                      }
                    }
                  },
                  "abotkibana": {
                    "image": "registry.connect.redhat.com/elastic/kibana@sha256:2be3542d50f51ad786ead958a399f26852339e9417279688aad194c6794336ce",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "abotlogstash": {
                    "image": "registry.connect.redhat.com/elastic/logstash@sha256:dbc25dc5c31dfee08e86221af9f549b50075b2ccf7f255779c5f022c54cd892e",
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "2000Mi"
                      }
                    }
                  },
                  "abotmongo": {
                    "image": "registry.connect.redhat.com/rebaca/abot-mongodb@sha256:be9059e7e8bc1636a84575416b823c20463228785598662ee81c27e4c3c987dc",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "abotnginx": {
                    "image": "registry.connect.redhat.com/rebaca/abot-nginx@sha256:3edc22b40cd30230768115efdceec3d054d71f78214b4100b7425d8484154378",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "abotrabbitmq": {
                    "image": "registry.connect.redhat.com/rebaca/abot-rabbitmq@sha256:e609bec067fe7740590bb272db27d1a7770db47a0fc802e2288c022e101c9e25",
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "abotsidecar": {
                    "alertmanagerimage": "registry.redhat.io/rhacm2/prometheus-alertmanager-rhel8@sha256:ecaee8ad24a9c7cb3b945ee6d2c1275b9de2a676e1868abc2687e687385b3f6a",
                    "kubestatemetricsimage": "registry.redhat.io/rhacm2/kube-state-metrics-rhel8@sha256:e6bb6fa0c155bebf64749368dd951c15f1dbfa3559e2ba6734d3661d3a95c00c",
                    "prometheusimage": "registry.connect.redhat.com/kasten/prometheus@sha256:1e490fcccdd50c1026150ba4cad7c29297d2b73b3d5c80240ff31509e666363d",
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "8000Mi"
                      },
                      "requests": {
                        "cpu": "1000m",
                        "memory": "6000Mi"
                      }
                    }
                  },
                  "elasticsearch": {
                    "image": "registry.connect.redhat.com/elastic/elasticsearch@sha256:f47fd7df80ab05635e31d733f6969d38205f9f81cf95da8594a047f3d07c59ae",
                    "replicas": 3,
                    "resources": {
                      "limits": {
                        "cpu": "1200m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "800m",
                        "memory": "1000Mi"
                      }
                    },
                    "volumeClaim": {
                      "storage": "50Gi"
                    }
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "initContainers": {
                    "image": "registry.access.redhat.com/ubi8/ubi@sha256:a965f33ee4ee57dc8e40a1f9350ddf28ed0727b6cf80db46cdad0486a7580f9d",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "protocol": "https",
                  "pvc": {
                    "abotvol": "100Gi",
                    "analyticsvol": "100Gi",
                    "databak": "10Gi",
                    "mongovol": "10Gi",
                    "prometheusvol": "10Gi"
                  },
                  "replicas": 1,
                  "security": {
                    "fsGroup": 1018450000,
                    "seLinuxOptions": {
                      "level": "s0:c123,c456"
                    }
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "name": "abot"
                  },
                  "storageclass": "ocs-storagecluster-cephfs"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/rebaca/abot-operator@sha256:d3e8e1b6988827b3c636bdfcb57cda26870089d90cf2a732be522f5474c0a734
          createdAt: "2024-10-03T20:32:24Z"
          description: The ABot Test Bench is a test automation tool for 4G, 5G and
            ORAN Telecom networks that follow specifications and use standards driven
            design and compliance as specified by the 3GPP consortium.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/abot-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/abot-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=1.0.0 <3.0.0'
          operatorframework.io/suggested-namespace: abot
          operators.openshift.io/valid-subscription: '["ABot Basic License for 15
            days"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: https://abot.rebaca.com/support/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Abot
            name: abots.abot.com
            version: v3
        description: A Helm based operator for deploying Abot application in Openshift
          cluster. The ABot Test Bench is a test automation tool for 4G, 5G and ORAN
          Telecom networks that follow specifications and use standards driven design
          and compliance as specified by the 3GPP consortium. It provides a platform
          for a user to quickly check for protocol compliance and test for load characteristics
          of a particular system under test.
        displayName: Abot Operator-v3.0.0
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 5G
        - 4G
        - ORAN
        - test tool
        - compliance
        - interop
        links:
        - name: Abot Operator
          url: https://abot-operator.domain
        maintainers:
        - email: suchismita.nag@rebaca.com
          name: Suchismita Nag
        - email: samir@rebaca.com
          name: Samir Chatterjee
        maturity: alpha
        provider:
          name: Rebaca Technologies Pvt Ltd
          url: https://www.rebaca.com/telecom/solutions
        relatedImages:
        - registry.connect.redhat.com/rebaca/abot-rabbitmq@sha256:e609bec067fe7740590bb272db27d1a7770db47a0fc802e2288c022e101c9e25
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel8@sha256:e6bb6fa0c155bebf64749368dd951c15f1dbfa3559e2ba6734d3661d3a95c00c
        - registry.connect.redhat.com/rebaca/abot-bdd@sha256:d5c6bfe12a548cbdb4e0d83d5f18e5ab0c17c3af5340e437b2d0c23740c9c9e6
        - registry.connect.redhat.com/rebaca/abot-baseimage@sha256:2cbe38298d26be5896e0dc071aae111dd921d0dbd05c3525290f6ff31c73e926
        - registry.connect.redhat.com/elastic/elasticsearch@sha256:f47fd7df80ab05635e31d733f6969d38205f9f81cf95da8594a047f3d07c59ae
        - registry.connect.redhat.com/rebaca/abot-analyticsbase@sha256:dcc2a417a889acd6ec63538a6dfa393d4c1942dac178a04e5d65189a1ae054dc
        - registry.connect.redhat.com/rebaca/abot-mongodb@sha256:be9059e7e8bc1636a84575416b823c20463228785598662ee81c27e4c3c987dc
        - registry.connect.redhat.com/rebaca/abot-nginx@sha256:3edc22b40cd30230768115efdceec3d054d71f78214b4100b7425d8484154378
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e34f912158985525f6efb3c4b675f20854fc677c867d12fe3cd2e971309428dc
        - registry.connect.redhat.com/rebaca/abot-operator@sha256:d3e8e1b6988827b3c636bdfcb57cda26870089d90cf2a732be522f5474c0a734
        - registry.redhat.io/rhel8/grafana@sha256:d894cd34f2d8cac30e0fab69bd869f4b0f258b5086a8fd7ead0c5886ce2c5f7a
        - registry.connect.redhat.com/elastic/logstash@sha256:dbc25dc5c31dfee08e86221af9f549b50075b2ccf7f255779c5f022c54cd892e
        - registry.connect.redhat.com/kasten/prometheus@sha256:1e490fcccdd50c1026150ba4cad7c29297d2b73b3d5c80240ff31509e666363d
        - registry.connect.redhat.com/rebaca/abot-analytics@sha256:c854d9eafbe26f834e10f5a97a587b2a2bf05558554897817bb8e2556124b75d
        - registry.connect.redhat.com/rebaca/abot-gui@sha256:78b512d6bd223a1f5d35f4cbd221f6637e2c9267726ce8899ee5d86aaac82b6f
        - registry.connect.redhat.com/elastic/kibana@sha256:2be3542d50f51ad786ead958a399f26852339e9417279688aad194c6794336ce
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel8@sha256:ecaee8ad24a9c7cb3b945ee6d2c1275b9de2a676e1868abc2687e687385b3f6a
        - registry.access.redhat.com/ubi8/ubi@sha256:a965f33ee4ee57dc8e40a1f9350ddf28ed0727b6cf80db46cdad0486a7580f9d
        version: 3.0.0
      entries:
      - name: abot-operator-rhmp.3.0.0
        version: 3.0.0
      - name: abot-operator.2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: abot-operator-rhmp
    provider:
      name: Rebaca Technologies Pvt Ltd
      url: https://www.rebaca.com/telecom/solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: secrets-store-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secrets-store-csi-driver-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9-operator@sha256:96f3a8df9ac5ff3828b0add052af00f7d9df0ebb622a6adfd9772658d55baeb8
          createdAt: "2023-06-12T00:00:00Z"
          description: Install and configure Secrets Store CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.13.0-0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/secrets-store-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SecretProviderClass is the Schema for the secretproviderclasses
              API
            displayName: Secret Provider Class
            kind: SecretProviderClass
            name: secretproviderclasses.secrets-store.csi.x-k8s.io
            version: v1
          - description: SecretProviderClassPodStatus is the Schema for the secretproviderclassespodstatus
              API
            displayName: Secret Provider Class Pod Status
            kind: SecretProviderClassPodStatus
            name: secretproviderclasspodstatuses.secrets-store.csi.x-k8s.io
            version: v1
        description: |
          Operator that installs and configures the CSI driver for Secrets Store. The Secrets Store CSI Driver Operator is a Technology Preview feature. For more information on Red Hat Technology Preview features support scope, see https://access.redhat.com/support/offerings/techpreview/.
        displayName: Secrets Store CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: Documentation
          url: https://github.com/openshift/secrets-store-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/secrets-store-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: preview
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0627c09fe3f5a015f117fb30af1f4d17340aefa48b23994737d3901a804b63d3
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:a5704d26d0b562dba9633d1c107f34e05c750c0e3a64dbcace4dab7b964fff8d
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9-operator@sha256:96f3a8df9ac5ff3828b0add052af00f7d9df0ebb622a6adfd9772658d55baeb8
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9@sha256:148672ccd882a43c5c1d52763da465558b5ed54310418ec40c7ca0b56e408b1a
        version: 4.16.0-202509240123
      entries:
      - name: secrets-store-csi-driver-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: secrets-store-csi-driver-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: secrets-store-csi-driver-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: secrets-store-csi-driver-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: secrets-store-csi-driver-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: secrets-store-csi-driver-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: secrets-store-csi-driver-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: secrets-store-csi-driver-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: secrets-store-csi-driver-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: secrets-store-csi-driver-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: secrets-store-csi-driver-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: secrets-store-csi-driver-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: secrets-store-csi-driver-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: secrets-store-csi-driver-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: secrets-store-csi-driver-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: secrets-store-csi-driver-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: secrets-store-csi-driver-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: secrets-store-csi-driver-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: secrets-store-csi-driver-operator.v4.16.0-202410031007
        version: 4.16.0-202410031007
      - name: secrets-store-csi-driver-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: secrets-store-csi-driver-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: secrets-store-csi-driver-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: secrets-store-csi-driver-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: secrets-store-csi-driver-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: secrets-store-csi-driver-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: secrets-store-csi-driver-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: secrets-store-csi-driver-operator.v4.16.0-202408021139
        version: 4.16.0-202408021139
      - name: secrets-store-csi-driver-operator.v4.16.0-202407190036
        version: 4.16.0-202407190036
      - name: secrets-store-csi-driver-operator.v4.16.0-202407111006
        version: 4.16.0-202407111006
      - name: secrets-store-csi-driver-operator.v4.16.0-202407081937
        version: 4.16.0-202407081937
      - name: secrets-store-csi-driver-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: secrets-store-csi-driver-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: secrets-store-csi-driver-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: secrets-store-csi-driver-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: preview
    defaultChannel: preview
    packageName: secrets-store-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/limitador-operator
    name: limitador-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: limitador-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "name": "toy_get_route",
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/limitador-rhel9-operator@sha256:edb28f112a0d3536f2f52467f38df685d953f46f1999cbd61f4fd4fa4805cd34
          createdAt: "2025-04-07T15:52:04Z"
          description: Enables rate limiting for Gateways and applications in a Gateway
            API network
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Limitador is the Schema for the limitadors API
            displayName: Limitador
            kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: Enables rate limiting for Gateways and applications in a Gateway
          API network. - Part of Red Hat Connectivity Link
        displayName: Red Hat - Limitador Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        - kuadrant
        links:
        - name: GitHub
          url: https://github.com/kuadrant/limitador-operator
        - name: Documentation
          url: https://github.com/Kuadrant/limitador/tree/main/docs
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - registry.redhat.io/rhcl-1/limitador-rhel9-operator@sha256:edb28f112a0d3536f2f52467f38df685d953f46f1999cbd61f4fd4fa4805cd34
        - registry.redhat.io/rhcl-1/limitador-rhel9@sha256:98f85a05da53fbf7f20845d1969441620407cc13a185fd52714308679f116425
        version: 1.1.0
      entries:
      - name: limitador-operator.v1.1.0
        version: 1.1.0
      - name: limitador-operator.v1.0.2
        version: 1.0.2
      - name: limitador-operator.v1.0.1
        version: 1.0.1
      - name: limitador-operator.v0.12.1
        version: 0.12.1
      name: stable
    defaultChannel: stable
    packageName: limitador-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/limitador-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicast-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicast-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:ccd64090b85536ddaf1b81ca07d40279c9f277b1c5911628f11d678b98dc3bad
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:ccd64090b85536ddaf1b81ca07d40279c9f277b1c5911628f11d678b98dc3bad
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:f0e9f64c353fdbee3bf7ec0eaf070c3b75330ef659a881ec4ebc874ba584d085
        version: 0.5.3
      entries:
      - name: apicast-operator.v0.5.3
        version: 0.5.3
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      name: threescale-2.11
    - currentCSV: apicast-operator.v0.6.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:8c7fe66b8d4044ad965e7b3d149e9d8e3846ac4e55465873f7095a25b297b3df
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:8c7fe66b8d4044ad965e7b3d149e9d8e3846ac4e55465873f7095a25b297b3df
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:12f7a8e69a1369a51bbcd2b34d4ec8c66dd1fc9fb515349a04e7b202953423e3
        version: 0.6.4
      entries:
      - name: apicast-operator.v0.6.4
        version: 0.6.4
      - name: apicast-operator.v0.6.3
        version: 0.6.3
      - name: apicast-operator.v0.6.2
        version: 0.6.2
      - name: apicast-operator.v0.6.1
        version: 0.6.1
      - name: apicast-operator.v0.6.0
        version: 0.6.0
      - name: apicast-operator.v0.5.3
        version: 0.5.3
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      name: threescale-2.12
    - currentCSV: apicast-operator.v0.7.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:bde44c0fb9f13f212184b24e0bef45db734a8439a92179a2089e5aebd6b87b63
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          olm.skipRange: '>=0.6.2 <0.7.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.13/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:bde44c0fb9f13f212184b24e0bef45db734a8439a92179a2089e5aebd6b87b63
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:88dc7a289c803248efbe151999707325b1bc0aac2d26827ea1707e46e5c71652
        version: 0.7.5
      entries:
      - name: apicast-operator.v0.7.5
        version: 0.7.5
      name: threescale-2.13
    - currentCSV: apicast-operator.v0.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:c5e49be7b43a0058f974a5496976255ab3fbd0677bb7bcf9c3438341cf55965a
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.2 <0.8.3'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.14/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:c5e49be7b43a0058f974a5496976255ab3fbd0677bb7bcf9c3438341cf55965a
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:f8cd25658fa6a154e7ea6e5adf0f9dae3fc596ec9989021de6a021ec94eef226
        version: 0.8.3
      entries:
      - name: apicast-operator.v0.8.3
        version: 0.8.3
      - name: apicast-operator.v0.8.2
        version: 0.8.2
      - name: apicast-operator.v0.8.1
        version: 0.8.1
      name: threescale-2.14
    - currentCSV: apicast-operator.v0.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel9-operator@sha256:c4d6acae4e4cc3236c5b8d8dcfb339b072688ab114785d8598bbe68b1f3571a9
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=0.8.0 <0.12.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.14/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel9-operator@sha256:c4d6acae4e4cc3236c5b8d8dcfb339b072688ab114785d8598bbe68b1f3571a9
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        version: 0.12.5
      entries:
      - name: apicast-operator.v0.12.5
        version: 0.12.5
      - name: apicast-operator.v0.12.4
        version: 0.12.4
      - name: apicast-operator.v0.12.3
        version: 0.12.3
      - name: apicast-operator.v0.12.2
        version: 0.12.2
      - name: apicast-operator.v0.12.1
        version: 0.12.1
      - name: apicast-operator.v0.12.0
        version: 0.12.0
      name: threescale-2.15
    - currentCSV: apicast-operator.v0.12.1-mas
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-mas/apicast-rhel9-operator@sha256:e9f22a8af53f881f1948f5d6a033c35a530895633f2eae522b2fb67f9dc6088b
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.17"}]'
          olm.skipRange: '>=0.7.9 <0.12.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-mas/apicast-gateway-rhel8@sha256:f174636a9ad43a2419fff5a9e6797e3185c540a9509188fd1606aef4acea7bdf
        - registry.redhat.io/3scale-mas/apicast-rhel9-operator@sha256:e9f22a8af53f881f1948f5d6a033c35a530895633f2eae522b2fb67f9dc6088b
        version: 0.12.1-mas
      entries:
      - name: apicast-operator.v0.12.1-mas
        version: 0.12.1-mas
      - name: apicast-operator.v0.7.9-mas
        version: 0.7.9-mas
      - name: apicast-operator.v0.7.8-mas
        version: 0.7.8-mas
      - name: apicast-operator.v0.7.7-mas
        version: 0.7.7-mas
      - name: apicast-operator.v0.7.6-mas
        version: 0.7.6-mas
      - name: apicast-operator.v0.7.5-mas
        version: 0.7.5-mas
      - name: apicast-operator.v0.7.4-mas
        version: 0.7.4-mas
      - name: apicast-operator.v0.7.3-mas
        version: 0.7.3-mas
      - name: apicast-operator.v0.7.2-mas
        version: 0.7.2-mas
      - name: apicast-operator.v0.7.1-mas
        version: 0.7.1-mas
      - name: apicast-operator.v0.7.0-mas
        version: 0.7.0-mas
      - name: apicast-operator.v0.6.0-mas
        version: 0.6.0-mas
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      name: threescale-mas
    defaultChannel: threescale-2.15
    packageName: apicast-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: cloud-bulldozer
      provider-url: ""
    name: ripsaw
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ripsaw.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery, Cloud Provider
          containerImage: quay.io/cloud-bulldozer/benchmark-operator:latest
          description: '[Benchmark-operator](https://github.com/cloud-bulldozer/ripsaw/)
            is a benchmarking tool to benchmark your kubernetes cluster. The intent
            of this Operator is to deploy common workloads to establish a performance
            baseline of Kubernetes cluster'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/cloud-bulldozer/benchmark-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Benchmark
            name: benchmarks.ripsaw.cloudbulldozer.io
            version: v1alpha1
        description: '[Benchmark-operator](https://github.com/cloud-bulldozer/benchmark-operator/)
          is a benchmarking tool to benchmark your kubernetes cluster. The intent
          of this Operator is to deploy common workloads to establish a performance
          baseline of Kubernetes cluster'
        displayName: benchmark-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - performance
        - benchmark
        links:
        - name: Site
          url: https://github.com/cloud-bulldozer/benchmark-operator
        - name: Documentation
          url: https://github.com/cloud-bulldozer/benchmark-operator/tree/master/docs
        maintainers:
        - email: rsevilla@redhat.com
          name: rsevilla87
        - email: jtaleric@redhat.com
          name: jtaleric
        - email: rzaleski@redhat.com
          name: dry939
        - email: kwhitley@redhat.com
          name: whitleykeith
        maturity: alpha
        provider:
          name: cloud-bulldozer
        relatedImages:
        - quay.io/cloud-bulldozer/benchmark-operator:latest
        - k8s.gcr.io/redis:v1
        version: 0.1.0
      entries:
      - name: ripsaw.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ripsaw
    provider:
      name: cloud-bulldozer
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: stolostron-engine
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stolostron-engine.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:b7d0c8413a80c251779fe3c1c9200f9fcfbbe085ccee3ab968ebd4a39209946c
          createdAt: "2024-06-28T02:42:26Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/work@sha256:4ef2518aa107ebd28d5fb1a98d63d82a8e0610544600169e0dd5f345599e518b
        - quay.io/stolostron/assisted-installer@sha256:a31ed7b8e4ff2f9e5ade2e61191cc759e1d47ddbba95a2af41bfa53fa6ddf855
        - quay.io/stolostron/cluster-api-provider-agent@sha256:c074ca1f28d03e86efe4f09200a38af4ac947247a19c46c5300eefcd4f337006
        - quay.io/stolostron/placement@sha256:0dbcc7c62569920afcad1443529f10e38d651a5fe5ae14ade89d6e42fb522e60
        - quay.io/stolostron/managed-serviceaccount@sha256:4c3f1e0b9c4f2937ceb46e6b3f63e3a07061f9b29fdda7dcfdf9ebd74c3bb30b
        - quay.io/stolostron/backplane-operator@sha256:b7d0c8413a80c251779fe3c1c9200f9fcfbbe085ccee3ab968ebd4a39209946c
        - quay.io/stolostron/assisted-installer-controller@sha256:213dc6fde247a5b9862f5076d40394164dd596cde7ca2061e58b73f4d2dbc89e
        - quay.io/stolostron/hive@sha256:fd39b8340cfaeb9849e83bd910e8a3575791fcc58dbbdeb3e66528f009b33d65
        - quay.io/stolostron/cluster-image-set-controller@sha256:632661c6cbfce7df0545bc2fccf1ef87257abb3cc92dcede054806613a3540e3
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:f0d6d46d3a4d65fba6b74bdc9eab99e8149133f83a5c8486f22a73ab92f78ed6
        - quay.io/stolostron/assisted-service@sha256:653728f209070c6adad846a4780449f56985b19006024cfe8137a3a1854fdc85
        - quay.io/stolostron/registration@sha256:8dca354dd63739d7578c24d57b2a72c0676c979344e05c6f2634a8a9fae3218d
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:26f8bb3c9b531e96e6538a5bff96569628f45c06eb3b991ebb421d4e90d7c7e3
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/registration-operator@sha256:4b6c31592c405db11375f4694a7f15bb13c5c9a770786a1607ea0be615396435
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/discovery-testserver@sha256:83d7d6cfaef629a4e3d5d0e9fe35456aa220f1eb68f6fe611d85232acf46bc73
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/discovery-operator-tests@sha256:c98c42b0c16306de86d7b365096cb809a0c797b0a4b7dc04a7de877d0ba5733a
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:b76a201edcad4805e32c40d3f0ad3f61584a879aa00c9e49aae3b2985bfb5fbc
        - quay.io/stolostron/addon-manager@sha256:87c829911438061a53aff2b297416bf43080fdd8d93558caefc198e6f9cb382f
        - quay.io/stolostron/backplane-operator-test@sha256:961d8c40aa3526ca368fd6833202800ba2eb344c97788a646d6fb350a5338397
        - quay.io/stolostron/hypershift-operator@sha256:7dd5eaf195e6b1ca4cecd50514f5b7045948184f9ee617341bc90edc5f9914f2
        - quay.io/stolostron/backplane-must-gather@sha256:2676ebb7438ca42f0d852100612715c429fe9fbe072297fc178d51c34378d1af
        - quay.io/stolostron/discovery-operator@sha256:acb2c4033da62055feb49ff3023b2850a2768996e1fa80f52588b3e8e458e8f5
        - quay.io/stolostron/assisted-image-service@sha256:0c29a74e0f82d1782d5170ff674f5bfcd89ecc5d180fe5b583a5e0f388b9504b
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:a39c52097faf7a21cc717768401851466f8c9341f08103e4d1fe8459ae7f5ea2
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/helloworld@sha256:abc0d8002c5fe6d95a1f24fce1b7899ca24fac576c894ea7443cf7e15e6a42c2
        - quay.io/stolostron/image-based-install-operator@sha256:b9af0ddb7ae8e6fd29b953dfb917546dfd676206e2ad143e892172b6e1cea5a1
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:b269d88e30551636e024aa1472ea4dd71b873b5a2778a199b32732488fd776fb
        - quay.io/stolostron/multicloud-manager@sha256:05da692b2780453115ef6c195658cadb2f28eb0f3a86dc0a586e459dd7c55afe
        - quay.io/stolostron/cluster-curator-controller@sha256:e3a91a4b06dd73fc4e401ddf034e61d4d5b2d324d2c5cea92941e1a56c4023c1
        - quay.io/stolostron/hypershift-addon-operator@sha256:47ed6a7c00036599051718d6015bee35818ea5bbaa14f1c01bedb338dce90668
        - quay.io/stolostron/assisted-installer-agent@sha256:a2efc62ca0080779acfea2e413e524608bc0e8f72b2b917e6c7e76864d08b2b6
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/provider-credential-controller@sha256:d3bd513c06591786216f2baab46c9f57e373d1939f038edacecb2180f673d888
        - quay.io/stolostron/cluster-proxy@sha256:6cd53c588bb9c5064ba0eb8eab03cd12c86976b353209102059f7d76ec59d7c3
        - quay.io/stolostron/cluster-proxy-addon@sha256:f70a26b276c67ad791d337e35cd102afa82f1005bc6ecbc60309932d2d163a6b
        - quay.io/stolostron/managedcluster-import-controller@sha256:925bbc6173501bdb60d329a1f5a134ea3e9230034509320f93382f1c570be275
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/console-mce@sha256:4e49f2b2fdb0491c225e4d5dc143a6d7c0a25cd855057b3792c7ee94a868ac85
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/assisted-service-el8@sha256:31422ffce8b94d662131986bd084fcba503809cb939389096f9c9dbb27cb58db
        - quay.io/stolostron/clusterclaims-controller@sha256:a0613147b0ec010b8a4442a3c515f7ffc48ec7d8137436a0e8b935443cac2f96
        version: 0.4.0
      entries:
      - name: stolostron-engine.v0.4.0
        version: 0.4.0
      name: community-0.4
    - currentCSV: stolostron-engine.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:30d5bcccc542988ac788dbca305e6f29fb678470b19e48adada9ca6f5803f04a
          createdAt: "2024-10-02T05:51:59Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.5.0 <0.5.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/assisted-installer@sha256:dcc4e28c7eea55dd2ac22dd75c2b7b96ee3552fcbeba8128ad8528fda235d879
        - quay.io/stolostron/cluster-curator-controller@sha256:fc2ad8e0b6701638ab89bb62a49700fdaadfe1149228cd0b3cd4a41810020c6d
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/work@sha256:4467a4bd00586ab837cb7e8bf56ea1faff82ed5ce2fffdeb7e51ee5fafb433b0
        - quay.io/stolostron/cluster-image-set-controller@sha256:654e218ecf0a3f67520c0521b544eea9d4552e92f54bf424e88ccab24e7c151e
        - quay.io/stolostron/assisted-installer-agent@sha256:16b040dd0690b62bc051aa6369272d004d2157a8e5230d318ba827adba8e8d63
        - quay.io/stolostron/hypershift-addon-operator@sha256:ba199ebfe0eea3ef6b07123f51116831e7b0a73860b0410ece723bcc9a3230af
        - quay.io/stolostron/placement@sha256:dc0ad451b085b3a747ee648475f8e8102b645df69b898fc88f1cb190d1b95a83
        - quay.io/stolostron/cluster-proxy@sha256:be09bb293a990600ec48aba2f9bf8b457d5bdbc6b136be319e5feebadb232a8c
        - quay.io/stolostron/discovery-operator@sha256:db56e2a78ed27a46b47b519e29db66636b9abb499fd507d99451bae47438e32b
        - quay.io/stolostron/hive@sha256:707ac9a2ad18dbc4c4686f1b72946077c7232afd05505d5e50e13f806a452392
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:4b5b2bea2aad6c7e16fe1520a2fb8454b0312f2749d2f398ca54e644e813055c
        - quay.io/stolostron/cluster-api-provider-agent@sha256:1458de75311675a5acec7361e507edb24b970ffa3403e916de3413f7061bfe36
        - quay.io/stolostron/clusterclaims-controller@sha256:1c0b3c4730ef1ab628bf2c1b7cf12b1b93b149abdefc4d908dbdc90e4b6c1c19
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:2fe53994fa5bde93112a2cc75758e0f68be013a34c8206066d0b315c2e6a6f47
        - quay.io/stolostron/assisted-service-el8@sha256:c8248f9a7ac02f47373d097953cee38b66125e98b98922f18c578a591f04bcca
        - quay.io/stolostron/console-mce@sha256:10a54c5012ef6439539297919c71d9b53aaaddc6342d90677113d6ce7b42e3fe
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/backplane-operator-test@sha256:048f21cf89fb5ed2557ffc90ae5d7104cdd5ce0e6a2a865ce6f4230d09400115
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:b76a201edcad4805e32c40d3f0ad3f61584a879aa00c9e49aae3b2985bfb5fbc
        - quay.io/stolostron/addon-manager@sha256:9384ae0e9bcc9f85d3fe36e44ab7b3a61777197846559b5add0c56f513e35333
        - quay.io/stolostron/assisted-image-service@sha256:703c12a918d3c66b62c29a5abd0e8c5032299289cfed459671d6ca82c4f35c73
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/assisted-service@sha256:6383ba508a0cac030a46df96ac37daaeddc27ad1fb41d8b4dca5a87cf4ab9c31
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:14a0648b76301de182154a8d7d6b7e7eca0bfbed226e6dcebddb75af5152451f
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/registration-operator@sha256:b42341143b1e1693748f4b7d69e97c56c8a5f9408abc5db1de96ba246b7b98a9
        - quay.io/stolostron/backplane-operator@sha256:30d5bcccc542988ac788dbca305e6f29fb678470b19e48adada9ca6f5803f04a
        - quay.io/stolostron/discovery-operator-tests@sha256:e8d3d1675242272cc139b0dba15605d00e25b17b4b1cc1b5b4f3ea8913ffb2da
        - quay.io/stolostron/discovery-testserver@sha256:0b23021d6c0c7d6c7aa97a92fd0c40f31fdff6363a5b10a1edc864b5d8d54108
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/multicloud-manager@sha256:6dd212840ea7fadd427b503ce69e6e32464881a0047e2af86f9f2c7562bd4703
        - quay.io/stolostron/managedcluster-import-controller@sha256:01ab0c8b664692d69c2a9b3363e007a5611c332ac0e3491234dd0cbfba31c88a
        - quay.io/stolostron/registration@sha256:4779cf430d479806efd2f3cb7e2a9ed3dce749287609c39ad45f1b06e220d023
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/managed-serviceaccount@sha256:667f823799984404c0c1589df7b857a6a35218c2653254905dc22ca90ce1542b
        - quay.io/stolostron/backplane-must-gather@sha256:92e992e23855375a8bae542480d789d1bb25a48959896bfb75f659df2a37dd98
        - quay.io/stolostron/image-based-install-operator@sha256:bd76503406e6d20693e636461411e158c8dde3d03c4904ecb8a1850deeefa4f5
        - quay.io/stolostron/hypershift-operator@sha256:a30fb6cf4d5e103820813019ce452d8134b0fee8ecd3c2166d5681199d4a0cad
        - quay.io/stolostron/cluster-proxy-addon@sha256:7aa09b0d53fdc0f5cec9c8a13c160277a9274466025f11b5b04da95c9a10e59c
        - quay.io/stolostron/provider-credential-controller@sha256:d240ad43c23de822f91a33849448ae70a4cf27570da3387bdecd101ca2f0b65d
        - quay.io/stolostron/assisted-installer-controller@sha256:01dd7c7fb9cb335215afefde204309d3127847b900658ee9744b9d31df847565
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/helloworld@sha256:96a9b8855fb8800a85aed6fdc4911295dc2c12629bbdcc34ed8dc8061827b17e
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        version: 0.5.3
      entries:
      - name: stolostron-engine.v0.5.3
        version: 0.5.3
      - name: stolostron-engine.v0.5.2
        version: 0.5.2
      - name: stolostron-engine.v0.5.1
        version: 0.5.1
      - name: stolostron-engine.v0.5.0
        version: 0.5.0
      name: community-0.5
    - currentCSV: stolostron-engine.v0.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:fa1fa6ef3958fd9d50bfb38197f033a635a05b47c32e653822246e708d8598cc
          createdAt: "2025-02-24T01:37:53Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.6.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/discovery-operator-tests@sha256:21449ecd291cf7203cc8e27a2a00ac60e74f5993708f9e7465e8efdfdca1c89f
        - quay.io/stolostron/assisted-installer@sha256:c88976ab4b35d458a39044e522d2d5ab1916d8485af07bbef373af1bcbbd6af0
        - quay.io/stolostron/assisted-service-el8@sha256:f50f4dece91100f45ef9e20b1e52557a55db45b17a380f32d365ee013d695d5c
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:b016a001a53bc37ae534b74cc832552759800746c27e16adcf0626ff9ad2e4c0
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:2781d87c1568f2e40409f5431bfd1a9a4c9ca35655537e7d4db65d9b881edd08
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        - quay.io/stolostron/placement@sha256:543f09446652bff8c4149945d94d715eff7a76d82783f2bad89e712798542c73
        - quay.io/stolostron/provider-credential-controller@sha256:a5ac3999d0aefb1dd9a239f9394485f93e4252a45eb4a43a43e603aaff80b3bb
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:44f8bb72b975456d4151a88a560fc525ad6f4da7968a38e94a86b6dd77d3c543
        - quay.io/stolostron/registration@sha256:9c221a40ffa119ac154d76f578e0919b5a63137c908afb0583519a9b8a399b2b
        - quay.io/stolostron/hypershift-operator@sha256:2d1809223ce5e75b4133a39b78a269703c6b5234f7ec2a7c0072124c345fae0a
        - quay.io/stolostron/clusterclaims-controller@sha256:b796cb2edd3d4c22b29fc6d2b45e5328a59b7ec1bdc52d5500e1a4ee74725d6c
        - quay.io/stolostron/cluster-api-provider-agent@sha256:dda3e9aef611c744c4c89c2689a12d0649a173c1a64954c3037de55ec8eacec9
        - quay.io/stolostron/multicloud-manager@sha256:0293f0852206562b891375b167cfe17843e2e17585f04c9bb6482d1fe78089d0
        - quay.io/stolostron/backplane-operator@sha256:fa1fa6ef3958fd9d50bfb38197f033a635a05b47c32e653822246e708d8598cc
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/discovery-testserver@sha256:f757a7a11108a88de4bc0ecc5e16f74b3bd9b5110ec94bfeef2074121ec1e47f
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/cluster-proxy-addon@sha256:7da9da3e4ecd1e0df636c9f72f9f30cc7ff24b81e247ba9825cef39f18e86182
        - quay.io/stolostron/image-based-install-operator@sha256:fecbff73eb97b28bedd7c17d5adbef4826f181a6838ad4cc72e7478acf9b4254
        - quay.io/stolostron/hive@sha256:1e1f60492c6a3b43b69fe8a0640f9b81f5a1cfafc0afcd6854c3cfe362ed063a
        - quay.io/stolostron/cluster-proxy@sha256:b341dd7058cdffccb00a026908031fa123fec5ba58ba9307dbb873b75d250e4d
        - quay.io/stolostron/cluster-image-set-controller@sha256:629361c1c0ec50fe6410fc6fe6417cba89957c49ab89201c6470c0f97c8fc91d
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/assisted-image-service@sha256:6846719f91dd954ce4b4c6e7c9654c7d79db72990270acd3c34bdd2a89af34fc
        - quay.io/stolostron/console-mce@sha256:bfaf9180cd1e5af4d34daf4f1dcdee89951067a605f21bfe6e1ce6bda85b0334
        - quay.io/stolostron/helloworld@sha256:4838f699ea2461f8358dcda32b3ac96ed05b55d4e1958eac2bdacae4807501d1
        - quay.io/stolostron/ose_cluster_api_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/assisted-installer-controller@sha256:3b0b43fdbbbbd5c1e7588fc449b002e1ed9ad0ea6daa774925a962bd499a4195
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/work@sha256:6ec081826f32467ea1e1a91f6624df349a382fb8a1cbbda906141281ac2bad5c
        - quay.io/stolostron/ose_aws_cluster_api_controllers_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/backplane-operator-test@sha256:b04661ad658423c9c7557b722adb2159d988c10445de1c59f39c44583262cb42
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/assisted-installer-agent@sha256:cd74c8b8888c100bf567dd07cfff40097c458d18fcc29a11ed3830f697156267
        - quay.io/stolostron/backplane-must-gather@sha256:99b558d687e1c17036b164d819cb1c9b96a814be79e3c87960ef7cdb521eee7d
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/discovery-operator@sha256:43f40b9c20b112248fcdb42d114fe4af73fdd45b05d9ff7855d827cdd5e971b4
        - quay.io/stolostron/managed-serviceaccount@sha256:0e4edd9ab95c1558fba40a0c41c5cc326996e9e19e296324ea10b5a442a7ef63
        - quay.io/stolostron/hypershift-addon-operator@sha256:65f2b7a356c22413307ce7b7de38ce00c31e4c43b6ee3b4b484c0bc61dc483e9
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:65180432218d035ffc7996a7643ddb5300eb8b09e47e09cb270cb9b961aecf7a
        - quay.io/stolostron/managedcluster-import-controller@sha256:ecfe7c1f2d24a75bfb8db2a0a2765fe5e9136f81f341882de896431570fd10e9
        - quay.io/stolostron/addon-manager@sha256:7e347a177d8c5b80a78411a3e8dbf04e49d157a4ca67e7fd5170970c2ba62da9
        - quay.io/stolostron/registration-operator@sha256:52289e6cf5f159d344717b49e6226ff9d4ee33942f63cc543d1b820d7ee24fc5
        - quay.io/stolostron/assisted-service@sha256:58271e54eda979c43fb474c3bb255f7328e380a7af9ba833e5f7558f323bd239
        - quay.io/stolostron/cluster-curator-controller@sha256:5d05462c2c9d93b21656a436c01fbda4256757bd057e5de7014eb6f320310a1e
        version: 0.6.3
      entries:
      - name: stolostron-engine.v0.6.3
        version: 0.6.3
      - name: stolostron-engine.v0.6.2
        version: 0.6.2
      - name: stolostron-engine.v0.6.1
        version: 0.6.1
      - name: stolostron-engine.v0.6.0
        version: 0.6.0
      name: community-0.6
    - currentCSV: stolostron-engine.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:9b39a5b6a502778d7884e276b276a9b44e6dbc88e7e82470e26d91ded3cb3839
          createdAt: "2025-06-09T15:04:58Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.7.1'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational components for managing multiple Kubernetes-based clusters. The deployment of the multicluster engine components is
              determined based on the configuration that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/cluster-image-set-controller@sha256:7f05ad13ffaa11daec066c5889b87df1b631e6eb3dc1c2913a2766c8e65218aa
        - quay.io/stolostron/cluster-proxy@sha256:1528e36dfa4fa2c9063d84c0f2b9aa8d2b549a750e812d4a2e2b8c7515bfe5cb
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:1e95f55d6918263e79de1c1e2be8a96cd3d35a346dde6d3e2fb484f48e567811
        - quay.io/stolostron/hive@sha256:86ff94d3e7c174ac4d6818183c66791edc7638965dd09e2d41172609c81f451f
        - quay.io/stolostron/cluster-api-provider-openshift-assisted-bootstrap@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/console-mce@sha256:7f538a21eb22ccd57765f47692da23be9cecfd3b7b65986f1413234bdf709530
        - quay.io/stolostron/assisted-installer-agent@sha256:ffceb06ce038013291e092324d3c49091cd1b2ff1569abe77c29484075e28107
        - quay.io/stolostron/hypershift-addon-operator@sha256:ed1c1580fee28310fc1b137c5c70271448e58ed29af4371515fe9c74512a2c7e
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/placement@sha256:6276e411c9f7cf7072d3d364569b6c3cd472742ab4ad48912f18cb4091d27a6f
        - quay.io/stolostron/ose_cluster_api_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/cluster-proxy-addon@sha256:c776405c4b1b623500c69a12b7d76563646b27cf23bab74dc8f1361534cfb81e
        - quay.io/stolostron/addon-manager@sha256:a245847f135f96c4b4fcaa673fc32013c8e407252b3259175c796e3c16111c71
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/discovery-testserver@sha256:faef24208a65283b4157521079a7bf23bc4ebd553d70a83a78c324b5cb0e24e4
        - quay.io/stolostron/discovery-operator@sha256:a6a37237bc06f0635a936b803d3c53f0f0caf5426c76b37951c4e157b181ecb3
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:1d4221cfc364743b58e15bad3afe9548f4151f06b0dbf0aec7c00715ba75abba
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/ip-address-manager@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/assisted-installer@sha256:17b78d73775a93b31ca872db79520ddc49d2be042fdf1599f419fbcbfd9a98ac
        - quay.io/stolostron/assisted-service-el8@sha256:f52c55691850a1f2f254b0e9377ba97493be337e012223e5a2b3190f03e1a349
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:44f8bb72b975456d4151a88a560fc525ad6f4da7968a38e94a86b6dd77d3c543
        - quay.io/stolostron/discovery-operator-tests@sha256:d96fe1f4f49ecfd4a27bedd87d7070a3c7f3f82b6ee8bb59e98a262d8cedcfc0
        - quay.io/stolostron/clusterclaims-controller@sha256:afab145b12a52f81e1f2eaa189731e3a6287b54f2f1f1ad2976ad378fa7dfa3a
        - quay.io/stolostron/ose_aws_cluster_api_controllers_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/backplane-operator@sha256:9b39a5b6a502778d7884e276b276a9b44e6dbc88e7e82470e26d91ded3cb3839
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        - quay.io/stolostron/assisted-image-service@sha256:b6df9c9fc8bbfdc58e2169b9e199e1eed188d3d847511eba01eb4920fabe3992
        - quay.io/stolostron/multicloud-manager@sha256:5a95a5d1dbb17456d7ce62f1a517ae49ca780e25f3ebb74b32346de1df23b4c9
        - quay.io/stolostron/helloworld@sha256:a790795409f6e781f506b6a63f39f7d344f01efa30c3de87b1727dc2010d5bbf
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/backplane-must-gather@sha256:f4f9596eade78333628199f62f8e71e4c2b3957c20da5bad807c12a2a02560db
        - quay.io/stolostron/managedcluster-import-controller@sha256:4d0be693db0576eb2e9e3d409bf1479cc60784f5747ed1b25b6a2443c5e91e36
        - quay.io/stolostron/cluster-api-provider-agent@sha256:2b22890b948ad1d597c424d73d4fe34c7c4d265ad6e54fa0ed87a7dec920ac57
        - quay.io/stolostron/ose-baremetal-cluster-api-controllers-rhel9@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/cluster-curator-controller@sha256:d667bd0892812eb4e1f724679759469db335db4ab801f9d57ea8818d8cd13188
        - quay.io/stolostron/image-based-install-operator@sha256:9d6796007b616de60560f55f14546295c4e48b3768d059098fce9367674a3368
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/registration@sha256:2db2c28af25007c77979cc456d929f5f077cd671cc4e0562c43efce2c42c4cd8
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/provider-credential-controller@sha256:3484cb0c2102f4a522efe31bc4c77910a202b42e1ad082a0a8709cc1a7698649
        - quay.io/stolostron/work@sha256:709c41d1a7a1eaf10b334dd0c0105a261d6db811693d5ae67eb55b9a26daed87
        - quay.io/stolostron/assisted-installer-controller@sha256:8ea9d12fa30825fe4f965a7d506af687e8de709f8d07dd75204747da2a6e09a8
        - quay.io/stolostron/assisted-service@sha256:af3532d487e03a077ce23e4178abbcb41cf45726b646d7660840c036677715e8
        - quay.io/stolostron/hypershift-operator@sha256:c9a8fc943ef35126434cc767f152abd09943a4d2d194a9966495fdc0870b417c
        - quay.io/stolostron/registration-operator@sha256:a9fdf837e149c8659fc99deab6a284975c8eb8785549381fa082303cdb172e44
        - quay.io/stolostron/managed-serviceaccount@sha256:4bcd5c9b2044254c28215fd78bf1e3fcd919578238986831fcea8055b2764696
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:1757b9ad635839087c4bef198d5e81f4abb7aee38bc74690021f9e8887366bcf
        - quay.io/stolostron/backplane-operator-test@sha256:8d6514fc0db5ee79c346ddc31bb094ad87fa375519da67e32398300d70d72f6e
        - quay.io/stolostron/cluster-api-provider-openshift-assisted-control-plane@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        version: 0.7.1
      entries:
      - name: stolostron-engine.v0.7.1
        version: 0.7.1
      - name: stolostron-engine.v0.7.0
        version: 0.7.0
      name: community-0.7
    defaultChannel: community-0.7
    packageName: stolostron-engine
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pulp Community
      provider-url: https://github.com/pulp/pulp-operator
    name: pulp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pulp-operator.v1.0.0-alpha.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache_enabled": true,
                  "content": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "ingress_type": "Route",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/"
                  },
                  "storage_type": "File",
                  "worker": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example-pulp"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true,
                    "redis_storage_class": "standard"
                  },
                  "content": {
                    "replicas": 1
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "pulp",
                  "file_storage_access_mode": "ReadWriteOnce",
                  "file_storage_size": "2Gi",
                  "file_storage_storage_class": "standard",
                  "image_version": "nightly",
                  "image_web_version": "nightly",
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/",
                    "telemetry": false
                  },
                  "web": {
                    "replicas": 1
                  },
                  "worker": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "galaxy-example"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "api": {
                    "replicas": 1
                  },
                  "content": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "galaxy",
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "file_storage_storage_class": "standard",
                  "image": "quay.io/pulp/galaxy-minimal",
                  "image_version": "nightly",
                  "image_web": "quay.io/pulp/galaxy-web",
                  "image_web_version": "nightly",
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/api/galaxy/pulp/",
                    "telemetry": false
                  },
                  "signing_scripts_configmap": "signing-scripts",
                  "signing_secret": "signing-galaxy",
                  "storage_type": "File",
                  "web": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "worker": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "galaxy-example",
                  "deployment_type": "galaxy",
                  "instance_name": "pulp",
                  "postgres_configuration_secret": "galaxy-example-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "galaxy-example"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.0.0-alpha.1
          createdAt: "2021-03-26 16:57:40"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1alpha1
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1alpha1
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1alpha1
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://docs.pulpproject.org/](https://docs.pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          Note that Pulp operator works with three different types of service containers (the operator itself, the main service and the web service):

          |           | Operator | Main | Web |
          | --------- | -------- | ---- | --- |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[pulp-minimal](https://quay.io/repository/pulp/pulp-minimal?tab=tags) | [pulp-web](https://quay.io/repository/pulp/pulp-web?tab=tags) |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[galaxy-minimal](https://quay.io/repository/pulp/galaxy-minimal?tab=tags) | [galaxy-web](https://quay.io/repository/pulp/galaxy-web?tab=tags) |

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://docs.pulpproject.org/pulp_operator/container/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://docs.pulpproject.org/pulp_operator/](https://docs.pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://docs.pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: beta
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - docker.io/library/postgres:13
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/pulp/pulp-operator:v1.0.0-alpha.1
        - quay.io/pulp/pulp-minimal:stable
        - quay.io/pulp/pulp-web:stable
        - docker.io/library/redis:latest
        version: 1.0.0-alpha.1
      entries:
      - name: pulp-operator.v1.0.0-alpha.1
        version: 1.0.0-alpha.1
      name: alpha
    - currentCSV: pulp-operator.v1.0.0-beta.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "Pulp",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true
                  },
                  "content": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/"
                  },
                  "worker": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "Pulp",
                "metadata": {
                  "name": "galaxy-example"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "api": {
                    "replicas": 1
                  },
                  "content": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "galaxy",
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "file_storage_storage_class": "standard",
                  "image": "quay.io/pulp/galaxy-minimal",
                  "image_version": "4.6.3",
                  "image_web": "quay.io/pulp/galaxy-web",
                  "image_web_version": "4.6.3",
                  "ingress_type": "nodeport",
                  "nodeport_port": 30000,
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "analytics": false,
                    "ansible_api_hostname": "http://nodeport.local:30000",
                    "api_root": "/api/galaxy/pulp/",
                    "content_origin": "http://nodeport.local:30000",
                    "galaxy_feature_flags": {
                      "ai_deny_index": true,
                      "execution_environments": true
                    },
                    "token_server": "http://nodeport.local:30000/token/"
                  },
                  "signing_scripts": "signing-scripts",
                  "signing_secret": "signing-galaxy",
                  "web": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "worker": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "galaxy-example",
                  "deployment_type": "galaxy",
                  "postgres_configuration_secret": "galaxy-example-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "galaxy-example"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.0.0-beta.4
          createdAt: "2024-01-18T20:33:37Z"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1beta2
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1beta2
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1beta2
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://docs.pulpproject.org/](https://docs.pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          Note that Pulp operator works with three different types of service containers (the operator itself, the main service and the web service):

          |           | Operator | Main | Web |
          | --------- | -------- | ---- | --- |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[pulp-minimal](https://quay.io/repository/pulp/pulp-minimal?tab=tags) | [pulp-web](https://quay.io/repository/pulp/pulp-web?tab=tags) |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[galaxy-minimal](https://quay.io/repository/pulp/galaxy-minimal?tab=tags) | [galaxy-web](https://quay.io/repository/pulp/galaxy-web?tab=tags) |

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://docs.pulpproject.org/pulp_operator/container/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://docs.pulpproject.org/pulp_operator/](https://docs.pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://docs.pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: beta
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.4
        - quay.io/pulp/pulp-operator:v1.0.0-beta.4
        - quay.io/pulp/pulp-minimal:stable
        - quay.io/pulp/pulp-web:stable
        - docker.io/library/redis:latest
        - docker.io/library/postgres:13
        version: 1.0.0-beta.4
      entries:
      - name: pulp-operator.v1.0.0-beta.4
        version: 1.0.0-beta.4
      - name: pulp-operator.v1.0.0-beta.3
        version: 1.0.0-beta.3
      - name: pulp-operator.v1.0.0-beta.2
        version: 1.0.0-beta.2
      - name: pulp-operator.v1.0.0-beta.1
        version: 1.0.0-beta.1
      - name: pulp-operator.v1.0.0-alpha.9
        version: 1.0.0-alpha.9
      - name: pulp-operator.v1.0.0-alpha.8
        version: 1.0.0-alpha.8
      - name: pulp-operator.v1.0.0-alpha.7
        version: 1.0.0-alpha.7
      - name: pulp-operator.v1.0.0-alpha.6
        version: 1.0.0-alpha.6
      - name: pulp-operator.v1.0.0-alpha.5
        version: 1.0.0-alpha.5
      - name: pulp-operator.v1.0.0-alpha.4
        version: 1.0.0-alpha.4
      - name: pulp-operator.v1.0.0-alpha.3
        version: 1.0.0-alpha.3
      - name: pulp-operator.v1.0.0-alpha.2
        version: 1.0.0-alpha.2
      - name: pulp-operator.v1.0.0-alpha.1
        version: 1.0.0-alpha.1
      name: beta
    defaultChannel: beta
    packageName: pulp-operator
    provider:
      name: Pulp Community
      url: https://github.com/pulp/pulp-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-common-service-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-common-service-operator.v4.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v3",
                "kind": "CommonService",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-common-service-operator",
                    "app.kubernetes.io/managed-by": "ibm-common-service-operator",
                    "app.kubernetes.io/name": "ibm-common-service-operator"
                  },
                  "name": "example-commonservice"
                },
                "spec": {
                  "size": "starterset"
                }
              }
            ]
          capabilities: Seamless Upgrades
          cloudPakThemesVersion: styles4100.css
          containerImage: registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
          createdAt: "2025-03-18T21:27:31Z"
          description: The IBM Cloud Pak foundational services operator is used to
            deploy IBM foundational services.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-common-service-operator
          nss.operator.ibm.com/managed-webhooks: ""
          olm.skipRange: '>=3.3.0 <4.12.0'
          operatorChannel: v4.12
          operatorVersion: 4.12.0
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/ibm-common-service-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CommonService is the Schema for the commonservices API. Documentation
              For additional details regarding install parameters check https://ibm.biz/icpfs39install.
              License By installing this product you accept the license terms https://ibm.biz/icpfs39license
            displayName: CommonService
            kind: CommonService
            name: commonservices.operator.ibm.com
            version: v3
        description: |-
          # Introduction

          The `IBM Cloud Pak foundational services` is a bridge to connect IBM Cloud Paks and Operand Deployment Lifecycle Manager (ODLM) with IBM Cloud Pak foundational services. You can also install the `IBM Cloud Pak foundational services` in stand-alone mode.

          When you install this operator, the operator completes the following tasks:

            - Installs ODLM operator namespace
            - Creates the Common Services `OperandRegistry` and `OperandConfig` in the service namespace

          # Details
          For more information about installing this operator and other Common Services operators, see [Installer documentation](http://ibm.biz/cpcs_opinstall). If you are using this operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For more information about IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).

          For more information about the available IBM Cloud Pak foundational services, see the [IBM Knowledge Center](http://ibm.biz/cpcsdocs).

          ## Supported platforms

          Red Hat OpenShift Container Platform 4.5 or newer installed on one of the following platforms:

           - Linux x86_64
           - Linux on Power (ppc64le)
           - Linux on IBM Z and LinuxONE

          ## Operator versions

           - 4.12.0

          ## Prerequisites

          Before you install this operator, you need to first install the operator prerequisites:

          - For the list of prerequisites for installing the operator, see the IBM Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).

          ### Resources Required

          #### Minimum scheduling capacity

          | Software                   | Memory (MB) | CPU (cores) | Disk (GB) | Nodes  |
          | -------------------------- | ----------- | ----------- | --------- | ------ |
          | ibm common service operator | 200          | 0.2        | 1          | worker |
          | **Total**                  | 200         | 0.2         | 1         |        |

          ## Documentation

          - If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak. For a list of IBM Cloud Paks, see [IBM Cloud Paks that use IBM Cloud Pak foundational services](http://ibm.biz/cpcs_cloudpaks).
          - If you are using the operator in stand-alone mode or with an IBM Containerized Software, see the IBM Cloud Pak foundational services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall).

          ## Installing

          For installation, see the [IBM Cloud Platform Common Services documentation](http://ibm.biz/cpcsdocs).

          ## Configuration

          For configuration, see the [IBM Cloud Platform Common Services documentation](http://ibm.biz/cpcsdocs).

          ### SecurityContextConstraints Requirements

          The IBM Common Service Operator supports running with the OpenShift Container Platform 4.3 default restricted Security Context Constraints (SCCs) and IBM Cloud Pak Security Context Constraints (SCCs).

          For more information about the OpenShift Container Platform Security Context Constraints, see [Managing Security Context Constraints](https://docs.openshift.com/container-platform/4.3/authentication/managing-security-context-constraints.html).

          For more information about the IBM Cloud Pak Security Context Constraints, see [Managing Security Context Constraints](https://ibm.biz/cpkspec-scc).

          ### Limitations

          No
        displayName: IBM Cloud Pak foundational services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Common Service Operator
          url: https://github.com/IBM/ibm-common-service-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90046eb6170e12439ccc5aa343f19a4d4fc99a10d98b383cf1dc0e020ebeed55
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:18f44ab15289625fa2c6b72ece6e06afb51058dcba30b8f38b4fdeedf8c9e51b
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:0fdf1beeb4888f648f6dbd17a29dcc53a715d2c4301bc714a4172b2daf8d13be
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:4209e78f12120b9c9f320496b780c3993a9317825ea17dc7d554452b28c99014
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:9a537eae7e581322e2f0764ed38a797ba7cd260724a285f24d5095058b7b47a0
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-operator-bundle@sha256:086eaeaaaf440a1d4a254984afacd8d04583aafb9ec11ac072feb42a87b986f7
        version: 4.12.0
      entries:
      - name: ibm-common-service-operator.v4.12.0
        version: 4.12.0
      name: v4.12
    defaultChannel: v4.12
    packageName: ibm-common-service-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: rebellions
      provider-url: https://rebellions.ai/
    name: rbln-npu-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: candidate
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: fast
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: rbln-npu-operator
    provider:
      name: rebellions
      url: https://rebellions.ai/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/component: operator
      app.kubernetes.io/managed-by: kustomize
      app.kubernetes.io/name: zabbix-operator-certified
      app.kubernetes.io/part-of: zabbix-operator-certified
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zabbix LLC
      provider-url: ""
    name: zabbix-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zabbix-operator-certified-rhmp.v6.0.40
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent",
                "metadata": {
                  "labels": {
                    "app": "agent",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "log_remote_commands": true,
                  "max_lines_per_second": 20,
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent2",
                "metadata": {
                  "labels": {
                    "app": "agent2",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent2"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "enable_persistent_buffer": false,
                  "enable_status_port": false,
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "persistent_buffer_claim": "",
                  "persistent_buffer_period": "1h",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixFull",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-full"
                },
                "spec": {
                  "builtin_report_service": true,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysql_volumeclaim": "mysql-full-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxyMysql",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-mysql"
                },
                "spec": {
                  "db_server_port": 3306,
                  "enable_snmp_traps": false,
                  "internal_db": true,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix_proxy",
                  "proxy": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "zabbix_mysql_volumeclaim": "mysql-proxy-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxySqlite",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-sqlite"
                },
                "spec": {
                  "db_data_volumeclaim": "zabbix-proxy-sqlite",
                  "enable_snmp_traps": false,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "proxy": {
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "node_name_as_db_name": true,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "proxy_size": 1,
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  }
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixServer",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-server"
                },
                "spec": {
                  "builtin_report_service": true,
                  "db_server_host": "mysql-server",
                  "db_server_port": 3306,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix",
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              }
            ]
          ansible.operator-sdk/reconcile-period: 5s
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          certified: "true"
          containerImage: registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
          createdAt: "2025-05-26T07:41:22Z"
          description: Zabbix operator with multiple deployment variants and different
            components
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/zabbix-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/zabbix-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <6.0.40
          operatorframework.io/suggested-namespace: zabbix
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: ""
          support: Zabbix
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Zabbix agent 2 is a new generation of Zabbix agent and may
              be used in place of Zabbix agent
            displayName: Zabbix agent 2
            kind: ZabbixAgent2
            name: zabbixagent2s.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix agent is deployed on a monitoring nodes to actively
              monitor local resources and applications
            displayName: Zabbix agent
            kind: ZabbixAgent
            name: zabbixagents.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database server, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Full
            kind: ZabbixFull
            name: zabbixfulls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with MySQL database server
            displayName: Zabbix proxy (MySQL)
            kind: ZabbixProxyMysql
            name: zabbixproxymysqls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with SQLite database
            displayName: Zabbix proxy (SQLite3)
            kind: ZabbixProxySqlite
            name: zabbixproxysqlites.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database support, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Server
            kind: ZabbixServer
            name: zabbixservers.kubernetes.zabbix.com
            version: v1alpha1
        description: "## About this Operator\n\nZabbix helps you to real-time monitoring
          of millions of metrics collected from tens of thousands of servers, virtual
          machines and network devices.\nThe Zabbix Operator allows users to easily
          deploy, manage, and maintain Zabbix deployments on OpenShift. By installing
          this integration you will be able to deploy Zabbix server / proxies and
          other components with a single command.\n## Supported Features\n* **Zabbix
          Server** - Simple Zabbix installation with included Zabbix server, Zabbix
          web-interface and Zabbix Java Gateway with MySQL database support. The feature
          does not provide MySQL service and requires an external MySQL database.
          \n* **Zabbix Server (Full)** - Zabbix installation with included Zabbix
          server, Zabbix web-interface, Zabbix Java Gateway and MySQL server instance.\n*
          **Zabbix proxy (SQLite3)** - Very simple way to gain power of Zabbix proxy.
          The feature has  SQLite3 support for Zabbix proxies and allow to specify
          amount of proxies. \n* **Zabbix proxy (MySQL)** - Another option of Zabbix
          proxy. The option support and deliver MySQL database.\n* **Zabbix agent**
          - Zabbix agent can be deployed on each available node for stability and
          performance monitoring on remote nodes. It allows to gather metrics with
          full automation!* \n## Prerequisites\nAll deployment options are require
          additional information during deployment. Please, check the following instructions
          and provide required configuration:\n* **Zabbix Server** - MySQL database
          host information and MySQL database credentials in specially formatted *Secret*.
          Additionally it is possible to specify SSL certificates for HTTPS support
          in *Secret*.\n* **Zabbix Server (Full)** - MySQL database credentials in
          specially formatted *Secret*. MySQL database volume name information. Additionally
          it is possible to specify SSL certificates for HTTPS support in *Secret*.\n*
          **Zabbix proxy (SQLite3)** - Zabbix server host information only.\n* **Zabbix
          proxy (MySQL)** - MySQL database credentials in specially formatted *Secrets*
          and Zabbix server host information.\n* **Zabbix agent** - Zabbix server
          host information only for outgoing and incoming connections. Zabbix agent
          uses \"privileged\" mode to monitor node resources! For example, running
          processes.* \n### MySQL credentials\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-full-secrets\ndata:\n  mysql_root_password: emFiYml4X3N1cGVyX3Jvb3Q=
          [1]\n  mysql_zabbix_password: emFiYml4X3VzZXJfcGFzc3dvcmQ= [2]\n  mysql_zabbix_username:
          emFiYml4 [3]\ntype: Opaque\n```\nUsing MySQL root password (*mysqlrootpassword*
          [1]) Zabbix server / proxy will try to create MySQL database schema with
          grant permissions to *mysqlzabbixusername* [2] and *mysqlzabbixpassword*
          [3].\n### SSL certificates for HTTPS\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-web-sslsecret\ndata:\n  ssl.crt: >-\n   < ssl.crt data>\n
          \ ssl.key: >-\n\t < ssl.key data >\n  dhparam.pem: >-\n   <  dhparam.pem
          data >\n```\nFiles *ssl.crt*, *ssl.key* and *dhparam.perm* are required
          for Zabbix web-interface for SSL support.\n"
        displayName: Zabbix Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - zabbix
        - appliance
        - agent
        - metrics
        links:
        - name: Zabbix
          url: https://www.zabbix.com
        - name: Zabbix Official Documentation
          url: https://www.zabbix.com/documentation/6.0/en/manual/quickstart
        - name: Zabbix Operator Installation
          url: https://www.zabbix.com/documentation/6.0/en/manual/installation/containers/openshift
        - name: Downloads
          url: https://www.zabbix.com/download
        maintainers:
        - email: alexey.pustovalov@zabbix.com
          name: Alexey Pustovalov
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Zabbix LLC
        relatedImages:
        - registry.connect.redhat.com/zabbix/zabbix-agent2-60@sha256:424003d52f3cb4360adf2b3619d8b9f8de9196c630ee7e4ceb5d177c1c5a36e4
        - registry.connect.redhat.com/zabbix/zabbix-server-mysql-60@sha256:637484faec6e9a55049c13dbd1f100c03d757c3a21749388c0605040f02d05ce
        - registry.connect.redhat.com/zabbix/zabbix-proxy-mysql-60@sha256:c2ae74d97c083ff348dc9c55932a33da4779cd9efdd9d7c5a79acfffae5b9ed0
        - registry.connect.redhat.com/zabbix/zabbix-java-gateway-60@sha256:a4feb725c12c4ac90eb9244c6fafdb6708679f2e0e7f6aa683c765127306412e
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
        - registry.connect.redhat.com/zabbix/zabbix-web-mysql-60@sha256:611d2670e1bf95b81226976c0a1f64434b2a7766f196dadbd646a7d8319f4cb1
        - registry.connect.redhat.com/zabbix/zabbix-web-service-60@sha256:4a333094e110713470a897b6406c5e340e962caf32a429dddf19865041fcc448
        - registry.redhat.io/rhel9/mysql-80@sha256:c4b5246d0c0d089a766ea07768c22d9d6be8049ac1379bd345db99390ffe2a36
        - registry.connect.redhat.com/zabbix/zabbix-snmptraps-60@sha256:e61e6b1c551dec8488cce0998f94d83d6d10b5d89a465d7119b088313032cb81
        - registry.connect.redhat.com/zabbix/zabbix-proxy-sqlite-60@sha256:62e62ebc8c1bc13fb610f7a83412e65d852f74b3c2c212baa75ea27fc7fa03eb
        - registry.connect.redhat.com/zabbix/zabbix-agent-60@sha256:2895b5e45785aaf528d4a6f78c8af0a0d494dbe9362403fdd470673d911d5004
        - quay.io/zabbix/zabbix-operator@sha256:43d26488a19967bdb2a8aafde5f43820be95011811da9cc5ec623c9029801675
        version: 6.0.40
      entries:
      - name: zabbix-operator-certified-rhmp.v6.0.40
        version: 6.0.40
      - name: zabbix-operator-certified-rhmp.v6.0.38
        version: 6.0.38
      - name: zabbix-operator-certified-rhmp.v6.0.36
        version: 6.0.36
      - name: zabbix-operator-certified-rhmp.v6.0.33
        version: 6.0.33
      - name: zabbix-operator-certified-rhmp.v6.0.29
        version: 6.0.29
      - name: zabbix-operator-certified-rhmp.v6.0.27
        version: 6.0.27
      - name: zabbix-operator-certified-rhmp.v6.0.26
        version: 6.0.26
      - name: zabbix-operator-certified-rhmp.v6.0.25
        version: 6.0.25
      - name: zabbix-operator-certified-rhmp.v6.0.17
        version: 6.0.17
      - name: zabbix-operator-certified-rhmp.v6.0.14
        version: 6.0.14
      - name: zabbix-operator-certified-rhmp.v6.0.13
        version: 6.0.13
      name: lts
    defaultChannel: lts
    packageName: zabbix-operator-certified-rhmp
    provider:
      name: Zabbix LLC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: run-once-duration-override-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runoncedurationoverrideoperator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "RunOnceDurationOverride",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "runOnceDurationOverride": {
                    "spec": {
                      "activeDeadlineSeconds": 3600
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certifiedLevel: "false"
          containerImage: registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9-operator@sha256:54993506a3a2fc1e88c9b5aac97ae90d2a04a06ae6e3a97bf487b5e8996b95fb
          createdAt: 2024/07/01
          description: An operator to manage the OpenShift RunOnceDurationOverride
            Mutating Admission Webhook Server
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.1.2'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/run-once-duration-override-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RunOnceDurationOverride is the schema for the runoncedurationoverrides
              API
            displayName: Run Once Duration Override
            kind: RunOnceDurationOverride
            name: runoncedurationoverrides.operator.openshift.io
            version: v1
        displayName: Run Once Duration Override Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mutating
        - webhook
        - workload
        - run-once
        links:
        - name: Source Code
          url: https://github.com/openshift/run-once-duration-override
        - name: Source Code
          url: https://github.com/openshift/run-once-duration-override-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9-operator@sha256:54993506a3a2fc1e88c9b5aac97ae90d2a04a06ae6e3a97bf487b5e8996b95fb
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9@sha256:3dc6e2a5900472e3e083abcb659a13fea3d08f7a9ff167df524b6f890a16dc8e
        version: 1.1.2
      entries:
      - name: runoncedurationoverrideoperator.v1.1.2
        version: 1.1.2
      - name: runoncedurationoverrideoperator.v1.1.1
        version: 1.1.1
      - name: runoncedurationoverrideoperator.v1.1.0
        version: 1.1.0
      - name: runoncedurationoverrideoperator.v1.0.2
        version: 1.0.2
      - name: runoncedurationoverrideoperator.v1.0.1
        version: 1.0.1
      - name: runoncedurationoverrideoperator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: run-once-duration-override-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/streamshub
    name: amq-streams-console
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: alpha
    - currentCSV: amq-streams-console.v2.8.0-13-0.1738265629.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:1b1b80710885b43600e22ff2a24930b5cab06cd91fad439ad8a282c00b4564b9
          createdAt: "2024-11-12T15:34:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-streams-console.v2.8.0-13
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.8
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.23.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:1b1b80710885b43600e22ff2a24930b5cab06cd91fad439ad8a282c00b4564b9
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:38b80d81e54d59667c1f7dfa4da4125d5c7430816da130ddc84678a4e3cff6cf
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:366d60ff48fb9ba32dc73d92a775046d7b1eb1df21c93e31605372c7858f16ca
        version: 2.8.0-13+0.1738265629.p
      entries:
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.8.x
    - currentCSV: amq-streams-console.v2.9.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
          createdAt: "2025-09-22T10:41:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0-0 <2.9.2-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.9
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:839368767bec735cd12f002874917c23f58550fe21df976858bdc92946985ec6
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:b7444b70a5dc4543a9976c08a619ff6978c9ff86ac7085d300b4870b7c6b7851
        version: 2.9.2-2
      entries:
      - name: amq-streams-console.v2.9.2-2
        version: 2.9.2-2
      - name: amq-streams-console.v2.9.2-1
        version: 2.9.2-1
      - name: amq-streams-console.v2.9.1-7
        version: 2.9.1-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.9.x
    - currentCSV: amq-streams-console.v2.9.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
          createdAt: "2025-09-22T10:41:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0-0 <2.9.2-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.9
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:839368767bec735cd12f002874917c23f58550fe21df976858bdc92946985ec6
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:b7444b70a5dc4543a9976c08a619ff6978c9ff86ac7085d300b4870b7c6b7851
        version: 2.9.2-2
      entries:
      - name: amq-streams-console.v2.9.2-2
        version: 2.9.2-2
      - name: amq-streams-console.v2.9.2-1
        version: 2.9.2-1
      - name: amq-streams-console.v2.9.1-7
        version: 2.9.1-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.x
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-3.0.x
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-3.x
    defaultChannel: alpha
    packageName: amq-streams-console
    provider:
      name: Red Hat
      url: https://github.com/streamshub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: 3scale-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: 3scale-operator.v0.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "ProductCRName": "product1-sample",
                  "Production": true
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/3scale-rhel7-operator@sha256:0751f9f5343a1264224c36db3db292e5a9e28c40cd96649ef48b5399c34668dd
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.16"}]'
          olm.skipRange: '>=0.10.0 <0.11.13'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/3scale-rhel7-operator@sha256:0751f9f5343a1264224c36db3db292e5a9e28c40cd96649ef48b5399c34668dd
        - registry.redhat.io/3scale-amp2/backend-rhel8@sha256:cdc53799ff9d022462a9a4afcbcd8043497cd40ccdc6fafdcb9ccf7ce69f43d1
        - registry.redhat.io/3scale-amp2/system-rhel7@sha256:61dcc91e5d3c91a0ea85c02f2599cadcb5496b1c09a8838b4e1c8e4efe126e8f
        - registry.redhat.io/3scale-amp2/memcached-rhel7@sha256:7e2ab1e24646fbfef878eb2b5a74ed865ae812b60aa0ba505045f838f26ebbfc
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:a95d09fc3b224f550dec3de3d23fd2dbfc0a220dc869b4ad9559ee2f85327daa
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:f8cd25658fa6a154e7ea6e5adf0f9dae3fc596ec9989021de6a021ec94eef226
        - registry.redhat.io/rhel8/redis-6@sha256:5250d644e80aea794b0f732e9285883a4348dace83557f177f64d78fe4dcabd0
        - registry.redhat.io/openshift4/ose-cli@sha256:1a8caae40be21f23ba3cb48b75809a73333907033a6041927aa568b5c9290f3c
        - registry.redhat.io/3scale-amp2/searchd-rhel7@sha256:8e2001c021fad554cb8a96bb34e9973010fe4206526f19aa040fa196709fb688
        - registry.redhat.io/3scale-amp2/zync-rhel9@sha256:0d2205d9f61a58414c4ba472ce8f19b12d2df722772847d5fb7442e664759627
        - registry.redhat.io/rhel8/mysql-80@sha256:08547a93659ad804d637b91e8e1eb3464b75783d494e6c40e1c921bd74f31645
        version: 0.11.13
      entries:
      - name: 3scale-operator.v0.11.13
        version: 0.11.13
      - name: 3scale-operator.v0.11.12
        version: 0.11.12
      name: threescale-2.14
    - currentCSV: 3scale-operator.v0.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                },
                "status": {
                  "deployments": {
                    "ready": [
                      "deployment x ready"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                },
                "status": {
                  "adminId": 1,
                  "tenantId": 2
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ApplicationAuth",
                "metadata": {
                  "name": "applicationauth-sample"
                },
                "spec": {
                  "applicationCRName": "application-cr-name",
                  "authSecretRef": {
                    "name": "auth-secret-reference"
                  },
                  "generateSecret": true
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "productCRName": "product1-sample",
                  "production": true
                },
                "status": {}
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/3scale-rhel9-operator@sha256:fcea6a77d0a48f35f4c3ce63e72dde104aac6b580f340950b311eb418fff5ec9
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=0.11.0 <0.12.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          rht_backend_redis_requirements: 6.2.0
          rht_mysql_requirements: 8.0.0
          rht_postgres_requirements: 10.0.0
          rht_system_redis_requirements: 6.2.0
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: ApplicationAuth is the Schema for the applicationauths API
            displayName: Application Auth
            kind: ApplicationAuth
            name: applicationauths.capabilities.3scale.net
            version: v1beta1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/3scale-rhel9-operator@sha256:fcea6a77d0a48f35f4c3ce63e72dde104aac6b580f340950b311eb418fff5ec9
        - registry.redhat.io/3scale-amp2/manticore-rhel9@sha256:ccf6baf5f2e9494a34739a79abe0f46e36d8ab9dc81c426f7a3a8964e8f39859
        - registry.redhat.io/3scale-amp2/backend-rhel8@sha256:eb2288de900dc42ead755b28cdb2509b25ecc8dd256781beaa4a2a86e7cbe532
        - registry.redhat.io/rhel9/memcached@sha256:8738110ef16c9e35deeb07c0577ffe09d6ef029761b504a8d20253fa715da394
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:a95d09fc3b224f550dec3de3d23fd2dbfc0a220dc869b4ad9559ee2f85327daa
        - registry.redhat.io/3scale-amp2/zync-rhel9@sha256:df9d62897ff286165772150a8b5e6952d2ce3b52a8afa07bd693a2d95c2c67ff
        - registry.redhat.io/openshift4/ose-cli@sha256:1a8caae40be21f23ba3cb48b75809a73333907033a6041927aa568b5c9290f3c
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/rhel8/mysql-80@sha256:ee9aca9096473da004f9140305742eb7b9a77ed81ce55d5a6503388cccab030f
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        - registry.redhat.io/3scale-amp2/system-rhel8@sha256:4c72dfeb1919a1de4fbc47f4611d0c830ef89cf9f17255ad55a9b370618e0e8f
        version: 0.12.5
      entries:
      - name: 3scale-operator.v0.12.5
        version: 0.12.5
      - name: 3scale-operator.v0.12.4
        version: 0.12.4
      - name: 3scale-operator.v0.12.3
        version: 0.12.3
      - name: 3scale-operator.v0.12.2
        version: 0.12.2
      - name: 3scale-operator.v0.12.1
        version: 0.12.1
      - name: 3scale-operator.v0.12.0
        version: 0.12.0
      name: threescale-2.15
    - currentCSV: 3scale-operator.v0.12.1-mas
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                },
                "status": {
                  "deployments": {
                    "ready": [
                      "deployment x ready"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                },
                "status": {
                  "adminId": 1,
                  "tenantId": 2
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ApplicationAuth",
                "metadata": {
                  "name": "applicationauth-sample"
                },
                "spec": {
                  "applicationCRName": "application-cr-name",
                  "authSecretRef": {
                    "name": "auth-secret-reference"
                  },
                  "generateSecret": true
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "productCRName": "product1-sample",
                  "production": true
                },
                "status": {}
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-mas/3scale-rhel9-operator@sha256:8f41ab6027616e4bb65b1fddbcf44cc8a25a79cdda19d5cc354ac4228a15141e
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.17"}]'
          olm.skipRange: '>=0.11.8 <0.12.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          rht_backend_redis_requirements: 6.2.0
          rht_mysql_requirements: 8.0.0
          rht_postgres_requirements: 10.0.0
          rht_system_redis_requirements: 6.2.0
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: ApplicationAuth is the Schema for the applicationauths API
            displayName: Application Auth
            kind: ApplicationAuth
            name: applicationauths.capabilities.3scale.net
            version: v1beta1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-mas/3scale-manticore-search-rhel9@sha256:7379186175c9c0f0a6d922cd2a0754b429c9728dfbe8979a9d6025789ebae0da
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:a95d09fc3b224f550dec3de3d23fd2dbfc0a220dc869b4ad9559ee2f85327daa
        - registry.redhat.io/3scale-mas/backend-rhel8@sha256:d0c561c6b9dfe9741722ba4a257fbd2df4d2a709457a0b0657ef7bcca1f2a5ca
        - registry.redhat.io/3scale-mas/zync-rhel9@sha256:78e5223a1a878f6527bd88d382948c125b7cb4a81540648244553af426d1a944
        - registry.redhat.io/3scale-mas/memcached-rhel7@sha256:7a1abee11b4120f399485b438bda1e5262f04de28d3e788b207eecc3a9de9ed2
        - registry.redhat.io/3scale-mas/3scale-rhel9-operator@sha256:8f41ab6027616e4bb65b1fddbcf44cc8a25a79cdda19d5cc354ac4228a15141e
        - registry.redhat.io/rhel8/mysql-80@sha256:754f6830e0bc4aadb973b1f613f2bc7278ec13ad0f0e1e87655a2bb2f76cb9b3
        - registry.redhat.io/openshift4/ose-cli@sha256:a688994ea404ce6ce4bb5ec1249f0e779de0d6c187b5078e2ff067e60b50937d
        - registry.redhat.io/3scale-mas/apicast-gateway-rhel8@sha256:f174636a9ad43a2419fff5a9e6797e3185c540a9509188fd1606aef4acea7bdf
        - registry.redhat.io/rhel8/redis-6@sha256:218f7fa2458c274cc34f79190423e0aba03027f55ef6b3334d916951a668b7b7
        - registry.redhat.io/3scale-mas/system-rhel8@sha256:0fa28aaf3467fbd22fc1335dc6a712043e4dead694c3f14b60e89740bb7eb13e
        version: 0.12.1-mas
      entries:
      - name: 3scale-operator.v0.12.1-mas
        version: 0.12.1-mas
      - name: 3scale-operator.v0.11.8-mas
        version: 0.11.8-mas
      - name: 3scale-operator.v0.11.7-mas
        version: 0.11.7-mas
      - name: 3scale-operator.v0.11.6-mas
        version: 0.11.6-mas
      - name: 3scale-operator.v0.11.5-mas
        version: 0.11.5-mas
      - name: 3scale-operator.v0.11.3-mas
        version: 0.11.3-mas
      - name: 3scale-operator.v0.11.2-mas
        version: 0.11.2-mas
      - name: 3scale-operator.v0.11.1-mas
        version: 0.11.1-mas
      - name: 3scale-operator.v0.11.0-mas
        version: 0.11.0-mas
      - name: 3scale-operator.v0.10.0-mas
        version: 0.10.0-mas
      name: threescale-mas
    defaultChannel: threescale-2.15
    packageName: 3scale-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: LINBIT
      provider-url: https://linbit.com
    name: linstor-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: linstor-operator.v2.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorCluster",
                "metadata": {
                  "name": "linstorcluster"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorNodeConnection",
                "metadata": {
                  "name": "example-connection"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatellite",
                "metadata": {
                  "name": "example-satellite"
                },
                "spec": {
                  "clusterRef": {
                    "name": "linstorcluster"
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatelliteConfiguration",
                "metadata": {
                  "name": "all-satellites"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          com.redhat.openshift.versions: v4.7
          containerImage: drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
          createdAt: "2025-09-25T13:44:22Z"
          description: LINSTOR Operator deploys LINBIT SDS in the cluster.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: linbit-sds
          operators.openshift.io/valid-subscription: '[''LINBIT SDS Subscription'']'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/internal-objects: '["linstorsatellites.piraeus.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster-wide configuration for LINSTOR.
            displayName: Linstor Cluster
            kind: LinstorCluster
            name: linstorclusters.piraeus.io
            version: v1
          - description: Configuration for connections between specific nodes.
            displayName: Linstor Node Connection
            kind: LinstorNodeConnection
            name: linstornodeconnections.piraeus.io
            version: v1
          - description: Configuration for specific Satellites/Nodes in the LINSTOR
              Cluster.
            displayName: Linstor Satellite Configuration
            kind: LinstorSatelliteConfiguration
            name: linstorsatelliteconfigurations.piraeus.io
            version: v1
          - description: Resource to configure a single satellite, managed by the
              Operator itself.
            displayName: Linstor Satellite
            kind: LinstorSatellite
            name: linstorsatellites.piraeus.io
            version: v1
          required:
          - kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
        description: |
          The LINSTOR Operator manages [LINSTOR](https://github.com/LINBIT/linstor-server) clusters in Kubernetes.

          All components of the LINSTOR software stack can be managed by the operator:
          * DRBD
          * LINSTOR
          * LINSTOR CSI driver
          * LINSTOR HA Controller
        displayName: LINSTOR Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: User Guide
          url: https://linbit.com/drbd-user-guide/linstor-guide-1_0-en/
        - name: LINSTOR Operator
          url: https://charts.linstor.io/
        - name: Source Code
          url: https://github.com/LINBIT/linstor-operator-builder
        maintainers:
        - email: support@linbit.com
          name: LINBIT
        maturity: stable
        minKubeVersion: 1.20.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: DaemonSet
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: storage.k8s.io
          kind: CSIDriver
          version: v1
        provider:
          name: LINBIT
          url: https://linbit.com
        relatedImages:
        - drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:f8cde5b5c289f7b3d233773f4e79990ba392fe6e14e0177de80714e7d3e98862
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ac3963bfd4246da386b82d810a348e9d42a9e4b746acaf4dbcbe0766643cc2fa
        - drbd.io/drbd-shutdown-guard@sha256:7ba049ecb205a0086e425d2ff37285b00449549d89e179b4b943ace6e59a9156
        - drbd.io/drbd9-rhel8@sha256:7cddfa770de66cefcb42c2b33da57d0f1adfcc5f1cc12522977aa6993722bb2a
        - drbd.io/linstor-k8s-ha-controller@sha256:ed0f0f86dad88feadf8c3c32d7d143eaa52b1130af3576315a562909b39d2afa
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:2669cde3db2268d37bc671f494eaa06480478a90b8943054af14cbffc050debf
        - drbd.io/drbd-reactor@sha256:77c51aeeefdcd788cb1583738f93778032fcc03c36fb289b692d3cf195c8b184
        - drbd.io/linstor-csi@sha256:9c7ef452b0997d3373803201614c3b8075687e75c397359738a9e4ef5530cab3
        - drbd.io/linstor-controller@sha256:841b3aa84cdf13aaaf1243bc5da8ac2205ca3fe0186fdfc7a04b69397f0020d6
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:624f61b1ae9d32f5a074ee8c12839d975fba9efcd3bfa7981cfe38febf5c4495
        - drbd.io/drbd9-rhel9@sha256:4eb553336ed64e127068a4566370b040166bf11301e21ac47ac5b92237ec9d2a
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:5a650dc21a6a038de6267f117d01adfd904ff3f5c94f428f6a13735f18bf94d6
        - drbd.io/drbd9-rhel7@sha256:0c96b216e55a128a9944276af9080aced8bd72fdf1fd424f024ed9ace2c1feb5
        - drbd.io/ktls-utils@sha256:3351a6e0fc048fe1544499095cdb8fdd221e832cecfa8a9a01f046c271991138
        - drbd.io/linstor-satellite@sha256:ce0a0b05494c6e6d5f16121e5762bf58e7b78f40c00b84780e2abba2a690af1b
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:700be620923fc99ed03f9f1c4a911e4084beb30685361270c49165b27ad1d374
        - drbd.io/drbd9-rhel10@sha256:da0a7015cf15724a7c73100d505891e62262a726141347df7f2d51ec79a0a1e1
        - drbd.io/linstor-affinity-controller@sha256:78e5b299ef1702114528678387b5e74821d7658bc02f05f0bd7edd9a176fba39
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:cd58f1789027e32eae08f00dac7b2275496c41020fb7cdebc10e9986c625d4f4
        version: 2.9.1
      entries:
      - name: linstor-operator.v2.9.1
        version: 2.9.1
      name: dev
    - currentCSV: linstor-operator.v2.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorCluster",
                "metadata": {
                  "name": "linstorcluster"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorNodeConnection",
                "metadata": {
                  "name": "example-connection"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatellite",
                "metadata": {
                  "name": "example-satellite"
                },
                "spec": {
                  "clusterRef": {
                    "name": "linstorcluster"
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatelliteConfiguration",
                "metadata": {
                  "name": "all-satellites"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          com.redhat.openshift.versions: v4.7
          containerImage: drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
          createdAt: "2025-09-25T13:44:22Z"
          description: LINSTOR Operator deploys LINBIT SDS in the cluster.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: linbit-sds
          operators.openshift.io/valid-subscription: '[''LINBIT SDS Subscription'']'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/internal-objects: '["linstorsatellites.piraeus.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster-wide configuration for LINSTOR.
            displayName: Linstor Cluster
            kind: LinstorCluster
            name: linstorclusters.piraeus.io
            version: v1
          - description: Configuration for connections between specific nodes.
            displayName: Linstor Node Connection
            kind: LinstorNodeConnection
            name: linstornodeconnections.piraeus.io
            version: v1
          - description: Configuration for specific Satellites/Nodes in the LINSTOR
              Cluster.
            displayName: Linstor Satellite Configuration
            kind: LinstorSatelliteConfiguration
            name: linstorsatelliteconfigurations.piraeus.io
            version: v1
          - description: Resource to configure a single satellite, managed by the
              Operator itself.
            displayName: Linstor Satellite
            kind: LinstorSatellite
            name: linstorsatellites.piraeus.io
            version: v1
          required:
          - kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
        description: |
          The LINSTOR Operator manages [LINSTOR](https://github.com/LINBIT/linstor-server) clusters in Kubernetes.

          All components of the LINSTOR software stack can be managed by the operator:
          * DRBD
          * LINSTOR
          * LINSTOR CSI driver
          * LINSTOR HA Controller
        displayName: LINSTOR Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: User Guide
          url: https://linbit.com/drbd-user-guide/linstor-guide-1_0-en/
        - name: LINSTOR Operator
          url: https://charts.linstor.io/
        - name: Source Code
          url: https://github.com/LINBIT/linstor-operator-builder
        maintainers:
        - email: support@linbit.com
          name: LINBIT
        maturity: stable
        minKubeVersion: 1.20.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: DaemonSet
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: storage.k8s.io
          kind: CSIDriver
          version: v1
        provider:
          name: LINBIT
          url: https://linbit.com
        relatedImages:
        - drbd.io/drbd9-rhel7@sha256:0c96b216e55a128a9944276af9080aced8bd72fdf1fd424f024ed9ace2c1feb5
        - drbd.io/linstor-controller@sha256:841b3aa84cdf13aaaf1243bc5da8ac2205ca3fe0186fdfc7a04b69397f0020d6
        - drbd.io/drbd9-rhel9@sha256:4eb553336ed64e127068a4566370b040166bf11301e21ac47ac5b92237ec9d2a
        - drbd.io/linstor-k8s-ha-controller@sha256:ed0f0f86dad88feadf8c3c32d7d143eaa52b1130af3576315a562909b39d2afa
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:f8cde5b5c289f7b3d233773f4e79990ba392fe6e14e0177de80714e7d3e98862
        - drbd.io/drbd-shutdown-guard@sha256:7ba049ecb205a0086e425d2ff37285b00449549d89e179b4b943ace6e59a9156
        - drbd.io/linstor-affinity-controller@sha256:78e5b299ef1702114528678387b5e74821d7658bc02f05f0bd7edd9a176fba39
        - drbd.io/linstor-satellite@sha256:ce0a0b05494c6e6d5f16121e5762bf58e7b78f40c00b84780e2abba2a690af1b
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:5a650dc21a6a038de6267f117d01adfd904ff3f5c94f428f6a13735f18bf94d6
        - drbd.io/linstor-csi@sha256:9c7ef452b0997d3373803201614c3b8075687e75c397359738a9e4ef5530cab3
        - drbd.io/ktls-utils@sha256:3351a6e0fc048fe1544499095cdb8fdd221e832cecfa8a9a01f046c271991138
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ac3963bfd4246da386b82d810a348e9d42a9e4b746acaf4dbcbe0766643cc2fa
        - drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:cd58f1789027e32eae08f00dac7b2275496c41020fb7cdebc10e9986c625d4f4
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:624f61b1ae9d32f5a074ee8c12839d975fba9efcd3bfa7981cfe38febf5c4495
        - drbd.io/drbd9-rhel8@sha256:7cddfa770de66cefcb42c2b33da57d0f1adfcc5f1cc12522977aa6993722bb2a
        - drbd.io/drbd-reactor@sha256:77c51aeeefdcd788cb1583738f93778032fcc03c36fb289b692d3cf195c8b184
        - drbd.io/drbd9-rhel10@sha256:da0a7015cf15724a7c73100d505891e62262a726141347df7f2d51ec79a0a1e1
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:700be620923fc99ed03f9f1c4a911e4084beb30685361270c49165b27ad1d374
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:2669cde3db2268d37bc671f494eaa06480478a90b8943054af14cbffc050debf
        version: 2.9.1
      entries:
      - name: linstor-operator.v2.9.1
        version: 2.9.1
      name: stable
    defaultChannel: dev
    packageName: linstor-operator
    provider:
      name: LINBIT
      url: https://linbit.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: https://opdev.github.io/
    name: loki-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.example.com/v1alpha1",
                "kind": "Loki",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/install-disable-crds": "true"
                  },
                  "name": "loki-sample"
                },
                "spec": {
                  "global": {
                    "clusterDomain": "cluster.local",
                    "dnsNamespace": "openshift-dns",
                    "dnsService": "dns-default"
                  },
                  "loki": {
                    "auth_enabled": false,
                    "commonConfig": {
                      "replication_factor": 1
                    },
                    "storage": {
                      "type": "filesystem"
                    }
                  },
                  "monitoring": {
                    "dashboards": {
                      "enabled": false
                    },
                    "lokiCanary": {
                      "enabled": false
                    },
                    "rules": {
                      "alerting": false,
                      "enabled": false
                    },
                    "selfMonitoring": {
                      "enabled": false,
                      "grafanaAgent": {
                        "installOperator": false
                      }
                    },
                    "serviceMonitor": {
                      "enabled": false,
                      "metricsInstance": {
                        "enabled": false
                      }
                    }
                  },
                  "rbac": {
                    "sccEnabled": true
                  },
                  "singleBinary": {
                    "replicas": 1
                  },
                  "test": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/loki-helm-operator-controller:0.0.1
          createdAt: "2023-07-14T15:06:17Z"
          description: An experimental operator that installs loki from its official
            helmchart.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys the loki operator helm chart
            displayName: Loki
            kind: Loki
            name: lokis.charts.example.com
            version: v1alpha1
        description: Loki operator based on the helm chart with single binary mode
          deployment optionality
        displayName: Loki Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loki
        - single-binary
        - demo
        - helm
        links:
        - name: Loki Helm Operator
          url: https://github.com/skattoju/loki-helm-operator
        maintainers:
        - email: skattoju@redhat.com
          name: Sid Kattoju
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: https://opdev.github.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/opdev/loki-helm-operator-controller:0.0.1
        version: 0.0.1
      entries:
      - name: loki-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: loki-helm-operator
    provider:
      name: opdev
      url: https://opdev.github.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat CCIT
      provider-url: ""
    name: reportportal-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: reportportal-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ReportPortal",
                "metadata": {
                  "name": "reportportal-sample"
                },
                "spec": {
                  "accesskey": "AKIAIOSFODNN7EXAMPLE",
                  "analyzer_image": "quay.io/waynesun09/service-auto-analyzer:5.3.5",
                  "api_image": "quay.io/waynesun09/service-api:5.3.5-rootless",
                  "api_replicas": 1,
                  "app_domain": "apps.test-example.com",
                  "bucket_prefix": null,
                  "default_bucket_name": null,
                  "enable_pg_restic_backup": "yes",
                  "es_backup_schedule": "@daily",
                  "es_s3_backup_dir": "s3_backup",
                  "es_snapshot_bucket": "es-snapshot-123123",
                  "gateway_replicas": 1,
                  "index_image": "quay.io/waynesun09/service-index:5.0.10",
                  "index_replicas": 1,
                  "integration_salt_path": null,
                  "migration_image": "quay.io/waynesun09/migrations:5.3.5",
                  "pg_restic_password": "rp_user",
                  "pg_restic_s3_bucket_init": "yes",
                  "pg_s3_bucket_name": "pgbackup-123123",
                  "s3_enabled": false,
                  "s3_endpoint": null,
                  "secretkey": "wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY",
                  "uat_image": "quay.io/waynesun09/service-authorization:5.3.5",
                  "ui_image": "quay.io/waynesun09/service-ui:5.3.5",
                  "ui_replicas": 1
                }
              },
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ReportPortalRestore",
                "metadata": {
                  "name": "reportportalrestore-sample"
                },
                "spec": {
                  "accesskey": "AKIAIOSFODNN7EXAMPLE",
                  "bucket_prefix": null,
                  "es_restore": "yes",
                  "es_restore_date": "2020-08-07",
                  "es_s3_backup_dir": "s3_backup",
                  "es_snapshot_bucket": "es-snapshot-123123",
                  "pg_recovery": "yes",
                  "pg_recovery_target_time": "2020-08-07 00:00:00",
                  "pg_restic_password": "rp_user",
                  "pg_s3_bucket_name": "pgbackup-123123",
                  "s3_endpoint": null,
                  "secretkey": "wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY"
                }
              },
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ServiceApiServiceMonitor",
                "metadata": {
                  "name": "serviceapiservicemonitor-sample"
                },
                "spec": {
                  "bearer_token_secret_key": "token",
                  "bearer_token_secret_name": "rp-user-auth"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/waynesun09/rp5-operator:v0.0.8
          createdAt: "2021-04-22 14:39:00"
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/waynesun09/reportportal-operator
          support: waynesun09
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ReportPortal Restore from s3 backup
            kind: ReportPortalRestore
            name: reportportalrestores.rp5.reportportal.io
            version: v1
          - description: ReportPortal v5 Deployment with ReportPortal Operator
            kind: ReportPortal
            name: reportportals.rp5.reportportal.io
            version: v1
          - description: ReportPortal service-api ServiceMonitor for metrics
            kind: ServiceApiServiceMonitor
            name: serviceapiservicemonitors.rp5.reportportal.io
            version: v1
        description: |-
          ReportPortal is a test automation analytics platform. ReportPortal Operator support instantiating the ReportPortal services and required dependencies.

          What services are included:

              - Report Portal v5.3
              - PostgreSQL
              - Elasticsearch
              - RabbitMQ
              - Minio

          The Operator also support data backup & restore and metrics monitoring.

          ## Install

          Install the Operator to a namespace, then create ReportPortal v5 instance under Provided APIs on the UI, all Report Portal v5 services and dependencies will be installed.

          ## Data backup and restore

          The Report Portal v5 data backup and restore in the operator include two parts, PostgreSQL database data and Elasticsearch index data.

          The Operator support automatically data backup and restore for both.

          ### PostgreSQL database WAL archiving and recovery

          The Operator supports PostgreSQL WAL archiving and recovery point-in-time via Minio S3 using restic.

          #### WAL archiving

          When deploy Report Portal v5 instance, make sure `enable_pg_restic_backup` option is set as `yes`, then the WAL archiving will be enabled when initiate the database, the WAL archive data will be automatically backup to S3 storage via Minio using restic command.

          Parameter `pg_restic_s3_bucket_init` is for init the bucket if not exist, `pg_s3_bucket_name` is the bucket name and `pg_restic_password` is the restic password for encryption.

          #### WAL recovery

          WAL archive recovery will download all WAL archive files to the Postgres container and restore point-in-time.

          To restore PostgreSQL database point-in-time, on Operator UI select `ReportPortalRestore` under Provided APIs to create ReportPortalRestore operand.

          After the ReportPortalRestore operand created, the PG WAL restore will get started, as restore time will base on the database size, the process will last in minutes, hours or days.

          **Note:** Current Postgres container liveless probe timeout is set as one hour in the postgresql statefulset, it might cause restore fail if restore time is longer than one hour, so it might need be adjusted to be longer.

          ### Elasticsearch snapshot and restore

          The operator supports elasticsearch snapshot and restore with data saved in S3 bucket via Minio, the elasticsearch container have installed the s3 repository plugin.

          #### Snapshot

          When deploy Report Portal v5 instance, make sure following elasticsearch parameters are set:

              es_s3_backup_dir: s3_backup
              es_snapshot_bucket: es-snapshot-123123
              es_backup_schedule: '@daily'

          #### Restore

          Elasticsearch restore support restore to certain date as default snapshot schedule is daily.

          To restore elasticsearch from snapshot, on Operator UI select `ReportPortalRestore` under Provided APIs to create ReportPortalRestore operand.

          The backup dir and bucket name should be same as when create the RP instance, `es_restore` need set to `yes` to enable the restore and `es_restore_date` is the restore target date.

          The restore time will also be depend on the snapshot size.

          ## Metrics

          The operator by default have enabled metrics, the default ansible operator port 8080 is used for the operator container.

          All Report Portal v5 depended services, rabbitmq, minio, postgresql and elasticsearch have metrics enabled in this operator.

          The Report Portal gatway service have metrics enabled.

          By default several ServiceMonitors have been defined when create ReportPortal v5 operand:

              $ oc get servicemonitor
              NAME                    AGE
              elasticsearch-metrics   1d21h
              gateway-metrics         1d21h
              minio-metrics           1d21h
              postgres-metrics        1d21h
              rabbitmq-metrics        1d21h

          To export Report Portal v5 service-api metrics with ServiceMonitor, create a `secret` with RP user api token. The api access token could be found with login to Report Portal -> Click user icon -> Profile.

          Then select `ServiceApiServiceMonitor` under provided APIs to Create Instance on operator UI.

          Input the secret name and key name that have been created, and click create, a api-metrics servicemonitor will be created.

          To check the metrics, if [user workload monitoring](https://docs.openshift.com/container-platform/4.5/monitoring/monitoring-your-own-services.html) is enabled on the cluster, click to Monitoring on OpenShift UI and navigate to Metrics, where you could query the metrics.

          You could also deploy Promethues Operator in the current project and query via it.

          ## Doc

          More detail with installation, data backup & restore and metrics could be found at https://github.com/waynesun09/reportportal-operator#reportportal-operator
        displayName: reportportal-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - reportportal
        - failure analysis
        - test automation
        links:
        - name: Documentation
          url: https://github.com/waynesun09/reportportal-operator/blob/master/README.md
        - name: ReportPortal.io
          url: https://reportportal.io
        maintainers:
        - email: gsun@redhat.com
          name: Wayne Sun
        maturity: alpha
        provider:
          name: Red Hat CCIT
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.7.0
        - quay.io/waynesun09/rp5-operator:v0.0.8
        version: 0.0.8
      entries:
      - name: reportportal-operator.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: reportportal-operator
    provider:
      name: Red Hat CCIT
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://www.redhat.com/
    name: analytics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: analytics-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyData",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalydata-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalydata",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalydata-sample"
                },
                "spec": {
                  "anomalyname": "etcd_object_namespaces",
                  "config": {
                    "max": 100,
                    "min": 60,
                    "query": "max(apiserver_storage_objects{resource=~\"namespaces\"}) by (resource)"
                  },
                  "method": "min_max",
                  "metricdata": {
                    "groupeddata": "{ \"resource\": \"namespaces\" }",
                    "latestvalue": 121,
                    "timestamp": 1695804362
                  }
                }
              },
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalyengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalyengine",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalyengine-sample"
                },
                "spec": {
                  "anomalyqueryconfiguration": "etcd_objects_all:\n  method: percentage_change\n  query: max(apiserver_storage_objects{resource!~\"certificatesigningrequests.certificates.k8s.io|installplans.operators.coreos.com|operators.operators.coreos.com|subscriptions.operators.coreos.com|anomalydata.observability-analytics.redhat.com|anomalyengines.observability-analytics.redhat.com\"}) by (resource)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: True\nfailed_pods:\n  method: min_max\n  query: count(kube_pod_status_phase{phase!~\"Succeeded|Pending|Running|Unknown\"}) by (phase, namespace)\n  min: 0\n  max: 10\n  have_multi_result_data: True\nworkload_memory_usage_bytes:\n  method: percentage_change\n  query: sum(container_memory_working_set_bytes{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nworkload_cpu_usage_cores:\n  method: percentage_change\n  query: sum(container_cpu_usage_seconds_total{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nrest_client_requests_total:\n  method: percentage_change\n  query: sum(rest_client_requests_total)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\n",
                  "cronjobconfig": {
                    "anomalyqueries": "etcd_objects_all,failed_pods,workload_memory_usage_bytes,workload_cpu_usage_cores,rest_client_requests_total",
                    "name": "osa-anomaly-detection",
                    "resource": {
                      "cpulimit": "256m",
                      "cpurequest": "128m",
                      "memorylimit": "512Mi",
                      "memoryrequest": "256Mi"
                    },
                    "schedule": "*/5 * * * *"
                  },
                  "namespace": "osa-anomaly-detection",
                  "serviceaccountrolebinding": {
                    "clusterrolename": "cluster-monitoring-view",
                    "name": "osa-in-cluster-thanos-querier"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
          createdAt: "2024-03-12T08:02:25Z"
          description: Deploys and Manages Observability Analytics Operator
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4-alpha
          repository: https://github.com/openshift/analytics-operator
          support: RedHat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnomalyData is the Schema for the anomalydata API
            displayName: Anomaly Data
            kind: AnomalyData
            name: anomalydata.observability-analytics.redhat.com
            version: v1alpha1
          - description: AnomalyEngine is the Schema for the anomalyengines API
            displayName: Anomaly Engine
            kind: AnomalyEngine
            name: anomalyengines.observability-analytics.redhat.com
            version: v1alpha1
        description: "### Introduction\nObservability Analytics for Red Hat OpenShift
          runs OpenShift analytics processes like anomaly detection.\n\n### API Backward
          Compatibility\n**NOTE:** Until the Operator supports **seamless upgrades**,
          a new version of the operator may introduce a change that is **NOT** backward-compatible
          with the previous version. Thus, to upgrade the operator, uninstall the
          already installed version completely (including CRDs) and install the new
          version.\n\n### Installtion Steps for the Anomaly Engine\n Create namespace
          if you would like to install this operator into separate namespace (You
          can utilize existing namespace as well). ```bash kubectl create ns <Namespace>
          ``` Install the Observability Analytics operator to specific `<Namespace>`,
          wait until you see the operator in the **Installed Operator** list. Click
          on the Operator name and go to Operator details.\n\nHere you should see
          a tab called **Anomaly Engine**. Click on it, and then click on the **Create
          AnomalyEngine** button, which will redirect you to a form where you need
          to provide the required inputs for the Anomaly Engine. Once successfully
          created, you should see a cronjob in the given namespace. This cronjob will
          identify anomalies and store them into **anomalydata**.\n\n### How to See
          Anomaly Data Captured by Anomaly Engine?\nWe are storing anomaly data into
          CR.\n ####Option 1:Through `kubectl`` command\nYou can check detected anomalies
          using the following command from your terminal. \n\n```bash\nkubectl get
          anomalydata -n <Namespace>\nkubectl describe anomalydata <NAME of the anomalydata>
          -n <Namespace> \n```\n #### Option 2: From Installed Operator UI Go to installed
          operator and select **Observability Analytics** operator, now go to **Anomaly
          Data** tab where you can able to see the detected anomaly, click on specific
          anomaly and you should be able to see details of the same. ### Documentation\nDocumentation
          around how to test Operator codebase as a developer can be found [here](https://github.com/openshift/analytics-operator#readme)\n\n###
          Contributing\nYou can contribute by:\n  * Raising [issues](https://github.com/openshift/analytics-operator/issues)\n
          \   related to observability-analytics-operator\n  * Fixing issues by opening
          [Pull Requests](https://github.com/openshift/analytics-operator/pulls)\n
          \ * Discussing observability-analytics-operator\n\n### License\nobservability-analytics-operator
          is licensed under the Apache 2.0 license.\n"
        displayName: Observability Analytics
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - observability-analytics-operator
        - analytics-operator
        - analytics
        - anomaly
        links:
        - name: Analytics Operator
          url: https://analytics-operator.domain
        maintainers:
        - email: rzalavad@redhat.com
          name: Raj
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
        version: 0.1.1
      entries:
      - name: analytics-operator.v0.1.1
        version: 0.1.1
      - name: analytics-operator.v0.1.0
        version: 0.1.0
      name: alpha
    - currentCSV: analytics-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyData",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalydata-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalydata",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalydata-sample"
                },
                "spec": {
                  "anomalyname": "etcd_object_namespaces",
                  "config": {
                    "max": 100,
                    "min": 60,
                    "query": "max(apiserver_storage_objects{resource=~\"namespaces\"}) by (resource)"
                  },
                  "method": "min_max",
                  "metricdata": {
                    "groupeddata": "{ \"resource\": \"namespaces\" }",
                    "latestvalue": 121,
                    "timestamp": 1695804362
                  }
                }
              },
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalyengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalyengine",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalyengine-sample"
                },
                "spec": {
                  "anomalyqueryconfiguration": "etcd_objects_all:\n  method: percentage_change\n  query: max(apiserver_storage_objects{resource!~\"certificatesigningrequests.certificates.k8s.io|installplans.operators.coreos.com|operators.operators.coreos.com|subscriptions.operators.coreos.com|anomalydata.observability-analytics.redhat.com|anomalyengines.observability-analytics.redhat.com\"}) by (resource)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: True\nfailed_pods:\n  method: min_max\n  query: count(kube_pod_status_phase{phase!~\"Succeeded|Pending|Running|Unknown\"}) by (phase, namespace)\n  min: 0\n  max: 10\n  have_multi_result_data: True\nworkload_memory_usage_bytes:\n  method: percentage_change\n  query: sum(container_memory_working_set_bytes{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nworkload_cpu_usage_cores:\n  method: percentage_change\n  query: sum(container_cpu_usage_seconds_total{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nrest_client_requests_total:\n  method: percentage_change\n  query: sum(rest_client_requests_total)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\n",
                  "cronjobconfig": {
                    "anomalyqueries": "etcd_objects_all,failed_pods,workload_memory_usage_bytes,workload_cpu_usage_cores,rest_client_requests_total",
                    "name": "osa-anomaly-detection",
                    "resource": {
                      "cpulimit": "256m",
                      "cpurequest": "128m",
                      "memorylimit": "512Mi",
                      "memoryrequest": "256Mi"
                    },
                    "schedule": "*/5 * * * *"
                  },
                  "namespace": "osa-anomaly-detection",
                  "serviceaccountrolebinding": {
                    "clusterrolename": "cluster-monitoring-view",
                    "name": "osa-in-cluster-thanos-querier"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
          createdAt: "2024-03-12T08:02:25Z"
          description: Deploys and Manages Observability Analytics Operator
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4-alpha
          repository: https://github.com/openshift/analytics-operator
          support: RedHat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnomalyData is the Schema for the anomalydata API
            displayName: Anomaly Data
            kind: AnomalyData
            name: anomalydata.observability-analytics.redhat.com
            version: v1alpha1
          - description: AnomalyEngine is the Schema for the anomalyengines API
            displayName: Anomaly Engine
            kind: AnomalyEngine
            name: anomalyengines.observability-analytics.redhat.com
            version: v1alpha1
        description: "### Introduction\nObservability Analytics for Red Hat OpenShift
          runs OpenShift analytics processes like anomaly detection.\n\n### API Backward
          Compatibility\n**NOTE:** Until the Operator supports **seamless upgrades**,
          a new version of the operator may introduce a change that is **NOT** backward-compatible
          with the previous version. Thus, to upgrade the operator, uninstall the
          already installed version completely (including CRDs) and install the new
          version.\n\n### Installtion Steps for the Anomaly Engine\n Create namespace
          if you would like to install this operator into separate namespace (You
          can utilize existing namespace as well). ```bash kubectl create ns <Namespace>
          ``` Install the Observability Analytics operator to specific `<Namespace>`,
          wait until you see the operator in the **Installed Operator** list. Click
          on the Operator name and go to Operator details.\n\nHere you should see
          a tab called **Anomaly Engine**. Click on it, and then click on the **Create
          AnomalyEngine** button, which will redirect you to a form where you need
          to provide the required inputs for the Anomaly Engine. Once successfully
          created, you should see a cronjob in the given namespace. This cronjob will
          identify anomalies and store them into **anomalydata**.\n\n### How to See
          Anomaly Data Captured by Anomaly Engine?\nWe are storing anomaly data into
          CR.\n ####Option 1:Through `kubectl`` command\nYou can check detected anomalies
          using the following command from your terminal. \n\n```bash\nkubectl get
          anomalydata -n <Namespace>\nkubectl describe anomalydata <NAME of the anomalydata>
          -n <Namespace> \n```\n #### Option 2: From Installed Operator UI Go to installed
          operator and select **Observability Analytics** operator, now go to **Anomaly
          Data** tab where you can able to see the detected anomaly, click on specific
          anomaly and you should be able to see details of the same. ### Documentation\nDocumentation
          around how to test Operator codebase as a developer can be found [here](https://github.com/openshift/analytics-operator#readme)\n\n###
          Contributing\nYou can contribute by:\n  * Raising [issues](https://github.com/openshift/analytics-operator/issues)\n
          \   related to observability-analytics-operator\n  * Fixing issues by opening
          [Pull Requests](https://github.com/openshift/analytics-operator/pulls)\n
          \ * Discussing observability-analytics-operator\n\n### License\nobservability-analytics-operator
          is licensed under the Apache 2.0 license.\n"
        displayName: Observability Analytics
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - observability-analytics-operator
        - analytics-operator
        - analytics
        - anomaly
        links:
        - name: Analytics Operator
          url: https://analytics-operator.domain
        maintainers:
        - email: rzalavad@redhat.com
          name: Raj
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
        version: 0.1.1
      entries:
      - name: analytics-operator.v0.1.1
        version: 0.1.1
      - name: analytics-operator.v0.1.0
        version: 0.1.0
      name: dev-preview
    defaultChannel: alpha
    packageName: analytics-operator
    provider:
      name: Red Hat Inc.
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-atlas-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-atlas-kubernetes.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupCompliancePolicy",
                "metadata": {
                  "name": "my-backup-compliance-policy"
                },
                "spec": {
                  "authorizedEmail": "john.doe@example.com",
                  "authorizedUserFirstName": "John",
                  "authorizedUserLastName": "Doe",
                  "copyProtectionEnabled": false,
                  "encryptionAtRestEnabled": false,
                  "onDemandPolicy": {
                    "retentionUnit": "weeks",
                    "retentionValue": 3
                  },
                  "overwriteBackupPolicies": false,
                  "pointInTimeEnabled": true,
                  "restoreWindowDays": 42,
                  "scheduledPolicyItems": [
                    {
                      "frequencyInterval": 2,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "hourly",
                      "retentionUnit": "days",
                      "retentionValue": 2
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "weekly",
                      "retentionUnit": "weeks",
                      "retentionValue": 4
                    },
                    {
                      "frequencyInterval": 15,
                      "frequencyType": "monthly",
                      "retentionUnit": "months",
                      "retentionValue": 12
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": false,
                  "copySettings": [
                    {
                      "cloudProvider": "AWS",
                      "frequencies": [
                        "WEEKLY"
                      ],
                      "regionName": "US_WEST_1",
                      "shouldCopyOplogs": false
                    }
                  ],
                  "policy": {
                    "name": "atlasbackuppolicy-sample"
                  },
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2,
                  "useOrgAndGroupNamesInExportPrefix": true
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasCustomRole",
                "metadata": {
                  "name": "shard-operator-role"
                },
                "spec": {
                  "projectRef": {
                    "name": "my-project"
                  },
                  "role": {
                    "actions": [
                      {
                        "name": "getShardMap",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "shardingState",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "connPoolStats",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "getLog",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      }
                    ],
                    "inheritedRoles": [
                      {
                        "database": "admin",
                        "name": "operator-role-1"
                      }
                    ],
                    "name": "my-role"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDataFederation",
                "metadata": {
                  "name": "my-data-federation",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "name": "my-data-federation",
                  "privateEndpoints": [
                    {
                      "endpointId": "vpce-03f9eeaa764e32454",
                      "provider": "AWS",
                      "type": "DATA_LAKE"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project",
                    "namespace": "mongodb-atlas-system"
                  },
                  "storage": {
                    "databases": [
                      {
                        "collections": [
                          {
                            "dataSources": [
                              {
                                "storeName": "http-test",
                                "urls": [
                                  "https://data.cityofnewyork.us/api/views/vfnx-vebw/rows.csv"
                                ]
                              }
                            ],
                            "name": "test-collection-1"
                          }
                        ],
                        "name": "test-db-1"
                      }
                    ],
                    "stores": [
                      {
                        "name": "http-test",
                        "provider": "http"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "the-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "theuser"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-deployment"
                },
                "spec": {
                  "backupRef": {
                    "name": "atlasbackupschedule-sample"
                  },
                  "deploymentSpec": {
                    "backupEnabled": true,
                    "clusterType": "REPLICASET",
                    "mongoDBMajorVersion": "7.0",
                    "name": "test-deployment",
                    "replicationSpecs": [
                      {
                        "regionConfigs": [
                          {
                            "electableSpecs": {
                              "instanceSize": "M10",
                              "nodeCount": 3
                            },
                            "priority": 7,
                            "providerName": "AWS",
                            "regionName": "US_EAST_1"
                          }
                        ],
                        "zoneName": "Zone 1"
                      }
                    ],
                    "tags": [
                      {
                        "key": "team",
                        "value": "atlas-operator"
                      }
                    ]
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasFederatedAuth",
                "metadata": {
                  "name": "atlas-default-federated-auth",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "connectionSecretRef": {
                    "name": "my-org-secret",
                    "namespace": "mongodb-atlas-system"
                  },
                  "domainAllowList": [
                    "my-org-domain.com"
                  ],
                  "domainRestrictionEnabled": true,
                  "enabled": true,
                  "postAuthRoleGrants": [
                    "ORG_MEMBER"
                  ],
                  "roleMappings": [
                    {
                      "externalGroupName": "org-admin",
                      "roleAssignments": [
                        {
                          "role": "ORG_OWNER"
                        }
                      ]
                    },
                    {
                      "externalGroupName": "dev-team",
                      "roleAssignments": [
                        {
                          "role": "ORG_GROUP_CREATOR"
                        },
                        {
                          "projectName": "dev-project",
                          "role": "GROUP_OWNER"
                        }
                      ]
                    }
                  ],
                  "ssoDebugEnabled": false
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasIPAccessList",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasipaccesslist-sample"
                },
                "spec": {
                  "entries": [
                    {
                      "cidrBlock": "192.168.1.0/24"
                    },
                    {
                      "deleteAfterDate": "2025-01-20T23:59:59+02:00",
                      "ipAddress": "10.0.0.1"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkContainer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasnetworkcontainer-sample"
                },
                "spec": {
                  "cidrBlock": "10.11.0.0/16",
                  "connectionSecret": {
                    "name": "atlas-connection-secret"
                  },
                  "externalProjectRef": {
                    "projectId": "66e2f2b621571b7e69a89b66"
                  },
                  "provider": "AWS",
                  "region": "US_EAST_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkPeering",
                "metadata": {
                  "name": "atlasnetworkpeering-sample"
                },
                "spec": {
                  "atlasCidrBlock": "10.12.0.0/16",
                  "awsConfiguration": {
                    "accepterRegionName": "US_EAST_1",
                    "awsAccountId": "23214235145",
                    "routeTableCidrBlock": "10.11.0.0/16",
                    "vpcId": "vpcid-7123782134"
                  },
                  "containerId": "623412394512350",
                  "containerRegion": "US_EAST_1",
                  "projectRef": {
                    "name": "atlas-project",
                    "namespace": "namespace"
                  },
                  "provider": "AWS"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasPrivateEndpoint",
                "metadata": {
                  "name": "atlasprivateendpoint-sample"
                },
                "spec": {
                  "awsConfiguration": {
                    "id": "vpce-f4k34w51d"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "provider": "AWS",
                  "region": "EU_CENTRAL_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    },
                    {
                      "cidrBlock": "203.0.113.0/24",
                      "comment": "CIDR block for Application Server B - D"
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasSearchIndexConfig",
                "metadata": {
                  "name": "atlassearchindexconfig-sample"
                },
                "spec": {
                  "analyzer": "lucene.standard",
                  "analyzers": [
                    {
                      "name": "test-analyzer",
                      "tokenFilters": [
                        {
                          "asciiFolding": {
                            "originalTokens": "omit"
                          },
                          "type": "asciiFolding"
                        }
                      ],
                      "tokenizer": {
                        "type": "whiteSpace",
                        "whitespace": {
                          "maxTokenLenght": 200
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamConnection",
                "metadata": {
                  "name": "my-streamconnection-kafka"
                },
                "spec": {
                  "kafkaConfig": {
                    "authentication": {
                      "credentials": {
                        "name": "kafka-userpass"
                      },
                      "mechanism": "SCRAM-512"
                    },
                    "bootstrapServers": "kafka.server1:9001,kafka.server2:9002,kafka.server3:9003",
                    "security": {
                      "certificate": {
                        "name": "kafka-cert"
                      },
                      "protocol": "SSL"
                    }
                  },
                  "name": "kafka-config",
                  "type": "Kafka"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamInstance",
                "metadata": {
                  "name": "my-streaminstance-sample"
                },
                "spec": {
                  "clusterConfig": {
                    "provider": "AWS",
                    "region": "VIRGINIA_USA",
                    "tier": "SP30"
                  },
                  "connectionRegistry": [
                    {
                      "name": "my-streamconnection-sample"
                    }
                  ],
                  "name": "my-instance",
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasTeam",
                "metadata": {
                  "name": "atlasteam-sample"
                },
                "spec": {
                  "name": "my-team",
                  "usernames": [
                    "username@mongodb.com"
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasThirdPartyIntegration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasintegration-sample"
                },
                "spec": {
                  "datadog": {
                    "apiKeySecretRef": {
                      "name": "secret-ref"
                    },
                    "region": "string",
                    "sendCollectionLatencyMetrics": "string",
                    "sendDatabaseMetrics": "string"
                  },
                  "projectRef": {
                    "name": "my-atlas-project"
                  },
                  "type": "DATADOG"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/mongodb/mongodb-atlas-kubernetes-operator:2.11.0
          createdAt: "2025-09-25T14:34:34Z"
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/mongodb/mongodb-atlas-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The AtlasBackupCompliancePolicy is a configuration that enforces
              specific backup and retention requirements
            displayName: Atlas Backup Compliance Policy
            kind: AtlasBackupCompliancePolicy
            name: atlasbackupcompliancepolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasCustomRole is the Schema for the AtlasCustomRole API
            displayName: Atlas Custom Role
            kind: AtlasCustomRole
            name: atlascustomroles.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDataFederation is the Schema for the Atlas Data Federation
              API
            displayName: Atlas Data Federation
            kind: AtlasDataFederation
            name: atlasdatafederations.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasdeployments API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasFederatedAuth is the Schema for the Atlasfederatedauth
              API
            displayName: Atlas Federated Auth
            kind: AtlasFederatedAuth
            name: atlasfederatedauths.atlas.mongodb.com
            version: v1
          - description: AtlasIPAccessList is the Schema for the atlasipaccesslists
              API.
            displayName: Atlas IPAccess List
            kind: AtlasIPAccessList
            name: atlasipaccesslists.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkContainer is the Schema for the AtlasNetworkContainer
              API
            displayName: Atlas Network Container
            kind: AtlasNetworkContainer
            name: atlasnetworkcontainers.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkPeering is the Schema for the AtlasNetworkPeering
              API
            displayName: Atlas Network Peering
            kind: AtlasNetworkPeering
            name: atlasnetworkpeerings.atlas.mongodb.com
            version: v1
          - displayName: Atlas Org Settings
            kind: AtlasOrgSettings
            name: atlasorgsettings.atlas.mongodb.com
            version: v1
          - description: "The AtlasPrivateEndpoint custom resource definition (CRD)
              defines a desired [Private Endpoint](https://www.mongodb.com/docs/atlas/security-private-endpoint/#std-label-private-endpoint-overview)
              configuration for an Atlas project. It allows a private connection between
              your cloud provider and Atlas that doesn't send information through
              a public network. \n You can use private endpoints to create a unidirectional
              connection to Atlas clusters from your virtual network."
            displayName: Atlas Private Endpoint
            kind: AtlasPrivateEndpoint
            name: atlasprivateendpoints.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
          - description: Atlas Search Index Config is the Schema for the Atlas Search
              API
            displayName: Atlas Search Index Config
            kind: AtlasSearchIndexConfig
            name: atlassearchindexconfigs.atlas.mongodb.com
            version: v1
          - description: AtlasStreamConnection is the Schema for the atlasstreamconnections
              API
            displayName: Atlas Stream Connection
            kind: AtlasStreamConnection
            name: atlasstreamconnections.atlas.mongodb.com
            version: v1
          - description: AtlasStreamInstance is the Schema for the atlasstreaminstances
              API
            displayName: Atlas Stream Instance
            kind: AtlasStreamInstance
            name: atlasstreaminstances.atlas.mongodb.com
            version: v1
          - description: AtlasTeam is the Schema for the Atlas Teams API
            displayName: Atlas Team
            kind: AtlasTeam
            name: atlasteams.atlas.mongodb.com
            version: v1
          - description: AtlasThirdPartyIntegration is the Schema for the atlas 3rd
              party integrations API.
            displayName: Atlas Third Party Integration
            kind: AtlasThirdPartyIntegration
            name: atlasthirdpartyintegrations.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB
          Atlas — the only multi-cloud document database service that gives you the versatility you need to build sophisticated
          and resilient applications that can adapt to changing customer demands and market trends.

          The full documentation for the Operator can be found [here](https://docs.atlas.mongodb.com/atlas-operator/)

          ## Getting Started

          ### Supported features

          * Create and configure a project, or connect to an existing one.
          * Deploy, manage, scale, and tear down clusters.
          * Support for serverless instances.
          * Create and edit database users.
          * Manage IP Access Lists, network peering and private endpoints.
          * Configure and control Atlas’s fully managed cloud backup.
          * Configure federated authentication for your Atlas organization
          * Integrate Atlas monitoring with Prometheus.

          ... and more.

          To view the list of custom resources and each CR's schema, visit our [reference](https://www.mongodb.com/docs/atlas/operator/stable/custom-resources/)
          documentation. See the [Quickstart](https://www.mongodb.com/docs/atlas/operator/stable/ak8so-quick-start/) to get
          started with Atlas Kubernetes Operator.
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator:2.11.0
        version: 2.11.0
      entries:
      - name: mongodb-atlas-kubernetes.v2.11.0
        version: 2.11.0
      - name: mongodb-atlas-kubernetes.v2.10.0
        version: 2.10.0
      - name: mongodb-atlas-kubernetes.v2.9.1
        version: 2.9.1
      - name: mongodb-atlas-kubernetes.v2.9.0
        version: 2.9.0
      - name: mongodb-atlas-kubernetes.v2.8.2
        version: 2.8.2
      - name: mongodb-atlas-kubernetes.v2.8.1
        version: 2.8.1
      - name: mongodb-atlas-kubernetes.v2.8.0
        version: 2.8.0
      - name: mongodb-atlas-kubernetes.v2.7.1
        version: 2.7.1
      - name: mongodb-atlas-kubernetes.v2.7.0
        version: 2.7.0
      - name: mongodb-atlas-kubernetes.v2.6.1
        version: 2.6.1
      - name: mongodb-atlas-kubernetes.v2.6.0
        version: 2.6.0
      - name: mongodb-atlas-kubernetes.v2.5.0
        version: 2.5.0
      - name: mongodb-atlas-kubernetes.v2.4.1
        version: 2.4.1
      - name: mongodb-atlas-kubernetes.v2.4.0
        version: 2.4.0
      - name: mongodb-atlas-kubernetes.v2.3.1
        version: 2.3.1
      - name: mongodb-atlas-kubernetes.v2.3.0
        version: 2.3.0
      - name: mongodb-atlas-kubernetes.v2.2.2
        version: 2.2.2
      - name: mongodb-atlas-kubernetes.v2.2.1
        version: 2.2.1
      - name: mongodb-atlas-kubernetes.v2.2.0
        version: 2.2.0
      - name: mongodb-atlas-kubernetes.v2.1.0
        version: 2.1.0
      - name: mongodb-atlas-kubernetes.v2.0.1
        version: 2.0.1
      - name: mongodb-atlas-kubernetes.v2.0.0
        version: 2.0.0
      - name: mongodb-atlas-kubernetes.v1.9.1
        version: 1.9.1
      - name: mongodb-atlas-kubernetes.v1.9.0
        version: 1.9.0
      - name: mongodb-atlas-kubernetes.v1.8.2
        version: 1.8.2
      - name: mongodb-atlas-kubernetes.v1.8.1
        version: 1.8.1
      - name: mongodb-atlas-kubernetes.v1.8.0
        version: 1.8.0
      - name: mongodb-atlas-kubernetes.v1.7.3
        version: 1.7.3
      - name: mongodb-atlas-kubernetes.v1.7.2
        version: 1.7.2
      - name: mongodb-atlas-kubernetes.v1.7.1
        version: 1.7.1
      - name: mongodb-atlas-kubernetes.v1.7.0
        version: 1.7.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-atlas-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OpenTelemetry Community
      provider-url: ""
    name: opentelemetry-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opentelemetry-operator.v0.136.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "Instrumentation",
                "metadata": {
                  "name": "instrumentation"
                },
                "spec": {
                  "exporter": {
                    "endpoint": "http://otel-collector-headless:4317"
                  }
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpAMPBridge",
                "metadata": {
                  "name": "opampbridge-sample"
                },
                "spec": {
                  "capabilities": {
                    "AcceptsOpAMPConnectionSettings": true,
                    "AcceptsOtherConnectionSettings": true,
                    "AcceptsRemoteConfig": true,
                    "AcceptsRestartCommand": true,
                    "ReportsEffectiveConfig": true,
                    "ReportsHealth": true,
                    "ReportsOwnLogs": true,
                    "ReportsOwnMetrics": true,
                    "ReportsOwnTraces": true,
                    "ReportsRemoteConfig": true,
                    "ReportsStatus": true
                  },
                  "componentsAllowed": {
                    "exporters": [
                      "debug"
                    ],
                    "processors": [
                      "memory_limiter"
                    ],
                    "receivers": [
                      "otlp"
                    ]
                  },
                  "endpoint": "ws://opamp-server:4320/v1/opamp"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": "receivers:\n  otlp:\n    protocols: \n      grpc: {}\n      http: {}\n\nexporters:\n  debug: {}\n\nservice:\n  pipelines:\n    traces:\n      receivers: [otlp]\n      exporters: [debug]\n"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1beta1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": {
                    "exporters": {
                      "debug": {}
                    },
                    "receivers": {
                      "otlp": {
                        "protocols": {
                          "grpc": {},
                          "http": {}
                        }
                      }
                    },
                    "service": {
                      "pipelines": {
                        "traces": {
                          "exporters": [
                            "debug"
                          ],
                          "receivers": [
                            "otlp"
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          certified: "false"
          containerImage: ghcr.io/open-telemetry/opentelemetry-operator/opentelemetry-operator
          createdAt: "2025-09-29T13:00:19Z"
          description: Provides the OpenTelemetry components, including the Collector
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/open-telemetry/opentelemetry-operator
          support: OpenTelemetry Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instrumentation is the spec for OpenTelemetry instrumentation.
            displayName: OpenTelemetry Instrumentation
            kind: Instrumentation
            name: instrumentations.opentelemetry.io
            version: v1alpha1
          - description: OpAMPBridge is the Schema for the opampbridges API.
            displayName: OpAMP Bridge
            kind: OpAMPBridge
            name: opampbridges.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1beta1
          - description: TargetAllocator is the Schema for the targetallocators API.
            displayName: Target Allocator
            kind: TargetAllocator
            name: targetallocators.opentelemetry.io
            version: v1alpha1
        description: |-
          OpenTelemetry is a collection of tools, APIs, and SDKs. You use it to instrument, generate, collect, and export telemetry data (metrics, logs, and traces) for analysis in order to understand your software's performance and behavior.

          ### Operator features

          * **Sidecar injection** - annotate your pods and let the operator inject a sidecar.
          * **Managed upgrades** - updating the operator will automatically update your OpenTelemetry collectors.
          * **Deployment modes** - your collector can be deployed as sidecar, daemon set, or regular deployment.
          * **Service port management** - the operator detects which ports need to be exposed based on the provided configuration.
        displayName: Community OpenTelemetry Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opentelemetry
        - tracing
        - logging
        - metrics
        - monitoring
        - troubleshooting
        links:
        - name: OpenTelemetry Operator
          url: https://github.com/open-telemetry/opentelemetry-operator
        maintainers:
        - email: jpkroehling@redhat.com
          name: Juraci Paixão Kröhling
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: OpenTelemetry Community
        relatedImages:
        - ghcr.io/open-telemetry/opentelemetry-operator/opentelemetry-operator:0.136.0
        - quay.io/brancz/kube-rbac-proxy:v0.13.1
        version: 0.136.0
      entries:
      - name: opentelemetry-operator.v0.136.0
        version: 0.136.0
      - name: opentelemetry-operator.v0.135.0
        version: 0.135.0
      - name: opentelemetry-operator.v0.134.0
        version: 0.134.0
      - name: opentelemetry-operator.v0.132.0
        version: 0.132.0
      - name: opentelemetry-operator.v0.131.0
        version: 0.131.0
      - name: opentelemetry-operator.v0.129.1
        version: 0.129.1
      - name: opentelemetry-operator.v0.127.0
        version: 0.127.0
      - name: opentelemetry-operator.v0.126.0
        version: 0.126.0
      - name: opentelemetry-operator.v0.125.0
        version: 0.125.0
      - name: opentelemetry-operator.v0.124.0
        version: 0.124.0
      - name: opentelemetry-operator.v0.123.0
        version: 0.123.0
      - name: opentelemetry-operator.v0.122.0
        version: 0.122.0
      - name: opentelemetry-operator.v0.121.0
        version: 0.121.0
      - name: opentelemetry-operator.v0.120.0
        version: 0.120.0
      - name: opentelemetry-operator.v0.119.0
        version: 0.119.0
      - name: opentelemetry-operator.v0.118.0
        version: 0.118.0
      - name: opentelemetry-operator.v0.117.0
        version: 0.117.0
      - name: opentelemetry-operator.v0.116.0
        version: 0.116.0
      - name: opentelemetry-operator.v0.115.0
        version: 0.115.0
      - name: opentelemetry-operator.v0.114.1
        version: 0.114.1
      - name: opentelemetry-operator.v0.112.0
        version: 0.112.0
      - name: opentelemetry-operator.v0.111.0
        version: 0.111.0
      - name: opentelemetry-operator.v0.110.0
        version: 0.110.0
      - name: opentelemetry-operator.v0.109.0
        version: 0.109.0
      - name: opentelemetry-operator.v0.108.0
        version: 0.108.0
      - name: opentelemetry-operator.v0.107.0
        version: 0.107.0
      - name: opentelemetry-operator.v0.106.0
        version: 0.106.0
      - name: opentelemetry-operator.v0.105.0
        version: 0.105.0
      name: alpha
    defaultChannel: alpha
    packageName: opentelemetry-operator
    provider:
      name: OpenTelemetry Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: StreamNative
      provider-url: https://streamnative.io
    name: sn-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc:latest",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway:0.10.0"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator:v0.10.0
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator:v0.10.0
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      - name: sn-operator.v0.9.0
        version: 0.9.0
      - name: sn-operator.v0.8.5
        version: 0.8.5
      - name: sn-operator.v0.7.10
        version: 0.7.10
      name: alpha
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc:latest",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway:0.10.0"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator:v0.10.0
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator:v0.10.0
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      - name: sn-operator.v0.9.0
        version: 0.9.0
      - name: sn-operator.v0.8.5
        version: 0.8.5
      - name: sn-operator.v0.7.10
        version: 0.7.10
      name: beta
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc:latest",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway:0.10.0"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator:v0.10.0
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator:v0.10.0
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      - name: sn-operator.v0.9.0
        version: 0.9.0
      - name: sn-operator.v0.8.5
        version: 0.8.5
      - name: sn-operator.v0.7.10
        version: 0.7.10
      name: stable
    defaultChannel: alpha
    packageName: sn-operator
    provider:
      name: StreamNative
      url: https://streamnative.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: operator-certification-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operator-certification-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "certification.redhat.com/v1alpha1",
                "kind": "OperatorPipeline",
                "metadata": {
                  "name": "operatorpipeline-sample"
                },
                "spec": {
                  "applyCIPipeline": true,
                  "applyHostedPipeline": false,
                  "applyReleasePipeline": false,
                  "gitHubSecretName": "github-api-token",
                  "kubeconfigSecretName": "kubeconfig",
                  "operatorPipelinesRelease": "main",
                  "pyxisSecretName": "pyxis-api-secret"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-06-23T16:02:00Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io.bundle.channel.default.v1: alpha
          operators.operatorframework.io.bundle.channels: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-openshift-ecosystem/operator-certification-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OperatorPipeline is the Schema for the operatorpipelines
              API
            displayName: Operator Pipeline
            kind: OperatorPipeline
            name: operatorpipelines.certification.redhat.com
            version: v1alpha1
        description: |-
          A Kubernetes operator to provision resources for the operator certification pipeline. This operator is installed in all namespaces which can support multi-tenant scenarios. **Note:** This operator should only be used by Red Hat partners attempting to certify their operator(s).


          **Requirements**


          The certification operator requires that you have the following tools installed, functional, and in your path.
          - [Install](https://docs.openshift.com/container-platform/4.8/cli_reference/openshift_cli/getting-started-cli.html#installing-openshift-cli) oc, the OpenShift CLI tool (tested with version 4.7.13)
          - [Install](https://tekton.dev/docs/cli/) tkn, the Tekton CLI tool (tested with version 0.19.1)
          - [Install](https://git-scm.com/downloads) git, the Git CLI tool (tested with 2.32.0)
          - The certification pipeline expects you to have the source files of your Operator bundle


          **Pre - Installation**


          The below steps exist for simplicity and to make the installation clearer. The operator watches all namespaces, so if secrets/configs/etc already exist in another namespace, feel free to use the existing namespace when following the operator installation steps.


          **Create a new namespace where the following secrets will be applied.**
          ```
          oc new-project oco
          ```


          **Add the kubeconfig secret which will be used to deploy the operator under test and run the certification checks.**
          * Open a terminal window
          * Set the `KUBECONFIG` environment variable
          ```
          export KUBECONFIG=/path/to/your/cluster/kubeconfig
          ```
          > *This kubeconfig will be used to deploy the Operator under test and run the certification checks.*
          ```
          oc create secret generic kubeconfig --from-file=kubeconfig=$KUBECONFIG
          ```


          **Configuring steps for submitting the results**
          - Add the github API token to the repo where the PR will be created
          ```
          oc create secret generic github-api-token --from-literal GITHUB_TOKEN=<github token>
          ```
          - Add RedHat Container API access key

            This API access key is specifically related to your unique partner account for Red Hat Connect portal. Instructions to obtain your API key can be found: [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/operator-cert-workflow.md#step-b---get-api-key)
          ```
          oc create secret generic pyxis-api-secret --from-literal pyxis_api_key=< API KEY >
          ```

          - Optional pipeline configurations can be found [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md#optional-configuration)


          **Execute the Pipeline (Development Iterations)**


          A pre-requisite to running a pipeline is that a `workspace-template.yaml` exists in the directory you want to execute the `tkn` commands from.

          To create a workspace-template.yaml
          ```
          cat <<EOF > workspace-template.yaml
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 5Gi
          EOF
          ```

          There are multiple ways to execute the Pipeline which can be found [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md#execute-the-pipeline-development-iterations)
        displayName: Operator Certification Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - certification
        - pipeline
        - openshift
        links:
        - name: Operator Certification CI Pipeline
          url: https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md
        - name: Red Hat OpenShift Pipelines
          url: https://github.com/openshift/tektoncd-operator
        maintainers:
        - email: certification-operator-devel@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/opdev/operator-certification-operator@sha256:7c84772f96650ff757c2552ba421355d54da41a5ed2155f72a14911b8a3f18ff
        version: 1.3.1
      entries:
      - name: operator-certification-operator.v1.3.1
        version: 1.3.1
      name: alpha
    defaultChannel: alpha
    packageName: operator-certification-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Instana
      provider-url: ""
    name: instana-agent-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: instana-agent-operator.v2.1.35
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instana.io/v1",
                "kind": "InstanaAgent",
                "metadata": {
                  "name": "instana-agent",
                  "namespace": "instana-agent"
                },
                "spec": {
                  "agent": {
                    "configuration_yaml": "# You can leave this empty, or use this to configure your instana agent.\n# See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.\n",
                    "endpointHost": "ingress-red-saas.instana.io",
                    "endpointPort": "443",
                    "env": {},
                    "key": "replace-key"
                  },
                  "cluster": {
                    "name": "my-cluster"
                  },
                  "opentelemetry": {
                    "grpc": {
                      "enabled": true
                    },
                    "http": {
                      "enabled": true
                    }
                  },
                  "zone": {
                    "name": "edited-zone"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring,Logging & Tracing,OpenShift Optional
          certified: "false"
          containerImage: icr.io/instana/instana-agent-operator@sha256:fe949ae0ca9cc05ffe3569cd7a6f4fde48f5e3ee64f27e20e0b3102733a6e809
          description: Fully automated Application Performance Monitoring (APM) for
            microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/instana/instana-agent-operator
          support: Instana
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InstanaAgent is the Schema for the agents API
            displayName: Instana Agent
            kind: InstanaAgent
            name: agents.instana.io
            version: v1
          - description: InstanaAgentRemote is the Schema for the agents API
            displayName: Instana Agent Remote
            kind: InstanaAgentRemote
            name: agentsremote.instana.io
            version: v1
        description: |
          # Instana

          Instana is an [APM solution](https://www.instana.com/product-overview/) built for microservices that enables IT Ops to build applications faster and deliver higher quality services by automating monitoring, tracing and root cause analysis. The solution is optimized for [Kubernetes](https://www.instana.com/automatic-kubernetes-monitoring/) and [OpenShift](https://www.instana.com/blog/automatic-root-cause-analysis-for-openshift-applications/).

          ## Instana Agent Operator

          This is the Kubernetes Operator for installing the Instana Agent on Kubernetes or OpenShift.

          ## Prerequisites for OpenShift

          Before the agent will be able to run in OpenShift, you need to perform a couple of extra configuration steps.

          You need to set up a project for the Instana Agent and configure it's permissions.

          The project you create here needs to be the namespace where you create the Instana Agent custom resource that the operator will use to deploy the agent.

          For example, create the `instana-agent` project:

              oc new-project instana-agent

          Then, ensure the `instana-agent` service account is in the privileged security context:

              oc adm policy add-scc-to-user privileged -z instana-agent

          The `instana-agent-remote` service account must be allowed to use any userid:

              oc adm policy add-scc-to-user anyuid -z instana-agent-remote -n instana-agent

          This service account will be created by the operator.

          Now you can proceed with installing the operator for the Instana agent.

          ## Installation and Configuration

          First, install this operator from [OperatorHub.io](https://operatorhub.io/), [OpenShift Container Platform](https://www.openshift.com/), or [OKD](https://www.okd.io/).

          Second, create a custom resource with the agent configuration in the target namespace (for now, this MUST always be the `instana-agent` namespace). The operator will pick up the custom resource and install the Instana agent accordingly.

          The following is a minimal template of the custom resource:

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgent
          metadata:
            name: instana-agent
            namespace: instana-agent
          spec:
            zone:
              name: my-zone # (optional) name of the zone of the host
            cluster:
              name: replace-me # replace with the name of your Kubernetes cluster
            agent:
              key: replace-me # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io # the monitoring ingress endpoint
              endpointPort: "443" # the monitoring ingress endpoint port, wrapped in quotes
              env:
                INSTANA_AGENT_TAGS: example
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://ibm.biz/monitoring-k8s
          ```

          Save the template in a file `instana-agent.yaml` and edit the following values:

          * If your target namespace is not `instana-agent`, replace the `namespace:` accordingly.
          * `agent.key` must be set with your Instana agent key.
          * `agent.endpointHost` must be set with the monitoring ingress endpoint, generally either `saas-us-west-2.instana.io` or `saas-eu-west-1.instana.io`.
          * `agent.endpointPort` must be set with the monitoring ingress port, generally "443" (wrapped in quotes).
          * `zone.name` should be set with the name of the Kubernetes cluster that is be displayed in Instana.

          For advanced configuration, you can edit the contents of the `configuration.yaml` file. View documentation [here](https://ibm.biz/monitoring-k8s).

          Apply the custom resource with `kubectl apply -f instana-agent.yaml`. After some time, you should see `instana-agent` Pods being created on each node of your cluster, and your cluster should show on the infrastructure map on your Instana Web interface.

          In addition to the InstanaAgent custom resource, a remote agent can be deployed. This agent will not monitor workload running inside the cluster, but can deploy a single replica of an agent which can get an independent configuration for running remote monitoring for certain sensors.

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgentRemote
          metadata:
            name: remote-agent
            namespace: "instana-agent"
          spec:
            zone:
              name: edited-zone # name of the zone of the host
            hostname:
              name: use-custom-hostname-for-remote-agent # (optional) give your remote agent pod a custom hostname
            agent:
              key: replace-key # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io
              endpointPort: "443"
              env: {}
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.
          ```

          ## Uninstalling

          In order to uninstall the Instana agent, simply remove the custom resource with `kubectl delete -f instana-agent.yaml`.

          ## Source Code

          The Instana agent operator is an open source project hosted on [https://github.com/instana/instana-agent-operator](https://github.com/instana/instana-agent-operator/).
        displayName: Instana Agent Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - APM
        - Instana
        links:
        - name: GitHub Repository
          url: https://github.com/instana/instana-agent-operator
        maintainers:
        - email: support@instana.com
          name: Instana
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: Instana
        relatedImages:
        - icr.io/instana/instana-agent-operator@sha256:fe949ae0ca9cc05ffe3569cd7a6f4fde48f5e3ee64f27e20e0b3102733a6e809
        - icr.io/instana/agent@sha256:ed71e6c3e5bdb155853a2db8c916460e6a867fb02627e3a332186188cbbd98dd
        version: 2.1.35
      entries:
      - name: instana-agent-operator.v2.1.35
        version: 2.1.35
      - name: instana-agent-operator.v2.1.34
        version: 2.1.34
      - name: instana-agent-operator.v2.1.33
        version: 2.1.33
      - name: instana-agent-operator.v2.1.32
        version: 2.1.32
      - name: instana-agent-operator.v2.1.31
        version: 2.1.31
      - name: instana-agent-operator.v2.1.30
        version: 2.1.30
      - name: instana-agent-operator.v2.1.29
        version: 2.1.29
      - name: instana-agent-operator.v2.1.28
        version: 2.1.28
      - name: instana-agent-operator.v2.1.27
        version: 2.1.27
      - name: instana-agent-operator.v2.1.25
        version: 2.1.25
      - name: instana-agent-operator.v2.1.24
        version: 2.1.24
      - name: instana-agent-operator.v2.1.23
        version: 2.1.23
      - name: instana-agent-operator.v2.1.22
        version: 2.1.22
      - name: instana-agent-operator.v2.1.21
        version: 2.1.21
      - name: instana-agent-operator.v2.1.20
        version: 2.1.20
      - name: instana-agent-operator.v2.1.19
        version: 2.1.19
      - name: instana-agent-operator.v2.1.18
        version: 2.1.18
      - name: instana-agent-operator.v2.1.17
        version: 2.1.17
      - name: instana-agent-operator.v2.1.16
        version: 2.1.16
      - name: instana-agent-operator.v2.1.15
        version: 2.1.15
      - name: instana-agent-operator.v2.1.14
        version: 2.1.14
      - name: instana-agent-operator.v2.1.13
        version: 2.1.13
      - name: instana-agent-operator.v2.1.12
        version: 2.1.12
      - name: instana-agent-operator.v2.1.11
        version: 2.1.11
      - name: instana-agent-operator.v2.1.10
        version: 2.1.10
      - name: instana-agent-operator.v2.1.9
        version: 2.1.9
      - name: instana-agent-operator.v2.1.8
        version: 2.1.8
      - name: instana-agent-operator.v2.1.7
        version: 2.1.7
      - name: instana-agent-operator.v2.1.6
        version: 2.1.6
      - name: instana-agent-operator.v2.1.5
        version: 2.1.5
      - name: instana-agent-operator.v2.1.4
        version: 2.1.4
      - name: instana-agent-operator.v2.1.3
        version: 2.1.3
      - name: instana-agent-operator.v2.1.2
        version: 2.1.2
      - name: instana-agent-operator.v2.1.1
        version: 2.1.1
      - name: instana-agent-operator.v2.1.0
        version: 2.1.0
      - name: instana-agent-operator.v2.0.32
        version: 2.0.32
      - name: instana-agent-operator.v2.0.31
        version: 2.0.31
      - name: instana-agent-operator.v2.0.30
        version: 2.0.30
      - name: instana-agent-operator.v2.0.27
        version: 2.0.27
      - name: instana-agent-operator.v2.0.19
        version: 2.0.19
      - name: instana-agent-operator.v2.0.18
        version: 2.0.18
      - name: instana-agent-operator.v2.0.17
        version: 2.0.17
      - name: instana-agent-operator.v2.0.13
        version: 2.0.13
      - name: instana-agent-operator.v2.0.12
        version: 2.0.12
      - name: instana-agent-operator.v2.0.11
        version: 2.0.11
      - name: instana-agent-operator.v2.0.10
        version: 2.0.10
      - name: instana-agent-operator.v2.0.9
        version: 2.0.9
      - name: instana-agent-operator.v2.0.8
        version: 2.0.8
      - name: instana-agent-operator.v2.0.7
        version: 2.0.7
      - name: instana-agent-operator.v2.0.5
        version: 2.0.5
      - name: instana-agent-operator.v2.0.4
        version: 2.0.4
      name: stable
    defaultChannel: stable
    packageName: instana-agent-operator
    provider:
      name: Instana
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hdfs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hdfs-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:e8b4ad2c3b127537239b003ae19a3f58b541d599fe3d5a74a6718612fe645d50
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:e8b4ad2c3b127537239b003ae19a3f58b541d599fe3d5a74a6718612fe645d50
        version: 24.11.1
      entries:
      - name: hdfs-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: hdfs-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:00c79036eed6dd6749736b964787acb638f1a6e0156c0bba5651b93738534377
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:00c79036eed6dd6749736b964787acb638f1a6e0156c0bba5651b93738534377
        version: 25.3.0
      entries:
      - name: hdfs-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: hdfs-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
        version: 25.7.0
      entries:
      - name: hdfs-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hdfs-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
        version: 25.7.0
      entries:
      - name: hdfs-operator.v25.7.0
        version: 25.7.0
      - name: hdfs-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hdfs-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: seldon
      provider-url: ""
    name: seldon-deploy-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: seldon-deploy-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1alpha1",
                "kind": "SeldonDeploy",
                "metadata": {
                  "name": "seldondeploy-sample"
                },
                "spec": {
                  "affinity": {},
                  "alibidetect": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306"
                  },
                  "batchjobs": {
                    "mc": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48"
                    },
                    "processor": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d"
                    },
                    "pvc": {
                      "defaultSize": "1Gi"
                    },
                    "serviceAccount": "workflow"
                  },
                  "defaultUserID": "",
                  "docker": {
                    "user": "unknown"
                  },
                  "elasticsearch": {
                    "basicAuth": true,
                    "secret": {
                      "name": "elastic-credentials",
                      "passwordKey": "password",
                      "userKey": "username"
                    },
                    "url": "https://elasticsearch-seldon-es-http.seldon-logs:9200"
                  },
                  "enableAppAnalytics": false,
                  "enableAppAuth": false,
                  "env": {
                    "USERID_CLAIM_KEY": "name"
                  },
                  "external": {
                    "protocol": "http"
                  },
                  "fullnameOverride": "",
                  "gitops": {
                    "argocd": {
                      "enabled": false,
                      "namespace": "argocd"
                    },
                    "fileFormat": "json",
                    "git": {
                      "email": "",
                      "secret": "git-creds",
                      "skipVerifyGit": true,
                      "token": "",
                      "user": "",
                      "webhook": {
                        "service": {
                          "create": true,
                          "loadBalancerSourceRanges": {}
                        }
                      }
                    }
                  },
                  "image": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e",
                    "pullPolicy": "Always"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": []
                      }
                    ],
                    "tls": []
                  },
                  "ingressGateway": {
                    "ingressNamespace": "istio-system",
                    "kfServingIngressService": "istio-ingressgateway",
                    "seldonIngressService": "istio-ingressgateway"
                  },
                  "kfserving": {
                    "curlForm": "MODEL_NAME={{ .ModelName }}\u003cbr\u003e\nCLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\nSERVICE_HOSTNAME=$(kubectl get inferenceservice {{ .ModelName }} -o jsonpath='{.status.url}' | cut -d \"/\" -f 3)\u003cbr\u003e\ncurl -v -H \"Host: ${SERVICE_HOSTNAME}\" {{ .KfServingProtocol }}://$CLUSTER_IP/v1/models/$MODEL_NAME:predict -d '{{ .Payload }}'\n",
                    "enabled": false,
                    "explainForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:explain",
                    "protocol": "http",
                    "requestForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:predict"
                  },
                  "loadtest": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c"
                  },
                  "metadata": {
                    "pg": {
                      "enabled": false,
                      "secret": "metadata-postgres"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "openshiftMarketplace": {
                    "cleanupClusterServiceVersions": false,
                    "kubectlCleanupImage": "registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382",
                    "prometheus": {
                      "monitorSpecs": {
                        "create": true
                      }
                    },
                    "seldonCore": {
                      "istioGateway": {
                        "create": true,
                        "name": "seldon-gateway",
                        "namespace": "istio-system"
                      },
                      "subscription": {
                        "apiVersion": "operators.coreos.com/v1alpha1",
                        "channel": "stable",
                        "create": true,
                        "istioEnabled": true,
                        "metricsPath": "/metrics",
                        "requestLoggerEndpoint": "http://broker-ingress.knative-eventing.svc.cluster.local/seldon-logs/default"
                      }
                    }
                  },
                  "prometheus": {
                    "knative": {
                      "url": "http://prometheus-system-np.knative-monitoring.svc.cluster.local:8080/api/v1/"
                    },
                    "seldon": {
                      "jwtSecretKey": "jwt-seldon.txt",
                      "jwtSecretName": "jwt-seldon",
                      "namespaceMetricName": "namespace",
                      "resourceMetricsUrl": "https://prometheus-k8s.openshift-monitoring:9091/api/v1/",
                      "serviceMetricName": "exported_service",
                      "url": "https://thanos-querier.openshift-monitoring.svc:9091/api/v1/"
                    }
                  },
                  "rbac": {
                    "clusterWide": true,
                    "create": true,
                    "readNamespaces": true
                  },
                  "replicaCount": 1,
                  "requestLogger": {
                    "create": true,
                    "elasticsearch": {
                      "host": "elasticsearch-seldon-es-http.seldon-logs",
                      "port": "9200",
                      "protocol": "https"
                    },
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398",
                    "imagePullPolicy": "IfNotPresent",
                    "namespace": {
                      "create": false,
                      "name": "seldon-logs"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "600m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "200Mi"
                      }
                    },
                    "trigger": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "broker": "default"
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "800m",
                      "memory": "800Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "seldon": {
                    "curlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions -d '{{ .Payload }}'\n",
                    "enabled": true,
                    "explainForm": "http://{{ .ModelName }}-{{ .Predictor }}-explainer.{{ .Namespace }}:9000/v1/models/{{ .ModelName }}:explain",
                    "kfservingV2CurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .GraphModelName }}/infer -d '{{ .Payload }}'\n",
                    "protocol": "http",
                    "seldonRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions",
                    "tensorFlowCurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict -d '{{ .Payload }}'\n",
                    "tensorflowRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict",
                    "v2RequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .ModelName }}/infer"
                  },
                  "service": {
                    "port": 80,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "create": true
                  },
                  "serviceAccountName": "seldon-deploy",
                  "skipVerifyHttpCalls": true,
                  "tolerations": [],
                  "virtualService": {
                    "create": true,
                    "gateways": [
                      "istio-system/seldon-gateway"
                    ],
                    "prefix": "/seldon-deploy/"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
          createdAt: "2020-12-11 15:00:00"
          description: The Seldon deploy operator allows the creation of Seldon Deploy
            applications to provide management, monitoring and updating of machine
            learning models from a range of technologies. It includes the ability
            to run explanations as well as outlier, drift and adversarial detection
            against running models. There is a full metrics and logging integration
            via Prometheus and ELK stacks as well as full GitOps operations to allow
            full auditing and recovery of updated models.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SeldonIO/seldon-deploy-operator
          support: Seldon Dev
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: SeldonDeploy
            name: seldondeploys.machinelearning.seldon.io
            version: v1alpha1
        description: The Seldon deploy operator allows the creation of Seldon Deploy
          applications to provide management, monitoring and updating of machine learning
          models from a range of technologies. It includes the ability to run explanations
          as well as outlier, drift and adversarial detection against running models.
          There is a full metrics and logging integration via Prometheus and ELK stacks
          as well as full GitOps operations to allow full auditing and recovery of
          updated models.
        displayName: Seldon Deploy Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - machine learning
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://deploy.seldon.io
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        provider:
          name: seldon
        relatedImages:
        - registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398
        - registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
        - registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c
        version: 1.2.1
      entries:
      - name: seldon-deploy-operator.v1.2.1
        version: 1.2.1
      - name: seldon-deploy-operator.v1.2.0
        version: 1.2.0
      - name: seldon-deploy-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: seldon-deploy-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1alpha1",
                "kind": "SeldonDeploy",
                "metadata": {
                  "name": "seldondeploy-sample"
                },
                "spec": {
                  "affinity": {},
                  "alibidetect": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306"
                  },
                  "batchjobs": {
                    "mc": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48"
                    },
                    "processor": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d"
                    },
                    "pvc": {
                      "defaultSize": "1Gi"
                    },
                    "serviceAccount": "workflow"
                  },
                  "defaultUserID": "",
                  "docker": {
                    "user": "unknown"
                  },
                  "elasticsearch": {
                    "basicAuth": true,
                    "secret": {
                      "name": "elastic-credentials",
                      "passwordKey": "password",
                      "userKey": "username"
                    },
                    "url": "https://elasticsearch-seldon-es-http.seldon-logs:9200"
                  },
                  "enableAppAnalytics": false,
                  "enableAppAuth": false,
                  "env": {
                    "USERID_CLAIM_KEY": "name"
                  },
                  "external": {
                    "protocol": "http"
                  },
                  "fullnameOverride": "",
                  "gitops": {
                    "argocd": {
                      "enabled": false,
                      "namespace": "argocd"
                    },
                    "fileFormat": "json",
                    "git": {
                      "email": "",
                      "secret": "git-creds",
                      "skipVerifyGit": true,
                      "token": "",
                      "user": "",
                      "webhook": {
                        "service": {
                          "create": true,
                          "loadBalancerSourceRanges": {}
                        }
                      }
                    }
                  },
                  "image": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e",
                    "pullPolicy": "Always"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": []
                      }
                    ],
                    "tls": []
                  },
                  "ingressGateway": {
                    "ingressNamespace": "istio-system",
                    "kfServingIngressService": "istio-ingressgateway",
                    "seldonIngressService": "istio-ingressgateway"
                  },
                  "kfserving": {
                    "curlForm": "MODEL_NAME={{ .ModelName }}\u003cbr\u003e\nCLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\nSERVICE_HOSTNAME=$(kubectl get inferenceservice {{ .ModelName }} -o jsonpath='{.status.url}' | cut -d \"/\" -f 3)\u003cbr\u003e\ncurl -v -H \"Host: ${SERVICE_HOSTNAME}\" {{ .KfServingProtocol }}://$CLUSTER_IP/v1/models/$MODEL_NAME:predict -d '{{ .Payload }}'\n",
                    "enabled": false,
                    "explainForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:explain",
                    "protocol": "http",
                    "requestForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:predict"
                  },
                  "loadtest": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c"
                  },
                  "metadata": {
                    "pg": {
                      "enabled": false,
                      "secret": "metadata-postgres"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "openshiftMarketplace": {
                    "cleanupClusterServiceVersions": false,
                    "kubectlCleanupImage": "registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382",
                    "prometheus": {
                      "monitorSpecs": {
                        "create": true
                      }
                    },
                    "seldonCore": {
                      "istioGateway": {
                        "create": true,
                        "name": "seldon-gateway",
                        "namespace": "istio-system"
                      },
                      "subscription": {
                        "apiVersion": "operators.coreos.com/v1alpha1",
                        "channel": "stable",
                        "create": true,
                        "istioEnabled": true,
                        "metricsPath": "/metrics",
                        "requestLoggerEndpoint": "http://broker-ingress.knative-eventing.svc.cluster.local/seldon-logs/default"
                      }
                    }
                  },
                  "prometheus": {
                    "knative": {
                      "url": "http://prometheus-system-np.knative-monitoring.svc.cluster.local:8080/api/v1/"
                    },
                    "seldon": {
                      "jwtSecretKey": "jwt-seldon.txt",
                      "jwtSecretName": "jwt-seldon",
                      "namespaceMetricName": "namespace",
                      "resourceMetricsUrl": "https://prometheus-k8s.openshift-monitoring:9091/api/v1/",
                      "serviceMetricName": "exported_service",
                      "url": "https://thanos-querier.openshift-monitoring.svc:9091/api/v1/"
                    }
                  },
                  "rbac": {
                    "clusterWide": true,
                    "create": true,
                    "readNamespaces": true
                  },
                  "replicaCount": 1,
                  "requestLogger": {
                    "create": true,
                    "elasticsearch": {
                      "host": "elasticsearch-seldon-es-http.seldon-logs",
                      "port": "9200",
                      "protocol": "https"
                    },
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398",
                    "imagePullPolicy": "IfNotPresent",
                    "namespace": {
                      "create": false,
                      "name": "seldon-logs"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "600m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "200Mi"
                      }
                    },
                    "trigger": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "broker": "default"
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "800m",
                      "memory": "800Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "seldon": {
                    "curlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions -d '{{ .Payload }}'\n",
                    "enabled": true,
                    "explainForm": "http://{{ .ModelName }}-{{ .Predictor }}-explainer.{{ .Namespace }}:9000/v1/models/{{ .ModelName }}:explain",
                    "kfservingV2CurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .GraphModelName }}/infer -d '{{ .Payload }}'\n",
                    "protocol": "http",
                    "seldonRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions",
                    "tensorFlowCurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict -d '{{ .Payload }}'\n",
                    "tensorflowRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict",
                    "v2RequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .ModelName }}/infer"
                  },
                  "service": {
                    "port": 80,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "create": true
                  },
                  "serviceAccountName": "seldon-deploy",
                  "skipVerifyHttpCalls": true,
                  "tolerations": [],
                  "virtualService": {
                    "create": true,
                    "gateways": [
                      "istio-system/seldon-gateway"
                    ],
                    "prefix": "/seldon-deploy/"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
          createdAt: "2020-12-11 15:00:00"
          description: The Seldon deploy operator allows the creation of Seldon Deploy
            applications to provide management, monitoring and updating of machine
            learning models from a range of technologies. It includes the ability
            to run explanations as well as outlier, drift and adversarial detection
            against running models. There is a full metrics and logging integration
            via Prometheus and ELK stacks as well as full GitOps operations to allow
            full auditing and recovery of updated models.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SeldonIO/seldon-deploy-operator
          support: Seldon Dev
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: SeldonDeploy
            name: seldondeploys.machinelearning.seldon.io
            version: v1alpha1
        description: The Seldon deploy operator allows the creation of Seldon Deploy
          applications to provide management, monitoring and updating of machine learning
          models from a range of technologies. It includes the ability to run explanations
          as well as outlier, drift and adversarial detection against running models.
          There is a full metrics and logging integration via Prometheus and ELK stacks
          as well as full GitOps operations to allow full auditing and recovery of
          updated models.
        displayName: Seldon Deploy Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - machine learning
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://deploy.seldon.io
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        provider:
          name: seldon
        relatedImages:
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d
        - registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48
        - registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c
        - registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e
        version: 1.2.1
      entries:
      - name: seldon-deploy-operator.v1.2.1
        version: 1.2.1
      - name: seldon-deploy-operator.v1.2.0
        version: 1.2.0
      - name: seldon-deploy-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: seldon-deploy-operator-rhmp
    provider:
      name: seldon
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: rhods-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhods-operator.2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
          createdAt: "2025-08-12T18:04:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "llamastackoperator": {
                      "managementState": "Removed"
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (including PyTorch, TensorFlow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack – unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.23
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fbe10282ada2fa6bf07ae97b497f46ebe735de130b4ce0182bf34301429999cb
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:2d2e7803b2b1e15017ca04961f9dcd25e89c47b07b101ca54cb9e793e6027202
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:e50adae13e9d773e7be8bf81b7525dac3f5e3864c9cdf4c287342bd568ee1d6b
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:bb7c352ccc10a8c60ef8d573368d0ada59e73cd131653d11bcebdf9f564acc10
        - registry.redhat.io/rhel9/mariadb-105@sha256:050dd5a7a32395b73b8680570e967e55050b152727412fdd73a25d8816e62d53
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:ccfc84028a95f1cc6a16e06ff6e9ed2ace74f74e5e9963a4fb33544c2ce005c2
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:2886e40f94fde72eb5959ad564bbdd466657208c09bac731b3860b0b54f104c6
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:b038d20def504c140e4365f2eb117a979534629332bec75a2ba750ebafdefe3d
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5fb7e6c3ee1e3b170cd146ea9cd172fa953326b08163897605ca74a3644f2819
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:795ab1842ef675e3d8b47deb8a7485a69064428e823dfeb930a6a7e7017cde0f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:80d284cac2ef6638fe028102b71a70f26041576c11a49414e77430954a3c1b99
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f23f8051886d0f0240937507471c818f440bca8703b8209a836d5dbe3ea2f7c8
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:16ff53f4f1a1e1e66826c8355c654038da83df545427a931485ec7a2d3764b16
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:23c098f7e791441cfcfa6bf37de51763511bd8bf4d543747b7d37ed591d68886
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:edcce7cddfa4c250108532e56541ddb95099612c180b2297e5934bc89d2bb46e
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:ff740fc3615edaa90dae976da1aa6455d64811e239eb66b422836a3aae791b41
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:240692b86bc8a4fe0b545e6e4cd7cf4ff0ca990406bc55f0446d565fd67e37f6
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:a4649e8ad4970bc3d5e967811c9694fa82d78ee9f8f025b8b2fbd37237eb14fc
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:68e74bdf518fc624cc25a59a18b2a87d0aa2a9b8aad30750ace7115d297a21ca
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:ca21e218e26c46e3c63d926241846f8f307fd4a586cc4b04147da49af6018ef5
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:857560240bb45bfc2db6d7743abf11a47f929a0d3035940256a6527afd313412
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:84c3213e11299ec032aa0e5a4dfa9b3055c4246b04b029e888d8d9f1caf5c30c
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:4847ac1be55284815670c2048c1edfc56c7227494280d3c6094aa23f1ed1a9d2
        - registry.redhat.io/ubi9/toolbox@sha256:4213516b6fe7957a01072720efd9ecbc7d4443238e151b5fe0263bd0c36792b1
        - registry.redhat.io/rhoai/odh-llama-stack-core-rhel9@sha256:43b60b1ee6f66fec38fe2ffbbe08dca8541ef162332e4bd8e422ecd24ee02646
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:467b04f4ee4118bbe0934d1ab8b80f72df8703dcdac61c2b53fdb5bd1b404c14
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:a1497e64349634b2567d28f685052de118f64f6b32e7d3ea469212e2004d7eb3
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:e0c9ac73891781ef421e0a020cfd5c962f7632d38020897183cf048b1819326a
        - registry.redhat.io/rhoai/odh-llama-stack-k8s-operator-rhel9@sha256:626b9ef09ee82c8d425ddb7ed3f2ffbb3d1019f541e31604db891787c881d5c9
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:3243caa4c8aac75069d19225dcee361bf9c012782328a6266b7016b3026f6d56
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:756b7bdaa7f6e7c4a18488e690188dd91153a33913d225bab001b0085382e514
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:da8debeb2b84a42aba0b98785ccc20e02ed520e125428fe012df3ca473cbb755
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:208005da389b7d4176ad8bffc703312011a1299954df151e2abbe8f3dec01d19
        version: 2.23.0
      entries:
      - name: rhods-operator.2.23.0
        version: 2.23.0
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.20.0
        version: 2.20.0
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.18.0
        version: 2.18.0
      - name: rhods-operator.2.17.0
        version: 2.17.0
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.15.0
        version: 2.15.0
      - name: rhods-operator.2.14.0
        version: 2.14.0
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: alpha
    - currentCSV: rhods-operator.2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhods/odh-rhel8-operator@sha256:a4d7f020509b9720063e87e1c2f19af04e4c9b4986b199be4dc617831b971434
          createdAt: "2023-11-21T08:37:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            Data Science
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:08c90a8f133b4cdb1477a742bc44be4d8fc4e88153b8bc2a2a148cbdf19f4e7d
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:6a56010de933f172b195a1a575855d37b70a4968be8edb35157f6ca193969ad2
          olm.skipRange: '>=1.33.0 <2.4.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "rhods",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "rhods",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Data Science"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:eb96f4b753606528750b3c28f7f9256582f81fa030450b05afdef684634903be
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift Data Science
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift Data Science (RHODS) is a complete platform for the entire lifecycle of your AI/ML projects. It is the flagship product of OpenShift AI.

          When using RHODS, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. RHODS can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * RHODS dashboard
          * Curated Notebook Images (incl CUDA, PyTorch, Tensorflow)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh (and Kserve) with a provided OpenVino Model Serving Runtime
          * Ability to use other runtimes for serving
          * Model Monitoring
        displayName: Red Hat OpenShift Data Science
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift Data Science
        - RHODS
        - rhods
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift Data Science
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-data-science
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhods/odh-ml-pipelines-api-server-rhel8@sha256:be57f237df156474a709667fa4fb4aa142118e3459963fed56648c1cb2e38cd6
        - registry.redhat.io/rhods/odh-kuberay-operator-controller-rhel8@sha256:bfc805aa1689a4bd6f7e617acf9eb1c2c307624dd49956e9e7442afac328151a
        - registry.redhat.io/rhods/odh-data-science-pipelines-operator-controller-rhel8@sha256:250db47cb873207cfb75c7e8a2f0de08b75b056a72db73362d2d79c08cb7b185
        - registry.redhat.io/rhods/odh-trustyai-service-operator-rhel8@sha256:68471555d0b2bb8150ebfae8d4f3d1dbdff9c5b961ff2305ef23ab7514cbbea1
        - registry.redhat.io/rhods/odh-codeflare-operator-rhel8@sha256:2841fd3051a50a1b0d9e3caafa20a8084b5c472d49393b372828960dbf156674
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhods/odh-model-controller-rhel8@sha256:fde989817cd1a6ec8eb408a7f6c73e28ba043ab2d88035586b1b7e55fb3405cb
        - registry.redhat.io/rhods/odh-modelmesh-runtime-adapter-rhel8@sha256:08478a1647d0413783bcbcd966a73b444b580464de46efedb39504420f22053e
        - registry.redhat.io/rhel8/mariadb-103@sha256:08c90a8f133b4cdb1477a742bc44be4d8fc4e88153b8bc2a2a148cbdf19f4e7d
        - registry.redhat.io/rhods/odh-trustyai-service-rhel8@sha256:fe17d6e07ebaa088f274dd8b5e83892dacc11cb336065b06582a3c055095dba5
        - registry.redhat.io/rhods/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:c5758102609821a6fdabec0277e03c4c728f82176093d1ade0570aba5f3cd8f9
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhods/odh-ml-pipelines-persistenceagent-rhel8@sha256:857a8eb444375505ccb34a11cce69c1794ceabbfb00df06933d7f12cd187319b
        - registry.redhat.io/ubi8/ubi-micro@sha256:6a56010de933f172b195a1a575855d37b70a4968be8edb35157f6ca193969ad2
        - registry.redhat.io/rhods/odh-rhel8-operator@sha256:a4d7f020509b9720063e87e1c2f19af04e4c9b4986b199be4dc617831b971434
        - registry.redhat.io/rhods/odh-mm-rest-proxy-rhel8@sha256:8d989f54009484d7f43fe79d7c37757418fc3c15cd4d7e80a2076677a9645bdf
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhods/odh-ml-pipelines-cache-rhel8@sha256:33aae32e3d348d382906f6e926281852f54006a94d7bdac188bb33debadec059
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhods/odh-modelmesh-rhel8@sha256:3827221f84d3aa22440d6b364b990a4ec0666dfb5963f174f7ed518b3650d223
        - registry.redhat.io/rhods/odh-kf-notebook-controller-rhel8@sha256:cd0ab8d70559be6074fbaf1caa1d50f8eb75e1bff1a22d6d69380f19190df0b6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:eb96f4b753606528750b3c28f7f9256582f81fa030450b05afdef684634903be
        - registry.redhat.io/rhods/odh-ml-pipelines-artifact-manager-rhel8@sha256:76c67d7015a2db6965219db6533e0cddb5b7cf699c88df3ef81155f9c192a3cc
        - registry.redhat.io/rhods/odh-modelmesh-serving-controller-rhel8@sha256:d94395fb251e45aaa73d09bc6106a2947c16b90d87fc3b64bfe4f2f600be5841
        - registry.redhat.io/rhods/odh-dashboard-rhel8@sha256:b98171c5543815abf7c095fba92796ceed9ea44bd8b78db143d2815c6bcfdcf6
        - registry.redhat.io/rhods/odh-notebook-controller-rhel8@sha256:753994fce63802f1e1121990a5c1dd722216446abee54cc1a7c50ab195134414
        version: 2.4.0
      entries:
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: beta
    - currentCSV: rhods-operator.2.21.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:7d0b8d69b6defb5f232da066573b0067ec810a8ccfb2f3cb91d472d2979ab7a7
          createdAt: "2025-09-24T14:42:33Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.21
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:687c8eeed55f021ecaab1307f0e88b5b16d91f72d63b3d7168d7bbee90e8947b
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:04859c878ffcc8c42550a98be625ac52212f9b79a6fd5b3baeafb3592926cbe9
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:2a16c353fd13dc1f3f5c4266f1b9786a8e7a100dc05f2d554bdaef082271063f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:239f226328e5ac77d6dee498217d55103a6d562f4829e38abf65ba55fda0aca2
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:7da004e158006dcbdcf0c3b78c40411b82c9d56ebf54b1c30c245b6550b8ee7d
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:2b4136569e10ad7d47c82bedf55c1705a833dc67c8f833656996bfe1fddbef3e
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:52e1ba8ebf4d55c64c94c9076185c5612b537e6e5ed4944fcd6ea87ccb9cd182
        - registry.redhat.io/ubi9/toolbox@sha256:7656f815a0d13ed37f56d11635559a87a3d00840253755c71b644b858f67d556
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:7d0b8d69b6defb5f232da066573b0067ec810a8ccfb2f3cb91d472d2979ab7a7
        - registry.redhat.io/rhel9/mariadb-105@sha256:c390b12f5c2c3ea7cd897b59eba7127ce22f474130cd8c1bbbbaadae870baca4
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:1b5452b87b36de2b62febc777efdc6233a168ffcc09e2067e482cc444463a57d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:cb47e165935f0d9d3c88088689c37e68a69519357a20d3c52ab6f71240dab507
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:003fdc6cd1821d620e1ce4050ee608e2d2a32f2b08c8fa9c147e2e14418a8d91
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:80fd53a580ce7bc4d02023d38a04912e84c2cd50679fc5ea034afba935623157
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:fe6593bc73e77d315d82b7921bb1975c19fda1e7630ef5b30b69f5901ce30178
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:e68492a6a75cb83be0a968b832c1478b97f14e661376328085fd5416a08cbc31
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:c3e283c0536266195dd461042ca0036ed98cc33650d806be132db4fba45239e4
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:152b425725228953f43b57a4a034a6725993a5ca556af328ae5e06aa3f1acbfc
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:a88696cff8cd2cc2e295c9ec082b7f25f61dfc9ac12f1c208529ff7036d5af29
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fc98ed817a4e9ab84e2a66063987304f0cdbe6e5620a58a7f25cccd61b34ebf9
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:5c0d56f434298f6e8bf6cf7061cd1ae6df9915225cbf04b010b6b59a64775cc9
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:cfaf4eb86ddb9c717340ab17fa4b83323e5b40977ac70265376c87329a3f6f20
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:4c57dcf4df09fe0d03578cc8b9a3ca2ea485b6af0ca0203bfb85e640058b7daf
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:4c192208c618b49f31cf31c9de1a33c622ce3f120204c2eec8ff7c439ea6785c
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:a936d87646843edea2a84eb429bb44eb730bf3e04e8900fecef66721426dba89
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:9c2d6ce05775ac34ee5695b3367d12bf2a9c14e318a9544d00a56653ea5f2007
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:dc0db686b7688f98b5f8595661022cdffa5d26a078289cc7aed528f0ee1d08a4
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:6a388ee3cbe3813ac955d319f0ab65a81da321d982c9881c6c81fe135311fe1d
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5f581957e4f8deebf6104a394272e6fbb1c54d813bf069e95875d865eac209ad
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:2638e3a90bbb242dd2483fefe9fc8b472f334c510d0139b9b7c875c30dab7984
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:85202cb75250a1ba2691d65b52b306fd77566660dd29cdf56bf8962e8bb4939c
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:1b7befcd29c9c8f783bdd3c5654b0412ea45cd88cd18eb5f8f5a411f1323c6bd
        version: 2.21.1
      entries:
      - name: rhods-operator.2.21.1
        version: 2.21.1
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: embedded
    - currentCSV: rhods-operator.2.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
          createdAt: "2025-09-24T13:27:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.16
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:22557a6fa52d2f311750a9ba253860f423ba697d26efa02ef8524a8258d2a909
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:d7d4fa406e0fcf0507894a7676532b27f45be742467e603a86f98ea5d2615df8
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:2ba2b2c4db8bb334c50f4dfb54059f060361186900a44c06eed00b7a3c43977e
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:7b70af8847b0806d8b43c399e2b3109f016fd864f5e9d30c44e2baca5d1359dc
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a0f72ffefb2b74b488dd949493f5d295a39bb9c97f578bf219d0138601f65468
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:6cf74044ae8d5308a2dfe03fa5d81086c89302db7f4cdbe2f4174a1c48b77869
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:de5f91180ead7d73a1825fe8b032fde9b8b01392569c9789f5ce1b4b9c08a98f
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:ee01e89f98feb185f6cd59c564e590a13e4d8d9ea760cca8de51426eb71b83a1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:04f305c44413ae7dcb7017e53570ee49a509701792c5f50efadd64f47395730b
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:c8456d98e90b6505957ab3686e9fd2f156e29f123c5558e581c206daf1e7d93a
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:f738aab1eab25854c93e1d8d4d98100a8ae7bb45a6b83f0326774e4220b1183b
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:5dcdcc2424602a69451f16d31dbfa1d43cb72c095ba561eb9076f0cd1e8182ed
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:efd039012559786deb3c521a130886c265e88d635e08baace2e573a9df270134
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:27b113d96453c2054d4c965a963d2badba6daac235eddaf23234c76e87bcb069
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:4c5ff3496b2a2a739939d94ee9dafc02b682100785d228dc2fde480fb597b7a5
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:4a7599f8a866eb169c9a62885906adbf6df0417c0be15857df1eef20cd9b1be2
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:633984b538f027c93c5886ea1045dcbe81eeda74acf80001d5fc5f765bdbe0be
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:23d307a36b69e0df04f72a7d3b35e28d8417a8bbe23dba31e8e977569785c078
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:11a301728573adf2b64ea72d0cb2d83ea5d4dbebea759f346e99f18c3d368c6e
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:5192f1dbd2a9ab92ae390c4ae506efbed0970545b6122e95b014728ac937e777
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:2eedd1e2465d2e78fafb39e234dbb413a82713f187c64d70a80340ec94807d30
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f37e4048f3a152798286793f3abfc6ed814453fcbe2667255a7e78eee483c5a3
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:97e2bd9b587f08e135a9aeb9b3e0dc6eafa1a9bdacbb5ecb681ce9bd5aa37fc9
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:65482864055021272a18b5b26792ab00cfa5fc9cc005d8d3a884cc82991506f3
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:4a212ec634225c14beac09be24ddf336e562f2aa9a13555fb1196f366ddae23c
        version: 2.16.3
      entries:
      - name: rhods-operator.2.16.3
        version: 2.16.3
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.8.5
        version: 2.8.5
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: eus-2.16
    - currentCSV: rhods-operator.2.8.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:3b9e8a71ba795a2ea91f4129253ccb4be3a79c7c1c6a87082f96ccc67eabbbac
          createdAt: "2024-11-07T07:17:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.8.0 <2.8.5'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:eeb8bdc94e7fa2b454fedf72eb75756052f8bfb9d92b116a8267a5584e2528c1
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:f9c474d454d0b006e06859601ec9a64b5e5c91e9a47bfab6ad6d8cce5aa6b89d
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:45498198fadfbc1457bf2a6084ae7ea515d454d86d148783aae3abdaca6e999d
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:60507a3dddbd88560fd85204699147facd2b6371fb25e0bf6701dad02c0b4dd4
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:df34e9d2d121a3def2dbb86f108b3fb6948bc7ba7cb44c70801dfd2e81d3fbb1
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:13dff890ec67fa56864c244a40fc404e9eac05f33573dd16d088b37f9340ec41
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:28ec01e514c15cae145499840597d1be1b750123bb6014de2ecc3ef0d60fa134
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:9aa5a93831992588766ac8594ae3c4dc7111d08a90fdbc8751c40484339e46d4
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:038ce76571217d6c624d614652e8b96ef6484562cf52ec1be88392b8ef8a3bbd
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:1a1074263ae26e4c8b0e3863228c35e50133169f377ece6c47b58a12e0fcf453
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:81ae2b95f2da740cb5264b9ba8b85f0f58bb4ab80d6431f66d52848ed0c21afb
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:1dbf09cb1624c9f961f8ac8e371c97aa7c3fbfcad4c4d02ab4fad0eb86d2aa95
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:61a506fbf263cd72951ac8bd151127e0c74fc74f0a05f7aaecaa54fde47ee490
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:3b9e8a71ba795a2ea91f4129253ccb4be3a79c7c1c6a87082f96ccc67eabbbac
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:13b850a5a3f8776c6c8db887c666d88df72901895b5ef1be9a8bd372cc02fcb8
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:8ffc5da798e783c65d5be7f344219b2951d8c2233d6494a4ae2bb5c86eccd699
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:d87e4b1cba72eb4f50b3f2d59863799929c22ca37b01a1ff4828a9ce0be31f25
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:2f6db789ceb3f90f9c9911c6392e62d0a8e4d61be68a7f7cbd781f47faab6389
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:4b1b9f05a94713292a7a2f6782fe1d3eea6f36214c36b38f402aa91a554ab18a
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:5c48474563d98df32e7243e33302428ea8836be5e1a2e30d20b4d796a62cbf6b
        version: 2.8.5
      entries:
      - name: rhods-operator.2.8.5
        version: 2.8.5
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: eus-2.8
    - currentCSV: rhods-operator.2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
          createdAt: "2025-08-12T18:04:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "llamastackoperator": {
                      "managementState": "Removed"
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (including PyTorch, TensorFlow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack – unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.23
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:edcce7cddfa4c250108532e56541ddb95099612c180b2297e5934bc89d2bb46e
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-llama-stack-k8s-operator-rhel9@sha256:626b9ef09ee82c8d425ddb7ed3f2ffbb3d1019f541e31604db891787c881d5c9
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:ff740fc3615edaa90dae976da1aa6455d64811e239eb66b422836a3aae791b41
        - registry.redhat.io/ubi9/toolbox@sha256:4213516b6fe7957a01072720efd9ecbc7d4443238e151b5fe0263bd0c36792b1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhel9/mariadb-105@sha256:050dd5a7a32395b73b8680570e967e55050b152727412fdd73a25d8816e62d53
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:68e74bdf518fc624cc25a59a18b2a87d0aa2a9b8aad30750ace7115d297a21ca
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:a1497e64349634b2567d28f685052de118f64f6b32e7d3ea469212e2004d7eb3
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:e50adae13e9d773e7be8bf81b7525dac3f5e3864c9cdf4c287342bd568ee1d6b
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:84c3213e11299ec032aa0e5a4dfa9b3055c4246b04b029e888d8d9f1caf5c30c
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:ccfc84028a95f1cc6a16e06ff6e9ed2ace74f74e5e9963a4fb33544c2ce005c2
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:bb7c352ccc10a8c60ef8d573368d0ada59e73cd131653d11bcebdf9f564acc10
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:2d2e7803b2b1e15017ca04961f9dcd25e89c47b07b101ca54cb9e793e6027202
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:b038d20def504c140e4365f2eb117a979534629332bec75a2ba750ebafdefe3d
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:3243caa4c8aac75069d19225dcee361bf9c012782328a6266b7016b3026f6d56
        - registry.redhat.io/rhoai/odh-llama-stack-core-rhel9@sha256:43b60b1ee6f66fec38fe2ffbbe08dca8541ef162332e4bd8e422ecd24ee02646
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:23c098f7e791441cfcfa6bf37de51763511bd8bf4d543747b7d37ed591d68886
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5fb7e6c3ee1e3b170cd146ea9cd172fa953326b08163897605ca74a3644f2819
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:857560240bb45bfc2db6d7743abf11a47f929a0d3035940256a6527afd313412
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:795ab1842ef675e3d8b47deb8a7485a69064428e823dfeb930a6a7e7017cde0f
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f23f8051886d0f0240937507471c818f440bca8703b8209a836d5dbe3ea2f7c8
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fbe10282ada2fa6bf07ae97b497f46ebe735de130b4ce0182bf34301429999cb
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:80d284cac2ef6638fe028102b71a70f26041576c11a49414e77430954a3c1b99
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:467b04f4ee4118bbe0934d1ab8b80f72df8703dcdac61c2b53fdb5bd1b404c14
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:ca21e218e26c46e3c63d926241846f8f307fd4a586cc4b04147da49af6018ef5
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:e0c9ac73891781ef421e0a020cfd5c962f7632d38020897183cf048b1819326a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:a4649e8ad4970bc3d5e967811c9694fa82d78ee9f8f025b8b2fbd37237eb14fc
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:208005da389b7d4176ad8bffc703312011a1299954df151e2abbe8f3dec01d19
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:240692b86bc8a4fe0b545e6e4cd7cf4ff0ca990406bc55f0446d565fd67e37f6
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:756b7bdaa7f6e7c4a18488e690188dd91153a33913d225bab001b0085382e514
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:2886e40f94fde72eb5959ad564bbdd466657208c09bac731b3860b0b54f104c6
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:16ff53f4f1a1e1e66826c8355c654038da83df545427a931485ec7a2d3764b16
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:4847ac1be55284815670c2048c1edfc56c7227494280d3c6094aa23f1ed1a9d2
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:da8debeb2b84a42aba0b98785ccc20e02ed520e125428fe012df3ca473cbb755
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        version: 2.23.0
      entries:
      - name: rhods-operator.2.23.0
        version: 2.23.0
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.20.0
        version: 2.20.0
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.18.0
        version: 2.18.0
      - name: rhods-operator.2.17.0
        version: 2.17.0
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.15.0
        version: 2.15.0
      - name: rhods-operator.2.14.0
        version: 2.14.0
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: fast
    - currentCSV: rhods-operator.2.22.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
          createdAt: "2025-09-25T11:28:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.22
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:7b3d7787ba95156de9c284ad78f2d63a0d3bcb525f4d8d11ebf0203d4c5c1fe0
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:7fd540586a1ca111370100a473923b3dc6cec4255e9e4f3654bfbabd9158a2bb
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:a44d32a18b1e906c1411cafc82102881573fcd04e040a4e2ba6bd1bd617b70fa
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:6234e15f001122ab222481321153fb0eba7f4d86bf3dfd64edf3f6cecbe7d62f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:abe1f9dd11c7313fdd55b678be511a5cdc3546654a68e2b53d42c1013fec8d00
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:b2ca229bb4c4dbf04b96766b8af8d9f55f2edef59ec4110de0d20f8c857e7d27
        - registry.redhat.io/rhel9/mariadb-105@sha256:5e77b33ce04c1eafc2e8840586ad6a5b51148dea8a46dda151071e28858331f0
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:2f3f305c97e50057781553416fb4026cde306ead0fbe0213627b9b5bbcc8a0c3
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:e6fbe6723b4801f6854303b3fd7d4a4a36727bdd28dc28002b7de4929739ccc0
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:2f1a0a0803b2f5118b167e9d9c2e86660add5f79ee6f47044f63b0e8abe0b75f
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:cb5be3dd7098f4c8d66f9e85ad04656460df6102652deddc6cccb884a12be733
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:9104cbc2d595933f0a2acb7313e41ed69905e76a9fae6bd9820e7717b56a632b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e3b90530c6e38f6940d7f9408e9fa2c81eb06a857ec2b0bab5d006b1e9a63a4f
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:32c9511846077ef6035eb19be7653f6a78e0f207eb420ab93b18f742ab5779b6
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:27415b1c763cf89b79391f0b039936c3e078a078bda4420f13bf10ef9be31d5e
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:463e9be9c4409ba1e5a42713bd65844a023e7a3e9032d4e862edbaa5fe35300a
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:1709fa3c79aad4ba7eb9be8299949396092c8e20210124e0c0936385bc04e839
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:b2c2ea9d35edbcd16e1ae559c88a453129b877fc049ab7a7cd6c2ebb8f0bcb59
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:bd6d1599b4ff9eb890be9808014715b28ca19247222d65f40df024db4dd3525a
        - registry.redhat.io/ubi9/toolbox@sha256:ea4ce8cb8f59cf1b0709f1606d7cab074c3806d4016964eedc0555410cb5dfeb
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:d59ed40de694bc0e0564bab159a5b586cffa0721b6d761fb2d81dd42e6225d1c
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:74d4e0174655cf6f6273857325a1e714a1531c5a06ad38bf139bee9c7d8f5f19
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:5d79f0bcca0a213b53aee6d88834e503d343821596b857b13d3a6362de8ebeab
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:5914f71e37c815ce2567826896e051303bdef52fae1acf8d02e672fc24c7eb9d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:2065b96bd6f0a830eab66e36217a927448cfc5a7dabf2a5e7d907aa6c40657a4
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:2b3aace0d80474470e71c2c8291b25738e096109229fe6d5d3e06e9885ee58b6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:13f3679ab289eea9de54fe129a7c83f6c8ec5d26b61b430637a7c4cfcd31a74a
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:ccba67a6a5aa3403e0018f29dcdfef483220dea8065d8b572e80c653ca992add
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:2fe9473cd1d6c445d2ca7cc357ee4bb009aa1838f9cb69ea418296b2a044816b
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:d7c4ac937add511460a36375d9b6b3c475a3a9a7cc96c00df2fbad6219aee0e9
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f9965177de24f3165ec4c001c14cf1e2c2c01ffe847839af837c442e3c96beae
        version: 2.22.2
      entries:
      - name: rhods-operator.2.22.2
        version: 2.22.2
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable
    - currentCSV: rhods-operator.2.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:8643f7ee8afacfe6ae08cbbf81da76ca97032dc3ce3da4fddf16e330e9e33a08
          createdAt: "2024-11-14T21:18:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.10.0 <2.10.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:a1d590d0e2b3fae9ecbd843a664dab083061e797f66a9f0a021b0c2c487d7e45
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:0af6a91553f40ea40e86d85edc204658af3e5a45e8ab94cdb6b767e545df59b9
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:214ca0608fcbc0fd585bf01768d28c7ba17082d1043349ec8445afd663095383
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:d2514bf4b4b2d9abc19facac3f1a55405eb3ad0d9e9203e8816bfcfa08c9a6c5
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:4876f14ca9cff0e2b998a56f2f56d0775edf3963de610f4b810b4d94fb12d719
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:24238abd46633709feb75f2a3e41cf6c9e8528d2c0e4be2fac6f56de44e3fb9b
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:d9b55e8f25b81050380b4aab040243dc18a7432c48912c9b660c97dcc5eea008
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:8643f7ee8afacfe6ae08cbbf81da76ca97032dc3ce3da4fddf16e330e9e33a08
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:9f7b1cb8623de8b0550b0ba124d3b50f5e960ed011b33a3262c6346dbe2dff95
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:807ec05e9e107afc969e718d0c2c980959e2fd462e777a5ff63e55156ea150a4
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:fdc2bc06b0dfa79ebc6451ac3793b5e8d9f165171633831dafe46ebc18c10c35
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:4da77afafde99b15fd357a4ab6609644bf6330a1d9c0c03e64726ed1b9633230
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:cba9f25ac6bb2734bb990a622b515c838b881edb688bf44ff196ebe7684ab74b
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:8911afbb65c8c4f10c3597624485be54b9103e87a5020b718b35ec3df2c9bf00
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:d709f65e1194bc877fa4f9ec681b598467703d94139d57b58c24490d3a59fb82
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:220cad7a22d40998063c6b4fb2f63e69330e64ca7118490a6680b05fc99b7733
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:8fed8f58732bff512289fff738d0a94be67286c41b24eaf5928bbcb30d88c88a
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:232ebac25084db056fdde49477b5c1e6a03a0228fea3758bcf6234b3d0eaa582
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:2007da5abf840712cd2edb50462b2d8d571a0193c3ae1550d4a279eddf2b335f
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:a4b9553fde02d9c67a9c98f063eaf82fb1d2689f4b1269e57bf5f12c8d4adecf
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:5d15175f934017c04d6a6057200ff55b2172a1f4bcbeb0f2cd9e9e545631d93f
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:f4d7d3ba4a0d3322bc9f327e095e9ec56073758f6dcd8da8c7bb185cf952ead5
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:7df2f8bc2c5c0f4f347d009b62a058b5107413a2abd789f6010eb4d6c9e4f8ee
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:a9c6e3afdf605465766fa93334cc0f6bb5e46a5fdfd7ec64d648e2ed72b05ccf
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:267d6795bf53f9a1955d73445776a428dd042af0cd3cb8bb5b91692a1bc1e5d2
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:65db5e8cb2fa46d020ed4837d4479d133b52167c373e65f81cb590eec5391af2
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:0ff5664c2eaadf3d3073b3d7fb711da47fb78d21f67710d8b8aad5452cb0355d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:12741379ab24d2286b5fae77f4383fb65713af7b3356fb43d8254ea85d4b0ee3
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:52e8deeeeaa21b17370a5d5cd644745da745b6e1cee8c6295d6df75c8d2aec04
        version: 2.10.2
      entries:
      - name: rhods-operator.2.10.2
        version: 2.10.2
      - name: rhods-operator.2.10.1
        version: 2.10.1
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.10
    - currentCSV: rhods-operator.2.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:28b5fff602ad6fb36d462e1905f453509ec22fc251a1e0cf64fedba86ff420d7
          createdAt: "2024-10-17T10:53:18Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.13.0 <2.13.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:cd6b535d95eb1b44fe261a5336270fd898b966f431c14e173944484381c26d25
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:c2bd493e9345eccd124bbe0440e45cce8cf57588827d8eb9533f8bbadaa3ebbe
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:74e6cfa0bc481588cb8dc61c206ad567fbd2b28f43c2a86bee4a8e0eff55447c
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:28b5fff602ad6fb36d462e1905f453509ec22fc251a1e0cf64fedba86ff420d7
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:77bca19180f30ef08f4d1598ac6c365e190bb7909ac4a1fb35d9ba9ba8d6ff7a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a37529eff3656a570d796083404e5a3f63603afecfe9e738da1a243c0cb3d7c2
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:5d6354e65106460ab9285b42124faf288298387a8b16317565bd591f009c39a6
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:fb630d6f5817512733b8ea05f7147121928917a24ac8d98239f43d920880c359
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:6b2c0d1dd40e2c276684e17818b14375bc778e6c88df0bdc3cc24bfc9294804f
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:8883bf71abf39b4a6985940c287eb6ac7dc137d491264805d48bb112b65bd008
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:9ccdd7fe8223057a2234a3a2eb3b492e7b69e6ce714cc26a805fe2dd82935195
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:850901511bde065cf73e4c18d787dc10b6e91ca6139fc3ba68da28c41dd28236
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:688389195aebf3f43e74e40d591d2fd5f9ae41994d8ac2db661f292f6eda012a
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:2566b73e98b783871fb500bb1a8cc4e3d58a513c1bb18abfb525da0818e6138b
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:fd4ce524acc1769475079c7231f5ea79ac29551c513855ef34e5694d2c9fbbd3
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:a55f73d72ad2b07f01c49949cdb00969bfe4789375b3d7cdf9814b969fcc808d
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:7d085942e4da73c2ff4885c62f10b904dc70fd08e819e62f925eb1b4ddfa1554
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:8faab013568ee023968a5ec5b3a79977d7ddae3deafff6328e6fd56a27198e24
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:71abac4fb3d613edef7f3ba117a73bbeea61f341d6182e8310360a76576c789e
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:030c53690fe25e15bc9907f51c3f04fab39b4d085ff2098d26316ba32cc8b09f
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:8c19c7fa535f36bdce3b73cf3300e321909d44954cbc63a8aef6321c5f2fa693
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:59e9cc9cd57ba48ca7e87ae1e797a902f826bb890cb198ccf549f76ea581ba6e
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:bf25fce160dd7b8cce11e68a501006e4fa81a7858645038cdefdbba03df7ecdb
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:a3048eaa8924ce3571c6e78014438183fae70d9e8c833a7c07774c675104e3d4
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:5c2ee4136ee0014d1ef395b5e3bcab125c846f5e258ece3631d912f431c8d208
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:7c895490d49beb6bd724f47ad1b17ce501650557b0889314799c4072632f2b99
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f5b17f44c5ab72d6cf658f72f24f15e06a5ac9be1790336f7228ab8e7cfc15c3
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:cf3779c64bbaef1988886ed629c516e7d6ef7bdc73b9c95ad4c087b62e75750d
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:363a35c332fc80b068157951d5e50e6b7209947649cc64803ffc02726e7581ab
        version: 2.13.1
      entries:
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.13
    - currentCSV: rhods-operator.2.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
          createdAt: "2025-09-24T13:27:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.16
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:97e2bd9b587f08e135a9aeb9b3e0dc6eafa1a9bdacbb5ecb681ce9bd5aa37fc9
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:efd039012559786deb3c521a130886c265e88d635e08baace2e573a9df270134
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:d7d4fa406e0fcf0507894a7676532b27f45be742467e603a86f98ea5d2615df8
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:7b70af8847b0806d8b43c399e2b3109f016fd864f5e9d30c44e2baca5d1359dc
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f37e4048f3a152798286793f3abfc6ed814453fcbe2667255a7e78eee483c5a3
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:4a7599f8a866eb169c9a62885906adbf6df0417c0be15857df1eef20cd9b1be2
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:4c5ff3496b2a2a739939d94ee9dafc02b682100785d228dc2fde480fb597b7a5
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:4a212ec634225c14beac09be24ddf336e562f2aa9a13555fb1196f366ddae23c
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:2ba2b2c4db8bb334c50f4dfb54059f060361186900a44c06eed00b7a3c43977e
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:f738aab1eab25854c93e1d8d4d98100a8ae7bb45a6b83f0326774e4220b1183b
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:23d307a36b69e0df04f72a7d3b35e28d8417a8bbe23dba31e8e977569785c078
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:04f305c44413ae7dcb7017e53570ee49a509701792c5f50efadd64f47395730b
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:2eedd1e2465d2e78fafb39e234dbb413a82713f187c64d70a80340ec94807d30
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:5dcdcc2424602a69451f16d31dbfa1d43cb72c095ba561eb9076f0cd1e8182ed
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:11a301728573adf2b64ea72d0cb2d83ea5d4dbebea759f346e99f18c3d368c6e
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:27b113d96453c2054d4c965a963d2badba6daac235eddaf23234c76e87bcb069
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:c8456d98e90b6505957ab3686e9fd2f156e29f123c5558e581c206daf1e7d93a
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:de5f91180ead7d73a1825fe8b032fde9b8b01392569c9789f5ce1b4b9c08a98f
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a0f72ffefb2b74b488dd949493f5d295a39bb9c97f578bf219d0138601f65468
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:6cf74044ae8d5308a2dfe03fa5d81086c89302db7f4cdbe2f4174a1c48b77869
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:633984b538f027c93c5886ea1045dcbe81eeda74acf80001d5fc5f765bdbe0be
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:5192f1dbd2a9ab92ae390c4ae506efbed0970545b6122e95b014728ac937e777
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:65482864055021272a18b5b26792ab00cfa5fc9cc005d8d3a884cc82991506f3
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:22557a6fa52d2f311750a9ba253860f423ba697d26efa02ef8524a8258d2a909
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:ee01e89f98feb185f6cd59c564e590a13e4d8d9ea760cca8de51426eb71b83a1
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        version: 2.16.3
      entries:
      - name: rhods-operator.2.16.3
        version: 2.16.3
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.16
    - currentCSV: rhods-operator.2.19.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:c605e5fdb3d66302781146347df6da5a91dee97e4f8f5d9ac33f105b8cdee383
          createdAt: "2025-09-24T13:27:42Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:b30d60cd458133430d4c92bf84911e03cecd02f60e88a58d1c6c003543cf833a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:f0ee0d27bb784e289f7d88cc8ee0e085ca70e88a5d126562105542f259a1ac01
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:8ce44de8c683f198bf24ba36cd17e89708153d11f5b42c0a27e77f8fdb233551
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.19
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:8a6732ac3ca60f13dcb1fd8f186d2fc5dc601b9b5981a58f039af9e237f20044
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:b20ee16b776871d7c49b23f2bc91726812998561129f5dbd14eddbfe57964c93
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:c605e5fdb3d66302781146347df6da5a91dee97e4f8f5d9ac33f105b8cdee383
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:b30d60cd458133430d4c92bf84911e03cecd02f60e88a58d1c6c003543cf833a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:f49483df1d342d8e39f9a8085920aad79313a475df02a2292a136ecb784402b5
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:c8c23c07038f59db848dea0d4247b62840dbe58e1c581dc0722fb9d7f9be3b75
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:4e8f6c6ca0ed516b8310af0332637683b5bfc8f2946eb935c647698dffbbcc79
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhel8/mariadb-103@sha256:f0ee0d27bb784e289f7d88cc8ee0e085ca70e88a5d126562105542f259a1ac01
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:d24e277f468b0a48f38801e8e0bcfeb8264c3f7c128073ddaa17934b8572cc3c
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:520ed21b80e5487b3a0a13b7e5c7d791d4dd28e0b6799d57abb65d53fca5a03f
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:6275a19ef06114b33d79833a94ef10cc8203368eee98e5d415e73d20cae1db85
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:87685c5cab105d876ff0d14b0e24d8dc2a1ff5b55bebe36028c5f3376b2a5271
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:6ad35cf5c4c62ed5ad4e192afe8cebead9796cb24a964401c76ff8656c47742f
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:b6039c3c33a5dc345ceb8e757437bb13023d2bf85cc2b4c10b9ab3887e78e0c0
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:6f01ac7cc62ab961b9c8a630a558338a61f6d7902301c4a1ad337495ac7ce85a
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:4313cae03c041d461bc276abc23d706b582b5d16feab316796ca43478b9dc705
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:b24c25dfdcefb4ec26e7f311c00798df108f7685def2b1762b7624fdfb160ebf
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:bbef4839f24facce111d21aca43898dcf4e6cf57adb16200e3f1590bd5f89eae
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:49c02285fcef6949892cecca20eeee49d03a9bb4d17bb9ceff7c7a0dbbb8ab40
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel8@sha256:6166f525948a018c41d735014d97fdfa52024f7be31e18976d85d1a0d88901a0
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:baefbe4e7dd5aa74c71cb81da4a7a2b888a02d3eb1398ec4419bd8003b77f0c2
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:8ce44de8c683f198bf24ba36cd17e89708153d11f5b42c0a27e77f8fdb233551
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:af4721d61dfb61c75488a1bd6a5f62b23e20d7b7586a9cf51ba062ece2837715
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:bf6231bfdbe52b597d86e03e4730ec105ebe7a639fcd529268b7c0423f2a890c
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:71cd15f0b1acba3a48ebedff037bf8a846b426c18363a4de8896a376b6b7b3d5
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:53ac36baa374159b9065c718a9ede821bbb61d9ebe9502b2243e0a9f7aca0d16
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:d47514cbceb3cc96e9d33b7db6fc1df60893e5c4118bf6086cdd27f04c472773
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:876ca7ce7b0b975688fb4a77d6cd1a6f87cf9eb40c2a9a5b84d8cffdc80eeebd
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:de0e4b7fb822101198817e83cd7e9f4f7406e75ec91d07a833a27a417cffe66b
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:c1c97b6beb759d3f6f2dd53c4258e4ccbb3bf35cf11873bd446b135defaa69d6
        version: 2.19.3
      entries:
      - name: rhods-operator.2.19.3
        version: 2.19.3
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.19
    - currentCSV: rhods-operator.2.22.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
          createdAt: "2025-09-25T11:28:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.22
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:13f3679ab289eea9de54fe129a7c83f6c8ec5d26b61b430637a7c4cfcd31a74a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:d59ed40de694bc0e0564bab159a5b586cffa0721b6d761fb2d81dd42e6225d1c
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:d7c4ac937add511460a36375d9b6b3c475a3a9a7cc96c00df2fbad6219aee0e9
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:b2ca229bb4c4dbf04b96766b8af8d9f55f2edef59ec4110de0d20f8c857e7d27
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:9104cbc2d595933f0a2acb7313e41ed69905e76a9fae6bd9820e7717b56a632b
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f9965177de24f3165ec4c001c14cf1e2c2c01ffe847839af837c442e3c96beae
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:2f3f305c97e50057781553416fb4026cde306ead0fbe0213627b9b5bbcc8a0c3
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:b2c2ea9d35edbcd16e1ae559c88a453129b877fc049ab7a7cd6c2ebb8f0bcb59
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:6234e15f001122ab222481321153fb0eba7f4d86bf3dfd64edf3f6cecbe7d62f
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:5d79f0bcca0a213b53aee6d88834e503d343821596b857b13d3a6362de8ebeab
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:a44d32a18b1e906c1411cafc82102881573fcd04e040a4e2ba6bd1bd617b70fa
        - registry.redhat.io/rhel9/mariadb-105@sha256:5e77b33ce04c1eafc2e8840586ad6a5b51148dea8a46dda151071e28858331f0
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:7fd540586a1ca111370100a473923b3dc6cec4255e9e4f3654bfbabd9158a2bb
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:2fe9473cd1d6c445d2ca7cc357ee4bb009aa1838f9cb69ea418296b2a044816b
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:27415b1c763cf89b79391f0b039936c3e078a078bda4420f13bf10ef9be31d5e
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:7b3d7787ba95156de9c284ad78f2d63a0d3bcb525f4d8d11ebf0203d4c5c1fe0
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:2065b96bd6f0a830eab66e36217a927448cfc5a7dabf2a5e7d907aa6c40657a4
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e3b90530c6e38f6940d7f9408e9fa2c81eb06a857ec2b0bab5d006b1e9a63a4f
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:e6fbe6723b4801f6854303b3fd7d4a4a36727bdd28dc28002b7de4929739ccc0
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/ubi9/toolbox@sha256:ea4ce8cb8f59cf1b0709f1606d7cab074c3806d4016964eedc0555410cb5dfeb
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:5914f71e37c815ce2567826896e051303bdef52fae1acf8d02e672fc24c7eb9d
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:1709fa3c79aad4ba7eb9be8299949396092c8e20210124e0c0936385bc04e839
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:cb5be3dd7098f4c8d66f9e85ad04656460df6102652deddc6cccb884a12be733
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:2b3aace0d80474470e71c2c8291b25738e096109229fe6d5d3e06e9885ee58b6
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:2f1a0a0803b2f5118b167e9d9c2e86660add5f79ee6f47044f63b0e8abe0b75f
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:ccba67a6a5aa3403e0018f29dcdfef483220dea8065d8b572e80c653ca992add
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:463e9be9c4409ba1e5a42713bd65844a023e7a3e9032d4e862edbaa5fe35300a
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:abe1f9dd11c7313fdd55b678be511a5cdc3546654a68e2b53d42c1013fec8d00
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:bd6d1599b4ff9eb890be9808014715b28ca19247222d65f40df024db4dd3525a
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:74d4e0174655cf6f6273857325a1e714a1531c5a06ad38bf139bee9c7d8f5f19
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:32c9511846077ef6035eb19be7653f6a78e0f207eb420ab93b18f742ab5779b6
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        version: 2.22.2
      entries:
      - name: rhods-operator.2.22.2
        version: 2.22.2
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.22
    - currentCSV: rhods-operator.2.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:b3563e201bf2720f60e73c0b04cb70566ea52eda7d09a28753f3a1487993c2e8
          createdAt: "2024-08-01T06:23:03Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.8.0 <2.8.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:eb4da3a645b53acfd6ce8758c043ae19b4126e62bcca09165114a141c2c8d014
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:92d9e4b53d580df6cc1f14a849b477749a1306941a2e8f9d1b6160f4911c3cff
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:b3563e201bf2720f60e73c0b04cb70566ea52eda7d09a28753f3a1487993c2e8
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:5f39b11f0cd6db2d7d6e51ee666c030e2643aee92a090b49656c1e7dfd8a5430
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:90833e85e0de3d7de870074c8b04319cd11b913b9652af6a5727aae41a91458c
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:91525fac5fadece75fbdafc40b94445fb92c88cfc8c2f06b596f8c6377e59c9e
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:ab8792fcd16b1aa0065444ba54335cb3e747379740c3ec2d56beb46ff80cb044
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:987c768c0688431a53e48a80591f2a41e074994169f9e691b28b89528d8e736f
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:1a6e29214e06a2665baadbed598ba758d6141bfa522984eb05f05a80558a403c
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:850b78a201effcafe33705dcd2670f1c470d8b46ba6e307032403627b55a7cfe
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:080377d42d782bf525b4072384d0cf8409b48019b3748ca7c855f340bcb94251
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:b6ecb3858c201498024e3bfc4a93e630699be4b1fdc3a65fd2d70e4b375eb9c2
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:df2edcaf4402b11ce24439263956e3b0cc43efbac2fdf152afacb1a398dcc59a
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:889f046b255b13d99801108588d3abf577611b828657b5240a18ea21f0e3cbbe
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:1c49888ba2e9449c04d5588040c099290fce4793c2ba8874c1037316e550d6e1
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:0b8665e8978deb73be2cc65830f08a74a5a48f3e3ee7b2b817f7bbacba257d22
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:68e1331f13dfaf6641a1e7d54f4cf29c0ef42663abc98b6dc69167892d2d8adb
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:06a6d5cbbe80d49bca9e4db9929571a69e35ae7951fcae84dcd6460f7ddf622c
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:9799003449bfb61ae8f59efd2a611b0da3bccadd9b20cc1519f6c9279d085337
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:9c90873fbd5af8fac8efc3c7c27fba742a06fad2d47f336319dd05d2b116867b
        version: 2.8.4
      entries:
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.8
    defaultChannel: stable
    packageName: rhods-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: ansible-automation-platform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aap-operator.v2.4.0-0.1758079821
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.1-0.1635279521 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the hub, including all hosted Ansible
              content, secrets and the database. By default, a persistent volume claim
              will be created using the default StorageClass on your cluster to store
              the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous hub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Automation Hub
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-24

          ## How to Upgrade

          Starting with AAP 2.4 in the May 2025 async release, the managed database was upgraded to PostgreSQL 15. It is recommended to take a backup of your deployment before upgrading.

          To take a backup, create AutomationControllerBackup and AutomationHubBackup objects for your deployments.
          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel.

          Note: Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8-operator@sha256:5b957031609c8aabf7575e675bc636c1152babc9564d503802301aff4642a404
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-runner-rhel8@sha256:e14286dc9605be3a8c6cb144d0b2e3f3bb3211129005009e7566ea6c16c91f52
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8@sha256:7ae8db2b5d12b619f40db1e581db1e982a695bc1a29a2628222fff09e7c25249
        - registry.redhat.io/ansible-automation-platform-24/ee-minimal-rhel8@sha256:7670e41037f023e3fc6ee95d2342787a80077273ab8d30ba6804029ff05b4edc
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-ui-rhel8@sha256:d229d4e09b2b788d1ba3413ea7ec15374bd1613d507c285ac1ddc10570b9473f
        - registry.redhat.io/ansible-automation-platform-24/hub-web-rhel8@sha256:6ff1f602f0d9cad201b3addb65acc570a9d527de005ca85fc861d3cb98138710
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8@sha256:bfd13d6b39c5ec3a67adbafb05a0482328f0b9b97a517b45c093f6c35a3a92d3
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8@sha256:04fb5cb6b0d7c79cf5d4b754fe917224ddfcb2b9a4994462db404c833ed69483
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-rhel8-operator@sha256:b120e6a5ef376662b4324f0458eae31627087dc91fc613788143bc3954a15318
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-24/ee-supported-rhel8@sha256:f1384eb558f7da1e4a35d51408b3cf5a5a9be08f4d50503f473b26909bc246fc
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8@sha256:cccb4db08735551c697a7b2cbf75812f34500e691f9cdbaea02b30edea292e47
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8-operator@sha256:5253c9fd587ff2ca94d251f2b5fc7f284080adc5eb64643df5299beb2015592f
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8-operator@sha256:3c28a1326ecb7495952a5bb8be13b0972a4378af82611b0b9d990c5fa9fa82c3
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8-operator@sha256:e20106a8a04abd568a4c56861cea39ae444c0bd204bc807ac01f1705e12937d3
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        version: 2.4.0+0.1758079821
      entries:
      - name: aap-operator.v2.4.0-0.1702352195
        version: 2.4.0+0.1702352195
      - name: aap-operator.v2.4.0-0.1707506910
        version: 2.4.0+0.1707506910
      - name: aap-operator.v2.4.0-0.1708974895
        version: 2.4.0+0.1708974895
      - name: aap-operator.v2.4.0-0.1711590540
        version: 2.4.0+0.1711590540
      - name: aap-operator.v2.4.0-0.1713466670
        version: 2.4.0+0.1713466670
      - name: aap-operator.v2.4.0-0.1715898994
        version: 2.4.0+0.1715898994
      - name: aap-operator.v2.4.0-0.1717558474
        version: 2.4.0+0.1717558474
      - name: aap-operator.v2.4.0-0.1718152210
        version: 2.4.0+0.1718152210
      - name: aap-operator.v2.4.0-0.1720482179
        version: 2.4.0+0.1720482179
      - name: aap-operator.v2.4.0-0.1725256739
        version: 2.4.0+0.1725256739
      - name: aap-operator.v2.4.0-0.1726111413
        version: 2.4.0+0.1726111413
      - name: aap-operator.v2.4.0-0.1727257740
        version: 2.4.0+0.1727257740
      - name: aap-operator.v2.4.0-0.1730152858
        version: 2.4.0+0.1730152858
      - name: aap-operator.v2.4.0-0.1733185647
        version: 2.4.0+0.1733185647
      - name: aap-operator.v2.4.0-0.1733943951
        version: 2.4.0+0.1733943951
      - name: aap-operator.v2.4.0-0.1737661496
        version: 2.4.0+0.1737661496
      - name: aap-operator.v2.4.0-0.1741194076
        version: 2.4.0+0.1741194076
      - name: aap-operator.v2.4.0-0.1742444794
        version: 2.4.0+0.1742444794
      - name: aap-operator.v2.4.0-0.1743629049
        version: 2.4.0+0.1743629049
      - name: aap-operator.v2.4.0-0.1746132523
        version: 2.4.0+0.1746132523
      - name: aap-operator.v2.4.0-0.1749069319
        version: 2.4.0+0.1749069319
      - name: aap-operator.v2.4.0-0.1750877268
        version: 2.4.0+0.1750877268
      - name: aap-operator.v2.4.0-0.1753232084
        version: 2.4.0+0.1753232084
      - name: aap-operator.v2.4.0-0.1755833431
        version: 2.4.0+0.1755833431
      - name: aap-operator.v2.4.0-0.1758079821
        version: 2.4.0+0.1758079821
      name: stable-2.4
    - currentCSV: aap-operator.v2.4.0-0.1758080293
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.1-0.1635283332 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the hub, including all hosted Ansible
              content, secrets and the database. By default, a persistent volume claim
              will be created using the default StorageClass on your cluster to store
              the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous hub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Automation Hub
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-24

          ## How to Upgrade

          Starting with AAP 2.4 in the May 2025 async release, the managed database was upgraded to PostgreSQL 15. It is recommended to take a backup of your deployment before upgrading.

          To take a backup, create AutomationControllerBackup and AutomationHubBackup objects for your deployments.
          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel.

          Note: Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8@sha256:bfd13d6b39c5ec3a67adbafb05a0482328f0b9b97a517b45c093f6c35a3a92d3
        - registry.redhat.io/ansible-automation-platform-24/ee-minimal-rhel8@sha256:7670e41037f023e3fc6ee95d2342787a80077273ab8d30ba6804029ff05b4edc
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8@sha256:04fb5cb6b0d7c79cf5d4b754fe917224ddfcb2b9a4994462db404c833ed69483
        - registry.redhat.io/ansible-automation-platform-24/ee-supported-rhel8@sha256:f1384eb558f7da1e4a35d51408b3cf5a5a9be08f4d50503f473b26909bc246fc
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-rhel8-operator@sha256:b120e6a5ef376662b4324f0458eae31627087dc91fc613788143bc3954a15318
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8-operator@sha256:5253c9fd587ff2ca94d251f2b5fc7f284080adc5eb64643df5299beb2015592f
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8-operator@sha256:3c28a1326ecb7495952a5bb8be13b0972a4378af82611b0b9d990c5fa9fa82c3
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8@sha256:7ae8db2b5d12b619f40db1e581db1e982a695bc1a29a2628222fff09e7c25249
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8@sha256:cccb4db08735551c697a7b2cbf75812f34500e691f9cdbaea02b30edea292e47
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8-operator@sha256:5b957031609c8aabf7575e675bc636c1152babc9564d503802301aff4642a404
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8-operator@sha256:e20106a8a04abd568a4c56861cea39ae444c0bd204bc807ac01f1705e12937d3
        - registry.redhat.io/ansible-automation-platform-24/hub-web-rhel8@sha256:6ff1f602f0d9cad201b3addb65acc570a9d527de005ca85fc861d3cb98138710
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-runner-rhel8@sha256:e14286dc9605be3a8c6cb144d0b2e3f3bb3211129005009e7566ea6c16c91f52
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-ui-rhel8@sha256:d229d4e09b2b788d1ba3413ea7ec15374bd1613d507c285ac1ddc10570b9473f
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        version: 2.4.0+0.1758080293
      entries:
      - name: aap-operator.v2.4.0-0.1702352614
        version: 2.4.0+0.1702352614
      - name: aap-operator.v2.4.0-0.1707507304
        version: 2.4.0+0.1707507304
      - name: aap-operator.v2.4.0-0.1708975279
        version: 2.4.0+0.1708975279
      - name: aap-operator.v2.4.0-0.1711591330
        version: 2.4.0+0.1711591330
      - name: aap-operator.v2.4.0-0.1713467078
        version: 2.4.0+0.1713467078
      - name: aap-operator.v2.4.0-0.1715899379
        version: 2.4.0+0.1715899379
      - name: aap-operator.v2.4.0-0.1717558885
        version: 2.4.0+0.1717558885
      - name: aap-operator.v2.4.0-0.1718152680
        version: 2.4.0+0.1718152680
      - name: aap-operator.v2.4.0-0.1720482740
        version: 2.4.0+0.1720482740
      - name: aap-operator.v2.4.0-0.1725257213
        version: 2.4.0+0.1725257213
      - name: aap-operator.v2.4.0-0.1726111804
        version: 2.4.0+0.1726111804
      - name: aap-operator.v2.4.0-0.1727259094
        version: 2.4.0+0.1727259094
      - name: aap-operator.v2.4.0-0.1730153340
        version: 2.4.0+0.1730153340
      - name: aap-operator.v2.4.0-0.1733186325
        version: 2.4.0+0.1733186325
      - name: aap-operator.v2.4.0-0.1733945743
        version: 2.4.0+0.1733945743
      - name: aap-operator.v2.4.0-0.1737662040
        version: 2.4.0+0.1737662040
      - name: aap-operator.v2.4.0-0.1741194557
        version: 2.4.0+0.1741194557
      - name: aap-operator.v2.4.0-0.1742445409
        version: 2.4.0+0.1742445409
      - name: aap-operator.v2.4.0-0.1743630487
        version: 2.4.0+0.1743630487
      - name: aap-operator.v2.4.0-0.1746133932
        version: 2.4.0+0.1746133932
      - name: aap-operator.v2.4.0-0.1749069693
        version: 2.4.0+0.1749069693
      - name: aap-operator.v2.4.0-0.1750878254
        version: 2.4.0+0.1750878254
      - name: aap-operator.v2.4.0-0.1753232791
        version: 2.4.0+0.1753232791
      - name: aap-operator.v2.4.0-0.1755833968
        version: 2.4.0+0.1755833968
      - name: aap-operator.v2.4.0-0.1758080293
        version: 2.4.0+0.1758080293
      name: stable-2.4-cluster-scoped
    - currentCSV: aap-operator.v2.5.0-0.1758147230
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727257740 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation gateway - Unify the management of AAP from a central location.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner


          ## New in AAP 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-25

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.5

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console. For example, change the channel from `stable-2.4` to `stable-2.5`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.

          After changing to the new 2.5 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place.

          #### Create the AnsibleAutomationPlatform custom resource

          Once that is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: myaap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8@sha256:3a989f148648a5a62e687058ec3ca772fa22aa10faec89453626d2b17b3fd9c2
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8-operator@sha256:43e731b9a435b3a412108170b4b39e6b6215be51d362a9f03c33fa412e97da53
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8-operator@sha256:8d8182e3a87c55eacc222e71df4a9bf7a2e51102d913a5824cc91e36056d4249
        - registry.redhat.io/ansible-automation-platform-25/hub-web-rhel8@sha256:e96832efcc03c6754031cb70f56c808135f62d1b80a882dd6d6fc9e1a4cfc69b
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8@sha256:697343aaea153ab22d7efa8f1bb5aed37a55cbb353fc206f4a355e8b878ae418
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8@sha256:ac0de2b7d354eb0c15a6c12861a4de3577a982c84af661fbab25517d8105c085
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8@sha256:aed3973eb953dda3571260dcf1b33bc7dcbc18c41e3350b8c361b7e635fb9527
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-runner-rhel8@sha256:ace1d2ab49ba5ff9f11ab18e8edce27c9e73ae7298bf7335569fa1740a44899b
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8-operator@sha256:acc0fcd509eeac46db307ab744dd0a0c26770a93e495cf2ec0ccb61a91c6fbc5
        - registry.redhat.io/ansible-automation-platform-25/ee-minimal-rhel8@sha256:41f944eb1450db56c3b14fc7baece8f148d5f10886a57bad730bec60fb28ac04
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8-operator@sha256:07aa7e257d04ae0eda6fcaa4af9121f1aece439231fce4f924e6467da91f0022
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-ui-rhel8@sha256:9ab35e80a6172018ed3f5cd179e0bcb6befc1378583d9e73c3a64a7cf9e8050f
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8@sha256:9f2867603b9747c2fe37862c2c5efac4a4530902f323a8bf7cc0717de4cdb5aa
        - registry.redhat.io/ansible-automation-platform-25/gateway-proxy-rhel8@sha256:789d55f355fa0d90f9509a912d199123bfc5a4041930a72a8fd4ed1d819f0d84
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8-operator@sha256:503fc7a20e2dd7de2cb55b9be5f97efa0d2b31f054ce326f23f3a11b4d683e64
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-rhel8-operator@sha256:74be9182723b41f87b3d13b3d86b6a9049bba88e2902cc9453264aa8a0f584e4
        - registry.redhat.io/ansible-automation-platform-25/ee-supported-rhel8@sha256:52d2c73fa2015af5f1b9a88ee48085c7c7bab44330fc9f95914f48b58e4b48bc
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-chatbot-rhel8@sha256:9a37c9739c488ccf211ec5efc6debb045223d508c9cb3fa5a8256eca98ff2f0c
        version: 2.5.0+0.1758147230
      entries:
      - name: aap-operator.v2.5.0-0.1727240614
        version: 2.5.0+0.1727240614
      - name: aap-operator.v2.5.0-0.1727875185
        version: 2.5.0+0.1727875185
      - name: aap-operator.v2.5.0-0.1728520175
        version: 2.5.0+0.1728520175
      - name: aap-operator.v2.5.0-0.1729741547
        version: 2.5.0+0.1729741547
      - name: aap-operator.v2.5.0-0.1731456196
        version: 2.5.0+0.1731456196
      - name: aap-operator.v2.5.0-0.1733193164
        version: 2.5.0+0.1733193164
      - name: aap-operator.v2.5.0-0.1733969526
        version: 2.5.0+0.1733969526
      - name: aap-operator.v2.5.0-0.1736893446
        version: 2.5.0+0.1736893446
      - name: aap-operator.v2.5.0-0.1737101025
        version: 2.5.0+0.1737101025
      - name: aap-operator.v2.5.0-0.1737675968
        version: 2.5.0+0.1737675968
      - name: aap-operator.v2.5.0-0.1739248042
        version: 2.5.0+0.1739248042
      - name: aap-operator.v2.5.0-0.1740093573
        version: 2.5.0+0.1740093573
      - name: aap-operator.v2.5.0-0.1740773472
        version: 2.5.0+0.1740773472
      - name: aap-operator.v2.5.0-0.1741368605
        version: 2.5.0+0.1741368605
      - name: aap-operator.v2.5.0-0.1742434024
        version: 2.5.0+0.1742434024
      - name: aap-operator.v2.5.0-0.1743660124
        version: 2.5.0+0.1743660124
      - name: aap-operator.v2.5.0-0.1746137767
        version: 2.5.0+0.1746137767
      - name: aap-operator.v2.5.0-0.1747343762
        version: 2.5.0+0.1747343762
      - name: aap-operator.v2.5.0-0.1749074128
        version: 2.5.0+0.1749074128
      - name: aap-operator.v2.5.0-0.1749604727
        version: 2.5.0+0.1749604727
      - name: aap-operator.v2.5.0-0.1750901111
        version: 2.5.0+0.1750901111
      - name: aap-operator.v2.5.0-0.1753900156
        version: 2.5.0+0.1753900156
      - name: aap-operator.v2.5.0-0.1755835086
        version: 2.5.0+0.1755835086
      - name: aap-operator.v2.5.0-0.1758147230
        version: 2.5.0+0.1758147230
      name: stable-2.5
    - currentCSV: aap-operator.v2.5.0-0.1758147817
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727259094 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation gateway - Unify the management of AAP from a central location.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner


          ## New in AAP 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-25

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.5

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console. For example, change the channel from `stable-2.4` to `stable-2.5`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.

          After changing to the new 2.5 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place.

          #### Create the AnsibleAutomationPlatform custom resource

          Once that is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: myaap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8-operator@sha256:acc0fcd509eeac46db307ab744dd0a0c26770a93e495cf2ec0ccb61a91c6fbc5
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-ui-rhel8@sha256:9ab35e80a6172018ed3f5cd179e0bcb6befc1378583d9e73c3a64a7cf9e8050f
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8-operator@sha256:43e731b9a435b3a412108170b4b39e6b6215be51d362a9f03c33fa412e97da53
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8-operator@sha256:8d8182e3a87c55eacc222e71df4a9bf7a2e51102d913a5824cc91e36056d4249
        - registry.redhat.io/ansible-automation-platform-25/hub-web-rhel8@sha256:e96832efcc03c6754031cb70f56c808135f62d1b80a882dd6d6fc9e1a4cfc69b
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-chatbot-rhel8@sha256:9a37c9739c488ccf211ec5efc6debb045223d508c9cb3fa5a8256eca98ff2f0c
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8@sha256:3a989f148648a5a62e687058ec3ca772fa22aa10faec89453626d2b17b3fd9c2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/ee-minimal-rhel8@sha256:41f944eb1450db56c3b14fc7baece8f148d5f10886a57bad730bec60fb28ac04
        - registry.redhat.io/ansible-automation-platform-25/gateway-proxy-rhel8@sha256:789d55f355fa0d90f9509a912d199123bfc5a4041930a72a8fd4ed1d819f0d84
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8@sha256:aed3973eb953dda3571260dcf1b33bc7dcbc18c41e3350b8c361b7e635fb9527
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8@sha256:697343aaea153ab22d7efa8f1bb5aed37a55cbb353fc206f4a355e8b878ae418
        - registry.redhat.io/ansible-automation-platform-25/ee-supported-rhel8@sha256:52d2c73fa2015af5f1b9a88ee48085c7c7bab44330fc9f95914f48b58e4b48bc
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8@sha256:ac0de2b7d354eb0c15a6c12861a4de3577a982c84af661fbab25517d8105c085
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-rhel8-operator@sha256:74be9182723b41f87b3d13b3d86b6a9049bba88e2902cc9453264aa8a0f584e4
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-runner-rhel8@sha256:ace1d2ab49ba5ff9f11ab18e8edce27c9e73ae7298bf7335569fa1740a44899b
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8@sha256:9f2867603b9747c2fe37862c2c5efac4a4530902f323a8bf7cc0717de4cdb5aa
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8-operator@sha256:07aa7e257d04ae0eda6fcaa4af9121f1aece439231fce4f924e6467da91f0022
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8-operator@sha256:503fc7a20e2dd7de2cb55b9be5f97efa0d2b31f054ce326f23f3a11b4d683e64
        version: 2.5.0+0.1758147817
      entries:
      - name: aap-operator.v2.5.0-0.1727241223
        version: 2.5.0+0.1727241223
      - name: aap-operator.v2.5.0-0.1727875896
        version: 2.5.0+0.1727875896
      - name: aap-operator.v2.5.0-0.1728520687
        version: 2.5.0+0.1728520687
      - name: aap-operator.v2.5.0-0.1729742145
        version: 2.5.0+0.1729742145
      - name: aap-operator.v2.5.0-0.1731456854
        version: 2.5.0+0.1731456854
      - name: aap-operator.v2.5.0-0.1733193761
        version: 2.5.0+0.1733193761
      - name: aap-operator.v2.5.0-0.1733970210
        version: 2.5.0+0.1733970210
      - name: aap-operator.v2.5.0-0.1736894220
        version: 2.5.0+0.1736894220
      - name: aap-operator.v2.5.0-0.1737101781
        version: 2.5.0+0.1737101781
      - name: aap-operator.v2.5.0-0.1737676792
        version: 2.5.0+0.1737676792
      - name: aap-operator.v2.5.0-0.1739248649
        version: 2.5.0+0.1739248649
      - name: aap-operator.v2.5.0-0.1740094176
        version: 2.5.0+0.1740094176
      - name: aap-operator.v2.5.0-0.1740774104
        version: 2.5.0+0.1740774104
      - name: aap-operator.v2.5.0-0.1741369251
        version: 2.5.0+0.1741369251
      - name: aap-operator.v2.5.0-0.1742434756
        version: 2.5.0+0.1742434756
      - name: aap-operator.v2.5.0-0.1743660958
        version: 2.5.0+0.1743660958
      - name: aap-operator.v2.5.0-0.1746138413
        version: 2.5.0+0.1746138413
      - name: aap-operator.v2.5.0-0.1747345055
        version: 2.5.0+0.1747345055
      - name: aap-operator.v2.5.0-0.1749074612
        version: 2.5.0+0.1749074612
      - name: aap-operator.v2.5.0-0.1749607543
        version: 2.5.0+0.1749607543
      - name: aap-operator.v2.5.0-0.1750901870
        version: 2.5.0+0.1750901870
      - name: aap-operator.v2.5.0-0.1753900674
        version: 2.5.0+0.1753900674
      - name: aap-operator.v2.5.0-0.1755835623
        version: 2.5.0+0.1755835623
      - name: aap-operator.v2.5.0-0.1758147817
        version: 2.5.0+0.1758147817
      name: stable-2.5-cluster-scoped
    - currentCSV: aap-operator.v2.6.0-0.1758684783
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727257740 <2.6.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Automation Gateway - Unify the management of AAP from a central location.
            * Automation Controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private Automation Hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Ansible Lightspeed: Deploy and manage the lifecycle of an Ansible Lightspeed application.
            * Platform Resource Operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## New in AAP 2.6 and 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.


          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-26

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5 and above, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.6

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console.

          **For customers on AAP 2.4:** Change the channel directly from `stable-2.4` to `stable-2.6` to upgrade directly to AAP 2.6, which will be released in October.

          **For customers on AAP 2.5:** Change the channel from `stable-2.5` to `stable-2.6`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.

          After changing to the new 2.6 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place. When upgrading from AAP 2.4 to 2.6, your external authentication configuration in Automation Controller will automatically be migrated to the new unified Platform UI.

          **Database:** AAP 2.6 continues to use PostgreSQL 15.

          #### Create the AnsibleAutomationPlatform custom resource (Required for upgrades from 2.4 only)

          **If upgrading from AAP 2.4:** Once the operator upgrade is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created.

          **If upgrading from AAP 2.5:** This step is not required as your AnsibleAutomationPlatform custom resource already exists.

          Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          Next, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda (lightspeed is disabled by default):


          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: aap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-26/gateway-rhel9-operator@sha256:fa1b102a0690767a1e08935a0e164c1eabb2d4cf9afccd310262ee9185544de7
        - registry.redhat.io/ansible-automation-platform-26/hub-rhel9@sha256:8041b141cd8921aaaa20cce15eab4c6189da9f449d99057189284891c071ac3b
        - registry.redhat.io/ansible-automation-platform-26/gateway-proxy-rhel9@sha256:13284d20c6fb4f2de24ec1715e7abdb5c4b0e4298c5034a7795f494b93634a02
        - registry.redhat.io/rhel9/postgresql-15@sha256:5e225c1646251d2509b3a6e45fad23abfb11dd5c45d3224b59178f90deb38a70
        - registry.redhat.io/ansible-automation-platform-26/controller-rhel9-operator@sha256:863e77e52ab01cec2a71672a224080a59d963d4fc08b9ed6f3cc7950729b1746
        - registry.redhat.io/ansible-automation-platform-26/hub-rhel9-operator@sha256:18114aab6db0d00780fdeaf9efa74ee44cab94824c2fd35f4b9998b6378ad8b3
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-rhel9-operator@sha256:cfbb98e17297134c187673f355342ac5ad60a56d73f4b0c918be2ff244286cc0
        - registry.redhat.io/ansible-automation-platform-26/platform-resource-rhel9-operator@sha256:969bbd033fbba33624329ee9347b3b89fb5a1f64e71892f5c6c7d74ed14a6d39
        - registry.redhat.io/ansible-automation-platform-26/controller-rhel9@sha256:64f5e77a7842965e9c6f7bc7990354b71bb19ffb06ab8ef7793dde956cb6c0f0
        - registry.redhat.io/ansible-automation-platform-26/ee-minimal-rhel9@sha256:261c324cac081e47d559c5fccbd48e7ec8f5bf64e3f27287fbe4fb8aff6b7d99
        - registry.redhat.io/ansible-automation-platform-26/hub-web-rhel9@sha256:aa4041672542a7f828d8a51c79e9c4dcea2a185388e4f20f828104db37e03a9b
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-rhel9@sha256:01e56bee772d15b120d6e4b02a59024b21eb8b266127b1a9b7f2c535ecdb2855
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-ui-rhel9@sha256:5b5f38a5f76781a042eec9cbf70b8be026878e4fa61c7fb90d1b0ad64bd21592
        - registry.redhat.io/ubi9/ubi-minimal@sha256:7c5495d5fad59aaee12abc3cbbd2b283818ee1e814b00dbc7f25bf2d14fa4f0c
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-chatbot-rhel9@sha256:8901fba14b21569b02108d2ab9050d558dee29f8177fae9d1bfcd5de3f4ce7c1
        - registry.redhat.io/ansible-automation-platform-26/mcp-tools-rhel9@sha256:0c1f6f6958f2570c8cd073287555d28c70ef9dded702ec014d3ead8ce572463e
        - registry.redhat.io/ansible-automation-platform-26/platform-resource-runner-rhel9@sha256:9983043386df38a17ff6fffef64dc187691e0b1fefec01eef3af39aa08e532ef
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-rhel9-operator@sha256:bfd55bffd8239e224d4ebba500d8517e8e66915078d6b59244f5e910fc5e10c8
        - registry.redhat.io/ansible-automation-platform-26/gateway-rhel9@sha256:c85dbbf6beaa264afec480dc1f06576aada531cdfd15f247f4db37dc9702be33
        - registry.redhat.io/rhel9/redis-6@sha256:0962d18bb5292c0e524354f319b63287c2298a4a2fd34556c68fa285df381427
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-rhel9@sha256:c5d4e2e76944e0991729e5c3b99f43b7f29ea08ad112e5e6513feb8fb1d899a1
        - registry.redhat.io/ansible-automation-platform-26/ee-supported-rhel9@sha256:2534d07f1ae41af32adb681e283cbaea465cbe0f9d234ed5478a7bfad7676e7f
        version: 2.6.0+0.1758684783
      entries:
      - name: aap-operator.v2.6.0-0.1758684783
        version: 2.6.0+0.1758684783
      name: stable-2.6
    - currentCSV: aap-operator.v2.6.0-0.1758685241
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727259094 <2.6.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Automation Gateway - Unify the management of AAP from a central location.
            * Automation Controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private Automation Hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Ansible Lightspeed: Deploy and manage the lifecycle of an Ansible Lightspeed application.
            * Platform Resource Operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## New in AAP 2.6 and 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.


          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-26

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5 and above, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.6

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console.

          **For customers on AAP 2.4:** Change the channel directly from `stable-2.4` to `stable-2.6` to upgrade directly to AAP 2.6, which will be released in October.

          **For customers on AAP 2.5:** Change the channel from `stable-2.5` to `stable-2.6`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.

          After changing to the new 2.6 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place. When upgrading from AAP 2.4 to 2.6, your external authentication configuration in Automation Controller will automatically be migrated to the new unified Platform UI.

          **Database:** AAP 2.6 continues to use PostgreSQL 15.

          #### Create the AnsibleAutomationPlatform custom resource (Required for upgrades from 2.4 only)

          **If upgrading from AAP 2.4:** Once the operator upgrade is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created.

          **If upgrading from AAP 2.5:** This step is not required as your AnsibleAutomationPlatform custom resource already exists.

          Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          Next, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda (lightspeed is disabled by default):


          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: aap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-26/gateway-rhel9@sha256:c85dbbf6beaa264afec480dc1f06576aada531cdfd15f247f4db37dc9702be33
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-ui-rhel9@sha256:5b5f38a5f76781a042eec9cbf70b8be026878e4fa61c7fb90d1b0ad64bd21592
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-rhel9-operator@sha256:cfbb98e17297134c187673f355342ac5ad60a56d73f4b0c918be2ff244286cc0
        - registry.redhat.io/ansible-automation-platform-26/controller-rhel9@sha256:64f5e77a7842965e9c6f7bc7990354b71bb19ffb06ab8ef7793dde956cb6c0f0
        - registry.redhat.io/ansible-automation-platform-26/ee-supported-rhel9@sha256:2534d07f1ae41af32adb681e283cbaea465cbe0f9d234ed5478a7bfad7676e7f
        - registry.redhat.io/ansible-automation-platform-26/ee-minimal-rhel9@sha256:261c324cac081e47d559c5fccbd48e7ec8f5bf64e3f27287fbe4fb8aff6b7d99
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-rhel9@sha256:c5d4e2e76944e0991729e5c3b99f43b7f29ea08ad112e5e6513feb8fb1d899a1
        - registry.redhat.io/ansible-automation-platform-26/hub-web-rhel9@sha256:aa4041672542a7f828d8a51c79e9c4dcea2a185388e4f20f828104db37e03a9b
        - registry.redhat.io/ansible-automation-platform-26/platform-resource-runner-rhel9@sha256:9983043386df38a17ff6fffef64dc187691e0b1fefec01eef3af39aa08e532ef
        - registry.redhat.io/ansible-automation-platform-26/gateway-rhel9-operator@sha256:fa1b102a0690767a1e08935a0e164c1eabb2d4cf9afccd310262ee9185544de7
        - registry.redhat.io/ansible-automation-platform-26/platform-resource-rhel9-operator@sha256:969bbd033fbba33624329ee9347b3b89fb5a1f64e71892f5c6c7d74ed14a6d39
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-chatbot-rhel9@sha256:8901fba14b21569b02108d2ab9050d558dee29f8177fae9d1bfcd5de3f4ce7c1
        - registry.redhat.io/ansible-automation-platform-26/mcp-tools-rhel9@sha256:0c1f6f6958f2570c8cd073287555d28c70ef9dded702ec014d3ead8ce572463e
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-rhel9-operator@sha256:bfd55bffd8239e224d4ebba500d8517e8e66915078d6b59244f5e910fc5e10c8
        - registry.redhat.io/ansible-automation-platform-26/hub-rhel9-operator@sha256:18114aab6db0d00780fdeaf9efa74ee44cab94824c2fd35f4b9998b6378ad8b3
        - registry.redhat.io/rhel9/postgresql-15@sha256:5e225c1646251d2509b3a6e45fad23abfb11dd5c45d3224b59178f90deb38a70
        - registry.redhat.io/ubi9/ubi-minimal@sha256:7c5495d5fad59aaee12abc3cbbd2b283818ee1e814b00dbc7f25bf2d14fa4f0c
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-rhel9@sha256:01e56bee772d15b120d6e4b02a59024b21eb8b266127b1a9b7f2c535ecdb2855
        - registry.redhat.io/ansible-automation-platform-26/gateway-proxy-rhel9@sha256:13284d20c6fb4f2de24ec1715e7abdb5c4b0e4298c5034a7795f494b93634a02
        - registry.redhat.io/ansible-automation-platform-26/hub-rhel9@sha256:8041b141cd8921aaaa20cce15eab4c6189da9f449d99057189284891c071ac3b
        - registry.redhat.io/rhel9/redis-6@sha256:0962d18bb5292c0e524354f319b63287c2298a4a2fd34556c68fa285df381427
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/ansible-automation-platform-26/controller-rhel9-operator@sha256:863e77e52ab01cec2a71672a224080a59d963d4fc08b9ed6f3cc7950729b1746
        version: 2.6.0+0.1758685241
      entries:
      - name: aap-operator.v2.6.0-0.1758685241
        version: 2.6.0+0.1758685241
      name: stable-2.6-cluster-scoped
    defaultChannel: stable-2.6-cluster-scoped
    packageName: ansible-automation-platform-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: cert-utils-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-utils-operator.v1.3.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/cert-utils-operator@sha256:0290e7b2800a5aa8eb57b0268908f2a20a93b16403c76df4c61021048161d6ff
          createdAt: "2024-01-23T12:29:04Z"
          description: Set of utilities for TLS certificates
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-utils-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/cert-utils-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Cert utils operator is a set of functionalities around certificates packaged in a [Kubernetes operator](https://github.com/operator-framework/operator-sdk).

          Certificates are assumed to be available in a [secret](https://kubernetes.io/docs/concepts/configuration/secret/) of type `kubernetes.io/tls` (other types of secrets are *ignored* by this operator).
          By convention this type of secrets have three optional entries:

          1. `tls.key`: the private key of the certificate.
          2. `tls.crt`: the actual certificate.
          3. `ca.crt`: the CA bundle that validates the certificate.

          The functionalities are the following:

          1. [Ability to populate route certificates](https://github.com/redhat-cop/cert-utils-operator#Populating-route-certificates)
          2. [Ability to create java keystore and truststore from the certificates](https://github.com/redhat-cop/cert-utils-operator#Creating-java-keystore-and-truststore)
          3. [Ability to show info regarding the certificates](https://github.com/redhat-cop/cert-utils-operator#Showing-info-on-the-certificates)
          4. [Ability to alert when a certificate is about to expire](https://github.com/redhat-cop/cert-utils-operator#Alerting-when-a-certificate-is-about-to-expire)
          5. [Ability to inject ca bundles in ValidatingWebhookConfiguration, MutatingWebhookConfiguration, CustomResourceDefinition object](https://github.com/redhat-cop/cert-utils-operator#CA-injection)

          All these feature are activated via opt-in annotations.
        displayName: Cert Utils Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Cert Utils Operator
          url: https://github.com/redhat-cop/cert-utils-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/cert-utils-operator
        maintainers:
        - email: rspazzoli@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/cert-utils-operator@sha256:0290e7b2800a5aa8eb57b0268908f2a20a93b16403c76df4c61021048161d6ff
        version: 1.3.12
      entries:
      - name: cert-utils-operator.v1.3.12
        version: 1.3.12
      - name: cert-utils-operator.v1.3.11
        version: 1.3.11
      - name: cert-utils-operator.v1.3.10
        version: 1.3.10
      - name: cert-utils-operator.v1.3.9
        version: 1.3.9
      - name: cert-utils-operator.v1.3.8
        version: 1.3.8
      - name: cert-utils-operator.v1.3.7
        version: 1.3.7
      - name: cert-utils-operator.v1.3.6
        version: 1.3.6
      - name: cert-utils-operator.v1.3.5
        version: 1.3.5
      - name: cert-utils-operator.v1.3.4
        version: 1.3.4
      - name: cert-utils-operator.v1.3.3
        version: 1.3.3
      - name: cert-utils-operator.v1.3.2
        version: 1.3.2
      - name: cert-utils-operator.v1.3.1
        version: 1.3.1
      - name: cert-utils-operator.v1.3.0
        version: 1.3.0
      - name: cert-utils-operator.v1.2.0
        version: 1.2.0
      - name: cert-utils-operator.v1.1.0
        version: 1.1.0
      - name: cert-utils-operator.v1.0.7
        version: 1.0.7
      - name: cert-utils-operator.v1.0.6
        version: 1.0.6
      - name: cert-utils-operator.v1.0.5
        version: 1.0.5
      - name: cert-utils-operator.v1.0.4
        version: 1.0.4
      - name: cert-utils-operator.v1.0.3
        version: 1.0.3
      - name: cert-utils-operator.v1.0.2
        version: 1.0.2
      - name: cert-utils-operator.v1.0.1
        version: 1.0.1
      - name: cert-utils-operator.v1.0.0
        version: 1.0.0
      - name: cert-utils-operator.v0.2.2
        version: 0.2.2
      - name: cert-utils-operator.v0.2.1
        version: 0.2.1
      - name: cert-utils-operator.v0.2.0
        version: 0.2.0
      - name: cert-utils-operator.v0.1.1
        version: 0.1.1
      - name: cert-utils-operator.v0.1.0
        version: 0.1.0
      - name: cert-utils-operator.v0.0.3
        version: 0.0.3
      - name: cert-utils-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cert-utils-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ses-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ses-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ses-controller:1.1.1
          createdAt: "2025-09-19T19:37:25Z"
          description: AWS SES controller is a service controller for managing SES
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConfigurationSet represents the state of an AWS ses ConfigurationSet
              resource.
            displayName: ConfigurationSet
            kind: ConfigurationSet
            name: configurationsets.ses.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Simple Email Service (SES) resources in AWS from within your Kubernetes cluster.

          **About Amazon SES**

          Amazon Simple Email Service (SES) is an email platform that provides an easy, cost-effective way for you to send and receive email using your own email addresses and domains.
          For example, you can send marketing emails such as special offers, transactional emails such as order confirmations, and other types of correspondence such as newsletters. When you use Amazon SES to receive mail, you can develop software solutions such as email autoresponders, email unsubscribe systems, and applications that generate customer support tickets from incoming emails.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SES
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ses
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SES Developer Resources
          url: https://aws.amazon.com/ses/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ses maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ses-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ses-controller.v1.1.1
        version: 1.1.1
      - name: ack-ses-controller.v1.1.0
        version: 1.1.0
      - name: ack-ses-controller.v1.0.13
        version: 1.0.13
      - name: ack-ses-controller.v1.0.12
        version: 1.0.12
      - name: ack-ses-controller.v1.0.11
        version: 1.0.11
      - name: ack-ses-controller.v1.0.10
        version: 1.0.10
      - name: ack-ses-controller.v1.0.9
        version: 1.0.9
      - name: ack-ses-controller.v1.0.8
        version: 1.0.8
      - name: ack-ses-controller.v1.0.7
        version: 1.0.7
      - name: ack-ses-controller.v1.0.6
        version: 1.0.6
      - name: ack-ses-controller.v1.0.5
        version: 1.0.5
      - name: ack-ses-controller.v1.0.3
        version: 1.0.3
      - name: ack-ses-controller.v1.0.2
        version: 1.0.2
      - name: ack-ses-controller.v1.0.1
        version: 1.0.1
      - name: ack-ses-controller.v1.0.0
        version: 1.0.0
      - name: ack-ses-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-ses-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco Networking - Compute Solutions
      provider-url: ""
    name: ucs-ci-solutions-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ucs-ci-solutions-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intersight.cisco.com/v1alpha1",
                "kind": "UCSCISolutions",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ucs-ci-solutions-operator",
                    "app.kubernetes.io/instance": "ucscisolutions-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ucscisolutions",
                    "app.kubernetes.io/part-of": "ucs-ci-solutions-operator"
                  },
                  "name": "ucscisolutions-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: docker.io/pkoppa/ucs-ci-solutions-operator:latest
          createdAt: "2023-07-15T15:19:39Z"
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: UCSCISolutions
            name: ucscisolutions.intersight.cisco.com
            version: v1alpha1
        description: The operator is used to automate the creation of  Pools, Policies
          and Server Profile Templates required to deploy Cisco UCS Converged Infrastructure
          Solution in Intersight.
        displayName: UCS-CI-Solutions-Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ucscisolutions
        - intersight
        - ucs
        - ci
        - flexpod
        - flashstack
        links:
        - name: Ucs Ci Solutions Operator
          url: https://ucs-ci-solutions-operator.domain
        maintainers:
        - email: pkoppa@cisco.com
          name: Paniraj Koppa
        - email: chrobrie@cisco.com
          name: Chris O'Brien
        maturity: alpha
        provider:
          name: Cisco Networking - Compute Solutions
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - docker.io/pkoppa/ucs-ci-solutions-operator:latest
        version: 0.0.1
      entries:
      - name: ucs-ci-solutions-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ucs-ci-solutions-operator
    provider:
      name: Cisco Networking - Compute Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-diagnostics
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-diagnostics.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Diagnostics",
                "metadata": {
                  "name": "diagnostics-sample"
                },
                "spec": {
                  "agentDiagnosticsService": {
                    "port": 8080
                  },
                  "auth": {
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "wapdbDiagnosticsService": {
                    "dumpedTables": "[ \"--regexpr cluster_logs_week_*\", \"current_week_cluster_logs\", \"config_metadata_dbs\", \"config_metadata_dbs_id_seq\", \"config_metadata_tags\", \"config_tags\", \"csv_reports_meta\", \"custom_import_handler_dbs\", \"data_point_db_models\", \"db_test_entries\", \"license_server_metadata\", \"operations\", \"result_file_metadata_dbs\", \"result_metadata_dbs\", \"result_metadata_tags\", \"result_tags\", \"session_meta\", \"session_types\", \"sessions\", \"stat_view_query_data\", \"test_run_configs\", \"test_run_infos\", \"test_runs\", \"wireless_globalsettings\", \"wireless_sessionconfigs\", \"wireless_templates\"]",
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:37:10Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Diagnostics
            name: diagnostics.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Diagnostics Operator
        displayName: Keysight WAP Diagnostics Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Diagnostics
          url: https://keysight-wap-diagnostics.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-diagnostics@sha256:97bf8affa7bf1f3a391f5f036313e7cd8196be2eb8e95debeab06c1b6c964c5e
        version: 0.1.21
      entries:
      - name: keysight-wap-diagnostics.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-diagnostics
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JFrog
      provider-url: ""
    name: openshiftartifactoryha-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: artifactory-ha-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftArtifactoryHa",
                "metadata": {
                  "name": "openshiftartifactoryha"
                },
                "spec": {
                  "artifactory-ha": {
                    "artifactory": {
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE",
                      "node": {
                        "replicaCount": 2,
                        "waitForPrimaryStartup": {
                          "enabled": false
                        }
                      }
                    },
                    "databaseUpgradeReady": true,
                    "database": {
                      "driver": "OVERRIDE",
                      "password": "OVERRIDE",
                      "type": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "nginx": {
                      "http": {
                        "externalPort": 80,
                        "internalPort": 8080
                      },
                      "https": {
                        "externalPort": 443,
                        "internalPort": 8443
                      },
                      "service": {
                        "ssloffload": false
                      },
                      "tlsSecretName": "OVERRIDE"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "waitForDatabase": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/artifactory-operator@sha256:de69bb16052d5ac6fed9882eb65d49446b6fa1cc5ba830dd2a39e882712b1347
          createdAt: 2024-03-23 00:00:00+00:00
          description: JFrog Artifactory Enterprise deploys Artifactory in a high
            availability environment across multiple pods
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Artifactory HA Instances
            displayName: Artifactory HA
            kind: OpenshiftArtifactoryHa
            name: openshiftartifactoryhas.charts.helm.k8s.io
            version: v1
        description: |2+
           ## Breaking change
          Please update to version 1.1.10 of the operator.
          Version 1.1.8 and 1.1.9 have issues. Please skip these versions.
          If you are using operator version < 1.2.0 there is a issue with upgrade where artifactory version is not updated if you update operator version. As workaround you can update the Custom Resource OpenshiftArtifactoryHa which has hardcored image value.After upgrading to version 1.2.0 you can remove the below parameters from your CR so that future upgrade works without workaround. ``` artifactory-ha.artifactory.image.registry artifactory-ha.artifactory.image.repository artifactory-ha.artifactory.image.tag ```
          ## Overview
          Openshift Operator to deploy JFrog Artifactory Enterprise into your Openshift cluster.
          ## Security Context Constraints
          To deploy this helm chart you will need to be a cluster admin w/ access to the anyuid scc and add the operator service account to the anyuid scc.
          ```
          oc adm policy add-scc-to-user anyuid -z artifactory-ha-operator -n openshift-operators
          ```
          Add the service account for the Artifactory chart to deploy successfully:
          ``` oc adm policy add-scc-to-user anyuid -z openshiftartifactoryha-artifactory-ha -n artifactory ```
          ### Usage

          An external DB is required. The operator will not deploy a DB but will require you to specify the configuration values to connect to it.

          Search for JFrog and click JFrog Artifactory Enterprise Operator to install.

          Go to the Installed Operators.

          Wait for the JFrog Artifactory Enterprise Operator to complete the installation.

          Open the Operator and click on the provided API: Artifactory HA.

          Click Create New Instance and provide the following parameters for your DB configuration:

          ```
          DATABASE_TYPE
          DATABASE_DRIVER
          DATABASE_URL
          DATABASE_USER
          DATABASE_PASSWORD
          ```
          Master key and Join key must be supplied. To generate a new key for each run the command below:

          ```
          # Create a key
          export JOIN_KEY=$(openssl rand -hex 32)
          echo ${JOIN_KEY}
          ```

          To use TLS you will need to first create a k8s tls secret to store your .crt and .key file into.
          Then supply the value of this k8s secret into the TLS_SECRET field.
          ``` oc create secret tls my_tls_secret --cert=tls.crt --key=tls.key --namespace=my_namespace ```
          Click Create for Artifactory Enterprise to deploy into OpenShift and connect to it on the external IP exposed by the load balancer.
          ### Create a route
          To expose Artifactory from Openshift we recommend you create a new route in Openshift.
          You can either use the oc command line tool or the Openshift web console to generate a new route.
          Depending upon where you terminate TLS you will need to either specify pass through or edge.
          Command Line (Edge):
          ``` oc create route edge --service=openshiftartifactory-ha --cert=tls.crt --key=tls.key --ca-cert=ca.crt --hostname=www.example.com ```
          Or for more information visit the official Openshift documentation on routes here:
          https://docs.openshift.com/container-platform/4.6/networking/routes/route-configuration.html

        displayName: JFrog Artifactory Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Artifact Repository
          url: https://jfrog.com/artifactory/
        - name: Artifactory Video
          url: https://www.youtube.com/watch?v=r2_A5CPo43U
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: alpha
        provider:
          name: JFrog
        relatedImages:
        - registry.redhat.io/rhel8/nginx-120@sha256:87f658744af695273c434fac8246b716728b691a17d5d0c5b89247362e61e45c
        - registry.connect.redhat.com/jfrog/artifactory-operator@sha256:de69bb16052d5ac6fed9882eb65d49446b6fa1cc5ba830dd2a39e882712b1347
        - registry.connect.redhat.com/jfrog/artifactory-pro@sha256:0986ca71419ae628892db11e7ca7da5ac1c60219c144ce9aeb2de6e683a45262
        version: 1.3.3
      entries:
      - name: artifactory-ha-operator.v1.3.3
        version: 1.3.3
      name: alpha
    - currentCSV: artifactory-ha-operator.v1.1.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftArtifactoryHa",
                "metadata": {
                  "name": "openshiftartifactoryha"
                },
                "spec": {
                  "artifactory-ha": {
                    "artifactory": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/artifactory-pro",
                        "tag": "7.27.3-1"
                      },
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE",
                      "uid": "1000721030",
                      "node": {
                        "replicaCount": 2,
                        "waitForPrimaryStartup": {
                          "enabled": false
                        }
                      }
                    },
                    "databaseUpgradeReady": true,
                    "database": {
                      "driver": "OVERRIDE",
                      "password": "OVERRIDE",
                      "type": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "initContainerImage": "registry.connect.redhat.com/jfrog/init@sha256:164b41720a37f75c6129c8a28d05cc56f821022313d44ff918cce4a2146d93e8",
                    "nginx": {
                      "uid": "1000720104",
                      "gid": "1000720107",
                      "http": {
                        "externalPort": 80,
                        "internalPort": 8080
                      },
                      "https": {
                        "externalPort": 443,
                        "internalPort": 8443
                      },
                      "image": {
                        "registry": "registry.redhat.io",
                        "repository": "rhel8/nginx-116",
                        "tag": "latest"
                      },
                      "service": {
                        "ssloffload": false
                      },
                      "tlsSecretName": "OVERRIDE"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "waitForDatabase": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/artifactory-operator@sha256:e292553e1b6e792ff19276c7f6565d7ed61dd95289eb176066329954bc93e371
          createdAt: 2020-03-25 00:00:00+00:00
          description: JFrog Artifactory Enterprise deploys Artifactory in a high
            availability environment across multiple pods
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Artifactory HA Instances
            displayName: Artifactory HA
            kind: OpenshiftArtifactoryHa
            name: openshiftartifactoryhas.charts.helm.k8s.io
            version: v1
        description: |2+
           ## Breaking change
          Please update to version 1.1.10 of the operator.
          Version 1.1.8 and 1.1.9 have issues. Please skip these versions.

          ## Overview
          Openshift Operator to deploy JFrog Artifactory Enterprise into your Openshift cluster.
          ## Security Context Constraints
          To deploy this helm chart you will need to be a cluster admin w/ access to the anyuid scc and add the operator service account to the anyuid scc.
          ```
          oc adm policy add-scc-to-user anyuid -z artifactory-ha-operator -n openshift-operators
          ```
          Add the service account for the Artifactory chart to deploy successfully:
          ``` oc adm policy add-scc-to-user anyuid -z openshiftartifactoryha-artifactory-ha -n artifactory ```
          ### Usage

          An external DB is required. The operator will not deploy a DB but will require you to specify the configuration values to connect to it.

          Search for JFrog and click JFrog Artifactory Enterprise Operator to install.

          Go to the Installed Operators.

          Wait for the JFrog Artifactory Enterprise Operator to complete the installation.

          Open the Operator and click on the provided API: Artifactory HA.

          Click Create New Instance and provide the following parameters for your DB configuration:

          ```
          DATABASE_TYPE
          DATABASE_DRIVER
          DATABASE_URL
          DATABASE_USER
          DATABASE_PASSWORD
          ```
          Master key and Join key must be supplied. To generate a new key for each run the command below:

          ```
          # Create a key
          export JOIN_KEY=$(openssl rand -hex 32)
          echo ${JOIN_KEY}
          ```

          To use TLS you will need to first create a k8s tls secret to store your .crt and .key file into.
          Then supply the value of this k8s secret into the TLS_SECRET field.
          ``` oc create secret tls my_tls_secret --cert=tls.crt --key=tls.key --namespace=my_namespace ```
          Click Create for Artifactory Enterprise to deploy into OpenShift and connect to it on the external IP exposed by the load balancer.
          ### Create a route
          To expose Artifactory from Openshift we recommend you create a new route in Openshift.
          You can either use the oc command line tool or the Openshift web console to generate a new route.
          Depending upon where you terminate TLS you will need to either specify pass through or edge.
          Command Line (Edge):
          ``` oc create route edge --service=openshiftartifactory-ha --cert=tls.crt --key=tls.key --ca-cert=ca.crt --hostname=www.example.com ```
          Or for more information visit the official Openshift documentation on routes here:
          https://docs.openshift.com/container-platform/4.6/networking/routes/route-configuration.html

        displayName: JFrog Artifactory Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Artifact Repository
          url: https://jfrog.com/artifactory/
        - name: Artifactory Video
          url: https://www.youtube.com/watch?v=r2_A5CPo43U
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: stable
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/init@sha256:164b41720a37f75c6129c8a28d05cc56f821022313d44ff918cce4a2146d93e8
        - registry.connect.redhat.com/jfrog/artifactory-operator@sha256:e292553e1b6e792ff19276c7f6565d7ed61dd95289eb176066329954bc93e371
        - registry.connect.redhat.com/jfrog/artifactory-pro@sha256:7289dde95d1e4d8e161a28cffd9c4a0cc96f92f7536d069ab63ec304a92254fd
        - registry.redhat.io/rhel8/nginx-116@sha256:d71d368d6094a5777000a8d99164959ccb97fca355e855e24bd92c13a8df04f6
        version: 1.1.18
      entries:
      - name: artifactory-ha-operator.v1.1.18
        version: 1.1.18
      name: stable
    defaultChannel: alpha
    packageName: openshiftartifactoryha-operator
    provider:
      name: JFrog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: advanced-cluster-management
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: advanced-cluster-management.v2.10.9
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-08-06T03:17:11Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0 <2.10.9'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/internal-objects: '[]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterHub defines the configuration for an instance
              of the MultiCluster Hub
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:95df18f4479b7f555a41508486624bde714778730921c20c665e9f23baadbabf
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:8f1bb3a8422d29960d8b90dd24a813a56875abf1f0393f78b862eaa48ddda783
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:50331cdd61332c8f4128b3dab237255463b307c5c40b0607730a7d3b33b9c6b5
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:5b162aeab4aa51185c880696d1472e152a20ada926a3baea0fb8494793ad238e
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:0ba2471b7a67532bdb90e02904671f0e724cdb39f74a45392ae6c693c25b354d
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:4b16da3070e30d879421aaa0d4a73aa2a1b0b852f62b8faabd801c03f22d5319
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:ced03229f208739867a43b9ed9bd158f19f4242dd1f5f045d3e5db81b2b1d171
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:4ad5ccf9ce682024b6e3025a9bcadd6f32692aa18d892afd670eacb7b4031860
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:1c2dbcf179508ea1c602f7817f85b0d3f58d6d85ae93ccc523e88a3936c9c690
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:23713ebb12d510d37de7e060d5ce4ff2198990f83cebbd2302e74bd83a55cecd
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:1aaba9ca65e363c7f9c07ef2f66f23cb8fff677204a56373634d319113d2c84b
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:e3c80e8e3af5c02c202dfb288c63b20c44c08f2a0474d2568cfb0b878f7fe888
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:9e78db48ff0b0fc0ac0347ca055a4b66e2f917073499adc6e2db2a9c16a70e01
        - registry.redhat.io/rhacm2/console-rhel9@sha256:1538a3a935aa01d8800d8145c45037eaab6987ad6e6e927017ca1cfea4bb111c
        - registry.redhat.io/rhel9/postgresql-13@sha256:31c77c19480ab82cb33525d32419b68cf59daccf0b9936e88e8238e48a2d47e5
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:fbc5b1456a37951c9f70f64357919222739f4a7e533fb6dde5b93672535fc87e
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:95890bb1aad544d97ae1b80e801a800c95ef51e1034b60e680b67bdb2f5aeb69
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:9e931cab5fcbf3df0ffe6fff8d12303800423ace3d749ee0e33b55300c816b19
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:dcd6a420467929b9af8c52c77c8717d22f3977e6b6b392b5dc90cbb04005b9f1
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:14094f09dfb166a6b342bd91e2db9162f0de88d55655ccce44d345047c700187
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:23b4d7fbc776c63dcd7007442a557124581939ce110257822e8233261d31fb7d
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:bbe57064cb1c40c35929526a082f11b2da6fd3746b6eef8af51ca63a4fd7d98b
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:46e3b8ccf3e7069b44eb3ad407329c03048b5219ef2886e237e1ab0dc3f7bae4
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:bb4b72c1510da84aad2d67516c58f37495d7682c51c8f2088e2c4a350fa36c50
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:dcbc4c5e2d46ec2b4911d8cd9f96a4c141bf7dd18b79aa4c25b77141e4f60b48
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:69a8faa2f4da2281c6c124508634628510d91ad40f725a9b41790e0821d64c14
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:500718f834e05c7a4531eb4f9d8ce560053c5125f76f6921c38c9cf150046a23
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:9db39230c528bfc675bc422a7acd40c394169b9632b088ca16358751994c9cac
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:335c903dde6f21e9c54f025c45666f6b4b6e7d455dfd5f85a6373b3947069f0a
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:3ccc45b0e9483f33be0d3b32a6d46fa794976131ad2f9861d96e319c8e35b9be
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:2f68a16a14f192b2cceaa07ed4840c08fbf413718a39bd584cacde866be7f2b1
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:9c4eb0ef8758d943796e13f65bc92b190db614eabd0fe7375ebfecc90e96e32c
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:da09087ba7d558611d733551b6ebc27ccf441581fc264efe7a5933db3816ef89
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:e02992cec8edcc9374c4cd30dfd7bccbda77d6d5cdd309862ab5265494f1278e
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:f4474ed21a71ad490778a6a378a856bec1884b4b9dc2c0e6c6140cc6c9d03b61
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:805e443fbfca4f816dc2dcea71f4cf723fd753ccea6d13f04d25945f29cb47dc
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:d8b8d130f75647cfeb9036db50d66fc4f1fca56fead6add4ba24c52f9213543e
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:13c33a997be0ddef59d1fb31835751367d4f901a1a6eac41284931bf7988035c
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:814f34f9e55b59d96021d3160b1dad7a68ffe208b3f4a999840b228aa7b3300b
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:47a8fa9541275f849a24e2ef1df87b7e9051dedd0ee8bd833d60defbd8724b46
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:3803d82d2b829c0e0883c0504176a96d2012cb7096cfcace316831a2b55bca3f
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:de68d598587a967d4b3f5e0e28c4675f74fd79cae7a5bb197c1b4b44bb4b31d9
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:35ff861165fccd848b8d328fae46cb8b3ac0c853c63062ec8672a6d41a23f333
        - registry.redhat.io/rhacm2/iam-policy-controller-rhel9@sha256:8c2b356bd9ff4ebfba9059a5f20cad8bd336af237d84cdc21f5c0aa35fabeabe
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:d989e9d2078f8a1e4b18de64f917e6d5e290c5cac769899f829b5f6bbfe6f0f2
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:87093f85c6235c99d4cca9d503b0e4198e3731e5fc3171506485091570e70bb5
        version: 2.10.9
      entries:
      - name: advanced-cluster-management.v2.10.9
        version: 2.10.9
      - name: advanced-cluster-management.v2.10.8
        version: 2.10.8
      - name: advanced-cluster-management.v2.10.7
        version: 2.10.7
      - name: advanced-cluster-management.v2.10.6
        version: 2.10.6
      - name: advanced-cluster-management.v2.10.5
        version: 2.10.5
      - name: advanced-cluster-management.v2.10.4
        version: 2.10.4
      - name: advanced-cluster-management.v2.10.3
        version: 2.10.3
      - name: advanced-cluster-management.v2.10.2
        version: 2.10.2
      - name: advanced-cluster-management.v2.10.1
        version: 2.10.1
      - name: advanced-cluster-management.v2.10.0
        version: 2.10.0
      name: release-2.10
    - currentCSV: advanced-cluster-management.v2.11.8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-08-01T05:25:37Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.10.0 <2.11.8'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:956d549e64c412f86f2083c8ddf97fea269d7806d8826aab93178bec9ed5d7af
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:9d821f4373722d9cb2b679cf5793ef22cb0b9ebcbb5a38060504c93c05c492c4
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:aef5ffbaa578d006bc7caec8fb35446dc263a51813f84c756aa57903b428329b
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:d248a5fad4ae3ee3210eb6dc4f58406d2f1e8f5c390f3c491040b437fe1adeba
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:447b1f6067a82476e14dfbb74a8f324fb86956420e55575f0017b91784f303a5
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:a081f531efb5b54d7d14d9ba0519552ecfc87c3a80d9d3ec4e598d2b445fc689
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:ffd3d9c363d8e15232bd6d12c5f1a77f7cb1ca18f94b7aea79d226568be4586e
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:95890bb1aad544d97ae1b80e801a800c95ef51e1034b60e680b67bdb2f5aeb69
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:f5c56a7381eb80c02a7c140f7c3f0ed08bb93697cd51732ee57498c080d8d3e5
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:8d15be5f40d135fd7069ac768dcf9188d8d8299c65db56df04e66f8a28941966
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:94c89ebd9257797ddc264511399689feb70d56e53332d9894c6d95fbdc51de40
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:6d4e3114494e051b3b875eeeae4bad4d48665fbcbbd3574a98f87b5a5742b41d
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:9be2d97fa3c1140db7d86157503a96ca2d3bdfb57125ab2230c8cf0649ea1e90
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:d4feeb3d3322210765d916ef1f49b47a1527f1c61041f94834c4395319a9ec32
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:49f0ec1aced72e6a2f0c862b0c6c0cd430d997a09ceb9eb0a7d25aaa1d9798b7
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:a9e0533af92ebadca27c9fe93fe4db9c68a1a8bc0060f4d13f67f26eb5691cb3
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:b3e0427341e3b1cbb1fe9166c56184eb0b0ec526d8a3992306e4861f65e4f0c0
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:ded47519fcf30e60cf38d44a101494b0318757969675782365de88b51cab24f0
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:968eb25c5cfcbdcea87b92dbff734d01b5954c13c13f5e794ad1db7bd886e2b4
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:d7552b2a3c864a3fc0d279a2519535e170fd8608195fe1f9abb4b538dfe2a3df
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:d6666c572294c7c85990304d21eaae3040b9c4e47fcb7b446289cb09aabc2d26
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:4a62d8706772ede443cffa005b984c26d137c35a33aafedaa7f01050b618c2a2
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:d4bfef313ff9044ae36aa5f832651a4dd09228d4b9b186803009aafe9774ee2c
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:1e8a26ebfcc33e60f60da4ba8d58cf7c152a99752eac6ebc9e114187e4cc2267
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:4dce3cebeb4ea0f1d3fc1c7e036dc6a3fd09e02f8f2ae4f3a753d9f4f97defe7
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:b713264826124df2c2b3a99dc268a3f3bd7eb24a9010b93c9876d83656e70e1a
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:e36832aa4852ea807d44f2b1f2c402bdaac6b7e2246bf9dd811b41f0b5e8e647
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:eaa719f9d9db8f91eced81e0e164bbc12bb502bf4760fc8c673cc542b4414771
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:45d4404abe02596611b29316bd10985f7a0684b1d305a57d3b24068ac776cf6e
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:251d3ee881c3e902e87f1752648ab084e8780aef79eb88a253ba79f984a01bcf
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:e347cd0f57e0691c61263e6554495c76ce56c7443bce3fd010cf76d068083fee
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:ad10d15efb6833b2daaf153210a080420395a7debfcb873dd757bb924197a654
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:36c90a52b12af24faaf28301d0d1a0c522723bd6cadd7391e5ee1f43f3019eef
        - registry.redhat.io/rhacm2/console-rhel9@sha256:2ed2f4add387caf790b2b41a65a3986e0046e08d0b8406ad17e4f0440ba0ac4a
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:5ba07381111651906ac3c93344e705152eb006826cb36bf5c05bd88723c4c999
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:7d39d0c29b02c399ffdcd83278693a8dd9c5ca3931e660af056030b126d10055
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:eec0e9bd6ea6f146939516da6c7ca9c3755888d96e81954f2f282fae63c1847c
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:827ca1beb707c02908914c517c8ab14b77b1c65ca7690dbbe1bcff600b7551d5
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:4e0c2cf5229fd68102a9e6d0f3155c651b1159b7af7b057151ce52990f747b08
        - registry.redhat.io/rhel9/postgresql-13@sha256:31c77c19480ab82cb33525d32419b68cf59daccf0b9936e88e8238e48a2d47e5
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:ef66846a3f178234618fd1e265067f863bbb56f14cbca0dfc96b0d741885dc9b
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:4b16da3070e30d879421aaa0d4a73aa2a1b0b852f62b8faabd801c03f22d5319
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:028b7258ec88d04397798552ed19a459110660ddbdf3ab375a7a1f74049c455a
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:147c567ce95673ba7151bcd290474756e13018ebc9f7275d61677ffe586a8a76
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:2878b5c41f8d58e986e86803bfd7eb3a05ccf23f7e49ebba56ad30245609b23d
        version: 2.11.8
      entries:
      - name: advanced-cluster-management.v2.11.8
        version: 2.11.8
      - name: advanced-cluster-management.v2.11.7
        version: 2.11.7
      - name: advanced-cluster-management.v2.11.6
        version: 2.11.6
      - name: advanced-cluster-management.v2.11.5
        version: 2.11.5
      - name: advanced-cluster-management.v2.11.4
        version: 2.11.4
      - name: advanced-cluster-management.v2.11.3
        version: 2.11.3
      - name: advanced-cluster-management.v2.11.2
        version: 2.11.2
      - name: advanced-cluster-management.v2.11.1
        version: 2.11.1
      - name: advanced-cluster-management.v2.11.0
        version: 2.11.0
      name: release-2.11
    - currentCSV: advanced-cluster-management.v2.12.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-21T21:16:17Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <2.12.4'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:b83e630e77bd011c374d73d801f10a3230f6798ede10d7f39bf2e8ff6437cf2c
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:3d3260ff206c5ee249443b1f8da49562584f7ee1abd284691b42f1728b6ed5dc
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:b85f3de5231f2f1bc3b4e9c95917f562f39e09dc9b7db35690024163e3a5afe5
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:6f2c4f1abc4388e76414544e65f08238c96fc4d6fbf71fa79330d9dca117d221
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:ec65a5744945ca20f804d94999a4002efab9ddad360a58f9498800bffa9fa0f4
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:187879345174e515863bc1b29a961a0a0d9ad44006a6e12de715b4508dec38a8
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:93df0586f9281fa2879eed0fdf2fb115be355d572cc24dc5c5c9a92137044877
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:58f08a93cd35bd2ac19f94a1779205c81a2fc3a552d1e77b7c659fd51c9dd34f
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:bd24cfabb81b464a31c3d8c21fa6371e46b83094d5c750acc33346203fe15e84
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:c2259b90decf24f8fb1cb058d4751c75fd1558e3c9fd1079857ad9c47ca9e8d2
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:22e262c0353c6f0d5d1b916074836c53437e22697f67c85a43e6d717daf651fa
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:c402a4ac61e11dc9e45fdc7bc7fcb5bf5ca3475a66d782aea60d9f6524e51a24
        - registry.redhat.io/rhel8/postgresql-13@sha256:826a1a75f4186f36217b5c1f1aa270838c4b55b1021d33a4876566bfc3b8e629
        - registry.redhat.io/rhacm2/console-rhel9@sha256:d239df44dafac3bde3eb7ae8c80cb342fa4bb4eabe22f27d2c92d283324e6e10
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:9c297b4c1f9ba17908594f1843ea6de6c5a5e0f3500258c23fbb8b5684249c9e
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:cdaf7be69a5d66cfa4777feb8266421dafd55353d6703e7e591bb9fb23017abc
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:7c548dd869f3458fb70a536708306cb615b9166c7e866cc1215e82f183b39529
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:1419e274b4e2e4c37985208c3b1b9ecd0cdef493427ace3c71ff35ed8fcc394b
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:a5c65dee1f7e4177050244109c10e384899e01d1924b41a540373aa7daf4f274
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:c46657831915ed5f42f4a55e725bc9d17e0671191f4509e610468fda160710c0
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:1e5c12ed0de090198e097cc5deb133ad075fba224f239b8b892e6e1672ec9883
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:fff4fb0924f866b305ad0ecc5d2682da6cb20de15448a2a8c640df182719b4e2
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:28730c8f2344e51b0b9ca603ac63b60885b39be86582943bd1cc200779eee0c5
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:1ccd17b482df844ec42936993ff162f665b5a473ad344270ff2425cd63d428fc
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:6a7c54cb7e6c3cc506ffb8a4f3ec7443ef80798e2c917e20b5f06d39bde02bc4
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:7375991ce3b9e64cb7542dc176414f0974047667d08beac184e2adadd9f83532
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:3ac71b363aef8db127b88e0a63ee616d0d06342142f83522e34c9cac4dd77b2e
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:c4c3bc4f45ae194b884d89319625f2a0970dada823877a35feecd8c733cb2c8d
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:7050f71c3a038fec16db8d0eee69d7d5a4294c043c13e9cf69c3cb2e28897f4c
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:70ddf96707d398d5444bb7bbde6c684275cff62b8a83c7cf6143ff529daf0dee
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:f3a00f55dee152e0345e2616ac14f652dcb4f37f3b9dfca6f049ac5522874c6e
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:f1c371fef66e96199c4a96e21192b47bca423b282d6f99e927644840aaedfd80
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:8ccbbc058d324429aef0c31c99de8c7b3a5abc02ec6f7a14cc60eef63a8f5c69
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:7b64803ddd3b7886ad3ec3de943a38cec5aeb9604732ab7a5ea5061748bf1411
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:5156129064dbe4e9989b41018b1bb250b2eb3f02e0c0afd75a874da9b54ad3c2
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:875e4bf42ad192cfb954e38c8474a529aa938855d69dc0d83c4ee4a4777c1028
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:3faa2318f159d264a9194fc01239d3a2098c184b365b153e3c56e555d5d4992e
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:7968e96b48b2ed63e7657793cf6cc2564055809df19cb43b7e284ac03f2cc26c
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:dca845ad05b02619b81f231eb57dd826720c221d22282fe7422ceebe57191371
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:794ef15591f217be629d6c680c56e84962923993f6dd77514ca5d2bdbd96217b
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:c10a72bf8c0d76b9b4d9d7c0bfca4d83a3f7dd7f6a37b4cc24ee7591638a2482
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:08dd3f6805fa02478671927982a2c77bdec0a5b7e7fd1c9faf20c8af2bb7b428
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:fbfade039739179995828b2ce4f4411b1315ac9a0e4a98e96a115012676a859d
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:b782190b3a285003d319321be09a25ee1376c951cb0cd146cbb634c9b1ca97b3
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:8e62fa4decf01a3beae7080f19469f33822f5010a688ada5db8d2d23b008e498
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:3f286e3c24fe73b4b42e6cf4b6769f91e333014450e1ab135b1e5b4a271cbaec
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:c2ef90b4704251dc9f7ef5288f5bad7aeb4cd2b3706ec531c60f6f4346073a77
        version: 2.12.4
      entries:
      - name: advanced-cluster-management.v2.12.4
        version: 2.12.4
      - name: advanced-cluster-management.v2.12.3
        version: 2.12.3
      - name: advanced-cluster-management.v2.12.2
        version: 2.12.2
      - name: advanced-cluster-management.v2.12.1
        version: 2.12.1
      - name: advanced-cluster-management.v2.12.0
        version: 2.12.0
      name: release-2.12
    - currentCSV: advanced-cluster-management.v2.13.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-09-08T18:37:07Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <2.13.4'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:dc271885425ab74e0554b531a2078133ce02006d0bb9688f3b54a850cfefd8f8
        - registry.redhat.io/rhel8/postgresql-16@sha256:89dfeaefc7e9e974e47a3365e22f354314f8a5e906e93a0152afe4e42f6a82d0
        - registry.redhat.io/rhacm2/console-rhel9@sha256:5526edf70b784bddfc2b7e1c154a44201c346fc20117635500bb54d325921c25
        - registry.redhat.io/rhacm2/acm-flightctl-worker-rhel9@sha256:57f69874b2df77c185374ddcea530af6783410a0f35528ee1a44cc2dee6f7fa2
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:0040791500337f88af7d922895487549915faa3b69dacba1a23540e7d6f2e580
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:ba51f7d077858a0abe32a254970390b50cd7d75433b5661e17f887c48c56a30f
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:65e1f0eb4e6c797c6dc6518674b28370db3fdaa6d7e17f4233c07207224dcb48
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:3028d011e7b88dd9597821128275f606fdba639e5b83e4e6a8783fa2616cba6f
        - registry.redhat.io/rhacm2/acm-flightctl-api-rhel9@sha256:8302b536c896290533f3f50084cb9047cb55af3a4404887b70704d23862619fa
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:d29ceda561f36faa581d76d8b889f7e01d253d33c88709078855ae26cada4917
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:0272215ab43b5688295d94cce0de4490a2fbabc1dd09ed41645be7acef1e3c9b
        - registry.redhat.io/rhacm2/acm-flightctl-ocp-ui-rhel9@sha256:89863ce0eb5fb73272bda8762d74461552f8502fbe262b16cbde01064cc533b3
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:84109be1814719c343b8b4676981cdbcea0356f012b6c854221c271bdb518fea
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:fc16203de469dd1ba16e428a9d7efbb16a5f511d866ef6fcc253530de1f032a6
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:ab1e4f8685896e2016a0b4c4479378058314f365e4f09ef42cba67f3649bb306
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:b3535378919524df259b85587ca83f5620433914418c9941374951c36e0530ae
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:1b7e6e6a4ae36209f98b1539c37c5975a82363b33fe5ac1e3a320e58b99be239
        - registry.redhat.io/rhacm2/acm-flightctl-periodic-rhel9@sha256:2c9b3c3d6cd7e435da337447580dfbfde8c4932308405eeae5d97ed5ec57fbfc
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:588678bea97237d87cd2e55a0e8000e73461cfc421cc91d554a2f1e352b68d14
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:7f742f024c787932932b8a4100a5329e70aad39537e0c20c2a4e245fc6245e25
        - registry.redhat.io/rhel9/redis-7@sha256:848f4298a9465dafb7ce9790e991bd8a11de2558e3a6685e1d7c4a6e0fc5f371
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:ccc9c0524149820d701c39355af89287ad749832053ea55d1585e77347fa0ec0
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:7b15c6b492e5cb2725b9b345aa3bfae4108e2815bc233ac063b81a67c6550fb6
        - registry.redhat.io/rhel8/postgresql-13@sha256:826a1a75f4186f36217b5c1f1aa270838c4b55b1021d33a4876566bfc3b8e629
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:b1620cebc62240f93b12e6937d15760440be119c421c127214974bdfa9d08ebd
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:0d8202cd10170bb0a138d52a1d71e630ff74df466541897ba86ed324a75ea1d8
        - registry.redhat.io/openshift4/ose-cli-artifacts-rhel9@sha256:67ff9f19edac8d623ce9080a8c05cde3da73afd68e7352e4ee5b22a9ccb596bc
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:3490daccc203fade26e7df348c74a907ab0ffc8901f4965606c8227849087c44
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:f74fb6b4d42f1d7431bfc9d2424f953732fed06d2abd8d1fc27b878b4d1322d5
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:ce342242c50c103fc5e4b0126551231977b35bd477eb627fb8f5d2a6510ddaa0
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:e23115b10a92c323b4cdaf0412919c273dbca33f19910120498bd49cb5810ff5
        - registry.redhat.io/rhacm2/acm-flightctl-ui-rhel9@sha256:86800d67ec8ed63e5408947ba6a441e5afc1050b72a86487b9357674ed495a7e
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:f1db9b10481319dc142ca6046e4e3141addb4c0afab482eeaf80643fad9ad039
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:96bdb2e25ae0f381de0c94a09a3cd11f3c0bdc1d8cfb4fd5ced2cd7b24fbe2c2
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:4d068813e3eb704ea7b2d342d5a2ce8fa2ac71793b1b37634c18704ef0a5e8d7
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:5fc35a944895cf7ea6afa5929c45304aaa9f17397b4f1e42cb8c799d9aead9d5
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:f9ba60c2a415a863c6bd6030fb4663913368a9c271bb40b8acfdc294dcf3a7eb
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:8b152051ef2c4c077379e72aa510c0bd613b219254cad66fab0cd3ec0d1d4bf1
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:04bcddae590e2debf35090055dd52ae4bcaac4b369727266be53ac64670203d0
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:49e99f41cbea3faefc3ae93385a10b544ce91df0d6639cfee65760be6ae0df00
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:81cc71faaf20a986d233fee799a7d2b954cd3f43875a2689f76c53920ccf4df0
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:67939d6e4342372820d7af6777f0ffd123e86e4e8186132c3d4892cb717a5e7c
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:6f95308a508499d2d995653d3fd2def94ac5170b4c3f391d32a1e0594250c33a
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:669e9e2178476bc7a4688f3ebf4c20402608bb4b10e4df6863ec821c7e572299
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:ae1f175ffbd3698f2b3036bf9b62b5c9b24729cfb8248345b8bdd1525ccc5bb4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:ab0192a51ee37b7508f3e4db66ffa53b904dcb1f76d9aa1ac0e80447d8ff9e6a
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:5aec572fc18c8ed533620553aba2c676453297ad2440336a5228a5a4f50f5782
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:a0218b9d5f1803f60d9b295d1504a870780895cffe8f1c5b3ce4c3553da5e4f9
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:a2ab651af1c97e54d2c14de78ff554c0f25e8d000af93a74fc7b7ed913f4e338
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:72c6896a5a4dbb7717d76d9eeade96130d97909eee07e8d3cc17af0658b306ab
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:9ef9f1e2010d77211126482c03e8c72dfb7458940dc95ac952ee8dd1dedd2991
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:6ed6f62f5c05a24938335287b785b80c253424e1ee20d83311cd87c6a2090895
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:2da264ca59278f63990e579939371d45afec3bc7d9922a608520e7647a0377ac
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:1393744757874e79d1fb058cb5d3fe7dcbbcd2f187bdcdcdf34e67227c61fbb9
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:296a31e1aad9f6288bdf80fb3191642e3dd404e9e8fcad198016bedbd3abe9b4
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:3cbb9d271c2f23d430a623c83606bbc8a80379c24d13e610886ad51fd21ce1e4
        version: 2.13.4
      entries:
      - name: advanced-cluster-management.v2.13.4
        version: 2.13.4
      - name: advanced-cluster-management.v2.13.3
        version: 2.13.3
      - name: advanced-cluster-management.v2.13.2
        version: 2.13.2
      - name: advanced-cluster-management.v2.13.1
        version: 2.13.1
      - name: advanced-cluster-management.v2.13.0
        version: 2.13.0
      name: release-2.13
    - currentCSV: advanced-cluster-management.v2.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-25T02:10:33Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.13.0 <2.14.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          repository: https://github.com/open-cluster-management/multiclusterhub-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MulticlusterHub defines the configuration
              for an instance of a multicluster hub, a central point for managing multiple
              Kubernetes-based clusters. The deployment of multicluster hub components
              is determined based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n### Where to find more installation
          information\nYou can find additional installation guidance in the [install
          documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:364c7416a9642dfb37f05e2eac4427918959b12b159ac90a9ca310c584d6434a
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:69f0b9a7fb494572adf2795be865d6d6ba4b67db0eed0e9d61f5dcb5e9bf6efe
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:c2ef90b4704251dc9f7ef5288f5bad7aeb4cd2b3706ec531c60f6f4346073a77
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:0e1eba23e0e8454f2dec240704e7ccc922351d58880f72514d3b2bc30c9e995e
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:c6ff6cae5bc6447d0ed24e458fc5d5fc709b2106fe93b143f9546bd962ea7f87
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:ad7f30c61bb11c0d32686e75b01f3b286a1fb06edd8d2ef6a9e2d924fba2b565
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:b489229190f9a3da66bd154e65490c3a35d1999d2db6707759f06eadb02b968d
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:8e6ee423acfafe28c002911e832d47dffc00a5227325bd6a5b1ab359c7eee5f4
        - registry.redhat.io/rhacm2/console-rhel9@sha256:d4c42a203110050a3834be8d5a0eee384ad7e9fbb991da589d75acd15055b8b1
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:0cecc3d87afcd0d63bdeae3ecae37f8a38534fbdc8bad5c548e0d4f4fb0156a4
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:8f4f59e237f366f08367cabbea3eb51e766ad52db0d35c5efc56182c13fc5c09
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:8253a9225f30b0f0a405e2c3293924acbfbf81f3d9c50bc4431fe4052a9fe908
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:55ef2c163b0c0deb3402e39acf37ff6bb05e2a5d268dcae5f8a400128d67c584
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:b7b7581e7743148d19283250d3300d292b96c48ecae96f6e6202a7ca8204e2ab
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:c329cfd004c58a7fcc40f4fbb986399dc2fe08d7922d2b4f210664f58751b1a6
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:143f49d2ce1f58282e276396722ba37165cd1bf2051d537bb4e6208786e5aed6
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:53deacec64238607afda29c27463d72ac3b582c9670e2e77a9c7e48851214795
        - registry.redhat.io/openshift4/ose-cli-rhel9@sha256:a48758b69b9a7451b7e0caea465a181f854b41f1e74ce86c15216ee1bb31fcb9
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:f9a9e10fde9065d650a7dc8eba83349e5db0a21727027464cf144724d0de8cfa
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:935baf67627d5e76c0766d5d2a9e36b398a12535f097749a987dd0c5de788570
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:0252cf95019717f03c7ac72fb034b5bd9ad7aef2da0a92dd21c56cf47f03b812
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:d56b997a9084d56c63f539147052e64ba30258afe91a9ff832f3673cbccbe467
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ee742e7baa8bcbbfccf79abcf15d22347d14f280eafaf22c42cef71051b10fba
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:0ab21d4189669901fe085c7f3a4b4c9778c853f8abd706bbfffcb59942fbc8e6
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:9c79b1153fc927e7582c5346ea0408e03f794eb983597f02a5e816ae961ba86b
        - registry.redhat.io/rhem/flightctl-periodic-rhel9@sha256:ee59f307e468e2b7b38f7c040898d587d441f8fb3d59cb022ae23e4717739e3f
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:2b1b1c707e0b5217c66a96fb82070de47fc0342747081d4be0049c4542944c8a
        - registry.redhat.io/rhel8/postgresql-16@sha256:d4abbd9492347e0aa7e1b0ff967e74d4b6a4edd6fd20d896f2f2ae03c623181c
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:46c5835a4bfc0e4767d0959bc47e939c195be197e9715aae17788d4d0e799d56
        - registry.redhat.io/rhel9/redis-7@sha256:ff31822cba556e40b6b00720839cc56f474fdb9257fb39b7604fafc11761d379
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:0ea80a1039d217e899ea012b77e99a7c5832e5b7e292dca473759c449bf2d1ea
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:e063c54c7182b70d78d24b5026f174436885383980c9b1e3e7664fa2adfad883
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:d550013d385552d6242b56c5b60f15b3ebb488c2033f808de54bcc639abc2f30
        - registry.redhat.io/rhem/flightctl-ui-ocp-rhel9@sha256:973eb3ca3da32c87ba308928a21fbd25514b2f6a6662300d4928479cb554cff3
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:bcb5949ebc31bbf6b45dfd34aae9c80c82a08531e8a3d9f6eef3354f942a5a5d
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:cee29bdb03927dc1866c267232d391ea5ee762b23c6b04a777e7d61d14a61bba
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:0cc23b5063c0d3264ed345dad8453901e4db18cc9bed1656f2799f72de75bd53
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:8151a89ebfebf2e6d336849fdd329b123b7f486938808ee80bf6bd8d44f6985c
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:7ee0aea84bf00b7b42027fa3e970c1f4b279e345a698b4050c9f413eb0383bb1
        - registry.redhat.io/rhem/flightctl-ui-rhel9@sha256:940c0564ced75b7912d3ffc923a1f6bab9aed0f4a52962f7f894d08937b7d37e
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:ce5ff191023b5e14c339b2a03aa1be05a12d024efa5302652065597e630996b6
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:90d399736ca81bb244415efaf05738beff806912a6ac7444a7d4839089b7a081
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:556d737195a4878877c650a93619af1528e52694f2a8e47e13463a603d2f4900
        - registry.redhat.io/rhem/flightctl-cli-artifacts-rhel9@sha256:3491ec932d90112e89f3a4cebe5bdd9ed6a7811f116df43bab8bb5f827129361
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:d9fa60b5b1d6b6c94c5d8b610c9a7b979202de8b46a344e0a65d06f6ef21a7ef
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:40b3e063435eea6e526a57d6bd1ae489ca703e81eeca82dda1c2e6aedec1262c
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:95085fc33a0150a39b6ef4e616fd23f0bddd9dcad0e75c4a89ed27113626e6eb
        - registry.redhat.io/rhel9/memcached@sha256:955138db9a7a23ee31198aabebd0811438fee11bb0b624c38ddb90efe52bd07c
        - registry.redhat.io/rhem/flightctl-worker-rhel9@sha256:1b1f6f85af0910f1acd029f54a608b25d3a2977c3d8b323266391543324305b4
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:f3e34b200e9480363c56da1cd88b2de37fc6a490691e4596a132217f8ea286d8
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:c70242a037e36690d8dd1030254dde778917356b4f1f61db2cecb03b6c2d7380
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:cb5346afe1c7b58eb758cdced94f2c1161c1621033b52f29757efe3a50320a65
        - registry.redhat.io/rhem/flightctl-api-rhel9@sha256:83c6c7819e04aa58d961518367ce6ad3159a28300ca097967f630044d181be34
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:8e57af77cc5dee3aaa83f6ae6e62e232c5fc4fbe4dca8eb755226351963058e3
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:44e305ca15e7c9b50bd7fa8bb391f5ee66906aa26ecc5f612d029aaee5bb4cb8
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:d65aef370a7ed01c0c7cc0ad7f0800359db7863db4ec8845de5b450992f26933
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:ad534f7b0efea45b9f4a0a026c76a2a030cb608acdb835dfea3852332acfcae1
        version: 2.14.0
      entries:
      - name: advanced-cluster-management.v2.14.0
        version: 2.14.0
      name: release-2.14
    defaultChannel: release-2.14
    packageName: advanced-cluster-management
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/component: kmm-hub
      app.kubernetes.io/name: kmm-hub
      app.kubernetes.io/part-of: kmm
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: kernel-module-management-hub
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kernel-module-management-hub.v2.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:888a3acad339cf38a324c718e0cbe1237ca5bf2f4636b01d001752658e4cca1b
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.0.2'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:888a3acad339cf38a324c718e0cbe1237ca5bf2f4636b01d001752658e4cca1b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62c44057a27abd41cf3daf60b8736da59a82dadb65fa6c32bef65d19bd49ea49
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:192e08b121a04ae69aea60c35199e8f5e3c42305e1ebd024a605a371de4ab472
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:94b4ddc45cb34d4ffe794a51b99c868cb709f90cc4eb8b28284d70c526a0a226
        version: 2.0.2
      entries:
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.0
    - currentCSV: kernel-module-management-hub.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:d1c1490d4c20dc2ec31effa1af067676c4d264e3b7d4670fc11c2baf8a400001
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.1.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:14e100e5853444114029116255ebfb7a54e5ed51503ba969740ad48b5ca80f70
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:1a9a19df17b464d1708b2905d03b8726e4480ca1f11aa97a15ff4c049a24ea4b
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:d1c1490d4c20dc2ec31effa1af067676c4d264e3b7d4670fc11c2baf8a400001
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:611556d999ea97f9dbb78fec5963baa02de7a26626d7096ba63b177e3bcf9707
        version: 2.1.1
      entries:
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.1
    - currentCSV: kernel-module-management-hub.v2.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:90c1946b3f577cde0435402b04baa52dc54e5aab3bb84bd834ac026a57453296
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.2.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:b1d98c0f14448ae8eb80821cbec236150c338674aae2a2b37f8d0ab0d5f1430c
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:53fddf0b226eb579c71c86112a8d9b6b1d90a1498d6f6985932b16b8be532f35
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:c3dc9ad4ebc68cd85a074b520903a9310e72a90a73bdf4aa7c42b7ed66b7bcba
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:90c1946b3f577cde0435402b04baa52dc54e5aab3bb84bd834ac026a57453296
        version: 2.2.1
      entries:
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.2
    - currentCSV: kernel-module-management-hub.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:f1d8f3c95f621c3faf41867c5dd4bc13fa2b9b291422d6184e6e42f1d1460259
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.3.0'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:f1d8f3c95f621c3faf41867c5dd4bc13fa2b9b291422d6184e6e42f1d1460259
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:c5005b74402f680d11c9ea682a9a0d3f74b6ab0d01d25298d232f910f1a29c34
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:f9fde334af68220943017a7491130e6a82f15e63522f5bb3931aa7d9b00ccc40
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:06ac089cc89b303e525fd71d6439b958f34ab944a72209f1cc8cb562ecfd1112
        version: 2.3.0
      entries:
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.3
    - currentCSV: kernel-module-management-hub.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm-hub",
                    "app.kubernetes.io/name": "kmm-hub",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:2efb81137ad2981b994c62d460c917749308133a7dcad5b9a2b251e18f3052ec
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        version: 2.4.1
      entries:
      - name: kernel-module-management-hub.v2.4.1
        version: 2.4.1
      - name: kernel-module-management-hub.v2.4.0
        version: 2.4.0
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.4
    - currentCSV: kernel-module-management-hub.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm-hub",
                    "app.kubernetes.io/name": "kmm-hub",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:2efb81137ad2981b994c62d460c917749308133a7dcad5b9a2b251e18f3052ec
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        version: 2.4.1
      entries:
      - name: kernel-module-management-hub.v2.4.1
        version: 2.4.1
      - name: kernel-module-management-hub.v2.4.0
        version: 2.4.0
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: kernel-module-management-hub
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.17.40-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5-beta.1 <8.17.40-beta.1'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBM Turbonomic software automates and simultaneously optimizes
              performance, compliance, and cost in real time. Turbonomic discovers
              and automates resource optimization for the complete application stack.
              Applications are continually resourced to perform while satisfying technical
              and business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
          - kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Optimization for Kubernetes
          IBM Turbonomic software automates and simultaneously optimizes performance, compliance, and cost in real time.
          Turbonomic discovers and automates resource optimization for the complete application stack. Applications are continually resourced to perform while satisfying technical and business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (proactive rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          Turbonomic assures application performance by giving workloads the resources they need when they need them.

          ### How Does It Work?
          Turbonomic uses a client agent called — KubeTurbo — that runs in your cluster to discover, monitor and optimize resource usage within your environment.
          KubeTurbo runs together with the default Kubernetes scheduler and sends data back to the Turbonomic for analysis.
          Turbonomic determines the right actions that drive continuous health for your workloads and your cluster.

          Turbonomic discovers Red Hat OpenShift Virtualization resources through a new or existing target configuration with Red Hat OpenShift. For more information refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=targets-red-hat-openshift-virtualization)

          ### Get Started
          Please follow the [Kubeturbo Documentation Page](https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets) for all deployment options and an overview of typical use cases.

          ### Known Issues
          For known issues refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=notes-known-issues)
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Kubeturbo Documentation Page
          url: https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets
        - name: Kubeturbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
        - name: IBM Turbonomic Application Resource Management
          url: https://www.ibm.com/products/turbonomic
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/kubeturbo-operator@sha256:7111c7c4002a447dc6977dbb73e96507bae56766fc0552fb2d036b20583722ba
        version: 8.17.40-beta.1
      entries:
      - name: kubeturbo-operator.v8.17.40-beta.1
        version: 8.17.40-beta.1
      name: beta
    - currentCSV: kubeturbo-operator.v8.17.40
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5 <8.17.40'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBM Turbonomic software automates and simultaneously optimizes
              performance, compliance, and cost in real time. Turbonomic discovers
              and automates resource optimization for the complete application stack.
              Applications are continually resourced to perform while satisfying technical
              and business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
          - kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Optimization for Kubernetes
          IBM Turbonomic software automates and simultaneously optimizes performance, compliance, and cost in real time.
          Turbonomic discovers and automates resource optimization for the complete application stack. Applications are continually resourced to perform while satisfying technical and business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (proactive rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          Turbonomic assures application performance by giving workloads the resources they need when they need them.

          ### How Does It Work?
          Turbonomic uses a client agent called — KubeTurbo — that runs in your cluster to discover, monitor and optimize resource usage within your environment.
          KubeTurbo runs together with the default Kubernetes scheduler and sends data back to the Turbonomic for analysis.
          Turbonomic determines the right actions that drive continuous health for your workloads and your cluster.

          Turbonomic discovers Red Hat OpenShift Virtualization resources through a new or existing target configuration with Red Hat OpenShift. For more information refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=targets-red-hat-openshift-virtualization)

          ### Get Started
          Please follow the [Kubeturbo Documentation Page](https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets) for all deployment options and an overview of typical use cases.

          ### Known Issues
          For known issues refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=notes-known-issues)
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Kubeturbo Documentation Page
          url: https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets
        - name: Kubeturbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
        - name: IBM Turbonomic Application Resource Management
          url: https://www.ibm.com/products/turbonomic
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/kubeturbo-operator@sha256:829dc6b9800efd3fd77d07f955a53610c7b2888b93495b2db32f1f48e93a05a6
        version: 8.17.40
      entries:
      - name: kubeturbo-operator.v8.17.40
        version: 8.17.40
      name: stable
    defaultChannel: stable
    packageName: kubeturbo-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://access.redhat.com/products/red-hat-integration
    name: red-hat-camel-k
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: red-hat-camel-k-operator.v1.10.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
          createdAt: "2025-01-08T17:00:43Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
        version: 1.10.10
      entries:
      - name: red-hat-camel-k-operator.v1.10.10
        version: 1.10.10
      - name: red-hat-camel-k-operator.v1.10.9
        version: 1.10.9
      - name: red-hat-camel-k-operator.v1.10.9-0.1743522834.p
        version: 1.10.9+0.1743522834.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1742828111.p
        version: 1.10.9+0.1742828111.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741884247.p
        version: 1.10.9+0.1741884247.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741875993.p
        version: 1.10.9+0.1741875993.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741580860.p
        version: 1.10.9+0.1741580860.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739988175.p
        version: 1.10.9+0.1739988175.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739388437.p
        version: 1.10.9+0.1739388437.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739387844.p
        version: 1.10.9+0.1739387844.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.10.x
    - currentCSV: red-hat-camel-k-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1beta1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:9e4fb5369e0479d4a68f60725e7d12abdc66c552babe59f1358f6e6b0f26357d
          createdAt: "2020-10-06T02:45:00Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:9e4fb5369e0479d4a68f60725e7d12abdc66c552babe59f1358f6e6b0f26357d
        version: 1.4.2
      entries:
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.4.x
    - currentCSV: red-hat-camel-k-operator.v1.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
          createdAt: "2022-08-15T20:17:25Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <=1.6.8'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
        version: 1.6.10
      entries:
      - name: red-hat-camel-k-operator.v1.6.10
        version: 1.6.10
      - name: red-hat-camel-k-operator.v1.6.9
        version: 1.6.9
      name: 1.6.x
    - currentCSV: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:e916f9040d3f4feb26cc1534d3ab2722890b34c7fe146cdce8b6085b47206e22
          createdAt: "2022-11-28T10:43:18Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <1.8.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:e916f9040d3f4feb26cc1534d3ab2722890b34c7fe146cdce8b6085b47206e22
        version: 1.8.2+0.1675913507.p
      entries:
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.8.x
    - currentCSV: red-hat-camel-k-operator.v1.10.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
          createdAt: "2025-01-08T17:00:43Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
        version: 1.10.10
      entries:
      - name: red-hat-camel-k-operator.v1.10.10
        version: 1.10.10
      - name: red-hat-camel-k-operator.v1.10.9
        version: 1.10.9
      - name: red-hat-camel-k-operator.v1.10.9-0.1743522834.p
        version: 1.10.9+0.1743522834.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1742828111.p
        version: 1.10.9+0.1742828111.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741884247.p
        version: 1.10.9+0.1741884247.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741875993.p
        version: 1.10.9+0.1741875993.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741580860.p
        version: 1.10.9+0.1741580860.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739988175.p
        version: 1.10.9+0.1739988175.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739388437.p
        version: 1.10.9+0.1739388437.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739387844.p
        version: 1.10.9+0.1739387844.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: candidate
    - currentCSV: red-hat-camel-k-operator.v1.10.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
          createdAt: "2025-01-08T17:00:43Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
        version: 1.10.10
      entries:
      - name: red-hat-camel-k-operator.v1.10.10
        version: 1.10.10
      - name: red-hat-camel-k-operator.v1.10.9
        version: 1.10.9
      - name: red-hat-camel-k-operator.v1.10.9-0.1743522834.p
        version: 1.10.9+0.1743522834.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1742828111.p
        version: 1.10.9+0.1742828111.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741884247.p
        version: 1.10.9+0.1741884247.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741875993.p
        version: 1.10.9+0.1741875993.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741580860.p
        version: 1.10.9+0.1741580860.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739988175.p
        version: 1.10.9+0.1739988175.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739388437.p
        version: 1.10.9+0.1739388437.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739387844.p
        version: 1.10.9+0.1739387844.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: latest
    - currentCSV: red-hat-camel-k-operator.v1.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
          createdAt: "2022-08-15T20:17:25Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <=1.6.8'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
        version: 1.6.10
      entries:
      - name: red-hat-camel-k-operator.v1.6.10
        version: 1.6.10
      - name: red-hat-camel-k-operator.v1.6.9
        version: 1.6.9
      name: stable
    defaultChannel: 1.10.x
    packageName: red-hat-camel-k
    provider:
      name: Red Hat
      url: https://access.redhat.com/products/red-hat-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MySQL
      provider-url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
    name: ndb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ndb-operator.v1.6.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "mysql.oracle.com/v1","kind": "NdbCluster","metadata":
            {"name": "example-ndb"},"spec": {"redundancyLevel": 2,"dataNode": {"nodeCount":
            2},"mysqlNode": {"nodeCount":2}}}]'
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: container-registry.oracle.com/mysql/community-ndb-operator@sha256:b562dc3303ffae9f4dab1d73fd0fe05833bdaa8b731af4227da81af617c75c6b
          createdAt: "2024-12-18T15:05:46.000Z"
          description: The MySQL NDB Operator is a Kubernetes operator for managing
            a MySQL NDB Cluster setup inside a Kubernetes Cluster.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Ndb Cluster
            displayName: NdbCluster
            kind: NdbCluster
            name: ndbclusters.mysql.oracle.com
            version: v1
        description: |
          ## NdbCluster Custom Resource
          The NdbCluster custom resource is defined with custom fields that map to the configuration of a MySQL Cluster. To deploy a MySQL Cluster inside the K8s Cluster, an NdbCluster object with the desired configuration of the MySQL Cluster has to be created in the K8s Cluster. The NDB Operator watches for any changes made to any NdbCluster objects and will receive an event when the NdbCluster object is created in the K8s Cluster. It will then start deploying the MySQL Cluster, with the specified configuration, inside the K8s Cluster using native Kubernetes workloads and resources. Once the MySQL Cluster is ready, the NdbCluster resource can be further modified to make changes to the MySQL Cluster configuration. The Operator will receive this event as well and will update the configuration of the already running MySQL Cluster.

          The documentation at [NdbCluster-CRD.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md) explains the various custom fields of the NdbCluster CRD and the effect they have on the MySQL Cluster configuration. The [examples](https://github.com/mysql/mysql-ndb-operator/tree/main/docs/examples) folder has a few examples of the NdbCluster resource objects which can be used to deploy MySQL Cluster inside the K8s Cluster.
          ## Deploy a simple MySQL Cluster

          The [examples/example-ndb.yaml](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/examples/example-ndb.yaml) file has a NdbCluster object example-ndb that defines a simple MySQL Cluster with 2 data nodes and 2 MySQL Servers. To create this object in the default namespace of the K8s Cluster, run this command from the root of the repo :
          ```
          kubectl apply -f docs/examples/example-ndb.yaml
          ```

          Various NdbCluster status fields report back the actual status of the MySQL Cluster running inside the K8s Cluster. Wait for the MySQL Cluster nodes to be started and become ready using the [UpToDate](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md#ndbclusterconditiontypestring-alias) Condition defined in the NdbCluster object status :
          ```
          kubectl wait --for=condition=UpToDate ndb example-ndb --timeout=10m
          ```
          The MySQL Cluster is ready for transactions once the NdbCluster is UpToDate, i.e. the UpToDate condition becomes true.

          At any point, a brief status of an NdbCluster object can be viewed using kubectl get ndb command.
          ```
          kubectl get ndb example-ndb
          ```

          The output of the command gives a short summary the MySQL Cluster nodes controlled by the respective NdbCluster object.
          ```
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE   UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:2/2    Ready:2/2 3m    True
          ```

          To list all the pods created by the NDB Operator, run :
          ```
          kubectl get pods -l mysql.oracle.com/v1=example-ndb
          ```

          For example-ndb NdbCluster object, two management nodes (*-mgmd-* pods), two multi-threaded data nodes (*-ndbmtd-* pods) and two MySQL Servers (*-mysqld-*) would have been created by the NDB operator :
          ```
          NAME                   READY   STATUS    RESTARTS   AGE
          example-ndb-mgmd-0     1/1     Running   0          3m53s
          example-ndb-mgmd-1     1/1     Running   0          2m5s
          example-ndb-mysqld-0   1/1     Running   0          78s
          example-ndb-mysqld-1   1/1     Running   0          78s
          example-ndb-ndbmtd-0   1/1     Running   0          3m53s
          example-ndb-ndbmtd-1   1/1     Running   0          3m53s
          ```
          If running inside a openshift cluster, pods from statefulset "example-ndb-mysqld" will not start up. By default openshift runs all the pod with "restricted" scc, but the statefulset "example-ndb-mysqld" required higher privilege. Use the below command to check the exact scc required
          ```
          oc get pods example-ndb-mysqld-0 -o yaml | oc adm policy scc-subject-review -f -
          ```
          Create a user account with the scc and set the user to the statefulset "example-ndb-mysqld". Below commands creates an user with "anyuid" and assign it to the statefulset.
          ```
          oc create sa any-uid-user
          oc adm policy add-scc-to-user anyuid -z any-uid-user
          oc set sa sts example-ndb-mysqld any-uid-user
          ```
          ## Connect to the MySQL Cluster

          The NDB Operator, by default, creates few Services to expose the services offered by the MySQL Cluster nodes within the Kubernetes Cluster.
          ```
          kubectl get services -l mysql.oracle.com/v1=example-ndb
          ```

          The above command will generate an output similar to :
          ```
          NAME                 TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S) AGE
          example-ndb-mgmd     ClusterIP   10.100.207.255   <none>        1186/TCP 5m
          example-ndb-mysqld   ClusterIP   10.103.182.157   <none>        3306/TCP 5m
          example-ndb-ndbmtd   ClusterIP   None             <none>        1186/TCP 5m
          ```

          The NDB Operator will also create a root user with a random password on all the MySQL Servers. The root password will be stored in a K8s Secret which can be retrieved from the NdbCluster object status :
          ```
          kubectl get ndb example-ndb -o jsonpath={.status.generatedRootPasswordSecretName}
          ```

          To extract and decode the password from the Secret :
          ```
          base64 -d <<< \
          $(kubectl get secret example-ndb-mysqld-root-password \
          -o jsonpath={.data.password})
          ```

          One can also specify the password and host to be used with the root account via the NdbCluster spec. See the [CRD documentation](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md#ndbmysqldspec) for more details on this.
          ## Access MySQL Cluster from inside K8s

          To connect to the Management and MySQL Servers from within the Kubernetes Cluster, you can use the example-ndb-mgmd and example-ndb-mysqld services as the MySQL Cluster connectstring and MySQL host respectively. You can also use the mgmd and mysqld pods' hostnames to directly connect to that particular pod. The mysql, ndb_mgm clients and any other ndb tools will work as expected.

          A demonstration is available at [connect-from-inside-k8s-demo.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/connect-from-inside-k8s-demo.md).
          ## Access MySQL Cluster from outside K8s

          By default, the Management and MySQL services created by the NDB Operator are of type [ClusterIP](https://kubernetes.io/docs/concepts/services-networking/service/#publishing-services-service-types) and are accessible only from within the K8s Cluster. To expose them outside the K8s Cluster, the services have to be upgraded into a LoadBalancer type. This can be done by setting the spec.managementNode.enableLoadBalancer and spec.mysqlNode.enableLoadBalancer fields of the NdbCluster resource to true. These options can be enabled when the NdbCluster resource object is created or via an update to the object when the MySQL Cluster is already running.

          Once enabled, the previously ClusterIP type example-ndb-mgmd and example-ndb-mysqld services will be upgraded to LoadBalancer type, and they will be available at the external IP assigned to them by the cloud provider.

          Another way to access these services without enabling the LoadBalancers support is to use the kubectl port-forward command.

          In both ways, only the mysql and ndb_mgm clients work. Any NDB tool which uses the NDBAPI to connect to the MySQL data nodes will not work as expected from outside the K8s Cluster.

          A demonstration is available at [connect-from-outside-k8s-demo.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/connect-from-outside-k8s-demo.md).
          ## Updating MySQL Cluster Configuration

          Once a MySQL Cluster has been deployed by the NDB Operator inside the K8s Cluster, its configuration can be updated by updating the spec of the NdbCluster resource object that represents it. The NDB Operator will pick up any changes made to the NdbCluster resource object and will apply the updated configuration to the respective MySQL Cluster. The NDB Operator takes care of updating the MySQL Cluster config file and restarting all the management/data nodes if required. When the update is being handled by the NdbOperator, it will set the UpToDate condition of the NdbCluster resource object to false and will not accept any further updates until it completes the current one.

          The kubectl get ndb example-ndb command will report the UpToDate status of an NdbCluster. It can be used to check if the current update has been completed.
          ```
          $ kubectl get ndb example-ndb
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE   UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:1/2    Ready:2/2 10m   False
          ```

          The UpToDate condition can also be used to wait for the operator to complete the update.
          ```
          kubectl wait --for=condition=UpToDate ndb example-ndb --timeout=10m
          ```

          Once the update is complete, the UpToDate condition will be set back to true by the operator.
          ```
          $ kubectl get ndb example-ndb
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE      UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:2/2    Ready:2/2 10m50s   True
          ```

          ## Delete a MySQL Cluster

          To stop and remove the MySQL Cluster running inside the K8s Cluster, delete the NdbCluster resource object.
          ```
          kubectl delete -f docs/examples/example-ndb.yaml
          ```

          (or)
          ```
          kubectl delete ndb example-ndb
          ```

          This will delete all the data, the pods running the MySQL Cluster nodes and also delete all other associated K8s resources created by the NDB Operator.
        displayName: Ndb Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - ndb
        - in-memory
        links:
        - name: Document
          url: https://dev.mysql.com/doc/ndb-operator/en/
        maintainers:
        - email: paulo.p.silva@oracle.com
          name: paulofasilva
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: MySQL
          url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
        relatedImages:
        - container-registry.oracle.com/mysql/community-ndb-operator@sha256:b562dc3303ffae9f4dab1d73fd0fe05833bdaa8b731af4227da81af617c75c6b
        version: 1.6.2
      entries:
      - name: ndb-operator.v1.6.2
        version: 1.6.2
      - name: ndb-operator.v1.6.1
        version: 1.6.1
      - name: ndb-operator.v1.6.0
        version: 1.6.0
      name: alpha
    defaultChannel: alpha
    packageName: ndb-operator
    provider:
      name: MySQL
      url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: VASTData
      provider-url: https://www.vastdata.com
    name: vast-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vast-csi-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"storage.vastdata.com/v1","kind":"VastCSIDriver","metadata":{"name":"csi.vastdata.com"},"spec":{"applySecurityContextConstraints":true,"attachRequired":true,"controller":{"dnsPolicy":"Default","extraArgs":{"csiAttacher":[],"csiProvisioner":[],"csiResizer":[],"csiSnapshotter":[]},"nodeAffinity":{},"nodeSelector":{},"podAffinity":{},"podAntiAffinity":{},"resources":{"csiAttacher":{"limits":{"memory":"200Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiProvisioner":{"limits":{"memory":"400Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiResizer":{"limits":{"memory":"300Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiSnapshotter":{"limits":{"memory":"200Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiVastPlugin":{"limits":{"memory":"400Mi"},"requests":{"cpu":"100m","memory":"50Mi"}}},"runOnControlPlane":false,"runOnMaster":false,"tolerations":[]},"deletionViewPolicy":null,"deletionVipPool":null,"dontUseTrashApi":false,"driverType":"nfs","endpoint":null,"image":{"csiAttacher":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-attacher@sha256:b4d611100ece2f9bc980d1cb19c2285b8868da261e3b1ee8f45448ab5512ab94"},"csiNodeDriverRegistrar":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac"},"csiProvisioner":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-provisioner@sha256:405a14e1aa702f7ea133cea459e8395fe40a6125c088c55569e696d48e1bd385"},"csiResizer":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-resizer@sha256:a541e6cc2d8b011bb21b1d4ffec6b090e85270cce6276ee302d86153eec0af43"},"csiSnapshotter":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-snapshotter@sha256:2e04046334baf9be425bb0fa1d04c2d1720d770825eedbdbcdb10d430da4ad8c"},"csiVastPlugin":{"imagePullPolicy":"IfNotPresent","repository":"registry.connect.redhat.com/vastdataorg/csi@sha256:7169f279d0dd7130535d74df5b938ed6b7944bfd4fe52d242ceee5178f2ea156"}},"imagePullSecrets":[],"kubeletPath":"/var/lib/kubelet","logLevel":5,"node":{"dnsPolicy":"Default","nodeAffinity":{},"nodeSelector":{},"podAffinity":{},"podAntiAffinity":{},"propagateHostMountOptions":true,"resources":{"csiVastPlugin":{"limits":{"memory":"400Mi"},"requests":{"cpu":"100m","memory":"50Mi"}},"nodeDriverRegistrar":{"limits":{"memory":"100Mi"},"requests":{"cpu":"10m","memory":"20Mi"}}},"tolerations":[]},"numWorkers":10,"operationRetryIntervalMax":60,"operationRetryIntervalStart":10,"operationTimeout":15,"secretName":null,"sslCertsSecretName":null,"truncateVolumeName":64,"useLocalIpForMount":null,"verifySsl":false}},{"apiVersion":"storage.vastdata.com/v1","kind":"VastStorage","metadata":{"name":"vastdata-filesystem"},"spec":{"allowVolumeExpansion":true,"blockingClones":false,"clusterName":"","createSnapshotClass":true,"driverType":"nfs","ephemeralVolumeNameFormat":"csi:{namespace}:{name}:{id}","fsType":"","hostEncryption":null,"mountOptions":[],"provisioner":"","qosPolicy":"","reclaimPolicy":"Delete","secretName":"","secretNamespace":"","setDefaultStorageClass":false,"snapshotClass":{"deletionPolicy":"Delete","setDefaultSnapshotClass":false,"snapshotNameFormat":"csi:{namespace}:{name}:{id}"},"storagePath":"","subsystem":"","tenantName":"","viewPolicy":"","vipPool":"","vipPoolFQDN":null,"volumeGroup":"","volumeNameFormat":"csi:{namespace}:{name}:{id}"}},{"apiVersion":"storage.vastdata.com/v1","kind":"VastCluster","metadata":{"name":"vastcluster-sample"},"spec":{"endpoint":null,"passphrase":null,"password":null,"sslCert":null,"tenant":null,"token":null,"username":null}}]'
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/vastdataorg/csi-operator@sha256:ba7f30a42e327004ecbc81ff5ccc453bf3046a29bab5785b9f8a9746e5c6a3fd
          createdAt: "2024-07-01T11:59:59Z"
          description: |-
            VAST CSI Operator supports seamless installation and upgrade of VAST Data's CSI Driver.
            The VAST CSI Driver allows container orchestration frameworks such as Kubernetes to dynamically provision, manage and scale volumes on the [VAST Data Platform](https://www.vastdata.com/platform/overview). VAST provides scalable, reliable, and fast persistent storage that can be accessed remotely by any Kubernetes application container.
            ### Installation
            Refer to the VASTData Operator for Kubernetes [official documentation](https://support.vastdata.com/s/document-item?bundleId=vast-csi-driver-2.6-administrator-s-guide&topicId=deploying-vast-csi-driver/deploying-vast-csi-driver-on-openshift-with-vast-csi-operator.html&_LANG=enus).
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: vast-csi
          operators.operatorframework.io/builder: operator-sdk-v1.3.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/vast-data/vast-csi
          support: VastData Infra Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VastCSIDriver is a common specification for VAST CSI Controller
              and VAST CSI Node plugins, typically intended for creation in a single
              instance.
            displayName: VastCSIDriver
            kind: VastCSIDriver
            name: vastcsidrivers.storage.vastdata.com
            version: v1
          - description: VastCluster encapsulates the connection details required
              for accessing a VAST cluster. This information is ultimately stored
              in a Kubernetes Secret.
            displayName: VastCluster
            kind: VastCluster
            name: vastclusters.storage.vastdata.com
            version: v1
          - description: Represents VAST storage. Each VastStorage instance generates
              a corresponding storage class that can be used in PVCs to create workloads.
            displayName: VastStorage
            kind: VastStorage
            name: vaststorages.storage.vastdata.com
            version: v1
        description: |-
          VAST CSI Operator supports seamless installation and upgrade of VAST Data's CSI Driver.
          The VAST CSI Driver allows container orchestration frameworks such as Kubernetes to dynamically provision, manage and scale volumes on the [VAST Data Platform](https://www.vastdata.com/platform/overview). VAST provides scalable, reliable, and fast persistent storage that can be accessed remotely by any Kubernetes application container.
          ### Installation
          Refer to the VASTData Operator for Kubernetes [official documentation](https://support.vastdata.com/s/document-item?bundleId=vast-csi-driver-2.6-administrator-s-guide&topicId=deploying-vast-csi-driver/deploying-vast-csi-driver-on-openshift-with-vast-csi-operator.html&_LANG=enus).
        displayName: VAST CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - VAST
        - VASTData
        - CSI
        - Multi Cluster
        - AI
        - High-Performance Computing
        links:
        - name: VAST CSI Plugin
          url: https://github.com/vast-data/vast-csi
        maintainers:
        - email: volodymyr.boiko@vastdata.com
          name: VAST Infrastructure Team
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: VASTData
          url: https://www.vastdata.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.connect.redhat.com/vastdataorg/csi-operator@sha256:ba7f30a42e327004ecbc81ff5ccc453bf3046a29bab5785b9f8a9746e5c6a3fd
        - registry.connect.redhat.com/vastdataorg/csi@sha256:7169f279d0dd7130535d74df5b938ed6b7944bfd4fe52d242ceee5178f2ea156
        - registry.k8s.io/sig-storage/csi-resizer@sha256:a541e6cc2d8b011bb21b1d4ffec6b090e85270cce6276ee302d86153eec0af43
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:2e04046334baf9be425bb0fa1d04c2d1720d770825eedbdbcdb10d430da4ad8c
        - registry.k8s.io/sig-storage/csi-attacher@sha256:b4d611100ece2f9bc980d1cb19c2285b8868da261e3b1ee8f45448ab5512ab94
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:405a14e1aa702f7ea133cea459e8395fe40a6125c088c55569e696d48e1bd385
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac
        version: 2.6.3
      entries:
      - name: vast-csi-operator.v2.6.3
        version: 2.6.3
      name: stable
    defaultChannel: stable
    packageName: vast-csi-operator
    provider:
      name: VASTData
      url: https://www.vastdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtc-operator.v1.7.19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigrationController",
                  "metadata": {
                    "name": "migration-controller",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "azure_resource_group": "",
                    "cluster_name": "host",
                    "migration_velero": true,
                    "migration_controller": true,
                    "migration_ui": true,
                    "migration_log_reader": true,
                    "olm_managed": true,
                    "restic_timeout": "1h",
                    "mig_pv_limit": "100",
                    "mig_pod_limit": "100",
                    "mig_namespace_limit": "10"
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageMigration",
                  "metadata": {
                    "name": "directimagemigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageStreamMigration",
                  "metadata": {
                    "name": "directimagestreammigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigAnalytic",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "plan": {
                      "name": "foo",
                      "namespace": "openshift-migration"
                    },
                    "analyzeK8SResources": true,
                    "analyzeImageCount": true,
                    "analyzePVCapacity": true
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigCluster",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectVolumeMigration",
                  "metadata": {
                    "name": "direct",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                      "kind": "DirectVolumeMigrationProgress",
                      "metadata": {
                        "name": "direct-progress",
                        "namespace": "openshift-migration"
                      },
                      "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigPlan",
                  "metadata": {
                    "name": "migplan",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigMigration",
                  "metadata": {
                    "name": "migmigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigStorage",
                  "metadata": {
                    "name": "migstorage",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:89ba5ccb1cd9c44b8c8d2189f1fc6b04e780d46b19ee8a22e20c0b8c9d94d91f
          createdAt: "2025-04-28T19:55:38Z"
          description: Facilitates migration of container workloads from OpenShift
            3.x to OpenShift 4.x
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.5.0 <1.7.19'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-migration
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:3b6168c6df0d86d76ae4de91755c631b7f6a1f84286257d18c97d5a6af312291
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/mig-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: a request to migrate PVCs from a set of namespaces directly
              to a target
            displayName: DirectVolumeMigration
            kind: DirectVolumeMigration
            name: directvolumemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate watch the progress of rsync pod migrating
              a PVC to a target
            displayName: DirectVolumeMigrationProgress
            kind: DirectVolumeMigrationProgress
            name: directvolumemigrationprogresses.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a set of namespaces
              directly to a target
            displayName: DirectImageMigration
            kind: DirectImageMigration
            name: directimagemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a single imagestream
              directly to a target
            displayName: DirectImageStreamMigration
            kind: DirectImageStreamMigration
            name: directimagestreammigrations.migration.openshift.io
            version: v1alpha1
          - description: A request to Analyze a MigPlan and its namespaces
            displayName: MigAnalytic
            kind: MigAnalytic
            name: miganalytics.migration.openshift.io
            version: v1alpha1
          - description: A cluster defined for migration
            displayName: MigCluster
            kind: MigCluster
            name: migclusters.migration.openshift.io
            version: v1alpha1
          - description: A migration process
            displayName: MigMigration
            kind: MigMigration
            name: migmigrations.migration.openshift.io
            version: v1alpha1
          - description: A migration plan
            displayName: MigPlan
            kind: MigPlan
            name: migplans.migration.openshift.io
            version: v1alpha1
          - description: A migration contoller deployment
            displayName: MigrationController
            kind: MigrationController
            name: migrationcontrollers.migration.openshift.io
            version: v1alpha1
          - description: A migration hook
            displayName: MigHook
            kind: MigHook
            name: mighooks.migration.openshift.io
            version: v1alpha1
          - description: A migration storage location
            displayName: MigStorage
            kind: MigStorage
            name: migstorages.migration.openshift.io
            version: v1alpha1
        description: |-
          The Migration Toolkit for Containers Operator enables installation of the application migration tool components.

          After you have installed the Operator, you can install components by creating a MigrationController CR.

          By default, the Operator installs the following components on an OpenShift Container Platform 4 target cluster:

          * velero and restic, to back up data on the source cluster and restore it on the target cluster
          * mig-controller, to coordinate the backup and restore processes
          * mig-ui, the web console for managing migrations

          If you experience any issues or have feature requests, please file an [issue](https://github.com/konveyor/mig-operator/issues)

          For more information, see the [OpenShift Container Platform Migration Guide](https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html).
        displayName: Migration Toolkit for Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - crane
        - mtc
        links:
        - name: Migrations
          url: https://github.com/konveyor/mig-controller
        - name: Documentation
          url: https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html
        - name: Crane Operator
          url: https://github.com/konveyor/mig-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhmtc/openshift-migration-hook-runner-rhel8@sha256:c8c932a486d9b83e5a93290058d9e84691b87cd4161f8132a1bb6de9a360605e
        - registry.redhat.io/rhmtc/openshift-migration-velero-restic-restore-helper-rhel8@sha256:932216fd08d8ad97baf48ac7b8c989806d30a191337c650df37da537c153dc17
        - registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:3b6168c6df0d86d76ae4de91755c631b7f6a1f84286257d18c97d5a6af312291
        - registry.redhat.io/rhmtc/openshift-migration-velero-rhel8@sha256:bc010bc31e8b553f86aacbca39e28a95e74994c56e85c9fc1b8b80d0791ecb16
        - registry.redhat.io/rhmtc/openshift-migration-log-reader-rhel8@sha256:a0c8ab3a5bbc78c158c2ccccb1d06430a20f40358649196aa4e7311e54f9bdf8
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-aws-rhel8@sha256:023465788745625ad3b4dc0574e35b93d6ac8eea17be2ad1fc88959149a0df2c
        - registry.redhat.io/rhmtc/openshift-migration-ui-rhel8@sha256:66b288170dca8007382d13dff3f2e1868e409fb8002e04ec1fc5c480f868d9b6
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-gcp-rhel8@sha256:cc70c5a63af77948760bc42bf695a5e4eb0bce6285d6b350bba89a53f4b4e7e9
        - registry.redhat.io/rhmtc/openshift-migration-registry-rhel8@sha256:49485dfe708fe1294dec3f0bf34fedc38845a22f91915bba523d728a586b0c30
        - registry.redhat.io/rhmtc/openshift-migration-controller-rhel8@sha256:9b1490b04fc235191515fcd07d7af2e81f846d865016831e17872bb82bf4eaf5
        - registry.redhat.io/rhmtc/openshift-migration-rsync-transfer-rhel8@sha256:23a2bc5ad0f99fc0e7517a0172fdf3597fa619543b0192757983219a578f4ee2
        - registry.redhat.io/rhmtc/openshift-velero-plugin-rhel8@sha256:02698c655a01df2237aee6c570db96f13c0018cee2ae58b96b9570d93c354617
        - registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:89ba5ccb1cd9c44b8c8d2189f1fc6b04e780d46b19ee8a22e20c0b8c9d94d91f
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-microsoft-azure-rhel8@sha256:dad8fa4dbce3b4562bf06756fd74c2007812ec4edbc308e4517ff123c5519767
        version: 1.7.19
      entries:
      - name: mtc-operator.v1.7.19
        version: 1.7.19
      - name: mtc-operator.v1.7.18
        version: 1.7.18
      - name: mtc-operator.v1.7.17
        version: 1.7.17
      - name: mtc-operator.v1.7.16
        version: 1.7.16
      - name: mtc-operator.v1.7.15
        version: 1.7.15
      - name: mtc-operator.v1.7.14
        version: 1.7.14
      - name: mtc-operator.v1.7.13
        version: 1.7.13
      - name: mtc-operator.v1.7.12
        version: 1.7.12
      - name: mtc-operator.v1.7.11
        version: 1.7.11
      - name: mtc-operator.v1.7.10
        version: 1.7.10
      - name: mtc-operator.v1.7.9
        version: 1.7.9
      - name: mtc-operator.v1.7.8
        version: 1.7.8
      - name: mtc-operator.v1.7.7
        version: 1.7.7
      - name: mtc-operator.v1.7.6
        version: 1.7.6
      - name: mtc-operator.v1.7.5
        version: 1.7.5
      - name: mtc-operator.v1.7.4
        version: 1.7.4
      - name: mtc-operator.v1.7.3
        version: 1.7.3
      - name: mtc-operator.v1.7.2
        version: 1.7.2
      - name: mtc-operator.v1.7.1
        version: 1.7.1
      - name: mtc-operator.v1.7.0
        version: 1.7.0
      name: release-v1.7
    - currentCSV: mtc-operator.v1.8.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigrationController",
                  "metadata": {
                    "name": "migration-controller",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "azure_resource_group": "",
                    "cluster_name": "host",
                    "migration_velero": true,
                    "migration_controller": true,
                    "migration_ui": true,
                    "migration_log_reader": true,
                    "olm_managed": true,
                    "restic_timeout": "1h",
                    "mig_pv_limit": "100",
                    "mig_pod_limit": "100",
                    "mig_namespace_limit": "10"
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageMigration",
                  "metadata": {
                    "name": "directimagemigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageStreamMigration",
                  "metadata": {
                    "name": "directimagestreammigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigAnalytic",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "plan": {
                      "name": "foo",
                      "namespace": "openshift-migration"
                    },
                    "analyzeK8SResources": true,
                    "analyzeImageCount": true,
                    "analyzePVCapacity": true
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigCluster",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectVolumeMigration",
                  "metadata": {
                    "name": "direct",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                      "kind": "DirectVolumeMigrationProgress",
                      "metadata": {
                        "name": "direct-progress",
                        "namespace": "openshift-migration"
                      },
                      "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigPlan",
                  "metadata": {
                    "name": "migplan",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigMigration",
                  "metadata": {
                    "name": "migmigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigStorage",
                  "metadata": {
                    "name": "migstorage",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:8a44fe2aea3e842dcccb1e20df4818f4ca573fe13f51427ae8b603d0000a9411
          createdAt: "2025-08-22T15:27:29Z"
          description: Facilitates migration of container workloads from OpenShift
            3.x to OpenShift 4.x
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.5.0 <1.8.10'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-migration
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:f9492269b75608057fbc5edd53603807f1f80cd8581dea12032c88e04cf8f4cb
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/mig-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: a request to migrate PVCs from a set of namespaces directly
              to a target
            displayName: DirectVolumeMigration
            kind: DirectVolumeMigration
            name: directvolumemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate watch the progress of rsync pod migrating
              a PVC to a target
            displayName: DirectVolumeMigrationProgress
            kind: DirectVolumeMigrationProgress
            name: directvolumemigrationprogresses.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a set of namespaces
              directly to a target
            displayName: DirectImageMigration
            kind: DirectImageMigration
            name: directimagemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a single imagestream
              directly to a target
            displayName: DirectImageStreamMigration
            kind: DirectImageStreamMigration
            name: directimagestreammigrations.migration.openshift.io
            version: v1alpha1
          - description: A request to Analyze a MigPlan and its namespaces
            displayName: MigAnalytic
            kind: MigAnalytic
            name: miganalytics.migration.openshift.io
            version: v1alpha1
          - description: A cluster defined for migration
            displayName: MigCluster
            kind: MigCluster
            name: migclusters.migration.openshift.io
            version: v1alpha1
          - description: A migration process
            displayName: MigMigration
            kind: MigMigration
            name: migmigrations.migration.openshift.io
            version: v1alpha1
          - description: A migration plan
            displayName: MigPlan
            kind: MigPlan
            name: migplans.migration.openshift.io
            version: v1alpha1
          - description: A migration contoller deployment
            displayName: MigrationController
            kind: MigrationController
            name: migrationcontrollers.migration.openshift.io
            version: v1alpha1
          - description: A migration hook
            displayName: MigHook
            kind: MigHook
            name: mighooks.migration.openshift.io
            version: v1alpha1
          - description: A migration storage location
            displayName: MigStorage
            kind: MigStorage
            name: migstorages.migration.openshift.io
            version: v1alpha1
        description: |-
          The Migration Toolkit for Containers Operator enables installation of the application migration tool components.

          After you have installed the Operator, you can install components by creating a MigrationController CR.

          By default, the Operator installs the following components on an OpenShift Container Platform 4 target cluster:

          * velero and restic, to back up data on the source cluster and restore it on the target cluster
          * mig-controller, to coordinate the backup and restore processes
          * mig-ui, the web console for managing migrations

          If you experience any issues or have feature requests, please file an [issue](https://github.com/konveyor/mig-operator/issues)

          For more information, see the [OpenShift Container Platform Migration Guide](https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html).
        displayName: Migration Toolkit for Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - crane
        - mtc
        links:
        - name: Migrations
          url: https://github.com/konveyor/mig-controller
        - name: Documentation
          url: https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html
        - name: Crane Operator
          url: https://github.com/konveyor/mig-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhmtc/openshift-migration-registry-rhel8@sha256:accef2a0dd69b254bd120232e5dc86a56aa8f25bcfdba3f9bab7e4127c157762
        - registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:f9492269b75608057fbc5edd53603807f1f80cd8581dea12032c88e04cf8f4cb
        - registry.redhat.io/rhmtc/openshift-migration-controller-rhel8@sha256:c776387c662e21aa61378d0136b2495826e1c157548f9ca9c95d044af38ee169
        - registry.redhat.io/rhmtc/openshift-migration-hook-runner-rhel8@sha256:adaa4f1afefb4d5853713821cb6e7de0d9f5a640f0243f20b41a44be84707caf
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-mtc-rhel8@sha256:5f46ba04146438f2abd6905468d2ecbeb19374a83771d364904ac2ee6afdbefa
        - registry.redhat.io/rhmtc/openshift-migration-openvpn-rhel8@sha256:d21b99d8edefa7a1bab30c85969ec7d7560d06be308bd787e698e409d54cbc19
        - registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:8a44fe2aea3e842dcccb1e20df4818f4ca573fe13f51427ae8b603d0000a9411
        - registry.redhat.io/rhmtc/openshift-migration-log-reader-rhel8@sha256:4f76bcf3df3bb7073b8aabd3da072cacfec05ecf4bfb47e23340074756f7fc30
        - registry.redhat.io/rhmtc/openshift-migration-rsync-transfer-rhel8@sha256:442e7c7f2cc075bf468403ace8846c96ad2fe43eb7f6562d566caa24e4f86873
        - registry.redhat.io/rhmtc/openshift-migration-ui-rhel8@sha256:86e747e0330b0156379d273d0c29c13a65dafbf4f536656723e2bc746872286c
        version: 1.8.10
      entries:
      - name: mtc-operator.v1.8.10
        version: 1.8.10
      - name: mtc-operator.v1.8.9
        version: 1.8.9
      - name: mtc-operator.v1.8.8
        version: 1.8.8
      - name: mtc-operator.v1.8.7
        version: 1.8.7
      - name: mtc-operator.v1.8.6
        version: 1.8.6
      - name: mtc-operator.v1.8.5
        version: 1.8.5
      - name: mtc-operator.v1.8.4
        version: 1.8.4
      - name: mtc-operator.v1.8.3
        version: 1.8.3
      - name: mtc-operator.v1.8.2
        version: 1.8.2
      - name: mtc-operator.v1.8.1
        version: 1.8.1
      - name: mtc-operator.v1.8.0
        version: 1.8.0
      name: release-v1.8
    defaultChannel: release-v1.8
    packageName: mtc-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Performance
      provider-url: ""
    name: cluster-impairment-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-impairment-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "all-impairments"
                },
                "spec": {
                  "duration": 480,
                  "egress": {
                    "bandwidth": 10000,
                    "latency": 50,
                    "loss": 0.02
                  },
                  "ingress": {
                    "bandwidth": 10000,
                    "latency": 50,
                    "loss": 0.02
                  },
                  "interfaces": [
                    "ens2f0"
                  ],
                  "linkFlapping": {
                    "downTime": 30,
                    "enable": true,
                    "upTime": 120
                  },
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "duplication"
                },
                "spec": {
                  "duration": 60,
                  "egress": {
                    "duplication": 25
                  },
                  "ingress": {
                    "duplication": 25
                  },
                  "interfaces": [
                    "ens3"
                  ],
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "two-min-flap"
                },
                "spec": {
                  "duration": 480,
                  "interfaces": [
                    "ens2f0"
                  ],
                  "linkFlapping": {
                    "downTime": 120,
                    "enable": true,
                    "upTime": 120
                  },
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "typical-scenario"
                },
                "spec": {
                  "duration": 30,
                  "egress": {
                    "latency": 50
                  },
                  "ingress": {
                    "latency": 50
                  },
                  "interfaces": [
                    "ens2f0",
                    "eno1"
                  ],
                  "loss": 0.02,
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "uneven-latency"
                },
                "spec": {
                  "duration": 60,
                  "egress": {
                    "latency": 100
                  },
                  "ingress": {
                    "latency": 10
                  },
                  "interfaces": [
                    "ens2f0"
                  ],
                  "nodeSelector": {
                    "key": "node-role.kubernetes.io/worker",
                    "value": ""
                  },
                  "startDelay": 5
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Developer Tools
          containerImage: quay.io/redhat-performance/cluster-impairment-operator:1.0.4
          createdAt: "2021-08-11 12:15:00"
          description: Run Impairments on Kubernetes/Openshift Nodes
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/redhat-performance/cluster-impairment-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterImpairment
            name: clusterimpairments.apps.redhat.com
            version: v1alpha1
        description: "cluster-impairment-operator is an operator designed to automate
          the application of impairments to the nodes in a cluster.\n\n## Features\n\nTraffic
          Direction:\n\n* Egress impairments affect the traffic going out.    \n*
          Ingress impairments affect the traffic coming in. (Ingress impairments require
          the kernel module IFB.)\n\n### Impairments\n\n| Impairment    | Description
          \                               | Unit    | Uses Netem | Correlation Supported
          |\n|---------------|--------------------------------------------|---------|------------|-----------------------|\n|
          Bandwidth     | The bandwidth limit                        | kbit/s  | Yes
          \       | No                    |\n| Latency       | The delay of the packets
          \                  | ms      | Yes        | Yes                   |\n| Packet
          Loss   | The percent of packets that are dropped    | percent | Yes        |
          Yes                   |\n| Corruption    | The percent of packets that are
          corrupted  | percent | Yes        | Yes                   |\n| Duplication
          \  | The percent of packets that are duplicated | percent | Yes        |
          Yes                   |\n| Link Flapping | Turns the interface on and off
          \            | bool    | No         | No                    |\n\n\nOn the
          tested environment (RHEL CoreOS 48.84), the impairments can be used alongside
          link flapping.\n\n\n#### Latency Options\n\nIn addition to simply delaying
          the packets, there are advanced latency options.\n\n##### Jitter\n\nSemi-randomly
          adds or subtracts from the latency according to the distribution up to the
          amount specified.\nIf latency is 100ms, and jitter is 10ms, the actual latency
          will vary 100±10ms.\n\n\n##### Distribution\n\nThe distribution of the jitter.
          The options are:\n* Normal\n* Uniform\n* Pareto\n* Paretonormal\n\n#####
          Jitter correlation\n\nThe percent chance that the next latency's jitter
          value will correlate with the preceeding latency.\n\n##### Reorder\n\nThe
          percentage of packets that are not delayed, causing a reorder due to them
          being sent before the delayed ones.\nApplying jitter itself has the potential
          to also cause reordering.\n\n##### Reorder correlation\n\nThe percet chance
          that the value for the next reorder will correlate with the preceeding value.\n\n####
          Loss options\n\n##### Correlation\n\nThe percent chance that the previous
          loss value for a packet correlates with the loss value for the next packet.\n\n####
          Corruption options\n\n###### Corruption\n\nThe percent chance that each
          packet will be corrupted.\n\n##### Corruption correlation\n\nThe percet
          chance that the value for the next corruption will correlate with the preceeding
          value.\n\n#### Duplication options\n\n##### Duplication\n\nThe percent chance
          that each packet will be duplicated.\n\n##### Duplication correlation\n\nThe
          percet chance that the value for the next duplication will correlate with
          the preceeding value.\n\n\n\n## Configuration\n\nHere is an example of the
          ClusterImpairment custom resource.\n\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: test-impairment-cr\nspec:\n  duration:
          30 # seconds\n  startDelay: 5 # seconds. It typically takes about 2-3 seconds
          for the Daemonset to run\n  interfaces:\n  - \"ens2f0\"\n  ingress: # uses
          ifb\n    bandwidth: 0 # kbit\n    latency: 10 # ms\n    latencyOptions:\n
          \     jitter: 5 # ms\n      jitterCorrelation: 25 # percent\n      distribution:
          normal\n      reorder: 25 # percent of packets that will skip the delay\n
          \     reorderCorrelation: 25 # percent\n    loss: 0 # percent\n    lossOptions:\n
          \     correlation: 25 # percent\n    corruption: 0.1 # percent\n    corruptionOptions:\n
          \     correlation: 25 # percent\n    duplication: 1 # percent\n    duplicationOptions:\n
          \     correlation: 25 # percent\n  egress:\n    bandwidth: 0 # kbit\n    latency:
          100 # ms\n    loss: 0 # percent\n  linkFlapping:\n    enable: false\n    downTime:
          3 # Seconds\n    upTime: 3 # Seconds\n  nodeSelector:\n    key: \"node-role.kubernetes.io/worker\"\n
          \   value: \"\"\n    \n```\n\n#### Interfaces\n\nThe interfaces configuration
          option is a list of all interfaces that the impairments should be applied
          to. The current implementation will error out once it hits an invalid interface.\n\nIf
          an invalid interface is found, it will print out the list of interfaces.\n\n####
          Node Selector\n\nThe default node selector is all worker nodes, but you
          can change it to whatever node selector you want by setting the key and
          value.\nThere is a toleration to allow it to run on master nodes if the
          node selector includes master nodes.\n\n#### Duration\n\nThe number of seconds
          the impairments should run. It will try to sync the start and end time between
          all pods.\nIf set to -1, it will run indefinitely (a year), until you remove
          the ClusterImpairment custom resource. This is good for testing that requires
          steady impairments.\n\nIf the script is link flapping, set this value to
          be short enough so that if the link flap interrupts communication between
          the nodes, the daemonset will remove itself.\n\n#### Start Delay\n\nThe
          delay before starting the script. If you want the pods to be in sync, a
          minimum of a few seconds should be used because the pods take time to start
          up.\n\nYou can also utilize this feature to run an separate impairment after
          another. Just apply two resources at the same time, but add the duration
          and start delay of the first to the start delay of the second.\n\n#### Ingress
          and Egress\n\nThe configuration sections \"ingress\" and \"egress\" apply
          to each direction.\n\n##### Examples:\n\n##### Example 1\n\n\nIn this example,
          egress latency is set to 100ms, but the ingress latency is set to 10ms.
          So the latency to the interface will end up being 10ms, but 100ms going
          out. When pinging, this will result in 110ms of latency.\n\n```\n\napiVersion:
          apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n  name: uneven-latency\nspec:\n
          \ duration: 60\n  startDelay: 5\n  interfaces:\n  - \"ens2f0\"\n  ingress:\n
          \   latency: 10 # ms\n  egress:\n    latency: 100 # ms\n\n```\n\n##### Example
          2\n\n\nIn this example, link flapping will be enabled, and it will turn
          the interface on and off every 2 minutes. That is long enough for kubernetes
          to determine that a node is unavailable.\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: two-min-flap\nspec:\n  duration: 480\n
          \ startDelay: 5\n  interfaces:\n  - \"ens2f0\"\n  linkFlapping:\n    enable:
          true\n    downTime: 120 # Seconds\n    upTime: 120 # Seconds\n\n```\n\n#####
          Example 3\n\n\nIn this example, a realistic set of impairments is applied
          to `ens2f0` and `eno1` for 30 seconds:\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: typical-scenario\nspec:\n  duration:
          30 # seconds\n  startDelay: 5 # seconds\n  interfaces:\n  - \"ens2f0\"\n
          \ - \"eno1\"\n  egress:\n    latency: 50 # ms. Bidirectional, so total of
          100ms\n  ingress:\n    latency: 50 # ms. Bidirectional, so total of 100ms\n
          \ loss: 0.02 # percent\n\n```\n\n##### Example 4\n\n\nIn this example, a
          combination of latency, packet loss, bandwidth, and link flapping is applied.\n\n```\n\napiVersion:
          apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n  name: all-impairments\nspec:\n
          \ duration: 480 # seconds\n  startDelay: 5 # seconds\n  interfaces:\n  -
          \"ens2f0\"\n  egress:\n    latency: 50 # ms. Bidirectional, so total of
          100ms\n    loss: 0.02 # percent\n    bandwidth: 1000 # 1000 kbit/s, about
          1 mbit/s\n  ingress:\n    latency: 50 # ms. Bidirectional, so total of 100ms\n
          \   loss: 0.02 # percent\n    bandwidth: 1000 # 1000 kbit/s, about 1 mbit/s\n
          \ linkFlapping:\n    enable: true\n    downTime: 30 # Seconds\n    upTime:
          120 # Seconds\n\n```\n\n##### Example 5\n\n\nIn this example, packet duplicaiton
          is applied.\n\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n
          \ name: duplication\nspec:\n  duration: 60\n  startDelay: 5\n  interfaces:\n
          \ - \"ens3\"\n  ingress:\n    duplication: 25 # percent\n  egress:\n    duplication:
          25 # percent\n\n```\n\n## Setup\n\n### Requirements\n\n1. You need `make`
          installed.\n2. You need access to the kubernetes cluster with a kubeconfig.\n\n###
          Installation\n\nTo run using the current latest image:\n1. Clone the repository.\n2.
          Run `make deploy` with KUBECONFIG environment variable set to the location
          of the kubeconfig.\n\nTo run with your own code, there are more steps.\n1.
          Fork the repository.\n2. Clone to a machine that has access to the Kubernetes
          cluster and the kubeconfig.\n3. Modify the makefile to change the `IMG`
          variable to your image repository. If you do not have podman installed,
          also change podman to docker.\n4. Run `make docker-build` then `make docker-push`.\n5.
          Then run `make deploy`\n\n### Deploying from operator-hub\n\n*Important*
          You must manually apply the serviceaccount!\n\nFirst, Follow the instructions
          on this site to deploy from here.\n\nNext, apply the manifest \"serviceaccount.yaml\"
          from the manifests folder.\n\nThen you can apply your ClusterImpairment
          resources.\n\n## Running impairments\n\nFirst, configure a ClusterImpairment
          type's spec with the information for the impairment you want to run. Most
          importantly, set the interface(s). You can list the interfaces with `ifconfig`.
          If an invalid interface is set, the pod's logs will show `ifconfig` output.\n\n\nOnce
          the clusterimpairment type is set, apply it and it will work.\n\n\nTo re-run
          an impairment, delete the impairment with `oc delete -f path/to/manifest.yaml`,
          then apply it again.\n\n\n## Limitations\n\n\n### Multiple Impairments\n\n\nYou
          should avoid any configurations that apply impairments to the same interface
          on the same node. There are potential conflicts. The worker pod will attempt
          to remove all impairments on that interface before applying new ones.\n\n\nInstead,
          take advantage of the full control of both ingress and egress impairments
          from within the same ClusterImpairment resource.\n\n\nThere is currently
          a configured limit of 20 ClusterImpairment resources working at the same
          time. If that is a limitation for your use case, you can change it in `config/manager/manager.yaml`
          or you can try to find a way to make each ClusterImpairment do more.\n\n###
          Traffic Control (TC)\n\nTraffic control is how cluster-impairment-operator
          applies the latency, bandwidth, and packet loss impairments. The limitation
          is due to Linux not being realtime operating system, so the impairment will
          not be perfectly consistent.\n\n### Link Flapping\n\nWhen link flapping,
          if you flap the link that Kubernetes uses to communicate with the pods,
          you may be unable to remove the pod until the link is on long enough for
          Kubernetes to communicate with the impaired node.\n\n\nIn this case, it
          is helpful to set the duration properly instead of running for an indefinite
          or large amount of time, because the node will properly go back to the unimpaired
          state at that time.\n\n\nIf the cluster becomes offline due to the link
          flapping when you do not want it to be offline, soft restarting the nodes
          after removing the custom resource should remove all impairments.\n"
        displayName: cluster-impairment-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network
        - impairment
        - kubernetes
        - openshift
        - tc
        - trafficcontrol
        links:
        - name: Cluster Impairment Operator
          url: https://github.com/redhat-performance/cluster-impairment-operator
        maintainers:
        - email: joconnel@redhat.com
          name: Jared O'Connell
        - email: akrzos@redhat.com
          name: Alex Krzos
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat Performance
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/redhat-performance/cluster-impairment-operator:1.0.4
        version: 1.0.4
      entries:
      - name: cluster-impairment-operator.v1.0.4
        version: 1.0.4
      - name: cluster-impairment-operator.v1.0.3
        version: 1.0.3
      - name: cluster-impairment-operator.v1.0.2
        version: 1.0.2
      - name: cluster-impairment-operator.v1.0.1
        version: 1.0.1
      name: beta
    defaultChannel: beta
    packageName: cluster-impairment-operator
    provider:
      name: Red Hat Performance
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: mercury-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mercury-operator.1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:1.0.1",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CustomerCreditRating",
                  "expose": ["http"]
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:1.0.1
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:1.0.1
        version: 1.0.1
      entries:
      - name: mercury-operator.1.0.1
        version: 1.0.1
      name: 1.0.x
    - currentCSV: mercury-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_CREDIT_RATING",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-offer",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-offer-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_OFFER",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-party-routing-profile",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/party-routing-profile-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "PARTY_ROUTING_PROFILE",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-example-sdi": {
                "description": "Example Service Domain Infrastructure"
            },
            {
             "example-customer-credit-rating": {
               "description": "Example Customer Credit Rating Service Domain"
            },
            {
             "example-customer-offer": {
               "description": "Example Customer Offer Service Domain"
            },
            {
             "example-party-routing-profile": {
               "description": "Example Party Routing Profile Service Domain"
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:v1.0.2
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - description: Represents a Kafka cluster
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Integration is the Schema for the Camel K integrations API
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:v1.0.2
        version: 1.0.2
      entries:
      - name: mercury-operator.v1.0.2
        version: 1.0.2
      name: dev
    - currentCSV: mercury-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_CREDIT_RATING",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-offer",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-offer-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_OFFER",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-party-routing-profile",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/party-routing-profile-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "PARTY_ROUTING_PROFILE",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-example-sdi": {
                "description": "Example Service Domain Infrastructure"
            },
            {
             "example-customer-credit-rating": {
               "description": "Example Customer Credit Rating Service Domain"
            },
            {
             "example-customer-offer": {
               "description": "Example Customer Offer Service Domain"
            },
            {
             "example-party-routing-profile": {
               "description": "Example Party Routing Profile Service Domain"
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:v1.0.2
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - description: Represents a Kafka cluster
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Integration is the Schema for the Camel K integrations API
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:v1.0.2
        version: 1.0.2
      entries:
      - name: mercury-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: 1.0.x
    packageName: mercury-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-ati-updates
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-ati-updates.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ati.loadcore.keysight.com/v1alpha1",
                "kind": "Updates",
                "metadata": {
                  "name": "updates-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "installWait": {
                    "enabled": true,
                    "image": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:40:21Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Updates
            name: updates.ati.loadcore.keysight.com
            version: v1alpha1
        description: Keysight ATI Updates Operator
        displayName: Keysight ATI Updates Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Ati Updates
          url: https://keysight-ati-updates.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-ati-updates@sha256:abda76eb50d77e5a95191c9dbb21eefe64af5b794511ceafdfda007c15449c6e
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        version: 0.1.21
      entries:
      - name: keysight-ati-updates.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-ati-updates
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: CyberArk
      provider-url: https://www.cyberark.com/
    name: vcp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vcp-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "installer.venafi.com/v1alpha1",
                "kind": "VenafiInstall",
                "metadata": {
                  "name": "install-venafi-enhanced-issuer"
                },
                "spec": {
                  "certManager": {
                    "install": true
                  },
                  "globals": {
                    "imagePullSecretNames": [
                      "venafi"
                    ]
                  },
                  "venafiConnection": {
                    "install": true
                  },
                  "venafiEnhancedIssuer": {
                    "install": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:9da79fd0941c21e433f22aaaccc43ebbd4646e685aea0050194954d8b4019c05
          createdAt: "2024-05-30T11:35:41Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              VenafiInstall is an object for installing Venafi components onto a Kubernetes
              cluster
            displayName: VenafiInstall
            kind: VenafiInstall
            name: venafiinstalls.installer.venafi.com
            version: v1alpha1
        description: |
          Venafi Control Plane Operator for Red Hat OpenShift simplifies deployment of Venafi's enterprise components and dependencies.

          The enterprise components include:
          - Enterprise cert-manager: An enhanced version of cert-manager designed for application certificate management within your clusters.
          - Venafi Enhanced Issuer: Enables seamless certificate enrollment for your clusters directly from the Venafi Control Plane.
          - Venafi Firefly: A high-performance, lightweight issuer for certificates, centrally managed by the Venafi Control Plane.
          - Venafi Kubernetes Agent: Provides real-time visibility into certificates and Kubernetes resources directly within the Venafi Control Plane.
        displayName: Venafi Control Plane Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - venafi
        - vcp-operator
        links:
        - name: Documentation
          url: https://docs.venafi.cloud/
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Venafi
          url: https://venafi.com
        relatedImages:
        - registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:9da79fd0941c21e433f22aaaccc43ebbd4646e685aea0050194954d8b4019c05
        version: 1.0.0
      entries:
      - name: vcp-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: vcp-operator.v1.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "installer.venafi.com/v1alpha1",
                "kind": "VenafiInstall",
                "metadata": {
                  "name": "install-venafi-enhanced-issuer"
                },
                "spec": {
                  "certManager": {
                    "install": true
                  },
                  "globals": {
                    "imagePullSecretNames": [
                      "venafi"
                    ]
                  },
                  "venafiConnection": {
                    "install": true
                  },
                  "venafiEnhancedIssuer": {
                    "install": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:eb2b06d191dbfbb02332375a7bca6bc010d5410e53b3bc5fb8c320101292408c
          createdAt: "2025-07-22T06:44:30Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              VenafiInstall is an object for installing Venafi components onto a Kubernetes
              cluster
            displayName: VenafiInstall
            kind: VenafiInstall
            name: venafiinstalls.installer.venafi.com
            version: v1alpha1
        description: |
          The operator, formerly known as the Venafi Control Plane Operator, simplifies the deployment of enterprise components part of the CyberArk Certificate Manager for Kubernetes.

          The components include:
          - cert-manager: An enterprise distribution of cert-manager designed for application certificate management within your clusters.
          - Enterprise Issuer: Enables seamless certificate enrollment directly from the Certificate Manager.
          - Workload Identity Manager: A high-performance issuer for certificates, centrally managed by the Certificate Manager.
          - Agent: Provides real-time visibility into cluster resources directly within the Certificate Manager.
        displayName: CyberArk Certificate Manager Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cyberark
        - vcp-operator
        links:
        - name: Documentation
          url: https://docs.venafi.cloud/vaas/k8s-components/c-vcpo-overview/
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:eb2b06d191dbfbb02332375a7bca6bc010d5410e53b3bc5fb8c320101292408c
        version: 1.8.1
      entries:
      - name: vcp-operator.v1.8.1
        version: 1.8.1
      - name: vcp-operator.v1.8.0
        version: 1.8.0
      - name: vcp-operator.v1.7.0
        version: 1.7.0
      - name: vcp-operator.v1.6.0
        version: 1.6.0
      - name: vcp-operator.v1.5.0
        version: 1.5.0
      - name: vcp-operator.v1.4.0
        version: 1.4.0
      - name: vcp-operator.v1.3.0
        version: 1.3.0
      - name: vcp-operator.v1.2.1
        version: 1.2.1
      - name: vcp-operator.v1.2.0
        version: 1.2.0
      - name: vcp-operator.v1.1.1
        version: 1.1.1
      name: stable
    defaultChannel: stable
    packageName: vcp-operator
    provider:
      name: CyberArk
      url: https://www.cyberark.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE Ezmeral
      provider-url: ""
    name: hpe-ezmeral-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-ezmeral-csi-operator.v1.0.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ezmeral.hpe.com/v1",
                "kind": "HPEEzmeralCSIDriver",
                "metadata": {
                  "name": "hpeezmeralcsidriver"
                },
                "spec": {
                  "controllerImage": "registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03",
                  "nodeImage": "registry.connect.redhat.com/maprtech/csi-kdfplugin@sha256:ffd273ee2e608d2ef64fd3b2bcbb02d5f0a6c173ed96a15eaf5549ea45fea8d1",
                  "pullPolicy": "Always"
                }
              },
              {
                "apiVersion": "ezmeral.hpe.com/v1",
                "kind": "HPEEzmeralNFSCSIDriver",
                "metadata": {
                  "name": "hpeezmeralnfscsidriver"
                },
                "spec": {
                  "controllerImage": "registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03",
                  "nodeImage": "registry.connect.redhat.com/maprtech/csi-nfsplugin@sha256:a5c954f62f1395bafe2d68454ac604bca4be97573a3482fcbd8e76b6361cc08a",
                  "pullPolicy": "Always"
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator@sha256:c9c243c51cd760c2c2a66b489807a5e22053fcbeb94fa29afd10d9c91e2d3908
          createdAt: "2025-06-25T11:51:05Z"
          description: A Container Storage Interface (CSI) driver for HPE Ezmeral
            Data Fabric platform. The CSI driver allows you to use HPE Ezmeral Data
            Fabric with your preferred container orchestrator.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mapr/mapr-csi
          support: HPE Ezmeral Data Fabric
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE Ezmeral CSI Driver (FUSE)
            displayName: HPEEzmeralCSIDriver
            kind: HPEEzmeralCSIDriver
            name: hpeezmeralcsidrivers.ezmeral.hpe.com
            version: v1
          - description: HPE Ezmeral CSI Driver (NFS)
            displayName: HPEEzmeralNFSCSIDriver
            kind: HPEEzmeralNFSCSIDriver
            name: hpeezmeralnfscsidrivers.ezmeral.hpe.com
            version: v1
        description: |
          The HPE Ezmeral CSI Operator for Kubernetes packages, deploys, manages, upgrades the HPE Ezmeral CSI Driver on Kubernetes and OpenShift for dynamic provisioning of persistent volumes on HPE Ezmeral Data Fabric platform.
          The HPE Ezmeral CSI Driver for Kubernetes leverages HPE Ezmeral Data Fabric platform to provide scalable and persistent storage for stateful applications. Please refer to our CSI driver [documentation](https://docs.datafabric.hpe.com/home/CSIdriver/csi_overview.html) for full list of supported CSI features.
          ## Installation
          Refer to the HPE Ezmeral CSI Operator for Kubernetes [documentation](https://docs.datafabric.hpe.com/home/CSIdriver/csi_installing_operator.html)
        displayName: HPE Ezmeral Data Fabric CSI Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE Ezmeral
        - HPE Ezmeral Data Fabric
        - Storage
        - FileSystem
        - BlockVolume
        links:
        - name: Documentation
          url: https://docs.datafabric.hpe.com/home/CSIdriver/csi_overview.html
        maintainers:
        - email: ezmeral-csi-support@hpe.com
          name: HPE Ezmeral Data Fabric
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: HPE Ezmeral
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6ba961c2c2a29750c0132fe6dd6fa9f6001010afbc5f19b98add87b31b54bcf6
        - registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator@sha256:c9c243c51cd760c2c2a66b489807a5e22053fcbeb94fa29afd10d9c91e2d3908
        - registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03
        - registry.connect.redhat.com/maprtech/csi-kdfplugin@sha256:ffd273ee2e608d2ef64fd3b2bcbb02d5f0a6c173ed96a15eaf5549ea45fea8d1
        - registry.connect.redhat.com/maprtech/csi-nfsplugin@sha256:a5c954f62f1395bafe2d68454ac604bca4be97573a3482fcbd8e76b6361cc08a
        version: 1.0.18
      entries:
      - name: hpe-ezmeral-csi-operator.v1.0.18
        version: 1.0.18
      name: stable
    defaultChannel: stable
    packageName: hpe-ezmeral-csi-operator
    provider:
      name: HPE Ezmeral
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzo-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzo-operator.v3.1.200
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzo.cambridgesemantics.com/v2",
                "kind": "Anzo",
                "metadata": {
                  "name": "no-persistence"
                },
                "spec": {
                  "jsonActivation": true,
                  "jvmMemory": "4096m",
                  "nodeConfig": {
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "app": "anzo"
                        }
                      },
                      "serviceName": "anzo-no-persistence",
                      "template": {
                        "metadata": {
                          "annotations": {
                            "cluster-autoscaler.kubernetes.io/safe-to-evict": "false"
                          },
                          "labels": {
                            "app": "anzo"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "registry.connect.redhat.com/cambridgesemantics/anzo@sha256:5b192604338f602fdabe55fecefd9eb82400a77c5aa7f04bd8e54aead8644c6d",
                              "imagePullPolicy": "Always",
                              "name": "anzo",
                              "resources": {
                                "limits": {
                                  "cpu": "2.0",
                                  "memory": "8.0Gi"
                                },
                                "requests": {
                                  "cpu": "2.0",
                                  "memory": "8.0Gi"
                                }
                              }
                            }
                          ],
                          "initContainers": [
                            {
                              "name": "anzo-init",
                              "resources": {
                                "limits": {
                                  "cpu": "0.5",
                                  "memory": "2Gi"
                                },
                                "requests": {
                                  "cpu": "0.5",
                                  "memory": "2Gi"
                                }
                              }
                            }
                          ],
                          "securityContext": {
                            "fsGroup": 1001
                          },
                          "serviceAccountName": "anzo-operator"
                        }
                      }
                    }
                  },
                  "role": "AnzoServer"
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/anzo-operator@sha256:4c0c8bf7acbfbb96f15459a20364608e5d0a0933c6326a89d219666ee0d88ca2
          createdAt: "2025-07-09T06:52:12Z"
          description: kubernetes operator for Altair Graph Studio
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Altair Customer Support
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoBackup is the Schema for the anzobackups API
            displayName: Altair Graph Studio Backup
            kind: AnzoBackup
            name: anzobackups.anzobackup.cambridgesemantics.com
            version: v1beta1
          - description: Altair Graph Studio is the Schema for the anzos API
            displayName: Altair Graph Studio Operator
            kind: Anzo
            name: anzos.anzo.cambridgesemantics.com
            version: v2
        description: |-
          The Altair Graph Studio Operator provides the way to install and configure an altair graph studio server and agent setup on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v3.1.2/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/deploy-k8s.htm )

          ### Support

          We offer Support to our customers [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Studio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        links:
        - name: Altair Graph Studio Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v3.1.2/README_openshift.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzo-operator@sha256:4c0c8bf7acbfbb96f15459a20364608e5d0a0933c6326a89d219666ee0d88ca2
        - registry.connect.redhat.com/cambridgesemantics/anzo@sha256:5b192604338f602fdabe55fecefd9eb82400a77c5aa7f04bd8e54aead8644c6d
        version: 3.1.200
      entries:
      - name: anzo-operator.v3.1.200
        version: 3.1.200
      - name: anzo-operator.v3.1.106
        version: 3.1.106
      - name: anzo-operator.v3.1.105
        version: 3.1.105
      name: stable
    defaultChannel: stable
    packageName: anzo-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: ""
    name: power-monitoring-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: power-monitoring-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "Kepler",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "kepler",
                    "app.kubernetes.io/name": "kepler",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "kepler"
                },
                "spec": {
                  "exporter": {
                    "deployment": {
                      "port": 9103
                    },
                    "redfish": null
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: registry.redhat.io/openshift-power-monitoring/power-monitoring-rhel9-operator@sha256:5eec15f6fe9bd4bdb1f5fb8057dc67c448c3711bcc7a0a4760f0d90c07688d1e
          createdAt: "2025-02-14T06:23:56Z"
          description: Deploys and Manages Kepler on OpenShift
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/internal-objects: |-
            [
              "keplerinternals.kepler.system.sustainable.computing.io"
            ]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift-power-monitoring/power-monitoring-operator
          support: Team Power Monitoring (team-powermon@redhat.com)
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KeplerInternal is the Schema for internal/unsupported API
            displayName: KeplerInternal
            kind: KeplerInternal
            name: keplerinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: Kepler is the Schema for the keplers API
            displayName: Kepler
            kind: Kepler
            name: keplers.kepler.system.sustainable.computing.io
            version: v1alpha1
        description: |
          ### Introduction
          Kepler (Kubernetes Efficient Power Level Exporter) uses eBPF to probe
          energy-related system stats and exports them as Prometheus metrics.

          ### NOTE
          To view the metrics exposed by Kepler, ensure that
          [User Project Monitoring](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/monitoring/configuring-user-workload-monitoring)
          is enabled in your cluster.

          ### Documentation
          Documentation and installation guide can be found below:
            * [Installation Guide](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/power_monitoring/installing-power-monitoring)
        displayName: Power monitoring for Red Hat OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kepler
        - power
        - energy
        - monitoring
        links:
        - name: Installation Guide
          url: https://docs.redhat.com/documentation/openshift_container_platform/latest/html/power_monitoring/installing-power-monitoring
        maintainers:
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: vimalkum@redhat.com
          name: Vimal Kumar
        - email: vprashar@redhat.com
          name: Vibhu Prashar
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift-power-monitoring/power-monitoring-rhel9-operator@sha256:5eec15f6fe9bd4bdb1f5fb8057dc67c448c3711bcc7a0a4760f0d90c07688d1e
        - registry.redhat.io/openshift-power-monitoring/kepler-rhel9@sha256:27d3cb0a98f1cffd649745e3886ad92995d83a0de6823ddc2d770bf036e2eeaf
        version: 0.4.0
      entries:
      - name: power-monitoring-operator.v0.4.0
        version: 0.4.0
      - name: power-monitoring-operator.v0.3.0
        version: 0.3.0
      - name: power-monitoring-operator.v0.2.0
        version: 0.2.0
      - name: power-monitoring-operator.v0.1.0
        version: 0.1.0
      name: tech-preview
    defaultChannel: tech-preview
    packageName: power-monitoring-operator
    provider:
      name: Red Hat, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-registry
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-operator.v1.1.3-v2.6.4.final
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-mem"
                },
                "spec": {
                  "configuration": {
                    "persistence": "mem"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-operator:1.1.3
          createdAt: "2024-09-17"
          description: Deploy and manage Apicurio Registry on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents an Apicurio Registry instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Apicurio Registry

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Provides a REST API and web UI to manage the artifacts and artifact meta-data
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, in-memory
          - Compatible with Confluent and IBM APIs
          - Runs on a lightweight Quarkus platform
          - Includes Maven plugin to integrate with Maven based builds

          ## Apicurio Registry Operator

          Provides a quick and easy way to deploy and manage Apicurio Registry on Kubernetes.

          **Features**
          - Supports installation and configuration of the Registry for all storage options
          - Easily perform a rolling upgrade of Apicurio Registry

          ## Prerequisites

          This operator does not deploy storage for Apicurio Registry. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/main/LICENSE)
        displayName: Apicurio Registry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-operator:1.1.3
        version: 1.1.3-v2.6.4.final
      entries:
      - name: apicurio-registry-operator.v1.1.3-v2.6.4.final
        version: 1.1.3-v2.6.4.final
      - name: apicurio-registry-operator.v1.1.2-v2.5.11.final
        version: 1.1.2-v2.5.11.final
      - name: apicurio-registry-operator.v1.1.1-v2.5.9.final
        version: 1.1.1-v2.5.9.final
      - name: apicurio-registry-operator.v1.1.0-v2.4.12.final
        version: 1.1.0-v2.4.12.final
      - name: apicurio-registry-operator.v1.0.0-v2.0.0.final
        version: 1.0.0-v2.0.0.final
      name: 2.x
    defaultChannel: 2.x
    packageName: apicurio-registry
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zettaset
      provider-url: ""
    name: xcrypt-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zts-xcrypt-operator.v0.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xcrypt.zettaset.com/v1alpha1",
                "kind": "XCrypt",
                "metadata": {
                  "name": "zts-masterset"
                },
                "spec": {
                  "status": {
                    "podNames": "zts-ca",
                    "replicas": "1,"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.connect.redhat.com/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
          createdAt: "2021-06-16T12:29:13.97Z"
          description: XCrypt Operator deploys Zettaset XCrypt Container Encryption  -
            the only software-defined solution that offers transparent high performance
            data-at-rest encryption of critical data in container environments.
          support: Zettaset
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XCrypt is the Schema for the xcrypts API
            kind: XCrypt
            name: xcrypts.xcrypt.zettaset.com
            version: v1alpha1
        description: "Zettaset’s RedHat Certified XCrypt Operator allows users to
          deploy multiple XCrypt components on OpenShift with just a few simple commands.
          The XCrypt Operator deployment provides software-defined encryption that
          transparently protects container data throughout your OpenShift environment.\n\n##
          Core Features\n* **Automated XCrypt provisioning** - Deploying a Zettaset
          XCrypt has never been easier. Adjust cluster specific configuration and
          let the Zettaset Operator take care of provisioning XCrypt services and
          setting up the XCrypt encryption to your exact specification. \n* **CSI
          Driver deployment**: Zettaset XCrypt Operator automatically deploys Zettaset
          XCrypt CSI Driver in order to provide automated and transparent data-at-rest
          \ encryption functionality compatible with Openshift and any other container
          orchestrator. Once the Operator is provisioned, the driver runs fine tuned
          and ready to work with other XCrypt components.\n* **Multi-node topology**:
          Encrypt as many nodes and volumes as you need to. XCrypt components can
          be installed on multiple nodes across the cluster. This is defined by two
          labels: zts-master and zts-worker. Normally zts-master node is a single
          node that runs Zettaset XCrypt major services required for the product operation.
          Multiple nodes can be marked as zts-worker and those are the nodes that
          can have encrypted volumes.\n* **Seamless Encrypted Volume mount**: Volumes
          are not only being encrypted, but also are mounted/unmounted to containers
          with Persistent Volume Claims. This happens as seamlessly as regular volumes
          provision and mount/unmount.\n* **Robust failover**: A node failure doesn’t
          affect XCrypt operations on other nodes of the cluster. Encryption on a
          restarted/recovered node gets back to normal functioning if the node stays
          healthy and data hasn’t been corrupted during the restart\n\n## About Zettaset
          XCrypt Kubernetes Encryption for OpenShift\nXCrypt Kubernetes Encryption
          for OpenShift is a software-only solution that enables encryption of data
          at rest stored in Red Hat OpenShift container environments without slowing
          down the DevSecOps process. In addition to providing a transparent and high-performance
          layer of security that flexibly protects container data across on-premise,
          in hybrid, and multi-cloud environments, it also helps your organization
          efficiently transition from DevOps to DevSecOps.  \n\nMore information about
          Zettaset XCrypt Kubernetes Encryption for OpenShift can be found in the
          [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)\n\n##
          RedHat OpenShift + Zettaset: The next generation of secure enterprise infrastructure
          for DevSecOps  \nBy combining RedHat’s Container Platform with Zettaset’s
          certified software-defined encryption, you can create the next generation
          of secure enterprise infrastructure that DevSecOps requires. Together they
          create the stable, consistent, and supported base that your organization
          needs to develop applications faster, simplify updates, drive innovation,
          and reduce the risk of potential security breaches and data theft.\n###
          Core Features\n* Software-only for simple and scalable deployments \n* Negligible
          performance impact\n* Transparent to developers and administrators \n* Works
          with AWS, Google, and Azure \n* Unique encryption key per container volume
          \n* Encrypted volumes are only available when in use\n* Direct integration
          with OpenShift \n* Automated encryption policy management Secure erase of
          volumes\n* Ability to securely decommission a node\n\n## How does XCrypt
          Kubernetes Encryption fit into OpenShift? \nXCrypt XCrypt Kubernetes Encryption
          for OpenShift Encryption makes it incredibly easy to protect container data
          in your OpenShift environment. \n### Use cases include:\n* Transitioning
          from DevOps to DevSecOps \n* Ensuring data protection in OpenShift container
          environments with negligible performance impact \n* Achieving compliance
          in regulated industries like healthcare and finance\n\n## Other Information\nRole-based
          access control needs to be configured prior to the Operator installation.
          Please find details in the [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)
          \   "
        displayName: Zettaset XCrypt Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - encryption
        - storage
        links:
        - name: Zettaset
          url: https://www.zettaset.com/
        maintainers:
        - email: info@zettaset.com
          name: Zettaset
        maturity: alpha
        provider:
          name: Zettaset
        relatedImages:
        - registry.connect.redhat.com/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
        - registry.connect.redhat.com/zts/xcrypt-license-server-1@sha256:bc5486892a41e70f5dc4c4e1d82fca8db2149598fb8b5950ba5da1ee2d9d245d
        - registry.connect.redhat.com/zts/xcrypt-host-manager-1@sha256:9faa4b597eb74df412761dffa057b64fb69dbf9fa016414962eb2099a343e734
        - registry.connect.redhat.com/zts/zts-csi-driver-1@sha256:c15fa9a7cc40891103f426206663dcebe1210c8ee3ab6037cbf64e1f86b9c7e5
        - registry.connect.redhat.com/zts/xcrypt-ca-1@sha256:9f1015dfb459b69f5dbf449efa0517452095a6cf4f20826ebae278f9f800ffe7
        - registry.connect.redhat.com/zts/xcrypt-kmip-server-1@sha256:9af5d21dd748dfbf4e107f27a5f10fa3c0ba928ade9b989a744b620ce34ec757
        version: 0.0.17
      entries:
      - name: zts-xcrypt-operator.v0.0.17
        version: 0.0.17
      name: alpha
    defaultChannel: alpha
    packageName: xcrypt-operator
    provider:
      name: Zettaset
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-spark-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spark-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:a6c82a6d6ebaa0d2572d644ac9013451e020f134c6fe46222fef37b1e9aa3cbb
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:a6c82a6d6ebaa0d2572d644ac9013451e020f134c6fe46222fef37b1e9aa3cbb
        version: 24.11.1
      entries:
      - name: spark-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: spark-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:8ad9948bf8264924a0eedd92c2e58269117487b9d57816384447bb6c999be681
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:8ad9948bf8264924a0eedd92c2e58269117487b9d57816384447bb6c999be681
        version: 25.3.0
      entries:
      - name: spark-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: spark-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkConnectServerSpec via
              `CustomResource`
            displayName: sparkconnectservers.spark.stackable.tech
            kind: SparkConnectServer
            name: sparkconnectservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
        version: 25.7.0
      entries:
      - name: spark-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: spark-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkConnectServerSpec via
              `CustomResource`
            displayName: sparkconnectservers.spark.stackable.tech
            kind: SparkConnectServer
            name: sparkconnectservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
        version: 25.7.0
      entries:
      - name: spark-operator.v25.7.0
        version: 25.7.0
      - name: spark-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-spark-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://skupper.io
    name: skupper-netobs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-netobs-operator.v2.1.1-rh-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability.skupper.io/v2alpha1",
                "kind": "NetworkObserver",
                "metadata": {
                  "name": "networkobserver-sample"
                },
                "spec": {
                  "auth": {
                    "basic": {
                      "create": true,
                      "secretName": ""
                    },
                    "openshift": {
                      "cookieSecretName": "",
                      "createCookieSecret": true,
                      "serviceAccount": {
                        "create": true,
                        "nameOverride": ""
                      }
                    },
                    "strategy": "openshift"
                  },
                  "extraArgs": null,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "registry.redhat.io/service-interconnect/skupper-network-observer-rhel9",
                    "tag": "2.1.1"
                  },
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nginx": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/ubi9/nginx-124",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": 9.6
                  },
                  "openshiftOauthProxy": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-oauth-proxy-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.19.0"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "prometheus": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-prometheus-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.16.0"
                  },
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "router": {
                    "certificate": {
                      "create": true,
                      "nameOverride": ""
                    },
                    "endpoint": "amqps://skupper-router-local"
                  },
                  "securityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    }
                  },
                  "service": {
                    "port": 443,
                    "type": "ClusterIP"
                  },
                  "skipManagementLabels": false,
                  "tls": {
                    "openshiftIssued": true,
                    "secretName": "",
                    "skupperIssued": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
          createdAt: "2025-08-27T20:21:18Z"
          description: Red Hat Service Interconnect Network Observer Operator provides
            the ability to monitor and observe a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkObserver
            name: networkobservers.observability.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect Network Observer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/ubi9/nginx-124@sha256:cbcca4743015b4ef167383a080f4711a6049e275e0b66f80a74f8b087b0dfc7f
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:d3056b35d9a205b9f2c48d924f199c5ac23904eb18d526e4bff229e7c7181415
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
        version: 2.1.1-rh-8
      entries:
      - name: skupper-netobs-operator.v2.1.1-rh-8
        version: 2.1.1-rh-8
      name: stable-2
    - currentCSV: skupper-netobs-operator.v2.1.1-rh-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability.skupper.io/v2alpha1",
                "kind": "NetworkObserver",
                "metadata": {
                  "name": "networkobserver-sample"
                },
                "spec": {
                  "auth": {
                    "basic": {
                      "create": true,
                      "secretName": ""
                    },
                    "openshift": {
                      "cookieSecretName": "",
                      "createCookieSecret": true,
                      "serviceAccount": {
                        "create": true,
                        "nameOverride": ""
                      }
                    },
                    "strategy": "openshift"
                  },
                  "extraArgs": null,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "registry.redhat.io/service-interconnect/skupper-network-observer-rhel9",
                    "tag": "2.1.1"
                  },
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nginx": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/ubi9/nginx-124",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": 9.6
                  },
                  "openshiftOauthProxy": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-oauth-proxy-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.19.0"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "prometheus": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-prometheus-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.16.0"
                  },
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "router": {
                    "certificate": {
                      "create": true,
                      "nameOverride": ""
                    },
                    "endpoint": "amqps://skupper-router-local"
                  },
                  "securityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    }
                  },
                  "service": {
                    "port": 443,
                    "type": "ClusterIP"
                  },
                  "skipManagementLabels": false,
                  "tls": {
                    "openshiftIssued": true,
                    "secretName": "",
                    "skupperIssued": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
          createdAt: "2025-08-27T20:21:18Z"
          description: Red Hat Service Interconnect Network Observer Operator provides
            the ability to monitor and observe a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkObserver
            name: networkobservers.observability.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect Network Observer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:d3056b35d9a205b9f2c48d924f199c5ac23904eb18d526e4bff229e7c7181415
        - registry.redhat.io/ubi9/nginx-124@sha256:cbcca4743015b4ef167383a080f4711a6049e275e0b66f80a74f8b087b0dfc7f
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        version: 2.1.1-rh-8
      entries:
      - name: skupper-netobs-operator.v2.1.1-rh-8
        version: 2.1.1-rh-8
      name: stable-2.1
    defaultChannel: stable-2
    packageName: skupper-netobs-operator
    provider:
      name: Red Hat
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/trustification/trusted-profile-analyzer-operator
    name: rhtpa-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhtpa-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtpa.io/v1",
                "kind": "TrustedProfileAnalyzer",
                "metadata": {
                  "name": "trustedprofileanalyzer-sample"
                },
                "spec": {
                  "appDomain": "change-me",
                  "collector": {},
                  "database": {},
                  "image": {
                    "fullName": "registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e",
                    "pullPolicy": "IfNotPresent"
                  },
                  "infrastructure": {
                    "port": 9010
                  },
                  "ingress": {},
                  "metrics": {
                    "enabled": false
                  },
                  "modules": {
                    "createDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "createImporters": {
                      "enabled": false,
                      "image": {},
                      "importers": {}
                    },
                    "importer": {
                      "concurrency": 1,
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {},
                      "workingDirectory": {
                        "size": "32Gi"
                      }
                    },
                    "migrateDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "server": {
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "ingress": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {}
                    }
                  },
                  "oidc": {
                    "clients": {
                      "cli": {
                        "clientSecret": null
                      },
                      "frontend": {}
                    }
                  },
                  "openshift": {
                    "useServiceCa": true
                  },
                  "partOf": "trustify",
                  "replicas": 1,
                  "rust": {},
                  "storage": {},
                  "tls": {},
                  "tracing": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-09-19T15:04:16Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Profile Analyzer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trustification/trusted-profile-analyzer-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: TrustedProfileAnalyzer
            name: trustedprofileanalyzers.rhtpa.io
            version: v1
        description: |
          [Tech Preview]
          Red Hat Trusted Profile Analyzer (RHTPA) provides a single source of truth for your software inventory across all your environments. You can verify the integrity of your in-house and third-party applications against their respective Software Bill of Materials (SBOM) document.  Trusted Profile Analyzer provides vulnerability, component and license transparency, strengthens your software supply chain security, and enables an accurate, up-to-date understanding of your software inventory.

          Features and Benefits:

          Enhanced Software Supply Chain Security: Offers critical transparency and integrity verification across the entire software lifecycle.

          Reduced Operational Risk: Proactively identifies and helps remediate vulnerabilities minimizing risk and compliance violations.

          Improved Compliance and Auditability: Simplifies the process of meeting regulatory requirements for software transparency and integrity by providing verifiable documentation and continuous monitoring.

          Faster Remediation and Incident Response: Provides targeted, actionable insights to prioritize and accelerate the patching and remediation of vulnerabilities.

          Increased Developer Productivity & Trust: Integrates security into existing workflows without significant overhead, allowing developers to build and deploy with greater confidence in the integrity of their code and components.

          Centralized Visibility and Control: Consolidates disparate security information into a single pane of glass, offering a holistic view of your application security posture.

          Red Hat Trusted Profile Analyzer is a crucial tool for any organization running applications on OpenShift that needs to confidently understand, verify, and secure the software they deploy, especially in an era of complex software supply chains and increasing regulatory scrutiny.

          [Documentation](https://docs.redhat.com/en/documentation/red_hat_trusted_profile_analyzer/2.1).
        displayName: Red Hat Trusted Profile Analyzer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rhtpa
        - trusted-profile-analyzer
        links:
        - name: Trusted Profile Analyzer Operator
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        maintainers:
        - email: rhtpa-support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        relatedImages:
        - registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
        - registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e
        version: 1.0.2
      entries:
      - name: rhtpa-operator.v1.0.2
        version: 1.0.2
      - name: rhtpa-operator.v1.0.1
        version: 1.0.1
      - name: rhtpa-operator.v1.0.0
        version: 1.0.0
      name: stable
    - currentCSV: rhtpa-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtpa.io/v1",
                "kind": "TrustedProfileAnalyzer",
                "metadata": {
                  "name": "trustedprofileanalyzer-sample"
                },
                "spec": {
                  "appDomain": "change-me",
                  "collector": {},
                  "database": {},
                  "image": {
                    "fullName": "registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e",
                    "pullPolicy": "IfNotPresent"
                  },
                  "infrastructure": {
                    "port": 9010
                  },
                  "ingress": {},
                  "metrics": {
                    "enabled": false
                  },
                  "modules": {
                    "createDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "createImporters": {
                      "enabled": false,
                      "image": {},
                      "importers": {}
                    },
                    "importer": {
                      "concurrency": 1,
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {},
                      "workingDirectory": {
                        "size": "32Gi"
                      }
                    },
                    "migrateDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "server": {
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "ingress": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {}
                    }
                  },
                  "oidc": {
                    "clients": {
                      "cli": {
                        "clientSecret": null
                      },
                      "frontend": {}
                    }
                  },
                  "openshift": {
                    "useServiceCa": true
                  },
                  "partOf": "trustify",
                  "replicas": 1,
                  "rust": {},
                  "storage": {},
                  "tls": {},
                  "tracing": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-09-19T15:04:16Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Profile Analyzer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trustification/trusted-profile-analyzer-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: TrustedProfileAnalyzer
            name: trustedprofileanalyzers.rhtpa.io
            version: v1
        description: |
          [Tech Preview]
          Red Hat Trusted Profile Analyzer (RHTPA) provides a single source of truth for your software inventory across all your environments. You can verify the integrity of your in-house and third-party applications against their respective Software Bill of Materials (SBOM) document.  Trusted Profile Analyzer provides vulnerability, component and license transparency, strengthens your software supply chain security, and enables an accurate, up-to-date understanding of your software inventory.

          Features and Benefits:

          Enhanced Software Supply Chain Security: Offers critical transparency and integrity verification across the entire software lifecycle.

          Reduced Operational Risk: Proactively identifies and helps remediate vulnerabilities minimizing risk and compliance violations.

          Improved Compliance and Auditability: Simplifies the process of meeting regulatory requirements for software transparency and integrity by providing verifiable documentation and continuous monitoring.

          Faster Remediation and Incident Response: Provides targeted, actionable insights to prioritize and accelerate the patching and remediation of vulnerabilities.

          Increased Developer Productivity & Trust: Integrates security into existing workflows without significant overhead, allowing developers to build and deploy with greater confidence in the integrity of their code and components.

          Centralized Visibility and Control: Consolidates disparate security information into a single pane of glass, offering a holistic view of your application security posture.

          Red Hat Trusted Profile Analyzer is a crucial tool for any organization running applications on OpenShift that needs to confidently understand, verify, and secure the software they deploy, especially in an era of complex software supply chains and increasing regulatory scrutiny.

          [Documentation](https://docs.redhat.com/en/documentation/red_hat_trusted_profile_analyzer/2.1).
        displayName: Red Hat Trusted Profile Analyzer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rhtpa
        - trusted-profile-analyzer
        links:
        - name: Trusted Profile Analyzer Operator
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        maintainers:
        - email: rhtpa-support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        relatedImages:
        - registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
        - registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e
        version: 1.0.2
      entries:
      - name: rhtpa-operator.v1.0.2
        version: 1.0.2
      - name: rhtpa-operator.v1.0.1
        version: 1.0.1
      - name: rhtpa-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable
    packageName: rhtpa-operator
    provider:
      name: Red Hat
      url: https://github.com/trustification/trusted-profile-analyzer-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: nova-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nova-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "Nova",
                "metadata": {
                  "name": "nova"
                },
                "spec": {
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaAPI",
                "metadata": {
                  "name": "novaapi-sample"
                },
                "spec": {
                  "apiDatabaseHostname": "openstack",
                  "cell0DatabaseHostname": "openstack",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "registeredCells": {},
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaCell",
                "metadata": {
                  "name": "novacell1-with-upcall-sample"
                },
                "spec": {
                  "apiDatabaseHostname": "openstack",
                  "cellDatabaseHostname": "openstack",
                  "cellName": "cell1",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaCompute",
                "metadata": {
                  "name": "nova-compute-ironic"
                },
                "spec": {
                  "novaComputeTemplates": {
                    "compute-ironic": {
                      "computeDriver": "ironic.IronicDriver",
                      "customServiceConfig": "",
                      "networkAttachments": [
                        "internalapi"
                      ],
                      "replicas": 1,
                      "resources": {}
                    }
                  }
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaConductor",
                "metadata": {
                  "name": "novaconductor-sample"
                },
                "spec": {
                  "cellName": "cell1",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaMetadata",
                "metadata": {
                  "name": "novametadata-sample"
                },
                "spec": {
                  "apiDatabaseAccount": "nova-api",
                  "apiDatabaseHostname": "openstack",
                  "cellDatabaseAccount": "nova-cell0",
                  "cellDatabaseHostname": "openstack",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "registeredCells": {},
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaNoVNCProxy",
                "metadata": {
                  "name": "novanovncproxy-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaScheduler",
                "metadata": {
                  "name": "novascheduler-sample"
                },
                "spec": {
                  "apiDatabaseHostname": "openstack",
                  "cell0DatabaseHostname": "openstack",
                  "customServiceConfig": "[DEFAULT]\ndebug=true\n",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "registeredCells": {},
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:21Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nova
            name: nova.nova.openstack.org
            version: v1beta1
          - description: NovaAPI is the Schema for the novaapis API
            displayName: Nova API
            kind: NovaAPI
            name: novaapis.nova.openstack.org
            version: v1beta1
          - description: NovaCell is the Schema for the novacells API
            displayName: Nova Cell
            kind: NovaCell
            name: novacells.nova.openstack.org
            version: v1beta1
          - description: NovaCompute is the Schema for the NovaCompute
            displayName: Nova Compute
            kind: NovaCompute
            name: novacomputes.nova.openstack.org
            version: v1beta1
          - description: NovaConductor is the Schema for the novaconductors API
            displayName: Nova Conductor
            kind: NovaConductor
            name: novaconductors.nova.openstack.org
            version: v1beta1
          - description: NovaMetadata is the Schema for the novametadata API
            displayName: Nova Metadata
            kind: NovaMetadata
            name: novametadata.nova.openstack.org
            version: v1beta1
          - description: NovaNoVNCProxy is the Schema for the novanovncproxies API
            displayName: Nova No VNCProxy
            kind: NovaNoVNCProxy
            name: novanovncproxies.nova.openstack.org
            version: v1beta1
          - description: NovaScheduler is the Schema for the novaschedulers API
            displayName: Nova Scheduler
            kind: NovaScheduler
            name: novaschedulers.nova.openstack.org
            version: v1beta1
        description: Nova Operator
        displayName: Nova Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Nova
        links:
        - name: Nova Operator
          url: https://github.com/openstack-k8s-operators/nova-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-nova-scheduler-rhel9@sha256:ba91196cd4f9b6e6f49868c91343b8eca7cc0d5016c36e83e2cc76f45cbe777b
        - registry.redhat.io/rhoso/openstack-nova-compute-rhel9@sha256:43cbe46ab263403fc303b6bc2388863029aad33465b237db18da7cd6d42078e1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/nova-rhel9-operator@sha256:f058c56df02a108505d51b9d19973894c2ee59055776c9f3fd7f7afe23a5106a
        - registry.redhat.io/rhoso/openstack-nova-api-rhel9@sha256:bbb8ebceb6f30b6c8beb120429e69f323bd231c09aaaddc5be9f15a527e928ed
        - registry.redhat.io/rhoso/openstack-nova-conductor-rhel9@sha256:500c36fd519611be8662e97ca31b1e59269c14d8aa94286dad9925e957462155
        - registry.redhat.io/rhoso/openstack-nova-novncproxy-rhel9@sha256:bf534f0c72f188d5f51e40b0392ace182db2055dff9ad14b01eeef13056febbf
        version: 1.0.6
      entries:
      - name: nova-operator.v1.0.6
        version: 1.0.6
      - name: nova-operator.v1.0.5
        version: 1.0.5
      - name: nova-operator.v1.0.4
        version: 1.0.4
      - name: nova-operator.v1.0.3
        version: 1.0.3
      - name: nova-operator.v1.0.2
        version: 1.0.2
      - name: nova-operator.v1.0.1
        version: 1.0.1
      - name: nova-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: nova-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq7-interconnect-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: 1.10.x
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: 1.2.0
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: latest
    defaultChannel: 1.10.x
    packageName: amq7-interconnect-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Splunk Inc.
      provider-url: www.splunk.com
    name: splunk-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: splunk-operator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "IndexerCluster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "indexercluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "LicenseMaster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "licensemaster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "MonitoringConsole",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "monitoringconsole-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "SearchHeadCluster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "searchheadcluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "Standalone",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "standalone-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "clustermanager-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "IndexerCluster",
                "metadata": {
                  "name": "indexercluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "LicenseManager",
                "metadata": {
                  "name": "licensemanager-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "MonitoringConsole",
                "metadata": {
                  "name": "monitoringconsole-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "SearchHeadCluster",
                "metadata": {
                  "name": "searchheadcluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "Standalone",
                "metadata": {
                  "name": "standalone-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Big Data, Logging & Tracing, Monitoring, Security, AI/Machine
            Learning
          containerImage: splunk/splunk-operator@sha256:ed02561995d46ff8eea143fa78f5da118ee6fdb500832e9a33b42d059a67e229
          createdAt: "2025-09-08T10:17:59Z"
          description: The Splunk Operator for Kubernetes enables you to quickly and
            easily deploy Splunk Enterprise on your choice of private or public cloud
            provider. The Operator simplifies scaling and management of Splunk Enterprise
            by automating administrative workflows using Kubernetes best practices.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=2.4.0 <3.0.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/splunk/splunk-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterManager is the Schema for the cluster manager API
            displayName: Cluster Manager
            kind: ClusterManager
            name: clustermanagers.enterprise.splunk.com
            version: v4
          - description: ClusterMaster is the Schema for the cluster manager API
            displayName: Cluster Master
            kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v3
          - kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v1
          - kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v2
          - description: IndexerCluster is the Schema for a Splunk Enterprise indexer
              cluster
            displayName: Indexer Cluster
            kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v3
          - description: IndexerCluster is the Schema for a Splunk Enterprise indexer
              cluster
            displayName: Indexer Cluster
            kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v4
          - kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v1
          - kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v2
          - description: LicenseManager is the Schema for a Splunk Enterprise license
              manager.
            displayName: License Manager
            kind: LicenseManager
            name: licensemanagers.enterprise.splunk.com
            version: v4
          - description: LicenseMaster is the Schema for a Splunk Enterprise license
              master.
            displayName: License Master
            kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v3
          - kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v1
          - kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v2
          - description: MonitoringConsole is the Schema for the monitoringconsole
              API
            displayName: Monitoring Console
            kind: MonitoringConsole
            name: monitoringconsoles.enterprise.splunk.com
            version: v3
          - description: MonitoringConsole is the Schema for the monitoringconsole
              API
            displayName: Monitoring Console
            kind: MonitoringConsole
            name: monitoringconsoles.enterprise.splunk.com
            version: v4
          - description: SearchHeadCluster is the Schema for a Splunk Enterprise search
              head cluster
            displayName: Search Head Cluster
            kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v3
          - description: SearchHeadCluster is the Schema for a Splunk Enterprise search
              head cluster
            displayName: Search Head Cluster
            kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v4
          - kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v1
          - kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v2
          - description: Standalone is the Schema for a Splunk Enterprise standalone
              instances.
            displayName: Standalone
            kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v3
          - description: Standalone is the Schema for a Splunk Enterprise standalone
              instances.
            displayName: Standalone
            kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v4
          - kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v1
          - kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v2
        description: The Splunk Operator for Kubernetes enables you to quickly and
          easily deploy Splunk Enterprise on your choice of private or public cloud
          provider. The Operator simplifies scaling and management of Splunk Enterprise
          by automating administrative workflows using Kubernetes best practices.
        displayName: Splunk Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Splunk
        - Enterprise
        - Operator
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://github.com/splunk/splunk-operator/tree/main/docs
        maintainers:
        - email: support@splunk.com
          name: Splunk Operator Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Splunk Inc.
          url: www.splunk.com
        relatedImages:
        - splunk/splunk-operator@sha256:ed02561995d46ff8eea143fa78f5da118ee6fdb500832e9a33b42d059a67e229
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - splunk/splunk@sha256:4a5ce5a7b4e84a87396f0664895095c6510a0f2bf52e8747a373c7357f3ee313
        version: 3.0.0
      entries:
      - name: splunk-operator.v3.0.0
        version: 3.0.0
      - name: splunk-operator.v2.8.1
        version: 2.8.1
      - name: splunk-operator.v2.8.0
        version: 2.8.0
      - name: splunk-operator.v2.7.1
        version: 2.7.1
      - name: splunk-operator.v2.7.0
        version: 2.7.0
      name: stable
    defaultChannel: stable
    packageName: splunk-operator
    provider:
      name: Splunk Inc.
      url: www.splunk.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-emrcontainers-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-emrcontainers-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "emrcontainers.services.k8s.aws/v1alpha1",
                "kind": "VirtualCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/emrcontainers-controller:1.1.1
          createdAt: "2025-09-19T19:36:26Z"
          description: AWS EMR on EKS controller is a service controller for managing
            EMR on EKS resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: JobRun represents the state of an AWS emrcontainers JobRun
              resource.
            displayName: JobRun
            kind: JobRun
            name: jobruns.emrcontainers.services.k8s.aws
            version: v1alpha1
          - description: VirtualCluster represents the state of an AWS emrcontainers
              VirtualCluster resource.
            displayName: VirtualCluster
            kind: VirtualCluster
            name: virtualclusters.emrcontainers.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EMR on EKS resources in AWS from within your Kubernetes cluster.

          **About Amazon EMR on EKS**

          Amazon EMR on EKS provides a deployment option for Amazon EMR that allows you to run open-source big data frameworks on Amazon Elastic Kubernetes Service (Amazon EKS). With this deployment option, you can focus on running analytics workloads while Amazon EMR on EKS builds, configures, and manages containers for open-source applications.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EMR on EKS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - emrcontainers
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EMR on EKS Developer Resources
          url: https://docs.aws.amazon.com/emr/latest/EMR-on-EKS-DevelopmentGuide/emr-eks.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: emr on eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/emrcontainers-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-emrcontainers-controller.v1.1.1
        version: 1.1.1
      - name: ack-emrcontainers-controller.v1.1.0
        version: 1.1.0
      - name: ack-emrcontainers-controller.v1.0.27
        version: 1.0.27
      - name: ack-emrcontainers-controller.v1.0.25
        version: 1.0.25
      - name: ack-emrcontainers-controller.v1.0.24
        version: 1.0.24
      - name: ack-emrcontainers-controller.v1.0.23
        version: 1.0.23
      - name: ack-emrcontainers-controller.v1.0.22
        version: 1.0.22
      - name: ack-emrcontainers-controller.v1.0.21
        version: 1.0.21
      - name: ack-emrcontainers-controller.v1.0.18
        version: 1.0.18
      - name: ack-emrcontainers-controller.v1.0.16
        version: 1.0.16
      - name: ack-emrcontainers-controller.v1.0.15
        version: 1.0.15
      - name: ack-emrcontainers-controller.v1.0.14
        version: 1.0.14
      - name: ack-emrcontainers-controller.v1.0.13
        version: 1.0.13
      - name: ack-emrcontainers-controller.v1.0.12
        version: 1.0.12
      - name: ack-emrcontainers-controller.v1.0.11
        version: 1.0.11
      - name: ack-emrcontainers-controller.v1.0.9
        version: 1.0.9
      - name: ack-emrcontainers-controller.v1.0.8
        version: 1.0.8
      - name: ack-emrcontainers-controller.v1.0.4
        version: 1.0.4
      - name: ack-emrcontainers-controller.v1.0.3
        version: 1.0.3
      - name: ack-emrcontainers-controller.v1.0.1
        version: 1.0.1
      - name: ack-emrcontainers-controller.v1.0.0
        version: 1.0.0
      - name: ack-emrcontainers-controller.v0.1.2
        version: 0.1.2
      - name: ack-emrcontainers-controller.v0.1.0
        version: 0.1.0
      - name: ack-emrcontainers-controller.v0.0.10
        version: 0.0.10
      - name: ack-emrcontainers-controller.v0.0.9
        version: 0.0.9
      - name: ack-emrcontainers-controller.v0.0.8
        version: 0.0.8
      - name: ack-emrcontainers-controller.v0.0.7
        version: 0.0.7
      - name: ack-emrcontainers-controller.v0.0.6
        version: 0.0.6
      - name: ack-emrcontainers-controller.v0.0.5
        version: 0.0.5
      - name: ack-emrcontainers-controller.v0.0.4
        version: 0.0.4
      - name: ack-emrcontainers-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-emrcontainers-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: volume-expander-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: volume-expander-operator.v0.3.6
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage,Monitoring
          certified: "false"
          containerImage: quay.io/redhat-cop/volume-expander-operator@sha256:e4059a66a9af7c786c8c651abb8a57a55a0e528342e3a59f8c264b40f87d6e25
          createdAt: "2022-05-05T13:25:53Z"
          description: This operator provides automation expanding pod's volumes when
            they are about to run out of space.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: volume-expander-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/volume-expander-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The purpose of the volume-expander-operator is to expand volumes when they are running out of space.
          This is achieved by using the [volume expansion feature](https://kubernetes.io/docs/concepts/storage/persistent-volumes/#expanding-persistent-volumes-claims).

          The operator periodically checks the `kubelet_volume_stats_used_bytes` and `kubelet_volume_stats_capacity_bytes published` by the kubelets to decide when to expand a volume.
          Notice that these metrics are generated only when a volume is mounted to a pod. Also the kubelet takes a minute or two to start generating accurate values for these metrics. The operator accounts for that.

          This operator works based on the following annotations to PersistentVolumeClaim resources:

          | Annotation | Default &nbsp;  &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;  | Description  |
          | :--- | :--- | :--- |
          | `volume-expander-operator.redhat-cop.io/autoexpand`  | N/A  | if set to "true" enables the volume-expander-operator to watch on this PVC  |
          | `volume-expander-operator.redhat-cop.io/polling-frequency`  | `"30s"` | How frequently to poll the volume metrics. Express this value as a valid golang [Duration](https://golang.org/pkg/time/#ParseDuration)  |
          | `volume-expander-operator.redhat-cop.io/expand-threshold-percent` | `"80"` | the percentage of used storage after which the volume will be expanded. This must be a positive integer. |
          | `volume-expander-operator.redhat-cop.io/expand-by-percent` | `"25"` | the percentage by which the volume will be expanded, relative to the current size. This must be an integer between 0 and 100 |
          | `volume-expander-operator.redhat-cop.io/expand-up-to` | MaxInt64 | the upper bound for this volume to be expanded to. The default value is the largest quantity representable and is intended to be interpreted as infinite. If the default is used it is recommend to ensure the namespace has a quota on the used storage class. |

          Note that not all of the storage driver implementations support volume expansion. It is a responsibility of the user/platform administrator to ensure that storage class and the persistent volume claim meet all the requirements needed for the volume expansion feature to work properly.

          This operator was tested with [OCS](https://www.redhat.com/en/technologies/cloud-computing/openshift-container-storage), but should work with any other storage driver that supports volume expansion.
        displayName: Volume Expander Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - volume
        - storage
        - csi
        - expansion
        links:
        - name: Volume Expander Operator
          url: https://github.com/redhat-cop/volume-expander-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/volume-expander-operator@sha256:e4059a66a9af7c786c8c651abb8a57a55a0e528342e3a59f8c264b40f87d6e25
        version: 0.3.6
      entries:
      - name: volume-expander-operator.v0.3.6
        version: 0.3.6
      - name: volume-expander-operator.v0.3.5
        version: 0.3.5
      - name: volume-expander-operator.v0.3.4
        version: 0.3.4
      - name: volume-expander-operator.v0.3.3
        version: 0.3.3
      - name: volume-expander-operator.v0.3.2
        version: 0.3.2
      - name: volume-expander-operator.v0.3.1
        version: 0.3.1
      - name: volume-expander-operator.v0.3.0
        version: 0.3.0
      - name: volume-expander-operator.v0.2.0
        version: 0.2.0
      - name: volume-expander-operator.v0.1.8
        version: 0.1.8
      - name: volume-expander-operator.v0.1.7
        version: 0.1.7
      - name: volume-expander-operator.v0.1.6
        version: 0.1.6
      - name: volume-expander-operator.v0.1.5
        version: 0.1.5
      - name: volume-expander-operator.v0.1.4
        version: 0.1.4
      - name: volume-expander-operator.v0.1.3
        version: 0.1.3
      - name: volume-expander-operator.v0.1.2
        version: 0.1.2
      - name: volume-expander-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: volume-expander-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-eventbridge-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-eventbridge-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Archive",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Endpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "EventBus",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/eventbridge-controller:1.1.1
          createdAt: "2025-09-19T20:51:55Z"
          description: AWS EventBridge controller is a service controller for managing
            EventBridge resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Archive represents the state of an AWS eventbridge Archive
              resource.
            displayName: Archive
            kind: Archive
            name: archives.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: Endpoint represents the state of an AWS eventbridge Endpoint
              resource.
            displayName: Endpoint
            kind: Endpoint
            name: endpoints.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: EventBus represents the state of an AWS eventbridge EventBus
              resource.
            displayName: EventBus
            kind: EventBus
            name: eventbuses.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: Rule represents the state of an AWS eventbridge Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.eventbridge.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EventBridge resources in AWS from within your Kubernetes cluster.

          **About Amazon EventBridge**

          EventBridge is a serverless service that uses events to connect application components together, making it easier for you to build scalable event-driven applications. Use it to route events from sources such as home-grown applications, AWS services, and third- party software to consumer applications across your organization. EventBridge provides a simple and consistent way to ingest, filter, transform, and deliver events so you can build new applications quickly.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EventBridge
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eventbridge
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EventBridge Developer Resources
          url: https://aws.amazon.com/eventbridge/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eventbridge maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/eventbridge-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-eventbridge-controller.v1.1.1
        version: 1.1.1
      - name: ack-eventbridge-controller.v1.1.0
        version: 1.1.0
      - name: ack-eventbridge-controller.v1.0.28
        version: 1.0.28
      - name: ack-eventbridge-controller.v1.0.27
        version: 1.0.27
      - name: ack-eventbridge-controller.v1.0.26
        version: 1.0.26
      - name: ack-eventbridge-controller.v1.0.25
        version: 1.0.25
      - name: ack-eventbridge-controller.v1.0.24
        version: 1.0.24
      - name: ack-eventbridge-controller.v1.0.23
        version: 1.0.23
      - name: ack-eventbridge-controller.v1.0.22
        version: 1.0.22
      - name: ack-eventbridge-controller.v1.0.21
        version: 1.0.21
      - name: ack-eventbridge-controller.v1.0.20
        version: 1.0.20
      - name: ack-eventbridge-controller.v1.0.18
        version: 1.0.18
      - name: ack-eventbridge-controller.v1.0.17
        version: 1.0.17
      - name: ack-eventbridge-controller.v1.0.16
        version: 1.0.16
      - name: ack-eventbridge-controller.v1.0.15
        version: 1.0.15
      - name: ack-eventbridge-controller.v1.0.14
        version: 1.0.14
      - name: ack-eventbridge-controller.v1.0.13
        version: 1.0.13
      - name: ack-eventbridge-controller.v1.0.12
        version: 1.0.12
      - name: ack-eventbridge-controller.v1.0.11
        version: 1.0.11
      - name: ack-eventbridge-controller.v1.0.6
        version: 1.0.6
      - name: ack-eventbridge-controller.v1.0.5
        version: 1.0.5
      - name: ack-eventbridge-controller.v1.0.4
        version: 1.0.4
      - name: ack-eventbridge-controller.v1.0.3
        version: 1.0.3
      - name: ack-eventbridge-controller.v1.0.1
        version: 1.0.1
      - name: ack-eventbridge-controller.v1.0.0
        version: 1.0.0
      - name: ack-eventbridge-controller.v0.0.7
        version: 0.0.7
      - name: ack-eventbridge-controller.v0.0.6
        version: 0.0.6
      - name: ack-eventbridge-controller.v0.0.5
        version: 0.0.5
      - name: ack-eventbridge-controller.v0.0.4
        version: 0.0.4
      - name: ack-eventbridge-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-eventbridge-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Oracle Corporation
      provider-url: https://github.com/oracle/coherence-operator
    name: coherence-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: coherence-operator.v3.5.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "coherence.oracle.com/v1",
                "kind": "Coherence",
                "metadata": {
                  "name": "coherence-sample"
                },
                "spec": {
                  "replicas": 3
                }
              },
              {
                "apiVersion": "coherence.oracle.com/v1",
                "kind": "CoherenceJob",
                "metadata": {
                  "name": "coherence-job-sample"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "false"
          containerImage: container-registry.oracle.com/middleware/coherence-operator:3.5.4
          createdAt: "2025-08-28T16:15:09Z"
          description: The Oracle Coherence Kubernetes Operator enables easy management
            of Coherence clusters in a Kubernetes environment.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/arch.amd64: supported
          operatorframework.io/arch.arm64: supported
          operatorframework.io/os.linux: supported
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/oracle/coherence-operator
          support: Oracle Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Coherence is the top level schema for the Coherence API and
              custom resource definition (CRD).
            displayName: Coherence
            kind: Coherence
            name: coherence.coherence.oracle.com
            version: v1
          - description: |-
              CoherenceJob is the top level schema for the CoherenceJob API and custom resource definition (CRD)
              for configuring Coherence Job workloads.
            displayName: Coherence Job
            kind: CoherenceJob
            name: coherencejob.coherence.oracle.com
            version: v1
        description: |
          The Oracle Coherence Kubernetes Operator enables easy management of Coherence clusters in a Kubernetes environment.

          ### Installation Guide & Configuration

          For the complete installation guide and all configuration options please refer to
          [Coherence Operator Documentation](https://oracle.github.io/coherence-operator/docs/latest/#/about/01_overview).
        displayName: Coherence Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Coherence
        - Oracle
        - In Memory Data
        - Caching
        links:
        - name: Coherence Kubernetes Operator
          url: https://oracle.github.io/coherence-operator/docs/latest
        - name: Oracle Coherence
          url: https://coherence.community
        maintainers:
        - email: jonathan.knight@oracle.com
          name: Jonathan Knight
        maturity: mature
        minKubeVersion: 1.26.0
        provider:
          name: Oracle Corporation
          url: https://github.com/oracle/coherence-operator
        relatedImages:
        - container-registry.oracle.com/middleware/coherence-operator:3.5.4
        version: 3.5.4
      entries:
      - name: coherence-operator.v3.5.4
        version: 3.5.4
      - name: coherence-operator.v3.5.2
        version: 3.5.2
      - name: coherence-operator.v3.5.0
        version: 3.5.0
      - name: coherence-operator.v3.4.3
        version: 3.4.3
      name: stable
    defaultChannel: stable
    packageName: coherence-operator
    provider:
      name: Oracle Corporation
      url: https://github.com/oracle/coherence-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nxrm-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nxrm-operator-certified.v3.58.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusRepo",
                "metadata": {
                  "name": "example-nexusrepo"
                },
                "spec": {
                  "config": {
                    "data": null,
                    "enabled": false,
                    "mountPath": "/sonatype-nexus-conf"
                  },
                  "deployment": {
                    "additionalContainers": null,
                    "additionalVolumeMounts": null,
                    "additionalVolumes": null,
                    "annotations": {},
                    "initContainers": null,
                    "postStart": {
                      "command": null
                    },
                    "preStart": {
                      "command": null
                    },
                    "terminationGracePeriodSeconds": 120
                  },
                  "deploymentStrategy": {
                    "type": "Recreate"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "path": "/",
                    "rules": null,
                    "tls": {
                      "enabled": true,
                      "secretName": "nexus-tls"
                    }
                  },
                  "nexus": {
                    "dockerPort": 5003,
                    "env": [
                      {
                        "name": "INSTALL4J_ADD_VM_PARAMS",
                        "value": "-Xms2703M -Xmx2703M -XX:MaxDirectMemorySize=2703M -XX:+UnlockExperimentalVMOptions -XX:+UseCGroupMemoryLimitForHeap"
                      }
                    ],
                    "hostAliases": [],
                    "imageName": "registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:06fe31f40ff3fd7556fc7b8f6b2efe80ada1d9079ec9aaad60d1fdc0ee0677f0",
                    "imagePullPolicy": "IfNotPresent",
                    "imagePullSecret": "",
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "path": "/",
                      "periodSeconds": 30
                    },
                    "nexusPort": 8081,
                    "podAnnotations": {},
                    "readinessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "path": "/",
                      "periodSeconds": 30
                    },
                    "resources": {},
                    "securityContext": {
                    },
                    "service": {
                      "type": "NodePort"
                    }
                  },
                  "nexusProxyRoute": {
                    "annotations": null,
                    "enabled": false,
                    "labels": null
                  },
                  "persistence": {
                    "accessMode": "ReadWriteOnce",
                    "enabled": true,
                    "storageSize": "8Gi"
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": null,
                    "enabled": false,
                    "labels": null,
                    "name": "docker",
                    "portName": "docker"
                  },
                  "secret": {
                    "data": null,
                    "enabled": false,
                    "mountPath": "/etc/secret-volume",
                    "readOnly": true
                  },
                  "service": {
                    "annotations": {},
                    "enabled": false,
                    "labels": {},
                    "ports": [
                      {
                        "name": "nexus-service",
                        "port": 80,
                        "targetPort": 80
                      }
                    ]
                  },
                  "statefulset": {
                    "enabled": false
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nxrm-operator-certified@sha256:4945ad993a10e27b358b2d9869d05f23ca761162ec088e5a833939022bc23192
          createdAt: "2023-07-14"
          description: |-
            Nexus Repository is the central source of control to efficiently manage all binaries
            and build artifacts across your DevOps pipeline.
          olm.skipRange: <3.58.1-2
          repository: https://github.com/sonatype/operator-nxrm3
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus Repository
            displayName: NexusRepo
            kind: NexusRepo
            name: nexusrepos.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Repository is the central source of control to efficiently manage all binaries
          and build artifacts across your DevOps pipeline.
          The flow of open source and third-party components into and through an organization
          creates a complex software supply chain.
          Nexus Repository delivers speed, efficiency, and quality to the governance
          and management of all dependencies, libraries, and applications for your DevOps teams.

          ## Core Capabilities

          * **Dependency Management**:
            Improves reliability with repeatable, fast access to secure dependencies
          * **Developer Productivity**:
            Streamline developer workflows by enabling the sharing of components and applications across teams
          * **Supply Chain Performance**:
            Improve speed-to-market and reduced build times with release advanced staging and component tagging
          * **CI/CD Integrations**:
            Increase DevOps scalability with integrations to the most popular build and deployment tools

          Version control systems and package registries do not scale when managing proprietary,
          open source, and third-party components.
          Organizations need a central binary and build artifact repository to manage dependencies
          across the entire software supply chain.

          ## Limitations

          High Availability Clustering (HA-C) is not supported for Nexus Repository Pro for OpenShift.

          This operator will be released on a quarterly basis.

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus Repository Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage

          Once the server instance is created by the operator and running,
          you'll want to expose the service as you see fit:
          1. Create a Route to that service for nexus.port (8081).

          By default, the Nexus Repository starts up in OSS mode until a license is installed.

          The Nexus Repository can be further configured via the NexusRepo custom resource definition:

          | Parameter                                   | Description                         | Default                                 |
          | ------------------------------------------  | ----------------------------------  | ----------------------------------------|
          | `statefulset.enabled`                       | Use statefulset instead of deployment | `false` |
          | `deploymentStrategy.type`                   | Deployment Strategy     |  `Recreate` |
          | `nexus.env`                                 | Nexus environment variables         | `See example.` |
          | `nexus.resources`                           | Nexus resource requests and limits  | `{}`                                    |
          | `nexus.dockerPort`                          | Port to access docker               | `5003`                                  |
          | `nexus.nexusPort`                           | Internal port for Nexus service     | `8081`                                  |
          | `nexus.service.type`                        | Service for Nexus                   |`NodePort`                                |
          | `nexus.service.clusterIp`                   | Specific cluster IP when service type is cluster IP. Use None for headless service |`nil`   |
          | `nexus.securityContext`                     | Security Context |
          | `nexus.labels`                              | Service labels                      | `{}`                                    |
          | `nexus.podAnnotations`                      | Pod Annotations                     | `{}`
          | `nexus.livenessProbe.initialDelaySeconds`   | LivenessProbe initial delay         | 30                                      |
          | `nexus.livenessProbe.periodSeconds`         | Seconds between polls               | 30                                      |
          | `nexus.livenessProbe.failureThreshold`      | Number of attempts before failure   | 6                                       |
          | `nexus.livenessProbe.timeoutSeconds`        | Time in seconds after liveness probe times out    | `nil`                     |
          | `nexus.livenessProbe.path`                  | Path for LivenessProbe              | /                                       |
          | `nexus.readinessProbe.initialDelaySeconds`  | ReadinessProbe initial delay        | 30                                      |
          | `nexus.readinessProbe.periodSeconds`        | Seconds between polls               | 30                                      |
          | `nexus.readinessProbe.failureThreshold`     | Number of attempts before failure   | 6                                       |
          | `nexus.readinessProbe.timeoutSeconds`       | Time in seconds after readiness probe times out    | `nil`                    |
          | `nexus.readinessProbe.path`                 | Path for ReadinessProbe             | /                                       |
          | `nexus.hostAliases`                         | Aliases for IPs in /etc/hosts       | []                                      |
          | `ingress.enabled`                           | Create an ingress for Nexus         | `true`                                  |
          | `ingress.annotations`                       | Annotations to enhance ingress configuration  | `{}`                          |
          | `ingress.tls.enabled`                       | Enable TLS                          | `true`                                 |
          | `ingress.tls.secretName`                    | Name of the secret storing TLS cert, `false` to use the Ingress' default certificate | `nexus-tls`                             |
          | `ingress.path`                              | Path for ingress rules. GCP users should set to `/*` | `/`                    |
          | `tolerations`                               | tolerations list                    | `[]`                                    |
          | `config.enabled`                            | Enable configmap                    | `false`                                 |
          | `config.mountPath`                          | Path to mount the config            | `/sonatype-nexus-conf`                  |
          | `config.data`                               | Configmap data                      | `nil`                                   |
          | `deployment.terminationGracePeriodSeconds`  | Time to allow for clean shutdown    | 120                                     |
          | `deployment.annotations`                    | Annotations to enhance deployment configuration  | `{}`                       |
          | `deployment.initContainers`                 | Init containers to run before main containers  | `nil`                        |
          | `deployment.postStart.command`              | Command to run after starting the nexus container  | `nil`                    |
          | `deployment.preStart.command`               | Command to run before starting the nexus container  | `nil`                   |
          | `deployment.additionalContainers`           | Add additional Container         | `nil`                                      |
          | `deployment.additionalVolumes`              | Add additional Volumes           | `nil`                                      |
          | `deployment.additionalVolumeMounts`         | Add additional Volume mounts     | `nil`                                      |
          | `secret.enabled`                            | Enable secret                    | `false`                                    |
          | `secret.mountPath`                          | Path to mount the secret         | `/etc/secret-volume`                       |
          | `secret.readOnly`                           | Secret readonly state            | `true`                                     |
          | `secret.data`                               | Secret data                      | `nil`                                      |
          | `service.enabled`                           | Enable additional service        | `nil`                                      |
          | `service.name`                              | Service name                     | `nil`                                      |
          | `service.portName`                          | Service port name                | `nil`                                      |
          | `service.labels`                            | Service labels                   | `nil`                                      |
          | `service.annotations`                       | Service annotations              | `nil`                                      |
          | `service.loadBalancerSourceRanges`          | Service LoadBalancer source IP whitelist | `nil`                              |
          | `service.targetPort`                        | Service port                     | `nil`                                      |
          | `service.port`                              | Port for exposing service        | `nil`                                      |
          | `route.enabled`         | Set to true to create route for additional service | `false` |
          | `route.name`            | Name of route                                      | `docker` |
          | `route.portName`        | Target port name of service                        | `docker` |
          | `route.labels`          | Labels to be added to route                        | `{}` |
          | `route.annotations`     | Annotations to be added to route                   | `{}` |
          | `route.path`            | Host name of Route e.g jenkins.example.com         | nil |
        displayName: Nexus Repository Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - repository
        - sonatype
        links:
        - name: Nexus Repository Manager
          url: https://www.sonatype.com/product-nexus-repository
        maintainers:
        - email: support@sonatype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:06fe31f40ff3fd7556fc7b8f6b2efe80ada1d9079ec9aaad60d1fdc0ee0677f0
        - registry.connect.redhat.com/sonatype/nxrm-operator-certified@sha256:4945ad993a10e27b358b2d9869d05f23ca761162ec088e5a833939022bc23192
        version: 3.58.1-2
      entries:
      - name: nxrm-operator-certified.v3.58.1-2
        version: 3.58.1-2
      name: stable
    defaultChannel: stable
    packageName: nxrm-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: AMD
      provider-url: ""
    name: amd-t2-january
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amd-t2-january.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovfect2.amd.com/v1",
                "kind": "SriovT2Card",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "amd-t2-january",
                    "app.kubernetes.io/instance": "sriovt2card-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sriovt2card",
                    "app.kubernetes.io/part-of": "amd-t2-january"
                  },
                  "name": "sriovt2card",
                  "namespace": "amd-t2"
                },
                "spec": {
                  "acceleratorSelector": {
                    "pciAddress": "0000:01:00.0"
                  },
                  "nodeSelector": {
                    "kubernetes.io/hostname": "7c-c2-55-85-ba-fe.amd.com"
                  },
                  "physicalFunction": {
                    "pfDriver": "qdma-pf",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-07-14T09:46:54Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovT2Card is the Schema for the sriovt2cards API
            displayName: Sriov T2 Card
            kind: SriovT2Card
            name: sriovt2cards.sriovfect2.amd.com
            version: v1
        description: The AMD-T2 Operator for accelerator enables the seamless integration
          and management of AMD T2 accelerator within Openshift/Kubernetes envirnoment.
        displayName: AMD-T2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - SRIOV
        - Xilinx
        - FPGA
        - Telco
        - T2
        - MultiArch
        - AMD
        - TelcoAccelerator
        links:
        - name: Amd T2 January
          url: https://amd-t2-january.domain
        maintainers:
        - email: dhirajs.lakhane@amd.com
          name: DhirajLakhane
        maturity: alpha
        provider:
          name: AMD
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:928e64203edad8f1bba23593c7be04f0f8410c6e4feb98d9e9c2d00a8ff59048
        - quay.io/amdt2operator/amd-t2-january@sha256:318f56e2670582c217b8973f69f0675e7fff7dc19d07875294f2ff2fc5403902
        version: 1.0.1
      entries:
      - name: amd-t2-january.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: amd-t2-january
    provider:
      name: AMD
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Cloudera
      provider-url: https://cloudera.com
    name: cfm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cfm-operator.v2.9.0-b96
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "Nifi",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifi-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifi",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifi-sample"
                },
                "spec": {
                  "hostName": "nifi.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifi-k8s",
                    "tag": "2.8.0-b94-nifi_1.25.0.2.3.13.0-36"
                  },
                  "replicas": 1,
                  "stateManagement": {
                    "clusterProvider": {
                      "class": "org.apache.nifi.kubernetes.state.provider.KubernetesConfigMapStateProvider",
                      "id": "kubernetes-provider"
                    }
                  },
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.8.0-b94"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              },
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "NifiRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifiregistry-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifiregistry",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifiregistry-sample"
                },
                "spec": {
                  "hostName": "registry.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s",
                    "tag": "2.8.0-b94-nifi_1.25.0.2.3.13.0-36"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.8.0-b94"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning,Big Data
          com.redhat.openshift.versions: v4.10
          containerImage: container.repository.cloudera.com/cloudera/cfm-operator@sha256:890d81188143ccf9ebe2c2fe5e039cdff3a2df53f7512555673f0468147e991e
          createdAt: "2024-12-06T19:03:24Z"
          description: Cloudera Flow Management - Kubernetes Operator for Apache NiFideploys
            and operates Apache NiFi
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: A valid Cloudera license is required.
            Contact a Cloudera account team to learn more.
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NifiRegistry is the Schema for the nifiregistries API.
            displayName: Nifi Registry
            kind: NifiRegistry
            name: nifiregistries.cfm.cloudera.com
            version: v1alpha1
          - description: Nifi is the Schema for the nifis API.
            displayName: Nifi
            kind: Nifi
            name: nifis.cfm.cloudera.com
            version: v1alpha1
        description: |-
          Cloudera Flow Management - Kubernetes Operator for Apache NiFi brings enterprise-grade NiFi and NiFi Registry deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of NiFi and NiFi Registry for enterprise deployments.
          To learn more about Cloudera Flow Management - Kubernetes Operator for Apache NiFi read this
          [deep dive blog post](https://community.cloudera.com/t5/Community-Articles/Cloudera-Flow-Management-Operator-A-technical-deep-dive/ta-p/390473).

          Cloudera Flow Management - Kubernetes Operator for Apache NiFi consists of and ships multiple components including Apache NiFi, Apache NiFi Registry,
          and a custom sidecar image for logging aggregation.

          ## Installation prerequisites

          Installing Cloudera Flow Management - Kubernetes Operator for Apache NiFi requires the a valid license as well as Cloudera credentials
          (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly.
          The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) from which installation artifacts are pulled.

          ### Obtaining a license
          Contact [Cloudera sales](https://cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets
          Before installation, create two `Secrets`. One containing your Cloudera credentials and one containing your license. Cloudera credentials are provided to you
          together with your license.

          1. Create a `Secret` containing your Cloudera credentials.
              `oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***]
               --docker-server=container.repository.cloudera.com`

          2. Create a `Secret` containing your license.
              `oc create secret generic cfm-operator-license --from-file=license.txt=[***PATH TO LICENSE FILE***]`

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces in which you
            plan to deploy NiFi clusters or NiFi Registry deployments. Cloudera recommends that you create the `Secret` in all required namespaces before installation
            if you know which namespaces you will be using to deploy NiFi clusters or NiFi Registry deployments.

          ## Learn more
          To learn more, visit the [Cloudera Flow Management - Kubernetes Operator for Apache NiFi documentation library](https://docs.cloudera.com/cfm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Flow Management - Kubernetes Operator for Apache NiFi, visit the following resources:
          * [Apache NiFi](https://nifi.apache.org/)
          * [Apache NiFi Registry](https://nifi.apache.org/registry.html)
        displayName: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        - nifiregistry
        - cloudera
        - cfm
        links:
        - name: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
          url: https://docs.cloudera.com/cfm-operator
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s@sha256:914dcde9ee97894fa6f5deec285e53bf81d196e071a997cabb40fedd52deccc3
        - container.repository.cloudera.com/cloudera/cfm-tini@sha256:6493d23e20fca8daf31823403fca3dd9b42da983464fbe788a6863e141e8b66b
        - container.repository.cloudera.com/cloudera/cfm-nifi-k8s@sha256:0d700b7eff096dfc362868345826c3a6460fa43a700273b1422e95015b7727ee
        - container.repository.cloudera.com/cloudera/cfm-operator@sha256:890d81188143ccf9ebe2c2fe5e039cdff3a2df53f7512555673f0468147e991e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86af06fb3b6c17b22be47118b0a8c797fa7fdd3c00dbe99a2d2a593ca8fc3d63
        version: 2.9.0-b96
      entries:
      - name: cfm-operator.v2.9.0-b96
        version: 2.9.0-b96
      name: alpha
    - currentCSV: cfm-operator.v2.10.0-b135
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "Nifi",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifi-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifi",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifi-sample"
                },
                "spec": {
                  "hostName": "nifi.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifi-k8s",
                    "tag": "2.9.0-b96-nifi_1.27.0.2.3.14.0-14"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.9.0-b96"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              },
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "NifiRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifiregistry-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifiregistry",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifiregistry-sample"
                },
                "spec": {
                  "hostName": "registry.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s",
                    "tag": "2.9.0-b96-nifi_1.27.0.2.3.14.0-14"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.9.0-b96"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning,Big Data
          com.redhat.openshift.versions: v4.10
          containerImage: container.repository.cloudera.com/cloudera/cfm-operator@sha256:0312b6dadd5904ea9507bb857ab89d7266bf244c93b86855e0528e72813b7a85
          createdAt: "2025-05-05T14:39:51Z"
          description: Cloudera Flow Management - Kubernetes Operator for Apache NiFideploys
            and operates Apache NiFi
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: A valid Cloudera license is required.
            Contact a Cloudera account team to learn more.
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NifiRegistry is the Schema for the nifiregistries API.
            displayName: Nifi Registry
            kind: NifiRegistry
            name: nifiregistries.cfm.cloudera.com
            version: v1alpha1
          - description: Nifi is the Schema for the nifis API.
            displayName: Nifi
            kind: Nifi
            name: nifis.cfm.cloudera.com
            version: v1alpha1
        description: |-
          Cloudera Flow Management - Kubernetes Operator for Apache NiFi brings enterprise-grade NiFi and NiFi Registry deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of NiFi and NiFi Registry for enterprise deployments.
          To learn more about Cloudera Flow Management - Kubernetes Operator for Apache NiFi read this
          [deep dive blog post](https://community.cloudera.com/t5/Community-Articles/Cloudera-Flow-Management-Operator-A-technical-deep-dive/ta-p/390473).

          Cloudera Flow Management - Kubernetes Operator for Apache NiFi consists of and ships multiple components including Apache NiFi, Apache NiFi Registry,
          and a custom sidecar image for logging aggregation.

          ## Installation prerequisites

          Installing Cloudera Flow Management - Kubernetes Operator for Apache NiFi requires the a valid license as well as Cloudera credentials
          (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly.
          The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) from which installation artifacts are pulled.

          ### Obtaining a license
          Contact [Cloudera sales](https://cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets
          Before installation, create two `Secrets`. One containing your Cloudera credentials and one containing your license. Cloudera credentials are provided to you
          together with your license.

          1. Create a `Secret` containing your Cloudera credentials.
              `oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***]
               --docker-server=container.repository.cloudera.com --namespace [***OPERATOR INSTALL NAMESPACE***]`

          2. Create a `Secret` containing your license.
              `oc create secret generic cfm-operator-license --from-file=license.txt=[***PATH TO LICENSE FILE***] --namespace [***OPERATOR INSTALL NAMESPACE***]`

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces in which you
            plan to deploy NiFi clusters or NiFi Registry deployments. Cloudera recommends that you create the `Secret` in all required namespaces before installation
            if you know which namespaces you will be using to deploy NiFi clusters or NiFi Registry deployments.

          ## Learn more
          To learn more, visit the [Cloudera Flow Management - Kubernetes Operator for Apache NiFi documentation library](https://docs.cloudera.com/cfm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Flow Management - Kubernetes Operator for Apache NiFi, visit the following resources:
          * [Apache NiFi](https://nifi.apache.org/)
          * [Apache NiFi Registry](https://nifi.apache.org/registry.html)
        displayName: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        - nifiregistry
        - cloudera
        - cfm
        links:
        - name: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
          url: https://docs.cloudera.com/cfm-operator
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/cfm-nifi-k8s@sha256:8607c81906422d65784c6dca74a37af3b7c87012bca65ce28a995366014aec86
        - container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s@sha256:52e00f102579b972500eb187e64a7662a6a46caa883df743e454b1d5111b34e5
        - container.repository.cloudera.com/cloudera/cfm-operator@sha256:0312b6dadd5904ea9507bb857ab89d7266bf244c93b86855e0528e72813b7a85
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86af06fb3b6c17b22be47118b0a8c797fa7fdd3c00dbe99a2d2a593ca8fc3d63
        - container.repository.cloudera.com/cloudera/cfm-tini@sha256:d42ad5f05a9bb283fed6dbbd820855714a31662f5b851e691f13703153e90d13
        version: 2.10.0-b135
      entries:
      - name: cfm-operator.v2.10.0-b135
        version: 2.10.0-b135
      name: stable
    defaultChannel: stable
    packageName: cfm-operator
    provider:
      name: Cloudera
      url: https://cloudera.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: placement-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: placement-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "placement.openstack.org/v1beta1",
                "kind": "PlacementAPI",
                "metadata": {
                  "name": "placement"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "placement",
                  "databaseInstance": "openstack",
                  "preserveJobs": false,
                  "replicas": 1,
                  "secret": "placement-secret",
                  "serviceUser": "placement"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:43Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PlacementAPI is the Schema for the placementapis API
            displayName: Placement API
            kind: PlacementAPI
            name: placementapis.placement.openstack.org
            version: v1beta1
        description: Placement Operator
        displayName: Placement Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Placement
        links:
        - name: Placement Operator
          url: https://github.com/openstack-k8s-operators/placement-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/placement-rhel9-operator@sha256:d3be48618b1f0f717aea162ae41876137fee8b35329d2e3e66e9d4eaccf62376
        - registry.redhat.io/rhoso/openstack-placement-api-rhel9@sha256:3308f2eccfbe7416a85443696273417bfb816df36b810ebbefa108dda12a5b19
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: placement-operator.v1.0.6
        version: 1.0.6
      - name: placement-operator.v1.0.5
        version: 1.0.5
      - name: placement-operator.v1.0.4
        version: 1.0.4
      - name: placement-operator.v1.0.3
        version: 1.0.3
      - name: placement-operator.v1.0.2
        version: 1.0.2
      - name: placement-operator.v1.0.1
        version: 1.0.1
      - name: placement-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: placement-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/cluster-nfd-operator
    name: nfd
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfd.4.16.0-202507211806
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nfd.openshift.io/v1",
                "kind": "NodeFeatureDiscovery",
                "metadata": {
                  "name": "nfd-instance",
                  "namespace": "openshift-nfd"
                },
                "spec": {
                  "customConfig": {
                    "configData": "#    - name: \"more.kernel.features\"\n#      matchOn:\n#      - loadedKMod: [\"example_kmod3\"]\n#    - name: \"more.features.by.nodename\"\n#      value: customValue\n#      matchOn:\n#      - nodename: [\"special-.*-node-.*\"]\n"
                  },
                  "operand": {
                    "image": "registry.redhat.io/openshift4/ose-node-feature-discovery-rhel9@sha256:bb65ec6a6f88688630305e081cb4bcb88aa538b5f6a3385ee28c61522dcd1662",
                    "imagePullPolicy": "IfNotPresent",
                    "servicePort": 12000
                  },
                  "workerConfig": {
                    "configData": "core:\n#  labelWhiteList:\n#  noPublish: false\n  sleepInterval: 60s\n#  sources: [all]\n#  klog:\n#    addDirHeader: false\n#    alsologtostderr: false\n#    logBacktraceAt:\n#    logtostderr: true\n#    skipHeaders: false\n#    stderrthreshold: 2\n#    v: 0\n#    vmodule:\n##   NOTE: the following options are not dynamically run-time \n##          configurable and require a nfd-worker restart to take effect\n##          after being changed\n#    logDir:\n#    logFile:\n#    logFileMaxSize: 1800\n#    skipLogHeaders: false\nsources:\n#  cpu:\n#    cpuid:\n##     NOTE: whitelist has priority over blacklist\n#      attributeBlacklist:\n#        - \"BMI1\"\n#        - \"BMI2\"\n#        - \"CLMUL\"\n#        - \"CMOV\"\n#        - \"CX16\"\n#        - \"ERMS\"\n#        - \"F16C\"\n#        - \"HTT\"\n#        - \"LZCNT\"\n#        - \"MMX\"\n#        - \"MMXEXT\"\n#        - \"NX\"\n#        - \"POPCNT\"\n#        - \"RDRAND\"\n#        - \"RDSEED\"\n#        - \"RDTSCP\"\n#        - \"SGX\"\n#        - \"SSE\"\n#        - \"SSE2\"\n#        - \"SSE3\"\n#        - \"SSE4.1\"\n#        - \"SSE4.2\"\n#        - \"SSSE3\"\n#      attributeWhitelist:\n#  kernel:\n#    kconfigFile: \"/path/to/kconfig\"\n#    configOpts:\n#      - \"NO_HZ\"\n#      - \"X86\"\n#      - \"DMI\"\n  pci:\n    deviceClassWhitelist:\n      - \"0200\"\n      - \"03\"\n      - \"12\"\n    deviceLabelFields:\n#      - \"class\"\n      - \"vendor\"\n#      - \"device\"\n#      - \"subsystem_vendor\"\n#      - \"subsystem_device\"\n#  usb:\n#    deviceClassWhitelist:\n#      - \"0e\"\n#      - \"ef\"\n#      - \"fe\"\n#      - \"ff\"\n#    deviceLabelFields:\n#      - \"class\"\n#      - \"vendor\"\n#      - \"device\"\n#  custom:\n#    - name: \"my.kernel.feature\"\n#      matchOn:\n#        - loadedKMod: [\"example_kmod1\", \"example_kmod2\"]\n#    - name: \"my.pci.feature\"\n#      matchOn:\n#        - pciId:\n#            class: [\"0200\"]\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#        - pciId :\n#            vendor: [\"8086\"]\n#            device: [\"1000\", \"1100\"]\n#    - name: \"my.usb.feature\"\n#      matchOn:\n#        - usbId:\n#          class: [\"ff\"]\n#          vendor: [\"03e7\"]\n#          device: [\"2485\"]\n#        - usbId:\n#          class: [\"fe\"]\n#          vendor: [\"1a6e\"]\n#          device: [\"089a\"]\n#    - name: \"my.combined.feature\"\n#      matchOn:\n#        - pciId:\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#          loadedKMod : [\"vendor_kmod1\", \"vendor_kmod2\"]\n"
                  }
                }
              },
              {
                "apiVersion": "nfd.openshift.io/v1alpha1",
                "kind": "NodeFeatureRule",
                "metadata": {
                  "name": "my-sample-rule-object"
                },
                "spec": {
                  "rules": [
                    {
                      "labels": {
                        "my-sample-feature": "true"
                      },
                      "matchFeatures": [
                        {
                          "feature": "kernel.loadedmodule",
                          "matchExpressions": {
                            "dummy": {
                              "op": "Exists"
                            }
                          }
                        },
                        {
                          "feature": "kernel.config",
                          "matchExpressions": {
                            "X86": {
                              "op": "In",
                              "value": [
                                "y"
                              ]
                            }
                          }
                        }
                      ],
                      "name": "my sample rule"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery,OpenShift Optional
          containerImage: registry.redhat.io/openshift4/ose-cluster-nfd-rhel9-operator@sha256:c1b6a5b8a59c08458d57c0fc2bb2bc9b056f006d15edb4d3161a3d1c0bc97ae6
          description: |-
            The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

            NFD consists  of the following software components:

            The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

            ##NFD-Master
            NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

            ##NFD-Worker
            NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

            ##NFD-Topology-Updater
            NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.0-202507211806'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-nfd
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/internal-objects: '["noderesourcetopologies.topology.node.k8s.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/cluster-nfd-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              The NodeFeatureDiscovery instance is the CustomResource being watched by the NFD-Operator, and holds all the needed information to setup the behaviour of the master and worker pods
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
          - description: |
              NodeFeatureRule resource specifies a configuration for feature-based customization of node objects, such as node labeling.
            kind: NodeFeatureRule
            name: nodefeaturerules.nfd.openshift.io
            version: v1alpha1
          - kind: NodeFeature
            name: nodefeatures.nfd.openshift.io
            version: v1alpha1
        description: |-
          The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

          NFD consists  of the following software components:

          The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

          ##NFD-Master
          NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

          ##NFD-Worker
          NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

          ##NFD-Topology-Updater
          NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
        displayName: Node Feature Discovery Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - feature-discovery
        - feature-detection
        - node-labels
        links:
        - name: Node Feature Discovery Operator
          url: https://docs.openshift.com/container-platform/4.16/hardware_enablement/psap-node-feature-discovery-operator.html
        - name: Node Feature Discovery Documentation
          url: https://kubernetes-sigs.github.io/node-feature-discovery/stable/get-started/index.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/cluster-nfd-operator
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:bcca89fb000b0fe8722ad7807b62a43ef3ae8fa76db1d790c285a975bfecb1aa
        - registry.redhat.io/openshift4/ose-cluster-nfd-rhel9-operator@sha256:c1b6a5b8a59c08458d57c0fc2bb2bc9b056f006d15edb4d3161a3d1c0bc97ae6
        - registry.redhat.io/openshift4/ose-node-feature-discovery-rhel9@sha256:bb65ec6a6f88688630305e081cb4bcb88aa538b5f6a3385ee28c61522dcd1662
        version: 4.16.0-202507211806
      entries:
      - name: nfd.4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: nfd.4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: nfd.4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: nfd.4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: nfd.4.16.0-202504271139
        version: 4.16.0-202504271139
      - name: nfd.4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: nfd.4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: nfd.4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: nfd.4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: nfd.4.16.0-202501271512
        version: 4.16.0-202501271512
      - name: nfd.4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: nfd.4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: nfd.4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: nfd.4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: nfd.4.16.0-202410251436
        version: 4.16.0-202410251436
      - name: nfd.4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: nfd.4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: nfd.4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: nfd.4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: nfd.4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: nfd.4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: nfd.4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: nfd.4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: nfd.4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: nfd.4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: nfd.4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: nfd.4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: nfd.4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: nfd.4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: nfd
    provider:
      name: Red Hat
      url: https://github.com/openshift/cluster-nfd-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhpam-kogito-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhpam-kogito-operator.7.13.5-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "ruleunit-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogitoinfra-sample"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "kafka.strimzi.io/v1beta2",
                    "kind": "Kafka",
                    "name": "kogito-kafka"
                  }
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator@sha256:19497d4aff0ea3ea114dccb0d0e265458bb14ba6d325353eb82b554547dbd2e4
          description: RHPAM Kogito Operator for deployment and management of Kogito
            services.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Process Automation
            Manager", "Red Hat Decision Manager"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/rhpam-kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.rhpam.kiegroup.org
            version: v1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.rhpam.kiegroup.org
            version: v1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.rhpam.kiegroup.org
            version: v1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.rhpam.kiegroup.org
            version: v1
        description: |-
          RHPAM Kogito Operator is designed for deployment and management of Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Builds a RHPAM Kogito service from a remote git repository and creates a custom image for the given runtime: Red Hat build of Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Deploy RHPAM Kogito service from user provided image.
          * Setup Grafana Dashboard when Grafana Operator is available in cluster.
          * Create Prometheus monitor object when Prometheus operator is available in cluster.

          **IMPORTANT!** RHPAM Kogito Operator may need Grafana and Prometheus Operator to configure monitoring in cases if you do not provide these services beforehand. If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the RHPAM Kogito Operator.
        displayName: RHPAM Kogito Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-process-automation-manager
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_process_automation_manager/7.13/html/getting_started_with_red_hat_build_of_kogito_in_red_hat_process_automation_manager/assembly-deploying-kogito-microservices-on-openshift
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator@sha256:19497d4aff0ea3ea114dccb0d0e265458bb14ba6d325353eb82b554547dbd2e4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9492c67eaa7ab37df6f5fa623c4aac955b7d7ffaab9ae2e57dea745e82907e40
        version: 7.13.5-4
      entries:
      - name: rhpam-kogito-operator.7.13.5-4
        version: 7.13.5-4
      - name: rhpam-kogito-operator.7.13.5-3
        version: 7.13.5-3
      - name: rhpam-kogito-operator.7.13.5-2
        version: 7.13.5-2
      - name: rhpam-kogito-operator.7.13.5-1
        version: 7.13.5-1
      - name: rhpam-kogito-operator.7.13.4-3
        version: 7.13.4-3
      - name: rhpam-kogito-operator.7.13.4-2
        version: 7.13.4-2
      - name: rhpam-kogito-operator.7.13.4-1
        version: 7.13.4-1
      - name: rhpam-kogito-operator.7.13.3-1
        version: 7.13.3-1
      - name: rhpam-kogito-operator.7.13.2-3
        version: 7.13.2-3
      - name: rhpam-kogito-operator.7.13.2-2
        version: 7.13.2-2
      - name: rhpam-kogito-operator.7.13.2-1
        version: 7.13.2-1
      - name: rhpam-kogito-operator.7.13.1-2
        version: 7.13.1-2
      - name: rhpam-kogito-operator.7.13.1-1
        version: 7.13.1-1
      - name: rhpam-kogito-operator.7.13.0-1
        version: 7.13.0-1
      - name: rhpam-kogito-operator.v7.12.0-4
        version: 7.12.0-4
      - name: rhpam-kogito-operator.v7.12.0-3
        version: 7.12.0-3
      - name: rhpam-kogito-operator.v7.12.0-2
        version: 7.12.0-2
      name: 7.x
    defaultChannel: 7.x
    packageName: rhpam-kogito-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://skupper.io
    name: skupper-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-operator.v1.4.8-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus@sha256:65602221cdff7f718d4149f4dead167d05dbbe0cea1eaa8bf4b98be588135583
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:c665677c639699d6a5ef24aed5a1a3d9c0a5ee25f077a6db42a9fc6247d93132
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:eedde9f2019d4e919612bcc18f427c0d88b4846e773751af231ba0e560d48869
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:1fd908e80ef21fe7d7051c400cb099f2f3ca51c8cdd698705ef9784d34688911
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:4c9cc9df83baa1ef37cfb429babe5b86a962b8bb096209b3fb74291c67c94000
        version: 1.4.8-rh-1
      entries:
      - name: skupper-operator.v1.4.8-rh-1
        version: 1.4.8-rh-1
      - name: skupper-operator.v1.4.7-rh-3
        version: 1.4.7-rh-3
      - name: skupper-operator.v1.4.7-rh-2
        version: 1.4.7-rh-2
      - name: skupper-operator.v1.4.7-rh-1
        version: 1.4.7-rh-1
      - name: skupper-operator.v1.4.5-rh-4
        version: 1.4.5-rh-4
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1
    - currentCSV: skupper-operator.v1.4.8-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:eedde9f2019d4e919612bcc18f427c0d88b4846e773751af231ba0e560d48869
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:1fd908e80ef21fe7d7051c400cb099f2f3ca51c8cdd698705ef9784d34688911
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:4c9cc9df83baa1ef37cfb429babe5b86a962b8bb096209b3fb74291c67c94000
        - registry.redhat.io/openshift4/ose-prometheus@sha256:65602221cdff7f718d4149f4dead167d05dbbe0cea1eaa8bf4b98be588135583
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:c665677c639699d6a5ef24aed5a1a3d9c0a5ee25f077a6db42a9fc6247d93132
        version: 1.4.8-rh-1
      entries:
      - name: skupper-operator.v1.4.8-rh-1
        version: 1.4.8-rh-1
      - name: skupper-operator.v1.4.7-rh-3
        version: 1.4.7-rh-3
      - name: skupper-operator.v1.4.7-rh-2
        version: 1.4.7-rh-2
      - name: skupper-operator.v1.4.7-rh-1
        version: 1.4.7-rh-1
      - name: skupper-operator.v1.4.5-rh-4
        version: 1.4.5-rh-4
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.4
    - currentCSV: skupper-operator.v1.5.5-rh-6
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:9e9e1f2288f0218e3ba2abf4bebac9731d7cf9f9c567e0f6de6767b0b66cbee4
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:a5d7c1b2b9db672548126c0873d062a55667d93291fd3e32809d3c9c900c548e
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:9e9e1f2288f0218e3ba2abf4bebac9731d7cf9f9c567e0f6de6767b0b66cbee4
        - registry.redhat.io/openshift4/ose-prometheus@sha256:53afaf53975807765fc3777314edd218c789c17680c1f57a8e33a8f9a76f38bf
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:3e78ec8cdb729a704d5a26e54697f285ba7c0b9afee6160b7ecc703876120dea
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:be9dbcd2a3a1e1e536921390dcfd283fc45dc4423d2705376ef37a9fb53f7cf1
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:1de071dabf620265252f3922c632f17f676ba716ce4c3df931a0ccb3871dbb6c
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:5f8621d0013febcecbed4d7c82db11a4dded99ad8ce6b6d83858b9bdbe79ee5a
        version: 1.5.5-rh-6
      entries:
      - name: skupper-operator.v1.5.5-rh-6
        version: 1.5.5-rh-6
      - name: skupper-operator.v1.5.5-rh-5
        version: 1.5.5-rh-5
      - name: skupper-operator.v1.5.5-rh-4
        version: 1.5.5-rh-4
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.5
    - currentCSV: skupper-operator.v1.8.7-rh-2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:b0d43628b779a101a24e4b271318a34c21f3b71bcba2bf20804f941a93d27455
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.8.7-rh-2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:56b8fde75a6e7df235c560a0f4b9114398da5ba8b20f75f1ea7338f48ce19506
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:d7d252cff08ba1589afd8707ed349f87d3b6214906a4fb49932bda0a4d19fee4
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:5ea36e96f878577104ab7601e57e7a2a233f11dbdc533454f95e2e14a066d21f
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:3adeef827eca0382cd410f7c3e99a0fee9b50791cb0e0c6fc130b72a2550684f
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:b0d43628b779a101a24e4b271318a34c21f3b71bcba2bf20804f941a93d27455
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:f798745e799515065af099a816057b9c98a741e951d30a0f45ed47ca50e6724e
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:61542c5f0989f51ac5d6c96f5d0e29aca75614b79e7626299c77c5a74636b564
        version: 1.8.7-rh-2
      entries:
      - name: skupper-operator.v1.8.7-rh-2
        version: 1.8.7-rh-2
      - name: skupper-operator.v1.8.7-rh-1
        version: 1.8.7-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.8
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.9
    - currentCSV: skupper-operator.v2.1.1-rh-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
          createdAt: "2025-08-19T20:43:19Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:36872037066fae806cbf88cab673229f25e592295972aa18227827233362eec5
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:c01d0eab0841b1a5fddb58a0ae110beb464b91390f02662590bebc7b2a7213b6
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e8df53454cf12a14338790af6ed209798010990ca3320096e8b44acf901c4123
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:1a290a5ac8b9d1b082e740054675fe432aa424c2f55dd662caf67c740044f229
        version: 2.1.1-rh-3
      entries:
      - name: skupper-operator.v2.1.1-rh-3
        version: 2.1.1-rh-3
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2
    - currentCSV: skupper-operator.v2.0.1-rh-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:b5c73dc59a9184cd120c2ee35256cc097c790b8ef488af0c732d5efe3192b768
          createdAt: "2025-05-29T13:35:40Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.0.1-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:92f44cff160874efe57a985609e599b13c49c0c4cef846dccc80833af2ae74f3
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:e239c859a1f49816ae3068455e703b8a12919f11158b0f6ee0e6bfb31e8a8082
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:f2a41674ed196c089748d6386c18c14580e502de33235ef23c37eddeb6350f41
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:95062ff1256135cfa17907e82b09b2a211371ea0e76ccc711b7c737e040d87e9
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:498b5bb648e5bbaad07663941600173abac7d60a10744305f619509ddf7563e0
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:bd3f72544841c91c2d573a65ac3c988756f3523e17e0988027793b0f5f3ec3d1
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:b5c73dc59a9184cd120c2ee35256cc097c790b8ef488af0c732d5efe3192b768
        version: 2.0.1-rh-2
      entries:
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2.0
    - currentCSV: skupper-operator.v2.1.1-rh-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
          createdAt: "2025-08-19T20:43:19Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:c01d0eab0841b1a5fddb58a0ae110beb464b91390f02662590bebc7b2a7213b6
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e8df53454cf12a14338790af6ed209798010990ca3320096e8b44acf901c4123
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:1a290a5ac8b9d1b082e740054675fe432aa424c2f55dd662caf67c740044f229
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:36872037066fae806cbf88cab673229f25e592295972aa18227827233362eec5
        version: 2.1.1-rh-3
      entries:
      - name: skupper-operator.v2.1.1-rh-3
        version: 2.1.1-rh-3
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2.1
    defaultChannel: stable-2
    packageName: skupper-operator
    provider:
      name: Red Hat
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://horreum.hyperfoil.io
    name: horreum-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: horreum-operator.v0.7.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hyperfoil.io/v1alpha1",
                "kind": "Horreum",
                "metadata": {
                  "name": "horreum"
                },
                "spec": {
                  "nodeHost": "127.0.0.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/hyperfoil/horreum-operator:0.7.9
          createdAt: "2023-06-13T10:42:39Z"
          description: Performance results repository
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/hyperfoil/horreum-operator
          support: Red Hat Application Runtimes Performance Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Horreum is the object configuring Horreum performance results
              repository
            displayName: Horreum
            kind: Horreum
            name: horreums.hyperfoil.io
            version: v1alpha1
        description: Performance results repository
        displayName: Horreum
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - performance
        - repository
        - database
        - results
        - hyperfoil
        - benchmark
        links:
        - name: Horreum Operator
          url: https://horreum-operator.domain
        maintainers:
        - email: rvansa@redhat.com
          name: Radim Vansa
        - email: spederse@redhat.com
          name: Ståle W. Pedersen
        - email: wreicher@redhat.com
          name: Will Reichert
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat, Inc.
          url: https://horreum.hyperfoil.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/hyperfoil/horreum-operator:0.7.9
        version: 0.7.9
      entries:
      - name: horreum-operator.v0.7.9
        version: 0.7.9
      - name: horreum-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: horreum-operator
    provider:
      name: Red Hat, Inc.
      url: https://horreum.hyperfoil.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/stolostron/multicluster-global-hub
    name: multicluster-global-hub-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-global-hub-operator.v1.4.0-dev.20250302
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2025-02-14T08:24:47Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: planning
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/postgresql-16:9.5-1732622748
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-inventory/inventory-api:latest
        - quay.io/prometheuscommunity/postgres-exporter:v0.15.0
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-relations/spicedb-operator:latest
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-relations/spicedb:latest
        - quay.io/stolostron/grafana:2.12.0-SNAPSHOT-2024-09-03-21-11-25
        - quay.io/stolostron/multicluster-global-hub-operator:v1.4.0
        - quay.io/stolostron/multicluster-global-hub-agent:v1.4.0
        - quay.io/stolostron/multicluster-global-hub-manager:v1.4.0
        version: 1.4.0-dev.20250302
      entries:
      - name: multicluster-global-hub-operator.v1.4.0-dev.20250302
        version: 1.4.0-dev.20250302
      - name: multicluster-global-hub-operator.v1.4.0-dev.20250225
        version: 1.4.0-dev.20250225
      - name: multicluster-global-hub-operator.v1.3.0-dev.20240911
        version: 1.3.0-dev.20240911
      - name: multicluster-global-hub-operator.v1.3.0-dev.20240724
        version: 1.3.0-dev.20240724
      - name: multicluster-global-hub-operator.v1.2.0-dev.20240516
        version: 1.2.0-dev.20240516
      - name: multicluster-global-hub-operator.v1.2.0-dev
        version: 1.2.0-dev
      - name: multicluster-global-hub-operator.v1.1.0-dev
        version: 1.1.0-dev
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: multicluster-global-hub-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-operator-globalhub-1-5@sha256:4db1251323bd64930461cb5025da41baffcf5c00c51ea4e49940b7130971b510
          createdAt: "2025-06-10T06:58:35Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-operator-globalhub-1-5@sha256:4db1251323bd64930461cb5025da41baffcf5c00c51ea4e49940b7130971b510
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-manager-globalhub-1-5@sha256:01253426d65f1297a1955a8393051e6a4b1df5e5249774fc5189a75f35e7d644
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-agent-globalhub-1-5@sha256:6f7520eac83abec7780f831d8473544d0b8db43fc6e77d59e59317b09eb1f6c1
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/glo-grafana-globalhub-1-5@sha256:73fdf00538e63595b7cc92d484e087ff815579a95d7259c579c6e7dc1beacff6
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/postgres-exporter-globalhub-1-5@sha256:1108572a8bdafde926254f1ff94c2b22e048706be660b07c793e9f9e038a0f4a
        version: 1.5.0
      entries:
      - name: multicluster-global-hub-operator.v1.5.0
        version: 1.5.0
      - name: multicluster-global-hub-operator.v1.4.0
        version: 1.4.0
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-1.5
    - currentCSV: multicluster-global-hub-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2024-03-05T02:13:54Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.32.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:latest
        version: 1.1.0
      entries:
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.10
    - currentCSV: multicluster-global-hub-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2024-07-03T06:52:02Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v1.2.0
        version: 1.2.1
      entries:
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.11
    - currentCSV: multicluster-global-hub-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator@sha256:636aea861da80fc36fb27a419a18c5b3fbb1bc072e8ae33c75e1cc47e2c05f75
          createdAt: "2024-10-10T05:42:48Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          kafka-zookeeper-0                                       1/1     Running   0          21m
          kafka-zookeeper-1                                       1/1     Running   0          21m
          kafka-zookeeper-2                                       1/1     Running   0          21m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgres-0                      2/2     Running   0          22m
          ```
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:636aea861da80fc36fb27a419a18c5b3fbb1bc072e8ae33c75e1cc47e2c05f75
        - quay.io/prometheuscommunity/postgres-exporter@sha256:386b12d19eab2a37d7cd8ca8b4c7491cc7a830d9581f49af6c98a393da9605e6
        - quay.io/stolostron/inventory-api@sha256:a15e62ee070dba5d74cf336b3b89b467593c5a33cfe3623d922f0af4fd7df392
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/release-globalhub-1-3/multicluster-global-hub-manager-globalhub-1-3@sha256:7656d01a4d7b39c61c3249695d68011d9e0df308627b9efd117686e901281ab1
        - quay.io/stolostron/multicluster-global-hub-manager@sha256:715cb5d0c6c407ec5c0303fc7e6d83fc77ce7e06775e169a475f7c075f1e7000
        - quay.io/stolostron/grafana@sha256:719b320a05d10158fadce5ae2fc6e4254a18957a6608b6f81e876cf46d3a6897
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        version: 1.3.0
      entries:
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.12
    - currentCSV: multicluster-global-hub-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator@sha256:085e9fc6cdc6a42df10a38856c8ccff25e5acffaba67ef95da72027e90e507fc
          createdAt: "2025-02-14T08:24:47Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:085e9fc6cdc6a42df10a38856c8ccff25e5acffaba67ef95da72027e90e507fc
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-inventory/inventory-api@sha256:6a55a1919ecd92d94e05c12be1b6450956f5fda2ce17af7027cac3806c4970a4
        - quay.io/prometheuscommunity/postgres-exporter@sha256:386b12d19eab2a37d7cd8ca8b4c7491cc7a830d9581f49af6c98a393da9605e6
        - quay.io/stolostron/multicluster-global-hub-manager@sha256:4cd602ff6f8dd951848dc8a95c856004295c9f4710f77d4dafcc0735c1492395
        - quay.io/stolostron/multicluster-global-hub-agent@sha256:8dfc33dfed33a85733ee5e02dcd4f05662c75263791e0961ba17cbb3d0c340c9
        - quay.io/stolostron/grafana@sha256:719b320a05d10158fadce5ae2fc6e4254a18957a6608b6f81e876cf46d3a6897
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        version: 1.4.0
      entries:
      - name: multicluster-global-hub-operator.v1.4.0
        version: 1.4.0
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.13
    - currentCSV: multicluster-global-hub-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "storage": {
                    "name": "storage-secret"
                  },
                  "transport": {
                    "name": "transport-secret"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.5.0
          createdAt: "2022-08-23T20:00:31Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha1
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.5 or later needs to be installed.\n- PostgreSQL needs to be installed
          and you must create database for the multicluster global hub. You must create
          a secret named postgresql-secret that contains the database access credential
          in `open-cluster-management` namespace. You can run the following command
          to create the secret: \n\n```\nkubectl create secret generic storage-secret
          -n open-cluster-management \\\n  --from-literal=database_uri=<postgresql-uri>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)
          to install postgres in `hoh-postgres` namespace and create the secret `storage-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n- Kafka needs to be installed,
          and you must create two sections, `spec` and `status`. You need to create
          a secret named kafka-secret that contains the kafka access information should
          be created in `open-cluster-management` namespace. You can run the following
          command to create the secret: \n\n```\nkubectl create secret generic transport-secret
          -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n  --from-literal=CA=<CA-for-kafka-server>\n```\n >_Note:_ There is a
          sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `transport-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n## How to Install\n\nInstall the
          Multicluster Global Hub Operator by following the instructions that are
          displayed after you select the Install button. The operator must be installed
          in the same namespace as Red Hat Advanced Cluster Management for Kubernetes
          (RHACM), by default, in `open-cluster-management` namespace.\n\nA pod will
          be created in `open-cluster-management` namespace\n\n```\n$ kubectl get
          pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          19s\n```\n\nThe operator is
          now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installing the operator,
          create an instance of the MulticlusterGlobalHub resource to install the
          Multicluster Global Hub.\n\nThe pods of Multicluster Global Hub are created
          in the `open-cluster-management` namespace:\n\n```\n$ kubectl get pods -n
          open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\nmulticluster-global-hub-rbac-77888b7dc4-z9vbp
          \                   1/1     Running   0          10s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.5.0
        version: 0.5.0
      entries:
      - name: multicluster-global-hub-operator.v0.5.0
        version: 0.5.0
      name: release-2.6
    - currentCSV: multicluster-global-hub-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha2",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "largeScale": {
                      "kafka": {
                        "name": "transport-secret"
                      },
                      "postgres": {
                        "name": "storage-secret"
                      }
                    },
                    "type": "largeScale"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.6.0
          createdAt: "2023-02-09T20:00:31Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha2
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.7 or later needs to be installed.\n- PostgreSQL needs to be installed
          and you must create database for the multicluster global hub. You must create
          a secret named postgresql-secret that contains the database access credential
          in `open-cluster-management` namespace. You can run the following command
          to create the secret: \n\n```\nkubectl create secret generic storage-secret
          -n open-cluster-management \\\n  --from-literal=database_uri=<postgresql-uri>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)
          to install postgres in `hoh-postgres` namespace and create the secret `storage-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n- Kafka needs to be installed,
          and you must create two sections, `spec` and `status`. You need to create
          a secret named kafka-secret that contains the kafka access information should
          be created in `open-cluster-management` namespace. You can run the following
          command to create the secret: \n\n```\nkubectl create secret generic transport-secret
          -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n  --from-literal=CA=<CA-for-kafka-server>\n```\n >_Note:_ There is a
          sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `transport-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n## How to Install\n\nInstall the
          Multicluster Global Hub Operator by following the instructions that are
          displayed after you select the Install button. The operator must be installed
          in the same namespace as Red Hat Advanced Cluster Management for Kubernetes
          (RHACM), by default, in `open-cluster-management` namespace.\n\nA pod will
          be created in `open-cluster-management` namespace\n\n```\n$ kubectl get
          pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          19s\n```\n\nThe operator is
          now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installing the operator,
          create an instance of the MulticlusterGlobalHub resource to install the
          Multicluster Global Hub.\n\nThe pods of Multicluster Global Hub are created
          in the `open-cluster-management` namespace:\n\n```\n$ kubectl get pods -n
          open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: multicluster-global-hub-operator.v0.6.0
        version: 0.6.0
      name: release-2.7
    - currentCSV: multicluster-global-hub-operator.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha3",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "largeScale": {
                      "kafka": {
                        "transportFormat": "cloudEvents"
                      }
                    },
                    "type": "largeScale"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.7.0
          createdAt: "2023-07-19T07:53:07Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha3
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.7 or later needs to be installed.\n## How to Install\n\nInstall the Multicluster
          Global Hub Operator by following the instructions that are displayed after
          you select the Install button. The operator must be installed in the same
          namespace as Red Hat Advanced Cluster Management for Kubernetes (RHACM),
          by default, in `open-cluster-management` namespace.\n\nA pod will be created
          in `open-cluster-management` namespace\n\n```\n$ kubectl get pods -n open-cluster-management\nNAME
          \                                                            READY   STATUS
          \   RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp                1/1
          \    Running   0          19s\n```\n\nThe operator is now providing new
          Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installed the operator,
          create an instance of the MulticlusterGlobalHub resource to instantiate
          the Multicluster Global Hub.\n\nYou need to prepare:\n\n- PostgreSQL needs
          to be installed and you must create database for the multicluster global
          hub. You must create a secret named multicluster-global-hub-storage that
          contains the database access credential in `open-cluster-management` namespace.
          You can run the following command to create the secret: \n\n```\nkubectl
          create secret generic multicluster-global-hub-storage -n open-cluster-management
          \\\n  --from-literal=database_uri=<postgresql-uri> \\\n --from-file=ca.crt=<CA-for-postgres-server>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)(Note:
          the client version of kubectl must be v1.21+) to install postgres in `hoh-postgres`
          namespace and create the secret `multicluster-global-hub-storage` in namespace
          `open-cluster-management` automatically. To override the secret namespace,
          set `TARGET_NAMESPACE` environment variable to the RHACM installation namespace
          before executing the script.\n\n- Kafka needs to be installed, and you must
          create three topics, `spec` `status` and `event`. You need to create a secret
          named multicluster-global-hub-transport that contains the kafka access information
          should be created in `open-cluster-management` namespace. You can run the
          following command to create the secret: \n\n```\nkubectl create secret generic
          multicluster-global-hub-transport -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n      --from-file=ca.crt=<CA-cert-for-kafka-server> --from-file=client.crt=<Client-cert-for-kafka-server>
          --from-file=client.key=<Client-key-for-kafka-server> \n```\n >_Note:_ There
          is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `multicluster-global-hub-transport`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\nOnce an instance of the MulticlusterGlobalHub
          is created, the following pods are created in the `open-cluster-management`
          namespace:\n\n```\n$ kubectl get pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-grafana-79f8d59f58-qwcjl
          \                1/1     Running   0          22s\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: beta
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.7.0
        version: 0.7.0
      entries:
      - name: multicluster-global-hub-operator.v0.7.0
        version: 0.7.0
      name: release-2.8
    - currentCSV: multicluster-global-hub-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v1.0.0
          createdAt: "2023-10-16T03:07:33Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the MultiCluster Global Hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.7 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```

          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v1.0.0
        version: 1.0.0
      entries:
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.9
    defaultChannel: release-1.5
    packageName: multicluster-global-hub-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/stolostron/multicluster-global-hub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhbk-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhbk-operator.v22.0.13-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
          createdAt: "2024-09-17T11:20:22Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:884afe0ca1698d85a02aa47b6d60b8222e50d70ad280db6f98e1a4bbdbe13c40
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
        version: 22.0.13-opr.1
      entries:
      - name: rhbk-operator.v22.0.13-opr.1
        version: 22.0.13-opr.1
      - name: rhbk-operator.v22.0.12-opr.1
        version: 22.0.12-opr.1
      - name: rhbk-operator.v22.0.11-opr.2
        version: 22.0.11-opr.2
      - name: rhbk-operator.v22.0.11-opr.1
        version: 22.0.11-opr.1
      - name: rhbk-operator.v22.0.10-opr.1
        version: 22.0.10-opr.1
      - name: rhbk-operator.v22.0.9-opr.1
        version: 22.0.9-opr.1
      - name: rhbk-operator.v22.0.8-opr.2
        version: 22.0.8-opr.2
      - name: rhbk-operator.v22.0.8-opr.1
        version: 22.0.8-opr.1
      - name: rhbk-operator.v22.0.7-opr.1
        version: 22.0.7-opr.1
      - name: rhbk-operator.v22.0.6-opr.1
        version: 22.0.6-opr.1
      name: stable-v22
    - currentCSV: rhbk-operator.v22.0.13-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
          createdAt: "2024-09-17T11:20:22Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:884afe0ca1698d85a02aa47b6d60b8222e50d70ad280db6f98e1a4bbdbe13c40
        version: 22.0.13-opr.1
      entries:
      - name: rhbk-operator.v22.0.13-opr.1
        version: 22.0.13-opr.1
      - name: rhbk-operator.v22.0.12-opr.1
        version: 22.0.12-opr.1
      - name: rhbk-operator.v22.0.11-opr.2
        version: 22.0.11-opr.2
      - name: rhbk-operator.v22.0.11-opr.1
        version: 22.0.11-opr.1
      - name: rhbk-operator.v22.0.10-opr.1
        version: 22.0.10-opr.1
      - name: rhbk-operator.v22.0.9-opr.1
        version: 22.0.9-opr.1
      - name: rhbk-operator.v22.0.8-opr.2
        version: 22.0.8-opr.2
      - name: rhbk-operator.v22.0.8-opr.1
        version: 22.0.8-opr.1
      - name: rhbk-operator.v22.0.7-opr.1
        version: 22.0.7-opr.1
      - name: rhbk-operator.v22.0.6-opr.1
        version: 22.0.6-opr.1
      name: stable-v22.0
    - currentCSV: rhbk-operator.v24.0.11-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
          createdAt: "2025-03-26T02:27:51Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:d339ecf788d1681ba9945a2ea63ca675caec863ec75f497ef916103f4c898003
        version: 24.0.11-opr.1
      entries:
      - name: rhbk-operator.v24.0.11-opr.1
        version: 24.0.11-opr.1
      - name: rhbk-operator.v24.0.10-opr.1
        version: 24.0.10-opr.1
      - name: rhbk-operator.v24.0.9-opr.1
        version: 24.0.9-opr.1
      - name: rhbk-operator.v24.0.8-opr.1
        version: 24.0.8-opr.1
      - name: rhbk-operator.v24.0.7-opr.1
        version: 24.0.7-opr.1
      - name: rhbk-operator.v24.0.6-opr.1
        version: 24.0.6-opr.1
      - name: rhbk-operator.v24.0.5-opr.1
        version: 24.0.5-opr.1
      - name: rhbk-operator.v24.0.4-opr.1
        version: 24.0.4-opr.1
      - name: rhbk-operator.v24.0.3-opr.1
        version: 24.0.3-opr.1
      name: stable-v24
    - currentCSV: rhbk-operator.v24.0.11-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
          createdAt: "2025-03-26T02:27:51Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:d339ecf788d1681ba9945a2ea63ca675caec863ec75f497ef916103f4c898003
        version: 24.0.11-opr.1
      entries:
      - name: rhbk-operator.v24.0.11-opr.1
        version: 24.0.11-opr.1
      - name: rhbk-operator.v24.0.10-opr.1
        version: 24.0.10-opr.1
      - name: rhbk-operator.v24.0.9-opr.1
        version: 24.0.9-opr.1
      - name: rhbk-operator.v24.0.8-opr.1
        version: 24.0.8-opr.1
      - name: rhbk-operator.v24.0.7-opr.1
        version: 24.0.7-opr.1
      - name: rhbk-operator.v24.0.6-opr.1
        version: 24.0.6-opr.1
      - name: rhbk-operator.v24.0.5-opr.1
        version: 24.0.5-opr.1
      - name: rhbk-operator.v24.0.4-opr.1
        version: 24.0.4-opr.1
      - name: rhbk-operator.v24.0.3-opr.1
        version: 24.0.3-opr.1
      name: stable-v24.0
    - currentCSV: rhbk-operator.v26.2.9-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
          createdAt: "2025-09-18T15:03:55Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:50a62b46af44ce0ff7e6c3a15e1bfb2a9611ba4a8b65b8d0ab65817064692c55
        version: 26.2.9-opr.1
      entries:
      - name: rhbk-operator.v26.2.9-opr.1
        version: 26.2.9-opr.1
      - name: rhbk-operator.v26.2.8-opr.1
        version: 26.2.8-opr.1
      - name: rhbk-operator.v26.2.7-opr.1
        version: 26.2.7-opr.1
      - name: rhbk-operator.v26.2.6-opr.1
        version: 26.2.6-opr.1
      - name: rhbk-operator.v26.2.5-opr.1
        version: 26.2.5-opr.1
      - name: rhbk-operator.v26.2.4-opr.2
        version: 26.2.4-opr.2
      - name: rhbk-operator.v26.2.4-opr.1
        version: 26.2.4-opr.1
      - name: rhbk-operator.v26.0.15-opr.1
        version: 26.0.15-opr.1
      - name: rhbk-operator.v26.0.14-opr.1
        version: 26.0.14-opr.1
      - name: rhbk-operator.v26.0.13-opr.1
        version: 26.0.13-opr.1
      - name: rhbk-operator.v26.0.12-opr.1
        version: 26.0.12-opr.1
      - name: rhbk-operator.v26.0.11-opr.1
        version: 26.0.11-opr.1
      - name: rhbk-operator.v26.0.10-opr.1
        version: 26.0.10-opr.1
      - name: rhbk-operator.v26.0.9-opr.1
        version: 26.0.9-opr.1
      - name: rhbk-operator.v26.0.8-opr.1
        version: 26.0.8-opr.1
      - name: rhbk-operator.v26.0.7-opr.1
        version: 26.0.7-opr.1
      - name: rhbk-operator.v26.0.6-opr.1
        version: 26.0.6-opr.1
      - name: rhbk-operator.v26.0.5-opr.1
        version: 26.0.5-opr.1
      name: stable-v26
    - currentCSV: rhbk-operator.v26.0.15-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:fc5fc8cfb2c100c5b3a33fc11a78dfca569f5f34f45d88439340392d2c60afb2
          createdAt: "2025-08-28T19:34:30Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:fc5fc8cfb2c100c5b3a33fc11a78dfca569f5f34f45d88439340392d2c60afb2
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:4bb4d3070fe118bdd7de24e8b68133b8b6e158cea48c9f1abad2b4297be0dfdb
        version: 26.0.15-opr.1
      entries:
      - name: rhbk-operator.v26.0.15-opr.1
        version: 26.0.15-opr.1
      - name: rhbk-operator.v26.0.14-opr.1
        version: 26.0.14-opr.1
      - name: rhbk-operator.v26.0.13-opr.1
        version: 26.0.13-opr.1
      - name: rhbk-operator.v26.0.12-opr.1
        version: 26.0.12-opr.1
      - name: rhbk-operator.v26.0.11-opr.1
        version: 26.0.11-opr.1
      - name: rhbk-operator.v26.0.10-opr.1
        version: 26.0.10-opr.1
      - name: rhbk-operator.v26.0.9-opr.1
        version: 26.0.9-opr.1
      - name: rhbk-operator.v26.0.8-opr.1
        version: 26.0.8-opr.1
      - name: rhbk-operator.v26.0.7-opr.1
        version: 26.0.7-opr.1
      - name: rhbk-operator.v26.0.6-opr.1
        version: 26.0.6-opr.1
      - name: rhbk-operator.v26.0.5-opr.1
        version: 26.0.5-opr.1
      name: stable-v26.0
    - currentCSV: rhbk-operator.v26.2.9-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
          createdAt: "2025-09-18T15:03:55Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:50a62b46af44ce0ff7e6c3a15e1bfb2a9611ba4a8b65b8d0ab65817064692c55
        version: 26.2.9-opr.1
      entries:
      - name: rhbk-operator.v26.2.9-opr.1
        version: 26.2.9-opr.1
      - name: rhbk-operator.v26.2.8-opr.1
        version: 26.2.8-opr.1
      - name: rhbk-operator.v26.2.7-opr.1
        version: 26.2.7-opr.1
      - name: rhbk-operator.v26.2.6-opr.1
        version: 26.2.6-opr.1
      - name: rhbk-operator.v26.2.5-opr.1
        version: 26.2.5-opr.1
      - name: rhbk-operator.v26.2.4-opr.2
        version: 26.2.4-opr.2
      - name: rhbk-operator.v26.2.4-opr.1
        version: 26.2.4-opr.1
      name: stable-v26.2
    defaultChannel: stable-v26.2
    packageName: rhbk-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-opensearchserverless-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-opensearchserverless-controller.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/opensearchserverless-controller:0.2.1
          createdAt: "2025-09-20T21:43:17Z"
          description: AWS OpenSearchServerless controller is a service controller
            for managing OpenSearchServerless resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Collection represents the state of an AWS opensearchserverless
              Collection resource.
            displayName: Collection
            kind: Collection
            name: collections.opensearchserverless.services.k8s.aws
            version: v1alpha1
          - description: SecurityPolicy represents the state of an AWS opensearchserverless
              SecurityPolicy resource.
            displayName: SecurityPolicy
            kind: SecurityPolicy
            name: securitypolicies.opensearchserverless.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon OpenSearchServerless resources in AWS from within your Kubernetes cluster.

          **About Amazon OpenSearchServerless**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon OpenSearchServerless
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opensearchserverless
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon OpenSearchServerless Developer Resources
          url: https://aws.amazon.com/OpenSearchServerless/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: opensearchserverless maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/opensearchserverless-controller:0.2.1
        version: 0.2.1
      entries:
      - name: ack-opensearchserverless-controller.v0.2.1
        version: 0.2.1
      - name: ack-opensearchserverless-controller.v0.2.0
        version: 0.2.0
      - name: ack-opensearchserverless-controller.v0.1.2
        version: 0.1.2
      - name: ack-opensearchserverless-controller.v0.1.1
        version: 0.1.1
      - name: ack-opensearchserverless-controller.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ack-opensearchserverless-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Automotive SIG CentOS
      provider-url: ""
    name: automotive-infra
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: automotive-infra.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "ContainerBuild",
                "metadata": {
                  "name": "cb-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "ImageBuild",
                "metadata": {
                  "name": "imagebuild-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "OperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "automotive-infra"
                  },
                  "name": "operatorconfig-sample",
                  "namespace": "automotive-infra-operator"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/fedora-sig-robotics/infra-ci-operator:v0.0.10
          createdAt: "2025-03-31T14:06:33Z"
          description: Streamline CI/CD for RHIVOS and its ecosystem using OpenShift
            Pipelines
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: automotive-infra-operator
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator
          support: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Client
            name: clients.jumpstarter.dev
            version: v1alpha1
          - kind: ContainerBuild
            name: containerbuilds.automotive.sig.centos.org
            version: v1
          - kind: ExporterAccessPolicy
            name: exporteraccesspolicies.jumpstarter.dev
            version: v1alpha1
          - kind: Exporter
            name: exporters.jumpstarter.dev
            version: v1alpha1
          - kind: ImageBuild
            name: imagebuilds.automotive.sig.centos.org
            version: v1
          - kind: Lease
            name: leases.jumpstarter.dev
            version: v1alpha1
          - kind: OperatorConfig
            name: operatorconfigs.automotive.sig.centos.org
            version: v1
        description: This operator is designed to streamline CI/CD for RHIVOS and
          its ecosystem using OpenShift Pipelines
        displayName: Automotive-CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Automotive
        - RHIVOS
        - Pipelines
        - CI
        links:
        - name: Automotive infra Operator
          url: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator
        maintainers:
        - email: automotive-sig@lists.centos.org
          name: Automotive SIG CentOS Community
        maturity: alpha
        provider:
          name: Automotive SIG CentOS
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/fedora-sig-robotics/infra-ci-operator:v0.0.10
        version: 0.0.10
      entries:
      - name: automotive-infra.v0.0.10
        version: 0.0.10
      - name: automotive-infra.v0.0.9
        version: 0.0.9
      - name: automotive-infra.v0.0.8
        version: 0.0.8
      - name: automotive-infra.v0.0.7
        version: 0.0.7
      - name: automotive-infra.v0.0.6
        version: 0.0.6
      - name: automotive-infra.v0.0.5
        version: 0.0.5
      - name: automotive-infra.v0.0.4
        version: 0.0.4
      - name: automotive-infra.v0.0.3
        version: 0.0.3
      - name: automotive-infra.v0.0.2
        version: 0.0.2
      - name: automotive-infra.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: automotive-infra
    provider:
      name: Automotive SIG CentOS
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: ibm-licensing-operator
      app.kubernetes.io/managed-by: ibm-licensing-operator
      app.kubernetes.io/name: ibm-licensing
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      intent: projected
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-licensing-operator-app
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-licensing-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "IBMLicensing",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-licensing-operator",
                    "app.kubernetes.io/managed-by": "ibm-licensing-operator",
                    "app.kubernetes.io/name": "ibm-licensing"
                  },
                  "name": "instance"
                },
                "spec": {
                  "version": "4.2.14",
                  "apiSecretToken": "ibm-licensing-token",
                  "datasource": "datacollector",
                  "httpsEnable": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
          createdAt: "2025-02-25T12:16:11Z"
          description: The IBM Licensing Operator provides a Kubernetes CRD-Based
            API to monitor the license usage of products.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-licensing-operator-app
          olm.skipRange: '>=1.0.0 <4.2.14'
          operatorframework.io/suggested-namespace: ibm-licensing
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/internal-objects: '["ibmlicensingmetadatas.operator.ibm.com",
            "ibmlicensingdefinitions.operator.ibm.com", "ibmlicensingquerysources.operator.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'IBMLicensing custom resource is used to create an instance
              of the License Service, used to collect information about license usage
              of IBM containerized products and IBM Cloud Paks per cluster. You can
              retrieve license usage data through a dedicated API call and generate
              an audit snapshot on demand. Documentation: For additional details regarding
              install parameters check: https://ibm.biz/icpfs39install. License: Please
              refer to the IBM Terms website (ibm.biz/lsvc-lic) to find the license
              terms for the particular IBM product for which you are deploying this
              component.'
            displayName: IBM License Service
            kind: IBMLicensing
            name: ibmlicensings.operator.ibm.com
            version: v1alpha1
          - kind: IBMLicensingDefinition
            name: ibmlicensingdefinitions.operator.ibm.com
            version: v1
          - description: 'IBMLicensingMetadata is the schema for IBM License Service.
              Thanks to IBMLicensingMetadata, you can track the license usage of Virtual
              Processor Core (VPC) metric by Pods that are managed by automation,
              for which licensing annotations are not defined at the time of deployment,
              such as the open source OpenLiberty. / For more information, see documentation:
              https://ibm.biz/icpfs39install. By installing the operator for IBM License
              Service, you accept the license terms: https://ibm.biz/icpfs39license.'
            displayName: IBMLicensing Metadata
            kind: IBMLicensingMetadata
            name: ibmlicensingmetadatas.operator.ibm.com
            version: v1alpha1
          - kind: IBMLicensingQuerySource
            name: ibmlicensingquerysources.operator.ibm.com
            version: v1
        description: |-
          **Important:**
          - If you are using the IBM Licensing Operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For the link to your IBM Cloud Pak documentation, see [IBM Cloud Paks that use Common Services](https://ibm.biz/BdyGwb).
          - If you are using a stand-alone IBM Container Software, you can use the IBM Licensing Operator directly. For more information, see [ibm-licensing-operator for stand-alone IBM Containerized Software](https://ibm.biz/BdyGwh).

          **IBM Licensing Operator overview**

          IBM Licensing Operator installs License Service. You can use License Service to collect information about license usage of IBM Containerized products and IBM Cloud Paks per cluster. You can retrieve license usage data through a dedicated API call and generate an audit snapshot on demand.

          **Supported platforms**

           Red Hat OpenShift Container Platform 4.10 or newer installed on Linux x86_64, Linux on Power (ppc64le), Linux on IBM Z and LinuxONE.

          **Prerequisites**

           Prerequisites depend on the integration of the License Service with an IBM Cloud Pak or IBM Containerized Software. For more information, see the applicable IBM Cloud Pak documentation or [ibm-licensing-operator for stand-alone IBM Containerized Software](https://ibm.biz/BdyGwh).
        displayName: IBM Licensing
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBMLicensing
        - IBM
        - Cloud
        - License
        - Service
        - License Service
        links:
        - name: GitHub
          url: https://github.com/IBM/ibm-licensing-operator
        maintainers:
        - email: talk2sam@us.ibm.com
          name: talk2sam
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9186c350fa54e8b792d72e47be6b54a67b610f726175c9641795010d45ad51f9
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:ec137115210dae9b69cb2899ed8a5437d94c8a658ae1ae5eb295232b34b0925e
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:96e6db93b2d16ba590252d6568b68f254548aaed73ec17e4bdd8d530ea668be6
        version: 4.2.14
      entries:
      - name: ibm-licensing-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-licensing-operator-app
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator-rhmp.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/dynatrace-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/dynatrace-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator-rhmp.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator-rhmp.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator-rhmp.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator-rhmp.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator-rhmp.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator-rhmp.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator-rhmp.v1.5.0
        version: 1.5.0
      - name: dynatrace-operator-rhmp.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator-rhmp.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator-rhmp.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator-rhmp
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Argo CD Community
      provider-url: ""
    name: argocd-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: argocd-operator.v0.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd-sample"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCDExport",
                "metadata": {
                  "name": "argocdexport-sample"
                },
                "spec": {
                  "argocd": "argocd-sample"
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd-sample"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/argoprojlabs/argocd-operator@sha256:223ad82fb697471ede79201c4a042115c9ca5c00a172d5bdcd5a780e58c46c79
          createdAt: "2025-09-11T06:29:33Z"
          description: Argo CD is a declarative, GitOps continuous delivery tool for
            Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/argoproj-labs/argocd-operator
          support: Argo CD
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: An ApplicationSet is a group or set of Application resources.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: ArgoCDExport is the Schema for the argocdexports API
            displayName: Argo CDExport
            kind: ArgoCDExport
            name: argocdexports.argoproj.io
            version: v1alpha1
          - description: ArgoCD is the Schema for the argocds API
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: ArgoCD is the Schema for the argocds API
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: NotificationsConfiguration is the Schema for the notificationsconfiguration
              API
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
        description: |
          ## Overview

          The Argo CD Operator manages the full lifecycle for [Argo CD](https://argoproj.github.io/argo-cd/) and it's
          components. The operator's goal is to automate the tasks required when operating an Argo CD cluster.

          Beyond installation, the operator helps to automate the process of upgrading, backing up and restoring as needed and
          remove the human as much as possible. In addition, the operator aims to provide deep insights into the Argo CD
          environment by configuring Prometheus and Grafana to aggregate, visualize and expose the metrics already exported by
          Argo CD.

          The operator aims to provide the following, and is a work in progress.

          * Easy configuration and installation of the Argo CD components with sane defaults to get up and running quickly.
          * Provide seamless upgrades to the Argo CD components.
          * Ability to back up and restore an Argo CD cluster from a point in time or on a recurring schedule.
          * Aggregate and expose the metrics for Argo CD and the operator itself using Prometheus and Grafana.
          * Autoscale the Argo CD components as necessary to handle variability in demand.

          ## Usage

          Deploy a basic Argo CD cluster by creating a new ArgoCD resource in the namespace where the operator is installed.

          ```
          apiVersion: argoproj.io/v1alpha1
          kind: ArgoCD
          metadata:
            name: example-argocd
          spec: {}
          ```

          ## Backup

          Backup the cluster above by creating a new ArgoCDExport resource in the namespace where the operator is installed.

          ```
          apiVersion: argoproj.io/v1alpha1
          kind: ArgoCDExport
          metadata:
            name: example-argocdexport
          spec:
            argocd: example-argocd
          ```

          See the [documentation](https://argocd-operator.readthedocs.io) and examples on
          [GitHub](https://github.com/argoproj-labs/argocd-operator) for more information.
        displayName: Argo CD
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - kubernetes
        links:
        - name: Argo CD Project
          url: https://argoproj.github.io/argo-cd/
        - name: Operator Documentation
          url: https://argocd-operator.readthedocs.io
        - name: Operator Source Code
          url: https://github.com/argoproj-labs/argocd-operator
        maintainers:
        - email: aveerama@redhat.com
          name: Abhishek Veeramalla
        maturity: alpha
        provider:
          name: Argo CD Community
        relatedImages:
        - quay.io/argoprojlabs/argocd-operator@sha256:223ad82fb697471ede79201c4a042115c9ca5c00a172d5bdcd5a780e58c46c79
        version: 0.15.0
      entries:
      - name: argocd-operator.v0.15.0
        version: 0.15.0
      - name: argocd-operator.v0.14.1
        version: 0.14.1
      - name: argocd-operator.v0.14.0
        version: 0.14.0
      - name: argocd-operator.v0.13.0
        version: 0.13.0
      - name: argocd-operator.v0.12.0
        version: 0.12.0
      - name: argocd-operator.v0.11.0
        version: 0.11.0
      - name: argocd-operator.v0.10.1
        version: 0.10.1
      - name: argocd-operator.v0.10.0
        version: 0.10.0
      - name: argocd-operator.v0.9.1
        version: 0.9.1
      - name: argocd-operator.v0.9.0
        version: 0.9.0
      - name: argocd-operator.v0.8.0
        version: 0.8.0
      - name: argocd-operator.v0.7.0
        version: 0.7.0
      - name: argocd-operator.v0.5.0
        version: 0.5.0
      - name: argocd-operator.v0.4.0
        version: 0.4.0
      - name: argocd-operator.v0.3.0
        version: 0.3.0
      - name: argocd-operator.v0.2.1
        version: 0.2.1
      - name: argocd-operator.v0.2.0
        version: 0.2.0
      - name: argocd-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: argocd-operator
    provider:
      name: Argo CD Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-framework
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-framework.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Framework",
                "metadata": {
                  "name": "framework-sample"
                },
                "spec": {
                  "kcos-eula": {
                    "image": {
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    },
                    "service": {
                      "nodePort": 0,
                      "type": "ClusterIP"
                    }
                  },
                  "kcos-logging-framework": {
                    "image": {
                      "remoteRepository": "docker.io",
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    },
                    "ingress": {
                      "annotations": {
                        "nginx.ingress.kubernetes.io/auth-method": "POST",
                        "nginx.ingress.kubernetes.io/auth-response-headers": "X-AUTH-USER",
                        "nginx.ingress.kubernetes.io/auth-url": "https://127.0.0.1/token/validate",
                        "nginx.ingress.kubernetes.io/cors-allow-credentials": "true",
                        "nginx.ingress.kubernetes.io/cors-allow-methods": "PUT, GET, POST, OPTIONS, PATCH",
                        "nginx.ingress.kubernetes.io/cors-allow-origin": "https://admin.example.com,https://localhost:4200",
                        "nginx.ingress.kubernetes.io/enable-cors": "true",
                        "nginx.ingress.kubernetes.io/proxy-body-size": "0",
                        "nginx.ingress.kubernetes.io/ssl-redirect": "true"
                      },
                      "enabled": true
                    }
                  },
                  "kcos-postgresql": {
                    "jobCopySecret": {
                      "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                    },
                    "postgresql": {
                      "image": {
                        "registry": "prod.harbor.keysight.digital/loadcore"
                      },
                      "initdbScripts": {
                        "init_script.sh": "#!/bin/sh\nexport PGUSER=postgres\npsql -c \"SELECT 1 FROM pg_database WHERE datname = 'db'\" | grep -q 1 || psql -c \"CREATE DATABASE db\"\npsql -c \"SELECT 1 FROM pg_database WHERE datname = 'kcoslicensingdb'\" | grep -q 1 || psql -c \"CREATE DATABASE kcoslicensingdb\"\n"
                      },
                      "metrics": {
                        "image": {
                          "registry": "docker.io"
                        }
                      },
                      "persistence": {
                        "storageClass": "managed-loadcore-storage"
                      },
                      "pgHbaConfiguration": "local all all trust\nhost all all 0.0.0.0/0 trust",
                      "securityContext": {
                        "enabled": false
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": false
                        }
                      },
                      "volumePermissions": {
                        "image": {
                          "registry": "docker.io"
                        },
                        "securityContext": {
                          "runAsUser": "auto"
                        }
                      }
                    }
                  },
                  "kcos-system-diagnostics": {
                    "environment": {
                      "ELASTICSEARCH_ENDPOINT": "elasticsearch.keysight-wap.svc.cluster.local.",
                      "IGNORE_DATA_PREFIX": "wireless_stats"
                    },
                    "image": {
                      "initContainer": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf",
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    }
                  },
                  "subcharts": {
                    "enable": {
                      "kcos-coredns-fixup": false,
                      "kcos-eula": true,
                      "kcos-local-storage": false,
                      "kcos-logging-framework": true,
                      "kcos-logging-rotate": false,
                      "kcos-naas": false,
                      "kcos-postgresql": true,
                      "kcos-shell": false,
                      "kcos-system-diagnostics": true,
                      "kcos-vital-service": false,
                      "nfs-server-provisioner": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:27:28Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Framework
            name: frameworks.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Framework Operator
        displayName: Keysight KCOS Framework Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Framework
          url: https://keysight-kcos-framework.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-framework@sha256:531f278def85e33d453479fce74127e12e3ed62a251872eab421c6561bedd0c5
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        version: 0.1.21
      entries:
      - name: keysight-kcos-framework.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-framework
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx-openshift-operator.v0.0.36
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
          createdAt: "2019-07-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX on JBoss EAP 7
            displayName: Joget DX on JBoss EAP 7
            kind: JogetDX
            name: jogetdx.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX7/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
        - registry.connect.redhat.com/joget/joget-dx7-eap7@sha256:7245501f5fccf011d75c034b7d55542cfe3bd2c64cc7e48b3224a48b623af282
        version: 0.0.36
      entries:
      - name: joget-dx-openshift-operator.v0.0.36
        version: 0.0.36
      - name: joget-dx-openshift-operator.v0.0.35
        version: 0.0.35
      - name: joget-dx-openshift-operator.v0.0.34
        version: 0.0.34
      - name: joget-dx-openshift-operator.v0.0.33
        version: 0.0.33
      - name: joget-dx-openshift-operator.v0.0.32
        version: 0.0.32
      - name: joget-dx-openshift-operator.v0.0.31
        version: 0.0.31
      - name: joget-dx-openshift-operator.v0.0.30
        version: 0.0.30
      - name: joget-dx-openshift-operator.v0.0.29
        version: 0.0.29
      - name: joget-dx-openshift-operator.v0.0.28
        version: 0.0.28
      - name: joget-dx-openshift-operator.v0.0.27
        version: 0.0.27
      - name: joget-dx-openshift-operator.v0.0.26
        version: 0.0.26
      - name: joget-dx-openshift-operator.v0.0.25
        version: 0.0.25
      - name: joget-dx-openshift-operator.v0.0.24
        version: 0.0.24
      - name: joget-dx-openshift-operator.v0.0.23
        version: 0.0.23
      - name: joget-dx-openshift-operator.v0.0.22
        version: 0.0.22
      - name: joget-dx-openshift-operator.v0.0.21
        version: 0.0.21
      - name: joget-dx-openshift-operator.v0.0.20
        version: 0.0.20
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Habana Labs Ltd.
      provider-url: https://habana.ai
    name: habana-ai-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: habana-ai-operator.v1.19.2-32-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.19.2-32"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.19.2"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.19.2-32"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.19.2-32"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.19.2-32"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.19.2-32"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.19.2-32"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.19.2-32"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.19.2-32"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.19.2-32"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.19.2-32"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:975ed2135061d3bdc145c83abaa15f99e4c9f6efb15c708de41aa00fbcb72333
          createdAt: "2025-02-10T07:33:30Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.19.2-32-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi AI accelerator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:ac16752b53df69573da257d3a3a717a214b2eb1656779a7ef94084d74363fb46
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:b12802e7ef05273fceaef3a9987c3c4ae760714e6f333363f665eb26be9175e9
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:f05cd78d7593b877ec8a75175a9faa72d9b991d81c74af9ecd92208fde296eca
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:170a15041fc3e3901b4e8e0b9a32d53fde11026776c17a6560d500dacb976c56
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:d728a91a36503f07dfacd49fdbfefe8083607ca33babb5f2dfc1e9de05386454
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:e667695b845cd2be887f117dac66861d2c5e74def009d2939559ae7c30574dc0
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:975ed2135061d3bdc145c83abaa15f99e4c9f6efb15c708de41aa00fbcb72333
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:eeb4573de20d3341a085c75e669b7d37c424e348421d7916440fbbfff12ea482
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:71d4c7181930c71c0ca4e6b696ea416a687eec6d798a43e854fcd1bfdc224a0c
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:72cc75353615369d12a07f5488bcdc041a8435906ca6e16c53b001016108e099
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:373b9caa27a96bb7979f7a6aec246795142ce3a28c6b97a8902035ec20814fa2
        version: 1.19.2-32-1
      entries:
      - name: habana-ai-operator.v1.19.2-32-1
        version: 1.19.2-32-1
      name: "1.19"
    - currentCSV: habana-ai-operator.v1.20.0-543-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.20.0-543"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.20.0-4"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.20.0-543-2"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.20.0-543"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.20.0-543"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.20.0-543-4"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.20.0-543"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.20.0-543-4"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:d5b37eafe48446e1ffa492e195f4519b77d4a11f6750e5c71f27c1536bdfd2cc
          createdAt: "2025-02-26T09:03:44Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.20.0-543-5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:9ab54a84a00e8afc078eec2a7d643f09b1bd674e0e230733fbce7108e392e009
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:d5b37eafe48446e1ffa492e195f4519b77d4a11f6750e5c71f27c1536bdfd2cc
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:317765eac4e0332a0c4f6bead5563b343a8c22e6e0eaa62276ff66be08608c30
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:c5beac6209d0671d3a49b63f1cf95401055b042430eafaefc3859b5c633f0e84
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:68d0f44170cbf3ea15b793e0503fb07ce7e32488011fc85ccb6c9234b3b9302f
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:3ea80aa8994e8b19fbbdcd18ba2c35800774da3ca4e27a2514d0f5f9676786fa
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:957702b58ee66d7339a20b2a5ece783606d6ccebde627c58951d5ecd6bb1b13f
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:cba03259b33c4cb9bcd22622ab0dda962846ea0a80c727779365cf55129ad308
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:d14c457106b6133fc41ef727ac40a6bf2f0e5c97f21659969d1daebdc1310b43
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:c662de244c5900c89b918b16982298676d9a6d553360a5716174b28d4d64a635
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:a673900f08c38c9bb40c90b3cddea580afcc6a99195d816cd0b5cae9831ca5c9
        version: 1.20.0-543-5
      entries:
      - name: habana-ai-operator.v1.20.0-543-5
        version: 1.20.0-543-5
      name: "1.20"
    - currentCSV: habana-ai-operator.v1.21.5-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.21.5-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.21.5-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.21.5-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:0741da99140d137474b9e7edba921672eab8ff6277e1dd8bf3a4179e87f01dd4
          createdAt: "2025-09-17T11:55:32Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.21.5-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:0741da99140d137474b9e7edba921672eab8ff6277e1dd8bf3a4179e87f01dd4
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:a216ec6a6e7df50e067c557e9ed5871a3308b2ba42a4e41c9642f0c2c4503d88
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:9692d079ef18a364ed235c187482f329b3a5764a773e363acb1238044ad4a884
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:c95d3ffd6c91be3418415740258d33a8dbcedad6fe7d6bfe68d4dd122027447b
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:bc3749ff153fff71e252820cbf22fcd0753b195b2204f9731e42af29fb7ddaa0
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:bd57e7059467f0ca1667d19a70bd9a6ad96b83bbcb53c09fdad5a262fee76af9
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:c84e997c7d8c22c4d4e902a29e45e5364f0ad306f17d03a7d8ddb4391142e167
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:256cd8bebec3ed835e187e925a14a6b605ce9b3a17fe7d8886e0b32cde163d90
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:1e488c8ad31784dab0fe70144f5146d5809e2fc7c86424d1b512ab887787e28f
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:c7147359dcd6a2e5adee95259d37743ff302b1deaeb13b184044f4029d33f44a
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:d7f64d15c379524597aa2b2c5b4586c890442fa32356692013cf665b01b0f092
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:a0b45ccfb6af54b1043ccc2336ad3d7990cc447af5ba3621f3eb1953cd40ffc5
        version: 1.21.5-6
      entries:
      - name: habana-ai-operator.v1.21.5-6
        version: 1.21.5-6
      name: "1.21"
    - currentCSV: habana-ai-operator.v1.22.1-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.22.1-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
          createdAt: "2025-09-17T10:15:21Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.22.1-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:1e15f3de21c2b1e20f54d1f8a973290c6d940624506880c5bc3eb880011bcc02
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:d1974d72774c626e77707e8f44bcc631ac142bfe2a5395714969a964c963e36e
        - vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer@sha256:2cf151c2a518e3e1c7c09efb8301e469d266d1f030fff2e6180b086cc2e04ee7
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:704b4bebdfe6b85693df4b213f67862859a554c21ce4e614736172f64b541ea7
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:c9ff6f3fe24ef87c7bfc71bdb1cd7900941df52f9ad1293badd7e14783ce7ef6
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:e34a6820a58e7e8090b521781c76acc3867c94f8291634de7b138f928e0bdb74
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:53c1844b5e8eb08d0cc29865a732100208c5dfc0fe55df8b32e589e319bcab6c
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:38c21eae6249e5571921998660418e59870bc0880e7b8c9b95a4a8db30ef2e7d
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:1925586fda34c6fa444a59b1fa7be20b3cf16e845bf56566ad2a41c276b0cfe9
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:392fcd2a4730299e6e3bab9f7e96fc3b710de8adb374a2c64b7316e097ed05f9
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:7140877b23291a5db3e5e28def3c9d5e7cb7be62f046e2fb7ac45eff369cea84
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:8f413edfff1978b35ba3998ed5f1c6293e6764b1d2cd9faa08f0bd81dc6e3e10
        version: 1.22.1-6
      entries:
      - name: habana-ai-operator.v1.22.1-6
        version: 1.22.1-6
      name: "1.22"
    - currentCSV: habana-ai-operator.v1.22.1-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.22.1-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
          createdAt: "2025-09-17T10:15:21Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.22.1-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:c9ff6f3fe24ef87c7bfc71bdb1cd7900941df52f9ad1293badd7e14783ce7ef6
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:392fcd2a4730299e6e3bab9f7e96fc3b710de8adb374a2c64b7316e097ed05f9
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:38c21eae6249e5571921998660418e59870bc0880e7b8c9b95a4a8db30ef2e7d
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:1925586fda34c6fa444a59b1fa7be20b3cf16e845bf56566ad2a41c276b0cfe9
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:704b4bebdfe6b85693df4b213f67862859a554c21ce4e614736172f64b541ea7
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:7140877b23291a5db3e5e28def3c9d5e7cb7be62f046e2fb7ac45eff369cea84
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:8f413edfff1978b35ba3998ed5f1c6293e6764b1d2cd9faa08f0bd81dc6e3e10
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:1e15f3de21c2b1e20f54d1f8a973290c6d940624506880c5bc3eb880011bcc02
        - vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer@sha256:2cf151c2a518e3e1c7c09efb8301e469d266d1f030fff2e6180b086cc2e04ee7
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:e34a6820a58e7e8090b521781c76acc3867c94f8291634de7b138f928e0bdb74
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:53c1844b5e8eb08d0cc29865a732100208c5dfc0fe55df8b32e589e319bcab6c
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:d1974d72774c626e77707e8f44bcc631ac142bfe2a5395714969a964c963e36e
        version: 1.22.1-6
      entries:
      - name: habana-ai-operator.v1.22.1-6
        version: 1.22.1-6
      name: stable
    defaultChannel: stable
    packageName: habana-ai-operator
    provider:
      name: Habana Labs Ltd.
      url: https://habana.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-multicluster-orchestrator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-multicluster-orchestrator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "items": [
                    {
                      "clusterName": "local-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    },
                    {
                      "clusterName": "spoke-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    }
                  ],
                  "manageS3": false,
                  "schedulingIntervals": [
                    "5m"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["odf-multicluster-console"]'
          createdAt: "2025-09-12T07:33:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          Orchestrator for OpenShift Data Foundation clusters running across multiple OpenShift clusters.
          It uses Red Hat Advanced Cluster Management for Kubernetes as the multicluster control plane.
        displayName: ODF Multicluster Orchestrator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: ODF Multicluster Orchestrator
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: uchapaga@redhat.com
          name: Umanga Chapagain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-multicluster-rhel9-operator@sha256:09e2c7a20ab92ddc5d589072bb0ab9ce2ed8706ed8364f1b11192d7a7396e138
        - registry.redhat.io/odf4/odf-multicluster-console-rhel9@sha256:7527f4a14aed0386b8f160fa8a92b7ff1f58114970e07c3ec6c433d2a53459b5
        version: 4.15.20-rhodf
      entries:
      - name: odf-multicluster-orchestrator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odf-multicluster-orchestrator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odf-multicluster-orchestrator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odf-multicluster-orchestrator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odf-multicluster-orchestrator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odf-multicluster-orchestrator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odf-multicluster-orchestrator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odf-multicluster-orchestrator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odf-multicluster-orchestrator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odf-multicluster-orchestrator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odf-multicluster-orchestrator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odf-multicluster-orchestrator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odf-multicluster-orchestrator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odf-multicluster-orchestrator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odf-multicluster-orchestrator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odf-multicluster-orchestrator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odf-multicluster-orchestrator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odf-multicluster-orchestrator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odf-multicluster-orchestrator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odf-multicluster-orchestrator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odf-multicluster-orchestrator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "items": [
                    {
                      "clusterName": "local-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    },
                    {
                      "clusterName": "spoke-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    }
                  ],
                  "manageS3": false,
                  "schedulingIntervals": [
                    "5m"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["odf-multicluster-console"]'
          createdAt: "2025-09-12T06:32:35Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          Orchestrator for OpenShift Data Foundation clusters running across multiple OpenShift clusters.
          It uses Red Hat Advanced Cluster Management for Kubernetes as the multicluster control plane.
        displayName: ODF Multicluster Orchestrator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: ODF Multicluster Orchestrator
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: uchapaga@redhat.com
          name: Umanga Chapagain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-multicluster-rhel9-operator@sha256:78d6bc14b95d1ee91ed529422062ec191afc489d6ab4e75906ae3c7fe58749cb
        - registry.redhat.io/odf4/odf-multicluster-console-rhel9@sha256:97d3645560388f6ae08fc12cafb58b6c766e38d83de04b645b3e6fb03fbba231
        version: 4.16.16-rhodf
      entries:
      - name: odf-multicluster-orchestrator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odf-multicluster-orchestrator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odf-multicluster-orchestrator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odf-multicluster-orchestrator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odf-multicluster-orchestrator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odf-multicluster-orchestrator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odf-multicluster-orchestrator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odf-multicluster-orchestrator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odf-multicluster-orchestrator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odf-multicluster-orchestrator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odf-multicluster-orchestrator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odf-multicluster-orchestrator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odf-multicluster-orchestrator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odf-multicluster-orchestrator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odf-multicluster-orchestrator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odf-multicluster-orchestrator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odf-multicluster-orchestrator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odf-multicluster-orchestrator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-access-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-access-operator.v23.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyAccess",
                "metadata": {
                  "name": "isva-sample"
                },
                "spec": {
                  "image": "icr.io/isva/verify-access-wrp:10.0.5.0",
                  "instance": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isva/verify-access-operator:23.3.0
          createdAt: Mon Mar  20 00:59:34 UTC 2023
          description: The IBM Security Verify Access Operator manages the lifecycle
            of IBM Security Verify Access worker containers.
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM-Security/verify-access-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyAccess is the Schema for the ibmsecurityverifyaccesses
              API.
            displayName: IBMSecurity Verify Access
            kind: IBMSecurityVerifyAccess
            name: ibmsecurityverifyaccesses.ibm.com
            version: v1
        description: |+
          In a world of highly fragmented access management environments, [IBM Security Verify Access](https://www.ibm.com/au-en/products/verify-access) helps you simplify your users' access while more securely adopting web, mobile and cloud technologies. This solution helps you strike a balance between usability and security through the use of risk-based access, single sign-on, integrated access management control, identity federation and its mobile multi-factor authentication capability, IBM Verify. Take back control of your access management with IBM Security Verify Access.

          The IBM Security Verify Access operator provides lifecycle management of the lightweight containers which are used to protect an environment, namely:
          * [Web Reverse Proxy](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-web-reverse-proxy)
          * [Runtime](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-runtime)
          * [Distributed Session Cache](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-distributed-session-cache)

          The operator will manage the deployment of these lightweight IBM Security Verify Access worker containers, and also control the rolling restart of these containers when a configuration snapshot is updated.

          See the project [Readme](https://github.com/IBM-Security/verify-access-operator/blob/master/README.md) for further information and details.

        displayName: IBM Security Verify Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Access Product Information
          url: https://www.ibm.com/au-en/products/verify-access
        - name: Verify Access Documentation
          url: https://www.ibm.com/docs/en/sva
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Access Development Team
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - icr.io/isva/verify-access-operator:23.3.0
        version: 23.3.0
      entries:
      - name: ibm-security-verify-access-operator.v23.3.0
        version: 23.3.0
      - name: ibm-security-verify-access-operator.v22.10.0
        version: 22.10.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-access-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: oadp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oadp-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "Velero",
                "metadata": {
                  "name": "example-velero",
                  "namespace":"openshift-adp"
                },
                "spec": {
                  "backupStorageLocations": [
                    {
                      "config": {
                        "profile": "default",
                        "region": "us-east-1"
                      },
                      "name": "default",
                      "objectStorage": {
                        "bucket": "myBucket",
                        "prefix": "velero"
                      },
                      "provider": "aws"
                    }
                  ],
                  "defaultVeleroPlugins": [
                    "aws",
                    "csi",
                    "openshift"
                  ],
                  "veleroFeatureFlags": [
                    "EnableCSI"
                  ],
                  "enableRestic": true,
                  "volumeSnapshotLocations": [
                    {
                      "config": {
                        "profile": "default",
                        "region": "us-west-1"
                      },
                      "name": "default",
                      "provider": "aws"
                    }
                  ]
                }
               },
                {
                  "apiVersion": "velero.io/v1",
                  "kind": "Backup",
                  "metadata": {
                    "name": "backup",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "BackupStorageLocation",
                  "metadata": {
                    "name": "backupstoragelocation",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "DeleteBackupRequest",
                  "metadata": {
                    "name": "deletebackuprequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "DownloadRequest",
                  "metadata": {
                    "name": "downloadrequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "PodVolumeBackup",
                  "metadata": {
                    "name": "podvolumebackup",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "PodVolumeRestore",
                  "metadata": {
                    "name": "podvolumerestore",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "ResticRepository",
                  "metadata": {
                    "name": "resticrepository",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "Restore",
                  "metadata": {
                    "name": "restore",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "Schedule",
                  "metadata": {
                    "name": "schedule",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "ServerStatusRequest",
                  "metadata": {
                    "name": "serverstatusrequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "VolumeSnapshotLocation",
                  "metadata": {
                    "name": "volumesnapshotlocation",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/konveyor/oadp-operator@sha256:ecb8534fbc245e890b9c7766946d01e1a8b4ce9be359c4ada4c95fd908b679b3
          createdAt: "2020-09-08T12:21:00Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          olm.skipRange: '>=0.0.0 <0.4.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Velero is a tool to safely backup and restore, perform disaster
              recovery and migrate Kubernetes cluster resources and persistent volumes.
            displayName: Velero
            kind: Velero
            name: veleroes.oadp.openshift.io
            version: v1alpha1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - description: Backup is a Velero resource that respresents the capture
              of Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Velero on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://oadp.konveyor.io/
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/konveyor/velero-restic-restore-helper@sha256:c4d12829e42cabe70b662a659875900677dfb415f912f931c3bf4c09a26aed5b
        - quay.io/konveyor/openshift-velero-plugin@sha256:aec8b4fc66cfcebd5a35ce99d8c985453ebf25c11453d30256089a8b425990b5
        - quay.io/konveyor/velero-plugin-for-aws@sha256:2197c6fef32910747526d33c1462b75cb7940aeace65791d0a5ddaa038480e52
        - quay.io/konveyor/velero-plugin-for-csi@sha256:5c2261effc902d8baa16e38265f9ad1fbe5284a3e6abd2bbc97761c161fa92cf
        - quay.io/konveyor/registry@sha256:b74ab3278b79eb89f99b6970c6aa0f5306221a8495b5496faad6cab2f74ff2cd
        - quay.io/konveyor/oadp-operator@sha256:ecb8534fbc245e890b9c7766946d01e1a8b4ce9be359c4ada4c95fd908b679b3
        - quay.io/konveyor/velero-plugin-for-microsoft-azure@sha256:62a344e5cf196e55dae92e537648987d758ccc75985bc45e9be6d0646a9e64d0
        - quay.io/konveyor/velero-plugin-for-gcp@sha256:3300458b1d6c1514dca7d24353ad03bd92db396b2a29f7d3522563160a5f5219
        - quay.io/konveyor/velero@sha256:e7865cc707664eb6ee767e77ccd933420604c2562736b53e16c8043a51df9d86
        version: 0.4.2
      entries:
      - name: oadp-operator.v0.4.2
        version: 0.4.2
      - name: oadp-operator.v0.4.1
        version: 0.4.1
      - name: oadp-operator.v0.4.0
        version: 0.4.0
      - name: oadp-operator.v0.3.0
        version: 0.3.0
      name: beta
    - currentCSV: oadp-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "restic": {
                      "enable": true
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ResticRepository",
                "metadata": {
                  "name": "resticrepository",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/konveyor/oadp-operator@sha256:90609044d52173619ea1e63ddef0ef3e7c1c939ac737b64bd427cbed5374dd27
          createdAt: "2020-09-08T12:21:00Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          olm.skipRange: '>=0.0.0 <0.5.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is a Velero resource that respresents the capture
              of Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Data Protection Applications on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://oadp.konveyor.io/
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/konveyor/registry@sha256:1be970879206258335357e3cf630138c24b8a5ccab35aae05eccb260e5e8ffb4
        - quay.io/konveyor/openshift-velero-plugin@sha256:221ab0bcd1d95fc166ef7319d187e196b59d2fbf9eefd29d91bd6435d45b5c3c
        - quay.io/konveyor/oadp-operator@sha256:90609044d52173619ea1e63ddef0ef3e7c1c939ac737b64bd427cbed5374dd27
        - quay.io/konveyor/velero@sha256:0ea5336dcbf0a1fab561697b24da01a0e5b533ee23553fbfd5f8e002f5a31c3e
        - quay.io/konveyor/velero-restic-restore-helper@sha256:c16248d027c09cd5b7360ae38484df88ca1cc01f9e470e3eade95076ee1a6b84
        - quay.io/konveyor/velero-plugin-for-aws@sha256:1e8bc27d6b60000e1406085d582910cb530af3fb901ef9bc780628df6c766288
        - quay.io/konveyor/velero-plugin-for-microsoft-azure@sha256:2d0089cbe715b482fb42e1acd9add70b975a8575636d8b143e6f90cef317ea13
        - quay.io/konveyor/velero-plugin-for-gcp@sha256:ee463ed2a8019c560961b118f8a87b587ed302f24fd25480c5d4ddad2557e92f
        - quay.io/konveyor/velero-plugin-for-csi@sha256:8c36ccfd294dafc415a7f113770cbbd66dbf6ecf8fef3bc4e73adc9878c606b0
        version: 0.5.6
      entries:
      - name: oadp-operator.v0.5.6
        version: 0.5.6
      - name: oadp-operator.v0.5.5
        version: 0.5.5
      - name: oadp-operator.v0.5.4
        version: 0.5.4
      - name: oadp-operator.v0.5.3
        version: 0.5.3
      - name: oadp-operator.v0.5.2
        version: 0.5.2
      - name: oadp-operator.v0.5.1
        version: 0.5.1
      - name: oadp-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: oadp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Trustification
      provider-url: ""
    name: trustify-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trustify-operator.v0.1.0-alpha.9
      currentCSVDesc:
        annotations:
          alm-examples: |
            [{
              "apiVersion": "org.trustify/v1alpha1",
              "kind": "Trustify",
              "metadata": {
                "name": "myapp"
              },
              "spec": { }
            }]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          containerImage: ghcr.io/trustification/trustify-operator:v0.1.0-alpha.9
          createdAt: "2025-03-07T11:57:13+00:00"
          description: An Operator for installing and managing Trustify
          repository: https://github.com/trustification/trustify-operator
          support: https://github.com/trustification/trustify-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Trustify instance
            displayName: Trustify
            kind: Trustify
            name: trustifies.org.trustify
            version: v1alpha1
        description: |+
          Trustify is vendor-neutral, thought-leadering, mostly informational collection of resources devoted to making Software Supply Chains easier to create, manage, consume and ultimately... to trust!


          You can use Trustify for:

          - Store and Manage all the SBOM (Software Bill of Materials) files of your company.
          - Understand which are the Vulnerabilities that affect each of your SBOM files
          - Understand exactly which Packages are included/shipped within each SBOM

          Trustify is a project within the [Trustification community](https://trustification.io/).


          ### Install
          Once you have successfully installed the Operator, proceed to deploy components by creating the required CR.

          By default, the Operator installs the following components on a target cluster:

          * Server
          * UI

          ### Documentation
          Documentation can be found on our [website](https://trustification.io/).

          ### Getting help
          If you encounter any issues while using Trustify, you can create an issue on our [Github repo](https://github.com/trustification/trustify/issues), for bugs, enhancements or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find
          * Fixing issues by opening [Pull Requests](https://github.com/trustification/trustify/pulls)

        displayName: Trustify Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trust
        links:
        - name: Website
          url: https://trustification.io/
        - name: Github
          url: https://github.com/trustification/trustify
        maintainers:
        - email: trustification@googlegroups.com
          name: Trustification
        maturity: alpha
        minKubeVersion: 1.23.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Trustification
        relatedImages:
        - ghcr.io/trustification/trustify-operator:v0.1.0-alpha.9
        version: 0.1.0-alpha.9
      entries:
      - name: trustify-operator.v0.1.0-alpha.9
        version: 0.1.0-alpha.9
      - name: trustify-operator.v0.1.0-alpha.8
        version: 0.1.0-alpha.8
      - name: trustify-operator.v0.1.0-alpha.7
        version: 0.1.0-alpha.7
      - name: trustify-operator.v0.1.0-alpha.6
        version: 0.1.0-alpha.6
      - name: trustify-operator.v0.1.0-alpha.5
        version: 0.1.0-alpha.5
      - name: trustify-operator.v0.1.0-alpha.4
        version: 0.1.0-alpha.4
      - name: trustify-operator.v0.1.0-alpha.3
        version: 0.1.0-alpha.3
      - name: trustify-operator.v0.1.0-alpha.2
        version: 0.1.0-alpha.2
      name: alpha
    defaultChannel: alpha
    packageName: trustify-operator
    provider:
      name: Trustification
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-directory-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-directory-operator.v23.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyDirectory",
                "metadata": {
                  "name": "ibmsecurityverifydirectory-sample"
                },
                "spec": {
                  "pods": {
                    "configMap": {
                      "proxy": {
                        "key": "config.yaml",
                        "name": "isvd-proxy-config"
                      },
                      "server": {
                        "key": "config.yaml",
                        "name": "isvd-server-config"
                      }
                    },
                    "image": {
                      "label": "latest",
                      "repo": "icr.io/isvd"
                    },
                    "proxy": {
                      "replicas": 1
                    },
                    "serviceAccountName": "isvd"
                  },
                  "replicas": {
                    "pvcs": [
                      "replica-1",
                      "replica-2"
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: icr.io/isvd/verify-directory-operator@sha256:d1dd7037584c6818dcf8d5aa69472278833bb8e37fba2aec5a69185ba88dcc34
          createdAt: "2023-04-18T22:10:32Z"
          description: The IBM Security Verify Directory operator is used to scale
            an IBM Security Verify Directory Server environment.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ibm-security/verify-directory-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyDirectory is the Schema for the ibmsecurityverifydirectories
              API
            displayName: IBMSecurity Verify Directory
            kind: IBMSecurityVerifyDirectory
            name: ibmsecurityverifydirectories.ibm.com
            version: v1
        description: "IBM Security Verify Directory is a scalable, standards-based
          identity directory that helps simplify identity and directory management.
          Verify Directory helps consolidate identity silos into a single identity
          source. Verify Directory is purpose-built to provide a directory foundation
          that can help provide a trusted identity data infrastructure that assists
          in enabling mission-critical security and authentication. It is designed
          to deliver a reliable, scalable, standards-based identity data platform
          that interoperates with a broad range of operating systems and applications.
          Verify Directory supports Lightweight Directory Access Protocol (LDAP) V3,
          offering a flexible and highly scalable LDAP infrastructure. \n\nThe operator
          can be used to manage deployed instances of the IBM Security Verify Directory
          server, providing an easy mechanism to scale the number of servers which
          are used to handle the workload.\n\nSee the project [Readme](https://github.com/ibm-security/verify-directory-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Security Verify Directory
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Security
        - Identity and Access
        links:
        - name: IBM Security Verify Directory Documentation
          url: https://www.ibm.com/docs/en/svd
        maintainers:
        - email: iamdev@us.ibm.com
          name: IBM Security Verify Directory Development Team
        maturity: stable
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - icr.io/isvd/verify-directory-operator@sha256:d1dd7037584c6818dcf8d5aa69472278833bb8e37fba2aec5a69185ba88dcc34
        version: 23.4.1
      entries:
      - name: ibm-security-verify-directory-operator.v23.4.1
        version: 23.4.1
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-directory-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Flux project
      provider-url: ""
    name: flux
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.toolkit.fluxcd.io/v2beta1",
                "kind": "HelmRelease",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "5m",
                    "chart": {
                      "spec": {
                          "chart": "podinfo",
                          "version": "4.0.x",
                          "sourceRef": {
                            "kind": "HelmRepository",
                            "name": "podinfo",
                            "namespace": "flux-system"
                          },
                          "interval": "1m"
                      }
                    },
                    "values": {
                      "replicaCount": 2
                    }
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "GitRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "1m",
                    "ref": {
                      "branch": "master"
                    },
                    "url": "https://github.com/stefanprodan/podinfo"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmRepository",
                "metadata": {
                    "name": "stable",
                    "namespace": "default"
                },
                "spec": {
                    "url": "https://kubernetes-charts.storage.googleapis.com/",
                    "interval": "10m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmChart",
                "metadata": {
                    "name": "redis",
                    "namespace": "default"
                },
                "spec": {
                    "chart": "redis",
                    "version": "10.5.7",
                    "sourceRef": {
                      "name": "stable",
                      "kind": "HelmRepository"
                    },
                    "interval": "5m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "Bucket",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "1m",
                    "provider": "generic",
                    "bucketName": "podinfo",
                    "endpoint": "minio.minio.svc.cluster.local:9000",
                    "insecure": true,
                    "secretRef": {
                      "name": "minio-credentials"
                    }
                }
              },
              {
                "apiVersion": "kustomize.toolkit.fluxcd.io/v1beta2",
                "kind": "Kustomization",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "5m",
                    "path": "./deploy/overlays/dev",
                    "prune": true,
                    "sourceRef": {
                      "kind": "GitRepository",
                      "name": "podinfo"
                    },
                    "timeout": "2m",
                    "wait": true
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "address": "https://github.com/stefanprodan/podinfo",
                    "secretRef": {
                      "name": "api-token"
                    }
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Alert",
                "metadata": {
                    "name": "webapp",
                    "namespace": "default"
                },
                "spec": {
                    "providerRef": {
                      "name": "on-call-slack"
                    },
                    "eventSeverity": "info",
                    "eventSources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "Bucket",
                          "name": "secrets"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-backend"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-frontend"
                      }
                    ]
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Receiver",
                "metadata": {
                    "name": "github-receiver",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "events": [
                      "ping",
                      "push"
                    ],
                    "secretRef": {
                      "name": "webhook-token"
                    },
                    "resources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "HelmRepository",
                          "name": "webapp"
                      }
                    ]
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "image": "ghcr.io/stefanprodan/podinfo",
                    "interval": "1m0s"
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImagePolicy",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "imageRepositoryRef": {
                      "name": "podinfo"
                    },
                    "policy": {
                      "semver": {
                          "range": "5.0.x"
                      }
                    }
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageUpdateAutomation",
                "metadata": {
                  "name": "flux-system",
                  "namespace": "flux-system"
                },
                "spec": {
                  "interval": "1m0s",
                  "sourceRef": {
                    "kind": "GitRepository",
                    "name": "flux-system"
                  },
                  "git": {
                    "checkout": {
                      "ref": {
                        "branch": "main"
                      }
                    },
                    "commit": {
                      "author": {
                        "email": "fluxcdbot@users.noreply.github.com",
                        "name": "fluxcdbot"
                      },
                      "messageTemplate": "{{range .Updated.Images}}{{println .}}{{end}}"
                    },
                    "push": {
                      "branch": "main"
                    }
                  },
                  "update": {
                    "path": "./clusters/my-cluster",
                    "strategy": "Setters"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.connect.redhat.com/flux/source-controller@sha256:b576e3f010e3ff07e716064ab0dad042aba93f0ea5c250535dd18bd678c45337
          description: Flux is a Continuous Delivery solution for Kubernetes.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/flux/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/flux/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: flux-system
          support: The Flux community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta1
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta2
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta2
          - description: OCIRepository
            displayName: OCIRepository
            kind: OCIRepository
            name: ocirepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta2
        description: |-
          [Flux](https://fluxcd.io/) is a Continuous Delivery solution for Kubernetes.
          Flux is constructed with the [GitOps Toolkit](https://toolkit.fluxcd.io/components/), a set of composable APIs and specialized tools for keeping Kubernetes clusters in sync with sources of configuration (like Git & Helm repositories), and automating updates to configuration when there is new code to deploy.

          Flux is a CNCF incubation project made for:

          - **cluster operators** who automate provision and configuration of clusters
          - **platform engineers** who build continuous delivery for developer teams
          - **app developers** who rely on continuous delivery to get their code live

          Features:

          - Source configuration from Git and Helm repositories, and S3-compatible buckets (e.g., Minio)
          - Kustomize and Helm support
          - Event-triggered and periodic reconciliation
          - Integration with Kubernetes RBAC
          - Health assessment (clusters and workloads)
          - Dependency management (infrastructure and workloads)
          - Secrets management (builtin support for Mozilla SOPS and Cloud KMS)
          - Alerting to external systems (webhook senders)
          - External events handling (webhook receivers)
          - Automated container image updates to Git (image scanning and patching)
          - Policy-driven validation (OPA, Kyverno, admission controllers)
          - Seamless integration with Git providers (GitHub, GitLab, Bitbucket)
          - Interoperability with Cluster API (CAPI) providers

          ### Steps to support bootstraping with Flux CLI

          SecComp Profile should be removed as it requires a special privilege to get the access.
          Also, it requires to enforce `runUserAs` to the same UID provided by the images to prevent OpenShift
          to alter the user id expected by our controllers.

          To do so, please specifiy a patch using the following Kustomization file, for example.

          ```yaml
          apiVersion: kustomize.config.k8s.io/v1beta1
          kind: Kustomization
          resources:
            - gotk-components.yaml
            - gotk-sync.yaml
          patches:
            - patch: |
                apiVersion: apps/v1
                kind: Deployment
                metadata:
                  name: all
                spec:
                  template:
                    spec:
                      containers:
                        - name: manager
                          securityContext:
                            runAsUser: 65534
                            seccompProfile:
                              $patch: delete
              target:
                kind: Deployment
                labelSelector: app.kubernetes.io/part-of=flux
          ```

          Then before installing Flux, please also perform the following steps to bind Flux to the `nonroot` SecurityContextConstraint (SCC). This applies to installation via OC's OLM or via Flux CLI.

          ```sh
          # replace default "flux-system" with the project/namespace name for your flux installation
          namespace="flux-system"
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:source-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:kustomize-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:helm-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:notification-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-automation-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-reflector-controller

          ```

          To see complete steps please refer to this [documentation](https://github.com/fluxcd/website/blob/main/content/en/docs/use-cases/openshift.md).

          If you've already installed Flux, you might need to uninstall it after applying the above pre-requisite steps.

          ### Network Policies

          This version of Flux is shipped without the normally provided network Policies.
          By default all traffic is enabled, and anyone can access the source-controller
          artifacts.

          It is not recommended to run it in production as is.

          Please visit [Flux's policies](https://github.com/fluxcd/flux2/tree/main/manifests/policies)
          to deploy the needed network policies and secure your Flux setup.

          ### Breaking Changes

          - Flux 0.18.x ships with v1beta2 of the kustomize.toolkit.fluxcd.io/Kustomization CRD.
          - Flux 0.25.x is the last Flux release where Kubernetes 1.19 is supported.

          ### Flux on OpenShift Use Cases

          - [The GitOps with Flux + OpenShift](https://www.youtube.com/watch?v=8ybCtNVP5To) talk by Andrew Block (Red Hat) showing how to setup multi-tenacy on OpenShift using Flux.
          - [Flux OpenShift Use Case](https://fluxcd.io/docs/use-cases/openshift/) on FluxCD.io
        displayName: Flux
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - gitops
        - cd
        - continuous delivery
        - fluxcd
        - helm
        - kustomize
        links:
        - name: Homepage
          url: https://fluxcd.io/
        - name: Documentation
          url: https://fluxcd.io/docs/
        - name: Slack
          url: https://cloud-native.slack.com/messages/flux
        - name: Mailing List
          url: https://lists.cncf.io/g/cncf-flux-dev
        - name: Source Code
          url: https://github.com/fluxcd/flux2
        maintainers:
        - email: support@weave.works
          name: The Flux maintainers
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: The Flux project
        relatedImages:
        - registry.connect.redhat.com/flux/helm-controller@sha256:cc24a7b73b023ad86a6c3a845088d153ea290dda7cab1ec03eb85aaf814e6049
        - registry.connect.redhat.com/flux/image-automation-controller@sha256:0c51595307e14acc15f22d50b414a87513ebb03530ef4a8f7d6cd9f8b757243e
        - registry.connect.redhat.com/flux/image-reflector-controller@sha256:3bef7fde22be641304e5ca69f77150e416eba5141f8aa13f68be7f810e55aa59
        - registry.connect.redhat.com/flux/kustomize-controller@sha256:246290908d899b9f6feb920f88913bd2d4673cf852a3020714fd4152c7f78237
        - registry.connect.redhat.com/flux/notification-controller@sha256:b3d81234ba3312e0106f2f7cabc73fbe1d7bf683eda36b9ed5bbdaa43789be40
        - registry.connect.redhat.com/flux/source-controller@sha256:b576e3f010e3ff07e716064ab0dad042aba93f0ea5c250535dd18bd678c45337
        version: 2.0.1
      entries:
      - name: flux.v2.0.1
        version: 2.0.1
      name: stable
    defaultChannel: stable
    packageName: flux
    provider:
      name: The Flux project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: telemetry-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: telemetry-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "Autoscaling",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "telemetry-operator",
                    "app.kubernetes.io/instance": "autoscaling-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "autoscaling",
                    "app.kubernetes.io/part-of": "telemetry-operator"
                  },
                  "name": "autoscaling"
                },
                "spec": {
                  "aodh": {
                    "databaseAccount": "aodh",
                    "databaseInstance": "openstack",
                    "memcachedInstance": "memcached",
                    "passwordSelectors": null,
                    "secret": "osp-secret"
                  }
                }
              },
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "Ceilometer",
                "metadata": {
                  "name": "ceilometer",
                  "namespace": "openstack"
                },
                "spec": {
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "Logging",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "telemetry-operator",
                    "app.kubernetes.io/instance": "logging-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "logging",
                    "app.kubernetes.io/part-of": "telemetry-operator"
                  },
                  "name": "logging"
                },
                "spec": {
                  "cloNamespace": "openshift-logging",
                  "port": 10514
                }
              },
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "MetricStorage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "telemetry-operator",
                    "app.kubernetes.io/instance": "metricstorage-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "metricstorage",
                    "app.kubernetes.io/part-of": "telemetry-operator"
                  },
                  "name": "metricstorage-sample"
                },
                "spec": {
                  "monitoringStack": {
                    "alertingEnabled": true,
                    "scrapeInterval": "30s",
                    "storage": {
                      "persistent": {
                        "pvcStorageRequest": "20G"
                      },
                      "retention": "24h",
                      "strategy": "persistent"
                    }
                  }
                }
              },
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "Telemetry",
                "metadata": {
                  "name": "telemetry"
                },
                "spec": {
                  "autoscaling": {
                    "aodh": {
                      "databaseAccount": "aodh",
                      "databaseInstance": "openstack",
                      "memcachedInstance": "memcached",
                      "passwordSelectors": null,
                      "secret": "osp-secret"
                    },
                    "enabled": false,
                    "heatInstance": "heat"
                  },
                  "ceilometer": {
                    "enabled": true,
                    "secret": "osp-secret"
                  },
                  "logging": {
                    "enabled": false,
                    "port": 10514
                  },
                  "metricStorage": {
                    "enabled": false,
                    "monitoringStack": {
                      "alertingEnabled": true,
                      "scrapeInterval": "30s",
                      "storage": {
                        "persistent": {
                          "pvcStorageRequest": "20G"
                        },
                        "retention": "24h",
                        "strategy": "persistent"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:47Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Autoscaling is the Schema for the autoscalings API
            displayName: Autoscaling
            kind: Autoscaling
            name: autoscalings.telemetry.openstack.org
            version: v1beta1
          - description: Ceilometer is the Schema for the ceilometers API
            displayName: Ceilometer
            kind: Ceilometer
            name: ceilometers.telemetry.openstack.org
            version: v1beta1
          - description: Logging is the Schema for the loggings API
            displayName: Logging
            kind: Logging
            name: loggings.telemetry.openstack.org
            version: v1beta1
          - description: MetricStorage is the Schema for the metricstorages API
            displayName: Metric Storage
            kind: MetricStorage
            name: metricstorages.telemetry.openstack.org
            version: v1beta1
          - description: Telemetry is the Schema for the telemetry API
            displayName: Telemetry
            kind: Telemetry
            name: telemetries.telemetry.openstack.org
            version: v1beta1
        description: Telemetry Operator
        displayName: Telemetry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Telemetry
        - Ceilometer
        links:
        - name: Telemetry Operator
          url: https://github.com/openstack-k8s-operators/telemetry-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-ceilometer-notification-rhel9@sha256:239e5376f65682fd644e8d9b10b27ad879747a056c8fd26ff19882e34c5bb056
        - registry.redhat.io/rhoso/openstack-ceilometer-ipmi-rhel9@sha256:86618d8f54df09210349e382188f3b6563ab96ee511fb4da6eb6dc80e4027aae
        - registry.redhat.io/rhoso/openstack-aodh-listener-rhel9@sha256:3599a25515a96b3ff4768f85fecb4bf955789fb9b5d8abc3181678062020ae64
        - registry.redhat.io/rhoso/openstack-ceilometer-compute-rhel9@sha256:b8a0412041c50e312bef39c17a9c2600e9dbb42ea380d55e58243cda641cef4f
        - registry.redhat.io/rhoso/openstack-aodh-evaluator-rhel9@sha256:16027b7d4f07385c87f97e1581d4454d4ce7ed654499720d9ba5bc6fc74fc246
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/telemetry-rhel9-operator@sha256:83962f19193f98b87c54136f02d18797a8047f65e0d75c11ce9518e50aabc050
        - registry.redhat.io/rhoso/openstack-ceilometer-central-rhel9@sha256:1c79d1edb2404a8614ea29f24e2a869695d75b1695ea0937eebc9c57250fc4ab
        - registry.redhat.io/rhoso/openstack-aodh-api-rhel9@sha256:5ab39a468ab014f2fec4cb9cd210b430ae1c1262f52b9e0f22881d6d8e9def97
        - registry.redhat.io/rhoso/openstack-aodh-notifier-rhel9@sha256:f729548ac3d718f2f5b6c85045e7c676e1656207ee54f43540d246bd97426894
        - registry.redhat.io/rhoso-operators/sg-core-rhel9@sha256:a6e4edc24409361cb770bb44fe5b475dc58748426ccc4003abad7e70f01f1b97
        - registry.redhat.io/ubi9/httpd-24@sha256:dbfbe3383e12d63df7178e4cdd5b1e6e51a84c781041b9192026ea6edb72fbf3
        version: 1.0.6
      entries:
      - name: telemetry-operator.v1.0.6
        version: 1.0.6
      - name: telemetry-operator.v1.0.5
        version: 1.0.5
      - name: telemetry-operator.v1.0.4
        version: 1.0.4
      - name: telemetry-operator.v1.0.3
        version: 1.0.3
      - name: telemetry-operator.v1.0.2
        version: 1.0.2
      - name: telemetry-operator.v1.0.1
        version: 1.0.1
      - name: telemetry-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: telemetry-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Snyk Ltd.
      provider-url: ""
    name: snyk-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: snyk-operator.v1.92.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1alpha1",
                "kind": "SnykMonitor",
                "metadata": {
                  "name": "snyk-monitor"
                },
                "spec": {
                  "clusterName": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "snyk/kubernetes-monitor",
                    "tag": "1.92.0-ubi8"
                  },
                  "integrationApi": "",
                  "monitorSecrets": "snyk-monitor",
                  "scope": "Cluster",
                  "temporaryStorageSize": "50Gi",
                  "pvc": {
                    "enabled": false,
                    "create": false,
                    "name": "snyk-monitor-pvc",
                    "storageClassName": null
                  },
                  "initContainerImage": {
                    "repository": "busybox",
                    "tag": "latest"
                  },
                  "requests": {
                    "memory": "400Mi"
                  },
                  "limits": {
                    "memory": "2Gi"
                  },
                  "nodeAffinity": {
                    "disableBetaArchNodeSelector": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Security
          certified: "false"
          containerImage: docker.io/snyk/kubernetes-operator:1.92.0
          createdAt: "2022-06-09T17:49:36Z"
          description: A Kubernetes Operator for creating and managing Snyk controller
            instances.
          olm.skipRange: <1.68.2
          repository: https://github.com/snyk/kubernetes-monitor
          support: Snyk Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a deployment of the Snyk controller, which scans
              container images in the cluster for vulnerabilities.
            displayName: Snyk Monitor
            kind: SnykMonitor
            name: snykmonitors.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          A Kubernetes Operator for creating and managing Snyk Kubernetes controller instances.

          Snyk integrates with Kubernetes, enabling you to import and test your running workloads and identify vulnerabilities in their associated images and configurations that might make those workloads less secure. Once imported, Snyk continues to monitor those workloads, identifying additional security issues as new images are deployed and the workload configuration changes.

          ## Prerequisites

          To start using the Operator and the Snyk controller that it manages, you will need to create a secret containing your Snyk integrationId and Docker config file:

          ```
          kubectl create namespace snyk-monitor
          kubectl create secret generic snyk-monitor -n snyk-monitor --from-literal=dockercfg.json={} --from-literal=integrationId=${SNYK_INTEGRATION_ID}
          ```

          Refer to the [Snyk documentation](https://support.snyk.io/hc/en-us/articles/360003916138-Kubernetes-integration-overview) for more details.
        displayName: Snyk Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - snyk
        - controller
        - monitor
        - operator
        - security
        - cluster
        - container
        - image
        - security
        - vulnerabilities
        - advisories
        - audit
        - scan
        - scanning
        links:
        - name: Website
          url: https://snyk.io
        - name: Documentation
          url: https://support.snyk.io/hc/en-us/articles/360003916138-Kubernetes-integration-overview
        maintainers:
        - email: support@snyk.io
          name: Snyk Ltd.
        maturity: stable
        provider:
          name: Snyk Ltd.
        relatedImages:
        - docker.io/snyk/kubernetes-operator:1.92.0
        version: 1.92.0
      entries:
      - name: snyk-operator.v1.92.0
        version: 1.92.0
      - name: snyk-operator.v1.90.2
        version: 1.90.2
      - name: snyk-operator.v1.90.1
        version: 1.90.1
      - name: snyk-operator.v1.88.4
        version: 1.88.4
      - name: snyk-operator.v1.85.2
        version: 1.85.2
      - name: snyk-operator.v1.82.1
        version: 1.82.1
      - name: snyk-operator.v1.81.1
        version: 1.81.1
      - name: snyk-operator.v1.80.0
        version: 1.80.0
      - name: snyk-operator.v1.73.0
        version: 1.73.0
      - name: snyk-operator.v1.72.0
        version: 1.72.0
      - name: snyk-operator.v1.70.4
        version: 1.70.4
      - name: snyk-operator.v1.70.1
        version: 1.70.1
      name: stable
    defaultChannel: stable
    packageName: snyk-operator
    provider:
      name: Snyk Ltd.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Aqua Security, Inc.
      provider-url: ""
    name: aqua
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aqua-operator.v2022.4.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaCsp",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "namespace": "aqua",
                    "version": "2022.4",
                    "requirements": true
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry",
                    "dbDiskSize": 10,
                    "databaseSecret": {
                      "key": "db-password",
                      "name": "aqua-database-password"
                    }
                  },
                  "database": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "database",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "gateway": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "gateway",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "server": {
                    "replicas": 1,
                    "service": "LoadBalancer",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "console",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "route": true,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaDatabase",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "database",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "diskSize": 10,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "image": {
                      "repository": "enforcer",
                      "registry": "registry.aquasec.com",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "IfNotPresent"
                    }
                  },
                  "runAsNonRoot": false,
                  "gateway": {
                    "host": "aqua-gateway",
                    "port": 8443
                  },
                  "token": "<<your-token>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaKubeEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "config": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "aqua-secure",
                    "imagePullSecret": "aqua-registry",
                    "kubebench_image": "docker.io/aquasec/kube-bench:v0.7.1"
                  },
                  "deploy": {
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "tag": "<<KUBE_ENFORCER_TAG>>",
                      "repository": "kube-enforcer",
                      "pullPolicy": "Always"
                    }
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "starboard-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  }
                }
              },
              {
              "apiVersion": "operator.aquasec.com/v1alpha1",
              "kind": "AquaLightning",
              "metadata": {
                "name": "aqua"
              },
              "spec": {
                "global": {
                  "gateway_address": "aqua-gateway.aqua:8443",
                  "cluster_name": "Default-cluster-name"
                },
                "common": {
                  "imagePullSecret": "aqua-registry"
                },
                "kubeEnforcer": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                  "allowAnyVersion": null,
                  "deploy": {
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "tag": "2022.4",
                      "repository": "kube-enforcer",
                      "pullPolicy": "Always"
                    }
                  },
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "starboard-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  },
                  "env": []
                },
                "enforcer": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "deploy": {
                    "image": {
                      "repository": "enforcer",
                      "registry": "registry.aquasec.com",
                      "tag": "2022.4",
                      "pullPolicy": "IfNotPresent"
                    }
                  },
                  "token": "<<ENFORCER_GROUP_TOKEN>>",
                  "env": [],
                  "secret": {
                    "name": null,
                    "key": null
                  },
                  "runAsNonRoot": null,
                  "rhcosVersion": null
                }
              }
            },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaGateway",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "gateway",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaScanner",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "replicas": 1,
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "scanner",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "runAsNonRoot": false,
                  "login": {
                    "username": "<<YOUR AQUA USER NAME>>",
                    "password": "<<YOUR AQUA USER PASSWORD>>",
                    "host": "http://aqua-server:8080",
                    "token": ""
                  }
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaServer",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "LoadBalancer",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "console",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "runAsNonRoot": false
                }
              },
              {
                  "apiVersion": "operator.aquasec.com/v1alpha1",
                  "kind": "AquaCloudConnector",
                  "metadata": {
                    "name": "aqua"
                  },
                  "spec": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "common": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "image": {
                        "repository": "aqua-cloud-connector",
                        "registry": "registry.aquasec.com",
                        "tag": "<<IMAGE TAG>>"
                      },
                      "replicas": 1
                    },
                    "runAsNonRoot": false,
                    "gateway": {
                      "host": "aqua-gateway",
                      "port": "8443"
                    },
                    "login": {
                      "host": "aqua-gateway:8443",
                      "password": "<<YOUR AQUA USER PASSWORD>>",
                      "token": "<<YOUR AQUA CLOUD CONNECTOR TOKEN>>",
                      "username": "<<YOUR AQUA USER NAME>>"
                    },
                    "tunnels": [
                      {
                        "host": "<account>.azurecr.io",
                        "port": "443"
                      },
                      {
                        "host": "gke.gcr.io",
                        "port": "443"
                      },
                      {
                        "host": "<account>.jfrog.io",
                        "port": "443"
                      },
                      {
                        "host": "1.1.1.1",
                        "port": "443"
                      },
                      {
                        "region": "us-east1",
                        "type": "ecr"
                      }
                    ]
                  }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: docker.io/aquasec/aqua-operator:2022.4.18
          createdAt: "2020-03-25T08:00:00Z"
          description: The Aqua Security Operator runs within a Openshift cluster
            and provides a means to deploy and manage Aqua Security cluster and components.
          operators.operatorframework.io/internal-objects: '["aquastarboards.aquasecurity.github.io","configauditreports.aquasecurity.github.io",
            "clusterconfigauditreports.aquasecurity.github.io"]'
          repository: https://github.com/aquasecurity/aqua-operator
          support: Aqua Security, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aqua Security CSP Deployment with Aqua Operator
            displayName: AquaCsp
            kind: AquaCsp
            name: aquacsps.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Lightning Enforcers Deployment with Aqua Operator
            displayName: AquaLightningEnforcers
            kind: AquaLightning
            name: aqualightnings.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Database Deployment with Aqua Operator
            displayName: AquaDatabase
            kind: AquaDatabase
            name: aquadatabases.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Enforcer Deployment with Aqua Operator
            displayName: AquaEnforcer
            kind: AquaEnforcer
            name: aquaenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security KubeEnforcer Deployment with Aqua Operator
            displayName: AquaKubeEnforcer
            kind: AquaKubeEnforcer
            name: aquakubeenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Gateway Deployment with Aqua Operator
            displayName: AquaGateway
            kind: AquaGateway
            name: aquagateways.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Scanner Deployment with Aqua Operator
            displayName: AquaScanner
            kind: AquaScanner
            name: aquascanners.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Server Deployment with Aqua Operator
            displayName: AquaServer
            kind: AquaServer
            name: aquaservers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Starboard for kube-enforcer.
            displayName: AquaStarboard
            kind: AquaStarboard
            name: aquastarboards.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ClusterConfigAuditReport
            kind: ClusterConfigAuditReport
            name: clusterconfigauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ConfigAuditReport
            kind: ConfigAuditReport
            name: configauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: Aqua Security Cloud Connector Deployment with Aqua Operator
            displayName: AquaCloudConnector
            kind: AquaCloudConnector
            name: aquacloudconnector.operator.aquasec.com
            version: v1alpha1
        description: |-
          The Aqua Security Operator runs within an OpenShift cluster and provides a means to deploy and manage the Aqua Security cluster and components :
          * Server (aka “console”)
          * Database (for production environments we recommend to use an external database and not the Aqua default database)
          * Gateway
          * Enforcer (aka “agent”)
          * KubeEnforcer
          * Scanner
          * CSP (package that contains the Server, Database, and Gateway)
          * LightningEnforcer (package that contains Enforcer and Kube Enforcer)

          Use the Aqua-Operator to
          * Deploy Aqua Security components on OpenShift
          * Scale up Aqua Security components with extra replicas
          * Assign metadata tags to Aqua Security components

          ## Before You Begin Using the Operator CRDs

          Obtain access to the Aqua registry - https://www.aquasec.com/about-us/contact-us/


          ## Instructions and exmaples

          Follow instructions and exmaples
          [here](https://github.com/aquasecurity/aqua-operator/blob/2022.4.0/docs/DeployOpenShiftOperator.md)
        displayName: Aqua Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aqua-security
        - scanning
        - security
        - runtime-security
        links:
        - name: Aqua Security
          url: https://www.aquasec.com/
        - name: Aqua Operator Github
          url: https://github.com/aquasecurity/aqua-operator
        maintainers:
        - email: support@aquasec.com
          name: Aqua Support, Aqua Security
        maturity: alpha
        provider:
          name: Aqua Security, Inc.
        relatedImages:
        - docker.io/aquasec/aqua-operator:2022.4.18
        version: 2022.4.18
      entries:
      - name: aqua-operator.v2022.4.18
        version: 2022.4.18
      - name: aqua-operator.v2022.4.17
        version: 2022.4.17
      name: 2022.4.0
    defaultChannel: 2022.4.0
    packageName: aqua
    provider:
      name: Aqua Security, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: The bpfman Community
      provider-url: https://bpfman.io/
    name: bpfman-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bpfman-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfapplication"
                  },
                  "name": "bpfapplication-sample"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:v0.5.6"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "kprobe": {
                        "bpffunctionname": "kprobe_counter",
                        "func_name": "try_to_wake_up",
                        "offset": 0,
                        "retprobe": false
                      },
                      "type": "Kprobe"
                    },
                    {
                      "tracepoint": {
                        "bpffunctionname": "tracepoint_kill_recorder",
                        "names": [
                          "syscalls/sys_enter_kill"
                        ]
                      },
                      "type": "Tracepoint"
                    },
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "containernames": [
                            "bpfman",
                            "bpfman-agent"
                          ],
                          "namespace": "bpfman",
                          "pods": {
                            "matchLabels": {
                              "name": "bpfman-daemon"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfNsApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfnsapplication"
                  },
                  "name": "bpfapplication-sample",
                  "namespace": "acme"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:v0.5.6"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FentryProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fentryprogram"
                  },
                  "name": "fentry-example"
                },
                "spec": {
                  "bpffunctionname": "test_fentry",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fentry:v0.5.6"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FexitProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fexitprogram"
                  },
                  "name": "fexit-example"
                },
                "spec": {
                  "bpffunctionname": "test_fexit",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fexit:v0.5.6"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "KprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kprobeprogram"
                  },
                  "name": "kprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_kprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/kprobe:v0.5.6"
                    }
                  },
                  "func_name": "try_to_wake_up",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "offset": 0,
                  "retprobe": false
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcnsprogram"
                  },
                  "name": "tc-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcprogram"
                  },
                  "name": "tc-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:v0.5.6"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxnsprogram"
                  },
                  "name": "tcx-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "tcx_next",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxprogram"
                  },
                  "name": "tcx-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "tcx_pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:v0.5.6"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TracepointProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tracepointprogram"
                  },
                  "name": "tracepoint-example"
                },
                "spec": {
                  "bpffunctionname": "enter_openat",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tracepoint:v0.5.6"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "names": [
                    "syscalls/sys_enter_openat"
                  ],
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobensprogram"
                  },
                  "name": "uprobe-example",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobeprogram"
                  },
                  "name": "uprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:v0.5.6"
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpnsprogram"
                  },
                  "name": "xdp-ns-pass-all-nodes",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpprogram"
                  },
                  "name": "xdp-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:v0.5.6"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/bpfman/bpfman-operator:v0.5.6
          createdAt: "2025-02-18T13:44:53Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: bpfman
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "bpfman",
                "labels": {
                  "pod-security.kubernetes.io/enforce": "privileged",
                  "pod-security.kubernetes.io/audit": "privileged",
                  "pod-security.kubernetes.io/warn": "privileged",
                },
                "annotations": {
                  "openshift.io/node-selector": ""
                },
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/bpfman/bpfman
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BpfApplication is the Schema for the bpfapplications API
            displayName: Bpf Application
            kind: BpfApplication
            name: bpfapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsApplication is the Schema for the BpfNsApplications
              API
            displayName: Bpf Namespaced Application
            kind: BpfNsApplication
            name: bpfnsapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsProgram is the Schema for the BpfNsProgram API
            displayName: Bpf Namespaced Program
            kind: BpfNsProgram
            name: bpfnsprograms.bpfman.io
            version: v1alpha1
          - description: BpfProgram is the Schema for the BpfProgram API
            displayName: Bpf Program
            kind: BpfProgram
            name: bpfprograms.bpfman.io
            version: v1alpha1
          - description: FentryProgram is the Schema for the Fentryprograms API
            displayName: Fentry Program
            kind: FentryProgram
            name: fentryprograms.bpfman.io
            version: v1alpha1
          - description: FexitProgram is the Schema for the Fexitprograms API
            displayName: Fexit Program
            kind: FexitProgram
            name: fexitprograms.bpfman.io
            version: v1alpha1
          - description: KprobeProgram is the Schema for the Kprobeprograms API
            displayName: Kprobe Program
            kind: KprobeProgram
            name: kprobeprograms.bpfman.io
            version: v1alpha1
          - description: TcNsProgram is the Schema for the Tcnsprograms API
            displayName: TC Namespaced Program
            kind: TcNsProgram
            name: tcnsprograms.bpfman.io
            version: v1alpha1
          - description: TcProgram is the Schema for the Tcprograms API
            displayName: TC Program
            kind: TcProgram
            name: tcprograms.bpfman.io
            version: v1alpha1
          - description: TcxNsProgram is the Schema for the Tcxnsprograms API
            displayName: TCX Namespaced Program
            kind: TcxNsProgram
            name: tcxnsprograms.bpfman.io
            version: v1alpha1
          - description: TcxProgram is the Schema for the Tcxprograms API
            displayName: TCX Program
            kind: TcxProgram
            name: tcxprograms.bpfman.io
            version: v1alpha1
          - description: TracepointProgram is the Schema for the Tracepointprograms
              API
            displayName: Tracepoint Program
            kind: TracepointProgram
            name: tracepointprograms.bpfman.io
            version: v1alpha1
          - description: UprobeNsProgram is the Schema for the Uprobensprograms API
            displayName: Uprobe Namespaced Program
            kind: UprobeNsProgram
            name: uprobensprograms.bpfman.io
            version: v1alpha1
          - description: UprobeProgram is the Schema for the Uprobeprograms API
            displayName: Uprobe Program
            kind: UprobeProgram
            name: uprobeprograms.bpfman.io
            version: v1alpha1
          - description: XdpNsProgram is the Schema for the Xdpnsprograms API
            displayName: Xdp Namespaced Program
            kind: XdpNsProgram
            name: xdpnsprograms.bpfman.io
            version: v1alpha1
          - description: XdpProgram is the Schema for the Xdpprograms API
            displayName: Xdp Program
            kind: XdpProgram
            name: xdpprograms.bpfman.io
            version: v1alpha1
        description: "The bpfman Operator is a Kubernetes Operator for deploying [bpfman](https://bpfman.netlify.app/),
          a system daemon\nfor managing eBPF programs. It deploys bpfman itself along
          with CRDs to make deploying\neBPF programs in Kubernetes much easier.\n\n##
          Quick Start\n\nTo get bpfman up and running quickly simply click 'install'
          to deploy the bpfman-operator in the bpfman namespace via operator-hub.\n##
          Configuration\n\nThe `bpfman-config` configmap is automatically created
          in the `bpfman` namespace and used to configure the bpfman deployment.\n\nTo
          edit the config simply run\n\n```bash\nkubectl edit cm bpfman-config\n```\n\nThe
          following fields are adjustable\n\n- `bpfman.agent.image`: The image used
          for the bpfman-agent`\n- `bpfman.image`: The image used for bpfman`\n- `bpfman.log.level`:
          the log level for bpfman, currently supports `debug`, `info`, `warn`, `error`,
          and `fatal`, defaults to `info`\n- `bpfman.agent.log.level`: the log level
          for the bpfman-agent currently supports `info`, `debug`, and `trace` \n\nThe
          bpfman operator deploys eBPF programs via CRDs. The following CRDs are currently
          available, \n\n- BpfApplication\n- FentryProgram\n- FexitProgram\n - KprobeProgram\n-
          TcProgram\n- TcxProgram\n- TracepointProgram\n- UprobeProgram\n- XdpProgram\n
          - BpfNsApplication\n- TcNsProgram\n- TcxNsProgram\n- UprobeNsProgram\n-
          XdpNsProgram\n- BpfProgram\n - BpfNsProgram\n\n ## More information\n\nPlease
          checkout the [bpfman community website](https://bpfman.io/) for more information."
        displayName: Bpfman Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ebpf
        - kubernetes
        links:
        - name: bpfman website
          url: https://bpfman.io/
        maintainers:
        - email: astoycos@redhat.com
          name: Andrew Stoycos
        - email: afredette@redhat.com
          name: Andre Fredette
        maturity: alpha
        provider:
          name: The bpfman Community
          url: https://bpfman.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/bpfman/bpfman-operator:v0.5.6
        version: 0.5.6
      entries:
      - name: bpfman-operator.v0.5.6
        version: 0.5.6
      - name: bpfman-operator.v0.5.5
        version: 0.5.5
      - name: bpfman-operator.v0.5.4
        version: 0.5.4
      - name: bpfman-operator.v0.5.2
        version: 0.5.2
      - name: bpfman-operator.v0.5.1
        version: 0.5.1
      - name: bpfman-operator.v0.5.0
        version: 0.5.0
      - name: bpfman-operator.v0.4.2
        version: 0.4.2
      - name: bpfman-operator.v0.4.1
        version: 0.4.1
      name: alpha
    defaultChannel: alpha
    packageName: bpfman-operator
    provider:
      name: The bpfman Community
      url: https://bpfman.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-organizations-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-organizations-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "organizations.services.k8s.aws/v1alpha1",
                "kind": "OrganizationalUnit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/organizations-controller:1.1.1
          createdAt: "2025-09-19T19:35:20Z"
          description: AWS Organizations controller is a service controller for managing
            Organizations resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Account represents the state of an AWS organizations Account
              resource.
            displayName: Account
            kind: Account
            name: accounts.organizations.services.k8s.aws
            version: v1alpha1
          - description: OrganizationalUnit represents the state of an AWS organizations
              OrganizationalUnit resource.
            displayName: OrganizationalUnit
            kind: OrganizationalUnit
            name: organizationalunits.organizations.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Organizations resources in AWS from within your Kubernetes cluster.

          **About Amazon Organizations**

          With AWS Organizations you can perform account management activities at scale by consolidating multiple AWS accounts into a single organization. Consolidating accounts simplifies how you use other AWS services. You can leverage the multi-account management services available in AWS Organizations with select AWS services to perform tasks on all accounts that are members of your organization.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Organizations
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - organizations
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Organizations Developer Resources
          url: https://aws.amazon.com/organizations/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: organizations maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/organizations-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-organizations-controller.v1.1.1
        version: 1.1.1
      - name: ack-organizations-controller.v1.0.13
        version: 1.0.13
      - name: ack-organizations-controller.v1.0.11
        version: 1.0.11
      - name: ack-organizations-controller.v1.0.10
        version: 1.0.10
      - name: ack-organizations-controller.v1.0.9
        version: 1.0.9
      - name: ack-organizations-controller.v1.0.8
        version: 1.0.8
      - name: ack-organizations-controller.v1.0.7
        version: 1.0.7
      - name: ack-organizations-controller.v1.0.6
        version: 1.0.6
      - name: ack-organizations-controller.v1.0.5
        version: 1.0.5
      - name: ack-organizations-controller.v1.0.4
        version: 1.0.4
      - name: ack-organizations-controller.v1.0.1
        version: 1.0.1
      - name: ack-organizations-controller.v1.0.0
        version: 1.0.0
      - name: ack-organizations-controller.v0.0.10
        version: 0.0.10
      - name: ack-organizations-controller.v0.0.9
        version: 0.0.9
      - name: ack-organizations-controller.v0.0.8
        version: 0.0.8
      - name: ack-organizations-controller.v0.0.7
        version: 0.0.7
      - name: ack-organizations-controller.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: ack-organizations-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-sds-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-sds-operator.v8.0.230
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.230"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleStoragePool",
                "metadata": {
                  "name": "infoscalestoragepool-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:f4947eab2b31fccc86b5d5d0d3fd4eb1f7513fd8325e379cc259f5ccd1dfc3e4
          createdAt: "2023-10-26T20:36:59Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          olm.skipRange: <8.0.230
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleStoragePool is the Schema for the infoscalestoragepools
              API
            displayName: InfoScale Storage Pool
            kind: InfoScaleStoragePool
            name: infoscalestoragepools.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.230"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:f4947eab2b31fccc86b5d5d0d3fd4eb1f7513fd8325e379cc259f5ccd1dfc3e4
        version: 8.0.230
      entries:
      - name: infoscale-sds-operator.v8.0.230
        version: 8.0.230
      - name: infoscale-sds-operator.v8.0.220
        version: 8.0.220
      - name: infoscale-sds-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-sds-operator.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-sds-operator.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.330"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:dab501b99373c840c7e8e3ab479fa273b39f9bf8a896f3f1e4e945f4ab81f64e
          createdAt: "2024-06-24T09:03:32Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.330"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:dab501b99373c840c7e8e3ab479fa273b39f9bf8a896f3f1e4e945f4ab81f64e
        version: 8.0.330
      entries:
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-sds-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.410"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
          createdAt: "2024-07-30T09:00:57Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Arctera InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/168721626-168743109-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.410"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Arctera recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
        version: 8.0.410
      entries:
      - name: infoscale-sds-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-sds-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-sds-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-sds-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.410"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
          createdAt: "2024-07-30T09:00:57Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Arctera InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/168721626-168743109-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.410"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Arctera recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
        version: 8.0.410
      entries:
      - name: infoscale-sds-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-sds-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-sds-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-sds-operator.v8.0.210
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.210"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:26573b98743245c0b607abbe0d981abce6ce9fb1c5c0cd9fe735351b7136f4e8
          createdAt: "2022-11-26T16:36:59Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleStoragePool is the Schema for the infoscalestoragepools
              API
            displayName: InfoScale Storage Pool
            kind: InfoScaleStoragePool
            name: infoscalestoragepools.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.210"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:1667227b5111fe52ed304ab19ff9fadfc677239e56b2c93a3b89b1bfbc31354d
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:26573b98743245c0b607abbe0d981abce6ce9fb1c5c0cd9fe735351b7136f4e8
        version: 8.0.210
      entries:
      - name: infoscale-sds-operator.v8.0.210
        version: 8.0.210
      - name: infoscale-sds-operator.v8.0.201
        version: 8.0.201
      - name: infoscale-sds-operator.v8.0.200
        version: 8.0.200
      - name: infoscale-sds-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-sds-operator.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-sds-operator
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: ovn-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ovn-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ovn.openstack.org/v1beta1",
                "kind": "OVNController",
                "metadata": {
                  "name": "ovncontroller-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "ovn.openstack.org/v1beta1",
                "kind": "OVNDBCluster",
                "metadata": {
                  "name": "ovndbcluster-nb-sample"
                },
                "spec": {
                  "dbType": "NB",
                  "logLevel": "info",
                  "storageClass": "local-storage",
                  "storageRequest": "10G"
                }
              },
              {
                "apiVersion": "ovn.openstack.org/v1beta1",
                "kind": "OVNDBCluster",
                "metadata": {
                  "name": "ovndbcluster-sb-sample"
                },
                "spec": {
                  "dbType": "SB",
                  "logLevel": "info",
                  "storageClass": "local-storage",
                  "storageRequest": "10G"
                }
              },
              {
                "apiVersion": "ovn.openstack.org/v1beta1",
                "kind": "OVNNorthd",
                "metadata": {
                  "name": "ovnnorthd-sample"
                },
                "spec": {
                  "logLevel": "info"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:45Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OVNController is the Schema for the ovncontrollers API
            displayName: OVNController
            kind: OVNController
            name: ovncontrollers.ovn.openstack.org
            version: v1beta1
          - description: OVNDBCluster is the Schema for the ovndbclusters API
            displayName: OVNDBCluster
            kind: OVNDBCluster
            name: ovndbclusters.ovn.openstack.org
            version: v1beta1
          - description: OVNNorthd is the Schema for the ovnnorthds API
            displayName: OVNNorthd
            kind: OVNNorthd
            name: ovnnorthds.ovn.openstack.org
            version: v1beta1
        description: OVN Operator
        displayName: OVN Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Network
        - Ovn
        links:
        - name: Ovn Operator
          url: https://github.com/openstack-k8s-operators/ovn-operator
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/ovn-rhel9-operator@sha256:bce19011b3623c66a59efe6ea874e2403440813bdc970cd6170853a8093076f3
        - registry.redhat.io/rhoso/openstack-ovn-base-rhel9@sha256:b0d424f6ea6de8192382541a95aa56fa32b9b95e8d433c41c4a6dadbe4127fae
        - registry.redhat.io/rhoso/openstack-ovn-nb-db-server-rhel9@sha256:e71c8aac00a78e3d786332aaada3657276f69483df1e441b70589601ca038a5b
        - registry.redhat.io/rhoso/openstack-ovn-sb-db-server-rhel9@sha256:7aa2832f80846d5ac71733a1f186d39dce3654ebb661d32b8e4580d92be8e338
        - registry.redhat.io/rhoso/openstack-ovn-northd-rhel9@sha256:f1e4ce04b67d888fa950b83d1f00459c67c602397336da81de3ff71d4052a5d8
        - registry.redhat.io/rhoso/openstack-ovn-controller-rhel9@sha256:6c7827dbe25c168d28250f197d5ea6f3f60cfbab4c974267ec5ea0f8d631c3f5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: ovn-operator.v1.0.6
        version: 1.0.6
      - name: ovn-operator.v1.0.5
        version: 1.0.5
      - name: ovn-operator.v1.0.4
        version: 1.0.4
      - name: ovn-operator.v1.0.3
        version: 1.0.3
      - name: ovn-operator.v1.0.2
        version: 1.0.2
      - name: ovn-operator.v1.0.1
        version: 1.0.1
      - name: ovn-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: ovn-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: fuse-apicurito
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-apicurito.v7.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1alpha1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:8d19ad579f874d6b9a08f127c8900906913a5a7c1e69cbe4b5cb0333b8d86152
          createdAt: 2022-02-28 09:35:38 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.10.3'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:8d19ad579f874d6b9a08f127c8900906913a5a7c1e69cbe4b5cb0333b8d86152
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:8dbeeb8b04c81f41d896f71db90813de57e67493fa85389b63ac23f390785043
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:c02c218ccfbb385ea29af98409995a3d9c26cb0b4c76caf2e0779fa233b5ff8c
        version: 7.10.3
      entries:
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.10.x
    - currentCSV: fuse-apicurito.v7.11.1-0.1687357785.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:2f8d4450187839f0e0c761e5ddcf61b03f5996acfe3d75b60120e76e5b0167c1
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.11.2'
          olm.substitutesFor: fuse-apicurito.v7.11.1-0.1682509045.p
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:2f8d4450187839f0e0c761e5ddcf61b03f5996acfe3d75b60120e76e5b0167c1
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:df61aafc4023b9068a8636c35eabb0d5fda400f523b64cb1fc5d8c480272c435
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:4c493a2d690d0d7c1684c53d15c06b47d691aeffdde6dd471b6e75fed2fec6a9
        version: 7.11.1+0.1687357785.p
      entries:
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.11.x
    - currentCSV: fuse-apicurito.v7.12.1-0.1706187670.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:7ccefee63165e7eaf9a609e8a9f978a06f27deff69a318fe399a29292ad99106
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.12.1'
          olm.substitutesFor: fuse-apicurito.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:7ccefee63165e7eaf9a609e8a9f978a06f27deff69a318fe399a29292ad99106
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:4cd612283e8136839f1f48a3a220dac7c6c38adf14b61bc03a981e514baefe9b
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:bee0eeaa7898af835f80c5217059d73361daa42197fa6cbbaf416e1ba373eb54
        version: 7.12.1+0.1706187670.p
      entries:
      - name: fuse-apicurito.v7.12.1-0.1706187670.p
        version: 7.12.1+0.1706187670.p
      - name: fuse-apicurito.v7.12.1
        version: 7.12.1
      - name: fuse-apicurito.v7.12.0
        version: 7.12.0
      - name: fuse-apicurito.v7.12.0-0.1697748407.p
        version: 7.12.0+0.1697748407.p
      - name: fuse-apicurito.v7.12.0-0.1697419531.p
        version: 7.12.0+0.1697419531.p
      - name: fuse-apicurito.v7.12.0-0.1694087055.p
        version: 7.12.0+0.1694087055.p
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.12.x
    - currentCSV: fuse-apicurito.v7.13.0-0.1753785079.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:d58d345d74059242d57d5c1a75f31517769230d3014a42243fda20aa93ebd5c8
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.7.0 <7.13.0'
          olm.substitutesFor: fuse-apicurito.v7.13.0-0.1751291353.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:d58d345d74059242d57d5c1a75f31517769230d3014a42243fda20aa93ebd5c8
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:f5bf9240b28b3664bdb11b1f57f611e6194ff8643ae939b6ab3b2b176705f487
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:4e64a5930a10b76147ac128a216cd7c0346da2ff3e9f31ed86311941ed193cc1
        version: 7.13.0+0.1753785079.p
      entries:
      - name: fuse-apicurito.v7.13.0-0.1739885035.p
        version: 7.13.0+0.1739885035.p
      - name: fuse-apicurito.v7.13.0-0.1741884705.p
        version: 7.13.0+0.1741884705.p
      - name: fuse-apicurito.v7.13.0-0.1753785079.p
        version: 7.13.0+0.1753785079.p
      - name: fuse-apicurito.v7.13.0-0.1721224330.p
        version: 7.13.0+0.1721224330.p
      - name: fuse-apicurito.v7.13.0-0.1753778625.p
        version: 7.13.0+0.1753778625.p
      - name: fuse-apicurito.v7.13.0-0.1753702605.p
        version: 7.13.0+0.1753702605.p
      - name: fuse-apicurito.v7.13.0-0.1721607497.p
        version: 7.13.0+0.1721607497.p
      - name: fuse-apicurito.v7.13.0-0.1751291353.p
        version: 7.13.0+0.1751291353.p
      - name: fuse-apicurito.v7.13.0-0.1748960166.p
        version: 7.13.0+0.1748960166.p
      - name: fuse-apicurito.v7.13.0-0.1745523192.p
        version: 7.13.0+0.1745523192.p
      - name: fuse-apicurito.v7.13.0-0.1744737027.p
        version: 7.13.0+0.1744737027.p
      - name: fuse-apicurito.v7.13.0-0.1744280419.p
        version: 7.13.0+0.1744280419.p
      - name: fuse-apicurito.v7.13.0-0.1741875975.p
        version: 7.13.0+0.1741875975.p
      - name: fuse-apicurito.v7.13.0-0.1739473863.p
        version: 7.13.0+0.1739473863.p
      - name: fuse-apicurito.v7.13.0-0.1738056248.p
        version: 7.13.0+0.1738056248.p
      - name: fuse-apicurito.v7.13.0-0.1731460815.p
        version: 7.13.0+0.1731460815.p
      - name: fuse-apicurito.v7.13.0-0.1729189341.p
        version: 7.13.0+0.1729189341.p
      - name: fuse-apicurito.v7.13.0-0.1727957847.p
        version: 7.13.0+0.1727957847.p
      - name: fuse-apicurito.v7.13.0-0.1727193594.p
        version: 7.13.0+0.1727193594.p
      - name: fuse-apicurito.v7.13.0-0.1725022690.p
        version: 7.13.0+0.1725022690.p
      - name: fuse-apicurito.v7.13.0-0.1724659732.p
        version: 7.13.0+0.1724659732.p
      - name: fuse-apicurito.v7.13.0-0.1724053718.p
        version: 7.13.0+0.1724053718.p
      - name: fuse-apicurito.v7.13.0-0.1723709434.p
        version: 7.13.0+0.1723709434.p
      - name: fuse-apicurito.v7.13.0
        version: 7.13.0
      - name: fuse-apicurito.v7.13.0-0.1720707249.p
        version: 7.13.0+0.1720707249.p
      - name: fuse-apicurito.v7.13.0-0.1718285973.p
        version: 7.13.0+0.1718285973.p
      - name: fuse-apicurito.v7.13.0-0.1717672390.p
        version: 7.13.0+0.1717672390.p
      - name: fuse-apicurito.v7.13.0-0.1753778936.p
        version: 7.13.0+0.1753778936.p
      - name: fuse-apicurito.v7.13.0-0.1753701210.p
        version: 7.13.0+0.1753701210.p
      - name: fuse-apicurito.v7.12.1-0.1706187670.p
        version: 7.12.1+0.1706187670.p
      - name: fuse-apicurito.v7.12.1
        version: 7.12.1
      - name: fuse-apicurito.v7.12.0-0.1697419531.p
        version: 7.12.0+0.1697419531.p
      - name: fuse-apicurito.v7.12.0-0.1697748407.p
        version: 7.12.0+0.1697748407.p
      - name: fuse-apicurito.v7.12.0-0.1694087055.p
        version: 7.12.0+0.1694087055.p
      - name: fuse-apicurito.v7.12.0
        version: 7.12.0
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.13.x
    - currentCSV: fuse-apicurito.v7.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
                    "apiVersion": "apicur.io/v1alpha1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
            }]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:05c9676aada2d34a985a59abcf55105a3765d75b0d3764089d2b32e9ca5d367c
          createdAt: "2020-10-06 12:48:47"
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.9.3'
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito Project
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CRD for Apicurito
            displayName: Apicurito CRD
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:05c9676aada2d34a985a59abcf55105a3765d75b0d3764089d2b32e9ca5d367c
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:5535864ce393ae0ae5ee52ed0c7661e0351fb7df2f9235da70608245519530b2
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:dc39b79c360511ff999eba5b2c299db3e08c9c244a0a1f3da8c90c02695e506d
        version: 7.9.3
      entries:
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.9.x
    defaultChannel: fuse-apicurito-7.13.x
    packageName: fuse-apicurito
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app: kube-green
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: kube-green Community
      provider-url: https://kube-green.dev
    name: kube-green
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kube-green.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kube-green.com/v1alpha1",
                "kind": "SleepInfo",
                "metadata": {
                  "labels": {
                    "app": "kube-green"
                  },
                  "name": "sleepinfo-sample"
                },
                "spec": {
                  "excludeRef": [
                    {
                      "apiVersion": "apps/v1",
                      "kind": "Deployment",
                      "name": "api-gateway"
                    }
                  ],
                  "sleepAt": "20:00",
                  "suspendCronJobs": true,
                  "timeZone": "Europe/Rome",
                  "wakeUpAt": "08:00",
                  "weekdays": "1-5"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          containerImage: docker.io/kubegreen/kube-green:0.7.1
          createdAt: "2025-06-29T15:45:17Z"
          description: Suspend your pods when no-one's using them to save energy and
            restore when necessary
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kube-green/kube-green
          support: kube-green Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SleepInfo is the Schema for the sleepinfos API
            displayName: SleepInfo
            kind: SleepInfo
            name: sleepinfos.kube-green.com
            version: v1alpha1
        description: |
          ## About this Operator

          [kube-green](https://kube-green.dev) shuts down (some of) your resources when you don't need them, reducing the consumption of resources of the cluster.

          Install this operator and apply the SleepInfo CRD to stop your dev namespaces from running during weekends, nights or when it is not necessary.
          See the [docs](https://kube-green.dev/docs/getting-started/) for more information.
          This operator is responsible to suspend and restart pods at specific times, defined by the SleepInfo CRD.

          An example of SleepInfo:

          ```yaml
          apiVersion: kube-green.com/v1alpha1
          kind: SleepInfo
          metadata:
            name: working-hours
          spec:
            weekdays: "1-5"
            sleepAt: "20:00"
            wakeUpAt: "08:00"
            timeZone: "Europe/Rome"
          ```

          This configuration stop pods everyday from monday to friday at 20:00 and restore them at 8:00.
          In this way, your cluster nodes could downscale, producing less CO2 and saving money.
        displayName: kube-green
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - green
        - green software
        - resources
        - downscale
        - autoscale
        - climate change
        links:
        - name: Homepage
          url: https://kube-green.dev
        - name: Source Code
          url: https://github.com/kube-green/kube-green
        maintainers:
        - email: bianchidavide12+kubegreen@gmail.com
          name: Davide
        maturity: alpha
        minKubeVersion: 1.27.0
        provider:
          name: kube-green Community
          url: https://kube-green.dev
        relatedImages:
        - docker.io/kubegreen/kube-green:0.7.1
        version: 0.7.1
      entries:
      - name: kube-green.v0.7.1
        version: 0.7.1
      - name: kube-green.v0.7.0
        version: 0.7.0
      - name: kube-green.v0.6.0
        version: 0.6.0
      - name: kube-green.v0.5.2
        version: 0.5.2
      - name: kube-green.v0.5.1
        version: 0.5.1
      - name: kube-green.v0.5.0
        version: 0.5.0
      - name: kube-green.v0.4.1
        version: 0.4.1
      - name: kube-green.v0.4.0
        version: 0.4.0
      - name: kube-green.v0.3.1
        version: 0.3.1
      - name: kube-green.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: kube-green
    provider:
      name: kube-green Community
      url: https://kube-green.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeRocketCI
      provider-url: https://docs.kuberocketci.io
    name: nexus-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-operator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexus-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexus",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexus-sample"
                },
                "spec": {
                  "secret": "nexus-secret",
                  "url": "https://nexus-example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRepository",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrepository-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrepository",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrepository-sample"
                },
                "spec": {
                  "go": {
                    "proxy": {
                      "cleanup": {
                        "policyNames": [
                          "cleanup-policy"
                        ]
                      },
                      "httpClient": {
                        "autoBlock": true,
                        "blocked": true,
                        "connection": {
                          "retries": 1440,
                          "timeout": 1440,
                          "userAgentSuffix": "user-agent-suffix"
                        }
                      },
                      "name": "go-proxy",
                      "negativeCache": {
                        "enabled": true,
                        "timeToLive": 1440
                      },
                      "online": true,
                      "proxy": {
                        "contentMaxAge": 1440,
                        "metadataMaxAge": 1440,
                        "remoteUrl": "https://test.com"
                      },
                      "routingRule": "routing-rule",
                      "storage": {
                        "blobStoreName": "blob-store-name",
                        "strictContentTypeValidation": true
                      }
                    }
                  },
                  "nexusRef": {
                    "name": "nexus-sample"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrole",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrole-sample"
                },
                "spec": {
                  "description": "test-role",
                  "id": "test-role",
                  "name": "test-role",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "privileges": [
                    "nx-blobstores-all"
                  ]
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexususer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexususer",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexususer-sample"
                },
                "spec": {
                  "email": "test@gmail.com",
                  "firstName": "John1",
                  "id": "test-user",
                  "lastName": "Doe",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "roles": [
                    "nx-admin"
                  ],
                  "secret": "$user-secret:password"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/nexus-operator:3.0.0
          createdAt: "2023-12-11T12:54:50Z"
          description: An Operator for managing Nexus entities.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-nexus-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus is the Schema for the nexus API.
            displayName: Nexus
            kind: Nexus
            name: nexuses.edp.epam.com
            version: v1alpha1
          - description: NexusRepository is the Schema for the nexusrepositories API.
            displayName: Nexus Repository
            kind: NexusRepository
            name: nexusrepositories.edp.epam.com
            version: v1alpha1
          - description: NexusRole is the Schema for the nexusroles API.
            displayName: Nexus Role
            kind: NexusRole
            name: nexusroles.edp.epam.com
            version: v1alpha1
          - description: NexusUser is the Schema for the nexususers API.
            displayName: Nexus User
            kind: NexusUser
            name: nexususers.edp.epam.com
            version: v1alpha1
        description: |
          Nexus Operator is an operator that is responsible for establishing
          a connection to provided NexusRepository Server, reconciling Nexus entities (groups,
          users, repositories, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Nexus Repository.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  nexus-secret
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Nexus` with Nexus instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1alpha1
              kind: Nexus
              metadata:
                name: nexus-sample
              spec:
                secret: nexus-secret                # Secret name
                url: https://nexus.example.com      # Nexus URL
              ```

              Wait for the `.status` field with  `status.connected: true`
        displayName: EDP Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - edp
        - nexus
        - operator
        - artifact
        - storage
        links:
        - name: Nexus Operator
          url: https://github.com/epam/edp-nexus-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/nexus-operator:3.0.0
        version: 3.0.0
      entries:
      - name: nexus-operator.v3.0.0
        version: 3.0.0
      name: alpha
    - currentCSV: nexus-operator.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexus-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexus",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexus-sample"
                },
                "spec": {
                  "secret": "nexus-secret",
                  "url": "https://nexus-example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusBlobStore",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusblobstore-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusblobstore",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusblobstore-sample"
                },
                "spec": {
                  "file": {
                    "path": "storage-blobstore/nexusblobstore-sample"
                  },
                  "name": "nexusblobstore-sample",
                  "nexusRef": {
                    "name": "nexus-sample"
                  },
                  "softQuota": {
                    "limit": 1000,
                    "type": "spaceUsedQuota"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusCleanupPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexuscleanuppolicy-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexuscleanuppolicy",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexuscleanuppolicy-sample"
                },
                "spec": {
                  "criteria": {
                    "lastBlobUpdated": 30
                  },
                  "description": "Cleanup policy for go",
                  "format": "go",
                  "name": "go-cleanup-policy",
                  "nexusRef": {
                    "name": "nexus-sample"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRepository",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrepository-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrepository",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrepository-sample"
                },
                "spec": {
                  "go": {
                    "proxy": {
                      "cleanup": {
                        "policyNames": [
                          "cleanup-policy"
                        ]
                      },
                      "httpClient": {
                        "autoBlock": true,
                        "blocked": true,
                        "connection": {
                          "retries": 1440,
                          "timeout": 1440,
                          "userAgentSuffix": "user-agent-suffix"
                        }
                      },
                      "name": "go-proxy",
                      "negativeCache": {
                        "enabled": true,
                        "timeToLive": 1440
                      },
                      "online": true,
                      "proxy": {
                        "contentMaxAge": 1440,
                        "metadataMaxAge": 1440,
                        "remoteUrl": "https://test.com"
                      },
                      "routingRule": "routing-rule",
                      "storage": {
                        "blobStoreName": "blob-store-name",
                        "strictContentTypeValidation": true
                      }
                    }
                  },
                  "nexusRef": {
                    "name": "nexus-sample"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrole",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrole-sample"
                },
                "spec": {
                  "description": "test-role",
                  "id": "test-role",
                  "name": "test-role",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "privileges": [
                    "nx-blobstores-all"
                  ]
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusScript",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusscript-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusscript",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusscript-sample"
                },
                "spec": {
                  "content": "security.setAnonymousAccess(Boolean.valueOf(args))\n",
                  "name": "anonymous",
                  "nexusRef": {
                    "name": "nexus-sample"
                  },
                  "payload": "true"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexususer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexususer",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexususer-sample"
                },
                "spec": {
                  "email": "test@gmail.com",
                  "firstName": "John1",
                  "id": "test-user",
                  "lastName": "Doe",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "roles": [
                    "nx-admin"
                  ],
                  "secret": "$user-secret:password"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/nexus-operator:3.5.0
          createdAt: "2025-04-22T11:21:24Z"
          description: An Operator for managing Nexus entities.
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/epam/edp-nexus-operator
          support: KubeRocketCI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NexusBlobStore is the Schema for the nexusblobstores API.
            displayName: Nexus Blob Store
            kind: NexusBlobStore
            name: nexusblobstores.edp.epam.com
            version: v1alpha1
          - description: NexusCleanupPolicy is the Schema for the cleanuppolicies
              API.
            displayName: Nexus Cleanup Policy
            kind: NexusCleanupPolicy
            name: nexuscleanuppolicies.edp.epam.com
            version: v1alpha1
          - description: Nexus is the Schema for the nexus API.
            displayName: Nexus
            kind: Nexus
            name: nexuses.edp.epam.com
            version: v1alpha1
          - description: NexusRepository is the Schema for the nexusrepositories API.
            displayName: Nexus Repository
            kind: NexusRepository
            name: nexusrepositories.edp.epam.com
            version: v1alpha1
          - description: NexusRole is the Schema for the nexusroles API.
            displayName: Nexus Role
            kind: NexusRole
            name: nexusroles.edp.epam.com
            version: v1alpha1
          - description: NexusScript is the Schema for the nexusscripts API.
            displayName: Nexus Script
            kind: NexusScript
            name: nexusscripts.edp.epam.com
            version: v1alpha1
          - description: NexusUser is the Schema for the nexususers API.
            displayName: Nexus User
            kind: NexusUser
            name: nexususers.edp.epam.com
            version: v1alpha1
        description: |
          Nexus Operator is an operator that is responsible for establishing
          a connection to provided NexusRepository Server, reconciling Nexus entities (groups,
          users, repositories, etc) according to the created CRs.

          ## Quick Start

          1. Login into Nexus and create user. Attach permissions to user such as scripts, rules, blobs etc. Insert user credentials into Kubernetes secret.

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name: nexus-admin-password
              data:
                password: cGFzcw==  # base64-encoded value of "pass"
                user:     dXNlcg==  # base64-encoded value of "user"
              ```

          2. Create Custom Resource `kind: Nexus` with Nexus instance URL and secret created on the previous step:

              ```yaml
              apiVersion: edp.epam.com/v1alpha1
              kind: Nexus
              metadata:
                name: nexus
              spec:
                secret: nexus-admin-password
                url: http://nexus.example.com
              ```

              Wait for the `.status` field with  `status.connected: true`

          3. Create Role using Custom Resources NexusRole:

              ```yaml
              apiVersion: edp.epam.com/v1alpha1
              kind: NexusRole
              metadata:
                name: edp-admin
              spec:
                description: Read and write access to all repos and scripts
                id: edp-admin
                name: edp-admin
                nexusRef:
                  kind: Nexus
                  name: nexus
                privileges:
                  - nx-apikey-all
                  - nx-repository-view-*-*-add
                  - nx-repository-view-*-*-browse
                  - nx-repository-view-*-*-edit
                  - nx-repository-view-*-*-read
                  - nx-script-*-add
                  - nx-script-*-delete
                  - nx-script-*-run
                  - nx-search-read
              ```
        displayName: EDP Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - artifact
        - edp
        - kuberocketci
        - nexus
        - operator
        - storage
        links:
        - name: Nexus Operator
          url: https://github.com/epam/edp-nexus-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: KubeRocketCI
          url: https://docs.kuberocketci.io
        relatedImages:
        - docker.io/epamedp/nexus-operator:3.5.0
        version: 3.5.0
      entries:
      - name: nexus-operator.v3.5.0
        version: 3.5.0
      - name: nexus-operator.v3.4.0
        version: 3.4.0
      - name: nexus-operator.v3.3.0
        version: 3.3.0
      name: stable
    defaultChannel: stable
    packageName: nexus-operator
    provider:
      name: KubeRocketCI
      url: https://docs.kuberocketci.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp, Inc.
      provider-url: https://www.netapp.com/
    name: trident-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trident-operator.v25.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "trident.netapp.io/v1",
                "kind": "TridentOrchestrator",
                "metadata": {
                  "name": "trident"
                },
                "spec": {
                  "IPv6": false,
                  "debug": true,
                  "nodePrep": [],
                  "imagePullSecrets": [],
                  "imageRegistry": "",
                  "namespace": "trident",
                  "silenceAutosupport": false
                }
              }
            ]
          categories: Storage
          containerImage: docker.io/netapp/trident-operator@sha256:c3448b6847550f45849a5cdbb9e0cedb35290d21131bc513d833d95eb42a43e0
          createdAt: "2025-09-03"
          description: Trident Operator, to manage NetApp Trident installations
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          operators.openshift.io/valid-subscription: No subscriptions are required
          repository: https://github.com/netapp/trident
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Used to deploy NetApp Trident.
            displayName: Trident Orchestrator
            kind: TridentOrchestrator
            name: tridentorchestrators.trident.netapp.io
            version: v1
          - description: Automates AWS FSxN backend configuration
            displayName: Trident Configurator
            kind: TridentConfigurator
            name: tridentconfigurators.trident.netapp.io
            version: v1
        description: |
          NetApp Trident is an open source storage provisioner and orchestrator maintained by NetApp. It enables you to create storage volumes for containerized applications managed by Docker and Kubernetes. For full release information, including patch release changes, see https://docs.netapp.com/us-en/trident/trident-rn.html.
        displayName: NetApp Trident
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - NetApp
        - ONTAP
        - Trident
        - storage
        - performance
        - open source
        - Astra
        - Cloud Native
        - AWS
        - Azure
        - GCP
        - Cloud Volumes
        links:
        - name: GitHub Repository
          url: https://github.com/NetApp/trident
        - name: Trident documentation
          url: https://docs.netapp.com/us-en/trident/index.html
        - name: Support policy
          url: https://mysupport.netapp.com/site/info/version-support
        - name: Release Notes
          url: https://docs.netapp.com/us-en/trident/trident-rn.html
        maintainers:
        - email: ng-trident-openshift-operator@netapp.com
          name: NetApp
        provider:
          name: NetApp, Inc.
          url: https://www.netapp.com/
        relatedImages:
        - docker.io/netapp/trident-operator@sha256:c3448b6847550f45849a5cdbb9e0cedb35290d21131bc513d833d95eb42a43e0
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:b88958c38883f5b1e3d307da90aac34a5ceafdb6bd3d989b062cb24ecafd1fbd
        - docker.io/netapp/trident-autosupport@sha256:21b3afbebfac3cc28c4eb319c2d385187bd8ebb728d8cfd82dbc437cc9b83b33
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:bdee8df520467b30312e7b14d7b6ef3963557ccd38269913de3fa51bb1bceec9
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:145ab76f9606e9d51dcd1e4e159c67021ecf53c24d8d80ed7bbeb2cc7fb5c42b
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6781d64bfb01e252595130eec814b9d2fe6426720a003cc61e82e9963767bfd4
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:b48bd5bc612c1bbe73d203c70e6caaae2b5360bd31eba20eb06cbd2102cc34fc
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:0ef3bf2a4dcbb5aa829ce9cc5f12efaccaf99b246b4b15abbb66b0135bec4cab
        - docker.io/netapp/trident@sha256:7925717979a84f93b62a2592682cde3443d6350ec4e88f8fd1a7d3ea812e8900
        version: 25.6.2
      entries:
      - name: trident-operator.v25.6.2
        version: 25.6.2
      - name: trident-operator.v25.6.1
        version: 25.6.1
      - name: trident-operator.v25.6.0
        version: 25.6.0
      - name: trident-operator.v25.2.1
        version: 25.2.1
      - name: trident-operator.v25.2.0
        version: 25.2.0
      name: stable
    defaultChannel: stable
    packageName: trident-operator
    provider:
      name: NetApp, Inc.
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: xspc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: xspc-operator.v1.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.hpe.com/v1",
                "kind": "XSPC",
                "metadata": {
                  "name": "xspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.connect.redhat.com/hpestorage/xspc-csi-driver-operator@sha256:8b8724998a8301057a3c4187c7b94bf4484aeb807065a12f84142cf844ea5534
          createdAt: "2024-12-17T03:53:30Z"
          description: An operator for managing HPE XP Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XSPC is the Schema for the xspcs API
            displayName: XSPC
            kind: XSPC
            name: xspcs.csi.hpe.com
            version: v1
        description: |-
          ## About
          HPE XP Storage Plug-in for Containers (XSPC) is a plugin that integrates HPE XP storage into Kubernetes based clusters.
          XSPC provides persistent volume provisioning capabilities from HPE's block storage arrays.
        displayName: HPE XP Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - XP7
        - XP8
        - HPE
        - Storage
        - Container Storage Interface
        - CSI
        links:
        - name: Knowledge Base
          url: https://www.hpe.com/us/en/storage/enterprise-xp-storage.html
        maintainers:
        - email: hpexpci@hpe.com
          name: HPE
        maturity: stable
        provider:
          name: HPE
        relatedImages:
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.connect.redhat.com/hpestorage/xspc-csi-driver@sha256:bdeac71c0b54c2779910a8d8faea03c1122000127704f17ceadba4938b685353
        - registry.connect.redhat.com/hpestorage/xspc-csi-driver-operator@sha256:8b8724998a8301057a3c4187c7b94bf4484aeb807065a12f84142cf844ea5534
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        version: 1.15.0
      entries:
      - name: xspc-operator.v1.15.0
        version: 1.15.0
      name: stable
    defaultChannel: stable
    packageName: xspc-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Entando, Inc
      provider-url: ""
    name: entando-k8s-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: entando-k8s-operator.v6.3.2-pr3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: beta
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - docker.io/entando/entando-k8s-service@sha256:775391f0005e64764288bfe007d9ea576f13618c02ea400d030a7502cfb551bd
        - docker.io/entando/entando-k8s-database-service-controller@sha256:58ca4ad2fca83a715aba9b68609f9056f633d556eb9845b3aff5c05769369539
        - docker.io/entando/app-builder@sha256:1ca90c09302b9cadf8950278d92ad4d7f3a14f02dea8704320794aa77ada8631
        - docker.io/entando/entando-k8s-app-controller@sha256:c68d3b129fc625f5d1ee10ed2a107e28c145049d4b513dcdb2c736555025e339
        - docker.io/entando/entando-k8s-plugin-controller@sha256:a20c753dda86e0f54f210e446b404733045ba426fc7186dc3936daed8224fbcb
        - docker.io/entando/entando-component-manager@sha256:668191fc86b090452baaba18ec7bbc3d07df5a99750e0c5b98759599d565cfbe
        - docker.io/entando/entando-redhat-sso@sha256:3692adb3694b3bf278bed7a678c9d0eaa0aba8e3e3ed4117ca04156525bcac72
        - registry.redhat.io/rhel8/mysql-80@sha256:570bce7dc649ac7ac85e7ecb196aabc281c77a94590aa7df7d9746869ce757ea
        - docker.io/entando/entando-k8s-dbjob@sha256:8ae47dbf93201965106663299f06daad544982fa244d4a0f365a1b717727e5ed
        - docker.io/entando/entando-k8s-composite-app-controller@sha256:4bc6797bb41b3e407b64530c3cff7a78f3fc2c22863caad9c606ccd650310b61
        - docker.io/entando/entando-k8s-app-plugin-link-controller@sha256:aa7ba0861705213434e610530c5bf5da1bdb33f1834d3c346174f5a28145c0bd
        - docker.io/entando/entando-de-app-eap@sha256:1f0e35d7f7ee8123721cf8e80ac96ac2f594f21a2b68cf9e720917d090c8095e
        - docker.io/entando/entando-k8s-keycloak-controller@sha256:f1ed42001260802353197433cea7df4c2d0e5fd9bcd926bb5ae7c654dfad2ffe
        - registry.redhat.io/rhel8/postgresql-12@sha256:f4e5c728b644bf1888ec8086424852ed74b5596a511be29e636fb10218fc9b6f
        - entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        - docker.io/entando/entando-k8s-cluster-infrastructure-controller@sha256:86dbea6696349d7b2393e12bacef4d1acec2dd3f8551e397fa26c92706a18dd3
        version: 6.3.2-pr3
      entries:
      - name: entando-k8s-operator.v6.3.2-pr3
        version: 6.3.2-pr3
      name: alpha
    - currentCSV: entando-k8s-operator.v6.3.2-pr3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: beta
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - docker.io/entando/entando-redhat-sso@sha256:3692adb3694b3bf278bed7a678c9d0eaa0aba8e3e3ed4117ca04156525bcac72
        - entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        - docker.io/entando/entando-de-app-eap@sha256:1f0e35d7f7ee8123721cf8e80ac96ac2f594f21a2b68cf9e720917d090c8095e
        - docker.io/entando/entando-k8s-composite-app-controller@sha256:4bc6797bb41b3e407b64530c3cff7a78f3fc2c22863caad9c606ccd650310b61
        - docker.io/entando/entando-k8s-app-controller@sha256:c68d3b129fc625f5d1ee10ed2a107e28c145049d4b513dcdb2c736555025e339
        - docker.io/entando/entando-k8s-plugin-controller@sha256:a20c753dda86e0f54f210e446b404733045ba426fc7186dc3936daed8224fbcb
        - docker.io/entando/entando-component-manager@sha256:668191fc86b090452baaba18ec7bbc3d07df5a99750e0c5b98759599d565cfbe
        - docker.io/entando/entando-k8s-app-plugin-link-controller@sha256:aa7ba0861705213434e610530c5bf5da1bdb33f1834d3c346174f5a28145c0bd
        - docker.io/entando/entando-k8s-database-service-controller@sha256:58ca4ad2fca83a715aba9b68609f9056f633d556eb9845b3aff5c05769369539
        - docker.io/entando/app-builder@sha256:1ca90c09302b9cadf8950278d92ad4d7f3a14f02dea8704320794aa77ada8631
        - docker.io/entando/entando-k8s-keycloak-controller@sha256:f1ed42001260802353197433cea7df4c2d0e5fd9bcd926bb5ae7c654dfad2ffe
        - registry.redhat.io/rhel8/mysql-80@sha256:570bce7dc649ac7ac85e7ecb196aabc281c77a94590aa7df7d9746869ce757ea
        - registry.redhat.io/rhel8/postgresql-12@sha256:f4e5c728b644bf1888ec8086424852ed74b5596a511be29e636fb10218fc9b6f
        - docker.io/entando/entando-k8s-cluster-infrastructure-controller@sha256:86dbea6696349d7b2393e12bacef4d1acec2dd3f8551e397fa26c92706a18dd3
        - docker.io/entando/entando-k8s-dbjob@sha256:8ae47dbf93201965106663299f06daad544982fa244d4a0f365a1b717727e5ed
        - docker.io/entando/entando-k8s-service@sha256:775391f0005e64764288bfe007d9ea576f13618c02ea400d030a7502cfb551bd
        version: 6.3.2-pr3
      entries:
      - name: entando-k8s-operator.v6.3.2-pr3
        version: 6.3.2-pr3
      name: beta
    - currentCSV: entando-k8s-operator.v6.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: registry.hub.docker.com/entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - registry.hub.docker.com/entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        version: 6.3.2
      entries:
      - name: entando-k8s-operator.v6.3.2
        version: 6.3.2
      name: stable
    defaultChannel: stable
    packageName: entando-k8s-operator-rhmp
    provider:
      name: Entando, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: submariner.io
      provider-url: ""
    name: submariner
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: submariner.v0.17.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:067ba6e58a6fe150a212881a29c90aa7cbd22f53a6877eebc413072f73f9179a
          createdAt: "2025-06-18 01:35:57"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.16.0 <0.17.6'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.17 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:d58214791a60cd9e4366c6ff534584eedbd37386da71b868867e5e212df270f9
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:a565d8c094f3755b6c9303c5b8ae03a5883796e0fea754b0202e5792f1e472c3
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:067ba6e58a6fe150a212881a29c90aa7cbd22f53a6877eebc413072f73f9179a
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:b76998e8bbef1d06aa8dbc49af4f42a3c5603d480a67e46baf11168514c7759a
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:56848aefe1d156147ab73e7353017b2663f63e2e01fe43ede57abc72cfc0c50f
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:e6dbc4bf3e8b1a3c1a7ed10eb463d134f9ed28a03c7bb523c82c5431c5c9ae90
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:5510cb1bb3e0ca5f2365f4a5fc147ca693d3c220f1a8e7835f8b13b66b957ae8
        version: 0.17.6
      entries:
      - name: submariner.v0.17.6
        version: 0.17.6
      - name: submariner.v0.17.5
        version: 0.17.5
      - name: submariner.v0.17.4
        version: 0.17.4
      - name: submariner.v0.17.2
        version: 0.17.2
      - name: submariner.v0.17.2-0.1725901206.p
        version: 0.17.2+0.1725901206.p
      - name: submariner.v0.17.1
        version: 0.17.1
      - name: submariner.v0.17.0
        version: 0.17.0
      name: stable-0.17
    - currentCSV: submariner.v0.18.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:d8d7f0147c7b65b41903e101240905703992a7bbadb33e302aa86b23000d2095
          createdAt: "2025-05-07 20:54:56"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.17.0 <0.18.5'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.18 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:e5c16f0d720e15d54781870a5e63c85540256355009584f7d1bd03ef20a2eae4
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:00072fea3fdb5e294a01da0ee44165fc52219a965a27bdf7d792276f6c3a36c3
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:e8366433d1f50e2062b86d7f749f17cd6566197560256e77f57b2c165f277733
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:a07ef1bfec7549b32ad94f2c44a168860d5f3b46fd13593c068c58bb137c77fc
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:d8d7f0147c7b65b41903e101240905703992a7bbadb33e302aa86b23000d2095
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:27f18015b8896b04457c356d03276d50c34b8c1d84142ec07342ff2c1c4e81fa
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:e913be7a440b6cc0c1dddbda93a5979fe28b99e63f12a7949966b69126fcee90
        version: 0.18.5
      entries:
      - name: submariner.v0.18.5
        version: 0.18.5
      - name: submariner.v0.18.4
        version: 0.18.4
      - name: submariner.v0.18.3
        version: 0.18.3
      - name: submariner.v0.18.2
        version: 0.18.2
      - name: submariner.v0.18.0
        version: 0.18.0
      name: stable-0.18
    - currentCSV: submariner.v0.19.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:4afaf6689e2225c5c7f6fbe002202a1cdf9444e23fd210cc2d86da71c9f78626
          createdAt: "2025-04-22 21:12:45"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.18.0 <0.19.4'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.19 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:d9d4d03013d65d11a4e8ecda411015f359e4215a3a53059a37a810acf2e3ea6c
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:4afaf6689e2225c5c7f6fbe002202a1cdf9444e23fd210cc2d86da71c9f78626
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:5b83c5889aa0d8f7838dd5afb8c901dfb0ce0fbe01175e8468ec50ddedef9223
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:70008af49a9d1ee516204715f66bd415c9caa7462f6b178dc1dc91e513605fcd
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:ef73b8964cd66ffb496191056594e0886afa9d17c03976b4159413c89ed569bf
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:5938d67a0a2653e92f0995440700512992ed82269820679f68483dcc1b2d98d9
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:1bbb624bbefbb75407071449d0144258732e219e3a7fc67ac709c3bcee560576
        version: 0.19.4
      entries:
      - name: submariner.v0.19.4
        version: 0.19.4
      - name: submariner.v0.19.3
        version: 0.19.3
      - name: submariner.v0.19.2
        version: 0.19.2
      - name: submariner.v0.19.0
        version: 0.19.0
      name: stable-0.19
    - currentCSV: submariner.v0.20.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "quay.io/submariner",
                  "version": "0.20.2"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "quay.io/submariner",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "0.20.2"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: $(SUBMARINER_OPERATOR_IMAGE)
          createdAt: "2025-09-30 14:21:54"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/submariner-io/submariner-operator
          support: Submariner
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.20 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: submariner.io
        relatedImages:
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:fd6c46b4a0570b5e7b6457c465e5a1b843ba6f9c209c1856dbebfee1068e35ea
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:70d563a6e6526b9a146a2973210b2abdae09d5acab403fa3f0b465ed5e3c9b11
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:77e0da2db6d5bb2831e6c51214e08685c9f7f3b8b1e6ac53b4fd1418644ad6a1
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:0eef6521c96b29722076c1e220770124c8ed214379abc4654d9567cff8b52d04
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:8ea02aba8f9dd901f1f0ccff97ec8d157bebe5a90a6c924affe5e01cf3454e72
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:ed521b241da60efabfe2e2b80c51639d1a82a43945945879decbfb192bbc6d23
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:4a4ac198e9d0f4a4aaed8f0ec0e1d79bd8fa712b671ee697f7ff9d5616a02d84
        version: 0.20.2
      entries:
      - name: submariner.v0.20.2
        version: 0.20.2
      - name: submariner.v0.20.1
        version: 0.20.1
      - name: submariner.v0.20.0
        version: 0.20.0
      name: stable-0.20
    - currentCSV: submariner.v0.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "quay.io/submariner",
                  "version": "v0.21.0"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "quay.io/submariner",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v0.21.0"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: $(SUBMARINER_OPERATOR_IMAGE)
          createdAt: "2025-08-13 21:57:14"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/submariner-io/submariner-operator
          support: Submariner
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.21 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: submariner.io
        relatedImages:
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:cd851eab4a2214ea7a699bd01d2abac1322d01562f27b35f58f3c0b8716f88fb
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:bcdeaa67664568c5e2a6928236ff69b175ce65ddf7a58c1664d091b926000052
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:b0ad338aed82367048cf024767b0b09e6bb09bbf7dd5128e6711e1feec6f696e
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:65765c7f08c5454b557fbda91c49efbd2cc0d4d973cf8acc7ff40ec8a381839d
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:f5909b1802986622ab169a82d5dd7877af72660903e961b3097cc29e30eb9799
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:b307f748125fd459c3474fdab645688675a06515a9924b47cedba80e038458db
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:0b74f058400f24a63cf9b57a388914cc6dda38e03d639196871f3fd76fb7f2ae
        version: 0.21.0
      entries:
      - name: submariner.v0.21.0
        version: 0.21.0
      name: stable-0.21
    defaultChannel: stable-0.21
    packageName: submariner
    provider:
      name: submariner.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microcks.io
      provider-url: ""
    name: microcks
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: microcks-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "microcks.github.io/v1alpha1",
                "kind": "MicrocksInstall",
                "metadata": {
                  "name": "my-microcksinstall"
                },
                "spec": {
                  "name": "my-microcksinstall",
                  "version": "1.10.0",
                  "microcks": {
                    "replicas": 1
                  },
                  "postman": {
                    "replicas": 1
                  },
                  "keycloak": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "1Gi"
                  },
                  "mongodb": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "2Gi"
                  }
                }
              },
              {
                "apiVersion": "microcks.github.io/v1alpha1",
                "kind": "MicrocksInstall",
                "metadata": {
                  "name": "my-microcksinstall-minikube"
                },
                "spec": {
                  "name": "my-microcksinstall-minikube",
                  "version": "1.10.0",
                  "microcks": {
                    "replicas": 2,
                    "url": "microcks.192.168.99.100.nip.io"
                  },
                  "postman": {
                    "replicas": 2
                  },
                  "keycloak": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "1Gi",
                    "url": "keycloak.192.168.99.100.nip.io"
                  },
                  "mongodb": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "2Gi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/microcks/microcks-ansible-operator:1.10.0
          createdAt: "2024-08-01T09:14:55Z"
          description: Open Source mocking and testing platform for API and microservices
          repository: https://github.com/microcks/microcks-ansible-operator
          support: microcks.github.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Microcks installation
            displayName: MicrocksInstall
            kind: MicrocksInstall
            name: microcksinstalls.microcks.github.io
            version: v1alpha1
        description: |
          Microcks is an Open source tool for API and microservices mocking and testing. It provides a platform for referencing, deploying mocks and allow contract-testing of your APIs. It can also be considered as a Service Virtualization solution because it will allow you to provide fake API or Service implementation before development being actually done. It supports Async APIs, REST APIs, gRPC, GraphQL and SOAP WebServices. It perfectly integrates into an iterative, contract-first delivery process.
          Microcks tries not reinventing the wheel and let you capitalize on already known tools of standards. It supports [AsyncAPI Specifications](https://asyncapi.com), [OpenAPI Specification](https://www.openapis.org/) `3.x`, [gRPC Protobuf](https://grpc.io/), [GraphQL Schema](https://graphql.org) out-of-the-box and can also relies on popular tools like [SoapUI](http://www.soapui.com) or [Postman](http://www.getpostman.io) to edit your API request / response samples as well as your test scripts.
          ### Operator features
          * **Create/Destroy**: Easily launch Microcks installations (`MicrocksInstall` custom resource) for your Kubernetes namespace
          ### MicrocksInstall options
          * **keycloak** component can be deployed as part of the install or you can reuse an existing instance. You can reference an existing `Secret` for providing TLS certificates for securing the `Ingress`,
          * **mongodb** component can be deployed as part of the install or you can reuse an existing instance, You can reference an existing `Secret` for providing connection credentials,
          * **features** component allow to activate optional features like the mocking of AsyncAPIs. In case you turn `features.async.enabled` to `true` and `features.async.kafka.install` to `true`, you must have the [Strimzi Operator](https://operatorhub.io/operator/strimzi-kafka-operator) already installed on your cluster/namespace.
          ### Documentation
          Documentation to the current _master_ branch as well as all releases can be found on our [GitHub](https://github.com/microcks/microcks-ansible-operator).
          ### License
          Microcks is licensed under the [Apache License, Version 2.0](https://github.com/microcks/microcks/blob/master/LICENSE).
        displayName: Microcks Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mock
        - test
        - API
        - microservice
        links:
        - name: Product Page
          url: https://microcks.io
        - name: Microcks Operator Source Code
          url: https://github.com/microcks/microcks-ansible-operator
        maintainers:
        - email: laurent.broudoux@gmail.com
          name: lbroudoux
        maturity: alpha
        provider:
          name: Microcks.io
        relatedImages:
        - quay.io/microcks/microcks-ansible-operator:1.10.0
        version: 1.10.0
      entries:
      - name: microcks-operator.v1.10.0
        version: 1.10.0
      - name: microcks-operator.v1.9.0
        version: 1.9.0
      - name: microcks-operator.v1.8.1
        version: 1.8.1
      - name: microcks-operator.v1.8.0
        version: 1.8.0
      - name: microcks-operator.v1.7.1
        version: 1.7.1
      - name: microcks-operator.v1.7.0
        version: 1.7.0
      - name: microcks-operator.v1.6.1
        version: 1.6.1
      - name: microcks-operator.v1.6.0
        version: 1.6.0
      - name: microcks-operator.v1.5.2
        version: 1.5.2
      - name: microcks-operator.v1.5.1
        version: 1.5.1
      - name: microcks-operator.v1.5.0
        version: 1.5.0
      - name: microcks-operator.v1.4.1
        version: 1.4.1
      name: stable
    defaultChannel: stable
    packageName: microcks
    provider:
      name: Microcks.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-wafv2-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-wafv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wafv2.services.k8s.aws/v1alpha1",
                "kind": "IPSet",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/wafv2-controller:1.1.1
          createdAt: "2025-09-20T21:41:21Z"
          description: AWS WAFV2 controller is a service controller for managing WAFV2
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IPSet represents the state of an AWS wafv2 IPSet resource.
            displayName: IPSet
            kind: IPSet
            name: ipsets.wafv2.services.k8s.aws
            version: v1alpha1
          - description: RuleGroup represents the state of an AWS wafv2 RuleGroup
              resource.
            displayName: RuleGroup
            kind: RuleGroup
            name: rulegroups.wafv2.services.k8s.aws
            version: v1alpha1
          - description: WebACL represents the state of an AWS wafv2 WebACL resource.
            displayName: WebACL
            kind: WebACL
            name: webacls.wafv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Web Application Firewall (Amazon WAFV2) resources in AWS from within your Kubernetes cluster.

          **About Amazon WAFV2**

          AWS WAF is a web application firewall that lets you monitor the HTTP and HTTPS requests that are forwarded to your protected web application resources. You can protect the following resource types:
          - Amazon CloudFront distribution
          - Amazon API Gateway REST API
          - Application Load Balancer
          - AWS AppSync GraphQL API
          - Amazon Cognito user pool
          - AWS App Runner service
          - AWS Verified Access instance

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon WAFV2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - wafv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon WAFV2 Developer Resources
          url: https://aws.amazon.com/waf/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: wafv2 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/wafv2-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-wafv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-wafv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-wafv2-controller.v1.0.14
        version: 1.0.14
      - name: ack-wafv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-wafv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-wafv2-controller.v1.0.10
        version: 1.0.10
      - name: ack-wafv2-controller.v1.0.9
        version: 1.0.9
      - name: ack-wafv2-controller.v1.0.8
        version: 1.0.8
      - name: ack-wafv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-wafv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-wafv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-wafv2-controller.v1.0.1
        version: 1.0.1
      - name: ack-wafv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-wafv2-controller.v0.0.4
        version: 0.0.4
      - name: ack-wafv2-controller.v0.0.3
        version: 0.0.3
      - name: ack-wafv2-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-wafv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/in-en/products/cloud-pak-for-data
    name: ibm-zen-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-zen-operator.v6.1.3
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.3 <6.1.3'
          operators.operatorframework.io/builder: operator-sdk-v1.6.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: represents zen instance
            kind: ZenService
            name: zenservices.zen.cpd.ibm.com
            version: v1
          - description: represents zen extension instance
            kind: ZenExtension
            name: zenextensions.zen.cpd.ibm.com
            version: v1
        description: |
          The Zen operator is a component in IBM® Cloud Pak foundational services. It serves a framework for user access management, defining user roles and for powering persona driven user experiences for Cloud Paks and Service offerings
        displayName: IBM Zen Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cloud Pak
        - IBM
        - Access management
        - Platform UI
        - Zen framework
        maintainers:
        - email: cp4d@ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/in-en/products/cloud-pak-for-data
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:443b51f8b10acc85bdefde7193e0f45b1c423fa7fbdcaa28342805815c43db3d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:3d5c6573faf4ae00cbb21cbfe9b2ad5cd8ffb658df0c5ba09bd98b3bf54f9979
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:fa01c7d11c23a47bb84375389e69619455c2879b5ed6bb5b9b753b945dcb02c3
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:45194b56f9fcb3af83158eb0297248364db40e8164fb4fe4f5656c58aeac3fa5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:a99b1f5a4eb0306b83562de0184fc826c34835ffbf46d929ceb9e70e0d325361
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:5f48081ffa453ec740c2562c24603d78b481b5e236d6bcc73e96203cabe153f2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:0522fc5482404fa9be179c7225a5692c8d71d755fda0b2d3928e452adcf030fb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:45ca5370b726e31c55f9d7d249374189cb48dc05ab89d675c35225d53b90e7c9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:aaee5240471872a8f1a77d4440a696d45041c5ce06849d516daccb90f5341b6d
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:abd96661823553d623eb0ea1fac8271d64391243dd0f13eb4a4fcc858dfe10a5
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:82afac013ef9b6b9d0eacb4575c979b65abd6f2821971182b90c71bd488ef366
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:b87fa8374976ccd3dad90ddbf6b65cd6ae5a028ba896562f284f089380ebe44a
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:24dff64a80b910f0b4c5442d4f611f9df7602a16c5231eff2c9d08f07b2a9f9a
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:ad027c91c719fd1d931cbac8c76735474f676b228f9ab23d1f243201246b3238
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:c6862158a2a5ae0a4e4206bfdc7f73e1c2359544966c72cf4fc3a617acb13697
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:10ede3c4e3c7e0cbc26e7b6e797cad35e39face1a1b9d4a4e2440ab0feeeb814
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:deb66a64288a370c5da5c7cc0ba4e2ba4ab816d7c5a080a0c505b05f55fa6072
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:7dbff355db7739152961cb6a97887d2e43bc960ac58837c126e03ed1a4480a3a
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:a72c2b70a9f0a93047d47247cddc1616138cbc87337289ddb208565af6e4369d
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:183ee860bf48076ea155db08032eed7004c1204a363625f54218734b00175441
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:525924934f74b28783dacf630a6401ccd4c88d253aafa2d312b7993d8b9a14c1
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:a7c050ae29cb93754f59d8f83676275ee9a91ae9a9956395b97abf8c5e20e96e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8da8b240176c00f8f6eadf32e5614c663cc1d56d546c73996cca614b715acf4a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:f06a1f50cd18d6f48762fdd0394307b44c0573d0cc91f53724c5b3c8c5e2b8e4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:c4ff5dae4560e5a943c7218df3a5aa9ca1a9ee2d1f7dfac73db05892725c34a9
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-bundle@sha256:552fc53664722903eccaa7d3782a3e84413509fd357725fe0b0f4a077c48ab2c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:719b0871f0d32da894ec80dc2a23483623b7802a5c6f935cf3672468f99faa40
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:c8baecad9c97ddbbb3f9f02c90abbf483d46974b7bc9605622a302cbeba1de74
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:b5a8fb14c3bb7c89ff970ca85f06dfcdca832e2b45ad86e8bb1b6c0a0e7e0764
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:8444703b2a7822f82141afcf4f36bf255650d755321b9e9fe2cbd1bdea41d5c7
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:e9173d81ef2a75839c0e1d677b72bb00b8e6d7ef0511afab103f08a884a03738
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:a8d2e42757c6ccbfe4372465b1827c633aad012c5194bf8b08afcd5691caa6f5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:6bbcb64cf9adb4d3b6a4c259eb910e7da5c3c58732643e60d416399b2083cc0d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:ea57c7cb1ccd5bce17f6cb35e326ab974718be8cd19b0cd38ea8112e91c51939
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ee161954766211651d7cb13f3f5a3c025ac04094ba742be453ce789c6f0e24a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:ecbc65705de76804a3c7642339ecfb4cc0f5a61891058e07c20bf2b63f1b82b8
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:1df782d0d923544767696d719bd0ac10802b8e06508a104af5e023a519f32dee
        version: 6.1.3
      entries:
      - name: ibm-zen-operator.v6.1.3
        version: 6.1.3
      name: v6.1
    defaultChannel: v6.1
    packageName: ibm-zen-operator
    provider:
      name: IBM
      url: https://www.ibm.com/in-en/products/cloud-pak-for-data
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: hpe-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-csi-operator.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.hpe.com/v1",
                "kind": "HPECSIDriver",
                "metadata": {
                  "name": "hpecsidriver-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "csp": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "disable": {
                    "alletra6000": false,
                    "alletra9000": false,
                    "alletraStorageMP": false,
                    "b10000FileService": false,
                    "nimble": false,
                    "primera": false
                  },
                  "disableHostDeletion": false,
                  "disableNodeConfiguration": false,
                  "disableNodeConformance": false,
                  "disableNodeGetVolumeStats": false,
                  "disableNodeMonitor": false,
                  "disablePreInstallHooks": false,
                  "imagePullPolicy": "IfNotPresent",
                  "images": {
                    "b10000FileServiceCSP": "quay.io/hpestorage/alletrastoragemp-b10000-nfs-csp@sha256:6a1855a2ef1a83b886bb1f45b596e8e420e5e13c84bbf446c8c202eaf27a2013",
                    "csiAttacher": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                    "csiControllerDriver": "quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676",
                    "csiExtensions": "quay.io/hpestorage/csi-extensions@sha256:27074aaa8df3b867cfa3d12318612c45f8e541d750c27db5e080734592160c89",
                    "csiNodeDriver": "quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676",
                    "csiNodeDriverRegistrar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                    "csiProvisioner": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                    "csiResizer": "registry.k8s.io/sig-storage/csi-resizer@sha256:8ddd178ba5d08973f1607f9b84619b58320948de494b31c9d7cd5375b316d6d4",
                    "csiSnapshotter": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769",
                    "csiVolumeGroupProvisioner": "quay.io/hpestorage/volume-group-provisioner@sha256:0a5f39f52113b963978a695bb8e5a5a3c4d387a02eb4d4a9647b4bc804de7399",
                    "csiVolumeGroupSnapshotter": "quay.io/hpestorage/volume-group-snapshotter@sha256:177d9abf8fbd3ac26ca4c12ca4b4adde20ebab308c79f10b04632f0171746fa8",
                    "csiVolumeMutator": "quay.io/hpestorage/volume-mutator@sha256:bf6beac1af5edcd3c9774f0dcc9579b79ef9cc8e4b74a4238082ab9aa443b5ed",
                    "nfsProvisioner": "quay.io/hpestorage/nfs-provisioner@sha256:5bf5952453ecca28e26724df85ebd5c8f0b7479ae4cb683d4a09c415da00e002",
                    "nimbleCSP": "quay.io/hpestorage/alletra-6000-and-nimble-csp@sha256:5480ffdebd74fdc547bc4a11b4434abac7bbfee3cee349f1ac893fbad84bb8d1",
                    "primera3parCSP": "quay.io/hpestorage/alletra-9000-primera-and-3par-csp@sha256:8af6e87075c47ba8d9714d55848ab2a0fef62b0768681e30d9e63ed7f396b0bf"
                  },
                  "iscsi": {
                    "chapSecretName": ""
                  },
                  "kubeletRootDir": "/var/lib/kubelet",
                  "logLevel": "info",
                  "maxVolumesPerNode": 100,
                  "node": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/hpestorage/csi-driver-operator@sha256:29e92f68cf0d7691a6c56876e7dde0d69a8cb3ff6ae6c9371c604e61cba16898
          createdAt: "2025-09-25T16:28:16Z"
          description: A Container Storage Interface (CSI) driver for HPE storage
            platforms. The CSI driver allows you to use HPE storage with your preferred
            container orchestrator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <3.0.1'
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/hpe-storage/co-deployments
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE CSI Driver for Kubernetes
            displayName: HPECSIDriver
            kind: HPECSIDriver
            name: hpecsidrivers.storage.hpe.com
            version: v1
        description: |
          The HPE CSI Operator for OpenShift deploys the HPE CSI Driver on Kubernetes and OpenShift for dynamic provisioning of persistent volumes on HPE storage systems.
          The HPE CSI Driver for Kubernetes leverages HPE storage platforms to provide scalable and persistent storage for stateful applications.
          Please refer to our CSI driver [documentation](https://scod.hpedev.io/csi_driver/index.html) for supported HPE platforms and full list of supported CSI features.
          ## Installation
          Refer to the HPE CSI Operator for OpenShift [documentation](https://scod.hpedev.io/csi_driver/deployment.html#operator).
        displayName: HPE CSI Operator for OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE CSI
        - HPE Storage
        - CSI
        links:
        - name: Documentation
          url: https://scod.hpedev.io/csi_driver/index.html
        maintainers:
        - email: dev-hi-containers@hpe.com
          name: HPE Storage Containers Team
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: HPE
        relatedImages:
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - quay.io/hpestorage/volume-mutator@sha256:bf6beac1af5edcd3c9774f0dcc9579b79ef9cc8e4b74a4238082ab9aa443b5ed
        - registry.connect.redhat.com/hpestorage/csi-driver-operator@sha256:29e92f68cf0d7691a6c56876e7dde0d69a8cb3ff6ae6c9371c604e61cba16898
        - quay.io/hpestorage/alletrastoragemp-b10000-nfs-csp@sha256:6a1855a2ef1a83b886bb1f45b596e8e420e5e13c84bbf446c8c202eaf27a2013
        - quay.io/hpestorage/alletra-9000-primera-and-3par-csp@sha256:8af6e87075c47ba8d9714d55848ab2a0fef62b0768681e30d9e63ed7f396b0bf
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - quay.io/hpestorage/nfs-provisioner@sha256:5bf5952453ecca28e26724df85ebd5c8f0b7479ae4cb683d4a09c415da00e002
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8ddd178ba5d08973f1607f9b84619b58320948de494b31c9d7cd5375b316d6d4
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769
        - quay.io/hpestorage/volume-group-provisioner@sha256:0a5f39f52113b963978a695bb8e5a5a3c4d387a02eb4d4a9647b4bc804de7399
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - quay.io/hpestorage/volume-group-snapshotter@sha256:177d9abf8fbd3ac26ca4c12ca4b4adde20ebab308c79f10b04632f0171746fa8
        - quay.io/hpestorage/csi-extensions@sha256:27074aaa8df3b867cfa3d12318612c45f8e541d750c27db5e080734592160c89
        - quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676
        - quay.io/hpestorage/alletra-6000-and-nimble-csp@sha256:5480ffdebd74fdc547bc4a11b4434abac7bbfee3cee349f1ac893fbad84bb8d1
        version: 3.0.1
      entries:
      - name: hpe-csi-operator.v3.0.1
        version: 3.0.1
      name: stable
    defaultChannel: stable
    packageName: hpe-csi-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SmartX
      provider-url: https://www.smartx.com
    name: iomesh-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: iomesh-operator.0.10.1-rc4
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
          createdAt: 2021-08-17 00:00:00+00:00
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/iomesh-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/iomesh-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/iomesh/docs
          support: SmartX, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: iomesh
            displayName: iomesh
            kind: IOMeshCluster
            name: iomeshclusters.iomesh.com
            version: v1alpha1
        description: "IOMesh (https://www.iomesh.com/) is a cloud-native storage product
          specifically designed and developed for Kubernetes. It takes ZBS as its
          core, a “production-ready” distributed storage system developed independently
          by SmartX, to provide production-level high performance and reliable persistent
          storage for business-critical applications such as MySQL, Cassandra, MongoDB,
          etc., contributing to the containerized transformation of stateful applications.
          \     \n## Pre-Requisites\n* According to https://docs.iomesh.com/deploy/prerequisites,
          make sure that your openshift cluster meets the \"Installation Requirements\"
          and set up the worker nodes correctly according to the \"Setup Worker Node\"
          chapter.\n* Run IOMesh Operator pre-installation script in an environment
          where oc or kubectl can be used to access the openshift cluster, the script
          will install the dependencies of IOMesh Operator and specific IOMesh settings
          for the openshift cluster:\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-pre-install-openshift.sh
          | sh -\n```\n## Install IOMesh Operator and IOMesh Cluster\n* Click install
          button on the current web page to install IOMesh Operator\n* On the **Installed
          Operators** -> **IOMesh Operator** -> **Create instance** -> **YAML view**,
          fill in an IOMesh Custom Resources according https://iomesh.run/iomesh.yaml,
          change the `spec.*.dataCIDR` to your own data network CIDR:\n## Post Install\n*
          Run IOMesh Operator post-installation script in an environment where oc
          or kubectl can be used to access the openshift cluster, the script will
          install IOMesh CSI Driver\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-post-install-openshift.sh
          | sh -\n```\n## Getting Started\n* Try to quickstart setup and use IOMesh:
          https://docs.iomesh.com/deploy/setup-iomesh\n"
        displayName: IOMesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iomesh-operator
        links:
        - name: IOMesh Operator
          url: https://iomesh.com
        maintainers:
        - email: na.wang@smartx.com
          name: Na Wang
        maturity: alpha
        provider:
          name: SmartX
          url: https://www.smartx.com
        relatedImages:
        - iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
        version: 0.10.1-rc4
      entries:
      - name: iomesh-operator.0.10.1-rc4
        version: 0.10.1-rc4
      name: stable
    defaultChannel: stable
    packageName: iomesh-operator-rhmp
    provider:
      name: SmartX
      url: https://www.smartx.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: openstack-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openstack-operator.v1.0.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openstack.org/v1beta1",
                "kind": "OpenStack",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openstack-operator",
                    "app.kubernetes.io/instance": "openstack",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "openstack",
                    "app.kubernetes.io/part-of": "openstack-operator"
                  },
                  "name": "openstack",
                  "namespace": "openstack-operators"
                },
                "spec": {
                  "operatorOverrides": [
                    {
                      "controllerManager": {
                        "resources": {
                          "limits": {
                            "cpu": "600m",
                            "memory": "2Gi"
                          }
                        }
                      },
                      "name": "infra",
                      "replicas": 1
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-02T13:33:09Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.14
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.openstack.org/v1beta1","kind":"OpenStack","metadata":{"name":"openstack","namespace":"openstack-operators"},"spec":{}}'
          operatorframework.io/suggested-namespace: openstack-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenStack is the Schema for the openstacks API
            displayName: Open Stack
            kind: OpenStack
            name: openstacks.operator.openstack.org
            version: v1beta1
        description: Install and configure OpenStack
        displayName: OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        links:
        - name: Openstack Operator
          url: https://github.com/openstack-k8s-operators/
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/glance-rhel9-operator@sha256:ee22cf417b7771930a442c2302277e6d37f368f5ae551fee9a50dbbe41c8fe3a
        - registry.redhat.io/rhoso-operators/mysqld-exporter-rhel9@sha256:b49e966cdd588071e0e1f2791ec6c62310078a7d39ea77f83843d3708a07c2b5
        - registry.redhat.io/rhoso/openstack-swift-proxy-server-rhel9@sha256:858551fd1c519797cae0e146e9c05412ed6e66dac01d0e68e62dd2d4bd335391
        - registry.redhat.io/rhoso/openstack-netutils-rhel9@sha256:b02ab1617922a2560bf739172bc2ce0f6b1fb3ef55054bd17e4b16bdf9512b4b
        - registry.redhat.io/rhoso/openstack-cinder-backup-rhel9@sha256:910693233ee3a1f3832ebe744f67a5139249523456de4abbd731c9d2f544b7c3
        - registry.redhat.io/rhoso/ironic-python-agent-rhel9@sha256:2ccd8e22a242059ca5c4e2bce853c36a71c36fdf40e8e36ab183a8218b0cc87b
        - registry.redhat.io/rhoso/openstack-frr-rhel9@sha256:f6516277149d1560c0bf2ac74a99cfb4ac2c84f9f090ab231b6d12bdb824de3b
        - registry.redhat.io/rhoso/openstack-octavia-housekeeping-rhel9@sha256:0a79259e398af746d0e8a6dda1aae5ee976c6bf8991b17f4a5da0d07dca1e36c
        - registry.redhat.io/rhoso/openstack-swift-object-rhel9@sha256:8fc2ab8caa9c7c35d94ef51c303763ea71e40b9c795c4555c17e6194344419b3
        - registry.redhat.io/ubi9/httpd-24@sha256:29382518403a4361c48fccc057163dc4a9f43ee159432754fa63c76485b223f5
        - registry.redhat.io/rhoso-operators/heat-rhel9-operator@sha256:583abb81a49db85949f7a4b846591ad8532afa79440ad7e2937ee00eacd972b3
        - registry.redhat.io/rhoso/openstack-barbican-keystone-listener-rhel9@sha256:cd8d4f0a94b4fe080ebf8e459f091db6637ceec6d83ac23cf04d66bd03c71e76
        - registry.redhat.io/rhoso/openstack-neutron-metadata-agent-ovn-rhel9@sha256:20323b151474703df610b6c24ccb9c748265196bbc70a7b5ece1b801ff540cd9
        - registry.redhat.io/rhoso-operators/test-rhel9-operator@sha256:df7566483abb18aa3c45d8bc4cd91531c7eb5b63345e511151a7cdca2dbcef17
        - registry.redhat.io/rhoso/openstack-iscsid-rhel9@sha256:508bd67021c271a10bb55ab4af033dc4e931708674b9c5ba94f72c07cd0c8ee2
        - registry.redhat.io/rhoso-operators/openstack-baremetal-agent-rhel9@sha256:08675b16d89d7931b1ef7b96e1e884c88e8e7fbfbe79a944f9213c52aba3ba83
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/rhoso/openstack-ironic-neutron-agent-rhel9@sha256:0abaa8f437025d2e7006c59c546ff6935556d707b57993719faef94142e1c43d
        - registry.redhat.io/rhoso/openstack-aodh-evaluator-rhel9@sha256:48c9192acf157df5b8f0827e35f0380d50cd4562676aba3c0d7791f92328af02
        - registry.redhat.io/rhoso/openstack-manila-scheduler-rhel9@sha256:8ce47f7fd8660167b3b292659d4fc58cf34ed671d5779693d24b71a9fddc1e7c
        - registry.redhat.io/rhoso/openstack-neutron-dhcp-agent-rhel9@sha256:cc9223ef705bcd5d0dcabbf71ad9f6e540ae7f003a6afd13161301f842dfa5fa
        - registry.redhat.io/rhoso/edpm-hardened-uefi-rhel9@sha256:c18ec7023fbcc4ab5183f689adc1f01a626a4923c15c61a37388f99904db7bc3
        - registry.redhat.io/rhoso/openstack-ceilometer-central-rhel9@sha256:e6d9f47406b7e3ff2a7f2256088cc64f043b8e497382b9f569eb90182598bbe9
        - registry.redhat.io/rhoso/openstack-cinder-api-rhel9@sha256:72a674caeeeecce3b60217f3607ad5b00651fa3d52f2f5d905f6e6395638dbc8
        - registry.redhat.io/rhoso/openstack-glance-api-rhel9@sha256:b1c0b09bb9aafe3bc31f43f8a99a92dde8d78cf90e781dd14300580986d8c980
        - registry.redhat.io/rhoso-operators/designate-rhel9-operator@sha256:d446af622bd32666eb4eb5f9b7f38f0c33ac81251a2230db9d5baf71081b9d6a
        - registry.redhat.io/rhoso/openstack-designate-producer-rhel9@sha256:04af4230f815a6e3fde22003cf382eebd9cbc3c3ec4056795f4474abc1e65fea
        - registry.redhat.io/rhoso-operators/mariadb-rhel9-operator@sha256:907e3a9d15941440d5289d7e448d9f0351053fc20fcb8cb3b9db0a84debd876f
        - registry.redhat.io/rhoso/openstack-manila-api-rhel9@sha256:e4dd3d5b66c5216b45a1515a975b85fd2ac4ef698a9350dd16e01417da462b32
        - registry.redhat.io/rhoso/openstack-cinder-volume-rhel9@sha256:8a09e80dac94fb4b3d35033e91812f95f0acac66532d9abc88aa1fb715d21822
        - registry.redhat.io/rhoso/openstack-designate-worker-rhel9@sha256:a6324a8a72722822df7692b8df0420388a1b4aecfff3da7e91da4afe4d9e2260
        - registry.redhat.io/rhoso/openstack-ceilometer-compute-rhel9@sha256:296f7f6c6bdcd7f5d07b8470308a4c71055f29a17c48d2ebb03b7f54a0e927ff
        - registry.redhat.io/rhoso/openstack-heat-api-cfn-rhel9@sha256:2789929ae3edd7c68579f34457d3071273e49a66f462043cd9c46e75bdda282a
        - registry.redhat.io/openshift4/ose-kube-state-metrics-rhel9@sha256:db6074b8aeedfd7603ae3adbe67071115f54e7f01a24997f1cd8fa8958aeddce
        - registry.redhat.io/rhoso-operators/cinder-rhel9-operator@sha256:12b8dd8fd3683250365eff6f07339cb976ee29da80642e591c10e9fa5ba627c9
        - registry.redhat.io/rhoso/openstack-heat-engine-rhel9@sha256:151e04b3e5a408aadef3a17954255acc3733d09961878a135e177188457447f0
        - registry.redhat.io/rhoso/openstack-rsyslog-rhel9@sha256:272e729accc2b64b8fcc0ece4ae5aa9ce6d454e34a25a8a8e2b99ac56b73361e
        - registry.redhat.io/rhoso/openstack-tempest-all-rhel9@sha256:16b1fbfd36ce9a27106bf09c414eaa93f14c9a6b11839068976c2fdb537d2ba7
        - registry.redhat.io/rhoso/openstack-octavia-api-rhel9@sha256:7b8d5927162ddab512096beb02e837ba51f6353c5187f3de72eb3db6e56ab6e7
        - registry.redhat.io/rhoso/openstack-keystone-rhel9@sha256:ac076c81247d774a9aeee1b63540ca6e18058213cf6aad10a6fd1f535a394ef4
        - registry.redhat.io/rhoso/openstack-cinder-scheduler-rhel9@sha256:2079ffe9b620e9277573ac796bf7e393b520631130e5bbb6f5ab51d872c943f3
        - registry.redhat.io/rhoso/openstack-aodh-notifier-rhel9@sha256:d2e89f27aac9afa504411ae6a6163442cd717f8c560e505a7da8e3fd9e2cff95
        - registry.redhat.io/rhoso/openstack-manila-share-rhel9@sha256:277174c885998e6f5b867d099664f6fa28f974bffb524903ff75283035cd8dc9
        - registry.redhat.io/rhoso/openstack-neutron-server-rhel9@sha256:2d3afde3201aa18b192ad0fb079c1e4f4f673565604f1af2b79aba8a7bc88bf2
        - registry.redhat.io/rhoso/openstack-ovn-northd-rhel9@sha256:668f742b2a18579bc8cfc610a862e43c5b877413cf65eabbe0ee36b3c7869446
        - registry.redhat.io/rhoso-operators/openstack-baremetal-rhel9-operator@sha256:03916e507b88943c0e00cedeb3cf20fa0787ccc762e47f38d305ff4d1b125b12
        - registry.redhat.io/rhoso/openstack-ovn-nb-db-server-rhel9@sha256:8c2ce1027779dc66d67a0e0d03e48b226d3278208cae0b2b81b35ad1b1f3ae72
        - registry.redhat.io/rhoso/openstack-neutron-ovn-agent-rhel9@sha256:bed1956c66b0e70064b51e34faac64036f77c7ecb14f46ceac4bc9ef52f0b9e0
        - registry.redhat.io/rhoso/openstack-nova-scheduler-rhel9@sha256:f0e5d3a693c9998ca3839b3ec7dfd5da0f9cf139e0b5b6913e92a10d5a13e6bc
        - registry.redhat.io/rhoso/openstack-designate-central-rhel9@sha256:d28d8d91d8bf75a1b77a5d650c7f2650a1791e2a1b0341450fa46978c955b92f
        - registry.redhat.io/rhoso/openstack-barbican-api-rhel9@sha256:bc942204843c45c6cb7ea841af8611f04efb003de7eeccc6813934aad556da73
        - registry.redhat.io/rhoso/openstack-octavia-health-manager-rhel9@sha256:8065851058019757b56f86ca4f030d44f169100eb2e37e08e0587a7b1e4d2e72
        - registry.redhat.io/rhoso/openstack-nova-api-rhel9@sha256:4116206f1fbe60fdd9458a5edee848571c94e2b0512d9c2cfcfd6890d60d4395
        - registry.redhat.io/rhoso-operators/keystone-rhel9-operator@sha256:879d4b1f9e720f8581b3a55c788a97bbe7b52f9108df3d5dce2e9bcb4200df6d
        - registry.redhat.io/rhoso/openstack-horizon-rhel9@sha256:82d0e6d24733afd305996e0f4a118df3e48621c40c9669b01f48669dbf40d868
        - registry.redhat.io/rhoso/openstack-designate-api-rhel9@sha256:b76ca7820553285bb4ee2bfb08e39aedbfe8c16b4edb478c06418a1e335a64ab
        - registry.redhat.io/rhoso/openstack-swift-container-rhel9@sha256:95fec49643fc18f402b8ba8057f2c35f214bf6138c5f0d7bbe1b96cbd78ef8d6
        - registry.redhat.io/rhoso-operators/swift-rhel9-operator@sha256:c99ef14338290d63e2bcbf129e953b434aea382e47072812bae31b1ea838e7df
        - registry.redhat.io/rhoso/openstack-ovn-base-rhel9@sha256:a3b8ec7ad1073facb8876acc787549a1a68b9bf81bebdc127f13eb6385e2d1da
        - registry.redhat.io/rhoso/openstack-ovn-bgp-agent-rhel9@sha256:a404c3416ab65e7134f438362e887529be474639e725e18f3542e86973edac87
        - registry.redhat.io/rhoso/openstack-nova-novncproxy-rhel9@sha256:04881867f9b24403bf6ef051eee3902a17e8476eee894ef579c0d5d130d2510d
        - registry.redhat.io/rhoso/openstack-ironic-inspector-rhel9@sha256:88357e83e8a4bbc3c5a37f0420ad75556821834a975cd8a1cf5fee56de2a4f99
        - registry.redhat.io/rhoso/openstack-ovn-controller-rhel9@sha256:3c42b7e0305c508d74602b9378143deb07a569443bba35850cdb7b6e3f5202e2
        - registry.redhat.io/rhoso/openstack-nova-conductor-rhel9@sha256:c67ca0e6c8126f56c6a83632b6a96bd4a35006f5f414a8779467b81864b31d12
        - registry.redhat.io/rhoso-operators/ee-openstack-ansible-ee-rhel9@sha256:61e729fb89c4872837cbe9cb1e3cea402b61f0da240751302605df487721f0aa
        - registry.redhat.io/rhoso-operators/ovn-rhel9-operator@sha256:66cd8add6078b63a36879c974db9e9f358593bb0651aa39fe85153d25404ae27
        - registry.redhat.io/rhoso/openstack-designate-backend-bind9-rhel9@sha256:82cca999357b0ab25bcc081bf802fab831587cf661db806af4ca453e0a626743
        - registry.redhat.io/rhoso/openstack-rabbitmq-rhel9@sha256:56bc2d2b201225fb98db43ea9ca68cceb4842f8f0f1a96ab3ab62ab785651d12
        - registry.redhat.io/rhoso-operators/prometheus-podman-exporter-rhel9@sha256:f33e7506db6dfdf530d3e2442baaf360d212bb1a47a7db887dba967ed0655cc2
        - registry.redhat.io/rhoso/openstack-nova-compute-rhel9@sha256:6264075310231af358803b53d5a09250dab32d63fc5966aa8fb957e5f56096a8
        - registry.redhat.io/rhoso-operators/neutron-rhel9-operator@sha256:30b6a812c47db9e6691211185c304c3e8fc6a8b9ba7ca79acf59c5ec07fb9fb0
        - registry.redhat.io/rhoso-operators/openstack-must-gather-rhel9@sha256:60b4355f4dbef3ad64103357086b10a5c0af126dd23798d6a9d725646f1e13b4
        - registry.redhat.io/rhoso/openstack-ironic-api-rhel9@sha256:50ccfe7d9ba22a50e0c1c8159cade6f4b1ac13ad8f5770f20c7be18fba58c8c3
        - registry.redhat.io/rhoso/openstack-ironic-pxe-rhel9@sha256:42e35ded74033f1f07287441917829cc934930a5df696c33cd8957be3c10fa41
        - registry.redhat.io/rhoso/openstack-openstackclient-rhel9@sha256:2952521d4b1799fd32ff773ef2d457a3abc354b639cf7f1e0f7f6615ca35ac8f
        - registry.redhat.io/rhoso-operators/manila-rhel9-operator@sha256:d9e69eb140b8ff76d5bf49deab8d700639ccdfd1ab5c8c7667bc2bbb50f3d60b
        - registry.redhat.io/rhoso-operators/horizon-rhel9-operator@sha256:9a2a13635588c6cc9949b272e1a1708c5fb53f19b05b8310ace703df5958ac63
        - registry.redhat.io/rhoso-operators/nova-rhel9-operator@sha256:3156dde249e49b40bed71092fb5213b112d43b663ba8c556e355aed524bec822
        - registry.redhat.io/rhoso-operators/ironic-rhel9-operator@sha256:dd67ffe3578ce619748735c1a5cd2a57d704d0c36c1b841d513ea36479aa4682
        - registry.redhat.io/rhoso/openstack-aodh-api-rhel9@sha256:4adc09cc1d922b387320da467de89d3f41a9c49b46d9b728458fec36505ec940
        - registry.redhat.io/rhoso/openstack-cron-rhel9@sha256:f09e724c69cec2c3b685783f9263b26c280a57b5909a723341416b342d18f9a4
        - registry.redhat.io/rhoso/openstack-ironic-conductor-rhel9@sha256:538da9fdeec546d794d192361d0a76f8b4e7682b6eb6d9cea371d8a52d603800
        - registry.redhat.io/rhoso/openstack-swift-account-rhel9@sha256:31e727f37db92fcc21319c44396576a03a118ba25d403715650162f7c0ead31e
        - registry.redhat.io/rhoso/openstack-redis-rhel9@sha256:c6dbeacbbd5aed9d13e540f1c987a574ea45747983c280446e5afadcf7c33b37
        - registry.redhat.io/rhoso/openstack-ovn-sb-db-server-rhel9@sha256:8ba4f8ed49e214f421e94133da0f17b54ce99a52bbbd367d0aef357de844ea01
        - registry.redhat.io/openshift4/ose-prometheus-node-exporter-rhel9@sha256:479c51de2fc8a3b3858d71bbe4eeafbd5b43263472346f9dde348115082894cc
        - registry.redhat.io/rhoso-operators/octavia-rhel9-operator@sha256:aa0f00f8e03dfdb547e53cf611581e9cc091eaf049558121c2bc5d702e556bcb
        - registry.redhat.io/rhoso/openstack-mariadb-rhel9@sha256:6115a7458cfa412c1659242fcec64b63f0a5ca414ecf3c5f890f0afec0bca069
        - registry.redhat.io/rhoso/openstack-multipathd-rhel9@sha256:6ff8ae176f8a633bcd78b1f8407924c74bb2f317eba16b90e7f0f9e53af74e8e
        - registry.redhat.io/rhoso/openstack-neutron-sriov-agent-rhel9@sha256:118d5f9430192ab48cf159ee2f7912f08e935ad56ef3364f18b3680210bb3e52
        - registry.redhat.io/rhoso-operators/barbican-rhel9-operator@sha256:b3b769e0dc32c12c5cc9578c2cac2622b973177ca14697bc1042d21d473e4e84
        - registry.redhat.io/rhoso/openstack-heat-api-rhel9@sha256:ccbc41b552bb34d223c9615a4e3b56bb516bc5bd8a694f048c15063ba45fca9f
        - registry.redhat.io/rhoso-operators/telemetry-rhel9-operator@sha256:0635515a5fed12f17de4cc5ac5f9567e3b3546fc2f69730c7beab25cf722f538
        - registry.redhat.io/rhoso/openstack-aodh-listener-rhel9@sha256:037845be6787f610e479345206b0fcfb2f1dc430540816bc5a54a1d26da2acbd
        - registry.redhat.io/rhoso/openstack-ceilometer-notification-rhel9@sha256:bd2b5f50a2bef4e77e3daa89aca51df6e0e55c3784886296d92c97616cd566d1
        - registry.redhat.io/rhoso-operators/sg-core-rhel9@sha256:b56923f869ff57396e6822133dfc3b21007c3df652cfe962c7ad3028214d988c
        - registry.redhat.io/rhoso/openstack-barbican-worker-rhel9@sha256:485525116fb3c22d8c60330dd87ebd3854cd71ded36bf88bcf18cee5957e22a3
        - registry.redhat.io/rhoso-operators/openstack-network-exporter-rhel9@sha256:8e8e775e37c28cb11b7aae9199405b85d5862dce195d165ca598fa1dc89bf914
        - registry.redhat.io/rhoso/openstack-placement-api-rhel9@sha256:0aa5e138947bf236f31c81bfb6f641d95af63812c06dca43141721b383c7e08e
        - registry.redhat.io/openshift-power-monitoring/kepler-rhel9@sha256:8e9e5bfeed61aaeee85bf5eefc3390556b8d4cbfc797a82cabe95bc0367126eb
        - registry.redhat.io/rhoso/openstack-ceilometer-ipmi-rhel9@sha256:2fb5723c68d0f2c3ae28b86e606ccae4b80f3bd9b16c01dd6c9adcae629db7f1
        - registry.redhat.io/rhoso-operators/openstack-rhel9-operator@sha256:64e2ad312f757101f083eedcb17e0b6c05cf6ebf6ed9e7db2e33e84c6aca9898
        - registry.redhat.io/rhoso/openstack-unbound-rhel9@sha256:7f17a719e8e49d457ff16437709c2b4eca5b9c2f58972d1e351c5c1c1ac38962
        - registry.redhat.io/rhoso-operators/infra-rhel9-operator@sha256:a2dc031a6ad2726d155e77cafb676f8e35a86e963b97ccdceed0da137e67ba0d
        - registry.redhat.io/rhoso/openstack-designate-mdns-rhel9@sha256:ce43dd1a30afd4753aa11d32df2b24ed4bf3d67847044aaae8f3e5c8901caf05
        - registry.redhat.io/rhoso/openstack-octavia-worker-rhel9@sha256:303e35926e43fe403673e18a25c2e815547c28a5f8c0f09d470cd4148a21c528
        - registry.redhat.io/rhoso-operators/placement-rhel9-operator@sha256:a16bd4fb929eb5bcf96f2aefcb5a7abf4143f4f9ed24e53de93f4a3caa89a507
        - registry.redhat.io/rhoso-operators/rabbitmq-cluster-rhel9-operator@sha256:e095e6cbda841b9c726c1ec0839a98964b894cf25ae9228052d314b1ee2f161d
        - registry.redhat.io/rhoso/openstack-memcached-rhel9@sha256:3e704b9664c65be39e83c2ed802722ffeaf969d8e571147a2be128d5c56331c1
        version: 1.0.14
      entries:
      - name: openstack-operator.v1.0.14
        version: 1.0.14
      - name: openstack-operator.v1.0.13
        version: 1.0.13
      - name: openstack-operator.v1.0.12
        version: 1.0.12
      - name: openstack-operator.v1.0.11
        version: 1.0.11
      - name: openstack-operator.v1.0.10
        version: 1.0.10
      - name: openstack-operator.v1.0.9
        version: 1.0.9
      - name: openstack-operator.v1.0.8
        version: 1.0.8
      - name: openstack-operator.v1.0.7
        version: 1.0.7
      - name: openstack-operator.v1.0.6
        version: 1.0.6
      - name: openstack-operator.v1.0.5
        version: 1.0.5
      - name: openstack-operator.v1.0.4
        version: 1.0.4
      - name: openstack-operator.v1.0.3
        version: 1.0.3
      - name: openstack-operator.v1.0.2
        version: 1.0.2
      - name: openstack-operator.v1.0.1
        version: 1.0.1
      - name: openstack-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: openstack-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Devfile
      provider-url: https://devfile.io
    name: devworkspace-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devworkspace-operator.v0.36.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/internal-objects: '["devworkspaceroutings.controller.devfile.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DevWorkspaceOperatorConfig
            name: devworkspaceoperatorconfigs.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspaceRouting
            name: devworkspaceroutings.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha2
          - kind: DevWorkspaceTemplate
            name: devworkspacetemplates.workspace.devfile.io
            version: v1alpha1
          - kind: DevWorkspaceTemplate
            name: devworkspacetemplates.workspace.devfile.io
            version: v1alpha2
        description: |
          The DevWorkspace Operator enables cluster-level support for the
          [Devfile 2.0 spec](https://docs.devfile.io), enabling static, reproducible
          configurations for launching cloud-based editors and IDEs in OpenShift and
          Kubernetes clusters.

          Leveraging the language server protocol and the Eclipse Theia web IDE, the
          DevWorkspace operator provides easy configuration of full development
          environments on the cloud with support for a wide variety of languages and
          technologies, including Go, Java, Typescript/Javascript, Python, and more.

          The DevWorkspace Operator is also used in the Web Terminal Operator to
          automatically provision Web Terminal environments.

          ## Installing the operator
          The DevWorkspace Operator can be installed directly from the OperatorHub UI and
          will be available in all namespaces on the cluster. DevWorkspace creation is
          driven by the DevWorkspace custom resource, which can be created in any
          namespace to provision a full development environment. To get started, browse
          the DevWorkspace [spec](https://devfile.io/docs/2.3.0/devfile-schema).

          Once a DevWorkspace is started, it can be accessed via the URL in its
          `.status.mainUrl` field.

          It's recommended to install the DevWorkspace Operator to the
          `openshift-operators` namespace for compatibility.

          ## Uninstalling the operator
          The DevWorkspace Operator utilizes finalizers on resources it creates and
          webhooks to restrict access to development resources. As a result, manual steps
          are required in uninstalling the operator. See the
          [documentation](https://github.com/devfile/devworkspace-operator/blob/main/docs/uninstall.md)
          for details.
        displayName: DevWorkspace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Devworkspace Operator
        - DevWorkspaces
        - Devfile
        links:
        - name: Devworkspace Operator
          url: https://github.com/devfile/devworkspace-operator
        maintainers:
        - email: dakwon@redhat.com
          name: David Kwon
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        maturity: alpha
        provider:
          name: Devfile
          url: https://devfile.io
        relatedImages:
        - registry.redhat.io/devworkspace/devworkspace-rhel9-operator@sha256:83b6abf6e57703da1125e037ea9353c224602bddc3146e5b00df47f1b1976e9a
        - registry.redhat.io/ubi9/ubi-minimal@sha256:a50731d3397a4ee28583f1699842183d4d24fadcc565c4688487af9ee4e13a44
        - registry.redhat.io/devworkspace/devworkspace-project-clone-rhel9@sha256:312a013201877cea03956e3db2ba19a3d7f842de2c3a65927c65c3e8d22a57fe
        version: 0.36.0
      entries:
      - name: devworkspace-operator.v0.36.0
        version: 0.36.0
      - name: devworkspace-operator.v0.35.1
        version: 0.35.1
      - name: devworkspace-operator.v0.34.0
        version: 0.34.0
      - name: devworkspace-operator.v0.33.0
        version: 0.33.0
      - name: devworkspace-operator.v0.32.1
        version: 0.32.1
      - name: devworkspace-operator.v0.31.2
        version: 0.31.2
      - name: devworkspace-operator.v0.31.1
        version: 0.31.1
      - name: devworkspace-operator.v0.31.0
        version: 0.31.0
      - name: devworkspace-operator.v0.30.2
        version: 0.30.2
      - name: devworkspace-operator.v0.30.1
        version: 0.30.1
      - name: devworkspace-operator.v0.30.0
        version: 0.30.0
      - name: devworkspace-operator.v0.29.0
        version: 0.29.0
      - name: devworkspace-operator.v0.28.0-0.1719314743.p
        version: 0.28.0+0.1719314743.p
      - name: devworkspace-operator.v0.28.0
        version: 0.28.0
      - name: devworkspace-operator.v0.27.0-0.1714987663.p
        version: 0.27.0+0.1714987663.p
      - name: devworkspace-operator.v0.27.0
        version: 0.27.0
      - name: devworkspace-operator.v0.26.0
        version: 0.26.0
      - name: devworkspace-operator.v0.25.0
        version: 0.25.0
      - name: devworkspace-operator.v0.24.0
        version: 0.24.0
      - name: devworkspace-operator.v0.23.0
        version: 0.23.0
      - name: devworkspace-operator.v0.22.0
        version: 0.22.0
      - name: devworkspace-operator.v0.21.2
        version: 0.21.2
      - name: devworkspace-operator.v0.21.1
        version: 0.21.1
      - name: devworkspace-operator.v0.20.0
        version: 0.20.0
      - name: devworkspace-operator.v0.19.1
        version: 0.19.1
      - name: devworkspace-operator.v0.19.1-0.1682321189.p
        version: 0.19.1+0.1682321189.p
      - name: devworkspace-operator.v0.19.1-0.1679521112.p
        version: 0.19.1+0.1679521112.p
      - name: devworkspace-operator.v0.18.1
        version: 0.18.1
      - name: devworkspace-operator.v0.18.1-0.1675929565.p
        version: 0.18.1+0.1675929565.p
      - name: devworkspace-operator.v0.17.0
        version: 0.17.0
      - name: devworkspace-operator.v0.16.0-0.1666668361.p
        version: 0.16.0+0.1666668361.p
      - name: devworkspace-operator.v0.16.0
        version: 0.16.0
      - name: devworkspace-operator.v0.15.2-0.1661828401.p
        version: 0.15.2+0.1661828401.p
      - name: devworkspace-operator.v0.15.2
        version: 0.15.2
      - name: devworkspace-operator.v0.14.1
        version: 0.14.1
      - name: devworkspace-operator.v0.13.0
        version: 0.13.0
      - name: devworkspace-operator.v0.12.0
        version: 0.12.0
      - name: devworkspace-operator.v0.11.0
        version: 0.11.0
      - name: devworkspace-operator.v0.10.0
        version: 0.10.0
      - name: devworkspace-operator.v0.9.0
        version: 0.9.0
      name: fast
    defaultChannel: fast
    packageName: devworkspace-operator
    provider:
      name: Devfile
      url: https://devfile.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: IBM
      provider-url: ""
    name: ibm-block-csi-operator-community
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12
    - currentCSV: ibm-block-csi-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
        version: 1.12.0
      entries:
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.0
    - currentCSV: ibm-block-csi-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
        version: 1.12.1
      entries:
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.1
    - currentCSV: ibm-block-csi-operator.v1.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:3fc8df547138a2009225d63c2a3111a672340bdf9f3cb24a889cbb69f2c33c2d
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:3fc8df547138a2009225d63c2a3111a672340bdf9f3cb24a889cbb69f2c33c2d
        version: 1.12.2
      entries:
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.2
    - currentCSV: ibm-block-csi-operator.v1.12.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:2bd461a1e40a1366c871585aa7a4e8039e686a66bf7192efdeb1c215209a071d
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:2bd461a1e40a1366c871585aa7a4e8039e686a66bf7192efdeb1c215209a071d
        version: 1.12.3
      entries:
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.3
    - currentCSV: ibm-block-csi-operator.v1.12.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e29db476de3778222bf049926223dee20f6bc34e55a04df5f0f3df38741ff23e
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e29db476de3778222bf049926223dee20f6bc34e55a04df5f0f3df38741ff23e
        version: 1.12.4
      entries:
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.4
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.5
    defaultChannel: stable
    packageName: ibm-block-csi-operator-community
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: ""
    name: pmem-csi-operator-os
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pmem-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pmem-csi.intel.com/v1beta1",
                "kind": "PmemCSIDeployment",
                "metadata": {
                  "name": "pmem-csi.intel.com"
                },
                "spec": {
                  "deviceMode": "lvm",
                  "nodeSelector": {
                    "feature.node.kubernetes.io/memory-nv.dax": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
          createdAt: "2021-08-26T10:45:23Z"
          description: An operator for deploying and managing the PMEM-CSI driver
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: ""
          repository: https://github.com/intel/pmem-csi/tree/release-1.0
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a PMEM-CSI driver deployment in the cluster.
            displayName: PmemCSIDeployment
            kind: PmemCSIDeployment
            name: pmemcsideployments.pmem-csi.intel.com
            version: v1beta1
        description: |2

          Intel PMEM-CSI is a [CSI](https://github.com/container-storage-interface/spec)
          storage driver for container orchestrators like Kubernetes. It makes local
          persistent memory ([PMEM](https://pmem.io/)) available as a filesystem volume to
          container applications. This is the operator to deploy and manage the
          [PMEM-CSI](https://intel.github.io/pmem-csi/1.0/README.html)
          driver on a Kubernetes cluster.

          Refer to the [PMEM-CSI documentation](https://intel.github.io/pmem-csi/1.0/README.html)
          for more details on deploying and using the PMEM-CSI driver.
        displayName: Operator for PMEM-CSI driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - PMEM-CSI
        - PMEM-CSI operator
        - Persistent Memory
        - PMEM
        links:
        - name: PMEM-CSI
          url: https://intel.github.io/pmem-csi/1.0
        maintainers:
        - email: olev.kartau@intel.com
          name: Olev Kartau
        - email: patrick.ohly@intel.com
          name: Patrick Ohly
        - email: amarnath.valluri@intel.com
          name: Amarnath Valluri
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Intel® Corporation
        relatedImages:
        - registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
        version: 1.0.0
      entries:
      - name: pmem-csi-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: pmem-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pmem-csi.intel.com/v1beta1",
                "kind": "PmemCSIDeployment",
                "metadata": {
                  "name": "pmem-csi.intel.com"
                },
                "spec": {
                  "deviceMode": "lvm",
                  "nodeSelector": {
                    "feature.node.kubernetes.io/memory-nv.dax": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
          createdAt: "2021-08-26T10:45:23Z"
          description: An operator for deploying and managing the PMEM-CSI driver
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: ""
          repository: https://github.com/intel/pmem-csi/tree/release-1.0
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a PMEM-CSI driver deployment in the cluster.
            displayName: PmemCSIDeployment
            kind: PmemCSIDeployment
            name: pmemcsideployments.pmem-csi.intel.com
            version: v1beta1
        description: |2

          Intel PMEM-CSI is a [CSI](https://github.com/container-storage-interface/spec)
          storage driver for container orchestrators like Kubernetes. It makes local
          persistent memory ([PMEM](https://pmem.io/)) available as a filesystem volume to
          container applications. This is the operator to deploy and manage the
          [PMEM-CSI](https://intel.github.io/pmem-csi/1.0/README.html)
          driver on a Kubernetes cluster.

          Refer to the [PMEM-CSI documentation](https://intel.github.io/pmem-csi/1.0/README.html)
          for more details on deploying and using the PMEM-CSI driver.
        displayName: Operator for PMEM-CSI driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - PMEM-CSI
        - PMEM-CSI operator
        - Persistent Memory
        - PMEM
        links:
        - name: PMEM-CSI
          url: https://intel.github.io/pmem-csi/1.0
        maintainers:
        - email: olev.kartau@intel.com
          name: Olev Kartau
        - email: patrick.ohly@intel.com
          name: Patrick Ohly
        - email: amarnath.valluri@intel.com
          name: Amarnath Valluri
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Intel® Corporation
        relatedImages:
        - registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
        version: 1.0.0
      entries:
      - name: pmem-csi-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pmem-csi-operator-os
    provider:
      name: Intel® Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OceanBase
      provider-url: https://oceanbase.com
    name: oceanbase-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oceanbase-operator.v2.2.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Full Lifecycle
          createdAt: "2024-08-14T16:21:06Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: OBClusterOperation
            name: obclusteroperations.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBCluster
            name: obclusters.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBParameter
            name: obparameters.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBResourceRescue
            name: obresourcerescues.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBServer
            name: observers.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantBackupPolicy
            name: obtenantbackuppolicies.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantBackup
            name: obtenantbackups.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantOperation
            name: obtenantoperations.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantRestore
            name: obtenantrestores.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenant
            name: obtenants.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBZone
            name: obzones.oceanbase.oceanbase.com
            version: v1alpha1
        description: Ob Operator description. TODO.
        displayName: Ob Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - oceanbase-operator
        - ob-operator
        links:
        - name: OceanBase Operator
          url: https://oceanbase.github.io/ob-operator
        maintainers:
        - email: hongxin.sun.vip@gmail.com
          name: Hongxin Sun
        maturity: alpha
        provider:
          name: OceanBase
          url: https://oceanbase.com
        relatedImages:
        - quay.io/oceanbasedev/oceanbase-operator@sha256:b11fe60bb8de545d3ed105acc50f2d6c5d6aa8093f3ae4ca7ebce58b1e5e7b48
        - quay.io/brancz/kube-rbac-proxy@sha256:18f12fc3a0085bbc303cfbd2ad1151127c1c959c4641e36027547b6b66eb4c92
        version: 2.2.2
      entries:
      - name: oceanbase-operator.v2.2.2
        version: 2.2.2
      name: alpha
    - currentCSV: oceanbase-operator.v2.2.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Full Lifecycle
          createdAt: "2024-08-14T16:21:06Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: OBClusterOperation
            name: obclusteroperations.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBCluster
            name: obclusters.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBParameter
            name: obparameters.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBResourceRescue
            name: obresourcerescues.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBServer
            name: observers.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantBackupPolicy
            name: obtenantbackuppolicies.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantBackup
            name: obtenantbackups.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantOperation
            name: obtenantoperations.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantRestore
            name: obtenantrestores.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenant
            name: obtenants.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBZone
            name: obzones.oceanbase.oceanbase.com
            version: v1alpha1
        description: Ob Operator description. TODO.
        displayName: Ob Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - oceanbase-operator
        - ob-operator
        links:
        - name: OceanBase Operator
          url: https://oceanbase.github.io/ob-operator
        maintainers:
        - email: hongxin.sun.vip@gmail.com
          name: Hongxin Sun
        maturity: alpha
        provider:
          name: OceanBase
          url: https://oceanbase.com
        relatedImages:
        - quay.io/oceanbasedev/oceanbase-operator@sha256:b11fe60bb8de545d3ed105acc50f2d6c5d6aa8093f3ae4ca7ebce58b1e5e7b48
        - quay.io/brancz/kube-rbac-proxy@sha256:18f12fc3a0085bbc303cfbd2ad1151127c1c959c4641e36027547b6b66eb4c92
        version: 2.2.2
      entries:
      - name: oceanbase-operator.v2.2.2
        version: 2.2.2
      name: stable
    defaultChannel: stable
    packageName: oceanbase-operator
    provider:
      name: OceanBase
      url: https://oceanbase.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.2.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"customImageNames":
            false,"externalArangoDBName": "arangodb","repository": "registry.connect.redhat.com/turbonomic","tag":"8.3.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/turbonomic/t8c-operator@sha256:d5164e6dd24a0feff10cab62530cafff1bc92d0e8ec2d16c6f4d880c6e2f50bd
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/t8c-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/t8c-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/turbonomic/t8c-operator@sha256:d5164e6dd24a0feff10cab62530cafff1bc92d0e8ec2d16c6f4d880c6e2f50bd
        version: 42.2.0
      entries:
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: stable
    defaultChannel: stable
    packageName: t8c-certified-rhmp
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ptp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ptp-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "NodePtpDevice",
                "metadata": {
                  "name": "node.example.com"
                },
                "spec": {}
              },
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "PtpConfig",
                "metadata": {
                  "name": "example-ptpconfig"
                },
                "spec": {
                  "profile": [
                    {
                      "interface": "ens787f1",
                      "name": "profile1",
                      "phc2sysOpts": "-a -r",
                      "ptp4lOpts": "-s -2",
                      "ptpClockThreshold": {}
                    }
                  ],
                  "recommend": [
                    {
                      "match": [
                        {
                          "nodeLabel": "node-role.kubernetes.io/worker=",
                          "nodeName": "node.example.com"
                        }
                      ],
                      "priority": 4,
                      "profile": "profile1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "PtpOperatorConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "daemonNodeSelector": {
                    "feature.node.kubernetes.io/ptp-capable": "yes",
                    "node-role.kubernetes.io/worker": ""
                  },
                  "ptpEventConfig": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-ptp-rhel9-operator@sha256:012334e6624d5440439824454bda1d1dc79804938174232a3b470176bf462b5e
          createdAt: "2024-03-13"
          description: This software enables configuration of Precision Time Protocol(PTP)
            on Kubernetes. It detects hardware capable PTP devices on each node, and
            configures linuxptp processes such as ptp4l, phc2sys and timemaster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.16.0-202509240123'
          operatorframework.io/suggested-namespace: openshift-ptp
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/ptp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodePtpDevice is the Schema for the nodeptpdevices API
            displayName: Node Ptp Device
            kind: NodePtpDevice
            name: nodeptpdevices.ptp.openshift.io
            version: v1
          - description: PtpConfig is the Schema for the ptpconfigs API
            displayName: Ptp Config
            kind: PtpConfig
            name: ptpconfigs.ptp.openshift.io
            version: v1
          - description: PtpOperatorConfig is the Schema for the ptpoperatorconfigs
              API
            displayName: Ptp Operator Config
            kind: PtpOperatorConfig
            name: ptpoperatorconfigs.ptp.openshift.io
            version: v1
        description: |
          # Precision Time Protocol (PTP) Operator for Openshift

          ## Introdution
          PTP Operator manages cluster wide PTP configuration.
          This operator has to run in namespace 'openshift-ptp'. An Operator Group is also required to install this operator:
          ```
          $ oc create -f - <<EOF
          apiVersion: v1
          kind: Namespace
          metadata:
            name: openshift-ptp
            labels:
              name: openshift-ptp
              openshift.io/cluster-monitoring: "true"
          EOF

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: ptp-operators
            namespace: openshift-ptp
          spec:
            targetNamespaces:
            - openshift-ptp
          EOF
          ```
        displayName: PTP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ptp
        links:
        - name: Documentation operator
          url: https://github.com/openshift/ptp-operator
        - name: Documentation operand
          url: https://github.com/openshift/linuxptp-daemon
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-cloud-event-proxy-rhel9@sha256:3fa2aad8cf026eba586645c851f07a1d752c85b30795fdcf037b138ef3bbb6aa
        - registry.redhat.io/openshift4/ose-ptp-rhel9-operator@sha256:012334e6624d5440439824454bda1d1dc79804938174232a3b470176bf462b5e
        - registry.redhat.io/openshift4/ose-ptp-rhel9@sha256:93b4717337b551603cbb6147d66d7b70a5580332953ddf42304f7092d8739c83
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        version: 4.16.0-202509240123
      entries:
      - name: ptp-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: ptp-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: ptp-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: ptp-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: ptp-operator.v4.16.0-202506201205
        version: 4.16.0-202506201205
      - name: ptp-operator.v4.16.0-202505291140
        version: 4.16.0-202505291140
      - name: ptp-operator.v4.16.0-202505161704
        version: 4.16.0-202505161704
      - name: ptp-operator.v4.16.0-202505020146
        version: 4.16.0-202505020146
      - name: ptp-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: ptp-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: ptp-operator.v4.16.0-202502260004
        version: 4.16.0-202502260004
      - name: ptp-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: ptp-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: ptp-operator.v4.16.0-202501201505
        version: 4.16.0-202501201505
      - name: ptp-operator.v4.16.0-202501072135
        version: 4.16.0-202501072135
      - name: ptp-operator.v4.16.0-202412311307
        version: 4.16.0-202412311307
      - name: ptp-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: ptp-operator.v4.16.0-202412102004
        version: 4.16.0-202412102004
      - name: ptp-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: ptp-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: ptp-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: ptp-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: ptp-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: ptp-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: ptp-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: ptp-operator.v4.16.0-202409131735
        version: 4.16.0-202409131735
      - name: ptp-operator.v4.16.0-202409062036
        version: 4.16.0-202409062036
      - name: ptp-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: ptp-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: ptp-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: ptp-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: ptp-operator.v4.16.0-202407311708
        version: 4.16.0-202407311708
      - name: ptp-operator.v4.16.0-202407301837
        version: 4.16.0-202407301837
      - name: ptp-operator.v4.16.0-202407160236
        version: 4.16.0-202407160236
      - name: ptp-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: ptp-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: ptp-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: ptp-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: ptp-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: ptp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fossul Framework
      provider-url: https://github.com/fossul/fossul
    name: fossul-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fossul-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n    \"kind\":
            \"Backup\",\n    \"metadata\": {\n      \"name\": \"mariadb-sample\"\n
            \   },\n    \"spec\": {\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\"\n    }\n  },\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n    \"kind\":
            \"Restore\",\n    \"metadata\": {\n      \"name\": \"mariadb-sample\"\n
            \   },\n    \"spec\": {\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\",\n      \"workflow_id\": \"xxxx\"\n    }\n  },        \n  {\n
            \   \"apiVersion\": \"fossul.io/v1\",\n    \"kind\": \"BackupConfig\",\n
            \   \"metadata\": {\n      \"name\": \"mariadb-sample\"\n    },\n    \"spec\":
            {\n      \"container_platform\": \"openshift\",\n      \"operator_controlled\":
            true,\n      \"auto_discovery\": false,\n      \"deployment_name\": \"mariadb\",\n
            \     \"deployment_type\": \"DeploymentConfig\",\n      \"job_retention\":
            50,\n      \"overwrite_pcv_on_restore\": true,\n      \"policies\": [\n
            \       {\n          \"policy\": \"hourly\",\n          \"retentionNumber\":
            3\n        },\n        {\n          \"policy\": \"daily\",\n          \"retentionNumber\":
            10\n        }\n      ],\n      \"pvc_deletion_timeout\": 300,\n      \"restore_to_new_pvc\":
            false,\n      \"snapshot_timeout\": 180,\n      \"storage_plugin\": \"csi.so\",\n
            \     \"app_plugin\": \"mariadb.so\"\n    }\n  },\n  {\n    \"apiVersion\":
            \"fossul.io/v1\",\n    \"kind\": \"BackupSchedule\",\n    \"metadata\":
            {\n      \"name\": \"mariadb-sample\"\n    },\n    \"spec\": {\n      \"cron_schedule\":
            \"59 23 * * *\",\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\"\n\n    }\n  },\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n
            \   \"kind\": \"Fossul\",\n    \"metadata\": {\n      \"name\": \"fossul-sample\"\n
            \   },\n    \"spec\": {\n      \"enable_routes\": false,\n      \"enable_virtualmachines\":
            false,\n      \"container_platform\": \"openshift\"\n    }\n  }\n]"
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/fossul/fossul-operator:v0.6.0
          createdAt: "2022-01-25T20:09:00Z"
          description: Fossul is a application centric container native backup and
            restore framework for databases or applications
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/fossul/fossul
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages application backup configurations
            kind: BackupConfig
            name: backupconfigs.fossul.io
            version: v1
          - description: Manages application backups
            kind: Backup
            name: backups.fossul.io
            version: v1
          - description: Manages application restores
            kind: Restore
            name: restores.fossul.io
            version: v1
          - description: Manages application backup scheduling
            kind: BackupSchedule
            name: backupschedules.fossul.io
            version: v1
          - description: Manages deployment of the Fossul Framework
            kind: Fossul
            name: fossuls.fossul.io
            version: v1
        description: "## Fossul Framework\nApplication backup and restore framework
          for container native applications on Kubernetes and OpenShift. Fossul provides
          application centric backup and recovery enabling integration with the application
          creating an application consistent backup. Since there are many applications
          and providers for storage or archive Fossul enables plugins that allow proper
          integration. Fossul provides plugins for postgresql, mariadb, mysql and
          mongo as well as plugin to integrate with CSI snapshots. Plugins can be
          written in any language. The Fossul operator deploys Fossul via the Fossul
          custom resource and additionally provides BackupConfig, Backup, BackupSchedule
          and Restore custom resources. Each user can configure and manage their own
          backups with simple custom resources that can be part of CI/CD or source
          control. Users can also opt to manage backups through Fossul CLI/API instead
          of the custom resources provided by the operator.\n\n## Installation\nIt
          is recommended to install the operator into the openshift-operators namespace
          which is the default location for operators. Once the operator is installed
          create a new project called ```fossul``` and create the Fossul custom resource
          under the fossul namespace. This will deploy the Fossul Framework. You can
          choose another namespace for the Fossul Framework but then also need to
          provide the optional 'fossul_namespace' parameter with the spec for all
          the additional custom resources. This is required so the Fossul operator
          can communicate with Fossul.\n\nOnce Fossul is deployed you can manage backups
          either through custom resources provided by the operator or the Fossul CLI/API.
          If using the customer resources, first create a BackupConfig custom resource
          for every application you want to backup. Once that is done, you can optionally
          create a backup by simply creating a Backup custom resource. You can also
          create a BackupSchedule custom resource using cron syntax which will schedule
          your backup to occur on a regular interval or do so via a Kubernetes job.
          Finally if a backup has been created you can perform a restore through a
          custom resource as well, providing the workflow id from the completed backup.
          If you are creating backups through the custom resource the workflow id
          will be appended to the spec once the backup completes. If you are using
          the Fossul CLI/API to create backups you will need to get the workflow id
          through the Fossul CLI/API directly using the backupList action.\n\nFor
          users of the cli, a cli pod is deployed with credentials and if desired
          you can create or manage backups from within that pod. To use the API you
          need the credentials stored in the fossul secret. The cli has already been
          pre-configured with these credentials. The API can also be optionally exposed
          via routes and API documentation is under ```/api/v1/index.html```.\n\n##
          Quick Setup\nWhen creating Fossul custom resources generally you will only
          need to update the ```deployment_name``` and ```deployment_type``` (DeploymentConfig,
          Deployment or VirtualMachine). All custom resources must be created in the
          namespace where the database or application exists. \n\n### Deploy Fossul
          in fossul namespace\n```$ oc new-project fossul```\n<pre>\n$ vi fossul.yaml\nkind:
          Fossul\napiVersion: fossul.io/v1\nmetadata:\n  name: fossul-sample\n  namespace:
          fossul\nspec: {}\n</pre>\n```$ oc create -f fossul.yaml```\n\n### Deploy
          Mariadb Database\nMake sure you create all Fossul custom resources in the
          same namespace as the database, application or virtual machine\n\n### Create
          MariaDB Fossul BackupConfig\nCreates a backup configuration for MariaDB
          or MySQL databases\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true\n  auto_discovery:
          false\n  deployment_name: mariadb\n  deployment_type: DeploymentConfig\n
          \ job_retention: 50\n  overwrite_pcv_on_restore: true\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc: false\n  snapshot_timeout:
          180\n  storage_plugin: csi.so\n  app_plugin: mariadb.so \n</pre>\n```$ oc
          create -f backupconfig.yaml```\n\n### Create PostgreSQL Fossul BackupConfig\nCreates
          a backup config for PostgreSQL databases, you need to ensure the user defined
          in secret has backup permissions\n\n```$ vi backupconfig.yaml```\n<pre>\nkind:
          BackupConfig\napiVersion: fossul.io/v1\nmetadata:\n  name: postgresql-sample\n
          \ namespace: databases\nspec:\n  container_platform: openshift\n  operator_controlled:
          true    \n  auto_discovery: false\n  deployment_name: postgresql \n  deployment_type:
          DeploymentConfig\n  job_retention: 50\n  overwrite_pcv_on_restore: true\n
          \ policies:\n    - policy: hourly\n      retentionNumber: 3\n    - policy:
          daily\n      retentionNumber: 10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc:
          false\n  snapshot_timeout: 180\n  storage_plugin: csi.so\n  app_plugin:
          postgres.so \n</pre>\n```$ oc create -f backupconfig.yaml```\n\n### Create
          Kubevirt Fossul BackupConfig\nCreates a backup configuration for virtual
          machines\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: rhel8-vm-sample\n  namespace: virtualmachines\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true    \n  auto_discovery:
          false\n  deployment_name: rhel8-vm\n  deployment_type: VirtualMachine\n
          \ job_retention: 50\n  overwrite_pcv_on_restore: true\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc: false\n  snapshot_timeout:
          180\n  storage_plugin: csi.so\n  app_plugin: kubevirt.so \n</pre>\n```$
          oc create -f backupconfig.yaml```\n\n### Create Tar Fossul BackupConfig\nCreates
          a backup config using tar to backup folders and files where csi snapshots
          are not desired\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: tar-sample\n  namespace: databases\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true    \n  auto_discovery:
          false\n  deployment_name: anyname\n  job_retention: 50\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  backup_source_paths: /opt/app-root/src/test\n  storage_plugin: tar.so\n</pre>\n```$
          oc create -f backupconfig.yaml```\n\n### Create BackupSchedule\nA backup
          can be scheduled per policy, defined by backup configuration using cron
          syntax\n\n```$ vi backupschedule.yaml```\n<pre>\nkind: BackupSchedule\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ cron_schedule: '59 23 * * *'\n  deployment_name: mariadb\n  policy: daily\n</pre>\n```$
          oc create -f backupschedule.yaml```\n\n### Create Backup\nA backup will
          be created according to policy and deleted either manually or as defined
          in retention policy\n\n```$ vi backup.yaml```\n<pre>\nkind: Backup\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ deployment_name: mariadb\n  policy: daily\n</pre>\n```$ oc create -f backup.yaml```\n\n###
          Perform Restore\nA restore requires a workflow_id located in the backup
          spec, upon a successful restore the custom resource will deleted automatically\n\n```$
          vi restore.yaml```\n<pre>\nkind: Restore\napiVersion: fossul.io/v1\nmetadata:\n
          \ name: mariadb-sample\n  namespace: databases\nspec:\n  deployment_name:
          mariadb\n  policy: daily\n  workflow_id: xxxx\n</pre>\n```$ oc create -f
          restore.yaml```\n"
        displayName: Fossul Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - application
        - consistent
        links:
        - name: Fossul Framework Community
          url: https://github.com/fossul/fossul
        maintainers:
        - email: ktenzer@redhat.com
          name: Keith Tenzer
        maturity: alpha
        provider:
          name: Fossul Framework
          url: https://github.com/fossul/fossul
        relatedImages:
        - quay.io/fossul/fossul-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: fossul-operator.v0.6.0
        version: 0.6.0
      - name: fossul-operator.v0.5.0
        version: 0.5.0
      - name: fossul-operator.v0.4.0
        version: 0.4.0
      - name: fossul-operator.v0.3.0
        version: 0.3.0
      - name: fossul-operator.v0.2.0
        version: 0.2.0
      - name: fossul-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: fossul-operator
    provider:
      name: Fossul Framework
      url: https://github.com/fossul/fossul
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Citrus Framework
      provider-url: ""
    name: yaks
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: yaks-operator.v0.17.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "yaks.citrusframework.org/v1alpha1",
                "kind": "Instance",
                "metadata": {
                  "labels": {
                    "app": "yaks"
                  },
                  "name": "example-instance"
                },
                "spec": {
                  "operator": {
                    "global": true
                  }
                }
              },
              {
                "apiVersion": "yaks.citrusframework.org/v1alpha1",
                "kind": "Test",
                "metadata": {
                  "labels": {
                    "app": "yaks"
                  },
                  "name": "example-test"
                },
                "spec": {
                  "source": {
                    "content": "Feature: hello world\n\n  Scenario: print slogan\n    Given YAKS does Cloud-Native BDD testing\n    Then YAKS rocks!",
                    "language": "feature",
                    "name": "example.feature"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/citrusframework/yaks:0.17.1
          createdAt: "2023-04-26T20:37:31Z"
          description: YAKS is a platform to enable Cloud Native BDD testing on Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["instances.yaks.citrusframework.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/citrusframework/yaks
          support: Citrus Framework
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance is the Schema for the yaks instance
            displayName: Instance
            kind: Instance
            name: instances.yaks.citrusframework.org
            version: v1alpha1
          - description: Test is the Schema for the tests API
            displayName: Test
            kind: Test
            name: tests.yaks.citrusframework.org
            version: v1alpha1
        description: |
          YAKS
          ====

          YAKS is a platform to enable Cloud Native BDD testing on Kubernetes.

          ## Running a YAKS test

          With the YAKS operator installed, you can run tests by creating a `Test` custom resource on the cluster.

          Tests are defined using [Gherkin](https://cucumber.io/docs/gherkin/) syntax. YAKS provides a set of predefined steps which
          help to connect with different messaging transports (Http REST, JMS, Kafka, Knative eventing) and verify responses with
          assertions on message header and body content.

          The example below defines a simple test resource.

          ```
          apiVersion: yaks.citrusframework.org/v1alpha1
          kind: Test
          metadata:
            name: example
          spec:
            source:
              name: example.feature
              content: |-
                Feature: hello world

                  Scenario: print slogan
                    Given YAKS does Cloud-Native BDD testing
                    Then YAKS rocks!
          ```

          Refer to the [YAKS repository](https://github.com/citrusframework/yaks) for more complex examples.
        displayName: YAKS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - yaks
        - testing
        - microservices
        - knative
        - kafka
        - serverless
        - camel
        - camel-k
        links:
        - name: Project page
          url: https://citrusframework.org/
        - name: YAKS source code repository
          url: https://github.com/citrusframework/yaks
        maintainers:
        - email: cdeppisch@redhat.com
          name: Christoph Deppisch
        - email: nferraro@redhat.com
          name: Nicola Ferraro
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Citrus Framework
        relatedImages:
        - docker.io/citrusframework/yaks:0.17.1
        version: 0.17.1
      entries:
      - name: yaks-operator.v0.17.1
        version: 0.17.1
      - name: yaks-operator.v0.17.0
        version: 0.17.0
      - name: yaks-operator.v0.16.0
        version: 0.16.0
      - name: yaks-operator.v0.15.1
        version: 0.15.1
      - name: yaks-operator.v0.15.0
        version: 0.15.0
      - name: yaks-operator.v0.14.3
        version: 0.14.3
      - name: yaks-operator.v0.14.2
        version: 0.14.2
      - name: yaks-operator.v0.14.1
        version: 0.14.1
      - name: yaks-operator.v0.13.0
        version: 0.13.0
      - name: yaks-operator.v0.12.2
        version: 0.12.2
      - name: yaks-operator.v0.11.0
        version: 0.11.0
      - name: yaks-operator.v0.10.0
        version: 0.10.0
      - name: yaks-operator.v0.9.0
        version: 0.9.0
      - name: yaks-operator.v0.8.0
        version: 0.8.0
      - name: yaks-operator.v0.7.0
        version: 0.7.0
      - name: yaks-operator.v0.6.0
        version: 0.6.0
      - name: yaks-operator.v0.5.0
        version: 0.5.0
      - name: yaks-operator.v0.4.0
        version: 0.4.0
      name: alpha
    defaultChannel: alpha
    packageName: yaks
    provider:
      name: Citrus Framework
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator.v1.5.0
        version: 1.5.0
      - name: dynatrace-operator.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetScaler
      provider-url: https://netscaler.com
    name: netscaler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netscaler-operator.v3.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCloudController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalercloudcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "azure": {
                    "clientsecret": "",
                    "image": "quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb",
                    "location": "",
                    "pullPolicy": "IfNotPresent",
                    "resourcegroupname": "",
                    "resources": {},
                    "subnetname": "",
                    "subscriptionid": "",
                    "tenantid": "",
                    "vnetname": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCpxWithIngressController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalercpxwithingresscontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "admSettings": {
                    "admIP": "",
                    "bandWidth": 1000,
                    "bandWidthLicense": false,
                    "cpxCores": 1,
                    "licenseEdition": "PLATINUM",
                    "licenseServerIP": "",
                    "licenseServerPort": 27000,
                    "loginSecret": "",
                    "platform": false,
                    "vCPULicense": false
                  },
                  "affinity": {},
                  "analyticsConfig": {
                    "distributedTracing": {
                      "enable": false,
                      "samplingRate": 100
                    },
                    "endpoint": {
                      "metrics": {
                        "service": ""
                      },
                      "transactions": {
                        "service": ""
                      }
                    },
                    "required": false,
                    "timeseries": {
                      "auditLogs": {
                        "enable": false
                      },
                      "events": {
                        "enable": false
                      },
                      "metrics": {
                        "enable": false,
                        "enableNativeScrape": false,
                        "exportFrequency": 30,
                        "mode": "avro",
                        "schemaFile": "schema.json"
                      },
                      "port": 5563
                    },
                    "transactions": {
                      "enable": false,
                      "port": 5557
                    }
                  },
                  "exporter": {
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4",
                    "ports": {
                      "containerPort": 8888
                    },
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {}
                  },
                  "extraVolumes": [],
                  "imagePullSecrets": [],
                  "ingressController": {
                    "defaultSSLCertSecret": "",
                    "defaultSSLSNICertSecret": "",
                    "disableAPIServerCertVerify": false,
                    "disableOpenshiftRoutes": true,
                    "enableLivenessProbe": true,
                    "enableReadinessProbe": true,
                    "entityPrefix": "k8s",
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "ingressClass": [],
                    "ipam": false,
                    "jsonLog": false,
                    "kubernetesURL": "",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "logLevel": "INFO",
                    "logProxy": "",
                    "namespaceLabels": "",
                    "openshift": true,
                    "optimizeEndpointBinding": false,
                    "profileHttpFrontend": {},
                    "profileSslFrontend": {},
                    "profileTcpFrontend": {},
                    "pullPolicy": "IfNotPresent",
                    "rbacRole": false,
                    "readinessProbe": {
                      "failureThreshold": 3,
                      "initialDelaySeconds": 50,
                      "periodSeconds": 60,
                      "successThreshold": 1
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "128m",
                        "memory": "500Mi"
                      }
                    },
                    "routeLabels": "",
                    "setAsDefaultIngressClass": false,
                    "updateIngressStatus": true
                  },
                  "netscalerCpx": {
                    "bgpSettings": {
                      "bgpConfig": [
                        {
                          "bgpRouter": {
                            "localAS": 100,
                            "neighbor": [
                              {
                                "ASOriginationInterval": 10,
                                "address": "",
                                "advertisementInterval": 10,
                                "remoteAS": 100
                              }
                            ]
                          }
                        }
                      ],
                      "bgpPort": 179
                    },
                    "cpxBgpRouter": false,
                    "cpxCommands": [],
                    "cpxLicenseAggregator": "",
                    "cpxShellCommands": [],
                    "daemonSet": false,
                    "enableLivenessProbe": true,
                    "enableStartupProbe": true,
                    "extraVolumeMounts": [],
                    "hostname": "",
                    "image": "quay.io/netscaler/netscaler-cpx@sha256:232596eaacda68c0e2e210f26b64cedc8d70ae12c5e8ae2e1a66896a479ae4f8",
                    "ingressIP": "",
                    "livenessProbe": {
                      "failureThreshold": 3,
                      "periodSeconds": 5,
                      "successThreshold": 1
                    },
                    "mgmtHttpPort": 9080,
                    "mgmtHttpsPort": 9443,
                    "nitroReadTimeout": 20,
                    "nsConfigDnsRec": false,
                    "nsCookieVersion": 0,
                    "nsDnsNameserver": "",
                    "nsEnableLabels": true,
                    "nsGateway": "192.168.1.1",
                    "nsHttp2ServerSide": "OFF",
                    "nsIP": "192.168.1.2",
                    "nsLbHashAlgo": {
                      "hashAlgorithm": "default",
                      "hashFingers": 500,
                      "required": false
                    },
                    "nsProtocol": "HTTP",
                    "nsSvcLbDnsRec": false,
                    "prometheusCredentialSecret": "",
                    "pullPolicy": "IfNotPresent",
                    "service": {
                      "annotations": {},
                      "spec": {}
                    },
                    "startupProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1
                    }
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "replicaCount": 1,
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerGslbController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalergslbcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "gslbController": {
                    "disableAPIServerCertVerify": false,
                    "entityPrefix": "",
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "kubernetesURL": "",
                    "localCluster": "",
                    "localRegion": "",
                    "logLevel": "Debug",
                    "openshift": true,
                    "pullPolicy": "IfNotPresent",
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "32m",
                        "memory": "128Mi"
                      }
                    },
                    "siteData": [
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteMask": "",
                        "siteName": "",
                        "sitePublicIp": "",
                        "siteRegion": ""
                      },
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteMask": "",
                        "siteName": "",
                        "sitePublicIp": "",
                        "siteRegion": ""
                      }
                    ]
                  },
                  "imagePullSecrets": [],
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsIP": "",
                    "nsPort": 443,
                    "nsProtocol": "HTTPS"
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerIngressController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscaleringresscontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "analyticsConfig": {
                    "distributedTracing": {
                      "enable": false,
                      "samplingRate": 100
                    },
                    "endpoint": {
                      "metrics": {
                        "service": ""
                      },
                      "transactions": {
                        "service": ""
                      }
                    },
                    "required": false,
                    "timeseries": {
                      "auditLogs": {
                        "enable": false
                      },
                      "events": {
                        "enable": false
                      },
                      "metrics": {
                        "enable": false,
                        "enableNativeScrape": false,
                        "exportFrequency": 50,
                        "mode": "avro",
                        "schemaFile": "schema.json"
                      },
                      "port": 50002
                    },
                    "transactions": {
                      "enable": false,
                      "port": 50005
                    }
                  },
                  "exporter": {
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4",
                    "ports": {
                      "containerPort": 8888
                    },
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {},
                    "serviceMonitorExtraLabels": {}
                  },
                  "extraVolumes": [],
                  "hostAliases": [],
                  "imagePullSecrets": [],
                  "ingressController": {
                    "clusterName": "",
                    "defaultSSLCertSecret": "",
                    "defaultSSLSNICertSecret": "",
                    "disableAPIServerCertVerify": false,
                    "disableOpenshiftRoutes": true,
                    "enableLivenessProbe": true,
                    "enableReadinessProbe": true,
                    "entityPrefix": "",
                    "extraVolumeMounts": [],
                    "ignoreNodeExternalIP": false,
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "ingressClass": [],
                    "ipam": false,
                    "jsonLog": false,
                    "kubernetesURL": "",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "logLevel": "INFO",
                    "logProxy": "",
                    "multiClusterPrefix": "",
                    "namespaceLabels": "",
                    "nodeLabels": "",
                    "nodeWatch": false,
                    "nsValidateCert": {},
                    "nsncPbr": false,
                    "openshift": true,
                    "optimizeEndpointBinding": false,
                    "podIPsforServiceGroupMembers": false,
                    "profileHttpFrontend": {},
                    "profileSslFrontend": {},
                    "profileTcpFrontend": {},
                    "pullPolicy": "IfNotPresent",
                    "rbacRole": false,
                    "readinessProbe": {
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 60,
                      "successThreshold": 1
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "32m",
                        "memory": "128Mi"
                      }
                    },
                    "routeLabels": "",
                    "serviceClass": [],
                    "setAsDefaultIngressClass": false,
                    "updateIngressStatus": true
                  },
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsConfigDnsRec": false,
                    "nsCookieVersion": 0,
                    "nsDnsNameserver": "",
                    "nsHttp2ServerSide": "OFF",
                    "nsIP": "",
                    "nsLbHashAlgo": {
                      "hashAlgorithm": "default",
                      "hashFingers": 256,
                      "required": false
                    },
                    "nsPort": 443,
                    "nsProtocol": "HTTPS",
                    "nsSNIPs": [],
                    "nsSvcLbDnsRec": false,
                    "nsVIP": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerIpamController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscaleripamcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "clusterName": "",
                  "image": "quay.io/netscaler/netscaler-ipam-controller@sha256:7c25b59a2e1df0e4f5615a779a9bea939c60a3eca065cb6c2295daf0d03831d8",
                  "localIPAM": {
                    "vipRange": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "reuseIngressVIP": true,
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerObservabilityExporter",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalerobservabilityexporter-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "elasticsearch": {
                    "enabled": false,
                    "server": "http://elasticsearch:9200"
                  },
                  "headless": false,
                  "image": "quay.io/netscaler/netscaler-observability-exporter@sha256:626e3a954c67bbb910477352dc740aaab21ff1c1cfb03354becf50beab17d4a4",
                  "jsonTransRateLimiting": {
                    "enabled": false,
                    "limit": 100,
                    "queueLimit": 1000,
                    "window": 5
                  },
                  "kafka": {
                    "auditLogs": "no",
                    "broker": "",
                    "dataFormat": "AVRO",
                    "enabled": false,
                    "events": "no",
                    "topic": "HTTP"
                  },
                  "nodePortRequired": false,
                  "nodeSelector": {},
                  "nsoeLogLevel": "INFO",
                  "nstracing": {
                    "enabled": false,
                    "server": "zipkin:9411/api/v1/spans"
                  },
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "replicas": 1,
                  "resources": {},
                  "splunk": {
                    "authToken": "",
                    "enabled": false,
                    "indexPrefix": "adc_nsoe",
                    "server": "http://splunk:8088"
                  },
                  "timeseries": {
                    "enabled": false
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-operator@sha256:e367c8c8ba6171e9ef5dd496592a5c62f4aa764f93d0cb78835aa29024c063d0
          createdAt: "2025-04-15T08:51:39Z"
          description: This is an operator to install Cloud Native portfolio provided
            by NetScaler
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler Cloud Controller for NetScaler VPX/MPX/SDX
              with Azure Redhat OpenShift in Azure.
            displayName: Netscaler Cloud Controller
            kind: NetscalerCloudController
            name: netscalercloudcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler CPX with Ingress Controller.
            displayName: Netscaler CPX With Ingress Controller
            kind: NetscalerCpxWithIngressController
            name: netscalercpxwithingresscontrollers.netscaler.com
            version: v1
          - description: To install NetScaler GSLB Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler GSLB Controller
            kind: NetscalerGslbController
            name: netscalergslbcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler Ingress Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler Ingress Controller
            kind: NetscalerIngressController
            name: netscaleringresscontrollers.netscaler.com
            version: v1
          - description: To install NetScaler IPAM Controller for NetScaler CPX/VPX/MPX/SDX.
            displayName: Netscaler IPAM Controller
            kind: NetscalerIpamController
            name: netscaleripamcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler Observability Exporter for NetScaler.
            displayName: Netscaler Observability Exporter
            kind: NetscalerObservabilityExporter
            name: netscalerobservabilityexporters.netscaler.com
            version: v1
        description: |
          NetScaler provides various products to empower Developer/DevOps managing your applications in OpenShift Cluster. Using this operator you can deploy below controllers for NetScaler appliance:
          1. [NetScaler Cloud Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. [NetScaler Ingress Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/unified-ingress-deployment)
          3. [NetScaler CPX with Ingress Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/two-tier-deployment)
          4. [NetScaler GSLB Controller ](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/gslb/gslb)
          5. [NetScaler Observability Exporter](https://docs.netscaler.com/en-us/netscaler-observability-exporter/deploy-coe)
          6. [NetScaler IPAM Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/ipam-installation.html)

          ### Installation
          Refer [installation instructions](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-operator-openshift).

          This operator version contains:
          1. NetScaler Ingress Controller version [2.2.10](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.2.10)
          2. NetScaler CPX version [14.1-25.111](https://docs.netscaler.com/en-us/updates?product=NetScaler-ADC-%28includes-NetScaler-Gateway%29&version=14.1&build=25.56)
          3. NetScaler GSLB Controller version [2.2.10](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.2.10)
          4. NetScaler Observability Exporter version [1.10.001](https://github.com/citrix/citrix-helm-charts/releases/tag/nsoe-v1.10.001)
          5. NetScaler Metrics Exporter [1.4.9](https://github.com/citrix/citrix-adc-metrics-exporter/releases/tag/1.4.9)
          6. NetScaler IPAM Controller [2.1.2](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.3.15)

          ### Support

          For debugging issues, please use [NetScaler Kubernetes Plugin](https://github.com/netscaler/modern-apps-toolkit/tree/main/netscaler-plugin).

          For questions and support, the following channels are available:
          - [NetScaler community portal](https://community.netscaler.com/s/)
          - [Contact us](https://www.netscaler.com/contact)
        displayName: NetScaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        - ADC
        - Kubernetes
        - OpenShift
        - Citrix
        - Controller
        - Observability
        - Metrices
        links:
        - name: Netscaler Operator
          url: https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-operator-openshift
        maintainers:
        - email: priyanka.sharma@cloud.com
          name: Priyanka Sharma
        - email: arijit.ray@cloud.com
          name: Arijit Ray
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: NetScaler
          url: https://netscaler.com
        relatedImages:
        - quay.io/netscaler/netscaler-cpx@sha256:232596eaacda68c0e2e210f26b64cedc8d70ae12c5e8ae2e1a66896a479ae4f8
        - quay.io/netscaler/netscaler-observability-exporter@sha256:626e3a954c67bbb910477352dc740aaab21ff1c1cfb03354becf50beab17d4a4
        - quay.io/netscaler/netscaler-ipam-controller@sha256:7c25b59a2e1df0e4f5615a779a9bea939c60a3eca065cb6c2295daf0d03831d8
        - quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f
        - quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:89f44c45373da73a51b0b9a02de7efe519dc2d5343b65fe382526cccc92f8f5c
        - registry.connect.redhat.com/citrix/netscaler-operator@sha256:e367c8c8ba6171e9ef5dd496592a5c62f4aa764f93d0cb78835aa29024c063d0
        - quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4
        version: 3.2.0
      entries:
      - name: netscaler-operator.v3.2.0
        version: 3.2.0
      - name: netscaler-operator.v3.1.2
        version: 3.1.2
      - name: netscaler-operator.v3.1.1
        version: 3.1.1
      - name: netscaler-operator.v3.0.1
        version: 3.0.1
      name: stable
    - currentCSV: netscaler-operator.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCloudController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netscaler-operator",
                    "app.kubernetes.io/instance": "nscc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscalercloudcontroller",
                    "app.kubernetes.io/part-of": "netscaler-operator"
                  },
                  "name": "nscc"
                },
                "spec": {
                  "azure": {
                    "clientsecret": "",
                    "image": "quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb",
                    "location": "",
                    "pullPolicy": "IfNotPresent",
                    "resourcegroupname": "",
                    "resources": {},
                    "subnetname": "",
                    "subscriptionid": "",
                    "tenantid": "",
                    "vnetname": ""
                  }
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerGslbController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netscaler-operator",
                    "app.kubernetes.io/instance": "nsgc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscalergslbcontroller",
                    "app.kubernetes.io/part-of": "netscaler-operator"
                  },
                  "name": "nsgc"
                },
                "spec": {
                  "gslbController": {
                    "acceptLicense": "Yes",
                    "cleanupGSLBSiteConfig": false,
                    "disableAPIServerCertVerify": false,
                    "entityPrefix": "",
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:88114c518ecceb54b000f38bfc330efe184bfeb5b5df0ee308950cf29e7a777f",
                    "imagePullSecrets": [],
                    "kubernetesURL": "",
                    "localCluster": "",
                    "localRegion": "",
                    "logLevel": "Debug",
                    "openshift": true,
                    "pullPolicy": "IfNotPresent",
                    "siteData": [
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteName": "",
                        "siteRegion": ""
                      },
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteName": "",
                        "siteRegion": ""
                      }
                    ]
                  },
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsIP": "",
                    "nsPort": 443,
                    "nsProtocol": "HTTPS"
                  },
                  "resources": {
                    "limits": {},
                    "requests": {
                      "cpu": "32m",
                      "memory": "128Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-operator@sha256:3f71f9a46d7b29e3da98a301965f635c03a022b5535b586d7d5c699341f68193
          createdAt: "2024-07-31T06:33:49Z"
          description: This is an operator to install Cloud Native portfolio provided
            by NetScaler
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler Cloud Controller for NetScaler VPX/MPX/SDX
              with Azure Redhat OpenShift in Azure.
            displayName: Netscaler Cloud Controller
            kind: NetscalerCloudController
            name: netscalercloudcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler GSLB Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler GSLB Controller
            kind: NetscalerGslbController
            name: netscalergslbcontrollers.netscaler.com
            version: v1
        description: |-
          NetScaler provides various products to empower Developer/DevOps managing your applications in OpenShift Cluster. Using this operator you can deploy below controllers for NetScaler appliance:
          1. [NetScaler Cloud Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. [NetScaler GSLB Controller ](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/gslb-openshift-operator)

          ### Installation
          Refer [installation instructions](https://github.com/netscaler/netscaler-k8s-ingress-controller/blob/master/docs/deploy/deploy-ns-operator.md).

          This operator version contains:
          1. NetScaler Cloud Controller version [1.2.0](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. NetScaler GSLB Controller version [1.43.7](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v1.43.7)

          ### Support

          For debugging issues, please use [NetScaler Kubernetes Plugin](https://github.com/netscaler/modern-apps-toolkit/tree/main/netscaler-plugin).

          For questions and support, the following channels are available:
          - [NetScaler community portal](https://community.netscaler.com/s/)
          - [Contact us](https://www.netscaler.com/contact)
        displayName: NetScaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        links:
        - name: Netscaler Operator
          url: https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/gslb-openshift-operator
        maintainers:
        - email: priyanka.sharma@cloud.com
          name: Priyanka Sharma
        - email: arijit.ray@cloud.com
          name: Arijit Ray
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: NetScaler
          url: https://netscaler.com
        relatedImages:
        - quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:88114c518ecceb54b000f38bfc330efe184bfeb5b5df0ee308950cf29e7a777f
        - quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:23405618050a4ceea8ea0c748cc7d9d16dbb80f10e4c39f7e82014a3ac80cec9
        - registry.connect.redhat.com/citrix/netscaler-operator@sha256:3f71f9a46d7b29e3da98a301965f635c03a022b5535b586d7d5c699341f68193
        version: 3.0.1
      entries:
      - name: netscaler-operator.v3.0.1
        version: 3.0.1
      name: tech-preview
    defaultChannel: stable
    packageName: netscaler-operator
    provider:
      name: NetScaler
      url: https://netscaler.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: open-liberty-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: open-liberty-operator.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4",
                  "expose": true,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/open-liberty-operator-controller@sha256:18f188e48725c785557c8edea46f69c464d6f81ab47860a3bb0ec978c20a26db
          createdAt: "2022-03-25T14:00:00Z"
          description: Deploy and manage applications running on Liberty
          olm.skipRange: <0.8.1
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          This advanced Operator can be used to deploy and manage Open Liberty applications with consistent, production-grade QoS. This operator is based on the [Runtime Component Operator](https://github.com/application-stacks/runtime-component-operator) and provides all of its capabilities in addition to Open Liberty specific features such as gathering traces and dumps (Day-2 operations) and easily configuring and managing the single sign-on information for your Open Liberty applications.

          Open Liberty Operator enables enterprise architects to govern the way their applications get deployed & managed in the cluster, while dramatically reducing the learning curve for developers to deploy into Kubernetes - allowing them to focus on writing the code!

          Here are some key features:

          #### Application Lifecyle
          You can deploy your Open Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the Operator will watch for any updates and will re-deploy the modified image.

          #### Custom RBAC
          This Operator is capable of using a custom ServiceAccount from the caller, allowing it to follow RBAC restrictions. By default it creates a ServiceAccount if one is not specified, which can also be bound with specific roles.

          #### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as: labels, annotations, and environment variables from a ConfigMap, a Secret or a value.

          #### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configuration, such as TLS settings, are also easily enabled.

          #### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto scaling to create (and delete) instances based on resource consumption.

          #### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you.
          This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart.
          We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          You can also easily configure and use a single storage for serviceability related Day-2 operations, such as gatherig server traces and dumps.

          #### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating kubernetes Secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to its required services without any intervention or interruption.

          #### Single Sign-On (SSO)
          Open Liberty provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows to easily configure and manage the single sign-on information for your applications.

          #### Exposing metrics to Prometheus
          Expose the Open Liberty application's metrics via the Prometheus Operator.
          You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `spec.monitoring` field to control configurations such as poll interval and security credentials.

          #### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply add the following configuration (to specify the volume size and the location to persist) to your OpenLibertyApplication CR:
          ``` storage: size: 2Gi mountPath: "/logs" ```

          #### Integration with OpenShift Serverless
          Deploy your serverless runtime component using a single toggle.  The Operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing your pod to automatically scale to 0 when it is idle.

          #### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and how they are connected.

          See our [**documentation**](https://github.com/OpenLiberty/open-liberty-operator/tree/main/doc/) for more information.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://github.com/OpenLiberty/open-liberty-operator/tree/main/doc
        - name: Open Liberty
          url: https://openliberty.io/
        - name: Open Liberty Container Images
          url: https://github.com/OpenLiberty/ci.docker
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: beta
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/open-liberty-operator-controller@sha256:18f188e48725c785557c8edea46f69c464d6f81ab47860a3bb0ec978c20a26db
        - registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4
        version: 0.8.1
      entries:
      - name: open-liberty-operator.v0.8.1
        version: 0.8.1
      - name: open-liberty-operator.v0.8.0
        version: 0.8.0
      name: beta2
    - currentCSV: open-liberty-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:d45c296d43dc4c4100f38bb736f073e04c8b4d72b68ddf1189e311aff8ad66e3
          createdAt: "2023-08-28T16:39:10Z"
          description: Deploy and manage containerized Liberty applications
          olm.skipRange: '>=0.8.0 <1.2.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:d45c296d43dc4c4100f38bb736f073e04c8b4d72b68ddf1189e311aff8ad66e3
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02
        version: 1.2.2
      entries:
      - name: open-liberty-operator.v1.2.2
        version: 1.2.2
      name: v1.2
    - currentCSV: open-liberty-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:d68bb66a8133702a198f7be655d7cff5f500bb620c460a81547971e00124ef73
          createdAt: "2024-08-01T20:19:13Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.3.3'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:d68bb66a8133702a198f7be655d7cff5f500bb620c460a81547971e00124ef73
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91
        version: 1.3.3
      entries:
      - name: open-liberty-operator.v1.3.3
        version: 1.3.3
      name: v1.3
    - currentCSV: open-liberty-operator.v1.4.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:67b5b46dfed008dab4ea7fd176d3bc31dae8940d72194a10aebb98b1dea9043e
          createdAt: "2025-07-18T20:45:06Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.4.4'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:67b5b46dfed008dab4ea7fd176d3bc31dae8940d72194a10aebb98b1dea9043e
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3
        version: 1.4.4
      entries:
      - name: open-liberty-operator.v1.4.4
        version: 1.4.4
      name: v1.4
    - currentCSV: open-liberty-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyPerformanceData",
                "metadata": {
                  "name": "openliberty-performance-data-sample"
                },
                "spec": {
                  "interval": 30,
                  "podName": "Specify_Pod_Name_Here",
                  "timespan": 240
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:c3b031fe14054f65d6612cc505b7681336873dc32ce1091451d8c7799034bc83
          createdAt: "2025-09-16T03:08:31Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.5.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server performance data
            displayName: OpenLibertyPerformanceData
            kind: OpenLibertyPerformanceData
            name: openlibertyperformancedata.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:c3b031fe14054f65d6612cc505b7681336873dc32ce1091451d8c7799034bc83
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8
        version: 1.5.0
      entries:
      - name: open-liberty-operator.v1.5.0
        version: 1.5.0
      name: v1.5
    defaultChannel: v1.5
    packageName: open-liberty-certified
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: RabbitMQ
      provider-url: https://rabbitmq.com
    name: rabbitmq-cluster-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-cluster-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rabbitmq.com/v1beta1",
                "kind": "RabbitmqCluster",
                "metadata": {
                  "name": "sample",
                  "namespace": "rabbitmq-system"
                },
                "spec": {
                  "replicas": 1,
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-07-10T12:52:23Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RabbitmqCluster is the Schema for the RabbitmqCluster API.
              Each instance of this object corresponds to a single RabbitMQ cluster.
            displayName: Rabbitmq Cluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: RabbitMQ Cluster Operator
        displayName: RabbitMQ Cluster Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        links:
        - name: Cluster Operator
          url: https://github.com/rabbitmq/cluster-operator
        maturity: alpha
        provider:
          name: RabbitMQ
          url: https://rabbitmq.com
        relatedImages:
        - registry.redhat.io/rhoso-operators/rabbitmq-cluster-rhel9-operator@sha256:e4b198236d70b7dfdd29d8ed45058dec3335b3530e0b676fcc862bfec2efce79
        version: 1.0.6
      entries:
      - name: rabbitmq-cluster-operator.v1.0.6
        version: 1.0.6
      - name: rabbitmq-cluster-operator.v1.0.5
        version: 1.0.5
      - name: rabbitmq-cluster-operator.v1.0.4
        version: 1.0.4
      - name: rabbitmq-cluster-operator.v1.0.3
        version: 1.0.3
      - name: rabbitmq-cluster-operator.v1.0.2
        version: 1.0.2
      - name: rabbitmq-cluster-operator.v1.0.1
        version: 1.0.1
      - name: rabbitmq-cluster-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: rabbitmq-cluster-operator
    provider:
      name: RabbitMQ
      url: https://rabbitmq.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-route53resolver-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-route53resolver-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "route53resolver.services.k8s.aws/v1alpha1",
                "kind": "ResolverEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "route53resolver.services.k8s.aws/v1alpha1",
                "kind": "ResolverRule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/route53resolver-controller:1.1.1
          createdAt: "2025-09-20T01:21:08Z"
          description: AWS Route53Resolver controller is a service controller for
            managing Route53Resolver resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ResolverEndpoint represents the state of an AWS route53resolver
              ResolverEndpoint resource.
            displayName: ResolverEndpoint
            kind: ResolverEndpoint
            name: resolverendpoints.route53resolver.services.k8s.aws
            version: v1alpha1
          - description: ResolverRule represents the state of an AWS route53resolver
              ResolverRule resource.
            displayName: ResolverRule
            kind: ResolverRule
            name: resolverrules.route53resolver.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Route53Resolver resources in AWS from within your Kubernetes cluster.

          **About Amazon Route53Resolver**

          Amazon Route 53 Resolver responds recursively to DNS queries from AWS resources for public records, Amazon VPC-specific DNS names, and Amazon Route 53 private hosted zones, and is available by default in all VPCs.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Route53Resolver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - route53resolver
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Route53Resolver Developer Resources
          url: https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/resolver.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: route53resolver maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/route53resolver-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-route53resolver-controller.v1.1.1
        version: 1.1.1
      - name: ack-route53resolver-controller.v1.0.15
        version: 1.0.15
      - name: ack-route53resolver-controller.v1.0.14
        version: 1.0.14
      - name: ack-route53resolver-controller.v1.0.13
        version: 1.0.13
      - name: ack-route53resolver-controller.v1.0.11
        version: 1.0.11
      - name: ack-route53resolver-controller.v1.0.10
        version: 1.0.10
      - name: ack-route53resolver-controller.v1.0.9
        version: 1.0.9
      - name: ack-route53resolver-controller.v1.0.7
        version: 1.0.7
      - name: ack-route53resolver-controller.v1.0.6
        version: 1.0.6
      - name: ack-route53resolver-controller.v1.0.5
        version: 1.0.5
      - name: ack-route53resolver-controller.v1.0.3
        version: 1.0.3
      - name: ack-route53resolver-controller.v1.0.2
        version: 1.0.2
      - name: ack-route53resolver-controller.v1.0.0
        version: 1.0.0
      - name: ack-route53resolver-controller.v0.0.11
        version: 0.0.11
      - name: ack-route53resolver-controller.v0.0.10
        version: 0.0.10
      - name: ack-route53resolver-controller.v0.0.9
        version: 0.0.9
      - name: ack-route53resolver-controller.v0.0.8
        version: 0.0.8
      - name: ack-route53resolver-controller.v0.0.7
        version: 0.0.7
      - name: ack-route53resolver-controller.v0.0.4
        version: 0.0.4
      - name: ack-route53resolver-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-route53resolver-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Weaveworks
      provider-url: ""
    name: tf-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tf-controller.v0.9.0-rc.8
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"infra.contrib.fluxcd.io/v1alpha1\",\n
            \   \"kind\": \"Terraform\",\n    \"metadata\": {\n        \"name\": \"helloworld-tf\",\n
            \       \"namespace\": \"flux-system\"\n    },\n    \"spec\": {\n        \"interval\":
            \"5m\",\n        \"approvePlan\": \"auto\",\n        \"path\": \"./\",\n
            \       \"sourceRef\": {\n          \"kind\": \"GitRepository\",\n          \"name\":
            \"helloworld\",\n          \"namespace\": \"flux-system\"\n        }      \n
            \   }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/weaveworks/tf-controller:v0.9.0-rc.8
          description: Weave TF-controller is a Flux controller that brings GitOps
            to Terraform resources.
          operatorframework.io/suggested-namespace: flux-system
          support: Weaveworks
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Terraform
            displayName: Terraform
            kind: Terraform
            name: terraforms.infra.contrib.fluxcd.io
            version: v1alpha1
        description: "**Weave GitOps Terraform Controller** (aka Weave TF-controller)
          is a Flux controller for reconciling Terraform resources in the GitOps way.
          \nWith the power of Flux together with Terraform, TF-controller allows you
          to GitOps-ify infrastructure, \nand application resources, in the Kubernetes
          and Terraform universe, together at your own pace.\n\nTF-controller offers
          many GitOps models:\n  1. **GitOps Automation Model:** GitOps your Terraform
          resources from the provision steps to the enforcement steps, like a whole
          EKS cluster.\n  2. **Hybrid GitOps Automation Model:** GitOps parts of your
          existing infrastructure resources. For example, you have an existing EKS
          cluster. You can choose to GitOps only its nodegroup, or its security group.\n
          \ 3. **State Enforcement Model:** You have a TFSTATE file, and you'd like
          to use GitOps enforce it, without changing anything else.\n  4. **Drift
          Detection Model:** You have a TFSTATE file, and you'd like to use GitOps
          just for drift detection, so you can decide to do things later when a drift
          occurs.\n\nFeatures\n  * GitOps Automation for Terraform\n  * Drift detection\n
          \ * Plan and Manual Approve"
        displayName: Weave GitOps Terraform Controller
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - weave
        - flux
        - cd
        - continuous delivery
        - terraform
        - gitops
        links:
        - name: Weave GitOps Terraform Controller
          url: https://github.com/weaveworks/tf-controller
        - name: Documentation
          url: https://weaveworks.github.io/tf-controller/
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Weaveworks
        relatedImages:
        - ghcr.io/weaveworks/tf-controller:v0.9.0-rc.8
        version: 0.9.0-rc.8
      entries:
      - name: tf-controller.v0.9.0-rc.8
        version: 0.9.0-rc.8
      - name: tf-controller.v0.9.0-rc.3
        version: 0.9.0-rc.3
      name: stable
    defaultChannel: stable
    packageName: tf-controller
    provider:
      name: Weaveworks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Randoli
      provider-url: https://www.randoli.io
    name: app-insights-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: app-insights-operator.v1.17.5-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "appinsights.randoli/v1beta1",
                "kind": "RandoliAgent",
                "metadata": {
                  "name": "randoli",
                  "namespace": "randoli-agents"
                },
                "spec": {
                  "costManagement": {
                    "opencost": {
                      "exporter": {
                        "image": {
                          "tag": "1.112.1"
                        }
                      }
                    },
                    "vpaOperator": {
                      "enabled": false
                    }
                  },
                  "logs": {
                    "vector": {
                      "agent": {
                        "containerSecurityContext": {
                          "privileged": true
                        }
                      }
                    }
                  },
                  "network": {
                    "standaloneConsole": {
                      "enable": true
                    }
                  },
                  "observability": {
                    "netobserv": {
                      "enabled": false
                    },
                    "otel": {
                      "operator": {
                        "enabled": false
                      }
                    }
                  },
                  "resources": {
                    "limits": {
                      "memory": "2Gi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "512Mi"
                    }
                  },
                  "securityContext": {
                    "fsGroup": 1001
                  },
                  "tags": {
                    "costManagement": false,
                    "observability": false
                  }
                }
              }
            ]
          categories: Developer Tools
          containerImage: docker.io/randoli/app-insights-operator:1.17.5-4@sha256:22e7150c29557f2a71186389e5d7d60bfe82a4ea022e1403033c6ad61bc95f7a
          createdAt: "2025-09-23T19:45:43Z"
          description: Randoli Agent connects OpenShift clusters to Randoli Observability
            & Cost Management.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[]'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Randoli
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: RandoliAgent
            name: randoliagents.appinsights.randoli
            version: v1beta1
        description: |-
          Randoli Agent connects OpenShift clusters to Randoli Observability & Cost Management platforms.

          This operator provides a container-based agent to be installed in your OpenShift clusters for connecting to Randoli Observability & Cost Management platforms

          **Prerequisites**

          If you are using the Randoli Observability platform, you will need to install the following Operators:
          - Red Hat build of OpenTelemetry Operator
          - Network Observability

          If you are using the Randoli Cost Management platform, you will need to install the following Operator:
          - VerticalPodAutoscaler

          For additional information please use our Openshift installation guide here: [https://docs.insights.randoli.io/agent/openshift](https://docs.insights.randoli.io/agent/openshift)

           **Randoli Observability for Kubernetes**

          [Product Page](https://www.randoli.io/product/kubernetes-observability)

          **Randoli Cost Management for Kubernetes**

          [Product Page](https://www.randoli.io/product/kubernetes-cost-management)
        displayName: Randoli Agent
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - developer tools
        - insights
        - monitoring
        links:
        - name: Randoli Observability for Kubernetes
          url: https://www.randoli.io/product/kubernetes-observability
        - name: Randoli Cost Management for Kubernetes
          url: https://www.randoli.io/product/kubernetes-cost-management
        maintainers:
        - email: contact@randoli.ca
          name: Randoli Inc.
        maturity: beta
        minKubeVersion: 1.24.0
        provider:
          name: Randoli
          url: https://www.randoli.io
        relatedImages:
        - docker.io/randoli/app-insights-operator:1.17.5-4@sha256:22e7150c29557f2a71186389e5d7d60bfe82a4ea022e1403033c6ad61bc95f7a
        version: 1.17.5-4
      entries:
      - name: app-insights-operator.v1.17.5-4
        version: 1.17.5-4
      - name: app-insights-operator.v1.17.5-3
        version: 1.17.5-3
      - name: app-insights-operator.v1.17.5-2
        version: 1.17.5-2
      - name: app-insights-operator.v1.17.5-1
        version: 1.17.5-1
      - name: app-insights-operator.v1.17.3-1
        version: 1.17.3-1
      - name: app-insights-operator.v1.17.2-1
        version: 1.17.2-1
      - name: app-insights-operator.v1.17.1-1
        version: 1.17.1-1
      - name: app-insights-operator.v1.17.0-2
        version: 1.17.0-2
      - name: app-insights-operator.v1.17.0-1
        version: 1.17.0-1
      name: stable
    defaultChannel: stable
    packageName: app-insights-operator
    provider:
      name: Randoli
      url: https://www.randoli.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ProphetStor Data Services, Inc.
      provider-url: ""
    name: federatorai-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: federatorai.v4.7.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "federatorai.containers.ai/v1alpha1",
                "kind": "AlamedaService",
                "metadata": {
                  "name": "my-alamedaservice"
                },
                "spec": {
                  "alamedaAi": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "10Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "alamedaInfluxdb": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "100Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "fedemeterInfluxdb": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "10Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "imageLocation": "registry.connect.redhat.com/prophetstor",
                  "storages": [
                    {
                      "accessMode": ["ReadWriteOnce"],
                      "size": "2Gi",
                      "type": "pvc",
                      "usage": "log"
                    }
                  ],
                  "version": "v4.7.1-b1760"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: registry.connect.redhat.com/prophetstor/federatorai-operator@sha256:bdd87f6b7d7ba9d12ed8ddb9c5151e5a89781e88d7a02b73814b16a766968047
          createdAt: "2021-10-12T01:14:00Z"
          description: Federator.ai Operator provides easy configuration and management
            of AI-based Kubernetes resource orchestrator
          operatorframework.io/suggested-namespace: federatorai
          repository: https://access.redhat.com/containers/#/registry.connect.redhat.com/prophetstor/federatorai-operator
          support: ProphetStor Data Services, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of Federator.ai.
            displayName: AlamedaService
            kind: AlamedaService
            name: alamedaservices.federatorai.containers.ai
            version: v1alpha1
        description: |-
          Federator.ai helps enterprises optimize cloud resources, maximize application performance, and save significant cost without excessive over-provisioning or under-provisioning of resources, meeting the service-level requirements of their applications.

          Enterprises often lack understanding of the resources needed to support their applications. This leads to either excessive over-provisioning or under-provisioning of resources (CPU, memory, storage). Using machine learning, Federator.ai determines the optimal cloud resources needed to support any workload on OpenShift and helps users find the best-cost instances from cloud providers for their applications.


          **Multi-layer workload prediction**

          Using machine learning and math-based algorithms, Federator.ai predicts containerized application and cluster node resource usage as the basis for resource recommendations at application level as well as at cluster node level. Federator.ai supports prediction for both physical/virtual CPUs and memories.


          **Auto-scaling via resource recommendation**

          Federator.ai utilizes the predicted resource usage to recommend the right number and size of pods for applications. Integrated with Datadog's WPA, applications are automatically scaled to meet the predicted resource usage.


          **Application-aware recommendation execution**

          Optimizing the resource usage and performance goals, Federator.ai uses application specific metrics for workload prediction and pod capacity estimation to auto-scale the right number of pods for best performance without overprovisioning.


          **Multi-cloud Cost Analysis**

          With resource usage prediction, Federator.ai analyzes potential cost of a cluster on different public cloud providers. It also recommend appropriate cluster nodes and instance types based on resource usage.


          **Custom Datadog/Sysdig Dashboards**

          Predefined custom Datadog/Sysdig Dashboards for workload prediction/recommendation visualization for cluster nodes and applications.


          **Additional resources**

          Want more product information? Explore detailed information about using this product and where to find additional help.

          * [Federator.ai Datasheet](https://prophetstor.com/wp-content/uploads/datasheets/Federator.ai.pdf)
          * [Quick Start Guide](https://prophetstor.com/wp-content/uploads/documentation/Federator.ai/Latest%20Version/ProphetStor%20Federator.ai%20Quick%20Installation%20Guide.pdf)
          * [Installation Guide](https://prophetstor.com/wp-content/uploads/2021/08/ProphetStor-Federator.ai-v4.7.0-Installation-Guide-v1.0.pdf)
          * [User Guide](https://prophetstor.com/wp-content/uploads/2021/08/Federator.ai-4.7.0-User-Guide.pdf)
          * [Release Notes](https://prophetstor.com/wp-content/uploads/2021/08/Federator.ai-4.7.0-Release-Notes.pdf)
          * [Company Information](https://www.prophetstor.com/)


          ### Prerequisite
          *  The [Openshift](https://www.openshift.com) version 4.4.x or later if using OpenShift platform.
          *  The [Kubernetes](https://kubernetes.io/) version 1.16 or later if using Kubernetes environment.
        displayName: Federator.ai
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - AI
        - Resource Orchestration
        - NoOps
        links:
        - name: Website
          url: https://prophetstor.com/federator-ai-2/
        - name: Quickstart guide
          url: https://prophetstor.com/wp-content/uploads/documentation/Federator.ai/Latest%20Version/ProphetStor%20Federator.ai%20Quick%20Installation%20Guide.pdf
        maintainers:
        - email: support@prophetstor.com
          name: ProphetStor Data Services, Inc.
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: ProphetStor Data Services, Inc.
        relatedImages:
        - registry.connect.redhat.com/prophetstor/federatorai-operator@sha256:bdd87f6b7d7ba9d12ed8ddb9c5151e5a89781e88d7a02b73814b16a766968047
        - registry.connect.redhat.com/prophetstor/alameda-recommender-ubi@sha256:13e941df9252cea3024ecb5e2247a696b64928015c1a3a4a04ee4e362388b1e4
        - registry.connect.redhat.com/prophetstor/federatorai-data-adapter@sha256:adf5d655fa030dc05133516d45642d7282fdd66e0aa8e7a7331ed614a41d3fe0
        - registry.connect.redhat.com/prophetstor/alameda-operator-ubi@sha256:9a863394cb1c89228223b26addb0d12bb91a97233fe2d5d19ecd5998e0cfe0d6
        - registry.connect.redhat.com/prophetstor/federatorai-agent-ubi@sha256:cfadc84f31f3776ded15d4d8dda92e00b7d9b1bce0ff5440277176292f9304d3
        - registry.connect.redhat.com/prophetstor/federatorai-agent-preloader@sha256:66daea6193e399f8363df5ba42e1fcd643db6df3a9075dbcd8c3464fdc91e776
        - registry.connect.redhat.com/prophetstor/federatorai-rest-ubi@sha256:3e92528615543ea900bf38387c4eb8439b60759be5be2ac5311a9dbaafc7026b
        - registry.connect.redhat.com/prophetstor/alameda-rabbitmq@sha256:b396d9aa0470e593a36f5799c04fc3c9e59b5723f703f4b019f7ef92528e6741
        - registry.connect.redhat.com/prophetstor/alameda-notifier-ubi@sha256:1f75034fb78d62924a87cb4a5435661b4afa990d6960431e784029742fbae0da
        - registry.connect.redhat.com/prophetstor/alameda-ai-dispatcher@sha256:6e0dc1136b0f8e5971d92b95ce0d6dfac475044c073bada50f23b315f4478ae4
        - registry.connect.redhat.com/prophetstor/alameda-ai@sha256:9293298bb2a72d0b0764050b6f6d8da707b46a83363a3b6aa7be60f512e2ec61
        - registry.connect.redhat.com/prophetstor/alameda-analyzer-ubi@sha256:6117ac4692793d49bc0204f13091dec119bdbc7b169e9deb86c243aca00ded27
        - registry.connect.redhat.com/prophetstor/federatorai-dashboard-backend@sha256:022dbf6696edb8eaf389fdbd1dff14c40a51c237c01f8a02a7ea65c60cd59f38
        - registry.connect.redhat.com/prophetstor/alameda-datahub-ubi@sha256:3819d6acce8d376dbe3b12469be2528d0ce9ac95c2a86507aeb76b6e3a5f5f38
        - registry.connect.redhat.com/prophetstor/alameda-executor-ubi@sha256:4512e3be7a89d0ca9f120b68f6ab9766d000bf2b01b279cd55b2dbc433aacfb6
        - registry.connect.redhat.com/prophetstor/fedemeter-api-ubi@sha256:451456207fe8d16b9824b3f64d00645766dcccbc8d685b6942e6c00d0d0c627e
        - registry.connect.redhat.com/prophetstor/federatorai-dashboard-frontend@sha256:1c1147425ca6302103363a1a3bcf27baeef13b10ec6780ef266b7f52fe27ffeb
        - registry.connect.redhat.com/prophetstor/fedemeter-influxdb@sha256:36904f003e8e29d4a9fd738655f401fbe840d60158d7d820cda79a09858dc2c2
        - registry.connect.redhat.com/prophetstor/alameda-influxdb@sha256:485c2119a6a56807e28ad03b2087129950118b55531126a66bb414063052ae76
        version: 4.7.1-1
      entries:
      - name: federatorai.v4.7.1-1
        version: 4.7.1-1
      - name: federatorai.v4.6.1-4
        version: 4.6.1-4
      name: stable
    defaultChannel: stable
    packageName: federatorai-certified
    provider:
      name: ProphetStor Data Services, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: redhat-oadp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oadp-operator.v1.0.19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "restic": {
                      "enable": true
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws",
                        "kubevirt"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ResticRepository",
                "metadata": {
                  "name": "resticrepository",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/oadp/oadp-rhel8-operator@sha256:caf87a9832c153833feb34affc2b797ad0e643362448742b0405d457e5f0ba41
          createdAt: "2025-04-14T13:25:43Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <1.0.19'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/oadp/oadp-mustgather-rhel8@sha256:af4c6af48f38510c7d52f678798a9affea2211dc488ed18f83a5d005b7ef4b6b
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is a Velero resource that represents the capture of
              Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Data Protection Applications on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://github.com/openshift/oadp-operator
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/oadp/oadp-velero-plugin-for-microsoft-azure-rhel8@sha256:3e81b5e55f96504ba76d9d47bbeec50c7b1fd90b7702fc31b1dba248050ce9b4
        - registry.redhat.io/oadp/oadp-velero-plugin-for-gcp-rhel8@sha256:05a164c27da94b374ce03451c851a607a3cd436b73b9e5b70723c6dc8cacb3d8
        - registry.redhat.io/oadp/oadp-mustgather-rhel8@sha256:af4c6af48f38510c7d52f678798a9affea2211dc488ed18f83a5d005b7ef4b6b
        - registry.redhat.io/oadp/oadp-rhel8-operator@sha256:caf87a9832c153833feb34affc2b797ad0e643362448742b0405d457e5f0ba41
        - registry.redhat.io/oadp/oadp-registry-rhel8@sha256:0886b3b4ce0295080bf2c5ee95f2b7b15a4483bcd94cb739e7e7154d21150397
        - registry.redhat.io/oadp/oadp-velero-restic-restore-helper-rhel8@sha256:5d288335b5ad114b9eb61410befb4a9d1f84f0ba0c01558098bf4567fda9a8c5
        - registry.redhat.io/oadp/oadp-velero-plugin-rhel8@sha256:694603b9c1fd5eec939b28f8727fd7c9c0d380620c5081006f2655e98358464f
        - registry.redhat.io/oadp/oadp-velero-plugin-for-csi-rhel8@sha256:05ec16bdf1f0fadc4ddad244b86934f73569d2e66cc19f03b3553e257441a662
        - registry.redhat.io/oadp/oadp-velero-rhel8@sha256:39dec8b9ccdcabeaab3e56e14b79cfed4d8574cb172a886d017813a437e4ffff
        - registry.redhat.io/oadp/oadp-velero-plugin-for-aws-rhel8@sha256:28045538fff9e5051cd01e56876971cf9401a448e1dd63bff17fbb900629f22f
        - registry.redhat.io/oadp/oadp-kubevirt-velero-plugin-rhel8@sha256:d862a1b0420fcfb91403f90047b00035efba112af6bc5c50adb723014a620767
        version: 1.0.19
      entries:
      - name: oadp-operator.v1.0.19
        version: 1.0.19
      - name: oadp-operator.v1.0.18
        version: 1.0.18
      - name: oadp-operator.v1.0.17
        version: 1.0.17
      - name: oadp-operator.v1.0.16
        version: 1.0.16
      - name: oadp-operator.v1.0.15
        version: 1.0.15
      - name: oadp-operator.v1.0.14
        version: 1.0.14
      - name: oadp-operator.v1.0.13
        version: 1.0.13
      - name: oadp-operator.v1.0.12
        version: 1.0.12
      - name: oadp-operator.v1.0.11
        version: 1.0.11
      - name: oadp-operator.v1.0.10
        version: 1.0.10
      - name: oadp-operator.v1.0.9
        version: 1.0.9
      - name: oadp-operator.v1.0.8
        version: 1.0.8
      - name: oadp-operator.v1.0.7
        version: 1.0.7
      - name: oadp-operator.v1.0.6
        version: 1.0.6
      - name: oadp-operator.v1.0.5
        version: 1.0.5
      - name: oadp-operator.v1.0.4
        version: 1.0.4
      - name: oadp-operator.v1.0.3
        version: 1.0.3
      - name: oadp-operator.v1.0.2
        version: 1.0.2
      - name: oadp-operator.v1.0.1
        version: 1.0.1
      - name: oadp-operator.v1.0.0
        version: 1.0.0
      name: stable-1.0
    - currentCSV: oadp-operator.v1.4.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "nodeAgent": {
                      "enable": true,
                      "uploaderType": "restic"
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws",
                        "kubevirt"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: |
            Cloud Provider,Developer Tools,Modernization & Migration,OpenShift Optional,Storage
          certified: "false"
          containerImage: registry.redhat.io/oadp/oadp-rhel9-operator@sha256:ee010fb34448b772690a8a98808aa25aa1578a81061618f948b1d16b5d63d333
          createdAt: "2025-07-10T04:59:11Z"
          description: OADP (OpenShift API for Data Protection) operator sets up and
            installs Data Protection Applications on the OpenShift platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <1.4.5'
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/oadp/oadp-mustgather-rhel9@sha256:0cee6fecf052bae3ede6bb6f032b25c0c7cd63ced1e26328f2b02144f24fbb12
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A backup repository is an indicator of a connection from
              the restic/kopia server to the backupstoragelocation.
            displayName: BackupRepository
            kind: BackupRepository
            name: backuprepositories.velero.io
            version: v1
          - description: Backup is a Velero resource that represents the capture of
              Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - description: The CloudStorage API automates the creation of a bucket for
              object storage.
            displayName: CloudStorage
            kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataDownload represents a data download of a volume snapshot.
              There is one DataDownload created per volume to be restored.
            displayName: DataDownload
            kind: DataDownload
            name: datadownloads.velero.io
            version: v2alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DataUpload acts as the protocol between data mover plugins
              and data mover controller for the datamover backup operation
            displayName: DataUpload
            kind: DataUpload
            name: datauploads.velero.io
            version: v2alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: |
          **OpenShift API for Data Protection (OADP)** operator sets up and installs
          Velero on the OpenShift platform, allowing users to backup and restore
          applications.

          Backup and restore Kubernetes resources and internal images, at the
          granularity of a namespace, using a version of Velero appropriate for
          the installed version of OADP.

          OADP backs up Kubernetes objects and internal images by saving them as
          an archive file on object storage. OADP backs up persistent volumes
          (PVs) by creating snapshots with the native cloud snapshot API or with
          the Container Storage Interface (CSI). For cloud providers that do not
          support snapshots, OADP backs up resources and PV data with Restic or Kopia.

          * [Installing OADP for application backup and restore][install-doc]

          * [Installing OADP on a ROSA cluster and using STS, please follow the Getting Started Steps 1-3 in order to obtain the role ARN needed for using the standardized STS configuration flow via OLM][rosa-sts-doc]

          * [Frequently Asked Questions][faq-doc]

          [install-doc]: https://docs.openshift.com/container-platform/latest/backup_and_restore/application_backup_and_restore/installing/about-installing-oadp.html
          [rosa-sts-doc]: https://mobb.ninja/docs/misc/oadp/rosa-sts/
          [faq-doc]: https://access.redhat.com/articles/5456281
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - disaster backup
        - disaster recovery
        - disaster restore
        - disaster
        - backup
        - restore
        - backup automation
        - recovery
        - data protection
        - data management
        - data
        - protection
        - management
        - application resilience
        - resilience
        - velero
        - openshift
        - oadp
        - cloud-native
        - replication
        - kopia
        - restic
        - network file system
        - nfs
        - simple cloud storage
        - s3
        - minio
        - cloud storage
        - data foundation
        - ceph
        - csi
        - container storage interface
        links:
        - name: OADP Operator
          url: https://github.com/openshift/oadp-operator
        maintainers:
        - email: oadp-maintainers@redhat.com
          name: OADP Operator maintainers
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/oadp/oadp-velero-plugin-for-legacy-aws-rhel9@sha256:ab5c8d33684a361fa3bb10a006ef489be92cc955ad3d173bbfa77949dcc42aa8
        - registry.redhat.io/oadp/oadp-kubevirt-velero-plugin-rhel9@sha256:c3b32cb9f5b82021c08cd70467e68aad66746ef211d73612ecfe0af625706267
        - registry.redhat.io/oadp/oadp-velero-plugin-for-aws-rhel9@sha256:79d39186cae5696949813a0ea6c46eeededad29e6bde171d7a9b0011c993bb26
        - registry.redhat.io/oadp/oadp-velero-plugin-for-microsoft-azure-rhel9@sha256:9f05592bf0e5b76322ca582bdede387f9f4aabbbdc436549156c05d652c951d0
        - registry.redhat.io/oadp/oadp-mustgather-rhel9@sha256:0cee6fecf052bae3ede6bb6f032b25c0c7cd63ced1e26328f2b02144f24fbb12
        - registry.redhat.io/oadp/oadp-rhel9-operator@sha256:ee010fb34448b772690a8a98808aa25aa1578a81061618f948b1d16b5d63d333
        - registry.redhat.io/oadp/oadp-velero-rhel9@sha256:17160f99940aa355dc91ab3d16ba6bf91faade6423434008cf8f6e5023b1e413
        - registry.redhat.io/oadp/oadp-velero-plugin-for-gcp-rhel9@sha256:00800ead11ea308794b1536d964f8a043e13ede9657cabb23afa42cba7cf2ce6
        - registry.redhat.io/oadp/oadp-velero-plugin-rhel9@sha256:3ee9d3e799cedc28c400d364b3a8790c7e88acd5f603b09ff512a22d91e265b3
        - registry.redhat.io/oadp/oadp-velero-restic-restore-helper-rhel9@sha256:e0647defc7c428f9a04a5f6d4532dc86e57b92f31b4f907253ae14af05c5993f
        version: 1.4.5
      entries:
      - name: oadp-operator.v1.4.5
        version: 1.4.5
      - name: oadp-operator.v1.4.4
        version: 1.4.4
      - name: oadp-operator.v1.4.3
        version: 1.4.3
      - name: oadp-operator.v1.4.2
        version: 1.4.2
      - name: oadp-operator.v1.4.1
        version: 1.4.1
      - name: oadp-operator.v1.4.0
        version: 1.4.0
      name: stable-1.4
    defaultChannel: stable-1.4
    packageName: redhat-oadp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Oracle
      provider-url: https://www.oracle.com/cloud/
    name: oci-ccm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oci-ccm-operator.v0.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ccm.oraclecloud.com/v1alpha1",
                "kind": "OCICloudManager",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "oci-ccm-operator",
                    "app.kubernetes.io/instance": "ocicloudmanager",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ocicloudmanager",
                    "app.kubernetes.io/part-of": "oci-ccm-operator"
                  },
                  "name": "ocicloudmanager"
                },
                "spec": {
                  "compartment": "CompartmentId",
                  "driverNamespace": "kube-system",
                  "loadBalancer": {
                    "securityList": "SecurityListId",
                    "securityListManagementMode": "Frontend",
                    "subnet": "SubnetId"
                  },
                  "vcn": "VcnId"
                }
              },
              {
                "apiVersion": "ccm.oraclecloud.com/v1alpha1",
                "kind": "OCICsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "oci-ccm-operator",
                    "app.kubernetes.io/instance": "ocicsidriver",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ocicsidriver",
                    "app.kubernetes.io/part-of": "oci-ccm-operator"
                  },
                  "name": "ocicsidriver"
                },
                "spec": {
                  "compartment": "CompartmentId",
                  "driverNamespace": "kube-system",
                  "vcn": "VcnId"
                }
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider, Storage
          containerImage: phx.ocir.io/ax1eoqvpvbtg/oci-ccm-operator:0.0.9
          createdAt: "2023-07-28T23:42:18Z"
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Oracle
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: OCICloudManager
            name: ocicloudmanagers.ccm.oraclecloud.com
            version: v1alpha1
          - kind: OCICsiDriver
            name: ocicsidrivers.ccm.oraclecloud.com
            version: v1alpha1
        description: OCI Kubernetes Cloud Controller Manager implementation (or out-of-tree
          cloud-provider) for Oracle Cloud Infrastructure (OCI).
        displayName: oci-ccm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OCI
        - CCM
        - OpenShift
        - Kubernetes
        links:
        - name: Oracle
          url: https://www.oracle.com/cloud/
        - name: OCI Cloud Controller Manager (CCM)
          url: https://github.com/oracle/oci-cloud-controller-manager
        maintainers:
        - email: david.d.campbell@oracle.com
          name: David Campbell
        - email: alex.blinov@oracle.com
          name: Alex Blinov
        maturity: alpha
        provider:
          name: Oracle
          url: https://www.oracle.com/cloud/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - phx.ocir.io/ax1eoqvpvbtg/oci-ccm-operator:0.0.9
        version: 0.0.9
      entries:
      - name: oci-ccm-operator.v0.0.9
        version: 0.0.9
      name: alpha
    defaultChannel: alpha
    packageName: oci-ccm-operator
    provider:
      name: Oracle
      url: https://www.oracle.com/cloud/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: eventing-kogito
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eventing-kogito.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kogito.knative.dev/v1alpha1",
                "kind": "KogitoSource",
                "metadata": {
                  "name": "kogito-order-processing"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "kind": "Broker",
                      "name": "default"
                    }
                  },
                  "subject": {
                    "apiVersion": "serving.knative.dev/v1",
                    "kind": "Service",
                    "name": "serverless-workflow-order-processing"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: gcr.io/knative-releases/knative.dev/eventing-kogito/cmd/controller:v1.2.0
          createdAt: "2021-08-25T12:00:00Z"
          olm.targetNamespaces: knative-kogito
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/knative-sandbox/eventing-kogito
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoSource is the representation of a Knative Eventing
              Source for a Kogito custom Service.
            displayName: Kogito Source
            kind: KogitoSource
            name: kogitosources.kogito.knative.dev
            version: v1alpha1
        description: |-
          Knative Eventing Source for Kogito Services. This controller will enable easy configuration of any Kogito custom Service
          with the Knative Eventing platform. Please find more information about using this source in the [Official Kogito Documentation](https://docs.jboss.org/kogito/release/latest/html_single/#_kogito_knative_eventing_add_on)

          **Note:** The admission webhooks are not installed via OLM at this time. Because of that, please ensure that you create the `KogitoSource` objects with the subbject and sink information with the appropriate namespace.
        displayName: Kogito Knative Eventing Source
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless workflow
        - kogito
        - knative
        - dmn
        - bpmn
        - eventing
        - cloudevents
        - messaging
        links:
        - name: Source Repository
          url: https://github.com/knative-sandbox/eventing-kogito
        - name: Kogito Knative Add-On Documentation
          url: https://docs.jboss.org/kogito/release/latest/html_single/#_kogito_knative_eventing_add_on
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - gcr.io/knative-releases/knative.dev/eventing-kogito/cmd/controller:v1.2.0
        version: 1.2.0
      entries:
      - name: eventing-kogito.v1.2.0
        version: 1.2.0
      - name: eventing-kogito.v1.1.0
        version: 1.1.0
      name: alpha
    defaultChannel: alpha
    packageName: eventing-kogito
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KEDA Community
      provider-url: ""
    name: keda
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keda.v2.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "KedaController",
                "metadata": {
                  "name": "keda",
                  "namespace": "keda"
                },
                "spec": {
                  "watchNamespace": "",
                  "operator": {
                    "logLevel": "info",
                    "logEncoder": "console"
                  },
                  "metricsServer": {
                    "logLevel": "0"
                  },
                  "serviceAccount": {}
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledJob",
                "metadata": {
                  "name": "scaledjob-sample"
                },
                "spec": {
                  "failedJobsHistoryLimit": 5,
                  "jobTargetRef": {
                    "activeDeadlineSeconds": 600,
                    "backoffLimit": 6,
                    "completions": 1,
                    "parallelism": 1,
                    "template": null
                  },
                  "maxReplicaCount": 100,
                  "pollingInterval": 30,
                  "successfulJobsHistoryLimit": 5,
                  "triggers": [
                    {
                      "metadata": {
                        "property": "examle-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledObject",
                "metadata": {
                  "name": "example-scaledobject"
                },
                "spec": {
                  "cooldownPeriod": 300,
                  "maxReplicaCount": 100,
                  "minReplicaCount": 0,
                  "pollingInterval": 30,
                  "scaleTargetRef": {
                    "name": "example-deployment"
                  },
                  "triggers": [
                    {
                      "metadata": {
                        "property": "examle-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "TriggerAuthentication",
                "metadata": {
                  "name": "example-triggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "exmaple-secret-parameter"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ClusterTriggerAuthentication",
                "metadata": {
                  "name": "example-clustertriggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "exmaple-secret-parameter"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Cloud Provider
          certified: "false"
          containerImage: ghcr.io/kedacore/keda-olm-operator:2.7.1
          createdAt: "2022-05-11T00:00:00.000Z"
          description: '[DEPRECATED] Use Custom Metrics Autoscaler Operator instead'
          operatorframework.io/suggested-namespace: keda
          operators.operatorframework.io/builder: operator-sdk-v1.1.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/kedacore/keda-olm-operator
          support: KEDA Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              Represents an installation of a particular version of KEDA Controller.
            displayName: KedaController
            kind: KedaController
            name: kedacontrollers.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your jobs and what the triggers
            displayName: ScaledJob
            kind: ScaledJob
            name: scaledjobs.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your application and what the
              triggers are
            displayName: ScaledObject
            kind: ScaledObject
            name: scaledobjects.keda.sh
            version: v1alpha1
          - description: Describes authentication parameters
            displayName: TriggerAuthentication
            kind: TriggerAuthentication
            name: triggerauthentications.keda.sh
            version: v1alpha1
          - description: Describes cluster wide authentication parameters
            displayName: ClusterTriggerAuthentication
            kind: ClusterTriggerAuthentication
            name: clustertriggerauthentications.keda.sh
            version: v1alpha1
        description: "## Deprecated\n This KEDA community operator is deprecated,
          please use `Custom Metrics Autoscaler` Operator instead.\n## About the managed
          application\nKEDA is a Kubernetes-based event driven autoscaler.  KEDA can
          monitor event sources like Kafka, RabbitMQ, or cloud event sources and feed
          the metrics from those sources into the Kubernetes horizontal pod autoscaler.
          \ With KEDA, you can have event driven and serverless scale of deployments
          within any Kubernetes cluster.\n## About this Operator\nThe KEDA Operator
          deploys and manages installation of KEDA Controller in the cluster. Install
          this operator and follow installation instructions on how to install KEDA
          in you cluster.\n\n## Prerequisites for enabling this Operator\nPlease note
          that you can not run both KEDA 1.x and 2.x on the same Kubernetes cluster.
          You need to uninstall KEDA 1.x first, in order to install and use KEDA 2.x.\n\nDon't
          forget to uninstall KEDA v1 CRDs as well, to ensure that, please run:\n\n```
          oc delete crd scaledobjects.keda.k8s.io ```\n\n``` oc delete crd triggerauthentications.keda.k8s.io
          \ ```\n## How to install KEDA in the cluster\nThe installation of KEDA is
          triggered by the creation of `KedaController` resource, please refer to
          [KedaController Example](https://github.com/kedacore/keda-olm-operator#kedacontroller-spec)
          for more deatils on available options.\n\nOnly resource named `keda` in
          namespace `keda` will trigger the installation, reconfiguration or removal
          of the KEDA Controller resource.\n\nThere could be only one KEDA Controller
          in the cluster. \n"
        displayName: '[DEPRECATED] KEDA'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keda
        - scaling
        - event
        links:
        - name: Homepage
          url: https://keda.sh
        - name: Source Code
          url: https://github.com/kedacore/keda
        - name: Slack Channel
          url: https://kubernetes.slack.com/messages/CKZJ36A5D
        maintainers:
        - email: cncf-keda-maintainers@lists.cncf.io
          name: KEDA Community
        maturity: deprecated
        minKubeVersion: 1.17.0
        provider:
          name: KEDA Community
        relatedImages:
        - ghcr.io/kedacore/keda-olm-operator:2.7.1
        version: 2.7.1
      entries:
      - name: keda.v2.7.1
        version: 2.7.1
      - name: keda.v2.6.1
        version: 2.6.1
      - name: keda.v2.6.0
        version: 2.6.0
      - name: keda.v2.5.0
        version: 2.5.0
      - name: keda.v2.4.0
        version: 2.4.0
      - name: keda.v2.3.0
        version: 2.3.0
      - name: keda.v2.2.0
        version: 2.2.0
      - name: keda.v2.1.0
        version: 2.1.0
      - name: keda.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: keda
    provider:
      name: KEDA Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: stolostron
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stolostron.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:3443da607b276d39ba28a8d6561d296f30e9bac5ab06d5c70715bddbde0a58b2
          createdAt: "2024-06-28T04:52:57Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/multiclusterhub-operator@sha256:3443da607b276d39ba28a8d6561d296f30e9bac5ab06d5c70715bddbde0a58b2
        - quay.io/stolostron/governance-policy-addon-controller@sha256:593dc0a689091ce7c916e6be764c222c1020629ccff7eafbf17524936e4f8262
        - quay.io/stolostron/observability-e2e-test@sha256:8d9341b1814cd085a20f9feceab181a8989d6684fd44d6c8e54a0c6df4dd7216
        - quay.io/stolostron/governance-policy-framework-addon@sha256:a1c0d40561befaf5e849060beaee4b201f42d77c5090823f23ef43a75106b00b
        - quay.io/stolostron/thanos-receive-controller@sha256:b1a191714cead24f66e8de7217db3c4d4a253ecf0238a912c8b88ba8b47c972d
        - quay.io/stolostron/console@sha256:c7900403ef8dba3aac52f999532d876bbfe2f66f00a255047850594b64021830
        - quay.io/stolostron/volsync-mover-rclone@sha256:b247b23944e20104f801c172cbbb620740a5b51d1777ecdbf98b4c128f1ef171
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/prometheus@sha256:8250af047c03c6b0c07b76420e0a5c2338f053d73b2dca1ecba4bb2f211715fe
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/search-collector@sha256:136ad44ff41a29dec48e87d10af5b608f335e4242c41ac17cba2c3c4c8205bd6
        - quay.io/stolostron/multicluster-operators-channel@sha256:d791aa2c6032c1de50cf4033c426646613b182e98c406a091b88fb9ffefc0c98
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/helloworld@sha256:81a0b06c7fcf915b1bb40d33d29d0bfd4d501404463c1f46bffc503e7d1cc45b
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:02e53b8a6529e3209809dada4113054e3398de18738948cb01ed1d24dcb8e8be
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/insights-metrics@sha256:49a177357e865d5f8347840fe695531d539d17e28d2b9971821e41ebbee014bd
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/rbac-query-proxy@sha256:24aafcba20ded0892209cfcc196bff88f877ed51eb928d5d350b17041d170893
        - quay.io/stolostron/volsync-mover-restic@sha256:284c3358f0f2158d599f27487d76a7fdb21842fc0e95bc639726dfdad14b7e7d
        - quay.io/stolostron/prometheus-config-reloader@sha256:bc93ca9f1b6d95e8b963fe1af4c17f87c8c7d938a348627c29cc3cf42b70ca3e
        - quay.io/stolostron/prometheus-alertmanager@sha256:8222bbdd37c246703de54ed3dc109fa8c2f6ca17c5bb90106948683858afdc65
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/multicloud-integrations@sha256:60be5b945ead8f845a29a8be18bd26922b83f009fde1412017f1ca985eb79b55
        - quay.io/stolostron/grafana@sha256:e40d66da749108ae1eecf909c95c87bf7a289328e8b373ae4402368c474f2da5
        - quay.io/stolostron/observatorium@sha256:af4e315e4c6da3607f84e24f31e0f37f8572d161736b4748b2e57e0a2c2dd36c
        - quay.io/stolostron/volsync-addon-controller@sha256:7774767ed20263bba4857a0d4be012e12215f640fce01fbd3a085bf860d087e7
        - quay.io/stolostron/volsync-mover-syncthing@sha256:b5fad266ca8d3cd8efcbf4e8241c446004747b551b2b8e06acae10900d7a0f1b
        - quay.io/stolostron/thanos@sha256:cc78cd786643d27ed85eb3670852722e915a20b20a21479e1477fb859dec6c5e
        - quay.io/stolostron/prometheus-operator@sha256:490b0056b9a23a1dc514a3036eac6593c7c8961de27d08ff24817d280325138c
        - quay.io/stolostron/cluster-backup-controller@sha256:9c77444bd42e6e99163f4dce9f9975276aea082100f53fd938cddf74ad3b1574
        - quay.io/stolostron/multicluster-operators-application@sha256:199b696dbd7589b77e2391b71504fbcb7d3ae785972dac3d001c4c1bffa342dd
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/kube-rbac-proxy@sha256:9626cb9b2ee8f00d2abdb14cd2ce56d77435859071c10d8394bc7ee2754ce87f
        - quay.io/stolostron/volsync-mover-rsync@sha256:f7affcc16ae9bf6d12f69511b7a269f5fd065d1d87392c4c38cb5d7af5b89e23
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/search-indexer@sha256:068c38489201316dc12d475de03c8abb11575338bd75e25c1262eca8aac49cb2
        - quay.io/stolostron/search-v2-api@sha256:6b97b2c7e64de7cd154fd83da3bc690a57b67f01c83170238fe9c2e764df91a8
        - quay.io/stolostron/cert-policy-controller@sha256:27673358610dc1ca308de472dbd794dcbb38250bb299b53c736549b1e8441a0f
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/metrics-collector@sha256:4101a2f2c4230bdc1420a58b546a241dfc5414cf92b57f3b73619c1be14dbc32
        - quay.io/stolostron/observatorium-operator@sha256:5da7f478ecf3609a3b86eb563e4dd22588f3d13efdc2e33f5b672fe95583ce68
        - quay.io/stolostron/grafana-dashboard-loader@sha256:cff97fe3e32dd14ed2f3b7b02cb205715cfee6cfd6dc7e7a630101ea5147e566
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:9d1d5e2b5313b898188d2c845dcda5d604b73f16e89c71915ea308281c1f56ff
        - quay.io/stolostron/klusterlet-addon-controller@sha256:01e5fca5e5c6caca6878bf8025471b7c77164ca7b4a2d764e20101fc79290df0
        - quay.io/stolostron/search-e2e@sha256:8dfbc899ba1ffaa8fb6eb2861f6db87a13d2565cf4551fd47df7fe8253217fb2
        - quay.io/stolostron/config-policy-controller@sha256:197d8491e7c889db8f2acb726e0146d384c979acc7f634c24a9d4b21f7d444f3
        - quay.io/stolostron/must-gather@sha256:1f7c129859bdaf6bd93dda77aec01f7cd03000fbfaccd809f26d33b9e3f66e51
        - quay.io/stolostron/search-v2-operator@sha256:c41517f372fc1874895b39e70cd1c07391110697a5e49ebcf56f4a5df499d810
        - quay.io/stolostron/kube-state-metrics@sha256:ef1d671ffa098ed95186de6f15ea91fe648d103c1c7a57871ed46af07c6f7881
        - quay.io/stolostron/submariner-addon@sha256:0f1b4ae59ff5c933542f9affed20bfa6694deb804f2e97f49ebd0f2e1b925eca
        - quay.io/stolostron/insights-client@sha256:aac5b9b66800c0b862e8bfa1fea533a4aec854d8c52baddbfa03977e1ed520cb
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:bf6bec399bc2072186e9d12ed865167fe940f3f619d69ffe6c178c14a3bdadbe
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/multicluster-observability-operator@sha256:245e1815a73b5c48f790ae5a5d85fac6642cce8a5a3bf122275809e97a4d519e
        - quay.io/stolostron/grc-policy-framework-tests@sha256:f2b01192fd900860428942c7d5b900604ffcab0d42731676a1acdeebd3716557
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/cluster-permission@sha256:716d2cbcb7f69b5640d04669b15884e0adf053597f278a0bb1d8ee77e4a7af4d
        - quay.io/stolostron/governance-policy-propagator@sha256:200558b5aa2366383bd298e159d9f7966a2c9b981038c79cde536edbfbe0092c
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/volsync@sha256:ba7399b316359118b1c1b915d667fca9a4a0d12a10beae69e6dee9c7236753be
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:7f6b6e45f0be0a14d7d9d79828572b9400f4d559f084028e3254cfd40948780d
        - quay.io/stolostron/multicluster-operators-subscription@sha256:8350df0a7d4f19c8e52eb58eecf5223bb2a0ccbe4d4aa00fcf8dff832cb43e5d
        version: 0.4.0
      entries:
      - name: stolostron.v0.4.0
        version: 0.4.0
      name: community-0.4
    - currentCSV: stolostron.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:781f6712369cd4b9bdc8ba0ee2f35a12db198ca190621e199dc10cfea3f5efc9
          createdAt: "2024-10-02T08:01:20Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.5.0 <0.5.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/governance-policy-framework-addon@sha256:5f2b326a279b7c42c233ea5f11b89ecaf4cfd7c549d021ddcc4e7b0c8926fc6d
        - quay.io/stolostron/rbac-query-proxy@sha256:2f1c78176d0ca0b21c237a9f7a3f77f7d857dd848f3a42254cee1e4e469b4c3c
        - quay.io/stolostron/multiclusterhub-operator@sha256:781f6712369cd4b9bdc8ba0ee2f35a12db198ca190621e199dc10cfea3f5efc9
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:d42ddd30570ccdf388ccd65cd3f7c72de2d857d328e05aef8e855508ccf91898
        - quay.io/stolostron/observatorium@sha256:8d0f02996b5bf02be02ff3743821909b6c2396f0f0519652dde0c5d1d76f8a05
        - quay.io/stolostron/prometheus@sha256:147f3a0cee8fa7b11c82e21ca95c28278e1ab53ccb5b70f3af0425bb8a74f61b
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/volsync-mover-syncthing@sha256:b5fad266ca8d3cd8efcbf4e8241c446004747b551b2b8e06acae10900d7a0f1b
        - quay.io/stolostron/search-collector@sha256:4082f7a37aa3f10c725bb9bebf94ee8dab97d22302cd5e59b35d0e3bcc888cef
        - quay.io/stolostron/volsync-mover-rclone@sha256:b247b23944e20104f801c172cbbb620740a5b51d1777ecdbf98b4c128f1ef171
        - quay.io/stolostron/insights-metrics@sha256:49a177357e865d5f8347840fe695531d539d17e28d2b9971821e41ebbee014bd
        - quay.io/stolostron/acm-cli@sha256:85a5c55dfa4708435807d1266b5607525e33f55c5b4672fd8ba78987a1c48ef3
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/thanos-receive-controller@sha256:58c97fe5ec2131f9dcfb70563b59ed0bcae94d8bddefa51cb4784e73b1d90e4f
        - quay.io/stolostron/cluster-backup-controller@sha256:688844c722db78f383d90986150bde93bb421068e7f5b4128d19043f3e01f93e
        - quay.io/stolostron/config-policy-controller@sha256:5af6a0e17da7b7261ac71bcb1e2d991856bbb5450686d188026798e05f0617ba
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/volsync-addon-controller@sha256:d0b15da92e0ff9792d750b7b48d72b38d6abb91488c20578c7be3382c561877f
        - quay.io/stolostron/observability-e2e-test@sha256:c361f5a0f806498c6c1150c0c0d5a42222efff13a18cb28dbfd5e4bd835cb4b7
        - quay.io/stolostron/siteconfig-operator@sha256:7f47f0c61d5b53411ed78046c5a02664c6693d940ef37d6db1853670196be334
        - quay.io/stolostron/insights-client@sha256:aac5b9b66800c0b862e8bfa1fea533a4aec854d8c52baddbfa03977e1ed520cb
        - quay.io/stolostron/kube-rbac-proxy@sha256:1b708bff7fb46391f2460d34be28d30add38501cdb084306df01faae75637232
        - quay.io/stolostron/multicluster-observability-operator@sha256:fb0519952ef6436cc66496409f00e89cd6e58a41776df96859a86176a7322d05
        - quay.io/stolostron/multicluster-operators-subscription@sha256:30b1b5bf6a7a12e525450905a474babdd7e38235d9be60f3cb04d60a745a1b96
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:a8d812a527902bd4d9dfd23f90ce90f415f5a6fabb007d1c41b4c998e1ca098a
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/observatorium-operator@sha256:38f22e30db7713e9f2e239463803cdb5cdfe7108a524ffa5d7e95201bae5300b
        - quay.io/stolostron/cluster-permission@sha256:716d2cbcb7f69b5640d04669b15884e0adf053597f278a0bb1d8ee77e4a7af4d
        - quay.io/stolostron/thanos@sha256:dd4d4f8fbb9a8f51134260fc728d19f61a93b511fd7889636089944ac46163bd
        - quay.io/stolostron/search-v2-operator@sha256:6df11a36ea19a1b3da191964f851dee95b87fd993e8f0ef8906b1e63dd450d17
        - quay.io/stolostron/metrics-collector@sha256:c580106b2c7173251afa6714b710fd109811023d88a8009fb741f3d00e486e76
        - quay.io/stolostron/volsync-mover-rsync@sha256:f7affcc16ae9bf6d12f69511b7a269f5fd065d1d87392c4c38cb5d7af5b89e23
        - quay.io/stolostron/search-v2-api@sha256:338634d767383716c7a88ed2bc9c58275e9d5953932293443725883d7efcffd5
        - quay.io/stolostron/multicluster-operators-application@sha256:329c82b92fe2448fe26923fbd82d2eadacb7f96c75689d96393c573f95e6d650
        - quay.io/stolostron/must-gather@sha256:362bbfd54059975369ce6d3869d34e346a0e76159163714cfcb3597022b7c88f
        - quay.io/stolostron/prometheus-config-reloader@sha256:337f7d2036c17459a41884e38c569a3601364b02660946b49d39ce40ef97bd48
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/governance-policy-addon-controller@sha256:9f387eb8b54d1c86fc20d205dbce199738849cb0ce15a47052ddbee8781d2d9b
        - quay.io/stolostron/volsync-mover-restic@sha256:284c3358f0f2158d599f27487d76a7fdb21842fc0e95bc639726dfdad14b7e7d
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/search-e2e@sha256:ddaa50ecfafcd7501903420a47495f733dce36fcb91c8ce9b8d94f4448ed501d
        - quay.io/stolostron/submariner-addon@sha256:07287987fb0041efc16557cb9d7403dc15b5959a5d0dd0d81cd72bc2658cba10
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/klusterlet-addon-controller@sha256:cf4cad3599c4ce538dc007cda9f5b06538799e8198912d907058ecdd4bbb2451
        - quay.io/stolostron/multicloud-integrations@sha256:d4cb3a0a50bc5d22f48c96ed3a10ea9843c6ef2f83daeae411a7ddb7ea332ea6
        - quay.io/stolostron/prometheus-operator@sha256:5ff1f7a4be0216f582f5cb40de3412cbe8126433144c8201750f6c3c0fee920e
        - quay.io/stolostron/governance-policy-propagator@sha256:69ab8f492fa418596db6da578b3b579fc2f1506c5bf780e6aba3ccd05d8b02b9
        - quay.io/stolostron/console@sha256:e42d2f7e39a93f13003193142a3d1f3f9e55966cbf55b94eac93c726373a2516
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/volsync@sha256:ba7399b316359118b1c1b915d667fca9a4a0d12a10beae69e6dee9c7236753be
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:dbdbec1600d57c440173aecf19c57709d69e945dcecb348bcf4a7c37b3700b70
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/search-indexer@sha256:7471f2808686d42c2eae9f0a537f9b520648f31e87072a376243f309fd044caa
        - quay.io/stolostron/grafana@sha256:c9869c7f590bc4c54151c7e4fa9ed76dfe35c24a0ea3c19a6a1b26e61924b6e1
        - quay.io/stolostron/cert-policy-controller@sha256:809cae1aef02dd3fb68143ebdba818b20c661bb628a7922ee1d705efe4947f35
        - quay.io/stolostron/helloworld@sha256:48724b53542a3fcaf342d6ae2056bdf846cbe2ee50879613953b045bfcbe1c0a
        - quay.io/stolostron/kube-state-metrics@sha256:dc534373feda042afa385026c7da674912ca310df2a4598d83451d4474bb3f44
        - quay.io/stolostron/grc-policy-framework-tests@sha256:57794e2b15949cedf3462df2d38b95e86ca1e140a5ae81e4a9b3e7e06072847d
        - quay.io/stolostron/prometheus-alertmanager@sha256:19f2e277cbdd2bd4814b8d8034167c1a7cfe9f1d698fb16f7604fa65cb8bcb1c
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/multicluster-operators-channel@sha256:260a11aae954299ce8fbb0f9c3102afd3b5cf3f1b972b5a71e56c273ba1d532f
        - quay.io/stolostron/grafana-dashboard-loader@sha256:85031d1a56514be239c585487aec05a8cbb08c17aeb60c0c84c685eb9d9f08b5
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:358ac0b0a3afdfac06f784f2e02a305cbef4e6f648e31a00981d3b097746412a
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        version: 0.5.3
      entries:
      - name: stolostron.v0.5.3
        version: 0.5.3
      - name: stolostron.v0.5.2
        version: 0.5.2
      - name: stolostron.v0.5.1
        version: 0.5.1
      - name: stolostron.v0.5.0
        version: 0.5.0
      name: community-0.5
    - currentCSV: stolostron.v0.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:09f1f4266938fba88376a50c3e2837c5517b4f3a947cf86638a17e52eb17c1f6
          createdAt: "2025-02-24T01:42:20Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.6.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/flightctl-worker@sha256:4dddded7d2d4cce63972c8ca81f213c2cedb551cc645ec85495a9692dc36bb4c
        - quay.io/stolostron/helloworld@sha256:56f0d641d4f07b922f1c28c536eacad747c56a564e0a65f084106c2ffc9bd99d
        - quay.io/stolostron/acm-cli@sha256:8c8c059ce2f6d131c410972bf8fd0b7e399cf17cd044c7ddc89a9668c0ceac4c
        - quay.io/stolostron/search-v2-api@sha256:694523242aabb8db1f27b5afcd8ab2f63a9cdb6056c33a79badf8e68a337738d
        - quay.io/stolostron/multicluster-observability-operator@sha256:bed0dd30646d7fce894a625e6cf52e7829c272f47dde0baf1f788e9f85f66e42
        - quay.io/stolostron/cert-policy-controller@sha256:98dd74d7375ca2b546075d5c8f06a9e2c4f2ce3b77247dccf36bcaa46bf564c7
        - quay.io/stolostron/governance-policy-propagator@sha256:c21574a2421fefe3fcabeef14640d6fe699d950106c9ed056b3ba9be3cbeabe0
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/flightctl-periodic@sha256:a3ed745dc65a5886c4cc05348cf4822bfbf2c6d28b85a39fa0b8f3e1f53fdf8a
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/search-indexer@sha256:4f9be902fe976fd88a77b737cc1f56f00d48740a6a62661468328cf3bb9da31a
        - quay.io/stolostron/search-v2-operator@sha256:85e53a97240f6fd1c4251101146cc2bcc51eb91d9b171efb5b58edade9e0244e
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/multicluster-operators-application@sha256:59802fcfae5b6fd0bebe58e98412885e71f32464b8dbbf85364ddaf090aa3e00
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:9c85fb91c4be3287cac2a376465836689df885902099096f0f76ffca50a6d311
        - quay.io/stolostron/search-e2e@sha256:43e1bbfb7259612ffb408108c71b825f6e61a268c7e405b8d855f6878b6f96dd
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/flightctl-api@sha256:64fb1b655bef6aeff03cf64b6450155d3cb570ed4713f084590fe515f4fbad02
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:51a8acc47098175db148ebc19a10cdba2803d4a9f5884dd38b1c839702398789
        - quay.io/stolostron/thanos@sha256:53eee303836070a7c7621bfb2a094e3809146497d1653dc65d16055aacea332c
        - quay.io/stolostron/siteconfig-operator@sha256:97dc128666780ec9a4c955f392ab98f81f0d6038ec2159e8b1d55aab393cf118
        - quay.io/stolostron/config-policy-controller@sha256:8cf54eb6f5b49a60ed165fa4b70e4984a1362f2f31e8304c3a22e8a6a953a606
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/governance-policy-addon-controller@sha256:fa0738abc0368f9dadd8b0845d868512d26be3525a1a07b9de8a9cefafaa42b5
        - quay.io/stolostron/klusterlet-addon-controller@sha256:39e6940eb956fe9640342e11ecbcbc7d6df0169e1b340a60afc19aff7dd713a5
        - quay.io/stolostron/thanos-receive-controller@sha256:d73f98e93a258adffc1a4bd875c79a9629166599e68ef777a790a1f3146a2929
        - quay.io/stolostron/cluster-permission@sha256:443d8b30e3bb83c139c64ddfc38b0d5e6d7a25a4fae783303ab2f2c27d015b26
        - quay.io/stolostron/grafana@sha256:51092f6754106a90aff28cdd80c2f0447c85c1bbc72db15878d0b65b93e1ead6
        - quay.io/stolostron/volsync-addon-controller@sha256:d9dad837f2f735674e5711c7fff1cfe2c0c92160fceb7d1548a62348dd43b81c
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:92369b6deb21ab3cb0a12ae5c74b455cfefb8698826bca39e26648e608e48a2b
        - quay.io/stolostron/observatorium@sha256:398ecee4d3f12e84e48b12009a4b02fc8b360dfc0c1607ffecc26e6bc04d0ea2
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/prometheus-config-reloader@sha256:2b26d2fb5c69e54aecfbcaab46e652d136d726f0604844031c8771874f1fd769
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/grc-policy-framework-tests@sha256:9b84dbafef35bd568b2bf79283b6850ba1ff57d70239a27cc0cc815fb1a55515
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/metrics-collector@sha256:c7433ee4ccdb56717802acf258530e2d6d237787d760a8fbcfd642790254d5ab
        - quay.io/stolostron/cluster-backup-controller@sha256:c85f88d42b42e4366f8ff7ad6c3e6b375e98c5b6a12a9c24a6333929e48f1811
        - quay.io/stolostron/observatorium-operator@sha256:5eb4c0da92c62b664aaf3bc187924870275e687440c3b0a3ad4f1c760c8b35e9
        - quay.io/stolostron/kube-rbac-proxy@sha256:5b521e4f0e5e54c53cea714020069aa796da23d1482d14e4b7c908098ac69ae9
        - quay.io/stolostron/observability-e2e-test@sha256:03403aeb59b364f578761ae4bcb302bd3589baff500d733775e8cb4f2f961236
        - quay.io/stolostron/insights-metrics@sha256:f2f4cf14090252af32832ff6df8a9f896e5314751199156c4e5ae0755e4515e6
        - quay.io/stolostron/search-collector@sha256:54ece9741641e1bb6e30f15d97a65c92b9c6408ba2572a2f6658f537134adff1
        - quay.io/stolostron/multicluster-operators-subscription@sha256:a056ce7d438f68d33ca3dfd42ec3d9a34b347cdf05499066412784109726e7ad
        - quay.io/stolostron/submariner-addon@sha256:9c72b5177de8ddc56dd3544997e9bd222895545b7c578e43e4051596fb1fa873
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/prometheus-alertmanager@sha256:b98480f8ba6ffd19fcee3b412757ad34428e6feab7e4196bb13f2f940306e2c9
        - quay.io/stolostron/insights-client@sha256:978d8b7c3e692272024b71b8c005b752c08dc643dde8258ad75961ddcdcce67f
        - quay.io/stolostron/prometheus@sha256:3d1810567f67ae1a7bb024a1c4e494ee510cfb14c7ed21a8b5e63e41fa7bcd68
        - quay.io/stolostron/flightctl-ocp-ui@sha256:6cff6425850d2b4a8b885a684afb01244d7cd9de7035d7ad71b1e10e5bab25e9
        - quay.io/stolostron/kube-state-metrics@sha256:cd24916e55ad905e386fa3df8a648272114353a0d75283224a4dddb1955b2687
        - quay.io/stolostron/flightctl-ui@sha256:fea717f3b8b4c343ec60ce7d74a691f1124b6847d55e30a73dc0d50ea610d370
        - quay.io/stolostron/multicluster-operators-channel@sha256:501a5053bd3fd5cc9de34e7e020890d731321b090a0e5bf11b8b3350d002f573
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/stolostron/multiclusterhub-operator@sha256:09f1f4266938fba88376a50c3e2837c5517b4f3a947cf86638a17e52eb17c1f6
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/prometheus-operator@sha256:f80e79ea87cde66955e1e46d264e414c072085da9db6876ee2ae68c87d2452e5
        - quay.io/stolostron/governance-policy-framework-addon@sha256:047124271d0b30176a94454b293c82c2eb67a6f06757fc22c54c455fd0913157
        - quay.io/stolostron/console@sha256:eed52f3ab860a3a224ce178d92765ca50d72d6261f04df040a39577881738252
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/multicloud-integrations@sha256:0081065f9e59c5ed41087344668ac86012f796cdbbf3575fd8491c7dfd24fbf8
        - quay.io/stolostron/must-gather@sha256:2bf9d3c15f32a6bfe0a7a8d39b9dc62ac6cbb6fe47b7cf1f6bc5e30b17e5e849
        - quay.io/stolostron/grafana-dashboard-loader@sha256:6c6c2aa528373536d5db66fffeb87a85b79f3cf723e5d9eeaefcb74f91c1a8a1
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:05da11b343938866c5afcfe844bcd15d9c513fdb10fd7ea0d3473b455b427c2a
        - quay.io/stolostron/rbac-query-proxy@sha256:096ad23c9b9c7e90451c09b48ded67a82575e75c8cc19f22c459d63ae8b5a786
        version: 0.6.3
      entries:
      - name: stolostron.v0.6.3
        version: 0.6.3
      - name: stolostron.v0.6.2
        version: 0.6.2
      - name: stolostron.v0.6.1
        version: 0.6.1
      - name: stolostron.v0.6.0
        version: 0.6.0
      name: community-0.6
    - currentCSV: stolostron.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:ab6660283a25bdf2ea9afe8c76bfbc32f4610d69af2bb13a7bca37a1f5961bdd
          createdAt: "2025-06-09T15:10:05Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.7.1'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MulticlusterHub defines the configuration
              for an instance of a multicluster hub, a central point for managing multiple
              Kubernetes-based clusters. The deployment of multicluster hub components
              is determined based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/rbac-query-proxy@sha256:1c198b6388a72ccfe2c5b68d57a2b22893fb65d977992ec44ee8cad6ee1ca976
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/stolostron/governance-policy-addon-controller@sha256:6ecc7557168177474423cea60eadc07d8382adb3e97dee669a8590859a41b08c
        - quay.io/stolostron/flightctl-api@sha256:64fb1b655bef6aeff03cf64b6450155d3cb570ed4713f084590fe515f4fbad02
        - quay.io/stolostron/cert-policy-controller@sha256:6deb41867f8781f4e34f69a54d4aeca190d588ba2991279d39cb5a82c2a5c5be
        - quay.io/stolostron/flightctl-ui@sha256:fea717f3b8b4c343ec60ce7d74a691f1124b6847d55e30a73dc0d50ea610d370
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/cluster-backup-controller@sha256:e576c2de1f931b7d33da2dbc03d71ce323766bd9158bf4b128221ddd1ab3fb9f
        - quay.io/stolostron/observatorium@sha256:18a06117ef114a1e68474fd9832adacd30b03f60ec1aa55374285a20e5311f13
        - quay.io/stolostron/insights-metrics@sha256:861dc03fff55c1d9aff96b5ed15b9d38d5932df1c00a4423e35ead19da6eda03
        - quay.io/stolostron/siteconfig-operator@sha256:16b00f6c4d81eea26e663ca5596526d85bf7777728d1f0ca7e4dd765c64d2bb5
        - quay.io/stolostron/search-indexer@sha256:58320b54847e3bb6baa558bda65e9d0ae2198df480ba27f46f96ba1454614508
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:4b29cc08c9e35054a4a820472ea19cd51e043ab1a90495418a36922df74b9540
        - quay.io/stolostron/multiclusterhub-operator@sha256:ab6660283a25bdf2ea9afe8c76bfbc32f4610d69af2bb13a7bca37a1f5961bdd
        - quay.io/stolostron/prometheus-alertmanager@sha256:a9da6f01cfa61fef803f09ef3f6b27e33a07b97fc1f1e11b6a6c8e13b8a4b90e
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/volsync-addon-controller@sha256:224a0a03a25e5ce61f4c50f6252370876ac0cbf4f9711e997e98a00d6a124fe9
        - quay.io/stolostron/origin-cli@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/search-v2-operator@sha256:2ae9ab48615966ccf91b2022ccda12913a3db3cfbad5351dd8a20648b01137d8
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/console@sha256:ba89196efe30c812a767d58a41170b56e2b04167a72e7f13180bf40474bdc824
        - quay.io/stolostron/prometheus-config-reloader@sha256:024ee6dda10c6d22ddeab57d7d6503b7aa9ac78981cfcec8ffae94b4914500e3
        - quay.io/stolostron/submariner-addon@sha256:51359ccdc0ea29bcaf7ec2b190878e9f3333b1530cb1128e142d893fb87eba8b
        - quay.io/stolostron/prometheus-operator@sha256:8c1c12562ec00da8eaedb7d428694ebdd3a749c0b1ce03fe2b5663ae40e1c701
        - quay.io/stolostron/flightctl-periodic@sha256:a3ed745dc65a5886c4cc05348cf4822bfbf2c6d28b85a39fa0b8f3e1f53fdf8a
        - quay.io/stolostron/governance-policy-framework-addon@sha256:b1b24c40d68415525e2bae21810fc439d3e96fcc5fa75c92e7f2eee284422bfd
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:05da11b343938866c5afcfe844bcd15d9c513fdb10fd7ea0d3473b455b427c2a
        - quay.io/stolostron/helloworld@sha256:4465550664bd210cf0df8edfd706de2fc51cacf01b66a962b5607935f9666ebc
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:54c6a2eec497eafdbf861fb95d78d961fb24906b197fc69b4e96d3e8b53e6496
        - quay.io/stolostron/multicluster-observability-operator@sha256:3d95297297a301fca1915268fec71d7a6ecef98c6c2ab8494d9befacb2e169fa
        - quay.io/stolostron/thanos@sha256:9d504e2aa1001135eb6e454c58c620d102f2b0a85016b7587452dd1595794c93
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/must-gather@sha256:ad8db72d559d06f4d467f78213d4f8f2cb00f8e76e942d8e75394eb7dadd29d2
        - quay.io/stolostron/acm-cli@sha256:2f0cec3bed4ad77f573e8cd51a7e2c62d68a59e4e261ea801d5ac203fd2724fc
        - quay.io/stolostron/grafana@sha256:cebaed7e5abe0064c65ea69bd4e14794d153de9172c4c3b6de4730f23fed7d24
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/metrics-collector@sha256:c7433ee4ccdb56717802acf258530e2d6d237787d760a8fbcfd642790254d5ab
        - quay.io/stolostron/observability-e2e-test@sha256:4e60d3e3e1b5dbab3bcc1c4d37a4940a2a42a065a9df539fb243fb174dde6716
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/search-v2-api@sha256:786866ca8c71a8263ddef45ca9bd9e48cfd1b3ad6215725fbf6f79967bda20ad
        - quay.io/stolostron/multicluster-operators-channel@sha256:318d885ab33d5afe17334e831c955f1c3cdf804281b7cc8523330820ac22eee2
        - quay.io/stolostron/multicloud-integrations@sha256:665e9385531674a5849aa661eb207a0c6e39ea563fc891c57678cddbf7878d8e
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/observatorium-operator@sha256:2f11444703addec163a00c99529c228594fb52df13de47f08faac2e446bec483
        - quay.io/stolostron/governance-policy-propagator@sha256:2b77422fb03177f647c04716e7242bc08f241a1ef9c8968e492ae57816f4c4fb
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/grafana-dashboard-loader@sha256:6c6c2aa528373536d5db66fffeb87a85b79f3cf723e5d9eeaefcb74f91c1a8a1
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/cluster-permission@sha256:e45139d459951b2e751018fe7ecabe1ff22b90ac87be81fb8f17a12f7b7f6a65
        - quay.io/stolostron/search-collector@sha256:2b3dfd21c0ba7b7b84fac390002e47fae927cd5fdf13e5d1d960703aa6c5bede
        - quay.io/stolostron/insights-client@sha256:2cbd20252f10a5a64f93af50f117f05a865aa9d578e5f6e55e4249470b761604
        - quay.io/stolostron/kube-rbac-proxy@sha256:e5555e256a518ca34bea01c959bbb5497582b53dbfaaf326e1852430c175637a
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:1efd8e470b88a992e2f35ce18bdf4970f91af7ba917b8d92627b47ce8c10eb40
        - quay.io/stolostron/flightctl-worker@sha256:4dddded7d2d4cce63972c8ca81f213c2cedb551cc645ec85495a9692dc36bb4c
        - quay.io/stolostron/config-policy-controller@sha256:b1718d5facb95c5572ac658c899f9fce0a7e8c08de93804de3327cd50cdb073b
        - quay.io/stolostron/redis-7-c9s@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/search-e2e@sha256:1c7dcc9f3588620b3570a493c2d744de059f425b572b9cb5725142b89414a904
        - quay.io/stolostron/flightctl-ocp-ui@sha256:6cff6425850d2b4a8b885a684afb01244d7cd9de7035d7ad71b1e10e5bab25e9
        - quay.io/stolostron/thanos-receive-controller@sha256:f445bbfc18bbc4646e0432842022d3be5305809e52ddcb9daf0ec000844d5854
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/klusterlet-addon-controller@sha256:4cb2f4f4e14cd67a4b2f9549cba714b635d9e1c2d8b869b0e0573fd7a41bc13e
        - quay.io/stolostron/multicluster-operators-application@sha256:9f804b4f2da910e0d7f122176f72c0b0ef4b30785fde0369a66f95f4bac58a82
        - quay.io/stolostron/prometheus@sha256:b959035ff754050ca092863bb74c35ba9622d9adc9e8eb7caf3b430cfe211730
        - quay.io/stolostron/grc-policy-framework-tests@sha256:327122b13ec79ea488d6f424806549e36ebfbabd055f9b40217ba541f4a118a9
        - quay.io/stolostron/multicluster-operators-subscription@sha256:d7c137379781ce6b697f73a464b89d04fbc20c3da00ce21a195ac79ab57a6dc7
        - quay.io/stolostron/kube-state-metrics@sha256:3503df61f3d9ebf3cbf66deef865450b32c7403839fdaca42504febaac00687c
        version: 0.7.1
      entries:
      - name: stolostron.v0.7.1
        version: 0.7.1
      - name: stolostron.v0.7.0
        version: 0.7.0
      name: community-0.7
    defaultChannel: community-0.7
    packageName: stolostron
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MinIO
      provider-url: https://min.io
    name: minio-object-store-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-object-store-operator.v2025.8.19175300
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aistor.min.io/v1",
                "kind": "ObjectStore",
                "metadata": {
                  "annotations": {
                    "prometheus.io/path": "/minio/v2/metrics/cluster",
                    "prometheus.io/port": "9000",
                    "prometheus.io/scrape": "true"
                  },
                  "labels": {
                    "app": "minio"
                  },
                  "name": "my-objectstore"
                },
                "spec": {
                  "cache": {
                    "enabled": false
                  },
                  "certificates": {
                    "disableAutoCert": false
                  },
                  "configuration": {
                    "name": "storage-configuration"
                  },
                  "firewall": {
                    "enabled": false
                  },
                  "mountPath": "/export",
                  "podManagementPolicy": "Parallel",
                  "pools": [
                    {
                      "name": "pool-0",
                      "securityContext": {
                        "fsGroupChangePolicy": "OnRootMismatch"
                      },
                      "servers": 4,
                      "volumeClaimTemplate": {
                        "apiVersion": "v1",
                        "kind": "persistentvolumeclaims",
                        "spec": {
                          "accessModes": [
                            "ReadWriteOnce"
                          ],
                          "resources": {
                            "requests": {
                              "storage": "1Ti"
                            }
                          },
                          "storageClassName": "standard"
                        }
                      },
                      "volumesPerServer": 4
                    }
                  ]
                }
              },
              {
                "apiVersion": "aistor.min.io/v1alpha1",
                "kind": "AdminJob",
                "metadata": {
                  "name": "minio-job"
                },
                "spec": {
                  "commands": [
                    {
                      "args": {
                        "name": "iac-bucket"
                      },
                      "op": "make-bucket"
                    }
                  ],
                  "containerSecurityContext": {},
                  "objectStore": {
                    "name": "my-objectstore",
                    "namespace": "my-objectstore"
                  },
                  "securityContext": {},
                  "serviceAccountName": "my-objectstore-sa"
                }
              },
              {
                "apiVersion": "sts.min.io/v1beta1",
                "kind": "PolicyBinding",
                "metadata": {
                  "name": "example-binding"
                },
                "spec": {
                  "application": {
                    "namespace": "my-objectstore",
                    "serviceaccount": "my-objectstore-sa"
                  },
                  "policies": [
                    "consoleAdmin"
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,AI/Machine Learning
          createdAt: "2025-08-27T14:50:56Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AdminJob
            name: adminjobs.aistor.min.io
            version: v1alpha1
          - kind: ObjectStore
            name: objectstores.aistor.min.io
            version: v1
          - kind: PolicyBinding
            name: policybindings.sts.min.io
            version: v1alpha1
          - kind: PolicyBinding
            name: policybindings.sts.min.io
            version: v1beta1
        description: "MinIO AIStor extends the most sophisticated object store on
          the market. It is designed for the exascale data \ninfrastructure challenges
          presented by modern AI workloads. AIStor provides new features, along with
          performance and\nscalability improvements, to enable enterprises to store
          all AI data in one infrastructure.\n\nMinIO is the standard for building
          <a href=\"https://min.io/solutions/object-storage-for-ai\">large scale AI
          data infrastructure</a>.\n\nThe MinIO AIStor is a software-defined, Amazon
          S3 compatible object store that will run anywhere - from the public cloud
          to the edge.<br><br>\n\nEnterprises use MinIO to deliver against artificial
          intelligence, machine learning, analytics, application, backup and archival
          workloads - all from a single platform.\n\nThe use cases include storage
          for HDFS replacements, AI datalakes, SIEM datalakes, analytics datalakes,
          object storage as a service, streaming and database warm/hot storage.<br><br>\n\nMinIO’s
          AIStor is the fastest object storage server, with production READ/WRITE
          speeds in excess of 2.2 TiB/s on commodity hardware.\n\nBuilt for the cloud
          operating model it is native to the technologies and architectures that
          define the cloud. These include containerization, orchestration with Kubernetes,
          microservices and multi-tenancy - making it the most widely integrated object
          store in the market today.<br><br>\n\nThe AIStor offers a host of enterprise
          features including inline erasure coding, bit-rot detection, state-of-the-art
          encryption, active-active replication, object locking, lifecycle management,
          observability, key-management, caching, firewall/load balancing, global
          console, search and identity + access management.<br><br>\n\nThe AIStor
          is priced on usable capacity and includes support through the MinIO Subscription
          Network.\n\nThe Subscription Network is designed to optimize deployments
          of MinIO and reduce the risk associated with data loss and data breach while
          providing 24/7/365 direct-to-engineering support.<br><br>\n\n<a href=\"https://resources.min.io/request-a-demo-marketplaces/\">Request
          a demo today</a>"
        displayName: Minio AIStor Object Store
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Storage
        - S3
        - MinIO
        - Object Storage
        links:
        - name: Minio AIStor
          url: https://min.io
        - name: Support
          url: https://subnet.min.io
        - name: Github
          url: https://github.com/minio/aistor
        maintainers:
        - email: dev@min.io
          name: MinIO
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: MinIO
          url: https://min.io
        relatedImages:
        - quay.io/minio/aistor/operator@sha256:f80139efa062fdd83f78318b20f371a827b68e854b03dd9706b9eba4c493aeb2
        version: 2025.8.19175300
      entries:
      - name: minio-object-store-operator.v2025.8.19175300
        version: 2025.8.19175300
      - name: minio-object-store-operator.v2025.8.1184927
        version: 2025.8.1184927
      - name: minio-object-store-operator.v2025.7.1011319
        version: 2025.7.1011319
      - name: minio-object-store-operator.v2025.6.3065135
        version: 2025.6.3065135
      - name: minio-object-store-operator.v2025.5.12190907
        version: 2025.5.12190907
      name: stable
    defaultChannel: stable
    packageName: minio-object-store-operator
    provider:
      name: MinIO
      url: https://min.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: https://stakater.com
    name: template-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: template-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-07-04T00:40:36Z"
          description: Operator to distribute plain kubernetes manifests, helm charts
            and, copy secrets and configmaps to multiple namespaces
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.0.1
          operatorframework.io/suggested-namespace: template-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterTemplateInstance is the Schema for the clustertemplateinstances
              API
            displayName: Cluster Template Instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateGroupInstance is the Schema for the TemplateGroupInstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.templates.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
        description: '"Operator to distribute plain kubernetes manifests, helm charts
          and, copy secrets and configmaps to multiple namespaces"'
        displayName: Template Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - templates
        - secrets
        - configmaps
        - distribution
        - stakater
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/template-operator@sha256:ad8312417e4e9621b74d3ad03020516840feef9c4b52c4f0d894dbb6dc6f906a
        version: 0.0.1
      entries:
      - name: template-operator.v0.0.1
        version: 0.0.1
      name: release-0.0
    - currentCSV: template-operator.v0.1.5
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-08-29T06:41:42Z"
          description: Operator to distribute plain kubernetes manifests, helm charts
            and, copy secrets and configmaps to multiple namespaces
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.1.5
          operatorframework.io/suggested-namespace: template-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterTemplateInstance is the Schema for the clustertemplateinstances
              API
            displayName: Cluster Template Instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateGroupInstance is the Schema for the TemplateGroupInstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.templates.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
        description: '"Operator to distribute plain kubernetes manifests, helm charts
          and, copy secrets and configmaps to multiple namespaces"'
        displayName: Template Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - templates
        - secrets
        - configmaps
        - distribution
        - stakater
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/template-operator@sha256:88f722baf22ac07704503f098f2b0ea7d76815a1ab0552474b1c9ba6639d8142
        version: 0.1.5
      entries:
      - name: template-operator.v0.1.5
        version: 0.1.5
      name: release-0.1
    defaultChannel: release-0.1
    packageName: template-operator
    provider:
      name: Stakater
      url: https://stakater.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Starburst Data, Inc.
      provider-url: https://www.starburstdata.com
    name: starburst-enterprise-helm-operator-paygo-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: starburst-enterprise-helm-operator-paygo.v402.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstEnterprise",
                "metadata": {
                  "name": "starburstenterprise-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "cache": {
                    "diskUsagePercentage": 80,
                    "enabled": false,
                    "ttl": "7d",
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "catalogs": {
                    "tpch": "connector.name=tpch\ntpch.splits-per-node=4"
                  },
                  "commonLabels": {},
                  "coordinator": {
                    "additionalProperties": "",
                    "affinity": {},
                    "deploymentAnnotations": {},
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=true\nnode-scheduler.include-coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://localhost:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "resources": {
                      "cpu": 16,
                      "memory": "60Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "tolerations": []
                  },
                  "environment": null,
                  "expose": {
                    "clusterIp": {
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      }
                    },
                    "ingress": {
                      "annotations": {},
                      "host": null,
                      "ingressClassName": null,
                      "ingressName": "coordinator-ingress",
                      "path": "/",
                      "pathType": "ImplementationSpecific",
                      "serviceName": "starburst",
                      "servicePort": 8080,
                      "tls": {
                        "enabled": true,
                        "secretName": null
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "nodePort": 30080,
                          "port": 8080
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "externalSecrets": {
                    "enabled": false,
                    "goDaddy": {
                      "backendType": "secretsManager"
                    },
                    "secretPrefix": "external0",
                    "type": "goDaddy"
                  },
                  "extraArguments": [],
                  "extraSecret": {
                    "file": null,
                    "name": null
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise",
                    "tag": "402-e.1"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "initImage": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise-init",
                    "tag": "1.3.2"
                  },
                  "internal": {
                    "ports": {
                      "http": {
                        "port": 8080
                      },
                      "https": {
                        "port": 8443
                      }
                    }
                  },
                  "internalTls": false,
                  "licenseManager": {
                    "cronjob": {
                      "affinity": {},
                      "nodeSelector": {},
                      "priorityClassName": null,
                      "securityContext": {},
                      "tolerations": []
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "registry.connect.redhat.com/starburst/license-manager",
                      "tag": "0.0.2"
                    },
                    "url": "https://license-api.dogfood.eng.starburstdata.net/test/license/checkout"
                  },
                  "livenessProbe": null,
                  "prometheus": {
                    "agent": {
                      "config": "/etc/starburst/telemetry/prometheus.yaml",
                      "port": 8081,
                      "version": "0.16.1"
                    },
                    "enabled": true,
                    "rules": [
                      {
                        "attrNameSnakeCase": true,
                        "name": "$1",
                        "pattern": "trino.execution<name=QueryManager><>(running_queries|queued_queries)",
                        "type": "GAUGE"
                      },
                      {
                        "name": "failed_queries",
                        "pattern": "trino.execution<name=QueryManager><>FailedQueries\\.TotalCount",
                        "type": "COUNTER"
                      }
                    ]
                  },
                  "query": {
                    "maxConcurrentQueries": 3
                  },
                  "readinessProbe": null,
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "securityContext": {},
                  "sharedSecret": null,
                  "siac": {
                    "additionalHeapSizePercentage": 15,
                    "enabled": false
                  },
                  "spilling": {
                    "enabled": false,
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "startupProbe": null,
                  "usageMetrics": {
                    "enabled": false,
                    "usageClient": {
                      "initialDelay": "1m",
                      "interval": "1m"
                    }
                  },
                  "userDatabase": {
                    "enabled": false,
                    "users": [
                      {
                        "password": "46991b33f7a75ff79213c0dc0e610610",
                        "username": "admin"
                      }
                    ]
                  },
                  "worker": {
                    "additionalProperties": "",
                    "affinity": {},
                    "autoscaling": {
                      "enabled": false,
                      "maxReplicas": 100,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "deploymentAnnotations": {},
                    "deploymentTerminationGracePeriodSeconds": 300,
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://coordinator:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "kedaScaler": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.connect.redhat.com/starburst/keda-trino-scaler",
                        "tag": "0.1.7"
                      },
                      "port": 8021,
                      "scaledObjectSpec": {
                        "advanced": {},
                        "cooldownPeriod": 300,
                        "idleReplicaCount": 0,
                        "maxReplicaCount": 100,
                        "minReplicaCount": 1,
                        "pollingInterval": 30,
                        "scaleTargetRef": {
                          "name": "worker"
                        },
                        "triggers": [
                          {
                            "metadata": {
                              "numberOfQueriesPerWorker": "10",
                              "scaleInToIdleReplicaCountIfNoQueuedQueriesLeft": "true",
                              "scaleMethod": "query_queue"
                            },
                            "type": "external"
                          }
                        ]
                      }
                    },
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "replicas": 2,
                    "resources": {
                      "cpu": 16,
                      "memory": "100Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "starburstWorkerShutdownGracePeriodSeconds": 120,
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstHive",
                "metadata": {
                  "name": "starbursthive-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "affinity": {},
                  "commonLabels": {},
                  "database": {
                    "external": {
                      "driver": null,
                      "jdbcUrl": null,
                      "password": null,
                      "user": null
                    },
                    "internal": {
                      "databaseName": "hive",
                      "databasePassword": "HivePass1234",
                      "databaseUser": "hive",
                      "driver": "org.postgresql.Driver",
                      "env": [],
                      "envFrom": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.redhat.io/rhel8/postgresql-10",
                        "tag": "1"
                      },
                      "port": 5432,
                      "resources": {
                        "limits": {
                          "cpu": 2,
                          "memory": "1Gi"
                        },
                        "requests": {
                          "cpu": 2,
                          "memory": "1Gi"
                        }
                      },
                      "volume": {
                        "emptyDir": {}
                      }
                    },
                    "type": "internal"
                  },
                  "deploymentAnnotations": {},
                  "env": [],
                  "envFrom": [],
                  "expose": {
                    "clusterIp": {
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "nodePort": 30083,
                          "port": 9083
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "extraArguments": [],
                  "hdfs": {
                    "hadoopUserName": null
                  },
                  "heapSizePercentage": 85,
                  "hiveMetastoreStorageAuthorization": true,
                  "hiveMetastoreWarehouseDir": null,
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/hive",
                    "tag": "3.1.2-e.22"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "nodeSelector": {},
                  "objectStorage": {
                    "adl": {
                      "oauth2": {
                        "clientId": null,
                        "credential": null,
                        "refreshUrl": null
                      }
                    },
                    "awsS3": {
                      "accessKey": null,
                      "endpoint": null,
                      "pathStyleAccess": false,
                      "region": null,
                      "secretKey": null
                    },
                    "azure": {
                      "abfs": {
                        "accessKey": {
                          "accessKey": null,
                          "storageAccount": null
                        },
                        "authType": "accessKey",
                        "oauth": {
                          "clientId": null,
                          "endpoint": null,
                          "secret": null
                        }
                      },
                      "wasb": {
                        "accessKey": null,
                        "storageAccount": null
                      }
                    },
                    "gs": {
                      "cloudKeyFileSecret": null
                    }
                  },
                  "podAnnotations": {},
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "resources": {
                    "limits": {
                      "cpu": 1,
                      "memory": "1Gi"
                    },
                    "requests": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "securityContext": {},
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
          description: Deploys StarburstEnterprise Helm Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-paygo-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-paygo-rhmp/support?utm_source=openshift_console
          marketplace.redhat.com/meterDefinition: |-
            {
              "apiVersion": "marketplace.redhat.com/v1beta1",
              "kind": "MeterDefinition",
              "metadata": {
                "name": "sep-vcpu-usage"
              },
              "spec": {
                "group": "charts.starburstdata.com",
                "kind": "StarburstEnterprise",
                "resourceFilters": [
                  {
                    "namespace": {
                      "useOperatorGroup": true
                    },
                    "ownerCRD": {
                      "apiVersion": "charts.starburstdata.com/v1alpha1",
                      "kind": "StarburstEnterprise"
                    },
                    "workloadType": "Pod"
                  }
                ],
                "meters": [
                  {
                    "aggregation": "sum",
                    "period": "1h",
                    "metricId": "container_cpu_usage_core_avg",
                    "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=~\"coordinator|worker\"}[5m])*100",
                    "workloadType": "Pod"
                  }
                ]
              }
            }
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Starburst Data
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: StarburstEnterprise
            name: starburstenterprises.charts.starburstdata.com
            version: v1alpha1
          - kind: StarburstHive
            name: starbursthives.charts.starburstdata.com
            version: v1alpha1
        description: CRs for deploying starburst enterprise and hive
        displayName: Starburst Enterprise Pay as you go Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - starburst
        - trino
        - sep
        - bigdata
        - data
        links:
        - name: Starburst Enterprise Helm Operator
          url: https://www.starburstdata.com
        maintainers:
        - email: hello@starburstdata.com
          name: Starburst Data, Inc.
        maturity: alpha
        provider:
          name: Starburst Data, Inc.
          url: https://www.starburstdata.com
        relatedImages:
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
        version: 402.1.0
      entries:
      - name: starburst-enterprise-helm-operator-paygo.v402.1.0
        version: 402.1.0
      name: alpha
    defaultChannel: alpha
    packageName: starburst-enterprise-helm-operator-paygo-rhmp
    provider:
      name: Starburst Data, Inc.
      url: https://www.starburstdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: loki-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-operator.v5.9.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:6728046841367066216b304ed06249b80c57c74151c6a75c9df04a6f7f76e03d
          createdAt: "2025-01-30T17:34:53Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.7.0-0 <5.9.15'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:0c3943fa009f0c649f6d592b0ac410653c8c44b4a90e41fabdfa5f3f7db1e43d
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:6728046841367066216b304ed06249b80c57c74151c6a75c9df04a6f7f76e03d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:2e164e377d4fba98e7f63f43f25b6b9dad441d2e2914c2bf02f3525d5c9aa483
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:542b28fd51194b46efc4e5af10c3241b52db4c71f0d5d1b1f184eb39764e6c53
        version: 5.9.15
      entries:
      - name: loki-operator.v5.9.15
        version: 5.9.15
      - name: loki-operator.v5.9.14
        version: 5.9.14
      - name: loki-operator.v5.9.13
        version: 5.9.13
      - name: loki-operator.v5.9.12
        version: 5.9.12
      - name: loki-operator.v5.9.11
        version: 5.9.11
      - name: loki-operator.v5.9.10
        version: 5.9.10
      - name: loki-operator.v5.9.9
        version: 5.9.9
      - name: loki-operator.v5.9.8
        version: 5.9.8
      - name: loki-operator.v5.9.7
        version: 5.9.7
      - name: loki-operator.v5.9.6
        version: 5.9.6
      - name: loki-operator.v5.9.5
        version: 5.9.5
      - name: loki-operator.v5.9.4
        version: 5.9.4
      - name: loki-operator.v5.9.3
        version: 5.9.3
      - name: loki-operator.v5.9.2
        version: 5.9.2
      - name: loki-operator.v5.9.1
        version: 5.9.1
      - name: loki-operator.v5.9.0
        version: 5.9.0
      name: stable
    - currentCSV: loki-operator.v5.9.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:6728046841367066216b304ed06249b80c57c74151c6a75c9df04a6f7f76e03d
          createdAt: "2025-01-30T17:34:53Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.7.0-0 <5.9.15'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:2e164e377d4fba98e7f63f43f25b6b9dad441d2e2914c2bf02f3525d5c9aa483
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:542b28fd51194b46efc4e5af10c3241b52db4c71f0d5d1b1f184eb39764e6c53
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:0c3943fa009f0c649f6d592b0ac410653c8c44b4a90e41fabdfa5f3f7db1e43d
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:6728046841367066216b304ed06249b80c57c74151c6a75c9df04a6f7f76e03d
        version: 5.9.15
      entries:
      - name: loki-operator.v5.9.15
        version: 5.9.15
      - name: loki-operator.v5.9.14
        version: 5.9.14
      - name: loki-operator.v5.9.13
        version: 5.9.13
      - name: loki-operator.v5.9.12
        version: 5.9.12
      - name: loki-operator.v5.9.11
        version: 5.9.11
      - name: loki-operator.v5.9.10
        version: 5.9.10
      - name: loki-operator.v5.9.9
        version: 5.9.9
      - name: loki-operator.v5.9.8
        version: 5.9.8
      - name: loki-operator.v5.9.7
        version: 5.9.7
      - name: loki-operator.v5.9.6
        version: 5.9.6
      - name: loki-operator.v5.9.5
        version: 5.9.5
      - name: loki-operator.v5.9.4
        version: 5.9.4
      - name: loki-operator.v5.9.3
        version: 5.9.3
      - name: loki-operator.v5.9.2
        version: 5.9.2
      - name: loki-operator.v5.9.1
        version: 5.9.1
      - name: loki-operator.v5.9.0
        version: 5.9.0
      name: stable-5.9
    - currentCSV: loki-operator.v6.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.14-v4.17
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:1cb1d2cad8632791529eb754938612f5e5ed373284e42f309d603059ab7ec7c5
          createdAt: "2025-03-12T12:05:37Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=5.8.0-0 <6.0.10'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.0
          operators.operatorframework.io.bundle.channels.v1: stable-6.0
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:9aa8d8e4a7fd7ddad5d842a22b09aecdbdfa31ca5eb8c4b17f0818636af5fbf8
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:69e69ab2628742774d12dd9dde50607625efdff50ea4671f2859692fbee5ecb4
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:98d2cf3437faa2400c01edb3c694816e0214c8f9dca0a0b0cdf5882a720248f3
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:1cb1d2cad8632791529eb754938612f5e5ed373284e42f309d603059ab7ec7c5
        version: 6.0.10
      entries:
      - name: loki-operator.v6.0.10
        version: 6.0.10
      - name: loki-operator.v6.0.9
        version: 6.0.9
      - name: loki-operator.v6.0.8
        version: 6.0.8
      - name: loki-operator.v6.0.7
        version: 6.0.7
      - name: loki-operator.v6.0.6
        version: 6.0.6
      - name: loki-operator.v6.0.5
        version: 6.0.5
      - name: loki-operator.v6.0.4
        version: 6.0.4
      - name: loki-operator.v6.0.3
        version: 6.0.3
      - name: loki-operator.v6.0.2
        version: 6.0.2
      - name: loki-operator.v6.0.1
        version: 6.0.1
      - name: loki-operator.v6.0.0
        version: 6.0.0
      name: stable-6.0
    - currentCSV: loki-operator.v6.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.15-v4.18
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:0bde3c1a0b667a693f443688ed3926eef1705c45ecf4cce3a3dedbddca7a2784
          createdAt: "2025-03-12T11:19:10Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=5.9.0-0 <6.1.8'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.1
          operators.operatorframework.io.bundle.channels.v1: stable-6.1
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:0bde3c1a0b667a693f443688ed3926eef1705c45ecf4cce3a3dedbddca7a2784
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:731f92c146f299bc3662b309a3068bde18fad51f65a7d29063e582b61e07042f
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:c34676932ddd14db213dbf42a1d341f51692f3f47f1b751d06187e292c99a170
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b94833f55644c289fe2c1297b68a4f9931baf977587c6a2c4c7a3e484c704f64
        version: 6.1.8
      entries:
      - name: loki-operator.v6.1.8
        version: 6.1.8
      - name: loki-operator.v6.1.7
        version: 6.1.7
      - name: loki-operator.v6.1.6
        version: 6.1.6
      - name: loki-operator.v6.1.5
        version: 6.1.5
      - name: loki-operator.v6.1.4
        version: 6.1.4
      - name: loki-operator.v6.1.3
        version: 6.1.3
      - name: loki-operator.v6.1.2
        version: 6.1.2
      - name: loki-operator.v6.1.1
        version: 6.1.1
      - name: loki-operator.v6.1.0
        version: 6.1.0
      name: stable-6.1
    - currentCSV: loki-operator.v6.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.16-v4.19
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f401be018c5ba35e8ea1e68831cc20067f0f4d1be031d05b222548c1bdfb5f7e
          createdAt: "2025-02-18T18:01:39Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=6.0.0-0 <6.2.5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.2
          operators.operatorframework.io.bundle.channels.v1: stable-6.2
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f401be018c5ba35e8ea1e68831cc20067f0f4d1be031d05b222548c1bdfb5f7e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:99d1e7c5c1f0609b33de576884782550f969b28dc622810e6661d4c60a664215
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:8955d8186ae98095623f2083b6d5389d3ac9f2612fcfcf695835219c3e1216c4
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b8c5b46a94a6c85c9d382eeeb65a1d2b4014a7dc765e9abfba706c92857462dc
        version: 6.2.5
      entries:
      - name: loki-operator.v6.2.5
        version: 6.2.5
      - name: loki-operator.v6.2.4
        version: 6.2.4
      - name: loki-operator.v6.2.3
        version: 6.2.3
      - name: loki-operator.v6.2.2
        version: 6.2.2
      - name: loki-operator.v6.2.1
        version: 6.2.1
      - name: loki-operator.v6.2.0
        version: 6.2.0
      name: stable-6.2
    defaultChannel: stable-6.2
    packageName: loki-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudtrail-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudtrail-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudtrail.services.k8s.aws/v1alpha1",
                "kind": "EventDataStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudtrail.services.k8s.aws/v1alpha1",
                "kind": "Trail",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudtrail-controller:1.1.1
          createdAt: "2025-09-19T21:16:25Z"
          description: AWS CloudTrail controller is a service controller for managing
            CloudTrail resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventDataStore represents the state of an AWS cloudtrail
              EventDataStore resource.
            displayName: EventDataStore
            kind: EventDataStore
            name: eventdatastores.cloudtrail.services.k8s.aws
            version: v1alpha1
          - description: Trail represents the state of an AWS cloudtrail Trail resource.
            displayName: Trail
            kind: Trail
            name: trails.cloudtrail.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Compute Cloud (CloudTrail) resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudTrail**

          AWS CloudTrail is an AWS service that helps you enable operational and risk auditing, governance, and compliance of your AWS account. Actions taken by a user, role, or an AWS service are recorded as events in CloudTrail. Events include actions taken in the AWS Management Console, AWS Command Line Interface, and AWS SDKs and APIs.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudTrail
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudtrail
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudTrail Developer Resources
          url: https://aws.amazon.com/cloudtrail/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudtrail maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudtrail-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-cloudtrail-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudtrail-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudtrail-controller.v1.0.28
        version: 1.0.28
      - name: ack-cloudtrail-controller.v1.0.26
        version: 1.0.26
      - name: ack-cloudtrail-controller.v1.0.25
        version: 1.0.25
      - name: ack-cloudtrail-controller.v1.0.24
        version: 1.0.24
      - name: ack-cloudtrail-controller.v1.0.23
        version: 1.0.23
      - name: ack-cloudtrail-controller.v1.0.22
        version: 1.0.22
      - name: ack-cloudtrail-controller.v1.0.21
        version: 1.0.21
      - name: ack-cloudtrail-controller.v1.0.20
        version: 1.0.20
      - name: ack-cloudtrail-controller.v1.0.19
        version: 1.0.19
      - name: ack-cloudtrail-controller.v1.0.16
        version: 1.0.16
      - name: ack-cloudtrail-controller.v1.0.15
        version: 1.0.15
      - name: ack-cloudtrail-controller.v1.0.14
        version: 1.0.14
      - name: ack-cloudtrail-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudtrail-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudtrail-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudtrail-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudtrail-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudtrail-controller.v1.0.5
        version: 1.0.5
      - name: ack-cloudtrail-controller.v1.0.4
        version: 1.0.4
      - name: ack-cloudtrail-controller.v1.0.3
        version: 1.0.3
      - name: ack-cloudtrail-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudtrail-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudtrail-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudtrail-controller.v0.1.3
        version: 0.1.3
      - name: ack-cloudtrail-controller.v0.1.2
        version: 0.1.2
      - name: ack-cloudtrail-controller.v0.1.1
        version: 0.1.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudtrail-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-route53-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-route53-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/route53-controller:1.1.1
          createdAt: "2025-09-19T19:15:31Z"
          description: AWS Route53 controller is a service controller for managing
            Route53 resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HealthCheck represents the state of an AWS route53 HealthCheck
              resource.
            displayName: HealthCheck
            kind: HealthCheck
            name: healthchecks.route53.services.k8s.aws
            version: v1alpha1
          - description: HostedZone represents the state of an AWS route53 HostedZone
              resource.
            displayName: HostedZone
            kind: HostedZone
            name: hostedzones.route53.services.k8s.aws
            version: v1alpha1
          - description: RecordSet represents the state of an AWS route53 RecordSet
              resource.
            displayName: RecordSet
            kind: RecordSet
            name: recordsets.route53.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Route53 resources in AWS from within your Kubernetes cluster.

          **About Amazon Route53**

          Amazon Route 53 is a highly available and scalable Domain Name System (DNS) web service. You can use Route 53 to perform three main functions in any combination: domain registration, DNS routing, and health checking.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Route53
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - route53
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Route53 Developer Resources
          url: https://docs.aws.amazon.com/route53/index.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/route53-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-route53-controller.v1.1.1
        version: 1.1.1
      - name: ack-route53-controller.v1.1.0
        version: 1.1.0
      - name: ack-route53-controller.v1.0.6
        version: 1.0.6
      - name: ack-route53-controller.v1.0.5
        version: 1.0.5
      - name: ack-route53-controller.v1.0.3
        version: 1.0.3
      - name: ack-route53-controller.v1.0.2
        version: 1.0.2
      - name: ack-route53-controller.v0.1.0
        version: 0.1.0
      - name: ack-route53-controller.v0.0.28
        version: 0.0.28
      - name: ack-route53-controller.v0.0.27
        version: 0.0.27
      - name: ack-route53-controller.v0.0.26
        version: 0.0.26
      - name: ack-route53-controller.v0.0.25
        version: 0.0.25
      - name: ack-route53-controller.v0.0.24
        version: 0.0.24
      - name: ack-route53-controller.v0.0.23
        version: 0.0.23
      - name: ack-route53-controller.v0.0.22
        version: 0.0.22
      - name: ack-route53-controller.v0.0.20
        version: 0.0.20
      - name: ack-route53-controller.v0.0.19
        version: 0.0.19
      - name: ack-route53-controller.v0.0.18
        version: 0.0.18
      - name: ack-route53-controller.v0.0.17
        version: 0.0.17
      - name: ack-route53-controller.v0.0.16
        version: 0.0.16
      - name: ack-route53-controller.v0.0.15
        version: 0.0.15
      - name: ack-route53-controller.v0.0.14
        version: 0.0.14
      - name: ack-route53-controller.v0.0.13
        version: 0.0.13
      - name: ack-route53-controller.v0.0.9
        version: 0.0.9
      - name: ack-route53-controller.v0.0.6
        version: 0.0.6
      - name: ack-route53-controller.v0.0.4
        version: 0.0.4
      - name: ack-route53-controller.v0.0.3
        version: 0.0.3
      - name: ack-route53-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-route53-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://www.medik8s.io/
    name: self-node-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: self-node-remediation.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/self-node-remediation-operator:v0.9.0
          createdAt: "2024-07-25 14:40:52"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.9.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Medik8s
          url: https://www.medik8s.io/
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/medik8s/self-node-remediation-operator:v0.9.0
        version: 0.9.0
      entries:
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      - name: self-node-remediation.v0.7.0
        version: 0.7.0
      - name: self-node-remediation.v0.6.1
        version: 0.6.1
      - name: self-node-remediation.v0.6.0
        version: 0.6.0
      - name: self-node-remediation.v0.5.0
        version: 0.5.0
      - name: self-node-remediation.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: self-node-remediation
    provider:
      name: Medik8s
      url: https://www.medik8s.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudfront-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudfront-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "CachePolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "Distribution",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudfront-controller:1.2.1
          createdAt: "2025-09-19T21:12:14Z"
          description: AWS CloudFront controller is a service controller for managing
            CloudFront resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CachePolicy represents the state of an AWS cloudfront CachePolicy
              resource.
            displayName: CachePolicy
            kind: CachePolicy
            name: cachepolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: Distribution represents the state of an AWS cloudfront Distribution
              resource.
            displayName: Distribution
            kind: Distribution
            name: distributions.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: Function represents the state of an AWS cloudfront Function
              resource.
            displayName: Function
            kind: Function
            name: functions.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: OriginAccessControl represents the state of an AWS cloudfront
              OriginAccessControl resource.
            displayName: OriginAccessControl
            kind: OriginAccessControl
            name: originaccesscontrols.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: OriginRequestPolicy represents the state of an AWS cloudfront
              OriginRequestPolicy resource.
            displayName: OriginRequestPolicy
            kind: OriginRequestPolicy
            name: originrequestpolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: ResponseHeadersPolicy represents the state of an AWS cloudfront
              ResponseHeadersPolicy resource.
            displayName: ResponseHeadersPolicy
            kind: ResponseHeadersPolicy
            name: responseheaderspolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: VPCOrigin represents the state of an AWS cloudfront VPCOrigin
              resource.
            displayName: VPCOrigin
            kind: VPCOrigin
            name: vpcorigins.cloudfront.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudFront resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudFront**

          Amazon CloudFront is a web service that speeds up distribution of your static and dynamic web content, such as .html, .css, .js, and image files, to your users. CloudFront delivers your content through a worldwide network of data centers called edge locations.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon CloudFront
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudfront
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudFront Developer Resources
          url: https://aws.amazon.com/cloudfront/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudfront maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudfront-controller:1.2.1
        version: 1.2.1
      entries:
      - name: ack-cloudfront-controller.v1.2.1
        version: 1.2.1
      - name: ack-cloudfront-controller.v1.2.0
        version: 1.2.0
      - name: ack-cloudfront-controller.v1.1.6
        version: 1.1.6
      - name: ack-cloudfront-controller.v1.1.4
        version: 1.1.4
      - name: ack-cloudfront-controller.v1.1.3
        version: 1.1.3
      - name: ack-cloudfront-controller.v1.1.2
        version: 1.1.2
      - name: ack-cloudfront-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudfront-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudfront-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudfront-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudfront-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudfront-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudfront-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudfront-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudfront-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudfront-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudfront-controller.v0.0.15
        version: 0.0.15
      - name: ack-cloudfront-controller.v0.0.14
        version: 0.0.14
      - name: ack-cloudfront-controller.v0.0.13
        version: 0.0.13
      - name: ack-cloudfront-controller.v0.0.12
        version: 0.0.12
      - name: ack-cloudfront-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudfront-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudfront-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudfront-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudfront-controller.v0.0.6
        version: 0.0.6
      - name: ack-cloudfront-controller.v0.0.5
        version: 0.0.5
      - name: ack-cloudfront-controller.v0.0.4
        version: 0.0.4
      - name: ack-cloudfront-controller.v0.0.3
        version: 0.0.3
      - name: ack-cloudfront-controller.v0.0.2
        version: 0.0.2
      - name: ack-cloudfront-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudfront-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/openshift/lightspeed-service
    name: lightspeed-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lightspeed-operator.v1.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ols.openshift.io/v1alpha1",
                "kind": "OLSConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightspeed-operator",
                    "app.kubernetes.io/instance": "olsconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "olsconfig",
                    "app.kubernetes.io/part-of": "lightspeed-operator"
                  },
                  "name": "cluster"
                },
                "spec": {
                  "llm": {
                    "providers": [
                      {
                        "credentialsSecretRef": {
                          "name": "credentials"
                        },
                        "models": [
                          {
                            "name": "gpt-3.5-turbo-1106"
                          }
                        ],
                        "name": "OpenAI"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-08-28T16:37:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lightspeed
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Kubernetes Engine", "OpenShift Virtualization
            Engine"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lightspeed-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Red Hat OpenShift Lightspeed instance. OLSConfig is the Schema
              for the olsconfigs API
            displayName: OLSConfig
            kind: OLSConfig
            name: olsconfigs.ols.openshift.io
            version: v1alpha1
        description: |-
          OpenShift Lightspeed Operator provides generative AI-based virtual assistant which integrates into the OpenShift web console. OpenShift Lightspeed can answer natural language questions related to OpenShift Container Platform.

          To configure the OpenShift Lightspeed Operator, you need a Large Language Model (LLM) . You can host the LLM using RHELAI or RHOAI or use one of the models from the following Software-as-a-Service providers:

          - OpenAI
          - Microsoft Azure OpenAI
          - IBM WatsonX

          For more information, see [About OpenShift Lightspeed](https://docs.redhat.com/en/documentation/red_hat_openshift_lightspeed/1.0tp1/html/about/ols-about-openshift-lightspeed) in the official product documentation.

          Questions, comments, concerns, or just want to talk to the OpenShift Lightspeed team? [Email us](mailto:openshift-lightspeed-contact-requests@redhat.com)!
        displayName: OpenShift Lightspeed Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ai
        - assistant
        - openshift
        - llm
        links:
        - name: Lightspeed Operator
          url: https://github.com/openshift/lightspeed-operator
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/lightspeed-service
        relatedImages:
        - registry.redhat.io/openshift-lightspeed/lightspeed-rhel9-operator@sha256:b12803642df4d9b3ad0cf800315021483f543c6a8827b3901e57bdb65f10dc48
        - registry.redhat.io/openshift-lightspeed/openshift-mcp-server-rhel9@sha256:f1e4521ff2cc076f361c04d8c9aeac09ffcedb89a2f1dbe71f85caa27472adab
        - registry.redhat.io/openshift-lightspeed/lightspeed-service-api-rhel9@sha256:ec359967a87bc7ee2c1d6a61e553acbdb0dc85d6f240a1865bf57d3ba5dc7193
        - registry.redhat.io/openshift-lightspeed/lightspeed-console-plugin-rhel9@sha256:dfc42c2651ac8793f61ff0e54334ecb006dc4a77487bc7c6878cdf47bd4ac080
        version: 1.0.5
      entries:
      - name: lightspeed-operator.v1.0.5
        version: 1.0.5
      name: alpha
    - currentCSV: lightspeed-operator.v1.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ols.openshift.io/v1alpha1",
                "kind": "OLSConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightspeed-operator",
                    "app.kubernetes.io/instance": "olsconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "olsconfig",
                    "app.kubernetes.io/part-of": "lightspeed-operator"
                  },
                  "name": "cluster"
                },
                "spec": {
                  "llm": {
                    "providers": [
                      {
                        "credentialsSecretRef": {
                          "name": "credentials"
                        },
                        "models": [
                          {
                            "name": "gpt-3.5-turbo-1106"
                          }
                        ],
                        "name": "OpenAI"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-08-28T16:37:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lightspeed
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Kubernetes Engine", "OpenShift Virtualization
            Engine"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lightspeed-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Red Hat OpenShift Lightspeed instance. OLSConfig is the Schema
              for the olsconfigs API
            displayName: OLSConfig
            kind: OLSConfig
            name: olsconfigs.ols.openshift.io
            version: v1alpha1
        description: |-
          OpenShift Lightspeed Operator provides generative AI-based virtual assistant which integrates into the OpenShift web console. OpenShift Lightspeed can answer natural language questions related to OpenShift Container Platform.

          To configure the OpenShift Lightspeed Operator, you need a Large Language Model (LLM) . You can host the LLM using RHELAI or RHOAI or use one of the models from the following Software-as-a-Service providers:

          - OpenAI
          - Microsoft Azure OpenAI
          - IBM WatsonX

          For more information, see [About OpenShift Lightspeed](https://docs.redhat.com/en/documentation/red_hat_openshift_lightspeed/1.0tp1/html/about/ols-about-openshift-lightspeed) in the official product documentation.

          Questions, comments, concerns, or just want to talk to the OpenShift Lightspeed team? [Email us](mailto:openshift-lightspeed-contact-requests@redhat.com)!
        displayName: OpenShift Lightspeed Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ai
        - assistant
        - openshift
        - llm
        links:
        - name: Lightspeed Operator
          url: https://github.com/openshift/lightspeed-operator
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/lightspeed-service
        relatedImages:
        - registry.redhat.io/openshift-lightspeed/lightspeed-service-api-rhel9@sha256:ec359967a87bc7ee2c1d6a61e553acbdb0dc85d6f240a1865bf57d3ba5dc7193
        - registry.redhat.io/openshift-lightspeed/lightspeed-console-plugin-rhel9@sha256:dfc42c2651ac8793f61ff0e54334ecb006dc4a77487bc7c6878cdf47bd4ac080
        - registry.redhat.io/openshift-lightspeed/openshift-mcp-server-rhel9@sha256:f1e4521ff2cc076f361c04d8c9aeac09ffcedb89a2f1dbe71f85caa27472adab
        - registry.redhat.io/openshift-lightspeed/lightspeed-rhel9-operator@sha256:b12803642df4d9b3ad0cf800315021483f543c6a8827b3901e57bdb65f10dc48
        version: 1.0.5
      entries:
      - name: lightspeed-operator.v1.0.5
        version: 1.0.5
      name: stable
    defaultChannel: stable
    packageName: lightspeed-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/openshift/lightspeed-service
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kubevirt-hyperconverged
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubevirt-hyperconverged-operator.v4.16.23
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:eaacb0d177fd64e2eefbcb2dc47633dadf4d54fcbe0f0f750b078ab3f4f189e1
          createdAt: "2025-09-29 21:23:00"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.14.1 <4.16.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:3e1c68d5695bab96c382661a33581e60597b59afd21660ec760d8fe5e20ff9e3
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.16.23-1,hyperconverged-cluster-operator-test-rhel9:v4.16.23-1,kubevirt-common-instancetypes-test-rhel9:v4.16.23-1,kubevirt-tekton-tasks-test-rhel9:v4.16.23-1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:e24191be577575fa9d462c87c6f51775a0cf0f5d1812d12c4c95a68ec34b34ff
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:6e4aa75bece3c2fe076e282b3f1ad8296299ec0551e379c7c88830a0ce8ed5bf
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:1be66eee321bc47dcc0e7ff4674a485b9a4552243080d796d5f3b0d417721dc4
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:28c0bdf60404fb435cfa0cfdc6d314f789417c44c3206ce8cd1e7b99031fb529
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:659abe1d740607b2cb3850c9ab29101fa5d5c821d2a09aab5994849697a504f3
        - registry.redhat.io/container-native-virtualization/mtq-operator-rhel9@sha256:9de5ded35edfbe98121bdd4404c84dcc72c8cba1742dc27df6382ef882db2183
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:0f0478e9357f8b6ace6d844e6e424db192398195a8cea2fa6398be8996d05be6
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:2843d199dc2748749d5f48fb9f90656962e011533d79dc7e0a4f3592b12aa69a
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:6b0c7cc17abd95fce45b6ec83ec291e843eedbc90dce0ecc449aec6e64562057
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:97d37927164f0ea986f209912e6247fdab37b680c214afc4bdbb7051caaf58b1
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:cc4ff488aa97e179c4df39ac5ae8085070017106adf1c5f2222b35d8f2234ea8
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:f8e2ea3988d4a77a4e809e1fd073ffa5f99393ca3eb69a2c1cb54dd4ab60e482
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:885670f2e9ad43224cbc96f34b23dd2642de00f688bade0f6b89dd3d3fb83dbd
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:251d6fb3197a7642d86193191aa69c85a45c5521d7fc509704edd46d5d4e90a7
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:fcbadecf01b81625d55f9219428baabc33ee7231b4ccf801c90dee3095907187
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:a0af0e5d4b142f28dc72c5636df5040c4ce92c6b12ab183d3a9bc7a9d2d2f418
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:c033ddd0ed30c0d25f4c2e9b30fca4e40959bb6d74029a1f452ff048b97de88a
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:7206ac7959c07f474f454ab48bcf3522ade95efb27fd9fbd3938d1387a1420af
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:468f117a1b55af59e72a9ff46499a8d0bb00884d450d57b692e2f7b557c9bf3d
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:e01871006b9e4cb1b6f477319de63b758eac672a70a298caed2032ec19986680
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:7587832b877f32621eb26e67a5c516728f293ecd98f74dd70d5f03474a5429d6
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:68b650d6779fc4d61f05d97e6d62b80fcd638721994153205eec68f8c1d999ce
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:19ca70dceb79b10ff458f0ea41faa6e5552a5f08828620ad7177efa34166a90e
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:84dbfa51755d60e34b194c1acf679ac083feb49038112212828fe8101dcec2a5
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:8462188ec7533d2105c0dc9154d8dcf1425b96d6945eef16ea774a5dff9bba1d
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:11030a8d8cd03efdf27fe3cfc9ebeadeb20851b6de1e9d8a93ba7f2bc6a77a96
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:8cf7993326ba220f650ae2fa5383d0d6cf3e2c7f3346f0a339a9d212b8df02f6
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:c18b6e315f5577cc18d46cd76595d48d4a49026a6b2f5763c3efb26919bde5eb
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:52b4f7c7a4f5c5c5d07a720c5eefa70709f520b1d939aec4cd971d8f3c9b12bc
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:ef8260d72239ad2126ca023250bfd976f5c050fe9c0727a0f7ef007446f5363b
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:18be4ef8b6c2cedab4a2e031939fc1f9cc23020bde7fccb7997e4222b673f57f
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:def80efdcc58d545aaf4b69b784a348cb36e97afd8e0124480521a2b980fa7c3
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:b031a12e73cb75e207203fd3785619e3a5a22d7d5a8549488c035f70374235be
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:14ea56d4149b72b98b6e7f5b33ba5bdd9115e7c643fd103c9df7a26b2a06700a
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:9e358366fd11dd167255abd930016aa4fc35b1b9dc69ce27316635488efc8ec4
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:6bf45534310768308958428692b29afa25f2740d0c62165d7182b0b3e9e944cf
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:eaacb0d177fd64e2eefbcb2dc47633dadf4d54fcbe0f0f750b078ab3f4f189e1
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:91b86228495f38d5528be1a8ffec96695ad72fafb76d2ead87f7bee25540d8b6
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:19b56bbb081db7198d570c3c691e142991aaa81857ee417d47903291ed0e7a1c
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:372652529b6a8705df581e7ab8acc2ec5c3cdf8fb273448052c59963abe4d984
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:7f89f2ba59309b26cf83f58ff3ba35569182097258fa543d5d88f3a8cd64ca00
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:bf1bb2fe09e028506fe7cf383368d76e358b8c59ecc74d3618bc250c1e97271c
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:a79579b18a7df9f23c584303593a9404c64df023458d55c715aeb71dcacc92a0
        - registry.redhat.io/container-native-virtualization/mtq-lock-server-rhel9@sha256:bb1dfc204e48a13985d6c2ec9cdf17f47330617725ee6dd854f0dae0f1901ba4
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:4005d165f1485eec98ca24ce41bef79cd8088002d43ffb709ec65f842a176bdc
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:3f3554ef043c80494db536b28eb03fa7043e53bb85ce426820c252fb999ce04e
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:531702869a34f76850d332efb84ccf571fafa281014afd1787497ca9de8aed37
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:a081e7e68def13b7de369d3f66469035d61fd1feef1bee4f95ba5f7b4041d960
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:e7672121156e050f05af771f75cfe26558699da8a074bf7051f710705a27d743
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:bd532e59edae5d2edac78b9249cd6a9f0779f44e25b019d824a1b9803f078bde
        - registry.redhat.io/container-native-virtualization/mtq-controller-rhel9@sha256:a774f0bd125c66b45e35e327c9c7e0afd7947d3c5ed222b3d5d223941366ba2c
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:52e76b4208ee4e7da8a0729b9bdb294729454d0f62580d64dd104b6921f464ed
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:3b57db7aaa3c61335044ec78cbd061f78ce99cdb4b75b3ce342e1a603da4bcef
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:7e19b42a648cee71c60cb69fbcf6ebca05c938942e16d23f1cd61a619cffdcb6
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:4046e8e862e0c55f09e9031b95e6a77af4d491593e39770c644287d44888b0bb
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:517c88a6be8842d4eb71ad30f99ce1afca954af1df2c1bf4a9584f4ec69fcf12
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:3e1c68d5695bab96c382661a33581e60597b59afd21660ec760d8fe5e20ff9e3
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:cca121312777d2727fe2c50a14583243ce69e6e356b499f133736787522095b2
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:ad24eb52bd65495224661d4894336d88fde342aeaab16b1084ceff6337f92059
        version: 4.16.23
      entries:
      - name: kubevirt-hyperconverged-operator.v4.16.23
        version: 4.16.23
      - name: kubevirt-hyperconverged-operator.v4.16.22
        version: 4.16.22
      - name: kubevirt-hyperconverged-operator.v4.16.21
        version: 4.16.21
      - name: kubevirt-hyperconverged-operator.v4.16.20
        version: 4.16.20
      - name: kubevirt-hyperconverged-operator.v4.16.19
        version: 4.16.19
      - name: kubevirt-hyperconverged-operator.v4.16.18
        version: 4.16.18
      - name: kubevirt-hyperconverged-operator.v4.16.17
        version: 4.16.17
      - name: kubevirt-hyperconverged-operator.v4.16.16
        version: 4.16.16
      - name: kubevirt-hyperconverged-operator.v4.16.15
        version: 4.16.15
      - name: kubevirt-hyperconverged-operator.v4.16.14
        version: 4.16.14
      - name: kubevirt-hyperconverged-operator.v4.16.13
        version: 4.16.13
      - name: kubevirt-hyperconverged-operator.v4.16.12
        version: 4.16.12
      - name: kubevirt-hyperconverged-operator.v4.16.11
        version: 4.16.11
      - name: kubevirt-hyperconverged-operator.v4.16.10
        version: 4.16.10
      - name: kubevirt-hyperconverged-operator.v4.16.9
        version: 4.16.9
      - name: kubevirt-hyperconverged-operator.v4.16.8
        version: 4.16.8
      name: candidate
    - currentCSV: kubevirt-hyperconverged-operator.v4.99.0-0.1723448771
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:5b2c64218342a5744182fc06f550bcdcf954e931275a9d1505ef0bbfbd0f8044
          createdAt: "2024-08-12 07:46:45"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.14.1 <4.16.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:d74c629e415d177069a0570fb8db9252d2b193d2ec7972c844686ffd14033f62
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.99.0-73,kubevirt-common-instancetypes-test-rhel9:v4.99.0-32,kubevirt-tekton-tasks-test-rhel9:v4.99.0-31
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:cc190d7c44b20f10ae61709fd2f5f7c0b248df73b768eb696d5a6cec90a34e59
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:41ee360ac43ee69ccee62d4ac3d2f07c8a8589e5b85c659b47018dd36c2b932f
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:b410508ecc71b0a3bf4825793efde4b07085e2243d8b3585da852fd1940c4be1
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:820a9e6bd0329df5c541ef0b67b3c7cdd9f81be33de87c5b5e5e43777bea197e
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:811e8e7e9872408794ef36229c42c6c437b1a4ace76c7134d91bb2331fe7815b
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:0769512df0fa1325d76a5f10a1a704cf60fc65f62f57b157ea53bc9ba682851f
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:8c64b526fd42a17c909bf45eb795121a81d502424a9cf11f8087118751333894
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:5f23fc8e9bc386622460c23b9644f58648d9f748f5670927a02888cd36c99a21
        - registry.redhat.io/container-native-virtualization/mtq-controller-rhel9@sha256:ad5938ebd5f07196ce5c35232a1c3b0a653301a83929ce84c3d1cb4dd1b91816
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:37e6b4fa5ff436572412ec0bdb1edc4a49e62c994e056c93578605f6b74adbcd
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:07435335fdbf6178bc00cc02b93cc0adb42784339151b2d9c31dc8da98b8bc6a
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:cffec516ab9ddc50cab77339d0cb3239ef4e7fd43bd2afdc73370aa487ea5704
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:dec430b74887e9b42d639f14c07d9f58a4881cac24e949dab98f96af150f1958
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:c61732be8024b4ede45eec8f786e51b198632ee5bfb0c2f5af84f37544584283
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:5ba79dd6bdf9e3778d3e301919b37e45cfc1eaf72fda77eba6588788829ba29b
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:c14cd037d14489321629908fc6b30248f7a20d866619177a165d72cfc1b6b7ef
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:83c03929cf11d7202e6c6cb9dd767bd4d7473f4fc1d562b79719ceb3d8382e68
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:6b7beed3cba9093b3aeb3109f23ec38c20f5568915dff774ad5e9e6bb5db3a58
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:e39732ebd3a4452fa7a678aeea5e64a7973fd72802eff193b16680c2f0129cb1
        - registry.redhat.io/container-native-virtualization/mtq-lock-server-rhel9@sha256:01cc6797800ccfb99453858ab36c0100c5ca343f1f71f975914b2437d1f757ba
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:9822f7a1590d830ccd64cc0729d2e30a2e04fb2b022f54498e1f1abff3748c2c
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:bf2ec130d1f3f32314d740037313f49f874a271e8b9a7f6cd2b93698a8423e8d
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:42a34caa18b5f5452b9d5603602ea74b1ee59a8bbb52bddf18901fecc27aed8a
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:18dfb38223647e11020e3ef1093a695454b90cc50922f80afb612212b7d13309
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:4c66a26b9f74d83a9bc60edb40ffd9d6ff2382d20500677d4567f8e141c3ddf0
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:758a9f6309a367d3f31cc03dd89356a8232a0fbd277570c01ebfa1fa30089a8d
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:2e5d5a2a70cda1f414f22b1c713d77d068c5fb16b1487a572019a9aa7130e6be
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:84b199e98e283c863d52a7b00044c20cb56ad563494adbec2229af265bf3f7c5
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:6baf67f3f410c071cc26f675d86edbbae5eaeb3410888cf622fad384b43b80ea
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:3a45cd2f3ea25f356167ce4fa20894a2a0706b4bcffa42c806f6017812053466
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:84679ffe8ce86c674818531897d122f5568d10c7e1e2ff919185d5771e5b4c39
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:ac33de246ee42bbfe735d7349d85c6df7e583e7599dea7dd7a2235bd0e59845f
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:e7145f016642b02bd651e21aa19edc7f93d0d1c2da3c6c7eed5d1174cf23619c
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:2564af054f52a8bb340e0270379c72da18658662a2a5bd26ac7bd979ddfccb56
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:4cff43cc145361c74c357055d51060ccc27e6bca58ccc3b5f5b983c7b8545114
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:09a30b59c906ddc4604d70f18210bcebef1f90f8bf965422b0caba363cd88989
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:63389195cf584a07ddbc2efec90cf6deaf737b91314b67db2b9f53dcc7acb57c
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:6193e569d4ab43806ac7b4870ef3634a5de3687527d6d190e59a9f735fa4778c
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:f4e7876fa8b27a8082efa99455d9bf7b0c479cff1e4c8a26136bf63092699c0b
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:d7e8e075a08b148ddfa91b44bb9cae1ef823756760279c4bb6d8489bc7e0eb2c
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:bf30b6c7173ebb6a4592246af16db87d7d391182f52286b559f8bad8f4b6cc5b
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:15179f0826831f051fdb867067be7c5528f01f442ae421b8782ae0cd1ed1b73d
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:4b3ac5ec2fafef0b7702a02847996f0903ec8d4262ccd03e8fa74885265ea4e9
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:d6dfd062df3d0d49f7efb928b835b9c8e0bcfb514a653ba2896a0152f24a0a21
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:185a83c994aced74546fe630cd71656f1e4112cb15be75f41dfcb1b9ed0966bd
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:b59e34cdb6214eacb9dc176a7b33125d7c1871f6b81696d59646294bfaf749e7
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:8e19d99232be4c37e857a4931627fb0aed721893e27b0d0e522fcf13bf67c1c9
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:06cc71822684c3ed094f327553451daf076d6bb97065c314da1a7d2adbbef3d2
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:4bdafc4404d5cfd844d05f32d88c262fba56dc931bfd5522b78c25d4ff253883
        - registry.redhat.io/container-native-virtualization/kubevirt-ipam-controller-rhel9@sha256:f33ac4d7477a730f273a462f171aed35dea38374d5ea7ef0bae2d7f74413a757
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:af5d2a2d4f0b6f0e1a13f1029073bb56980d8a70503b21628f18e8986e9b531c
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:5b2c64218342a5744182fc06f550bcdcf954e931275a9d1505ef0bbfbd0f8044
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:88f02cbebd6c3a2add6e0f8ca744e42bbd182f5a07d4fac59cf5cf07a4edfd45
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:fa0478fbf91248166f0eed8ec5f8dbb5b31d26031438376d9301836b6a069643
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:9bfdaf1cca6eac15b63a4ea6b7de28f2328516c5cf8bd5727935dd729e45657d
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:30c669770b0b992a3c84bb81c892e70a24fb14af187e8a4ee19cb3e0dd56ddbb
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:c5e8f44f647f593b00d3427172932d7aa694379d124209b9c66b4c1ffe3d4116
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:e6d9a6c41433360abd42d1118f5af3df9592d6d36af5f65082ae214cc8513c1b
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:d74c629e415d177069a0570fb8db9252d2b193d2ec7972c844686ffd14033f62
        - registry.redhat.io/container-native-virtualization/mtq-operator-rhel9@sha256:033bf7a6ef8ced1ac7275a9ddea8b16a10340a0fb0fb39619d89354d661f05b6
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:442e5cbcc63961642b662544032240bffb3863227324128a67f0bb6a05a1c364
        version: 4.99.0-0.1723448771
      entries:
      - name: kubevirt-hyperconverged-operator.v4.99.0-0.1723448771
        version: 4.99.0-0.1723448771
      name: dev-preview
    - currentCSV: kubevirt-hyperconverged-operator.v4.16.21
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:08f15657a1eee02927c5a6ef661e6e703d0064e091751bd2d8fbad5f118d9672
          createdAt: "2025-09-20 19:53:21"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.14.1 <4.16.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:88863a8dd2145bf17f7c5378c81e7c0be7a01dc47ee75bd7e5203a8fed8895b9
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.16.21-3,hyperconverged-cluster-operator-test-rhel9:v4.16.21-1,kubevirt-common-instancetypes-test-rhel9:v4.16.21-1,kubevirt-tekton-tasks-test-rhel9:v4.16.21-1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:9afe1ec9f776854caae38b89198a82e8e47a45d6c08f8dbd65a06c85f741c6a0
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:611d60b6ea4285fbc8a9f945fa5d06c09679a8f7ec1d70ec18edae7eec1a6bb3
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:299825fa33ab3f491cf954dd2e4ba23fdba2a96723b9ac3d98b3264b0b9fbb28
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:f6b075a56380143a7892e1fe5fdb212f8aa1c974f6a2d7e7e9b940789e44c010
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:d346688e02d03ddd9918f7a344d6eb3d5ddac3f1802dc82daf72bc56e7a9d496
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:406681aff7ea0a17941449593f839418edffdb073e2afffd53174f0df6292f94
        - registry.redhat.io/container-native-virtualization/mtq-controller-rhel9@sha256:253ef56b2867d33b3a475a456878b4a60f56473122eeed8c169568a9724ed30c
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:e54d87e6735151ed3c5295b23ca2f7ec666afb6421923636ef836bca1cf883d6
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:6385b346b3dd8f2bb3284871d326c851cf743a5e3fc3c49a1e51a1d43b0eafb4
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:8cd1052b9f69d0a0d8db41e90414756e83cbd07d7bd03c2faa85c41393d6539c
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:b80e68822fb0b5e78a995a97e02fe2ec448d67a5aa2224497fdf12b812182a0b
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:51a1fb0eb77e4f5c8bdfaf3b5074369b93ef63f5a7af62e22ed0789731f8dc73
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:361a48df213d52ced2a9963c2a8bde72a0981f1c497235fb918ea8c065580eeb
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:355e0946cfa963aeb5b6c3e8d476a9c3aee202baf77b95ce3f69c8e4b9e73836
        - registry.redhat.io/container-native-virtualization/mtq-lock-server-rhel9@sha256:c447251fea804340047bbaff885ad56be7a0d2612af1de26765f6cfc37da25e6
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:3daa460ed347e6eaab290b0a7fb0a08da6f210fb8be88af33f6e256511af80b6
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:dcc73f518ca7557cbd84b236367b3ecb7baab405268d54543783b4b772e14568
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:f7e5e04adb0467eb0aa4691a68b0d6485db9a52f99494097bf5b598e08f2edfb
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:2a109f94282b76ca0ff929adc35c93c0a98e34b43aba248d2975957fa149133f
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:cca121312777d2727fe2c50a14583243ce69e6e356b499f133736787522095b2
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:a0af0e5d4b142f28dc72c5636df5040c4ce92c6b12ab183d3a9bc7a9d2d2f418
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:88863a8dd2145bf17f7c5378c81e7c0be7a01dc47ee75bd7e5203a8fed8895b9
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:8cf7993326ba220f650ae2fa5383d0d6cf3e2c7f3346f0a339a9d212b8df02f6
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:edf12760e7fe4c8d37c22e172c0da270c219b7ff16d4ebb6b62729545309f6fa
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:a040696505e266e8f06a5f06b108f2db5ad3a6ed803d547d736455c1928a6a0e
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:d9560ed2758d3793139d3e13b7e5ce9d0f4ac0f8a6a8767449822c033cb2a456
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:9d997314bc7c6b30c2c3ff54b27bcbc1cdeac6dba300cab8a898b97fbce05ff7
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:dbff3d1592b192253a288fb3230a0ff125564540a30a6dd5ac881e79ed93b60a
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:9ae64f17191b98bb75f613969a9415103dad304e4fd20fb444f4fc4e57a57e0e
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:14952f38ab1d7f0abe4d54883e2c93d01ec1240ae809aefa9ca8cb4aaef080e1
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:15a971cef680f6f43e6b8ad7e0f5dab572ebacebc35cfbc6f7949a4ccd7789c2
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:d41ece8df9482ff860706edcdf27b53ac97d7c87407b3acfe8b1d9adfbf40e38
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:7581cd2330c20be2b77621e0e14da183d2b94f40fa5d3f448dede5f8c182759e
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:6fbdfdf3b9206058d5efc6653194bf6c42c07655817e3c97044da678e84b310a
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:3aa57a7ee9b342844ea4aa20bbbdf7a43b5d41f7866baa619f25473a230011f3
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:91b86228495f38d5528be1a8ffec96695ad72fafb76d2ead87f7bee25540d8b6
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:2e240514425d70039f5b48fe1908c56fae6efcc3eff93c63ec67f4aa3318a5fb
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:69388282df18b68c1d937ac98a5026b34221a54e5ce4b9cf4b89379b66623488
        - registry.redhat.io/container-native-virtualization/mtq-operator-rhel9@sha256:e29ffbe332d5da66ab4612a50ed73813ccbde92622474273b617cccd3bc24cd7
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:d253f1dbb54c70cf5874c2177bcd1cc3697381916cd4cf825bda934bfdb3d53c
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:7752f41a0f380fcd0d2652d286d4aac1ebb94a8a5fbb655b2e7058054f2eca4d
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:b20230f9a74652cf676bcf05dad81fc1ff64f9cff7232782c62cf585db48b115
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:0c5e1ee8ae0f19237a2c8b93466a7fc217a112cf1fb6dd55e2be27f95c9b66a3
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:bf7562a3bbc7655565b69fc96f2deae15729bbdef158cd90b97fc5ecc58d5bd8
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:fd2b4a337a8e45061cb6f455de3eefe27b004844c724fd821efc3b47215b74d3
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:77211068f3a8d331f1b21c64918861398f2cdbb2a3b760a5e8df51c89efea5b9
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:cddcdbf1d02b6e55cd93787991b5cf8704d0fb3a4a0bef5f58c918b57e26ee16
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:387fba10dc8f16fbfbc8c84274c5f1f5f6d01cf11f17f1b54fa0775a3300c6de
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:53ae4db672daad95d99f2955545ba335095c96630efc2a2ff19c5019a9413e87
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:63262574fa7d51d3b8c26b79e22d78f462c70239746aec6b8b3b951a02919e81
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:4046e8e862e0c55f09e9031b95e6a77af4d491593e39770c644287d44888b0bb
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:42fc0d1d112b30f65d6a48fa0442e2f41872658d2431ac8618dadf0aa7a41208
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:9e87e7d8679211a317d11e588868b40235da8c668bfb2ee10b564f6850fb91d1
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:38e767770ea0a74c4d594a382eed72bf9efa527e45ac6438acb79a1039f57fbd
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:75d76d7274f5169a414d8024125d4f0d8709888ac2b5d6121360635c7967d878
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:96092abe951f495f4f7dd0bdcb371b7efcb03a78a15e3ed8f1b37502203a6316
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:08f15657a1eee02927c5a6ef661e6e703d0064e091751bd2d8fbad5f118d9672
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:7e19b42a648cee71c60cb69fbcf6ebca05c938942e16d23f1cd61a619cffdcb6
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:878e1c8bd72cb3002457b6fbcddca99c0f4b00f6231fea52387d540b12e6329f
        version: 4.16.21
      entries:
      - name: kubevirt-hyperconverged-operator.v4.16.21
        version: 4.16.21
      - name: kubevirt-hyperconverged-operator.v4.16.17
        version: 4.16.17
      - name: kubevirt-hyperconverged-operator.v4.16.13
        version: 4.16.13
      - name: kubevirt-hyperconverged-operator.v4.16.7
        version: 4.16.7
      - name: kubevirt-hyperconverged-operator.v4.16.6
        version: 4.16.6
      - name: kubevirt-hyperconverged-operator.v4.16.5
        version: 4.16.5
      - name: kubevirt-hyperconverged-operator.v4.16.4
        version: 4.16.4
      - name: kubevirt-hyperconverged-operator.v4.16.3
        version: 4.16.3
      - name: kubevirt-hyperconverged-operator.v4.16.2
        version: 4.16.2
      - name: kubevirt-hyperconverged-operator.v4.16.1
        version: 4.16.1
      - name: kubevirt-hyperconverged-operator.v4.16.0
        version: 4.16.0
      - name: kubevirt-hyperconverged-operator.v4.15.2
        version: 4.15.2
      - name: kubevirt-hyperconverged-operator.v4.15.1
        version: 4.15.1
      - name: kubevirt-hyperconverged-operator.v4.15.0
        version: 4.15.0
      - name: kubevirt-hyperconverged-operator.v4.14.3
        version: 4.14.3
      - name: kubevirt-hyperconverged-operator.v4.14.2
        version: 4.14.2
      - name: kubevirt-hyperconverged-operator.v4.14.1
        version: 4.14.1
      - name: kubevirt-hyperconverged-operator.v4.14.0
        version: 4.14.0
      - name: kubevirt-hyperconverged-operator.v4.13.4
        version: 4.13.4
      - name: kubevirt-hyperconverged-operator.v4.13.3
        version: 4.13.3
      - name: kubevirt-hyperconverged-operator.v4.13.2
        version: 4.13.2
      - name: kubevirt-hyperconverged-operator.v4.13.1
        version: 4.13.1
      - name: kubevirt-hyperconverged-operator.v4.13.0
        version: 4.13.0
      - name: kubevirt-hyperconverged-operator.v4.12.2
        version: 4.12.2
      - name: kubevirt-hyperconverged-operator.v4.12.1
        version: 4.12.1
      - name: kubevirt-hyperconverged-operator.v4.12.0
        version: 4.12.0
      name: stable
    defaultChannel: stable
    packageName: kubevirt-hyperconverged
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: http://marketplace.redhat.com
    name: marketplace-games-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marketplace-games-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "game.marketplace.redhat.com/v1",
                "kind": "Arcade",
                "metadata": {
                  "name": "arcade-sample"
                },
                "spec": {
                  "size": 1
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Arcade is the Schema for the arcades API
            displayName: Arcade Instance
            kind: Arcade
            name: arcades.game.marketplace.redhat.com
            version: v1
        description: RHM Arcade Operator - a simple operator used to showcase how
          to build an operator around an existing web application that can can run
          on Openshift, Vanilla k8s, or on-boarded to Red Hat Marketplace.
        displayName: Arcade Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - rhm
        - red hat
        - marketplace
        - games
        - arcade
        links:
        - name: Marketplace Games Operator
          url: https://github.com/redhat-marketplace/marketplace-games-operator
        maturity: alpha
        provider:
          name: Red Hat Marketplace
          url: http://marketplace.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.connect.redhat.com/rhm-labs/arcade-operator@sha256:d6959be6693728d2dc4cd38b69251551d23c5002cb52db2684df31a0a8f7b0ab
        version: 1.0.0
      entries:
      - name: marketplace-games-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: marketplace-games-operator
    provider:
      name: Red Hat Marketplace
      url: http://marketplace.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Provider Name
      provider-url: https://your.domain
    name: smilecdr-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smilecdr-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.smilecdr.com/v1alpha1",
                "kind": "Smilecdr",
                "metadata": {
                  "name": "smilecdr"
                },
                "spec": {
                  "labels": {
                    "application": "smilecdr",
                    "client": "smilecdr",
                    "env": "dev",
                    "version": "one"
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Smilecdr
            name: smilecdrs.operator.smilecdr.com
            version: v1alpha1
        description: Smilecdr Operator description. TODO.
        displayName: Smilecdr Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - smilecdr-operator
        links:
        - name: Smilecdr Operator
          url: https://smilecdr-operator.domain
        maintainers:
        - email: your@email.com
          name: Maintainer Name
        maturity: alpha
        provider:
          name: Provider Name
          url: https://your.domain
        relatedImages:
        - quay.io/jibran_shaukat/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - quay.io/jibran_shaukat/smilecdr@sha256:cd6b2cec6f5b9ad9e0f0434532f8fb00b6919fcd840e686d96ce08874f2154db
        version: 0.0.1
      entries:
      - name: smilecdr-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: smilecdr-operator
    provider:
      name: Provider Name
      url: https://your.domain
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: model-builder-for-vision-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-modelbuilder-for-vision-rhmp.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "Modelbuilder",
                "metadata": {
                  "name": "modelbuilder-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "backup_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "metadata_storage": {
                    "storage_class": "ibmc-block-bronze",
                    "storage_size": "20G"
                  },
                  "photo_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "500G"
                  },
                  "in_memory_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              },
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "ModelbuilderAWS",
                "metadata": {
                  "name": "modelbuilder-aws-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "postgres": {
                    "storage_size": "20G",
                    "storage_class": "gp2",
                    "backup_storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
          createdAt: "2021-08-28T12:00:00Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/model-builder-for-vision-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/model-builder-for-vision-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.0.3
          operators.operatorframework.io/builder: operator-sdk-v1.11.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/ibm-modelbuilder-for-vision/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Model Builder for Vision on AWS
            displayName: Deploy-Modelbuilder on AWS
            kind: ModelbuilderAWS
            name: modelbuildersaws.modelbuilder.com
            version: v1alpha1
          - description: Deploy Model Builder for Vision on IBM
            displayName: Deploy-Modelbuilder on IBM
            kind: Modelbuilder
            name: modelbuilders.modelbuilder.com
            version: v1alpha1
        description: IBM Services Software Model Builder for Vision is an AI training
          platform that can use the capabilities of IBM Cloud and Amazon Web Services
          and their GPUs to quickly train computer vision models for compatible mobile
          apps. IBM Services Software Inspection Workbench iOS and iPadOS app will
          be used as the exclusive UI for labeling and training. Specifically, IBM
          Model Builder for Vision will train computer vision models that can then
          be deployed for use on the IBM Services Software Inspector Portable , IBM
          Services Software Inspector Wearable and the IBM Maximo Visual Inspection
          Mobile apps.
        displayName: IBM Model Builder for Vision
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Model Builder
        - AI
        - Computer Vision
        - Inspection
        - Mobile
        - Core ML
        - TensorFlow Lite
        links:
        - name: IBM Model Builder for Vision
          url: http://ibm.biz/model_builder
        maintainers:
        - email: isshelp@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.connect.redhat.com/ibm/modelbuilder-augmentor@sha256:ef5d0dde3eb334d919c2773af63007f1ef5e735ae040ef77fffe3440f854f068
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/rhel8/redis-6@sha256:2f65f46b25d84a3cb41a6a10f93786bcab47b7354ca9ac13a8e602ea833eba60
        - registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
        - registry.connect.redhat.com/ibm/modelbuilder-image-server@sha256:ecdb665af651417c50335e7259b89ef8096521abec431976bd7abe4c1edc49d6
        - registry.connect.redhat.com/ibm/modelbuilder-auth-service@sha256:f96b444fdcf7781775c9eba942bcc5e5ebdcf5f50820447f5eb26061c4c674cd
        - registry.connect.redhat.com/ibm/modelbuilder-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/modelbuilder-cv-training-service@sha256:c3c9080b08b87022c784fd98ac3e2774ec89ce29a2dd19d91dee039e3fdb2b36
        version: 1.0.3
      entries:
      - name: ibm-modelbuilder-for-vision-rhmp.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: model-builder-for-vision-certified-rhmp
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: fuse-console
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-console.v7.10.3-0.1655306919.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "replicas": 1,
                  "version": "1.10",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "100Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:37dc38375f5d553ac1d9f20f66ce005ed9495354b81d7915ae3bd7dd6af132f3
          createdAt: "2020-09-14T10:59:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.10.3'
          olm.substitutesFor: fuse-console.v7.10.3-0.1652362834.p
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.10/html/managing_fuse
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:37dc38375f5d553ac1d9f20f66ce005ed9495354b81d7915ae3bd7dd6af132f3
        version: 7.10.3+0.1655306919.p
      entries:
      - name: fuse-console.v7.10.3
        version: 7.10.3
      - name: fuse-console.v7.10.3-0.1652362834.p
        version: 7.10.3+0.1652362834.p
      - name: fuse-console.v7.10.3-0.1655306919.p
        version: 7.10.3+0.1655306919.p
      name: 7.10.x
    - currentCSV: fuse-console.v7.11.1-0.1681224695.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.11",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:518b5ace9178e2aafd7b6f247fd41d9392abd1a7bccb4cb53783531820611193
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.11.1'
          olm.substitutesFor: fuse-console.v7.11.1-0.1675972857.p
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.11/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:518b5ace9178e2aafd7b6f247fd41d9392abd1a7bccb4cb53783531820611193
        version: 7.11.1+0.1681224695.p
      entries:
      - name: fuse-console.v7.11.1
        version: 7.11.1
      - name: fuse-console.v7.11.1-0.1669922044.p
        version: 7.11.1+0.1669922044.p
      - name: fuse-console.v7.11.1-0.1675972857.p
        version: 7.11.1+0.1675972857.p
      - name: fuse-console.v7.11.1-0.1681224695.p
        version: 7.11.1+0.1681224695.p
      name: 7.11.x
    - currentCSV: fuse-console.v7.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.12",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:4e6889bf76d323956e0b594db7381e8162fea6b108f53f9e6dd0b38cf3596c8d
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.12.1'
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.12/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:4e6889bf76d323956e0b594db7381e8162fea6b108f53f9e6dd0b38cf3596c8d
        version: 7.12.1
      entries:
      - name: fuse-console.v7.12.1
        version: 7.12.1
      name: 7.12.x
    - currentCSV: fuse-console.v7.13.0-0.1757604404.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.13",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:72d98254a821bce237b3dde896f2825628cb7b6df564e23949337590586e8bea
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.7.0 <7.13.0'
          olm.substitutesFor: fuse-console.v7.13.0-0.1754559584.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.13/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:72d98254a821bce237b3dde896f2825628cb7b6df564e23949337590586e8bea
        version: 7.13.0+0.1757604404.p
      entries:
      - name: fuse-console.v7.13.0-0.1744281421.p
        version: 7.13.0+0.1744281421.p
      - name: fuse-console.v7.13.0-0.1723709432.p
        version: 7.13.0+0.1723709432.p
      - name: fuse-console.v7.13.0-0.1757604404.p
        version: 7.13.0+0.1757604404.p
      - name: fuse-console.v7.13.0-0.1721224332.p
        version: 7.13.0+0.1721224332.p
      - name: fuse-console.v7.13.0-0.1744737026.p
        version: 7.13.0+0.1744737026.p
      - name: fuse-console.v7.13.0-0.1724659726.p
        version: 7.13.0+0.1724659726.p
      - name: fuse-console.v7.13.0-0.1725026290.p
        version: 7.13.0+0.1725026290.p
      - name: fuse-console.v7.13.0-0.1748960166.p
        version: 7.13.0+0.1748960166.p
      - name: fuse-console.v7.13.0-0.1738056254.p
        version: 7.13.0+0.1738056254.p
      - name: fuse-console.v7.13.0-0.1739885034.p
        version: 7.13.0+0.1739885034.p
      - name: fuse-console.v7.13.0-0.1741875974.p
        version: 7.13.0+0.1741875974.p
      - name: fuse-console.v7.13.0-0.1741884704.p
        version: 7.13.0+0.1741884704.p
      - name: fuse-console.v7.13.0-0.1719822976.p
        version: 7.13.0+0.1719822976.p
      - name: fuse-console.v7.13.0
        version: 7.13.0
      - name: fuse-console.v7.13.0-0.1731460809.p
        version: 7.13.0+0.1731460809.p
      - name: fuse-console.v7.13.0-0.1749640277.p
        version: 7.13.0+0.1749640277.p
      - name: fuse-console.v7.13.0-0.1751291353.p
        version: 7.13.0+0.1751291353.p
      - name: fuse-console.v7.13.0-0.1753702606.p
        version: 7.13.0+0.1753702606.p
      - name: fuse-console.v7.13.0-0.1753778626.p
        version: 7.13.0+0.1753778626.p
      - name: fuse-console.v7.13.0-0.1753785080.p
        version: 7.13.0+0.1753785080.p
      - name: fuse-console.v7.13.0-0.1753860418.p
        version: 7.13.0+0.1753860418.p
      - name: fuse-console.v7.13.0-0.1754303647.p
        version: 7.13.0+0.1754303647.p
      - name: fuse-console.v7.13.0-0.1754559584.p
        version: 7.13.0+0.1754559584.p
      - name: fuse-console.v7.13.0-0.1754567932.p
        version: 7.13.0+0.1754567932.p
      - name: fuse-console.v7.12.1
        version: 7.12.1
      name: 7.13.x
    - currentCSV: fuse-console.v7.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "replicas": 1,
                  "version": "1.9",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "100Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:846d57f3fd9f307a98b0169b6733c26299209d99ed189f219c48f36ba771b198
          createdAt: "2020-09-14T10:59:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.9.0'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.9/html/managing_fuse
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:846d57f3fd9f307a98b0169b6733c26299209d99ed189f219c48f36ba771b198
        version: 7.9.0
      entries:
      - name: fuse-console.v7.9.0
        version: 7.9.0
      name: 7.9.x
    defaultChannel: 7.13.x
    packageName: fuse-console
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: opentelemetry-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opentelemetry-operator.v0.135.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "Instrumentation",
                "metadata": {
                  "name": "instrumentation"
                },
                "spec": {
                  "exporter": {
                    "endpoint": "http://otel-collector-headless:4317"
                  }
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpAMPBridge",
                "metadata": {
                  "name": "opampbridge-sample"
                },
                "spec": {
                  "capabilities": {
                    "AcceptsOpAMPConnectionSettings": true,
                    "AcceptsOtherConnectionSettings": true,
                    "AcceptsRemoteConfig": true,
                    "AcceptsRestartCommand": true,
                    "ReportsEffectiveConfig": true,
                    "ReportsHealth": true,
                    "ReportsOwnLogs": true,
                    "ReportsOwnMetrics": true,
                    "ReportsOwnTraces": true,
                    "ReportsRemoteConfig": true,
                    "ReportsStatus": true
                  },
                  "componentsAllowed": {
                    "exporters": [
                      "debug"
                    ],
                    "processors": [
                      "memory_limiter"
                    ],
                    "receivers": [
                      "otlp"
                    ]
                  },
                  "endpoint": "ws://opamp-server:4320/v1/opamp"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": "receivers:\n  otlp:\n    protocols: \n      grpc: {}\n      http: {}\n\nexporters:\n  debug: {}\n\nservice:\n  pipelines:\n    traces:\n      receivers: [otlp]\n      exporters: [debug]\n"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1beta1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": {
                    "exporters": {
                      "debug": {}
                    },
                    "receivers": {
                      "otlp": {
                        "protocols": {
                          "grpc": {},
                          "http": {}
                        }
                      }
                    },
                    "service": {
                      "pipelines": {
                        "traces": {
                          "exporters": [
                            "debug"
                          ],
                          "receivers": [
                            "otlp"
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/rhosdt/opentelemetry-rhel8-operator@sha256:45c6c52744e9a17ffd883a2864cb7722ce3f9580e522c976d4f55623118f3504
          createdAt: 24 Sep 2025, 10:19
          description: Provides the OpenTelemetry components, including the Collector
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.33.0 <0.135.0-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-opentelemetry-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-telemetry/opentelemetry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instrumentation is the spec for OpenTelemetry instrumentation.
            displayName: OpenTelemetry Instrumentation
            kind: Instrumentation
            name: instrumentations.opentelemetry.io
            version: v1alpha1
          - description: OpAMPBridge is the Schema for the opampbridges API.
            displayName: OpAMP Bridge
            kind: OpAMPBridge
            name: opampbridges.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1beta1
          - description: TargetAllocator is the Schema for the targetallocators API.
            displayName: Target Allocator
            kind: TargetAllocator
            name: targetallocators.opentelemetry.io
            version: v1alpha1
        description: |-
          Red Hat build of OpenTelemetry is a collection of tools, APIs, and SDKs. You use it to instrument, generate, collect, and export telemetry data (metrics, logs, and traces) for analysis in order to understand your software's performance and behavior.
          This operator was previously called Red Hat OpenShift distributed tracing data collection.

          ### Operator features

          * **Sidecar injection** - annotate your pods and let the operator inject a sidecar.
          * **Managed upgrades** - updating the operator will automatically update your OpenTelemetry collectors.
          * **Deployment modes** - your collector can be deployed as sidecar, daemon set, or regular deployment.
          * **Service port management** - the operator detects which ports need to be exposed based on the provided configuration.

          ## Examples

          There is a list of examples to help you deploy different configurations in [this GitHub repository](https://github.com/os-observability/redhat-rhosdt-samples)

          ### Support & Troubleshooting

          Red Hat build of OpenTelemetry is available as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/4.17/html/red_hat_build_of_opentelemetry/index) is recommended when installing, configuring, and managing the Operator and instances.
        displayName: Red Hat build of OpenTelemetry
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opentelemetry
        - tracing
        - logging
        - metrics
        - monitoring
        - troubleshooting
        links:
        - name: OpenTelemetry Operator
          url: https://github.com/open-telemetry/opentelemetry-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/opentelemetry-rhel8-operator@sha256:45c6c52744e9a17ffd883a2864cb7722ce3f9580e522c976d4f55623118f3504
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5a6edfecf1636805e6f0bcad02df10bceec764f0580f3ac8fe62aba978d50006
        - registry.redhat.io/rhosdt/opentelemetry-collector-rhel8@sha256:5f31caf1f4a12cf2832ac3e84877f4fb0c856073820498c8a5be13698348c38a
        - registry.redhat.io/rhosdt/opentelemetry-target-allocator-rhel8@sha256:f77a26edafba06a4fca1a07083e06a908664852cf228b1814455ffa8155cc979
        version: 0.135.0-1
      entries:
      - name: opentelemetry-operator.v0.135.0-1
        version: 0.135.0-1
      - name: opentelemetry-operator.v0.127.0-2
        version: 0.127.0-2
      - name: opentelemetry-operator.v0.127.0-1
        version: 0.127.0-1
      - name: opentelemetry-operator.v0.119.0-2
        version: 0.119.0-2
      - name: opentelemetry-operator.v0.119.0-1
        version: 0.119.0-1
      - name: opentelemetry-operator.v0.113.0-2
        version: 0.113.0-2
      - name: opentelemetry-operator.v0.113.0-1
        version: 0.113.0-1
      - name: opentelemetry-operator.v0.107.0-5
        version: 0.107.0-5
      - name: opentelemetry-operator.v0.107.0-4-0.1727168137.p
        version: 0.107.0-4+0.1727168137.p
      - name: opentelemetry-operator.v0.107.0-4
        version: 0.107.0-4
      - name: opentelemetry-operator.v0.102.0-3
        version: 0.102.0-3
      - name: opentelemetry-operator.v0.102.0-2
        version: 0.102.0-2
      - name: opentelemetry-operator.v0.100.1-3
        version: 0.100.1-3
      - name: opentelemetry-operator.v0.93.0-3
        version: 0.93.0-3
      - name: opentelemetry-operator.v0.93.0-2
        version: 0.93.0-2
      - name: opentelemetry-operator.v0.89.0-3
        version: 0.89.0-3
      - name: opentelemetry-operator.v0.81.1-5
        version: 0.81.1-5
      - name: opentelemetry-operator.v0.74.0-5-0.1687199949.p
        version: 0.74.0-5+0.1687199949.p
      - name: opentelemetry-operator.v0.74.0-5
        version: 0.74.0-5
      - name: opentelemetry-operator.v0.60.0-2
        version: 0.60.0-2
      - name: opentelemetry-operator.v0.56.0-1
        version: 0.56.0-1
      name: stable
    defaultChannel: stable
    packageName: opentelemetry-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://www.medik8s.io/
    name: customized-user-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: customized-user-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediation",
                "metadata": {
                  "name": "customizeduserremediation-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediationConfig",
                "metadata": {
                  "name": "customizeduserremediationconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediationTemplate",
                "metadata": {
                  "name": "customizeduserremediationtemplate-sample",
                  "namespace": "openshift-workload-availability"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/customized-user-remediation-operator:v0.1.0
          createdAt: "2024-03-25 16:17:57"
          description: Customized User Remediation Operator will remediate unhealthy
            nodes by running a pre defined user script on the unhealthy node. Since
            the user has full control over the script content, is may be quite risky
            so apply discretion.
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/customized-user-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CustomizedUserRemediationConfig is the Schema for the customizeduserremediationconfigs
              API
            displayName: Customized User Remediation Config
            kind: CustomizedUserRemediationConfig
            name: customizeduserremediationconfigs.customized-user-remediation.medik8s.io
            version: v1alpha1
          - description: CustomizedUserRemediation is the Schema for the customizeduserremediations
              API
            displayName: Customized User Remediation
            kind: CustomizedUserRemediation
            name: customizeduserremediations.customized-user-remediation.medik8s.io
            version: v1alpha1
          - description: CustomizedUserRemediationTemplate is the Schema for the customizeduserremediationtemplates
              API
            displayName: Customized User Remediation Template
            kind: CustomizedUserRemediationTemplate
            name: customizeduserremediationtemplates.customized-user-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Customized User Remediation (CUR) works with Node Health Check to remediate unhealthy nodes by running a user customized script on the unhealthy node.
          This operator does not preform any fencing on the remediated node so users should be careful with the scripts that are being used.

          ### Compatibility
          CUR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Flexibility - CUR is quite flexible sine it runs in the remediation process a user customized script.
        displayName: Customized User Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - script
        - CUR
        links:
        - name: Customized User Remediation
          url: https://customized-user-remediation.domain
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Medik8s
          url: https://www.medik8s.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/medik8s/customized-user-remediation-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: customized-user-remediation.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: customized-user-remediation
    provider:
      name: Medik8s
      url: https://www.medik8s.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-loadcore-agents
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-loadcore-agents.v0.2.25
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2024-11-20T18:10:53Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <=10.0.0'
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Agents
            name: agents.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight LoadCore Agents Operator
        displayName: Keysight LoadCore Agents Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Loadcore Agents
          url: https://keysight-loadcore-agents.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:7efeeb8b29872a6f0271f651d7ae02c91daea16d853c50e374c310f044d8c76c
        - prod.harbor.keysight.digital/loadcore/keysight-loadcore-agents@sha256:96e18b6270097c23bad5445a15613e467d06c6fbdbaf39545edb79915aacc04f
        version: 0.2.25
      entries:
      - name: keysight-loadcore-agents.v0.2.25
        version: 0.2.25
      name: alpha
    defaultChannel: alpha
    packageName: keysight-loadcore-agents
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-secret-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secret-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/tools@sha256:47204b2697dfbab5eaf9d6dcb4c160dd427a148d6b840727239b773810e74fc9
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:a6d6631d35c32b874c995b6af889636ac8dd1f39dcb68810d9f9b10a4e2c1987
        - quay.io/stackable/tools@sha256:47204b2697dfbab5eaf9d6dcb4c160dd427a148d6b840727239b773810e74fc9
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:161e754437428c68490333337303fb30b695a9ac677c3a00a1dc27565b855ab8
        - quay.io/stackable/secret-operator@sha256:1267311830649f928b528512a3cdc843efad6dd32b851d99c84639fbe70e35a3
        version: 24.11.1
      entries:
      - name: secret-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: secret-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:6eb08b99259d62514d87844e1b7470382d500696fecc9a1232f6bb536ed897fa
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:dd730457133f619d8759269abcfa79d7aeb817e01ba7af2d3aa1417df406ea56
        - quay.io/stackable/secret-operator@sha256:6eb08b99259d62514d87844e1b7470382d500696fecc9a1232f6bb536ed897fa
        version: 25.3.0
      entries:
      - name: secret-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: secret-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
          - description: Trust stores managed by the operator.
            displayName: Trust stores managed by the operator.
            kind: TrustStore
            name: truststores.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
        version: 25.7.0
      entries:
      - name: secret-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: secret-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
          - description: Trust stores managed by the operator.
            displayName: Trust stores managed by the operator.
            kind: TrustStore
            name: truststores.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        version: 25.7.0
      entries:
      - name: secret-operator.v25.7.0
        version: 25.7.0
      - name: secret-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-secret-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zettaset
      provider-url: ""
    name: xcrypt-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zts-xcrypt-operator.v0.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xcrypt.zettaset.com/v1alpha1",
                "kind": "XCrypt",
                "metadata": {
                  "name": "zts-masterset"
                },
                "spec": {
                  "status": {
                    "podNames": "zts-ca",
                    "replicas": "1,"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
          createdAt: "2021-06-16T12:29:13.97Z"
          description: XCrypt Operator deploys Zettaset XCrypt Container Encryption  -
            the only software-defined solution that offers transparent high performance
            data-at-rest encryption of critical data in container environments.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/xcrypt-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/xcrypt-operator-rhmp/support?utm_source=openshift_console
          support: Zettaset
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XCrypt is the Schema for the xcrypts API
            kind: XCrypt
            name: xcrypts.xcrypt.zettaset.com
            version: v1alpha1
        description: "Zettaset’s RedHat Certified XCrypt Operator allows users to
          deploy multiple XCrypt components on OpenShift with just a few simple commands.
          The XCrypt Operator deployment provides software-defined encryption that
          transparently protects container data throughout your OpenShift environment.\n\n##
          Core Features\n* **Automated XCrypt provisioning** - Deploying a Zettaset
          XCrypt has never been easier. Adjust cluster specific configuration and
          let the Zettaset Operator take care of provisioning XCrypt services and
          setting up the XCrypt encryption to your exact specification. \n* **CSI
          Driver deployment**: Zettaset XCrypt Operator automatically deploys Zettaset
          XCrypt CSI Driver in order to provide automated and transparent data-at-rest
          \ encryption functionality compatible with Openshift and any other container
          orchestrator. Once the Operator is provisioned, the driver runs fine tuned
          and ready to work with other XCrypt components.\n* **Multi-node topology**:
          Encrypt as many nodes and volumes as you need to. XCrypt components can
          be installed on multiple nodes across the cluster. This is defined by two
          labels: zts-master and zts-worker. Normally zts-master node is a single
          node that runs Zettaset XCrypt major services required for the product operation.
          Multiple nodes can be marked as zts-worker and those are the nodes that
          can have encrypted volumes.\n* **Seamless Encrypted Volume mount**: Volumes
          are not only being encrypted, but also are mounted/unmounted to containers
          with Persistent Volume Claims. This happens as seamlessly as regular volumes
          provision and mount/unmount.\n* **Robust failover**: A node failure doesn’t
          affect XCrypt operations on other nodes of the cluster. Encryption on a
          restarted/recovered node gets back to normal functioning if the node stays
          healthy and data hasn’t been corrupted during the restart\n\n## About Zettaset
          XCrypt Kubernetes Encryption for OpenShift\nXCrypt Kubernetes Encryption
          for OpenShift is a software-only solution that enables encryption of data
          at rest stored in Red Hat OpenShift container environments without slowing
          down the DevSecOps process. In addition to providing a transparent and high-performance
          layer of security that flexibly protects container data across on-premise,
          in hybrid, and multi-cloud environments, it also helps your organization
          efficiently transition from DevOps to DevSecOps.  \n\nMore information about
          Zettaset XCrypt Kubernetes Encryption for OpenShift can be found in the
          [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)\n\n##
          RedHat OpenShift + Zettaset: The next generation of secure enterprise infrastructure
          for DevSecOps  \nBy combining RedHat’s Container Platform with Zettaset’s
          certified software-defined encryption, you can create the next generation
          of secure enterprise infrastructure that DevSecOps requires. Together they
          create the stable, consistent, and supported base that your organization
          needs to develop applications faster, simplify updates, drive innovation,
          and reduce the risk of potential security breaches and data theft.\n###
          Core Features\n* Software-only for simple and scalable deployments \n* Negligible
          performance impact\n* Transparent to developers and administrators \n* Works
          with AWS, Google, and Azure \n* Unique encryption key per container volume
          \n* Encrypted volumes are only available when in use\n* Direct integration
          with OpenShift \n* Automated encryption policy management Secure erase of
          volumes\n* Ability to securely decommission a node\n\n## How does XCrypt
          Kubernetes Encryption fit into OpenShift? \nXCrypt XCrypt Kubernetes Encryption
          for OpenShift Encryption makes it incredibly easy to protect container data
          in your OpenShift environment. \n### Use cases include:\n* Transitioning
          from DevOps to DevSecOps \n* Ensuring data protection in OpenShift container
          environments with negligible performance impact \n* Achieving compliance
          in regulated industries like healthcare and finance\n\n## Other Information\nRole-based
          access control needs to be configured prior to the Operator installation.
          Please find details in the [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)
          \   "
        displayName: Zettaset XCrypt Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - encryption
        - storage
        links:
        - name: Zettaset
          url: https://www.zettaset.com/
        maintainers:
        - email: info@zettaset.com
          name: Zettaset
        maturity: alpha
        provider:
          name: Zettaset
        relatedImages:
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-ca-1@sha256:9f1015dfb459b69f5dbf449efa0517452095a6cf4f20826ebae278f9f800ffe7
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-kmip-server-1@sha256:9af5d21dd748dfbf4e107f27a5f10fa3c0ba928ade9b989a744b620ce34ec757
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-license-server-1@sha256:bc5486892a41e70f5dc4c4e1d82fca8db2149598fb8b5950ba5da1ee2d9d245d
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-host-manager-1@sha256:9faa4b597eb74df412761dffa057b64fb69dbf9fa016414962eb2099a343e734
        - registry.marketplace.redhat.com/rhm/zts/zts-csi-driver-1@sha256:c15fa9a7cc40891103f426206663dcebe1210c8ee3ab6037cbf64e1f86b9c7e5
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
        version: 0.0.17
      entries:
      - name: zts-xcrypt-operator.v0.0.17
        version: 0.0.17
      name: alpha
    defaultChannel: alpha
    packageName: xcrypt-operator-rhmp
    provider:
      name: Zettaset
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: costmanagement-metrics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: costmanagement-metrics-operator.4.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "true"
          containerImage: registry.redhat.io/costmanagement/costmanagement-metrics-rhel9-operator@sha256:c5bb5291052d57390b9f4bdc660c1926827072b6dcbe09d85f5b6fc375562182
          createdAt: "2025-08-18T12:10:15Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: costmanagement-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Cost Management Metrics Operator
          ## Introduction
          The `costmanagement-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Cost Management Metrics Operator (`costmanagement-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created CostManagementMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The CostManagementMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **FIPS Compliance:** Supports deployment in high-security environments when run on an OpenShift cluster with FIPS mode enabled.

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created CostManagementMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the CostManagementMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `CostManagementMetricsConfig`, creating a new `CostManagementMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `costmanagement-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: costmanagement-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the CostManagementMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the CostManagementMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon CostManagementMetricsConfig creation** (This parameter will start to appear in CostManagementMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the CostManagementMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the costmanagement-metrics-operator
          **Note** There are separate instructions for configuring the `costmanagement-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `costmanagement-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the CostManagementMetricsConfig
          Configure the costmanagement-metrics-operator by creating a `CostManagementMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `costmanagement-metrics-operator` -> `CostManagementMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the costmanagement-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `costmanagement-metrics-operator-data` with 10Gi of storage. To configure the costmanagement-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `costmanagement-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `costmanagement-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the costmanagement-metrics-operator for a restricted network
          ##### Create the CostManagementMetricsConfig
          Configure the costmanagement-metrics-operator by creating a `CostManagementMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `costmanagement-metrics-operator` -> `CostManagementMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `costmanagement-metrics-operator-data` with 10Gi of storage. To configure the costmanagement-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `costmanagement-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: costmanagement-metrics-operator
                labels:
                  app: costmanagement-metrics-operator
              spec:
                volumes:
                - name: costmanagement-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: costmanagement-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: costmanagement-metrics-operator-reports
                    mountPath: /tmp/costmanagement-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/costmanagement-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/costmanagement-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `costmanagement-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the CostManagementMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Cost Management Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Cost Management Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/costmanagement/costmanagement-metrics-rhel9-operator@sha256:c5bb5291052d57390b9f4bdc660c1926827072b6dcbe09d85f5b6fc375562182
        version: 4.1.0
      entries:
      - name: costmanagement-metrics-operator.4.1.0
        version: 4.1.0
      - name: costmanagement-metrics-operator.4.0.0
        version: 4.0.0
      - name: costmanagement-metrics-operator.3.3.2
        version: 3.3.2
      - name: costmanagement-metrics-operator.3.3.1
        version: 3.3.1
      - name: costmanagement-metrics-operator.3.3.0
        version: 3.3.0
      - name: costmanagement-metrics-operator.3.2.1
        version: 3.2.1
      - name: costmanagement-metrics-operator.3.2.0
        version: 3.2.0
      - name: costmanagement-metrics-operator.3.1.1
        version: 3.1.1
      - name: costmanagement-metrics-operator.3.1.0
        version: 3.1.0
      - name: costmanagement-metrics-operator.3.0.1
        version: 3.0.1
      - name: costmanagement-metrics-operator.3.0.0
        version: 3.0.0
      - name: costmanagement-metrics-operator.2.0.0
        version: 2.0.0
      - name: costmanagement-metrics-operator.1.2.0
        version: 1.2.0
      - name: costmanagement-metrics-operator.1.1.9
        version: 1.1.9
      - name: costmanagement-metrics-operator.1.1.8
        version: 1.1.8
      - name: costmanagement-metrics-operator.1.1.7
        version: 1.1.7
      - name: costmanagement-metrics-operator.1.1.6
        version: 1.1.6
      - name: costmanagement-metrics-operator.1.1.5
        version: 1.1.5
      - name: costmanagement-metrics-operator.1.1.4
        version: 1.1.4
      - name: costmanagement-metrics-operator.1.1.3
        version: 1.1.3
      - name: costmanagement-metrics-operator.1.1.0
        version: 1.1.0
      - name: costmanagement-metrics-operator.1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: costmanagement-metrics-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-applicationautoscaling-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-applicationautoscaling-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "applicationautoscaling.services.k8s.aws/v1alpha1",
                "kind": "ScalableTarget",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "applicationautoscaling.services.k8s.aws/v1alpha1",
                "kind": "ScalingPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/applicationautoscaling-controller:1.1.1
          createdAt: "2025-09-20T21:43:01Z"
          description: AWS Application Auto Scaling controller is a service controller
            for managing Application Auto Scaling resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ScalableTarget represents the state of an AWS applicationautoscaling
              ScalableTarget resource.
            displayName: ScalableTarget
            kind: ScalableTarget
            name: scalabletargets.applicationautoscaling.services.k8s.aws
            version: v1alpha1
          - description: ScalingPolicy represents the state of an AWS applicationautoscaling
              ScalingPolicy resource.
            displayName: ScalingPolicy
            kind: ScalingPolicy
            name: scalingpolicies.applicationautoscaling.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Application Auto Scaling resources in AWS from within your Kubernetes cluster.
          **About Amazon Application Auto Scaling**
          Application Auto Scaling is a web service for developers and system administrators who need a solution for automatically scaling their scalable resources for individual AWS services beyond Amazon EC2. Application Auto Scaling allows you to configure automatic scaling for the following resources:
          - AppStream 2.0 fleets
          - Aurora replicas
          - Amazon Comprehend document classification and entity recognizer endpoints
          - DynamoDB tables and global secondary indexes
          - Amazon Elastic Container Service (ECS) services
          - Amazon EMR clusters
          - Amazon Keyspaces (for Apache Cassandra) tables
          - Lambda function provisioned concurrency
          - Amazon Managed Streaming for Apache Kafka (MSK) broker storage
          - SageMaker endpoint variants
          - Spot Fleet requests
          - Custom resources provided by your own applications or services. For more information, see the [GitHub repository](https://github.com/aws/aws-auto-scaling-custom-resource)

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Application Auto Scaling
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - applicationautoscaling
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Application Auto Scaling User Guide
          url: https://docs.aws.amazon.com/autoscaling/application/userguide/what-is-application-auto-scaling.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: application auto scaling maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/applicationautoscaling-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-applicationautoscaling-controller.v1.1.1
        version: 1.1.1
      - name: ack-applicationautoscaling-controller.v1.1.0
        version: 1.1.0
      - name: ack-applicationautoscaling-controller.v1.0.30
        version: 1.0.30
      - name: ack-applicationautoscaling-controller.v1.0.29
        version: 1.0.29
      - name: ack-applicationautoscaling-controller.v1.0.27
        version: 1.0.27
      - name: ack-applicationautoscaling-controller.v1.0.26
        version: 1.0.26
      - name: ack-applicationautoscaling-controller.v1.0.25
        version: 1.0.25
      - name: ack-applicationautoscaling-controller.v1.0.24
        version: 1.0.24
      - name: ack-applicationautoscaling-controller.v1.0.23
        version: 1.0.23
      - name: ack-applicationautoscaling-controller.v1.0.22
        version: 1.0.22
      - name: ack-applicationautoscaling-controller.v1.0.21
        version: 1.0.21
      - name: ack-applicationautoscaling-controller.v1.0.19
        version: 1.0.19
      - name: ack-applicationautoscaling-controller.v1.0.18
        version: 1.0.18
      - name: ack-applicationautoscaling-controller.v1.0.17
        version: 1.0.17
      - name: ack-applicationautoscaling-controller.v1.0.16
        version: 1.0.16
      - name: ack-applicationautoscaling-controller.v1.0.15
        version: 1.0.15
      - name: ack-applicationautoscaling-controller.v1.0.11
        version: 1.0.11
      - name: ack-applicationautoscaling-controller.v1.0.9
        version: 1.0.9
      - name: ack-applicationautoscaling-controller.v1.0.8
        version: 1.0.8
      - name: ack-applicationautoscaling-controller.v1.0.7
        version: 1.0.7
      - name: ack-applicationautoscaling-controller.v1.0.6
        version: 1.0.6
      - name: ack-applicationautoscaling-controller.v1.0.5
        version: 1.0.5
      - name: ack-applicationautoscaling-controller.v1.0.4
        version: 1.0.4
      - name: ack-applicationautoscaling-controller.v1.0.3
        version: 1.0.3
      - name: ack-applicationautoscaling-controller.v1.0.2
        version: 1.0.2
      - name: ack-applicationautoscaling-controller.v1.0.0
        version: 1.0.0
      - name: ack-applicationautoscaling-controller.v0.2.16
        version: 0.2.16
      - name: ack-applicationautoscaling-controller.v0.2.15
        version: 0.2.15
      - name: ack-applicationautoscaling-controller.v0.2.14
        version: 0.2.14
      - name: ack-applicationautoscaling-controller.v0.2.13
        version: 0.2.13
      - name: ack-applicationautoscaling-controller.v0.2.12
        version: 0.2.12
      - name: ack-applicationautoscaling-controller.v0.2.10
        version: 0.2.10
      - name: ack-applicationautoscaling-controller.v0.2.9
        version: 0.2.9
      - name: ack-applicationautoscaling-controller.v0.2.8
        version: 0.2.8
      - name: ack-applicationautoscaling-controller.v0.2.7
        version: 0.2.7
      - name: ack-applicationautoscaling-controller.v0.2.6
        version: 0.2.6
      - name: ack-applicationautoscaling-controller.v0.2.5
        version: 0.2.5
      - name: ack-applicationautoscaling-controller.v0.2.4
        version: 0.2.4
      - name: ack-applicationautoscaling-controller.v0.2.3
        version: 0.2.3
      - name: ack-applicationautoscaling-controller.v0.2.2
        version: 0.2.2
      - name: ack-applicationautoscaling-controller.v0.2.1
        version: 0.2.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-applicationautoscaling-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-acm-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-acm-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acm.services.k8s.aws/v1alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/acm-controller:1.1.1
          createdAt: "2025-09-20T01:19:56Z"
          description: AWS ACM controller is a service controller for managing ACM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Certificate represents the state of an AWS acm Certificate
              resource.
            displayName: Certificate
            kind: Certificate
            name: certificates.acm.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage AWS Certificate Manager (ACM) resources in AWS from within your Kubernetes cluster.

          **About Amazon ACM**

          AWS Certificate Manager (ACM) handles the complexity of creating, storing, and renewing public and private SSL/TLS X.509 certificates and keys that protect your AWS websites and applications. You can provide certificates for your [integrated AWS services](https://docs.aws.amazon.com/acm/latest/userguide/acm-services.html) either by issuing them directly with ACM or by [importing](https://docs.aws.amazon.com/acm/latest/userguide/import-certificate.html) third-party certificates into the ACM management system. ACM certificates can secure singular domain names, multiple specific domain names, wildcard domains, or combinations of these. ACM wildcard certificates can protect an unlimited number of subdomains. You can also [export](https://docs.aws.amazon.com/acm/latest/userguide/export-private.html) ACM certificates signed by AWS Private CA for use anywhere in your internal PKI.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon ACM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ACM Developer Resources
          url: https://aws.amazon.com/certificate-manager/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: acm maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/acm-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-acm-controller.v1.1.1
        version: 1.1.1
      - name: ack-acm-controller.v1.1.0
        version: 1.1.0
      - name: ack-acm-controller.v1.0.13
        version: 1.0.13
      - name: ack-acm-controller.v1.0.12
        version: 1.0.12
      - name: ack-acm-controller.v1.0.11
        version: 1.0.11
      - name: ack-acm-controller.v1.0.10
        version: 1.0.10
      - name: ack-acm-controller.v1.0.9
        version: 1.0.9
      - name: ack-acm-controller.v1.0.8
        version: 1.0.8
      - name: ack-acm-controller.v1.0.7
        version: 1.0.7
      - name: ack-acm-controller.v1.0.6
        version: 1.0.6
      - name: ack-acm-controller.v1.0.5
        version: 1.0.5
      - name: ack-acm-controller.v1.0.4
        version: 1.0.4
      - name: ack-acm-controller.v1.0.1
        version: 1.0.1
      - name: ack-acm-controller.v1.0.0
        version: 1.0.0
      - name: ack-acm-controller.v0.0.20
        version: 0.0.20
      - name: ack-acm-controller.v0.0.19
        version: 0.0.19
      - name: ack-acm-controller.v0.0.18
        version: 0.0.18
      - name: ack-acm-controller.v0.0.17
        version: 0.0.17
      - name: ack-acm-controller.v0.0.16
        version: 0.0.16
      - name: ack-acm-controller.v0.0.12
        version: 0.0.12
      - name: ack-acm-controller.v0.0.10
        version: 0.0.10
      - name: ack-acm-controller.v0.0.9
        version: 0.0.9
      - name: ack-acm-controller.v0.0.6
        version: 0.0.6
      - name: ack-acm-controller.v0.0.5
        version: 0.0.5
      - name: ack-acm-controller.v0.0.4
        version: 0.0.4
      - name: ack-acm-controller.v0.0.2
        version: 0.0.2
      - name: ack-acm-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-acm-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ricardo Maraschini
      provider-url: https://github.com/ricardomaraschini/tagger
    name: tagger
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tagger.v2.1.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tagger.dev/v1beta1",
                "kind": "Tag",
                "metadata": {
                  "name": "nginx"
                },
                "spec": {
                  "from": "docker.io/library/nginx:latest",
                  "generation": 0,
                  "mirror": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          containerImage: quay.io/tagger/operator:v2.1.17
          createdAt: "2021-08-02T00:00:00Z"
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/ricardomaraschini/tagger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tag is a Container Image representation as a Custom Resource.
            displayName: Tag
            kind: Tag
            name: tags.tagger.dev
            version: v1beta1
        description: |-
          A Container Image Orchestrator. Adds mirroring and Container Image manipulation capabilities
          to a Kubernetes cluster.

          Keeping track of all Container Images in use in a Kubernetes cluster is a complicated task.
          Container Images may come from numerous different Image Registries. In some cases, controlling
          how a stable version of a given Container Image looks escapes the user's authority. To add to
          this, Container Runtimes rely on remote registries (from the cluster's point of view) when
          obtaining Container Images, potentially making the process of pulling their blobs (manifests,
          config, and layers) slower.

          The notion of indexing Container Image versions by Tags is helpful. Still, it does not provide
          users with the right confidence to always use the intended Container Image – today's "latest"
          tag might not be tomorrow's "latest" tag. In addition to that, these Image Registries allow
          access to Container Images by their Manifest content's hash (i.e., usually sha256), which gives
          users the confidence at a cost in semantics.

          When releasing a new version of an application to Push and to Deploy are split into two
          distinct steps. Both the pusher and the puller need access to the same Image Registry, adding
          complexity.  Credentials are one example of the concerns. Other factors may pop up when
          running, for instance, in an air-gapped environment, where the cluster may not reach external
          Image Registries.

          Tagger aims to overcome these caveats by providing an image management abstraction layer. For
          instance, by providing a direct mapping between a Container Image tag (e.g., "latest") and its
          correspondent Manifest content's hash, users can then refer to the Container Image by its tag
          – and be sure to use that specific version. More than that, when allied with an Internal Image
          Registry, Tagger can also automatically mirror Container Images into the cluster.

          While using Tagger, Deployments can refer to Container Images by an arbitrarily defined name,
          such as "my-app", and Tagger will make sure that they use the right Container Image through its
          internal "tag to Manifest content's hash" mapping.

          For each new "release" of a given Container Image, Tagger creates a new Generation for it,
          making it easy to roll back to previously pushed "releases" of the same Container Image in case
          of problems.

          When integrated with an Internal Registry, Tagger allows users to push or pull Images directly
          without requiring an external Image Registry. It works as a layer between the user and the
          Internal Registry. Every time a new "release" of Container Images is pushed, all Deployments
          are updated automatically. Users don't need to know about the Internal Registry existence, if
          they are logged-in to the Kubernetes cluster, they can obtain old or provide new "Generations"
          for a Container Image.

          In summary, Tagger mirrors remote Container Images into a Kubernetes cluster indexing them in
          different Generations (allowing easy navigation through these multiple Generations), provides
          an interface allowing users to pull and push images directly to the Kubernetes cluster and
          provides full integration with Kubernetes Deployments (automatic triggers new rollouts on
          Container Image changes).

          By default Tagger won't be able to mirror until you provide it with a mirror registry config.
          After install you can configure the mirror by editing the Secret mirror-registry-config in the
          operator namespace. Follow below an example of a valid mirror-registry-config secret (you gonna
          have to provide your own address, username, password, etc).

          ```yaml
          apiVersion: v1
          kind: Secret
          metadata:
            name: mirror-registry-config
          data:
            address: cmVnaXN0cnkuaW8=
            username: YWRtaW4=
            password: d2hhdCB3ZXJlIHlvdSB0aGlua2luZz8K
            token: YW4gb3B0aW9uYWwgdG9rZW4gZm9yIHRva2VuIGJhc2VkIGF1dGg=
            insecure: dHJ1ZQ==
          ```

          To make pushing an pulling images easier you should go ahead and grab the `kubectl` plugin
          available [here](https://github.com/ricardomaraschini/tagger/releases). You can download a
          full version of the README with usage examples and configuration options
          [here](https://github.com/ricardomaraschini/tagger/releases/download/v2.1.17/README.pdf).
        displayName: Tagger
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tagger
        - mirroring
        - image
        links:
        - name: Documentation
          url: https://github.com/ricardomaraschini/tagger/releases/download/v2.1.17/README.pdf
        - name: Plugin (kubectl)
          url: https://github.com/ricardomaraschini/tagger/releases/tag/v2.1.17
        maintainers:
        - email: rmarasch@redhat.com
          name: Ricardo Maraschini
        maturity: alpha
        provider:
          name: Ricardo Maraschini
          url: https://github.com/ricardomaraschini/tagger
        relatedImages:
        - quay.io/tagger/operator:v2.1.17
        version: 2.1.17
      entries:
      - name: tagger.v2.1.17
        version: 2.1.17
      name: alpha
    defaultChannel: alpha
    packageName: tagger
    provider:
      name: Ricardo Maraschini
      url: https://github.com/ricardomaraschini/tagger
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecs-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "Service",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "TaskDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecs-controller:1.1.1
          createdAt: "2025-09-19T21:02:44Z"
          description: AWS ECS controller is a service controller for managing ECS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster represents the state of an AWS ecs Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.ecs.services.k8s.aws
            version: v1alpha1
          - description: Service represents the state of an AWS ecs Service resource.
            displayName: Service
            kind: Service
            name: services.ecs.services.k8s.aws
            version: v1alpha1
          - description: TaskDefinition represents the state of an AWS ecs TaskDefinition
              resource.
            displayName: TaskDefinition
            kind: TaskDefinition
            name: taskdefinitions.ecs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Container Service (ECS) resources in AWS from within your Kubernetes cluster.

          **About Amazon ECS**

          Amazon Elastic Container Service (Amazon ECS) is a fully managed container orchestration service that helps you easily deploy, manage, and scale containerized applications. As a fully managed service, Amazon ECS comes with AWS configuration and operational best practices built-in. It's integrated with both AWS and third-party tools, such as Amazon Elastic Container Registry and Docker. This integration makes it easier for teams to focus on building the applications, not the environment. You can run and scale your container workloads across AWS Regions in the cloud, and on-premises, without the complexity of managing a control plane.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECS Developer Resources
          url: https://aws.amazon.com/ecs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecs-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ecs-controller.v1.1.1
        version: 1.1.1
      - name: ack-ecs-controller.v1.0.15
        version: 1.0.15
      - name: ack-ecs-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecs-controller.v1.0.13
        version: 1.0.13
      - name: ack-ecs-controller.v1.0.12
        version: 1.0.12
      - name: ack-ecs-controller.v1.0.11
        version: 1.0.11
      - name: ack-ecs-controller.v1.0.10
        version: 1.0.10
      - name: ack-ecs-controller.v1.0.9
        version: 1.0.9
      - name: ack-ecs-controller.v1.0.8
        version: 1.0.8
      - name: ack-ecs-controller.v1.0.7
        version: 1.0.7
      - name: ack-ecs-controller.v1.0.6
        version: 1.0.6
      - name: ack-ecs-controller.v1.0.2
        version: 1.0.2
      - name: ack-ecs-controller.v1.0.1
        version: 1.0.1
      - name: ack-ecs-controller.v1.0.0
        version: 1.0.0
      - name: ack-ecs-controller.v0.0.8
        version: 0.0.8
      - name: ack-ecs-controller.v0.0.7
        version: 0.0.7
      - name: ack-ecs-controller.v0.0.6
        version: 0.0.6
      - name: ack-ecs-controller.v0.0.5
        version: 0.0.5
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseWeb
      provider-url: ""
    name: cnfv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cnfv-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eweb.enterpriseweb.com/v1",
                "kind": "CloudNFV",
                "metadata": {
                  "name": "cloudnfv-sample"
                },
                "spec": {
                  "size": 1
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CloudNFV is the Schema for the cloudnfvs API
            displayName: Cloud NFV
            kind: CloudNFV
            name: cloudnfvs.eweb.enterpriseweb.com
            version: v1
        description: Manages a cluster of CloudNFV Runtime Modules
        displayName: CloudNFV
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - automation
        - NFV
        - controller
        links:
        - name: Cnfv Operator
          url: https://cnfv-operator.domain
        maintainers:
        - email: bill@enterpriseweb.com
          name: Bill Malyk
        maturity: alpha
        provider:
          name: EnterpriseWeb
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61a164ad711478ac603242338f7b3c55986319479878b092be128581a823598e
        - quay.io/wmalyk28/cnfv-op@sha256:e5913ad5dd20cb1b2299d2387237ef15e645c9a8c9d4f5fe0044b86a6b3dc1df
        - quay.io/wmalyk28/cloudnfv-runtime@sha256:fcde8375576c9bff602d3a16ef1ba8ae88b780166649ba6ae355737b1d35a03f
        version: 0.0.1
      entries:
      - name: cnfv-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cnfv-operator
    provider:
      name: EnterpriseWeb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzounstructured-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzounstructured-operator.v3.1.200
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzounstructured.clusters.cambridgesemantics.com/v1",
                "kind": "AnzoUnstructured",
                "metadata": {
                  "name": "au01"
                },
                "spec": {
                  "auWorker": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-w",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-worker@sha256:964633d6b16ddc517a09ac86f3e3698f5a161a72cfa50ea9ba4e301fa631e347",
                                "name": "w",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  },
                  "msLeader": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-ms",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzo-microservices-leader@sha256:0637603c00a99533d57e16009516ec73d40703ae151e631fc281f51f1f0ca93b",
                                "name": "ms",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/unstructured-operator@sha256:633a8992e35601d99c0d690dd71af314605c223b57c2038fa7090be06977b2f9
          createdAt: "2025-07-08T09:47:55Z"
          description: Kubernetes Operator for Altair Graph Studio Unstructured
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoUnstructured is the Schema for the anzounstructureds
              API
            displayName: Altair Graph Studio Unstructured Operator
            kind: AnzoUnstructured
            name: anzounstructureds.anzounstructured.clusters.cambridgesemantics.com
            version: v1
        description: |-
          The Altair Graph Studio Unstructured Operator provides the way to install and configure an anzo unstructured setup on Red Hat K8S environment.
          Currently, this is possible only through existing Anzo installation.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v3.1.2/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Studio Unstructured Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Database
        - Big Data
        links:
        - name: Altair Graph Studio Unstructured Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v3.1.2/README.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/unstructured-operator@sha256:633a8992e35601d99c0d690dd71af314605c223b57c2038fa7090be06977b2f9
        - registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-worker@sha256:964633d6b16ddc517a09ac86f3e3698f5a161a72cfa50ea9ba4e301fa631e347
        - registry.connect.redhat.com/cambridgesemantics/anzo-microservices-leader@sha256:0637603c00a99533d57e16009516ec73d40703ae151e631fc281f51f1f0ca93b
        version: 3.1.200
      entries:
      - name: anzounstructured-operator.v3.1.200
        version: 3.1.200
      - name: anzounstructured-operator.v3.1.104
        version: 3.1.104
      - name: anzounstructured-operator.v3.1.103
        version: 3.1.103
      name: stable
    defaultChannel: stable
    packageName: anzounstructured-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: openshift-secondary-scheduler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secondaryscheduleroperator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "SecondaryScheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-secondary-scheduler-operator"
                },
                "spec": {
                  "managementState": "Managed",
                  "schedulerConfig": "secondary-scheduler-config"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-rhel9-operator@sha256:210cc51760c8d3a9e1e5ba2b41f575b667e0d541b995362426e5a28f3f666600
          createdAt: 2024/10/01
          description: Runs a secondary scheduler in an OpenShift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.2.0 <1.3.1'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/secondary-scheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SecondaryScheduler is the schema for the secondaryschedulers
              API
            displayName: Secondary Scheduler
            kind: SecondaryScheduler
            name: secondaryschedulers.operator.openshift.io
            version: v1
        description: |
          The Secondary Scheduler Operator provides the ability to use a customized scheduler image that was developed using the scheduler plugin framework as a secondary scheduler in OpenShift.

          ## Secondary Scheduler Configuration

          Once the Secondary Scheduler Operator is installed, you can configure the plugins to run for your secondary scheduler.

          The configuration of the secondary scheduler is defined through a config map that wraps the KubeSchedulerConfiguration YAML file in "config.yaml".

          ## Additional Parameters

          Additionally, the following parameters can be configured:

          * `schedulerConfig` - Set the config map configuration for the secondary scheduler.
          * `schedulerImage` - Set the default operand image for the secondary scheduler.
        displayName: Secondary Scheduler Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - secondary
        - scheduling
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/secondary-scheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-rhel9-operator@sha256:210cc51760c8d3a9e1e5ba2b41f575b667e0d541b995362426e5a28f3f666600
        version: 1.3.1
      entries:
      - name: secondaryscheduleroperator.v1.3.1
        version: 1.3.1
      - name: secondaryscheduleroperator.v1.3.0
        version: 1.3.0
      - name: secondaryscheduleroperator.v1.2.2
        version: 1.2.2
      - name: secondaryscheduleroperator.v1.2.1
        version: 1.2.1
      - name: secondaryscheduleroperator.v1.2.0
        version: 1.2.0
      - name: secondaryscheduleroperator.v1.1.4
        version: 1.1.4
      - name: secondaryscheduleroperator.v1.1.2
        version: 1.1.2
      name: stable
    defaultChannel: stable
    packageName: openshift-secondary-scheduler-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: klusterlet
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: klusterlet.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "Klusterlet",
                "metadata": {
                  "name": "klusterlet"
                },
                "spec": {
                  "clusterName": "cluster1",
                  "deployOption": {
                    "mode": "Singleton"
                  },
                  "externalServerURLs": [
                    {
                      "url": "https://localhost"
                    }
                  ],
                  "imagePullSpec": "quay.io/open-cluster-management/registration-operator:v0.13.1",
                  "namespace": "open-cluster-management-agent",
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/registration-operator:v0.13.1
          createdAt: "2024-04-10T15:27:49Z"
          description: Manages the installation and upgrade of the Klusterlet.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management-io/ocm
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Klusterlet defines the configurations of agents running on
              the managed cluster
            displayName: Klusterlet
            kind: Klusterlet
            name: klusterlets.operator.open-cluster-management.io
            version: v1
        description: "Note: This operator will stop updating, please use the latest
          operator in OperatorHub.io https://operatorhub.io/operator/klusterlet. \n\nThe
          Klusterlet proivdes the registation to the Hub clusters as a managed cluster.
          This operator supports the installation and upgrade of Klusterlet.\n\n##
          Prerequisites\nYou need a Hub cluster which has installed ClusterManager.\n\n##
          How to Install\n* Install Klusterlet operator by following instructions
          in top right button `Install`.\n* Optional: Create an agent namespace used
          to deploy Klusterlet. The default namespace is `open-cluster-management-agent`.\n*
          Create a secret `bootstrap-hub-kubeconfig` with the kubeconfig file of hub
          cluster in the agent namespace.\n\n  ```\n  $ kubectl create secret generic
          bootstrap-hub-kubeconfig --from-file=kubeconfig=<kubeconfig file of hub
          clsuter> -n <agent namespace>\n  ```\n\n* Create an instance of the Klusterlet
          resource to install the Klusterlet.\n\nThe pods of Klusterlet are created
          in agent namespace.\n\n```\n$ kubectl get pods -n open-cluster-management-agent\n
          \ NAME                                READY   STATUS    RESTARTS   AGE\n
          \ klusterlet-agent-55c6574876-nfxm9   1/1     Running   0          8m38s\n```\n\nThere
          is a ManagedCluster created in the Hub cluster.\n\n```\n$ kubectl get managedclusters.cluster.open-cluster-management.io\nNAME
          \                  HUB ACCEPTED   MANAGED CLUSTER URLS     JOINED   AVAILABLE
          \  AGE\ncluster1               false                                                        27s\n```\n"
        displayName: Klusterlet
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open-cluster-management
        - Klusterlet
        links:
        - name: Registration Operator
          url: https://github.com/open-cluster-management-io/ocm
        maintainers:
        - email: zyin@redhat.com
          name: Zhiwei Yin
        - email: deads@redhat.com
          name: David Eads
        - email: jqiu@redhat.com
          name: Jian Qiu
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/open-cluster-management/registration-operator:v0.13.1
        version: 0.13.1
      entries:
      - name: klusterlet.v0.13.1
        version: 0.13.1
      - name: klusterlet.v0.13.0
        version: 0.13.0
      - name: klusterlet.v0.12.0
        version: 0.12.0
      - name: klusterlet.v0.11.0
        version: 0.11.0
      - name: klusterlet.v0.10.0
        version: 0.10.0
      - name: klusterlet.v0.9.1
        version: 0.9.1
      - name: klusterlet.v0.8.0
        version: 0.8.0
      - name: klusterlet.v0.7.0
        version: 0.7.0
      - name: klusterlet.v0.6.0
        version: 0.6.0
      - name: klusterlet.v0.5.0
        version: 0.5.0
      - name: klusterlet.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: klusterlet
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-api-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-api-controller.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "example-registry-with-studio"
                },
                "spec": {
                  "app": {
                    "host": "example-registry-with-studio-app.apps.yourcluster.example"
                  },
                  "studioUi": {
                    "enabled": true,
                    "host": "example-registry-with-studio-studio-ui.apps.yourcluster.example"
                  },
                  "ui": {
                    "host": "example-registry-with-studio-ui.apps.yourcluster.example"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-operator:3.0.5
          createdAt: "2024-12-03T11:30:52Z"
          description: Deploy and manage Apicurio API Controller on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Apicurio API
              Controller.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Apicurio API Controller

          Apicurio API Controller consists of 2 components:
           - Apicurio Registry 3
           - Apicurio Studio

          Storage options: PostgreSQL, Kafka, MySQL, SQL Server, in-memory.

          ## Apicurio API Controller Operator

          Provides a quick and easy way to deploy and manage Apicurio API Controller on Kubernetes.

          **Features**
          - Supports installation and configuration of Apicurio API Controller
          - Perform a rolling upgrade of Apicurio API Controller

          ## Prerequisites

          This operator does not deploy storage for Apicurio API Controller. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio API Controller and Apicurio API Controller Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Apicurio API Controller
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - API Management
        - Apicurio Registry
        - Apicurio Studio
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-operator:3.0.5
        version: 0.0.1
      entries:
      - name: apicurio-api-controller.v0.0.1
        version: 0.0.1
      name: 0.x
    defaultChannel: 0.x
    packageName: apicurio-api-controller
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: unsupported
      operatorframework.io/arch.s390x: unsupported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: https://www.konveyor.io
    name: konveyor-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: konveyor-operator.v0.8.0-alpha.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:c1d1942b3bb3c4e99d9b5484493e27a432455608fee4548f1d352be34a8d6ded
          createdAt: "2025-08-13T23:43:34Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0-alpha.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-hub@sha256:6be8fe8307e6686b178ace42f6db5cab8f04c5e2e29e02d2edcb5dc28ee337e9
        - quay.io/konveyor/tackle2-addon-platform@sha256:12efa5b7fbf6d829879b19e2820dcf782e648059f51544ebd25251dfc67fc7ca
        - quay.io/konveyor/postgresql-15-c9s@sha256:d75c5a1a0fa686ccf701e5e55633aa73d0cbb2cd9b5d8d4d4e80d7c46d78c121
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-operator@sha256:c1d1942b3bb3c4e99d9b5484493e27a432455608fee4548f1d352be34a8d6ded
        - quay.io/konveyor/tackle2-addon-discovery@sha256:ba865a74287ff94375ed316f348ee2c1d962d8d350c8afae5dbb0aee9b64fdea
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:7a5750fb76935e55b7b59e8363f21f5f9b48b39492d7f3e58e2e6a84d07ab027
        - quay.io/konveyor/kai-solution-server@sha256:83987d12d8290e085b6b66de56d347a8bf262815a5f99796177cb9120513ee3a
        - quay.io/konveyor/java-external-provider@sha256:42a156f1d81352f5e09fa0ffe2b51a4d93a1f476755eb0bac9a1d24bd3a89631
        - quay.io/konveyor/kantra@sha256:06ae932d9c2f51035bc5a488dd61e8a31c344e72615b013419199168701ea3fb
        - quay.io/konveyor/tackle2-ui@sha256:867bf24dd1687a79f2a911c22c1059ca7bd025a307d290290265653a696b8e98
        - quay.io/konveyor/tackle-keycloak-init@sha256:07c2ca07922d273bd2a3d9156ad8af0e5bf973aeb316496ee65d7201c381c1b1
        - quay.io/konveyor/origin-oauth-proxy@sha256:2c811da843057119b246784f21dc7eec490669a8569e8295d8e914cdce555b24
        - quay.io/konveyor/generic-external-provider@sha256:d58e867b51d5e76cd3597cfb7f2478d221b966f84509dbdd44c609ff5d9169be
        version: 0.8.0-alpha.4
      entries:
      - name: konveyor-operator.v0.8.0-alpha.4
        version: 0.8.0-alpha.4
      - name: konveyor-operator.v0.8.0-alpha.2
        version: 0.8.0-alpha.2
      - name: konveyor-operator.v0.8.0-alpha.1
        version: 0.8.0-alpha.1
      - name: konveyor-operator.v0.7.0-alpha.2
        version: 0.7.0-alpha.2
      - name: konveyor-operator.v0.7.0-alpha.1
        version: 0.7.0-alpha.1
      - name: konveyor-operator.v0.6.0-alpha.2
        version: 0.6.0-alpha.2
      - name: konveyor-operator.v0.6.0-alpha.1
        version: 0.6.0-alpha.1
      - name: konveyor-operator.v0.5.0-alpha.1
        version: 0.5.0-alpha.1
      - name: konveyor-operator.v0.4.0-alpha.1
        version: 0.4.0-alpha.1
      - name: konveyor-operator.v0.3.0-alpha.6
        version: 0.3.0-alpha.6
      - name: konveyor-operator.v0.3.0-alpha.5
        version: 0.3.0-alpha.5
      - name: konveyor-operator.v0.3.0-alpha.4
        version: 0.3.0-alpha.4
      - name: konveyor-operator.v0.3.0-alpha.3
        version: 0.3.0-alpha.3
      - name: konveyor-operator.v0.3.0-alpha.2
        version: 0.3.0-alpha.2
      - name: konveyor-operator.v0.3.0-alpha.1
        version: 0.3.0-alpha.1
      - name: konveyor-operator.v0.2.0-alpha.3
        version: 0.2.0-alpha.3
      - name: konveyor-operator.v0.2.0-alpha.2
        version: 0.2.0-alpha.2
      - name: konveyor-operator.v0.2.0-alpha.1
        version: 0.2.0-alpha.1
      name: alpha
    - currentCSV: konveyor-operator.v0.8.0-beta.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:019a79dec40445b6289948e173e5bf2ebfa9b495a7bb077abd634c0667953a54
          createdAt: "2025-09-17T19:50:24Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0-beta.5'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/kai-solution-server@sha256:1e69cf8a342090a03a9dddb1b28ae3627052b6d636ffa290a11761b94fd10648
        - quay.io/konveyor/tackle2-ui@sha256:78398832e092557639531230cf6d023bb07b589f3587afeaacf8c68640abc454
        - quay.io/konveyor/tackle2-operator@sha256:019a79dec40445b6289948e173e5bf2ebfa9b495a7bb077abd634c0667953a54
        - quay.io/konveyor/tackle-keycloak-init@sha256:0b4bfc1fea756f10a0657861c4c422303a2349d8d9b04a23238a32ceaf106ac5
        - quay.io/konveyor/origin-oauth-proxy@sha256:8988f14db7dde3edcaed0541cecb0b4fd58fab9dbaffb79a6c1c4aee13b9bd73
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/kantra@sha256:ad6f5c05e5e950e33a745accd03ad933fb7bb4907f44455f88a2b5c27e6dd15e
        - quay.io/konveyor/tackle2-addon-platform@sha256:d0280ade44ba5e15210c07342608576b38e6993af2e479740675a4706cd2c875
        - quay.io/konveyor/tackle2-addon-discovery@sha256:a29a0248e222d9ac3225e9d3df603f1383fa8844856bf68ea618d303f0b79883
        - quay.io/konveyor/java-external-provider@sha256:0cbdd91cc5057889b8daa2f7611fa0c944fab951c73bdf88a3e28795e9a5fc33
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:fed49c5ff14bb232e63d2dc180d3b530da63f11bf97055fc9d86cbb9e47104a4
        - quay.io/konveyor/generic-external-provider@sha256:a6c550bad7a09a1c51e226247f11c00d8453987698832cb566582842219f1f3c
        - quay.io/konveyor/tackle2-hub@sha256:1611913558c9f1203a844b43969d7b1539dab3578b83460e8383ee6757c3c34c
        - quay.io/konveyor/postgresql-15-c9s@sha256:157a2941ae971d8f72b2f8af1d4a29deeb5548012a978fa986c1c6e52395da4d
        version: 0.8.0-beta.5
      entries:
      - name: konveyor-operator.v0.8.0-beta.5
        version: 0.8.0-beta.5
      - name: konveyor-operator.v0.8.0-beta.4
        version: 0.8.0-beta.4
      - name: konveyor-operator.v0.8.0-beta.2
        version: 0.8.0-beta.2
      - name: konveyor-operator.v0.8.0-beta.1
        version: 0.8.0-beta.1
      - name: konveyor-operator.v0.6.0-beta.2
        version: 0.6.0-beta.2
      - name: konveyor-operator.v0.6.0-beta.1
        version: 0.6.0-beta.1
      - name: konveyor-operator.v0.5.0-beta.3
        version: 0.5.0-beta.3
      - name: konveyor-operator.v0.5.0-beta.2
        version: 0.5.0-beta.2
      - name: konveyor-operator.v0.5.0-beta.1
        version: 0.5.0-beta.1
      - name: konveyor-operator.v0.3.0-beta.3
        version: 0.3.0-beta.3
      - name: konveyor-operator.v0.3.0-beta.2
        version: 0.3.0-beta.2
      - name: konveyor-operator.v0.3.0-beta.1.2
        version: 0.3.0-beta.1.2
      - name: konveyor-operator.v0.3.0-beta.1.1
        version: 0.3.0-beta.1.1
      - name: konveyor-operator.v0.3.0-beta.1
        version: 0.3.0-beta.1
      name: beta
    - currentCSV: konveyor-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "move2kube-plan",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "image": "quay.io/konveyor/tackle-addon-move2kube-plan:latest"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:9688d0763d8c7812beee715a061c3c198006448855d7b891d45833f62cf869fb
          createdAt: "2023-05-25T15:23:45Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.1.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)

        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:9688d0763d8c7812beee715a061c3c198006448855d7b891d45833f62cf869fb
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/postgresql-12-centos7@sha256:1287d947464fbf5275feb31d106d2feb856c6117f0eaec8d947f2e03bd8618c5
        - quay.io/konveyor/tackle2-addon@sha256:35be344bbd3393f5634f52aafd21e9c3964d1e0ba5011ecad1561600d476bbe9
        - quay.io/konveyor/tackle-keycloak-init@sha256:b0347e5a608352c4b7bb796b65a67d417e98ba4896db71a487fa26243214cfcf
        - quay.io/konveyor/tackle2-hub@sha256:3d15d158ffb6f1fc4ca5aaf29a12173fe344c289773902c9161056bdf69acc92
        - quay.io/konveyor/tackle2-ui@sha256:99650804e65e2b6cb15a37c00189d7f463e4d19a0f866a79e9c09c8420a4d028
        - quay.io/konveyor/tackle2-addon-windup@sha256:81f0625cb82eb24ded51c7258f229f57778820c62e3be6ec05fb2b12d2970825
        - quay.io/konveyor/tackle-pathfinder@sha256:f9c5aa49f1d27f477cdd82c21441ae0287acdcdc9b2b3fcf197884007dc5f7dd
        version: 0.1.2
      entries:
      - name: konveyor-operator.v0.1.2
        version: 0.1.2
      - name: konveyor-operator.v0.1.1
        version: 0.1.1
      - name: konveyor-operator.v0.1.0
        version: 0.1.0
      name: konveyor-0.1
    - currentCSV: konveyor-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "move2kube-plan",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "image": "quay.io/konveyor/tackle-addon-move2kube-plan:latest"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:1860bb099ac64d9a2af27692f0671b7240e4d19746fffa9f9ae2356ff2818e10
          createdAt: "2023-07-06T06:17:40Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.2.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)

        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:1860bb099ac64d9a2af27692f0671b7240e4d19746fffa9f9ae2356ff2818e10
        - quay.io/konveyor/postgresql-12-centos7@sha256:6016530ab1fce71d29421ce9a97b3a89fbbb126b87b1b079c6aadebb7ff5664c
        - quay.io/konveyor/tackle2-hub@sha256:d9c4f21f44098a2526474968b6d88094dca8b5eaeb9042716332c9cb67df47cd
        - quay.io/konveyor/tackle2-ui@sha256:d10c1e717d160c5eb2581649a8dd7cdf4922a74d92db014f72f42a48a3134341
        - quay.io/konveyor/tackle2-addon-windup@sha256:538dfdb134d53a876d126dbd16679638584a606ef2b9e8fbe69ed1f29733c03a
        - quay.io/konveyor/tackle-keycloak-init@sha256:ca6c85bdf8417d4be116c2098b8d8ad5ae8b7ab0b636c4cbe4a392aa424e2d47
        - quay.io/konveyor/tackle2-addon@sha256:11693e16363bcd98249209d63a29f0c6cfffab3a4071649cbeea2fd276bc2440
        - quay.io/konveyor/tackle-pathfinder@sha256:190269b061b2065ca8ba9357bf9b959a2f587a547a44769ae0afd1ac86b4b197
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        version: 0.2.1
      entries:
      - name: konveyor-operator.v0.2.1
        version: 0.2.1
      - name: konveyor-operator.v0.2.0
        version: 0.2.0
      name: konveyor-0.2
    - currentCSV: konveyor-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:cd2305373f50ce927393e43f5735915e46c32d6f8a76091ee2808218885bba72
          createdAt: "2024-05-02T22:12:42Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.3.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:cd2305373f50ce927393e43f5735915e46c32d6f8a76091ee2808218885bba72
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:51bda0c6577c15684583d607a4f0cebef760814102423f4b2025016093a19fc1
        - quay.io/konveyor/origin-oauth-proxy@sha256:05f6bd1216b921b1f47a3821f0e22b84ed6b141bb2ada474c12bd9fde613108c
        - quay.io/konveyor/tackle2-hub@sha256:bf8803721d06d1a39251e22892343a013d5ef0fb4c1883c95e1eb55aa683f486
        - quay.io/konveyor/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle-keycloak-init@sha256:a63702cc0cb36830a7830ed70ca344a9aacc201ba6f3c94806ca2e93160df4dd
        - quay.io/konveyor/tackle2-ui@sha256:bbf58d2ea613489f91efae450421afafb9052416b0d7971606e2729e89fd7de6
        version: 0.3.2
      entries:
      - name: konveyor-operator.v0.3.2
        version: 0.3.2
      - name: konveyor-operator.v0.3.1
        version: 0.3.1
      - name: konveyor-operator.v0.3.0
        version: 0.3.0
      name: konveyor-0.3
    - currentCSV: konveyor-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:496db8ca859608d04008ac36cf7bdc9cafd7214294d33a9c5646083c361ffcb6
          createdAt: "2024-06-05T03:08:51Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.4.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:496db8ca859608d04008ac36cf7bdc9cafd7214294d33a9c5646083c361ffcb6
        - quay.io/konveyor/tackle-keycloak-init@sha256:4c7e07b1fce05f7dca47edecc0ba0ce373cc38e48879cc7eed72510d2865a18e
        - quay.io/konveyor/tackle2-ui@sha256:cb1255fc5fa4608fc659c3455f3164664b960cc2d9a3b2208967798e3532df78
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:e84ea36d206a8c143794e3986940377e77c7618225cad594bdf9a4047d7823dc
        - quay.io/konveyor/origin-oauth-proxy@sha256:05f6bd1216b921b1f47a3821f0e22b84ed6b141bb2ada474c12bd9fde613108c
        - quay.io/konveyor/tackle2-hub@sha256:b87acd8e2ee4a7edc4369bc375f64e2b98bc479d09556439e2ee4d74b74970a9
        - quay.io/konveyor/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        version: 0.4.0
      entries:
      - name: konveyor-operator.v0.4.0
        version: 0.4.0
      name: konveyor-0.4
    - currentCSV: konveyor-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:474d739ed470701b628f1f9e971925ac7fbe0b543f3e36305d932b963b1225fd
          createdAt: "2024-08-27T15:48:02Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.5.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-hub@sha256:faf91ca68e24979642a44fa013b1fb9edfee415eee835ccef9dce2eb418e2909
        - quay.io/konveyor/java-external-provider@sha256:33fb5d24b998f01c481a0a03e3f1bf3f7a6c93f5b662885102a0c966026e3375
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle-keycloak-init@sha256:5f652be5914ae6e23537b9fe02ccc9cd638c08de5462d9f58135a405c81f5287
        - quay.io/konveyor/tackle2-ui@sha256:36f29fe01d3078f0e100b24d1ed126850c5a773f4b0ae922fd4c4487de1ac29d
        - quay.io/konveyor/generic-external-provider@sha256:334a631538d31efc8d82adae4f52303336800fa3037b5055c72400918d81955c
        - quay.io/konveyor/postgresql-15-c9s@sha256:039ce764cadd03b4cfea954700139b8ceff415a9a0970baa7f2cf60f7f929174
        - quay.io/konveyor/tackle2-operator@sha256:474d739ed470701b628f1f9e971925ac7fbe0b543f3e36305d932b963b1225fd
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:2afbe8bea09904e084e026cd0a294fbe1c7b05d0e387f6633b09c5eaf4665e02
        - quay.io/konveyor/tackle2-addon-discovery@sha256:461f960bb1d29e719fb472fe9be9cf3deca17d984687957f54546a09047b0d3c
        version: 0.5.3
      entries:
      - name: konveyor-operator.v0.5.3
        version: 0.5.3
      - name: konveyor-operator.v0.5.1
        version: 0.5.1
      - name: konveyor-operator.v0.5.0
        version: 0.5.0
      name: konveyor-0.5
    - currentCSV: konveyor-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:8ee143c473570efd27b6f93332dc272871e11c66cc20f381c3a1e6e5688abf68
          createdAt: "2025-03-05T21:14:52Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.6.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle-keycloak-init@sha256:a3741269aa15edd2f9fb7df853a03a942d16a88eec114f3e54f82ef63b82adc4
        - quay.io/konveyor/tackle2-ui@sha256:12b0eda13a63508b9c13d4f9309bbfbe19d564627cf0de672307af1a6ec993fb
        - quay.io/konveyor/tackle2-operator@sha256:8ee143c473570efd27b6f93332dc272871e11c66cc20f381c3a1e6e5688abf68
        - quay.io/konveyor/java-external-provider@sha256:15148cde9a6fedef2207fe36f0ac3bd1ecdc6e3ac76f1cbec9eb97ffcf247231
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle2-hub@sha256:52a8b9be038fbfbc4e22172a4a8c7191d2b06fd8670e5df2ec442e3ce00d8c8a
        - quay.io/konveyor/tackle2-addon-discovery@sha256:3eb6f395c451e2c9a867871b3984a25f9f8e6ddcfe388f37fbd6d46a61a70b1e
        - quay.io/konveyor/generic-external-provider@sha256:27c1731859a4a639d2750cff8d1246281f4b0466bb5b069169c3e170c3c19b30
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:0b5cbb322fbe96335522410376c2e8505a8302088c0e454f3e5e7e28ba7c96a6
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/postgresql-15-c9s@sha256:55f8014c1edafab8f9085228a0d207c4e98820c615be34b935ce98e942d388b1
        version: 0.6.1
      entries:
      - name: konveyor-operator.v0.6.1
        version: 0.6.1
      - name: konveyor-operator.v0.6.0
        version: 0.6.0
      name: konveyor-0.6
    - currentCSV: konveyor-operator.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:f5870bc26b2a07ddacf00ed5b5f7df7d1068489010a80bcc543686c6c9576fbb
          createdAt: "2025-07-10T15:16:17Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:f5870bc26b2a07ddacf00ed5b5f7df7d1068489010a80bcc543686c6c9576fbb
        - quay.io/konveyor/tackle-keycloak-init@sha256:dab4cb1229c8a729d14b4cecb4aa3b5c4d48e94c36e84d785c922fc08d120617
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:7ed52ea9b95650241b5c2c01914775e29be23f7ed75fabb6e7c71969f3f0728f
        - quay.io/konveyor/generic-external-provider@sha256:facca8861559a658710ad3aef8b325731595bf944cce1907715d6efd33887072
        - quay.io/konveyor/tackle2-hub@sha256:52f66607f3134a4165ce691c6e9ce992d93927e9591055daafdc1bfc1dd09dd9
        - quay.io/konveyor/tackle2-ui@sha256:f87b04762fd163f40c49f3bcffd473f0c3f80a643576c6a3e2c2b4a14395c09f
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/postgresql-15-c9s@sha256:41460103a0bab189cd77acb6e741d9c5a6d3506479f7e56df40eb156d9b2cadd
        - quay.io/konveyor/tackle2-addon-discovery@sha256:60951801aa539ee59cd43f3c5f42a2ca751ca70e995cbb258b7d55ebb24d8450
        - quay.io/konveyor/java-external-provider@sha256:a0ef733e9fce3845d02b5360440add85427acad697d2157ef92242dec06bd73b
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/origin-oauth-proxy@sha256:2c811da843057119b246784f21dc7eec490669a8569e8295d8e914cdce555b24
        version: 0.7.1
      entries:
      - name: konveyor-operator.v0.7.1
        version: 0.7.1
      - name: konveyor-operator.v0.7.0
        version: 0.7.0
      name: konveyor-0.7
    - currentCSV: konveyor-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:68b6672fccdf54c65cc83f4c82db023053d6b61409bea880f58b53ccfea7b1ed
          createdAt: "2025-09-29T20:41:09Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-addon-platform@sha256:d0280ade44ba5e15210c07342608576b38e6993af2e479740675a4706cd2c875
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:8c82ad880ca2a2d5cd4b08125feca16d13d014d15f4f77677306086d1067f9aa
        - quay.io/konveyor/generic-external-provider@sha256:c7b6618bfff81d73f1b455963107ffdaf19334e6552ce076c009476da3f37f21
        - quay.io/konveyor/postgresql-15-c9s@sha256:af275f35a307367dcee9a339a6903cf79e3a8a41cabf79fb7cea7015895f0b58
        - quay.io/konveyor/origin-oauth-proxy@sha256:8988f14db7dde3edcaed0541cecb0b4fd58fab9dbaffb79a6c1c4aee13b9bd73
        - quay.io/konveyor/kai-solution-server@sha256:bab6e224c50de5b44cc37ce0297a3b91255efc08c1ad0fccf14fef9e5182775f
        - quay.io/konveyor/tackle2-operator@sha256:68b6672fccdf54c65cc83f4c82db023053d6b61409bea880f58b53ccfea7b1ed
        - quay.io/konveyor/tackle-keycloak-init@sha256:c0051b8c3a3c891fe12012623aee775080ecb47af8a4e75b8be8b0b4855d9f2c
        - quay.io/konveyor/tackle2-ui@sha256:f3f7380be65ff1fb3459c37a3d4865354bab685dc836893f78221f16ba579f70
        - quay.io/konveyor/kantra@sha256:a5b5591a51143574a6ce6678b141b9cd3e14328c6fad773412dde21b442fe364
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-hub@sha256:b3e6db1ffc195143b8e5d54b60949256e154b0f8b667736512a084e015cdf2a7
        - quay.io/konveyor/tackle2-addon-discovery@sha256:96a5b37834fbed9506014bcbd909258081fd833ef9e16daa104effd0e3a3c6ea
        - quay.io/konveyor/java-external-provider@sha256:8eab08a04e98051f89352250d63523a5b591f35a6497f5d69a5a9d556a94abf9
        version: 0.8.0
      entries:
      - name: konveyor-operator.v0.8.0
        version: 0.8.0
      name: konveyor-0.8
    - currentCSV: konveyor-operator.v0.7.0-rc.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
          createdAt: "2025-05-21T17:12:48Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.0-rc.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-ui@sha256:afe9141366a23c072286517e9af5436d14fd1df775aaa2fa13c3911764515777
        - quay.io/konveyor/postgresql-15-c9s@sha256:012cd6611aa9d3f4e0a82636d6817f381c19e3b0809199216fbe7fa3396a2ef5
        - quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/generic-external-provider@sha256:de55f78939576c110eb6f084a7ca51338fedba699fbb5135f23392783a5d7b95
        - quay.io/konveyor/tackle2-addon-discovery@sha256:89d2b0f8636897eb4dc1957a74a5b5041c0f3267a747adcf2bc6ca56b17b9937
        - quay.io/konveyor/origin-oauth-proxy@sha256:6c2847fd9667938b1d1294d988038e885e43cd9fb378862dae3419ce4f21631f
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle-keycloak-init@sha256:c3124971dc31be9201de2a6a1f6b2f886adf4c96731188dd5e750798cde3cfe7
        - quay.io/konveyor/tackle2-hub@sha256:bed834eb5488eef23b0d1e8af194b0646cf74c39c197e0d57acc5aed2299d9f5
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:46150725a2adbde8dfdc2250614d437a50e64bbf0b4562dabc73715a287de67b
        - quay.io/konveyor/java-external-provider@sha256:61ab20e37916abfeeadde685336660f0e8bbf519403d1f7100d43c1e42a72fbd
        version: 0.7.0-rc.1
      entries:
      - name: konveyor-operator.v0.7.0-rc.1
        version: 0.7.0-rc.1
      - name: konveyor-operator.v0.6.0-rc.2
        version: 0.6.0-rc.2
      - name: konveyor-operator.v0.6.0-rc.1
        version: 0.6.0-rc.1
      - name: konveyor-operator.v0.5.0-rc.1
        version: 0.5.0-rc.1
      - name: konveyor-operator.v0.3.0-rc.4
        version: 0.3.0-rc.4
      - name: konveyor-operator.v0.3.0-rc.3
        version: 0.3.0-rc.3
      - name: konveyor-operator.v0.3.0-rc.2
        version: 0.3.0-rc.2
      - name: konveyor-operator.v0.3.0-rc.1
        version: 0.3.0-rc.1
      name: rc
    - currentCSV: konveyor-operator.v0.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:33cf01f19a7043bcc5bdb12a3ed6bde7a9094981aa2383554111ebec9b4ba498
          createdAt: "2025-03-26T20:25:41Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.6.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/generic-external-provider@sha256:bc1112352f1a2c3172c5e7d5e637b5c5a69bf7d98b78f494fa593e8872e5cfd1
        - quay.io/konveyor/tackle2-ui@sha256:3c568c6047a044095080b6013b643a4041deae7b45936a443d4c4d708a81b945
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:734752be111a97e4b7e0030d563c32883f419ad031b2f350cb8e7333f2f37de3
        - quay.io/konveyor/tackle2-hub@sha256:0df49087d0c9341c2120a7c3aa2ba7907c056f94e1602ffd0496ca7113c75382
        - quay.io/konveyor/postgresql-15-c9s@sha256:7957d3b272b9f4e469bdc0c514ec9e70e8fbdc34a7250b4f4c62f9bb73cbf08d
        - quay.io/konveyor/tackle2-addon-discovery@sha256:1b89266d6bdfa497ba8137e0d33ff53785ac476dd96bf4fd9a185ef0ddf513ce
        - quay.io/konveyor/java-external-provider@sha256:fe85689b57125d678df685012b93c8a567fcb9251777cd711e9b82bdb22a734c
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/tackle2-operator@sha256:33cf01f19a7043bcc5bdb12a3ed6bde7a9094981aa2383554111ebec9b4ba498
        - quay.io/konveyor/tackle-keycloak-init@sha256:e32fa48351f410361e8bc100303deeb40d8a66d7eefb0cf7b1bc16e067a9a795
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        version: 0.6.2
      entries:
      - name: konveyor-operator.v0.6.2
        version: 0.6.2
      name: release-0.6
    - currentCSV: konveyor-operator.v0.7.0-rc.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
          createdAt: "2025-05-21T17:12:48Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.0-rc.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/generic-external-provider@sha256:de55f78939576c110eb6f084a7ca51338fedba699fbb5135f23392783a5d7b95
        - quay.io/konveyor/origin-oauth-proxy@sha256:6c2847fd9667938b1d1294d988038e885e43cd9fb378862dae3419ce4f21631f
        - quay.io/konveyor/tackle2-addon-discovery@sha256:89d2b0f8636897eb4dc1957a74a5b5041c0f3267a747adcf2bc6ca56b17b9937
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-hub@sha256:bed834eb5488eef23b0d1e8af194b0646cf74c39c197e0d57acc5aed2299d9f5
        - quay.io/konveyor/postgresql-15-c9s@sha256:012cd6611aa9d3f4e0a82636d6817f381c19e3b0809199216fbe7fa3396a2ef5
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:46150725a2adbde8dfdc2250614d437a50e64bbf0b4562dabc73715a287de67b
        - quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
        - quay.io/konveyor/tackle-keycloak-init@sha256:c3124971dc31be9201de2a6a1f6b2f886adf4c96731188dd5e750798cde3cfe7
        - quay.io/konveyor/tackle2-ui@sha256:afe9141366a23c072286517e9af5436d14fd1df775aaa2fa13c3911764515777
        - quay.io/konveyor/java-external-provider@sha256:61ab20e37916abfeeadde685336660f0e8bbf519403d1f7100d43c1e42a72fbd
        version: 0.7.0-rc.1
      entries:
      - name: konveyor-operator.v0.7.0-rc.1
        version: 0.7.0-rc.1
      name: release-0.7
    defaultChannel: konveyor-0.8
    packageName: konveyor-operator
    provider:
      name: Konveyor
      url: https://www.konveyor.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: sriov-network-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sriov-network-operator.v4.13.0-202309190544
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovIBNetwork",
                "metadata": {
                  "name": "example-sriovibnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "linkState": "auto",
                  "networkNamespace": "default",
                  "resourceName": "mellanoxnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetwork",
                "metadata": {
                  "name": "example-sriovnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "intelnics",
                  "spoofChk": "on",
                  "trust": "off",
                  "vlan": 0
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodePolicy",
                "metadata": {
                  "name": "policy-1",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "deviceType": "vfio-pci",
                  "mtu": 9000,
                  "nicSelector": {
                    "pfNames": [
                      "eth1"
                    ],
                    "rootDevices": [
                      "0000:01:00.1"
                    ],
                    "vendor": "8086"
                  },
                  "nodeSelector": {
                    "feature.node.kubernetes.io/network-sriov.capable": "true"
                  },
                  "numVfs": 6,
                  "priority": 99,
                  "resourceName": "intelnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodeState",
                "metadata": {
                  "name": "minikube",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "interfaces": []
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkPoolConfig",
                "metadata": {
                  "name": "sriovnetworkpoolconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovOperatorConfig",
                "metadata": {
                  "name": "default",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "enableInjector": true,
                  "enableOperatorWebhook": true,
                  "logLevel": 2
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: quay.io/pliurh/sriov-network-operator-manager:bd1
          createdAt: "2023-08-02T09:22:34Z"
          description: An operator for configuring SR-IOV components and initializing
            SRIOV network devices in Openshift cluster.
          olm.skipRange: '>=4.3.0-0 <4.13.0-202309190544'
          operatorframework.io/suggested-namespace: openshift-sriov-network-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "cni"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/k8snetworkplumbingwg/sriov-network-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovIBNetwork is the Schema for the sriovibnetworks API
            displayName: Sriov IBNetwork
            kind: SriovIBNetwork
            name: sriovibnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodePolicy is the Schema for the sriovnetworknodepolicies
              API
            displayName: Sriov Network Node Policy
            kind: SriovNetworkNodePolicy
            name: sriovnetworknodepolicies.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodeState is the Schema for the sriovnetworknodestates
              API
            displayName: Sriov Network Node State
            kind: SriovNetworkNodeState
            name: sriovnetworknodestates.sriovnetwork.openshift.io
            version: v1
          - kind: SriovNetworkPoolConfig
            name: sriovnetworkpoolconfigs.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetwork is the Schema for the sriovnetworks API
            displayName: Sriov Network
            kind: SriovNetwork
            name: sriovnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovOperatorConfig is the Schema for the sriovoperatorconfigs
              API
            displayName: Sriov Operator Config
            kind: SriovOperatorConfig
            name: sriovoperatorconfigs.sriovnetwork.openshift.io
            version: v1
        description: |
          # SR-IOV Network Operator for Openshift

          ## Introduction
          The sriov-network-operator is generally responsible for configuring the sriov components in a openshift cluster.

          ### Supported Features
          * Initialize the SR-IOV NICs on nodes.
          * provision SR-IOV device plugin on selected node.
          * provision SR-IOV CNI plugin on selected nodes.
          * manage configuration of SR-IOV device plugin.
          * generate net-att-def CRs for SR-IOV CNI plugin.

          This operator has to run in namespace 'openshift-sriov-network-operator'. An Operator Group is also required to install this operator:

          ```
          $ oc create namespace openshift-sriov-network-operator

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: sriov-network-operators
            namespace: openshift-sriov-network-operator
          spec:
            targetNamespaces:
            - openshift-sriov-network-operator
          EOF
          ```
        displayName: SR-IOV Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - network
        - sr-iov
        - sriov
        links:
        - name: Source Code
          url: https://github.com/k8snetworkplumbingwg/sriov-network-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-sriov-network-config-daemon@sha256:a2053c5366eb89112dfbc7a9994685fc198cce6fa87fdd6d1c0a07727e6eee03
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin@sha256:82d4846f8dfb9e0d02e7ff3f532f8f2bf44c9aba63c37d5ce2efb36969e5c453
        - registry.redhat.io/openshift4/ose-sriov-network-operator@sha256:3314ee580113b985e275583585b1f2adf381c5a0885b51100b3a01872a77a555
        - registry.redhat.io/openshift4/ose-sriov-network-webhook@sha256:bfb84d00dbb6b18d596ca2720b52d922c589249d2f988679189ea0eea985e58a
        - registry.redhat.io/openshift4/sriov-cni-rhel9@sha256:a4f0ec210cd065cb7f77dd25257aadf8345f47819cc728c5b6229acb79a09bb3
        - registry.redhat.io/openshift4/ose-sriov-dp-admission-controller@sha256:8dfaa07a5775a7156d303019879e41b17833e5643794f6e1087881d2a4dd1781
        - registry.redhat.io/openshift4/ose-sriov-infiniband-cni@sha256:716aa0b9ecfddfea0a62615f424f4abeab5362899c695e9ecc9df4df05f1add5
        version: 4.13.0-202309190544
      entries:
      - name: sriov-network-operator.v4.13.0-202309190544
        version: 4.13.0-202309190544
      name: "4.13"
    - currentCSV: sriov-network-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovIBNetwork",
                "metadata": {
                  "name": "example-sriovibnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "linkState": "auto",
                  "networkNamespace": "default",
                  "resourceName": "mellanoxnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetwork",
                "metadata": {
                  "name": "example-sriovnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "intelnics",
                  "spoofChk": "on",
                  "trust": "off",
                  "vlan": 0
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodePolicy",
                "metadata": {
                  "name": "policy-1",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "deviceType": "vfio-pci",
                  "mtu": 9000,
                  "nicSelector": {
                    "pfNames": [
                      "eth1"
                    ],
                    "rootDevices": [
                      "0000:01:00.1"
                    ],
                    "vendor": "8086"
                  },
                  "nodeSelector": {
                    "feature.node.kubernetes.io/network-sriov.capable": "true"
                  },
                  "numVfs": 6,
                  "priority": 99,
                  "resourceName": "intelnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodeState",
                "metadata": {
                  "name": "minikube",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "interfaces": []
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkPoolConfig",
                "metadata": {
                  "name": "sriovnetworkpoolconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovOperatorConfig",
                "metadata": {
                  "name": "default",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "enableInjector": true,
                  "enableOperatorWebhook": true,
                  "logLevel": 2
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-sriov-network-rhel9-operator@sha256:8dcae2a77e904713fba9c5907789f9b22ffb1f941f793adc4b604798d29a44f2
          createdAt: "2024-05-07T23:50:52Z"
          description: An operator for configuring SR-IOV components and initializing
            SRIOV network devices in Openshift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.16.0-202509240123'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "sriovnetwork.openshift.io/v1",
              "kind": "SriovOperatorConfig",
              "metadata": {
                "name": "default"
              },
              "spec": {
                "configurationMode": "daemon",
                "disableDrain": false,
                "enableInjector": true,
                "enableOperatorWebhook": true,
                "logLevel": 2
              }
            }
          operatorframework.io/suggested-namespace: openshift-sriov-network-operator
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-sriov-network-operator",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["sriovnetworknodestates.sriovnetwork.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/k8snetworkplumbingwg/sriov-network-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovIBNetwork is the Schema for the sriovibnetworks API
            displayName: Sriov IBNetwork
            kind: SriovIBNetwork
            name: sriovibnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodePolicy is the Schema for the sriovnetworknodepolicies
              API
            displayName: Sriov Network Node Policy
            kind: SriovNetworkNodePolicy
            name: sriovnetworknodepolicies.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodeState is the Schema for the sriovnetworknodestates
              API
            displayName: Sriov Network Node State
            kind: SriovNetworkNodeState
            name: sriovnetworknodestates.sriovnetwork.openshift.io
            version: v1
          - kind: SriovNetworkPoolConfig
            name: sriovnetworkpoolconfigs.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetwork is the Schema for the sriovnetworks API
            displayName: Sriov Network
            kind: SriovNetwork
            name: sriovnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovOperatorConfig is the Schema for the sriovoperatorconfigs
              API
            displayName: Sriov Operator Config
            kind: SriovOperatorConfig
            name: sriovoperatorconfigs.sriovnetwork.openshift.io
            version: v1
        description: |
          # SR-IOV Network Operator for Openshift

          ## Introduction
          The sriov-network-operator is generally responsible for configuring the sriov components in a openshift cluster.

          ### Supported Features
          * Initialize the SR-IOV NICs on nodes.
          * provision SR-IOV device plugin on selected node.
          * provision SR-IOV CNI plugin on selected nodes.
          * manage configuration of SR-IOV device plugin.
          * generate net-att-def CRs for SR-IOV CNI plugin.

          This operator has to run in namespace 'openshift-sriov-network-operator'. An Operator Group is also required to install this operator:

          ```
          $ oc create namespace openshift-sriov-network-operator

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: sriov-network-operators
            namespace: openshift-sriov-network-operator
          spec:
            targetNamespaces:
            - openshift-sriov-network-operator
          EOF
          ```
        displayName: SR-IOV Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - network
        - sr-iov
        - sriov
        links:
        - name: Source Code
          url: https://github.com/k8snetworkplumbingwg/sriov-network-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-sriov-network-config-daemon-rhel9@sha256:61afbb271a08dc3ab3b5509df535d9456f9b2c1b267554386b4bf859b6bccbe4
        - registry.redhat.io/openshift4/ose-sriov-infiniband-cni-rhel9@sha256:f828e415db73bf2df3fa7b26238a2f5c8a53cda6b396c11a0337d688108e6647
        - registry.redhat.io/openshift4/sriov-cni-rhel9@sha256:cac67ac25a0d319b5ea68cb3f63eac058339f1981811c7a3466b9daa2534a38d
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin-rhel9@sha256:5b3bc02e31eecd0d3468db729b03239d518ef73112b0fdb5ec3f05a788baa0be
        - registry.redhat.io/openshift4/ose-sriov-network-rhel9-operator@sha256:8dcae2a77e904713fba9c5907789f9b22ffb1f941f793adc4b604798d29a44f2
        - registry.redhat.io/openshift4/ose-sriov-network-webhook-rhel9@sha256:a6a5cc76d88b6b7aa86ee24b882ce43be33c8c6d4bcfae5f1b5490ead71ce8b6
        - registry.redhat.io/openshift4/ose-sriov-network-metrics-exporter-rhel9@sha256:a0580d8156de0ca1e5a119441196e9babbeb4d7edc502659c7ac66a49b1928d5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-sriov-dp-admission-controller-rhel9@sha256:20fd8a0081ffdea05796c8f6a0539a0e246da24f739f7bbf0722c90acfdeed15
        version: 4.16.0-202509240123
      entries:
      - name: sriov-network-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: sriov-network-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: sriov-network-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: sriov-network-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: sriov-network-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: sriov-network-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: sriov-network-operator.v4.16.0-202504290837
        version: 4.16.0-202504290837
      - name: sriov-network-operator.v4.16.0-202504151638
        version: 4.16.0-202504151638
      - name: sriov-network-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: sriov-network-operator.v4.16.0-202502262005
        version: 4.16.0-202502262005
      - name: sriov-network-operator.v4.16.0-202502171404
        version: 4.16.0-202502171404
      - name: sriov-network-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: sriov-network-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: sriov-network-operator.v4.16.0-202501081506
        version: 4.16.0-202501081506
      - name: sriov-network-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: sriov-network-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: sriov-network-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: sriov-network-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: sriov-network-operator.v4.16.0-202410300306
        version: 4.16.0-202410300306
      - name: sriov-network-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: sriov-network-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: sriov-network-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: sriov-network-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: sriov-network-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: sriov-network-operator.v4.16.0-202409111535
        version: 4.16.0-202409111535
      - name: sriov-network-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: sriov-network-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: sriov-network-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: sriov-network-operator.v4.16.0-202408120337
        version: 4.16.0-202408120337
      - name: sriov-network-operator.v4.16.0-202407100107
        version: 4.16.0-202407100107
      - name: sriov-network-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: sriov-network-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: sriov-network-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: sriov-network-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: sriov-network-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: The cert-manager maintainers
      provider-url: https://cert-manager.io/
    name: cert-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-manager.v1.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "example-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "example-clusterissuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example-certificate"
                },
                "spec": {
                  "dnsNames": [
                    "example.com"
                  ],
                  "issuerRef": {
                    "name": "example-issuer"
                  },
                  "secretName": "example-certificate-tls"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "name": "example-certificaterequest"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "name": "example-order"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "example-challenge"
                },
                "spec": {}
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: quay.io/jetstack/cert-manager-controller:v1.16.5
          createdAt: 2025-05-02T09:11:31
          olm.skipRange: '>=1.16.0 <1.16.5'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/internal-objects: |-
            [
              "challenges.acme.cert-manager.io",
              "orders.acme.cert-manager.io"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/cert-manager/cert-manager
          support: The cert-manager maintainers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A CertificateRequest is used to request a signed certificate from one of the
              configured issuers.

              All fields within the CertificateRequest's `spec` are immutable after creation.
              A CertificateRequest will either succeed or fail, as denoted by its `Ready` status
              condition and its `status.failureTime` field.

              A CertificateRequest is a one-shot resource, meaning it represents a single
              point in time request for a certificate and cannot be re-used.
            displayName: CertificateRequest
            kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - description: |-
              A Certificate resource should be created to ensure an up to date and signed
              X.509 certificate is stored in the Kubernetes Secret resource named in `spec.secretName`.

              The stored certificate will be renewed before it expires (as configured by `spec.renewBefore`).
            displayName: Certificate
            kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: Challenge is a type to represent a Challenge request with
              an ACME server
            displayName: (Internal) Challenge
            kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - description: |-
              A ClusterIssuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is similar to an Issuer, however it is cluster-scoped and therefore can
              be referenced by resources that exist in *any* namespace, not just the same
              namespace as the referent.
            displayName: ClusterIssuer
            kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - description: |-
              An Issuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is scoped to a single namespace and can therefore only be referenced by
              resources within the same namespace.
            displayName: Issuer
            kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - description: Order is a type to represent an Order with an ACME server
            displayName: (Internal) Order
            kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          cert-manager is a Kubernetes add-on to automate the management and issuance of TLS certificates from various issuing sources.
          It will ensure certificates are valid and up to date periodically, and attempt to renew certificates at an appropriate time before expiry.

          cert-manager has the following [built in issuers][]:

          * ACME: sign certificates using Let's Encrypt (or any other [ACME][] compliant signer).
          * Vault: sign certificates using HashiCorp Vault.
          * Venafi: sign certificates using Venafi Cloud or Venafi Trust Protection Platform.
          * SelfSigned: the private key of the certificate will be used to sign the certificate.
          * CA: a private key stored as a Kubernetes Secret will be used to sign the certificate.

          [built in issuers]: https://cert-manager.io/docs/configuration/#supported-issuer-types
          [ACME]: https://datatracker.ietf.org/doc/html/rfc8555

          cert-manager also supports [external issuers][],
          which are out-of-tree components that provide support for other certificate authorities, such as:
          [Google Cloud Certificate Authority](https://github.com/jetstack/google-cas-issuer),
          [AWS Private Certificate Authority](https://github.com/cert-manager/aws-privateca-issuer), and
          [Cloudflare Origin CA](https://github.com/cloudflare/origin-ca-issuer)

          [external issuers]: https://cert-manager.io/docs/configuration/external/

          In order to configure cert-manager to issue certificates, you must first create an Issuer or ClusterIssuer.
          These resources represent a particular signing authority and tell cert-manager how to handle your certificate requests.
          The [Issuer Documentation][] has more detail.

          [configure cert-manager]: https://cert-manager.io/docs/configuration/
          [Issuer Documentation]: https://cert-manager.io/docs/concepts/issuer/

          Once an Issuer has been configured, you can create a Certificate resource,
          which is a definition of a certificate request that is to be honored by an issuer
          and which is to be kept up-to-date.
          This is the usual way that you will interact with cert-manager to request signed certificates.
          The [Certificate Documentation][] has more detail.

          [Certificate Documentation]: https://cert-manager.io/docs/usage/certificate/
        displayName: cert-manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Documentation
          url: https://cert-manager.io/
        - name: Release Notes
          url: https://cert-manager.io/docs/release-notes
        maintainers:
        - email: cert-manager-maintainers@googlegroups.com
          name: The cert-manager maintainers
        maturity: stable
        minKubeVersion: 1.19.0-0
        provider:
          name: The cert-manager maintainers
          url: https://cert-manager.io/
        relatedImages:
        - quay.io/jetstack/cert-manager-controller:v1.16.5
        - quay.io/jetstack/cert-manager-cainjector:v1.16.5
        - quay.io/jetstack/cert-manager-webhook:v1.16.5
        version: 1.16.5
      entries:
      - name: cert-manager.v1.16.5
        version: 1.16.5
      - name: cert-manager.v1.16.1
        version: 1.16.1
      - name: cert-manager.v1.16.0-beta.0
        version: 1.16.0-beta.0
      - name: cert-manager.v1.15.2
        version: 1.15.2
      - name: cert-manager.v1.15.0
        version: 1.15.0
      - name: cert-manager.v1.14.2
        version: 1.14.2
      - name: cert-manager.v1.14.2-rc1
        version: 1.14.2-rc1
      - name: cert-manager.v1.14.1-rc1
        version: 1.14.1-rc1
      - name: cert-manager.v1.13.3
        version: 1.13.3
      - name: cert-manager.v1.13.3-rc1
        version: 1.13.3-rc1
      - name: cert-manager.v1.13.1
        version: 1.13.1
      - name: cert-manager.v1.13.1-rc1
        version: 1.13.1-rc1
      - name: cert-manager.v1.12.2
        version: 1.12.2
      - name: cert-manager.v1.12.2-rc1
        version: 1.12.2-rc1
      - name: cert-manager.v1.11.4
        version: 1.11.4
      - name: cert-manager.v1.11.4-rc1
        version: 1.11.4-rc1
      - name: cert-manager.v1.11.3-rc1
        version: 1.11.3-rc1
      - name: cert-manager.v1.11.0
        version: 1.11.0
      - name: cert-manager.v1.11.0-rc1
        version: 1.11.0-rc1
      - name: cert-manager.v1.10.2
        version: 1.10.2
      - name: cert-manager.v1.10.2-rc1
        version: 1.10.2-rc1
      - name: cert-manager.v1.10.1
        version: 1.10.1
      - name: cert-manager.v1.10.1-rc1
        version: 1.10.1-rc1
      - name: cert-manager.v1.10.0
        version: 1.10.0
      - name: cert-manager.v1.10.0-rc1
        version: 1.10.0-rc1
      - name: cert-manager.v1.9.1
        version: 1.9.1
      - name: cert-manager.v1.9.1-rc1
        version: 1.9.1-rc1
      - name: cert-manager.v1.8.2
        version: 1.8.2
      - name: cert-manager.v1.8.2-rc1
        version: 1.8.2-rc1
      - name: cert-manager.v1.8.0
        version: 1.8.0
      - name: cert-manager.v1.8.0-rc1
        version: 1.8.0-rc1
      - name: cert-manager.v1.7.2
        version: 1.7.2
      - name: cert-manager.v1.7.1
        version: 1.7.1
      - name: cert-manager.v1.7.1-rc1
        version: 1.7.1-rc1
      - name: cert-manager.v1.6.2
        version: 1.6.2
      - name: cert-manager.v1.6.2-rc1
        version: 1.6.2-rc1
      name: candidate
    - currentCSV: cert-manager.v1.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "example-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "example-clusterissuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example-certificate"
                },
                "spec": {
                  "dnsNames": [
                    "example.com"
                  ],
                  "issuerRef": {
                    "name": "example-issuer"
                  },
                  "secretName": "example-certificate-tls"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "name": "example-certificaterequest"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "name": "example-order"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "example-challenge"
                },
                "spec": {}
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: quay.io/jetstack/cert-manager-controller:v1.16.5
          createdAt: 2025-05-02T09:11:31
          olm.skipRange: '>=1.16.0 <1.16.5'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/internal-objects: |-
            [
              "challenges.acme.cert-manager.io",
              "orders.acme.cert-manager.io"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/cert-manager/cert-manager
          support: The cert-manager maintainers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A CertificateRequest is used to request a signed certificate from one of the
              configured issuers.

              All fields within the CertificateRequest's `spec` are immutable after creation.
              A CertificateRequest will either succeed or fail, as denoted by its `Ready` status
              condition and its `status.failureTime` field.

              A CertificateRequest is a one-shot resource, meaning it represents a single
              point in time request for a certificate and cannot be re-used.
            displayName: CertificateRequest
            kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - description: |-
              A Certificate resource should be created to ensure an up to date and signed
              X.509 certificate is stored in the Kubernetes Secret resource named in `spec.secretName`.

              The stored certificate will be renewed before it expires (as configured by `spec.renewBefore`).
            displayName: Certificate
            kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: Challenge is a type to represent a Challenge request with
              an ACME server
            displayName: (Internal) Challenge
            kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - description: |-
              A ClusterIssuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is similar to an Issuer, however it is cluster-scoped and therefore can
              be referenced by resources that exist in *any* namespace, not just the same
              namespace as the referent.
            displayName: ClusterIssuer
            kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - description: |-
              An Issuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is scoped to a single namespace and can therefore only be referenced by
              resources within the same namespace.
            displayName: Issuer
            kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - description: Order is a type to represent an Order with an ACME server
            displayName: (Internal) Order
            kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          cert-manager is a Kubernetes add-on to automate the management and issuance of TLS certificates from various issuing sources.
          It will ensure certificates are valid and up to date periodically, and attempt to renew certificates at an appropriate time before expiry.

          cert-manager has the following [built in issuers][]:

          * ACME: sign certificates using Let's Encrypt (or any other [ACME][] compliant signer).
          * Vault: sign certificates using HashiCorp Vault.
          * Venafi: sign certificates using Venafi Cloud or Venafi Trust Protection Platform.
          * SelfSigned: the private key of the certificate will be used to sign the certificate.
          * CA: a private key stored as a Kubernetes Secret will be used to sign the certificate.

          [built in issuers]: https://cert-manager.io/docs/configuration/#supported-issuer-types
          [ACME]: https://datatracker.ietf.org/doc/html/rfc8555

          cert-manager also supports [external issuers][],
          which are out-of-tree components that provide support for other certificate authorities, such as:
          [Google Cloud Certificate Authority](https://github.com/jetstack/google-cas-issuer),
          [AWS Private Certificate Authority](https://github.com/cert-manager/aws-privateca-issuer), and
          [Cloudflare Origin CA](https://github.com/cloudflare/origin-ca-issuer)

          [external issuers]: https://cert-manager.io/docs/configuration/external/

          In order to configure cert-manager to issue certificates, you must first create an Issuer or ClusterIssuer.
          These resources represent a particular signing authority and tell cert-manager how to handle your certificate requests.
          The [Issuer Documentation][] has more detail.

          [configure cert-manager]: https://cert-manager.io/docs/configuration/
          [Issuer Documentation]: https://cert-manager.io/docs/concepts/issuer/

          Once an Issuer has been configured, you can create a Certificate resource,
          which is a definition of a certificate request that is to be honored by an issuer
          and which is to be kept up-to-date.
          This is the usual way that you will interact with cert-manager to request signed certificates.
          The [Certificate Documentation][] has more detail.

          [Certificate Documentation]: https://cert-manager.io/docs/usage/certificate/
        displayName: cert-manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Documentation
          url: https://cert-manager.io/
        - name: Release Notes
          url: https://cert-manager.io/docs/release-notes
        maintainers:
        - email: cert-manager-maintainers@googlegroups.com
          name: The cert-manager maintainers
        maturity: stable
        minKubeVersion: 1.19.0-0
        provider:
          name: The cert-manager maintainers
          url: https://cert-manager.io/
        relatedImages:
        - quay.io/jetstack/cert-manager-controller:v1.16.5
        - quay.io/jetstack/cert-manager-cainjector:v1.16.5
        - quay.io/jetstack/cert-manager-webhook:v1.16.5
        version: 1.16.5
      entries:
      - name: cert-manager.v1.16.5
        version: 1.16.5
      - name: cert-manager.v1.16.1
        version: 1.16.1
      - name: cert-manager.v1.15.2
        version: 1.15.2
      - name: cert-manager.v1.15.0
        version: 1.15.0
      - name: cert-manager.v1.14.2
        version: 1.14.2
      - name: cert-manager.v1.13.3
        version: 1.13.3
      - name: cert-manager.v1.13.1
        version: 1.13.1
      - name: cert-manager.v1.13.1-rc1
        version: 1.13.1-rc1
      - name: cert-manager.v1.12.2
        version: 1.12.2
      - name: cert-manager.v1.11.4
        version: 1.11.4
      - name: cert-manager.v1.11.0
        version: 1.11.0
      - name: cert-manager.v1.10.2
        version: 1.10.2
      - name: cert-manager.v1.10.1
        version: 1.10.1
      - name: cert-manager.v1.10.0
        version: 1.10.0
      - name: cert-manager.v1.9.1
        version: 1.9.1
      - name: cert-manager.v1.8.2
        version: 1.8.2
      - name: cert-manager.v1.8.0
        version: 1.8.0
      - name: cert-manager.v1.7.2
        version: 1.7.2
      - name: cert-manager.v1.7.1
        version: 1.7.1
      - name: cert-manager.v1.6.2
        version: 1.6.2
      - name: cert-manager.v1.6.1
        version: 1.6.1
      - name: cert-manager.v1.6.0
        version: 1.6.0
      - name: cert-manager.v1.5.4
        version: 1.5.4
      - name: cert-manager.v1.5.3
        version: 1.5.3
      - name: cert-manager.v1.4.4
        version: 1.4.4
      - name: cert-manager.v1.4.3
        version: 1.4.3
      - name: cert-manager.v1.4.2
        version: 1.4.2
      - name: cert-manager.v1.4.1
        version: 1.4.1
      - name: cert-manager.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: cert-manager
    provider:
      name: The cert-manager maintainers
      url: https://cert-manager.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.12
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-06-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:9aed98a061b2c76f3c512c1d6453df5920f0c779b33d7168c1e7ab9d8bc928a4
        version: 0.0.12
      entries:
      - name: joget-dx8-openshift-operator.v0.0.12
        version: 0.0.12
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Dell Technologies
      provider-url: https://github.com/dell/csm-operator
    name: dell-csm-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dell-csm-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "authorization",
                  "namespace": "authorization"
                },
                "spec": {
                  "modules": [
                    {
                      "components": [
                        {
                          "enabled": true,
                          "name": "nginx"
                        },
                        {
                          "enabled": true,
                          "name": "cert-manager"
                        },
                        {
                          "authorizationController": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231",
                          "authorizationControllerReplicas": 1,
                          "certificate": "",
                          "controllerReconcileInterval": "5m",
                          "enabled": true,
                          "hostname": "csm-authorization.com",
                          "leaderElection": true,
                          "name": "proxy-server",
                          "opa": "docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095",
                          "opaKubeMgmt": "docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517",
                          "openTelemetryCollectorAddress": "",
                          "privateKey": "",
                          "proxyServerIngress": [
                            {
                              "annotations": {},
                              "hosts": [],
                              "ingressClassName": "nginx"
                            }
                          ],
                          "proxyService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682",
                          "proxyServiceReplicas": 1,
                          "roleService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c",
                          "roleServiceReplicas": 1,
                          "storageService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5",
                          "storageServiceReplicas": 1,
                          "tenantService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358",
                          "tenantServiceReplicas": 1
                        },
                        {
                          "commander": "docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a",
                          "name": "redis",
                          "redis": "docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f",
                          "redisCommander": "rediscommander",
                          "redisName": "redis-csm",
                          "redisReplicas": 5,
                          "redisSecretProviderClass": [
                            {
                              "redisPasswordKey": "",
                              "redisSecretName": "",
                              "redisUsernameKey": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "sentinel": "sentinel"
                        },
                        {
                          "configSecretProviderClass": [
                            {
                              "configSecretName": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "name": "config"
                        },
                        {
                          "name": "storage-system-credentials",
                          "secretProviderClasses": {
                            "vault": [
                              "secret-provider-class-1",
                              "secret-provider-class-2"
                            ]
                          }
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": true,
                      "forceRemoveModule": true,
                      "name": "authorization-proxy-server"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "isilon",
                  "namespace": "isilon"
                },
                "spec": {
                  "driver": {
                    "authSecret": "isilon-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VERBOSE",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_ISI_PORT",
                          "value": "8080"
                        },
                        {
                          "name": "X_CSI_ISI_PATH",
                          "value": "/ifs/data/csi"
                        },
                        {
                          "name": "X_CSI_ISI_NO_PROBE_ON_START",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_AUTOPROBE",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_AUTH_TYPE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_CUSTOM_TOPOLOGY_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOISILON_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "AZ_RECONCILE_INTERVAL",
                          "value": "1h"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_ISI_QUOTA_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_ACCESS_ZONE",
                          "value": "System"
                        },
                        {
                          "name": "X_CSI_ISI_VOLUME_PATH_PERMISSIONS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_IGNORE_UNRESOLVABLE_HOSTS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "isilon",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerscale"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSCALE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_PERFORMANCE_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSCALE_QUOTA_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "ISICLIENT_INSECURE",
                              "value": "true"
                            },
                            {
                              "name": "ISICLIENT_AUTH_TYPE",
                              "value": "1"
                            },
                            {
                              "name": "ISICLIENT_VERBOSE",
                              "value": "0"
                            },
                            {
                              "name": "POWERSCALE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSCALE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0",
                          "name": "metrics-powerscale"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-isilon/csi_sock",
                            "--mode=node",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powermax",
                  "namespace": "powermax"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powermax-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_MANAGED_ARRAYS",
                          "value": "000000000001,000000000002"
                        },
                        {
                          "name": "X_CSI_POWERMAX_PORTGROUPS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_TRANSPORT_PROTOCOL",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_K8S_CLUSTER_PREFIX",
                          "value": "CSM"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "X_CSI_VSPHERE_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VSPHERE_PORTGROUP",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VSPHERE_HOSTNAME",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VCENTER_HOST",
                          "value": ""
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "X_CSI_POWERMAX_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "X_CSI_POWERMAX_DEBUG",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_REVPROXY_USE_SECRET",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powermax",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERMAX_ISCSI_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_TOPOLOGY_CONTROL_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": [
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/memory-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/disk-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/network-unavailable",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REVPROXY_TLS_SECRET",
                              "value": "csirevproxy-tls-secret"
                            },
                            {
                              "name": "X_CSI_REVPROXY_PORT",
                              "value": "2222"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            },
                            {
                              "name": "DeployAsSidecar",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265",
                          "name": "csipowermax-reverseproxy"
                        }
                      ],
                      "configVersion": "v2.14.0",
                      "name": "csireverseproxy"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powermax"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERMAX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERMAX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414",
                          "name": "metrics-powermax"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/powermax.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powerstore",
                  "namespace": "powerstore"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powerstore-config",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_NODE_NAME_PREFIX",
                          "value": "csi-node"
                        },
                        {
                          "name": "X_CSI_FC_PORTS_FILTER_FILE_PATH",
                          "value": "/etc/fc-ports-filter"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOPOWERSTORE_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSM_AUTH_ENABLED",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_MULTI_NAS_FAILURE_THRESHOLD",
                          "value": "5"
                        },
                        {
                          "name": "X_CSI_MULTI_NAS_COOLDOWN_PERIOD",
                          "value": "5m"
                        },
                        {
                          "name": "X_CSI_NFS_ACLS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerstore",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-powerstore.dellemc.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerstore"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSTORE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_SPACE_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_ARRAY_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_FILE_SYSTEM_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSTORE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSTORE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "ZIPKIN_URI",
                              "value": ""
                            },
                            {
                              "name": "ZIPKIN_SERVICE_NAME",
                              "value": "metrics-powerstore"
                            },
                            {
                              "name": "ZIPKIN_PROBABILITY",
                              "value": "0.0"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60",
                          "name": "metrics-powerstore"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "unity",
                  "namespace": "unity"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_UNITY_ALLOW_MULTI_POD_ACCESS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_EPHEMERAL_STAGING_PATH",
                          "value": "/var/lib/kubelet/plugins/kubernetes.io/csi/pv/"
                        },
                        {
                          "name": "X_CSI_ISCSI_CHROOT",
                          "value": "/noderoot"
                        },
                        {
                          "name": "X_CSI_UNITY_SYNC_NODEINFO_INTERVAL",
                          "value": "15"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOUNITY_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOUNITY_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "TENANT_NAME",
                          "value": ""
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_UNITY_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "unity",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "vxflexos",
                  "namespace": "vxflexos"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLELISTVOLUMESNAPSHOT",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLESNAPSHOTCGDELETE",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "INFO"
                        },
                        {
                          "name": "GOSCALEIO_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOSCALEIO_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_QUOTA_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "INTERFACE_NAMES",
                          "value": null
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERFLEX_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "File",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerflex",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "initContainers": [
                      {
                        "envs": [
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          },
                          {
                            "name": "REPO_ADDRESS",
                            "value": "sftp://0.0.0.0"
                          },
                          {
                            "name": "REPO_USER",
                            "value": "sdcRepoUser"
                          },
                          {
                            "name": "MODULE_SIGCHECK",
                            "value": "0"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "imagePullPolicy": "IfNotPresent",
                        "name": "sdc"
                      }
                    ],
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_SDC_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_APPROVE_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_PREFIX",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_SDC_SFTP_REPO_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 1,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "enabled": false,
                        "envs": [
                          {
                            "name": "HOST_PID",
                            "value": "1"
                          },
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "name": "sdc-monitor"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "csi-external-health-monitor-controller"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERFLEX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_SDC_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_SDC_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERFLEX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d",
                          "name": "metrics-powerflex"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerflex"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--csisock=unix:/var/lib/kubelet/plugins/vxflexos.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
          createdAt: "2025-09-05T13:18:42Z"
          description: Easily install and manage Dell’s CSI Drivers and CSM
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/dell/csm-operator
          support: Dell Technologies
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ContainerStorageModule is the Schema for the containerstoragemodules
              API
            displayName: Container Storage Module
            kind: ContainerStorageModule
            name: containerstoragemodules.storage.dell.com
            version: v1
        description: "Dell Container Storage Modules (CSM) Operator is a Kubernetes
          Operator which can be \nused to install and manage Dell’s CSI drivers and
          CSM modules. \nBy using Dell CSM Operator, enterprises can quickly and easily
          deploy the \nCSM modules for CSI drivers making it easy for DevOps \nteams
          to build and optimize their Kubernetes workloads using enterprise grade
          storage \nwhile consistently aligning with the desired outcomes of the organization.\nThis
          operator is available as both community and certified operator for Openshift
          \nand as a community operator for upstream Kubernetes.\nOperator can be
          deployed via OperatorHub.io. using Operator Lifecycle Manager (OLM)\n\n###
          About Dell CSM\n\nDell CSM brings powerful enterprise storage features and
          functionality to Kubernetes for \neasier adoption of cloud native workloads,
          improved productivity and scalable operations. \n\nCSM delivers software
          modules for storage management that provide developers access to \nbuild
          automation for enhanced IT needs and other critical enterprise storage features.
          \n\nCurrently, the following modules are available as part of Dell CSM.
          \n*\tObservability - Delivers a high-level view of the storage capacity
          \n  and performance usage via Grafana dashboards to the Kubernetes users\n*\tReplication
          - Enables array replication capabilities for Kubernetes users\n*\tAuthorization
          - Provides both storage and Kubernetes administrators the ability \n  to
          apply RBAC for Dell CSI Drivers\n*\tResiliency - Designed to make Kubernetes
          applications more resilient to \n  node failures\n\nThese modules are available
          for various storage platforms in the Dell portfolio, \nsuch as Dell PowerStore,
          Dell PowerMax, Dell PowerFlex, Dell PowerScale, and Dell Unity XT.\n\nFor
          more information, visit [Dell’s Container Storage Module web page](https://www.dell.com/en-us/dt/solutions/kubernetes-containers/kubernetes-storage.htm)\n\nAdditional
          Materials\n*\t[Documentation](https://dell.github.io/csm-docs)\n*\t[CSM
          Solution Brief](https://www.delltechnologies.com/asset/en-us/products/storage/briefs-summaries/csm-solution-brief.pdf)\n"
        displayName: Dell Container Storage Modules
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dell Container Storage Modules
        - Dell CSI Driver
        - Dell CSM Modules
        - Powerflex
        - Powerscale
        - Powerstore
        - Unity
        - Authorization
        - Observability
        - Replication
        links:
        - name: Documentation
          url: https://dell.github.io/csm-docs
        maintainers:
        - email: container.storage.modules@dell.com
          name: Dell Container Storage Modules
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Dell Technologies
          url: https://github.com/dell/csm-operator
        relatedImages:
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0
        - registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa
        - docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265
        - registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea
        - docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682
        - registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334
        - docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f
        - quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:59335acaca515fceed5f0350c46cfcfd6037475f0cf35ae8ab41d37333039e2e
        - ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e
        - quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540
        - registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf
        - registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16
        - registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe
        version: 1.10.0
      entries:
      - name: dell-csm-operator-certified.v1.10.0
        version: 1.10.0
      - name: dell-csm-operator-certified.v1.9.1
        version: 1.9.1
      - name: dell-csm-operator-certified.v1.9.0
        version: 1.9.0
      - name: dell-csm-operator-certified.v1.8.1
        version: 1.8.1
      - name: dell-csm-operator-certified.v1.8.0
        version: 1.8.0
      - name: dell-csm-operator-certified.v1.7.0
        version: 1.7.0
      - name: dell-csm-operator-certified.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: dell-csm-operator-certified
    provider:
      name: Dell Technologies
      url: https://github.com/dell/csm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: external-dns-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: external-dns-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "provider": {
                    "type": "AWS",
                    "aws": {
                      "credentials": {
                        "name": "aws-access-key"
                      }
                    }
                  },
                  "source": {
                    "type": "Service"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:6be1403e418f9b63ec8ebcc8858e602409bc41fdee4586acc627623a78b50eb5
          createdAt: 2022-03-01T21:00:46
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## Prerequisites and Requirements
           The ExternalDNS Operator has minimal cluster level permissions. Unfortunately OLM doesn't allow the local role creation in the operand namespace yet.
           Therefore make sure the following commands are run against your cluster before the installation of the operator:
           * Create the operand namespace: `oc create ns external-dns`
           * Grant the operator access to manage operand resources: `oc apply -f https://raw.githubusercontent.com/openshift/external-dns-operator/release-0.1/config/rbac/extra-roles.yaml`
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b6fde43491c2e0733c8fe6ee8660f7f6435fe20c0b6d435cfc7f86eeab8b4748
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:6be1403e418f9b63ec8ebcc8858e602409bc41fdee4586acc627623a78b50eb5
        - registry.redhat.io/edo/external-dns-rhel8@sha256:c1134bb46172997ef7278b6cefbb0da44e72a9f808a7cd67b3c65d464754cab9
        version: 0.1.2
      entries:
      - name: external-dns-operator.v0.1.2
        version: 0.1.2
      name: alpha
    - currentCSV: external-dns-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
          createdAt: 2023-11-15T09:24:04
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e33874a51971a90917cf437fc6cbeea1da569ba07217f66f1bba96d594f58aed
        - registry.redhat.io/edo/external-dns-rhel8@sha256:6263a78b2a170e56dfca2173f14dbf14f0a5bddb6ea600bfdef248c7d5f20e4b
        version: 1.2.0
      entries:
      - name: external-dns-operator.v1.2.0
        version: 1.2.0
      name: stable-v1
    - currentCSV: external-dns-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:0877eb0a111efe0bff63173a2dfeb7a06a61afd0efbcdba053322a539e0dc667
          createdAt: 2024-01-22T09:26:37
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/edo/external-dns-rhel8@sha256:17f4127cbb9a9304769e9655c19cbea2260107c7e9371a99a9aa3e85d759c00b
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:0877eb0a111efe0bff63173a2dfeb7a06a61afd0efbcdba053322a539e0dc667
        version: 1.0.1
      entries:
      - name: external-dns-operator.v1.0.1
        version: 1.0.1
      - name: external-dns-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: external-dns-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:9ca751e84b8843cc62dc60550e5c4af741bf09a381ef62220e7eb3760ffe8524
          createdAt: 2024-01-22T09:56:14
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:9ca751e84b8843cc62dc60550e5c4af741bf09a381ef62220e7eb3760ffe8524
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/edo/external-dns-rhel8@sha256:ffcc3f2e9b9feb889e7882465f7f034aba6b9ee0699a1b6c24e83f501a63ecca
        version: 1.1.1
      entries:
      - name: external-dns-operator.v1.1.1
        version: 1.1.1
      - name: external-dns-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: external-dns-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
          createdAt: 2023-11-15T09:24:04
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e33874a51971a90917cf437fc6cbeea1da569ba07217f66f1bba96d594f58aed
        - registry.redhat.io/edo/external-dns-rhel8@sha256:6263a78b2a170e56dfca2173f14dbf14f0a5bddb6ea600bfdef248c7d5f20e4b
        version: 1.2.0
      entries:
      - name: external-dns-operator.v1.2.0
        version: 1.2.0
      name: stable-v1.2
    defaultChannel: stable-v1
    packageName: external-dns-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: rook-ceph-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rook-ceph-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephCluster",
                "metadata": {
                  "name": "my-rook-ceph",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "cephVersion": {
                    "image": "registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d"
                  },
                  "dataDirHostPath": "/var/lib/rook",
                  "mon": {
                    "count": 3
                  },
                  "dashboard": {
                    "enabled": true
                  },
                  "network": {
                    "hostNetwork": false
                  },
                  "rbdMirroring": {
                    "workers": 0
                  },
                  "storage": {
                    "useAllNodes": true,
                    "useAllDevices": true
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephBlockPool",
                "metadata": {
                  "name": "replicapool",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "failureDomain": "host",
                  "replicated": {
                    "size": 3
                  },
                  "annotations": null
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStore",
                "metadata": {
                  "name": "my-store",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "metadataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "dataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "gateway": {
                    "type": "s3",
                    "sslCertificateRef": null,
                    "port": 8080,
                    "securePort": null,
                    "instances": 1,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStoreUser",
                "metadata": {
                  "name": "my-user",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "store": "my-store",
                  "displayName": "my display name"
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephNFS",
                "metadata": {
                  "name": "my-nfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "rados": {
                    "pool": "myfs-data0",
                    "namespace": "nfs-ns"
                  },
                  "server": {
                    "active": 3,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephClient",
                "metadata": {
                  "name": "cinder",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "caps": {
                    "mon": "profile rbd",
                    "osd": "profile rbd pool=volumes, profile rbd pool=vms, profile rbd-read-only pool=images"
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephFilesystem",
                "metadata": {
                  "name": "myfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "dataPools": [
                    {
                      "compressionMode": "",
                      "crushRoot": "",
                      "deviceClass": "",
                      "erasureCoded": {
                        "algorithm": "",
                        "codingChunks": 0,
                        "dataChunks": 0
                      },
                      "failureDomain": "host",
                      "replicated": {
                        "requireSafeReplicaSize": false,
                        "size": 1,
                        "targetSizeRatio": 0.5
                      }
                    }
                  ],
                  "metadataPool": {
                    "compressionMode": "",
                    "crushRoot": "",
                    "deviceClass": "",
                    "erasureCoded": {
                      "algorithm": "",
                      "codingChunks": 0,
                      "dataChunks": 0
                    },
                    "failureDomain": "",
                    "replicated": {
                      "requireSafeReplicaSize": false,
                      "size": 1,
                      "targetSizeRatio": 0
                    }
                  },
                  "metadataServer": {
                    "activeCount": 1,
                    "activeStandby": true,
                    "placement": {},
                    "resources": {}
                  },
                  "preservePoolsOnDelete": false,
                  "preserveFilesystemOnDelete": false
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephRBDMirror",
                "metadata": {
                  "name": "my-rbd-mirror",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "annotations": null,
                  "count": 1,
                  "placement": {
                    "topologyKey": "kubernetes.io/hostname"
                  },
                  "resources": null
                }
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:cf7f72469d25fb37301a8159406613673f8968c82af1035fd45a470e0ae8acfa
          externalClusterScript: |-
            IiIiCkNvcHlyaWdodCAyMDIwIFRoZSBSb29rIEF1dGhvcnMuIEFsbCByaWdodHMgcmVzZXJ2ZWQu
            CgpMaWNlbnNlZCB1bmRlciB0aGUgQXBhY2hlIExpY2Vuc2UsIFZlcnNpb24gMi4wICh0aGUgIkxp
            Y2Vuc2UiKTsKeW91IG1heSBub3QgdXNlIHRoaXMgZmlsZSBleGNlcHQgaW4gY29tcGxpYW5jZSB3
            aXRoIHRoZSBMaWNlbnNlLgpZb3UgbWF5IG9idGFpbiBhIGNvcHkgb2YgdGhlIExpY2Vuc2UgYXQK
            CglodHRwOi8vd3d3LmFwYWNoZS5vcmcvbGljZW5zZXMvTElDRU5TRS0yLjAKClVubGVzcyByZXF1
            aXJlZCBieSBhcHBsaWNhYmxlIGxhdyBvciBhZ3JlZWQgdG8gaW4gd3JpdGluZywgc29mdHdhcmUK
            ZGlzdHJpYnV0ZWQgdW5kZXIgdGhlIExpY2Vuc2UgaXMgZGlzdHJpYnV0ZWQgb24gYW4gIkFTIElT
            IiBCQVNJUywKV0lUSE9VVCBXQVJSQU5USUVTIE9SIENPTkRJVElPTlMgT0YgQU5ZIEtJTkQsIGVp
            dGhlciBleHByZXNzIG9yIGltcGxpZWQuClNlZSB0aGUgTGljZW5zZSBmb3IgdGhlIHNwZWNpZmlj
            IGxhbmd1YWdlIGdvdmVybmluZyBwZXJtaXNzaW9ucyBhbmQKbGltaXRhdGlvbnMgdW5kZXIgdGhl
            IExpY2Vuc2UuCiIiIgoKaW1wb3J0IGVycm5vCmltcG9ydCBzeXMKaW1wb3J0IGpzb24KaW1wb3J0
            IGFyZ3BhcnNlCmltcG9ydCByZQppbXBvcnQgc3VicHJvY2VzcwppbXBvcnQgaG1hYwpmcm9tIGhh
            c2hsaWIgaW1wb3J0IHNoYTEgYXMgc2hhCmZyb20gb3MgaW1wb3J0IGxpbmVzZXAgYXMgTElORVNF
            UApmcm9tIG9zIGltcG9ydCBwYXRoCmZyb20gZW1haWwudXRpbHMgaW1wb3J0IGZvcm1hdGRhdGUK
            aW1wb3J0IHJlcXVlc3RzCmZyb20gcmVxdWVzdHMuYXV0aCBpbXBvcnQgQXV0aEJhc2UKCnB5M2sg
            PSBGYWxzZQppZiBzeXMudmVyc2lvbl9pbmZvLm1ham9yID49IDM6CiAgICBweTNrID0gVHJ1ZQog
            ICAgaW1wb3J0IHVybGxpYi5wYXJzZQogICAgZnJvbSBpcGFkZHJlc3MgaW1wb3J0IGlwX2FkZHJl
            c3MsIElQdjRBZGRyZXNzCgpNb2R1bGVOb3RGb3VuZEVycm9yID0gSW1wb3J0RXJyb3IKCnRyeToK
            ICAgIGltcG9ydCByYWRvcwpleGNlcHQgTW9kdWxlTm90Rm91bmRFcnJvciBhcyBub01vZEVycjoK
            ICAgIHByaW50KGYiRXJyb3I6IHtub01vZEVycn1cbkV4aXRpbmcgdGhlIHNjcmlwdC4uLiIpCiAg
            ICBzeXMuZXhpdCgxKQoKdHJ5OgogICAgaW1wb3J0IHJiZApleGNlcHQgTW9kdWxlTm90Rm91bmRF
            cnJvciBhcyBub01vZEVycjoKICAgIHByaW50KGYiRXJyb3I6IHtub01vZEVycn1cbkV4aXRpbmcg
            dGhlIHNjcmlwdC4uLiIpCiAgICBzeXMuZXhpdCgxKQoKdHJ5OgogICAgIyBmb3IgMi43LngKICAg
            IGZyb20gU3RyaW5nSU8gaW1wb3J0IFN0cmluZ0lPCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9y
            OgogICAgIyBmb3IgMy54CiAgICBmcm9tIGlvIGltcG9ydCBTdHJpbmdJTwoKdHJ5OgogICAgIyBm
            b3IgMi43LngKICAgIGZyb20gdXJscGFyc2UgaW1wb3J0IHVybHBhcnNlCiAgICBmcm9tIHVybGxp
            YiBpbXBvcnQgdXJsZW5jb2RlIGFzIHVybGVuY29kZQpleGNlcHQgTW9kdWxlTm90Rm91bmRFcnJv
            cjoKICAgICMgZm9yIDMueAogICAgZnJvbSB1cmxsaWIucGFyc2UgaW1wb3J0IHVybHBhcnNlCiAg
            ICBmcm9tIHVybGxpYi5wYXJzZSBpbXBvcnQgdXJsZW5jb2RlIGFzIHVybGVuY29kZQoKdHJ5Ogog
            ICAgZnJvbSBiYXNlNjQgaW1wb3J0IGVuY29kZXN0cmluZwpleGNlcHQ6CiAgICBmcm9tIGJhc2U2
            NCBpbXBvcnQgZW5jb2RlYnl0ZXMgYXMgZW5jb2Rlc3RyaW5nCgoKY2xhc3MgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbihFeGNlcHRpb24pOgogICAgcGFzcwoKCiMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMgRHVtbXlSYWRv
            cyAjIyMjIyMjIyMjIyMjIyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjCiMgdGhpcyBpcyBtYWlubHkgZm9yIHRlc3RpbmcgYW5kIGNvdWxkIGJlIHVz
            ZWQgd2hlcmUgJ3JhZG9zJyBpcyBub3QgYXZhaWxhYmxlCgoKY2xhc3MgRHVtbXlSYWRvcyhvYmpl
            Y3QpOgogICAgZGVmIF9faW5pdF9fKHNlbGYpOgogICAgICAgIHNlbGYucmV0dXJuX3ZhbCA9IDAK
            ICAgICAgICBzZWxmLmVycl9tZXNzYWdlID0gIiIKICAgICAgICBzZWxmLnN0YXRlID0gImNvbm5l
            Y3RlZCIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwID0ge30KICAgICAgICBzZWxmLmNtZF9u
            YW1lcyA9IHt9CiAgICAgICAgc2VsZi5faW5pdF9jbWRfb3V0cHV0X21hcCgpCiAgICAgICAgc2Vs
            Zi5kdW1teV9ob3N0X2lwX21hcCA9IHt9CgogICAgZGVmIF9pbml0X2NtZF9vdXRwdXRfbWFwKHNl
            bGYpOgogICAgICAgIGpzb25fZmlsZV9uYW1lID0gInRlc3QtZGF0YS9jZXBoLXN0YXR1cy1vdXQi
            CiAgICAgICAgc2NyaXB0X2RpciA9IHBhdGguYWJzcGF0aChwYXRoLmRpcm5hbWUoX19maWxlX18p
            KQogICAgICAgIGNlcGhfc3RhdHVzX3N0ciA9ICIiCiAgICAgICAgd2l0aCBvcGVuKAogICAgICAg
            ICAgICBwYXRoLmpvaW4oc2NyaXB0X2RpciwganNvbl9maWxlX25hbWUpLCBtb2RlPSJyIiwgZW5j
            b2Rpbmc9IlVURi04IgogICAgICAgICkgYXMganNvbl9maWxlOgogICAgICAgICAgICBjZXBoX3N0
            YXR1c19zdHIgPSBqc29uX2ZpbGUucmVhZCgpCiAgICAgICAgc2VsZi5jbWRfbmFtZXNbImZzIGxz
            Il0gPSAiIiJ7ImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJmcyBscyJ9IiIiCiAgICAgICAg
            c2VsZi5jbWRfbmFtZXNbInF1b3J1bV9zdGF0dXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3Jt
            YXQiOiAianNvbiIsICJwcmVmaXgiOiAicXVvcnVtX3N0YXR1cyJ9IiIiCiAgICAgICAgKQogICAg
            ICAgIHNlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXSA9ICgKICAgICAgICAgICAgIiIieyJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAibWdyIHNlcnZpY2VzIn0iIiIKICAgICAgICApCiAg
            ICAgICAgIyBhbGwgdGhlIGNvbW1hbmRzIGFuZCB0aGVpciBvdXRwdXQKICAgICAgICBzZWxmLmNt
            ZF9vdXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJmcyBscyJdXSA9ICgKICAgICAgICAgICAgIiIi
            W3sibmFtZSI6Im15ZnMiLCJtZXRhZGF0YV9wb29sIjoibXlmcy1tZXRhZGF0YSIsIm1ldGFkYXRh
            X3Bvb2xfaWQiOjIsImRhdGFfcG9vbF9pZHMiOlszXSwiZGF0YV9wb29scyI6WyJteWZzLXJlcGxp
            Y2F0ZWQiXX1dIiIiCiAgICAgICAgKQogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbc2VsZi5j
            bWRfbmFtZXNbInF1b3J1bV9zdGF0dXMiXV0gPSAoCiAgICAgICAgICAgICIiInsiZWxlY3Rpb25f
            ZXBvY2giOjMsInF1b3J1bSI6WzBdLCJxdW9ydW1fbmFtZXMiOlsiYSJdLCJxdW9ydW1fbGVhZGVy
            X25hbWUiOiJhIiwicXVvcnVtX2FnZSI6MTQzODUsImZlYXR1cmVzIjp7InF1b3J1bV9jb24iOiI0
            NTQwMTM4MjkyODM2Njk2MDYzIiwicXVvcnVtX21vbiI6WyJrcmFrZW4iLCJsdW1pbm91cyIsIm1p
            bWljIiwib3NkbWFwLXBydW5lIiwibmF1dGlsdXMiLCJvY3RvcHVzIl19LCJtb25tYXAiOnsiZXBv
            Y2giOjEsImZzaWQiOiJhZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5ZDBmMWMiLCJtb2Rp
            ZmllZCI6IjIwMjAtMDUtMDdUMDM6MzY6MzkuOTE4MDM1WiIsImNyZWF0ZWQiOiIyMDIwLTA1LTA3
            VDAzOjM2OjM5LjkxODAzNVoiLCJtaW5fbW9uX3JlbGVhc2UiOjE1LCJtaW5fbW9uX3JlbGVhc2Vf
            bmFtZSI6Im9jdG9wdXMiLCJmZWF0dXJlcyI6eyJwZXJzaXN0ZW50IjpbImtyYWtlbiIsImx1bWlu
            b3VzIiwibWltaWMiLCJvc2RtYXAtcHJ1bmUiLCJuYXV0aWx1cyIsIm9jdG9wdXMiXSwib3B0aW9u
            YWwiOltdfSwibW9ucyI6W3sicmFuayI6MCwibmFtZSI6ImEiLCJwdWJsaWNfYWRkcnMiOnsiYWRk
            cnZlYyI6W3sidHlwZSI6InYyIiwiYWRkciI6IjEwLjExMC4yMDUuMTc0OjMzMDAiLCJub25jZSI6
            MH0seyJ0eXBlIjoidjEiLCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OSIsIm5vbmNlIjowfV19
            LCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OS8wIiwicHVibGljX2FkZHIiOiIxMC4xMTAuMjA1
            LjE3NDo2Nzg5LzAiLCJwcmlvcml0eSI6MCwid2VpZ2h0IjowfV19fSIiIgogICAgICAgICkKICAg
            ICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXV0g
            PSAoCiAgICAgICAgICAgICIiInsiZGFzaGJvYXJkIjoiaHR0cHM6Ly9jZXBoLWRhc2hib2FyZDo4
            NDQzLyIsInByb21ldGhldXMiOiJodHRwOi8vY2VwaC1kYXNoYm9hcmQtZGI6OTI4My8ifSIiIgog
            ICAgICAgICkKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNh
            cHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMiLCAib3Nk
            IiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0
            YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250
            cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJdLCAiZW50aXR5Ijog
            ImNsaWVudC5oZWFsdGhjaGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRo
            IGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuaGVh
            bHRoY2hlY2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRB
            PT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cyIs
            Im9zZCI6InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5j
            b250cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJ9fV0iIiIKICAg
            ICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIs
            ICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAib3NkIiwgInBy
            b2ZpbGUgcmJkIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtbm9kZSIsICJmb3JtYXQiOiAi
            anNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIi
            W3siZW50aXR5IjoiY2xpZW50LmNzaS1yYmQtbm9kZSIsImtleSI6IkFRQk9nck5lSGJLMUF4QUF1
            YllCZVY4UzFVL0dQenE1U1ZlcTZnPT0iLCJjYXBzIjp7Im1vbiI6InByb2ZpbGUgcmJkLCBhbGxv
            dyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsIm9zZCI6InByb2ZpbGUgcmJkIn19XSIiIgogICAg
            ICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsibW9uIiwg
            InByb2ZpbGUgcmJkLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxs
            b3cgcnciLCAib3NkIiwgInByb2ZpbGUgcmJkIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQt
            cHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNy
            ZWF0ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktcmJkLXByb3Zp
            c2lvbmVyIiwia2V5IjoiQVFCTmdyTmUxZ2V5S3hBQThla1ZpUmRFK2hzczVPd2VZQmt3Tmc9PSIs
            ImNhcHMiOnsibWdyIjoiYWxsb3cgcnciLCJtb24iOiJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFu
            ZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJwcm9maWxlIHJiZCJ9fV0iIiIKICAgICAgICBzZWxm
            LmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyBy
            LCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3Nk
            IiwgImFsbG93IHJ3IHRhZyBjZXBoZnMgKj0qIiwgIm1kcyIsICJhbGxvdyBydyJdLCAiZW50aXR5
            IjogImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4Ijog
            ImF1dGggZ2V0LW9yLWNyZWF0ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVu
            dC5jc2ktY2VwaGZzLW5vZGUiLCJrZXkiOiJBUUJPZ3JOZUVOdW5LeEFBUENtZ0U3UjZHOERjWG5h
            SjFGMzJxZz09IiwiY2FwcyI6eyJtZHMiOiJhbGxvdyBydyIsIm1nciI6ImFsbG93IHJ3IiwibW9u
            IjoiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBy
            dyB0YWcgY2VwaGZzICo9KiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAg
            ICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kICdvc2Qg
            YmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3NkIiwgImFsbG93IHJ3IHRhZyBjZXBo
            ZnMgbWV0YWRhdGE9KiJdLCAiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVy
            IiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgog
            ICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIs
            ImtleSI6IkFRQk9nck5lQUZnY0dCQUF2R3FLT0FEMEQzeHhtVlkwUjkxMmRnPT0iLCJjYXBzIjp7
            Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2Nr
            bGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoifX1dIiIiCiAgICAg
            ICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAi
            YWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdyIiwgImFsbG93IHJ3
            IiwgIm9zZCIsICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiXSwgImVudGl0eSI6ICJj
            bGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVuc2hpZnQtc3RvcmFnZSIsICJmb3JtYXQi
            OiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0g
            IiIiW3siZW50aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0
            b3JhZ2UiLCJrZXkiOiJCUUJPZ3JOZUFGZ2NHQkFBdkdxS09BRDBEM3h4bVZZMFI5MTJkZz09Iiwi
            Y2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29z
            ZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIn19XSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsi
            bW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciIsICJh
            bGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT1teWZzIl0sICJl
            bnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0b3JhZ2Ut
            bXlmcyIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0i
            IiIKICAgICAgICBdID0gIiIiW3siZW50aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9u
            ZXItb3BlbnNoaWZ0LXN0b3JhZ2UtbXlmcyIsImtleSI6IkNRQk9nck5lQUZnY0dCQUF2R3FLT0FE
            MEQzeHhtVlkwUjkxMmRnPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cg
            ciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2Vw
            aGZzIG1ldGFkYXRhPW15ZnMifX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAg
            ICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCBxdW9y
            dW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIiwgImFsbG93IGNvbW1hbmQg
            Y29uZmlnIiwgIm9zZCIsICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFsbG93IHJ3eCBwb29sPWRl
            ZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9ZGVm
            YXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJndy5sb2csIGFsbG93IHgg
            cG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4Il0sICJlbnRpdHkiOiAiY2xpZW50LmhlYWx0
            aGNoZWNrZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0
            ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5oZWFsdGhjaGVja2VyIiwi
            a2V5IjoiQVFERmtiTmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsi
            bW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFu
            ZCB2ZXJzaW9uIiwgIm1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsICJvc2QiOiAicHJvZmls
            ZSByYmQtcmVhZC1vbmx5LCBhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhLCBhbGxvdyBy
            IHBvb2w9LnJndy5yb290LCBhbGxvdyBydyBwb29sPWRlZmF1bHQucmd3LmNvbnRyb2wsIGFsbG93
            IHJ4IHBvb2w9ZGVmYXVsdC5yZ3cubG9nLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0
            cy5pbmRleCJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAi
            IiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMs
            IGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cgY29tbWFuZCBjb25maWciLCAi
            b3NkIiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5j
            b250cm9sLCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1
            bHQucmd3LmJ1Y2tldHMuaW5kZXgiXSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIs
            ICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBjYXBzIn0iIiIKICAgICAgICBdID0g
            IiIiW3siZW50aXR5IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0NWJG
            UkFBVG5kTE5VU1JLcnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIjog
            ImFsbG93IGNvbW1hbmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFs
            bG93IHJ3eCBwb29sPWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFs
            bG93IHJ3IHBvb2w9ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJn
            dy5sb2csIGFsbG93IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4In19XSIiIgogICAg
            ICAgIHNlbGYuY21kX291dHB1dF9tYXBbIiIieyJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            bWdyIHNlcnZpY2VzIn0iIiJdID0gKAogICAgICAgICAgICAiIiJ7ImRhc2hib2FyZCI6ICJodHRw
            Oi8vcm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6NzAwMC8iLCAicHJvbWV0aGV1cyI6
            ICJodHRwOi8vcm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6OTI4My8ifSIiIgogICAg
            ICAgICkKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0
            eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            YXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJ7ImRhc2hib2FyZCI6ICJodHRwOi8vcm9vay1j
            ZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6NzAwMC8iLCAicHJvbWV0aGV1cyI6ICJodHRwOi8v
            cm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6OTI4My8ifSIiIgogICAgICAgIHNlbGYu
            Y21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNsaWVudC5oZWFsdGhj
            aGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldCJ9IiIiCiAgICAg
            ICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5oZWFsdGhjaGVja2VyIiwia2V5IjoiQVFERmti
            TmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjogImFsbG93
            IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFuZCB2ZXJzaW9uIiwg
            Im1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsICJvc2QiOiAicHJvZmlsZSByYmQtcmVhZC1v
            bmx5LCBhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhLCBhbGxvdyByIHBvb2w9LnJndy5y
            b290LCBhbGxvdyBydyBwb29sPWRlZmF1bHQucmd3LmNvbnRyb2wsIGFsbG93IHJ4IHBvb2w9ZGVm
            YXVsdC5yZ3cubG9nLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJ9fV0i
            IiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0eSI6
            ICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJh
            dXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0
            X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtbm9kZSIsICJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJb
            XSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5
            IjogImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRf
            b3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMt
            cHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIK
            ICAgICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAg
            ICAgICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVuc2hp
            ZnQtc3RvcmFnZSIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgog
            ICAgICAgIF0gPSAiIiJbXSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAg
            ICAgICIiInsiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlm
            dC1zdG9yYWdlLW15ZnMiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0i
            IiIKICAgICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAg
            ICAgICAgICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJb
            eyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsImtleSI6IkFRREZrYk5l
            ZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRBPT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIi
            LCAibWdyIjoiYWxsb3cgcnciLCAib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0q
            In19XSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2Fw
            cyI6IFsibW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1n
            ciIsICJhbGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIl0s
            ICJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpz
            b24iLCAicHJlZml4IjogImF1dGggY2FwcyJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmtiTmVmdDViRlJBQVRu
            ZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cgciwgIGFsbG93IGNv
            bW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciI6ImFsbG93IHJ3IiwgIm9zZCI6ImFsbG93IHJ3
            IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFw
            Wyd7ImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJzdGF0dXMifSddID0gY2VwaF9zdGF0dXNf
            c3RyCgogICAgZGVmIHNodXRkb3duKHNlbGYpOgogICAgICAgIHBhc3MKCiAgICBkZWYgZ2V0X2Zz
            aWQoc2VsZik6CiAgICAgICAgcmV0dXJuICJhZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5
            ZDBmMWMiCgogICAgZGVmIGNvbmZfcmVhZF9maWxlKHNlbGYpOgogICAgICAgIHBhc3MKCiAgICBk
            ZWYgY29ubmVjdChzZWxmKToKICAgICAgICBwYXNzCgogICAgZGVmIHBvb2xfZXhpc3RzKHNlbGYs
            IHBvb2xfbmFtZSk6CiAgICAgICAgcmV0dXJuIFRydWUKCiAgICBkZWYgbW9uX2NvbW1hbmQoc2Vs
            ZiwgY21kLCBvdXQpOgogICAgICAgIGpzb25fY21kID0ganNvbi5sb2FkcyhjbWQpCiAgICAgICAg
            anNvbl9jbWRfc3RyID0ganNvbi5kdW1wcyhqc29uX2NtZCwgc29ydF9rZXlzPVRydWUpCiAgICAg
            ICAgY21kX291dHB1dCA9IHNlbGYuY21kX291dHB1dF9tYXBbanNvbl9jbWRfc3RyXQogICAgICAg
            IHJldHVybiBzZWxmLnJldHVybl92YWwsIGNtZF9vdXRwdXQsIHN0cihzZWxmLmVycl9tZXNzYWdl
            LmVuY29kZSgidXRmLTgiKSkKCiAgICBkZWYgX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoc2VsZiwg
            aG9zdF9uYW1lKToKICAgICAgICBpcF9yZWdfeCA9IHJlLmNvbXBpbGUociJcZHsxLDN9LlxkezEs
            M30uXGR7MSwzfS5cZHsxLDN9IikKICAgICAgICAjIGlmIHByb3ZpZGVkIGhvc3QgaXMgZGlyZWN0
            bHkgYW4gSVAgYWRkcmVzcywgcmV0dXJuIHRoZSBzYW1lCiAgICAgICAgaWYgaXBfcmVnX3gubWF0
            Y2goaG9zdF9uYW1lKToKICAgICAgICAgICAgcmV0dXJuIGhvc3RfbmFtZQogICAgICAgIGltcG9y
            dCByYW5kb20KCiAgICAgICAgaG9zdF9pcCA9IHNlbGYuZHVtbXlfaG9zdF9pcF9tYXAuZ2V0KGhv
            c3RfbmFtZSwgIiIpCiAgICAgICAgaWYgbm90IGhvc3RfaXA6CiAgICAgICAgICAgIGhvc3RfaXAg
            PSBmIjE3Mi45LntyYW5kb20ucmFuZGludCgwLCAyNTQpfS57cmFuZG9tLnJhbmRpbnQoMCwgMjU0
            KX0iCiAgICAgICAgICAgIHNlbGYuZHVtbXlfaG9zdF9pcF9tYXBbaG9zdF9uYW1lXSA9IGhvc3Rf
            aXAKICAgICAgICBkZWwgcmFuZG9tCiAgICAgICAgcmV0dXJuIGhvc3RfaXAKCiAgICBAY2xhc3Nt
            ZXRob2QKICAgIGRlZiBSYWRvcyhjb25mZmlsZT1Ob25lKToKICAgICAgICByZXR1cm4gRHVtbXlS
            YWRvcygpCgoKY2xhc3MgUzNBdXRoKEF1dGhCYXNlKToKICAgICIiIkF0dGFjaGVzIEFXUyBBdXRo
            ZW50aWNhdGlvbiB0byB0aGUgZ2l2ZW4gUmVxdWVzdCBvYmplY3QuIiIiCgogICAgc2VydmljZV9i
            YXNlX3VybCA9ICJzMy5hbWF6b25hd3MuY29tIgoKICAgIGRlZiBfX2luaXRfXyhzZWxmLCBhY2Nl
            c3Nfa2V5LCBzZWNyZXRfa2V5LCBzZXJ2aWNlX3VybD1Ob25lKToKICAgICAgICBpZiBzZXJ2aWNl
            X3VybDoKICAgICAgICAgICAgc2VsZi5zZXJ2aWNlX2Jhc2VfdXJsID0gc2VydmljZV91cmwKICAg
            ICAgICBzZWxmLmFjY2Vzc19rZXkgPSBzdHIoYWNjZXNzX2tleSkKICAgICAgICBzZWxmLnNlY3Jl
            dF9rZXkgPSBzdHIoc2VjcmV0X2tleSkKCiAgICBkZWYgX19jYWxsX18oc2VsZiwgcik6CiAgICAg
            ICAgIyBDcmVhdGUgZGF0ZSBoZWFkZXIgaWYgaXQgaXMgbm90IGNyZWF0ZWQgeWV0LgogICAgICAg
            IGlmICJkYXRlIiBub3QgaW4gci5oZWFkZXJzIGFuZCAieC1hbXotZGF0ZSIgbm90IGluIHIuaGVh
            ZGVyczoKICAgICAgICAgICAgci5oZWFkZXJzWyJkYXRlIl0gPSBmb3JtYXRkYXRlKHRpbWV2YWw9
            Tm9uZSwgbG9jYWx0aW1lPUZhbHNlLCB1c2VnbXQ9VHJ1ZSkKICAgICAgICBzaWduYXR1cmUgPSBz
            ZWxmLmdldF9zaWduYXR1cmUocikKICAgICAgICBpZiBweTNrOgogICAgICAgICAgICBzaWduYXR1
            cmUgPSBzaWduYXR1cmUuZGVjb2RlKCJ1dGYtOCIpCiAgICAgICAgci5oZWFkZXJzWyJBdXRob3Jp
            emF0aW9uIl0gPSBmIkFXUyB7c2VsZi5hY2Nlc3Nfa2V5fTp7c2lnbmF0dXJlfSIKICAgICAgICBy
            ZXR1cm4gcgoKICAgIGRlZiBnZXRfc2lnbmF0dXJlKHNlbGYsIHIpOgogICAgICAgIGNhbm9uaWNh
            bF9zdHJpbmcgPSBzZWxmLmdldF9jYW5vbmljYWxfc3RyaW5nKHIudXJsLCByLmhlYWRlcnMsIHIu
            bWV0aG9kKQogICAgICAgIGlmIHB5M2s6CiAgICAgICAgICAgIGtleSA9IHNlbGYuc2VjcmV0X2tl
            eS5lbmNvZGUoInV0Zi04IikKICAgICAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZy5lbmNv
            ZGUoInV0Zi04IikKICAgICAgICBlbHNlOgogICAgICAgICAgICBrZXkgPSBzZWxmLnNlY3JldF9r
            ZXkKICAgICAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZwogICAgICAgIGggPSBobWFjLm5l
            dyhrZXksIG1zZywgZGlnZXN0bW9kPXNoYSkKICAgICAgICByZXR1cm4gZW5jb2Rlc3RyaW5nKGgu
            ZGlnZXN0KCkpLnN0cmlwKCkKCiAgICBkZWYgZ2V0X2Nhbm9uaWNhbF9zdHJpbmcoc2VsZiwgdXJs
            LCBoZWFkZXJzLCBtZXRob2QpOgogICAgICAgIHBhcnNlZHVybCA9IHVybHBhcnNlKHVybCkKICAg
            ICAgICBvYmplY3RrZXkgPSBwYXJzZWR1cmwucGF0aFsxOl0KCiAgICAgICAgYnVja2V0ID0gcGFy
            c2VkdXJsLm5ldGxvY1s6IC1sZW4oc2VsZi5zZXJ2aWNlX2Jhc2VfdXJsKV0KICAgICAgICBpZiBs
            ZW4oYnVja2V0KSA+IDE6CiAgICAgICAgICAgICMgcmVtb3ZlIGxhc3QgZG90CiAgICAgICAgICAg
            IGJ1Y2tldCA9IGJ1Y2tldFs6LTFdCgogICAgICAgIGludGVyZXN0aW5nX2hlYWRlcnMgPSB7ImNv
            bnRlbnQtbWQ1IjogIiIsICJjb250ZW50LXR5cGUiOiAiIiwgImRhdGUiOiAiIn0KICAgICAgICBm
            b3Iga2V5IGluIGhlYWRlcnM6CiAgICAgICAgICAgIGxrID0ga2V5Lmxvd2VyKCkKICAgICAgICAg
            ICAgdHJ5OgogICAgICAgICAgICAgICAgbGsgPSBsay5kZWNvZGUoInV0Zi04IikKICAgICAgICAg
            ICAgZXhjZXB0OgogICAgICAgICAgICAgICAgcGFzcwogICAgICAgICAgICBpZiBoZWFkZXJzW2tl
            eV0gYW5kICgKICAgICAgICAgICAgICAgIGxrIGluIGludGVyZXN0aW5nX2hlYWRlcnMua2V5cygp
            IG9yIGxrLnN0YXJ0c3dpdGgoIngtYW16LSIpCiAgICAgICAgICAgICk6CiAgICAgICAgICAgICAg
            ICBpbnRlcmVzdGluZ19oZWFkZXJzW2xrXSA9IGhlYWRlcnNba2V5XS5zdHJpcCgpCgogICAgICAg
            ICMgSWYgeC1hbXotZGF0ZSBpcyB1c2VkIGl0IHN1cGVyc2VkZXMgdGhlIGRhdGUgaGVhZGVyLgog
            ICAgICAgIGlmIG5vdCBweTNrOgogICAgICAgICAgICBpZiAieC1hbXotZGF0ZSIgaW4gaW50ZXJl
            c3RpbmdfaGVhZGVyczoKICAgICAgICAgICAgICAgIGludGVyZXN0aW5nX2hlYWRlcnNbImRhdGUi
            XSA9ICIiCiAgICAgICAgZWxzZToKICAgICAgICAgICAgaWYgIngtYW16LWRhdGUiIGluIGludGVy
            ZXN0aW5nX2hlYWRlcnM6CiAgICAgICAgICAgICAgICBpbnRlcmVzdGluZ19oZWFkZXJzWyJkYXRl
            Il0gPSAiIgoKICAgICAgICBidWYgPSBmInttZXRob2R9XG4iCiAgICAgICAgZm9yIGtleSBpbiBz
            b3J0ZWQoaW50ZXJlc3RpbmdfaGVhZGVycy5rZXlzKCkpOgogICAgICAgICAgICB2YWwgPSBpbnRl
            cmVzdGluZ19oZWFkZXJzW2tleV0KICAgICAgICAgICAgaWYga2V5LnN0YXJ0c3dpdGgoIngtYW16
            LSIpOgogICAgICAgICAgICAgICAgYnVmICs9IGYie2tleX06e3ZhbH1cbiIKICAgICAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgICAgIGJ1ZiArPSBmInt2YWx9XG4iCgogICAgICAgICMgYXBwZW5k
            IHRoZSBidWNrZXQgaWYgaXQgZXhpc3RzCiAgICAgICAgaWYgYnVja2V0ICE9ICIiOgogICAgICAg
            ICAgICBidWYgKz0gZiIve2J1Y2tldH0iCgogICAgICAgICMgYWRkIHRoZSBvYmplY3RrZXkuIGV2
            ZW4gaWYgaXQgZG9lc24ndCBleGlzdCwgYWRkIHRoZSBzbGFzaAogICAgICAgIGJ1ZiArPSBmIi97
            b2JqZWN0a2V5fSIKCiAgICAgICAgcmV0dXJuIGJ1ZgoKCmNsYXNzIFJhZG9zSlNPTjoKICAgIEVY
            VEVSTkFMX1VTRVJfTkFNRSA9ICJjbGllbnQuaGVhbHRoY2hlY2tlciIKICAgIEVYVEVSTkFMX1JH
            V19BRE1JTl9PUFNfVVNFUl9OQU1FID0gInJndy1hZG1pbi1vcHMtdXNlciIKICAgIEVNUFRZX09V
            VFBVVF9MSVNUID0gIkVtcHR5IG91dHB1dCBsaXN0IgogICAgREVGQVVMVF9SR1dfUE9PTF9QUkVG
            SVggPSAiZGVmYXVsdCIKICAgIERFRkFVTFRfTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUID0gIjky
            ODMiCgogICAgQGNsYXNzbWV0aG9kCiAgICBkZWYgZ2VuX2FyZ19wYXJzZXIoY2xzLCBhcmdzX3Rv
            X3BhcnNlPU5vbmUpOgogICAgICAgIGFyZ1AgPSBhcmdwYXJzZS5Bcmd1bWVudFBhcnNlcigpCgog
            ICAgICAgIGNvbW1vbl9ncm91cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJjb21tb24iKQog
            ICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoIi0tdmVyYm9zZSIsICItdiIsIGFjdGlv
            bj0ic3RvcmVfdHJ1ZSIsIGRlZmF1bHQ9RmFsc2UpCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9h
            cmd1bWVudCgKICAgICAgICAgICAgIi0tY2VwaC1jb25mIiwgIi1jIiwgaGVscD0iUHJvdmlkZSBh
            IGNlcGggY29uZiBmaWxlLiIsIHR5cGU9c3RyCiAgICAgICAgKQogICAgICAgIGNvbW1vbl9ncm91
            cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLWtleXJpbmciLCAiLWsiLCBoZWxwPSJQYXRo
            IHRvIGNlcGgga2V5cmluZyBmaWxlLiIsIHR5cGU9c3RyCiAgICAgICAgKQogICAgICAgIGNvbW1v
            bl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJ1bi1hcy11c2VyIiwKICAgICAg
            ICAgICAgIi11IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgdHlwZT1zdHIs
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIGEgdXNlciBuYW1lIHRvIGNoZWNrIHRoZSBjbHVz
            dGVyJ3MgaGVhbHRoIHN0YXR1cywgbXVzdCBiZSBwcmVmaXhlZCBieSAnY2xpZW50LiciLAogICAg
            ICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1j
            bHVzdGVyLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJL
            dWJlcm5ldGVzIGNsdXN0ZXIgbmFtZShsZWdhY3kgZmxhZyksIE5vdGU6IEVpdGhlciB1c2UgdGhp
            cyBvciAtLWs4cy1jbHVzdGVyLW5hbWUiLAogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAu
            YWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1rOHMtY2x1c3Rlci1uYW1lIiwgZGVmYXVsdD0i
            IiwgaGVscD0iS3ViZXJuZXRlcyBjbHVzdGVyIG5hbWUiCiAgICAgICAgKQogICAgICAgIGNvbW1v
            bl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLW5hbWVzcGFjZSIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIGhlbHA9Ik5hbWVzcGFjZSB3aGVyZSBDZXBoQ2x1
            c3RlciBpcyBydW5uaW5nIiwKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tcmd3LXBvb2wtcHJlZml4IiwgZGVmYXVsdD0iIiwgaGVscD0i
            UkdXIFBvb2wgcHJlZml4IgogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3Vt
            ZW50KAogICAgICAgICAgICAiLS1yZXN0cmljdGVkLWF1dGgtcGVybWlzc2lvbiIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlJlc3RyaWN0IGNlcGhDU0lLZXly
            aW5ncyBhdXRoIHBlcm1pc3Npb25zIHRvIHNwZWNpZmljIHBvb2xzLCBjbHVzdGVyLiIKICAgICAg
            ICAgICAgKyAiTWFuZGF0b3J5IGZsYWdzIHRoYXQgbmVlZCB0byBiZSBzZXQgYXJlIC0tcmJkLWRh
            dGEtcG9vbC1uYW1lLCBhbmQgLS1rOHMtY2x1c3Rlci1uYW1lLiIKICAgICAgICAgICAgKyAiLS1j
            ZXBoZnMtZmlsZXN5c3RlbS1uYW1lIGZsYWcgY2FuIGFsc28gYmUgcGFzc2VkIGluIGNhc2Ugb2Yg
            Y2VwaGZzIHVzZXIgcmVzdHJpY3Rpb24sIHNvIGl0IGNhbiByZXN0cmljdCB1c2VyIHRvIHBhcnRp
            Y3VsYXIgY2VwaGZzIGZpbGVzeXN0ZW0iCiAgICAgICAgICAgICsgInNhbXBsZSBydW46IGBweXRo
            b24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwtY2x1c3Rlci1yZXNvdXJjZXMucHkgLS1jZXBo
            ZnMtZmlsZXN5c3RlbS1uYW1lIG15ZnMgLS1yYmQtZGF0YS1wb29sLW5hbWUgcmVwbGljYXBvb2wg
            LS1rOHMtY2x1c3Rlci1uYW1lIHJvb2tzdG9yYWdlIC0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Np
            b24gdHJ1ZWAiCiAgICAgICAgICAgICsgIk5vdGU6IFJlc3RyaWN0aW5nIHRoZSBjc2ktdXNlcnMg
            cGVyIHBvb2wsIGFuZCBwZXIgY2x1c3RlciB3aWxsIHJlcXVpcmUgY3JlYXRpbmcgbmV3IGNzaS11
            c2VycyBhbmQgbmV3IHNlY3JldHMgZm9yIHRoYXQgY3NpLXVzZXJzLiIKICAgICAgICAgICAgKyAi
            U28gYXBwbHkgdGhlc2Ugc2VjcmV0cyBvbmx5IHRvIG5ldyBgQ29uc3VtZXIgY2x1c3RlcmAgZGVw
            bG95bWVudCB3aGlsZSB1c2luZyB0aGUgc2FtZSBgU291cmNlIGNsdXN0ZXJgLiIsCiAgICAgICAg
            KQogICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXYyLXBv
            cnQtZW5hYmxlIiwKICAgICAgICAgICAgYWN0aW9uPSJzdG9yZV90cnVlIiwKICAgICAgICAgICAg
            ZGVmYXVsdD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iRW5hYmxlIHYyIG1vbiBwb3J0KDMzMDAp
            IGZvciBtb25zIiwKICAgICAgICApCgogICAgICAgIG91dHB1dF9ncm91cCA9IGFyZ1AuYWRkX2Fy
            Z3VtZW50X2dyb3VwKCJvdXRwdXQiKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQo
            CiAgICAgICAgICAgICItLWZvcm1hdCIsCiAgICAgICAgICAgICItdCIsCiAgICAgICAgICAgIGNo
            b2ljZXM9WyJqc29uIiwgImJhc2giXSwKICAgICAgICAgICAgZGVmYXVsdD0ianNvbiIsCiAgICAg
            ICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBvdXRwdXQgZm9ybWF0IChqc29uIHwgYmFzaCkiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1vdXRwdXQiLAogICAgICAgICAgICAiLW8iLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAg
            ICAgICAgICBoZWxwPSJPdXRwdXQgd2lsbCBiZSBzdG9yZWQgaW50byB0aGUgcHJvdmlkZWQgZmls
            ZSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAog
            ICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgQ2VwaCBmaWxlc3lzdGVt
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tY2VwaGZzLW1ldGFkYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIs
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBjZXBoZnMgbWV0YWRh
            dGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAg
            ICAgICAgICAgICItLWNlcGhmcy1kYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9
            IiIsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBjZXBoZnMgZGF0
            YSBwb29sIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAg
            ICAgICAgICAgIi0tcmJkLWRhdGEtcG9vbC1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwK
            ICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRo
            ZSBuYW1lIG9mIHRoZSBSQkQgZGF0YXBvb2wiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3Jv
            dXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1hbGlhcy1yYmQtZGF0YS1wb29sLW5hbWUi
            LAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAg
            ICAgICAgICAgaGVscD0iUHJvdmlkZXMgYW4gYWxpYXMgZm9yIHRoZSAgUkJEIGRhdGEgcG9vbCBu
            YW1lLCBuZWNlc3NhcnkgaWYgYSBzcGVjaWFsIGNoYXJhY3RlciBpcyBwcmVzZW50IGluIHRoZSBw
            b29sIG5hbWUgc3VjaCBhcyBhIHBlcmlvZCBvciB1bmRlcnNjb3JlIiwKICAgICAgICApCiAgICAg
            ICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmd3LWVuZHBvaW50
            IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAg
            ICAgICAgICAgIGhlbHA9IlJBRE9TIEdhdGV3YXkgZW5kcG9pbnQgKGluIGA8SVB2ND46PFBPUlQ+
            YCBvciBgPFtJUHY2XT46PFBPUlQ+YCBvciBgPEZRRE4+OjxQT1JUPmAgZm9ybWF0KSIsCiAgICAg
            ICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJn
            dy10bHMtY2VydC1wYXRoIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVx
            dWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlJBRE9TIEdhdGV3YXkgZW5kcG9pbnQgVExT
            IGNlcnRpZmljYXRlIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVu
            dCgKICAgICAgICAgICAgIi0tcmd3LXNraXAtdGxzIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFs
            c2UsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9Iklnbm9yZSBU
            TFMgY2VydGlmaWNhdGlvbiB2YWxpZGF0aW9uIHdoZW4gYSBzZWxmLXNpZ25lZCBjZXJ0aWZpY2F0
            ZSBpcyBwcm92aWRlZCAoTk9UIFJFQ09NTUVOREVEIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tbW9uaXRvcmluZy1lbmRwb2ludCIs
            CiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAg
            ICAgICAgICBoZWxwPSJDZXBoIE1hbmFnZXIgcHJvbWV0aGV1cyBleHBvcnRlciBlbmRwb2ludHMg
            KGNvbW1hIHNlcGFyYXRlZCBsaXN0IG9mIChmb3JtYXQgYDxJUHY0PmAgb3IgYDxbSVB2Nl0+YCBv
            ciBgPEZRRE4+YCkgZW50cmllcyBvZiBhY3RpdmUgYW5kIHN0YW5kYnkgbWdycykiLAogICAgICAg
            ICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1tb25p
            dG9yaW5nLWVuZHBvaW50LXBvcnQiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAg
            ICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iQ2VwaCBNYW5hZ2VyIHByb21ldGhl
            dXMgZXhwb3J0ZXIgcG9ydCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJn
            dW1lbnQoCiAgICAgICAgICAgICItLXNraXAtbW9uaXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAg
            ICAgICAgIGhlbHA9IkRvIG5vdCBjaGVjayBmb3IgYSBtb25pdG9yaW5nIGVuZHBvaW50IGZvciB0
            aGUgQ2VwaCBjbHVzdGVyIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tcmJkLW1ldGFkYXRhLWVjLXBvb2wtbmFtZSIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBo
            ZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiBlcmFzdXJlIGNvZGVkIFJCRCBtZXRhZGF0YSBwb29s
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tZHJ5LXJ1biIsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFj
            dGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGhlbHA9IkRyeSBydW4gcHJpbnRzIHRoZSBl
            eGVjdXRlZCBjb21tYW5kcyB3aXRob3V0IHJ1bm5pbmcgdGhlbSIsCiAgICAgICAgKQogICAgICAg
            IG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJhZG9zLW5hbWVzcGFj
            ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAog
            ICAgICAgICAgICBoZWxwPSJEaXZpZGVzIGEgcG9vbCBpbnRvIHNlcGFyYXRlIGxvZ2ljYWwgbmFt
            ZXNwYWNlcywgdXNlZCBmb3IgY3JlYXRpbmcgUkJEIFBWQyBpbiBhIENlcGhCbG9ja1Bvb2xSYWRv
            c05hbWVzcGFjZSAoc2hvdWxkIGJlIGxvd2VyIGNhc2UpIiwKICAgICAgICApCiAgICAgICAgb3V0
            cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tc3Vidm9sdW1lLWdyb3VwIiwK
            ICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAg
            ICAgICAgIGhlbHA9InByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBzdWJ2b2x1bWUgZ3JvdXAiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1yZ3ctcmVhbG0tbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJl
            cXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUg
            cmd3LXJlYWxtIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgK
            ICAgICAgICAgICAgIi0tcmd3LXpvbmUtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAg
            ICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUg
            bmFtZSBvZiB0aGUgcmd3LXpvbmUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS1yZ3ctem9uZWdyb3VwLW5hbWUiLAogICAgICAgICAg
            ICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVs
            cD0icHJvdmlkZXMgdGhlIG5hbWUgb2YgdGhlIHJndy16b25lZ3JvdXAiLAogICAgICAgICkKICAg
            ICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS10b3BvbG9neS1w
            b29scyIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNl
            LAogICAgICAgICAgICBoZWxwPSJjb21tYS1zZXBhcmF0ZWQgbGlzdCBvZiB0b3BvbG9neS1jb25z
            dHJhaW5lZCByYmQgcG9vbHMiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2Fy
            Z3VtZW50KAogICAgICAgICAgICAiLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi1sYWJlbCIsCiAg
            ICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAg
            ICAgICBoZWxwPSJrOHMgY2x1c3RlciBmYWlsdXJlIGRvbWFpbiBsYWJlbCAoZXhhbXBsZTogem9u
            ZSwgcmFjaywgb3IgaG9zdCkgZm9yIHRoZSB0b3BvbG9neS1wb29scyB0aGF0IG1hdGNoIHRoZSBj
            ZXBoIGRvbWFpbiIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQo
            CiAgICAgICAgICAgICItLXRvcG9sb2d5LWZhaWx1cmUtZG9tYWluLXZhbHVlcyIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBo
            ZWxwPSJjb21tYS1zZXBhcmF0ZWQgbGlzdCBvZiB0aGUgazhzIGNsdXN0ZXIgZmFpbHVyZSBkb21h
            aW4gdmFsdWVzIGNvcnJlc3BvbmRpbmcgdG8gZWFjaCBvZiB0aGUgcG9vbHMgaW4gdGhlIGB0b3Bv
            bG9neS1wb29sc2AgbGlzdCIsCiAgICAgICAgKQoKICAgICAgICB1cGdyYWRlX2dyb3VwID0gYXJn
            UC5hZGRfYXJndW1lbnRfZ3JvdXAoInVwZ3JhZGUiKQogICAgICAgIHVwZ3JhZGVfZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS11cGdyYWRlIiwKICAgICAgICAgICAgYWN0aW9uPSJz
            dG9yZV90cnVlIiwKICAgICAgICAgICAgZGVmYXVsdD1GYWxzZSwKICAgICAgICAgICAgaGVscD0i
            VXBncmFkZXMgdGhlIGNlcGhDU0lLZXlyaW5ncyhGb3IgZXhhbXBsZTogY2xpZW50LmNzaS1jZXBo
            ZnMtcHJvdmlzaW9uZXIpIGFuZCBjbGllbnQuaGVhbHRoY2hlY2tlciBjZXBoIHVzZXJzIHdpdGgg
            bmV3IHBlcm1pc3Npb25zIG5lZWRlZCBmb3IgdGhlIG5ldyBjbHVzdGVyIHZlcnNpb24gYW5kIG9s
            ZGVyIHBlcm1pc3Npb24gd2lsbCBzdGlsbCBiZSBhcHBsaWVkLiIKICAgICAgICAgICAgKyAiU2Ft
            cGxlIHJ1bjogYHB5dGhvbjMgL2V0Yy9jZXBoL2NyZWF0ZS1leHRlcm5hbC1jbHVzdGVyLXJlc291
            cmNlcy5weSAtLXVwZ3JhZGVgLCB0aGlzIHdpbGwgdXBncmFkZSBhbGwgdGhlIGRlZmF1bHQgY3Np
            IHVzZXJzKG5vbi1yZXN0cmljdGVkKSIKICAgICAgICAgICAgKyAiRm9yIHJlc3RyaWN0ZWQgdXNl
            cnMoRm9yIGV4YW1wbGU6IGNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1z
            dG9yYWdlLW15ZnMpLCB1c2VycyBjcmVhdGVkIHVzaW5nIC0tcmVzdHJpY3RlZC1hdXRoLXBlcm1p
            c3Npb24gZmxhZyBuZWVkIHRvIHBhc3MgbWFuZGF0b3J5IGZsYWdzIgogICAgICAgICAgICArICJt
            YW5kYXRvcnkgZmxhZ3M6ICctLXJiZC1kYXRhLXBvb2wtbmFtZSwgLS1rOHMtY2x1c3Rlci1uYW1l
            IGFuZCAtLXJ1bi1hcy11c2VyJyBmbGFncyB3aGlsZSB1cGdyYWRpbmciCiAgICAgICAgICAgICsg
            ImluIGNhc2Ugb2YgY2VwaGZzIHVzZXJzIGlmIHlvdSBoYXZlIHBhc3NlZCAtLWNlcGhmcy1maWxl
            c3lzdGVtLW5hbWUgZmxhZyB3aGlsZSBjcmVhdGluZyB1c2VyIHRoZW4gd2hpbGUgdXBncmFkaW5n
            IGl0IHdpbGwgYmUgbWFuZGF0b3J5IHRvbyIKICAgICAgICAgICAgKyAiU2FtcGxlIHJ1bjogYHB5
            dGhvbjMgL2V0Yy9jZXBoL2NyZWF0ZS1leHRlcm5hbC1jbHVzdGVyLXJlc291cmNlcy5weSAtLXVw
            Z3JhZGUgLS1yYmQtZGF0YS1wb29sLW5hbWUgcmVwbGljYXBvb2wgLS1rOHMtY2x1c3Rlci1uYW1l
            IHJvb2tzdG9yYWdlICAtLXJ1bi1hcy11c2VyIGNsaWVudC5jc2ktcmJkLW5vZGUtcm9va3N0b3Jh
            Z2UtcmVwbGljYXBvb2xgIgogICAgICAgICAgICArICJQUzogQW4gZXhpc3Rpbmcgbm9uLXJlc3Ry
            aWN0ZWQgdXNlciBjYW5ub3QgYmUgY29udmVydGVkIHRvIGEgcmVzdHJpY3RlZCB1c2VyIGJ5IHVw
            Z3JhZGluZy4iCiAgICAgICAgICAgICsgIlVwZ3JhZGUgZmxhZyBzaG91bGQgb25seSBiZSB1c2Vk
            IHRvIGFwcGVuZCBuZXcgcGVybWlzc2lvbnMgdG8gdXNlcnMsIGl0IHNob3VsZG4ndCBiZSB1c2Vk
            IGZvciBjaGFuZ2luZyB1c2VyIGFscmVhZHkgYXBwbGllZCBwZXJtaXNzaW9uLCBmb3IgZXhhbXBs
            ZSB5b3Ugc2hvdWxkbid0IGNoYW5nZSBpbiB3aGljaCBwb29sIHVzZXIgaGFzIGFjY2VzcyIsCiAg
            ICAgICAgKQoKICAgICAgICBpZiBhcmdzX3RvX3BhcnNlOgogICAgICAgICAgICBhc3NlcnQgKAog
            ICAgICAgICAgICAgICAgdHlwZShhcmdzX3RvX3BhcnNlKSA9PSBsaXN0CiAgICAgICAgICAgICks
            ICJBcmd1bWVudCB0byAnZ2VuX2FyZ19wYXJzZXInIHNob3VsZCBiZSBhIGxpc3QiCiAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgYXJnc190b19wYXJzZSA9IHN5cy5hcmd2WzE6XQogICAgICAgIHJl
            dHVybiBhcmdQLnBhcnNlX2FyZ3MoYXJnc190b19wYXJzZSkKCiAgICBkZWYgdmFsaWRhdGVfcmJk
            X21ldGFkYXRhX2VjX3Bvb2xfbmFtZShzZWxmKToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2Vy
            LnJiZF9tZXRhZGF0YV9lY19wb29sX25hbWU6CiAgICAgICAgICAgIHJiZF9tZXRhZGF0YV9lY19w
            b29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUKICAg
            ICAgICAgICAgcmJkX3Bvb2xfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9u
            YW1lCgogICAgICAgICAgICBpZiByYmRfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiRmxh
            ZyAnLS1yYmQtZGF0YS1wb29sLW5hbWUnIHNob3VsZCBub3QgYmUgZW1wdHkiCiAgICAgICAgICAg
            ICAgICApCgogICAgICAgICAgICBpZiByYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lID09ICIiOgog
            ICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAg
            ICAgICAgICAgICAiRmxhZyAnLS1yYmQtbWV0YWRhdGEtZWMtcG9vbC1uYW1lJyBzaG91bGQgbm90
            IGJlIGVtcHR5IgogICAgICAgICAgICAgICAgKQoKICAgICAgICAgICAgY21kX2pzb24gPSB7InBy
            ZWZpeCI6ICJvc2QgZHVtcCIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgICAgIHJldF92YWws
            IGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikK
            ICAgICAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAg
            ICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAg
            ICAgICAgZiJ7Y21kX2pzb25bJ3ByZWZpeCddfSBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAg
            ICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVN
            UFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgbWV0YWRhdGFf
            cG9vbF9leGlzdCwgcG9vbF9leGlzdCA9IEZhbHNlLCBGYWxzZQoKICAgICAgICAgICAgZm9yIGtl
            eSBpbiBqc29uX291dFsicG9vbHMiXToKICAgICAgICAgICAgICAgICMgaWYgZXJhc3VyZV9jb2Rl
            X3Byb2ZpbGUgaXMgZW1wdHkgYW5kIHBvb2wgbmFtZSBleGlzdHMgdGhlbiBpdCByZXBsaWNhIHBv
            b2wKICAgICAgICAgICAgICAgIGlmICgKICAgICAgICAgICAgICAgICAgICBrZXlbImVyYXN1cmVf
            Y29kZV9wcm9maWxlIl0gPT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQga2V5WyJwb29sX25h
            bWUiXSA9PSByYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lCiAgICAgICAgICAgICAgICApOgogICAg
            ICAgICAgICAgICAgICAgIG1ldGFkYXRhX3Bvb2xfZXhpc3QgPSBUcnVlCiAgICAgICAgICAgICAg
            ICAjIGlmIGVyYXN1cmVfY29kZV9wcm9maWxlIGlzIG5vdCBlbXB0eSBhbmQgcG9vbCBuYW1lIGV4
            aXN0cyB0aGVuIGl0IGlzIGVjIHBvb2wKICAgICAgICAgICAgICAgIGlmIGtleVsiZXJhc3VyZV9j
            b2RlX3Byb2ZpbGUiXSBhbmQga2V5WyJwb29sX25hbWUiXSA9PSByYmRfcG9vbF9uYW1lOgogICAg
            ICAgICAgICAgICAgICAgIHBvb2xfZXhpc3QgPSBUcnVlCgogICAgICAgICAgICBpZiBub3QgbWV0
            YWRhdGFfcG9vbF9leGlzdDoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVF
            eGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIlByb3ZpZGVkIHJiZF9lY19tZXRhZGF0YV9w
            b29sIG5hbWUsIgogICAgICAgICAgICAgICAgICAgIGYiIHtyYmRfbWV0YWRhdGFfZWNfcG9vbF9u
            YW1lfSwgZG9lcyBub3QgZXhpc3QiCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmIG5v
            dCBwb29sX2V4aXN0OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICBmIlByb3ZpZGVkIHJiZF9kYXRhX3Bvb2wgbmFtZSwg
            e3JiZF9wb29sX25hbWV9LCBkb2VzIG5vdCBleGlzdCIKICAgICAgICAgICAgICAgICkKICAgICAg
            ICAgICAgcmV0dXJuIHJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUKCiAgICBkZWYgZHJ5X3J1bihz
            ZWxmLCBtc2cpOgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAg
            ICAgcHJpbnQoIkV4ZWN1dGU6ICIgKyAiJyIgKyBtc2cgKyAiJyIpCgogICAgZGVmIHZhbGlkYXRl
            X3Jnd19lbmRwb2ludF90bHNfY2VydChzZWxmKToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd190bHNfY2VydF9wYXRoOgogICAgICAgICAgICB3aXRoIG9wZW4oc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfdGxzX2NlcnRfcGF0aCwgZW5jb2Rpbmc9InV0ZjgiKSBhcyBmOgogICAgICAgICAgICAg
            ICAgY29udGVudHMgPSBmLnJlYWQoKQogICAgICAgICAgICAgICAgcmV0dXJuIGNvbnRlbnRzLnJz
            dHJpcCgpCgogICAgZGVmIF9jaGVja19jb25mbGljdGluZ19vcHRpb25zKHNlbGYpOgogICAgICAg
            IGlmIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnVwZ3JhZGUgYW5kIG5vdCBzZWxmLl9hcmdfcGFyc2Vy
            LnJiZF9kYXRhX3Bvb2xfbmFtZToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICJFaXRoZXIgJy0tdXBncmFkZScgb3IgJy0tcmJkLWRh
            dGEtcG9vbC1uYW1lIDxwb29sX25hbWU+JyBzaG91bGQgYmUgc3BlY2lmaWVkIgogICAgICAgICAg
            ICApCgogICAgZGVmIF9pbnZhbGlkX2VuZHBvaW50KHNlbGYsIGVuZHBvaW50X3N0cik6CiAgICAg
            ICAgIyBleHRyYWN0IHRoZSBwb3J0IGJ5IGdldHRpbmcgdGhlIGxhc3Qgc3BsaXQgb24gYDpgIGRl
            bGltaXRlcgogICAgICAgIHRyeToKICAgICAgICAgICAgZW5kcG9pbnRfc3RyX2lwLCBwb3J0ID0g
            ZW5kcG9pbnRfc3RyLnJzcGxpdCgiOiIsIDEpCiAgICAgICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAg
            ICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oZiJOb3QgYSBwcm9wZXIg
            ZW5kcG9pbnQ6IHtlbmRwb2ludF9zdHJ9IikKCiAgICAgICAgdHJ5OgogICAgICAgICAgICBpZiBl
            bmRwb2ludF9zdHJfaXBbMF0gPT0gIlsiOgogICAgICAgICAgICAgICAgZW5kcG9pbnRfc3RyX2lw
            ID0gZW5kcG9pbnRfc3RyX2lwWzEgOiBsZW4oZW5kcG9pbnRfc3RyX2lwKSAtIDFdCiAgICAgICAg
            ICAgIGlwX3R5cGUgPSAoCiAgICAgICAgICAgICAgICAiSVB2NCIgaWYgdHlwZShpcF9hZGRyZXNz
            KGVuZHBvaW50X3N0cl9pcCkpIGlzIElQdjRBZGRyZXNzIGVsc2UgIklQdjYiCiAgICAgICAgICAg
            ICkKICAgICAgICBleGNlcHQgVmFsdWVFcnJvcjoKICAgICAgICAgICAgaXBfdHlwZSA9ICJGUURO
            IgogICAgICAgIGlmIG5vdCBwb3J0LmlzZGlnaXQoKToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0
            aW9uRmFpbHVyZUV4Y2VwdGlvbihmIlBvcnQgbm90IHZhbGlkOiB7cG9ydH0iKQogICAgICAgIGlu
            dFBvcnQgPSBpbnQocG9ydCkKICAgICAgICBpZiBpbnRQb3J0IDwgMSBvciBpbnRQb3J0ID4gMioq
            MTYgLSAxOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKGYiT3V0
            IG9mIHJhbmdlIHBvcnQgbnVtYmVyOiB7cG9ydH0iKQoKICAgICAgICByZXR1cm4gaXBfdHlwZQoK
            ICAgIGRlZiBlbmRwb2ludF9kaWFsKHNlbGYsIGVuZHBvaW50X3N0ciwgaXBfdHlwZSwgdGltZW91
            dD0zLCBjZXJ0PU5vbmUpOgogICAgICAgICMgaWYgdGhlICdjbHVzdGVyJyBpbnN0YW5jZSBpcyBh
            IGR1bW15IG9uZSwKICAgICAgICAjIGRvbid0IHRyeSB0byByZWFjaCBvdXQgdG8gdGhlIGVuZHBv
            aW50CiAgICAgICAgaWYgaXNpbnN0YW5jZShzZWxmLmNsdXN0ZXIsIER1bW15UmFkb3MpOgogICAg
            ICAgICAgICByZXR1cm4gIiIsICIiLCAiIgogICAgICAgIGlmIGlwX3R5cGUgPT0gIklQdjYiOgog
            ICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBlbmRwb2ludF9zdHJfaXAsIGVuZHBvaW50
            X3N0cl9wb3J0ID0gZW5kcG9pbnRfc3RyLnJzcGxpdCgiOiIsIDEpCiAgICAgICAgICAgIGV4Y2Vw
            dCBWYWx1ZUVycm9yOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICBmIk5vdCBhIHByb3BlciBlbmRwb2ludDoge2VuZHBv
            aW50X3N0cn0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmIGVuZHBvaW50X3N0cl9p
            cFswXSAhPSAiWyI6CiAgICAgICAgICAgICAgICBlbmRwb2ludF9zdHJfaXAgPSAiWyIgKyBlbmRw
            b2ludF9zdHJfaXAgKyAiXSIKICAgICAgICAgICAgZW5kcG9pbnRfc3RyID0gIjoiLmpvaW4oW2Vu
            ZHBvaW50X3N0cl9pcCwgZW5kcG9pbnRfc3RyX3BvcnRdKQoKICAgICAgICBwcm90b2NvbHMgPSBb
            Imh0dHAiLCAiaHR0cHMiXQogICAgICAgIHJlc3BvbnNlX2Vycm9yID0gTm9uZQogICAgICAgIGZv
            ciBwcmVmaXggaW4gcHJvdG9jb2xzOgogICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBl
            cCA9IGYie3ByZWZpeH06Ly97ZW5kcG9pbnRfc3RyfSIKICAgICAgICAgICAgICAgIHZlcmlmeSA9
            IE5vbmUKICAgICAgICAgICAgICAgICMgSWYgdmVyaWZ5IGlzIHNldCB0byBhIHBhdGggdG8gYSBk
            aXJlY3RvcnksCiAgICAgICAgICAgICAgICAjIHRoZSBkaXJlY3RvcnkgbXVzdCBoYXZlIGJlZW4g
            cHJvY2Vzc2VkIHVzaW5nIHRoZSBjX3JlaGFzaCB1dGlsaXR5IHN1cHBsaWVkIHdpdGggT3BlblNT
            TC4KICAgICAgICAgICAgICAgIGlmIHByZWZpeCA9PSAiaHR0cHMiIGFuZCBzZWxmLl9hcmdfcGFy
            c2VyLnJnd19za2lwX3RsczoKICAgICAgICAgICAgICAgICAgICB2ZXJpZnkgPSBGYWxzZQogICAg
            ICAgICAgICAgICAgICAgIHIgPSByZXF1ZXN0cy5oZWFkKGVwLCB0aW1lb3V0PXRpbWVvdXQsIHZl
            cmlmeT1GYWxzZSkKICAgICAgICAgICAgICAgIGVsaWYgcHJlZml4ID09ICJodHRwcyIgYW5kIGNl
            cnQ6CiAgICAgICAgICAgICAgICAgICAgdmVyaWZ5ID0gY2VydAogICAgICAgICAgICAgICAgICAg
            IHIgPSByZXF1ZXN0cy5oZWFkKGVwLCB0aW1lb3V0PXRpbWVvdXQsIHZlcmlmeT1jZXJ0KQogICAg
            ICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgICAgICByID0gcmVxdWVzdHMuaGVhZChl
            cCwgdGltZW91dD10aW1lb3V0KQogICAgICAgICAgICAgICAgaWYgci5zdGF0dXNfY29kZSA9PSAy
            MDA6CiAgICAgICAgICAgICAgICAgICAgcmV0dXJuIHByZWZpeCwgdmVyaWZ5LCAiIgogICAgICAg
            ICAgICBleGNlcHQgRXhjZXB0aW9uIGFzIGVycjoKICAgICAgICAgICAgICAgIHJlc3BvbnNlX2Vy
            cm9yID0gZXJyCiAgICAgICAgICAgICAgICBjb250aW51ZQogICAgICAgIHN5cy5zdGRlcnIud3Jp
            dGUoCiAgICAgICAgICAgIGYidW5hYmxlIHRvIGNvbm5lY3QgdG8gZW5kcG9pbnQ6IHtlbmRwb2lu
            dF9zdHJ9LCBmYWlsZWQgZXJyb3I6IHtyZXNwb25zZV9lcnJvcn0iCiAgICAgICAgKQogICAgICAg
            IHJldHVybiAoCiAgICAgICAgICAgICIiLAogICAgICAgICAgICAiIiwKICAgICAgICAgICAgKCIt
            MSIpLAogICAgICAgICkKCiAgICBkZWYgX19pbml0X18oc2VsZiwgYXJnX2xpc3Q9Tm9uZSk6CiAg
            ICAgICAgc2VsZi5vdXRfbWFwID0ge30KICAgICAgICBzZWxmLl9leGNsdWRlZF9rZXlzID0gc2V0
            KCkKICAgICAgICBzZWxmLl9hcmdfcGFyc2VyID0gc2VsZi5nZW5fYXJnX3BhcnNlcihhcmdzX3Rv
            X3BhcnNlPWFyZ19saXN0KQogICAgICAgIHNlbGYuX2NoZWNrX2NvbmZsaWN0aW5nX29wdGlvbnMo
            KQogICAgICAgIHNlbGYucnVuX2FzX3VzZXIgPSBzZWxmLl9hcmdfcGFyc2VyLnJ1bl9hc191c2Vy
            CiAgICAgICAgc2VsZi5vdXRwdXRfZmlsZSA9IHNlbGYuX2FyZ19wYXJzZXIub3V0cHV0CiAgICAg
            ICAgc2VsZi5jZXBoX2NvbmYgPSBzZWxmLl9hcmdfcGFyc2VyLmNlcGhfY29uZgogICAgICAgIHNl
            bGYuY2VwaF9rZXlyaW5nID0gc2VsZi5fYXJnX3BhcnNlci5rZXlyaW5nCiAgICAgICAgIyBpZiB1
            c2VyIG5vdCBwcm92aWRlZCwgZ2l2ZSBhIGRlZmF1bHQgdXNlcgogICAgICAgIGlmIG5vdCBzZWxm
            LnJ1bl9hc191c2VyIGFuZCBub3Qgc2VsZi5fYXJnX3BhcnNlci51cGdyYWRlOgogICAgICAgICAg
            ICBzZWxmLnJ1bl9hc191c2VyID0gc2VsZi5FWFRFUk5BTF9VU0VSX05BTUUKICAgICAgICBpZiBu
            b3Qgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXggYW5kIG5vdCBzZWxmLl9hcmdfcGFy
            c2VyLnVwZ3JhZGU6CiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4
            ID0gc2VsZi5ERUZBVUxUX1JHV19QT09MX1BSRUZJWAogICAgICAgIGlmIHNlbGYuY2VwaF9jb25m
            OgogICAgICAgICAgICBrd2FyZ3MgPSB7fQogICAgICAgICAgICBpZiBzZWxmLmNlcGhfa2V5cmlu
            ZzoKICAgICAgICAgICAgICAgIGt3YXJnc1siY29uZiJdID0geyJrZXlyaW5nIjogc2VsZi5jZXBo
            X2tleXJpbmd9CiAgICAgICAgICAgIHNlbGYuY2x1c3RlciA9IHJhZG9zLlJhZG9zKGNvbmZmaWxl
            PXNlbGYuY2VwaF9jb25mLCAqKmt3YXJncykKICAgICAgICBlbHNlOgogICAgICAgICAgICBzZWxm
            LmNsdXN0ZXIgPSByYWRvcy5SYWRvcygpCiAgICAgICAgICAgIHNlbGYuY2x1c3Rlci5jb25mX3Jl
            YWRfZmlsZSgpCiAgICAgICAgc2VsZi5jbHVzdGVyLmNvbm5lY3QoKQoKICAgIGRlZiBzaHV0ZG93
            bihzZWxmKToKICAgICAgICBpZiBzZWxmLmNsdXN0ZXIuc3RhdGUgPT0gImNvbm5lY3RlZCI6CiAg
            ICAgICAgICAgIHNlbGYuY2x1c3Rlci5zaHV0ZG93bigpCgogICAgZGVmIGdldF9mc2lkKHNlbGYp
            OgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJu
            IHNlbGYuZHJ5X3J1bigiY2VwaCBmc2lkIikKICAgICAgICByZXR1cm4gc3RyKHNlbGYuY2x1c3Rl
            ci5nZXRfZnNpZCgpKQoKICAgIGRlZiBfY29tbW9uX2NtZF9qc29uX2dlbihzZWxmLCBjbWRfanNv
            bik6CiAgICAgICAgY21kID0ganNvbi5kdW1wcyhjbWRfanNvbiwgc29ydF9rZXlzPVRydWUpCiAg
            ICAgICAgcmV0X3ZhbCwgY21kX291dCwgZXJyX21zZyA9IHNlbGYuY2x1c3Rlci5tb25fY29tbWFu
            ZChjbWQsIGIiIikKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnZlcmJvc2U6CiAgICAgICAg
            ICAgIHByaW50KGYiQ29tbWFuZCBJbnB1dDoge2NtZH0iKQogICAgICAgICAgICBwcmludCgKICAg
            ICAgICAgICAgICAgIGYiUmV0dXJuIFZhbDoge3JldF92YWx9XG5Db21tYW5kIE91dHB1dDoge2Nt
            ZF9vdXR9XG4iCiAgICAgICAgICAgICAgICBmIkVycm9yIE1lc3NhZ2U6IHtlcnJfbXNnfVxuLS0t
            LS0tLS0tLVxuIgogICAgICAgICAgICApCiAgICAgICAganNvbl9vdXQgPSB7fQogICAgICAgICMg
            aWYgdGhlcmUgaXMgbm8gZXJyb3IgKGkuZTsgcmV0X3ZhbCBpcyBaRVJPKSBhbmQgJ2NtZF9vdXQn
            IGlzIG5vdCBlbXB0eQogICAgICAgICMgdGhlbiBjb252ZXJ0ICdjbWRfb3V0JyB0byBhIGpzb24g
            b3V0cHV0CiAgICAgICAgaWYgcmV0X3ZhbCA9PSAwIGFuZCBjbWRfb3V0OgogICAgICAgICAgICBq
            c29uX291dCA9IGpzb24ubG9hZHMoY21kX291dCkKICAgICAgICByZXR1cm4gcmV0X3ZhbCwganNv
            bl9vdXQsIGVycl9tc2cKCiAgICBkZWYgZ2V0X2NlcGhfZXh0ZXJuYWxfbW9uX2RhdGEoc2VsZik6
            CiAgICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJxdW9ydW1fc3RhdHVzIiwgImZvcm1hdCI6
            ICJqc29uIn0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAg
            IHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGggIiArIGNtZF9qc29uWyJwcmVmaXgiXSkKICAgICAg
            ICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4o
            Y21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwK
            ICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgIidxdW9y
            dW1fc3RhdHVzJyBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAgICAgIGYiRXJyb3I6IHtl
            cnJfbXNnIGlmIHJldF92YWwgIT0gMCBlbHNlIHNlbGYuRU1QVFlfT1VUUFVUX0xJU1R9IgogICAg
            ICAgICAgICApCiAgICAgICAgcV9sZWFkZXJfbmFtZSA9IGpzb25fb3V0WyJxdW9ydW1fbGVhZGVy
            X25hbWUiXQogICAgICAgIHFfbGVhZGVyX2RldGFpbHMgPSB7fQogICAgICAgIHFfbGVhZGVyX21h
            dGNoaW5nX2xpc3QgPSBbCiAgICAgICAgICAgIGwgZm9yIGwgaW4ganNvbl9vdXRbIm1vbm1hcCJd
            WyJtb25zIl0gaWYgbFsibmFtZSJdID09IHFfbGVhZGVyX25hbWUKICAgICAgICBdCiAgICAgICAg
            aWYgbGVuKHFfbGVhZGVyX21hdGNoaW5nX2xpc3QpID09IDA6CiAgICAgICAgICAgIHJhaXNlIEV4
            ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oIk5vIG1hdGNoaW5nICdtb24nIGRldGFpbHMgZm91bmQi
            KQogICAgICAgIHFfbGVhZGVyX2RldGFpbHMgPSBxX2xlYWRlcl9tYXRjaGluZ19saXN0WzBdCiAg
            ICAgICAgIyBnZXQgdGhlIGFkZHJlc3MgdmVjdG9yIG9mIHRoZSBxdW9ydW0tbGVhZGVyCiAgICAg
            ICAgcV9sZWFkZXJfYWRkcnZlYyA9IHFfbGVhZGVyX2RldGFpbHMuZ2V0KCJwdWJsaWNfYWRkcnMi
            LCB7fSkuZ2V0KCJhZGRydmVjIiwgW10pCiAgICAgICAgaXBfYWRkciA9IHN0cihxX2xlYWRlcl9k
            ZXRhaWxzWyJwdWJsaWNfYWRkciJdLnNwbGl0KCIvIilbMF0pCgogICAgICAgIGlmIHNlbGYuX2Fy
            Z19wYXJzZXIudjJfcG9ydF9lbmFibGU6CiAgICAgICAgICAgIGlmIHFfbGVhZGVyX2FkZHJ2ZWNb
            MF1bInR5cGUiXSA9PSAidjIiOgogICAgICAgICAgICAgICAgaXBfYWRkciA9IHFfbGVhZGVyX2Fk
            ZHJ2ZWNbMF1bImFkZHIiXQogICAgICAgICAgICBlbGlmIGxlbihxX2xlYWRlcl9hZGRydmVjKSA+
            IDEgYW5kIHFfbGVhZGVyX2FkZHJ2ZWNbMV1bInR5cGUiXSA9PSAidjIiOgogICAgICAgICAgICAg
            ICAgaXBfYWRkciA9IHFfbGVhZGVyX2FkZHJ2ZWNbMV1bImFkZHIiXQogICAgICAgICAgICBlbHNl
            OgogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZSgKICAgICAgICAgICAgICAgICAgICAi
            J3YyJyBhZGRyZXNzIHR5cGUgbm90IHByZXNlbnQsIGFuZCAndjItcG9ydC1lbmFibGUnIGZsYWcg
            aXMgcHJvdmlkZWQiCiAgICAgICAgICAgICAgICApCgogICAgICAgIHJldHVybiBmIntzdHIocV9s
            ZWFkZXJfbmFtZSl9PXtpcF9hZGRyfSIKCiAgICBkZWYgX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAo
            c2VsZiwgaG9zdF9uYW1lLCBwb3J0LCBpcF90eXBlKToKICAgICAgICAjIGlmICdjbHVzdGVyJyBp
            bnN0YW5jZSBpcyBhIGR1bW15IHR5cGUsCiAgICAgICAgIyBjYWxsIHRoZSBkdW1teSBpbnN0YW5j
            ZSdzICJjb252ZXJ0IiBtZXRob2QKICAgICAgICBpZiBub3QgaG9zdF9uYW1lOgogICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKCJFbXB0eSBob3N0bmFtZSBwcm92aWRl
            ZCIpCiAgICAgICAgaWYgaXNpbnN0YW5jZShzZWxmLmNsdXN0ZXIsIER1bW15UmFkb3MpOgogICAg
            ICAgICAgICByZXR1cm4gc2VsZi5jbHVzdGVyLl9jb252ZXJ0X2hvc3RuYW1lX3RvX2lwKGhvc3Rf
            bmFtZSkKCiAgICAgICAgaWYgaXBfdHlwZSA9PSAiRlFETiI6CiAgICAgICAgICAgICMgY2hlY2sg
            d2hpY2ggaXAgRlFETiBzaG91bGQgYmUgY29udmVydGVkIHRvLCBJUHY0IG9yIElQdjYKICAgICAg
            ICAgICAgIyBjaGVjayB0aGUgaG9zdCBpcCwgdGhlIGVuZHBvaW50IGlwIHR5cGUgd291bGQgYmUg
            c2ltaWxhciB0byBob3N0IGlwCiAgICAgICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAib3Jj
            aCBob3N0IGxzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICAgICAgcmV0X3ZhbCwganNvbl9v
            dXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAg
            ICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAogICAgICAgICAgICBp
            ZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICAgICAgcmFp
            c2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJ29yY2gg
            aG9zdCBscycgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAgICAgICAgICAgICAgICAgZiJFcnJvcjog
            e2Vycl9tc2cgaWYgcmV0X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAg
            ICAgICAgICAgICAgICApCiAgICAgICAgICAgIGhvc3RfYWRkciA9IGpzb25fb3V0WzBdWyJhZGRy
            Il0KICAgICAgICAgICAgIyBhZGQgOjgwIHNhbXBsZSBwb3J0IGluIGlwX3R5cGUsIGFzIF9pbnZh
            bGlkX2VuZHBvaW50IGFsc28gdmVyaWZ5IHBvcnQKICAgICAgICAgICAgaG9zdF9pcF90eXBlID0g
            c2VsZi5faW52YWxpZF9lbmRwb2ludChob3N0X2FkZHIgKyAiOjgwIikKICAgICAgICAgICAgaW1w
            b3J0IHNvY2tldAoKICAgICAgICAgICAgIyBleGFtcGxlIG91dHB1dCBbKDxBZGRyZXNzRmFtaWx5
            LkFGX0lORVQ6IDI+LCA8U29ja2V0S2luZC5TT0NLX1NUUkVBTTogMT4sIDYsICcnLCAoJzkzLjE4
            NC4yMTYuMzQnLCA4MCkpLCAuLi5dCiAgICAgICAgICAgICMgd2UgbmVlZCB0byBnZXQgOTMuMTg0
            LjIxNi4zNCBzbyBpdCB3b3VsZCBiZSBpcFswXVs0XVswXQogICAgICAgICAgICBpZiBob3N0X2lw
            X3R5cGUgPT0gIklQdjYiOgogICAgICAgICAgICAgICAgaXAgPSBzb2NrZXQuZ2V0YWRkcmluZm8o
            CiAgICAgICAgICAgICAgICAgICAgaG9zdF9uYW1lLCBwb3J0LCBmYW1pbHk9c29ja2V0LkFGX0lO
            RVQ2LCBwcm90bz1zb2NrZXQuSVBQUk9UT19UQ1AKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgZWxpZiBob3N0X2lwX3R5cGUgPT0gIklQdjQiOgogICAgICAgICAgICAgICAgaXAgPSBzb2Nr
            ZXQuZ2V0YWRkcmluZm8oCiAgICAgICAgICAgICAgICAgICAgaG9zdF9uYW1lLCBwb3J0LCBmYW1p
            bHk9c29ja2V0LkFGX0lORVQsIHByb3RvPXNvY2tldC5JUFBST1RPX1RDUAogICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICBkZWwgc29ja2V0CiAgICAgICAgICAgIHJldHVybiBpcFswXVs0XVsw
            XQogICAgICAgIHJldHVybiBob3N0X25hbWUKCiAgICBkZWYgZ2V0X2FjdGl2ZV9hbmRfc3RhbmRi
            eV9tZ3JzKHNlbGYpOgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAg
            ICAgICAgcmV0dXJuICIiLCBzZWxmLmRyeV9ydW4oImNlcGggc3RhdHVzIikKICAgICAgICBtb25p
            dG9yaW5nX2VuZHBvaW50X3BvcnQgPSBzZWxmLl9hcmdfcGFyc2VyLm1vbml0b3JpbmdfZW5kcG9p
            bnRfcG9ydAogICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdCA9IHNlbGYuX2FyZ19w
            YXJzZXIubW9uaXRvcmluZ19lbmRwb2ludAogICAgICAgIHN0YW5kYnlfbWdycyA9IFtdCiAgICAg
            ICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdDoKICAgICAgICAgICAgY21kX2pz
            b24gPSB7InByZWZpeCI6ICJzdGF0dXMiLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgICAgICBy
            ZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21k
            X2pzb24pCiAgICAgICAgICAgICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQs
            CiAgICAgICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAgICAg
            ICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAg
            ICAgICAgICInbWdyIHNlcnZpY2VzJyBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAgICAg
            ICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09V
            VFBVVF9MSVNUfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRw
            b2ludCA9ICgKICAgICAgICAgICAgICAgIGpzb25fb3V0LmdldCgibWdybWFwIiwge30pLmdldCgi
            c2VydmljZXMiLCB7fSkuZ2V0KCJwcm9tZXRoZXVzIiwgIiIpCiAgICAgICAgICAgICkKICAgICAg
            ICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnQ6CiAgICAgICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJjYW4ndCBmaW5k
            IG1vbml0b3JpbmdfZW5kcG9pbnQsIHByb21ldGhldXMgbW9kdWxlIG1pZ2h0IG5vdCBiZSBlbmFi
            bGVkLCAiCiAgICAgICAgICAgICAgICAgICAgImVuYWJsZSB0aGUgbW9kdWxlIGJ5IHJ1bm5pbmcg
            J2NlcGggbWdyIG1vZHVsZSBlbmFibGUgcHJvbWV0aGV1cyciCiAgICAgICAgICAgICAgICApCiAg
            ICAgICAgICAgICMgbm93IGNoZWNrIHRoZSBzdGFuZC1ieSBtZ3ItcwogICAgICAgICAgICBzdGFu
            ZGJ5X2FyciA9IGpzb25fb3V0LmdldCgibWdybWFwIiwge30pLmdldCgic3RhbmRieXMiLCBbXSkK
            ICAgICAgICAgICAgZm9yIGVhY2hfc3RhbmRieSBpbiBzdGFuZGJ5X2FycjoKICAgICAgICAgICAg
            ICAgIGlmICJuYW1lIiBpbiBlYWNoX3N0YW5kYnkua2V5cygpOgogICAgICAgICAgICAgICAgICAg
            IHN0YW5kYnlfbWdycy5hcHBlbmQoZWFjaF9zdGFuZGJ5WyJuYW1lIl0pCiAgICAgICAgICAgIHRy
            eToKICAgICAgICAgICAgICAgIHBhcnNlZF9lbmRwb2ludCA9IHVybHBhcnNlKG1vbml0b3Jpbmdf
            ZW5kcG9pbnQpCiAgICAgICAgICAgIGV4Y2VwdCBWYWx1ZUVycm9yOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBmImlu
            dmFsaWQgZW5kcG9pbnQ6IHttb25pdG9yaW5nX2VuZHBvaW50fSIKICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0ID0gcGFyc2VkX2VuZHBvaW50
            Lmhvc3RuYW1lCiAgICAgICAgICAgIGlmIG5vdCBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQ6CiAg
            ICAgICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQgPSBzdHIocGFyc2VkX2VuZHBv
            aW50LnBvcnQpCgogICAgICAgICMgaWYgbW9uaXRvcmluZyBlbmRwb2ludCBwb3J0IGlzIG5vdCBz
            ZXQsIHB1dCBhIGRlZmF1bHQgbW9uIHBvcnQKICAgICAgICBpZiBub3QgbW9uaXRvcmluZ19lbmRw
            b2ludF9wb3J0OgogICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQgPSBzZWxmLkRF
            RkFVTFRfTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUCgogICAgICAgICMgdXNlciBjb3VsZCBnaXZl
            IGNvbW1hIGFuZCBzcGFjZSBzZXBhcmF0ZWQgaW5wdXRzIChsaWtlIC0tbW9uaXRvcmluZy1lbmRw
            b2ludD0iPGlwMT4sIDxpcDI+IikKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Qg
            PSBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3QucmVwbGFjZSgiLCIsICIgIikKICAgICAgICBt
            b25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXQgPSBtb25pdG9yaW5nX2VuZHBvaW50X2lw
            X2xpc3Quc3BsaXQoKQogICAgICAgICMgaWYgbW9uaXRvcmluZy1lbmRwb2ludCBjb3VsZCBub3Qg
            YmUgZm91bmQsIHJhaXNlIGFuIGVycm9yCiAgICAgICAgaWYgbGVuKG1vbml0b3JpbmdfZW5kcG9p
            bnRfaXBfbGlzdF9zcGxpdCkgPT0gMDoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVy
            ZUV4Y2VwdGlvbigiTm8gJ21vbml0b3JpbmctZW5kcG9pbnQnIGZvdW5kIikKICAgICAgICAjIGZp
            cnN0IGlwIGlzIHRyZWF0ZWQgYXMgdGhlIG1haW4gbW9uaXRvcmluZy1lbmRwb2ludAogICAgICAg
            IG1vbml0b3JpbmdfZW5kcG9pbnRfaXAgPSBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3Bs
            aXRbMF0KICAgICAgICAjIHJlc3Qgb2YgdGhlIGlwLXMgYXJlIGFkZGVkIHRvIHRoZSAnc3RhbmRi
            eV9tZ3JzJyBsaXN0CiAgICAgICAgc3RhbmRieV9tZ3JzLmV4dGVuZChtb25pdG9yaW5nX2VuZHBv
            aW50X2lwX2xpc3Rfc3BsaXRbMTpdKQogICAgICAgIGZhaWxlZF9pcCA9IG1vbml0b3JpbmdfZW5k
            cG9pbnRfaXAKCiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludCA9ICI6Ii5qb2luKAogICAgICAg
            ICAgICBbbW9uaXRvcmluZ19lbmRwb2ludF9pcCwgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0XQog
            ICAgICAgICkKICAgICAgICBpcF90eXBlID0gc2VsZi5faW52YWxpZF9lbmRwb2ludChtb25pdG9y
            aW5nX2VuZHBvaW50KQogICAgICAgIHRyeToKICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2lu
            dF9pcCA9IHNlbGYuX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoCiAgICAgICAgICAgICAgICBtb25p
            dG9yaW5nX2VuZHBvaW50X2lwLCBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQsIGlwX3R5cGUKICAg
            ICAgICAgICAgKQogICAgICAgICAgICAjIGNvbGxlY3QgYWxsIHRoZSAnc3RhbmQtYnknIG1nciBp
            cHMKICAgICAgICAgICAgbWdyX2lwcyA9IFtdCiAgICAgICAgICAgIGZvciBlYWNoX3N0YW5kYnlf
            bWdyIGluIHN0YW5kYnlfbWdyczoKICAgICAgICAgICAgICAgIGZhaWxlZF9pcCA9IGVhY2hfc3Rh
            bmRieV9tZ3IKICAgICAgICAgICAgICAgIG1ncl9pcHMuYXBwZW5kKAogICAgICAgICAgICAgICAg
            ICAgIHNlbGYuX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoCiAgICAgICAgICAgICAgICAgICAgICAg
            IGVhY2hfc3RhbmRieV9tZ3IsIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydCwgaXBfdHlwZQogICAg
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgICkKICAgICAgICBleGNlcHQ6CiAgICAg
            ICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBm
            IkNvbnZlcnNpb24gb2YgaG9zdDoge2ZhaWxlZF9pcH0gdG8gSVAgZmFpbGVkLiAiCiAgICAgICAg
            ICAgICAgICAiUGxlYXNlIGVudGVyIHRoZSBJUCBhZGRyZXNzZXMgb2YgYWxsIHRoZSBjZXBoLW1n
            cnMgd2l0aCB0aGUgJy0tbW9uaXRvcmluZy1lbmRwb2ludCcgZmxhZyIKICAgICAgICAgICAgKQoK
            ICAgICAgICBfLCBfLCBlcnIgPSBzZWxmLmVuZHBvaW50X2RpYWwobW9uaXRvcmluZ19lbmRwb2lu
            dCwgaXBfdHlwZSkKICAgICAgICBpZiBlcnIgPT0gIi0xIjoKICAgICAgICAgICAgcmFpc2UgRXhl
            Y3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihlcnIpCiAgICAgICAgIyBhZGQgdGhlIHZhbGlkYXRlZCBh
            Y3RpdmUgbWdyIElQIGludG8gdGhlIGZpcnN0IGluZGV4CiAgICAgICAgbWdyX2lwcy5pbnNlcnQo
            MCwgbW9uaXRvcmluZ19lbmRwb2ludF9pcCkKICAgICAgICBhbGxfbWdyX2lwc19zdHIgPSAiLCIu
            am9pbihtZ3JfaXBzKQogICAgICAgIHJldHVybiBhbGxfbWdyX2lwc19zdHIsIG1vbml0b3Jpbmdf
            ZW5kcG9pbnRfcG9ydAoKICAgIGRlZiBjaGVja191c2VyX2V4aXN0KHNlbGYsIHVzZXIpOgogICAg
            ICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAiYXV0aCBnZXQiLCAiZW50aXR5IjogZiJ7dXNlcn0i
            LCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBfID0gc2VsZi5f
            Y29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3Ig
            bGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICByZXR1cm4gIiIKICAgICAgICByZXR1cm4g
            c3RyKGpzb25fb3V0WzBdWyJrZXkiXSkKCiAgICBkZWYgZ2V0X2NlcGhmc19wcm92aXNpb25lcl9j
            YXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gImNsaWVudC5jc2ktY2VwaGZz
            LXByb3Zpc2lvbmVyIgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAgICJtb24iOiAiYWxsb3cg
            ciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLAogICAgICAgICAgICAibWdyIjogImFs
            bG93IHJ3IiwKICAgICAgICAgICAgIm9zZCI6ICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRh
            PSoiLAogICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0
            aF9wZXJtaXNzaW9uOgogICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3Bh
            cnNlci5rOHNfY2x1c3Rlcl9uYW1lCiAgICAgICAgICAgIGlmIGs4c19jbHVzdGVyX25hbWUgPT0g
            IiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAg
            ICAgICAgICAgICAgICAgICJrOHNfY2x1c3Rlcl9uYW1lIG5vdCBmb3VuZCwgcGxlYXNlIHNldCB0
            aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgY2VwaGZzX2ZpbGVzeXN0ZW0gPSBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVt
            X25hbWUKICAgICAgICAgICAgaWYgY2VwaGZzX2ZpbGVzeXN0ZW0gPT0gIiI6CiAgICAgICAgICAg
            ICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9uYW1lfSIKICAgICAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0te2s4c19jbHVzdGVyX25h
            bWV9LXtjZXBoZnNfZmlsZXN5c3RlbX0iCiAgICAgICAgICAgICAgICBjYXBzWyJvc2QiXSA9IGYi
            YWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT17Y2VwaGZzX2ZpbGVzeXN0ZW19IgoKICAgICAg
            ICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVmIGdldF9jZXBoZnNfbm9kZV9jYXBzX2FuZF9l
            bnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiCiAg
            ICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJhbGxvdyByLCBhbGxvdyBjb21tYW5k
            ICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJtZ3IiOiAiYWxsb3cgcnciLAogICAgICAg
            ICAgICAib3NkIjogImFsbG93IHJ3IHRhZyBjZXBoZnMgKj0qIiwKICAgICAgICAgICAgIm1kcyI6
            ICJhbGxvdyBydyIsCiAgICAgICAgfQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJp
            Y3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAgIGs4c19jbHVzdGVyX25hbWUgPSBzZWxm
            Ll9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAgICAgICAgICAgaWYgazhzX2NsdXN0ZXJf
            bmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRp
            b24oCiAgICAgICAgICAgICAgICAgICAgIms4c19jbHVzdGVyX25hbWUgbm90IGZvdW5kLCBwbGVh
            c2Ugc2V0IHRoZSAnLS1rOHMtY2x1c3Rlci1uYW1lJyBmbGFnIgogICAgICAgICAgICAgICAgKQog
            ICAgICAgICAgICBjZXBoZnNfZmlsZXN5c3RlbSA9IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2Zp
            bGVzeXN0ZW1fbmFtZQogICAgICAgICAgICBpZiBjZXBoZnNfZmlsZXN5c3RlbSA9PSAiIjoKICAg
            ICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0te2s4c19jbHVzdGVyX25hbWV9IgogICAg
            ICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17azhzX2Ns
            dXN0ZXJfbmFtZX0te2NlcGhmc19maWxlc3lzdGVtfSIKICAgICAgICAgICAgICAgIGNhcHNbIm9z
            ZCJdID0gZiJhbGxvdyBydyB0YWcgY2VwaGZzICo9e2NlcGhmc19maWxlc3lzdGVtfSIKCiAgICAg
            ICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfZW50aXR5KAogICAgICAgIHNlbGYs
            CiAgICAgICAgZW50aXR5LAogICAgICAgIHJiZF9wb29sX25hbWUsCiAgICAgICAgYWxpYXNfcmJk
            X3Bvb2xfbmFtZSwKICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lLAogICAgICAgIHJhZG9zX25hbWVz
            cGFjZSwKICAgICk6CiAgICAgICAgaWYgKAogICAgICAgICAgICByYmRfcG9vbF9uYW1lLmNvdW50
            KCIuIikgIT0gMAogICAgICAgICAgICBvciByYmRfcG9vbF9uYW1lLmNvdW50KCJfIikgIT0gMAog
            ICAgICAgICAgICBvciBhbGlhc19yYmRfcG9vbF9uYW1lICE9ICIiCiAgICAgICAgICAgICMgY2hl
            Y2tpbmcgYWxpYXNfcmJkX3Bvb2xfbmFtZSBpcyBub3QgZW1wdHkgYXMgdGhlcmUgbWF5YmUgYSBz
            cGVjaWFsIGNoYXJhY3RlciB1c2VkIG90aGVyIHRoYW4gLiBvciBfCiAgICAgICAgKToKICAgICAg
            ICAgICAgaWYgYWxpYXNfcmJkX3Bvb2xfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgInBsZWFzZSBz
            ZXQgdGhlICctLWFsaWFzLXJiZC1kYXRhLXBvb2wtbmFtZScgZmxhZyBhcyB0aGUgcmJkIGRhdGEg
            cG9vbCBuYW1lIGNvbnRhaW5zICcuJyBvciAnXyciCiAgICAgICAgICAgICAgICApCiAgICAgICAg
            ICAgIGlmICgKICAgICAgICAgICAgICAgIGFsaWFzX3JiZF9wb29sX25hbWUuY291bnQoIi4iKSAh
            PSAwCiAgICAgICAgICAgICAgICBvciBhbGlhc19yYmRfcG9vbF9uYW1lLmNvdW50KCJfIikgIT0g
            MAogICAgICAgICAgICApOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJy0tYWxpYXMtcmJkLWRhdGEtcG9vbC1uYW1l
            JyBmbGFnIHZhbHVlIHNob3VsZCBub3QgY29udGFpbiAnLicgb3IgJ18nIgogICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9uYW1lfS17
            YWxpYXNfcmJkX3Bvb2xfbmFtZX0iCiAgICAgICAgZWxzZToKICAgICAgICAgICAgZW50aXR5ID0g
            ZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0te3JiZF9wb29sX25hbWV9IgoKICAgICAgICBp
            ZiByYWRvc19uYW1lc3BhY2U6CiAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0te3JhZG9z
            X25hbWVzcGFjZX0iCiAgICAgICAgcmV0dXJuIGVudGl0eQoKICAgIGRlZiBnZXRfcmJkX3Byb3Zp
            c2lvbmVyX2NhcHNfYW5kX2VudGl0eShzZWxmKToKICAgICAgICBlbnRpdHkgPSAiY2xpZW50LmNz
            aS1yYmQtcHJvdmlzaW9uZXIiCiAgICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJw
            cm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLAogICAgICAgICAgICAi
            bWdyIjogImFsbG93IHJ3IiwKICAgICAgICAgICAgIm9zZCI6ICJwcm9maWxlIHJiZCIsCiAgICAg
            ICAgfQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Np
            b246CiAgICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRh
            X3Bvb2xfbmFtZQogICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3Bh
            cnNlci5hbGlhc19yYmRfZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFt
            ZSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQogICAgICAgICAgICByYWRvc19u
            YW1lc3BhY2UgPSBzZWxmLl9hcmdfcGFyc2VyLnJhZG9zX25hbWVzcGFjZQogICAgICAgICAgICBp
            ZiByYmRfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAibWFuZGF0b3J5IGZsYWcgbm90IGZv
            dW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1yYmQtZGF0YS1wb29sLW5hbWUnIGZsYWciCiAgICAgICAg
            ICAgICAgICApCiAgICAgICAgICAgIGlmIGs4c19jbHVzdGVyX25hbWUgPT0gIiI6CiAgICAgICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLWs4cy1jbHVz
            dGVyLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVudGl0eSA9IHNl
            bGYuZ2V0X2VudGl0eSgKICAgICAgICAgICAgICAgIGVudGl0eSwKICAgICAgICAgICAgICAgIHJi
            ZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lLAogICAgICAg
            ICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAgICAgICAgICAgIHJhZG9zX25hbWVzcGFj
            ZSwKICAgICAgICAgICAgKQogICAgICAgICAgICBpZiByYWRvc19uYW1lc3BhY2UgIT0gIiI6CiAg
            ICAgICAgICAgICAgICBjYXBzWyJvc2QiXSA9ICgKICAgICAgICAgICAgICAgICAgICBmInByb2Zp
            bGUgcmJkIHBvb2w9e3JiZF9wb29sX25hbWV9IG5hbWVzcGFjZT17cmFkb3NfbmFtZXNwYWNlfSIK
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGNhcHNb
            Im9zZCJdID0gZiJwcm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9uYW1lfSIKCiAgICAgICAgcmV0
            dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfcmJkX25vZGVfY2Fwc19hbmRfZW50aXR5KHNl
            bGYpOgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3NpLXJiZC1ub2RlIgogICAgICAgIGNhcHMg
            PSB7CiAgICAgICAgICAgICJtb24iOiAicHJvZmlsZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBi
            bG9ja2xpc3QnIiwKICAgICAgICAgICAgIm9zZCI6ICJwcm9maWxlIHJiZCIsCiAgICAgICAgfQog
            ICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb246CiAg
            ICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xf
            bmFtZQogICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5h
            bGlhc19yYmRfZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNl
            bGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQogICAgICAgICAgICByYWRvc19uYW1lc3Bh
            Y2UgPSBzZWxmLl9hcmdfcGFyc2VyLnJhZG9zX25hbWVzcGFjZQogICAgICAgICAgICBpZiByYmRf
            cG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAibWFuZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBw
            bGVhc2Ugc2V0IHRoZSAnLS1yYmQtZGF0YS1wb29sLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAg
            ICApCiAgICAgICAgICAgIGlmIGs4c19jbHVzdGVyX25hbWUgPT0gIiI6CiAgICAgICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJt
            YW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLWs4cy1jbHVzdGVyLW5h
            bWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVudGl0eSA9IHNlbGYuZ2V0
            X2VudGl0eSgKICAgICAgICAgICAgICAgIGVudGl0eSwKICAgICAgICAgICAgICAgIHJiZF9wb29s
            X25hbWUsCiAgICAgICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lLAogICAgICAgICAgICAg
            ICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAgICAgICAgICAgIHJhZG9zX25hbWVzcGFjZSwKICAg
            ICAgICAgICAgKQogICAgICAgICAgICBpZiByYWRvc19uYW1lc3BhY2UgIT0gIiI6CiAgICAgICAg
            ICAgICAgICBjYXBzWyJvc2QiXSA9ICgKICAgICAgICAgICAgICAgICAgICBmInByb2ZpbGUgcmJk
            IHBvb2w9e3JiZF9wb29sX25hbWV9IG5hbWVzcGFjZT17cmFkb3NfbmFtZXNwYWNlfSIKICAgICAg
            ICAgICAgICAgICkKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJd
            ID0gZiJwcm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9uYW1lfSIKCiAgICAgICAgcmV0dXJuIGNh
            cHMsIGVudGl0eQoKICAgIGRlZiBnZXRfZGVmYXVsdFVzZXJfY2Fwc19hbmRfZW50aXR5KHNlbGYp
            OgogICAgICAgIGVudGl0eSA9IHNlbGYucnVuX2FzX3VzZXIKICAgICAgICBjYXBzID0gewogICAg
            ICAgICAgICAibW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxs
            b3cgY29tbWFuZCB2ZXJzaW9uIiwKICAgICAgICAgICAgIm1nciI6ICJhbGxvdyBjb21tYW5kIGNv
            bmZpZyIsCiAgICAgICAgICAgICJvc2QiOiBmInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cg
            cnd4IHBvb2w9e3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cubWV0YSwgYWxs
            b3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD17c2VsZi5fYXJnX3BhcnNlci5yZ3df
            cG9vbF9wcmVmaXh9LnJndy5jb250cm9sLCBhbGxvdyByeCBwb29sPXtzZWxmLl9hcmdfcGFyc2Vy
            LnJnd19wb29sX3ByZWZpeH0ucmd3LmxvZywgYWxsb3cgeCBwb29sPXtzZWxmLl9hcmdfcGFyc2Vy
            LnJnd19wb29sX3ByZWZpeH0ucmd3LmJ1Y2tldHMuaW5kZXgiLAogICAgICAgIH0KCiAgICAgICAg
            cmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfY2Fwc19hbmRfZW50aXR5KHNlbGYsIHVz
            ZXJfbmFtZSk6CiAgICAgICAgaWYgImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiBpbiB1
            c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIg
            IT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVk
            X2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZ2V0X2NlcGhm
            c19wcm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoKQogICAgICAgIGlmICJjbGllbnQuY3NpLWNl
            cGhmcy1ub2RlIiBpbiB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQuY3NpLWNlcGhm
            cy1ub2RlIiAhPSB1c2VyX25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJl
            c3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5n
            ZXRfY2VwaGZzX25vZGVfY2Fwc19hbmRfZW50aXR5KCkKICAgICAgICBpZiAiY2xpZW50LmNzaS1y
            YmQtcHJvdmlzaW9uZXIiIGluIHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVudC5jc2kt
            cmJkLXByb3Zpc2lvbmVyIiAhPSB1c2VyX25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdf
            cGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1
            cm4gc2VsZi5nZXRfcmJkX3Byb3Zpc2lvbmVyX2NhcHNfYW5kX2VudGl0eSgpCiAgICAgICAgaWYg
            ImNsaWVudC5jc2ktcmJkLW5vZGUiIGluIHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVu
            dC5jc2ktcmJkLW5vZGUiICE9IHVzZXJfbmFtZToKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19w
            YXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVlCiAgICAgICAgICAgIHJldHVy
            biBzZWxmLmdldF9yYmRfbm9kZV9jYXBzX2FuZF9lbnRpdHkoKQogICAgICAgIGlmICJjbGllbnQu
            aGVhbHRoY2hlY2tlciIgaW4gdXNlcl9uYW1lOgogICAgICAgICAgICBpZiAiY2xpZW50LmhlYWx0
            aGNoZWNrZXIiICE9IHVzZXJfbmFtZToKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIu
            cmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVlCiAgICAgICAgICAgIHJldHVybiBzZWxm
            LmdldF9kZWZhdWx0VXNlcl9jYXBzX2FuZF9lbnRpdHkoKQoKICAgICAgICByYWlzZSBFeGVjdXRp
            b25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICBmIm5vIHVzZXIgZm91bmQgd2l0aCB1c2Vy
            X25hbWU6IHt1c2VyX25hbWV9LCAiCiAgICAgICAgICAgICJnZXRfY2Fwc19hbmRfZW50aXR5IGNv
            bW1hbmQgZmFpbGVkLlxuIgogICAgICAgICkKCiAgICBkZWYgY3JlYXRlX2NlcGhDU0lLZXlyaW5n
            X3VzZXIoc2VsZiwgdXNlcik6CiAgICAgICAgIiIiCiAgICAgICAgY29tbWFuZDogY2VwaCBhdXRo
            IGdldC1vci1jcmVhdGUgY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIgbW9uICdhbGxvdyBy
            JyBtZ3IgJ2FsbG93IHJ3JyBvc2QgJ2FsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KicKICAg
            ICAgICAiIiIKICAgICAgICBjYXBzLCBlbnRpdHkgPSBzZWxmLmdldF9jYXBzX2FuZF9lbnRpdHko
            dXNlcikKICAgICAgICBjbWRfanNvbiA9IHsKICAgICAgICAgICAgInByZWZpeCI6ICJhdXRoIGdl
            dC1vci1jcmVhdGUiLAogICAgICAgICAgICAiZW50aXR5IjogZW50aXR5LAogICAgICAgICAgICAi
            Y2FwcyI6IFtjYXAgZm9yIGNhcF9saXN0IGluIGxpc3QoY2Fwcy5pdGVtcygpKSBmb3IgY2FwIGlu
            IGNhcF9saXN0XSwKICAgICAgICAgICAgImZvcm1hdCI6ICJqc29uIiwKICAgICAgICB9CgogICAg
            ICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuICgKICAg
            ICAgICAgICAgICAgIHNlbGYuZHJ5X3J1bigKICAgICAgICAgICAgICAgICAgICAiY2VwaCAiCiAg
            ICAgICAgICAgICAgICAgICAgKyBjbWRfanNvblsicHJlZml4Il0KICAgICAgICAgICAgICAgICAg
            ICArICIgIgogICAgICAgICAgICAgICAgICAgICsgY21kX2pzb25bImVudGl0eSJdCiAgICAgICAg
            ICAgICAgICAgICAgKyAiICIKICAgICAgICAgICAgICAgICAgICArICIgIi5qb2luKGNtZF9qc29u
            WyJjYXBzIl0pCiAgICAgICAgICAgICAgICApLAogICAgICAgICAgICAgICAgIiIsCiAgICAgICAg
            ICAgICkKICAgICAgICAjIGNoZWNrIGlmIHVzZXIgYWxyZWFkeSBleGlzdAogICAgICAgIHVzZXJf
            a2V5ID0gc2VsZi5jaGVja191c2VyX2V4aXN0KGVudGl0eSkKICAgICAgICBpZiB1c2VyX2tleSAh
            PSAiIjoKICAgICAgICAgICAgcmV0dXJuIHVzZXJfa2V5LCBmIntlbnRpdHkuc3BsaXQoJy4nLCAx
            KVsxXX0iCiAgICAgICAgICAgICMgZW50aXR5LnNwbGl0KCcuJywxKVsxXSB0byByZW5hbWUgZW50
            aXR5KGNsaWVudC5jc2ktcmJkLW5vZGUpIGFzIGNzaS1yYmQtbm9kZQoKICAgICAgICByZXRfdmFs
            LCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24p
            CiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwKICAgICAgICBp
            ZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiInYXV0aCBnZXQtb3It
            Y3JlYXRlIHt1c2VyfScgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAgICAgICAgICAgICBmIkVycm9y
            OiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIK
            ICAgICAgICAgICAgKQogICAgICAgIHJldHVybiBzdHIoanNvbl9vdXRbMF1bImtleSJdKSwgZiJ7
            ZW50aXR5LnNwbGl0KCcuJywgMSlbMV19IgogICAgICAgICMgZW50aXR5LnNwbGl0KCcuJywxKVsx
            XSB0byByZW5hbWUgZW50aXR5KGNsaWVudC5jc2ktcmJkLW5vZGUpIGFzIGNzaS1yYmQtbm9kZQoK
            ICAgIGRlZiBnZXRfY2VwaGZzX2RhdGFfcG9vbF9kZXRhaWxzKHNlbGYpOgogICAgICAgIGNtZF9q
            c29uID0geyJwcmVmaXgiOiAiZnMgbHMiLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIGlmIHNl
            bGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigi
            Y2VwaCAiICsgY21kX2pzb25bInByZWZpeCJdKQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBl
            cnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICAjIGlm
            IHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LCByZXBvcnQgYW4gZXJyb3IKICAgICAg
            ICBpZiByZXRfdmFsICE9IDA6CiAgICAgICAgICAgICMgaWYgZnMgYW5kIGRhdGFfcG9vbCBhcmd1
            bWVudHMgYXJlIG5vdCBzZXQsIHNpbGVudGx5IHJldHVybgogICAgICAgICAgICBpZiAoCiAgICAg
            ICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUgPT0gIiIK
            ICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFt
            ZSA9PSAiIgogICAgICAgICAgICApOgogICAgICAgICAgICAgICAgcmV0dXJuCiAgICAgICAgICAg
            ICMgaWYgdXNlciBoYXMgcHJvdmlkZWQgYW55IG9mIHRoZQogICAgICAgICAgICAjICctLWNlcGhm
            cy1maWxlc3lzdGVtLW5hbWUnIG9yICctLWNlcGhmcy1kYXRhLXBvb2wtbmFtZScgYXJndW1lbnRz
            LAogICAgICAgICAgICAjIHJhaXNlIGFuIGV4Y2VwdGlvbiBhcyB3ZSBhcmUgdW5hYmxlIHRvIHZl
            cmlmeSB0aGUgYXJncwogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KAogICAgICAgICAgICAgICAgZiInZnMgbHMnIGNlcGggY2FsbCBmYWlsZWQgd2l0aCBlcnJvcjog
            e2Vycl9tc2d9IgogICAgICAgICAgICApCgogICAgICAgIG1hdGNoaW5nX2pzb25fb3V0ID0ge30K
            ICAgICAgICAjIGlmICctLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUnIGFyZ3VtZW50IGlzIHByb3Zp
            ZGVkLAogICAgICAgICMgY2hlY2sgd2hldGhlciB0aGUgcHJvdmlkZWQgZmlsZXN5c3RlbS1uYW1l
            IGV4aXN0cyBvciBub3QKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lz
            dGVtX25hbWU6CiAgICAgICAgICAgICMgZ2V0IHRoZSBtYXRjaGluZyBsaXN0CiAgICAgICAgICAg
            IG1hdGNoaW5nX2pzb25fb3V0X2xpc3QgPSBbCiAgICAgICAgICAgICAgICBtYXRjaGVkCiAgICAg
            ICAgICAgICAgICBmb3IgbWF0Y2hlZCBpbiBqc29uX291dAogICAgICAgICAgICAgICAgaWYgc3Ry
            KG1hdGNoZWRbIm5hbWUiXSkgPT0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9u
            YW1lCiAgICAgICAgICAgIF0KICAgICAgICAgICAgIyB1bmFibGUgdG8gZmluZCBhIG1hdGNoaW5n
            IGZzLW5hbWUsIHJhaXNlIGFuIGVycm9yCiAgICAgICAgICAgIGlmIGxlbihtYXRjaGluZ19qc29u
            X291dF9saXN0KSA9PSAwOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBmIkZpbGVzeXN0ZW0gcHJvdmlkZWQsICd7c2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lfScsICIKICAgICAgICAgICAgICAg
            ICAgICBmImlzIG5vdCBmb3VuZCBpbiB0aGUgZnMtbGlzdDoge1tzdHIoeFsnbmFtZSddKSBmb3Ig
            eCBpbiBqc29uX291dF19IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBtYXRjaGluZ19q
            c29uX291dCA9IG1hdGNoaW5nX2pzb25fb3V0X2xpc3RbMF0KICAgICAgICAjIGlmIGNlcGhmcyBm
            aWxlc3lzdGVtIG5hbWUgaXMgbm90IHByb3ZpZGVkLAogICAgICAgICMgdHJ5IHRvIGdldCBhIGRl
            ZmF1bHQgZnMgbmFtZSBieSBkb2luZyB0aGUgZm9sbG93aW5nCiAgICAgICAgZWxzZToKICAgICAg
            ICAgICAgIyBhLiBjaGVjayBpZiB0aGVyZSBpcyBvbmx5IG9uZSBmaWxlc3lzdGVtIGlzIHByZXNl
            bnQKICAgICAgICAgICAgaWYgbGVuKGpzb25fb3V0KSA9PSAxOgogICAgICAgICAgICAgICAgbWF0
            Y2hpbmdfanNvbl9vdXQgPSBqc29uX291dFswXQogICAgICAgICAgICAjIGIuIG9yIGVsc2UsIGNo
            ZWNrIGlmIGRhdGFfcG9vbCBuYW1lIGlzIHByb3ZpZGVkCiAgICAgICAgICAgIGVsaWYgc2VsZi5f
            YXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWU6CiAgICAgICAgICAgICAgICAjIGFuZCBp
            ZiBwcmVzZW50LCBjaGVjayB3aGV0aGVyIHRoZXJlIGV4aXN0cyBhIGZzIHdoaWNoIGhhcyB0aGUg
            ZGF0YV9wb29sCiAgICAgICAgICAgICAgICBmb3IgZWFjaEogaW4ganNvbl9vdXQ6CiAgICAgICAg
            ICAgICAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUgaW4g
            ZWFjaEpbImRhdGFfcG9vbHMiXToKICAgICAgICAgICAgICAgICAgICAgICAgbWF0Y2hpbmdfanNv
            bl9vdXQgPSBlYWNoSgogICAgICAgICAgICAgICAgICAgICAgICBicmVhawogICAgICAgICAgICAg
            ICAgIyBpZiB0aGVyZSBpcyBubyBtYXRjaGluZyBmcyBleGlzdHMsIHRoYXQgbWVhbnMgcHJvdmlk
            ZWQgZGF0YV9wb29sIG5hbWUgaXMgaW52YWxpZAogICAgICAgICAgICAgICAgaWYgbm90IG1hdGNo
            aW5nX2pzb25fb3V0OgogICAgICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVF
            eGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgICAgIGYiUHJvdmlkZWQgZGF0YV9wb29sIG5h
            bWUsIHtzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZX0sIgogICAgICAgICAg
            ICAgICAgICAgICAgICAiIGRvZXMgbm90IGV4aXN0cyIKICAgICAgICAgICAgICAgICAgICApCiAg
            ICAgICAgICAgICMgYy4gaWYgbm90aGluZyBpcyBzZXQgYW5kIGNvdWxkbid0IGZpbmQgYSBkZWZh
            dWx0LAogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgIyBqdXN0IHJldHVybiBzaWxl
            bnRseQogICAgICAgICAgICAgICAgcmV0dXJuCgogICAgICAgIGlmIG1hdGNoaW5nX2pzb25fb3V0
            OgogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUgPSBz
            dHIobWF0Y2hpbmdfanNvbl9vdXRbIm5hbWUiXSkKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5jZXBoZnNfbWV0YWRhdGFfcG9vbF9uYW1lID0gc3RyKAogICAgICAgICAgICAgICAgbWF0Y2hp
            bmdfanNvbl9vdXRbIm1ldGFkYXRhX3Bvb2wiXQogICAgICAgICAgICApCgogICAgICAgIGlmIGlz
            aW5zdGFuY2UobWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXSwgbGlzdCk6CiAgICAgICAg
            ICAgICMgaWYgdGhlIHVzZXIgaGFzIGFscmVhZHkgcHJvdmlkZWQgZGF0YS1wb29sLW5hbWUsCiAg
            ICAgICAgICAgICMgdGhyb3VnaCAtLWNlcGhmcy1kYXRhLXBvb2wtbmFtZQogICAgICAgICAgICBp
            ZiBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZToKICAgICAgICAgICAgICAg
            ICMgaWYgdGhlIHByb3ZpZGVkIG5hbWUgaXMgbm90IG1hdGNoaW5nIHdpdGggdGhlIG9uZSBpbiB0
            aGUgbGlzdAogICAgICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2Fy
            Z19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgICAgICAgICAgbm90IGlu
            IG1hdGNoaW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0KICAgICAgICAgICAgICAgICk6CiAgICAg
            ICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAg
            ICAgICAgICAgICAgICAgZiJQcm92aWRlZCBkYXRhLXBvb2wtbmFtZTogJ3tzZWxmLl9hcmdfcGFy
            c2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZX0nLCAiCiAgICAgICAgICAgICAgICAgICAgICAgICJk
            b2Vzbid0IG1hdGNoIGZyb20gdGhlIGRhdGEtcG9vbHMgbGlzdDogIgogICAgICAgICAgICAgICAg
            ICAgICAgICBmIntbc3RyKHgpIGZvciB4IGluIG1hdGNoaW5nX2pzb25fb3V0WydkYXRhX3Bvb2xz
            J11dfSIKICAgICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICMgaWYgZGF0YV9wb29sIG5h
            bWUgaXMgbm90IHByb3ZpZGVkLAogICAgICAgICAgICAjIHRoZW4gdHJ5IHRvIGZpbmQgYSBkZWZh
            dWx0IGRhdGEgcG9vbCBuYW1lCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICAjIGlm
            IG5vIGRhdGFfcG9vbHMgZXhpc3QsIHNpbGVudGx5IHJldHVybgogICAgICAgICAgICAgICAgaWYg
            bGVuKG1hdGNoaW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0pID09IDA6CiAgICAgICAgICAgICAg
            ICAgICAgcmV0dXJuCiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRh
            X3Bvb2xfbmFtZSA9IHN0cigKICAgICAgICAgICAgICAgICAgICBtYXRjaGluZ19qc29uX291dFsi
            ZGF0YV9wb29scyJdWzBdCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICMgaWYgdGhlcmUg
            YXJlIG1vcmUgdGhhbiBvbmUgJ2RhdGFfcG9vbHMnIGV4aXN0LAogICAgICAgICAgICAjIHRoZW4g
            d2FybiB0aGUgdXNlciB0aGF0IHdlIGFyZSB1c2luZyB0aGUgc2VsZWN0ZWQgbmFtZQogICAgICAg
            ICAgICBpZiBsZW4obWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXSkgPiAxOgogICAgICAg
            ICAgICAgICAgcHJpbnQoCiAgICAgICAgICAgICAgICAgICAgIldBUk5JTkc6IE11bHRpcGxlIGRh
            dGEgcG9vbHMgZGV0ZWN0ZWQ6ICIKICAgICAgICAgICAgICAgICAgICBmIntbc3RyKHgpIGZvciB4
            IGluIG1hdGNoaW5nX2pzb25fb3V0WydkYXRhX3Bvb2xzJ11dfVxuIgogICAgICAgICAgICAgICAg
            ICAgIGYiVXNpbmcgdGhlIGRhdGEtcG9vbDogJ3tzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRh
            X3Bvb2xfbmFtZX0nXG4iCiAgICAgICAgICAgICAgICApCgogICAgZGVmIGNyZWF0ZV9jaGVja2Vy
            S2V5KHNlbGYsIHVzZXIpOgogICAgICAgIGNhcHMsIGVudGl0eSA9IHNlbGYuZ2V0X2NhcHNfYW5k
            X2VudGl0eSh1c2VyKQogICAgICAgIGNtZF9qc29uID0gewogICAgICAgICAgICAicHJlZml4Ijog
            ImF1dGggZ2V0LW9yLWNyZWF0ZSIsCiAgICAgICAgICAgICJlbnRpdHkiOiBlbnRpdHksCiAgICAg
            ICAgICAgICJjYXBzIjogW2NhcCBmb3IgY2FwX2xpc3QgaW4gbGlzdChjYXBzLml0ZW1zKCkpIGZv
            ciBjYXAgaW4gY2FwX2xpc3RdLAogICAgICAgICAgICAiZm9ybWF0IjogImpzb24iLAogICAgICAg
            IH0KCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1
            cm4gc2VsZi5kcnlfcnVuKAogICAgICAgICAgICAgICAgImNlcGggIgogICAgICAgICAgICAgICAg
            KyBjbWRfanNvblsicHJlZml4Il0KICAgICAgICAgICAgICAgICsgIiAiCiAgICAgICAgICAgICAg
            ICArIGNtZF9qc29uWyJlbnRpdHkiXQogICAgICAgICAgICAgICAgKyAiICIKICAgICAgICAgICAg
            ICAgICsgIiAiLmpvaW4oY21kX2pzb25bImNhcHMiXSkKICAgICAgICAgICAgKQogICAgICAgICMg
            Y2hlY2sgaWYgdXNlciBhbHJlYWR5IGV4aXN0CiAgICAgICAgdXNlcl9rZXkgPSBzZWxmLmNoZWNr
            X3VzZXJfZXhpc3QoZW50aXR5KQogICAgICAgIGlmIHVzZXJfa2V5ICE9ICIiOgogICAgICAgICAg
            ICByZXR1cm4gdXNlcl9rZXkKCiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBz
            ZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICMgaWYgdGhlcmUgaXMg
            YW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihq
            c29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgICAgIGYiJ2F1dGggZ2V0LW9yLWNyZWF0ZSB7c2VsZi5ydW5fYXNfdXNl
            cn0nIGNvbW1hbmQgZmFpbGVkXG4iCiAgICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBp
            ZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAg
            KQogICAgICAgIHJldHVybiBzdHIoanNvbl9vdXRbMF1bImtleSJdKQoKICAgIGRlZiBnZXRfY2Vw
            aF9kYXNoYm9hcmRfbGluayhzZWxmKToKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4IjogIm1n
            ciBzZXJ2aWNlcyIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNl
            ci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKCJjZXBoICIgKyBjbWRf
            anNvblsicHJlZml4Il0pCiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIF8gPSBzZWxmLl9jb21t
            b25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNj
            ZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkg
            PT0gMDoKICAgICAgICAgICAgcmV0dXJuIE5vbmUKICAgICAgICBpZiAiZGFzaGJvYXJkIiBub3Qg
            aW4ganNvbl9vdXQ6CiAgICAgICAgICAgIHJldHVybiBOb25lCiAgICAgICAgcmV0dXJuIGpzb25f
            b3V0WyJkYXNoYm9hcmQiXQoKICAgIGRlZiBjcmVhdGVfcmd3X2FkbWluX29wc191c2VyKHNlbGYp
            OgogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAgICAgICAg
            ICAidXNlciIsCiAgICAgICAgICAgICJjcmVhdGUiLAogICAgICAgICAgICAiLS11aWQiLAogICAg
            ICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9PUFNfVVNFUl9OQU1FLAogICAgICAgICAg
            ICAiLS1kaXNwbGF5LW5hbWUiLAogICAgICAgICAgICAiUm9vayBSR1cgQWRtaW4gT3BzIHVzZXIi
            LAogICAgICAgICAgICAiLS1jYXBzIiwKICAgICAgICAgICAgImJ1Y2tldHM9Kjt1c2Vycz0qO3Vz
            YWdlPXJlYWQ7bWV0YWRhdGE9cmVhZDt6b25lPXJlYWQiLAogICAgICAgICAgICAiLS1yZ3ctcmVh
            bG0iLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lLAogICAgICAg
            ICAgICAiLS1yZ3ctem9uZWdyb3VwIiwKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3df
            em9uZWdyb3VwX25hbWUsCiAgICAgICAgICAgICItLXJndy16b25lIiwKICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1lLAogICAgICAgIF0KICAgICAgICBpZiBzZWxmLl9h
            cmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGgg
            IiArICIgIi5qb2luKGNtZCkpCiAgICAgICAgdHJ5OgogICAgICAgICAgICBvdXRwdXQgPSBzdWJw
            cm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAg
            ZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yIGFzIGV4ZWNFcnI6CiAgICAgICAg
            ICAgICMgaWYgdGhlIHVzZXIgYWxyZWFkeSBleGlzdHMsIHdlIGp1c3QgcXVlcnkgaXQKICAgICAg
            ICAgICAgaWYgZXhlY0Vyci5yZXR1cm5jb2RlID09IGVycm5vLkVFWElTVDoKICAgICAgICAgICAg
            ICAgIGNtZCA9IFsKICAgICAgICAgICAgICAgICAgICAicmFkb3Nndy1hZG1pbiIsCiAgICAgICAg
            ICAgICAgICAgICAgInVzZXIiLAogICAgICAgICAgICAgICAgICAgICJpbmZvIiwKICAgICAgICAg
            ICAgICAgICAgICAiLS11aWQiLAogICAgICAgICAgICAgICAgICAgIHNlbGYuRVhURVJOQUxfUkdX
            X0FETUlOX09QU19VU0VSX05BTUUsCiAgICAgICAgICAgICAgICAgICAgIi0tcmd3LXJlYWxtIiwK
            ICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lLAogICAg
            ICAgICAgICAgICAgICAgICItLXJndy16b25lZ3JvdXAiLAogICAgICAgICAgICAgICAgICAgIHNl
            bGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lLAogICAgICAgICAgICAgICAgICAgICIt
            LXJndy16b25lIiwKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25l
            X25hbWUsCiAgICAgICAgICAgICAgICBdCiAgICAgICAgICAgICAgICB0cnk6CiAgICAgICAgICAg
            ICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3Vi
            cHJvY2Vzcy5QSVBFKQogICAgICAgICAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJv
            Y2Vzc0Vycm9yIGFzIGV4ZWNFcnI6CiAgICAgICAgICAgICAgICAgICAgZXJyX21zZyA9ICgKICAg
            ICAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5kIHtjbWR9LiBP
            dXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICAgICAgZiJDb2Rl
            OiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtleGVjRXJyLnN0ZGVycn0iCiAgICAgICAg
            ICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21z
            ZykKICAgICAgICAgICAgICAgICAgICByZXR1cm4gTm9uZSwgTm9uZSwgRmFsc2UsICItMSIKICAg
            ICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVycl9tc2cgPSAoCiAgICAgICAgICAgICAg
            ICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5kIHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJy
            Lm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICBmIkNvZGU6IHtleGVjRXJyLnJldHVybmNv
            ZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykKICAgICAgICAgICAgICAgIHJldHVybiBO
            b25lLCBOb25lLCBGYWxzZSwgIi0xIgoKICAgICAgICAjIGlmIGl0IGlzIHB5dGhvbjIsIGRvbid0
            IGNoZWNrIGZvciBjZXBoIHZlcnNpb24gZm9yIGFkZGluZyBgaW5mbz1yZWFkYCBjYXAocmd3X3Zh
            bGlkYXRpb24pCiAgICAgICAgaWYgc3lzLnZlcnNpb25faW5mby5tYWpvciA8IDM6CiAgICAgICAg
            ICAgIGpzb25vdXRwdXQgPSBqc29uLmxvYWRzKG91dHB1dCkKICAgICAgICAgICAgcmV0dXJuICgK
            ICAgICAgICAgICAgICAgIGpzb25vdXRwdXRbImtleXMiXVswXVsiYWNjZXNzX2tleSJdLAogICAg
            ICAgICAgICAgICAganNvbm91dHB1dFsia2V5cyJdWzBdWyJzZWNyZXRfa2V5Il0sCiAgICAgICAg
            ICAgICAgICBGYWxzZSwKICAgICAgICAgICAgICAgICIiLAogICAgICAgICAgICApCgogICAgICAg
            ICMgc2VwYXJhdGVseSBhZGQgaW5mbz1yZWFkIGNhcHMgZm9yIHJndy1lbmRwb2ludCBpcCB2YWxp
            ZGF0aW9uCiAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkID0gVHJ1ZQogICAgICAgIGNtZCA9IFsK
            ICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAgICAgICAgICAiY2FwcyIsCiAgICAgICAg
            ICAgICJhZGQiLAogICAgICAgICAgICAiLS11aWQiLAogICAgICAgICAgICBzZWxmLkVYVEVSTkFM
            X1JHV19BRE1JTl9PUFNfVVNFUl9OQU1FLAogICAgICAgICAgICAiLS1jYXBzIiwKICAgICAgICAg
            ICAgImluZm89cmVhZCIsCiAgICAgICAgICAgICItLXJndy1yZWFsbSIsCiAgICAgICAgICAgIHNl
            bGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUsCiAgICAgICAgICAgICItLXJndy16b25lZ3Jv
            dXAiLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSwKICAg
            ICAgICAgICAgIi0tcmd3LXpvbmUiLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196
            b25lX25hbWUsCiAgICAgICAgXQogICAgICAgIHRyeToKICAgICAgICAgICAgb3V0cHV0ID0gc3Vi
            cHJvY2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAg
            IGV4Y2VwdCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJvciBhcyBleGVjRXJyOgogICAgICAg
            ICAgICAjIGlmIHRoZSBjZXBoIHZlcnNpb24gbm90IHN1cHBvcnRlZCBmb3IgYWRkaW5nIGBpbmZv
            PXJlYWRgIGNhcChyZ3dfdmFsaWRhdGlvbikKICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAg
            ICAgImNvdWxkIG5vdCBhZGQgY2FwczogdW5hYmxlIHRvIGFkZCBjYXBzOiBpbmZvPXJlYWRcbiIK
            ICAgICAgICAgICAgICAgIGluIGV4ZWNFcnIuc3RkZXJyLmRlY29kZSgidXRmLTgiKQogICAgICAg
            ICAgICAgICAgYW5kIGV4ZWNFcnIucmV0dXJuY29kZSA9PSAyNDQKICAgICAgICAgICAgKToKICAg
            ICAgICAgICAgICAgIGluZm9fY2FwX3N1cHBvcnRlZCA9IEZhbHNlCiAgICAgICAgICAgIGVsc2U6
            CiAgICAgICAgICAgICAgICBlcnJfbXNnID0gKAogICAgICAgICAgICAgICAgICAgIGYiZmFpbGVk
            IHRvIGV4ZWN1dGUgY29tbWFuZCB7Y21kfS4gT3V0cHV0OiB7ZXhlY0Vyci5vdXRwdXR9LiAiCiAg
            ICAgICAgICAgICAgICAgICAgZiJDb2RlOiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtl
            eGVjRXJyLnN0ZGVycn0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICBzeXMuc3Rk
            ZXJyLndyaXRlKGVycl9tc2cpCiAgICAgICAgICAgICAgICByZXR1cm4gTm9uZSwgTm9uZSwgRmFs
            c2UsICItMSIKCiAgICAgICAganNvbm91dHB1dCA9IGpzb24ubG9hZHMob3V0cHV0KQogICAgICAg
            IHJldHVybiAoCiAgICAgICAgICAgIGpzb25vdXRwdXRbImtleXMiXVswXVsiYWNjZXNzX2tleSJd
            LAogICAgICAgICAgICBqc29ub3V0cHV0WyJrZXlzIl1bMF1bInNlY3JldF9rZXkiXSwKICAgICAg
            ICAgICAgaW5mb19jYXBfc3VwcG9ydGVkLAogICAgICAgICAgICAiIiwKICAgICAgICApCgogICAg
            ZGVmIHZhbGlkYXRlX3JiZF9wb29sKHNlbGYsIHBvb2xfbmFtZSk6CiAgICAgICAgaWYgbm90IHNl
            bGYuY2x1c3Rlci5wb29sX2V4aXN0cyhwb29sX25hbWUpOgogICAgICAgICAgICByYWlzZSBFeGVj
            dXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcG9v
            bCwgJ3twb29sX25hbWV9JywgZG9lcyBub3QgZXhpc3QiCiAgICAgICAgICAgICkKCiAgICBkZWYg
            aW5pdF9yYmRfcG9vbChzZWxmLCByYmRfcG9vbF9uYW1lKToKICAgICAgICBpZiBpc2luc3RhbmNl
            KHNlbGYuY2x1c3RlciwgRHVtbXlSYWRvcyk6CiAgICAgICAgICAgIHJldHVybgogICAgICAgIGlv
            Y3R4ID0gc2VsZi5jbHVzdGVyLm9wZW5faW9jdHgocmJkX3Bvb2xfbmFtZSkKICAgICAgICByYmRf
            aW5zdCA9IHJiZC5SQkQoKQogICAgICAgIHJiZF9pbnN0LnBvb2xfaW5pdChpb2N0eCwgVHJ1ZSkK
            CiAgICBkZWYgdmFsaWRhdGVfcmFkb3NfbmFtZXNwYWNlKHNlbGYpOgogICAgICAgIHJiZF9wb29s
            X25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgIHJhZG9z
            X25hbWVzcGFjZSA9IHNlbGYuX2FyZ19wYXJzZXIucmFkb3NfbmFtZXNwYWNlCiAgICAgICAgaWYg
            cmFkb3NfbmFtZXNwYWNlID09ICIiOgogICAgICAgICAgICByZXR1cm4KICAgICAgICBpZiByYWRv
            c19uYW1lc3BhY2UuaXNsb3dlcigpID09IEZhbHNlOgogICAgICAgICAgICByYWlzZSBFeGVjdXRp
            b25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmFkb3Mg
            TmFtZXNwYWNlLCAne3JhZG9zX25hbWVzcGFjZX0nLCAiCiAgICAgICAgICAgICAgICBmImNvbnRh
            aW5zIHVwcGVyIGNhc2UiCiAgICAgICAgICAgICkKICAgICAgICByYmRfaW5zdCA9IHJiZC5SQkQo
            KQogICAgICAgIGlvY3R4ID0gc2VsZi5jbHVzdGVyLm9wZW5faW9jdHgocmJkX3Bvb2xfbmFtZSkK
            ICAgICAgICBpZiByYmRfaW5zdC5uYW1lc3BhY2VfZXhpc3RzKGlvY3R4LCByYWRvc19uYW1lc3Bh
            Y2UpIGlzIEZhbHNlOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmFkb3MgTmFtZXNwYWNlLCAne3JhZG9z
            X25hbWVzcGFjZX0nLCAiCiAgICAgICAgICAgICAgICBmImlzIG5vdCBmb3VuZCBpbiB0aGUgcG9v
            bCAne3JiZF9wb29sX25hbWV9JyIKICAgICAgICAgICAgKQoKICAgIGRlZiBnZXRfb3JfY3JlYXRl
            X3N1YnZvbHVtZV9ncm91cChzZWxmLCBzdWJ2b2x1bWVfZ3JvdXAsIGNlcGhmc19maWxlc3lzdGVt
            X25hbWUpOgogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgImNlcGgiLAogICAgICAgICAgICAi
            ZnMiLAogICAgICAgICAgICAic3Vidm9sdW1lZ3JvdXAiLAogICAgICAgICAgICAiZ2V0cGF0aCIs
            CiAgICAgICAgICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgIHN1YnZvbHVt
            ZV9ncm91cCwKICAgICAgICBdCiAgICAgICAgdHJ5OgogICAgICAgICAgICBfID0gc3VicHJvY2Vz
            cy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2Vw
            dCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJvcjoKICAgICAgICAgICAgY21kID0gWwogICAg
            ICAgICAgICAgICAgImNlcGgiLAogICAgICAgICAgICAgICAgImZzIiwKICAgICAgICAgICAgICAg
            ICJzdWJ2b2x1bWVncm91cCIsCiAgICAgICAgICAgICAgICAiY3JlYXRlIiwKICAgICAgICAgICAg
            ICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgICAgICBzdWJ2b2x1bWVfZ3Jv
            dXAsCiAgICAgICAgICAgIF0KICAgICAgICAgICAgdHJ5OgogICAgICAgICAgICAgICAgXyA9IHN1
            YnByb2Nlc3MuY2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAgICAg
            ICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBm
            InN1YnZvbHVtZSBncm91cCB7c3Vidm9sdW1lX2dyb3VwfSBpcyBub3QgYWJsZSB0byBnZXQgY3Jl
            YXRlZCIKICAgICAgICAgICAgICAgICkKCiAgICBkZWYgcGluX3N1YnZvbHVtZSgKICAgICAgICBz
            ZWxmLCBzdWJ2b2x1bWVfZ3JvdXAsIGNlcGhmc19maWxlc3lzdGVtX25hbWUsIHBpbl90eXBlLCBw
            aW5fc2V0dGluZwogICAgKToKICAgICAgICBjbWQgPSBbCiAgICAgICAgICAgICJjZXBoIiwKICAg
            ICAgICAgICAgImZzIiwKICAgICAgICAgICAgInN1YnZvbHVtZWdyb3VwIiwKICAgICAgICAgICAg
            InBpbiIsCiAgICAgICAgICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgIHN1
            YnZvbHVtZV9ncm91cCwKICAgICAgICAgICAgcGluX3R5cGUsCiAgICAgICAgICAgIHBpbl9zZXR0
            aW5nLAogICAgICAgIF0KICAgICAgICB0cnk6CiAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNo
            ZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1
            YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJzdWJ2b2x1bWUgZ3JvdXAge3N1YnZv
            bHVtZV9ncm91cH0gaXMgbm90IGFibGUgdG8gZ2V0IHBpbm5lZCIKICAgICAgICAgICAgKQoKICAg
            IGRlZiBnZXRfcmd3X2ZzaWQoc2VsZiwgYmFzZV91cmwsIHZlcmlmeSk6CiAgICAgICAgYWNjZXNz
            X2tleSA9IHNlbGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX0FDQ0VTU19LRVkiXQogICAg
            ICAgIHNlY3JldF9rZXkgPSBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9TRUNSRVRf
            S0VZIl0KICAgICAgICByZ3dfZW5kcG9pbnQgPSBzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2lu
            dAogICAgICAgIGJhc2VfdXJsID0gYmFzZV91cmwgKyAiOi8vIiArIHJnd19lbmRwb2ludCArICIv
            YWRtaW4vaW5mbz8iCiAgICAgICAgcGFyYW1zID0geyJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAg
            cmVxdWVzdF91cmwgPSBiYXNlX3VybCArIHVybGVuY29kZShwYXJhbXMpCgogICAgICAgIHRyeToK
            ICAgICAgICAgICAgciA9IHJlcXVlc3RzLmdldCgKICAgICAgICAgICAgICAgIHJlcXVlc3RfdXJs
            LAogICAgICAgICAgICAgICAgYXV0aD1TM0F1dGgoYWNjZXNzX2tleSwgc2VjcmV0X2tleSwgcmd3
            X2VuZHBvaW50KSwKICAgICAgICAgICAgICAgIHZlcmlmeT12ZXJpZnksCiAgICAgICAgICAgICkK
            ICAgICAgICBleGNlcHQgcmVxdWVzdHMuZXhjZXB0aW9ucy5UaW1lb3V0OgogICAgICAgICAgICBz
            eXMuc3RkZXJyLndyaXRlKAogICAgICAgICAgICAgICAgZiJpbnZhbGlkIGVuZHBvaW50Oiwgbm90
            IGFibGUgdG8gY2FsbCBhZG1pbi1vcHMgYXBpe3Jnd19lbmRwb2ludH0iCiAgICAgICAgICAgICkK
            ICAgICAgICAgICAgcmV0dXJuICIiLCAiLTEiCiAgICAgICAgcjEgPSByLmpzb24oKQogICAgICAg
            IGlmIHIxIGlzIE5vbmUgb3IgcjEuZ2V0KCJpbmZvIikgaXMgTm9uZToKICAgICAgICAgICAgc3lz
            LnN0ZGVyci53cml0ZSgKICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHJndyBFbmRwb2lu
            dCwgJ3tzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludH0nLCBpcyBpbnZhbGlkLiIKICAgICAg
            ICAgICAgKQogICAgICAgICAgICByZXR1cm4gKAogICAgICAgICAgICAgICAgIiIsCiAgICAgICAg
            ICAgICAgICAiLTEiLAogICAgICAgICAgICApCgogICAgICAgIHJldHVybiByMVsiaW5mbyJdWyJz
            dG9yYWdlX2JhY2tlbmRzIl1bMF1bImNsdXN0ZXJfaWQiXSwgIiIKCiAgICBkZWYgdmFsaWRhdGVf
            cmd3X2VuZHBvaW50KHNlbGYsIGluZm9fY2FwX3N1cHBvcnRlZCk6CiAgICAgICAgIyBpZiB0aGUg
            J2NsdXN0ZXInIGluc3RhbmNlIGlzIGEgZHVtbXkgb25lLAogICAgICAgICMgZG9uJ3QgdHJ5IHRv
            IHJlYWNoIG91dCB0byB0aGUgZW5kcG9pbnQKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1
            c3RlciwgRHVtbXlSYWRvcyk6CiAgICAgICAgICAgIHJldHVybgoKICAgICAgICByZ3dfZW5kcG9p
            bnQgPSBzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludAoKICAgICAgICAjIHZhbGlkYXRlIHJn
            dyBlbmRwb2ludCBvbmx5IGlmIGlwIGFkZHJlc3MgaXMgcGFzc2VkCiAgICAgICAgaXBfdHlwZSA9
            IHNlbGYuX2ludmFsaWRfZW5kcG9pbnQocmd3X2VuZHBvaW50KQoKICAgICAgICAjIGNoZWNrIGlm
            IHRoZSByZ3cgZW5kcG9pbnQgaXMgcmVhY2hhYmxlCiAgICAgICAgY2VydCA9IE5vbmUKICAgICAg
            ICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNlci5yZ3dfc2tpcF90bHMgYW5kIHNlbGYudmFsaWRhdGVf
            cmd3X2VuZHBvaW50X3Rsc19jZXJ0KCk6CiAgICAgICAgICAgIGNlcnQgPSBzZWxmLl9hcmdfcGFy
            c2VyLnJnd190bHNfY2VydF9wYXRoCiAgICAgICAgYmFzZV91cmwsIHZlcmlmeSwgZXJyID0gc2Vs
            Zi5lbmRwb2ludF9kaWFsKHJnd19lbmRwb2ludCwgaXBfdHlwZSwgY2VydD1jZXJ0KQogICAgICAg
            IGlmIGVyciAhPSAiIjoKICAgICAgICAgICAgcmV0dXJuICItMSIKCiAgICAgICAgIyBjaGVjayBp
            ZiB0aGUgcmd3IGVuZHBvaW50IGJlbG9uZ3MgdG8gdGhlIHNhbWUgY2x1c3RlcgogICAgICAgICMg
            b25seSBjaGVjayBpZiBgaW5mb2AgY2FwIGlzIHN1cHBvcnRlZAogICAgICAgIGlmIGluZm9fY2Fw
            X3N1cHBvcnRlZDoKICAgICAgICAgICAgZnNpZCA9IHNlbGYuZ2V0X2ZzaWQoKQogICAgICAgICAg
            ICByZ3dfZnNpZCwgZXJyID0gc2VsZi5nZXRfcmd3X2ZzaWQoYmFzZV91cmwsIHZlcmlmeSkKICAg
            ICAgICAgICAgaWYgZXJyID09ICItMSI6CiAgICAgICAgICAgICAgICByZXR1cm4gIi0xIgogICAg
            ICAgICAgICBpZiBmc2lkICE9IHJnd19mc2lkOgogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53
            cml0ZSgKICAgICAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCByZ3cgRW5kcG9pbnQsICd7
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfZW5kcG9pbnR9JywgaXMgaW52YWxpZC4gV2UgYXJlIHZhbGlk
            YXRpbmcgYnkgY2FsbGluZyB0aGUgYWRtaW5vcHMgYXBpIHRocm91Z2ggcmd3LWVuZHBvaW50IGFu
            ZCB2YWxpZGF0aW5nIHRoZSBjbHVzdGVyX2lkICd7cmd3X2ZzaWR9JyBpcyBlcXVhbCB0byB0aGUg
            Y2VwaCBjbHVzdGVyIGZzaWQgJ3tmc2lkfSciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAg
            ICAgICByZXR1cm4gIi0xIgoKICAgICAgICAjIGNoZWNrIGlmIHRoZSByZ3cgZW5kcG9pbnQgcG9v
            bCBleGlzdAogICAgICAgICMgb25seSB2YWxpZGF0ZSBpZiByZ3dfcG9vbF9wcmVmaXggaXMgcGFz
            c2VkIGVsc2UgaXQgd2lsbCB0YWtlIGRlZmF1bHQgdmFsdWUgYW5kIHdlIGRvbid0IGNyZWF0ZSB0
            aGVzZSBkZWZhdWx0IHBvb2xzCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9w
            cmVmaXggIT0gImRlZmF1bHQiOgogICAgICAgICAgICByZ3dfcG9vbHNfdG9fdmFsaWRhdGUgPSBb
            CiAgICAgICAgICAgICAgICBmIntzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3
            Lm1ldGEiLAogICAgICAgICAgICAgICAgIi5yZ3cucm9vdCIsCiAgICAgICAgICAgICAgICBmIntz
            ZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmNvbnRyb2wiLAogICAgICAgICAg
            ICAgICAgZiJ7c2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5sb2ciLAogICAg
            ICAgICAgICBdCiAgICAgICAgICAgIGZvciBfcmd3X3Bvb2xfdG9fdmFsaWRhdGUgaW4gcmd3X3Bv
            b2xzX3RvX3ZhbGlkYXRlOgogICAgICAgICAgICAgICAgaWYgbm90IHNlbGYuY2x1c3Rlci5wb29s
            X2V4aXN0cyhfcmd3X3Bvb2xfdG9fdmFsaWRhdGUpOgogICAgICAgICAgICAgICAgICAgIHN5cy5z
            dGRlcnIud3JpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHBvb2ws
            ICd7X3Jnd19wb29sX3RvX3ZhbGlkYXRlfScsIGRvZXMgbm90IGV4aXN0IgogICAgICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgICAgICAgICByZXR1cm4gIi0xIgoKICAgICAgICByZXR1cm4g
            IiIKCiAgICBkZWYgdmFsaWRhdGVfcmd3X211bHRpc2l0ZShzZWxmLCByZ3dfbXVsdGlzaXRlX2Nv
            bmZpZ19uYW1lLCByZ3dfbXVsdGlzaXRlX2NvbmZpZyk6CiAgICAgICAgaWYgcmd3X211bHRpc2l0
            ZV9jb25maWcgIT0gIiI6CiAgICAgICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgICAgICJyYWRv
            c2d3LWFkbWluIiwKICAgICAgICAgICAgICAgIHJnd19tdWx0aXNpdGVfY29uZmlnLAogICAgICAg
            ICAgICAgICAgImdldCIsCiAgICAgICAgICAgICAgICAiLS1yZ3ctIiArIHJnd19tdWx0aXNpdGVf
            Y29uZmlnLAogICAgICAgICAgICAgICAgcmd3X211bHRpc2l0ZV9jb25maWdfbmFtZSwKICAgICAg
            ICAgICAgXQogICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBfID0gc3VicHJvY2Vzcy5j
            aGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAgICBleGNl
            cHQgc3VicHJvY2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3IgYXMgZXhlY0VycjoKICAgICAgICAgICAg
            ICAgIGVycl9tc2cgPSAoCiAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBj
            b21tYW5kIHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAgICAgICAgICAg
            ICAgICBmIkNvZGU6IHtleGVjRXJyLnJldHVybmNvZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJy
            fSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJy
            X21zZykKICAgICAgICAgICAgICAgIHJldHVybiAiLTEiCiAgICAgICAgcmV0dXJuICIiCgogICAg
            ZGVmIGNvbnZlcnRfY29tbWFfc2VwYXJhdGVkX3RvX2FycmF5KHNlbGYsIHZhbHVlKToKICAgICAg
            ICByZXR1cm4gdmFsdWUuc3BsaXQoIiwiKQoKICAgIGRlZiByYWlzZV9leGNlcHRpb25faWZfYW55
            X3RvcG9sb2d5X2ZsYWdfaXNfbWlzc2luZyhzZWxmKToKICAgICAgICBpZiAoCiAgICAgICAgICAg
            ICgKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMgIT0gIiIK
            ICAgICAgICAgICAgICAgIGFuZCAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCA9PSAiIgogICAgICAgICAgICAgICAgICAg
            IG9yIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfZmFpbHVyZV9kb21haW5fdmFsdWVzID09ICIi
            CiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICkKICAgICAgICAgICAgb3IgKAogICAgICAg
            ICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCAh
            PSAiIgogICAgICAgICAgICAgICAgYW5kICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdf
            cGFyc2VyLnRvcG9sb2d5X3Bvb2xzID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5f
            YXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMgPT0gIiIKICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgKQogICAgICAgICAgICBvciAoCiAgICAgICAgICAgICAgICBz
            ZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX3ZhbHVlcyAhPSAiIgogICAg
            ICAgICAgICAgICAgYW5kICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRv
            cG9sb2d5X3Bvb2xzID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5fYXJnX3BhcnNl
            ci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCA9PSAiIgogICAgICAgICAgICAgICAgKQog
            ICAgICAgICAgICApCiAgICAgICAgKToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVy
            ZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICJwcm92aWRlIGFsbCB0aGUgdG9wb2xvZ3kgZmxh
            Z3MgLS10b3BvbG9neS1wb29scywgLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi1sYWJlbCwgLS10
            b3BvbG9neS1mYWlsdXJlLWRvbWFpbi12YWx1ZXMiCiAgICAgICAgICAgICkKCiAgICBkZWYgdmFs
            aWRhdGVfdG9wb2xvZ3lfdmFsdWVzKHNlbGYsIHRvcG9sb2d5X3Bvb2xzLCB0b3BvbG9neV9mZCk6
            CiAgICAgICAgaWYgbGVuKHRvcG9sb2d5X3Bvb2xzKSAhPSBsZW4odG9wb2xvZ3lfZmQpOgogICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ZiJUaGUgcHJvdmlkZWQgdG9wb2xvZ3kgcG9vbHMsICd7dG9wb2xvZ3lfcG9vbHN9JywgYW5kICIK
            ICAgICAgICAgICAgICAgIGYidG9wb2xvZ3kgZmFpbHVyZSBkb21haW4sICd7dG9wb2xvZ3lfZmR9
            JywiCiAgICAgICAgICAgICAgICBmImFyZSBvZiBkaWZmZXJlbnQgbGVuZ3RoLCAne2xlbih0b3Bv
            bG9neV9wb29scyl9JyBhbmQgJ3tsZW4odG9wb2xvZ3lfZmQpfScgcmVzcGN0aXZlbHkiCiAgICAg
            ICAgICAgICkKICAgICAgICByZXR1cm4KCiAgICBkZWYgdmFsaWRhdGVfdG9wb2xvZ3lfcmJkX3Bv
            b2xzKHNlbGYsIHRvcG9sb2d5X3JiZF9wb29scyk6CiAgICAgICAgZm9yIHBvb2wgaW4gdG9wb2xv
            Z3lfcmJkX3Bvb2xzOgogICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3JiZF9wb29sKHBvb2wpCgog
            ICAgZGVmIGluaXRfdG9wb2xvZ3lfcmJkX3Bvb2xzKHNlbGYsIHRvcG9sb2d5X3JiZF9wb29scyk6
            CiAgICAgICAgZm9yIHBvb2wgaW4gdG9wb2xvZ3lfcmJkX3Bvb2xzOgogICAgICAgICAgICBzZWxm
            LmluaXRfcmJkX3Bvb2wocG9vbCkKCiAgICBkZWYgX2dlbl9vdXRwdXRfbWFwKHNlbGYpOgogICAg
            ICAgIGlmIHNlbGYub3V0X21hcDoKICAgICAgICAgICAgcmV0dXJuCiAgICAgICAgIyBzdXBwb3J0
            IGxlZ2FjeSBmbGFnIHdpdGggdXBncmFkZXMKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNs
            dXN0ZXJfbmFtZToKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1l
            ID0gc2VsZi5fYXJnX3BhcnNlci5jbHVzdGVyX25hbWUKICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            Lms4c19jbHVzdGVyX25hbWUgPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuazhzX2Ns
            dXN0ZXJfbmFtZS5sb3dlcigpCiAgICAgICAgKSAgIyBhbHdheXMgY29udmVydCBjbHVzdGVyIG5h
            bWUgdG8gbG93ZXJjYXNlIGNoYXJhY3RlcnMKICAgICAgICBzZWxmLnZhbGlkYXRlX3JiZF9wb29s
            KHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lKQogICAgICAgIHNlbGYuaW5pdF9y
            YmRfcG9vbChzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZSkKICAgICAgICBzZWxm
            LnZhbGlkYXRlX3JhZG9zX25hbWVzcGFjZSgpCiAgICAgICAgc2VsZi5fZXhjbHVkZWRfa2V5cy5h
            ZGQoIks4U19DTFVTVEVSX05BTUUiKQogICAgICAgIHNlbGYuZ2V0X2NlcGhmc19kYXRhX3Bvb2xf
            ZGV0YWlscygpCiAgICAgICAgc2VsZi5vdXRfbWFwWyJOQU1FU1BBQ0UiXSA9IHNlbGYuX2FyZ19w
            YXJzZXIubmFtZXNwYWNlCiAgICAgICAgc2VsZi5vdXRfbWFwWyJLOFNfQ0xVU1RFUl9OQU1FIl0g
            PSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBb
            IlJPT0tfRVhURVJOQUxfRlNJRCJdID0gc2VsZi5nZXRfZnNpZCgpCiAgICAgICAgc2VsZi5vdXRf
            bWFwWyJST09LX0VYVEVSTkFMX1VTRVJOQU1FIl0gPSBzZWxmLnJ1bl9hc191c2VyCiAgICAgICAg
            c2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0NFUEhfTU9OX0RBVEEiXSA9IHNlbGYuZ2V0X2Nl
            cGhfZXh0ZXJuYWxfbW9uX2RhdGEoKQogICAgICAgIHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5B
            TF9VU0VSX1NFQ1JFVCJdID0gc2VsZi5jcmVhdGVfY2hlY2tlcktleSgKICAgICAgICAgICAgImNs
            aWVudC5oZWFsdGhjaGVja2VyIgogICAgICAgICkKICAgICAgICBzZWxmLm91dF9tYXBbIlJPT0tf
            RVhURVJOQUxfREFTSEJPQVJEX0xJTksiXSA9IHNlbGYuZ2V0X2NlcGhfZGFzaGJvYXJkX2xpbmso
            KQogICAgICAgICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VDUkVU
            Il0sCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FIl0s
            CiAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3VzZXIoImNsaWVudC5jc2kt
            cmJkLW5vZGUiKQogICAgICAgICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BS
            T1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9SQkRfUFJP
            VklTSU9ORVJfU0VDUkVUX05BTUUiXSwKICAgICAgICApID0gc2VsZi5jcmVhdGVfY2VwaENTSUtl
            eXJpbmdfdXNlcigiY2xpZW50LmNzaS1yYmQtcHJvdmlzaW9uZXIiKQogICAgICAgIHNlbGYub3V0
            X21hcFsiQ0VQSEZTX1BPT0xfTkFNRSJdID0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9w
            b29sX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBbIkNFUEhGU19NRVRBREFUQV9QT09MX05BTUUi
            XSA9ICgKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfbWV0YWRhdGFfcG9vbF9u
            YW1lCiAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSA9IHNl
            bGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgIHNlbGYub3V0X21h
            cFsiUkVTVFJJQ1RFRF9BVVRIX1BFUk1JU1NJT04iXSA9ICgKICAgICAgICAgICAgc2VsZi5fYXJn
            X3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbgogICAgICAgICkKICAgICAgICBzZWxm
            Lm91dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdID0gc2VsZi5fYXJnX3BhcnNlci5yYWRvc19uYW1l
            c3BhY2UKICAgICAgICBzZWxmLm91dF9tYXBbIlNVQlZPTFVNRV9HUk9VUCJdID0gc2VsZi5fYXJn
            X3BhcnNlci5zdWJ2b2x1bWVfZ3JvdXAKICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNf
            Tk9ERV9TRUNSRVQiXSA9ICIiCiAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJ
            U0lPTkVSX1NFQ1JFVCJdID0gIiIKICAgICAgICAjIGNyZWF0ZSBDZXBoRlMgbm9kZSBhbmQgcHJv
            dmlzaW9uZXIga2V5cmluZyBvbmx5IHdoZW4gTURTIGV4aXN0cwogICAgICAgIGlmIHNlbGYub3V0
            X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSBhbmQgc2VsZi5vdXRfbWFwWyJDRVBIRlNfUE9PTF9OQU1F
            Il06CiAgICAgICAgICAgICgKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhG
            U19OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZT
            X05PREVfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lL
            ZXlyaW5nX3VzZXIoImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiKQogICAgICAgICAgICAoCiAgICAg
            ICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUIl0s
            CiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VD
            UkVUX05BTUUiXSwKICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3Vz
            ZXIoImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIikKICAgICAgICAgICAgIyBjcmVhdGUg
            dGhlIGRlZmF1bHQgImNzaSIgc3Vidm9sdW1lZ3JvdXAKICAgICAgICAgICAgc2VsZi5nZXRfb3Jf
            Y3JlYXRlX3N1YnZvbHVtZV9ncm91cCgKICAgICAgICAgICAgICAgICJjc2kiLCBzZWxmLl9hcmdf
            cGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICAgICAgKQogICAgICAgICAgICAj
            IHBpbiB0aGUgZGVmYXVsdCAiY3NpIiBzdWJ2b2x1bWVncm91cAogICAgICAgICAgICBzZWxmLnBp
            bl9zdWJ2b2x1bWUoCiAgICAgICAgICAgICAgICAiY3NpIiwgc2VsZi5fYXJnX3BhcnNlci5jZXBo
            ZnNfZmlsZXN5c3RlbV9uYW1lLCAiZGlzdHJpYnV0ZWQiLCAiMSIKICAgICAgICAgICAgKQogICAg
            ICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlNVQlZPTFVNRV9HUk9VUCJdOgogICAgICAgICAgICAg
            ICAgc2VsZi5nZXRfb3JfY3JlYXRlX3N1YnZvbHVtZV9ncm91cCgKICAgICAgICAgICAgICAgICAg
            ICBzZWxmLl9hcmdfcGFyc2VyLnN1YnZvbHVtZV9ncm91cCwKICAgICAgICAgICAgICAgICAgICBz
            ZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgICAgICBzZWxmLnBpbl9zdWJ2b2x1bWUoCiAgICAgICAgICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci5zdWJ2b2x1bWVfZ3JvdXAsCiAgICAgICAgICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lLAogICAgICAgICAgICAgICAgICAg
            ICJkaXN0cmlidXRlZCIsCiAgICAgICAgICAgICAgICAgICAgIjEiLAogICAgICAgICAgICAgICAg
            KQogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1RMU19DRVJUIl0gPSAiIgogICAgICAgIHNlbGYu
            b3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdID0gIiIKICAgICAgICBzZWxmLm91dF9tYXBb
            Ik1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJdID0gIiIKICAgICAgICBpZiBub3Qgc2VsZi5fYXJn
            X3BhcnNlci5za2lwX21vbml0b3JpbmdfZW5kcG9pbnQ6CiAgICAgICAgICAgICgKICAgICAgICAg
            ICAgICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdLAogICAgICAgICAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UX1BPUlQiXSwKICAgICAgICAgICAg
            KSA9IHNlbGYuZ2V0X2FjdGl2ZV9hbmRfc3RhbmRieV9tZ3JzKCkKICAgICAgICBzZWxmLm91dF9t
            YXBbIlJCRF9QT09MX05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1l
            CiAgICAgICAgc2VsZi5vdXRfbWFwWyJSQkRfTUVUQURBVEFfRUNfUE9PTF9OQU1FIl0gPSAoCiAg
            ICAgICAgICAgIHNlbGYudmFsaWRhdGVfcmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZSgpCiAgICAg
            ICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9PTFMiXSA9IHNlbGYuX2FyZ19w
            YXJzZXIudG9wb2xvZ3lfcG9vbHMKICAgICAgICBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxV
            UkVfRE9NQUlOX0xBQkVMIl0gPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xv
            Z3lfZmFpbHVyZV9kb21haW5fbGFiZWwKICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJU
            T1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiXSA9ICgKICAgICAgICAgICAgc2VsZi5fYXJn
            X3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMKICAgICAgICApCiAgICAgICAg
            aWYgKAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X3Bvb2xzICE9ICIiCiAg
            ICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX2xh
            YmVsICE9ICIiCiAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1
            cmVfZG9tYWluX3ZhbHVlcyAhPSAiIgogICAgICAgICk6CiAgICAgICAgICAgIHNlbGYudmFsaWRh
            dGVfdG9wb2xvZ3lfdmFsdWVzKAogICAgICAgICAgICAgICAgc2VsZi5jb252ZXJ0X2NvbW1hX3Nl
            cGFyYXRlZF90b19hcnJheShzZWxmLm91dF9tYXBbIlRPUE9MT0dZX1BPT0xTIl0pLAogICAgICAg
            ICAgICAgICAgc2VsZi5jb252ZXJ0X2NvbW1hX3NlcGFyYXRlZF90b19hcnJheSgKICAgICAgICAg
            ICAgICAgICAgICBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX1ZBTFVFUyJd
            CiAgICAgICAgICAgICAgICApLAogICAgICAgICAgICApCiAgICAgICAgICAgIHNlbGYudmFsaWRh
            dGVfdG9wb2xvZ3lfcmJkX3Bvb2xzKAogICAgICAgICAgICAgICAgc2VsZi5jb252ZXJ0X2NvbW1h
            X3NlcGFyYXRlZF90b19hcnJheShzZWxmLm91dF9tYXBbIlRPUE9MT0dZX1BPT0xTIl0pCiAgICAg
            ICAgICAgICkKICAgICAgICAgICAgc2VsZi5pbml0X3RvcG9sb2d5X3JiZF9wb29scygKICAgICAg
            ICAgICAgICAgIHNlbGYuY29udmVydF9jb21tYV9zZXBhcmF0ZWRfdG9fYXJyYXkoc2VsZi5vdXRf
            bWFwWyJUT1BPTE9HWV9QT09MUyJdKQogICAgICAgICAgICApCiAgICAgICAgZWxzZToKICAgICAg
            ICAgICAgc2VsZi5yYWlzZV9leGNlcHRpb25faWZfYW55X3RvcG9sb2d5X2ZsYWdfaXNfbWlzc2lu
            ZygpCgogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1BPT0xfUFJFRklYIl0gPSBzZWxmLl9hcmdf
            cGFyc2VyLnJnd19wb29sX3ByZWZpeAogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX0VORFBPSU5U
            Il0gPSAiIgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50OgogICAgICAg
            ICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgICAgICBzZWxmLmNy
            ZWF0ZV9yZ3dfYWRtaW5fb3BzX3VzZXIoKQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAg
            ICAgaWYgKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25h
            bWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9u
            ZWdyb3VwX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfem9uZV9uYW1lICE9ICIiCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAg
            ICAgIGVyciA9IHNlbGYudmFsaWRhdGVfcmd3X211bHRpc2l0ZSgKICAgICAgICAgICAgICAgICAg
            ICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwgInJlYWxtIgogICAgICAgICAg
            ICAgICAgICAgICkKICAgICAgICAgICAgICAgICAgICBlcnIgPSBzZWxmLnZhbGlkYXRlX3Jnd19t
            dWx0aXNpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pv
            bmVncm91cF9uYW1lLCAiem9uZWdyb3VwIgogICAgICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgICAgICAgICBlcnIgPSBzZWxmLnZhbGlkYXRlX3Jnd19tdWx0aXNpdGUoCiAgICAgICAgICAg
            ICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVfbmFtZSwgInpvbmUiCiAgICAg
            ICAgICAgICAgICAgICAgKQoKICAgICAgICAgICAgICAgIGlmICgKICAgICAgICAgICAgICAgICAg
            ICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lID09ICIiCiAgICAgICAgICAgICAgICAg
            ICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lID09ICIiCiAgICAgICAg
            ICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVfbmFtZSA9PSAiIgogICAg
            ICAgICAgICAgICAgKSBvciAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5y
            Z3dfcmVhbG1fbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFy
            c2VyLnJnd196b25lZ3JvdXBfbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd196b25lX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICk6CiAgICAg
            ICAgICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJH
            V19BRE1JTl9PUFNfVVNFUl9BQ0NFU1NfS0VZIl0sCiAgICAgICAgICAgICAgICAgICAgICAgIHNl
            bGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX1NFQ1JFVF9LRVkiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkLAogICAgICAgICAgICAgICAgICAgICAgICBl
            cnIsCiAgICAgICAgICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRlX3Jnd19hZG1pbl9vcHNfdXNl
            cigpCiAgICAgICAgICAgICAgICAgICAgZXJyID0gc2VsZi52YWxpZGF0ZV9yZ3dfZW5kcG9pbnQo
            aW5mb19jYXBfc3VwcG9ydGVkKQogICAgICAgICAgICAgICAgICAgIGlmIHNlbGYuX2FyZ19wYXJz
            ZXIucmd3X3Rsc19jZXJ0X3BhdGg6CiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYub3V0X21h
            cFsiUkdXX1RMU19DRVJUIl0gPSAoCiAgICAgICAgICAgICAgICAgICAgICAgICAgICBzZWxmLnZh
            bGlkYXRlX3Jnd19lbmRwb2ludF90bHNfY2VydCgpCiAgICAgICAgICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgICAgICAgICAjIGlmIHRoZXJlIGlzIG5vIGVycm9yLCBzZXQgdGhlIFJHV19F
            TkRQT0lOVAogICAgICAgICAgICAgICAgICAgIGlmIGVyciAhPSAiLTEiOgogICAgICAgICAgICAg
            ICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdID0gc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfZW5kcG9pbnQKICAgICAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICAgICAg
            ZXJyID0gIlBsZWFzZSBwcm92aWRlIGFsbCB0aGUgUkdXIG11bHRpc2l0ZSBwYXJhbWV0ZXJzIG9y
            IG5vbmUgb2YgdGhlbSIKICAgICAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKGVycikK
            CiAgICBkZWYgZ2VuX3NoZWxsX291dChzZWxmKToKICAgICAgICBzZWxmLl9nZW5fb3V0cHV0X21h
            cCgpCiAgICAgICAgc2hPdXRJTyA9IFN0cmluZ0lPKCkKICAgICAgICBmb3IgaywgdiBpbiBzZWxm
            Lm91dF9tYXAuaXRlbXMoKToKICAgICAgICAgICAgaWYgdiBhbmQgayBub3QgaW4gc2VsZi5fZXhj
            bHVkZWRfa2V5czoKICAgICAgICAgICAgICAgIHNoT3V0SU8ud3JpdGUoZiJleHBvcnQge2t9PXt2
            fXtMSU5FU0VQfSIpCiAgICAgICAgc2hPdXQgPSBzaE91dElPLmdldHZhbHVlKCkKICAgICAgICBz
            aE91dElPLmNsb3NlKCkKICAgICAgICByZXR1cm4gc2hPdXQKCiAgICBkZWYgZ2VuX2pzb25fb3V0
            KHNlbGYpOgogICAgICAgIHNlbGYuX2dlbl9vdXRwdXRfbWFwKCkKICAgICAgICBpZiBzZWxmLl9h
            cmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiAiIgogICAgICAgIGpzb25fb3V0
            ID0gWwogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFtZSI6ICJyb29rLWNlcGgtbW9u
            LWVuZHBvaW50cyIsCiAgICAgICAgICAgICAgICAia2luZCI6ICJDb25maWdNYXAiLAogICAgICAg
            ICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiBzZWxmLm91dF9t
            YXBbIlJPT0tfRVhURVJOQUxfQ0VQSF9NT05fREFUQSJdLAogICAgICAgICAgICAgICAgICAgICJt
            YXhNb25JZCI6ICIwIiwKICAgICAgICAgICAgICAgICAgICAibWFwcGluZyI6ICJ7fSIsCiAgICAg
            ICAgICAgICAgICB9LAogICAgICAgICAgICB9LAogICAgICAgICAgICB7CiAgICAgICAgICAgICAg
            ICAibmFtZSI6ICJyb29rLWNlcGgtbW9uIiwKICAgICAgICAgICAgICAgICJraW5kIjogIlNlY3Jl
            dCIsCiAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAiYWRtaW4t
            c2VjcmV0IjogImFkbWluLXNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImZzaWQiOiBzZWxm
            Lm91dF9tYXBbIlJPT0tfRVhURVJOQUxfRlNJRCJdLAogICAgICAgICAgICAgICAgICAgICJtb24t
            c2VjcmV0IjogIm1vbi1zZWNyZXQiLAogICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgfSwK
            ICAgICAgICAgICAgewogICAgICAgICAgICAgICAgIm5hbWUiOiAicm9vay1jZXBoLW9wZXJhdG9y
            LWNyZWRzIiwKICAgICAgICAgICAgICAgICJraW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAg
            ICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJS
            T09LX0VYVEVSTkFMX1VTRVJOQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgInVzZXJLZXkiOiBz
            ZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAg
            IH0sCiAgICAgICAgICAgIH0sCiAgICAgICAgXQoKICAgICAgICAjIGlmICdNT05JVE9SSU5HX0VO
            RFBPSU5UJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ21vbml0b3JpbmctZW5kcG9pbnQnIHRvIENs
            dXN0ZXIKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19F
            TkRQT0lOVCJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9J
            TlRfUE9SVCJdCiAgICAgICAgKToKICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAg
            ICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjogIm1vbml0b3JpbmctZW5kcG9p
            bnQiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIkNlcGhDbHVzdGVyIiwKICAgICAgICAg
            ICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgIk1vbml0b3JpbmdF
            bmRwb2ludCI6IHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdLAogICAgICAgICAg
            ICAgICAgICAgICAgICAiTW9uaXRvcmluZ1BvcnQiOiBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdf
            RU5EUE9JTlRfUE9SVCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9
            CiAgICAgICAgICAgICkKCiAgICAgICAgIyBpZiAnQ1NJX1JCRF9OT0RFX1NFQ1JFVCcgZXhpc3Rz
            LCB0aGVuIG9ubHkgYWRkICdyb29rLWNzaS1yYmQtcHJvdmlzaW9uZXInIFNlY3JldAogICAgICAg
            IGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VDUkVUIl0KICAg
            ICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FIl0KICAg
            ICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAg
            ICAgICAgICAgICAgICAgICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9O
            T0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQi
            LAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAi
            dXNlcklEIjogc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VDUkVUX05BTUUiXSwKICAgICAg
            ICAgICAgICAgICAgICAgICAgInVzZXJLZXkiOiBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9T
            RUNSRVQiXSwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAg
            ICAgICApCiAgICAgICAgIyBpZiAnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQnIGV4aXN0cywg
            dGhlbiBvbmx5IGFkZCAncm9vay1jc2ktcmJkLXByb3Zpc2lvbmVyJyBTZWNyZXQKICAgICAgICBp
            ZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQi
            XQogICAgICAgICAgICBhbmQgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JF
            VF9OQU1FIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAg
            ICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21h
            cFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAg
            ICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAg
            ICAgICAgICAgICAgICAgICAgInVzZXJJRCI6IHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJ
            T05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNl
            bGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAg
            ICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBpZiAnQ1NJ
            X0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAncm9vay1j
            c2ktY2VwaGZzLXByb3Zpc2lvbmVyJyBTZWNyZXQKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNl
            bGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVQiXQogICAgICAgICAgICBh
            bmQgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0KICAg
            ICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAg
            ICAgICAgICAgICAgICAgICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhG
            U19QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQi
            OiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAg
            ICAgICAgICAgImFkbWluSUQiOiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJf
            U0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgImFkbWluS2V5Ijogc2VsZi5v
            dXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAg
            ICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdDU0lf
            Q0VQSEZTX05PREVfU0VDUkVUJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ3Jvb2stY3NpLWNlcGhm
            cy1ub2RlJyBTZWNyZXQKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJ
            X0NFUEhGU19OT0RFX1NFQ1JFVCJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9D
            RVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJdCiAgICAgICAgKToKICAgICAgICAgICAganNvbl9vdXQu
            YXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjogZiJy
            b29rLXtzZWxmLm91dF9tYXBbJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAg
            ICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0
            YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImFkbWluSUQiOiBzZWxmLm91dF9tYXBbIkNT
            SV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAiYWRt
            aW5LZXkiOiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVQiXSwKICAgICAgICAg
            ICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBp
            ZiAnUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSycgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdy
            b29rLWNlcGgtZGFzaGJvYXJkLWxpbmsnIFNlY3JldAogICAgICAgIGlmIHNlbGYub3V0X21hcFsi
            Uk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSyJdOgogICAgICAgICAgICBqc29uX291dC5hcHBl
            bmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAicm9vay1j
            ZXBoLWRhc2hib2FyZC1saW5rIiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQi
            LAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAi
            dXNlcklEIjogImNlcGgtZGFzaGJvYXJkLWxpbmsiLAogICAgICAgICAgICAgICAgICAgICAgICAi
            dXNlcktleSI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSyJdLAog
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAg
            ICAgICAjIGlmICdSQURPU19OQU1FU1BBQ0UnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCB0aGUgIlJB
            RE9TX05BTUVTUEFDRSIgbmFtZXNwYWNlCiAgICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLm91
            dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIlJF
            U1RSSUNURURfQVVUSF9QRVJNSVNTSU9OIl0KICAgICAgICAgICAgYW5kIG5vdCBzZWxmLm91dF9t
            YXBbIlJCRF9NRVRBREFUQV9FQ19QT09MX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpz
            b25fb3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFt
            ZSI6ICJyYWRvcy1uYW1lc3BhY2UiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIkNlcGhC
            bG9ja1Bvb2xSYWRvc05hbWVzcGFjZSIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAg
            ICAgICAgICAgICAgICAgICAgICAgICJyYWRvc05hbWVzcGFjZU5hbWUiOiBzZWxmLm91dF9tYXBb
            IlJBRE9TX05BTUVTUEFDRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYu
            b3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAg
            ICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAg
            ICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjogImNlcGgtcmJkLXJhZG9z
            LW5hbWVzcGFjZSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwK
            ICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgInBv
            b2wiOiBzZWxmLm91dF9tYXBbIlJCRF9QT09MX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAg
            ICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNyZXQtbmFtZSI6IGYicm9vay17
            c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAg
            ICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250cm9sbGVyLWV4cGFuZC1z
            ZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NF
            Q1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5p
            by9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRf
            Tk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAg
            ICAgIH0KICAgICAgICAgICAgKQogICAgICAgIGVsc2U6CiAgICAgICAgICAgIGlmIHNlbGYub3V0
            X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdOgogICAgICAgICAgICAgICAganNvbl9v
            dXQuYXBwZW5kKAogICAgICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAgICAg
            Im5hbWUiOiAiY2VwaC1yYmQiLAogICAgICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9y
            YWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAg
            ICAgICAgICAgICAgICAgICJkYXRhUG9vbCI6IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJd
            LAogICAgICAgICAgICAgICAgICAgICAgICAgICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIlJCRF9N
            RVRBREFUQV9FQ19QT09MX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJjc2ku
            c3RvcmFnZS5rOHMuaW8vcHJvdmlzaW9uZXItc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0
            X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNlY3Jl
            dC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVU
            X05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5p
            by9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRf
            Tk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAg
            ICAgICAgICAgICAgfQogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBlbHNlOgogICAgICAg
            ICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgICAgIHsKICAgICAgICAg
            ICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaC1yYmQiLAogICAgICAgICAgICAgICAgICAgICAg
            ICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAgICAgICAiZGF0YSI6
            IHsKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJwb29sIjogc2VsZi5vdXRfbWFwWyJSQkRf
            UE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2Uuazhz
            LmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9S
            QkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAg
            ICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYi
            cm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwK
            ICAgICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vbm9kZS1zdGFn
            ZS1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX05PREVfU0VDUkVU
            X05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgICAg
            IH0KICAgICAgICAgICAgICAgICkKCiAgICAgICAgIyBpZiAnVE9QT0xPR1lfUE9PTFMnLCAnVE9Q
            T0xPR1lfRkFJTFVSRV9ET01BSU5fTEFCRUwnLCAnVE9QT0xPR1lfRkFJTFVSRV9ET01BSU5fVkFM
            VUVTJyAgZXhpc3RzLAogICAgICAgICMgdGhlbiBvbmx5IGFkZCAndG9wb2xvZ3knIFN0b3JhZ2VD
            bGFzcwogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09M
            UyJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlO
            X0xBQkVMIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfRkFJTFVSRV9E
            T01BSU5fVkFMVUVTIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAg
            ICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaC1yYmQtdG9w
            b2xvZ3kiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAg
            ICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJ0b3BvbG9n
            eUZhaWx1cmVEb21haW5MYWJlbCI6IHNlbGYub3V0X21hcFsKICAgICAgICAgICAgICAgICAgICAg
            ICAgICAgICJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9MQUJFTCIKICAgICAgICAgICAgICAgICAg
            ICAgICAgXSwKICAgICAgICAgICAgICAgICAgICAgICAgInRvcG9sb2d5RmFpbHVyZURvbWFpblZh
            bHVlcyI6IHNlbGYub3V0X21hcFsKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJUT1BPTE9H
            WV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiCiAgICAgICAgICAgICAgICAgICAgICAgIF0sCiAgICAg
            ICAgICAgICAgICAgICAgICAgICJ0b3BvbG9neVBvb2xzIjogc2VsZi5vdXRfbWFwWyJUT1BPTE9H
            WV9QT09MUyJdLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3By
            b3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJP
            VklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0
            b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxm
            Lm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUi
            OiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAg
            ICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCgogICAg
            ICAgICMgaWYgJ0NFUEhGU19GU19OQU1FJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ2NlcGhmcycg
            U3RvcmFnZUNsYXNzCiAgICAgICAgaWYgc2VsZi5vdXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdOgog
            ICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAg
            ICAgICAgICAgIm5hbWUiOiAiY2VwaGZzIiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJT
            dG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAg
            ICAgICAgICAgICAiZnNOYW1lIjogc2VsZi5vdXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdLAogICAg
            ICAgICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYub3V0X21hcFsiQ0VQSEZTX1BPT0xfTkFN
            RSJdLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lv
            bmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9DRVBIRlNfUFJPVklT
            SU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3Jh
            Z2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91
            dF9tYXBbJ0NTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUi
            OiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FJ119IiwK
            ICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAg
            ICAgICAgIyBpZiAnUkdXX0VORFBPSU5UJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ2NlcGgtcmd3
            JyBTdG9yYWdlQ2xhc3MKICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdOgog
            ICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAg
            ICAgICAgICAgIm5hbWUiOiAiY2VwaC1yZ3ciLAogICAgICAgICAgICAgICAgICAgICJraW5kIjog
            IlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAg
            ICAgICAgICAgICAgICJlbmRwb2ludCI6IHNlbGYub3V0X21hcFsiUkdXX0VORFBPSU5UIl0sCiAg
            ICAgICAgICAgICAgICAgICAgICAgICJwb29sUHJlZml4Ijogc2VsZi5vdXRfbWFwWyJSR1dfUE9P
            TF9QUkVGSVgiXSwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAg
            ICAgICAgICApCiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsK
            ICAgICAgICAgICAgICAgICAgICAibmFtZSI6ICJyZ3ctYWRtaW4tb3BzLXVzZXIiLAogICAgICAg
            ICAgICAgICAgICAgICJraW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEi
            OiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJhY2Nlc3NLZXkiOiBzZWxmLm91dF9tYXBbIlJH
            V19BRE1JTl9PUFNfVVNFUl9BQ0NFU1NfS0VZIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJz
            ZWNyZXRLZXkiOiBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9TRUNSRVRfS0VZIl0s
            CiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQog
            ICAgICAgICMgaWYgJ1JHV19UTFNfQ0VSVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkIHRoZSAiY2Vw
            aC1yZ3ctdGxzLWNlcnQiIHNlY3JldAogICAgICAgIGlmIHNlbGYub3V0X21hcFsiUkdXX1RMU19D
            RVJUIl06CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAg
            ICAgICAgICAgICAgICAgICAibmFtZSI6ICJjZXBoLXJndy10bHMtY2VydCIsCiAgICAgICAgICAg
            ICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsK
            ICAgICAgICAgICAgICAgICAgICAgICAgImNlcnQiOiBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VS
            VCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAg
            ICkKCiAgICAgICAgcmV0dXJuIGpzb24uZHVtcHMoanNvbl9vdXQpICsgTElORVNFUAoKICAgIGRl
            ZiB1cGdyYWRlX3VzZXJzX3Blcm1pc3Npb25zKHNlbGYpOgogICAgICAgIHVzZXJzID0gWwogICAg
            ICAgICAgICAiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIsCiAgICAgICAgICAgICJjbGllbnQuY3Np
            LWNlcGhmcy1wcm92aXNpb25lciIsCiAgICAgICAgICAgICJjbGllbnQuY3NpLXJiZC1ub2RlIiwK
            ICAgICAgICAgICAgImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwKICAgICAgICAgICAgImNs
            aWVudC5oZWFsdGhjaGVja2VyIiwKICAgICAgICBdCiAgICAgICAgaWYgc2VsZi5ydW5fYXNfdXNl
            ciAhPSAiIiBhbmQgc2VsZi5ydW5fYXNfdXNlciBub3QgaW4gdXNlcnM6CiAgICAgICAgICAgIHVz
            ZXJzLmFwcGVuZChzZWxmLnJ1bl9hc191c2VyKQogICAgICAgIGZvciB1c2VyIGluIHVzZXJzOgog
            ICAgICAgICAgICBzZWxmLnVwZ3JhZGVfdXNlcl9wZXJtaXNzaW9ucyh1c2VyKQoKICAgIGRlZiBn
            ZXRfcmd3X3Bvb2xfbmFtZV9kdXJpbmdfdXBncmFkZShzZWxmLCB1c2VyLCBjYXBzKToKICAgICAg
            ICBpZiB1c2VyID09ICJjbGllbnQuaGVhbHRoY2hlY2tlciI6CiAgICAgICAgICAgICMgd2hlbiBh
            ZG1pbiBoYXMgbm90IHByb3ZpZGVkIHJndyBwb29sIG5hbWUgZHVyaW5nIHVwZ3JhZGUsCiAgICAg
            ICAgICAgICMgZ2V0IHRoZSByZ3cgcG9vbCBuYW1lIGZyb20gY2xpZW50LmhlYWx0aGNoZWNrZXIg
            dXNlciB3aGljaCB3YXMgdXNlZCBkdXJpbmcgY29ubmVjdGlvbgogICAgICAgICAgICBpZiBub3Qg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXg6CiAgICAgICAgICAgICAgICAjIFRvIGdl
            dCB2YWx1ZSAnZGVmYXVsdCcgd2hpY2ggaXMgcmd3IHBvb2wgbmFtZSBmcm9tICdhbGxvdyByd3gg
            cG9vbD1kZWZhdWx0LnJndy5tZXRhJwogICAgICAgICAgICAgICAgcGF0dGVybiA9IHIicG9vbD0o
            Lio/KVwucmd3XC5tZXRhIgogICAgICAgICAgICAgICAgbWF0Y2ggPSByZS5zZWFyY2gocGF0dGVy
            biwgY2FwcykKICAgICAgICAgICAgICAgIGlmIG1hdGNoOgogICAgICAgICAgICAgICAgICAgIHNl
            bGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4ID0gbWF0Y2guZ3JvdXAoMSkKICAgICAgICAg
            ICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAgICAgImZhaWxlZCB0byBnZXQgcmd3IHBvb2wg
            bmFtZSBmb3IgdXBncmFkZSIKICAgICAgICAgICAgICAgICAgICApCgogICAgZGVmIHVwZ3JhZGVf
            dXNlcl9wZXJtaXNzaW9ucyhzZWxmLCB1c2VyKToKICAgICAgICAjIGNoZWNrIHdoZXRoZXIgdGhl
            IGdpdmVuIHVzZXIgZXhpc3RzIG9yIG5vdAogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAi
            YXV0aCBnZXQiLCAiZW50aXR5IjogZiJ7dXNlcn0iLCAiZm9ybWF0IjogImpzb24ifQogICAgICAg
            IHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihj
            bWRfanNvbikKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgog
            ICAgICAgICAgICBwcmludChmInVzZXIge3VzZXJ9IG5vdCBmb3VuZCBmb3IgdXBncmFkaW5nLiIp
            CiAgICAgICAgICAgIHJldHVybgogICAgICAgIGV4aXN0aW5nX2NhcHMgPSBqc29uX291dFswXVsi
            Y2FwcyJdCiAgICAgICAgc2VsZi5nZXRfcmd3X3Bvb2xfbmFtZV9kdXJpbmdfdXBncmFkZSh1c2Vy
            LCBzdHIoZXhpc3RpbmdfY2FwcykpCiAgICAgICAgbmV3X2NhcCwgXyA9IHNlbGYuZ2V0X2NhcHNf
            YW5kX2VudGl0eSh1c2VyKQogICAgICAgIGNhcF9rZXlzID0gWyJtb24iLCAibWdyIiwgIm9zZCIs
            ICJtZHMiXQogICAgICAgIGNhcHMgPSBbXQogICAgICAgIGZvciBlYWNoQ2FwIGluIGNhcF9rZXlz
            OgogICAgICAgICAgICBjdXJfY2FwX3ZhbHVlcyA9IGV4aXN0aW5nX2NhcHMuZ2V0KGVhY2hDYXAs
            ICIiKQogICAgICAgICAgICBuZXdfY2FwX3ZhbHVlcyA9IG5ld19jYXAuZ2V0KGVhY2hDYXAsICIi
            KQogICAgICAgICAgICBjdXJfY2FwX3Blcm1fbGlzdCA9IFsKICAgICAgICAgICAgICAgIHguc3Ry
            aXAoKSBmb3IgeCBpbiBjdXJfY2FwX3ZhbHVlcy5zcGxpdCgiLCIpIGlmIHguc3RyaXAoKQogICAg
            ICAgICAgICBdCiAgICAgICAgICAgIG5ld19jYXBfcGVybV9saXN0ID0gWwogICAgICAgICAgICAg
            ICAgeC5zdHJpcCgpIGZvciB4IGluIG5ld19jYXBfdmFsdWVzLnNwbGl0KCIsIikgaWYgeC5zdHJp
            cCgpCiAgICAgICAgICAgIF0KICAgICAgICAgICAgIyBhcHBlbmQgbmV3X2NhcF9saXN0IHRvIGN1
            cl9jYXBfbGlzdCB0byBtYWludGFpbiB0aGUgb3JkZXIgb2YgY2FwcwogICAgICAgICAgICBjdXJf
            Y2FwX3Blcm1fbGlzdC5leHRlbmQobmV3X2NhcF9wZXJtX2xpc3QpCiAgICAgICAgICAgICMgZWxp
            bWluYXRlIGR1cGxpY2F0ZXMgd2l0aG91dCB1c2luZyAnc2V0JwogICAgICAgICAgICAjIHNldCBy
            ZS1vcmRlcnMgaXRlbXMgaW4gdGhlIGxpc3QgYW5kIHdlIGhhdmUgdG8ga2VlcCB0aGUgb3JkZXIK
            ICAgICAgICAgICAgbmV3X2NhcF9saXN0ID0gW10KICAgICAgICAgICAgW25ld19jYXBfbGlzdC5h
            cHBlbmQoeCkgZm9yIHggaW4gY3VyX2NhcF9wZXJtX2xpc3QgaWYgeCBub3QgaW4gbmV3X2NhcF9s
            aXN0XQogICAgICAgICAgICBleGlzdGluZ19jYXBzW2VhY2hDYXBdID0gIiwgIi5qb2luKG5ld19j
            YXBfbGlzdCkKICAgICAgICAgICAgaWYgZXhpc3RpbmdfY2Fwc1tlYWNoQ2FwXToKICAgICAgICAg
            ICAgICAgIGNhcHMuYXBwZW5kKGVhY2hDYXApCiAgICAgICAgICAgICAgICBjYXBzLmFwcGVuZChl
            eGlzdGluZ19jYXBzW2VhY2hDYXBdKQogICAgICAgIGNtZF9qc29uID0gewogICAgICAgICAgICAi
            cHJlZml4IjogImF1dGggY2FwcyIsCiAgICAgICAgICAgICJlbnRpdHkiOiB1c2VyLAogICAgICAg
            ICAgICAiY2FwcyI6IGNhcHMsCiAgICAgICAgICAgICJmb3JtYXQiOiAianNvbiIsCiAgICAgICAg
            fQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9q
            c29uX2dlbihjbWRfanNvbikKICAgICAgICBpZiByZXRfdmFsICE9IDA6CiAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIidhdXRoIGNh
            cHMge3VzZXJ9JyBjb21tYW5kIGZhaWxlZC5cbiBFcnJvcjoge2Vycl9tc2d9IgogICAgICAgICAg
            ICApCiAgICAgICAgcHJpbnQoZiJVcGRhdGVkIHVzZXIge3VzZXJ9IHN1Y2Nlc3NmdWxseS4iKQoK
            ICAgIGRlZiBtYWluKHNlbGYpOgogICAgICAgIGdlbmVyYXRlZF9vdXRwdXQgPSAiIgogICAgICAg
            IGlmIHNlbGYuX2FyZ19wYXJzZXIudXBncmFkZToKICAgICAgICAgICAgc2VsZi51cGdyYWRlX3Vz
            ZXJzX3Blcm1pc3Npb25zKCkKICAgICAgICBlbGlmIHNlbGYuX2FyZ19wYXJzZXIuZm9ybWF0ID09
            ICJqc29uIjoKICAgICAgICAgICAgZ2VuZXJhdGVkX291dHB1dCA9IHNlbGYuZ2VuX2pzb25fb3V0
            KCkKICAgICAgICBlbGlmIHNlbGYuX2FyZ19wYXJzZXIuZm9ybWF0ID09ICJiYXNoIjoKICAgICAg
            ICAgICAgZ2VuZXJhdGVkX291dHB1dCA9IHNlbGYuZ2VuX3NoZWxsX291dCgpCiAgICAgICAgZWxz
            ZToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAg
            ICAgICAgIGYiVW5zdXBwb3J0ZWQgZm9ybWF0OiB7c2VsZi5fYXJnX3BhcnNlci5mb3JtYXR9Igog
            ICAgICAgICAgICApCiAgICAgICAgcHJpbnQoZ2VuZXJhdGVkX291dHB1dCkKICAgICAgICBpZiBz
            ZWxmLm91dHB1dF9maWxlIGFuZCBnZW5lcmF0ZWRfb3V0cHV0OgogICAgICAgICAgICBmT3V0ID0g
            b3BlbihzZWxmLm91dHB1dF9maWxlLCBtb2RlPSJ3IiwgZW5jb2Rpbmc9IlVURi04IikKICAgICAg
            ICAgICAgZk91dC53cml0ZShnZW5lcmF0ZWRfb3V0cHV0KQogICAgICAgICAgICBmT3V0LmNsb3Nl
            KCkKCgojIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMKIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIE1BSU4gIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIwppZiBfX25hbWVfXyA9PSAiX19t
            YWluX18iOgogICAgcmpPYmogPSBSYWRvc0pTT04oKQogICAgdHJ5OgogICAgICAgIHJqT2JqLm1h
            aW4oKQogICAgZXhjZXB0IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24gYXMgZXJyOgogICAgICAg
            IHByaW50KGYiRXhlY3V0aW9uIEZhaWxlZDoge2Vycn0iKQogICAgICAgIHJhaXNlIGVycgogICAg
            ZXhjZXB0IEtleUVycm9yIGFzIGtFcnI6CiAgICAgICAgcHJpbnQoZiJLZXlFcnJvcjoge2tFcnJ9
            IikKICAgIGV4Y2VwdCBPU0Vycm9yIGFzIG9zRXJyOgogICAgICAgIHByaW50KGYiRXJyb3Igd2hp
            bGUgdHJ5aW5nIHRvIG91dHB1dCB0aGUgZGF0YToge29zRXJyfSIpCiAgICBmaW5hbGx5OgogICAg
            ICAgIHJqT2JqLnNodXRkb3duKCkK
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/red-hat-storage/rook
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Ceph cluster.
            displayName: Ceph Cluster
            kind: CephCluster
            name: cephclusters.ceph.rook.io
            version: v1
          - description: Represents a Ceph Block Pool.
            displayName: Ceph Block Pool
            kind: CephBlockPool
            name: cephblockpools.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store.
            displayName: Ceph Object Store
            kind: CephObjectStore
            name: cephobjectstores.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store User.
            displayName: Ceph Object Store User
            kind: CephObjectStoreUser
            name: cephobjectstoreusers.ceph.rook.io
            version: v1
          - description: Represents a cluster of Ceph NFS ganesha gateways.
            displayName: Ceph NFS
            kind: CephNFS
            name: cephnfses.ceph.rook.io
            version: v1
          - description: Represents a Ceph User.
            displayName: Ceph Client
            kind: CephClient
            name: cephclients.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem.
            displayName: Ceph Filesystem
            kind: CephFilesystem
            name: cephfilesystems.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem Mirror.
            displayName: Ceph Filesystem Mirror
            kind: CephFilesystemMirror
            name: cephfilesystemmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph RBD Mirror.
            displayName: Ceph RBD Mirror
            kind: CephRBDMirror
            name: cephrbdmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Realm.
            displayName: Ceph Object Store Realm
            kind: CephObjectRealm
            name: cephobjectrealms.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone Group.
            displayName: Ceph Object Store Zone Group
            kind: CephObjectZoneGroup
            name: cephobjectzonegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone.
            displayName: Ceph Object Store Zone
            kind: CephObjectZone
            name: cephobjectzones.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Notification.
            displayName: Ceph Bucket Notification
            kind: CephBucketNotification
            name: cephbucketnotifications.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Topic.
            displayName: Ceph Bucket Topic
            kind: CephBucketTopic
            name: cephbuckettopics.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem SubVolumeGroup.
            displayName: Ceph Filesystem SubVolumeGroup
            kind: CephFilesystemSubVolumeGroup
            name: cephfilesystemsubvolumegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph BlockPool Rados Namespace.
            displayName: Ceph BlockPool Rados Namespace
            kind: CephBlockPoolRadosNamespace
            name: cephblockpoolradosnamespaces.ceph.rook.io
            version: v1
          - description: Represents a Ceph COSI Driver.
            displayName: Ceph COSI Driver
            kind: CephCOSIDriver
            name: cephcosidrivers.ceph.rook.io
            version: v1
        description: |2

          The Rook-Ceph storage operator packages, deploys, manages, upgrades and scales Ceph storage for providing persistent storage to infrastructure services (Logging, Metrics, Registry) as well as stateful applications in Kubernetes clusters.

          ## Rook-Ceph Storage Operator

          Rook runs as a cloud-native service in Kubernetes clusters for optimal integration with applications in need of storage, and handles the heavy-lifting behind the scenes such as provisioning and management.
          Rook orchestrates battle-tested open-source storage technology Ceph, which has years of production deployments and runs some of the worlds largest clusters.

          Ceph is a massively scalable, software-defined, cloud native storage platform that offers block, file and object storage services.
          Ceph can be used to back a wide variety of applications including relational databases, NoSQL databases, CI/CD tool-sets, messaging, AI/ML and analytics applications.
          Ceph is a proven storage platform that backs some of the world's largest storage deployments and has a large vibrant open source community backing the project.

          ## Supported features
          * **High Availability and resiliency** - Ceph has no single point of failures (SPOF) and all its components work natively in a highly available fashion
          * **Data Protection** - Ceph periodically scrub for inconsistent objects and repair them if necessary, making sure your replicas are always coherent
          * **Consistent storage platform across hybrid cloud** - Ceph can be deployed anywhere (on-premise or bare metal) and thus offers a similar experience regardless
          * **Block, File & Object storage service** - Ceph can expose your data through several storage interfaces, solving all the application use cases
          * **Scale up/down** - addition and removal of storage is fully covered by the operator.
          * **Dashboard** - The Operator deploys a dashboard for monitoring and introspecting your cluster.

          ## Before you start
          https://rook.io/docs/rook/v1.0/k8s-pre-reqs.html
        displayName: Rook-Ceph
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rook
        - ceph
        - storage
        - object storage
        - open source
        - block storage
        - shared filesystem
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/rook
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:cf7f72469d25fb37301a8159406613673f8968c82af1035fd45a470e0ae8acfa
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:80c01633a6bfca44df37ac38f764035be6998fce965c8709e4f6d54d17eeaa7b
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6fee56851da3a87b9f7f477fc59e3385e118dc0b398861a8f13459c998a5af77
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c11a599bfd09b9b22888e03eabdbe1054d4049718a591d2a45bf0ed89800111e
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:82a4b481028195cfd75471656fe40c1cbbade4c7b2d170d6bf9422d9d3b84fa7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:693a60d6a1162659213ae554b1c6ee200b0300a7da72c01415044cf507acab49
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:0afcb516eb528857887c6ee1a9cba081dc8f6b7d47d9f26a0dbb2dca0cefe07c
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:8d8326c74795c9c6bd8f33a9f9e81d2304a70b1f9be3c2f8d50e76be4b0292ff
        version: 4.16.16-rhodf
      entries:
      - name: rook-ceph-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: rook-ceph-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: rook-ceph-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: rook-ceph-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: rook-ceph-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: rook-ceph-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: rook-ceph-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: rook-ceph-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: rook-ceph-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: rook-ceph-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: rook-ceph-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: rook-ceph-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: rook-ceph-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: rook-ceph-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: rook-ceph-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: rook-ceph-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: rook-ceph-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: rook-ceph-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: logic-operator-rhel8
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: logic-operator-rhel8.v1.36.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "host/namespace",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          createdAt: "2025-05-26T14:10:15Z"
          description: OpenShift Serverless Logic Kubernetes Operator for deploying
            workflow applications based on the CNCF Serverless Workflow specification
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-serverless-logic
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kubesmarts/kie-tools
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: Sonata Flow Build
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: Sonata Flow Cluster Platform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: Sonata Flow Platform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: Sonata Flow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          OpenShift Serverless Logic Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification v0.8](https://serverlessworkflow.io/).
        displayName: OpenShift Serverless Logic Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        - workflows
        links:
        - name: Documentation
          url: https://docs.openshift.com/serverless/
        - name: Source Repository
          url: https://github.com/kiegroup/kogito-serverless-operator/
        maintainers:
        - email: serverless-logic-team@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/logic-rhel8-operator@sha256:8d3682448ebdac3aeabb2d23842b7e67a252b95f959c408af805037f9728fd3c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4564ca3dc5bac80d6faddaf94c817fbbc270698a9399d8a21ee1005d85ceda56
        version: 1.36.0
      entries:
      - name: logic-operator-rhel8.v1.36.0
        version: 1.36.0
      - name: logic-operator-rhel8.v1.35.0
        version: 1.35.0
      - name: logic-operator-rhel8.v1.34.0
        version: 1.34.0
      - name: logic-operator-rhel8.v1.33.0
        version: 1.33.0
      name: alpha
    defaultChannel: alpha
    packageName: logic-operator-rhel8
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudwatchlogs-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudwatchlogs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudwatchlogs.services.k8s.aws/v1alpha1",
                "kind": "LogGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudwatchlogs-controller:1.1.1
          createdAt: "2025-09-19T20:51:16Z"
          description: AWS CloudWatch Logs controller is a service controller for
            managing CloudWatch Logs resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LogGroup represents the state of an AWS cloudwatchlogs LogGroup
              resource.
            displayName: LogGroup
            kind: LogGroup
            name: loggroups.cloudwatchlogs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudWatch Logs resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudWatch Logs**

          You can use Amazon CloudWatch Logs to monitor, store, and access your log files from Amazon Elastic Compute Cloud (Amazon EC2) instances, AWS CloudTrail, Route 53, and other sources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudWatch Logs
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudwatchlogs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudWatch Logs Developer Resources
          url: https://aws.amazon.com/cloudwatch/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudwatch maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudwatchlogs-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-cloudwatchlogs-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudwatchlogs-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudwatchlogs-controller.v1.0.14
        version: 1.0.14
      - name: ack-cloudwatchlogs-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudwatchlogs-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudwatchlogs-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudwatchlogs-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudwatchlogs-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudwatchlogs-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudwatchlogs-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudwatchlogs-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudwatchlogs-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudwatchlogs-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudwatchlogs-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudwatchlogs-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudwatchlogs-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudwatchlogs-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudwatchlogs-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudwatchlogs-controller.v0.0.7
        version: 0.0.7
      - name: ack-cloudwatchlogs-controller.v0.0.4
        version: 0.0.4
      - name: ack-cloudwatchlogs-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudwatchlogs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: TrendMicro
      provider-url: trendmicro.com
    name: visionone-containersecurity
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: visionone-containersecurity.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "visionone-containersecurity-sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          com.redhat.openshift.versions: v4.12-v4.20
          containerImage: public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
          createdAt: "2025-09-18T22:43:41Z"
          description: Trend Vision One™ Container Security delivers full lifecycle
            protection for containers with real-time threat detection, policy enforcement,
            compliance assurance, and rapid incident response.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trendmicro-system
          operatorhub.io/ui-metadata-max-k8s-version: "1.26"
          operators.openshift.io/valid-subscription: VisionOneContainerSecurity Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trendmicro/visionone-container-security-helm
          support: TrendMicro Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trend Vision One™ Container Security delivers full lifecycle
              protection for containers with real-time threat detection, policy enforcement,
              compliance assurance, and rapid incident response.
            displayName: VisionOneContainerSecurity
            kind: VisionOneContainerSecurity
            name: visiononecontainersecurities.container-security.trendmicro.com
            version: v1alpha1
        description: "[Trend Vision One™ Container Security](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-container-security)
          delivers comprehensive, end-to-end\n protection for containerized workloads
          across OpenShift and Kubernetes environments.\n It continuously monitors
          for threats, vulnerabilities, policy violations, and\n exposed secrets ensuring
          compliance and enabling rapid, automated response. Purpose-built\n for hybrid
          cloud, it instantly isolates compromised containers to prevent lateral\n
          movement and minimize impact. Seamlessly integrating into OpenShift DevSecOps\n
          pipelines, it empowers security teams to enforce policies while enabling
          developers\n to innovate and deploy with confidence without disrupting workflows
          or compromising\n agility.\n\n **Installation**\n \n - Use the recommended
          `trendmicro-system` namespace.\n - Refer to the [Trend Vision One™ Documentation](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator)
          for registration and installation steps.\n"
        displayName: Trend Vision One™ Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trendmicro
        - visionone
        - container-security
        links:
        - name: Documentation
          url: https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator
        maintainers:
        - email: support@trendmicro.com
          name: TrendMicro VisionOne Support Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: TrendMicro
          url: trendmicro.com
        relatedImages:
        - public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
        version: 0.0.2
      entries:
      - name: visionone-containersecurity.v0.0.2
        version: 0.0.2
      - name: visionone-containersecurity.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: visionone-containersecurity.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "visionone-containersecurity-sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          com.redhat.openshift.versions: v4.12-v4.20
          containerImage: public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
          createdAt: "2025-09-18T22:43:41Z"
          description: Trend Vision One™ Container Security delivers full lifecycle
            protection for containers with real-time threat detection, policy enforcement,
            compliance assurance, and rapid incident response.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trendmicro-system
          operatorhub.io/ui-metadata-max-k8s-version: "1.26"
          operators.openshift.io/valid-subscription: VisionOneContainerSecurity Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trendmicro/visionone-container-security-helm
          support: TrendMicro Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trend Vision One™ Container Security delivers full lifecycle
              protection for containers with real-time threat detection, policy enforcement,
              compliance assurance, and rapid incident response.
            displayName: VisionOneContainerSecurity
            kind: VisionOneContainerSecurity
            name: visiononecontainersecurities.container-security.trendmicro.com
            version: v1alpha1
        description: "[Trend Vision One™ Container Security](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-container-security)
          delivers comprehensive, end-to-end\n protection for containerized workloads
          across OpenShift and Kubernetes environments.\n It continuously monitors
          for threats, vulnerabilities, policy violations, and\n exposed secrets ensuring
          compliance and enabling rapid, automated response. Purpose-built\n for hybrid
          cloud, it instantly isolates compromised containers to prevent lateral\n
          movement and minimize impact. Seamlessly integrating into OpenShift DevSecOps\n
          pipelines, it empowers security teams to enforce policies while enabling
          developers\n to innovate and deploy with confidence without disrupting workflows
          or compromising\n agility.\n\n **Installation**\n \n - Use the recommended
          `trendmicro-system` namespace.\n - Refer to the [Trend Vision One™ Documentation](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator)
          for registration and installation steps.\n"
        displayName: Trend Vision One™ Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trendmicro
        - visionone
        - container-security
        links:
        - name: Documentation
          url: https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator
        maintainers:
        - email: support@trendmicro.com
          name: TrendMicro VisionOne Support Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: TrendMicro
          url: trendmicro.com
        relatedImages:
        - public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
        version: 0.0.2
      entries:
      - name: visionone-containersecurity.v0.0.2
        version: 0.0.2
      name: stable
    defaultChannel: stable
    packageName: visionone-containersecurity
    provider:
      name: TrendMicro
      url: trendmicro.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Docling Maintainers
      provider-url: https://docling-project.github.io/docling/
    name: docling-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: docling-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2025-05-19T20:20:14Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DoclingServe is the Schema for the doclingserves API
            displayName: Docling Serve
            kind: DoclingServe
            name: doclingserves.docling.github.io
            version: v1alpha1
        description: |-
          **Overview**

          The Docling Operator is responsible for managing the full lifecycle of Docling Serve instances in your Kubernetes or OpenShift container platforms.


          **Docling Serve**

          The Docling Serve application deploys a REST API for consuming and integrating Docling in your workflows.

          The API offers endpoints for multiple types of interaction, i.e. both synchronous and asynchronous endpoints, as well as status subscriptions via websockets.

          The options for the Docling conversion can be customized at runtime by the users (e.g. which OCR engine, which enrichment models, etc). Users can also choose which of the many DoclingDocument export formats (JSON, Markdown, HTML, DocTags, etc) to receive in the result response.

          Docling Serve allows to choose between a local compute engine, i.e. running Docling in the same container as the API server, or a distributed compute engine which leverages Kubeflow Pipelines to parallelize large-scale ingestion jobs.

          **Docling**

          Docling is a powerful library which simplifies document processing, parsing diverse formats — including advanced PDF understanding — and providing seamless integrations with the gen AI ecosystem.

          Learn more about the Docling Project by visiting the GitHub organization at [https://github.com/docling-project](https://github.com/docling-project).


          **Sample DoclingServe Custom Resources**


          To deploy with the `local` compute engine with the UI enabled and Route exposed:
          ```yaml
          apiVersion: docling.github.io/v1alpha1
          kind: DoclingServe
          metadata:
            labels:
              app.kubernetes.io/name: docling-operator
              app.kubernetes.io/managed-by: kustomize
            name: doclingserve-local-sample
          spec:
            apiServer:
              image: "quay.io/docling-project/docling-serve:latest"
              enableUI: true
              instances: 1
            engine:
              local:
                numWorkers: 2
            route:
              enabled: true
          ```

          To deploy with the `KFP` compute engine with the UI enabled and Route exposed. **Note: Kubeflow Pipeline must be installed and setup in the cluster to use this option.**:
          ```yaml
          apiVersion: docling.github.io/v1alpha1
          kind: DoclingServe
          metadata:
            labels:
              app.kubernetes.io/name: docling-operator
              app.kubernetes.io/managed-by: kustomize
            name: doclingserve-kfp-sample
          spec:
            apiServer:
              image: "quay.io/docling-project/docling-serve:latest"
              enableUI: true
              instances: 1
            engine:
              kfp:
                endpoint: "https://NAME.NAMESPACE.svc.cluster.local:8888"
            route:
              enabled: true
          ```

          **Additional Configurations**


          The CRD exposes `apiServer.configMapName` to enable additional configurations of the `docling-serve` workload. For a full list of configurations can be found on docling-serve's
          [Configuration](https://github.com/docling-project/docling-serve/blob/main/docs/configuration.md) page.
        displayName: Docling Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - document
        - processing
        links:
        - name: Docling Operator
          url: https://docling-project.github.io/docling/
        maintainers:
        - email: docling-technical-discuss@lists.lfaidata.foundation
          name: Docling Maintainers
        maturity: alpha
        provider:
          name: Docling Maintainers
          url: https://docling-project.github.io/docling/
        relatedImages:
        - quay.io/docling-project/docling-operator@sha256:200d8db51874c4b1e3d20d8d1423e9a4807843fbb29f514cd5d8ba872e47f4af
        version: 0.0.1
      entries:
      - name: docling-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: docling-operator
    provider:
      name: Docling Maintainers
      url: https://docling-project.github.io/docling/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubemq.io
      provider-url: ""
    name: kubemq-operator-marketplace
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubemq-operator-marketplace.1.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
              "apiVersion": "core.k8s.kubemq.io/v1beta1",
              "kind": "KubemqCluster",
              "metadata": {
                "name": "kubemq-cluster"
              },
              "spec": {
                "key": "<Your-KubeMQ-License-Key>",
                "replicas": 3
              }
            },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqConnector",
                "metadata": {
                  "name": "kubemq-bridges"
                },
                "spec": {
                  "type": "bridges",
                  "config": "bindings: null"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.connect.redhat.com/kubemq/kubemq-operator@sha256:3a9c87e462eb1caa81925dee72407f73fca41f19a3f9bcaf4e76e569db1dfbab
          createdAt: 2020-03-11 12:00:00+00:00
          description: Kubernetes Messaging Platform
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          support: support@kubemq.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kubemq Cluster
            displayName: Kubemq Cluster
            kind: KubemqCluster
            name: kubemqclusters.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Connector
            displayName: Kubemq Connector
            kind: KubemqConnector
            name: kubemqconnectors.core.k8s.kubemq.io
            version: v1beta1
        description: |2-

          ## Kubernetes Messaging Platform
          A message-based architecture for microservices providing flexible connectivity for hybrid environments: multi-cloud, on-premises, and at the edge. Powered by enterprise-grade message broker and message queue with ready to use connectors, bridges, and control center.

          Messaging platform components:
            - **KubeMQ Cluster** - Enterprise-grade message broker and message queue, scalable, high available, and secured.
            - **KubeMQ Targets Connector** - connects a KubeMQ cluster with external systems and cloud services.
            - **KubeMQ Bridges Connector** - bridge, replicate, aggregate, and transform messages between KubeMQ clusters.
            - **KubeMQ Sources Connector** - transform external systems and cloud services data to KubeMQ cluster messages.
          \n### Kubernetes Native
          Innovative and modern message queue and message broker in a lightweight container developed to run in Kubernetes, certified in the CNCF landscape, and connect natively to the cloud-native ecosystem.

          ### Ease of use

          Simple deployment in Kubernetes in less than 1 minute. Developer friendly by simple to use SDKs and elimination of the many developers and DevOps-centered challenges to define exchanges, brokers, channels, routes, and predefined topics

          ### All messaging patterns

          KubeMQ is available with all messaging patterns such as Queue, Stream, Pub/Sub, and RPC, Saving the need to maintain multiple messaging platforms for different use cases.

          ### Hybrid Environments

          KubeMQ enables the microservices from multiple environments to communicate and build one hybrid infrastructure solution across clouds, on-prem, and at the edge.

          ### KubeMQ Cluster key features

          - All-batteries included Messaging Queue Broker for the Kubernetes environment
          - Deployed with Operator for full life cycle operation
          - Blazing fast (written in Go), small and lightweight Docker container
          - Asynchronous and Synchronous messaging with support for  `At Most Once Delivery` and `At Least Once Delivery` models
          - Supports durable FIFO based Queue, Publish-Subscribe Events, Publish-Subscribe with Persistence (Events Store), RPC Command and Query messaging patterns
          - Supports gRPC, Rest and WebSocket Transport protocols with TLS support (both RPC and Stream modes)
          - Supports Access control Authorization and Authentication
          - Supports message masticating and smart routing
          - No Message broker configuration needed (i.e., queues, exchanges)
          - .Net, Java, Python, Go, and NodeJS SDKs

          ### KubeMQ Targets Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major services types (databases, cache, messaging, serverless, HTTP, etc.)
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-targets)

          ### KubeMQ Bridges Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Build Any Topology - connects KubeMQ clusters in 1:1, 1:n , n:1, n:n
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-bridges)

          ### KubeMQ Sources Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major messaging services such as Kafka, ActiveMQ, SQS, and more.
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-sources)

          ## Getting Started

          [Start with OpenShift](https://docs.kubemq.io/getting-started/create-cluster/openshift)

          ## Get Enterprise Trial License

          [Register](https://account.kubemq.io/login/register)

          ## Request a Demo

          [See KubeMQ in Action](https://kubemq.io/request-a-demo/)
        displayName: Kubemq Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - messaging
        - queue
        - broker
        - pubsub
        - rpc
        links:
        - name: Learn more about Kubemq
          url: https://kubemq.io/
        - name: Register for free license
          url: https://account.kubemq.io/login/register
        - name: Contact Sales
          url: https://kubemq.io/contact-sales/
        - name: Documentation
          url: https://docs.kubemq.io/
        - name: Get Help
          url: https://kubemq.io/contact-support/
        maintainers:
        - email: support@kubemq.io
          name: kubemq
        maturity: alpha
        provider:
          name: Kubemq.io
        relatedImages:
        - registry.connect.redhat.com/kubemq/kubemq-operator@sha256:3a9c87e462eb1caa81925dee72407f73fca41f19a3f9bcaf4e76e569db1dfbab
        - registry.connect.redhat.com/kubemq/kubemq-uni@sha256:51154b2a112c5cdc17ee8d338524c02ccd995f9386180b1cc232b7a875a23e73
        - registry.connect.redhat.com/kubemq/kubemq-targets@sha256:2a09e1718a03729a121728e4e3683e7e1f0690789d7f9857d76bebf1cc01d436
        - registry.connect.redhat.com/kubemq/kubemq-sources@sha256:5736575b6cae39d129346f83fe2d8ab9cdbd1c1d2a99a66b4b9c01631fb5db21
        - registry.connect.redhat.com/kubemq/kubemq-sources@sha256:8ed531e58c60760888ffb18165883386bda9cc5624645353406c94750f5c3008
        version: 1.8.6
      entries:
      - name: kubemq-operator-marketplace.1.8.6
        version: 1.8.6
      - name: kubemq-operator-marketplace.1.8.5
        version: 1.8.5
      - name: kubemq-operator-marketplace.v1.8.4
        version: 1.8.4
      - name: kubemq-operator-marketplace.v1.8.3
        version: 1.8.3
      name: alpha
    defaultChannel: alpha
    packageName: kubemq-operator-marketplace
    provider:
      name: Kubemq.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: candidate
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: fast
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: stable
    defaultChannel: stable
    packageName: stackgres-certified
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzounstructured-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzounstructured-operator.v2.0.101
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzounstructured.clusters.cambridgesemantics.com/v1",
                "kind": "AnzoUnstructured",
                "metadata": {
                  "name": "au01"
                },
                "spec": {
                  "auWorker": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-w",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-unstructured-worker@sha256:4db61bdf6dc8b42824a06d2f8d8aeecddde765d123b40a63a7bbf87845719fb2",
                                "name": "w",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  },
                  "msLeader": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-ms",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-microservices-leader@sha256:0579fd97e81e436eaff3b58c5395f13aa924940a3d51f0bdea5163d327620286",
                                "name": "ms",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/unstructured-operator@sha256:a47af068a7aa55c12be84c80d263feeed5597ecc434986b4f2fff782b568bb04
          createdAt: "2021-08-02"
          description: Kubernetes Operator for Anzo Unstructured
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzounstructured-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzounstructured-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoUnstructured is the Schema for the anzounstructureds
              API
            displayName: AnzoUnstructured Operator
            kind: AnzoUnstructured
            name: anzounstructureds.anzounstructured.clusters.cambridgesemantics.com
            version: v1
        description: |-
          The Anzo Unstructured Operator provides the way to install and configure an anzo unstructured setup on Red Hat K8S environment.
          Currently, this is possible only through existing Anzo installation.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v2.0.0/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers through [ Cambridge Semantics Customer Center ]( https://customercenter.cambridgesemantics.com/ ).
        displayName: Anzo Unstructured Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Database
        - Big Data
        links:
        - name: Anzo Unstructured Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v2.0.0/README.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/unstructured-operator@sha256:a47af068a7aa55c12be84c80d263feeed5597ecc434986b4f2fff782b568bb04
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-microservices-leader@sha256:0579fd97e81e436eaff3b58c5395f13aa924940a3d51f0bdea5163d327620286
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-unstructured-worker@sha256:4db61bdf6dc8b42824a06d2f8d8aeecddde765d123b40a63a7bbf87845719fb2
        version: 2.0.0
      entries:
      - name: anzounstructured-operator.v2.0.101
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anzounstructured-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
kind: List
metadata:
  resourceVersion: ""
apiVersion: v1
items:
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC
      provider-url: https://www.nec.com
    name: sanstoragecsi-operator-bundle
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sanstoragecsi-operator-bundle.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.nec.com/v1",
                "kind": "SANStorageCSI",
                "metadata": {
                  "name": "sanstoragecsi"
                },
                "spec": {
                  "driverName": "san.csi.nec.com",
                  "images": {
                    "csiDriver": "",
                    "externalAttacher": "registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:bd8262cc4b6c405e175c39623323dd98a74bbf70e6aa90851a72612056c184ff",
                    "externalProvisioner": "registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:85f1cc8f21568e7885e492b29660f29e09796fb23ebdf5fd526aa2e99d6b2fa8",
                    "livenessProbe": "registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:1a0d15012ba70955af754b03a7488b7c6ace813646ec7c3ce845f80597050c3c",
                    "nodeRegistrar": "registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:dc7f0992be6c2853dc0e798afda368e36005a94ad76b21fa2f09790621318864"
                  },
                  "parameters": {
                    "controller": {
                      "logLevel": 5
                    },
                    "node": {
                      "livenessProbePort": 9808,
                      "logLevel": 5,
                      "maxVolumesPerNode": 1024
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          createdAt: "2022-06-15T00:00:00Z"
          description: An operator for managing the NEC Storage M Series CSI Driver
          operatorframework.io/suggested-namespace: nec-san-csi-operator-system
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: NEC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SANStorageCSI is the Schema for the sanstoragecsis API
            displayName: NEC Storage M Series CSI Driver
            kind: SANStorageCSI
            name: sanstoragecsis.csi.nec.com
            version: v1
        description: |
          This operator deploys the NEC Storage M Series CSI Driver on OpenShift Container Platform.

          NEC Storage M Series CSI Driver provides persistent storage for stateful applications using NEC Storage M Series.
          Please refer to the documentation of the NEC Storage M Series CSI Driver for detail information about supported NEC Storage Platforms and supported CSI features.
        displayName: NEC Storage M Series CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - NEC
        - M Series
        - SAN Storage
        - CSI Driver
        links:
        - name: NEC Storage M Series
          url: https://www.nec.com/en/global/prod/storage/product/san/index.html
        maturity: stable
        provider:
          name: NEC
          url: https://www.nec.com
        relatedImages:
        - registry.connect.redhat.com/nec/sanstoragecsi-operator@sha256:6656c2b5036bdc013712ecb640bbbebe0e3e90cf625c73697dbb49e3e201e2ac
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:bd8262cc4b6c405e175c39623323dd98a74bbf70e6aa90851a72612056c184ff
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:85f1cc8f21568e7885e492b29660f29e09796fb23ebdf5fd526aa2e99d6b2fa8
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:1a0d15012ba70955af754b03a7488b7c6ace813646ec7c3ce845f80597050c3c
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:dc7f0992be6c2853dc0e798afda368e36005a94ad76b21fa2f09790621318864
        version: 1.1.0
      entries:
      - name: sanstoragecsi-operator-bundle.v1.1.0
        version: 1.1.0
      - name: sanstoragecsi-operator-bundle.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: sanstoragecsi-operator-bundle
    provider:
      name: NEC
      url: https://www.nec.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nxiq-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nxiq-operator-certified.v1.193.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusIQ",
                "metadata": {
                  "name": "example-nexusiq"
                },
                "spec": {
                  "deployment": {
                    "postStart": {
                      "command": null
                    }
                  },
                  "deploymentStrategy": {
                    "type": "Recreate"
                  },
                  "fullnameOverride": "",
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [],
                    "path": "/",
                    "tls": []
                  },
                  "iq": {
                    "adminPort": 8071,
                    "applicationPort": 8070,
                    "env": [
                      {
                        "name": "JAVA_OPTS",
                        "value": "-Djava.util.prefs.userRoot=$(SONATYPE_WORK)/javaprefs -Djavax.net.ssl.trustStore=/etc/nexus-iq-server/jks/keystore.jks"
                      }
                    ],
                    "configYaml": {
                      "sonatypeWork": "/sonatype-work",
                      "createSampleData": true,
                      "server": {
                        "adminConnectors": [
                          {
                            "port": 8071,
                            "type": "http"
                          }
                        ],
                        "applicationConnectors": [
                          {
                            "port": 8070,
                            "type": "http"
                          }
                        ],
                        "requestLog": {
                          "appenders": [
                            {
                              "type": "file",
                              "currentLogFilename": "/var/log/nexus-iq-server/request.log",
                              "logFormat": "%clientHost %l %user [%date] \"%requestURL\" %statusCode %bytesSent %elapsedTime \"%header{User-Agent}\"",
                              "archivedLogFilenamePattern": "/var/log/nexus-iq-server/request-%d.log.gz",
                              "archivedFileCount": 50
                            }
                          ]
                        }
                      },
                      "logging": {
                        "level": "DEBUG",
                        "loggers": {
                          "com.sonatype.insight.scan": "INFO",
                          "eu.medsea.mimeutil.MimeUtil2": "INFO",
                          "org.apache.http": "INFO",
                          "org.apache.http.wire": "ERROR",
                          "org.eclipse.birt.report.engine.layout.pdf.font.FontConfigReader": "WARN",
                          "org.eclipse.jetty": "INFO",
                          "org.apache.shiro.web.filter.authc.BasicHttpAuthenticationFilter": "INFO",
                          "com.sonatype.insight.audit": {
                            "appenders": [
                              {
                                "type": "file",
                                "currentLogFilename": "/var/log/nexus-iq-server/audit.log",
                                "archivedLogFilenamePattern": "/var/log/nexus-iq-server/audit-%d.log.gz",
                                "archivedFileCount": 50
                              }
                            ]
                          },
                          "com.sonatype.insight.policy.violation": {
                            "appenders": [
                              {
                                "type": "console",
                                "threshold": "INFO"
                              }
                            ]
                          }
                        },
                        "appenders": [
                          {
                            "type": "console",
                            "threshold": "INFO",
                            "logFormat": "%d{'yyyy-MM-dd HH:mm:ss,SSSZ'} %level [%thread] %X{username} %logger - %msg%n"
                          },
                          {
                            "type": "file",
                            "threshold": "ALL",
                            "logFormat": "%d{'yyyy-MM-dd HH:mm:ss,SSSZ'} %level [%thread] %X{username} %logger - %msg%n",
                            "currentLogFilename": "/var/log/nexus-iq-server/clm-server.log",
                            "archivedLogFilenamePattern": "/var/log/nexus-iq-server/clm-server-%d.log.gz",
                            "archivedFileCount": 50
                          }
                        ]
                      }
                    },
                    "imageName": "registry.connect.redhat.com/sonatype/nexus-iq-server@sha256:d65df6bdff1f32bf32a87e5cb38f5e9f98b75b81e9acccbffbd967e1f71691f0",
                    "imagePullPolicy": "IfNotPresent",
                    "imagePullSecret": "",
                    "licenseSecret": "",
                    "memory": "1Gi",
                    "name": "nxiq",
                    "certificates": {}
                  },
                  "nameOverride": "",
                  "persistence": {
                    "accessMode": "ReadWriteOnce",
                    "storageSize": "1Gi"
                  },
                  "service": {
                    "annotations": {},
                    "enabled": false,
                    "labels": {},
                    "ports": [
                      {
                        "name": "nexus-service",
                        "port": 80,
                        "targetPort": 80
                      }
                    ]
                  },
                  "serviceAccount": {
                    "create": true,
                    "name": null
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nxiq-operator-certified@sha256:59c52729189da87d9b5c3de8ef85326662d94c99835ab99d6e8d7fefdcf53ef0
          createdAt: "2020-07-20"
          description: |-
            Nexus Lifecycle is an open source governance platform that continuously identifies risk,
            automatically enforces policy, and provides visibility throughout the entire SDLC.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.193.0-1
          repository: https://github.com/sonatype/operator-iq
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus IQ Server
            displayName: NexusIQ
            kind: NexusIQ
            name: nexusiqs.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Lifecycle is an open source governance platform that continuously identifies risk,
          automatically enforces policy, and provides visibility throughout the entire SDLC.
          Developers leveraging open source require a solution that helps them make safer choices
          while still delivering at DevOps speed.
          Nexus Lifecycle empowers developers and security teams to choose the best components
          and govern open source usage, ensuring organizations continue to innovate with less risk.

          ## Core Capabilities

          * **Dependency Management**:
            Generate a precise software bill of materials for all applications
            to identify open source components and gather contextual direct
            and transitive dependency information for faster action.
          * **One-Click Remediation**:
            Upgrade to the best components and versions based on real-time intelligence.
            Available in your preferred IDE or a GitHub pull request.
          * **Automated Open Source Policy Enforcement**:
            Create custom security, license, and architectural policies based
            on application type or organization and contextually enforce those policies
            across every stage of the SDLC.
          * **Continuous Monitoring**:
            Real-time notifications of newly discovered defects,
            as well as vulnerabilities based on component,
            risk level or application/container affected.

          ## Extras

          * IDE Integrations: Eclipse, IntelliJ, and Visual Studio
          * Source Control Integrations: GitHub, GitLab, and Bitbucket
          * Chrome Browser Extension
          * Red Hat Clair for Container Scanning

          ## Limitations

          This operator will be released on a quarterly basis.

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus IQ Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage

          Once the server instance is created by the operator and running,
          you'll want to expose the service as you see fit:
          1. Create a Route to the new service for iq.applicationPort (8070).
          2. Visit the URL provided by the Route, login, and set new credentials.
            The default credentials are `admin`/`admin123`.

          The Nexus IQ Server can be further configured via the NexusIQ custom resource definition:

          | Parameter            | Description                                                  | Default           |
          | -------------------- | ------------------------------------------------------------ | ----------------- |
          | `iq.applicationPort` | Port of the application connector. Must match the value in the `configYaml` property | `8070`            |
          | `iq.adminPort`       | Port of the application connector. Must match the value in the `configYaml` property | `8071`            |
          | `iq.memory`          | The amount of RAM to allocate                                | `1Gi`             |
          | `iq.env`             | Customize the server environment, including `JAVA_OPTS`      | See example YAML show when ceating a NexusIQ. |
          | `iq.licenseSecret`   | The base-64 encoded license file to be installed at startup  | `""`              |
          | `iq.configYaml`      | A YAML block which will be used as a configuration block for IQ Server. | See example YAML shown when creating a NexusIQ. |
          | `iq.certificates`   | Add your own certificates for your own trusted Certificate Authorities with names and plain armored text. | `{}` |
          | `ingress.enabled`                           | Create an ingress for Nexus         | `true`                                  |
          | `ingress.annotations`                       | Annotations to enhance ingress configuration  | `{}`                          |
          | `ingress.tls.enabled`                       | Enable TLS                          | `true`                                 |
          | `ingress.tls.secretName`                    | Name of the secret storing TLS cert, `false` to use the Ingress' default certificate | `nexus-tls`                             |
          | `ingress.path`                              | Path for ingress rules. GCP users should set to `/*` | `/`                    |
          | `deploymentStrategy.type`                   | Deployment Strategy                 |  `Recreate` |
          | `initContainers`                            | Containers to run before starting the server | `nil` |
          | `deployment.postStart.command`              | Command to run after starting the IQ Server container  | `nil`                    |
          | `deployment.terminationGracePeriodSeconds` | Time to allow for clean shutdown                        | 120                      |
          | `persistence.storageClass` | The provisioner class                        | `-` (disables dynamic provisioning)            |
          | `persistence.storageSize` | The amount of drive space to allocate                        | `1Gi`             |
          | `persistence.accessMode` | Default access mode                        | `ReadWriteOnce`             |
          | `persistence.volumeConfiguration` | A YAML block to configure the persistent volume type. Defaults to `hostPath` which should not be used in production | `hostPath`             |

          ## Configuring IQ Server

          You can define the `config.yml` for IQ Server in your CRD on startup.
          It is the `iq.configYaml` property. For more details, see the [Configuring IQ Server](https://help.sonatype.com/iqserver/configuring) help page.

          ## Installing the License

          The license file can be installed via the UI when IQ server is running, or it can be done as a part of the deploy.
          If you leave the `licenseFile` field empty/commented, IQ Server will start and prompt you to manually install the license
          when you first enter the GUI.

          ### Installing the License Automatically
          To do it automatically, first encode your `.lic` file in Base 64 with no line breaks, eg:

          ```bash
          base64 --wrap=0 mylicense.lic > lic.base64
          ```

          Then add this value to your CRD file as `iq.licenseSecret`, eg:

          ```yaml
          iq:
            licenseSecret: bXkgc2FtcGxlIGxpY2Vuc2U=
          ```

          Specify the `licenseFile` path in `iq.configYaml` as:

          ```yaml
          iq:
            configYaml:
              sonatypeWork: /sonatype-work
              createSampleData: true
              # add this line and the `licenseSecret` above to autoconfigure licensing
              licenseFile: /etc/nexus-iq-license/license_lic
              server:
                applicationConnectors:
                  - type: http
                    port: 8070
                adminConnectors:
                  - type: http
                    port: 8071
          ```

          ## Adding Certificates for Your Own Trusted Certificate Authorities

          Specify a set of certificates to be imported into the Java trust store
          of the running IQ Server. More than one certificate can be provided.

          ```yaml
          iq:
            certificates:
              my-company-cert: |
                -----BEGIN CERTIFICATE-----
                ...
                -----END CERTIFICATE-----
          ```

          ## Running Commands Before Initialization of the Server

          If you need custom initialization, you can use standard Kubernetes
          [`initContainers`](https://kubernetes.io/docs/concepts/workloads/pods/init-containers/)
          to run any series of containers and commands you need.

          ```yaml
          initContainers:
            - name: an-init-container
              image: busybox:1.28
              command: ['/bin/sh', '-c', 'echo "initializing"']
          ```
        displayName: Nexus IQ Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - sonatype
        - licensing
        - policy
        - components
        links:
        - name: Nexus Lifecycle
          url: https://www.sonatype.com/product-nexus-lifecycle
        maintainers:
        - email: support@sonaype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nxiq-operator-certified@sha256:59c52729189da87d9b5c3de8ef85326662d94c99835ab99d6e8d7fefdcf53ef0
        - registry.connect.redhat.com/sonatype/nexus-iq-server@sha256:d65df6bdff1f32bf32a87e5cb38f5e9f98b75b81e9acccbffbd967e1f71691f0
        version: 1.193.0-1
      entries:
      - name: nxiq-operator-certified.v1.193.0-1
        version: 1.193.0-1
      name: stable
    defaultChannel: stable
    packageName: nxiq-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com/
    name: openstack-baremetal-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openstack-baremetal-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "baremetal.openstack.org/v1beta1",
                "kind": "OpenStackBaremetalSet",
                "metadata": {
                  "name": "openstackbaremetalset"
                },
                "spec": {
                  "baremetalHosts": {
                    "compute-0": "172.22.0.100/24",
                    "compute-1": "172.22.0.101/24"
                  },
                  "ctlplaneGateway": "172.22.0.3",
                  "ctlplaneInterface": "enp1s0",
                  "deploymentSSHSecret": "baremetalset-deploy-secret",
                  "domainName": "osptest.openstack.org",
                  "osImage": "edpm-hardened-uefi.qcow2",
                  "passwordSecret": "baremetalset-password-secret"
                }
              },
              {
                "apiVersion": "baremetal.openstack.org/v1beta1",
                "kind": "OpenStackProvisionServer",
                "metadata": {
                  "name": "openstackprovisionserver"
                },
                "spec": {
                  "interface": "enp1s0",
                  "osImage": "edpm-hardened-uefi.qcow2",
                  "port": 8080
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:07Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenStackBaremetalSet is the Schema for the openstackbaremetalsets
              API
            displayName: OpenStack BaremetalSet
            kind: OpenStackBaremetalSet
            name: openstackbaremetalsets.baremetal.openstack.org
            version: v1beta1
          - description: OpenStackProvisionServer used to serve custom images for
              baremetal provisioning with Metal3
            displayName: OpenStackProvisionServer
            kind: OpenStackProvisionServer
            name: openstackprovisionservers.baremetal.openstack.org
            version: v1beta1
        description: An operator for handling baremetal provisioning concerns for
          OpenStack
        displayName: OpenStack Baremetal Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cn-openstack
        - openstack
        links:
        - name: Openstack Baremetal Operator
          url: https://github.com/openstack-k8s-operators/openstack-baremetal-operator
        maturity: beta
        provider:
          name: Red Hat
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/ubi9/httpd-24@sha256:dbfbe3383e12d63df7178e4cdd5b1e6e51a84c781041b9192026ea6edb72fbf3
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/openstack-baremetal-rhel9-operator@sha256:a6c69c03e59faf50973d07ccea32b80cb9c46ceb60ed00ffe7311ca42d501296
        - registry.redhat.io/rhoso/edpm-hardened-uefi-rhel9@sha256:e4bfacb9d8de056a251bee16a66b69631b96fb19fb926d8e33acd379d40a92d6
        - registry.redhat.io/rhoso-operators/openstack-baremetal-agent-rhel9@sha256:8f6f8e99042710819aa639eae6513f36c4642190ecd85d58a2ed819f1b55e38f
        version: 1.0.6
      entries:
      - name: openstack-baremetal-operator.v1.0.6
        version: 1.0.6
      - name: openstack-baremetal-operator.v1.0.5
        version: 1.0.5
      - name: openstack-baremetal-operator.v1.0.4
        version: 1.0.4
      - name: openstack-baremetal-operator.v1.0.3
        version: 1.0.3
      - name: openstack-baremetal-operator.v1.0.2
        version: 1.0.2
      - name: openstack-baremetal-operator.v1.0.1
        version: 1.0.1
      - name: openstack-baremetal-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: openstack-baremetal-operator
    provider:
      name: Red Hat
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-msg-broker
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-msg-broker.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Messagebroker",
                "metadata": {
                  "name": "messagebroker-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "chiselServer": {
                    "enabled": false,
                    "port": 8080
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": false
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "natsCore": {
                    "reloader": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats-server-config-reloader",
                      "tag": "0.8.3-3747-30"
                    },
                    "server": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats",
                      "tag": "2.9.0-3747-21"
                    }
                  },
                  "natsHttpProxy": {
                    "enabled": false,
                    "port": 8080
                  },
                  "natsStreaming": {
                    "image": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats-streaming",
                      "tag": "0.23.2-3747-31"
                    }
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "tagSuffix": "-ocp"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:31:32Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Messagebroker
            name: messagebrokers.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Message Broker Operator
        displayName: Keysight WAP Message Broker Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Msg Broker
          url: https://keysight-wap-msg-broker.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-msg-broker@sha256:4701b6e456b099697bf1e25aeb4bc8e32b1892bb0956e36d84f9f80666b3a1fc
        version: 0.1.21
      entries:
      - name: keysight-wap-msg-broker.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-msg-broker
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: VictoriaMetrics
      provider-url: ""
    name: victoriametrics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: victoriametrics-operator.v0.63.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLAgent",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "persistentVolumeClaimRetentionPolicy": {
                    "whenDeleted": "Delete"
                  },
                  "remoteWrite": [
                    {
                      "maxDiskUsage": "10GB",
                      "url": "http://vlsingle-example-0.default.svc:9428/internal/insert"
                    }
                  ],
                  "remoteWriteSettings": {
                    "maxBlockSize": "30MB"
                  },
                  "replicaCount": 2,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "350Mi"
                    }
                  },
                  "storage": {
                    "volumeClaimTemplate": {
                      "spec": {
                        "resources": {
                          "requests": {
                            "storage": "50Gi"
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "victoriametrics-operator"
                  },
                  "name": "example"
                },
                "spec": {
                  "requestsLoadBalancer": {
                    "enabled": true,
                    "spec": {}
                  },
                  "vlinsert": {
                    "replicaCount": 1
                  },
                  "vlselect": {
                    "replicaCount": 1
                  },
                  "vlstorage": {
                    "replicaCount": 2,
                    "retentionPeriod": "1y",
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "50Gi"
                            }
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLSingle",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "victoriametrics-operator"
                  },
                  "name": "sample"
                },
                "spec": {
                  "retentionMaxDiskSpaceUsageBytes": "20GB",
                  "retentionPeriod": "1y",
                  "storage": {
                    "resources": {
                      "requests": {
                        "storage": "50Gi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VLogs",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "retentionPeriod": "2"
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAgent",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "extraArgs": {
                    "memory.allowedPercent": "40"
                  },
                  "remoteWrite": [
                    {
                      "maxDiskUsage": "10GB",
                      "url": "http://vmsingle-dc-1.default.svc:8429/api/v1/write"
                    },
                    {
                      "maxDiskUsage": "15GB",
                      "url": "http://vmsingle-dc-2.default.svc:8429/api/v1/write"
                    },
                    {
                      "url": "http://vminsert-main.default.svc:8480/insert/1:1/prometheus/api/v1/write"
                    }
                  ],
                  "remoteWriteSettings": {
                    "maxDiskUsagePerURL": "5GB"
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "350Mi"
                    }
                  },
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlert",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "datasource": {
                    "url": "http://vmsingle-example-pvc.default.svc:8429"
                  },
                  "evaluationInterval": "30s",
                  "notifier": {
                    "url": "http://vmalertmanager-example.default.svc:9093"
                  },
                  "replicaCount": 1,
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlertmanager",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "configSecret": "am-global-config",
                  "replicaCount": 1,
                  "selectAllByDefault": true,
                  "storage": {
                    "volumeClaimTemplate": {
                      "spec": {
                        "resources": {
                          "requests": {
                            "storage": "2Gi"
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlertmanagerConfig",
                "metadata": {
                  "name": "example-email-web"
                },
                "spec": {
                  "receivers": [
                    {
                      "email_configs": [
                        {
                          "from": "alerting@example.com",
                          "smarthost": "example.com:25",
                          "text": "ALARM",
                          "to": "some-email@example.com"
                        }
                      ],
                      "name": "email",
                      "webhook_configs": [
                        {
                          "url": "http://some-other-wh"
                        }
                      ]
                    }
                  ],
                  "route": {
                    "group_interval": "1m",
                    "receiver": "email",
                    "routes": [
                      {
                        "receiver": "email"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAuth",
                "metadata": {
                  "name": "vmauth-sample"
                },
                "spec": {
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "victoriametrics/vmauth"
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "250m",
                      "memory": "350Mi"
                    }
                  },
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMCluster",
                "metadata": {
                  "name": "example-persistent"
                },
                "spec": {
                  "replicationFactor": 2,
                  "retentionPeriod": "4",
                  "vminsert": {
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "0.5",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "vmselect": {
                    "cacheMountPath": "/select-cache",
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "0.5",
                        "memory": "500Mi"
                      }
                    },
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "2Gi"
                            }
                          }
                        }
                      }
                    }
                  },
                  "vmstorage": {
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1500Mi"
                      }
                    },
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "10Gi"
                            }
                          }
                        }
                      }
                    },
                    "storageDataPath": "/vm-data"
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMNodeScrape",
                "metadata": {
                  "name": "cadvisor-metrics"
                },
                "spec": {
                  "bearerTokenFile": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                  "interval": "10s",
                  "relabelConfigs": [
                    {
                      "action": "labelmap",
                      "regex": "__meta_kubernetes_node_label_(.+)"
                    },
                    {
                      "replacement": "kubernetes.default.svc:443",
                      "targetLabel": "__address__"
                    },
                    {
                      "regex": "(.+)",
                      "replacement": "/api/v1/nodes/$1/proxy/metrics/cadvisor",
                      "sourceLabels": [
                        "__meta_kubernetes_node_name"
                      ],
                      "targetLabel": "__metrics_path__"
                    }
                  ],
                  "scheme": "https",
                  "scrapeTimeout": "2s",
                  "selector": {},
                  "tlsConfig": {
                    "caFile": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                    "insecureSkipVerify": true
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMPodScrape",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "podMetricsEndpoints": [
                    {
                      "port": "metrics",
                      "scheme": "http"
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app.kubernetes.io/name": "example-application",
                      "monitored-by": "vm-operator"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMProbe",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "interval": "2s",
                  "jobName": "static-probe",
                  "module": "http_2xx",
                  "targets": {
                    "staticConfig": {
                      "targets": [
                        "vmagent-example.default.svc:8429/health"
                      ]
                    }
                  },
                  "vmProberSpec": {
                    "url": "prometheus-blackbox-exporter.default.svc:9115"
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMRule",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "groups": [
                    {
                      "name": "kafka",
                      "rules": [
                        {
                          "alert": "coordinator down",
                          "annotations": {
                            "description": "kafka coordinator is down",
                            "value": "{{ $value }}"
                          },
                          "expr": "ml_app_gauge{exec_context=\"consumer_group_state\"} == 0",
                          "for": "60s",
                          "labels": {
                            "job": "{{ $labels.job }}",
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMScrapeConfig",
                "metadata": {
                  "name": "mongodb"
                },
                "spec": {
                  "consulSDConfigs": [
                    {
                      "server": "https://consul-dns:8500",
                      "services": [
                        "mongodb"
                      ]
                    }
                  ],
                  "relabelConfigs": [
                    {
                      "action": "replace",
                      "sourceLabels": [
                        "__meta_consul_service"
                      ],
                      "targetLabel": "job"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMServiceScrape",
                "metadata": {
                  "name": "example-scrape"
                },
                "spec": {
                  "endpoints": [
                    {
                      "port": "http"
                    }
                  ],
                  "namespaceSelector": {},
                  "selector": {
                    "matchLabels": {
                      "app.kubernetes.io/name": "example-application",
                      "monitored-by": "vm-operator"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMSingle",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "retentionPeriod": "1"
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMSingle",
                "metadata": {
                  "name": "example-pvc"
                },
                "spec": {
                  "removePvcAfterDelete": true,
                  "retentionPeriod": "1",
                  "storage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMStaticScrape",
                "metadata": {
                  "name": "vmstaticscrape-sample"
                },
                "spec": {
                  "jobName": "static",
                  "targetEndpoints": [
                    {
                      "labels": {
                        "env": "dev",
                        "project": "operator"
                      },
                      "targets": [
                        "192.168.0.1:9100",
                        "196.168.0.50:9100"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMUser",
                "metadata": {
                  "name": "vmuser-tenant-1"
                },
                "spec": {
                  "bearerToken": "some-token",
                  "targetRefs": [
                    {
                      "crd": {
                        "kind": "VMCluster/vminsert",
                        "name": "test-persistent",
                        "namespace": "vm"
                      },
                      "target_path_suffix": "/insert/1"
                    },
                    {
                      "crd": {
                        "kind": "VMCluster/vmselect",
                        "name": "test-persistent",
                        "namespace": "vm"
                      },
                      "target_path_suffix": "/select/1"
                    },
                    {
                      "paths": [
                        "/internal/resetRollupResultCache"
                      ],
                      "static": {
                        "url": "http://vmselect-test-persistent.default.svc:8481/"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: docker.io/victoriametrics/operator:v0.63.0
          createdAt: "2025-09-12T12:58:29Z"
          description: Provides monitoring capabilities for kubernetes clusters and
            applications
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: "1.30"
          operators.openshift.io/infrastructure-features: '[fips]'
          operators.operatorframework.io.bundle.channel.default.v1: beta
          operators.operatorframework.io.bundle.channels.v1: beta
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/VictoriaMetrics/operator
          support: VictoriaMetrics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VLAgent - is a tiny but brave agent, which helps you collect
              logs from various sources and stores them in VictoriaLogs.
            displayName: VLAgent
            kind: VLAgent
            name: vlagents.operator.victoriametrics.com
            version: v1
          - description: VLCluster is fast, cost-effective and scalable logs database.
            displayName: VLCluster
            kind: VLCluster
            name: vlclusters.operator.victoriametrics.com
            version: v1
          - description: |-
              VLogs is fast, cost-effective and scalable logs database.
              VLogs is the Schema for the vlogs API
            displayName: VLogs
            kind: VLogs
            name: vlogs.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VLSingle is fast, cost-effective and scalable logs database.
              VLSingle is the Schema for the API
            displayName: VLSingle
            kind: VLSingle
            name: vlsingles.operator.victoriametrics.com
            version: v1
          - description: |-
              VMAgent - is a tiny but brave agent, which helps you collect metrics from various sources and stores them in VictoriaMetrics
              or any other Prometheus-compatible storage system that supports the remote_write protocol.
            displayName: VMAgent
            kind: VMAgent
            name: vmagents.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlertmanagerConfig is the Schema for the vmalertmanagerconfigs
              API
            displayName: VMAlertmanager Config
            kind: VMAlertmanagerConfig
            name: vmalertmanagerconfigs.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlertmanager represents Victoria-Metrics deployment for
              Alertmanager.
            displayName: VMAlertmanager
            kind: VMAlertmanager
            name: vmalertmanagers.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlert  executes a list of given alerting or recording rules
              against configured address.
            displayName: VMAlert
            kind: VMAlert
            name: vmalerts.operator.victoriametrics.com
            version: v1beta1
          - description: VMAnomaly is the Schema for the vmanomalies API.
            displayName: VMAnomaly
            kind: VMAnomaly
            name: vmanomalies.operator.victoriametrics.com
            version: v1
          - description: VMAuth is the Schema for the vmauths API
            displayName: VMAuth
            kind: VMAuth
            name: vmauths.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMCluster is fast, cost-effective and scalable time-series database.
              Cluster version with
            displayName: VMCluster
            kind: VMCluster
            name: vmclusters.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMNodeScrape defines discovery for targets placed on kubernetes nodes,
              usually its node-exporters and other host services.
              InternalIP is used as __address__ for scraping.
            displayName: VMNode Scrape
            kind: VMNodeScrape
            name: vmnodescrapes.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMPodScrape is scrape configuration for pods,
              it generates vmagent's config for scraping pod targets
              based on selectors.
            displayName: VMPod Scrape
            kind: VMPodScrape
            name: vmpodscrapes.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMProbe defines a probe for targets, that will be executed with prober,
              like blackbox exporter.
              It helps to monitor reachability of target with various checks.
            displayName: VMProbe
            kind: VMProbe
            name: vmprobes.operator.victoriametrics.com
            version: v1beta1
          - description: VMRule defines rule records for vmalert application
            displayName: VMRule
            kind: VMRule
            name: vmrules.operator.victoriametrics.com
            version: v1beta1
          - description: VMScrapeConfig specifies a set of targets and parameters
              describing how to scrape them.
            displayName: VMScrape Config
            kind: VMScrapeConfig
            name: vmscrapeconfigs.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMServiceScrape is scrape configuration for endpoints associated with
              kubernetes service,
              it generates scrape configuration for vmagent based on selectors.
              result config will scrape service endpoints
            displayName: VMService Scrape
            kind: VMServiceScrape
            name: vmservicescrapes.operator.victoriametrics.com
            version: v1beta1
          - description: VMSingle  is fast, cost-effective and scalable time-series
              database.
            displayName: VMSingle
            kind: VMSingle
            name: vmsingles.operator.victoriametrics.com
            version: v1beta1
          - description: VMStaticScrape  defines static targets configuration for
              scraping.
            displayName: VMStatic Scrape
            kind: VMStaticScrape
            name: vmstaticscrapes.operator.victoriametrics.com
            version: v1beta1
          - description: VMUser is the Schema for the vmusers API
            displayName: VMUser
            kind: VMUser
            name: vmusers.operator.victoriametrics.com
            version: v1beta1
          - description: VTCluster is fast, cost-effective and scalable traces database.
            displayName: VTCluster
            kind: VTCluster
            name: vtclusters.operator.victoriametrics.com
            version: v1
          - description: |-
              VTSingle is fast, cost-effective and scalable traces database.
              VTSingle is the Schema for the API
            displayName: VTSingle
            kind: VTSingle
            name: vtsingles.operator.victoriametrics.com
            version: v1
        description: |
          Operator manages VictoriaMetrics applications and provides monitoring features for applications running inside and outside kubernetes cluster. It has support for prometheus-operator objects and
          provides migration mechanism.

          Operator designed to use existing installation of VictoriaMetrics inside or outside kubernetes cluster or it could
          bring own resources and fully manages them.

           You can read more about operator at docs:
            - quick start [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/quick-start.md)
            - high availability [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/high-availability.md)
            - design and description of implementation [design](https://github.com/VictoriaMetrics/operator/blob/master/docs/design.md)
            - operator objects description [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/api.md)

          # VictoriaMetrics
          * VictoriaMetrics can be used as long-term storage for Prometheus or for [vmagent](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmagent/README.md).
          See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#prometheus-setup) for details.
          * Supports [Prometheus querying API](https://prometheus.io/docs/prometheus/latest/querying/api/), so it can be used as Prometheus drop-in replacement in Grafana.
          VictoriaMetrics implements [MetricsQL](https://github.com/VictoriaMetrics/VictoriaMetrics/wiki/MetricsQL) query language, which is inspired by PromQL.
          * Supports global query view. Multiple Prometheus instances may write data into VictoriaMetrics. Later this data may be used in a single query.
          * High performance and good scalability for both [inserts](https://medium.com/@valyala/high-cardinality-tsdb-benchmarks-victoriametrics-vs-timescaledb-vs-influxdb-13e6ee64dd6b)
          and [selects](https://medium.com/@valyala/when-size-matters-benchmarking-victoriametrics-vs-timescale-and-influxdb-6035811952d4).
          [Outperforms InfluxDB and TimescaleDB by up to 20x](https://medium.com/@valyala/measuring-vertical-scalability-for-time-series-databases-in-google-cloud-92550d78d8ae).
          * [Uses 10x less RAM than InfluxDB](https://medium.com/@valyala/insert-benchmarks-with-inch-influxdb-vs-victoriametrics-e31a41ae2893) when working with millions of unique time series (aka high cardinality).
          * Optimized for time series with high churn rate. Think about [prometheus-operator](https://github.com/coreos/prometheus-operator) metrics from frequent deployments in Kubernetes.
          * High data compression, so [up to 70x more data points](https://medium.com/@valyala/when-size-matters-benchmarking-victoriametrics-vs-timescale-and-influxdb-6035811952d4)
          may be crammed into limited storage comparing to TimescaleDB.
          * Optimized for storage with high-latency IO and low IOPS (HDD and network storage in AWS, Google Cloud, Microsoft Azure, etc). See [graphs from these benchmarks](https://medium.com/@valyala/high-cardinality-tsdb-benchmarks-victoriametrics-vs-timescaledb-vs-influxdb-13e6ee64dd6b).
          * A single-node VictoriaMetrics may substitute moderately sized clusters built with competing solutions such as Thanos, M3DB, Cortex, InfluxDB or TimescaleDB.
          See [vertical scalability benchmarks](https://medium.com/@valyala/measuring-vertical-scalability-for-time-series-databases-in-google-cloud-92550d78d8ae),
          [comparing Thanos to VictoriaMetrics cluster](https://medium.com/@valyala/comparing-thanos-to-victoriametrics-cluster-b193bea1683)
          and [Remote Write Storage Wars](https://promcon.io/2019-munich/talks/remote-write-storage-wars/) talk
          from [PromCon 2019](https://promcon.io/2019-munich/talks/remote-write-storage-wars/).
          * Easy operation:
              * VictoriaMetrics consists of a single [small executable](https://medium.com/@valyala/stripping-dependency-bloat-in-victoriametrics-docker-image-983fb5912b0d) without external dependencies.
              * All the configuration is done via explicit command-line flags with reasonable defaults.
              * All the data is stored in a single directory pointed by `-storageDataPath` flag.
              * Easy and fast backups from [instant snapshots](https://medium.com/@valyala/how-victoriametrics-makes-instant-snapshots-for-multi-terabyte-time-series-data-e1f3fb0e0282)
              to S3 or GCS with [vmbackup](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmbackup/README.md) / [vmrestore](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmrestore/README.md).
              See [this article](https://medium.com/@valyala/speeding-up-backups-for-big-time-series-databases-533c1a927883) for more details.
          * Storage is protected from corruption on unclean shutdown (i.e. OOM, hardware reset or `kill -9`) thanks to [the storage architecture](https://medium.com/@valyala/how-victoriametrics-makes-instant-snapshots-for-multi-terabyte-time-series-data-e1f3fb0e0282).
          * Supports metrics' scraping, ingestion and [backfilling](https://github.com/VictoriaMetrics/VictoriaMetrics#backfilling) via the following protocols:
            * [Metrics from Prometheus exporters](https://github.com/prometheus/docs/blob/master/content/docs/instrumenting/exposition_formats.md#text-based-format)
              such as [node_exporter](https://github.com/prometheus/node_exporter). See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-scrape-prometheus-exporters-such-as-node-exporter) for details.
              * [Prometheus remote write API](https://prometheus.io/docs/prometheus/latest/configuration/configuration/#remote_write)
              * [InfluxDB line protocol](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-send-data-from-influxdb-compatible-agents-such-as-telegraf) over HTTP, TCP and UDP.
              * [Graphite plaintext protocol](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-send-data-from-graphite-compatible-agents-such-as-statsd) with [tags](https://graphite.readthedocs.io/en/latest/tags.html#carbon)
              if `-graphiteListenAddr` is set.
              * [OpenTSDB put message](https://github.com/VictoriaMetrics/VictoriaMetrics#sending-data-via-telnet-put-protocol) if `-opentsdbListenAddr` is set.
              * [HTTP OpenTSDB /api/put requests](https://github.com/VictoriaMetrics/VictoriaMetrics#sending-opentsdb-data-via-http-apiput-requests) if `-opentsdbHTTPListenAddr` is set.
              * [How to import time series data](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-time-series-data).
              * [Prometheus exposition format](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-data-in-prometheus-exposition-format).
              * [Arbitrary CSV data](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-csv-data).
          * Supports metrics' relabeling. See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#relabeling) for details.
          * Ideally works with big amounts of time series data from Kubernetes, IoT sensors, connected cars, industrial telemetry, financial data and various Enterprise workloads.
        displayName: VictoriaMetrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - prometheus
        - victoriametrics
        links:
        - name: VictoriaMetrics
          url: https://victoriametrics.com
        maintainers:
        - email: nik@victoriametrics.com
          name: f41gh7
        - email: info@victoriametrics.com
          name: info
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: VictoriaMetrics
        relatedImages:
        - docker.io/victoriametrics/operator:v0.63.0
        version: 0.63.0
      entries:
      - name: victoriametrics-operator.v0.63.0
        version: 0.63.0
      - name: victoriametrics-operator.v0.62.0
        version: 0.62.0
      - name: victoriametrics-operator.v0.61.2
        version: 0.61.2
      - name: victoriametrics-operator.v0.61.1
        version: 0.61.1
      - name: victoriametrics-operator.v0.61.0
        version: 0.61.0
      - name: victoriametrics-operator.v0.60.2
        version: 0.60.2
      - name: victoriametrics-operator.v0.60.1
        version: 0.60.1
      - name: victoriametrics-operator.v0.60.0
        version: 0.60.0
      - name: victoriametrics-operator.v0.59.2
        version: 0.59.2
      - name: victoriametrics-operator.v0.59.1
        version: 0.59.1
      - name: victoriametrics-operator.v0.59.0
        version: 0.59.0
      - name: victoriametrics-operator.v0.58.1
        version: 0.58.1
      - name: victoriametrics-operator.v0.58.0
        version: 0.58.0
      - name: victoriametrics-operator.v0.57.0
        version: 0.57.0
      - name: victoriametrics-operator.v0.56.0
        version: 0.56.0
      - name: victoriametrics-operator.v0.55.0
        version: 0.55.0
      - name: victoriametrics-operator.v0.54.1
        version: 0.54.1
      - name: victoriametrics-operator.v0.53.0
        version: 0.53.0
      - name: victoriametrics-operator.v0.52.0
        version: 0.52.0
      - name: victoriametrics-operator.v0.51.1
        version: 0.51.1
      - name: victoriametrics-operator.v0.49.0
        version: 0.49.0
      - name: victoriametrics-operator.v0.48.4
        version: 0.48.4
      - name: victoriametrics-operator.v0.48.3
        version: 0.48.3
      - name: victoriametrics-operator.v0.48.1
        version: 0.48.1
      - name: victoriametrics-operator.v0.47.3
        version: 0.47.3
      - name: victoriametrics-operator.v0.47.1
        version: 0.47.1
      - name: victoriametrics-operator.v0.47.0
        version: 0.47.0
      - name: victoriametrics-operator.v0.46.4
        version: 0.46.4
      - name: victoriametrics-operator.v0.46.3
        version: 0.46.3
      - name: victoriametrics-operator.v0.39.4
        version: 0.39.4
      - name: victoriametrics-operator.v0.33.0
        version: 0.33.0
      - name: victoriametrics-operator.v0.32.1
        version: 0.32.1
      - name: victoriametrics-operator.v0.32.0
        version: 0.32.0
      - name: victoriametrics-operator.v0.29.1
        version: 0.29.1
      - name: victoriametrics-operator.v0.27.2
        version: 0.27.2
      - name: victoriametrics-operator.v0.26.2
        version: 0.26.2
      - name: victoriametrics-operator.v0.26.0
        version: 0.26.0
      - name: victoriametrics-operator.v0.25.1
        version: 0.25.1
      - name: victoriametrics-operator.v0.25.0
        version: 0.25.0
      - name: victoriametrics-operator.v0.24.0
        version: 0.24.0
      - name: victoriametrics-operator.v0.23.3
        version: 0.23.3
      - name: victoriametrics-operator.v0.22.1
        version: 0.22.1
      - name: victoriametrics-operator.v0.22.0
        version: 0.22.0
      - name: victoriametrics-operator.v0.21.0
        version: 0.21.0
      - name: victoriametrics-operator.v0.20.1
        version: 0.20.1
      - name: victoriametrics-operator.v0.19.1
        version: 0.19.1
      - name: victoriametrics-operator.v0.19.0
        version: 0.19.0
      - name: victoriametrics-operator.v0.18.2
        version: 0.18.2
      - name: victoriametrics-operator.v0.18.1
        version: 0.18.1
      - name: victoriametrics-operator.v0.18.0
        version: 0.18.0
      name: beta
    defaultChannel: beta
    packageName: victoriametrics-operator
    provider:
      name: VictoriaMetrics
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: F5 Networks Inc.
      provider-url: ""
    name: f5-bigip-ctlr-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: alpha
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: beta
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: stable
    defaultChannel: stable
    packageName: f5-bigip-ctlr-operator
    provider:
      name: F5 Networks Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dns-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dns-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSHealthCheckProbe",
                "metadata": {
                  "name": "$NAME"
                },
                "spec": {
                  "additionalHeadersRef": {
                    "name": "headers"
                  },
                  "address": "192.168.0.16",
                  "allowInsecureCertificate": true,
                  "failureThreshold": 5,
                  "hostname": "test.com",
                  "interval": "60s",
                  "path": "/healthz",
                  "port": 443,
                  "protocol": "HTTPS"
                }
              },
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSRecord",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dns-operator",
                    "app.kubernetes.io/instance": "dnsrecord-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dnsrecord",
                    "app.kubernetes.io/part-of": "dns-operator"
                  },
                  "name": "dnsrecord-sample"
                },
                "spec": {
                  "endpoints": [
                    {
                      "dnsName": "dnsrecord-simple.example.com",
                      "recordTTL": 60,
                      "recordType": "A",
                      "targets": [
                        "52.215.108.61",
                        "52.30.101.221"
                      ]
                    }
                  ],
                  "providerRef": {
                    "name": "dns-provider-creds"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/dns-rhel9-operator@sha256:8ad3876e36ba2f76355499d98a97672f54f26754fc9d219d9cb1c5b9f9874430
          createdAt: "2025-04-07T08:13:58Z"
          description: Configures how north-south traffic from outside the network
            is balanced and reaches Gateways
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kuadrant/dns-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSHealthCheckProbe is the Schema for the dnshealthcheckprobes
              API
            displayName: DNSHealthCheckProbe
            kind: DNSHealthCheckProbe
            name: dnshealthcheckprobes.kuadrant.io
            version: v1alpha1
          - description: DNSRecord is the Schema for the dnsrecords API
            displayName: DNSRecord
            kind: DNSRecord
            name: dnsrecords.kuadrant.io
            version: v1alpha1
        description: Configures how north-south traffic from outside the network is
          balanced and reaches Gateways - Part of Red Hat Connectivity Link
        displayName: Red Hat - DNS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dns
        - kuadrant
        links:
        - name: GitHub
          url: https://github.com/kuadrant/dns-operator
        - name: Documentation
          url: https://github.com/Kuadrant/dns-operator/tree/main/docs
        maintainers:
        - email: mnairn@redhat.com
          name: Michael Nairn
        - email: pbrookes@redhat.com
          name: Phil Brookes
        - email: cbrookes@redhat.com
          name: Craig Brookes
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhcl-1/dns-rhel9-operator@sha256:8ad3876e36ba2f76355499d98a97672f54f26754fc9d219d9cb1c5b9f9874430
        version: 1.1.0
      entries:
      - name: dns-operator.v1.1.0
        version: 1.1.0
      - name: dns-operator.v1.0.2
        version: 1.0.2
      - name: dns-operator.v1.0.1
        version: 1.0.1
      - name: dns-operator.v0.12.0
        version: 0.12.0
      name: stable
    defaultChannel: stable
    packageName: dns-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecr-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecr-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.services.k8s.aws/v1alpha1",
                "kind": "Repository",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecr-controller:1.3.1
          createdAt: "2025-09-20T01:21:24Z"
          description: AWS ECR controller is a service controller for managing ECR
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PullThroughCacheRule represents the state of an AWS ecr PullThroughCacheRule
              resource.
            displayName: PullThroughCacheRule
            kind: PullThroughCacheRule
            name: pullthroughcacherules.ecr.services.k8s.aws
            version: v1alpha1
          - description: Repository represents the state of an AWS ecr Repository
              resource.
            displayName: Repository
            kind: Repository
            name: repositories.ecr.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Container Registry (ECR) resources in AWS from within your Kubernetes cluster.

          **About Amazon ECR**

          Amazon Elastic Container Registry (Amazon ECR) is an AWS managed container image registry service that is secure, scalable, and reliable. Amazon ECR supports private repositories with resource-based permissions using AWS IAM. This is so that specified users or Amazon EC2 instances can access your container repositories and images. You can use your preferred CLI to push, pull, and manage Docker images, Open Container Initiative (OCI) images, and OCI compatible artifacts.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECR
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECR Developer Resources
          url: https://aws.amazon.com/ecr/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecr maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecr-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-ecr-controller.v1.3.1
        version: 1.3.1
      - name: ack-ecr-controller.v1.3.0
        version: 1.3.0
      - name: ack-ecr-controller.v1.2.0
        version: 1.2.0
      - name: ack-ecr-controller.v1.1.0
        version: 1.1.0
      - name: ack-ecr-controller.v1.0.33
        version: 1.0.33
      - name: ack-ecr-controller.v1.0.32
        version: 1.0.32
      - name: ack-ecr-controller.v1.0.31
        version: 1.0.31
      - name: ack-ecr-controller.v1.0.30
        version: 1.0.30
      - name: ack-ecr-controller.v1.0.29
        version: 1.0.29
      - name: ack-ecr-controller.v1.0.28
        version: 1.0.28
      - name: ack-ecr-controller.v1.0.27
        version: 1.0.27
      - name: ack-ecr-controller.v1.0.25
        version: 1.0.25
      - name: ack-ecr-controller.v1.0.24
        version: 1.0.24
      - name: ack-ecr-controller.v1.0.23
        version: 1.0.23
      - name: ack-ecr-controller.v1.0.22
        version: 1.0.22
      - name: ack-ecr-controller.v1.0.21
        version: 1.0.21
      - name: ack-ecr-controller.v1.0.20
        version: 1.0.20
      - name: ack-ecr-controller.v1.0.19
        version: 1.0.19
      - name: ack-ecr-controller.v1.0.17
        version: 1.0.17
      - name: ack-ecr-controller.v1.0.16
        version: 1.0.16
      - name: ack-ecr-controller.v1.0.15
        version: 1.0.15
      - name: ack-ecr-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecr-controller.v1.0.12
        version: 1.0.12
      - name: ack-ecr-controller.v1.0.10
        version: 1.0.10
      - name: ack-ecr-controller.v1.0.9
        version: 1.0.9
      - name: ack-ecr-controller.v1.0.8
        version: 1.0.8
      - name: ack-ecr-controller.v1.0.7
        version: 1.0.7
      - name: ack-ecr-controller.v1.0.6
        version: 1.0.6
      - name: ack-ecr-controller.v1.0.4
        version: 1.0.4
      - name: ack-ecr-controller.v1.0.3
        version: 1.0.3
      - name: ack-ecr-controller.v1.0.2
        version: 1.0.2
      - name: ack-ecr-controller.v1.0.0
        version: 1.0.0
      - name: ack-ecr-controller.v0.1.9
        version: 0.1.9
      - name: ack-ecr-controller.v0.1.7
        version: 0.1.7
      - name: ack-ecr-controller.v0.1.6
        version: 0.1.6
      - name: ack-ecr-controller.v0.1.5
        version: 0.1.5
      - name: ack-ecr-controller.v0.1.4
        version: 0.1.4
      - name: ack-ecr-controller.v0.1.3
        version: 0.1.3
      - name: ack-ecr-controller.v0.1.2
        version: 0.1.2
      - name: ack-ecr-controller.v0.1.1
        version: 0.1.1
      - name: ack-ecr-controller.v0.1.0
        version: 0.1.0
      - name: ack-ecr-controller.v0.0.27
        version: 0.0.27
      - name: ack-ecr-controller.v0.0.26
        version: 0.0.26
      - name: ack-ecr-controller.v0.0.25
        version: 0.0.25
      - name: ack-ecr-controller.v0.0.24
        version: 0.0.24
      - name: ack-ecr-controller.v0.0.23
        version: 0.0.23
      - name: ack-ecr-controller.v0.0.22
        version: 0.0.22
      - name: ack-ecr-controller.v0.0.21
        version: 0.0.21
      - name: ack-ecr-controller.v0.0.20
        version: 0.0.20
      - name: ack-ecr-controller.v0.0.19
        version: 0.0.19
      - name: ack-ecr-controller.v0.0.18
        version: 0.0.18
      - name: ack-ecr-controller.v0.0.17
        version: 0.0.17
      - name: ack-ecr-controller.v0.0.16
        version: 0.0.16
      - name: ack-ecr-controller.v0.0.15
        version: 0.0.15
      - name: ack-ecr-controller.v0.0.14
        version: 0.0.14
      - name: ack-ecr-controller.v0.0.13
        version: 0.0.13
      - name: ack-ecr-controller.v0.0.12
        version: 0.0.12
      - name: ack-ecr-controller.v0.0.11
        version: 0.0.11
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecr-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-elbv2-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-elbv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "LoadBalancer",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "TargetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/elbv2-controller:1.1.1
          createdAt: "2025-09-19T22:16:38Z"
          description: AWS ELB controller is a service controller for managing ELB
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Listener represents the state of an AWS elbv2 Listener resource.
            displayName: Listener
            kind: Listener
            name: listeners.elbv2.services.k8s.aws
            version: v1alpha1
          - description: LoadBalancer represents the state of an AWS elbv2 LoadBalancer
              resource.
            displayName: LoadBalancer
            kind: LoadBalancer
            name: loadbalancers.elbv2.services.k8s.aws
            version: v1alpha1
          - description: Rule represents the state of an AWS elbv2 Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.elbv2.services.k8s.aws
            version: v1alpha1
          - description: TargetGroup represents the state of an AWS elbv2 TargetGroup
              resource.
            displayName: TargetGroup
            kind: TargetGroup
            name: targetgroups.elbv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Load Balancing (ELB) resources in AWS from within your Kubernetes cluster.

          **About Amazon ELB**

          Elastic Load Balancing automatically distributes your incoming traffic across multiple targets, such as EC2 instances, containers, and IP addresses, in one or more Availability Zones. It monitors the health of its registered targets, and routes traffic only to the healthy targets. Elastic Load Balancing scales your load balancer capacity automatically in response to changes in incoming traffic.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ELB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elbv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ELB Developer Resources
          url: https://aws.amazon.com/elasticloadbalancing
        maintainers:
        - email: ack-maintainers@amazon.com
          name: elb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/elbv2-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-elbv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-elbv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-elbv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-elbv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-elbv2-controller.v1.0.10
        version: 1.0.10
      - name: ack-elbv2-controller.v1.0.9
        version: 1.0.9
      - name: ack-elbv2-controller.v1.0.8
        version: 1.0.8
      - name: ack-elbv2-controller.v1.0.7
        version: 1.0.7
      - name: ack-elbv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-elbv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-elbv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-elbv2-controller.v1.0.3
        version: 1.0.3
      - name: ack-elbv2-controller.v1.0.2
        version: 1.0.2
      - name: ack-elbv2-controller.v1.0.1
        version: 1.0.1
      - name: ack-elbv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-elbv2-controller.v0.0.11
        version: 0.0.11
      - name: ack-elbv2-controller.v0.0.10
        version: 0.0.10
      - name: ack-elbv2-controller.v0.0.9
        version: 0.0.9
      - name: ack-elbv2-controller.v0.0.8
        version: 0.0.8
      - name: ack-elbv2-controller.v0.0.6
        version: 0.0.6
      - name: ack-elbv2-controller.v0.0.5
        version: 0.0.5
      - name: ack-elbv2-controller.v0.0.4
        version: 0.0.4
      - name: ack-elbv2-controller.v0.0.3
        version: 0.0.3
      - name: ack-elbv2-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-elbv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hazelcast, Inc
      provider-url: ""
    name: hazelcast-platform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hazelcast-platform-operator.v5.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "cache-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "CronHotBackup",
                "metadata": {
                  "name": "cronhotbackup-sample"
                },
                "spec": {
                  "hotBackupTemplate": {
                    "metadata": {
                      "labels": {
                        "cron-hotbackup": "true"
                      }
                    },
                    "spec": {
                      "hazelcastResourceName": "hazelcast"
                    }
                  },
                  "schedule": "*/30 * * * *"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast"
                },
                "spec": {
                  "clusterSize": 3,
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast-sample"
                },
                "spec": {
                  "clusterSize": 3,
                  "jet": {
                    "bucketConfig": {
                      "bucketURI": "gs://operator-user-code/jetJobs",
                      "secretName": "br-secret-gcp"
                    },
                    "enabled": true,
                    "resourceUploadEnabled": true
                  },
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HazelcastEndpoint",
                "metadata": {
                  "name": "hazelcastendpoint-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "port": 5701,
                  "type": "Discovery"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HotBackup",
                "metadata": {
                  "name": "hot-backup"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-longrun"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-longrun-2.0.0.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-1.0.2.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJobSnapshot",
                "metadata": {
                  "name": "jetjobsnapshot-sample"
                },
                "spec": {
                  "cancelJob": false,
                  "jetJobResourceName": "jet-job"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ManagementCenter",
                "metadata": {
                  "name": "managementcenter"
                },
                "spec": {
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "hazelcastClusters": [
                    {
                      "address": "hazelcast",
                      "name": "dev"
                    }
                  ],
                  "licenseKeySecretName": "hazelcast-license-key",
                  "persistence": {
                    "enabled": true,
                    "size": "10Gi"
                  },
                  "repository": "hazelcast/management-center"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Map",
                "metadata": {
                  "name": "map"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "MultiMap",
                "metadata": {
                  "name": "multimap-sample"
                },
                "spec": {
                  "backupCount": 3,
                  "binary": true,
                  "collectionType": "LIST",
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "queue-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ReplicatedMap",
                "metadata": {
                  "name": "replicatedmap-sample"
                },
                "spec": {
                  "asyncFillup": true,
                  "hazelcastResourceName": "hazelcast",
                  "inMemoryFormat": "OBJECT"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "topic-sample"
                },
                "spec": {
                  "globalOrderingEnabled": true,
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "UserCodeNamespace",
                "metadata": {
                  "name": "ucn-sample"
                },
                "spec": {
                  "bucketConfig": {
                    "bucketURI": "gs://operator-user-code/userCodeNamespace",
                    "secretName": "br-secret-gcp"
                  },
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanReplication",
                "metadata": {
                  "name": "wanreplication-sample"
                },
                "spec": {
                  "endpoints": "35.192.33.252",
                  "resources": [
                    {
                      "kind": "Map",
                      "name": "map-sample"
                    },
                    {
                      "kind": "Hazelcast",
                      "name": "hz-sample"
                    }
                  ],
                  "targetClusterName": "dev"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanSync",
                "metadata": {
                  "name": "wansync-sample"
                },
                "spec": {
                  "wanReplicationResourceName": "wan-replication-sample"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: docker.io/hazelcast/hazelcast-platform-operator@sha256:99005afda5c71ceeac200928dfed4ea35480f871053101e3aeaa44005e3fda14
          createdAt: "2024-07-29T15:35:56Z"
          description: '[DEPRECATED] Use the certified Hazelcast Operator instead'
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hazelcast/hazelcast-platform-operator/
          support: Hazelcast, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.hazelcast.com
            version: v1alpha1
          - description: CronHotBackup is the Schema for the cronhotbackups API
            displayName: Cron Hot Backup
            kind: CronHotBackup
            name: cronhotbackups.hazelcast.com
            version: v1alpha1
          - description: HazelcastEndpoint is the Schema for the hazelcastendpoints
              API
            displayName: Hazelcast Endpoint
            kind: HazelcastEndpoint
            name: hazelcastendpoints.hazelcast.com
            version: v1alpha1
          - description: Hazelcast is the Schema for the hazelcasts API
            displayName: Hazelcast
            kind: Hazelcast
            name: hazelcasts.hazelcast.com
            version: v1alpha1
          - description: HotBackup is the Schema for the hot backup API
            displayName: Hot Backup
            kind: HotBackup
            name: hotbackups.hazelcast.com
            version: v1alpha1
          - description: JetJob is the Schema for the jetjobs API
            displayName: Jet Job
            kind: JetJob
            name: jetjobs.hazelcast.com
            version: v1alpha1
          - description: JetJobSnapshot is the Schema for the jetjobsnapshots API
            displayName: Jet Job Snapshot
            kind: JetJobSnapshot
            name: jetjobsnapshots.hazelcast.com
            version: v1alpha1
          - description: ManagementCenter is the Schema for the managementcenters
              API
            displayName: Management Center
            kind: ManagementCenter
            name: managementcenters.hazelcast.com
            version: v1alpha1
          - description: Map is the Schema for the maps API
            displayName: Map
            kind: Map
            name: maps.hazelcast.com
            version: v1alpha1
          - description: MultiMap is the Schema for the multimaps API
            displayName: Multi Map
            kind: MultiMap
            name: multimaps.hazelcast.com
            version: v1alpha1
          - description: Queue is the Schema for the queues API
            displayName: Queue
            kind: Queue
            name: queues.hazelcast.com
            version: v1alpha1
          - description: ReplicatedMap is the Schema for the replicatedmaps API
            displayName: Replicated Map
            kind: ReplicatedMap
            name: replicatedmaps.hazelcast.com
            version: v1alpha1
          - description: Topic is the Schema for the topics API
            displayName: Topic
            kind: Topic
            name: topics.hazelcast.com
            version: v1alpha1
          - description: UserCodeNamespace is the Schema for the usercodenamespaces
              API
            displayName: User Code Namespace
            kind: UserCodeNamespace
            name: usercodenamespaces.hazelcast.com
            version: v1alpha1
          - description: WanReplication is the Schema for the wanreplications API
            displayName: Wan Replication
            kind: WanReplication
            name: wanreplications.hazelcast.com
            version: v1alpha1
          - description: WanSync is the Schema for the wansyncs API
            displayName: Wan Sync
            kind: WanSync
            name: wansyncs.hazelcast.com
            version: v1alpha1
        description: |
          ## Deprecated
          Use the certified Hazelcast Operator instead.
          # Hazelcast Platform Operator #

          Easily deploy Hazelcast clusters and Management Center into Kubernetes environments and manage their lifecycles.

          ## Before You Start

          If you are planning to create Hazelcast Platform Enterprise clusters, you need to [create a secret](https://docs.hazelcast.com/operator/latest/get-started#step-2-start-the-hazelcast-cluster) for the license. You can request a trial license key from [here](https://trialrequest.hazelcast.com).

          For Hazelcast Platform clusters, you can simply continue.

          ## Documentation

          1. [Get started](https://docs.hazelcast.com/operator/latest/get-started) with the Operator
          2. [Connect to the cluster from outside Kubernetes](https://guides.hazelcast.org/hazelcast-platform-operator-expose-externally)
          3. [Restore a Cluster from Cloud Storage with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-external-backup-restore)
          4. [Replicate Data between Two Hazelcast Clusters with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-wan-replication)

          ## Features

          Hazelcast Platform Operator supports the features below:

          * Custom resource for Hazelcast Platform (Enterprise) and Management Center
          * Observe status of Hazelcast clusters and Management Center instances
          * High Availability Mode configuration to create clusters that are resilient to node and zone failures
          * Support for TLS connections between members using self-signed certificates
          * Scale up and down Hazelcast clusters
          * Expose Hazelcast cluster to external clients ([Smart & Unisocket](https://docs.hazelcast.com/hazelcast/latest/clients/java#java-client-operation-modes))
          * Backup Hazelcast persistence data to cloud storage with the possibility of scheduling it and restoring the data accordingly
          * WAN Replication feature when you need to synchronize multiple Hazelcast clusters, which are connected by WANs
          * Full and Delta WAN Sync support
          * Tiered Storage support for Hazelcast and Map CRs
          * CP Subsystem configuration support for Hazelcast CR
          * User Code Deployment feature, which allows you to deploy custom and domain classes from cloud storages and URLs to Hazelcast members
          * The User Code Namespaces feature allows you to deploy custom and domain classes at runtime without a members restart
          * Support the configuration of advanced networking options
          * Support Multi-namespace configuration
          * ExecutorService and EntryProcessor support
          * Support several data structures like Map, Topic, MultiMap, ReplicatedMap, Queue and Cache which can be created dynamically via specific Custom Resources
          * MapStore, Near Cache and off-heap memory (HD memory and native memory) support for the Map CR
          * Native Memory support for the Cache CR
          * Support Jet configuration and Jet Job submission using the JetJob CR
          * Support for exporting the snapshots of JetJob CRs using JetJobSnapshot CR
          * Support for custom configurations using ConfigMap
        displayName: '[DEPRECATED] Hazelcast Platform Operator'
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hazelcast
        - keyvalue
        - in-memory
        - database
        - caching
        links:
        - name: Documentation
          url: https://docs.hazelcast.com/operator/latest/
        - name: Source Code
          url: https://github.com/hazelcast/hazelcast-platform-operator
        maintainers:
        - email: cloudnative@hazelcast.com
          name: Hazelcast Cloud Native Team
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Hazelcast, Inc
        relatedImages:
        - docker.io/hazelcast/hazelcast-platform-operator@sha256:99005afda5c71ceeac200928dfed4ea35480f871053101e3aeaa44005e3fda14
        version: 5.13.1
      entries:
      - name: hazelcast-platform-operator.v5.13.1
        version: 5.13.1
      - name: hazelcast-platform-operator.v5.13.0
        version: 5.13.0
      - name: hazelcast-platform-operator.v5.12.0
        version: 5.12.0
      - name: hazelcast-platform-operator.v5.11.0
        version: 5.11.0
      - name: hazelcast-platform-operator.v5.10.0
        version: 5.10.0
      - name: hazelcast-platform-operator.v5.9.0
        version: 5.9.0
      - name: hazelcast-platform-operator.v5.8.0
        version: 5.8.0
      - name: hazelcast-platform-operator.v5.7.0
        version: 5.7.0
      - name: hazelcast-platform-operator.v5.6.0
        version: 5.6.0
      - name: hazelcast-platform-operator.v5.5.0
        version: 5.5.0
      - name: hazelcast-platform-operator.v5.4.0
        version: 5.4.0
      - name: hazelcast-platform-operator.v5.3.0
        version: 5.3.0
      - name: hazelcast-platform-operator.v5.2.0
        version: 5.2.0
      - name: hazelcast-platform-operator.v5.1.0
        version: 5.1.0
      - name: hazelcast-platform-operator.v5.0.0
        version: 5.0.0
      name: alpha
    defaultChannel: alpha
    packageName: hazelcast-platform-operator
    provider:
      name: Hazelcast, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-storage
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-storage.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Storage",
                "metadata": {
                  "name": "storage-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore"
                  },
                  "minio": {
                    "clusterDomain": "cluster.local",
                    "helmKubectlJqImage": {
                      "repository": "docker.io/bskim45/helm-kubectl-jq"
                    },
                    "image": {
                      "repository": "prod.harbor.keysight.digital/loadcore/wap-minio",
                      "tag": "RELEASE.2021-12-29T06-49-06Z-3747-37-ocp"
                    },
                    "mcImage": {
                      "repository": "prod.harbor.keysight.digital/loadcore/wap-minio-mc",
                      "tag": "RELEASE.2021-12-29T06-52-55Z-3747-37-ocp"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "securityContext": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:32:20Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Storage
            name: storages.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Storage Operator
        displayName: Keysight WAP Storage Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Storage
          url: https://keysight-wap-storage.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-storage@sha256:9f7710cfe72658cd5456fd54f86ca6be5c9a4de2a01114c0f7531b70712bd981
        version: 0.1.21
      entries:
      - name: keysight-wap-storage.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-storage
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: patch-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: patch-operator.v0.1.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Patch",
                "metadata": {
                  "name": "test-complex-patch"
                },
                "spec": {
                  "patches": {
                    "test-complex-patch": {
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 1).metadata.name }}: {{ (index . 2).metadata.name }}    \n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "kind": "Namespace",
                          "name": "test-patch-operator"
                        },
                        {
                          "apiVersion": "v1",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "test-patch-operator"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "test-patch-operator"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/patch-operator@sha256:030ade9b9428deca3ac6390618dc4d6e6d728577d1467a22dc4d4b68c9848886
          createdAt: "2024-05-10T14:07:19Z"
          description: An operator to apply patches to Kubernetes objects in a declarative
            way.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: patch-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/patch-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Patch
            name: patches.redhatcop.redhat.io
            version: v1alpha1
        description: "The patch operator helps with defining patches in a declarative
          way. This operator has two main features:\n\n1. [ability to patch an object
          at creation time via a mutating webhook](#creation-time-patch-injection)\n2.
          [ability to enforce patches on one or more objects via a controller](#runtime-patch-enforcement)\n
          \n## Creation-time patch injection\n\nWhy apply a patch at creation time
          when you could directly create the correct object? The reason is that sometime
          the correct value depends on configuration set on the specific cluster in
          which the object is being deployed. For example, an ingress/route hostname
          might depend on the specific cluster. Consider the following example based
          on cert-manager:\n \n```yaml\napiVersion: cert-manager.io/v1\nkind: ClusterIssuer\nmetadata:\n
          \ name: letsencrypt-issuer\nspec:\n  acme:\n    server: 'https://acme-v02.api.letsencrypt.org/directory'\n
          \   email: {{ .Values.letsencrypt.email }}\n    privateKeySecretRef:\n      name:
          letsencrypt-staging\n    solvers:  \n    - dns01:\n        route53:\n          accessKeyID:
          << access_key >>\n          secretAccessKeySecretRef:\n            name:
          cert-manager-dns-credentials\n            key: aws_secret_access_key\n          region:
          << region >>\n          hostedZoneID: << hosted_zone_id >>\n ```\n\nIn this
          example the fields: `<< access_key >>`, `<< region >>` and `<< hosted_zone_id
          >>` are dependent on the specific region in which the cluster is being deployed
          and in many cases they are discoverable from other configurations already
          present in the cluster. If you want to deploy the above Cluster Issuer object
          with a gitops approach, then there is no easy way to discover those values.
          The solution so far is to manually discover those values and create a different
          gitops configuration for each cluster. But consider if you could look up
          values at deploy time based on the cluster you are deploying to. Here is
          how this object might look:\n\n```yaml\napiVersion: cert-manager.io/v1\nkind:
          ClusterIssuer\nmetadata:\n  name: letsencrypt-issuer\n  namespace: {{ .Release.Namespace
          }}\n  annotations:\n    \"redhat-cop.redhat.io/patch\": |\n      spec:\n
          \       acme:\n        - dns01:\n            route53:\n              accessKeyID:
          {{ (lookup \"v1\" \"Secret\" .metadata.namespace \"cert-manager-dns-credentials\").data.aws_access_key_id
          | b64dec }}\n              secretAccessKeySecretRef:\n                name:
          cert-manager-dns-credentials\n                key: aws_secret_access_key\n
          \             region: {{ (lookup \"config.openshift.io/v1\" \"Infrastructure\"
          \"\" \"cluster\").status.platformStatus.aws.region }}\n              hostedZoneID:
          {{ (lookup \"config.openshift.io/v1\" \"DNS\" \"\" \"cluster\").spec.publicZone.id
          }} \nspec:\n  acme:\n    server: 'https://acme-v02.api.letsencrypt.org/directory'\n
          \   email: {{ .Values.letsencrypt.email }}\n    privateKeySecretRef:\n      name:
          letsencrypt-staging\n    solvers:  \n    - dns01:\n        route53:\n          accessKeyID:
          << access_key >>\n          secretAccessKeySecretRef:\n            name:
          cert-manager-dns-credentials\n            key: aws_secret_access_key\n          region:
          << region >>\n          hostedZoneID: << hosted_zone_id >>\n ```\n\nThe
          annotation specifies a patch that will be applied by a [MutatingWebhook](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/),
          as you can see the three values are now looked up from different configurations
          in the cluster.\n\nTwo annotations influences the behavior of this MutatingWebhook\n
          \n1. \"redhat-cop.redhat.io/patch\" : this is the patch itself. The patch
          is evaluated as a template with the object itself as it's only parameter.
          The template is expressed in golang template notation and supports the same
          functions as helm template including the [lookup](https://helm.sh/docs/chart_template_guide/functions_and_pipelines/#using-the-lookup-function)
          function which plays a major role here. The patch must be expressed in yaml
          for readability. It will be converted to json by the webhook logic.\n2.
          \"redhat-cop.redhat.io/patch-type\" : this is the type of json patch. The
          possible values are: `application/json-patch+json`, `application/merge-patch+json`
          and `application/strategic-merge-patch+json`. If this annotation is omitted
          it defaults to strategic merge.\n\n### Security Considerations\n \nThe lookup
          function, if used by the template, is executed with a client which impersonates
          the user issuing the object creation/update request. This should prevent
          security permission leakage.\n\n### Installing the creation time webhook\n
          \nThe creation time webhook is not installed by the operator. This is because
          there is no way to know which specific object type should be intercepted
          and intercepting all of the types would be too inefficient. It's up to the
          administrator then to install the webhook. Here is some guidance.\n\nIf
          you installed the operator via OLM, use the following webhook template:\n\n```yaml\napiVersion:
          admissionregistration.k8s.io/v1\n kind: MutatingWebhookConfiguration\nmetadata:\n
          \ name: patch-operator-inject\n  annotations:\n    service.beta.openshift.io/inject-cabundle:
          \"true\"\nwebhooks:\n - admissionReviewVersions:\n  - v1\n  clientConfig:\n
          \   service:\n      name: patch-operator-webhook-service\n      namespace:
          patch-operator\n      path: /inject\n  failurePolicy: Fail\n  name: patch-operator-inject.redhatcop.redhat.io\n
          \ rules:\n  - << add your intercepted objects here >>\n  sideEffects: None\n```\n\nIf
          you installed the operator via the Helm chart and are using cert-manager,
          use the following webhook template:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\n
          kind: MutatingWebhookConfiguration\nmetadata:\n  name: patch-operator-inject\n
          \ annotations:\n    cert-manager.io/inject-ca-from: '{{ .Release.Namespace
          }}/webhook-server-cert'\n webhooks:\n- admissionReviewVersions:\n  - v1\n
          \ clientConfig:\n    service:\n      name: patch-operator-webhook-service\n
          \     namespace: patch-operator\n      path: /inject\n  failurePolicy: Fail\n
          \ name: patch-operator-inject.redhatcop.redhat.io\n  rules:\n  - << add
          your intercepted objects here >>\n  sideEffects: None\n```  \n\nYou should
          need to enable the webhook only for `CREATE` operations. So for example
          to enable the webhook on configmaps:\n\n```yaml\n  rules:\n  - apiGroups:\n
          \   - \"\"\n    apiVersions:\n    - v1\n    operations:\n    - CREATE\n
          \   resources:\n    - configmaps\n```\n\n## Runtime patch enforcement\n\nThere
          are situations when we need to patch pre-existing objects. Again this is
          a use case that is hard to model with gitops operators which will work only
          on object that they own. Especially with sophisticated Kubernetes distributions,
          it is not uncommon that a Kubernetes instance, at installation time, is
          configured with some default settings. Changing those configurations means
          patching those objects. For example, let's take the case of OpenShift Oauth
          configuration. This object is present by default and it is expected to be
          patched with any newly enabled authentication mechanism. This is how it
          looks like after installation:\n\n```yaml\napiVersion: config.openshift.io/v1\n
          kind: OAuth\nmetadata:\n  name: cluster\n  ownerReferences:\n    - apiVersion:
          config.openshift.io/v1\n      kind: ClusterVersion\n      name: version\n
          \     uid: 9a9d450b-3076-4e30-ac05-a889d6341fc3\n  resourceVersion: '20405124'\n
          spec: []\n```\n\nIf we need to patch it we can use the patch controller
          and the `Patch` object as follows:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: Patch\nmetadata:\n  name: gitlab-ocp-oauth-provider\n  namespace:
          openshift-config\nspec:\n  serviceAccountRef:\n    name: default\n  patches:\n
          \   gitlab-ocp-oauth-provider:\n      targetObjectRef:\n        apiVersion:
          config.openshift.io/v1\n        kind: OAuth\n        name: cluster\n      patchTemplate:
          |\n        spec:\n          identityProviders:\n          - name: my-github
          \n            mappingMethod: claim \n            type: GitHub\n            github:\n
          \             clientID: \"{{ (index . 1).data.client_id | b64dec }}\" \n
          \             clientSecret: \n                name: ocp-github-app-credentials\n
          \             organizations: \n              - my-org\n              teams:
          []            \n      patchType: application/merge-patch+json\n      sourceObjectRefs:\n
          \     - apiVersion: v1\n        kind: Secret\n        name: ocp-github-app-credentials\n
          \       namespace: openshift-config\n```\n\nThis will cause the OAuth object
          to be patched and the patch to be enforced. That means that if anything
          changes on the secret that we use a parameter (which may be rotated) or
          in Oauth object itself, the patch will be reapplied. In this case we are
          adding a gitlab authentication provider.\n\n A `patch` has the following
          fields:\n\n`targetObjectRef` this refers to the object(s) receiving the
          patch. Mutliple object can be selected based on the following rules:\n \n|
          Namespaced Type | Namespace | Name | Selection type |\n| --- | --- | ---
          | --- |\n| yes | null | null | multiple selection across namespaces |\n|
          yes | null | not null | multiple selection across namespaces where the name
          corresponds to the passed name |\n| yes | not null | null | multiple selection
          within a namespace |\n| yes | not null | not nul | single selection |\n|
          no | N/A | null | multiple selection  |\n| no | N/A | not null | single
          selection |\n\nSelection can be further narrowed down by filtering by labels
          and/or annotations using the `labelSelector` and `annotationSelector` fields.
          The patch will be applied to all of the selected instances.\n\n`sourceObjectRefs`
          these are the objects that will be watched and become part of the parameters
          of the patch template. Name and Namespace of sourceRefObjects are interpreted
          as golang templates with the current target instance and the only parameter.
          This allows to select different source object for each target object.\n
          \nSo, for example, with this patch:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: Patch\nmetadata:\n  name: multiple-namespaced-targets-patch\nspec:\n
          \ serviceAccountRef:\n    name: default\n  patches:\n    multiple-namespaced-targets-patch:\n
          \     targetObjectRef:\n        apiVersion: v1\n        kind: ServiceAccount\n
          \       name: deployer\n      patchTemplate: |\n        metadata:\n          annotations:\n
          \           {{ (index . 1).metadata.uid }}: {{ (index . 1) }}\n      patchType:
          application/strategic-merge-patch+json\n      sourceObjectRefs:\n      -
          apiVersion: v1\n        kind: ServiceAccount\n        name: default\n        namespace:
          \"{{ .metadata.namespace }}\"\n        fieldPath: $.metadata.uid\n```\n\nThe
          `deployer` service accounts from all namespaces are selected as target of
          this patch, each patch template will receive a different parameter and that
          is the `default` service account of the same namespace as the namespace
          of the `deployer` service account being processed.\n\n`sourceObjectRefs`
          also have the `fieldPath` field which can contain a jsonpath expression.
          If a value is passed the jsonpath expression will be calculate for the current
          source object and the result will be passed as parameter of the template.\n\n`patchTemplate`
          This is the the template that will be evaluated. The result must be a valid
          patch compatible with the requested type and expressed in yaml for readability.
          The parameters passed to the template are the target object and then the
          all of the source object. So if you want to refer to the target object in
          the template you can use this expression `(index . 0)`. Higher indexes refer
          to the sourceObjectRef array. The template is expressed in golang template
          notation and supports the same functions as helm template.\n\n`patchType`
          is the type of the json patch. The possible values are: `application/json-patch+json`,
          `application/merge-patch+json` and `application/strategic-merge-patch+json`.
          If this annotation is omitted it defaults to strategic merge.\n\n### Patch
          Controller Security Considerations\n\nThe patch enforcement enacted by the
          patch controller is executed with a client which uses the service account
          referenced by the `serviceAccountRef` field. So before a patch object can
          actually work an administrator must have granted the needed permissions
          to a service account in the same namespace. The `serviceAccountRef` will
          default to the `default` service account if not specified.\n\n### Patch
          Controller Performance Considerations\n\nThe patch controller will create
          a controller-manager and per `Patch` object and a reconciler for each of
          the `PatchSpec` defined in the array on patches in the `Patch` object.\nThese
          reconcilers share the same cached client. In order to be able to watch changes
          on target and source objects of a `PatchSpec`, all of the target and source
          object type instances will be cached by the client. This is a normal behavior
          of a controller-manager client, but it implies that if you create patches
          on object types that have many instances in etcd (Secrets, ServiceAccounts,
          Namespaces for example), the patch operator instance will require a significant
          amount of memory. A way to contain this issue is to try to aggregate together
          `PatchSpec` that deal with the same object types. This will cause those
          object type instances to cached only once.  \n"
        displayName: Patch Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - patch
        - enforcing
        - declarative
        links:
        - name: Patch Operator
          url: https://github.com/redhat-cop/patch-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/patch-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/patch-operator@sha256:030ade9b9428deca3ac6390618dc4d6e6d728577d1467a22dc4d4b68c9848886
        version: 0.1.11
      entries:
      - name: patch-operator.v0.1.11
        version: 0.1.11
      - name: patch-operator.v0.1.10
        version: 0.1.10
      - name: patch-operator.v0.1.9
        version: 0.1.9
      - name: patch-operator.v0.1.8
        version: 0.1.8
      - name: patch-operator.v0.1.7
        version: 0.1.7
      - name: patch-operator.v0.1.6
        version: 0.1.6
      - name: patch-operator.v0.1.5
        version: 0.1.5
      - name: patch-operator.v0.1.4
        version: 0.1.4
      - name: patch-operator.v0.1.3
        version: 0.1.3
      - name: patch-operator.v0.1.2
        version: 0.1.2
      - name: patch-operator.v0.1.1
        version: 0.1.1
      - name: patch-operator.v0.1.0
        version: 0.1.0
      - name: patch-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: patch-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-bedrockagent-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-bedrockagent-controller.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bedrockagent.services.k8s.aws/v1alpha1",
                "kind": "Agent",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/bedrockagent-controller:0.1.1
          createdAt: "2025-09-20T01:18:56Z"
          description: AWS Bedrock Agent controller is a service controller for managing
            Bedrock Agent resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Agent represents the state of an AWS bedrockagent Agent resource.
            displayName: Agent
            kind: Agent
            name: agents.bedrockagent.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Bedrock Agent resources in AWS from within your Kubernetes cluster.

          **About Amazon Bedrock Agent**

          Amazon Bedrock is a fully managed service that makes high-performing foundation models (FMs) from leading AI companies and Amazon available for your use through a unified API. You can choose from a wide range of foundation models to find the model that is best suited for your use case. Amazon Bedrock also offers a broad set of capabilities to build generative AI applications with security, privacy, and responsible AI. Using Amazon Bedrock, you can easily experiment with and evaluate top foundation models for your use cases, privately customize them with your data using techniques such as fine-tuning and Retrieval Augmented Generation (RAG), and build agents that execute tasks using your enterprise systems and data sources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Bedrock Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bedrockagent
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Bedrock Agent Developer Resources
          url: https://aws.amazon.com/bedrock/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: bedrock-agent maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/bedrockagent-controller:0.1.1
        version: 0.1.1
      entries:
      - name: ack-bedrockagent-controller.v0.1.1
        version: 0.1.1
      - name: ack-bedrockagent-controller.v0.1.0
        version: 0.1.0
      - name: ack-bedrockagent-controller.v0.0.6
        version: 0.0.6
      - name: ack-bedrockagent-controller.v0.0.5
        version: 0.0.5
      - name: ack-bedrockagent-controller.v0.0.4
        version: 0.0.4
      - name: ack-bedrockagent-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-bedrockagent-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: devopstales
      provider-url: devopstales.github.io
    name: community-trivy-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: community-trivy-operator.v2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "trivy-operator.devopstales.io/v1",
                "kind": "NamespaceScanner",
                "metadata": {
                  "name": "trivy-operator-main-config"
                },
                "spec": {
                  "crontab": "*/5 * * * *",
                  "namespace_selector": "trivy-scan",
                  "clusterWide": "False",
                  "policyreport": "False",
                  "registry": [
                    {
                      "name": "docker.io",
                      "user": "",
                      "password": ""
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: devopstales/trivy-operator:2.4.1
          createdAt: "2022-07-11T15:00:00Z"
          description: Trivy Operator for scheduled imagescans and an Admission Control.
          repository: https://github.com/devopstales/trivy-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Confifuration for trivy-operator.
            displayName: NamespaceScanner
            kind: NamespaceScanner
            name: namespace-scanners.trivy-operator.devopstales.io
            version: v1
          - description: |
              VulnerabilityReport summarizes vulnerabilities in application dependencies and operating system packages
              built into container images.
            displayName: VulnerabilityReport
            kind: VulnerabilityReport
            name: vulnerabilityreports.trivy-operator.devopstales.io
            version: v1
        description: |
          Trivy Operator for scheduled imagescans and an Admission Control.

          ### Add Github token to extend your Github API Rate Limit
          The Operator use Github to download Trivy database. If your Github API Rate Limit Exceeded add your Github token to the operator deployment as GITHUB_TOKEN variable.

          ```yaml
          ...
          env:
          - name: GITHUB_TOKEN
            value: "githubToken"
          ```

          ### Scheduled Image scans
          Default every 5 minutes execute a scan script. It will get image list from all namespaces with the label `trivy-scan=true`, and then scan this images with trivy, finally we will get metrics on `http://[pod-ip]:9115/metrics`

          You can chane the schedule of the image scanning by setting the `crontab` variable in the namespace-scanner object.


          ### Cache Scan results
          From the 2.4 version you can use an external redis to store scanning results and speed up the scanning process.

          ```bash
          ...
          env:
          - name: REDIS_ENABLED
            value: "True"
          - name: REDIS_BACKEND
            value: redis://redis-cache:6379
          ```

          ### Persist Trivy database cache
          To Persist Trivy cache you need to create a pvc and mount to the Deployment under `/home/trivy-operator/trivy-cache`

          ```yaml
          ...
            volumeMounts:
            - name: cache
              mountPath: "/home/trivy-operator/trivy-cache"
          volumes:
          - name: cache
            persistentVolumeClaim:
              claimName: trivy-cache
          ```

          ### Trivy Image Validator
          The admission controller function can be configured as a ValidatingWebhook in a k8s cluster. Kubernetes will send requests to the admission server when a Pod creation is initiated. The admission controller checks the image using trivy if it is in a namespace with the lable `trivy-operator-validation=true`.

          To enable the admission controller you need to set the `clusterWide` variable to `true` in the namespace-scanner object.

          Enable the scheduled scan and the admission controller for a specific namespace:
          ```bash
          kubectl label namespaces test-apps trivy-scan=true
          # or
          kubectl label namespaces test-apps trivy-operator-validation=true
          ```

          You can define policy to the Admission Controller, by adding annotation to the pod trough the deployment:

          ```bash
          spec:
            ...
            template:
              metadata:
                annotations:
                  trivy.security.devopstales.io/medium: "5"
                  trivy.security.devopstales.io/low: "10"
                  trivy.security.devopstales.io/critical: "2"
          ...
          ```

          ### Monitorng
          Trivy operator has a prometheus endpoint wher yo can get the resoults of the namespace scans and the adission controller.

          ```bash
          curl -s http://10.43.179.39:9115/metrics | grep so_vulnerabilities

          # HELP so_vulnerabilities Container vulnerabilities
          # TYPE so_vulnerabilities gauge
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="UNKNOWN"} 0
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="LOW"} 23
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="MEDIUM"} 93
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="HIGH"} 76
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="CRITICAL"} 25
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="UNKNOWN"} 0
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="LOW"} 23
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="MEDIUM"} 88
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="HIGH"} 60
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="CRITICAL"} 8
          ```

          ```bash
          curl -s http://10.43.179.39:9115/metrics | grep ac_vulnerabilities

          # HELP ac_vulnerabilities Admission Controller vulnerabilities
          # TYPE ac_vulnerabilities gauge
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="UNKNOWN"} 0.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="LOW"} 83.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="MEDIUM"} 6.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="HIGH"} 6.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="CRITICAL"} 4.0
          ```

          ### Policy Report reation
          The PolicyReport object is a protopype object probosed by the Kubernetes policy work group. The Policy Report Custom Resource Definition (CRD) can be used as a common way to provide policy results to Kubernetes cluster administrators and users, using native tools.

          Add the PolicyReport CRDs to your cluster (v1alpha2):
          ```bash
          kubectl create -f https://github.com/kubernetes-sigs/wg-policy-prototypes/raw/master/policy-report/crd/v1alpha2/wgpolicyk8s.io_policyreports.yaml
          ```

          You can enable the Policy Report creatin by setting the `policyreport` to `True` in the NamespaceScanner object.

          This objects can be visualized by the Policy Reporter UI. The Policy Reporter UI is a monitoring and Observability Tool for the PolicyReport CRD with an optional UI. It is created by Kyverno. The main goal was a tool to visualize the resoluts of the Kyverno policies, but because it uses the PolicyReports CRD it can visualize the resoults of the trivy-operator scans.
        displayName: Community Trivy Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trivy
        - security
        links:
        - name: Documentation
          url: https://devopstales.github.io/trivy-operator
        - name: Blog
          url: https://devopstales.github.io
        maintainers:
        - email: devopstales@protonmail.com
          name: devopstales
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: devopstales
          url: devopstales.github.io
        relatedImages:
        - devopstales/trivy-operator:2.4.1
        version: 2.4.0
      entries:
      - name: community-trivy-operator.v2.4.0
        version: 2.4.0
      - name: community-trivy-operator.v2.3.0
        version: 2.3.0
      - name: community-trivy-operator.v2.2.0
        version: 2.2.0
      - name: community-trivy-operator.v2.1.1
        version: 2.1.1
      name: stable
    defaultChannel: stable
    packageName: community-trivy-operator
    provider:
      name: devopstales
      url: devopstales.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: www.ibm.com
    name: ibm-elasticsearch-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-elasticsearch-operator.v1.1.2667
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
          createdAt: "2025-01-30"
          description: Elasticsearch operator for use within Cloudpak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-elasticsearch-operator
          olm.skipRange: <0.0.0 >1.1.0
          operators.openshift.io/valid-subscription: '["IBM Elasticsearch"]'
          operators.operatorframework.io/internal-objects: '["elasticsearchclusters.elasticsearch.opencontent.ibm.com"]'
          operators.operatorframework.io/operator-type: non-standalone
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ElasticsearchCluster
            name: elasticsearchclusters.elasticsearch.opencontent.ibm.com
            version: v1
        description: |-
          # Introduction
          An Elasticsearch operator for use with IBM Cloudpak services.
          # Details
          ibm-elasticsearch-operator supports installing Elasticsearch major version 6 and 7. This operator supports two CRDs Elasticsearch and Elasticsearchdetails. Currently Elasticsearch is an abstraction of Elasticsearchdetails, but there is no difference between both the CRDs. Creating either of them provisions elasticsearch instances. This abstraction layer has been created to support any changes to the operator with out breaking the CRD/API for adopting products.

          ## Prerequisites
          - Openshift 3.11+
          ## Resources Required
          In short the answer is it depends. Especially if you are running Elasticsearch 6 you need 3 or a greater odd number of replicas to make a cluster and not have split brain. Each replica will have a sidecar that needs 0.1 cpu and 2gb of ram. You should size Elasticsearch based on your use-case. The default is a full cpu and 4 cpu and 64 gb ram. To assist sizing requirements we have included a copy of [esrally](https://esrally.readthedocs.io/en/stable/) on all the docker images.
          ## Installing
          ElasticsearchCluster is the strategic Kind moving forward. There are two flavors of CRD supported that will soon be deprecated. A simple Elasticseach "All in one" crd and a more detailed CRD for teams who need greater control. Upon deploy the elasticsearch is available from the service (not the headless) created by the operator. Only https will be available on port 443. the service will also be in the ElasticsearchDetails status.

          ### Case install
          To install the component using case you must either use the native install or catalog install. Upon installing the catalog you can then use either OLM or ODLM to install but must create the OperatorGroup & Subscription for either method to avoid conflict.

          # Configuration
          Architecture:
            - x86
            - s390x
            - ppc64le
          Storage: RWO Encrypted Block Storage with 120+ IOPS/GB
          Clouds:
          - IBM CLoud (ibm-block-gold, Gen2 custom iops storage class)
          - AWS (EBS)
          - Azure (For Azure please ensure you are using a higher tier of AzureDisk storage)
          - OCS (RBD not Ceph)
          - Portworx (Non-shared replacated)
          - Please do not use with NFS

          ### Limitations
          This product is not deployable standalone and it is part of the LICCR already accepted
          ## SecurityContextConstraints Requirements
          This component requires a restricted SCC.
        displayName: IBM OpenContent Elasticsearch
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - internal
        - Elasticsearch
        maintainers:
        - email: noreply@ibm.com
          name: IBM CloudpakOpen Community
        maturity: alpha
        provider:
          name: IBM
          url: www.ibm.com
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:2c39ff1850b7a9d32a11e99393039ef74ab363f3fffec1e4fd542f2ae148a6b4
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:130872119947fe91cbfb8dd82b3baaf78f7adcaca5933ca394eca3f9ee84c8cc
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:47a3745c78280f98f2ee352bb9192cdcd2fcd6d4c1a3627e7aa42178a0f107fd
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:30a5b2c4e3c231373fa18d8aaab16d3505cc11a8a5d4be36eaf99d7c0dee35f2
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:2668a562eb8df1117c40acbc3600d6735469d210e5aae4aa03bf9f54d0ebfe46
        version: 1.1.2667
      entries:
      - name: ibm-elasticsearch-operator.v1.1.2667
        version: 1.1.2667
      name: v1.1
    defaultChannel: v1.1
    packageName: ibm-elasticsearch-operator
    provider:
      name: IBM
      url: www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: ovms-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openvino-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "ModelServer",
                "metadata": {
                  "name": "model-server-sample"
                },
                "spec": {
                  "image_name": "registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0",
                  "deployment_parameters": {
                    "replicas": 1
                  },
                  "service_parameters": {
                    "grpc_port": 8080,
                    "rest_port": 8081
                  },
                  "models_settings": {
                    "single_model_mode": true,
                    "config_configmap_name": "",
                    "config_path": "",
                    "model_name": "resnet",
                    "model_path": "gs://<bucket_name>/<model_dir>",
                    "nireq": 0,
                    "plugin_config": "{\"PERFORMANCE_HINT\":\"LATENCY\"}",
                    "batch_size": "",
                    "shape": "",
                    "model_version_policy": "{\"latest\": { \"num_versions\":1 }}",
                    "layout": "",
                    "target_device": "CPU",
                    "is_stateful": false,
                    "idle_sequence_cleanup": false,
                    "low_latency_transformation": true,
                    "max_sequence_number": 0
                  },
                  "server_settings": {
                    "file_system_poll_wait_seconds": 0,
                    "sequence_cleaner_poll_wait_minutes": 0,
                    "log_level": "INFO",
                    "grpc_workers": 1,
                    "rest_workers": 0
                  },
                  "models_repository": {
                    "https_proxy": "",
                    "http_proxy": "",
                    "storage_type": "google",
                    "models_host_path": "",
                    "models_volume_claim": "",
                    "aws_secret_access_key": "",
                    "aws_access_key_id": "",
                    "aws_region": "",
                    "s3_compat_api_endpoint": "",
                    "gcp_creds_secret_name": "",
                    "azure_storage_connection_string": ""
                  }
                }
              },
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "Notebook",
                "metadata": {
                  "name": "v2024.1"
                },
                "spec": {
                  "git_ref": "2024.1",
                  "auto_update_image": false
                }
              }
            ]
          alm-examples-metadata: '{ "model-server-sample":{"description":"Example
            of the model server configuration"}, "v2024":{"description":"Default notebook
            configuration"} }'
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
          createdAt: "2024-05-14T09:29:35Z"
          description: OpenVINO Toolkit Operator manages OpenVINO components in OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.34.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Provides inference serving through gRPC & REST
            displayName: Model Server
            kind: ModelServer
            name: modelservers.intel.com
            version: v1alpha1
          - description: Adds a collection of ready-to-run Jupyter notebooks for learning
              and experimenting with the OpenVINO™ Toolkit to Red Hat® OpenShift®
              Data Science
            displayName: Notebook
            kind: Notebook
            name: notebooks.intel.com
            version: v1alpha1
        description: |
          OpenVINO Toolkit Operator manages OpenVINO components in OpenShift.

          Currently there available components are ModelServer and Notebook.

          # Model Server
          [OpenVINO™ Model Server](https://github.com/openvinotoolkit/model_server) (OVMS) is a scalable, high-performance solution for serving machine learning models optimized for Intel® architectures. The server provides an inference service via gRPC or REST API for any models trained in a framework that is supported by [OpenVINO](https://software.intel.com/en-us/openvino-toolki://docs.openvino.ai/latest/index.html).
          Model Server configuration parameters is explained [here](https://github.com/openvinotoolkit/operator/blob/main/docs/modelserver_params.md).
          ## Using the cluster
          OpenVINO Model Server can be consumed as a `Service`. It is called like with `ModelServer` resource with `-ovms` suffix.
          The suffix is omitted when `ovms` phrase is included in the name.
          The service exposes gRPC and REST API interfaces to run inference requests.
          ```
          oc get pods
          NAME                                        READY   STATUS    RESTARTS   AGE
          model-server-sample-ovms-586f6f76df-dpps4   1/1     Running   0          8h

          oc get services
          NAME                       TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)             AGE
          model-server-sample-ovms   ClusterIP   172.25.199.210   <none>        8080/TCP,8081/TCP   8h
          ```

          # Notebook
          The Notebook resource integrates JupyterHub from OpenShift Data Science or Open Data Hub with a container image that includes [developer
          tools](https://github.com/openvinotoolkit/openvino/blob/master/docs/install_guides/pypi-openvino-dev.md)
          from the OpenVINO toolkit and a set of Jupyter notebook tutorials. It enables selecting a defined image `OpenVINO™ Toolkit` from
          the Jupyter Spawner choice list.

          Create the `Notebook` resource in the same project with JupyterHub and RedHat OpenShift Data Science operator as seen in [documentation](https://github.com/openvinotoolkit/operator/blob/main/docs/notebook_in_rhods.md).
          It builds the image in the cluster based on Dockerfile from [openvino_notebooks](https://github.com/openvinotoolkit/openvino_notebooks).

          ## References
          OpenVINO Model Server on [Github](https://github.com/openvinotoolkit/model_server)

          OpenVINO Model Server Operator on [Github](https://github.com/openvinotoolkit/operator)
        displayName: OpenVINO Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Inference OpenVINO
        links:
        - name: OpenVINO Toolkit Operator
          url: https://github.com/openvinotoolkit/operator
        - name: OpenVINO Model Server
          url: https://github.com/openvinotoolkit/model_server
        - name: OpenVINO Notebooks
          url: https://github.com/openvinotoolkit/openvino_notebooks
        maintainers:
        - email: openvino-operator-support@intel.com
          name: Operator Support
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5714b93ac39277affdc10c3ba142d8c032ed0ddb2ba61c04cbddac710464c4
        - registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
        - registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0
        version: 1.2.0
      entries:
      - name: openvino-operator.v1.2.0
        version: 1.2.0
      - name: openvino-operator.v1.1.0
        version: 1.1.0
      - name: openvino-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: ovms-operator
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx-openshift-operator.v0.0.33
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
          createdAt: "2019-07-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx-operator-rhmp/support?utm_source=openshift_console
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX on JBoss EAP 7
            displayName: Joget DX on JBoss EAP 7
            kind: JogetDX
            name: jogetdx.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX7/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
        - registry.connect.redhat.com/joget/joget-dx7-eap7@sha256:7245501f5fccf011d75c034b7d55542cfe3bd2c64cc7e48b3224a48b623af282
        version: 0.0.33
      entries:
      - name: joget-dx-openshift-operator.v0.0.33
        version: 0.0.33
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx-operator-rhmp
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/cluster-nfd-operator
    name: openshift-nfd-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfd.v4.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nfd.openshift.io/v1",
                "kind": "NodeFeatureDiscovery",
                "metadata": {
                  "name": "nfd-instance",
                  "namespace": "openshift-nfd"
                },
                "spec": {
                  "customConfig": {
                    "configData": "#    - name: \"more.kernel.features\"\n#      matchOn:\n#      - loadedKMod: [\"example_kmod3\"]\n#    - name: \"more.features.by.nodename\"\n#      value: customValue\n#      matchOn:\n#      - nodename: [\"special-.*-node-.*\"]\n"
                  },
                  "operand": {
                    "image": "quay.io/openshift/origin-node-feature-discovery:4.10",
                    "servicePort": 12000
                  },
                  "workerConfig": {
                    "configData": "core:\n#  labelWhiteList:\n#  noPublish: false\n  sleepInterval: 60s\n#  sources: [all]\n#  klog:\n#    addDirHeader: false\n#    alsologtostderr: false\n#    logBacktraceAt:\n#    logtostderr: true\n#    skipHeaders: false\n#    stderrthreshold: 2\n#    v: 0\n#    vmodule:\n##   NOTE: the following options are not dynamically run-time \n##          configurable and require a nfd-worker restart to take effect\n##          after being changed\n#    logDir:\n#    logFile:\n#    logFileMaxSize: 1800\n#    skipLogHeaders: false\nsources:\n#  cpu:\n#    cpuid:\n##     NOTE: whitelist has priority over blacklist\n#      attributeBlacklist:\n#        - \"BMI1\"\n#        - \"BMI2\"\n#        - \"CLMUL\"\n#        - \"CMOV\"\n#        - \"CX16\"\n#        - \"ERMS\"\n#        - \"F16C\"\n#        - \"HTT\"\n#        - \"LZCNT\"\n#        - \"MMX\"\n#        - \"MMXEXT\"\n#        - \"NX\"\n#        - \"POPCNT\"\n#        - \"RDRAND\"\n#        - \"RDSEED\"\n#        - \"RDTSCP\"\n#        - \"SGX\"\n#        - \"SSE\"\n#        - \"SSE2\"\n#        - \"SSE3\"\n#        - \"SSE4.1\"\n#        - \"SSE4.2\"\n#        - \"SSSE3\"\n#      attributeWhitelist:\n#  kernel:\n#    kconfigFile: \"/path/to/kconfig\"\n#    configOpts:\n#      - \"NO_HZ\"\n#      - \"X86\"\n#      - \"DMI\"\n  pci:\n    deviceClassWhitelist:\n      - \"0200\"\n      - \"03\"\n      - \"12\"\n    deviceLabelFields:\n#      - \"class\"\n      - \"vendor\"\n#      - \"device\"\n#      - \"subsystem_vendor\"\n#      - \"subsystem_device\"\n#  usb:\n#    deviceClassWhitelist:\n#      - \"0e\"\n#      - \"ef\"\n#      - \"fe\"\n#      - \"ff\"\n#    deviceLabelFields:\n#      - \"class\"\n#      - \"vendor\"\n#      - \"device\"\n#  custom:\n#    - name: \"my.kernel.feature\"\n#      matchOn:\n#        - loadedKMod: [\"example_kmod1\", \"example_kmod2\"]\n#    - name: \"my.pci.feature\"\n#      matchOn:\n#        - pciId:\n#            class: [\"0200\"]\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#        - pciId :\n#            vendor: [\"8086\"]\n#            device: [\"1000\", \"1100\"]\n#    - name: \"my.usb.feature\"\n#      matchOn:\n#        - usbId:\n#          class: [\"ff\"]\n#          vendor: [\"03e7\"]\n#          device: [\"2485\"]\n#        - usbId:\n#          class: [\"fe\"]\n#          vendor: [\"1a6e\"]\n#          device: [\"089a\"]\n#    - name: \"my.combined.feature\"\n#      matchOn:\n#        - pciId:\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#          loadedKMod : [\"vendor_kmod1\", \"vendor_kmod2\"]\n"
                  }
                }
              },
              {
                "apiVersion": "nfd.openshift.io/v1alpha1",
                "kind": "NodeFeatureRule",
                "metadata": {
                  "name": "my-sample-rule-object"
                },
                "spec": {
                  "rules": [
                    {
                      "labels": {
                        "my-sample-feature": "true"
                      },
                      "matchFeatures": [
                        {
                          "feature": "kernel.loadedmodule",
                          "matchExpressions": {
                            "dummy": {
                              "op": "Exists"
                            }
                          }
                        },
                        {
                          "feature": "kernel.config",
                          "matchExpressions": {
                            "X86": {
                              "op": "In",
                              "value": [
                                "y"
                              ]
                            }
                          }
                        }
                      ],
                      "name": "my sample rule"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery,OpenShift Optional
          containerImage: quay.io/openshift/origin-cluster-nfd-operator:4.10
          description: |-
            The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

            NFD consists  of the following software components:

            The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

            ##NFD-Master
            NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

            ##NFD-Worker
            NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

            ##NFD-Topology-Updater
            NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
          olm.skipRange: '>=4.6.0 <4.11.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-nfd
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/internal-objects: '["noderesourcetopologies.topology.node.k8s.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/cluster-nfd-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              The NodeFeatureDiscovery instance is the CustomResource being watched by the NFD-Operator, and holds all the needed information to setup the behaviour of the master and worker pods
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
          - description: |
              NodeFeatureRule resource specifies a configuration for feature-based customization of node objects, such as node labeling.
            kind: NodeFeatureRule
            name: nodefeaturerules.nfd.openshift.io
            version: v1alpha1
          - description: |
              NodeResourceTopology resource describes node resources and their topology.
            kind: NodeResourceTopology
            name: noderesourcetopologies.topology.node.k8s.io
            version: v1alpha1
        description: |-
          The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

          NFD consists  of the following software components:

          The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

          ##NFD-Master
          NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

          ##NFD-Worker
          NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

          ##NFD-Topology-Updater
          NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
        displayName: Node Feature Discovery Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - feature-discovery
        - feature-detection
        - node-labels
        links:
        - name: Node Feature Discovery Operator
          url: https://docs.openshift.com/container-platform/4.9/hardware_enablement/psap-node-feature-discovery-operator.html
        - name: Node Feature Discovery Documentation
          url: https://kubernetes-sigs.github.io/node-feature-discovery/stable/get-started/index.html
        maintainers:
        - email: openshift-psap@redhat.com
          name: openshift-psap
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/cluster-nfd-operator
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy
        - quay.io/openshift/origin-cluster-nfd-operator:4.10
        version: 4.10.0
      entries:
      - name: nfd.v4.10.0
        version: 4.10.0
      - name: nfd.v4.9.0
        version: 4.9.0
      - name: nfd.v4.8.0
        version: 4.8.0
      name: stable
    defaultChannel: stable
    packageName: openshift-nfd-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/cluster-nfd-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kepler Operator Contributors
      provider-url: https://sustainable-computing.io/
    name: kepler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kepler-operator.v0.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "Kepler",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "kepler",
                    "app.kubernetes.io/name": "kepler",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "kepler"
                },
                "spec": {
                  "exporter": {
                    "deployment": {
                      "port": 9103
                    },
                    "redfish": null
                  }
                }
              },
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "PowerMonitor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "powermonitor",
                    "app.kubernetes.io/name": "powermonitor",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "power-monitor"
                },
                "spec": {
                  "kepler": {
                    "config": {
                      "logLevel": "info"
                    },
                    "deployment": {
                      "security": {
                        "mode": "none"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: quay.io/sustainable_computing_io/kepler-operator:0.20.0
          createdAt: "2025-07-18T06:13:30Z"
          description: Deploys and Manages Kepler on Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: |-
            [
              "keplerinternals.kepler.system.sustainable.computing.io",
              "powermonitorinternals.kepler.system.sustainable.computing.io"
            ]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/sustainable-computing-io/kepler-operator
          support: https://github.com/sustainable-computing-io/kepler-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KeplerInternal is the Schema for internal/unsupported API
            displayName: KeplerInternal
            kind: KeplerInternal
            name: keplerinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: Kepler is deprecated and will be removed in a future release.
              Please use PowerMonitor instead.
            displayName: Kepler
            kind: Kepler
            name: keplers.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitorInternal is the Schema for internal/unsupported
              API
            displayName: PowerMonitorInternal
            kind: PowerMonitorInternal
            name: powermonitorinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitor is the Schema for the powermonitors API
            displayName: PowerMonitor
            kind: PowerMonitor
            name: powermonitors.kepler.system.sustainable.computing.io
            version: v1alpha1
        description: |-
          Kepler (Kubernetes-based Efficient Power Level Exporter) is a Prometheus exporter that
          measures energy consumption metrics at the container, pod, and node level in Kubernetes clusters.

          ### Deprecation Notice
          **Important:** The Kepler CR (Custom Resource) is deprecated and will be removed in a future release. Please use the new PowerMonitor instead.

          ### Prerequisites
          To view the metrics exposed by Kepler, **[User Workload Monitoring](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/monitoring/configuring-user-workload-monitoring)**
          MUST be enabled in your cluster.

          ### Documentation
          * [User Documentation](https://github.com/sustainable-computing-io/kepler-operator/blob/v1alpha1/README.md)
          * [Developer Guide](https://github.com/sustainable-computing-io/kepler-operator/tree/v1alpha1/docs/developer)

          ### Contributing
          * [Report Bugs](https://github.com/sustainable-computing-io/kepler-operator/issues/new?template=bug_report.md) - Found a bug? Let us know!
          * [Browse Issues](https://github.com/sustainable-computing-io/kepler-operator/issues) - See what we're working on

          ### License
          Licensed under [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)
        displayName: Kepler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kepler
        - power
        - energy
        - monitoring
        links:
        - name: Installation Guide
          url: https://sustainable-computing.io/installation/community-operator/
        - name: Kepler Project
          url: https://sustainable-computing.io/
        maintainers:
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: vprashar@redhat.com
          name: Vibhu Prashar
        - email: vimalkum@redhat.com
          name: Vimal Kumar
        - email: kaliu@redhat.com
          name: Kaiyi Liu
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Kepler Operator Contributors
          url: https://sustainable-computing.io/
        relatedImages:
        - quay.io/sustainable_computing_io/kepler:v0.10.2
        - quay.io/brancz/kube-rbac-proxy:v0.19.0
        - quay.io/sustainable_computing_io/kepler-operator:0.20.0
        version: 0.20.0
      entries:
      - name: kepler-operator.v0.20.0
        version: 0.20.0
      - name: kepler-operator.v0.19.0
        version: 0.19.0
      - name: kepler-operator.v0.18.0
        version: 0.18.0
      - name: kepler-operator.v0.17.0
        version: 0.17.0
      - name: kepler-operator.v0.16.0
        version: 0.16.0
      - name: kepler-operator.v0.15.0
        version: 0.15.0
      name: alpha
    defaultChannel: alpha
    packageName: kepler-operator
    provider:
      name: Kepler Operator Contributors
      url: https://sustainable-computing.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MuleSoft
      provider-url: anypoint.mulesoft.com
    name: runtime-fabric-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runtime-fabric-operator.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rtf.operators.mulesoft.com/v1",
                "kind": "RuntimeFabric",
                "metadata": {
                  "name": "runtime-fabric"
                },
                "spec": {
                  "activationData": null,
                  "agent": {
                    "replicas": 1,
                    "app": {
                      "resources": {
                        "limits": {
                          "cpu": "1000m",
                          "memory": "500Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      }
                    },
                    "rtfd": {
                      "resources": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "150m",
                          "memory": "200Mi"
                        }
                      },
                      "watch": {
                        "enabled": true,
                        "components": {
                          "deployments": {
                            "enabled": true
                          },
                          "ingress": {
                            "enabled": true
                          },
                          "secrets": {
                            "enabled": true
                          },
                          "persistencegateways": {
                            "enabled": true
                          },
                          "nodes": {
                            "enabled": true
                          },
                          "jobs": {
                            "enabled": true
                          },
                          "routes": {
                            "enabled": true
                          }
                        }
                      }
                    }
                  },
                  "resource-cache": {
                    "replicas": 1,
                    "nginx": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "250Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    },
                    "fetcher": {
                      "resources": {
                        "limits": {
                          "cpu": "50m",
                          "memory": "100Mi"
                        },
                        "requests": {
                          "cpu": "50m",
                          "memory": "50Mi"
                        }
                      }
                    }
                  },
                  "mule-clusterip-service": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "global": {
                    "cluster": {
                      "flavor": "byok",
                      "type": "openshift"
                    },
                    "containerLogPaths": [
                      "/var/lib/docker/containers",
                      "/var/log/containers",
                      "/var/log/pods"
                    ],
                    "core": {
                      "affinity": {},
                      "tolerations": [],
                      "topologySpreadConstraints": [
                        {
                          "maxSkew": 1,
                          "topologyKey": "topology.kubernetes.io/zone",
                          "whenUnsatisfiable": "ScheduleAnyway"
                        }
                      ]
                    },
                    "worker": {
                      "affinity": {},
                      "tolerations": []
                    },
                    "deployment": {
                      "nodeSelector": ""
                    },
                    "image": {
                      "provider": "aws",
                      "pullSecretName": "rtf-pull-secret",
                      "rtfRegistry": "rtf-runtime-registry.kprod.msap.io",
                      "repositoryNamespace": "mulesoft"
                    },
                    "service": {
                      "clusterIp": null
                    },
                    "authorizedNamespaces": false,
                    "fipsEnabled": false,
                    "crds": {
                      "install": true,
                      "keep": true,
                      "flex": {
                        "enabled": false
                      }
                    },
                    "nodeWatcherEnabled": true,
                    "nodeReconciliation": true,
                    "deploymentRateLimitPerSecond": 1
                  },
                  "muleLicense": null,
                  "proxy": {
                    "http_no_proxy": null,
                    "http_proxy": null,
                    "monitoring_proxy": null
                  },
                  "customLog4jEnabled": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
          createdAt: "2025-07-31T21:12:41Z"
          description: Runtime Fabric Operator Application from Mulesoft
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mulesoft/rtf-ocp-operator
          support: MuleSoft
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Runtime Fabric Operator from Mulesoft
              **Activation data**
              Used to install Runtime Fabric on a Kubernetes service. Copy this value from the Runtime Fabric setup page in Runtime Manager.
              **Registry URL**
              The registry from which to retrieve the images needed to install and run Runtime Fabric. If you're using a local registry, specify the URL here.
                * Default registry URLs:
                    - US control plane: rtf-runtime-registry.kprod.msap.io
                    - EU control plane: rtf-runtime-registry.kprod-eu.msap.io
              **Registry pull secret**
              Used to retrieve images from the registry. Change this value only if you did so during the operator installation.
              **Mule license key**
              Your company’s Mule license key. The key is not required during install, but it is required to deploy applications to Runtime Fabric. Must be Base64 encoded.
            displayName: Runtime Fabric
            kind: RuntimeFabric
            name: runtimefabrics.rtf.operators.mulesoft.com
            version: v1
        description: |-
          Anypoint Runtime Fabric is a container service that you install in your existing Red Hat OpenShift cluster that enables you to run Mule applications, composite APIs, and API gateways across any environment with centralized management via Anypoint Platform.
          Capabilities of Runtime Fabric include:
          * Isolation between applications by running a separate Mule runtime server per application.
          * Ability to run multiple versions of Mule runtime server on the same set of resources.
          * Scaling applications across multiple replicas.
          * Automated application fail-over.
          * Application management with Anypoint Runtime Manager.
          Runtime Fabric supports the following Red Hat OpenShift deployment options:
          * Red Hat OpenShift Service on AWS
          * Microsoft Azure Red Hat OpenShift
          * Red Hat OpenShift Dedicated
          * Red Hat OpenShift on IBM Cloud
          * Self-managed Red Hat OpenShift editions (Performance Plus, OCP, Kubernetes engine)
          Runtime Fabric supports Red Hat OpenShift versions 4.16 to 4.19.
        displayName: Runtime Fabric Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - agent
        - rtf
        - runtime
        - fabric
        - mulesoft
        links:
        - name: Runtime Fabric Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/62901512fefb1b137ba23946
        - name: Documentation
          url: https://docs.mulesoft.com/release-notes/runtime-fabric/runtime-fabric
        maintainers:
        - email: support@mulesoft.com
          name: Mulesoft
        maturity: stable
        provider:
          name: MuleSoft
          url: anypoint.mulesoft.com
        relatedImages:
        - quay.io/rtf/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
        version: 2.11.0
      entries:
      - name: runtime-fabric-operator.v2.11.0
        version: 2.11.0
      - name: runtime-fabric-operator.v2.10.34
        version: 2.10.34
      - name: runtime-fabric-operator.v2.10.17
        version: 2.10.17
      - name: runtime-fabric-operator.v2.10.0
        version: 2.10.0
      - name: runtime-fabric-operator.v2.9.2
        version: 2.9.2
      - name: runtime-fabric-operator.v2.8.2
        version: 2.8.2
      - name: runtime-fabric-operator.v2.7.0
        version: 2.7.0
      - name: runtime-fabric-operator.v2.6.52
        version: 2.6.52
      name: alpha
    - currentCSV: runtime-fabric-operator.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rtf.operators.mulesoft.com/v1",
                "kind": "RuntimeFabric",
                "metadata": {
                  "name": "runtime-fabric"
                },
                "spec": {
                  "activationData": null,
                  "agent": {
                    "replicas": 1,
                    "app": {
                      "resources": {
                        "limits": {
                          "cpu": "1000m",
                          "memory": "500Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      }
                    },
                    "rtfd": {
                      "resources": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "150m",
                          "memory": "200Mi"
                        }
                      },
                      "watch": {
                        "enabled": true,
                        "components": {
                          "deployments": {
                            "enabled": true
                          },
                          "ingress": {
                            "enabled": true
                          },
                          "secrets": {
                            "enabled": true
                          },
                          "persistencegateways": {
                            "enabled": true
                          },
                          "nodes": {
                            "enabled": true
                          },
                          "jobs": {
                            "enabled": true
                          },
                          "routes": {
                            "enabled": true
                          }
                        }
                      }
                    }
                  },
                  "resource-cache": {
                    "replicas": 1,
                    "nginx": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "250Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    },
                    "fetcher": {
                      "resources": {
                        "limits": {
                          "cpu": "50m",
                          "memory": "100Mi"
                        },
                        "requests": {
                          "cpu": "50m",
                          "memory": "50Mi"
                        }
                      }
                    }
                  },
                  "mule-clusterip-service": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "global": {
                    "cluster": {
                      "flavor": "byok",
                      "type": "openshift"
                    },
                    "containerLogPaths": [
                      "/var/lib/docker/containers",
                      "/var/log/containers",
                      "/var/log/pods"
                    ],
                    "core": {
                      "affinity": {},
                      "tolerations": [],
                      "topologySpreadConstraints": [
                        {
                          "maxSkew": 1,
                          "topologyKey": "topology.kubernetes.io/zone",
                          "whenUnsatisfiable": "ScheduleAnyway"
                        }
                      ]
                    },
                    "worker": {
                      "affinity": {},
                      "tolerations": []
                    },
                    "deployment": {
                      "nodeSelector": ""
                    },
                    "image": {
                      "provider": "aws",
                      "pullSecretName": "rtf-pull-secret",
                      "rtfRegistry": "rtf-runtime-registry.kprod.msap.io",
                      "repositoryNamespace": "mulesoft"
                    },
                    "service": {
                      "clusterIp": null
                    },
                    "authorizedNamespaces": false,
                    "fipsEnabled": false,
                    "crds": {
                      "install": true,
                      "keep": true,
                      "flex": {
                        "enabled": false
                      }
                    },
                    "nodeWatcherEnabled": true,
                    "nodeReconciliation": true,
                    "deploymentRateLimitPerSecond": 1
                  },
                  "muleLicense": null,
                  "proxy": {
                    "http_no_proxy": null,
                    "http_proxy": null,
                    "monitoring_proxy": null
                  },
                  "customLog4jEnabled": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
          createdAt: "2025-07-31T21:12:41Z"
          description: Runtime Fabric Operator Application from Mulesoft
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mulesoft/rtf-ocp-operator
          support: MuleSoft
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Runtime Fabric Operator from Mulesoft
              **Activation data**
              Used to install Runtime Fabric on a Kubernetes service. Copy this value from the Runtime Fabric setup page in Runtime Manager.
              **Registry URL**
              The registry from which to retrieve the images needed to install and run Runtime Fabric. If you're using a local registry, specify the URL here.
                * Default registry URLs:
                    - US control plane: rtf-runtime-registry.kprod.msap.io
                    - EU control plane: rtf-runtime-registry.kprod-eu.msap.io
              **Registry pull secret**
              Used to retrieve images from the registry. Change this value only if you did so during the operator installation.
              **Mule license key**
              Your company’s Mule license key. The key is not required during install, but it is required to deploy applications to Runtime Fabric. Must be Base64 encoded.
            displayName: Runtime Fabric
            kind: RuntimeFabric
            name: runtimefabrics.rtf.operators.mulesoft.com
            version: v1
        description: |-
          Anypoint Runtime Fabric is a container service that you install in your existing Red Hat OpenShift cluster that enables you to run Mule applications, composite APIs, and API gateways across any environment with centralized management via Anypoint Platform.
          Capabilities of Runtime Fabric include:
          * Isolation between applications by running a separate Mule runtime server per application.
          * Ability to run multiple versions of Mule runtime server on the same set of resources.
          * Scaling applications across multiple replicas.
          * Automated application fail-over.
          * Application management with Anypoint Runtime Manager.
          Runtime Fabric supports the following Red Hat OpenShift deployment options:
          * Red Hat OpenShift Service on AWS
          * Microsoft Azure Red Hat OpenShift
          * Red Hat OpenShift Dedicated
          * Red Hat OpenShift on IBM Cloud
          * Self-managed Red Hat OpenShift editions (Performance Plus, OCP, Kubernetes engine)
          Runtime Fabric supports Red Hat OpenShift versions 4.16 to 4.19.
        displayName: Runtime Fabric Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - agent
        - rtf
        - runtime
        - fabric
        - mulesoft
        links:
        - name: Runtime Fabric Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/62901512fefb1b137ba23946
        - name: Documentation
          url: https://docs.mulesoft.com/release-notes/runtime-fabric/runtime-fabric
        maintainers:
        - email: support@mulesoft.com
          name: Mulesoft
        maturity: stable
        provider:
          name: MuleSoft
          url: anypoint.mulesoft.com
        relatedImages:
        - quay.io/rtf/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
        version: 2.11.0
      entries:
      - name: runtime-fabric-operator.v2.11.0
        version: 2.11.0
      - name: runtime-fabric-operator.v2.10.34
        version: 2.10.34
      - name: runtime-fabric-operator.v2.10.17
        version: 2.10.17
      - name: runtime-fabric-operator.v2.10.0
        version: 2.10.0
      - name: runtime-fabric-operator.v2.9.2
        version: 2.9.2
      - name: runtime-fabric-operator.v2.8.2
        version: 2.8.2
      - name: runtime-fabric-operator.v2.7.0
        version: 2.7.0
      - name: runtime-fabric-operator.v2.6.52
        version: 2.6.52
      name: stable
    defaultChannel: stable
    packageName: runtime-fabric-operator
    provider:
      name: MuleSoft
      url: anypoint.mulesoft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-codeartifact-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-codeartifact-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/codeartifact-controller:1.1.1
          createdAt: "2025-09-19T22:00:24Z"
          description: AWS codeartifact controller is a service controller for managing
            codeartifact resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Domain represents the state of an AWS codeartifact Domain
              resource.
            displayName: Domain
            kind: Domain
            name: domains.codeartifact.services.k8s.aws
            version: v1alpha1
          - description: PackageGroup represents the state of an AWS codeartifact
              PackageGroup resource.
            displayName: PackageGroup
            kind: PackageGroup
            name: packagegroups.codeartifact.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon codeartifact resources in AWS from within your Kubernetes cluster.

          **About Amazon codeartifact**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon codeartifact
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - codeartifact
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon codeartifact Developer Resources
          url: https://aws.amazon.com/codeartifact/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: codeartifact maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/codeartifact-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-codeartifact-controller.v1.1.1
        version: 1.1.1
      - name: ack-codeartifact-controller.v1.1.0
        version: 1.1.0
      - name: ack-codeartifact-controller.v1.0.13
        version: 1.0.13
      - name: ack-codeartifact-controller.v1.0.11
        version: 1.0.11
      name: alpha
    defaultChannel: alpha
    packageName: ack-codeartifact-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-elasticache-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-elasticache-controller.v1.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "CacheParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "CacheSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "ReplicationGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "Snapshot",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "UserGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/elasticache-controller:1.2.3
          createdAt: "2025-09-19T19:19:33Z"
          description: AWS ElastiCache controller is a service controller for managing
            ElastiCache resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CacheCluster represents the state of an AWS elasticache CacheCluster
              resource.
            displayName: CacheCluster
            kind: CacheCluster
            name: cacheclusters.elasticache.services.k8s.aws
            version: v1alpha1
          - description: CacheParameterGroup represents the state of an AWS elasticache
              CacheParameterGroup resource.
            displayName: CacheParameterGroup
            kind: CacheParameterGroup
            name: cacheparametergroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: CacheSubnetGroup represents the state of an AWS elasticache
              CacheSubnetGroup resource.
            displayName: CacheSubnetGroup
            kind: CacheSubnetGroup
            name: cachesubnetgroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ReplicationGroup represents the state of an AWS elasticache
              ReplicationGroup resource.
            displayName: ReplicationGroup
            kind: ReplicationGroup
            name: replicationgroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCache represents the state of an AWS elasticache
              ServerlessCache resource.
            displayName: ServerlessCache
            kind: ServerlessCache
            name: serverlesscaches.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCacheSnapshot represents the state of an AWS elasticache
              ServerlessCacheSnapshot resource.
            displayName: ServerlessCacheSnapshot
            kind: ServerlessCacheSnapshot
            name: serverlesscachesnapshots.elasticache.services.k8s.aws
            version: v1alpha1
          - description: Snapshot represents the state of an AWS elasticache Snapshot
              resource.
            displayName: Snapshot
            kind: Snapshot
            name: snapshots.elasticache.services.k8s.aws
            version: v1alpha1
          - description: UserGroup represents the state of an AWS elasticache UserGroup
              resource.
            displayName: UserGroup
            kind: UserGroup
            name: usergroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS elasticache User resource.
            displayName: User
            kind: User
            name: users.elasticache.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ElastiCache resources in AWS from within your Kubernetes cluster.

          **About Amazon ElastiCache**

          Amazon ElastiCache allows you to seamlessly set up, run, and scale popular open-source compatible in-memory data stores in the cloud. Build data-intensive apps or boost the performance of your existing databases by retrieving data from high throughput and low latency in-memory data stores. Amazon ElastiCache is a popular choice for real-time use cases like Caching, Session Stores, Gaming, Geospatial Services, Real-Time Analytics, and Queuing.

          Amazon ElastiCache offers fully managed Redis, voted the most loved database by developers in the Stack Overflow 2020 Developer Survey, and Memcached for your most demanding applications that require sub-millisecond response times.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ElastiCache
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticache
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ElastiCache Developer Resources
          url: https://aws.amazon.com/elasticache/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: elasticache maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/elasticache-controller:1.2.3
        version: 1.2.3
      entries:
      - name: ack-elasticache-controller.v1.2.3
        version: 1.2.3
      - name: ack-elasticache-controller.v1.2.2
        version: 1.2.2
      - name: ack-elasticache-controller.v1.2.1
        version: 1.2.1
      - name: ack-elasticache-controller.v1.2.0
        version: 1.2.0
      - name: ack-elasticache-controller.v1.1.0
        version: 1.1.0
      - name: ack-elasticache-controller.v1.0.7
        version: 1.0.7
      - name: ack-elasticache-controller.v1.0.6
        version: 1.0.6
      - name: ack-elasticache-controller.v1.0.5
        version: 1.0.5
      - name: ack-elasticache-controller.v1.0.4
        version: 1.0.4
      - name: ack-elasticache-controller.v1.0.3
        version: 1.0.3
      - name: ack-elasticache-controller.v1.0.2
        version: 1.0.2
      - name: ack-elasticache-controller.v1.0.1
        version: 1.0.1
      - name: ack-elasticache-controller.v1.0.0
        version: 1.0.0
      - name: ack-elasticache-controller.v0.2.6
        version: 0.2.6
      - name: ack-elasticache-controller.v0.2.5
        version: 0.2.5
      - name: ack-elasticache-controller.v0.2.4
        version: 0.2.4
      - name: ack-elasticache-controller.v0.2.3
        version: 0.2.3
      - name: ack-elasticache-controller.v0.2.2
        version: 0.2.2
      - name: ack-elasticache-controller.v0.2.1
        version: 0.2.1
      - name: ack-elasticache-controller.v0.2.0
        version: 0.2.0
      - name: ack-elasticache-controller.v0.1.1
        version: 0.1.1
      - name: ack-elasticache-controller.v0.1.0
        version: 0.1.0
      - name: ack-elasticache-controller.v0.0.31
        version: 0.0.31
      - name: ack-elasticache-controller.v0.0.30
        version: 0.0.30
      - name: ack-elasticache-controller.v0.0.29
        version: 0.0.29
      - name: ack-elasticache-controller.v0.0.28
        version: 0.0.28
      - name: ack-elasticache-controller.v0.0.27
        version: 0.0.27
      - name: ack-elasticache-controller.v0.0.26
        version: 0.0.26
      - name: ack-elasticache-controller.v0.0.25
        version: 0.0.25
      - name: ack-elasticache-controller.v0.0.24
        version: 0.0.24
      - name: ack-elasticache-controller.v0.0.23
        version: 0.0.23
      - name: ack-elasticache-controller.v0.0.22
        version: 0.0.22
      - name: ack-elasticache-controller.v0.0.21
        version: 0.0.21
      - name: ack-elasticache-controller.v0.0.20
        version: 0.0.20
      - name: ack-elasticache-controller.v0.0.19
        version: 0.0.19
      - name: ack-elasticache-controller.v0.0.18
        version: 0.0.18
      - name: ack-elasticache-controller.v0.0.17
        version: 0.0.17
      - name: ack-elasticache-controller.v0.0.16
        version: 0.0.16
      - name: ack-elasticache-controller.v0.0.15
        version: 0.0.15
      - name: ack-elasticache-controller.v0.0.14
        version: 0.0.14
      - name: ack-elasticache-controller.v0.0.13
        version: 0.0.13
      - name: ack-elasticache-controller.v0.0.12
        version: 0.0.12
      - name: ack-elasticache-controller.v0.0.11
        version: 0.0.11
      - name: ack-elasticache-controller.v0.0.10
        version: 0.0.10
      - name: ack-elasticache-controller.v0.0.9
        version: 0.0.9
      - name: ack-elasticache-controller.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: ack-elasticache-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ember-csi-community-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ember-csi-community-operator.v0.9.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ember-csi.io/v1alpha1",
                "kind": "EmberStorageBackend",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "config": {
                    "envVars": {
                      "X_CSI_PERSISTENCE_CONFIG": {
                        "storage": "crd"
                      },
                      "X_CSI_DEBUG_MODE": "",
                      "X_CSI_DEFAULT_MOUNT_FS": "ext4",
                      "X_CSI_EMBER_CONFIG": {
                        "grpc_workers": 30,
                        "disable_logs": false,
                        "plugin_name": "",
                        "slow_operations": true,
                        "disabled": [],
                        "enable_probe": false,
                        "debug": false,
                        "user_id": "ember_csi.io",
                        "project_id": "ember_csi.io"
                      },
                      "X_CSI_BACKEND_CONFIG": {
                        "enable_unsupported_driver": false,
                        "driver": "ACCESSIscsi",
                        "name": "",
                        "driver__ACCESSIscsi__vrts_lun_sparse": true,
                        "driver__ACCESSIscsi__vrts_target_config": "/etc/cinder/vrts_target.xml",
                        "driver__AS13000__as13000_ipsan_pools__transform_csv": "Pool0",
                        "driver__AS13000__as13000_meta_pool": "",
                        "driver__AS13000__as13000_token_available_time": 3300,
                        "driver__Acs5000ISCSI__acs5000_copy_interval": 5,
                        "driver__Acs5000ISCSI__acs5000_target": 0,
                        "driver__Acs5000ISCSI__acs5000_volpool_name__transform_csv": "pool01",
                        "driver__DPLFC__dpl_pool": "",
                        "driver__DPLFC__dpl_port": 8357,
                        "driver__DPLISCSI__dpl_pool": "",
                        "driver__DPLISCSI__dpl_port": 8357,
                        "driver__DSWARE__dsware_rest_url": "",
                        "driver__DSWARE__dsware_storage_pools": "",
                        "driver__DSWARE__manager_ips__transform_csv_kvs": "",
                        "driver__Datera__backend_availability_zone": "",
                        "driver__Datera__chap_password": "",
                        "driver__Datera__chap_username": "",
                        "driver__Datera__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__Datera__driver_client_cert": "",
                        "driver__Datera__driver_client_cert_key": "",
                        "driver__Datera__driver_data_namespace": "",
                        "driver__Datera__driver_ssl_cert_path": "",
                        "driver__Datera__driver_ssl_cert_verify": false,
                        "driver__Datera__driver_use_ssl": false,
                        "driver__Datera__filter_function": "",
                        "driver__Datera__goodness_function": "",
                        "driver__Datera__iscsi_iotype": "fileio",
                        "driver__Datera__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__Datera__iscsi_target_flags": "",
                        "driver__Datera__iscsi_write_cache": "on",
                        "driver__Datera__num_shell_tries": 3,
                        "driver__Datera__num_volume_device_scan_tries": 3,
                        "driver__Datera__report_discard_supported": false,
                        "driver__Datera__storage_protocol": "iscsi",
                        "driver__Datera__target_helper": "tgtadm",
                        "driver__Datera__target_ip_address": "$my_ip",
                        "driver__Datera__target_port": 3260,
                        "driver__Datera__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__Datera__target_protocol": "iscsi",
                        "driver__Datera__trace_flags__transform_csv": "",
                        "driver__Datera__use_chap_auth": false,
                        "driver__Datera__volume_backend_name": "",
                        "driver__Datera__volume_clear": "zero",
                        "driver__Datera__volume_clear_ionice": "",
                        "driver__Datera__volume_clear_size": 0,
                        "driver__Datera__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__Datera__volume_copy_bps_limit": 0,
                        "driver__Datera__volume_dd_blocksize": "1M",
                        "driver__Datera__volumes_dir": "$state_path/volumes",
                        "driver__FJDXFC__cinder_eternus_config_file": "/etc/cinder/cinder_fujitsu_eternus_dx.xml",
                        "driver__FJDXISCSI__cinder_eternus_config_file": "/etc/cinder/cinder_fujitsu_eternus_dx.xml",
                        "driver__FlashSystemFC__flashsystem_connection_protocol": "FC",
                        "driver__FlashSystemFC__flashsystem_multihostmap_enabled": true,
                        "driver__FlashSystemISCSI__flashsystem_connection_protocol": "FC",
                        "driver__FlashSystemISCSI__flashsystem_multihostmap_enabled": true,
                        "driver__GPFS__gpfs_images_dir": "",
                        "driver__GPFS__gpfs_images_share_mode__transform_empty_none": "",
                        "driver__GPFS__gpfs_max_clone_depth": 0,
                        "driver__GPFS__gpfs_mount_point_base": "",
                        "driver__GPFS__gpfs_sparse_volumes": true,
                        "driver__GPFS__gpfs_storage_pool": "system",
                        "driver__GPFSRemote__gpfs_hosts__transform_csv": "",
                        "driver__GPFSRemote__gpfs_hosts_key_file": "$state_path/ssh_known_hosts",
                        "driver__GPFSRemote__gpfs_images_dir": "",
                        "driver__GPFSRemote__gpfs_images_share_mode__transform_empty_none": "",
                        "driver__GPFSRemote__gpfs_max_clone_depth": 0,
                        "driver__GPFSRemote__gpfs_mount_point_base": "",
                        "driver__GPFSRemote__gpfs_private_key": "",
                        "driver__GPFSRemote__gpfs_sparse_volumes": true,
                        "driver__GPFSRemote__gpfs_ssh_port": 22,
                        "driver__GPFSRemote__gpfs_storage_pool": "system",
                        "driver__GPFSRemote__gpfs_strict_host_key_policy": false,
                        "driver__GPFSRemote__gpfs_user_login": "root",
                        "driver__GPFSRemote__gpfs_user_password": "",
                        "driver__HBSDFC__backend_availability_zone": "",
                        "driver__HBSDFC__chap_password": "",
                        "driver__HBSDFC__chap_username": "",
                        "driver__HBSDFC__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__HBSDFC__driver_client_cert": "",
                        "driver__HBSDFC__driver_client_cert_key": "",
                        "driver__HBSDFC__driver_data_namespace": "",
                        "driver__HBSDFC__driver_ssl_cert_path": "",
                        "driver__HBSDFC__driver_ssl_cert_verify": false,
                        "driver__HBSDFC__driver_use_ssl": false,
                        "driver__HBSDFC__filter_function": "",
                        "driver__HBSDFC__goodness_function": "",
                        "driver__HBSDFC__iscsi_iotype": "fileio",
                        "driver__HBSDFC__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__HBSDFC__iscsi_target_flags": "",
                        "driver__HBSDFC__iscsi_write_cache": "on",
                        "driver__HBSDFC__num_shell_tries": 3,
                        "driver__HBSDFC__num_volume_device_scan_tries": 3,
                        "driver__HBSDFC__report_discard_supported": false,
                        "driver__HBSDFC__storage_protocol": "iscsi",
                        "driver__HBSDFC__target_helper": "tgtadm",
                        "driver__HBSDFC__target_ip_address": "$my_ip",
                        "driver__HBSDFC__target_port": 3260,
                        "driver__HBSDFC__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__HBSDFC__target_protocol": "iscsi",
                        "driver__HBSDFC__trace_flags__transform_csv": "",
                        "driver__HBSDFC__use_chap_auth": false,
                        "driver__HBSDFC__volume_backend_name": "",
                        "driver__HBSDFC__volume_clear": "zero",
                        "driver__HBSDFC__volume_clear_ionice": "",
                        "driver__HBSDFC__volume_clear_size": 0,
                        "driver__HBSDFC__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__HBSDFC__volume_copy_bps_limit": 0,
                        "driver__HBSDFC__volume_dd_blocksize": "1M",
                        "driver__HBSDFC__volumes_dir": "$state_path/volumes",
                        "driver__HBSDISCSI__backend_availability_zone": "",
                        "driver__HBSDISCSI__chap_password": "",
                        "driver__HBSDISCSI__chap_username": "",
                        "driver__HBSDISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__HBSDISCSI__driver_client_cert": "",
                        "driver__HBSDISCSI__driver_client_cert_key": "",
                        "driver__HBSDISCSI__driver_data_namespace": "",
                        "driver__HBSDISCSI__driver_ssl_cert_path": "",
                        "driver__HBSDISCSI__driver_ssl_cert_verify": false,
                        "driver__HBSDISCSI__driver_use_ssl": false,
                        "driver__HBSDISCSI__filter_function": "",
                        "driver__HBSDISCSI__goodness_function": "",
                        "driver__HBSDISCSI__iscsi_iotype": "fileio",
                        "driver__HBSDISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__HBSDISCSI__iscsi_target_flags": "",
                        "driver__HBSDISCSI__iscsi_write_cache": "on",
                        "driver__HBSDISCSI__num_shell_tries": 3,
                        "driver__HBSDISCSI__num_volume_device_scan_tries": 3,
                        "driver__HBSDISCSI__report_discard_supported": false,
                        "driver__HBSDISCSI__storage_protocol": "iscsi",
                        "driver__HBSDISCSI__target_helper": "tgtadm",
                        "driver__HBSDISCSI__target_ip_address": "$my_ip",
                        "driver__HBSDISCSI__target_port": 3260,
                        "driver__HBSDISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__HBSDISCSI__target_protocol": "iscsi",
                        "driver__HBSDISCSI__trace_flags__transform_csv": "",
                        "driver__HBSDISCSI__use_chap_auth": false,
                        "driver__HBSDISCSI__volume_backend_name": "",
                        "driver__HBSDISCSI__volume_clear": "zero",
                        "driver__HBSDISCSI__volume_clear_ionice": "",
                        "driver__HBSDISCSI__volume_clear_size": 0,
                        "driver__HBSDISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__HBSDISCSI__volume_copy_bps_limit": 0,
                        "driver__HBSDISCSI__volume_dd_blocksize": "1M",
                        "driver__HBSDISCSI__volumes_dir": "$state_path/volumes",
                        "driver__HPE3PARFC__hpe3par_api_url": "",
                        "driver__HPE3PARFC__hpe3par_cpg__transform_csv": "OpenStack",
                        "driver__HPE3PARFC__hpe3par_cpg_snap": "",
                        "driver__HPE3PARFC__hpe3par_debug": false,
                        "driver__HPE3PARFC__hpe3par_iscsi_chap_enabled": false,
                        "driver__HPE3PARFC__hpe3par_iscsi_ips__transform_csv": "",
                        "driver__HPE3PARFC__hpe3par_password": "",
                        "driver__HPE3PARFC__hpe3par_snapshot_expiration": "",
                        "driver__HPE3PARFC__hpe3par_snapshot_retention": "",
                        "driver__HPE3PARFC__hpe3par_target_nsp": "",
                        "driver__HPE3PARFC__hpe3par_username": "",
                        "driver__HPE3PARFC__san_ip": "",
                        "driver__HPE3PARFC__san_login": "admin",
                        "driver__HPE3PARFC__san_password": "",
                        "driver__HPE3PARFC__san_private_key": "",
                        "driver__HPE3PARFC__san_ssh_port": 22,
                        "driver__HPE3PARFC__ssh_conn_timeout": 30,
                        "driver__HPE3PARFC__target_ip_address": "$my_ip",
                        "driver__HPE3PARFC__target_port": 3260,
                        "driver__HPE3PARFC__unique_fqdn_network": true,
                        "driver__HPE3PARISCSI__hpe3par_api_url": "",
                        "driver__HPE3PARISCSI__hpe3par_cpg__transform_csv": "OpenStack",
                        "driver__HPE3PARISCSI__hpe3par_cpg_snap": "",
                        "driver__HPE3PARISCSI__hpe3par_debug": false,
                        "driver__HPE3PARISCSI__hpe3par_iscsi_chap_enabled": false,
                        "driver__HPE3PARISCSI__hpe3par_iscsi_ips__transform_csv": "",
                        "driver__HPE3PARISCSI__hpe3par_password": "",
                        "driver__HPE3PARISCSI__hpe3par_snapshot_expiration": "",
                        "driver__HPE3PARISCSI__hpe3par_snapshot_retention": "",
                        "driver__HPE3PARISCSI__hpe3par_target_nsp": "",
                        "driver__HPE3PARISCSI__hpe3par_username": "",
                        "driver__HPE3PARISCSI__san_ip": "",
                        "driver__HPE3PARISCSI__san_login": "admin",
                        "driver__HPE3PARISCSI__san_password": "",
                        "driver__HPE3PARISCSI__san_private_key": "",
                        "driver__HPE3PARISCSI__san_ssh_port": 22,
                        "driver__HPE3PARISCSI__ssh_conn_timeout": 30,
                        "driver__HPE3PARISCSI__target_ip_address": "$my_ip",
                        "driver__HPE3PARISCSI__target_port": 3260,
                        "driver__HPE3PARISCSI__unique_fqdn_network": true,
                        "driver__HPMSAFC__driver_ssl_cert_path": "",
                        "driver__HPMSAFC__driver_ssl_cert_verify": false,
                        "driver__HPMSAFC__driver_use_ssl": false,
                        "driver__HPMSAFC__hpmsa_pool_name": "A",
                        "driver__HPMSAFC__hpmsa_pool_type": "virtual",
                        "driver__HPMSAFC__san_ip": "",
                        "driver__HPMSAFC__san_login": "admin",
                        "driver__HPMSAFC__san_password": "",
                        "driver__HPMSAISCSI__driver_ssl_cert_path": "",
                        "driver__HPMSAISCSI__driver_ssl_cert_verify": false,
                        "driver__HPMSAISCSI__driver_use_ssl": false,
                        "driver__HPMSAISCSI__hpmsa_iscsi_ips__transform_csv": "",
                        "driver__HPMSAISCSI__hpmsa_pool_name": "A",
                        "driver__HPMSAISCSI__hpmsa_pool_type": "virtual",
                        "driver__HPMSAISCSI__san_ip": "",
                        "driver__HPMSAISCSI__san_login": "admin",
                        "driver__HPMSAISCSI__san_password": "",
                        "driver__HuaweiFC__cinder_huawei_conf_file": "/etc/cinder/cinder_huawei_conf.xml",
                        "driver__HuaweiFC__hypermetro_devices": "",
                        "driver__HuaweiFC__metro_domain_name": "",
                        "driver__HuaweiFC__metro_san_address": "",
                        "driver__HuaweiFC__metro_san_password": "",
                        "driver__HuaweiFC__metro_san_user": "",
                        "driver__HuaweiFC__metro_storage_pools": "",
                        "driver__HuaweiISCSI__cinder_huawei_conf_file": "/etc/cinder/cinder_huawei_conf.xml",
                        "driver__HuaweiISCSI__hypermetro_devices": "",
                        "driver__HuaweiISCSI__metro_domain_name": "",
                        "driver__HuaweiISCSI__metro_san_address": "",
                        "driver__HuaweiISCSI__metro_san_password": "",
                        "driver__HuaweiISCSI__metro_san_user": "",
                        "driver__HuaweiISCSI__metro_storage_pools": "",
                        "driver__IBMStorage__chap": "disabled",
                        "driver__IBMStorage__connection_type": "iscsi",
                        "driver__IBMStorage__management_ips": "",
                        "driver__IBMStorage__proxy": "cinder.volume.drivers.ibm.ibm_storage.proxy.IBMStorageProxy",
                        "driver__InStorageMCSFC__instorage_mcs_allow_tenant_qos": false,
                        "driver__InStorageMCSFC__instorage_mcs_localcopy_rate": 50,
                        "driver__InStorageMCSFC__instorage_mcs_localcopy_timeout": 120,
                        "driver__InStorageMCSFC__instorage_mcs_vol_autoexpand": true,
                        "driver__InStorageMCSFC__instorage_mcs_vol_compression": false,
                        "driver__InStorageMCSFC__instorage_mcs_vol_grainsize": 256,
                        "driver__InStorageMCSFC__instorage_mcs_vol_intier": true,
                        "driver__InStorageMCSFC__instorage_mcs_vol_iogrp": "0",
                        "driver__InStorageMCSFC__instorage_mcs_vol_rsize": 2,
                        "driver__InStorageMCSFC__instorage_mcs_vol_warning": 0,
                        "driver__InStorageMCSFC__instorage_mcs_volpool_name__transform_csv": "volpool",
                        "driver__InStorageMCSFC__instorage_san_secondary_ip": "",
                        "driver__InStorageMCSISCSI__instorage_mcs_allow_tenant_qos": false,
                        "driver__InStorageMCSISCSI__instorage_mcs_localcopy_rate": 50,
                        "driver__InStorageMCSISCSI__instorage_mcs_localcopy_timeout": 120,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_autoexpand": true,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_compression": false,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_grainsize": 256,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_intier": true,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_iogrp": "0",
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_rsize": 2,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_warning": 0,
                        "driver__InStorageMCSISCSI__instorage_mcs_volpool_name__transform_csv": "volpool",
                        "driver__InStorageMCSISCSI__instorage_san_secondary_ip": "",
                        "driver__InfiniboxVolume__chap_password": "",
                        "driver__InfiniboxVolume__chap_username": "",
                        "driver__InfiniboxVolume__infinidat_iscsi_netspaces__transform_csv": "",
                        "driver__InfiniboxVolume__infinidat_pool_name": "",
                        "driver__InfiniboxVolume__infinidat_storage_protocol": "fc",
                        "driver__InfiniboxVolume__infinidat_use_compression": false,
                        "driver__InfiniboxVolume__num_volume_device_scan_tries": 3,
                        "driver__InfiniboxVolume__san_ip": "",
                        "driver__InfiniboxVolume__san_login": "admin",
                        "driver__InfiniboxVolume__san_password": "",
                        "driver__InfiniboxVolume__san_thin_provision": true,
                        "driver__InfiniboxVolume__use_chap_auth": false,
                        "driver__InfiniboxVolume__volume_dd_blocksize": "1M",
                        "driver__InfortrendCLIFC__infortrend_cli_cache": false,
                        "driver__InfortrendCLIFC__infortrend_cli_max_retries": 5,
                        "driver__InfortrendCLIFC__infortrend_cli_path": "/opt/bin/Infortrend/raidcmd_ESDS10.jar",
                        "driver__InfortrendCLIFC__infortrend_cli_timeout": 60,
                        "driver__InfortrendCLIFC__infortrend_iqn_prefix": "iqn.2002-10.com.infortrend",
                        "driver__InfortrendCLIFC__infortrend_pools_name__transform_csv": "",
                        "driver__InfortrendCLIFC__infortrend_slots_a_channels_id__transform_csv": "",
                        "driver__InfortrendCLIFC__infortrend_slots_b_channels_id__transform_csv": "",
                        "driver__InfortrendCLIFC__java_path": "/usr/bin/java",
                        "driver__InfortrendCLIISCSI__infortrend_cli_cache": false,
                        "driver__InfortrendCLIISCSI__infortrend_cli_max_retries": 5,
                        "driver__InfortrendCLIISCSI__infortrend_cli_path": "/opt/bin/Infortrend/raidcmd_ESDS10.jar",
                        "driver__InfortrendCLIISCSI__infortrend_cli_timeout": 60,
                        "driver__InfortrendCLIISCSI__infortrend_iqn_prefix": "iqn.2002-10.com.infortrend",
                        "driver__InfortrendCLIISCSI__infortrend_pools_name__transform_csv": "",
                        "driver__InfortrendCLIISCSI__infortrend_slots_a_channels_id__transform_csv": "",
                        "driver__InfortrendCLIISCSI__infortrend_slots_b_channels_id__transform_csv": "",
                        "driver__InfortrendCLIISCSI__java_path": "/usr/bin/java",
                        "driver__JovianISCSI__backend_availability_zone": "",
                        "driver__JovianISCSI__chap_password": "",
                        "driver__JovianISCSI__chap_username": "",
                        "driver__JovianISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__JovianISCSI__driver_client_cert": "",
                        "driver__JovianISCSI__driver_client_cert_key": "",
                        "driver__JovianISCSI__driver_data_namespace": "",
                        "driver__JovianISCSI__driver_ssl_cert_path": "",
                        "driver__JovianISCSI__driver_ssl_cert_verify": false,
                        "driver__JovianISCSI__driver_use_ssl": false,
                        "driver__JovianISCSI__filter_function": "",
                        "driver__JovianISCSI__goodness_function": "",
                        "driver__JovianISCSI__iscsi_iotype": "fileio",
                        "driver__JovianISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__JovianISCSI__iscsi_target_flags": "",
                        "driver__JovianISCSI__iscsi_write_cache": "on",
                        "driver__JovianISCSI__num_shell_tries": 3,
                        "driver__JovianISCSI__num_volume_device_scan_tries": 3,
                        "driver__JovianISCSI__report_discard_supported": false,
                        "driver__JovianISCSI__storage_protocol": "iscsi",
                        "driver__JovianISCSI__target_helper": "tgtadm",
                        "driver__JovianISCSI__target_ip_address": "$my_ip",
                        "driver__JovianISCSI__target_port": 3260,
                        "driver__JovianISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__JovianISCSI__target_protocol": "iscsi",
                        "driver__JovianISCSI__trace_flags__transform_csv": "",
                        "driver__JovianISCSI__use_chap_auth": false,
                        "driver__JovianISCSI__volume_backend_name": "",
                        "driver__JovianISCSI__volume_clear": "zero",
                        "driver__JovianISCSI__volume_clear_ionice": "",
                        "driver__JovianISCSI__volume_clear_size": 0,
                        "driver__JovianISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__JovianISCSI__volume_copy_bps_limit": 0,
                        "driver__JovianISCSI__volume_dd_blocksize": "1M",
                        "driver__JovianISCSI__volumes_dir": "$state_path/volumes",
                        "driver__KaminarioISCSI__disable_discovery": false,
                        "driver__KaminarioISCSI__san_ip": "",
                        "driver__KaminarioISCSI__san_login": "admin",
                        "driver__KaminarioISCSI__san_password": "",
                        "driver__KaminarioISCSI__unique_fqdn_network": true,
                        "driver__KaminarioISCSI__volume_dd_blocksize": "1M",
                        "driver__LVMVolume__iscsi_iotype": "fileio",
                        "driver__LVMVolume__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__LVMVolume__iscsi_target_flags": "",
                        "driver__LVMVolume__iscsi_write_cache": "on",
                        "driver__LVMVolume__lvm_conf_file": "/etc/cinder/lvm.conf",
                        "driver__LVMVolume__lvm_mirrors": 0,
                        "driver__LVMVolume__lvm_suppress_fd_warnings": false,
                        "driver__LVMVolume__lvm_type": "auto",
                        "driver__LVMVolume__nvmet_port_id": 1,
                        "driver__LVMVolume__scst_target_driver": "iscsi",
                        "driver__LVMVolume__scst_target_iqn_name": "",
                        "driver__LVMVolume__spdk_max_queue_depth": 64,
                        "driver__LVMVolume__spdk_rpc_ip": "",
                        "driver__LVMVolume__spdk_rpc_password": "",
                        "driver__LVMVolume__spdk_rpc_port": 8000,
                        "driver__LVMVolume__spdk_rpc_username": "",
                        "driver__LVMVolume__target_helper": "tgtadm",
                        "driver__LVMVolume__target_ip_address": "$my_ip",
                        "driver__LVMVolume__target_port": 3260,
                        "driver__LVMVolume__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__LVMVolume__target_protocol": "iscsi",
                        "driver__LVMVolume__volume_clear": "zero",
                        "driver__LVMVolume__volume_clear_size": 0,
                        "driver__LVMVolume__volume_dd_blocksize": "1M",
                        "driver__LVMVolume__volume_group": "cinder-volumes",
                        "driver__LVMVolume__volumes_dir": "$state_path/volumes",
                        "driver__LenovoFC__lenovo_pool_name": "A",
                        "driver__LenovoFC__lenovo_pool_type": "virtual",
                        "driver__LenovoISCSI__lenovo_iscsi_ips__transform_csv": "",
                        "driver__LenovoISCSI__lenovo_pool_name": "A",
                        "driver__LenovoISCSI__lenovo_pool_type": "virtual",
                        "driver__LinstorDrbd__linstor_autoplace_count": 0,
                        "driver__LinstorDrbd__linstor_controller_diskless": true,
                        "driver__LinstorDrbd__linstor_default_blocksize": 4096,
                        "driver__LinstorDrbd__linstor_default_storage_pool_name": "DfltStorPool",
                        "driver__LinstorDrbd__linstor_default_uri": "linstor://localhost",
                        "driver__LinstorDrbd__linstor_default_volume_group_name": "drbd-vg",
                        "driver__LinstorDrbd__linstor_volume_downsize_factor__transform_string_float": "4096",
                        "driver__LinstorIscsi__linstor_autoplace_count": 0,
                        "driver__LinstorIscsi__linstor_controller_diskless": true,
                        "driver__LinstorIscsi__linstor_default_blocksize": 4096,
                        "driver__LinstorIscsi__linstor_default_storage_pool_name": "DfltStorPool",
                        "driver__LinstorIscsi__linstor_default_uri": "linstor://localhost",
                        "driver__LinstorIscsi__linstor_default_volume_group_name": "drbd-vg",
                        "driver__LinstorIscsi__linstor_volume_downsize_factor__transform_string_float": "4096",
                        "driver__MStorageFC__nec_actual_free_capacity": false,
                        "driver__MStorageFC__nec_auto_accesscontrol": true,
                        "driver__MStorageFC__nec_backend_max_ld_count": 1024,
                        "driver__MStorageFC__nec_backup_ldname_format": "LX:%s",
                        "driver__MStorageFC__nec_backup_pools__transform_csv": "",
                        "driver__MStorageFC__nec_cv_ldname_format": "LX:__ControlVolume_%xh",
                        "driver__MStorageFC__nec_diskarray_name": "",
                        "driver__MStorageFC__nec_ismcli_fip": "",
                        "driver__MStorageFC__nec_ismcli_password": "",
                        "driver__MStorageFC__nec_ismcli_privkey": "",
                        "driver__MStorageFC__nec_ismcli_user": "",
                        "driver__MStorageFC__nec_ismview_alloptimize": false,
                        "driver__MStorageFC__nec_ismview_dir": "/tmp/nec/cinder",
                        "driver__MStorageFC__nec_ldname_format": "LX:%s",
                        "driver__MStorageFC__nec_ldset": "",
                        "driver__MStorageFC__nec_pools__transform_csv": "",
                        "driver__MStorageFC__nec_queryconfig_view": false,
                        "driver__MStorageFC__nec_ssh_pool_port_number": 22,
                        "driver__MStorageFC__nec_unpairthread_timeout": 3600,
                        "driver__MStorageISCSI__nec_actual_free_capacity": false,
                        "driver__MStorageISCSI__nec_auto_accesscontrol": true,
                        "driver__MStorageISCSI__nec_backend_max_ld_count": 1024,
                        "driver__MStorageISCSI__nec_backup_ldname_format": "LX:%s",
                        "driver__MStorageISCSI__nec_backup_pools__transform_csv": "",
                        "driver__MStorageISCSI__nec_cv_ldname_format": "LX:__ControlVolume_%xh",
                        "driver__MStorageISCSI__nec_diskarray_name": "",
                        "driver__MStorageISCSI__nec_ismcli_fip": "",
                        "driver__MStorageISCSI__nec_ismcli_password": "",
                        "driver__MStorageISCSI__nec_ismcli_privkey": "",
                        "driver__MStorageISCSI__nec_ismcli_user": "",
                        "driver__MStorageISCSI__nec_ismview_alloptimize": false,
                        "driver__MStorageISCSI__nec_ismview_dir": "/tmp/nec/cinder",
                        "driver__MStorageISCSI__nec_ldname_format": "LX:%s",
                        "driver__MStorageISCSI__nec_ldset": "",
                        "driver__MStorageISCSI__nec_pools__transform_csv": "",
                        "driver__MStorageISCSI__nec_queryconfig_view": false,
                        "driver__MStorageISCSI__nec_ssh_pool_port_number": 22,
                        "driver__MStorageISCSI__nec_unpairthread_timeout": 3600,
                        "driver__MacroSANFC__backend_availability_zone": "",
                        "driver__MacroSANFC__chap_password": "",
                        "driver__MacroSANFC__chap_username": "",
                        "driver__MacroSANFC__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__MacroSANFC__driver_client_cert": "",
                        "driver__MacroSANFC__driver_client_cert_key": "",
                        "driver__MacroSANFC__driver_data_namespace": "",
                        "driver__MacroSANFC__driver_ssl_cert_path": "",
                        "driver__MacroSANFC__driver_ssl_cert_verify": false,
                        "driver__MacroSANFC__driver_use_ssl": false,
                        "driver__MacroSANFC__filter_function": "",
                        "driver__MacroSANFC__goodness_function": "",
                        "driver__MacroSANFC__iscsi_iotype": "fileio",
                        "driver__MacroSANFC__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__MacroSANFC__iscsi_target_flags": "",
                        "driver__MacroSANFC__iscsi_write_cache": "on",
                        "driver__MacroSANFC__num_shell_tries": 3,
                        "driver__MacroSANFC__num_volume_device_scan_tries": 3,
                        "driver__MacroSANFC__report_discard_supported": false,
                        "driver__MacroSANFC__storage_protocol": "iscsi",
                        "driver__MacroSANFC__target_helper": "tgtadm",
                        "driver__MacroSANFC__target_ip_address": "$my_ip",
                        "driver__MacroSANFC__target_port": 3260,
                        "driver__MacroSANFC__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__MacroSANFC__target_protocol": "iscsi",
                        "driver__MacroSANFC__trace_flags__transform_csv": "",
                        "driver__MacroSANFC__use_chap_auth": false,
                        "driver__MacroSANFC__volume_backend_name": "",
                        "driver__MacroSANFC__volume_clear": "zero",
                        "driver__MacroSANFC__volume_clear_ionice": "",
                        "driver__MacroSANFC__volume_clear_size": 0,
                        "driver__MacroSANFC__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__MacroSANFC__volume_copy_bps_limit": 0,
                        "driver__MacroSANFC__volume_dd_blocksize": "1M",
                        "driver__MacroSANFC__volumes_dir": "$state_path/volumes",
                        "driver__MacroSANISCSI__backend_availability_zone": "",
                        "driver__MacroSANISCSI__chap_password": "",
                        "driver__MacroSANISCSI__chap_username": "",
                        "driver__MacroSANISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__MacroSANISCSI__driver_client_cert": "",
                        "driver__MacroSANISCSI__driver_client_cert_key": "",
                        "driver__MacroSANISCSI__driver_data_namespace": "",
                        "driver__MacroSANISCSI__driver_ssl_cert_path": "",
                        "driver__MacroSANISCSI__driver_ssl_cert_verify": false,
                        "driver__MacroSANISCSI__driver_use_ssl": false,
                        "driver__MacroSANISCSI__filter_function": "",
                        "driver__MacroSANISCSI__goodness_function": "",
                        "driver__MacroSANISCSI__iscsi_iotype": "fileio",
                        "driver__MacroSANISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__MacroSANISCSI__iscsi_target_flags": "",
                        "driver__MacroSANISCSI__iscsi_write_cache": "on",
                        "driver__MacroSANISCSI__num_shell_tries": 3,
                        "driver__MacroSANISCSI__num_volume_device_scan_tries": 3,
                        "driver__MacroSANISCSI__report_discard_supported": false,
                        "driver__MacroSANISCSI__storage_protocol": "iscsi",
                        "driver__MacroSANISCSI__target_helper": "tgtadm",
                        "driver__MacroSANISCSI__target_ip_address": "$my_ip",
                        "driver__MacroSANISCSI__target_port": 3260,
                        "driver__MacroSANISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__MacroSANISCSI__target_protocol": "iscsi",
                        "driver__MacroSANISCSI__trace_flags__transform_csv": "",
                        "driver__MacroSANISCSI__use_chap_auth": false,
                        "driver__MacroSANISCSI__volume_backend_name": "",
                        "driver__MacroSANISCSI__volume_clear": "zero",
                        "driver__MacroSANISCSI__volume_clear_ionice": "",
                        "driver__MacroSANISCSI__volume_clear_size": 0,
                        "driver__MacroSANISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__MacroSANISCSI__volume_copy_bps_limit": 0,
                        "driver__MacroSANISCSI__volume_dd_blocksize": "1M",
                        "driver__MacroSANISCSI__volumes_dir": "$state_path/volumes",
                        "driver__NetAppCmodeFibreChannel__netapp_vserver": "",
                        "driver__NetAppCmodeISCSI__netapp_vserver": "",
                        "driver__NexentaISCSI__nexenta_blocksize": 4096,
                        "driver__NexentaISCSI__nexenta_dataset_compression": "on",
                        "driver__NexentaISCSI__nexenta_dataset_dedup": "off",
                        "driver__NexentaISCSI__nexenta_dataset_description": "",
                        "driver__NexentaISCSI__nexenta_folder": "",
                        "driver__NexentaISCSI__nexenta_group_snapshot_template": "group-snapshot-%s",
                        "driver__NexentaISCSI__nexenta_host": "",
                        "driver__NexentaISCSI__nexenta_host_group_prefix": "cinder",
                        "driver__NexentaISCSI__nexenta_iscsi_target_host_group": "all",
                        "driver__NexentaISCSI__nexenta_iscsi_target_portal_groups": "",
                        "driver__NexentaISCSI__nexenta_iscsi_target_portal_port": 3260,
                        "driver__NexentaISCSI__nexenta_iscsi_target_portals": "",
                        "driver__NexentaISCSI__nexenta_lu_writebackcache_disabled": false,
                        "driver__NexentaISCSI__nexenta_luns_per_target": 100,
                        "driver__NexentaISCSI__nexenta_ns5_blocksize": 32,
                        "driver__NexentaISCSI__nexenta_origin_snapshot_template": "origin-snapshot-%s",
                        "driver__NexentaISCSI__nexenta_rest_backoff_factor__transform_string_float": "0.5",
                        "driver__NexentaISCSI__nexenta_rest_connect_timeout__transform_string_float": "30",
                        "driver__NexentaISCSI__nexenta_rest_protocol": "auto",
                        "driver__NexentaISCSI__nexenta_rest_read_timeout__transform_string_float": "300",
                        "driver__NexentaISCSI__nexenta_rest_retry_count": 3,
                        "driver__NexentaISCSI__nexenta_sparse": false,
                        "driver__NexentaISCSI__nexenta_target_group_prefix": "cinder",
                        "driver__NexentaISCSI__nexenta_target_prefix": "iqn.1986-03.com.sun:02:cinder",
                        "driver__NexentaISCSI__nexenta_use_https": true,
                        "driver__NexentaISCSI__nexenta_volume": "cinder",
                        "driver__NexentaISCSI__nexenta_volume_group": "iscsi",
                        "driver__NimbleFC__nimble_pool_name": "default",
                        "driver__NimbleFC__nimble_subnet_label": "*",
                        "driver__NimbleFC__nimble_verify_cert_path": "",
                        "driver__NimbleFC__nimble_verify_certificate": false,
                        "driver__NimbleISCSI__nimble_pool_name": "default",
                        "driver__NimbleISCSI__nimble_subnet_label": "*",
                        "driver__NimbleISCSI__nimble_verify_cert_path": "",
                        "driver__NimbleISCSI__nimble_verify_certificate": false,
                        "driver__PVMEFC__driver_ssl_cert_path": "",
                        "driver__PVMEFC__driver_ssl_cert_verify": false,
                        "driver__PVMEFC__driver_use_ssl": false,
                        "driver__PVMEFC__pvme_pool_name": "A",
                        "driver__PVMEFC__san_ip": "",
                        "driver__PVMEFC__san_login": "admin",
                        "driver__PVMEFC__san_password": "",
                        "driver__PVMEISCSI__driver_ssl_cert_path": "",
                        "driver__PVMEISCSI__driver_ssl_cert_verify": false,
                        "driver__PVMEISCSI__driver_use_ssl": false,
                        "driver__PVMEISCSI__pvme_iscsi_ips__transform_csv": "",
                        "driver__PVMEISCSI__pvme_pool_name": "A",
                        "driver__PVMEISCSI__san_ip": "",
                        "driver__PVMEISCSI__san_login": "admin",
                        "driver__PVMEISCSI__san_password": "",
                        "driver__PowerFlex__powerflex_allow_migration_during_rebuild": false,
                        "driver__PowerFlex__powerflex_allow_non_padded_volumes": false,
                        "driver__PowerFlex__powerflex_max_over_subscription_ratio__transform_string_float": "10.0",
                        "driver__PowerFlex__powerflex_rest_server_port": 443,
                        "driver__PowerFlex__powerflex_round_volume_capacity": true,
                        "driver__PowerFlex__powerflex_server_api_version": "",
                        "driver__PowerFlex__powerflex_storage_pools": "",
                        "driver__PowerFlex__powerflex_unmap_volume_before_deletion": false,
                        "driver__PowerMaxFC__driver_ssl_cert_verify": false,
                        "driver__PowerMaxFC__initiator_check": false,
                        "driver__PowerMaxFC__interval": 3,
                        "driver__PowerMaxFC__load_balance": false,
                        "driver__PowerMaxFC__load_balance_real_time": false,
                        "driver__PowerMaxFC__load_data_format": "Avg",
                        "driver__PowerMaxFC__load_look_back": 60,
                        "driver__PowerMaxFC__load_look_back_real_time": 1,
                        "driver__PowerMaxFC__port_group_load_metric": "PercentBusy",
                        "driver__PowerMaxFC__port_load_metric": "PercentBusy",
                        "driver__PowerMaxFC__powermax_array": "",
                        "driver__PowerMaxFC__powermax_array_tag_list__transform_csv": "",
                        "driver__PowerMaxFC__powermax_port_group_name_template": "portGroupName",
                        "driver__PowerMaxFC__powermax_port_groups__transform_csv": "",
                        "driver__PowerMaxFC__powermax_service_level": "",
                        "driver__PowerMaxFC__powermax_short_host_name_template": "shortHostName",
                        "driver__PowerMaxFC__powermax_snapvx_unlink_limit": 3,
                        "driver__PowerMaxFC__powermax_srp": "",
                        "driver__PowerMaxFC__retries": 200,
                        "driver__PowerMaxFC__san_ip": "",
                        "driver__PowerMaxFC__san_login": "admin",
                        "driver__PowerMaxFC__san_password": "",
                        "driver__PowerMaxFC__u4p_failover_autofailback": true,
                        "driver__PowerMaxFC__u4p_failover_backoff_factor": 1,
                        "driver__PowerMaxFC__u4p_failover_retries": 3,
                        "driver__PowerMaxFC__u4p_failover_target__transform_csv_kvs": "",
                        "driver__PowerMaxFC__u4p_failover_timeout": 20,
                        "driver__PowerMaxFC__vmax_workload": "",
                        "driver__PowerMaxISCSI__chap_password": "",
                        "driver__PowerMaxISCSI__chap_username": "",
                        "driver__PowerMaxISCSI__driver_ssl_cert_verify": false,
                        "driver__PowerMaxISCSI__initiator_check": false,
                        "driver__PowerMaxISCSI__interval": 3,
                        "driver__PowerMaxISCSI__load_balance": false,
                        "driver__PowerMaxISCSI__load_balance_real_time": false,
                        "driver__PowerMaxISCSI__load_data_format": "Avg",
                        "driver__PowerMaxISCSI__load_look_back": 60,
                        "driver__PowerMaxISCSI__load_look_back_real_time": 1,
                        "driver__PowerMaxISCSI__port_group_load_metric": "PercentBusy",
                        "driver__PowerMaxISCSI__port_load_metric": "PercentBusy",
                        "driver__PowerMaxISCSI__powermax_array": "",
                        "driver__PowerMaxISCSI__powermax_array_tag_list__transform_csv": "",
                        "driver__PowerMaxISCSI__powermax_port_group_name_template": "portGroupName",
                        "driver__PowerMaxISCSI__powermax_port_groups__transform_csv": "",
                        "driver__PowerMaxISCSI__powermax_service_level": "",
                        "driver__PowerMaxISCSI__powermax_short_host_name_template": "shortHostName",
                        "driver__PowerMaxISCSI__powermax_snapvx_unlink_limit": 3,
                        "driver__PowerMaxISCSI__powermax_srp": "",
                        "driver__PowerMaxISCSI__retries": 200,
                        "driver__PowerMaxISCSI__san_ip": "",
                        "driver__PowerMaxISCSI__san_login": "admin",
                        "driver__PowerMaxISCSI__san_password": "",
                        "driver__PowerMaxISCSI__u4p_failover_autofailback": true,
                        "driver__PowerMaxISCSI__u4p_failover_backoff_factor": 1,
                        "driver__PowerMaxISCSI__u4p_failover_retries": 3,
                        "driver__PowerMaxISCSI__u4p_failover_target__transform_csv_kvs": "",
                        "driver__PowerMaxISCSI__u4p_failover_timeout": 20,
                        "driver__PowerMaxISCSI__use_chap_auth": false,
                        "driver__PowerMaxISCSI__vmax_workload": "",
                        "driver__PowerStore__powerstore_appliances__transform_csv": "",
                        "driver__PowerStore__powerstore_ports__transform_csv": "",
                        "driver__PureFC__driver_ssl_cert_path": "",
                        "driver__PureFC__driver_ssl_cert_verify": false,
                        "driver__PureFC__pure_api_token": "",
                        "driver__PureFC__pure_automatic_max_oversubscription_ratio": true,
                        "driver__PureFC__pure_eradicate_on_delete": false,
                        "driver__PureFC__pure_host_personality__transform_empty_none": "",
                        "driver__PureFC__pure_iscsi_cidr": "0.0.0.0/0",
                        "driver__PureFC__pure_replica_interval_default": 3600,
                        "driver__PureFC__pure_replica_retention_long_term_default": 7,
                        "driver__PureFC__pure_replica_retention_long_term_per_day_default": 3,
                        "driver__PureFC__pure_replica_retention_short_term_default": 14400,
                        "driver__PureFC__pure_replication_pg_name": "cinder-group",
                        "driver__PureFC__pure_replication_pod_name": "cinder-pod",
                        "driver__PureFC__san_ip": "",
                        "driver__PureFC__use_chap_auth": false,
                        "driver__PureISCSI__driver_ssl_cert_path": "",
                        "driver__PureISCSI__driver_ssl_cert_verify": false,
                        "driver__PureISCSI__pure_api_token": "",
                        "driver__PureISCSI__pure_automatic_max_oversubscription_ratio": true,
                        "driver__PureISCSI__pure_eradicate_on_delete": false,
                        "driver__PureISCSI__pure_host_personality__transform_empty_none": "",
                        "driver__PureISCSI__pure_iscsi_cidr": "0.0.0.0/0",
                        "driver__PureISCSI__pure_replica_interval_default": 3600,
                        "driver__PureISCSI__pure_replica_retention_long_term_default": 7,
                        "driver__PureISCSI__pure_replica_retention_long_term_per_day_default": 3,
                        "driver__PureISCSI__pure_replica_retention_short_term_default": 14400,
                        "driver__PureISCSI__pure_replication_pg_name": "cinder-group",
                        "driver__PureISCSI__pure_replication_pod_name": "cinder-pod",
                        "driver__PureISCSI__san_ip": "",
                        "driver__PureISCSI__use_chap_auth": false,
                        "driver__QnapISCSI__chap_password": "",
                        "driver__QnapISCSI__chap_username": "",
                        "driver__QnapISCSI__driver_ssl_cert_verify": false,
                        "driver__QnapISCSI__qnap_management_url": "",
                        "driver__QnapISCSI__qnap_poolname": "",
                        "driver__QnapISCSI__qnap_storage_protocol": "iscsi",
                        "driver__QnapISCSI__san_login": "admin",
                        "driver__QnapISCSI__san_password": "",
                        "driver__QnapISCSI__target_ip_address": "$my_ip",
                        "driver__QnapISCSI__use_chap_auth": false,
                        "driver__Quobyte__quobyte_client_cfg": "",
                        "driver__Quobyte__quobyte_mount_point_base": "$state_path/mnt",
                        "driver__Quobyte__quobyte_overlay_volumes": false,
                        "driver__Quobyte__quobyte_qcow2_volumes": true,
                        "driver__Quobyte__quobyte_sparsed_volumes": true,
                        "driver__Quobyte__quobyte_volume_from_snapshot_cache": false,
                        "driver__Quobyte__quobyte_volume_url": "",
                        "driver__RBD__deferred_deletion_delay": 0,
                        "driver__RBD__deferred_deletion_purge_interval": 60,
                        "driver__RBD__enable_deferred_deletion": false,
                        "driver__RBD__rados_connect_timeout": -1,
                        "driver__RBD__rados_connection_interval": 5,
                        "driver__RBD__rados_connection_retries": 3,
                        "driver__RBD__rbd_ceph_conf": "",
                        "driver__RBD__rbd_cluster_name": "ceph",
                        "driver__RBD__rbd_exclusive_cinder_pool": false,
                        "driver__RBD__rbd_flatten_volume_from_snapshot": false,
                        "driver__RBD__rbd_keyring_conf": "",
                        "driver__RBD__rbd_max_clone_depth": 5,
                        "driver__RBD__rbd_pool": "rbd",
                        "driver__RBD__rbd_secret_uuid": "",
                        "driver__RBD__rbd_store_chunk_size": 4,
                        "driver__RBD__rbd_user": "",
                        "driver__RBD__report_dynamic_total_capacity": true,
                        "driver__RBD__volume_dd_blocksize": "1M",
                        "driver__RBDISCSI__deferred_deletion_delay": 0,
                        "driver__RBDISCSI__deferred_deletion_purge_interval": 60,
                        "driver__RBDISCSI__driver_ssl_cert_verify": false,
                        "driver__RBDISCSI__enable_deferred_deletion": false,
                        "driver__RBDISCSI__rados_connect_timeout": -1,
                        "driver__RBDISCSI__rados_connection_interval": 5,
                        "driver__RBDISCSI__rados_connection_retries": 3,
                        "driver__RBDISCSI__rbd_ceph_conf": "",
                        "driver__RBDISCSI__rbd_cluster_name": "ceph",
                        "driver__RBDISCSI__rbd_exclusive_cinder_pool": false,
                        "driver__RBDISCSI__rbd_flatten_volume_from_snapshot": false,
                        "driver__RBDISCSI__rbd_iscsi_api_debug": false,
                        "driver__RBDISCSI__rbd_iscsi_api_password": "",
                        "driver__RBDISCSI__rbd_iscsi_api_url": "",
                        "driver__RBDISCSI__rbd_iscsi_api_user": "",
                        "driver__RBDISCSI__rbd_iscsi_target_iqn": "",
                        "driver__RBDISCSI__rbd_max_clone_depth": 5,
                        "driver__RBDISCSI__rbd_pool": "rbd",
                        "driver__RBDISCSI__rbd_secret_uuid": "",
                        "driver__RBDISCSI__rbd_store_chunk_size": 4,
                        "driver__RBDISCSI__rbd_user": "",
                        "driver__RBDISCSI__report_dynamic_total_capacity": true,
                        "driver__RBDISCSI__suppress_requests_ssl_warnings": false,
                        "driver__RBDISCSI__volume_dd_blocksize": "1M",
                        "driver__RSD__podm_password": "",
                        "driver__RSD__podm_url": "",
                        "driver__RSD__podm_username": "",
                        "driver__SCFC__dell_api_async_rest_timeout": 15,
                        "driver__SCFC__dell_api_sync_rest_timeout": 30,
                        "driver__SCFC__dell_sc_api_port": 3033,
                        "driver__SCFC__dell_sc_server_folder": "openstack",
                        "driver__SCFC__dell_sc_ssn": 64702,
                        "driver__SCFC__dell_sc_verify_cert": false,
                        "driver__SCFC__dell_sc_volume_folder": "openstack",
                        "driver__SCFC__dell_server_os": "Red Hat Linux 6.x",
                        "driver__SCFC__excluded_domain_ips__transform_csv": "",
                        "driver__SCFC__included_domain_ips__transform_csv": "",
                        "driver__SCFC__secondary_san_ip": "",
                        "driver__SCFC__secondary_san_login": "Admin",
                        "driver__SCFC__secondary_san_password": "",
                        "driver__SCFC__secondary_sc_api_port": 3033,
                        "driver__SCISCSI__dell_api_async_rest_timeout": 15,
                        "driver__SCISCSI__dell_api_sync_rest_timeout": 30,
                        "driver__SCISCSI__dell_sc_api_port": 3033,
                        "driver__SCISCSI__dell_sc_server_folder": "openstack",
                        "driver__SCISCSI__dell_sc_ssn": 64702,
                        "driver__SCISCSI__dell_sc_verify_cert": false,
                        "driver__SCISCSI__dell_sc_volume_folder": "openstack",
                        "driver__SCISCSI__dell_server_os": "Red Hat Linux 6.x",
                        "driver__SCISCSI__excluded_domain_ips__transform_csv": "",
                        "driver__SCISCSI__included_domain_ips__transform_csv": "",
                        "driver__SCISCSI__secondary_san_ip": "",
                        "driver__SCISCSI__secondary_san_login": "Admin",
                        "driver__SCISCSI__secondary_san_password": "",
                        "driver__SCISCSI__secondary_sc_api_port": 3033,
                        "driver__SdsISCSI__backend_availability_zone": "",
                        "driver__SdsISCSI__chap_password": "",
                        "driver__SdsISCSI__chap_username": "",
                        "driver__SdsISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__SdsISCSI__driver_client_cert": "",
                        "driver__SdsISCSI__driver_client_cert_key": "",
                        "driver__SdsISCSI__driver_data_namespace": "",
                        "driver__SdsISCSI__driver_ssl_cert_path": "",
                        "driver__SdsISCSI__driver_ssl_cert_verify": false,
                        "driver__SdsISCSI__driver_use_ssl": false,
                        "driver__SdsISCSI__filter_function": "",
                        "driver__SdsISCSI__goodness_function": "",
                        "driver__SdsISCSI__iscsi_iotype": "fileio",
                        "driver__SdsISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__SdsISCSI__iscsi_target_flags": "",
                        "driver__SdsISCSI__iscsi_write_cache": "on",
                        "driver__SdsISCSI__num_shell_tries": 3,
                        "driver__SdsISCSI__num_volume_device_scan_tries": 3,
                        "driver__SdsISCSI__report_discard_supported": false,
                        "driver__SdsISCSI__storage_protocol": "iscsi",
                        "driver__SdsISCSI__target_helper": "tgtadm",
                        "driver__SdsISCSI__target_ip_address": "$my_ip",
                        "driver__SdsISCSI__target_port": 3260,
                        "driver__SdsISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__SdsISCSI__target_protocol": "iscsi",
                        "driver__SdsISCSI__trace_flags__transform_csv": "",
                        "driver__SdsISCSI__use_chap_auth": false,
                        "driver__SdsISCSI__volume_backend_name": "",
                        "driver__SdsISCSI__volume_clear": "zero",
                        "driver__SdsISCSI__volume_clear_ionice": "",
                        "driver__SdsISCSI__volume_clear_size": 0,
                        "driver__SdsISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__SdsISCSI__volume_copy_bps_limit": 0,
                        "driver__SdsISCSI__volume_dd_blocksize": "1M",
                        "driver__SdsISCSI__volumes_dir": "$state_path/volumes",
                        "driver__SolidFire__driver_ssl_cert_verify": false,
                        "driver__SolidFire__san_ip": "",
                        "driver__SolidFire__san_login": "admin",
                        "driver__SolidFire__san_password": "",
                        "driver__SolidFire__sf_account_prefix": "",
                        "driver__SolidFire__sf_allow_tenant_qos": false,
                        "driver__SolidFire__sf_api_port": 443,
                        "driver__SolidFire__sf_api_request_timeout": 30,
                        "driver__SolidFire__sf_cluster_pairing_timeout": 60,
                        "driver__SolidFire__sf_emulate_512": true,
                        "driver__SolidFire__sf_enable_vag": false,
                        "driver__SolidFire__sf_provisioning_calc": "maxProvisionedSpace",
                        "driver__SolidFire__sf_svip": "",
                        "driver__SolidFire__sf_volume_clone_timeout": 600,
                        "driver__SolidFire__sf_volume_create_timeout": 60,
                        "driver__SolidFire__sf_volume_pairing_timeout": 3600,
                        "driver__SolidFire__sf_volume_prefix": "UUID-",
                        "driver__StorPool__storpool_replication": 3,
                        "driver__StorPool__storpool_template": "",
                        "driver__StorwizeSVCFC__cycle_period_seconds": 300,
                        "driver__StorwizeSVCFC__storwize_peer_pool": "",
                        "driver__StorwizeSVCFC__storwize_preferred_host_site__transform_csv_kvs": "",
                        "driver__StorwizeSVCFC__storwize_san_secondary_ip": "",
                        "driver__StorwizeSVCFC__storwize_svc_allow_tenant_qos": false,
                        "driver__StorwizeSVCFC__storwize_svc_flashcopy_rate": 50,
                        "driver__StorwizeSVCFC__storwize_svc_flashcopy_timeout": 120,
                        "driver__StorwizeSVCFC__storwize_svc_mirror_pool": "",
                        "driver__StorwizeSVCFC__storwize_svc_multipath_enabled": false,
                        "driver__StorwizeSVCFC__storwize_svc_retain_aux_volume": false,
                        "driver__StorwizeSVCFC__storwize_svc_stretched_cluster_partner": "",
                        "driver__StorwizeSVCFC__storwize_svc_vol_autoexpand": true,
                        "driver__StorwizeSVCFC__storwize_svc_vol_compression": false,
                        "driver__StorwizeSVCFC__storwize_svc_vol_easytier": true,
                        "driver__StorwizeSVCFC__storwize_svc_vol_grainsize": 256,
                        "driver__StorwizeSVCFC__storwize_svc_vol_iogrp": "0",
                        "driver__StorwizeSVCFC__storwize_svc_vol_nofmtdisk": false,
                        "driver__StorwizeSVCFC__storwize_svc_vol_rsize": 2,
                        "driver__StorwizeSVCFC__storwize_svc_vol_warning": 0,
                        "driver__StorwizeSVCFC__storwize_svc_volpool_name__transform_csv": "volpool",
                        "driver__StorwizeSVCISCSI__cycle_period_seconds": 300,
                        "driver__StorwizeSVCISCSI__storwize_peer_pool": "",
                        "driver__StorwizeSVCISCSI__storwize_preferred_host_site__transform_csv_kvs": "",
                        "driver__StorwizeSVCISCSI__storwize_san_secondary_ip": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_allow_tenant_qos": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_flashcopy_rate": 50,
                        "driver__StorwizeSVCISCSI__storwize_svc_flashcopy_timeout": 120,
                        "driver__StorwizeSVCISCSI__storwize_svc_iscsi_chap_enabled": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_mirror_pool": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_retain_aux_volume": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_stretched_cluster_partner": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_autoexpand": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_compression": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_easytier": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_grainsize": 256,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_iogrp": "0",
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_nofmtdisk": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_rsize": 2,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_warning": 0,
                        "driver__StorwizeSVCISCSI__storwize_svc_volpool_name__transform_csv": "volpool",
                        "driver__SynoISCSI__chap_password": "",
                        "driver__SynoISCSI__chap_username": "",
                        "driver__SynoISCSI__driver_use_ssl": false,
                        "driver__SynoISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__SynoISCSI__synology_admin_port": 5000,
                        "driver__SynoISCSI__synology_device_id": "",
                        "driver__SynoISCSI__synology_one_time_pass": "",
                        "driver__SynoISCSI__synology_password": "",
                        "driver__SynoISCSI__synology_pool_name": "",
                        "driver__SynoISCSI__synology_ssl_verify": true,
                        "driver__SynoISCSI__synology_username": "admin",
                        "driver__SynoISCSI__target_ip_address": "$my_ip",
                        "driver__SynoISCSI__target_port": 3260,
                        "driver__SynoISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__SynoISCSI__target_protocol": "iscsi",
                        "driver__SynoISCSI__use_chap_auth": false,
                        "driver__Unity__remove_empty_host": false,
                        "driver__Unity__unity_io_ports__transform_csv": "",
                        "driver__Unity__unity_storage_pool_names__transform_csv": "",
                        "driver__VNX__default_timeout": 31536000,
                        "driver__VNX__destroy_empty_storage_group": false,
                        "driver__VNX__force_delete_lun_in_storagegroup": true,
                        "driver__VNX__ignore_pool_full_threshold": false,
                        "driver__VNX__initiator_auto_deregistration": false,
                        "driver__VNX__initiator_auto_registration": false,
                        "driver__VNX__io_port_list__transform_csv": "",
                        "driver__VNX__iscsi_initiators": "",
                        "driver__VNX__max_luns_per_storage_group": 255,
                        "driver__VNX__naviseccli_path": "",
                        "driver__VNX__storage_vnx_authentication_type": "global",
                        "driver__VNX__storage_vnx_pool_names__transform_csv": "",
                        "driver__VNX__storage_vnx_security_file_dir": "",
                        "driver__VNX__vnx_async_migrate": true,
                        "driver__VZStorage__vzstorage_default_volume_format": "raw",
                        "driver__VZStorage__vzstorage_mount_options__transform_csv": "",
                        "driver__VZStorage__vzstorage_mount_point_base": "$state_path/mnt",
                        "driver__VZStorage__vzstorage_shares_config": "/etc/cinder/vzstorage_shares",
                        "driver__VZStorage__vzstorage_sparsed_volumes": true,
                        "driver__VZStorage__vzstorage_used_ratio__transform_string_float": "0.95",
                        "driver__WindowsISCSI__windows_iscsi_lun_path": "C:\\iSCSIVirtualDisks",
                        "driver__WindowsSmbfs__smbfs_default_volume_format": "vhd",
                        "driver__WindowsSmbfs__smbfs_mount_point_base": "C:\\OpenStack\\_mnt",
                        "driver__WindowsSmbfs__smbfs_pool_mappings__transform_csv_kvs": "",
                        "driver__WindowsSmbfs__smbfs_shares_config": "C:\\OpenStack\\smbfs_shares.txt",
                        "driver__XtremIOFC__driver_ssl_cert_path": "",
                        "driver__XtremIOFC__driver_ssl_cert_verify": false,
                        "driver__XtremIOFC__san_ip": "",
                        "driver__XtremIOFC__san_login": "admin",
                        "driver__XtremIOFC__san_password": "",
                        "driver__XtremIOFC__xtremio_array_busy_retry_count": 5,
                        "driver__XtremIOFC__xtremio_array_busy_retry_interval": 5,
                        "driver__XtremIOFC__xtremio_clean_unused_ig": false,
                        "driver__XtremIOFC__xtremio_cluster_name": "",
                        "driver__XtremIOISCSI__driver_ssl_cert_path": "",
                        "driver__XtremIOISCSI__driver_ssl_cert_verify": false,
                        "driver__XtremIOISCSI__san_ip": "",
                        "driver__XtremIOISCSI__san_login": "admin",
                        "driver__XtremIOISCSI__san_password": "",
                        "driver__XtremIOISCSI__xtremio_array_busy_retry_count": 5,
                        "driver__XtremIOISCSI__xtremio_array_busy_retry_interval": 5,
                        "driver__XtremIOISCSI__xtremio_clean_unused_ig": false,
                        "driver__XtremIOISCSI__xtremio_cluster_name": "",
                        "driver__ZadaraVPSAISCSI__zadara_access_key": "",
                        "driver__ZadaraVPSAISCSI__zadara_default_snap_policy": false,
                        "driver__ZadaraVPSAISCSI__zadara_ssl_cert_verify": true,
                        "driver__ZadaraVPSAISCSI__zadara_use_iser": true,
                        "driver__ZadaraVPSAISCSI__zadara_vol_encrypt": false,
                        "driver__ZadaraVPSAISCSI__zadara_vol_name_template": "OS_%s",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_host": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_poolname": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_port": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_use_ssl": false
                      }
                    },
                    "sysFiles": {
                      "name": ""
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/embercsi/ember-csi-operator:latest
          createdAt: 2021-02-22:17:47:00
          description: Multi-vendor CSI plugin supporting 80+ storage drivers
          repository: https://github.com/embercsi/ember-csi-operator
          support: http://readthedocs.org/projects/ember-csi/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an Ember CSI storage backend
            displayName: Storage Backends
            kind: EmberStorageBackend
            name: embercsis.ember-csi.io
            version: v1alpha1
        description: |
          Ember-CSI is a multi-vendor CSI plugin driver supporting
          over 80 storage drivers in a single plugin to provide block and mount
          storage to container orchestration systems.

          The operator adds support for management of multiple Ember CSI instances
          within a cluster and simplifies storage backend deployments.

          ### Supported drivers
          Ember-CSI includes a good number of storage drivers, but due to limitation on hardware availability only a small number of them have been validated at one point or another. In alphabetical order they are:

          * HPE3PARFC
          * HPE3PARISCSI
          * KaminarioISCSI
          * LVMVolume
          * PowerMaxFC
          * PowerMaxISCSI
          * PureFC
          * PureISCSI
          * QnapISCSI
          * RBD
          * SolidFire
          * SynoISCSI
          * XtremIOFC
          * XtremIOISCSI

          ### Getting started
          Please have a look at our quick start documentation on how to get started:
          [github.com/embercsi/ember-csi-operator#quick-start](https://github.com/embercsi/ember-csi-operator#quick-start).
          We recommend to use the user friendly "Edit Form" when creating a new
          EmberStorageBackend. You'll find the link in the upper right corner of the
          editor.

          ### Documentation
          Documentation on Ember CSI itself can be found at [docs.ember-csi.io](http://docs.ember-csi.io/).

          ### Getting help
          Please contact us if you encounter any issues while using Ember CSI.
          Get in touch through the #ember-csi channel on Freenode, also accessible through
          [Kiwi IRC](https://kiwiirc.com/nextclient/irc.freenode.net/ember-csi).

          You can also create an issue on the [Github project](https://github.com/embercsi/ember-csi-operator)
          for bugs, enhancements, or other requests.
        displayName: Ember CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Ember-CSI
        - CSI
        links:
        - name: Learn more about the project
          url: http://ember-csi.io/
        - name: Documentation
          url: http://readthedocs.org/projects/ember-csi/
        - name: Ember-CSI Source Code
          url: https://github.com/embercsi/ember-csi
        - name: Ember-CSI Operator Source Code
          url: https://github.com/embercsi/ember-csi-operator
        maintainers:
        - email: geguileo@redhat.com
          name: Gorka Eguileor
        - email: cschwede@redhat.com
          name: Christian Schwede
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/embercsi/ember-csi-operator:latest
        version: 0.9.9
      entries:
      - name: ember-csi-community-operator.v0.9.9
        version: 0.9.9
      - name: ember-csi-community-operator.v0.9.8
        version: 0.9.8
      - name: ember-csi-community-operator.v0.9.7
        version: 0.9.7
      name: beta
    defaultChannel: beta
    packageName: ember-csi-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-dynamodb-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-dynamodb-controller.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "GlobalTable",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "Table",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/dynamodb-controller:1.5.2
          createdAt: "2025-09-19T19:19:47Z"
          description: AWS DynamoDB controller is a service controller for managing
            DynamoDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup represents the state of an AWS dynamodb Backup resource.
            displayName: Backup
            kind: Backup
            name: backups.dynamodb.services.k8s.aws
            version: v1alpha1
          - description: GlobalTable represents the state of an AWS dynamodb GlobalTable
              resource.
            displayName: GlobalTable
            kind: GlobalTable
            name: globaltables.dynamodb.services.k8s.aws
            version: v1alpha1
          - description: Table represents the state of an AWS dynamodb Table resource.
            displayName: Table
            kind: Table
            name: tables.dynamodb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon DynamoDB resources in AWS from within your Kubernetes cluster.

          **About Amazon DynamoDB**

          Amazon DynamoDB is a fully managed NoSQL database service that provides fast and predictable performance with seamless scalability. DynamoDB lets you offload the administrative burdens of operating and scaling a distributed database so that you don't have to worry about hardware provisioning, setup and configuration, replication, software patching, or cluster scaling. DynamoDB also offers encryption at rest, which eliminates the operational burden and complexity involved in protecting sensitive data. For more information, see [DynamoDB Encryption at Rest](https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/EncryptionAtRest.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon DynamoDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dynamodb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon DynamoDB Developer Resources
          url: https://aws.amazon.com/dynamodb/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: dynamodb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/dynamodb-controller:1.5.2
        version: 1.5.2
      entries:
      - name: ack-dynamodb-controller.v1.5.2
        version: 1.5.2
      - name: ack-dynamodb-controller.v1.5.1
        version: 1.5.1
      - name: ack-dynamodb-controller.v1.5.0
        version: 1.5.0
      - name: ack-dynamodb-controller.v1.4.10
        version: 1.4.10
      - name: ack-dynamodb-controller.v1.4.9
        version: 1.4.9
      - name: ack-dynamodb-controller.v1.4.7
        version: 1.4.7
      - name: ack-dynamodb-controller.v1.4.6
        version: 1.4.6
      - name: ack-dynamodb-controller.v1.4.5
        version: 1.4.5
      - name: ack-dynamodb-controller.v1.4.4
        version: 1.4.4
      - name: ack-dynamodb-controller.v1.4.3
        version: 1.4.3
      - name: ack-dynamodb-controller.v1.4.2
        version: 1.4.2
      - name: ack-dynamodb-controller.v1.4.1
        version: 1.4.1
      - name: ack-dynamodb-controller.v1.3.0
        version: 1.3.0
      - name: ack-dynamodb-controller.v1.2.23
        version: 1.2.23
      - name: ack-dynamodb-controller.v1.2.22
        version: 1.2.22
      - name: ack-dynamodb-controller.v1.2.21
        version: 1.2.21
      - name: ack-dynamodb-controller.v1.2.20
        version: 1.2.20
      - name: ack-dynamodb-controller.v1.2.15
        version: 1.2.15
      - name: ack-dynamodb-controller.v1.2.14
        version: 1.2.14
      - name: ack-dynamodb-controller.v1.2.13
        version: 1.2.13
      - name: ack-dynamodb-controller.v1.2.12
        version: 1.2.12
      - name: ack-dynamodb-controller.v1.2.11
        version: 1.2.11
      - name: ack-dynamodb-controller.v1.2.9
        version: 1.2.9
      - name: ack-dynamodb-controller.v1.2.8
        version: 1.2.8
      - name: ack-dynamodb-controller.v1.2.6
        version: 1.2.6
      - name: ack-dynamodb-controller.v1.2.5
        version: 1.2.5
      - name: ack-dynamodb-controller.v1.2.4
        version: 1.2.4
      - name: ack-dynamodb-controller.v1.2.2
        version: 1.2.2
      - name: ack-dynamodb-controller.v1.2.1
        version: 1.2.1
      - name: ack-dynamodb-controller.v1.2.0
        version: 1.2.0
      - name: ack-dynamodb-controller.v1.1.1
        version: 1.1.1
      - name: ack-dynamodb-controller.v1.1.0
        version: 1.1.0
      - name: ack-dynamodb-controller.v1.0.3
        version: 1.0.3
      - name: ack-dynamodb-controller.v1.0.2
        version: 1.0.2
      - name: ack-dynamodb-controller.v1.0.1
        version: 1.0.1
      - name: ack-dynamodb-controller.v1.0.0
        version: 1.0.0
      - name: ack-dynamodb-controller.v0.1.10
        version: 0.1.10
      - name: ack-dynamodb-controller.v0.1.9
        version: 0.1.9
      - name: ack-dynamodb-controller.v0.1.8
        version: 0.1.8
      - name: ack-dynamodb-controller.v0.1.7
        version: 0.1.7
      - name: ack-dynamodb-controller.v0.1.6
        version: 0.1.6
      - name: ack-dynamodb-controller.v0.1.5
        version: 0.1.5
      - name: ack-dynamodb-controller.v0.1.4
        version: 0.1.4
      - name: ack-dynamodb-controller.v0.1.3
        version: 0.1.3
      - name: ack-dynamodb-controller.v0.1.1
        version: 0.1.1
      - name: ack-dynamodb-controller.v0.1.0
        version: 0.1.0
      - name: ack-dynamodb-controller.v0.0.20
        version: 0.0.20
      - name: ack-dynamodb-controller.v0.0.19
        version: 0.0.19
      - name: ack-dynamodb-controller.v0.0.18
        version: 0.0.18
      - name: ack-dynamodb-controller.v0.0.17
        version: 0.0.17
      - name: ack-dynamodb-controller.v0.0.16
        version: 0.0.16
      - name: ack-dynamodb-controller.v0.0.15
        version: 0.0.15
      - name: ack-dynamodb-controller.v0.0.14
        version: 0.0.14
      - name: ack-dynamodb-controller.v0.0.13
        version: 0.0.13
      - name: ack-dynamodb-controller.v0.0.12
        version: 0.0.12
      - name: ack-dynamodb-controller.v0.0.11
        version: 0.0.11
      - name: ack-dynamodb-controller.v0.0.10
        version: 0.0.10
      - name: ack-dynamodb-controller.v0.0.9
        version: 0.0.9
      - name: ack-dynamodb-controller.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: ack-dynamodb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/limitador-operator
    name: limitador-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: limitador-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/limitador-operator:v0.5.0
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: The Limitador operator installs and maintains limitador instances
        displayName: Limitador
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        links:
        - name: Limitador Operator
          url: https://github.com/Kuadrant/limitador-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - quay.io/kuadrant/limitador-operator:v0.5.0
        - quay.io/kuadrant/limitador:v1.2.0
        version: 0.5.0
      entries:
      - name: limitador-operator.v0.5.0
        version: 0.5.0
      - name: limitador-operator.v0.4.0
        version: 0.4.0
      - name: limitador-operator.v0.3.0
        version: 0.3.0
      name: alpha
    - currentCSV: limitador-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "name": "toy_get_route",
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/limitador-operator:v0.11.0
          createdAt: "2024-10-02T10:46:31Z"
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: The Limitador operator installs and maintains limitador instances
        displayName: Limitador
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        links:
        - name: Limitador Operator
          url: https://github.com/Kuadrant/limitador-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - quay.io/kuadrant/limitador-operator:v0.11.0
        - quay.io/kuadrant/limitador:v1.6.0
        version: 0.11.0
      entries:
      - name: limitador-operator.v0.11.0
        version: 0.11.0
      - name: limitador-operator.v0.10.0
        version: 0.10.0
      - name: limitador-operator.v0.6.0
        version: 0.6.0
      - name: limitador-operator.v0.5.0
        version: 0.5.0
      - name: limitador-operator.v0.4.0
        version: 0.4.0
      - name: limitador-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: limitador-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/limitador-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Commvault
      provider-url: ""
    name: hedvig-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hedvig-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\":\"hedvig.io/v1\",\n    \"kind\":\"HedvigDeploy\",\n
            \   \"metadata\": {\n      \"name\": \"testdeploy\",\n      \"namespace\":
            \"hedvig\"\n    },\n    \"spec\": {\n      \"cluster\": {\n        \"name\":
            \"hedvigcl\",\n        \"seeds\": [\n          \"hedvigcl1.hedvig.io\",\n
            \         \"hedvigcl2.hedvig.io\",\n          \"hedvigcl3.hedvig.io\"\n
            \       ]\n      },\n      \"k8s\": {\n        \"name\": \"k8scl\"\n      },\n
            \     \"proxy\": {\n        \"repository\": \"hedvig/hedvig-proxy\",\n
            \       \"tag\": \"v-4.3.1\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"controller\": {\n        \"repository\": \"hedvig/hedvig-csi-volume-driver\",\n
            \       \"tag\": \"v-2.0\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"node\": {\n        \"repository\": \"hedvig/hedvig-csi-volume-driver\",\n
            \       \"tag\": \"v-2.0\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"sidecars\": [\n        {\n          \"name\": \"csi-provisioner\",\n
            \         \"repository\": \"quay.io/k8scsi/csi-provisioner\",\n          \"tag\":
            \"v1.6.0\",\n          \"imagePullPolicy\": \"IfNotPresent\"\n        },\n
            \       {\n          \"name\": \"csi-snapshotter\",\n          \"repository\":
            \"quay.io/k8scsi/csi-snapshotter\",\n          \"tag\": \"v2.1.0\",\n
            \         \"imagePullPolicy\": \"IfNotPresent\"\n        },              \n
            \       {\n          \"name\": \"csi-resizer\",\n          \"repository\":
            \"quay.io/k8scsi/csi-resizer\",\n          \"tag\": \"v0.5.0\",\n          \"imagePullPolicy\":
            \"IfNotPresent\"\n        },              \n        {\n          \"name\":
            \"csi-node-driver-registrar\",\n          \"repository\": \"quay.io/k8scsi/csi-node-driver-registrar\",\n
            \         \"tag\": \"v1.0-canary\",\n          \"imagePullPolicy\": \"IfNotPresent\"\n
            \       }\n      ]\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: hedvig/hedvig-operator:2.0.0-ubi
          createdAt: ""
          description: Hedvig Operator is the official operator to deploy and manage
            Hedvig components in container orchestrators such as Kubernetes and Openshift.
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hedvig Deploy
            displayName: HedvigDeploy
            kind: HedvigDeploy
            name: hedvigdeploys.hedvig.io
            version: v1
        description: "The Hedvig Distributed Storage Platform is a software-defined,
          scale-out storage solution. It distributes data simultaneously across multiple
          locations – from on-premises data centers to the cloud – and scales capacity
          on-demand by leveraging the storage of commodity x86/ARM servers. The distributed
          write maximizes availability and protects data from hardware failures ranging
          from a single disk to an entire site, improving disaster recovery planning.\n\nHedvig
          provides a CSI Driver for accelerating your journey into the container ecosystem.
          CSI Driver enables containerized applications to maintain state by dynamically
          provisioning and consuming Hedvig virtual disks as persistent volumes. \n###
          \n\nHedvig Operator deploys the Hedvig Storage Proxy and the Hedvig CSI
          Driver to manage the life cycle of persistent storage in Kubernetes and
          Openshift.\n\n**Supported Container Orchestrators:**\n* Kubernetes 1.16+\n*
          Red Hat OpenShift 4.6+\n\n**Supported Hedvig Releases:**\n* Hedvig 4.0+\nPlease
          refer to the [documentation](https://documentation.commvault.com/commvault/hedvig/others/pdf/Hedvig_Operator_User_Guide.pdf)
          for steps to install and use the operator.\n\n\n\n\n\n"
        displayName: Hedvig Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - CSI
        - Hedvig
        - Cloud
        - Block
        - NFS
        - iSCSI
        - Snapshot
        - Volume
        links:
        - name: Documentation
          url: https://documentation.commvault.com/commvault/hedvig/article
        - name: Hedvig CSI User Guide
          url: https://documentation.commvault.com/commvault/hedvig/others/pdf/Hedvig_CSI_User_Guide.pdf
        maintainers:
        - email: ashenoy@commvault.com
          name: Abhijith Shenoy
        maturity: stable
        provider:
          name: Commvault
        relatedImages:
        - hedvig/hedvig-operator:2.0.0-ubi
        version: 1.0.2
      entries:
      - name: hedvig-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: hedvig-operator
    provider:
      name: Commvault
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: External Secrets
      provider-url: https://external-secrets.io
    name: external-secrets-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: external-secrets-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": "1234"
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "key": "provider-key",
                      "property": "provider-key-property",
                      "version": "provider-key-version"
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterExternalSecret",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "externalSecretName": "hello-world-es",
                  "externalSecretSpec": {
                    "data": [
                      {
                        "remoteRef": {
                          "key": "provider-key",
                          "property": "provider-key-property",
                          "version": "provider-key-version"
                        },
                        "secretKey": "secret-key-to-be-managed"
                      }
                    ],
                    "dataFrom": [
                      {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    ],
                    "refreshInterval": "1h",
                    "secretStoreRef": {
                      "kind": "SecretStore",
                      "name": "secret-store-name"
                    },
                    "target": {
                      "creationPolicy": "Merge",
                      "name": "my-secret",
                      "template": {
                        "data": {
                          "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                        },
                        "metadata": {
                          "annotations": {},
                          "labels": {}
                        },
                        "templateFrom": [
                          {
                            "configMap": {
                              "items": [
                                {
                                  "key": "alertmanager.yaml"
                                }
                              ],
                              "name": "alertmanager"
                            }
                          }
                        ],
                        "type": "kubernetes.io/dockerconfigjson"
                      }
                    }
                  },
                  "namespaceSelector": {
                    "matchLabels": {
                      "cool": "label"
                    }
                  },
                  "refreshTime": "1m"
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": 1234
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "extract": {
                        "conversionStrategy": "Default",
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    },
                    {
                      "find": {
                        "conversionStrategy": "Unicode",
                        "name": {
                          "regexp": ".*foobar.*"
                        },
                        "path": "path-to-filter",
                        "tags": {
                          "foo": "bar"
                        }
                      }
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "deletionPolicy": "Retain",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.external-secrets.io/v1alpha1",
                "kind": "OperatorConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "requeueInterval": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  },
                  "concurrent": 1,
                  "controllerClass": "",
                  "crds": {
                    "createClusterExternalSecret": true,
                    "createClusterSecretStore": true
                  },
                  "createOperator": true,
                  "deploymentAnnotations": {},
                  "extraArgs": {},
                  "extraEnv": [],
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/external-secrets/external-secrets",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "installCRDs": false,
                  "leaderElect": false,
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "priorityClassName": "",
                  "processClusterExternalSecret": true,
                  "processClusterStore": true,
                  "prometheus": {
                    "enabled": false,
                    "service": {
                      "port": 8080
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "scopedNamespace": "",
                  "scopedRBAC": false,
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "certCheckInterval": "5m",
                    "certDir": "/tmp/certs",
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
          createdAt: "2024-12-03T10:44:19Z"
          description: Operator to configure external-secrets helm-chart based operator
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/external-secrets/external-secrets-helm-operator
          support: external-secrets
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ACRAccessToken
            name: acraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster scoped resource that allows you to manage ExternalSecrets
              across namespaces
            displayName: ClusterExternalSecret
            kind: ClusterExternalSecret
            name: clusterexternalsecrets.external-secrets.io
            version: v1beta1
          - kind: ClusterGenerator
            name: clustergenerators.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1alpha1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1beta1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1beta1
          - kind: ECRAuthorizationToken
            name: ecrauthorizationtokens.generators.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1alpha1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1beta1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1beta1
          - kind: Fake
            name: fakes.generators.external-secrets.io
            version: v1alpha1
          - kind: GCRAccessToken
            name: gcraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - kind: GithubAccessToken
            name: githubaccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Configures external-secrets helm-chart based operator
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.external-secrets.io
            version: v1alpha1
          - kind: Password
            name: passwords.generators.external-secrets.io
            version: v1alpha1
          - kind: PushSecret
            name: pushsecrets.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1alpha1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1beta1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1beta1
          - kind: STSSessionToken
            name: stssessiontokens.generators.external-secrets.io
            version: v1alpha1
          - kind: UUID
            name: uuids.generators.external-secrets.io
            version: v1alpha1
          - kind: VaultDynamicSecret
            name: vaultdynamicsecrets.generators.external-secrets.io
            version: v1alpha1
          - kind: Webhook
            name: webhooks.generators.external-secrets.io
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK (Helm version) to configure **[official external-secrets operator helm chart](https://github.com/external-secrets/external-secrets)**, so it can be installed via OLM without having to do any change on current Helm Charts.

          Before any other resources provided by this Operator can be deployed, it is essential to create an
          OperatorConfig resource.

          The usual Helm Chart file `values.yaml`, like:
          ```yaml
          prometheus:
            enabled: true
            service:
              port: 8080
          resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          need to be encapsulated into a new custom resource called `OperatorConfig`:
          ```yaml
          apiVersion: operator.external-secrets.io/v1alpha1
          kind: OperatorConfig
          metadata:
            name: cluster
          spec:
            prometheus:
              enabled: true
              service:
                port: 8080
            resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          Once the OperatorConfig resource is deployed, the operator will create all helm chart resources, using the custom resource name as a preffix for all resources names, like for example a `Deployment` called `cluster-external-secrets`.

          ## Documentation

          * [Upstream Documentation](https://external-secrets.io)
          * [Helm Operator Install](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/install.md)
          * [Helm Operator Development](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/development.md)
          * [Helm Operator Release](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/release.md)

          ## Contributing

          You can contribute by:

          * Raising any issues you find using External Secrets Operator
          * Fixing issues by opening [Pull Requests](https://github.com/external-secrets/external-secrets-helm-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving documentation
          * Talking about External Secrets Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/external-secrets/external-secrets-helm-operator/issues).

          ## License

          External Secrets Operator is under Apache 2.0 license. See the [LICENSE](LICENSE) file for details.
        displayName: External Secrets Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secrets
        - secrets-manager
        - vault
        - aws
        - azure
        - google
        - ibm
        - akeyless
        - yandex
        - gitlab
        - oracle
        links:
        - name: External Secrets Operator
          url: https://external-secrets.io
        - name: GitHub
          url: https://github.com/external-secrets/external-secrets
        - name: Slack
          url: https://kubernetes.slack.com/messages/external-secrets
        maintainers:
        - email: contact@external-secrets.io
          name: external-secrets community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: External Secrets
          url: https://external-secrets.io
        relatedImages:
        - ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
        version: 0.11.0
      entries:
      - name: external-secrets-operator.v0.11.0
        version: 0.11.0
      - name: external-secrets-operator.v0.10.7
        version: 0.10.7
      - name: external-secrets-operator.v0.10.6
        version: 0.10.6
      - name: external-secrets-operator.v0.10.5
        version: 0.10.5
      - name: external-secrets-operator.v0.10.4
        version: 0.10.4
      - name: external-secrets-operator.v0.10.3
        version: 0.10.3
      - name: external-secrets-operator.v0.10.2
        version: 0.10.2
      - name: external-secrets-operator.v0.10.1
        version: 0.10.1
      - name: external-secrets-operator.v0.10.0
        version: 0.10.0
      - name: external-secrets-operator.v0.9.20
        version: 0.9.20
      - name: external-secrets-operator.v0.9.19
        version: 0.9.19
      - name: external-secrets-operator.v0.9.18
        version: 0.9.18
      - name: external-secrets-operator.v0.9.17
        version: 0.9.17
      - name: external-secrets-operator.v0.9.16
        version: 0.9.16
      - name: external-secrets-operator.v0.9.15-2
        version: 0.9.15-2
      - name: external-secrets-operator.v0.9.14
        version: 0.9.14
      - name: external-secrets-operator.v0.9.13
        version: 0.9.13
      - name: external-secrets-operator.v0.9.12
        version: 0.9.12
      - name: external-secrets-operator.v0.9.11
        version: 0.9.11
      - name: external-secrets-operator.v0.9.10
        version: 0.9.10
      - name: external-secrets-operator.v0.9.9
        version: 0.9.9
      - name: external-secrets-operator.v0.9.8
        version: 0.9.8
      - name: external-secrets-operator.v0.9.7
        version: 0.9.7
      - name: external-secrets-operator.v0.9.6
        version: 0.9.6
      - name: external-secrets-operator.v0.9.5
        version: 0.9.5
      - name: external-secrets-operator.v0.9.4
        version: 0.9.4
      - name: external-secrets-operator.v0.9.3
        version: 0.9.3
      - name: external-secrets-operator.v0.9.2
        version: 0.9.2
      - name: external-secrets-operator.v0.9.1
        version: 0.9.1
      - name: external-secrets-operator.v0.9.0
        version: 0.9.0
      - name: external-secrets-operator.v0.8.7
        version: 0.8.7
      - name: external-secrets-operator.v0.8.6
        version: 0.8.6
      - name: external-secrets-operator.v0.8.5
        version: 0.8.5
      - name: external-secrets-operator.v0.8.4
        version: 0.8.4
      - name: external-secrets-operator.v0.8.3
        version: 0.8.3
      - name: external-secrets-operator.v0.8.2
        version: 0.8.2
      - name: external-secrets-operator.v0.8.1
        version: 0.8.1
      - name: external-secrets-operator.v0.8.0
        version: 0.8.0
      - name: external-secrets-operator.v0.7.2
        version: 0.7.2
      - name: external-secrets-operator.v0.7.1
        version: 0.7.1
      - name: external-secrets-operator.v0.7.0
        version: 0.7.0
      - name: external-secrets-operator.v0.7.0-rc1
        version: 0.7.0-rc1
      - name: external-secrets-operator.v0.6.1
        version: 0.6.1
      - name: external-secrets-operator.v0.6.0
        version: 0.6.0
      - name: external-secrets-operator.v0.6.0-rc1
        version: 0.6.0-rc1
      - name: external-secrets-operator.v0.5.9
        version: 0.5.9
      - name: external-secrets-operator.v0.5.8
        version: 0.5.8
      - name: external-secrets-operator.v0.5.7
        version: 0.5.7
      - name: external-secrets-operator.v0.5.6
        version: 0.5.6
      - name: external-secrets-operator.v0.5.5
        version: 0.5.5
      - name: external-secrets-operator.v0.5.4
        version: 0.5.4
      - name: external-secrets-operator.v0.5.3
        version: 0.5.3
      - name: external-secrets-operator.v0.5.2
        version: 0.5.2
      - name: external-secrets-operator.v0.5.1
        version: 0.5.1
      - name: external-secrets-operator.v0.5.0
        version: 0.5.0
      - name: external-secrets-operator.v0.4.4
        version: 0.4.4
      - name: external-secrets-operator.v0.4.3
        version: 0.4.3
      - name: external-secrets-operator.v0.4.2
        version: 0.4.2
      - name: external-secrets-operator.v0.4.1
        version: 0.4.1
      - name: external-secrets-operator.v0.3.11
        version: 0.3.11
      - name: external-secrets-operator.v0.3.10
        version: 0.3.10
      - name: external-secrets-operator.v0.3.9
        version: 0.3.9
      - name: external-secrets-operator.v0.3.8
        version: 0.3.8
      name: alpha
    - currentCSV: external-secrets-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": "1234"
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "key": "provider-key",
                      "property": "provider-key-property",
                      "version": "provider-key-version"
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterExternalSecret",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "externalSecretName": "hello-world-es",
                  "externalSecretSpec": {
                    "data": [
                      {
                        "remoteRef": {
                          "key": "provider-key",
                          "property": "provider-key-property",
                          "version": "provider-key-version"
                        },
                        "secretKey": "secret-key-to-be-managed"
                      }
                    ],
                    "dataFrom": [
                      {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    ],
                    "refreshInterval": "1h",
                    "secretStoreRef": {
                      "kind": "SecretStore",
                      "name": "secret-store-name"
                    },
                    "target": {
                      "creationPolicy": "Merge",
                      "name": "my-secret",
                      "template": {
                        "data": {
                          "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                        },
                        "metadata": {
                          "annotations": {},
                          "labels": {}
                        },
                        "templateFrom": [
                          {
                            "configMap": {
                              "items": [
                                {
                                  "key": "alertmanager.yaml"
                                }
                              ],
                              "name": "alertmanager"
                            }
                          }
                        ],
                        "type": "kubernetes.io/dockerconfigjson"
                      }
                    }
                  },
                  "namespaceSelector": {
                    "matchLabels": {
                      "cool": "label"
                    }
                  },
                  "refreshTime": "1m"
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": 1234
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "extract": {
                        "conversionStrategy": "Default",
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    },
                    {
                      "find": {
                        "conversionStrategy": "Unicode",
                        "name": {
                          "regexp": ".*foobar.*"
                        },
                        "path": "path-to-filter",
                        "tags": {
                          "foo": "bar"
                        }
                      }
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "deletionPolicy": "Retain",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.external-secrets.io/v1alpha1",
                "kind": "OperatorConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "requeueInterval": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  },
                  "concurrent": 1,
                  "controllerClass": "",
                  "crds": {
                    "createClusterExternalSecret": true,
                    "createClusterSecretStore": true
                  },
                  "createOperator": true,
                  "deploymentAnnotations": {},
                  "extraArgs": {},
                  "extraEnv": [],
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/external-secrets/external-secrets",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "installCRDs": false,
                  "leaderElect": false,
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "priorityClassName": "",
                  "processClusterExternalSecret": true,
                  "processClusterStore": true,
                  "prometheus": {
                    "enabled": false,
                    "service": {
                      "port": 8080
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "scopedNamespace": "",
                  "scopedRBAC": false,
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "certCheckInterval": "5m",
                    "certDir": "/tmp/certs",
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
          createdAt: "2024-12-03T10:44:19Z"
          description: Operator to configure external-secrets helm-chart based operator
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/external-secrets/external-secrets-helm-operator
          support: external-secrets
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ACRAccessToken
            name: acraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster scoped resource that allows you to manage ExternalSecrets
              across namespaces
            displayName: ClusterExternalSecret
            kind: ClusterExternalSecret
            name: clusterexternalsecrets.external-secrets.io
            version: v1beta1
          - kind: ClusterGenerator
            name: clustergenerators.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1alpha1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1beta1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1beta1
          - kind: ECRAuthorizationToken
            name: ecrauthorizationtokens.generators.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1alpha1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1beta1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1beta1
          - kind: Fake
            name: fakes.generators.external-secrets.io
            version: v1alpha1
          - kind: GCRAccessToken
            name: gcraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - kind: GithubAccessToken
            name: githubaccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Configures external-secrets helm-chart based operator
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.external-secrets.io
            version: v1alpha1
          - kind: Password
            name: passwords.generators.external-secrets.io
            version: v1alpha1
          - kind: PushSecret
            name: pushsecrets.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1alpha1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1beta1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1beta1
          - kind: STSSessionToken
            name: stssessiontokens.generators.external-secrets.io
            version: v1alpha1
          - kind: UUID
            name: uuids.generators.external-secrets.io
            version: v1alpha1
          - kind: VaultDynamicSecret
            name: vaultdynamicsecrets.generators.external-secrets.io
            version: v1alpha1
          - kind: Webhook
            name: webhooks.generators.external-secrets.io
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK (Helm version) to configure **[official external-secrets operator helm chart](https://github.com/external-secrets/external-secrets)**, so it can be installed via OLM without having to do any change on current Helm Charts.

          Before any other resources provided by this Operator can be deployed, it is essential to create an
          OperatorConfig resource.

          The usual Helm Chart file `values.yaml`, like:
          ```yaml
          prometheus:
            enabled: true
            service:
              port: 8080
          resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          need to be encapsulated into a new custom resource called `OperatorConfig`:
          ```yaml
          apiVersion: operator.external-secrets.io/v1alpha1
          kind: OperatorConfig
          metadata:
            name: cluster
          spec:
            prometheus:
              enabled: true
              service:
                port: 8080
            resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          Once the OperatorConfig resource is deployed, the operator will create all helm chart resources, using the custom resource name as a preffix for all resources names, like for example a `Deployment` called `cluster-external-secrets`.

          ## Documentation

          * [Upstream Documentation](https://external-secrets.io)
          * [Helm Operator Install](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/install.md)
          * [Helm Operator Development](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/development.md)
          * [Helm Operator Release](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/release.md)

          ## Contributing

          You can contribute by:

          * Raising any issues you find using External Secrets Operator
          * Fixing issues by opening [Pull Requests](https://github.com/external-secrets/external-secrets-helm-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving documentation
          * Talking about External Secrets Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/external-secrets/external-secrets-helm-operator/issues).

          ## License

          External Secrets Operator is under Apache 2.0 license. See the [LICENSE](LICENSE) file for details.
        displayName: External Secrets Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secrets
        - secrets-manager
        - vault
        - aws
        - azure
        - google
        - ibm
        - akeyless
        - yandex
        - gitlab
        - oracle
        links:
        - name: External Secrets Operator
          url: https://external-secrets.io
        - name: GitHub
          url: https://github.com/external-secrets/external-secrets
        - name: Slack
          url: https://kubernetes.slack.com/messages/external-secrets
        maintainers:
        - email: contact@external-secrets.io
          name: external-secrets community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: External Secrets
          url: https://external-secrets.io
        relatedImages:
        - ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
        version: 0.11.0
      entries:
      - name: external-secrets-operator.v0.11.0
        version: 0.11.0
      - name: external-secrets-operator.v0.10.7
        version: 0.10.7
      - name: external-secrets-operator.v0.10.6
        version: 0.10.6
      - name: external-secrets-operator.v0.10.5
        version: 0.10.5
      - name: external-secrets-operator.v0.10.4
        version: 0.10.4
      - name: external-secrets-operator.v0.10.3
        version: 0.10.3
      - name: external-secrets-operator.v0.10.2
        version: 0.10.2
      - name: external-secrets-operator.v0.10.1
        version: 0.10.1
      - name: external-secrets-operator.v0.10.0
        version: 0.10.0
      - name: external-secrets-operator.v0.9.20
        version: 0.9.20
      - name: external-secrets-operator.v0.9.19
        version: 0.9.19
      - name: external-secrets-operator.v0.9.18
        version: 0.9.18
      - name: external-secrets-operator.v0.9.17
        version: 0.9.17
      - name: external-secrets-operator.v0.9.16
        version: 0.9.16
      - name: external-secrets-operator.v0.9.15-2
        version: 0.9.15-2
      - name: external-secrets-operator.v0.9.14
        version: 0.9.14
      - name: external-secrets-operator.v0.9.13
        version: 0.9.13
      - name: external-secrets-operator.v0.9.12
        version: 0.9.12
      - name: external-secrets-operator.v0.9.11
        version: 0.9.11
      - name: external-secrets-operator.v0.9.10
        version: 0.9.10
      - name: external-secrets-operator.v0.9.9
        version: 0.9.9
      - name: external-secrets-operator.v0.9.8
        version: 0.9.8
      - name: external-secrets-operator.v0.9.7
        version: 0.9.7
      - name: external-secrets-operator.v0.9.6
        version: 0.9.6
      - name: external-secrets-operator.v0.9.5
        version: 0.9.5
      - name: external-secrets-operator.v0.9.4
        version: 0.9.4
      - name: external-secrets-operator.v0.9.3
        version: 0.9.3
      - name: external-secrets-operator.v0.9.2
        version: 0.9.2
      - name: external-secrets-operator.v0.9.1
        version: 0.9.1
      - name: external-secrets-operator.v0.9.0
        version: 0.9.0
      - name: external-secrets-operator.v0.8.7
        version: 0.8.7
      - name: external-secrets-operator.v0.8.6
        version: 0.8.6
      - name: external-secrets-operator.v0.8.5
        version: 0.8.5
      - name: external-secrets-operator.v0.8.4
        version: 0.8.4
      - name: external-secrets-operator.v0.8.3
        version: 0.8.3
      - name: external-secrets-operator.v0.8.2
        version: 0.8.2
      - name: external-secrets-operator.v0.8.1
        version: 0.8.1
      - name: external-secrets-operator.v0.8.0
        version: 0.8.0
      - name: external-secrets-operator.v0.7.2
        version: 0.7.2
      - name: external-secrets-operator.v0.7.1
        version: 0.7.1
      - name: external-secrets-operator.v0.7.0
        version: 0.7.0
      - name: external-secrets-operator.v0.7.0-rc1
        version: 0.7.0-rc1
      - name: external-secrets-operator.v0.6.1
        version: 0.6.1
      - name: external-secrets-operator.v0.6.0
        version: 0.6.0
      - name: external-secrets-operator.v0.5.9
        version: 0.5.9
      - name: external-secrets-operator.v0.5.8
        version: 0.5.8
      - name: external-secrets-operator.v0.5.7
        version: 0.5.7
      - name: external-secrets-operator.v0.5.6
        version: 0.5.6
      - name: external-secrets-operator.v0.5.5
        version: 0.5.5
      - name: external-secrets-operator.v0.5.4
        version: 0.5.4
      - name: external-secrets-operator.v0.5.3
        version: 0.5.3
      - name: external-secrets-operator.v0.5.2
        version: 0.5.2
      - name: external-secrets-operator.v0.5.1
        version: 0.5.1
      - name: external-secrets-operator.v0.5.0
        version: 0.5.0
      - name: external-secrets-operator.v0.4.4
        version: 0.4.4
      - name: external-secrets-operator.v0.4.2
        version: 0.4.2
      - name: external-secrets-operator.v0.4.1
        version: 0.4.1
      - name: external-secrets-operator.v0.3.11
        version: 0.3.11
      - name: external-secrets-operator.v0.3.10
        version: 0.3.10
      - name: external-secrets-operator.v0.3.9
        version: 0.3.9
      name: stable
    defaultChannel: alpha
    packageName: external-secrets-operator
    provider:
      name: External Secrets
      url: https://external-secrets.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/kuadrant-operator
    name: rhcl-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhcl-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "authpolicy-sample"
                },
                "spec": {
                  "rules": {
                    "authentication": {
                      "apikey": {
                        "apiKey": {
                          "selector": {}
                        },
                        "credentials": {
                          "authorizationHeader": {
                            "prefix": "APIKEY"
                          }
                        }
                      }
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "DNSPolicy",
                "metadata": {
                  "name": "dnspolicy-sample"
                },
                "spec": {
                  "healthCheck": {
                    "protocol": "HTTP"
                  },
                  "providerRefs": [
                    {
                      "name": "provider-ref"
                    }
                  ],
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "ratelimitpolicy-sample"
                },
                "spec": {
                  "limits": {
                    "toys": {
                      "rates": [
                        {
                          "limit": 50,
                          "window": "1m"
                        }
                      ]
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "TLSPolicy",
                "metadata": {
                  "name": "tlspolicy-sample"
                },
                "spec": {
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "self-signed-ca"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/rhcl-rhel9-operator@sha256:92f14fd8de74bc72c407947cf5fa4739a7b455e5d351baeb64521e4c07f02a9e
          createdAt: "2025-04-28T13:10:44Z"
          description: Red Hat Connectivity Link enables you to secure, protect, and
            connect your APIs and applications in multicluster, multicloud, and hybrid
            cloud environments
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthPolicy enables authentication and authorization for service
              workloads in a Gateway API network
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1
          - description: DNSPolicy configures how North-South based traffic should
              be balanced and reach the gateways
            displayName: DNSPolicy
            kind: DNSPolicy
            name: dnspolicies.kuadrant.io
            version: v1
          - description: Kuadrant configures installations of Kuadrant Service Protection
              components
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: RateLimitPolicy enables rate limiting for service workloads
              in a Gateway API network
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1
          - description: TLSPolicy provides tls for gateway listeners by managing
              the lifecycle of tls certificates
            displayName: TLSPolicy
            kind: TLSPolicy
            name: tlspolicies.kuadrant.io
            version: v1
        description: Red Hat Connectivity Link enables you to secure, protect, and
          connect your APIs and applications in multicluster, multicloud, and hybrid
          cloud environments
        displayName: Red Hat Connectivity Link
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: GitHub
          url: https://github.com/kuadrant/kuadrant-operator
        - name: Documentation
          url: https://github.com/Kuadrant/kuadrant-operator/tree/main/docs
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - registry.redhat.io/rhcl-1/rhcl-console-plugin-rhel9@sha256:ae92d638d993d782986e761d5aa3c7c5da2b8af0aabc1decb332a4f1487dacd3
        - registry.redhat.io/rhcl-1/rhcl-rhel9-operator@sha256:92f14fd8de74bc72c407947cf5fa4739a7b455e5d351baeb64521e4c07f02a9e
        - registry.redhat.io/rhcl-1/wasm-shim-rhel9@sha256:845bb8af57f3d219aa09b9c0bb20fa945306f3b413e34e196760c8cdf624a532
        version: 1.1.0
      entries:
      - name: rhcl-operator.v1.1.0
        version: 1.1.0
      - name: rhcl-operator.v1.0.2
        version: 1.0.2
      - name: rhcl-operator.v1.0.1
        version: 1.0.1
      - name: rhcl-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: rhcl-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/kuadrant-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sqs-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sqs-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sqs.services.k8s.aws/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sqs-controller:1.2.1
          createdAt: "2025-09-19T19:14:25Z"
          description: AWS SQS controller is a service controller for managing SQS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Queue represents the state of an AWS sqs Queue resource.
            displayName: Queue
            kind: Queue
            name: queues.sqs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Simple Queue Service (SQS) resources in AWS from within your Kubernetes cluster.

          **About Amazon SQS**

          Amazon Simple Queue Service (Amazon SQS) offers a secure, durable, and available hosted queue that lets you integrate and decouple distributed software systems and components. Amazon SQS offers common constructs such as [dead-letter queues](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGuide/sqs-dead-letter-queues.html) and [cost allocation tags](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGuide/sqs-queue-tags.html). It provides a generic web services API that you can access using any programming language that the AWS SDK supports.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SQS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sqs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SQS Developer Resources
          url: https://aws.amazon.com/sqs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sqs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sqs-controller:1.2.1
        version: 1.2.1
      entries:
      - name: ack-sqs-controller.v1.2.1
        version: 1.2.1
      - name: ack-sqs-controller.v1.2.0
        version: 1.2.0
      - name: ack-sqs-controller.v1.1.16
        version: 1.1.16
      - name: ack-sqs-controller.v1.1.14
        version: 1.1.14
      - name: ack-sqs-controller.v1.1.13
        version: 1.1.13
      - name: ack-sqs-controller.v1.1.12
        version: 1.1.12
      - name: ack-sqs-controller.v1.1.11
        version: 1.1.11
      - name: ack-sqs-controller.v1.1.10
        version: 1.1.10
      - name: ack-sqs-controller.v1.1.9
        version: 1.1.9
      - name: ack-sqs-controller.v1.1.8
        version: 1.1.8
      - name: ack-sqs-controller.v1.1.7
        version: 1.1.7
      - name: ack-sqs-controller.v1.1.6
        version: 1.1.6
      - name: ack-sqs-controller.v1.1.5
        version: 1.1.5
      - name: ack-sqs-controller.v1.1.4
        version: 1.1.4
      - name: ack-sqs-controller.v1.1.3
        version: 1.1.3
      - name: ack-sqs-controller.v1.1.2
        version: 1.1.2
      - name: ack-sqs-controller.v1.1.1
        version: 1.1.1
      - name: ack-sqs-controller.v1.1.0
        version: 1.1.0
      - name: ack-sqs-controller.v1.0.17
        version: 1.0.17
      - name: ack-sqs-controller.v1.0.16
        version: 1.0.16
      - name: ack-sqs-controller.v1.0.15
        version: 1.0.15
      - name: ack-sqs-controller.v1.0.14
        version: 1.0.14
      - name: ack-sqs-controller.v1.0.13
        version: 1.0.13
      - name: ack-sqs-controller.v1.0.12
        version: 1.0.12
      - name: ack-sqs-controller.v1.0.10
        version: 1.0.10
      - name: ack-sqs-controller.v1.0.7
        version: 1.0.7
      - name: ack-sqs-controller.v1.0.5
        version: 1.0.5
      - name: ack-sqs-controller.v1.0.3
        version: 1.0.3
      - name: ack-sqs-controller.v1.0.2
        version: 1.0.2
      - name: ack-sqs-controller.v1.0.1
        version: 1.0.1
      - name: ack-sqs-controller.v1.0.0
        version: 1.0.0
      - name: ack-sqs-controller.v0.0.7
        version: 0.0.7
      - name: ack-sqs-controller.v0.0.6
        version: 0.0.6
      - name: ack-sqs-controller.v0.0.5
        version: 0.0.5
      - name: ack-sqs-controller.v0.0.4
        version: 0.0.4
      - name: ack-sqs-controller.v0.0.2
        version: 0.0.2
      - name: ack-sqs-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-sqs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-rds-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-rds-controller.v1.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rds.services.k8s.aws/v1alpha1",
                "kind": "DBParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "rds.services.k8s.aws/v1alpha1",
                "kind": "DBSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/rds-controller:1.6.2
          createdAt: "2025-09-22T17:27:42Z"
          description: AWS RDS controller is a service controller for managing RDS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DBClusterEndpoint represents the state of an AWS rds DBClusterEndpoint
              resource.
            displayName: DBClusterEndpoint
            kind: DBClusterEndpoint
            name: dbclusterendpoints.rds.services.k8s.aws
            version: v1alpha1
          - description: DBClusterParameterGroup represents the state of an AWS rds
              DBClusterParameterGroup resource.
            displayName: DBClusterParameterGroup
            kind: DBClusterParameterGroup
            name: dbclusterparametergroups.rds.services.k8s.aws
            version: v1alpha1
          - description: DBCluster represents the state of an AWS rds DBCluster resource.
            displayName: DBCluster
            kind: DBCluster
            name: dbclusters.rds.services.k8s.aws
            version: v1alpha1
          - description: DBClusterSnapshot represents the state of an AWS rds DBClusterSnapshot
              resource.
            displayName: DBClusterSnapshot
            kind: DBClusterSnapshot
            name: dbclustersnapshots.rds.services.k8s.aws
            version: v1alpha1
          - description: DBInstance represents the state of an AWS rds DBInstance
              resource.
            displayName: DBInstance
            kind: DBInstance
            name: dbinstances.rds.services.k8s.aws
            version: v1alpha1
          - description: DBParameterGroup represents the state of an AWS rds DBParameterGroup
              resource.
            displayName: DBParameterGroup
            kind: DBParameterGroup
            name: dbparametergroups.rds.services.k8s.aws
            version: v1alpha1
          - description: DBProxy represents the state of an AWS rds DBProxy resource.
            displayName: DBProxy
            kind: DBProxy
            name: dbproxies.rds.services.k8s.aws
            version: v1alpha1
          - description: DBSnapshot represents the state of an AWS rds DBSnapshot
              resource.
            displayName: DBSnapshot
            kind: DBSnapshot
            name: dbsnapshots.rds.services.k8s.aws
            version: v1alpha1
          - description: DBSubnetGroup represents the state of an AWS rds DBSubnetGroup
              resource.
            displayName: DBSubnetGroup
            kind: DBSubnetGroup
            name: dbsubnetgroups.rds.services.k8s.aws
            version: v1alpha1
          - description: GlobalCluster represents the state of an AWS rds GlobalCluster
              resource.
            displayName: GlobalCluster
            kind: GlobalCluster
            name: globalclusters.rds.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Relational Database Service ("RDS") resources in AWS from within your Kubernetes cluster.

          **About Amazon RDS**

          Amazon Relational Database Service (Amazon RDS) makes it easy to set up, operate, and scale a relational database in the cloud. It provides cost-efficient and resizable capacity while automating time-consuming administration tasks such as hardware provisioning, database setup, patching and backups. It frees you to focus on your applications so you can give them the fast performance, high availability, security and compatibility they need.

          Amazon RDS is available on several database instance types - optimized for memory, performance or I/O -and provides you with six familiar database engines to choose from, including Amazon Aurora, PostgreSQL, MySQL, MariaDB, Oracle Database, and SQL Server. You can use the AWS Database Migration Service to easily migrate or replicate your existing databases to Amazon RDS.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon RDS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rds
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon RDS Developer Resources
          url: https://aws.amazon.com/rds/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: rds maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/rds-controller:1.6.2
        version: 1.6.2
      entries:
      - name: ack-rds-controller.v1.6.2
        version: 1.6.2
      - name: ack-rds-controller.v1.6.1
        version: 1.6.1
      - name: ack-rds-controller.v1.6.0
        version: 1.6.0
      - name: ack-rds-controller.v1.5.0
        version: 1.5.0
      - name: ack-rds-controller.v1.4.26
        version: 1.4.26
      - name: ack-rds-controller.v1.4.25
        version: 1.4.25
      - name: ack-rds-controller.v1.4.24
        version: 1.4.24
      - name: ack-rds-controller.v1.4.22
        version: 1.4.22
      - name: ack-rds-controller.v1.4.21
        version: 1.4.21
      - name: ack-rds-controller.v1.4.20
        version: 1.4.20
      - name: ack-rds-controller.v1.4.19
        version: 1.4.19
      - name: ack-rds-controller.v1.4.18
        version: 1.4.18
      - name: ack-rds-controller.v1.4.17
        version: 1.4.17
      - name: ack-rds-controller.v1.4.16
        version: 1.4.16
      - name: ack-rds-controller.v1.4.15
        version: 1.4.15
      - name: ack-rds-controller.v1.4.14
        version: 1.4.14
      - name: ack-rds-controller.v1.4.13
        version: 1.4.13
      - name: ack-rds-controller.v1.4.12
        version: 1.4.12
      - name: ack-rds-controller.v1.4.11
        version: 1.4.11
      - name: ack-rds-controller.v1.4.10
        version: 1.4.10
      - name: ack-rds-controller.v1.4.8
        version: 1.4.8
      - name: ack-rds-controller.v1.4.7
        version: 1.4.7
      - name: ack-rds-controller.v1.4.6
        version: 1.4.6
      - name: ack-rds-controller.v1.4.4
        version: 1.4.4
      - name: ack-rds-controller.v1.4.3
        version: 1.4.3
      - name: ack-rds-controller.v1.4.2
        version: 1.4.2
      - name: ack-rds-controller.v1.4.1
        version: 1.4.1
      - name: ack-rds-controller.v1.4.0
        version: 1.4.0
      - name: ack-rds-controller.v1.3.0
        version: 1.3.0
      - name: ack-rds-controller.v1.2.5
        version: 1.2.5
      - name: ack-rds-controller.v1.2.4
        version: 1.2.4
      - name: ack-rds-controller.v1.2.3
        version: 1.2.3
      - name: ack-rds-controller.v1.2.2
        version: 1.2.2
      - name: ack-rds-controller.v1.2.1
        version: 1.2.1
      - name: ack-rds-controller.v1.2.0
        version: 1.2.0
      - name: ack-rds-controller.v1.1.11
        version: 1.1.11
      - name: ack-rds-controller.v1.1.9
        version: 1.1.9
      - name: ack-rds-controller.v1.1.8
        version: 1.1.8
      - name: ack-rds-controller.v1.1.7
        version: 1.1.7
      - name: ack-rds-controller.v1.1.6
        version: 1.1.6
      - name: ack-rds-controller.v1.1.5
        version: 1.1.5
      - name: ack-rds-controller.v1.1.4
        version: 1.1.4
      - name: ack-rds-controller.v1.1.3
        version: 1.1.3
      - name: ack-rds-controller.v1.1.2
        version: 1.1.2
      - name: ack-rds-controller.v1.1.1
        version: 1.1.1
      - name: ack-rds-controller.v1.1.0
        version: 1.1.0
      - name: ack-rds-controller.v1.0.0
        version: 1.0.0
      - name: ack-rds-controller.v0.1.3
        version: 0.1.3
      - name: ack-rds-controller.v0.1.1
        version: 0.1.1
      - name: ack-rds-controller.v0.1.0
        version: 0.1.0
      - name: ack-rds-controller.v0.0.30
        version: 0.0.30
      - name: ack-rds-controller.v0.0.29
        version: 0.0.29
      - name: ack-rds-controller.v0.0.28
        version: 0.0.28
      - name: ack-rds-controller.v0.0.27
        version: 0.0.27
      - name: ack-rds-controller.v0.0.26
        version: 0.0.26
      - name: ack-rds-controller.v0.0.25
        version: 0.0.25
      - name: ack-rds-controller.v0.0.24
        version: 0.0.24
      - name: ack-rds-controller.v0.0.23
        version: 0.0.23
      - name: ack-rds-controller.v0.0.22
        version: 0.0.22
      - name: ack-rds-controller.v0.0.21
        version: 0.0.21
      - name: ack-rds-controller.v0.0.20
        version: 0.0.20
      - name: ack-rds-controller.v0.0.19
        version: 0.0.19
      - name: ack-rds-controller.v0.0.18
        version: 0.0.18
      - name: ack-rds-controller.v0.0.17
        version: 0.0.17
      - name: ack-rds-controller.v0.0.16
        version: 0.0.16
      - name: ack-rds-controller.v0.0.15
        version: 0.0.15
      - name: ack-rds-controller.v0.0.14
        version: 0.0.14
      - name: ack-rds-controller.v0.0.13
        version: 0.0.13
      - name: ack-rds-controller.v0.0.12
        version: 0.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-rds-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: integrity-shield-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: integrity-shield-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/open-cluster-management/integrity-shield-observer",
                    "imagePullPolicy": "Always",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "provenanece": true,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  resp.status_code == 200\n  result := json.unmarshal(resp.raw_body)\n  result.allow == false\n  not is_detect_mode\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"tls_insecure_skip_verify\": true\n  })\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "requestHandlerConfig": "defaultConstraintAction:\n  audit:\n    inform: true\n  admissionControl:\n    enforce: false\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/open-cluster-management/integrity-shield-api",
                    "imagePullPolicy": "Always",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/open-cluster-management/integrity-shield-operator:0.3.0
          createdAt: "2021-10-14T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.

          [K8s Integrity Shield](https://github.com/open-cluster-management/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.

          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Detect (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster

          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).

          ## Preparations before installation

          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).


          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.

          If you want to change the settings such as default run mode (detection/enforcement) or audit interval,  please check [here](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).

          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-7b7f768bf7-fhrpg                           1/1     Running   0          20s
          integrity-shield-observer-5bc66f75f7-tn8fw                      1/1     Running   0          25s
          integrity-shield-operator-controller-manager-65b7fb58f7-j25zd   2/2     Running   0          3h5m
          ```

          After installation, you can protect Kubernetes resources by following this [document](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).

          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:

          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)

          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/open-cluster-management/integrity-shield
        - name: Operator Documentation
          url: https://github.com/open-cluster-management/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - quay.io/open-cluster-management/integrity-shield-operator:0.3.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        version: 0.3.0
      entries:
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.0
    - currentCSV: integrity-shield-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/open-cluster-management/integrity-shield-observer",
                    "imagePullPolicy": "Always",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  resp.status_code == 200\n  result := json.unmarshal(resp.raw_body)\n  result.allow == false\n  not is_detect_mode\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"tls_insecure_skip_verify\": true\n  })\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: detect\n  admissionOnly: false\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/open-cluster-management/integrity-shield-api",
                    "imagePullPolicy": "Always",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/open-cluster-management/integrity-shield-operator:0.3.1
          createdAt: "2021-10-22T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.

          [K8s Integrity Shield](https://github.com/open-cluster-management/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.

          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Detect (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster

          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).

          ## Preparations before installation

          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).


          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.

          If you want to change the settings such as default run mode (detection/enforcement) or audit interval,  please check [here](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).

          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-7b7f768bf7-fhrpg                           1/1     Running   0          20s
          integrity-shield-observer-5bc66f75f7-tn8fw                      1/1     Running   0          25s
          integrity-shield-operator-controller-manager-65b7fb58f7-j25zd   2/2     Running   0          3h5m
          ```

          After installation, you can protect Kubernetes resources by following this [document](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).

          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:

          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)

          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/open-cluster-management/integrity-shield
        - name: Operator Documentation
          url: https://github.com/open-cluster-management/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/open-cluster-management/integrity-shield-operator:0.3.1
        version: 0.3.1
      entries:
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.1
    - currentCSV: integrity-shield-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/stolostron/integrity-shield-observer",
                    "imagePullPolicy": "IfNotPresent",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  should_be_denied(resp)\n  not is_detect_mode\n  result := get_result(resp)\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"timeout\": \"2500ms\",\n    \"raise_error\": false,\n    \"tls_ca_cert\": REPLACE_WITH_CA_CERT\n  })\n}\n\n# get decision result from API response\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  not is_null(raw_body)\n  result := json.unmarshal(raw_body)\n}\n# if error, return it as result\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  is_null(raw_body)\n  error_msg := object.get(resp, \"error\", \"\")\n  result := {\"http_error\":error_msg}\n}\n\n# if API response is null, deny\nshould_be_denied(resp) {\n  is_null(resp)\n}\n\n# if API response has status_code other than 200, deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  resp.status_code != 200\n}\n\n# if decision result is \"allow == false\" or not set, then deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  result := get_result(resp)\n  allow := object.get(result, \"allow\", false)\n  allow == false\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "reporter": {
                    "image": "quay.io/stolostron/integrity-shield-reporter",
                    "imagePullPolicy": "IfNotPresent",
                    "intervalSeconds": "10",
                    "name": "integrity-shield-reporter",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: inform\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\ndecisionReporterConfig:\n  enabled: true\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/stolostron/integrity-shield-api",
                    "imagePullPolicy": "IfNotPresent",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/stolostron/integrity-shield-operator:0.3.2
          createdAt: "2022-01-25T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.15.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.
          [K8s Integrity Shield](https://github.com/stolostron/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.
          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Inform (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster
          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).
          ## Preparations before installation
          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).
          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.
          If you want to change the settings such as default run mode (inform/enforce) or audit interval,  please check [here](https://github.com/stolostron/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).
          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-5799f69896-zbcjs                           2/2     Running   0          88s
          integrity-shield-observer-757f58f7d7-7pf5h                      1/1     Running   0          55s
          integrity-shield-operator-controller-manager-6f94d7d684-dg5bt   2/2     Running   0          4m27s
          ```
          After installation, you can protect Kubernetes resources by following this [document](https://github.com/stolostron/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).
          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:
          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)
          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/stolostron/integrity-shield
        - name: Operator Documentation
          url: https://github.com/stolostron/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/stolostron/integrity-shield-operator:0.3.2
        version: 0.3.2
      entries:
      - name: integrity-shield-operator.v0.3.2
        version: 0.3.2
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.2
    - currentCSV: integrity-shield-operator.v0.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": false,
                    "image": "quay.io/stolostron/integrity-shield-observer",
                    "imagePullPolicy": "IfNotPresent",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  should_be_denied(resp)\n  not is_detect_mode\n  result := get_result(resp)\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"timeout\": \"2500ms\",\n    \"raise_error\": false,\n    \"tls_ca_cert\": REPLACE_WITH_CA_CERT\n  })\n}\n\n# get decision result from API response\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  not is_null(raw_body)\n  result := json.unmarshal(raw_body)\n}\n# if error, return it as result\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  is_null(raw_body)\n  error_msg := object.get(resp, \"error\", \"\")\n  result := {\"http_error\":error_msg}\n}\n\n# if API response is null, deny\nshould_be_denied(resp) {\n  is_null(resp)\n}\n\n# if API response has status_code other than 200, deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  resp.status_code != 200\n}\n\n# if decision result is \"allow == false\" or not set, then deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  result := get_result(resp)\n  allow := object.get(result, \"allow\", false)\n  allow == false\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "reporter": {
                    "image": "quay.io/stolostron/integrity-shield-reporter",
                    "imagePullPolicy": "IfNotPresent",
                    "intervalSeconds": "10",
                    "name": "integrity-shield-reporter",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: inform\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\ndecisionReporterConfig:\n  enabled: true\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/stolostron/integrity-shield-api",
                    "imagePullPolicy": "IfNotPresent",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/stolostron/integrity-shield-operator:0.3.3
          createdAt: "2022-03-18T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.15.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.
          [K8s Integrity Shield](https://github.com/stolostron/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.
          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Inform (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster
          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).
          ## Preparations before installation
          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).
          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.
          If you want to change the settings such as default run mode (inform/enforce) or audit interval,  please check [here](https://github.com/stolostron/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).
          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-5799f69896-zbcjs                           2/2     Running   0          88s
          integrity-shield-observer-757f58f7d7-7pf5h                      1/1     Running   0          55s
          integrity-shield-operator-controller-manager-6f94d7d684-dg5bt   2/2     Running   0          4m27s
          ```
          After installation, you can protect Kubernetes resources by following this [document](https://github.com/stolostron/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).
          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:
          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)
          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/stolostron/integrity-shield
        - name: Operator Documentation
          url: https://github.com/stolostron/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/stolostron/integrity-shield-operator:0.3.3
        version: 0.3.3
      entries:
      - name: integrity-shield-operator.v0.3.3
        version: 0.3.3
      - name: integrity-shield-operator.v0.3.2
        version: 0.3.2
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.3
    defaultChannel: alpha-0.3.3
    packageName: integrity-shield-operator
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: akaris
      provider-url: https://github.com/andreaskaris/sosreport-operator
    name: sosreport-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sosreport-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "support.openshift.io/v1alpha1",
                "kind": "Sosreport",
                "metadata": {
                  "name": "sosreport-sample"
                },
                "spec": {
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Sosreport is the Schema for the sosreports API
            displayName: Sosreport
            kind: Sosreport
            name: sosreports.support.openshift.io
            version: v1alpha1
        description: Operator that facilitates the generation of sosreports
        displayName: sosreport-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sosreport
        - debugging
        - troubleshooting
        links:
        - name: Sosreport Operator
          url: https://github.com/andreaskaris/sosreport-operator
        maintainers:
        - email: ak.karis@gmail.com
          name: Andreas Karis
        maturity: alpha
        provider:
          name: akaris
          url: https://github.com/andreaskaris/sosreport-operator
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/akaris/sosreport-operator:0.0.2
        version: 0.0.2
      entries:
      - name: sosreport-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: sosreport-operator
    provider:
      name: akaris
      url: https://github.com/andreaskaris/sosreport-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: awss3-operator-registry
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: awss3operator.1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/screeley44/aws-s3-provisioner:v1.0.0
          createdAt: "2019-05-02 16:12:00"
          description: Manage the full lifecycle of installing, configuring and managing
            AWS S3 Provisioner.
          repository: https://github.com/yard-turkey/aws-s3-provisioner
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - description: instance of an AWS S3 Bucket
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
          - description: Request for an AWS S3 Bucket
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
        description: |
          AWS S3 Operator will deploy the AWS S3 Provisioner which will dynamically or statically
          provision AWS S3 Bucket storage and access. The operator deploys the ObjectBucket (OB) and ObjectBucketClaim (OBC)
          CustomResourceDefinitions. The OB/OBC model follows the traditional Kubernetes PV/PVC pattern, when an OBC is detected
          the operator will act on the OBC to either provision a brand new S3 Bucket in AWS or gain access to an existing
          S3 Bucket in AWS. The operator produces an OB and ConfigMap and Secret which can then be consumed by application pods.

          **Important Note**: Currently, while in preview, this operator does not
          support automatic upgrades. You must remove the old version of the operator
          manually before installing a new version.
          ## Using AWS S3 Operator

          ### Administrator Creates Secret
          This secret will contain the elevated/admin privileges needed by the provisioner
          to properly access and create S3 Buckets and IAM users and policies. The AWS Access ID
          and AWS Secret Key will be needed for this.
          1. Create the Kubernetes Secret for the Provisioner's Owner Access.
          ```yaml
          apiVersion: v1
          kind: Secret
          metadata:
            name: s3-bucket-owner [1]
            namespace: s3-provisioner [2]
          type: Opaque
          data:
            AWS_ACCESS_KEY_ID: *base64 encoded value* [3]
            AWS_SECRET_ACCESS_KEY: *base64 encoded value* [4]
          ```
          1. Name of the secret, this will be referenced in StorageClass.
          1. Namespace where the Secret will exist.
          1. Your AWS_ACCESS_KEY_ID base64 encoded.
          1. Your AWS_SECRET_ACCESS_KEY base64 encoded.
          ```
           # kubectl create -f creds.yaml
          secret/s3-bucket-owner created
          ```
          ### Administrator Creates StorageClass
          The StorageClass defines the name of the provisioner and holds other properties that are needed to provision a new bucket, including
          the Owner Secret and Namespace, and the AWS Region.
          #### Greenfield Example:
          For Greenfield, a new, dynamic bucket will be generated.
          1. Create the Kubernetes StorageClass for the Provisioner.
          ```yaml
          kind: StorageClass
          apiVersion: storage.k8s.io/v1
          metadata:
            name: s3-buckets [1]
          provisioner: aws-s3.io/bucket [2]
          parameters:
            region: us-west-1 [3]
            secretName: s3-bucket-owner [4]
            secretNamespace: s3-provisioner [5]
          reclaimPolicy: Delete [6]
          ```
          1. Name of the StorageClass, this will be referenced in the User ObjectBucketClaim.
          1. Provisioner name
          1. AWS Region that the StorageClass will serve
          1. Name of the bucket owner Secret created above
          1. Namespace where the Secret will exist
          1. reclaimPolicy (Delete or Retain) indicates if the bucket can be deleted when the OBC is deleted.
          **NOTE:** the absence of the `bucketName` Parameter key in the storage class indicates this is a new bucket and its name is based on the bucket name fields in the OBC.
          ```
           # kubectl create -f storageclass-greenfield.yaml
          storageclass.storage.k8s.io/s3-buckets created
          ```
          #### Brownfield Example:
          For brownfield, the StorageClass defines the name of the provisioner and the name of the existing bucket. It also includes other properties needed by the target
          provisioner, including: the Owner Secret and Namespace, and the AWS Region
          1. Create the Kubernetes StorageClass for the Provisioner.
          ```yaml
          kind: StorageClass
          apiVersion: storage.k8s.io/v1
          metadata:
            name: s3-existing-buckets [1]
          provisioner: aws-s3.io/bucket [2]
          parameters:
            bucketName: my-existing-bucket [3]
            region: us-west-1 [4]
            secretName: s3-bucket-owner [5]
            secretNamespace: s3-provisioner [6]
          ```
          1. Name of the StorageClass, this will be referenced in the User ObjectBucketClaim.
          1. Provisioner name
          1. Name of the existing bucket
          1. AWS Region that the StorageClass will serve
          1. Name of the bucket owner Secret created above
          1. Namespace for that bucket owner secret
          **NOTE:** the storage class's `reclaimPolicy` is ignored for existing buckets.
          ```
           # kubectl create -f storageclass-brownfield.yaml
          storageclass.storage.k8s.io/s3-buckets created
          ```
          ### User Creates ObjectBucketClaim
          An ObjectBucketClaim follows the same concept as a PVC, in that
          it is a request for Object Storage, the user doesn't need to
          concern him/herself with the underlying storage, just that
          they need access to it. The user will work with the cluster/storage
          administrator to get the proper StorageClass needed and will
          then request access via the OBC.
          #### Greenfield Request Example:
          1. Create the ObjectBucketClaim.
          ```yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucketClaim
          metadata:
            name: myobc [1]
            namespace: s3-provisioner [2]
          spec:
            generateBucketName: mybucket [3]
            bucketName: my-awesome-bucket [4]
            storageClassName: s3-buckets [5]
          ```
          1. Name of the OBC
          1. Namespace of the OBC
          1. Name prepended to a random string used to generate a bucket name. It is ignored if bucketName is defined
          1. Name of new bucket which must be unique across all AWS regions, otherwise an error occurs when creating the bucket. If present, this name overrides `generateName`
          1. StorageClass name
          **NOTE:** if both `generateBucketName` and `bucketName` are omitted, and the storage class does _not_ define a bucket name, then a new, random bucket name is generated with no prefix.
          ```
           # kubectl create -f obc-brownfield.yaml
          objectbucketclaim.objectbucket.io/myobc created
          ```
          #### Brownfield Request Example:
          1. Create the ObjectBucketClaim.
          ```yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucketClaim
          metadata:
            name: myobc [1]
            namespace: s3-provisioner [2]
          spec:
            storageClassName: s3-existing-buckets [3]
          ```
          1. Name of the OBC
          1. Namespace of the OBC
          1. StorageClass name
          **NOTE:** in the OBC here there is no reference to the bucket's name. This is defined in the storage class and is not a concern of the user creating the claim to this bucket.  An OBC does have fields for defining a bucket name for greenfield use only.
          ```
           # kubectl create -f obc-brownfield.yaml
          objectbucketclaim.objectbucket.io/myobc created
          ```
          ### Results and Recap
          Let's pause for a moment and digest what just happened.
          After creating the OBC, and assuming the S3 provisioner is running, we now have
          the following Kubernetes resources:
          .  a global ObjectBucket (OB) which contains: bucket endpoint info (including region and bucket name), a reference to the OBC, and a reference to the storage class. Unique to S3, the OB also contains the bucket Amazon Resource Name (ARN).Note: there is always a 1:1 relationship between an OBC and an OB.
          .  a ConfigMap in the same namespace as the OBC, which contains the same endpoint data found in the OB.
          .  a Secret in the same namespace as the OBC, which contains the AWS key-pairs needed to access the bucket.
          And of course, we have a *new* AWS S3 Bucket which you should be able to see via the AWS Console.
          *ObjectBucket*
          ```yaml
           # kubectl get ob obc-s3-provisioner-my-awesome-bucket -o yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucket
          metadata:
            creationTimestamp: "2019-04-03T15:42:22Z"
            generation: 1
            name: obc-s3-provisioner-my-awesome-bucket
            resourceVersion: "15057"
            selfLink: /apis/objectbucket.io/v1alpha1/objectbuckets/obc-s3-provisioner-my-awesome-bucket
            uid: 0bfe8e84-576d-4c4e-984b-f73c4460f736
          spec:
            Connection:
              additionalState:
                ARN: arn:aws:iam::<accountid>:policy/my-awesome-bucket-vSgD5 [1]
                UserName: my-awesome-bucket-vSgD5 [2]
              endpoint:
                additionalConfig: null
                bucketHost: s3-us-west-1.amazonaws.com
                bucketName: my-awesome-bucket [3]
                bucketPort: 443
                region: us-west-1
                ssl: true
                subRegion: ""
            claimRef: null [4]
            reclaimPolicy: null
            storageClassName: s3-buckets [5]
          ```
          1. The AWS Policy created for this user and bucket.
          1. The new user generated by the Provisioner to access this existing bucket.
          1. The bucket name.
          1. The reference to the OBC (not filled in yet).
          1. The reference to the StorageClass used.
          *ConfigMap*
          ```yaml
           # kubectl get cm myobc -n s3-provisioner -o yaml
          apiVersion: v1
          data:
            BUCKET_HOST: s3-us-west-1.amazonaws.com [1]
            BUCKET_NAME: my-awesome-bucket [2]
            BUCKET_PORT: "443"
            BUCKET_REGION: us-west-1
            BUCKET_SSL: "true"
            BUCKET_SUBREGION: ""
          kind: ConfigMap
          metadata:
            creationTimestamp: "2019-04-01T19:11:38Z"
            finalizers:
            - objectbucket.io/finalizer
            name: my-awesome-bucket
            namespace: s3-provisioner
            resourceVersion: "892"
            selfLink: /api/v1/namespaces/s3-provisioner/configmaps/my-awesome-bucket
            uid: 2edcc58a-aff8-4a29-814a-ffbb6439a9cd
          ```
          1. The AWS S3 host.
          1. The name of the new bucket we are gaining access to.
          *Secret*
          ```yaml
           # kubectl get secret my-awesome-bucket -n s3-provisioner -o yaml
          apiVersion: v1
          data:
            AWS_ACCESS_KEY_ID: *the_new_access_id* [1]
            AWS_SECRET_ACCESS_KEY: *the_new_access_key_value* [2]
          kind: Secret
          metadata:
            creationTimestamp: "2019-04-03T15:42:22Z"
            finalizers:
            - objectbucket.io/finalizer
            name: my-awesome-bucket
            namespace: s3-provisioner
            resourceVersion: "15058"
            selfLink: /api/v1/namespaces/s3-provisioner/secrets/screeley-provb-5
            uid: 225c71a5-9d75-4ccc-b41f-bfe91b272a13
          type: Opaque
          ```
          1. The new generated AWS Access Key ID.
          1. The new generated AWS Secret Access Key.
          What happened in AWS? The first thing we do on any OBC request is
          create a new IAM user and generate Access ID and Secret Keys.
          This allows us to also better control ACLs. We also create a policy
          in IAM which we then attach to the user and bucket. We also created a new bucket, called *my-awesome-bucket*.
          When the OBC is deleted all of its Kubernetes and AWS resources will also be deleted, which includes:
          the generated OB, Secret, ConfigMap, IAM user, and policy.
          If the _retainPolicy_ on the StorageClass for this bucket is *"Delete"*, then, in addition to the above cleanup, the physical bucket is also deleted.
          **NOTE:** The actual bucket is only deleted if the Storage Class's _reclaimPolicy_ is "Delete".
          ### User Creates Pod
          Now that we have our bucket and connection/access information, a pod
          can be used to access the bucket. This can be done in several different
          ways, but the key here is that the provisioner has provided the proper
          endpoints and keys to access the bucket. The user then simply references
          the keys.
          1. Create a Sample Pod to Access the Bucket.
          ```yaml
          apiVersion: v1
          kind: Pod
          metadata:
            name: photo1
            labels:
              name: photo1
          spec:
            containers:
            - name: photo1
              image: docker.io/screeley44/photo-gallery:latest
              imagePullPolicy: Always
              envFrom:
              - configMapRef:
                  name: my-awesome-bucket <1>
              - secretRef:
                  name: my-awesome-bucket <2>
              ports:
              - containerPort: 3000
                protocol: TCP
          ```
          1. Name of the generated configmap from the provisioning process
          1. Name of the generated secret from the provisioning process
          *[Note]* Generated ConfigMap and Secret are same name as the OBC!
          Lastly, expose the pod as a service so you can access the url from a browser. In this example,
          I exposed as a LoadBalancer
          ```
            # kubectl expose pod photo1 --type=LoadBalancer --name=photo1 -n your-namespace
          ```
          To access via a url use the EXTERNAL-IP
          ```
            # kubectl get svc photo1
            NAME                         TYPE           CLUSTER-IP       EXTERNAL-IP                                                               PORT(S)          AGE
            photo1                       LoadBalancer   100.66.124.105   a00c53ccb3c5411e9b6550a7c0e50a2a-2010797808.us-east-1.elb.amazonaws.com   3000:32344/TCP   6d
          ```
          **NOTE:** This is just one example of a Pod that can utilize the bucket information,
          there are several ways that these pod applications can be developed and therefore
          the method of getting the actual values needed from the Secrets and ConfigMaps
          will vary greatly, but the idea remains the same, that the pod consumes the generated
          ConfigMap and Secret created by the provisioner.
        displayName: AWS S3 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - awss3provisioner
        - aws
        - S3
        - provisioner
        - buckets
        links:
        - name: AWS S3 Provisioner
          url: https://github.com/yard-turkey/aws-s3-provisioner
        - name: Provisioner Examples
          url: https://github.com/yard-turkey/aws-s3-provisioner/tree/master/examples
        - name: Operator Documentation
          url: https://github.com/yard-turkey/awss3operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - quay.io/screeley44/aws-s3-provisioner:v1.0.0
        version: 1.0.1
      entries:
      - name: awss3operator.1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: awss3-operator-registry
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Managed OpenShift Black Belt
      provider-url: https://github.com/rh-mobb/ocm-operator
    name: ocm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocm-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "GitLabIdentityProvider",
                "metadata": {
                  "name": "gitlab-sample"
                },
                "spec": {
                  "clientID": "gitlab",
                  "clientSecret": {
                    "name": "gitlab"
                  },
                  "clusterName": "my-cluster",
                  "displayName": "gitlab-sample",
                  "mappingMethod": "claim",
                  "url": "https://gitlab.com"
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "LDAPIdentityProvider",
                "metadata": {
                  "name": "ldap-sample"
                },
                "spec": {
                  "bindDN": "CN=test,OU=Users,DC=example,DC=com",
                  "bindPassword": {
                    "name": "test"
                  },
                  "clusterName": "my-cluster",
                  "displayName": "ldap-sample",
                  "mappingMethod": "claim",
                  "url": "ldap://test.example.com:389"
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "MachinePool",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "aws": {
                    "spotInstances": {
                      "enabled": false
                    }
                  },
                  "clusterName": "my-cluster",
                  "instanceType": "m5.xlarge",
                  "labels": {
                    "owner": "my-user"
                  },
                  "maximumNodesPerZone": 1,
                  "minimumNodesPerZone": 1,
                  "wait": false
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "ROSACluster",
                "metadata": {
                  "name": "rosa-sample"
                },
                "spec": {
                  "accountID": "111111111111",
                  "defaultMachinePool": {
                    "instanceType": "m5.xlarge",
                    "labels": {
                      "owner": "my-owner"
                    },
                    "minimumNodesPerZone": 2
                  },
                  "displayName": "my-cluster",
                  "iam": {
                    "userRole": "arn:aws:iam::111111111111:role/ManagedOpenShift-User-my_user-Role"
                  },
                  "tags": {
                    "owner": "my-owner"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ocm-operator:v0.1.0
          createdAt: "2023-06-02T17:32:25Z"
          operatorframework.io/suggested-namespace: ocm-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ocm-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLabIdentityProvider is the Schema for the gitlabidentityproviders
              API.
            displayName: Git Lab Identity Provider
            kind: GitLabIdentityProvider
            name: gitlabidentityproviders.ocm.mobb.redhat.com
            version: v1alpha1
          - description: LDAPIdentityProvider is the Schema for the ldapidentityproviders
              API.
            displayName: LDAPIdentity Provider
            kind: LDAPIdentityProvider
            name: ldapidentityproviders.ocm.mobb.redhat.com
            version: v1alpha1
          - description: MachinePool is the Schema for the machinepools API.
            displayName: Machine Pool
            kind: MachinePool
            name: machinepools.ocm.mobb.redhat.com
            version: v1alpha1
          - description: ROSACluster is the Schema for the clusters API.
            displayName: ROSACluster
            kind: ROSACluster
            name: rosaclusters.ocm.mobb.redhat.com
            version: v1alpha1
        description: "# Quick Start\n\n## Prerequisites\n\n- [oc](https://docs.openshift.com/container-platform/4.8/cli_reference/openshift_cli/getting-started-cli.html)\n-
          [aws](https://docs.aws.amazon.com/cli/latest/userguide/getting-started-install.html)\n-
          [ROSA Cluster (4.12+)](https://mobb.ninja/docs/rosa/sts/)\n\n> **Warning:**\n>
          This is only tested on ROSA, but may work on other Kubernetes and OpenShift
          variants.  It requires\n> an AWS STS-enabled cluster.  Additionally, OpenShift
          versions 4.12+ are recommended, as they \n> enable [CRD CEL validation](https://kubernetes.io/blog/2022/09/23/crd-validation-rules-beta/).
          \ Versions\n> prior to 4.12 provide a lower version of Kubernetes that does
          not enable this feature.  They may \n> work, but provide no input validation
          when submitting custom resources to the cluster.\n\nBefore installing this
          operator, there are a couple secrets that must be created.\n\n1. OCM Token:
          used to authenticate against the OCM API.  The controller will not start
          without\nthis token as it used for all calls into OCM.\n2. AWS Credentials:
          uses an assume role to a role created in the instructions below in order\nto
          access AWS resources.  This is needed for cluster creation.  See \n[Create
          AWS IAM Policies and Roles](#create-aws-iam-policies-and-roles) for more
          details.\n\n> **Note:**\n> Certain custom resources have their own specific
          prerequisites.  Please use `oc explain` \n> or read the docs [here](https://github.com/rh-mobb/ocm-operator/tree/main/docs)
          for more details.\n\n### Create OCM Token Secret\n\n1. Create a namespace
          where you wish to install the operator:\n\n```bash\noc new-project ocm-operator\n```\n\n2.
          Create a secret containing the OCM_TOKEN.  This token can be obtained form
          \nhttps://console.redhat.com/openshift/token and is used by the operator
          to authenticate \nagainst the OCM API.  This token must exist in the same
          namespace that the operator \nis running and be named `ocm-token`.  It also
          expects the key to be called `OCM_TOKEN` \nas the operator is expecting
          this value as an environment variable.\n\n```bash\noc create secret generic
          ocm-token \\\n  --namespace=ocm-operator \\\n  --from-literal=OCM_TOKEN=${MY_OCM_TOKEN}\n```\n\n###
          Create AWS IAM Policies and Roles\n\nThe operator will need to elevate privileges
          in order to perform things like \ncreating the operator-roles for the clusters.
          \ Because of this, the operator \nmust have a specific role created to allow
          it these permissions.\n\n**NOTE:** please understand what you are doing
          if you deviate from the known good \npolicies.  If errors or more stringent
          security lockdowns are found, please submit a PR \nso that we can get this
          fixed.\n\n1. Set `ACCOUNT_ID` environment variable to define your AWS account
          ID:\n\n```bash\nexport ACCOUNT_ID=111111111111\n```\n\n2. Create the permissions
          boundary.  Because the operator needs to create policies and \nroles, the
          boundary ensures that the operator is not allowed to create additional\npermissions
          outside of the defined boundary.  The sample permission set is located \nat
          `test/aws/boundary.json` in this repository:\n\n```bash\nboundary=$(curl
          -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/boundary.json\")\naws
          iam create-policy \\\n  --policy-name \"OCMOperatorBoundary\" \\\n  --policy-document
          \"$boundary\"\n```\n\n3. Create the policy.  This policy sets what the operator
          is allowed to do.  For any \n`iam` permission, the boundary created above
          is used.  The sample permission set is \nlocated at `test/aws/iam_policy.json`
          in this repository:\n\n```bash\npolicy=$(curl -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/iam_policy.json\")\naws
          iam create-policy \\\n  --policy-name \"OCMOperator\" \\\n  --policy-document
          \"$policy\"\n```\n\n4. Create the role using a trust policy and attach the
          previously created role.  The trust \npolicy is located at `test/aws/trust_policy.json`
          in this repository.  Please note that \nthe trust policy requires an OIDC
          configuration.  The OIDC configuration refers to \n**where the operator
          is running, NOT what it is provisioning**:\n\n```bash\ntrust_policy=$(curl
          -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/trust_policy.json\")\naws
          iam create-role \\\n    --role-name OCMOperator \\\n    --assume-role-policy-document
          \"$trust_policy\"\naws iam attach-role-policy \\\n    --role-name OCMOperator
          \\\n    --policy-arn arn:aws:iam::$ACCOUNT_ID:policy/OCMOperator\n```\n\n5.
          Finally, create the secret containing the assume role credentials.  The
          previous steps allow \nthe operator to assume the role you created in the
          previous step with the permissions created \nvia the previous policies.\n\n```bash\ncat
          <<EOF > /tmp/credentials\n[default]\nrole_arn = arn:aws:iam::$ACCOUNT_ID:role/OCMOperator\nweb_identity_token_file
          = /var/run/secrets/openshift/serviceaccount/token\nEOF\n\noc create secret
          generic aws-credentials \\\n  --namespace=ocm-operator \\\n  --from-file=credentials=/tmp/credentials\n```\n\n##
          Install the Operator\n\n```bash\ncat <<EOF | oc apply -f -\napiVersion:
          operators.coreos.com/v1alpha1\nkind: Subscription\nmetadata:\n  name: ocm-operator\n
          \ namespace: ocm-operator\nspec:\n  channel: v0.1\n  installPlanApproval:
          Automatic\n  name: ocm-operator\n  source: community-operators\n  sourceNamespace:
          ocm-operator\n  startingCSV: ocm-operator.v0.1.0\nEOF\n```\n\n## Provision
          Resources\n\nOnce the operator is available and running, you can provision
          any of the \nresources that it manages.  Also note that documentation is
          always available \nonce the operator is installed as well by using the `oc
          explain` command.  For \nexample, `oc explain rosacluster.spec.clusterName`
          will give you detailed documentation \nabout what the field does.\n\nSee
          the following documentation for details:\n\n* [ROSA Clusters](https://github.com/rh-mobb/ocm-operator/blob/main/docs/clusters.md)\n*
          [Machine Pools](https://github.com/rh-mobb/ocm-operator/blob/main/docs/machinepools.md)\n*
          [Identity Providers](https://github.com/rh-mobb/ocm-operator/blob/main/docs/identityproviders.md)\n"
        displayName: OpenShift Cluster Manager Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ocm
        - rosa
        - redhat
        - machinepool
        - mobb
        links:
        - name: OCM Machine Pool Operator
          url: https://github.com/rh-mobb/ocm-operator
        maintainers:
        - email: dscott@redhat.com
          name: Dustin Scott
        maturity: alpha
        provider:
          name: Managed OpenShift Black Belt
          url: https://github.com/rh-mobb/ocm-operator
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/mobb/ocm-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: ocm-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ocm-operator
    provider:
      name: Managed OpenShift Black Belt
      url: https://github.com/rh-mobb/ocm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SAP LinuxLab - Red Hat
      provider-url: https://github.com/sap-linuxlab
    name: sap-hana-express-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sap-hana-express-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "db.sap-redhat.io/v1alpha1",
                "kind": "HanaExpress",
                "metadata": {
                  "name": "hanaexpress-sample"
                },
                "spec": {
                  "credential": {
                    "key": "hxepasswd.json",
                    "name": "hxepasswd"
                  },
                  "pvcSize": "1Gi"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/redhat-sap-cop/sap-hanaexpress-operator:0.1.4
          createdAt: "2023-07-16T20:53:02Z"
          description: Manage the lifecycle of SAP Hana Express
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HanaExpress is the Schema for the hanaexpresses API
            displayName: Hana Express
            kind: HanaExpress
            name: hanaexpresses.db.sap-redhat.io
            version: v1alpha1
        description: Operator for managing the lifecycle of SAP Hana Express on Kubernetes
          platform
        displayName: SAP Hana Express Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Database
        - SAP
        links:
        - name: Sap Hana Express Operator
          url: https://github.com/redhat-sap/sap-hana-express-operator
        maintainers:
        - email: mjiao@redhat.com
          name: RedHat@SAPLinuxLab
        maturity: alpha
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/redhat-sap-cop/sap-hanaexpress-operator:0.1.4
        version: 1.0.2
      entries:
      - name: sap-hana-express-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: sap-hana-express-operator
    provider:
      name: SAP LinuxLab - Red Hat
      url: https://github.com/sap-linuxlab
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-registry-3
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-3.v3.0.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "simple"
                },
                "spec": {
                  "app": {
                    "ingress": {
                      "host": "simple-app.apps.cluster.example"
                    }
                  },
                  "ui": {
                    "ingress": {
                      "host": "simple-ui.apps.cluster.example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-3-operator:3.0.15
          createdAt: "2025-09-22T13:04:40Z"
          description: Deploy and manage Apicurio Registry 3 on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Apicurio Registry
              version 3.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Apicurio Registry 3

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Allows you manage event schemas and API designs using the Apicurio Registry web console, REST API, Maven plug-in, or Java client
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, MySQL, SQL Server, in-memory
          - Compatible with Confluent APIs
          - Runs on a lightweight Quarkus platform

          ## Apicurio Registry 3 Operator

          Provides a quick and easy way to deploy and manage Apicurio Registry 3 on Kubernetes.

          **Features**
          - Supports installation and configuration of Apicurio Registry 3
          - Perform a rolling upgrade of Apicurio Registry 3

          ## Prerequisites

          This operator does not deploy storage for Apicurio Registry 3. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio Registry 3 and Apicurio Registry 3 Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Apicurio Registry 3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-3-operator:3.0.15
        version: 3.0.15
      entries:
      - name: apicurio-registry-3.v3.0.15
        version: 3.0.15
      - name: apicurio-registry-3.v3.0.14
        version: 3.0.14
      - name: apicurio-registry-3.v3.0.12
        version: 3.0.12
      - name: apicurio-registry-3.v3.0.9
        version: 3.0.9
      - name: apicurio-registry-3.v3.0.8
        version: 3.0.8
      - name: apicurio-registry-3.v3.0.7
        version: 3.0.7
      name: 3.x
    defaultChannel: 3.x
    packageName: apicurio-registry-3
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM Edge
      provider-url: ""
    name: infrastructure-asset-orchestrator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infrastructure-asset-orchestrator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infrastructure.asset.orchestrator.com/v1",
                "kind": "Generatekey",
                "metadata": {
                  "name": "modelbuilder-token"
                },
                "spec": {}
              },
              {
                "apiVersion": "infrastructure.asset.orchestrator.com/v1",
                "kind": "Mb-broker-service",
                "metadata": {
                  "name": "mb-broker-service-sample"
                },
                "spec": {
                  "env_type": "prod",
                  "cloud_type": "IBM",
                  "external_db": {
                    "enabled": false
                  },
                  "postgres": {
                    "enable_pg_tls": true,
                    "storage_class": "",
                    "storage_size": "20G",
                    "backup_type": "incremental",
                    "backup_frequency": "@daily"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/iao-operator@sha256:5cc395f22d86641e175339deb134af6c74de92b81470a65a5c736c8005d645fd
          createdAt: "2021-08-19T12:00:00Z"
          olm.skipRange: <1.0.2
          operators.operatorframework.io/builder: operator-sdk-v1.12.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/infrastructure-asset-orchestrator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Generate key CR
            displayName: GenerateKey
            kind: Generatekey
            name: generatekeys.infrastructure.asset.orchestrator.com
            version: v1
          - description: Deploy MB broker deployment
            displayName: MBBrokerService
            kind: Mb-broker-service
            name: mb-broker-services.infrastructure.asset.orchestrator.com
            version: v1
        description: Infrastructure Asset Orchestrator provides users access to cloud
          resources and services. Leverages the dynamic nature of the cloud platforms
          to provide significant cost savings and efficiencies. Using the Infrastructure
          Asset Orchestrator user can provision, configure, utilize and deprovision
          variety of services on the IBM Cloud Catalog.
        displayName: Infrastructure Asset Orchestrator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cloud-service
        - mbbroker-service
        links:
        - name: Infrastructure Asset Orchestrator
          url: https://infrastructure-asset-orchestrator.domain
        maintainers:
        - email: edge@us.ibm.com
          name: IBM-Edge
        maturity: alpha
        provider:
          name: IBM Edge
        relatedImages:
        - registry.connect.redhat.com/ibm/iao-operator@sha256:5cc395f22d86641e175339deb134af6c74de92b81470a65a5c736c8005d645fd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:609c19d1718947ee814b829334c37e1d720acf5fb642b0712b8ab1b14d57f53f
        - registry.connect.redhat.com/ibm/iao-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/iao-orchestrator-service@sha256:145928929fd0c2b7d9ed5fb374d7cc0e6b48269847e810c5472580a1538ad44c
        - registry.connect.redhat.com/ibm/iao-mb-broker-service@sha256:6531b4196c51246958027453c2d9c7dbcdb4a1f55c826d1da238d256cdbc8a10
        version: 1.0.2
      entries:
      - name: infrastructure-asset-orchestrator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: infrastructure-asset-orchestrator-certified
    provider:
      name: IBM Edge
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datasance
      provider-url: ""
    name: datasance-pot-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datasance-pot-operator.v3.4.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasance.com/v3",
                "kind": "ControlPlane",
                "metadata": {
                  "name": "pot"
                },
                "spec": {
                  "replicas": {
                    "controller": 2
                  },
                  "database": {
                    "provider": "postgres",
                    "user": "user",
                    "host": "postgre.host",
                    "port": 5432,
                    "password": "password",
                    "databaseName": "dbname"
                  },
                  "auth": {
                    "url": "auth.url",
                    "realm": "realm",
                    "realmKey": "realmKey",
                    "ssl": "none",
                    "controllerClient": "controllerClient",
                    "controllerSecret": "controllerClientSecret",
                    "viewerClient": "viewerClient"
                  },
                  "images": {
                    "pullSecret": "pullsecret",
                    "controller": "ghcr.io/datasance/controller@sha256:aa896bae6c3e6eda70938a03aaa649226356d9e4a1d03ba427d4e4ee3090ca6c",
                    "portManager": "ghcr.io/datasance/port-manager@sha256:4bf3793e9e3cd10df9b46c459e983631393825ca788ffbbb3ec3659158ad61fe",
                    "proxy": "ghcr.io/datasance/proxy@sha256:276dbed0c8ef4f5b03fdf29d5c863b6f3efcae50444ce8123220c5673272860f",
                    "router": "ghcr.io/datasance/router@sha256:82657fc8f39d33d8ef0161cf27d705cbdc0a97dca07b22a5fe78fd4213ef0d31"
                  },
                  "services": {
                    "controller": {
                      "type": "LoadBalancer",
                      "annotations": {
                        "service.beta.kubernetes.io/azure-load-balancer-internal": "false"
                      }
                    },
                    "proxy": {
                      "type": "LoadBalancer",
                      "annotations": {
                        "service.beta.kubernetes.io/azure-load-balancer-internal": "false"
                      }
                    },
                    "router": {
                      "type": "LoadBalancer",
                      "annotations": {
                        "service.beta.kubernetes.io/azure-load-balancer-internal": "false"
                      }
                    }
                  },
                  "controller": {
                    "ecnViewerUrl": "pot.demo.datasance.com",
                    "https": true,
                    "secretName": "secretName"
                  },
                  "ingresses": {
                    "controller": {
                      "annotations": {
                        "cert-manager.io/cluster-issuer": "letsencrypt"
                      },
                      "ingressClassName": "nginx",
                      "host": "pot.demo.datasance.com",
                      "secretName": "secretName"
                    },
                    "router": {
                      "address": "pot.demo.datasance.com",
                      "messagePort": 5672,
                      "interiorPort": 55672,
                      "edgePort": 45672
                    },
                    "httpProxy": {
                      "address": "pot.demo.datasance.com"
                    },
                    "tcpProxy": {
                      "address": "pot.demo.datasance.com"
                    }
                  }
                },
                "status": {
                  "conditions": [
                    {
                      "lastTransitionTime": "2022-04-10T22:44:09Z",
                      "message": "",
                      "reason": "initial_status",
                      "status": "True",
                      "type": "deploying"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          containerImage: ghcr.io/datasance/operator@sha256:143b9307c2520ef16b40c4c92e6faee8b24b7fb3a97682231e53236ddd532687
          createdAt: "2024-11-04T20:45:43Z"
          description: Datasance PoT is an Enteprise version of Eclipse ioFog Platform.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Datasance PoT Enterprise Open
            Source Subscriptions based on number of Edge device that ControlPlane
            manages"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/datasance/iofog-operator
          support: Datasance Teknoloji A.S.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Datasance PoT supported ioFog ControlPlane CRD for managing
              container workloads on Edge device
            kind: ControlPlane
            name: controlplanes.datasance.com
            version: v3
        description: ioFog Control Plane on Kubernetes
        displayName: iofog-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iofog
        - datasance
        - pot
        links:
        - name: Datasance PoT // Iofog Operator
          url: https://docs.datasance.com
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Datasance
        relatedImages:
        - ghcr.io/datasance/operator@sha256:143b9307c2520ef16b40c4c92e6faee8b24b7fb3a97682231e53236ddd532687
        - ghcr.io/datasance/controller@sha256:aa896bae6c3e6eda70938a03aaa649226356d9e4a1d03ba427d4e4ee3090ca6c
        - ghcr.io/datasance/router@sha256:82657fc8f39d33d8ef0161cf27d705cbdc0a97dca07b22a5fe78fd4213ef0d31
        - ghcr.io/datasance/proxy@sha256:276dbed0c8ef4f5b03fdf29d5c863b6f3efcae50444ce8123220c5673272860f
        - ghcr.io/datasance/port-manager@sha256:4bf3793e9e3cd10df9b46c459e983631393825ca788ffbbb3ec3659158ad61fe
        version: 3.4.9
      entries:
      - name: datasance-pot-operator.v3.4.9
        version: 3.4.9
      name: stable
    defaultChannel: stable
    packageName: datasance-pot-operator
    provider:
      name: Datasance
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: gcp-filestore-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gcp-filestore-csi-driver-operator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel8-operator@sha256:1ec61d2988b95a10260c7932e9dc6a9d1d5b3eb4148b22afd438aa4ee75dc9b3
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure GCP Filestore CSI driver.
          olm.skipRange: '>=4.11.0-0 <4.13.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/gcp-filestore-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Operator that installs and configures the CSI driver for GCP Filestore. The GCP Filestore CSI Driver Operator is a Technology Preview feature. For more information on Red Hat Technology Preview features support scope, see https://access.redhat.com/support/offerings/techpreview/.
        displayName: GCP Filestore CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - filestore
        links:
        - name: Documentation
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: preview
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:700c0e50f86a70f506403aba75e020a17530ada284af7eee77d4b4ce3dbf4102
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:4f1e873080754ae9aebc361cd77f965e1ade369403346aa0cf4cf502082d18e4
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:ee45e11a6fe8cf60f3ac285cba1b23459ca95068bd9769042ae052c778c3e080
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:537fbc8fe2b212dc798361099d2c6fd981cc846c1cf5454b703ba08d488b61f8
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel8@sha256:a1a4e7b93d6609561a7a2cd732ce772bf0047448e5019298d66c4b985baecfb2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:767682dd3bd65651a8204c9fb732b9c48b99127189992b43abc6ecce027f4589
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel8-operator@sha256:1ec61d2988b95a10260c7932e9dc6a9d1d5b3eb4148b22afd438aa4ee75dc9b3
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:6f1b497772b953468f0b816face8810ace8ea44fa91dd4a3b3e583da2cac350c
        version: 4.13.0-202309181427
      entries:
      - name: gcp-filestore-csi-driver-operator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: preview
    - currentCSV: gcp-filestore-csi-driver-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9-operator@sha256:2dc13de58b391237dc48b9335d88b99f5e26d0e5edc5a664c7d39f16ca9b297d
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure GCP Filestore CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.11.0-0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/gcp-filestore-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Operator that installs and configures the CSI driver for GCP Filestore.
        displayName: GCP Filestore CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - filestore
        links:
        - name: Documentation
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:a5704d26d0b562dba9633d1c107f34e05c750c0e3a64dbcace4dab7b964fff8d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9@sha256:8e801d924071c8f2513555450e688bff60d4100b2ca56d2ba9517038a1d48d4d
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:9c3c3d2b857ce289324aee56df1ba4ab5e5ea1e76cd86ac988233463915834b4
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:78e02d937740d742eb33b08c7317f211c25020c375069d5a97344107716608d3
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:b28def976fc155dff57ff3003cc462a8e9923b325af0b319b65220c66688699e
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0627c09fe3f5a015f117fb30af1f4d17340aefa48b23994737d3901a804b63d3
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9-operator@sha256:2dc13de58b391237dc48b9335d88b99f5e26d0e5edc5a664c7d39f16ca9b297d
        version: 4.16.0-202509240123
      entries:
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202508270905
        version: 4.16.0-202508270905
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202504121405
        version: 4.16.0-202504121405
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: gcp-filestore-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-prometheusservice-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-prometheusservice-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "prometheusservice.services.k8s.aws/v1alpha1",
                "kind": "RuleGroupsNamespace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "prometheusservice.services.k8s.aws/v1alpha1",
                "kind": "Workspace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/prometheusservice-controller:1.3.1
          createdAt: "2025-09-19T21:11:38Z"
          description: AWS Prometheus controller is a service controller for managing
            Prometheus resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertManagerDefinition represents the state of an AWS prometheusservice
              AlertManagerDefinition resource.
            displayName: AlertManagerDefinition
            kind: AlertManagerDefinition
            name: alertmanagerdefinitions.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: LoggingConfiguration represents the state of an AWS prometheusservice
              LoggingConfiguration resource.
            displayName: LoggingConfiguration
            kind: LoggingConfiguration
            name: loggingconfigurations.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: RuleGroupsNamespace represents the state of an AWS prometheusservice
              RuleGroupsNamespace resource.
            displayName: RuleGroupsNamespace
            kind: RuleGroupsNamespace
            name: rulegroupsnamespaces.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: Workspace represents the state of an AWS prometheusservice
              Workspace resource.
            displayName: Workspace
            kind: Workspace
            name: workspaces.prometheusservice.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Prometheus resources in AWS from within your Kubernetes cluster.

          **About Amazon Managed Service for Prometheus**

          Amazon Managed Service for Prometheus is a serverless, Prometheus-compatible monitoring service for container metrics that makes it easier to securely monitor container environments at scale. With Amazon Managed Service for Prometheus, you can use the same open-source Prometheus data model and query language that you use today to monitor the performance of your containerized workloads, and also enjoy improved scalability, availability, and security without having to manage the underlying infrastructure.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Prometheus
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - prometheusservice
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Prometheus Developer Resources
          url: https://aws.amazon.com/prometheus/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: prometheus maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/prometheusservice-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-prometheusservice-controller.v1.3.1
        version: 1.3.1
      - name: ack-prometheusservice-controller.v1.3.0
        version: 1.3.0
      - name: ack-prometheusservice-controller.v1.2.27
        version: 1.2.27
      - name: ack-prometheusservice-controller.v1.2.26
        version: 1.2.26
      - name: ack-prometheusservice-controller.v1.2.24
        version: 1.2.24
      - name: ack-prometheusservice-controller.v1.2.23
        version: 1.2.23
      - name: ack-prometheusservice-controller.v1.2.22
        version: 1.2.22
      - name: ack-prometheusservice-controller.v1.2.20
        version: 1.2.20
      - name: ack-prometheusservice-controller.v1.2.19
        version: 1.2.19
      - name: ack-prometheusservice-controller.v1.2.17
        version: 1.2.17
      - name: ack-prometheusservice-controller.v1.2.16
        version: 1.2.16
      - name: ack-prometheusservice-controller.v1.2.15
        version: 1.2.15
      - name: ack-prometheusservice-controller.v1.2.14
        version: 1.2.14
      - name: ack-prometheusservice-controller.v1.2.13
        version: 1.2.13
      - name: ack-prometheusservice-controller.v1.2.12
        version: 1.2.12
      - name: ack-prometheusservice-controller.v1.2.11
        version: 1.2.11
      - name: ack-prometheusservice-controller.v1.2.10
        version: 1.2.10
      - name: ack-prometheusservice-controller.v1.2.9
        version: 1.2.9
      - name: ack-prometheusservice-controller.v1.2.6
        version: 1.2.6
      - name: ack-prometheusservice-controller.v1.2.5
        version: 1.2.5
      - name: ack-prometheusservice-controller.v1.2.4
        version: 1.2.4
      - name: ack-prometheusservice-controller.v1.2.3
        version: 1.2.3
      - name: ack-prometheusservice-controller.v1.2.2
        version: 1.2.2
      - name: ack-prometheusservice-controller.v1.2.1
        version: 1.2.1
      - name: ack-prometheusservice-controller.v1.2.0
        version: 1.2.0
      - name: ack-prometheusservice-controller.v1.1.0
        version: 1.1.0
      - name: ack-prometheusservice-controller.v1.0.0
        version: 1.0.0
      - name: ack-prometheusservice-controller.v0.1.3
        version: 0.1.3
      - name: ack-prometheusservice-controller.v0.1.2
        version: 0.1.2
      - name: ack-prometheusservice-controller.v0.1.1
        version: 0.1.1
      - name: ack-prometheusservice-controller.v0.1.0
        version: 0.1.0
      - name: ack-prometheusservice-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-prometheusservice-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: networkservicemesh.io
      provider-url: networkservicemesh.io
    name: nsm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nsm-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nsm.networkservicemesh.io/v1alpha1",
                "kind": "NSM",
                "metadata": {
                  "name": "nsm-sample",
                  "namespace": "nsm"
                },
                "spec": {
                  "forwarders": [
                    {
                      "image": "ghcr.io/networkservicemesh/cmd-forwarder-vpp:v1.8.0",
                      "type": "vpp"
                    }
                  ],
                  "nsmLogLevel": "TRACE",
                  "nsmPullPolicy": "IfNotPresent",
                  "registry": {
                    "image": "ghcr.io/networkservicemesh/cmd-registry-memory:v1.8.0",
                    "type": "memory"
                  },
                  "version": "v1.8.0",
                  "webhook": {
                    "image": "ghcr.io/networkservicemesh/admission-webhook-k8s:v1.8.0"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: quay.io/acmenezes/nsm-operator:v1.8.0
          createdAt: "2023-03-02T19:36:46Z"
          operatorframework.io/suggested-namespace: nsm
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/networkservicemesh/nsm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkServiceEndpoint
            name: networkserviceendpoints.networkservicemesh.io
            version: v1
          - kind: NetworkService
            name: networkservices.networkservicemesh.io
            version: v1
          - kind: NSM
            name: nsms.nsm.networkservicemesh.io
            version: v1alpha1
        description: |-
          The Network Service Mesh Operator manages and install network service meshes.

          Check networkservicemesh.io

          Install requirements:

          Before installing the operator it's necessary to have installed [spire](https://github.com/spiffe/spire)

          We have a ready to install yaml file for spire [here](https://github.com/networkservicemesh/nsm-operator/blob/master/config/spire/spire.yaml)
        displayName: nsm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network
        - service mesh
        links:
        - name: Nsm Operator
          url: https://nsm-operator.domain
        maintainers:
        - email: adcmenezes@gmail.com
          name: Alexandre Menezes
        maturity: alpha
        provider:
          name: networkservicemesh.io
          url: networkservicemesh.io
        relatedImages:
        - quay.io/acmenezes/nsm-operator:v1.8.0
        version: 1.8.0
      entries:
      - name: nsm-operator.v1.8.0
        version: 1.8.0
      name: alpha
    defaultChannel: alpha
    packageName: nsm-operator
    provider:
      name: networkservicemesh.io
      url: networkservicemesh.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: observability-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/observability-operator:0.0.28
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      name: candidate
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      - name: observability-operator.v0.0.22
        version: 0.0.22
      name: development
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      name: stable
    defaultChannel: stable
    packageName: observability-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-atlas-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-atlas-kubernetes.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "id": "1",
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "WEEKLY",
                      "id": "2",
                      "retentionUnit": "DAYS",
                      "retentionValue": 6
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": true,
                  "policies": [
                    {
                      "name": "atlas-default-backuppolicy",
                      "namespace": "mongodb-atlas-system"
                    },
                    {
                      "name": "atlas-default-backuppolicy2",
                      "namespace": "mongodb-atlas-system"
                    }
                  ],
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "my-database-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "david"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-cluster"
                },
                "spec": {
                  "deploymentSpec": {
                    "name": "test-cluster",
                    "providerSettings": {
                      "instanceSizeName": "M10",
                      "providerName": "AWS",
                      "regionName": "US_EAST_1"
                    }
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          containerImage: registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator@sha256:3cfbd700e4771c2c7e75c2c41cf6de57c9ef71ecd286e96a21d2567e43424bc3
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasclusters API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB Atlas —
          the only multi-cloud document database service that gives you the versatility you need to build sophisticated and resilient applications that can adapt to changing customer demands and market trends.

          > Current Status: *Beta*. The Operator gives users the ability to provision
          > Atlas projects, clusters and database users using Kubernetes Specifications and bind connection information
          > into applications deployed to Kubernetes.

          ## Quick Start guide
          ### Step 1. Deploy Kubernetes operator by clicking Install button.

          ### Step 2. Create Atlas Cluster

          **1.** Create an Atlas API Key Secret
          In order to work with the Atlas Operator you need to provide [authentication information](https://docs.atlas.mongodb.com/configure-api-access)
           to allow the Atlas Operator to communicate with Atlas API. Once you have generated a Public and Private key in Atlas, you can create a Kuberentes Secret with:
          ```
          kubectl create secret generic mongodb-atlas-operator-api-key \
                   --from-literal="orgId=<the_atlas_organization_id>" \
                   --from-literal="publicApiKey=<the_atlas_api_public_key>" \
                   --from-literal="privateApiKey=<the_atlas_api_private_key>" \
                   -n openshift-operators
          ```
          (Note, that you should use the namespace where the Operator was installed - it's `openshift-operators` by default)

          **2.** Create an `AtlasProject` Custom Resource

          The `AtlasProject` CustomResource represents Atlas Projects in our Kubernetes cluster. You need to specify
          `projectIpAccessList` with the IP addresses or CIDR blocks of any hosts that will connect to the Atlas Cluster.
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasProject
          metadata:
            name: my-project
          spec:
            name: Test Atlas Operator Project
            projectIpAccessList:
              - ipAddress: "192.0.2.15"
                comment: "IP address for Application Server A"
              - ipAddress: "203.0.113.0/24"
                comment: "CIDR block for Application Server B - D"
          ```
          **3.** Create an `AtlasDeployment` Custom Resource.
          The example below is a minimal configuration to create an M10 Atlas cluster in the AWS US East region. For a full list of properties, check
          `atlasclusters.atlas.mongodb.com` [CRD specification](config/crd/bases/atlas.mongodb.com_atlasclusters.yaml)):
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasDeployment
          metadata:
            name: my-atlas-cluster
          spec:
            name: "Test-cluster"
            projectRef:
              name: my-project
            providerSettings:
              instanceSizeName: M10
              providerName: AWS
              regionName: US_EAST_1
          ```

          **4.** Create a database user password Kubernetes Secret
          The Secret must be created in the same namespace as the `AtlasDeployment` and `AtlasProject` were created.
          ```
          kubectl create secret generic the-user-password --from-literal="password=P@@sword%"
          ```

          **5.** Create an `AtlasDatabaseUser` Custom Resource

          In order to connect to an Atlas Cluster the database user needs to be created. `AtlasDatabaseUser` resource should reference
          the password Kubernetes Secret created in the previous step.
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasDatabaseUser
          metadata:
            name: my-database-user
          spec:
            roles:
              - roleName: "readWriteAnyDatabase"
                databaseName: "admin"
            projectRef:
              name: my-project
            username: theuser
            passwordSecretRef:
              name: the-user-password
          ```
          **6.** Wait for the `AtlasDatabaseUser` Custom Resource to be ready

          Wait until the AtlasDatabaseUser resource gets to "ready" status (it will wait until the cluster is created that may take around 10 minutes):
          ```
          kubectl get atlasdatabaseusers my-database-user -o=jsonpath='{.status.conditions[?(@.type=="Ready")].status}'
          True
          ```
          ### Step 3. Connect your application to the Atlas Cluster

          The Atlas Operator will create a Kubernetes Secret with the information necessary to connect to the Atlas Cluster created
          in the previous step. An application in the same Kubernetes Cluster can mount and use the Secret:

          ```
          ...
          containers:
                - name: test-app
                  env:
                   - name: "CONNECTION_STRING"
                     valueFrom:
                       secretKeyRef:
                         name: test-atlas-operator-project-test-cluster-theuser
                         key: connectionStringStandardSrv

          ```
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator@sha256:3cfbd700e4771c2c7e75c2c41cf6de57c9ef71ecd286e96a21d2567e43424bc3
        version: 1.0.0
      entries:
      - name: mongodb-atlas-kubernetes.v1.0.0
        version: 1.0.0
      name: beta
    - currentCSV: mongodb-atlas-kubernetes.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupCompliancePolicy",
                "metadata": {
                  "name": "my-backup-compliance-policy"
                },
                "spec": {
                  "authorizedEmail": "john.doe@example.com",
                  "authorizedUserFirstName": "John",
                  "authorizedUserLastName": "Doe",
                  "copyProtectionEnabled": false,
                  "encryptionAtRestEnabled": false,
                  "onDemandPolicy": {
                    "retentionUnit": "weeks",
                    "retentionValue": 3
                  },
                  "overwriteBackupPolicies": false,
                  "pointInTimeEnabled": true,
                  "restoreWindowDays": 42,
                  "scheduledPolicyItems": [
                    {
                      "frequencyInterval": 2,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "hourly",
                      "retentionUnit": "days",
                      "retentionValue": 2
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "weekly",
                      "retentionUnit": "weeks",
                      "retentionValue": 4
                    },
                    {
                      "frequencyInterval": 15,
                      "frequencyType": "monthly",
                      "retentionUnit": "months",
                      "retentionValue": 12
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": false,
                  "copySettings": [
                    {
                      "cloudProvider": "AWS",
                      "frequencies": [
                        "WEEKLY"
                      ],
                      "regionName": "US_WEST_1",
                      "shouldCopyOplogs": false
                    }
                  ],
                  "policy": {
                    "name": "atlasbackuppolicy-sample"
                  },
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2,
                  "useOrgAndGroupNamesInExportPrefix": true
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasCustomRole",
                "metadata": {
                  "name": "shard-operator-role"
                },
                "spec": {
                  "projectRef": {
                    "name": "my-project"
                  },
                  "role": {
                    "actions": [
                      {
                        "name": "getShardMap",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "shardingState",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "connPoolStats",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "getLog",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      }
                    ],
                    "inheritedRoles": [
                      {
                        "database": "admin",
                        "name": "operator-role-1"
                      }
                    ],
                    "name": "my-role"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDataFederation",
                "metadata": {
                  "name": "my-data-federation",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "name": "my-data-federation",
                  "privateEndpoints": [
                    {
                      "endpointId": "vpce-03f9eeaa764e32454",
                      "provider": "AWS",
                      "type": "DATA_LAKE"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project",
                    "namespace": "mongodb-atlas-system"
                  },
                  "storage": {
                    "databases": [
                      {
                        "collections": [
                          {
                            "dataSources": [
                              {
                                "storeName": "http-test",
                                "urls": [
                                  "https://data.cityofnewyork.us/api/views/vfnx-vebw/rows.csv"
                                ]
                              }
                            ],
                            "name": "test-collection-1"
                          }
                        ],
                        "name": "test-db-1"
                      }
                    ],
                    "stores": [
                      {
                        "name": "http-test",
                        "provider": "http"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "the-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "theuser"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-deployment"
                },
                "spec": {
                  "backupRef": {
                    "name": "atlasbackupschedule-sample"
                  },
                  "deploymentSpec": {
                    "backupEnabled": true,
                    "clusterType": "REPLICASET",
                    "mongoDBMajorVersion": "7.0",
                    "name": "test-deployment",
                    "replicationSpecs": [
                      {
                        "regionConfigs": [
                          {
                            "electableSpecs": {
                              "instanceSize": "M10",
                              "nodeCount": 3
                            },
                            "priority": 7,
                            "providerName": "AWS",
                            "regionName": "US_EAST_1"
                          }
                        ],
                        "zoneName": "Zone 1"
                      }
                    ],
                    "tags": [
                      {
                        "key": "team",
                        "value": "atlas-operator"
                      }
                    ]
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasFederatedAuth",
                "metadata": {
                  "name": "atlas-default-federated-auth",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "connectionSecretRef": {
                    "name": "my-org-secret",
                    "namespace": "mongodb-atlas-system"
                  },
                  "domainAllowList": [
                    "my-org-domain.com"
                  ],
                  "domainRestrictionEnabled": true,
                  "enabled": true,
                  "postAuthRoleGrants": [
                    "ORG_MEMBER"
                  ],
                  "roleMappings": [
                    {
                      "externalGroupName": "org-admin",
                      "roleAssignments": [
                        {
                          "role": "ORG_OWNER"
                        }
                      ]
                    },
                    {
                      "externalGroupName": "dev-team",
                      "roleAssignments": [
                        {
                          "role": "ORG_GROUP_CREATOR"
                        },
                        {
                          "projectName": "dev-project",
                          "role": "GROUP_OWNER"
                        }
                      ]
                    }
                  ],
                  "ssoDebugEnabled": false
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasIPAccessList",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasipaccesslist-sample"
                },
                "spec": {
                  "entries": [
                    {
                      "cidrBlock": "192.168.1.0/24"
                    },
                    {
                      "deleteAfterDate": "2025-01-20T23:59:59+02:00",
                      "ipAddress": "10.0.0.1"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkContainer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasnetworkcontainer-sample"
                },
                "spec": {
                  "cidrBlock": "10.11.0.0/16",
                  "connectionSecret": {
                    "name": "atlas-connection-secret"
                  },
                  "externalProjectRef": {
                    "projectId": "66e2f2b621571b7e69a89b66"
                  },
                  "provider": "AWS",
                  "region": "US_EAST_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkPeering",
                "metadata": {
                  "name": "atlasnetworkpeering-sample"
                },
                "spec": {
                  "atlasCidrBlock": "10.12.0.0/16",
                  "awsConfiguration": {
                    "accepterRegionName": "US_EAST_1",
                    "awsAccountId": "23214235145",
                    "routeTableCidrBlock": "10.11.0.0/16",
                    "vpcId": "vpcid-7123782134"
                  },
                  "containerId": "623412394512350",
                  "containerRegion": "US_EAST_1",
                  "projectRef": {
                    "name": "atlas-project",
                    "namespace": "namespace"
                  },
                  "provider": "AWS"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasPrivateEndpoint",
                "metadata": {
                  "name": "atlasprivateendpoint-sample"
                },
                "spec": {
                  "awsConfiguration": {
                    "id": "vpce-f4k34w51d"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "provider": "AWS",
                  "region": "EU_CENTRAL_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    },
                    {
                      "cidrBlock": "203.0.113.0/24",
                      "comment": "CIDR block for Application Server B - D"
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasSearchIndexConfig",
                "metadata": {
                  "name": "atlassearchindexconfig-sample"
                },
                "spec": {
                  "analyzer": "lucene.standard",
                  "analyzers": [
                    {
                      "name": "test-analyzer",
                      "tokenFilters": [
                        {
                          "asciiFolding": {
                            "originalTokens": "omit"
                          },
                          "type": "asciiFolding"
                        }
                      ],
                      "tokenizer": {
                        "type": "whiteSpace",
                        "whitespace": {
                          "maxTokenLenght": 200
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamConnection",
                "metadata": {
                  "name": "my-streamconnection-kafka"
                },
                "spec": {
                  "kafkaConfig": {
                    "authentication": {
                      "credentials": {
                        "name": "kafka-userpass"
                      },
                      "mechanism": "SCRAM-512"
                    },
                    "bootstrapServers": "kafka.server1:9001,kafka.server2:9002,kafka.server3:9003",
                    "security": {
                      "certificate": {
                        "name": "kafka-cert"
                      },
                      "protocol": "SSL"
                    }
                  },
                  "name": "kafka-config",
                  "type": "Kafka"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamInstance",
                "metadata": {
                  "name": "my-streaminstance-sample"
                },
                "spec": {
                  "clusterConfig": {
                    "provider": "AWS",
                    "region": "VIRGINIA_USA",
                    "tier": "SP30"
                  },
                  "connectionRegistry": [
                    {
                      "name": "my-streamconnection-sample"
                    }
                  ],
                  "name": "my-instance",
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasTeam",
                "metadata": {
                  "name": "atlasteam-sample"
                },
                "spec": {
                  "name": "my-team",
                  "usernames": [
                    "username@mongodb.com"
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasThirdPartyIntegration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasintegration-sample"
                },
                "spec": {
                  "datadog": {
                    "apiKeySecretRef": {
                      "name": "secret-ref"
                    },
                    "region": "string",
                    "sendCollectionLatencyMetrics": "string",
                    "sendDatabaseMetrics": "string"
                  },
                  "projectRef": {
                    "name": "my-atlas-project"
                  },
                  "type": "DATADOG"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:8ac3f6b1c0f31af5f32562ddeed37dd6586ac47e1e7e111947df3ed570c5bed9
          createdAt: "2025-09-25T14:34:34Z"
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/mongodb/mongodb-atlas-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The AtlasBackupCompliancePolicy is a configuration that enforces
              specific backup and retention requirements
            displayName: Atlas Backup Compliance Policy
            kind: AtlasBackupCompliancePolicy
            name: atlasbackupcompliancepolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasCustomRole is the Schema for the AtlasCustomRole API
            displayName: Atlas Custom Role
            kind: AtlasCustomRole
            name: atlascustomroles.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDataFederation is the Schema for the Atlas Data Federation
              API
            displayName: Atlas Data Federation
            kind: AtlasDataFederation
            name: atlasdatafederations.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasdeployments API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasFederatedAuth is the Schema for the Atlasfederatedauth
              API
            displayName: Atlas Federated Auth
            kind: AtlasFederatedAuth
            name: atlasfederatedauths.atlas.mongodb.com
            version: v1
          - description: AtlasIPAccessList is the Schema for the atlasipaccesslists
              API.
            displayName: Atlas IPAccess List
            kind: AtlasIPAccessList
            name: atlasipaccesslists.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkContainer is the Schema for the AtlasNetworkContainer
              API
            displayName: Atlas Network Container
            kind: AtlasNetworkContainer
            name: atlasnetworkcontainers.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkPeering is the Schema for the AtlasNetworkPeering
              API
            displayName: Atlas Network Peering
            kind: AtlasNetworkPeering
            name: atlasnetworkpeerings.atlas.mongodb.com
            version: v1
          - displayName: Atlas Org Settings
            kind: AtlasOrgSettings
            name: atlasorgsettings.atlas.mongodb.com
            version: v1
          - description: "The AtlasPrivateEndpoint custom resource definition (CRD)
              defines a desired [Private Endpoint](https://www.mongodb.com/docs/atlas/security-private-endpoint/#std-label-private-endpoint-overview)
              configuration for an Atlas project. It allows a private connection between
              your cloud provider and Atlas that doesn't send information through
              a public network. \n You can use private endpoints to create a unidirectional
              connection to Atlas clusters from your virtual network."
            displayName: Atlas Private Endpoint
            kind: AtlasPrivateEndpoint
            name: atlasprivateendpoints.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
          - description: Atlas Search Index Config is the Schema for the Atlas Search
              API
            displayName: Atlas Search Index Config
            kind: AtlasSearchIndexConfig
            name: atlassearchindexconfigs.atlas.mongodb.com
            version: v1
          - description: AtlasStreamConnection is the Schema for the atlasstreamconnections
              API
            displayName: Atlas Stream Connection
            kind: AtlasStreamConnection
            name: atlasstreamconnections.atlas.mongodb.com
            version: v1
          - description: AtlasStreamInstance is the Schema for the atlasstreaminstances
              API
            displayName: Atlas Stream Instance
            kind: AtlasStreamInstance
            name: atlasstreaminstances.atlas.mongodb.com
            version: v1
          - description: AtlasTeam is the Schema for the Atlas Teams API
            displayName: Atlas Team
            kind: AtlasTeam
            name: atlasteams.atlas.mongodb.com
            version: v1
          - description: AtlasThirdPartyIntegration is the Schema for the atlas 3rd
              party integrations API.
            displayName: Atlas Third Party Integration
            kind: AtlasThirdPartyIntegration
            name: atlasthirdpartyintegrations.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB
          Atlas — the only multi-cloud document database service that gives you the versatility you need to build sophisticated
          and resilient applications that can adapt to changing customer demands and market trends.

          The full documentation for the Operator can be found [here](https://docs.atlas.mongodb.com/atlas-operator/)

          ## Getting Started

          ### Supported features

          * Create and configure a project, or connect to an existing one.
          * Deploy, manage, scale, and tear down clusters.
          * Support for serverless instances.
          * Create and edit database users.
          * Manage IP Access Lists, network peering and private endpoints.
          * Configure and control Atlas’s fully managed cloud backup.
          * Configure federated authentication for your Atlas organization
          * Integrate Atlas monitoring with Prometheus.

          ... and more.

          To view the list of custom resources and each CR's schema, visit our [reference](https://www.mongodb.com/docs/atlas/operator/stable/custom-resources/)
          documentation. See the [Quickstart](https://www.mongodb.com/docs/atlas/operator/stable/ak8so-quick-start/) to get
          started with Atlas Kubernetes Operator.
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:8ac3f6b1c0f31af5f32562ddeed37dd6586ac47e1e7e111947df3ed570c5bed9
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:9eb6465e723ede43ac47eed6de22042d26c0efb276301a4424a6a3948a8f5b15
        version: 2.11.0
      entries:
      - name: mongodb-atlas-kubernetes.v2.11.0
        version: 2.11.0
      - name: mongodb-atlas-kubernetes.v2.10.0
        version: 2.10.0
      - name: mongodb-atlas-kubernetes.v2.9.1
        version: 2.9.1
      - name: mongodb-atlas-kubernetes.v2.9.0
        version: 2.9.0
      - name: mongodb-atlas-kubernetes.v2.8.2
        version: 2.8.2
      - name: mongodb-atlas-kubernetes.v2.8.1
        version: 2.8.1
      - name: mongodb-atlas-kubernetes.v2.8.0
        version: 2.8.0
      - name: mongodb-atlas-kubernetes.v2.7.1
        version: 2.7.1
      - name: mongodb-atlas-kubernetes.v2.7.0
        version: 2.7.0
      - name: mongodb-atlas-kubernetes.v2.6.1
        version: 2.6.1
      - name: mongodb-atlas-kubernetes.v2.6.0
        version: 2.6.0
      - name: mongodb-atlas-kubernetes.v2.5.0
        version: 2.5.0
      - name: mongodb-atlas-kubernetes.v2.4.1
        version: 2.4.1
      - name: mongodb-atlas-kubernetes.v2.4.0
        version: 2.4.0
      - name: mongodb-atlas-kubernetes.v2.3.1
        version: 2.3.1
      - name: mongodb-atlas-kubernetes.v2.3.0
        version: 2.3.0
      - name: mongodb-atlas-kubernetes.v2.2.2
        version: 2.2.2
      - name: mongodb-atlas-kubernetes.v2.2.1
        version: 2.2.1
      - name: mongodb-atlas-kubernetes.v2.2.0
        version: 2.2.0
      - name: mongodb-atlas-kubernetes.v2.1.0
        version: 2.1.0
      - name: mongodb-atlas-kubernetes.v2.0.1
        version: 2.0.1
      - name: mongodb-atlas-kubernetes.v2.0.0
        version: 2.0.0
      - name: mongodb-atlas-kubernetes.v1.9.1
        version: 1.9.1
      - name: mongodb-atlas-kubernetes.v1.9.0
        version: 1.9.0
      - name: mongodb-atlas-kubernetes.v1.8.2
        version: 1.8.2
      - name: mongodb-atlas-kubernetes.v1.8.1
        version: 1.8.1
      - name: mongodb-atlas-kubernetes.v1.8.0
        version: 1.8.0
      - name: mongodb-atlas-kubernetes.v1.7.3
        version: 1.7.3
      - name: mongodb-atlas-kubernetes.v1.7.2
        version: 1.7.2
      - name: mongodb-atlas-kubernetes.v1.7.1
        version: 1.7.1
      - name: mongodb-atlas-kubernetes.v1.7.0
        version: 1.7.0
      - name: mongodb-atlas-kubernetes.v1.6.1
        version: 1.6.1
      - name: mongodb-atlas-kubernetes.v1.6.0
        version: 1.6.0
      - name: mongodb-atlas-kubernetes.v1.5.1
        version: 1.5.1
      - name: mongodb-atlas-kubernetes.v1.5.0
        version: 1.5.0
      - name: mongodb-atlas-kubernetes.v1.4.1
        version: 1.4.1
      - name: mongodb-atlas-kubernetes.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-atlas-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco AppDynamics
      provider-url: https://www.appdynamics.com
    name: appdynamics-cloud-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appdynamics-cloud-operator.v24.7.0-133
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clustermon",
                "metadata": {
                  "name": "clustermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "clusterName": "test",
                  "image": "appdynamics/appdynamics-cloud-k8s-monitoring:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Containermon",
                "metadata": {
                  "name": "containermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "name": "ContainerMon"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "DbCollector",
                "metadata": {
                  "name": "dbcollector",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-db-collector:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "DbConfig",
                "metadata": {
                  "name": "dbconfig",
                  "namespace": "appdynamics"
                },
                "spec": null
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Inframon",
                "metadata": {
                  "name": "inframon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-k8s-monitoring:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "LogCollector",
                "metadata": {
                  "name": "logcollector-sample",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-log-collector-agent:latest",
                  "resources": {
                    "limits": {
                      "cpu": "400m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "150Mi"
                    }
                  }
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Servermon",
                "metadata": {
                  "name": "servermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "name": "ServerMon"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-08-06T08:33:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clustermon is the Schema for the clustermons API
            displayName: Clustermon
            kind: Clustermon
            name: clustermons.cluster.appdynamics.com
            version: v1alpha1
          - description: Containermon is the Schema for the containermons API
            displayName: Containermon
            kind: Containermon
            name: containermons.cluster.appdynamics.com
            version: v1alpha1
          - kind: DbCollector
            name: dbcollectors.cluster.appdynamics.com
            version: v1alpha1
          - kind: DbConfig
            name: dbconfigs.cluster.appdynamics.com
            version: v1alpha1
          - description: Inframon is the Schema for the inframons API
            displayName: Inframon
            kind: Inframon
            name: inframons.cluster.appdynamics.com
            version: v1alpha1
          - kind: LogCollector
            name: logcollectors.cluster.appdynamics.com
            version: v1alpha1
          - description: Servermon is the Schema for the servermons API
            displayName: Servermon
            kind: Servermon
            name: servermons.cluster.appdynamics.com
            version: v1alpha1
        description: The Appdynamics Cloud Operator is used to install Cluster Collector,
          Container Collector, Server Collector, Log Collector and Cluster Agent
        displayName: Cisco AppDynamics Cloud Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - appdynamics
        links:
        - name: Appdynamics Cloud Operator
          url: https://appdynamics-cloud-operator.domain
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: Cisco AppDynamics
          url: https://www.appdynamics.com
        relatedImages:
        - registry.connect.redhat.com/appdynamics/appdynamics-cloud-operator@sha256:2710d63425942f60af21f060a02eb3c40f5a5f0c952ea6593b3b9769c7ca1014
        version: 24.7.0-133
      entries:
      - name: appdynamics-cloud-operator.v24.7.0-133
        version: 24.7.0-133
      name: alpha
    defaultChannel: alpha
    packageName: appdynamics-cloud-operator
    provider:
      name: Cisco AppDynamics
      url: https://www.appdynamics.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: smb-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smb-csi-driver-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9-operator@sha256:1019812a3969f844a0f3fcd45091fccc7d0d816207e230a2f399556c41ebc705
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure CIFS/SMB CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.15.0-0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/csi-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          CIFS/SMB CSI Driver Operator provides Server Message Block (SMB) CSI Driver that enables you to create and mount CIFS/SMB PersistentVolumes.
        displayName: CIFS/SMB CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - smb
        - cifs
        links:
        - name: Documentation
          url: https://github.com/openshift/csi-operator
        - name: Source Repository
          url: https://github.com/openshift/csi-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: preview
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9-operator@sha256:1019812a3969f844a0f3fcd45091fccc7d0d816207e230a2f399556c41ebc705
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0627c09fe3f5a015f117fb30af1f4d17340aefa48b23994737d3901a804b63d3
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:a5704d26d0b562dba9633d1c107f34e05c750c0e3a64dbcace4dab7b964fff8d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9@sha256:9f516d06c279cf1f82e8948ce2290395eec5f6611abf5ea919e4d6548d81ec10
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:9c3c3d2b857ce289324aee56df1ba4ab5e5ea1e76cd86ac988233463915834b4
        version: 4.16.0-202509240123
      entries:
      - name: smb-csi-driver-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: smb-csi-driver-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: smb-csi-driver-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: smb-csi-driver-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: smb-csi-driver-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: smb-csi-driver-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: smb-csi-driver-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: smb-csi-driver-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: smb-csi-driver-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: smb-csi-driver-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: smb-csi-driver-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: smb-csi-driver-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: smb-csi-driver-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: smb-csi-driver-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: smb-csi-driver-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: smb-csi-driver-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: smb-csi-driver-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: smb-csi-driver-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: smb-csi-driver-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: smb-csi-driver-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: smb-csi-driver-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: smb-csi-driver-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: smb-csi-driver-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: smb-csi-driver-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: smb-csi-driver-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: smb-csi-driver-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: smb-csi-driver-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: smb-csi-driver-operator.v4.16.0-202406251906
        version: 4.16.0-202406251906
      - name: smb-csi-driver-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: smb-csi-driver-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: preview
    defaultChannel: preview
    packageName: smb-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: node-maintenance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-maintenance-operator.v5.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:58846ab62346b1004e11f36bd901c96f8d315958eb39fbb272d8b328faebcf04
          createdAt: "2023-10-20 15:36:51"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=4.10.0 <5.0.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-node-maintenance-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:58846ab62346b1004e11f36bd901c96f8d315958eb39fbb272d8b328faebcf04
        version: 5.0.1
      entries:
      - name: node-maintenance-operator.v5.0.1
        version: 5.0.1
      name: 4.12-eus
    - currentCSV: node-maintenance-operator.v5.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:2efd1976d2fb110391e0e5937b90476d3eeaacb4dd29cd52375f2b8566f17b41
          createdAt: "2023-10-20 15:50:49"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=5.0.0 <5.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:2efd1976d2fb110391e0e5937b90476d3eeaacb4dd29cd52375f2b8566f17b41
        version: 5.2.1
      entries:
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.14-eus
    - currentCSV: node-maintenance-operator.v5.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:7ca5853885534198544d23067087d80047b0e1421903085a2cb328bd00def147
          createdAt: "2024-08-10 19:55:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.2.0 <5.3.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:7ca5853885534198544d23067087d80047b0e1421903085a2cb328bd00def147
        version: 5.3.1
      entries:
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.16-eus
    - currentCSV: node-maintenance-operator.v5.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-10 10:59:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.3.0 <5.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.4/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel9-operator@sha256:752bff8bebcaf3df1f9d126dc4b07b622a44b02bfcf0f82b9100a7693627079a
        version: 5.4.1
      entries:
      - name: node-maintenance-operator.v5.4.1
        version: 5.4.1
      - name: node-maintenance-operator.v5.4.0
        version: 5.4.0
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.18-eus
    - currentCSV: node-maintenance-operator.v5.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-10 10:59:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.3.0 <5.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.4/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel9-operator@sha256:752bff8bebcaf3df1f9d126dc4b07b622a44b02bfcf0f82b9100a7693627079a
        version: 5.4.1
      entries:
      - name: node-maintenance-operator.v5.4.1
        version: 5.4.1
      - name: node-maintenance-operator.v5.4.0
        version: 5.4.0
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: stable
    defaultChannel: stable
    packageName: node-maintenance-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sagemaker-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sagemaker-controller.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "App",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "DataQualityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Domain",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Endpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "EndpointConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "FeatureGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "HyperParameterTuningJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Model",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelBiasJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelExplainabilityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelPackage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelPackageGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelQualityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "MonitoringSchedule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "NotebookInstance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "NotebookInstanceLifecycleConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ProcessingJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "TrainingJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "TransformJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "UserProfile",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sagemaker-controller:1.4.1
          createdAt: "2025-09-22T19:07:36Z"
          description: ACK service controller for Amazon SageMaker manages SageMaker
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: App represents the state of an AWS sagemaker App resource.
            displayName: App
            kind: App
            name: apps.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: DataQualityJobDefinition represents the state of an AWS sagemaker
              DataQualityJobDefinition resource.
            displayName: DataQualityJobDefinition
            kind: DataQualityJobDefinition
            name: dataqualityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Domain represents the state of an AWS sagemaker Domain resource.
            displayName: Domain
            kind: Domain
            name: domains.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: EndpointConfig represents the state of an AWS sagemaker EndpointConfig
              resource.
            displayName: EndpointConfig
            kind: EndpointConfig
            name: endpointconfigs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Endpoint represents the state of an AWS sagemaker Endpoint
              resource.
            displayName: Endpoint
            kind: Endpoint
            name: endpoints.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: FeatureGroup represents the state of an AWS sagemaker FeatureGroup
              resource.
            displayName: FeatureGroup
            kind: FeatureGroup
            name: featuregroups.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: HyperParameterTuningJob represents the state of an AWS sagemaker
              HyperParameterTuningJob resource.
            displayName: HyperParameterTuningJob
            kind: HyperParameterTuningJob
            name: hyperparametertuningjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: InferenceComponent represents the state of an AWS sagemaker
              InferenceComponent resource.
            displayName: InferenceComponent
            kind: InferenceComponent
            name: inferencecomponents.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: LabelingJob represents the state of an AWS sagemaker LabelingJob
              resource.
            displayName: LabelingJob
            kind: LabelingJob
            name: labelingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelBiasJobDefinition represents the state of an AWS sagemaker
              ModelBiasJobDefinition resource.
            displayName: ModelBiasJobDefinition
            kind: ModelBiasJobDefinition
            name: modelbiasjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelExplainabilityJobDefinition represents the state of
              an AWS sagemaker ModelExplainabilityJobDefinition resource.
            displayName: ModelExplainabilityJobDefinition
            kind: ModelExplainabilityJobDefinition
            name: modelexplainabilityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelPackageGroup represents the state of an AWS sagemaker
              ModelPackageGroup resource.
            displayName: ModelPackageGroup
            kind: ModelPackageGroup
            name: modelpackagegroups.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelPackage represents the state of an AWS sagemaker ModelPackage
              resource.
            displayName: ModelPackage
            kind: ModelPackage
            name: modelpackages.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelQualityJobDefinition represents the state of an AWS
              sagemaker ModelQualityJobDefinition resource.
            displayName: ModelQualityJobDefinition
            kind: ModelQualityJobDefinition
            name: modelqualityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Model represents the state of an AWS sagemaker Model resource.
            displayName: Model
            kind: Model
            name: models.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: MonitoringSchedule represents the state of an AWS sagemaker
              MonitoringSchedule resource.
            displayName: MonitoringSchedule
            kind: MonitoringSchedule
            name: monitoringschedules.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: NotebookInstanceLifecycleConfig represents the state of an
              AWS sagemaker NotebookInstanceLifecycleConfig resource.
            displayName: NotebookInstanceLifecycleConfig
            kind: NotebookInstanceLifecycleConfig
            name: notebookinstancelifecycleconfigs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: NotebookInstance represents the state of an AWS sagemaker
              NotebookInstance resource.
            displayName: NotebookInstance
            kind: NotebookInstance
            name: notebookinstances.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: PipelineExecution represents the state of an AWS sagemaker
              PipelineExecution resource.
            displayName: PipelineExecution
            kind: PipelineExecution
            name: pipelineexecutions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Pipeline represents the state of an AWS sagemaker Pipeline
              resource.
            displayName: Pipeline
            kind: Pipeline
            name: pipelines.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ProcessingJob represents the state of an AWS sagemaker ProcessingJob
              resource.
            displayName: ProcessingJob
            kind: ProcessingJob
            name: processingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: TrainingJob represents the state of an AWS sagemaker TrainingJob
              resource.
            displayName: TrainingJob
            kind: TrainingJob
            name: trainingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: TransformJob represents the state of an AWS sagemaker TransformJob
              resource.
            displayName: TransformJob
            kind: TransformJob
            name: transformjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: UserProfile represents the state of an AWS sagemaker UserProfile
              resource.
            displayName: UserProfile
            kind: UserProfile
            name: userprofiles.sagemaker.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon SageMaker resources in AWS from within your Kubernetes cluster.

          **About Amazon SageMaker**

          Amazon SageMaker helps data scientists and developers to prepare, build, train, and deploy high-quality machine learning (ML) models quickly by bringing together a broad set of capabilities purpose-built for ML.

          For more information on Amazon SageMaker, visit the [product page](https://aws.amazon.com/sagemaker/).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SageMaker
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sagemaker
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SageMaker Developer Resources
          url: https://aws.amazon.com/sagemaker/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sagemaker maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sagemaker-controller:1.4.1
        version: 1.4.1
      entries:
      - name: ack-sagemaker-controller.v1.4.1
        version: 1.4.1
      - name: ack-sagemaker-controller.v1.4.0
        version: 1.4.0
      - name: ack-sagemaker-controller.v1.3.3
        version: 1.3.3
      - name: ack-sagemaker-controller.v1.3.2
        version: 1.3.2
      - name: ack-sagemaker-controller.v1.3.1
        version: 1.3.1
      - name: ack-sagemaker-controller.v1.3.0
        version: 1.3.0
      - name: ack-sagemaker-controller.v1.2.20
        version: 1.2.20
      - name: ack-sagemaker-controller.v1.2.19
        version: 1.2.19
      - name: ack-sagemaker-controller.v1.2.18
        version: 1.2.18
      - name: ack-sagemaker-controller.v1.2.17
        version: 1.2.17
      - name: ack-sagemaker-controller.v1.2.16
        version: 1.2.16
      - name: ack-sagemaker-controller.v1.2.13
        version: 1.2.13
      - name: ack-sagemaker-controller.v1.2.12
        version: 1.2.12
      - name: ack-sagemaker-controller.v1.2.11
        version: 1.2.11
      - name: ack-sagemaker-controller.v1.2.9
        version: 1.2.9
      - name: ack-sagemaker-controller.v1.2.8
        version: 1.2.8
      - name: ack-sagemaker-controller.v1.2.5
        version: 1.2.5
      - name: ack-sagemaker-controller.v1.2.4
        version: 1.2.4
      - name: ack-sagemaker-controller.v1.2.3
        version: 1.2.3
      - name: ack-sagemaker-controller.v1.2.2
        version: 1.2.2
      - name: ack-sagemaker-controller.v1.2.1
        version: 1.2.1
      - name: ack-sagemaker-controller.v1.2.0
        version: 1.2.0
      - name: ack-sagemaker-controller.v1.1.0
        version: 1.1.0
      - name: ack-sagemaker-controller.v1.0.0
        version: 1.0.0
      - name: ack-sagemaker-controller.v0.6.0
        version: 0.6.0
      - name: ack-sagemaker-controller.v0.5.1
        version: 0.5.1
      - name: ack-sagemaker-controller.v0.5.0
        version: 0.5.0
      - name: ack-sagemaker-controller.v0.4.5
        version: 0.4.5
      - name: ack-sagemaker-controller.v0.4.4
        version: 0.4.4
      - name: ack-sagemaker-controller.v0.4.3
        version: 0.4.3
      - name: ack-sagemaker-controller.v0.4.2
        version: 0.4.2
      - name: ack-sagemaker-controller.v0.4.0
        version: 0.4.0
      - name: ack-sagemaker-controller.v0.3.4
        version: 0.3.4
      - name: ack-sagemaker-controller.v0.3.3
        version: 0.3.3
      - name: ack-sagemaker-controller.v0.3.2
        version: 0.3.2
      - name: ack-sagemaker-controller.v0.3.1
        version: 0.3.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-sagemaker-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudwatch-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudwatch-controller.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudwatch-controller:1.3.0
          createdAt: "2025-10-02T23:05:17Z"
          description: AWS CloudWatch controller is a service controller for managing
            CloudWatch resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Dashboard represents the state of an AWS cloudwatch Dashboard
              resource.
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.cloudwatch.services.k8s.aws
            version: v1alpha1
          - description: MetricAlarm represents the state of an AWS cloudwatch MetricAlarm
              resource.
            displayName: MetricAlarm
            kind: MetricAlarm
            name: metricalarms.cloudwatch.services.k8s.aws
            version: v1alpha1
          - description: MetricStream represents the state of an AWS cloudwatch MetricStream
              resource.
            displayName: MetricStream
            kind: MetricStream
            name: metricstreams.cloudwatch.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudWatch resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudWatch**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudWatch
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudwatch
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudWatch Developer Resources
          url: https://aws.amazon.com/CloudWatch/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudwatch maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudwatch-controller:1.3.0
        version: 1.3.0
      entries:
      - name: ack-cloudwatch-controller.v1.3.0
        version: 1.3.0
      - name: ack-cloudwatch-controller.v1.2.1
        version: 1.2.1
      - name: ack-cloudwatch-controller.v1.2.0
        version: 1.2.0
      - name: ack-cloudwatch-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudwatch-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudwatch-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudwatch-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudwatch-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudwatch-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudwatch-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudwatch-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudwatch-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudwatch-controller.v1.0.5
        version: 1.0.5
      - name: ack-cloudwatch-controller.v1.0.4
        version: 1.0.4
      - name: ack-cloudwatch-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudwatch-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudwatch-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudwatch-controller.v0.0.12
        version: 0.0.12
      - name: ack-cloudwatch-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudwatch-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudwatch-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudwatch-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudwatch-controller.v0.0.3
        version: 0.0.3
      - name: ack-cloudwatch-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudwatch-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OpenShift Database Access
      provider-url: ""
    name: dbaas-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dbaas-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSConnection",
                "metadata": {
                  "name": "dbaasconnection-example"
                },
                "spec": {
                  "databaseServiceID": "instance-example",
                  "databaseServiceType": "instance",
                  "inventoryRef": {
                    "name": "inventory-example",
                    "namespace": "openshift-dbaas-operator"
                  }
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSInstance",
                "metadata": {
                  "name": "dbaasinstance-sample"
                },
                "spec": {
                  "cloudProvider": "AWS",
                  "cloudRegion": "US_EAST_1",
                  "inventoryRef": {
                    "name": "inventory-example",
                    "namespace": "openshift-dbaas-operator"
                  },
                  "name": "my-cluster-free"
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSInventory",
                "metadata": {
                  "name": "inventory-example"
                },
                "spec": {
                  "credentialsRef": {
                    "name": "dbaas-vendor-credentials"
                  },
                  "providerRef": {
                    "name": "provider-example-registration"
                  }
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSPolicy",
                "metadata": {
                  "name": "dbaaspolicy-sample"
                },
                "spec": {
                  "disableProvisions": false
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSProvider",
                "metadata": {
                  "labels": {
                    "related-to": "dbaas-operator",
                    "type": "dbaas-provider-registration"
                  },
                  "name": "provider-example-registration"
                },
                "spec": {
                  "allowsFreeTrial": true,
                  "connectionKind": "ProviderConnection",
                  "credentialFields": [
                    {
                      "displayName": "ProviderAPIKeyD",
                      "helpText": "This is the API Key for the example provider to show on UI",
                      "key": "APIKey",
                      "required": true,
                      "type": "maskedstring"
                    }
                  ],
                  "externalProvisionDescription": "Follow the guide to start a free Provider Serverless (beta) cluster",
                  "externalProvisionURL": "https://www.exmample.com/docs/provider/quickstart.html",
                  "groupVersion": "dbaas.redhat.com/v1beta1",
                  "instanceKind": "ProviderInstance",
                  "inventoryKind": "ProviderInventory",
                  "provider": {
                    "displayDescription": "This is an example for providers on how to implement their operator to integrate with DBaaS.",
                    "displayName": "DBaaS Provider Example ",
                    "icon": {
                      "base64data": "SGVsbG8sIHdvcmxkLg==",
                      "mediatype": "image/png"
                    },
                    "name": "Provider Example"
                  },
                  "provisioningParameters": {
                    "cloudProvider": {
                      "conditionalData": [
                        {
                          "defaultValue": "GCP",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "FREETRIAL"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        },
                        {
                          "defaultValue": "AWS",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "SERVERLESS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Amazon Web Services",
                              "value": "AWS"
                            },
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        },
                        {
                          "defaultValue": "AWS",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Amazon Web Services",
                              "value": "AWS"
                            },
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        }
                      ],
                      "displayName": "Cloud Provider"
                    },
                    "dedicatedLocationLabel": {
                      "displayName": "Select regions \u0026 nodes",
                      "helpText": "Select the geographical region where you want the database instance to run, and set the number of nodes you want running in this dedicated cluster."
                    },
                    "hardwareLabel": {
                      "displayName": "Hardware per node",
                      "helpText": "Select the compute and storage requirements for this database instance."
                    },
                    "machineType": {
                      "conditionalData": [
                        {
                          "defaultValue": "m5.large",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "AWS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "2 vCPU, 8 GiB RAM",
                              "value": "m5.large"
                            },
                            {
                              "displayValue": "4 vCPU, 16 GiB RAM",
                              "value": "m5.xlarge"
                            },
                            {
                              "displayValue": "8 vCPU, 32 GiB RAM",
                              "value": "m5.2xlarge"
                            },
                            {
                              "displayValue": "16 vCPU, 64 GiB RAM",
                              "value": "m5.4xlarge"
                            },
                            {
                              "displayValue": "32 vCPU, 128 GiB RAM",
                              "value": "m5.8xlarge"
                            }
                          ]
                        },
                        {
                          "defaultValue": "n1-standard-2",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "GCP"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "2 vCPU, 7.5 GiB RAM",
                              "value": "n1-standard-2"
                            },
                            {
                              "displayValue": "4 vCPU, 15 GiB RAM",
                              "value": "n1-standard-4"
                            },
                            {
                              "displayValue": "8 vCPU, 30 GiB RAM",
                              "value": "n1-standard-8"
                            },
                            {
                              "displayValue": "16 vCPU, 60 GiB RAM",
                              "value": "n1-standard-16"
                            },
                            {
                              "displayValue": "32 vCPU, 120 GiB RAM",
                              "value": "n1-standard-32"
                            }
                          ]
                        }
                      ],
                      "displayName": "Compute"
                    },
                    "name": {
                      "displayName": "Cluster name"
                    },
                    "nodes": {
                      "displayName": "nodes"
                    },
                    "plan": {
                      "conditionalData": [
                        {
                          "defaultValue": "SERVERLESS",
                          "options": [
                            {
                              "displayValue": "Free trial",
                              "value": "FREETRIAL"
                            },
                            {
                              "displayValue": "Serverless",
                              "value": "SERVERLESS"
                            },
                            {
                              "displayValue": "Dedicated",
                              "value": "DEDICATED"
                            }
                          ]
                        }
                      ],
                      "displayName": "Hosting plan"
                    },
                    "planLabel": {
                      "displayName": "Select a plan"
                    },
                    "regions": {
                      "displayName": "us-west-2"
                    },
                    "serverlessLocationLabel": {
                      "displayName": "Select regions",
                      "helpText": "Select the geographical region where you want the database instance to run."
                    },
                    "spendLimit": {
                      "conditionalData": [
                        {
                          "defaultValue": "0",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "SERVERLESS"
                            }
                          ]
                        }
                      ],
                      "displayName": "Spend limit"
                    },
                    "spendLimitLabel": {
                      "displayName": "Spend limit",
                      "helpText": "Set a spending limit on resources for this database instance."
                    },
                    "storageGib": {
                      "conditionalData": [
                        {
                          "defaultValue": "15",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "AWS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "15 GiB",
                              "value": "15"
                            },
                            {
                              "displayValue": "35 GiB",
                              "value": "35"
                            },
                            {
                              "displayValue": "75 GiB",
                              "value": "75"
                            },
                            {
                              "displayValue": "150 GiB",
                              "value": "150"
                            },
                            {
                              "displayValue": "300 GiB",
                              "value": "300"
                            },
                            {
                              "displayValue": "600 GiB",
                              "value": "600"
                            }
                          ]
                        },
                        {
                          "defaultValue": "15",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "GCP"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "15 GiB",
                              "value": "15"
                            },
                            {
                              "displayValue": "35 GiB",
                              "value": "35"
                            },
                            {
                              "displayValue": "75 GiB",
                              "value": "75"
                            },
                            {
                              "displayValue": "150 GiB",
                              "value": "150"
                            },
                            {
                              "displayValue": "300 GiB",
                              "value": "300"
                            },
                            {
                              "displayValue": "600 GiB",
                              "value": "600"
                            }
                          ]
                        }
                      ],
                      "displayName": "Storage"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: quay.io/ecosystem-appeng/dbaas-operator:v0.5.0
          description: Discover & connect with database instances hosted on 3rd-party
            ISV cloud platforms.
          operatorframework.io/suggested-namespace: openshift-dbaas-operator
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/internal-objects: '["dbaasconnections.dbaas.redhat.com","dbaasproviders.dbaas.redhat.com","dbaasplatforms.dbaas.redhat.com","dbaasinstances.dbaas.redhat.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/RHEcosystemAppEng/dbaas-operator
          support: rhoda-requests@redhat.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DBaaSConnection
            name: dbaasconnections.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSConnection defines the schema for the DBaaSConnection
              API.
            displayName: DBaaSConnection
            kind: DBaaSConnection
            name: dbaasconnections.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSInstance
            name: dbaasinstances.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSInstance defines the schema for the DBaaSInstance API.
            displayName: DBaaSInstance
            kind: DBaaSInstance
            name: dbaasinstances.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSInventory
            name: dbaasinventories.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSInventory defines the schema for the DBaaSInventory
              API. Inventory objects must be created in a valid namespace, determined
              by the existence of a DBaaSPolicy object.
            displayName: Provider Account
            kind: DBaaSInventory
            name: dbaasinventories.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSPlatform
            name: dbaasplatforms.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSPlatform defines the schema for the DBaaSPlatform API.
            displayName: DBaaSPlatform
            kind: DBaaSPlatform
            name: dbaasplatforms.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSPolicy
            name: dbaaspolicies.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSPolicy enables administrative capabilities within a
              namespace, and sets a default inventory policy. Policy defaults can
              be overridden on a per-inventory basis.
            displayName: Provider Account Policy
            kind: DBaaSPolicy
            name: dbaaspolicies.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSProvider
            name: dbaasproviders.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSProvider defines the schema for the DBaaSProvider API.
            displayName: DBaaSProvider
            kind: DBaaSProvider
            name: dbaasproviders.dbaas.redhat.com
            version: v1beta1
        description: |
          The OpenShift Database Access Operator enables OpenShift users to discover & connect with database instances
          hosted on 3rd-party ISV cloud platforms such as CrunchyData Bridge & CockroachCloud.

          ## Feedback & Support

          Please contact us at [rhoda-requests@redhat.com](mailto:rhoda-requests@redhat.com)

          For more information on how to use our service, please refer to our [Quick Starts](quickstart?keyword=database).
        displayName: OpenShift Database Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rhoda
        - oda
        - dbaas
        - database
        - db
        - data
        - access
        - crunchy
        - bridge
        - cockroach
        - cockroachdb
        - cloud
        - ack
        - amazon
        - rds
        - relational
        links:
        - name: GitHub
          url: https://github.com/RHEcosystemAppEng/dbaas-operator
        maintainers:
        - email: ecosystem-appeng@redhat.com
          name: Red Hat Ecosystem App Engineering
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: OpenShift Database Access
        relatedImages:
        - registry.developers.crunchydata.com/crunchydata/crunchy-bridge-operator-catalog:v0.0.6
        - gcr.io/cockroach-shared/ccapi-k8s-operator-catalog:v0.0.5
        - quay.io/ecosystem-appeng/dbaas-operator:v0.5.0
        - quay.io/openshift/origin-kube-rbac-proxy:4.10.0
        - quay.io/ecosystem-appeng/rds-dbaas-operator-catalog:v0.4.0
        - quay.io/ecosystem-appeng/dbaas-dynamic-plugin:0.5.0
        version: 0.5.0
      entries:
      - name: dbaas-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: dbaas-operator
    provider:
      name: OpenShift Database Access
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ControlPlane Ltd.
      provider-url: https://control-plane.io
    name: flux-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux-operator.v0.31.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "FluxInstance",
                "metadata": {
                  "name": "flux",
                  "namespace": "flux-system",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcileEvery": "1h",
                    "fluxcd.controlplane.io/reconcileTimeout": "5m"
                  }
                },
                "spec": {
                  "distribution": {
                    "version": "2.x",
                    "registry": "ghcr.io/fluxcd"
                  },
                  "components": [
                    "source-controller",
                    "kustomize-controller",
                    "helm-controller",
                    "notification-controller",
                    "image-reflector-controller",
                    "image-automation-controller"
                  ],
                  "cluster": {
                    "type": "openshift",
                    "size": "medium",
                    "domain": "cluster.local",
                    "networkPolicy": true,
                    "multitenant": true,
                    "tenantDefaultServiceAccount": "flux"
                  }
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "FluxReport",
                "metadata": {
                  "name": "flux",
                  "namespace": "flux-system"
                },
                "spec": {
                  "distribution": {
                    "entitlement": "Issued by controlplane",
                    "managedBy": "flux-operator",
                    "status": "Installed",
                    "version": "v2.3.0"
                  },
                "components": [
                    {
                      "image": "ghcr.io/fluxcd/kustomize-controller:v1.3.0@sha256:48a032574dd45c39750ba0f1488e6f1ae36756a38f40976a6b7a588d83acefc1",
                      "name": "kustomize-controller",
                      "ready": true,
                      "status": "Current Deployment is available. Replicas: 1"
                    },
                    {
                      "image": "ghcr.io/fluxcd/source-controller:v1.3.0@sha256:161da425b16b64dda4b3cec2ba0f8d7442973aba29bb446db3b340626181a0bc",
                      "name": "source-controller",
                      "ready": true,
                      "status": "Current Deployment is available. Replicas: 1"
                    }
                  ],
                "reconcilers": [
                    {
                      "apiVersion": "kustomize.toolkit.fluxcd.io/v1",
                      "kind": "Kustomization",
                      "stats": {
                        "failing": 0,
                        "running": 1,
                        "suspended": 0
                      }
                    },
                    {
                      "apiVersion": "source.toolkit.fluxcd.io/v1",
                      "kind": "GitRepository",
                      "stats": {
                        "failing": 0,
                        "running": 1,
                        "suspended": 0,
                        "totalSize": "3.7 MiB"
                      }
                    }
                  ],
                  "sync": {
                    "ready": true,
                    "id": "kustomization/flux-system",
                    "path": "clusters/production",
                    "source": "https://github.com/my-org/my-fleet.git",
                    "status": "Applied revision: refs/heads/main@sha1:a90cd1ac35de01c175f7199315d3f4cd60195911"
                  }
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "ResourceSet",
                "metadata": {
                  "name": "podinfo",
                  "namespace": "default",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcile": "enabled",
                    "fluxcd.controlplane.io/reconcileEvery": "30m",
                    "fluxcd.controlplane.io/reconcileTimeout": "5m"
                  }
                },
                "spec": {
                  "commonMetadata": {
                    "labels": {
                      "app.kubernetes.io/name": "podinfo"
                    }
                  },
                  "inputs": [
                    {
                      "tenant": "team1",
                      "version": "6.7.x",
                      "replicas": "2"
                    },
                    {
                      "tenant": "team2",
                      "version": "6.6.x",
                      "replicas": "3"
                    }
                  ],
                  "resources": [
                    {
                      "apiVersion": "source.toolkit.fluxcd.io/v1beta2",
                      "kind": "OCIRepository",
                      "metadata": {
                        "name": "podinfo-<< inputs.tenant >>",
                        "namespace": "default"
                      },
                      "spec": {
                        "interval": "10m",
                        "url": "oci://ghcr.io/stefanprodan/charts/podinfo",
                        "ref": {
                          "semver": "<< inputs.version | quote >>"
                        }
                      }
                    },
                    {
                      "apiVersion": "helm.toolkit.fluxcd.io/v2",
                      "kind": "HelmRelease",
                      "metadata": {
                        "name": "podinfo-<< inputs.tenant >>",
                        "namespace": "default"
                      },
                      "spec": {
                        "interval": "1h",
                        "releaseName": "podinfo-<< inputs.tenant >>",
                        "chartRef": {
                          "kind": "OCIRepository",
                          "name": "podinfo-<< inputs.tenant >>"
                        },
                        "values": {
                          "replicaCount": "<< inputs.replicas | int >>"
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "ResourceSetInputProvider",
                "metadata": {
                  "name": "demo",
                  "namespace": "default",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcile": "enabled",
                    "fluxcd.controlplane.io/reconcileEvery": "10m",
                    "fluxcd.controlplane.io/reconcileTimeout": "1m"
                  }
                },
                "spec": {
                  "type": "GitHubPullRequest",
                  "url": "https://github.com/fluxcd-testing/pr-testing",
                  "filter": {
                    "limit": 10,
                    "includeBranch": "^stefanprodan-patch-.*$",
                    "labels": [
                      "enhancement"
                    ]
                  },
                  "skip": {
                    "labels": [
                        "deploy/flux-preview-pause",
                        "!test-build-push/passed"
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/controlplaneio-fluxcd/flux-operator:v0.31.0-ubi
          createdAt: "2025-10-06T20:33:03Z"
          description: The Flux Operator manages the lifecycle of the CNCF Flux project.
          operatorframework.io/suggested-namespace: flux-system
          repository: https://github.com/controlplaneio-fluxcd/flux-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Flux Instance
            displayName: FluxInstance
            kind: FluxInstance
            name: fluxinstances.fluxcd.controlplane.io
            version: v1
          - description: Flux Report (Autogenerated)
            displayName: FluxReport
            kind: FluxReport
            name: fluxreports.fluxcd.controlplane.io
            version: v1
          - description: Resource Set
            displayName: ResourceSet
            kind: ResourceSet
            name: resourcesets.fluxcd.controlplane.io
            version: v1
          - description: ResourceSet Input Provider
            displayName: ResourceSetInputProvider
            kind: ResourceSetInputProvider
            name: resourcesetinputproviders.fluxcd.controlplane.io
            version: v1
        description: |-
          [Flux](https://fluxcd.io) is a leading GitOps Continuous Delivery tool used to streamline and automate application deployments on Kubernetes. Flux enables teams to achieve a reliable and auditable delivery process while promoting collaboration and traceability across different environments.
          Flux is powered by the GitOps Toolkit, a set of composable APIs and specialized tools that enable a wide range of continuous delivery use-cases, from simple Kubernetes deployment pipelines to multi-tenant and multi-cluster progressive delivery rollouts.
          For more information about the CNCF-graduated Flux project, please see the [Flux architecture overview](https://fluxcd.control-plane.io/guides/flux-architecture/).
          ### Operator Capabilities
          The [Flux Operator](https://github.com/controlplaneio-fluxcd/flux-operator) provides a declarative API for the installation, configuration and upgrade of CNCF Flux and the ControlPlane [enterprise distribution](https://fluxcd.control-plane.io/).
          The operator allows the configuration of Flux multi-tenancy lockdown, network policies, persistent storage, sharding, vertical scaling, custom patches, and the synchronization of the cluster state from Git repositories, OCI artifacts and S3-compatible storage.
          The Flux Operator supervises the Flux controllers and provides a unified view of all the Flux resources that define the GitOps workflows for the target cluster. The operator generates reports, emits events, and exports Prometheus metrics to help with monitoring and troubleshooting Flux.
          ### Operator Features
          The operator extends Flux with self-service capabilities, deployment windows, and preview environments for GitHub, GitLab and Azure DevOps pull requests testing.
          The operator ResourceSet API enables platform teams to define their own application standard as a group of Flux and Kubernetes resources that can be templated, parameterized and deployed as a single unit on self-service environments. For more information, please see the [ResourceSet documentation](https://fluxcd.control-plane.io/operator/resourcesets/introduction/).
          ### OpenShift Support
          The Flux Operator should be installed in a dedicated namespace, e.g. `flux-system`.
          To deploy Flux on OpenShift clusters, create a `FluxInstance` custom resource named `flux` in the same namespace as the operator and set the `.spec.cluster.type` field to `openshift`.
          For more information on how to configure the Flux instance, please see the [Flux Operator documentation](https://fluxcd.control-plane.io/operator/flux-config/).
        displayName: Flux Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - fluxcd
        - gitops
        - continuous delivery
        links:
        - name: Source Code
          url: https://github.com/controlplaneio-fluxcd/flux-operator
        - name: Documentation
          url: https://fluxcd.control-plane.io/operator/
        - name: Enterprise Support
          url: https://fluxcd.control-plane.io/pricing/
        maintainers:
        - email: flux-enterprise@control-plane.io
          name: ControlPlane Flux Team
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: ControlPlane Ltd.
          url: https://control-plane.io
        relatedImages:
        - ghcr.io/controlplaneio-fluxcd/flux-operator:v0.31.0-ubi
        version: 0.31.0
      entries:
      - name: flux-operator.v0.31.0
        version: 0.31.0
      - name: flux-operator.v0.30.0
        version: 0.30.0
      - name: flux-operator.v0.28.0
        version: 0.28.0
      - name: flux-operator.v0.27.0
        version: 0.27.0
      - name: flux-operator.v0.26.0
        version: 0.26.0
      - name: flux-operator.v0.24.1
        version: 0.24.1
      - name: flux-operator.v0.24.0
        version: 0.24.0
      - name: flux-operator.v0.23.0
        version: 0.23.0
      - name: flux-operator.v0.22.0
        version: 0.22.0
      - name: flux-operator.v0.21.0
        version: 0.21.0
      - name: flux-operator.v0.20.0
        version: 0.20.0
      - name: flux-operator.v0.19.0
        version: 0.19.0
      - name: flux-operator.v0.18.0
        version: 0.18.0
      name: stable
    defaultChannel: stable
    packageName: flux-operator
    provider:
      name: ControlPlane Ltd.
      url: https://control-plane.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-load-core
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-load-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Loadcore",
                "metadata": {
                  "name": "loadcore-sample"
                },
                "spec": {
                  "accessRoute": {
                    "enabled": true,
                    "host": "loadcore.apps.cluster.perf.keysight.lab"
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "reportTemplates": {
                    "enabled": false
                  },
                  "wirelessDataModel": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:42:44Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Loadcore
            name: loadcores.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight LoadCore Operator
        displayName: Keysight LoadCore Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Load Core
          url: https://keysight-load-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-load-core@sha256:245d9be60ff6e9a6e5e42a026a4ee7b8168f391a0491de23ae827d1906af88e5
        version: 0.1.21
      entries:
      - name: keysight-load-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-load-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CloudBees
      provider-url: https://www.cloudbees.com/products/continuous-integration
    name: cloudbees-ci-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudbees-ci.v3.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.cloudbees.com/v1alpha1",
                "kind": "CloudBeesCI",
                "metadata": {
                  "name": "cloudbeesci-sample"
                },
                "spec": {
                  "Agents": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-agent:2.289.3.2"
                    },
                    "ImagePullSecrets": null,
                    "SeparateNamespace": {
                      "Create": false,
                      "Enabled": false,
                      "Name": null
                    }
                  },
                  "Hibernation": {
                    "Enabled": false,
                    "Image": {
                      "dockerImage": "cloudbees/managed-master-hibernation-monitor:247.c5dfce00a179",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "NodeSelector": {},
                    "Tolerations": []
                  },
                  "Master": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-mm:2.289.3.2"
                    },
                    "JavaOpts": null,
                    "OperationsCenterNamespace": null
                  },
                  "NetworkPolicy": {
                    "Enabled": false,
                    "JMXSelectors": [],
                    "ingressControllerSelector": []
                  },
                  "OperationsCenter": {
                    "AgentListenerPort": 50000,
                    "Annotations": {},
                    "CSRF": {
                      "ProxyCompatibility": false
                    },
                    "CasC": {
                      "ConfigMapName": "oc-casc-bundle",
                      "Enabled": false
                    },
                    "ContainerPort": 8080,
                    "ContextPath": null,
                    "Enabled": true,
                    "ExtraConfigMaps": [],
                    "ExtraContainers": [],
                    "ExtraGroovyConfiguration": {},
                    "ExtraVolumeMounts": [],
                    "ExtraVolumes": [],
                    "HealthProbeLivenessFailureThreshold": 12,
                    "HealthProbes": true,
                    "HostName": null,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-cloud-core-oc:2.289.3.2",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "Ingress": {
                      "Annotations": {
                        "kubernetes.io/tls-acme": "false"
                      },
                      "Class": "nginx",
                      "tls": {
                        "Enable": false,
                        "SecretName": null
                      }
                    },
                    "JavaOpts": null,
                    "JenkinsOpts": null,
                    "LoadBalancerIP": null,
                    "LoadBalancerSourceRanges": [
                      "0.0.0.0/0"
                    ],
                    "Name": "cjoc",
                    "NodeSelector": {},
                    "Platform": "standard",
                    "Protocol": "http",
                    "Resources": {
                      "Limits": {
                        "Cpu": 1,
                        "Memory": "2G"
                      },
                      "Requests": {
                        "Cpu": 1,
                        "Memory": "2G"
                      }
                    },
                    "Route": {
                      "tls": {
                        "CACertificate": null,
                        "Certificate": null,
                        "DestinationCACertificate": null,
                        "Enable": false,
                        "InsecureEdgeTerminationPolicy": "Redirect",
                        "Key": null,
                        "Termination": "edge"
                      }
                    },
                    "ServiceAgentListenerPort": 50000,
                    "ServiceAnnotations": {},
                    "ServicePort": 80,
                    "ServiceType": "ClusterIP",
                    "Tolerations": []
                  },
                  "Persistence": {
                    "AccessMode": "ReadWriteOnce",
                    "Annotations": {},
                    "Size": "20Gi",
                    "StorageClass": null
                  },
                  "PodSecurityPolicy": {
                    "Annotations": {},
                    "Enabled": false
                  },
                  "Subdomain": false,
                  "ingress-nginx": {
                    "Enabled": false,
                    "controller": {
                      "admissionWebhooks": {
                        "patch": {
                          "nodeSelector": {
                            "kubernetes.io/os": "linux"
                          }
                        }
                      },
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "nginx-ingress": {
                    "Enabled": false,
                    "controller": {
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "rbac": {
                    "agentsServiceAccountName": "jenkins-agents",
                    "hibernationMonitorServiceAccountName": "managed-master-hibernation-monitor",
                    "install": true,
                    "masterServiceAccountName": "jenkins",
                    "serviceAccountName": "cjoc"
                  },
                  "sidecarinjector": {
                    "Enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cloudbees-ci-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cloudbees-ci-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudBeesCI
            name: cloudbeescis.charts.cloudbees.com
            version: v1alpha1
        description: CloudBees CI
        displayName: CloudBees CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudbees
        - ci
        - enterprise
        - continuous
        - integration
        - jenkins
        - automation
        - pipeline
        - devops
        links:
        - name: CloudBees CI
          url: https://www.cloudbees.com/products/continuous-integration
        maintainers:
        - email: support@cloudbees.com
          name: Support
        - email: sales@cloudbees.com
          name: Sales
        maturity: alpha
        provider:
          name: CloudBees
          url: https://www.cloudbees.com/products/continuous-integration
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.marketplace.redhat.com/rhm/cloudbees/cloudbees-ci-operator@sha256:bd0bd72e9e7807b0fa307de8a93a56967221fe2dd181bac02383f0f9be9120e4
        version: 3.34.1
      entries:
      - name: cloudbees-ci.v3.34.1
        version: 3.34.1
      name: alpha
    defaultChannel: alpha
    packageName: cloudbees-ci-rhmp
    provider:
      name: CloudBees
      url: https://www.cloudbees.com/products/continuous-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: bamoe-kogito-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bamoe-kogito-operator.8.0.5-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "ruleunit-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogitoinfra-sample"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "kafka.strimzi.io/v1beta2",
                    "kind": "Kafka",
                    "name": "kogito-kafka"
                  }
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator@sha256:8a99fcd217502b7e850af1e1097de58b6cb19d1e786394811a8e075c4d871279
          description: IBM BAMOE Kogito Operator for deployment and management of
            Kogito services.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["IBM Business Automation Manager
            Open Edition"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: https://www.ibm.com/mysupport
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.rhpam.kiegroup.org
            version: v1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.rhpam.kiegroup.org
            version: v1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.rhpam.kiegroup.org
            version: v1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.rhpam.kiegroup.org
            version: v1
        description: |-
          IBM BAMOE Kogito Operator is designed for deployment and management of Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Builds a IBM BAMOE Kogito service from a remote git repository and creates a custom image for the given runtime: Red Hat build of Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Deploy IBM BAMOE Kogito service from user provided image.
          * Setup Grafana Dashboard when Grafana Operator is available in cluster.
          * Create Prometheus monitor object when Prometheus operator is available in cluster.

          **IMPORTANT!** IBM BAMOE Kogito Operator may need Grafana and Prometheus Operator to configure monitoring in cases if you do not provide these services beforehand. If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the IBM BAMOE Kogito Operator.
        displayName: IBM BAMOE Kogito Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://ibm.com
        - name: Documentation
          url: https://www.ibm.com/docs/en/ibamoe
        - name: Source
          url: https://github.com/kiegroup/kogito-operator
        maintainers:
        - email: mysphelp@us.ibm.com
          name: IBM
        maturity: beta
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9492c67eaa7ab37df6f5fa623c4aac955b7d7ffaab9ae2e57dea745e82907e40
        - registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator@sha256:8a99fcd217502b7e850af1e1097de58b6cb19d1e786394811a8e075c4d871279
        version: 8.0.5-3
      entries:
      - name: bamoe-kogito-operator.8.0.5-3
        version: 8.0.5-3
      - name: bamoe-kogito-operator.8.0.5-2
        version: 8.0.5-2
      - name: bamoe-kogito-operator.8.0.5-1
        version: 8.0.5-1
      - name: bamoe-kogito-operator.8.0.4-2
        version: 8.0.4-2
      - name: bamoe-kogito-operator.8.0.4-1
        version: 8.0.4-1
      - name: bamoe-kogito-operator.8.0.3-1
        version: 8.0.3-1
      - name: bamoe-kogito-operator.8.0.2-3
        version: 8.0.2-3
      - name: bamoe-kogito-operator.8.0.2-2
        version: 8.0.2-2
      - name: bamoe-kogito-operator.8.0.2-1
        version: 8.0.2-1
      - name: bamoe-kogito-operator.8.0.1-2
        version: 8.0.1-2
      - name: bamoe-kogito-operator.8.0.1-1
        version: 8.0.1-1
      - name: bamoe-kogito-operator.8.0.0-1
        version: 8.0.0-1
      name: 8.x
    defaultChannel: 8.x
    packageName: bamoe-kogito-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc
      provider-url: ""
    name: openshift-gitops-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-gitops-operator.v1.10.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:b0dc693854319ada341e2e3872b34d7db9823b470ed369edfde88d3ae73a40e1
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.8.19\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:a74e0a1fe3589ebe9f226a2bad20e9a5d5601a0e62c5f3290902f466f55b7160
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:2bec31b0c5407ffcecfd8b860d8afb1629eaab31a250f60191f94a824e0a52f4
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:c779ad87a7f71d851c2af818abf16e5151c43d3fd54e64eac2fc7117c00753f5
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:b0dc693854319ada341e2e3872b34d7db9823b470ed369edfde88d3ae73a40e1
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:b5c8a6c7438a45125fc7fd970946f715ab7193b7446d625fae80dde9d1f3f914
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:4c842f0ea8f0e9b53eb16d94488fc2f4298bd8e4be2f48713be9157da24549e6
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:ece2fee4227ad1becadca47028b78ceaf8717a98ec84e3f1b5188a135c4b093c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d35dff5d59e748ef0404e7e76b0237ea96b2bb19b86fa04a3a34df7a379277b2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5d5061dbc2ec5082cf14b6c600fb5400b83cf91d7ccebfa80680a238d275db
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:0905ba3f5cf87780a1e95b6a00ea7cd340d4ed1b5054e4b2c82406270baeaefa
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ec9f60018694dcc5d431ba47d5536b761b71cb3f66684978fe6bb74c157679ac
        version: 1.10.6
      entries:
      - name: openshift-gitops-operator.v1.10.6
        version: 1.10.6
      - name: openshift-gitops-operator.v1.10.5
        version: 1.10.5
      - name: openshift-gitops-operator.v1.10.4
        version: 1.10.4
      - name: openshift-gitops-operator.v1.10.3
        version: 1.10.3
      - name: openshift-gitops-operator.v1.10.2
        version: 1.10.2
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.10
    - currentCSV: openshift-gitops-operator.v1.11.7-0.1724840231.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d6ad2ee900bf99a3672049cc8ce9c94870d41176501ded1abbc1b229b9553401
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          olm.substitutesFor: openshift-gitops-operator.v1.11.7
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.9.21\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:4d22a7a6cc2f0442889512267ae3d9356f05ff0dfdbbe137c3d8ebebc24818c8
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:42f809f32f1d43186b849e3cc78063d3ebb2b54c55b2a1b77908af09a5ce044b
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:0b926fc9eae775425faa1ab9709a36fdac15afe4bfe523720e17ecf5df45bb28
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:0df37cf0a0e48efae02a0711c6da61c6165e7856db97097e98d56c8b6a137bb9
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:eade50b8321c1b2e0a8ac4dcb34596d43dea0a137b8b97b6eb68c6d469c8b200
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:20566173a8e907980e603d82fae4d96fcdfd5c43d9004c2e506cf0e09a47b91e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/rhel8/redis-6@sha256:c4a98010f7dd5faec2047bb9803061704233a8efc930be2a02386ac16d062d86
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c94c2167dc957856682a104b9a6187d3d321dec71ae794acc0dc37794d2cc373
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:7ea96000768b3cf7711150531c4c1c7edff2c45a2f29fe10bd9811ab6b281909
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d6ad2ee900bf99a3672049cc8ce9c94870d41176501ded1abbc1b229b9553401
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:959de6a95ef8978f66fdbe1033bcaa0ddeff9a376023e5d1d2698c5acb417305
        version: 1.11.7+0.1724840231.p
      entries:
      - name: openshift-gitops-operator.v1.11.7-0.1724840231.p
        version: 1.11.7+0.1724840231.p
      - name: openshift-gitops-operator.v1.11.7
        version: 1.11.7
      - name: openshift-gitops-operator.v1.11.6
        version: 1.11.6
      - name: openshift-gitops-operator.v1.11.5
        version: 1.11.5
      - name: openshift-gitops-operator.v1.11.4
        version: 1.11.4
      - name: openshift-gitops-operator.v1.11.3
        version: 1.11.3
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.11
    - currentCSV: openshift-gitops-operator.v1.12.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:48dd1254fd3a887aef2eae9239f68637eae18514a50f72ea504414d83dc727b8
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.10.18\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:dbf3038daf7c797123c511f9a609c3485cb1a5bcae53c614fad17e0a65fa7ee3
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a1d4e26ea755181818401697f18260b9e876aa441be14830f9142081ee22a24d
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:73918f3b3c507fc64006b7a500ecfa9d58151d694852f696c41a4412edffc78e
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:33bbb88343285a0510670384b87d0226dd6de8aa12c78b68921362c4d6f5a6aa
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:48dd1254fd3a887aef2eae9239f68637eae18514a50f72ea504414d83dc727b8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9e0fee30842a16d7b8a39d336123c80c5271cb2ceb50759fc7ee51a4ed09ceb3
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:06a6637c4776f4285bcb295d327ee60b1aed61b7adc5f7ab52898d79281ced70
        - registry.redhat.io/rhel8/redis-6@sha256:a848cc4edb4405846c673187dd68ce66c328dd1ad42f193f5737288c7971399d
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:b5736d1106db29fcdf42328ca20b96382d4339f90b97a07120ea73ee174ea1ad
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:7d82c719b1cef64607d30e8aaa5ab293955771c5b1dbd430e0befc988a43356d
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:1cf485a889c7586cb4944f000b53c0e97d08f97ba14e580bec57dad7cb13712e
        version: 1.12.6
      entries:
      - name: openshift-gitops-operator.v1.12.6
        version: 1.12.6
      - name: openshift-gitops-operator.v1.12.5-0.1727371523.p
        version: 1.12.5+0.1727371523.p
      - name: openshift-gitops-operator.v1.12.5
        version: 1.12.5
      - name: openshift-gitops-operator.v1.12.4
        version: 1.12.4
      - name: openshift-gitops-operator.v1.12.3
        version: 1.12.3
      - name: openshift-gitops-operator.v1.12.2
        version: 1.12.2
      - name: openshift-gitops-operator.v1.12.1
        version: 1.12.1
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.12
    - currentCSV: openshift-gitops-operator.v1.13.3-0.1741683398.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d41034bcf5c28bcb8afef08678793d0cc27a4032e2811a0c496c3bd0f10ecc17
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          olm.substitutesFor: openshift-gitops-operator.v1.13.3-0.1738142397.p
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.11.12\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_gitops/
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d41034bcf5c28bcb8afef08678793d0cc27a4032e2811a0c496c3bd0f10ecc17
        - registry.redhat.io/rhel8/redis-6@sha256:a848cc4edb4405846c673187dd68ce66c328dd1ad42f193f5737288c7971399d
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:b982df02ea7c45a7603d83fe82cac12e1faac303a1b72af6a727512bb97441df
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3f2804bc2af437ff1f06c214b48a1a780dfcfab995650d901c03270057eb9f3c
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:06a6637c4776f4285bcb295d327ee60b1aed61b7adc5f7ab52898d79281ced70
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a7c3e31ce414e6c6fecf1652343b22c52502902676bf4d7a75bf42bdabfca06
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:a366229e6ee26ebe2d7f765b2a2da6057a29e783835655b5d97545ac563b541a
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:d4d5724f67859d24c32f0f5735f5af3e325f77db97f58bc7f17ed4b831b7e534
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c02dc2c79ee24ce5450ea4e19493ac2a1997eeb5bfd359291642c529826d0cdd
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:e87515d016565bb69cb2c2ac40432e0127dd2cb20a490ac82019836032ba15e5
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:bde11fa83b7e708677b08ea995041f480b88edfb93e4a62c10380d101c0a349c
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        version: 1.13.3+0.1741683398.p
      entries:
      - name: openshift-gitops-operator.v1.13.3-0.1741683398.p
        version: 1.13.3+0.1741683398.p
      - name: openshift-gitops-operator.v1.13.3-0.1738142397.p
        version: 1.13.3+0.1738142397.p
      - name: openshift-gitops-operator.v1.13.3
        version: 1.13.3
      - name: openshift-gitops-operator.v1.13.2
        version: 1.13.2
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.13
    - currentCSV: openshift-gitops-operator.v1.14.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:7cf924af77d9c56e70710b99e080ed8aee162f05ec472f32085a76ef85cf9879
          createdAt: "2025-05-22 12:40:45"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.12.13-1\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_gitops/
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:2c45cfc01f9ec7577afc09cadc4c1a6f470836f6a0a7c24d66b791913f89bb05
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:42a79aad641e8607fc339e4ce7f9cf1d258be01f863dbc31a94d2c27edb645bd
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:97e32825f59c33cca7d8dba3580095205c20435ccd15b91cf0aff37db9841e0e
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:7cf924af77d9c56e70710b99e080ed8aee162f05ec472f32085a76ef85cf9879
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:b7d3c882fc25fbbea28951afb197d4d47ace253864889339396e20f0c6897a7f
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:bcf6b3136a5e71803cb518c8763c165bb46ee37ef7bf55503911d7fcb36b3a8f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c369df980874aae270ca7bc6455f1ca45467d75d25fe80d7bcb06a7826bdeaa3
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:cd4965d91c163160bcf43a02c54ca60b8704281c95de904f904df942bfd83bcd
        - registry.redhat.io/rhel8/redis-6@sha256:a9f30ed08abcab0122556a5485fbcf6e0fc03860f19954dcebdbf5e754876a45
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:7358787959af7e7e589edfc043a97d80ebab87536bde365d33be7a600071bf0e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1ed3b514bf9af6ac0d386e3739874d1d964e667d54e041c2d67edd769b0009fb
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:8096f824bd96e3148cf4b6936766c888d1fff0af9e9d0be440ae4feb6adc03f2
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:1112160243f227a5d393eeee165bd4fed0d3f42646a9280cbe9af60ef728aeb2
        version: 1.14.4
      entries:
      - name: openshift-gitops-operator.v1.14.4
        version: 1.14.4
      - name: openshift-gitops-operator.v1.14.3-0.1746016855.p
        version: 1.14.3+0.1746016855.p
      - name: openshift-gitops-operator.v1.14.3-0.1742836272.p
        version: 1.14.3+0.1742836272.p
      - name: openshift-gitops-operator.v1.14.3
        version: 1.14.3
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.14
    - currentCSV: openshift-gitops-operator.v1.15.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ad49a3d14f843bde0d2365093d00ffa6d3b9d59a75c81e0c56710ad37b53bd92
          createdAt: "2024-11-25T03:57:37Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v2.13.9\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:d41454824632475f0e88aa6e0031ec643bd37e1d575df36cd3ee63daf1403281
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:6cec76d7ea70d77923cf4285177e375d34e96e3aecbe77b68cd5ff769e6e540b
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:911733d78ae6f36f642b4d63f044cd102b7eee170508eaf518ac7acf0f339698
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ad49a3d14f843bde0d2365093d00ffa6d3b9d59a75c81e0c56710ad37b53bd92
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:41717c27915251bda8afeef651b609b8a9ed410e0fa09c4cf8fa7e91d6a94292
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:0c7f72546d0d918bfd21aff7157b983ff7c832b269f5efb84d2ed723a09bf612
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:5bda26e2c16beebf66dd8f3c3c55d48e2f00db870f2d9d4d44570d9d3b8bddb0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:a9e2360f00c5c9840a100a6748bd905b8f34169140a33fb0ca4af9cf6db71642
        version: 1.15.4
      entries:
      - name: openshift-gitops-operator.v1.15.4
        version: 1.15.4
      - name: openshift-gitops-operator.v1.15.3
        version: 1.15.3
      - name: openshift-gitops-operator.v1.15.2
        version: 1.15.2
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.15
    - currentCSV: openshift-gitops-operator.v1.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ead539d7cf54992cf1bfeae6a73a453059a7dffef44505991ff33319c8d40c50
          createdAt: "2025-03-07T13:13:07Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v2.14.16\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:58284a12174d19c2d04723fa1f551d96b490872e0fa2ad94af1b3e52dc6ad955
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a5e2a12ec5dea607a6651462d8e7920cc73f3272dba4041c397a14f2c73d7532
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:03d0d9093422396330aa8611adf77abdc9be759f3d2e23850e4abe778bd6002b
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:cab809fd1288d31c93f4f5c4facb7efaff9f844e21a859d1e4557831575a13c8
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c37aa141387c847eac825363dddda3eb59203d3694120dba439b00cf0c66f320
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:2598a3339bc920b91018d3f7acbe78ccbee684b5182c874e48296744d2373f9d
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ead539d7cf54992cf1bfeae6a73a453059a7dffef44505991ff33319c8d40c50
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:ceff1cde6d043daae599b578eec9759c6d49cd5cf770460b9d075fa879a16e81
        version: 1.16.3
      entries:
      - name: openshift-gitops-operator.v1.16.3
        version: 1.16.3
      - name: openshift-gitops-operator.v1.16.2
        version: 1.16.2
      - name: openshift-gitops-operator.v1.16.1
        version: 1.16.1
      - name: openshift-gitops-operator.v1.16.0-0.1746014725.p
        version: 1.16.0+0.1746014725.p
      - name: openshift-gitops-operator.v1.16.0
        version: 1.16.0
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.16
    - currentCSV: openshift-gitops-operator.v1.17.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:1b86eb4e071b1c526931b6d18c3d8f51522293699059c514488602c320271049
          createdAt: "2025-07-30T13:03:16Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.0.14\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:3fc4941fa4a6f2dbe9dd82622a0d3fb794ecd5a9dee286a38b60a7f6a947d685
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:cc45e212e7eb4e7e8daac6ea74071d5d98260de0b1f6d1cea0e7876111ccea8c
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:2f5f997bce924445de735ae0508dca1a7bba561bc4acdacf659928488233cb8a
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:5c9ea426cd60e7b8d1d8e4fe763909200612434c65596855334054e26cbfe3d0
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:70e5e3900d0c89999c69be1d58adf84c1b3664aacbffa9c5e0efc8ae1b6ef9c1
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:1b86eb4e071b1c526931b6d18c3d8f51522293699059c514488602c320271049
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:d73b4f5b355f556208751281923bb8b43edffc061a673c04d53b88f0e50ecb26
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:a2ae66772d90e734bb3385a59852e6016325a21b916a2580a33cf0da7cb0b527
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:c0a513ed54b631629efafe6e94b775d5bd0a943c44fbe96f02a0d5bb1f989099
        version: 1.17.1
      entries:
      - name: openshift-gitops-operator.v1.17.1
        version: 1.17.1
      - name: openshift-gitops-operator.v1.17.0
        version: 1.17.0
      name: gitops-1.17
    - currentCSV: openshift-gitops-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
          createdAt: "2025-08-21T01:20:45Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.18.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: NamespaceManagement
            name: namespacemanagements.argoproj.io
            version: v1beta1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.1.6\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/rhel9/redis-7@sha256:b4f58ad83ab8719a2e102f38ee526546cd1d56428901afc3add45c9a0e2a2f12
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:1eef9ecd58cf8391f29360ef74dd8aa5dcb48e1582d1500e445706d8302c951e
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a0544c14823492165550d83a6d8ba79dd632b46144d3fdcb543793726111d76
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:6284873bcf1593c210003e0292ac5583f03f03e80d937098d808d48fa0d01cf6
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:5a09fcba3d9e6fc462719e81a46af6022cf7f70a05bcb387ed959a49f6f37dbc
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:adca9c55b80d899d500bf0a7234cf61ae30ffed1b2c927ddea8d9fd07c9acb6a
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7f6ceae56c60fd432425c36f93a9e9bbbb96ef9422333ad6ea41ff3a2eaf8374
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:640e3298c7cd2faa84971b0d861446c4892ab8602108af05b8b5b049a4de624f
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c2283997d47298d49378e8200b149ea613c7735317ef12e74b7915a590f41ce2
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:08efd3e7a79c2945693c99ed31ec912223becff7238b1e28d1b1da5d63f93019
        version: 1.18.0
      entries:
      - name: openshift-gitops-operator.v1.18.0
        version: 1.18.0
      name: gitops-1.18
    - currentCSV: openshift-gitops-operator.v1.6.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:104cc7564fcc3261faa3fbf9c95720b4434e15103b876c080ed012b244ef1dae
          createdAt: "2021-09-15T22:31:21Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD v2.4.28\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces. \n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc.
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:104cc7564fcc3261faa3fbf9c95720b4434e15103b876c080ed012b244ef1dae
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:0cc4fa704e581cd5581c816668183f602fae6ef9ab3c04ce0d99d561287ef140
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:06f1150114d4c0bd236d2a5d4716c9c3e95a9733058297790389d6ba2a1bf264
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:e9e2f950cda8153cff112e43872b98e4977d3cbdf59713b8de92defef6d34b6d
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:dbcad68b70ab7918e944099aef4177c9bc3378e9ca49499dbb94a175927d3209
        - registry.redhat.io/rhel8/redis-6@sha256:53598a6effeb90e4f1b005b2521beffd2fa2b0c52d0e7f2347ee2abd2577cab3
        version: 1.6.6
      entries:
      - name: openshift-gitops-operator.v1.6.6
        version: 1.6.6
      - name: openshift-gitops-operator.v1.6.5
        version: 1.6.5
      - name: openshift-gitops-operator.v1.6.4
        version: 1.6.4
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.6
    - currentCSV: openshift-gitops-operator.v1.7.4-0.1690486082.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:e0967a4058fe5a9bfe7e998a3347529c99536cf3a44de9e40f9632035ff5a2fe
          createdAt: "2021-09-15T22:31:21Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.substitutesFor: openshift-gitops-operator.v1.7.4
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD v2.5.16\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficient privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces. \n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/rhel8/redis-6@sha256:53598a6effeb90e4f1b005b2521beffd2fa2b0c52d0e7f2347ee2abd2577cab3
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:e0967a4058fe5a9bfe7e998a3347529c99536cf3a44de9e40f9632035ff5a2fe
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c55aadf45397083bec4e42d6c0514ed79169a07fd3d9d75671904df2c0520e52
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:45789c49d9334e0ed062641313714c101bd0117dc5633316b1ec63e7a6c76c15
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:26050fb95978e90e8d00e895032045e77cb236008a0e1417459e1aeca8a48c69
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:5c11be8ec9db54331893f2f298aa27006e77b0eaded2d88d0531c77da91a421f
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:f1371d2012796e9aaca105112190ff89ec3d0be95968e05b19db101480f67975
        version: 1.7.4+0.1690486082.p
      entries:
      - name: openshift-gitops-operator.v1.7.4-0.1690486082.p
        version: 1.7.4+0.1690486082.p
      - name: openshift-gitops-operator.v1.7.4
        version: 1.7.4
      - name: openshift-gitops-operator.v1.7.3
        version: 1.7.3
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.7
    - currentCSV: openshift-gitops-operator.v1.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f32df77b3c23d6a833ff887386285ca7a5bd436641f0ec1ffc934c8247fa315f
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.skipRange: ""
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.6.153\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:8cc59901689858e06f5c2942f8c975449c17454fa8364da6153d9b5a3538a985
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:f77594bc053be144b33ff9603ee9675c7e82946ec0dbfb04d8f942c8d73155da
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f32df77b3c23d6a833ff887386285ca7a5bd436641f0ec1ffc934c8247fa315f
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:8aa68fddef88921aeeede24a849d3dabbceff01dc03185342c21e6ee58033eaf
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:16d9184345e6d5a19f326a7b52f76d8a59b96763f233c50546ff017a625e5ff5
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:0d13c03048b4435d9f5045a6fb7bd4b8808461b26aeca60c0047fa5937e1d45d
        version: 1.8.6
      entries:
      - name: openshift-gitops-operator.v1.8.6
        version: 1.8.6
      - name: openshift-gitops-operator.v1.8.5
        version: 1.8.5
      - name: openshift-gitops-operator.v1.8.4-0.1693214052.p
        version: 1.8.4+0.1693214052.p
      - name: openshift-gitops-operator.v1.8.4
        version: 1.8.4
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.8
    - currentCSV: openshift-gitops-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "argo-rollout"
                },
                "spec": null
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:11a6084e11aa32511a34bd6cfb4f4995aec690c1f5ed61861e99f6e42d62afb4
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.skipRange: ""
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.7.16\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:80f92842fe4cba008a873ad2ab6059a3a24e76e16671d8ece034dbbe666996bc
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:3d17786e2f4063154b9e3a40b96bf8ec0a14b86395cea5fe7ca18c5b43764478
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a568c8c647e2e2d6abdf12138eeef8f8bdde1fdcbf786e9ca8c8f93bc79acaf9
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:43bb961d847477902e66a92edd1539322cac8d5f74706b048a458ad037a0c7a3
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:11a6084e11aa32511a34bd6cfb4f4995aec690c1f5ed61861e99f6e42d62afb4
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ec9f60018694dcc5d431ba47d5536b761b71cb3f66684978fe6bb74c157679ac
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:3873d1f18b5643e1de2601419410270342650a21428bb26a4cbee20c0c63f98b
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:86012145d4b8798dc12f6f4f07a9316a193188aaac865e324b63f4e60e8ec584
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:136d3a936759e112d3ea2e78f7c02bd882b0fd3011f4c452f45c32c43822772d
        version: 1.9.4
      entries:
      - name: openshift-gitops-operator.v1.9.4
        version: 1.9.4
      - name: openshift-gitops-operator.v1.9.3
        version: 1.9.3
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.9
    - currentCSV: openshift-gitops-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
          createdAt: "2025-08-21T01:20:45Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.18.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: NamespaceManagement
            name: namespacemanagements.argoproj.io
            version: v1beta1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.1.6\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:adca9c55b80d899d500bf0a7234cf61ae30ffed1b2c927ddea8d9fd07c9acb6a
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7f6ceae56c60fd432425c36f93a9e9bbbb96ef9422333ad6ea41ff3a2eaf8374
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:6284873bcf1593c210003e0292ac5583f03f03e80d937098d808d48fa0d01cf6
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:08efd3e7a79c2945693c99ed31ec912223becff7238b1e28d1b1da5d63f93019
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:640e3298c7cd2faa84971b0d861446c4892ab8602108af05b8b5b049a4de624f
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c2283997d47298d49378e8200b149ea613c7735317ef12e74b7915a590f41ce2
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:5a09fcba3d9e6fc462719e81a46af6022cf7f70a05bcb387ed959a49f6f37dbc
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:1eef9ecd58cf8391f29360ef74dd8aa5dcb48e1582d1500e445706d8302c951e
        - registry.redhat.io/rhel9/redis-7@sha256:b4f58ad83ab8719a2e102f38ee526546cd1d56428901afc3add45c9a0e2a2f12
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a0544c14823492165550d83a6d8ba79dd632b46144d3fdcb543793726111d76
        version: 1.18.0
      entries:
      - name: openshift-gitops-operator.v1.18.0
        version: 1.18.0
      - name: openshift-gitops-operator.v1.17.1
        version: 1.17.1
      - name: openshift-gitops-operator.v1.17.0
        version: 1.17.0
      - name: openshift-gitops-operator.v1.16.2
        version: 1.16.2
      - name: openshift-gitops-operator.v1.16.1
        version: 1.16.1
      - name: openshift-gitops-operator.v1.16.0-0.1746014725.p
        version: 1.16.0+0.1746014725.p
      - name: openshift-gitops-operator.v1.16.0
        version: 1.16.0
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: latest
    defaultChannel: latest
    packageName: openshift-gitops-operator
    provider:
      name: Red Hat Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: node-maintenance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-maintenance-operator.v0.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-maintenance-operator:v0.17.0
          createdAt: "2024-03-13 18:56:47"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=0.12.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/medik8s/node-maintenance-operator:v0.17.0
        version: 0.17.0
      entries:
      - name: node-maintenance-operator.v0.17.0
        version: 0.17.0
      - name: node-maintenance-operator.v0.16.1
        version: 0.16.1
      - name: node-maintenance-operator.v0.16.0
        version: 0.16.0
      - name: node-maintenance-operator.v0.15.0
        version: 0.15.0
      - name: node-maintenance-operator.v0.14.0
        version: 0.14.0
      - name: node-maintenance-operator.v0.13.0
        version: 0.13.0
      - name: node-maintenance-operator.v0.12.1
        version: 0.12.1
      name: stable
    defaultChannel: stable
    packageName: node-maintenance-operator
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://www.redhat.com
    name: sap-commerce-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sap-commerce-operator.v0.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hybris.hybris.org/v1alpha1",
                "kind": "HybrisApp",
                "metadata": {
                  "name": "hybrisapp-sample"
                },
                "spec": {
                  "aJPServicePort": 30000,
                  "apachejvmRouteName": "",
                  "baseImageName": "hybris-base",
                  "baseImageTag": "v0.2",
                  "hybrisANTTaskNames": "",
                  "podHealthyProbePath": "/platform/init",
                  "sourceRepoLocalPropertiesOverride": "",
                  "sourceRepoRef": "main",
                  "sourceRepoURL": "https://github.com/redhat-sap/sap-commerce-app-example.git",
                  "startupProbeFailureThreshold": 100,
                  "startupProbePeriodSecond": 60
                }
              },
              {
                "apiVersion": "hybris.hybris.org/v1alpha1",
                "kind": "HybrisBase",
                "metadata": {
                  "name": "hybrisbase-sample"
                },
                "spec": {
                  "URL": "http://nexus.redhat.com/nexus/service/local/repositories/releases/content/org/hybris/hybris/1.0/hybris-1.0.tar.gz",
                  "buildSourceRepo": "https://github.com/redhat-sap/sap-commerce-operator",
                  "buildSourceRepoBranch": "main",
                  "imageName": "hybris-base",
                  "imageTag": "v0.2",
                  "jdkURL": "https://github.com/SAP/SapMachine/releases/download/sapmachine-11.0.5/sapmachine-jdk-11.0.5-1.x86_64.rpm",
                  "password": "admin123",
                  "username": "admin"
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          containerImage: quay.io/redhat-sap-cop/hybris-operator:0.0.4
          description: Builds and runs SAP Commerce in OpenShift
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-sap/sap-commerce-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HybrisApp is the Schema for the hybrisapps API
            displayName: Hybris Application
            kind: HybrisApp
            name: hybrisapps.hybris.hybris.org
            version: v1alpha1
          - description: HybrisBase is the Schema for the Hybris Base API
            displayName: Hybris Base
            kind: HybrisBase
            name: hybrisbases.hybris.hybris.org
            version: v1alpha1
        description: |-
          Builds and runs SAP Commerce in OpenShift

          * The operator provides APIs to initialize SAP Commerce with the provided SAP Commerce Server. Though there are no restrictions from the operator on the base package to be used from SAP, the default SAP base image provides a server preconfigured with an application server based on Apache Tomcat. Both HSQLDB and Tomcat are included in the SAP Commerce package, and are designed to be used for testing, development, and demonstration purposes. The custom classes created for various business needs along with an external RDBMS configuration can be used in a HybrisApp creation to execute SAP Commerce in a production setup.

          * The operator also enables the use of SAP Commerce in a clustered model. A SAP Commerce Cluster is a number of individual SAP Commerce installations using a common set of data on one database. A cluster configuration lets you balance the load between the nodes to scale SAP Commerce in a way that is transparent to the user.

          * The clustering offers few communication protocols as options (UDP: both multicast or unicast, or JGroups). In order to support multicast for SAP Commerce clustering the corresponding OpenShift namespace has to be enabled for multicast usage.

          For more details on SAP Commerce consult the product documentation: [https://help.sap.com/viewer/product/SAP_COMMERCE](https://help.sap.com/viewer/product/SAP_COMMERCE)
        displayName: SAP Commerce Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sap-commerce-operator
        links:
        - name: SAP Commerce Operator
          url: https://github.com/redhat-sap/sap-commerce-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
          url: https://www.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/redhat-sap-cop/hybris-operator:0.0.4
        version: 0.0.4
      entries:
      - name: sap-commerce-operator.v0.0.4
        version: 0.0.4
      - name: sap-commerce-operator.v0.0.3
        version: 0.0.3
      - name: sap-commerce-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: sap-commerce-operator
    provider:
      name: Red Hat, Inc.
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cognitoidentityprovider-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cognitoidentityprovider-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cognitoidentityprovider-controller:1.1.1
          createdAt: "2025-09-19T20:58:49Z"
          description: AWS Cognito Identity Provider controller is a service controller
            for managing Cognito Identity Provider resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: UserPool represents the state of an AWS cognitoidentityprovider
              UserPool resource.
            displayName: UserPool
            kind: UserPool
            name: userpools.cognitoidentityprovider.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Cognito Identity Provider resources in AWS from within your Kubernetes cluster.

          **About Amazon Cognito Identity Provider**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Cognito Identity Provider
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cognitoidentityprovider
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Cognito Identity Provider Developer Resources
          url: https://aws.amazon.com/Cognito Identity Provider/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cognitoidentityprovider maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cognitoidentityprovider-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-cognitoidentityprovider-controller.v1.1.1
        version: 1.1.1
      - name: ack-cognitoidentityprovider-controller.v1.1.0
        version: 1.1.0
      - name: ack-cognitoidentityprovider-controller.v1.0.12
        version: 1.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-cognitoidentityprovider-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sumo Logic
      provider-url: https://www.sumologic.com/
    name: sumologic-kubernetes-collection-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "helm-operator.sumologic.com/v1alpha1",
                    "kind": "SumologicCollection",
                    "metadata": {
                        "name": "collection"
                    },
                    "spec": {
                        "sumologic": {
                            "accessId": "",
                            "accessKey": "",
                            "clusterName": "kubernetes",
                            "scc": {
                                "create": true
                            }
                        },
                        "kube-prometheus-stack": {
                            "prometheus-node-exporter": {
                                "service": {
                                    "port": 9200,
                                    "targetPort": 9200
                                }
                            }
                        },
                        "otellogs": {
                            "daemonset": {
                                "containers": {
                                    "otelcol": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                },
                                "initContainers": {
                                    "changeowner": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                }
                            }
                        }
                    }
                }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing,Monitoring
          containerImage: registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
          createdAt: "2024-07-19T10:48:00Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator.git
          support: Sumo Logic
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configuration for Sumo Logic Kubernetes Collection
            displayName: SumologicCollection
            kind: SumologicCollection
            name: sumologiccollections.helm-operator.sumologic.com
            version: v1alpha1
        description: Helm Operator for the Sumo Logic Kubernetes Collection Helm Chart
          v4.13.0-2
        displayName: Sumo Logic Kubernetes Collection Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - collection
        - logging
        - metrics
        - monitoring
        links:
        - name: Sumo Logic Kubernetes Collection Helm Operator
          url: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator
        maintainers:
        - email: collection@sumologic.com
          name: Sumo Logic
        maturity: alpha
        provider:
          name: Sumo Logic
          url: https://www.sumologic.com/
        relatedImages:
        - registry.connect.redhat.com/sumologic/thanos@sha256:16faf7191668e86d8b576bba055dd0cbdeb009635976b9e07bd00f926032f11c
        - registry.connect.redhat.com/sumologic/autoinstrumentation-java@sha256:242958b7d8905f4ac02f6ad38897ba58b9badf2a6213a389e67dcc71fee3630c
        - registry.connect.redhat.com/sumologic/busybox@sha256:209bc8e0b1d958fc699f3bb75b1248544e372ad35cdc46c991bd1698f641e1f9
        - registry.connect.redhat.com/sumologic/opentelemetry-operator@sha256:5587fe5b7ebfa57ad320ad2521103f908317388d1741c40e5f7be5ed9e5f9ad7
        - registry.connect.redhat.com/sumologic/prometheus@sha256:0ef6eaf570a40499dd5de92a79593e950bc771b014e0ee6ff7419b4b82a36fcc
        - registry.connect.redhat.com/sumologic/nginx-unprivileged@sha256:7260555d16c94c4fe1e016626b4cb8a5c0de6b700e2eea28744331fb298f71ae
        - registry.connect.redhat.com/sumologic/node-exporter@sha256:11e9e604432a66d9b388cb1bf70eac077c7f0372d1264342ef927d76b8ef0272
        - registry.connect.redhat.com/sumologic/prometheus-config-reloader@sha256:0b605d36782afd8f50fd631d0e9c7b19d39ce962af89b7a737b2f09a8ceff291
        - registry.connect.redhat.com/sumologic/kubernetes-tools-kubectl@sha256:d3155b864cc74f327adedf7d013a8f0fcb11c2e73a5655b7cafa848824e80172
        - registry.connect.redhat.com/sumologic/telegraf-operator@sha256:ff958222be01c0c3966474463e43bbd5bcf348790dd34bb761718cb1cea2d793
        - registry.connect.redhat.com/sumologic/prometheus-operator@sha256:58a329e6582456b32abebb470086191f8366b927af4703741613993465f116b8
        - registry.connect.redhat.com/sumologic/sumologic-otel-collector@sha256:4554d43dba204e3ec8e91b36442a438f10234fa0cb4db7fb7880b257a6485112
        - registry.connect.redhat.com/sumologic/tailing-sidecar@sha256:738d653a78dc415243ab3627db9135768175fb7f51d0803566f5d9e824c8744b
        - registry.connect.redhat.com/sumologic/autoinstrumentation-python@sha256:f0fa45fa45e64c7df88f21181ae18ee96fe83cd3988258663d0806ea6f39e6bb
        - registry.connect.redhat.com/sumologic/kubernetes-setup@sha256:f07eaf68fa11ffc708e4485a83bd3763fc5712f7d73d29edfe5b9c9e08f05df6
        - registry.connect.redhat.com/sumologic/autoinstrumentation-dotnet@sha256:f4f385c4489c74b5c43f2fe6b6620a9eb8d9ecab470567236b3c45f636106640
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:e1f49fcf5b7e86d9eb49c16dbbf9484cb486f55f999fd51d7c738c3b1f2666cf
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
        - registry.connect.redhat.com/sumologic/kube-state-metrics@sha256:3551ededa327a35db798ec8286cd843580e3ae395d2231e041e8c4b7621127ee
        - registry.connect.redhat.com/sumologic/autoinstrumentation-nodejs@sha256:c206eef4fdecaa58b4f01e5e7f0dbd74e9606eab2b4debf47461d36191e248d1
        - registry.connect.redhat.com/sumologic/metrics-server@sha256:65d2a3ddc4b4183f438fa18b655079c93c905cb9e3b8c54977336205dc1fff07
        - registry.connect.redhat.com/sumologic/tailing-sidecar-operator@sha256:c16b8b18e9b263fc3eb24059bba512bcfc3201b32105af6e5518392826ddd66c
        - registry.connect.redhat.com/sumologic/telegraf@sha256:75efd2e509c6017ea0fc5c9103ce068cbc89087b16506a4ef48cb5ff00eab84e
        version: 4.13.0-2
      entries:
      - name: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
        version: 4.13.0-2
      name: alpha
    defaultChannel: alpha
    packageName: sumologic-kubernetes-collection-helm-operator
    provider:
      name: Sumo Logic
      url: https://www.sumologic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: rh-service-binding-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-binding-operator.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ServiceBinding expresses intent to bind an operator-backed
              service with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding for Kubernetes. Describes the connection
              between a Provisioned Service and an Application Projection. Implements
              Service Binding Specifcation for Kubernetes (https://github.com/k8s-service-bindings/spec).
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: alpha
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:fe6629401bc8447da95ec55949b5a27e620364bb6e02f86230262699ac8a5d55
        version: 0.9.1
      entries:
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: beta
    - currentCSV: service-binding-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindablekinds-sample"
                }
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:e799bd06043028d214baeeffc683710ad2eed7b83272fc653e3b89f69711319b
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:e799bd06043028d214baeeffc683710ad2eed7b83272fc653e3b89f69711319b
        version: 0.11.0
      entries:
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: candidate
    - currentCSV: service-binding-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:3d7a3648e86848402577dd19c1f8fd130a5556ebdfbc0f663a72f9e141024812
        version: 0.10.0
      entries:
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: preview
    - currentCSV: service-binding-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindable-kinds"
                },
                "spec": {}
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  },
                  "type": "postgres",
                  "workload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: registry.redhat.io/ocp-tools-4/service-binding-rhel8-operator@sha256:9c8fb0c8f9d2f54685561af21e595aa88bf309eaebebdf23fb6997b75d026b18
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'.
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1alpha3
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1beta1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
          - description: Service Binding implementing community specification (https://github.com/k8s-service-bindings/spec)
              describing the connection between a backing service and an application
              workload.
            displayName: Service Binding (Spec API)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1beta1
        description: |
          The Service Binding Operator manages the data plane for applications and backing services. It enables developers to connect their
          applications to backing services (REST API, databases, event buses and others) with a consistent and predictable
          experience. The operator reads data made available by the control plane of backing services and projects the bindings
          data to applications either as accessible files mounted into the application’s container or environment variables.

          In Kubernetes, each service suggests a different way to access their secrets and developers will consume them in their
          application in a custom way. While this provides a good deal of flexibility, it also creates discrepancies and
          complexity dealing with each specific services, which leads large development teams to lose overall velocity.

          Service Binding Operator removes the error-prone manual configuration of binding information, provides service operators
          a low-touch administrative experience to provision and manage access to services and enriches development lifecycle.

          For additional details refer to the project [documentation](https://redhat-developer.github.io/service-binding-operator).

          Get Started with Service Binding Operator
          * Find out about SBO’s [prerequisites](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/installing-service-binding.html).
          * Refer to the [Quick Start Guide](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/quick-start.html) to see SBO in action in a simple scenario.

          Need Help
          * Raise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)

          Licence
          * Service Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE)

          Service Binding Operator is providing and supporting two different APIs.
          * `binding.operators.coreos.com/v1alpha1`: This API is compliant with the Service Binding specification. It is fully supported and considered as the mature API. It’s the API we recommend in production environments.
          * `servicebinding.io/v1beta1`: This API is the one defined in the Service Binding specification. It is available in Service Binding Operator 1.x.

          We’ll consider supporting only the API from the specification, once the Service Binding specification will be officially GA. Of course, that will be a breaking change and handle in a safe way for you.
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Documentation
          url: https://redhat-developer.github.io/service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4/service-binding-rhel8-operator@sha256:9c8fb0c8f9d2f54685561af21e595aa88bf309eaebebdf23fb6997b75d026b18
        version: 1.3.3
      entries:
      - name: service-binding-operator.v1.3.3
        version: 1.3.3
      - name: service-binding-operator.v1.3.1
        version: 1.3.1
      - name: service-binding-operator.v1.3.0
        version: 1.3.0
      - name: service-binding-operator.v1.2.0
        version: 1.2.0
      - name: service-binding-operator.v1.1.1
        version: 1.1.1
      - name: service-binding-operator.v1.1.0
        version: 1.1.0
      - name: service-binding-operator.v1.0.1
        version: 1.0.1
      - name: service-binding-operator.v1.0.0
        version: 1.0.0
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: rh-service-binding-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cluster-logging
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-logging.v5.9.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "name": "remote-elasticsearch",
                      "secret": {
                        "name": "elasticsearch"
                      },
                      "type": "elasticsearch",
                      "url": "remote.example.org:9200"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "audit",
                        "infrastructure"
                      ],
                      "name": "enable-default-log-store",
                      "outputRefs": [
                        "default"
                      ]
                    },
                    {
                      "inputRefs": [
                        "application"
                      ],
                      "name": "forward-to-remote",
                      "outputRefs": [
                        "remote-elasticsearch"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogging",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "collection": {
                    "type": "vector"
                  },
                  "logStore": {
                    "elasticsearch": {
                      "nodeCount": 3,
                      "redundancyPolicy": "SingleRedundancy",
                      "resources": {
                        "requests": {
                          "memory": "2Gi"
                        }
                      },
                      "storage": {
                        "size": "200G"
                      }
                    },
                    "retentionPolicy": {
                      "application": {
                        "maxAge": "7d"
                      }
                    },
                    "type": "elasticsearch"
                  },
                  "managementState": "Managed",
                  "visualization": {
                    "kibana": {
                      "replicas": 1
                    },
                    "type": "kibana"
                  }
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          console.openshift.io/plugins: '["logging-view-plugin"]'
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:54054faf0d2f375dd6d7fc7929a2dbb529508e1b3c004a56fc97c2a682d16be2
          createdAt: "2025-03-14T20:19:21Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing your aggregated logging stack.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.7.0-0 <5.9.15'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Logging (team-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.


              There are built-in input names for common log categories, and you can
              define custom inputs to do additional filtering.


              There is a built-in output name for the default openshift log store, but
              you can define your own outputs with a URL and other connection information
              to forward logs to other stores or processors, inside or outside the cluster.


              For more details see the documentation on the API fields.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.logging.openshift.io
            version: v1
          - description: A Red Hat OpenShift Logging instance. ClusterLogging is the
              Schema for the clusterloggings API
            displayName: Cluster Logging
            kind: ClusterLogging
            name: clusterloggings.logging.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates and manages the aggregated logging stack as a cluster-wide service.

          ##Features
          * **Create/Destroy**: Launch and create an aggregated logging stack to support the entire OCP cluster.
          * **Simplified Configuration**: Configure your aggregated logging cluster's structure like components and end points easily.
          * Configure log forwarding to an in-cluster, Red Hat managed log storage solution based on LokiStack or Elasticsearch

          ## Prerequisites
          ### Red Hat OpenShift Logging Namespace
          Cluster logging and the Red Hat OpenShift Logging Operator is only deployable to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - fluentd
        - logging
        - aggregated
        - efk
        - vector
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Fluentd
          url: https://www.fluentd.org/
        - name: Vector
          url: https://vector.dev/
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:54054faf0d2f375dd6d7fc7929a2dbb529508e1b3c004a56fc97c2a682d16be2
        - registry.redhat.io/openshift-logging/logging-view-plugin-rhel9@sha256:8023b072b45e9a21ddef83a56f4760d476053250d5a920d8eb5f03eecab8b1b3
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:54e3a5efaae5ecc2a08db03a421aa79e323b48c764115ea84d83a37c21d1d997
        - registry.redhat.io/openshift-logging/fluentd-rhel9@sha256:a676c9164ec2523b2002dc5eba234ce774926dbf7fe87de6d72b53f99cbf1db5
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:865572275b79462f14e84c69438a61afe9465cfe45ec164caf30965730b276f7
        version: 5.9.15
      entries:
      - name: cluster-logging.v5.9.15
        version: 5.9.15
      - name: cluster-logging.v5.9.14
        version: 5.9.14
      - name: cluster-logging.v5.9.13
        version: 5.9.13
      - name: cluster-logging.v5.9.12
        version: 5.9.12
      - name: cluster-logging.v5.9.11
        version: 5.9.11
      - name: cluster-logging.v5.9.10
        version: 5.9.10
      - name: cluster-logging.v5.9.9
        version: 5.9.9
      - name: cluster-logging.v5.9.8
        version: 5.9.8
      - name: cluster-logging.v5.9.7
        version: 5.9.7
      - name: cluster-logging.v5.9.6
        version: 5.9.6
      - name: cluster-logging.v5.9.5
        version: 5.9.5
      - name: cluster-logging.v5.9.4
        version: 5.9.4
      - name: cluster-logging.v5.9.3
        version: 5.9.3
      - name: cluster-logging.v5.9.2
        version: 5.9.2
      - name: cluster-logging.v5.9.1
        version: 5.9.1
      - name: cluster-logging.v5.9.0
        version: 5.9.0
      name: stable
    - currentCSV: cluster-logging.v5.8.22
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "name": "remote-elasticsearch",
                      "secret": {
                        "name": "elasticsearch"
                      },
                      "type": "elasticsearch",
                      "url": "remote.example.org:9200"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "audit",
                        "infrastructure"
                      ],
                      "name": "enable-default-log-store",
                      "outputRefs": [
                        "default"
                      ]
                    },
                    {
                      "inputRefs": [
                        "application"
                      ],
                      "name": "forward-to-remote",
                      "outputRefs": [
                        "remote-elasticsearch"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogging",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "collection": {
                    "type": "vector"
                  },
                  "logStore": {
                    "elasticsearch": {
                      "nodeCount": 3,
                      "redundancyPolicy": "SingleRedundancy",
                      "resources": {
                        "requests": {
                          "memory": "2Gi"
                        }
                      },
                      "storage": {
                        "size": "200G"
                      }
                    },
                    "retentionPolicy": {
                      "application": {
                        "maxAge": "7d"
                      }
                    },
                    "type": "elasticsearch"
                  },
                  "managementState": "Managed",
                  "visualization": {
                    "kibana": {
                      "replicas": 1
                    },
                    "type": "kibana"
                  }
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.12-v4.17
          console.openshift.io/plugins: '["logging-view-plugin"]'
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:8e13bd3e1654fe26c1d4d8d4a1bfb12e04cee7899cf16b5fa07b62b8f7494af5
          createdAt: "2024-10-17T17:45:03Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing your aggregated logging stack.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.22'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-5.8
          operators.operatorframework.io.bundle.channels.v1: stable-5.8
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Logging (team-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "ClusterLogForwarder is an API to configure forwarding logs.
              \n You configure forwarding by specifying a list of `pipelines`, which
              forward from a set of named inputs to a set of named outputs. \n There
              are built-in input names for common log categories, and you can define
              custom inputs to do additional filtering. \n There is a built-in output
              name for the default openshift log store, but you can define your own
              outputs with a URL and other connection information to forward logs
              to other stores or processors, inside or outside the cluster. \n For
              more details see the documentation on the API fields."
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.logging.openshift.io
            version: v1
          - description: A Red Hat OpenShift Logging instance. ClusterLogging is the
              Schema for the clusterloggings API
            displayName: Cluster Logging
            kind: ClusterLogging
            name: clusterloggings.logging.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          ### Notice for OCP 4.16 and 4.17 Only

          Logging 5.8 is supported on OCP 4.16 and 4.17 to help customers upgrade and migrate to Logging 6.0. **Logging 5.8 will be End of Life in November 2025.**  At that time, it will no longer be available or supported.

          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates and manages the aggregated logging stack as a cluster-wide service.

          ##Features
          * **Create/Destroy**: Launch and create an aggregated logging stack to support the entire OCP cluster.
          * **Simplified Configuration**: Configure your aggregated logging cluster's structure like components and end points easily.
          * Configure log forwarding to an in-cluster, Red Hat managed log storage solution based on LokiStack or Elasticsearch

          ## Prerequisites
          ### Red Hat OpenShift Logging Namespace
          Cluster logging and the Red Hat OpenShift Logging Operator is only deployable to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - fluentd
        - logging
        - aggregated
        - efk
        - vector
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Fluentd
          url: https://www.fluentd.org/
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/logging-view-plugin-rhel9@sha256:f00d4e7ce6b37f1a053c5dbe554359d844989c2243574000ada0f5bc678ffff8
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:8e13bd3e1654fe26c1d4d8d4a1bfb12e04cee7899cf16b5fa07b62b8f7494af5
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:f26518dd3dbf6b6a9e9e20e40741687996b1160c6ad8ad417989197800c2b3fe
        - registry.redhat.io/openshift-logging/fluentd-rhel9@sha256:9296e50fccae60dea580186bc5ae2a387172dc90ece35663d96c4553a2446a01
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:8f29d6eaeb4a11c9a61e3d763605e3c9c8655df7fbb4e39e5b8311d18449f04a
        version: 5.8.22
      entries:
      - name: cluster-logging.v5.8.22
        version: 5.8.22
      - name: cluster-logging.v5.8.21
        version: 5.8.21
      - name: cluster-logging.v5.8.20
        version: 5.8.20
      - name: cluster-logging.v5.8.19
        version: 5.8.19
      - name: cluster-logging.v5.8.18
        version: 5.8.18
      - name: cluster-logging.v5.8.17
        version: 5.8.17
      - name: cluster-logging.v5.8.16
        version: 5.8.16
      - name: cluster-logging.v5.8.15
        version: 5.8.15
      - name: cluster-logging.v5.8.14
        version: 5.8.14
      - name: cluster-logging.v5.8.13
        version: 5.8.13
      - name: cluster-logging.v5.8.12
        version: 5.8.12
      - name: cluster-logging.v5.8.11
        version: 5.8.11
      - name: cluster-logging.v5.8.10
        version: 5.8.10
      - name: cluster-logging.v5.8.9
        version: 5.8.9
      - name: cluster-logging.v5.8.8
        version: 5.8.8
      - name: cluster-logging.v5.8.7
        version: 5.8.7
      - name: cluster-logging.v5.8.6
        version: 5.8.6
      - name: cluster-logging.v5.8.5
        version: 5.8.5
      - name: cluster-logging.v5.8.4
        version: 5.8.4
      - name: cluster-logging.v5.8.3
        version: 5.8.3
      - name: cluster-logging.v5.8.2
        version: 5.8.2
      - name: cluster-logging.v5.8.1
        version: 5.8.1
      - name: cluster-logging.v5.8.0
        version: 5.8.0
      name: stable-5.8
    - currentCSV: cluster-logging.v5.9.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "name": "remote-elasticsearch",
                      "secret": {
                        "name": "elasticsearch"
                      },
                      "type": "elasticsearch",
                      "url": "remote.example.org:9200"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "audit",
                        "infrastructure"
                      ],
                      "name": "enable-default-log-store",
                      "outputRefs": [
                        "default"
                      ]
                    },
                    {
                      "inputRefs": [
                        "application"
                      ],
                      "name": "forward-to-remote",
                      "outputRefs": [
                        "remote-elasticsearch"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogging",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "collection": {
                    "type": "vector"
                  },
                  "logStore": {
                    "elasticsearch": {
                      "nodeCount": 3,
                      "redundancyPolicy": "SingleRedundancy",
                      "resources": {
                        "requests": {
                          "memory": "2Gi"
                        }
                      },
                      "storage": {
                        "size": "200G"
                      }
                    },
                    "retentionPolicy": {
                      "application": {
                        "maxAge": "7d"
                      }
                    },
                    "type": "elasticsearch"
                  },
                  "managementState": "Managed",
                  "visualization": {
                    "kibana": {
                      "replicas": 1
                    },
                    "type": "kibana"
                  }
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          console.openshift.io/plugins: '["logging-view-plugin"]'
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:54054faf0d2f375dd6d7fc7929a2dbb529508e1b3c004a56fc97c2a682d16be2
          createdAt: "2025-03-14T20:19:21Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing your aggregated logging stack.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.7.0-0 <5.9.15'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Logging (team-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.


              There are built-in input names for common log categories, and you can
              define custom inputs to do additional filtering.


              There is a built-in output name for the default openshift log store, but
              you can define your own outputs with a URL and other connection information
              to forward logs to other stores or processors, inside or outside the cluster.


              For more details see the documentation on the API fields.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.logging.openshift.io
            version: v1
          - description: A Red Hat OpenShift Logging instance. ClusterLogging is the
              Schema for the clusterloggings API
            displayName: Cluster Logging
            kind: ClusterLogging
            name: clusterloggings.logging.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates and manages the aggregated logging stack as a cluster-wide service.

          ##Features
          * **Create/Destroy**: Launch and create an aggregated logging stack to support the entire OCP cluster.
          * **Simplified Configuration**: Configure your aggregated logging cluster's structure like components and end points easily.
          * Configure log forwarding to an in-cluster, Red Hat managed log storage solution based on LokiStack or Elasticsearch

          ## Prerequisites
          ### Red Hat OpenShift Logging Namespace
          Cluster logging and the Red Hat OpenShift Logging Operator is only deployable to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - fluentd
        - logging
        - aggregated
        - efk
        - vector
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Fluentd
          url: https://www.fluentd.org/
        - name: Vector
          url: https://vector.dev/
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:54054faf0d2f375dd6d7fc7929a2dbb529508e1b3c004a56fc97c2a682d16be2
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:54e3a5efaae5ecc2a08db03a421aa79e323b48c764115ea84d83a37c21d1d997
        - registry.redhat.io/openshift-logging/fluentd-rhel9@sha256:a676c9164ec2523b2002dc5eba234ce774926dbf7fe87de6d72b53f99cbf1db5
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:865572275b79462f14e84c69438a61afe9465cfe45ec164caf30965730b276f7
        - registry.redhat.io/openshift-logging/logging-view-plugin-rhel9@sha256:8023b072b45e9a21ddef83a56f4760d476053250d5a920d8eb5f03eecab8b1b3
        version: 5.9.15
      entries:
      - name: cluster-logging.v5.9.15
        version: 5.9.15
      - name: cluster-logging.v5.9.14
        version: 5.9.14
      - name: cluster-logging.v5.9.13
        version: 5.9.13
      - name: cluster-logging.v5.9.12
        version: 5.9.12
      - name: cluster-logging.v5.9.11
        version: 5.9.11
      - name: cluster-logging.v5.9.10
        version: 5.9.10
      - name: cluster-logging.v5.9.9
        version: 5.9.9
      - name: cluster-logging.v5.9.8
        version: 5.9.8
      - name: cluster-logging.v5.9.7
        version: 5.9.7
      - name: cluster-logging.v5.9.6
        version: 5.9.6
      - name: cluster-logging.v5.9.5
        version: 5.9.5
      - name: cluster-logging.v5.9.4
        version: 5.9.4
      - name: cluster-logging.v5.9.3
        version: 5.9.3
      - name: cluster-logging.v5.9.2
        version: 5.9.2
      - name: cluster-logging.v5.9.1
        version: 5.9.1
      - name: cluster-logging.v5.9.0
        version: 5.9.0
      name: stable-5.9
    - currentCSV: cluster-logging.v6.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.14-v4.17
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:3e36ea9a808945d3817d169f736b1758875f2fbc6a7ae845266e1545a1b713c7
          createdAt: "2025-05-14T12:32:39Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.8.0-0 <6.0.10'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.0
          operators.operatorframework.io.bundle.channels.v1: stable-6.0
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:3e36ea9a808945d3817d169f736b1758875f2fbc6a7ae845266e1545a1b713c7
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:4e9936c090818d1d8f812487a9a8218fda5b11f8c06167fffe35bae1bafcadf9
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:15487d7b9d8f83d0df0b6fbf380f20df3518dfb1b831fac4c3209e811494b436
        version: 6.0.10
      entries:
      - name: cluster-logging.v6.0.10
        version: 6.0.10
      - name: cluster-logging.v6.0.9
        version: 6.0.9
      - name: cluster-logging.v6.0.8
        version: 6.0.8
      - name: cluster-logging.v6.0.7
        version: 6.0.7
      - name: cluster-logging.v6.0.6
        version: 6.0.6
      - name: cluster-logging.v6.0.5
        version: 6.0.5
      - name: cluster-logging.v6.0.4
        version: 6.0.4
      - name: cluster-logging.v6.0.3
        version: 6.0.3
      - name: cluster-logging.v6.0.2
        version: 6.0.2
      - name: cluster-logging.v6.0.1
        version: 6.0.1
      - name: cluster-logging.v6.0.0
        version: 6.0.0
      name: stable-6.0
    - currentCSV: cluster-logging.v6.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.15-v4.18
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f7101c962de4e5643bd7f760104a41cc48f020889159a97e2d6711703fbbec64
          createdAt: "2025-05-14T12:38:55Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.9.0-0 <6.1.8'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.1
          operators.operatorframework.io.bundle.channels.v1: stable-6.1
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f7101c962de4e5643bd7f760104a41cc48f020889159a97e2d6711703fbbec64
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:96e438dd5c080cfa47080410f375dc71da24f6d046cf609db41c8b873d85404f
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:76037c744b34b1c4fccf18f17888813d9670bb6c061c2c3de843aa43c0497170
        version: 6.1.8
      entries:
      - name: cluster-logging.v6.1.8
        version: 6.1.8
      - name: cluster-logging.v6.1.7
        version: 6.1.7
      - name: cluster-logging.v6.1.6
        version: 6.1.6
      - name: cluster-logging.v6.1.5
        version: 6.1.5
      - name: cluster-logging.v6.1.4
        version: 6.1.4
      - name: cluster-logging.v6.1.3
        version: 6.1.3
      - name: cluster-logging.v6.1.2
        version: 6.1.2
      - name: cluster-logging.v6.1.1
        version: 6.1.1
      - name: cluster-logging.v6.1.0
        version: 6.1.0
      name: stable-6.1
    - currentCSV: cluster-logging.v6.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.16-v4.19
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
          createdAt: "2025-07-29T15:25:15Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0-0 <6.2.5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.2
          operators.operatorframework.io.bundle.channels.v1: stable-6.2
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Installing logging](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html-single/logging/index#installing-logging-6-2)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:c76c853565fc93da88ba0116425839cfb86cab63cbde9856d7cca78278d7758a
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:502a5d2891049bbc32115fda28086841351cd9e6a3110b4c3c412f683f2b2491
        version: 6.2.5
      entries:
      - name: cluster-logging.v6.2.5
        version: 6.2.5
      - name: cluster-logging.v6.2.4
        version: 6.2.4
      - name: cluster-logging.v6.2.3
        version: 6.2.3
      - name: cluster-logging.v6.2.2
        version: 6.2.2
      - name: cluster-logging.v6.2.1
        version: 6.2.1
      - name: cluster-logging.v6.2.0
        version: 6.2.0
      name: stable-6.2
    defaultChannel: stable-6.2
    packageName: cluster-logging
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhacs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhacs-operator.v3.74.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
          createdAt: "2024-02-06T23:06:13.271846+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.73.0 < 3.74.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:2c0840f4573373a19eb8c9a26501138ef0e76efbd508fd69a6260e78194d86c8
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:0482d4464d6d39c01abab8df882604dd227c52b5420d36fe7414aaff8067cbec
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:71b9fdcaf4cc761b1499be8e6ccff6c463cc6b9856e9deb1f93baad45ea59eb0
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:74624389ac3261ed699b1f64e5717353da68895709d30a011bc8813b61a4a249
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:120b9891ee9db839ab0b2029a62e67ec4747c47c6ffe15af38e9035160f57b78
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:22bf00e1457b4a9e63901892aa4375a1d4a9d1b2c067c1f054d918a24b222b10
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:a108e9a7967001e0bdfed7124b01e9a3dbf1dec308444762973030aec57e1d5a
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:87eb62fe8aa7952d0d69dc13c4867ac2f51e821f56d149a9b588c826caec4c16
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d546f2d3f668ce2f20e594394ade0a86d808f6c8c88cadbba227083d4abac940
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:264102d4bca7d9bd384dc04e5bbff1a7b43343ea8e59cc79047cdda50a3f299f
        version: 3.74.9
      entries:
      - name: rhacs-operator.v3.74.9
        version: 3.74.9
      - name: rhacs-operator.v3.74.8
        version: 3.74.8
      - name: rhacs-operator.v3.74.7
        version: 3.74.7
      - name: rhacs-operator.v3.74.6
        version: 3.74.6
      - name: rhacs-operator.v3.74.5
        version: 3.74.5
      - name: rhacs-operator.v3.74.4
        version: 3.74.4
      - name: rhacs-operator.v3.74.3
        version: 3.74.3
      - name: rhacs-operator.v3.74.2
        version: 3.74.2
      - name: rhacs-operator.v3.74.1
        version: 3.74.1
      - name: rhacs-operator.v3.74.0
        version: 3.74.0
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: latest
    - currentCSV: rhacs-operator.v3.62.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:56905c2e7df0e29ca2b6301cd81b4f9bbe02b483374708f7f7da219a4b29428d
          createdAt: "2021-08-13T17:10:01.410173+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.61.0 < 3.62.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.5.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main page before deploying, or consult the RHACS
              documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create an cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead. If you want to deploy Secured Cluster Service in another namespace, additionally pass the `-n <namespace>` flag.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e30f7835cf52dbce9afdbaf55e5a93d8baf9c4b3a3ae488d1d42c1adba7f46af
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:56905c2e7df0e29ca2b6301cd81b4f9bbe02b483374708f7f7da219a4b29428d
        - registry.redhat.io/rh-acs/collector@sha256:98cfab15684506f1d0978eb701e8de478d91fc5bda3013ffd3b436aa95f07cc0
        - registry.redhat.io/rh-acs/main@sha256:9932304bb62b084ee1ae7e75bb21d6dfb737260c851f2d8daf05082f79230c66
        - registry.redhat.io/rh-acs/scanner@sha256:266a575c2547a3223890e8b73db0f14d69dc5a22217c7e4e3dc79644e6ab0258
        - registry.redhat.io/rh-acs/scanner-db@sha256:ff877f503e08ecca0ce7a7d8db760b2e4004e8e8ae2f27fb9b9c00a05b58c93a
        - registry.redhat.io/rh-acs/collector@sha256:558c0dabdc52efaee7bc889ef397ac2ac30f7c54562c2cc0070d0ef991a8b31b
        version: 3.62.1
      entries:
      - name: rhacs-operator.v3.62.1
        version: 3.62.1
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.62
    - currentCSV: rhacs-operator.v3.64.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fa478167f2f738f244bed6a06450679be73f8464c40f86acde4ffd390505692b
          createdAt: "2021-09-22T05:10:01.286657+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.63.0 < 3.64.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0cacd00c0a147cc8295138565ef2dc2ddbd9ad1fb25f6be817719a75ea4c6448
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fa478167f2f738f244bed6a06450679be73f8464c40f86acde4ffd390505692b
        - registry.redhat.io/rh-acs/scanner@sha256:8b7121bce8f9c7133d29245ebb265a4b8f9831a66a57f11b2d68b9b8b5cf79cc
        - registry.redhat.io/rh-acs/scanner-db@sha256:67e81bf896267b7acab8cb6755cde552173de4f0881c38eaca8e491bd926cc53
        - registry.redhat.io/rh-acs/collector@sha256:2b3a8bd3af62516a6714fe612e5fdc0433b91d873182e290d3ddc2b32ce291f0
        - registry.redhat.io/rh-acs/collector@sha256:3eb099da081cf208d46be5ccfa7b5e9ac0e2cad49f127fa4db91e49e3f26db42
        - registry.redhat.io/rh-acs/main@sha256:6aeb2223124540ac217c38b016dea99d770ab897dde4906a821d8776d91ea03d
        version: 3.64.2
      entries:
      - name: rhacs-operator.v3.64.2
        version: 3.64.2
      - name: rhacs-operator.v3.64.1
        version: 3.64.1
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.64
    - currentCSV: rhacs-operator.v3.65.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:229c7e5f967c70dfbc3cf638d521267ea6acf875b055499dfd30cb19c0b1f69c
          createdAt: "2021-09-21T13:26:28.711610+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.64.0 < 3.65.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/collector@sha256:3eb099da081cf208d46be5ccfa7b5e9ac0e2cad49f127fa4db91e49e3f26db42
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0cacd00c0a147cc8295138565ef2dc2ddbd9ad1fb25f6be817719a75ea4c6448
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:229c7e5f967c70dfbc3cf638d521267ea6acf875b055499dfd30cb19c0b1f69c
        - registry.redhat.io/rh-acs/main@sha256:e672ef9a1239065f8cf0f331757861fea082e971a9e9760d1e65795f399cbde1
        - registry.redhat.io/rh-acs/scanner@sha256:bfe9b2205bfca6aebed14827b4bda098bfad574e67ffa967ad339543062200fe
        - registry.redhat.io/rh-acs/scanner-db@sha256:20f258500ac2ba4e626df92dbcd193cf7ab8687ac251591be2442b8a46871835
        - registry.redhat.io/rh-acs/collector@sha256:2b3a8bd3af62516a6714fe612e5fdc0433b91d873182e290d3ddc2b32ce291f0
        version: 3.65.1
      entries:
      - name: rhacs-operator.v3.65.1
        version: 3.65.1
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.65
    - currentCSV: rhacs-operator.v3.66.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:10cdb0ff3e67b65737cde420536494bd84c2b7ccf65192bbed2eb43eaed66a41
          createdAt: "2021-10-20T16:38:38.091764+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.65.0 < 3.66.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bd74f0b95ff2881faf6eb4520feecd067b69a759f09bbfecd467447ce2a2a25e
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:10cdb0ff3e67b65737cde420536494bd84c2b7ccf65192bbed2eb43eaed66a41
        - registry.redhat.io/rh-acs/main@sha256:97230c01e76d90c7fe35ffeaa591be3d9175f350a8fd88599aded09a1edf930c
        - registry.redhat.io/rh-acs/scanner@sha256:323ee86d6a9cbeb32270c659c1154f3a62c36535c6ea894e4e29e93eb1f4a84f
        - registry.redhat.io/rh-acs/scanner-db@sha256:490b0bf49161e4bc6de2504726aaf360e0ca9a214aea6b886fff13153bdb263d
        - registry.redhat.io/rh-acs/collector@sha256:5673c27554a71502e06e08945a4424c6ce5da27f489854f3aa452e6d31240aee
        - registry.redhat.io/rh-acs/collector@sha256:d93e13ee1d1cc4b7f82e644a7d43e560aac42bd1e06d25b476cea9a4338bf9ce
        version: 3.66.1
      entries:
      - name: rhacs-operator.v3.66.1
        version: 3.66.1
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.66
    - currentCSV: rhacs-operator.v3.67.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:893ebcb2d0c96647e56bce3b16fe32851df59cd71f2c0fc8f6e66ca3a669bb70
          createdAt: "2021-12-16T12:54:10.117625+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.66.0 < 3.67.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/scanner@sha256:2860ed271da6064149bb53568c5ad60e129918d55a2f244f9ade4a0bea4c4933
        - registry.redhat.io/rh-acs/scanner-db@sha256:6f8d47277732eb56317d576c2d06fb1f539d46d16520e6ca34abec290392cef9
        - registry.redhat.io/rh-acs/collector@sha256:57c2dfa3dbc8368076291fbc627baa01dcb5ebc50f838e02ce7ac38a128d027b
        - registry.redhat.io/rh-acs/collector@sha256:e4648723ccbd43fa5acf005d9734bfc06d41e19e1be0bde839880e28fa361e16
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86e5fa1fa294987114be200890c2e516501e424aee0fb98ece25c95e7716295b
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:893ebcb2d0c96647e56bce3b16fe32851df59cd71f2c0fc8f6e66ca3a669bb70
        - registry.redhat.io/rh-acs/main@sha256:8db7630d094b661411ebb3f35248848d3147c735677d0922f7a04de9a41eacc9
        version: 3.67.2
      entries:
      - name: rhacs-operator.v3.67.2
        version: 3.67.2
      - name: rhacs-operator.v3.67.1
        version: 3.67.1
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.67
    - currentCSV: rhacs-operator.v3.68.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:195a694dc96096166fe82e81ecb3ceb493f96cfb5029112e576e44ab39b56abd
          createdAt: "2022-06-15T22:13:22.427644+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.67.0 < 3.68.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:195a694dc96096166fe82e81ecb3ceb493f96cfb5029112e576e44ab39b56abd
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:f0d2f5f11c2a96d34d6b4baf249cbf821fe064eade73c3f95c22f3c6981ac4d1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:105d10bd50a6062c998ed107dea388dd78f0caddb0b57713c8692b1d07042774
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:081f2198ad70bfb2ec66153b50b48b950b9a5550ba996de413c5a2aff4052783
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c9d0a5a2871c24c0b1058e419f2b090429363a319856397855412fd2c0037395
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:9b9b9146d9683d6da0ff2abbc8ea82b96d945b7a8c3740d72d8c018fb2b04fba
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:0bb76363a3a25916dd4493e177e9b2d36eac9633473b07fe0fffb853fd072da3
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:711293438bab35b17d2e1f162e7f74f667a6e16dd32e45122f917b6817fefb4e
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:b5fc3dcaa72bbc8dea005189c940e1cf2d00c94a6cf7676cf0e1790e1191d45c
        version: 3.68.2
      entries:
      - name: rhacs-operator.v3.68.2
        version: 3.68.2
      - name: rhacs-operator.v3.68.1
        version: 3.68.1
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.68
    - currentCSV: rhacs-operator.v3.69.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9b2beceb255124855ba1b352d16009b2f2cd422901a19c311f0783dab1fcc3a0
          createdAt: "2022-06-21T09:07:40.705944+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.68.0 < 3.69.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:08a7d8c892fad69c589b81ad8fbd21d36c8fa4078530dba361cfa9a8f9f05950
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:9ae03669cfdc0f4fe56b021d911edb17159c07757720ae30a1f22d7f7a897c07
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:f167e928740d1c614e26948ea58f49e569f6a4fddc438cd0abe52f60538432d0
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:867613178d491f416e57975843817251490c5fbca2fc69d8b55b169bea675279
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:8db5206532e95a7ad7cb23f389155d41757aaf5f445ddce92b008a06627d3fd5
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:77f5db63e17b140a8d6b84e3822fd77463035be9c0c832b55b4977696da13ccc
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:da6dd486a20fdba61ea12d2c77ea299d9795a9f22d9057d65ff992b1f5896e07
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:b254746ed3faa1f1cbcafdf1343a1957bf0214311b36f4e892be3000cc2aecf3
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9b2beceb255124855ba1b352d16009b2f2cd422901a19c311f0783dab1fcc3a0
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:2afb2823edc6f3434e18db60ee37381b5b79f51234c1a261b6d0f8ed74cc8b63
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:cf5629823f2dd583495bcbbcf3b83b4f29f855ace3c388f5cb7b24a6d34ed267
        version: 3.69.2
      entries:
      - name: rhacs-operator.v3.69.2
        version: 3.69.2
      - name: rhacs-operator.v3.69.1
        version: 3.69.1
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.69
    - currentCSV: rhacs-operator.v3.70.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fb903bf9e110ba03e9c63977ebaf707e6884dd8fd413bcd6b793a0ed795cdd0e
          createdAt: "2022-06-20T12:30:26.999648+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.69.0 < 3.70.1'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:dfeb239931f7a33ade5c829a102cabc01f5c348e1aa35a7ddc24757cfd549629
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:df8cfdaa40b50ff2f2670d97c2115b47ae4bea443ae2831caaa468dd62f3c75f
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fb903bf9e110ba03e9c63977ebaf707e6884dd8fd413bcd6b793a0ed795cdd0e
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:6c32b86ed8bc18e384706eb6188406e66910bf34c8869bd155a60c3525d0fea7
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:d2e827a209840b7197960c8115720972237532eb9cf9d6fee7a36107053408ce
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:3d7c8b77671740e324ee6c309a84bbaa4d925e6750979fe4e46c561bc79f1b4d
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:b47c96af1c9ffe4d8728da510a93efe21adc221225fe2b822cfcace6c6672ba2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:08a7d8c892fad69c589b81ad8fbd21d36c8fa4078530dba361cfa9a8f9f05950
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:ee55a233d2a8fd77d89ec2c69f357721c7debead6470a9f494b3cd6a001829c4
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:6b14e850cda9a54618847858292ea669abff599a40692e27fed0e337c4a60526
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:e71f2f2347dd46dec790405b1bb98d54a348a4583949f142f5703be9a790f96e
        version: 3.70.1
      entries:
      - name: rhacs-operator.v3.70.1
        version: 3.70.1
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.70
    - currentCSV: rhacs-operator.v3.71.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:469b6c1592226eb9a521131ef758a33e6a6ac7c96bc67fcc903cea30469c125e
          createdAt: "2022-12-07T08:31:36.777654+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.70.0 < 3.71.3'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4a0dc8c337299d7f6440ff75c2d5219d3ab68183f4f117f008ef065757dd5d4b
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:469b6c1592226eb9a521131ef758a33e6a6ac7c96bc67fcc903cea30469c125e
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:f16f92e074f7125b88e9e2033c7c12b44421145b93d040034fa4688626869af1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:c3ec88cd99908894d0e67153779062f90ac1b007980d66040a79b2be13798a10
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:4d8bac9b3857af6ba351c7768f70ef30ef7cb99ae4df60cb226b5a793353a052
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:7f279bb6e0f34a5208d8711ba4d7b62c78f08b683815ce33f6aa28c23c4246da
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:9bc25e81c21cda6d8b535d8ac67c1286fe57a1596001850e730f8f73bf371111
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:e282a8e5e606003ea344d44ac5ee7699563848f7160eeae183aa58ad740e0e94
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:cb2d0f470968c2b725b0d176d423cc1866fd56893344d077d28f2c16581d3614
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:df7b41999b542a477848823a1d0e7e5c29ae1b00fbac035007a5a1f33fac139f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:5fdfd3690af60bd5ecf04741cd1f73259491da501700b066120976960567f0db
        version: 3.71.3
      entries:
      - name: rhacs-operator.v3.71.3
        version: 3.71.3
      - name: rhacs-operator.v3.71.2
        version: 3.71.2
      - name: rhacs-operator.v3.71.1
        version: 3.71.1
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.71
    - currentCSV: rhacs-operator.v3.72.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:3ff61e1cc77b66a22c8a8c0ab306f0e122cd74e762005285aede23ebb6c2f6a2
          createdAt: "2023-02-28T17:02:08.443613+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.71.0 < 3.72.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:575b6c3e2914751df1668ffff83cbda187ce6e56f32214414cdf036b7a62bbd8
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d56a00b3e20e60b7169f5c35d03fa6a1bd3b9980dd3ebc9d89d759e1d6d20ad2
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:cd475a6c2356d63cfe736a11114de7fedfbd8fdcfecc56224a00b6c819103be6
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:42e5f6c4ba789f124ac3ebe7b6d1a607ede8e763a35888c97ccbd0259bec4a1c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:648b14e4158da2a85eedb46b3456fb502d680d52d375ad34bf45b4b94f8b2ef6
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:3ff61e1cc77b66a22c8a8c0ab306f0e122cd74e762005285aede23ebb6c2f6a2
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:cd0c205ee44eddcdf99b84e80fcf4403851d956efc50532b9d12013bafe44988
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:0b69595405e4e598183bbf4197ac5b7593724fa6162413f225aa01a542af18c9
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:622f30b7f591194d86a5bf95deede0225e66f7b624db197f6289a5fbbbe887a9
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:c07fae3616cbc681e4d31839abea6623b4bf4dade93273201c8db7ba0d635aad
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:cb5df858e0c5cae8f77b285e23daeb83062bd72570c90ed5ea3ce4a02d2f5f08
        version: 3.72.4
      entries:
      - name: rhacs-operator.v3.72.4
        version: 3.72.4
      - name: rhacs-operator.v3.72.3
        version: 3.72.3
      - name: rhacs-operator.v3.72.2
        version: 3.72.2
      - name: rhacs-operator.v3.72.1
        version: 3.72.1
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.72
    - currentCSV: rhacs-operator.v3.73.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a8486b66f977a639d77ef110492e5ecdde47be8acaefbaa1cbb5417085c56839
          createdAt: "2023-05-26T21:27:04.033276+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.72.0 < 3.73.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:7bf9415aa4b0746d11ca5a59d4179bc382a0fac4ded969f0c4294171fde27d58
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:e6687c12d070069f655c17832d94c619c18274a259ccc59715036f9908e162f7
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:ee2a9e4ab74f5a554d1e4ff148833a53bdaab70ac1e2668cc9ee2fceca290ba9
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:082bd6f4b40209e0591be001dea358f328cd86a08691cdceb7f71ea59d861d9d
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:4d0ead8969f120a629104a92a454842bb7103aec137deb6c8209b7db9c48a1b2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:648b14e4158da2a85eedb46b3456fb502d680d52d375ad34bf45b4b94f8b2ef6
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a8486b66f977a639d77ef110492e5ecdde47be8acaefbaa1cbb5417085c56839
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:e2c89f735969d438f75a89ff63c404a8ec4361a03fde48e880fee73006b9c669
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:e4a27d3ffb173c87728fb637ca7fde92a9e78ea9ee3edcecccc2a5e3d3748b8c
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:dce5db4f162dd40c09cad114cb802cd6d7c72d0c6986581f2e74c977493f75b0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:c020a80ff82bf4d9b5cec3eeac5de9990de76041e22c444e664a20e15c46ce3b
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:5df359156b01505ce43e5f14c3b08722856bdb86f3d0d9b4fba785a2e580f346
        version: 3.73.5
      entries:
      - name: rhacs-operator.v3.73.5
        version: 3.73.5
      - name: rhacs-operator.v3.73.4
        version: 3.73.4
      - name: rhacs-operator.v3.73.3
        version: 3.73.3
      - name: rhacs-operator.v3.73.2
        version: 3.73.2
      - name: rhacs-operator.v3.73.1
        version: 3.73.1
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.73
    - currentCSV: rhacs-operator.v3.74.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
          createdAt: "2024-02-06T23:06:13.271846+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.73.0 < 3.74.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d546f2d3f668ce2f20e594394ade0a86d808f6c8c88cadbba227083d4abac940
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:264102d4bca7d9bd384dc04e5bbff1a7b43343ea8e59cc79047cdda50a3f299f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:0482d4464d6d39c01abab8df882604dd227c52b5420d36fe7414aaff8067cbec
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:2c0840f4573373a19eb8c9a26501138ef0e76efbd508fd69a6260e78194d86c8
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:74624389ac3261ed699b1f64e5717353da68895709d30a011bc8813b61a4a249
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:87eb62fe8aa7952d0d69dc13c4867ac2f51e821f56d149a9b588c826caec4c16
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:120b9891ee9db839ab0b2029a62e67ec4747c47c6ffe15af38e9035160f57b78
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:22bf00e1457b4a9e63901892aa4375a1d4a9d1b2c067c1f054d918a24b222b10
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:a108e9a7967001e0bdfed7124b01e9a3dbf1dec308444762973030aec57e1d5a
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:71b9fdcaf4cc761b1499be8e6ccff6c463cc6b9856e9deb1f93baad45ea59eb0
        version: 3.74.9
      entries:
      - name: rhacs-operator.v3.74.9
        version: 3.74.9
      - name: rhacs-operator.v3.74.8
        version: 3.74.8
      - name: rhacs-operator.v3.74.7
        version: 3.74.7
      - name: rhacs-operator.v3.74.6
        version: 3.74.6
      - name: rhacs-operator.v3.74.5
        version: 3.74.5
      - name: rhacs-operator.v3.74.4
        version: 3.74.4
      - name: rhacs-operator.v3.74.3
        version: 3.74.3
      - name: rhacs-operator.v3.74.2
        version: 3.74.2
      - name: rhacs-operator.v3.74.1
        version: 3.74.1
      - name: rhacs-operator.v3.74.0
        version: 3.74.0
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.74
    - currentCSV: rhacs-operator.v4.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9bf3a3e0fd573f91f03794eaf5291364a90392e5e7d4b5d4ed550ef4600ab826
          createdAt: "2023-10-20T07:04:50.531740+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.74.0 < 4.0.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:08f537cba262605a0eb0a510b6faf78e3ba26465f636d03cc1f7a6ed6b40d355
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:eee1aa054d4d15b23be293f31d3ea659e0ffc1d05174ef89a6c977dd3d4377cc
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:4c848eb7465e928251e05d32ffbc93cd14e7bb77972d22b8721bd6eaa4a93d2f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:3250d89e78de852a2627593f29469f0b3ff639ea3644b4b7d6d057d0c1ad3659
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:5a492b856e0b6d9fdb020b556d13186ef5e7095b0c736a2031500c9c00dad0c1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:139d80d398e7e5cfc85839e3039a3630b48c7ae6650d755754430fccd94d6a27
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0aedd8e865f72ffa8b79fd2bab48974caec3eea3e3a6f77c89fe6dacb8947057
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9bf3a3e0fd573f91f03794eaf5291364a90392e5e7d4b5d4ed550ef4600ab826
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:cfa10bb6b461f302674f185b26c8d6efaaeba7f3f65d2cf05ae1f069aa2710d3
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:c39621837cbc8419d187a43a9575ec8cb4021c58ea1b56f215003d7264eb5d03
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:09e79b5db8977520f9951cc43c339c8e40b4659d7da513347d5a88b2500f809d
        version: 4.0.5
      entries:
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.0
    - currentCSV: rhacs-operator.v4.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:8ffb6f0ff1bf016c3919c1eefc6c3db3d4a8fbe0e79919f72b4c281afe6d4b1c
          createdAt: "2024-01-13T18:55:22.011709+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 4.0.0 < 4.1.6'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:364c107c34694ca33ac724de78db3e41ee55f9098d9ade0775a51b11f1533065
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:8ffb6f0ff1bf016c3919c1eefc6c3db3d4a8fbe0e79919f72b4c281afe6d4b1c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a8ad0e76ce4cae9f9b36aecf125c4ed56fc2d776fe3df113d41bfc1c2f8b3602
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:8547b76cf0533dfe6da2aafbbd063021bad4ff78b32b2dfed4a7607496e0a904
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:8760d6dc68c86f0c5b53d8f0bfada0df012f80bc8e0023ba3522fe6829a05a41
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:762ad216e79be0837e98873202b0072c7fd399925ee94bde57cc61c68cbd5f4e
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:01b9f9798d3ecbefe3da1ac2ac782e8e8b930e7eeceb82fe3473b5bfc1ccd40a
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:7ac504b7824d17bd7ac55b1e59c26be308506f315a0ab0d8c45d3b4934c0a16c
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:2ec62f3da93a1346ad215744257820377ccf4ead84e7471c52e348a31a431d3c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:bdbe08ef2d01b114702780fab76905e6eb1851684692ffa0a4c84604bda3ddb3
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:34694b9f4fe7fffc0c547be4e311579b5bb0c1fa92d8257bf6d1581ffd655c46
        version: 4.1.6
      entries:
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.1
    - currentCSV: rhacs-operator.v4.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4253eca84cbb0c5fc033507ddc2deef235b054c0ba944fa271ebbba21545b872
          createdAt: "2024-03-13T14:51:55.084122+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.1.0 < 4.2.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:b5a15f2985b114bf9bc24a0d9942db5d0cb818e60fd15a2f9094aa00904cce77
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4253eca84cbb0c5fc033507ddc2deef235b054c0ba944fa271ebbba21545b872
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:12db245af3b8475c9b5680bb0035c566109819e05cfe62bd1dfdfb5d6ed25bea
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:b82f7f654235688fc07fb037422a73e1106e27024d39a68ef7fac1422b69496d
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:ff27f03064633d6d08f8383649069e39c868def4aeca8f2622750397e049d0ec
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:2f326cabf03f780a3e6120995d5442641764641f62430d82e05edbe21b30c7b2
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:3858ab936d4626a11bbf895d41bad181364ee5df3312f6dd29ebeb9b7d31ec77
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:e2b07cb7d690477d78a2b8e1875545998d4bdb946f80dd0533a14187da99dbbf
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:59f1862c8ce188906499f08201ae05deeb47bd4fc5eab4be0386ca44ec065da8
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:e79a5e5681ebfe2d6349e2763ee0c1fa521b20319846958e402d23229b1f257b
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:4b35f5f39865238a8e95d3cba7ebb5f15ccfbf6a2563c34f57d91a0b5c927066
        version: 4.2.5
      entries:
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.2
    - currentCSV: rhacs-operator.v4.3.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:266cbbc270efafa26d5d15f3e333c96eaedee3b47854ff337ae1c572d34901ce
          createdAt: "2024-06-06T22:31:19.365878+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.2.0 < 4.3.8'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c36a557c9dc2ae4479c432933a9a4dbef38f2be53360a68a5b6befb614d01dac
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:b0c9b3858fc0547dcb4eb3065dedd5bbcf1e1f6707a8ab812ed06882d4f6d6ad
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:b7eb17fbcf6c4a6e4205ad66f1ffd9b10a4634f6ef5045d407c266566b1f255d
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:bdf9785bff51194636c6e8a518448c6d12a297bd31c4bb2e79d07084409c18b4
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:266cbbc270efafa26d5d15f3e333c96eaedee3b47854ff337ae1c572d34901ce
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:9fba889021a876ab627ff8c496e4b6c46c221c61142cd0c76daf33c89762e2c1
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:10d0cc82b6eddd089f3a7af37b872c33248513d0b4d3f094b6f62c01f4460825
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:6b55a25eb3a611c648d9be883828834592229c16f4166fd28ef40a25fc667061
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:2c35cb6f518393d002edd2bf5f714c1d053cd0fab55f6152a2f35676e4979202
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:2bb643651c8d51525223df3b2b675947174bbe9f1eaea8688e1fa83044cfa41f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:5092c3fbb4a38c06a2f42d125dc718ef0e1005f304ad8046afdd2a5e7ccbc580
        version: 4.3.8
      entries:
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.3
    - currentCSV: rhacs-operator.v4.4.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:2b87569d7f4ef5e06c989b273c378e007c6d87b26eec41774ac1fc67dcd9e4e4
          createdAt: "2025-02-10T19:50:13.597164+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.3.0 < 4.4.8'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:65707fdad55767ff5598a8d3679fd0779a37f59501e58971b81bebcb23f1f9fc
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:d6b4f8caf342b105c8df14c5c9180cb36123e08d34756ee8fe4f2f76c34f8064
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:664a5dc40dc1fa06c47a11e13b0ab5e0612a981d7b69bcfb29422abf3d1a13b2
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:39cfd01236e905034f6645a73063427436c14e351e0e0bc735e73e62aa5a1c3c
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:467d069ee957aea07882b83b60ae9fcc62ccca159f5d643aac13e531ccf7df88
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:2b87569d7f4ef5e06c989b273c378e007c6d87b26eec41774ac1fc67dcd9e4e4
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:0df4e95c7406aee7812d145cdd4d8c61cec32857fd2079dbc4244954a9d8288a
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:183c75920ab8e7cb947cb966f1543e39b6fdb5e1bb0caa1b0cf5f89671f0328f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:76f9dc2e2ee5595c36530d97dec216e2dbefada2fe0747e3f6a3a725b75722ab
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:763e7568704971cb4656727dc223374de13e82b4304dd3f5387ddc0f4e42b8b6
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:69c12230a785619c765e3749a7d56f76dce9ce4f530c20e8a9387b99eb65627b
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:ba6e25a413c79131b966a097f72c36478933ea52deb55a62acda7011a96b3983
        version: 4.4.8
      entries:
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.4
    - currentCSV: rhacs-operator.v4.5.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a7dfbe3590668f91fe6694335b5a35b2fca07b4735348cc5202921755d2f654d
          createdAt: "2025-04-09T22:51:15.157672+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.4.0 < 4.5.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:5914548f6ce2160e92c5d96580ffba23cdf1ae940c1f73f3f6424d15c6e97c37
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:dc7f23341c8b209f1846f29f889d00ef232fb58da9d5b4bd561f3933093b2786
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:d928600f75461465598a10de8df2495a275fd5232a92b078f609d1f5774eb778
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:1817cf8d2c08b04391792321702ef53d16a4b1ca06459c00c9c053128f534358
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:c2656b794ec4236e36f48f71bf26bdedd9de27fc4acef39deba2bdcfcddab03b
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:cc4b329cc795752ecbb178d2e4efc152bde54b04181334eb5bd70304c3354cad
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:0ff948b5517df3773cae2df02b708c348d8707409899b1de4108588e31601c0c
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a7dfbe3590668f91fe6694335b5a35b2fca07b4735348cc5202921755d2f654d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:5109b304eb9458f97af85c2ba63e08f51672fa3f9406226b7f357b86b5828dc0
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:fde78d34ecd2944c7c16b4e823c76b3667bf6c5b468e9991128ecc4272ec0e2e
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:ed33e7396357f1db03a065573c020846a5872b12a21284439770686689222434
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:f07cba0738897e0e7ae74971be0470ce2a855657f133a6e1cdd8a7092f6151b0
        version: 4.5.9
      entries:
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.5
    - currentCSV: rhacs-operator.v4.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:c4d434e1ff052508e5594d2b95dafa8726fd8adf5317754774c013a46de224d8
          createdAt: "2025-09-26T09:24:28.046251+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.5.0 < 4.6.10'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:b5cd08349931669055f03d21331499ef2ea04597f8a90fb201615d43000271bc
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:ffe41363dc00b05b0daa0c6d16efb30e534086aa4e3d7ccc599c86d33813138c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:713e61d3c8b55de1edc9d7834d3e8576718e1327665f94a6b58178fa7496c0cc
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:65793c05aec8085a53fab8bcf33a85192c17b46c91a7d86e0a457d1e3a113878
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:899da9c23000bb821bc9d544ebb56bcda4f3245ceee5b1b9c12d09f580b4b9a1
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:8fad8ab2efdcbc99d9efa736a90cca6f236f0a9d45fdb355ae0b6b4dc42d77f0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:90c47c6f95571cfcca04aa4e05a18befad75c1b44a6f3a055fcdfdab749d29c5
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:16b61ca1b9ee4e0821e7fc82fe4def75ed22950e502730f5704ec4fe63bb2e59
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:c4d434e1ff052508e5594d2b95dafa8726fd8adf5317754774c013a46de224d8
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:b630d53ad0b862151be7e14f0483ca5adac570216deae66f60e75af87a4dfd46
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:19db8c5d2425dcf18ae63b025f946fc46adaec0578e9505415cb585797069e95
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:6da59c5b82625894f6efa3fa0bc5ea0378edd4f95cf42dee772e9c8a565dba81
        version: 4.6.10
      entries:
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.6
    - currentCSV: rhacs-operator.v4.7.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:5b60d8271d6fa70df3c8f7e0ecd0d64feecfdf674dc84036500fe5f145c9415a
          createdAt: "2025-09-26T11:05:05.900682+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.6.0 < 4.7.7'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:def71e1ddea5b8c1e54c8582f39dc9ecca855b38aa9423690150aa6ef0ff189f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8882c00c9be9b579654a9ee8cd0c38537cf92d2f4bd8ebded1678aa0b6ae827
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:34ea97c6e9fc2d23427d0c6f2628075c0eaa0e151ca1c381539b039c249217f7
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:98553e6f18d3c03a29212b4f7ea1d30a42bcf8b52ca12aecbcbc47769fc35b53
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:bac79889d4717233e89bb6c8fae6732ed7bec5b9a55fd3b3de1f1311d870d2b1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:6bb454d64a9e339b1aeb81ca17b26f33db4a84f19ced4ad81858e40fd01878dc
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:5b60d8271d6fa70df3c8f7e0ecd0d64feecfdf674dc84036500fe5f145c9415a
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:7596529cbdce7118226b79d69e0667ab35ee692569e4014b74abdf0fa7415222
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:5ed9902b52830be93436d7977c10e73d346b6290532f6d7245cfe7cca9c93b31
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:0c32318ee4cbd547e2c1fd64433cd23a629d1b06523d7c0f07ddb98c2baaa708
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:d8fcfe1267495801d23a0b94101c7b4d06b48346a949c97946ebcbffe148f4b6
        version: 4.7.7
      entries:
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.7
    - currentCSV: rhacs-operator.v4.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
          createdAt: "2025-09-12T11:06:11.361461+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.7.0 < 4.8.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:adf47cb6615d344533fcca0aee41774f5e8df9b952b9b89a57c2be2b7e0fc03c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:92ff7691963e620dc5bf6da0a7694e6bcd68c2545afc1642a59a443ba2ab5611
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:db691065e46f604b980b0b202f6c2e14c8b204094d784c4a15db1e4fecb2965e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:97df59280c12727a99fdec0c9bf56eb874e6a1cacd12d849709797d4be635ce9
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:e5c200f28187c1f6971562e41cae969183ee15f3d0feb076bb202b22ff67cbcb
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8964b1441251607b23f92918e9158724174013dc4af423988f56d6098dfb2d5
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:e949830f3034c41aa5434a3654b2b1ccba7dac0373bf43d7c50a9ea43bb686fb
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:16912df2c26e5fc2e6051c0501170722856db274415063608de716cf53851038
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:d1758b76c32f03b5a1a44951d18fb2bb0c4fd0a99d24f8c42fe9e6b059574889
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0c92d5b5508032d77559606927e45686179eaafc59e16d5496e088f2f212d258
        version: 4.8.4
      entries:
      - name: rhacs-operator.v4.8.4
        version: 4.8.4
      - name: rhacs-operator.v4.8.3
        version: 4.8.3
      - name: rhacs-operator.v4.8.2
        version: 4.8.2
      - name: rhacs-operator.v4.8.1
        version: 4.8.1
      - name: rhacs-operator.v4.8.0
        version: 4.8.0
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.8
    - currentCSV: rhacs-operator.v4.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
          createdAt: "2025-09-12T11:06:11.361461+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.7.0 < 4.8.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:e5c200f28187c1f6971562e41cae969183ee15f3d0feb076bb202b22ff67cbcb
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:e949830f3034c41aa5434a3654b2b1ccba7dac0373bf43d7c50a9ea43bb686fb
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:16912df2c26e5fc2e6051c0501170722856db274415063608de716cf53851038
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:adf47cb6615d344533fcca0aee41774f5e8df9b952b9b89a57c2be2b7e0fc03c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:97df59280c12727a99fdec0c9bf56eb874e6a1cacd12d849709797d4be635ce9
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:d1758b76c32f03b5a1a44951d18fb2bb0c4fd0a99d24f8c42fe9e6b059574889
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8964b1441251607b23f92918e9158724174013dc4af423988f56d6098dfb2d5
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0c92d5b5508032d77559606927e45686179eaafc59e16d5496e088f2f212d258
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:db691065e46f604b980b0b202f6c2e14c8b204094d784c4a15db1e4fecb2965e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:92ff7691963e620dc5bf6da0a7694e6bcd68c2545afc1642a59a443ba2ab5611
        version: 4.8.4
      entries:
      - name: rhacs-operator.v4.8.4
        version: 4.8.4
      - name: rhacs-operator.v4.8.3
        version: 4.8.3
      - name: rhacs-operator.v4.8.2
        version: 4.8.2
      - name: rhacs-operator.v4.8.1
        version: 4.8.1
      - name: rhacs-operator.v4.8.0
        version: 4.8.0
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: stable
    defaultChannel: stable
    packageName: rhacs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dns-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dns-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSRecord",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dns-operator",
                    "app.kubernetes.io/instance": "dnsrecord-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dnsrecord",
                    "app.kubernetes.io/part-of": "dns-operator"
                  },
                  "name": "dnsrecord-sample"
                },
                "spec": {
                  "endpoints": [
                    {
                      "dnsName": "dnsrecord-simple.example.com",
                      "recordTTL": 60,
                      "recordType": "A",
                      "targets": [
                        "52.215.108.61",
                        "52.30.101.221"
                      ]
                    }
                  ],
                  "providerRef": {
                    "name": "dns-provider-creds"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/dns-operator:v0.6.0
          createdAt: "2024-09-13T09:55:39Z"
          description: A Kubernetes Operator to manage the lifecycle of DNS resources
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kuadrant/dns-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSRecord is the Schema for the dnsrecords API
            displayName: DNSRecord
            kind: DNSRecord
            name: dnsrecords.kuadrant.io
            version: v1alpha1
        description: A Kubernetes Operator to manage the lifecycle of DNS resources
        displayName: DNS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dns
        - kuadrant
        links:
        - name: DNS Operator
          url: https://github.com/Kuadrant/dns-operator
        maintainers:
        - email: mnairn@redhat.com
          name: Michael Nairn
        - email: pbrookes@redhat.com
          name: Phil Brookes
        - email: cbrookes@redhat.com
          name: Craig Brookes
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/dns-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: dns-operator.v0.6.0
        version: 0.6.0
      name: stable
    defaultChannel: stable
    packageName: dns-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nexus-repository-ha-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-repository-ha-operator-certified.v3.84.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusRepo",
                "metadata": {
                  "name": "nexusrepo-sample"
                },
                "spec": {
                  "ingress": {
                    "additionalRules": null,
                    "annotations": null,
                    "defaultRule": false,
                    "dockerIngress": {
                      "annotations": null,
                      "enabled": false,
                      "host": "example.com",
                      "name": "nexus-docker-ingress"
                    },
                    "dockerSubdomain": false,
                    "enabled": false,
                    "host": "example.com",
                    "name": "nexus-ingress",
                    "tls": {
                      "enabled": false,
                      "secretName": "tlsSecretName"
                    }
                  },
                  "license": {
                    "fileContentsBase64": "your_license_file_contents_in_base_64",
                    "secretName": "nexus-repo-license"
                  },
                  "secret": {
                    "nexusSecret": {
                      "enabled" : false,
                      "name": "nexus-secret.json",
                      "nexusSecretsKeyId": "super_secret_key_id",
                      "secretKeyfileContentsBase64": "secretKeyfileContentsBase64",
                      "mountPath": "/var/nexus-repo-secrets"
                    }
                  },
                  "nexusData": {
                    "pvc": {
                      "accessMode": "ReadWriteOnce",
                      "size": "2Gi"
                    },
                    "storageClass": {
                      "enabled": false,
                      "name": "nexusrepo-storage",
                      "parameters": {},
                      "provisioner": "provisioner",
                      "reclaimPolicy": "Retain",
                      "volumeBindingMode": "WaitForFirstConsumer"
                    },
                    "volumeClaimTemplate": {
                      "enabled": false
                    }
                  },
                  "service": {
                    "docker": {
                      "enabled": false,
                      "name": "nexus-repo-docker-service",
                      "port": 9090,
                      "protocol": "TCP",
                      "targetPort": 9090,
                      "type": "NodePort"
                    },
                    "nexus": {
                      "enabled": false,
                      "name": "nexus-repo-service",
                      "port": 80,
                      "protocol": "TCP",
                      "targetPort": 8081,
                      "type": "NodePort"
                    }
                  },
                  "statefulset": {
                    "container": {
                      "containerPort": 8081,
                      "env": {
                        "clustered": false,
                        "install4jAddVmParams": "-Xms2703m -Xmx2703m",
                        "jdbcUrl": null,
                        "nexusInitialPassword": "nexusabc",
                        "password": "nexus",
                        "user": "nexus",
                        "zeroDowntimeEnabled": false
                      },
                      "imageName": "registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:fbad488ff2e368845473a11b4955992fe826a94811da3abf5cc3c6a33ff3974f",
                      "pullPolicy": "IfNotPresent",
                      "resources": {
                        "limits": {
                          "cpu": 4,
                          "memory": "8Gi"
                        },
                        "requests": {
                          "cpu": 4,
                          "memory": "8Gi"
                        }
                      },
                      "terminationGracePeriod": 120
                    },
                    "imagePullSecrets": {},
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 240,
                      "path": "/",
                      "periodSeconds": 60
                    },
                    "name": "nexusrepo-statefulset",
                    "readinessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 240,
                      "path": "/",
                      "periodSeconds": 60
                    },
                    "replicaCount": 1
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-certified@sha256:99f30aa5ba8469d96de3efe69d1b21e73ff022172fc79144bba496d1119ec8a2
          createdAt: "2023-08-14"
          description: |-
            Nexus Repository is the central source of control to efficiently manage all binaries
            and build artifacts across your DevOps pipeline.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.replaces: 3.80.0-2
          olm.skipRange: <3.84.0-1
          repository: https://github.com/sonatype/operator-nexus-repository
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus Repository
            displayName: NexusRepo
            kind: NexusRepo
            name: nexusrepos.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Repository is the central source of control to efficiently manage all binaries
          and build artifacts across your DevOps pipeline.
          The flow of open source and third-party components into and through an organization
          creates a complex software supply chain.
          Nexus Repository delivers speed, efficiency, and quality to the governance
          and management of all dependencies, libraries, and applications for your DevOps teams.

          ## Core Capabilities

          * **Dependency Management**:
            Improves reliability with repeatable, fast access to secure dependencies
          * **Developer Productivity**:
            Streamline developer workflows by enabling the sharing of components and applications across teams
          * **Supply Chain Performance**:
            Improve speed-to-market and reduced build times with release advanced staging and component tagging
          * **CI/CD Integrations**:
            Increase DevOps scalability with integrations to the most popular build and deployment tools

          Version control systems and package registries do not scale when managing proprietary,
          open source, and third-party components.
          Organizations need a central binary and build artifact repository to manage dependencies
          across the entire software supply chain.

          ## Limitations

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus Repository Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage
          By default, the example Custom Resource starts up a single instance of Nexus Repository in non-Pro mode using an H2 database.

          ### Connect to a PostgreSQL database
          We strongly recommend using Nexus Repository with a PostgreSQL database. To connect a Nexus Repository instance to a PostgreSQL database, do as follows:
          * Update the `license.fileContentsBase64` parameter with a base 64 representation of your Nexus Repository Pro license.
          * Set the `statefulset.container.env.jdbcUrl` parameter to your database url.
          * Set the `statefulset.container.env.user` parameter to your database user name.
          * Set the `statefulset.container.env.password` parameter to your database password.

          ### High Availability (HA) mode
          To start Nexus Repository in HA mode, you must edit the Custom Resource YAML as follows:

          * Update the `license.fileContentsBase64` parameter with a base 64 representation of your Nexus Repository Pro license.
          * Set the `statefulset.container.env.jdbcUrl` parameter to your database url.
          * Set the `statefulset.container.env.user` parameter to your database user name.
          * Set the `statefulset.container.env.password` parameter to your database password.
          * Set the `statefulset.container.env.clustered` parameter to true.
          * Set the `statefulset.replicaCount` parameter to the number of Nexus Repository pods you require in your HA cluster
          * You must use a persistent volume as described in the 'Storage' section below.

          ### Storage
          The default configuration uses an emptyDir volume for storing Nexus Repository logs. However, we strongly recommend that
          you configure dynamic provisioning of persistent storage based on your deployment environment as explained below.

          #### Cloud deployments (AWS/Azure)
          * Ensure the appropriate Container Storage Interface (CSI) driver(s) are installed for your chosen OpenShift cloud deployment.
            Please, refer to OpenShift documentation for details on configuring CSI drivers.
          Note: if you're using Red Hat OpenShift on AWS (ROSA), CSI drivers for dynamically provisioning EBS volumes are installed
          by default; you should see associated storage classes for them in your cluster web console.

          #### On premise deployments
          1. Attach separate disks (i.e. separate from the root disk) to your worker nodes.
          2. Install the Local Storage Operator. Please refer to Open Shift persistent storage documentation:
             https://docs.openshift.com/container-platform/4.13/storage/persistent_storage/persistent_storage_local/persistent-storage-local.html
          and the Local Storage Operator documentation: https://github.com/openshift/local-storage-operator.
          3. Use the Local Storage Operator to automatically create persistent volumes for your chosen storage class name as documented here:
             https://docs.openshift.com/container-platform/4.13/storage/persistent_storage/persistent_storage_local/persistent-storage-local.html
          4. Specify the storage class name from step 3 as the value of the `nexusData.storageClass.name` property and also set `nexusData.volumeClaimTemplate.enabled` to true.

          #### Configuring the operator for dynamic persistent volume provisioning
          * Set the `nexusData.storageClass.name` parameter to a storage class name. This could be one of the default storage classes
            automatically created in your OpenShift cluster (e.g., if you're using ROSA) or one that you would like the operator to create.
          * If you would like to create a dedicated storage class (i.e., you don't want to use the default), then in addition to
            specifying a value for `nexusData.storageClass.name`, you must also set `nexusData.storageClass.enabled` parameter to `true`.
          * Set the `nexusData.volumeClaimTemplate.enabled` parameter to `true`.

          ### Additional configuration
          * Once the operator has created the server instance and it is running, you can further configure Nexus Repository via the NexusRepo custom resource:

          | Parameter                                                   | Description                         | Default                                 |
          | ----------------------------------------------------------- | ----------------------------------  | ----------------------------------------|
          | `statefulset.name`                                          |  The suffix for the statefulset's service name    | nexusrepo-statefulset     |
          | `statefulset.replicaCount`                                  |  The desired number of Nexus Repository pods    | 3                           |
          | `statefulset.container.imageName`                           |  The Nexus repository image registry url    |
          | | | registry.connect.redhat.com/sonatype/nexus-repository-manager:<app version>-ubi-1|
          | `statefulset.container.resources.requests.cpu`              |  The minimum cpu the Nexus repository pod can request    | 4 |
          | `statefulset.container.resources.requests.memory`           |  The minimum memory the Nexus repository pod can request    | 8Gi|
          | `statefulset.container.resources.limits.cpu`                |  The maximum cpu the Nexus repository pod may get.    | 4|
          | `statefulset.container.resources.limits.memory`             |  The maximum memory the Nexus repository pod may get.  | 8Gi|
          | `statefulset.container.containerPort`                       |  The Nexus Repository container's HTTP port  | 8081|
          | `statefulset.container.pullPolicy`                          |  The Nexus Repository docker image pull policy  | IfNotPresent|
          | `statefulset.container.terminationGracePeriod`              |  The time given for the pod to gracefully shut down  | 120 seconds|
          | `statefulset.container.env.clustered`                       |  Controls whether Nexus Repository should start in clustered mode  | False|
          | `statefulset.container.env.install4jAddVmParams`            |  Xmx and Xms settings for JVM  | -Xms2703m -Xmx2703m|
          | `statefulset.container.env.user`                            |  The database user name | nexus (can be overridden via Kubernetes secret. See above)|
          | `statefulset.container.env.password`                        |  The database password  | nexus (can be overridden via Kubernetes secret. See above)|
          | `statefulset.container.env.jdbcUrl`                         |  The database url |
          | | | jdbc:postgresql://your_database_server:5432/nexus (must be overridden via Kubernetes |
          | | | secret object which feeds into env variable variables defined in the operator.yaml . See above)|
          | `statefulset.container.env.nexusInitialPassword`            |  The initial admin password for Nexus Repository UI. This is for first time |
          | | installs only. Subsequent installs/restarts on the same node ignores this setting | nexusabc|
          | `statefulset.livenessProbe.initialDelaySeconds`             |  The initial delay before the Nexus Repository application is pinged for a |
          | | response i.e. to determine whether the application is still alive | 240|
          | `statefulset.livenessProbe.path`                            |  The path to ping the application | |
          | `statefulset.livenessProbe.periodSeconds`                   |  The interval for checking that the application is still alive i.e. responsive, not deadlocked etc | 60|
          | `statefulset.livenessProbe.failureThreshold`                |  The maximum number of times the application ping can fail before the |
          | | application is considered unresponsive. Kubernetes will attempt to restart the application's pod in this case | 6|
          | `statefulset.readinessProbe.initialDelaySeconds`            |  The initial delay before the Nexus Repository application is checked |
          | | whether it is ready to receive traffic | 240|
          | `statefulset.readinessProbe.path`                           |  The path where the Nexus Repository application is checked for traffic readiness | |
          | `statefulset.readinessProbe.periodSeconds`                  |  The interval for checking the Nexus Repository application for traffic readiness | 60|
          | `statefulset.readinessProbe.failureThreshold`               |  The maximum number of times to check the Nexus Repository application for traffic readiness | 6|
          | `statefulset.imagePullSecrets`                              |  The pull secret for private image registries | {}|
          | `ingress.name`                                              |  The name of an Ingress object which listens for requests to Nexus Repository on port 80 | nexus-ingress|
          | `ingress.enabled`                                           |  Whether or not to create the Ingress | false|
          | `ingress.host`                                              |  Ingress host | "example.com"|
          | `ingress.defaultRule`                                       |  Whether or not to add a default rule for the Nexus Repository |
          | | Ingress which forwards traffic to a Service object  | false|
          | `ingress.additionalRules`                                   |  Additional rules to add to the ingress | null|
          | `ingress.tls.enabled`                                       |  Whether or not TLS should be enabled for the Ingress | false|
          | `ingress.tls.secretName`                                    | The name of a Secret object in which to store the TLS secret | tlsSecretName|
          | `ingress.dockerSubdomain`                                   |  Whether or not to add rules for docker subdomains | false|
          | `ingress.annotations`                                       |  Annotations for the Ingress object | null |
          | `license.fileContentsBase64`                                |  Your Base64 encoded Nexus Repository license | your_license_file_contents_in_base_64 |
          | `license.secretName`                                        |  The name of a secret object in which to store your |
          | | Base64-encoded Nexus Repository license | nexus-repo-license|
          | `ingress.dockerIngress.name`                                |  The name of an Ingress object which listens for requests to an |
          | | existing docker repository in Nexus Repository | nexus-docker-ingress|
          | `ingress.dockerIngress.enabled`                             |  Whether or not to create an Ingress for a docker repository | false|
          | `ingress.dockerIngress.host`                                |  The docker Ingress's host | "example.com"|
          | `ingress.dockerIngress.annotations`                         |  Annotations for the docker Ingress object | null |
          | `nexusData.pvc.size`                                        |  The volume size to request for storing Nexus logs | 2Gi |
          | `nexusData.pvc.accessMode`                                  |  The persistent volume claim access mode | ReadWriteOnce |
          | `nexusData.storageClass.name`                               |  The name of the storage class to use for dynamically requesting that |
          | | storage be provisioned when using dynamic provisioning or to randomly select a pre-created persisted volume when using the Local Storage Operator. | nexusrepo-storage |
          | `nexusData.storageClass.reclaimPolicy`                      |  The reclaim policy for persistent volume's created by the storage class | Retain |
          | `nexusData.storageClass.enabled`                            |  Whether or not a storage class object should be created. | false |
          | `nexusData.storageClass.volumeBindingMode`                  |  It controls when volume binding and dynamic provisioning should occur | WaitForFirstConsumer |
          | `nexusData.storageClass.provisioner`                        |  The name of the volume provisioner | provisioner |
          | `nexusData.storageClass.parameters`                         |  The additional parameters for the storage class | {} |
          | `nexusData.volumeClaimTemplate.enabled`                           |  Whether or not volume should be requested using a claim template. |
          | | You should set this property to true when using dynamic volume provisioning (cloud deployments) |
          | | or if you're using Local Storage Operator to automatically create persistent volumes for locally attached disks (i.e. not root disk) | false |
          | `license.secretName | A name for the Kubernetes secret object which stores the base64 representation of your |
          | | Nexus Repository license specified via `license.fileContentsBase64` is stored | nexus-repo-license |
          | `license.fileContentsBase64 | A base64 representation of your Nexus Repository license | your_license_file_contents_in_base_64 |
          | `service.nexus.name`                                              |  The name of a kubernetes Service Object which allows access to |
          | | Nexus Repository's main HTTP port | nexus-repo-service |
          | `service.nexus.enabled`                                           |  Whether or not to create the Service object | false|
          | `service.nexus.type`                                              |  The type of the Kubernetes Service | "NodePort"|
          | `service.nexus.protocol`                                       |  The protocol  | TCP |
          | `service.nexus.port`                                   |  The port to listen for incoming requests | 80|
          | `service.nexus.targetPort`                                   |  The port to forward requests to | 8081 |
          | `service.docker.name`                                              |  The name of a kubernetes Service Object which allows access to a |
          | | docker repository within Nexus Repository | nexus-repo-docker-service |
          | `service.docker.enabled`                                           |  Whether or not to create a Kubernetes Service object for a given |
          | | docker repository within Nexus Repository | false |
          | `service.docker.type`                                              |  The type of the Kubernetes Service | "NodePort"|
          | `service.docker.protocol`                                       |  The protocol  | TCP |
          | `service.docker.port`                                   |  The port to listen for incoming requests | 9090 |
          | `service.docker.targetPort`                                   |  The port to forward requests to | 9090 |
        displayName: Nexus Repository HA Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - repository
        - sonatype
        links:
        - name: Nexus Repository Manager
          url: https://www.sonatype.com/product-nexus-repository
        maintainers:
        - email: support@sonatype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-certified@sha256:99f30aa5ba8469d96de3efe69d1b21e73ff022172fc79144bba496d1119ec8a2
        - registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:fbad488ff2e368845473a11b4955992fe826a94811da3abf5cc3c6a33ff3974f
        version: 3.84.0-1
      entries:
      - name: nexus-repository-ha-operator-certified.v3.84.0-1
        version: 3.84.0-1
      name: stable
    defaultChannel: stable
    packageName: nexus-repository-ha-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-kubernetes.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "ClusterMongoDBRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-enterprise"
                  },
                  "name": "clustermongodbrole-sample"
                },
                "spec": {
                  "db": "admin",
                  "role": "rootMonitor",
                  "roles": [
                    {
                      "db": "admin",
                      "role": "root"
                    },
                    {
                      "db": "admin",
                      "role": "clusterMonitor"
                    }
                  ]
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "my-replica-set"
                },
                "spec": {
                  "credentials": "my-credentials",
                  "members": 3,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "sample-sharded-cluster"
                },
                "spec": {
                  "configServerCount": 3,
                  "credentials": "my-credentials",
                  "mongodsPerShardCount": 3,
                  "mongosCount": 2,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "shardCount": 1,
                  "type": "ShardedCluster",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBMultiCluster",
                "metadata": {
                  "name": "multi-replica-set"
                },
                "spec": {
                  "clusterSpecList": [
                    {
                      "clusterName": "e2e.cluster1.mongokubernetes.com",
                      "members": 2
                    },
                    {
                      "clusterName": "e2e.cluster2.mongokubernetes.com",
                      "members": 1
                    },
                    {
                      "clusterName": "e2e.cluster3.mongokubernetes.com",
                      "members": 2
                    }
                  ],
                  "credentials": "my-credentials",
                  "duplicateServiceObjects": false,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBOpsManager",
                "metadata": {
                  "name": "ops-manager"
                },
                "spec": {
                  "adminCredentials": "ops-manager-admin",
                  "applicationDatabase": {
                    "members": 3,
                    "podSpec": {
                      "cpu": 1
                    }
                  },
                  "configuration": {
                    "mms.fromEmailAddr": "admin@thecompany.com"
                  },
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "version": "6.0.3"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBUser",
                "metadata": {
                  "name": "my-replica-set-x509-user"
                },
                "spec": {
                  "db": "$external",
                  "mongodbResourceRef": {
                    "name": "my-replica-set"
                  },
                  "roles": [
                    {
                      "db": "admin",
                      "name": "dbOwner"
                    }
                  ],
                  "username": "CN=my-replica-set-x509-user,OU=cloud,O=MongoDB,L=New York,ST=New York,C=US"
                },
                "status": {
                  "phase": "Updated"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/mongodb/mongodb-kubernetes@sha256:745cc2c6bb893c1cb1e75d79b6f3bfc7b125a6a62f7431bdc43cf6cacd373998
          createdAt: "2025-09-16T14:56:17Z"
          description: The MongoDB Controllers for Kubernetes enable easy deploys
            of MongoDB into Kubernetes clusters, using our management, monitoring
            and backup platforms, Ops Manager and Cloud Manager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mongodb/mongodb-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterMongoDBRole
            name: clustermongodbroles.mongodb.com
            version: v1
          - description: MongoDB Deployment
            displayName: MongoDB Deployment
            kind: MongoDB
            name: mongodb.mongodb.com
            version: v1
          - kind: MongoDBCommunity
            name: mongodbcommunity.mongodbcommunity.mongodb.com
            version: v1
          - description: MongoDB Multi Deployment
            displayName: MongoDB Multi Deployment
            kind: MongoDBMultiCluster
            name: mongodbmulticluster.mongodb.com
            version: v1
          - kind: MongoDBSearch
            name: mongodbsearch.mongodb.com
            version: v1
          - description: MongoDB x509 User
            displayName: MongoDB User
            kind: MongoDBUser
            name: mongodbusers.mongodb.com
            version: v1
          - description: MongoDB Ops Manager
            displayName: MongoDB Ops Manager
            kind: MongoDBOpsManager
            name: opsmanagers.mongodb.com
            version: v1
        description: |
          MongoDB Controllers for Kubernetes (MCK) is an open-source project. It represents a merge of the previous
          MongoDB Community Operator (MCO) and MongoDB Enterprise Kubernetes Operator (MEKO).

          This release brings MongoDB Community and Enterprise editions together under a single, unified operator,
          making it easier to manage, scale, and upgrade your deployments. While the first version simply brings
          the capabilities of both into a single Operator, future changes will build on this to more closely align
          how Community and Enterprise are managed in Kubernetes, to offer an even more seamless and streamlined experience.
          As an open-source project, it now allows for community contributions, helping drive quicker bug fixes and ongoing innovation.
          The MongoDB Controllers for Kubernetes enable easy deploys of MongoDB
          into Kubernetes clusters, using our management, monitoring and backup
          platforms, Ops Manager and Cloud Manager.

          ## Before You Start

          To start using the operator you''ll need an account in MongoDB Cloud Manager or
          a MongoDB Ops Manager deployment.

          * [Create a Secret with your OpsManager API key](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-operator-credentials/#procedure)

          * [Create a ConfigMap with your OpsManager project ID and URL](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-project-using-configmap/)

          By installing this integration, you will be able to deploy MongoDB instances
          with a single simple command.

          ## Required Parameters

          * `opsManager` or `cloudManager` - Enter the name of the ConfigMap containing project information
          * `credentials` - Enter the name of the Secret containing your OpsManager credentials
          * `type` - Enter MongoDB Deployment Types ("Standalone", "ReplicaSet", "ShardedCluster"

          ## Supported MongoDB Deployment Types ##

          * Standalone: An instance of mongod that is running as a single server and
          not as part of a replica set, this is, it does not do any kind of
          replication.

          * Replica Set: A replica set in MongoDB is a group of mongod processes that
          maintain the same data set. Replica sets provide redundancy and high
          availability, and are the basis for all production deployments. This section
          introduces replication in MongoDB as well as the components and architecture
          of replica sets. The section also provides tutorials for common tasks
          related to replica sets.

          * Sharded Cluster: The set of nodes comprising a sharded MongoDB deployment.
          A sharded cluster consists of config servers, shards, and one or more mongos
          routing processes. Sharding is a A database architecture that partitions
          data by key ranges and distributes the data among two or more database
          instances. Sharding enables horizontal scaling.

          ## Requirements for deploying MongoDB OpsManager

          * In order to deploy resources of type MongoDB OpsManager, you will need to
          create a secret containing the [credentials](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/plan-om-resource/#om-rsrc-prereqs)
          for the Global Owner user

          ## Security ##

          The operator can enable TLS for all traffic between servers and also between
          clients and servers. Before enabling `security.tls.enabled` to `true` you
          should create your certificates.  or you can leave the operator to create all
          the certificates for you. The operator ability to create certs is been
          deprecated due to Kubernetes API changes.

          For more information, please read the official MongoDB
          Kubernetes Operator  [docs](https://docs.mongodb.com/kubernetes-operator/stable/).
        displayName: MongoDB Controllers for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mongodb
        - databases
        - nosql
        links:
        - name: Documentation
          url: https://docs.opsmanager.mongodb.com/current/tutorial/install-k8s-operator/index.html
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: stable
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f2d9a3264a8814f33143b73c56624572fb80afeab184133d3a901713a31c1a87
        - quay.io/mongodb/mongodb-enterprise-server@sha256:376ad7e5fe42651d946cba5705c0eee44aa636c1b4ff258210e4ce2fe4205b4d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:fb8397cbcdbf7c33f6ad142492792771127f0d9a8c46388f350b6927405ee590
        - quay.io/mongodb/mongodb-agent@sha256:d8a2ef79d767f03162ec57fe028f9072294b292d376af8eb97d2967eb6d7047b
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:09788976bdc1327acbf18cedf8efd5fca2159e43b506faa40e03ee42ad94928a
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:cd869ff7d661059fad7593e30141d74e8cc751295b1579577f7a853396eb5d0b
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:78f11c9cd5e63f2b0e10c67dc1491f4ce1279fa90178c108814725414defab9a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:15e31285c8949a4666f7702df53416b0905d76945a28575a6028a876cc8a6f0f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1a77851485ad3b2566e765b4bb051ff0ddfa48703492cef6c6db021dde84d87f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:146c4f5dc443f94f54d72808b4e4df93bb904e23589b222d17700166780776b4
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa8683500e8d3632c4588bbcdddb4e3186a01f48aefe1b6d95f8860dae8b6c76
        - quay.io/mongodb/mongodb-kubernetes-init-appdb@sha256:afb0d463304fc0bfb3d8dade91459414daac6bf6fcf742347a2012cbc9fea354
        - quay.io/mongodb/mongodb-search@sha256:7149a7b6f1ee84646d3972837920f80fc5db49503c598406ce92e2f794bd5f0d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:85c1d669c073caf9cbaac815e4c8ea2ca18b71deefa03bf84cd6d8b8ebd53e12
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:5d299820bb903f0cdfdf2626a3c44099d2eaee16188cd8234d2ff3a57259b3a7
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5145c97d2281aabcd8538ff515c3833aa48acb66b0b739fb1881120c1985d88a
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:5e3982493e85ac1e40536758c3145b7c2fff9a7b4be9d053ee41201363a89920
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b7764c316eeacffd3f9ddae62d9fd2a74eb58484d2cfd5021e460095d604764f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d49417ea85ee1ddb84429552a5aac87a82c2843904d8c0300a43b657b7a49e0e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:19925e1297ab652db59c35f53e7b7c8b2b277f8ecdc4215682961aaf2b3e924e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7e5361c2a1604dec3fc671d629806fe6ddc6df524b69a441b86933a4b30093
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:adeab0dc766a27138db7df5f37cbb1da860c14179ef631b2e001e5bb875325e4
        - quay.io/mongodb/mongodb-agent@sha256:a8358ce465adf463d8e73916d7944765d7aaf2b0ed10bd3b8119a87fc9080667
        - quay.io/mongodb/mongodb-agent@sha256:6e826bbe7dd812c101464e75295370c4b10e4e9c31477db89cfceb96759853e9
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:85b4e954c9c4bbe1019774bad1e05e9c8dc175648b01d3badf0b64143f182019
        - quay.io/mongodb/mongodb-enterprise-server@sha256:bfbcfcfc5f0de22e20520b2d96f826638e424adb028a909ccc192013d11f5833
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a86c4d8a0b9ca3c20cc524809676eb03a944ada43378ec98c107cd72cc64700f
        - quay.io/mongodb/mongodb-agent@sha256:ee9f3906d893004bb2aaaa6412646f58288b0684502b9e90f33cafadcfce4349
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0fde236fe26d5b20210d6f246aadacd7845874712def516cbe2eec11d79d5181
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:702f529cb9d8d9042e54263f8bc37b40c062ac2695efc3e7150819779886e323
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f1e72679b2468aed2f3c3106568327911e6594c9ee15f7348e953eaa0d8fd44a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7885ab6e88214f9289d8babf5189136dde0dd4c84a15a5481ccd3ef512b2e66f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:d81eac2129b6ab8346be1749c56ea67f6da64d4ee0fb9f3ff4171ba0f61918eb
        - quay.io/mongodb/mongodb-agent@sha256:fa30370a75e8e4ac648650ceb39a7a259a754aab8a48d7e2f2801cd1b777e91b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:67238129df09d33ee4746ff1acb5515d699a29d9f42350eae3f3f69ccb7fbf54
        - quay.io/mongodb/mongodb-enterprise-server@sha256:523566f44cbb6bbf0cd12b1adae939a883f2254c78dc4b8c9d0098581337b1db
        - quay.io/mongodb/mongodb-kubernetes-database@sha256:f8c981a6223f70d7c6bbd30c503fc8b2497f351830b2b74c30e6f81863d78d2f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:33bbd29b74f95baad522b380d70bcc9f933a3fee57642c37b09376c86179c40f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c752a7f0726d21d55f204a359b85fa29cd727a5fe571b37289ae3ea52d6206b0
        - quay.io/mongodb/mongodb-agent@sha256:da77dcb60b927bf6059c2c9bf008f8c3c0eec45cf7d1b40c34a0c263e8cc18b8
        - quay.io/mongodb/mongodb-kubernetes-init-database@sha256:3ea9106c0e35c015bc8f15657c76d974bb1cadb9220dc58be2076c9699f26fb7
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b00ac773623e1bf422b47e6d783263d2d08737865abe9abbe106d86d1ea25165
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9bef77e905dadd1e61d0439dd53b21ed1f3606e49affd9ffc4ba8156578e8f0f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:37af0913742b6de339c7e664b30f35dcef04506529e4a09a9bbbdd57643f7149
        - quay.io/mongodb/mongodb-agent@sha256:2d5652d0e263aa94554e7a068c93e4fd94b84ec223f077361d030d172bc93f67
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8445d59d491a6d3890ea9c8b421214bcd0fdcfbc420ed40497a790b79f0bc89e
        - quay.io/mongodb/mongodb-agent@sha256:b387b7458f5640f907c9074fcb08ae4d4785be31d4bb65f2d7c459cd6762b2d6
        - quay.io/mongodb/mongodb-enterprise-server@sha256:720056b985c1a16d48b2aecb7a88a15cba3d8186e55b3c6c24729699a85c8232
        - quay.io/mongodb/mongodb-agent@sha256:b422829aac63f068aed4a6948fd14cbd3c3bfad6d47ab76523f6839bf20b0203
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c8aec83b76dfecc05b8f5d8f40fd61b24980c0287a07294f75f321c2d42c32b4
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa300abf739e26350a258f643fa8e57aedacd041055c7b923bd22b21ef554cdb
        - quay.io/mongodb/mongodb-agent@sha256:3fa5bcfda3228e0ea880434a60ced829e3872e74763b347636c18475d866d0e1
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3fe62f3ad52707ccf0e9993c69891e44c569a4b4e1a3244cf60e3b5fdde57bd2
        - quay.io/mongodb/mongodb-agent@sha256:6c23febaa9dd52bc29ac1e52d087fc4b2b47c578f4d6ac3a1674bb750a38cb07
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3add8b68bfbaa42596fe5dbbee4623ba5ff1a826b18bb71301525adf574dcb94
        - quay.io/mongodb/mongodb-enterprise-server@sha256:655896bac9e82de8e9729d59a36c3a31f66235f22b93449260a2ccb7709de084
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b9ab7147b59b62e40a69c6400f87d0c7faecfa5462bed6e68ca653b00e50b887
        - quay.io/mongodb/mongodb-enterprise-server@sha256:e7aeedb752dda77cf0e49df3430843b5a2bba64c8404628a983f46640d316808
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ef732e3eea91cc86caf2eabedc0c9115ed235a8d204f33e74bbf090a0b0259fb
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:895dbdfb87ddc2e9693671c7108a31ab28e16fec2e5be6d046bdd2e30319eade
        - quay.io/mongodb/mongodb-enterprise-server@sha256:454b1585c342375dea643b10ead688d0f62347841f5f30059fcd4f92e3297ee9
        - quay.io/mongodb/mongodb-kubernetes-init-ops-manager@sha256:93c4633cbc84e9bfa3563771ac2efbf5e6515f380dec2c085494cb8326546377
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9a3a6496be0d2d446e6d654eb56386c7726e0cd2d7232f6dcec579923edf0014
        - quay.io/mongodb/mongodb-agent@sha256:5b064c470335df840b050b714b28b75b6061444843250f1eb9215cb2b8c1a149
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7a93a0276531ff9be4c90bb8fe8d104e0a9e930c29792aafe03cc6a76a9fa89c
        - quay.io/mongodb/mongodb-enterprise-server@sha256:33a30d11969eb21952681108092fe7f916e25c8e1de43e6a4e40e126025a2bec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1d418c65d82ccad205b7889be4ba161d3cbfa8b1fb75b21c8fcf02de38fad9ca
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d323c1bcd14118cc537ac282996ea5a58756d20b6b73f1eec0efa8710d246077
        - quay.io/mongodb/mongodb-enterprise-server@sha256:172e338a43df01ee5408d8d1e3cb038d2c55e77761c1a2ac6ed37959d1525140
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7266b405ebb2dd24a9f7c12d3921024f547c3d8c01831e6bcfc1d551d63be1
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a48a4d5c275fd669a27469649fb39ea38c728dbcf9f0a9da4e7cf27a2ce912ed
        - quay.io/mongodb/mongodb-enterprise-server@sha256:11ab324698f919f84a5dab5a964d93fe63f98a48146bb3b86974cba7606d30ec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c48631742ad1fec8a9ee745fc3cc9ca1fe7d86ef7fa68ad8e8337c7a4566dc87
        - quay.io/mongodb/mongodb-agent@sha256:89819704d70a3b6b4d2358745a22060ef740b33ad5c6debf8e6ac4d4a6fd63ba
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:a4590af027182305913f0de1a2e37ad8612bd73b8b31f4df11068b80089021ff
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b8aa1edbc5fb02af312a863bc7db8535a282dfd692a07b1e321e5a1acbfeb6a1
        - quay.io/mongodb/mongodb-enterprise-server@sha256:dfd4277f0c16ed476ac7de51894739e8c8d13a65ea5024b9a97b3dce0f6c6975
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d2d8b59c1f39c41951e7878963f8496eb494688720366f8edc7060887cdbe6a0
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:3bbd3d7fd45d107621e7770f2b4c2a758e148cd472e0ee21719442f71bf6ade0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3316015ac0699d66ac73f714e5dc0784738654e0349e80b1e1f4e953b3c88964
        - quay.io/mongodb/mongodb-agent@sha256:5f5031e4c2a3729f618d026b0a7f34fd4509ea73d5a4e3068b551041dc291583
        - quay.io/mongodb/mongodb-enterprise-server@sha256:df6081325ea2fe01d1b27d0c70005dec8232aaf51a43d91661b722b8a1761263
        - quay.io/mongodb/mongodb-agent@sha256:6394391f78da5e3218bb2c694e7ec9f6682fd17a6dd7e7250b0c0b5e1729796d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5c8553e5ba115657fb878c0fe08704af9fd5decec572a87d89d2c8c156ffc9f9
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:fa4aa97767bd19cf3711a74bf21ecb9cbbc25fed6b6b09a2eb241ebd3fbc8f3a
        - quay.io/mongodb/mongodb-agent@sha256:6cc4e953df3f09d793f0dbc6e57e55a3c5277ef63177a6d8f691a41227facf64
        - quay.io/mongodb/mongodb-kubernetes@sha256:745cc2c6bb893c1cb1e75d79b6f3bfc7b125a6a62f7431bdc43cf6cacd373998
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d21c65780cadc9d8668d2c28935d492625d6993f13ba562722200a819281cbcc
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:439d6fefb469613432e2d05149dfc96c32b16469d743835d5f402a2ada8be8cc
        - quay.io/mongodb/mongodb-agent@sha256:af3de4f631d0b00cae7f7e80c3900378ba7e449e01830bf9dc643b873d2e6d33
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:f71b027fccf649f18f63f8df7a59b1bac421a2cdab6712e0c985a93c08ed970a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:778f9e29e36d6687cd17dd95ec1f8127d2a732bde1a2dd5821212eb55c1c998a
        version: 1.4.0
      entries:
      - name: mongodb-kubernetes.v1.4.0
        version: 1.4.0
      - name: mongodb-kubernetes.v1.3.0
        version: 1.3.0
      - name: mongodb-kubernetes.v1.2.0
        version: 1.2.0
      - name: mongodb-kubernetes.v1.1.0
        version: 1.1.0
      - name: mongodb-kubernetes.v1.0.1
        version: 1.0.1
      - name: mongodb-kubernetes.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-apigateway-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-apigateway-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Resource",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "RestAPI",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Stage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "VPCLink",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/apigateway-controller:1.3.1
          createdAt: "2025-09-19T21:13:12Z"
          description: AWS API Gateway controller is a service controller for managing
            API Gateway resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIIntegrationResponse represents the state of an AWS apigateway
              APIIntegrationResponse resource.
            displayName: APIIntegrationResponse
            kind: APIIntegrationResponse
            name: apiintegrationresponses.apigateway.services.k8s.aws
            version: v1alpha1
          - description: APIKey represents the state of an AWS apigateway APIKey resource.
            displayName: APIKey
            kind: APIKey
            name: apikeys.apigateway.services.k8s.aws
            version: v1alpha1
          - description: APIMethodResponse represents the state of an AWS apigateway
              APIMethodResponse resource.
            displayName: APIMethodResponse
            kind: APIMethodResponse
            name: apimethodresponses.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Authorizer represents the state of an AWS apigateway Authorizer
              resource.
            displayName: Authorizer
            kind: Authorizer
            name: authorizers.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Deployment represents the state of an AWS apigateway Deployment
              resource.
            displayName: Deployment
            kind: Deployment
            name: deployments.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Integration represents the state of an AWS apigateway Integration
              resource.
            displayName: Integration
            kind: Integration
            name: integrations.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Method represents the state of an AWS apigateway Method resource.
            displayName: Method
            kind: Method
            name: methods.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Resource represents the state of an AWS apigateway Resource
              resource.
            displayName: Resource
            kind: Resource
            name: resources.apigateway.services.k8s.aws
            version: v1alpha1
          - description: RestAPI represents the state of an AWS apigateway RestAPI
              resource.
            displayName: RestAPI
            kind: RestAPI
            name: restapis.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Stage represents the state of an AWS apigateway Stage resource.
            displayName: Stage
            kind: Stage
            name: stages.apigateway.services.k8s.aws
            version: v1alpha1
          - description: VPCLink represents the state of an AWS apigateway VPCLink
              resource.
            displayName: VPCLink
            kind: VPCLink
            name: vpclinks.apigateway.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon API Gateway resources in AWS from within your Kubernetes cluster.

          **About Amazon API Gateway**

          Amazon API Gateway is an AWS service for creating, publishing, maintaining, monitoring, and securing REST, HTTP, and WebSocket APIs at any scale. API developers can create APIs that access AWS or other web services, as well as data stored in the AWS Cloud. As an API Gateway API developer, you can create APIs for use in your own client applications. Or you can make your APIs available to third-party app developers.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon API Gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apigateway
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon API Gateway Developer Resources
          url: https://aws.amazon.com/api-gateway/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: apigateway maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/apigateway-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-apigateway-controller.v1.3.1
        version: 1.3.1
      - name: ack-apigateway-controller.v1.3.0
        version: 1.3.0
      - name: ack-apigateway-controller.v1.2.5
        version: 1.2.5
      - name: ack-apigateway-controller.v1.2.4
        version: 1.2.4
      - name: ack-apigateway-controller.v1.2.3
        version: 1.2.3
      - name: ack-apigateway-controller.v1.2.2
        version: 1.2.2
      - name: ack-apigateway-controller.v1.2.1
        version: 1.2.1
      - name: ack-apigateway-controller.v1.1.0
        version: 1.1.0
      - name: ack-apigateway-controller.v1.0.6
        version: 1.0.6
      - name: ack-apigateway-controller.v1.0.5
        version: 1.0.5
      - name: ack-apigateway-controller.v1.0.4
        version: 1.0.4
      - name: ack-apigateway-controller.v1.0.2
        version: 1.0.2
      - name: ack-apigateway-controller.v1.0.1
        version: 1.0.1
      - name: ack-apigateway-controller.v1.0.0
        version: 1.0.0
      - name: ack-apigateway-controller.v0.0.3
        version: 0.0.3
      - name: ack-apigateway-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-apigateway-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: geunwookim
      provider-url: ""
    name: ncn-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ncn-operator.v0.3.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-1"
                },
                "spec": null
              },
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "label-1"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
          - description: Label is the Schema for the labels API
            displayName: Label
            kind: Label
            name: labels.cxl.antline.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: jwahn@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Antline
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/rasen708/ncn-operator:v0.4.0
        version: 0.3.18
      entries:
      - name: ncn-operator.v0.3.18
        version: 0.3.18
      - name: ncn-operator.v0.3.14
        version: 0.3.14
      name: alpha
    - currentCSV: ncn-operator.v0.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-1"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-19T01:21:08Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: jwahn@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        provider:
          name: Antline
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/rasen708/ncn-operator:v0.2.2
        version: 0.2.2
      entries:
      - name: ncn-operator.v0.2.2
        version: 0.2.2
      - name: ncn-operator.v0.2.1
        version: 0.2.1
      - name: ncn-operator.v0.2.0
        version: 0.2.0
      name: beta
    - currentCSV: ncn-operator.v0.9.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cmmd.gw.com/v1",
                "kind": "CMMD",
                "metadata": {
                  "name": "my-app",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "allocateMode": "auto",
                  "enable": true,
                  "payload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "metadata": {
                      "name": "my-app"
                    },
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "name": "my-app"
                        }
                      },
                      "template": {
                        "metadata": {
                          "labels": {
                            "name": "my-app"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "quay.io/my-account/my-app:latest",
                              "name": "my-app-01"
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "cmmd.gw.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "machineconfiguration.openshift.io/role": "cmmd"
                  },
                  "name": "cmmd-node-label",
                  "namespace": "socmmd"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          console.openshift.io/plugins: '["socmmd-console-plugin"]'
          containerImage: quay.io/rasen708/ncn-operator@sha256:a09963745bd9c1419532d42d91237b7496c7a78e6a1c460a8e714ea1a0ce4d45
          createdAt: "2024-08-01T05:45:38Z"
          description: Provides the ncn extend
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "cmmd.gw.com/v1",
              "kind": "Label",
              "metadata": {
                "name": "cmmd-node-label",
                "namespace": "socmmd",
                "labels": {}
              },
              "spec": {}
            }
          operatorframework.io/suggested-namespace: socmmd-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/geunwookim/ncn-operator
          support: geunwookim
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CMMD is the Schema for the cmmds API
            displayName: CMMD
            kind: CMMD
            name: cmmds.cmmd.gw.com
            version: v1
          - description: Label is the Schema for the labels API
            displayName: Node Label
            kind: Label
            name: labels.cmmd.gw.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: geunwookim@gmail.com
          name: geunwookim
        maturity: betav1
        provider:
          name: geunwookim
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/rasen708/ncn-operator@sha256:a09963745bd9c1419532d42d91237b7496c7a78e6a1c460a8e714ea1a0ce4d45
        version: 0.9.8
      entries:
      - name: ncn-operator.v0.9.8
        version: 0.9.8
      - name: ncn-operator.v0.9.2
        version: 0.9.2
      - name: ncn-operator.v0.8.1
        version: 0.8.1
      - name: ncn-operator.v0.5.6
        version: 0.5.6
      - name: ncn-operator.v0.4.3
        version: 0.4.3
      - name: ncn-operator.v0.4.2
        version: 0.4.2
      - name: ncn-operator.v0.3.18
        version: 0.3.18
      - name: ncn-operator.v0.3.14
        version: 0.3.14
      name: betav1
    defaultChannel: betav1
    packageName: ncn-operator
    provider:
      name: geunwookim
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CNCF
      provider-url: ""
    name: etcd
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: etcdoperator.v0.6.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          description: etcd is a distributed key value store providing a reliable
            way to store data across a cluster of machines.
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
        description: |
          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader. Your applications can read and write data into etcd.
          A simple use-case is to store database connection details or feature flags within etcd as key value pairs. These values can be watched, allowing your app to reconfigure itself when they change. Advanced uses take advantage of the consistency guarantees to implement database leader elections or do distributed locking across a cluster of workers.

          _The etcd Open Cloud Service is Public Alpha. The goal before Beta is to fully implement backup features._

          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` or with the API using the automatically generated Kubernetes Service.

          [Read the complete guide to using the etcd Open Cloud Service](https://coreos.com/tectonic/docs/latest/alm/etcd-ocs.html)

          ### Supported Features
          **High availability**
          Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.
          **Automated updates**
          Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.
          **Backups included**
          Coming soon, the ability to schedule backups to happen on or off cluster.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: support@coreos.com
          name: CoreOS, Inc
        maturity: alpha
        provider:
          name: CoreOS, Inc
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:bd944a211eaf8f31da5e6d69e8541e7cada8f16a9f7a5a570b22478997819943
        version: 0.6.1
      entries:
      - name: etcdoperator.v0.6.1
        version: 0.6.1
      name: alpha
    - currentCSV: etcdoperator.v0.9.4-clusterwide
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdCluster",
                "metadata": {
                  "name": "example",
                  "annotations": {
                    "etcd.database.coreos.com/scope": "clusterwide"
                  }
                },
                "spec": {
                  "size": 3,
                  "version": "3.2.13"
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdRestore",
                "metadata": {
                  "name": "example-etcd-cluster-restore"
                },
                "spec": {
                  "etcdCluster": {
                    "name": "example-etcd-cluster"
                  },
                  "backupStorageType": "S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdBackup",
                "metadata": {
                  "name": "example-etcd-cluster-backup"
                },
                "spec": {
                  "etcdEndpoints": ["<etcd-cluster-endpoints>"],
                  "storageType":"S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
          createdAt: "2019-02-28 01:03:00"
          description: Create and maintain highly-available etcd clusters on Kubernetes
          repository: https://github.com/coreos/etcd-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to backup an etcd cluster.
            displayName: etcd Backup
            kind: EtcdBackup
            name: etcdbackups.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to restore an etcd cluster from a backup.
            displayName: etcd Restore
            kind: EtcdRestore
            name: etcdrestores.etcd.database.coreos.com
            version: v1beta2
        description: |
          The etcd Operater creates and maintains highly-available etcd clusters on Kubernetes, allowing engineers to easily deploy and manage etcd clusters for their applications.

          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader.


          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` via port forwarding:

              $ kubectl --namespace default port-forward service/example-client 2379:2379
              $ etcdctl --endpoints http://127.0.0.1:2379 get /

          Or directly to the API using the automatically generated Kubernetes Service:

              $ etcdctl --endpoints http://example-client.default.svc:2379 get /

          Be sure to secure your etcd cluster (see Common Configurations) before exposing it outside of the namespace or cluster.


          ### Supported Features

          * **High availability** - Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.

          * **Automated updates** - Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.

          * **Backups included** - Create etcd backups and restore them through the etcd Operator.

          ### Common Configurations

          * **Configure TLS** - Specify [static TLS certs](https://github.com/coreos/etcd-operator/blob/master/doc/user/cluster_tls.md) as Kubernetes secrets.

          * **Set Node Selector and Affinity** - [Spread your etcd Pods](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-node-selector-and-anti-affinity-across-nodes) across Nodes and availability zones.

          * **Set Resource Limits** - [Set the Kubernetes limit and request](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-resource-requirement) values for your etcd Pods.

          * **Customize Storage** - [Set a custom StorageClass](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#custom-persistentvolumeclaim-definition) that you would like to use.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: etcd-dev@googlegroups.com
          name: etcd Community
        maturity: alpha
        provider:
          name: CNCF
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
        version: 0.9.4-clusterwide
      entries:
      - name: etcdoperator.v0.9.4-clusterwide
        version: 0.9.4-clusterwide
      - name: etcdoperator.v0.9.2-clusterwide
        version: 0.9.2-clusterwide
      - name: etcdoperator.v0.9.0
        version: 0.9.0
      name: clusterwide-alpha
    - currentCSV: etcdoperator.v0.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "size": 3,
                  "version": "3.2.13"
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdRestore",
                "metadata": {
                  "name": "example-etcd-cluster-restore"
                },
                "spec": {
                  "etcdCluster": {
                    "name": "example-etcd-cluster"
                  },
                  "backupStorageType": "S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdBackup",
                "metadata": {
                  "name": "example-etcd-cluster-backup"
                },
                "spec": {
                  "etcdEndpoints": ["<etcd-cluster-endpoints>"],
                  "storageType":"S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
          createdAt: "2019-02-28 01:03:00"
          description: Create and maintain highly-available etcd clusters on Kubernetes
          repository: https://github.com/coreos/etcd-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to backup an etcd cluster.
            displayName: etcd Backup
            kind: EtcdBackup
            name: etcdbackups.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to restore an etcd cluster from a backup.
            displayName: etcd Restore
            kind: EtcdRestore
            name: etcdrestores.etcd.database.coreos.com
            version: v1beta2
        description: |
          The etcd Operater creates and maintains highly-available etcd clusters on Kubernetes, allowing engineers to easily deploy and manage etcd clusters for their applications.

          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader.


          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` via port forwarding:

              $ kubectl --namespace default port-forward service/example-client 2379:2379
              $ etcdctl --endpoints http://127.0.0.1:2379 get /

          Or directly to the API using the automatically generated Kubernetes Service:

              $ etcdctl --endpoints http://example-client.default.svc:2379 get /

          Be sure to secure your etcd cluster (see Common Configurations) before exposing it outside of the namespace or cluster.


          ### Supported Features

          * **High availability** - Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.

          * **Automated updates** - Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.

          * **Backups included** - Create etcd backups and restore them through the etcd Operator.

          ### Common Configurations

          * **Configure TLS** - Specify [static TLS certs](https://github.com/coreos/etcd-operator/blob/master/doc/user/cluster_tls.md) as Kubernetes secrets.

          * **Set Node Selector and Affinity** - [Spread your etcd Pods](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-node-selector-and-anti-affinity-across-nodes) across Nodes and availability zones.

          * **Set Resource Limits** - [Set the Kubernetes limit and request](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-resource-requirement) values for your etcd Pods.

          * **Customize Storage** - [Set a custom StorageClass](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#custom-persistentvolumeclaim-definition) that you would like to use.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: etcd-dev@googlegroups.com
          name: etcd Community
        maturity: alpha
        provider:
          name: CNCF
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
        version: 0.9.4
      entries:
      - name: etcdoperator.v0.9.4
        version: 0.9.4
      - name: etcdoperator.v0.9.2
        version: 0.9.2
      - name: etcdoperator.v0.9.0
        version: 0.9.0
      name: singlenamespace-alpha
    defaultChannel: singlenamespace-alpha
    packageName: etcd
    provider:
      name: CNCF
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataStax
      provider-url: ""
    name: cass-operator-community
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cass-operator.v1.23.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cassandra.datastax.com/v1beta1",
                "kind": "CassandraDatacenter",
                "metadata": {
                  "name": "dc1"
                },
                "spec": {
                  "clusterName": "development",
                  "config": {
                    "cassandra-yaml": {
                      "authenticator": "PasswordAuthenticator",
                      "authorizer": "CassandraAuthorizer",
                      "num_tokens": 16,
                      "role_manager": "CassandraRoleManager"
                    },
                    "jvm-server-options": {
                      "initial_heap_size": "1G",
                      "max_heap_size": "1G"
                    }
                  },
                  "managementApiAuth": {
                    "insecure": {}
                  },
                  "podTemplateSpec": {
                    "spec": {
                      "containers": [
                        {
                          "name": "cassandra"
                        }
                      ],
                      "securityContext": {}
                    }
                  },
                  "racks": [
                    {
                      "name": "rack1"
                    },
                    {
                      "name": "rack2"
                    },
                    {
                      "name": "rack3"
                    }
                  ],
                  "resources": {
                    "requests": {
                      "cpu": "1000m",
                      "memory": "2Gi"
                    }
                  },
                  "serverType": "cassandra",
                  "serverVersion": "4.1.2",
                  "size": 3,
                  "storageConfig": {
                    "cassandraDataVolumeClaimSpec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "server-storage"
                    }
                  }
                }
              },
              {
                "apiVersion": "control.k8ssandra.io/v1alpha1",
                "kind": "CassandraTask",
                "metadata": {
                  "name": "example-cleanup"
                },
                "spec": {
                  "datacenter": {
                    "name": "dc2",
                    "namespace": "cass-operator"
                  },
                  "jobs": [
                    {
                      "command": "cleanup",
                      "name": "cleanup-run"
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database,Big Data,OpenShift Optional
          certified: "true"
          containerImage: cr.k8ssandra.io/k8ssandra/cass-operator:v1.23.2
          createdAt: "2025-03-20"
          description: |
            Simple provisioning, turn-key operations, and automated remediation of Apache Cassandra clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/k8ssandra/cass-operator
          support: DataStax
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CassandraDatacenter is the Schema for the cassandradatacenters
              API
            displayName: Cassandra Datacenter
            kind: CassandraDatacenter
            name: cassandradatacenters.cassandra.datastax.com
            version: v1beta1
          - description: CassandraTask is the Schema for the cassandrajobs API
            displayName: Cassandra Task
            kind: CassandraTask
            name: cassandratasks.control.k8ssandra.io
            version: v1alpha1
        description: "## Apache Cassandra\n\nThe Apache Cassandra database is the
          right choice when you need scalability and\nhigh availability without compromising
          performance. Linear scalability and\nproven fault-tolerance on commodity
          hardware or cloud infrastructure make it the\nperfect platform for mission-critical
          data. Cassandra's support for replicating\nacross multiple datacenters is
          best-in-class, providing lower latency for your\nusers and the peace of
          mind of knowing that you can survive regional outages.\n\n## DataStax Enterprise\n\nThe
          most advanced distribution of Apache Cassandra™ on the market, with the\nenterprise
          functionality needed for serious production systems and backed up and\nsupported
          by the best distributed-experts in the world. It's one platform for\nall
          types of applications anywhere, any cloud, any model: key-value, graph,\ntabular,
          JSON.\n\nDataStax Enterprise is a fully integrated and optimized database,
          with graph,\nanalytics, and search included, all with a unified security
          model. Simply put,\nit's the only database capable of meeting today's demanding
          requirements\n\n## Operator Details\n\n`cass-operator` is designed as a
          modular operator for Apache Cassandra and\nderived  distributions. Apache
          Cassandra is a distributed database consisting of\nmultiple nodes working
          in concert to store data and process queries along a\nnumber of fault domains.
          `cass-operator` has the deployment of a Cassandra\ncluster around the logical
          domain of a datacenter with the `CassandraDatacenter`\ncustom resource.\n
          \   \nUpon submission of one of these resources it handles provisioning
          the underlying\nstateful sets (analogous to C\\* logical racks), services,
          and configuration.\nAdditionally through monitoring pod state via Kubernetes
          callbacks it handles day to day\noperations such as restarting failed processes,
          scaling clusters up, and deploying\nconfiguration changes in a rolling,
          non-disruptive, fashion.\n    \nThis operator is designed to be `Namespace`
          scoped. A single Kubernetes cluster may\nbe running multiple instances of
          this operator, in separate namespaces, to support\na number of C\\* clusters
          and environments. Configuration is simple with the usage of\nYAML based
          overrides in the Custom Resource paired with an `init` container.\n    \nIn
          C\\* clusters ordering and timing of certain operations are important to
          keep the system\nevenly distributed. `cass-operator` takes advantage of
          a sidecar process within the\nmain container to handle the orchestration
          of starting our main C* process."
        displayName: DataStax Kubernetes Operator for Apache Cassandra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - k8ssandra
        - cass-operator
        - DataStax Enterprise
        - Apache Cassandra
        - DSE
        - C*
        - Cassandra
        links:
        - name: DataStax
          url: https://datastax.com
        - name: Operator GitHub Repo
          url: https://github.com/k8ssandra/cass-operator
        - name: Config Builder GitHub Repo
          url: https://github.com/datastax/cass-config-builder
        - name: Documentation [1]
          url: https://github.com/k8ssandra/cass-operator/tree/master/docs/user
        - name: Documentation [2]
          url: https://docs.datastax.com/en/cass-operator/doc/cass-operator/cassOperatorTOC.html
        maintainers:
        - email: cass-operator@datastax.com
          name: DataStax Operator Team
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: DataStax
        relatedImages:
        - cr.k8ssandra.io/k8ssandra/cass-operator:v1.23.2
        version: 1.23.2
      entries:
      - name: cass-operator.v1.23.2
        version: 1.23.2
      - name: cass-operator.v1.23.1
        version: 1.23.1
      - name: cass-operator.v1.23.0
        version: 1.23.0
      - name: cass-operator.v1.21.1
        version: 1.21.1
      - name: cass-operator.v1.19.1
        version: 1.19.1
      - name: cass-operator.v1.17.1
        version: 1.17.1
      - name: cass-operator.v1.16.0
        version: 1.16.0
      - name: cass-operator.v1.14.0
        version: 1.14.0
      - name: cass-operator.v1.13.0
        version: 1.13.0
      - name: cass-operator.v1.10.4
        version: 1.10.4
      - name: cass-operator.v1.10.1
        version: 1.10.1
      name: stable
    defaultChannel: stable
    packageName: cass-operator-community
    provider:
      name: DataStax
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: openstack-ansibleee-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openstack-ansibleee-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ansibleee.openstack.org/v1beta1",
                "kind": "OpenStackAnsibleEE",
                "metadata": {
                  "name": "openstackansibleee-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-09-23T20:57:34Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.3
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenStackAnsibleEE is the Schema for the openstackansibleees
              API
            displayName: OpenStack Ansible EE
            kind: OpenStackAnsibleEE
            name: openstackansibleees.ansibleee.openstack.org
            version: v1beta1
        description: The OpenStack Ansible Execution Environment Operator
        displayName: OpenStackAnsibleEE
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - execution
        links:
        - name: OpenStack Ansible EE Operator
          url: https://github.com/openstack-k8s-operators/openstack-ansibleee-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:24b468b09b43dd6cf77f5b08a68f2c6359d115722d28982fd8cdb1923e8122de
        - registry.redhat.io/rhoso-operators/openstack-ansibleee-rhel9-operator@sha256:bf096518547e92fb6fe0ecd526813bdc4a15f928a6d89eacadb71ffaad91db69
        - registry.redhat.io/rhoso-operators/ee-openstack-ansible-ee-rhel9@sha256:640efe595fbdb993ecc1c55c14457edc2b0e3e1c6dde8ad9ea1acf4b4503339f
        version: 1.0.3
      entries:
      - name: openstack-ansibleee-operator.v1.0.3
        version: 1.0.3
      - name: openstack-ansibleee-operator.v1.0.2
        version: 1.0.2
      - name: openstack-ansibleee-operator.v1.0.1
        version: 1.0.1
      - name: openstack-ansibleee-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: openstack-ansibleee-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: ""
    name: red-hat-hawtio-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: red-hat-hawtio-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              },
              {
                "apiVersion": "hawt.io/v2",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24,
                    "internalSSL": true
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
          createdAt: "2025-08-07T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v2
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
        version: 1.2.1
      entries:
      - name: red-hat-hawtio-operator.v1.2.1
        version: 1.2.1
      - name: red-hat-hawtio-operator.v1.2.0-0.1754874242.p
        version: 1.2.0+0.1754874242.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754338551.p
        version: 1.2.0+0.1754338551.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754317392.p
        version: 1.2.0+0.1754317392.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1753863723.p
        version: 1.2.0+0.1753863723.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752679472.p
        version: 1.2.0+0.1752679472.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752452299.p
        version: 1.2.0+0.1752452299.p
      - name: red-hat-hawtio-operator.v1.2.0
        version: 1.2.0
      - name: red-hat-hawtio-operator.v1.1.1
        version: 1.1.1
      - name: red-hat-hawtio-operator.v1.1.0
        version: 1.1.0
      - name: red-hat-hawtio-operator.v1.1.0-0.1742200932.p
        version: 1.1.0+0.1742200932.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1740358952.p
        version: 1.1.0+0.1740358952.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739994212.p
        version: 1.1.0+0.1739994212.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739992853.p
        version: 1.1.0+0.1739992853.p
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: latest
    - currentCSV: red-hat-hawtio-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              },
              {
                "apiVersion": "hawt.io/v2",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24,
                    "internalSSL": true
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
          createdAt: "2025-08-07T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v2
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
        version: 1.2.1
      entries:
      - name: red-hat-hawtio-operator.v1.2.1
        version: 1.2.1
      - name: red-hat-hawtio-operator.v1.2.0-0.1754874242.p
        version: 1.2.0+0.1754874242.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754338551.p
        version: 1.2.0+0.1754338551.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754317392.p
        version: 1.2.0+0.1754317392.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1753863723.p
        version: 1.2.0+0.1753863723.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752679472.p
        version: 1.2.0+0.1752679472.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752452299.p
        version: 1.2.0+0.1752452299.p
      - name: red-hat-hawtio-operator.v1.2.0
        version: 1.2.0
      - name: red-hat-hawtio-operator.v1.1.1
        version: 1.1.1
      - name: red-hat-hawtio-operator.v1.1.0
        version: 1.1.0
      - name: red-hat-hawtio-operator.v1.1.0-0.1742200932.p
        version: 1.1.0+0.1742200932.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1740358952.p
        version: 1.1.0+0.1740358952.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739994212.p
        version: 1.1.0+0.1739994212.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739992853.p
        version: 1.1.0+0.1739992853.p
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: stable-v1
    - currentCSV: red-hat-hawtio-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel8-operator@sha256:f04d81d24c757db7b678464382ad8c5182b1580e714123a9f4f5fc9fbd9b8c57
          createdAt: "2024-01-30T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-rhel8-operator@sha256:f04d81d24c757db7b678464382ad8c5182b1580e714123a9f4f5fc9fbd9b8c57
        version: 1.0.3
      entries:
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1
    packageName: red-hat-hawtio-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: namespace-configuration-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: namespace-configuration-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupConfig",
                "metadata": {
                  "name": "test-groupconfig"
                },
                "spec": {
                  "labelSelector": {
                    "matchLabels": {
                      "type": "team"
                    }
                  },
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-dev\n  labels:\n    group: {{ .Name }}  \n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-qa\n  labels:\n    group: {{ .Name }}          \n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-prod\n  labels:\n    group: {{ .Name }}                           \n"
                    },
                    {
                      "objectTemplate": "apiVersion: quota.openshift.io/v1\nkind: ClusterResourceQuota\nmetadata:\n  name: {{ .Name }}-quota\nspec:\n  quota:\n    hard:\n      pods: \"4\" \n      requests.cpu: \"1\" \n      requests.memory: 1Gi \n      requests.ephemeral-storage: 2Gi \n      limits.cpu: \"2\" \n      limits.memory: 2Gi \n      limits.ephemeral-storage: 4Gi\n  selector:\n    labels:\n      matchLabels:\n        group: {{ .Name }}\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NamespaceConfig",
                "metadata": {
                  "name": "test-namespaceconfig"
                },
                "spec": {
                  "labelSelector": {
                    "matchLabels": {
                      "type": "secure"
                    }
                  },
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: ResourceQuota\nmetadata:\n  name: standard-sandbox\n  namespace: {{ .Name }}\nspec:\n  hard:\n    pods: \"4\" \n    requests.cpu: \"1\" \n    requests.memory: 1Gi \n    requests.ephemeral-storage: 2Gi \n    limits.cpu: \"2\" \n    limits.memory: 2Gi \n    limits.ephemeral-storage: 4Gi\n"
                    },
                    {
                      "objectTemplate": "kind: EgressNetworkPolicy\napiVersion: network.openshift.io/v1\nmetadata:\n  name: air-gapped\n  namespace: {{ .Name }}  \nspec:\n  egress:\n  - type: Deny\n    to:\n      cidrSelector: 0.0.0.0/0\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "UserConfig",
                "metadata": {
                  "name": "test-user-config"
                },
                "spec": {
                  "providerName": "my-provider",
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-sandbox\n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: ResourceQuota\nmetadata:\n  name: standard-sandbox\n  namespace: {{ .Name }}-sandbox\nspec:\n  hard:\n    pods: \"4\" \n    requests.cpu: \"1\" \n    requests.memory: 1Gi \n    requests.ephemeral-storage: 2Gi \n    limits.cpu: \"2\" \n    limits.memory: 2Gi \n    limits.ephemeral-storage: 4Gi\n"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/namespace-configuration-operator@sha256:49ed7d6155342adaa2b12fd80c6761c3081d8e6149d187cb7ff91a247cdf2e7a
          createdAt: "2023-11-22T15:58:50Z"
          description: This operator provides a facility to define and enforce namespace
            configurations
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: namespace-configuration-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/namespace-configuration-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GroupConfig is the Schema for the groupconfigs API
            displayName: Group Config
            kind: GroupConfig
            name: groupconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: NamespaceConfig is the Schema for the namespaceconfigs API
            displayName: Namespace Config
            kind: NamespaceConfig
            name: namespaceconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: UserConfig is the Schema for the userconfigs API
            displayName: User Config
            kind: UserConfig
            name: userconfigs.redhatcop.redhat.io
            version: v1alpha1
        description: "The `namespace-configuration-operator` helps keeping configurations
          related to Users, Groups and Namespaces aligned with one of more policies
          specified as a CRs. The purpose is to provide the foundational building
          block to create an end-to-end onboarding process. \nBy onboarding process
          we mean all the provisioning steps needed to a developer team working on
          one or more applications to OpenShift.\n This usually involves configuring
          resources such as: Groups, RoleBindings, Namespaces, ResourceQuotas, NetworkPolicies,
          EgressNetworkPolicies, etc.... . Depending on the specific environment the
          list could continue.\nNaturally such a process should be as automatic and
          scalable as possible.\n\nWith the namespace-configuration-operator one can
          create rules that will react to the creation of Users, Groups and Namespace
          and will create and enforce a set of resources.\n\nHere are some examples
          of the type of onboarding processes that one could support:\n\n1. [developer
          sandbox](https://github.com/redhat-cop/namespace-configuration-operator/blob/master/examples/user-sandbox/readme.md)\n
          2. [team onboarding](https://github.com/redhat-cop/namespace-configuration-operator//blob/master/examples/team-onboarding/readme.md)
          with support of the entire SDLC in a multitentant environment.\n\nPolicies
          can be expressed with the following CRDs:\n\n| Watched Resource | CRD |\n|--|--|\n
          | Groups | [GroupConfig](#GroupConfig) |\n| Users | [UserConfig](#UserConfig)
          |\n| Namespace | [NamespaceConfig](#NamespaceConfig) |\n\nThese CRDs all
          share some commonalities:\n\n1. Templated Resources\n2. List of ignored
          jason path\n \n### Templated Resources\n\nEach has a parameter called `templatedResources`,
          which is an array. Each element of the array has two fields `objectTemplate`
          and `excludedPaths` (see below).\n\nThe `objectTemplate` field must contains
          a [go template](https://golang.org/pkg/text/template/) that resolves to
          a single API Resource expressed in `yaml`. The template is merged with the
          object selected by the CR. For example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: UserConfig\nmetadata:\n  name: test-user-config\nspec:\n  provider:
          corp-ldap\n  templates:\n  - objectTemplate: |\n      apiVersion: v1\n      kind:
          Namespace\n      metadata:\n        name: {{ .Name }}-sandbox\n```\n\nThis
          create a rule in which every time a user from the `corp-ldap` provider is
          created, a namespace called `<username>-sandbox` is also created.\n\n###
          Excluded Paths\n\nThe logic of the `namespace-configuration-operator` is
          to enforce that the resources resolved by processing the templates \"stays
          in place\". In other words if those resources are changed and/or deleted
          they will be reset by the operator.\nBut there are situations in which at
          least part of a resource is allowed to change. Common use cases are: annotations
          and in general the metadata section of a resource can be updated by the
          various operators watching that resources. The status field is often updated
          by the main operator managing that resources. Finally, when applicable the
          `spec.replicas` field should also be allowed to change.\n\nTo handle special
          use case, one can also specify additional *jsonpaths* that should be ignored
          when comparing the desired resource and the current resource and making
          a decision on whether that resource should be reset.\n\nThe following paths
          are always included:\n \n1. `.metadata`\n2. `.status`\n3. `.spec.replicas`\n\n##
          NamespaceConfig\n\n The `NamespaceConfig` CR allows specifying one or more
          objects that will be created in the selected namespaces.\n\nNamespaces can
          be selected by labels or annotations via a label selector for example:\n\n```yaml\napiVersion:
          redhatcop.redhat.io/v1alpha1\n kind: NamespaceConfig\nmetadata:\n  name:
          small-namespace\n  selector:\n    matchLabels:\n      size: small  \n  resources:\n
          \ - apiVersion: v1\n    kind: ResourceQuota\n    metadata:\n      name:
          small-size  \n    spec:\n      hard:\n        requests.cpu: \"4\"\n        requests.memory:
          \"2Gi\"\n```\n\nHere is a `NamespaceConfig` object using a `matchExpressions`
          selector:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n kind: NamespaceConfig\nmetadata:\n
          \ name: tier-config\nspec:\n  annotationSelector:\n    matchExpressions:\n
          \     - {key: tier, operator: In, values: [gold,silver]}\n ```\n\nAlthough
          not enforced by the operator the general expectation is that the NamespaceConfig
          CR will be used to create objects inside the selected namespace.\n \nExamples
          of NamespaceConfig usages can be found [here](https://github.com/redhat-cop/namespace-configuration-operator/blob/master/examples/namespace-config/readme.md)\n
          \n## GroupConfig\n\nThe `GroupConfig` CR allows specifying one or more objects
          that will be created in the selected Group.\nGroups can be selected by labels
          or annotations via a label selector, similarly to the `NamespaceConfig`.\n\nOften
          groups are created in OpenShift by a job that synchronizes an Identity Provider
          with OCP. So the idea is that when new groups are added or deleted the configuration
          in OpenShift will adapt automatically.\n\nAlthough not enforced by the operator,
          GroupConfig are expected to create cluster-scoped resources like Namespaces,
          ClusterResourceQuotas and potentially some namespaced resources like RoleBindings.\n\n##
          UserConfig\n \nIn OpenShift an external user is defined by two entities:
          Users and Identities. There is a relationship of on to many between Users
          and Identities. Given one user, there can be one Identity per authentication
          mechanism.\n\nThe `UserConfig` CR allows specifying one or more objects
          that will be created in the selected User.\nUsers can be selected by label
          or annotation like `NamespaceConfig` and `UserConfig`.\nUSers can also be
          selected by provider name (the name of the authentication mechanism) and
          identity extra field.\n\nHere is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          UserConfig\nmetadata:\n  name: test-user-config\nspec:\n  providerName:
          okta-provider\n  identityExtraFieldSelector:\n    matchLabels:\n      sandbox_enabled:
          \"true\"\n  templates:\n  - objectTemplate: |\n      apiVersion: v1\n      kind:
          Namespace\n      metadata:\n        name: {{ .Name }}-sandbox\n ```\n\nUser
          will be selected by this `UserConfig` only if they login via the *okta-provider*
          and if the extra field was populate with the label `sandbox_enabled: \"true\"
          `. Note that not all authentication provider allow populating the extra
          fields in the Identity object.\n"
        displayName: Namespace Configuration Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - namespace
        - user
        - group
        - configuration
        - policy
        - management
        links:
        - name: Namespace Configuration Operator
          url: https://github.com/redhat-cop/namespace-configuration-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/namespace-configuration-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/namespace-configuration-operator@sha256:49ed7d6155342adaa2b12fd80c6761c3081d8e6149d187cb7ff91a247cdf2e7a
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.2.6
      entries:
      - name: namespace-configuration-operator.v1.2.6
        version: 1.2.6
      - name: namespace-configuration-operator.v1.2.5
        version: 1.2.5
      - name: namespace-configuration-operator.v1.2.4
        version: 1.2.4
      - name: namespace-configuration-operator.v1.2.3
        version: 1.2.3
      - name: namespace-configuration-operator.v1.2.2
        version: 1.2.2
      - name: namespace-configuration-operator.v1.2.1
        version: 1.2.1
      - name: namespace-configuration-operator.v1.2.0
        version: 1.2.0
      - name: namespace-configuration-operator.v1.1.0
        version: 1.1.0
      - name: namespace-configuration-operator.v1.0.7
        version: 1.0.7
      - name: namespace-configuration-operator.v1.0.6
        version: 1.0.6
      - name: namespace-configuration-operator.v1.0.5
        version: 1.0.5
      - name: namespace-configuration-operator.v1.0.4
        version: 1.0.4
      - name: namespace-configuration-operator.v1.0.3
        version: 1.0.3
      - name: namespace-configuration-operator.v1.0.2
        version: 1.0.2
      - name: namespace-configuration-operator.v1.0.1
        version: 1.0.1
      - name: namespace-configuration-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: namespace-configuration-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator-certified.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a",
                  "imagePullPolicy": "Always",
                  "initImage": "registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:3c03fb18a5d73876576b80aca2b61e7c84ca052839ccb467d14e15f6d1f9d834
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a57926b53eef862b70a53818a7ccb0e9756f878f30d4128b608090b109bdc585
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator-certified.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator-certified.v1.20.0
        version: 1.20.0
      - name: percona-server-mongodb-operator-certified.v1.19.1
        version: 1.19.1
      - name: percona-server-mongodb-operator-certified.v1.19.0
        version: 1.19.0
      - name: percona-server-mongodb-operator-certified.v1.18.0
        version: 1.18.0
      - name: percona-server-mongodb-operator-certified.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator-certified
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Cloudera
      provider-url: https://cloudera.com
    name: cloudera-streams-messaging-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster"
                },
                "spec": {
                  "kafka": {
                    "version": "3.9.0.1.4",
                    "replicas": 3,
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2,
                      "inter.broker.protocol.version": "3.9"
                    },
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "zookeeper": {
                    "replicas": 3,
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "3.9.0.1.4",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
          createdAt: "2025-07-14 10:09:41"
          description: Run Apache Kafka with Cloudera Streams Messaging - Kubernetes
            Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: A valid license and current subscription
            for Cloudera Streaming - Kubernetes Operator is required. Contact Cloudera
            sales to obtain a license. https://www.cloudera.com/contact-sales.html
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |-
          Cloudera Streams Messaging - Kubernetes Operator brings enterprise-grade Kafka deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of Kafka for enterprise deployments. To learn more about Cloudera Streaming, [read this deep dive blog post](https://blog.cloudera.com/accelerating-deployments-of-streaming-pipelines-announcing-data-in-motion-on-kubernetes/).

          Cloudera Streams Messaging - Kubernetes Operator consists of and ships multiple components including Apache Kafka, Apache ZooKeeper, Cruise Control, Strimzi, and others.

          ## Installation prerequisites

          Installing Cloudera Streams Messaging - Kubernetes Operator requires a valid license as well as Cloudera credentials (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly. The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) where installation artifacts are pulled from.

          ### Obtaining a license

          Contact [Cloudera Sales](https://www.cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets

          Before installation, create  two `Secrets`. One containing your Cloudera credentials, and one containing your license. Cloudera credentials are provided to you together with your license.

          Create a `Secret` containing your Cloudera credentials.
          ```
          oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***] --docker-server=container.repository.cloudera.com
          ```
          Create a `Secret` containing your license.
          ```
          oc create secret generic csm-op-license --from-file=license.txt=[***PATH TO LICENSE FILE***]
          ```

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces where you plan to deploy Kafka or Kafka Connect clusters. Cloudera recommends that you create the `Secret` in all required namespaces before installation if you know what namespaces you will be using to deploy Kafka or Kafka Connect.

          ## Learn more
          To learn more, visit the [Cloudera Streams Messaging - Kubernetes Operator documentation library](https://docs.cloudera.com/csm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Streams Messaging - Kubernetes Operator, visit the following resources:

          * [Strimzi](https://strimzi.io/)
          * [Apache Kafka](https://kafka.apache.org/)
          * [Apache ZooKeeper](https://zookeeper.apache.org/)
          * [Cruise Control](https://github.com/linkedin/cruise-control)
        displayName: Cloudera Streams Messaging - Kubernetes Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Documentation
          url: https://docs.cloudera.com/csm-operator
        - name: Cloudera Community Forum
          url: https://community.cloudera.com/
        - name: Cloudera Stream Processing
          url: https://www.cloudera.com/products/stream-processing.html
        - name: Contact Cloudera
          url: https://www.cloudera.com/contact-sales.html
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
        - container.repository.cloudera.com/cloudera/kafka@sha256:aa64e161b9672153a6f6559f9ae9ef88d0e40d6040a80e21bd3dc8ee39df6523
        - container.repository.cloudera.com/cloudera/kafka@sha256:b675e7426a558c4681e53d8e03f278264d7ad555f7c02422017683a29708ac2b
        - container.repository.cloudera.com/cloudera/maven-builder@sha256:9f18203b175193e5079fdc0c2b71c14887641d997f92b4a1c918546f777166df
        version: 1.4.0-b199
      entries:
      - name: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
        version: 1.4.0-b199
      name: 1.2.x
    - currentCSV: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster"
                },
                "spec": {
                  "kafka": {
                    "version": "3.9.0.1.4",
                    "replicas": 3,
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2,
                      "inter.broker.protocol.version": "3.9"
                    },
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "zookeeper": {
                    "replicas": 3,
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "3.9.0.1.4",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
          createdAt: "2025-07-14 10:09:41"
          description: Run Apache Kafka with Cloudera Streams Messaging - Kubernetes
            Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: A valid license and current subscription
            for Cloudera Streaming - Kubernetes Operator is required. Contact Cloudera
            sales to obtain a license. https://www.cloudera.com/contact-sales.html
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |-
          Cloudera Streams Messaging - Kubernetes Operator brings enterprise-grade Kafka deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of Kafka for enterprise deployments. To learn more about Cloudera Streaming, [read this deep dive blog post](https://blog.cloudera.com/accelerating-deployments-of-streaming-pipelines-announcing-data-in-motion-on-kubernetes/).

          Cloudera Streams Messaging - Kubernetes Operator consists of and ships multiple components including Apache Kafka, Apache ZooKeeper, Cruise Control, Strimzi, and others.

          ## Installation prerequisites

          Installing Cloudera Streams Messaging - Kubernetes Operator requires a valid license as well as Cloudera credentials (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly. The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) where installation artifacts are pulled from.

          ### Obtaining a license

          Contact [Cloudera Sales](https://www.cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets

          Before installation, create  two `Secrets`. One containing your Cloudera credentials, and one containing your license. Cloudera credentials are provided to you together with your license.

          Create a `Secret` containing your Cloudera credentials.
          ```
          oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***] --docker-server=container.repository.cloudera.com
          ```
          Create a `Secret` containing your license.
          ```
          oc create secret generic csm-op-license --from-file=license.txt=[***PATH TO LICENSE FILE***]
          ```

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces where you plan to deploy Kafka or Kafka Connect clusters. Cloudera recommends that you create the `Secret` in all required namespaces before installation if you know what namespaces you will be using to deploy Kafka or Kafka Connect.

          ## Learn more
          To learn more, visit the [Cloudera Streams Messaging - Kubernetes Operator documentation library](https://docs.cloudera.com/csm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Streams Messaging - Kubernetes Operator, visit the following resources:

          * [Strimzi](https://strimzi.io/)
          * [Apache Kafka](https://kafka.apache.org/)
          * [Apache ZooKeeper](https://zookeeper.apache.org/)
          * [Cruise Control](https://github.com/linkedin/cruise-control)
        displayName: Cloudera Streams Messaging - Kubernetes Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Documentation
          url: https://docs.cloudera.com/csm-operator
        - name: Cloudera Community Forum
          url: https://community.cloudera.com/
        - name: Cloudera Stream Processing
          url: https://www.cloudera.com/products/stream-processing.html
        - name: Contact Cloudera
          url: https://www.cloudera.com/contact-sales.html
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
        - container.repository.cloudera.com/cloudera/kafka@sha256:aa64e161b9672153a6f6559f9ae9ef88d0e40d6040a80e21bd3dc8ee39df6523
        - container.repository.cloudera.com/cloudera/kafka@sha256:b675e7426a558c4681e53d8e03f278264d7ad555f7c02422017683a29708ac2b
        - container.repository.cloudera.com/cloudera/maven-builder@sha256:9f18203b175193e5079fdc0c2b71c14887641d997f92b4a1c918546f777166df
        version: 1.4.0-b199
      entries:
      - name: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
        version: 1.4.0-b199
      name: stable
    defaultChannel: stable
    packageName: cloudera-streams-messaging-kubernetes-operator
    provider:
      name: Cloudera
      url: https://cloudera.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: vertical-pod-autoscaler
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: verticalpodautoscaler.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "autoscaling.openshift.io/v1",
                "kind": "VerticalPodAutoscalerController",
                "metadata": {
                  "name": "default",
                  "namespace": "openshift-vertical-pod-autoscaler"
                },
                "spec": {
                  "podMinCPUMillicores": 25,
                  "podMinMemoryMb": 250,
                  "recommendationOnly": false,
                  "safetyMarginFraction": 0.15,
                  "minReplicas": 2
                }
              },
              {
                 "apiVersion": "autoscaling.k8s.io/v1",
                 "kind": "VerticalPodAutoscaler",
                 "metadata": {
                    "name": "myapp-vpa"
                 },
                 "spec": {
                    "targetRef": {
                       "apiVersion": "apps/v1",
                       "kind": "Deployment",
                       "name": "myapp-deployment"
                    },
                    "resourcePolicy": {
                       "containerPolicies": [
                          {
                             "containerName": "*",
                             "minAllowed": {
                                "cpu": "100m",
                                "memory": "50Mi"
                             },
                             "maxAllowed": {
                                "cpu": 1,
                                "memory": "500Mi"
                             },
                             "controlledResources": [
                                "cpu",
                                "memory"
                             ]
                          }
                       ]
                    }
                 }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          certifiedLevel: Primed
          containerImage: registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9-operator@sha256:edf76334695c9a4aee4a9c0297af40475a16dec7e35ec7113a35b8507724536e
          createdAt: 2019-10-14 12:17:13+00:00
          description: An operator to run the OpenShift Vertical Pod Autoscaler. Vertical
            Pod Autoscaler (VPA) can be configured to monitor a workload's resource
            utilization, and then adjust its CPU and memory limits by updating the
            pod (future) or restarting the pod with updated limits.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          healthIndex: B
          olm.skipRange: '>=4.5.0 <4.16.0-202509240123'
          operatorframework.io/suggested-namespace: openshift-vertical-pod-autoscaler
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/vertical-pod-autoscaler-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of the set of VPA controllers
            displayName: VPA Controller
            kind: VerticalPodAutoscalerController
            name: verticalpodautoscalercontrollers.autoscaling.openshift.io
            version: v1
          - description: An application whose resources are managed by the VPA
            displayName: VerticalPodAutoscaler
            kind: VerticalPodAutoscaler
            name: verticalpodautoscalers.autoscaling.k8s.io
            version: v1
          - description: Represents VPA checkpoints used by the VPA controllers to
              manage workloads
            displayName: VPA Checkpoint
            kind: VerticalPodAutoscalerCheckpoint
            name: verticalpodautoscalercheckpoints.autoscaling.k8s.io
            version: v1
        description: An operator to run the OpenShift Vertical Pod Autoscaler. Vertical
          Pod Autoscaler (VPA) can be configured to monitor a workload's resource
          utilization, and then adjust its CPU and memory limits by updating the pod
          (future) or restarting the pod with updated limits.
        displayName: VerticalPodAutoscaler
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Vertical Pod Autoscaler Documentation
          url: https://docs.openshift.com/container-platform/latest/nodes/pods/nodes-pods-vertical-autoscaler.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9-operator@sha256:edf76334695c9a4aee4a9c0297af40475a16dec7e35ec7113a35b8507724536e
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9@sha256:e6c8f34895184179bb87f1659c4ff30b96b131787fa7a64e7bc6b6adbb336c16
        version: 4.16.0-202509240123
      entries:
      - name: verticalpodautoscaler.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: verticalpodautoscaler.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: verticalpodautoscaler.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: verticalpodautoscaler.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: verticalpodautoscaler.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: verticalpodautoscaler.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: verticalpodautoscaler.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: verticalpodautoscaler.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: verticalpodautoscaler.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: verticalpodautoscaler.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: verticalpodautoscaler.v4.16.0-202501260334
        version: 4.16.0-202501260334
      - name: verticalpodautoscaler.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: verticalpodautoscaler.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: verticalpodautoscaler.v4.16.0-202412021735
        version: 4.16.0-202412021735
      - name: verticalpodautoscaler.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: verticalpodautoscaler.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: verticalpodautoscaler.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: verticalpodautoscaler.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: verticalpodautoscaler.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: verticalpodautoscaler.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: verticalpodautoscaler.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: verticalpodautoscaler.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: verticalpodautoscaler.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: verticalpodautoscaler.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: verticalpodautoscaler.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: verticalpodautoscaler.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: verticalpodautoscaler.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: verticalpodautoscaler.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: verticalpodautoscaler.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: verticalpodautoscaler.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: vertical-pod-autoscaler
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-businessautomation: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: businessautomation-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: businessautomation-operator.7.13.5-16
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.kiegroup.org/v2","kind":"KieApp","metadata":{"name":"rhpam-trial"},"spec":{"environment":"rhpam-trial"}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/rhpam-7/rhpam-rhel8-operator@sha256:3b4f420def9e5d2c768d5209264bd5f9af6d6a37b22ef9a988a167ff7d9959cf
          createdAt: "2025-07-30 19:54:27"
          description: Deploys and manages Red Hat Process Automation Manager and
            Red Hat Decision Manager environments.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Process Automation
            Manager", "Red Hat Decision Manager"]'
          repository: https://github.com/kiegroup/kie-cloud-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A project prescription running an RHPAM/RHDM environment.
            displayName: KieApp
            kind: KieApp
            name: kieapps.app.kiegroup.org
            version: v2
        description: |-
          Deploys and manages Red Hat Process Automation Manager and Red Hat Decision Manager environments.

          * **Red Hat Process Automation Manager** is a platform for developing containerized microservices and applications that automate business decisions and processes. It includes business process management (BPM), business rules management (BRM), and business resource optimization and complex event processing (CEP) technologies. It also includes a user experience platform to create engaging user interfaces for process and decision services with minimal coding.

          [See more](https://www.redhat.com/en/products/process-automation).
        displayName: Business Automation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kieapp
        - pam
        - decision
        - kie
        - cloud
        - bpm
        - process
        - automation
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-process-automation-manager
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_process_automation_manager/7.13/#category-deploying-red-hat-process-automation-manager-on-openshift
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-monitoring-rhel8@sha256:8d4481a11269565f1690ad007f152e24326bae4b2099cb26ddf7b21be67031e8
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:cb37360843e9b062c59755593d9d412e1bf2c1dc1d6902e8a65a2825812c0989
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhpam-7/rhpam-kieserver-rhel8@sha256:2526d12da63b67ae9ed9294c914e9406d833c090ffa777623303ea0be22755d8
        - registry.redhat.io/rhpam-7/rhpam-process-migration-rhel8@sha256:3d7bdbe38895e50549e97573bd40da47b11691932164db7dafff0155a051aab2
        - registry.redhat.io/rhpam-7/rhpam-controller-rhel8@sha256:3a901b54a9d35d38b5aca275df2160e3190a874e139fcb606b9969b38284b3dc
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        - registry.redhat.io/rhpam-7/rhpam-kieserver-rhel8@sha256:3fbef82d1c5da8da1606d1a2d5dcfd20000da574737eb32b772d4f0123b179cb
        - registry.redhat.io/rhpam-7/rhpam-smartrouter-rhel8@sha256:d1e5461db8983a13f98259152dc3cfa2aa677c29f799104190e5cefa83cf97b5
        - registry.redhat.io/rhpam-7/rhpam-dashbuilder-rhel8@sha256:46396dc8060f2f9cc6f87ad905ef55be5b8e5216e5162bdf54df98c9d70e29ef
        - registry.redhat.io/rhscl/mysql-80-rhel7@sha256:e4f535c78bc2fdb58914c52fb805185035ff31515452b4df714f9ab623ba383a
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-rhel8@sha256:228c6ab5b83e215a98f6eb76250feb2ae3096fd21398c44e5b2701257c9c3493
        - registry.redhat.io/rhpam-7/rhpam-smartrouter-rhel8@sha256:9d7efdd97f163770d8a84b1ce5236bd4f31508154dced1f11dc6efe75a7b5089
        - registry.redhat.io/rhscl/postgresql-13-rhel7@sha256:338dd09681dcf860e6a976a3bc48a1474c2376d698815e4eb611ae9c5aa68966
        - registry.redhat.io/rhpam-7/rhpam-rhel8-operator@sha256:3b4f420def9e5d2c768d5209264bd5f9af6d6a37b22ef9a988a167ff7d9959cf
        - registry.redhat.io/rhpam-7/rhpam-dashbuilder-rhel8@sha256:4edac282119898ab1395030eecaee584990064e88fc8e65fb999f19ea2ef3f8b
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-rhel8@sha256:1abbaba9d4b2ecaec38205bcccf5ef4c7b6388ed9725b4f4c6dc8dc664588eb1
        - registry.redhat.io/rhpam-7/rhpam-controller-rhel8@sha256:989dc6dc64a2d5b6723fb5c000b0411fbdc6f78c0056666d831ffeabeab4b612
        - registry.redhat.io/openshift4/ose-cli@sha256:c3eb8807caf73e30012ae53c6d4f5114471290a5596929c27acef5a699a5b813
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/rhpam-7/rhpam-process-migration-rhel8@sha256:bd5c6586b4859dec23ad9530ee98bab33497ab251d544f863fc0af01cbd45101
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:457fc73423fa94591bc7e35729a25c58d3465583afccf342fb338b6e11188575
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-monitoring-rhel8@sha256:32912270eba0f2464da88f315bde8bf1ef5bbded814db156811d7a415a458f86
        version: 7.13.5-16
      entries:
      - name: businessautomation-operator.7.13.5-16
        version: 7.13.5-16
      - name: businessautomation-operator.7.13.5-15
        version: 7.13.5-15
      - name: businessautomation-operator.7.13.5-14
        version: 7.13.5-14
      - name: businessautomation-operator.7.13.5-13
        version: 7.13.5-13
      - name: businessautomation-operator.7.13.5-12
        version: 7.13.5-12
      - name: businessautomation-operator.7.13.5-11
        version: 7.13.5-11
      - name: businessautomation-operator.7.13.5-10
        version: 7.13.5-10
      - name: businessautomation-operator.7.13.5-9
        version: 7.13.5-9
      - name: businessautomation-operator.7.13.5-8
        version: 7.13.5-8
      - name: businessautomation-operator.7.13.5-7
        version: 7.13.5-7
      - name: businessautomation-operator.7.13.5-6
        version: 7.13.5-6
      - name: businessautomation-operator.7.13.5-5
        version: 7.13.5-5
      - name: businessautomation-operator.7.13.5-4
        version: 7.13.5-4
      - name: businessautomation-operator.7.13.5-3
        version: 7.13.5-3
      - name: businessautomation-operator.7.13.5-2
        version: 7.13.5-2
      - name: businessautomation-operator.7.13.5-1
        version: 7.13.5-1
      - name: businessautomation-operator.7.13.4-4
        version: 7.13.4-4
      - name: businessautomation-operator.7.13.4-3
        version: 7.13.4-3
      - name: businessautomation-operator.7.13.4-2
        version: 7.13.4-2
      - name: businessautomation-operator.7.13.4-1
        version: 7.13.4-1
      - name: businessautomation-operator.7.13.3-5
        version: 7.13.3-5
      - name: businessautomation-operator.7.13.3-4
        version: 7.13.3-4
      - name: businessautomation-operator.7.13.3-3
        version: 7.13.3-3
      - name: businessautomation-operator.7.13.3-2
        version: 7.13.3-2
      - name: businessautomation-operator.7.13.3-1
        version: 7.13.3-1
      - name: businessautomation-operator.7.13.2-3
        version: 7.13.2-3
      - name: businessautomation-operator.7.13.2-2
        version: 7.13.2-2
      - name: businessautomation-operator.7.13.2-1
        version: 7.13.2-1
      - name: businessautomation-operator.7.13.1-2
        version: 7.13.1-2
      - name: businessautomation-operator.7.13.1-1
        version: 7.13.1-1
      - name: businessautomation-operator.7.13.0-1
        version: 7.13.0-1
      - name: businessautomation-operator.7.12.1-4
        version: 7.12.1-4
      - name: businessautomation-operator.7.12.1-3
        version: 7.12.1-3
      - name: businessautomation-operator.7.12.1-2
        version: 7.12.1-2
      - name: businessautomation-operator.7.12.1-1
        version: 7.12.1-1
      - name: businessautomation-operator.7.12.0-2
        version: 7.12.0-2
      - name: businessautomation-operator.7.12.0-1
        version: 7.12.0-1
      - name: businessautomation-operator.7.11.1-3
        version: 7.11.1-3
      - name: businessautomation-operator.7.11.1-2
        version: 7.11.1-2
      - name: businessautomation-operator.7.11.1-1
        version: 7.11.1-1
      - name: businessautomation-operator.7.11.0-2
        version: 7.11.0-2
      - name: businessautomation-operator.7.11.0-1
        version: 7.11.0-1
      - name: businessautomation-operator.7.10.1-2
        version: 7.10.1-2
      - name: businessautomation-operator.7.10.1-1
        version: 7.10.1-1
      - name: businessautomation-operator.7.10.0-2
        version: 7.10.0-2
      - name: businessautomation-operator.7.10.0-1
        version: 7.10.0-1
      - name: businessautomation-operator.7.9.1-1
        version: 7.9.1-1
      - name: businessautomation-operator.7.9.0
        version: 7.9.0
      - name: businessautomation-operator.7.9.0-2
        version: 7.9.0-2
      - name: businessautomation-operator.7.8.1
        version: 7.8.1
      - name: businessautomation-operator.7.8.0
        version: 7.8.0
      - name: businessautomation-operator.1.4.1
        version: 1.4.1
      - name: businessautomation-operator.1.4.0
        version: 1.4.0
      - name: businessautomation-operator.1.3.0
        version: 1.3.0
      - name: businessautomation-operator.1.2.1
        version: 1.2.1
      - name: businessautomation-operator.1.2.0
        version: 1.2.0
      name: stable
    defaultChannel: stable
    packageName: businessautomation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: quay-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:d73321dfa6dcced943405fae45c2ccf841663e0ab57e8be8cc1ddba251b7ce08
          createdAt: 2021-11-02 11:11 UTC
          description: Opinionated deployment of Red Hat Quay on Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:d73321dfa6dcced943405fae45c2ccf841663e0ab57e8be8cc1ddba251b7ce08
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:0e6bef51caf479074925115fa11621c41bcada375cc1925d0511a73084b93543
        - registry.redhat.io/rhel8/postgresql-10@sha256:b2695e9af379624621bef9ec785f0cb618bfdabae98c5f17eea82ff5f8d2d5cb
        - registry.redhat.io/rhel8/redis-5@sha256:8d7e66927ce4b3050ca15ab4df4dbaeb593cb855e16fb0a6f5dde950e9db7db9
        - registry.redhat.io/quay/quay-rhel8@sha256:ed8c4249d376b4226e35bf088ffe5bf688bf68b059718c5e218aa1184fe048d0
        - registry.redhat.io/quay/clair-rhel8@sha256:af5e2fd1543365b365b1a6cce00b86e44fe5e66cde6491e9b113324c21f629c8
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:2a471300f70d3995ab7e9fcc3df35c5513e73fcc1d140fb4d651a1c3eae949eb
        version: 3.4.7
      entries:
      - name: quay-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: quay-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:1f5e4cc9ec2d9b8b89d82f1a698acfa1ac2764ca4ebc5c85b3417da7730b5867
          createdAt: 2021-09-09 20:09 UTC
          description: Opinionated deployment of Red Hat Quay on Kubernetes.
          olm.skipRange: '>=3.4.x <3.5.7'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.5.7
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7fe408a1e428b2d0660cb578c1dcd4b961f838e37166fe7b21b2bed70d288594
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:5d72577ccfd63c6d9ad712b376b604c40d2d3ecb53e391e2ee0d18caa6e0225a
        - registry.redhat.io/rhel8/postgresql-10@sha256:a723bb3a7a5046527b9a5b628ddfb211c526e3814cc4242262c4f8a016f38edd
        - registry.redhat.io/rhel8/redis-5@sha256:7916760ff03af5ca230df5e96b2bf353fbf1829386de1c063c05f623837be7f7
        - registry.redhat.io/quay/quay-rhel8@sha256:c599892160ac20c744d833603d9375923af7aec1edfd86982b7513e02fb6d463
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:1f5e4cc9ec2d9b8b89d82f1a698acfa1ac2764ca4ebc5c85b3417da7730b5867
        - registry.redhat.io/quay/clair-rhel8@sha256:d242bd9d37f95b2a5a8267b3f1eba02c671d46121712baae43466c86d1da89f9
        version: 3.5.7
      entries:
      - name: quay-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: quay-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:3c0201e3cf88e711c921870324002a02cbd42c0d2eecafb4f242aa0601e24602
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.10.15
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:fc20b8bdac88c1bfce9aec993bdce6dd2cc132f8a9d4062d174b098819c74013
        - registry.redhat.io/quay/clair-rhel8@sha256:efa6e69500ea643042b544e6205f088214d79ab09a36864a612e73d07979d1c5
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:3c0201e3cf88e711c921870324002a02cbd42c0d2eecafb4f242aa0601e24602
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7e720c6a332ec4032e8576c31421af404f61d9baaf8e88f69f8333c52e5d1683
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:774cd1f181527787ba8631c7c14437eb2250c15ddb4a9c75fdfa35c5c1a19a2e
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        version: 3.10.15
      entries:
      - name: quay-operator.v3.10.15
        version: 3.10.15
      - name: quay-operator.v3.10.14
        version: 3.10.14
      - name: quay-operator.v3.10.13
        version: 3.10.13
      - name: quay-operator.v3.10.12
        version: 3.10.12
      - name: quay-operator.v3.10.11
        version: 3.10.11
      - name: quay-operator.v3.10.10
        version: 3.10.10
      - name: quay-operator.v3.10.9
        version: 3.10.9
      - name: quay-operator.v3.10.8
        version: 3.10.8
      - name: quay-operator.v3.10.7
        version: 3.10.7
      - name: quay-operator.v3.10.6
        version: 3.10.6
      - name: quay-operator.v3.10.5
        version: 3.10.5
      - name: quay-operator.v3.10.4
        version: 3.10.4
      - name: quay-operator.v3.10.3
        version: 3.10.3
      - name: quay-operator.v3.10.2
        version: 3.10.2
      - name: quay-operator.v3.10.1
        version: 3.10.1
      - name: quay-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:cb3a41caa3a572b7981d424be2a32b8e881127cc03b557ee0a36ab13f430b897
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.11.13
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:7230fa1d4970e877cb6de8716e99a9bd43302c0ff23b4501f2aa806a8d496d9b
        - registry.redhat.io/quay/clair-rhel8@sha256:3746677ef1cb0ddd4f237ab598b14cf0bc76dcce080ffde127b5bba87ffe6693
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:1a083cd72502609427428ed22fb60846613ddb41a4613496c5cddc797b382d3a
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:11b8c9d64c42f0f0c7b2df491bba859d0410666a8ef1d2fd41eb77e4ad9d2dc8
        - registry.redhat.io/rhel8/postgresql-13@sha256:35d86936c539007a78a89869a14b87b76788fe117df442a670f63936b978ed5b
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:cb3a41caa3a572b7981d424be2a32b8e881127cc03b557ee0a36ab13f430b897
        version: 3.11.13
      entries:
      - name: quay-operator.v3.11.13
        version: 3.11.13
      - name: quay-operator.v3.11.12
        version: 3.11.12
      - name: quay-operator.v3.11.11
        version: 3.11.11
      - name: quay-operator.v3.11.10
        version: 3.11.10
      - name: quay-operator.v3.11.9
        version: 3.11.9
      - name: quay-operator.v3.11.8
        version: 3.11.8
      - name: quay-operator.v3.11.7
        version: 3.11.7
      - name: quay-operator.v3.11.6
        version: 3.11.6
      - name: quay-operator.v3.11.5
        version: 3.11.5
      - name: quay-operator.v3.11.4
        version: 3.11.4
      - name: quay-operator.v3.11.3
        version: 3.11.3
      - name: quay-operator.v3.11.2
        version: 3.11.2
      - name: quay-operator.v3.11.1
        version: 3.11.1
      - name: quay-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:accb705b6327fb2dcc1ce657a99f61e343eca752c9873d63d89fea89c10c3187
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.12.12
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:accb705b6327fb2dcc1ce657a99f61e343eca752c9873d63d89fea89c10c3187
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:ccbd83e05aea0bb96bf099f63420175b36ed413bfc8a618f4c90b6a2abc53895
        - registry.redhat.io/quay/clair-rhel8@sha256:dc3d5fb35f2bd5d00480351dca705fe260a225d9a97958cf0276801af08d8307
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:8e5dd171e6e94de0137bb6a972533b1f93b56efc1731e19b80869d5e9b8a26a6
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:8cdd677449f3c723dad79bcd9c4191c8917d6eb0beeeeeda68fe19a0988bb2a8
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        version: 3.12.12
      entries:
      - name: quay-operator.v3.12.12
        version: 3.12.12
      - name: quay-operator.v3.12.11
        version: 3.12.11
      - name: quay-operator.v3.12.10
        version: 3.12.10
      - name: quay-operator.v3.12.9
        version: 3.12.9
      - name: quay-operator.v3.12.8
        version: 3.12.8
      - name: quay-operator.v3.12.7
        version: 3.12.7
      - name: quay-operator.v3.12.6
        version: 3.12.6
      - name: quay-operator.v3.12.5
        version: 3.12.5
      - name: quay-operator.v3.12.4
        version: 3.12.4
      - name: quay-operator.v3.12.3
        version: 3.12.3
      - name: quay-operator.v3.12.2
        version: 3.12.2
      - name: quay-operator.v3.12.1
        version: 3.12.1
      - name: quay-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:5481a4b914ce4aed6ba3ceee8432eddab6e264cad2caef528404795f904100eb
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.13.8
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/clair-rhel8@sha256:069ccf36346cb14f4e4ea54de1d59afb72548912b140cfef67a6bcb7534d25e9
        - registry.redhat.io/rhel8/postgresql-13@sha256:f17fd8dd4d56d8e916cb2ad3f403d05ad259f79fd6918d67a38a6e9e4a8896cf
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:966a3dafa2dd31ac1a39e826e08c1b988fe694189fe79449bbe14a48b2631213
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:5481a4b914ce4aed6ba3ceee8432eddab6e264cad2caef528404795f904100eb
        - registry.redhat.io/quay/quay-rhel8@sha256:8887c17547c7a4be61471523eade9820e5095cf966b1885a3758081da559871c
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:31a9095d8f291ce68e8e688877692e74a17143bdda8d998fe3a07bcee32575dd
        - registry.redhat.io/rhel8/postgresql-15@sha256:a5f36e7893a519c12752c6118233c92d4e83618670bc9ea7bd3a8a34a63bd50b
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:f62cbd6632454c227b7cbb01dadc8e501e8288b2d8e0abf8e7842b806fa3195c
        version: 3.13.8
      entries:
      - name: quay-operator.v3.13.8
        version: 3.13.8
      - name: quay-operator.v3.13.7
        version: 3.13.7
      - name: quay-operator.v3.13.6
        version: 3.13.6
      - name: quay-operator.v3.13.5
        version: 3.13.5
      - name: quay-operator.v3.13.4
        version: 3.13.4
      - name: quay-operator.v3.13.3
        version: 3.13.3
      - name: quay-operator.v3.13.2
        version: 3.13.2
      - name: quay-operator.v3.13.1
        version: 3.13.1
      - name: quay-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:8ebc1f0511f7d394ceaa39a4c993e955c560dc04588072cb3f8c8fab8c85e32b
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.14.5
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/clair-rhel8@sha256:e11e8a819d17bd14c67f093b5303b4958855e731f89ce67dcca70fcb7b77f45a
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:8ebc1f0511f7d394ceaa39a4c993e955c560dc04588072cb3f8c8fab8c85e32b
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-rhel8@sha256:33e984892650d315065313fcb5f8808e145aadec5a287219c7031d85b713235e
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:43ef8adddbf9b007b6e2396f53bd6a5475150d28c2f4254f031a0fe6c7de5826
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:a12a0918f6dea114b0aae231d7119a4acfd0cd297074d895f4ecf190d2aec43a
        version: 3.14.5
      entries:
      - name: quay-operator.v3.14.5
        version: 3.14.5
      - name: quay-operator.v3.14.4
        version: 3.14.4
      - name: quay-operator.v3.14.3
        version: 3.14.3
      - name: quay-operator.v3.14.2
        version: 3.14.2
      - name: quay-operator.v3.14.1
        version: 3.14.1
      - name: quay-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:05fb981d9848281b97b3683093f752b1c1dc5ced439a2b4ba18b26ff8e11231e
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.15.2
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:873a3daf1dadf2438eb96e151072326c31ba34ed48cac041614b604413d5f8e7
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:956f3110f944d880bc27bb6df9472d0804cd75f5dcf4ed11bcf721e9d1d4542b
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:aa8c9f7b6ce57441ca1fe9dbaaefaf2bb83fd44aa367647b663bb822bf890362
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:05fb981d9848281b97b3683093f752b1c1dc5ced439a2b4ba18b26ff8e11231e
        - registry.redhat.io/quay/clair-rhel8@sha256:861be31d30142bae125cd059bb716d79b28694e9c4336333c68adf4a3823b61d
        version: 3.15.2
      entries:
      - name: quay-operator.v3.15.2
        version: 3.15.2
      - name: quay-operator.v3.15.1
        version: 3.15.1
      - name: quay-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-operator.v3.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:7a6e592d7f5f760b4451f9c974dbb96f56b1d3de92f039adcbf284f78ba13a02
          createdAt: 2021-10-12 22:10 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.6.10
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:ec36241f0a9a2983b1894cd8501898e6b70379af51a81b1e09b0ba0e9b54b33f
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:75df6b8e6bef14844a08f7d994470fa6ec4ffa66d4b15faaa0a3ba373f612fbb
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:7a6e592d7f5f760b4451f9c974dbb96f56b1d3de92f039adcbf284f78ba13a02
        - registry.redhat.io/rhel8/postgresql-10@sha256:4b912c80085b88a03309aeb7907efcc29dd3342fa3952b6ea067afb1914bfe53
        - registry.redhat.io/rhel8/redis-6@sha256:d7c7852338717308cbb59e9303e1ea35cc8e5c01ceb2818569be20c15f3f943d
        - registry.redhat.io/quay/quay-rhel8@sha256:e75e4d084eba9bb65b9cc72e38294af314d49316b12a0beb75db808dc057830b
        - registry.redhat.io/quay/clair-rhel8@sha256:cbafddd1eb50c1c0f9159ffafd86f44a8c0b976194903c96c6f144bfe62e749d
        version: 3.6.10
      entries:
      - name: quay-operator.v3.6.10
        version: 3.6.10
      - name: quay-operator.v3.6.9
        version: 3.6.9
      - name: quay-operator.v3.6.8
        version: 3.6.8
      - name: quay-operator.v3.6.7
        version: 3.6.7
      - name: quay-operator.v3.6.6
        version: 3.6.6
      - name: quay-operator.v3.6.5
        version: 3.6.5
      - name: quay-operator.v3.6.4
        version: 3.6.4
      - name: quay-operator.v3.6.2
        version: 3.6.2
      - name: quay-operator.v3.6.1
        version: 3.6.1
      - name: quay-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: quay-operator.v3.7.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:a8283c2be661e07c5f2f2e5b683c520846bcb91d580d9a3a0aa27ce994ddfff6
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.7.14
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:b15335d88b240bd6bbedeb7fe26f0a1de2f6c330af9758b0fb2edab7ca69f6df
        - registry.redhat.io/quay/quay-rhel8@sha256:7bc9e115e9bace07505ffd225242cab83d5eb6d50358188d18fd9161bdf1a5a6
        - registry.redhat.io/quay/clair-rhel8@sha256:d676af83e96740220e6f18721c98817ba071fd023a62ebdc3815424e94838083
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:c54c77e09c95dc7d80d7b31e796857c8d06f2f733426629b53d3298206d23b20
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:04ac405c152bffb43b33fb2e7f45df8b3940eb21e055ebf6aa17bd81ebdeeab4
        - registry.redhat.io/rhel8/postgresql-10@sha256:a52ad402458ec8ef3f275972c6ebed05ad64398f884404b9bb8e3010c5c95291
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:a8283c2be661e07c5f2f2e5b683c520846bcb91d580d9a3a0aa27ce994ddfff6
        version: 3.7.14
      entries:
      - name: quay-operator.v3.7.14
        version: 3.7.14
      - name: quay-operator.v3.7.13
        version: 3.7.13
      - name: quay-operator.v3.7.12
        version: 3.7.12
      - name: quay-operator.v3.7.11
        version: 3.7.11
      - name: quay-operator.v3.7.10
        version: 3.7.10
      name: stable-3.7
    - currentCSV: quay-operator.v3.8.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:abec0fa7ee641c8e424e995b5fe07b98e522ad1eb92c3b90426f69941f33eb2e
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.8.15
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7e4e57071439d671971b0de46b597cf6cff8ef3879ab711125a6ae7ecb9c6d29
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:3c047c7b22fb42557bd3a9b9c9ac333f0df7d5e4a4d333a2b02a4313f57085be
        - registry.redhat.io/rhel8/postgresql-10@sha256:b7fa644277d5e406b642685ff2ace1b1de7aea2e2a1421b3cbcda7d60405f44d
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:abec0fa7ee641c8e424e995b5fe07b98e522ad1eb92c3b90426f69941f33eb2e
        - registry.redhat.io/rhel8/redis-6@sha256:444200241cb9186cd97d564d7acf077d83fc997a8ee5ee7fdae41b604895085c
        - registry.redhat.io/quay/quay-rhel8@sha256:de2b65506f99eebb5eaafca6bb42b4ebec08e9479dd0d50aa1286f197e42d6de
        - registry.redhat.io/quay/clair-rhel8@sha256:b327fcfd7472cdb9d83a564904976b26b285fa576dc0a76e5590990ec77fd871
        version: 3.8.15
      entries:
      - name: quay-operator.v3.8.15
        version: 3.8.15
      - name: quay-operator.v3.8.14
        version: 3.8.14
      - name: quay-operator.v3.8.13
        version: 3.8.13
      - name: quay-operator.v3.8.12
        version: 3.8.12
      - name: quay-operator.v3.8.11
        version: 3.8.11
      - name: quay-operator.v3.8.10
        version: 3.8.10
      - name: quay-operator.v3.8.9
        version: 3.8.9
      - name: quay-operator.v3.8.8
        version: 3.8.8
      - name: quay-operator.v3.8.7
        version: 3.8.7
      - name: quay-operator.v3.8.6
        version: 3.8.6
      - name: quay-operator.v3.8.5
        version: 3.8.5
      - name: quay-operator.v3.8.4
        version: 3.8.4
      - name: quay-operator.v3.8.3
        version: 3.8.3
      - name: quay-operator.v3.8.2
        version: 3.8.2
      - name: quay-operator.v3.8.1
        version: 3.8.1
      - name: quay-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:8a16803946272156b7aeb00eb2202ab6cb87d875fe4a7c327554f770f0f2ecb5
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.9.16
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/clair-rhel8@sha256:6c67163c783d7d9fa4d187963379885a8b76685347ecf1894c4f8ebd6656544b
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:af8ada0c88b6f19991e61ee86d04b49a148c19dacb57c282e4ab98630e7499b4
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:5a42689198701d845ff3f05502cbbebceb7e6ad7b12762ccc038c0131cf6479e
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:f87b7e7ee6c31f6ae80b08aef0e10729b21fe1b8656fe18f6dee2310bae28ec1
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:8a16803946272156b7aeb00eb2202ab6cb87d875fe4a7c327554f770f0f2ecb5
        version: 3.9.16
      entries:
      - name: quay-operator.v3.9.16
        version: 3.9.16
      - name: quay-operator.v3.9.15
        version: 3.9.15
      - name: quay-operator.v3.9.14
        version: 3.9.14
      - name: quay-operator.v3.9.13
        version: 3.9.13
      - name: quay-operator.v3.9.12
        version: 3.9.12
      - name: quay-operator.v3.9.11
        version: 3.9.11
      - name: quay-operator.v3.9.10
        version: 3.9.10
      - name: quay-operator.v3.9.9
        version: 3.9.9
      - name: quay-operator.v3.9.8
        version: 3.9.8
      - name: quay-operator.v3.9.7
        version: 3.9.7
      - name: quay-operator.v3.9.6
        version: 3.9.6
      - name: quay-operator.v3.9.5
        version: 3.9.5
      - name: quay-operator.v3.9.4
        version: 3.9.4
      - name: quay-operator.v3.9.3
        version: 3.9.3
      - name: quay-operator.v3.9.2
        version: 3.9.2
      - name: quay-operator.v3.9.1
        version: 3.9.1
      - name: quay-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: quay-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MinIO
      provider-url: https://min.io
    name: minio-key-manager-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-key-manager-operator.v2025.8.19175300
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aistor.min.io/v1alpha1",
                "kind": "KeyManager",
                "metadata": {
                  "name": "my-kms",
                  "namespace": "ns-1"
                },
                "spec": {
                  "apiKeySecret": {
                    "name": "my-api-key"
                  },
                  "configuration": {
                    "name": "my-kms-server-config"
                  },
                  "hsmSecret": {
                    "name": "my-kms-hsm"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "registry-creds"
                    }
                  ],
                  "replicas": 2,
                  "volumeClaimTemplate": {
                    "metadata": {
                      "name": "key-manager-volume"
                    },
                    "spec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "100Mi"
                        }
                      },
                      "storageClassName": "standard"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security, Storage
          createdAt: "2025-08-27T14:50:59Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KeyManager
            name: keymanagers.aistor.min.io
            version: v1alpha1
        description: |-
          Highly available, powerful and operationally simple, MinIO's Enterprise Key Management Server is optimized for large storage infrastructures where billions of cryptographic keys are required.

          MinIO's AIStor KMS establishes its foundational trust using the concept of an hardware security module (but given KES is software, this is only a concept). That module assumes a pivotal role in sealing and unsealing the KMS root encryption key. The module responsibility extends to safeguarding the integrity of KMS by allowing the unsealing of its encrypted on-disk state and facilitating communication among nodes within a KMS cluster.

          It solves the challenges associated with billions of cryptographic keys and hundreds of thousands of cryptographic operations per node per second - which are commonplace in larger deployments.

          <b>High Availability and Fault Tolerance</b>
          In the dynamic landscape of large-scale systems, network or node outages are inevitable. Taking down a cluster for maintenance is rarely feasible. MinIO's AIStor KMS ensures uninterrupted availability, even when faced with such disruptions, mitigating cascading effects that can take down the entire storage infrastructure. Specifically, you could lose all but one node of a cluster and still handle any encryption, decryption or data key generation requests.

          <b>Scalability</b>
          While the amount of data usually only increases, the load on a large-scale storage system may vary significantly from time to time. MinIO's AIStor KMS supports dynamic cluster resizing and nodes can be added or removed at any point without incurring any downtime.

          <b>Multi-Tenancy</b>
          Large-scale storage infrastructures are often used by many applications and teams across the entire organization. Isolating teams and groups into their own namespaces is a core requirement. MinIO's AIStor KMS supports namespacing in the form of enclaves. Each tenant can be assigned its own enclave which is completely independent and isolated from all other enclaves on the KMS cluster.

          <b>Predictable Behavior</b>
          MinIO's AIStor KMS is designed to be easily managed, providing operators with the ability to comprehend its state intuitively. Due to its simple design, MinIO's AIStor KMS is significantly easier to operate than similar solutions that rely on more complex consensus algorithms like Raft, or Paxos.

          <b>Consistent and Performant</b>
          The responsiveness of the KMS for GET/PUT operations directly influences the overall efficiency and speed of the storage system. MinIO's AIStor KMS nodes don’t have to coordinate when handling such requests from the storage system. Therefore, the performance of a MinIO's AIStor KMS cluster increases linearly with the number of nodes. Further, MinIO's AIStor KMS supports request pipelining to handle hundreds of thousands of cryptographic operations per node and second.

          <b>Simplicity</b>
          Operating a KMS cluster does not require expertise in cryptography or distributed systems. Everything can be done from the AIStor Console.

          <a href="https://resources.min.io/request-a-demo-marketplaces/">Request a demo today</a>
        displayName: Minio AIStor Key management System (KMS)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Storage
        - MinIO
        links:
        - name: Minio AIStor
          url: https://min.io
        - name: Support
          url: https://subnet.min.io
        - name: Github
          url: https://github.com/minio/aistor
        maintainers:
        - email: dev@min.io
          name: MinIO
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: MinIO
          url: https://min.io
        relatedImages:
        - quay.io/minio/aistor/operator@sha256:f80139efa062fdd83f78318b20f371a827b68e854b03dd9706b9eba4c493aeb2
        version: 2025.8.19175300
      entries:
      - name: minio-key-manager-operator.v2025.8.19175300
        version: 2025.8.19175300
      - name: minio-key-manager-operator.v2025.8.1184927
        version: 2025.8.1184927
      - name: minio-key-manager-operator.v2025.7.1011319
        version: 2025.7.1011319
      - name: minio-key-manager-operator.v2025.6.3065135
        version: 2025.6.3065135
      - name: minio-key-manager-operator.v2025.5.12190907
        version: 2025.5.12190907
      name: stable
    defaultChannel: stable
    packageName: minio-key-manager-operator
    provider:
      name: MinIO
      url: https://min.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-acmpca-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-acmpca-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acmpca.services.k8s.aws/v1alpha1",
                "kind": "CertificateAuthority",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/acmpca-controller:1.1.1
          createdAt: "2025-09-19T20:51:07Z"
          description: AWS ACM PCA controller is a service controller for managing
            ACM PCA resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CertificateAuthority represents the state of an AWS acmpca
              CertificateAuthority resource.
            displayName: CertificateAuthority
            kind: CertificateAuthority
            name: certificateauthorities.acmpca.services.k8s.aws
            version: v1alpha1
          - description: CertificateAuthorityActivation represents the state of an
              AWS acmpca CertificateAuthorityActivation resource.
            displayName: CertificateAuthorityActivation
            kind: CertificateAuthorityActivation
            name: certificateauthorityactivations.acmpca.services.k8s.aws
            version: v1alpha1
          - description: Certificate represents the state of an AWS acmpca Certificate
              resource.
            displayName: Certificate
            kind: Certificate
            name: certificates.acmpca.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ACM PCA resources in AWS from within your Kubernetes cluster.

          **About Amazon ACM PCA**

          AWS Private CA enables creation of private certificate authority (CA) hierarchies, including root and subordinate CAs, without the investment and maintenance costs of operating an on-premises CA. Your private CAs can issue end-entity X.509 certificates useful in scenarios including:
          - Creating encrypted TLS communication channels
          - Authenticating users, computers, API endpoints, and IoT devices
          - Cryptographically signing code
          - Implementing Online Certificate Status Protocol (OCSP) for obtaining certificate revocation status

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ACM PCA
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acmpca
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ACM PCA Developer Resources
          url: https://aws.amazon.com/private-ca/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: acmpca maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/acmpca-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-acmpca-controller.v1.1.1
        version: 1.1.1
      - name: ack-acmpca-controller.v1.1.0
        version: 1.1.0
      - name: ack-acmpca-controller.v1.0.10
        version: 1.0.10
      - name: ack-acmpca-controller.v1.0.9
        version: 1.0.9
      - name: ack-acmpca-controller.v1.0.8
        version: 1.0.8
      - name: ack-acmpca-controller.v1.0.7
        version: 1.0.7
      - name: ack-acmpca-controller.v1.0.6
        version: 1.0.6
      - name: ack-acmpca-controller.v1.0.5
        version: 1.0.5
      - name: ack-acmpca-controller.v1.0.4
        version: 1.0.4
      - name: ack-acmpca-controller.v1.0.3
        version: 1.0.3
      - name: ack-acmpca-controller.v1.0.2
        version: 1.0.2
      - name: ack-acmpca-controller.v1.0.1
        version: 1.0.1
      - name: ack-acmpca-controller.v1.0.0
        version: 1.0.0
      - name: ack-acmpca-controller.v0.0.20
        version: 0.0.20
      - name: ack-acmpca-controller.v0.0.16
        version: 0.0.16
      - name: ack-acmpca-controller.v0.0.15
        version: 0.0.15
      - name: ack-acmpca-controller.v0.0.14
        version: 0.0.14
      - name: ack-acmpca-controller.v0.0.13
        version: 0.0.13
      - name: ack-acmpca-controller.v0.0.12
        version: 0.0.12
      - name: ack-acmpca-controller.v0.0.11
        version: 0.0.11
      - name: ack-acmpca-controller.v0.0.10
        version: 0.0.10
      - name: ack-acmpca-controller.v0.0.9
        version: 0.0.9
      - name: ack-acmpca-controller.v0.0.8
        version: 0.0.8
      - name: ack-acmpca-controller.v0.0.5
        version: 0.0.5
      - name: ack-acmpca-controller.v0.0.4
        version: 0.0.4
      - name: ack-acmpca-controller.v0.0.3
        version: 0.0.3
      - name: ack-acmpca-controller.v0.0.2
        version: 0.0.2
      - name: ack-acmpca-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-acmpca-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MayaData
      provider-url: ""
    name: openebs
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openebsoperator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "openebs.io/v1alpha1",
                "kind": "OpenEBSInstallTemplate",
                "metadata": {
                  "name": "oebs",
                  "namespace": "openebs"
                },
                "spec": {
                  "rbac": {
                    "create": true,
                    "pspEnabled": false,
                    "kyvernoEnabled": false
                  },
                  "serviceAccount": {
                    "create": true,
                    "name": "openebs-maya-operator"
                  },
                  "imagePullSecrets": [],
                  "release": {
                    "version": "3.0.0"
                  },
                  "legacy": {
                    "enabled": false
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": ""
                  },
                  "apiserver": {
                    "enabled": true,
                    "image": "openebs/m-apiserver",
                    "imageTag": "2.12.2",
                    "replicas": 1,
                    "ports": {
                      "externalPort": 5656,
                      "internalPort": 5656
                    },
                    "sparse": {
                      "enabled": "false"
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "defaultStorageConfig": {
                    "enabled": "true"
                  },
                  "varDirectoryPath": {
                    "baseDir": "/var/openebs"
                  },
                  "provisioner": {
                    "enabled": true,
                    "image": "openebs/openebs-k8s-provisioner",
                    "imageTag": "2.12.2",
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "patchJivaNodeAffinity": "enabled",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "localprovisioner": {
                    "enabled": true,
                    "image": "openebs/provisioner-localpv",
                    "imageTag": "3.0.0",
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "enableDeviceClass": true,
                    "enableHostpathClass": true,
                    "basePath": "/var/openebs/local",
                    "waitForBDBindTimeoutRetryCount": "12",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "snapshotOperator": {
                    "enabled": true,
                    "controller": {
                      "image": "openebs/snapshot-controller",
                      "imageTag": "2.12.2",
                      "resources": {}
                    },
                    "provisioner": {
                      "image": "openebs/snapshot-provisioner",
                      "imageTag": "2.12.2",
                      "resources": {}
                    },
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "upgradeStrategy": "Recreate",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    }
                  },
                  "ndm": {
                    "enabled": true,
                    "image": "openebs/node-disk-manager",
                    "imageTag": "1.7.0",
                    "sparse": {
                      "path": "/var/openebs/sparse",
                      "size": "10737418240",
                      "count": "0"
                    },
                    "filters": {
                      "enableOsDiskExcludeFilter": true,
                      "osDiskExcludePaths": "/,/etc/hosts,/boot",
                      "enableVendorFilter": true,
                      "excludeVendors": "CLOUDBYT,OpenEBS",
                      "enablePathFilter": true,
                      "includePaths": "",
                      "excludePaths": "/dev/loop,/dev/fd0,/dev/sr0,/dev/ram,/dev/dm-,/dev/md,/dev/rbd,/dev/zd"
                    },
                    "probes": {
                      "enableSeachest": false
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "ndmOperator": {
                    "enabled": true,
                    "image": "openebs/node-disk-operator",
                    "imageTag": "1.7.0",
                    "replicas": 1,
                    "upgradeStrategy": "Recreate",
                    "nodeSelector": {},
                    "tolerations": [],
                    "healthCheck": {
                      "initialDelaySeconds": 15,
                      "periodSeconds": 20
                    },
                    "readinessCheck": {
                      "initialDelaySeconds": 5,
                      "periodSeconds": 10
                    },
                    "resources": {}
                  },
                  "ndmExporter": {
                    "enabled": false,
                    "image": {
                      "registry": null,
                      "repository": "openebs/node-disk-exporter",
                      "pullPolicy": "IfNotPresent",
                      "tag": "1.7.0"
                    },
                    "nodeExporter": {
                      "name": "ndm-node-exporter",
                      "podLabels": {
                        "name": "openebs-ndm-node-exporter"
                      },
                      "metricsPort": 9101
                    },
                    "clusterExporter": {
                      "name": "ndm-cluster-exporter",
                      "podLabels": {
                        "name": "openebs-ndm-cluster-exporter"
                      },
                      "metricsPort": 9100
                    }
                  },
                  "webhook": {
                    "enabled": true,
                    "image": "openebs/admission-server",
                    "imageTag": "2.12.2",
                    "failurePolicy": "Fail",
                    "replicas": 1,
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "hostNetwork": false,
                    "resources": {}
                  },
                  "helper": {
                    "image": "openebs/linux-utils",
                    "helperImageTag": "3.0.0"
                  },
                  "featureGates": {
                    "enabled": true,
                    "GPTBasedUUID": {
                      "enabled": true,
                      "featureGateFlag": "GPTBasedUUID"
                    },
                    "APIService": {
                      "enabled": false,
                      "featureGateFlag": "APIService",
                      "address": "0.0.0.0:9115"
                    },
                    "UseOSDisk": {
                      "enabled": false,
                      "featureGateFlag": "UseOSDisk"
                    },
                    "ChangeDetection": {
                      "enabled": false,
                      "featureGateFlag": "ChangeDetection"
                    }
                  },
                  "crd": {
                    "enableInstall": true
                  },
                  "policies": {
                    "monitoring": {
                      "enabled": true,
                      "image": "openebs/m-exporter",
                      "imageTag": "2.12.2"
                    }
                  },
                  "analytics": {
                    "enabled": true,
                    "pingInterval": "24h"
                  },
                  "jiva": {
                    "image": "openebs/jiva",
                    "imageTag": "2.12.2",
                    "replicas": 3,
                    "defaultStoragePath": "/var/openebs",
                    "enabled": false,
                    "openebsLocalpv": {
                      "enabled": false
                    },
                    "localpv-provisioner": {
                      "openebsNDM": {
                        "enabled": false
                      }
                    }
                  },
                  "cstor": {
                    "pool": {
                      "image": "openebs/cstor-pool",
                      "imageTag": "2.12.2"
                    },
                    "poolMgmt": {
                      "image": "openebs/cstor-pool-mgmt",
                      "imageTag": "2.12.2"
                    },
                    "target": {
                      "image": "openebs/cstor-istgt",
                      "imageTag": "2.12.2"
                    },
                    "volumeMgmt": {
                      "image": "openebs/cstor-volume-mgmt",
                      "imageTag": "2.12.2"
                    },
                    "enabled": false,
                    "openebsNDM": {
                      "enabled": false
                    }
                  },
                  "openebs-ndm": {
                    "enabled": false
                  },
                  "localpv-provisioner": {
                    "enabled": false,
                    "openebsNDM": {
                      "enabled": false
                    }
                  },
                  "zfs-localpv": {
                    "enabled": false
                  },
                  "lvm-localpv": {
                    "enabled": false
                  },
                  "nfs-provisioner": {
                    "enabled": false
                  },
                  "cleanup": {
                    "image": {
                      "registry": "",
                      "repository": "bitnami/kubectl",
                      "tag": "",
                      "imagePullSecrets": []
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: index.docker.io/openebs/helm-operator:v0.0.9
          createdAt: "2021-11-02T02:56:58Z"
          description: Creates and maintains OpenEBS Control Plane deployments
          repository: https://github.com/openebs/helm-operator
          support: https://slack.openebs.io/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a OpenEBS Install Operator
            displayName: OpenEBS Install Template
            kind: OpenEBSInstallTemplate
            name: openebsinstalltemplates.openebs.io
            version: v1alpha1
        description: |
          **OpenEBS** is a leading container attached storage solution that enables the use of containers for mission-critical, persistent workloads and for other stateful workloads such as logging or Prometheus for example.
          OpenEBS itself is deployed as just another container on your host and enables storage services that can be designated on a per pod, application, cluster or container level, including:
          * Data persistence across nodes
          * Synchronization of data across availability zones and cloud providers
          * A common layer whether you are running on the cloud, or your bare metal
          * Integration with Kubernetes, so developer and application intent flows into OpenEBS
          * Management of tiering to and from S3 and other targets.
          ## OpenEBS Operator
          OpenEBS primarily provides container attached block storage (iSCSI volumes) by leveraging/aggregating the storage on the nodes, with the storage controller itself running as a container. Different storage engines (Jiva & cStor) are supported, with tools available to dynamically provision Kubernetes Local PVs. The volumes are dynamically provisioned via PersistentVolumeClaims and are managed by a control plane component called "maya", which also runs as a deployment in the K8s cluster. In addition to maya, a typical OpenEBS installation comprises several other resources, which aid with various functionalities, ranging from snapshotting to disk management. All these components are described briefly below:
          * **Maya-API-Server** - A storage orchestrator which integrates into Kubernetes workflow to help provision
            and manage OpenEBS Jiva & cStor (storage engine) volumes

          * **Dynamic-OpenEBS-Provisioner** - A Kubernetes external storage provisioner that utilizes APIs exposed by maya-apiserver
            to perform provision & delete operations of Jiva & cStor volumes

          * **Dynamic-LocalPV-Provisioner** - A dynamic provisioner for Kubernetes Local PVs
          * **OpenEBS-Snapshot-Operator** - A Kubernetes snapshot controller that creates & restores OpenEBS volume snapshots
          * **Node-Disk-Manager** - A disk management controller which identifies available disks, maintains inventory, and
            dynamically attaches/detaches disks to backend storage pods

          The helm-based OpenEBS Operator eases the setup of all the above mentioned components, with a simple custom resource provided to define the install options, thereby enabling applications to start using the OpenEBS storageclasses in their PVCs. The OpenEBSInstallTemplate CR can be used to specify start-up parameters & also update/overwrite the definitions post install.
          ## Pre-Requisites
          Before installing OpenEBS control plane, perform the following steps to ensure successful deployment of the Node-Disk-Manager & the OpenEBS volume replicas, respectively.
          * Configure the OpenEBS service account on the openshift-operators namespace/project to use the privileged security context constraint.

            **Note**: The serviceaccount name is same as the one specified in the `spec.serviceAccount.name` field of the OpenEBSInstallTemplate CR.


            ```
            oc adm policy add-scc-to-user privileged system:serviceaccount:openshift-operators:openebs-maya-operator
            ```

          * Configure the default service account on the namespace/project in which the volume replicas are deployed to use privileged
            security context constraint.


            ```
            oc adm policy add-scc-to-user privileged system:serviceaccount:openshift-operators:default
            ```

          ## Getting Started
          * Try the quickstart [guide](https://github.com/openebs/helm-operator/blob/master/olm/README.md)
          * To learn how to contribute, please read the [contribution guide](https://github.com/openebs/helm-operator/blob/master/CONTRIBUTING.md)
          * OpenEBS welcomes your feedback and contributions in any form possible. [Join our Community](https://openebs.org/community)

          ## License
          OpenEBS is distributed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt)
        displayName: OpenEBS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenEBS
        - ContainerAttachedStorage
        links:
        - name: OpenEBS Website
          url: https://openebs.io
        - name: Operator Source Code
          url: https://github.com/openebs/helm-operator
        - name: Install Instructions
          url: https://github.com/openebs/helm-operator/blob/master/olm/README.md
        maintainers:
        - email: kiran.mova@openebs.io
          name: kmova
        maturity: alpha
        minKubeVersion: 1.12.0
        provider:
          name: MayaData
        relatedImages:
        - index.docker.io/openebs/helm-operator:v0.0.9
        version: 3.0.0
      entries:
      - name: openebsoperator.v3.0.0
        version: 3.0.0
      - name: openebsoperator.v2.12.2
        version: 2.12.2
      name: alpha
    defaultChannel: alpha
    packageName: openebs
    provider:
      name: MayaData
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-core
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Core",
                "metadata": {
                  "name": "core-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configService": {
                    "port": 8080
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "helper": {
                    "port": 8080
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "licenseService": {
                    "port": 8080
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "restStatsService": {
                    "port": 8080
                  },
                  "resultService": {
                    "port": 8080,
                    "storage": "64Gi"
                  },
                  "sessionManager": {
                    "port": 8080
                  },
                  "statsService": {
                    "port": 8080
                  },
                  "testResultsService": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:34:00Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Core
            name: cores.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Core Operator
        displayName: Keysight WAP Core Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Core
          url: https://keysight-wap-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-core@sha256:001b0d08cd84a9ebc13cd8fd35b68d431b96f84d71e933c1123ed6c0d244fd37
        version: 0.1.21
      entries:
      - name: keysight-wap-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-nats-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-nats-operator.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2023-09-19T12:44:10Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NatsCluster
            name: natsclusters.nats.io
            version: v1alpha2
          - kind: NatsServiceRole
            name: natsserviceroles.nats.io
            version: v1alpha2
        description: Keysight Nats Operator
        displayName: Keysight Nats Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Nats Operator
          url: https://keysight-nats-operator.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-nats-operator@sha256:2ee5dc8814274e83a5f82e89eabd0b643e8cde8e3d555a56535a61d05c72fd75
        version: 0.1.21
      entries:
      - name: keysight-nats-operator.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-nats-operator
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: github.com/stakater-ab/hibernation-operator
    name: hibernation-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hibernation-operator.v0.1.103
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Monitoring
          createdAt: "2025-10-02T05:44:42Z"
          description: Operator to engage and disengage hibernation on k8s resources
            using cron expressions
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.1.103
          operatorframework.io/suggested-namespace: hibernation-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Cluster Resource Supervisor
            kind: ClusterResourceSupervisor
            name: clusterresourcesupervisors.hibernation.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.hibernation.stakater.com
            version: v1beta1
        description: Schedule k8s resources using cron expressions via Hibernation
          Operator
        displayName: Hibernation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - redhat
        - k8s
        - stakater
        - hibernation
        - operator
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maintainers:
        - email: hello@stakater.com
          name: StakaterAB
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Stakater
          url: github.com/stakater-ab/hibernation-operator
        relatedImages:
        - registry.connect.redhat.com/stakater/hibernation-operator@sha256:3738f82834d53aec4f8fba5caad6258c7adb4bdc23a57bde47c7ec22030fae5c
        version: 0.1.103
      entries:
      - name: hibernation-operator.v0.1.103
        version: 0.1.103
      name: release-0.1
    defaultChannel: release-0.1
    packageName: hibernation-operator
    provider:
      name: Stakater
      url: github.com/stakater-ab/hibernation-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: runtime-component-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runtime-component.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/runtime-component-operator@sha256:d37b3f62575c9ad1ffcb22c89e2c3886275c326bf27aa49fb1808acba475853d
          createdAt: "2022-03-25T14:00:00Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          olm.skipRange: <0.8.1
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          This advanced Operator is capable of deploying any runtime component image with consistent, production-grade QoS. It enables enterprise architects to govern the way their applications get deployed & managed in the cluster, while dramatically reducing the learning curve for developers to deploy into Kubernetes - allowing them to focus on writing the code!
          Here are some key features:

          #### Application Lifecyle
          You can deploy your runtime component container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the Operator will watch for any updates and will re-deploy the modified image.

          #### Custom RBAC
          This Operator is capable of using a custom ServiceAccount from the caller, allowing it to follow RBAC restrictions. By default it creates a ServiceAccount if one is not specified, which can also be bound with specific roles.

          #### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as: labels, annotations, and environment variables from a ConfigMap, a Secret or a value.

          #### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configuration, such as TLS settings, are also easily enabled.

          #### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto scaling to create (and delete) instances based on resource consumption.

          #### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you.
          This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart.
          We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.

          #### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating kubernetes Secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to its required services without any intervention or interruption.

          #### Exposing metrics to Prometheus
          The Runtime Component Operator exposes the runtime container's metrics via the Prometheus Operator.
          You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `spec.monitoring` field to control configurations such as the poll interval and security credentials.

          #### Easily mount logs and transaction directories
          If you need to mount the logs and transaction data from your runtime component to an external volume such as NFS (or any storage supported in your cluster), simply add the following (to specify the volume size and the location to persist) to your RuntimeComponent CR:
          ``` storage: size: 2Gi mountPath: "/logs" ```

          #### Integration with OpenShift Serverless
          Deploy your serverless runtime component using a single toggle.  The Operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing your pod to automatically scale to 0 when it is idle.

          #### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and how they are connected.

          See our [**documentation**](https://github.com/application-stacks/runtime-component-operator/tree/main/doc/) for more information.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://github.com/application-stacks/runtime-component-operator/tree/main/doc
        - name: Repository
          url: https://github.com/application-stacks/runtime-component-operator
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: beta
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4
        - registry.connect.redhat.com/ibm/runtime-component-operator@sha256:d37b3f62575c9ad1ffcb22c89e2c3886275c326bf27aa49fb1808acba475853d
        version: 0.8.1
      entries:
      - name: runtime-component.v0.8.1
        version: 0.8.1
      - name: runtime-component.v0.8.0
        version: 0.8.0
      name: beta2
    - currentCSV: runtime-component.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:1b818f1a9c27baab6a39ae42d86f78bac085d772cae5cd97b357d0ee5799cf85
          createdAt: "2023-08-28T16:46:50Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          olm.skipRange: '>=0.8.0 <1.2.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/runtime-component-operator@sha256:1b818f1a9c27baab6a39ae42d86f78bac085d772cae5cd97b357d0ee5799cf85
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02
        version: 1.2.2
      entries:
      - name: runtime-component.v1.2.2
        version: 1.2.2
      name: v1.2
    - currentCSV: runtime-component.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:682c5cb3e24566bf2933b1d4a4f6bf520d9b546edcedc1ee3214ec1ac4d48629
          createdAt: "2024-08-01T20:26:42Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.3.3'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/runtime-component-operator@sha256:682c5cb3e24566bf2933b1d4a4f6bf520d9b546edcedc1ee3214ec1ac4d48629
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91
        version: 1.3.3
      entries:
      - name: runtime-component.v1.3.3
        version: 1.3.3
      name: v1.3
    - currentCSV: runtime-component.v1.4.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:6f527c5173421c28d175b42d8c690eb63c4d1c5092c31d5350fd7a5e93661eaa
          createdAt: "2025-07-18T20:38:10Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.4.4'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/runtime-component-operator@sha256:6f527c5173421c28d175b42d8c690eb63c4d1c5092c31d5350fd7a5e93661eaa
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3
        version: 1.4.4
      entries:
      - name: runtime-component.v1.4.4
        version: 1.4.4
      name: v1.4
    - currentCSV: runtime-component.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:58bfceaed2fc828bce1c0e4e21a583d25f6d72f88abdc78df2510867a4f2020f
          createdAt: "2025-09-15T18:08:27Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.5.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/runtime-component-operator@sha256:58bfceaed2fc828bce1c0e4e21a583d25f6d72f88abdc78df2510867a4f2020f
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8
        version: 1.5.0
      entries:
      - name: runtime-component.v1.5.0
        version: 1.5.0
      name: v1.5
    defaultChannel: v1.5
    packageName: runtime-component-operator-certified
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: glance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: glance-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "glance.openstack.org/v1beta1",
                "kind": "Glance",
                "metadata": {
                  "name": "glance"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\nenabled_backends=foo:bar,foo1:bar1\n",
                  "databaseAccount": "glance",
                  "databaseInstance": "openstack",
                  "secret": "osp-secret",
                  "serviceUser": "glance",
                  "storage": {
                    "storageClass": "local-storage",
                    "storageRequest": "10G"
                  }
                }
              },
              {
                "apiVersion": "glance.openstack.org/v1beta1",
                "kind": "GlanceAPI",
                "metadata": {
                  "annotation": {
                    "keystoneEndpoint": "true"
                  },
                  "name": "default"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\nenabled_backends=foo:bar,foo1:bar1\n[foo]\nhere-foo-config\n[foo1]\nhere-foo1-config\n",
                  "databaseAccount": "glance",
                  "databaseHostname": "glance",
                  "preserveJobs": false,
                  "replicas": 1,
                  "secret": "glance-secret",
                  "serviceAccount": "glance",
                  "serviceUser": "glance",
                  "storage": {
                    "storageRequest": "10G"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:45Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GlanceAPI is the Schema for the glanceapis API
            displayName: Glance API
            kind: GlanceAPI
            name: glanceapis.glance.openstack.org
            version: v1beta1
          - description: Glance is the Schema for the glances API
            displayName: Glance
            kind: Glance
            name: glances.glance.openstack.org
            version: v1beta1
        description: Glance Operator
        displayName: Glance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Image
        - Glance
        links:
        - name: Glance Operator
          url: https://github.com/openstack-k8s-operators/glance-operator
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/glance-rhel9-operator@sha256:74f9907b59493239fbe9c5f534399ff513c693a2bda11a037aa2cc5dc8cc3e91
        - registry.redhat.io/rhoso/openstack-glance-api-rhel9@sha256:b8f84f9b432a70f7ca7aebda60455a4890347529f6bed9a440a3b197e377e0a1
        version: 1.0.6
      entries:
      - name: glance-operator.v1.0.6
        version: 1.0.6
      - name: glance-operator.v1.0.5
        version: 1.0.5
      - name: glance-operator.v1.0.4
        version: 1.0.4
      - name: glance-operator.v1.0.3
        version: 1.0.3
      - name: glance-operator.v1.0.2
        version: 1.0.2
      - name: glance-operator.v1.0.1
        version: 1.0.1
      - name: glance-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: glance-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kafka-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kafka-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kafka-controller:1.2.1
          createdAt: "2025-09-20T21:41:07Z"
          description: AWS Kafka controller is a service controller for managing Kafka
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster represents the state of an AWS kafka Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.kafka.services.k8s.aws
            version: v1alpha1
          - description: Configuration represents the state of an AWS kafka Configuration
              resource.
            displayName: Configuration
            kind: Configuration
            name: configurations.kafka.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCluster represents the state of an AWS kafka ServerlessCluster
              resource.
            displayName: ServerlessCluster
            kind: ServerlessCluster
            name: serverlessclusters.kafka.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Kafka resources in AWS from within your Kubernetes cluster.

          **About Amazon Kafka**

          Amazon Managed Streaming for Apache Kafka (Amazon MSK) is a fully managed AWS streaming data service makes it easy for you to build and run applications that use Apache Kafka to process streaming data without needing to become an expert in operating Apache Kafka clusters. Amazon MSK manages the provisioning, configuration, and maintenance of Apache Kafka clusters and Apache ZooKeeper nodes for you. Amazon MSK also shows key Apache Kafka performance metrics in the AWS Console.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Kafka Developer Resources
          url: https://aws.amazon.com/msk/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kafka maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/kafka-controller:1.2.1
        version: 1.2.1
      entries:
      - name: ack-kafka-controller.v1.2.1
        version: 1.2.1
      - name: ack-kafka-controller.v1.2.0
        version: 1.2.0
      - name: ack-kafka-controller.v1.1.3
        version: 1.1.3
      - name: ack-kafka-controller.v1.1.2
        version: 1.1.2
      - name: ack-kafka-controller.v1.1.1
        version: 1.1.1
      - name: ack-kafka-controller.v1.1.0
        version: 1.1.0
      - name: ack-kafka-controller.v1.0.14
        version: 1.0.14
      - name: ack-kafka-controller.v1.0.13
        version: 1.0.13
      - name: ack-kafka-controller.v1.0.12
        version: 1.0.12
      - name: ack-kafka-controller.v1.0.11
        version: 1.0.11
      - name: ack-kafka-controller.v1.0.10
        version: 1.0.10
      - name: ack-kafka-controller.v1.0.8
        version: 1.0.8
      - name: ack-kafka-controller.v1.0.7
        version: 1.0.7
      - name: ack-kafka-controller.v1.0.6
        version: 1.0.6
      - name: ack-kafka-controller.v1.0.5
        version: 1.0.5
      - name: ack-kafka-controller.v1.0.4
        version: 1.0.4
      - name: ack-kafka-controller.v1.0.3
        version: 1.0.3
      - name: ack-kafka-controller.v1.0.2
        version: 1.0.2
      - name: ack-kafka-controller.v1.0.1
        version: 1.0.1
      - name: ack-kafka-controller.v1.0.0
        version: 1.0.0
      - name: ack-kafka-controller.v0.0.14
        version: 0.0.14
      - name: ack-kafka-controller.v0.0.12
        version: 0.0.12
      - name: ack-kafka-controller.v0.0.11
        version: 0.0.11
      - name: ack-kafka-controller.v0.0.10
        version: 0.0.10
      - name: ack-kafka-controller.v0.0.9
        version: 0.0.9
      - name: ack-kafka-controller.v0.0.8
        version: 0.0.8
      - name: ack-kafka-controller.v0.0.7
        version: 0.0.7
      - name: ack-kafka-controller.v0.0.6
        version: 0.0.6
      - name: ack-kafka-controller.v0.0.5
        version: 0.0.5
      - name: ack-kafka-controller.v0.0.4
        version: 0.0.4
      - name: ack-kafka-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-kafka-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Apache Software Foundation
      provider-url: ""
    name: camel-k
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: camel-k-operator.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.0.1
          createdAt: "2023-09-04T12:32:38Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
          - kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.0.1
        version: 2.0.1
      entries:
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: candidate
    - currentCSV: camel-k-operator.v2.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "setBody": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationProfile",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "debug"
                },
                "spec": {
                  "traits": {
                    "logging": {
                      "level": "DEBUG"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example-source"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "setBody": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "timer-to-log-example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "log-sink"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world",
                      "period": 2000
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "timer-source"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.8.0
          createdAt: "2025-09-21T09:07:43Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: IntegrationProfile is the Schema for the integrationprofiles
              API
            displayName: Integration Profile
            kind: IntegrationProfile
            name: integrationprofiles.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Pipe is the Schema for the pipes API
            displayName: Pipe
            kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers:
          the easiest way to build and manage your Camel applications on Kubernetes.

          ## Installation

          You may need to specify the container registry where to store the container images created for the Camel applications.
          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          spec:
            build:
              registry:
                address: docker.io/my-org
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step when using Openshift cluster.

          ## Running an Integration

          Run a Camel application on the cluster by creating a sample `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            annotations:
              camel.apache.org/operator.id: camel-k
            name: test
          spec:
            flows:
            - from:
                parameters:
                  period: "1000"
                steps:
                - setBody:
                    simple: Hello Camel from ${routeId}
                - log: ${body}
                uri: timer:yaml
          ```
          Log the result of the running application:
          ```
          kubectl wait --for=condition=ready integration test --timeout=180s
          kubectl logs deployment/test -f
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.8.0
        version: 2.8.0
      entries:
      - name: camel-k-operator.v2.8.0
        version: 2.8.0
      - name: camel-k-operator.v2.7.0
        version: 2.7.0
      - name: camel-k-operator.v2.6.0
        version: 2.6.0
      - name: camel-k-operator.v2.5.1
        version: 2.5.1
      - name: camel-k-operator.v2.5.0
        version: 2.5.0
      - name: camel-k-operator.v2.4.0
        version: 2.4.0
      - name: camel-k-operator.v2.3.3
        version: 2.3.3
      - name: camel-k-operator.v2.3.2
        version: 2.3.2
      - name: camel-k-operator.v2.3.1
        version: 2.3.1
      - name: camel-k-operator.v2.3.0
        version: 2.3.0
      - name: camel-k-operator.v2.2.0
        version: 2.2.0
      - name: camel-k-operator.v2.1.0
        version: 2.1.0
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: latest
    - currentCSV: camel-k-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.12.0
          createdAt: "2023-02-27T09:01:55Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.12.0
        version: 1.12.0
      entries:
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable
    - currentCSV: camel-k-operator.v1.10.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.10.4
          createdAt: "2023-01-28T13:49:05Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.10.4
        version: 1.10.4
      entries:
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.10
    - currentCSV: camel-k-operator.v1.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.11.1
          createdAt: "2023-01-30T08:45:04Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.11.1
        version: 1.11.1
      entries:
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.11
    - currentCSV: camel-k-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.12.1
          createdAt: "2023-06-12T08:46:58Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.12.1
        version: 1.12.1
      entries:
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.12
    - currentCSV: camel-k-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.5.1
          createdAt: "2021-08-31T09:03:41Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.5.1
        version: 1.5.1
      entries:
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.5
    - currentCSV: camel-k-operator.v1.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.6.1
          createdAt: "2021-10-27T21:06:20Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.6.1
        version: 1.6.1
      entries:
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.6
    - currentCSV: camel-k-operator.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.7.0
          createdAt: "2021-11-16T10:54:12Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.8.1+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.7.0
        version: 1.7.0
      entries:
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.7
    - currentCSV: camel-k-operator.v1.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.8.2
          createdAt: "2022-03-07T07:13:19Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.8.2
        version: 1.8.2
      entries:
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.8
    - currentCSV: camel-k-operator.v1.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.9.2
          createdAt: "2022-05-23T09:05:22Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.9.2
        version: 1.9.2
      entries:
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.9
    - currentCSV: camel-k-operator.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.0.1
          createdAt: "2023-09-04T12:32:38Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
          - kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.0.1
        version: 2.0.1
      entries:
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-2.0
    - currentCSV: camel-k-operator.v2.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "setBody": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationProfile",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "debug"
                },
                "spec": {
                  "traits": {
                    "logging": {
                      "level": "DEBUG"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example-source"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "setBody": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "timer-to-log-example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "log-sink"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world",
                      "period": 2000
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "timer-source"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.8.0
          createdAt: "2025-09-21T09:07:43Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: IntegrationProfile is the Schema for the integrationprofiles
              API
            displayName: Integration Profile
            kind: IntegrationProfile
            name: integrationprofiles.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Pipe is the Schema for the pipes API
            displayName: Pipe
            kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers:
          the easiest way to build and manage your Camel applications on Kubernetes.

          ## Installation

          You may need to specify the container registry where to store the container images created for the Camel applications.
          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          spec:
            build:
              registry:
                address: docker.io/my-org
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step when using Openshift cluster.

          ## Running an Integration

          Run a Camel application on the cluster by creating a sample `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            annotations:
              camel.apache.org/operator.id: camel-k
            name: test
          spec:
            flows:
            - from:
                parameters:
                  period: "1000"
                steps:
                - setBody:
                    simple: Hello Camel from ${routeId}
                - log: ${body}
                uri: timer:yaml
          ```
          Log the result of the running application:
          ```
          kubectl wait --for=condition=ready integration test --timeout=180s
          kubectl logs deployment/test -f
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.8.0
        version: 2.8.0
      entries:
      - name: camel-k-operator.v2.8.0
        version: 2.8.0
      - name: camel-k-operator.v2.7.0
        version: 2.7.0
      - name: camel-k-operator.v2.6.0
        version: 2.6.0
      - name: camel-k-operator.v2.5.1
        version: 2.5.1
      - name: camel-k-operator.v2.5.0
        version: 2.5.0
      - name: camel-k-operator.v2.4.0
        version: 2.4.0
      - name: camel-k-operator.v2.3.3
        version: 2.3.3
      - name: camel-k-operator.v2.3.2
        version: 2.3.2
      - name: camel-k-operator.v2.3.1
        version: 2.3.1
      - name: camel-k-operator.v2.3.0
        version: 2.3.0
      - name: camel-k-operator.v2.2.0
        version: 2.2.0
      - name: camel-k-operator.v2.1.0
        version: 2.1.0
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-v2
    defaultChannel: stable-v2
    packageName: camel-k
    provider:
      name: The Apache Software Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-cpd-ccs
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-ccs.v10.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ccs.cpd.ibm.com/v1beta1",
                "kind": "CCS",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ccs-cr",
                    "app.kubernetes.io/managed-by": "ibm-cpd-ccs-operator",
                    "app.kubernetes.io/name": "ibm-cpd-ccs-operator"
                  },
                  "name": "ccs-cr"
                },
                "spec": {
                  "scaleConfig": "small",
                  "license": {
                    "accept": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
          createdAt: "2025-04-18"
          description: Common Core Services
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-ccs
          olm.skipRange: '>=7.0.0 <10.3.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM watsonx.ai"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: |-
            [
              "runtimeassemblies.runtimes.ibm.com"
            ]
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation Runtime Manager CRD. This product is not deployable
              standalone and it is part of the LICCR already accepted.  For additional
              details regarding install parameters check: http://ibm.biz/-readme.
              License By installing this product you accept the license terms https://ibm.biz/BdaaLC
              or https://ibm.biz/BdaaL9.'
            kind: RuntimeAssembly
            name: runtimeassemblies.runtimes.ibm.com
            version: v1alpha1
          - description: 'Documentation This product is not deployable standalone
              and it is part of the LICCR already accepted.  For additional details
              regarding install parameters check: http://ibm.biz/-readme. License
              By installing this product you accept the license terms https://ibm.biz/BdaaLC
              or https://ibm.biz/BdaaL9.'
            kind: CCS
            name: ccs.ccs.cpd.ibm.com
            version: v1beta1
        description: |-
          # Introduction Common Core Services Operator
          Put stuff here

          # Details
          Various Details
          ## Prerequisites
          Various Prerequisites
          ## Resources Required
          Various Resources Required
          ## Installing
          Various Installation details
          # Configuration
          Various Configuration information
          ### Limitations
          Various Limitations
          ### SecurityContextConstraints Requirements
          All pods will use the restricted SCC, I think.
        displayName: Common Core Services
        installModes:
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: OwnNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commonc-core-services
        - cpd40
        links:
        - name: Ccs
          url: https://ccs.domain
        maintainers:
        - email: test@ibm.com
          name: test
        maturity: v1
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:ad19b65042fb26cc1244a92f81a96e5ceb4313f8c823181efe28b66e9fecf9c8
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:81d8b25221d827160dd5e88efd383453f532057ad02d7a2ee3dcd9c6eb16ba65
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:0621eb9b83c9b2b8bfe5cd09797fe08fa37a854a5506914935ff6f0df599b631
        - registry.marketplace.redhat.com/cp/cpd/runtime-manager-api@sha256:a37773df07fe757dd698e3f09f8687902e7b3592d97855e703dd92ecc832aab0
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:6fe90f051d9261ad224dccf4d9f6bc9f19ce0e8d22fa010a46b2752374bce2b3
        - registry.marketplace.redhat.com/cp/cpd/spaces-ui-job@sha256:95dade51444e30ec034c12207ecbef675b70889b81970f49f10aa75160b2f947
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-bundle@sha256:94c7a1dd290e2abd55c5b8998cf08ad174d3fbfc659ae6fa6fb6cbab61d5fe14
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:f5527ddbbd0b9878555fce852996b13020fd91b1604d8c03d34a61ca2fdb2d1d
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:0e0ae96aff8540323da8a2379af014d3af1770e4b0043b84b9b415dda44501f7
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:6f96b8b6afb00cb26fcf903b8e4c30b0d6344bb86d8bce2f2026fc8cb0b575ec
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:740a1d87a43e61e32cfdea032d852af33100d959c8b57ff2f165f4baf77fc011
        - registry.marketplace.redhat.com/cp/cpd/task-credentials@sha256:e162828d5ad07ceae4db0589f930979842714e3b766552102bee373a3a570af3
        - registry.marketplace.redhat.com/cp/cpd/environment-specifications-sync@sha256:894b4624d8ead7286cf9068c6477300992c32d96339644fe764bcb29b1c8ef5c
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:c268d2852580286e5c6770857fcfc3b06a698521f5acd17ed83b8e2cb9ab55cf
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:bcdf9b7eea8753b3e463973a2690140ae7aa8c02c3c7726511f116bd751ec2d8
        - registry.marketplace.redhat.com/cp/cpd/ngp-projects-api@sha256:df420e9f616ad8a123d9eedd117fb6cb007b30a772138db534f47ad8801fa96d
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:044b663e4b7508c62dbb28b7fb2aa267c9771fc62ab509b08d46ea1499d95431
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:257bf301338f7b123099e2ffe31c3f3b793bf28d6a0570a25b889271239579ba
        - registry.marketplace.redhat.com/cp/cpd/dap-base-job@sha256:f208686a3f56d2980e7b8640b6ebdea9057109e7520c6817a731e68d08506a36
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:4fb2127ba7c525a959bbbcf024de364ef266ebb3ca7e696f74d990d2552cda59
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:6cffef7631ee54612fba71efc0eff4e17bc08defc3e2137aee9881d685444298
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:8a6ce0dada706056efda73ad0cc1547566003e0a7aab8ed379b50e4776861b94
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:21c81ff7c4b8809accaf62b242fd9b3d52f235489ae427556b7f71cc21bf8926
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e12599ff99db058fdc8f9c1eebfba31160c636ff092138deb1f6f6705f44fd62
        - registry.marketplace.redhat.com/cp/cpd/environments-ui@sha256:2923664475c7a8ea5e63055c90748db37825fe76c7975b845f1185d4b32ee4ea
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
        - registry.marketplace.redhat.com/cp/cpd/environments-api@sha256:6dd60b530f81d7a2d55a534211f7f128814abf87b8cd3575d528dba53442c44e
        - registry.marketplace.redhat.com/cp/cpd/jobs-api@sha256:173b88f90078e1817686ac949f48387af70a706e23e6fdab4a7c6217db5ad86e
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:d023f35b65816f244b6d699b41362e6031553f7d4c416efcee4a7f2395a61355
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:722944bfd0eba280255f4c198f65f562423a26e2c4ad3580591ee71cb6e9c2ef
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:56315ce0a0a70f5899191a7eab856105489143f0e889c8cdd1b6b8c225f02eb7
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:a3a0d58f895fca4d832e602d0dee80018df4ad7bad8800ad4058eea07fda4817
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:fbc6eab07fcfc2d0c1895f18e17f751e9481ad111875959e9ff08e14f81b9a35
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:523dcc671f1c05e690d5e578a90300d28d76570f605d4de7f80645b590f4a462
        - registry.marketplace.redhat.com/cp/cpd/environments-init-job@sha256:6af8f900af7c11eee6894e2c3e52f740b3c700dac0ee753fadef330230bd42dd
        - registry.marketplace.redhat.com/cp/cpd/jobs-ui@sha256:49daf607918b378593485da16ac417056ee295bca94f1a5cac37b334381ae096
        - registry.marketplace.redhat.com/cp/cpd/job-trigger@sha256:f2ea487072393a24e4c6864b2107c1fce072ebba07f9dbc2c139bd37fbd062da
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:562281dbb2bec5191a7c6b595ab853092e6fe3ff1432e77b2471f71ccdffcbc8
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:6f2e7cc917adb54b3112dc4909f645ac08c33e8c2c33266fbc8f5133cbcd1973
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:351375de35d7fc9d940b905b867ce8a00b7ff1416e70d3a100918797fe3d8c56
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:54c86367bfe0ab658611856cc0a6c5de24c14a336bd129ac9f16293945a55238
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:0ae22914dbaf09d6679bd06bc9b6a5728e744c9ce2309cbd5f01e28308c403ea
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:3f2f97e6b2b18526255e88ff02d1039b24312eca51ee22b7cbf76d5680f7aee1
        version: 10.3.0
      entries:
      - name: ibm-cpd-ccs.v10.3.0
        version: 10.3.0
      name: v10.3
    defaultChannel: v10.3
    packageName: ibm-cpd-ccs
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: swift-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: swift-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "swift.openstack.org/v1beta1",
                "kind": "Swift",
                "metadata": {
                  "name": "swift"
                },
                "spec": {
                  "swiftProxy": {
                    "passwordSelectors": {
                      "service": "SwiftPassword"
                    },
                    "replicas": 1
                  },
                  "swiftRing": {
                    "ringReplicas": 1
                  },
                  "swiftStorage": {
                    "replicas": 1,
                    "storageClass": "local-storage"
                  }
                }
              },
              {
                "apiVersion": "swift.openstack.org/v1beta1",
                "kind": "SwiftProxy",
                "metadata": {
                  "name": "swift-proxy"
                },
                "spec": {
                  "passwordSelectors": {
                    "service": "SwiftPassword"
                  },
                  "replicas": 1,
                  "secret": "osp-secret",
                  "serviceUser": "swift"
                }
              },
              {
                "apiVersion": "swift.openstack.org/v1beta1",
                "kind": "SwiftRing",
                "metadata": {
                  "name": "swift-ring"
                },
                "spec": {
                  "ringReplicas": 1
                }
              },
              {
                "apiVersion": "swift.openstack.org/v1beta1",
                "kind": "SwiftStorage",
                "metadata": {
                  "name": "swift-storage"
                },
                "spec": {
                  "replicas": 1,
                  "storageClass": "local-storage",
                  "storageRequest": "10G"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:34Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SwiftProxy is the Schema for the swiftproxies API
            displayName: Swift Proxy
            kind: SwiftProxy
            name: swiftproxies.swift.openstack.org
            version: v1beta1
          - description: SwiftRing is the Schema for the swiftrings API
            displayName: Swift Ring
            kind: SwiftRing
            name: swiftrings.swift.openstack.org
            version: v1beta1
          - description: Swift is the Schema for the swifts API
            displayName: Swift
            kind: Swift
            name: swifts.swift.openstack.org
            version: v1beta1
          - description: SwiftStorage is the Schema for the swiftstorages API
            displayName: Swift Storage
            kind: SwiftStorage
            name: swiftstorages.swift.openstack.org
            version: v1beta1
        description: OpenStack Swift operator
        displayName: Swift operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Swift
        - Object Storage
        links:
        - name: Swift Operator
          url: https://github.com/openstack-k8s-operators/swift-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-swift-object-rhel9@sha256:656f23b25ca0bd80d7b77ba717a21455366dfaf2049df1bd6d6536925a18b94b
        - registry.redhat.io/rhoso/openstack-swift-proxy-server-rhel9@sha256:6e654db83d570a3f8feedb912d5d9dff35ecba4d53169481a3b4982d3a832798
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/swift-rhel9-operator@sha256:18207f02c049447836211b326f09ce660782c384cda92e592db256692e3c8e73
        - registry.redhat.io/rhoso/openstack-swift-account-rhel9@sha256:ecebf3eb430188392f7a7b8cda3d1c0c0f658bcbbaff531d99eba88a7e0c9889
        - registry.redhat.io/rhoso/openstack-swift-container-rhel9@sha256:782866f8e9661de50d6a74bc52262f4546d7e0055d0c4da83ba6b891a50c35ea
        version: 1.0.6
      entries:
      - name: swift-operator.v1.0.6
        version: 1.0.6
      - name: swift-operator.v1.0.5
        version: 1.0.5
      - name: swift-operator.v1.0.4
        version: 1.0.4
      - name: swift-operator.v1.0.3
        version: 1.0.3
      - name: swift-operator.v1.0.2
        version: 1.0.2
      - name: swift-operator.v1.0.1
        version: 1.0.1
      - name: swift-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: swift-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/component: kmm
      app.kubernetes.io/name: kmm
      app.kubernetes.io/part-of: kmm
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: kernel-module-management
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kernel-module-management.v2.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:1022205d712a2ed9d6e76968961c180d4e66a2e0fb991221b0b02ed94ed830d2
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.0.2'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel-module-management
        - out-of-tree
        - kernel
        - kmm
        - module
        - modules
        - oot
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:1022205d712a2ed9d6e76968961c180d4e66a2e0fb991221b0b02ed94ed830d2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62c44057a27abd41cf3daf60b8736da59a82dadb65fa6c32bef65d19bd49ea49
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:25f39f20f325f64676bf8ef14cd8a60cfbdc77116753a89d4bb667ea2a86172c
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:192e08b121a04ae69aea60c35199e8f5e3c42305e1ebd024a605a371de4ab472
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:94b4ddc45cb34d4ffe794a51b99c868cb709f90cc4eb8b28284d70c526a0a226
        version: 2.0.2
      entries:
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.0
    - currentCSV: kernel-module-management.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a85037778c53c3766f39efe4c4d2523d01ae7743e5c5693b05bd8b2cbb50b1d8
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.1.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a85037778c53c3766f39efe4c4d2523d01ae7743e5c5693b05bd8b2cbb50b1d8
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:611556d999ea97f9dbb78fec5963baa02de7a26626d7096ba63b177e3bcf9707
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:14e100e5853444114029116255ebfb7a54e5ed51503ba969740ad48b5ca80f70
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:1a9a19df17b464d1708b2905d03b8726e4480ca1f11aa97a15ff4c049a24ea4b
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:ea3e5fbf7eaf6bd29031e2876d7e24859779fb4d3d7ef27c26a1d310f137ad17
        version: 2.1.1
      entries:
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.1
    - currentCSV: kernel-module-management.v2.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a224708950f8defb259ef0da49011d361c5c0378404350640600c6e8fbdbd606
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.2.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:b1d98c0f14448ae8eb80821cbec236150c338674aae2a2b37f8d0ab0d5f1430c
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:53fddf0b226eb579c71c86112a8d9b6b1d90a1498d6f6985932b16b8be532f35
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:c3dc9ad4ebc68cd85a074b520903a9310e72a90a73bdf4aa7c42b7ed66b7bcba
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:ce1d68fd96f5003285f9fbc0c7abd8aa23830dbdd9b76629d38d4d2238e30032
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a224708950f8defb259ef0da49011d361c5c0378404350640600c6e8fbdbd606
        version: 2.2.1
      entries:
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.2
    - currentCSV: kernel-module-management.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:750dcc75f908cdda86932386fa1a28edaf65d7b5379ef6d0447586a064db906f
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.3.0'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:750dcc75f908cdda86932386fa1a28edaf65d7b5379ef6d0447586a064db906f
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:c5005b74402f680d11c9ea682a9a0d3f74b6ab0d01d25298d232f910f1a29c34
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:055c2d6e80bd1ec225f0cb2041572ebb28cfd1a4ed82ba47ff0d638d52ded2a8
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:f9fde334af68220943017a7491130e6a82f15e63522f5bb3931aa7d9b00ccc40
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:06ac089cc89b303e525fd71d6439b958f34ab944a72209f1cc8cb562ecfd1112
        version: 2.3.0
      entries:
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.3
    - currentCSV: kernel-module-management.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:41Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:8e0a3cfb1cb92d6d3c9ab5a3030593d8df1378c3ce755d598fb77b08ea5793d3
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        version: 2.4.1
      entries:
      - name: kernel-module-management.v2.4.1
        version: 2.4.1
      - name: kernel-module-management.v2.4.0
        version: 2.4.0
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.4
    - currentCSV: kernel-module-management.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:41Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:8e0a3cfb1cb92d6d3c9ab5a3030593d8df1378c3ce755d598fb77b08ea5793d3
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        version: 2.4.1
      entries:
      - name: kernel-module-management.v2.4.1
        version: 2.4.1
      - name: kernel-module-management.v2.4.0
        version: 2.4.0
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: kernel-module-management
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio Project
      provider-url: ""
    name: apicurito
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: 1.0.x
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: candidate
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: latest
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: latest
    packageName: apicurito
    provider:
      name: Apicurio Project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: mongodb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"opstreelabs.in/v1alpha1","kind":"MongoDB","metadata":{"name":"mongodb"},"spec":{"kubernetesConfig":{"image":"quay.io/opstree/mongo:v5.0.6","imagePullPolicy":"IfNotPresent","securityContext":{"fsGroup":1001}},"storage":{"accessModes":["ReadWriteOnce"],"storageSize":"1Gi","storageClass":"gp2"},"mongoDBSecurity":{"mongoDBAdminUser":"admin","secretRef":{"name":"mongodb-secret","key":"password"}}}},{"apiVersion":"opstreelabs.in/v1alpha1","kind":"MongoDBCluster","metadata":{"name":"mongodb"},"spec":{"clusterSize":3,"kubernetesConfig":{"image":"quay.io/opstree/mongo:v5.0.6","imagePullPolicy":"IfNotPresent","securityContext":{"fsGroup":1001}},"storage":{"accessModes":["ReadWriteOnce"],"storageSize":"1Gi","storageClass":"gp2"},"mongoDBSecurity":{"mongoDBAdminUser":"admin","secretRef":{"name":"mongodb-secret","key":"password"}}}}]'
          capabilities: Seamless Upgrades
          categories: Database,  OpenShift Optional, Storage
          certified: "false"
          containerImage: quay.io/opstree/mongodb-operator:v0.3.0
          createdAt: 05-04-2022
          description: A golang based operator to create and manage MongoDB standalone
            setup and cluster setup
          k8sMaxVersion: ""
          k8sMinVersion: 1.17.0
          repository: https://github.com/ot-container-kit/mongodb-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Mongo DB Cluster
            displayName: MongoDBCluster
            kind: MongoDBCluster
            name: mongodbclusters.opstreelabs.in
            version: v1alpha1
          - description: Mongo DB
            displayName: MongoDB
            kind: MongoDB
            name: mongodbs.opstreelabs.in
            version: v1alpha1
        description: |+
          MongoDB Operator is an operator created in Golang to create, update, and manage MongoDB standalone, replicated, and arbiter replicated setup on Kubernetes and Openshift clusters. This operator is capable of doing the setup for MongoDB with all the required best practices.

          ### Supported Features

          * MongoDB replicated cluster setup
          * MongoDB standalone setup
          * MongoDB replicated cluster failover and recovery
          * Monitoring support with MongoDB Exporter
          * Password based authentication for MongoDB
          * Kubernetes's resources for MongoDB standalone and cluster
          * Customizable configuration changes in MongoDB
          * Insightful Grafana dashboards

          ### Documentation

          **[Documentation](https://ot-mongodb-operator.netlify.app/)**
          ### Getting Help

          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.
          ### Contributing
          You can contribute to this project by:-

          * Raising Bugs and Feature Requests
          * Fixing issue by opening a Pull Request
          * Improving Documentation

          ### License

          MongoDB  Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/mongodb-operator/blob/master/LICENSE)


        displayName: MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - MongoDB
        - Opstree
        - Operatork
        - Cloud Native
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-mongodb-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: alpha
        minKubeVersion: 1.17.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/opstree/mongodb-operator:v0.3.0
        version: 0.3.0
      entries:
      - name: mongodb-operator.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: mongodb-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: https://konveyor.io
    name: move2kube-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: move2kube-operator.v0.3.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {
                  "deployment": {
                    "api": {
                      "cpu": "300m",
                      "imageTag": "",
                      "initContainer": false,
                      "memory": "1Gi",
                      "privilegedPods": false,
                      "replicas": 1
                    },
                    "authServer": {
                      "cpu": "300m",
                      "enable": false,
                      "memory": "1Gi",
                      "replicas": 1
                    },
                    "database": {
                      "cpu": "300m",
                      "enable": false,
                      "memory": "1Gi",
                      "replicas": 1,
                      "startEmpty": true
                    }
                  },
                  "ingress": {
                    "enable": true,
                    "host": "mydomain.com",
                    "preferRoute": false,
                    "tlsSecretName": ""
                  },
                  "persistentVolumeClaim": {
                    "createNew": true,
                    "enable": false,
                    "name": "",
                    "readWriteMany": true,
                    "storageClassName": "",
                    "storageSize": "1Gi"
                  },
                  "secret": {
                    "api": {
                      "configYAML": "",
                      "createNew": true,
                      "enable": false,
                      "name": ""
                    },
                    "authServer": {
                      "adminPassword": "password",
                      "adminUsername": "admin",
                      "createNew": true,
                      "databasePassword": "password",
                      "databaseUsername": "auth-server",
                      "name": "",
                      "realmJSON": "",
                      "standaloneHAXML": ""
                    }
                  },
                  "securityContextContraints": {
                    "enable": false
                  },
                  "serviceAccount": {
                    "createNew": true,
                    "enable": false,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.10
          createdAt: "2023-09-25T02:56:15Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/konveyor/move2kube-operator:v0.3.10
        version: 0.3.10
      entries:
      - name: move2kube-operator.v0.3.10
        version: 0.3.10
      - name: move2kube-operator.v0.3.10-rc.0
        version: 0.3.10-rc.0
      - name: move2kube-operator.v0.3.9
        version: 0.3.9
      - name: move2kube-operator.v0.3.8
        version: 0.3.8
      name: alpha
    - currentCSV: move2kube-operator.v0.3.15-rc.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.15-rc.0
          createdAt: "2024-10-18T16:10:56Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/konveyor/move2kube-operator:v0.3.15-rc.0
        version: 0.3.15-rc.0
      entries:
      - name: move2kube-operator.v0.3.15-rc.0
        version: 0.3.15-rc.0
      - name: move2kube-operator.v0.3.14-rc.0
        version: 0.3.14-rc.0
      - name: move2kube-operator.v0.3.13-rc.0
        version: 0.3.13-rc.0
      - name: move2kube-operator.v0.3.11-rc.0
        version: 0.3.11-rc.0
      name: prerelease
    - currentCSV: move2kube-operator.v0.3.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.15
          createdAt: "2025-03-07T06:09:48Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - quay.io/konveyor/move2kube-operator:v0.3.15
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 0.3.15
      entries:
      - name: move2kube-operator.v0.3.15
        version: 0.3.15
      - name: move2kube-operator.v0.3.14
        version: 0.3.14
      - name: move2kube-operator.v0.3.13
        version: 0.3.13
      - name: move2kube-operator.v0.3.11
        version: 0.3.11
      name: stable
    defaultChannel: stable
    packageName: move2kube-operator
    provider:
      name: Konveyor
      url: https://konveyor.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Dell Technologies
      provider-url: https://github.com/dell/csm-operator
    name: dell-csm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dell-csm-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "authorization",
                  "namespace": "authorization"
                },
                "spec": {
                  "modules": [
                    {
                      "components": [
                        {
                          "enabled": true,
                          "name": "nginx"
                        },
                        {
                          "enabled": true,
                          "name": "cert-manager"
                        },
                        {
                          "authorizationController": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231",
                          "authorizationControllerReplicas": 1,
                          "certificate": "",
                          "controllerReconcileInterval": "5m",
                          "enabled": true,
                          "hostname": "csm-authorization.com",
                          "leaderElection": true,
                          "name": "proxy-server",
                          "opa": "docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095",
                          "opaKubeMgmt": "docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517",
                          "openTelemetryCollectorAddress": "",
                          "privateKey": "",
                          "proxyServerIngress": [
                            {
                              "annotations": {},
                              "hosts": [],
                              "ingressClassName": "nginx"
                            }
                          ],
                          "proxyService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682",
                          "proxyServiceReplicas": 1,
                          "roleService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c",
                          "roleServiceReplicas": 1,
                          "storageService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5",
                          "storageServiceReplicas": 1,
                          "tenantService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358",
                          "tenantServiceReplicas": 1
                        },
                        {
                          "commander": "docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a",
                          "name": "redis",
                          "redis": "docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f",
                          "redisCommander": "rediscommander",
                          "redisName": "redis-csm",
                          "redisReplicas": 5,
                          "redisSecretProviderClass": [
                            {
                              "redisPasswordKey": "",
                              "redisSecretName": "",
                              "redisUsernameKey": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "sentinel": "sentinel"
                        },
                        {
                          "configSecretProviderClass": [
                            {
                              "configSecretName": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "name": "config"
                        },
                        {
                          "name": "storage-system-credentials",
                          "secretProviderClasses": {
                            "vault": [
                              "secret-provider-class-1",
                              "secret-provider-class-2"
                            ]
                          }
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": true,
                      "forceRemoveModule": true,
                      "name": "authorization-proxy-server"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "isilon",
                  "namespace": "isilon"
                },
                "spec": {
                  "driver": {
                    "authSecret": "isilon-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VERBOSE",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_ISI_PORT",
                          "value": "8080"
                        },
                        {
                          "name": "X_CSI_ISI_PATH",
                          "value": "/ifs/data/csi"
                        },
                        {
                          "name": "X_CSI_ISI_NO_PROBE_ON_START",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_AUTOPROBE",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_AUTH_TYPE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_CUSTOM_TOPOLOGY_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOISILON_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "AZ_RECONCILE_INTERVAL",
                          "value": "1h"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_ISI_QUOTA_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_ACCESS_ZONE",
                          "value": "System"
                        },
                        {
                          "name": "X_CSI_ISI_VOLUME_PATH_PERMISSIONS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_IGNORE_UNRESOLVABLE_HOSTS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "isilon",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerscale"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSCALE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_PERFORMANCE_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSCALE_QUOTA_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "ISICLIENT_INSECURE",
                              "value": "true"
                            },
                            {
                              "name": "ISICLIENT_AUTH_TYPE",
                              "value": "1"
                            },
                            {
                              "name": "ISICLIENT_VERBOSE",
                              "value": "0"
                            },
                            {
                              "name": "POWERSCALE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSCALE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0",
                          "name": "metrics-powerscale"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-isilon/csi_sock",
                            "--mode=node",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powermax",
                  "namespace": "powermax"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powermax-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_MANAGED_ARRAYS",
                          "value": "000000000001,000000000002"
                        },
                        {
                          "name": "X_CSI_POWERMAX_PORTGROUPS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_TRANSPORT_PROTOCOL",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_K8S_CLUSTER_PREFIX",
                          "value": "CSM"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "X_CSI_VSPHERE_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VSPHERE_PORTGROUP",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VSPHERE_HOSTNAME",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VCENTER_HOST",
                          "value": ""
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "X_CSI_POWERMAX_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "X_CSI_POWERMAX_DEBUG",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_REVPROXY_USE_SECRET",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powermax",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERMAX_ISCSI_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_TOPOLOGY_CONTROL_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": [
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/memory-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/disk-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/network-unavailable",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REVPROXY_TLS_SECRET",
                              "value": "csirevproxy-tls-secret"
                            },
                            {
                              "name": "X_CSI_REVPROXY_PORT",
                              "value": "2222"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            },
                            {
                              "name": "DeployAsSidecar",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265",
                          "name": "csipowermax-reverseproxy"
                        }
                      ],
                      "configVersion": "v2.14.0",
                      "name": "csireverseproxy"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powermax"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERMAX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERMAX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414",
                          "name": "metrics-powermax"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/powermax.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powerstore",
                  "namespace": "powerstore"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powerstore-config",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_NODE_NAME_PREFIX",
                          "value": "csi-node"
                        },
                        {
                          "name": "X_CSI_FC_PORTS_FILTER_FILE_PATH",
                          "value": "/etc/fc-ports-filter"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOPOWERSTORE_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSM_AUTH_ENABLED",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_MULTI_NAS_FAILURE_THRESHOLD",
                          "value": "5"
                        },
                        {
                          "name": "X_CSI_MULTI_NAS_COOLDOWN_PERIOD",
                          "value": "5m"
                        },
                        {
                          "name": "X_CSI_NFS_ACLS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerstore",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-powerstore.dellemc.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerstore"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSTORE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_SPACE_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_ARRAY_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_FILE_SYSTEM_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSTORE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSTORE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "ZIPKIN_URI",
                              "value": ""
                            },
                            {
                              "name": "ZIPKIN_SERVICE_NAME",
                              "value": "metrics-powerstore"
                            },
                            {
                              "name": "ZIPKIN_PROBABILITY",
                              "value": "0.0"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60",
                          "name": "metrics-powerstore"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "unity",
                  "namespace": "unity"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_UNITY_ALLOW_MULTI_POD_ACCESS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_EPHEMERAL_STAGING_PATH",
                          "value": "/var/lib/kubelet/plugins/kubernetes.io/csi/pv/"
                        },
                        {
                          "name": "X_CSI_ISCSI_CHROOT",
                          "value": "/noderoot"
                        },
                        {
                          "name": "X_CSI_UNITY_SYNC_NODEINFO_INTERVAL",
                          "value": "15"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOUNITY_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOUNITY_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "TENANT_NAME",
                          "value": ""
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_UNITY_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "unity",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "vxflexos",
                  "namespace": "vxflexos"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLELISTVOLUMESNAPSHOT",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLESNAPSHOTCGDELETE",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "INFO"
                        },
                        {
                          "name": "GOSCALEIO_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOSCALEIO_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_QUOTA_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "INTERFACE_NAMES",
                          "value": null
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERFLEX_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "File",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerflex",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "initContainers": [
                      {
                        "envs": [
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          },
                          {
                            "name": "REPO_ADDRESS",
                            "value": "sftp://0.0.0.0"
                          },
                          {
                            "name": "REPO_USER",
                            "value": "sdcRepoUser"
                          },
                          {
                            "name": "MODULE_SIGCHECK",
                            "value": "0"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "imagePullPolicy": "IfNotPresent",
                        "name": "sdc"
                      }
                    ],
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_SDC_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_APPROVE_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_PREFIX",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_SDC_SFTP_REPO_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 1,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "enabled": false,
                        "envs": [
                          {
                            "name": "HOST_PID",
                            "value": "1"
                          },
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "name": "sdc-monitor"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "csi-external-health-monitor-controller"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERFLEX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_SDC_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_SDC_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERFLEX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d",
                          "name": "metrics-powerflex"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerflex"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--csisock=unix:/var/lib/kubelet/plugins/vxflexos.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
          createdAt: "2025-09-05T13:18:42Z"
          description: Easily install and manage Dell’s CSI Drivers and CSM
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/dell/csm-operator
          support: Dell Technologies
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ContainerStorageModule is the Schema for the containerstoragemodules
              API
            displayName: Container Storage Module
            kind: ContainerStorageModule
            name: containerstoragemodules.storage.dell.com
            version: v1
        description: "Dell Container Storage Modules (CSM) Operator is a Kubernetes
          Operator which can be \nused to install and manage Dell’s CSI drivers and
          CSM modules. \nBy using Dell CSM Operator, enterprises can quickly and easily
          deploy the \nCSM modules for CSI drivers making it easy for DevOps \nteams
          to build and optimize their Kubernetes workloads using enterprise grade
          storage \nwhile consistently aligning with the desired outcomes of the organization.\nThis
          operator is available as both community and certified operator for Openshift
          \nand as a community operator for upstream Kubernetes.\nOperator can be
          deployed via OperatorHub.io. using Operator Lifecycle Manager (OLM)\n\n###
          About Dell CSM\n\nDell CSM brings powerful enterprise storage features and
          functionality to Kubernetes for \neasier adoption of cloud native workloads,
          improved productivity and scalable operations. \n\nCSM delivers software
          modules for storage management that provide developers access to \nbuild
          automation for enhanced IT needs and other critical enterprise storage features.
          \n\nCurrently, the following modules are available as part of Dell CSM.
          \n*\tObservability - Delivers a high-level view of the storage capacity
          \n  and performance usage via Grafana dashboards to the Kubernetes users\n*\tReplication
          - Enables array replication capabilities for Kubernetes users\n*\tAuthorization
          - Provides both storage and Kubernetes administrators the ability \n  to
          apply RBAC for Dell CSI Drivers\n*\tResiliency - Designed to make Kubernetes
          applications more resilient to \n  node failures\n\nThese modules are available
          for various storage platforms in the Dell portfolio, \nsuch as Dell PowerStore,
          Dell PowerMax, Dell PowerFlex, Dell PowerScale, and Dell Unity XT.\n\nFor
          more information, visit [Dell’s Container Storage Module web page](https://www.dell.com/en-us/dt/solutions/kubernetes-containers/kubernetes-storage.htm)\n\nAdditional
          Materials\n*\t[Documentation](https://dell.github.io/csm-docs)\n*\t[CSM
          Solution Brief](https://www.delltechnologies.com/asset/en-us/products/storage/briefs-summaries/csm-solution-brief.pdf)\n"
        displayName: Dell Container Storage Modules
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dell Container Storage Modules
        - Dell CSI Driver
        - Dell CSM Modules
        - Powerflex
        - Powerscale
        - Powerstore
        - Unity
        - Authorization
        - Observability
        - Replication
        links:
        - name: Documentation
          url: https://dell.github.io/csm-docs
        maintainers:
        - email: container.storage.modules@dell.com
          name: Dell Container Storage Modules
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Dell Technologies
          url: https://github.com/dell/csm-operator
        relatedImages:
        - registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da
        - quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540
        - registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa
        - registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
        - docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517
        - registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea
        - registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682
        - registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0
        - ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd
        - registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265
        - docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a
        - registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d
        - docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f
        - quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16
        - docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:59335acaca515fceed5f0350c46cfcfd6037475f0cf35ae8ab41d37333039e2e
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414
        version: 1.10.0
      entries:
      - name: dell-csm-operator.v1.10.0
        version: 1.10.0
      - name: dell-csm-operator.v1.9.1
        version: 1.9.1
      - name: dell-csm-operator.v1.9.0
        version: 1.9.0
      - name: dell-csm-operator.v1.8.1
        version: 1.8.1
      - name: dell-csm-operator.v1.8.0
        version: 1.8.0
      - name: dell-csm-operator.v1.7.0
        version: 1.7.0
      - name: dell-csm-operator.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: dell-csm-operator
    provider:
      name: Dell Technologies
      url: https://github.com/dell/csm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: raftt.io
      provider-url: https:///www.raftt.io
    name: raftt-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: raftt-operator.v1.25.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "raftt.io/v1beta1",
                "kind": "Raftt",
                "metadata": {
                  "name": "raftt",
                  "namespace": "raftt"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/raftt/raftt-operator@sha256:73edc6d8416ee3815d7e59bb26c33e3d35d97cf1ceadd8fa457041be13e175ca
          createdAt: "2025-04-29T11:31:02Z"
          description: This operator installs and configures a Raftt instance
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/raftt-operator/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/raftt-operator/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: raftt
          operators.openshift.io/valid-subscription: Raftt commercial license
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Raftt systems
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Raftt represents an installation of Raftt cluster controller
              & supporting services
            displayName: Raftt
            kind: Raftt
            name: raftts.raftt.io
            version: v1beta1
        description: |
          Raftt allows fast code iterations and debugging over complex Kubernetes environments.
          Raftt provides a localhost development experience for applications running on Kubernetes.
          You can use Raftt to orchestrate the environments for you, so you get the optimal development experience -
          iterations that take under a second, OOTB debugging of your code as it runs in the cluster, easy introspection, and more.
        displayName: Raftt
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dx
        - development
        - debugging
        links:
        - name: Raftt Operator
          url: https://raftt-operator.domain
        - name: Documentation
          url: https://docs.raftt.io
        maintainers:
        - email: ops@raftt.io
          name: Raftt
        maturity: alpha
        provider:
          name: raftt.io
          url: https:///www.raftt.io
        relatedImages:
        - quay.io/raftt/raftt-operator@sha256:73edc6d8416ee3815d7e59bb26c33e3d35d97cf1ceadd8fa457041be13e175ca
        version: 1.25.2
      entries:
      - name: raftt-operator.v1.25.2
        version: 1.25.2
      name: alpha
    defaultChannel: alpha
    packageName: raftt-operator
    provider:
      name: raftt.io
      url: https:///www.raftt.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HashiCorp
      provider-url: https://www.hashicorp.com/
    name: vault-secrets-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vault-secrets-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultConnection",
                "metadata": {
                  "name": "vaultconnection-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "address": "http://vault.vault.svc.cluster.local:8200"
                }
              },
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultAuth",
                "metadata": {
                  "name": "vaultauth-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "vaultConnectionRef": "vaultconnection-sample",
                  "method": "kubernetes",
                  "mount": "kubernetes",
                  "kubernetes": {
                    "role": "sample",
                    "serviceAccount": "default"
                  }
                }
              },
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultStaticSecret",
                "metadata": {
                  "name": "vaultstaticsecret-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "vaultAuthRef": "vaultauth-sample",
                  "mount": "kvv2",
                  "type": "kv-v2",
                  "path": "secret",
                  "refreshAfter": "5s",
                  "destination": {
                    "create": true,
                    "name": "secret1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.connect.redhat.com/hashicorp/vault-secrets-operator@sha256:84ac7003c4cb90c35f1e296585bb58187a86d64f972133e2661839bdb532ddb8
          createdAt: "2025-09-26T21:50:51Z"
          description: The Vault Secrets Operator (VSO) allows Pods to consume Vault
            secrets natively from Kubernetes Secrets.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/hashicorp/vault-secrets-operator
          support: HashiCorp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CSISecrets
            name: csisecrets.secrets.hashicorp.com
            version: v1beta1
          - description: HCPAuth is the Schema for the hcpauths API
            displayName: HCPAuth
            kind: HCPAuth
            name: hcpauths.secrets.hashicorp.com
            version: v1beta1
          - description: HCPVaultSecretsApp is the Schema for the hcpvaultsecretsapps
              API
            displayName: HCPVault Secrets App
            kind: HCPVaultSecretsApp
            name: hcpvaultsecretsapps.secrets.hashicorp.com
            version: v1beta1
          - description: SecretTransformation is the Schema for the secrettransformations
              API
            displayName: Secret Transformation
            kind: SecretTransformation
            name: secrettransformations.secrets.hashicorp.com
            version: v1beta1
          - description: VaultAuthGlobal is the Schema for the vaultauthglobals API
            displayName: Vault Auth Global
            kind: VaultAuthGlobal
            name: vaultauthglobals.secrets.hashicorp.com
            version: v1beta1
          - description: VaultAuth is the Schema for the vaultauths API
            displayName: Vault Auth
            kind: VaultAuth
            name: vaultauths.secrets.hashicorp.com
            version: v1beta1
          - description: VaultConnection is the Schema for the vaultconnections API
            displayName: Vault Connection
            kind: VaultConnection
            name: vaultconnections.secrets.hashicorp.com
            version: v1beta1
          - description: VaultDynamicSecret is the Schema for the vaultdynamicsecrets
              API
            displayName: Vault Dynamic Secret
            kind: VaultDynamicSecret
            name: vaultdynamicsecrets.secrets.hashicorp.com
            version: v1beta1
          - description: VaultPKISecret is the Schema for the vaultpkisecrets API
            displayName: Vault PKISecret
            kind: VaultPKISecret
            name: vaultpkisecrets.secrets.hashicorp.com
            version: v1beta1
          - description: VaultStaticSecret is the Schema for the vaultstaticsecrets
              API
            displayName: Vault Static Secret
            kind: VaultStaticSecret
            name: vaultstaticsecrets.secrets.hashicorp.com
            version: v1beta1
        description: |-
          The Vault Secrets Operator (VSO) allows Pods to consume Vault secrets
          natively from Kubernetes Secrets.

          ## Overview

          The Vault Secrets Operator operates by watching for changes to its supported set of Custom Resource Definitions (CRD).
          Each CRD provides the specification required to allow the *Operator* to synchronize a Vault Secrets to a Kubernetes Secret.
          The *Operator* writes the *source* Vault secret data directly to the *destination* Kubernetes Secret, ensuring that any
          changes made to the *source* are replicated to the *destination* over its lifetime. In this way, an application only needs
          to have access to the *destination* secret in order to make use of the secret data contained within.

          See the developer docs for more info [here](https://developer.hashicorp.com/vault/docs/platform/k8s/vso), including
          [examples](https://developer.hashicorp.com/vault/docs/platform/k8s/vso/examples) and a
          [tutorial](https://developer.hashicorp.com/vault/tutorials/kubernetes/vault-secrets-operator)
          for getting started.
        displayName: Vault Secrets Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vault
        - security
        - secrets
        links:
        - name: Vault Secrets Operator
          url: https://github.com/hashicorp/vault-secrets-operator
        - name: Documentation
          url: https://developer.hashicorp.com/vault/docs/platform/k8s/vso
        maintainers:
        - email: support@hashicorp.com
          name: HashiCorp Support
        maturity: stable
        provider:
          name: HashiCorp
          url: https://www.hashicorp.com/
        relatedImages:
        - registry.connect.redhat.com/hashicorp/vault-secrets-operator@sha256:84ac7003c4cb90c35f1e296585bb58187a86d64f972133e2661839bdb532ddb8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:503d5ee0fbcd5bd3492785f4246e19cbdb92960be0d05f3c6f123a32fe0fe84d
        version: 1.0.1
      entries:
      - name: vault-secrets-operator.v1.0.1
        version: 1.0.1
      - name: vault-secrets-operator.v1.0.0
        version: 1.0.0
      - name: vault-secrets-operator.v0.10.0
        version: 0.10.0
      - name: vault-secrets-operator.v0.9.1
        version: 0.9.1
      - name: vault-secrets-operator.v0.9.0
        version: 0.9.0
      - name: vault-secrets-operator.v0.8.1
        version: 0.8.1
      - name: vault-secrets-operator.v0.8.0
        version: 0.8.0
      - name: vault-secrets-operator.v0.7.0
        version: 0.7.0
      - name: vault-secrets-operator.v0.6.0
        version: 0.6.0
      - name: vault-secrets-operator.v0.5.2
        version: 0.5.2
      - name: vault-secrets-operator.v0.5.1
        version: 0.5.1
      - name: vault-secrets-operator.v0.5.0
        version: 0.5.0
      - name: vault-secrets-operator.v0.4.3
        version: 0.4.3
      name: stable
    defaultChannel: stable
    packageName: vault-secrets-operator
    provider:
      name: HashiCorp
      url: https://www.hashicorp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ab Initio
      provider-url: ""
    name: abinitio-runtime-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: abinitio-runtime-operator.v4.3.3-1
      currentCSVDesc:
        annotations:
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: abinitio-system
          operators.operatorframework.io/builder: operator-sdk-v1.8.0-ocp
          operators.operatorframework.io/internal-objects: '["cooperatingsystemruntimepoolmetrics.cloud.abinitio.com",
            "cooperatingsystemruntimeprofiles.cloud.abinitio.com"]'
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Defines a *runtime claim*, an interface for reserving a runtime
              object (CoOperatingSystemRuntime) in a runtime pool (CoOperatingSystemRuntimePool)
              in order to launch a job.
            displayName: Co>Operating System Runtime Claim
            kind: CoOperatingSystemRuntimeClaim
            name: cooperatingsystemruntimeclaims.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Pool Metric
            kind: CoOperatingSystemRuntimePoolMetric
            name: cooperatingsystemruntimepoolmetrics.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime pool*, a set of templated runtime objects
              (CoOperatingSystemRuntime) that are available on demand for running
              jobs. A runtime pool's capacity is configurable and can be scheduled
              in response to usage patterns.
            displayName: Co>Operating System Runtime Pool
            kind: CoOperatingSystemRuntimePool
            name: cooperatingsystemruntimepools.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Profile
            kind: CoOperatingSystemRuntimeProfile
            name: cooperatingsystemruntimeprofiles.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime object*, a set of coordinated resources
              used to run a data processing job in parallel across a launcher pod
              and a dynamically scaled set of worker pods. Such a job leverages the
              full power of the Co>Operating System for processing complex enterprise
              data.
            displayName: Co>Operating System Runtime
            kind: CoOperatingSystemRuntime
            name: cooperatingsystemruntimes.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime template*, a reusable specification (CoOperatingSystemRuntimeSpec)
              that can be referenced and overridden by runtime objects as needed,
              making it possible to specify jobs concisely. The runtime objects in
              a runtime pool (CoOperatingSystemRuntimePool) are created from a runtime
              template.
            displayName: Co>Operating System Runtime Template
            kind: CoOperatingSystemRuntimeTemplate
            name: cooperatingsystemruntimetemplates.cloud.abinitio.com
            version: v1
        description: |-
          The Co>Operating System runtime operator brings the power of Ab Initio's Co>Operating
          System to Kubernetes by defining a set of custom API resources for provisioning and
          running data processing jobs.

          The operator is installed along with the following custom resource definitions (CRDs):

          - CoOperatingSystemRuntime
          - CoOperatingSystemRuntimeClaim
          - CoOperatingSystemRuntimePool
          - CoOperatingSystemRuntimeTemplate

          In combination, these resources give you significant flexibility in scheduling, running,
          and monitoring your jobs, which can run in parallel across a dynamically scaled set of
          containers. For reference documentation for these resources, see
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-FOUR-Reference/Co-Operating-System-runtime-operator-API-reference)
          in the Ab Initio Help.

          The following topics describe the basic requirements and procedures for creating custom
          resources for running jobs after you have installed the runtime operator.

          ## Requirements for using custom resources

          The runtime operator is fully configured when you install it in your OpenShift cluster
          through OperatorHub. Importantly, the operator is installed in unrestricted mode,
          which means it can manage the CoOperatingSystemRuntime family of CRDs in any namespace
          in your cluster.

          For creating custom resources in order to run jobs using the operator, you must meet
          the requirements and configure the settings described in the following sections.

          ### Container images for job agents

          To run jobs of kind CoOperatingSystemRuntime, you must have a way to build the following
          OCI container images:

          - A *processing* image with the Co>Operating System server software installed in it.
          - A *sandbox workspace* image with the graphs and other files needed to run your jobs.
          This image is based on the Co>Operating System image and functions as the job's launcher
          agent, which coordinates with a set of worker agents (processing containers) in order to
          run a job in parallel.

          For building and managing these container images, you need the following:

          - A Co>Operating System installation package.
          - Access to a technical repository containing the projects that make up your application.
          - A container-building tool.
          - Access to a private container registry for hosting the images you build. You need pull
          and push access to this registry from your container build context and pull access from
          your OpenShift cluster (for example, via a Kubernetes Secret).

          > WARNING! Never push container images with Ab Initio software to a public registry. Push
          them to a private container registry only.

          The Co>Operating System includes reference Dockerfiles and example scripts for building
          container images. Typically, you build these images as part of your CI/CD pipeline. For
          more information, contact your Ab Initio administrator or Ab Initio Support.

          ### Ab Initio key servers

          The key servers that you use to enable jobs must be reachable from your cluster. You
          specify these key servers via the `AB_HOSTNAME_KEYSERVER_URLS` configuration variable in
          the pod specifications for launcher and worker agents in the CoOperatingSystemRuntime
          resource.

          To discuss the software activation key requirements for jobs that run in Kubernetes,
          contact your key server administrator or Ab Initio Support.

          ### Namespace and role considerations

          The runtime operator is installed with the cluster-wide RBAC roles and role bindings that
          it needs in order to manage custom resources and run jobs.

          You typically use a separate namespace or namespaces for creating and managing custom
          resources to run jobs. Each of these namespaces must be equipped with a service
          account that can list pod limits and get and patch CoOperatingSystemRuntime resources.

          For each namespace in which you intend to create custom resources, create the needed RBAC
          role and role binding by applying the following Role and RoleBinding manifests:

          ```yaml
          apiVersion: rbac.authorization.k8s.io/v1
          kind: Role
          metadata:
            name: cooperatingsystemruntime-role
          rules:
          - apiGroups:
            - ""
            resources:
            - limitranges
            verbs:
            - list
          - apiGroups:
            - cloud.abinitio.com
            resources:
            - cooperatingsystemruntimes
            verbs:
            - get
            - patch
          ---
          apiVersion: rbac.authorization.k8s.io/v1
          kind: RoleBinding
          metadata:
            name: run-cooperatingsystemruntime-rolebinding
          subjects:
          - kind: ServiceAccount
            name: default
          roleRef:
            kind: Role
            name: cooperatingsystemruntime-role
            apiGroup: rbac.authorization.k8s.io
          ```

          One way to do this is to save these manifests to a file and specify a namespace when
          using `oc` to apply them to the cluster. In the following example, the manifests are
          saved to a file named `runtime-rbac.yaml` and the role and role binding are to be created
          for the `default` service account in the `hello-world` namespace:

          ```
          oc -n hello-world apply -f runtime-rbac.yaml
          ```

          To bind the role to a service account other than `default`, simply update the value of
          `subjects[0].name` in the provided RoleBinding manifest.

          ### Dynamic volume provisioning

          The launcher and worker template sections of the CoOperatingSystemRuntime resource
          require you to specify persistent volume claims (PVCs) for job recovery data. To ensure
          that these PVCs can be bound to approriate storage volumes, dynamic volume provisioning
          must be enabled in your cluster.

          Work with your cluster administrator to identify the storage needs of your jobs and to
          ensure that dynamic volume provisioning is enabled so that PVCs can be matched to
          persistent volumes of the right storage class.

          For more information about dynamic storage provisioning, see the
          [OpenShift documentation](https://docs.openshift.com/container-platform/latest/storage/dynamic-provisioning.html).

          ## Quick start and examples

          For an overview of the key concepts associated with running jobs in Kubernetes, see
          ["About running jobs in Kubernetes"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-TWO-Running-jobs-in-Kubernetes/About-running-jobs-in-Kubernetes)
          in the Ab Initio Help.

          As described in this topic, you might begin by creating a CoOperatingSystemRuntime
          manifest that defines the configuration and compute resources needed for a standalone test
          job. You could then use the specification for this job to create a CoOperatingSystemRuntimeTemplate
          resource, which would collect common settings and make it possible to concisely specify
          CoOperatingSystemRuntime jobs.

          You could also use a runtime pool (of kind CoOperatingSystemRuntime) to extend these basic
          resources: such a pool consists of an actively running group of templated runtime objects whose
          pods can be used to run jobs on demand. Finally, for reserving resources in a pool, you could
          create a CoOperatingSystemRuntimeClaim resource.

          For more information about these custom resources, refer to the
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-FOUR-Reference/Co-Operating-System-runtime-operator-API-reference).

          For a set of integrated examples (including complete YAML manifests) that show you how to
          create and use the custom resources, see ["Example manifests for running jobs"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-TWO-Running-jobs-in-Kubernetes/Configuring-and-running-Co-Operating-System-jobs-in-Kubernetes/Example-manifests-for-CoOperatingSystemRuntime-jobs).
        displayName: Co>Operating System Runtime Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - abinitio
        - runtime
        - operator
        links:
        - name: Ab Initio
          url: https://www.abinitio.com/
        maintainers:
        - email: support@abinitio.com
          name: Ab Initio Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Ab Initio
        relatedImages:
        - registry.connect.redhat.com/abinitio/cooperating-system-runtime-operator@sha256:665efa8cc2ba97ad31f219161cdd53e15c1aef16b6c05ac60506be9bbc9ffc9a
        version: 4.3.3-1
      entries:
      - name: abinitio-runtime-operator.v4.3.3-1
        version: 4.3.3-1
      - name: abinitio-runtime-operator.v4.3.1-0
        version: 4.3.1-0
      name: release-4.3
    - currentCSV: abinitio-runtime-operator.v4.2.3-0
      currentCSVDesc:
        annotations:
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: abinitio-system
          operators.operatorframework.io/builder: operator-sdk-v1.8.0-ocp
          operators.operatorframework.io/internal-objects: '["cooperatingsystemruntimepoolmetrics.cloud.abinitio.com",
            "cooperatingsystemruntimeprofiles.cloud.abinitio.com"]'
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Defines a *runtime claim*, an interface for reserving a runtime
              object (CoOperatingSystemRuntime) in a runtime pool (CoOperatingSystemRuntimePool)
              in order to launch a job.
            displayName: Co>Operating System Runtime Claim
            kind: CoOperatingSystemRuntimeClaim
            name: cooperatingsystemruntimeclaims.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Pool Metric
            kind: CoOperatingSystemRuntimePoolMetric
            name: cooperatingsystemruntimepoolmetrics.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime pool*, a set of templated runtime objects
              (CoOperatingSystemRuntime) that are available on demand for running
              jobs. A runtime pool's capacity is configurable and can be scheduled
              in response to usage patterns.
            displayName: Co>Operating System Runtime Pool
            kind: CoOperatingSystemRuntimePool
            name: cooperatingsystemruntimepools.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Profile
            kind: CoOperatingSystemRuntimeProfile
            name: cooperatingsystemruntimeprofiles.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime object*, a set of coordinated resources
              used to run a data processing job in parallel across a launcher pod
              and a dynamically scaled set of worker pods. Such a job leverages the
              full power of the Co>Operating System for processing complex enterprise
              data.
            displayName: Co>Operating System Runtime
            kind: CoOperatingSystemRuntime
            name: cooperatingsystemruntimes.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime template*, a reusable specification (CoOperatingSystemRuntimeSpec)
              that can be referenced and overridden by runtime objects as needed,
              making it possible to specify jobs concisely. The runtime objects in
              a runtime pool (CoOperatingSystemRuntimePool) are created from a runtime
              template.
            displayName: Co>Operating System Runtime Template
            kind: CoOperatingSystemRuntimeTemplate
            name: cooperatingsystemruntimetemplates.cloud.abinitio.com
            version: v1
        description: |-
          The Co>Operating System runtime operator brings the power of Ab Initio's Co>Operating
          System to Kubernetes by defining a set of custom API resources for provisioning and
          running data processing jobs.

          The operator is installed along with the following custom resource definitions (CRDs):

          - CoOperatingSystemRuntime
          - CoOperatingSystemRuntimeClaim
          - CoOperatingSystemRuntimePool
          - CoOperatingSystemRuntimeTemplate

          In combination, these resources give you significant flexibility in scheduling, running,
          and monitoring your jobs, which can run in parallel across a dynamically scaled set of
          containers. For reference documentation for these resources, see
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Co_Operating_System_runtime_operator_API_reference)
          in the Ab Initio Help.

          The following topics describe the basic requirements and procedures for creating custom
          resources for running jobs after you have installed the runtime operator.

          ## Requirements for using custom resources

          The runtime operator is fully configured when you install it in your OpenShift cluster
          through OperatorHub. Importantly, the operator is installed in unrestricted mode,
          which means it can manage the CoOperatingSystemRuntime family of CRDs in any namespace
          in your cluster.

          For creating custom resources in order to run jobs using the operator, you must meet
          the requirements and configure the settings described in the following sections.

          ### Container images for job agents

          To run jobs of kind CoOperatingSystemRuntime, you must have a way to build the following
          OCI container images:

          - A *processing* image with the Co>Operating System server software installed in it.
          - A *sandbox workspace* image with the graphs and other files needed to run your jobs.
          This image is based on the Co>Operating System image and functions as the job's launcher
          agent, which coordinates with a set of worker agents (processing containers) in order to
          run a job in parallel.

          For building and managing these container images, you need the following:

          - A Co>Operating System installation package.
          - Access to a technical repository containing the projects that make up your application.
          - A container-building tool.
          - Access to a private container registry for hosting the images you build. You need pull
          and push access to this registry from your container build context and pull access from
          your OpenShift cluster (for example, via a Kubernetes Secret).

          > WARNING! Never push container images with Ab Initio software to a public registry. Push
          them to a private container registry only.

          The Co>Operating System includes reference Dockerfiles and example scripts for building
          container images. Typically, you build these images as part of your CI/CD pipeline. For
          more information, contact your Ab Initio administrator or Ab Initio Support.

          ### Ab Initio key servers

          The key servers that you use to enable jobs must be reachable from your cluster. You
          specify these key servers via the `AB_HOSTNAME_KEYSERVER_URLS` configuration variable in
          the pod specifications for launcher and worker agents in the CoOperatingSystemRuntime
          resource.

          To discuss the software activation key requirements for jobs that run in Kubernetes,
          contact your key server administrator or Ab Initio Support.

          ### Namespace and role considerations

          The runtime operator is installed with the cluster-wide RBAC roles and role bindings that
          it needs in order to manage custom resources and run jobs.

          You typically use a separate namespace or namespaces for creating and managing custom
          resources to run jobs. Each of these namespaces must be equipped with a service
          account that can list pod limits and get and patch CoOperatingSystemRuntime resources.

          For each namespace in which you intend to create custom resources, create the needed RBAC
          role and role binding by applying the following Role and RoleBinding manifests:

          ```yaml
          apiVersion: rbac.authorization.k8s.io/v1
          kind: Role
          metadata:
            name: cooperatingsystemruntime-role
          rules:
          - apiGroups:
            - ""
            resources:
            - limitranges
            verbs:
            - list
          - apiGroups:
            - cloud.abinitio.com
            resources:
            - cooperatingsystemruntimes
            verbs:
            - get
            - patch
          ---
          apiVersion: rbac.authorization.k8s.io/v1
          kind: RoleBinding
          metadata:
            name: run-cooperatingsystemruntime-rolebinding
          subjects:
          - kind: ServiceAccount
            name: default
          roleRef:
            kind: Role
            name: cooperatingsystemruntime-role
            apiGroup: rbac.authorization.k8s.io
          ```

          One way to do this is to save these manifests to a file and specify a namespace when
          using `oc` to apply them to the cluster. In the following example, the manifests are
          saved to a file named `runtime-rbac.yaml` and the role and role binding are to be created
          for the `default` service account in the `hello-world` namespace:

          ```
          oc -n hello-world apply -f runtime-rbac.yaml
          ```

          To bind the role to a service account other than `default`, simply update the value of
          `subjects[0].name` in the provided RoleBinding manifest.

          ### Dynamic volume provisioning

          The launcher and worker template sections of the CoOperatingSystemRuntime resource
          require you to specify persistent volume claims (PVCs) for job recovery data. To ensure
          that these PVCs can be bound to approriate storage volumes, dynamic volume provisioning
          must be enabled in your cluster.

          Work with your cluster administrator to identify the storage needs of your jobs and to
          ensure that dynamic volume provisioning is enabled so that PVCs can be matched to
          persistent volumes of the right storage class.

          For more information about dynamic storage provisioning, see the
          [OpenShift documentation](https://docs.openshift.com/container-platform/latest/storage/dynamic-provisioning.html).

          ## Quick start and examples

          For an overview of the key concepts associated with running jobs in Kubernetes, see
          ["About running jobs in Kubernetes"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Configuring_and_running_Co_Operating_System_jobs_in_Kubernetes/About_running_jobs_in_Kubernetes)
          in the Ab Initio Help.

          As described in this topic, you might begin by creating a CoOperatingSystemRuntime
          manifest that defines the configuration and compute resources needed for a standalone test
          job. You could then use the specification for this job to create a CoOperatingSystemRuntimeTemplate
          resource, which would collect common settings and make it possible to concisely specify
          CoOperatingSystemRuntime jobs.

          You could also use a runtime pool (of kind CoOperatingSystemRuntime) to extend these basic
          resources: such a pool consists of an actively running group of templated runtime objects whose
          pods can be used to run jobs on demand. Finally, for reserving resources in a pool, you could
          create a CoOperatingSystemRuntimeClaim resource.

          For more information about these custom resources, refer to the
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Co_Operating_System_runtime_operator_API_reference).

          For a set of integrated examples (including complete YAML manifests) that show you how to
          create and use the custom resources, see ["Example manifests for running jobs"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Configuring_and_running_Co_Operating_System_jobs_in_Kubernetes/Creating_custom_resources_for_running_jobs/Example_manifests_for_running_jobs).
        displayName: Co>Operating System Runtime Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - abinitio
        - runtime
        - operator
        links:
        - name: Ab Initio
          url: https://www.abinitio.com/
        maintainers:
        - email: support@abinitio.com
          name: Ab Initio Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Ab Initio
        relatedImages:
        - registry.connect.redhat.com/abinitio/cooperating-system-runtime-operator@sha256:ee30210a153e8b3ba09080737216d881cc6a07e1a888e9684be0d17fd4ff791d
        version: 4.2.3-0
      entries:
      - name: abinitio-runtime-operator.v4.2.3-0
        version: 4.2.3-0
      - name: abinitio-runtime-operator.v4.2.2
        version: 4.2.2
      name: stable
    defaultChannel: release-4.3
    packageName: abinitio-runtime-operator
    provider:
      name: Ab Initio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: recipe
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: recipe.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "Recipe",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "recipe",
                    "app.kubernetes.io/instance": "recipe-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "recipe",
                    "app.kubernetes.io/part-of": "recipe"
                  },
                  "name": "recipe-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          createdAt: "2025-09-12T06:34:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Recipe is the Schema for the recipes API
            displayName: Recipe
            kind: Recipe
            name: recipes.ramendr.openshift.io
            version: v1alpha1
        description: |
          Recipe describes a workflow used for capturing or recovering Kubernetes resources. This can be referred to by a Ramen DRPlacementControl object when kubeObjectProtection is enabled.
        displayName: Recipe
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - recipe
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/recipe
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:af262b0890300183db1766bc392017b0bda723a9ba841dc25e3d928aa413db9a
        version: 4.16.16-rhodf
      entries:
      - name: recipe.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: recipe.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: recipe.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: recipe.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: recipe.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: recipe.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: recipe.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: recipe.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: recipe.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: recipe.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: recipe.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: recipe.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: recipe.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: recipe.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: recipe.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: recipe.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: recipe.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: recipe
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: New England Research Cloud
      provider-url: https://nerc.mghpcc.org/
    name: keycloak-permissions-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keycloak-permissions-operator.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.nerc.mghpcc.org/v1",
                "kind": "KeycloakAuthorization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keycloak-permissions-operator",
                    "app.kubernetes.io/instance": "keycloakauthorization-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "keycloakauthorization",
                    "app.kubernetes.io/part-of": "keycloak-permissions-operator"
                  },
                  "name": "example"
                },
                "spec": {
                  "cleanupUndefinedAuthorizations": false,
                  "description": "",
                  "keycloak": {
                    "adminSecret": {
                      "name": "keycloak-initial-admin",
                      "passwordKey": "password",
                      "usernameKey": "username"
                    }
                  },
                  "managementState": "Managed",
                  "realm": null,
                  "validateCerts": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          com.redhat.openshift.versions: v4.11
          containerImage: quay.io/nerc-images/keycloak-permissions-operator:1.4.1
          createdAt: "2024-05-07T21:09:02Z"
          description: Managing Keycloak resources, scopes, policies, and permissions
            for fine-grained resource permissions
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/nerc-images/keycloak-permissions-operator
          support: New England Research Cloud
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Keycloak resources, scopes, policies, and permissions
              for fine-grained resource permissions.
            displayName: Keycloak Authorization
            kind: KeycloakAuthorization
            name: keycloakauthorizations.keycloak.nerc.mghpcc.org
            version: v1
        description: "An OpenShift Operator for managing Keycloak resources, scopes,
          policies,\nand permissions for fine-grained resource permissions.\nAccording
          to the [Keycloak Authorization Services documentation](https://www.keycloak.org/docs/23.0.7/authorization_services/):\n\n-
          A resource can be a web page, a RESTFul resource, a file in your file system,
          or other entity you wish to authorize access to.\n- Scopes usually represent
          the actions that can be performed on a resource, but they are not limited
          to that. You can also use scopes to represent one or more attributes within
          a resource.\n- Policies define the conditions that must be satisfied to
          access or perform operations on something (resource or scope), but they
          are not tied to what they are protecting. They are generic and can be reused
          to build permissions or even more complex policies.\n- Permissions are coupled
          with the resource they are protecting. Here you specify what you want to
          protect (resource or scope) and the policies that must be satisfied to grant
          or deny permission.\n\n## Getting started with the Keycloak Permissions
          Operator\n\nStart by defining the name of your KeycloakAuthorization and
          the same namespace where Keycloak is deployed.\n\n```yaml\napiVersion: keycloak.nerc.mghpcc.org/v1\nkind:
          KeycloakAuthorization\nmetadata:\n  name: example\n  namespace: keycloak\n```\n\n##
          KeycloakAuthorization spec\n\nNext, define the spec. \n\n- `validateCerts:
          false` If your Keycloak does not have valid certificates, if you are deploying
          to OpenShift Local for example. You don't need this line if Keycloak has
          valid certificates.\n- `cleanupUndefinedAuthorizations: true` if you would
          like to remove any undefined authorizations that may have been added manually.
          This is useful for keeping in sync with GitOps. If you do not want the operator
          removing any resources, policies, or permissions, then remove this line.\n\n```yaml\nspec:\n
          \ validateCerts: false\n  cleanupUndefinedAuthorizations: true\n```\n\n##
          Keycloak admin credentials\n\nSpecify the details to connect to keycloak,
          including the `baseUrl`, and the Secret name containing the Keycloak admin
          usernameKey and passwordKey.\n\n```yaml\n  keycloak:\n    baseUrl: https://keycloak-service:8443\n
          \   adminSecret:\n      name: keycloak-initial-admin\n      usernameKey:
          username\n      passwordKey: password\n```\n\n## Realm\n\nSpecify the realm
          id you wish to add authorization configuration to. If you want to add authorization
          configuration to multiple realms, then create a KeycloakAuthorization resource
          for each realm.\n\n```yaml\n  realm:\n    id: NERC\n```\n\n## Clients\n\nFor
          each client in the realm you wish to configure, add authorizationSettings.\n\n```yaml\n
          \   clients:\n      - id: ai4cloudops\n        authorizationSettings:\n```\n\n##
          Authorization Scopes\n\nList the scopes you wish to define for the client.\n\n```yaml\n
          \         scopes:\n            - id: prod\n              name: prod\n              displayName:
          Production\n            - id: test\n              name: test\n              displayName:
          Test\n```\n\n## Authorization Resources\n\nList the resources and corresponding
          scopes you wish to define for the client.\n\n```yaml\n          resources:\n
          \           - name: cluster\n              displayName: cluster\n              scopes:\n
          \               - id: prod\n                  name: prod\n                  displayName:
          Production\n                - id: test\n                  name: test\n                  displayName:
          Test\n```\n\n## Authorization Policies\n\nDefine the policies you wish to
          use in your permissions. You can choose from all the different types of
          permissions, including: aggregated, client, client-scope, group, regex,
          role, time, and user.\n\n```yaml\n          policies:\n            - id:
          client-example\n              name: client-example\n              logic:
          POSITIVE\n              decisionStrategy: UNANIMOUS\n              type:\n
          \               client:\n                  clients:\n                    -
          example\n            - id: group-example\n              name: group-example\n
          \             logic: POSITIVE\n              decisionStrategy: UNANIMOUS\n
          \             type:\n                group:\n                  groups:\n
          \                   - id: example\n```\n\n## Authorization Permissions\n\nDefine
          the permissions, connecting the resources, scopes, and policies together.\n\n```yaml\n
          \         permissions:\n            - name: group-example-cluster-test\n
          \             client: test\n              policy: group-example\n              resource:
          cluster\n            - name: client-example-cluster-test\n              client:
          example\n              policy: client-test\n              resource: cluster\n
          \             scope: test\n```\n\n## Other details\n\nResources, scopes,
          policies, and permissions will all be applied to Keycloak if you have configured
          the admin credentials to Keycloak successfully.\nWhen the KeycloakAuthorization
          is deleted, then the configured resources, policies, and permissions will
          be deleted as well. "
        displayName: Keycloak Permissions Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - fine-grained resource permissions
        - keycloak
        - resources
        - scopes
        - policies
        - permissions
        links:
        - name: Keycloak Permissions Operator
          url: https://keycloak-permissions-operator.domain
        maintainers:
        - email: computate@computate.org
          name: Christopher Tate
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: New England Research Cloud
          url: https://nerc.mghpcc.org/
        relatedImages:
        - quay.io/nerc-images/keycloak-permissions-operator:1.4.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 1.4.1
      entries:
      - name: keycloak-permissions-operator.v1.4.1
        version: 1.4.1
      - name: keycloak-permissions-operator.v1.3.0
        version: 1.3.0
      - name: keycloak-permissions-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: keycloak-permissions-operator
    provider:
      name: New England Research Cloud
      url: https://nerc.mghpcc.org/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: service-registry-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-registry-operator.v2.0.7-0.1662549961.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-kafkasql"
                },
                "spec": {
                  "configuration": {
                    "kafkasql": {
                      "bootstrapServers": "\u003cservice name\u003e.\u003cnamespace\u003e.svc:9092"
                    },
                    "persistence": "kafkasql"
                  }
                }
              },
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-sql"
                },
                "spec": {
                  "configuration": {
                    "persistence": "sql",
                    "sql": {
                      "dataSource": {
                        "password": "\u003cpassword\u003e",
                        "url": "jdbc:postgresql://\u003cservice name\u003e.\u003cnamespace\u003e.svc:5432/\u003cdatabase name\u003e",
                        "userName": "postgres"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/integration/service-registry-rhel8-operator@sha256:6dfd539951a739ec06fb6620b79c42c6159d4ee148c0d9d336547e786cd4780c
          createdAt: "2022-06-24"
          description: Deploy and manage Red Hat Integration service registry on OpenShift.
          olm.substitutesFor: service-registry-operator.v2.0.7-0.1661784420.p
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents a Red Hat Integration Service
              Registry instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Service Registry
          Service Registry stores and retrieves API designs and event schemas, and gives you control of their evolution.
          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf - Provides a REST API and web UI to manage the artifacts and artifact meta-data - Includes Serializers and Deserializers for Kafka client integration - Configurable rules to control schema validity and evolution (compatibility) - Storage options: Kafka, PostgreSQL - Compatible with Confluent and IBM APIs - Runs on a lightweight Quarkus platform - Includes Maven plugin to integrate with Maven based builds
          ## Service Registry Operator
          Provides a quick and easy way to deploy and manage service registry on OpenShift.
          **Features**
          - Supports installation and configuration of the Registry for all storage options - Easily perform a rolling upgrade of the Registry
          ## Prerequisites
          This operator does not deploy storage for the Registry. Therefore, some storage options require that the chosen persistence service is already set up.
          ## License
          Red Hat Service Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - Service Registry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - service
        - service-registry
        links:
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/integration/service-registry-rhel8-operator@sha256:6dfd539951a739ec06fb6620b79c42c6159d4ee148c0d9d336547e786cd4780c
        - registry.redhat.io/integration/service-registry-kafkasql-rhel8@sha256:fc97a2a08b2f77d620ea86ec77bb4f750bfc0109ede3ad1080c2cb694019f8b8
        - registry.redhat.io/integration/service-registry-sql-rhel8@sha256:6c97fa8ead1f6f0323811f8f35a80ae33f0e160f6c58f2bdb31713dbeaa3f014
        version: 2.0.7+0.1662549961.p
      entries:
      - name: service-registry-operator.v2.0.7-0.1662549961.p
        version: 2.0.7+0.1662549961.p
      - name: service-registry-operator.v2.0.7-0.1661784420.p
        version: 2.0.7+0.1661784420.p
      - name: service-registry-operator.v2.0.7
        version: 2.0.7
      - name: service-registry-operator.v2.0.6
        version: 2.0.6
      - name: service-registry-operator.v2.0.5-0.1654503709.p
        version: 2.0.5+0.1654503709.p
      - name: service-registry-operator.v2.0.5-0.1652887585.p
        version: 2.0.5+0.1652887585.p
      - name: service-registry-operator.v2.0.5-0.1646309885.p
        version: 2.0.5+0.1646309885.p
      - name: service-registry-operator.v2.0.5-0.1647938417.p
        version: 2.0.5+0.1647938417.p
      - name: service-registry-operator.v2.0.5-0.1648813064.p
        version: 2.0.5+0.1648813064.p
      - name: service-registry-operator.v2.0.5-0.1651047781.p
        version: 2.0.5+0.1651047781.p
      - name: service-registry-operator.v2.0.5-0.1652429755.p
        version: 2.0.5+0.1652429755.p
      - name: service-registry-operator.v2.0.5
        version: 2.0.5
      - name: service-registry-operator.v2.0.5-0.1655396591.p
        version: 2.0.5+0.1655396591.p
      - name: service-registry-operator.v2.0.4
        version: 2.0.4
      - name: service-registry-operator.v2.0.3
        version: 2.0.3
      - name: service-registry-operator.v2.0.2
        version: 2.0.2
      - name: service-registry-operator.v2.0.1
        version: 2.0.1
      - name: service-registry-operator.v2.0.0
        version: 2.0.0
      name: 2.0.x
    - currentCSV: service-registry-operator.v2.6.13-r2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-kafkasql"
                },
                "spec": {
                  "configuration": {
                    "kafkasql": {
                      "bootstrapServers": "\u003cservice name\u003e.\u003cnamespace\u003e.svc:9092"
                    },
                    "persistence": "kafkasql"
                  }
                }
              },
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-sql"
                },
                "spec": {
                  "configuration": {
                    "persistence": "sql",
                    "sql": {
                      "dataSource": {
                        "password": "\u003cpassword\u003e",
                        "url": "jdbc:postgresql://\u003cservice name\u003e.\u003cnamespace\u003e.svc:5432/\u003cdatabase name\u003e",
                        "userName": "postgres"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/apicurio/apicurio-registry-rhel8-operator@sha256:61f00395b51819312dda7abcdaa47190a11a83b1e2b523101be0ba776b1c5b7c
          createdAt: "2025-09-15"
          description: Deploy and manage Red Hat build of Apicurio Registry on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations",
            "Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents a Red Hat build of Apicurio Registry
              instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Red Hat build of Apicurio Registry

          Red Hat build of Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Provides a REST API and web UI to manage the artifacts and artifact meta-data
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL
          - Compatible with Confluent and IBM APIs
          - Runs on a lightweight Quarkus platform
          - Includes Maven plugin to integrate with Maven based builds

          ## Red Hat build of Apicurio Registry Operator

          Provides a quick and easy way to deploy and manage Red Hat build of Apicurio Registry on OpenShift.

          **Features**
          - Supports installation and configuration of the Registry for all storage options
          - Easily perform a rolling upgrade of Red Hat build of Apicurio Registry

          ## Prerequisites

          This operator does not deploy storage for Red Hat build of Apicurio Registry. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Red Hat build of Apicurio Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/main/LICENSE)
        displayName: Red Hat build of Apicurio Registry
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - service
        - service-registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/apicurio/apicurio-registry-rhel8-operator@sha256:61f00395b51819312dda7abcdaa47190a11a83b1e2b523101be0ba776b1c5b7c
        - registry.redhat.io/apicurio/apicurio-registry-kafkasql-rhel8@sha256:f2da7cd92e126187d91043e2ae40521f8574586f8b3975be27d8370fe31b6732
        - registry.redhat.io/apicurio/apicurio-registry-sql-rhel8@sha256:a55425a9af1a16959ab5e528a5740d47ea06f6935593cc925e7727cded4e7d13
        version: 2.6.13-r2
      entries:
      - name: service-registry-operator.v2.6.13-r2
        version: 2.6.13-r2
      - name: service-registry-operator.v2.6.13-r1
        version: 2.6.13-r1
      - name: service-registry-operator.v2.6.12-r1
        version: 2.6.12-r1
      - name: service-registry-operator.v2.6.11-r1
        version: 2.6.11-r1
      - name: service-registry-operator.v2.6.10
        version: 2.6.10
      - name: service-registry-operator.v2.6.9
        version: 2.6.9
      - name: service-registry-operator.v2.6.8-0.1740477005.p
        version: 2.6.8+0.1740477005.p
      - name: service-registry-operator.v2.6.8-0.1739457262.p
        version: 2.6.8+0.1739457262.p
      - name: service-registry-operator.v2.6.8
        version: 2.6.8
      - name: service-registry-operator.v2.6.6-0.1736942038.p
        version: 2.6.6+0.1736942038.p
      - name: service-registry-operator.v2.6.6
        version: 2.6.6
      - name: service-registry-operator.v2.6.5-0.1731946870.p
        version: 2.6.5+0.1731946870.p
      - name: service-registry-operator.v2.6.5
        version: 2.6.5
      - name: service-registry-operator.v2.6.5-0.1729511936.p
        version: 2.6.5+0.1729511936.p
      - name: service-registry-operator.v2.6.5-0.1730982111.p
        version: 2.6.5+0.1730982111.p
      - name: service-registry-operator.v2.6.3
        version: 2.6.3
      - name: service-registry-operator.v2.6.1-0.1724927626.p
        version: 2.6.1+0.1724927626.p
      - name: service-registry-operator.v2.6.1
        version: 2.6.1
      - name: service-registry-operator.v2.3.5
        version: 2.3.5
      - name: service-registry-operator.v2.3.4-0.1716987426.p
        version: 2.3.4+0.1716987426.p
      - name: service-registry-operator.v2.3.4
        version: 2.3.4
      - name: service-registry-operator.v2.3.3
        version: 2.3.3
      - name: service-registry-operator.v2.3.2
        version: 2.3.2
      - name: service-registry-operator.v2.3.1-0.1706181988.p
        version: 2.3.1+0.1706181988.p
      - name: service-registry-operator.v2.3.1
        version: 2.3.1
      - name: service-registry-operator.v2.3.0
        version: 2.3.0
      - name: service-registry-operator.v2.2.3
        version: 2.2.3
      - name: service-registry-operator.v2.2.3-0.1698369839.p
        version: 2.2.3+0.1698369839.p
      - name: service-registry-operator.v2.2.2
        version: 2.2.2
      - name: service-registry-operator.v2.2.2-0.1697116886.p
        version: 2.2.2+0.1697116886.p
      - name: service-registry-operator.v2.2.1
        version: 2.2.1
      - name: service-registry-operator.v2.2.0
        version: 2.2.0
      - name: service-registry-operator.v2.1.5
        version: 2.1.5
      - name: service-registry-operator.v2.1.4
        version: 2.1.4
      - name: service-registry-operator.v2.1.3
        version: 2.1.3
      - name: service-registry-operator.v2.1.2-0.1676389596.p
        version: 2.1.2+0.1676389596.p
      - name: service-registry-operator.v2.1.2
        version: 2.1.2
      - name: service-registry-operator.v2.1.1
        version: 2.1.1
      - name: service-registry-operator.v2.1.0
        version: 2.1.0
      - name: service-registry-operator.v2.0.7
        version: 2.0.7
      - name: service-registry-operator.v2.0.7-0.1662549961.p
        version: 2.0.7+0.1662549961.p
      - name: service-registry-operator.v2.0.7-0.1661784420.p
        version: 2.0.7+0.1661784420.p
      - name: service-registry-operator.v2.0.6
        version: 2.0.6
      - name: service-registry-operator.v2.0.5-0.1648813064.p
        version: 2.0.5+0.1648813064.p
      - name: service-registry-operator.v2.0.5-0.1646309885.p
        version: 2.0.5+0.1646309885.p
      - name: service-registry-operator.v2.0.5-0.1654503709.p
        version: 2.0.5+0.1654503709.p
      - name: service-registry-operator.v2.0.5-0.1652887585.p
        version: 2.0.5+0.1652887585.p
      - name: service-registry-operator.v2.0.5-0.1652429755.p
        version: 2.0.5+0.1652429755.p
      - name: service-registry-operator.v2.0.5-0.1651047781.p
        version: 2.0.5+0.1651047781.p
      - name: service-registry-operator.v2.0.5-0.1655396591.p
        version: 2.0.5+0.1655396591.p
      - name: service-registry-operator.v2.0.5-0.1647938417.p
        version: 2.0.5+0.1647938417.p
      - name: service-registry-operator.v2.0.5
        version: 2.0.5
      - name: service-registry-operator.v2.0.4
        version: 2.0.4
      - name: service-registry-operator.v2.0.3
        version: 2.0.3
      - name: service-registry-operator.v2.0.2
        version: 2.0.2
      - name: service-registry-operator.v2.0.1
        version: 2.0.1
      - name: service-registry-operator.v2.0.0
        version: 2.0.0
      name: 2.x
    defaultChannel: 2.x
    packageName: service-registry-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microsoft
      provider-url: ""
    name: azure-service-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: azure-service-operator.v1.0.59040
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "APIMgmtAPI",
                "metadata": {
                  "name": "aso-apimgmtapi-sample"
                },
                "spec": {
                  "apiId": "apiId0",
                  "apiService": "apimservice-sample",
                  "location": "westus",
                  "properties": {
                    "apiVersionDescription": "API version description",
                    "description": "API description",
                    "displayName": "aso-apimgmtapi-sample",
                    "isCurrent": true,
                    "isOnline": true,
                    "path": "/api/test",
                    "protocols": [
                      "http"
                    ],
                    "subscriptionRequired": false
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ApimService",
                "metadata": {
                  "name": "apimservice-sample"
                },
                "spec": {
                  "appInsightsName": "appinsights-sample-wmm",
                  "appInsightsResourceGroup": "resourcegroup-azure-operators-wmm",
                  "location": "westus",
                  "publisherEmail": {
                    "Your email": ""
                  },
                  "publisherName": {
                    "Your name": ""
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "tier": "premium",
                  "vnetName": "virtualnetwork-sample",
                  "vnetResourceGroup": "resourcegroup-azure-operators",
                  "vnetSubnetName": "test1",
                  "vnetType": "none"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AppInsights",
                "metadata": {
                  "name": "appinsights-sample"
                },
                "spec": {
                  "applicationType": "other",
                  "kind": "other",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AppInsightsApiKey",
                "metadata": {
                  "name": "key-sample"
                },
                "spec": {
                  "appInsights": "appinsights-sample",
                  "authSDKControlChannel": true,
                  "readTelemetry": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "writeAnnotations": true
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureLoadBalancer",
                "metadata": {
                  "name": "lb-sample-01"
                },
                "spec": {
                  "backendAddressPoolName": "test",
                  "backendPort": 3000,
                  "frontendPortRangeEnd": 2000,
                  "frontendPortRangeStart": 1001,
                  "inboundNatPoolName": "test",
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "lb-pip-sample-01",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureNetworkInterface",
                "metadata": {
                  "name": "nic-sample-01"
                },
                "spec": {
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "nic-pip-sample-01",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnetName": "test1",
                  "vnetName": "vnet-sample-01"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureNetworkInterface",
                "metadata": {
                  "name": "nic-sample-02"
                },
                "spec": {
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "nic-pip-sample-02",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnetName": "test1",
                  "vnetName": "vnet-sample-02"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "lb-pip-sample-01"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "nic-pip-sample-01"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "nic-pip-sample-02"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLManagedUser",
                "metadata": {
                  "name": "azuresqlmanageduser-sample"
                },
                "spec": {
                  "dbName": "azuresqldatabase-sample",
                  "managedIdentityClientId": null,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "db_owner"
                  ],
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLUser",
                "metadata": {
                  "name": "sqluser-sample"
                },
                "spec": {
                  "dbName": "azuresqldatabase-sample1",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "db_owner"
                  ],
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLVNetRule",
                "metadata": {
                  "name": "azuresqlvnetrule-sample"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-vnet",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlAction",
                "metadata": {
                  "name": "azuresqlaction-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverName": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "azuresqldatabase-sample1"
                },
                "spec": {
                  "edition": 0,
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "valueA",
                    "tag2": "valueB"
                  },
                  "name": "azuresqldatabase-sample3"
                },
                "spec": {
                  "edition": 0,
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlFailoverGroup",
                "metadata": {
                  "name": "azuresqlfailovergroup-sample"
                },
                "spec": {
                  "databaselist": [
                    "azuresqldatabase-sample1",
                    "azuresqldatabase-sample2"
                  ],
                  "failovergraceperiod": 30,
                  "failoverpolicy": "Automatic",
                  "location": "eastus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "secondaryserver": "sqlserver-samplesec",
                  "secondaryserverresourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-samplepri"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlFirewallRule",
                "metadata": {
                  "name": "sqlf-allowazuresvcaccess"
                },
                "spec": {
                  "endipaddress": "0.0.0.0",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "startipaddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample-777"
                },
                "spec": {
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample777"
                },
                "spec": {
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVMScaleSet",
                "metadata": {
                  "name": "vmss-sample-01"
                },
                "spec": {
                  "adminUserName": "azureuser7",
                  "backendAddressPoolName": "test",
                  "capacity": 3,
                  "inboundNatPoolName": "test",
                  "loadBalancerName": "lb-sample-01",
                  "location": "SouthCentralUS",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": "{ssh public key}",
                  "subnetName": "test1",
                  "virtualNetworkName": "vnet-sample-01",
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachine",
                "metadata": {
                  "name": "hpfvm1"
                },
                "spec": {
                  "adminUserName": "azureuser",
                  "location": "SouthCentralUS",
                  "networkInterfaceName": "nic-sample-02",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": {
                    "ssh key": ""
                  },
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachine",
                "metadata": {
                  "name": "hpfvm20"
                },
                "spec": {
                  "adminUserName": "azureuser",
                  "location": "SouthCentralUS",
                  "networkInterfaceName": "nic-sample-01",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": "{ssh public key}",
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachineExtension",
                "metadata": {
                  "name": "vmext20a"
                },
                "spec": {
                  "autoUpgradeMinorVersion": true,
                  "forceUpdateTag": "Test2",
                  "location": "SouthCentralUS",
                  "protectedSettings": "{}",
                  "publisher": "Microsoft.Azure.Extensions",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "settings": "{\"commandToExecute\":\"echo 'hello, world'\"}",
                  "typeHandlerVersion": "2.1",
                  "typeName": "CustomScript",
                  "vmName": "vm20"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "BlobContainer",
                "metadata": {
                  "name": "blobcontainer-sample"
                },
                "spec": {
                  "accesslevel": "Container",
                  "accountname": "storageaccountsample777",
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ConsumerGroup",
                "metadata": {
                  "name": "consumergroup-sample-1"
                },
                "spec": {
                  "eventHub": "eventhub-sample-1",
                  "namespace": "eventhubnamespace-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "CosmosDB",
                "metadata": {
                  "name": "cosmosdb-sample-1"
                },
                "spec": {
                  "kind": "GlobalDocumentDB",
                  "location": "westus",
                  "properties": {
                    "databaseAccountOfferType": "Standard",
                    "isVirtualNetworkFilterEnabled": true
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "CosmosDB",
                "metadata": {
                  "labels": {
                    "CosmosAccountType": "Production"
                  },
                  "name": "cosmosdb-sample-2"
                },
                "spec": {
                  "ipRules": [
                    "104.42.195.92,40.76.54.131,52.176.6.30,52.169.50.45,52.187.184.26"
                  ],
                  "kind": "GlobalDocumentDB",
                  "location": "eastus",
                  "locations": [
                    {
                      "failoverPriority": 0,
                      "isZoneRedundant": true,
                      "locationName": "eastus"
                    },
                    {
                      "failoverPriority": 1,
                      "locationName": "westus"
                    }
                  ],
                  "properties": {
                    "databaseAccountOfferType": "Standard",
                    "enableMultipleWriteLocations": true
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "Eventhub",
                "metadata": {
                  "name": "eventhub-sample-1"
                },
                "spec": {
                  "authorizationRule": {
                    "name": "RootManageSharedAccessKey",
                    "rights": [
                      "Listen",
                      "Manage",
                      "Send"
                    ]
                  },
                  "location": "westus",
                  "namespace": "eventhubnamespace-sample-1",
                  "properties": {
                    "messageRetentionInDays": 7,
                    "partitionCount": 2
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "Eventhub",
                "metadata": {
                  "name": "eventhub-sample-2307-10-capture"
                },
                "spec": {
                  "authorizationRule": {
                    "name": "RootManageSharedAccessKey",
                    "rights": [
                      "Listen",
                      "Manage",
                      "Send"
                    ]
                  },
                  "location": "westus",
                  "namespace": "eventhubnamespace-sample-1",
                  "properties": {
                    "captureDescription": {
                      "destination": {
                        "archiveNameFormat": "{Namespace}/{EventHub}/{PartitionId}/{Year}/{Month}/{Day}/{Hour}/{Minute}/{Second}",
                        "blobContainer": "capturecontainer",
                        "name": "EventHubArchive.AzureBlockBlob",
                        "storageAccount": {
                          "accountName": "storageaccountauv1",
                          "resourceGroup": "my-resource-group"
                        }
                      },
                      "enabled": true,
                      "intervalInSeconds": 120,
                      "sizeLimitInBytes": 524288000
                    },
                    "messageRetentionInDays": 7,
                    "partitionCount": 2
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "EventhubNamespace",
                "metadata": {
                  "name": "eventhubnamespace-sample-1"
                },
                "spec": {
                  "location": "westus",
                  "properties": {
                    "isAutoInflateEnabled": false,
                    "kafkaEnabled": false,
                    "maximumThroughputUnits": 0
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "capacity": 1,
                    "name": "Standard",
                    "tier": "Standard"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVault",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "keyvaultsample123"
                },
                "spec": {
                  "accessPolicies": [
                    {
                      "objectID": "\u003cobjectID\u003e",
                      "permissions": {
                        "certificates": [
                          "list",
                          "get"
                        ],
                        "keys": [
                          "list",
                          "get"
                        ],
                        "secrets": [
                          "list",
                          "get"
                        ],
                        "storage": [
                          "list",
                          "get"
                        ]
                      },
                      "tenantID": "\u003ctenantID\u003e"
                    }
                  ],
                  "enableSoftDelete": false,
                  "location": "westus",
                  "networkPolicies": {
                    "bypass": "AzureServices",
                    "defaultAction": "Allow",
                    "ipRules": [
                      "172.16.0.0/24",
                      "172.16.1.0/24"
                    ],
                    "virtualNetworkRules": [
                      "/subscriptions/\u003csubid\u003e/resourceGroups/rg1/providers/Microsoft.Network/virtualNetworks/test-vnet/subnets/subnet1",
                      "/subscriptions/\u003csubid\u003e/resourceGroups/rg1/providers/Microsoft.Network/virtualNetworks/test-vnet/subnets/subnet2"
                    ]
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "name": "standard"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVault",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "keyvaultsample123-simple"
                },
                "spec": {
                  "enableSoftDelete": false,
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVaultKey",
                "metadata": {
                  "name": "keyvaultkey-sample"
                },
                "spec": {
                  "keySize": 4096,
                  "keyVault": "keyvaultsample123",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "type": "RSA"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLDatabase",
                "metadata": {
                  "name": "mysqldatabase-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLDatabase",
                "metadata": {
                  "name": "mysqldatabase-sample2"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLFirewallRule",
                "metadata": {
                  "name": "mysqlfirewallrule-sample"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLFirewallRule",
                "metadata": {
                  "name": "mysqlfirewallrule-sample2"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample2",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-replica"
                },
                "spec": {
                  "createMode": "Replica",
                  "location": "eastus2",
                  "replicaProperties": {
                    "sourceServerId": "/subscriptions/{SUBID}/resourceGroups/resourcegroup-azure-operators/providers/Microsoft.DBforMySQL/servers/mysqlserver-sample"
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-sample2"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLUser",
                "metadata": {
                  "name": "mysqluser-sample"
                },
                "spec": {
                  "dbName": "mysqldatabase-sample",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "SELECT"
                  ],
                  "server": "mysqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLVNetRule",
                "metadata": {
                  "name": "mysqlvnetrule-sample"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-vnet",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLDatabase",
                "metadata": {
                  "name": "postgresqldatabase-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLDatabase",
                "metadata": {
                  "name": "postgresqldatabase-sample2"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLFirewallRule",
                "metadata": {
                  "name": "postgresqlfirewallrule-sample"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLFirewallRule",
                "metadata": {
                  "name": "postgresqlfirewallrule-sample2"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample2",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-replica3"
                },
                "spec": {
                  "createMode": "Replica",
                  "location": "eastus",
                  "replicaProperties": {
                    "sourceServerId": "/subscriptions/{subscription ID}/resourceGroups/resourcegroup-azure-operators/providers/Microsoft.DBforPostgreSQL/servers/postgresqlserver-sample"
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-sample2"
                },
                "spec": {
                  "location": "southcentralus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 2,
                    "family": "Gen5",
                    "name": "B_Gen5_2",
                    "size": "51200",
                    "tier": "Basic"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLUser",
                "metadata": {
                  "name": "psqluser-sample"
                },
                "spec": {
                  "dbName": "postgresqldatabase-sample",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "azure_pg_admin"
                  ],
                  "server": "postgresqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLVNetRule",
                "metadata": {
                  "name": "postgresqlvnetrule-sample1"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-azure-operators",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCache",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "rediscache-sample-1"
                },
                "spec": {
                  "location": "westus",
                  "properties": {
                    "enableNonSslPort": true,
                    "sku": {
                      "capacity": 1,
                      "family": "C",
                      "name": "Basic"
                    }
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCacheAction",
                "metadata": {
                  "name": "rediscacheaction-sample-1"
                },
                "spec": {
                  "actionName": "rollallkeys",
                  "cacheName": "rediscache-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCacheFirewallRule",
                "metadata": {
                  "name": "rediscachefirewallrule"
                },
                "spec": {
                  "properties": {
                    "endIP": "0.0.0.0",
                    "startIP": "0.0.0.0"
                  },
                  "redisCache": "rediscache-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ResourceGroup",
                "metadata": {
                  "name": "resourcegroup-azure-operators"
                },
                "spec": {
                  "location": "westus"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "StorageAccount",
                "metadata": {
                  "name": "storageaccountsample777"
                },
                "spec": {
                  "accessTier": "Hot",
                  "kind": "StorageV2",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "name": "Standard_RAGRS"
                  },
                  "supportsHttpsTrafficOnly": true
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "VirtualNetwork",
                "metadata": {
                  "name": "vnet-sample-01"
                },
                "spec": {
                  "addressSpace": "10.0.0.0/8",
                  "location": "SouthCentralUS",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnets": [
                    {
                      "subnetAddressPrefix": "10.1.0.0/16",
                      "subnetName": "test1"
                    },
                    {
                      "subnetAddressPrefix": "10.2.0.0/16",
                      "subnetName": "test2"
                    }
                  ]
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "VirtualNetwork",
                "metadata": {
                  "name": "vnet-sample-02"
                },
                "spec": {
                  "addressSpace": "10.0.0.0/8",
                  "location": "SouthCentralUS",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnets": [
                    {
                      "subnetAddressPrefix": "10.1.0.0/16",
                      "subnetName": "test1"
                    },
                    {
                      "subnetAddressPrefix": "10.2.0.0/16",
                      "subnetName": "test2"
                    }
                  ]
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "BlobContainer",
                "metadata": {
                  "name": "blobcontainer-sample"
                },
                "spec": {
                  "accessLevel": "Container",
                  "accountName": "storageaccountsample777",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "MySQLServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "mysqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "postgresqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "azuresqldatabase-sample"
                },
                "spec": {
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "sku": {
                    "name": "Basic",
                    "tier": "Basic"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlFailoverGroup",
                "metadata": {
                  "name": "azuresqlfailovergroup-sample"
                },
                "spec": {
                  "databaseList": [
                    "azuresqldatabase-sample",
                    "azuresqldatabase-sample2"
                  ],
                  "failoverGracePeriod": 30,
                  "failoverPolicy": "automatic",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "secondaryServer": "sqlserver-samplesec",
                  "secondaryServerResourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-samplepri"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlFirewallRule",
                "metadata": {
                  "name": "sqlf-allowazuresvcaccess"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample-777"
                },
                "spec": {
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Cloud Provider
          certified: "false"
          containerImage: mcr.microsoft.com/k8s/azureserviceoperator:1.0.59040
          createdAt: Mon 12 Jun 2023 04:24:02 UTC
          description: The Azure Service Operator helps you provision Azure resources
            and connect your applications to them from within Kubernetes.
          displayName: Azure Service Operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Azure/azure-service-operator
          support: https://github.com/Azure/azure-service-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Creates an API with the specified properties in the specified
              API Management service.
            displayName: APIMgmtAPI
            kind: APIMgmtAPI
            name: apimgmtapis.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an API Management instance into a specified Resource
              Group at the specified location. It also provides the option to link
              to an Application Insights instance for logging, and to place the API
              Management instance in a specified Virtual Network.
            displayName: ApimService
            kind: ApimService
            name: apimservices.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Application Insights instance into a specified
              Resource Group at the specified location.
            displayName: AppInsights
            kind: AppInsights
            name: appinsights.azure.microsoft.com
            version: v1alpha1
          - description: Creates an Api Key to access a given Application Insights
              instance.
            displayName: AppInsightsApiKey
            kind: AppInsightsApiKey
            name: appinsightsapikeys.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Load Balancer (LB) into a specified Resource
              Group at the specified location.
            displayName: AzureLoadBalancer
            kind: AzureLoadBalancer
            name: azureloadbalancers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Network Interface (NIC) into a specified
              Resource Group at the specified location. Users can specify underlying
              public IP address and virtual network configurations in their NIC setup.
            displayName: AzureNetworkInterface
            kind: AzureNetworkInterface
            name: azurenetworkinterfaces.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Public IP Address (PIP) into a specified
              Resource Group at the specified location. Users can specify IP allocation
              method, idle timeout, IP address version, and SKU.
            displayName: AzurePublicIPAddress
            kind: AzurePublicIPAddress
            name: azurepublicipaddresses.azure.microsoft.com
            version: v1alpha1
          - description: Allows you to roll the password for the specified SQL server.
            displayName: AzureSqlAction
            kind: AzureSqlAction
            name: azuresqlactions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an SQL database given the SQL server.
            displayName: AzureSqlDatabase
            kind: AzureSqlDatabase
            name: azuresqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an SQL database given the SQL server
            displayName: AzureSqlDatabase
            kind: AzureSqlDatabase
            name: azuresqldatabases.azure.microsoft.com
            version: v1beta1
          - description: Deploys a failover group on a specified Azure SQL server
              given the secondary server and the databases to failover.
            displayName: AzureSqlFailoverGroup
            kind: AzureSqlFailoverGroup
            name: azuresqlfailovergroups.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a failover group on a specified Azure SQL server
              given the secondary server and the databases to failover.
            displayName: AzureSqlFailoverGroup
            kind: AzureSqlFailoverGroup
            name: azuresqlfailovergroups.azure.microsoft.com
            version: v1beta1
          - description: Deploys a firewall rule to allow access to the SQL server
              from specific IPs.
            displayName: AzureSqlFirewallRule
            kind: AzureSqlFirewallRule
            name: azuresqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the SQL server
              from specific IPs.
            displayName: AzureSqlFirewallRule
            kind: AzureSqlFirewallRule
            name: azuresqlfirewallrules.azure.microsoft.com
            version: v1beta1
          - description: Binds a specified Managed Identity as a user on the SQL database
            displayName: AzureSQLManagedUser
            kind: AzureSQLManagedUser
            name: azuresqlmanagedusers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure SQL server given the location and Resource
              Group.
            displayName: AzureSqlServer
            kind: AzureSqlServer
            name: azuresqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure SQL server given the location and Resource
              Group.
            displayName: AzureSqlServer
            kind: AzureSqlServer
            name: azuresqlservers.azure.microsoft.com
            version: v1beta1
          - description: Creates an user on the specified Azure SQL database and stores
              the username/password as secrets.
            displayName: AzureSQLUser
            kind: AzureSQLUser
            name: azuresqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a VirtualNetwork rule to allow access to the SQL
              server from specific VNets.
            displayName: AzureSQLVNetRule
            kind: AzureSQLVNetRule
            name: azuresqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys Azure Virtual Machine Extensions to provide post-deployment
              configuration and automation on Azure VMs.
            displayName: AzureVirtualMachineExtension
            kind: AzureVirtualMachineExtension
            name: azurevirtualmachineextensions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Machine (VM) into a specified Resource
              Group at the specified location. Users can specify platform image, size,
              user name and public SSH key, etc. for the VM.
            displayName: AzureVirtualMachine
            kind: AzureVirtualMachine
            name: azurevirtualmachines.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Machine Scale Set (VMSS) into a
              specified Resource Group at the specified location. Users can specify
              platform image, size, user name and public SSH key, etc. for the VMSS.
            displayName: AzureVMScaleSet
            kind: AzureVMScaleSet
            name: azurevmscalesets.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a Blob Container into a specified Resource Group
              at the specified location. A Storage Account must first be created prior
              to creating a Blob Container.
            displayName: BlobContainer
            kind: BlobContainer
            name: blobcontainers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a Blob Container into a specified Resource Group
              at the specified location. A Storage Account must first be created prior
              to creating a Blob Container.
            displayName: BlobContainer
            kind: BlobContainer
            name: blobcontainers.azure.microsoft.com
            version: v1alpha2
          - description: Deploys a consumer group given the Event Hub, Event Hubs
              namespace and Resource Group.
            displayName: ConsumerGroup
            kind: ConsumerGroup
            name: consumergroups.azure.microsoft.com
            version: v1alpha1
          - description: Provision a CosmosDB instance given the Cosmos DB instance
              type, location, and Resource Group.
            displayName: CosmosDB
            kind: CosmosDB
            name: cosmosdbs.azure.microsoft.com
            version: v1alpha1
          - kind: CosmosDBSQLDatabase
            name: cosmosdbsqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Event Hubs namespace given the Resource Group
              and location. Also has the ability to configure SKU, properties, and
              network rules.
            displayName: EventhubNamespace
            kind: EventhubNamespace
            name: eventhubnamespaces.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Event Hubs instance given the Event Hubs namespace,
              Resource Group and Location.
            displayName: Eventhub
            kind: Eventhub
            name: eventhubs.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Key Vault key given the location and Resource
              Group.
            displayName: KeyVaultKey
            kind: KeyVaultKey
            name: keyvaultkeys.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Key Vault given the location and Resource
              Group.
            displayName: KeyVault
            kind: KeyVault
            name: keyvaults.azure.microsoft.com
            version: v1alpha1
          - description: Add an AAD user to an existing MySQL database.
            kind: MySQLAADUser
            name: mysqlaadusers.azure.microsoft.com
            version: v1alpha1
          - description: Add an AAD user to an existing MySQL database.
            kind: MySQLAADUser
            name: mysqlaadusers.azure.microsoft.com
            version: v1alpha2
          - description: Deploys a database under the given Azure Database for MySQL
              server
            displayName: MySQLDatabase
            kind: MySQLDatabase
            name: mysqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the Azure Database
              for MySQL server from the specified IP range.
            displayName: MySQLFirewallRule
            kind: MySQLFirewallRule
            name: mysqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Create an AAD-based admininistrative user in a MySQL server.
            kind: MySQLServerAdministrator
            name: mysqlserveradministrators.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for MySQL server given the Location,
              Resource Group and other properties. Also helps creating read replicas
              for MySQL server.
            displayName: MySQLServer
            kind: MySQLServer
            name: mysqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for MySQL server given the Location,
              Resource Group and other properties. Also helps creating read replicas
              for MySQL server.
            displayName: MySQLServer
            kind: MySQLServer
            name: mysqlservers.azure.microsoft.com
            version: v1alpha2
          - description: Add a new user to an existing MySQL database.
            displayName: MySQLUser
            kind: MySQLUser
            name: mysqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Add a new user to an existing MySQL database.
            kind: MySQLUser
            name: mysqlusers.azure.microsoft.com
            version: v1alpha2
          - description: Add virtual network rules to the MySQL server.
            displayName: MySQLVNetRule
            kind: MySQLVNetRule
            name: mysqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a database under the given Azure Database for PostgreSQL
              server.
            displayName: PostgreSQLDatabase
            kind: PostgreSQLDatabase
            name: postgresqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the Azure Database
              for PostgreSQL server from the specified IP range.
            displayName: PostgreSQLFirewallRule
            kind: PostgreSQLFirewallRule
            name: postgresqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for PostgreSQL server given the
              Location and Resource Group.
            displayName: PostgreSQLServer
            kind: PostgreSQLServer
            name: postgresqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for PostgreSQL server given the
              Location and Resource Group.
            displayName: PostgreSQLServer
            kind: PostgreSQLServer
            name: postgresqlservers.azure.microsoft.com
            version: v1alpha2
          - description: Create and Manage Users for Azure Database for PostgreSQL
              server.
            displayName: PostgreSQLUser
            kind: PostgreSQLUser
            name: postgresqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Create and manage VNet service endpoints and VNet rules in
              Azure Database for PostgreSQL.
            displayName: PostgreSQLVNetRule
            kind: PostgreSQLVNetRule
            name: postgresqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Allows you to regenerate keys and reboot the RedisCache cluster.
            displayName: RedisCacheAction
            kind: RedisCacheAction
            name: rediscacheactions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the RedisCache
              from the specified IP range.
            displayName: RedisCacheFirewallRule
            kind: RedisCacheFirewallRule
            name: rediscachefirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Cache for Redis into a specified Resource
              Group at the specified location.
            displayName: RedisCache
            kind: RedisCache
            name: rediscaches.azure.microsoft.com
            version: v1alpha1
          - description: Provisions a Resource Group given the location and a name.
            displayName: ResourceGroup
            kind: ResourceGroup
            name: resourcegroups.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Storage Account into a specified Resource
              Group at the specified location.
            displayName: StorageAccount
            kind: StorageAccount
            name: storageaccounts.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Network into a specified Resource
              Group at the specified location, in the address space provided. Users
              are also able to add subnets to their virtual network through the operator.
            displayName: VirtualNetwork
            kind: VirtualNetwork
            name: virtualnetworks.azure.microsoft.com
            version: v1alpha1
        description: |
          ## Overview

          The Azure Service Operator comprises of:

          - The Custom Resource Definitions (CRDs) for each of the Azure services a Kubernetes user can provision.
          - The Kubernetes controller that watches for requests to create Custom Resources for each of these CRDs and creates them.

          The project was built using [Kubebuilder](https://book.kubebuilder.io/).

          Curious to see how it all works? Check out our [control flow diagrams](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/controlflow.md).

          **Note**: The current version of Azure Service Operator only supports watching all namespaces in the cluster.

          ## Supported Azure Services

          - [Resource Group](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/resourcegroup/resourcegroup.md)
          - [Event Hubs](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/eventhub/eventhub.md)
          - [Azure SQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/azuresql/azuresql.md)
          - [Azure Database for PostgreSQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/postgresql/postgresql.md)
          - [Azure Database for MySQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/mysql/mysql.md)
          - [Azure Key Vault](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/keyvault/keyvault.md)
          - [Azure Cache for Redis](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/rediscache/rediscache.md)
          - [Storage Account](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/storage/storageaccount.md)
          - [Blob Storage](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/storage/blobcontainer.md)
          - [Virtual Network](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/virtualnetwork/virtualnetwork.md)
          - [Application Insights](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/appinsights/appinsights.md)
          - [API Management](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/apimgmt/apimgmt.md)
          - [Cosmos DB](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/cosmosdb/cosmosdb.md)
          - [Virtual Machine](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/virtualmachine/virtualmachine.md)
          - [Virtual Machine Scale Set](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/vmscaleset/vmscaleset.md)

          ## Instructions - Please read before installing

          Before you begin, verify you're running the Azure CLI version 2.0.53 or later. Run `az --version` to find the version. If you need to install or upgrade, see [Install Azure CLI](https://docs.microsoft.com/en-us/cli/azure/install-azure-cli).

          1. Create an Azure Service Principal. You'll need this to grant Azure Service Operator permissions to create resources in your subscription.

              First, set the following environment variables to your Azure Tenant ID and Subscription ID with your values:
              ```yaml
              AZURE_TENANT_ID=<your-tenant-id-goes-here>
              AZURE_SUBSCRIPTION_ID=<your-subscription-id-goes-here>
              ```

              You can find these values by using the [Azure CLI](https://docs.microsoft.com/en-us/cli/azure/?view=azure-cli-latest):
              ```sh
              az account show
              ```

          2. Next, we'll create a Service Principal with Contributor permissions for your subscription, so ASO can create resources in your subscription on your behalf. Note that the [ServicePrincipal](https://docs.microsoft.com/en-us/cli/azure/create-an-azure-service-principal-azure-cli) you pass to the command below needs to have access to create resources in your subscription. If you'd like to use Managed Identity for authorization instead, check out instructions [here](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/managedidentity.md).

              ```sh
              az ad sp create-for-rbac -n "azure-service-operator" --role contributor \
                  --scopes /subscriptions/$AZURE_SUBSCRIPTION_ID
              ```

              This should give you output like the following:
              ```sh
              "appId": "xxxxxxxxxx",
              "displayName": "azure-service-operator",
              "name": "http://azure-service-operator",
              "password": "xxxxxxxxxxx",
              "tenant": "xxxxxxxxxxxxx"
              ```

          3.  Once you have created a Service Principal, gather the following values:

              `AZURE_TENANT_ID` is the Tenant ID from Step 1.

              `AZURE_SUBSCRIPTION_ID` is the Subscription ID from Step 1.

              `AZURE_CLIENT_ID` is the appID from the Service Principal created in Step 2.

              `AZURE_CLIENT_SECRET` is the password from the service Principal we created in Step 2.

              `AZURE_CLOUD_ENV` is the Azure Environment you'd like to use, i.e. AzurePublicCloud, AzureUSGovernmentCloud, AzureChinaCloud, AzureGermanCloud.

          4.  Set those values in a `Secret` called `azureoperatorsettings` within the `operators` Namespace:
              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name: azureoperatorsettings
                namespace: operators
              stringData:
                AZURE_TENANT_ID: <your-tenant-id-goes-here>
                AZURE_SUBSCRIPTION_ID: <your-subscription-id-goes-here>
                AZURE_CLIENT_ID: <your-client-id>
                AZURE_CLIENT_SECRET: <your-client-secret>
                AZURE_CLOUD_ENV: <your-azure-cloud-environment>
              ```

          5. Now you can proceed to install the Azure Service Operator to the `operators` Namespace via the Install button on the top right of this page. After the operator is installed, you will then see the Azure Service Operator pod running in your cluster:

              ```console
              $ kubectl get pods -n operators
              NAME                                                READY   STATUS    RESTARTS   AGE
              azureoperator-controller-manager-7dd75bbd97-mk4s9   2/2     Running   0          35s
              ```

          ## About the project

          This project maintains [releases of the Azure Service Operator](https://github.com/Azure/azure-service-operator/releases).

          Please see the [FAQ](https://github.com/Azure/azure-service-operator/blob/main/docs/faq.md) for answers to commonly asked questions about the Azure Service Operator.

          Have more questions? Feel free to consult our documentation [here](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/contents.md).

          ## Contributing

          The [contribution guide](https://github.com/Azure/azure-service-operator/blob/main/CONTRIBUTING.md) covers everything you need to know about how you can contribute to Azure Service Operators. The [developer guide](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/contents.md#developing-azure-service-operator) will help you onboard as a developer.

          ## Support

          Azure Service Operator is an open source project that is **not** covered by the [Microsoft Azure support policy](https://support.microsoft.com/en-us/help/2941892/support-for-linux-and-open-source-technology-in-azure). Please search open issues [here](https://github.com/Azure/azure-service-operator/issues). If your issue isn't already represented, please [open a new one](https://github.com/Azure/azure-service-operator/issues/new/choose). The Azure Service Operator project maintainers will respond to the best of their abilities.

          ## Code of conduct

          This project has adopted the [Microsoft Open Source Code of Conduct](https://opensource.microsoft.com/codeofconduct/). For more information, see the [Code of Conduct FAQ](https://opensource.microsoft.com/codeofconduct/faq) or contact [opencode@microsoft.com](mailto:opencode@microsoft.com) with any additional questions or comments.
        displayName: Azure Service Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - azure, microsoft, cloud service broker
        links:
        - name: Azure Service Operator
          url: https://github.com/Azure/azure-service-operator
        maintainers:
        - email: christian.muirhead@microsoft.com
          name: Christian Muirhead
        maturity: stable
        provider:
          name: Microsoft
        relatedImages:
        - mcr.microsoft.com/k8s/azureserviceoperator:1.0.59040
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 1.0.59040
      entries:
      - name: azure-service-operator.v1.0.59040
        version: 1.0.59040
      name: stable
    defaultChannel: stable
    packageName: azure-service-operator
    provider:
      name: Microsoft
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/kuadrant-operator
    name: kuadrant-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kuadrant-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "authScheme": {
                    "identity": [
                      {
                        "apiKey": {
                          "allNamespaces": true,
                          "selector": {
                            "matchLabels": {
                              "app": "toystore"
                            }
                          }
                        },
                        "credentials": {
                          "in": "authorization_header",
                          "keySelector": "APIKEY"
                        },
                        "name": "friends"
                      }
                    ],
                    "response": [
                      {
                        "json": {
                          "properties": [
                            {
                              "name": "userID",
                              "valueFrom": {
                                "authJSON": "auth.identity.metadata.annotations.secret\\.kuadrant\\.io/user-id"
                              }
                            }
                          ]
                        },
                        "name": "rate-limit-apikey",
                        "wrapper": "envoyDynamicMetadata",
                        "wrapperKey": "ext_auth_data"
                      }
                    ]
                  },
                  "rules": null,
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "rateLimits": [
                    {
                      "configurations": [
                        {
                          "actions": [
                            {
                              "metadata": {
                                "default_value": "no-user",
                                "descriptor_key": "userID",
                                "metadata_key": {
                                  "key": "envoy.filters.http.ext_authz",
                                  "path": [
                                    {
                                      "segment": {
                                        "key": "ext_auth_data"
                                      }
                                    },
                                    {
                                      "segment": {
                                        "key": "userID"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          ]
                        }
                      ],
                      "limits": [
                        {
                          "conditions": [],
                          "maxValue": 2,
                          "seconds": 10,
                          "variables": [
                            "userID"
                          ]
                        }
                      ]
                    }
                  ],
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/kuadrant-operator:v0.3.1
          createdAt: "2023-06-01T12:24:17Z"
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Enable AuthN and AuthZ based access control on workloads
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1beta1
          - description: Kuadrant is the Schema for the kuadrants API
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: Enable access control on workloads based on HTTP rate limiting
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1beta1
        description: A Kubernetes Operator to manage the lifecycle of the Kuadrant
          system
        displayName: Kuadrant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: Kuadrant Operator
          url: https://github.com/Kuadrant/kuadrant-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - quay.io/kuadrant/kuadrant-operator:v0.3.1
        - oci://quay.io/kuadrant/wasm-shim:v0.2.0
        version: 0.3.1
      entries:
      - name: kuadrant-operator.v0.3.1
        version: 0.3.1
      - name: kuadrant-operator.v0.3.0
        version: 0.3.0
      - name: kuadrant-operator.v0.2.1
        version: 0.2.1
      - name: kuadrant-operator.v0.2.0
        version: 0.2.0
      name: alpha
    - currentCSV: kuadrant-operator.v0.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSPolicy",
                "metadata": {
                  "name": "dnspolicy-sample"
                },
                "spec": {
                  "healthCheck": {
                    "endpoint": "/",
                    "protocol": "HTTP"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "TLSPolicy",
                "metadata": {
                  "name": "tlspolicy-sample"
                },
                "spec": {
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "self-signed-ca"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "kuadrant.io/v1beta2",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "rules": {
                    "authentication": {
                      "apikey": {
                        "apiKey": {
                          "selector": {}
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta2",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "limits": {
                    "toys": {
                      "rates": [
                        {
                          "duration": 1,
                          "limit": 50,
                          "unit": "minute"
                        }
                      ]
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/kuadrant-operator:v0.11.0
          createdAt: "2024-10-02T20:30:45Z"
          description: '[DEPRECATED] Follow https://docs.kuadrant.io instead'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthPolicy enables authentication and authorization for service
              workloads in a Gateway API network
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1beta2
          - description: DNSPolicy configures how North-South based traffic should
              be balanced and reach the gateways
            displayName: DNSPolicy
            kind: DNSPolicy
            name: dnspolicies.kuadrant.io
            version: v1alpha1
          - description: Kuadrant configures installations of Kuadrant Service Protection
              components
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: RateLimitPolicy enables rate limiting for service workloads
              in a Gateway API network
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1beta2
          - description: TLSPolicy provides tls for gateway listeners by managing
              the lifecycle of tls certificates
            displayName: TLSPolicy
            kind: TLSPolicy
            name: tlspolicies.kuadrant.io
            version: v1alpha1
        description: |-
          ## Deprecated
          This Kuadrant community operator is deprecated, Follow https://docs.kuadrant.io instead.
        displayName: '[DEPRECATED] Kuadrant Operator'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: Kuadrant Operator
          url: https://github.com/Kuadrant/kuadrant-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - quay.io/kuadrant/kuadrant-operator:v0.11.0
        - oci://quay.io/kuadrant/wasm-shim:v0.6.0
        version: 0.11.1
      entries:
      - name: kuadrant-operator.v0.11.1
        version: 0.11.1
      - name: kuadrant-operator.v0.11.0
        version: 0.11.0
      - name: kuadrant-operator.v0.10.0
        version: 0.10.0
      - name: kuadrant-operator.v0.7.2
        version: 0.7.2
      - name: kuadrant-operator.v0.4.1
        version: 0.4.1
      - name: kuadrant-operator.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: kuadrant-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/kuadrant-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Nutanix
      provider-url: ""
    name: nutanixcsioperator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nutanixcsioperator.v2.6.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crd.nutanix.com/v1alpha1",
                "kind": "NutanixCsiStorage",
                "metadata": {
                  "name": "nutanixcsistorage"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:859b6d1b2b36a48391e347c2a7431cafa08decafb56f801e21e544c3460a4b4b
          createdAt: "2024-10-30T17:10:05Z"
          description: An Operator for deploying and managing the Nutanix CSI Driver
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2.6.11
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi","disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/nutanix/csi-operator
          support: Nutanix Support Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a new instance of Nutanix CSI Driver
            displayName: Nutanix CSI Storage
            kind: NutanixCsiStorage
            name: nutanixcsistorages.crd.nutanix.com
            version: v1alpha1
        description: "# Overview \n\nThe Nutanix CSI Operator for Kubernetes packages,
          deploys, manages, upgrades the Nutanix CSI Driver on Kubernetes and OpenShift
          for dynamic provisioning of persistent volumes on Nutanix Enterprise Cloud
          platform.\n\nThe Nutanix Container Storage Interface (CSI) Driver for Kubernetes
          leverages Nutanix Volumes and Nutanix Files to provide scalable and persistent
          storage for stateful applications.\n\nWith the Nutanix CSI Provider you
          can:\n\n*  **Provide persistent storage to your containers**\n    * Leverage
          PVC resources to consume dynamically Nutanix storage\n    * With Files storage
          classes, applications on multiple pods can access the same storage, and
          also have the benefit of multi-pod read and write access. \n\n## Warning\n\nIf
          you upgrade from a previous version take care to validate your existing
          StorageClass parameters against the corresponding [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v2_6:CSI-Volume-Driver-v2_6).\n\n##
          Supported Access mode\n\n| Storage Class mode | ReadWriteOnce | ReadOnlyMany
          | ReadWriteMany |\n|--------------------|---------------|--------------|---------------|\n|
          Nutanix Volumes    | yes           | yes          | no            |\n| Nutanix
          Files      | yes           | yes          | yes           |\n\n\n## Supported
          orchestration platforms\nThe following table details orchestration platforms
          suitable for deployment of the Nutanix CSI driver.\n\n| Orchestration platform
          | Version | Architecture |\n|------------------------|---------|--------------|\n|
          Red Hat OpenShift      | 4.12    | x86          |\n| Red Hat OpenShift      |
          4.13    | x86          |\n| Red Hat OpenShift      | 4.14    | x86          |\n|
          Red Hat OpenShift      | 4.15    | x86          |\n| Red Hat OpenShift      |
          4.16    | x86          |\n| Red Hat OpenShift      | 4.17    | x86          |\n\n##
          Configuring k8s secret and storage class\nIn order to use this driver, deploy
          your NutanixCsiStorage instance and create the relevant storage classes
          and secrets, by following the [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v2_6:CSI-Volume-Driver-v2_6)
          or the example in below section:\n\n**1. Create a NutanixCsiStorage resource
          to deploy your driver**\n\nYou can do it directly inside the Operator UI
          with the `Create instance` button or with the following ressource\n```yaml\napiVersion:
          crd.nutanix.com/v1alpha1\nkind: NutanixCsiStorage\nmetadata:\n  name: nutanixcsistorage\n
          \ namespace: openshift-cluster-csi-drivers\nspec: {}\n```\n\n**2. Create
          a Nutanix secret**\n```yaml\napiVersion: v1\nkind: Secret\nmetadata:\n  name:
          ntnx-secret\n  namespace: openshift-cluster-csi-drivers\nstringData:\n  #
          prism-ip:prism-port:username:password.\n  key: 1.2.3.4:9440:admin:password\n```\n\n**3.
          Create storage classes**\n\n- volume mode example\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n  name: nutanix-volume\nprovisioner: csi.nutanix.com\nparameters:\n
          \ csi.storage.k8s.io/provisioner-secret-name: ntnx-secret\n  csi.storage.k8s.io/provisioner-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/node-publish-secret-name:
          ntnx-secret\n  csi.storage.k8s.io/node-publish-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/controller-expand-secret-name: ntnx-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/fstype: ext4\n  storageContainer:
          default-container\n  storageType: NutanixVolumes\n  #description: \"description
          added to each storage object created by the driver\"\n  #isSegmentedIscsiNetwork:
          \"false\"\n  #whitelistIPMode: ENABLED\n  #chapAuth: ENABLED\n  #isLVMVolume:
          \"false\"\n  #numLVMDisks: 4\nallowVolumeExpansion: true\nreclaimPolicy:
          Delete\n```\n\n- dynamic files mode example\n\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n    name: nutanix-dynfiles\nprovisioner: csi.nutanix.com\nparameters:\n
          \ dynamicProv: ENABLED\n  nfsServerName: fs\n  csi.storage.k8s.io/provisioner-secret-name:
          ntnx-secret\n  csi.storage.k8s.io/provisioner-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/node-publish-secret-name: ntnx-secret\n  csi.storage.k8s.io/node-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-expand-secret-name:
          ntnx-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace: openshift-cluster-csi-drivers\n
          \ storageType: NutanixFiles\n  squashType: \"none\"\n  #description: \"description
          added to each storage object created by the driver\"\nallowVolumeExpansion:
          true\n```"
        displayName: Nutanix CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Nutanix
        - CSI
        - Volumes
        - Files
        - Storage
        links:
        - name: Documentation
          url: https://portal.nutanix.com/page/documents/list?type=software&filterKey=product&filterVal=CSI
        maintainers:
        - email: cloudnative@nutanix.com
          name: Nutanix Cloud Native
        maturity: beta
        provider:
          name: Nutanix
        relatedImages:
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:04c55b93a032ad8075cae540d60fc740a72c31006590dbdb61ad4ab9813959d0
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:2cddcc716c1930775228d56b0d2d339358647629701047edfdad5fcdfaf4ebcb
        - registry.connect.redhat.com/nutanix/ntnx-csi-os@sha256:92962320ad682a7113aeadc8f9de1c7849996e06819acf00925aa8aa3bf51068
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-resizer@sha256:e998f22243869416f9860fc6a1fb07d4202eac8846defc1b85ebd015c1207605
        - registry.k8s.io/sig-storage/livenessprobe@sha256:82adbebdf5d5a1f40f246aef8ddbee7f89dea190652aefe83336008e69f9a89f
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:9af9bf28430b00a0cedeb2ec29acadce45e6afcecd8bdf31c793c624cfa75fa7
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:76a47d100de1f2afa23803bec503d9440605db03c7179b231f963e5663a3d12a
        - registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:859b6d1b2b36a48391e347c2a7431cafa08decafb56f801e21e544c3460a4b4b
        version: 2.6.11
      entries:
      - name: nutanixcsioperator.v2.6.11
        version: 2.6.11
      name: stable
    - currentCSV: nutanixcsioperator.v3.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crd.nutanix.com/v1alpha1",
                "kind": "NutanixCsiStorage",
                "metadata": {
                  "name": "nutanixcsistorage"
                },
                "spec": {
                  "ntnxInitConfigMap": {
                    "usePC": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:9cc0ffc6e9c0a82320e207dad6b6b3ddc30ab8f83fa9e099a34450d5537ca435
          createdAt: "2025-06-18T11:04:03Z"
          description: An Operator for deploying and managing the Nutanix CSI Driver
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.3.4
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi","disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/nutanix/csi-operator
          support: Nutanix Support Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a new instance of Nutanix CSI Driver
            displayName: Nutanix CSI Storage
            kind: NutanixCsiStorage
            name: nutanixcsistorages.crd.nutanix.com
            version: v1alpha1
        description: "# Important\nPlease read the following instructions carefully
          before upgrading from 2.6.x to 3.3.4, for more information please refer
          to [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)
          \n- Please do not upgrade to the CSI 3.x operator if:\n    * You are using
          LVM volumes.\n    \n- To upgrade from the CSI 2.6.x to CSI 3.3 (interacting
          with Prism Central) operator\n    * Create a Nutanix Prism Central secret
          as explained below.\n    * Delete the csidriver object from the cluster:\n\n
          \       ```\n        oc delete csidriver csi.nutanix.com\n        ```\n\n
          \   * In the installed operators, go to Nutanix CSI Operator and change
          the subscription channel from stable to stable-3.x. \n    If you have installed
          the operator with automatic update approval, the operator will be automatically
          upgraded to CSI 3.3, and then the nutanixcsistorage resource will be upgraded.\n
          \   An update plan will be generated for manual updates. Upon approval,
          the operator will be successfully upgraded.\n    \n- Direct upgrades from
          CSI 2.6.x to CSI 3.3 interacting with Prism Element are not supported. \n
          \   The only solution is to recreate the nutanixcsistorage instance by following
          the below procedure:\n    - In the installed operators, go to Nutanix CSI
          Operator and delete the nutanixcsistorage instance.\n    - Next change the
          subscription channel from stable to stable-3.x.\n    - Verify the following
          points:\n        - Ensure a Nutanix Prism Element secret is present in the
          namespace.\n        - Ensure that all the storage classes with provisioner:
          csi.nutanix.com have a controller publish secret as explained below.\n\n
          \           ```\n            csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n            csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n            ```\n\n            If this secret
          is not present in the storage class please delete and recreate the storage
          classes with the required secrets.\n    - Create a new instance of nutanixcsistorage
          from this operator by specifying `usePC: false` in YAML spec section.\n
          \   - Caution: Moving from CSI driver interacting with Prism Central to
          CSI driver interacting with Prism Element is not supported.\n    \n- Troubleshooting:\n
          \   \n    If the upgrade was unsuccessful and you want to revert to version
          CSI 2.6.x, please delete the csidriver object as explained above, uninstall
          the operator (no need to delete the nutanixcsistorage custom resource),
          and install version CSI 2.6.x from the stable channel.\n# Overview \nThe
          Nutanix CSI Operator for Kubernetes packages, deploys, manages, and upgrades
          the Nutanix CSI Driver on Kubernetes and OpenShift for dynamic provisioning
          of persistent volumes on the Nutanix Enterprise Cloud platform.\nThe Nutanix
          Container Storage Interface (CSI) Driver for Kubernetes leverages Nutanix
          Volumes and Nutanix Files to provide scalable and persistent storage for
          stateful applications.\nWith the Nutanix CSI Provider, you can:\n*  **Provide
          persistent storage to your containers**\n    * Leverage PVC resources to
          consume dynamically Nutanix storage\n    * With Files storage classes, applications
          on multiple pods can access the same storage and have the benefit of multi-pod
          read and write access. \n## Warning\nIf you upgrade from a previous version,
          make sure to validate your existing StorageClass parameters against the
          corresponding [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)\n##
          Supported Access mode\n| Storage Class mode | ReadWriteOnce | ReadOnlyMany
          | ReadWriteMany |\n|--------------------|---------------|--------------|---------------|\n|
          Nutanix Volumes    | yes           | yes          | no            |\n| Nutanix
          Files      | yes           | yes          | yes           |\n## Supported
          orchestration platforms\nThe following table details orchestration platforms
          suitable for deployment of the Nutanix CSI driver.\n| Orchestration platform
          | Version | Architecture |\n|------------------------|---------|--------------|\n|
          Red Hat OpenShift      | 4.12    | x86          |\n| Red Hat OpenShift      |
          4.13    | x86          |\n| Red Hat OpenShift      | 4.14    | x86          |\n|
          Red Hat OpenShift      | 4.15    | x86          |\n| Red Hat OpenShift      |
          4.16    | x86          |\n| Red Hat OpenShift      | 4.17    | x86          |\n|
          Red Hat OpenShift      | 4.18    | x86          |\n| Red Hat OpenShift      |
          4.19    | x86          |\n## Configuring k8s secret and storage class\nTo
          use this driver, create the relevant storage classes and secrets and deploy
          your NutanixCsiStorage instance by following the [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)
          or the example in below section:\n**1. Create a Nutanix PE secret**\n```yaml\napiVersion:
          v1\nkind: Secret\nmetadata:\n  name: ntnx-pe-secret\n  namespace: openshift-cluster-csi-drivers\nstringData:\n
          \ # prism-ip:prism-port:username:password.\n  key: 1.2.3.4:9440:admin:password\n
          \ files-key: \"fileserver01.sample.com:csi:password1\"\n```\n**2. Create
          a Nutanix PC secret**\n```yaml\napiVersion: v1\nkind: Secret\nmetadata:\n
          \ name: ntnx-pc-secret\n  namespace: openshift-cluster-csi-drivers\nstringData:\n
          \ # prism-ip:prism-port:username:password.\n  key: 1.2.3.4:9440:admin:password\n```\n**3.
          Create a NutanixCsiStorage resource to deploy your driver**\nYou can do
          it directly inside the Operator UI with the `Create instance` button or
          with the following resource\n```yaml\napiVersion: crd.nutanix.com/v1alpha1\nkind:
          NutanixCsiStorage\nmetadata:\n  name: nutanixcsistorage\n  namespace: openshift-cluster-csi-drivers\nspec:
          \n  ntnxInitConfigMap:\n    usePC : true (false for PE based installations.
          Default is true.)\n```\n**4. Create storage classes**\n- volume mode example
          for 3.x installation\n```yaml\nkind: StorageClass\napiVersion: storage.k8s.io/v1\nmetadata:\n
          \ name: nutanix-volume\nprovisioner: csi.nutanix.com\nparameters:\n  csi.storage.k8s.io/provisioner-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/node-publish-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-expand-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/fstype: ext4\n  storageContainer:
          default-container\n  storageType: NutanixVolumes\n  #description: \"description
          added to each storage object created by the driver\"\n  #hypervisorAttached:
          ENABLED\nallowVolumeExpansion: true\nreclaimPolicy: Delete\n```\n- volume
          mode example for 2.x installation\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n  name: nutanix-volume\nprovisioner: csi.nutanix.com\nparameters:\n
          \ csi.storage.k8s.io/provisioner-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/node-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/controller-expand-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/fstype: ext4\n   storageContainer:
          default-container\n   storageType: NutanixVolumes\n   #description: \"description
          added to each storage object created by the driver\"\n   #isSegmentedIscsiNetwork:
          \"false\"\n   #whitelistIPMode: ENABLED\n   #chapAuth: ENABLED\n   #isLVMVolume:
          \"false\"\n   #numLVMDisks: 4\n allowVolumeExpansion: true\n reclaimPolicy:
          Delete\n```\n- dynamic files mode example\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n    name: nutanix-dynfiles\nprovisioner: csi.nutanix.com\nparameters:\n
          \ dynamicProv: ENABLED\n  nfsServerName: fs\n  csi.storage.k8s.io/provisioner-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/node-publish-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-expand-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  storageType: NutanixFiles\n  squashType:
          \"none\"\n  #description: \"description added to each storage object created
          by the driver\"\nallowVolumeExpansion: true\n```"
        displayName: Nutanix CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Nutanix
        - CSI
        - Volumes
        - Files
        - Storage
        links:
        - name: Documentation
          url: https://portal.nutanix.com/page/documents/list?type=software&filterKey=product&filterVal=CSI
        maintainers:
        - email: cloudnative@nutanix.com
          name: Nutanix Cloud Native
        maturity: beta
        provider:
          name: Nutanix
        relatedImages:
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0f64602ea791246712b51df334bbd701a0f31df9950a4cb9c28c059f367baa9e
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:f1bd6ee18c4021c1c94f29edfab89b49b6a4d1b800936c19dbef2d75f8202f2d
        - registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:9cc0ffc6e9c0a82320e207dad6b6b3ddc30ab8f83fa9e099a34450d5537ca435
        - registry.connect.redhat.com/nutanix/ntnx-csi-os@sha256:f0e191e7bf98829dbf718978fc971444436a1ecb9957ee22a5d8f9ba01a931b7
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:04c55b93a032ad8075cae540d60fc740a72c31006590dbdb61ad4ab9813959d0
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:9e581022a4e063b9d8b4f9466a36eabe4d0ced3badc8e353f13b700074ce4553
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:9af9bf28430b00a0cedeb2ec29acadce45e6afcecd8bdf31c793c624cfa75fa7
        - registry.k8s.io/sig-storage/csi-attacher@sha256:d7325367ab72b2d469a5091d87b4fc01142d2d13d1a28b2defbbe3e6fdbc4611
        - registry.connect.redhat.com/nutanix/ntnx-csi-os@sha256:9f121929227acf5581a813dc2f952105f517647c11a39cd7ea0863e768082586
        - registry.k8s.io/sig-storage/livenessprobe@sha256:82adbebdf5d5a1f40f246aef8ddbee7f89dea190652aefe83336008e69f9a89f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:3c116f543f0590aeff3299c8bb0683f250817d11a77d9e9071b15a0bffdabcd9
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:10624570c0aceb03f55f1eb07147b0c537e4676869cca2e9bd4bab113f810ac4
        version: 3.3.4
      entries:
      - name: nutanixcsioperator.v3.3.4
        version: 3.3.4
      name: stable-3.x
    defaultChannel: stable-3.x
    packageName: nutanixcsioperator
    provider:
      name: Nutanix
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/confidential-containers
    name: trustee-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trustee-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          containerImage: registry.redhat.io/build-of-trustee/trustee-rhel9-operator@sha256:fe8e23de14ad088c79b1bd5e7aab08a1b88a86679effcbf174be81e761bd9d6d
          createdAt: "2024-06-07T10:06:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trustee-operator-system
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/trustee-operator
          support: Confidential Containers Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KbsConfig is the Schema for the kbsconfigs API
            displayName: Kbs Config
            kind: KbsConfig
            name: kbsconfigs.confidentialcontainers.org
            version: v1alpha1
        description: Remote attestation operator for confidential containers
        displayName: Red Hat build of Trustee
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trustee
        - Confidential Computing
        - Attestation
        links:
        - name: Red Hat build of Trustee
          url: https://github.com/openshift/trustee-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
          url: https://github.com/confidential-containers
        relatedImages:
        - registry.redhat.io/build-of-trustee/trustee-rhel9-operator@sha256:fe8e23de14ad088c79b1bd5e7aab08a1b88a86679effcbf174be81e761bd9d6d
        version: 0.4.2
      entries:
      - name: trustee-operator.v0.4.2
        version: 0.4.2
      - name: trustee-operator.v0.4.1
        version: 0.4.1
      - name: trustee-operator.v0.4.0
        version: 0.4.0
      - name: trustee-operator.v0.3.0
        version: 0.3.0
      - name: trustee-operator.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: trustee-operator
    provider:
      name: Red Hat
      url: https://github.com/confidential-containers
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-keyspaces-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-keyspaces-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keyspaces.services.k8s.aws/v1alpha1",
                "kind": "Keyspace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "keyspaces.services.k8s.aws/v1alpha1",
                "kind": "Table",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/keyspaces-controller:1.1.1
          createdAt: "2025-09-20T01:23:26Z"
          description: AWS Keyspaces controller is a service controller for managing
            Keyspaces resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Keyspace represents the state of an AWS keyspaces Keyspace
              resource.
            displayName: Keyspace
            kind: Keyspace
            name: keyspaces.keyspaces.services.k8s.aws
            version: v1alpha1
          - description: Table represents the state of an AWS keyspaces Table resource.
            displayName: Table
            kind: Table
            name: tables.keyspaces.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Keyspaces resources in AWS from within your Kubernetes cluster.

          **About Amazon Keyspaces**

          Amazon Keyspaces (for Apache Cassandra) is a scalable, highly available, and managed Apache Cassandra–compatible database service. With Amazon Keyspaces, you don’t have to provision, patch, or manage servers, and you don’t have to install, maintain, or operate software.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Keyspaces
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keyspaces
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Keyspaces Developer Resources
          url: https://aws.amazon.com/keyspaces/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: keyspaces maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/keyspaces-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-keyspaces-controller.v1.1.1
        version: 1.1.1
      - name: ack-keyspaces-controller.v1.1.0
        version: 1.1.0
      - name: ack-keyspaces-controller.v1.0.14
        version: 1.0.14
      - name: ack-keyspaces-controller.v1.0.12
        version: 1.0.12
      - name: ack-keyspaces-controller.v1.0.11
        version: 1.0.11
      - name: ack-keyspaces-controller.v1.0.10
        version: 1.0.10
      - name: ack-keyspaces-controller.v1.0.9
        version: 1.0.9
      - name: ack-keyspaces-controller.v1.0.7
        version: 1.0.7
      - name: ack-keyspaces-controller.v1.0.6
        version: 1.0.6
      - name: ack-keyspaces-controller.v1.0.5
        version: 1.0.5
      - name: ack-keyspaces-controller.v1.0.3
        version: 1.0.3
      - name: ack-keyspaces-controller.v1.0.2
        version: 1.0.2
      - name: ack-keyspaces-controller.v1.0.1
        version: 1.0.1
      - name: ack-keyspaces-controller.v1.0.0
        version: 1.0.0
      - name: ack-keyspaces-controller.v0.0.16
        version: 0.0.16
      - name: ack-keyspaces-controller.v0.0.15
        version: 0.0.15
      - name: ack-keyspaces-controller.v0.0.14
        version: 0.0.14
      - name: ack-keyspaces-controller.v0.0.13
        version: 0.0.13
      - name: ack-keyspaces-controller.v0.0.12
        version: 0.0.12
      - name: ack-keyspaces-controller.v0.0.11
        version: 0.0.11
      - name: ack-keyspaces-controller.v0.0.10
        version: 0.0.10
      - name: ack-keyspaces-controller.v0.0.9
        version: 0.0.9
      - name: ack-keyspaces-controller.v0.0.8
        version: 0.0.8
      - name: ack-keyspaces-controller.v0.0.7
        version: 0.0.7
      - name: ack-keyspaces-controller.v0.0.6
        version: 0.0.6
      - name: ack-keyspaces-controller.v0.0.5
        version: 0.0.5
      - name: ack-keyspaces-controller.v0.0.4
        version: 0.0.4
      - name: ack-keyspaces-controller.v0.0.3
        version: 0.0.3
      - name: ack-keyspaces-controller.v0.0.2
        version: 0.0.2
      - name: ack-keyspaces-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-keyspaces-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Carlos Eduardo de Paula
      provider-url: ""
    name: lbconfig-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lbconfig-operator.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-infra-sample",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "http",
                    "path": "/healthz",
                    "port": 1936
                  },
                  "ports": [
                    80,
                    443
                  ],
                  "provider": {
                    "creds": "f5-creds",
                    "host": "https://192.168.1.35",
                    "partition": "Common",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "F5_BigIP"
                  },
                  "type": "infra",
                  "vip": "192.168.1.45"
                }
              },
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-infra-sample-shard",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "http",
                    "path": "/healthz",
                    "port": 1936
                  },
                  "nodelabels": {
                    "kubernetes.io/region": "DC1",
                    "node-role.kubernetes.io/infra": ""
                  },
                  "ports": [
                    80,
                    443
                  ],
                  "provider": {
                    "creds": "f5-creds",
                    "host": "https://192.168.1.35",
                    "partition": "Common",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "F5_BigIP"
                  },
                  "type": "infra",
                  "vip": "192.168.1.46"
                }
              },
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-master-sample",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "https",
                    "path": "/healthz",
                    "port": 6443
                  },
                  "ports": [
                    6443
                  ],
                  "provider": {
                    "creds": "netscaler-creds",
                    "host": "https://192.168.1.36",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "Citrix_ADC"
                  },
                  "type": "master",
                  "vip": "192.168.1.40"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Networking
          certified: "false"
          containerImage: quay.io/carlosedp/lbconfig-operator:v0.5.1
          createdAt: "2025-01-09T15:45:28Z"
          description: Manage External Load Balancers allowing creation/update for
            VIPs and Servers dynamically via API.
          k8sMaxVersion: ""
          k8sMinVersion: ""
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/carlosedp/lbconfig-operator
          support: https://github.com/carlosedp/lbconfig-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalLoadBalancer is the Schema for the externalloadbalancers
              API
            displayName: ExternalLoadBalancer Instance
            kind: ExternalLoadBalancer
            name: externalloadbalancers.lb.lbconfig.carlosedp.com
            version: v1
          required:
          - description: ExternalLoadBalancer represents a configured instance of
              an external Load-Balancer for a specific group of nodes of the cluster.
              The Instance has a VIP and ports to be balanced to the cluster nodes
              based on a set of node labels.
            displayName: External Load-Balancer Configuration Instance
            kind: ExternalLoadBalancer
            name: externalloadbalancers.lb.lbconfig.carlosedp.com
            version: v1
        description: |
          ## About the Operator

          The LBConfig Operator, manages the configuration of External Load Balancer instances (on third-party equipment via it's API) and creates VIPs and IP Pools with Monitors for a set of OpenShift or Kubernetes nodes like Master-nodes (Control-Plane), Infra nodes (where the Routers or Ingress controllers are located) or based on it's roles and/or labels.

          The operator dynamically handles creating, updating or deleting the IPs of the pools in the Load Balancer based on the Node IPs for each role or label. On every change of the operator configuration (CRDs) or addition/change/removal or cluster Nodes, the operator updates the Load Balancer properly.

          ## Main Use Cases

          The main users for this operator is enterprise deployments or clusters composed of multiple nodes having an external load-balancer providing the balancing and high-availability to access the cluster in both API and Application levels.

          ### Create ExternalLoadBalancer Instances

          Create the instances for each Load Balancer instance you need (for example one for Master Nodes and another for the Infra Nodes). **If installing on OpenShift or Kubernetes with OLM (or in a different namespace), adjust the sample YAMLs to match the created namespace**.

          **The provider `vendor` field can be (case-insensitive):**

          * **`F5_BigIP`** - Tested on F5 BigIP version 15
          * **`Citrix_ADC`** - Tested on Citrix ADC (Netscaler) version 13
          * **`HAProxy`** - HAProxy with Dataplane API. ([Docs](./docs/haproxy/))
          * **`Dummy`** - Dummy backend used for testing to only print log messages on operations

          Create the secret holding the Load Balancer API user and password:

          ```sh
          oc create secret generic f5-creds --from-literal=username=admin --from-literal=password=admin123 --namespace lbconfig-operator-system
          ```

          #### Sample CRDs and Available Fields

          Master Nodes using a Citrix ADC LB:

          ```yaml
          apiVersion: lb.lbconfig.carlosedp.com/v1
          kind: ExternalLoadBalancer
          metadata:
            name: externalloadbalancer-master-sample
            namespace: lbconfig-operator-system
          spec:
            vip: "192.168.1.40"
            type: "master"
            ports:
              - 6443
            monitor:
              path: "/healthz"
              port: 6443
              monitortype: "https"
            provider:
              vendor: Citrix_ADC
              host: "https://192.168.1.36"
              port: 443
              creds: netscaler-creds
              validatecerts: false
          ```

          Infra Nodes using a F5 BigIP LB:

          ```yaml
          apiVersion: lb.lbconfig.carlosedp.com/v1
          kind: ExternalLoadBalancer
          metadata:
            name: externalloadbalancer-infra-sample
            namespace: lbconfig-operator-system
          spec:
            vip: "192.168.1.45"
            type: "infra"
            ports:
              - 80
              - 443
            monitor:
              path: "/healthz"
              port: 1936
              monitortype: http
            provider:
              vendor: F5_BigIP
              host: "https://192.168.1.35"
              port: 443
              creds: f5-creds
              partition: "Common"
              validatecerts: false
          ```

          To choose the nodes which will be part of the server pool, you can set either `type` or `nodelabels` fields. The yaml field `type: "master"` or `type: "infra"` selects nodes with the role label `"node-role.kubernetes.io/master"` and `"node-role.kubernetes.io/infra"` respectively. If the field `nodelabels` array is used instead, the operator will use nodes which match all labels.

          Clusters with sharded routers or using arbitrary labels to determine where the Ingress Controllers run can be configured like:

          ```yaml
          spec:
            vip: "10.0.0.6"
            ports:
              - 80
            nodelabels:
              "node.kubernetes.io/ingress-controller": "production"
              "kubernetes.io/region": "DC1"
            ...
          ```

          ## Getting Help

          This project fully open-source and is hosted at [GitHub](https://github.com/carlosedp/lbconfig-operator). Always check the project page and issues for the latest updates and/or solutions to problems encountered. If a problem is detected, please fill an issue. Feature request and improvement issues are welcome.

          Some fields inside `providers` are optional and depend on the used backend. Check the [API docs](https://pkg.go.dev/github.com/carlosedp/lbconfig-operator/api/lb.lbconfig.carlosedp.com/v1?utm_source=gopls#Provider) which fields are backend-specific.

          ## Disclaimers

          * The operator does not check if the requested configuration (names, IPs) already exists and/or conflicts with existing configuration in the Load Balancer. The user is responsible for these checks before deployment;
          * I am not responsible if the operator changes/deletes existing configuration on the Load Balancer if existing names are already configured.
          * The operator creates the entries(Pools, VIPs, Monitors) in the provided Load Balancer with the `name` of the instance configured in the CustomResource prefixed with the type. Eg. For a CR with name `externalloadbalancer-master-sample`, the operator creates a server pool named `Pool-externalloadbalancer-master-sample-6443` (suffixed with the port), a monitor named `Monitor-externalloadbalancer-master-sample` and a VIP named `VIP-externalloadbalancer-master-sample-6443` (suffixed with the port).
        displayName: External Load-Balancer Configuration Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 'load-balance '
        - infrastructure
        links:
        - name: GitHub
          url: https://github.com/carlosedp/lbconfig-operator
        - name: Documentation
          url: https://github.com/carlosedp/lbconfig-operator/docs
        - name: Issues
          url: https://github.com/carlosedp/lbconfig-operator/issues
        maintainers:
        - email: carlosedp@gmail.com
          name: Carlos Eduardo de Paula
        maturity: beta
        minKubeVersion: 1.18.0
        provider:
          name: Carlos Eduardo de Paula
        relatedImages:
        - quay.io/carlosedp/kube-rbac-proxy:v0.15.0
        - quay.io/carlosedp/lbconfig-operator:v0.5.1
        version: 0.5.1
      entries:
      - name: lbconfig-operator.v0.5.1
        version: 0.5.1
      - name: lbconfig-operator.v0.5.0
        version: 0.5.0
      - name: lbconfig-operator.v0.4.1
        version: 0.4.1
      - name: lbconfig-operator.v0.4.0
        version: 0.4.0
      - name: lbconfig-operator.v0.2.0
        version: 0.2.0
      name: beta
    defaultChannel: beta
    packageName: lbconfig-operator
    provider:
      name: Carlos Eduardo de Paula
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: group-sync-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: group-sync-operator.v0.0.36
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "azure-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "azure": {
                        "credentialsSecret": {
                          "name": "azure-group-sync",
                          "namespace": "group-sync-operator"
                        }
                      },
                      "name": "azure"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "github-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "github": {
                        "credentialsSecret": {
                          "name": "github-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "organization": "redhat-cop"
                      },
                      "name": "github"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "gitlab-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "gitlab": {
                        "credentialsSecret": {
                          "name": "gitlab-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "url": "https://gitlab.apps.openshift.com"
                      },
                      "name": "gitlab"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "keycloak-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "keycloak": {
                        "credentialsSecret": {
                          "name": "keycloak-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "insecure": true,
                        "realm": "ocp",
                        "url": "https://keycloak.apps.openshift.com"
                      },
                      "name": "keycloak"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "ldap-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "ldap": {
                        "credentialsSecret": {
                          "name": "ldap-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "insecure": true,
                        "rfc2307": {
                          "groupMembershipAttributes": [
                            "member"
                          ],
                          "groupNameAttributes": [
                            "cn"
                          ],
                          "groupUIDAttribute": "dn",
                          "groupsQuery": {
                            "baseDN": "ou=Groups,dc=example,dc=com",
                            "derefAliases": "never",
                            "filter": "(objectClass=groupofnames)",
                            "scope": "sub"
                          },
                          "tolerateMemberNotFoundErrors": true,
                          "tolerateMemberOutOfScopeErrors": true,
                          "userNameAttributes": [
                            "cn"
                          ],
                          "userUIDAttribute": "dn",
                          "usersQuery": {
                            "baseDN": "ou=Users,dc=example,dc=com",
                            "derefAliases": "never",
                            "scope": "sub"
                          }
                        },
                        "url": "ldap://localhost:389"
                      },
                      "name": "ldap"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "okta-sync"
                },
                "spec": {
                  "providers": [
                    {
                      "name": "okta",
                      "okta": {
                        "appId": "okta-sync-app-id",
                        "credentialsSecret": {
                          "name": "okta-api-token",
                          "namespace": "group-sync-operator"
                        },
                        "extractLoginUsername": true,
                        "profileKey": "login",
                        "url": "https://example.okta.com/"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/group-sync-operator@sha256:43ef79efc2a77ee5edd4e3123f9dab156a30d3e00fd1fba9f89600b96f0f68d6
          createdAt: "2025-08-12T19:19:21Z"
          description: Synchronize groups and users from external providers
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: group-sync-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/group-sync-operator
          support: Red Hat Community of Practice
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GroupSync is the Schema for the groupsyncs API
            displayName: Group Sync
            kind: GroupSync
            name: groupsyncs.redhatcop.redhat.io
            version: v1alpha1
        description: |
          Synchronizes groups from external providers into OpenShift

          ## Overview

          The OpenShift Container Platform contains functionality to synchronize groups found in external identity providers into the platform. Currently, the functionality that is included in OpenShift to limited to synchronizing LDAP only. This operator is designed to integrate with external providers in order to provide new solutions.

          Group Synchronization is facilitated by creating a `GroupSync` resource. The following describes the high level schema for this resource:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: example-groupsync
          spec:
            providers:
              - <One or more providers to synchronize>
          ```

          ## Authentication

          In most cases, authentication details must be provided in order to communicate with providers. Authentication details are provider specific with regards to the required values. In supported providers, the secret can be referenced in the `credentialsSecret` by name and namespace where it has been created as shown below:

          ```
          credentialsSecret:
            name: <secret_name>
            namespace: <secret_namespace>
          ```

          ## Providers

          Integration with external systems is made possible through a set of pluggable external providers. The following providers are currently supported:

          * [Azure](https://azure.microsoft.com/)
          * [GitHub](https://github.com)
          * [GitLab](https://gitlab.com)
          * [LDAP](https://en.wikipedia.org/wiki/Lightweight_Directory_Access_Protocol)
          * [Keycloak](https://www.keycloak.org/)/[Red Hat Single Sign On](https://access.redhat.com/products/red-hat-single-sign-on)
          * [Okta](https://www.okta.com/)
          * [IBM Security Verify](https://docs.verify.ibm.com/verify)

          The following sections describe the configuration options available for each provider


          ### Azure

          Groups contained within Azure Active Directory can be synchronized into OpenShift. The following table describes the set of configuration options for the Azure provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `authorityHost` | Azure Active Directory Endpoint | `https://login.microsoftonline.com` | No |
          | `baseGroups` | List of groups to start searching from instead of listing all groups in the directory | | No |
          | `credentialsSecret` | Name of the secret containing authentication details (See below) | | Yes |
          | `filter` | Graph API filter | | No |
          | `groups` | List of groups to filter against | | No |
          | `userNameAttributes` | Fields on a user record to use as the User Name | `userPrincipalName` | No |
          | `prune` | Prune Whether to prune groups that are no longer in Azure | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Azure provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: azure-groupsync
          spec:
            providers:
            - name: azure
              azure:
                credentialsSecret:
                  name: azure-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to Azure

          Authentication to Azure can be performed using Application Registration with access to query group information in Azure Active Directory.

          The App Registration must be granted access to the following Microsoft Graph API's:

          * Group.Read.All
          * GroupMember.Read.All
          * User.Read.All

          A secret must be created in the same namespace that contains the `GroupSync` resource:

          The following keys must be defined in the secret

          * `AZURE_TENANT_ID` - Tenant ID
          * `AZURE_CLIENT_ID` - Client ID
          * `AZURE_CLIENT_SECRET` - Client Secret

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic azure-group-sync --from-literal=AZURE_SUBSCRIPTION_ID=<AZURE_SUBSCRIPTION_ID> --from-literal=AZURE_TENANT_ID=<AZURE_TENANT_ID> --from-literal=AZURE_CLIENT_ID=<AZURE_CLIENT_ID> --from-literal=AZURE_CLIENT_SECRET=<AZURE_CLIENT_SECRET>
          ```

          ### GitHub

          Teams stored within a GitHub organization can be synchronized into OpenShift. The following table describes the set of configuration options for the GitHub provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `caSecret` | Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `organization` | Organization to synchronize against | | Yes |
          | `teams` | List of teams to filter against | | No |
          | `url` | Base URL for the GitHub or GitHub Enterprise host (Must contain a trailing slash) | | No |
          | `prune` | Prune Whether to prune groups that are no longer in GitHub | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Github provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: github-groupsync
          spec:
            providers:
            - name: github
              github:
                organization: ocp
                credentialsSecret:
                  name: github-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to GitHub

          Authentication to GitLab can be performed using a [Token](https://docs.gitlab.com/ee/security/token_overview.html) or a Username and Password (Note: 2FA not supported). A secret must be created in the same namespace that contains the `GroupSync` resource:

          When using an OAuth token, the following token types are supported:

          * Personal Access Token
          * OAuth Token
          * Job Token

          the following key is required:

          * `token` - OAuth token

          Optionally, the `tokenType` key can be specified to indicate the type of token being provided from the following values:

          * OAuth - `oauth`
          * Personal Access Token - `personal`
          * Job Token - `job`

          If no `tokenType` is provided, `oauth` is used by default

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token>
          ```

          To specify a token type, such as a Personal Access Token, the following command can be executed:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token> --from-literal=tokenType=personal
          ```

          The following keys are required for username and password:

          * `username` - Username for authenticating with GitHub
          * `password` - Password for authenticating with GitHub

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic github-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### GitLab

          Groups stored within a GitLab can be synchronized into OpenShift. The following table describes the set of configuration options for the GitLab provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `groups` | List of groups to filter against | | No |
          | `prune` | Prune Whether to prune groups that are no longer in GitLab | `false` | No |
          | `scope` | Scope for group synchronization. Options are `one` for one level or `sub` to include subgroups | `sub` | No |
          | `url` | Base URL for the GitLab instance | `https://gitlab.com` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Github provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: gitlab-groupsync
          spec:
            providers:
            - name: gitlab
              gitlab:
                credentialsSecret:
                  name: gitlab-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to GitLab

          Authentication to GitLab can be performed using an OAuth Personal Access Token or a Username and Password (Note: 2FA not supported). A secret must be created in the same namespace that contains the `GroupSync` resource:

          When using an OAuth token, the following key is required:

          * `token` - OAuth token

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token>
          ```

          The following keys are required for username and password:

          * `username` - Username for authenticating with GitLab
          * `password` - Password for authenticating with GitLab

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### LDAP

          Groups stored within an [LDAP](https://en.wikipedia.org/wiki/Lightweight_Directory_Access_Protocol) server can be synchronized into OpenShift. The LDAP provider implements the included features of the [Syncing LDAP groups](https://docs.openshift.com/container-platform/latest/authentication/ldap-syncing.html) feature and makes use of the libraries from the [OpenShift Command Line](https://docs.openshift.com/container-platform/latest/cli_reference/openshift_cli/getting-started-cli.html) tool to streamline the migration to this operator based implementation.

          The configurations of the three primary schemas (`rfc2307`, `activeDirectory` and `augmentedActiveDirectory`) can be directly migrated as is without any modification.

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | No |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `groupUIDNameMapping` | User defined name mapping | | No |
          | `rfc2307` | Configuration using the [rfc2307](https://docs.openshift.com/container-platform/latest/authentication/ldap-syncing.html#ldap-syncing-rfc2307_ldap-syncing-groups) schema | | No |
          | `activeDirectory` | Configuration using the [activeDirectory](https://docs.openshift.com/container-platform/4.5/authentication/ldap-syncing.html#ldap-syncing-activedir_ldap-syncing-groups) schema | | No |
          | `augmentedActiveDirectory` | Configuration using the [activeDirectory](https://docs.openshift.com/container-platform/4.5/authentication/ldap-syncing.html#ldap-syncing-augmented-activedir_ldap-syncing-groups) schema | | No |
          | `url` | Connection URL for the LDAP server | `ldap://ldapserver:389` | No |
          | `whitelist` | Explicit list of groups to synchronize |  | No |
          | `blacklist` | Explicit list of groups to not synchronize |  | No |
          | `prune` | Prune Whether to prune groups that are no longer in LDAP | `false` | No |

          The following is an example using the `rfc2307` schema:

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
                credentialsSecret:
                  name: ldap-group-sync
                  namespace: group-sync-operator
                insecure: true
                rfc2307:
                  groupMembershipAttributes:
                  - member
                  groupNameAttributes:
                  - cn
                  groupUIDAttribute: dn
                  groupsQuery:
                    baseDN: ou=Groups,dc=example,dc=com
                    derefAliases: never
                    filter: (objectClass=groupofnames)
                    scope: sub
                  tolerateMemberNotFoundErrors: true
                  tolerateMemberOutOfScopeErrors: true
                  userNameAttributes:
                  - cn
                  userUIDAttribute: dn
                  usersQuery:
                    baseDN: ou=Users,dc=example,dc=com
                    derefAliases: never
                    scope: sub
                url: ldap://ldapserver:389
              name: ldap
          ```

          The examples provided in the OpenShift documented referenced previously can be used to construct the schemas for the other LDAP synchronization types.

          #### Authenticating to LDAP

          If authentication is required in order to communicate with the LDAP server, a secret should be created in the same namespace that contains the `GroupSync` resource. The following keys can be defined:

          * `username` - Username (Bind DN) for authenticating with the LDAP server
          * `password` - Password for authenticating with the LDAP server

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic ldap-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          #### Whitelists and Blacklists

          Groups can be explicitly whitelisted or blacklisted in order to control the groups that are eligible to be synchronized into OpenShift. When running LDAP group synchronization using the command line, this configuration is referenced via separate files, but these are instead specified in the `blacklist` and `whitelist` properties as shown below:

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
          ...
                whitelist:
                - cn=Online Corporate Banking,ou=Groups,dc=example,dc=com
          ...
              name: ldap
          ```

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
          ...
                blacklist:
                - cn=Finance,ou=Groups,dc=example,dc=com
          ...
              name: ldap
          ```

          ### Keycloak

          Groups stored within Keycloak can be synchronized into OpenShift. The following table describes the set of configuration options for the Keycloak provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `groups` | List of groups to filter against | | No |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `loginRealm` | Realm to authenticate against | `master` | No |
          | `realm` | Realm to synchronize | | Yes |
          | `scope` | Scope for group synchronization. Options are `one` for one level or `sub` to include subgroups | `sub` | No |
          | `url` | Base URL for the Keycloak server. Older versions (<17.0.0) including Red Hat SSO should include the context path `/auth` appended to the hostname  | | Yes |
          | `prune` | Prune Whether to prune groups that are no longer in Keycloak | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Keycloak provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                url: https://keycloak-keycloak-operator.apps.openshift.com/auth
          ```

          #### Authenticating to Keycloak

          A user with rights to query for Keycloak groups must be available. The following permissions must be associated to the user:

          * Password must be set (Temporary option unselected) on the _Credentials_ tab
          * On the _Role Mappings_ tab, select _master-realm_ or _realm-management_ next to the _Client Roles_ dropdown and then select **Query Groups** and **Query Users**.

          A secret must be created in the same namespace that contains the `GroupSync` resource. It must contain the following keys for the user previously created:

          * `username` - Username for authenticating with Keycloak
          * `password` - Password for authenticating with Keycloak

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic keycloak-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### Okta

          [Okta Groups](https://help.okta.com/en/prod/Content/Topics/users-groups-profiles/usgp-main.htm) assigned to [Okta Applications](https://help.okta.com/en/prod/Content/Topics/Apps/Apps_Apps.htm) can be synchronized into OpenShift. The developer docs for the Okta API that the Okta Syncer uses can be found [here](https://developer.okta.com/docs/reference/api/apps/#list-groups-assigned-to-application).
          The following table describes the set of configuration options for the Okta provider:

            | Name | Description | Defaults | Required |
            | ----- | ---------- | -------- | ----- |
            | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
            | `groups` | List of groups to filter against | | No |
            | `url` | Okta organization URL (Must contain a trailing slash) | | Yes |
            | `appId` | Application ID of App Groups are assigned to | | Yes |
            | `extractLoginUsername` | Bool to determine if you should extract username from okta login | | No |
            | `profileKey` | Attribute field on Okta User Profile you would like to use as identity | `login` | No |
            | `groupLimit` | Integer to set the maximum number of groups to retrieve from OKTA per request. | `1000` | No |
            | `prune` | Prune Whether to prune groups that are no longer in OKTA | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with an Okta provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: okta-sync
          spec:
            providers:
              - name: okta
                okta:
                  credentialsSecret:
                    name: okta-api-token
                    namespace: group-sync-operator
                  url: "https://example.okta.com/"
                  appId: okta-sync-app-id
          ```

          #### Authenticating to Okta

          A secret must be created in the same namespace as the group-sync-operator pod. It must contain the following key:

            * `okta-api-token` - Okta API Token for interacting with Okta

          The secret can be created by executing the following command:

            ```shell
            oc create secret generic okta-api-token --from-literal=okta-api-token=<OKTA_API_TOKEN> -n group-sync-operator
            ```

          ### IBM Security Verify

          Groups defined in [IBM Security Verify](https://help.okta.com/en/prod/Content/Topics/users-groups-profiles/usgp-main.htm) (ISV) can be synchronized into OpenShift. Currently only the `userName` field from ISV will be synchronized. The developer docs for the ISV API can be found [here](https://docs.verify.ibm.com/verify/page/api-documentation).
          The following table describes the set of configuration options for the provider:

            | Name | Description | Defaults | Required |
            | ----- | ---------- | -------- | ----- |
            | `credentialsSecret` | Reference to a secret containing authentication details (see below) | `''`  | Yes |
            | `groups` | List of groups to synchronize (see below) | `nil`  | Yes |
            | `tenantUrl` | The ISV tenant URL, for example `https://my-isv.verify.ibm.com`) | `''`  | Yes |

          The following is an example of a minimal configuration that can be applied to integrate with an Okta provider:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ibmsecurityverify-sync
          spec:
              providers:
              - name: ibmsecurityverify
                ibmsecurityverify:
                  credentialsSecret:
                    name: isv-group-sync
                    namespace: group-sync-operator
                  tenantUrl: https://my-isv.verify.ibm.com
                  groups:
                    - name: 'application owners'
                      id: 645001V3V9
                    - name: developer
                      id: 645001V3VA
          ```

          #### Group Objects
          Each group object in the `groups` array must contain an `id` field. The group ID can be retrieved by pulling the group information from the ISV API. Optionally, the object may also contain a `name` which corresponds to the group's display name. When defined, the operator will confirm that the name defined in the YAML matches that received from the API when synchronization occurs; as the group IDs are not human-friendly, using the name can confirm the correct groups are configured. If the names do not match an error will be logged.

          #### Group Names
          The name of each groups created in OpenShift will match the group name in ISV. Any whitespace in the ISV group name will be replaced with a hyphen.

          #### Authenticating to IBM Security Verify

          A secret must be created in the same namespace as the group-sync-operator pod. It must contain the following keys:

          * `clientId` - The API client ID.
          * `clientSecret`- The API client secret.

          See the IBM Security Verify [API documentation](https://docs.verify.ibm.com/verify/docs/api-access) for setting up authentication.

          ### Support for Additional Metadata (Beta)

          Additional metadata based on Keycloak group are also added to the OpenShift groups as Annotations including:

          * Parent/child relationship between groups and their subgroups
          * Group attributes

          ## CA Certificates

          Several providers allow for certificates to be provided in either a _ConfigMap_ or _Secret_ to communicate securely to the target host through the use of a property called `ca`.

          The certificate can be added to a Secret called _keycloak-certs_ using the key `ca.crt` representing the certificate using the following command.

          ```
          oc create secret generic keycloak-certs --from-file=ca.crt=<file>
          ```

          An example of how the CA certificate can be added to the Keycloak provider is shown below:


          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                caSecret:
                  kind: Secret
                  name: keycloak-certs
                  namespace: group-sync-operator
                  key: ca.crt
                url: https://keycloak-keycloak-operator.apps.openshift.com
          ```

          Alteratively, a _ConfigMap_ can be used instead instead of a _Secret_. This is useful when using the [Certificate injection using Operators](https://docs.openshift.com/container-platform/latest/networking/configuring-a-custom-pki.html#certificate-injection-using-operators_configuring-a-custom-pki) feature.

          The following command can be used to create a _ConfigMap_ containing the certificate:

          ```
          oc create configmap keycloak-certs --from-file=ca.crt=<file>
          ````

          An example of how the CA certificate can be added to the Keycloak provider is shown below:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                caSecret:
                  kind: ConfigMap
                  name: keycloak-certs
                  namespace: group-sync-operator
                  key: ca.crt
                url: https://keycloak-keycloak-operator.apps.openshift.com
          ```

          ## Scheduled Execution

          A cron style expression can be specified for which a synchronization event will occur. The following specifies that a synchronization should occur nightly at 3AM

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            schedule: "0 3 * * *"
            providers:
            - ...
          ```

          If a schedule is not provided, synchronization will occur only when the object is reconciled by the platform.

          ## Metrics

          Prometheus compatible metrics are exposed by the Operator and can be integrated into OpenShift's default cluster monitoring. To enable OpenShift cluster monitoring, label the namespace the operator is deployed in with the label `openshift.io/cluster-monitoring="true"`.

          ```
          oc label namespace <namespace> openshift.io/cluster-monitoring="true"
          ```
        displayName: Group Sync Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - group
        - user
        - security
        - authorization
        links:
        - name: Group Sync Operator
          url: https://github.com/redhat-cop/group-sync-operator
        maintainers:
        - email: ablock@redhat.com
          name: Andrew Block
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:b8a174be91f3df75dff9656d4c72de28af51a731008e9e96db5eef3c7af58aac
        - quay.io/redhat-cop/group-sync-operator@sha256:43ef79efc2a77ee5edd4e3123f9dab156a30d3e00fd1fba9f89600b96f0f68d6
        version: 0.0.36
      entries:
      - name: group-sync-operator.v0.0.36
        version: 0.0.36
      - name: group-sync-operator.v0.0.35
        version: 0.0.35
      - name: group-sync-operator.v0.0.34
        version: 0.0.34
      - name: group-sync-operator.v0.0.33
        version: 0.0.33
      - name: group-sync-operator.v0.0.32
        version: 0.0.32
      - name: group-sync-operator.v0.0.31
        version: 0.0.31
      - name: group-sync-operator.v0.0.30
        version: 0.0.30
      - name: group-sync-operator.v0.0.29
        version: 0.0.29
      - name: group-sync-operator.v0.0.28
        version: 0.0.28
      - name: group-sync-operator.v0.0.27
        version: 0.0.27
      - name: group-sync-operator.v0.0.26
        version: 0.0.26
      - name: group-sync-operator.v0.0.25
        version: 0.0.25
      - name: group-sync-operator.v0.0.24
        version: 0.0.24
      - name: group-sync-operator.v0.0.23
        version: 0.0.23
      - name: group-sync-operator.v0.0.22
        version: 0.0.22
      - name: group-sync-operator.v0.0.21
        version: 0.0.21
      - name: group-sync-operator.v0.0.20
        version: 0.0.20
      - name: group-sync-operator.v0.0.19
        version: 0.0.19
      - name: group-sync-operator.v0.0.18
        version: 0.0.18
      - name: group-sync-operator.v0.0.17
        version: 0.0.17
      - name: group-sync-operator.v0.0.16
        version: 0.0.16
      - name: group-sync-operator.v0.0.15
        version: 0.0.15
      - name: group-sync-operator.v0.0.14
        version: 0.0.14
      - name: group-sync-operator.v0.0.13
        version: 0.0.13
      name: alpha
    defaultChannel: alpha
    packageName: group-sync-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubecost
      provider-url: https://www.kubecost.com
    name: kubecost-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubecost-operator.v2.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kubecost.com/v1alpha1",
                "kind": "CostAnalyzer",
                "metadata": {
                  "name": "costanalyzer-sample"
                },
                "spec": {
                  "clusterController": {
                    "image": {
                      "repository": "registry.connect.redhat.com/kubecost/kubecost-cluster-controller@sha256",
                      "tag": "1a6f11dcd7341ab3f24eba23ec06c7b75646a534a7ed960153601806425469d1"
                    }
                  },
                  "forecasting": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-modeling@sha256:8f59f8fd8696964c1b5322ca79fae78befe97323fa56c2b509bdcc67ee4c0edd"
                  },
                  "global": {
                    "platforms": {
                      "openshift": {
                        "createMonitoringClusterRoleBinding": true,
                        "createMonitoringResourceReaderRoleBinding": true,
                        "enabled": true,
                        "scc": {
                          "clusterController": true,
                          "networkCosts": true,
                          "nodeExporter": true
                        }
                      }
                    },
                    "prometheus": {
                      "enabled": false,
                      "fqdn": "https://prometheus-k8s.openshift-monitoring.svc.cluster.local:9091",
                      "kubeRBACProxy": true
                    }
                  },
                  "grafana": {
                    "image": {
                      "repository": "registry.redhat.io/rhel9/grafana@sha256",
                      "tag": "9f71de028c18034e07a8d15dac13ef68afb91cb899b9e8e7d39703547dd0eac1"
                    },
                    "sidecar": {
                      "image": {
                        "repository": "quay.io/kiwigrid/k8s-sidecar@sha256",
                        "tag": "d3749777d5bedf62549b452a7e2ccb205da53588dc7f04927efc75d0da5b6d23"
                      }
                    }
                  },
                  "kubecostAggregator": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef"
                  },
                  "kubecostFrontend": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-frontend@sha256:f9b2196dc6ac017b8074f8fe40f1a58e7a2630282c5b7a019ca9f3dbdb670dc8"
                  },
                  "kubecostModel": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef"
                  },
                  "networkCosts": {
                    "image": {
                      "repository": "registry.connect.redhat.com/kubecost/kubecost-network-costs@sha256",
                      "tag": "c444d475be4f1aa36680bd84358e1e07303dcfe5d13a95544ad9e59f80d4a5f0"
                    }
                  },
                  "prometheusRule": {
                    "enabled": true
                  },
                  "serviceMonitor": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          createdAt: "2025-08-08T19:34:34Z"
          description: Deploys Kubecost 2.8.2 via the Helm chart.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Contact Kubecost for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Kubecost.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The CostAnalyzer API describes a deployment of Kubecost and
              its dependencies.
            displayName: CostAnalyzer
            kind: CostAnalyzer
            name: costanalyzers.charts.kubecost.com
            version: v1alpha1
        description: |2-

          IMPORTANT - Before installation, you must run the following command to allow monitoring of the Kubecost namespace by the local OpenShift Promtheus. If you install in a namespace other than "Kubecost", adjust accordingly.

          `oc label namespace kubecost openshift.io/cluster-monitoring=true`

          Kubecost provides real-time cost visibility and insights for teams using Kubernetes, helping you continuously reduce your cloud costs.

          This OpenShift Operator deploys Kubecost from the official Helm chart.

          Kubecost allows you to:

          * See allocated spend across all native Kubernetes concepts and provide your teams with transparent, accurate cost data reconciled with your actual cloud bill.
          * Join in-cluster costs like CPU and memory with out-of-cluster spend from other cloud services.
          * Get complete cost visibility across OpenShift, On-Prem, AWS, GCP, Azure, Oracle, AliCloud + more.
          * Gain insights you can use to save 30-50%+ on your infrastructure spend, without exposing your private information — your data never gets shared externally, even with us.
          * Have real-time alerts and recurring reports, empowering teams to take control of their Kubernetes-enabled infrastructure and stay within budgeted limits.

          Visit the Kubecost website for more details including features, pricing, documentation, and more at www.kubecost.com.
        displayName: Kubecost
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kubernetes
        - cost
        - kubernetes costs
        - kubernetes optimization
        - k8s
        - cost monitoring
        - cloud costs
        - k8s cost
        links:
        - name: Kubecost
          url: https://www.kubecost.com
        maintainers:
        - email: support@kubecost.com
          name: Kubecost Support
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Kubecost
          url: https://www.kubecost.com
        relatedImages:
        - registry.connect.redhat.com/kubecost/kubecost-network-costs@sha256:c444d475be4f1aa36680bd84358e1e07303dcfe5d13a95544ad9e59f80d4a5f0
        - registry.redhat.io/rhel9/grafana@sha256:9f71de028c18034e07a8d15dac13ef68afb91cb899b9e8e7d39703547dd0eac1
        - quay.io/kiwigrid/k8s-sidecar@sha256:d3749777d5bedf62549b452a7e2ccb205da53588dc7f04927efc75d0da5b6d23
        - registry.connect.redhat.com/kubecost/kubecost-operator-manager@sha256:582223a66a77f35d1f7ca33460032cfca1581a8ff30f75ead289d8cd4cf62bb9
        - registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef
        - registry.connect.redhat.com/kubecost/kubecost-frontend@sha256:f9b2196dc6ac017b8074f8fe40f1a58e7a2630282c5b7a019ca9f3dbdb670dc8
        - registry.connect.redhat.com/kubecost/kubecost-modeling@sha256:8f59f8fd8696964c1b5322ca79fae78befe97323fa56c2b509bdcc67ee4c0edd
        - registry.connect.redhat.com/kubecost/kubecost-cluster-controller@sha256:1a6f11dcd7341ab3f24eba23ec06c7b75646a534a7ed960153601806425469d1
        version: 2.8.2
      entries:
      - name: kubecost-operator.v2.8.2
        version: 2.8.2
      - name: kubecost-operator.v2.8.1
        version: 2.8.1
      - name: kubecost-operator.v2.8.0
        version: 2.8.0
      - name: kubecost-operator.v2.7.2
        version: 2.7.2
      - name: kubecost-operator.v2.7.1
        version: 2.7.1
      - name: kubecost-operator.v2.5.4
        version: 2.5.4
      name: alpha
    defaultChannel: alpha
    packageName: kubecost-operator
    provider:
      name: Kubecost
      url: https://www.kubecost.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-access-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-access-operator.v25.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyAccess",
                "metadata": {
                  "name": "ivia-sample"
                },
                "spec": {
                  "image": "icr.io/ivia/ivia-wrp@sha256:c2dd5ec2b362136896e47ccbd0f3ee033af945fc2794ddaf075da603e4d4d7b2",
                  "instance": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isva/verify-access-operator@sha256:ca1d7be14b4dd313d4675f18541b2cc670ea0a0adf3ec87e347e5e20a86bc39c
          createdAt: "2025-05-01T01:27:17Z"
          description: The IBM Verify Identity Access Operator manages the lifecycle
            of IBM Verify Identity Access worker containers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM-Security/verify-access-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyAccess is the Schema for the ibmsecurityverifyaccesses
              API.
            displayName: IBMSecurity Verify Access
            kind: IBMSecurityVerifyAccess
            name: ibmsecurityverifyaccesses.ibm.com
            version: v1
        description: |+
          In a world of highly fragmented access management environments, [IBM Verify Identity Access](https://www.ibm.com/au-en/products/verify-access) helps you simplify your users' access while more securely adopting web, mobile and cloud technologies. This solution helps you strike a balance between usability and security through the use of risk-based access, single sign-on, integrated access management control, identity federation and its mobile multi-factor authentication capability, IBM Verify. Take back control of your access management with IBM Verify Identity Access.

          The IBM Verify Identity Access operator provides lifecycle management of the lightweight containers which are used to protect an environment, namely:
          * [Web Reverse Proxy](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-web-reverse-proxy)
          * [Runtime](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-runtime)
          * [Distributed Session Cache](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-distributed-session-cache)

          The operator will manage the deployment of these lightweight IBM Verify Identity Access worker containers, and also control the rolling restart of these containers when a configuration snapshot is updated.

          See the project [Readme](https://github.com/IBM-Security/verify-access-operator/blob/master/README.md) for further information and details.

        displayName: IBM Verify Identity Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Identity Access Product Information
          url: https://www.ibm.com/au-en/products/verify-access
        - name: Verify Identity Access Documentation
          url: https://www.ibm.com/docs/en/sva
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Identity Access Development Team
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - icr.io/isva/verify-access-operator@sha256:ca1d7be14b4dd313d4675f18541b2cc670ea0a0adf3ec87e347e5e20a86bc39c
        - icr.io/ivia/ivia-wrp@sha256:c2dd5ec2b362136896e47ccbd0f3ee033af945fc2794ddaf075da603e4d4d7b2
        version: 25.4.0
      entries:
      - name: ibm-security-verify-access-operator.v25.4.0
        version: 25.4.0
      - name: ibm-security-verify-access-operator.v24.12.0
        version: 24.12.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-access-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Redis
      provider-url: ""
    name: redis-enterprise-operator-cert
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-enterprise-operator.v6.2.10-45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
        - registry.connect.redhat.com/redislabs/services-manager@sha256:8b46f2aacfb3e1b243c0b64975accc55fd80b4de9c9ac4769c3c67a5676addc5
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:ffdf10a1b08115fd4de06f1cd751d36644a9bc9a8c11034cd432e9b66a5d2f73
        version: 6.2.10-45
      entries:
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: 6.2.10
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.22.0
    - currentCSV: redis-enterprise-operator.v7.22.0-11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:0cfb07f774db131b5b74e5aa91291579c916c4d4cd1a693d7aaed554f0177a68
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:088e4b57c6e1ca1680b359078e5eb1bba67343cdbd9a539af26a0ea23a830593
        - registry.connect.redhat.com/redislabs/services-manager@sha256:3b96cdc65c3e42fd22eb165950be96b907725903ac74fe4f82cfa0073e28e95a
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:d56ba97e95e799210e76615095409143a5dddd5c5a1e1be1d551d7e3f9e4d3ef
        version: 7.22.0-11.3
      entries:
      - name: redis-enterprise-operator.v7.22.0-11.3
        version: 7.22.0-11.3
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.22.0-11.3
    - currentCSV: redis-enterprise-operator.v7.4.6-7.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:473b5661049a6c8830ba113504975726154c7b172376ba3b63d891674d59c78f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5f6f32a63f4dd8be85b9d469f0fedd26d6aa76e32b7735bfa7ab7203cb272cdb
        - registry.connect.redhat.com/redislabs/services-manager@sha256:ba20ab644a473bc742af671171d5313537292d88942b16336d03a6386c061d8a
        version: 7.4.6-7.2
      entries:
      - name: redis-enterprise-operator.v7.4.6-7.2
        version: 7.4.6-7.2
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      name: 7.4.6
    - currentCSV: redis-enterprise-operator.v7.8.2-6.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:acf817c90b5f03bc2796fbc305cf43ce5c0644dcbe1b685752090b3f600a0f7b
        - registry.connect.redhat.com/redislabs/services-manager@sha256:c6e1f2109a0dac76b51c25c1cbf6030b4894b21b98147e2ad1db2dd8e7693576
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:36c7a36dabb23a4c003e112085cda2873296bfb22e5d11b470d7c1aba55b7f0f
        version: 7.8.2-6.8
      entries:
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.2
    - currentCSV: redis-enterprise-operator.v7.8.4-9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:4ba0c414a74268bed9c6757e447b340db7c912afa864cfb096ddf88f11feeaf9
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:66ec124da16cc9d0abd66687e088be0b358e20cfd48f5774d34bf5874e6a4e65
        - registry.connect.redhat.com/redislabs/services-manager@sha256:803be04534e59cdc7b8898b3bb336676aad1cb77cf266051e226be82679fcf6a
        version: 7.8.4-9.0
      entries:
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.4
    - currentCSV: redis-enterprise-operator.v7.8.6-5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:1fff01daece34aeb7befcb0760d5112c8c7d261f2b86388ece416ac94076a573
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5125e83d3c93b224d1983c599994e43899a0ba0c6f2fc4eed31476f755b5fc0b
        - registry.connect.redhat.com/redislabs/services-manager@sha256:88d483f5e37b5a901d1ae3d965304edc000c172d2e5e307297d5b7bcd77ae73b
        version: 7.8.6-5.1
      entries:
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.6
    - currentCSV: redis-enterprise-operator.v6.2.10-41
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/outbound",
                      "versionTag": "6.2.10-41"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager",
                      "versionTag": "6.2.10-41"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise",
                    "versionTag": "6.2.14-9.rhel8-openshift"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/outbound@sha256:3c251754ae8650157465a56ef3a7b41d5a0ad4e9db09ef042e0014076f25e8a2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/outbound@sha256:3c251754ae8650157465a56ef3a7b41d5a0ad4e9db09ef042e0014076f25e8a2
        version: 6.2.10-41
      entries:
      - name: redis-enterprise-operator.v6.2.10-41
        version: 6.2.10-41
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: preview
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: production
    defaultChannel: production
    packageName: redis-enterprise-operator-cert
    provider:
      name: Redis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: candidate
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: fast
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: stable
    defaultChannel: stable
    packageName: stackgres
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NooBaa
      provider-url: ""
    name: mcg-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mcg-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[{"kind":"NooBaa","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"noobaa","creationTimestamp":null},"spec":{"cleanupPolicy":{},"security":{"kms":{}},"loadBalancerSourceSubnets":{},"autoscaler":{}},"status":{"lastKeyRotateTime":null}},{"kind":"BackingStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"NamespaceStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"BucketClass","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{},"status":{}},{"kind":"NooBaaAccount","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"allow_bucket_creation":false},"status":{}}]'
          capabilities: Deep Insights
          categories: Storage,Big Data
          certified: "false"
          containerImage: registry.redhat.io/odf4/mcg-rhel9-operator@sha256:43ddea9d3c1cde5c8b38f74e8840a871dd0441ed9fd074528bacf78b439f024f
          createdAt: "2019-07-08T13:10:20.940Z"
          description: NooBaa is an object data service for hybrid and multi cloud
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operators.openshift.io/infrastructure-features: ֿ'["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/noobaa/noobaa-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A NooBaa system - Create this to start
            displayName: NooBaa
            kind: NooBaa
            name: noobaas.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos,
              PV's and more. Used in BucketClass to construct data placement policies.
            displayName: Backing Store
            kind: BackingStore
            name: backingstores.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos
              and more. Used in BucketClass to construct namespace policies.
            displayName: Namespace Store
            kind: NamespaceStore
            name: namespacestores.noobaa.io
            version: v1alpha1
          - description: Storage policy spec  tiering, mirroring, spreading, namespace
              policy. Combines BackingStores Or NamespaceStores. Referenced by ObjectBucketClaims.
            displayName: Bucket Class
            kind: BucketClass
            name: bucketclasses.noobaa.io
            version: v1alpha1
          - kind: NooBaaAccount
            name: noobaaaccounts.noobaa.io
            version: v1alpha1
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information.
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: "The noobaa operator creates and reconciles a NooBaa system in
          a Kubernetes/Openshift cluster.\n\nNooBaa provides an S3 object-store service
          abstraction and data placement policies to create hybrid and multi cloud
          data solutions.\n\nFor more information on using NooBaa refer to [Github](https://github.com/noobaa/noobaa-core)
          / [Website](https://www.noobaa.io) / [Articles](https://noobaa.desk.com).
          \n\n## How does it work?\n\n- The operator deploys the noobaa core pod and
          two services - Mgmt (UI/API) and S3 (object-store).\n- Both services require
          credentials which you will get from a secret that the operator creates -
          use describe noobaa to locate it.\n- The service addresses will also appear
          in the describe output - pick the one that is suitable for your client:\n
          \   - minikube - use the NodePort address.\n    - remote cluster - probably
          need one of the External addresses.\n    - connect an application on the
          same cluster - use Internal DNS (though any address should work)\n    \n-
          Feel free to email us or open github issues on any question.\n\n## Getting
          Started\n\n### Notes:\n- The following instructions are for **minikube**
          but it works on any Kubernetes/Openshift clusters.\n- This will setup noobaa
          in the **my-noobaa-operator** namespace.\n- You will need **jq**, **curl**,
          **kubectl** or **oc**, **aws-cli**.\n\n### 1. Install OLM (if you don't
          have it already):\n```\ncurl -sL https://github.com/operator-framework/operator-lifecycle-manager/releases/download/0.12.0/install.sh
          | bash -s 0.12.0\n```\n\n### 2. Install noobaa-operator:\n```\nkubectl create
          -f https://operatorhub.io/install/noobaa-operator.yaml\n```\nWait for it
          to be ready:\n```\nkubectl wait pod -n my-noobaa-operator -l noobaa-operator
          --for=condition=ready\n```\n\n### 3. Create noobaa system:\n```\ncurl -sL
          https://operatorhub.io/api/operator?packageName=noobaa-operator | \n    jq
          '.operator.customResourceDefinitions[0].yamlExample | .metadata.namespace=\"my-noobaa-operator\"'
          |\n    kubectl create -f -\n```\nWait for it to be ready:\n```\nkubectl
          wait pod -n my-noobaa-operator -l noobaa-core --for=condition=ready\nkubectl
          get noobaa -n my-noobaa-operator -w\n# NAME     PHASE   MGMT-ENDPOINTS                  S3-ENDPOINTS
          \                   IMAGE                    AGE\n# noobaa   **Ready**   [https://192.168.64.12:31121]
          \  [https://192.168.64.12:32557]   noobaa/noobaa-core:4.0   19m\n```\n\n###
          4. Get system information to your shell:\n```\nNOOBAA_SECRET=$(kubectl get
          noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.accounts.admin.secretRef.name'
          )\nNOOBAA_MGMT=$(kubectl get noobaa noobaa -n my-noobaa-operator -o json
          | jq -r '.status.services.serviceMgmt.nodePorts[0]' )\nNOOBAA_S3=$(kubectl
          get noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.services.serviceS3.nodePorts[0]'
          )\nNOOBAA_ACCESS_KEY=$(kubectl get secret $NOOBAA_SECRET -n my-noobaa-operator
          -o json | jq -r '.data.AWS_ACCESS_KEY_ID|@base64d')\nNOOBAA_SECRET_KEY=$(kubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq -r '.data.AWS_SECRET_ACCESS_KEY|@base64d')\n```\n\n###
          5. Connect to Mgmt UI:\n```\n# show email/password from the secret:\nkubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq '.data|map_values(@base64d)'\n\n#
          open mgmt UI login:\nopen $NOOBAA_MGMT\n```\n\n### 6. Connect to S3 with
          aws-cli:\n```\nalias s3='AWS_ACCESS_KEY_ID=$NOOBAA_ACCESS_KEY AWS_SECRET_ACCESS_KEY=$NOOBAA_SECRET_KEY
          aws --endpoint $NOOBAA_S3 --no-verify-ssl s3'\ns3 ls\ns3 sync /var/log/
          s3://first.bucket\ns3 ls s3://first.bucket\n```\n"
        displayName: NooBaa Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - noobaa
        - kubernetes
        - openshift
        - cloud
        - hybrid
        - multi
        - data
        - storage
        - s3
        - tiering
        - mirroring
        links:
        - name: Github
          url: https://github.com/noobaa/noobaa-core
        - name: Website
          url: https://www.noobaa.io
        - name: Articles
          url: https://noobaa.desk.com
        maintainers:
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: etamir@redhat.com
          name: Eran Tamir
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: NooBaa
        relatedImages:
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:f37e2c611f4cda32172d4e8938938e488076fd8a8f19c252c56a80c48044bc02
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhel8/postgresql-12@sha256:af1ab437b165fa4ee9dfd6ade637d7de5fa556d0b3851fb8edc78afe0e6fa95e
        - registry.redhat.io/odf4/mcg-rhel9-operator@sha256:43ddea9d3c1cde5c8b38f74e8840a871dd0441ed9fd074528bacf78b439f024f
        - registry.redhat.io/odf4/odf-cosi-sidecar-rhel9@sha256:bce36c7d33a85a11cd6b6b24f92887cffe6f38af70340e54d10e4f6f524aa782
        version: 4.15.20-rhodf
      entries:
      - name: mcg-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: mcg-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: mcg-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: mcg-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: mcg-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: mcg-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: mcg-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: mcg-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: mcg-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: mcg-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: mcg-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: mcg-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: mcg-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: mcg-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: mcg-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: mcg-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: mcg-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: mcg-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: mcg-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: mcg-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: mcg-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[{"kind":"NooBaa","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"noobaa","creationTimestamp":null},"spec":{"cleanupPolicy":{},"security":{"kms":{}},"loadBalancerSourceSubnets":{},"autoscaler":{}},"status":{"lastKeyRotateTime":null}},{"kind":"BackingStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"NamespaceStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"BucketClass","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{},"status":{}},{"kind":"NooBaaAccount","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"allow_bucket_creation":false},"status":{}}]'
          capabilities: Deep Insights
          categories: Storage,Big Data
          certified: "false"
          containerImage: registry.redhat.io/odf4/mcg-rhel9-operator@sha256:fdcdc8870b12d853973000877818463c3b4a5ed1d3316ff72a04a299577ed23d
          createdAt: "2019-07-08T13:10:20.940Z"
          description: NooBaa is an object data service for hybrid and multi cloud
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/infrastructure-features: ֿ'["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/noobaa/noobaa-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A NooBaa system - Create this to start
            displayName: NooBaa
            kind: NooBaa
            name: noobaas.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos,
              PV's and more. Used in BucketClass to construct data placement policies.
            displayName: Backing Store
            kind: BackingStore
            name: backingstores.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos
              and more. Used in BucketClass to construct namespace policies.
            displayName: Namespace Store
            kind: NamespaceStore
            name: namespacestores.noobaa.io
            version: v1alpha1
          - description: Storage policy spec  tiering, mirroring, spreading, namespace
              policy. Combines BackingStores Or NamespaceStores. Referenced by ObjectBucketClaims.
            displayName: Bucket Class
            kind: BucketClass
            name: bucketclasses.noobaa.io
            version: v1alpha1
          - kind: NooBaaAccount
            name: noobaaaccounts.noobaa.io
            version: v1alpha1
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information.
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: "The noobaa operator creates and reconciles a NooBaa system in
          a Kubernetes/Openshift cluster.\n\nNooBaa provides an S3 object-store service
          abstraction and data placement policies to create hybrid and multi cloud
          data solutions.\n\nFor more information on using NooBaa refer to [Github](https://github.com/noobaa/noobaa-core)
          / [Website](https://www.noobaa.io) / [Articles](https://noobaa.desk.com).
          \n\n## How does it work?\n\n- The operator deploys the noobaa core pod and
          two services - Mgmt (UI/API) and S3 (object-store).\n- Both services require
          credentials which you will get from a secret that the operator creates -
          use describe noobaa to locate it.\n- The service addresses will also appear
          in the describe output - pick the one that is suitable for your client:\n
          \   - minikube - use the NodePort address.\n    - remote cluster - probably
          need one of the External addresses.\n    - connect an application on the
          same cluster - use Internal DNS (though any address should work)\n    \n-
          Feel free to email us or open github issues on any question.\n\n## Getting
          Started\n\n### Notes:\n- The following instructions are for **minikube**
          but it works on any Kubernetes/Openshift clusters.\n- This will setup noobaa
          in the **my-noobaa-operator** namespace.\n- You will need **jq**, **curl**,
          **kubectl** or **oc**, **aws-cli**.\n\n### 1. Install OLM (if you don't
          have it already):\n```\ncurl -sL https://github.com/operator-framework/operator-lifecycle-manager/releases/download/0.12.0/install.sh
          | bash -s 0.12.0\n```\n\n### 2. Install noobaa-operator:\n```\nkubectl create
          -f https://operatorhub.io/install/noobaa-operator.yaml\n```\nWait for it
          to be ready:\n```\nkubectl wait pod -n my-noobaa-operator -l noobaa-operator
          --for=condition=ready\n```\n\n### 3. Create noobaa system:\n```\ncurl -sL
          https://operatorhub.io/api/operator?packageName=noobaa-operator | \n    jq
          '.operator.customResourceDefinitions[0].yamlExample | .metadata.namespace=\"my-noobaa-operator\"'
          |\n    kubectl create -f -\n```\nWait for it to be ready:\n```\nkubectl
          wait pod -n my-noobaa-operator -l noobaa-core --for=condition=ready\nkubectl
          get noobaa -n my-noobaa-operator -w\n# NAME     PHASE   MGMT-ENDPOINTS                  S3-ENDPOINTS
          \                   IMAGE                    AGE\n# noobaa   **Ready**   [https://192.168.64.12:31121]
          \  [https://192.168.64.12:32557]   noobaa/noobaa-core:4.0   19m\n```\n\n###
          4. Get system information to your shell:\n```\nNOOBAA_SECRET=$(kubectl get
          noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.accounts.admin.secretRef.name'
          )\nNOOBAA_MGMT=$(kubectl get noobaa noobaa -n my-noobaa-operator -o json
          | jq -r '.status.services.serviceMgmt.nodePorts[0]' )\nNOOBAA_S3=$(kubectl
          get noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.services.serviceS3.nodePorts[0]'
          )\nNOOBAA_ACCESS_KEY=$(kubectl get secret $NOOBAA_SECRET -n my-noobaa-operator
          -o json | jq -r '.data.AWS_ACCESS_KEY_ID|@base64d')\nNOOBAA_SECRET_KEY=$(kubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq -r '.data.AWS_SECRET_ACCESS_KEY|@base64d')\n```\n\n###
          5. Connect to Mgmt UI:\n```\n# show email/password from the secret:\nkubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq '.data|map_values(@base64d)'\n\n#
          open mgmt UI login:\nopen $NOOBAA_MGMT\n```\n\n### 6. Connect to S3 with
          aws-cli:\n```\nalias s3='AWS_ACCESS_KEY_ID=$NOOBAA_ACCESS_KEY AWS_SECRET_ACCESS_KEY=$NOOBAA_SECRET_KEY
          aws --endpoint $NOOBAA_S3 --no-verify-ssl s3'\ns3 ls\ns3 sync /var/log/
          s3://first.bucket\ns3 ls s3://first.bucket\n```\n"
        displayName: NooBaa Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - noobaa
        - kubernetes
        - openshift
        - cloud
        - hybrid
        - multi
        - data
        - storage
        - s3
        - tiering
        - mirroring
        links:
        - name: Github
          url: https://github.com/noobaa/noobaa-core
        - name: Website
          url: https://www.noobaa.io
        - name: Articles
          url: https://noobaa.desk.com
        maintainers:
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: etamir@redhat.com
          name: Eran Tamir
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: NooBaa
        relatedImages:
        - registry.redhat.io/odf4/mcg-rhel9-operator@sha256:fdcdc8870b12d853973000877818463c3b4a5ed1d3316ff72a04a299577ed23d
        - registry.redhat.io/odf4/odf-cosi-sidecar-rhel9@sha256:9fed589373d1586577d57f23b7a4f39ace5052fc55c99a03cb256da63a26c441
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:0084d9053ad2c74926ca02d2ede0461abd59227eee98d122c45e08a5ebaf629a
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhel8/postgresql-12@sha256:af1ab437b165fa4ee9dfd6ade637d7de5fa556d0b3851fb8edc78afe0e6fa95e
        version: 4.16.16-rhodf
      entries:
      - name: mcg-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: mcg-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: mcg-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: mcg-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: mcg-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: mcg-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: mcg-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: mcg-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: mcg-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: mcg-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: mcg-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: mcg-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: mcg-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: mcg-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: mcg-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: mcg-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: mcg-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: mcg-operator
    provider:
      name: NooBaa
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-eks-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-eks-controller.v1.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "FargateProfile",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Nodegroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/eks-controller:1.9.3
          createdAt: "2025-10-03T21:43:09Z"
          description: AWS EKS controller is a service controller for managing EKS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessEntry represents the state of an AWS eks AccessEntry
              resource.
            displayName: AccessEntry
            kind: AccessEntry
            name: accessentries.eks.services.k8s.aws
            version: v1alpha1
          - description: Addon represents the state of an AWS eks Addon resource.
            displayName: Addon
            kind: Addon
            name: addons.eks.services.k8s.aws
            version: v1alpha1
          - description: Cluster represents the state of an AWS eks Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.eks.services.k8s.aws
            version: v1alpha1
          - description: FargateProfile represents the state of an AWS eks FargateProfile
              resource.
            displayName: FargateProfile
            kind: FargateProfile
            name: fargateprofiles.eks.services.k8s.aws
            version: v1alpha1
          - description: IdentityProviderConfig represents the state of an AWS eks
              IdentityProviderConfig resource.
            displayName: IdentityProviderConfig
            kind: IdentityProviderConfig
            name: identityproviderconfigs.eks.services.k8s.aws
            version: v1alpha1
          - description: Nodegroup represents the state of an AWS eks Nodegroup resource.
            displayName: Nodegroup
            kind: Nodegroup
            name: nodegroups.eks.services.k8s.aws
            version: v1alpha1
          - description: PodIdentityAssociation represents the state of an AWS eks
              PodIdentityAssociation resource.
            displayName: PodIdentityAssociation
            kind: PodIdentityAssociation
            name: podidentityassociations.eks.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Kubernetes Service (EKS) resources in AWS from within your Kubernetes cluster.

          **About Amazon EKS**

          Amazon Elastic Kubernetes Service (Amazon EKS) is a managed service that you can use to run Kubernetes on AWS without needing to install, operate, and maintain your own Kubernetes control plane or nodes. Kubernetes is an open-source system for automating the deployment, scaling, and management of containerized applications.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EKS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eks
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EKS Developer Resources
          url: https://aws.amazon.com/eks/getting-started/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/eks-controller:1.9.3
        version: 1.9.3
      entries:
      - name: ack-eks-controller.v1.9.3
        version: 1.9.3
      - name: ack-eks-controller.v1.9.2
        version: 1.9.2
      - name: ack-eks-controller.v1.9.1
        version: 1.9.1
      - name: ack-eks-controller.v1.9.0
        version: 1.9.0
      - name: ack-eks-controller.v1.8.0
        version: 1.8.0
      - name: ack-eks-controller.v1.7.6
        version: 1.7.6
      - name: ack-eks-controller.v1.7.5
        version: 1.7.5
      - name: ack-eks-controller.v1.7.4
        version: 1.7.4
      - name: ack-eks-controller.v1.7.3
        version: 1.7.3
      - name: ack-eks-controller.v1.7.2
        version: 1.7.2
      - name: ack-eks-controller.v1.6.4
        version: 1.6.4
      - name: ack-eks-controller.v1.6.3
        version: 1.6.3
      - name: ack-eks-controller.v1.6.2
        version: 1.6.2
      - name: ack-eks-controller.v1.6.1
        version: 1.6.1
      - name: ack-eks-controller.v1.6.0
        version: 1.6.0
      - name: ack-eks-controller.v1.5.5
        version: 1.5.5
      - name: ack-eks-controller.v1.5.4
        version: 1.5.4
      - name: ack-eks-controller.v1.5.3
        version: 1.5.3
      - name: ack-eks-controller.v1.5.2
        version: 1.5.2
      - name: ack-eks-controller.v1.5.1
        version: 1.5.1
      - name: ack-eks-controller.v1.5.0
        version: 1.5.0
      - name: ack-eks-controller.v1.4.7
        version: 1.4.7
      - name: ack-eks-controller.v1.4.6
        version: 1.4.6
      - name: ack-eks-controller.v1.4.5
        version: 1.4.5
      - name: ack-eks-controller.v1.4.4
        version: 1.4.4
      - name: ack-eks-controller.v1.4.3
        version: 1.4.3
      - name: ack-eks-controller.v1.4.2
        version: 1.4.2
      - name: ack-eks-controller.v1.4.1
        version: 1.4.1
      - name: ack-eks-controller.v1.4.0
        version: 1.4.0
      - name: ack-eks-controller.v1.3.5
        version: 1.3.5
      - name: ack-eks-controller.v1.3.3
        version: 1.3.3
      - name: ack-eks-controller.v1.3.1
        version: 1.3.1
      - name: ack-eks-controller.v1.3.0
        version: 1.3.0
      - name: ack-eks-controller.v1.2.3
        version: 1.2.3
      - name: ack-eks-controller.v1.2.2
        version: 1.2.2
      - name: ack-eks-controller.v1.2.1
        version: 1.2.1
      - name: ack-eks-controller.v1.2.0
        version: 1.2.0
      - name: ack-eks-controller.v1.1.3
        version: 1.1.3
      - name: ack-eks-controller.v1.1.2
        version: 1.1.2
      - name: ack-eks-controller.v1.1.1
        version: 1.1.1
      - name: ack-eks-controller.v1.1.0
        version: 1.1.0
      - name: ack-eks-controller.v1.0.5
        version: 1.0.5
      - name: ack-eks-controller.v1.0.4
        version: 1.0.4
      - name: ack-eks-controller.v1.0.3
        version: 1.0.3
      - name: ack-eks-controller.v1.0.2
        version: 1.0.2
      - name: ack-eks-controller.v1.0.1
        version: 1.0.1
      - name: ack-eks-controller.v1.0.0
        version: 1.0.0
      - name: ack-eks-controller.v0.1.9
        version: 0.1.9
      - name: ack-eks-controller.v0.1.8
        version: 0.1.8
      - name: ack-eks-controller.v0.1.7
        version: 0.1.7
      - name: ack-eks-controller.v0.1.6
        version: 0.1.6
      - name: ack-eks-controller.v0.1.5
        version: 0.1.5
      - name: ack-eks-controller.v0.1.4
        version: 0.1.4
      - name: ack-eks-controller.v0.1.3
        version: 0.1.3
      - name: ack-eks-controller.v0.1.2
        version: 0.1.2
      - name: ack-eks-controller.v0.1.1
        version: 0.1.1
      - name: ack-eks-controller.v0.1.0
        version: 0.1.0
      - name: ack-eks-controller.v0.0.15
        version: 0.0.15
      - name: ack-eks-controller.v0.0.14
        version: 0.0.14
      - name: ack-eks-controller.v0.0.13
        version: 0.0.13
      - name: ack-eks-controller.v0.0.12
        version: 0.0.12
      - name: ack-eks-controller.v0.0.11
        version: 0.0.11
      - name: ack-eks-controller.v0.0.10
        version: 0.0.10
      - name: ack-eks-controller.v0.0.9
        version: 0.0.9
      - name: ack-eks-controller.v0.0.8
        version: 0.0.8
      - name: ack-eks-controller.v0.0.7
        version: 0.0.7
      - name: ack-eks-controller.v0.0.6
        version: 0.0.6
      - name: ack-eks-controller.v0.0.5
        version: 0.0.5
      - name: ack-eks-controller.v0.0.4
        version: 0.0.4
      - name: ack-eks-controller.v0.0.3
        version: 0.0.3
      - name: ack-eks-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-eks-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ODH
      provider-url: ""
    name: opendatahub-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opendatahub-operator.v2.35.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Managed",
                      "registriesNamespace": "odh-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "opendatahub",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "opendatahub"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v2.35.0
          createdAt: "2025-10-01T13:12:27Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",
            "workbenches.components.platform.opendatahub.io", "monitorings.services.platform.opendatahub.io",
            "servicemeshes.services.platform.opendatahub.io", "modelcontrollers.components.platform.opendatahub.io",
            "feastoperators.components.platform.opendatahub.io", "llamastackoperators.components.platform.opendatahub.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/opendatahub-io/opendatahub-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: ServiceMesh is the Schema for the servicemesh API
            displayName: Service Mesh
            kind: ServiceMesh
            name: servicemeshes.services.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack - Unified open-source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odh
        - notebooks
        - serving
        - training
        - pipelines
        - modelmesh
        - workbenches
        - dashboard
        - kserve
        - distributed workloads
        - trustyai
        - modelregistry
        - feast
        - featurestore
        - llamastack
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ODH
        relatedImages:
        - quay.io/opendatahub/opendatahub-operator:v2.35.0
        version: 2.35.0
      entries:
      - name: opendatahub-operator.v2.35.0
        version: 2.35.0
      - name: opendatahub-operator.v2.34.0
        version: 2.34.0
      - name: opendatahub-operator.v2.33.0
        version: 2.33.0
      - name: opendatahub-operator.v2.32.0
        version: 2.32.0
      - name: opendatahub-operator.v2.31.0
        version: 2.31.0
      - name: opendatahub-operator.v2.30.0
        version: 2.30.0
      - name: opendatahub-operator.v2.29.0
        version: 2.29.0
      - name: opendatahub-operator.v2.28.0
        version: 2.28.0
      - name: opendatahub-operator.v2.27.0
        version: 2.27.0
      - name: opendatahub-operator.v2.26.0
        version: 2.26.0
      - name: opendatahub-operator.v2.25.0
        version: 2.25.0
      - name: opendatahub-operator.v2.24.0
        version: 2.24.0
      - name: opendatahub-operator.v2.23.0
        version: 2.23.0
      - name: opendatahub-operator.v2.22.0
        version: 2.22.0
      - name: opendatahub-operator.v2.21.0
        version: 2.21.0
      - name: opendatahub-operator.v2.20.0
        version: 2.20.0
      - name: opendatahub-operator.v2.19.0
        version: 2.19.0
      - name: opendatahub-operator.v2.18.2
        version: 2.18.2
      - name: opendatahub-operator.v2.18.1
        version: 2.18.1
      - name: opendatahub-operator.v2.18.0
        version: 2.18.0
      - name: opendatahub-operator.v2.17.0
        version: 2.17.0
      - name: opendatahub-operator.v2.16.0
        version: 2.16.0
      - name: opendatahub-operator.v2.15.0
        version: 2.15.0
      - name: opendatahub-operator.v2.14.0
        version: 2.14.0
      - name: opendatahub-operator.v2.13.0
        version: 2.13.0
      - name: opendatahub-operator.v2.12.1
        version: 2.12.1
      - name: opendatahub-operator.v2.12.0
        version: 2.12.0
      - name: opendatahub-operator.v2.11.1
        version: 2.11.1
      - name: opendatahub-operator.v2.11.0
        version: 2.11.0
      - name: opendatahub-operator.v2.10.0
        version: 2.10.0
      name: fast
    - currentCSV: opendatahub-operator.v2.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "opendatahub",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "opendatahub"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v2.8.0
          createdAt: "2024-2-20T00:00:00Z"
          operatorframework.io/initialization-resource: |-
            {
            "apiVersion": "dscinitialization.opendatahub.io/v1",
            "kind": "DSCInitialization",
            "metadata": {
            "labels": {
             "app.kubernetes.io/created-by": "opendatahub-operator",
             "app.kubernetes.io/instance": "default",
             "app.kubernetes.io/managed-by": "kustomize",
             "app.kubernetes.io/name": "dscinitialization",
             "app.kubernetes.io/part-of": "opendatahub-operator"
            },
            "name": "default-dsci"
            },
            "spec": {
            "applicationsNamespace": "opendatahub",
            "monitoring": {
             "managementState": "Managed",
             "namespace": "opendatahub"
            },
            "serviceMesh": {
             "controlPlane": {
               "metricsCollection": "Istio",
               "name": "data-science-smcp",
               "namespace": "istio-system"
             },
             "managementState": "Managed"
            },
            "trustedCABundle": {
             "customCABundle": "",
             "managementState": "Managed"
            }
            }
            }
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: "The Open Data Hub is a machine-learning-as-a-service platform
          built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open
          Data Hub integrates multiple AI/ML open source components into one operator
          that can easily be downloaded and installed by OpenShift users.\n\nOpen
          Data Hub operator allows users to install and manage components of the Open
          Data Hub. Users can mix and match tools from each project to fulfill the
          needs of their use case. Each of the projects share some components, but
          can be mostly seen as an extension of each other to provide a complete solution
          for both novice and skilled enterprise users.\n    \n### Components\n\n*
          Open Data Hub Dashboard - A web dashboard that displays installed Open Data
          Hub components with easy access to component UIs and documentation\n* ODH
          Notebook Controller - Secure management of Jupyter Notebook in Kubernetes
          environments built on top of Kubeflow Notebook Controller with support for
          OAuth\n* Jupyter Notebooks - JupyterLab notebook that provide Python support
          for GPU workloads\n* Data Science Pipelines - Pipeline solution for end
          to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton\n*
          Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh,
          a general-purpose model serving management/routing layer\n* Distributed
          Workloads(Incubation) - Stack built to make managing distributed compute
          infrastructure in the cloud easy and intuitive for Data Scientists. This
          stack consists of three components \n                                      Codeflare
          , KubeRay and Kueue.\n* Kserve - Kserve is the Controller for for serving
          machine learning (ML) models on arbitrary frameworks"
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odh
        - notebooks
        - serving
        - training
        - pipelines
        - modelmesh
        - workbenches
        - dashboard
        - kserve
        - distributed-workloads
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: ODH
        relatedImages:
        - quay.io/opendatahub/opendatahub-operator:v2.8.0
        version: 2.8.1
      entries:
      - name: opendatahub-operator.v2.8.1
        version: 2.8.1
      name: odh-2.8.z
    - currentCSV: opendatahub-operator.v1.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kfdef.apps.kubeflow.org/v1",
                "kind": "KfDef",
                "metadata": {
                  "name": "opendatahub",
                  "namespace": "opendatahub"
                },
                "spec": {
                  "applications": [
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-common"
                        }
                      },
                      "name": "odh-common"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-dashboard"
                        }
                      },
                      "name": "odh-dashboard"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "data-science-pipelines-operator/"
                        }
                      },
                      "name": "data-science-pipelines-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-notebook-controller"
                        }
                      },
                      "name": "odh-notebook-controller"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "notebook-images"
                        }
                      },
                      "name": "notebook-images"
                    },
                    {
                      "kustomizeConfig": {
                        "parameters": [
                          {
                            "name": "monitoring-namespace",
                            "value": "opendatahub"
                          }
                        ],
                        "repoRef": {
                          "name": "manifests",
                          "path": "model-mesh"
                        }
                      },
                      "name": "model-mesh"
                    }
                  ],
                  "repos": [
                    {
                      "name": "manifests",
                      "uri": "file:///opt/manifests/odh-manifests.tar.gz"
                    }
                  ]
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhApplication",
                "metadata": {
                  "name": "example-odhapplication",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started,Model development,Model training,Notebook environments,Package management,Data Management"
                  }
                },
                "spec": {
                  "displayName": "ODHAPPLICATION DISPLAY NAME",
                  "provider": "PROVIDER_NAME",
                  "description": "Description of your ODHApplication",
                  "kfdefApplications": [
                    "<NAME OF THE ODH APPLICATION IN KFDEF>"
                  ],
                  "route": "<APPLICATION_OCP_ROUTE_OBJECT_NAME>",
                  "img": "<svg xmlns=\"http://www.w3.org/2000/svg\" role=\"img\" viewBox=\"-1.13 -1.13 58.50 48.94\"><title>Open Data Hub logo</title><g><g><g fill=\"#4d5258\" stroke-width=\".006\"><path fill=\"#bbb\" d=\"M112.30231 22.797083c-.2034 0-.35737.143517-.35737.337407 0 .192939.15397.339307.35737.339307.20339 0 .35736-.147318.35736-.339307 0-.19294-.15397-.337407-.35736-.337407zm.002.131161c.1112 0 .20149.08934.20149.206246 0 .116904-.0903.208146-.20149.208146-.1112 0-.2053-.09124-.2053-.208146 0-.116904.0931-.206246.2053-.206246zM113.04141 22.801836h-.28228v.665308h.15017v-.190088h.13211c.16918 0 .26612-.08934.26612-.244263 0-.147319-.0969-.230957-.26612-.230957zm-.008.349762h-.12451v-.225255h.12451c.0836 0 .13116.03707.13116.111202 0 .07603-.0475.114053-.13116.114053zM113.40328 22.801836v.665308h.51703v-.124508h-.36687v-.147318h.32125V23.07081h-.32125v-.144467h.35642v-.124507zM114.03814 22.801836v.665308h.14352v-.426748l.3222.426748h.13686v-.665308h-.14257v.427698L114.175 22.801836z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/><path d=\"M115.29685 22.801836h-.28228v.665308h.27658c.21004 0 .35546-.136863.35546-.332654 0-.196741-.14351-.332654-.34976-.332654zm.004.538899h-.13591v-.412491h.12926c.11405 0 .19959.08459.19959.207196 0 .121657-.0817.205295-.19294.205295zM116.21878 23.467144h.16062l-.28418-.665308h-.15492l-.28894.665308h.15493l.0523-.12926h.30794zm-.3108-.249966l.1055-.25852.10455.25852zM116.31525 22.801836v.127359h.20244v.537949h.15017v-.537949h.2034v-.127359zM117.37527 23.467144h.16063l-.28419-.665308h-.15492l-.28893.665308h.15492l.0523-.12926h.30795zm-.31079-.249966l.1055-.25852.10454.25852zM117.83433 22.801836v.665308h.15017v-.262322h.30129v.262322h.15017v-.665308h-.15017v.277528h-.30129v-.277528zM118.72025 23.188665v-.386829h-.15017v.386829c0 .175831.11785.285132.30224.285132.18343 0 .29844-.109301.29844-.285132v-.386829h-.15017v.386829c0 .09599-.0589.153971-.14827.153971-.0893 0-.15207-.05798-.15207-.153971zM119.61059 22.801836h-.30604v.665308h.31649c.15492 0 .25092-.07033.25092-.184385 0-.08744-.0532-.152071-.13876-.168228.0713-.01806.115-.07413.115-.146368 0-.102648-.0912-.166327-.23761-.166327zm-.008.268024h-.14826v-.145418h.14826c.0561 0 .0913.02566.0913.07128 0 .04657-.0352.07413-.0913.07413zm0 .274677h-.14826v-.158723h.14826c.0713 0 .11406.02946.11406.07794 0 .05132-.0428.08079-.11406.08079z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g><g fill=\"#58595b\" stroke-width=\".01\"><path d=\"M112.19192 24.15027h.0381l-.13717-.306759h-.0364l-.1376.306759h.0372l.0342-.07844h.16696zm-.18712-.110871l.0688-.157763.0697.157763zM112.27858 23.843511v.306759h.0351v-.306759zM112.64205 23.843511h-.11525v.306759h.0351v-.09729h.0802c.0758 0 .11964-.039.11964-.10649 0-.0653-.0438-.102983-.11964-.102983zm-.001.177044h-.0789V23.87594h.0789c.0552 0 .0872.02454.0872.07143 0 .0482-.032.07318-.0872.07318zM112.81775 23.825105v.325165h.0342v-.325165zM113.1137 24.15027l-.00044-.152503c-.00044-.04908-.032-.07932-.089-.07932-.0394 0-.0644.01139-.0942.03111l.0145.0241c.0241-.01709.0478-.02673.0732-.02673.0408 0 .0618.0206.0618.05434v.01446h-.0758c-.0543.000439-.0863.02586-.0863.06617 0 .03944.0307.06968.0784.06968.0368 0 .0662-.01227.0846-.03725v.03593zm-.11131-.02717c-.0324 0-.053-.01709-.053-.04251 0-.02585.0202-.03769.0587-.03769h.0719v.02805c-.009.03287-.0386.05215-.0776.05215zM113.29915 24.108201c-.0127.0074-.025.01139-.0368.01183-.0197 0-.0307-.01139-.0307-.03944v-.127086h.0732v-.02673h-.0732v-.06048h-.0337v.06048h-.0333v.02673h.0333v.132783c0 .04558.025.06573.0587.0653.0184 0 .0359-.0061.0521-.01709zM113.44228 23.854466c.0118 0 .0245.0044.0351.01183l.014-.02629c-.014-.01052-.032-.01622-.0495-.01622-.0386-.000438-.0741.02717-.0741.07713v.02586h-.0298v.02673h.0298v.196764h.0337v-.196764h.0662v-.02673h-.0662v-.03024c0-.02717.0184-.04251.0408-.04207zM113.59236 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM113.80323 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM114.22747 23.918448c-.0443.000438-.0767.01972-.0911.06004-.01-.03813-.0394-.06004-.0828-.06004-.0416.000438-.0727.01753-.0881.05346v-.05215h-.0337v.230508h.0337v-.113501c0-.04996.0285-.08458.0758-.08589.0403 0 .064.02454.064.06617v.133221h.0346v-.113501c0-.04996.0276-.08458.0754-.08589.0399 0 .064.02454.064.06617v.133221h.0342V24.00828c0-.05609-.032-.08984-.0859-.08984zM114.63474 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0487.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11088-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM114.89934 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM115.02844 23.919762l.0863.230508h.0355l.0675-.188876.0679.188876h.0355l.0855-.230508h-.0355l-.0675.195888-.0679-.195888h-.0359l-.067.195888-.0684-.195888zM115.53009 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM115.73075 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM115.95034 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11482.116569.0386 0 .071-.01402.0929-.03813l-.0188-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18843c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM116.29474 23.825105v.144615c-.0171-.03287-.0491-.05171-.0903-.05171-.0653 0-.11087.04777-.11087.116569 0 .0688.0456.117445.11131.117445.0412 0 .0727-.01884.0898-.05127v.04952h.0342v-.325165zm-.0833.297557c-.0491 0-.0837-.03637-.0837-.08721 0-.05083.0346-.08765.0837-.08765.0486 0 .0833.03681.0833.08765 0 .05083-.0346.08721-.0833.08721zM116.65984 23.918448c-.0403 0-.0719.01884-.089.05083v-.144177h-.0342v.325165h.0342v-.04952c.0171.03199.0486.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM116.89657 24.183138l.10605-.263376h-.0351l-.0732.191944-.0767-.191944h-.0359l.0955.228756-.01.02191c-.009.02191-.0241.03331-.0421.03331-.0114 0-.0219-.0035-.0324-.01095l-.0145.02717c.014.01139.0294.01665.0486.01665.0307 0 .0548-.01841.0692-.05346zM117.29947 23.842634c-.0898 0-.16083.06793-.16083.154257 0 .08633.071.154694.16083.154694.0894 0 .16039-.06836.16039-.154694 0-.08633-.071-.154257-.16039-.154257zm0 .03243c.0684 0 .12446.0539.12446.121828 0 .06792-.0561.122703-.12446.122703-.0684 0-.12533-.05478-.12533-.122703 0-.06793.057-.121828.12533-.121828zM117.64414 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0486.05083.0898.05083.0653 0 .10999-.04777.10999-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM117.9061 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM118.19616 23.918448c-.0421.000438-.0732.01709-.0894.05215v-.05084h-.0342v.230508h.0342v-.12358c.004-.0447.032-.07494.0776-.07537.0408 0 .0653.02454.0653.06573v.133221h.0342v-.141986c0-.05609-.0324-.08984-.0876-.08984zM118.65836 23.901795l.0149-.03111c-.0254-.01709-.0618-.02848-.0973-.02848-.0635 0-.10605.03068-.10605.07888 0 .109995.17266.06661.17222.148559 0 .0298-.0276.04777-.0719.04777-.0346 0-.0736-.01578-.1021-.04295l-.0158.03024c.0289.02761.0727.04689.11745.04689.0657 0 .11087-.03287.11087-.0837.00044-.11131-.17223-.06968-.17223-.149436 0-.02629.0259-.04207.0653-.04207.0245 0 .057.0075.0846.02542zM118.84168 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .11701-.04777.11701-.117007 0-.06836-.0487-.11613-.11701-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM119.18446 23.919762v.113063c0 .04996-.0281.08502-.0758.08589-.0399 0-.064-.02454-.064-.06573v-.133222h-.0342v.141986c0 .05565.0324.08984.0863.08984.0416-.000438.0723-.01753.0876-.05346v.05215h.0342v-.230508zM119.33586 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM119.62557 23.976294l.0188-.02366c-.021-.02147-.0517-.03418-.089-.03418-.0666 0-.11438.0482-.11438.116568 0 .0688.0473.116569.11438.116569.0399 0 .0719-.0149.0925-.03988l-.0197-.02016c-.0162.01928-.0412.0298-.0714.0298-.0478 0-.082-.03506-.082-.08633 0-.05083.0342-.08589.082-.08589.028 0 .0513.01008.0688.02717zM119.78997 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g></g><g><g><path fill=\"#a586f3\" d=\"m 7808.8125,2097.1875 c -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 v 329.875 c -0.8781,10.2195 7.1804,19 17.4375,19 10.2571,0 19.0227,-8.7805 18.1446,-19 v -77.7324 l 0.4179,-6.1426 c 0,-55.1597 44.2153,-99.5625 99.375,-99.5625 55.1596,0 99.5625,44.4028 99.5625,99.5625 0.01,0.1837 0.015,0.3674 0.027,0.5508 -0.026,0.4409 -0.035,0.8826 -0.027,1.3242 v 110 c 0,9.6822 7.849,17.5312 17.5312,17.5312 9.6823,0 17.5313,-7.849 17.5313,-17.5312 v -110 c 0,-0.3245 -0.012,-0.6488 -0.031,-0.9727 0.018,-0.3004 0.029,-0.6013 0.031,-0.9023 1e-4,-74.075 -60.55,-134.4374 -134.625,-134.4375 -39.9177,1e-4 -75.8465,17.5354 -100.5,45.293 V 2115 c 0.1749,-9.8356 -7.7881,-17.8833 -17.625,-17.8125 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><path fill=\"#5ed3ee\" d=\"m 7765.8027,2098.123 c -0.038,5e-4 -0.077,0 -0.1152,0 -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 V 2361 c 0,41.9333 -26.0162,79.2671 -65.4375,93.5625 -39.4213,14.2954 -83.5594,2.3741 -110.4375,-29.8125 -26.878,-32.1865 -30.6579,-77.5719 -9.5625,-113.8125 21.0955,-36.2406 62.4262,-55.2897 103.6875,-47.8125 9.5173,1.7397 18.6463,-4.5547 20.4043,-14.0686 1.758,-9.514 -4.5187,-18.6551 -14.0293,-20.4314 -55.6354,-10.082 -111.8054,15.8219 -140.25,64.6875 -28.4446,48.8656 -23.3039,110.5381 12.9375,153.9375 36.2415,43.3993 95.908,59.4005 149.0625,40.125 44.7667,-16.2337 77.0286,-54.658 86.0977,-100.1562 0,0 1.9812,-4.7686 2.5898,-9.1563 0.6086,-4.3877 0,-17.0625 0,-17.0625 v -245.0625 c 0.1769,-9.865 -7.8313,-17.9263 -17.6973,-17.8145 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><g fill=\"none\" stroke-width=\".365\" transform=\"translate(-449.265 -83.342) scale(5.91262)\"><circle cx=\"79.677\" cy=\"16.823\" r=\".645\" stroke=\"#c6ea70\"/><path stroke=\"#dff3b1\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 80.808656,18.234303 a 1.8088735,1.8088735 0 0 1 -2.309183,-0.03848 1.8088735,1.8088735 0 0 1 -0.389241,-2.276466 1.8088735,1.8088735 0 0 1 2.165211,-0.803532\"/><path stroke=\"#f2fadf\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 82.216272,17.248644 a 2.455857,2.4425189 0 0 1 -2.468226,2.016391 2.455857,2.4425189 0 0 1 -2.383743,-2.114547 2.455857,2.4425189 0 0 1 1.733108,-2.668494\"/></g></g></g></g></svg>",
                  "category": "Open Data Hub",
                  "support": "open data hub",
                  "docsLink": "http://REPLACE.THIS.URL",
                  "quickStart": "example-odhquickstart",
                  "getStartedLink": "http://REPLACE.THIS.URL",
                  "getStartedMarkDown": "# Header for Getting Started MarkDown section\nHere is where you add information about your ODHApplication in markdown format\n\n1. List item\n1. List item\n1. List item"
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhDocument",
                "metadata": {
                  "name": "example-odhdocument",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHDocument",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "type": "how-to",
                  "description": "This is an example of an ODHDocument custom resource",
                  "url": "http://REPLACE.THIS.URL",
                  "durationMinutes": 5
                }
              },
              {
                "apiVersion": "console.openshift.io/v1",
                "kind": "OdhQuickStart",
                "metadata": {
                  "name": "example-odhquickstart",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHQuickstart Name",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "durationMinutes": 5,
                  "description": "Description of an example quickstart",
                  "introduction": "### Introduction Header\nThis is the body of the introduction for an ODHQuickstart",
                  "tasks": [
                    {
                      "title": "Title of the first task",
                      "description": "### Description header for the first task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the first task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    },
                    {
                      "title": "Title of the second task",
                      "description": "### Description header for the second task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the second task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    }
                  ],
                  "conclusion": "Message for the conclusion of the ODHQuickstart",
                  "nextQuickStart": null
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v1.11.0
          createdAt: "2023-10-31T00:00:00Z"
          description: Open Data Hub Operator for deployment and management of Open
            Data Hub
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
          support: Open Data Hub
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AcceleratorProfile is the Schema for the acceleratorprofiles
            displayName: AcceleratorProfile
            kind: AcceleratorProfile
            name: acceleratorprofiles.dashboard.opendatahub.io
            version: v1alpha
          - description: KfDef is the Schema for the kfdefs API
            displayName: Kf Def
            kind: KfDef
            name: kfdefs.kfdef.apps.kubeflow.org
            version: v1
          - description: OdhApplication is the Schema for the odhapplications
            displayName: Open Data Hub Dashboard Application
            kind: OdhApplication
            name: odhapplications.dashboard.opendatahub.io
            version: v1
          - description: OdhDashboardConfig is the Schema for the odhdashboardconfigs
            displayName: Open Data Hub Dashboard Configuration
            kind: OdhDashboardConfig
            name: odhdashboardconfigs.opendatahub.io
            version: v1alpha
          - description: OdhDocument is the Schema for the odhdocuments
            displayName: Open Data Hub Dashboard Documentation
            kind: OdhDocument
            name: odhdocuments.dashboard.opendatahub.io
            version: v1
          - description: Extension for guiding user through various workflows in the
              Open Data Hub dashboard.
            displayName: Open Data Hub Dashboard QuickStart
            kind: OdhQuickStart
            name: odhquickstarts.console.openshift.io
            version: v1
        description: |
          The Open Data Hub is a machine-learning-as-a-service platform built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open Data Hub integrates multiple AI/ML open source components into one operator that can easily be downloaded and installed by OpenShift users.

          Open Data Hub operator allows users to install and manage components of the Open Data Hub. Users can mix and match tools from each project to fulfill the needs of their use case. Each of the
          projects share some components, but can be mostly seen as an extension of each other to provide a complete solution for both novice and skilled enterprise users.

          ### Components
          * Open Data Hub Dashboard - A web dashboard that displays installed Open Data Hub components with easy access to component UIs and documentation
          * ODH Notebook Controller - Secure management of Jupyter Notebook in Kubernetes environments built on top of Kubeflow Notebook Controller with support for OAuth
          * Jupyter Notebooks - JupyterLab notebook that provide Python support for GPU workloads
          * Data Science Pipelines - Pipeline solution for end to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton
          * Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh, a general-purpose model serving management/routing layer.
          * TrustyAI - TrustyAI is an open source AI Explainability (XAI) Toolkit

          To install one or multiple of these components use the default KfDef provided with the operator.
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Open Data Hub
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/opendatahub/opendatahub-operator:v1.11.0
        version: 1.11.0
      entries:
      - name: opendatahub-operator.v1.11.0
        version: 1.11.0
      - name: opendatahub-operator.v1.10.1
        version: 1.10.1
      - name: opendatahub-operator.v1.10.0
        version: 1.10.0
      - name: opendatahub-operator.v1.9.0
        version: 1.9.0
      - name: opendatahub-operator.v1.8.0
        version: 1.8.0
      - name: opendatahub-operator.v1.7.0
        version: 1.7.0
      - name: opendatahub-operator.v1.6.0
        version: 1.6.0
      - name: opendatahub-operator.v1.5.0
        version: 1.5.0
      - name: opendatahub-operator.v1.4.2
        version: 1.4.2
      - name: opendatahub-operator.v1.4.1
        version: 1.4.1
      - name: opendatahub-operator.v1.4.0
        version: 1.4.0
      - name: opendatahub-operator.v1.3.0
        version: 1.3.0
      - name: opendatahub-operator.v1.2.0
        version: 1.2.0
      - name: opendatahub-operator.v1.1.2
        version: 1.1.2
      - name: opendatahub-operator.v1.1.1
        version: 1.1.1
      name: rolling
    - currentCSV: opendatahub-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kfdef.apps.kubeflow.org/v1",
                "kind": "KfDef",
                "metadata": {
                  "name": "opendatahub",
                  "namespace": "opendatahub"
                },
                "spec": {
                  "applications": [
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-common"
                        }
                      },
                      "name": "odh-common"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-dashboard"
                        }
                      },
                      "name": "odh-dashboard"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "prometheus/cluster"
                        }
                      },
                      "name": "prometheus-cluster"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "prometheus/operator"
                        }
                      },
                      "name": "prometheus-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "grafana/cluster"
                        }
                      },
                      "name": "grafana-cluster"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "grafana/grafana"
                        }
                      },
                      "name": "grafana-instance"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-notebook-controller"
                        }
                      },
                      "name": "odh-notebook-controller"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "notebook-images"
                        }
                      },
                      "name": "notebook-images"
                    },
                    {
                      "kustomizeConfig": {
                        "overlays": [
                          "odh-model-controller"
                        ],
                        "repoRef": {
                          "name": "manifests",
                          "path": "model-mesh"
                        }
                      },
                      "name": "model-mesh"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "data-science-pipelines-operator"
                        }
                      },
                      "name": "data-science-pipelines-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "trustyai-service"
                        }
                      },
                      "name": "trustyai"
                    }
                  ],
                  "repos": [
                    {
                      "name": "manifests",
                      "uri": "https://github.com/opendatahub-io/odh-manifests/tarball/v1.5"
                    }
                  ]
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhApplication",
                "metadata": {
                  "name": "example-odhapplication",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started,Model development,Model training,Notebook environments,Package management,Data Management"
                  }
                },
                "spec": {
                  "displayName": "ODHAPPLICATION DISPLAY NAME",
                  "provider": "PROVIDER_NAME",
                  "description": "Description of your ODHApplication",
                  "kfdefApplications": [
                    "<NAME OF THE ODH APPLICATION IN KFDEF>"
                  ],
                  "route": "<APPLICATION_OCP_ROUTE_OBJECT_NAME>",
                  "img": "<svg xmlns=\"http://www.w3.org/2000/svg\" role=\"img\" viewBox=\"-1.13 -1.13 58.50 48.94\"><title>Open Data Hub logo</title><g><g><g fill=\"#4d5258\" stroke-width=\".006\"><path fill=\"#bbb\" d=\"M112.30231 22.797083c-.2034 0-.35737.143517-.35737.337407 0 .192939.15397.339307.35737.339307.20339 0 .35736-.147318.35736-.339307 0-.19294-.15397-.337407-.35736-.337407zm.002.131161c.1112 0 .20149.08934.20149.206246 0 .116904-.0903.208146-.20149.208146-.1112 0-.2053-.09124-.2053-.208146 0-.116904.0931-.206246.2053-.206246zM113.04141 22.801836h-.28228v.665308h.15017v-.190088h.13211c.16918 0 .26612-.08934.26612-.244263 0-.147319-.0969-.230957-.26612-.230957zm-.008.349762h-.12451v-.225255h.12451c.0836 0 .13116.03707.13116.111202 0 .07603-.0475.114053-.13116.114053zM113.40328 22.801836v.665308h.51703v-.124508h-.36687v-.147318h.32125V23.07081h-.32125v-.144467h.35642v-.124507zM114.03814 22.801836v.665308h.14352v-.426748l.3222.426748h.13686v-.665308h-.14257v.427698L114.175 22.801836z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/><path d=\"M115.29685 22.801836h-.28228v.665308h.27658c.21004 0 .35546-.136863.35546-.332654 0-.196741-.14351-.332654-.34976-.332654zm.004.538899h-.13591v-.412491h.12926c.11405 0 .19959.08459.19959.207196 0 .121657-.0817.205295-.19294.205295zM116.21878 23.467144h.16062l-.28418-.665308h-.15492l-.28894.665308h.15493l.0523-.12926h.30794zm-.3108-.249966l.1055-.25852.10455.25852zM116.31525 22.801836v.127359h.20244v.537949h.15017v-.537949h.2034v-.127359zM117.37527 23.467144h.16063l-.28419-.665308h-.15492l-.28893.665308h.15492l.0523-.12926h.30795zm-.31079-.249966l.1055-.25852.10454.25852zM117.83433 22.801836v.665308h.15017v-.262322h.30129v.262322h.15017v-.665308h-.15017v.277528h-.30129v-.277528zM118.72025 23.188665v-.386829h-.15017v.386829c0 .175831.11785.285132.30224.285132.18343 0 .29844-.109301.29844-.285132v-.386829h-.15017v.386829c0 .09599-.0589.153971-.14827.153971-.0893 0-.15207-.05798-.15207-.153971zM119.61059 22.801836h-.30604v.665308h.31649c.15492 0 .25092-.07033.25092-.184385 0-.08744-.0532-.152071-.13876-.168228.0713-.01806.115-.07413.115-.146368 0-.102648-.0912-.166327-.23761-.166327zm-.008.268024h-.14826v-.145418h.14826c.0561 0 .0913.02566.0913.07128 0 .04657-.0352.07413-.0913.07413zm0 .274677h-.14826v-.158723h.14826c.0713 0 .11406.02946.11406.07794 0 .05132-.0428.08079-.11406.08079z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g><g fill=\"#58595b\" stroke-width=\".01\"><path d=\"M112.19192 24.15027h.0381l-.13717-.306759h-.0364l-.1376.306759h.0372l.0342-.07844h.16696zm-.18712-.110871l.0688-.157763.0697.157763zM112.27858 23.843511v.306759h.0351v-.306759zM112.64205 23.843511h-.11525v.306759h.0351v-.09729h.0802c.0758 0 .11964-.039.11964-.10649 0-.0653-.0438-.102983-.11964-.102983zm-.001.177044h-.0789V23.87594h.0789c.0552 0 .0872.02454.0872.07143 0 .0482-.032.07318-.0872.07318zM112.81775 23.825105v.325165h.0342v-.325165zM113.1137 24.15027l-.00044-.152503c-.00044-.04908-.032-.07932-.089-.07932-.0394 0-.0644.01139-.0942.03111l.0145.0241c.0241-.01709.0478-.02673.0732-.02673.0408 0 .0618.0206.0618.05434v.01446h-.0758c-.0543.000439-.0863.02586-.0863.06617 0 .03944.0307.06968.0784.06968.0368 0 .0662-.01227.0846-.03725v.03593zm-.11131-.02717c-.0324 0-.053-.01709-.053-.04251 0-.02585.0202-.03769.0587-.03769h.0719v.02805c-.009.03287-.0386.05215-.0776.05215zM113.29915 24.108201c-.0127.0074-.025.01139-.0368.01183-.0197 0-.0307-.01139-.0307-.03944v-.127086h.0732v-.02673h-.0732v-.06048h-.0337v.06048h-.0333v.02673h.0333v.132783c0 .04558.025.06573.0587.0653.0184 0 .0359-.0061.0521-.01709zM113.44228 23.854466c.0118 0 .0245.0044.0351.01183l.014-.02629c-.014-.01052-.032-.01622-.0495-.01622-.0386-.000438-.0741.02717-.0741.07713v.02586h-.0298v.02673h.0298v.196764h.0337v-.196764h.0662v-.02673h-.0662v-.03024c0-.02717.0184-.04251.0408-.04207zM113.59236 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM113.80323 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM114.22747 23.918448c-.0443.000438-.0767.01972-.0911.06004-.01-.03813-.0394-.06004-.0828-.06004-.0416.000438-.0727.01753-.0881.05346v-.05215h-.0337v.230508h.0337v-.113501c0-.04996.0285-.08458.0758-.08589.0403 0 .064.02454.064.06617v.133221h.0346v-.113501c0-.04996.0276-.08458.0754-.08589.0399 0 .064.02454.064.06617v.133221h.0342V24.00828c0-.05609-.032-.08984-.0859-.08984zM114.63474 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0487.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11088-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM114.89934 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM115.02844 23.919762l.0863.230508h.0355l.0675-.188876.0679.188876h.0355l.0855-.230508h-.0355l-.0675.195888-.0679-.195888h-.0359l-.067.195888-.0684-.195888zM115.53009 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM115.73075 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM115.95034 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11482.116569.0386 0 .071-.01402.0929-.03813l-.0188-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18843c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM116.29474 23.825105v.144615c-.0171-.03287-.0491-.05171-.0903-.05171-.0653 0-.11087.04777-.11087.116569 0 .0688.0456.117445.11131.117445.0412 0 .0727-.01884.0898-.05127v.04952h.0342v-.325165zm-.0833.297557c-.0491 0-.0837-.03637-.0837-.08721 0-.05083.0346-.08765.0837-.08765.0486 0 .0833.03681.0833.08765 0 .05083-.0346.08721-.0833.08721zM116.65984 23.918448c-.0403 0-.0719.01884-.089.05083v-.144177h-.0342v.325165h.0342v-.04952c.0171.03199.0486.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM116.89657 24.183138l.10605-.263376h-.0351l-.0732.191944-.0767-.191944h-.0359l.0955.228756-.01.02191c-.009.02191-.0241.03331-.0421.03331-.0114 0-.0219-.0035-.0324-.01095l-.0145.02717c.014.01139.0294.01665.0486.01665.0307 0 .0548-.01841.0692-.05346zM117.29947 23.842634c-.0898 0-.16083.06793-.16083.154257 0 .08633.071.154694.16083.154694.0894 0 .16039-.06836.16039-.154694 0-.08633-.071-.154257-.16039-.154257zm0 .03243c.0684 0 .12446.0539.12446.121828 0 .06792-.0561.122703-.12446.122703-.0684 0-.12533-.05478-.12533-.122703 0-.06793.057-.121828.12533-.121828zM117.64414 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0486.05083.0898.05083.0653 0 .10999-.04777.10999-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM117.9061 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM118.19616 23.918448c-.0421.000438-.0732.01709-.0894.05215v-.05084h-.0342v.230508h.0342v-.12358c.004-.0447.032-.07494.0776-.07537.0408 0 .0653.02454.0653.06573v.133221h.0342v-.141986c0-.05609-.0324-.08984-.0876-.08984zM118.65836 23.901795l.0149-.03111c-.0254-.01709-.0618-.02848-.0973-.02848-.0635 0-.10605.03068-.10605.07888 0 .109995.17266.06661.17222.148559 0 .0298-.0276.04777-.0719.04777-.0346 0-.0736-.01578-.1021-.04295l-.0158.03024c.0289.02761.0727.04689.11745.04689.0657 0 .11087-.03287.11087-.0837.00044-.11131-.17223-.06968-.17223-.149436 0-.02629.0259-.04207.0653-.04207.0245 0 .057.0075.0846.02542zM118.84168 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .11701-.04777.11701-.117007 0-.06836-.0487-.11613-.11701-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM119.18446 23.919762v.113063c0 .04996-.0281.08502-.0758.08589-.0399 0-.064-.02454-.064-.06573v-.133222h-.0342v.141986c0 .05565.0324.08984.0863.08984.0416-.000438.0723-.01753.0876-.05346v.05215h.0342v-.230508zM119.33586 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM119.62557 23.976294l.0188-.02366c-.021-.02147-.0517-.03418-.089-.03418-.0666 0-.11438.0482-.11438.116568 0 .0688.0473.116569.11438.116569.0399 0 .0719-.0149.0925-.03988l-.0197-.02016c-.0162.01928-.0412.0298-.0714.0298-.0478 0-.082-.03506-.082-.08633 0-.05083.0342-.08589.082-.08589.028 0 .0513.01008.0688.02717zM119.78997 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g></g><g><g><path fill=\"#a586f3\" d=\"m 7808.8125,2097.1875 c -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 v 329.875 c -0.8781,10.2195 7.1804,19 17.4375,19 10.2571,0 19.0227,-8.7805 18.1446,-19 v -77.7324 l 0.4179,-6.1426 c 0,-55.1597 44.2153,-99.5625 99.375,-99.5625 55.1596,0 99.5625,44.4028 99.5625,99.5625 0.01,0.1837 0.015,0.3674 0.027,0.5508 -0.026,0.4409 -0.035,0.8826 -0.027,1.3242 v 110 c 0,9.6822 7.849,17.5312 17.5312,17.5312 9.6823,0 17.5313,-7.849 17.5313,-17.5312 v -110 c 0,-0.3245 -0.012,-0.6488 -0.031,-0.9727 0.018,-0.3004 0.029,-0.6013 0.031,-0.9023 1e-4,-74.075 -60.55,-134.4374 -134.625,-134.4375 -39.9177,1e-4 -75.8465,17.5354 -100.5,45.293 V 2115 c 0.1749,-9.8356 -7.7881,-17.8833 -17.625,-17.8125 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><path fill=\"#5ed3ee\" d=\"m 7765.8027,2098.123 c -0.038,5e-4 -0.077,0 -0.1152,0 -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 V 2361 c 0,41.9333 -26.0162,79.2671 -65.4375,93.5625 -39.4213,14.2954 -83.5594,2.3741 -110.4375,-29.8125 -26.878,-32.1865 -30.6579,-77.5719 -9.5625,-113.8125 21.0955,-36.2406 62.4262,-55.2897 103.6875,-47.8125 9.5173,1.7397 18.6463,-4.5547 20.4043,-14.0686 1.758,-9.514 -4.5187,-18.6551 -14.0293,-20.4314 -55.6354,-10.082 -111.8054,15.8219 -140.25,64.6875 -28.4446,48.8656 -23.3039,110.5381 12.9375,153.9375 36.2415,43.3993 95.908,59.4005 149.0625,40.125 44.7667,-16.2337 77.0286,-54.658 86.0977,-100.1562 0,0 1.9812,-4.7686 2.5898,-9.1563 0.6086,-4.3877 0,-17.0625 0,-17.0625 v -245.0625 c 0.1769,-9.865 -7.8313,-17.9263 -17.6973,-17.8145 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><g fill=\"none\" stroke-width=\".365\" transform=\"translate(-449.265 -83.342) scale(5.91262)\"><circle cx=\"79.677\" cy=\"16.823\" r=\".645\" stroke=\"#c6ea70\"/><path stroke=\"#dff3b1\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 80.808656,18.234303 a 1.8088735,1.8088735 0 0 1 -2.309183,-0.03848 1.8088735,1.8088735 0 0 1 -0.389241,-2.276466 1.8088735,1.8088735 0 0 1 2.165211,-0.803532\"/><path stroke=\"#f2fadf\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 82.216272,17.248644 a 2.455857,2.4425189 0 0 1 -2.468226,2.016391 2.455857,2.4425189 0 0 1 -2.383743,-2.114547 2.455857,2.4425189 0 0 1 1.733108,-2.668494\"/></g></g></g></g></svg>",
                  "category": "Open Data Hub",
                  "support": "open data hub",
                  "docsLink": "http://REPLACE.THIS.URL",
                  "quickStart": "example-odhquickstart",
                  "getStartedLink": "http://REPLACE.THIS.URL",
                  "getStartedMarkDown": "# Header for Getting Started MarkDown section\nHere is where you add information about your ODHApplication in markdown format\n\n1. List item\n1. List item\n1. List item"
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhDocument",
                "metadata": {
                  "name": "example-odhdocument",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHDocument",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "type": "how-to",
                  "description": "This is an example of an ODHDocument custom resource",
                  "url": "http://REPLACE.THIS.URL",
                  "durationMinutes": 5
                }
              },
              {
                "apiVersion": "console.openshift.io/v1",
                "kind": "OdhQuickStart",
                "metadata": {
                  "name": "example-odhquickstart",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHQuickstart Name",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "durationMinutes": 5,
                  "description": "Description of an example quickstart",
                  "introduction": "### Introduction Header\nThis is the body of the introduction for an ODHQuickstart",
                  "tasks": [
                    {
                      "title": "Title of the first task",
                      "description": "### Description header for the first task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the first task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    },
                    {
                      "title": "Title of the second task",
                      "description": "### Description header for the second task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the second task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    }
                  ],
                  "conclusion": "Message for the conclusion of the ODHQuickstart",
                  "nextQuickStart": null
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v1.5.0
          createdAt: "2022-10-18T00:00:00Z"
          description: Open Data Hub Operator for deployment and management of Open
            Data Hub
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
          support: Open Data Hub
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KfDef is the Schema for the kfdefs API
            displayName: Kf Def
            kind: KfDef
            name: kfdefs.kfdef.apps.kubeflow.org
            version: v1
          - description: OdhApplication is the Schema for the odhapplications
            displayName: Open Data Hub Dashboard Application
            kind: OdhApplication
            name: odhapplications.dashboard.opendatahub.io
            version: v1
          - description: OdhDashboardConfig is the Schema for the odhdashboardconfigs
            displayName: Open Data Hub Dashboard Configuration
            kind: OdhDashboardConfig
            name: odhdashboardconfigs.opendatahub.io
            version: v1alpha
          - description: OdhDocument is the Schema for the odhdocuments
            displayName: Open Data Hub Dashboard Documentation
            kind: OdhDocument
            name: odhdocuments.dashboard.opendatahub.io
            version: v1
          - description: Extension for guiding user through various workflows in the
              Open Data Hub dashboard.
            displayName: Open Data Hub Dashboard QuickStart
            kind: OdhQuickStart
            name: odhquickstarts.console.openshift.io
            version: v1
        description: |
          The Open Data Hub is a machine-learning-as-a-service platform built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open Data Hub integrates multiple AI/ML open source components into one operator that can easily be downloaded and installed by OpenShift users.

          Open Data Hub operator allows users to install and manage components of the Open Data Hub. Users can mix and match tools from each project to fulfill the needs of their use case. Each of the
          projects share some components, but can be mostly seen as an extension of each other to provide a complete solution for both novice and skilled enterprise users.

          ### Components
          * Open Data Hub Dashboard - A web dashboard that displays installed Open Data Hub components with easy access to component UIs and documentation
          * ODH Notebook Controller - Secure management of Jupyter Notebook in Kubernetes environments built on top of Kubeflow Notebook Controller with support for OAuth
          * Jupyter Notebooks - JupyterLab notebook that provide Python support for GPU workloads
          * Data Science Pipelines - Pipeline solution for end to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton
          * Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh, a general-purpose model serving management/routing layer.

          To install one or multiple of these components use the default KfDef provided with the operator.

          ### Available Channels

          #### Stable

          Channel `stable` offers the major releases of Open Data Hub operator and ODH manifests. It is based on Kubeflow Operator and Kustomize for deployment configuration. This version is compatible with [Open Data Hub manifests](https://github.com/opendatahub-io/odh-manifests/) as well as [Kubeflow manifests](https://github.com/opendatahub-io/manifests). This channel will offer new components and features that have been thoroughly tested by consumers of the `rolling` channel

          NOTE: The `stable` channel will be deprecated starting with the ODH 1.6 release and we will only be publishing updates on the `rolling` channel for all future releases

          #### Rolling

          Channel `rolling` will offer the latest release of Open Data Hub operator and manifests. Subscribing to this branch will provide faster updates and access to components that will be in developement and testing for the next stable release
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io
        maintainers:
        - email: maintainers@lists.opendatahub.io
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Open Data Hub
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/opendatahub/opendatahub-operator:v1.5.0
        version: 1.5.0
      entries:
      - name: opendatahub-operator.v1.5.0
        version: 1.5.0
      - name: opendatahub-operator.v1.4.2
        version: 1.4.2
      - name: opendatahub-operator.v1.4.1
        version: 1.4.1
      - name: opendatahub-operator.v1.4.0
        version: 1.4.0
      - name: opendatahub-operator.v1.3.0
        version: 1.3.0
      - name: opendatahub-operator.v1.2.0
        version: 1.2.0
      - name: opendatahub-operator.v1.1.2
        version: 1.1.2
      - name: opendatahub-operator.v1.1.1
        version: 1.1.1
      name: stable
    defaultChannel: fast
    packageName: opendatahub-operator
    provider:
      name: ODH
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Thales Group
      provider-url: https://cpl.thalesgroup.com/data-protection
    name: cte-k8s-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cte-k8s-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cte-k8s-operator.csi.cte.cpl.thalesgroup.com/v1",
                "kind": "CteK8sOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cte-k8s-operator"
                  },
                  "name": "ctek8soperator-sample"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-06-05T08:24:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CteK8sOperator is the Schema for the ctek8soperators API
            displayName: Cte K8s Operator
            kind: CteK8sOperator
            name: ctek8soperators.cte-k8s-operator.csi.cte.cpl.thalesgroup.com
            version: v1
        description: |-
          # Kubernetes Operator to deploy CTE for Kubernetes on an OCP/K8s Cluster.
          This Operator deploys CTE for Kubernetes on the OCP/K8s Cluster.

          ## About this operator

          It is a GO based operator that manages the deployment, upgrade and
          deletion of the CTE for Kubernetes on all the nodes of the Cluster on
          which it is installed.

          ## Prerequisites

          Applicable only on OCP cluster
            For an OCP cluster the operator requires a SecurityContextConstraints (SCC)
            to be created before installation. To create the SecurityContextConstraints
            you can use a definition provided by the deployment scripts at
            https://github.com/thalescpl-io/ciphertrust-transparent-encryption-kubernetes

            Look for cte-csi-scc.yaml file in deploy/kubernetes/X.Y.Z/operator-deploy/
            Where X.Y.Z is the last released version of the deploy scripts.

          For more information refer to the guide at:
            https://thalesdocs.com/ctp/cte-con/cte-k8s/latest/cte-4-k8s/index.html
        displayName: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Encryption
        - Data Protection
        - Security
        - Kubernetes
        - Openshift
        links:
        - name: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          url: https://cpl.thalesgroup.com/encryption/ciphertrust-transparent-data-encryption-kubernetes
        maintainers:
        - email: technical.support.DIS@thalesgroup.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Thales Group
          url: https://cpl.thalesgroup.com/data-protection
        relatedImages:
        - docker.io/thalesciphertrust/ciphertrust-transparent-encryption-kubernetes-operator@sha256:d58cbb65d758fd0b9d4cfba4cb48a35bdaebe2e239fb25edb9b491cd8c2c3f73
        version: 1.6.5
      entries:
      - name: cte-k8s-operator.v1.6.5
        version: 1.6.5
      name: candidate
    - currentCSV: cte-k8s-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cte-k8s-operator.csi.cte.cpl.thalesgroup.com/v1",
                "kind": "CteK8sOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cte-k8s-operator"
                  },
                  "name": "ctek8soperator-sample"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-06-05T08:24:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CteK8sOperator is the Schema for the ctek8soperators API
            displayName: Cte K8s Operator
            kind: CteK8sOperator
            name: ctek8soperators.cte-k8s-operator.csi.cte.cpl.thalesgroup.com
            version: v1
        description: |-
          # Kubernetes Operator to deploy CTE for Kubernetes on an OCP/K8s Cluster.
          This Operator deploys CTE for Kubernetes on the OCP/K8s Cluster.

          ## About this operator

          It is a GO based operator that manages the deployment, upgrade and
          deletion of the CTE for Kubernetes on all the nodes of the Cluster on
          which it is installed.

          ## Prerequisites

          Applicable only on OCP cluster
            For an OCP cluster the operator requires a SecurityContextConstraints (SCC)
            to be created before installation. To create the SecurityContextConstraints
            you can use a definition provided by the deployment scripts at
            https://github.com/thalescpl-io/ciphertrust-transparent-encryption-kubernetes

            Look for cte-csi-scc.yaml file in deploy/kubernetes/X.Y.Z/operator-deploy/
            Where X.Y.Z is the last released version of the deploy scripts.

          For more information refer to the guide at:
            https://thalesdocs.com/ctp/cte-con/cte-k8s/latest/cte-4-k8s/index.html
        displayName: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Encryption
        - Data Protection
        - Security
        - Kubernetes
        - Openshift
        links:
        - name: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          url: https://cpl.thalesgroup.com/encryption/ciphertrust-transparent-data-encryption-kubernetes
        maintainers:
        - email: technical.support.DIS@thalesgroup.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Thales Group
          url: https://cpl.thalesgroup.com/data-protection
        relatedImages:
        - docker.io/thalesciphertrust/ciphertrust-transparent-encryption-kubernetes-operator@sha256:d58cbb65d758fd0b9d4cfba4cb48a35bdaebe2e239fb25edb9b491cd8c2c3f73
        version: 1.6.5
      entries:
      - name: cte-k8s-operator.v1.6.5
        version: 1.6.5
      name: stable
    defaultChannel: stable
    packageName: cte-k8s-operator
    provider:
      name: Thales Group
      url: https://cpl.thalesgroup.com/data-protection
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-team-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-team-operator.v2024.10.2-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": null,
                        "host": null,
                        "process": null
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master",
                        "operator": "",
                        "value": ""
                      }
                    ]
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": null
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:12c1d2e9347ec447196a618b481a0970dacac7a8a72a3b38ec22efa69668d180
          createdAt: "2024-10-25T01:47:52Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2024.10.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2023.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                   operator: ""
                   value: ""

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:12c1d2e9347ec447196a618b481a0970dacac7a8a72a3b38ec22efa69668d180
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:7c4868db021beaffa762eb0de43739bbc865a4bf5900f1d17f78caed8612c073
        version: 2024.10.2-12
      entries:
      - name: uma-team-operator.v2024.10.2-12
        version: 2024.10.2-12
      name: "2024.10"
    - currentCSV: uma-team-operator.v2025.1.3-33
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": null,
                        "host": null,
                        "process": null
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.1.1.33,990033",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master",
                        "operator": "",
                        "value": ""
                      }
                    ]
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": null
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:78b66a291f58aa94123cdd59f2aa07dce4928162c481fe9de5ce524782bc93de
          createdAt: "2025-02-18T05:35:06Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.1.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2023.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.1.1.26,990026"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                   operator: ""
                   value: ""

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:ae91a28d57c8a359f81757e4bf79472f2ac189ac0ae3ee0d87e8ff9c433f1ccc
        - registry.connect.redhat.com/ca/uma-operator@sha256:78b66a291f58aa94123cdd59f2aa07dce4928162c481fe9de5ce524782bc93de
        version: 2025.1.3-33
      entries:
      - name: uma-team-operator.v2025.1.3-33
        version: 2025.1.3-33
      name: "2025.1"
    - currentCSV: uma-team-operator.v2025.2.1-25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.2.1.25,990025",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master",
                        "operator": "",
                        "value": ""
                      }
                    ]
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:d0a0090a871ff7e23d3c97b620a1b0d004a8317efe1581ba0437d073c5656825
          createdAt: "2025-03-05T05:51:44Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.2.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2023.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.1.1.26,990026"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                   operator: ""
                   value: ""

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:d0a0090a871ff7e23d3c97b620a1b0d004a8317efe1581ba0437d073c5656825
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:8bcf0dac889f38a0bf77eb9babe3f726e3b2564d5988c05bdc5c2d91ae189c61
        version: 2025.2.1-25
      entries:
      - name: uma-team-operator.v2025.2.1-25
        version: 2025.2.1-25
      name: "2025.2"
    - currentCSV: uma-team-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:60fb1f76b4d4aab569eec5f2c20f7a06d813759f5fb6ac952156b41c4d295a52
          createdAt: "2025-04-16T06:57:55Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.4.5
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.4.2.5,990005"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:60fb1f76b4d4aab569eec5f2c20f7a06d813759f5fb6ac952156b41c4d295a52
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        version: 2025.4.5-5
      entries:
      - name: uma-team-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-team-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:052055a6280f60a69caccf89f50fd8bf92da61cd59f1962f61d21467c4191970
          createdAt: "2025-07-24T14:54:58Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.5.2.12,990012"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:052055a6280f60a69caccf89f50fd8bf92da61cd59f1962f61d21467c4191970
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        version: 2025.5.4-12
      entries:
      - name: uma-team-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-team-operator.v2025.7.3-21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:e3ae2240aeb618f36f41073a5f65b03a7ad5dfdccc7f9feb830506fbd2bfd604
          createdAt: "2025-08-23T09:01:14Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.7.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.7.1.23,990023"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:e3ae2240aeb618f36f41073a5f65b03a7ad5dfdccc7f9feb830506fbd2bfd604
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.3-21
      entries:
      - name: uma-team-operator.v2025.7.3-21
        version: 2025.7.3-21
      name: "2025.7"
    - currentCSV: uma-team-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "hostmonitoring": {
                        "linux": {
                          "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:b8422d0f3a9e576948ced8244a91abc3283b26217e4c7efc7e2669ca5508930e
          createdAt: "2025-08-15T03:20:57Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 hostmonitoring:
                   linux:
                   # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                   groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.8.1.9,990009"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:b8422d0f3a9e576948ced8244a91abc3283b26217e4c7efc7e2669ca5508930e
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        version: 2025.8.2-9
      entries:
      - name: uma-team-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-team-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
          createdAt: "2025-09-24T12:24:28Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.12,990012"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-team-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-team-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "hostmonitoring": {
                        "linux": {
                          "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:8c12315745ad3564035a4db923698bf63fe32dddbc7d246f5a3867945305713e
          createdAt: "2025-07-30T08:27:50Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 hostmonitoring:
                   linux:
                   # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                   groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "99.99.0.serp_F152813,19"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:8c12315745ad3564035a4db923698bf63fe32dddbc7d246f5a3867945305713e
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        version: 2025.7.2-19
      entries:
      - name: uma-team-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-team-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:89e0505b5e526beffda9b13a1ca7b49e5ba144fd0166a7e171e9d9b084915a56
          createdAt: "2025-09-11T11:16:37Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.8,990008"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:89e0505b5e526beffda9b13a1ca7b49e5ba144fd0166a7e171e9d9b084915a56
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-team-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-team-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
          createdAt: "2025-09-24T12:24:28Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.12,990012"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-team-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-team-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ocs-client-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocs-client-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          console.openshift.io/plugins: '["odf-client-console"]'
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageClassClaim is the Schema for the storageclassclaims
              API
            displayName: Storage Class Claim
            kind: StorageClassClaim
            name: storageclassclaims.ocs.openshift.io
            version: v1alpha1
          - description: StorageClient is the Schema for the storageclients API
            displayName: Storage Client
            kind: StorageClient
            name: storageclients.ocs.openshift.io
            version: v1alpha1
        description: This operator is not intended to be directly installed
        displayName: OpenShift Data Foundation Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: OCS Client Operator
          url: https://github.com/red-hat-storage/ocs-client-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:975529d3fbdcf6f78cc1602333832f543da2918ac5c018d1ee6bea2b315106be
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:ef4276b80cacae825d4831de88d3fe6ba7cc324464087b7b28734bd2b8de3315
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel8@sha256:d83b44a1257e8015409cd261e8eed5bd8c4e581b788e8452f19114ba696a30c1
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:bc637bbbed3457c45fbb5cb1e8940b2a1dacfa64959e5bd0bfb58c5d669427fd
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:1dd7c0ef17d39ebf3311aee326202e09038cece1b7554cea23c1efe582620d0b
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:c9e5b48d84febf10ff4c9b78a8210df36c90905fbb92b6a6b5b32e0fd9a6a41d
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:9a0008f662b6afa9e2677a8a83bd662d806ad6f4d03fb49bccb78ed487ce4c7a
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:028120f2212738917f49e0330b60be5f79a208ce2974bea22b71f533aa6ce3be
        - registry.redhat.io/odf4/ocs-client-console-rhel9@sha256:ebf3f6098a9a0a06706d20fe8fc86513be34b98604bc70fd82910f0fcd17f191
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:ffcc0806ec2afb8b40d75c70ef2a6d98bb887879f5b3f573e900861316ce9e41
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel8@sha256:05cf61192ee75dc11f5777bebeacdd209232042445553329e14f542ac0084c71
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:e206dcc33863f4d8d0aa6df173a86cd50ee1d52e6bcf9efc43b39831acc20573
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:0dc0733df67bb49886ac9715b81bd4ec4a36330fe1afdde657587e1883f5ca7a
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel8@sha256:647074ada8cb88f7ed99ca14621d3b754db6a60346f12b7561cd035c05f283f7
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:390b527d10e9c7a8a9191251b0a1ba501128057cefaff35e950f0efe04ef3122
        - registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:cb93b68adb857c82f23f292cdee47d52656dc4fc5bb83fac8581110869cbe841
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:645d4fd09628f7ddcfc765e4ddacaa79fa6a7c35c03292ca2f6e317d48765d6b
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel8@sha256:1f6a17909f1fe6e09a01cad38b6a50b41d850b1c5cffd24ef9a87c8798ff15e7
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:1711a0aabd28937aac1424adebe96619a48b91f7f888f0588117f80891001da4
        version: 4.15.20-rhodf
      entries:
      - name: ocs-client-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: ocs-client-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: ocs-client-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: ocs-client-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: ocs-client-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: ocs-client-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: ocs-client-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: ocs-client-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: ocs-client-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: ocs-client-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: ocs-client-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: ocs-client-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: ocs-client-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: ocs-client-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: ocs-client-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: ocs-client-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: ocs-client-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: ocs-client-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: ocs-client-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: ocs-client-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: ocs-client-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-client-console"]'
          containerImage: registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:bd8454b5a0b235669732650ae8502831811793d93fbfbd0d0f53278fcdfed459
          createdAt: "2025-09-12T06:29:47Z"
          description: OpenShift Data Foundation client operator enables consumption
            of storage services from a remote centralized OpenShift Data Foundation
            provider cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["storageclaims.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/ocs-client-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageClaim is the Schema for the storageclaims API
            displayName: Storage Claim
            kind: StorageClaim
            name: storageclaims.ocs.openshift.io
            version: v1alpha1
          - description: StorageClient is the Schema for the storageclients API
            displayName: Storage Client
            kind: StorageClient
            name: storageclients.ocs.openshift.io
            version: v1alpha1
        description: OpenShift Data Foundation client operator enables consumption
          of storage services from a remote centralized OpenShift Data Foundation
          provider cluster.
        displayName: OpenShift Data Foundation Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: OCS Client Operator
          url: https://github.com/red-hat-storage/ocs-client-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:80c01633a6bfca44df37ac38f764035be6998fce965c8709e4f6d54d17eeaa7b
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:82a4b481028195cfd75471656fe40c1cbbade4c7b2d170d6bf9422d9d3b84fa7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel8@sha256:d83b44a1257e8015409cd261e8eed5bd8c4e581b788e8452f19114ba696a30c1
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel8@sha256:1f6a17909f1fe6e09a01cad38b6a50b41d850b1c5cffd24ef9a87c8798ff15e7
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:975529d3fbdcf6f78cc1602333832f543da2918ac5c018d1ee6bea2b315106be
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c11a599bfd09b9b22888e03eabdbe1054d4049718a591d2a45bf0ed89800111e
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6fee56851da3a87b9f7f477fc59e3385e118dc0b398861a8f13459c998a5af77
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:0afcb516eb528857887c6ee1a9cba081dc8f6b7d47d9f26a0dbb2dca0cefe07c
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:8d8326c74795c9c6bd8f33a9f9e81d2304a70b1f9be3c2f8d50e76be4b0292ff
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/odf4/ocs-client-console-rhel9@sha256:beb4f82d3923171b80aacb90dfe345a22626453bf2f3921b03a16839b204d377
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:1711a0aabd28937aac1424adebe96619a48b91f7f888f0588117f80891001da4
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:ef4276b80cacae825d4831de88d3fe6ba7cc324464087b7b28734bd2b8de3315
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:028120f2212738917f49e0330b60be5f79a208ce2974bea22b71f533aa6ce3be
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:645d4fd09628f7ddcfc765e4ddacaa79fa6a7c35c03292ca2f6e317d48765d6b
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:bc637bbbed3457c45fbb5cb1e8940b2a1dacfa64959e5bd0bfb58c5d669427fd
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:ffcc0806ec2afb8b40d75c70ef2a6d98bb887879f5b3f573e900861316ce9e41
        - registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:bd8454b5a0b235669732650ae8502831811793d93fbfbd0d0f53278fcdfed459
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:693a60d6a1162659213ae554b1c6ee200b0300a7da72c01415044cf507acab49
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:c9e5b48d84febf10ff4c9b78a8210df36c90905fbb92b6a6b5b32e0fd9a6a41d
        version: 4.16.16-rhodf
      entries:
      - name: ocs-client-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: ocs-client-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: ocs-client-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: ocs-client-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: ocs-client-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: ocs-client-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: ocs-client-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: ocs-client-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: ocs-client-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: ocs-client-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: ocs-client-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: ocs-client-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: ocs-client-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: ocs-client-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: ocs-client-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: ocs-client-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: ocs-client-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: ocs-client-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://your.domain
    name: amq-streams-proxy
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: alpha
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: amq-streams-3.0.x
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: amq-streams-3.x
    defaultChannel: alpha
    packageName: amq-streams-proxy
    provider:
      name: Red Hat
      url: https://your.domain
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-iam-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-iam-controller.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Role",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Group",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "OpenIDConnectProvider",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Policy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/iam-controller:1.5.2
          createdAt: "2025-10-03T22:22:57Z"
          description: AWS IAM controller is a service controller for managing IAM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Group represents the state of an AWS iam Group resource.
            displayName: Group
            kind: Group
            name: groups.iam.services.k8s.aws
            version: v1alpha1
          - description: InstanceProfile represents the state of an AWS iam InstanceProfile
              resource.
            displayName: InstanceProfile
            kind: InstanceProfile
            name: instanceprofiles.iam.services.k8s.aws
            version: v1alpha1
          - description: OpenIDConnectProvider represents the state of an AWS iam
              OpenIDConnectProvider resource.
            displayName: OpenIDConnectProvider
            kind: OpenIDConnectProvider
            name: openidconnectproviders.iam.services.k8s.aws
            version: v1alpha1
          - description: Policy represents the state of an AWS iam Policy resource.
            displayName: Policy
            kind: Policy
            name: policies.iam.services.k8s.aws
            version: v1alpha1
          - description: Role represents the state of an AWS iam Role resource.
            displayName: Role
            kind: Role
            name: roles.iam.services.k8s.aws
            version: v1alpha1
          - description: ServiceLinkedRole represents the state of an AWS iam ServiceLinkedRole
              resource.
            displayName: ServiceLinkedRole
            kind: ServiceLinkedRole
            name: servicelinkedroles.iam.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS iam User resource.
            displayName: User
            kind: User
            name: users.iam.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Identity and Access Management (IAM) resources in AWS from within your Kubernetes cluster.

          **About Amazon IAM**

          AWS Identity and Access Management (IAM) is a web service that helps you securely control access to AWS resources. You use IAM to control who is authenticated (signed in) and authorized (has permissions) to use resources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon IAM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iam
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon IAM Developer Resources
          url: https://aws.amazon.com/iam/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: iam maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/iam-controller:1.5.2
        version: 1.5.2
      entries:
      - name: ack-iam-controller.v1.5.2
        version: 1.5.2
      - name: ack-iam-controller.v1.5.1
        version: 1.5.1
      - name: ack-iam-controller.v1.5.0
        version: 1.5.0
      - name: ack-iam-controller.v1.4.4
        version: 1.4.4
      - name: ack-iam-controller.v1.4.3
        version: 1.4.3
      - name: ack-iam-controller.v1.4.2
        version: 1.4.2
      - name: ack-iam-controller.v1.4.1
        version: 1.4.1
      - name: ack-iam-controller.v1.4.0
        version: 1.4.0
      - name: ack-iam-controller.v1.3.22
        version: 1.3.22
      - name: ack-iam-controller.v1.3.21
        version: 1.3.21
      - name: ack-iam-controller.v1.3.20
        version: 1.3.20
      - name: ack-iam-controller.v1.3.19
        version: 1.3.19
      - name: ack-iam-controller.v1.3.18
        version: 1.3.18
      - name: ack-iam-controller.v1.3.17
        version: 1.3.17
      - name: ack-iam-controller.v1.3.15
        version: 1.3.15
      - name: ack-iam-controller.v1.3.14
        version: 1.3.14
      - name: ack-iam-controller.v1.3.13
        version: 1.3.13
      - name: ack-iam-controller.v1.3.12
        version: 1.3.12
      - name: ack-iam-controller.v1.3.11
        version: 1.3.11
      - name: ack-iam-controller.v1.3.10
        version: 1.3.10
      - name: ack-iam-controller.v1.3.9
        version: 1.3.9
      - name: ack-iam-controller.v1.3.8
        version: 1.3.8
      - name: ack-iam-controller.v1.3.6
        version: 1.3.6
      - name: ack-iam-controller.v1.3.2
        version: 1.3.2
      - name: ack-iam-controller.v1.3.1
        version: 1.3.1
      - name: ack-iam-controller.v1.3.0
        version: 1.3.0
      - name: ack-iam-controller.v1.2.6
        version: 1.2.6
      - name: ack-iam-controller.v1.2.5
        version: 1.2.5
      - name: ack-iam-controller.v1.2.3
        version: 1.2.3
      - name: ack-iam-controller.v1.2.2
        version: 1.2.2
      - name: ack-iam-controller.v1.2.1
        version: 1.2.1
      - name: ack-iam-controller.v1.2.0
        version: 1.2.0
      - name: ack-iam-controller.v1.1.3
        version: 1.1.3
      - name: ack-iam-controller.v1.1.2
        version: 1.1.2
      - name: ack-iam-controller.v1.1.1
        version: 1.1.1
      - name: ack-iam-controller.v1.1.0
        version: 1.1.0
      - name: ack-iam-controller.v1.0.0
        version: 1.0.0
      - name: ack-iam-controller.v0.1.2
        version: 0.1.2
      - name: ack-iam-controller.v0.1.1
        version: 0.1.1
      - name: ack-iam-controller.v0.1.0
        version: 0.1.0
      - name: ack-iam-controller.v0.0.22
        version: 0.0.22
      - name: ack-iam-controller.v0.0.21
        version: 0.0.21
      - name: ack-iam-controller.v0.0.20
        version: 0.0.20
      - name: ack-iam-controller.v0.0.19
        version: 0.0.19
      - name: ack-iam-controller.v0.0.18
        version: 0.0.18
      - name: ack-iam-controller.v0.0.17
        version: 0.0.17
      - name: ack-iam-controller.v0.0.16
        version: 0.0.16
      - name: ack-iam-controller.v0.0.15
        version: 0.0.15
      - name: ack-iam-controller.v0.0.14
        version: 0.0.14
      - name: ack-iam-controller.v0.0.13
        version: 0.0.13
      - name: ack-iam-controller.v0.0.12
        version: 0.0.12
      - name: ack-iam-controller.v0.0.11
        version: 0.0.11
      - name: ack-iam-controller.v0.0.10
        version: 0.0.10
      - name: ack-iam-controller.v0.0.9
        version: 0.0.9
      - name: ack-iam-controller.v0.0.8
        version: 0.0.8
      - name: ack-iam-controller.v0.0.7
        version: 0.0.7
      - name: ack-iam-controller.v0.0.6
        version: 0.0.6
      - name: ack-iam-controller.v0.0.5
        version: 0.0.5
      - name: ack-iam-controller.v0.0.4
        version: 0.0.4
      - name: ack-iam-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-iam-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: crunchy-postgres-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.connect.redhat.com/crunchydata/crunchy-pgbackrest@sha256:b70da0035b74379f2507cfabdd9d1cc1e3c3eecb9066c22f292981c9a27916e1
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:0e283616c2d32e53258d62d02702bc4e179ef8c03933592154babed1a70bf666
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:c559c896e174674d151f93543ac92a1b5647eb4676313f7dc6cbb8faee786f18
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:56a5e3a5758982e086c9f78201a7137e0d934316f058a31cf30df44639a569a2
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:a7f62f367817fc08294a0b8b82ac9826a42c0ae668efdfacaa158a4287b8a688
        - registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
        - registry.connect.redhat.com/crunchydata/crunchy-pgadmin4@sha256:0d84acb1ce6a9b046e3f8b222a583ddcc91dda0a117c9f637ef173d0ac992f57
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:b2c3c74fdcc0e08f555e50c527f0d322292333c5d3dce061a0dee3680a1fbc56
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-exporter@sha256:3b231902b2116cd38de05c7378202716c95813b9961fb8cee811b61d36cddf86
        - registry.connect.redhat.com/crunchydata/crunchy-upgrade@sha256:68bc93cce61d9b1e44822ae3e3ac8f7a9c2b712eddfaea090eabefacd60c67f6
        - registry.connect.redhat.com/crunchydata/crunchy-pgbouncer@sha256:17cb58dec78447c2f70a414a2aeff6facbf921b42a0f366d0dbf7be95aa8987a
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:4bdd0588ffbb3f628e7df1c72670e7a5087269438343b68779012bc3a13dea31
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.4
        version: 5.4.4
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.4.0
        version: 5.4.0
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: crunchy-postgres-operator
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hbase-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hbase-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:e56051b66b0f419874d113c0320232dba7dd2547357256cf80eeaf18bbecb298
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hbase-operator@sha256:e56051b66b0f419874d113c0320232dba7dd2547357256cf80eeaf18bbecb298
        version: 24.11.1
      entries:
      - name: hbase-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: hbase-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:232ff30ffdf4bdc33f49d87950864e15436957510ba47f352dc0b133cb43279d
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hbase-operator@sha256:232ff30ffdf4bdc33f49d87950864e15436957510ba47f352dc0b133cb43279d
        version: 25.3.0
      entries:
      - name: hbase-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: hbase-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
        version: 25.7.0
      entries:
      - name: hbase-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hbase-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
        version: 25.7.0
      entries:
      - name: hbase-operator.v25.7.0
        version: 25.7.0
      - name: hbase-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hbase-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-secretsmanager-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-secretsmanager-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secretsmanager.services.k8s.aws/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/secretsmanager-controller:1.1.1
          createdAt: "2025-09-19T21:14:43Z"
          description: AWS Secrets Manager controller is a service controller for
            managing Secrets Manager resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret represents the state of an AWS secretsmanager Secret
              resource.
            displayName: Secret
            kind: Secret
            name: secrets.secretsmanager.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Secrets Manager resources in AWS from within your Kubernetes cluster.

          **About Amazon Secrets Manager**

          AWS Secrets Manager helps you manage, retrieve, and rotate database credentials, application credentials, OAuth tokens, API keys, and other secrets throughout their lifecycles. [Many AWS services that use secrets](https://docs.aws.amazon.com/secretsmanager/latest/userguide/integrating.html) store them in Secrets Manager.

          Secrets Manager helps you improve your security posture, because you no longer need hard-coded credentials in application source code. Storing the credentials in Secrets Manager helps avoid possible compromise by anyone who can inspect your application or the components. You replace hard-coded credentials with a runtime call to the Secrets Manager service to retrieve credentials dynamically when you need them.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Secrets Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secretsmanager
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Secrets Manager Developer Resources
          url: https://aws.amazon.com/secrets-manager/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: secretsmanager maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/secretsmanager-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-secretsmanager-controller.v1.1.1
        version: 1.1.1
      - name: ack-secretsmanager-controller.v1.1.0
        version: 1.1.0
      - name: ack-secretsmanager-controller.v1.0.14
        version: 1.0.14
      - name: ack-secretsmanager-controller.v1.0.13
        version: 1.0.13
      - name: ack-secretsmanager-controller.v1.0.12
        version: 1.0.12
      - name: ack-secretsmanager-controller.v1.0.11
        version: 1.0.11
      - name: ack-secretsmanager-controller.v1.0.10
        version: 1.0.10
      - name: ack-secretsmanager-controller.v1.0.9
        version: 1.0.9
      - name: ack-secretsmanager-controller.v1.0.8
        version: 1.0.8
      - name: ack-secretsmanager-controller.v1.0.7
        version: 1.0.7
      - name: ack-secretsmanager-controller.v1.0.6
        version: 1.0.6
      - name: ack-secretsmanager-controller.v1.0.5
        version: 1.0.5
      - name: ack-secretsmanager-controller.v1.0.4
        version: 1.0.4
      - name: ack-secretsmanager-controller.v1.0.3
        version: 1.0.3
      - name: ack-secretsmanager-controller.v1.0.1
        version: 1.0.1
      - name: ack-secretsmanager-controller.v1.0.0
        version: 1.0.0
      - name: ack-secretsmanager-controller.v0.0.11
        version: 0.0.11
      - name: ack-secretsmanager-controller.v0.0.10
        version: 0.0.10
      - name: ack-secretsmanager-controller.v0.0.9
        version: 0.0.9
      - name: ack-secretsmanager-controller.v0.0.8
        version: 0.0.8
      - name: ack-secretsmanager-controller.v0.0.7
        version: 0.0.7
      - name: ack-secretsmanager-controller.v0.0.2
        version: 0.0.2
      - name: ack-secretsmanager-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-secretsmanager-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Accuknox Inc.
      provider-url: https://accuknox.com
    name: accuknox-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: accuknox-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.accuknox.com/v1",
                "kind": "AccuknoxConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "accuknox-operator",
                    "app.kubernetes.io/instance": "accuknoxconfig",
                    "app.kubernetes.io/name": "accuknoxconfig-sample",
                    "app.kubernetes.io/part-of": "accuknox-operator"
                  },
                  "name": "accuknoxconfig-sample",
                  "namespace": "accuknox-agents"
                },
                "spec": {
                  "ppsHost": "PLEASE_FILL_IN",
                  "knoxGateway": "PLEASE_FILL_IN",
                  "spireHost": "PLEASE_FILL_IN"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Security
          containerImage: public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:b6fae25e229fb7fca757872fc0460642d52c4f9906f776afac50c54a69b05b68
          createdAt: "2024-02-09T17:06:30Z"
          operatorframework.io/suggested-namespace: accuknox-agents
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: public.ecr.aws/k9v9d5v2/accuknox-operator
          support: Accuknox Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccuknoxConfig is the Schema for the accuknoxconfigs API
            kind: AccuknoxConfig
            name: accuknoxconfigs.operator.accuknox.com
            version: v1
        description: |2-

          The AccuKnox Operator brings native K8s support for deploying and managing AccuKnox Components on a Kubernetes cluster.
          The AccuKnox Operator helps to easily onboard a k8s cluster to the AccuKnox SaaS.

          AccuKnox Operator deploys and manages the components:

          1. Discovery Engine

          2. Policy Enforcement Agent

          3. Shared Informer Agent

          4. Feeder Service

          5. Agents Operator

          ## Prerequisites for using this Operator

          1. Get AccuKnox licence - https://app.accuknox.com/sign-up

          2. After Successful Sign-up into AccuKnox Application.

             Please naviagate to Settings >> Manager Cluster >> Onboard Cluster >>

             After successful onboarding, Copy the join token created by AccuKnox

          3. Create a secret using jointoken by executing the below command:

          ```
          oc create secret generic accuknox-jointoken --from-literal=token=<joinToken> -n accuknox-agents
          ```

          ## Deploy AccuKnox Operator using AccuknoxConfig Custom Resource

          All the information to configure the AccuKnox Operator will be provided via Accuknox during Cluster Onboarding steps.

          ### AccuknoxConfig Spec

          | Spec | Description |
          | ---- | ----------- |
          | spireHost | "spire host url" |
          | ppsHost   |  "pps host url" |
          | knoxGateway | "knox gateway url" |
          | joinTokenSecretName | "join token secret name if it's different than accuknox-jointoken" |
          | joinTokenSecretKey | "key name used to store the jointoken with the secret" |

          *Example*

          ```
          apiVersion: operator.accuknox.com/v1
          kind: AccuknoxConfig
          metadata:
            name: accuknoxconfig-default
            namespace: accuknox-agents
          spec:
            ppsHost: <ppsHost url>
            knoxGateway: <knox gateway url>
            spireHost: <spire host url>
          ```
        displayName: Accuknox Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CNAPP
        - CSPM
        - CWPP
        - Cloud Native
        - Runtime Security
        - Zero-Trust
        maturity: alpha
        provider:
          name: Accuknox Inc.
          url: https://accuknox.com
        relatedImages:
        - public.ecr.aws/k9v9d5v2/discovery-engine-discover@sha256:052efbe1064503853adab497415b9407f51578e1456931a8a0e60353aa53e0c2
        - public.ecr.aws/k9v9d5v2/discovery-engine-sumengine@sha256:4ca78f90b28b38badb0520f148ce702a3a6cd6b02bb62e8b1a6782a5b82e0868
        - public.ecr.aws/k9v9d5v2/shared-informer-agent@sha256:6d18d13312d4309b6e32b18c063d5b26f15fbf057fda2b9e6b7e8ebb1a5885e9
        - public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:b6fae25e229fb7fca757872fc0460642d52c4f9906f776afac50c54a69b05b68
        - public.ecr.aws/k9v9d5v2/discovery-engine-offloader@sha256:3c2291310e30e35a931d2af389f7ade656ce86f609d2e42ecdabf1000764695c
        - public.ecr.aws/k9v9d5v2/policy-enforcement-agent@sha256:f004fc3572a4d66da1bc0db9e05a279d761a42d6ed62f95de7a2bb42ce71e17b
        - public.ecr.aws/k9v9d5v2/agents-operator@sha256:656d6bac4371adef3e86658011ba68ae2f98514f6ddaea0667d4269eb88dd02e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        - public.ecr.aws/k9v9d5v2/feeder-service@sha256:f5bc340d9d3e34804dd54a03701d272e0ce5c04160d1aa41984cb3d877eb5b76
        - public.ecr.aws/k9v9d5v2/discovery-engine-hardening@sha256:3eb28d44ef93e6ba331e9085e64475458e1c234017c77dc30b8bc802def434c3
        version: 0.2.1
      entries:
      - name: accuknox-operator.v0.2.1
        version: 0.2.1
      name: alpha
    - currentCSV: accuknox-operator.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.accuknox.com/v1",
                "kind": "AccuknoxConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "accuknox-operator",
                    "app.kubernetes.io/instance": "accuknoxconfig",
                    "app.kubernetes.io/name": "accuknoxconfig-sample",
                    "app.kubernetes.io/part-of": "accuknox-operator"
                  },
                  "name": "accuknoxconfig-sample",
                  "namespace": "accuknox-agents"
                },
                "spec": {
                  "ppsHost": "PLEASE_FILL_IN",
                  "knoxGateway": "PLEASE_FILL_IN",
                  "spireHost": "PLEASE_FILL_IN"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Security
          containerImage: public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:a935ad806e82760bb838bada30d3ed1a54c8c61b95fe5c6109f48985c6faf4d7
          createdAt: "2024-08-23T04:25:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: accuknox-agents
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: docker.io/accuknox/accuknox-operator
          support: Accuknox Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccuknoxConfig is the Schema for the accuknoxconfigs API
            kind: AccuknoxConfig
            name: accuknoxconfigs.operator.accuknox.com
            version: v1
        description: |2-

          The AccuKnox Operator brings native K8s support for deploying and managing AccuKnox Components on a Kubernetes cluster.
          The AccuKnox Operator helps to easily onboard a k8s cluster to the AccuKnox SaaS.

          AccuKnox Operator deploys and manages the components:

          1. Discovery Engine

          2. Policy Enforcement Agent

          3. Shared Informer Agent

          4. Feeder Service

          5. Agents Operator

          ## Prerequisites for using this Operator

          1. Get AccuKnox licence - https://app.accuknox.com/sign-up

          2. After Successful Sign-up into AccuKnox Application.

             Please naviagate to Settings >> Manager Cluster >> Onboard Cluster >>

             After successful onboarding, Copy the join token created by AccuKnox

          3. Create a secret using jointoken by executing the below command:

          ```
          oc create secret generic accuknox-jointoken --from-literal=token=<joinToken> -n accuknox-agents
          ```

          ## Deploy AccuKnox Operator using AccuknoxConfig Custom Resource

          All the information to configure the AccuKnox Operator will be provided via Accuknox during Cluster Onboarding steps.

          ### AccuknoxConfig Spec

          | Spec | Description |
          | ---- | ----------- |
          | spireHost | "spire host url" |
          | ppsHost   |  "pps host url" |
          | knoxGateway | "knox gateway url" |
          | joinTokenSecretName | "join token secret name if it's different than accuknox-jointoken" |
          | joinTokenSecretKey | "key name used to store the jointoken with the secret" |

          *Example*

          ```
          apiVersion: operator.accuknox.com/v1
          kind: AccuknoxConfig
          metadata:
            name: accuknoxconfig-default
            namespace: accuknox-agents
          spec:
            ppsHost: <ppsHost url>
            knoxGateway: <knox gateway url>
            spireHost: <spire host url>
          ```
        displayName: Accuknox Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CNAPP
        - CSPM
        - CWPP
        - Cloud Native
        - Runtime Security
        - Zero-Trust
        maturity: stable
        provider:
          name: Accuknox Inc.
          url: https://accuknox.com
        relatedImages:
        - docker.io/accuknox/feeder-service@sha256:295e897b6f27ce61c45fb0179b734ef6d4c9e91864bcc5d2884a2d1dd4db81b6
        - docker.io/accuknox/policy-enforcement-agent@sha256:a0364343afeac4537589bfc247f64fffd7aac16f8afe07ce050bc58b2af248fb
        - docker.io/accuknox/discovery-engine-discover@sha256:e2d37bfbd0e9d23c129c143eb7c9aab1c3ed55e9571b44888eea2b842f8a2361
        - docker.io/accuknox/discovery-engine-sumengine@sha256:f7927128b86fdfb635377bd599df06f0eb9b9f0d1b068a0047468e480855443e
        - docker.io/accuknox/discovery-engine-offloader@sha256:59cc594009de89a574cd569e701b6765d59a6899a03ad07eee1c025331df4331
        - docker.io/accuknox/agents-operator@sha256:b0ae2728f15ebae904efb778cd29215ccf07ffca945a59deb5484f6c19c1b21f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        - docker.io/accuknox/shared-informer-agent@sha256:9f53d7dba9569487c42c6cf9e88988a1adb4c0971742e8b521700dce2d22b8cc
        - docker.io/accuknox/discovery-engine-hardening@sha256:e596aaaabaec94e0abb4f92ad9266dc010c0e88a89d600e8750a360099cae3cc
        - public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:a935ad806e82760bb838bada30d3ed1a54c8c61b95fe5c6109f48985c6faf4d7
        version: 0.7.1
      entries:
      - name: accuknox-operator.v0.7.1
        version: 0.7.1
      name: stable
    defaultChannel: stable
    packageName: accuknox-operator-certified
    provider:
      name: Accuknox Inc.
      url: https://accuknox.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pachyderm, Inc.
      provider-url: https://pachyderm.com
    name: pachyderm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pachyderm-operator.v0.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "Pachyderm",
                "metadata": {
                  "name": "pachyderm-sample"
                },
                "spec": {
                  "console": {
                    "disable": true
                  },
                  "pachd": {
                    "metrics": {
                      "disable": false
                    },
                    "storage": {
                      "amazon": {
                        "credentialSecretName": "pachyderm-aws-secret"
                      },
                      "backend": "AMAZON"
                    }
                  },
                  "version": "v2.1.6"
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermExport",
                "metadata": {
                  "name": "pachydermexport-sample"
                },
                "spec": {
                  "backup": {
                    "target": "pachyderm-sample"
                  }
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermImport",
                "metadata": {
                  "name": "pachydermimport-sample"
                },
                "spec": {
                  "backup": "backups/pachyderm-backup-20220610.tar.gz",
                  "destination": {
                    "name": "pachyderm-restore",
                    "namespace": "pachyderm-workloads"
                  },
                  "storageSecret": "s3-credentials-secret"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
          createdAt: "2021-05-20T08:05:00Z"
          description: Pachyderm provides the data layer that allows data science
            teams to productionize and scale their machine learning lifecycle with
            data driven automation, petabyte scalability and end-to-end reproducibility.
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pachyderm/openshift-operator
          support: Pachyderm, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PachydermExport is the Schema for the pachydermexports API
            displayName: Pachyderm Export
            kind: PachydermExport
            name: pachydermexports.aiml.pachyderm.com
            version: v1beta1
          - description: PachydermImport is the Schema for the pachydermimports API
            displayName: Pachyderm Import
            kind: PachydermImport
            name: pachydermimports.aiml.pachyderm.com
            version: v1beta1
          - description: Pachyderm is the Schema for the pachyderms API
            displayName: Pachyderm
            kind: Pachyderm
            name: pachyderms.aiml.pachyderm.com
            version: v1beta1
        description: "The Pachyderm operator aims to make deploying and managing Pachyderm
          instances on Openshift container platform easier.\n\nPachyderm is the data
          foundation for machine learning.  Pachyderm provides industry leading data
          versioning, pipelines and lineage that allow data science teams to automate
          the machine learning lifecycle and optimize their machine learning operations
          (MLOps). \n\n## Getting started\n\n1. Pachyderm uses an object store to
          store your data. Set the pachd storage backend at `pachyderm.spec.pachd.storage.backend`
          to `AMAZON`, `MICROSOFT`, `GOOGLE` or `MINIO` depending on your storage
          provider. Additionally, create a user with read/write access to the object
          store.\n\n2. Setup the user credentials in the section, `pachyderm.spec.pachd.storage`.
          Required fields depend on the backend type set in step 1 above.\n\n3. Finally,
          deploy the pachyderm cluster in its own namespace.\n\nFor more information
          on getting started, reference the documentation.\n"
        displayName: Pachyderm
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pachyderm
        - versioning
        - machine learning
        - artificial intelligence
        links:
        - name: Documentation
          url: https://docs.pachyderm.com/latest/
        maintainers:
        - email: kyle@pachyderm.io
          name: Kyle White
        - email: eochieng@redhat.com
          name: Edmund Ochieng
        maturity: beta
        minKubeVersion: v1.19.0
        provider:
          name: Pachyderm, Inc.
          url: https://pachyderm.com
        relatedImages:
        - registry.connect.redhat.com/pachyderm/pachd@sha256:116477fefdf413ea02c6ac3d013f1f91d3f6e4b16aed2408c9890832cdb41430
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
        - registry.connect.redhat.com/pachyderm/pgbouncer@sha256:c3b6949227df24abb8934d111a34b44e87fb16194ac34a5d0fe284cff02e9ac0
        - registry.connect.redhat.com/pachyderm/init-utils@sha256:5c0258bfa0203079925648c0a7dda6d05a29b6f461854821a1eec541777ff359
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.redhat.io/rhel8/postgresql-13@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - registry.connect.redhat.com/pachyderm/backup-handler@sha256:e2ef6bd8191e8ad58c765c0db8f5bb21ea7123caed0c0f58fe681a024201306d
        - registry.connect.redhat.com/pachyderm/worker@sha256:374f83d92899707bc87b0318fda6bffa78d2e9e523b827bb3d40693a2d1bdecd
        - registry.connect.redhat.com/pachyderm/etcd@sha256:7f330b51874379b1d4b71d94e37e0ecfa52b86d628296136970688cf204f53b5
        version: 0.1.4
      entries:
      - name: pachyderm-operator.v0.1.4
        version: 0.1.4
      name: stable
    defaultChannel: stable
    packageName: pachyderm-operator
    provider:
      name: Pachyderm, Inc.
      url: https://pachyderm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-listener-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: listener-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/tools@sha256:47204b2697dfbab5eaf9d6dcb4c160dd427a148d6b840727239b773810e74fc9
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/tools@sha256:47204b2697dfbab5eaf9d6dcb4c160dd427a148d6b840727239b773810e74fc9
        - quay.io/stackable/listener-operator@sha256:b9c7957e7e562889bddd08e679abff05bb677672b3f735932f68514b77eacfd5
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:a6d6631d35c32b874c995b6af889636ac8dd1f39dcb68810d9f9b10a4e2c1987
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:161e754437428c68490333337303fb30b695a9ac677c3a00a1dc27565b855ab8
        version: 24.11.1
      entries:
      - name: listener-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: listener-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/listener-operator@sha256:da53970954b889bf4d874ae3df7e411b9abf41950ef7d50d6289831826f045a0
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/listener-operator@sha256:da53970954b889bf4d874ae3df7e411b9abf41950ef7d50d6289831826f045a0
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:71f8465e35a67f4b96e4c4904e67ccac9bbb9ae347426afb8bbfaabd6709e983
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        version: 25.3.0
      entries:
      - name: listener-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: listener-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        version: 25.7.0
      entries:
      - name: listener-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: listener-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
        version: 25.7.0
      entries:
      - name: listener-operator.v25.7.0
        version: 25.7.0
      - name: listener-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: "25.7"
    packageName: stackable-listener-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      pod-security.kubernetes.io/audit: privileged
      pod-security.kubernetes.io/enforce: privileged
      pod-security.kubernetes.io/warn: privileged
      provider: NVIDIA Corporation
      provider-url: ""
    name: gpu-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gpu-operator-certified.v25.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "kernelModuleType": "auto",
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": []
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
          createdAt: Thu Sep 18 18:11:36 PDT 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <25.3.4'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:c22dc67465c33d2f17932931229e61a6e7d9e32c5ed2ecefe8d10e873ffedfeb
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:5c4e61f7ba83d7a64ff2523d447c209ce5bde1ddc79acaf1f32f19620b4912d6
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:619e7bf48168d76a7e087c3bf6190fda9b10449c9839309fa4f9ed5b8a9e8804
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b689dfb7806ae51b0072620b47c26114f05477e8ad9725892ee10366e5d93d8d
        - nvcr.io/nvidia/cuda@sha256:d19fe621624c4eb6ac931b8558daa3ecc0c3f07f1e2a52e0267e083d22dceade
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:119de9a331a47203858b99901f44d0c4a8052961b4e60327f4b100d0ab8c9df0
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:6bb786c41fa7b1a6238c023566fdeaeeea3ba842a8daec015a7bd85a8037b249
        - nvcr.io/nvidia/driver@sha256:43aa99352e1dc47e9218f36340ae6328a6282ed1e7e9a61b4204475e980230ef
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3191cc0684e92d655e72fe9faf3da95fe6b8a6c2eb4ad396f75ba9b98c021aaf
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:66c3d68ca2b35d47b5562300a10ff8711b953ba3a15df7dbc409d81613ebe5d0
        - nvcr.io/nvidia/k8s-device-plugin@sha256:3c54348fe5a57e5700e7d8068e7531d2ef2d5f3ccb70c8f6bac0953432527abd
        - nvcr.io/nvidia/driver@sha256:44422f10e899d3419fbd611a02fc1c6127e071a2067bd36e3debb043bf251de0
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:51c8f71d3b3c08ae4eb4853697e3f8e6f11e435e666e08210178e6a1faf8028f
        - nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
        - nvcr.io/nvidia/driver@sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735
        version: 25.3.4
      entries:
      - name: gpu-operator-certified.v25.3.4
        version: 25.3.4
      - name: gpu-operator-certified.v25.3.3
        version: 25.3.3
      - name: gpu-operator-certified.v25.3.2
        version: 25.3.2
      - name: gpu-operator-certified.v25.3.1
        version: 25.3.1
      - name: gpu-operator-certified.v25.3.0
        version: 25.3.0
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: stable
    - currentCSV: gpu-operator-certified.v1.10.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "dcgmExporter": {
                    "config": {
                      "name": ""
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                  },
                  "devicePlugin": {
                  },
                  "driver": {
                    "enabled": true,
                    "use_ocp_driver_toolkit": true,
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "gfd": {
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "operator": {
                    "defaultRuntime": "crio",
                    "deployGFD": true,
                    "initContainer": {
                    }
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:c7f9074c1a7f58947c807f23f2eece3a8b04e11175127919156f8e864821d45a
          createdAt: Mon Apr 18 10:06:40 PDT 2022
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=v1.9.0 <v1.10.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:c2e3d0fe41a0d227dbb70caec03c780cc76317515e5ab3875f31d50c63f41c66
        - nvcr.io/nvidia/cuda@sha256:7f0f7dc8c993a86865d4d9f9b0e36d113ea15603d5d5279d2879fa2bc21c8236
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:5b16056257acc51b517d9cdb1da3218693cefc214af93789e6e214fd2b4cacf1
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:67c48a8370c34f68b46eb57339e87d01c04198368a0423553fab7565c862c4df
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:e8f06ce53415e8572f633b2169289d973993cb1e49d4dcc686ac134194d88f33
        - nvcr.io/nvidia/driver@sha256:f99b66855e92451a59a722355dd52bf97eb985cc8a647172273de8178aeaa93c
        - nvcr.io/nvidia/gpu-operator@sha256:c7f9074c1a7f58947c807f23f2eece3a8b04e11175127919156f8e864821d45a
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:9abb063e6b74085728fead5ed5bbe3bd7b3f57de60e87f6b964bba2d46f1bc67
        - nvcr.io/nvidia/k8s-device-plugin@sha256:aa95c16106280e36a5e32d2fe4c66e8b70f5a114860c6f4ed5b1a4085c63601b
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:24d804e8f005d7aeca8343aa13e5f92295d8642a4c47cb24e3ac86a22543bc37
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:ac13256c295126b854a07421c45de0721e7a57974b38b3e66eb1fdc3a4e29f2e
        version: 1.10.1
      entries:
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v1.10
    - currentCSV: gpu-operator-certified.v1.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    },
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "dcgmExporter": {
                    "config": {
                      "name": ""
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                  },
                  "devicePlugin": {
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": true
                  },
                  "vgpuDeviceManager": {
                    "enabled": true,
                    "config": {
                      "name": "vgpu-devices-config",
                      "default": "default"
                      }
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:53180f81fb7e1e4bcd49a556bc6285c2262abbab8083b29d2688b02d4200dd2d
          createdAt: Thu Aug  4 11:57:56 PDT 2022
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=v1.9.0 <v1.11.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7a0ccbcbe8d379bfc441e06504eb45d0d433a37e2a9e6dde83d13347296ebe81
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:8490d555131a57a11311561667ea9ec8cf554a399d2492a6f9e6677c01ec4b9a
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:10d825ccbfcf87f8c74f841fa89ae345f2d381fb47a5b2b31f61b72ebbbb0915
        - nvcr.io/nvidia/driver@sha256:cf8ce04907d0e8decf90a8dccfa6134af9c39509d2f8a21c355ed71223acacdb
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:16ee5b1a8d9217f761c40a06c7ae598cbc893fbc35d99ff9b84fa886ea6b9b37
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:f48e8b430248a9cadf716a9211bb0050d4a663d1f1386ea107e72917f82e5b5f
        - nvcr.io/nvidia/driver@sha256:8c9fabc1cd773e9526eb7e8526084f065f6cca9d0fd8dc2a672a0bb717f1cc60
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:e88e610d09f1a7528ab32cb54259aee340eba2b2bf66b92a28157626daadc82d
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:2d6e86db0e2b0db64c00566938e1851e25d4bb1019b1e7b696e2b199cba33397
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:39127d04c9faff3f22068dcf463ed48d02be684cd18e3f14db843bddbdd783a7
        - nvcr.io/nvidia/gpu-operator@sha256:53180f81fb7e1e4bcd49a556bc6285c2262abbab8083b29d2688b02d4200dd2d
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3804e2b03c44d25dd410e3882b7fb6d1fde41f55110d7702af47136b042a63ff
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:b90a519e9f30fbc013e1806bc081268724e9e80b4035c67c167ab3b43a37704a
        - nvcr.io/nvidia/driver@sha256:b8799e7397fae5a27446737443091eed05b918816f53c400a48be71125ee22e8
        - nvcr.io/nvidia/cuda@sha256:4400b56f81095b24fc9dde93ad7ad7d0ca6a7e9736c4c7f1d4e2bcbf0afbe45c
        - nvcr.io/nvidia/driver@sha256:6e2902d73c54cf57bec94edf166d206cfeb0b948b5aab70800ccbe025064b3a1
        version: 1.11.1
      entries:
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v1.11
    - currentCSV: gpu-operator-certified.v22.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:54d5846c0291db621931a8544c9cb6437512470a93278ca9130ba73f2e2f0055
          createdAt: Mon Jan 30 15:43:24 PST 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <22.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:b860ef04cc6cd8c481810fa19bef7221071f5a3862640b2ad6c2704728dc77c8
        - nvcr.io/nvidia/cuda@sha256:5f2a2d8977f2c59abe88394f05cc3c044194554f90994d8554b0e1117ab5900d
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:9177a0ae30798b42d0387f6a20cd3ce3cd1799a91b7866bf812368764b05b1af
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:bec9f026d9b3d9404c78d6091817a359015c6a7aa411735b34138c1518853b5d
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:50c2a2b998e467c82716efa79fe136aa6f7ea95fd23576cf384d251bb9628640
        - nvcr.io/nvidia/gpu-operator@sha256:54d5846c0291db621931a8544c9cb6437512470a93278ca9130ba73f2e2f0055
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:64d757b4c80b910e64647a84a1d592fab2ea3313ff6dce30c25c3a08e180bd74
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:0d47dad29d2ef445b301c5c64717758eed43a606345b79f97bce2e64b40a91a8
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:efb88937f73434994d1bbadc87b492a1df047aa9f8d6e9f5ec3b09536e6e7691
        - nvcr.io/nvidia/driver@sha256:f250d1e4989fd4821e81bad9227eeeaa531ca8810f3a40e996f1ca2af7945f96
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:c3cf59dd5d6160eba5d816ade2e81b35ebb10f4884df67971f6ace36f8e6efc1
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:00f1476548fbed9ee01961443a73bf65396c2e8bb2b84426f949dd56cb4d14cd
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:9a00cdfdddb73327ef8e8e0fa60e50926a388d380c551bba6a3d3012be40401d
        - nvcr.io/nvidia/driver@sha256:f2ca4612f3418e0e3963be8c3771ffe463cdc1fc1b810437a3aad9d76bb6499d
        - nvcr.io/nvidia/driver@sha256:c5a62bc2a6cce339798e4b7cf9cec47e289c2d30b40f451ba8c9e958a68e3e55
        - nvcr.io/nvidia/k8s-device-plugin@sha256:9c17d3a907eb77eb8f7b4f3faf52d8352e4252af92003f828083f80d629bd2c3
        - nvcr.io/nvidia/driver@sha256:268c26781c46b36fe691638e807eb5c3a2cd077029ddb095ac75264adac04925
        version: 22.9.2
      entries:
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v22.9
    - currentCSV: gpu-operator-certified.v23.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:08ca33f491195b7ba867be1e6642b5ff8aff6a4b0b7cfc0ec87619252ccfddc7
          createdAt: Mon May 22 15:22:15 PDT 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <23.3.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:2d7e32e3d30c2415b4eb0b48ff4ce5a4ccabaf69ede0486305ed51d26cab7713
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:03e43d6b1460cadb33509d5ee07a5e27603ab89d7ec35f67d8a51bb78745b463
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:5489125ceae5864280e4d6a9ab52ab0f650b3179349a7298c4a204feb60b661a
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:8f03f74076b5aa6672a5d75a398fa7515fcde57fdb1cf8079e015b3366efd3b5
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:21dfc9c56b5f8bce73e60361d6e83759c3fa14dc6afc2d5ebdf1b891a936daf6
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:f10170b076badda8f669af8b0f7b47eb5470abaf35a67dda9c4e182d66138177
        - nvcr.io/nvidia/driver@sha256:917f92c589c3124dba54f66f89c44b93b0beaf418f01d0da6a546e1670fdf3a9
        - nvcr.io/nvidia/driver@sha256:8ff0ee88fe6c4ff031ef642593f93625bfd9a48fe7be8a79a5c2736117e36941
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:cab21c93987a5c884075efe0fb4a8abaa1997e1696cbc773ba69889f42f8329b
        - nvcr.io/nvidia/k8s-device-plugin@sha256:80871275fbcf888d4683c58638c7a75b3c46091040cafc6210a8c93513adb146
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:84ce86490d0d313ed6517f2ac3a271e1179d7478d86c772da3846727d7feddc3
        - nvcr.io/nvidia/driver@sha256:73b8928a1b1fb650d2ca8d6451a7d7c551fd0509f7511c2fd5bfdcef83ae4f5d
        - nvcr.io/nvidia/gpu-operator@sha256:08ca33f491195b7ba867be1e6642b5ff8aff6a4b0b7cfc0ec87619252ccfddc7
        - nvcr.io/nvidia/driver@sha256:3df118f2330e95db3c76173cca35e542b2da120ec0bb8f6aff83c4c96c0c2daf
        - nvcr.io/nvidia/cuda@sha256:62d845b4bd6de4be65f0b6daaae6831c29d4529955f99db5d26f5f7179106285
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:0d47dad29d2ef445b301c5c64717758eed43a606345b79f97bce2e64b40a91a8
        version: 23.3.2
      entries:
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.3
    - currentCSV: gpu-operator-certified.v23.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:d2e0f8bd86e8e9f2ff3514e113657658584ca4a2d0b82dac2e47364cc0b30648
          createdAt: Thu Aug 31 12:22:44 PDT 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <23.6.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:bbdc4415fe2e71eeea1b8a6e50ee562e11704dabb46db1db4aea807102886a35
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:00a91252be8e62da6f28cdf1ec3721f2e945ae296938b0b925ef9db2bf6dcc8b
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:bb845160b32fd12eb3fae3e830d2e6a7780bc7405e0d8c5b816242d48be9daa8
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:60086c33226bbe7f04b611c20d308fed5825226f8a0a5635bc58d8ed55cc08a5
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:b30f5c64791dc5406b57eaf51ac40195dc1051a18cd1f38fd0a43c65f1d6dba9
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:a229beba8540549c234a046e975bf4bf8c50d521dfe551af56f6e042b25ec017
        - nvcr.io/nvidia/driver@sha256:3382e254056f28831767bc6729bc2594353a5ff2a28fe9f2d94396c597bb23d8
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:bf06c8ad380a36e5014607bd3ab05b80e8e5e7e77d23d0dd89fe53a5564152fa
        - nvcr.io/nvidia/k8s-device-plugin@sha256:f1617c3fe54fd9f6d4a00358e5da1a9fa9e8963a4ef6b3ed06f47d4971ddff4b
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:878a6162a4fa0b61dae02c7cfc13ba4d28bb06f92fd62d0c521633ab507d675a
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:9484110986c80ab83bc404066ca4b7be115124ec04ca16bce775403e92bfd890
        - nvcr.io/nvidia/gpu-operator@sha256:d2e0f8bd86e8e9f2ff3514e113657658584ca4a2d0b82dac2e47364cc0b30648
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:0379c59b43e11861837787a0da22e70b528718373bb10de9e8b7752849bd4e7a
        - nvcr.io/nvidia/driver@sha256:b4517810f2884f345b4f2323e83e41d28f54cdc6e2caf0690bb7ed4f3947dbcc
        - nvcr.io/nvidia/cuda@sha256:64c0ff5fd4a27d4b1d46378634f7a7398c75f87c18c70c988dde18910d461560
        version: 23.6.1
      entries:
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.6
    - currentCSV: gpu-operator-certified.v23.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:7481a3697783dcdca9ae78e7b548a6900e86ea33ab49ec14f0ba55db2fdb1a2e",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:3988b3b570cf65b0d338cb29bd24b0b9bf1885e1563fbb398ef23cba0b59e6bb
          createdAt: Wed Mar 6 16:43:36 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <23.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cuda@sha256:714547d54e5fe4191019a1e5f1daffc7fab7481b619b79c378541dafc76c9e5d
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:387021553dfb16aab633228d42f63f04fa932b4f46add07527f296dfe97e5148
        - nvcr.io/nvidia/driver@sha256:abda1ac56371d55917b96ff330109980f468e133c9d5705da0ef87429f14ccd7
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:8d1ec78f2b1ddb7f0c47453d0427231190747bda411733a7dd0c8f5196f09e9c
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:9aefef081c3ab1123556374d2b15d0429f3990af2fbaccc3c9827801e1042703
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:a67d8e92861a2dce5649105c07561e4422e9fe4ba81a6525dc0d70a7ef85f9c0
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:011fb450af3fa2e8fe5d28d590e4c653631447bc23d149591ced3d89089c4f2c
        - nvcr.io/nvidia/k8s-device-plugin@sha256:bb81ae8d034d1e1c6a342091a5f1dda5884dd43ab47e96624487a133d1edd4b3
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:6a05d6a1923fda756aed0dddf7ed23a83c30cf1e6c519fc39dd70c0309ec8257
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:59a3875e7a37eb370385e654184efa3a1b193c9ea352165818496b19cbe14aa4
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:27c44f4720a4abf780217bd5e7903e4a008ebdbcf71238c4f106a0c22654776c
        - nvcr.io/nvidia/driver@sha256:56c79482582cdfbc58d3134e8672637c5bf05f328880f76898f526143d04c6af
        - nvcr.io/nvidia/gpu-operator@sha256:3988b3b570cf65b0d338cb29bd24b0b9bf1885e1563fbb398ef23cba0b59e6bb
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:925f4a47710e4318ed457930f5406174c1f6d28b1bf6b1bc310687fec0fde712
        - nvcr.io/nvidia/driver@sha256:7481a3697783dcdca9ae78e7b548a6900e86ea33ab49ec14f0ba55db2fdb1a2e
        version: 23.9.2
      entries:
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.9
    - currentCSV: gpu-operator-certified.v24.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:96f25c67e5b1072d5981080e12d65ec37eb9ef2fc0494499416aa801b0a34da3",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:041e75a3df84039c2dbbd4b9d67763bd212138822dbb6dbc0008858c1c6eff8d
          createdAt: Tue Apr 30 14:18:26 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.3.0'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:2442fd2ec0ee29746abbced727c53741f017e8f3f615321c1ec42c1c3ffa0b12
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:0c77725de8c42d248ed825453efd2e005f4900f4be384fd23084f6b721ddd0e0
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:f0e4e14a3081417be8e6a5d855fb0cb69e1b63bc54a74f17e6d7084abe275588
        - nvcr.io/nvidia/gpu-operator@sha256:041e75a3df84039c2dbbd4b9d67763bd212138822dbb6dbc0008858c1c6eff8d
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:2cc2119d20515f8ca7a61a0f3932578d69a45bcdea49e6c320a89c56f105e7d9
        - nvcr.io/nvidia/k8s-device-plugin@sha256:80674b19898ecf9ed6949e39674da769d6feeeb01bea54b914ef9ff502834f49
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:ecb867c5787e44f1889b257e6c03b07748689a7b293c8f1affb1d4cb449b235c
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:7798eb9b9424e09959fc808596f87bdb68ac5740174123c6646dee2166d3e7d2
        - nvcr.io/nvidia/driver@sha256:a836ccbe21da735aee1c39b81060ed5e2fdb4ffa339874baaf4634f1e9259f74
        - nvcr.io/nvidia/driver@sha256:56c79482582cdfbc58d3134e8672637c5bf05f328880f76898f526143d04c6af
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:2edc1d4ed555830e70010c82558936198f5faa86fc29ecf5698219145102cfcc
        - nvcr.io/nvidia/cuda@sha256:ae0623ec8634b6c88f815b88037763def160cbbac15013b77ddef257fc276c9a
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:d8774d6afd37110199512636172823ae8749e5ea2e13760f57c255a74f47018c
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:1ae0b923bc57f47bab046b50c50110f6914bbaffbfef704df34b3fe332db2e31
        - nvcr.io/nvidia/driver@sha256:96f25c67e5b1072d5981080e12d65ec37eb9ef2fc0494499416aa801b0a34da3
        version: 24.3.0
      entries:
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.3
    - currentCSV: gpu-operator-certified.v24.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:858de27c152669f5a3cf4287406405b16dd5bb70c0373324eb735511997bb415",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:8e0969cffc030a89c4acd68e64d41dd54e3bce8a794106b178d4dbd636a07f1c
          createdAt: Wed Sep 25 08:25:16 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.6.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:15dab1273345df4a5844c4c761d064dbc4b592101251dc39174e597137123027
        - nvcr.io/nvidia/driver@sha256:07e11f85d54d49ec9648fb06e148b8d832ee1f9c3549a915eee853c9ef2949c2
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:ae63fac9a4057a7646f0cf0ee0566e8928529adde05c4c0a017cda0599e381b2
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:781fb47e264d9e0fbc8da5bd046e5e678316c866bc36ddd4b56d4eb0de682d5b
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7ad2c9f71fe06f9f7745ac8635f46740fbdff4f11edd468addfab81afcdfa534
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:969147c01d63be5d1fe458f32f1cc0c7408cf3062531db91408e2fc57b4d8a67
        - nvcr.io/nvidia/cuda@sha256:de5b598bca89850567c4c104411d66bb52f47c9179199e6a3be6829b7ac586a2
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:96380b95396b7f29d2ed2ec8ef1ad5a8bcf9f55051db723295dc0a20db845331
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:740abc3ff657545c10effd5354f09af525200ed9a1b7623f0c2e8c7bd9e4a4e2
        - nvcr.io/nvidia/driver@sha256:858de27c152669f5a3cf4287406405b16dd5bb70c0373324eb735511997bb415
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:21f4c8b88716e8e6f732f9fb4c2efaef937c227491a8631c5e55036f80f39a4d
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:33de74eb590f071403c17b6c210c02963245851971168bc0c07c06c100a9f376
        - nvcr.io/nvidia/driver@sha256:a6d12fb5753f267dda25dfd38910f972bc632c006a24107fa50e20bba3642d7c
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:7bcc188703f2fac630f7a4ff8960e6733ac3a29adf6a946533b796d9a27b8acf
        - nvcr.io/nvidia/gpu-operator@sha256:8e0969cffc030a89c4acd68e64d41dd54e3bce8a794106b178d4dbd636a07f1c
        version: 24.6.2
      entries:
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.6
    - currentCSV: gpu-operator-certified.v24.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:6fe74322562c726c8fade184d8c45ebae3da7b1ea0a21f0ff9dc42c66c65e692",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:92f5262bb4934e0ea29eb10ec0ad4fc293164f2f2f94d48a935922791abbfdac
          createdAt: Mon Jan 27 16:10:52 PST 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/gpu-operator@sha256:92f5262bb4934e0ea29eb10ec0ad4fc293164f2f2f94d48a935922791abbfdac
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7089559ce6153018806857f5049085bae15b3bf6f1c8bd19d8b12f707d087dea
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:857f7669ae5ce2a43a3b5691fa970085098321ef8e90a896fa8dda3dee5b1d2b
        - nvcr.io/nvidia/driver@sha256:6fe74322562c726c8fade184d8c45ebae3da7b1ea0a21f0ff9dc42c66c65e692
        - nvcr.io/nvidia/driver@sha256:d55b57938866e538acc3a71ca32f8cf87e71c591abd4a34695ee428e7ec2fa73
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:7edd7a0413dcb39b6e3bcefaf06812f3293c8e480ca10783e821a561ed686200
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:cf39d78ced7fb5727a9668ee2cd44b14bb7a23a95b83d5464b7d755740e02121
        - nvcr.io/nvidia/cuda@sha256:b3b340847db9abd0c39703278f5638d3e65e1369e9a526a19b3138d5a6563527
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:d934033f9ff89087d555850eb1da150eaaa27b11e5ad21db1b874b8bc0b8d35b
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:40830d3561c14743f484d45b498141f9e86b1308e16fae3978110783927264ab
        - nvcr.io/nvidia/driver@sha256:38b66a8d44cab9e2c62da9e101f32cd9dbcb5e02d8e57b47671284d374ca3695
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:34eab99d7992a57c35803ec5e0afd538d16e997341e951fef4c4019afb08793e
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b072c5793be65eee556eaff1b9cbbd115a1ef29982be95b2959adfcb4bc72382
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:4ffa1cd2a6497eb647a89ed259dcfb007554737b9d80f69bc173a2c3cd72a1da
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:e8398b2451996e5c64fc3855e7a3c93b667a85f69c0e259049d3b0012c21545d
        - nvcr.io/nvidia/driver@sha256:026f8f2d29b7058ecaaa5b98666e28e1b3646b73e7d7cda3dc8026d47a929152
        version: 24.9.2
      entries:
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.9
    - currentCSV: gpu-operator-certified.v25.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "kernelModuleType": "auto",
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": []
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
          createdAt: Thu Sep 18 18:11:36 PDT 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <25.3.4'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b689dfb7806ae51b0072620b47c26114f05477e8ad9725892ee10366e5d93d8d
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:66c3d68ca2b35d47b5562300a10ff8711b953ba3a15df7dbc409d81613ebe5d0
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:51c8f71d3b3c08ae4eb4853697e3f8e6f11e435e666e08210178e6a1faf8028f
        - nvcr.io/nvidia/driver@sha256:44422f10e899d3419fbd611a02fc1c6127e071a2067bd36e3debb043bf251de0
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3191cc0684e92d655e72fe9faf3da95fe6b8a6c2eb4ad396f75ba9b98c021aaf
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:619e7bf48168d76a7e087c3bf6190fda9b10449c9839309fa4f9ed5b8a9e8804
        - nvcr.io/nvidia/driver@sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735
        - nvcr.io/nvidia/cuda@sha256:d19fe621624c4eb6ac931b8558daa3ecc0c3f07f1e2a52e0267e083d22dceade
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:119de9a331a47203858b99901f44d0c4a8052961b4e60327f4b100d0ab8c9df0
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:5c4e61f7ba83d7a64ff2523d447c209ce5bde1ddc79acaf1f32f19620b4912d6
        - nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
        - nvcr.io/nvidia/driver@sha256:43aa99352e1dc47e9218f36340ae6328a6282ed1e7e9a61b4204475e980230ef
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:c22dc67465c33d2f17932931229e61a6e7d9e32c5ed2ecefe8d10e873ffedfeb
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:6bb786c41fa7b1a6238c023566fdeaeeea3ba842a8daec015a7bd85a8037b249
        - nvcr.io/nvidia/k8s-device-plugin@sha256:3c54348fe5a57e5700e7d8068e7531d2ef2d5f3ccb70c8f6bac0953432527abd
        version: 25.3.4
      entries:
      - name: gpu-operator-certified.v25.3.4
        version: 25.3.4
      - name: gpu-operator-certified.v25.3.3
        version: 25.3.3
      - name: gpu-operator-certified.v25.3.2
        version: 25.3.2
      - name: gpu-operator-certified.v25.3.1
        version: 25.3.1
      - name: gpu-operator-certified.v25.3.0
        version: 25.3.0
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v25.3
    defaultChannel: v25.3
    packageName: gpu-operator-certified
    provider:
      name: NVIDIA Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Yugabyte Inc
      provider-url: ""
    name: yugabyte-platform-operator-bundle-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: yugabyte-platform-operator.v0.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "yugabyte.com/v1alpha1",
                "kind": "YBPlatform",
                "metadata": {
                  "name": "ybplatform-sample"
                },
                "spec": {
                  "domainName": "cluster.local",
                  "helm": {
                    "package": "/opt/yugabyte/helm/yugabyte-latest.tgz",
                    "timeout": 900
                  },
                  "helm2Legacy": false,
                  "image": {
                    "commonRegistry": "",
                    "nginx": {
                      "name": "nginxinc/nginx-unprivileged",
                      "registry": "",
                      "tag": "1.17.4-amd64"
                    },
                    "postgres": {
                      "name": "postgres",
                      "registry": "",
                      "tag": 11.5
                    },
                    "prometheus": {
                      "name": "prom/prometheus",
                      "registry": "",
                      "tag": "v2.2.1"
                    },
                    "pullPolicy": "IfNotPresent",
                    "pullSecret": "yugabyte-k8s-pull-secret",
                    "repository": "registry.connect.redhat.com/yugabytedb/yugabyte-platform",
                    "tag": "latest",
                    "thirdparty-deps": {
                      "name": "yugabyte/thirdparty-deps",
                      "registry": "quay.io",
                      "tag": "latest"
                    }
                  },
                  "ip_version_support": "v4_only",
                  "ocpCompatibility": {
                    "enabled": true
                  },
                  "rbac": {
                    "create": true
                  },
                  "securityContext": {
                    "enabled": false,
                    "fsGroup": 1001,
                    "runAsUser": 1001
                  },
                  "tls": {
                    "certificate": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tDQpNSUlHTHpDQ0JSZWdBd0lCQWdJUUQ2R1ZmL0ViUDhSMFQvcWxjRGZyM3pBTkJna3Foa2lHOXcwQkFRc0ZBREJlDQpNUXN3Q1FZRFZRUUdFd0pWVXpFVk1CTUdBMVVFQ2hNTVJHbG5hVU5sY25RZ1NXNWpNUmt3RndZRFZRUUxFeEIzDQpkM2N1WkdsbmFXTmxjblF1WTI5dE1SMHdHd1lEVlFRREV4UlNZWEJwWkZOVFRDQlNVMEVnUTBFZ01qQXhPREFlDQpGdzB4T1RBNE1ERXdNREF3TURCYUZ3MHlNVEEzTXpFeE1qQXdNREJhTUI4eEhUQWJCZ05WQkFNTUZDb3VZMnh2DQpkV1F1ZVhWbllXSjVkR1V1WTI5dE1JSUJJakFOQmdrcWhraUc5dzBCQVFFRkFBT0NBUThBTUlJQkNnS0NBUUVBDQp3SmswU25ub1pRYW1hWm0rMVVJdjdRYjBLalhZeVEwckpJSGpsbE03elhFYUFoL0JqVjZ1M0h1TWxrTlZSekVFDQpGeEtGUm5nSlpXTVlPcy9sWDU1QmtBS0RocXNmaDl0ZVA4ZFVQU1RTVTN6bWRVc0dTcDJIMGpKbVhuRWNvMFpzDQpwYlVyYWhlcWlxSUUyZ1gwQlpvYldvWGRHc1EydXJiUGxyMGlVSnUxVnp4SGRoTS9nTnFQWEN5bE5YdENrbVI2DQo0eXpIa0Rob21pRTAySTMwTXBlOERxczl4d3lKd3F6T3J2ZnA0eUE0Ukhna2g1clovK1Erc0VQNkIyVzBHQXc2DQpzZDNTYytOTytYVUZoVnlFRDhqZkNsTndPVHZodi9keE5IWHZ4WHd0RWNjZTd1TVBBTFJzSFh1RW1kTkN5N2pvDQpibzJuUWFObzZBQ1c0U2s5RWJVZXB3SURBUUFCbzRJREpqQ0NBeUl3SHdZRFZSMGpCQmd3Rm9BVVU4b1hXZnhyDQp3QU1oTHhxdTVLcW9ISUpXMm5Vd0hRWURWUjBPQkJZRUZHemw3bjdOMy9GbFdDSDFvdWFCWldKMjI1UStNQjhHDQpBMVVkRVFRWU1CYUNGQ291WTJ4dmRXUXVlWFZuWVdKNWRHVXVZMjl0TUE0R0ExVWREd0VCL3dRRUF3SUZvREFkDQpCZ05WSFNVRUZqQVVCZ2dyQmdFRkJRY0RBUVlJS3dZQkJRVUhBd0l3UGdZRFZSMGZCRGN3TlRBem9ER2dMNFl0DQphSFIwY0RvdkwyTmtjQzV5WVhCcFpITnpiQzVqYjIwdlVtRndhV1JUVTB4U1UwRkRRVEl3TVRndVkzSnNNRXdHDQpBMVVkSUFSRk1FTXdOd1lKWUlaSUFZYjliQUVDTUNvd0tBWUlLd1lCQlFVSEFnRVdIR2gwZEhCek9pOHZkM2QzDQpMbVJwWjJsalpYSjBMbU52YlM5RFVGTXdDQVlHWjRFTUFRSUJNSFVHQ0NzR0FRVUZCd0VCQkdrd1p6QW1CZ2dyDQpCZ0VGQlFjd0FZWWFhSFIwY0RvdkwzTjBZWFIxY3k1eVlYQnBaSE56YkM1amIyMHdQUVlJS3dZQkJRVUhNQUtHDQpNV2gwZEhBNkx5OWpZV05sY25SekxuSmhjR2xrYzNOc0xtTnZiUzlTWVhCcFpGTlRURkpUUVVOQk1qQXhPQzVqDQpjblF3Q1FZRFZSMFRCQUl3QURDQ0FYNEdDaXNHQVFRQjFua0NCQUlFZ2dGdUJJSUJhZ0ZvQUhVQTdrdTl0M1hPDQpZTHJoUW1rZnErR2VacU1QZmwrd2N0aURBTVI3aVhxby9jc0FBQUZzVHEwL013QUFCQU1BUmpCRUFpQlg3NXJ0DQpwNkdQbUQxbVl1S1hvcXVRVytIcUtRNE8rV1RXbTNCbVBteWF1d0lnQUoya01WdmduaUx1dEh5Y3AyYzBwdDRsDQptL3M5SWN2Yi9aN09jMU82b2xJQWRnQ0hkYi9uV1h6NGpFT1pYNzN6YnY5V2pVZFdOdjlLdFdEQnRPci9YcUNEDQpEd0FBQVd4T3JUOXNBQUFFQXdCSE1FVUNJSHphY3NCcmpEN1FER1ZRSHA5YUZPMnJ4VEx4NU9GOVE2MTFYaUU2DQpZUEk3QWlFQW9xd3pTd1BVaGt4bVhYNzN0SDVEcVNwaUxuOHdrMXFWeWhZOTVmSmxZa0VBZHdCRWxHVXVzTzdPDQpyOFJBQjlpby9pakEydWFDdnRqTE1iVS8wek9XdGJhQnFBQUFBV3hPclQ1M0FBQUVBd0JJTUVZQ0lRQ1ZLaHQ5DQpWTWhlNTUvZEwrSGN6Qzk5a2xheG9DcW5zOStRNGVOTDZ2Smd4QUloQUsyZmp4a1RadGw4WFZ4WHEwc2NSTWNJDQpiSktHcGJYdEV0N3dFVzRkZDNGS01BMEdDU3FHU0liM0RRRUJDd1VBQTRJQkFRQm1GcDNCK1g2Y21Cb0FLWkhrDQpIWWxzanhIZk01bDQyZWJtSFdGdlQyTi9aVkZ3MEZ3YndBNExyNzlvd01QdU5FbVNNa2pmVVhxR011Q2VlVTdODQo0MlRwUFdPaGNzVHF6ZkxVa0hhMU9rK011L3RUZHBNcmhROEVqbkhMY3FKVE9MbmRSVy8zcE5IM2xrK2FJejNoDQpDVGFueElRVEE2cUxRNThRakszWmFMVzdqWVVNcVJFcnBGVDlBVXkxMDAxMjlOam5HVkhtYmU1bW1ZbEwvQlhMDQpmdHBJUUJJZnhCNXpHbHpTcjRHT3p5RDBES05YU040RU0yM2pIM1ljeUg4eFQrODU4ZkxSaXQ4SFlOQ3RQbkloDQoyRzR6UjV3emZFVDlnZUh6aWtuY0ZMWUpMMEpKMVIrZGVNRE91bmJWOUJqcmZJb3NNSk53TUV2ejFZczlBWXhHDQpxSFNYDQotLS0tLUVORCBDRVJUSUZJQ0FURS0tLS0tDQotLS0tLUJFR0lOIENFUlRJRklDQVRFLS0tLS0KTUlJRXNUQ0NBNW1nQXdJQkFnSVFDS1dpUnMxTFhJeUQxd0swdTZ0VFNUQU5CZ2txaGtpRzl3MEJBUXNGQURCaApNUXN3Q1FZRFZRUUdFd0pWVXpFVk1CTUdBMVVFQ2hNTVJHbG5hVU5sY25RZ1NXNWpNUmt3RndZRFZRUUxFeEIzCmQzY3VaR2xuYVdObGNuUXVZMjl0TVNBd0hnWURWUVFERXhkRWFXZHBRMlZ5ZENCSGJHOWlZV3dnVW05dmRDQkQKUVRBZUZ3MHhOekV4TURZeE1qSXpNek5hRncweU56RXhNRFl4TWpJek16TmFNRjR4Q3pBSkJnTlZCQVlUQWxWVApNUlV3RXdZRFZRUUtFd3hFYVdkcFEyVnlkQ0JKYm1NeEdUQVhCZ05WQkFzVEVIZDNkeTVrYVdkcFkyVnlkQzVqCmIyMHhIVEFiQmdOVkJBTVRGRkpoY0dsa1UxTk1JRkpUUVNCRFFTQXlNREU0TUlJQklqQU5CZ2txaGtpRzl3MEIKQVFFRkFBT0NBUThBTUlJQkNnS0NBUUVBNVMyb2loRW85bm5wZXpvemlEdHg0V1dMTENsbC9lMHQxRVllbUU1bgorTWdQNXZpYUhMeStWcEhQK25kWDVEMThJTkl1dUFWOHdGcTI2S0Y1VTBXTklaaVFwNm1MdElXalVlV0RQQTI4Ck9leWhUbGo5VExrMmJleXRidEZVNnlwYnBXVWx0bXZZNVY4bmdzcEM3bkZSTkNqcGZuREVEMmtSeUp6Tzh5b0sKTUZ6NEo0SkU4TjdOQTF1SndVRUZNVXZITHMwc2NMb1Baa0tjZXdJUm0xUlYyQXhtRlF4SmtkZjdZTjlQY2traQpmMlhnbTNiNDhCWm4wemYwcVhzU2VHdTg0dWE5Z3d6anpJN3RiVEJqYXlUcFQrL1hwV3VCVnY2ZnZhckk2YmlrCktCODU5T1NHUXV3NzNYWGdldUZ3RVBIVElSb1V0a3p1My9FUStMdHd6bmtrZFFJREFRQUJvNElCWmpDQ0FXSXcKSFFZRFZSME9CQllFRkZQS0YxbjhhOEFESVM4YXJ1U3FxQnlDVnRwMU1COEdBMVVkSXdRWU1CYUFGQVBlVURWVwowVXk3WnZDajRoc2J3NWV5UGRGVk1BNEdBMVVkRHdFQi93UUVBd0lCaGpBZEJnTlZIU1VFRmpBVUJnZ3JCZ0VGCkJRY0RBUVlJS3dZQkJRVUhBd0l3RWdZRFZSMFRBUUgvQkFnd0JnRUIvd0lCQURBMEJnZ3JCZ0VGQlFjQkFRUW8KTUNZd0pBWUlLd1lCQlFVSE1BR0dHR2gwZEhBNkx5OXZZM053TG1ScFoybGpaWEowTG1OdmJUQkNCZ05WSFI4RQpPekE1TURlZ05hQXpoakZvZEhSd09pOHZZM0pzTXk1a2FXZHBZMlZ5ZEM1amIyMHZSR2xuYVVObGNuUkhiRzlpCllXeFNiMjkwUTBFdVkzSnNNR01HQTFVZElBUmNNRm93TndZSllJWklBWWI5YkFFQ01Db3dLQVlJS3dZQkJRVUgKQWdFV0hHaDBkSEJ6T2k4dmQzZDNMbVJwWjJsalpYSjBMbU52YlM5RFVGTXdDd1lKWUlaSUFZYjliQUVCTUFnRwpCbWVCREFFQ0FUQUlCZ1puZ1F3QkFnSXdEUVlKS29aSWh2Y05BUUVMQlFBRGdnRUJBSDRqeC9MS05XNVprbEZjCllXczhFamJtMG55ektlWkMyS09WWVI3UDhnZXZLeXNsV200WG80QlN6S3IyMzVGc0o0YUZ0NnlBaXYxZVkwdFoKL1pOMThiT0dTR1N0b0VjL0pFNG9jSXpyOFA1TWcxMWtSWUhibWdZbnIxUnhla2k1bVNlYjM5REd4VHBKRDRrRwpoczVsWE5vbzRjb25VaWlKd0thcUg3dmgyYmFyeWQ4cE1JU2FnODNKVXF5VkdjMnRXUHBPMDMyOS9DV3Eya3J5CnF2NjZPU01qd3VsVXowZFhmNE9IUWFzUjdDTmZJcis0S1NjYzZBQmxRNVJERjg2UEdlRTZrZHdTUWtGaUIvY1EKeXNOeXEwakVEUVRrZmEycGptdVd0TUNOYkJuaEZYQlllamZ1YkloYVViRXYyRk9RQjNkQ2F2K0ZQZzVlRXZlWApUVnlNbkdvPQotLS0tLUVORCBDRVJUSUZJQ0FURS0tLS0tCg==",
                    "enabled": false,
                    "hostname": "localhost",
                    "key": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQpNSUlFb3dJQkFBS0NBUUVBd0prMFNubm9aUWFtYVptKzFVSXY3UWIwS2pYWXlRMHJKSUhqbGxNN3pYRWFBaC9CCmpWNnUzSHVNbGtOVlJ6RUVGeEtGUm5nSlpXTVlPcy9sWDU1QmtBS0RocXNmaDl0ZVA4ZFVQU1RTVTN6bWRVc0cKU3AySDBqSm1YbkVjbzBac3BiVXJhaGVxaXFJRTJnWDBCWm9iV29YZEdzUTJ1cmJQbHIwaVVKdTFWenhIZGhNLwpnTnFQWEN5bE5YdENrbVI2NHl6SGtEaG9taUUwMkkzME1wZThEcXM5eHd5Sndxek9ydmZwNHlBNFJIZ2toNXJaCi8rUStzRVA2QjJXMEdBdzZzZDNTYytOTytYVUZoVnlFRDhqZkNsTndPVHZodi9keE5IWHZ4WHd0RWNjZTd1TVAKQUxSc0hYdUVtZE5DeTdqb2JvMm5RYU5vNkFDVzRTazlFYlVlcHdJREFRQUJBb0lCQUJDbDdaZzJlM2tvSHdCbgowdEFhWFl3dW1ySi9YeExaK1kyNVVKUHc4KzFwUm1kQ21YZC81THVZVG1tNnZrZGVIRTFtUTNENDlVNzNSdVBtCmpwdzZVTTZDK1IyTjh6YktKcFo2dzNSWmwzaXlCalhEVzNMZHRSemRVRzMvZHd0Q0RpZ3d4ZVR3ME1RVkJ6aFYKWldlS2o2eStmNE5OekxGTkZocWdsTFBGRWVMTlk3WUgvYUdHdTZwaVpERkFMT0liMlFVUUVvVVRkUEJabDMxSwpKL0Y0cHE1RkV6ekQ0aHMwYXJzSlIxeUd4L0NCRWxqQWRSNWE4Y3FHdGhxYjFMdXJTMVcxT0dNUGkzNFdDblJKCnYwTC9zSU9ORCt5cjdjd3ZQaFFzcmI3Tk9NbGxRSDYwUGdyQ0JyOUVwalFiRm1ZN3NCTS9jYUZTWHVDTmZnTzMKT3hTQjF0RUNnWUVBNHpSMTh2aXl3Ykt5aE5xeEhkNzJ1UkxHS2k4SWZhVmhsNGxFeGdsR0pLNHQ0KzYreXhObAp2OFBkdGg3elJjL1pyQUlBUEtBRnZxa1RnVzkrc2llL3JPcE9PUzA2UnVIWVpJMjdpS3ZiL054N29YTm9TM0Y2CitXRlNJSVJqRFo0eHc2SU1iNUJFS0hOVGtKU1pBMXBkaXhYeFgzeDVTM2hCWTlmekl2NGdVVDhDZ1lFQTJRSDAKbnU5REZ0MDNmWmpvcjRMUkZCQS84NEV1dmdBL2RxMkZ4TzVvMjIrdk43Qk9LRW9IWTRuTzhYdFN2L2MzaVFZdApmTUdGZUR2eXdqRnFPU05SS2JxenMza3ZmY2gxNEhHdFltT29DVTF6Smd2VkF4SytWOFF4YWxMbmhlbjZsaTgwCnJKYmtWWUJXYnFHZHU1d2FPM1RyeGsvbHBMejFvSnliK0FQbmNKa0NnWUVBeHd3anRTeEtOc1hkL2VRMXZReWoKcTdvVjM4ajZzRkRkOGNzeGRDbVZ5UGJMZjErZXlQeUNiSUVrMExQR3ZzZDhWR1FnWHZ4MmEvVm1BRVIwZzVHdApLOGpOcE92Sm5XV29sVmRTRDBjNDRBc3RTR0Y1NG5aVm5QcTAzNFg3M1dkVkE2WUpMUWYvbmt5Q3ppeFE3UTRTCjI1SHF0NU91RS84VGttd1ROdm5aeURNQ2dZQlUxV0xmendFWEt4dktTY0NNL0VMbjZRSlhHeVUrdWJJVm52Z3cKYUZNdGxaZVYyYUowTlp4MTJTWEJXVWljS3ZUVzZFZFlBZjgvWlpBZFFsd1lRaW8wZWZ4T0tRM2J5a1hxbkFYLwo5SjE0SllMaVdMWEpJY2F2WU05QWxYQlU5WWMwdHZlTWtPQnFYR1Z0a2xpclF2Y3VUeUxzdjhjeWczd3ljWWEzCmlpOHI0UUtCZ0NaODZtcEZuUlN2aTdpTWw1MUdybDFQZER2YzZVMURRVkI5empTQlVzak1lamNTQXk4WmVJZlQKUDVoSGc3REFaUEhmUGpYLzZxUnpNQW8zWlFlZVBMdTNEYVgwbFBoekNhalF1bkpIbWZvYlA2QjFOUG45T056bQo0dE0wdEU1YTJVVmJuZW5haVZyaDRhSXlJTTFxVTdHT1ArcDNMNi9hNzVnSVQ2SVJRVlloCi0tLS0tRU5EIFJTQSBQUklWQVRFIEtFWS0tLS0tCg==",
                    "sslProtocols": ""
                  },
                  "yugaware": {
                    "health": {
                      "email": "",
                      "password": "",
                      "username": ""
                    },
                    "multiTenant": false,
                    "replicas": 1,
                    "service": {
                      "enabled": true,
                      "ip": "",
                      "type": "LoadBalancer"
                    },
                    "serviceAccount": "yugaware",
                    "storage": "100Gi",
                    "storageClass": "standard"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/yugabytedb/yugabyte-platform-operator@sha256:f227c9e2c625b04160d952349ddba552ac66dae8e7da35f56746f3447c270ce3
          description: Operator for Yugabyte Platform makes it easy to deploy and
            upgrade Yugabyte Platform on OpenShift environments.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/yugabyte-platform-operator-bundle-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/yugabyte-platform-operator-bundle-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/yugabyte/yugabyte-platform-operator
          support: YugabyteDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy and manage Yugabyte Platform as a Kubernetes native
              custom resource.
            displayName: Yugabyte Platform
            kind: YBPlatform
            name: ybplatforms.yugabyte.com
            version: v1alpha1
        description: |2

          Operator for Yugabyte Platform makes it easy to deploy and upgrade Yugabyte Platform on OpenShift environments.
          Use Yugabyte Platform's orchestration and monitoring to manage YugabyteDB universes. Yugabyte Platform can create a YugabyteDB universe with many instances (VMs, pods, machines, etc., provided by IaaS), logically grouped together to form one logical distributed database. Each universe includes one or more clusters. A universe is comprised of one primary cluster and, optionally, one or more read replica clusters. All instances belonging to a cluster run on the same type of cloud provider instance type.

          ### Yugabyte Platform Features
          * **YugabyteDB-as-a-Service on Kubernetes** - Deploy your own YugabyteDB-as-a-Service running on any Kubernetes infrastructure, Yugabyte Platform provides central management for multiple database clusters and Includes 24/7 production support.
          * **Monitoring & Alerting** - Comprehensive monitoring & alerting without any configuration.
          * **Scheduled Backups** - Use Yugabyte Platform to perform a regular scheduled backup of a full YugabyteDB universe or selected tables.

          ### Getting Started
          Here are a few resources for getting started with Yugabyte platform
          * [Quick start guide for Yugabyte Platform](https://docs.yugabyte.com/latest/yugabyte-platform/deploy/) - Install Yugabyte platform on OpenShift.
          * [Quick Start for YugabyteDB](http://docs.yugabyte.com/latest/quick-start/) - install, create a local cluster and read/write from YugabyteDB.
          * [Explore core features](https://docs.yugabyte.com/latest/explore/) - automatic sharding & re-balancing, linear scalability, fault tolerance, tunable reads etc.
          * [Ecosystem integrations](https://docs.yugabyte.com/latest/develop/ecosystem-integrations/) - integrations with Apache Kafka/KSQL, Apache Spark, JanusGraph, KairosDB, Presto and more.
          * [Real world apps](https://docs.yugabyte.com/latest/develop/realworld-apps/) - sample real-world, end-to-end applications built using YugabyteDB.

          ### Help
          Cannot find what you are looking for? Have a question? We love to hear from you - [please file a GitHub issue](https://github.com/YugaByte/yugabyte-db/issues) or talk to us live at [Slack](https://www.yugabyte.com/slack).
        displayName: Yugabyte Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - distributed sql
        links:
        - name: Yugabyte Platform Operator
          url: https://www.yugabyte.com/
        maintainers:
        - email: oss-maintainers@yugabyte.com
          name: Yugabyte
        maturity: alpha
        provider:
          name: Yugabyte Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9211b70edbb93d718428e1c9a992e5fc74a20cb213f8a0592ffe340c362890fe
        - registry.marketplace.redhat.com/rhm/yugabytedb/yugabyte-platform-operator@sha256:f227c9e2c625b04160d952349ddba552ac66dae8e7da35f56746f3447c270ce3
        version: 0.0.6
      entries:
      - name: yugabyte-platform-operator.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: yugabyte-platform-operator-bundle-rhmp
    provider:
      name: Yugabyte Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: cinder-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cinder-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cinder.openstack.org/v1beta1",
                "kind": "Cinder",
                "metadata": {
                  "name": "cinder",
                  "namespace": "openstack"
                },
                "spec": {
                  "cinderAPI": {},
                  "cinderBackup": {},
                  "cinderScheduler": {},
                  "cinderVolumes": {
                    "volume1": {}
                  },
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "cinder",
                  "databaseInstance": "openstack",
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "cinder-secret",
                  "serviceUser": "cinder"
                }
              },
              {
                "apiVersion": "cinder.openstack.org/v1beta1",
                "kind": "CinderAPI",
                "metadata": {
                  "name": "cinderapi-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cinder.openstack.org/v1beta1",
                "kind": "CinderBackup",
                "metadata": {
                  "name": "cinderbackup-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cinder.openstack.org/v1beta1",
                "kind": "CinderScheduler",
                "metadata": {
                  "name": "cinderscheduler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cinder.openstack.org/v1beta1",
                "kind": "CinderVolume",
                "metadata": {
                  "name": "cindervolume-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:43Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CinderAPI is the Schema for the cinderapis API
            displayName: Cinder API
            kind: CinderAPI
            name: cinderapis.cinder.openstack.org
            version: v1beta1
          - description: CinderBackup is the Schema for the cinderbackups API
            displayName: Cinder Backup
            kind: CinderBackup
            name: cinderbackups.cinder.openstack.org
            version: v1beta1
          - description: Cinder is the Schema for the cinders API
            displayName: Cinder
            kind: Cinder
            name: cinders.cinder.openstack.org
            version: v1beta1
          - description: CinderScheduler is the Schema for the cinderschedulers API
            displayName: Cinder Scheduler
            kind: CinderScheduler
            name: cinderschedulers.cinder.openstack.org
            version: v1beta1
          - description: CinderVolume is the Schema for the cindervolumes API
            displayName: Cinder Volume
            kind: CinderVolume
            name: cindervolumes.cinder.openstack.org
            version: v1beta1
        description: Cinder Operator
        displayName: Cinder Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Block Storage
        - Cinder
        links:
        - name: Cinder Operator
          url: https://github.com/openstack-k8s-operators/cinder-operator
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/cinder-rhel9-operator@sha256:e0e457500af9db0831c7080b10e010d11213cd800d89e40683974aaf7966a68d
        - registry.redhat.io/rhoso/openstack-cinder-api-rhel9@sha256:ec6835efb7ffc4fa7eed10bd029440263237e8f792e7e43d4cda337d00fecdb6
        - registry.redhat.io/rhoso/openstack-cinder-backup-rhel9@sha256:936e10d80300c325d0f58b4d6ca290fcf3c5cad774eb5cb696330a20c0151561
        - registry.redhat.io/rhoso/openstack-cinder-scheduler-rhel9@sha256:753775cc575321c727b9a6371daedc8774ba3a2287c96f3d44ad8f7025f63841
        - registry.redhat.io/rhoso/openstack-cinder-volume-rhel9@sha256:47b19d993bdfc790c508861dbcc87f0ec51302db8975698af8e1476a0f5209c7
        version: 1.0.6
      entries:
      - name: cinder-operator.v1.0.6
        version: 1.0.6
      - name: cinder-operator.v1.0.5
        version: 1.0.5
      - name: cinder-operator.v1.0.4
        version: 1.0.4
      - name: cinder-operator.v1.0.3
        version: 1.0.3
      - name: cinder-operator.v1.0.2
        version: 1.0.2
      - name: cinder-operator.v1.0.1
        version: 1.0.1
      - name: cinder-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: cinder-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: node-healthcheck-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-healthcheck-operator.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:60d95ea8240ecf701ba974b6df3ee296a7fb3ef5393e0dcf43af15e782f2d153
          createdAt: "2023-11-02 11:44:50"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.3.0 <0.4.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://docs.openshift.com/container-platform/4.12/nodes/nodes/eco-node-health-check-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:60d95ea8240ecf701ba974b6df3ee296a7fb3ef5393e0dcf43af15e782f2d153
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:8d32dee5c72ec1c94e4e92f8d1da935a384e40e379f237c432086a2781f4ecb9
        version: 0.4.1
      entries:
      - name: node-healthcheck-operator.v0.4.1
        version: 0.4.1
      name: 4.12-eus
    - currentCSV: node-healthcheck-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:90cceb176fa903cc8e9aa4c0f7316560a8a354a3138dc26abbfc9884c1c576b5
          createdAt: "2023-10-30 14:21:41"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.4.0 <0.6.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/node-health-check-operator
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:90cceb176fa903cc8e9aa4c0f7316560a8a354a3138dc26abbfc9884c1c576b5
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:b8d5db068bfc81031c2eef86e8649e4fc008b7480e08c9cfab2ab4497098eecf
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel8@sha256:85054b34e119c16cc95cda9dc3f5d63f2d7bfc70634f70973410b2a1b1fa58d4
        version: 0.6.1
      entries:
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.14-eus
    - currentCSV: node-healthcheck-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:4811046e26846b1f12a3b20ddc6209980897c11b171ddd61d26b43630cc1d4b4
          createdAt: "2024-08-09T14:54:37Z"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.8.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          SNR is installed automatically when installing NHC.

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/node-health-check-operator
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:4811046e26846b1f12a3b20ddc6209980897c11b171ddd61d26b43630cc1d4b4
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:942f69efc7ae05d6c4d9c33335e4be71bfae2e6f9ef062dcd7887eebfe02120e
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel8@sha256:062cd9def39a087ae3802cd5208e2f0d5bfca29ba76ebd5139d88348e8c9e923
        version: 0.8.2
      entries:
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.16-eus
    - currentCSV: node-healthcheck-operator.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:f174ada653bb5fb7e58219510d0b10d868273a70db5c486e0b3ad05a1be50640
          createdAt: "2025-08-04 12:38:05"
          description: Detect failed Nodes and trigger remediation with a remediation
            operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <0.9.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.7/html/remediation_fencing_and_maintenance
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel9@sha256:8e9926558c3d8082f298e9e5f58bc5b473020acd44c2ce853c2916cdad71d649
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:f174ada653bb5fb7e58219510d0b10d868273a70db5c486e0b3ad05a1be50640
        - registry.redhat.io/workload-availability/node-remediation-console-rhel9@sha256:37f62407284fe574b973a1a685371d191b710ad70a6bfe8172f19a4568c11cf6
        version: 0.9.1
      entries:
      - name: node-healthcheck-operator.v0.9.1
        version: 0.9.1
      - name: node-healthcheck-operator.v0.9.0
        version: 0.9.0
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.18-eus
    - currentCSV: node-healthcheck-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:8e1dbc35126101e1025bd4c0eac4f4c614eb779ccafbb55050541a27ba7e9706
          createdAt: "2022-08-15 12:34:34"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.2.0 <0.3.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-node-health-check-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b5786bbbef725badf3dfcc2c2c7a86ead5ebb584c978c47aae8b9a62e241b80d
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:8e1dbc35126101e1025bd4c0eac4f4c614eb779ccafbb55050541a27ba7e9706
        version: 0.3.1
      entries:
      - name: node-healthcheck-operator.v0.3.1
        version: 0.3.1
      name: candidate
    - currentCSV: node-healthcheck-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:2160f2688e5b098c0bae7597cebcd380019b82ae02f4506d07c2ed85f0b3664d
          createdAt: "2025-09-01 18:07:56"
          description: Detect failed Nodes and trigger remediation with a remediation
            operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.8/html/remediation_fencing_and_maintenance
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:2160f2688e5b098c0bae7597cebcd380019b82ae02f4506d07c2ed85f0b3664d
        - registry.redhat.io/workload-availability/node-remediation-console-rhel9@sha256:cc0f671b126cde10476f8cc2061c6903932e7bf5e3dbf4e8bdda9add70d5deea
        version: 0.10.0
      entries:
      - name: node-healthcheck-operator.v0.10.0
        version: 0.10.0
      - name: node-healthcheck-operator.v0.9.1
        version: 0.9.1
      - name: node-healthcheck-operator.v0.9.0
        version: 0.9.0
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: stable
    defaultChannel: stable
    packageName: node-healthcheck-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-mq-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-mq-controller.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mq.services.k8s.aws/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/mq-controller:1.1.2
          createdAt: "2025-10-03T18:49:51Z"
          description: AWS MQ controller is a service controller for managing MQ resources
            in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker represents the state of an AWS mq Broker resource.
            displayName: Broker
            kind: Broker
            name: brokers.mq.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon MQ resources in AWS from within your Kubernetes cluster.

          **About Amazon MQ**

          Amazon MQ is a managed message broker service for Apache ActiveMQ and RabbitMQ that makes it easy to set up and operate message brokers on AWS. Amazon MQ reduces your operational responsibilities by managing the provisioning, setup, and maintenance of message brokers for you. Because Amazon MQ connects to your current applications with industry-standard APIs and protocols, you can easily migrate to AWS without having to rewrite code.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon MQ
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mq
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon MQ Developer Resources
          url: https://docs.aws.amazon.com/amazon-mq/latest/developer-guide/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: mq maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/mq-controller:1.1.2
        version: 1.1.2
      entries:
      - name: ack-mq-controller.v1.1.2
        version: 1.1.2
      - name: ack-mq-controller.v1.1.1
        version: 1.1.1
      - name: ack-mq-controller.v1.1.0
        version: 1.1.0
      - name: ack-mq-controller.v1.0.13
        version: 1.0.13
      - name: ack-mq-controller.v1.0.11
        version: 1.0.11
      - name: ack-mq-controller.v1.0.10
        version: 1.0.10
      - name: ack-mq-controller.v1.0.9
        version: 1.0.9
      - name: ack-mq-controller.v1.0.8
        version: 1.0.8
      - name: ack-mq-controller.v1.0.7
        version: 1.0.7
      - name: ack-mq-controller.v1.0.6
        version: 1.0.6
      - name: ack-mq-controller.v1.0.5
        version: 1.0.5
      - name: ack-mq-controller.v1.0.2
        version: 1.0.2
      - name: ack-mq-controller.v1.0.1
        version: 1.0.1
      - name: ack-mq-controller.v1.0.0
        version: 1.0.0
      - name: ack-mq-controller.v0.0.36
        version: 0.0.36
      - name: ack-mq-controller.v0.0.35
        version: 0.0.35
      - name: ack-mq-controller.v0.0.34
        version: 0.0.34
      - name: ack-mq-controller.v0.0.31
        version: 0.0.31
      - name: ack-mq-controller.v0.0.29
        version: 0.0.29
      - name: ack-mq-controller.v0.0.27
        version: 0.0.27
      - name: ack-mq-controller.v0.0.26
        version: 0.0.26
      - name: ack-mq-controller.v0.0.25
        version: 0.0.25
      - name: ack-mq-controller.v0.0.24
        version: 0.0.24
      - name: ack-mq-controller.v0.0.23
        version: 0.0.23
      - name: ack-mq-controller.v0.0.22
        version: 0.0.22
      - name: ack-mq-controller.v0.0.21
        version: 0.0.21
      - name: ack-mq-controller.v0.0.20
        version: 0.0.20
      - name: ack-mq-controller.v0.0.19
        version: 0.0.19
      - name: ack-mq-controller.v0.0.18
        version: 0.0.18
      - name: ack-mq-controller.v0.0.17
        version: 0.0.17
      - name: ack-mq-controller.v0.0.16
        version: 0.0.16
      - name: ack-mq-controller.v0.0.15
        version: 0.0.15
      - name: ack-mq-controller.v0.0.14
        version: 0.0.14
      - name: ack-mq-controller.v0.0.13
        version: 0.0.13
      - name: ack-mq-controller.v0.0.12
        version: 0.0.12
      - name: ack-mq-controller.v0.0.11
        version: 0.0.11
      - name: ack-mq-controller.v0.0.10
        version: 0.0.10
      - name: ack-mq-controller.v0.0.9
        version: 0.0.9
      - name: ack-mq-controller.v0.0.8
        version: 0.0.8
      - name: ack-mq-controller.v0.0.7
        version: 0.0.7
      name: alpha
    defaultChannel: alpha
    packageName: ack-mq-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: volsync-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: volsync-product.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
          createdAt: "2025-08-27T13:12:05Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.13.1'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
        version: 0.13.1
      entries:
      - name: volsync-product.v0.13.1
        version: 0.13.1
      - name: volsync-product.v0.13.0
        version: 0.13.0
      - name: volsync-product.v0.12.2
        version: 0.12.2
      - name: volsync-product.v0.12.1
        version: 0.12.1
      - name: volsync-product.v0.12.1-0.1749677898.p
        version: 0.12.1+0.1749677898.p
      - name: volsync-product.v0.12.0
        version: 0.12.0
      name: stable
    - currentCSV: volsync-product.v0.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:d345825e953c4c170db45a5d92639182b7ee18933287aaeea035f2a2bad13c37
          createdAt: 09 Jan 2025, 00:47
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.10.2'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:d345825e953c4c170db45a5d92639182b7ee18933287aaeea035f2a2bad13c37
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 0.10.2
      entries:
      - name: volsync-product.v0.10.2
        version: 0.10.2
      - name: volsync-product.v0.10.1
        version: 0.10.1
      - name: volsync-product.v0.10.1-0.1730236570.p
        version: 0.10.1+0.1730236570.p
      - name: volsync-product.v0.10.0
        version: 0.10.0
      name: stable-0.10
    - currentCSV: volsync-product.v0.11.2-0.1749674925.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:9da8572c79b75767952cf699e5b1699b9f46d83586aa2d6a17fb23c092713b62
          createdAt: 16 Apr 2025, 06:42
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.11.2'
          olm.substitutesFor: volsync-product.v0.11.2
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:9da8572c79b75767952cf699e5b1699b9f46d83586aa2d6a17fb23c092713b62
        version: 0.11.2+0.1749674925.p
      entries:
      - name: volsync-product.v0.11.2
        version: 0.11.2
      - name: volsync-product.v0.11.2-0.1749674925.p
        version: 0.11.2+0.1749674925.p
      - name: volsync-product.v0.11.1
        version: 0.11.1
      - name: volsync-product.v0.11.0
        version: 0.11.0
      name: stable-0.11
    - currentCSV: volsync-product.v0.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:7d1ac69fe3d0fddc66acdfce9361a4fc8dd3f549f66c98450f7a0e4c5cc893a0
          createdAt: "2025-07-07T13:46:18Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.12.2'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:7d1ac69fe3d0fddc66acdfce9361a4fc8dd3f549f66c98450f7a0e4c5cc893a0
        version: 0.12.2
      entries:
      - name: volsync-product.v0.12.2
        version: 0.12.2
      - name: volsync-product.v0.12.1
        version: 0.12.1
      - name: volsync-product.v0.12.1-0.1749677898.p
        version: 0.12.1+0.1749677898.p
      - name: volsync-product.v0.12.0
        version: 0.12.0
      name: stable-0.12
    - currentCSV: volsync-product.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
          createdAt: "2025-08-27T13:12:05Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.13.1'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
        version: 0.13.1
      entries:
      - name: volsync-product.v0.13.1
        version: 0.13.1
      - name: volsync-product.v0.13.0
        version: 0.13.0
      name: stable-0.13
    - currentCSV: volsync-product.v0.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:ef44675fe70ee9ec8109fb91c6559f0c5518b45219e19175a18a5004e40b65f7
          createdAt: 22 Aug 2024, 22:06
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.9.3'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ReplicationDestination defines the destination for a replicated
              volume
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: ReplicationSource defines the source for a replicated volume
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:ef44675fe70ee9ec8109fb91c6559f0c5518b45219e19175a18a5004e40b65f7
        version: 0.9.3
      entries:
      - name: volsync-product.v0.9.3
        version: 0.9.3
      - name: volsync-product.v0.9.2
        version: 0.9.2
      - name: volsync-product.v0.9.1
        version: 0.9.1
      - name: volsync-product.v0.9.0
        version: 0.9.0
      name: stable-0.9
    defaultChannel: stable
    packageName: volsync-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: ""
    name: gitops-primer
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitops-primer.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "primer.gitops.io/v1alpha1",
                "kind": "Export",
                "metadata": {
                  "name": "primer"
                },
                "spec": {
                  "method": "download"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/migtools/gitops-primer:v0.0.11
          operatorframework.io/suggested-namespace: gitops-primer-system
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Export is the Schema for the exports API
            displayName: Export
            kind: Export
            name: exports.primer.gitops.io
            version: v1alpha1
        description: Export and normalize Kubernetes YAML files for GitOps
        displayName: gitops-primer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - operator
        links:
        - name: Gitops Primer
          url: https://github.com/migtools/gitops-primer
        maturity: alpha
        provider:
          name: Konveyor
        relatedImages:
        - quay.io/migtools/gitops-primer-export:v0.0.11
        - quay.io/openshift/origin-oauth-proxy:4.9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6d57bfd91fac9b68eb72d27226bc297472ceb136c996628b845ecc54a48b31cb
        - quay.io/migtools/gitops-primer:v0.0.11
        - quay.io/migtools/gitops-primer-webhook:v0.0.11
        - quay.io/migtools/gitops-primer-downloader:v0.0.11
        version: 0.0.11
      entries:
      - name: gitops-primer.v0.0.11
        version: 0.0.11
      - name: gitops-primer.v0.0.10
        version: 0.0.10
      - name: gitops-primer.v0.0.9
        version: 0.0.9
      - name: gitops-primer.v0.0.8
        version: 0.0.8
      - name: gitops-primer.v0.0.7
        version: 0.0.7
      - name: gitops-primer.v0.0.6
        version: 0.0.6
      - name: gitops-primer.v0.0.5
        version: 0.0.5
      - name: gitops-primer.v0.0.4
        version: 0.0.4
      - name: gitops-primer.v0.0.3
        version: 0.0.3
      - name: gitops-primer.v0.0.2
        version: 0.0.2
      - name: gitops-primer.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: gitops-primer
    provider:
      name: Konveyor
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Craig Trought
      provider-url: ""
    name: prometheus
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometheusoperator.0.56.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "Prometheus",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "replicas": 2,
                  "serviceAccountName": "prometheus-k8s",
                  "serviceMonitorSelector": {},
                  "ruleSelector": {},
                  "podMonitorSelector": {},
                  "probeSelector": {},
                  "alerting": {
                    "alertmanagers": [
                      {
                        "namespace": "monitoring",
                        "name": "alertmanager-main",
                        "port": "web"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "ServiceMonitor",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "operated-prometheus": "true"
                    }
                  },
                  "endpoints": [
                    {
                      "port": "web",
                      "interval": "30s"
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "PodMonitor",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "app": "prometheus"
                    }
                  },
                  "podMetricsEndpoints": [
                    {
                      "port": "web",
                      "interval": "30s"
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "Alertmanager",
                "metadata": {
                  "name": "alertmanager-main"
                },
                "spec": {
                  "replicas": 3,
                  "alertmanagerConfigSelector": {}
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "PrometheusRule",
                "metadata": {
                  "name": "prometheus-example-rules"
                },
                "spec": {
                  "groups": [
                    {
                      "name": "./example.rules",
                      "rules": [
                        {
                          "alert": "ExampleAlert",
                          "expr": "vector(1)"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1alpha1",
                "kind": "AlertmanagerConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "receivers": [
                    {
                      "name": "example"
                    }
                  ],
                  "route": {
                    "receiver": "example"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/prometheus-operator/prometheus-operator:v0.56.3
          createdAt: "2022-05-24T9:40:00Z"
          description: Manage the full lifecycle of configuring and managing Prometheus
            and Alertmanager servers.
          repository: https://github.com/prometheus-operator/prometheus-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A running Prometheus instance
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.coreos.com
            version: v1
          - description: A Prometheus Rule configures groups of sequentially evaluated
              recording and alerting rules.
            displayName: Prometheus Rule
            kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - description: Configures prometheus to monitor a particular k8s service
            displayName: Service Monitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - description: Configures prometheus to monitor a particular pod
            displayName: Pod Monitor
            kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
          - description: Configures an Alertmanager for the namespace
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.coreos.com
            version: v1
          - description: Configures subsections of Alertmanager configuration specific
              to each namespace
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.coreos.com
            version: v1alpha1
          - description: Defines how groups of ingresses and static targets should
              be monitored.
            displayName: Probe
            kind: Probe
            name: probes.monitoring.coreos.com
            version: v1
          - description: Allows recording and alerting rules to be processed across
              multiple Prometheus instances.
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.coreos.com
            version: v1
        description: |
          The Prometheus Operator for Kubernetes provides easy monitoring definitions for Kubernetes services and deployment and management of Prometheus instances.

          Once installed, the Prometheus Operator provides the following features:

          * **Create/Destroy**: Easily launch a Prometheus instance for your Kubernetes namespace, a specific application or team easily using the Operator.
          * **Simple Configuration**: Configure the fundamentals of Prometheus like versions, persistence, retention policies, and replicas from a native Kubernetes resource.
          * **Target Services via Labels**: Automatically generate monitoring target configurations based on familiar Kubernetes label queries; no need to learn a Prometheus specific configuration language.

          ### Other Supported Features

          **High availability**

          Multiple instances are run across failure zones and data is replicated. This keeps your monitoring available during an outage, when you need it most.

          **Updates via automated operations**

          New Prometheus versions are deployed using a rolling update with no downtime, making it easy to stay up to date.

          **Handles the dynamic nature of containers**

          Alerting rules are attached to groups of containers instead of individual instances, which is ideal for the highly dynamic nature of container deployment.
        displayName: Prometheus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - prometheus
        - monitoring
        - tsdb
        - alerting
        - observability
        links:
        - name: Prometheus
          url: https://www.prometheus.io/
        - name: Documentation
          url: https://github.com/prometheus-operator/prometheus-operator/tree/master/Documentation
        - name: Prometheus Operator
          url: https://github.com/prometheus-operator/prometheus-operator
        maintainers:
        - email: k8s@trought.ca
          name: Craig Trought
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Craig Trought
        relatedImages:
        - quay.io/prometheus-operator/prometheus-operator:v0.56.3
        version: 0.56.3
      entries:
      - name: prometheusoperator.0.56.3
        version: 0.56.3
      - name: prometheusoperator.0.47.0
        version: 0.47.0
      name: beta
    defaultChannel: beta
    packageName: prometheus
    provider:
      name: Craig Trought
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC
      provider-url: ""
    name: nspc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nspc-operator.v1.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.nec.com/v1",
                "kind": "NSPC",
                "metadata": {
                  "name": "nspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.connect.redhat.com/nec/nspc-operator@sha256:50fe10f05d8aa7db1b97280e5842bb0b67c6f9bfb01c0d17c5be3cd11b78e853
          createdAt: "2025-07-24T00:00:00Z"
          description: An operator for managing NEC Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: NEC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NSPC is the Schema for the nspcs API
            displayName: NSPC
            kind: NSPC
            name: nspcs.csi.nec.com
            version: v1
        description: |-
          ## About
          NEC Storage Plug-in for Containers (NSPC) is a plugin that integrates NEC Storage V Series into Kubernetes based clusters.
          NSPC provides persistent volume provisioning capabilities from NEC Storage V Series arrays.

          For full documentation, refer to the reference guide.

          ## Requirements

          ### Supported Driver Version

          * NSPC v3.16.0
        displayName: NEC Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - NEC
        - Storage
        - iStorage
        - V Series
        - Container Storage Interface
        - CSI
        maturity: stable
        provider:
          name: NEC
        relatedImages:
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.connect.redhat.com/nec/nspc-csi-driver@sha256:e50d60e025b3c8ae522a8c5ae78a07f8a624a1210ed7a078c8084efd76ccc646
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:81d32545fb1b2c319bfeb36a70db4b4d3aea9c51c0574b209b8f64750b32d2f0
        - registry.connect.redhat.com/nec/nspc-operator@sha256:50fe10f05d8aa7db1b97280e5842bb0b67c6f9bfb01c0d17c5be3cd11b78e853
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        version: 1.16.0
      entries:
      - name: nspc-operator.v1.16.0
        version: 1.16.0
      name: stable
    defaultChannel: stable
    packageName: nspc-operator
    provider:
      name: NEC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: tempo-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tempo-operator.v0.18.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoMonolithic",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "jaegerui": {
                    "enabled": true,
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    }
                  },
                  "storage": {
                    "traces": {
                      "backend": "memory"
                    }
                  }
                }
              },
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoStack",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "resources": {
                    "total": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2Gi"
                      }
                    }
                  },
                  "storage": {
                    "secret": {
                      "name": "my-storage-secret",
                      "type": "s3"
                    }
                  },
                  "storageSize": "1Gi",
                  "template": {
                    "queryFrontend": {
                      "jaegerQuery": {
                        "enabled": true,
                        "ingress": {
                          "type": "route"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/rhosdt/tempo-rhel8-operator@sha256:9bfa5a2ffb09b501f001a4275379210621f06620adcf6454a9327c65538872f5
          createdAt: 22 Sep 2025, 11:15
          description: Create and manage deployments of Tempo, a high-scale distributed
            tracing backend.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.18.0-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-tempo-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/tempo-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TempoMonolithic manages a Tempo deployment in monolithic
              mode.
            displayName: TempoMonolithic
            kind: TempoMonolithic
            name: tempomonolithics.tempo.grafana.com
            version: v1alpha1
          - description: TempoStack manages a Tempo deployment in microservices mode.
            displayName: TempoStack
            kind: TempoStack
            name: tempostacks.tempo.grafana.com
            version: v1alpha1
        description: |-
          Red Hat OpenShift distributed tracing platform based on Tempo. Tempo is an open-source, easy-to-use, and highly scalable distributed tracing backend. It provides observability for microservices architectures by allowing developers to track requests as they flow through distributed systems. Tempo is optimized to handle large volumes of trace data and is designed to be highly performant even under heavy loads.
          It can ingest common open source tracing protocols including Jaeger, Zipkin, and OpenTelemetry and requires only object storage to operate.

          The Tempo Operator supports Tempo deployments in Microservices mode (`TempoStack` CR) and Monolithic mode (`TempoMonolithic` CR).

          ### Core capabilities
          Tempo is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed transaction monitoring
          * Root cause analysis
          * Performance / latency optimization

          ### Operator features
          * **Resource Limits** - Specify overall resource requests and limits in the `TempoStack` CR; the operator assigns fractions of it to each component
          * **AuthN and AuthZ** - Supports OpenID Control (OIDC) and role-based access control (RBAC)
          * **Managed upgrades** - Updating the operator will automatically update all managed Tempo clusters
          * **Multitenancy** - Multiple tenants can send traces to the same Tempo cluster
          * **mTLS** - Communication between the Tempo components can be secured via mTLS
          * **Jaeger UI** - Traces can be visualized in Jaeger UI and exposed via Ingress or OpenShift Route
          * **Observability** - The operator and `TempoStack` operands expose telemetry (metrics, traces) and integrate with Prometheus `ServiceMonitor` and `PrometheusRule`

          ### Before you start
          `TempoStack` requires object storage to store its traces. `TempoMonolithic` can store traces in-memory, in a Persistent Volume and in object storage.
          Please ensure that your system has a compatible object storage solution that is supported, such as OpenShift Data Foundation, Minio, AWS S3, Azure Storage, or Google Cloud Storage.

          ### Examples
          There is a list of examples to help you deploy different configurations in [this GitHub repository](https://github.com/os-observability/redhat-rhosdt-samples).

          ### Support & Troubleshooting
          Tempo Operator is available and supported as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/distributed_tracing) is recommended when installing, configuring, and managing the Operator and instances.
        displayName: Tempo Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tempo
        - tracing
        - observability
        - monitoring
        - database
        links:
        - name: Tempo Operator
          url: https://github.com/grafana/tempo-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6bcfa981a7a75c242a16afa6dab36bf6e3b8a1442fa863875919ad551f48eceb
        - registry.redhat.io/rhosdt/tempo-rhel8@sha256:530a0ef3aaea3f07d6210a683a771c416cea8ce5accac8f5f751dcd7232cb341
        - registry.redhat.io/rhosdt/tempo-query-rhel8@sha256:5b89188bc587a81af3ecdca3caabd6e48356e9ba4f9070860f5160544ce30eb9
        - registry.redhat.io/rhosdt/tempo-rhel8-operator@sha256:9bfa5a2ffb09b501f001a4275379210621f06620adcf6454a9327c65538872f5
        - registry.redhat.io/rhosdt/tempo-jaeger-query-rhel8@sha256:53b62fea39ce211842026e8365d61f579b8e01e8587a1ce639423d72f3b44f45
        - registry.redhat.io/rhosdt/tempo-gateway-rhel8@sha256:095ec83219660bb1955758b94a9431b3add572da4dc2cba61b42b59fa6b4bba0
        - registry.redhat.io/rhosdt/tempo-gateway-opa-rhel8@sha256:22e876837d00371d24feac3f989ead3d4f034400b292f043b18f3839bfde564a
        version: 0.18.0-1
      entries:
      - name: tempo-operator.v0.18.0-1
        version: 0.18.0-1
      - name: tempo-operator.v0.16.0-2
        version: 0.16.0-2
      - name: tempo-operator.v0.16.0-1
        version: 0.16.0-1
      - name: tempo-operator.v0.15.4-1
        version: 0.15.4-1
      - name: tempo-operator.v0.15.3-1
        version: 0.15.3-1
      - name: tempo-operator.v0.14.1-2
        version: 0.14.1-2
      - name: tempo-operator.v0.14.1-1
        version: 0.14.1-1
      - name: tempo-operator.v0.13.0-1-0.1730817419.p
        version: 0.13.0-1+0.1730817419.p
      - name: tempo-operator.v0.13.0-1
        version: 0.13.0-1
      - name: tempo-operator.v0.10.0-8
        version: 0.10.0-8
      - name: tempo-operator.v0.10.0-7
        version: 0.10.0-7
      - name: tempo-operator.v0.10.0-6
        version: 0.10.0-6
      - name: tempo-operator.v0.8.0-3
        version: 0.8.0-3
      - name: tempo-operator.v0.8.0-2
        version: 0.8.0-2
      - name: tempo-operator.v0.6.0-2
        version: 0.6.0-2
      - name: tempo-operator.v0.3.1-3
        version: 0.3.1-3
      - name: tempo-operator.v0.1.0-6
        version: 0.1.0-6
      - name: tempo-operator.v0.1.0-6-0.1687199947.p
        version: 0.1.0-6+0.1687199947.p
      name: stable
    defaultChannel: stable
    packageName: tempo-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: deployment-validation-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: deployment-validation-operator.v0.7.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Application Runtime, Monitoring, Security
          certified: "false"
          containerImage: quay.io/deployment-validation-operator/dv-operator:0.7.12
          createdAt: "2025-09-11T00:00:00Z"
          description: The deployment validation operator
          ignore-check.kube-linter.io/minimum-three-replicas: This deployment uses
            1 pod as currently replicating does not replicate metric data causing
            installation issues
          repository: https://github.com/app-sre/deployment-validation-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Deployment Validation Operator (DVO) checks deployments and other resources against a curated collection of best practices.
          These best practices focus mainly on ensuring that the applications are fault-tolerant. DVO reports failed validations via Prometheus metrics.
          If the best-practice check has failed, the metrics will report `1`.
        displayName: Deployment Validation Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: false
          type: MultiNamespace
        keywords:
        - dvo
        links:
        - name: repository
          url: https://github.com/app-sre/deployment-validation-operator
        - name: containerImage
          url: https://quay.io/deployment-validation-operator/dv-operator:0.7.12
        maintainers:
        - email: dvo-owners@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/deployment-validation-operator/dv-operator:0.7.12
        version: 0.7.12
      entries:
      - name: deployment-validation-operator.v0.7.12
        version: 0.7.12
      - name: deployment-validation-operator.v0.7.9
        version: 0.7.9
      - name: deployment-validation-operator.v0.7.8
        version: 0.7.8
      - name: deployment-validation-operator.v0.7.7
        version: 0.7.7
      - name: deployment-validation-operator.v0.7.6
        version: 0.7.6
      - name: deployment-validation-operator.v0.7.5
        version: 0.7.5
      - name: deployment-validation-operator.v0.7.4
        version: 0.7.4
      - name: deployment-validation-operator.v0.7.3
        version: 0.7.3
      - name: deployment-validation-operator.v0.7.2
        version: 0.7.2
      - name: deployment-validation-operator.v0.7.1
        version: 0.7.1
      - name: deployment-validation-operator.v0.7.0
        version: 0.7.0
      - name: deployment-validation-operator.v0.6.0
        version: 0.6.0
      - name: deployment-validation-operator.v0.5.0
        version: 0.5.0
      - name: deployment-validation-operator.v0.4.0
        version: 0.4.0
      - name: deployment-validation-operator.v0.3.0
        version: 0.3.0
      - name: deployment-validation-operator.v0.2.2
        version: 0.2.2
      name: alpha
    defaultChannel: alpha
    packageName: deployment-validation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Google
      provider-url: https://cloud.google.com/alloydb/omni
    name: alloydb-omni-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: alloydb-omni-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "backupPlanRef": "backupplan1",
                  "dbclusterRef": "dbcluster-sample",
                  "manual": true,
                  "physicalBackupSpec": {
                    "backupType": "full"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "BackupPlan",
                "metadata": {
                  "name": "backupplan1"
                },
                "spec": {
                  "backupRetainDays": 14,
                  "backupSchedules": {
                    "full": "0 0 * * 0",
                    "incremental": "0 21 * * *"
                  },
                  "dbclusterRef": "dbcluster-sample",
                  "paused": false
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "clone1"
                },
                "spec": {
                  "clonedDBClusterConfig": {
                    "dbclusterName": "new-dbcluster-sample"
                  },
                  "pointInTime": "2024-02-23T19:59:43Z",
                  "sourceDBCluster": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "DBCluster",
                "metadata": {
                  "name": "dbcluster-sample"
                },
                "spec": {
                  "databaseVersion": "16.3.0",
                  "primarySpec": {
                    "adminUser": {
                      "passwordRef": {
                        "name": "db-pw-dbcluster-sample"
                      }
                    },
                    "resources": {
                      "cpu": 1,
                      "disks": [
                        {
                          "name": "DataDisk",
                          "size": "10Gi"
                        }
                      ],
                      "memory": "5Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "DBInstance",
                "metadata": {
                  "name": "dbcluster-sample-rp-1"
                },
                "spec": {
                  "dbcParent": {
                    "name": "dbcluster-sample"
                  },
                  "instanceType": "ReadPool",
                  "nodeCount": 2,
                  "resources": {
                    "cpu": 2,
                    "disks": [
                      {
                        "name": "DataDisk",
                        "size": "15Gi"
                      }
                    ],
                    "memory": "6Gi"
                  },
                  "schedulingconfig": {
                    "nodeaffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "preference": {
                            "matchExpressions": [
                              {
                                "key": "another-node-label-key",
                                "operator": "In",
                                "values": [
                                  "another-node-label-value"
                                ]
                              }
                            ]
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "podAffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "podAffinityTerm": {
                            "labelSelector": {
                              "matchExpressions": [
                                {
                                  "key": "app",
                                  "operator": "In",
                                  "values": [
                                    "store"
                                  ]
                                }
                              ]
                            },
                            "topologyKey": "kubernetes.io/hostname"
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "podAntiAffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "podAffinityTerm": {
                            "labelSelector": {
                              "matchExpressions": [
                                {
                                  "key": "security",
                                  "operator": "In",
                                  "values": [
                                    "S1"
                                  ]
                                }
                              ]
                            },
                            "topologyKey": "topology.kubernetes.io/zone"
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane",
                        "operator": "Exists"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Failover",
                "metadata": {
                  "name": "failover-sample"
                },
                "spec": {
                  "dbclusterRef": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "PgBouncer",
                "metadata": {
                  "name": "mypgbouncer"
                },
                "spec": {
                  "allowSuperUserAccess": true,
                  "dbclusterRef": "dbcluster-sample",
                  "parameters": {
                    "default_pool_size": "15",
                    "ignore_startup_parameters": "extra_float_digits",
                    "max_client_conn": "800",
                    "max_db_connections": "160",
                    "pool_mode": "transaction"
                  },
                  "podSpec": {
                    "image": "gcr.io/alloydb-omni/operator/g-pgbouncer:1.5.0",
                    "resources": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "replicaCount": 1,
                  "serviceOptions": {
                    "type": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Replication",
                "metadata": {
                  "name": "replication-downstream-sample"
                },
                "spec": {
                  "dbcluster": {
                    "name": "dbcluster-sample"
                  },
                  "downstream": {
                    "control": "setup",
                    "host": "10.10.10.10",
                    "password": {
                      "name": "ha-rep-pw-dbcluster-sample"
                    },
                    "port": 5432,
                    "replicationSlotName": "dbcluster_sample_replication_upstream_sample",
                    "username": "alloydbreplica"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Replication",
                "metadata": {
                  "name": "replication-upstream-sample"
                },
                "spec": {
                  "dbcluster": {
                    "name": "dbcluster-sample"
                  },
                  "upstream": {}
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backup": "backup1",
                  "sourceDBCluster": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Sidecar",
                "metadata": {
                  "name": "sidecar-sample"
                },
                "spec": {
                  "sidecars": [
                    {
                      "args": [
                        "-c",
                        "while [ true ]\ndo\ndate\nset -x\nls -lh /logs/diagnostic\nset +x\ndone\n"
                      ],
                      "command": [
                        "/bin/sh"
                      ],
                      "image": "busybox",
                      "name": "sidecar-sample",
                      "volumeMounts": [
                        {
                          "mountPath": "/logs",
                          "name": "obsdisk"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Switchover",
                "metadata": {
                  "name": "switchover-sample"
                },
                "spec": {
                  "dbclusterRef": "dbcluster-sample",
                  "newPrimary": "aaaa-dbcluster-sample"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: gcr.io/alloydb-omni/operator/fleet-operator:1.5.0
          createdAt: "2025-07-25T04:12:04Z"
          operatorframework.io/suggested-namespace: alloydb-omni-system
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["backuprepositories.alloydbomni.internal.dbadmin.goog",
            "failovers.alloydbomni.internal.dbadmin.goog", "instancebackupplans.alloydbomni.internal.dbadmin.goog","instancebackups.alloydbomni.internal.dbadmin.goog","instancerestores.alloydbomni.internal.dbadmin.goog","instances.alloydbomni.internal.dbadmin.goog","instanceswitchovers.alloydbomni.internal.dbadmin.goog","instanceuserdefinedauthentications.alloydbomni.internal.dbadmin.goog","lrojobs.alloydbomni.internal.dbadmin.goog","replicationconfigs.alloydbomni.internal.dbadmin.goog","sidecars.alloydbomni.internal.dbadmin.goog","createstandbyjobs.alloydbomni.internal.dbadmin.goog","deletestandbyjobs.alloydbomni.internal.dbadmin.goog"]'
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BackupPlan is the Schema for the backupplans API.
            displayName: Backup Plan
            kind: BackupPlan
            name: backupplans.alloydbomni.dbadmin.goog
            version: v1
          - description: Backup is the Schema for the backups API.
            displayName: Backup
            kind: Backup
            name: backups.alloydbomni.dbadmin.goog
            version: v1
          - description: CreateStandbyJob is an internal workflow tracking object.
              Users should not directly interact with this.
            displayName: (Internal) Create Standby Job
            kind: CreateStandbyJob
            name: createstandbyjobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: DBCluster is the Schema for the DBClusters API.
            displayName: DB Cluster
            kind: DBCluster
            name: dbclusters.alloydbomni.dbadmin.goog
            version: v1
          - description: DBInstance is the Schema for the DBInstances API.
            displayName: DB Instance
            kind: DBInstance
            name: dbinstances.alloydbomni.dbadmin.goog
            version: v1
          - description: DeleteStandbyJob is an internal workflow tracking object.
              Users should not directly interact with this.
            displayName: (Internal) Delete Standby Job
            kind: DeleteStandbyJob
            name: deletestandbyjobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Failover represents the parameters and status of a single
              failover operation.
            displayName: Failover
            kind: Failover
            name: failovers.alloydbomni.dbadmin.goog
            version: v1
          - description: PgBouncer is the Schema for the PgBouncer API.
            displayName: PgBouncer
            kind: PgBouncer
            name: pgbouncers.alloydbomni.dbadmin.goog
            version: v1
          - description: Replication is the Schema for the Replications API.
            displayName: Replication
            kind: Replication
            name: replications.alloydbomni.dbadmin.goog
            version: v1
          - description: Restore is the Schema for the Restores API.
            displayName: Restore
            kind: Restore
            name: restores.alloydbomni.dbadmin.goog
            version: v1
          - description: Sidecar is the Schema for the Sidecar API.
            displayName: Sidecar
            kind: Sidecar
            name: sidecars.alloydbomni.dbadmin.goog
            version: v1
          - description: Switchover represents the parameters and status of a single
              switchover.
            displayName: Switchover
            kind: Switchover
            name: switchovers.alloydbomni.dbadmin.goog
            version: v1
          - description: UserDefinedAuthentication is the Schema for the UserDefinedAuthentications
              API.
            displayName: User Defined Authentication
            kind: UserDefinedAuthentication
            name: userdefinedauthentications.alloydbomni.dbadmin.goog
            version: v1
          - description: BackupRepository is the Schema for the backuprepositories
              API.
            displayName: (Internal) Backup Repository
            kind: BackupRepository
            name: backuprepositories.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Failover is the Schema for the failover API.
            displayName: (Internal) Failover
            kind: Failover
            name: failovers.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceBackupPlan is the Schema for the InstanceBackupPlan
              API.
            displayName: (Internal) Instance Backup Plan
            kind: InstanceBackupPlan
            name: instancebackupplans.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceBackup is the Schema for the InstanceBackup API.
            displayName: (Internal) Instance Backup
            kind: InstanceBackup
            name: instancebackups.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceRestore is the Schema for the InstanceRestore API.
            displayName: (Internal) Instance Restore
            kind: InstanceRestore
            name: instancerestores.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Instance is the Schema for the Instance API.
            displayName: (Internal) Instance
            kind: Instance
            name: instances.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceSwitchover is the Schema for the InstanceSwitchover
              API.
            displayName: (Internal) Instance Switchover
            kind: InstanceSwitchover
            name: instanceswitchovers.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceUserDefinedAuthentication is the Schema for the InstanceUserDefinedAuthentication
              API.
            displayName: (Internal) Instance User Defined Authentication
            kind: InstanceUserDefinedAuthentication
            name: instanceuserdefinedauthentications.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: LROJob is the Schema for the LROJob API.
            displayName: (Internal) LRO Job
            kind: LROJob
            name: lrojobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: ReplicationConfig is the Schema for the ReplicationConfig
              API.
            displayName: (Internal) Replication Config
            kind: ReplicationConfig
            name: replicationconfigs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Sidecar is the Schema for the Sidecar API.
            displayName: (Internal) Sidecar
            kind: Sidecar
            name: sidecars.alloydbomni.internal.dbadmin.goog
            version: v1
        description: |-
          AlloyDB Omni is a database that lets you deploy a streamlined edition of AlloyDB for PostgreSQL in your own computing environment. While being 100% compatible with PostgreSQL applications and tools, AlloyDB Omni delivers higher performance for both OLTP and OLAP workloads and makes administration easy with a number of autopilots. AlloyDB AI provides powerful tools to build generative AI applications with the ability to create, store, and index vector embeddings right in your database.

          ## Post-installation

          After installing the operator, you need to manually create the cert-manager resources in your cluster as this is a strict dependency for our operator. For this, follow these steps:

          ```
          kubectl create ns ${NAMESPACE:?}
          ```

          ```
          kubectl apply -f - <<EOF
          apiVersion: cert-manager.io/v1
          kind: ClusterIssuer
          metadata:
            name: alloydbomni-selfsigned-cluster-issuer
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Issuer
          metadata:
            name: fleet-selfsigned-issuer
            namespace: ${NAMESPACE:?}
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Certificate
          metadata:
            name: fleet-serving-cert
            namespace: ${NAMESPACE:?}
          spec:
            dnsNames:
            - fleet-webhook-service.alloydb-omni-system.svc
            - fleet-webhook-service.alloydb-omni-system.svc.cluster.local
            issuerRef:
              kind: Issuer
              name: fleet-selfsigned-issuer
            secretName: fleet-webhook-server-cert
          ---
          apiVersion: cert-manager.io/v1
          kind: Issuer
          metadata:
            name: local-selfsigned-issuer
            namespace: ${NAMESPACE:?}
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Certificate
          metadata:
            name: local-serving-cert
            namespace: ${NAMESPACE:?}
          spec:
            dnsNames:
            - local-webhook-service.alloydb-omni-system.svc
            - local-webhook-service.alloydb-omni-system.svc.cluster.local
            issuerRef:
              kind: Issuer
              name: local-selfsigned-issuer
            secretName: local-webhook-server-cert
          EOF
          ```

          Replace the following:
          * `NAMESPACE` : the namespace where you have installed your operator. Suggested namespace: `alloydb-omni-system`.

          ## Creating a database cluster

          After you install the AlloyDB Omni operator on your Kubernetes cluster, you can create an AlloyDB Omni database cluster. For each database cluster you create, you must first create a password secret using the following format:

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: db-pw-${DB_CLUSTER_NAME:s?}
            namespace: ${DB_CLUSTER_NAMESPACE:?}
          type: Opaque
          data:
            ${DB_CLUSTER_NAME:?}: "ENCODED_PASSWORD"
          ```

          Replace the following:
          * `DB_CLUSTER_NAME`: the name of this database cluster. For example, `my-db-cluster`.
          * `DB_CLUSTER_NAMESPACE` (Optional): the namespace where you want to create the database cluster. For example, `my-db-cluster-namespace`.
          * `ENCODED_PASSWORD`: the database login password for the default postgres user role, encoded as a base64 string. For example, `Q2hhbmdlTWUxMjM=` for password `ChangeMe123`.

          Follow [these][database-cluster] steps for creating a database cluster.

          ## Connect to AlloyDB Omni running on Kubernetes
          Follow [these][connect] steps to connect to AlloyDB Omni.

          [database-cluster]: https://cloud.google.com/alloydb/omni/docs/deploy-kubernetes#create
          [connect]: https://cloud.google.com/alloydb/omni/docs/run-connect#connect-kubernetes
        displayName: AlloyDB Omni Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - sql
        - postgres
        - vector
        - ai
        - analytics
        - postgresql
        - alloydb
        links:
        - name: Documentation
          url: https://cloud.google.com/alloydb/omni/docs
        - name: Release Notes
          url: https://cloud.google.com/alloydb/docs/release-notes
        maintainers:
        - email: alloydb-omni-ext@google.com
          name: Google Cloud
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Google
          url: https://cloud.google.com/alloydb/omni
        relatedImages:
        - gcr.io/alloydb-omni/operator/fleet-operator:1.5.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.1
        - gcr.io/alloydb-omni/operator/local-operator:1.5.0
        version: 1.5.0
      entries:
      - name: alloydb-omni-operator.v1.5.0
        version: 1.5.0
      - name: alloydb-omni-operator.v1.4.1
        version: 1.4.1
      - name: alloydb-omni-operator.v1.4.0
        version: 1.4.0
      - name: alloydb-omni-operator.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: alloydb-omni-operator
    provider:
      name: Google
      url: https://cloud.google.com/alloydb/omni
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Trilio
      provider-url: https://trilio.io/
    name: tvo-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tvo-operator.6.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOControlPlane",
                "metadata": {
                  "name": "tvocontrolplane-sample"
                },
                "spec": {
                  "crdType": "TVOControlPlane",
                  "common": {
                    "dmapi_workers": "16",
                    "memcached_servers": "",
                    "my_ip": {
                      "host_interface": null
                    },
                    "nova_group_id": "42436",
                    "nova_user_id": "42436",
                    "oslo_enable_proxy_headers_parsing": false,
                    "trustee_role": "creator"
                  },
                  "database": {
                    "common": {
                      "host": "EDIT_DB_HOSTNAME",
                      "root_password": "EDIT_PASSWORD",
                      "root_user_name": "root"
                    },
                    "datamover_api": {
                      "database": "dmapi",
                      "password": "EDIT_PASSWORD",
                      "user": "dmapi"
                    },
                    "wlm_api": {
                      "database": "workloadmgr",
                      "password": "EDIT_PASSWORD",
                      "user": "workloadmgr"
                    }
                  },
                  "helm3_hook": true,
                  "images": {
                    "triliovault_datamover_api": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_datamover_db_sync": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_cloud_trust": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_cron": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_db_sync": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_scheduler": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_workloads": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1"
                  },
                  "keystone": {
                    "common": {
                      "admin_role_name": "admin",
                      "auth_uri": "",
                      "auth_url": "",
                      "ca_cert": "CERT MULTIPLE LINES\n",
                      "cloud_admin_domain_name": "Default",
                      "cloud_admin_project_name": "admin",
                      "cloud_admin_user_name": "admin",
                      "cloud_admin_user_password": "",
                      "is_self_signed_ssl_cert": true,
                      "keystone_auth_host": "",
                      "keystone_auth_port": "",
                      "keystone_auth_protocol": "",
                      "region_name": "regionOne",
                      "service_project_domain_name": "Default",
                      "service_project_name": "service"
                    },
                    "datamover_api": {
                      "internal_endpoint": "ENDPOINT_PROTOCOL_INTERNAL://ENDPOINT_HOST_INTERNAL:8784/v2",
                      "password": "EDIT_PASSWORD",
                      "public_auth_host": "",
                      "public_endpoint": "ENDPOINT_PROTOCOL_PUBLIC://ENDPOINT_HOST_PUBLIC:8784/v2",
                      "service_desc": "TrilioVault Datamover Api Service",
                      "service_name": "dmapi",
                      "service_type": "datamover",
                      "user": "dmapi"
                    },
                    "identity": null,
                    "keystone_interface": "internal",
                    "wlm_api": {
                      "internal_endpoint": "ENDPOINT_PROTOCOL_INTERNAL://ENDPOINT_HOST_INTERNAL:8781/v1/$(tenant_id)s",
                      "password": "EDIT_PASSWORD",
                      "public_auth_host": "",
                      "public_endpoint": "ENDPOINT_PROTOCOL_PUBLIC://ENDPOINT_HOST_PUBLIC:8781/v1/$(tenant_id)s",
                      "service_desc": "TrilioVault Workloadmgr Api Service",
                      "service_name": "TrilioVaultWLM",
                      "service_type": "workloads",
                      "user": "triliovault"
                    }
                  },
                  "labels": {
                    "datamover_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_cron": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_scheduler": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_workloads": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    }
                  },
                  "pod": {
                    "replicas": {
                      "triliovault_datamover_api": 3,
                      "triliovault_object_store": 3,
                      "triliovault_wlm_api": 3,
                      "triliovault_wlm_cron": 1,
                      "triliovault_wlm_scheduler": 3,
                      "triliovault_wlm_workloads": 3
                    },
                    "security_context": {
                      "triliovault": {
                        "container": {
                          "triliovault_datamover_api": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 0,
                            "runAsUser": 63630
                          },
                          "triliovault_datamover_api_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 63630,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_api": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_cron": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_scheduler": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_workloads": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "wlm_cloud_trust_creation": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 42436
                          }
                        }
                      }
                    }
                  },
                  "rabbitmq": {
                    "common": {
                      "admin_password": "EDIT_PASSWORD",
                      "admin_user": "admin",
                      "driver": "messagingv2",
                      "host": "EDIT_HOST",
                      "port": "5672"
                    },
                    "datamover_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "",
                      "user": "dmapi",
                      "vhost": "dmapi"
                    },
                    "wlm_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "",
                      "user": "workloadmgr",
                      "vhost": "workloadmgr"
                    }
                  },
                  "triliovault_backup_targets": [
                    {
                      "backup_target_name": "S3_BT1",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY1",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET1",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL1",
                      "s3_region_name": "REGION1",
                      "s3_secret_key": "SECRETKEY1",
                      "s3_self_signed_cert": false,
                      "s3_signature_version": "default",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "S3_BT2",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY2",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET2",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL2",
                      "s3_region_name": "REGION2",
                      "s3_secret_key": "SECRETKEY2",
                      "s3_self_signed_cert": true,
                      "s3_signature_version": "default",
                      "s3_ssl_ca_cert": "SSL_CERT_LINE_1\n",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "NFS_BT3",
                      "backup_target_type": "nfs",
                      "is_default": false,
                      "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10",
                      "nfs_shares": "11.30.1.10:/mnt/share"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:3d80cad27a1043fb61735e7a6fb156c98b269cfcc68f8b9712713a7d4f6ad6bd
          createdAt: "2025-01-14T11:20:30Z"
          description: Cloud-Native Data Protection for OpenStack
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trilio-openstack
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trilioData/triliovault-cfg-scripts
          support: Trilio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trilio for OpenStack Control Plane. Includes all control
              plane components like Datamover Api, Trilio object store and Workloadmgr
              services.
            displayName: Trilio for OpenStack Control Plane
            kind: TVOControlPlane
            name: tvocontrolplanes.tvo.trilio.io
            version: v1
          - description: Trilio for OpenStack Backup Target. This CR is responsible
              to add a new backup target.
            displayName: Trilio for OpenStack Backup Target
            kind: TVOBackupTarget
            name: tvobackuptargets.tvo.trilio.io
            version: v1
        description: |-
          Trilio for OpenStack provides scalable, native, agentless, multi-tenant backup and recovery solutions, ensuring seamless protection of workloads with flexible storage options for dynamic cloud environments.

            * OpenStack Backup and Recovery: Recover any OpenStack application quickly in the event of data loss or misconfiguration. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Such as One-Click-Restore to entire project recovery on different clusters.
            * Disaster Recovery: Recover any OpenStack application quickly in the event of a disaster. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Trilio makes OpenStack disaster recovery easy.

            * Migration: A tenant or administrator can capture an application and its data and non-disruptively migrate that exact point-in-time to another OpenStack cloud. The need to migrate OpenStack workloads can be driven by economics, security or test/dev.

            * Network Functions Virtualization backup and recovery: Business critical services like NFV can quickly be configured to any state, assuring the optimal quality of service. Trilio provides recovery from data damage and allows restoration of entire VNFs, selected VNFs, or individual data items using a point-in-time working copy. The platform also allows recovery from a disaster by restoring the OpenStack network topology and populating it with a working copy of the entire NFV.

          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Backup
        - Recovery
        - RHOSO Backup
        - Disaster Recovery
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/openstack-backup-and-recovery/
        - name: Operator Documentation
          url: https://docs.trilio.io/openstack
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        provider:
          name: Trilio
          url: https://trilio.io/
        relatedImages:
        - registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:3d80cad27a1043fb61735e7a6fb156c98b269cfcc68f8b9712713a7d4f6ad6bd
        - registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82
        - registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1
        version: 6.1.1
      entries:
      - name: tvo-operator.6.1.1
        version: 6.1.1
      name: 6.0.x
    - currentCSV: tvo-operator.6.1.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOControlPlane",
                "metadata": {
                  "name": "tvocontrolplane-sample",
                  "namespace": "trilio-openstack"
                },
                "spec": {
                  "common": {
                    "dmapi_workers": "16",
                    "memcached_servers": "memcached-0.memcached.openstack.svc:11212",
                    "nova_group_id": "42436",
                    "nova_user_id": "42436",
                    "oslo_enable_proxy_headers_parsing": false,
                    "trustee_role": "member"
                  },
                  "crdType": "TVOControlPlane",
                  "database": {
                    "common": {
                      "host": "openstack.openstack.svc",
                      "port": 3306,
                      "root_password": "openstack",
                      "root_user_name": "root"
                    },
                    "datamover_api": {
                      "database": "dmapi",
                      "password": "dsw21aSyH5gHH6XE72k11Y3t3C308jY5rP02ndwJW",
                      "user": "dmapi"
                    },
                    "wlm_api": {
                      "database": "workloadmgr",
                      "password": "dfeeYb72OwUzX5YuRKcVjxXJaELV8YT0jZczIQwys0v",
                      "user": "workloadmgr"
                    }
                  },
                  "helm3_hook": true,
                  "images": {
                    "triliovault_datamover_api": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_datamover_db_sync": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_cloud_trust": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_cron": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_db_sync": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_scheduler": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_workloads": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd"
                  },
                  "keystone": {
                    "common": {
                      "admin_role_name": "admin",
                      "auth_uri": "https://keystone-internal.openstack.svc:5000/v3",
                      "auth_url": "https://keystone-internal.openstack.svc:5000",
                      "cloud_admin_domain_name": "Default",
                      "cloud_admin_project_name": "admin",
                      "cloud_admin_user_name": "admin",
                      "cloud_admin_user_password": "openstack",
                      "is_self_signed_ssl_cert": true,
                      "keystone_auth_host": "keystone-internal.openstack.svc",
                      "keystone_auth_port": "5000",
                      "keystone_auth_protocol": "https",
                      "region_name": "regionOne",
                      "service_project_domain_name": "Default",
                      "service_project_name": "service"
                    },
                    "datamover_api": {
                      "internal_endpoint": "https://triliovault-datamover-internal.trilio-openstack.svc:8784/v2",
                      "password": "defajgh48zgt88wFzzhwEt3xHGeCOdExYbLGQN2XRm",
                      "public_auth_host": "triliovault-datamover-public-trilio-openstack.apps.trilio.trilio.bos2",
                      "public_endpoint": "https://triliovault-datamover-public-trilio-openstack.apps.trilio.trilio.bos2/v2",
                      "service_desc": "TrilioVault Datamover Api Service",
                      "service_name": "dmapi",
                      "service_type": "datamover",
                      "user": "dmapi"
                    },
                    "keystone_interface": "internal",
                    "wlm_api": {
                      "internal_endpoint": "https://triliovault-wlm-internal.trilio-openstack.svc:8781/v1/$(tenant_id)s",
                      "password": "dsdeJl8axy0gv14fURwUUxy0qUOQbugUUOfXikz6j",
                      "public_auth_host": "triliovault-wlm-public-trilio-openstack.apps.trilio.trilio.bos2",
                      "public_endpoint": "https://triliovault-wlm-public-trilio-openstack.apps.trilio.trilio.bos2/v1/$(tenant_id)s",
                      "service_desc": "TrilioVault Workloadmgr Api Service",
                      "service_name": "TrilioVaultWLM",
                      "service_type": "workloads",
                      "user": "triliovault"
                    }
                  },
                  "labels": {
                    "datamover_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_cron": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_scheduler": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_workloads": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    }
                  },
                  "pod": {
                    "replicas": {
                      "triliovault_datamover_api": 3,
                      "triliovault_object_store": 3,
                      "triliovault_wlm_api": 3,
                      "triliovault_wlm_cron": 1,
                      "triliovault_wlm_scheduler": 3,
                      "triliovault_wlm_workloads": 3
                    },
                    "security_context": {
                      "triliovault": {
                        "container": {
                          "triliovault_datamover_api": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 0,
                            "runAsUser": 63630
                          },
                          "triliovault_datamover_api_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 63630,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_api": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_cron": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_scheduler": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_workloads": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "wlm_cloud_trust_creation": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 42436
                          }
                        }
                      }
                    }
                  },
                  "rabbitmq": {
                    "common": {
                      "admin_password": "EDIT_PASSWORD",
                      "admin_user": "admin",
                      "driver": "messagingv2",
                      "host": "EDIT_HOST",
                      "port": "15671",
                      "ssl": true
                    },
                    "datamover_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "rabbit://dmapi:${PASSWORD}@${RABBITMQ_HOST}:5671/dmapi?ssl=1",
                      "user": "dmapi",
                      "vhost": "dmapi"
                    },
                    "wlm_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "rabbit://workloadmgr:${PASSWORD}@${RABBITMQ_HOST}:5671/workloadmgr?ssl=1",
                      "user": "workloadmgr",
                      "vhost": "workloadmgr"
                    }
                  },
                  "triliovault_backup_targets": [
                    {
                      "backup_target_name": "BT1_S3",
                      "backup_target_type": "s3",
                      "is_default": false,
                      "s3_access_key": "",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "",
                      "s3_region_name": "",
                      "s3_secret_key": "",
                      "s3_self_signed_cert": false,
                      "s3_signature_version": "default",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "BT2_Other_S3",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY2",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET2",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL2",
                      "s3_region_name": "REGION2",
                      "s3_secret_key": "SECRETKEY2",
                      "s3_self_signed_cert": true,
                      "s3_signature_version": "default",
                      "s3_ssl_ca_cert": "SSL_CERT_LINE_1\nSSL_CERT_LINE_2\nSSL_CERT_LINE_N\n",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "other_s3"
                    },
                    {
                      "backup_target_name": "BT3_NFS",
                      "backup_target_type": "nfs",
                      "is_default": false,
                      "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10",
                      "nfs_shares": "192.168.2.3:/nfs/share"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOBackupTarget",
                "metadata": {
                  "name": "tvobackuptarget-sample",
                  "namespace": "trilio-openstack"
                },
                "spec": {
                  "crdType": "TVOBackupTarget",
                  "images": {
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd"
                  },
                  "triliovault_backup_target": {
                    "backup_target_name": "S3_BT4",
                    "backup_target_type": "s3",
                    "is_default": false,
                    "s3_type": "other_s3",
                    "s3_access_key": "",
                    "s3_secret_key": "",
                    "s3_region_name": "",
                    "s3_bucket": "",
                    "s3_endpoint_url": "",
                    "s3_signature_version": "default",
                    "s3_auth_version": "DEFAULT",
                    "s3_bucket_object_lock_enabled": false,
                    "s3_ssl_enabled": true,
                    "s3_ssl_verify": true,
                    "s3_self_signed_cert": true,
                    "s3_ssl_ca_cert": "SSL_CERT_LINE_1\nSSL_CERT_LINE_2\nSSL_CERT_LINE_N",
                    "nfs_shares": "192.168.2.30:/nfs/share",
                    "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:e282bb056a1780e5170db3e63efdf5099a5ac5240ad44162bc7af83cac893ac7
          createdAt: "2025-01-14T11:20:30Z"
          description: Cloud-Native Data Protection for OpenStack
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: tvo-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trilioData/triliovault-cfg-scripts
          support: Trilio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trilio for OpenStack Control Plane. Includes all control
              plane components like Datamover Api, Trilio object store and Workloadmgr
              services.
            displayName: Trilio for OpenStack Control Plane
            kind: TVOControlPlane
            name: tvocontrolplanes.tvo.trilio.io
            version: v1
          - description: Trilio for OpenStack Backup Target. This CR is responsible
              to add a new backup target.
            displayName: Trilio for OpenStack Backup Target
            kind: TVOBackupTarget
            name: tvobackuptargets.tvo.trilio.io
            version: v1
        description: |-
          Trilio for OpenStack provides scalable, native, agentless, multi-tenant backup and recovery solutions, ensuring seamless protection of workloads with flexible storage options for dynamic cloud environments.

            * OpenStack Backup and Recovery: Recover any OpenStack application quickly in the event of data loss or misconfiguration. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Such as One-Click-Restore to entire project recovery on different clusters.
            * Disaster Recovery: Recover any OpenStack application quickly in the event of a disaster. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Trilio makes OpenStack disaster recovery easy.

            * Migration: A tenant or administrator can capture an application and its data and non-disruptively migrate that exact point-in-time to another OpenStack cloud. The need to migrate OpenStack workloads can be driven by economics, security or test/dev.

            * Network Functions Virtualization backup and recovery: Business critical services like NFV can quickly be configured to any state, assuring the optimal quality of service. Trilio provides recovery from data damage and allows restoration of entire VNFs, selected VNFs, or individual data items using a point-in-time working copy. The platform also allows recovery from a disaster by restoring the OpenStack network topology and populating it with a working copy of the entire NFV.

          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Backup
        - Recovery
        - RHOSO Backup
        - Disaster Recovery
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/openstack-backup-and-recovery/
        - name: Operator Documentation
          url: https://docs.trilio.io/openstack
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        provider:
          name: Trilio
          url: https://trilio.io/
        relatedImages:
        - registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82
        - registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd
        - registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:e282bb056a1780e5170db3e63efdf5099a5ac5240ad44162bc7af83cac893ac7
        version: 6.1.9
      entries:
      - name: tvo-operator.6.1.9
        version: 6.1.9
      name: 6.1.x
    defaultChannel: 6.0.x
    packageName: tvo-operator
    provider:
      name: Trilio
      url: https://trilio.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: elasticsearch-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: elasticsearch-operator.v5.8.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Elasticsearch",
                "metadata": {
                  "name": "elasticsearch"
                },
                "spec": {
                  "indexManagement": {
                    "mappings": [
                      {
                        "aliases": [
                          "infra",
                          "logs.infra"
                        ],
                        "name": "infra",
                        "policyRef": "infra-policy"
                      }
                    ],
                    "policies": [
                      {
                        "name": "infra-policy",
                        "phases": {
                          "delete": {
                            "minAge": "2d",
                            "namespaceSpec": [
                              {
                                "minAge": "5h",
                                "namespace": "openshift-monitoring"
                              }
                            ],
                            "pruneNamespacesInterval": "24h"
                          },
                          "hot": {
                            "actions": {
                              "rollover": {
                                "maxAge": "8h"
                              }
                            }
                          }
                        },
                        "pollInterval": "30m"
                      }
                    ]
                  },
                  "managementState": "Managed",
                  "nodeSpec": {
                    "resources": {
                      "limits": {
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "512Mi"
                      }
                    }
                  },
                  "nodes": [
                    {
                      "nodeCount": 1,
                      "roles": [
                        "client",
                        "data",
                        "master"
                      ],
                      "storage": {
                        "size": "20G"
                      }
                    }
                  ],
                  "redundancyPolicy": "ZeroRedundancy"
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Kibana",
                "metadata": {
                  "name": "kibana"
                },
                "spec": {
                  "managementState": "Managed",
                  "nodeSelector": {},
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "memory": "512Mi"
                    },
                    "requests": {
                      "memory": "512Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
          createdAt: "2024-09-19T15:14:33Z"
          description: |
            The Elasticsearch Operator for OCP provides a means for configuring and managing an Elasticsearch cluster for tracing and cluster logging.
            ## Prerequisites and Requirements
            ### Elasticsearch Operator Namespace
            The Elasticsearch Operator must be deployed to the global operator group namespace
            ### Memory Considerations
            Elasticsearch is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Elasticsearch cluster.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory. Each ES node can operate with a
            lower memory setting though this is not recommended for production deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.21'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Cluster Logging, Jaeger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Elasticsearch cluster instance
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.openshift.io
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.openshift.io
            version: v1
        description: "### IN OCP 4.16 AND FORWARD \n\nELASTICSEARCH IS ONLY SUPPORTED
          FOR **SERVICEMESH, TRACING, AND KIALI**, AND THIS OPERATOR **WILL BE REMOVED
          FROM THE OPENSHIFT OPERATOR CATALOG IN NOVEMBER 2025**.  ELASTICSEARCH IS
          **NO LONGER SUPPORTED FOR LOG STORAGE**. **KIBANA IS NO LONGER SUPPORTED**
          IN OCP 4.16 AND FORWARD. \n\n### About \n\nThe Elasticsearch Operator for
          OCP provides a means for configuring and managing an Elasticsearch cluster
          for use in tracing \nand cluster logging as well as a Kibana instance to
          connect to it.\nThis operator only supports OCP Cluster Logging and Jaeger.
          \ It is tightly coupled to each and is not currently capable of\nbeing used
          as a general purpose manager of Elasticsearch clusters running on OCP.\n\nPlease
          note: For a general purpose Elasticsearch operator, please use Elastic's
          Elasticsearch (ECK) Operator [here](https://catalog.redhat.com/software/containers/elastic/eck-operator/5fabf6d1ecb52450895164be?container-tabs=gti)\n\nIt
          is recommended that this operator be installed in the `openshift-operators-redhat`
          namespace to \nproperly support the Cluster Logging and Jaeger use cases.\n\nOnce
          installed, the operator provides the following features for **Elasticsearch**:\n*
          **Create/Destroy**: Deploy an Elasticsearch cluster to the same namespace
          in which the elasticsearch CR is created.\n* **Update**: Changes to the
          elasticsearch CR will be scheduled and applied to the cluster in a controlled
          manner (most often as a rolling upgrade).\n* **Cluster health**: The operator
          will periodically poll the cluster to evaluate its current health (such
          as the number of active shards and if any cluster nodes have reached their
          storage watermark usage).\n* **Redeploys**: In the case where the provided
          secrets are updated, the Elasticsearch Operator will schedule and perform
          a full cluster restart.\n* **Index management**: The Elasticsearch Operator
          will create cronjobs to perform index management such as roll over and deletion.\n\nOnce
          installed, the operator provides the following features for **Kibana**:\n*
          **Create/Destroy**: Deploy a Kibana instance to the same namespace in which
          the kibana CR is created (this should be the same namespace as the elasticsearch
          CR).\n* **Update**: Changes to the kibana CR will be scheduled and applied
          to the cluster in a controlled manner.\n* **Redeploys**: In the case where
          the provided secrets are updated, the Elasticsearch Operator will perform
          a restart.\n\n### Additionally provided features\n* Out of the box multitenancy
          that is integrated with OCP user access control.\n* Document Level Security\n*
          mTLS communication between Elasticsearch, Kibana, Index Management cronjobs,
          and CLO's Fluentd\n* OCP prometheus dashboard for Elasticsearch clusters\n*
          Prometheus Alerting rules  \n"
        displayName: OpenShift Elasticsearch Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - jaeger
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Elasticsearch Operator
          url: https://github.com/openshift/elasticsearch-operator
        maintainers:
        - email: aos-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8d603100978c655020a8f4d81c527162348f468d93989edbc00463b8297fb8a5
        - registry.redhat.io/openshift-logging/kibana6-rhel8@sha256:a2c200667738fdb079ba11ed3b0df3c5901373428d5e560bc1efea9d5e1ddd67
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
        - registry.redhat.io/openshift-logging/logging-curator5-rhel9@sha256:394e0f46fb8be44c720b29aed716443747a636d2758226c9386f930f95758470
        - registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:b26d2218aa6b1efea83a1717d62f81009c97e8082434b918937c609f5cdab912
        version: 5.8.21
      entries:
      - name: elasticsearch-operator.v5.8.21
        version: 5.8.21
      - name: elasticsearch-operator.v5.8.20
        version: 5.8.20
      - name: elasticsearch-operator.v5.8.19
        version: 5.8.19
      - name: elasticsearch-operator.v5.8.18
        version: 5.8.18
      - name: elasticsearch-operator.v5.8.17
        version: 5.8.17
      - name: elasticsearch-operator.v5.8.16
        version: 5.8.16
      - name: elasticsearch-operator.v5.8.15
        version: 5.8.15
      - name: elasticsearch-operator.v5.8.14
        version: 5.8.14
      - name: elasticsearch-operator.v5.8.13
        version: 5.8.13
      - name: elasticsearch-operator.v5.8.12
        version: 5.8.12
      - name: elasticsearch-operator.v5.8.11
        version: 5.8.11
      - name: elasticsearch-operator.v5.8.10
        version: 5.8.10
      - name: elasticsearch-operator.v5.8.9
        version: 5.8.9
      - name: elasticsearch-operator.v5.8.8
        version: 5.8.8
      - name: elasticsearch-operator.v5.8.7
        version: 5.8.7
      - name: elasticsearch-operator.v5.8.6
        version: 5.8.6
      - name: elasticsearch-operator.v5.8.5
        version: 5.8.5
      - name: elasticsearch-operator.v5.8.4
        version: 5.8.4
      - name: elasticsearch-operator.v5.8.3
        version: 5.8.3
      - name: elasticsearch-operator.v5.8.2
        version: 5.8.2
      - name: elasticsearch-operator.v5.8.1
        version: 5.8.1
      - name: elasticsearch-operator.v5.8.0
        version: 5.8.0
      name: stable
    - currentCSV: elasticsearch-operator.v5.8.22
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Elasticsearch",
                "metadata": {
                  "name": "elasticsearch"
                },
                "spec": {
                  "indexManagement": {
                    "mappings": [
                      {
                        "aliases": [
                          "infra",
                          "logs.infra"
                        ],
                        "name": "infra",
                        "policyRef": "infra-policy"
                      }
                    ],
                    "policies": [
                      {
                        "name": "infra-policy",
                        "phases": {
                          "delete": {
                            "minAge": "2d",
                            "namespaceSpec": [
                              {
                                "minAge": "5h",
                                "namespace": "openshift-monitoring"
                              }
                            ],
                            "pruneNamespacesInterval": "24h"
                          },
                          "hot": {
                            "actions": {
                              "rollover": {
                                "maxAge": "8h"
                              }
                            }
                          }
                        },
                        "pollInterval": "30m"
                      }
                    ]
                  },
                  "managementState": "Managed",
                  "nodeSpec": {
                    "resources": {
                      "limits": {
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "512Mi"
                      }
                    }
                  },
                  "nodes": [
                    {
                      "nodeCount": 1,
                      "roles": [
                        "client",
                        "data",
                        "master"
                      ],
                      "storage": {
                        "size": "20G"
                      }
                    }
                  ],
                  "redundancyPolicy": "ZeroRedundancy"
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Kibana",
                "metadata": {
                  "name": "kibana"
                },
                "spec": {
                  "managementState": "Managed",
                  "nodeSelector": {},
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "memory": "512Mi"
                    },
                    "requests": {
                      "memory": "512Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.12-v4.17
          containerImage: registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:ce1f01f7a7fd8fa3d52f6bca62f63f65aebcb4c9a8ff8f6afaca7e5d121f0406
          createdAt: "2025-09-23T09:54:59Z"
          description: |
            The Elasticsearch Operator for OCP provides a means for configuring and managing an Elasticsearch cluster for tracing and cluster logging.
            ## Prerequisites and Requirements
            ### Elasticsearch Operator Namespace
            The Elasticsearch Operator must be deployed to the global operator group namespace
            ### Memory Considerations
            Elasticsearch is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Elasticsearch cluster.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory. Each ES node can operate with a
            lower memory setting though this is not recommended for production deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.22'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-5.8
          operators.operatorframework.io.bundle.channels.v1: stable-5.8
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: elasticsearch-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Cluster Logging, Jaeger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Elasticsearch cluster instance
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.openshift.io
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.openshift.io
            version: v1
        description: |
          ### IN OCP 4.16 AND FORWARD

          IN OCP 4.16 AND FORWARD ELASTICSEARCH and KIBANA ARE **NO LONGER SUPPORTED
          FOR LOG STORAGE** AND ELASTICSEARCH IS ONLY SUPPORTED FOR **SERVICEMESH,
          TRACING, AND KIALI**

          THIS OPERATOR **WILL REACH END OF LIFE ON NOVEMBER 2025**

          ### About

          The Elasticsearch Operator for OCP provides a means for configuring and
          managing an Elasticsearch cluster for use in tracing and cluster logging as
          well as a Kibana instance to connect to it. This operator only supports OCP
          Cluster Logging and Jaeger.  It is tightly coupled to each and is not
          currently capable of being used as a general purpose manager of
          Elasticsearch clusters running on OCP.

          Please note: For a general purpose Elasticsearch operator, please use
          Elastic's Elasticsearch (ECK) Operator
          [here](https://catalog.redhat.com/software/containers/elastic/eck-operator/5fabf6d1ecb52450895164be?container-tabs=gti)

          It is recommended that this operator be installed in the
          `openshift-operators-redhat` namespace to properly support the Cluster
          Logging and Jaeger use cases.

          Once installed, the operator provides the following features for
          **Elasticsearch**: * **Create/Destroy**: Deploy an Elasticsearch cluster to
          the same namespace in which the elasticsearch CR is created.
          * **Update**: Changes to the elasticsearch CR will be scheduled and applied
          to the cluster in a controlled manner (most often as a rolling upgrade).
          * **Cluster health**: The operator will periodically poll the cluster to
          evaluate its current health (such as the number of active shards and if any
          cluster nodes have reached their storage watermark usage).
          * **Redeploys**: In the case where the provided secrets are updated, the
          Elasticsearch Operator will schedule and perform a full cluster restart.
          * **Index management**: The Elasticsearch Operator will create cronjobs to
          perform index management such as roll over and deletion.

          Once installed, the operator provides the following features for **Kibana**:

          * **Create/Destroy**: Deploy a Kibana instance to the same namespace in
          which the kibana CR is created (this should be the same namespace as the
          elasticsearch CR).
          * **Update**: Changes to the kibana CR will be scheduled and applied to the
          cluster in a controlled manner.
          * **Redeploys**: In the case where the provided secrets are updated, the
          Elasticsearch Operator will perform a restart.

          ### Additionally provided features

          * Out of the box multitenancy that is integrated with OCP user access
          control.
          * Document Level Security
          * mTLS communication between Elasticsearch, Kibana, Index Management
          cronjobs, and CLO's Fluentd
          * OCP prometheus dashboard for Elasticsearch clusters
          * Prometheus Alerting rules
        displayName: OpenShift Elasticsearch Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - jaeger
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Elasticsearch Operator
          url: https://github.com/openshift/elasticsearch-operator
        maintainers:
        - email: aos-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:6388562dc496e524a297c733720dea23a2889468fa25394812b3247f5f08c907
        - registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:69878c9e40178530baf837a28d3f78e63eb6fba54865dbf05f30dda1a661d8d0
        - registry.redhat.io/openshift-logging/kibana6-rhel8@sha256:7cfb21861cbc3c6d8c6b6f0192ac3c687e4ac9b16fbccf56841e179a08553c84
        - registry.redhat.io/openshift-logging/logging-curator5-rhel9@sha256:dd1cef6a91ca25231af46e0417c7929663098a3b4ad1230dddf4a8171871ce0e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:ce1f01f7a7fd8fa3d52f6bca62f63f65aebcb4c9a8ff8f6afaca7e5d121f0406
        version: 5.8.22
      entries:
      - name: elasticsearch-operator.v5.8.22
        version: 5.8.22
      - name: elasticsearch-operator.v5.8.21
        version: 5.8.21
      - name: elasticsearch-operator.v5.8.20
        version: 5.8.20
      - name: elasticsearch-operator.v5.8.19
        version: 5.8.19
      - name: elasticsearch-operator.v5.8.18
        version: 5.8.18
      - name: elasticsearch-operator.v5.8.17
        version: 5.8.17
      - name: elasticsearch-operator.v5.8.16
        version: 5.8.16
      - name: elasticsearch-operator.v5.8.15
        version: 5.8.15
      - name: elasticsearch-operator.v5.8.14
        version: 5.8.14
      - name: elasticsearch-operator.v5.8.13
        version: 5.8.13
      - name: elasticsearch-operator.v5.8.12
        version: 5.8.12
      - name: elasticsearch-operator.v5.8.11
        version: 5.8.11
      - name: elasticsearch-operator.v5.8.10
        version: 5.8.10
      - name: elasticsearch-operator.v5.8.9
        version: 5.8.9
      - name: elasticsearch-operator.v5.8.8
        version: 5.8.8
      - name: elasticsearch-operator.v5.8.7
        version: 5.8.7
      - name: elasticsearch-operator.v5.8.6
        version: 5.8.6
      - name: elasticsearch-operator.v5.8.5
        version: 5.8.5
      - name: elasticsearch-operator.v5.8.4
        version: 5.8.4
      - name: elasticsearch-operator.v5.8.3
        version: 5.8.3
      - name: elasticsearch-operator.v5.8.2
        version: 5.8.2
      - name: elasticsearch-operator.v5.8.1
        version: 5.8.1
      - name: elasticsearch-operator.v5.8.0
        version: 5.8.0
      name: stable-5.8
    defaultChannel: stable-5.8
    packageName: elasticsearch-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: logging-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: logging-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Elasticsearch","metadata":{"name":"elasticsearch"},"spec":{"esClusterName":"prod","esVersion":"7.17.0","esSecurity":{"autoGeneratePassword":true,"tlsEnabled":true}}},{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Fluentd","metadata":{"name":"fluentd"},"spec":{"esCluster":{"host":"elasticsearch-master"},"esSecurity":{"tlsEnabled":true,"existingSecret":"elasticsearch-password"},"indexNameStrategy":"namespace_name"}},{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Kibana","metadata":{"name":"kibana"},"spec":{"replicas":1,"esCluster":{"host":"https://elasticsearch-master:9200","esVersion":"7.16.0","clusterName":"elasticsearch"},"esSecurity":{"tlsEnabled":true,"existingSecret":"elasticsearch-password"}}}]'
          capabilities: Seamless Upgrades
          categories: Database,  Monitoring, Big Data
          certified: "false"
          containerImage: quay.io/opstree/logging-operator:v0.4.0
          createdAt: 04-07-2022
          description: A golang-based operator to create and manage EFK (Elasticsearch,
            Fluentd, and Kibana) stack on Kubernetes.
          k8sMaxVersion: ""
          k8sMinVersion: ""
          repository: https://github.com/ot-container-kit/logging-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Elasticsearch
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Index Life Cycle
            displayName: IndexLifeCycle
            kind: IndexLifeCycle
            name: indexlifecycles.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Index Template
            displayName: IndexTemplate
            kind: IndexTemplate
            name: indextemplates.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Fluentd
            displayName: Fluentd
            kind: Fluentd
            name: fluentds.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Kibana
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.logging.opstreelabs.in
            version: v1beta1
        description: |2+

          Logging Operator is an operator created in Golang to set up and manage EFK(Elasticsearch, Fluentd, and Kibana) cluster inside Kubernetes and Openshift environment. This operator is capable of setting up each individual component of EFK cluster separately.
          ### Supported Features
          - Setup of elasticsearch cluster with different nodes type:- master, data, ingestion, and client.
          - Customizable configuration for elasticsearch, fluentd stack.
          - Setup of fluentd as light-weight log-shipper as Daemonset.
          - Kibana will be set up as visualization tool for elastic stack.
          - Seamless upgrades of elasticsearch, fluentd, and kibana.
          - Security best practices support for complete stack such as TLS, elastic security.
          - Kubernetes resources objects support like:- resources, securityContext, affinity, tolerations, etc.
          - Elasticsearch plugins and keystore seamless support with operator
          ### Documentation
          **[Documentation](https://ot-logging-operator.netlify.app/)**
          ### Getting Help
          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.
          ### Contributing
          You can contribute to this project by:-
          * Raising Bugs and Feature Requests
          * Fixing issue by opening a Pull Request
          * Improving Documentation
          ### License
          MongoDB Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/logging-operator/blob/master/LICENSE)

        displayName: Logging Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Operator
        - Elasticsearch
        - Fluentd
        - Kibana
        - Opstree
        - Logging
        - Monitoring
        - Cloud Native
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-logging-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep
        maturity: beta
        minKubeVersion: 1.17.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/opstree/logging-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: logging-operator.v0.4.0
        version: 0.4.0
      - name: logging-operator.v0.3.0
        version: 0.3.0
      name: beta
    defaultChannel: beta
    packageName: logging-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-licensing-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-licensing-operator.v8.0.220
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Developer"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:feeabfe4ecd36205306e3f3bf6c11517cfee735d7ae7eb4350abd01e9d76610b
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          olm.skipRange: <8.0.220
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:2f3de2bf46920cfe8c251011fc99841b43999561c1d3fe44850b205e94585a68
        version: 8.0.220
      entries:
      - name: infoscale-licensing-operator.v8.0.220
        version: 8.0.220
      - name: infoscale-licensing-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-licensing-operator.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-licensing-operator.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:b357b2203466516684c8fbc209757a98a45ec2376981f8699faf7dd5e37486a3
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:b357b2203466516684c8fbc209757a98a45ec2376981f8699faf7dd5e37486a3
        version: 8.0.330
      entries:
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-licensing-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Arctera
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Arctera
        - InfoScale
        - Licensing
        - Arctera Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
        version: 8.0.410
      entries:
      - name: infoscale-licensing-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-licensing-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-licensing-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-licensing-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Arctera
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Arctera
        - InfoScale
        - Licensing
        - Arctera Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
        version: 8.0.410
      entries:
      - name: infoscale-licensing-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-licensing-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-licensing-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-licensing-operator.v8.0.110
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Developer"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:e7b8d3c44df6151e61ded91336273e35eeaec04bd19124cf6313dc16d1224660
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:e7b8d3c44df6151e61ded91336273e35eeaec04bd19124cf6313dc16d1224660
        version: 8.0.110
      entries:
      - name: infoscale-licensing-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-licensing-operator.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-licensing-operator
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kedify
      provider-url: kedify.io
    name: kedify-agent-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kedify-agent-helm-operator.v0.2.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kedify.io/v1alpha1",
                "kind": "KedifyAgent",
                "metadata": {
                  "name": "kedifyagent-sample"
                },
                "spec": {
                  "agent": {
                    "agentId": "",
                    "apiKey": "kfy_xxxxxxxxxxxxxxxxxxxxx",
                    "createCrds": true,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v0.2.16@sha256:bfcc0b24ff75de17464ec213406f48bc211a932308e888c450ed23ee1786d37a"
                    },
                    "kedifyProxy": {
                      "globalValues": {
                        "openshift": true
                      },
                      "image": {
                        "tag": "v1.34.1@sha256:683e2f32809657c8d4a1508648545481130ac5409cd1ce63ccc795ad8c71202e"
                      }
                    },
                    "kubectlImage": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3"
                    },
                    "orgId": "00000000-0000-0000-0000-000000000000",
                    "securityContext": {
                      "readOnlyRootFilesystem": true
                    }
                  },
                  "keda": {
                    "enabled": true,
                    "image": {
                      "keda": {
                        "tag": "v2.17.2-0@sha256:802592b691a7b81a249055fa117994927c2fd376b5023a3888bcc7973c65ba52"
                      },
                      "metricsApiServer": {
                        "tag": "v2.17.2-0@sha256:df44bc1de75324158017d623a9d0adf3e3e9f1d64db6afa7133f777adc7b6389"
                      },
                      "pullPolicy": "IfNotPresent",
                      "webhooks": {
                        "tag": "v2.17.2-0@sha256:73e180bd5793dec3785fcb9ecd7e7184e645f7c37d561a5728970e5303c19a41"
                      }
                    }
                  },
                  "keda-add-ons-http": {
                    "enabled": true,
                    "interceptor": {
                      "pullPolicy": "IfNotPresent",
                      "scaledObject": {
                        "waitForCrd": true
                      }
                    },
                    "kubectlImage": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3"
                    },
                    "pinnedImages": {
                      "interceptor": {
                        "tag": "v0.10.0-17@sha256:30ccf20e79c6a62c14781872a73ad0a811063add7136b77df986bf403d6ca6bb"
                      },
                      "scaler": {
                        "tag": "v0.10.0-17@sha256:5863ee0ca09efeca225966c2eb8781f90f2dfb4aed872d84a6075b711ab042b2"
                      }
                    },
                    "scaler": {
                      "pullPolicy": "IfNotPresent"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: '{"kedify-openshift-default":{"description":"Default
            Kedify Agent CR for OpenShift. Register at https://dashboard.kedify.io
            to get started."}}'
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:d61332d50587ce2c75aff29033fdd5cb1bf0ec49d199ab72323a27d9d19ff402
          createdAt: "2025-08-07T14:04:11Z"
          description: Kedify Autoscaler, powered by proven open source technologies
            and KEDA, is an enterprise-grade service that allows teams to autoscale
            any workload using various event sources including HTTP events
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: keda
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: http://github.com/kedify/kedify-agent-helm-operator
          support: support@kedify.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KedifyAgent
            name: kedifyagents.charts.kedify.io
            version: v1alpha1
        description: Kedify Autoscaler, powered by proven open source technologies
          and KEDA, is an enterprise-grade service that allows teams to autoscale
          any workload using various event sources including HTTP events
        displayName: kedify-agent-helm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - autoscaling
        - keda
        - kedify
        links:
        - name: Kedify Homepage
          url: https://kedify.io
        maintainers:
        - email: support@kedify.io
          name: Kedify
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Kedify
          url: kedify.io
        relatedImages:
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:73e180bd5793dec3785fcb9ecd7e7184e645f7c37d561a5728970e5303c19a41
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:4b7277ad53dfb01bf5f2201dd31d94f24519da00d008638c3c36026b96ea78b4
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:8e3a458912ec6540cb414446425ad7b1546528f04a5d84fcd64d8ee5424392e6
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:cb40290842976fd09f2008d8e2e64ebf90a215f0584656d4886122d322006873
        - envoyproxy/envoy:v1.34.1@sha256:007da57c2c328a90bd4e6d99b70bd899132f1b4a9426ccafe25437cf84a60c14
        - envoyproxy/envoy:v1.34.1@sha256:683e2f32809657c8d4a1508648545481130ac5409cd1ce63ccc795ad8c71202e
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:9fcc76f983035eb0ceff7beb941bf32fe7fc28b341ec1aac5191ffc1471f401b
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:5863ee0ca09efeca225966c2eb8781f90f2dfb4aed872d84a6075b711ab042b2
        - ghcr.io/kedify/agent:v0.2.16@sha256:bfcc0b24ff75de17464ec213406f48bc211a932308e888c450ed23ee1786d37a
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:9300a5382373b179f7f541a663cd30e50284e54f352993e8cca34460acbd890d
        - envoyproxy/envoy:v1.34.1@sha256:75545b91d050aa51b52cb0673be172f2f263c971dae3d2a82a70d128d92e87c2
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:d61332d50587ce2c75aff29033fdd5cb1bf0ec49d199ab72323a27d9d19ff402
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:df44bc1de75324158017d623a9d0adf3e3e9f1d64db6afa7133f777adc7b6389
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:720b90a4f112a91a94b1bdafe445df1681657ddaa4fb20aa6d2904a0eb664a2a
        - ghcr.io/kedify/agent:v0.2.16@sha256:9a57d5eafe368aab1d94f0eec3101e02f0ff2e356d8499984030ab89161e42ae
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:850922994153104a8e382416561c251b46a95ff2b2a216aacbd7fcb4be7f6382
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:a38fdaf3a49eb4c71c4f3ab1211befa22fa5de56391f88efc11335f0bb3b1d29
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:c1c093a6fda7efc9bc30f110c55f8bb075250db929dbdb2534b9071efbfc3169
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:ea6b22bb12c8ff7cbab4633c2f4a0e49eaf3883b07ab1f47dbebff2a3e07f029
        - ghcr.io/kedify/agent:v0.2.16@sha256:c59d765091a31659236b8d80f4033912f5cf47d86c2e2da96bcfa133b9f6fe25
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:4b63409c022ba137760e7a534b6308fc1e3639f06fdb53ee1f6df32787b68400
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:82b041027dc900fdab717e8db6a522139067801c04cd1ebfe950572e714938ab
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:33cd5a8c6dd5e90614341dc1863e1f041f2c2475bd61edf5176a2a0d98ea54f9
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:802592b691a7b81a249055fa117994927c2fd376b5023a3888bcc7973c65ba52
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:30ccf20e79c6a62c14781872a73ad0a811063add7136b77df986bf403d6ca6bb
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:fc91d1f7f8b632614f31dea14c0c8d4a5f8311a83fdef82303857aa8b43b5782
        version: 0.2.16
      entries:
      - name: kedify-agent-helm-operator.v0.2.16
        version: 0.2.16
      name: stable
    defaultChannel: stable
    packageName: kedify-agent-helm-operator
    provider:
      name: Kedify
      url: kedify.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-operator.v2024.10.2-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "24.10.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "false",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": null
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": null
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:f4b9eb4f209f5428b0940056c760efcd9939553f3baac29924d1554701bc2dce
          createdAt: "2024-10-25T01:46:32Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2024.10.2
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2024.3.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "99.99.0.nextgen_uma_beta,3"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:f4b9eb4f209f5428b0940056c760efcd9939553f3baac29924d1554701bc2dce
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:7c4868db021beaffa762eb0de43739bbc865a4bf5900f1d17f78caed8612c073
        version: 2024.10.2-12
      entries:
      - name: uma-operator.v2024.10.2-12
        version: 2024.10.2-12
      name: "2024.10"
    - currentCSV: uma-operator.v2025.1.3-33
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.1.1.33,990033",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "false",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": null
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": null
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperDataQueueSize": 300,
                            "scrapperProcessingRoutines": 50,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:4a1bc124f22cf05dc6b42edbcc971d6309862e7be9c39b2ab6b1e3d50de61f3f
          createdAt: "2025-02-18T05:33:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.1.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2024.3.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       scrapperProcessingRoutines: 50
                       scrapperDataQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.1.1.26,990026"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:4a1bc124f22cf05dc6b42edbcc971d6309862e7be9c39b2ab6b1e3d50de61f3f
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:ae91a28d57c8a359f81757e4bf79472f2ac189ac0ae3ee0d87e8ff9c433f1ccc
        version: 2025.1.3-33
      entries:
      - name: uma-operator.v2025.1.3-33
        version: 2025.1.3-33
      name: "2025.1"
    - currentCSV: uma-operator.v2025.2.1-25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.2.1.25,990025",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mixedMode": "false",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperDataQueueSize": 300,
                            "scrapperProcessingRoutines": 50,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:d8f4ca750252f1570a2dd89cabf018cc4f0b2c0cd6cadfedbf48507a34d4aee9
          createdAt: "2025-03-05T05:49:42Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.2.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2024.3.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       scrapperProcessingRoutines: 50
                       scrapperDataQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.1.1.26,990026"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:d8f4ca750252f1570a2dd89cabf018cc4f0b2c0cd6cadfedbf48507a34d4aee9
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:8bcf0dac889f38a0bf77eb9babe3f726e3b2564d5988c05bdc5c2d91ae189c61
        version: 2025.2.1-25
      entries:
      - name: uma-operator.v2025.2.1-25
        version: 2025.2.1-25
      name: "2025.2"
    - currentCSV: uma-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:8ff4aab6d0043ccd2542a86b7e9c24d60670a3c376021ef2d02fc1a289130275
          createdAt: "2025-04-16T06:54:48Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.4.5
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.4.2.5,990005"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        - registry.connect.redhat.com/ca/uma-operator@sha256:8ff4aab6d0043ccd2542a86b7e9c24d60670a3c376021ef2d02fc1a289130275
        version: 2025.4.5-5
      entries:
      - name: uma-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:90b4b21d93bce41f2530699cec01f03e5daa73bfbf4f30cbcc8fbd6e75e97da6
          createdAt: "2025-07-24T15:07:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.5.2.12,990012""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:90b4b21d93bce41f2530699cec01f03e5daa73bfbf4f30cbcc8fbd6e75e97da6
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        version: 2025.5.4-12
      entries:
      - name: uma-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-operator.v2025.7.4-23
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:35f74109aa1916524c1fb483972caa31cb84381c3cff7c3b26fd9b676f9d748c
          createdAt: "2025-08-25T09:00:19Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.3
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.7.1.23,990023"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:35f74109aa1916524c1fb483972caa31cb84381c3cff7c3b26fd9b676f9d748c
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.4-23
      entries:
      - name: uma-operator.v2025.7.4-23
        version: 2025.7.4-23
      name: "2025.7"
    - currentCSV: uma-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:adea186632253865406e95a0062abbb312dee23c92ff108f5914e325dc0b8833
          createdAt: "2025-08-15T03:19:46Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.8.1.9,990009"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:adea186632253865406e95a0062abbb312dee23c92ff108f5914e325dc0b8833
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        version: 2025.8.2-9
      entries:
      - name: uma-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
          createdAt: "2025-09-24T12:22:41Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.12,990012"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        - registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
        version: 2025.9.1-12
      entries:
      - name: uma-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:c892787785c8c5e0de1692709a9e3f54eef36f1f00c53f1b9feabb613769a48c
          createdAt: "2025-07-30T08:06:11Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "99.99.0.serp_F152813,19"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:c892787785c8c5e0de1692709a9e3f54eef36f1f00c53f1b9feabb613769a48c
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        version: 2025.7.2-19
      entries:
      - name: uma-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:dc15dfdc8b82ee159b52036cc9d057e6525b08c478ed759939cf5c3fc7daeeec
          createdAt: "2025-09-11T11:15:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.8,990008"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:dc15dfdc8b82ee159b52036cc9d057e6525b08c478ed759939cf5c3fc7daeeec
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
          createdAt: "2025-09-24T12:22:41Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.12,990012"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: EDB
      provider-url: ""
    name: cloud-native-postgresql
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloud-native-postgresql.v1.27.0-rc1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:5f5bebbf4d8021dafd0fb7485d629b30aaa362e899e69eb52fa5e74757af6e36
          createdAt: "2025-07-30T09:28:05Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:5f5bebbf4d8021dafd0fb7485d629b30aaa362e899e69eb52fa5e74757af6e36
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.27.0-rc1
      entries:
      - name: cloud-native-postgresql.v1.27.0-rc1
        version: 1.27.0-rc1
      - name: cloud-native-postgresql.v1.26.0-rc3
        version: 1.26.0-rc3
      - name: cloud-native-postgresql.v1.26.0-rc2
        version: 1.26.0-rc2
      - name: cloud-native-postgresql.v1.26.0-rc1
        version: 1.26.0-rc1
      name: candidate
    - currentCSV: cloud-native-postgresql.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
          createdAt: "2025-08-19T12:00:36Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.27.0
      entries:
      - name: cloud-native-postgresql.v1.27.0
        version: 1.27.0
      - name: cloud-native-postgresql.v1.26.1
        version: 1.26.1
      - name: cloud-native-postgresql.v1.26.0
        version: 1.26.0
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      - name: cloud-native-postgresql.v1.24.1
        version: 1.24.1
      - name: cloud-native-postgresql.v1.24.0
        version: 1.24.0
      - name: cloud-native-postgresql.v1.23.3
        version: 1.23.3
      - name: cloud-native-postgresql.v1.23.2
        version: 1.23.2
      - name: cloud-native-postgresql.v1.23.1
        version: 1.23.1
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      - name: cloud-native-postgresql.v1.21.1
        version: 1.21.1
      - name: cloud-native-postgresql.v1.21.0
        version: 1.21.0
      - name: cloud-native-postgresql.v1.20.2
        version: 1.20.2
      - name: cloud-native-postgresql.v1.20.1
        version: 1.20.1
      - name: cloud-native-postgresql.v1.20.0
        version: 1.20.0
      - name: cloud-native-postgresql.v1.19.1
        version: 1.19.1
      - name: cloud-native-postgresql.v1.19.0
        version: 1.19.0
      name: fast
    - currentCSV: cloud-native-postgresql.v1.25.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
          createdAt: "2025-07-26T06:26:53Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.25.3
      entries:
      - name: cloud-native-postgresql.v1.25.3
        version: 1.25.3
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      - name: cloud-native-postgresql.v1.22.9
        version: 1.22.9
      - name: cloud-native-postgresql.v1.22.8
        version: 1.22.8
      - name: cloud-native-postgresql.v1.22.7
        version: 1.22.7
      - name: cloud-native-postgresql.v1.22.6
        version: 1.22.6
      - name: cloud-native-postgresql.v1.22.5
        version: 1.22.5
      - name: cloud-native-postgresql.v1.22.4
        version: 1.22.4
      - name: cloud-native-postgresql.v1.22.3
        version: 1.22.3
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      - name: cloud-native-postgresql.v1.18.13
        version: 1.18.13
      - name: cloud-native-postgresql.v1.18.12
        version: 1.18.12
      - name: cloud-native-postgresql.v1.18.11
        version: 1.18.11
      - name: cloud-native-postgresql.v1.18.10
        version: 1.18.10
      - name: cloud-native-postgresql.v1.18.9
        version: 1.18.9
      - name: cloud-native-postgresql.v1.18.8
        version: 1.18.8
      - name: cloud-native-postgresql.v1.18.7
        version: 1.18.7
      name: stable
    - currentCSV: cloud-native-postgresql.v1.18.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:69e439a2a961881309cc122174eaca273cf6d7873e9e871efbdb2d60804f40e4
          createdAt: "2024-06-13T16:08:55Z"
          description: EDB Postgres for Kubernetes is an open source operator designed
            to manage highly available PostgreSQL databases with a primary/standby
            architecture on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 1.18.0 < 1.18.13'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloud
        - native
        - edb
        - enterprisedb
        - cnp
        - pg4k
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: john.long@enterprisedb.com
          name: John Long
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: The EDB Postgres for Kubernetes Contributors
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:69e439a2a961881309cc122174eaca273cf6d7873e9e871efbdb2d60804f40e4
        version: 1.18.13
      entries:
      - name: cloud-native-postgresql.v1.18.13
        version: 1.18.13
      - name: cloud-native-postgresql.v1.18.12
        version: 1.18.12
      - name: cloud-native-postgresql.v1.18.11
        version: 1.18.11
      - name: cloud-native-postgresql.v1.18.10
        version: 1.18.10
      - name: cloud-native-postgresql.v1.18.9
        version: 1.18.9
      - name: cloud-native-postgresql.v1.18.8
        version: 1.18.8
      - name: cloud-native-postgresql.v1.18.7
        version: 1.18.7
      name: stable-v1.18
    - currentCSV: cloud-native-postgresql.v1.19.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:2a717edd0523798bd4d0b14c557b8d6b19ae3ad1f019c0cc4f61db49f10c3dad
          createdAt: "2023-11-08T19:02:47Z"
          description: Operator to manage Postgres high availability clusters with
            a primary/standby architecture.
          olm.skipRange: '>=1.15.3 < 1.19.6'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup manager
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster manager
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PgBouncer Pooler Manager
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL backups schedule
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: "EDB Postgres for Kubernetes is an operator designed, developed,
          and supported by EDB that covers the full \nlifecycle of a highly available
          Postgres database clusters with a primary/standby architecture, using native\nstreaming
          replication. It is based on the open source CloudNativePG operator, and
          provides additional value\nsuch as compatibility with Oracle using EDB Postgres
          Advanced Server and additional supported platforms such\nas IBM Power and
          OpenShift.\n\nKey features available include:\n\n* Kubernetes API integration
          for high availability\n* Self-healing through failover and automated recreation
          of replicas\n* Capacity management with scale up/down capabilities\n* Planned
          switchovers for scheduled maintenance\n* Read-only and read-write Kubernetes
          services definitions\n* Rolling updates for Postgres minor versions and
          operator upgrades\n* Continuous backup and point-in-time recovery\n* Connection
          Pooling with PgBouncer\n* Integrated metrics exporter out of the box\n*
          PostgreSQL replication across multiple Kubernetes clusters\n* Red Hat certified
          operator for OpenShift\n\nThe operator has been renamed from Cloud Native
          PostgreSQL. Existing users of Cloud Native PostgreSQL will not\nexperience
          any change, as the underlying components and resources have not changed.\n"
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloud
        - native
        - edb
        - enterprisedb
        - cnp
        - pg4k
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: sales@enterprisedb.com
          name: Sales
        - email: support@enterprisedb.com
          name: Support
        maturity: stable
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:2a717edd0523798bd4d0b14c557b8d6b19ae3ad1f019c0cc4f61db49f10c3dad
        version: 1.19.6
      entries:
      - name: cloud-native-postgresql.v1.19.6
        version: 1.19.6
      - name: cloud-native-postgresql.v1.19.5
        version: 1.19.5
      - name: cloud-native-postgresql.v1.19.4
        version: 1.19.4
      - name: cloud-native-postgresql.v1.19.3
        version: 1.19.3
      - name: cloud-native-postgresql.v1.19.2
        version: 1.19.2
      - name: cloud-native-postgresql.v1.19.1
        version: 1.19.1
      - name: cloud-native-postgresql.v1.19.0
        version: 1.19.0
      name: stable-v1.19
    - currentCSV: cloud-native-postgresql.v1.20.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:919c6669f8bfa333b19fdd1ca03aaeceebe0c5ca74d91cb66b2369680422e916
          createdAt: "2024-02-03T17:26:23Z"
          description: Operator to manage Postgres high availability clusters with
            a primary/standby architecture.
          olm.skipRange: '>=1.15.3 < 1.20.6'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup manager
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster manager
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PgBouncer Pooler Manager
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL backups schedule
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: "EDB Postgres for Kubernetes is an operator designed, developed,
          and supported by EDB that covers the full \nlifecycle of a highly available
          Postgres database clusters with a primary/standby architecture, using native\nstreaming
          replication. It is based on the open source CloudNativePG operator, and
          provides additional value\nsuch as compatibility with Oracle using EDB Postgres
          Advanced Server and additional supported platforms such\nas IBM Power and
          OpenShift.\n\nKey features available include:\n\n* Kubernetes API integration
          for high availability\n* Self-healing through failover and automated recreation
          of replicas\n* Capacity management with scale up/down capabilities\n* Planned
          switchovers for scheduled maintenance\n* Read-only and read-write Kubernetes
          services definitions\n* Rolling updates for Postgres minor versions and
          operator upgrades\n* Continuous backup and point-in-time recovery\n* Connection
          Pooling with PgBouncer\n* Integrated metrics exporter out of the box\n*
          PostgreSQL replication across multiple Kubernetes clusters\n* Red Hat certified
          operator for OpenShift\n\nThe operator has been renamed from Cloud Native
          PostgreSQL. Existing users of Cloud Native PostgreSQL will not\nexperience
          any change, as the underlying components and resources have not changed.\n"
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloud
        - native
        - edb
        - enterprisedb
        - cnp
        - pg4k
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: sales@enterprisedb.com
          name: Sales
        - email: support@enterprisedb.com
          name: Support
        maturity: stable
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:919c6669f8bfa333b19fdd1ca03aaeceebe0c5ca74d91cb66b2369680422e916
        version: 1.20.6
      entries:
      - name: cloud-native-postgresql.v1.20.6
        version: 1.20.6
      - name: cloud-native-postgresql.v1.20.5
        version: 1.20.5
      - name: cloud-native-postgresql.v1.20.4
        version: 1.20.4
      - name: cloud-native-postgresql.v1.20.3
        version: 1.20.3
      - name: cloud-native-postgresql.v1.20.2
        version: 1.20.2
      - name: cloud-native-postgresql.v1.20.1
        version: 1.20.1
      - name: cloud-native-postgresql.v1.20.0
        version: 1.20.0
      name: stable-v1.20
    - currentCSV: cloud-native-postgresql.v1.21.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:8cc328b1fc8efa3911d3dba689bc824ff789b305f7eb3284bf21ffcdc52a9683
          createdAt: "2024-06-13T16:41:26Z"
          description: EDB Postgres for Kubernetes is an open source operator designed
            to manage highly available PostgreSQL databases with a primary/standby
            architecture on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 1.18.0 < 1.21.6'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: john.long@enterprisedb.com
          name: John Long
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: The EDB Postgres for Kubernetes Contributors
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:8cc328b1fc8efa3911d3dba689bc824ff789b305f7eb3284bf21ffcdc52a9683
        version: 1.21.6
      entries:
      - name: cloud-native-postgresql.v1.21.6
        version: 1.21.6
      - name: cloud-native-postgresql.v1.21.5
        version: 1.21.5
      - name: cloud-native-postgresql.v1.21.4
        version: 1.21.4
      - name: cloud-native-postgresql.v1.21.3
        version: 1.21.3
      - name: cloud-native-postgresql.v1.21.2
        version: 1.21.2
      - name: cloud-native-postgresql.v1.21.1
        version: 1.21.1
      - name: cloud-native-postgresql.v1.21.0
        version: 1.21.0
      name: stable-v1.21
    - currentCSV: cloud-native-postgresql.v1.22.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:4d55ad54cc7e7807bc99f5a6f96bc92e88b4060f424ffd2cb6552be9505f38f4
          createdAt: "2025-07-25T19:02:56Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images. Only
              available in 1.25 or later
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster.
              Only available in 1.25 or later
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster. Only available in 1.25 or later
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication.
              Only available in 1.25 or later
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:4d55ad54cc7e7807bc99f5a6f96bc92e88b4060f424ffd2cb6552be9505f38f4
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.22.11
      entries:
      - name: cloud-native-postgresql.v1.22.11
        version: 1.22.11
      - name: cloud-native-postgresql.v1.22.10
        version: 1.22.10
      - name: cloud-native-postgresql.v1.22.9
        version: 1.22.9
      - name: cloud-native-postgresql.v1.22.8
        version: 1.22.8
      - name: cloud-native-postgresql.v1.22.7
        version: 1.22.7
      - name: cloud-native-postgresql.v1.22.6
        version: 1.22.6
      - name: cloud-native-postgresql.v1.22.5
        version: 1.22.5
      - name: cloud-native-postgresql.v1.22.4
        version: 1.22.4
      - name: cloud-native-postgresql.v1.22.3
        version: 1.22.3
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      name: stable-v1.22
    - currentCSV: cloud-native-postgresql.v1.23.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:934e01ed5cd35cf681414b9136a337514a1b73c0110871f1c78ba98447badb0f
          createdAt: "2024-12-23T16:37:54Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 1.18.0 < 1.23.6'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:934e01ed5cd35cf681414b9136a337514a1b73c0110871f1c78ba98447badb0f
        version: 1.23.6
      entries:
      - name: cloud-native-postgresql.v1.23.6
        version: 1.23.6
      - name: cloud-native-postgresql.v1.23.5
        version: 1.23.5
      - name: cloud-native-postgresql.v1.23.4
        version: 1.23.4
      - name: cloud-native-postgresql.v1.23.3
        version: 1.23.3
      - name: cloud-native-postgresql.v1.23.2
        version: 1.23.2
      - name: cloud-native-postgresql.v1.23.1
        version: 1.23.1
      name: stable-v1.23
    - currentCSV: cloud-native-postgresql.v1.24.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:a1acaaf5a413db0f33f3824a15aefe598974d1593172bcbccd0fc3d39d7b46d0
          createdAt: "2025-05-23T15:20:37Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:a1acaaf5a413db0f33f3824a15aefe598974d1593172bcbccd0fc3d39d7b46d0
        version: 1.24.4
      entries:
      - name: cloud-native-postgresql.v1.24.4
        version: 1.24.4
      - name: cloud-native-postgresql.v1.24.3
        version: 1.24.3
      - name: cloud-native-postgresql.v1.24.2
        version: 1.24.2
      - name: cloud-native-postgresql.v1.24.1
        version: 1.24.1
      - name: cloud-native-postgresql.v1.24.0
        version: 1.24.0
      name: stable-v1.24
    - currentCSV: cloud-native-postgresql.v1.25.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
          createdAt: "2025-07-26T06:26:53Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.25.3
      entries:
      - name: cloud-native-postgresql.v1.25.3
        version: 1.25.3
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      name: stable-v1.25
    - currentCSV: cloud-native-postgresql.v1.26.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:47d70233908ea855ddab3ff79c7a226719c3f2be5b8a5018c7934cd44bf0f546
          createdAt: "2025-07-26T07:59:14Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:47d70233908ea855ddab3ff79c7a226719c3f2be5b8a5018c7934cd44bf0f546
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        version: 1.26.1
      entries:
      - name: cloud-native-postgresql.v1.26.1
        version: 1.26.1
      - name: cloud-native-postgresql.v1.26.0
        version: 1.26.0
      name: stable-v1.26
    - currentCSV: cloud-native-postgresql.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
          createdAt: "2025-08-19T12:00:36Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.27.0
      entries:
      - name: cloud-native-postgresql.v1.27.0
        version: 1.27.0
      name: stable-v1.27
    defaultChannel: fast
    packageName: cloud-native-postgresql
    provider:
      name: EDB
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: http://marketplace.redhat.com
    name: marketplace-games-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marketplace-games-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "game.marketplace.redhat.com/v1",
                "kind": "Arcade",
                "metadata": {
                  "name": "arcade-sample"
                },
                "spec": {
                  "size": 1
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/marketplace-games-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/marketplace-games-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Arcade is the Schema for the arcades API
            displayName: Arcade Instance
            kind: Arcade
            name: arcades.game.marketplace.redhat.com
            version: v1
        description: RHM Arcade Operator - a simple operator used to showcase how
          to build an operator around an existing web application that can can run
          on Openshift, Vanilla k8s, or on-boarded to Red Hat Marketplace.
        displayName: Arcade Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - rhm
        - red hat
        - marketplace
        - games
        - arcade
        links:
        - name: Marketplace Games Operator
          url: https://github.com/redhat-marketplace/marketplace-games-operator
        maturity: alpha
        provider:
          name: Red Hat Marketplace
          url: http://marketplace.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/rhm-labs/arcade-operator@sha256:d6959be6693728d2dc4cd38b69251551d23c5002cb52db2684df31a0a8f7b0ab
        version: 1.0.0
      entries:
      - name: marketplace-games-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: marketplace-games-operator-rhmp
    provider:
      name: Red Hat Marketplace
      url: http://marketplace.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: application-services-metering-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: application-services-metering-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "applicationservices.redhat.com/v1",
                "kind": "Meter",
                "metadata": {
                  "name": "app-services-metrics-example"
                },
                "spec": {
                  "includeInfrastructure": false,
                  "meterCollectionEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Application Runtime
          certified: "false"
          containerImage: quay.io/redhat-developer/application-services-metering-operator:0.6.1
          createdAt: "2021-09-28T13:00:00Z"
          description: Collect the core usage of products from the Application Services
            portfolio into a single metric labeled by Runtimes, Integration, and Automation.
          repository: https://github.com/redhat-developer/application-services-metering-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Meter
            displayName: Meter
            kind: Meter
            name: meters.applicationservices.redhat.com
            version: v1
        description: |-
          ## About the managed application

          There is no managed application created with this operator.

          ## About this Operator

          The metering operator watches pods for predetermined labels, across a cluster or namespace(s).
          The label schema for Red Hat Application Services is in the format:

          * `com.company=Red_Hat`
          * `rht.comp=PAM`
          * `rht.comp_ver=7.10`
          * `rht.prod_name=Red_Hat_Process_Automation`
          * `rht.prod_ver=7.11`
          * `rht.subcomp=rhpam-kogito-runtime`
          * `rht.subcomp_t=application`

          It's important not to remove labels as the operator will cease monitoring those pods.

          The operator exposes a metric named: `appsvcs_cpu_usage_cores`
          The metric has a label for product category, these are Runtimes, Integration and Automation.
          There is a prometheus rule using the metric to group by product name, `appsvcs_cores_by_product:sum`.

          The operator utilises the OpenShift Monitoring Stack to provide usage observability via cloud.redhat.com

          ## Prerequisites for enabling this Operator

          None.
        displayName: Application Services Metering Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - metering
        - operator
        links:
        - name: Red Hat Application Services portfolio
          url: https://www.redhat.com/en/products/middleware
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/redhat-developer/application-services-metering-operator:0.6.1
        version: 0.6.1
      entries:
      - name: application-services-metering-operator.v0.6.1
        version: 0.6.1
      - name: application-services-metering-operator.v0.6.0
        version: 0.6.0
      name: alpha
    defaultChannel: alpha
    packageName: application-services-metering-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CloudBees
      provider-url: https://www.cloudbees.com/products/continuous-integration
    name: cloudbees-ci
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudbees-ci.v3.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.cloudbees.com/v1alpha1",
                "kind": "CloudBeesCI",
                "metadata": {
                  "name": "cloudbeesci-sample"
                },
                "spec": {
                  "Agents": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-agent:2.289.3.2"
                    },
                    "ImagePullSecrets": null,
                    "SeparateNamespace": {
                      "Create": false,
                      "Enabled": false,
                      "Name": null
                    }
                  },
                  "Hibernation": {
                    "Enabled": false,
                    "Image": {
                      "dockerImage": "cloudbees/managed-master-hibernation-monitor:247.c5dfce00a179",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "NodeSelector": {},
                    "Tolerations": []
                  },
                  "Master": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-mm:2.289.3.2"
                    },
                    "JavaOpts": null,
                    "OperationsCenterNamespace": null
                  },
                  "NetworkPolicy": {
                    "Enabled": false,
                    "JMXSelectors": [],
                    "ingressControllerSelector": []
                  },
                  "OperationsCenter": {
                    "AgentListenerPort": 50000,
                    "Annotations": {},
                    "CSRF": {
                      "ProxyCompatibility": false
                    },
                    "CasC": {
                      "ConfigMapName": "oc-casc-bundle",
                      "Enabled": false
                    },
                    "ContainerPort": 8080,
                    "ContextPath": null,
                    "Enabled": true,
                    "ExtraConfigMaps": [],
                    "ExtraContainers": [],
                    "ExtraGroovyConfiguration": {},
                    "ExtraVolumeMounts": [],
                    "ExtraVolumes": [],
                    "HealthProbeLivenessFailureThreshold": 12,
                    "HealthProbes": true,
                    "HostName": null,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-cloud-core-oc:2.289.3.2",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "Ingress": {
                      "Annotations": {
                        "kubernetes.io/tls-acme": "false"
                      },
                      "Class": "nginx",
                      "tls": {
                        "Enable": false,
                        "SecretName": null
                      }
                    },
                    "JavaOpts": null,
                    "JenkinsOpts": null,
                    "LoadBalancerIP": null,
                    "LoadBalancerSourceRanges": [
                      "0.0.0.0/0"
                    ],
                    "Name": "cjoc",
                    "NodeSelector": {},
                    "Platform": "standard",
                    "Protocol": "http",
                    "Resources": {
                      "Limits": {
                        "Cpu": 1,
                        "Memory": "2G"
                      },
                      "Requests": {
                        "Cpu": 1,
                        "Memory": "2G"
                      }
                    },
                    "Route": {
                      "tls": {
                        "CACertificate": null,
                        "Certificate": null,
                        "DestinationCACertificate": null,
                        "Enable": false,
                        "InsecureEdgeTerminationPolicy": "Redirect",
                        "Key": null,
                        "Termination": "edge"
                      }
                    },
                    "ServiceAgentListenerPort": 50000,
                    "ServiceAnnotations": {},
                    "ServicePort": 80,
                    "ServiceType": "ClusterIP",
                    "Tolerations": []
                  },
                  "Persistence": {
                    "AccessMode": "ReadWriteOnce",
                    "Annotations": {},
                    "Size": "20Gi",
                    "StorageClass": null
                  },
                  "PodSecurityPolicy": {
                    "Annotations": {},
                    "Enabled": false
                  },
                  "Subdomain": false,
                  "ingress-nginx": {
                    "Enabled": false,
                    "controller": {
                      "admissionWebhooks": {
                        "patch": {
                          "nodeSelector": {
                            "kubernetes.io/os": "linux"
                          }
                        }
                      },
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "nginx-ingress": {
                    "Enabled": false,
                    "controller": {
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "rbac": {
                    "agentsServiceAccountName": "jenkins-agents",
                    "hibernationMonitorServiceAccountName": "managed-master-hibernation-monitor",
                    "install": true,
                    "masterServiceAccountName": "jenkins",
                    "serviceAccountName": "cjoc"
                  },
                  "sidecarinjector": {
                    "Enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudBeesCI
            name: cloudbeescis.charts.cloudbees.com
            version: v1alpha1
        description: CloudBees CI
        displayName: CloudBees CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudbees
        - ci
        - enterprise
        - continuous
        - integration
        - jenkins
        - automation
        - pipeline
        - devops
        links:
        - name: CloudBees CI
          url: https://www.cloudbees.com/products/continuous-integration
        maintainers:
        - email: support@cloudbees.com
          name: Support
        - email: sales@cloudbees.com
          name: Sales
        maturity: alpha
        provider:
          name: CloudBees
          url: https://www.cloudbees.com/products/continuous-integration
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.connect.redhat.com/cloudbees/cloudbees-ci-operator@sha256:bd0bd72e9e7807b0fa307de8a93a56967221fe2dd181bac02383f0f9be9120e4
        version: 3.34.1
      entries:
      - name: cloudbees-ci.v3.34.1
        version: 3.34.1
      name: alpha
    defaultChannel: alpha
    packageName: cloudbees-ci
    provider:
      name: CloudBees
      url: https://www.cloudbees.com/products/continuous-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: https://github.com/Mellanox/network-operator/
    name: nvidia-network-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nvidia-network-operator.v25.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "doca3.1.0-25.07-0.9.7.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "nvcr.io/nvidia/mellanox",
                    "version": "sha256:a87096761d155eeb6f470e042d2d167bb466d57e63b4aba957f57d745e15a9b2"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:51270965284cceeda659fb1098e70a888b39664ceade7e9d455fb95ac85914b7
          createdAt: "2025-08-21T19:12:20Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:9a831bfdf85f313b1f5749b7c9b2673bb8fff18b4ff768c9242dabaa4468e449
        - nvcr.io/nvidia/mellanox/sriov-network-device-plugin@sha256:7508a160f3f124e0b78d2b4d5a52deca6b9f56e226fee957c4d156e425eb101a
        - nvcr.io/nvidia/mellanox/k8s-rdma-shared-dev-plugin@sha256:a87096761d155eeb6f470e042d2d167bb466d57e63b4aba957f57d745e15a9b2
        - nvcr.io/nvidia/mellanox/network-operator-init-container@sha256:02ee60ac6ee163acc97706bd4229d268f608a95bcbcb6fccee90ceb1d8a6c62b
        - nvcr.io/nvidia/mellanox/nvidia-k8s-ipam@sha256:3f692273c1598525b9c698c9d9cd463e5ddc55b638c6c031639f38786fc67475
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:13c2a2d055e501c03c6451413b85476ceafbc2fbffc7443952294c196f3f89f3
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:30354799264480e1aee057397eef2c52b277bbc6872184e2b7b811f4152b3a67
        - nvcr.io/nvidia/mellanox/nic-feature-discovery@sha256:1330e3a7ea2491bc310bfc4766769389ceb56d050831bf9c145b7bfb6aba36c4
        - nvcr.io/nvidia/mellanox/nic-configuration-operator@sha256:03f34f7fde3f65ca7b214b66c6fdbc1250d4f3863f9c02b0fba78e8aa84efca3
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:51270965284cceeda659fb1098e70a888b39664ceade7e9d455fb95ac85914b7
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:423aea08b4a0f01e2e19b6a709e14347bd08f8274f40002e0224090dfa4904fc
        - nvcr.io/nvidia/mellanox/ib-kubernetes@sha256:1fa251979b5f5201af50bc7aa6936829103d742f42bfbf752f20852f5fe44dfb
        - nvcr.io/nvidia/mellanox/ipoib-cni@sha256:7564ad1142c863e61f1dd84f5714deba9a7c6f6d85efb383ac3a2566eb4f4b71
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:5ad37cd5ae96153e6f87670f437ff5f6e44febe7bb6eaeb1a0f219e28284e53d
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:d41eea65b102dc5bb5792871ae212b59a65dae492ca109c52138d5626eba1ffd
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:f9d4b8358c68ba28f849ca710bf769c4327a639733c0ff30282ca33a744078f6
        - nvcr.io/nvidia/mellanox/nic-configuration-operator-daemon@sha256:b78c15904cecd343fa48374757b3619b5cbdc22059d5ee9731da87e35b94a2e2
        version: 25.7.0
      entries:
      - name: nvidia-network-operator.v25.7.0
        version: 25.7.0
      - name: nvidia-network-operator.v25.4.0
        version: 25.4.0
      name: stable
    - currentCSV: nvidia-network-operator.v24.10.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "24.10-0.7.0.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "v1.5.2"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:23aeb481d662496210d1c13083c65cfa4ffdcf67f973e5f884b1b936f693c0f5
          createdAt: "2025-01-27T08:16:45Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - ghcr.io/mellanox/network-operator-init-container@sha256:1699d23027ea30c9fa59575a914114bdfd5a87a359caf8c0a9b16d409ec0d068
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:47e1bb84ac97f9af49f97b49c2de44b4ba82b890b487880edc5c9adf86a176f7
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:23aeb481d662496210d1c13083c65cfa4ffdcf67f973e5f884b1b936f693c0f5
        - ghcr.io/mellanox/ib-kubernetes@sha256:100e6addf99d5b44e81cb3aa8d2fc892fced1de3183190931961b92ee924207b
        - ghcr.io/mellanox/nic-feature-discovery@sha256:92dda9434519de39be0f94c80787766e3399e8b238d14839360aff4d6232a09b
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:e36c46cf2c813c081bf7cc22b47fb4cf391a3d91804e47a6ce87f31d62f42535
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:4bb13a8b850c91011013509334d5ae4f4b7aa1d1b0c12958625e161cd50309d6
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:ecb2cfdc510b7ed09a59daed4817ee353e83a94d2d8581c3c052f4fa4b990574
        - ghcr.io/mellanox/ipoib-cni@sha256:81646c07e27ce1df2603050a75dd267f264685fdc64c285eca3925ae98ed5392
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:5a87d04872d5d4530cab3cbb3db528722fa24435f3375debd8ed102b04e631c2
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:db5797121735eeb935d8f11e305b50c80ba07aa83c291d76d05dea7e303bc740
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:77e3b946c20f45728dfa6dd1d115356de5ea317f91cddfd1540b7121550f66dc
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:9f468fdc4449e65e4772575f83aa85840a00f97165f9a00ba34695c91d610fbd
        version: 24.10.1
      entries:
      - name: nvidia-network-operator.v24.10.1
        version: 24.10.1
      name: v24.10
    - currentCSV: nvidia-network-operator.v24.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "24.10-0.7.0.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "v1.5.2"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:f258e8665d094c09159654759bc9ae41fbe0f7fffe95139ec02964759c3d8a87
          createdAt: "2024-11-26T09:58:46Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:f258e8665d094c09159654759bc9ae41fbe0f7fffe95139ec02964759c3d8a87
        - ghcr.io/mellanox/ipoib-cni@sha256:81646c07e27ce1df2603050a75dd267f264685fdc64c285eca3925ae98ed5392
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:77e3b946c20f45728dfa6dd1d115356de5ea317f91cddfd1540b7121550f66dc
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:e36c46cf2c813c081bf7cc22b47fb4cf391a3d91804e47a6ce87f31d62f42535
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:ecb2cfdc510b7ed09a59daed4817ee353e83a94d2d8581c3c052f4fa4b990574
        - ghcr.io/mellanox/ib-kubernetes@sha256:100e6addf99d5b44e81cb3aa8d2fc892fced1de3183190931961b92ee924207b
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:47e1bb84ac97f9af49f97b49c2de44b4ba82b890b487880edc5c9adf86a176f7
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:9f468fdc4449e65e4772575f83aa85840a00f97165f9a00ba34695c91d610fbd
        - ghcr.io/mellanox/nic-feature-discovery@sha256:92dda9434519de39be0f94c80787766e3399e8b238d14839360aff4d6232a09b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:db5797121735eeb935d8f11e305b50c80ba07aa83c291d76d05dea7e303bc740
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:5a87d04872d5d4530cab3cbb3db528722fa24435f3375debd8ed102b04e631c2
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:4bb13a8b850c91011013509334d5ae4f4b7aa1d1b0c12958625e161cd50309d6
        - ghcr.io/mellanox/network-operator-init-container@sha256:1699d23027ea30c9fa59575a914114bdfd5a87a359caf8c0a9b16d409ec0d068
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        version: 24.10.0
      entries:
      - name: nvidia-network-operator.v24.10.0
        version: 24.10.0
      name: v24.10.0
    - currentCSV: nvidia-network-operator.v24.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "docaTelemetryService": {
                    "image": "doca_telemetry",
                    "repository": "nvcr.io/nvidia/doca",
                    "version": "1.16.5-doca2.6.0-host"
                  },
                  "nicFeatureDiscovery": {
                    "image": "nic-feature-discovery",
                    "repository": "ghcr.io/mellanox",
                    "version": "v0.0.1"
                  },
                  "nvIpam": {
                    "enableWebhook": false,
                    "image": "nvidia-k8s-ipam",
                    "repository": "ghcr.io/mellanox",
                    "version": "v0.2.0"
                  },
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "24.07-0.6.1.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "v1.5.1"
                  },
                  "secondaryNetwork": {
                    "ipoib": {
                      "image": "ipoib-cni",
                      "repository": "ghcr.io/mellanox",
                      "version": "v1.2.0"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:fcee44995d64058cb7bf14c02da604879443e191a382402b8a6ebfe10a3a2601
          createdAt: "2024-08-29T13:23:34Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:8324c009c3d694fe7702af36305ede78b903d839190b1e4506e82362f9aa5c02
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:93e856dde88abb0aeb8e0b991e5f6d9f41884d8beea792f4d70cda06f18e783f
        - ghcr.io/mellanox/ib-kubernetes@sha256:b56bb88963dadaa2290a17cf0b57b86c55c09e7c075b7271c184959532555c91
        - ghcr.io/mellanox/ipoib-cni@sha256:81646c07e27ce1df2603050a75dd267f264685fdc64c285eca3925ae98ed5392
        - ghcr.io/mellanox/network-operator-init-container@sha256:1699d23027ea30c9fa59575a914114bdfd5a87a359caf8c0a9b16d409ec0d068
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:fcee44995d64058cb7bf14c02da604879443e191a382402b8a6ebfe10a3a2601
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:47e1bb84ac97f9af49f97b49c2de44b4ba82b890b487880edc5c9adf86a176f7
        - ghcr.io/mellanox/nic-feature-discovery@sha256:92dda9434519de39be0f94c80787766e3399e8b238d14839360aff4d6232a09b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:c41eaf76c06186e1dabb0d870225d946ad3ecda2a072fa4be197dd131b7695f2
        version: 24.7.0
      entries:
      - name: nvidia-network-operator.v24.7.0
        version: 24.7.0
      name: v24.7.0
    - currentCSV: nvidia-network-operator.v24.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "docaTelemetryService": {
                    "image": "doca_telemetry",
                    "repository": "nvcr.io/nvidia/doca",
                    "version": "1.16.5-doca2.6.0-host"
                  },
                  "nicFeatureDiscovery": {
                    "image": "nic-feature-discovery",
                    "repository": "ghcr.io/mellanox",
                    "version": "v0.0.1"
                  },
                  "nvIpam": {
                    "enableWebhook": false,
                    "image": "nvidia-k8s-ipam",
                    "repository": "ghcr.io/mellanox",
                    "version": "v0.2.0"
                  },
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "24.07-0.6.1.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "v1.5.1"
                  },
                  "secondaryNetwork": {
                    "ipoib": {
                      "image": "ipoib-cni",
                      "repository": "ghcr.io/mellanox",
                      "version": "v1.2.0"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:b8a07696d05fb8e5292991ba731758579313d6983ed3d75dd95e7356024d987c
          createdAt: "2024-09-04T06:38:33Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:b8a07696d05fb8e5292991ba731758579313d6983ed3d75dd95e7356024d987c
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:47e1bb84ac97f9af49f97b49c2de44b4ba82b890b487880edc5c9adf86a176f7
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:93e856dde88abb0aeb8e0b991e5f6d9f41884d8beea792f4d70cda06f18e783f
        - ghcr.io/mellanox/nic-feature-discovery@sha256:92dda9434519de39be0f94c80787766e3399e8b238d14839360aff4d6232a09b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:7facd05f60e2be29925ffce9d0a6b2c712c0670c8889b815425582c46c82efdc
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:676c74fcbe6d2c2341a092913b1afca65ba1a2f568ee6796b371542657e7a174
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:22e6ff1e9245f0cdf67e7904235d1a6181ee2bd3f39cedf76b39d4aeaa4a6c1d
        - ghcr.io/mellanox/ipoib-cni@sha256:81646c07e27ce1df2603050a75dd267f264685fdc64c285eca3925ae98ed5392
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:c41eaf76c06186e1dabb0d870225d946ad3ecda2a072fa4be197dd131b7695f2
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:10684c0c36364b1d897a03a346abb2f691a18541ae1242b596d002f34f0d7390
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:403be1bcf708984f90b3ccfdf24f58594554336d69f878d3843535f4fa61ceca
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:8324c009c3d694fe7702af36305ede78b903d839190b1e4506e82362f9aa5c02
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:7f87324adb351513d87fa27dfd0c9d9371ae42529d76cce8ecd68ce41c4d263f
        - ghcr.io/mellanox/ib-kubernetes@sha256:b56bb88963dadaa2290a17cf0b57b86c55c09e7c075b7271c184959532555c91
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:d5a163d36ce60630e14ae8620ab437510ff0821924e043b88f0d7b67cee3bd77
        - ghcr.io/mellanox/network-operator-init-container@sha256:1699d23027ea30c9fa59575a914114bdfd5a87a359caf8c0a9b16d409ec0d068
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:b3f503b1ac39ebc55840a436973f142eb7cf67e48a070284129fd1e30390c032
        version: 24.7.1
      entries:
      - name: nvidia-network-operator.v24.7.1
        version: 24.7.1
      name: v24.7.1
    - currentCSV: nvidia-network-operator.v25.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "25.01-0.6.0.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "sha256:9f468fdc4449e65e4772575f83aa85840a00f97165f9a00ba34695c91d610fbd"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:12f0576f372f5adc17b60795952ae47388e50d8fad563b9595f3091eadd8eb29
          createdAt: "2025-02-26T16:55:35Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:f5c1741cc6368fcec5b250285552d9c80ff7254414f68102cc89162e12ac0400
        - ghcr.io/mellanox/network-operator-init-container@sha256:67e93ccf3ecb61f17597567faf0f72e1b8ddcf73c5d7440baeadcc1cb6bb811b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:a257aae9c4e02b62f9f4ddeea66c4349c517519a544dd02fe4c8d3fee5229521
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:9f468fdc4449e65e4772575f83aa85840a00f97165f9a00ba34695c91d610fbd
        - ghcr.io/mellanox/nic-feature-discovery@sha256:92dda9434519de39be0f94c80787766e3399e8b238d14839360aff4d6232a09b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:68b3735c4edca2d889c456beba80ac2077b0cc0edba11a32769dacd3ece64892
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:3edf0cec91eea10df17c9f182b2db430cb383887ac144e27ac44043b308284d7
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:12f0576f372f5adc17b60795952ae47388e50d8fad563b9595f3091eadd8eb29
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:47e1bb84ac97f9af49f97b49c2de44b4ba82b890b487880edc5c9adf86a176f7
        - ghcr.io/mellanox/ib-kubernetes@sha256:100e6addf99d5b44e81cb3aa8d2fc892fced1de3183190931961b92ee924207b
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:cabce074d10a0f1d62135e2cc5442d65b49094b95b8297fdd024a1a5f461319f
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:5b91974c74f08fc5b1ad2edd4ca97d81b2dc6cc897045c3eac777826fe5d40d8
        - ghcr.io/mellanox/ipoib-cni@sha256:56f19687b0b956ba45bf079cf3a655b0d1bad0314da9d5ecfe0f4d98ff31fc86
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:7604a30cbd0927eba76ddc74ea1cfbe113563ccf068e6ab3fe78123e4deec28d
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        version: 25.1.0
      entries:
      - name: nvidia-network-operator.v25.1.0
        version: 25.1.0
      name: v25.1
    - currentCSV: nvidia-network-operator.v25.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "25.04-0.6.1.0-2"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "sha256:1eb5ab7f4b6f0cf89c954cf7ac8eb2e6fb2290a4f5f57417136a5302f20b12c8"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:93c2063f4da6a93810784be95df0083f07618486699d93245c97a29a63861dee
          createdAt: "2025-05-28T19:40:27Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:1b20b78f889339834ed74e0da621fc5da582719b2537b36d8967ddc6a04679b8
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:4a620bd39bc70928cd6e249855176ced0f4feb585c8c7624c0b9f5a3e57664c9
        - ghcr.io/mellanox/nic-feature-discovery@sha256:35b1820fe94a833a43613ef5c9db1f58b293ea0dbcf96ec0742b0987f23f133c
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:93c2063f4da6a93810784be95df0083f07618486699d93245c97a29a63861dee
        - ghcr.io/mellanox/ib-kubernetes@sha256:9774f774df6ede9ddf780b9193ca9d919618056e682d985e6b41e477b13c3432
        - ghcr.io/mellanox/nic-configuration-operator@sha256:353ab27f5f5fba08197cbe8968222e1c115431cc1b9b8d8445f1facb3c3c4d7b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:76529c24ea17e68b89a2c0d4823d66504fa6f53759760a449308fda396574097
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:69e6ab6a0249be918f60c1e04fc53a07f280b33563f8edc953806538a9cc3c30
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:35ff3b173b2694198e57b61720545be67e944c1da18ac004007a5b4e68562adc
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:cabce074d10a0f1d62135e2cc5442d65b49094b95b8297fdd024a1a5f461319f
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:9bc09051aca763f83cde99ee8909cf989b1d6c5d6ab683fbb1ca441d47f61fdf
        - ghcr.io/mellanox/network-operator-init-container@sha256:67e93ccf3ecb61f17597567faf0f72e1b8ddcf73c5d7440baeadcc1cb6bb811b
        - ghcr.io/mellanox/ipoib-cni@sha256:513d88d639a01ef98b65365f711cee7fbaa166137a56245b7ab32818f745418e
        - ghcr.io/mellanox/nic-configuration-operator-daemon@sha256:0e8478898e0b30aec13030b11081b205308666c2a9c5c5bf2c00121079901b41
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:56a41010b4dbee0dc31e8f9c1467a739241a2d248ed8b5054d4b418f7f79f04d
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:1eb5ab7f4b6f0cf89c954cf7ac8eb2e6fb2290a4f5f57417136a5302f20b12c8
        version: 25.4.0
      entries:
      - name: nvidia-network-operator.v25.4.0
        version: 25.4.0
      name: v25.4
    defaultChannel: stable
    packageName: nvidia-network-operator
    provider:
      name: NVIDIA
      url: https://github.com/Mellanox/network-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: node-observability-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-observability-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservability",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "labels": {
                    "kubernetes.io/hostname": "worker-a"
                  },
                  "type": "crio-kubelet"
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservabilityMachineConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "debug": {
                    "enableCrioProfiling": true
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservabilityRun",
                "metadata": {
                  "name": "nodeobservabilityrun-sample",
                  "namespace": "node-observability-operator"
                },
                "spec": {
                  "nodeObservabilityRef": {
                    "name": "cluster"
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservability",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "nodeSelector": {
                    "kubernetes.io/hostname": "worker-a"
                  },
                  "type": "crio-kubelet"
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservabilityMachineConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "debug": {
                    "enableCrioProfiling": true
                  },
                  "nodeSelector": {
                    "kubernetes.io/hostname": "worker-a"
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservabilityRun",
                "metadata": {
                  "name": "nodeobservabilityrun-sample",
                  "namespace": "node-observability-operator"
                },
                "spec": {
                  "nodeObservabilityRef": {
                    "name": "cluster"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/noo/node-observability-rhel8-operator@sha256:53ba77f648154ac8aca1ae6c83815078dfde211dba1e42f597e69844e4954c01
          createdAt: 2022-12-05T09:25:09
          olm.skipRange: <0.2.0
          operatorframework.io/suggested-namespace: node-observability-operator
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeObservability prepares a subset of worker nodes (identified
              by a label) for running node observability queries, such as profiling
            displayName: Node Observability
            kind: NodeObservability
            name: nodeobservabilities.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservability prepares a subset of worker nodes (identified
              by a label) for running node observability queries, such as profiling
            displayName: Node Observability
            kind: NodeObservability
            name: nodeobservabilities.nodeobservability.olm.openshift.io
            version: v1alpha2
          - description: NodeObservabilityMachineConfig is the Schema for the nodeobservabilitymachineconfigs
              API
            displayName: Node Observability Machine Config
            kind: NodeObservabilityMachineConfig
            name: nodeobservabilitymachineconfigs.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservabilityMachineConfig is the Schema for the nodeobservabilitymachineconfigs
              API
            displayName: Node Observability Machine Config
            kind: NodeObservabilityMachineConfig
            name: nodeobservabilitymachineconfigs.nodeobservability.olm.openshift.io
            version: v1alpha2
          - description: NodeObservabilityRun is a request to run observability actions
              on the nodes previously selected in NodeObservability resource
            displayName: Node Observability Run
            kind: NodeObservabilityRun
            name: nodeobservabilityruns.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservabilityRun is a request to run observability actions
              on the nodes previously selected in NodeObservability resource
            displayName: Node Observability Run
            kind: NodeObservabilityRun
            name: nodeobservabilityruns.nodeobservability.olm.openshift.io
            version: v1alpha2
        description: |-
          An Operator that will be able to gather debugging/profiling data over
          a custom period of time which would be helpful to troubleshoot and resolve issues
          for OpenShift customers.
        displayName: Node Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node-observability-operator
        links:
        - name: Source Code
          url: https://github.com/openshift/node-observability-operator
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/noo/node-observability-rhel8-operator@sha256:53ba77f648154ac8aca1ae6c83815078dfde211dba1e42f597e69844e4954c01
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/noo/node-observability-agent-rhel8@sha256:a8e60d9f63e2ed0cf1e27309aac54014e3f1efc947b51ca0a99376338d220875
        version: 0.2.0
      entries:
      - name: node-observability-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: node-observability-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/
    name: varnish-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: varnish-operator.v0.27.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "caching.ibm.com/v1alpha1",
                "kind": "VarnishCluster",
                "metadata": {
                  "name": "varnishcluster-sample"
                },
                "spec": {
                  "backend": {
                    "port": "web",
                    "selector": {
                      "app": "nginx"
                    }
                  },
                  "replicas": 1,
                  "service": {
                    "port": 80
                  },
                  "varnish": {
                    "args": [
                      "-p",
                      "default_ttl=3600",
                      "-p",
                      "default_grace=3600"
                    ]
                  },
                  "vcl": {
                    "configMapName": "vcl-files",
                    "entrypointFileName": "entrypoint.vcl"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: ibmcom/varnish-operator:0.27.2
          createdAt: 2021-02-15T11:15:31+0000
          description: Run and manage Varnish clusters on Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/varnish-operator
          support: https://ibm.github.io/varnish-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VarnishCluster is the Schema for the varnishclusters API
            displayName: Varnish Cluster
            kind: VarnishCluster
            name: varnishclusters.caching.ibm.com
            version: v1alpha1
        description: |
          Run and manage Varnish clusters on Kubernetes

          The operator manages the whole lifecycle of the cluster: creating, deleting and keeping the cluster configuration up to date.

          The operator is responsible for building the VCL configuration using templates defined by the users and keeping the configuration up to date when relevant events occur (backend pod failure, scaling of the deployment, VCL configuration change).

          Example of a simple `VarnishCluster`:

          ```yaml
          apiVersion: caching.ibm.com/v1alpha1
          kind: VarnishCluster
          metadata:
            name: varnishcluster-sample
            namespace: varnish-ns
          spec:
            vcl:
              configMapName: vcl-files
              entrypointFileName: entrypoint.vcl
            replicas: 3
            backend:
              selector:
                app: nginx
            service:
              port: 80
          ```

          Full documentation can be found [here](https://ibm.github.io/varnish-operator/)
        displayName: Varnish Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - varnish
        - cache
        links:
        - name: Varnish Operator Docs
          url: https://ibm.github.io/varnish-operator
        - name: Varnish Operator Source Code
          url: https://github.com/IBM/varnish-operator
        maintainers:
        - email: tomash.sidei@ibm.com
          name: Tomash Sidei
        - email: cingram@us.ibm.com
          name: Craig Ingram
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/
        relatedImages:
        - ibmcom/varnish-operator:0.27.2
        version: 0.27.2
      entries:
      - name: varnish-operator.v0.27.2
        version: 0.27.2
      name: alpha
    defaultChannel: alpha
    packageName: varnish-operator
    provider:
      name: IBM
      url: https://www.ibm.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Odigos
      provider-url: https://odigos.io
    name: odigos-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odigos-operator.v1.0.212
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.odigos.io/v1alpha1",
                "kind": "Odigos",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "odigos-operator"
                  },
                  "name": "odigos-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing
          containerImage: registry.connect.redhat.com/odigos/odigos-certified-operator-ubi9@sha256:c221eb2be22c2fcabcee728429bbca1255e0e31642c572e4dc40096d1d35e0d1
          createdAt: "2025-07-18T13:23:10Z"
          description: Odigos enables automatic distributed tracing with OpenTelemetry
            and eBPF.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Odigos Enterprise subscription
            (for enterprise features)
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Odigos
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Odigos is the Schema for the odigos API
            displayName: Odigos
            kind: Odigos
            name: odigos.operator.odigos.io
            version: v1alpha1
        description: |-
          The Odigos Operator provides options for installing and configuring Odigos.

          Upon creation of an Odigos resource, the Odigos Operator installs the following Odigos components:

          - Instrumentor
          - Odiglet
          - Autoscaler
          - Scheduler
          - Frontend UI

          With Odigos installed, follow the Odigos docs at docs.odigos.io for instructions on instrumenting applications and configuring destinations.
        displayName: Odigos Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - OpenTelemetry
        - eBPF
        - tracing
        - observability
        - distributed tracing
        - otel
        - monitoring
        links:
        - name: Odigos
          url: https://odigos.io
        - name: Odigos Documentation
          url: https://docs.odigos.io
        - name: Odigos on Github
          url: https://github.com/odigos-io/odigos
        maintainers:
        - email: mike@odigos.io
          name: Mike Dame
        maturity: alpha
        minKubeVersion: 1.20.15
        provider:
          name: Odigos
          url: https://odigos.io
        relatedImages:
        - registry.connect.redhat.com/odigos/odigos-autoscaler-ubi9@sha256:692193851b40f72ca79f7c82903136686d721c7fed77ef599037269ac6e55fdb
        - registry.connect.redhat.com/odigos/odigos-scheduler-ubi9@sha256:9e613c9b22d2fbfa3d21cc24f3b1e2dbc2e7287790e1f4286f1817c92565c865
        - registry.connect.redhat.com/odigos/odigos-enterprise-odiglet-ubi9@sha256:aec9b66461375994dc9aa92e5b2cc7d4151cf0332150726840cec049a7cbe7d9
        - registry.connect.redhat.com/odigos/odigos-enterprise-instrumentor-ubi9@sha256:647e526cdbe38023bb221a802b180f0b95cb25734a0b39fa11cbbe22a07482a3
        - registry.connect.redhat.com/odigos/odigos-collector-ubi9@sha256:b67138d3ffd9e721602c5bf83f0c03ae1dcee230eca5e15335dcb9abced281ba
        - registry.connect.redhat.com/odigos/odigos-instrumentor-ubi9@sha256:16712373c8e927b7208e43e9fe6e21180289975dbf3a81df60578e8550cf7150
        - registry.connect.redhat.com/odigos/odigos-odiglet-ubi9@sha256:770193901b4e5abcec59833f56f1530b43d4ec9514f3b5a5cb0bbb4356e06c67
        - registry.connect.redhat.com/odigos/odigos-ui-ubi9@sha256:b2a27cdfc94f1968a06ae18de1977127bb7d4e7d553e875d09a212693302f9c6
        - registry.connect.redhat.com/odigos/odigos-certified-operator-ubi9@sha256:c221eb2be22c2fcabcee728429bbca1255e0e31642c572e4dc40096d1d35e0d1
        version: 1.0.212
      entries:
      - name: odigos-operator.v1.0.212
        version: 1.0.212
      name: alpha
    defaultChannel: alpha
    packageName: odigos-operator
    provider:
      name: Odigos
      url: https://odigos.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-superset-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: superset-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:e0c9be4ac01a3d02f908f71f47f96bb1423fee14ba454c1bc627b35654d9cae3
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:e0c9be4ac01a3d02f908f71f47f96bb1423fee14ba454c1bc627b35654d9cae3
        version: 24.11.1
      entries:
      - name: superset-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: superset-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:f12cc3ed35135c14a82d49fcafed55465f0e66d46c2a36d532e5e09b9dece31e
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:f12cc3ed35135c14a82d49fcafed55465f0e66d46c2a36d532e5e09b9dece31e
        version: 25.3.0
      entries:
      - name: superset-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: superset-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
        version: 25.7.0
      entries:
      - name: superset-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: superset-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
        version: 25.7.0
      entries:
      - name: superset-operator.v25.7.0
        version: 25.7.0
      - name: superset-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-superset-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: LINBIT
      provider-url: ""
    name: linstor-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: linstor-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorController",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerImage": "",
                  "dbConnectionURL": "etcd://linstor-etcd:2379",
                  "drbdRepoCred": "",
                  "priorityClassName": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorCSIDriver",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerReplicas": 1,
                  "controllerEndpoint": "http://linstor:3370",
                  "csiControllerServiceAccountName": "csi-controller",
                  "csiNodeServiceAccountName": "csi-node",
                  "linstorPluginImage": "",
                  "imagePullSecret": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorSatelliteSet",
                "metadata": {
                  "name": "linstor-satellites"
                },
                "spec": {
                  "automaticStorageType": "None",
                  "controllerEndpoint": "http://linstor:3370",
                  "drbdRepoCred": "",
                  "kernelModuleInjectionMode": "ShippedModules",
                  "priorityClassName": "",
                  "satelliteImage": ""
                },
                "status": {
                  "errors": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          createdAt: "2021-06-22T07:57:53+00:00"
          description: LINSTOR Kubernetes Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/piraeusdatastore/piraeus-operator
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LinstorController is the Schema for the linstorcontrollersets
              API
            kind: LinstorController
            name: linstorcontrollers.linstor.linbit.com
            version: v1
          - description: LinstorCSIDriver is the Schema for the linstorcsidrivers
              API
            kind: LinstorCSIDriver
            name: linstorcsidrivers.linstor.linbit.com
            version: v1
          - description: LinstorSatelliteSet is the Schema for the linstornodesets
              API
            kind: LinstorSatelliteSet
            name: linstorsatellitesets.linstor.linbit.com
            version: v1
        description: |-
          LINSTOR is a configuration management system for storage on Linux systems.
          It manages LVM logical volumes and/or ZFS ZVOLs on a cluster of nodes.
          It leverages DRBD for replication between different nodes and to provide block
          storage devices to users and applications. It manages snapshots, encryption and
          caching of HDD backed data in SSDs via bcache.

          LINBIT provides a certified LINSTOR operator to ease deployment of LINSTOR
          on Openshift by installing DRBD, managing Satellite and Controller pods,
          and integrating with Openshift to provision persistent storage for your workloads.

          For detailed instructions and more configuration options, see our [user guide].

          [user guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift

          ## Install

          Unlike deployment via the helm chart, the certified Openshift
          operator does not deploy the needed etcd cluster. You must deploy this
          yourself ahead of time. We do this via the etcd operator available in the
          OperatorHub.

          IMPORTANT: It is advised that the etcd deployment uses persistent
          storage of some type. Either use an existing storage provisioner with
          a default `StorageClass` or simply use `hostPath` volumes.

          ### Installing the operator

          Hit "Install", select the stable update channel and a namespace for the
          operator. Use of a new namespace is recommended.

          Hit "Install" again. At this point you should have just one pod, the
          operator pod, running. Next we needs to configure the remaining provided APIs.

          #### A note on operator namespaces
          The LINSTOR operator can only watch for events and manage
          custom resources that are within the same namespace it is deployed
          within (OwnNamsespace). This means the LINSTOR Controller, LINSTOR
          Satellites, and LINSTOR CSI Driver pods all need to be deployed in the
          same namsepace as the LINSTOR Operator pod.

          ### Deploying the LINSTOR Controller

          Navigate to the left-hand control pane of the Openshift Web
          Console. Expand the "Operators" section, selecting "Installed Operators".
          Find the entry for the "Linstor Operator", then select the "LinstorController"
          from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorController". Click the "Create LinstorController" button.

          Here you will be presented with options to configure the LINSTOR
          Controller. Either via the web-form view or the YAML View. Regardless
          of which view you select, make sure that the `dbConnectionURL` matches
          the endpoint provided from your etcd deployment. Otherwise, the
          defaults are usually fine for most purposes.

          Lastly hit "Create", you should now see a linstor-controller pod
          running.

          ### Deploying the LINSTOR Satellites

          Next we need to deploy the Satellites Set. Just as before navigate
          to the left-hand control pane of the Openshift Web Console. Expand the
          "Operators" section, but this time select "Installed Operators". Find
          the entry for the "Linstor Operator", then select the
          "LinstorSatelliteSet" from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorSatelliteSet". Click the "Create LinstorSatelliteSet" button.

          Here you will be presented with the options to configure the LINSTOR
          Satellites. The defaults should be enough to get you started.
          Make sure the `controllerEndpoint` matches what is available in the
          openshift endpoints. The default is usually correct here.

          You can edit the `storagePools` section to configure LINSTOR storage pools,
          including preparing the backing devices. See our [storage guide].

          [storage guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-storage

          Below is an example manifest:

          ```yaml
          apiVersion: linstor.linbit.com/v1
          kind: LinstorSatelliteSet
          metadata:
            name: linstor-satellites
          spec:
            satelliteImage: ''
            automaticStorageType: None
            storagePools:
              lvmThinPools:
              - name: openshift-pool
                volumeGroup: ""
                thinVolume: openshift
                devicePaths:
                - /dev/vdb
            drbdRepoCred: ''
            kernelModuleInjectionMode: ShippedModules
            controllerEndpoint: 'http://linstor:3370'
            priorityClassName: ''
          ```

          Lastly hit "Create", you should now see a linstor-node pod
          running on every worker node.

          ### Deploying the LINSTOR CSI driver

          Last bit left is the CSI pods to bridge the layer between the CSI and
          LINSTOR. Just as before navigate to the left-hand control pane of the
          Openshift Web Console. Expand the "Operators" section, but this time
          select "Installed Operators". Find the entry for the "Linstor Operator",
          then select the "LinstorCSIDriver" from the "Provided APIs" column on the
          right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorCSIDriver". Click the "Create LinstorCSIDriver" button.

          Again, you will be presented with the options. Make sure that the
          `controllerEndpoint` is correct. Otherwise the defaults are fine for
          most use cases.

          Lastly hit "Create". You will now see a single "linstor-csi-controller" pod,
          as well as a "linstor-csi-node" pod on all worker nodes.

          ## Interacting with LINSTOR in Openshift.

          The Controller pod includes a LINSTOR Client, making it easy to interact directly with LINSTOR.
          For instance:

          ```
          oc exec deployment/linstor-cs-controller -- linstor storage-pool list
          ```

          This should only be necessary for investigating problems and accessing advanced functionality.
          Regular operation such as creating volumes should be achieved via the [Openshift/Kubernetes integration].

          [Openshift/Kubernetes integration]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-basic-configuration-and-deployment
        displayName: Linstor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - DRBD
        links:
        - name: LINBIT
          url: https://www.linbit.com/
        - name: User Guide
          url: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift
        maintainers:
        - email: sales@linbit.com
          name: LINBIT
        maturity: beta
        provider:
          name: LINBIT
        relatedImages:
        - registry.marketplace.redhat.com/rhm/linbit/linstor-operator@sha256:75e3ec9a9331072dc88c13979f0cacdf6232f8b3a6bfaf55c95b2ee3a03a5d5b
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:395a5a4aa4cfe3a0093d2225ce2e67acdcec0fd894e4b61e30a750f22931448d
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:bba872052e42fe0d6cbb2eea6541bf57f035575cbc7d838a3260b8582b8fe572
        - registry.marketplace.redhat.com/rhm/linbit/linstor-controller@sha256:1084ba14eaec7511682dd87c562398ce5e6690ded700c964373188b73bfe3b17
        - registry.marketplace.redhat.com/rhm/linbit/drbd-9@sha256:6337b3adfa3de87e74adcc974edafa09b618cb8e0315658c1cbdf8dd1c6c4c32
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:ba217e7e51f0748b54f477486263ff4983b0029635fd1f7ccc67160a09f81608
        - registry.marketplace.redhat.com/rhm/linbit/linstor-csi@sha256:034e3641376aa7ad47ed309fb389083c3ff82974f246f4d71b8789be8f8f618d
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:53b0f48eeecaa3ff1e211a2a413cd49266b7be85a5aa3a0431ab72d32e3a4e5c
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:8e52a6b31f9f3834d709dbca3614c2110b8ebb4af382c76b7056921221c87600
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:91cfe63acc12c91c3053ba7cc12d72cce404d33aab4afcfac285311f5ddfc7fe
        - registry.marketplace.redhat.com/rhm/linbit/linstor-satellite@sha256:4a3503092c8cb9435ccfb4837ee6a9183b680eb18a35c7d40227fa5c8f4da996
        version: 1.5.1
      entries:
      - name: linstor-operator.v1.5.1
        version: 1.5.1
      name: alpha
    - currentCSV: linstor-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorController",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerImage": "",
                  "dbConnectionURL": "etcd://linstor-etcd:2379",
                  "drbdRepoCred": "",
                  "priorityClassName": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorCSIDriver",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerReplicas": 1,
                  "controllerEndpoint": "http://linstor:3370",
                  "csiControllerServiceAccountName": "csi-controller",
                  "csiNodeServiceAccountName": "csi-node",
                  "linstorPluginImage": "",
                  "imagePullSecret": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorSatelliteSet",
                "metadata": {
                  "name": "linstor-satellites"
                },
                "spec": {
                  "automaticStorageType": "None",
                  "controllerEndpoint": "http://linstor:3370",
                  "drbdRepoCred": "",
                  "kernelModuleInjectionMode": "ShippedModules",
                  "priorityClassName": "",
                  "satelliteImage": ""
                },
                "status": {
                  "errors": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          createdAt: "2021-06-22T07:57:53+00:00"
          description: LINSTOR Kubernetes Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/piraeusdatastore/piraeus-operator
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LinstorController is the Schema for the linstorcontrollersets
              API
            kind: LinstorController
            name: linstorcontrollers.linstor.linbit.com
            version: v1
          - description: LinstorCSIDriver is the Schema for the linstorcsidrivers
              API
            kind: LinstorCSIDriver
            name: linstorcsidrivers.linstor.linbit.com
            version: v1
          - description: LinstorSatelliteSet is the Schema for the linstornodesets
              API
            kind: LinstorSatelliteSet
            name: linstorsatellitesets.linstor.linbit.com
            version: v1
        description: |-
          LINSTOR is a configuration management system for storage on Linux systems.
          It manages LVM logical volumes and/or ZFS ZVOLs on a cluster of nodes.
          It leverages DRBD for replication between different nodes and to provide block
          storage devices to users and applications. It manages snapshots, encryption and
          caching of HDD backed data in SSDs via bcache.

          LINBIT provides a certified LINSTOR operator to ease deployment of LINSTOR
          on Openshift by installing DRBD, managing Satellite and Controller pods,
          and integrating with Openshift to provision persistent storage for your workloads.

          For detailed instructions and more configuration options, see our [user guide].

          [user guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift

          ## Install

          Unlike deployment via the helm chart, the certified Openshift
          operator does not deploy the needed etcd cluster. You must deploy this
          yourself ahead of time. We do this via the etcd operator available in the
          OperatorHub.

          IMPORTANT: It is advised that the etcd deployment uses persistent
          storage of some type. Either use an existing storage provisioner with
          a default `StorageClass` or simply use `hostPath` volumes.

          ### Installing the operator

          Hit "Install", select the stable update channel and a namespace for the
          operator. Use of a new namespace is recommended.

          Hit "Install" again. At this point you should have just one pod, the
          operator pod, running. Next we needs to configure the remaining provided APIs.

          #### A note on operator namespaces
          The LINSTOR operator can only watch for events and manage
          custom resources that are within the same namespace it is deployed
          within (OwnNamsespace). This means the LINSTOR Controller, LINSTOR
          Satellites, and LINSTOR CSI Driver pods all need to be deployed in the
          same namsepace as the LINSTOR Operator pod.

          ### Deploying the LINSTOR Controller

          Navigate to the left-hand control pane of the Openshift Web
          Console. Expand the "Operators" section, selecting "Installed Operators".
          Find the entry for the "Linstor Operator", then select the "LinstorController"
          from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorController". Click the "Create LinstorController" button.

          Here you will be presented with options to configure the LINSTOR
          Controller. Either via the web-form view or the YAML View. Regardless
          of which view you select, make sure that the `dbConnectionURL` matches
          the endpoint provided from your etcd deployment. Otherwise, the
          defaults are usually fine for most purposes.

          Lastly hit "Create", you should now see a linstor-controller pod
          running.

          ### Deploying the LINSTOR Satellites

          Next we need to deploy the Satellites Set. Just as before navigate
          to the left-hand control pane of the Openshift Web Console. Expand the
          "Operators" section, but this time select "Installed Operators". Find
          the entry for the "Linstor Operator", then select the
          "LinstorSatelliteSet" from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorSatelliteSet". Click the "Create LinstorSatelliteSet" button.

          Here you will be presented with the options to configure the LINSTOR
          Satellites. The defaults should be enough to get you started.
          Make sure the `controllerEndpoint` matches what is available in the
          openshift endpoints. The default is usually correct here.

          You can edit the `storagePools` section to configure LINSTOR storage pools,
          including preparing the backing devices. See our [storage guide].

          [storage guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-storage

          Below is an example manifest:

          ```yaml
          apiVersion: linstor.linbit.com/v1
          kind: LinstorSatelliteSet
          metadata:
            name: linstor-satellites
          spec:
            satelliteImage: ''
            automaticStorageType: None
            storagePools:
              lvmThinPools:
              - name: openshift-pool
                volumeGroup: ""
                thinVolume: openshift
                devicePaths:
                - /dev/vdb
            drbdRepoCred: ''
            kernelModuleInjectionMode: ShippedModules
            controllerEndpoint: 'http://linstor:3370'
            priorityClassName: ''
          ```

          Lastly hit "Create", you should now see a linstor-node pod
          running on every worker node.

          ### Deploying the LINSTOR CSI driver

          Last bit left is the CSI pods to bridge the layer between the CSI and
          LINSTOR. Just as before navigate to the left-hand control pane of the
          Openshift Web Console. Expand the "Operators" section, but this time
          select "Installed Operators". Find the entry for the "Linstor Operator",
          then select the "LinstorCSIDriver" from the "Provided APIs" column on the
          right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorCSIDriver". Click the "Create LinstorCSIDriver" button.

          Again, you will be presented with the options. Make sure that the
          `controllerEndpoint` is correct. Otherwise the defaults are fine for
          most use cases.

          Lastly hit "Create". You will now see a single "linstor-csi-controller" pod,
          as well as a "linstor-csi-node" pod on all worker nodes.

          ## Interacting with LINSTOR in Openshift.

          The Controller pod includes a LINSTOR Client, making it easy to interact directly with LINSTOR.
          For instance:

          ```
          oc exec deployment/linstor-cs-controller -- linstor storage-pool list
          ```

          This should only be necessary for investigating problems and accessing advanced functionality.
          Regular operation such as creating volumes should be achieved via the [Openshift/Kubernetes integration].

          [Openshift/Kubernetes integration]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-basic-configuration-and-deployment
        displayName: Linstor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - DRBD
        links:
        - name: LINBIT
          url: https://www.linbit.com/
        - name: User Guide
          url: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift
        maintainers:
        - email: sales@linbit.com
          name: LINBIT
        maturity: beta
        provider:
          name: LINBIT
        relatedImages:
        - registry.marketplace.redhat.com/rhm/linbit/linstor-csi@sha256:034e3641376aa7ad47ed309fb389083c3ff82974f246f4d71b8789be8f8f618d
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:8e52a6b31f9f3834d709dbca3614c2110b8ebb4af382c76b7056921221c87600
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:395a5a4aa4cfe3a0093d2225ce2e67acdcec0fd894e4b61e30a750f22931448d
        - registry.marketplace.redhat.com/rhm/linbit/linstor-controller@sha256:1084ba14eaec7511682dd87c562398ce5e6690ded700c964373188b73bfe3b17
        - registry.marketplace.redhat.com/rhm/linbit/linstor-satellite@sha256:4a3503092c8cb9435ccfb4837ee6a9183b680eb18a35c7d40227fa5c8f4da996
        - registry.marketplace.redhat.com/rhm/linbit/linstor-operator@sha256:75e3ec9a9331072dc88c13979f0cacdf6232f8b3a6bfaf55c95b2ee3a03a5d5b
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:53b0f48eeecaa3ff1e211a2a413cd49266b7be85a5aa3a0431ab72d32e3a4e5c
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:bba872052e42fe0d6cbb2eea6541bf57f035575cbc7d838a3260b8582b8fe572
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:ba217e7e51f0748b54f477486263ff4983b0029635fd1f7ccc67160a09f81608
        - registry.marketplace.redhat.com/rhm/linbit/drbd-9@sha256:6337b3adfa3de87e74adcc974edafa09b618cb8e0315658c1cbdf8dd1c6c4c32
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:91cfe63acc12c91c3053ba7cc12d72cce404d33aab4afcfac285311f5ddfc7fe
        version: 1.5.1
      entries:
      - name: linstor-operator.v1.5.1
        version: 1.5.1
      name: stable
    defaultChannel: stable
    packageName: linstor-operator-rhmp
    provider:
      name: LINBIT
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lib-bucket-provisioner
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lib-bucket-provisioner.v2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage,Big Data
          certified: "false"
          containerImage: quay.io/noobaa/pause
          createdAt: "2020-07-01T06:28:48Z"
          description: Library for the dynamic provisioning of object store buckets
            to be used by object store providers.
          repository: https://github.com/kube-object-storage/lib-bucket-provisioner
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: ObjectBucketClaim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information
            displayName: ObjectBucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: |
          ### CRD-only Operator

          This operator package is **CRD-only** and the operator is a no-op operator.

          Instead, bucket provisioners using this library are using these CRD's and using CSV [required-crds](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/Documentation/design/building-your-csv.md#required-crds) them so that OLM can install it as a dependency.

          ### Generic Bucket Provisioning

          Kubernetes natively supports dynamic provisioning for many types of file and block storage, but lacks support for object bucket provisioning.

          This repo is a placeholder for an object store bucket provisioning library, very similar to the Kubernetes [sig-storage-lib-external-provisioner](https://github.com/kubernetes-sigs/sig-storage-lib-external-provisioner/blob/master/controller/controller.go) library.

          ### Known Provisioners
          - https://github.com/noobaa/noobaa-operator (NooBaa)
          - https://github.com/rook/rook (Rook-Ceph)
          - https://github.com/yard-turkey/aws-s3-provisioner (AWS-S3)
        displayName: lib-bucket-provisioner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - object
        - bucket
        - storage
        - cloud
        - s3
        links:
        - name: Github
          url: https://github.com/kube-object-storage/lib-bucket-provisioner
        maintainers:
        - email: jcope@redhat.com
          name: Jon Cope
        - email: jvance@redhat.com
          name: Jeff Vance
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: dzaken@redhat.com
          name: Danny Zaken
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/noobaa/pause
        version: 2.0.0
      entries:
      - name: lib-bucket-provisioner.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: lib-bucket-provisioner
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: dapr.io
      provider-url: https://dapr.io
    name: dapr-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dapr-kubernetes-operator.v0.0.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprControlPlane",
                "metadata": {
                  "name": "dapr-control-plane",
                  "namespace": "dapr-system"
                },
                "spec": {
                  "values": {}
                }
              },
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprCruiseControl",
                "metadata": {
                  "name": "dapr-cruise-control",
                  "namespace": "dapr-system"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprInstance",
                "metadata": {
                  "name": "dapr-instance",
                  "namespace": "dapr-system"
                },
                "spec": {
                  "values": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          containerImage: docker.io/daprio/kubernetes-operator@sha256:da599a273943bf28812314a11df1052411237dd1887edf0fe6104c5857f439b0
          createdAt: "2025-10-06T18:47:59Z"
          description: Dapr Control Plane Operator
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/dapr/kubernetes-operator
          support: contact@dapr.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DaprControlPlane is the Schema for the Dapr ControlPlane
              API.
            displayName: DaprControlPlane
            kind: DaprControlPlane
            name: daprcontrolplanes.operator.dapr.io
            version: v1alpha1
          - kind: DaprCruiseControl
            name: daprcruiscontrols.operator.dapr.io
            version: v1alpha1
          - description: DaprInstance is the Schema for the Dapr Instance API.
            displayName: DaprInstance
            kind: DaprInstance
            name: daprinstances.operator.dapr.io
            version: v1alpha1
        description: Dapr Control Plane Operator
        displayName: Dapr Control Plane Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dapr
        links:
        - name: Dapr
          url: https://dapr.io
        - name: Dapr Kubernetes Operator
          url: https://github.com/dapr/kubernetes-operator
        maintainers:
        - email: lburgazzoli@gmail.com
          name: Luca Burgazzoli
        - email: salaboy@gmail.com
          name: Mauricio Salatino
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: dapr.io
          url: https://dapr.io
        relatedImages:
        - docker.io/daprio/kubernetes-operator@sha256:da599a273943bf28812314a11df1052411237dd1887edf0fe6104c5857f439b0
        version: 0.0.12
      entries:
      - name: dapr-kubernetes-operator.v0.0.12
        version: 0.0.12
      - name: dapr-kubernetes-operator.v0.0.11
        version: 0.0.11
      - name: dapr-kubernetes-operator.v0.0.10
        version: 0.0.10
      - name: dapr-kubernetes-operator.v0.0.9
        version: 0.0.9
      - name: dapr-kubernetes-operator.v0.0.8
        version: 0.0.8
      - name: dapr-kubernetes-operator.v0.0.6
        version: 0.0.6
      - name: dapr-kubernetes-operator.v0.0.5
        version: 0.0.5
      - name: dapr-kubernetes-operator.v0.0.4
        version: 0.0.4
      - name: dapr-kubernetes-operator.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: dapr-kubernetes-operator
    provider:
      name: dapr.io
      url: https://dapr.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: gitwebhook-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitwebhook-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitWebhook",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "gitwebhook-operator",
                    "app.kubernetes.io/instance": "gitwebhook-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gitwebhook",
                    "app.kubernetes.io/part-of": "gitwebhook-operator"
                  },
                  "name": "gitwebhook-sample"
                },
                "spec": null
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/redhat-cop/gitwebhook-operator@sha256:2810d75eebb24069dbbedf6a82285136c6f16b4cc58b59f1b0fbdc38560cf8a8
          createdAt: "2024-05-10T18:17:13Z"
          description: An operator to create git webhooks declaratively
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: gitwebhook-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/gitwebhook-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitWebhook is the Schema for the gitwebhooks API
            displayName: Git Webhook
            kind: GitWebhook
            name: gitwebhooks.redhatcop.redhat.io
            version: v1alpha1
        description: "This operator automates the setting up of git webhooks. Its
          main purpose is to stream line the configuration of operators such as Tekton
          or ArgoCD which rely on git webhooks for some of their functionalities.\n\n##
          The GitWebhook CRD\n \nA CRD is provided by this operator to create a git
          web hook, here is an example for github\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GitWebhook\nmetadata:\n  name: gitwebhook-github\nspec:\n  gitHub:\n    gitServerCredentials:\n
          \     name: github-pat\n  repositoryOwner: ${repo_owner}\n  ownerType: user\n
          \ repositoryName: ${repo_name}\n  webhookURL: https://hellowebhook.com\n
          \ insecureSSL: false\n  webhookSecret:\n    name: webhook-secret\n  events:\n
          \   - push  \n  contentType: json\n  active: true  \n```    \n\nhere is
          an explanation of each field:\n\n- `gihub` specifies how to connect to the
          git api server. It also requires a local reference to a secret (in the same
          namespace) containing a key `token` with a valid github token to be used
          to authenticate. A similar `gitLab` section exists when connecting to gitlab.
          Only one of `gitLab` or `gitHub` can be defined. \n- `repositoryOwner` and
          `repositoryName` identify the repository for which we want to receive events.\n-
          `ownerType` can have two values: `user` and `organization` and identifies
          the kind of owner.\n- `webhookURL` is the URL for to be called.\n - `insecureSSL`
          defines whether the target URL certificate should be validated (default
          `false`). \n- `webhookSecret` defines a local reference to a secret containing
          the `secret` key. The value is a shared secret between the webhook caller
          and the received for farther validation or identification of the caller.\n-
          `events` is the list of the repo-level events that the webhook should generate.
          The list of valid events for github can be found [here](https://docs.github.com/en/developers/webhooks-and-events/webhooks/webhook-events-and-payloads).
          The list of valid events for gitlab can be found [here](https://docs.gitlab.com/ee/user/project/integrations/webhook_events.html).\n
          - `contentType` defines the format of the webhook payload (default `json`)
          (github).\n - `active` whether the webhook should be turned on (default
          `true`) (github).\n - `pushEventBranchFilter` a regular expression to filter
          from which branches push events should be generated (gitlab only).\n\n##
          Security Considerations\n\nThis operator does not own credentials for the
          git server, but instead always allocate a new connection based on the credentials
          referenced in the CR and every reconcile cycle. As a result there is no
          risk of security escalation or credential leaking between tenants of a cluster
          using this operator. On the other hand it is the responsibility of the namespace
          owners or the platform owner to ensure that valid git credentials are always
          available in the namespace where the GitWebhook CRs need to defined.\n\n##
          Current support\n\nCurrently this operator support creating repo-level webhooks
          for github and gitlab. Potentially this operator could be extended to support
          org-level webhook or other git systems. Contributions are welcome.  "
        displayName: GitWebhook operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - webhook
        - github
        - gitlab
        links:
        - name: GitWbehook Operator
          url: https://github.com/redhat-cop/gitwebhook-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/gitwebhook-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - quay.io/redhat-cop/gitwebhook-operator@sha256:2810d75eebb24069dbbedf6a82285136c6f16b4cc58b59f1b0fbdc38560cf8a8
        version: 0.1.2
      entries:
      - name: gitwebhook-operator.v0.1.2
        version: 0.1.2
      - name: gitwebhook-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: gitwebhook-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ec2-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ec2-controller.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "DHCPOptions",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "ElasticIPAddress",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "Instance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "InternetGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "NATGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "RouteTable",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "SecurityGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "Subnet",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "TransitGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "VPC",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "VPCEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ec2-controller:1.7.0
          createdAt: "2025-09-27T00:04:55Z"
          description: AWS EC2 controller is a service controller for managing EC2
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CapacityReservation represents the state of an AWS ec2 CapacityReservation
              resource.
            displayName: CapacityReservation
            kind: CapacityReservation
            name: capacityreservations.ec2.services.k8s.aws
            version: v1alpha1
          - description: DHCPOptions represents the state of an AWS ec2 DHCPOptions
              resource.
            displayName: DHCPOptions
            kind: DHCPOptions
            name: dhcpoptions.ec2.services.k8s.aws
            version: v1alpha1
          - description: ElasticIPAddress represents the state of an AWS ec2 ElasticIPAddress
              resource.
            displayName: ElasticIPAddress
            kind: ElasticIPAddress
            name: elasticipaddresses.ec2.services.k8s.aws
            version: v1alpha1
          - description: FlowLog represents the state of an AWS ec2 FlowLog resource.
            displayName: FlowLog
            kind: FlowLog
            name: flowlogs.ec2.services.k8s.aws
            version: v1alpha1
          - description: Instance represents the state of an AWS ec2 Instance resource.
            displayName: Instance
            kind: Instance
            name: instances.ec2.services.k8s.aws
            version: v1alpha1
          - description: InternetGateway represents the state of an AWS ec2 InternetGateway
              resource.
            displayName: InternetGateway
            kind: InternetGateway
            name: internetgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: LaunchTemplate represents the state of an AWS ec2 LaunchTemplate
              resource.
            displayName: LaunchTemplate
            kind: LaunchTemplate
            name: launchtemplates.ec2.services.k8s.aws
            version: v1alpha1
          - description: NATGateway represents the state of an AWS ec2 NATGateway
              resource.
            displayName: NATGateway
            kind: NATGateway
            name: natgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: NetworkACL represents the state of an AWS ec2 NetworkACL
              resource.
            displayName: NetworkACL
            kind: NetworkACL
            name: networkacls.ec2.services.k8s.aws
            version: v1alpha1
          - description: RouteTable represents the state of an AWS ec2 RouteTable
              resource.
            displayName: RouteTable
            kind: RouteTable
            name: routetables.ec2.services.k8s.aws
            version: v1alpha1
          - description: SecurityGroup represents the state of an AWS ec2 SecurityGroup
              resource.
            displayName: SecurityGroup
            kind: SecurityGroup
            name: securitygroups.ec2.services.k8s.aws
            version: v1alpha1
          - description: Subnet represents the state of an AWS ec2 Subnet resource.
            displayName: Subnet
            kind: Subnet
            name: subnets.ec2.services.k8s.aws
            version: v1alpha1
          - description: TransitGateway represents the state of an AWS ec2 TransitGateway
              resource.
            displayName: TransitGateway
            kind: TransitGateway
            name: transitgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: TransitGatewayVPCAttachment represents the state of an AWS
              ec2 TransitGatewayVPCAttachment resource.
            displayName: TransitGatewayVPCAttachment
            kind: TransitGatewayVPCAttachment
            name: transitgatewayvpcattachments.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCEndpoint represents the state of an AWS ec2 VPCEndpoint
              resource.
            displayName: VPCEndpoint
            kind: VPCEndpoint
            name: vpcendpoints.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCEndpointServiceConfiguration represents the state of an
              AWS ec2 VPCEndpointServiceConfiguration resource.
            displayName: VPCEndpointServiceConfiguration
            kind: VPCEndpointServiceConfiguration
            name: vpcendpointserviceconfigurations.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCPeeringConnection represents the state of an AWS ec2 VPCPeeringConnection
              resource.
            displayName: VPCPeeringConnection
            kind: VPCPeeringConnection
            name: vpcpeeringconnections.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPC represents the state of an AWS ec2 VPC resource.
            displayName: VPC
            kind: VPC
            name: vpcs.ec2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Compute Cloud (EC2) resources in AWS from within your Kubernetes cluster.

          **About Amazon EC2**

          Amazon Elastic Compute Cloud (Amazon EC2) provides scalable computing capacity in the Amazon Web Services (AWS) Cloud. Using Amazon EC2 eliminates your need to invest in hardware up front, so you can develop and deploy applications faster. You can use Amazon EC2 to launch as many or as few virtual servers as you need, configure security and networking, and manage storage. Amazon EC2 enables you to scale up or down to handle changes in requirements or spikes in popularity, reducing your need to forecast traffic.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EC2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ec2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EC2 Developer Resources
          url: https://aws.amazon.com/ec2/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ec2 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ec2-controller:1.7.0
        version: 1.7.0
      entries:
      - name: ack-ec2-controller.v1.7.0
        version: 1.7.0
      - name: ack-ec2-controller.v1.6.2
        version: 1.6.2
      - name: ack-ec2-controller.v1.6.1
        version: 1.6.1
      - name: ack-ec2-controller.v1.6.0
        version: 1.6.0
      - name: ack-ec2-controller.v1.4.11
        version: 1.4.11
      - name: ack-ec2-controller.v1.4.10
        version: 1.4.10
      - name: ack-ec2-controller.v1.4.9
        version: 1.4.9
      - name: ack-ec2-controller.v1.4.8
        version: 1.4.8
      - name: ack-ec2-controller.v1.4.7
        version: 1.4.7
      - name: ack-ec2-controller.v1.4.6
        version: 1.4.6
      - name: ack-ec2-controller.v1.4.5
        version: 1.4.5
      - name: ack-ec2-controller.v1.4.4
        version: 1.4.4
      - name: ack-ec2-controller.v1.4.2
        version: 1.4.2
      - name: ack-ec2-controller.v1.4.1
        version: 1.4.1
      - name: ack-ec2-controller.v1.4.0
        version: 1.4.0
      - name: ack-ec2-controller.v1.3.7
        version: 1.3.7
      - name: ack-ec2-controller.v1.3.6
        version: 1.3.6
      - name: ack-ec2-controller.v1.3.5
        version: 1.3.5
      - name: ack-ec2-controller.v1.3.4
        version: 1.3.4
      - name: ack-ec2-controller.v1.3.3
        version: 1.3.3
      - name: ack-ec2-controller.v1.3.1
        version: 1.3.1
      - name: ack-ec2-controller.v1.3.0
        version: 1.3.0
      - name: ack-ec2-controller.v1.2.28
        version: 1.2.28
      - name: ack-ec2-controller.v1.2.27
        version: 1.2.27
      - name: ack-ec2-controller.v1.2.26
        version: 1.2.26
      - name: ack-ec2-controller.v1.2.25
        version: 1.2.25
      - name: ack-ec2-controller.v1.2.24
        version: 1.2.24
      - name: ack-ec2-controller.v1.2.23
        version: 1.2.23
      - name: ack-ec2-controller.v1.2.22
        version: 1.2.22
      - name: ack-ec2-controller.v1.2.21
        version: 1.2.21
      - name: ack-ec2-controller.v1.2.20
        version: 1.2.20
      - name: ack-ec2-controller.v1.2.17
        version: 1.2.17
      - name: ack-ec2-controller.v1.2.16
        version: 1.2.16
      - name: ack-ec2-controller.v1.2.15
        version: 1.2.15
      - name: ack-ec2-controller.v1.2.14
        version: 1.2.14
      - name: ack-ec2-controller.v1.2.13
        version: 1.2.13
      - name: ack-ec2-controller.v1.2.12
        version: 1.2.12
      - name: ack-ec2-controller.v1.2.11
        version: 1.2.11
      - name: ack-ec2-controller.v1.2.10
        version: 1.2.10
      - name: ack-ec2-controller.v1.2.9
        version: 1.2.9
      - name: ack-ec2-controller.v1.2.8
        version: 1.2.8
      - name: ack-ec2-controller.v1.2.7
        version: 1.2.7
      - name: ack-ec2-controller.v1.2.6
        version: 1.2.6
      - name: ack-ec2-controller.v1.2.4
        version: 1.2.4
      - name: ack-ec2-controller.v1.2.1
        version: 1.2.1
      - name: ack-ec2-controller.v1.2.0
        version: 1.2.0
      - name: ack-ec2-controller.v1.1.2
        version: 1.1.2
      - name: ack-ec2-controller.v1.1.1
        version: 1.1.1
      - name: ack-ec2-controller.v1.1.0
        version: 1.1.0
      - name: ack-ec2-controller.v1.0.7
        version: 1.0.7
      - name: ack-ec2-controller.v1.0.6
        version: 1.0.6
      - name: ack-ec2-controller.v1.0.4
        version: 1.0.4
      - name: ack-ec2-controller.v1.0.3
        version: 1.0.3
      - name: ack-ec2-controller.v1.0.2
        version: 1.0.2
      - name: ack-ec2-controller.v1.0.1
        version: 1.0.1
      - name: ack-ec2-controller.v1.0.0
        version: 1.0.0
      - name: ack-ec2-controller.v0.1.1
        version: 0.1.1
      - name: ack-ec2-controller.v0.1.0
        version: 0.1.0
      - name: ack-ec2-controller.v0.0.21
        version: 0.0.21
      - name: ack-ec2-controller.v0.0.20
        version: 0.0.20
      - name: ack-ec2-controller.v0.0.19
        version: 0.0.19
      - name: ack-ec2-controller.v0.0.18
        version: 0.0.18
      - name: ack-ec2-controller.v0.0.17
        version: 0.0.17
      - name: ack-ec2-controller.v0.0.16
        version: 0.0.16
      - name: ack-ec2-controller.v0.0.15
        version: 0.0.15
      - name: ack-ec2-controller.v0.0.14
        version: 0.0.14
      - name: ack-ec2-controller.v0.0.13
        version: 0.0.13
      - name: ack-ec2-controller.v0.0.12
        version: 0.0.12
      - name: ack-ec2-controller.v0.0.11
        version: 0.0.11
      - name: ack-ec2-controller.v0.0.10
        version: 0.0.10
      - name: ack-ec2-controller.v0.0.9
        version: 0.0.9
      - name: ack-ec2-controller.v0.0.8
        version: 0.0.8
      - name: ack-ec2-controller.v0.0.7
        version: 0.0.7
      - name: ack-ec2-controller.v0.0.6
        version: 0.0.6
      - name: ack-ec2-controller.v0.0.5
        version: 0.0.5
      - name: ack-ec2-controller.v0.0.4
        version: 0.0.4
      - name: ack-ec2-controller.v0.0.3
        version: 0.0.3
      - name: ack-ec2-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-ec2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: global-load-balancer-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: global-load-balancer-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSRecord",
                "metadata": {
                  "name": "hello-global-record"
                },
                "spec": {
                  "endpoints": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    }
                  ],
                  "globalZoneRef": {
                    "name": "external-dns-zone"
                  },
                  "loadBalancingPolicy": "Multivalue",
                  "name": "hello.my.base.domain",
                  "ttl": 60
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSRecord",
                "metadata": {
                  "name": "route53-geoproximity-global-record-with-healthcheck"
                },
                "spec": {
                  "endpoints": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    }
                  ],
                  "globalZoneRef": {
                    "name": "route53-global-dns-zone"
                  },
                  "healthCheck": {
                    "failureThreshold": 3,
                    "httpGet": {
                      "host": "geoproximity-hc.${global_base_domain}",
                      "path": "/healthz",
                      "port": 80,
                      "scheme": "HTTP"
                    },
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 2
                  },
                  "loadBalancingPolicy": "Geoproximity",
                  "name": "geoproximity-hc.my.base.domain",
                  "ttl": 60
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSZone",
                "metadata": {
                  "name": "external-dns-zone"
                },
                "spec": {
                  "domain": "myglobalzone.mybasedomain.io",
                  "provider": {
                    "externalDNS": {
                      "annotations": {
                        "type": "global"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSZone",
                "metadata": {
                  "name": "route53-global-dns-zone"
                },
                "spec": {
                  "domain": "myglobalzone.mybasedomain.io",
                  "provider": {
                    "route53": {
                      "credentialsSecretRef": {
                        "name": "route53-global-zone-credentials",
                        "namespace": "mynamespace"
                      },
                      "zoneID": "Z083892239ROKNEP70BDY"
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalRouteDiscovery",
                "metadata": {
                  "name": "external-dns-globalroutediscovery"
                },
                "spec": {
                  "clusters": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3"
                    }
                  ],
                  "defaultLoadBalancingPolicy": "Multivalue",
                  "defaultTTL": 30,
                  "globalZoneRef": {
                    "name": "external-dns-zone"
                  },
                  "routeSelector": {
                    "matchLabels": {
                      "route-type": "global"
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalRouteDiscovery",
                "metadata": {
                  "name": "route53-globalroutediscovery"
                },
                "spec": {
                  "clusters": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3"
                    }
                  ],
                  "defaultLoadBalancingPolicy": "Multivalue",
                  "defaultTTL": 30,
                  "globalZoneRef": {
                    "name": "route53-global-dns-zone"
                  },
                  "routeSelector": {
                    "matchLabels": {
                      "route-type": "global"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Cloud Provider
          certified: "false"
          containerImage: quay.io/redhat-cop/global-load-balancer-operator@sha256:c0b6ac9b03571a74a56b28302d720e8aa4442233f7867d05ed365e022f7cb8f5
          createdAt: "2021-12-02T17:33:23Z"
          description: This operator creates automation around a DNS to operate as
            a global load balancer for a set of OpenShift clusters.
          operatorframework.io/suggested-namespace: global-load-balancer-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/global-load-balancer-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DNSEndpoint
            name: dnsendpoints.externaldns.k8s.io
            version: v1alpha1
          - description: GlobalDNSRecord is the Schema for the globaldnsrecords API
            displayName: Global DNSRecord
            kind: GlobalDNSRecord
            name: globaldnsrecords.redhatcop.redhat.io
            version: v1alpha1
          - description: GlobalDNSZone is the Schema for the globaldnszones API
            displayName: Global DNSZone
            kind: GlobalDNSZone
            name: globaldnszones.redhatcop.redhat.io
            version: v1alpha1
          - description: GlobalRouteDiscovery is the Schema for the globalroutediscoveries
              API
            displayName: Global Route Discovery
            kind: GlobalRouteDiscovery
            name: globalroutediscoveries.redhatcop.redhat.io
            version: v1alpha1
        description: "The global-load-balancer-operator implements automation to program
          a DNS service to act as global load balancer for applications deployed to
          multiple OpenShift clusters.\nThis operator is designed to be deployed to
          a control cluster which will watch the load balanced clusters (controlled
          clusters).\nThere are two main concepts (APIs) provided by this operator:\n\n1.
          GlobalDNSZone\n2. GlobalDNSRecord\n\n## GlobalDNSZone\n\nThe `GlobalDNSZone`
          CR allows you to configure a zone which will contain global load balanced
          records and the provider used to populate it.\nHere is an example of GlobalDNSZone:\n\n```yaml\napiVersion:
          redhatcop.redhat.io/v1alpha1\nkind: GlobalDNSZone\nmetadata:\n  name: external-dns-zone\nspec:\n
          \ # Add fields here\n  domain: global.myzone.io\n  provider:\n    externalDNS:\n
          \     annotations:\n        type: global\n```\n\nHere is a table summarizing
          the supported providers and their capabilities:\n\n| Provider  | Supports
          Health Checks  | Supports Multivalue LB | Supports Latency LB  | Supports
          GeoProximity LB  |\n|:--:|:--:|:---:|:---:|:---:|\n| External-dns  | no
          \ | yes | no  | no  |\n| Route53  | yes | yes(*)  | yes(*)  | yes(*)  |\n\n(*)
          only if all controlled clusters run on AWS.\n\n## GlobalDNSRecord\n\nThe
          `GlobalDNSRecord` CR allows you to specify the intention to create a global
          dns record. Here is an example\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GlobalDNSRecord\nmetadata:\n  name: hello-global-record\nspec:\n  name:
          hello.global.myzone.io\n  endpoints:\n  - clusterName: cluster1\n    clusterCredentialRef:\n
          \     name: kubeconfig\n      namespace: cluster1\n    loadBalancerServiceRef:\n
          \     name: router-default\n      namespace: openshift-ingress\n  - clusterName:
          cluster2\n    clusterCredentialRef:\n      name: kubeconfig\n      namespace:
          cluster2\n    loadBalancerServiceRef:\n      name: router-default\n      namespace:
          openshift-ingress\n  - clusterName: cluster3\n    clusterCredentialRef:\n
          \     name: kubeconfig\n      namespace: cluster3\n    loadBalancerServiceRef:\n
          \     name: router-default\n      namespace: openshift-ingress\n  ttl: 60\n
          \ loadBalancingPolicy: Multivalue\n  globalZoneRef:\n    name: external-dns-zone\n
          \ healthCheck:\n    failureThreshold: 3\n    httpGet:\n      host: hello.global.myzone.io\n
          \     httpHeaders:\n        - name: ciao\n          value: hello\n      path:
          /readyz\n      port: 80\n      scheme: HTTP\n    periodSeconds: 10\n    successThreshold:
          1\n    timeoutSeconds: 1     \n```\n\nThe list of endpoints identifies the
          set of `LoadBalancer` type of services to watch on the remote clusters (at
          the moment only LoadBalancer services are supported). These are the `LoadBalancer`
          services created by the [ingress controller](https://docs.openshift.com/container-platform/4.5/networking/ingress-operator.html)
          on which the routers rely. Each entry of this list requires a reference
          to the loadbalancer service on the remote cluster and a reference to a local
          secret containing the credential to connect to the remote cluster.\n\nThis
          secret must contain one entry called `kubeconfig` with a kubeconfig file
          with a default context pointing to the remote cluster. The account used
          by the kubeconfig file (presumably a service account) will require at a
          minimum cluster-level permissions as described in this [cluster role](./docs/scripts/remote-account-cluster-role.yaml).\n\nThe
          `globalZoneRef` field refers to a local (same namespace) `GlobalZone` CR.
          The DNS record represented by this `GlobalDNSRecord`, will be created in
          the referenced zone.\n\n`ttl` is the TTL of the crated DNS record.\n\n`loadBalancingPolicy`
          is the load balancing policy for this global record. It must match one of
          the policy supported by the provider of the referenced `GlobalZone`.\n\nFinally,
          `healthcheck` represent a [probe](https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.18/#probe-v1-core)
          to be used to test the health of a record. This field will be ignored if
          the provider does not support health checks.\n\n## External DNS Provider\n\nThe
          [external-dns](https://github.com/kubernetes-sigs/external-dns) provider
          delegates to external-dns the creation of the actual DNS records by creating
          a `DNSEndpoint` object.\nThe `DNSEndpoint` object will be created in the
          same namespace as the `GlobalDNSRecord` and will be owned by it.\nThe `DNSEdnpoint`
          object will have the same labels as the `GlobalDNSRecord` and the annotations
          specified in the GlobalDNSZone configuration.\nExternal-dns should be configured
          to watch for DNSEnpoints at the cluster level and to point to the desired
          provider.\nDetails on configuration can be found at the external-dns git
          repository.\nThe External-dns should be used as a fall back option when
          other options are not available as it does not support health checks and
          advanced load balancing policies.\n\n## AWS Route53 provider\n\nAWS Route53
          provider uses the Route53 service as a global loadbalancer and offers advanced
          routing capabilities via [route53 traffic policies](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/traffic-flow.html)
          (note that traffic policies will trigger an expense).\nThe following routing
          polices are currently supported:\n\n1. [Multivalue](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-multivalue)\n2.
          [Geoproximity](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-geoproximity)\n3.
          [Latency](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-latency)\n\nAWS
          Route53 provider at the moment requires that all the controlled clusters
          run in AWS.\n\nIf health checks are defined, a route53 health check originating
          from any reason (you have to ensure connectivity) will be created for each
          of the endpoint. Because the endpoint represent s a shared ELB (shared with
          other apps, that is) and the health check is app specific, we cannot use
          the ELB health check, so the route53 endpoint is created with one of the
          two IP exposed by the ELB. This is suboptimal, but it works in most situations.\n\n##
          Global Route Auto Discovery\n\nThe aboev examples showed how to create global
          DNS records. This can be good in some situations, but most of the times
          in an openshift deployment global DNS records will point to routes that
          are intended to be global.\nThe global-load-balancer operator can auto-discover
          these routes and automatically create the corresponding `GloablDNSRecord`.
          The `GlobalRouteDiscovery` CRD is used to configure the discovery process,
          here is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GlobalRouteDiscovery\nmetadata:\n  name: route53-globalroutediscovery\nspec:\n
          \ clusters:\n  - clusterName: cluster1\n    clusterCredentialRef:\n      name:
          ${cluster1_secret_name}\n      namespace: cluster1\n...\n  routeSelector:\n
          \   matchLabels:\n      route-type: global\n  defaultLoadBalancingPolicy:
          Multivalue\n  defaultTTL: 30 \n  globalZoneRef:\n    name: route53-global-dns-zone\n```\n\nThis
          global discovery route will discover routes in the provided list of cluster.
          Only the routes that match the route selector will be considered global.
          \nThe default load balancing policy and default TTL can be expressed in
          the `GlobalRouteDiscovery` CR. However with the following annotations, it's
          possible to configure route-specific values:\n\n* `global-load-balancer-operator.redhat-cop.io/load-balancing-policy`
          to set the load balancing policy\n* `global-load-balancer-operator.redhat-cop.io/ttl`
          to set the TTL\n\nThe globalZoneRef refers to the global zone to be used
          for the created `GlobalDNSRecords`.\n\nHealth checks will also be automatically
          discovered. If the pods behind the route expose a readiness check of httpGet
          kind, that configuration will be used to create the GlobalDNSRecord health
          check.\nWhen more than one container is present in the pod, by default the
          first container will be examined for health check. This behavior can be
          overridden with the this annotation on the route: `global-load-balancer-operator.redhat-cop.io/container-probe`
          where the value will container the name of the container with teh correct
          readiness probe.\n\nIf routes with the same namespace and name exists in
          multiple cluster, the following conditions must be met:\n\n* all host names
          must be the same\n* all load balancing policy must be the same\n* all TTLs
          must be the same\n* all discovered readiness checks must be the same\n\n##
          Examples\n\nThese examples are intended to help you setting up working configuration
          with each of the providers\n\n### Cluster Set up\n\nTwo approaches for cluster
          setup are provided\n\n1. [One cluster, three ingress-gateways.](https://github.com/redhat-cop/global-load-balancer-operator/docs/one-cluster-three-ingress-gateways.md)
          This approach is intended for development purposes and has the objective
          to keep resource consumption at the minimum.\n2. [Control cluster and three
          controlled clusters in different regions.](https://github.com/redhat-cop/global-load-balancer-operator/docs/three-clusters.md).
          This approach represents a more realistic set-up albeit it consumes more
          resources.\n\nYou can also set up the cluster on your own, at the end the
          following conditions must be met:\n\nThree namespace `cluster1` `cluster2`
          `cluster3` are created.\nthe following environment variables are initialized
          for each cluster:\n\n1. <cluster>_secret_name. Pointing to a secret in each
          of the cluster namespaces containing a valid kubeconfig fot that cluster\n2.
          <cluster>_service_name.  Pointing to the name of the loadbalancer service
          to be used for that cluster.\n3. <cluster>_service_namespace. Pointing to
          the namespace of the loadbalancer service to be used for that cluster.\n\nHere
          are examples for the supported provider:\n\n1. [Setting up external-dns
          as provider](https://github.com/redhat-cop/global-load-balancer-operator/docs/external-dns-provider.md)\n2.
          [Setting up route53 as a provider](https://github.com/redhat-cop/global-load-balancer-operator/docs/aws-route53-provider.md)\n"
        displayName: Global Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Networking
        - Cloud Provider
        - Global Load Balancing
        - GLB
        - GTM
        - route53
        - DNS
        links:
        - name: GigHub Repository
          url: https://github.com/redhat-cop/global-load-balancer-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/global-load-balancer-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/global-load-balancer-operator@sha256:c0b6ac9b03571a74a56b28302d720e8aa4442233f7867d05ed365e022f7cb8f5
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.4.2
      entries:
      - name: global-load-balancer-operator.v1.4.2
        version: 1.4.2
      - name: global-load-balancer-operator.v1.4.1
        version: 1.4.1
      - name: global-load-balancer-operator.v1.4.0
        version: 1.4.0
      - name: global-load-balancer-operator.v1.3.0
        version: 1.3.0
      - name: global-load-balancer-operator.v1.2.0
        version: 1.2.0
      - name: global-load-balancer-operator.v1.1.0
        version: 1.1.0
      - name: global-load-balancer-operator.v1.0.7
        version: 1.0.7
      - name: global-load-balancer-operator.v1.0.6
        version: 1.0.6
      - name: global-load-balancer-operator.v1.0.5
        version: 1.0.5
      - name: global-load-balancer-operator.v1.0.4
        version: 1.0.4
      - name: global-load-balancer-operator.v1.0.3
        version: 1.0.3
      - name: global-load-balancer-operator.v1.0.2
        version: 1.0.2
      - name: global-load-balancer-operator.v1.0.1
        version: 1.0.1
      - name: global-load-balancer-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: global-load-balancer-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Netscaler
      provider-url: https://www.netscaler.com
    name: netscaler-adm-agent-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netscaler-adm-agent-operator.v141.37.44
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "AdmAgent",
                "metadata": {
                  "name": "admagent"
                },
                "spec": {
                  "accessSecret": "",
                  "admURL": "adm.cloud.com",
                  "affinity": {},
                  "clusterPrefix": "",
                  "core": {
                    "livenessProbe": {
                      "failureThreshold": 5,
                      "initialDelaySeconds": 600,
                      "periodSeconds": 60,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "readinessProbe": {
                      "failureThreshold": 30,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 30,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "100m",
                        "memory": "2000Mi"
                      }
                    }
                  },
                  "fullnameOverride": "",
                  "helper": {
                    "image": "{{ .Values.helper.imageRegistry }}/{{ .Values.helper.imageRepository }}:{{ .Values.helper.imageTag }}",
                    "imageRegistry": "quay.io",
                    "imageRepository": "citrix/adm-agent-helper",
                    "imageTag": "2.0.1",
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "200Mi"
                      }
                    }
                  },
                  "hpa": {
                    "maxReplicas": 10,
                    "minReplicas": 1,
                    "required": false,
                    "targetCPUUtilization": "1000m",
                    "targetMemoryUtilization": "4000Mi"
                  },
                  "image": "{{ .Values.imageRegistry }}/{{ .Values.imageRepository }}:{{ .Values.imageTag }}",
                  "imagePullSecrets": [],
                  "imageRegistry": "quay.io",
                  "imageRepository": "citrix/adm-agent",
                  "imageTag": "14.1-37.44",
                  "kad": {
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "50m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "lic": {
                    "livenessProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 600,
                      "periodSeconds": 20,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "100m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "loginSecret": "",
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "redis": {
                    "livenessProbe": {
                      "failureThreshold": 5,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "readinessProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 20,
                      "periodSeconds": 10,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true
                  },
                  "sharding": {
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-adm-agent-operator@sha256:1c462c2d6161877f87d5eed4888f8d94c733501e67a94757ab0873d983fb3a4a
          createdAt: "2024-11-18T06:40:58Z"
          description: This is an operator to install ADM Agent
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler ADM Agent Operator
            displayName: NetScaler ADM Agent Operator
            kind: AdmAgent
            name: admagents.netscaler.com
            version: v1
        description: Netscaler ADM Agent Operator
        displayName: Netscaler ADM Agent Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        - ADC
        - Kubernetes
        - OpenShift
        - Citrix
        - Controller
        - Observability
        - Metrices
        - adm
        - agent
        - monitoring
        - proxy
        links:
        - name: Netscaler Adm Agent Operator
          url: https://github.com/netscaler/netscaler-helm-charts/blob/master/adm-agent/README.md
        maintainers:
        - email: pavan.belani@cloud.com
          name: Pavan Belani
        - email: swapnil.khanderaokakade@cloud.com
          name: Swapnil Khanderao
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Netscaler
          url: https://www.netscaler.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6b8c101124579dd46086ec83715618dd9a6aebe66176d676242e6c0b0900a53c
        - registry.connect.redhat.com/citrix/netscaler-adm-agent-operator@sha256:1c462c2d6161877f87d5eed4888f8d94c733501e67a94757ab0873d983fb3a4a
        - quay.io/citrix/adm-agent@sha256:45cab90524687c76a6d53f014b55063fa25c8aaeec0261f0f4f0efa6bff42968
        - quay.io/citrix/adm-agent-helper@sha256:f567a0637e8024f9d1674977f6c013741cc0fd12cf78703f8dbe01126561c3c4
        version: 141.37.44
      entries:
      - name: netscaler-adm-agent-operator.v141.37.44
        version: 141.37.44
      - name: netscaler-adm-agent-operator.v141.31.38
        version: 141.31.38
      - name: netscaler-adm-agent-operator.v141.25.33
        version: 141.25.33
      - name: netscaler-adm-agent-operator.v141.23.31
        version: 141.23.31
      - name: netscaler-adm-agent-operator.v141.21.41
        version: 141.21.41
      - name: netscaler-adm-agent-operator.v141.16.39
        version: 141.16.39
      - name: netscaler-adm-agent-operator.v141.13.23
        version: 141.13.23
      name: stable
    defaultChannel: stable
    packageName: netscaler-adm-agent-operator
    provider:
      name: Netscaler
      url: https://www.netscaler.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: pg4k-pgd
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pg4k-pgd.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgd.k8s.enterprisedb.io/v1beta1",
                "kind": "PGDGroup",
                "metadata": {
                  "name": "region-a"
                },
                "spec": {
                  "backup": null,
                  "cnp": {
                    "storage": {
                      "size": "1Gi"
                    },
                    "walStorage": {
                      "size": "1Gi"
                    }
                  },
                  "connectivity": {
                    "dns": {
                      "additional": [
                        {
                          "domain": "alternate.domain"
                        },
                        {
                          "domain": "my.domain",
                          "hostSuffix": "-dc1"
                        }
                      ]
                    },
                    "tls": {
                      "clientCert": {
                        "caCertSecret": "client-ca-key-pair",
                        "certManager": {
                          "spec": {
                            "issuerRef": {
                              "group": "cert-manager.io",
                              "kind": "Issuer",
                              "name": "client-ca-issuer"
                            }
                          }
                        }
                      },
                      "mode": "verify-ca",
                      "serverCert": {
                        "caCertSecret": "server-ca-key-pair",
                        "certManager": {
                          "spec": {
                            "issuerRef": {
                              "group": "cert-manager.io",
                              "kind": "Issuer",
                              "name": "server-ca-issuer"
                            }
                          }
                        }
                      }
                    }
                  },
                  "instances": 2,
                  "pgd": {
                    "parentGroup": {
                      "create": true,
                      "name": "world"
                    }
                  },
                  "proxyInstances": 2,
                  "witness": {
                    "storage": {
                      "size": "1Gi"
                    },
                    "walStorage": {
                      "size": "1Gi"
                    }
                  },
                  "witnessInstances": 1
                }
              },
              {
                "apiVersion": "pgd.k8s.enterprisedb.io/v1beta1",
                "kind": "PGDGroupCleanup",
                "metadata": {
                  "name": "region-b-cleanup"
                },
                "spec": {
                  "executor": "region-a",
                  "target": "region-b"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: docker.enterprisedb.com/k8s_enterprise_pgd/pg4k-pgd-openshift@sha256:5127d913011619d61fd06b27be3c709df2236a4b5c0254fa58632698d58b50fb
          createdAt: "2025-09-26T01:49:29Z"
          description: EDB Postgres Distributed for Kubernetes is an operator designed
            to manage EDB Postgres Distributed (PGD) v5 workloads on Kubernetes both
            in a single Kubernetes cluster or across multiple ones.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterImageCatalog
            name: clusterimagecatalogs.pgd.k8s.enterprisedb.io
            version: v1beta1
          - kind: ImageCatalog
            name: imagecatalogs.pgd.k8s.enterprisedb.io
            version: v1beta1
          - description: Object to clean up a PGD Group
            displayName: PGD Group Cleanup
            kind: PGDGroupCleanup
            name: pgdgroupcleanups.pgd.k8s.enterprisedb.io
            version: v1beta1
          - description: A group of Postgres Clusters
            displayName: PGD Group
            kind: PGDGroup
            name: pgdgroups.pgd.k8s.enterprisedb.io
            version: v1beta1
        description: |
          EDB Postgres Distributed for Kubernetes is an operator designed to manage EDB Postgres Distributed (PGD) v5 workloads on Kubernetes both in a single Kubernetes cluster or across multiple ones.

          EDB Postgres Distributed is an extension for PostgreSQL that provides multi-master replication and data distribution with advanced conflict management, data-loss protection, and throughput up to 5X faster than native logical replication. It enables distributed PostgreSQL clusters with high availability up to five 9s.

          This operator requires access to the EDB private repositories.
        displayName: EDB Postgres Distributed for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - distributed
        - postgres
        - database
        - sql
        - cloud
        - native
        - edb
        - enterprisedb
        - cnp
        - pg4k
        - pgd
        links:
        - name: 'Documentation: EDB Postgres Distributed v5'
          url: https://www.enterprisedb.com/docs/pgd/latest/
        - name: 'Documentation: EDB Postgres Distributed for Kubernetes'
          url: https://www.enterprisedb.com/docs/postgres_distributed_for_kubernetes/latest/
        maintainers:
        - email: sales@enterprisedb.com
          name: Sales
        - email: support@enterprisedb.com
          name: Support
        maturity: alpha
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - docker.enterprisedb.com/k8s_enterprise_pgd/pg4k-pgd-openshift@sha256:5127d913011619d61fd06b27be3c709df2236a4b5c0254fa58632698d58b50fb
        version: 1.1.3
      entries:
      - name: pg4k-pgd.v1.1.3
        version: 1.1.3
      - name: pg4k-pgd.v1.1.2
        version: 1.1.2
      name: stable
    defaultChannel: stable
    packageName: pg4k-pgd
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Mantisnet
      provider-url: https://www.mantisnet.com
    name: tawon-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tawon-operator.v2.29.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Directive",
                "metadata": {
                  "name": "amf-h2"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "amf"
                    }
                  },
                  "container": {
                    "resources": {
                      "limits": {
                        "cpu": "500m"
                      }
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  },
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 8080"
                      },
                      "task": "payload"
                    },
                    {
                      "task": "h2"
                    },
                    {
                      "config": {
                        "addr": "nats://nats:4222",
                        "topic": "hello"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "TopologyAggregator",
                "metadata": {
                  "name": "topology-aggregator"
                },
                "spec": {
                  "messaging": {
                    "nats": {
                      "url": "nats://nats:4222"
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-11-13T01:16:29Z"
          olm.skipRange: '>=2.23.0 <2.29.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Directive allows to define a Tawon task pipeline that will
              be applied to a Kubernetes cluster.
            displayName: Tawon Directive
            kind: Directive
            name: directives.tawon.mantisnet.com
            version: v1alpha1
          - description: TopologyAggregator is the Schema for the topologyaggregators
              API
            displayName: Tawon Topology Aggregator
            kind: TopologyAggregator
            name: topologyaggregators.tawon.mantisnet.com
            version: v1alpha1
        description: Tawon is a cloud-native observability fabric for deeply instrumenting
          microservice-based applications.
        displayName: Tawon
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - networking
        - metrics
        - ebpf
        links:
        - name: Tawon Docs
          url: https://docs.tawon.xmantis.net
        maintainers:
        - email: elliott@mantisnet.com
          name: Elliott Starin
        - email: nicolas@mantisnet.com
          name: Nicolas Paton
        - email: thiago@mantisnet.com
          name: Thiago Navarro
        maturity: alpha
        provider:
          name: Mantisnet
          url: https://www.mantisnet.com
        relatedImages:
        - quay.io/mantisnet/kernel-headers@sha256:242e837af831b6eed98d522b6bae7416928b9fbe451fb6e02ffa678686771068
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - quay.io/mantisnet/tawon-operator@sha256:d5669dde46bb30411964baa86844d97721655886769b96809626246a4123caad
        - quay.io/mantisnet/kernel-headers@sha256:2aa3bc7871fe0fc0a79ac812d941e55cd0d25a5f0d1d9b3e2435aab6f3d70c67
        - quay.io/mantisnet/tawonctl@sha256:d436ef902bb08d830f64f34fdb86b20d2bd18d86233c8240014f0722c60e5e94
        - quay.io/mantisnet/tawon@sha256:83575079fe00ae6e53bb13005b94ef0047cc1974d95015c19e64dd53e13bcb93
        - quay.io/mantisnet/nats@sha256:6afc7ced92181a7c12d709e9693740847297dc5d60ab78cbac7b9b6c859ecdd1
        version: 2.29.1
      entries:
      - name: tawon-operator.v2.29.1
        version: 2.29.1
      name: alpha
    - currentCSV: tawon-operator.v2.39.27-rc6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "ClusterDirective",
                "metadata": {
                  "name": "capture-coredns"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "coredns"
                    }
                  },
                  "debug": {
                    "logLevel": "debug"
                  },
                  "streams": [
                    {
                      "maxage": "6h0m0s",
                      "maxmsgs": 1000,
                      "name": "coredns",
                      "retentionPolicy": "Delete"
                    }
                  ],
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 53"
                      },
                      "task": "capture"
                    },
                    {
                      "config": {
                        "name": "coredns",
                        "type": "stream"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Dashboard",
                "metadata": {
                  "name": "tawon-dashboard"
                },
                "spec": null
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Directive",
                "metadata": {
                  "name": "capture-coredns"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "coredns"
                    }
                  },
                  "debug": {
                    "logLevel": "debug"
                  },
                  "streams": [
                    {
                      "maxage": "6h0m0s",
                      "maxmsgs": 1000,
                      "name": "coredns",
                      "retentionPolicy": "Delete"
                    }
                  ],
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 53"
                      },
                      "task": "capture"
                    },
                    {
                      "config": {
                        "name": "coredns",
                        "type": "stream"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Stream",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "tawon-operator",
                    "app.kubernetes.io/instance": "stream-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "stream",
                    "app.kubernetes.io/part-of": "tawon-operator"
                  },
                  "name": "stream-sample"
                },
                "spec": {
                  "maxage": "6h",
                  "maxbytes": "1G",
                  "maxmsgs": 100000,
                  "store": {
                    "name": "mystreamstore"
                  }
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "StreamStore",
                "metadata": {
                  "name": "mystreamstore"
                },
                "spec": {
                  "defaults": {
                    "maxage": "240h",
                    "maxbytes": "1Gi",
                    "maxsize": 1000000
                  },
                  "limits": {
                    "max": 10000000,
                    "maxage": "2400h",
                    "maxbytes": "2Gi"
                  },
                  "persistentVolumeClaimRetentionPolicy": {
                    "whenDeleted": "Delete"
                  },
                  "volumeClaimTemplate": {
                    "metadata": {
                      "name": "mypvclaim"
                    },
                    "spec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "ssd-csi"
                    }
                  }
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "TopologyAggregator",
                "metadata": {
                  "name": "topology-aggregator"
                },
                "spec": {
                  "messaging": {
                    "nats": {
                      "url": "nats://nats:4222"
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing, Monitoring, Networking
          containerImage: quay.io/mantisnet/tawon-operator@sha256:d8ce5c8966d3590ce5f286d8e778463b85582b970ad7730ddf7828bc68843a90
          createdAt: "2025-03-08T18:58:04Z"
          description: Tawon is a cloud-native observability fabric for deeply instrumenting
            microservice-based applications.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Mantisnet Tawon Operator Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: MantisNet
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterDirective is the Schema for the clusterdirectives
              API.
            displayName: Cluster Tawon Directive
            kind: ClusterDirective
            name: clusterdirectives.tawon.mantisnet.com
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.tawon.mantisnet.com
            version: v1alpha1
          - description: A Directive allows to define a Tawon task pipeline that will
              be applied to a Kubernetes cluster.
            displayName: Tawon Directive
            kind: Directive
            name: directives.tawon.mantisnet.com
            version: v1alpha1
          - description: Stream is a persistent Tawon data stream.
            displayName: Stream
            kind: Stream
            name: streams.tawon.mantisnet.com
            version: v1alpha1
          - description: StreamStore is the Schema for the streamstores API
            displayName: Stream Store
            kind: StreamStore
            name: streamstores.tawon.mantisnet.com
            version: v1alpha1
          - description: TopologyAggregator is the Schema for the topologyaggregators
              API
            displayName: Tawon Topology Aggregator
            kind: TopologyAggregator
            name: topologyaggregators.tawon.mantisnet.com
            version: v1alpha1
        description: Tawon is a cloud-native observability fabric for deeply instrumenting
          microservice-based applications.
        displayName: Tawon
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - networking
        - metrics
        - ebpf
        links:
        - name: Tawon Docs
          url: https://docs.tawon.xmantis.net
        maintainers:
        - email: elliott@mantisnet.com
          name: Elliott Starin
        - email: nicolas@mantisnet.com
          name: Nicolas Paton
        - email: thiago@mantisnet.com
          name: Thiago Navarro
        maturity: alpha
        provider:
          name: Mantisnet
          url: https://www.mantisnet.com
        relatedImages:
        - quay.io/mantisnet/tawonctl@sha256:8bf40ca0cc9984a252363e87d6135e43ac52839113b762399cc54a085af171d8
        - quay.io/mantisnet/nats@sha256:5f9e7f8f685fd0c1279a3465b61c7215f4600c99d81785dda86a71c966cf829a
        - quay.io/mantisnet/kernel-headers@sha256:ca3baeba7ab8ce968c1ee129112532fc8b8d6fb6741e42831b6100f0a2451629
        - quay.io/openshift/origin-oauth-proxy@sha256:28856aae9e481fe763776df713b4b335632645687e6fd2e31619f76d11e0f47c
        - quay.io/mantisnet/tawon-tests@sha256:cd49c24deb7cc9f4b439cb9baa885208dfe91de385cd8c5d312fd08280615c2c
        - quay.io/mantisnet/tawon@sha256:46e654c6f275be2be604254102aaee9d927859cfd05dfa16b7cb571d36b0f1a2
        - quay.io/mantisnet/tawon-operator@sha256:d8ce5c8966d3590ce5f286d8e778463b85582b970ad7730ddf7828bc68843a90
        - quay.io/mantisnet/sengat@sha256:00c3e1194c2c3e343d16c1543b7ce2a1ee9829f6ddd1c2870785624596b46ebf
        - quay.io/mantisnet/kernel-headers@sha256:5d00bfa9f60c95afcfa363ede59a9625fef9925cd1cb4235327a1de94a0fc632
        - quay.io/mantisnet/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/mantisnet/kernel-headers@sha256:9290eb56b186ee0edf65da615c1a8aabc66f7d290168e49ad8764ab8fc2d209a
        - quay.io/mantisnet/diagnose@sha256:ab15f4eaeda32f510d81a3cc865b0fd77405b55f6a0c6f285131331dcc38f2ca
        - quay.io/mantisnet/tawon@sha256:5a8bf475ff18abacb96f631e4d3ea01c94bef8a38f84c00d1dc4047073b6cc14
        version: 2.39.27-rc6
      entries:
      - name: tawon-operator.v2.39.27-rc6
        version: 2.39.27-rc6
      name: candidate-v2
    - currentCSV: tawon-operator.v2.39.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "ClusterDirective",
                "metadata": {
                  "name": "capture-coredns"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "coredns"
                    }
                  },
                  "debug": {
                    "logLevel": "debug"
                  },
                  "streams": [
                    {
                      "maxage": "6h0m0s",
                      "maxmsgs": 1000,
                      "name": "coredns",
                      "retentionPolicy": "Delete"
                    }
                  ],
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 53"
                      },
                      "task": "capture"
                    },
                    {
                      "config": {
                        "name": "coredns",
                        "type": "stream"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Dashboard",
                "metadata": {
                  "name": "tawon-dashboard"
                },
                "spec": null
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Directive",
                "metadata": {
                  "name": "capture-coredns"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "coredns"
                    }
                  },
                  "debug": {
                    "logLevel": "debug"
                  },
                  "streams": [
                    {
                      "maxage": "6h0m0s",
                      "maxmsgs": 1000,
                      "name": "coredns",
                      "retentionPolicy": "Delete"
                    }
                  ],
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 53"
                      },
                      "task": "capture"
                    },
                    {
                      "config": {
                        "name": "coredns",
                        "type": "stream"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Stream",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "tawon-operator",
                    "app.kubernetes.io/instance": "stream-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "stream",
                    "app.kubernetes.io/part-of": "tawon-operator"
                  },
                  "name": "stream-sample"
                },
                "spec": {
                  "maxage": "6h",
                  "maxbytes": "1G",
                  "maxmsgs": 100000,
                  "store": {
                    "name": "mystreamstore"
                  }
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "StreamStore",
                "metadata": {
                  "name": "mystreamstore"
                },
                "spec": {
                  "defaults": {
                    "maxage": "240h",
                    "maxbytes": "1Gi",
                    "maxsize": 1000000
                  },
                  "limits": {
                    "max": 10000000,
                    "maxage": "2400h",
                    "maxbytes": "2Gi"
                  },
                  "persistentVolumeClaimRetentionPolicy": {
                    "whenDeleted": "Delete"
                  },
                  "volumeClaimTemplate": {
                    "metadata": {
                      "name": "mypvclaim"
                    },
                    "spec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "ssd-csi"
                    }
                  }
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "TopologyAggregator",
                "metadata": {
                  "name": "topology-aggregator"
                },
                "spec": {
                  "messaging": {
                    "nats": {
                      "url": "nats://nats:4222"
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing, Monitoring, Networking
          containerImage: quay.io/mantisnet/tawon-operator@sha256:f831abf0f9fb3a4e096743d4ff7356984738da25790e305325e740ca97f0be5b
          createdAt: "2025-07-18T16:24:07Z"
          description: Tawon is a cloud-native observability fabric for deeply instrumenting
            microservice-based applications.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.23.0 <2.39.32'
          operators.openshift.io/valid-subscription: Mantisnet Tawon Operator Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: MantisNet
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterDirective is the Schema for the clusterdirectives
              API.
            displayName: Cluster Tawon Directive
            kind: ClusterDirective
            name: clusterdirectives.tawon.mantisnet.com
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.tawon.mantisnet.com
            version: v1alpha1
          - description: A Directive allows to define a Tawon task pipeline that will
              be applied to a Kubernetes cluster.
            displayName: Tawon Directive
            kind: Directive
            name: directives.tawon.mantisnet.com
            version: v1alpha1
          - description: Stream is a persistent Tawon data stream.
            displayName: Stream
            kind: Stream
            name: streams.tawon.mantisnet.com
            version: v1alpha1
          - description: StreamStore is the Schema for the streamstores API
            displayName: Stream Store
            kind: StreamStore
            name: streamstores.tawon.mantisnet.com
            version: v1alpha1
          - description: TopologyAggregator is the Schema for the topologyaggregators
              API
            displayName: Tawon Topology Aggregator
            kind: TopologyAggregator
            name: topologyaggregators.tawon.mantisnet.com
            version: v1alpha1
        description: Tawon is a cloud-native observability fabric for deeply instrumenting
          microservice-based applications.
        displayName: Tawon
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - networking
        - metrics
        - ebpf
        links:
        - name: Tawon Docs
          url: https://docs.tawon.xmantis.net
        maintainers:
        - email: elliott@mantisnet.com
          name: Elliott Starin
        - email: nicolas@mantisnet.com
          name: Nicolas Paton
        - email: thiago@mantisnet.com
          name: Thiago Navarro
        maturity: alpha
        provider:
          name: Mantisnet
          url: https://www.mantisnet.com
        relatedImages:
        - quay.io/mantisnet/tawon-tests@sha256:294df1580816919502dc592c1c28ffff32b13ee892686f783eff181863fa9d18
        - quay.io/mantisnet/kernel-headers@sha256:9290eb56b186ee0edf65da615c1a8aabc66f7d290168e49ad8764ab8fc2d209a
        - quay.io/mantisnet/diagnose@sha256:ab15f4eaeda32f510d81a3cc865b0fd77405b55f6a0c6f285131331dcc38f2ca
        - quay.io/mantisnet/kernel-headers@sha256:ca3baeba7ab8ce968c1ee129112532fc8b8d6fb6741e42831b6100f0a2451629
        - quay.io/mantisnet/kernel-headers@sha256:b0975a44cc94bc1d75679f288a9e23a65c2c0b350ce1fb2f5eb421a76a9567c6
        - quay.io/mantisnet/nats@sha256:5f9e7f8f685fd0c1279a3465b61c7215f4600c99d81785dda86a71c966cf829a
        - quay.io/mantisnet/tawonctl@sha256:88ac81b4fe00d360c5ff127f77456f9e36990e435442a9194d3dbf50b20e7206
        - quay.io/mantisnet/tawon@sha256:846d03ada7e6f6648fc74fd0ec9e68c3c6b17e4c155df22829a3db3ca556c395
        - quay.io/mantisnet/kernel-headers@sha256:5d00bfa9f60c95afcfa363ede59a9625fef9925cd1cb4235327a1de94a0fc632
        - quay.io/openshift/origin-oauth-proxy@sha256:28856aae9e481fe763776df713b4b335632645687e6fd2e31619f76d11e0f47c
        - quay.io/mantisnet/alpine@sha256:ec1b05d1eac264d9204a57f4ad9d4dc35e9e756e9fedaea0674aefc7edb1d6a4
        - quay.io/mantisnet/tawon@sha256:1640758505d9b4e0f684ab502ef3344135a270192054e256e7e711209fa51adf
        - quay.io/mantisnet/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/mantisnet/tawon-operator@sha256:f831abf0f9fb3a4e096743d4ff7356984738da25790e305325e740ca97f0be5b
        - quay.io/mantisnet/sengat@sha256:483fbce2bacbd95b7458773e2278b936f06a19e65acb8987adffa8403d76e40a
        - quay.io/mantisnet/nginx@sha256:c9f91949187fa1c2b4615b88d3acf7902c7e2d4a2557f33ca0cf90164269a7ae
        version: 2.39.32
      entries:
      - name: tawon-operator.v2.39.32
        version: 2.39.32
      - name: tawon-operator.v2.39.31
        version: 2.39.31
      - name: tawon-operator.v2.39.29
        version: 2.39.29
      - name: tawon-operator.v2.39.28
        version: 2.39.28
      - name: tawon-operator.v2.39.27
        version: 2.39.27
      name: fast-v2
    defaultChannel: fast-v2
    packageName: tawon-operator
    provider:
      name: Mantisnet
      url: https://www.mantisnet.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Perficient Inc
      provider-url: http://www.perficient.com
    name: devopsinabox
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devopsinabox.v0.0.45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "DevSecOpsManager",
                "metadata": {
                  "name": "devsecopsmanager-sample"
                },
                "spec": {
                  "config": {
                    "repoPath": "",
                    "repoUrl": ""
                  },
                  "forwarding": {
                    "fluentd": {
                      "affinity": {},
                      "caFile": "files/default-openshift-logging-fluentd.crt",
                      "image": "registry.redhat.io/openshift4/ose-logging-fluentd:v4.4",
                      "keyFile": "files/default-openshift-logging-fluentd.key",
                      "loglevel": "warn",
                      "nodeSelector": {},
                      "passphrase": "",
                      "persistence": {
                        "accessMode": "ReadWriteOnce",
                        "enabled": false,
                        "size": "10Gi",
                        "storageClass": ""
                      },
                      "port": 24224,
                      "replicas": 2,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "512Mi"
                        }
                      },
                      "sharedkey": "splunkforwarding",
                      "ssl": true,
                      "tolerations": [],
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      }
                    },
                    "preserveDefault": true
                  },
                  "fullnameOverride": "",
                  "git": {
                    "password": "",
                    "server": "https://gitlab.com",
                    "token": "",
                    "username": ""
                  },
                  "nameOverride": "",
                  "prometheus": {
                    "token": ""
                  }
                }
              },
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "ManagedNamespace",
                "metadata": {
                  "name": "managednamespace-sample"
                },
                "spec": {
                  "config": {
                    "repoBranch": "master",
                    "repoPath": "",
                    "repoUrl": ""
                  },
                  "container": {
                    "password": "",
                    "registry": "https://quay.io",
                    "username": ""
                  },
                  "fullnameOverride": "",
                  "git": {
                    "password": "",
                    "server": "https://gitlab.com",
                    "token": "",
                    "username": ""
                  },
                  "manager": {
                    "namespace": ""
                  },
                  "mirrors": {
                    "maven": ""
                  },
                  "nameOverride": ""
                }
              },
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "SpringBootApp",
                "metadata": {
                  "name": "springbootapp-sample"
                },
                "spec": {
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "config": {
                    "repo": ""
                  },
                  "container": {
                    "image": "placeholder",
                    "registry": ""
                  },
                  "fullnameOverride": "",
                  "git": {
                    "branch": "master",
                    "module": "",
                    "url": ""
                  },
                  "management": {
                    "port": 8081
                  },
                  "maven": {
                    "mirror": ""
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "replicaCount": 1,
                  "resources": {},
                  "service": {
                    "port": 8080
                  },
                  "serviceAccount": {
                    "annotations": {}
                  },
                  "version": {
                    "branch": "",
                    "image": "placeholder",
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration,Developer Tools,Application Runtime
          containerImage: quay.io/mrethers/devopsinabox-operator:v0.0.45
          createdAt: 2021-03-03 10:51:09Z
          description: Perficient's DevSecOps Manager For Kubernetes Applications
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/prft-rh/devopsinabox-operator
          support: Matthieu Rethers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DevSecOpsManager
            name: devsecopsmanagers.apimc.com.perficient
            version: v1alpha1
          - kind: ManagedNamespace
            name: managednamespaces.apimc.com.perficient
            version: v1alpha1
          - kind: SpringBootApp
            name: springbootapps.apimc.com.perficient
            version: v1alpha1
        description: "The devops-in-a-box Operator manages the entire lifecycle of
          your applications in Kubernetes. It encapsulates all the best devsecops
          practices gathered by Perficient over the years working on many large projects.
          Use it to migrate existing applications with minimum changes or deploy new
          ones with little to no knowledge of Kubernetes.\n\nOnce installed, the Devops-in-a-box
          operator provides the following features:\n\n* **Create/Destroy**: Easily
          launch a new DevSecOpsManager instance to bootstrap your cluster and start
          deploying new applications with total confidence.\n\n* **Gitops Simplified**:
          Configure your cluster and individual applications through Git and watch
          your changes propagate instantly. Just point to where your code is and let
          the operator do the rest with preconfigured cloud-native CI/CD pipelines.\n\n*
          **Production-Ready**: Automatically register your applications with Openshift
          logging and monitoring systems and enjoy special additions for Spring Boot
          like custom dashboards and searchable logs.\n\n*Contact us on Github for
          installation and usage instruction*\n\n### Other Supported Features\n\n**Blue/Green,
          Canary, A/B Deployments**\n\nRun multiple versions of your applications
          simultaneously and configure traffic rules to target specific groups of
          users using Service Mesh.\n\n**Updates via automated operations**\n\nGet
          the latest enhancements automatically through the Operator Lifecycle Manager
          and upgrade all your running applications at once. "
        displayName: Devops-in-a-box
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devops
        - devsecops
        - spring
        - boot
        - java
        - gitops
        - cicd
        - argo
        links:
        - name: Devops-in-a-box Operator
          url: https://github.com/prft-rh/devopsinabox-operator
        maintainers:
        - email: matthieu.rethers@perficient.com
          name: Matthieu Rethers
        maturity: alpha
        provider:
          name: Perficient Inc
          url: http://www.perficient.com
        relatedImages:
        - quay.io/mrethers/devopsinabox-operator:v0.0.45
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        version: 0.0.45
      entries:
      - name: devopsinabox.v0.0.45
        version: 0.0.45
      name: alpha
    defaultChannel: alpha
    packageName: devopsinabox
    provider:
      name: Perficient Inc
      url: http://www.perficient.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc
      provider-url: ""
    name: bare-metal-event-relay
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bare-metal-event-relay.v4.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "event.redhat-cne.org/v1alpha1",
                "kind": "HardwareEvent",
                "metadata": {
                  "name": "hardwareevent-sample"
                },
                "spec": {
                  "nodeSelector": {
                    "node-role.kubernetes.io/worker": ""
                  },
                  "storageType": "emptyDir",
                  "transportHost": "http://hw-event-publisher-service.openshift-bare-metal-events.svc.cluster.local:9043"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/openshift4/baremetal-hardware-event-proxy-rhel9@sha256:aaaea040164e1fd801e13c74024b2307997cdde954c1a5e35efd90023d2d7376
          createdAt: "2023-02-23"
          description: This software manages the lifecycle of the hw-event-proxy container.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.22.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/hw-event-proxy
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HardwareEvent is the Schema for the hardwareevents API
            displayName: Hardware Event
            kind: HardwareEvent
            name: hardwareevents.event.redhat-cne.org
            version: v1alpha1
        description: |
          ## Funtionality
          Install and manage a hardware event delivery framework.
          ## Installation
          To start using Hardware event proxy, install the operator and then create the following HardwareEvent:
          ```
          apiVersion: "event.redhat-cne.org/v1alpha1"
          kind: "HardwareEvent"
          metadata:
            name: "hardware-event"
          spec:
            nodeSelector:
              node-role.kubernetes.io/worker: ""
            transportHost: "http://hw-event-publisher-service.openshift-bare-metal-events.svc.cluster.local:9043"

          ```
          This operator needs a secret to be created to access Redfish
          Message Registry.
          The secret name and spec the operand expects are under the same namespace
          the operand is deployed on. For example:
          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: redfish-basic-auth
          type: Opaque
          stringData:
            username: <bmc_username>
            password: <bmc_password>
            # BMC host DNS or IP address
            hostaddr: <bmc_host_ip_address>
          ```
        displayName: Bare Metal Event Relay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hardware
        - event
        - cloudevents
        links:
        - name: Hardware Event Proxy Operator
          url: https://github.com/redhat-cne/hw-event-proxy-operator
        maturity: alpha
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:6b7beed3cba9093b3aeb3109f23ec38c20f5568915dff774ad5e9e6bb5db3a58
        - registry.redhat.io/openshift4/bare-metal-event-relay-rhel9-operator@sha256:2219f2f623ee27ef83518d28857995287de0a70cefa193792f28acb352ed8fc0
        - registry.redhat.io/openshift4/baremetal-hardware-event-proxy-rhel9@sha256:aaaea040164e1fd801e13c74024b2307997cdde954c1a5e35efd90023d2d7376
        - registry.redhat.io/openshift4/ose-cloud-event-proxy-rhel9@sha256:c99b448820cf03a33e9237030ac74c1d1ad6a4ce17f1b2840c69b039097c5733
        version: 4.16.1
      entries:
      - name: bare-metal-event-relay.v4.16.1
        version: 4.16.1
      name: "4.16"
    - currentCSV: bare-metal-event-relay.v4.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "event.redhat-cne.org/v1alpha1",
                "kind": "HardwareEvent",
                "metadata": {
                  "name": "hardwareevent-sample"
                },
                "spec": {
                  "nodeSelector": {
                    "node-role.kubernetes.io/worker": ""
                  },
                  "storageType": "emptyDir",
                  "transportHost": "http://hw-event-publisher-service.openshift-bare-metal-events.svc.cluster.local:9043"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/openshift4/baremetal-hardware-event-proxy-rhel9@sha256:aaaea040164e1fd801e13c74024b2307997cdde954c1a5e35efd90023d2d7376
          createdAt: "2023-02-23"
          description: This software manages the lifecycle of the hw-event-proxy container.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.22.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/hw-event-proxy
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HardwareEvent is the Schema for the hardwareevents API
            displayName: Hardware Event
            kind: HardwareEvent
            name: hardwareevents.event.redhat-cne.org
            version: v1alpha1
        description: |
          ## Funtionality
          Install and manage a hardware event delivery framework.
          ## Installation
          To start using Hardware event proxy, install the operator and then create the following HardwareEvent:
          ```
          apiVersion: "event.redhat-cne.org/v1alpha1"
          kind: "HardwareEvent"
          metadata:
            name: "hardware-event"
          spec:
            nodeSelector:
              node-role.kubernetes.io/worker: ""
            transportHost: "http://hw-event-publisher-service.openshift-bare-metal-events.svc.cluster.local:9043"

          ```
          This operator needs a secret to be created to access Redfish
          Message Registry.
          The secret name and spec the operand expects are under the same namespace
          the operand is deployed on. For example:
          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: redfish-basic-auth
          type: Opaque
          stringData:
            username: <bmc_username>
            password: <bmc_password>
            # BMC host DNS or IP address
            hostaddr: <bmc_host_ip_address>
          ```
        displayName: Bare Metal Event Relay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hardware
        - event
        - cloudevents
        links:
        - name: Hardware Event Proxy Operator
          url: https://github.com/redhat-cne/hw-event-proxy-operator
        maturity: alpha
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:6b7beed3cba9093b3aeb3109f23ec38c20f5568915dff774ad5e9e6bb5db3a58
        - registry.redhat.io/openshift4/bare-metal-event-relay-rhel9-operator@sha256:2219f2f623ee27ef83518d28857995287de0a70cefa193792f28acb352ed8fc0
        - registry.redhat.io/openshift4/ose-cloud-event-proxy-rhel9@sha256:c99b448820cf03a33e9237030ac74c1d1ad6a4ce17f1b2840c69b039097c5733
        - registry.redhat.io/openshift4/baremetal-hardware-event-proxy-rhel9@sha256:aaaea040164e1fd801e13c74024b2307997cdde954c1a5e35efd90023d2d7376
        version: 4.16.1
      entries:
      - name: bare-metal-event-relay.v4.16.1
        version: 4.16.1
      name: stable
    defaultChannel: stable
    packageName: bare-metal-event-relay
    provider:
      name: Red Hat Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ocs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocs-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |2-

            [
                {
                    "apiVersion": "ocs.openshift.io/v1",
                    "kind": "StorageCluster",
                    "metadata": {
                        "name": "example-storagecluster",
                        "namespace": "openshift-storage"
                    },
                    "spec": {
                        "manageNodes": false,
                        "monPVCTemplate": {
                            "spec": {
                                "accessModes": [
                                    "ReadWriteOnce"
                                ],
                                "resources": {
                                    "requests": {
                                        "storage": "10Gi"
                                    }
                                },
                                "storageClassName": "gp2-csi"
                            }
                        },
                        "storageDeviceSets": [
                            {
                                "count": 3,
                                "dataPVCTemplate": {
                                    "spec": {
                                        "accessModes": [
                                            "ReadWriteOnce"
                                        ],
                                        "resources": {
                                            "requests": {
                                                "storage": "1Ti"
                                            }
                                        },
                                        "storageClassName": "gp2-csi",
                                        "volumeMode": "Block"
                                    }
                                },
                                "name": "example-deviceset",
                                "placement": {},
                                "portable": true,
                                "resources": {}
                            }
                        ]
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Storage
          containerImage: registry.redhat.io/odf4/ocs-rhel9-operator@sha256:b1a9492bd9b314f86b4e192cf2c83144885ff863d9932e386a41691f50be6193
          description: Red Hat OpenShift Container Storage provides hyperconverged
            storage for applications within an OpenShift cluster.
          external.features.ocs.openshift.io/export-script: |-
            IiIiCkNvcHlyaWdodCAyMDIwIFRoZSBSb29rIEF1dGhvcnMuIEFsbCByaWdodHMgcmVzZXJ2ZWQu
            CgpMaWNlbnNlZCB1bmRlciB0aGUgQXBhY2hlIExpY2Vuc2UsIFZlcnNpb24gMi4wICh0aGUgIkxp
            Y2Vuc2UiKTsKeW91IG1heSBub3QgdXNlIHRoaXMgZmlsZSBleGNlcHQgaW4gY29tcGxpYW5jZSB3
            aXRoIHRoZSBMaWNlbnNlLgpZb3UgbWF5IG9idGFpbiBhIGNvcHkgb2YgdGhlIExpY2Vuc2UgYXQK
            CglodHRwOi8vd3d3LmFwYWNoZS5vcmcvbGljZW5zZXMvTElDRU5TRS0yLjAKClVubGVzcyByZXF1
            aXJlZCBieSBhcHBsaWNhYmxlIGxhdyBvciBhZ3JlZWQgdG8gaW4gd3JpdGluZywgc29mdHdhcmUK
            ZGlzdHJpYnV0ZWQgdW5kZXIgdGhlIExpY2Vuc2UgaXMgZGlzdHJpYnV0ZWQgb24gYW4gIkFTIElT
            IiBCQVNJUywKV0lUSE9VVCBXQVJSQU5USUVTIE9SIENPTkRJVElPTlMgT0YgQU5ZIEtJTkQsIGVp
            dGhlciBleHByZXNzIG9yIGltcGxpZWQuClNlZSB0aGUgTGljZW5zZSBmb3IgdGhlIHNwZWNpZmlj
            IGxhbmd1YWdlIGdvdmVybmluZyBwZXJtaXNzaW9ucyBhbmQKbGltaXRhdGlvbnMgdW5kZXIgdGhl
            IExpY2Vuc2UuCiIiIgoKaW1wb3J0IGVycm5vCmltcG9ydCBzeXMKaW1wb3J0IGpzb24KaW1wb3J0
            IGFyZ3BhcnNlCmltcG9ydCByZQppbXBvcnQgc3VicHJvY2VzcwppbXBvcnQgaG1hYwpmcm9tIGhh
            c2hsaWIgaW1wb3J0IHNoYTEgYXMgc2hhCmZyb20gb3MgaW1wb3J0IGxpbmVzZXAgYXMgTElORVNF
            UApmcm9tIG9zIGltcG9ydCBwYXRoCmZyb20gZW1haWwudXRpbHMgaW1wb3J0IGZvcm1hdGRhdGUK
            aW1wb3J0IHJlcXVlc3RzCmZyb20gcmVxdWVzdHMuYXV0aCBpbXBvcnQgQXV0aEJhc2UKCnB5M2sg
            PSBGYWxzZQppZiBzeXMudmVyc2lvbl9pbmZvLm1ham9yID49IDM6CiAgICBweTNrID0gVHJ1ZQog
            ICAgaW1wb3J0IHVybGxpYi5wYXJzZQogICAgZnJvbSBpcGFkZHJlc3MgaW1wb3J0IGlwX2FkZHJl
            c3MsIElQdjRBZGRyZXNzCgpNb2R1bGVOb3RGb3VuZEVycm9yID0gSW1wb3J0RXJyb3IKCnRyeToK
            ICAgIGltcG9ydCByYWRvcwpleGNlcHQgTW9kdWxlTm90Rm91bmRFcnJvciBhcyBub01vZEVycjoK
            ICAgIHByaW50KGYiRXJyb3I6IHtub01vZEVycn1cbkV4aXRpbmcgdGhlIHNjcmlwdC4uLiIpCiAg
            ICBzeXMuZXhpdCgxKQoKdHJ5OgogICAgaW1wb3J0IHJiZApleGNlcHQgTW9kdWxlTm90Rm91bmRF
            cnJvciBhcyBub01vZEVycjoKICAgIHByaW50KGYiRXJyb3I6IHtub01vZEVycn1cbkV4aXRpbmcg
            dGhlIHNjcmlwdC4uLiIpCiAgICBzeXMuZXhpdCgxKQoKdHJ5OgogICAgIyBmb3IgMi43LngKICAg
            IGZyb20gU3RyaW5nSU8gaW1wb3J0IFN0cmluZ0lPCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9y
            OgogICAgIyBmb3IgMy54CiAgICBmcm9tIGlvIGltcG9ydCBTdHJpbmdJTwoKdHJ5OgogICAgIyBm
            b3IgMi43LngKICAgIGZyb20gdXJscGFyc2UgaW1wb3J0IHVybHBhcnNlCiAgICBmcm9tIHVybGxp
            YiBpbXBvcnQgdXJsZW5jb2RlIGFzIHVybGVuY29kZQpleGNlcHQgTW9kdWxlTm90Rm91bmRFcnJv
            cjoKICAgICMgZm9yIDMueAogICAgZnJvbSB1cmxsaWIucGFyc2UgaW1wb3J0IHVybHBhcnNlCiAg
            ICBmcm9tIHVybGxpYi5wYXJzZSBpbXBvcnQgdXJsZW5jb2RlIGFzIHVybGVuY29kZQoKdHJ5Ogog
            ICAgZnJvbSBiYXNlNjQgaW1wb3J0IGVuY29kZXN0cmluZwpleGNlcHQ6CiAgICBmcm9tIGJhc2U2
            NCBpbXBvcnQgZW5jb2RlYnl0ZXMgYXMgZW5jb2Rlc3RyaW5nCgoKY2xhc3MgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbihFeGNlcHRpb24pOgogICAgcGFzcwoKCiMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMgRHVtbXlSYWRv
            cyAjIyMjIyMjIyMjIyMjIyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjCiMgdGhpcyBpcyBtYWlubHkgZm9yIHRlc3RpbmcgYW5kIGNvdWxkIGJlIHVz
            ZWQgd2hlcmUgJ3JhZG9zJyBpcyBub3QgYXZhaWxhYmxlCgoKY2xhc3MgRHVtbXlSYWRvcyhvYmpl
            Y3QpOgogICAgZGVmIF9faW5pdF9fKHNlbGYpOgogICAgICAgIHNlbGYucmV0dXJuX3ZhbCA9IDAK
            ICAgICAgICBzZWxmLmVycl9tZXNzYWdlID0gIiIKICAgICAgICBzZWxmLnN0YXRlID0gImNvbm5l
            Y3RlZCIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwID0ge30KICAgICAgICBzZWxmLmNtZF9u
            YW1lcyA9IHt9CiAgICAgICAgc2VsZi5faW5pdF9jbWRfb3V0cHV0X21hcCgpCiAgICAgICAgc2Vs
            Zi5kdW1teV9ob3N0X2lwX21hcCA9IHt9CgogICAgZGVmIF9pbml0X2NtZF9vdXRwdXRfbWFwKHNl
            bGYpOgogICAgICAgIGpzb25fZmlsZV9uYW1lID0gInRlc3QtZGF0YS9jZXBoLXN0YXR1cy1vdXQi
            CiAgICAgICAgc2NyaXB0X2RpciA9IHBhdGguYWJzcGF0aChwYXRoLmRpcm5hbWUoX19maWxlX18p
            KQogICAgICAgIGNlcGhfc3RhdHVzX3N0ciA9ICIiCiAgICAgICAgd2l0aCBvcGVuKAogICAgICAg
            ICAgICBwYXRoLmpvaW4oc2NyaXB0X2RpciwganNvbl9maWxlX25hbWUpLCBtb2RlPSJyIiwgZW5j
            b2Rpbmc9IlVURi04IgogICAgICAgICkgYXMganNvbl9maWxlOgogICAgICAgICAgICBjZXBoX3N0
            YXR1c19zdHIgPSBqc29uX2ZpbGUucmVhZCgpCiAgICAgICAgc2VsZi5jbWRfbmFtZXNbImZzIGxz
            Il0gPSAiIiJ7ImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJmcyBscyJ9IiIiCiAgICAgICAg
            c2VsZi5jbWRfbmFtZXNbInF1b3J1bV9zdGF0dXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3Jt
            YXQiOiAianNvbiIsICJwcmVmaXgiOiAicXVvcnVtX3N0YXR1cyJ9IiIiCiAgICAgICAgKQogICAg
            ICAgIHNlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXSA9ICgKICAgICAgICAgICAgIiIieyJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAibWdyIHNlcnZpY2VzIn0iIiIKICAgICAgICApCiAg
            ICAgICAgIyBhbGwgdGhlIGNvbW1hbmRzIGFuZCB0aGVpciBvdXRwdXQKICAgICAgICBzZWxmLmNt
            ZF9vdXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJmcyBscyJdXSA9ICgKICAgICAgICAgICAgIiIi
            W3sibmFtZSI6Im15ZnMiLCJtZXRhZGF0YV9wb29sIjoibXlmcy1tZXRhZGF0YSIsIm1ldGFkYXRh
            X3Bvb2xfaWQiOjIsImRhdGFfcG9vbF9pZHMiOlszXSwiZGF0YV9wb29scyI6WyJteWZzLXJlcGxp
            Y2F0ZWQiXX1dIiIiCiAgICAgICAgKQogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbc2VsZi5j
            bWRfbmFtZXNbInF1b3J1bV9zdGF0dXMiXV0gPSAoCiAgICAgICAgICAgICIiInsiZWxlY3Rpb25f
            ZXBvY2giOjMsInF1b3J1bSI6WzBdLCJxdW9ydW1fbmFtZXMiOlsiYSJdLCJxdW9ydW1fbGVhZGVy
            X25hbWUiOiJhIiwicXVvcnVtX2FnZSI6MTQzODUsImZlYXR1cmVzIjp7InF1b3J1bV9jb24iOiI0
            NTQwMTM4MjkyODM2Njk2MDYzIiwicXVvcnVtX21vbiI6WyJrcmFrZW4iLCJsdW1pbm91cyIsIm1p
            bWljIiwib3NkbWFwLXBydW5lIiwibmF1dGlsdXMiLCJvY3RvcHVzIl19LCJtb25tYXAiOnsiZXBv
            Y2giOjEsImZzaWQiOiJhZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5ZDBmMWMiLCJtb2Rp
            ZmllZCI6IjIwMjAtMDUtMDdUMDM6MzY6MzkuOTE4MDM1WiIsImNyZWF0ZWQiOiIyMDIwLTA1LTA3
            VDAzOjM2OjM5LjkxODAzNVoiLCJtaW5fbW9uX3JlbGVhc2UiOjE1LCJtaW5fbW9uX3JlbGVhc2Vf
            bmFtZSI6Im9jdG9wdXMiLCJmZWF0dXJlcyI6eyJwZXJzaXN0ZW50IjpbImtyYWtlbiIsImx1bWlu
            b3VzIiwibWltaWMiLCJvc2RtYXAtcHJ1bmUiLCJuYXV0aWx1cyIsIm9jdG9wdXMiXSwib3B0aW9u
            YWwiOltdfSwibW9ucyI6W3sicmFuayI6MCwibmFtZSI6ImEiLCJwdWJsaWNfYWRkcnMiOnsiYWRk
            cnZlYyI6W3sidHlwZSI6InYyIiwiYWRkciI6IjEwLjExMC4yMDUuMTc0OjMzMDAiLCJub25jZSI6
            MH0seyJ0eXBlIjoidjEiLCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OSIsIm5vbmNlIjowfV19
            LCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OS8wIiwicHVibGljX2FkZHIiOiIxMC4xMTAuMjA1
            LjE3NDo2Nzg5LzAiLCJwcmlvcml0eSI6MCwid2VpZ2h0IjowfV19fSIiIgogICAgICAgICkKICAg
            ICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXV0g
            PSAoCiAgICAgICAgICAgICIiInsiZGFzaGJvYXJkIjoiaHR0cHM6Ly9jZXBoLWRhc2hib2FyZDo4
            NDQzLyIsInByb21ldGhldXMiOiJodHRwOi8vY2VwaC1kYXNoYm9hcmQtZGI6OTI4My8ifSIiIgog
            ICAgICAgICkKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNh
            cHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMiLCAib3Nk
            IiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0
            YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250
            cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJdLCAiZW50aXR5Ijog
            ImNsaWVudC5oZWFsdGhjaGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRo
            IGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuaGVh
            bHRoY2hlY2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRB
            PT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cyIs
            Im9zZCI6InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5j
            b250cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJ9fV0iIiIKICAg
            ICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIs
            ICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAib3NkIiwgInBy
            b2ZpbGUgcmJkIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtbm9kZSIsICJmb3JtYXQiOiAi
            anNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIi
            W3siZW50aXR5IjoiY2xpZW50LmNzaS1yYmQtbm9kZSIsImtleSI6IkFRQk9nck5lSGJLMUF4QUF1
            YllCZVY4UzFVL0dQenE1U1ZlcTZnPT0iLCJjYXBzIjp7Im1vbiI6InByb2ZpbGUgcmJkLCBhbGxv
            dyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsIm9zZCI6InByb2ZpbGUgcmJkIn19XSIiIgogICAg
            ICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsibW9uIiwg
            InByb2ZpbGUgcmJkLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxs
            b3cgcnciLCAib3NkIiwgInByb2ZpbGUgcmJkIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQt
            cHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNy
            ZWF0ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktcmJkLXByb3Zp
            c2lvbmVyIiwia2V5IjoiQVFCTmdyTmUxZ2V5S3hBQThla1ZpUmRFK2hzczVPd2VZQmt3Tmc9PSIs
            ImNhcHMiOnsibWdyIjoiYWxsb3cgcnciLCJtb24iOiJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFu
            ZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJwcm9maWxlIHJiZCJ9fV0iIiIKICAgICAgICBzZWxm
            LmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyBy
            LCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3Nk
            IiwgImFsbG93IHJ3IHRhZyBjZXBoZnMgKj0qIiwgIm1kcyIsICJhbGxvdyBydyJdLCAiZW50aXR5
            IjogImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4Ijog
            ImF1dGggZ2V0LW9yLWNyZWF0ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVu
            dC5jc2ktY2VwaGZzLW5vZGUiLCJrZXkiOiJBUUJPZ3JOZUVOdW5LeEFBUENtZ0U3UjZHOERjWG5h
            SjFGMzJxZz09IiwiY2FwcyI6eyJtZHMiOiJhbGxvdyBydyIsIm1nciI6ImFsbG93IHJ3IiwibW9u
            IjoiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBy
            dyB0YWcgY2VwaGZzICo9KiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAg
            ICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kICdvc2Qg
            YmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3NkIiwgImFsbG93IHJ3IHRhZyBjZXBo
            ZnMgbWV0YWRhdGE9KiJdLCAiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVy
            IiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgog
            ICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIs
            ImtleSI6IkFRQk9nck5lQUZnY0dCQUF2R3FLT0FEMEQzeHhtVlkwUjkxMmRnPT0iLCJjYXBzIjp7
            Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2Nr
            bGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoifX1dIiIiCiAgICAg
            ICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAi
            YWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdyIiwgImFsbG93IHJ3
            IiwgIm9zZCIsICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiXSwgImVudGl0eSI6ICJj
            bGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVuc2hpZnQtc3RvcmFnZSIsICJmb3JtYXQi
            OiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0g
            IiIiW3siZW50aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0
            b3JhZ2UiLCJrZXkiOiJCUUJPZ3JOZUFGZ2NHQkFBdkdxS09BRDBEM3h4bVZZMFI5MTJkZz09Iiwi
            Y2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29z
            ZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIn19XSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsi
            bW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciIsICJh
            bGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT1teWZzIl0sICJl
            bnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0b3JhZ2Ut
            bXlmcyIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0i
            IiIKICAgICAgICBdID0gIiIiW3siZW50aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9u
            ZXItb3BlbnNoaWZ0LXN0b3JhZ2UtbXlmcyIsImtleSI6IkNRQk9nck5lQUZnY0dCQUF2R3FLT0FE
            MEQzeHhtVlkwUjkxMmRnPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cg
            ciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2Vw
            aGZzIG1ldGFkYXRhPW15ZnMifX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAg
            ICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCBxdW9y
            dW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIiwgImFsbG93IGNvbW1hbmQg
            Y29uZmlnIiwgIm9zZCIsICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFsbG93IHJ3eCBwb29sPWRl
            ZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9ZGVm
            YXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJndy5sb2csIGFsbG93IHgg
            cG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4Il0sICJlbnRpdHkiOiAiY2xpZW50LmhlYWx0
            aGNoZWNrZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0
            ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5oZWFsdGhjaGVja2VyIiwi
            a2V5IjoiQVFERmtiTmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsi
            bW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFu
            ZCB2ZXJzaW9uIiwgIm1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsICJvc2QiOiAicHJvZmls
            ZSByYmQtcmVhZC1vbmx5LCBhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhLCBhbGxvdyBy
            IHBvb2w9LnJndy5yb290LCBhbGxvdyBydyBwb29sPWRlZmF1bHQucmd3LmNvbnRyb2wsIGFsbG93
            IHJ4IHBvb2w9ZGVmYXVsdC5yZ3cubG9nLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0
            cy5pbmRleCJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAi
            IiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMs
            IGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cgY29tbWFuZCBjb25maWciLCAi
            b3NkIiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5j
            b250cm9sLCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1
            bHQucmd3LmJ1Y2tldHMuaW5kZXgiXSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIs
            ICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBjYXBzIn0iIiIKICAgICAgICBdID0g
            IiIiW3siZW50aXR5IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0NWJG
            UkFBVG5kTE5VU1JLcnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIjog
            ImFsbG93IGNvbW1hbmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFs
            bG93IHJ3eCBwb29sPWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFs
            bG93IHJ3IHBvb2w9ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJn
            dy5sb2csIGFsbG93IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4In19XSIiIgogICAg
            ICAgIHNlbGYuY21kX291dHB1dF9tYXBbIiIieyJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            bWdyIHNlcnZpY2VzIn0iIiJdID0gKAogICAgICAgICAgICAiIiJ7ImRhc2hib2FyZCI6ICJodHRw
            Oi8vcm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6NzAwMC8iLCAicHJvbWV0aGV1cyI6
            ICJodHRwOi8vcm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6OTI4My8ifSIiIgogICAg
            ICAgICkKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0
            eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            YXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJ7ImRhc2hib2FyZCI6ICJodHRwOi8vcm9vay1j
            ZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6NzAwMC8iLCAicHJvbWV0aGV1cyI6ICJodHRwOi8v
            cm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6OTI4My8ifSIiIgogICAgICAgIHNlbGYu
            Y21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNsaWVudC5oZWFsdGhj
            aGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldCJ9IiIiCiAgICAg
            ICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5oZWFsdGhjaGVja2VyIiwia2V5IjoiQVFERmti
            TmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjogImFsbG93
            IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFuZCB2ZXJzaW9uIiwg
            Im1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsICJvc2QiOiAicHJvZmlsZSByYmQtcmVhZC1v
            bmx5LCBhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhLCBhbGxvdyByIHBvb2w9LnJndy5y
            b290LCBhbGxvdyBydyBwb29sPWRlZmF1bHQucmd3LmNvbnRyb2wsIGFsbG93IHJ4IHBvb2w9ZGVm
            YXVsdC5yZ3cubG9nLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJ9fV0i
            IiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0eSI6
            ICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJh
            dXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0
            X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtbm9kZSIsICJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJb
            XSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5
            IjogImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRf
            b3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMt
            cHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIK
            ICAgICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAg
            ICAgICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVuc2hp
            ZnQtc3RvcmFnZSIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgog
            ICAgICAgIF0gPSAiIiJbXSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAg
            ICAgICIiInsiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlm
            dC1zdG9yYWdlLW15ZnMiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0i
            IiIKICAgICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAg
            ICAgICAgICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJb
            eyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsImtleSI6IkFRREZrYk5l
            ZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRBPT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIi
            LCAibWdyIjoiYWxsb3cgcnciLCAib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0q
            In19XSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2Fw
            cyI6IFsibW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1n
            ciIsICJhbGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIl0s
            ICJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpz
            b24iLCAicHJlZml4IjogImF1dGggY2FwcyJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmtiTmVmdDViRlJBQVRu
            ZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cgciwgIGFsbG93IGNv
            bW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciI6ImFsbG93IHJ3IiwgIm9zZCI6ImFsbG93IHJ3
            IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFw
            Wyd7ImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJzdGF0dXMifSddID0gY2VwaF9zdGF0dXNf
            c3RyCgogICAgZGVmIHNodXRkb3duKHNlbGYpOgogICAgICAgIHBhc3MKCiAgICBkZWYgZ2V0X2Zz
            aWQoc2VsZik6CiAgICAgICAgcmV0dXJuICJhZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5
            ZDBmMWMiCgogICAgZGVmIGNvbmZfcmVhZF9maWxlKHNlbGYpOgogICAgICAgIHBhc3MKCiAgICBk
            ZWYgY29ubmVjdChzZWxmKToKICAgICAgICBwYXNzCgogICAgZGVmIHBvb2xfZXhpc3RzKHNlbGYs
            IHBvb2xfbmFtZSk6CiAgICAgICAgcmV0dXJuIFRydWUKCiAgICBkZWYgbW9uX2NvbW1hbmQoc2Vs
            ZiwgY21kLCBvdXQpOgogICAgICAgIGpzb25fY21kID0ganNvbi5sb2FkcyhjbWQpCiAgICAgICAg
            anNvbl9jbWRfc3RyID0ganNvbi5kdW1wcyhqc29uX2NtZCwgc29ydF9rZXlzPVRydWUpCiAgICAg
            ICAgY21kX291dHB1dCA9IHNlbGYuY21kX291dHB1dF9tYXBbanNvbl9jbWRfc3RyXQogICAgICAg
            IHJldHVybiBzZWxmLnJldHVybl92YWwsIGNtZF9vdXRwdXQsIHN0cihzZWxmLmVycl9tZXNzYWdl
            LmVuY29kZSgidXRmLTgiKSkKCiAgICBkZWYgX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoc2VsZiwg
            aG9zdF9uYW1lKToKICAgICAgICBpcF9yZWdfeCA9IHJlLmNvbXBpbGUociJcZHsxLDN9LlxkezEs
            M30uXGR7MSwzfS5cZHsxLDN9IikKICAgICAgICAjIGlmIHByb3ZpZGVkIGhvc3QgaXMgZGlyZWN0
            bHkgYW4gSVAgYWRkcmVzcywgcmV0dXJuIHRoZSBzYW1lCiAgICAgICAgaWYgaXBfcmVnX3gubWF0
            Y2goaG9zdF9uYW1lKToKICAgICAgICAgICAgcmV0dXJuIGhvc3RfbmFtZQogICAgICAgIGltcG9y
            dCByYW5kb20KCiAgICAgICAgaG9zdF9pcCA9IHNlbGYuZHVtbXlfaG9zdF9pcF9tYXAuZ2V0KGhv
            c3RfbmFtZSwgIiIpCiAgICAgICAgaWYgbm90IGhvc3RfaXA6CiAgICAgICAgICAgIGhvc3RfaXAg
            PSBmIjE3Mi45LntyYW5kb20ucmFuZGludCgwLCAyNTQpfS57cmFuZG9tLnJhbmRpbnQoMCwgMjU0
            KX0iCiAgICAgICAgICAgIHNlbGYuZHVtbXlfaG9zdF9pcF9tYXBbaG9zdF9uYW1lXSA9IGhvc3Rf
            aXAKICAgICAgICBkZWwgcmFuZG9tCiAgICAgICAgcmV0dXJuIGhvc3RfaXAKCiAgICBAY2xhc3Nt
            ZXRob2QKICAgIGRlZiBSYWRvcyhjb25mZmlsZT1Ob25lKToKICAgICAgICByZXR1cm4gRHVtbXlS
            YWRvcygpCgoKY2xhc3MgUzNBdXRoKEF1dGhCYXNlKToKICAgICIiIkF0dGFjaGVzIEFXUyBBdXRo
            ZW50aWNhdGlvbiB0byB0aGUgZ2l2ZW4gUmVxdWVzdCBvYmplY3QuIiIiCgogICAgc2VydmljZV9i
            YXNlX3VybCA9ICJzMy5hbWF6b25hd3MuY29tIgoKICAgIGRlZiBfX2luaXRfXyhzZWxmLCBhY2Nl
            c3Nfa2V5LCBzZWNyZXRfa2V5LCBzZXJ2aWNlX3VybD1Ob25lKToKICAgICAgICBpZiBzZXJ2aWNl
            X3VybDoKICAgICAgICAgICAgc2VsZi5zZXJ2aWNlX2Jhc2VfdXJsID0gc2VydmljZV91cmwKICAg
            ICAgICBzZWxmLmFjY2Vzc19rZXkgPSBzdHIoYWNjZXNzX2tleSkKICAgICAgICBzZWxmLnNlY3Jl
            dF9rZXkgPSBzdHIoc2VjcmV0X2tleSkKCiAgICBkZWYgX19jYWxsX18oc2VsZiwgcik6CiAgICAg
            ICAgIyBDcmVhdGUgZGF0ZSBoZWFkZXIgaWYgaXQgaXMgbm90IGNyZWF0ZWQgeWV0LgogICAgICAg
            IGlmICJkYXRlIiBub3QgaW4gci5oZWFkZXJzIGFuZCAieC1hbXotZGF0ZSIgbm90IGluIHIuaGVh
            ZGVyczoKICAgICAgICAgICAgci5oZWFkZXJzWyJkYXRlIl0gPSBmb3JtYXRkYXRlKHRpbWV2YWw9
            Tm9uZSwgbG9jYWx0aW1lPUZhbHNlLCB1c2VnbXQ9VHJ1ZSkKICAgICAgICBzaWduYXR1cmUgPSBz
            ZWxmLmdldF9zaWduYXR1cmUocikKICAgICAgICBpZiBweTNrOgogICAgICAgICAgICBzaWduYXR1
            cmUgPSBzaWduYXR1cmUuZGVjb2RlKCJ1dGYtOCIpCiAgICAgICAgci5oZWFkZXJzWyJBdXRob3Jp
            emF0aW9uIl0gPSBmIkFXUyB7c2VsZi5hY2Nlc3Nfa2V5fTp7c2lnbmF0dXJlfSIKICAgICAgICBy
            ZXR1cm4gcgoKICAgIGRlZiBnZXRfc2lnbmF0dXJlKHNlbGYsIHIpOgogICAgICAgIGNhbm9uaWNh
            bF9zdHJpbmcgPSBzZWxmLmdldF9jYW5vbmljYWxfc3RyaW5nKHIudXJsLCByLmhlYWRlcnMsIHIu
            bWV0aG9kKQogICAgICAgIGlmIHB5M2s6CiAgICAgICAgICAgIGtleSA9IHNlbGYuc2VjcmV0X2tl
            eS5lbmNvZGUoInV0Zi04IikKICAgICAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZy5lbmNv
            ZGUoInV0Zi04IikKICAgICAgICBlbHNlOgogICAgICAgICAgICBrZXkgPSBzZWxmLnNlY3JldF9r
            ZXkKICAgICAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZwogICAgICAgIGggPSBobWFjLm5l
            dyhrZXksIG1zZywgZGlnZXN0bW9kPXNoYSkKICAgICAgICByZXR1cm4gZW5jb2Rlc3RyaW5nKGgu
            ZGlnZXN0KCkpLnN0cmlwKCkKCiAgICBkZWYgZ2V0X2Nhbm9uaWNhbF9zdHJpbmcoc2VsZiwgdXJs
            LCBoZWFkZXJzLCBtZXRob2QpOgogICAgICAgIHBhcnNlZHVybCA9IHVybHBhcnNlKHVybCkKICAg
            ICAgICBvYmplY3RrZXkgPSBwYXJzZWR1cmwucGF0aFsxOl0KCiAgICAgICAgYnVja2V0ID0gcGFy
            c2VkdXJsLm5ldGxvY1s6IC1sZW4oc2VsZi5zZXJ2aWNlX2Jhc2VfdXJsKV0KICAgICAgICBpZiBs
            ZW4oYnVja2V0KSA+IDE6CiAgICAgICAgICAgICMgcmVtb3ZlIGxhc3QgZG90CiAgICAgICAgICAg
            IGJ1Y2tldCA9IGJ1Y2tldFs6LTFdCgogICAgICAgIGludGVyZXN0aW5nX2hlYWRlcnMgPSB7ImNv
            bnRlbnQtbWQ1IjogIiIsICJjb250ZW50LXR5cGUiOiAiIiwgImRhdGUiOiAiIn0KICAgICAgICBm
            b3Iga2V5IGluIGhlYWRlcnM6CiAgICAgICAgICAgIGxrID0ga2V5Lmxvd2VyKCkKICAgICAgICAg
            ICAgdHJ5OgogICAgICAgICAgICAgICAgbGsgPSBsay5kZWNvZGUoInV0Zi04IikKICAgICAgICAg
            ICAgZXhjZXB0OgogICAgICAgICAgICAgICAgcGFzcwogICAgICAgICAgICBpZiBoZWFkZXJzW2tl
            eV0gYW5kICgKICAgICAgICAgICAgICAgIGxrIGluIGludGVyZXN0aW5nX2hlYWRlcnMua2V5cygp
            IG9yIGxrLnN0YXJ0c3dpdGgoIngtYW16LSIpCiAgICAgICAgICAgICk6CiAgICAgICAgICAgICAg
            ICBpbnRlcmVzdGluZ19oZWFkZXJzW2xrXSA9IGhlYWRlcnNba2V5XS5zdHJpcCgpCgogICAgICAg
            ICMgSWYgeC1hbXotZGF0ZSBpcyB1c2VkIGl0IHN1cGVyc2VkZXMgdGhlIGRhdGUgaGVhZGVyLgog
            ICAgICAgIGlmIG5vdCBweTNrOgogICAgICAgICAgICBpZiAieC1hbXotZGF0ZSIgaW4gaW50ZXJl
            c3RpbmdfaGVhZGVyczoKICAgICAgICAgICAgICAgIGludGVyZXN0aW5nX2hlYWRlcnNbImRhdGUi
            XSA9ICIiCiAgICAgICAgZWxzZToKICAgICAgICAgICAgaWYgIngtYW16LWRhdGUiIGluIGludGVy
            ZXN0aW5nX2hlYWRlcnM6CiAgICAgICAgICAgICAgICBpbnRlcmVzdGluZ19oZWFkZXJzWyJkYXRl
            Il0gPSAiIgoKICAgICAgICBidWYgPSBmInttZXRob2R9XG4iCiAgICAgICAgZm9yIGtleSBpbiBz
            b3J0ZWQoaW50ZXJlc3RpbmdfaGVhZGVycy5rZXlzKCkpOgogICAgICAgICAgICB2YWwgPSBpbnRl
            cmVzdGluZ19oZWFkZXJzW2tleV0KICAgICAgICAgICAgaWYga2V5LnN0YXJ0c3dpdGgoIngtYW16
            LSIpOgogICAgICAgICAgICAgICAgYnVmICs9IGYie2tleX06e3ZhbH1cbiIKICAgICAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgICAgIGJ1ZiArPSBmInt2YWx9XG4iCgogICAgICAgICMgYXBwZW5k
            IHRoZSBidWNrZXQgaWYgaXQgZXhpc3RzCiAgICAgICAgaWYgYnVja2V0ICE9ICIiOgogICAgICAg
            ICAgICBidWYgKz0gZiIve2J1Y2tldH0iCgogICAgICAgICMgYWRkIHRoZSBvYmplY3RrZXkuIGV2
            ZW4gaWYgaXQgZG9lc24ndCBleGlzdCwgYWRkIHRoZSBzbGFzaAogICAgICAgIGJ1ZiArPSBmIi97
            b2JqZWN0a2V5fSIKCiAgICAgICAgcmV0dXJuIGJ1ZgoKCmNsYXNzIFJhZG9zSlNPTjoKICAgIEVY
            VEVSTkFMX1VTRVJfTkFNRSA9ICJjbGllbnQuaGVhbHRoY2hlY2tlciIKICAgIEVYVEVSTkFMX1JH
            V19BRE1JTl9PUFNfVVNFUl9OQU1FID0gInJndy1hZG1pbi1vcHMtdXNlciIKICAgIEVNUFRZX09V
            VFBVVF9MSVNUID0gIkVtcHR5IG91dHB1dCBsaXN0IgogICAgREVGQVVMVF9SR1dfUE9PTF9QUkVG
            SVggPSAiZGVmYXVsdCIKICAgIERFRkFVTFRfTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUID0gIjky
            ODMiCgogICAgQGNsYXNzbWV0aG9kCiAgICBkZWYgZ2VuX2FyZ19wYXJzZXIoY2xzLCBhcmdzX3Rv
            X3BhcnNlPU5vbmUpOgogICAgICAgIGFyZ1AgPSBhcmdwYXJzZS5Bcmd1bWVudFBhcnNlcigpCgog
            ICAgICAgIGNvbW1vbl9ncm91cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJjb21tb24iKQog
            ICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoIi0tdmVyYm9zZSIsICItdiIsIGFjdGlv
            bj0ic3RvcmVfdHJ1ZSIsIGRlZmF1bHQ9RmFsc2UpCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9h
            cmd1bWVudCgKICAgICAgICAgICAgIi0tY2VwaC1jb25mIiwgIi1jIiwgaGVscD0iUHJvdmlkZSBh
            IGNlcGggY29uZiBmaWxlLiIsIHR5cGU9c3RyCiAgICAgICAgKQogICAgICAgIGNvbW1vbl9ncm91
            cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLWtleXJpbmciLCAiLWsiLCBoZWxwPSJQYXRo
            IHRvIGNlcGgga2V5cmluZyBmaWxlLiIsIHR5cGU9c3RyCiAgICAgICAgKQogICAgICAgIGNvbW1v
            bl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJ1bi1hcy11c2VyIiwKICAgICAg
            ICAgICAgIi11IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgdHlwZT1zdHIs
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIGEgdXNlciBuYW1lIHRvIGNoZWNrIHRoZSBjbHVz
            dGVyJ3MgaGVhbHRoIHN0YXR1cywgbXVzdCBiZSBwcmVmaXhlZCBieSAnY2xpZW50LiciLAogICAg
            ICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1j
            bHVzdGVyLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJL
            dWJlcm5ldGVzIGNsdXN0ZXIgbmFtZShsZWdhY3kgZmxhZyksIE5vdGU6IEVpdGhlciB1c2UgdGhp
            cyBvciAtLWs4cy1jbHVzdGVyLW5hbWUiLAogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAu
            YWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1rOHMtY2x1c3Rlci1uYW1lIiwgZGVmYXVsdD0i
            IiwgaGVscD0iS3ViZXJuZXRlcyBjbHVzdGVyIG5hbWUiCiAgICAgICAgKQogICAgICAgIGNvbW1v
            bl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLW5hbWVzcGFjZSIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIGhlbHA9Ik5hbWVzcGFjZSB3aGVyZSBDZXBoQ2x1
            c3RlciBpcyBydW5uaW5nIiwKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tcmd3LXBvb2wtcHJlZml4IiwgZGVmYXVsdD0iIiwgaGVscD0i
            UkdXIFBvb2wgcHJlZml4IgogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3Vt
            ZW50KAogICAgICAgICAgICAiLS1yZXN0cmljdGVkLWF1dGgtcGVybWlzc2lvbiIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlJlc3RyaWN0IGNlcGhDU0lLZXly
            aW5ncyBhdXRoIHBlcm1pc3Npb25zIHRvIHNwZWNpZmljIHBvb2xzLCBjbHVzdGVyLiIKICAgICAg
            ICAgICAgKyAiTWFuZGF0b3J5IGZsYWdzIHRoYXQgbmVlZCB0byBiZSBzZXQgYXJlIC0tcmJkLWRh
            dGEtcG9vbC1uYW1lLCBhbmQgLS1rOHMtY2x1c3Rlci1uYW1lLiIKICAgICAgICAgICAgKyAiLS1j
            ZXBoZnMtZmlsZXN5c3RlbS1uYW1lIGZsYWcgY2FuIGFsc28gYmUgcGFzc2VkIGluIGNhc2Ugb2Yg
            Y2VwaGZzIHVzZXIgcmVzdHJpY3Rpb24sIHNvIGl0IGNhbiByZXN0cmljdCB1c2VyIHRvIHBhcnRp
            Y3VsYXIgY2VwaGZzIGZpbGVzeXN0ZW0iCiAgICAgICAgICAgICsgInNhbXBsZSBydW46IGBweXRo
            b24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwtY2x1c3Rlci1yZXNvdXJjZXMucHkgLS1jZXBo
            ZnMtZmlsZXN5c3RlbS1uYW1lIG15ZnMgLS1yYmQtZGF0YS1wb29sLW5hbWUgcmVwbGljYXBvb2wg
            LS1rOHMtY2x1c3Rlci1uYW1lIHJvb2tzdG9yYWdlIC0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Np
            b24gdHJ1ZWAiCiAgICAgICAgICAgICsgIk5vdGU6IFJlc3RyaWN0aW5nIHRoZSBjc2ktdXNlcnMg
            cGVyIHBvb2wsIGFuZCBwZXIgY2x1c3RlciB3aWxsIHJlcXVpcmUgY3JlYXRpbmcgbmV3IGNzaS11
            c2VycyBhbmQgbmV3IHNlY3JldHMgZm9yIHRoYXQgY3NpLXVzZXJzLiIKICAgICAgICAgICAgKyAi
            U28gYXBwbHkgdGhlc2Ugc2VjcmV0cyBvbmx5IHRvIG5ldyBgQ29uc3VtZXIgY2x1c3RlcmAgZGVw
            bG95bWVudCB3aGlsZSB1c2luZyB0aGUgc2FtZSBgU291cmNlIGNsdXN0ZXJgLiIsCiAgICAgICAg
            KQogICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXYyLXBv
            cnQtZW5hYmxlIiwKICAgICAgICAgICAgYWN0aW9uPSJzdG9yZV90cnVlIiwKICAgICAgICAgICAg
            ZGVmYXVsdD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iRW5hYmxlIHYyIG1vbiBwb3J0KDMzMDAp
            IGZvciBtb25zIiwKICAgICAgICApCgogICAgICAgIG91dHB1dF9ncm91cCA9IGFyZ1AuYWRkX2Fy
            Z3VtZW50X2dyb3VwKCJvdXRwdXQiKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQo
            CiAgICAgICAgICAgICItLWZvcm1hdCIsCiAgICAgICAgICAgICItdCIsCiAgICAgICAgICAgIGNo
            b2ljZXM9WyJqc29uIiwgImJhc2giXSwKICAgICAgICAgICAgZGVmYXVsdD0ianNvbiIsCiAgICAg
            ICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBvdXRwdXQgZm9ybWF0IChqc29uIHwgYmFzaCkiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1vdXRwdXQiLAogICAgICAgICAgICAiLW8iLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAg
            ICAgICAgICBoZWxwPSJPdXRwdXQgd2lsbCBiZSBzdG9yZWQgaW50byB0aGUgcHJvdmlkZWQgZmls
            ZSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAog
            ICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgQ2VwaCBmaWxlc3lzdGVt
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tY2VwaGZzLW1ldGFkYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIs
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBjZXBoZnMgbWV0YWRh
            dGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAg
            ICAgICAgICAgICItLWNlcGhmcy1kYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9
            IiIsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBjZXBoZnMgZGF0
            YSBwb29sIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAg
            ICAgICAgICAgIi0tcmJkLWRhdGEtcG9vbC1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwK
            ICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRo
            ZSBuYW1lIG9mIHRoZSBSQkQgZGF0YXBvb2wiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3Jv
            dXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1hbGlhcy1yYmQtZGF0YS1wb29sLW5hbWUi
            LAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAg
            ICAgICAgICAgaGVscD0iUHJvdmlkZXMgYW4gYWxpYXMgZm9yIHRoZSAgUkJEIGRhdGEgcG9vbCBu
            YW1lLCBuZWNlc3NhcnkgaWYgYSBzcGVjaWFsIGNoYXJhY3RlciBpcyBwcmVzZW50IGluIHRoZSBw
            b29sIG5hbWUgc3VjaCBhcyBhIHBlcmlvZCBvciB1bmRlcnNjb3JlIiwKICAgICAgICApCiAgICAg
            ICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmd3LWVuZHBvaW50
            IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAg
            ICAgICAgICAgIGhlbHA9IlJBRE9TIEdhdGV3YXkgZW5kcG9pbnQgKGluIGA8SVB2ND46PFBPUlQ+
            YCBvciBgPFtJUHY2XT46PFBPUlQ+YCBvciBgPEZRRE4+OjxQT1JUPmAgZm9ybWF0KSIsCiAgICAg
            ICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJn
            dy10bHMtY2VydC1wYXRoIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVx
            dWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlJBRE9TIEdhdGV3YXkgZW5kcG9pbnQgVExT
            IGNlcnRpZmljYXRlIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVu
            dCgKICAgICAgICAgICAgIi0tcmd3LXNraXAtdGxzIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFs
            c2UsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9Iklnbm9yZSBU
            TFMgY2VydGlmaWNhdGlvbiB2YWxpZGF0aW9uIHdoZW4gYSBzZWxmLXNpZ25lZCBjZXJ0aWZpY2F0
            ZSBpcyBwcm92aWRlZCAoTk9UIFJFQ09NTUVOREVEIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tbW9uaXRvcmluZy1lbmRwb2ludCIs
            CiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAg
            ICAgICAgICBoZWxwPSJDZXBoIE1hbmFnZXIgcHJvbWV0aGV1cyBleHBvcnRlciBlbmRwb2ludHMg
            KGNvbW1hIHNlcGFyYXRlZCBsaXN0IG9mIChmb3JtYXQgYDxJUHY0PmAgb3IgYDxbSVB2Nl0+YCBv
            ciBgPEZRRE4+YCkgZW50cmllcyBvZiBhY3RpdmUgYW5kIHN0YW5kYnkgbWdycykiLAogICAgICAg
            ICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1tb25p
            dG9yaW5nLWVuZHBvaW50LXBvcnQiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAg
            ICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iQ2VwaCBNYW5hZ2VyIHByb21ldGhl
            dXMgZXhwb3J0ZXIgcG9ydCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJn
            dW1lbnQoCiAgICAgICAgICAgICItLXNraXAtbW9uaXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAg
            ICAgICAgIGhlbHA9IkRvIG5vdCBjaGVjayBmb3IgYSBtb25pdG9yaW5nIGVuZHBvaW50IGZvciB0
            aGUgQ2VwaCBjbHVzdGVyIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tcmJkLW1ldGFkYXRhLWVjLXBvb2wtbmFtZSIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBo
            ZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiBlcmFzdXJlIGNvZGVkIFJCRCBtZXRhZGF0YSBwb29s
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tZHJ5LXJ1biIsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFj
            dGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGhlbHA9IkRyeSBydW4gcHJpbnRzIHRoZSBl
            eGVjdXRlZCBjb21tYW5kcyB3aXRob3V0IHJ1bm5pbmcgdGhlbSIsCiAgICAgICAgKQogICAgICAg
            IG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJhZG9zLW5hbWVzcGFj
            ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAog
            ICAgICAgICAgICBoZWxwPSJEaXZpZGVzIGEgcG9vbCBpbnRvIHNlcGFyYXRlIGxvZ2ljYWwgbmFt
            ZXNwYWNlcywgdXNlZCBmb3IgY3JlYXRpbmcgUkJEIFBWQyBpbiBhIENlcGhCbG9ja1Bvb2xSYWRv
            c05hbWVzcGFjZSAoc2hvdWxkIGJlIGxvd2VyIGNhc2UpIiwKICAgICAgICApCiAgICAgICAgb3V0
            cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tc3Vidm9sdW1lLWdyb3VwIiwK
            ICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAg
            ICAgICAgIGhlbHA9InByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBzdWJ2b2x1bWUgZ3JvdXAiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1yZ3ctcmVhbG0tbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJl
            cXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUg
            cmd3LXJlYWxtIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgK
            ICAgICAgICAgICAgIi0tcmd3LXpvbmUtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAg
            ICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUg
            bmFtZSBvZiB0aGUgcmd3LXpvbmUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS1yZ3ctem9uZWdyb3VwLW5hbWUiLAogICAgICAgICAg
            ICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVs
            cD0icHJvdmlkZXMgdGhlIG5hbWUgb2YgdGhlIHJndy16b25lZ3JvdXAiLAogICAgICAgICkKCiAg
            ICAgICAgdXBncmFkZV9ncm91cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJ1cGdyYWRlIikK
            ICAgICAgICB1cGdyYWRlX2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tdXBncmFk
            ZSIsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9
            RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlVwZ3JhZGVzIHRoZSBjZXBoQ1NJS2V5cmluZ3MoRm9y
            IGV4YW1wbGU6IGNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyKSBhbmQgY2xpZW50LmhlYWx0
            aGNoZWNrZXIgY2VwaCB1c2VycyB3aXRoIG5ldyBwZXJtaXNzaW9ucyBuZWVkZWQgZm9yIHRoZSBu
            ZXcgY2x1c3RlciB2ZXJzaW9uIGFuZCBvbGRlciBwZXJtaXNzaW9uIHdpbGwgc3RpbGwgYmUgYXBw
            bGllZC4iCiAgICAgICAgICAgICsgIlNhbXBsZSBydW46IGBweXRob24zIC9ldGMvY2VwaC9jcmVh
            dGUtZXh0ZXJuYWwtY2x1c3Rlci1yZXNvdXJjZXMucHkgLS11cGdyYWRlYCwgdGhpcyB3aWxsIHVw
            Z3JhZGUgYWxsIHRoZSBkZWZhdWx0IGNzaSB1c2Vycyhub24tcmVzdHJpY3RlZCkiCiAgICAgICAg
            ICAgICsgIkZvciByZXN0cmljdGVkIHVzZXJzKEZvciBleGFtcGxlOiBjbGllbnQuY3NpLWNlcGhm
            cy1wcm92aXNpb25lci1vcGVuc2hpZnQtc3RvcmFnZS1teWZzKSwgdXNlcnMgY3JlYXRlZCB1c2lu
            ZyAtLXJlc3RyaWN0ZWQtYXV0aC1wZXJtaXNzaW9uIGZsYWcgbmVlZCB0byBwYXNzIG1hbmRhdG9y
            eSBmbGFncyIKICAgICAgICAgICAgKyAibWFuZGF0b3J5IGZsYWdzOiAnLS1yYmQtZGF0YS1wb29s
            LW5hbWUsIC0tazhzLWNsdXN0ZXItbmFtZSBhbmQgLS1ydW4tYXMtdXNlcicgZmxhZ3Mgd2hpbGUg
            dXBncmFkaW5nIgogICAgICAgICAgICArICJpbiBjYXNlIG9mIGNlcGhmcyB1c2VycyBpZiB5b3Ug
            aGF2ZSBwYXNzZWQgLS1jZXBoZnMtZmlsZXN5c3RlbS1uYW1lIGZsYWcgd2hpbGUgY3JlYXRpbmcg
            dXNlciB0aGVuIHdoaWxlIHVwZ3JhZGluZyBpdCB3aWxsIGJlIG1hbmRhdG9yeSB0b28iCiAgICAg
            ICAgICAgICsgIlNhbXBsZSBydW46IGBweXRob24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwt
            Y2x1c3Rlci1yZXNvdXJjZXMucHkgLS11cGdyYWRlIC0tcmJkLWRhdGEtcG9vbC1uYW1lIHJlcGxp
            Y2Fwb29sIC0tazhzLWNsdXN0ZXItbmFtZSByb29rc3RvcmFnZSAgLS1ydW4tYXMtdXNlciBjbGll
            bnQuY3NpLXJiZC1ub2RlLXJvb2tzdG9yYWdlLXJlcGxpY2Fwb29sYCIKICAgICAgICAgICAgKyAi
            UFM6IEFuIGV4aXN0aW5nIG5vbi1yZXN0cmljdGVkIHVzZXIgY2Fubm90IGJlIGNvbnZlcnRlZCB0
            byBhIHJlc3RyaWN0ZWQgdXNlciBieSB1cGdyYWRpbmcuIgogICAgICAgICAgICArICJVcGdyYWRl
            IGZsYWcgc2hvdWxkIG9ubHkgYmUgdXNlZCB0byBhcHBlbmQgbmV3IHBlcm1pc3Npb25zIHRvIHVz
            ZXJzLCBpdCBzaG91bGRuJ3QgYmUgdXNlZCBmb3IgY2hhbmdpbmcgdXNlciBhbHJlYWR5IGFwcGxp
            ZWQgcGVybWlzc2lvbiwgZm9yIGV4YW1wbGUgeW91IHNob3VsZG4ndCBjaGFuZ2UgaW4gd2hpY2gg
            cG9vbCB1c2VyIGhhcyBhY2Nlc3MiLAogICAgICAgICkKCiAgICAgICAgaWYgYXJnc190b19wYXJz
            ZToKICAgICAgICAgICAgYXNzZXJ0ICgKICAgICAgICAgICAgICAgIHR5cGUoYXJnc190b19wYXJz
            ZSkgPT0gbGlzdAogICAgICAgICAgICApLCAiQXJndW1lbnQgdG8gJ2dlbl9hcmdfcGFyc2VyJyBz
            aG91bGQgYmUgYSBsaXN0IgogICAgICAgIGVsc2U6CiAgICAgICAgICAgIGFyZ3NfdG9fcGFyc2Ug
            PSBzeXMuYXJndlsxOl0KICAgICAgICByZXR1cm4gYXJnUC5wYXJzZV9hcmdzKGFyZ3NfdG9fcGFy
            c2UpCgogICAgZGVmIHZhbGlkYXRlX3JiZF9tZXRhZGF0YV9lY19wb29sX25hbWUoc2VsZik6CiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lOgogICAg
            ICAgICAgICByYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5yYmRf
            bWV0YWRhdGFfZWNfcG9vbF9uYW1lCiAgICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9h
            cmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQoKICAgICAgICAgICAgaWYgcmJkX3Bvb2xfbmFt
            ZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24o
            CiAgICAgICAgICAgICAgICAgICAgIkZsYWcgJy0tcmJkLWRhdGEtcG9vbC1uYW1lJyBzaG91bGQg
            bm90IGJlIGVtcHR5IgogICAgICAgICAgICAgICAgKQoKICAgICAgICAgICAgaWYgcmJkX21ldGFk
            YXRhX2VjX3Bvb2xfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIkZsYWcgJy0tcmJkLW1ldGFkYXRh
            LWVjLXBvb2wtbmFtZScgc2hvdWxkIG5vdCBiZSBlbXB0eSIKICAgICAgICAgICAgICAgICkKCiAg
            ICAgICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAib3NkIGR1bXAiLCAiZm9ybWF0IjogImpz
            b24ifQogICAgICAgICAgICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1v
            bl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBs
            ZW4oanNvbl9vdXQpID09IDA6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJl
            RXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYie2NtZF9qc29uWydwcmVmaXgnXX0gY29t
            bWFuZCBmYWlsZWQuXG4iCiAgICAgICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYg
            cmV0X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAgICAgICAg
            ICApCiAgICAgICAgICAgIG1ldGFkYXRhX3Bvb2xfZXhpc3QsIHBvb2xfZXhpc3QgPSBGYWxzZSwg
            RmFsc2UKCiAgICAgICAgICAgIGZvciBrZXkgaW4ganNvbl9vdXRbInBvb2xzIl06CiAgICAgICAg
            ICAgICAgICAjIGlmIGVyYXN1cmVfY29kZV9wcm9maWxlIGlzIGVtcHR5IGFuZCBwb29sIG5hbWUg
            ZXhpc3RzIHRoZW4gaXQgcmVwbGljYSBwb29sCiAgICAgICAgICAgICAgICBpZiAoCiAgICAgICAg
            ICAgICAgICAgICAga2V5WyJlcmFzdXJlX2NvZGVfcHJvZmlsZSJdID09ICIiCiAgICAgICAgICAg
            ICAgICAgICAgYW5kIGtleVsicG9vbF9uYW1lIl0gPT0gcmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFt
            ZQogICAgICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgICAgICBtZXRhZGF0YV9wb29sX2V4
            aXN0ID0gVHJ1ZQogICAgICAgICAgICAgICAgIyBpZiBlcmFzdXJlX2NvZGVfcHJvZmlsZSBpcyBu
            b3QgZW1wdHkgYW5kIHBvb2wgbmFtZSBleGlzdHMgdGhlbiBpdCBpcyBlYyBwb29sCiAgICAgICAg
            ICAgICAgICBpZiBrZXlbImVyYXN1cmVfY29kZV9wcm9maWxlIl0gYW5kIGtleVsicG9vbF9uYW1l
            Il0gPT0gcmJkX3Bvb2xfbmFtZToKICAgICAgICAgICAgICAgICAgICBwb29sX2V4aXN0ID0gVHJ1
            ZQoKICAgICAgICAgICAgaWYgbm90IG1ldGFkYXRhX3Bvb2xfZXhpc3Q6CiAgICAgICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJQ
            cm92aWRlZCByYmRfZWNfbWV0YWRhdGFfcG9vbCBuYW1lLCIKICAgICAgICAgICAgICAgICAgICBm
            IiB7cmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZX0sIGRvZXMgbm90IGV4aXN0IgogICAgICAgICAg
            ICAgICAgKQogICAgICAgICAgICBpZiBub3QgcG9vbF9leGlzdDoKICAgICAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgZiJQcm92
            aWRlZCByYmRfZGF0YV9wb29sIG5hbWUsIHtyYmRfcG9vbF9uYW1lfSwgZG9lcyBub3QgZXhpc3Qi
            CiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIHJldHVybiByYmRfbWV0YWRhdGFfZWNfcG9v
            bF9uYW1lCgogICAgZGVmIGRyeV9ydW4oc2VsZiwgbXNnKToKICAgICAgICBpZiBzZWxmLl9hcmdf
            cGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHByaW50KCJFeGVjdXRlOiAiICsgIiciICsgbXNn
            ICsgIiciKQoKICAgIGRlZiB2YWxpZGF0ZV9yZ3dfZW5kcG9pbnRfdGxzX2NlcnQoc2VsZik6CiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZ3dfdGxzX2NlcnRfcGF0aDoKICAgICAgICAgICAg
            d2l0aCBvcGVuKHNlbGYuX2FyZ19wYXJzZXIucmd3X3Rsc19jZXJ0X3BhdGgsIGVuY29kaW5nPSJ1
            dGY4IikgYXMgZjoKICAgICAgICAgICAgICAgIGNvbnRlbnRzID0gZi5yZWFkKCkKICAgICAgICAg
            ICAgICAgIHJldHVybiBjb250ZW50cy5yc3RyaXAoKQoKICAgIGRlZiBfY2hlY2tfY29uZmxpY3Rp
            bmdfb3B0aW9ucyhzZWxmKToKICAgICAgICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNlci51cGdyYWRl
            IGFuZCBub3Qgc2VsZi5fYXJnX3BhcnNlci5yYmRfZGF0YV9wb29sX25hbWU6CiAgICAgICAgICAg
            IHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAiRWl0aGVy
            ICctLXVwZ3JhZGUnIG9yICctLXJiZC1kYXRhLXBvb2wtbmFtZSA8cG9vbF9uYW1lPicgc2hvdWxk
            IGJlIHNwZWNpZmllZCIKICAgICAgICAgICAgKQoKICAgIGRlZiBfaW52YWxpZF9lbmRwb2ludChz
            ZWxmLCBlbmRwb2ludF9zdHIpOgogICAgICAgICMgZXh0cmFjdCB0aGUgcG9ydCBieSBnZXR0aW5n
            IHRoZSBsYXN0IHNwbGl0IG9uIGA6YCBkZWxpbWl0ZXIKICAgICAgICB0cnk6CiAgICAgICAgICAg
            IGVuZHBvaW50X3N0cl9pcCwgcG9ydCA9IGVuZHBvaW50X3N0ci5yc3BsaXQoIjoiLCAxKQogICAg
            ICAgIGV4Y2VwdCBWYWx1ZUVycm9yOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJl
            RXhjZXB0aW9uKGYiTm90IGEgcHJvcGVyIGVuZHBvaW50OiB7ZW5kcG9pbnRfc3RyfSIpCgogICAg
            ICAgIHRyeToKICAgICAgICAgICAgaWYgZW5kcG9pbnRfc3RyX2lwWzBdID09ICJbIjoKICAgICAg
            ICAgICAgICAgIGVuZHBvaW50X3N0cl9pcCA9IGVuZHBvaW50X3N0cl9pcFsxIDogbGVuKGVuZHBv
            aW50X3N0cl9pcCkgLSAxXQogICAgICAgICAgICBpcF90eXBlID0gKAogICAgICAgICAgICAgICAg
            IklQdjQiIGlmIHR5cGUoaXBfYWRkcmVzcyhlbmRwb2ludF9zdHJfaXApKSBpcyBJUHY0QWRkcmVz
            cyBlbHNlICJJUHY2IgogICAgICAgICAgICApCiAgICAgICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAg
            ICAgICAgICAgIGlwX3R5cGUgPSAiRlFETiIKICAgICAgICBpZiBub3QgcG9ydC5pc2RpZ2l0KCk6
            CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oZiJQb3J0IG5vdCB2
            YWxpZDoge3BvcnR9IikKICAgICAgICBpbnRQb3J0ID0gaW50KHBvcnQpCiAgICAgICAgaWYgaW50
            UG9ydCA8IDEgb3IgaW50UG9ydCA+IDIqKjE2IC0gMToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0
            aW9uRmFpbHVyZUV4Y2VwdGlvbihmIk91dCBvZiByYW5nZSBwb3J0IG51bWJlcjoge3BvcnR9IikK
            CiAgICAgICAgcmV0dXJuIGlwX3R5cGUKCiAgICBkZWYgZW5kcG9pbnRfZGlhbChzZWxmLCBlbmRw
            b2ludF9zdHIsIGlwX3R5cGUsIHRpbWVvdXQ9MywgY2VydD1Ob25lKToKICAgICAgICAjIGlmIHRo
            ZSAnY2x1c3RlcicgaW5zdGFuY2UgaXMgYSBkdW1teSBvbmUsCiAgICAgICAgIyBkb24ndCB0cnkg
            dG8gcmVhY2ggb3V0IHRvIHRoZSBlbmRwb2ludAogICAgICAgIGlmIGlzaW5zdGFuY2Uoc2VsZi5j
            bHVzdGVyLCBEdW1teVJhZG9zKToKICAgICAgICAgICAgcmV0dXJuICIiLCAiIiwgIiIKICAgICAg
            ICBpZiBpcF90eXBlID09ICJJUHY2IjoKICAgICAgICAgICAgdHJ5OgogICAgICAgICAgICAgICAg
            ZW5kcG9pbnRfc3RyX2lwLCBlbmRwb2ludF9zdHJfcG9ydCA9IGVuZHBvaW50X3N0ci5yc3BsaXQo
            IjoiLCAxKQogICAgICAgICAgICBleGNlcHQgVmFsdWVFcnJvcjoKICAgICAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgZiJOb3Qg
            YSBwcm9wZXIgZW5kcG9pbnQ6IHtlbmRwb2ludF9zdHJ9IgogICAgICAgICAgICAgICAgKQogICAg
            ICAgICAgICBpZiBlbmRwb2ludF9zdHJfaXBbMF0gIT0gIlsiOgogICAgICAgICAgICAgICAgZW5k
            cG9pbnRfc3RyX2lwID0gIlsiICsgZW5kcG9pbnRfc3RyX2lwICsgIl0iCiAgICAgICAgICAgIGVu
            ZHBvaW50X3N0ciA9ICI6Ii5qb2luKFtlbmRwb2ludF9zdHJfaXAsIGVuZHBvaW50X3N0cl9wb3J0
            XSkKCiAgICAgICAgcHJvdG9jb2xzID0gWyJodHRwIiwgImh0dHBzIl0KICAgICAgICByZXNwb25z
            ZV9lcnJvciA9IE5vbmUKICAgICAgICBmb3IgcHJlZml4IGluIHByb3RvY29sczoKICAgICAgICAg
            ICAgdHJ5OgogICAgICAgICAgICAgICAgZXAgPSBmIntwcmVmaXh9Oi8ve2VuZHBvaW50X3N0cn0i
            CiAgICAgICAgICAgICAgICB2ZXJpZnkgPSBOb25lCiAgICAgICAgICAgICAgICAjIElmIHZlcmlm
            eSBpcyBzZXQgdG8gYSBwYXRoIHRvIGEgZGlyZWN0b3J5LAogICAgICAgICAgICAgICAgIyB0aGUg
            ZGlyZWN0b3J5IG11c3QgaGF2ZSBiZWVuIHByb2Nlc3NlZCB1c2luZyB0aGUgY19yZWhhc2ggdXRp
            bGl0eSBzdXBwbGllZCB3aXRoIE9wZW5TU0wuCiAgICAgICAgICAgICAgICBpZiBwcmVmaXggPT0g
            Imh0dHBzIiBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfc2tpcF90bHM6CiAgICAgICAgICAgICAg
            ICAgICAgdmVyaWZ5ID0gRmFsc2UKICAgICAgICAgICAgICAgICAgICByID0gcmVxdWVzdHMuaGVh
            ZChlcCwgdGltZW91dD10aW1lb3V0LCB2ZXJpZnk9RmFsc2UpCiAgICAgICAgICAgICAgICBlbGlm
            IHByZWZpeCA9PSAiaHR0cHMiIGFuZCBjZXJ0OgogICAgICAgICAgICAgICAgICAgIHZlcmlmeSA9
            IGNlcnQKICAgICAgICAgICAgICAgICAgICByID0gcmVxdWVzdHMuaGVhZChlcCwgdGltZW91dD10
            aW1lb3V0LCB2ZXJpZnk9Y2VydCkKICAgICAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAg
            ICAgICAgciA9IHJlcXVlc3RzLmhlYWQoZXAsIHRpbWVvdXQ9dGltZW91dCkKICAgICAgICAgICAg
            ICAgIGlmIHIuc3RhdHVzX2NvZGUgPT0gMjAwOgogICAgICAgICAgICAgICAgICAgIHJldHVybiBw
            cmVmaXgsIHZlcmlmeSwgIiIKICAgICAgICAgICAgZXhjZXB0IEV4Y2VwdGlvbiBhcyBlcnI6CiAg
            ICAgICAgICAgICAgICByZXNwb25zZV9lcnJvciA9IGVycgogICAgICAgICAgICAgICAgY29udGlu
            dWUKICAgICAgICBzeXMuc3RkZXJyLndyaXRlKAogICAgICAgICAgICBmInVuYWJsZSB0byBjb25u
            ZWN0IHRvIGVuZHBvaW50OiB7ZW5kcG9pbnRfc3RyfSwgZmFpbGVkIGVycm9yOiB7cmVzcG9uc2Vf
            ZXJyb3J9IgogICAgICAgICkKICAgICAgICByZXR1cm4gKAogICAgICAgICAgICAiIiwKICAgICAg
            ICAgICAgIiIsCiAgICAgICAgICAgICgiLTEiKSwKICAgICAgICApCgogICAgZGVmIF9faW5pdF9f
            KHNlbGYsIGFyZ19saXN0PU5vbmUpOgogICAgICAgIHNlbGYub3V0X21hcCA9IHt9CiAgICAgICAg
            c2VsZi5fZXhjbHVkZWRfa2V5cyA9IHNldCgpCiAgICAgICAgc2VsZi5fYXJnX3BhcnNlciA9IHNl
            bGYuZ2VuX2FyZ19wYXJzZXIoYXJnc190b19wYXJzZT1hcmdfbGlzdCkKICAgICAgICBzZWxmLl9j
            aGVja19jb25mbGljdGluZ19vcHRpb25zKCkKICAgICAgICBzZWxmLnJ1bl9hc191c2VyID0gc2Vs
            Zi5fYXJnX3BhcnNlci5ydW5fYXNfdXNlcgogICAgICAgIHNlbGYub3V0cHV0X2ZpbGUgPSBzZWxm
            Ll9hcmdfcGFyc2VyLm91dHB1dAogICAgICAgIHNlbGYuY2VwaF9jb25mID0gc2VsZi5fYXJnX3Bh
            cnNlci5jZXBoX2NvbmYKICAgICAgICBzZWxmLmNlcGhfa2V5cmluZyA9IHNlbGYuX2FyZ19wYXJz
            ZXIua2V5cmluZwogICAgICAgICMgaWYgdXNlciBub3QgcHJvdmlkZWQsIGdpdmUgYSBkZWZhdWx0
            IHVzZXIKICAgICAgICBpZiBub3Qgc2VsZi5ydW5fYXNfdXNlciBhbmQgbm90IHNlbGYuX2FyZ19w
            YXJzZXIudXBncmFkZToKICAgICAgICAgICAgc2VsZi5ydW5fYXNfdXNlciA9IHNlbGYuRVhURVJO
            QUxfVVNFUl9OQU1FCiAgICAgICAgaWYgbm90IHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJl
            Zml4IGFuZCBub3Qgc2VsZi5fYXJnX3BhcnNlci51cGdyYWRlOgogICAgICAgICAgICBzZWxmLl9h
            cmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeCA9IHNlbGYuREVGQVVMVF9SR1dfUE9PTF9QUkVGSVgK
            ICAgICAgICBpZiBzZWxmLmNlcGhfY29uZjoKICAgICAgICAgICAga3dhcmdzID0ge30KICAgICAg
            ICAgICAgaWYgc2VsZi5jZXBoX2tleXJpbmc6CiAgICAgICAgICAgICAgICBrd2FyZ3NbImNvbmYi
            XSA9IHsia2V5cmluZyI6IHNlbGYuY2VwaF9rZXlyaW5nfQogICAgICAgICAgICBzZWxmLmNsdXN0
            ZXIgPSByYWRvcy5SYWRvcyhjb25mZmlsZT1zZWxmLmNlcGhfY29uZiwgKiprd2FyZ3MpCiAgICAg
            ICAgZWxzZToKICAgICAgICAgICAgc2VsZi5jbHVzdGVyID0gcmFkb3MuUmFkb3MoKQogICAgICAg
            ICAgICBzZWxmLmNsdXN0ZXIuY29uZl9yZWFkX2ZpbGUoKQogICAgICAgIHNlbGYuY2x1c3Rlci5j
            b25uZWN0KCkKCiAgICBkZWYgc2h1dGRvd24oc2VsZik6CiAgICAgICAgaWYgc2VsZi5jbHVzdGVy
            LnN0YXRlID09ICJjb25uZWN0ZWQiOgogICAgICAgICAgICBzZWxmLmNsdXN0ZXIuc2h1dGRvd24o
            KQoKICAgIGRlZiBnZXRfZnNpZChzZWxmKToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRy
            eV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGggZnNpZCIpCiAgICAg
            ICAgcmV0dXJuIHN0cihzZWxmLmNsdXN0ZXIuZ2V0X2ZzaWQoKSkKCiAgICBkZWYgX2NvbW1vbl9j
            bWRfanNvbl9nZW4oc2VsZiwgY21kX2pzb24pOgogICAgICAgIGNtZCA9IGpzb24uZHVtcHMoY21k
            X2pzb24sIHNvcnRfa2V5cz1UcnVlKQogICAgICAgIHJldF92YWwsIGNtZF9vdXQsIGVycl9tc2cg
            PSBzZWxmLmNsdXN0ZXIubW9uX2NvbW1hbmQoY21kLCBiIiIpCiAgICAgICAgaWYgc2VsZi5fYXJn
            X3BhcnNlci52ZXJib3NlOgogICAgICAgICAgICBwcmludChmIkNvbW1hbmQgSW5wdXQ6IHtjbWR9
            IikKICAgICAgICAgICAgcHJpbnQoCiAgICAgICAgICAgICAgICBmIlJldHVybiBWYWw6IHtyZXRf
            dmFsfVxuQ29tbWFuZCBPdXRwdXQ6IHtjbWRfb3V0fVxuIgogICAgICAgICAgICAgICAgZiJFcnJv
            ciBNZXNzYWdlOiB7ZXJyX21zZ31cbi0tLS0tLS0tLS1cbiIKICAgICAgICAgICAgKQogICAgICAg
            IGpzb25fb3V0ID0ge30KICAgICAgICAjIGlmIHRoZXJlIGlzIG5vIGVycm9yIChpLmU7IHJldF92
            YWwgaXMgWkVSTykgYW5kICdjbWRfb3V0JyBpcyBub3QgZW1wdHkKICAgICAgICAjIHRoZW4gY29u
            dmVydCAnY21kX291dCcgdG8gYSBqc29uIG91dHB1dAogICAgICAgIGlmIHJldF92YWwgPT0gMCBh
            bmQgY21kX291dDoKICAgICAgICAgICAganNvbl9vdXQgPSBqc29uLmxvYWRzKGNtZF9vdXQpCiAg
            ICAgICAgcmV0dXJuIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnCgogICAgZGVmIGdldF9jZXBo
            X2V4dGVybmFsX21vbl9kYXRhKHNlbGYpOgogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAi
            cXVvcnVtX3N0YXR1cyIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgaWYgc2VsZi5fYXJnX3Bh
            cnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKCJjZXBoICIgKyBj
            bWRfanNvblsicHJlZml4Il0pCiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBz
            ZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICMgaWYgdGhlcmUgaXMg
            YW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihq
            c29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgICAgICIncXVvcnVtX3N0YXR1cycgY29tbWFuZCBmYWlsZWQuXG4iCiAg
            ICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxm
            LkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgKQogICAgICAgIHFfbGVhZGVyX25hbWUg
            PSBqc29uX291dFsicXVvcnVtX2xlYWRlcl9uYW1lIl0KICAgICAgICBxX2xlYWRlcl9kZXRhaWxz
            ID0ge30KICAgICAgICBxX2xlYWRlcl9tYXRjaGluZ19saXN0ID0gWwogICAgICAgICAgICBsIGZv
            ciBsIGluIGpzb25fb3V0WyJtb25tYXAiXVsibW9ucyJdIGlmIGxbIm5hbWUiXSA9PSBxX2xlYWRl
            cl9uYW1lCiAgICAgICAgXQogICAgICAgIGlmIGxlbihxX2xlYWRlcl9tYXRjaGluZ19saXN0KSA9
            PSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKCJObyBtYXRj
            aGluZyAnbW9uJyBkZXRhaWxzIGZvdW5kIikKICAgICAgICBxX2xlYWRlcl9kZXRhaWxzID0gcV9s
            ZWFkZXJfbWF0Y2hpbmdfbGlzdFswXQogICAgICAgICMgZ2V0IHRoZSBhZGRyZXNzIHZlY3RvciBv
            ZiB0aGUgcXVvcnVtLWxlYWRlcgogICAgICAgIHFfbGVhZGVyX2FkZHJ2ZWMgPSBxX2xlYWRlcl9k
            ZXRhaWxzLmdldCgicHVibGljX2FkZHJzIiwge30pLmdldCgiYWRkcnZlYyIsIFtdKQogICAgICAg
            ICMgaWYgdGhlIHF1b3J1bS1sZWFkZXIgaGFzIG9ubHkgb25lIGFkZHJlc3MgaW4gdGhlIGFkZHJl
            c3MtdmVjdG9yCiAgICAgICAgIyBhbmQgaXQgaXMgb2YgdHlwZSAndjInIChpZTsgd2l0aCA8SVA+
            OjMzMDApLAogICAgICAgICMgcmFpc2UgYW4gZXhjZXB0aW9uIHRvIG1ha2UgdXNlciBhd2FyZSB0
            aGF0CiAgICAgICAgIyB0aGV5IGhhdmUgdG8gZW5hYmxlICd2MScgKGllOyB3aXRoIDxJUD46Njc4
            OSkgdHlwZSBhcyB3ZWxsCiAgICAgICAgaWYgbGVuKHFfbGVhZGVyX2FkZHJ2ZWMpID09IDEgYW5k
            IHFfbGVhZGVyX2FkZHJ2ZWNbMF1bInR5cGUiXSA9PSAidjIiOgogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgIk9ubHkgJ3YyJyBhZGRy
            ZXNzIHR5cGUgaXMgZW5hYmxlZCwgdXNlciBzaG91bGQgYWxzbyBlbmFibGUgJ3YxJyB0eXBlIGFz
            IHdlbGwiCiAgICAgICAgICAgICkKICAgICAgICBpcF9hZGRyID0gc3RyKHFfbGVhZGVyX2RldGFp
            bHNbInB1YmxpY19hZGRyIl0uc3BsaXQoIi8iKVswXSkKCiAgICAgICAgaWYgc2VsZi5fYXJnX3Bh
            cnNlci52Ml9wb3J0X2VuYWJsZToKICAgICAgICAgICAgaWYgbGVuKHFfbGVhZGVyX2FkZHJ2ZWMp
            ID4gMToKICAgICAgICAgICAgICAgIGlmIHFfbGVhZGVyX2FkZHJ2ZWNbMF1bInR5cGUiXSA9PSAi
            djIiOgogICAgICAgICAgICAgICAgICAgIGlwX2FkZHIgPSBxX2xlYWRlcl9hZGRydmVjWzBdWyJh
            ZGRyIl0KICAgICAgICAgICAgICAgIGVsaWYgcV9sZWFkZXJfYWRkcnZlY1sxXVsidHlwZSJdID09
            ICJ2MiI6CiAgICAgICAgICAgICAgICAgICAgaXBfYWRkciA9IHFfbGVhZGVyX2FkZHJ2ZWNbMV1b
            ImFkZHIiXQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0
            ZSgKICAgICAgICAgICAgICAgICAgICAiJ3YyJyBhZGRyZXNzIHR5cGUgbm90IHByZXNlbnQsIGFu
            ZCAndjItcG9ydC1lbmFibGUnIGZsYWcgaXMgcHJvdmlkZWQiCiAgICAgICAgICAgICAgICApCgog
            ICAgICAgIHJldHVybiBmIntzdHIocV9sZWFkZXJfbmFtZSl9PXtpcF9hZGRyfSIKCiAgICBkZWYg
            X2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoc2VsZiwgaG9zdF9uYW1lLCBwb3J0LCBpcF90eXBlKToK
            ICAgICAgICAjIGlmICdjbHVzdGVyJyBpbnN0YW5jZSBpcyBhIGR1bW15IHR5cGUsCiAgICAgICAg
            IyBjYWxsIHRoZSBkdW1teSBpbnN0YW5jZSdzICJjb252ZXJ0IiBtZXRob2QKICAgICAgICBpZiBu
            b3QgaG9zdF9uYW1lOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KCJFbXB0eSBob3N0bmFtZSBwcm92aWRlZCIpCiAgICAgICAgaWYgaXNpbnN0YW5jZShzZWxmLmNs
            dXN0ZXIsIER1bW15UmFkb3MpOgogICAgICAgICAgICByZXR1cm4gc2VsZi5jbHVzdGVyLl9jb252
            ZXJ0X2hvc3RuYW1lX3RvX2lwKGhvc3RfbmFtZSkKCiAgICAgICAgaWYgaXBfdHlwZSA9PSAiRlFE
            TiI6CiAgICAgICAgICAgICMgY2hlY2sgd2hpY2ggaXAgRlFETiBzaG91bGQgYmUgY29udmVydGVk
            IHRvLCBJUHY0IG9yIElQdjYKICAgICAgICAgICAgIyBjaGVjayB0aGUgaG9zdCBpcCwgdGhlIGVu
            ZHBvaW50IGlwIHR5cGUgd291bGQgYmUgc2ltaWxhciB0byBob3N0IGlwCiAgICAgICAgICAgIGNt
            ZF9qc29uID0geyJwcmVmaXgiOiAib3JjaCBob3N0IGxzIiwgImZvcm1hdCI6ICJqc29uIn0KICAg
            ICAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pz
            b25fZ2VuKGNtZF9qc29uKQogICAgICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1
            bCBhdHRlbXB0LAogICAgICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9
            PSAwOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAg
            ICAgICAgICAgICAgICAgICAiJ29yY2ggaG9zdCBscycgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAg
            ICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYgcmV0X3ZhbCAhPSAwIGVsc2Ugc2Vs
            Zi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGhvc3Rf
            YWRkciA9IGpzb25fb3V0WzBdWyJhZGRyIl0KICAgICAgICAgICAgIyBhZGQgOjgwIHNhbXBsZSBw
            b3J0IGluIGlwX3R5cGUsIGFzIF9pbnZhbGlkX2VuZHBvaW50IGFsc28gdmVyaWZ5IHBvcnQKICAg
            ICAgICAgICAgaG9zdF9pcF90eXBlID0gc2VsZi5faW52YWxpZF9lbmRwb2ludChob3N0X2FkZHIg
            KyAiOjgwIikKICAgICAgICAgICAgaW1wb3J0IHNvY2tldAoKICAgICAgICAgICAgIyBleGFtcGxl
            IG91dHB1dCBbKDxBZGRyZXNzRmFtaWx5LkFGX0lORVQ6IDI+LCA8U29ja2V0S2luZC5TT0NLX1NU
            UkVBTTogMT4sIDYsICcnLCAoJzkzLjE4NC4yMTYuMzQnLCA4MCkpLCAuLi5dCiAgICAgICAgICAg
            ICMgd2UgbmVlZCB0byBnZXQgOTMuMTg0LjIxNi4zNCBzbyBpdCB3b3VsZCBiZSBpcFswXVs0XVsw
            XQogICAgICAgICAgICBpZiBob3N0X2lwX3R5cGUgPT0gIklQdjYiOgogICAgICAgICAgICAgICAg
            aXAgPSBzb2NrZXQuZ2V0YWRkcmluZm8oCiAgICAgICAgICAgICAgICAgICAgaG9zdF9uYW1lLCBw
            b3J0LCBmYW1pbHk9c29ja2V0LkFGX0lORVQ2LCBwcm90bz1zb2NrZXQuSVBQUk9UT19UQ1AKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxpZiBob3N0X2lwX3R5cGUgPT0gIklQdjQiOgog
            ICAgICAgICAgICAgICAgaXAgPSBzb2NrZXQuZ2V0YWRkcmluZm8oCiAgICAgICAgICAgICAgICAg
            ICAgaG9zdF9uYW1lLCBwb3J0LCBmYW1pbHk9c29ja2V0LkFGX0lORVQsIHByb3RvPXNvY2tldC5J
            UFBST1RPX1RDUAogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBkZWwgc29ja2V0CiAgICAg
            ICAgICAgIHJldHVybiBpcFswXVs0XVswXQogICAgICAgIHJldHVybiBob3N0X25hbWUKCiAgICBk
            ZWYgZ2V0X2FjdGl2ZV9hbmRfc3RhbmRieV9tZ3JzKHNlbGYpOgogICAgICAgIGlmIHNlbGYuX2Fy
            Z19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuICIiLCBzZWxmLmRyeV9ydW4oImNl
            cGggc3RhdHVzIikKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQgPSBzZWxmLl9hcmdf
            cGFyc2VyLm1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydAogICAgICAgIG1vbml0b3JpbmdfZW5kcG9p
            bnRfaXBfbGlzdCA9IHNlbGYuX2FyZ19wYXJzZXIubW9uaXRvcmluZ19lbmRwb2ludAogICAgICAg
            IHN0YW5kYnlfbWdycyA9IFtdCiAgICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBf
            bGlzdDoKICAgICAgICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJzdGF0dXMiLCAiZm9ybWF0
            IjogImpzb24ifQogICAgICAgICAgICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYu
            X2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgICAgICMgaWYgdGhlcmUgaXMg
            YW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBs
            ZW4oanNvbl9vdXQpID09IDA6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJl
            RXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICInbWdyIHNlcnZpY2VzJyBjb21tYW5kIGZh
            aWxlZC5cbiIKICAgICAgICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFs
            ICE9IDAgZWxzZSBzZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgICAgICkKICAg
            ICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludCA9ICgKICAgICAgICAgICAgICAgIGpzb25fb3V0
            LmdldCgibWdybWFwIiwge30pLmdldCgic2VydmljZXMiLCB7fSkuZ2V0KCJwcm9tZXRoZXVzIiwg
            IiIpCiAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnQ6
            CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAg
            ICAgICAgICAgICAgICJjYW4ndCBmaW5kIG1vbml0b3JpbmdfZW5kcG9pbnQsIHByb21ldGhldXMg
            bW9kdWxlIG1pZ2h0IG5vdCBiZSBlbmFibGVkLCAiCiAgICAgICAgICAgICAgICAgICAgImVuYWJs
            ZSB0aGUgbW9kdWxlIGJ5IHJ1bm5pbmcgJ2NlcGggbWdyIG1vZHVsZSBlbmFibGUgcHJvbWV0aGV1
            cyciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICMgbm93IGNoZWNrIHRoZSBzdGFuZC1i
            eSBtZ3ItcwogICAgICAgICAgICBzdGFuZGJ5X2FyciA9IGpzb25fb3V0LmdldCgibWdybWFwIiwg
            e30pLmdldCgic3RhbmRieXMiLCBbXSkKICAgICAgICAgICAgZm9yIGVhY2hfc3RhbmRieSBpbiBz
            dGFuZGJ5X2FycjoKICAgICAgICAgICAgICAgIGlmICJuYW1lIiBpbiBlYWNoX3N0YW5kYnkua2V5
            cygpOgogICAgICAgICAgICAgICAgICAgIHN0YW5kYnlfbWdycy5hcHBlbmQoZWFjaF9zdGFuZGJ5
            WyJuYW1lIl0pCiAgICAgICAgICAgIHRyeToKICAgICAgICAgICAgICAgIHBhcnNlZF9lbmRwb2lu
            dCA9IHVybHBhcnNlKG1vbml0b3JpbmdfZW5kcG9pbnQpCiAgICAgICAgICAgIGV4Y2VwdCBWYWx1
            ZUVycm9yOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigK
            ICAgICAgICAgICAgICAgICAgICBmImludmFsaWQgZW5kcG9pbnQ6IHttb25pdG9yaW5nX2VuZHBv
            aW50fSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9p
            cF9saXN0ID0gcGFyc2VkX2VuZHBvaW50Lmhvc3RuYW1lCiAgICAgICAgICAgIGlmIG5vdCBtb25p
            dG9yaW5nX2VuZHBvaW50X3BvcnQ6CiAgICAgICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50
            X3BvcnQgPSBzdHIocGFyc2VkX2VuZHBvaW50LnBvcnQpCgogICAgICAgICMgaWYgbW9uaXRvcmlu
            ZyBlbmRwb2ludCBwb3J0IGlzIG5vdCBzZXQsIHB1dCBhIGRlZmF1bHQgbW9uIHBvcnQKICAgICAg
            ICBpZiBub3QgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0OgogICAgICAgICAgICBtb25pdG9yaW5n
            X2VuZHBvaW50X3BvcnQgPSBzZWxmLkRFRkFVTFRfTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUCgog
            ICAgICAgICMgdXNlciBjb3VsZCBnaXZlIGNvbW1hIGFuZCBzcGFjZSBzZXBhcmF0ZWQgaW5wdXRz
            IChsaWtlIC0tbW9uaXRvcmluZy1lbmRwb2ludD0iPGlwMT4sIDxpcDI+IikKICAgICAgICBtb25p
            dG9yaW5nX2VuZHBvaW50X2lwX2xpc3QgPSBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3QucmVw
            bGFjZSgiLCIsICIgIikKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXQg
            PSBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Quc3BsaXQoKQogICAgICAgICMgaWYgbW9uaXRv
            cmluZy1lbmRwb2ludCBjb3VsZCBub3QgYmUgZm91bmQsIHJhaXNlIGFuIGVycm9yCiAgICAgICAg
            aWYgbGVuKG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdF9zcGxpdCkgPT0gMDoKICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigiTm8gJ21vbml0b3JpbmctZW5kcG9p
            bnQnIGZvdW5kIikKICAgICAgICAjIGZpcnN0IGlwIGlzIHRyZWF0ZWQgYXMgdGhlIG1haW4gbW9u
            aXRvcmluZy1lbmRwb2ludAogICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfaXAgPSBtb25pdG9y
            aW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXRbMF0KICAgICAgICAjIHJlc3Qgb2YgdGhlIGlwLXMg
            YXJlIGFkZGVkIHRvIHRoZSAnc3RhbmRieV9tZ3JzJyBsaXN0CiAgICAgICAgc3RhbmRieV9tZ3Jz
            LmV4dGVuZChtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXRbMTpdKQogICAgICAgIGZh
            aWxlZF9pcCA9IG1vbml0b3JpbmdfZW5kcG9pbnRfaXAKCiAgICAgICAgbW9uaXRvcmluZ19lbmRw
            b2ludCA9ICI6Ii5qb2luKAogICAgICAgICAgICBbbW9uaXRvcmluZ19lbmRwb2ludF9pcCwgbW9u
            aXRvcmluZ19lbmRwb2ludF9wb3J0XQogICAgICAgICkKICAgICAgICBpcF90eXBlID0gc2VsZi5f
            aW52YWxpZF9lbmRwb2ludChtb25pdG9yaW5nX2VuZHBvaW50KQogICAgICAgIHRyeToKICAgICAg
            ICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcCA9IHNlbGYuX2NvbnZlcnRfaG9zdG5hbWVfdG9f
            aXAoCiAgICAgICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X2lwLCBtb25pdG9yaW5nX2Vu
            ZHBvaW50X3BvcnQsIGlwX3R5cGUKICAgICAgICAgICAgKQogICAgICAgICAgICAjIGNvbGxlY3Qg
            YWxsIHRoZSAnc3RhbmQtYnknIG1nciBpcHMKICAgICAgICAgICAgbWdyX2lwcyA9IFtdCiAgICAg
            ICAgICAgIGZvciBlYWNoX3N0YW5kYnlfbWdyIGluIHN0YW5kYnlfbWdyczoKICAgICAgICAgICAg
            ICAgIGZhaWxlZF9pcCA9IGVhY2hfc3RhbmRieV9tZ3IKICAgICAgICAgICAgICAgIG1ncl9pcHMu
            YXBwZW5kKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAo
            CiAgICAgICAgICAgICAgICAgICAgICAgIGVhY2hfc3RhbmRieV9tZ3IsIG1vbml0b3JpbmdfZW5k
            cG9pbnRfcG9ydCwgaXBfdHlwZQogICAgICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAg
            ICkKICAgICAgICBleGNlcHQ6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNl
            cHRpb24oCiAgICAgICAgICAgICAgICBmIkNvbnZlcnNpb24gb2YgaG9zdDoge2ZhaWxlZF9pcH0g
            dG8gSVAgZmFpbGVkLiAiCiAgICAgICAgICAgICAgICAiUGxlYXNlIGVudGVyIHRoZSBJUCBhZGRy
            ZXNzZXMgb2YgYWxsIHRoZSBjZXBoLW1ncnMgd2l0aCB0aGUgJy0tbW9uaXRvcmluZy1lbmRwb2lu
            dCcgZmxhZyIKICAgICAgICAgICAgKQoKICAgICAgICBfLCBfLCBlcnIgPSBzZWxmLmVuZHBvaW50
            X2RpYWwobW9uaXRvcmluZ19lbmRwb2ludCwgaXBfdHlwZSkKICAgICAgICBpZiBlcnIgPT0gIi0x
            IjoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihlcnIpCiAgICAg
            ICAgIyBhZGQgdGhlIHZhbGlkYXRlZCBhY3RpdmUgbWdyIElQIGludG8gdGhlIGZpcnN0IGluZGV4
            CiAgICAgICAgbWdyX2lwcy5pbnNlcnQoMCwgbW9uaXRvcmluZ19lbmRwb2ludF9pcCkKICAgICAg
            ICBhbGxfbWdyX2lwc19zdHIgPSAiLCIuam9pbihtZ3JfaXBzKQogICAgICAgIHJldHVybiBhbGxf
            bWdyX2lwc19zdHIsIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydAoKICAgIGRlZiBjaGVja191c2Vy
            X2V4aXN0KHNlbGYsIHVzZXIpOgogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAiYXV0aCBn
            ZXQiLCAiZW50aXR5IjogZiJ7dXNlcn0iLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIHJldF92
            YWwsIGpzb25fb3V0LCBfID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAg
            ICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICBy
            ZXR1cm4gIiIKICAgICAgICByZXR1cm4gc3RyKGpzb25fb3V0WzBdWyJrZXkiXSkKCiAgICBkZWYg
            Z2V0X2NlcGhmc19wcm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50
            aXR5ID0gImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIgogICAgICAgIGNhcHMgPSB7CiAg
            ICAgICAgICAgICJtb24iOiAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCci
            LAogICAgICAgICAgICAibWdyIjogImFsbG93IHJ3IiwKICAgICAgICAgICAgIm9zZCI6ICJhbGxv
            dyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiLAogICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9h
            cmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uOgogICAgICAgICAgICBrOHNfY2x1
            c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1lCiAgICAgICAgICAg
            IGlmIGs4c19jbHVzdGVyX25hbWUgPT0gIiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRp
            b25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJrOHNfY2x1c3Rlcl9uYW1l
            IG5vdCBmb3VuZCwgcGxlYXNlIHNldCB0aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW0gPSBzZWxmLl9hcmdf
            cGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICAgICAgaWYgY2VwaGZzX2ZpbGVz
            eXN0ZW0gPT0gIiI6CiAgICAgICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1
            c3Rlcl9uYW1lfSIKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYi
            e2VudGl0eX0te2s4c19jbHVzdGVyX25hbWV9LXtjZXBoZnNfZmlsZXN5c3RlbX0iCiAgICAgICAg
            ICAgICAgICBjYXBzWyJvc2QiXSA9IGYiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT17Y2Vw
            aGZzX2ZpbGVzeXN0ZW19IgoKICAgICAgICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVmIGdl
            dF9jZXBoZnNfbm9kZV9jYXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gImNs
            aWVudC5jc2ktY2VwaGZzLW5vZGUiCiAgICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6
            ICJhbGxvdyByLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJt
            Z3IiOiAiYWxsb3cgcnciLAogICAgICAgICAgICAib3NkIjogImFsbG93IHJ3IHRhZyBjZXBoZnMg
            Kj0qIiwKICAgICAgICAgICAgIm1kcyI6ICJhbGxvdyBydyIsCiAgICAgICAgfQogICAgICAgIGlm
            IHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAg
            IGs4c19jbHVzdGVyX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAg
            ICAgICAgICAgaWYgazhzX2NsdXN0ZXJfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIms4c19jbHVz
            dGVyX25hbWUgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1rOHMtY2x1c3Rlci1uYW1lJyBm
            bGFnIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBjZXBoZnNfZmlsZXN5c3RlbSA9IHNl
            bGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgICAgICBpZiBjZXBo
            ZnNfZmlsZXN5c3RlbSA9PSAiIjoKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0t
            e2s4c19jbHVzdGVyX25hbWV9IgogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgZW50
            aXR5ID0gZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0te2NlcGhmc19maWxlc3lzdGVtfSIK
            ICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJdID0gZiJhbGxvdyBydyB0YWcgY2VwaGZzICo9e2Nl
            cGhmc19maWxlc3lzdGVtfSIKCiAgICAgICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBn
            ZXRfZW50aXR5KAogICAgICAgIHNlbGYsCiAgICAgICAgZW50aXR5LAogICAgICAgIHJiZF9wb29s
            X25hbWUsCiAgICAgICAgYWxpYXNfcmJkX3Bvb2xfbmFtZSwKICAgICAgICBrOHNfY2x1c3Rlcl9u
            YW1lLAogICAgICAgIHJhZG9zX25hbWVzcGFjZSwKICAgICk6CiAgICAgICAgaWYgKAogICAgICAg
            ICAgICByYmRfcG9vbF9uYW1lLmNvdW50KCIuIikgIT0gMAogICAgICAgICAgICBvciByYmRfcG9v
            bF9uYW1lLmNvdW50KCJfIikgIT0gMAogICAgICAgICAgICBvciBhbGlhc19yYmRfcG9vbF9uYW1l
            ICE9ICIiCiAgICAgICAgICAgICMgY2hlY2tpbmcgYWxpYXNfcmJkX3Bvb2xfbmFtZSBpcyBub3Qg
            ZW1wdHkgYXMgdGhlcmUgbWF5YmUgYSBzcGVjaWFsIGNoYXJhY3RlciB1c2VkIG90aGVyIHRoYW4g
            LiBvciBfCiAgICAgICAgKToKICAgICAgICAgICAgaWYgYWxpYXNfcmJkX3Bvb2xfbmFtZSA9PSAi
            IjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAg
            ICAgICAgICAgICAgICAgInBsZWFzZSBzZXQgdGhlICctLWFsaWFzLXJiZC1kYXRhLXBvb2wtbmFt
            ZScgZmxhZyBhcyB0aGUgcmJkIGRhdGEgcG9vbCBuYW1lIGNvbnRhaW5zICcuJyBvciAnXyciCiAg
            ICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmICgKICAgICAgICAgICAgICAgIGFsaWFzX3Ji
            ZF9wb29sX25hbWUuY291bnQoIi4iKSAhPSAwCiAgICAgICAgICAgICAgICBvciBhbGlhc19yYmRf
            cG9vbF9uYW1lLmNvdW50KCJfIikgIT0gMAogICAgICAgICAgICApOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJy0t
            YWxpYXMtcmJkLWRhdGEtcG9vbC1uYW1lJyBmbGFnIHZhbHVlIHNob3VsZCBub3QgY29udGFpbiAn
            Licgb3IgJ18nIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRp
            dHl9LXtrOHNfY2x1c3Rlcl9uYW1lfS17YWxpYXNfcmJkX3Bvb2xfbmFtZX0iCiAgICAgICAgZWxz
            ZToKICAgICAgICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0te3Ji
            ZF9wb29sX25hbWV9IgoKICAgICAgICBpZiByYWRvc19uYW1lc3BhY2U6CiAgICAgICAgICAgIGVu
            dGl0eSA9IGYie2VudGl0eX0te3JhZG9zX25hbWVzcGFjZX0iCiAgICAgICAgcmV0dXJuIGVudGl0
            eQoKICAgIGRlZiBnZXRfcmJkX3Byb3Zpc2lvbmVyX2NhcHNfYW5kX2VudGl0eShzZWxmKToKICAg
            ICAgICBlbnRpdHkgPSAiY2xpZW50LmNzaS1yYmQtcHJvdmlzaW9uZXIiCiAgICAgICAgY2FwcyA9
            IHsKICAgICAgICAgICAgIm1vbiI6ICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJs
            b2NrbGlzdCciLAogICAgICAgICAgICAibWdyIjogImFsbG93IHJ3IiwKICAgICAgICAgICAgIm9z
            ZCI6ICJwcm9maWxlIHJiZCIsCiAgICAgICAgfQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIu
            cmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBz
            ZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgICAgICBhbGlhc19yYmRf
            cG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5hbGlhc19yYmRfZGF0YV9wb29sX25hbWUKICAg
            ICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJf
            bmFtZQogICAgICAgICAgICByYWRvc19uYW1lc3BhY2UgPSBzZWxmLl9hcmdfcGFyc2VyLnJhZG9z
            X25hbWVzcGFjZQogICAgICAgICAgICBpZiByYmRfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAg
            ICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAg
            ICAibWFuZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1yYmQtZGF0YS1w
            b29sLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmIGs4c19jbHVz
            dGVyX25hbWUgPT0gIiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhj
            ZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBs
            ZWFzZSBzZXQgdGhlICctLWs4cy1jbHVzdGVyLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgIGVudGl0eSA9IHNlbGYuZ2V0X2VudGl0eSgKICAgICAgICAgICAgICAgIGVu
            dGl0eSwKICAgICAgICAgICAgICAgIHJiZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBhbGlh
            c19yYmRfcG9vbF9uYW1lLAogICAgICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAg
            ICAgICAgICAgIHJhZG9zX25hbWVzcGFjZSwKICAgICAgICAgICAgKQogICAgICAgICAgICBpZiBy
            YWRvc19uYW1lc3BhY2UgIT0gIiI6CiAgICAgICAgICAgICAgICBjYXBzWyJvc2QiXSA9ICgKICAg
            ICAgICAgICAgICAgICAgICBmInByb2ZpbGUgcmJkIHBvb2w9e3JiZF9wb29sX25hbWV9IG5hbWVz
            cGFjZT17cmFkb3NfbmFtZXNwYWNlfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxz
            ZToKICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJdID0gZiJwcm9maWxlIHJiZCBwb29sPXtyYmRf
            cG9vbF9uYW1lfSIKCiAgICAgICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfcmJk
            X25vZGVfY2Fwc19hbmRfZW50aXR5KHNlbGYpOgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3Np
            LXJiZC1ub2RlIgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAgICJtb24iOiAicHJvZmlsZSBy
            YmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwKICAgICAgICAgICAgIm9zZCI6ICJw
            cm9maWxlIHJiZCIsCiAgICAgICAgfQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJp
            Y3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9h
            cmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9u
            YW1lID0gc2VsZi5fYXJnX3BhcnNlci5hbGlhc19yYmRfZGF0YV9wb29sX25hbWUKICAgICAgICAg
            ICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQog
            ICAgICAgICAgICByYWRvc19uYW1lc3BhY2UgPSBzZWxmLl9hcmdfcGFyc2VyLnJhZG9zX25hbWVz
            cGFjZQogICAgICAgICAgICBpZiByYmRfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAibWFu
            ZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1yYmQtZGF0YS1wb29sLW5h
            bWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmIGs4c19jbHVzdGVyX25h
            bWUgPT0gIiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KAogICAgICAgICAgICAgICAgICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBz
            ZXQgdGhlICctLWs4cy1jbHVzdGVyLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAg
            ICAgICAgIGVudGl0eSA9IHNlbGYuZ2V0X2VudGl0eSgKICAgICAgICAgICAgICAgIGVudGl0eSwK
            ICAgICAgICAgICAgICAgIHJiZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBhbGlhc19yYmRf
            cG9vbF9uYW1lLAogICAgICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAgICAgICAg
            ICAgIHJhZG9zX25hbWVzcGFjZSwKICAgICAgICAgICAgKQogICAgICAgICAgICBpZiByYWRvc19u
            YW1lc3BhY2UgIT0gIiI6CiAgICAgICAgICAgICAgICBjYXBzWyJvc2QiXSA9ICgKICAgICAgICAg
            ICAgICAgICAgICBmInByb2ZpbGUgcmJkIHBvb2w9e3JiZF9wb29sX25hbWV9IG5hbWVzcGFjZT17
            cmFkb3NfbmFtZXNwYWNlfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxzZToKICAg
            ICAgICAgICAgICAgIGNhcHNbIm9zZCJdID0gZiJwcm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9u
            YW1lfSIKCiAgICAgICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfZGVmYXVsdFVz
            ZXJfY2Fwc19hbmRfZW50aXR5KHNlbGYpOgogICAgICAgIGVudGl0eSA9IHNlbGYucnVuX2FzX3Vz
            ZXIKICAgICAgICBjYXBzID0gewogICAgICAgICAgICAibW9uIjogImFsbG93IHIsIGFsbG93IGNv
            bW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFuZCB2ZXJzaW9uIiwKICAgICAgICAgICAg
            Im1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsCiAgICAgICAgICAgICJvc2QiOiBmInByb2Zp
            bGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9e3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bv
            b2xfcHJlZml4fS5yZ3cubWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9v
            bD17c2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5jb250cm9sLCBhbGxvdyBy
            eCBwb29sPXtzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmxvZywgYWxsb3cg
            eCBwb29sPXtzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmJ1Y2tldHMuaW5k
            ZXgiLAogICAgICAgIH0KCiAgICAgICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRf
            Y2Fwc19hbmRfZW50aXR5KHNlbGYsIHVzZXJfbmFtZSk6CiAgICAgICAgaWYgImNsaWVudC5jc2kt
            Y2VwaGZzLXByb3Zpc2lvbmVyIiBpbiB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQu
            Y3NpLWNlcGhmcy1wcm92aXNpb25lciIgIT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAg
            ICAgcmV0dXJuIHNlbGYuZ2V0X2NlcGhmc19wcm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoKQog
            ICAgICAgIGlmICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiBpbiB1c2VyX25hbWU6CiAgICAgICAg
            ICAgIGlmICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiAhPSB1c2VyX25hbWU6CiAgICAgICAgICAg
            ICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQog
            ICAgICAgICAgICByZXR1cm4gc2VsZi5nZXRfY2VwaGZzX25vZGVfY2Fwc19hbmRfZW50aXR5KCkK
            ICAgICAgICBpZiAiY2xpZW50LmNzaS1yYmQtcHJvdmlzaW9uZXIiIGluIHVzZXJfbmFtZToKICAg
            ICAgICAgICAgaWYgImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiAhPSB1c2VyX25hbWU6CiAg
            ICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9u
            ID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5nZXRfcmJkX3Byb3Zpc2lvbmVyX2NhcHNf
            YW5kX2VudGl0eSgpCiAgICAgICAgaWYgImNsaWVudC5jc2ktcmJkLW5vZGUiIGluIHVzZXJfbmFt
            ZToKICAgICAgICAgICAgaWYgImNsaWVudC5jc2ktcmJkLW5vZGUiICE9IHVzZXJfbmFtZToKICAg
            ICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24g
            PSBUcnVlCiAgICAgICAgICAgIHJldHVybiBzZWxmLmdldF9yYmRfbm9kZV9jYXBzX2FuZF9lbnRp
            dHkoKQogICAgICAgIGlmICJjbGllbnQuaGVhbHRoY2hlY2tlciIgaW4gdXNlcl9uYW1lOgogICAg
            ICAgICAgICBpZiAiY2xpZW50LmhlYWx0aGNoZWNrZXIiICE9IHVzZXJfbmFtZToKICAgICAgICAg
            ICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVl
            CiAgICAgICAgICAgIHJldHVybiBzZWxmLmdldF9kZWZhdWx0VXNlcl9jYXBzX2FuZF9lbnRpdHko
            KQoKICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICBm
            Im5vIHVzZXIgZm91bmQgd2l0aCB1c2VyX25hbWU6IHt1c2VyX25hbWV9LCAiCiAgICAgICAgICAg
            ICJnZXRfY2Fwc19hbmRfZW50aXR5IGNvbW1hbmQgZmFpbGVkLlxuIgogICAgICAgICkKCiAgICBk
            ZWYgY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3VzZXIoc2VsZiwgdXNlcik6CiAgICAgICAgIiIiCiAg
            ICAgICAgY29tbWFuZDogY2VwaCBhdXRoIGdldC1vci1jcmVhdGUgY2xpZW50LmNzaS1jZXBoZnMt
            cHJvdmlzaW9uZXIgbW9uICdhbGxvdyByJyBtZ3IgJ2FsbG93IHJ3JyBvc2QgJ2FsbG93IHJ3IHRh
            ZyBjZXBoZnMgbWV0YWRhdGE9KicKICAgICAgICAiIiIKICAgICAgICBjYXBzLCBlbnRpdHkgPSBz
            ZWxmLmdldF9jYXBzX2FuZF9lbnRpdHkodXNlcikKICAgICAgICBjbWRfanNvbiA9IHsKICAgICAg
            ICAgICAgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUiLAogICAgICAgICAgICAiZW50aXR5
            IjogZW50aXR5LAogICAgICAgICAgICAiY2FwcyI6IFtjYXAgZm9yIGNhcF9saXN0IGluIGxpc3Qo
            Y2Fwcy5pdGVtcygpKSBmb3IgY2FwIGluIGNhcF9saXN0XSwKICAgICAgICAgICAgImZvcm1hdCI6
            ICJqc29uIiwKICAgICAgICB9CgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoK
            ICAgICAgICAgICAgcmV0dXJuICgKICAgICAgICAgICAgICAgIHNlbGYuZHJ5X3J1bigKICAgICAg
            ICAgICAgICAgICAgICAiY2VwaCAiCiAgICAgICAgICAgICAgICAgICAgKyBjbWRfanNvblsicHJl
            Zml4Il0KICAgICAgICAgICAgICAgICAgICArICIgIgogICAgICAgICAgICAgICAgICAgICsgY21k
            X2pzb25bImVudGl0eSJdCiAgICAgICAgICAgICAgICAgICAgKyAiICIKICAgICAgICAgICAgICAg
            ICAgICArICIgIi5qb2luKGNtZF9qc29uWyJjYXBzIl0pCiAgICAgICAgICAgICAgICApLAogICAg
            ICAgICAgICAgICAgIiIsCiAgICAgICAgICAgICkKICAgICAgICAjIGNoZWNrIGlmIHVzZXIgYWxy
            ZWFkeSBleGlzdAogICAgICAgIHVzZXJfa2V5ID0gc2VsZi5jaGVja191c2VyX2V4aXN0KGVudGl0
            eSkKICAgICAgICBpZiB1c2VyX2tleSAhPSAiIjoKICAgICAgICAgICAgcmV0dXJuIHVzZXJfa2V5
            LCBmIntlbnRpdHkuc3BsaXQoJy4nLCAxKVsxXX0iCiAgICAgICAgICAgICMgZW50aXR5LnNwbGl0
            KCcuJywxKVsxXSB0byByZW5hbWUgZW50aXR5KGNsaWVudC5jc2ktcmJkLW5vZGUpIGFzIGNzaS1y
            YmQtbm9kZQoKICAgICAgICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1v
            bl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nl
            c3NmdWwgYXR0ZW1wdCwKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9
            PSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAg
            ICAgICAgICAgZiInYXV0aCBnZXQtb3ItY3JlYXRlIHt1c2VyfScgY29tbWFuZCBmYWlsZWQuXG4i
            CiAgICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBz
            ZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgKQogICAgICAgIHJldHVybiBzdHIo
            anNvbl9vdXRbMF1bImtleSJdKSwgZiJ7ZW50aXR5LnNwbGl0KCcuJywgMSlbMV19IgogICAgICAg
            ICMgZW50aXR5LnNwbGl0KCcuJywxKVsxXSB0byByZW5hbWUgZW50aXR5KGNsaWVudC5jc2ktcmJk
            LW5vZGUpIGFzIGNzaS1yYmQtbm9kZQoKICAgIGRlZiBnZXRfY2VwaGZzX2RhdGFfcG9vbF9kZXRh
            aWxzKHNlbGYpOgogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAiZnMgbHMiLCAiZm9ybWF0
            IjogImpzb24ifQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAg
            ICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigiY2VwaCAiICsgY21kX2pzb25bInByZWZpeCJdKQogICAg
            ICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dl
            bihjbWRfanNvbikKICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0
            LCByZXBvcnQgYW4gZXJyb3IKICAgICAgICBpZiByZXRfdmFsICE9IDA6CiAgICAgICAgICAgICMg
            aWYgZnMgYW5kIGRhdGFfcG9vbCBhcmd1bWVudHMgYXJlIG5vdCBzZXQsIHNpbGVudGx5IHJldHVy
            bgogICAgICAgICAgICBpZiAoCiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhm
            c19maWxlc3lzdGVtX25hbWUgPT0gIiIKICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFy
            c2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZSA9PSAiIgogICAgICAgICAgICApOgogICAgICAgICAg
            ICAgICAgcmV0dXJuCiAgICAgICAgICAgICMgaWYgdXNlciBoYXMgcHJvdmlkZWQgYW55IG9mIHRo
            ZQogICAgICAgICAgICAjICctLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUnIG9yICctLWNlcGhmcy1k
            YXRhLXBvb2wtbmFtZScgYXJndW1lbnRzLAogICAgICAgICAgICAjIHJhaXNlIGFuIGV4Y2VwdGlv
            biBhcyB3ZSBhcmUgdW5hYmxlIHRvIHZlcmlmeSB0aGUgYXJncwogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiInZnMgbHMnIGNlcGgg
            Y2FsbCBmYWlsZWQgd2l0aCBlcnJvcjoge2Vycl9tc2d9IgogICAgICAgICAgICApCgogICAgICAg
            IG1hdGNoaW5nX2pzb25fb3V0ID0ge30KICAgICAgICAjIGlmICctLWNlcGhmcy1maWxlc3lzdGVt
            LW5hbWUnIGFyZ3VtZW50IGlzIHByb3ZpZGVkLAogICAgICAgICMgY2hlY2sgd2hldGhlciB0aGUg
            cHJvdmlkZWQgZmlsZXN5c3RlbS1uYW1lIGV4aXN0cyBvciBub3QKICAgICAgICBpZiBzZWxmLl9h
            cmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWU6CiAgICAgICAgICAgICMgZ2V0IHRoZSBt
            YXRjaGluZyBsaXN0CiAgICAgICAgICAgIG1hdGNoaW5nX2pzb25fb3V0X2xpc3QgPSBbCiAgICAg
            ICAgICAgICAgICBtYXRjaGVkCiAgICAgICAgICAgICAgICBmb3IgbWF0Y2hlZCBpbiBqc29uX291
            dAogICAgICAgICAgICAgICAgaWYgc3RyKG1hdGNoZWRbIm5hbWUiXSkgPT0gc2VsZi5fYXJnX3Bh
            cnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lCiAgICAgICAgICAgIF0KICAgICAgICAgICAgIyB1
            bmFibGUgdG8gZmluZCBhIG1hdGNoaW5nIGZzLW5hbWUsIHJhaXNlIGFuIGVycm9yCiAgICAgICAg
            ICAgIGlmIGxlbihtYXRjaGluZ19qc29uX291dF9saXN0KSA9PSAwOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBmIkZp
            bGVzeXN0ZW0gcHJvdmlkZWQsICd7c2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9u
            YW1lfScsICIKICAgICAgICAgICAgICAgICAgICBmImlzIG5vdCBmb3VuZCBpbiB0aGUgZnMtbGlz
            dDoge1tzdHIoeFsnbmFtZSddKSBmb3IgeCBpbiBqc29uX291dF19IgogICAgICAgICAgICAgICAg
            KQogICAgICAgICAgICBtYXRjaGluZ19qc29uX291dCA9IG1hdGNoaW5nX2pzb25fb3V0X2xpc3Rb
            MF0KICAgICAgICAjIGlmIGNlcGhmcyBmaWxlc3lzdGVtIG5hbWUgaXMgbm90IHByb3ZpZGVkLAog
            ICAgICAgICMgdHJ5IHRvIGdldCBhIGRlZmF1bHQgZnMgbmFtZSBieSBkb2luZyB0aGUgZm9sbG93
            aW5nCiAgICAgICAgZWxzZToKICAgICAgICAgICAgIyBhLiBjaGVjayBpZiB0aGVyZSBpcyBvbmx5
            IG9uZSBmaWxlc3lzdGVtIGlzIHByZXNlbnQKICAgICAgICAgICAgaWYgbGVuKGpzb25fb3V0KSA9
            PSAxOgogICAgICAgICAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXQgPSBqc29uX291dFswXQogICAg
            ICAgICAgICAjIGIuIG9yIGVsc2UsIGNoZWNrIGlmIGRhdGFfcG9vbCBuYW1lIGlzIHByb3ZpZGVk
            CiAgICAgICAgICAgIGVsaWYgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWU6
            CiAgICAgICAgICAgICAgICAjIGFuZCBpZiBwcmVzZW50LCBjaGVjayB3aGV0aGVyIHRoZXJlIGV4
            aXN0cyBhIGZzIHdoaWNoIGhhcyB0aGUgZGF0YV9wb29sCiAgICAgICAgICAgICAgICBmb3IgZWFj
            aEogaW4ganNvbl9vdXQ6CiAgICAgICAgICAgICAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5j
            ZXBoZnNfZGF0YV9wb29sX25hbWUgaW4gZWFjaEpbImRhdGFfcG9vbHMiXToKICAgICAgICAgICAg
            ICAgICAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXQgPSBlYWNoSgogICAgICAgICAgICAgICAgICAg
            ICAgICBicmVhawogICAgICAgICAgICAgICAgIyBpZiB0aGVyZSBpcyBubyBtYXRjaGluZyBmcyBl
            eGlzdHMsIHRoYXQgbWVhbnMgcHJvdmlkZWQgZGF0YV9wb29sIG5hbWUgaXMgaW52YWxpZAogICAg
            ICAgICAgICAgICAgaWYgbm90IG1hdGNoaW5nX2pzb25fb3V0OgogICAgICAgICAgICAgICAgICAg
            IHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgICAg
            IGYiUHJvdmlkZWQgZGF0YV9wb29sIG5hbWUsIHtzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRh
            X3Bvb2xfbmFtZX0sIgogICAgICAgICAgICAgICAgICAgICAgICAiIGRvZXMgbm90IGV4aXN0cyIK
            ICAgICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICMgYy4gaWYgbm90aGluZyBpcyBzZXQg
            YW5kIGNvdWxkbid0IGZpbmQgYSBkZWZhdWx0LAogICAgICAgICAgICBlbHNlOgogICAgICAgICAg
            ICAgICAgIyBqdXN0IHJldHVybiBzaWxlbnRseQogICAgICAgICAgICAgICAgcmV0dXJuCgogICAg
            ICAgIGlmIG1hdGNoaW5nX2pzb25fb3V0OgogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNl
            cGhmc19maWxlc3lzdGVtX25hbWUgPSBzdHIobWF0Y2hpbmdfanNvbl9vdXRbIm5hbWUiXSkKICAg
            ICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfbWV0YWRhdGFfcG9vbF9uYW1lID0gc3Ry
            KAogICAgICAgICAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXRbIm1ldGFkYXRhX3Bvb2wiXQogICAg
            ICAgICAgICApCgogICAgICAgIGlmIGlzaW5zdGFuY2UobWF0Y2hpbmdfanNvbl9vdXRbImRhdGFf
            cG9vbHMiXSwgbGlzdCk6CiAgICAgICAgICAgICMgaWYgdGhlIHVzZXIgaGFzIGFscmVhZHkgcHJv
            dmlkZWQgZGF0YS1wb29sLW5hbWUsCiAgICAgICAgICAgICMgdGhyb3VnaCAtLWNlcGhmcy1kYXRh
            LXBvb2wtbmFtZQogICAgICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bv
            b2xfbmFtZToKICAgICAgICAgICAgICAgICMgaWYgdGhlIHByb3ZpZGVkIG5hbWUgaXMgbm90IG1h
            dGNoaW5nIHdpdGggdGhlIG9uZSBpbiB0aGUgbGlzdAogICAgICAgICAgICAgICAgaWYgKAogICAg
            ICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lCiAg
            ICAgICAgICAgICAgICAgICAgbm90IGluIG1hdGNoaW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0K
            ICAgICAgICAgICAgICAgICk6CiAgICAgICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAgICAgZiJQcm92aWRlZCBkYXRhLXBv
            b2wtbmFtZTogJ3tzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZX0nLCAiCiAg
            ICAgICAgICAgICAgICAgICAgICAgICJkb2Vzbid0IG1hdGNoIGZyb20gdGhlIGRhdGEtcG9vbHMg
            bGlzdDogIgogICAgICAgICAgICAgICAgICAgICAgICBmIntbc3RyKHgpIGZvciB4IGluIG1hdGNo
            aW5nX2pzb25fb3V0WydkYXRhX3Bvb2xzJ11dfSIKICAgICAgICAgICAgICAgICAgICApCiAgICAg
            ICAgICAgICMgaWYgZGF0YV9wb29sIG5hbWUgaXMgbm90IHByb3ZpZGVkLAogICAgICAgICAgICAj
            IHRoZW4gdHJ5IHRvIGZpbmQgYSBkZWZhdWx0IGRhdGEgcG9vbCBuYW1lCiAgICAgICAgICAgIGVs
            c2U6CiAgICAgICAgICAgICAgICAjIGlmIG5vIGRhdGFfcG9vbHMgZXhpc3QsIHNpbGVudGx5IHJl
            dHVybgogICAgICAgICAgICAgICAgaWYgbGVuKG1hdGNoaW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xz
            Il0pID09IDA6CiAgICAgICAgICAgICAgICAgICAgcmV0dXJuCiAgICAgICAgICAgICAgICBzZWxm
            Ll9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZSA9IHN0cigKICAgICAgICAgICAgICAg
            ICAgICBtYXRjaGluZ19qc29uX291dFsiZGF0YV9wb29scyJdWzBdCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgICMgaWYgdGhlcmUgYXJlIG1vcmUgdGhhbiBvbmUgJ2RhdGFfcG9vbHMnIGV4
            aXN0LAogICAgICAgICAgICAjIHRoZW4gd2FybiB0aGUgdXNlciB0aGF0IHdlIGFyZSB1c2luZyB0
            aGUgc2VsZWN0ZWQgbmFtZQogICAgICAgICAgICBpZiBsZW4obWF0Y2hpbmdfanNvbl9vdXRbImRh
            dGFfcG9vbHMiXSkgPiAxOgogICAgICAgICAgICAgICAgcHJpbnQoCiAgICAgICAgICAgICAgICAg
            ICAgIldBUk5JTkc6IE11bHRpcGxlIGRhdGEgcG9vbHMgZGV0ZWN0ZWQ6ICIKICAgICAgICAgICAg
            ICAgICAgICBmIntbc3RyKHgpIGZvciB4IGluIG1hdGNoaW5nX2pzb25fb3V0WydkYXRhX3Bvb2xz
            J11dfVxuIgogICAgICAgICAgICAgICAgICAgIGYiVXNpbmcgdGhlIGRhdGEtcG9vbDogJ3tzZWxm
            Ll9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZX0nXG4iCiAgICAgICAgICAgICAgICAp
            CgogICAgZGVmIGNyZWF0ZV9jaGVja2VyS2V5KHNlbGYsIHVzZXIpOgogICAgICAgIGNhcHMsIGVu
            dGl0eSA9IHNlbGYuZ2V0X2NhcHNfYW5kX2VudGl0eSh1c2VyKQogICAgICAgIGNtZF9qc29uID0g
            ewogICAgICAgICAgICAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0ZSIsCiAgICAgICAgICAg
            ICJlbnRpdHkiOiBlbnRpdHksCiAgICAgICAgICAgICJjYXBzIjogW2NhcCBmb3IgY2FwX2xpc3Qg
            aW4gbGlzdChjYXBzLml0ZW1zKCkpIGZvciBjYXAgaW4gY2FwX2xpc3RdLAogICAgICAgICAgICAi
            Zm9ybWF0IjogImpzb24iLAogICAgICAgIH0KCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5k
            cnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKAogICAgICAgICAgICAgICAg
            ImNlcGggIgogICAgICAgICAgICAgICAgKyBjbWRfanNvblsicHJlZml4Il0KICAgICAgICAgICAg
            ICAgICsgIiAiCiAgICAgICAgICAgICAgICArIGNtZF9qc29uWyJlbnRpdHkiXQogICAgICAgICAg
            ICAgICAgKyAiICIKICAgICAgICAgICAgICAgICsgIiAiLmpvaW4oY21kX2pzb25bImNhcHMiXSkK
            ICAgICAgICAgICAgKQogICAgICAgICMgY2hlY2sgaWYgdXNlciBhbHJlYWR5IGV4aXN0CiAgICAg
            ICAgdXNlcl9rZXkgPSBzZWxmLmNoZWNrX3VzZXJfZXhpc3QoZW50aXR5KQogICAgICAgIGlmIHVz
            ZXJfa2V5ICE9ICIiOgogICAgICAgICAgICByZXR1cm4gdXNlcl9rZXkKCiAgICAgICAgcmV0X3Zh
            bCwganNvbl9vdXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29u
            KQogICAgICAgICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAg
            aWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmFpc2Ug
            RXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgIGYiJ2F1dGggZ2V0LW9y
            LWNyZWF0ZSB7c2VsZi5ydW5fYXNfdXNlcn0nIGNvbW1hbmQgZmFpbGVkXG4iCiAgICAgICAgICAg
            ICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09V
            VFBVVF9MSVNUfSIKICAgICAgICAgICAgKQogICAgICAgIHJldHVybiBzdHIoanNvbl9vdXRbMF1b
            ImtleSJdKQoKICAgIGRlZiBnZXRfY2VwaF9kYXNoYm9hcmRfbGluayhzZWxmKToKICAgICAgICBj
            bWRfanNvbiA9IHsicHJlZml4IjogIm1nciBzZXJ2aWNlcyIsICJmb3JtYXQiOiAianNvbiJ9CiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2Vs
            Zi5kcnlfcnVuKCJjZXBoICIgKyBjbWRfanNvblsicHJlZml4Il0pCiAgICAgICAgcmV0X3ZhbCwg
            anNvbl9vdXQsIF8gPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAg
            ICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3Zh
            bCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmV0dXJuIE5vbmUKICAg
            ICAgICBpZiAiZGFzaGJvYXJkIiBub3QgaW4ganNvbl9vdXQ6CiAgICAgICAgICAgIHJldHVybiBO
            b25lCiAgICAgICAgcmV0dXJuIGpzb25fb3V0WyJkYXNoYm9hcmQiXQoKICAgIGRlZiBjcmVhdGVf
            cmd3X2FkbWluX29wc191c2VyKHNlbGYpOgogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgInJh
            ZG9zZ3ctYWRtaW4iLAogICAgICAgICAgICAidXNlciIsCiAgICAgICAgICAgICJjcmVhdGUiLAog
            ICAgICAgICAgICAiLS11aWQiLAogICAgICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9P
            UFNfVVNFUl9OQU1FLAogICAgICAgICAgICAiLS1kaXNwbGF5LW5hbWUiLAogICAgICAgICAgICAi
            Um9vayBSR1cgQWRtaW4gT3BzIHVzZXIiLAogICAgICAgICAgICAiLS1jYXBzIiwKICAgICAgICAg
            ICAgImJ1Y2tldHM9Kjt1c2Vycz0qO3VzYWdlPXJlYWQ7bWV0YWRhdGE9cmVhZDt6b25lPXJlYWQi
            LAogICAgICAgICAgICAiLS1yZ3ctcmVhbG0iLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd19yZWFsbV9uYW1lLAogICAgICAgICAgICAiLS1yZ3ctem9uZWdyb3VwIiwKICAgICAgICAg
            ICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZWdyb3VwX25hbWUsCiAgICAgICAgICAgICItLXJn
            dy16b25lIiwKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1lLAogICAg
            ICAgIF0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJl
            dHVybiBzZWxmLmRyeV9ydW4oImNlcGggIiArICIgIi5qb2luKGNtZCkpCiAgICAgICAgdHJ5Ogog
            ICAgICAgICAgICBvdXRwdXQgPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1z
            dWJwcm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vy
            cm9yIGFzIGV4ZWNFcnI6CiAgICAgICAgICAgICMgaWYgdGhlIHVzZXIgYWxyZWFkeSBleGlzdHMs
            IHdlIGp1c3QgcXVlcnkgaXQKICAgICAgICAgICAgaWYgZXhlY0Vyci5yZXR1cm5jb2RlID09IGVy
            cm5vLkVFWElTVDoKICAgICAgICAgICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgICAgICAgICAi
            cmFkb3Nndy1hZG1pbiIsCiAgICAgICAgICAgICAgICAgICAgInVzZXIiLAogICAgICAgICAgICAg
            ICAgICAgICJpbmZvIiwKICAgICAgICAgICAgICAgICAgICAiLS11aWQiLAogICAgICAgICAgICAg
            ICAgICAgIHNlbGYuRVhURVJOQUxfUkdXX0FETUlOX09QU19VU0VSX05BTUUsCiAgICAgICAgICAg
            ICAgICAgICAgIi0tcmd3LXJlYWxtIiwKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFy
            c2VyLnJnd19yZWFsbV9uYW1lLAogICAgICAgICAgICAgICAgICAgICItLXJndy16b25lZ3JvdXAi
            LAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1l
            LAogICAgICAgICAgICAgICAgICAgICItLXJndy16b25lIiwKICAgICAgICAgICAgICAgICAgICBz
            ZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUsCiAgICAgICAgICAgICAgICBdCiAgICAgICAg
            ICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVj
            a19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAgICAgICAgZXhj
            ZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yIGFzIGV4ZWNFcnI6CiAgICAgICAgICAg
            ICAgICAgICAgZXJyX21zZyA9ICgKICAgICAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8g
            ZXhlY3V0ZSBjb21tYW5kIHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAg
            ICAgICAgICAgICAgICAgICAgZiJDb2RlOiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtl
            eGVjRXJyLnN0ZGVycn0iCiAgICAgICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAg
            IHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykKICAgICAgICAgICAgICAgICAgICByZXR1cm4gTm9u
            ZSwgTm9uZSwgRmFsc2UsICItMSIKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVy
            cl9tc2cgPSAoCiAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5k
            IHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICBm
            IkNvZGU6IHtleGVjRXJyLnJldHVybmNvZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykK
            ICAgICAgICAgICAgICAgIHJldHVybiBOb25lLCBOb25lLCBGYWxzZSwgIi0xIgoKICAgICAgICAj
            IGlmIGl0IGlzIHB5dGhvbjIsIGRvbid0IGNoZWNrIGZvciBjZXBoIHZlcnNpb24gZm9yIGFkZGlu
            ZyBgaW5mbz1yZWFkYCBjYXAocmd3X3ZhbGlkYXRpb24pCiAgICAgICAgaWYgc3lzLnZlcnNpb25f
            aW5mby5tYWpvciA8IDM6CiAgICAgICAgICAgIGpzb25vdXRwdXQgPSBqc29uLmxvYWRzKG91dHB1
            dCkKICAgICAgICAgICAgcmV0dXJuICgKICAgICAgICAgICAgICAgIGpzb25vdXRwdXRbImtleXMi
            XVswXVsiYWNjZXNzX2tleSJdLAogICAgICAgICAgICAgICAganNvbm91dHB1dFsia2V5cyJdWzBd
            WyJzZWNyZXRfa2V5Il0sCiAgICAgICAgICAgICAgICBGYWxzZSwKICAgICAgICAgICAgICAgICIi
            LAogICAgICAgICAgICApCgogICAgICAgICMgc2VwYXJhdGVseSBhZGQgaW5mbz1yZWFkIGNhcHMg
            Zm9yIHJndy1lbmRwb2ludCBpcCB2YWxpZGF0aW9uCiAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVk
            ID0gVHJ1ZQogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAg
            ICAgICAgICAiY2FwcyIsCiAgICAgICAgICAgICJhZGQiLAogICAgICAgICAgICAiLS11aWQiLAog
            ICAgICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9PUFNfVVNFUl9OQU1FLAogICAgICAg
            ICAgICAiLS1jYXBzIiwKICAgICAgICAgICAgImluZm89cmVhZCIsCiAgICAgICAgICAgICItLXJn
            dy1yZWFsbSIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUsCiAg
            ICAgICAgICAgICItLXJndy16b25lZ3JvdXAiLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd196b25lZ3JvdXBfbmFtZSwKICAgICAgICAgICAgIi0tcmd3LXpvbmUiLAogICAgICAgICAg
            ICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUsCiAgICAgICAgXQogICAgICAgIHRyeToK
            ICAgICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9
            c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2VwdCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NF
            cnJvciBhcyBleGVjRXJyOgogICAgICAgICAgICAjIGlmIHRoZSBjZXBoIHZlcnNpb24gbm90IHN1
            cHBvcnRlZCBmb3IgYWRkaW5nIGBpbmZvPXJlYWRgIGNhcChyZ3dfdmFsaWRhdGlvbikKICAgICAg
            ICAgICAgaWYgKAogICAgICAgICAgICAgICAgImNvdWxkIG5vdCBhZGQgY2FwczogdW5hYmxlIHRv
            IGFkZCBjYXBzOiBpbmZvPXJlYWRcbiIKICAgICAgICAgICAgICAgIGluIGV4ZWNFcnIuc3RkZXJy
            LmRlY29kZSgidXRmLTgiKQogICAgICAgICAgICAgICAgYW5kIGV4ZWNFcnIucmV0dXJuY29kZSA9
            PSAyNDQKICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgIGluZm9fY2FwX3N1cHBvcnRlZCA9
            IEZhbHNlCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICBlcnJfbXNnID0gKAogICAg
            ICAgICAgICAgICAgICAgIGYiZmFpbGVkIHRvIGV4ZWN1dGUgY29tbWFuZCB7Y21kfS4gT3V0cHV0
            OiB7ZXhlY0Vyci5vdXRwdXR9LiAiCiAgICAgICAgICAgICAgICAgICAgZiJDb2RlOiB7ZXhlY0Vy
            ci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtleGVjRXJyLnN0ZGVycn0iCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKGVycl9tc2cpCiAgICAgICAgICAgICAg
            ICByZXR1cm4gTm9uZSwgTm9uZSwgRmFsc2UsICItMSIKCiAgICAgICAganNvbm91dHB1dCA9IGpz
            b24ubG9hZHMob3V0cHV0KQogICAgICAgIHJldHVybiAoCiAgICAgICAgICAgIGpzb25vdXRwdXRb
            ImtleXMiXVswXVsiYWNjZXNzX2tleSJdLAogICAgICAgICAgICBqc29ub3V0cHV0WyJrZXlzIl1b
            MF1bInNlY3JldF9rZXkiXSwKICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkLAogICAgICAg
            ICAgICAiIiwKICAgICAgICApCgogICAgZGVmIHZhbGlkYXRlX3JiZF9wb29sKHNlbGYpOgogICAg
            ICAgIGlmIG5vdCBzZWxmLmNsdXN0ZXIucG9vbF9leGlzdHMoc2VsZi5fYXJnX3BhcnNlci5yYmRf
            ZGF0YV9wb29sX25hbWUpOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0
            aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcG9vbCwgJ3tzZWxmLl9hcmdfcGFy
            c2VyLnJiZF9kYXRhX3Bvb2xfbmFtZX0nLCBkb2VzIG5vdCBleGlzdCIKICAgICAgICAgICAgKQoK
            ICAgIGRlZiBpbml0X3JiZF9wb29sKHNlbGYpOgogICAgICAgIGlmIGlzaW5zdGFuY2Uoc2VsZi5j
            bHVzdGVyLCBEdW1teVJhZG9zKToKICAgICAgICAgICAgcmV0dXJuCiAgICAgICAgcmJkX3Bvb2xf
            bmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgaW9jdHgg
            PSBzZWxmLmNsdXN0ZXIub3Blbl9pb2N0eChyYmRfcG9vbF9uYW1lKQogICAgICAgIHJiZF9pbnN0
            ID0gcmJkLlJCRCgpCiAgICAgICAgcmJkX2luc3QucG9vbF9pbml0KGlvY3R4LCBUcnVlKQoKICAg
            IGRlZiB2YWxpZGF0ZV9yYWRvc19uYW1lc3BhY2Uoc2VsZik6CiAgICAgICAgcmJkX3Bvb2xfbmFt
            ZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgcmFkb3NfbmFt
            ZXNwYWNlID0gc2VsZi5fYXJnX3BhcnNlci5yYWRvc19uYW1lc3BhY2UKICAgICAgICBpZiByYWRv
            c19uYW1lc3BhY2UgPT0gIiI6CiAgICAgICAgICAgIHJldHVybgogICAgICAgIGlmIHJhZG9zX25h
            bWVzcGFjZS5pc2xvd2VyKCkgPT0gRmFsc2U6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCByYWRvcyBOYW1l
            c3BhY2UsICd7cmFkb3NfbmFtZXNwYWNlfScsICIKICAgICAgICAgICAgICAgIGYiY29udGFpbnMg
            dXBwZXIgY2FzZSIKICAgICAgICAgICAgKQogICAgICAgIHJiZF9pbnN0ID0gcmJkLlJCRCgpCiAg
            ICAgICAgaW9jdHggPSBzZWxmLmNsdXN0ZXIub3Blbl9pb2N0eChyYmRfcG9vbF9uYW1lKQogICAg
            ICAgIGlmIHJiZF9pbnN0Lm5hbWVzcGFjZV9leGlzdHMoaW9jdHgsIHJhZG9zX25hbWVzcGFjZSkg
            aXMgRmFsc2U6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAg
            ICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCByYWRvcyBOYW1lc3BhY2UsICd7cmFkb3NfbmFt
            ZXNwYWNlfScsICIKICAgICAgICAgICAgICAgIGYiaXMgbm90IGZvdW5kIGluIHRoZSBwb29sICd7
            cmJkX3Bvb2xfbmFtZX0nIgogICAgICAgICAgICApCgogICAgZGVmIGdldF9vcl9jcmVhdGVfc3Vi
            dm9sdW1lX2dyb3VwKHNlbGYsIHN1YnZvbHVtZV9ncm91cCwgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFt
            ZSk6CiAgICAgICAgY21kID0gWwogICAgICAgICAgICAiY2VwaCIsCiAgICAgICAgICAgICJmcyIs
            CiAgICAgICAgICAgICJzdWJ2b2x1bWVncm91cCIsCiAgICAgICAgICAgICJnZXRwYXRoIiwKICAg
            ICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAgICAgICAgICAgc3Vidm9sdW1lX2dy
            b3VwLAogICAgICAgIF0KICAgICAgICB0cnk6CiAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNo
            ZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1
            YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICBjbWQgPSBbCiAgICAgICAg
            ICAgICAgICAiY2VwaCIsCiAgICAgICAgICAgICAgICAiZnMiLAogICAgICAgICAgICAgICAgInN1
            YnZvbHVtZWdyb3VwIiwKICAgICAgICAgICAgICAgICJjcmVhdGUiLAogICAgICAgICAgICAgICAg
            Y2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAgICAgICAgICAgICAgIHN1YnZvbHVtZV9ncm91cCwK
            ICAgICAgICAgICAgXQogICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBfID0gc3VicHJv
            Y2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAg
            ICBleGNlcHQgc3VicHJvY2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3I6CiAgICAgICAgICAgICAgICBy
            YWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYic3Vi
            dm9sdW1lIGdyb3VwIHtzdWJ2b2x1bWVfZ3JvdXB9IGlzIG5vdCBhYmxlIHRvIGdldCBjcmVhdGVk
            IgogICAgICAgICAgICAgICAgKQoKICAgIGRlZiBwaW5fc3Vidm9sdW1lKAogICAgICAgIHNlbGYs
            IHN1YnZvbHVtZV9ncm91cCwgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwgcGluX3R5cGUsIHBpbl9z
            ZXR0aW5nCiAgICApOgogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgImNlcGgiLAogICAgICAg
            ICAgICAiZnMiLAogICAgICAgICAgICAic3Vidm9sdW1lZ3JvdXAiLAogICAgICAgICAgICAicGlu
            IiwKICAgICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAgICAgICAgICAgc3Vidm9s
            dW1lX2dyb3VwLAogICAgICAgICAgICBwaW5fdHlwZSwKICAgICAgICAgICAgcGluX3NldHRpbmcs
            CiAgICAgICAgXQogICAgICAgIHRyeToKICAgICAgICAgICAgXyA9IHN1YnByb2Nlc3MuY2hlY2tf
            b3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAgICAgICBleGNlcHQgc3VicHJv
            Y2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3I6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1
            cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmInN1YnZvbHVtZSBncm91cCB7c3Vidm9sdW1l
            X2dyb3VwfSBpcyBub3QgYWJsZSB0byBnZXQgcGlubmVkIgogICAgICAgICAgICApCgogICAgZGVm
            IGdldF9yZ3dfZnNpZChzZWxmLCBiYXNlX3VybCwgdmVyaWZ5KToKICAgICAgICBhY2Nlc3Nfa2V5
            ID0gc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJfQUNDRVNTX0tFWSJdCiAgICAgICAg
            c2VjcmV0X2tleSA9IHNlbGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX1NFQ1JFVF9LRVki
            XQogICAgICAgIHJnd19lbmRwb2ludCA9IHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50CiAg
            ICAgICAgYmFzZV91cmwgPSBiYXNlX3VybCArICI6Ly8iICsgcmd3X2VuZHBvaW50ICsgIi9hZG1p
            bi9pbmZvPyIKICAgICAgICBwYXJhbXMgPSB7ImZvcm1hdCI6ICJqc29uIn0KICAgICAgICByZXF1
            ZXN0X3VybCA9IGJhc2VfdXJsICsgdXJsZW5jb2RlKHBhcmFtcykKCiAgICAgICAgdHJ5OgogICAg
            ICAgICAgICByID0gcmVxdWVzdHMuZ2V0KAogICAgICAgICAgICAgICAgcmVxdWVzdF91cmwsCiAg
            ICAgICAgICAgICAgICBhdXRoPVMzQXV0aChhY2Nlc3Nfa2V5LCBzZWNyZXRfa2V5LCByZ3dfZW5k
            cG9pbnQpLAogICAgICAgICAgICAgICAgdmVyaWZ5PXZlcmlmeSwKICAgICAgICAgICAgKQogICAg
            ICAgIGV4Y2VwdCByZXF1ZXN0cy5leGNlcHRpb25zLlRpbWVvdXQ6CiAgICAgICAgICAgIHN5cy5z
            dGRlcnIud3JpdGUoCiAgICAgICAgICAgICAgICBmImludmFsaWQgZW5kcG9pbnQ6LCBub3QgYWJs
            ZSB0byBjYWxsIGFkbWluLW9wcyBhcGl7cmd3X2VuZHBvaW50fSIKICAgICAgICAgICAgKQogICAg
            ICAgICAgICByZXR1cm4gIiIsICItMSIKICAgICAgICByMSA9IHIuanNvbigpCiAgICAgICAgaWYg
            cjEgaXMgTm9uZSBvciByMS5nZXQoImluZm8iKSBpcyBOb25lOgogICAgICAgICAgICBzeXMuc3Rk
            ZXJyLndyaXRlKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmd3IEVuZHBvaW50LCAn
            e3NlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50fScsIGlzIGludmFsaWQuIgogICAgICAgICAg
            ICApCiAgICAgICAgICAgIHJldHVybiAoCiAgICAgICAgICAgICAgICAiIiwKICAgICAgICAgICAg
            ICAgICItMSIsCiAgICAgICAgICAgICkKCiAgICAgICAgcmV0dXJuIHIxWyJpbmZvIl1bInN0b3Jh
            Z2VfYmFja2VuZHMiXVswXVsiY2x1c3Rlcl9pZCJdLCAiIgoKICAgIGRlZiB2YWxpZGF0ZV9yZ3df
            ZW5kcG9pbnQoc2VsZiwgaW5mb19jYXBfc3VwcG9ydGVkKToKICAgICAgICAjIGlmIHRoZSAnY2x1
            c3RlcicgaW5zdGFuY2UgaXMgYSBkdW1teSBvbmUsCiAgICAgICAgIyBkb24ndCB0cnkgdG8gcmVh
            Y2ggb3V0IHRvIHRoZSBlbmRwb2ludAogICAgICAgIGlmIGlzaW5zdGFuY2Uoc2VsZi5jbHVzdGVy
            LCBEdW1teVJhZG9zKToKICAgICAgICAgICAgcmV0dXJuCgogICAgICAgIHJnd19lbmRwb2ludCA9
            IHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50CgogICAgICAgICMgdmFsaWRhdGUgcmd3IGVu
            ZHBvaW50IG9ubHkgaWYgaXAgYWRkcmVzcyBpcyBwYXNzZWQKICAgICAgICBpcF90eXBlID0gc2Vs
            Zi5faW52YWxpZF9lbmRwb2ludChyZ3dfZW5kcG9pbnQpCgogICAgICAgICMgY2hlY2sgaWYgdGhl
            IHJndyBlbmRwb2ludCBpcyByZWFjaGFibGUKICAgICAgICBjZXJ0ID0gTm9uZQogICAgICAgIGlm
            IG5vdCBzZWxmLl9hcmdfcGFyc2VyLnJnd19za2lwX3RscyBhbmQgc2VsZi52YWxpZGF0ZV9yZ3df
            ZW5kcG9pbnRfdGxzX2NlcnQoKToKICAgICAgICAgICAgY2VydCA9IHNlbGYuX2FyZ19wYXJzZXIu
            cmd3X3Rsc19jZXJ0X3BhdGgKICAgICAgICBiYXNlX3VybCwgdmVyaWZ5LCBlcnIgPSBzZWxmLmVu
            ZHBvaW50X2RpYWwocmd3X2VuZHBvaW50LCBpcF90eXBlLCBjZXJ0PWNlcnQpCiAgICAgICAgaWYg
            ZXJyICE9ICIiOgogICAgICAgICAgICByZXR1cm4gIi0xIgoKICAgICAgICAjIGNoZWNrIGlmIHRo
            ZSByZ3cgZW5kcG9pbnQgYmVsb25ncyB0byB0aGUgc2FtZSBjbHVzdGVyCiAgICAgICAgIyBvbmx5
            IGNoZWNrIGlmIGBpbmZvYCBjYXAgaXMgc3VwcG9ydGVkCiAgICAgICAgaWYgaW5mb19jYXBfc3Vw
            cG9ydGVkOgogICAgICAgICAgICBmc2lkID0gc2VsZi5nZXRfZnNpZCgpCiAgICAgICAgICAgIHJn
            d19mc2lkLCBlcnIgPSBzZWxmLmdldF9yZ3dfZnNpZChiYXNlX3VybCwgdmVyaWZ5KQogICAgICAg
            ICAgICBpZiBlcnIgPT0gIi0xIjoKICAgICAgICAgICAgICAgIHJldHVybiAiLTEiCiAgICAgICAg
            ICAgIGlmIGZzaWQgIT0gcmd3X2ZzaWQ6CiAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRl
            KAogICAgICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHJndyBFbmRwb2ludCwgJ3tzZWxm
            Ll9hcmdfcGFyc2VyLnJnd19lbmRwb2ludH0nLCBpcyBpbnZhbGlkLiBXZSBhcmUgdmFsaWRhdGlu
            ZyBieSBjYWxsaW5nIHRoZSBhZG1pbm9wcyBhcGkgdGhyb3VnaCByZ3ctZW5kcG9pbnQgYW5kIHZh
            bGlkYXRpbmcgdGhlIGNsdXN0ZXJfaWQgJ3tyZ3dfZnNpZH0nIGlzIGVxdWFsIHRvIHRoZSBjZXBo
            IGNsdXN0ZXIgZnNpZCAne2ZzaWR9JyIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAg
            IHJldHVybiAiLTEiCgogICAgICAgICMgY2hlY2sgaWYgdGhlIHJndyBlbmRwb2ludCBwb29sIGV4
            aXN0CiAgICAgICAgIyBvbmx5IHZhbGlkYXRlIGlmIHJnd19wb29sX3ByZWZpeCBpcyBwYXNzZWQg
            ZWxzZSBpdCB3aWxsIHRha2UgZGVmYXVsdCB2YWx1ZSBhbmQgd2UgZG9uJ3QgY3JlYXRlIHRoZXNl
            IGRlZmF1bHQgcG9vbHMKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZp
            eCAhPSAiZGVmYXVsdCI6CiAgICAgICAgICAgIHJnd19wb29sc190b192YWxpZGF0ZSA9IFsKICAg
            ICAgICAgICAgICAgIGYie3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cubWV0
            YSIsCiAgICAgICAgICAgICAgICAiLnJndy5yb290IiwKICAgICAgICAgICAgICAgIGYie3NlbGYu
            X2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cuY29udHJvbCIsCiAgICAgICAgICAgICAg
            ICBmIntzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmxvZyIsCiAgICAgICAg
            ICAgIF0KICAgICAgICAgICAgZm9yIF9yZ3dfcG9vbF90b192YWxpZGF0ZSBpbiByZ3dfcG9vbHNf
            dG9fdmFsaWRhdGU6CiAgICAgICAgICAgICAgICBpZiBub3Qgc2VsZi5jbHVzdGVyLnBvb2xfZXhp
            c3RzKF9yZ3dfcG9vbF90b192YWxpZGF0ZSk6CiAgICAgICAgICAgICAgICAgICAgc3lzLnN0ZGVy
            ci53cml0ZSgKICAgICAgICAgICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcG9vbCwgJ3tf
            cmd3X3Bvb2xfdG9fdmFsaWRhdGV9JywgZG9lcyBub3QgZXhpc3QiCiAgICAgICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICAgICAgICAgIHJldHVybiAiLTEiCgogICAgICAgIHJldHVybiAiIgoK
            ICAgIGRlZiB2YWxpZGF0ZV9yZ3dfbXVsdGlzaXRlKHNlbGYsIHJnd19tdWx0aXNpdGVfY29uZmln
            X25hbWUsIHJnd19tdWx0aXNpdGVfY29uZmlnKToKICAgICAgICBpZiByZ3dfbXVsdGlzaXRlX2Nv
            bmZpZyAhPSAiIjoKICAgICAgICAgICAgY21kID0gWwogICAgICAgICAgICAgICAgInJhZG9zZ3ct
            YWRtaW4iLAogICAgICAgICAgICAgICAgcmd3X211bHRpc2l0ZV9jb25maWcsCiAgICAgICAgICAg
            ICAgICAiZ2V0IiwKICAgICAgICAgICAgICAgICItLXJndy0iICsgcmd3X211bHRpc2l0ZV9jb25m
            aWcsCiAgICAgICAgICAgICAgICByZ3dfbXVsdGlzaXRlX2NvbmZpZ19uYW1lLAogICAgICAgICAg
            ICBdCiAgICAgICAgICAgIHRyeToKICAgICAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNoZWNr
            X291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgICAgIGV4Y2VwdCBz
            dWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJvciBhcyBleGVjRXJyOgogICAgICAgICAgICAgICAg
            ZXJyX21zZyA9ICgKICAgICAgICAgICAgICAgICAgICBmImZhaWxlZCB0byBleGVjdXRlIGNvbW1h
            bmQge2NtZH0uIE91dHB1dDoge2V4ZWNFcnIub3V0cHV0fS4gIgogICAgICAgICAgICAgICAgICAg
            IGYiQ29kZToge2V4ZWNFcnIucmV0dXJuY29kZX0uIEVycm9yOiB7ZXhlY0Vyci5zdGRlcnJ9Igog
            ICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZShlcnJfbXNn
            KQogICAgICAgICAgICAgICAgcmV0dXJuICItMSIKICAgICAgICByZXR1cm4gIiIKCiAgICBkZWYg
            X2dlbl9vdXRwdXRfbWFwKHNlbGYpOgogICAgICAgIGlmIHNlbGYub3V0X21hcDoKICAgICAgICAg
            ICAgcmV0dXJuCiAgICAgICAgIyBzdXBwb3J0IGxlZ2FjeSBmbGFnIHdpdGggdXBncmFkZXMKICAg
            ICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNsdXN0ZXJfbmFtZToKICAgICAgICAgICAgc2VsZi5f
            YXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5jbHVzdGVyX25h
            bWUKICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUgPSAoCiAgICAgICAg
            ICAgIHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZS5sb3dlcigpCiAgICAgICAgKSAg
            IyBhbHdheXMgY29udmVydCBjbHVzdGVyIG5hbWUgdG8gbG93ZXJjYXNlIGNoYXJhY3RlcnMKICAg
            ICAgICBzZWxmLnZhbGlkYXRlX3JiZF9wb29sKCkKICAgICAgICBzZWxmLmluaXRfcmJkX3Bvb2wo
            KQogICAgICAgIHNlbGYudmFsaWRhdGVfcmFkb3NfbmFtZXNwYWNlKCkKICAgICAgICBzZWxmLl9l
            eGNsdWRlZF9rZXlzLmFkZCgiSzhTX0NMVVNURVJfTkFNRSIpCiAgICAgICAgc2VsZi5nZXRfY2Vw
            aGZzX2RhdGFfcG9vbF9kZXRhaWxzKCkKICAgICAgICBzZWxmLm91dF9tYXBbIk5BTUVTUEFDRSJd
            ID0gc2VsZi5fYXJnX3BhcnNlci5uYW1lc3BhY2UKICAgICAgICBzZWxmLm91dF9tYXBbIks4U19D
            TFVTVEVSX05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQogICAgICAg
            IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9GU0lEIl0gPSBzZWxmLmdldF9mc2lkKCkKICAg
            ICAgICBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUk5BTUUiXSA9IHNlbGYucnVuX2Fz
            X3VzZXIKICAgICAgICBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfQ0VQSF9NT05fREFUQSJd
            ID0gc2VsZi5nZXRfY2VwaF9leHRlcm5hbF9tb25fZGF0YSgpCiAgICAgICAgc2VsZi5vdXRfbWFw
            WyJST09LX0VYVEVSTkFMX1VTRVJfU0VDUkVUIl0gPSBzZWxmLmNyZWF0ZV9jaGVja2VyS2V5KAog
            ICAgICAgICAgICAiY2xpZW50LmhlYWx0aGNoZWNrZXIiCiAgICAgICAgKQogICAgICAgIHNlbGYu
            b3V0X21hcFsiUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSyJdID0gc2VsZi5nZXRfY2VwaF9k
            YXNoYm9hcmRfbGluaygpCiAgICAgICAgKAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9S
            QkRfTk9ERV9TRUNSRVQiXSwKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVf
            U0VDUkVUX05BTUUiXSwKICAgICAgICApID0gc2VsZi5jcmVhdGVfY2VwaENTSUtleXJpbmdfdXNl
            cigiY2xpZW50LmNzaS1yYmQtbm9kZSIpCiAgICAgICAgKAogICAgICAgICAgICBzZWxmLm91dF9t
            YXBbIkNTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUIl0sCiAgICAgICAgICAgIHNlbGYub3V0X21h
            cFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICkgPSBzZWxmLmNy
            ZWF0ZV9jZXBoQ1NJS2V5cmluZ191c2VyKCJjbGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIpCiAg
            ICAgICAgc2VsZi5vdXRfbWFwWyJDRVBIRlNfUE9PTF9OQU1FIl0gPSBzZWxmLl9hcmdfcGFyc2Vy
            LmNlcGhmc19kYXRhX3Bvb2xfbmFtZQogICAgICAgIHNlbGYub3V0X21hcFsiQ0VQSEZTX01FVEFE
            QVRBX1BPT0xfTkFNRSJdID0gKAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19t
            ZXRhZGF0YV9wb29sX25hbWUKICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJDRVBIRlNf
            RlNfTkFNRSJdID0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lCiAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJSRVNUUklDVEVEX0FVVEhfUEVSTUlTU0lPTiJdID0gKAogICAgICAg
            ICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uCiAgICAgICAg
            KQogICAgICAgIHNlbGYub3V0X21hcFsiUkFET1NfTkFNRVNQQUNFIl0gPSBzZWxmLl9hcmdfcGFy
            c2VyLnJhZG9zX25hbWVzcGFjZQogICAgICAgIHNlbGYub3V0X21hcFsiU1VCVk9MVU1FX0dST1VQ
            Il0gPSBzZWxmLl9hcmdfcGFyc2VyLnN1YnZvbHVtZV9ncm91cAogICAgICAgIHNlbGYub3V0X21h
            cFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVCJdID0gIiIKICAgICAgICBzZWxmLm91dF9tYXBbIkNT
            SV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUIl0gPSAiIgogICAgICAgICMgY3JlYXRlIENlcGhG
            UyBub2RlIGFuZCBwcm92aXNpb25lciBrZXlyaW5nIG9ubHkgd2hlbiBNRFMgZXhpc3RzCiAgICAg
            ICAgaWYgc2VsZi5vdXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdIGFuZCBzZWxmLm91dF9tYXBbIkNF
            UEhGU19QT09MX05BTUUiXToKICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgc2VsZi5vdXRf
            bWFwWyJDU0lfQ0VQSEZTX05PREVfU0VDUkVUIl0sCiAgICAgICAgICAgICAgICBzZWxmLm91dF9t
            YXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICApID0gc2VsZi5j
            cmVhdGVfY2VwaENTSUtleXJpbmdfdXNlcigiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIpCiAgICAg
            ICAgICAgICgKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9WSVNJ
            T05FUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19Q
            Uk9WSVNJT05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICApID0gc2VsZi5jcmVhdGVfY2Vw
            aENTSUtleXJpbmdfdXNlcigiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiKQogICAgICAg
            ICAgICAjIGNyZWF0ZSB0aGUgZGVmYXVsdCAiY3NpIiBzdWJ2b2x1bWVncm91cAogICAgICAgICAg
            ICBzZWxmLmdldF9vcl9jcmVhdGVfc3Vidm9sdW1lX2dyb3VwKAogICAgICAgICAgICAgICAgImNz
            aSIsIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgICAgICAp
            CiAgICAgICAgICAgICMgcGluIHRoZSBkZWZhdWx0ICJjc2kiIHN1YnZvbHVtZWdyb3VwCiAgICAg
            ICAgICAgIHNlbGYucGluX3N1YnZvbHVtZSgKICAgICAgICAgICAgICAgICJjc2kiLCBzZWxmLl9h
            cmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUsICJkaXN0cmlidXRlZCIsICIxIgogICAg
            ICAgICAgICApCiAgICAgICAgICAgIGlmIHNlbGYub3V0X21hcFsiU1VCVk9MVU1FX0dST1VQIl06
            CiAgICAgICAgICAgICAgICBzZWxmLmdldF9vcl9jcmVhdGVfc3Vidm9sdW1lX2dyb3VwKAogICAg
            ICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuc3Vidm9sdW1lX2dyb3VwLAogICAgICAg
            ICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgIHNlbGYucGluX3N1YnZvbHVtZSgKICAgICAg
            ICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnN1YnZvbHVtZV9ncm91cCwKICAgICAgICAg
            ICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAg
            ICAgICAgICAgICAgICAgImRpc3RyaWJ1dGVkIiwKICAgICAgICAgICAgICAgICAgICAiMSIsCiAg
            ICAgICAgICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJSR1dfVExTX0NFUlQiXSA9ICIi
            CiAgICAgICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UIl0gPSAiIgogICAgICAg
            IHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUIl0gPSAiIgogICAgICAgIGlm
            IG5vdCBzZWxmLl9hcmdfcGFyc2VyLnNraXBfbW9uaXRvcmluZ19lbmRwb2ludDoKICAgICAgICAg
            ICAgKAogICAgICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UIl0s
            CiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJd
            LAogICAgICAgICAgICApID0gc2VsZi5nZXRfYWN0aXZlX2FuZF9zdGFuZGJ5X21ncnMoKQogICAg
            ICAgIHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdID0gc2VsZi5fYXJnX3BhcnNlci5yYmRf
            ZGF0YV9wb29sX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBbIlJCRF9NRVRBREFUQV9FQ19QT09M
            X05BTUUiXSA9ICgKICAgICAgICAgICAgc2VsZi52YWxpZGF0ZV9yYmRfbWV0YWRhdGFfZWNfcG9v
            bF9uYW1lKCkKICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJSR1dfUE9PTF9QUkVGSVgi
            XSA9IHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4CiAgICAgICAgc2VsZi5vdXRfbWFw
            WyJSR1dfRU5EUE9JTlQiXSA9ICIiCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZ3dfZW5k
            cG9pbnQ6CiAgICAgICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAg
            ICAgICAgIHNlbGYuY3JlYXRlX3Jnd19hZG1pbl9vcHNfdXNlcigpCiAgICAgICAgICAgIGVsc2U6
            CiAgICAgICAgICAgICAgICBpZiAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfcmVhbG1fbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdf
            cGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBz
            ZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICk6CiAg
            ICAgICAgICAgICAgICAgICAgZXJyID0gc2VsZi52YWxpZGF0ZV9yZ3dfbXVsdGlzaXRlKAogICAg
            ICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lLCAicmVh
            bG0iCiAgICAgICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAgIGVyciA9IHNlbGYu
            dmFsaWRhdGVfcmd3X211bHRpc2l0ZSgKICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJn
            X3BhcnNlci5yZ3dfem9uZWdyb3VwX25hbWUsICJ6b25lZ3JvdXAiCiAgICAgICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICAgICAgICAgIGVyciA9IHNlbGYudmFsaWRhdGVfcmd3X211bHRpc2l0
            ZSgKICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1l
            LCAiem9uZSIKICAgICAgICAgICAgICAgICAgICApCgogICAgICAgICAgICAgICAgaWYgKAogICAg
            ICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUgPT0gIiIKICAg
            ICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZWdyb3VwX25hbWUg
            PT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZV9u
            YW1lID09ICIiCiAgICAgICAgICAgICAgICApIG9yICgKICAgICAgICAgICAgICAgICAgICBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lICE9ICIiCiAgICAgICAgICAgICAgICAgICAgYW5k
            IHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lICE9ICIiCiAgICAgICAgICAgICAg
            ICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVfbmFtZSAhPSAiIgogICAgICAgICAg
            ICAgICAgKToKICAgICAgICAgICAgICAgICAgICAoCiAgICAgICAgICAgICAgICAgICAgICAgIHNl
            bGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX0FDQ0VTU19LRVkiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJfU0VDUkVUX0tFWSJd
            LAogICAgICAgICAgICAgICAgICAgICAgICBpbmZvX2NhcF9zdXBwb3J0ZWQsCiAgICAgICAgICAg
            ICAgICAgICAgICAgIGVyciwKICAgICAgICAgICAgICAgICAgICApID0gc2VsZi5jcmVhdGVfcmd3
            X2FkbWluX29wc191c2VyKCkKICAgICAgICAgICAgICAgICAgICBlcnIgPSBzZWxmLnZhbGlkYXRl
            X3Jnd19lbmRwb2ludChpbmZvX2NhcF9zdXBwb3J0ZWQpCiAgICAgICAgICAgICAgICAgICAgaWYg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfdGxzX2NlcnRfcGF0aDoKICAgICAgICAgICAgICAgICAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJSR1dfVExTX0NFUlQiXSA9ICgKICAgICAgICAgICAgICAgICAgICAg
            ICAgICAgIHNlbGYudmFsaWRhdGVfcmd3X2VuZHBvaW50X3Rsc19jZXJ0KCkKICAgICAgICAgICAg
            ICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAgICMgaWYgdGhlcmUgaXMgbm8gZXJyb3Is
            IHNldCB0aGUgUkdXX0VORFBPSU5UCiAgICAgICAgICAgICAgICAgICAgaWYgZXJyICE9ICItMSI6
            CiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiUkdXX0VORFBPSU5UIl0gPSBz
            ZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludAogICAgICAgICAgICAgICAgZWxzZToKICAgICAg
            ICAgICAgICAgICAgICBlcnIgPSAiUGxlYXNlIHByb3ZpZGUgYWxsIHRoZSBSR1cgbXVsdGlzaXRl
            IHBhcmFtZXRlcnMgb3Igbm9uZSBvZiB0aGVtIgogICAgICAgICAgICAgICAgICAgIHN5cy5zdGRl
            cnIud3JpdGUoZXJyKQoKICAgIGRlZiBnZW5fc2hlbGxfb3V0KHNlbGYpOgogICAgICAgIHNlbGYu
            X2dlbl9vdXRwdXRfbWFwKCkKICAgICAgICBzaE91dElPID0gU3RyaW5nSU8oKQogICAgICAgIGZv
            ciBrLCB2IGluIHNlbGYub3V0X21hcC5pdGVtcygpOgogICAgICAgICAgICBpZiB2IGFuZCBrIG5v
            dCBpbiBzZWxmLl9leGNsdWRlZF9rZXlzOgogICAgICAgICAgICAgICAgc2hPdXRJTy53cml0ZShm
            ImV4cG9ydCB7a309e3Z9e0xJTkVTRVB9IikKICAgICAgICBzaE91dCA9IHNoT3V0SU8uZ2V0dmFs
            dWUoKQogICAgICAgIHNoT3V0SU8uY2xvc2UoKQogICAgICAgIHJldHVybiBzaE91dAoKICAgIGRl
            ZiBnZW5fanNvbl9vdXQoc2VsZik6CiAgICAgICAgc2VsZi5fZ2VuX291dHB1dF9tYXAoKQogICAg
            ICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuICIiCiAg
            ICAgICAganNvbl9vdXQgPSBbCiAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICJuYW1lIjog
            InJvb2stY2VwaC1tb24tZW5kcG9pbnRzIiwKICAgICAgICAgICAgICAgICJraW5kIjogIkNvbmZp
            Z01hcCIsCiAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAiZGF0
            YSI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9DRVBIX01PTl9EQVRBIl0sCiAgICAgICAg
            ICAgICAgICAgICAgIm1heE1vbklkIjogIjAiLAogICAgICAgICAgICAgICAgICAgICJtYXBwaW5n
            IjogInt9IiwKICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgIH0sCiAgICAgICAgICAgIHsK
            ICAgICAgICAgICAgICAgICJuYW1lIjogInJvb2stY2VwaC1tb24iLAogICAgICAgICAgICAgICAg
            ImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAg
            ICAgICAgICJhZG1pbi1zZWNyZXQiOiAiYWRtaW4tc2VjcmV0IiwKICAgICAgICAgICAgICAgICAg
            ICAiZnNpZCI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9GU0lEIl0sCiAgICAgICAgICAg
            ICAgICAgICAgIm1vbi1zZWNyZXQiOiAibW9uLXNlY3JldCIsCiAgICAgICAgICAgICAgICB9LAog
            ICAgICAgICAgICB9LAogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFtZSI6ICJyb29r
            LWNlcGgtb3BlcmF0b3ItY3JlZHMiLAogICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwK
            ICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICJ1c2VySUQiOiBz
            ZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUk5BTUUiXSwKICAgICAgICAgICAgICAgICAg
            ICAidXNlcktleSI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9VU0VSX1NFQ1JFVCJdLAog
            ICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgfSwKICAgICAgICBdCgogICAgICAgICMgaWYg
            J01PTklUT1JJTkdfRU5EUE9JTlQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAnbW9uaXRvcmluZy1l
            bmRwb2ludCcgdG8gQ2x1c3RlcgogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFw
            WyJNT05JVE9SSU5HX0VORFBPSU5UIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsiTU9O
            SVRPUklOR19FTkRQT0lOVF9QT1JUIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291dC5h
            cHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAibW9u
            aXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiQ2VwaENsdXN0
            ZXIiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAg
            ICAiTW9uaXRvcmluZ0VuZHBvaW50Ijogc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5U
            Il0sCiAgICAgICAgICAgICAgICAgICAgICAgICJNb25pdG9yaW5nUG9ydCI6IHNlbGYub3V0X21h
            cFsiTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUIl0sCiAgICAgICAgICAgICAgICAgICAgfSwKICAg
            ICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQoKICAgICAgICAjIGlmICdDU0lfUkJEX05PREVf
            U0VDUkVUJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ3Jvb2stY3NpLXJiZC1wcm92aXNpb25lcicg
            U2VjcmV0CiAgICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9E
            RV9TRUNSRVQiXQogICAgICAgICAgICBhbmQgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VD
            UkVUX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAg
            ICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9vay17c2VsZi5vdXRf
            bWFwWydDU0lfUkJEX05PREVfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICJr
            aW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAg
            ICAgICAgICAgICAgICJ1c2VySUQiOiBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9TRUNSRVRf
            TkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNlbGYub3V0X21hcFsi
            Q1NJX1JCRF9OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAg
            ICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdDU0lfUkJEX1BST1ZJU0lPTkVSX1NF
            Q1JFVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdyb29rLWNzaS1yYmQtcHJvdmlzaW9uZXInIFNl
            Y3JldAogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJ
            U0lPTkVSX1NFQ1JFVCJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9SQkRfUFJP
            VklTSU9ORVJfU0VDUkVUX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFw
            cGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9v
            ay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAg
            ICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJk
            YXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJD
            U0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAg
            ICJ1c2VyS2V5Ijogc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAog
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAg
            ICAgICAjIGlmICdDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCcgZXhpc3RzLCB0aGVuIG9u
            bHkgYWRkICdyb29rLWNzaS1jZXBoZnMtcHJvdmlzaW9uZXInIFNlY3JldAogICAgICAgIGlmICgK
            ICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJd
            CiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VD
            UkVUX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAg
            ICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9vay17c2VsZi5vdXRf
            bWFwWydDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAg
            ICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewog
            ICAgICAgICAgICAgICAgICAgICAgICAiYWRtaW5JRCI6IHNlbGYub3V0X21hcFsiQ1NJX0NFUEhG
            U19QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAiYWRt
            aW5LZXkiOiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUIl0sCiAg
            ICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAg
            ICAgICMgaWYgJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAn
            cm9vay1jc2ktY2VwaGZzLW5vZGUnIFNlY3JldAogICAgICAgIGlmICgKICAgICAgICAgICAgc2Vs
            Zi5vdXRfbWFwWyJDU0lfQ0VQSEZTX05PREVfU0VDUkVUIl0KICAgICAgICAgICAgYW5kIHNlbGYu
            b3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FIl0KICAgICAgICApOgogICAgICAg
            ICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAg
            ICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9O
            QU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAg
            ICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAiYWRtaW5JRCI6IHNl
            bGYub3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FIl0sCiAgICAgICAgICAgICAg
            ICAgICAgICAgICJhZG1pbktleSI6IHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JF
            VCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAg
            ICkKICAgICAgICAjIGlmICdST09LX0VYVEVSTkFMX0RBU0hCT0FSRF9MSU5LJyBleGlzdHMsIHRo
            ZW4gb25seSBhZGQgJ3Jvb2stY2VwaC1kYXNoYm9hcmQtbGluaycgU2VjcmV0CiAgICAgICAgaWYg
            c2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0RBU0hCT0FSRF9MSU5LIl06CiAgICAgICAgICAg
            IGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAi
            bmFtZSI6ICJyb29rLWNlcGgtZGFzaGJvYXJkLWxpbmsiLAogICAgICAgICAgICAgICAgICAgICJr
            aW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAg
            ICAgICAgICAgICAgICJ1c2VySUQiOiAiY2VwaC1kYXNoYm9hcmQtbGluayIsCiAgICAgICAgICAg
            ICAgICAgICAgICAgICJ1c2VyS2V5Ijogc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0RBU0hC
            T0FSRF9MSU5LIl0sCiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAg
            ICAgICAgICAgKQogICAgICAgICMgaWYgJ1JBRE9TX05BTUVTUEFDRScgZXhpc3RzLCB0aGVuIG9u
            bHkgYWRkIHRoZSAiUkFET1NfTkFNRVNQQUNFIiBuYW1lc3BhY2UKICAgICAgICBpZiAoCiAgICAg
            ICAgICAgIHNlbGYub3V0X21hcFsiUkFET1NfTkFNRVNQQUNFIl0KICAgICAgICAgICAgYW5kIHNl
            bGYub3V0X21hcFsiUkVTVFJJQ1RFRF9BVVRIX1BFUk1JU1NJT04iXQogICAgICAgICAgICBhbmQg
            bm90IHNlbGYub3V0X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdCiAgICAgICAgKToK
            ICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAg
            ICAgICAgICAgICJuYW1lIjogInJhZG9zLW5hbWVzcGFjZSIsCiAgICAgICAgICAgICAgICAgICAg
            ImtpbmQiOiAiQ2VwaEJsb2NrUG9vbFJhZG9zTmFtZXNwYWNlIiwKICAgICAgICAgICAgICAgICAg
            ICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgInJhZG9zTmFtZXNwYWNlTmFtZSI6
            IHNlbGYub3V0X21hcFsiUkFET1NfTkFNRVNQQUNFIl0sCiAgICAgICAgICAgICAgICAgICAgICAg
            ICJwb29sIjogc2VsZi5vdXRfbWFwWyJSQkRfUE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAg
            ICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICAgICBqc29uX291
            dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAi
            Y2VwaC1yYmQtcmFkb3MtbmFtZXNwYWNlIiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJT
            dG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAg
            ICAgICAgICAgICAicG9vbCI6IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAogICAgICAg
            ICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1u
            YW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05B
            TUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRy
            b2xsZXItZXhwYW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRf
            UFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3Np
            LnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0
            X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICB9
            LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgZWxzZToKICAgICAgICAg
            ICAgaWYgc2VsZi5vdXRfbWFwWyJSQkRfTUVUQURBVEFfRUNfUE9PTF9OQU1FIl06CiAgICAgICAg
            ICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICAgICAgewogICAgICAgICAg
            ICAgICAgICAgICAgICAibmFtZSI6ICJjZXBoLXJiZCIsCiAgICAgICAgICAgICAgICAgICAgICAg
            ICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAgICAgICAgICAgICJkYXRhIjog
            ewogICAgICAgICAgICAgICAgICAgICAgICAgICAgImRhdGFQb29sIjogc2VsZi5vdXRfbWFwWyJS
            QkRfUE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYu
            b3V0X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdLAogICAgICAgICAgICAgICAgICAg
            ICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNyZXQtbmFtZSI6IGYi
            cm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwK
            ICAgICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vY29udHJvbGxl
            ci1leHBhbmQtc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9QUk9W
            SVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiY3Np
            LnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0
            X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAg
            ICAgfSwKICAgICAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICAgICApCiAgICAgICAgICAg
            IGVsc2U6CiAgICAgICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICAg
            ICAgewogICAgICAgICAgICAgICAgICAgICAgICAibmFtZSI6ICJjZXBoLXJiZCIsCiAgICAgICAg
            ICAgICAgICAgICAgICAgICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAgICAg
            ICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAgICAgInBvb2wiOiBzZWxm
            Lm91dF9tYXBbIlJCRF9QT09MX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJj
            c2kuc3RvcmFnZS5rOHMuaW8vcHJvdmlzaW9uZXItc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYu
            b3V0X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAg
            ICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNl
            Y3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VD
            UkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4
            cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9S
            QkRfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgICAgIH0sCiAgICAg
            ICAgICAgICAgICAgICAgfQogICAgICAgICAgICAgICAgKQoKICAgICAgICAjIGlmICdDRVBIRlNf
            RlNfTkFNRScgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdjZXBoZnMnIFN0b3JhZ2VDbGFzcwogICAg
            ICAgIGlmIHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXToKICAgICAgICAgICAganNvbl9v
            dXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjog
            ImNlcGhmcyIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwKICAg
            ICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImZzTmFt
            ZSI6IHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAg
            ICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIkNFUEhGU19QT09MX05BTUUiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNyZXQtbmFtZSI6
            IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1F
            J119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250cm9s
            bGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfQ0VQSEZT
            X1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNz
            aS5zdG9yYWdlLms4cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91
            dF9tYXBbJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAg
            ICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICMgaWYgJ1JHV19F
            TkRQT0lOVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdjZXBoLXJndycgU3RvcmFnZUNsYXNzCiAg
            ICAgICAgaWYgc2VsZi5vdXRfbWFwWyJSR1dfRU5EUE9JTlQiXToKICAgICAgICAgICAganNvbl9v
            dXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjog
            ImNlcGgtcmd3IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAog
            ICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAiZW5k
            cG9pbnQiOiBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdLAogICAgICAgICAgICAgICAgICAg
            ICAgICAicG9vbFByZWZpeCI6IHNlbGYub3V0X21hcFsiUkdXX1BPT0xfUFJFRklYIl0sCiAgICAg
            ICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAg
            ICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAg
            ICAgIm5hbWUiOiAicmd3LWFkbWluLW9wcy11c2VyIiwKICAgICAgICAgICAgICAgICAgICAia2lu
            ZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAg
            ICAgICAgICAgICAiYWNjZXNzS2V5Ijogc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJf
            QUNDRVNTX0tFWSJdLAogICAgICAgICAgICAgICAgICAgICAgICAic2VjcmV0S2V5Ijogc2VsZi5v
            dXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJfU0VDUkVUX0tFWSJdLAogICAgICAgICAgICAgICAg
            ICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdSR1df
            VExTX0NFUlQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCB0aGUgImNlcGgtcmd3LXRscy1jZXJ0IiBz
            ZWNyZXQKICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VSVCJdOgogICAgICAgICAg
            ICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAg
            Im5hbWUiOiAiY2VwaC1yZ3ctdGxzLWNlcnQiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjog
            IlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAg
            ICAgICAgICJjZXJ0Ijogc2VsZi5vdXRfbWFwWyJSR1dfVExTX0NFUlQiXSwKICAgICAgICAgICAg
            ICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCgogICAgICAgIHJldHVy
            biBqc29uLmR1bXBzKGpzb25fb3V0KSArIExJTkVTRVAKCiAgICBkZWYgdXBncmFkZV91c2Vyc19w
            ZXJtaXNzaW9ucyhzZWxmKToKICAgICAgICB1c2VycyA9IFsKICAgICAgICAgICAgImNsaWVudC5j
            c2ktY2VwaGZzLW5vZGUiLAogICAgICAgICAgICAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9u
            ZXIiLAogICAgICAgICAgICAiY2xpZW50LmNzaS1yYmQtbm9kZSIsCiAgICAgICAgICAgICJjbGll
            bnQuY3NpLXJiZC1wcm92aXNpb25lciIsCiAgICAgICAgICAgICJjbGllbnQuaGVhbHRoY2hlY2tl
            ciIsCiAgICAgICAgXQogICAgICAgIGlmIHNlbGYucnVuX2FzX3VzZXIgIT0gIiIgYW5kIHNlbGYu
            cnVuX2FzX3VzZXIgbm90IGluIHVzZXJzOgogICAgICAgICAgICB1c2Vycy5hcHBlbmQoc2VsZi5y
            dW5fYXNfdXNlcikKICAgICAgICBmb3IgdXNlciBpbiB1c2VyczoKICAgICAgICAgICAgc2VsZi51
            cGdyYWRlX3VzZXJfcGVybWlzc2lvbnModXNlcikKCiAgICBkZWYgZ2V0X3Jnd19wb29sX25hbWVf
            ZHVyaW5nX3VwZ3JhZGUoc2VsZiwgdXNlciwgY2Fwcyk6CiAgICAgICAgaWYgdXNlciA9PSAiY2xp
            ZW50LmhlYWx0aGNoZWNrZXIiOgogICAgICAgICAgICAjIHdoZW4gYWRtaW4gaGFzIG5vdCBwcm92
            aWRlZCByZ3cgcG9vbCBuYW1lIGR1cmluZyB1cGdyYWRlLAogICAgICAgICAgICAjIGdldCB0aGUg
            cmd3IHBvb2wgbmFtZSBmcm9tIGNsaWVudC5oZWFsdGhjaGVja2VyIHVzZXIgd2hpY2ggd2FzIHVz
            ZWQgZHVyaW5nIGNvbm5lY3Rpb24KICAgICAgICAgICAgaWYgbm90IHNlbGYuX2FyZ19wYXJzZXIu
            cmd3X3Bvb2xfcHJlZml4OgogICAgICAgICAgICAgICAgIyBUbyBnZXQgdmFsdWUgJ2RlZmF1bHQn
            IHdoaWNoIGlzIHJndyBwb29sIG5hbWUgZnJvbSAnYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YScKICAgICAgICAgICAgICAgIHBhdHRlcm4gPSByInBvb2w9KC4qPylcLnJnd1wubWV0YSIK
            ICAgICAgICAgICAgICAgIG1hdGNoID0gcmUuc2VhcmNoKHBhdHRlcm4sIGNhcHMpCiAgICAgICAg
            ICAgICAgICBpZiBtYXRjaDoKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJn
            d19wb29sX3ByZWZpeCA9IG1hdGNoLmdyb3VwKDEpCiAgICAgICAgICAgICAgICBlbHNlOgogICAg
            ICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAg
            ICAgICAgICAgICAgICAgICJmYWlsZWQgdG8gZ2V0IHJndyBwb29sIG5hbWUgZm9yIHVwZ3JhZGUi
            CiAgICAgICAgICAgICAgICAgICAgKQoKICAgIGRlZiB1cGdyYWRlX3VzZXJfcGVybWlzc2lvbnMo
            c2VsZiwgdXNlcik6CiAgICAgICAgIyBjaGVjayB3aGV0aGVyIHRoZSBnaXZlbiB1c2VyIGV4aXN0
            cyBvciBub3QKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4IjogImF1dGggZ2V0IiwgImVudGl0
            eSI6IGYie3VzZXJ9IiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICByZXRfdmFsLCBqc29uX291
            dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAg
            aWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgcHJpbnQo
            ZiJ1c2VyIHt1c2VyfSBub3QgZm91bmQgZm9yIHVwZ3JhZGluZy4iKQogICAgICAgICAgICByZXR1
            cm4KICAgICAgICBleGlzdGluZ19jYXBzID0ganNvbl9vdXRbMF1bImNhcHMiXQogICAgICAgIHNl
            bGYuZ2V0X3Jnd19wb29sX25hbWVfZHVyaW5nX3VwZ3JhZGUodXNlciwgc3RyKGV4aXN0aW5nX2Nh
            cHMpKQogICAgICAgIG5ld19jYXAsIF8gPSBzZWxmLmdldF9jYXBzX2FuZF9lbnRpdHkodXNlcikK
            ICAgICAgICBjYXBfa2V5cyA9IFsibW9uIiwgIm1nciIsICJvc2QiLCAibWRzIl0KICAgICAgICBj
            YXBzID0gW10KICAgICAgICBmb3IgZWFjaENhcCBpbiBjYXBfa2V5czoKICAgICAgICAgICAgY3Vy
            X2NhcF92YWx1ZXMgPSBleGlzdGluZ19jYXBzLmdldChlYWNoQ2FwLCAiIikKICAgICAgICAgICAg
            bmV3X2NhcF92YWx1ZXMgPSBuZXdfY2FwLmdldChlYWNoQ2FwLCAiIikKICAgICAgICAgICAgY3Vy
            X2NhcF9wZXJtX2xpc3QgPSBbCiAgICAgICAgICAgICAgICB4LnN0cmlwKCkgZm9yIHggaW4gY3Vy
            X2NhcF92YWx1ZXMuc3BsaXQoIiwiKSBpZiB4LnN0cmlwKCkKICAgICAgICAgICAgXQogICAgICAg
            ICAgICBuZXdfY2FwX3Blcm1fbGlzdCA9IFsKICAgICAgICAgICAgICAgIHguc3RyaXAoKSBmb3Ig
            eCBpbiBuZXdfY2FwX3ZhbHVlcy5zcGxpdCgiLCIpIGlmIHguc3RyaXAoKQogICAgICAgICAgICBd
            CiAgICAgICAgICAgICMgYXBwZW5kIG5ld19jYXBfbGlzdCB0byBjdXJfY2FwX2xpc3QgdG8gbWFp
            bnRhaW4gdGhlIG9yZGVyIG9mIGNhcHMKICAgICAgICAgICAgY3VyX2NhcF9wZXJtX2xpc3QuZXh0
            ZW5kKG5ld19jYXBfcGVybV9saXN0KQogICAgICAgICAgICAjIGVsaW1pbmF0ZSBkdXBsaWNhdGVz
            IHdpdGhvdXQgdXNpbmcgJ3NldCcKICAgICAgICAgICAgIyBzZXQgcmUtb3JkZXJzIGl0ZW1zIGlu
            IHRoZSBsaXN0IGFuZCB3ZSBoYXZlIHRvIGtlZXAgdGhlIG9yZGVyCiAgICAgICAgICAgIG5ld19j
            YXBfbGlzdCA9IFtdCiAgICAgICAgICAgIFtuZXdfY2FwX2xpc3QuYXBwZW5kKHgpIGZvciB4IGlu
            IGN1cl9jYXBfcGVybV9saXN0IGlmIHggbm90IGluIG5ld19jYXBfbGlzdF0KICAgICAgICAgICAg
            ZXhpc3RpbmdfY2Fwc1tlYWNoQ2FwXSA9ICIsICIuam9pbihuZXdfY2FwX2xpc3QpCiAgICAgICAg
            ICAgIGlmIGV4aXN0aW5nX2NhcHNbZWFjaENhcF06CiAgICAgICAgICAgICAgICBjYXBzLmFwcGVu
            ZChlYWNoQ2FwKQogICAgICAgICAgICAgICAgY2Fwcy5hcHBlbmQoZXhpc3RpbmdfY2Fwc1tlYWNo
            Q2FwXSkKICAgICAgICBjbWRfanNvbiA9IHsKICAgICAgICAgICAgInByZWZpeCI6ICJhdXRoIGNh
            cHMiLAogICAgICAgICAgICAiZW50aXR5IjogdXNlciwKICAgICAgICAgICAgImNhcHMiOiBjYXBz
            LAogICAgICAgICAgICAiZm9ybWF0IjogImpzb24iLAogICAgICAgIH0KICAgICAgICByZXRfdmFs
            LCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24p
            CiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiInYXV0aCBjYXBzIHt1c2VyfScgY29tbWFu
            ZCBmYWlsZWQuXG4gRXJyb3I6IHtlcnJfbXNnfSIKICAgICAgICAgICAgKQogICAgICAgIHByaW50
            KGYiVXBkYXRlZCB1c2VyIHt1c2VyfSBzdWNjZXNzZnVsbHkuIikKCiAgICBkZWYgbWFpbihzZWxm
            KToKICAgICAgICBnZW5lcmF0ZWRfb3V0cHV0ID0gIiIKICAgICAgICBpZiBzZWxmLl9hcmdfcGFy
            c2VyLnVwZ3JhZGU6CiAgICAgICAgICAgIHNlbGYudXBncmFkZV91c2Vyc19wZXJtaXNzaW9ucygp
            CiAgICAgICAgZWxpZiBzZWxmLl9hcmdfcGFyc2VyLmZvcm1hdCA9PSAianNvbiI6CiAgICAgICAg
            ICAgIGdlbmVyYXRlZF9vdXRwdXQgPSBzZWxmLmdlbl9qc29uX291dCgpCiAgICAgICAgZWxpZiBz
            ZWxmLl9hcmdfcGFyc2VyLmZvcm1hdCA9PSAiYmFzaCI6CiAgICAgICAgICAgIGdlbmVyYXRlZF9v
            dXRwdXQgPSBzZWxmLmdlbl9zaGVsbF9vdXQoKQogICAgICAgIGVsc2U6CiAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlVuc3VwcG9y
            dGVkIGZvcm1hdDoge3NlbGYuX2FyZ19wYXJzZXIuZm9ybWF0fSIKICAgICAgICAgICAgKQogICAg
            ICAgIHByaW50KGdlbmVyYXRlZF9vdXRwdXQpCiAgICAgICAgaWYgc2VsZi5vdXRwdXRfZmlsZSBh
            bmQgZ2VuZXJhdGVkX291dHB1dDoKICAgICAgICAgICAgZk91dCA9IG9wZW4oc2VsZi5vdXRwdXRf
            ZmlsZSwgbW9kZT0idyIsIGVuY29kaW5nPSJVVEYtOCIpCiAgICAgICAgICAgIGZPdXQud3JpdGUo
            Z2VuZXJhdGVkX291dHB1dCkKICAgICAgICAgICAgZk91dC5jbG9zZSgpCgoKIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyBNQUlOICMjIyMjIyMjIyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMKaWYgX19uYW1lX18gPT0gIl9fbWFpbl9fIjoKICAgIHJqT2Jq
            ID0gUmFkb3NKU09OKCkKICAgIHRyeToKICAgICAgICByak9iai5tYWluKCkKICAgIGV4Y2VwdCBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uIGFzIGVycjoKICAgICAgICBwcmludChmIkV4ZWN1dGlv
            biBGYWlsZWQ6IHtlcnJ9IikKICAgICAgICByYWlzZSBlcnIKICAgIGV4Y2VwdCBLZXlFcnJvciBh
            cyBrRXJyOgogICAgICAgIHByaW50KGYiS2V5RXJyb3I6IHtrRXJyfSIpCiAgICBleGNlcHQgT1NF
            cnJvciBhcyBvc0VycjoKICAgICAgICBwcmludChmIkVycm9yIHdoaWxlIHRyeWluZyB0byBvdXRw
            dXQgdGhlIGRhdGE6IHtvc0Vycn0iKQogICAgZmluYWxseToKICAgICAgICByak9iai5zaHV0ZG93
            bigpCg==
          external.features.ocs.openshift.io/supported-platforms: '["BareMetal", "None",
            "VSphere", "OpenStack", "oVirt"]'
          external.features.ocs.openshift.io/validation: '{"secrets":["rook-ceph-operator-creds",
            "rook-csi-rbd-node", "rook-csi-rbd-provisioner"], "configMaps": ["rook-ceph-mon-endpoints",
            "rook-ceph-mon"], "storageClasses": ["ceph-rbd"], "cephClusters": ["monitoring-endpoint"]}'
          features.ocs.openshift.io/disabled: '["ss-list", "install-wizard", "block-pool",
            "mcg-resource", "odf-dashboard", "common", "storage-provider", "storage-provisioner",
            "dashboard-resources", "csv-actions", "inventory-item", "alert-actions"]'
          features.ocs.openshift.io/enabled: '["kms", "arbiter", "flexible-scaling",
            "multus", "pool-management", "namespace-store", "mcg-standalone", "taint-nodes",
            "vault-sa-kms", "hpcs-kms"]'
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operatorframework.io/initialization-resource: "\n    {\n        \"apiVersion\":
            \"ocs.openshift.io/v1\",\n        \"kind\": \"StorageCluster\",\n        \"metadata\":
            {\n            \"name\": \"example-storagecluster\",\n            \"namespace\":
            \"openshift-storage\"\n        },\n        \"spec\": {\n            \"manageNodes\":
            false,\n            \"monPVCTemplate\": {\n                \"spec\": {\n
            \                   \"accessModes\": [\n                        \"ReadWriteOnce\"\n
            \                   ],\n                    \"resources\": {\n                        \"requests\":
            {\n                            \"storage\": \"10Gi\"\n                        }\n
            \                   },\n                    \"storageClassName\": \"gp2-csi\"\n
            \               }\n            },\n            \"storageDeviceSets\":
            [\n                {\n                    \"count\": 3,\n                    \"dataPVCTemplate\":
            {\n                        \"spec\": {\n                            \"accessModes\":
            [\n                                \"ReadWriteOnce\"\n                            ],\n
            \                           \"resources\": {\n                                \"requests\":
            {\n                                    \"storage\": \"1Ti\"\n                                }\n
            \                           },\n                            \"storageClassName\":
            \"gp2-csi\",\n                            \"volumeMode\": \"Block\"\n
            \                       }\n                    },\n                    \"name\":
            \"example-deviceset\",\n                    \"placement\": {},\n                    \"portable\":
            true,\n                    \"resources\": {}\n                }\n            ]\n
            \       }\n    }\n\t"
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:03e50284f7fe090c57eddf90feea8ff48d78aff8ce95d1b297aa91e71c3ee577
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.4
          operators.operatorframework.io/internal-objects: '["ocsinitializations.ocs.openshift.io","storageclassrequests.ocs.openshift.io","storageconsumers.ocs.openshift.io","storageprofiles.ocs.openshift.io","cephclusters.ceph.rook.io","cephobjectstores.ceph.rook.io","cephobjectstoreusers.ceph.rook.io","cephnfses.ceph.rook.io","cephclients.ceph.rook.io","cephfilesystems.ceph.rook.io","cephfilesystemmirrors.ceph.rook.io","cephrbdmirrors.ceph.rook.io","cephobjectrealms.ceph.rook.io","cephobjectzonegroups.ceph.rook.io","cephobjectzones.ceph.rook.io","cephbucketnotifications.ceph.rook.io","cephbuckettopics.ceph.rook.io","cephfilesystemsubvolumegroups.ceph.rook.io","cephblockpoolradosnamespaces.ceph.rook.io","cephcosidrivers.ceph.rook.io"]'
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/ocs-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OCSInitialization represents the initial data to be created
              when the operator is installed.
            displayName: OCS Initialization
            kind: OCSInitialization
            name: ocsinitializations.ocs.openshift.io
            version: v1
          - description: StorageClassRequest is the Schema for the StorageClassRequests
              API
            displayName: Storage Class Request
            kind: StorageClassRequest
            name: storageclassrequests.ocs.openshift.io
            version: v1alpha1
          - description: StorageCluster represents a cluster including Ceph Cluster,
              NooBaa and all the storage and compute resources required.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.ocs.openshift.io
            version: v1
          - description: StorageConsumer is the Schema for the storageconsumers API
            displayName: Storage Consumer
            kind: StorageConsumer
            name: storageconsumers.ocs.openshift.io
            version: v1alpha1
          - kind: StorageProfile
            name: storageprofiles.ocs.openshift.io
            version: v1
          - description: Represents a Ceph cluster.
            displayName: Ceph Cluster
            kind: CephCluster
            name: cephclusters.ceph.rook.io
            version: v1
          - description: Represents a Ceph Block Pool.
            displayName: Block Pools
            kind: CephBlockPool
            name: cephblockpools.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store.
            displayName: Ceph Object Store
            kind: CephObjectStore
            name: cephobjectstores.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store User.
            displayName: Ceph Object Store User
            kind: CephObjectStoreUser
            name: cephobjectstoreusers.ceph.rook.io
            version: v1
          - description: Represents a cluster of Ceph NFS ganesha gateways.
            displayName: Ceph NFS
            kind: CephNFS
            name: cephnfses.ceph.rook.io
            version: v1
          - description: Represents a Ceph User.
            displayName: Ceph Client
            kind: CephClient
            name: cephclients.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem.
            displayName: Ceph Filesystem
            kind: CephFilesystem
            name: cephfilesystems.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem Mirror.
            displayName: Ceph Filesystem Mirror
            kind: CephFilesystemMirror
            name: cephfilesystemmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph RBD Mirror.
            displayName: Ceph RBD Mirror
            kind: CephRBDMirror
            name: cephrbdmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Realm.
            displayName: Ceph Object Store Realm
            kind: CephObjectRealm
            name: cephobjectrealms.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone Group.
            displayName: Ceph Object Store Zone Group
            kind: CephObjectZoneGroup
            name: cephobjectzonegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone.
            displayName: Ceph Object Store Zone
            kind: CephObjectZone
            name: cephobjectzones.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Notification.
            displayName: Ceph Bucket Notification
            kind: CephBucketNotification
            name: cephbucketnotifications.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Topic.
            displayName: Ceph Bucket Topic
            kind: CephBucketTopic
            name: cephbuckettopics.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem SubVolumeGroup.
            displayName: Ceph Filesystem SubVolumeGroup
            kind: CephFilesystemSubVolumeGroup
            name: cephfilesystemsubvolumegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph BlockPool Rados Namespace.
            displayName: Ceph BlockPool Rados Namespace
            kind: CephBlockPoolRadosNamespace
            name: cephblockpoolradosnamespaces.ceph.rook.io
            version: v1
          - description: Represents a Ceph COSI Driver.
            displayName: Ceph COSI Driver
            kind: CephCOSIDriver
            name: cephcosidrivers.ceph.rook.io
            version: v1
        description: |
          **Red Hat OpenShift Container Storage** deploys three operators.

          ### OpenShift Container Storage operator

          The OpenShift Container Storage operator is the primary operator for OpenShift Container Storage. It serves to facilitate the other operators in OpenShift Container Storage by performing administrative tasks outside their scope as well as watching and configuring their CustomResources.

          ### Rook

          [Rook][1] deploys and manages Ceph on OpenShift, which provides block and file storage.

          # Core Capabilities

          * **Self-managing service:** No matter which supported storage technologies you choose, OpenShift Container Storage ensures that resources can be deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Container Storage you can either build dedicated storage clusters or hyper-converged clusters where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Container Storage:** OpenShift Container Storage integrates Ceph with multiple storage presentations including object storage (compatible with S3), block storage, and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Container Storage efficiently distributes and replicates your data across your cluster to minimize the risk of data loss. With snapshots, cloning, and versioning, no more losing sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data storage for your workloads, using a single namespace.

          [1]: https://rook.io
        displayName: OpenShift Container Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - rook
        - ceph
        - block storage
        - shared filesystem
        - object storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ocs-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:1711a0aabd28937aac1424adebe96619a48b91f7f888f0588117f80891001da4
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:f37e2c611f4cda32172d4e8938938e488076fd8a8f19c252c56a80c48044bc02
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:028120f2212738917f49e0330b60be5f79a208ce2974bea22b71f533aa6ce3be
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:f75a12fc0c449008028d7fc8e18ca7c1afc37ba585885ac754fed860047794a0
        - registry.redhat.io/odf4/ocs-metrics-exporter-rhel9@sha256:c58c67263c08d84a68c8e45cfd77d59a35aee2151bf79b44377016ca0fb429ee
        - registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:9a0008f662b6afa9e2677a8a83bd662d806ad6f4d03fb49bccb78ed487ce4c7a
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:645d4fd09628f7ddcfc765e4ddacaa79fa6a7c35c03292ca2f6e317d48765d6b
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:ef4276b80cacae825d4831de88d3fe6ba7cc324464087b7b28734bd2b8de3315
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:0dc0733df67bb49886ac9715b81bd4ec4a36330fe1afdde657587e1883f5ca7a
        - registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:03e50284f7fe090c57eddf90feea8ff48d78aff8ce95d1b297aa91e71c3ee577
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:ffcc0806ec2afb8b40d75c70ef2a6d98bb887879f5b3f573e900861316ce9e41
        - registry.redhat.io/odf4/ocs-rhel9-operator@sha256:b1a9492bd9b314f86b4e192cf2c83144885ff863d9932e386a41691f50be6193
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:f8bfbf64b4bd030474e239aed5e03af9dfe920eeae444c5cc8ac3dcfa640fcc1
        version: 4.15.20-rhodf
      entries:
      - name: ocs-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: ocs-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: ocs-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: ocs-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: ocs-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: ocs-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: ocs-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: ocs-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: ocs-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: ocs-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: ocs-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: ocs-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: ocs-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: ocs-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: ocs-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: ocs-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: ocs-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: ocs-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: ocs-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: ocs-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: ocs-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |2-

            [
                {
                    "apiVersion": "ocs.openshift.io/v1",
                    "kind": "StorageCluster",
                    "metadata": {
                        "name": "example-storagecluster",
                        "namespace": "openshift-storage"
                    },
                    "spec": {
                        "manageNodes": false,
                        "monPVCTemplate": {
                            "spec": {
                                "accessModes": [
                                    "ReadWriteOnce"
                                ],
                                "resources": {
                                    "requests": {
                                        "storage": "10Gi"
                                    }
                                },
                                "storageClassName": "gp2-csi"
                            }
                        },
                        "storageDeviceSets": [
                            {
                                "count": 3,
                                "dataPVCTemplate": {
                                    "spec": {
                                        "accessModes": [
                                            "ReadWriteOnce"
                                        ],
                                        "resources": {
                                            "requests": {
                                                "storage": "1Ti"
                                            }
                                        },
                                        "storageClassName": "gp2-csi",
                                        "volumeMode": "Block"
                                    }
                                },
                                "name": "example-deviceset",
                                "placement": {},
                                "portable": true,
                                "resources": {}
                            }
                        ]
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Storage
          containerImage: registry.redhat.io/odf4/ocs-rhel9-operator@sha256:3ef1c2679e9229d51ebc04bd2f2b22ce9b11e8f7fdd7bddee69630fe5b1de47c
          description: Red Hat OpenShift Container Storage provides hyperconverged
            storage for applications within an OpenShift cluster.
          external.features.ocs.openshift.io/supported-platforms: '["BareMetal", "None",
            "VSphere", "OpenStack", "oVirt"]'
          external.features.ocs.openshift.io/validation: '{"secrets":["rook-ceph-operator-creds",
            "rook-csi-rbd-node", "rook-csi-rbd-provisioner"], "configMaps": ["rook-ceph-mon-endpoints",
            "rook-ceph-mon"], "storageClasses": ["ceph-rbd"], "cephClusters": ["monitoring-endpoint"]}'
          features.ocs.openshift.io/disabled: '["ss-list", "install-wizard", "block-pool",
            "mcg-resource", "odf-dashboard", "common", "storage-provider", "storage-provisioner",
            "dashboard-resources", "csv-actions", "inventory-item", "alert-actions"]'
          features.ocs.openshift.io/enabled: '["kms", "arbiter", "flexible-scaling",
            "multus", "pool-management", "namespace-store", "mcg-standalone", "taint-nodes",
            "vault-sa-kms", "hpcs-kms"]'
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operatorframework.io/initialization-resource: "\n    {\n        \"apiVersion\":
            \"ocs.openshift.io/v1\",\n        \"kind\": \"StorageCluster\",\n        \"metadata\":
            {\n            \"name\": \"example-storagecluster\",\n            \"namespace\":
            \"openshift-storage\"\n        },\n        \"spec\": {\n            \"manageNodes\":
            false,\n            \"monPVCTemplate\": {\n                \"spec\": {\n
            \                   \"accessModes\": [\n                        \"ReadWriteOnce\"\n
            \                   ],\n                    \"resources\": {\n                        \"requests\":
            {\n                            \"storage\": \"10Gi\"\n                        }\n
            \                   },\n                    \"storageClassName\": \"gp2-csi\"\n
            \               }\n            },\n            \"storageDeviceSets\":
            [\n                {\n                    \"count\": 3,\n                    \"dataPVCTemplate\":
            {\n                        \"spec\": {\n                            \"accessModes\":
            [\n                                \"ReadWriteOnce\"\n                            ],\n
            \                           \"resources\": {\n                                \"requests\":
            {\n                                    \"storage\": \"1Ti\"\n                                }\n
            \                           },\n                            \"storageClassName\":
            \"gp2-csi\",\n                            \"volumeMode\": \"Block\"\n
            \                       }\n                    },\n                    \"name\":
            \"example-deviceset\",\n                    \"placement\": {},\n                    \"portable\":
            true,\n                    \"resources\": {}\n                }\n            ]\n
            \       }\n    }\n\t"
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:b7758b197cf423573de96bdca0cb3045852b20f34fc9c17ad1632b9c3a3cf3d0
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.4
          operators.operatorframework.io/internal-objects: '["ocsinitializations.ocs.openshift.io","storageconsumers.ocs.openshift.io","storageprofiles.ocs.openshift.io","storagerequests.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/ocs-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OCSInitialization represents the initial data to be created
              when the operator is installed.
            displayName: OCS Initialization
            kind: OCSInitialization
            name: ocsinitializations.ocs.openshift.io
            version: v1
          - description: StorageCluster represents a cluster including Ceph Cluster,
              NooBaa and all the storage and compute resources required.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.ocs.openshift.io
            version: v1
          - description: StorageConsumer is the Schema for the storageconsumers API
            displayName: Storage Consumer
            kind: StorageConsumer
            name: storageconsumers.ocs.openshift.io
            version: v1alpha1
          - description: StorageProfile is the Schema for the storageprofiles API
            displayName: Storage Profile
            kind: StorageProfile
            name: storageprofiles.ocs.openshift.io
            version: v1
          - description: StorageRequest is the Schema for the StorageRequests API
            displayName: Storage Request
            kind: StorageRequest
            name: storagerequests.ocs.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Container Storage** deploys three operators.

          ### OpenShift Container Storage operator

          The OpenShift Container Storage operator is the primary operator for OpenShift Container Storage. It serves to facilitate the other operators in OpenShift Container Storage by performing administrative tasks outside their scope as well as watching and configuring their CustomResources.

          ### Rook

          [Rook][1] deploys and manages Ceph on OpenShift, which provides block and file storage.

          # Core Capabilities

          * **Self-managing service:** No matter which supported storage technologies you choose, OpenShift Container Storage ensures that resources can be deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Container Storage you can either build dedicated storage clusters or hyper-converged clusters where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Container Storage:** OpenShift Container Storage integrates Ceph with multiple storage presentations including object storage (compatible with S3), block storage, and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Container Storage efficiently distributes and replicates your data across your cluster to minimize the risk of data loss. With snapshots, cloning, and versioning, no more losing sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data storage for your workloads, using a single namespace.

          [1]: https://rook.io
        displayName: OpenShift Container Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - rook
        - ceph
        - block storage
        - shared filesystem
        - object storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ocs-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d
        - registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:b7758b197cf423573de96bdca0cb3045852b20f34fc9c17ad1632b9c3a3cf3d0
        - registry.redhat.io/odf4/ocs-rhel9-operator@sha256:3ef1c2679e9229d51ebc04bd2f2b22ce9b11e8f7fdd7bddee69630fe5b1de47c
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:70a3c6bf4bde84c7e43fd58c60abc5612d25c792042b902f6abee94bfaa63f80
        - registry.redhat.io/odf4/ocs-metrics-exporter-rhel9@sha256:3332a93e3ea2e83af15a7edab02a9e00d70e81e44338b63a0eacdea54142ad00
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:0084d9053ad2c74926ca02d2ede0461abd59227eee98d122c45e08a5ebaf629a
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:cf7f72469d25fb37301a8159406613673f8968c82af1035fd45a470e0ae8acfa
        version: 4.16.16-rhodf
      entries:
      - name: ocs-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: ocs-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: ocs-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: ocs-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: ocs-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: ocs-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: ocs-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: ocs-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: ocs-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: ocs-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: ocs-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: ocs-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: ocs-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: ocs-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: ocs-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: ocs-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: ocs-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: ocs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Redis
      provider-url: ""
    name: redis-enterprise-operator-cert-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-enterprise-operator.v6.2.10-45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:ffdf10a1b08115fd4de06f1cd751d36644a9bc9a8c11034cd432e9b66a5d2f73
        - registry.connect.redhat.com/redislabs/services-manager@sha256:8b46f2aacfb3e1b243c0b64975accc55fd80b4de9c9ac4769c3c67a5676addc5
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
        version: 6.2.10-45
      entries:
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: 6.2.10
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.22.0
    - currentCSV: redis-enterprise-operator.v7.22.0-11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:088e4b57c6e1ca1680b359078e5eb1bba67343cdbd9a539af26a0ea23a830593
        - registry.connect.redhat.com/redislabs/services-manager@sha256:3b96cdc65c3e42fd22eb165950be96b907725903ac74fe4f82cfa0073e28e95a
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:d56ba97e95e799210e76615095409143a5dddd5c5a1e1be1d551d7e3f9e4d3ef
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:0cfb07f774db131b5b74e5aa91291579c916c4d4cd1a693d7aaed554f0177a68
        version: 7.22.0-11.3
      entries:
      - name: redis-enterprise-operator.v7.22.0-11.3
        version: 7.22.0-11.3
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.22.0-11.3
    - currentCSV: redis-enterprise-operator.v7.4.6-7.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:473b5661049a6c8830ba113504975726154c7b172376ba3b63d891674d59c78f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5f6f32a63f4dd8be85b9d469f0fedd26d6aa76e32b7735bfa7ab7203cb272cdb
        - registry.connect.redhat.com/redislabs/services-manager@sha256:ba20ab644a473bc742af671171d5313537292d88942b16336d03a6386c061d8a
        version: 7.4.6-7.2
      entries:
      - name: redis-enterprise-operator.v7.4.6-7.2
        version: 7.4.6-7.2
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      name: 7.4.6
    - currentCSV: redis-enterprise-operator.v7.8.2-6.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:36c7a36dabb23a4c003e112085cda2873296bfb22e5d11b470d7c1aba55b7f0f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:c6e1f2109a0dac76b51c25c1cbf6030b4894b21b98147e2ad1db2dd8e7693576
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:acf817c90b5f03bc2796fbc305cf43ce5c0644dcbe1b685752090b3f600a0f7b
        version: 7.8.2-6.8
      entries:
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.2
    - currentCSV: redis-enterprise-operator.v7.8.4-9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:803be04534e59cdc7b8898b3bb336676aad1cb77cf266051e226be82679fcf6a
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:4ba0c414a74268bed9c6757e447b340db7c912afa864cfb096ddf88f11feeaf9
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:66ec124da16cc9d0abd66687e088be0b358e20cfd48f5774d34bf5874e6a4e65
        version: 7.8.4-9.0
      entries:
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.4
    - currentCSV: redis-enterprise-operator.v7.8.6-5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:1fff01daece34aeb7befcb0760d5112c8c7d261f2b86388ece416ac94076a573
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5125e83d3c93b224d1983c599994e43899a0ba0c6f2fc4eed31476f755b5fc0b
        - registry.connect.redhat.com/redislabs/services-manager@sha256:88d483f5e37b5a901d1ae3d965304edc000c172d2e5e307297d5b7bcd77ae73b
        version: 7.8.6-5.1
      entries:
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: 7.8.6
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      - name: redis-enterprise-operator.v7.8.4-8.0
        version: 7.8.4-8.0
      - name: redis-enterprise-operator.v7.8.2-6.8
        version: 7.8.2-6.8
      - name: redis-enterprise-operator.v7.8.2-6.3
        version: 7.8.2-6.3
      - name: redis-enterprise-operator.v7.4.6-6.4
        version: 7.4.6-6.4
      - name: redis-enterprise-operator.v7.4.6-6.3
        version: 7.4.6-6.3
      - name: redis-enterprise-operator.v7.4.6-6.1
        version: 7.4.6-6.1
      - name: redis-enterprise-operator.v7.4.6-2.8
        version: 7.4.6-2.8
      - name: redis-enterprise-operator.v7.4.6-2.4
        version: 7.4.6-2.4
      - name: redis-enterprise-operator.v7.4.6-2.3
        version: 7.4.6-2.3
      - name: redis-enterprise-operator.v7.4.6-2.0
        version: 7.4.6-2.0
      name: production
    defaultChannel: production
    packageName: redis-enterprise-operator-cert-rhmp
    provider:
      name: Redis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Sigstore Authors
      provider-url: https://sigstore.dev
    name: sigstore-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sigstore-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.sigstore.dev/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "name": "rekor"
                },
                "spec": {
                  "server": {
                    "ingress": {
                      "hostname": "rekor.example.com"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          containerImage: ghcr.io/sigstore/sigstore-helm-operator@sha256:d9c6eb162b97a1aa1ad3bf515db65d01a96e55b6f8bcd646864d3e64a3ee2b9e
          createdAt: "2021-09-01T21:20:28Z"
          description: Helm Charts for Sigstore Project.
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/sigstore/sigstore-helm-operator
          support: The Sigstore Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Rekor
            name: rekors.helm.sigstore.dev
            version: v1alpha1
        description: |-
          Helm Charts for the sigstore project

          ## Charts

          The following charts are available:

          * [Rekor](https://github.com/sigstore/helm-charts/tree/main/charts/rekor)
        displayName: sigstore
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        links:
        - name: Sigstore Helm Operator
          url: https://github.com/sabre1041/sigstore-helm-operator
        maintainers:
        - email: ablock@redhat.com
          name: Andrew Block
        maturity: alpha
        provider:
          name: The Sigstore Authors
          url: https://sigstore.dev
        relatedImages:
        - quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - ghcr.io/sigstore/sigstore-helm-operator@sha256:d9c6eb162b97a1aa1ad3bf515db65d01a96e55b6f8bcd646864d3e64a3ee2b9e
        version: 0.0.1
      entries:
      - name: sigstore-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: sigstore-helm-operator
    provider:
      name: The Sigstore Authors
      url: https://sigstore.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Loki SIG Operator
      provider-url: ""
    name: loki-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: docker.io/grafana/loki-operator:0.8.0
          createdAt: "2025-03-17T11:00:08Z"
          description: The Community Loki Operator provides Kubernetes native deployment
            and management of Loki and related logging components.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/loki/tree/main/operator
          support: Grafana Loki SIG Operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |-
          The Community Loki Operator provides Kubernetes native deployment and management of Loki and related logging components.
          The purpose of this project is to simplify and automate the configuration of a Loki based logging stack for Kubernetes clusters.

          ### Operator features

          The Loki Operator includes, but is not limited to, the following features:

          * Kubernetes Custom Resources: Use Kubernetes custom resources to deploy and manage Loki, Alerting rules, Recording rules, and related components.
          * Simplified Deployment Configuration: Configure the fundamentals of Loki like tenants, limits, replication factor and storage from a native Kubernetes resource.

          ### Feature Gates

          The Loki Operator Bundle provides a set of feature gates that enable/disable specific feature depending on the target Kubernetes distribution. The following feature gates are enabled by default:
          * `serviceMonitors`: Enables creating a Prometheus-Operator managed ServiceMonitor resource per LokiStack component.
          * `serviceMonitorTlsEndpoints`: Enables TLS for the ServiceMonitor endpoints.
          * `lokiStackAlerts`: Enables creating PrometheusRules for common Loki alerts.
          * `httpEncryption`: Enables TLS encryption for all HTTP LokiStack services.
          * `grpcEncryption`: Enables TLS encryption for all GRPC LokiStack services.
          * `builtInCertManagement`: Enables the built-in facility for generating and rotating TLS client and serving certificates for all LokiStack services and internal clients
          * `lokiStackGateway`: Enables reconciling the reverse-proxy lokistack-gateway component for multi-tenant authentication/authorization traffic control to Loki.
          * `runtimeSeccompProfile`: Enables the restricted seccomp profile on all Lokistack components.
          * `defaultNodeAffinity`: Enable the operator will set a default node affinity on all pods. This will limit scheduling of the pods to Nodes with Linux.
          * `lokiStackWebhook`: Enables the LokiStack CR validation and conversion webhooks.
          * `alertingRuleWebhook`: Enables the AlertingRule CR validation webhook.
          * `recordingRuleWebhook`: Enables the RecordingRule CR validation webhook.
          * `rulerConfigWebhook`: Enables the RulerConfig CR validation webhook.

          In addition it enables the following OpenShift-only related feature gates:
          * `servingCertsService`: Enables OpenShift ServiceCA annotations on the lokistack-gateway service only.
          * `ruleExtendedValidation`: Enables extended validation of AlertingRule and RecordingRule to enforce tenancy in an OpenShift context.
          * `clusterTLSPolicy`: Enables usage of TLS policies set in the API Server.
          * `clusterProxy`: Enables usage of the proxy variables set in the proxy resource.

          ### Before you start

          1. Ensure that the appropriate object storage solution, that will be used by Loki, is avaliable and configured.
        displayName: Community Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Documentation
          url: https://loki-operator.dev/
        maintainers:
        - email: loki-operator-team@googlegroups.com
          name: Grafana Loki SIG Operator
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Grafana Loki SIG Operator
        relatedImages:
        - quay.io/observatorium/opa-openshift:latest
        - docker.io/grafana/loki-operator:0.8.0
        - quay.io/openshift/origin-kube-rbac-proxy:latest
        - docker.io/grafana/loki:3.4.2
        - quay.io/observatorium/api:latest
        version: 0.8.0
      entries:
      - name: loki-operator.v0.8.0
        version: 0.8.0
      - name: loki-operator.v0.7.1
        version: 0.7.1
      - name: loki-operator.v0.7.0
        version: 0.7.0
      - name: loki-operator.v0.6.2
        version: 0.6.2
      - name: loki-operator.v0.6.1
        version: 0.6.1
      - name: loki-operator.v0.6.0
        version: 0.6.0
      - name: loki-operator.v0.5.0
        version: 0.5.0
      - name: loki-operator.v0.4.0
        version: 0.4.0
      - name: loki-operator.v0.3.0
        version: 0.3.0
      - name: loki-operator.v0.2.0
        version: 0.2.0
      - name: loki-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: loki-operator
    provider:
      name: Grafana Loki SIG Operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-license-service-reporter-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-license-service-reporter-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "IBMLicenseServiceReporter",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-license-service-reporter-operator",
                    "app.kubernetes.io/instance": "ibmlicenseservicereporter-instance",
                    "app.kubernetes.io/managed-by": "operator",
                    "app.kubernetes.io/name": "ibmlicenseservicereporter",
                    "app.kubernetes.io/part-of": "ibm-license-service-reporter-operator"
                  },
                  "name": "instance"
                },
                "spec": {
                  "authentication": {
                    "useradmin": {
                      "enabled": true
                    }
                  },
                  "license": {
                    "accept": false
                  },
                  "version": "4.2.14"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:277e2ab849907553e4992c2dccc11f03ae1e0302f20aae48c620968aa97d6afc
          createdAt: "2025-02-26T12:10:52Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <4.2.14'
          operatorframework.io/suggested-namespace: ibm-ls-reporter
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "IBM License Service Reporter is a singleton which may be
              used in a single or multi-cluster environment. It aggregates data pushed
              from IBM License Services, deployed on clusters and from ILMT. \n Documentation:
              https://ibm.biz/lsvc-rprtr. \n License: Please refer to the IBM Terms
              website (ibm.biz/lsvc-lic) to find the license terms for the particular
              IBM product for which you are deploying this component. \n IBM License
              Service Reporter is a free, optionally installed add-on – one of services
              of Cloud Pak Foundational Services. Thanks to the IBM License Service
              Reporter customer can: \n - see IBM software deployments on a single
              dashboard, verify and maintain license compliance and avoid audit infractions,
              \n - see historical data on IBM software deployments to support making
              informed decisions for future purchases, \n - see details of software
              deployments, including source (ILMT for VMs, License Service for containers),
              cluster and quantities of licenses deployed, \n - identify which software
              deployments are VMs vs. containers and can use this information to evaluate
              workloads for modernization."
            displayName: IBMLicense Service Reporter
            kind: IBMLicenseServiceReporter
            name: ibmlicenseservicereporters.operator.ibm.com
            version: v1alpha1
        description: |-
          IBM License Service Reporter is a free, optionally installed add-on – one of services of Cloud Pak Foundational Services. Thanks to the IBM License Service Reporter customer can:
           * see IBM software deployments on a single dashboard, verify and maintain license compliance and avoid audit infractions,
           * see historical data on IBM software deployments to support making informed decisions for future purchases,
           * see details of software deployments, including source (ILMT for VMs, License Service for containers), cluster and quantities of licenses deployed,
           * identify which software deployments are VMs vs. containers and can use this information to evaluate workloads for modernization.

          IBM License Service Reporter is a singleton which may be used in a single or multi-cluster environment. It aggregates data pushed from IBM License Services, deployed on clusters and from ILMT.

          **Supported platforms**

          Red Hat OpenShift Container Platform 4.10 or newer installed on Linux x86_64.

          **Prerequisites**

          Prerequisites depend on the integration of the License Service Reporter with an IBM Cloud Pak. For more information, see the applicable IBM Cloud Pak documentation.
        displayName: IBM License Service Reporter
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - licensing
        - usage
        - IBMLicenseServiceReporter
        - IBM
        - Cloud
        - License
        - License Service
        - Service
        - Reporter
        - License Service Reporter
        - License Reporter
        - Service Reporter
        maintainers:
        - email: talk2sam@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-postgresql@sha256:b6e11080774863f56dd16d5748d03c3f89f6f4872f662ee218d6bf600491a35f
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter@sha256:87e62695d733a60bfc2084cd751810df6801a5fbe1ab87b0a421f6814f1b887a
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-ui@sha256:213889e3c4c1d7bb10c88926ac8276d7e78c558397993d098e36f6c5da617de1
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:277e2ab849907553e4992c2dccc11f03ae1e0302f20aae48c620968aa97d6afc
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-oauth2-proxy@sha256:c403c24857e29ddc306ba5d9bc28c3d556b181c5265fee16879fd2b4d946f899
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator-bundle@sha256:648f92bacd52fb51655925e34da1d0bdd0da720da8489b79064e73a47dcdd54a
        version: 4.2.14
      entries:
      - name: ibm-license-service-reporter-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-license-service-reporter-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: orchestrator-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: orchestrator-operator.v1.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Orchestrator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "orchestrator-sample"
                  },
                  "name": "orchestrator-sample"
                },
                "spec": {
                  "argocd": {
                    "enabled": false,
                    "namespace": "orchestrator-gitops"
                  },
                  "platform": {
                    "eventing": {
                      "broker": {}
                    },
                    "monitoring": {
                      "enabled": false
                    },
                    "namespace": "sonataflow-infra",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "64Mi"
                      }
                    }
                  },
                  "postgres": {
                    "authSecret": {
                      "name": "sonataflow-psql-postgresql",
                      "passwordKey": "postgres-password",
                      "userKey": "postgres-username"
                    },
                    "database": "sonataflow",
                    "name": "sonataflow-psql-postgresql",
                    "namespace": "sonataflow-infra"
                  },
                  "rhdh": {
                    "devMode": true,
                    "installOperator": true,
                    "name": "my-rhdh",
                    "namespace": "rhdh",
                    "plugins": {
                      "notificationsEmail": {
                        "enabled": false,
                        "port": 587,
                        "replyTo": "",
                        "sender": ""
                      }
                    }
                  },
                  "serverless": {
                    "installOperator": true
                  },
                  "serverlessLogic": {
                    "installOperator": true
                  },
                  "tekton": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-08-15T00:51:04Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/rhdhorchestrator/orchestrator-go-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Orchestrator
            name: orchestrators.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub Orchestrator is a plugin that enables serverless asynchronous workflows to Backstage.

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://rhdhorchestrator.io)  displayName: Orchestrator Operator
        displayName: Orchestrator Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Orchestrator
        - RHDH
        links:
        - name: Orchestrator Operator
          url: https://github.com/rhdhorchestrator/orchestrator-go-operator
        maintainers:
        - email: jubah@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/rhdh-orchestrator-dev-preview-beta/controller-rhel9-operator@sha256:9f5423829ac55faf2395e2d0d3683b238339dc03285696e1b350c5c5a1c4df3a
        version: 1.6.1
      entries:
      - name: orchestrator-operator.v1.6.1
        version: 1.6.1
      - name: orchestrator-operator.v1.6.0
        version: 1.6.0
      - name: orchestrator-operator.v1.5.2
        version: 1.5.2
      - name: orchestrator-operator.v1.5.1
        version: 1.5.1
      - name: orchestrator-operator.v1.5.0
        version: 1.5.0
      - name: orchestrator-operator.v1.4.0
        version: 1.4.0
      - name: orchestrator-operator.v1.3.1
        version: 1.3.1
      - name: orchestrator-operator.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: orchestrator-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtv-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:d0158c37686bf378347dea5368c0c7838eec9674fbb0ff11f369e6fd99753652
          createdAt: "2024-09-23T16:42:11Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.7.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.5 is supported on OpenShift 4.12 and 4.13

          Migration Toolkit for Virtualization 2.6 is supported on OpenShift 4.14 and 4.15

          Migration Toolkit for Virtualization 2.7 is supported on OpenShift 4.15 and 4.16

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:d0158c37686bf378347dea5368c0c7838eec9674fbb0ff11f369e6fd99753652
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:f8b8d2ce23c1be9dd74d18c7a90324288213e91630da529ded561fbfe5d59a5d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:fffa2fd9f8c804d021264748a910ceb17cee388b4d55cc763a8598234db87f15
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:31e20712c82ae77807eb5dfba5e8616c045280ff164027443214ab8f259d964a
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:ac13419a8c9993bbe656563f478095526dedd34a93d6463979d26852d08ed217
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:19909db45872b2d665eac31f20dac2f0fa3bd9b9a343f2be839747106de6bc5b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:90742a924a1782dced8ae639e68aef93df3c14e4a7b65e2e6b1b73e22860e7a7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:3a07a1e7c74783b0de3c190717f17091d4c9d573776419b46c2ef0c55328d769
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:045cfaec032648cb1a734398fd625de7ac1cc25bfd621b7558e641da82316b58
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:8ce4bba6a69de762f221792c03f6a759afcd16dc576de4b9b1f341d660cb82b0
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:1b06a93c3458b92d40d8eb437043c9abbb9ec5290fe85b2f6ef12297bea0c838
        version: 2.7.0
      entries:
      - name: mtv-operator.v2.7.0
        version: 2.7.0
      name: candidate
    - currentCSV: mtv-operator.2.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui": "true",
                  "feature_validation": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "plan": {
                      "name": "example-plan",
                      "namespace": "openshift-mtv"
                   }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {}
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "type": "vmware",
                   "secret": {
                      "name": "",
                      "namespace": ""
                   },
                   "url": ""
                }
              },
              {
                "kind": "Host",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                     "namespace": "",
                     "name": ""
                   },
                   "id": "",
                   "ipAddress": ""
                }
              },
              {
                "kind": "NetworkMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-networkmap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "type": "pod",
                            "namespace": "",
                            "name": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "StorageMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-storagemap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "storageClass": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "Hook",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "kind": "Provisioner",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "cinder",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "name": "kubernetes.io/cinder",
                  "volumeModes": [
                    {
                      "name": "Block",
                      "priority": 0,
                      "accessModes": [
                        {
                          "name": "ReadWriteOnce",
                          "priority": 0
                        },
                        {
                          "name": "ReadWriteMany",
                          "priority": 1
                        }
                      ]
                    },
                    {
                      "name": "Filesystem",
                      "priority": 1,
                      "accessModes": [
                        {
                          "name": "ReadWriteOnce",
                          "priority": 0
                        },
                        {
                          "name": "ReadWriteMany",
                          "priority": 1
                        }
                      ]
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:74a740878707470ea79c7bf03a7f045c2ea77d924d0890042ba3554dc5613a33
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.1.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui": "true",
                "feature_validation": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          repository: https://github.com/konveyor/forklift-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: Storage Provisioner Capabilities
            displayName: Provisioner
            kind: Provisioner
            name: provisioners.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator fully manages the deployment and life cycle of Migration Toolkit for Virtualization on OpenShift.

          If you experience any issues or have feature requests, please [file an issue in our Github repository](https://github.com/konveyor/forklift-operator/issues)
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        links:
        - name: Konveyor Forklift Operator
          url: https://github.com/konveyor/forklift-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:584a696270554f4fb64755cdc5b525bc6852925e6780f776fb309b1bd8d81d0f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ui-rhel8@sha256:8089b0889f8956f2abf5a69f4ab8f4382919bc15a48663b98551fb0b865d8af7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel8@sha256:aead9b494562985e20fa466888f28119ab302d64e93d7f23bc7d8da1c60267e7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel8@sha256:8c6758c2d81f807c3875f41cec14d9e39f1fe7710584794de57a1449510110db
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:74a740878707470ea79c7bf03a7f045c2ea77d924d0890042ba3554dc5613a33
        version: 2.1.0
      entries:
      - name: mtv-operator.2.1.0
        version: 2.1.0
      name: release-v2.1.0
    - currentCSV: mtv-operator.v2.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui": "true",
                  "feature_validation": "true",
                  "feature_must_gather_api": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "plan": {
                      "name": "example-plan",
                      "namespace": "openshift-mtv"
                   }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {}
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "type": "vmware",
                   "secret": {
                      "name": "",
                      "namespace": ""
                   },
                   "url": ""
                }
              },
              {
                "kind": "Host",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                     "namespace": "",
                     "name": ""
                   },
                   "id": "",
                   "ipAddress": ""
                }
              },
              {
                "kind": "NetworkMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-networkmap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "type": "pod",
                            "namespace": "",
                            "name": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "StorageMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-storagemap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "storageClass": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "Hook",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "kind": "Provisioner",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "cinder",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "name": "kubernetes.io/cinder",
                  "volumeModes": [
                    {
                      "name": "Block",
                      "priority": 0,
                      "accessModes": [
                        {
                          "name": "ReadWriteOnce",
                          "priority": 0
                        },
                        {
                          "name": "ReadWriteMany",
                          "priority": 1
                        }
                      ]
                    },
                    {
                      "name": "Filesystem",
                      "priority": 1,
                      "accessModes": [
                        {
                          "name": "ReadWriteOnce",
                          "priority": 0
                        },
                        {
                          "name": "ReadWriteMany",
                          "priority": 1
                        }
                      ]
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:3d28138500d6b715e7d11983684db12ced28dfef9c97c619044f0b830534f4ed
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.2.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui": "true",
                "feature_validation": "true",
                "feature_must_gather_api": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          repository: https://github.com/konveyor/forklift-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: Storage Provisioner Capabilities
            displayName: Provisioner
            kind: Provisioner
            name: provisioners.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation
          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.
          * Must-gather-api, a service to generate targeted must-gather archives.

          ### Compatibility
          Migration Toolkit for Virtualization 2.0 is supported in OpenShift 4.7
          Migration Toolkit for Virtualization 2.1 is supported in OpenShift 4.8
          Migration Toolkit for Virtualization 2.2 is supported in OpenShift 4.9

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [Konveyor Community](https://github.com/konveyor/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/konveyor/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Konveyor Forklift Operator
          url: https://github.com/konveyor/forklift-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:3d28138500d6b715e7d11983684db12ced28dfef9c97c619044f0b830534f4ed
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel8@sha256:097a3c667e594060844cbe5d16adcc278d35402d1965ff3fca249c8f84a688dc
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:02ec1b0dbc12ba31fce85ee164c97560ba3a3184316324ab7c871989a319bf8d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-api-rhel8@sha256:d6bf2bb018c471f8f86f85c62ef52f4a99510b748ac8b6b7b177fc44fa53fb44
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ui-rhel8@sha256:697b20d6423ac25c0d2f41b2598f4909b6863e85265a4f52bc23b8ce77afc732
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel8@sha256:220d8f97173292dc8389bf4b1a350843757f01ea7861bbcebcf75749366b450b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel8@sha256:6f332be2991b026d5ce2b22e5bdef8d17da06c8b437fb3e5b26e8d35b75e3f95
        version: 2.2.0
      entries:
      - name: mtv-operator.v2.2.0
        version: 2.2.0
      name: release-v2.2.0
    - currentCSV: mtv-operator.v2.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui": "true",
                  "feature_validation": "true",
                  "feature_must_gather_api": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "plan": {
                      "name": "example-plan",
                      "namespace": "openshift-mtv"
                   }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {}
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "type": "vmware",
                   "secret": {
                      "name": "",
                      "namespace": ""
                   },
                   "url": ""
                }
              },
              {
                "kind": "Host",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                     "namespace": "",
                     "name": ""
                   },
                   "id": "",
                   "ipAddress": ""
                }
              },
              {
                "kind": "NetworkMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-networkmap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "type": "pod",
                            "namespace": "",
                            "name": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "StorageMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-storagemap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "storageClass": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "Hook",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:fc6a4503d00719a0d22736a78689b5828a8ebaf8c7895689a56df5cf7b43b97f
          createdAt: "2022-11-08T18:56:59Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.3.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui": "true",
                "feature_validation": "true",
                "feature_must_gather_api": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/forklift-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.
          * Must-gather-api, a service to generate targeted must-gather archives.

          ### Compatibility

          Migration Toolkit for Virtualization 2.2 is supported in OpenShift 4.9

          Migration Toolkit for Virtualization 2.3 is supported in OpenShift 4.10

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [Konveyor Community](https://github.com/konveyor/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/konveyor/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/konveyor/forklift-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ui-rhel8@sha256:e6c2337e610e0f634e57a200a92f492ebe1f111d9fccaa019f8756ba907b067b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel8@sha256:4423034aac33409721ce1b5344f6d5330c174c9b47fc6837b8de7b47018036d7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel8@sha256:872e2a803742071bf81b70d9a58aafd0ceb6ce1d80e651d876690820835854c0
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:fc6a4503d00719a0d22736a78689b5828a8ebaf8c7895689a56df5cf7b43b97f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel8@sha256:59ad5149ce8691eb7944fa291be75e5701c8e74160a0b9a74d098872be704a67
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:a0c201d3474996a08ce662b074ebb069e5082da36f3e7977cf54e3c4e1e54b1d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-api-rhel8@sha256:a7a2d6216ee87284a4f9e07d4c1973b9e001292c504c3b933515826e90073167
        version: 2.3.4
      entries:
      - name: mtv-operator.v2.3.4
        version: 2.3.4
      name: release-v2.3
    - currentCSV: mtv-operator.v2.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui": "true",
                  "feature_validation": "true",
                  "feature_must_gather_api": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "plan": {
                      "name": "example-plan",
                      "namespace": "openshift-mtv"
                   }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {}
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "type": "vmware",
                   "secret": {
                      "name": "",
                      "namespace": ""
                   },
                   "url": ""
                }
              },
              {
                "kind": "Host",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                     "namespace": "",
                     "name": ""
                   },
                   "id": "",
                   "ipAddress": ""
                }
              },
              {
                "kind": "NetworkMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-networkmap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "type": "pod",
                            "namespace": "",
                            "name": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "StorageMap",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                   "name": "example-storagemap",
                   "namespace": "openshift-mtv"
                },
                "spec": {
                   "provider": {
                      "namespace": "",
                      "name": ""
                   },
                   "map": [
                      {
                         "source": {
                            "id": ""
                         },
                         "destination": {
                            "storageClass": ""
                         }
                      }
                   ]
                }
              },
              {
                "kind": "Hook",
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:fc6a4503d00719a0d22736a78689b5828a8ebaf8c7895689a56df5cf7b43b97f
          createdAt: "2022-11-08T18:56:59Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.3.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui": "true",
                "feature_validation": "true",
                "feature_must_gather_api": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/forklift-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.
          * Must-gather-api, a service to generate targeted must-gather archives.

          ### Compatibility

          Migration Toolkit for Virtualization 2.2 is supported in OpenShift 4.9

          Migration Toolkit for Virtualization 2.3 is supported in OpenShift 4.10

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [Konveyor Community](https://github.com/konveyor/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/konveyor/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/konveyor/forklift-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel8@sha256:4423034aac33409721ce1b5344f6d5330c174c9b47fc6837b8de7b47018036d7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel8@sha256:872e2a803742071bf81b70d9a58aafd0ceb6ce1d80e651d876690820835854c0
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel8@sha256:59ad5149ce8691eb7944fa291be75e5701c8e74160a0b9a74d098872be704a67
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:a0c201d3474996a08ce662b074ebb069e5082da36f3e7977cf54e3c4e1e54b1d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-api-rhel8@sha256:a7a2d6216ee87284a4f9e07d4c1973b9e001292c504c3b933515826e90073167
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ui-rhel8@sha256:e6c2337e610e0f634e57a200a92f492ebe1f111d9fccaa019f8756ba907b067b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:fc6a4503d00719a0d22736a78689b5828a8ebaf8c7895689a56df5cf7b43b97f
        version: 2.3.4
      entries:
      - name: mtv-operator.v2.3.4
        version: 2.3.4
      name: release-v2.3.0
    - currentCSV: mtv-operator.v2.4.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_must_gather_api": "true",
                  "feature_ui": "false",
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:f17fb7a2a840bda51551683e18d01ad1fae3ea73c293de1711f382ed309b8668
          createdAt: "2023-10-23T12:16:09Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.4.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui": "false",
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_must_gather_api": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.
          * Must-gather-api, a service to generate targeted must-gather archives.

          ### Compatibility

          Migration Toolkit for Virtualization 2.2 is supported in OpenShift 4.9

          Migration Toolkit for Virtualization 2.3 is supported in OpenShift 4.10 and 4.11

          Migration Toolkit for Virtualization 2.4 is supported in OpenShift 4.11 and 4.12

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-api-rhel8@sha256:cf22690c6bb420b9acc09ee5f90329ece6a35c18e5d3d6f6e40a320e2385ad85
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:7b4eb4de5dbc7e1af9c74d47fb6bba39f4973dc2d41531989c9fcf454b715b14
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:dc7ed113fae6ea3d5baaf165b08b5404a15f3a70f08089caca65b837623af4fb
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:8d3aeb8f6cfb17f5694f92f56a2361e08d220e2608183fb39ac94e92187b3fc7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-warm-rhel8@sha256:a6f91bc0ffc5e2f545a18b8888715281db7a751e1eff5f9eb89e8d319268a3ec
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:f17fb7a2a840bda51551683e18d01ad1fae3ea73c293de1711f382ed309b8668
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:8e7a7b3c2dc31b90e2102a82633f1eda95355496ff657964fedb26157f9ec229
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:ad9bacdfa4f10e020902d8e0381bdd84e48ce89ffab0f4144bd3c03fc91f7259
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:86716db780d3e43fa1e3029ab095ae6a557cf023797f4d79565a1aa5ebcb9b93
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:a7caa55c9d211a869b42e6f0101cd5099d64f5f3fd423f8a06091bad98d84aff
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ui-rhel8@sha256:54638e44269175e112a510fe2558f9a1dda4758bf3b3fb5e5b54c6617a5917bd
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:20a7b16f9f5d3dd58dbc7c061567590618a24bd595911ed089034095cc919d96
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:9aec00a487545eef76dcc4701680c5ec5bdc5e74302edbdfbc12c821b8ab4997
        version: 2.4.3
      entries:
      - name: mtv-operator.v2.4.3
        version: 2.4.3
      name: release-v2.4
    - currentCSV: mtv-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_must_gather_api": "true",
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:7dac245977fa0beeb7539b8ea5b3be68c457da132c979fd36405509fc879c7e1
          createdAt: "2023-08-31T15:03:42Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          olm.skipRange: '>=0.0.0 <2.5.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_must_gather_api": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.
          * Must-gather-api, a service to generate targeted must-gather archives.

          ### Compatibility

          Migration Toolkit for Virtualization 2.4 is supported on OpenShift 4.11 and 4.12

          Migration Toolkit for Virtualization 2.5 is supported on OpenShift 4.12 and 4.13

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:00fac23d8a514208c906ab2c911f188947368cb5e3076aacf3dc049e27b6c366
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:7dac245977fa0beeb7539b8ea5b3be68c457da132c979fd36405509fc879c7e1
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:8cc711f92258cad8f0747d3b98fef432d20661b90c4b67df8aa91650f2be3388
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-warm-rhel8@sha256:b5e0a88008c5c355844ed16b0889b0d73e63748fffafe57d0d1c52bc39c50f87
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:349cb43a620b2db9e724d874b231ef2ad896260009e3b0a93539121ac9cbe84b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:c157d6ebfb12f37a646f5f3bfbfdc90b7c8c669afe08ad342d34e474c39aa094
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:68cde15ed754189d74a528fc61cc2191861a27a4a9be112fb3ce1de3e836a77c
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-api-rhel8@sha256:c702dc948e42b9ba1e87b988f603be3df249a44dfdd6a11b0b9b49cb25f6de0c
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:55d773cad8fcf2919f5b0a5387dea62feb4ce676fb94eba310fc4d459f729d73
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:18591334cd99eb9e9f7e3d2609ff8666a9c0625e996d2f8751a14ebb9386457a
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:6097cc8e88905793bc61e48a62d77decffbbc59c1d4de83bb2eb5340ea1e8697
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:e2fca2e132bfc9a90a7c1f7e8e6b6d6c4fd8a63ec4177f5dea3f06c0c529e722
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:e665d8c5da798dc0dcd66b691db087de49564333e3c6e238d16a356c8c62d080
        version: 2.5.0
      entries:
      - name: mtv-operator.v2.5.0
        version: 2.5.0
      name: release-v2.5
    - currentCSV: mtv-operator.v2.6.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:1dfbb0993d3009608a6b272f178197f8a88edb047db4f4ff84aee8c023191d65
          createdAt: "2024-09-13T15:42:40Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.6.7'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.5 is supported on OpenShift 4.12 and 4.13

          Migration Toolkit for Virtualization 2.6 is supported on OpenShift 4.14 and 4.15

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:d34826b2a9370a1b6502d8e74a1254b2bd786684b25d90c2f0a960be0e99777d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-warm-rhel8@sha256:a46eb455027d248bcf591b612f34eb2ff4cd887cbbc3d116f78187b579ac31ab
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:3fef1c267f1f89d51ba7cd54bf39e17d7fe8ba5a4b6ecf9cdedcbe6d49d28bcc
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:1dfbb0993d3009608a6b272f178197f8a88edb047db4f4ff84aee8c023191d65
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:537f5fed16dcc46ee9cea05285a68c0e037c9b439289ff537590352865eced92
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:7dbf0f8d3d11cdb73033377552f999951331ed61d28c98cfd836903faf5601f9
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:a30fca5c5f260b94181c03a8c4c1cfe340048af87515e86571e9676e836d4b9f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:c98635ac256743873c4d74d63f4278cd26794dbaa1d56a82b15089dab510e905
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:a24acbfeca235270e75bd78046aa1dded9ff2a325a725cb46fc52ed3ff6880b8
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:403a5f8fd64926a9638342dbe568b557a7e3dc4b700cde79e64ecd9d948002a7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:6e0b0a7aca1a4afef62cdf511195e6a567feb644d3e5128c9899afc351807212
        version: 2.6.7
      entries:
      - name: mtv-operator.v2.6.7
        version: 2.6.7
      - name: mtv-operator.v2.6.6
        version: 2.6.6
      - name: mtv-operator.v2.6.5
        version: 2.6.5
      - name: mtv-operator.v2.6.4
        version: 2.6.4
      - name: mtv-operator.v2.6.3
        version: 2.6.3
      - name: mtv-operator.v2.6.2
        version: 2.6.2
      - name: mtv-operator.v2.6.1
        version: 2.6.1
      - name: mtv-operator.v2.6.0
        version: 2.6.0
      name: release-v2.6
    - currentCSV: mtv-operator.v2.7.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:db4cbc7214cdc7bf176ff16e562120111219ad8e663b7d4385dfd4c164253b7f
          createdAt: "2025-04-22T18:23:33Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.7.12'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.7 is supported on OpenShift 4.15, 4.16 and 4.17

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:a0360971c4cf448f99eb9ca79c9a153aec479e59e5985b91e019ee478382510b
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:ed8cefd01f3b03df7faaf1ba0774cc18b7fc08b25f359d1e52c794b4ad60441e
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:33bae3e4c5b96df3319287d409d8d25db7db33704c7e9f8a58aec797bf20d12a
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:db4cbc7214cdc7bf176ff16e562120111219ad8e663b7d4385dfd4c164253b7f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:97ba4a4b479957e57eb296598d0c8797233f3fd0dbac20fb4686db02d3a8ed9f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:1622759bca392e2f7f66611df7988e7c63138807f917f8b34aefcc60467051ec
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:a32fba94e62f4c979e3b3596ab3a70750bd3bbf841e6004b5fd5305ac796027f
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:51fdb9274911e93ae9dff9ab616a04a8a2c8fce4b6b21c05cf5f0145b1f93de0
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:a0b713409ad1d5018832cd50a2ffa3ce70ebf6cf1a4e532b3676b6f039316b52
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:ec2b67b3aff9b74f2246c310f1871198bbea8428b4e563d0e8c254a785b46a27
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:112a465da1695340286a7ff6e4dd66144209ce4389134f03813f786d30ba7508
        version: 2.7.12
      entries:
      - name: mtv-operator.v2.7.12
        version: 2.7.12
      - name: mtv-operator.v2.7.11
        version: 2.7.11
      - name: mtv-operator.v2.7.10
        version: 2.7.10
      - name: mtv-operator.v2.7.9
        version: 2.7.9
      - name: mtv-operator.v2.7.8
        version: 2.7.8
      - name: mtv-operator.v2.7.7
        version: 2.7.7
      - name: mtv-operator.v2.7.6
        version: 2.7.6
      - name: mtv-operator.v2.7.5
        version: 2.7.5
      - name: mtv-operator.v2.7.4
        version: 2.7.4
      - name: mtv-operator.v2.7.3
        version: 2.7.3
      - name: mtv-operator.v2.7.2
        version: 2.7.2
      - name: mtv-operator.v2.7.1
        version: 2.7.1
      - name: mtv-operator.v2.7.0
        version: 2.7.0
      name: release-v2.7
    - currentCSV: mtv-operator.v2.8.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "\u003cbase64-encoded-playbook\u003e"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "konveyor-forklift"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:e4145efb22dfc336d8b81f24f5b488db602d6d547479d615bf82b4fdce96c784
          createdAt: "2025-08-26T14:06:48Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: VSphere Xcopy Volume Populator
            displayName: VSphereXcopyVolumePopulator
            kind: VSphereXcopyVolumePopulator
            name: vspherexcopyvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.8 is supported on OpenShift 4.16, 4.17 and 4.18

          Migration Toolkit for Virtualization 2.9 is supported on OpenShift 4.17, 4.18 and 4.19

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/support/policy/updates/migration-toolkit-for-virtualization).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:e4145efb22dfc336d8b81f24f5b488db602d6d547479d615bf82b4fdce96c784
        version: 2.8.7
      entries:
      - name: mtv-operator.v2.8.7
        version: 2.8.7
      - name: mtv-operator.v2.8.6
        version: 2.8.6
      - name: mtv-operator.v2.8.5
        version: 2.8.5
      - name: mtv-operator.v2.8.4
        version: 2.8.4
      - name: mtv-operator.v2.8.3
        version: 2.8.3
      - name: mtv-operator.v2.8.2
        version: 2.8.2
      - name: mtv-operator.v2.8.1
        version: 2.8.1
      - name: mtv-operator.v2.8.0
        version: 2.8.0
      name: release-v2.8
    defaultChannel: release-v2.8
    packageName: mtv-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: serverless-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: serverless-operator.v1.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.36.0 <1.36.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:52ba6b1252b6f3d41e2a0b3b880cb768010ab9dd925ac34596f6adaedcbbe07c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-source-rhel8@sha256:88323da16360981f27b8e4686fbbb64053e4e0e3c79c65d695edb317b334e1e0
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:3fe326ad9a00b112fc888ffd345c8d5042f0a22cf32ea575931192c39ce1fece
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:259e5dab50d36cdbfc117a42325128c4e16f34c048ba64b36d4a89132ddb4c1a
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:4344716eb2940c6a3fa9a278250f936602bb1549fece6037912837719c178b24
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-transform-jsonata-rhel8@sha256:3af4c48030263eb5769e5f36a6dd13fd390d8566efaab37532f39e1df1c95c67
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sns-sink-rhel8@sha256:51b5e18e8d4c1ca81d7c5ddfea6352a3887f757acfcc53bd2f8bd3b01337ed42
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-ddb-streams-source-rhel8@sha256:a903480159ebf1d8652eeaa702d9885fb76bb106c02d75cec5c82c31d77d6410
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:e3ae1cc6012f19631c012a07b113fb015977d2ee4f9faab3207fbc2464c25aa7
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:596a7272d6de714d7dbc6e00e9e50a2872846c269547465252d953172ef0b009
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:f4d86b196457cdeebdc380b6f1b707c10f8dee789c7779c90ee1332e511af8c4
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:f2d112ba9fdfccea988589b99b463c8a7157475f7a3f03678e8999e65b678cc3
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-sink-rhel8@sha256:afa634e029bcf79a972f49372498d44b45bcfbdb6eb28ecaf4f0b0b2a0a1824b
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:dcfe5fbc57b9f8520adbff368c8bf29eb491c9c26d1a1737a1722cffea83eb1b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-source-rhel8@sha256:27e74287774cda208f667bc2a386debe6b6255e8991abfb37c959c8525f2af53
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:9b328fff0cc8545fb9bb1f24f11d78cfac3ee76dedc11fa4def4262d79a60e8c
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:82e69a1890042f5636ac7bb645b1aaeaba0a4894851eccc1fa1c63296918e31a
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:3d780ad1ab280b5092b90e277485d589da2a18e8b08090af7d95aecd17af545d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:631eb3a8dd9d83ceafa7d8a4da4b1e2d4f2740e6017c766828dcc7575cf1322f
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-log-sink-rhel8@sha256:30b8045ef5f925a36c58941386f1899bec125d005a08aeeeba5edd49b0859149
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:c7c2e738e0d9da1d2d1993f68938e72c4709b1f28a54fa92224ef6a0e456b0aa
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:a1834b620f875681388f0ae224dd0771773a8b61b1d8e5214a942f232854b87b
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:074ebfc4494581611647ba90049655a470ca43743610a27be77db66a8633bac6
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:429d5d9568cca31fbe83c663276a300286af4cdbf3d35685afdbe5f6f2f8c6e0
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:5352e4eb4e426a41521276decadf204b6d23b4762a298e4d3dc73969e3eddc68
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:bf82b37ab8fd29c7a7cd846008a16d14b4cfdc1c09df4f731d71be3957e90123
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:6b7316cce3432d391adc6ac38c92ff0ee86dd7ae0101957cc65cd338cc69e352
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:3b04e9d8cd4b5e43a07ece838d10478723a6c6529350f010550b4a0d1d0950d1
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:6210a946f373b9c36c9b31293debe6fed090008b8234c0ba17d37fa81646172e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:c39cb88afaddc295d93a3774b8a20d001b82392df540fb502ee02316d82460a9
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:0c2cdf917bbad1f5bca8a2a2ef12a96cfcd0c77c190cc35687eb57169b54ab6f
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:187fb9d64b53c55a029ffce3c0f9e7c3304683fa8a0e832cfb7134e8a3d0c237
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:d6f338e08fa937c8a4ccfec692424de4acded565f7ca3971ea507dffeb351503
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:f3d87a7fa5b5f2818effcac3304bd56eddcdb0db97a18ac0f35a06fc7e223e7d
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:4aeb68a22f63300e54391ec6b4cda8c4e665d904b5dcdd3c40155040525742a0
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:ed60e889a677fb139f2eb4bd9ab4fda6ecc39fd050ca2b6cc1c19717e428535a
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:77e4689b2ce7a713e905553e0e42b8e128045f40cb6c6d4ee082f126cbc46b00
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:4274004b6289fa9bea3fa6e56d64077a328a6b5e63a9f0e9c41f6d561f682bd7
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:a6f565d73375e827d830dca85838f4c4cb7f08b0c078acb0c0d567b25f0d256a
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-timer-source-rhel8@sha256:8fed5ef9a1f39a55077bb5026c8854d0ee7fc8d8ff5da277ce9c1bfc2ab2474b
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:78fee4c596f2848487ca704da3fa1074e3c9c99b3f96f66833aeba63836d7b93
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:a47b6b5237a87ccd90e5a49a53d06485fd55c87eebf12d01c335b87fc3a8c494
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-sink-rhel8@sha256:d9b944e64391a996104afbe8a21f66316a948e9fb7b91734e861847285be59c7
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:ee6533b71cf280877c9d7489dfd821c9209c6802bed7fa3fdb4f8f00fbeacb09
        version: 1.36.1
      entries:
      - name: serverless-operator.v1.36.1
        version: 1.36.1
      - name: serverless-operator.v1.36.0
        version: 1.36.0
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable
    - currentCSV: serverless-operator.v1.29.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2023-07-28T05:42:31Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.29.0 <1.29.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of a particular version of Knative
              Serving
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Eventing
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Kafka components
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:a3c5d5fd8014c432830be912aedebbc58f9c60a8cabd87ae6c42c7b8d27db8f9
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:a59ba5aeb8e66db3d8ba9074b8e553603c59e6b0e35b57b82c4ce3de589b94cc
        - registry.redhat.io/ocp-tools-4-tech-preview/source-to-image-rhel8@sha256:98d8cb3a255641ca6a1bce854e5e2460c20de9fb9b28e3cc67eb459f122873dd
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:56e48afb1dcbf4e266ace944105f3c45042550d8d4a0528c982225a793d26d56
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:31d45f10a8519addec9aac115c5edc8a2e14f9621c875d95f123d1291776ef4b
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:b780634fffe2257eb845f8c1d25498d9c654d0327c1f8841b40d90c176a2bfe0
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:04866e16881f64ff2b5edd379eab0731c9d66a47c25ebfe79c8d8ad33d8116e6
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:61fb98802490323bb03ff4a7060334d8fbda9e9f95cd6366b29b1b6494e4ef8b
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:3b7449287adf98f0cf0a1dc74e03b69218be3bcf5538327bc34ef776cc1c69c4
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:10788ca67809e5a2d76d6d72df7dd64ea4407773999707a7681e0cf1b5c4695f
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:216f58dc8c8166c0c69fc0312b667daad14cf39b7aeb7bc0fe1e313376f1a5fd
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:aea5f200a845e1d4dc718d76465fdbb1eeac2b20ecae224d5d20a98214ad0357
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:e5db11a0adb889f907cad797dad9cf2de09c433e5522fa333193cb35575802d8
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:c4cff5ad5b0ac79e71a16b88640a312ab320817041317f50caf575f7e12a864e
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:c2cb80bce34a74afa76151d2b44dc25ece73fb274cdfc6fdf243b1b0cb7d989f
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:3aab995872336e6ccb590ef6488c097cff38a91348055031145427a00aa6b732
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:b520305189c7567b2f8ac867b20468259fb22cd3edc5b4576af2bfcdb0c11855
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:8273842d8e1714fa03c7e0809924b1ffc5ef282f79b9606bd8ac88c03e7a16bd
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a10e76aa15707e708f82c8d9e8b5179bf67780d99f3c3e2e99b84ae2a0aa4ecb
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:6342358e6aed2339333ebf37fc223a41a0a3e199a0e4a840f3f66e5accf8f233
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:594c3ac24bf59802e2d3fa127e4051277bc07b6444f934f535fddeb668ccc144
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:4408d1fcf34c0ffe9327d4c9c0bc772afa969d831bc5de1097164d10b4e8d900
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:03a3d6f0314acf4da4f3d3d85f1e273a3ef25828d28f5b22f394441d55a645f3
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:7dd11d385b3d5a178566d901650ecd45e4af395faec2141d57b98af8fc1e5151
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:a4943c0563c45ceccfda0a01e26be3a01c1437c89d927714c577bf6060495958
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:fd86ea2a34de43014dea3ba2b8d705a3d94347d3a24a370a8bfd032ed239dc56
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:f37f494956adf0892ffca34d2a4c9b839ce28dc9f2779ca32e8801fb4f0e9105
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:557e73127326be051aa6257e7daeb9721e2af04d2f06cc26cfc7a1073666b989
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:fe916630003194dbcd3894dd880cad8b15a041a5e534cd8f76fde6296db00797
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:397eb425f383c1273c7976bea69a603530676a00a61f461ceb5f48d67901c4cd
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:bd134e64c156dc5c81d14522c9c9efcc1f56b2e972ac8ecb1bb1b130c76adc60
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:ab90ec9b3cd9203fdcdd7c43abf95f48c9d7479a9c7af479eb1e4b6599ba9642
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:c12c471902b8e2596588d9e553ad59b501d979dd04f301d45f7d45e794737a8c
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:de0634d14962bdb8701a8c29e80b7b70684824462bc5203e5fa7f139839cb73b
        - registry.redhat.io/rhel8/buildah@sha256:a1e5cc0fb334e333e5eab69689223e8bd1f0c060810d260603b26cf8c0da2023
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:0229a0a4f026f04f8fa0a1cd8e7afea10117fe3f12e21986ec6089769c8afcaf
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ad399867cbb712111db6affaf5c9e3573699ed0eb0def8ce6057de02054ced4f
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:3d7801f7ffe6420595779673a170235921a9090333f8ca51d5b7f664e19c82b2
        version: 1.29.1
      entries:
      - name: serverless-operator.v1.29.1
        version: 1.29.1
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.29
    - currentCSV: serverless-operator.v1.30.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2023-10-30T14:03:25Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.30.0 <1.30.2'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of a particular version of Knative
              Serving
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Eventing
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Kafka components
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:9a28cb1d4c4f4dd7d494829eeeb7122db2174a27747f915244672c95dc8f9d12
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:4800c85e8de031bf77f6fae0e4c8e6c1a9912ccdfb77799cc2ea574c37d870ac
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:66149df8549380673c24e1487d0c2c8eb638b7e502455671ca64ed6f32ff82c3
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:a0c5a91348ec71697747113990652feb024143fbbd0cdd67e8ae672730383a79
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:2d3e314a6584fc6974517c1a992fb59f5133f54c2f6cfaa42f579c0d332e0300
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:ce4b36c90955a356a5774410356aafdac2b60795526a3488b0b76ac2ec1a018c
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:1f7e65009168383cdcb6b2caa22fe4db49e03c71574973a8086f6c332dc5e26e
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:32445e4a73bc7c1e8007775adeb147175d934b17e7d3d4dcf422104dbe80a52b
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:c9353a022e6f883b7c0b2d87b1bc4653172feb13806ebf0aa075d6f9f43c264e
        - registry.redhat.io/rhel8/buildah@sha256:a1e5cc0fb334e333e5eab69689223e8bd1f0c060810d260603b26cf8c0da2023
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:92a39ca90984cebb1938447a31c9170b52994ae98f83caa860acdb2c3ebb75c7
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:06e9e073f092dce67d4227e282fd8eea6dc9f38b90a8d444cf448d80d027b616
        - registry.redhat.io/openshift-serverless-1-tech-preview/eventing-istio-controller-rhel8@sha256:b6ef281fca2b80d379b1e28fe3f23119df8ff765fb2cca291d514053fb512da6
        - registry.redhat.io/ubi8/python-39@sha256:0ff58549041d32b685cc850aaa85da32d1ef73f99cd4b6e0f6882dad5f0399bf
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:21a638b481639c3c58e8b026d875917c30a5874b08fa4ceac40d447076b2bfa4
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:124087e6402caa375b24b976c3b9fd7d53890e20ac3b0b57ba3ac646e352c7ce
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:16939b802f4c16a3d9d941917d0363325f40cdef28f6e1a982bc95401117674c
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:e9505aba7c44cffc3f261a5dc7617a0e54ed98e78f73ea3c9c7a508560b792ed
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:e8c756b659f6187790bc1a5115e0f71791bff96630859a905bc5b75d14596e0e
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:a9c95fc208cc03625cf49bcebdaac73da823d1fa3bfae4239d7db20d3a0b8723
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:3fb6e7626d2b0c92c2a2fef35e9b8171382b4bdf54393f2ab1281ae7ab441761
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/ocp-tools-4-tech-preview/source-to-image-rhel8@sha256:98d8cb3a255641ca6a1bce854e5e2460c20de9fb9b28e3cc67eb459f122873dd
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:12bc62246142a605f4efed0f6bb8843aa2df7699cc1c5ab53debb94b0682e9d1
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:3c8bf480dbb05e09e8d07cab12e4736643f61bf9b5d98efa6bc0557eb31887bc
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:1455606c8edf611d996a48e29ead53e0bb4078f55384b7689058bb25fc59fb2c
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:6057a2d18406e00b0b7b1731870f707844e39193d4cce6de50137b77f17d220b
        - registry.redhat.io/ubi8/nodejs-16@sha256:8abeddc6ccfffbe313f87e5e069e18cb7047f2c7880ad63cf15bcd0a79ba71ff
        - registry.redhat.io/ubi8/openjdk-17@sha256:af305e087061cbaa4107ce0b304a1ef987f6b54bae3f75bfd529cfd084cba9b5
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:2264d2cd4ee2a720667d625ff42b6f8be4f2cd65b57a91eb6044da086a110740
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:69ea22c646c09208ca68345d4b3508453d8c49f7f7fd9a4ff4c4cb78be7bff43
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:a5154f9d13a1c373b8ba47e2bcfb9d4474fd03e08db4134c92aa4f049536f1a8
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:96d6f52711ba433ab2f273be8080840a88e626e989027480424d19c2fe1e9454
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ef116868c9c60e5ee246060688da49832571ff6a7901cfe143a42d048743736e
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:01c8fee92889a84a19136b0843e4fed5db31c1f2a6d997669f103373fd995e36
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:5c933641efc6bf919b1945be5a01dfb7774ae865975b1bb1f44653b3e4ca9f1f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d03f2bf8f5e7cba47102ba26ea0149ae6719076c645d5e451058392d6044b738
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:3508434d5ab6b38c8a7466256f1158ebcd75e8b751afcc5255da8c21566804f0
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:cd1abeb2067afa44a0f1eca301cce37307582d56c1147820ff86b9b64a40588b
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:85a369d5696771d0932b61760336e52ae3f7c66f9a3def5312aefcac03ba516e
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:96c135ced2e3bff434addb012368d96d0778ef1d294ba507ae7e624811e20d46
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:e97eb2bb7eb529cbfe67c14e823ffa1269d614e2d35723896e89efa12a499c6a
        version: 1.30.2
      entries:
      - name: serverless-operator.v1.30.2
        version: 1.30.2
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.30
    - currentCSV: serverless-operator.v1.31.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-02-01T19:15:42Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.31.0 <1.31.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:9221a833ecc66fe2354aee7aaec2a2470b21e2d89cb6ed287ece32edab2a2edb
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:22ffe08d6a0d5461158613f2b2a2ac6936025357e282a09197a750c42b5e6922
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:3dab790fa845ed7eda088e5abdc87799da1354884a53ff1ff728739b8e278418
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:1ee675125f7141e91f4ee57ebce082967834bbdabe1a870c221f31aeb8750be4
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:486bf60d17a5d265ebd65d590a6ee1a7e13bf3953ea797b2f6c2b5910789858f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:bf693eaa845373e1a06041626a9314ec00519a6fd2437b12f7f4f3f104864039
        - registry.redhat.io/ubi8/openjdk-17@sha256:a8165bc2cd5051a96d6937e25ed03155bbd4b731da6e58cebfe2ea83209c16d8
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:565becc77dbe3bf03d89b1399d6e1853253276a71e80b6b14069285d03cb14b2
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:433821e6c7f43da5ca2e38c0550722feb179a0bab021decd9a97ad020cb78b76
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:e49220eaf3143a8958f0cb1f22115b2e834fb7177755433bdc592b753a54a8ec
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:1fcecc11b55007b9c6969a24940a391968378e4102eebf7ceabbe7efda138575
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:b37dd2cf4b459fecb6d681cac36918e486c3bd22501df50b90e20b15fe164d56
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:b2480154951109a159acfb88c5472a312a515d6a6b060d832b3f7bfc16f90ac5
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:887480662b1bfe09626fbca8ec7c5da879d097c46dca11764ea98474ec78ab5d
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:cda04e3b7c2c1f37164733ad162ea0c4e52f14fdc8eb8bdef8507f0ed5454640
        - registry.redhat.io/ubi8/nodejs-16@sha256:f90b28179d1d5de68d42ef1a9166d70a9dfea85a1c6a4bf12dce2c59eeb2fbb1
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:677be039353b2cee51d5faa5d5a2abd6a7bd46992ffe9521eb033eea9925c785
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:3181ad775ead55f317f96466c2719b2b73c5b35d31933733da21273955d160c4
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:9e120ecdbea54eecbc5e0412c1774f8a2ee70d7fca682ecd0757b8627f70680b
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:eacdad8db61b23634c3f89bffa3f1c30ac5c7bbb5e938c6208003ff00f9595f1
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:df1aee67a161fa19d7f49ba56d53e0682cc1996436e985a12a2068fa419b54ea
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:3b01ebdb0704dc6a1a993ac076819489495ab3fe32e3c14609e42d1358856be2
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:d8bdd96c6a9b31f48c961ca438c254511db785fd34bd10b98084b702baf0ebab
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:63e08279d2852d494d4f8836c54e2ce942a342f452a23ef437df9640bb2f5bc9
        - registry.redhat.io/openshift-serverless-1-tech-preview/eventing-istio-controller-rhel8@sha256:5b38e5c71e422161e40f9c37dd2f46212cb80e8e37b00247767276b957080397
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:80361e052d366ffffe9a99893cd83965bd69a9ff81d8c9262976de145bbc6184
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:d2a34100f6309c3cbac7c9ccd452b7bdab2a533afc99463bfbf91e3ba2706d54
        - registry.redhat.io/rhel8/buildah@sha256:b48f410efa0ff8ab0db6ead420a5d8d866d64af846fece5efb185230d7ecf591
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:0a2989f688b966277652957b312df736ada43fff78fd89b17f60b2137e38f115
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c078071cd51c7fd5584039b54a2004331c05a897ee8b0689bfcb9204a8ebae80
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:8353ccb4f0d70597d61e4fd1f799d4bb74e5bc038f6b9916059e712aa70b97c1
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:3be4fc173b489d0beb768739f60274e07d95ffee86d70a8423e1f59678c805ea
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:3a4e7412322755b4c278540c965a403f9e178f138bd4ad7929c1c24a8036dd14
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:de62762efa0df95ed23df56f5520490d8a231e47ce13c433c512254ceff81844
        - registry.redhat.io/ubi8/python-39@sha256:4f35cbcc6f4e108c69f7d2b70cce5afb138741b0c961506376d524e001b922f4
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:99f5dcb92a33c047f0b68236a3a42ffaec78037dd0a5eb79d1144d62d6bfcf43
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:2572d98363cca2dd94984d6cefa573c5054fc206ee70048f901ad4b846d8dd52
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:dbd9c0cac86f79b8d91ab65ae7543854144781d4cb17ef16183931ec9bc234a7
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:7941b421b3e18bda8c565305241621fcdf1212cbbe0dc02f38f4ab0565d2bfcb
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:af5c30fc9f631d9ed110641b815315d8b4aef539f687277a3821d9939ca5dd46
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:ef7d28580e75d746449a90cc0e5343ef3f4ddd75bd9fb866d97836bfbc3b314d
        version: 1.31.1
      entries:
      - name: serverless-operator.v1.31.1
        version: 1.31.1
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.31
    - currentCSV: serverless-operator.v1.32.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-08-01T10:56:58Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.32.0 <1.32.2'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:16c026f09b377c44253cdb902b9a5fa0ae766fd0185cca1863798a212e0daee1
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:2f0f84767f7d76dbbf4387ca2e272766267a2c46b5bc29f5d01f05dfa39cb944
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:20566173a8e907980e603d82fae4d96fcdfd5c43d9004c2e506cf0e09a47b91e
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:04e5b3882223d45a5f5383ce6481efb7b2cbedea57ebd894caafb1454ad8c09d
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:e1e2273fdf6384d95e1128e54c0794d7d620291a272bfb1f91e0c2723be7f1b7
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:22b4d47725ed8cdd663ee26a03f31198fff9e71dc703532e70c42bdfb4840811
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:e7e6cbb46720eeca3f0f4b9829dfeeb3db373a83c89cd808b667ded4d5e139a4
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:a5a7b2be50f211d4d9d555b2a2e2302096c139ff5011d984cc386f1530007f16
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:0729de0f3f03ed220216c6cb718fc34beece0ceacd926b5e2480bdfcf44df5cd
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:ab9a1b952e9603b4ecc5f197715a1d871abc25ebb0c5e6470d4f0b365ae94276
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:2a30ae808d5c32d812406ff4ac9b472e82b6bb0555e62eee4c2c6a972f8750c2
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:b008da825ee2429f04cf39258389740c68056f82cf4bde0cbd3db850b6256885
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:5930caa25ea875d29dff0ec5e235eaa13cb6dc98980d8045335eba1a09ae761d
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:da5f813dea324ffe9679524ea7f3801536fa48f7dbbd8da56a8492b916b15fc3
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:990962ad245a6d62cd11551bd4e1f8b14dc08abbebe47beddb2f4874f78d4f47
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:b0db9f0fc85fcdcd5da980633a3a44be2bbabae068f89d0965263de38789e812
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:9d36cde68bd822016e967148eebd2ef3e7ffab9dcdd37d17f33017e9b703d9b2
        - registry.redhat.io/ubi8/openjdk-17@sha256:a166bd2fedb99cb62ebf4695ab46e19c7e451f887940cdf0b8245c81e403aea2
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:9b6524f9c0e7f78e6ffb372a8587eaf5f4f1281eaf64f93dd00b63be84a4c21e
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:42f31eb8f12dd36541c8c0b6572d942c62e2064d5418428482d676d117a6c426
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:7c8cefd5823ad953f112b905c651f12a97491aafc008b2ba0111f0c9e83e50ae
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:3be20d63d7e82f35ac88274532d6ab7839f9a7306ca272e7d532ac03215faea9
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:46b8dc1c8ea6c40401d4d1a9868ab3edbe8b9e92364a43add6722489e2eefa04
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:ab09a1710aa2331e24df9e7283f3909fdb104c1b314dea88e919d0a396f334d3
        - registry.redhat.io/ubi8/openjdk-21@sha256:bb95ea866269ac8652f666ee1587d71f4f43da33493c2be7f6a9548bcfd2407c
        - registry.redhat.io/ubi8/nodejs-16@sha256:69439fe83fa957089cb6fd3c1042ab71d719f81dd457a5d1fbc56c1e3ea600da
        - registry.redhat.io/ubi8/python-39@sha256:8c96e8648cc6651d0cefbd20496a89f208621601f9664554d3fdeef8ced25453
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:e55c70591c0498b6b85f0d68e304e88b830ea0b35219ba8881143555757be4b7
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:25cbdf78b67e7f5a360313ccfbe3b6e2c98873b4e9ddee46c115a623e77c7abb
        - registry.redhat.io/rhel8/buildah@sha256:55f5c7d88c77e1ab5a2ab8e877172b2b66cb196e21f10dc45148470d0ee8bbae
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:1a1de677297525c0c5d317fe8a7a42684fe3f1b422a7d6bea1f70f0924653b08
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:0507ea8745267fdb53329a5274d59d23c387225d2594af51dd2603b4ad99db0f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:4c9faea6fd162db3340aee8c11af11a08c9d79bbed104967c25df26c4933c8ae
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:14ca32e999767776c3e585ff5eb1aa568f7d76fd3e5ae24cae05ce7564678080
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:8033b7c1e903ba313134645c626eb3bf7ab1a0d68dc2575bc8e26884ad4f805f
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:e70dd3a9e067075eb56e5f2772642545e308348cd49b89357c02de3979b1a5be
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:3ffcf746df914f0949477b46f90cb2992bb63f8198d73ffd54208e7a777a9cb1
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:1befc9969edd2062b3df69f45aee9771b33a2659c3a14db0aaaf5c3df985af58
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:0551f1657dc158bd41eb2338dd6514d55c39fc77300cdf1744d8f5ab14e9ed1b
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:4933b76d95d411d597d4dc460fc46c92875d28e107cda661946206a9d934b8b5
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:70c26c7ac48d1eaaf1f95ca5c5b8ba8ee6b047a3c9c4a2c2db9c49a506bb2205
        version: 1.32.2
      entries:
      - name: serverless-operator.v1.32.2
        version: 1.32.2
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.32
    - currentCSV: serverless-operator.v1.33.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-11-20T08:38:34Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.33.0 <1.33.3'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:e1cb1c24333605e52364456f9bd8d86d787d1bae17ef75dd04be61257fdb2256
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:6dbfa2573a2c78e450d3c1c04c6e7cbc50d9272233e828fe9b42e575ecf386c0
        - registry.redhat.io/ubi8/nodejs-20-minimal@sha256:acee3b00ccf8ff49bcebcf10e028e34a84d45f28f855363bafb3b93c75e84f11
        - registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:e1cb1c24333605e52364456f9bd8d86d787d1bae17ef75dd04be61257fdb2256
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:eb95deb2105f14c6a9f3bb2cdd7306b446109a9b142a7bfc726b6429f270d9e5
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:ea9e4de6228d5c3ed494c2ce5317ec290543ad40bcf7c43c2a3d6565e5d94e5d
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:359f8b9fbb4c3c2f79e8c70f3b50c994693b930f6f078c34c87112ebb93037e3
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:c3dee2598088bda36ef474f4748943a29169ed9589aeb0f77bc79a7f6bef9907
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:672022b6f2a080483d9d6f6f4a5f42e52e522a5797105f3465731809c6cf4947
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:28c9d05a3ba73b76f293e8f4902b574deb44f77a2bae8cf92473b8a8b10cee92
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:a0399a993155e755409355740b260d9b0019559a62be3b6253d8693a4fb59286
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:42d16e4d9f14a427e7a5b7e3a4ea07c258e0f2c74984baeee3b4926a8c5e2640
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:a8821b59d495e24598b27c6510b32ef19de2bb04ea1e568fc5afb9135d772036
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:9b850afa750a67120f68a49f4834ba7f105563df22cbcbc563317fe1545640d1
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:09060de091d617bd4606f7a289b4528265ae0b7a04737f383e79efd5b0d3eb79
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:dd1f7c850f7cf2a4d70403deedd8a57ecc6e0ee5455d68fdfc49ff175fe42dde
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:93b2f5211e79abc38bf92f0b8fb8adf046303efa2fe62c47a2fef9dfb5662f55
        - registry.redhat.io/ubi8/openjdk-21@sha256:095175ef8573de8055e67bd5c8a21098448ac73db4e0d6ccd5269dd36b4d263c
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:b871bd08b782c9168e9a54340efffc0b0effadde8079915b848e4eff72bb3e95
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:879fca768b46676f1df89de1cdc4ded8dac310209c146c2f381ed1deb19b06be
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:993597ff83a4d62a23f642fadc92f182944b953770716b73b7d19492b5a462c3
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:eab4aa93b9d175045b1f01bff09681107d839285e0d633cd5146d2a2efab373c
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:af66cc7dbed88907d96a390fe5c8807a2a60bbe3276b8cd13b4b061cbaf7b162
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:b12272e1a9cb4f4bb62c4badc62b0bab110abb99484612dc781c46bfeac13cc0
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:b2be693c71cfc4ce934d5b01efb863dc6e391f93a6b15fe7eeb15914063a8294
        - registry.redhat.io/openshift-serverless-1-tech-preview/backstage-plugins-eventmesh-rhel8@sha256:28ae309e4cf7feabaf78ac6067e1173f9f5b33a71c06d6c27cf25e01a0b1ed64
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:05e3dd6255294b691783523f7722cfaf13aca1a1e1dab344a25eeea72aaf9f9e
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:e9b38f1df225238836aa829f3a2016372859fb8709cdd2b30496498669a90be1
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:510e98837a549f489b2d74939f3b9c7ed90a8260ba914910f681d5c0dec7655c
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:d8ae7bc888f11121fbbba65074389b879bc22453bbabc551d79d0cee68fc33a3
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:5c7f29554d04eb9348932953344f2ea0806d14e7704355619381bf395d993f58
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:76a47d100de1f2afa23803bec503d9440605db03c7179b231f963e5663a3d12a
        - registry.redhat.io/ubi8/python-39@sha256:ac12e4706e0fdc464e791a3f274474ac1bb14d5af8f217c440fbb0c76d6a5d00
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:ace72435e3e63ab21dff2345f958cf6bd171537ef2846c5d5d91cf821877140d
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:9b94875e0e5d7abdb3b9ba7b04e217a2a9b7b4410d38d1608c86d1c21f929cc1
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:56c6ac59a68976741d2ba676a53e59fc494d04dc44b513d72fdf8a9ad894a54d
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:62bf4bee0ea17ecc60366ecc71431adc79d6f0a5b9f59489913b474128c4adb0
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ccbda564735555785815f27f8f25ad9c6c27dabe4e262586fc51609331692527
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:a9c84ac7e7c3fb1e5dedaf949d54a3aee8a62397d2bd6022094671de00b651da
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:2bd241898bd54dcb248ae86470430042dc545d45e7519ed470a1a9c8c63bdc37
        version: 1.33.3
      entries:
      - name: serverless-operator.v1.33.3
        version: 1.33.3
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.33
    - currentCSV: serverless-operator.v1.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-11-20T08:53:53Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.34.0 <1.34.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:a4b3d75903de1929326041e9885213074473f07fda91b299548d45680c295c00
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi8/python-39@sha256:ac12e4706e0fdc464e791a3f274474ac1bb14d5af8f217c440fbb0c76d6a5d00
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:633ea30ea3b6d5152bbdb61acb4271e5905d50fbe3c731fb4c8d8640a57f1970
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:586fc8ed88f91298dbff31bc2fc4c63e0d6d9356e3358b2f853f49511c3ff9bf
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:d8b94bbce98325fe8079c11c0b7a6c59bad8a7edd0532a7d27ef1632ff0de708
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:ef7f34a025e40f93fbd51f3a4596998a6dc508db824f137ffa1fbaf90e88f4a8
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:2adb0124f04d6c4a552ade79dd1499c8ca997edb45855bf438ca8043b2709e19
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:f94f0e8d21a6b6341b1db00513a1a2ed5cd84fcd898a376ecde942e940a518e7
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:01ea7a06332d60ac16ae0a27adcaaede3f680a98a22ff995259a6c42ac0e358d
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:b07010595201d7cfe0ec4f18e612d323eb38464c74eb187f275617a744d9df12
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:144883b09d8a999fd4d2a109d074e44d7d8b959987512379aa71c71e4060ca28
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:a9e8583c19ca3e01593f8c0e54731cb73df9b46b2ed47a2b1c711828c2295f82
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:76a47d100de1f2afa23803bec503d9440605db03c7179b231f963e5663a3d12a
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:6d042c614431aa9a9e77cf8ba0064423e903f7c418e367ecc23bc02a24a796c0
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:86ffb8be1762782d643c147cefcfbc2b4723b0f97d017314a99bd2275d05a1cf
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:30cb41f09067959cfe3792e070fca62820fcca4035683e5c43dd46d1cce97c01
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:45bc50a993d04b775a153e985a43d587d5455883fc4c6f31b772408297cf9040
        - registry.redhat.io/openshift-serverless-1-tech-preview/backstage-plugins-eventmesh-rhel8@sha256:09dd94090eea9cac7a1935a6d671cb97fd9fd5714c715c9d2aed461e468ca54a
        - registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:a4b3d75903de1929326041e9885213074473f07fda91b299548d45680c295c00
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:946daeff59b41a6f8c4842f8c2668f97ff1b49e2a8afae5d64ba3baa916b1d95
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:75228e70fb1383bbdbcc64d0cf24bf5d977c09112e3f32316824188264c3a2bc
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:33d62d13e16f47db9951ac826671f39fdc048fe7c3fb019eeb4bf9bd0f814222
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:d5ad3c9db7ce8e4fc950261f7c388476214a03d92e0043c7bbc4f0add2b1e69e
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:b9d85323e110d8eb6600223c38f9d4cbe89b50d59e03bf1e6b8808f7cef9a3a5
        - registry.redhat.io/ubi8/nodejs-20-minimal@sha256:acee3b00ccf8ff49bcebcf10e028e34a84d45f28f855363bafb3b93c75e84f11
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:512c4d93dc13cc53d923a6990770a7527d6fb657278e2c5d8f0ece3cd5020155
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:0f126893074cd0398654c4018df6e882b7610e148c82a41ac814f4bf3830bf9d
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:a5950f81cbc64d1edb64ae7f7f120c3435d36afcd14ffc34ae522af297d87e1e
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:3d17f12cc9da578549254ba1b8934b5afc66e453b1d5d667bb90281e45b8795c
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:c0376504edae12e7af6b6cce02ac53b6b374f5058ebefcbdbd32628b551bb0c4
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:1c82d20ccf5eb5f8e7725839f1060f4efdfd1c1b4773a99bd905e690a6e36f38
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:1252960c0c0a4dc7b39090cde0fd5742a19453b1a65a08b737bc41d044a0882a
        - registry.redhat.io/ubi8/openjdk-21@sha256:095175ef8573de8055e67bd5c8a21098448ac73db4e0d6ccd5269dd36b4d263c
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:d40daf67b36387faa9f55e70e50d10d02accfd19470c0ac11b78bf969947fed1
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:84994eabeb7342a0771c7cb1bdd70e7ee3928dd45874eafa64cde5bb385fe964
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:0f8d006d0e7468783739942a1c82c7aec376c9948f23841132c8ddece8ecba06
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:b801ed7b1cb32d4a1522a959e97f280f2eff6fc92e47fea2ea022f180d37831a
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:e520d1b6b87f85dfffab5e53cfb70562ab7edf450061c07609da39e3b2498fdd
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:4b6f161c14c68536ec2c7679d30681afa9a1fa2d512548f0f591791a10a8b894
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:76480201188dd142206040af845bd55b44155af92e0c5ee9bcdb8660bea0cdb8
        version: 1.34.1
      entries:
      - name: serverless-operator.v1.34.1
        version: 1.34.1
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.34
    - currentCSV: serverless-operator.v1.35.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.35.0 <1.35.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:f23d3ab058667c6fe9ee44dc249aa7e8449c1bb9581380ddbc66624511b286d2
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:b2de6b658a1595084a202203c90704c3f9de018a7a904ca276112ad606e374f1
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:0751b9ea7ca197480c4cbca17654d0653d1c0c760ca3c818c2a1018dc952cdc8
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:da8d975527ba100313b92e87d6d2690fddae9988aa8d82d728221295955b7224
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:ab8f85bb18311b0bfee00561ef6b5d9faf810358c954ac2c6f39cf2a7739e07c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:6b842b4eddfc71a2c8f6c1fcebb8f23a05f50c95be1ba2d20a1be660d3014c8e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:e743dd93b7ae855825716403979db3f34cd300caad130e7d3011172b1aa4eaf7
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:0e58edf5ee340fdfb282faeb6c9bee8cd6323e7f29a9b282dbb91c2622131a9c
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:477b1ac0db0abf1e16ae31bb22fdb73dd98f4dbd243fa0a7e6e94852b16a813c
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:b66bb9531f05cd46360f787955c849208496e3e45541cb50b23bfb6b68570aa8
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:9b176477be4c782ce60d81a72bacb566069a499579ce6df5a987a0841f05ad5c
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:86386b54cc10743023884193b7d6c7159b90052bc5e4a0608a6cdfd47865c966
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:cb2ebe98481808e788ca70e7f0a9b3402818b622158f327b99a1e87c23226d2d
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:d0650e4d7e72241e07a3796f20931adba231fbf5e401a45d794761f2c5afae48
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:2cbe84afebac23b5397d144b4d5660af33290df451141fa3625607b6033983e0
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:020264063c50f2a4590f3d7524b271f0f77350a1f8fa3c35ca6e28f0162aa9a5
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:d00a0059f6134b2519f9b1e034b81a943cae7ce391109703f24db21e56ee923d
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:d68fae6fc07efa9da9af0532aadb49f897a15afcdf9a98dc7517ff5e17735454
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:3130a5159adb9b4eaab75a0209d2ac639156d5e9f2346a8b2e6e0e22fad4b45f
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:80a155d2637f47e249fd69760b250c9b7a1eeb6781de9308b54162bb3e1bce71
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:ffaad2214fa897e89e92e574eb5cd1c671edab7f6eab87cf2ffd4ebf162b3cf4
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:b5db9a51bbb46640b23712c61e9c5a1c7bbdbcd9a53b0b92fa36b80dcaffffc0
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:f23d3ab058667c6fe9ee44dc249aa7e8449c1bb9581380ddbc66624511b286d2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:56e7a0158a713c70765ead847732db4e5f8439aacf31e9a4a22672b583028064
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:e7aa54986c3d388c1bb00828c9ccb294ca0b6c26376c7593ed6bfddda0ef4e5d
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:ab0342b325ee67cc2a7d321aa1eda9c6797dc91fe2b6f26e3cbab53fadf564c7
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:732827652b7a7a49181e86d561831001c28b72b7e12d5e3c1d71b2e496003229
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:018459cbba2e3169bc4be6564cf3747c9cd0677445912abf237f8fe6d58e46ac
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:6f4fa254d923f3260ab0f7d6a0b459554f83b1abfcb575a5ce2c0135fba32427
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:6722aecbbbefa80dfaf93143e033c562cea0982514e2eea27654dd4638e1d072
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:2980eca437e4c1cd22a2b6c2ccae9c4d6b55921b3d575fbd4b9241f1f0a8a3a4
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:dfcd11686f161d44786a205d12f9d628d9e847d40ffdb628ec9aa4a0da0b86f0
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:acb28d84cf03b8551efa32646864005dc3bc9bb341a696544892dfd0c87fbda6
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:818dc7bd69b453ba5882b4ccb82841f2d48c7552368417bac20bef1df64f52ff
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:4ca1bfaf2d7e9477583cddfa40f0b2ce59c20ff9ab20022038774dc3cc49662b
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:f56d21089566330ecbd9aa8acf14da332d3db7360aaf6da00c9a19df83f7334c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:8a4e5e0934197c32e0dfa1efc922d750940587fb61e0df49d476a2d035c53413
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:86074674b7926ad9bfc95556e50820792fcbc5e11ab02879378625ffabdd1889
        version: 1.35.1
      entries:
      - name: serverless-operator.v1.35.1
        version: 1.35.1
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.35
    - currentCSV: serverless-operator.v1.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.36.0 <1.36.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-ddb-streams-source-rhel8@sha256:a903480159ebf1d8652eeaa702d9885fb76bb106c02d75cec5c82c31d77d6410
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-timer-source-rhel8@sha256:8fed5ef9a1f39a55077bb5026c8854d0ee7fc8d8ff5da277ce9c1bfc2ab2474b
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:77e4689b2ce7a713e905553e0e42b8e128045f40cb6c6d4ee082f126cbc46b00
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:3d780ad1ab280b5092b90e277485d589da2a18e8b08090af7d95aecd17af545d
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:429d5d9568cca31fbe83c663276a300286af4cdbf3d35685afdbe5f6f2f8c6e0
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:4344716eb2940c6a3fa9a278250f936602bb1549fece6037912837719c178b24
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:f2d112ba9fdfccea988589b99b463c8a7157475f7a3f03678e8999e65b678cc3
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:5352e4eb4e426a41521276decadf204b6d23b4762a298e4d3dc73969e3eddc68
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:187fb9d64b53c55a029ffce3c0f9e7c3304683fa8a0e832cfb7134e8a3d0c237
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:6210a946f373b9c36c9b31293debe6fed090008b8234c0ba17d37fa81646172e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-transform-jsonata-rhel8@sha256:3af4c48030263eb5769e5f36a6dd13fd390d8566efaab37532f39e1df1c95c67
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-sink-rhel8@sha256:d9b944e64391a996104afbe8a21f66316a948e9fb7b91734e861847285be59c7
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:631eb3a8dd9d83ceafa7d8a4da4b1e2d4f2740e6017c766828dcc7575cf1322f
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:a6f565d73375e827d830dca85838f4c4cb7f08b0c078acb0c0d567b25f0d256a
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:a47b6b5237a87ccd90e5a49a53d06485fd55c87eebf12d01c335b87fc3a8c494
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:ed60e889a677fb139f2eb4bd9ab4fda6ecc39fd050ca2b6cc1c19717e428535a
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:074ebfc4494581611647ba90049655a470ca43743610a27be77db66a8633bac6
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:dcfe5fbc57b9f8520adbff368c8bf29eb491c9c26d1a1737a1722cffea83eb1b
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:f3d87a7fa5b5f2818effcac3304bd56eddcdb0db97a18ac0f35a06fc7e223e7d
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:4274004b6289fa9bea3fa6e56d64077a328a6b5e63a9f0e9c41f6d561f682bd7
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-source-rhel8@sha256:27e74287774cda208f667bc2a386debe6b6255e8991abfb37c959c8525f2af53
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:3b04e9d8cd4b5e43a07ece838d10478723a6c6529350f010550b4a0d1d0950d1
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:ee6533b71cf280877c9d7489dfd821c9209c6802bed7fa3fdb4f8f00fbeacb09
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:6b7316cce3432d391adc6ac38c92ff0ee86dd7ae0101957cc65cd338cc69e352
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:c39cb88afaddc295d93a3774b8a20d001b82392df540fb502ee02316d82460a9
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:82e69a1890042f5636ac7bb645b1aaeaba0a4894851eccc1fa1c63296918e31a
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:52ba6b1252b6f3d41e2a0b3b880cb768010ab9dd925ac34596f6adaedcbbe07c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-sink-rhel8@sha256:afa634e029bcf79a972f49372498d44b45bcfbdb6eb28ecaf4f0b0b2a0a1824b
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:f4d86b196457cdeebdc380b6f1b707c10f8dee789c7779c90ee1332e511af8c4
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:259e5dab50d36cdbfc117a42325128c4e16f34c048ba64b36d4a89132ddb4c1a
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-log-sink-rhel8@sha256:30b8045ef5f925a36c58941386f1899bec125d005a08aeeeba5edd49b0859149
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:3fe326ad9a00b112fc888ffd345c8d5042f0a22cf32ea575931192c39ce1fece
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:e3ae1cc6012f19631c012a07b113fb015977d2ee4f9faab3207fbc2464c25aa7
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:bf82b37ab8fd29c7a7cd846008a16d14b4cfdc1c09df4f731d71be3957e90123
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:d6f338e08fa937c8a4ccfec692424de4acded565f7ca3971ea507dffeb351503
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:4aeb68a22f63300e54391ec6b4cda8c4e665d904b5dcdd3c40155040525742a0
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:596a7272d6de714d7dbc6e00e9e50a2872846c269547465252d953172ef0b009
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:0c2cdf917bbad1f5bca8a2a2ef12a96cfcd0c77c190cc35687eb57169b54ab6f
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:c7c2e738e0d9da1d2d1993f68938e72c4709b1f28a54fa92224ef6a0e456b0aa
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:a1834b620f875681388f0ae224dd0771773a8b61b1d8e5214a942f232854b87b
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:9b328fff0cc8545fb9bb1f24f11d78cfac3ee76dedc11fa4def4262d79a60e8c
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:78fee4c596f2848487ca704da3fa1074e3c9c99b3f96f66833aeba63836d7b93
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sns-sink-rhel8@sha256:51b5e18e8d4c1ca81d7c5ddfea6352a3887f757acfcc53bd2f8bd3b01337ed42
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-source-rhel8@sha256:88323da16360981f27b8e4686fbbb64053e4e0e3c79c65d695edb317b334e1e0
        version: 1.36.1
      entries:
      - name: serverless-operator.v1.36.1
        version: 1.36.1
      - name: serverless-operator.v1.36.0
        version: 1.36.0
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.36
    defaultChannel: stable
    packageName: serverless-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-pipes-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-pipes-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pipes.services.k8s.aws/v1alpha1",
                "kind": "Pipe",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/pipes-controller:1.1.1
          createdAt: "2025-09-19T21:15:33Z"
          description: AWS EventBridge Pipes controller is a service controller for
            managing Pipes resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pipe represents the state of an AWS pipes Pipe resource.
            displayName: Pipe
            kind: Pipe
            name: pipes.pipes.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EventBridge Pipes resources in AWS from within your Kubernetes cluster.

          **About Amazon EventBridge Pipes**

          Amazon EventBridge Pipes connects sources to targets. Pipes are intended for point-to-point integrations between supported [sources](https://docs.aws.amazon.com/eventbridge/latest/userguide/eb-pipes-event-source.html) and [targets](https://docs.aws.amazon.com/eventbridge/latest/userguide/eb-pipes-event-target.html), with support for advanced transformations and [enrichment](https://docs.aws.amazon.com/eventbridge/latest/userguide/pipes-enrichment.html). It reduces the need for specialized knowledge and integration code when developing event-driven architectures, fostering consistency across your company’s applications. To set up a pipe, you choose the source, add optional filtering, define optional enrichment, and choose the target for the event data.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EventBridge Pipes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pipes
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EventBridge Pipes Developer Resources
          url: https://aws.amazon.com/eventbridge/pipes/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: pipes maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/pipes-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-pipes-controller.v1.1.1
        version: 1.1.1
      - name: ack-pipes-controller.v1.1.0
        version: 1.1.0
      - name: ack-pipes-controller.v1.0.25
        version: 1.0.25
      - name: ack-pipes-controller.v1.0.24
        version: 1.0.24
      - name: ack-pipes-controller.v1.0.23
        version: 1.0.23
      - name: ack-pipes-controller.v1.0.22
        version: 1.0.22
      - name: ack-pipes-controller.v1.0.21
        version: 1.0.21
      - name: ack-pipes-controller.v1.0.20
        version: 1.0.20
      - name: ack-pipes-controller.v1.0.18
        version: 1.0.18
      - name: ack-pipes-controller.v1.0.17
        version: 1.0.17
      - name: ack-pipes-controller.v1.0.16
        version: 1.0.16
      - name: ack-pipes-controller.v1.0.15
        version: 1.0.15
      - name: ack-pipes-controller.v1.0.14
        version: 1.0.14
      - name: ack-pipes-controller.v1.0.13
        version: 1.0.13
      - name: ack-pipes-controller.v1.0.12
        version: 1.0.12
      - name: ack-pipes-controller.v1.0.11
        version: 1.0.11
      - name: ack-pipes-controller.v1.0.10
        version: 1.0.10
      - name: ack-pipes-controller.v1.0.9
        version: 1.0.9
      - name: ack-pipes-controller.v1.0.8
        version: 1.0.8
      - name: ack-pipes-controller.v1.0.6
        version: 1.0.6
      - name: ack-pipes-controller.v1.0.3
        version: 1.0.3
      - name: ack-pipes-controller.v1.0.2
        version: 1.0.2
      - name: ack-pipes-controller.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-pipes-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: PingCAP
      provider-url: ""
    name: tidb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tidb-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pingcap.com/v1alpha1",
                "kind": "TidbCluster",
                "metadata": {
                  "name": "basic"
                },
                "spec": {
                  "version": "v5.4.0",
                  "timezone": "UTC",
                  "pvReclaimPolicy": "Delete",
                  "discovery": {},
                  "pd": {
                    "baseImage": "pingcap/pd",
                    "replicas": 1,
                    "requests": {
                      "storage": "1Gi"
                    },
                    "config": {}
                  },
                  "tikv": {
                    "baseImage": "pingcap/tikv",
                    "replicas": 1,
                    "requests": {
                      "storage": "1Gi"
                    },
                    "config": {
                      "storage": {
                        "reserve-space": "0MB"
                      }
                    }
                  },
                  "tidb": {
                    "baseImage": "pingcap/tidb",
                    "replicas": 1,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "config": {}
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: pingcap/tidb-operator:v1.3.1
          createdAt: 2021-2-15T07-18-39Z
          description: TiDB Operator manages TiDB clusters on Kubernetes and automates
            tasks related to operating a TiDB cluster. It makes TiDB a truly cloud-native
            database.
          repository: https://github.com/pingcap/tidb-operator
          support: KanShiori
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup
            displayName: Backup
            kind: Backup
            name: backups.pingcap.com
            version: v1alpha1
          - description: Backup Schedule
            displayName: BackupSchedule
            kind: BackupSchedule
            name: backupschedules.pingcap.com
            version: v1alpha1
          - description: DM Cluster
            displayName: DMCluster
            kind: DMCluster
            name: dmclusters.pingcap.com
            version: v1alpha1
          - description: Restore
            displayName: Restore
            kind: Restore
            name: restores.pingcap.com
            version: v1alpha1
          - description: Tidb Cluster Auto Scaler
            displayName: TidbClusterAutoScaler
            kind: TidbClusterAutoScaler
            name: tidbclusterautoscalers.pingcap.com
            version: v1alpha1
          - description: Tidb Cluster
            displayName: TidbCluster
            kind: TidbCluster
            name: tidbclusters.pingcap.com
            version: v1alpha1
          - description: Tidb Initializer
            displayName: TidbInitializer
            kind: TidbInitializer
            name: tidbinitializers.pingcap.com
            version: v1alpha1
          - description: Tidb Monitor
            displayName: TidbMonitor
            kind: TidbMonitor
            name: tidbmonitors.pingcap.com
            version: v1alpha1
          - description: Tidb NG Monitoring
            displayName: TidbNGMonitoring
            kind: TidbNGMonitoring
            name: tidbngmonitorings.pingcap.com
            version: v1alpha1
        description: |
          ## About the managed application
          TiDB ("Ti" stands for Titanium) is an open-source, distributed, NewSQL database that supports Hybrid Transactional and Analytical Processing (HTAP) workloads. It is MySQL compatible and features horizontal scalability, strong consistency, and high availability. TiDB can be deployed on-premise or in-cloud.
          ## About this Operator

          TiDB Operator manages TiDB clusters on Kubernetes and automates tasks related to operating a TiDB cluster. It makes TiDB a truly cloud-native database.

          ## Features

          * Safely scaling the TiDB cluster

          TiDB Operator empowers TiDB with horizontal scalability on the cloud.

          * Rolling update of the TiDB cluster

          Gracefully perform rolling updates for the TiDB cluster in order, achieving zero-downtime of the TiDB cluster.

          * Multi-tenant support

          Users can deploy and manage multiple TiDB clusters on a single Kubernetes cluster easily.

          * Automatic failover

          TiDB Operator automatically performs failover for your TiDB cluster when node failures occur.

          * Kubernetes package manager support

          By embracing Kubernetes package manager Helm, users can easily deploy TiDB clusters with only one command.

          * Automatically monitoring TiDB cluster at creating

          Automatically deploy Prometheus, Grafana for TiDB cluster monitoring.

          ## Quick Start
          You can follow our [Get Started](https://docs.pingcap.com/tidb-in-kubernetes/stable/get-started) guide to quickly start a testing Kubernetes cluster and play with TiDB Operator on your own machine.

          ## Documentation

          You can see our documentation at PingCAP website for more in-depth installation and instructions for production:

          - [English](https://docs.pingcap.com/tidb-in-kubernetes/stable)
          - [简体中文](https://docs.pingcap.com/zh/tidb-in-kubernetes/stable)

          All the TiDB Operator documentation is maintained in the [docs-tidb-operator repository](https://github.com/pingcap/docs-tidb-operator).

          ## Community

          Feel free to reach out if you have any questions. The maintainers of this project are reachable via:

          - [TiDB Community Slack](https://pingcap.com/tidbslack/) in the [#sig-k8s](https://tidbcommunity.slack.com/archives/CHD0HA3LZ) channel
          - [Filing an issue](https://github.com/pingcap/tidb-operator/issue) against this repo

          Pull Requests are welcome! Check the [issue tracker](https://github.com/pingcap/tidb-operator/issue) for `status/help-wanted` issues if you're unsure where to start.

          If you're planning a new feature, please file an issue or join [#sig-k8s](https://tidbcommunity.slack.com/archives/CHD0HA3LZ) channel to discuss first.
          ## Prerequisites for enabling this Operator
        displayName: TiDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TiDB
        - SQL
        - Database
        links:
        - name: GitHub
          url: https://github.com/pingcap/tidb-operator
        - name: Documentation
          url: https://docs.pingcap.com/tidb-in-kubernetes/stable
        maintainers:
        - email: zhanghailong@pingcap.com
          name: Daniel Zhang
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: PingCAP
        relatedImages:
        - pingcap/tidb-operator:v1.3.1
        version: 1.3.1
      entries:
      - name: tidb-operator.v1.3.1
        version: 1.3.1
      name: stable
    defaultChannel: stable
    packageName: tidb-operator
    provider:
      name: PingCAP
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v0.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: gcr.io/datadoghq/operator:0.8.4
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          olm.skipRange: <0.8.4
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - gcr.io/datadoghq/operator:0.8.4
        version: 0.8.4
      entries:
      - name: datadog-operator.v0.8.4
        version: 0.8.4
      - name: datadog-operator.v0.8.2
        version: 0.8.2
      - name: datadog-operator.v0.8.1
        version: 0.8.1
      - name: datadog-operator.v0.8.0
        version: 0.8.0
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: gcr.io/datadoghq/operator:1.18.0
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - gcr.io/datadoghq/operator:1.18.0
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.2
        version: 1.4.2
      - name: datadog-operator.v1.4.1
        version: 1.4.1
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      - name: datadog-operator.v1.2.0
        version: 1.2.0
      - name: datadog-operator.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: datadog-operator
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: servicemeshoperator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: servicemeshoperator.v2.6.10
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"maistra.io/v2\",\n    \"kind\":
            \"ServiceMeshControlPlane\",\n    \"metadata\": {\n      \"name\": \"basic\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"version\": \"v2.6\",\n      \"policy\": {\n         \"type\":
            \"Istiod\"\n      },\n      \"telemetry\": {\n         \"type\": \"Istiod\"\n
            \     },\n      \"addons\": {\n        \"prometheus\": {\n          \"enabled\":
            true\n        },\n        \"kiali\": {\n          \"enabled\": true\n
            \       },\n        \"grafana\": {\n          \"enabled\": true\n        }\n
            \     }\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n    \"kind\":
            \"ServiceMeshMemberRoll\",\n    \"metadata\": {\n      \"name\": \"default\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"members\": [\n        \"your-project\",\n        \"another-of-your-projects\"
            \n      ]\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n
            \   \"kind\": \"ServiceMeshMember\",\n    \"metadata\": {\n      \"name\":
            \"default\",\n      \"namespace\": \"application-namespace\"\n    },\n
            \   \"spec\": {\n      \"controlPlaneRef\": {\n        \"name\": \"basic\",\n
            \       \"namespace\": \"control-plane-namespace\"\n      }\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
          createdAt: 2025-08-11T09:32:44CEST
          description: The OpenShift Service Mesh 2 Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh 2.6
            and earlier releases. OpenShift Service Mesh 2 is based on the open source
            Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2.6.10-0'
          operators.openshift.io/infrastructure-features: '["Disconnected","fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/maistra/istio-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Istio control plane installation
            displayName: Istio Service Mesh Control Plane
            kind: ServiceMeshControlPlane
            name: servicemeshcontrolplanes.maistra.io
            version: v2
          - description: Marks the containing namespace as a member of the referenced
              Service Mesh
            displayName: Istio Service Mesh Member
            kind: ServiceMeshMember
            name: servicemeshmembers.maistra.io
            version: v1
          - description: A list of namespaces in Service Mesh
            displayName: Istio Service Mesh Member Roll
            kind: ServiceMeshMemberRoll
            name: servicemeshmemberrolls.maistra.io
            version: v1
        description: |-
          Red Hat OpenShift Service Mesh 2 is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh 2, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh 2
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh 2 provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh 2 supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### Joining Projects Into a Mesh

          Once an instance of Red Hat OpenShift Service Mesh 2 has been installed, it will only exercise control over services within its own
          project.  Other projects may be added into the mesh using one of two methods:

          1. A **ServiceMeshMember** resource may be created in other projects to add those projects into the mesh.  The
            **ServiceMeshMember** specifies a reference to the **ServiceMeshControlPlane** object that was used to install
            the control plane.  The user creating the **ServiceMeshMember** resource must have permission to *use* the
            **ServiceMeshControlPlane** object.  The adminstrator for the project containing the control plane can grant
            individual users or groups the *use* permissions.

          2. A **ServiceMeshMemberRoll** resource may be created in the project containing the control plane.  This resource
            contains a single *members* list of all the projects that should belong in the mesh.  The resource must be named
            *default*.  The user creating the resource must have *edit* or *admin* permissions for all projects in the
            *members* list.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 2
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - maistra
        - servicemesh
        links:
        - name: Red Hat OpenShift Service Mesh 2
          url: https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html
        - name: Istio
          url: https://istio.io/
        - name: Operator Source Code
          url: https://github.com/Maistra/istio-operator
        - name: Bugs
          url: https://issues.redhat.com/projects/OSSM
        maintainers:
        - email: istio-feedback@redhat.com
          name: Red Hat, OpenShift Service Mesh
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:4ed63a5440753b1b22d44191246c82e7671e3f545c1dbdae4cb4fad163ca3a46
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6b4623c416fc85bd8e38904c16494fa3c4aeae25c90dc0da9763a0a87141bc04
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:515c5d9e6700fae981b0ea3807436b250acb1534e9ecf50043e5df8474e66dce
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:b134578e0677963197e353d8acd0ac3330df1b6d2752aa41cb072e86b5b4b91b
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6fb7d50b188dc27ffb9a396edabbfb21767fe438344fd70d4754d1ec7cf27957
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:e734ca310e09f737b4f06f3ecfeb9ac12971c63884fdcc27e3bebef9d5b18722
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:a36d5db6097749c55e946bba3b8ddd13a84b1f6f68b62d0ccbad5a803e9b46fb
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:fc500a282c6668ff987a632efbbd19ecc3aebdc021f831a7b386e360a5c8324d
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:57eb7ba03f24781501a7ceed2e160ff95317869af6c2caf8f9a83990e6a6c7b9
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:1ca80927334c6af8e386de8d44704f568b4b35d3f7afbe785a999d0fa4efd0bf
        - registry.redhat.io/openshift4/ose-prometheus@sha256:423ae2e81cfe660b7217f7662754f366a91d457d6dd1589b4fc40733896eb9d6
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:d16a60204d266bf4cc266c005fabf175b533b827500f171013884a713c88857a
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel8@sha256:3d773d16fed72278949b9e3b54dc59e346e95341dc94b715bcfc328d3c9e83b2
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:6a420228e684596ce2b900614a874f3e1284a6f2bbbba4c5a71ade158867affe
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:c77183e50a34d33b5e2fede60350f54b18ab3536c3189e331569d5df862d9979
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:a3d65983d46dcf47b94f26e7265ed32082208ce08abcf3587a9a4e15c19b274a
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:b241452cce5c9af5aa0f12d1db3a7094cc004ce355988b8a5207091810a6f84b
        - registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:3dc1bddb070fd530ce6251666989b9a849de90591718a578da4a9900f98d93ec
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:c34b4cd9a82fef40708a25f0568f6dc632878affba35d5d631a48ec31da08cd3
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:c1cadf40bdb9511d612a0257c12eb3a9738823592ebb6e32a4f51b3a8dfe08db
        - registry.redhat.io/openshift-service-mesh/prometheus-rhel8@sha256:71e193a2d403bb06a7a76686b376985315aa0c32f495b8b9dacd73c7ccce2af3
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d95762c62087072cadee5948c706619f09202a349e1310f0ed81bfb42f11203f
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:4669724b430683bf2e895c5cf941c2d99536117815999d38131e5b6f162bbef0
        version: 2.6.10-0
      entries:
      - name: servicemeshoperator.v2.6.10
        version: 2.6.10-0
      - name: servicemeshoperator.v2.6.9
        version: 2.6.9-0
      - name: servicemeshoperator.v2.6.8
        version: 2.6.8-0
      - name: servicemeshoperator.v2.6.7
        version: 2.6.7-0
      - name: servicemeshoperator.v2.6.6
        version: 2.6.6-0
      - name: servicemeshoperator.v2.6.5
        version: 2.6.5-0
      - name: servicemeshoperator.v2.6.4
        version: 2.6.4-0
      - name: servicemeshoperator.v2.6.3
        version: 2.6.3-0
      - name: servicemeshoperator.v2.6.2
        version: 2.6.2-0
      - name: servicemeshoperator.v2.6.1
        version: 2.6.1-0
      - name: servicemeshoperator.v2.6.0
        version: 2.6.0-0
      - name: servicemeshoperator.v2.5.2
        version: 2.5.2-0
      - name: servicemeshoperator.v2.5.1
        version: 2.5.1-0
      - name: servicemeshoperator.v2.5.0
        version: 2.5.0-0
      - name: servicemeshoperator.v2.4.5
        version: 2.4.5-0
      - name: servicemeshoperator.v2.4.4
        version: 2.4.4-0
      - name: servicemeshoperator.v2.4.3
        version: 2.4.3-0
      - name: servicemeshoperator.v2.4.2
        version: 2.4.2-0
      - name: servicemeshoperator.v2.4.1
        version: 2.4.1-0
      - name: servicemeshoperator.v2.4.0
        version: 2.4.0-0
      - name: servicemeshoperator.v2.3.3
        version: 2.3.3-0
      - name: servicemeshoperator.v2.3.2
        version: 2.3.2-0
      - name: servicemeshoperator.v2.3.1
        version: 2.3.1-0
      - name: servicemeshoperator.v2.3.0
        version: 2.3.0-0
      - name: servicemeshoperator.v2.2.3
        version: 2.2.3-0
      - name: servicemeshoperator.v2.2.2
        version: 2.2.2-0
      - name: servicemeshoperator.v2.2.1
        version: 2.2.1-0
      - name: servicemeshoperator.v2.2.0
        version: 2.2.0-0
      - name: servicemeshoperator.v2.1.3
        version: 2.1.3-0
      - name: servicemeshoperator.v2.1.2
        version: 2.1.2-0
      - name: servicemeshoperator.v2.1.1
        version: 2.1.1-0
      - name: servicemeshoperator.v2.1.0
        version: 2.1.0-0
      - name: servicemeshoperator.v2.0.8
        version: 2.0.8-0
      name: "1.0"
    - currentCSV: servicemeshoperator.v2.6.10
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"maistra.io/v2\",\n    \"kind\":
            \"ServiceMeshControlPlane\",\n    \"metadata\": {\n      \"name\": \"basic\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"version\": \"v2.6\",\n      \"policy\": {\n         \"type\":
            \"Istiod\"\n      },\n      \"telemetry\": {\n         \"type\": \"Istiod\"\n
            \     },\n      \"addons\": {\n        \"prometheus\": {\n          \"enabled\":
            true\n        },\n        \"kiali\": {\n          \"enabled\": true\n
            \       },\n        \"grafana\": {\n          \"enabled\": true\n        }\n
            \     }\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n    \"kind\":
            \"ServiceMeshMemberRoll\",\n    \"metadata\": {\n      \"name\": \"default\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"members\": [\n        \"your-project\",\n        \"another-of-your-projects\"
            \n      ]\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n
            \   \"kind\": \"ServiceMeshMember\",\n    \"metadata\": {\n      \"name\":
            \"default\",\n      \"namespace\": \"application-namespace\"\n    },\n
            \   \"spec\": {\n      \"controlPlaneRef\": {\n        \"name\": \"basic\",\n
            \       \"namespace\": \"control-plane-namespace\"\n      }\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
          createdAt: 2025-08-11T09:32:44CEST
          description: The OpenShift Service Mesh 2 Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh 2.6
            and earlier releases. OpenShift Service Mesh 2 is based on the open source
            Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2.6.10-0'
          operators.openshift.io/infrastructure-features: '["Disconnected","fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/maistra/istio-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Istio control plane installation
            displayName: Istio Service Mesh Control Plane
            kind: ServiceMeshControlPlane
            name: servicemeshcontrolplanes.maistra.io
            version: v2
          - description: Marks the containing namespace as a member of the referenced
              Service Mesh
            displayName: Istio Service Mesh Member
            kind: ServiceMeshMember
            name: servicemeshmembers.maistra.io
            version: v1
          - description: A list of namespaces in Service Mesh
            displayName: Istio Service Mesh Member Roll
            kind: ServiceMeshMemberRoll
            name: servicemeshmemberrolls.maistra.io
            version: v1
        description: |-
          Red Hat OpenShift Service Mesh 2 is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh 2, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh 2
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh 2 provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh 2 supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### Joining Projects Into a Mesh

          Once an instance of Red Hat OpenShift Service Mesh 2 has been installed, it will only exercise control over services within its own
          project.  Other projects may be added into the mesh using one of two methods:

          1. A **ServiceMeshMember** resource may be created in other projects to add those projects into the mesh.  The
            **ServiceMeshMember** specifies a reference to the **ServiceMeshControlPlane** object that was used to install
            the control plane.  The user creating the **ServiceMeshMember** resource must have permission to *use* the
            **ServiceMeshControlPlane** object.  The adminstrator for the project containing the control plane can grant
            individual users or groups the *use* permissions.

          2. A **ServiceMeshMemberRoll** resource may be created in the project containing the control plane.  This resource
            contains a single *members* list of all the projects that should belong in the mesh.  The resource must be named
            *default*.  The user creating the resource must have *edit* or *admin* permissions for all projects in the
            *members* list.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 2
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - maistra
        - servicemesh
        links:
        - name: Red Hat OpenShift Service Mesh 2
          url: https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html
        - name: Istio
          url: https://istio.io/
        - name: Operator Source Code
          url: https://github.com/Maistra/istio-operator
        - name: Bugs
          url: https://issues.redhat.com/projects/OSSM
        maintainers:
        - email: istio-feedback@redhat.com
          name: Red Hat, OpenShift Service Mesh
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:c1cadf40bdb9511d612a0257c12eb3a9738823592ebb6e32a4f51b3a8dfe08db
        - registry.redhat.io/openshift4/ose-prometheus@sha256:423ae2e81cfe660b7217f7662754f366a91d457d6dd1589b4fc40733896eb9d6
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:4ed63a5440753b1b22d44191246c82e7671e3f545c1dbdae4cb4fad163ca3a46
        - registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:57eb7ba03f24781501a7ceed2e160ff95317869af6c2caf8f9a83990e6a6c7b9
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:3dc1bddb070fd530ce6251666989b9a849de90591718a578da4a9900f98d93ec
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel8@sha256:3d773d16fed72278949b9e3b54dc59e346e95341dc94b715bcfc328d3c9e83b2
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:a3d65983d46dcf47b94f26e7265ed32082208ce08abcf3587a9a4e15c19b274a
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:6a420228e684596ce2b900614a874f3e1284a6f2bbbba4c5a71ade158867affe
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:c77183e50a34d33b5e2fede60350f54b18ab3536c3189e331569d5df862d9979
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:d16a60204d266bf4cc266c005fabf175b533b827500f171013884a713c88857a
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:1ca80927334c6af8e386de8d44704f568b4b35d3f7afbe785a999d0fa4efd0bf
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6b4623c416fc85bd8e38904c16494fa3c4aeae25c90dc0da9763a0a87141bc04
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:515c5d9e6700fae981b0ea3807436b250acb1534e9ecf50043e5df8474e66dce
        - registry.redhat.io/openshift-service-mesh/prometheus-rhel8@sha256:71e193a2d403bb06a7a76686b376985315aa0c32f495b8b9dacd73c7ccce2af3
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d95762c62087072cadee5948c706619f09202a349e1310f0ed81bfb42f11203f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:e734ca310e09f737b4f06f3ecfeb9ac12971c63884fdcc27e3bebef9d5b18722
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:a36d5db6097749c55e946bba3b8ddd13a84b1f6f68b62d0ccbad5a803e9b46fb
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:b134578e0677963197e353d8acd0ac3330df1b6d2752aa41cb072e86b5b4b91b
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6fb7d50b188dc27ffb9a396edabbfb21767fe438344fd70d4754d1ec7cf27957
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:4669724b430683bf2e895c5cf941c2d99536117815999d38131e5b6f162bbef0
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:fc500a282c6668ff987a632efbbd19ecc3aebdc021f831a7b386e360a5c8324d
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:c34b4cd9a82fef40708a25f0568f6dc632878affba35d5d631a48ec31da08cd3
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:b241452cce5c9af5aa0f12d1db3a7094cc004ce355988b8a5207091810a6f84b
        version: 2.6.10-0
      entries:
      - name: servicemeshoperator.v2.6.10
        version: 2.6.10-0
      - name: servicemeshoperator.v2.6.9
        version: 2.6.9-0
      - name: servicemeshoperator.v2.6.8
        version: 2.6.8-0
      - name: servicemeshoperator.v2.6.7
        version: 2.6.7-0
      - name: servicemeshoperator.v2.6.6
        version: 2.6.6-0
      - name: servicemeshoperator.v2.6.5
        version: 2.6.5-0
      - name: servicemeshoperator.v2.6.4
        version: 2.6.4-0
      - name: servicemeshoperator.v2.6.3
        version: 2.6.3-0
      - name: servicemeshoperator.v2.6.2
        version: 2.6.2-0
      - name: servicemeshoperator.v2.6.1
        version: 2.6.1-0
      - name: servicemeshoperator.v2.6.0
        version: 2.6.0-0
      - name: servicemeshoperator.v2.5.2
        version: 2.5.2-0
      - name: servicemeshoperator.v2.5.1
        version: 2.5.1-0
      - name: servicemeshoperator.v2.5.0
        version: 2.5.0-0
      - name: servicemeshoperator.v2.4.5
        version: 2.4.5-0
      - name: servicemeshoperator.v2.4.4
        version: 2.4.4-0
      - name: servicemeshoperator.v2.4.3
        version: 2.4.3-0
      - name: servicemeshoperator.v2.4.2
        version: 2.4.2-0
      - name: servicemeshoperator.v2.4.1
        version: 2.4.1-0
      - name: servicemeshoperator.v2.4.0
        version: 2.4.0-0
      - name: servicemeshoperator.v2.3.3
        version: 2.3.3-0
      - name: servicemeshoperator.v2.3.2
        version: 2.3.2-0
      - name: servicemeshoperator.v2.3.1
        version: 2.3.1-0
      - name: servicemeshoperator.v2.3.0
        version: 2.3.0-0
      - name: servicemeshoperator.v2.2.3
        version: 2.2.3-0
      - name: servicemeshoperator.v2.2.2
        version: 2.2.2-0
      - name: servicemeshoperator.v2.2.1
        version: 2.2.1-0
      - name: servicemeshoperator.v2.2.0
        version: 2.2.0-0
      - name: servicemeshoperator.v2.1.3
        version: 2.1.3-0
      - name: servicemeshoperator.v2.1.2
        version: 2.1.2-0
      - name: servicemeshoperator.v2.1.1
        version: 2.1.1-0
      - name: servicemeshoperator.v2.1.0
        version: 2.1.0-0
      - name: servicemeshoperator.v2.0.8
        version: 2.0.8-0
      name: stable
    defaultChannel: stable
    packageName: servicemeshoperator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Labs
      provider-url: https://grafana.com
    name: grafana-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: grafana-operator.v5.19.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "Grafana",
                "metadata": {
                  "labels": {
                    "dashboards": "grafana-a",
                    "folders": "grafana-a"
                  },
                  "name": "grafana-a"
                },
                "spec": {
                  "config": {
                    "auth": {
                      "disable_login_form": "false"
                    },
                    "log": {
                      "mode": "console"
                    },
                    "security": {
                      "admin_password": "start",
                      "admin_user": "root"
                    }
                  }
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaAlertRuleGroup",
                "metadata": {
                  "name": "grafanaalertrulegroup-sample"
                },
                "spec": {
                  "folderRef": "test-folder-from-operator",
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "interval": "5m",
                  "rules": [
                    {
                      "condition": "B",
                      "data": [
                        {
                          "datasourceUid": "grafanacloud-demoinfra-prom",
                          "model": {
                            "datasource": {
                              "type": "prometheus",
                              "uid": "grafanacloud-demoinfra-prom"
                            },
                            "editorMode": "code",
                            "expr": "weather_temp_c{}",
                            "instant": true,
                            "intervalMs": 1000,
                            "legendFormat": "__auto",
                            "maxDataPoints": 43200,
                            "range": false,
                            "refId": "A"
                          },
                          "refId": "A",
                          "relativeTimeRange": {
                            "from": 600
                          }
                        },
                        {
                          "datasourceUid": "__expr__",
                          "model": {
                            "conditions": [
                              {
                                "evaluator": {
                                  "params": [
                                    0
                                  ],
                                  "type": "gt"
                                },
                                "operator": {
                                  "type": "and"
                                },
                                "query": {
                                  "params": [
                                    "C"
                                  ]
                                },
                                "reducer": {
                                  "params": [],
                                  "type": "last"
                                },
                                "type": "query"
                              }
                            ],
                            "datasource": {
                              "type": "__expr__",
                              "uid": "__expr__"
                            },
                            "expression": "A",
                            "intervalMs": 1000,
                            "maxDataPoints": 43200,
                            "refId": "B",
                            "type": "threshold"
                          },
                          "refId": "B",
                          "relativeTimeRange": {
                            "from": 600
                          }
                        }
                      ],
                      "execErrState": "Error",
                      "for": "5m0s",
                      "noDataState": "NoData",
                      "title": "Temperature below freezing",
                      "uid": "4843de5c-4f8a-4af0-9509-23526a04faf8"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaContactPoint",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "grafana-operator",
                    "app.kubernetes.io/instance": "grafanacontactpoint-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "grafanacontactpoint",
                    "app.kubernetes.io/part-of": "grafana-operator"
                  },
                  "name": "grafanacontactpoint-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "name": "grafanacontactpoint-sample",
                  "settings": {
                    "email": null
                  },
                  "type": "email"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaDashboard",
                "metadata": {
                  "name": "grafanadashboard-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "json": "{\n\n  \"id\": null,\n  \"title\": \"Simple Dashboard\",\n  \"tags\": [],\n  \"style\": \"dark\",\n  \"timezone\": \"browser\",\n  \"editable\": true,\n  \"hideControls\": false,\n  \"graphTooltip\": 1,\n  \"panels\": [],\n  \"time\": {\n    \"from\": \"now-6h\",\n    \"to\": \"now\"\n  },\n  \"timepicker\": {\n    \"time_options\": [],\n    \"refresh_intervals\": []\n  },\n  \"templating\": {\n    \"list\": []\n  },\n  \"annotations\": {\n    \"list\": []\n  },\n  \"refresh\": \"5s\",\n  \"schemaVersion\": 17,\n  \"version\": 0,\n  \"links\": []\n}\n"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaDatasource",
                "metadata": {
                  "name": "grafanadatasource-sample"
                },
                "spec": {
                  "datasource": {
                    "access": "proxy",
                    "isDefault": true,
                    "jsonData": {
                      "timeInterval": "5s",
                      "tlsSkipVerify": true
                    },
                    "name": "prometheus",
                    "type": "prometheus",
                    "url": "http://prometheus-service:9090"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "plugins": [
                    {
                      "name": "grafana-clock-panel",
                      "version": "1.3.0"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaFolder",
                "metadata": {
                  "name": "grafanafolder-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "title": "Example Folder"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaLibraryPanel",
                "metadata": {
                  "name": "grafana-library-panel-inline-envs"
                },
                "spec": {
                  "envs": [
                    {
                      "name": "CUSTOM_RANGE_ENV",
                      "value": "now - 12h"
                    }
                  ],
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "jsonnet": "local myRange = std.extVar('CUSTOM_RANGE_ENV'); {\n\n  model: {}\n}\n",
                  "plugins": [
                    {
                      "name": "grafana-piechart-panel",
                      "version": "1.3.9"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaMuteTiming",
                "metadata": {
                  "name": "mutetiming-sample"
                },
                "spec": {
                  "editable": false,
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "name": "mutetiming-sample",
                  "time_intervals": [
                    {
                      "days_of_month": [
                        "1",
                        "15"
                      ],
                      "location": "Asia/Shanghai",
                      "times": [
                        {
                          "end_time": "06:00",
                          "start_time": "00:00"
                        }
                      ],
                      "weekdays": [
                        "saturday"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationPolicy",
                "metadata": {
                  "name": "grafananotificationpolicy-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "route": {
                    "group_by": [
                      "grafana_folder",
                      "alertname"
                    ],
                    "receiver": "Grafana Cloud OnCall",
                    "routes": [
                      {
                        "object_matchers": [
                          [
                            "foo",
                            "=",
                            "bar"
                          ]
                        ],
                        "receiver": "grafana-default-email",
                        "routes": [
                          {
                            "object_matchers": [
                              [
                                "severity",
                                "=",
                                "critical"
                              ]
                            ],
                            "receiver": "Grafana Cloud OnCall"
                          }
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationPolicyRoute",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "grafana-operator",
                    "app.kubernetes.io/instance": "grafananotificationpolicyroute-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "grafananotificationpolicyroute",
                    "app.kubernetes.io/part-of": "grafana-operator"
                  },
                  "name": "grafananotificationpolicyroute-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationTemplate",
                "metadata": {
                  "name": "test"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "name": "test",
                  "template": "{{ define \"SlackAlert\" }}\n  [{{.Status}}] {{ .Labels.alertname }}\n  {{ .Annotations.AlertValues }}\n{{ end }}\n\n{{ define \"SlackAlertMessage\" }}\n  {{ if gt (len .Alerts.Firing) 0 }}\n    {{ len .Alerts.Firing }} firing:\n    {{ range .Alerts.Firing }} {{ template \"SlackAlert\" . }} {{ end }}\n  {{ end }}\n  {{ if gt (len .Alerts.Resolved) 0 }}\n    {{ len .Alerts.Resolved }} resolved:\n    {{ range .Alerts.Resolved }} {{ template \"SlackAlert\" . }} {{ end }}\n  {{ end }}\n{{ end }}\n\n{{ template \"SlackAlertMessage\" . }}\n"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: ghcr.io/grafana/grafana-operator@sha256:b678bc3faad11f4ef53c240e2a650ce939b9e8920dd72a47681b2efcadc7b837
          createdAt: "2025-08-21T08:38:00Z"
          description: Deploys and manages Grafana instances, dashboards and data
            sources
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/grafana/grafana-operator
          support: Grafana Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GrafanaAlertRuleGroup is the Schema for the grafanaalertrulegroups
              API
            displayName: Grafana Alert Rule Group
            kind: GrafanaAlertRuleGroup
            name: grafanaalertrulegroups.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaContactPoint is the Schema for the grafanacontactpoints
              API
            displayName: Grafana Contact Point
            kind: GrafanaContactPoint
            name: grafanacontactpoints.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaDashboard is the Schema for the grafanadashboards
              API
            displayName: Grafana Dashboard
            kind: GrafanaDashboard
            name: grafanadashboards.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaDatasource is the Schema for the grafanadatasources
              API
            displayName: Grafana Datasource
            kind: GrafanaDatasource
            name: grafanadatasources.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaFolder is the Schema for the grafanafolders API
            displayName: Grafana Folder
            kind: GrafanaFolder
            name: grafanafolders.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaLibraryPanel
            name: grafanalibrarypanels.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaMuteTiming
            name: grafanamutetimings.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaNotificationPolicy is the Schema for the GrafanaNotificationPolicy
              API
            displayName: Grafana Notification policy
            kind: GrafanaNotificationPolicy
            name: grafananotificationpolicies.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaNotificationPolicyRoute is the Schema for the grafananotificationpolicyroutes
              API
            displayName: Grafana Notification Policy Route
            kind: GrafanaNotificationPolicyRoute
            name: grafananotificationpolicyroutes.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaNotificationTemplate
            name: grafananotificationtemplates.grafana.integreatly.org
            version: v1beta1
          - description: Grafana is the Schema for the grafanas API
            displayName: Grafana
            kind: Grafana
            name: grafanas.grafana.integreatly.org
            version: v1beta1
        description: Deploys and manages Grafana instances, dashboards and data sources
        displayName: Grafana Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Grafana
        - Metrics
        - Observability
        links:
        - name: Grafana Operator
          url: https://grafana.github.io/grafana-operator
        maintainers:
        - email: grafana-operator@grafana.com
          name: Grafana-operator maintainers
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Grafana Labs
          url: https://grafana.com
        relatedImages:
        - ghcr.io/grafana/grafana-operator@sha256:b678bc3faad11f4ef53c240e2a650ce939b9e8920dd72a47681b2efcadc7b837
        - docker.io/grafana/grafana@sha256:6ac590e7cabc2fbe8d7b8fc1ce9c9f0582177b334e0df9c927ebd9670469440f
        version: 5.19.4
      entries:
      - name: grafana-operator.v5.19.4
        version: 5.19.4
      - name: grafana-operator.v5.18.0
        version: 5.18.0
      - name: grafana-operator.v5.17.1
        version: 5.17.1
      - name: grafana-operator.v5.17.0
        version: 5.17.0
      - name: grafana-operator.v5.16.0
        version: 5.16.0
      - name: grafana-operator.v5.15.1
        version: 5.15.1
      - name: grafana-operator.v5.14.0
        version: 5.14.0
      name: v5
    defaultChannel: v5
    packageName: grafana-operator
    provider:
      name: Grafana Labs
      url: https://grafana.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: netobserv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netobserv-operator.v1.6.3-community
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowMetric",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netobserv-operator",
                    "app.kubernetes.io/instance": "flowmetric-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "flowmetric",
                    "app.kubernetes.io/part-of": "netobserv-operator"
                  },
                  "name": "flowmetric-sample"
                },
                "spec": {
                  "charts": [
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "",
                          "promQL": "sum(rate($METRIC[2m]))"
                        }
                      ],
                      "title": "External ingress traffic",
                      "type": "SingleStat",
                      "unit": "Bps"
                    },
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "{{DstK8S_Namespace}} / {{DstK8S_OwnerName}}",
                          "promQL": "sum(rate($METRIC{DstK8S_Namespace!=\"\"}[2m])) by (DstK8S_Namespace, DstK8S_OwnerName)"
                        }
                      ],
                      "sectionName": "External",
                      "title": "Top external ingress traffic per workload",
                      "type": "StackArea",
                      "unit": "Bps"
                    }
                  ],
                  "direction": "Ingress",
                  "filters": [
                    {
                      "field": "SrcSubnetLabel",
                      "matchType": "Absence"
                    }
                  ],
                  "labels": [
                    "DstK8S_HostName",
                    "DstK8S_Namespace",
                    "DstK8S_OwnerName",
                    "DstK8S_OwnerType"
                  ],
                  "metricName": "cluster_external_ingress_bytes_total",
                  "type": "Counter",
                  "valueField": "Bytes"
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "app"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "infra"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "enable": true,
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "includeList": [
                        "node_ingress_bytes_total",
                        "workload_ingress_bytes_total",
                        "namespace_flows_total"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta2",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "metrics": {
                        "server": {
                          "port": 9400
                        }
                      },
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "eBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "Disabled"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "\"app\""
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "\"infra\""
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "\"Pod\"",
                          "src_kind": "\"Pod\""
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "\"Service\""
                        },
                        "name": "Services network"
                      }
                    ],
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    }
                  },
                  "deploymentModel": "Direct",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "enable": true,
                    "lokiStack": {
                      "name": "loki"
                    },
                    "mode": "Monolithic",
                    "monolithic": {
                      "tenantID": "netobserv",
                      "tls": {
                        "caCert": {
                          "certFile": "service-ca.crt",
                          "name": "loki-gateway-ca-bundle",
                          "type": "configmap"
                        },
                        "enable": false
                      },
                      "url": "http://loki.netobserv.svc:3100/"
                    },
                    "readTimeout": "30s",
                    "writeBatchSize": 10485760,
                    "writeBatchWait": "1s",
                    "writeTimeout": "10s"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "Flows",
                    "metrics": {
                      "disableAlerts": [],
                      "server": {
                        "port": 9401
                      }
                    },
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "prometheus": {
                    "querier": {
                      "enable": true,
                      "mode": "Auto",
                      "timeout": "30s"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:1.6.2-community
          createdAt: "2024-08-22T18:26:56Z"
          description: Network flows collector and monitoring solution
          operatorframework.io/initialization-resource: '{"apiVersion":"flows.netobserv.io/v1beta2",
            "kind":"FlowCollector","metadata":{"name":"cluster"},"spec": {}}'
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
          support: NetObserv team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta2
          - description: '`FlowMetric` is the schema for the custom metrics API, which
              allows to generate more metrics out of flow logs. You can find examples
              here: https://github.com/netobserv/network-observability-operator/tree/1.6.2-community/config/samples/flowmetrics'
            displayName: Flow Metric
            kind: FlowMetric
            name: flowmetrics.flows.netobserv.io
            version: v1alpha1
        description: |-
          DEPRECATED: this distribution channel is deprecated and doesn't receive new updates. Please read https://github.com/orgs/netobserv/discussions/996 to learn more.

          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline that consists in:
          - an eBPF agent, that generates network flows from captured packets
          - flowlogs-pipeline, a component that collects, enriches and exports these flows
          - when used in OpenShift, a Console plugin for flows visualization with powerful filtering options, a topology representation and more

          Flow data is then available in multiple ways, each optional:

          - As Prometheus metrics
          - As raw flow logs stored in Grafana Loki
          - As raw flow logs exported to a collector

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, can optionally be used as the backend to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 10GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          If you prefer to not use Loki, you must set `spec.loki.enable` to `false` in `FlowCollector`.
          In that case, you can still get the Prometheus metrics or export raw flows to a custom collector. But be aware that some of the Console plugin features will be disabled. For instance, you will not be able to view raw flows there, and the metrics / topology will have a more limited level of details, missing information such as pods or IPs.

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for example [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/config/samples/flows_v1beta2_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide on every node, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you might have to configure differently if you used another installation method. Make sure to disable it (`spec.loki.enable`) if you don't want to use Loki.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`) an optional list of exporters to which to send enriched flows. KAFKA and IPFIX exporters are supported. This allows you to define any custom storage or processing that can read from Kafka or use the IPFIX standard.

          - To enable availability zones awareness, set `spec.processor.addZone` to `true`.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - More information on [configuring metrics](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/Metrics.md).
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: '[DEPRECATED] NetObserv Operator'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/network-observability-operator:1.6.2-community
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/netobserv/netobserv-ebpf-agent:v1.6.2-community
        - quay.io/netobserv/flowlogs-pipeline:v1.6.2-community
        - quay.io/netobserv/network-observability-console-plugin:v1.6.2-community
        version: 1.6.3-community
      entries:
      - name: netobserv-operator.v1.6.3-community
        version: 1.6.3-community
      - name: netobserv-operator.v1.6.2-community
        version: 1.6.2-community
      - name: netobserv-operator.v1.6.1-community
        version: 1.6.1-community
      - name: netobserv-operator.v1.6.0-community
        version: 1.6.0-community
      - name: netobserv-operator.v1.5.0-community
        version: 1.5.0-community
      name: community
    - currentCSV: netobserv-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "ipfix": {
                      "cacheActiveTimeout": "20s",
                      "cacheMaxFlows": 400,
                      "clusterNetworkOperator": {
                        "namespace": "openshift-network-operator"
                      },
                      "ovnKubernetes": {
                        "containerName": "ovnkube-node",
                        "daemonSetName": "ovnkube-node",
                        "namespace": "ovn-kubernetes"
                      },
                      "sampling": 400
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "staticLabels": {
                      "app": "netobserv-flowcollector"
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "enableKubeProbes": true,
                    "healthPort": 8080,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:0.2.1
          createdAt: "2022-12-07T09:00:48Z"
          description: Network flows collector and monitoring solution
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
        description: |-
          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline to collect and enrich network flows. These flows can be produced by the NetObserv eBPF agent, or by any device or CNI able to export flows in IPFIX format, such as OVN-Kubernetes.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When used in OpenShift, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For a normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionnaly be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/0.2.1/config/samples/flows_v1alpha1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Agent (`spec.agent.type`) can be `EBPF` (default) or `IPFIX`. eBPF is recommended, as it should work in more situations and offers better performances. If you can't, or don't want to use eBPF, note that the IPFIX option is fully functional only when using [OVN-Kubernetes](https://github.com/ovn-org/ovn-kubernetes/) CNI. Other CNIs are not officially supported, but you may still be able to configure them manually if they allow IPFIX exports.

          - Sampling (`spec.agent.ebpf.sampling` and `spec.agent.ipfix.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50) for eBPF and 400 (1:400) for IPFIX. Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub more more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - A [F.A.Q](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: NetObserv Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: mmaciasl@redhat.com
          name: Mario Macias
        - email: ocazade@redhat.com
          name: Olivier Cazade
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/network-observability-console-plugin:v0.1.6
        - quay.io/netobserv/network-observability-operator:0.2.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/netobserv/netobserv-ebpf-agent:v0.2.2
        - quay.io/netobserv/flowlogs-pipeline:v0.1.5
        version: 0.2.1
      entries:
      - name: netobserv-operator.v0.2.1
        version: 0.2.1
      - name: netobserv-operator.v0.2.0
        version: 0.2.0
      - name: netobserv-operator.v0.1.4
        version: 0.1.4
      - name: netobserv-operator.v0.1.3
        version: 0.1.3
      - name: netobserv-operator.v0.1.2
        version: 0.1.2
      - name: netobserv-operator.v0.1.1
        version: 0.1.1
      - name: netobserv-operator.v0.1.0
        version: 0.1.0
      name: latest
    - currentCSV: netobserv-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "ipfix": {
                      "cacheActiveTimeout": "20s",
                      "cacheMaxFlows": 400,
                      "clusterNetworkOperator": {
                        "namespace": "openshift-network-operator"
                      },
                      "ovnKubernetes": {
                        "containerName": "ovnkube-node",
                        "daemonSetName": "ovnkube-node",
                        "namespace": "ovn-kubernetes"
                      },
                      "sampling": 400
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "staticLabels": {
                      "app": "netobserv-flowcollector"
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "enableKubeProbes": true,
                    "healthPort": 8080,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:0.2.1
          createdAt: "2022-12-07T09:00:48Z"
          description: Network flows collector and monitoring solution
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
        description: |-
          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline to collect and enrich network flows. These flows can be produced by the NetObserv eBPF agent, or by any device or CNI able to export flows in IPFIX format, such as OVN-Kubernetes.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When used in OpenShift, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For a normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionnaly be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/0.2.1/config/samples/flows_v1alpha1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Agent (`spec.agent.type`) can be `EBPF` (default) or `IPFIX`. eBPF is recommended, as it should work in more situations and offers better performances. If you can't, or don't want to use eBPF, note that the IPFIX option is fully functional only when using [OVN-Kubernetes](https://github.com/ovn-org/ovn-kubernetes/) CNI. Other CNIs are not officially supported, but you may still be able to configure them manually if they allow IPFIX exports.

          - Sampling (`spec.agent.ebpf.sampling` and `spec.agent.ipfix.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50) for eBPF and 400 (1:400) for IPFIX. Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub more more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - A [F.A.Q](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: NetObserv Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: mmaciasl@redhat.com
          name: Mario Macias
        - email: ocazade@redhat.com
          name: Olivier Cazade
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/network-observability-operator:0.2.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/netobserv/netobserv-ebpf-agent:v0.2.2
        - quay.io/netobserv/flowlogs-pipeline:v0.1.5
        - quay.io/netobserv/network-observability-console-plugin:v0.1.6
        version: 0.2.1
      entries:
      - name: netobserv-operator.v0.2.1
        version: 0.2.1
      - name: netobserv-operator.v0.2.0
        version: 0.2.0
      - name: netobserv-operator.v0.1.4
        version: 0.1.4
      - name: netobserv-operator.v0.1.3
        version: 0.1.3
      - name: netobserv-operator.v0.1.2
        version: 0.1.2
      - name: netobserv-operator.v0.1.1
        version: 0.1.1
      - name: netobserv-operator.v0.1.0
        version: 0.1.0
      name: v0.2.x
    defaultChannel: community
    packageName: netobserv-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-license-service-scanner-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-license-service-scanner-operator.v4.2.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1",
                "kind": "IBMLicenseServiceScanner",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-licensing-scanner-operator",
                    "app.kubernetes.io/instance": "ibm-licensing-scanner-instance",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ibm-licensing-scanner-instance",
                    "app.kubernetes.io/part-of": "ibm-licensing-scanner-operator"
                  },
                  "name": "ibm-licensing-scanner-instance"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "license-service-upload-secret": "ibm-licensing-scanner-upload-secret",
                  "log-level": "INFO",
                  "registry-pull-secret": "ibm-licensing-scanner-registry-pull-secret",
                  "scan": {
                    "namespaces": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:77c9db26a13dba25438a6fc5a41290606418481e08ea2ecc0a9a397062af74fb
          createdAt: "2025-05-14T12:59:00Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <4.2.16'
          operatorframework.io/suggested-namespace: ibm-licensing-scanner
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMLicenseServiceScanner is the Schema for the ibmlicenseservicescanners
              API
            displayName: IBMLicense Service Scanner
            kind: IBMLicenseServiceScanner
            name: ibmlicenseservicescanners.operator.ibm.com
            version: v1
        description: IBM License Service Scanner
        displayName: IBM License Service Scanner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - licensing
        - usage
        - IBMLicenseServiceScanner
        - IBM
        - Cloud
        - License
        - License Service
        - Service
        - Scanner
        - License Service Scanner
        - License Scanner
        - Service Scanner
        links:
        - name: Ibm License Service Scanner Operator
          url: https://ibm-license-service-scanner-operator.domain
        maintainers:
        - email: talk2sam@us.ibm.com
          name: IBM
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:77c9db26a13dba25438a6fc5a41290606418481e08ea2ecc0a9a397062af74fb
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:f3503f72a90ff8abbadaef3e78a16db6e1fbff4c90b475c33ead17bf1e69990c
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-bundle@sha256:b6cb9e5976b749dd7818ac19d1cfd5e24673f3c052a41d6e2fe1d2d7094694b4
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-catalog@sha256:c016904e8c5cfcf8270241ad5e28b52f528859984ed894cdd5e1283fa90cd6e8
        version: 4.2.16
      entries:
      - name: ibm-license-service-scanner-operator.v4.2.16
        version: 4.2.16
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-license-service-scanner-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/component: operator
      app.kubernetes.io/managed-by: kustomize
      app.kubernetes.io/name: zabbix-operator-certified
      app.kubernetes.io/part-of: zabbix-operator-certified
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zabbix LLC
      provider-url: ""
    name: zabbix-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zabbix-operator-certified.v6.0.40
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent",
                "metadata": {
                  "labels": {
                    "app": "agent",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "log_remote_commands": true,
                  "max_lines_per_second": 20,
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent2",
                "metadata": {
                  "labels": {
                    "app": "agent2",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent2"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "enable_persistent_buffer": false,
                  "enable_status_port": false,
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "persistent_buffer_claim": "",
                  "persistent_buffer_period": "1h",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixFull",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-full"
                },
                "spec": {
                  "builtin_report_service": true,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysql_volumeclaim": "mysql-full-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxyMysql",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-mysql"
                },
                "spec": {
                  "db_server_port": 3306,
                  "enable_snmp_traps": false,
                  "internal_db": true,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix_proxy",
                  "proxy": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "zabbix_mysql_volumeclaim": "mysql-proxy-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxySqlite",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-sqlite"
                },
                "spec": {
                  "db_data_volumeclaim": "zabbix-proxy-sqlite",
                  "enable_snmp_traps": false,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "proxy": {
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "node_name_as_db_name": true,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "proxy_size": 1,
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  }
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixServer",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-server"
                },
                "spec": {
                  "builtin_report_service": true,
                  "db_server_host": "mysql-server",
                  "db_server_port": 3306,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix",
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              }
            ]
          ansible.operator-sdk/reconcile-period: 5s
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          certified: "true"
          containerImage: registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
          createdAt: "2025-05-26T07:41:22Z"
          description: Zabbix operator with multiple deployment variants and different
            components
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <6.0.40
          operatorframework.io/suggested-namespace: zabbix
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: ""
          support: Zabbix
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Zabbix agent 2 is a new generation of Zabbix agent and may
              be used in place of Zabbix agent
            displayName: Zabbix agent 2
            kind: ZabbixAgent2
            name: zabbixagent2s.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix agent is deployed on a monitoring nodes to actively
              monitor local resources and applications
            displayName: Zabbix agent
            kind: ZabbixAgent
            name: zabbixagents.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database server, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Full
            kind: ZabbixFull
            name: zabbixfulls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with MySQL database server
            displayName: Zabbix proxy (MySQL)
            kind: ZabbixProxyMysql
            name: zabbixproxymysqls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with SQLite database
            displayName: Zabbix proxy (SQLite3)
            kind: ZabbixProxySqlite
            name: zabbixproxysqlites.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database support, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Server
            kind: ZabbixServer
            name: zabbixservers.kubernetes.zabbix.com
            version: v1alpha1
        description: "## About this Operator\n\nZabbix helps you to real-time monitoring
          of millions of metrics collected from tens of thousands of servers, virtual
          machines and network devices.\nThe Zabbix Operator allows users to easily
          deploy, manage, and maintain Zabbix deployments on OpenShift. By installing
          this integration you will be able to deploy Zabbix server / proxies and
          other components with a single command.\n## Supported Features\n* **Zabbix
          Server** - Simple Zabbix installation with included Zabbix server, Zabbix
          web-interface and Zabbix Java Gateway with MySQL database support. The feature
          does not provide MySQL service and requires an external MySQL database.
          \n* **Zabbix Server (Full)** - Zabbix installation with included Zabbix
          server, Zabbix web-interface, Zabbix Java Gateway and MySQL server instance.\n*
          **Zabbix proxy (SQLite3)** - Very simple way to gain power of Zabbix proxy.
          The feature has  SQLite3 support for Zabbix proxies and allow to specify
          amount of proxies. \n* **Zabbix proxy (MySQL)** - Another option of Zabbix
          proxy. The option support and deliver MySQL database.\n* **Zabbix agent**
          - Zabbix agent can be deployed on each available node for stability and
          performance monitoring on remote nodes. It allows to gather metrics with
          full automation!* \n## Prerequisites\nAll deployment options are require
          additional information during deployment. Please, check the following instructions
          and provide required configuration:\n* **Zabbix Server** - MySQL database
          host information and MySQL database credentials in specially formatted *Secret*.
          Additionally it is possible to specify SSL certificates for HTTPS support
          in *Secret*.\n* **Zabbix Server (Full)** - MySQL database credentials in
          specially formatted *Secret*. MySQL database volume name information. Additionally
          it is possible to specify SSL certificates for HTTPS support in *Secret*.\n*
          **Zabbix proxy (SQLite3)** - Zabbix server host information only.\n* **Zabbix
          proxy (MySQL)** - MySQL database credentials in specially formatted *Secrets*
          and Zabbix server host information.\n* **Zabbix agent** - Zabbix server
          host information only for outgoing and incoming connections. Zabbix agent
          uses \"privileged\" mode to monitor node resources! For example, running
          processes.* \n### MySQL credentials\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-full-secrets\ndata:\n  mysql_root_password: emFiYml4X3N1cGVyX3Jvb3Q=
          [1]\n  mysql_zabbix_password: emFiYml4X3VzZXJfcGFzc3dvcmQ= [2]\n  mysql_zabbix_username:
          emFiYml4 [3]\ntype: Opaque\n```\nUsing MySQL root password (*mysqlrootpassword*
          [1]) Zabbix server / proxy will try to create MySQL database schema with
          grant permissions to *mysqlzabbixusername* [2] and *mysqlzabbixpassword*
          [3].\n### SSL certificates for HTTPS\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-web-sslsecret\ndata:\n  ssl.crt: >-\n   < ssl.crt data>\n
          \ ssl.key: >-\n\t < ssl.key data >\n  dhparam.pem: >-\n   <  dhparam.pem
          data >\n```\nFiles *ssl.crt*, *ssl.key* and *dhparam.perm* are required
          for Zabbix web-interface for SSL support.\n"
        displayName: Zabbix Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - zabbix
        - appliance
        - agent
        - metrics
        links:
        - name: Zabbix
          url: https://www.zabbix.com
        - name: Zabbix Official Documentation
          url: https://www.zabbix.com/documentation/6.0/en/manual/quickstart
        - name: Zabbix Operator Installation
          url: https://www.zabbix.com/documentation/6.0/en/manual/installation/containers/openshift
        - name: Downloads
          url: https://www.zabbix.com/download
        maintainers:
        - email: alexey.pustovalov@zabbix.com
          name: Alexey Pustovalov
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Zabbix LLC
        relatedImages:
        - registry.connect.redhat.com/zabbix/zabbix-server-mysql-60@sha256:637484faec6e9a55049c13dbd1f100c03d757c3a21749388c0605040f02d05ce
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
        - registry.connect.redhat.com/zabbix/zabbix-java-gateway-60@sha256:a4feb725c12c4ac90eb9244c6fafdb6708679f2e0e7f6aa683c765127306412e
        - registry.connect.redhat.com/zabbix/zabbix-agent-60@sha256:2895b5e45785aaf528d4a6f78c8af0a0d494dbe9362403fdd470673d911d5004
        - registry.connect.redhat.com/zabbix/zabbix-agent2-60@sha256:424003d52f3cb4360adf2b3619d8b9f8de9196c630ee7e4ceb5d177c1c5a36e4
        - quay.io/zabbix/zabbix-operator@sha256:43d26488a19967bdb2a8aafde5f43820be95011811da9cc5ec623c9029801675
        - registry.redhat.io/rhel9/mysql-80@sha256:c4b5246d0c0d089a766ea07768c22d9d6be8049ac1379bd345db99390ffe2a36
        - registry.connect.redhat.com/zabbix/zabbix-snmptraps-60@sha256:e61e6b1c551dec8488cce0998f94d83d6d10b5d89a465d7119b088313032cb81
        - registry.connect.redhat.com/zabbix/zabbix-web-mysql-60@sha256:611d2670e1bf95b81226976c0a1f64434b2a7766f196dadbd646a7d8319f4cb1
        - registry.connect.redhat.com/zabbix/zabbix-proxy-mysql-60@sha256:c2ae74d97c083ff348dc9c55932a33da4779cd9efdd9d7c5a79acfffae5b9ed0
        - registry.connect.redhat.com/zabbix/zabbix-proxy-sqlite-60@sha256:62e62ebc8c1bc13fb610f7a83412e65d852f74b3c2c212baa75ea27fc7fa03eb
        - registry.connect.redhat.com/zabbix/zabbix-web-service-60@sha256:4a333094e110713470a897b6406c5e340e962caf32a429dddf19865041fcc448
        version: 6.0.40
      entries:
      - name: zabbix-operator-certified.v6.0.40
        version: 6.0.40
      - name: zabbix-operator-certified.v6.0.38
        version: 6.0.38
      - name: zabbix-operator-certified.v6.0.36
        version: 6.0.36
      - name: zabbix-operator-certified.v6.0.33
        version: 6.0.33
      - name: zabbix-operator-certified.v6.0.29
        version: 6.0.29
      - name: zabbix-operator-certified.v6.0.27
        version: 6.0.27
      - name: zabbix-operator-certified.v6.0.26
        version: 6.0.26
      - name: zabbix-operator-certified.v6.0.17
        version: 6.0.17
      - name: zabbix-operator-certified.v6.0.14
        version: 6.0.14
      - name: zabbix-operator-certified.v6.0.13
        version: 6.0.13
      name: lts
    defaultChannel: lts
    packageName: zabbix-operator-certified
    provider:
      name: Zabbix LLC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/multiarch-tuning-operator
    name: multiarch-tuning-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multiarch-tuning-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multiarch.openshift.io/v1beta1",
                "kind": "ClusterPodPlacementConfig",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logVerbosity": "Normal",
                  "namespaceSelector": {
                    "matchExpressions": [
                      {
                        "key": "multiarch.openshift.io/exclude-pod-placement",
                        "operator": "DoesNotExist"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Other
          console.openshift.io/disable-operand-delete: "false"
          containerImage: registry.redhat.io/multiarch-tuning/multiarch-tuning-rhel9-operator@sha256:1a1d3f5258d4b91f1a0380b0fe2c809823e323d9ff74ae4171468168d255e7e0
          createdAt: "2025-03-05T12:27:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operator.openshift.io/uninstall-message: You must remove all Operands before
            uninstalling the operator.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-multiarch-tuning-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/multiarch-tuning-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPodPlacementConfig defines the configuration for the
              architecture aware pod placement operand. Users can only deploy a single
              object named "cluster". Creating the object enables the operand.
            displayName: Cluster Pod Placement Config
            kind: ClusterPodPlacementConfig
            name: clusterpodplacementconfigs.multiarch.openshift.io
            version: v1alpha1
          - description: ClusterPodPlacementConfig defines the configuration for the
              architecture aware pod placement operand. Users can only deploy a single
              object named "cluster". Creating the object enables the operand.
            displayName: Cluster Pod Placement Config
            kind: ClusterPodPlacementConfig
            name: clusterpodplacementconfigs.multiarch.openshift.io
            version: v1beta1
        description: |
          The Multiarch Tuning Operator optimizes workload management within multi-architecture clusters and in
          single-architecture clusters transitioning to multi-architecture environments.

          By default, the scheduler does not consider the architecture of a pod's container images
          when determining the placement of new pods onto nodes.

          The Multiarch Tuning Operator introduces the ClusterPodPlacementConfig custom resource to
          enable architecture-aware scheduling, ensuring that pods are assigned to nodes with
          a corresponding architecture for their container images.

          To enable this functionality, you must create a ClusterPodPlacementConfig resource.

          When the ClusterPodPlacementConfig object is created, the Multiarch Tuning Operator deploys the necessary operands
          to support architecture-aware workload scheduling.
        displayName: Multiarch Tuning Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multiarch-tuning-operator
        - pod-placement
        links:
        - name: Multiarch Tuning Operator Documentation
          url: https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/postinstallation_configuration/configuring-multi-architecture-compute-machines-on-an-openshift-cluster#multiarch-tuning-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/multiarch-tuning-operator
        relatedImages:
        - registry.redhat.io/multiarch-tuning/multiarch-tuning-rhel9-operator@sha256:1a1d3f5258d4b91f1a0380b0fe2c809823e323d9ff74ae4171468168d255e7e0
        version: 1.1.0
      entries:
      - name: multiarch-tuning-operator.v1.1.0
        version: 1.1.0
      - name: multiarch-tuning-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: multiarch-tuning-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/multiarch-tuning-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: infra-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infra-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instanceha.openstack.org/v1beta1",
                "kind": "InstanceHa",
                "metadata": {
                  "name": "instanceha"
                },
                "spec": {
                  "caBundleSecretName": "combined-ca-bundle"
                }
              },
              {
                "apiVersion": "memcached.openstack.org/v1beta1",
                "kind": "Memcached",
                "metadata": {
                  "name": "memcached"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "network.openstack.org/v1beta1",
                "kind": "DNSData",
                "metadata": {
                  "name": "allhosts"
                },
                "spec": {
                  "hosts": [
                    {
                      "hostnames": [
                        "edpm-compute-0.ctlplane"
                      ],
                      "ip": "192.168.122.100"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-0.internalapi"
                      ],
                      "ip": "172.17.0.100"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-0.storage"
                      ],
                      "ip": "172.18.0.100"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-0.tenant"
                      ],
                      "ip": "172.19.0.100"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-1.ctlplane"
                      ],
                      "ip": "192.168.122.101"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-1.internalapi"
                      ],
                      "ip": "172.17.0.101"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-1.storage"
                      ],
                      "ip": "172.18.0.101"
                    },
                    {
                      "hostnames": [
                        "edpm-compute-1.tenant"
                      ],
                      "ip": "172.19.0.101"
                    },
                    {
                      "hostnames": [
                        "keystone-internal.openstack.svc",
                        "placement-internal.openstack.svc",
                        "cinder-internal.openstack.svc",
                        "glance-internal.openstack.svc",
                        "nova-internal.openstack.svc",
                        "nova-metadata-internal.openstack.svc",
                        "neutron-internal.openstack.svc"
                      ],
                      "ip": "172.17.0.80"
                    },
                    {
                      "hostnames": [
                        "rabbitmq.openstack.svc"
                      ],
                      "ip": "172.17.0.85"
                    },
                    {
                      "hostnames": [
                        "rabbitmq-cell1.openstack.svc"
                      ],
                      "ip": "172.17.0.86"
                    }
                  ]
                }
              },
              {
                "apiVersion": "network.openstack.org/v1beta1",
                "kind": "DNSMasq",
                "metadata": {
                  "name": "dnsmasq"
                },
                "spec": {
                  "debug": {
                    "service": false
                  },
                  "options": [
                    {
                      "key": "server",
                      "values": [
                        "192.168.122.1"
                      ]
                    },
                    {
                      "key": "local",
                      "values": [
                        "/example.com/"
                      ]
                    }
                  ],
                  "override": {
                    "service": {
                      "metadata": {
                        "annotations": {
                          "metallb.universe.tf/address-pool": "ctlplane",
                          "metallb.universe.tf/allow-shared-ip": "ctlplane",
                          "metallb.universe.tf/loadBalancerIPs": "192.168.122.80"
                        }
                      },
                      "spec": {
                        "type": "LoadBalancer"
                      }
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "network.openstack.org/v1beta1",
                "kind": "IPSet",
                "metadata": {
                  "name": "edpm-compute-0"
                },
                "spec": {
                  "networks": [
                    {
                      "defaultRoute": true,
                      "name": "CtlPlane",
                      "subnetName": "subnet1"
                    },
                    {
                      "name": "InternalApi",
                      "subnetName": "subnet1"
                    },
                    {
                      "name": "Storage",
                      "subnetName": "subnet1"
                    },
                    {
                      "name": "Tenant",
                      "subnetName": "subnet1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "network.openstack.org/v1beta1",
                "kind": "NetConfig",
                "metadata": {
                  "name": "netconfig"
                },
                "spec": {
                  "networks": [
                    {
                      "dnsDomain": "ctlplane.example.com",
                      "name": "ctlplane",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "192.168.122.120",
                              "start": "192.168.122.100"
                            },
                            {
                              "end": "192.168.122.200",
                              "start": "192.168.122.150"
                            }
                          ],
                          "cidr": "192.168.122.0/24",
                          "gateway": "192.168.122.1",
                          "name": "subnet1"
                        }
                      ]
                    },
                    {
                      "dnsDomain": "internalapi.example.com",
                      "name": "internalapi",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "172.17.0.250",
                              "start": "172.17.0.100"
                            }
                          ],
                          "cidr": "172.17.0.0/24",
                          "name": "subnet1",
                          "vlan": 20
                        }
                      ]
                    },
                    {
                      "dnsDomain": "external.example.com",
                      "name": "external",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "10.0.0.250",
                              "start": "10.0.0.100"
                            }
                          ],
                          "cidr": "10.0.0.0/24",
                          "gateway": "10.0.0.1",
                          "name": "subnet1"
                        }
                      ]
                    },
                    {
                      "dnsDomain": "storage.example.com",
                      "name": "storage",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "172.18.0.250",
                              "start": "172.18.0.100"
                            }
                          ],
                          "cidr": "172.18.0.0/24",
                          "name": "subnet1",
                          "vlan": 21
                        }
                      ]
                    },
                    {
                      "dnsDomain": "storagemgmt.example.com",
                      "name": "storagemgmt",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "172.20.0.250",
                              "start": "172.20.0.100"
                            }
                          ],
                          "cidr": "172.20.0.0/24",
                          "name": "subnet1",
                          "vlan": 23
                        }
                      ]
                    },
                    {
                      "dnsDomain": "tenant.example.com",
                      "name": "tenant",
                      "subnets": [
                        {
                          "allocationRanges": [
                            {
                              "end": "172.19.0.250",
                              "start": "172.19.0.100"
                            }
                          ],
                          "cidr": "172.19.0.0/24",
                          "name": "subnet1",
                          "vlan": 22
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "network.openstack.org/v1beta1",
                "kind": "Reservation",
                "metadata": {
                  "name": "edpm-compute-0"
                },
                "spec": {
                  "ipSetRef": {
                    "name": "edpm-compute-0",
                    "namespace": "openstack",
                    "uid": "56753ffb-e486-43ab-a79a-f85f27b2ac14"
                  },
                  "reservation": {
                    "CtlPlane": {
                      "address": "192.168.122.100",
                      "network": "CtlPlane",
                      "subnet": "subnet1"
                    },
                    "InternalApi": {
                      "address": "172.17.0.100",
                      "network": "InternalApi",
                      "subnet": "subnet1"
                    },
                    "Storage": {
                      "address": "172.18.0.100",
                      "network": "Storage",
                      "subnet": "subnet1"
                    },
                    "StorageMgmt": {
                      "address": "172.20.0.100",
                      "network": "StorageMgmt",
                      "subnet": "subnet1"
                    },
                    "Tenant": {
                      "address": "172.19.0.100",
                      "network": "Tenant",
                      "subnet": "subnet1"
                    }
                  }
                }
              },
              {
                "apiVersion": "rabbitmq.openstack.org/v1beta1",
                "kind": "TransportURL",
                "metadata": {
                  "name": "transporturl"
                },
                "spec": {
                  "rabbitmqClusterName": "rabbitmq"
                }
              },
              {
                "apiVersion": "redis.openstack.org/v1beta1",
                "kind": "Redis",
                "metadata": {
                  "name": "redis"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:45Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSData is the Schema for the dnsdata API
            displayName: DNSData
            kind: DNSData
            name: dnsdata.network.openstack.org
            version: v1beta1
          - description: DNSMasq is the Schema for the dnsmasqs API
            displayName: DNSMasq
            kind: DNSMasq
            name: dnsmasqs.network.openstack.org
            version: v1beta1
          - description: InstanceHa is the Schema for the instancehas API
            displayName: Instance Ha
            kind: InstanceHa
            name: instancehas.instanceha.openstack.org
            version: v1beta1
          - description: IPSet is the Schema for the ipsets API
            displayName: IPSet
            kind: IPSet
            name: ipsets.network.openstack.org
            version: v1beta1
          - description: Memcached is the Schema for the memcacheds API
            displayName: Memcached
            kind: Memcached
            name: memcacheds.memcached.openstack.org
            version: v1beta1
          - description: NetConfig is the Schema for the netconfigs API
            displayName: Net Config
            kind: NetConfig
            name: netconfigs.network.openstack.org
            version: v1beta1
          - description: Redis is the Schema for the redises API
            displayName: Redis
            kind: Redis
            name: redises.redis.openstack.org
            version: v1beta1
          - description: Reservation is the Schema for the reservations API
            displayName: Reservation
            kind: Reservation
            name: reservations.network.openstack.org
            version: v1beta1
          - description: TransportURL is the Schema for the transporturls API
            displayName: Transport URL
            kind: TransportURL
            name: transporturls.rabbitmq.openstack.org
            version: v1beta1
        description: OpenStack Infrastructure
        displayName: OpenStack Infra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Infrastructure
        links:
        - name: Infra Operator
          url: https://github.com/openstack-k8s-operators/infra-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/infra-rhel9-operator@sha256:177077029ad8baa959e41a785cd29ed63745e8f39be618aa97509e6867fbff7c
        - registry.redhat.io/rhoso/openstack-memcached-rhel9@sha256:962a314226009a8ca5c95fb66a75b15176506e02110847d0f42aeed6dd26a751
        - registry.redhat.io/rhoso/openstack-redis-rhel9@sha256:b83d23df9f2b367e604f404d97574fd6db9e0ed530eda23e9e0bbe01292c8db8
        - registry.redhat.io/rhoso/openstack-neutron-server-rhel9@sha256:806e5e009a77a29c82699487e539669e1ec2136514a0332d222b5cb3fed6f452
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: infra-operator.v1.0.6
        version: 1.0.6
      - name: infra-operator.v1.0.5
        version: 1.0.5
      - name: infra-operator.v1.0.4
        version: 1.0.4
      - name: infra-operator.v1.0.3
        version: 1.0.3
      - name: infra-operator.v1.0.2
        version: 1.0.2
      - name: infra-operator.v1.0.1
        version: 1.0.1
      - name: infra-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: infra-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: service-binding-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-binding-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:b355d2370926063a3de8c7a09c0ca8c2348a8d158d6a7095d0e5e8381cc95be5
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/redhat-developer/servicebinding-operator@sha256:b355d2370926063a3de8c7a09c0ca8c2348a8d158d6a7095d0e5e8381cc95be5
        version: 0.10.0
      entries:
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: beta
    - currentCSV: service-binding-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindablekinds-sample"
                }
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:c9c9e45321db2d4d3cdc2d364337af844f69c9893d0cc74d2fa56c846ffcf3da
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/redhat-developer/servicebinding-operator@sha256:c9c9e45321db2d4d3cdc2d364337af844f69c9893d0cc74d2fa56c846ffcf3da
        version: 0.11.0
      entries:
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: candidate
    - currentCSV: service-binding-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindable-kinds"
                },
                "spec": {}
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  },
                  "type": "postgres",
                  "workload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:ac47f496fb7ecdcbc371f8c809fad2687ec0c35bbc8c522a7ab63b3e5ffd90ea
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'.
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1alpha3
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1beta1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
          - description: Service Binding implementing community specification (https://github.com/k8s-service-bindings/spec)
              describing the connection between a backing service and an application
              workload.
            displayName: Service Binding (Spec API)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1beta1
        description: |
          The Service Binding Operator manages the data plane for applications and backing services. It enables developers to connect their
          applications to backing services (REST API, databases, event buses and others) with a consistent and predictable
          experience. The operator reads data made available by the control plane of backing services and projects the bindings
          data to applications either as accessible files mounted into the application’s container or environment variables.

          In Kubernetes, each service suggests a different way to access their secrets and developers will consume them in their
          application in a custom way. While this provides a good deal of flexibility, it also creates discrepancies and
          complexity dealing with each specific services, which leads large development teams to lose overall velocity.

          Service Binding Operator removes the error-prone manual configuration of binding information, provides service operators
          a low-touch administrative experience to provision and manage access to services and enriches development lifecycle.

          For additional details refer to the project [documentation](https://redhat-developer.github.io/service-binding-operator).

          Get Started with Service Binding Operator
          * Find out about SBO’s [prerequisites](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/installing-service-binding.html).
          * Refer to the [Quick Start Guide](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/quick-start.html) to see SBO in action in a simple scenario.

          Need Help
          * Raise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)

          Licence
          * Service Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE)

          Service Binding Operator is providing and supporting two different APIs.
          * `binding.operators.coreos.com/v1alpha1`: This API is compliant with the Service Binding specification. It is fully supported and considered as the mature API. It’s the API we recommend in production environments.
          * `servicebinding.io/v1beta1`: This API is the one defined in the Service Binding specification. It is available in Service Binding Operator 1.x.

          We’ll consider supporting only the API from the specification, once the Service Binding specification will be officially GA. Of course, that will be a breaking change and handle in a safe way for you.
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Documentation
          url: https://redhat-developer.github.io/service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/redhat-developer/servicebinding-operator@sha256:ac47f496fb7ecdcbc371f8c809fad2687ec0c35bbc8c522a7ab63b3e5ffd90ea
        version: 1.4.0
      entries:
      - name: service-binding-operator.v1.4.0
        version: 1.4.0
      - name: service-binding-operator.v1.3.4
        version: 1.3.4
      - name: service-binding-operator.v1.3.3
        version: 1.3.3
      - name: service-binding-operator.v1.3.2
        version: 1.3.2
      - name: service-binding-operator.v1.3.1
        version: 1.3.1
      - name: service-binding-operator.v1.3.0
        version: 1.3.0
      - name: service-binding-operator.v1.2.0
        version: 1.2.0
      - name: service-binding-operator.v1.1.1
        version: 1.1.1
      - name: service-binding-operator.v1.1.0
        version: 1.1.0
      - name: service-binding-operator.v1.0.1
        version: 1.0.1
      - name: service-binding-operator.v1.0.0
        version: 1.0.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: service-binding-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Skupper Project
      provider-url: https://skupper.io
    name: skupper-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: alpha
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1
    - currentCSV: skupper-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:e5f8b2ddf378cf3e40f4c7b7ec8482eb3321947cd46481c868d727932c449d78
          createdAt: "2024-02-26T18:22:07Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/prometheus/prometheus@sha256:d2ab0a27783fd4ad96a8853e2847b99a0be0043687b8a5d1ebfb2dd3fa4fd1b8
        - quay.io/skupper/site-controller@sha256:e5f8b2ddf378cf3e40f4c7b7ec8482eb3321947cd46481c868d727932c449d78
        - quay.io/openshift/origin-oauth-proxy@sha256:a7dff785d821acc4b972d94964947713fade45f007dedb5deee17a2869bf5d24
        - quay.io/skupper/skupper-router@sha256:bf8d5955ce4e406275a4934278efe51c1af9a1a269a70e8fe642b7d94d80cb81
        - quay.io/skupper/service-controller@sha256:345edf2d9d72b5a97bcce0e8cea8cda82d120f7bf83756d73f9ac114c2fcced8
        - quay.io/skupper/config-sync@sha256:659fe28d1df7c99bba3bb854b1e9991c715ae4de99d91c88d5db5e4f9ef4c852
        - quay.io/skupper/flow-collector@sha256:3870c5bc89e623b7cebe21856bfefc22de76313705efa0ec551a5963789fff3b
        version: 1.6.0
      entries:
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.6
    - currentCSV: skupper-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:4809db7263d77a95797061798e910f1bdbb5d5fe2db2bb3c743b97f5d0c7b3af
          createdAt: "2024-06-12T14:36:59Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/site-controller@sha256:4809db7263d77a95797061798e910f1bdbb5d5fe2db2bb3c743b97f5d0c7b3af
        - quay.io/openshift/origin-oauth-proxy@sha256:a7dff785d821acc4b972d94964947713fade45f007dedb5deee17a2869bf5d24
        - quay.io/skupper/skupper-router@sha256:5eb258a0927e7706e7c035f1def2eb5211c8c33dd371fafa7210517fd894f68b
        - quay.io/skupper/service-controller@sha256:624e88e695297df28e485db7e802945ad1978ac6b0c9e9edca8b5e6013af8991
        - quay.io/skupper/config-sync@sha256:a884b26a32179af5a2bfa85118bf873852eec6e2930e92da778b57506c0a0880
        - quay.io/skupper/flow-collector@sha256:bf646b0cb75e1c29bbe0261aa9943e06d33118f5e6a5ccb7503c60c974167692
        - quay.io/prometheus/prometheus@sha256:d2ab0a27783fd4ad96a8853e2847b99a0be0043687b8a5d1ebfb2dd3fa4fd1b8
        version: 1.7.1
      entries:
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.7
    - currentCSV: skupper-operator.v1.8.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:ffe3ea5067d350b64b69120bdf6d9e3ee9a30e8a2559b77170a10d66635623b8
          createdAt: "2025-03-27T20:11:53Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/skupper-router@sha256:7633b143f1844ee559322e4642ca80e8248a08711c8e814d489c26c39ecf9bd0
        - quay.io/skupper/service-controller@sha256:a45de211b8f8a8294a3588932a0a72766cf0d5ce64b01f4f744832c8d09aaa46
        - quay.io/skupper/config-sync@sha256:9e96d6a585f037f60ca3d89b5efee27ec92444d353e8b289ca35b94aeed1e309
        - quay.io/skupper/flow-collector@sha256:252244ab6aaf6b80118bfc06b690d4f5e8d7829b999fca60d194a84452fee7f6
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/openshift/origin-oauth-proxy@sha256:c15baf57dbd17dcc517cd2fa2299462d62124f689d6f5a48a3c9fd154fc77ccf
        - quay.io/skupper/site-controller@sha256:ffe3ea5067d350b64b69120bdf6d9e3ee9a30e8a2559b77170a10d66635623b8
        version: 1.8.4
      entries:
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.8
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.9
    defaultChannel: stable
    packageName: skupper-operator
    provider:
      name: Skupper Project
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Harness Inc.
      provider-url: https://harness.io
    name: harness-gitops-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: harness-gitops-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gitops.harness.io/v1alpha1",
                "kind": "GitopsAgent",
                "metadata": {
                  "annotations": {
                    "containerImage": "harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde",
                    "createdAt": "2025-07-31T12:11:18Z"
                  },
                  "name": "gitops-agent-sample"
                },
                "spec": {
                  "identity": {
                    "accountIdentifier": "accountIdentifier",
                    "agentIdentifier": "agentIdentifier",
                    "agentName": "agentName"
                  },
                  "replicas": 1,
                  "secrets": {
                    "agentSecret": "agentSecret"
                  },
                  "target": {
                    "http": "https://app.harness.io/gitops/",
                    "protocol": "HTTP1"
                  },
                  "upgrade": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde
          createdAt: "2025-07-31T06:41:18Z"
          description: Harness operator to manage and provision GitOps Agent
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Harness Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Schema for the Gitops Agents API
            displayName: Gitops Agent
            kind: GitopsAgent
            name: gitopsagents.gitops.harness.io
            version: v1alpha1
        description: |-
          ## Overview
          Harness Gitops Operator manages the full lifecycle for [Gitops Agent](https://developer.harness.io/docs/continuous-delivery/gitops/connect-and-manage/install-a-harness-git-ops-agent/)

          The operator aims to provide the following:

            * Easy configuration and installation of Harness GitOps Agent.
            * Provide seamless upgrades for Harness GitOps Agent.
            * Monitoring and autorecovery for GitOps Agent.

          ## Usage
          Deploy gitops-agent in namespace where operator is installed with basic setup.

          ### Prerequisites:
            * Go to your Harness account and create a new gitops agent
            * Copy account, org and project identifiers
            * Retrieve the agent secret key

          ```
          apiVersion: gitops.harness.io/v1alpha1
          kind: GitopsAgent
          metadata:
            name: gitops-agent-sample
          spec:
            identity:
              accountIdentifier: accountIdentifier
              agentIdentifier: agentIdentifier
              agentName: agentName
            secrets:
              agentSecret: agentSecret
            target:
              http: https://app.harness.io/gitops/
              protocol: HTTP1
            replicas: 1
            upgrade:
              enabled: true
          ```
          **Note:** Gitops Agent require an active ArgoCD instance running in the same namespace where agent is installed.

          See [Harness Documentation](https://developer.harness.io/) for more information.
        displayName: Harness GitOps Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - harness
        - agent
        - cd
        - argocd
        - kubernetes
        links:
        - name: Harness Gitops Operator
          url: https://harness-gitops-operator.domain
        maintainers:
        - email: alexandru.bosii@harness.io
          name: Alexandru Bosii
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Harness Inc.
          url: https://harness.io
        relatedImages:
        - harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde
        - docker.io/harness/gitops-agent@sha256:c9f57338be294d32a73a4cbecf90dfff50f5d36a699f3ec32d962dcb64d8d085
        version: 0.3.0
      entries:
      - name: harness-gitops-operator.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: harness-gitops-operator
    provider:
      name: Harness Inc.
      url: https://harness.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: maximilianopizarro
      provider-url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
    name: jhipster-online-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jhipster-online-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maximilianopizarro.github.io/v1alpha1",
                "kind": "JhipsterOnline",
                "metadata": {
                  "name": "jhipsteronline-sample"
                },
                "spec": {
                  "affinity": {},
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "env": {
                    "APPLICATION_GITHUB_CLIENT-ID": "CLIENT-ID",
                    "APPLICATION_GITHUB_CLIENT-SECRET": "CLIENT-SECRET",
                    "APPLICATION_GITHUB_HOST": "https://github.com",
                    "APPLICATION_JHIPSTER-CMD_CMD": "jhipster-quarkus",
                    "JAVA_APP_JAR": "jhonline-2.33.0.war",
                    "SPRING_DATASOURCE_PASSWORD": "jhipster",
                    "SPRING_DATASOURCE_URL": "jdbc:mariadb://mariadb:3306/jhipsteronline",
                    "SPRING_DATASOURCE_USERNAME": "jhipster"
                  },
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "quay.io/maximilianopizarro/jhipster-online",
                    "tag": "quarkus"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "livenessProbe": {
                    "httpGet": {
                      "path": "/jdl-studio/",
                      "port": 8080
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "readinessProbe": {
                    "httpGet": {
                      "path": "/jdl-studio/",
                      "port": 8080
                    }
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "securityContext": {},
                  "service": {
                    "port": 8080,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automount": false,
                    "create": false,
                    "name": ""
                  },
                  "tolerations": [],
                  "volumeMounts": [],
                  "volumes": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/maximilianopizarro/jhipster-online-operator:v0.1.0
          createdAt: "2025-05-30T22:42:19Z"
          description: JHipster Online provides an intuitive web interface for generating
            JHipster projects, allowing you to define your application architecture
            using a JDL (JHipster Domain Language) model and generate the code directly
            into a GitHub repository.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/maximilianoPizarro/jhipster-online-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: JhipsterOnline
            name: jhipsteronlines.maximilianopizarro.github.io
            version: v1alpha1
        description: |
          JHipster Online Operator: This repository hosts the JHipster Online Operator, designed to seamlessly deploy JHipster Online on Red Hat OpenShift and Kubernetes environments.

          JHipster Online provides an intuitive web interface for generating JHipster projects, allowing you to define your application architecture using a JDL (JHipster Domain Language) model and generate the code directly into a GitHub repository.

          Overview: The JHipster Online Operator manages the deployment and lifecycle of the JHipster Online application. It leverages the power of Kubernetes Operators to provide a native Kubernetes experience for JHipster Online users.

          This artifact includes:
          - JHipster 8.8.0. for generate Spring Boot 3.4.1 projects.
          - generator-jhipster-quarkus 3.4.0 for generate Quarkus 3.11.1 projects.
          - JDL Studio for add JDL models files by PR on your repo.

          Usage: Once the operator is running, you can create instances of JHipster Online by defining `JhipsterOnline` custom resources. An example `JhipsterOnline` custom resource is provided in the `alm-examples`. You would apply this (or a similar) YAML file to your cluster to create an instance of JHipster Online. Feel free to customize the `JhipsterOnline` Custom Resource Definition (CRD) instance to suit your needs, such as configuring GitHub integration, database connections, and resource allocations.

          Configuration: The `JhipsterOnline` custom resource supports various configuration options including `replicaCount`, `image`, `imagePullSecrets`, `serviceAccount`, `podAnnotations`, `podLabels`, `podSecurityContext`, `securityContext`, `service` type and port, `ingress` settings (enabled, `className`, `annotations`, `hosts`, `tls`), OpenShift `Route` (enabled), `autoscaling` (enabled, `minReplicas`, `maxReplicas`, `targetCPUUtilizationPercentage`), `resource` requests and limits, `nodeSelector`, `tolerations`, `affinity`, `liveness` and `readiness` probes, `env` variables, `volumes`, and `volumeMounts`. Refer to the detailed `README.md` for a full table of parameters.
        displayName: JHipster Online Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - java
        - spring-boot
        - jhipster
        - quarkus
        links:
        - name: Jhipster Online Operator
          url: https://maximilianopizarro.github.io/jhipster-online-operator/
        maintainers:
        - email: maximiliano.pizarro.5@gmail.com
          name: Maximiliano Pizarro
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: maximilianopizarro
          url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/maximilianopizarro/jhipster-online-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: jhipster-online-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: jhipster-online-operator
    provider:
      name: maximilianopizarro
      url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: postgresql
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: registry.developers.crunchydata.com/crunchydata/postgres-operator:ubi9-5.8.3-0
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.developers.crunchydata.com/crunchydata/postgres-operator:ubi9-5.8.3-0
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: postgresql
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Vertica
      provider-url: http://www.vertica.com
    name: verticadb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: verticadb-operator.v1.11.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "EventTrigger",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "verticadb-operator",
                    "app.kubernetes.io/instance": "eventtrigger-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eventtrigger",
                    "app.kubernetes.io/part-of": "verticadb-operator"
                  },
                  "name": "eventtrigger-sample"
                },
                "spec": {
                  "matches": [
                    {
                      "condition": {
                        "status": "True",
                        "type": "DBInitialized"
                      }
                    }
                  ],
                  "references": [
                    {
                      "object": {
                        "apiVersion": "vertica.com/v1beta1",
                        "kind": "VerticaDB",
                        "name": "verticadb-sample"
                      }
                    }
                  ],
                  "template": {
                    "metadata": {
                      "generateName": "create-user-table-"
                    },
                    "spec": {
                      "template": {
                        "spec": {
                          "containers": [
                            {
                              "command": [
                                "/opt/vertica/bin/vsql",
                                "-h",
                                "verticadb-sample-defaultsubcluster",
                                "-c",
                                "CREATE TABLE T1 (C1 INT);"
                              ],
                              "image": "vertica/vertica-k8s:latest",
                              "name": "main"
                            }
                          ],
                          "restartPolicy": "OnFailure"
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "serviceName": "defaultsubcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaDB",
                "metadata": {
                  "name": "verticadb-sample"
                },
                "spec": {
                  "communal": {
                    "endpoint": "https://s3.amazonaws.com"
                  },
                  "image": "vertica/vertica-k8s:23.3.0-0-minimal",
                  "subclusters": [
                    {
                      "name": "default_subcluster",
                      "size": 3
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: docker.io/vertica/verticadb-operator:1.11.2
          createdAt: "2023-08-02T11:29:39Z"
          description: Operator that manages a Vertica Eon Mode database.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/vertica/vertica-kubernetes
          support: Vertica
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventTrigger is the Schema for the eventtriggers API
            displayName: Event Trigger
            kind: EventTrigger
            name: eventtriggers.vertica.com
            version: v1beta1
          - description: VerticaAutoscaler is a CR that allows you to autoscale one
              or more subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1beta1
          - description: VerticaDB is the CR that defines a Vertica Eon mode cluster
              that is managed by the verticadb-operator.
            displayName: Vertica DB
            kind: VerticaDB
            name: verticadbs.vertica.com
            version: v1beta1
        description: |-
          ### What is Vertica?
          Vertica is a unified analytics platform, based on a massively scalable architecture with the broadest set of analytical functions spanning event and time series, pattern matching, geospatial and end-to-end in-database machine learning. Vertica enables you to easily apply these powerful functions to the largest and most demanding analytical workloads, arming you and your customers with predictive business insights faster than any analytics data warehouse in the market. Vertica provides a unified analytics platform across major public clouds and on-premises data centers and integrates data in cloud object storage and HDFS without forcing you to move any of your data.

          ### What does this operator do?
          This operator manages a Vertica Eon Mode database, and automates the following administrative tasks:

          * Installing Vertica
          * Creating and reviving a Vertica database
          * Restarting and rescheduling DOWN pods to maintain quorum
          * Upgrading Vertica to a new version while keeping the database online
          * Subcluster scaling, both manually and automatically with the Horizontal Pod Autoscaler
          * Service management and health monitoring for pods
          * Load balancing for internal and external traffic
          * Expose a metrics endpoint to integrate with Prometheus to monitor the operator's health

          For a brief overview on how to create a cluster, see the [Vertica GitHub repository](https://github.com/vertica/vertica-kubernetes). For an in-depth look at Vertica on Kubernetes, see the [Vertica documentation](https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm).
        displayName: VerticaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - database
        - vertica
        links:
        - name: Documentation
          url: https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm
        - name: Vertica Container Images
          url: https://hub.docker.com/u/vertica
        maintainers:
        - email: matt.spilchen@vertica.com
          name: Matt Spilchen
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Vertica
          url: http://www.vertica.com
        relatedImages:
        - docker.io/vertica/verticadb-operator:1.11.2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 1.11.2
      entries:
      - name: verticadb-operator.v1.11.2
        version: 1.11.2
      - name: verticadb-operator.v1.11.1
        version: 1.11.1
      - name: verticadb-operator.v1.11.0
        version: 1.11.0
      - name: verticadb-operator.v1.10.2
        version: 1.10.2
      - name: verticadb-operator.v1.10.1
        version: 1.10.1
      - name: verticadb-operator.v1.10.0
        version: 1.10.0
      - name: verticadb-operator.v1.9.0
        version: 1.9.0
      - name: verticadb-operator.v1.8.0
        version: 1.8.0
      - name: verticadb-operator.v1.7.0
        version: 1.7.0
      - name: verticadb-operator.v1.6.0
        version: 1.6.0
      - name: verticadb-operator.v1.5.0
        version: 1.5.0
      - name: verticadb-operator.v1.4.0
        version: 1.4.0
      - name: verticadb-operator.v1.3.1
        version: 1.3.1
      - name: verticadb-operator.v1.3.0
        version: 1.3.0
      - name: verticadb-operator.v1.2.0
        version: 1.2.0
      name: stable
    - currentCSV: verticadb-operator.v25.3.1-0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vertica.com/v1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "customAutoscaler": {},
                  "scalingGranularity": "Subcluster",
                  "serviceName": "default-subcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1",
                "kind": "VerticaDB",
                "metadata": {
                  "name": "verticadb-sample"
                },
                "spec": {
                  "communal": {
                    "endpoint": "https://s3.amazonaws.com"
                  },
                  "image": "opentext/vertica-k8s:24.1.0-0-minimal",
                  "subclusters": [
                    {
                      "name": "default_subcluster",
                      "size": 3
                    }
                  ]
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "EventTrigger",
                "metadata": {
                  "name": "eventtrigger-sample"
                },
                "spec": {
                  "matches": [
                    {
                      "condition": {
                        "status": "True",
                        "type": "DBInitialized"
                      }
                    }
                  ],
                  "references": [
                    {
                      "object": {
                        "apiVersion": "vertica.com/v1beta1",
                        "kind": "VerticaDB",
                        "name": "verticadb-sample"
                      }
                    }
                  ],
                  "template": {
                    "metadata": {
                      "generateName": "create-user-table-"
                    },
                    "spec": {
                      "template": {
                        "spec": {
                          "containers": [
                            {
                              "command": [
                                "/opt/vertica/bin/vsql",
                                "-h",
                                "verticadb-sample-defaultsubcluster",
                                "-c",
                                "CREATE TABLE T1 (C1 INT);"
                              ],
                              "image": "opentext/vertica-k8s:23.4.0-0-minimal",
                              "name": "main"
                            }
                          ],
                          "restartPolicy": "OnFailure"
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "customAutoscaler": {},
                  "scalingGranularity": "Subcluster",
                  "serviceName": "default-subcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaReplicator",
                "metadata": {
                  "name": "verticareplicator-sample"
                },
                "spec": {
                  "source": {
                    "verticaDB": "sourceDBName"
                  },
                  "target": {
                    "verticaDB": "targetDBName"
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaRestorePointsQuery",
                "metadata": {
                  "name": "verticarestorepointsquery-sample"
                },
                "spec": {
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaScrutinize",
                "metadata": {
                  "name": "verticascrutinize-sample"
                },
                "spec": {
                  "verticaDBName": "verticadb-sample"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: docker.io/opentext/verticadb-operator:25.3.1-0
          createdAt: "2025-09-24T00:58:06Z"
          description: Operator that manages a Vertica Eon Mode database.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/vertica/vertica-kubernetes
          support: Vertica
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventTrigger is the Schema for the eventtriggers API
            displayName: Event Trigger
            kind: EventTrigger
            name: eventtriggers.vertica.com
            version: v1beta1
          - description: |-
              VerticaAutoscaler is a CR that allows you to autoscale one or more
              subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1
          - description: |-
              VerticaAutoscaler is a CR that allows you to autoscale one or more
              subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1beta1
          - description: VerticaDB is the CR that defines a Vertica Eon mode cluster
              that is managed by the verticadb-operator.
            displayName: Vertica DB
            kind: VerticaDB
            name: verticadbs.vertica.com
            version: v1
          - description: VerticaReplicator is the Schema for the verticareplicators
              API
            displayName: Vertica Replicator
            kind: VerticaReplicator
            name: verticareplicators.vertica.com
            version: v1beta1
          - description: VerticaRestorePointsQuery is the Schema for the verticarestorepointsqueries
              API
            displayName: Vertica Restore Points Query
            kind: VerticaRestorePointsQuery
            name: verticarestorepointsqueries.vertica.com
            version: v1beta1
          - description: VerticaScrutinize is the schema for verticascrutinize API
            displayName: Vertica Scrutinize
            kind: VerticaScrutinize
            name: verticascrutinizers.vertica.com
            version: v1beta1
        description: |-
          ### What is Vertica?
          Vertica is a unified analytics platform, based on a massively scalable architecture with the broadest set of analytical functions spanning event and time series, pattern matching, geospatial and end-to-end in-database machine learning. Vertica enables you to easily apply these powerful functions to the largest and most demanding analytical workloads, arming you and your customers with predictive business insights faster than any analytics data warehouse in the market. Vertica provides a unified analytics platform across major public clouds and on-premises data centers and integrates data in cloud object storage and HDFS without forcing you to move any of your data.

          ### What does this operator do?
          This operator manages a Vertica Eon Mode database, and automates the following administrative tasks:

          * Installing Vertica
          * Creating and reviving a Vertica database
          * Restarting and rescheduling DOWN pods to maintain quorum
          * Upgrading Vertica to a new version while keeping the database online
          * Subcluster scaling, both manually and automatically with the Horizontal Pod Autoscaler
          * Service management and health monitoring for pods
          * Load balancing for internal and external traffic
          * Expose a metrics endpoint to integrate with Prometheus to monitor the operator's health

          For a brief overview on how to create a cluster, see the [Vertica GitHub repository](https://github.com/vertica/vertica-kubernetes). For an in-depth look at Vertica on Kubernetes, see the [Vertica documentation](https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm).
        displayName: VerticaDB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - vertica
        links:
        - name: Documentation
          url: https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm
        - name: Vertica Container Images
          url: https://hub.docker.com/u/vertica
        maintainers:
        - email: cchen8@opentext.com
          name: Cai Chen
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Vertica
          url: http://www.vertica.com
        relatedImages:
        - docker.io/opentext/verticadb-operator:25.3.1-0
        version: 25.3.1-0
      entries:
      - name: verticadb-operator.v25.3.1-0
        version: 25.3.1-0
      - name: verticadb-operator.v25.3.0-0
        version: 25.3.0-0
      - name: verticadb-operator.v25.2.1-0
        version: 25.2.1-0
      - name: verticadb-operator.v25.2.0-0
        version: 25.2.0-0
      - name: verticadb-operator.v25.1.0-0
        version: 25.1.0-0
      - name: verticadb-operator.v24.4.0-0
        version: 24.4.0-0
      - name: verticadb-operator.v2.2.0
        version: 2.2.0
      - name: verticadb-operator.v2.1.3
        version: 2.1.3
      - name: verticadb-operator.v2.1.2
        version: 2.1.2
      - name: verticadb-operator.v2.1.1
        version: 2.1.1
      - name: verticadb-operator.v2.1.0
        version: 2.1.0
      - name: verticadb-operator.v2.0.2
        version: 2.0.2
      - name: verticadb-operator.v2.0.1
        version: 2.0.1
      - name: verticadb-operator.v2.0.0
        version: 2.0.0
      name: v2-stable
    defaultChannel: v2-stable
    packageName: verticadb-operator
    provider:
      name: Vertica
      url: http://www.vertica.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: F5 Networks Inc.
      provider-url: ""
    name: f5-ipam-controller-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: f5-ipam-controller-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fic.f5.com/v1",
                "kind": "F5IpamCtlr",
                "metadata": {
                  "name": "f5ipamctlr-sample"
                },
                "spec": {
                  "args": {
                    "infoblox_grid_host": "IP_ADDRESS_OF_SERVER",
                    "infoblox_labels": "CIDR_Labels",
                    "infoblox_netview": "NetView",
                    "infoblox_password": "PASSWORD",
                    "infoblox_username": "USERNAME",
                    "infoblox_wapi_version": "VERSION",
                    "insecure": true,
                    "ip_range": "IP_RANGE_LABELS",
                    "log_level": "DEBUG_OR_INFO",
                    "orchestration": "kubernetes",
                    "provider": "INFOBLOX_OR_F5-IP-PROVIDER"
                  },
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "f5-ipam-controller",
                    "user": "f5networks",
                    "version": "latest"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "serviceAccount": {
                    "create": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
          createdAt: "2025-08-08T03:48:02Z"
          description: Operator to install F5 IPAM Controller.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/f5-ipam-controller
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This Operator installs F5 IPAM controller(FIC)
            kind: F5IpamCtlr
            name: f5ipamctlrs.fic.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 IPAM Controller in your OpenShift Cluster. This enables to configure and deploy FIC using Helm Charts.
          ## F5 IPAM Controller
          The F5 IPAM Controller is a Docker container that runs in an orchestration environment and interfaces with an IPAM system. It allocates IP addresses from an IPAM system’s address pool for hostnames in an orchestration environment. The F5 IPAM Controller watches orchestration-specific resources and consumes the hostnames within each resource.
          ## Documentation
          Refer to F5 IPAM Controller documentation
          - https://github.com/F5Networks/f5-ipam-controller/blob/main/README.md
          ## Prerequisites
          ### Configure Infoblox with network and netview [Provider: Infoblox]
          Refer Infoblox documentation
          ### Create persistent volume and persistent volume claim [Provider: f5-ip-provider]
          Create persistent volume and persistent volume claim
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/f5-ipam-controller/main/docs/config_examples/f5-ip-provider/localstorage-pv-pvc-example.yaml
          ```
        displayName: F5 Ipam controller
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipam
        - ip
        - ipaddress
        - controller
        - ipv4
        - ipv6
        - F5
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/userguide/ipam/
        - name: Github Repo
          url: https://github.com/F5Networks/f5-ipam-controller
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
        - registry.connect.redhat.com/f5networks/f5-ipam-controller@sha256:09a358db300d82e00fd3b4e255917df82c02e1bc58425f8cff21cc2c77eba1b6
        version: 0.0.8
      entries:
      - name: f5-ipam-controller-operator.v0.0.8
        version: 0.0.8
      name: alpha
    - currentCSV: f5-ipam-controller-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fic.f5.com/v1",
                "kind": "F5IpamCtlr",
                "metadata": {
                  "name": "f5ipamctlr-sample"
                },
                "spec": {
                  "args": {
                    "infoblox_grid_host": "IP_ADDRESS_OF_SERVER",
                    "infoblox_labels": "CIDR_Labels",
                    "infoblox_netview": "NetView",
                    "infoblox_password": "PASSWORD",
                    "infoblox_username": "USERNAME",
                    "infoblox_wapi_version": "VERSION",
                    "insecure": true,
                    "ip_range": "IP_RANGE_LABELS",
                    "log_level": "DEBUG_OR_INFO",
                    "orchestration": "kubernetes",
                    "provider": "INFOBLOX_OR_F5-IP-PROVIDER"
                  },
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "f5-ipam-controller",
                    "user": "f5networks",
                    "version": "latest"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "serviceAccount": {
                    "create": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
          createdAt: "2025-08-08T03:48:02Z"
          description: Operator to install F5 IPAM Controller.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/f5-ipam-controller
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This Operator installs F5 IPAM controller(FIC)
            kind: F5IpamCtlr
            name: f5ipamctlrs.fic.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 IPAM Controller in your OpenShift Cluster. This enables to configure and deploy FIC using Helm Charts.
          ## F5 IPAM Controller
          The F5 IPAM Controller is a Docker container that runs in an orchestration environment and interfaces with an IPAM system. It allocates IP addresses from an IPAM system’s address pool for hostnames in an orchestration environment. The F5 IPAM Controller watches orchestration-specific resources and consumes the hostnames within each resource.
          ## Documentation
          Refer to F5 IPAM Controller documentation
          - https://github.com/F5Networks/f5-ipam-controller/blob/main/README.md
          ## Prerequisites
          ### Configure Infoblox with network and netview [Provider: Infoblox]
          Refer Infoblox documentation
          ### Create persistent volume and persistent volume claim [Provider: f5-ip-provider]
          Create persistent volume and persistent volume claim
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/f5-ipam-controller/main/docs/config_examples/f5-ip-provider/localstorage-pv-pvc-example.yaml
          ```
        displayName: F5 Ipam controller
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipam
        - ip
        - ipaddress
        - controller
        - ipv4
        - ipv6
        - F5
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/userguide/ipam/
        - name: Github Repo
          url: https://github.com/F5Networks/f5-ipam-controller
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
        - registry.connect.redhat.com/f5networks/f5-ipam-controller@sha256:09a358db300d82e00fd3b4e255917df82c02e1bc58425f8cff21cc2c77eba1b6
        version: 0.0.8
      entries:
      - name: f5-ipam-controller-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: f5-ipam-controller-operator
    provider:
      name: F5 Networks Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: smart-gateway-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smart-gateway-operator.v5.0.1757650606
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "smartgateway.infra.watch/v2",
                "kind": "SmartGateway",
                "metadata": {
                  "name": "demo"
                },
                "spec": {
                  "applications": [
                    {
                      "config": "host: localhost\nport: 8081\nwithtimestamp: false\n",
                      "name": "prometheus"
                    }
                  ],
                  "blockEventBus": false,
                  "bridge": {
                    "amqpBlock": false,
                    "amqpUrl": "amqp://amq-interconnect:5672/collectd",
                    "enabled": true,
                    "ringBufferCount": 15000,
                    "ringBufferSize": 2048,
                    "socketBlock": true,
                    "statsPeriod": 60,
                    "stopCount": 0,
                    "unixSocketPath": "/tmp",
                    "verbose": true
                  },
                  "handleErrors": false,
                  "logLevel": "info",
                  "pluginDir": "/usr/lib64/sg-core",
                  "services": [
                    {
                      "name": "prometheus",
                      "ports": [
                        {
                          "name": "prom-http",
                          "port": 8081,
                          "protocol": "TCP",
                          "targetPort": 8081
                        }
                      ]
                    }
                  ],
                  "size": 1,
                  "transports": [
                    {
                      "config": "path: /tmp/smartgateway\n",
                      "handlers": [
                        {
                          "config": "",
                          "name": "collectd-metrics"
                        }
                      ],
                      "name": "socket"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/stf/smart-gateway-rhel9-operator@sha256:3363abcb91319f8642212b9b80ca49bb9f90cae69255b1f75f03929af3d3a8c3
          createdAt: "2025-09-12T04:17:59Z"
          description: Operator for managing the Smart Gateway Custom Resources, resulting
            in deployments of the Smart Gateway.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenStack Platform", "Cloud
            Infrastructure", "Cloud Suite"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/infrawatch/smart-gateway-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of the Smart Gateway
            displayName: SmartGateway
            kind: SmartGateway
            name: smartgateways.smartgateway.infra.watch
            version: v2
        description: |-
          # Smart Gateway for Service Telemetry Framework

          The Smart Gateway is middleware for OpenShift that takes metrics and events
          data streams from an AMQP 1.x message bus, and provides a method to having
          that data stream stored within supported data stores such as Prometheus and
          ElasticSearch.

          ## Core capabilities

          The Smart Gateway provides two modes:

          * metrics: provides an HTTP scrape endpoint for Prometheus
          * events: writes events directly to an ElasticSearch endpoint

          ## Operator features

          * **Two Service Types** - Supports the `metrics` and `events` service types which defines the type of Smart Gateway to be deployed

          * **Configuration** - Configuration of the Smart Gateway to allow for it to be connected to an AMQP 1.x bus and pointed at the appropriate storage mechanism based on service type.

          ## Before getting started

          1. Ensure an AMQP 1.x bus has been setup (such as AMQ Interconnect)
          2. If the Smart Gateway is running in events mode, be sure the ElasticSearch Operator and ElasticSearch instance have been pre-deployed.
        displayName: Smart Gateway Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - servicetelemetry
        - monitoring
        - telemetry
        - notifications
        - smartgateway
        links:
        - name: Source Code
          url: https://github.com/infrawatch/smart-gateway-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/stf/smart-gateway-rhel9-operator@sha256:3363abcb91319f8642212b9b80ca49bb9f90cae69255b1f75f03929af3d3a8c3
        - registry.redhat.io/stf/sg-core-rhel9@sha256:0e12903da0b15c64421102b5b914ca4c82d107295485c9b76ed19e91a45a4420
        - registry.redhat.io/stf/sg-bridge-rhel9@sha256:7fae342cf6cecc676eaa0a8fb64ee380ad101d667d6e8f0d89c943938bb6baf4
        - registry.redhat.io/openshift4/ose-oauth-proxy:latest@sha256:dc24aa382e02c4ac058d4b824a21f0e90a6259d6d817d0341b2a41809b5daa78
        version: 5.0.1757650606
      entries:
      - name: smart-gateway-operator.v5.0.1757650606
        version: 5.0.1757650606
      - name: smart-gateway-operator.v5.0.1743671602
        version: 5.0.1743671602
      - name: smart-gateway-operator.v5.0.1740051852
        version: 5.0.1740051852
      - name: smart-gateway-operator.v5.0.1731336910
        version: 5.0.1731336910
      - name: smart-gateway-operator.v5.0.1725953602
        version: 5.0.1725953602
      name: stable-1.5
    defaultChannel: stable-1.5
    packageName: smart-gateway-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-pipelines-operator-rh
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-pipelines-operator-rh.v1.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    },
                    {
                      "name": "legacyPipelineRbac",
                      "value": "true"
                    },
                    {
                      "name": "createCABundleConfigMaps",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.19.0 <1.20.0'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.3.2
          - Tekton Triggers: v0.33.0
          - Pipelines as Code: v0.37.0
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.22.3
          - Tekton Results (tech-preview): v0.16.0
          - Manual Approval Gate (tech-preview): v0.7.0
          - Tekton Pruner (tech-preview): v0.2.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.20/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:5021158a4752a26f8b33adea7da66695b22d2ad155d346d35dde5f971bc1a55b
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:5683321a5376e63c3ff3932270a1850d5e261f76212020ce867334166f90d2f7
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:a712bb411d2433b01793dff25785c1e265fa865a52f7c8732c9fb4418ab61757
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e7998e113f75e8f29602bae210c8b66ed06634b699f1d99c5875ef879220d733
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:96db721939b22572d7d8d8b95dcb5c92fdedb98e9cd0aabb7d17b703a915b07f
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:115eea1a5ca87f3b6b72db1df05b7386bf73dc70c9c2e4dc5e1a66700d960c37
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:980ed7e81ea55522e1681dc1e9b170b68d16f07061dd1f1ffd142367d59552f1
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:1b6aecfce45ed3163227b845b212e83385cdbe999edfc2e9387a0c76f3f8d8e9
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:161df929e2bd4c5acff594be31d5458365de5ae0d2189bb2f647c3fc6657fa28
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:bc10e7d6122ede5cf4f2882c200447683255465a1080e826761e90bfb983506e
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:df15eb1f1f5f237bfe481c16ab3d40bf7fb4123f3c930bcfb121e2cc8dc42882
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:d56d6b46331b4569077ab1e7a7b609ba01f0a1dd62b695d62f8aa0f32ea1b621
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:c0f438e6095aa81f3f7d820614e0414cc43374014c1505d7641bc1dba69a3903
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:818bfa4a3a874974b5f7f8bcf2faeb6c05e1795381ee1ceacf89f17b1ec0ef7f
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:9415e889ef98df6d2b2dca1e5d798fb765eeea0d9828e48424374d5267a6b0d1
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:441e1c7124c1674e3d4c38c3e82a4fe5568e76d13c89c04bf009d6b97435de09
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:7eafffa6e4e04f10c1d4e84e02d51c00eb8649c60f8f9a13da14492f24634891
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:bc403f93ef6c5526fa30152eab384c4a25bb0bc6ab4158a1e24f637b87a57259
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:2d496752b19f765f1ea36ec94b873ebd8fa604764a56aa67d3ed94b33e5fe75c
        - registry.redhat.io/rhel9/buildah@sha256:18453ab0a62154283aaf4c45efd7543d1c8ecca1aebae46dedc7ddc0f410b232
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:036250bcdc6fb95273bf07948dfdd4dce765ab84c803ceca0d0b4abf6ca19dec
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:8f0494341244a75546a4b80b065702be04e6d200421aa4e7ec90385c9dabc6ab
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:c4d05e546c466c57379fff502816012f615be073eb8cdce54677b4ab206a3f5c
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:bbd1918660be7b3c78777dcccc87d3b26bf5adfc077844ceffbbdd53696dc0a3
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:e3e64abef53151f6b15f15bcbff85668167a571e98badbd8a2ca252ff3a5b8dc
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:dd941acd357ae2c0fd7e15fb8d32d7ee6b2847435101dd1e1f107db48af18bf3
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:823e018e0329d764ac694f177ce88af9074f364c29a1c17929f9dd0b596d03dd
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:a7163a9e1fe891de021b1003558e6e64fe17ed17ab0820ba16bba09aa112401e
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:e63298d07ccec5804dfab4e42bcc8163c29cb3ac371c91248777423a4edd31ce
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:3cf990fb2fc422f3527a7c8b43d06ccd4d32bb5487ec8cb2d4a97f0f5d13f170
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:c3013c347bca8a520443a44ac2817fbba5744a9e91069ac5f09e91447ebf6b28
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:901b5108ceb669c985e75f5fb033c496e9f028863df99ab32651d1b08433a352
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:14dab8d87323fc7d112213f7f4db3ee7cf3841747de669ce710ae6a794053457
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:458f865f1b246c5f0682f5956795771a0515b89edfcffbbc0da1a3d3f532645e
        - registry.redhat.io/ubi9/ubi-minimal@sha256:2f06ae0e6d3d9c4f610d32c480338eef474867f435d8d28625f2985e8acde6e8
        - registry.redhat.io/ubi9/openjdk-17@sha256:555c8bcd2596637befd1a6da4aa949c5a2640cf66e09c2dfb932afe37e39095e
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:c31af237afb84214bf2a4c06a7636ca071fe97203b41da90d7878109ccd64784
        - registry.redhat.io/openshift-pipelines/pipelines-opc-rhel9@sha256:c1bee4a2bb495de6d4da962b6e7b2447caa51cc36c3005e4083bd31420ef5102
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:ec74a562c18fefe36ec7ae3ba083e4eb76c56b562f916578900f8767ae98e6c2
        - registry.redhat.io/rhel9/skopeo@sha256:a65a413f8a2864389a09dc750690d97afbcdc5c70821e0f85e99e8adba7954e0
        version: 1.20.0
      entries:
      - name: openshift-pipelines-operator-rh.v1.20.0
        version: 1.20.0
      - name: openshift-pipelines-operator-rh.v1.19.3
        version: 1.19.3
      - name: openshift-pipelines-operator-rh.v1.19.2
        version: 1.19.2
      - name: openshift-pipelines-operator-rh.v1.19.1
        version: 1.19.1
      - name: openshift-pipelines-operator-rh.v1.19.0
        version: 1.19.0
      - name: openshift-pipelines-operator-rh.v1.18.1
        version: 1.18.1
      - name: openshift-pipelines-operator-rh.v1.18.0
        version: 1.18.0
      - name: openshift-pipelines-operator-rh.v1.17.1
        version: 1.17.1
      - name: openshift-pipelines-operator-rh.v1.17.0
        version: 1.17.0
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      - name: openshift-pipelines-operator-rh.v1.14.1
        version: 1.14.1
      - name: openshift-pipelines-operator-rh.v1.13.0
        version: 1.13.0
      - name: openshift-pipelines-operator-rh.v1.12.2
        version: 1.12.2
      - name: openshift-pipelines-operator-rh.v1.12.1
        version: 1.12.1
      - name: openshift-pipelines-operator-rh.v1.12.0
        version: 1.12.0
      - name: openshift-pipelines-operator-rh.v1.11.1
        version: 1.11.1
      name: latest
    - currentCSV: openshift-pipelines-operator-rh.v1.14.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "db_user": "result",
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.14.6'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.56.4
          - Tekton Triggers: v0.26.2
          - Pipelines as Code: v0.24.6
          - Tekton Chains: v0.20.1
          - Tekton Hub (tech-preview): v1.16.0
          - Tekton Results (tech-preview): v0.9.2

          ## Getting Started
          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/container-platform/OPENSHIFT_DOCS_VERSION/cicd/pipelines/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)
          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi8/openjdk-17@sha256:a8165bc2cd5051a96d6937e25ed03155bbd4b731da6e58cebfe2ea83209c16d8
        - registry.redhat.io/ubi8/ubi-minimal@sha256:6825946f10e8c81586fa7815c4f06016af05cfc231d921ec9699402a6655d201
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:059c08316152c209de13017e322d6fe5eacfde81c3b0039e55b5de8f2f27183a
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:d01ba7ead670d5ab9c1355b90010657988e46b4d1aa0bb68b194eaf4001b2d7a
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:6327385e3d3c9df1ea50e803955d980b4cd35fb15593fcfec0d477d91a45021f
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel8@sha256:6e7c7a1ca458e08ad85e1146815ec00a30016f709b45e28b464043e875c2a3e8
        - registry.redhat.io/rhel8/skopeo@sha256:e297b84ba6f1d5eb2c37e2c3b31c9a9a4c39f9590e7d7c32388165245bde4835
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:ae0260abd2725c2f4bb25c0c07d86290978892a49050e0fd23f466401aba668c
        - registry.redhat.io/rhel8/buildah@sha256:b48f410efa0ff8ab0db6ead420a5d8d866d64af846fece5efb185230d7ecf591
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:2261f4f657f70759f4d0c898e44ffafceb8dbe2144063d50227140463bdbf2c5
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:7a63e7a762b9c7b15057ed9d5bcbea08acb4c8933a300c6ec336824d49a6e937
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:8d4450d68ac03d8275e337d2c55692ef35e954400d0c9f041a943d163a8b7680
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:bf6cf2e87fb19f7aa9a490b83c16af69834c0721220a643710a1b077959e91ca
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:aa6eb4c7ade9d6add1e4ab94197c2ce72a253fea0d7461698ca0e37854f1dcbe
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:71d91c7b80b6d196025f68749bd8518b8b0affd8d5064a7c5c70d125656fbc36
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:db3ff043410d1dd0621fef4483761727f7241d788b136e446fb861953350a2cb
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:5e7f5a2813a591e7b13fea9fc4e2f599b5ec855a015cd5a4c287c4e178358756
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:9c29f6eff1f7019993446d85d226f04c5614a18351dfb4042a8bcb62276810de
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:d9972bc64dfa7e260bd9b949bbb85ef5ae532e7c6df8d874d3d3963af1097726
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:bf0c38beefa24ca3d04ebfa6140a1cc5a9a56fba81ae05799279a559093b9033
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel8@sha256:b3bedee6084e9d9da549d3e637d2bf176aa365b5818e6b6377525577edd710dc
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:82c8a91d521d7e6d81fb462fa2b804f420f52ee062df5cda39545165a2f19e0c
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:ce0ea093c522af329acdc85148c2473ea45cd44270c15f8dd32965b0d56f304f
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:abf4442b6acc714c2392b86af046d5f32d247ee0c3763bb18c57d0473edfcf02
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:e2bd929246d62038a24a0685cb3e4873c131c9b7713d6a5268ca09a8614538be
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:83913744a1bb7ff23091cfa541b807cfb82b38dde0c37e0f1dd9c17cade99775
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:63fc1351191a54d12d4e463a1bb4cc732620cba3000af86ba7f7a579514f5fe4
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:e23c901868229e4f6a4016aa7798f1aba879ff35ffe594f5b0e4d39f1bff9e52
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:f68dc18a4dc6a1886507097551f4a6fce39062d03c6eb86c067d317812b0d34a
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:f9fed20cdfebbb226e035cc62f037174f5ecdb3139a5edc7e39f019f6d2b1e98
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:ddf0cb64ea57de872515e19de63b2dec5f56738c9d7eaa283a4a489545ef4902
        - registry.redhat.io/rhel8/postgresql-13@sha256:3903d562425dd564986123280dc076548fb3d24c74138dcd857652f53052e0a7
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:2e75096ffe20d1261e8a645fefd32e37295d727aa6a280f10041a44443917469
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:57fb86ba112d79a228cbacc96cd0525c123464bcac956354722e3f2aeff42508
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:ffa8c6d8fb5bacb59da27f4ebde87385359b3f97a6a023b4a4488ced2d46c574
        version: 1.14.6
      entries:
      - name: openshift-pipelines-operator-rh.v1.14.6
        version: 1.14.6
      - name: openshift-pipelines-operator-rh.v1.14.5
        version: 1.14.5
      - name: openshift-pipelines-operator-rh.v1.14.4
        version: 1.14.4
      - name: openshift-pipelines-operator-rh.v1.14.3
        version: 1.14.3
      - name: openshift-pipelines-operator-rh.v1.14.2
        version: 1.14.2
      - name: openshift-pipelines-operator-rh.v1.14.1
        version: 1.14.1
      - name: openshift-pipelines-operator-rh.v1.13.0
        version: 1.13.0
      - name: openshift-pipelines-operator-rh.v1.12.2
        version: 1.12.2
      - name: openshift-pipelines-operator-rh.v1.12.1
        version: 1.12.1
      - name: openshift-pipelines-operator-rh.v1.12.0
        version: 1.12.0
      - name: openshift-pipelines-operator-rh.v1.11.1
        version: 1.11.1
      name: pipelines-1.14
    - currentCSV: openshift-pipelines-operator-rh.v1.15.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "db_user": "result",
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.15.3'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.59.6
          - Tekton Triggers: v0.27.0
          - Pipelines as Code: v0.27.2
          - Tekton Chains: v0.20.1
          - Tekton Hub (tech-preview): v1.17.1
          - Tekton Results (tech-preview): v0.10.0
          - Manual Approval Gate (tech-preview): v0.2.2

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.15/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:322ef2385620661a7914f47c73a6d145ec0c9ce1ccde30645bb66148545ddc96
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:49c346bf7fc69a9796690236dc8597e35a874343b02fb74c5d2f1c05d356e41c
        - registry.redhat.io/rhel8/skopeo@sha256:29465586e92ed04d2878dcdeda2d508c7e8e7539762e10296b52d631929960aa
        - registry.redhat.io/rhel8/buildah@sha256:aac6629389db17e99894c5bee0da01d4c8065d11d8c6f6e1602f9484290baa70
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:bf94c350923da706c6e0a427e2d91af3e13e751a12388a4ed75d1c34dd7b3177
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:ac1a6e057a5599fda7e048ff77030e98c92b5d51cac2472c1ef3910964fc17e6
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:61c22119f37e620a3807056cde35bc07c90867765b149294056326b1d11494f1
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:abac279a0a5ecde57952b6bef634f8ba7527cd8c8956a2ea30e9449174ad52bd
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:166789a990df9f6ab04fd3df0a8cd301bc19b9bedde94215cb27306060eb1e2c
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:83eb08bb9579cc19132af282c455bca32a0fd4f5436a35e7ad23b5efa2faa3cf
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:479e3eed93e0aa13e3489514fe00375a519eb45969d099c490b4acd056ea5adf
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:d1f86458079449c86407ac716c3cbf608cad06de29edeecbde6586f76628daa9
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:e749db81c9d7e11e2802833a2a02e93afc357c83d1a2a827e537262df281a4f7
        - registry.redhat.io/ubi8/ubi-minimal@sha256:f729a7f5685ea823e87ffd68aff988f2b8ff8d52126ade4e6de7c68088f28ebd
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:7c81ae208a16e17d0102074867add37dfb1207ee95cbbc6c732d88e61bbc59fb
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:7152c3c174d25f2a5239456ba677ac5315bfb82bc3b3ac5ee0eb8f9f74fb6632
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:aca0024cac7a8c41b1ed151eb1e1188f09bc636866f4e03f1b973f90890b37cf
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:bf6cf2e87fb19f7aa9a490b83c16af69834c0721220a643710a1b077959e91ca
        - registry.redhat.io/rhel8/postgresql-13@sha256:a92a579f1aef66ac188d24fd489c456a1a3e311d95dcce652da6b81e28fbf725
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:7f00ff1b407d9ea107d3d969988a6279b28bde86943f490a007fa922dd40a1c3
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:eed464a3b4e23f0b2623aa1b923c10232d235f4272f562955926c1748e295d2b
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:7958925cecce1b4b02ac4649e3bfd7fa3b5361c16ab3593700b8da93fdf1c79a
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:312d1f90765d05a86845856c677e1ece4c57e81a9bfe3fd1e76868d624e3d72d
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:1300d171dbf09c63071ba5586e223491a13acfc6dbab27eff04648aeb788d455
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:538c63c8e7af3cba91526f2f9ac99fb8f6b0f4781db31771c959612c209dd090
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:601fcac1abc6c8efa1248c5ca229213d6660708470a1c416d41c6aff0ba4691a
        - registry.redhat.io/ubi8/openjdk-17@sha256:e8cc2e476282b75d89c73057bfa713db22d72bdb2808d62d981a84c33beb2575
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:e7b000a2735bbf98ee79bd093f7b703ed09d7c1e11ed2a5b22c4b8ac8d440386
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:827313c2691fb8833e42418b65b4a3360a11cabc2c92a7d8e7b20183f06cbf68
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:f02af8475c33f6f38a2f90ff1be1fb1b2094860c4d5e300aefdd7eb84b4a529a
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:acf8017029f4b2e7a1c3da9bc62c65239d313d24e3012712356a191bea9a2e6f
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:75d6228eb656b8778fb3bb8dc68c686ac1c9b787e5343ae3c2467f8e724af83e
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:1db14f23077785108660dde38c98e2e30c8301f8e2e1c594dd4636e0ba29f84f
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:70bb0ebcfe5da45e5c239aa011dc5f1ef48844116d7346bb05433b503107268a
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:f04ea12675e0eb2354635cd2f9e99d6e59619d1c611e0ea95c3ce8222d615322
        version: 1.15.3
      entries:
      - name: openshift-pipelines-operator-rh.v1.15.3
        version: 1.15.3
      - name: openshift-pipelines-operator-rh.v1.15.2
        version: 1.15.2
      - name: openshift-pipelines-operator-rh.v1.15.1
        version: 1.15.1
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      - name: openshift-pipelines-operator-rh.v1.14.1
        version: 1.14.1
      - name: openshift-pipelines-operator-rh.v1.13.0
        version: 1.13.0
      - name: openshift-pipelines-operator-rh.v1.12.2
        version: 1.12.2
      - name: openshift-pipelines-operator-rh.v1.12.1
        version: 1.12.1
      - name: openshift-pipelines-operator-rh.v1.12.0
        version: 1.12.0
      - name: openshift-pipelines-operator-rh.v1.11.1
        version: 1.11.1
      name: pipelines-1.15
    - currentCSV: openshift-pipelines-operator-rh.v1.16.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.16.4'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.62.9
          - Tekton Triggers: v0.29.2
          - Pipelines as Code: v0.28.2
          - Tekton Chains: v0.22.3
          - Tekton Hub (tech-preview): v1.18.2
          - Tekton Results (tech-preview): v0.12.3
          - Manual Approval Gate (tech-preview): v0.3.1

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.16/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:81a8b903d0fc4b81e6a23fac862d18da4cc18b14cca8213665df37dfaabbc2e9
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:5672f3a375c6a274200b423eb2643c818aba75488dec0b210fec2532814ba51e
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:47f094ccd12db7a43c8a91c1a3d2a190bc2b83f1745e7eb8ce097d0c8e1e2c88
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:80d46bc5ce8c931af22a97ccbce05b59740e313167c88ad910febcc2f2dc0320
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:90e2a7ea498da756e09c7e4d19c7fb91d429cc28c57f97c6958958e3466cb58c
        - registry.redhat.io/rhel8/skopeo@sha256:1dcb4189c92da21ee58b3ca5433615e0208194c8412f7f1fb45fa491179aaf9b
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:d2bb3f9f79ef719d8fdafef3a639e3eb8bc199ee56fcfb33e85975d536d2a6ac
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:7edbd11e5a449408d673cae1fc2b9b5bab492ecff2ac6e640c8cd69a9b90ff08
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:a5a7b2be50f211d4d9d555b2a2e2302096c139ff5011d984cc386f1530007f16
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:f9122d71c3862cd2afafc8eb7f1a39c83eee306e5baf38514e6e39215c70a72b
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel8@sha256:8cb87920e6fc8185ead49aafb7799fead6a4a510eafdae0ee29e85248dae3945
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:aacdabe243dc573f3161a3d04fbf014b77d9962207305d70fcafce475f7eea93
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:92042dcacd4ab0f2a78aab03fd91a9a39d7d7366b5398fd797ecb145acf5fce4
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:2fea7d8e29d70c5893ddb6caf2bf1d4144787f5f0de8739351d2097ed09ee751
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/rhel8/buildah@sha256:5c7cd7c9a3d49e8905fc98693f6da605aeafae36bde5622dc78e12f31db3cd59
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:474658c262b539dd1c83ac22bb090b3fdac5de9cffc06587d36dbc6fa4ce208c
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:e1fe4a86211cf22e5ad2b23c989d24b69f597db4c3604e84777dbd0c585ec186
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:6acc31f179b08658976cddcb9ef42f377edb29933d1d83a2ecebfd455bb68571
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:f3e6df07fd86acb91037d55a1970e5f8c6ab12a7517752692c224c7687022812
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:8129e62906c68f04075a44c939375a7bafdeaeac673a20a4fa74ffd6285187ac
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:7d92dca9da27c066cb530172e1811493d8c916e67d8068f5ddc471fa3f3f7856
        - registry.redhat.io/ubi8/ubi-minimal@sha256:a47c89f02b39a98290f88204ed3d162845db0a0c464b319c2596cfd1e94b444e
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:b94bd566ae3d9ce5f47273d867164486826b19fca0ce11d1a7d18312413e87ba
        - registry.redhat.io/ubi8/openjdk-17@sha256:632e78989471859ca4ed1148b951a911f7e3b6e6516482be20967c4171612c55
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel8@sha256:46635d636ea2e9a6ba0760bf8c528028700ee40b0ba8466797034371185b16d0
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:856229e9e37683c926e3449658d76f93d910384eff62c2ca192768a653b71533
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:c6d654c71aba35c3c7f5c1a2ca5b8401e95adf6f649c599a9aff25177c2b3323
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:549d74a18ec52db91fa949cac9b8e3d943d478d5dec9850d2ba0cb1bb526f578
        - registry.redhat.io/rhel8/postgresql-13@sha256:a308263f5d1071be8ac5c13fd84b40609988ca30ed56c22a8bc6769e107ce09b
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:b41b888fc1a3b7215e6b71ee97bf64a1816914a8f33e96ad3930a4e2d5b5ffa7
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:4e0e4c358faa9cafc3964d6d40ee4aa73f8dfe8b6338dfb8206613d7c5241842
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:dcab4a3d519975340ade59a0ca0222104e68acd052a61f2aeb8c736be4b672cb
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:e84c2d80e92815c6bf384e1504a194dec9b796a3aceac2eed555fff241249571
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:d36dbb60c7b5caca7f3be08705f8ab91d7f01abaf8bfee4aee73bdbbb3e7e71a
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:7da7f929d2711fc55dda01fe51a474bfa7adf5aaf78be6d65a6d65850958d6e7
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:19b3dd69c1d7e17890b7163aef322c1efbc76fed7f2b1faf26aaa05fd8161184
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:aeeed543412cde0bd3a9baec965cb99d80f2b6acb77ae6356ed51b4b1262d006
        version: 1.16.4
      entries:
      - name: openshift-pipelines-operator-rh.v1.16.4
        version: 1.16.4
      - name: openshift-pipelines-operator-rh.v1.16.3
        version: 1.16.3
      - name: openshift-pipelines-operator-rh.v1.16.2
        version: 1.16.2
      - name: openshift-pipelines-operator-rh.v1.16.1
        version: 1.16.1
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      - name: openshift-pipelines-operator-rh.v1.14.1
        version: 1.14.1
      - name: openshift-pipelines-operator-rh.v1.13.0
        version: 1.13.0
      - name: openshift-pipelines-operator-rh.v1.12.2
        version: 1.12.2
      - name: openshift-pipelines-operator-rh.v1.12.1
        version: 1.12.1
      - name: openshift-pipelines-operator-rh.v1.12.0
        version: 1.12.0
      - name: openshift-pipelines-operator-rh.v1.11.1
        version: 1.11.1
      name: pipelines-1.16
    - currentCSV: openshift-pipelines-operator-rh.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.16.0 <1.17.2'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.65.7
          - Tekton Triggers: v0.30.1
          - Pipelines as Code: v0.29.1
          - Tekton Chains: v0.23.1
          - Tekton Hub (tech-preview): v1.19.2
          - Tekton Results (tech-preview): v0.13.4
          - Manual Approval Gate (tech-preview): v0.4.1

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.17/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:a14270967460883f283ec1ce17ff8b82db1e0777dda4f6d690f15ac38dcbad85
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:650391ffd3e4c7b173bf9d718405b2375c6ff2027abc0b7ce4f67e79e7ef4b81
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:37109f6de75dda750496f6668164d0dceaf9db5278e8c034708436f39de3276e
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:a392d2c85baedd7428c579e91d9f8acf8232399b90523b1093dc3447aaf56a16
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:8e52b4b31b0d9ec772abe6e6ef9c6698e89893395f364968216bebc3367da4b2
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:4b87782b35e7284a365c49f1c094aae98531ac630db1958294009c041a2d8440
        - registry.redhat.io/ubi8/openjdk-17@sha256:e5fc9e34382da41b10d4ea566639d31707493abff9e8de81f5accfd66876409e
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:4224040b00cfd76e85f80aecc5095098564cc62e9c477fe2f7a428b30df7d4d9
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:41ec6db921812fd1e80b05d51c224ddbce63bf0c57f02ccf07461d2202e97b90
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:4fabae1312c1aaf8a57bd2de63bd040956faa0c728453f2a4b4002705fba0f0c
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:7abc88bea01dd6513f6e48334f0b98501f909a1dc1ab9e83efeb3c1a834f8f92
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:eab798472b76bd51235198fd3b95189740f8d650ebc6bb01fe653c025c1defaa
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:61c9e8b1660d8935ee316ecc113e8141daf05149d172685cabd861ae8aeed3eb
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:4240501c076605a01d969328bbb8ad885c598b44ec1dc1c066312491a0eb6715
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:7c710882752f12be3c22bedc8d32096c111e42c5b9b392163f61cf8d72be3ec0
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:66309ae437c4647f48efe3d6f2b696df83859323700427d0025b94795751a859
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:41e2f2c7397a8b9e093e18977666e93943734f51780c769c99958ae1d20af53e
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:4a96f1bb7e3a5c7ca62dff2199fa15ef964b129e4e36e9619837c1221538f496
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:3696cf724ce7865626ae408d82fd87727c741e4d3eb3475c271ec6d56f0882a4
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:44da76d75f567c2ba623bae646e1d552d2c0fded1accdde3aab4f4bcdf3adc28
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:26f7b695fc1430359ab2bfd2915cf0b7e88b06d8695e26761bdc2ce8935f0fb3
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:0b6ee6cd98cb7e835cfc9544996c6f85288d7252071c100bed853eb593d6befb
        - registry.redhat.io/rhel8/buildah@sha256:6d2dcb651ba680cf4ec74331f8349dec43d071d420625a1703370acc8d984e9e
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:d851a2dd8ec95bfd7ac6ec5515c8cdadfcbfc6e6ec84b40630b4fe595f978fa5
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:429bc3d6044ca7dab780ff4b5eb91d42d38a29eb7ef428a25c53b7dda49871bf
        - registry.redhat.io/rhel8/postgresql-13@sha256:757d8dd6f243717d820846da7144a49913f28f223b74ec305b2215f54720f8a0
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel8@sha256:b5cceead6b8e9c99655c9ba810679b9d6fb978e5ff18d020a46d475c4c26fa29
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:c92ebe828b6ced6c93a3df5f4ccdfba022471729ad08b184de2f7410ecf592cc
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:df11cdbab2a1c6b9b800ef124c73f2fb5e3981402b21bc822ccc13e95836197a
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:db27cb4f0c350d99546b0f1bdf02d1ba271a8053a4552d88ff302adca60a3e29
        - registry.redhat.io/rhel8/skopeo@sha256:7fd65200b4847fc84fba0466519b87f915e28c0e9bc90d3022beecfebffd4af8
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:99f5383d8c6bc046a0bf48fd6c2064a5d055645085248327305f3ee4ae71979c
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:e673c20fff2fc684f5098282dd6ac6cb265045945249e995b81289532ce8c856
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:9668d852123b160e49511242371edc6e941a8e8b8c2ffa60080d91cc6b97d3d7
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:84eed8ba1189448e02ffd284c7d913e7b3cb7efe7a8e5b0edf890f7eb17fc439
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:922b9e86a674a6f3eb82f510569ab9cde97a1ad833c263335cd4164bf4164b66
        - registry.redhat.io/ubi8/ubi-minimal@sha256:c12e67af6a7e15113d76bc72f10bef2045c026c71ec8b7124c8a075458188a83
        version: 1.17.2
      entries:
      - name: openshift-pipelines-operator-rh.v1.17.2
        version: 1.17.2
      - name: openshift-pipelines-operator-rh.v1.17.1
        version: 1.17.1
      - name: openshift-pipelines-operator-rh.v1.17.0
        version: 1.17.0
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      - name: openshift-pipelines-operator-rh.v1.14.1
        version: 1.14.1
      - name: openshift-pipelines-operator-rh.v1.13.0
        version: 1.13.0
      - name: openshift-pipelines-operator-rh.v1.12.2
        version: 1.12.2
      - name: openshift-pipelines-operator-rh.v1.12.1
        version: 1.12.1
      - name: openshift-pipelines-operator-rh.v1.12.0
        version: 1.12.0
      - name: openshift-pipelines-operator-rh.v1.11.1
        version: 1.11.1
      name: pipelines-1.17
    - currentCSV: openshift-pipelines-operator-rh.v1.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.17.0 <1.18.1'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.68.1
          - Tekton Triggers: v0.31.0
          - Pipelines as Code: v0.33.2
          - Tekton Chains: v0.24.0
          - Tekton Hub (tech-preview): v1.20.1
          - Tekton Results: v0.14.0
          - Manual Approval Gate (tech-preview): v0.5.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.18/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi9/openjdk-17@sha256:5d619e202695acf76a5f32c07ace6709f5ee5d3999bf977e1bf95ad34197f44a
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:86ea94b12e8923035e3e4f39b6a53b3ab32c5f77d7eea170aa9b34aecc6f8577
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:0a60ce46c05256c74f770901e0bb79ad89e88ba882b643c1ff7b2a1738463acc
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:3964c3248dfb1006ea9a46295baf7f8cfa71952d6b9863b52f87e8f3d7863b1e
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:36a4c6a3273f2371ab5e679f1955221c286ea753e29e6ab8da7c983681a68f02
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:b550a8bc2a87d07a5b22a3ec3fb83efd8eb26ca6e80d404f2e453cc9b8434da7
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:cd28c2c90dd340324b0d8709640cd7636f28286114b0023649b6e21be13e03c2
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:fa5c456fc91319c55fac9a011dc7bb32ef85db293c73d74436e903b76ceb4349
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:409c14649d8e7007d4b72b6feffb71dabd26322bf2a7c77f5307f2546fdfe81b
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:96f5e1922fb9d6dc17e271576dda9bd032bd6dcc04e7496fb15a2f6b5871cd67
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:16d799503674641d4690822edc9c399c61429ccd31621d295eee4a4d1ef1c076
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:cbd89c531953a43b8055bda72f7696be80f57fd1d782567a91d82e0fbcb3d26d
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:8fec8e7a8b3e5182c67104de4e3abe32593b682c4a8986d1e9ab835dafb95ff4
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:3696cf724ce7865626ae408d82fd87727c741e4d3eb3475c271ec6d56f0882a4
        - registry.redhat.io/rhel9/skopeo@sha256:d077e35100d6d7b3795f8e3594293dbf8b65c6bc18b909aa555fe56f048b356a
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:1b57b64657eca0002145100bae262a19f02eb40068e07994df467e28a1618f90
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:20e19c02efdecf94a7b16483047e5db81c50dc2a0aa59d4895067e6c7769efcc
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e2b4c3b505c4b5aa36f8a5e661389b0de4e799c74c2c7f16a065e5522b438e89
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:80d441ec5d07c5059277a3a9173a35dfc0b8372d6e4d7823514131a94bae3fd7
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:d72842a7c6a1e50a68576c93f60047afec801cf83db64d7d3e34b25bf3c753ec
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:e6a7738cdb0e8aa9e1af381273b1c18ec1414b08a66fc851f6e11e5aa7a6706f
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:fcd5fd10de7334bd59a8da4997f91008ced95c6cde40d56d50f3b3e32c0f19e1
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:c20ab6fa719506bca942669ea84b113335b05df3ed275257b88be234cade7c8d
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:2ba6782af0d7e5c4d623de31814f43f4f7c3c38d09b57705021957d334067f87
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:d055d2b35a663aef3e1aafdbed0b12957867c0670c946ebae66e9c44a768bda2
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:72901cf4e091fe5f6a577144fdffaea8570712e9887a8bd715d5ae5bc346ff1d
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:b9675b22bc8759903472c1eb0e03f90c75460f02799d894294c161641cad9414
        - registry.redhat.io/rhel9/postgresql-13@sha256:ecd5a5583db922c517c735797db3fca587c0ef0b410e7dd494ace2267d89cf05
        - registry.redhat.io/rhel9/buildah@sha256:6886997da8c92214e99e3e8da47a0f08239172d2544b4edae058d5cb018effab
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:c90a9c3a20f088241980b76986ba4f43b0243fd6d94a826fa7b5888e01c13093
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:776d34b0ea2e35c11ad4050e6b40cf4b40c84dbe0d5f3e6ad669c494a210ea8e
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:8b4bf043e0fb237553503d63b129c432600bb4e65b2096b246b257b404b42be9
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:eaad37453a5ebef51c28d4a068f56e9f394e710c72bb86342cd2189001420836
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:8ae0ca6411e4fcd5c0605344bc6930cf122ffd80e71199509d58423a8b41de1d
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:dbc462cb50385dec3a220b5d0caed0e0211c166182026f86e3c51a1351ba0de4
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:6abc903c784bb455e1f2fde45a48afe775239668e1d0dbc1b3cc6afca3c0b477
        - registry.redhat.io/ubi9/ubi-minimal@sha256:ac61c96b93894b9169221e87718733354dd3765dd4a62b275893c7ff0d876869
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:e27a30d471dda4983a5b8192709d283a9d082558f3caf31b30db2214e418b017
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:e63d956a0534f2e726301e00e877c4fe2f34ca4559e59d2df7567c04ded311a3
        version: 1.18.1
      entries:
      - name: openshift-pipelines-operator-rh.v1.18.1
        version: 1.18.1
      - name: openshift-pipelines-operator-rh.v1.18.0
        version: 1.18.0
      name: pipelines-1.18
    - currentCSV: openshift-pipelines-operator-rh.v1.19.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.18.0 <1.19.3'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.0.0
          - Tekton Triggers: v0.32.0
          - Pipelines as Code: v0.35.3
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.21.0
          - Tekton Results (tech-preview): v0.15.3
          - Manual Approval Gate (tech-preview): v0.6.0
          - Tekton Pruner (tech-preview): v0.1.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.19/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:22cfadc42e793553182acf05a34fb6a7ae0d7055431c2e393699ecfeb171b173
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:7a0c6794e9284209dc131f3087ffc01fd5946e20a9261422808c6dd3938407de
        - registry.redhat.io/rhel9/skopeo@sha256:993f799d58c48151ecae7a84e63860f6c0039e8e18efed5ad67e7752f4b7e408
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:fd6f4cd7801d7c20cf5a6e546022136404212da5d9997750e643f6e640b00e3d
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:2c3074ecc0c74d508020bf1e11667d868150915b55d436a43c2b4fc22815bf48
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:80af5fa2d87e7c405f16575519293ca7fccc49d635b4f3d0d334dea4fbd28721
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:25bcdb30f0fe34300c4f61f62513f4853f65fe11bafde58bcf19b7b3c3fb9af7
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:251c87e2c48e7f377bcb399ed64d50dff11de3459629a73a63bac17c353eb16b
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:363e28a82ed022e507c3976d3a42ee63c053eb7871712c213b12235d8234b7ef
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:b247e9f3a2db32b575999387ef4f91d685a67c1f3950a6e265e1fe14c1fac196
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:6aaabb9af37e46a856371e29f141de6b0fac623128099bb927fc30f8f61e9db4
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:d000daec36a6cdbe41a85ee4eac24a712f7628441a412f751baf8d7778626f62
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:8bcaeaa751c61eb927d6353984f81d6a92b4bd4f9a5be9dc620a11b06ff9e6ba
        - registry.redhat.io/rhel9/buildah@sha256:a6e04c061b8d261dba85d01dcb64d3f7dba8c0b9765e03f05ec79850a248d75a
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:41d669fbe8bb48481890fb9fb52de9d4cac98cb74e6a6b284d046a8d21b06e58
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:28258903330a0715141e4685b3760ff18b0d75da0694985bba7c3345d7808b5f
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:90a5eb2c0decb72234cbea5354bb7737cebc53c467229b2c16a1425b4bf408dc
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:3f7e1d83eb1e21860cc2ae0bfe3abe8c9058e92a69698dba9d5871069016e90d
        - registry.redhat.io/ubi9/openjdk-17@sha256:cdbd64ba1b7826d09bc180fa19aa6fbc09ddb5bff5e02b7e1351353bbabebb78
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:a90a6c269b3fc4c77652d4862878211eee5eca1d8047e561d07e3ef1292e52f5
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:31a40f8095a3f8cd04750810d964572339965ae0b2d357c57c0b671a95e1cdbe
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:1f870e70669ad584db1ee31b1dcd822d8b8d0cc03028c7cb09238fb7abf19a42
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:b8cdf2a1873a3df1fb8b8f6d58d06235cef28d228abd70d0bb83cdcfbe3bc230
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:3b1c6f8acdf17a5185413147e0248baa559378ed79e08cb098c724036020576e
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:c9fbb018a70df9a542236ace556fa3343ae350f43c0629f13e525854508f73f9
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:92b8c92eb0bec102215b97f1d60d76c7a57747aeec9594ce1e0e46f7c27391d6
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:5f39f9e61ddbce92629e6502bbef737365c8a1bb2dc2ecaf9f9d8092dea844c9
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:aa1b6f374643ed4e89ba194afde8a5138e401c8cdeb45a6bb5df5b608f58358b
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:22095c436562ff92771d9349f57d007d82415dae519e3da1de7d51e3f1042a93
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:198cf9e5c9641757a0453dec2747a24b018a9b85e10eb0dfdc7a2e75ed09cf6d
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:ba596c629fe8b5b70db97923ff11392dc12115c8e7afefee0d48c12440b8dd86
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:10c620aa3a9c4bfbca185fb460e2521473a5fe6d518342b89a3dada5cf14e65a
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:13608cfe2b87535776c65728aa40cfaa84ccf63afe64d1f2e799ea0e26819239
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:68d9007c5abb2ad3b7666249a61d9428c2cf07165bd8090a6e6def0b7cf70fa3
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:37e2f1fb8a2b4c7ac3d5390ba68a9bf98a0c83a38567b6a9778764061b3d2bab
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:c3ba6b9652205407cc4d4d47b4a928fe8fa86ce4427e6b44ee1d53faf3455f21
        - registry.redhat.io/ubi9/ubi-minimal@sha256:14f14e03d68f7fd5f2b18a13478b6b127c341b346c86b6e0b886ed2b7573b8e0
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:d21cc7e094aa46ba7f6ea717a3d7927da489024a46a6c1224c0b3c5834dcb7a6
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:2d2499b7b12aace07c4b5ab4121caad445aab5bc8dd73377171683b36d417d73
        version: 1.19.3
      entries:
      - name: openshift-pipelines-operator-rh.v1.19.3
        version: 1.19.3
      - name: openshift-pipelines-operator-rh.v1.19.2
        version: 1.19.2
      - name: openshift-pipelines-operator-rh.v1.19.1
        version: 1.19.1
      - name: openshift-pipelines-operator-rh.v1.19.0
        version: 1.19.0
      name: pipelines-1.19
    - currentCSV: openshift-pipelines-operator-rh.v1.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    },
                    {
                      "name": "legacyPipelineRbac",
                      "value": "true"
                    },
                    {
                      "name": "createCABundleConfigMaps",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.19.0 <1.20.0'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.3.2
          - Tekton Triggers: v0.33.0
          - Pipelines as Code: v0.37.0
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.22.3
          - Tekton Results (tech-preview): v0.16.0
          - Manual Approval Gate (tech-preview): v0.7.0
          - Tekton Pruner (tech-preview): v0.2.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.20/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:bc403f93ef6c5526fa30152eab384c4a25bb0bc6ab4158a1e24f637b87a57259
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:980ed7e81ea55522e1681dc1e9b170b68d16f07061dd1f1ffd142367d59552f1
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:bc10e7d6122ede5cf4f2882c200447683255465a1080e826761e90bfb983506e
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:c3013c347bca8a520443a44ac2817fbba5744a9e91069ac5f09e91447ebf6b28
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:1b6aecfce45ed3163227b845b212e83385cdbe999edfc2e9387a0c76f3f8d8e9
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:5683321a5376e63c3ff3932270a1850d5e261f76212020ce867334166f90d2f7
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/ubi9/ubi-minimal@sha256:2f06ae0e6d3d9c4f610d32c480338eef474867f435d8d28625f2985e8acde6e8
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:7eafffa6e4e04f10c1d4e84e02d51c00eb8649c60f8f9a13da14492f24634891
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:d56d6b46331b4569077ab1e7a7b609ba01f0a1dd62b695d62f8aa0f32ea1b621
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:e63298d07ccec5804dfab4e42bcc8163c29cb3ac371c91248777423a4edd31ce
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:8f0494341244a75546a4b80b065702be04e6d200421aa4e7ec90385c9dabc6ab
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:c31af237afb84214bf2a4c06a7636ca071fe97203b41da90d7878109ccd64784
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:115eea1a5ca87f3b6b72db1df05b7386bf73dc70c9c2e4dc5e1a66700d960c37
        - registry.redhat.io/rhel9/skopeo@sha256:a65a413f8a2864389a09dc750690d97afbcdc5c70821e0f85e99e8adba7954e0
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:dd941acd357ae2c0fd7e15fb8d32d7ee6b2847435101dd1e1f107db48af18bf3
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:5021158a4752a26f8b33adea7da66695b22d2ad155d346d35dde5f971bc1a55b
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:823e018e0329d764ac694f177ce88af9074f364c29a1c17929f9dd0b596d03dd
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:14dab8d87323fc7d112213f7f4db3ee7cf3841747de669ce710ae6a794053457
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:bbd1918660be7b3c78777dcccc87d3b26bf5adfc077844ceffbbdd53696dc0a3
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:a7163a9e1fe891de021b1003558e6e64fe17ed17ab0820ba16bba09aa112401e
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e7998e113f75e8f29602bae210c8b66ed06634b699f1d99c5875ef879220d733
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:ec74a562c18fefe36ec7ae3ba083e4eb76c56b562f916578900f8767ae98e6c2
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:c4d05e546c466c57379fff502816012f615be073eb8cdce54677b4ab206a3f5c
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:c0f438e6095aa81f3f7d820614e0414cc43374014c1505d7641bc1dba69a3903
        - registry.redhat.io/rhel9/buildah@sha256:18453ab0a62154283aaf4c45efd7543d1c8ecca1aebae46dedc7ddc0f410b232
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:818bfa4a3a874974b5f7f8bcf2faeb6c05e1795381ee1ceacf89f17b1ec0ef7f
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:9415e889ef98df6d2b2dca1e5d798fb765eeea0d9828e48424374d5267a6b0d1
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:a712bb411d2433b01793dff25785c1e265fa865a52f7c8732c9fb4418ab61757
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:3cf990fb2fc422f3527a7c8b43d06ccd4d32bb5487ec8cb2d4a97f0f5d13f170
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:458f865f1b246c5f0682f5956795771a0515b89edfcffbbc0da1a3d3f532645e
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:96db721939b22572d7d8d8b95dcb5c92fdedb98e9cd0aabb7d17b703a915b07f
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:036250bcdc6fb95273bf07948dfdd4dce765ab84c803ceca0d0b4abf6ca19dec
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:161df929e2bd4c5acff594be31d5458365de5ae0d2189bb2f647c3fc6657fa28
        - registry.redhat.io/ubi9/openjdk-17@sha256:555c8bcd2596637befd1a6da4aa949c5a2640cf66e09c2dfb932afe37e39095e
        - registry.redhat.io/openshift-pipelines/pipelines-opc-rhel9@sha256:c1bee4a2bb495de6d4da962b6e7b2447caa51cc36c3005e4083bd31420ef5102
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:901b5108ceb669c985e75f5fb033c496e9f028863df99ab32651d1b08433a352
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:df15eb1f1f5f237bfe481c16ab3d40bf7fb4123f3c930bcfb121e2cc8dc42882
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:441e1c7124c1674e3d4c38c3e82a4fe5568e76d13c89c04bf009d6b97435de09
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:e3e64abef53151f6b15f15bcbff85668167a571e98badbd8a2ca252ff3a5b8dc
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:2d496752b19f765f1ea36ec94b873ebd8fa604764a56aa67d3ed94b33e5fe75c
        version: 1.20.0
      entries:
      - name: openshift-pipelines-operator-rh.v1.20.0
        version: 1.20.0
      name: pipelines-1.20
    defaultChannel: latest
    packageName: openshift-pipelines-operator-rh
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataTrucker.IO
      provider-url: https://www.datatrucker.io/
    name: datatrucker-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datatrucker-operator.v2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datatrucker.datatrucker.io/v1",
                "kind": "DatatruckerConfig",
                "metadata": {
                  "name": "datatruckerconfig-sample"
                },
                "spec": {
                  "API": {
                    "crypto": {
                      "CredentailEncryption": {
                        "Algorithm": "aes-256-cbc",
                        "IV": "eThWmZq4t7w!z$C\u0026",
                        "Key": "JaNdRgUkXp2s5u8x/A?D(G+KbPeShVmY"
                      },
                      "JWT": {
                        "authorization_keys": {
                          "private_key": "privateJWT.key",
                          "public_key": "publicJWT.key"
                        },
                        "signOptions": {
                          "algorithm": "RS256",
                          "expiresIn": "60m",
                          "issuer": "datatrucker"
                        },
                        "verify": {
                          "issuer": "datatrucker"
                        }
                      },
                      "keycloak": {
                        "client_id": "nodejs",
                        "client_secret": "7c5542ae-7264-420d-bf5f-766f4ab39ad3",
                        "jwt_publickey_verifier": "keycloak.key",
                        "realm": "TestRealm",
                        "url": "http://localhost:30015"
                      },
                      "passwordHash": {
                        "Algorithm": "sha256",
                        "SaltLength": 16
                      }
                    },
                    "cryptokeys": {
                      "keycloakkey": "LS0tLS1CRUdJTiBQVUJMSUMgS0VZLS0tLS0KTUlJQklqQU5CZ2txaGtpRzl3MEJBUUVGQUFPQ0FROEFNSUlCQ2dLQ0FRRUF4ZnhGY1gxb3RmTUN6SWlkZ09tMndZakZWZXZHN1BQT1h5dTNrTEFuTWtTVkUxUXl4TzJ4SFFvSGcydDFVcUU4SVNXeTN3ZkZTbERzdU1VK0J1VVkvaWI5L04vSlpBV3k5MGRET0YxREdIeGVwM21HMVBTUllMQWFGQ0I3NTVsbWZjL1dna3VzZFdLdmxOTmZuNkd6S1J6OXBmV3RNUnhWWkFrejhmNlVUeG0zWE02Q0tOMUludThubEcxU3ZzaFc5cmt3Lyt6UmVvOFVCV2pPcTh5RjdPV0tnS0ZZZlhENFREa090OWFxN3RaaSsxaExYOE5iV2xxRnJZYmhJcFZINXlLSko3ZTA5cUVIRkVEelg5QXJzVVNHamNjbFFFYTl1dmJxU2d4Rzc3RUZVbk82aDZxUkNnU3lNNnVYdk9qcWlIMHp4RTNmRHB4R2NhY2pvTW1EMlFJREFRQUIKLS0tLS1FTkQgUFVCTElDIEtFWS0tLS0t",
                      "privateJWTkey": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQ0KTUlJRW9RSUJBQUtDQVFCVWd1ZDN1OFlXcUJCWWtYZ3M2K25pMS84bnM5YjRGNkZudzA4RDRKdGRlalg5L1ZaMg0KWFdOamJ1ekMxUkRHMnVseFM4QjZIdTZYeEhYWU9XU2ZzOEY0UHZ2eC9TUWdmRWtWUlNDL1pQN3pzM1ZETllaMA0Kc1JScDNSZzUza29qUG80T3o2azFaeGprbFRFZEpMWUtqYlZud3U1M3A3MTFsVTJkV3B4R3VEeGYxVTJuUWgrWA0KaUhKZkhyUmxTZVA1aVhYQkpxd21VTDFkK25lTU5BOXlGakR0MnNZMG5iSDRySDVxbk1VRk4xMEpsZ3V4NFdiUw0KaGFUNklNN3h1dUVHbDRDckVENmdTejBNNGF1UVk4ekZxQWxJNlFWUTlML28zQnBCRmxYUUMwWmowb3lwQVFKLw0KQ1VLaU9FcEhLRHgram02RU9KOE1aejlMQ01yTmdLME5KSWlaQWdNQkFBRUNnZ0VBRTBMMnNLSjZob1Z0bkZJag0KZXRHV01CNmN5QmJFSXkxM215ZTNwYUdXcDVxSUhjaVU0VXR0c0tjQ2pYRjBibTVNa0liOGFwa01Oa2tQdHVWMQ0KMnJSa254eHozZmFkT2s0NWRnQ1ljeWI0dU9DcEZGVnB2TVR6YXlGNk9RNUNCWUJuYmxiRU1TaTdVclh5U0xzcQ0KUm9OLzJoTml1S0RobktBWDNZcGVYZnYxdnIyaWIrMGtiRDVvMVZLcnB6SlJ5emlsWHFwQWJwNDQ4RExFZHVGeA0KTnJKc0tWb1dXMVhmS1FvR2VDMXZqNFJudzdhWkNyU1BjMEhvMUdKb29xSXlrNjRXZnhGWjhpVElsWDJMU1gzbQ0KQm1pN1ZXZnVJK0QvUXRtWEVpN05oTE1jUmExVXB4QkdiK1ZPZXlucEVyMEpxM3VZSmlva1VIZzloOGIzSDM1Yg0KK0lJaEFRS0JnUUNtbzUrRUNpOEJUcFRHdlpqQXBtdmxGZ055NkpWd0RvVHJLSGdCcHJHVHpsVGt6Q0tLYzJhTw0KTTMvQmVIVUQ1NVFqcTdUM094eGgzQUk3cG9kT0tCWS84eXgzenlYUExQUmVCMlR0eTNVQjE5YU5CclhNUGQ5RQ0KTG5hbUQ0SmZQeTV0WHpWWEloWVB0UzBiNjdGbjZnZFl4MlNBbGFYYWFjejU4eURLZnlZL3VRS0JnUUNCMUxNYg0KTFZNTU1Pd1U4R2xsVVB1V0Q5VnBsMnlURytCN2Nnd0ZiYTlrWVlVSFQ4TmlvaUdxQzlLZWxqeklKNnlUdyt3aA0KQisvNGNJU1o5Y0JDNHB0VnRlTWJwUkg3UTQwK20xU3M1SmQzU1VFcDBMOW5VbmdVNDhrMzlpSTV5aEgxU3JTRw0KREtBamVka2Rjb3hNVWFwUHB6cVprOVNKdGVHckVsTHY5WmcvNFFLQmdHcExQTmt3MWNMenFSc01uWDlscUxVOA0KVVg3ZXRWUUJMdkFCcEdzSk5ZU2V2TEFsTm5NeUlWemJwTVo0Vzh2ZDBEVGd0bEp0dGpFRjhiYXpCdVBuMWlsOA0KcUFUVWNiMnhZejhIWUczK3RlZ3pFWHhycDFOY3RDbjlxVmxMMDFHR0w4czhTLzRFbENYNDM0bEwwcUlDT2lWNw0KS3JTS3dIeU51RjVIbzhjSFRCS2hBb0dBVjhVRVpENEs0N1g3RTlqSFY3QWN0UUQxUkpJUzI0ekFBSlJFSGpOZA0KbkIzMVFkRWhoSCtaQkkxRGtPUitLTTN4ZEYvQ05UNVpjM29WbXNPeGUwWVl2eVBjYWViNU5nVWY2cEExK1QvOA0KU2E4ak4zS2d3dWZVdUVLWjJKUXo1UjkvYm5WWGZ6dnYwcXZnSkhFNCthdEgyTnR6MlRsQU9PelZwOElXZ1pSbA0KbmVFQ2dZQXpVS0RLS1ZKaXBYdC9TRC8xd2ZzWElCazFxYWpmdENHOFVTcFlGWGp5VzVYMlZ5bTlVTldmVEdJcA0KQlJkVldXamdFbWhyOU5Da1UrVEU4WEVjTXJNRnhWbXBEeTQ4YmxsQWltWFJtOEhNMlhKODQ4TlArTG5GdEtHdA0KTWswT3JuUHdPYitualZGcXdWUllNQ0QzeFN0Z2RnZUppUHFYcWd3SDhjR0M0OGVmZ0E9PQ0KLS0tLS1FTkQgUlNBIFBSSVZBVEUgS0VZLS0tLS0=",
                      "publicJWTkey": "LS0tLS1CRUdJTiBQVUJMSUMgS0VZLS0tLS0NCk1JSUJJVEFOQmdrcWhraUc5dzBCQVFFRkFBT0NBUTRBTUlJQkNRS0NBUUJVZ3VkM3U4WVdxQkJZa1hnczYrbmkNCjEvOG5zOWI0RjZGbncwOEQ0SnRkZWpYOS9WWjJYV05qYnV6QzFSREcydWx4UzhCNkh1Nlh4SFhZT1dTZnM4RjQNClB2dngvU1FnZkVrVlJTQy9aUDd6czNWRE5ZWjBzUlJwM1JnNTNrb2pQbzRPejZrMVp4amtsVEVkSkxZS2piVm4NCnd1NTNwNzExbFUyZFdweEd1RHhmMVUyblFoK1hpSEpmSHJSbFNlUDVpWFhCSnF3bVVMMWQrbmVNTkE5eUZqRHQNCjJzWTBuYkg0ckg1cW5NVUZOMTBKbGd1eDRXYlNoYVQ2SU03eHV1RUdsNENyRUQ2Z1N6ME00YXVRWTh6RnFBbEkNCjZRVlE5TC9vM0JwQkZsWFFDMFpqMG95cEFRSi9DVUtpT0VwSEtEeCtqbTZFT0o4TVp6OUxDTXJOZ0swTkpJaVoNCkFnTUJBQUU9DQotLS0tLUVORCBQVUJMSUMgS0VZLS0tLS0=",
                      "truckerapicrt": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSUQxVENDQXIyZ0F3SUJBZ0lVUFVzZ2NYVXZRYWtuQldjeUFDeEw2dmt3b204d0RRWUpLb1pJaHZjTkFRRUwKQlFBd2RqRUxNQWtHQTFVRUJoTUNRMEV4Q3pBSkJnTlZCQWdNQWs5T01ROHdEUVlEVlFRSERBWk5hV3gwYjI0eApGekFWQmdOVkJBb01Ea1JoZEdGVWNuVmphMlZ5TGtsUE1SUXdFZ1lEVlFRTERBdEZibWRwYm1WbGNtbHVaekVhCk1CZ0dBMVVFQXd3UlJHRjBZVlJ5ZFdOclpYSlNiMjkwUTBFd0hoY05NakF4TURFeU1UWXlOalV5V2hjTk1qTXcKTVRFMU1UWXlOalV5V2pCM01Rc3dDUVlEVlFRR0V3SkRRVEVMTUFrR0ExVUVDQXdDVDA0eER6QU5CZ05WQkFjTQpCazFwYkhSdmJqRVhNQlVHQTFVRUNnd09SR0YwWVZSeWRXTnJaWEl1U1U4eEZEQVNCZ05WQkFzTUMwVnVaMmx1ClpXVnlhVzVuTVJzd0dRWURWUVFEREJKa1lYUmhkSEoxWTJ0bGNpMWhjR2t1YVc4d2dnRWlNQTBHQ1NxR1NJYjMKRFFFQkFRVUFBNElCRHdBd2dnRUtBb0lCQVFEakM4ZmVCVEg4L3NnejFQcTJQYU9QQjc5YWtyeUZvWVpGV292bQpWOXNvMmJNRHp4NVBZSFJydVFGYmJQRzlVVlc2QWZpSHdEb1p2Z3NTVXpyS3ZyR1RlYjluNUhSVlVOMTl0NExTCjdtUFppNzFRbGE4bVR3QmdvTnE4TFppcnZUdzV0S2I4K3NmcFJxQUlUQkVsRmxVUEZISzllcDl2eFdvaHhwdGgKS1dPVkwvV2ZtWldhWXlRdjlLNDBkWUJpRWpGcTFKZVB4MDlnUFdHM2pOT1VYTTZQUVlMaHJDUXFFcy9VME9GWQpwVFRHUmNXZEx4TFVFZVNkcHJKK1ZEc3lEK1lVeUFRR2phYk14VmVTVlRUN1FGbFBjLzc3OVZsRkdXMC8xWVVaCmpqakNsSkFLUWZOTHg5bG85cWJERVBZL3grZmhnRnYyWUVaSTNTdGVyZWUvVDJoWEFnTUJBQUdqV2pCWU1COEcKQTFVZEl3UVlNQmFBRkNJVy9JYkZIUUNBeEdxZC9FVFpXbVhGeGMvQ01Ba0dBMVVkRXdRQ01BQXdDd1lEVlIwUApCQVFEQWdUd01CMEdBMVVkRVFRV01CU0NFbVJoZEdGMGNuVmphMlZ5TFdGd2FTNXBiekFOQmdrcWhraUc5dzBCCkFRc0ZBQU9DQVFFQUxrMlJDbnVweDA5Wmx0SmN3M1pUaFIveUZGNS9KVlVsWkdHTnc3MlAwQmR6SEloQncrVk4KRkNucGFPMjJ0WGF1NlJVeTVNUWhwc3JrbFNKYVdJT25jOUg0Uk0vbGh3YWsyRXN0MkFhdUFwNUdYSVFJMDVBeAo3NmtDZXEyMnUyalNXRWpFS1hGRnRiZXVDZnFpWlpqNWFLcGF1YlZWRG4ydEZTWjJRMFRib2RRN2dpTFdNUTBLCk5ycVgwQWNSU2J0ZnRSNjBhdGQ1YW41aFNXY0dyZUg5ZHdFUVRvVVNFblg2RFRZSURoeElsZStjQ2V5VjVSWk8KVS9hck5RZ0lFY3ZrZUM1d1lZaFFFL1FUN3lNb1lNajFnckxmOTZSZkRhbEVFNE1seUtDWm9wYWNJWUdzSmhuKwpLMWFISEQvbGZxTC9zbnNtY3pXdHlLUncxTDF1WkxPcjJnPT0KLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=",
                      "truckerapikey": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQpNSUlFb3dJQkFBS0NBUUVBNHd2SDNnVXgvUDdJTTlUNnRqMmpqd2UvV3BLOGhhR0dSVnFMNWxmYktObXpBODhlClQyQjBhN2tCVzJ6eHZWRlZ1Z0g0aDhBNkdiNExFbE02eXI2eGszbS9aK1IwVlZEZGZiZUMwdTVqMll1OVVKV3YKSms4QVlLRGF2QzJZcTcwOE9iU20vUHJINlVhZ0NFd1JKUlpWRHhSeXZYcWZiOFZxSWNhYllTbGpsUy8xbjVtVgptbU1rTC9TdU5IV0FZaEl4YXRTWGo4ZFBZRDFodDR6VGxGek9qMEdDNGF3a0toTFAxTkRoV0tVMHhrWEZuUzhTCjFCSGtuYWF5ZmxRN01nL21GTWdFQm8ybXpNVlhrbFUwKzBCWlQzUCsrL1ZaUlJsdFA5V0ZHWTQ0d3BTUUNrSHoKUzhmWmFQYW13eEQyUDhmbjRZQmI5bUJHU04wclhxM252MDlvVndJREFRQUJBb0lCQUZSbFo2aiswMk1RcmdpZQpUWWQ2aUJhOXNZS1FjdVZWY2x6TFdKK0ozdHhqZCtpZGo4Mi9zK2JyQVRjajJ3UWRmTGxFekVaaHpQY09ybnprClZVRlIreDR2cmxlSEV6czNjTzJ2VGtEci9rKzZZdWZXakNHNFVoS1pHRlhFdVRCMWlqS0ZGV0dLTXM5aXdrL3oKUHBoRG9lQ1dJRVNRbFQxdDY1S2NpZHRuRUM0K3oxaURxR21ZZEVEa2xEZFFIVHorRGpseS80VzV2MFpVQWN2Vwo1Unkvd0JldmNURlBXQy9uZGFrN3hGUnVrU2xNZUlhTWhvTlllOU5uVCt2Nmh5VVh1MXp6cG45UEFJRTVpQWpOCnBsZkFEZitwakwrZFJDbW9sWW1LR3NQVjEzSmtCY1p2cEVoUWgrUmJnaWFwczJLc2plQ0NmTmtST0hndHc0MU0KeGxOTU1Ha0NnWUVBOHM1akpuZ2FMblIvTzZtakZEb0wvd1p2UWdnT1h3Z1RBN2tsQ0trbjBYM0xJS0hhaTZESgpyOVRGQktMdkhUazVFK1Y4NkhIZ29VMUFQQnYyMWJpaHhuQXhUYm8wSHZMdG5QOXU3YnZJejFNOWMyd2ZlWUgzCmliMG1FNkV0Y2wvNnVIUHdMVlZtRGh5bllneEZpeTRRRXp1ZHZwaFIvSnlPazQvUytQUzh2ZXNDZ1lFQTcySW8KWDlkY0xvUmYzSk0vRTF1SXhrRGRLZm1VakN6Mk1NN3pyRkpNUERobTFSUlJhQjdidHNxUzl4Mm5KeTI5amtoSQpndTR0Y2s5STdncm9WRXY0cEtkU0lOVzVNVUg0MXJPckxTbFBobCtMaTM5QXNrTERFVXVQbkFmYko3YWZibFFHCjNFWE1Ka1d2WFQxRG90aE11MUpUMTFPMjhmYlBmb3R3YVRucXFFVUNnWUE1RnArblh0dkJDbkVUNVJuTDJxNFIKWTBEcmhKdGthckNWMG9WTG1RSWN0bmY1RHpVa1dVU3lpOU5hK0xoQjViOXUxeWxjT2RJMytzTGNuektWSTVrZAo0Q2tuZmwxYzNYOVN4NWdjVzJlbW40cTJSYmpNeCtzY3ZwVmlUaVl5V3pESEI3Tno1VDFlbWd3YytJZXRac0l1CjAzaU9aSi82UmhFYnp0Z3E0dGd4eXdLQmdCanhwSDFnSm9YZVcxTFRzUlhtU3RsRnR3WUJzTDN1alhUR2k5T2YKKzd1bXNPaHhKbzV2OExzYVhyZ0xTZmltSm9ObCt0QlIyMEIzRElWcGM0NGRJSHFmeU1vTkd6QUx6TThxaDZLOApGRUpQUUc0M3cxa1A1ZDJ1ZnZkbVJpcmNONzY2V01DcE1UMm5zejlHKy81REJiUGJQMHRIdU45N3RXRmRURWhhCnRaMzFBb0dCQU5xZnVycEV4ZHozdXRNUm91RnJjTmVkV2s4VlVpVVZIV1hZMjhNNWQvRjcxMExIOTZuOW12V08Kc1cwcDFjcXBJMEsyQko2WlYwNVpybnlyRHBsVVpVdVpSQWxGR3hYR1dLU2trQzlmWm90SkFOWDBkR29KSitqTwpEYUNLNGxlcXdOMHNPRE1YZFErMlFPaUtlSUxkNU5KcDNUSW1YV1Z0RFlXUkRwOW9iSUdlCi0tLS0tRU5EIFJTQSBQUklWQVRFIEtFWS0tLS0tCg=="
                    },
                    "jobsServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "debug"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": true,
                        "FileSFTP": true,
                        "IOTKafka": true,
                        "IOTProxy": false,
                        "IOTRedis": true,
                        "Login": false,
                        "ScriptJS": true,
                        "ScriptSSH": true,
                        "ScriptShell": true,
                        "UtilEcho": true,
                        "UtilFuzzy": true,
                        "UtilSentiment": true,
                        "credentials": false,
                        "jobs": true,
                        "resources": false,
                        "statuschecks": true,
                        "ui": false,
                        "usermanagement": false
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "loginServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "info"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": false,
                        "FileSFTP": false,
                        "IOTKafka": false,
                        "IOTProxy": false,
                        "IOTRedis": false,
                        "Login": true,
                        "ScriptJS": false,
                        "ScriptSSH": false,
                        "ScriptShell": false,
                        "UtilEcho": false,
                        "UtilFuzzy": false,
                        "UtilSentiment": false,
                        "credentials": false,
                        "jobs": false,
                        "resources": false,
                        "statuschecks": true,
                        "ui": false,
                        "usermanagement": false
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "managementServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "debug"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": false,
                        "FileSFTP": false,
                        "IOTKafka": false,
                        "IOTProxy": false,
                        "IOTRedis": false,
                        "Login": true,
                        "ScriptJS": false,
                        "ScriptSSH": false,
                        "ScriptShell": false,
                        "UtilEcho": false,
                        "UtilFuzzy": false,
                        "UtilSentiment": false,
                        "credentials": true,
                        "jobs": false,
                        "resources": false,
                        "statuschecks": true,
                        "ui": true,
                        "usermanagement": true
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "name": "API",
                    "resource": {
                      "Cache": {
                        "Creds": false,
                        "CredsFlush": 10,
                        "Resource": false,
                        "ResourceFlush": 10
                      },
                      "Templates": {
                        "Scripts": "scripts",
                        "Shell": "shell",
                        "keys": "keys",
                        "resourcedefinitions": "resourcedefinitions"
                      }
                    }
                  },
                  "DB": {
                    "databasename": "userdb",
                    "hostname": "db",
                    "password": "password",
                    "port": 5432,
                    "type": "pg",
                    "user": "testuser"
                  },
                  "TempDB": {
                    "enabled": false,
                    "pvc": "samplepvc"
                  }
                }
              },
              {
                "apiVersion": "datatrucker.datatrucker.io/v1",
                "kind": "DatatruckerFlow",
                "metadata": {
                  "name": "datatruckerflow-sample"
                },
                "spec": {
                  "API": {
                    "Image": {
                      "imageName": "datatrucker-server/datatrucker-api",
                      "imageNameUI": "datatruckerio/datatrucker-ui",
                      "repository": "registry.gitlab.com/datatruckerio",
                      "tagName": "latest"
                    },
                    "name": "api"
                  },
                  "DatatruckerConfig": "datatruckerconfig-sample",
                  "JobDefinitions": [
                    {
                      "name": "echoapi",
                      "restmethod": "GET",
                      "tenant": "Admin",
                      "type": "Util-Echo",
                      "validations": {
                        "properties": {
                          "voila": {
                            "enum": [
                              "true",
                              "false"
                            ],
                            "type": "string"
                          }
                        },
                        "type": "object"
                      }
                    }
                  ],
                  "Keys": {
                    "configmap": "placeholder"
                  },
                  "Replicas": 1,
                  "Resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "256Mi"
                    },
                    "requests": {
                      "cpu": "250m",
                      "memory": "256Mi"
                    }
                  },
                  "Scripts": {
                    "configmap": "placeholder"
                  },
                  "Type": "Job"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/datatruckerio/datatrucker-operator:2.0.0
          createdAt: "2021-09-08 01:01:01"
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: hhttps://gitlab.com/datatruckerio/datatrucker-server.git
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              - creates secrets to hold server configs
              - create tempDB for protoypting
              - create configs for seperating Login, management and Jobs endpoints
              - Create crypto configs
            displayName: Datatrucker Config
            kind: DatatruckerConfig
            name: datatruckerconfigs.datatrucker.datatrucker.io
            version: v1
          - description: '- creates an API as per Job Definitions'
            displayName: Datatrucker Flow
            kind: DatatruckerFlow
            name: datatruckerflows.datatrucker.datatrucker.io
            version: v1
        description: "## DataTrucker.IO\n\nDataTrucker.IO is a simple no-code / less-code
          API Backend\n\n### What does it do ?\n\nDatatrucker.IO is a product capable
          of reading simple json/yaml configs and building th code necessary for converting
          it into an API.\nAlong with building the code, it will also host the code
          base on a nodejs server , i.e. it will immediately make it available for
          consumption.\n\n### What does it offload from a developer ?\n\nDataTrucker
          is capable of removing the most common activities which a developer needs
          to do on every new project A few of such common activities are\n\n- Creating
          an API endpoint with a specified business logic (using simple plugins)\n-
          Applying standard RBAC\n- Applying Authorization logic\n- Applying hardening
          on endpoints\n- Log management\n- Connecting to variety of systems\n- Modularizing
          business logic\n- The best of doing it with little to no code\n\n### watch
          the quick elevator pitch to understand datatrucker.io\n\n https://youtu.be/DKLFDqhjs5M\n "
        displayName: DataTrucker.IO
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - datatrucker
        - operator
        - nocode
        - apibuilder
        links:
        - name: Website
          url: https://www.datatrucker.io/
        - name: Git Repo
          url: https://gitlab.com/datatruckerio/datatrucker-server.git
        maintainers:
        - email: gauravshankar.can@gmail.com
          name: Gaurav Shankar
        maturity: alpha
        provider:
          name: DataTrucker.IO
          url: https://www.datatrucker.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - docker.io/datatruckerio/datatrucker-operator:2.0.0
        version: 2.0.0
      entries:
      - name: datatrucker-operator.v2.0.0
        version: 2.0.0
      - name: datatrucker-operator.v1.3.2
        version: 1.3.2
      - name: datatrucker-operator.v1.3.1
        version: 1.3.1
      - name: datatrucker-operator.v1.3.0
        version: 1.3.0
      name: alpha
    defaultChannel: alpha
    packageName: datatrucker-operator
    provider:
      name: DataTrucker.IO
      url: https://www.datatrucker.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Syndesis Team
      provider-url: https://github.com/syndesisio/syndesis
    name: syndesis
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: syndesis.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta2",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.12
          createdAt: "2021-08-11T12:00:00Z"
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta2
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.12
        - centos/postgresql-10-centos7
        version: 1.12.0
      entries:
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: 1.12.x
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.13
        - centos/postgresql-10-centos7
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: 1.13.x
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.13
        - centos/postgresql-10-centos7
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: candidate
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.13
        - centos/postgresql-10-centos7
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: latest
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.13
        - centos/postgresql-10-centos7
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: stable
    defaultChannel: latest
    packageName: syndesis
    provider:
      name: Syndesis Team
      url: https://github.com/syndesisio/syndesis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.22.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.5"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.22
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.22
        version: 42.22.0
      entries:
      - name: t8c-operator.v42.22.0
        version: 42.22.0
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: stable
    - currentCSV: t8c-operator.v42.10.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.10
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.10
        version: 42.10.0
      entries:
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.10.0
    - currentCSV: t8c-operator.v42.10.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.10
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.10
        version: 42.10.0
      entries:
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.12.0
    - currentCSV: t8c-operator.v42.13.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.7"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.13
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.13
        version: 42.13.0
      entries:
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.13.0
    - currentCSV: t8c-operator.v42.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.7"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.14
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.14
        version: 42.14.0
      entries:
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.14.0
    - currentCSV: t8c-operator.v42.15.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.6.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.15
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.15
        version: 42.15.0
      entries:
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.15.0
    - currentCSV: t8c-operator.v42.16.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.16
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.16
        version: 42.16.0
      entries:
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.16.0
    - currentCSV: t8c-operator.v42.17.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.6.6"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.17
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.17
        version: 42.17.0
      entries:
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.17.0
    - currentCSV: t8c-operator.v42.18.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.1"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.18
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.18
        version: 42.18.0
      entries:
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.18.0
    - currentCSV: t8c-operator.v42.19.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.2"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.19
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.19
        version: 42.19.0
      entries:
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.19.0
    - currentCSV: t8c-operator.v42.20.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.20
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.20
        version: 42.20.0
      entries:
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.20.0
    - currentCSV: t8c-operator.v42.21.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.21
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.21
        version: 42.21.0
      entries:
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.21.0
    - currentCSV: t8c-operator.v42.22.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.5"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.22
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.22
        version: 42.22.0
      entries:
      - name: t8c-operator.v42.22.0
        version: 42.22.0
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.22.0
    - currentCSV: t8c-operator.v42.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.4.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.6
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.6
        version: 42.6.0
      entries:
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.6.0
    - currentCSV: t8c-operator.v42.7.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.4.6"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.7
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.7
        version: 42.7.0
      entries:
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.7.0
    - currentCSV: t8c-operator.v42.8.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.0"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.8
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.8
        version: 42.8.0
      entries:
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.8.0
    - currentCSV: t8c-operator.v42.9.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.1"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.9
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.9
        version: 42.9.0
      entries:
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.9.0
    defaultChannel: stable
    packageName: t8c
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-db
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-db.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Db",
                "metadata": {
                  "name": "db-sample"
                },
                "spec": {
                  "clusterDomain": "cluster.local",
                  "databases": [
                    {
                      "name": "wapdb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "grafanadb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "brokerdb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "nimbusmosaic",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    }
                  ],
                  "image": {
                    "migrationJobImage": "wap-ose-cli",
                    "migrationJobRepository": "prod.harbor.keysight.digital/loadcore",
                    "migrationJobTag": "v4.13.0-3747-35-ocp",
                    "pgImage": "postgresql-ubi-image",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "pgTag": "14.2.0-3747-32"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "postgresql": {
                    "image": {
                      "pullSecrets": [
                        "load-core-secret"
                      ],
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "postgresql-ubi-image",
                      "tag": "14.2.0-3747-32"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "pgHbaConfiguration": "local all all trust\nhost all all 0.0.0.0/0 trust",
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "storageClass": "managed-loadcore-storage"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "volumePermissions": {
                        "containerSecurityContext": {
                          "runAsUser": "auto"
                        }
                      }
                    },
                    "serviceAccount": {
                      "automountServiceAccountToken": false,
                      "name": "load-core-service-account"
                    }
                  },
                  "storage": "8Gi"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:30:46Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Db
            name: dbs.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP DB Operator
        displayName: Keysight WAP DB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Db
          url: https://keysight-wap-db.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-db@sha256:5689d6a4da2fdd97223bc64aa61a8574bba810f93d09ab804cb45b4475f80ff5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-db.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-db
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-certified-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "tag": "5.4.6",
                  "oem": "",
                  "imagePullSecrets": "",
                  "psp": false,
                  "rbac": true,
                  "serviceAccount": "default",
                  "leastPrivilege": false,
                  "global": {
                    "cattle": {
                      "url": ""
                    }
                  },
                  "bootstrapPassword": "",
                  "autoGenerateCert": true,
                  "defaultValidityPeriod": 365,
                  "internal": {
                    "certmanager": {
                      "enabled": false,
                      "secretname": "neuvector-internal"
                    },
                    "autoGenerateCert": true,
                    "autoRotateCert": true
                  },
                  "controller": {
                    "enabled": true,
                    "annotations": {},
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": "registry.connect.redhat.com/neuvector/controller@sha256:65ee3723c7c007db5ce3cd0493a785d4163d972cf957de6dec956b842807a551",
                    "imagePullPolicy": "IfNotPresent",
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "searchRegistries": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "nodePort": "",
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      }
                    },
                    "ranchersso": {
                      "enabled": false
                    },
                    "pvc": {
                      "enabled": false,
                      "existingClaim": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {
                        "userinitcfg.yaml": {
                          "users": [
                            {
                              "Fullname": "admin",
                              "Password": "",
                              "Role": "admin"
                            }
                          ]
                        }
                      }
                    },
                    "certupgrader": {
                      "env": [],
                      "schedule": "",
                      "imagePullPolicy": "IfNotPresent",
                      "timeout": 3600,
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "prime": {
                      "enabled": false,
                      "image": {
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "1.0.7"
                      }
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/enforcer@sha256:72a628a5c811c95bb267bbd0a5ffc9d74dbaca03a92aae615d43c7601b4d3b43",
                    "imagePullPolicy": "IfNotPresent",
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "env": [],
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/etcd"
                      }
                    ],
                    "resources": {},
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "manager": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/manager@sha256:89a42c9ee86f98ef56b5e695e23bdcc6334361ebd29cfb7729d7c86f04da98b0",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "",
                    "env": {
                      "ssl": true,
                      "envs": []
                    },
                    "svc": {
                      "type": "ClusterIP",
                      "nodePort": "",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": "",
                      "tls": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "topologySpreadConstraints": [],
                    "affinity": {},
                    "podLabels": {},
                    "podAnnotations": {},
                    "tolerations": [],
                    "nodeSelector": {},
                    "runAsUser": "",
                    "probes": {
                      "enabled": false,
                      "timeout": 1,
                      "periodSeconds": 10,
                      "startupFailureThreshold": 30
                    }
                  },
                  "cve": {
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "cacert": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                      "image": "registry.access.redhat.com/ubi8",
                      "imagePullPolicy": "IfNotPresent",
                      "schedule": "0 0 * * *",
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": "registry.connect.redhat.com/neuvector/scanner@sha256:b632f1d31998a4c8bad9f97216e8aa56cd7e6446bd70d2089d6c66845b964e8d",
                      "imagePullPolicy": "Always",
                      "priorityClassName": "",
                      "resources": {},
                      "topologySpreadConstraints": [],
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      },
                      "volumes": "",
                      "volumeMounts": ""
                    }
                  },
                  "resources": {},
                  "runtimePath": "",
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhooksvc": {
                    "enabled": true
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  },
                  "lease": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "true"
          containerImage: registry.connect.redhat.com/neuvector/neuvector-operator@sha256:e2f19c50ebc4127a83c943dbcfa1312ec46dec71bd36f6d99178b320d8da970b
          createdAt: "2020-12-02 14:19:21"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: Neuvector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm/tree/master/charts/core) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart.  NeuVector Operator versions are tied to NeuVector product versions. Version 1.5.2 of the NeuVector Certified Operator deploys version 5.4.6 of NeuVector.

          **Complete below steps to Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Privileged SCC is added to Service Account specified in the deploment yaml by Operator version 1.3.6 and above in new deployment. In case of upgrading NeuVector Operator from previous version to 1.3.6, please delete Privileged SCC before upgrading.

                   oc delete rolebinding -nneuvector system:openshift:scc:privileged

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s                    neuvector/default


          #Deploying the NeuVector Operator#


          Please refer to the instructions [here](https://github.com/neuvector/neuvector-operator/blob/master/README.md)


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Networking
        - Security
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - registry.access.redhat.com/ubi8@sha256:534c2c0efa4150ede18e3f9d7480d3b9ec2a52e62bc91cd54e08ee7336819619
        - registry.connect.redhat.com/neuvector/neuvector-operator@sha256:e2f19c50ebc4127a83c943dbcfa1312ec46dec71bd36f6d99178b320d8da970b
        - registry.connect.redhat.com/neuvector/controller@sha256:65ee3723c7c007db5ce3cd0493a785d4163d972cf957de6dec956b842807a551
        - registry.connect.redhat.com/neuvector/enforcer@sha256:72a628a5c811c95bb267bbd0a5ffc9d74dbaca03a92aae615d43c7601b4d3b43
        - registry.connect.redhat.com/neuvector/manager@sha256:89a42c9ee86f98ef56b5e695e23bdcc6334361ebd29cfb7729d7c86f04da98b0
        - registry.connect.redhat.com/neuvector/scanner@sha256:b632f1d31998a4c8bad9f97216e8aa56cd7e6446bd70d2089d6c66845b964e8d
        version: 1.5.2
      entries:
      - name: neuvector-operator.v1.5.2
        version: 1.5.2
      - name: neuvector-operator.v1.5.1
        version: 1.5.1
      - name: neuvector-operator.v1.5.0
        version: 1.5.0
      - name: neuvector-operator.v1.4.9
        version: 1.4.9
      - name: neuvector-operator.v1.4.8
        version: 1.4.8
      - name: neuvector-operator.v1.4.7
        version: 1.4.7
      - name: neuvector-operator.v1.4.6
        version: 1.4.6
      - name: neuvector-operator.v1.4.5
        version: 1.4.5
      - name: neuvector-operator.v1.4.4
        version: 1.4.4
      - name: neuvector-operator.v1.4.3
        version: 1.4.3
      - name: neuvector-operator.v1.4.2
        version: 1.4.2
      - name: neuvector-operator.v1.4.1
        version: 1.4.1
      - name: neuvector-operator.v1.4.0
        version: 1.4.0
      - name: neuvector-operator.v1.3.9
        version: 1.3.9
      - name: neuvector-operator.v1.3.8
        version: 1.3.8
      - name: neuvector-operator.v1.3.7
        version: 1.3.7
      - name: neuvector-operator.v1.3.6
        version: 1.3.6
      - name: neuvector-operator.v1.3.5
        version: 1.3.5
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-certified-operator
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: model-builder-for-vision-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-modelbuilder-for-vision.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "Modelbuilder",
                "metadata": {
                  "name": "modelbuilder-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "backup_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "metadata_storage": {
                    "storage_class": "ibmc-block-bronze",
                    "storage_size": "20G"
                  },
                  "photo_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "500G"
                  },
                  "in_memory_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              },
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "ModelbuilderAWS",
                "metadata": {
                  "name": "modelbuilder-aws-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "postgres": {
                    "storage_size": "20G",
                    "storage_class": "gp2",
                    "backup_storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
          createdAt: "2021-08-28T12:00:00Z"
          olm.skipRange: <1.0.3
          operators.operatorframework.io/builder: operator-sdk-v1.11.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/ibm-modelbuilder-for-vision/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Model Builder for Vision on AWS
            displayName: Deploy-Modelbuilder on AWS
            kind: ModelbuilderAWS
            name: modelbuildersaws.modelbuilder.com
            version: v1alpha1
          - description: Deploy Model Builder for Vision on IBM
            displayName: Deploy-Modelbuilder on IBM
            kind: Modelbuilder
            name: modelbuilders.modelbuilder.com
            version: v1alpha1
        description: IBM Services Software Model Builder for Vision is an AI training
          platform that can use the capabilities of IBM Cloud and Amazon Web Services
          and their GPUs to quickly train computer vision models for compatible mobile
          apps. IBM Services Software Inspection Workbench iOS and iPadOS app will
          be used as the exclusive UI for labeling and training. Specifically, IBM
          Model Builder for Vision will train computer vision models that can then
          be deployed for use on the IBM Services Software Inspector Portable , IBM
          Services Software Inspector Wearable and the IBM Maximo Visual Inspection
          Mobile apps.
        displayName: IBM Model Builder for Vision
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Model Builder
        - AI
        - Computer Vision
        - Inspection
        - Mobile
        - Core ML
        - TensorFlow Lite
        links:
        - name: IBM Model Builder for Vision
          url: http://ibm.biz/model_builder
        maintainers:
        - email: isshelp@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/modelbuilder-image-server@sha256:ecdb665af651417c50335e7259b89ef8096521abec431976bd7abe4c1edc49d6
        - registry.connect.redhat.com/ibm/modelbuilder-cv-training-service@sha256:c3c9080b08b87022c784fd98ac3e2774ec89ce29a2dd19d91dee039e3fdb2b36
        - registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d4c60531f59953f399c6a2b0f8239128cdc2640f72f2d1e9ac4d6add618f9f01
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.connect.redhat.com/ibm/modelbuilder-augmentor@sha256:ef5d0dde3eb334d919c2773af63007f1ef5e735ae040ef77fffe3440f854f068
        - registry.connect.redhat.com/ibm/modelbuilder-auth-service@sha256:f96b444fdcf7781775c9eba942bcc5e5ebdcf5f50820447f5eb26061c4c674cd
        - registry.redhat.io/rhel8/redis-6@sha256:2f65f46b25d84a3cb41a6a10f93786bcab47b7354ca9ac13a8e602ea833eba60
        - registry.connect.redhat.com/ibm/modelbuilder-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        version: 1.0.3
      entries:
      - name: ibm-modelbuilder-for-vision.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: model-builder-for-vision-certified
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Coroot Inc.
      provider-url: https://coroot.com
    name: coroot-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: coroot-operator.1.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "coroot.com/v1",
                "kind": "Coroot",
                "metadata": {
                  "name": "coroot",
                  "namespace": "coroot"
                },
                "spec": {
                  "enterpriseEdition": {
                    "licenseKey": "<COROOT_LICENSE_KEY> from https://coroot.com/account"
                  },
                  "ingress": {
                    "host": "coroot.example.com"
                  },
                  "storage": {
                    "size": "50Gi"
                  },
                  "clickhouse": {
                    "replicas": 2,
                    "shards": 2,
                    "storage": {
                      "size": "100Gi"
                    }
                  },
                  "prometheus": {
                    "storage": {
                      "size": "100Gi"
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring, Logging & Tracing
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/coroot/coroot-operator
          support: support@coroot.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Coroot is the Schema for the coroots API
            displayName: Coroot
            kind: Coroot
            name: coroots.coroot.com
            version: v1
        description: An open-source observability platform using eBPF to gather telemetry
          data, providing actionable insights for quick issue detection and resolution
          in Kubernetes and VM environments.
        displayName: Coroot
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - coroot
        - observability
        links:
        - name: Documentation
          url: https://docs.coroot.com
        maintainers:
        - email: support@coroot.com
          name: Coroot Inc.
        maturity: stable
        provider:
          name: Coroot Inc.
          url: https://coroot.com
        relatedImages:
        - ghcr.io/coroot/prometheus@sha256:e7117bd86e1088fc02ac12758bc94dba1029de54a5bc1ee04954569d8b79625d
        - ghcr.io/coroot/kube-state-metrics@sha256:080a85e77b1e530df56f84c8e0ee2523c1f740adea152015367b87edbfac078d
        - ghcr.io/coroot/coroot-ee@sha256:09b9177ac07c2d26f4bd16050d07bb1e2132431055c32e047563083ee09194a8
        - ghcr.io/coroot/coroot-node-agent@sha256:3d552e98ca2ff8c6e125496857306172ec6f8e134ff13886c7f799abef29d1b8
        - ghcr.io/coroot/coroot-operator@sha256:13f1a4a07877c8fe8230a0bb7282ff47451007c9cf689a995a37d16697b9268a
        - ghcr.io/coroot/coroot-cluster-agent@sha256:674bcb147354873eaa0ee9c1328ddf758d43adbbe532c9fbbe41ddae7addc114
        - ghcr.io/coroot/clickhouse@sha256:afed1054626598371975c948e5b9e4dcaac54e8059b7ad9f0b905ef1c99f2260
        version: 1.1.5
      entries:
      - name: coroot-operator.1.1.5
        version: 1.1.5
      name: stable
    defaultChannel: stable
    packageName: coroot-operator
    provider:
      name: Coroot Inc.
      url: https://coroot.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: cpd-platform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cpd-platform-operator.v6.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cpd.ibm.com/v1",
                "kind": "Ibmcpd",
                "metadata": {
                  "name": "ibmcpd-sample"
                },
                "spec": {
                  "license": {
                    "accept": false,
                    "license": "EE"
                  },
                  "storageClass": "managed-nfs-storage"
                }
              }
            ]
          capabilities: Seamless Upgrades
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.0 <6.1.3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ibmcpd
            name: ibmcpds.cpd.ibm.com
            version: v1
          required:
          - kind: CommonService
            name: commonservices.operator.ibm.com
            version: v3
        description: "IBM Cloud Pak for Data is a cloud-native solution that enables
          you to put your data to work quickly and efficiently. \nCloud Pak for Data
          includes a suite of services that support you on your journey to AI.\n#
          Introduction\nWith Cloud Pak for Data, your users can collaborate from a
          single, unified interface that supports a large number of \nservices that
          are designed to work together. You can use the Cloud Pak for Data Operator
          to quickly provision services for \nimmediate use. These are just a few
          of the services that are available in Cloud Pak for Data:     \n- Watson
          Studio\n- Watson Machine Learning\n- Watson Knowledge Catalog\n- Watson
          OpenScale\n- Watson Machine Learning Accelerator\n- Data Virtualization\n-
          Db2 Data Management Console\n- Db2 Warehouse\n- Cognos® Dashboards\n- Analytics
          Engine Powered by Apache Spark\n- IBM Match 360 with Watson\n- Data Refinery\n-
          Db2® Big SQL\n- Decision Optimization\n- SPSS® Modeler\n- Db2 Data Gate\n\n#
          Details\n\n## Prerequisites\n- Red Hat OpenShift Container Platform 4.6
          or newer installed on x86_64\n- A user with cluster administrator role\n-
          No storage is required to install the operator, but you need storage and
          resources to install Cloud Pak for Data services. For more information,
          see [Storage considerations](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=planning-storage-considerations).\n-
          See [Pre-installation](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-pre-installation-tasks)\n-
          To install the Cloud Pak for Data control plane, review the [system\nrequirements
          for IBM Cloud Pak for Data](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=planning-system-requirements).\n-
          Before you install Cloud Pak for Data services, review the [system requirements\nfor
          services](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=requirements-system-services).\n-
          For air-gapped installations, see [Installing the Cloud Pak for Data Operator
          on air-gapped clusters](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=tasks-mirroring-images-your-container-registry).\n##
          Resources Required\n- The operator requires 100m CPU and 256Mi memory.\n-
          For the minimum requirements for services, see the [system requirements
          for services](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=requirements-system-services).\n##
          Installing\n- [Installing Cloud Pak for Data Control Plane] (https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-cloud-pak-data)\n-
          [Installing Cloud Pak for Data Service Subscriptions](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=tasks-creating-operator-subscriptions)\n-
          [Installing Cloud Pak for Data Services](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=integrations-services)\n##
          Configuration\n See [Cloud Pak for Data Post Installation](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-post-installation-tasks)\n##
          Limitations\n- `x86-64`, `PPCLE`, and `x390` platforms are supported by
          this operator.\n- See [limitations](https://www.ibm.com/docs/en/search/limitations?scope=SSQNUZ_4.0)
          in Cloud Pak for Data services\n## SecurityContextConstraints Requirements\nThe
          `restricted` security context constraint (SCC) is used for the operator.\n##
          Documentation\nSee the [Cloud Pak for Data documentation](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest)
          and [license information](https://ibm.biz/BdfVHT).\n"
        displayName: Cloud Pak for Data Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cloud Pak
        - Data
        - Analytics
        - IBM
        - AI
        - Machine Learning
        - Cloud Pak for Data
        - Auto AI
        - Db2
        - Watson
        - Data Governance
        links:
        - name: Cloud Pak for Data Operator
          url: https://www.ibm.com/products/cloud-pak-for-data/deployment-model-options
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:d76476d2ecd88ea7b4fd857e591cc9b71be9d31fe0fb0057baa1b21e6113f730
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:75ca37701dea1bd079df8f8cf302cf95e8dd4ed817fe126b192ca368e92fb78c
        - registry.marketplace.redhat.com/cp/cpd/job-trigger@sha256:f2ea487072393a24e4c6864b2107c1fce072ebba07f9dbc2c139bd37fbd062da
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:ebcfc8941330da6991480dc18ac34d6df4fe879e562ee537e5c36df73f4bc536
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:ac128758328325ae56273800ea010f9559902281cc60226c8f99de51b636b180
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:21c81ff7c4b8809accaf62b242fd9b3d52f235489ae427556b7f71cc21bf8926
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:562281dbb2bec5191a7c6b595ab853092e6fe3ff1432e77b2471f71ccdffcbc8
        - registry.marketplace.redhat.com/cpopen/ds-operator-bundle@sha256:be2ddb821274938e5ab926627bd998279b6d19b7789d9683513e85a4b05bc2d8
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:e673f1d0c5a0e32e9abc7ac40eff905ce411f9f3d8ca0b03160e23ba5661e041
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:8d7625a0ba656425f601b6ec9a398a6d30511fe1de119cf0e29f680c704ea225
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:e76b61f9f0efe16facbddd80808d80bc32d0d41b4c16aed566116f1366152987
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:a7c050ae29cb93754f59d8f83676275ee9a91ae9a9956395b97abf8c5e20e96e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-bundle@sha256:842f2ac568db4db79cf0450b413b4b09b5f2180ef3541385b46fb650ac7066c0
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:65d327900e3b7a874dfe4bfbd0f2c7a1ee402346aba0e8cbc628b406e5b5329b
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:ec137115210dae9b69cb2899ed8a5437d94c8a658ae1ae5eb295232b34b0925e
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:17f016c11039c34f90f67279b45ee488740cc3d033112a3bd589bb1a60ce6812
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:582ced3f60212f736f720558524225b0d4a1c79ad4890656c47dd58009ffe0de
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:4794a0fdae9d060def1e0d75a98560e713ca18de2625e4d9e5b5fa4c4877fac2
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:99d3b5a88d20380bcb4161e3b182f305af2e4cb405460d0f2a91965dd6f99f80
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:bdbbd76c93f0fb8f838edd9f124b02cc646a6f735a68a430a00f025a06c5f407
        - registry.marketplace.redhat.com/cp/cpd/environment-specifications-sync@sha256:894b4624d8ead7286cf9068c6477300992c32d96339644fe764bcb29b1c8ef5c
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:b7468b824d519e46e3ea23f5a803a8ace030b74d71d4cd282819027665ff0e12
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:10a935e41cc2dc09f8f3939cd512d9f6bc71d68457f77793a6957c5420cf4bf5
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:973eda4612c905661ff5ebc28c7cac6d9fe16fe5edaa544548d2311496ed938e
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:b2ce782efc5349a49d9c7326e2f72b2c1d99d30eccfb6975f7cba10008f9b2d3
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:62dda55cc870279d491679f6b43be177986991b6eb85d9dfeba3ed2e72b3f10e
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-hook-utils@sha256:ec44be69b4c585e9a3d507e29bd287f5b464360d82c8e0969c5fb4ed3e95abc0
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:0b640bda3afa672011d5da390af798bbd3e6dd9126029a9eef03846d41582743
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:31470f873a0931d130484d5255692d445e3127398aae0e777a3c3cb702917743
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:cb9a33db3707c0c6045e6c823ecb21b0f229c17ff868da7d36602d0026eff351
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:697baab21258166a1897ce4e056e39bc4667ff837bb97cadf92dad5863242ceb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:716b5823661ac8f1eacda11ffbe5f0ff08a0e9d2296d04acffb646b7526128a9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:5d889be3dff96d1e888eff94bba94228aa46f4e94e5f30ba33077ef00692509b
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9186c350fa54e8b792d72e47be6b54a67b610f726175c9641795010d45ad51f9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:a99b1f5a4eb0306b83562de0184fc826c34835ffbf46d929ceb9e70e0d325361
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:da0190fee910cb8245cf7a61d6e7aa2bf599d41c679eeedc183950918ea2611f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:525924934f74b28783dacf630a6401ccd4c88d253aafa2d312b7993d8b9a14c1
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:838f7649df5a7c13d6155394e211ac565aa86233bd30a3a2aab9706ec4a74eec
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:7dbff355db7739152961cb6a97887d2e43bc960ac58837c126e03ed1a4480a3a
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:0ceaaab514f65f9a1c40717e97ca652506f8b3595186222d470a99d653afcead
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:5aecbb4de6fca34f026de72b7480f35dbb2ee8d8ef1f22e524a85a71bea99a1d
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:d904761e70d3969a4ccacd71a60ad4d77dda939c556615df8d65782ccc3273aa
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:c411456aa41fe705b47dec60eb49c6d65d4b7bd1a64d1e0545817efc95afc67f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:474187fe2837da4b86ed9482a96fbca0a7f83db149859f8ef23db84de9a6d7fb
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:4a5ed208b68613d5d88936687f6912b3aad0e93c75f0d311120a8e55200f5795
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:b797e00eb32ff9a6a56e71d355bea653ecdecfecc121ddb197a6e347b89dbf8c
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:fa01c7d11c23a47bb84375389e69619455c2879b5ed6bb5b9b753b945dcb02c3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:4f0cd4920e7cef64e9f49ae1abc5d87303102fc96819e0c3b699d86b8f8a3980
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:37853d2979d0ca5f88fed76fed5194b628507ef0480e668dfd6eb034fa1d2023
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:cdfb097f2438b5739bb88905b8318d15d15c4c64d37cfa6fdb06f44dfe976e6d
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:ff5a5be492347e75b309194661d7742d1c909c3765c9947c8b41f4c3af62b08c
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:c730fd4a7baa3f46c03f6c73e2c6c7c0ff24a8900b79dfc0f2fd319909edce74
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:aba8f1d08d143a8901b4d2cd8790e067fc7b9eca9987696a1ec6f9f505a686ba
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter@sha256:be5fd8a8b43ae3792c58de1342cca91836287fb584bab003f7f9aeeeaca3f65f
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:4e65e96603d5043e78a517a4c6e5debc5df7e3747fad4d8025ef1fdf1cd6a9da
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d51033417b786dd2dfbc94e648db00ba68f316c23b06a0bc9e36a5c22f579124
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:f662be48eeb6dc93a8a4b0c3374f59bc342b5b8f5360437978cbac8f74c3a084
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:6a996123e69c7f2c3254f334e278ee0daf75fc53980183e1a72358323e41349c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:2698150edf65a99a1d24bf3e8336f02b57fd964fcad283e182794634b033af89
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:427e431d5b8b56011a91bc4f1e1586cd5490944615011fa6ac20abc8f45f92a7
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:e8eee5083cbe317b368c002f47ce09f98e0c5e3407e5a4712e1c5473c4090ba3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:53d1fba68aa0227c2f4c8de26837f6c114c56ca4a397826e1f2e93bb857e5b7a
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:f00158405adec00c7e992efa7d7a19a520df485cbbcf39972e579ed3c6e4d289
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:08dcf99453b4dc6dfb5620410b8f7e4eeb42fa824aedb2e480e2fb40ec8ce347
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:257bf301338f7b123099e2ffe31c3f3b793bf28d6a0570a25b889271239579ba
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:b9c8a65c1742c72c6351b1c8f4a858940c2bfa00ac5ec271c57b2ee7e5e7a66b
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:411a4c7164b59ab3eed86e67aa457e45bd028067c937d35d0a227de63870cf87
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-api@sha256:dd9e4909317512a6fada5ecfc7cdb47c6807f266fa456a01b60d7b3486faeeb7
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90d45b7a3c5ffcd30a1ca1447693ba4aedfb400983cf49779a8421a42dc0eecd
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:75595bf2753fb43391e2a496fb6d59faa8785babebb4aeb1da34cd5e72c93699
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:719b0871f0d32da894ec80dc2a23483623b7802a5c6f935cf3672468f99faa40
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:ad027c91c719fd1d931cbac8c76735474f676b228f9ab23d1f243201246b3238
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:4ccd786884e39641a74096d783742c219fb277caafe80b667a80396c3a9cf6c3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:36f4c57c6594f997383c4731e1e62c4765b8642c0b5c7373ba22118220928d09
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:bcc579eac4be59609102836252ded82a17ee0088e5ce28c684fff41c80232a0e
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:fada47a5f4824d3cd5bdc262b92018ca0e697e18dd373f92fd93186fb8719a1e
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:9b91aaeacb196806646a1f196d4adc83b5159de767fad1a2eab699f735938911
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:6df671937bc55c410a9df13aba15687b2832c7c19c0eff4686c3f83648528068
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-bundle@sha256:cd92fd1c85a72560241fffcb3bb62cf965f96bf2a4e1f5cca84c0b52624ae962
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:f5248b19c4a9b605a9160ecdb5dbbe0906841b667ede6fd6cf8e1d6c4369fc8e
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:690e9b5218a200de93a7c8477fd2980d8d341f22be75002967201db09b5a5524
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:978cf4bf299cd3e8cde9d7ee7aa75b73fd9d9e6755567651d6c1c49036b8978b
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:31edab3e5662489dc2fe9528d078966663c8d49836425e648b7e112f871808dc
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:ca4da5a8a3c1e7add6ac630c3fe4713a08739f6b891b2290bb17026b94c31317
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:d6cdb71edd7b7451a206d1620b35c321f88b3ba8028468329595cd538a45abf6
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:8a6ce0dada706056efda73ad0cc1547566003e0a7aab8ed379b50e4776861b94
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:483bc8f74d02efcbfb40ceb6abb0ab0b6581523cdea576b34512709b98879233
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:26500e941ccdddc08dea07d3a4b9aa9eae7840fd5b683191320060f6dfc88c97
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:24dff64a80b910f0b4c5442d4f611f9df7602a16c5231eff2c9d08f07b2a9f9a
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:d733c5a10a21cac2df2fff6378035151e5d0df57e3ed1f96bb3ac17c0c95b8eb
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:f6e8de9f11696ca22302b8576ffd5bc2c024df0c8dc3734c404cc799565f351a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:e8e0ea5cb1a070443011794988f2f6c9d67758605dc2847bf0183e19bfde52eb
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:06403c4ad99655cfc10fde46b21e84f90c05ae143bb6d08464dbc8954486897a
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:7ca8d085dd9aeb568eec2c4da974c03aa72bbfe4648f5d66a62d0444a2d1c839
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:a64a93104dee64e0e7e3e381144340511f903155f72b7f619c82c1d440712227
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:245f81b08980bfd4790cc3661e609e2cf53f3dd4d4c1f988e5db4dcf316d90fb
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:cab91a231068a6bbef03ee8db32fadf11654317c4573c6714721a15d0e2a98dd
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:849ecfeaa36fe365c7dd8f1a76fa0d263a1196fd16af805dc66035643e8e9d9e
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:17f7039e49f28abb7239e0c4f3689de58c75f34c9bde9e7b33a6ffb93ba92677
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:ec5e4a002cfc63c1ff1bec340111e8a1755c8624066445db36c23fdd1c3061f0
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:6e0a23c3434046195428792c45c95869d81fb4ac4ba75444e4343963fda6014b
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:fbc6eab07fcfc2d0c1895f18e17f751e9481ad111875959e9ff08e14f81b9a35
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:96d382b1b26a1e1e4074b7a1aa48b973cd4865bb5c3cf02795306e92c752a8b4
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:f40a9375fe5722972f995a82b50c00f254e3fe37f202b0aa1bebf66b0504099f
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:164f6063f3a50a36e28a240ae7c060c4f90266b7ed799bfdc075f90628bea7ff
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:0fb78efb9e92dd0c2e52ca2488e6f1e7431e197f87e6994e15f7d621b90319f2
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:c078bd65fb18b31ecfe03c832643bd6fd85f7330b94a1ca26f557d32d89903c1
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:dd76a8484628970b62238334e122768aeea820ab18e24d62d23a6328c83b868c
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:2472053d912f693a5fcdd00b7d9f0a7ce98697c47eae955e31c0b602fd8a4a37
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:9c24d02b1cbb3857d017453744bbb2e68fbe133210c35e7da5ce0341921d5b08
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:ab686e9e4c38b7909994f9fb496943b745e1a07492fe61a61b9e8a78f756eb1c
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:abd96661823553d623eb0ea1fac8271d64391243dd0f13eb4a4fcc858dfe10a5
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:d4738f64e36a8f0c88d40043eebaae1d561193895d438fb6d8bf1a042d0a1007
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:ee15594f82d374da81dab59c559567af74153524a6d9bfeb28ee085bd4109b37
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:866ebafb730e660062cf7401023d82c5242a2b80cca0126d47a68193c84670e0
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:e857323acd10f7b99c5b7fb4b0a5b89a808f68036a1d9a224eff55a7ae7744b6
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:2011f32eb1e68ed5b75a4e8a99c992eac546995e219701d17a511596a5a09361
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-velero-plugin@sha256:2dd02d8f15dca1dce7cecad098cd8791cf1cbeddd2abd8e8c6d8f066d85f9e2e
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:c330a321438943fdd264b292b68fdfb8be89c3bde93a7bd987bd92c9dd2512b0
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:5f086cdc7a272384c8a9b680f8f90fd471799e9e96175b40c7f16126b73169e7
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:c2aa3566f8fce992f00f43867173e5f47593e8618357369b23044fdf3c67ba53
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:78206c6634f2be2bd36806eb717bb8e4db2c755263cefa6d026a425a19329324
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:65108559a5edc9a7c4f82962ab2293017e0873872730d60d8487963970761941
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-bundle@sha256:94c7a1dd290e2abd55c5b8998cf08ad174d3fbfc659ae6fa6fb6cbab61d5fe14
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:eb157bfe50005f9ffd3560ce1ef7ee0fc283ddfe265c42b35a71f18409e3f561
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:3112fa03dc9a530642d329666a0cc64c06a04308beb1ddca0f7472915719ef5d
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:b341d9af40ad6aceb4ecc792ec75295313128a1eea3a0974aac879aed24a89ba
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:73c40c6c72b7f2399fb0ed7e462baa72c3e8b5aa246fd2fe1082904b4afc0796
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4dd54195a3283a820d260eaeaf83463722e49bee05d875333c1b99201a7e2b85
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:9021a04805bb7f21d68fc6211f844a21c1d06f700c332ec3238073065b7c224b
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:a3dc5fcff6a2aeee7c7e562fe9f6e0cf0652fd7ad4ec61496021d366206e2544
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-bundle@sha256:552fc53664722903eccaa7d3782a3e84413509fd357725fe0b0f4a077c48ab2c
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:b5a8fb14c3bb7c89ff970ca85f06dfcdca832e2b45ad86e8bb1b6c0a0e7e0764
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:e4888fadfadb3462e3a9ad0395a526a2631872f70caf291dc18b60269ded199f
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:deb66a64288a370c5da5c7cc0ba4e2ba4ab816d7c5a080a0c505b05f55fa6072
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:96e6db93b2d16ba590252d6568b68f254548aaed73ec17e4bdd8d530ea668be6
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:2f45aabc99eb9646db4c74686569a7f2af3f27cc44362a30c92d171540bb274c
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:f3f8c6c395ae860fed5761f67ce092a122e7073edbc205c9167c27a4ba85094f
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:de7094acd65a80daa67669d8eab809220dc715b9cd6aaa96c58e5fcbe34c460a
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:a6e98064effd12ffda0613db98e795e056b83ef77b5deee932a762b178ceec3f
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:aadebf0598c492639e7c1109ce681256731215b686fec808578bb09558a40315
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:39700bf47b93c7f0db1894148c16c60eaf406196034351f54573f0a94c4c8aac
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:209ea4b77b3b0863c591bcdec6f53c4c39dafcb10ff13e27ae4e1cb986a59727
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:1f68dadaf2ab779dece300037f3abda1b0bb33603436b84336f8fb7d53d33644
        - registry.marketplace.redhat.com/cp/cpd/spaces-ui-job@sha256:95dade51444e30ec034c12207ecbef675b70889b81970f49f10aa75160b2f947
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:b82a10ce8f75d4441e869bf62fbfb5da349a2f228b52178bb29723af62f13938
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:4b6bc5640d9e68519a7f543ddbe3f16ca200f32a74ce9c07e996cf122da3b59f
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a24f4cb635e8b5759a5e2bda20ddefc6d5ea9b16fb1081592885c54aefb5480c
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:847751338f3b78cc06e29a4a26a18dc0f5d0bc24cbda23b765198a6a5c4503c1
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:ded16bf3cc993bfa5af8907a1517cc34ecb511948e4d9ac73bab68888af5543b
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:7de2c2f7cd478baba2011461c8d28c1a4c70aa485404d4523d1166dcb94c0a79
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:c957787ae587a31e6a497b46649da9cd46102a2a3dff29243fd5cbac96e7e2d5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:e3d8d4d5285d058130345d6108ed4056ea9c871ed8c0575601a94629e86ec691
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:0dc719e58694fb501b8b36f3a375b3c0462624d27fe49afe6b835821e31393f3
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:46bbfdc1f5e418db95ba806228fa5696be22c4cf9eefc185ce27bfa68ec1d247
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:cf2b88be76a4aad436f45a79694f32d2e2175587c3630543a6d24d7725b60c5a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:6bbcb64cf9adb4d3b6a4c259eb910e7da5c3c58732643e60d416399b2083cc0d
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:b8c83fe5215b7412765181aaba84162b6cb188b5765e45972a6f29b2a9f19826
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:d28e6a3c65b580051ffbe287492dfad28dbb1c669cb76aa3df23bd9428f4bbe9
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:b372540b630507f4cb74f5eab38d0335d7b5bd64197e3f83852b2c7c5ba9067a
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-oadp@sha256:9aec2bab8c7a6941e97c18352cb9e838a99aec08b06269e8053ac764638a90b9
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:670a36f3400164cc9eec5d23422d184340e0c1edfb13a634e58f0345c325c21b
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:a0273c013c5943c9b9a223cf1aafbc96b3cdd2dbb147c73d2a9ed49befdbdfc2
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:6caf5203489492755e5cf6bffcba53be0bd32662e8cd331dada68a85b2c083ca
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:ab99b62d40ba05cef8e699ebc7cc65271ea6fcb5900a76fe17526e7d55b50b08
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:6bec687b2bd5c4d04403a25f56d2245f822e914907771f26c25166462976258e
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:63d4a9fa16231cc0c01f1b8bb9363221ac1e93f1571ea132866f170b156731e8
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:a3a0d58f895fca4d832e602d0dee80018df4ad7bad8800ad4058eea07fda4817
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e12599ff99db058fdc8f9c1eebfba31160c636ff092138deb1f6f6705f44fd62
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:9cd2ba7ab3ecab7094c892086a5276bbde22ff96d26ce0d087df2c1791235b02
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:45194b56f9fcb3af83158eb0297248364db40e8164fb4fe4f5656c58aeac3fa5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:55d4c6a13452f76bbab485363288efcace23975623b8a4b6b83f3c549f844c58
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:2c39ff1850b7a9d32a11e99393039ef74ab363f3fffec1e4fd542f2ae148a6b4
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:6d903fd4bd0ef3ef361d22b209ba7b686601b49d6f8d4e29b5d030da8d949dd5
        - registry.marketplace.redhat.com/cpopen/cpfs/cpdbr-hook-utils@sha256:63a5606df81013b8a7b5cb07bd1f29bf032c8998db843885184fe86a76fb93bb
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:ad19b65042fb26cc1244a92f81a96e5ceb4313f8c823181efe28b66e9fecf9c8
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-oadp@sha256:35594a16332ba3ac00226414be9b85243bb25dd0648e256d7ef9d42137dcf914
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:83038bb7e3f4b4637f0d5095b42dc906e7fc27bbffcbd66f6ad93dc9b1cd6bbb
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:130872119947fe91cbfb8dd82b3baaf78f7adcaca5933ca394eca3f9ee84c8cc
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:c268d2852580286e5c6770857fcfc3b06a698521f5acd17ed83b8e2cb9ab55cf
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:786b21db2a9ce6cb0eabd5cb7c93dc41a8b1db6c1f2d07725a5bdd75b3a1b7d5
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e924abf025141f3b3437ed89830957c5915152d2610213090e33fde988bd879e
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:1019e13e6f072c07da6d31277c8ac639e8ff2200599ece1666a3e34745e763d4
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:be5679938413dfb7766f765fba7923d3079d2d9b6f464485a8309e8e306f19e6
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:bc39ff644108a458d1b7e29fcd88bf53982ab4070ce2ea0e75b0908678b9205d
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-bundle@sha256:94662b9dba5d4bafeec53b8ad334e7a1235385b21ac0952e12975f4f0a7623a3
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:bbb457d4f6b7bab065e909756c2dae31af7162dce10bfb92f274aedbc98d0762
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:d06ef794046dac2c58163c201f55489b69081dd15dd91cab16387510a50e24b2
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:21ce62d80df2575c3aa7c8dfde706cf8af01b2fc86f72ea68f767bc1770fe733
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:66c11636fdf68631e37bda7467536093b9402e883999e7989dbdcc4b4d2ba7d7
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:0cc0283b768c2968849f24cbf7178213926e084da5d0c90c18d38e2260e19138
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:4fb2127ba7c525a959bbbcf024de364ef266ebb3ca7e696f74d990d2552cda59
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:113fdfb9395c96f7eb65f9a20e18ede8a2eae4aa6c91660be8fd3e517656e156
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:7cec55c00827b747f23c2f1703d99e1ca5b30a0c445f8ad0471b25b0c9cca57c
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:3be3ee728cbae6bb8cf29d47c37ed1a4cd5791b54e1262135089424e5bda7ea1
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:351358f551dfcf817c5936c0b43c2e28db61adcde65d7d2c81cae666630426b8
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:1cf983f9b9cda18257066fc82e199bd3e22e2fcba13daf0f60c92b50a4a0e078
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:772f98763527c5452c662b2ae516de59c1f6edcf339da46231f4be8aab6c580d
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:ef03bb6ed647d92b9c84d34810376da7049aad88589eb2e35798d82a22f5b994
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:f7e68137e1ca58a63924967a8e4c872912557fa68af8a7392e3e0d4aaaef1e78
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:c43770715b76e0aa3fd9b3c2b21998ee369f2142e3c5d04320c677c9e9b29add
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:e29885c1ee083a73b952b0de8d1cd675b2b2830abf44d69a8577f56972a601fb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:ea57c7cb1ccd5bce17f6cb35e326ab974718be8cd19b0cd38ea8112e91c51939
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7d9b2d61a4eaac645490d1c9e6db97ba193d65015b6efa8d5a1f3ec6cccb1ad1
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:4f0165926685f49cde22a93387fea401bc1799ee846041fe54bca148dfc2f62f
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:4deb399f28d3385c50a256c6e927cf27115c311be5943021733e4b281c12a419
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:1279b116fbcebb1c052eb27a72c67011b44a12fe673a86c1ece28202afd911d4
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:aea8fcb72664a41c3881839086a47fef1b54dd672300d4aaf583fd0b575a64b4
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:a6d5bd253959be1f91b50f5f2c56ed77e1c82b06483e62094aa86ff57cd46eec
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:c2b9d370b6c3e3c8e30079465502ce376055261cbf0f95c48d8a2ae10ca36c76
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:a31e0fb7a05e969e31c10a74cc858c483d5d4da14ce3abc36f5318100e60e792
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:c59837981b9a2a645c9ffd84c40d4f9ddc3797911f7a785ce4e3a2bfdd0bbe5c
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:4e3ec61a34fc7c71d213999b68b1f5c3f866f97e19c12fb001d506d4166a50f1
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:2f40d8188dc8b41fb072256ce8d109167095469af416b322ac023071814cda8f
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:5436a25b6c5b419b77643a76eb073bbbf8f09517299dbf4a992ae14c1965972c
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:51e1275f6e8297418997b0ca9615642153371c29b3466f155e15d7fb9e80cda9
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:0a32e127dd305fc445808d0469147db6c9704cd15d924d57d791e035475ecb60
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:07c13b18197ed0bdd8b97d65312a342f95d7a7fc43d50af746d788032a67f83a
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:56315ce0a0a70f5899191a7eab856105489143f0e889c8cdd1b6b8c225f02eb7
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:12eece1fc0a2e14244a2f57dcf7c198a792ced9ad0d02c9606645bce1f295877
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:369f66883e040accb73c3deae984ed479e90eed242051b7035723f0ba21b51ac
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:60dbe69693314c937a69bd213de438c3af0a8a290aa9b1cfaad8a57e9d04c448
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:54d8ceaaa06196e1ff524b7be595cbe90811f3be1a57c767b1ba74f53edb2f0e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:e6ce3c998642b235080daa6c0b954733bd3fe973fc8d85a10ef0cdae2a73aeb0
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:5f0974565de8b045d3faf6acac7c8ebce6cd35327dd76290b39fe48b27ad5b62
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:740a1d87a43e61e32cfdea032d852af33100d959c8b57ff2f165f4baf77fc011
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:b368c22956707b8e594441c0e9bac20981339a5713d4fdfbe3c64bbfdf677b35
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:52ce08ac2bce7c02dc30e2c0aa74df4bde8f156debccc59960f02b4b0040ffea
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:67515618d631b0ec00ad95ecb9bcb48d7f199855af9a64e86c4b46e7ae3be69e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:3ddc6bbed1ddae1be71043d177cbcb58830d0dad542b06567fcfa254dd11e626
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:edfb732843a53254ebf7d4a2f109af109623be8fb9bbae94366d30febd2d0494
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:631a18f4b437e5092b9671fb4c86af05e99effe5540c4d6f964dc626466764a3
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:2712755961a1c1e23f30d49341c037858c0965dbfc065aebda88c84526291f21
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:62e2b059f565f87beded59b36962c457c9662d1e401a78766fda328ea1aeb78f
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:119be0bf4c60230c43bd64c0ea252108a0173adfffc9f05e5257ec260b335599
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:2565a8ba9aa103956d6902f2bcb53ac13328b647a2d7a150bf80ea2749d5b972
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:13a9ea6ee24d9b0ac1d5fb09c3e42eff731779867e7e71ce40429b43ba2779f9
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:70d23c45cc636af2148fadc1c7a08802692fed14e1a31893214b8d2d82f06e1e
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:48828f128c98e73f8e345d87bc228f0920a3ff2cdc325d4a4ffe73ad95c5c5e0
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:8f9d3438f58604cf29d643691b125a4138275b2c09cca4afc8e4b29289febd77
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:6298c1561c066eae3d420f20212a3c7a8f9069a6996408685437588bf8335dec
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:6ccbb20c9fc7ce00ddfbac19c8808417957481473eeb1ee62ecda165e0e275cc
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:0621eb9b83c9b2b8bfe5cd09797fe08fa37a854a5506914935ff6f0df599b631
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:c41e4d3314fac0d2cf5f08e7db85801cb2098d37d0aa8104a347e9e82526e914
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ee161954766211651d7cb13f3f5a3c025ac04094ba742be453ce789c6f0e24a
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:777e090e0394ffdf39abdde285b5b2749d6835e61458f117dbdac108186c34c8
        - registry.marketplace.redhat.com/cp/cpd/task-credentials@sha256:e162828d5ad07ceae4db0589f930979842714e3b766552102bee373a3a570af3
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:7a0e784b0722ab0a4c62754a4e817a1f25e5a2d5e960de76387971fa1ff7da41
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:8305f26c2ebc493b54abfea0757d65b679f6c96bc3a0840c89f08a034327ed9c
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:27195e9758afdf8521971866fd60ea3dafed91a146c3c4165876e5ea476a4622
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:f42395ea365edc2c1bbef914a55b091e6442d17bd87bf3730a8f9f6d401bf4f5
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-operator-bundle@sha256:086eaeaaaf440a1d4a254984afacd8d04583aafb9ec11ac072feb42a87b986f7
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:276d5d06661ed401f0b0672dbcfa3d9f07fa2cde662fbb532fcdffbe3e332227
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:0e563d51d601adbd710833cbde26ab8f857092f85686213adfd6fb3fe5e2828f
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:4a87dbcb7d2ca2d121512fbf57b645eaa9ab745439b79722f474b84efa9eb391
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:8d744af7dfeb9892378b7cc67d0e85e29d12b6f79c8b6b2a53f7a53aeffe8231
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:fb15cbade10fb37b14afe063a93decefd9a40583895b7b6b436f36f08bd73aee
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:650e68c41252300836740ad495ad6776a7da94331e8070520e60a1c5e2c5f3ca
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:bd237f42078822b038c13271f35e6710273c2a4f3fe2f3eb74429c93cd2a0e41
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:b63e2c10478735b51b6828bb11709569a2fce62a19b70157a8a5b1964820164d
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:7ac32270044a9ec08ddaa8dee62b61618d9363c8773fe44ba1dd86e1b53acbd5
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:a2949c3868c2c91bf35e50c64b51fcc24372986744b55520e21e9ce31fde59f0
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-bundle@sha256:98cf2a6c7b7af5d7cbfbc61ce140201272b30b12edabaec2b2d4287b0690d89e
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d4dff1a8968c18d984e16c4c9f1517c3a10e23a11bf166c64d734773324d4c9d
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:d72cef9b19ea48f03f7658dd55cabe6e118be0209852025845d426b1fbf0e0ca
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:20ceaa70a3ab5ed8a2e451a1e2960d04c201f0087fba1679fb80c57c6be6117a
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:81d8b25221d827160dd5e88efd383453f532057ad02d7a2ee3dcd9c6eb16ba65
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:dadfa9ef372b8d165276fc5f1ee9c63c3f01367d156700810d63e1ba19f3da79
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:25420a21686a32c62e83b99fcb215df015dd13fec10f73f397f927c61226f964
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:a28b71ef64c1967cdd36a16e01b611fff6a387b6a741dbcaf7556ed6be9b1f69
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:8d9e35c3f29f8f56d55d81a2e6cc4d1b0cc352ddc8d33d83dced83c4a111a0cb
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:00cc7d072ea6b6f5e9f3e5b95b9cb4ab8ad7bd942d8c259e70f8ba0e9c09a464
        - registry.marketplace.redhat.com/cp/cpd/jobs-api@sha256:173b88f90078e1817686ac949f48387af70a706e23e6fdab4a7c6217db5ad86e
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:b87fa8374976ccd3dad90ddbf6b65cd6ae5a028ba896562f284f089380ebe44a
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:489d7f80d83c9890600e3c9c4cb9442a6f42b9c6c4f13dafdf0f614b72e67143
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:8cb37bb0f0461efc603cab89b5420c56990aa104fe39c9b77f0393bc6b641172
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:23ba1b277cf76bc28f952ce6f5bb02d6177f76eca8a8b4ff7925ef119a994a85
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:2e70036d7299168331c94af70a22e3b2063c56160f616362b1a2187c12d003a4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:fd35c33c7361bf04a66e69112b05a078ad65fdbe8f8ca8b7f4ecc2143b1643e2
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:9f835bb2984329db9db7eb2ac5521371ce0ced216a7875e45ef3ec31d255b648
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0a773245bec514cb893f730086f39a3789d0204b82cd11f5d11a5c270b92f70f
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:894d67753dd0c5b8625369e53635a1ff53066d3910b73d91e5625dd8907bff43
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:9ad526bd56829ef24834b8a790b29c50809866c1dfb60cf8e0600fb9a903d388
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:ad892ce5e6a3f69ef484eed68e92d9c972597a86420e22ff80c247689ba02905
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:00f8bcee4ba7e0c5cf5f0c52e9e7707df367041c9f4dd5508dbefe0dbdba8877
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:f9c9202e106d5a159c8f6b8fd883da4d6ac0d81e75b558fa52e0f7776269404c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ed81ba4ab2caa7ffd1cb90e11aa8915df1c5bc1b76aeb4c87532c84097bb102
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:b87b6c32a2050f1d5ff311d163a7b45efa7d5de1d352984b2f70fa578301a952
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:e296fa15dc14ee827d5046e8b1434d7f6838059c41d355471e1275e09f087411
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:9d4ae40d34e38a48ad2ad6556e997a42f6308006d5f6b8638cb7bc4ce5b1e40a
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:7ae4dac24ea054ce527a507dea6c77a3233aca0047672ba5a367fd5742448517
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:61ba2b05a6826f618853ba048496e2c9558e509a01648c74b50cce6d28d0d764
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:811d1cbda1d0287dab48cff90c7091dd1bc8aea385f562c896f2b09f075cd1f6
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:46893c166cabdd8b3422a190e220ee2e34861e797aa068100ccf53da349e3fe7
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:54c86367bfe0ab658611856cc0a6c5de24c14a336bd129ac9f16293945a55238
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8ccecc00f3a9656a3dd1b07af76ef065e301e4101bda4ae8dc5bb91097cad7ad
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:d34d561bbf8afbfba9362d2a9a58f2919faed1502abf72e11882dd9cd4fc9711
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:3d5c6573faf4ae00cbb21cbfe9b2ad5cd8ffb658df0c5ba09bd98b3bf54f9979
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:932825fc7af8b96c381d504cd0976c751661ce078257430e5c79d565f416bacb
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:47a3745c78280f98f2ee352bb9192cdcd2fcd6d4c1a3627e7aa42178a0f107fd
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:787e01e29abcbb3d1fe58e0b6051067b309085459983f38d6bd18d3811d608d9
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:5ec262d1ada13a7d5d519ad98cf82dfbbd915664bb7560408558001ea6bb8451
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:9a537eae7e581322e2f0764ed38a797ba7cd260724a285f24d5095058b7b47a0
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:523dcc671f1c05e690d5e578a90300d28d76570f605d4de7f80645b590f4a462
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:351375de35d7fc9d940b905b867ce8a00b7ff1416e70d3a100918797fe3d8c56
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:fb850a3d660729fe8281967c2cb466fbad3f01fdc302db17988c27682217ad29
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:ab31869bdab5180b816a89c45008328cb7013c26b09c531bdbe5b00382f0dd7d
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:095aae63ced410e033d8f7f3d27bd14dc5b93a10068ef5a389f181115c5d07c2
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:8b7941a3f518e952766a2e6f88acfa7c9914b6d2bcd4ad9ddd54df313fa30347
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:ad2215adfb8dd2dc600bf5097b7f88c257d5fbaf541409c01b36cdb67c130236
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:7a90c8e0180da12172b23f3b54536bd1afe35affe37d908ddd78996a81435751
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:68a98f52b7eba7e41c78aa91602c4e88d21d4a2a2a11b9b7592f181ce30e41ea
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:268c763fec7defe810624bbd31c96407fa37fef47b904758bce4c58d95aac043
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:b10ca594f31751faad4050b4f7717e0e98c7977fb7e26d4866e1c14125422225
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:d7ca81372914a28c8298de5f9d3c35cdbd6d8d8582ac9b8fda5a85d22e3c0def
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:8542b9e449b054e455abf6e7884dcfda708a6213015e932844568d1853b65eb2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:2346f67ffdc1a802720b325e010018cdb908aa458f372fe50ac6d8a92d8fb733
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:9c2093089ca0651bcabc1d01ee2b69d2fb0cf2a277a8b966cec9533950fdcdae
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator-bundle@sha256:c743dd7de2edbd08c6c73e20cf15f3b147491ad77a81677339e88caffb93e84b
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:2f2590734343c99183dba7aa9fa4debc73377ee707f9bb090680750e62acdd68
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:ce57bc323b95db548dcd2f90ec938e708988cf6103579e793ee39d36359d60cf
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:e0feb4d305c5ec9501b30ccdf8f5826cd8eb1ba5ba1d62d4d685101c785a708f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:07c5e4b3dc18b921339a7a5c78b8fe9179205678eb662377e8124e46479d7f25
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9669ed5ba26ef32ad36a6784b2658553c00ac2a89cf4e9caddadf7bf46c1592e
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:19bb32e0389361e8314e627bcad2e9bd93984289485cf632466834353d9c3449
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:66c1d2493e6721094e85a0c65011523004f9589f58b7f68cc6f33c5e5adb6984
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:85669f2a35ecc4ebf6b46fd793ba6a6bdf996164c00662e65cd7ceece8d60cda
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:aed24dcc7ec24707c978a17b86501b65c6f0a77bd3ea0d173c28684b9df9cd91
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:5f63e062489b8e1bf49b5943e7a836174fb47af8dbbc8092ace5c8b72baabb99
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:be55d9ad6ae76bd061046af448d032e65926abaeef95c831c978e01e2b0355c6
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:9a970756cd76ec748b555d5bed97889b8c9f410db6765eedee84fad4976cd523
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:3af0967b0fb0a28a83a3e007b2c78fe27ad56a5e9c153806c6a7b17d7d9b1df5
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:5e5176376a664755fa776efe623a1e11bc5d3c31a3d0af980b165e54cf06620d
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:7693312b2bcee7b8502de0da8786be985607fbfe6616bc9b0da4af8a6fdedea8
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:b2ca6bc1119571cce732f1ce2ccb9ce406b298550b55144f950bb46cf985df3a
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:6fe90f051d9261ad224dccf4d9f6bc9f19ce0e8d22fa010a46b2752374bce2b3
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:132670412a0ce6dd010869e569504d2e89aee0983ff221da371c9e30d5f8d8a9
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:9fa6eea66a3daff39b7c0c7839ef5349b483b29dc219e1db384aef6c4c31ecb4
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:993f16d744a81c95e73339d3c8d5880b5cdbcf29beb914fcb2c6403af6d628a1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:ca175109562cd9140d98bea3be88d6b34490ccb8d79f0cbbde08c821db9c5fe9
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:c8786cffb4a27decb5dff3d2a9682d07d11b7096a689bda0d7cf105aa3156299
        - registry.marketplace.redhat.com/cp/cpd/ngp-projects-api@sha256:df420e9f616ad8a123d9eedd117fb6cb007b30a772138db534f47ad8801fa96d
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:57e3433f860ec2b65f5c5f86574614fa9273c09b6b37db96a9e1888f4f6932f4
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:7861185d55a5fd18d896dbc57501ae5cbc1de0bc02f1ae3fe558a36f0f80b4fc
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-velero-plugin@sha256:fe562192e86558a17d1c88a8fe21148bc000fa93c8928f7a3a897acce62e6798
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:a72c2b70a9f0a93047d47247cddc1616138cbc87337289ddb208565af6e4369d
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:a16b822b41f5a745b475ceb353a459e998f9f43bb9bb867ca1640866ba51780e
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:7b2e9c373f9bfd91e16eff68ef2313171c69b94fe5a21d3d3158470c74f54a05
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-ui@sha256:b9a255ea269a11c85f52fe556291b0e869efbff9918b488b65e4297b958c6059
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:24ca69c2f80a4369e29aa382a94d64fd3546b110787df1142c45a20b8246e844
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-usage@sha256:8438c2229b500fc9c40629280701856345e228fc645a5b468bee2135c61b2387
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:4aa81bd0043d0bac60c9f1b0ab56e45a3035c88314614e60eaea48b2a29ffa16
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:7655f531c354cfcce438c61cc5cde2c7690bfee4e2eadf9f99e4b157f163cd73
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:095c969026e191bc4fceb020b5760e7595e439cca91b88f0d8c3e422591d56aa
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:04f9ed67696fc193877f9da3d73ee3f6f4cadc7b8d09fd07fa12d0ee15f1152d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:42ffa27ae47b577cdda214e0fbc46c10bd108211e7dc58777d8236c59cb78766
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:b8886b9b54e6ee4c6b2981ef092e54348a0286577d9924cf136cee881f5bf1f4
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:6cffef7631ee54612fba71efc0eff4e17bc08defc3e2137aee9881d685444298
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:6cd1825ba40071bca7f46c1af3e813ae7ea44a393dd3d10e3c83e62834563ffd
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-usage@sha256:fe5bc4a40472926ec0df3023777b2aa8105507128bd9ac02eac2c68ad08723c1
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:d6575de3156c7a9b0553887abaf8713a758d9c9916f9cdcef611923c50632867
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:21dffe4500172f74345fdafb9beb476e0ccffa55f7bca59f372f312f8d28d247
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:67a02a9264d9c6e82edc15ae36d4f99b3fa9099697c550088d27c76511189f37
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:c6862158a2a5ae0a4e4206bfdc7f73e1c2359544966c72cf4fc3a617acb13697
        - registry.marketplace.redhat.com/cp/cpd/runtime-manager-api@sha256:a37773df07fe757dd698e3f09f8687902e7b3592d97855e703dd92ecc832aab0
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:3c1e08c56430e9bfc0c831be2394fa28e395a3341ce5cb5201b08828298deb60
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:4184f25155643f01850f286f15cc92294a92d8fdc8e5020b1b522977ba429c2c
        - registry.marketplace.redhat.com/cp/cpd/dap-base-job@sha256:f208686a3f56d2980e7b8640b6ebdea9057109e7520c6817a731e68d08506a36
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:0ab7d6651cdeda78fd0416137eff72eb58a621c8cd55bbbcb16f269808dec7cd
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:1a1bc54977b160b70cc00cc4eb1380250dcd639fa3d741363212b8081b35f72b
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:b0c8bb70a01bb19c685d019e89f83d8ed327bfa670a13b78d3e803fed77499f3
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:b0735e396c09fbd08f328ce0686cd55fe9eafd5b759f36cf5677326b24cbb859
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:ae5e6858ed698d67b4b3917c0d41d0b4f7e2564585694f834099584c21a3c012
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:234a8dcd8a9c5a68533bce6e0f793324e3ad7a660fdedf95895c5ed73a4cee3d
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:25c98c67d09cfaa3a4d332c461fb6766700a65876a48a1aabbd73b58bc2653f3
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:bb45d0217b4bc4352b209e3771756cb82923b8d260846c75b33ec270b4a8b8dd
        - registry.marketplace.redhat.com/cpopen/cpd/cpd-toolbox@sha256:1b3469f6fcd037aacabde3e05605b908e22f012d331d7365d0450f6d933d3efd
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:7d8fb7d8e2551bf949edd33589393fd12f634871b486e26464cd089e3e4b429d
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:fcb79d76f965e74f7e3459abb92e3561ba2f016b34e2ebb6f88298bb19f7bc11
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:a1754e003811afb578dace78f9eaef91e7917e5468e590fcb41920fa71bcc22d
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:93f73afa23f9fe8b73fd389081281fe1b0fd74bba6aa9bf7a6df663053afd3b8
        - registry.marketplace.redhat.com/cpopen/cpd/ibm-network-performance@sha256:30e0b58531f876a47174c14bc6f94a3c38a40f757403775a06745d4ae992060c
        - registry.marketplace.redhat.com/cpopen/cpd/ibm-service-functionality@sha256:ea4c32a19b1549698c8a68fa030d7036d3ea3cf2d000f3523b6e9fc7c6c708f1
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:6f96b8b6afb00cb26fcf903b8e4c30b0d6344bb86d8bce2f2026fc8cb0b575ec
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a4daf28771178b0af391965684337805f8de0019a677753112961710f6011991
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:673320deb0ae96c6132545032f6b3e569d40204d3a5565a568dd97c5c1a4a552
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:0711e1c6001586c950e74fbdeef8f04d20ce7b93de2ad033e82d2d5421338936
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:fb37fe521693bdb8b96400fd491a39c573545f35b4131fdfab501b7ba29713f6
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:cff7466aa6063a2bca677815df01208822930f8c12784c1c406fa352d6bd6cb4
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:67bfdf9115fbee782e35ac61dd93801bcb9e622d441806f1144208b4447f1430
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:7dc67256c002e86ee416c60ebb2942b35d8dfe0a5f96b4db495eff7361d21466
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:3d83ebdfddcbe6eca38705fc3d575f2e70580cf9f7e8735941afee0a33d3b5c5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:96451d4a9d1c4b759d82dff2eaa9aba2a6fd58518f6ac6306cc082ebb7766a64
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:077107dd9cede2c49beec429decb66760372a229c476ca1089e42e4ed5ef40d9
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:267b7da5936493b0daf8ce824e3e6b50ef3b2bcb6e3126ed18fa9e09cdb0da43
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:f5527ddbbd0b9878555fce852996b13020fd91b1604d8c03d34a61ca2fdb2d1d
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:ef24cc79265473a8b6fea231b1930c5763293dca79eedd46c89ec63dd95812c5
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:3be006596f8e1d36865a9000c77df3223c264e807bf0aae6049a23307e8079e8
        - registry.marketplace.redhat.com/cpopen/operand-deployment-lifecycle-manager-bundle@sha256:68f71f2767a40ca9333f0a670e07493e8a8c13a5c7ca877ea5bd8a544b352d9a
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:749dd365ce85a4c9ac8e6ea7f90a34b791a7bc5300319daa85d427b04f40af65
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:e4eb6a75e0d207232e479693d9c99ba863f8e62830f3191ca4d1e7ceaf3586d1
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:481391c273cffdeca36f91543732ba1f082dc409ee851ae95d3c1d205a7a6a08
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:173ef1b032594c4e070e2e013db469115ec58d970fa8b62f3b748163b46d6d3e
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:78682d9b9f19eae363cec0b6d31e696b2a5fd1bd1f8429df6f65a197310b6778
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:107911ed6a60a024d151ec3724ddd6d2d1402aa70a519b9dbec1a519d88231fe
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:bcadf3d367f76d3454e119a9d39f6fdeeb6cb806c282f21a1bc3c38bed41b427
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:25a4bb9cea1c3bdbeeeb205622512bcba98163f9629eba35625756346ce9d652
        - registry.marketplace.redhat.com/cpopen/cpd/cpdtool@sha256:714caed1042ee03461efc9256a3da1046bd8f43819caf3a91ec38b118d459869
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:0164d8fbea591bbd9d67d84603162c3781b08551eddcc2a8a842aafc828af18c
        - registry.marketplace.redhat.com/cpopen/cpd/cpd-toolbox@sha256:dcd74c128502dc301482eb302b34c21dc02173ae1cf522356d5b500ba42b4e25
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:23337786ca20ab624eb428b0d53490975e2b08a2db642ccc76d974dc0e455f55
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:6e055b6e07cb1adaa652b8e7515c4ce91438593a56ab0ae051435f55373d96ab
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:3a872814d01ded278afc8dfa6b1569d9e933e8fbda28c55a991da010b4304c69
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:62102f79bec5a8f66c600d184478f2b42c8acfde059fc71bc262b16dd46975cb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:686700e561181ebe1431f71eb9aa11b550bc4361dab8a3668ce0c7ead217f79a
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:40064d2893ce54b358c02023f4cc3603d3d28202ba56a177b20636abd74cb559
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:86cdb5b6543e7184361cbf0b0fa556277232460737ad254e920756c6d365e47c
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e1b10435f94bbb7165703720f58aa8c68acbdaa18006a2c36d1e24ffdd72cfc3
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:8fe85b93b9dcab9c33306ae358e6910685e109f32662d02e2b6bcd8dae243968
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-oauth2-proxy@sha256:bf4052b099393b563332acd5dcba6ba37ac424aa431d2d4f21532838b5ed6d6f
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:6c226a0ccde697b56c81427a970062872929eab73d99791c9dc670f4aa176ab8
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:3f6e9eeb659857799e30011f222924cef35301c59eaa5327ab5e8d212af75ce4
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:47cd9776a46ae5fdac3a101d79a779baefc802ddf8e1f8fb02216f47feb55fa9
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:009d98ff8e3cf353cf8ba71a973228e1e409e4de21373498534a6cce01c07ba5
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:19ee4cb8e3e60d9c1e8fa63b0bf3e218bd683b6c05b9e9553d9477c68c493a9d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:14da175e95c91864142ae00e9087ac2fe0e2083fc1a3bcb20a6b29dbab99d8a3
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:381170c9c497b1bda33eef3ddeb1bafefa07be1e6ca75e75c9eaef9dc87b8d4a
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:83ca1235dc054bf9ead64932c85333ea0dbade076b4a0cd1c3fe2b6bc2e05896
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:f2f256a2fb787419b25e99c765003d25d1edbbe60042151d91b46a5d029fc0e7
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:68ae72930a346d0fa0871cff573c64285266311208f2bf64101697a540f60272
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90046eb6170e12439ccc5aa343f19a4d4fc99a10d98b383cf1dc0e020ebeed55
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:20863de342e07dddc9aa8eeb214fba5cb17478bb88613e4e1f167d77cd0af214
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:d023f35b65816f244b6d699b41362e6031553f7d4c416efcee4a7f2395a61355
        - registry.marketplace.redhat.com/cpopen/ds-operator-bundle@sha256:e30b09797e341a5fdcc3ec377657e0e65be5855fb850dfc31f35b200f43130d8
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:ceee171c5ce0a8e580dfeccfe0dbfb73474c109becd45a562411678f3ab08595
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:36e3b8e88cbb895ca89288ed385fb800b888743844f50fb2233ebe5e6c7498c3
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-api@sha256:a76d3114418a3df1d4d3cfe03da587c37028ca1697795e2535cac9f6f190457d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:99589ee11b50af8ed6355a5f9c2272ce1fc7f9a13e80fda8eda5883a332782a3
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:20f05c662659be9b2b54407c02f1805b93f09a3c204011981b7f17a76caad8ee
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:bcdf9b7eea8753b3e463973a2690140ae7aa8c02c3c7726511f116bd751ec2d8
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:1d6dc7a3039e6fa044053d4e89ef54ba2e83ceadd57bcc4613a6d447d027ae2f
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:166a3918b942a4ea7b6c371a1e30fd4c928586d7ffd5b6919c2fc728a086ccbb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:fa3d8a54b1ba33465171240d66f62271286cb120c51c54267d2b4f8b731c6177
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:17bac3307ad3843afea77ed213279603b5f775f24dcf4857095baf5029409c18
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:80c0913e2c84530d9b7c6a2152de76ded5754c134f6768754d00944f46d0ed4f
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:186e849f3cf782f0d78660eb4433853e1e61296234d230a9dc5921fa4bd41060
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:f561f089c4e5037f7c04971d96fa024ba4df5e6d17834f0c44545d0e53c772d7
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:c0b6ad3b88f216379fc8f14aa5f075f5adc06ad06cc46495d3a40cc765f3deba
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:f289ac0d44803ff12d41c3659e6fdf04c404e0b3b233b6bc44eec671abb1b47e
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:3129d9c7b096f9fce9cc12002eafb5c3d731a162ddef61578455dcf8acd31fac
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:cb3620eb9aec24f7ee54c3ec6709fc223b685177500c503af8ffeb23642dcdb5
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:4eab846d6e80d68ce488e2cda36329cec1a6bf2f81feadf666b1005fed294e48
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:dc257df716b270858ca5d07909bda9ab635a388c9887d30b015391f60cc7b5c1
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:443b51f8b10acc85bdefde7193e0f45b1c423fa7fbdcaa28342805815c43db3d
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:c8d2825596b7e675c247158bdd06911e1cc0ad9af3364c9c175a5b4cf16d5d8e
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:545b41cc6c044d6807bfeb7453a045e074a5670cfff9f5c311f4918946901fcb
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:2668a562eb8df1117c40acbc3600d6735469d210e5aae4aa03bf9f54d0ebfe46
        - registry.marketplace.redhat.com/cp/cpd/environments-api@sha256:6dd60b530f81d7a2d55a534211f7f128814abf87b8cd3575d528dba53442c44e
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:0ae22914dbaf09d6679bd06bc9b6a5728e744c9ce2309cbd5f01e28308c403ea
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:0e0ae96aff8540323da8a2379af014d3af1770e4b0043b84b9b415dda44501f7
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:61155acdb8982648ae5748be9f5944ba913db1590ec3e773206d6fbb1d8952cb
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:da38999d66199df0eeabd7c9e934205fb858cf2c52be8bbaf6b34adb894dd38e
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:9ced60f18740063ab281b836d4caf21035242ac3f659c6bb022ddb546770bff9
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:0341c499ae144ee81dc2b92e8b99ee313ff6759374b7c1a6469156c20f522921
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bb074bdf89cbc7c703fadb17639608599ab4f66db9be5617e5f79a0550676790
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:4b3ffae1ddb7a2c28f98199be6343ed826f8fd21b64fc9c584566341f44e44d9
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:2d320dee224cc4818a84493723f2262d259891bef89dedef09e1f1e6ae5686cb
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:82f60c881048618473c07970d6463b3186518bc18e4b263e16ca61d15448ab30
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:ef38f12712c406f8206742dffae9378e344d66c774e5b37b4fcb0b657db456ca
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:887348b5035f5a70f1c31aede992a5c76397f5eddd2aebeb27074c3ea6248c00
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:a65ed5ac55cd5be99cc43c5a8d33bf127e8694917fce5f83fb8401c04e31b1f1
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:508412a22fbd65074964150d6b270f38d6df898e17eb7ec5e28f279c4c2e1062
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:48b7cfa31b9a185176c20afc119e37346a6f9380d74000edf6c729b4677ca44e
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:aa8d647f007d921db565a505d52583511e25de4a6a1187fc6b4df15427944364
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:39e6aee0bb64fe6dc18ce81496936b113de6d6746221655cea102e04c11dfa33
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:88d3c1ee02acbffedbf6ddcc27b80be6727f85425c361d45cc5e5c4056cf6733
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:0f46b288d060a15a48ecad8f328a8046bf232b8629c3d79da6a82ea712b363f4
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:603a6bf0e616f16b609bf157b9e21415802478aefc9c606e124008d0224a1188
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:2d8f924a7e4ef373df9760b4f1f3ad8050facec7de0c0efccd8c27d46b756e91
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:d0e0144eaf27e03aec1499e53cae5366e11a3da284536cd3b27ba5691b4f70d0
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:934bc42fb29cbe616edcd93d2a060dbffab5e5a8a2ef4caa7b2e538f26167ee3
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:075b6bc4866b812c4b5394c8902b012978449a710ce71a30f55231fead941105
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:587af68ec8b4974b29da46ede525c70054703fce8d5934f0cbe68739c6058fe4
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:010bd5bb9943ffdffa8fd00cc8226e59d9583bb3baaeacfe70e5b00e805878f1
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:9c160303552a9b8a97b951311018de2ada410883468680dd61fb41519b9a9e7a
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:758f9e49dccaec5ebd125fc0c383fb5a656ba9f2f301984f993d50b103208d12
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:d2680c5f05c886b44df644aef6cb2fdc9534bb6a238066f6ebbb97bbeaac431e
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:e9e408a13bd103fb46536f896c51e7a27cba38d9b224231b20328a28efe1e616
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:e3138406cfe81021370924992dca891cba22b3be2eef84546faa0e8f8aa57ad6
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:b646ea3330f3fd9bb3174224de69f20ef58589a464d50d1c9399211141f67bd5
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:c0459e5d3acb3b2a6d9863efe6ca1a24e529b38a6ddd3cc1e13d143924e872e7
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:00c30fe372524fa82b5ed4a637509dd9d4373dc727a9abc706221f832858f9bd
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:bb2f19e1797a900d60fb9f05d9f115faf20d514000aa3acac55fdba701581d81
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:a95e42a4f9138f1f97937a74982c98d466c171d0e3e79c522ededba6c25b39a3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:78e33fff4bfc2e33ea179eaed9902df9bbd11463b9152d0281d1e0924fe09563
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:d61b7b8e6a9785b59630d9c5776f72a29ffbf2e81eb753d7b11dcf2d9d77f1f0
        - registry.marketplace.redhat.com/cp/cpd/environments-init-job@sha256:6af8f900af7c11eee6894e2c3e52f740b3c700dac0ee753fadef330230bd42dd
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:b2034fae4f2287c8997104154064f2712b98d3c4382d1faea3df777338711f23
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:a34169228887ff93161584753c990bb82a5520ad7945dfa1ff4078e70aede592
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:c91fc0992cff4b674f5e3690398e7dfb70662c807b930b9b0e65ba507ff925d2
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:a477d47eaf8f8e569601d3c23c1def66f31c098ac3f5a1ae350618c460917dd8
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:498ad7832e989e9356aeffbee96f310e3edaa6a7466320fda9b4f75034e73eec
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:3743af183b7b801f72b279a658b0d32da0bdbfa526a00d191a9a44de6df2284f
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:99c863002ae1e93e97792fa606128d05063abbedad9e9404acf28c745ae1a5c4
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:8bd9fe1d35785fa0125b8fb3a60333219a50ee8be7a7ec07386e8e9d65316061
        - registry.marketplace.redhat.com/cpopen/cpd/cpdtool@sha256:8ed09b7d65178aab93b77b9d39371bc86b1022f32c5e63a9192baef02c71016d
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:48d237cd3d363da8a206ca5e092d0f91e8a5bc931ea81763498dcf6f28c0434b
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:722944bfd0eba280255f4c198f65f562423a26e2c4ad3580591ee71cb6e9c2ef
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:f1aa6add76f59afeccdff84c9ea89a46b49c6696f62acd75ef8ff2f31a3080be
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:860c17d2b2c6b0e013d0f4c035a388e45d8ef1483f53aa7ad022913b025f73b0
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:1b29e04308bb8e3c990a8c41a5d3d1e854e197bbdf0cffb2833fea4866772ba9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:ecbc65705de76804a3c7642339ecfb4cc0f5a61891058e07c20bf2b63f1b82b8
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:d5feb4aecf8f36531db4c52e672558c04f7e314a1a791b27b5638202f9477bf5
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:b6500764d4f6d281f1b42458eb223d21f6a413b5db55f2a2ba1a7973aad8465c
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:6fc976bc622c2165dd8bf93c9244e1c29b7c973a4da7a5ce4d0a097a1c1ea4fb
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:f0c68715f78bd20c352c97dd48ac5a912cad58e19a00159b5e6878f6beeb00fb
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:bb85f2567aaf7b612343fb9819fe05e2e771ccca8383316d34282ec37fcd1383
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:9460c98c7202ab368e732951c13bb6f5064cf18f69d36b3d26eedcfed233ec81
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:bcd879a7cc3442ef62bb9677cc5534a2bfc9e7ea248e566aa4f1f6a0242431a8
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8da8b240176c00f8f6eadf32e5614c663cc1d56d546c73996cca614b715acf4a
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:7505eb74c6bf165d40f6a8f3511f8e891cc34c0201fee38240c693a08e002ab1
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:8160c688f6b66206f8428f78dcf5e21b65c3a1cadc38b765f22ba8403a87f4d6
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:ecac5e0327d939ca84d304b2fe95739ae43a06745ef775ee59e4e5378ced65c2
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:f6f3758287dd31889933ed0461fcf02dfa85d872c3d193669608573ca5269524
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:c70d3c86d4c61bd87a427f7a7571f37fc4d0966ff4f420b73ac4a2af73ed1379
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:30a5b2c4e3c231373fa18d8aaab16d3505cc11a8a5d4be36eaf99d7c0dee35f2
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:5e161ff695c0552d1333a1f993048bd055168a9be9e0a75adb7c4464c9c2b603
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:044b663e4b7508c62dbb28b7fb2aa267c9771fc62ab509b08d46ea1499d95431
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:27210840cbff8a7c61acac2ccc328a415a405b4738422b4f47eb9a0110127d22
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:e21a346d15b286956f62bdd1fbb6d7b77a554ee10a475cb698a8796c42f3a267
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr@sha256:43cf735b51e2ef45a911ac7032f03c74cf0125e2d2801258d4aa04831e1c6ce0
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator-bundle@sha256:e4340bd9481da998a6b89e94c884c46e17cfb50d9d0c1099f39a3c54e500d2b2
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:9782a9de863ad10ed36341d07ea5087e40f6f213cd4453c3ac2ec100d8733ce2
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:4ca30e2017a0ff2d64100a03a84bee8d601c765f0e624e1f485cbb97748bb8a2
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:8444703b2a7822f82141afcf4f36bf255650d755321b9e9fe2cbd1bdea41d5c7
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:2a50cd6ee3cb352ba186a57ba73fb9df321cf98f0b79b3d81024cbe57e0c7f35
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:aa0068aec14e37986620c982295c083df6fc6bbe95facba582ae3c6f769a8e3a
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:d30397b74780cdc14930a3a02dfcf6b99264847702afe0513ce536983c32e2b1
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:0f01a67524fa46829e455f654d3b9b00ceac44f3d870b21c022051dfcc8a00eb
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:abec4867d6b9294d674cc02b9f7f254e65e7de8eea0231fa94f50513f2b7c45d
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:4d24fb274b5429ba61fc4bc6833e93b14b92b8d525556db64e2172b45433de31
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:07142ce0d259412d0a9908b63dd78d845ef6b7c5f8a24442b93964b4c38372cd
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:8cba5f256747002b62fcedb9066f7ea6dfdb1071eb359e9137fd413d20b29ff8
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:f88a7294aa200d3c633c0a25c4e4ccda52c75de174ba508d847ffa3092d5b8d1
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:6f2e7cc917adb54b3112dc4909f645ac08c33e8c2c33266fbc8f5133cbcd1973
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:99d0b3192e6e46167423505f48655e7d49af1a461ae00b95c9c7c9d4209e5b65
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:0157fe80f2c67786cacaaf1302c334a6e0ec0b941afd870e0c4ba163ce394216
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:7661b04bc59d64a4fafb2f660cffb99e3d326772749759ed0d2a1940ba147078
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:0cdba17b95a959b6d04889179d105515ed3176ea3449bc104f18b86374876048
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bcf193d6f30ce1c7f9337ba8efa8d9b853eb7961e0008aa82635aa383fbe9d76
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:9c635194a9c63fed440e7c0ab0be1fe6cf59b1c3f954216dfb9150fe9e0fc207
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:9a0c76fd9ef9b2dd9af5b0f29e1813567ab7f7d35c1efdf71da1a315f2942b45
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:3c18707afb90a7dc1bd2ed1fff05c53cce89b7c85062e72261f538a12834ab37
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:1f82599d91ab20d2024abe134d2f56caca1f9bca1f02d64352c81b5f1e0f1730
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:e6fc9d7a70877100b42e290cd091011f2e1c674a27985fce180ee245373391dc
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:c35eaf4921bfd046985ea7a03fc325f79766bf73e3ddba69408f4467fbd7994f
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:18f44ab15289625fa2c6b72ece6e06afb51058dcba30b8f38b4fdeedf8c9e51b
        - registry.marketplace.redhat.com/cp/cpd/environments-ui@sha256:2923664475c7a8ea5e63055c90748db37825fe76c7975b845f1185d4b32ee4ea
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:9f7e6a1ca3faa5af1db2afe8693f6bf9182296be2a23e305ba6c990aee3fc301
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:183ee860bf48076ea155db08032eed7004c1204a363625f54218734b00175441
        - registry.marketplace.redhat.com/cp/cpd/jobs-ui@sha256:49daf607918b378593485da16ac417056ee295bca94f1a5cac37b334381ae096
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:78fffd8a74eb6495f104ef346bdd03a5636134e2d5e8da9ef5c2647a9ebc7624
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:97b9ecef8caef156a0eb073b6eb2a34b63209b44511f5694a243f11285d88662
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:3f2f97e6b2b18526255e88ff02d1039b24312eca51ee22b7cbf76d5680f7aee1
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:74aadd4fe99e66b89471222f853e442fa8338a23d9c24b4e908d38536c66456a
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:2754384c6936b36ffd64f89df751cbe47993f73dd700a184af35f412b3408d72
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:bcfb9b7d3b379ee2bb0b449cdf5af3fbc88f99c77641441ca392ebed04403e00
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:b19bbc36033c3e60f387a9924711aab1ed81f150ca9d8f435e5433fed7f6a5c8
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:b3387bae415fe3a66db2264ee766cf43aa8f31e2b615b200baa9228216596a0a
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:96ebe954a21f697ebaecfe5aa2c497c60b2ddec17e9ec4b376f435e6c53b3c29
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:a10a15f07b856a7803ce0dcd30a542741dcb591804dc74c2a5fd50c55de0a983
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:bb6492c452fa6b481551628da26bb85bad151d12afd34a89c1a272a9587ecd1c
        version: 6.1.3
      entries:
      - name: cpd-platform-operator.v6.1.3
        version: 6.1.3
      name: v6.1
    defaultChannel: v6.1
    packageName: cpd-platform-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Arkmq-org
      provider-url: https://arkmq-org.io
    name: activemq-artemis-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: activemq-artemis-operator.v2.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/arkmq-org/activemq-artemis-operator:2.0.5
          createdAt: "2025-04-04 14:35:00"
          description: An operator for managing the Apache ActiveMQ Artemis message
            broker
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/arkmq-org/activemq-artemis-operator
          support: arkmq-org
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: ArkMQ Operator manages ActiveMQ Artemis messaging broker in the
          cloud
        displayName: ArkMQ Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - messaging
        - cloud
        - kubernetes
        - activemq
        - artemis
        - integration
        - operator
        links:
        - name: Documentation
          url: https://arkmq-org.io/docs
        - name: Repository
          url: https://github.com/arkmq-org/activemq-artemis-operator
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Arkmq-org
          url: https://arkmq-org.io
        relatedImages:
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:49f34122ab46ab5d7ead094b6e40817b62c75b6a809f67db07d68bce750da51e
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.0
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:b71d1a57c8b3820a8133cf7a25a6edde7b29e1c4d32d7c30325183415a8bacd1
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:4bba8828e72fd7c9d38c22ba1c2a4f7130817a4e7eac8fb63c13e86465a8e1e7
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:decb3932ebb5b848b90a4a23b083f0cb5237eec2844095d8c58a6798da8bb999
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:42230ce9d11a1949a5bb73b6ee218b1581f585bcd553485ee438e1c5b50fcf8b
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:ddaf00914e3a412d21ec3a814e24f3d4cc382b5c4d6c21e9b1b94dde16ddf194
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:52f337cd5d7d5d199eebf4fc722f8c5279f652506f009efbc0f8f6de88e18718
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.7
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:6f2da9ecd8f5021964d0dfa85144830efcfbd0462d18f18f036971e5f17b3bb6
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:5e768aefade6d18cd65cb1eb5ee233844f9f7e7b52631f93a0982b67d382239d
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:4be60cfa2a300b2eef9b5f10ea756bd3f55d75fbd8890232d907f8d64080449b
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:1fd99a46758cd29fce846d0d590021758b2a7a3efedb817bb205359caa7d4f67
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:f25db980c3229e66cf48690aa175f6f93e3dce3974c0e70ab23af6462e7d1893
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:634f3d8fd620607b56e0360a191d23a9cb8a32ba0ba1f24fe20791ec52343586
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:264b33a9aad90d32a41258904f35092a141d76dcef0c115d414475444100aace
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:93168fd8cd4fb32f041795afb8fb0e21cf81081d6cc463a44356e153dc1ad032
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:a4628c2591ec06caf21a3467f6e3faeaede7fa402b4f137d943b25b4dd4ad343
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:b293c1269166cc6181f68b8d803fdfcf20fdbd5df2787ed0453724b22e298fb2
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.0
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:fc9b9a4c7550d84fd90335dd4f9e0d8d39c09136c6034b28490a368368cb11f1
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.2
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:a4a0d0453f0761b9e114d0d4cc7c14091a2a9c30790f18ffa6bf1e365d5083fe
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:9e1fefc9a94d72a7b5a3f4e2d53ebd23e5ab221df17a5afde42ce973277838f6
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:9bbb915855f29ea16950a3fd50d0b9bac25f82cba7d197579e04d9c33a5b02c4
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:f10983c68aff358e285d89d827c27b3487a3ee3b8e121cdaa295bf5fa8501145
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.7
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.2
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:719178251893def955e444e706ff07042ea98d294671315d95da504c1f47de24
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:8fbdd1ba728058d5f437186b9055ca28596140068ebc69ecff4fa52717b19a82
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:de82344daa0caeabc6559656da0af285436a687e8f371744c0c51bdbc309167d
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:8f592b2454d092c82b5f61b4e04c3aff173db246539b3c06e15a71960d3bd235
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:b6f61188379def8c796c738549c099bd40bcd9bb44f5c30682c5edd136347aca
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:88a0098a628fe43a9b422f8ebb774086fb65ed389a195e2779482a4a63447e28
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:70d4c799c7d905fac27457d63e662ea72bca9e899a692e425609e4eaa45129e6
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.2
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:37ab7354273f4dcd19bad6936f10e92220d98b8c0266002221477a0e2d146c81
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:81590ec33b2f00775d3c6596f610887b670de7c0b89148b09eab7584f5b0b0ec
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.10
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:ffa0d7cb4068df84319776dbfb66210297da1b8466ccd5451318fc3de79736c3
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:56bae938a6b4cd87f9f5284931f440f9b9d116cda3b63b13f83a4b1c6c641959
        - quay.io/arkmq-org/activemq-artemis-operator:2.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.4
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:da95bff69deb63545d9782ed20cbdd17c1a79c6b16165580bf96a1d5191e6a59
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:2895b2cf6292c951a0cdde78e755514f5d7cb43ce9f7b6e7dc381d8dfeec31ab
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:3bbfb6177e4839087f615a01ee54f9e06a95acaddbb083e11851b6af9315dbf0
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:e438bcacf099a92b143793e88c00c6ac0afb31effe4b0eed49a935c9265ad943
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.1
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.0
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:f2c853ff91809282c04d5b1381ceae366518bf07b16649a61aa6c60c47f10714
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:2d155bffd246f26f516a6075ec170babc1373c72bb664cba4e3c58521ffd3558
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.7
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:e2438670831bad39bb08147bbebe67ef9465899fb7750f1a5a185569f99d5fb5
        version: 2.0.5
      entries:
      - name: activemq-artemis-operator.v2.0.5
        version: 2.0.5
      name: stable
    defaultChannel: stable
    packageName: activemq-artemis-operator
    provider:
      name: Arkmq-org
      url: https://arkmq-org.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: ScaleOps
      provider-url: https://scaleops.com
    name: scaleops-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: scaleops-operator.v1.7.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1alpha1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops-sample"
                },
                "spec": {
                  "scaleopsToken": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: ghcr.io/scaleops-sh/operator-bundle@sha256:ee627042deefa4faceb63efd56aa6f6338c90f6c13877f29d7161efe603e02f4
          createdAt: "2024-09-19T07:04:10Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: ScaleOps
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1alpha1
        description: ScaleOps Cluster Optimization
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: vlad@scaleops.com
          name: Vladislav Shub
        - email: guy@scaleops.com
          name: Guy Baron
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - ghcr.io/scaleops-sh/kube-state-metrics/ubi@sha256:d8cb65795861ce257f08726b7a3aa60a869cc44ed152f559431191bc79e98908
        - ghcr.io/scaleops-sh/operator-bundle@sha256:ee627042deefa4faceb63efd56aa6f6338c90f6c13877f29d7161efe603e02f4
        - ghcr.io/scaleops-sh/operator@sha256:7de0651a76d6eb63fee81e9177df93f126906d50063d7891f12c4a72c817dc86
        - ghcr.io/scaleops-sh/scaleops-operator-bundle@sha256:128771d92c852f7ace78103493db079a474e06d1352f06cbde7a7d4666684610
        - ghcr.io/scaleops-sh/configmap-reload/ubi@sha256:fed19f1aad43228560912abc36372a6c124d9ddf5e3e39476269dc51604d26de
        - ghcr.io/scaleops-sh/prometheus/ubi@sha256:8aeeae2a33251493d4d6e2f02a2fff1d4c2f32ab54554e6a84619ecd8b4aa69a
        version: 0.0.1
      entries:
      - name: scaleops-operator.v1.7.5
        version: 0.0.1
      name: alpha
    - currentCSV: scaleops-operator.v1.21.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops",
                  "namespace": "scaleops-system"
                },
                "spec": {
                  "scaleopsToken": "${SCALEOPS_TOKEN}"
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: registry.connect.redhat.com/scaleops-sh/operator@sha256:528f2e09d2dad4f48e01a272b51baa1753c12069515831223e3f97c7dfe03e3c
          createdAt: "2025-09-29T13:35:24Z"
          description: |-
            ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

            ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

            Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

            Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

            Contact [us](mailto:team@scaleops.com) to start your free trial today
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.21.8
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://scaleops.com
          support: ScaleOps Support support@scaleops.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This will install an instance of the ScaleOps Platform on
              your cluster
            displayName: ScaleOps Platform
            kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1
        description: |-
          ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

          ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

          Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

          Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

          Contact [us](mailto:team@scaleops.com) to start your free trial today
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        - application runtime
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: support@scaleops.com
          name: ScaleOps
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - registry.connect.redhat.com/scaleops-sh/operator@sha256:528f2e09d2dad4f48e01a272b51baa1753c12069515831223e3f97c7dfe03e3c
        - registry.connect.redhat.com/scaleops-sh/scaleops-gpu@sha256:724aef7683ea4cb245ab0a464517a6d542efa4634e5469ff35b7a7b7f00b563e
        - registry.connect.redhat.com/scaleops-sh/scaleops-sh@sha256:57d5d06f4f4b5470562c4f24ea5d37203633e1f4b65cd497e71d9066fd2b9e41
        version: 1.21.8
      entries:
      - name: scaleops-operator.v1.21.8
        version: 1.21.8
      - name: scaleops-operator.v1.21.7
        version: 1.21.7
      - name: scaleops-operator.v1.21.6
        version: 1.21.6
      - name: scaleops-operator.v1.21.5
        version: 1.21.5
      - name: scaleops-operator.v1.21.3
        version: 1.21.3
      - name: scaleops-operator.v1.21.2
        version: 1.21.2
      - name: scaleops-operator.v1.21.1
        version: 1.21.1
      - name: scaleops-operator.v1.21.0
        version: 1.21.0
      - name: scaleops-operator.v1.20.9
        version: 1.20.9
      - name: scaleops-operator.v1.20.8
        version: 1.20.8
      - name: scaleops-operator.v1.20.7
        version: 1.20.7
      - name: scaleops-operator.v1.20.6
        version: 1.20.6
      - name: scaleops-operator.v1.20.5
        version: 1.20.5
      - name: scaleops-operator.v1.20.4
        version: 1.20.4
      - name: scaleops-operator.v1.20.2
        version: 1.20.2
      - name: scaleops-operator.v1.20.1
        version: 1.20.1
      - name: scaleops-operator.v1.19.8
        version: 1.19.8
      - name: scaleops-operator.v1.19.7
        version: 1.19.7
      - name: scaleops-operator.v1.19.6
        version: 1.19.6
      - name: scaleops-operator.v1.19.5
        version: 1.19.5
      - name: scaleops-operator.v1.19.4
        version: 1.19.4
      - name: scaleops-operator.v1.19.1
        version: 1.19.1
      - name: scaleops-operator.v1.19.0
        version: 1.19.0
      - name: scaleops-operator.v1.18.8
        version: 1.18.8
      - name: scaleops-operator.v1.18.7
        version: 1.18.7
      - name: scaleops-operator.v1.18.6
        version: 1.18.6
      - name: scaleops-operator.v1.18.5
        version: 1.18.5
      - name: scaleops-operator.v1.18.4
        version: 1.18.4
      - name: scaleops-operator.v1.18.3
        version: 1.18.3
      - name: scaleops-operator.v1.18.2
        version: 1.18.2
      name: stable
    - currentCSV: scaleops-operator.v1.10.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops",
                  "namespace": "scaleops-system"
                },
                "spec": {
                  "scaleopsToken": "${SCALEOPS_TOKEN}"
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: registry.connect.redhat.com/scaleops-sh/operator@sha256:8def662a546babd6740e5417401836d647d92d52b15d559b1b3e2cebd7f77f33
          createdAt: "2025-02-12T19:06:04Z"
          description: |-
            ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

            ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

            Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

            Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

            Contact [us](mailto:team@scaleops.com) to start your free trial today
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.10.32
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://scaleops.com
          support: ScaleOps Support support@scaleops.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This will install an instance of the ScaleOps Platform on
              your cluster
            displayName: ScaleOps Platform
            kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1
        description: |-
          ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

          ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

          Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

          Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

          Contact [us](mailto:team@scaleops.com) to start your free trial today
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        - application runtime
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: support@scaleops.com
          name: ScaleOps
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - registry.connect.redhat.com/scaleops-sh/operator@sha256:8def662a546babd6740e5417401836d647d92d52b15d559b1b3e2cebd7f77f33
        - registry.connect.redhat.com/scaleops-sh/scaleops-sh@sha256:5d77e81d7959483121c208072eb61b063eb3ad9abc49e7df3c7efb8e2b77a3ae
        version: 1.10.32
      entries:
      - name: scaleops-operator.v1.10.32
        version: 1.10.32
      name: stable-1.10
    - currentCSV: scaleops-operator.v1.12.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops",
                  "namespace": "scaleops-system"
                },
                "spec": {
                  "scaleopsToken": "${SCALEOPS_TOKEN}"
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: registry.connect.redhat.com/scaleops-sh/operator@sha256:4bd74a56a8680f47aa1da7a377f74529fe7ca1568995a4e74018af0cf67763ed
          createdAt: "2025-03-11T18:10:07Z"
          description: |-
            ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

            ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

            Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

            Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

            Contact [us](mailto:team@scaleops.com) to start your free trial today
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.12.6
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://scaleops.com
          support: ScaleOps Support support@scaleops.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This will install an instance of the ScaleOps Platform on
              your cluster
            displayName: ScaleOps Platform
            kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1
        description: |-
          ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

          ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

          Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

          Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

          Contact [us](mailto:team@scaleops.com) to start your free trial today
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        - application runtime
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: support@scaleops.com
          name: ScaleOps
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - registry.connect.redhat.com/scaleops-sh/scaleops-sh@sha256:5ea4e26f2fecee3570d63adc3e2588d34f9610345d703975d4febac9c94b4bd0
        - registry.connect.redhat.com/scaleops-sh/operator@sha256:4bd74a56a8680f47aa1da7a377f74529fe7ca1568995a4e74018af0cf67763ed
        version: 1.12.6
      entries:
      - name: scaleops-operator.v1.12.6
        version: 1.12.6
      name: stable-1.12
    - currentCSV: scaleops-operator.v1.14.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops",
                  "namespace": "scaleops-system"
                },
                "spec": {
                  "scaleopsToken": "${SCALEOPS_TOKEN}"
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: registry.connect.redhat.com/scaleops-sh/operator@sha256:9267f49ce774ee4d401458531599bbd6729c6313cecc4f2bcd3bf0b5eecd92e1
          createdAt: "2025-06-08T16:15:46Z"
          description: |-
            ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

            ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

            Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

            Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

            Contact [us](mailto:team@scaleops.com) to start your free trial today
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.14.14
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://scaleops.com
          support: ScaleOps Support support@scaleops.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This will install an instance of the ScaleOps Platform on
              your cluster
            displayName: ScaleOps Platform
            kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1
        description: |-
          ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

          ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

          Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

          Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

          Contact [us](mailto:team@scaleops.com) to start your free trial today
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        - application runtime
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: support@scaleops.com
          name: ScaleOps
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - registry.connect.redhat.com/scaleops-sh/scaleops-sh@sha256:9c6395202ccb70e8edae3a1b540de31991ec9232b7b235dd92755588b8491720
        - registry.connect.redhat.com/scaleops-sh/operator@sha256:9267f49ce774ee4d401458531599bbd6729c6313cecc4f2bcd3bf0b5eecd92e1
        version: 1.14.14
      entries:
      - name: scaleops-operator.v1.14.14
        version: 1.14.14
      name: stable-1.14
    defaultChannel: stable
    packageName: scaleops-operator
    provider:
      name: ScaleOps
      url: https://scaleops.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataStax
      provider-url: ""
    name: cass-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cass-operator.v1.23.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cassandra.datastax.com/v1beta1",
                "kind": "CassandraDatacenter",
                "metadata": {
                  "name": "dc1"
                },
                "spec": {
                  "clusterName": "development",
                  "config": {
                    "cassandra-yaml": {
                      "authenticator": "PasswordAuthenticator",
                      "authorizer": "CassandraAuthorizer",
                      "num_tokens": 16,
                      "role_manager": "CassandraRoleManager"
                    },
                    "jvm-server-options": {
                      "initial_heap_size": "1G",
                      "max_heap_size": "1G"
                    }
                  },
                  "managementApiAuth": {
                    "insecure": {}
                  },
                  "podTemplateSpec": {
                    "spec": {
                      "containers": [
                        {
                          "name": "cassandra"
                        }
                      ],
                      "securityContext": {}
                    }
                  },
                  "racks": [
                    {
                      "name": "rack1"
                    },
                    {
                      "name": "rack2"
                    },
                    {
                      "name": "rack3"
                    }
                  ],
                  "resources": {
                    "requests": {
                      "cpu": "1000m",
                      "memory": "2Gi"
                    }
                  },
                  "serverType": "cassandra",
                  "serverVersion": "4.1.2",
                  "size": 3,
                  "storageConfig": {
                    "cassandraDataVolumeClaimSpec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "server-storage"
                    }
                  }
                }
              },
              {
                "apiVersion": "control.k8ssandra.io/v1alpha1",
                "kind": "CassandraTask",
                "metadata": {
                  "name": "example-cleanup"
                },
                "spec": {
                  "datacenter": {
                    "name": "dc2",
                    "namespace": "cass-operator"
                  },
                  "jobs": [
                    {
                      "command": "cleanup",
                      "name": "cleanup-run"
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database,Big Data,OpenShift Optional
          certified: "true"
          containerImage: registry.connect.redhat.com/datastax/cass-operator@sha256:6abc6253b50a0d7f6bdd4b394d2bf96320a8845952aeb8f0bbed99863192ec52
          createdAt: "2025-03-20"
          description: |
            Simple provisioning, turn-key operations, and automated remediation of Apache Cassandra clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/k8ssandra/cass-operator
          support: DataStax
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CassandraDatacenter is the Schema for the cassandradatacenters
              API
            displayName: Cassandra Datacenter
            kind: CassandraDatacenter
            name: cassandradatacenters.cassandra.datastax.com
            version: v1beta1
          - description: CassandraTask is the Schema for the cassandrajobs API
            displayName: Cassandra Task
            kind: CassandraTask
            name: cassandratasks.control.k8ssandra.io
            version: v1alpha1
        description: "## Apache Cassandra\n\nThe Apache Cassandra database is the
          right choice when you need scalability and\nhigh availability without compromising
          performance. Linear scalability and\nproven fault-tolerance on commodity
          hardware or cloud infrastructure make it the\nperfect platform for mission-critical
          data. Cassandra's support for replicating\nacross multiple datacenters is
          best-in-class, providing lower latency for your\nusers and the peace of
          mind of knowing that you can survive regional outages.\n\n## DataStax Enterprise\n\nThe
          most advanced distribution of Apache Cassandra™ on the market, with the\nenterprise
          functionality needed for serious production systems and backed up and\nsupported
          by the best distributed-experts in the world. It's one platform for\nall
          types of applications anywhere, any cloud, any model: key-value, graph,\ntabular,
          JSON.\n\nDataStax Enterprise is a fully integrated and optimized database,
          with graph,\nanalytics, and search included, all with a unified security
          model. Simply put,\nit's the only database capable of meeting today's demanding
          requirements\n\n## Operator Details\n\n`cass-operator` is designed as a
          modular operator for Apache Cassandra and\nderived  distributions. Apache
          Cassandra is a distributed database consisting of\nmultiple nodes working
          in concert to store data and process queries along a\nnumber of fault domains.
          `cass-operator` has the deployment of a Cassandra\ncluster around the logical
          domain of a datacenter with the `CassandraDatacenter`\ncustom resource.\n
          \   \nUpon submission of one of these resources it handles provisioning
          the underlying\nstateful sets (analogous to C\\* logical racks), services,
          and configuration.\nAdditionally through monitoring pod state via Kubernetes
          callbacks it handles day to day\noperations such as restarting failed processes,
          scaling clusters up, and deploying\nconfiguration changes in a rolling,
          non-disruptive, fashion.\n    \nThis operator is designed to be `Namespace`
          scoped. A single Kubernetes cluster may\nbe running multiple instances of
          this operator, in separate namespaces, to support\na number of C\\* clusters
          and environments. Configuration is simple with the usage of\nYAML based
          overrides in the Custom Resource paired with an `init` container.\n    \nIn
          C\\* clusters ordering and timing of certain operations are important to
          keep the system\nevenly distributed. `cass-operator` takes advantage of
          a sidecar process within the\nmain container to handle the orchestration
          of starting our main C* process."
        displayName: DataStax Kubernetes Operator for Apache Cassandra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - k8ssandra
        - cass-operator
        - DataStax Enterprise
        - Apache Cassandra
        - DSE
        - C*
        - Cassandra
        links:
        - name: DataStax
          url: https://datastax.com
        - name: Operator GitHub Repo
          url: https://github.com/k8ssandra/cass-operator
        - name: Config Builder GitHub Repo
          url: https://github.com/datastax/cass-config-builder
        - name: Documentation [1]
          url: https://github.com/k8ssandra/cass-operator/tree/master/docs/user
        - name: Documentation [2]
          url: https://docs.datastax.com/en/cass-operator/doc/cass-operator/cassOperatorTOC.html
        maintainers:
        - email: cass-operator@datastax.com
          name: DataStax Operator Team
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: DataStax
        relatedImages:
        - registry.connect.redhat.com/datastax/cass-operator@sha256:6abc6253b50a0d7f6bdd4b394d2bf96320a8845952aeb8f0bbed99863192ec52
        - registry.connect.redhat.com/datastax/system-logger@sha256:3408620786e6265aa835be56f85aadc48b20060b2038dc21126f614685f77c6b
        version: 1.23.2
      entries:
      - name: cass-operator.v1.23.2
        version: 1.23.2
      name: stable
    defaultChannel: stable
    packageName: cass-operator
    provider:
      name: DataStax
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
    name: aikit-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aikit-operator.v2021.2.102120
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aikit.intel/v1alpha1",
                "kind": "AIKitContainer",
                "metadata": {
                  "name": "intel-aikit-container",
                  "namespace": "redhat-ods-applications"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagestream": {
                    "name": "oneapi-aikit",
                    "namespace": "redhat-ods-applications",
                    "registry": {
                      "root": "registry.connect.redhat.com",
                      "repo": "intel",
                      "name": "oneapi-aikit",
                      "version": "latest"
                    }
                  },
                  "nameOverride": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aikit-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aikit-operator-rhmp/support?utm_source=openshift_console
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AIKitContainer
            name: aikitcontainers.aikit.intel
            version: v1alpha1
        description: |
          The Intel® oneAPI AI Analytics Toolkit gives data scientists, AI developers, and researchers familiar Python* tools and frameworks to accelerate end-to-end data science and machine learning pipelines on Intel® architectures. The components are built using oneAPI libraries for low-level compute optimizations. This toolkit maximizes performance from preprocessing through machine and deep learning phases and provides interoperability for efficient model development and deployment across single and multinodes.
        displayName: Intel® oneAPI AI Analytics Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aikit-operator
        links:
        - name: Aikit Operator
          url: https://aikit-operator.domain
        maintainers:
        - email: kam.d.kasravi@intel.com
          name: Kam Kasravi
        - email: abolfazl.shahbazi@intel.com
          name: Abolfazl Shahbazi
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Intel
          url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
        relatedImages:
        - registry.marketplace.redhat.com/rhm/intel/aikit-operator@sha256:7abb5e6164ffb3e4b87c39061f45d09848218fcfb74b48bf5e1035d56b41ee0c
        version: 2021.2.102120
      entries:
      - name: aikit-operator.v2021.2.102120
        version: 2021.2.102120
      - name: aikit-operator.v2021.2.100410
        version: 2021.2.100410
      - name: aikit-operator.v2021.2.92911
        version: 2021.2.92911
      - name: aikit-operator.v2021.2.92814
        version: 2021.2.92814
      - name: aikit-operator.v2021.2.92808-dev-a60e5ba
        version: 2021.2.92808-dev-a60e5ba
      name: alpha
    defaultChannel: alpha
    packageName: aikit-operator-rhmp
    provider:
      name: Intel
      url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cluster-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-manager.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "cluster-manager"
                },
                "spec": {
                  "addOnManagerImagePullSpec": "quay.io/open-cluster-management/addon-manager:v0.13.1",
                  "deployOption": {
                    "mode": "Default"
                  },
                  "placementImagePullSpec": "quay.io/open-cluster-management/placement:v0.13.1",
                  "registrationConfiguration": {
                    "featureGates": [
                      {
                        "feature": "DefaultClusterSet",
                        "mode": "Enable"
                      }
                    ]
                  },
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "cluster-manager"
                },
                "spec": {
                  "addOnManagerImagePullSpec": "quay.io/open-cluster-management/addon-manager:v0.13.1",
                  "deployOption": {
                    "hosted": {
                      "registrationWebhookConfiguration": {
                        "address": "management-control-plane",
                        "port": 30443
                      },
                      "workWebhookConfiguration": {
                        "address": "management-control-plane",
                        "port": 31443
                      }
                    },
                    "mode": "Hosted"
                  },
                  "placementImagePullSpec": "quay.io/open-cluster-management/placement:v0.13.1",
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/registration-operator:v0.13.1
          createdAt: "2024-04-10T15:27:49Z"
          description: Manages the installation and upgrade of the ClusterManager.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management-io/ocm
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterManager defines the configurations of controllers
              running hub that govern registration and work distribution for attached
              Klusterlets
            displayName: ClusterManager
            kind: ClusterManager
            name: clustermanagers.operator.open-cluster-management.io
            version: v1
        description: "Note: This operator will stop updating, please use the latest
          operator in OperatorHub.io https://operatorhub.io/operator/cluster-manager.
          \n\nThe Cluster Manager provides the multicluster hub, which can manage
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. This operator supports the installation and upgrade of ClusterManager.\n\n##
          How to Install\nInstall ClusterManager operator by following instructions
          in top right button `Install`. After installing the operator, create an
          instance of the ClusterManager resource to install the ClusterManager.\n\nThe
          pods of ClusterManager are created in the `open-cluster-management-hub`
          namespace:\n\n```\n$ kubectl get pods -n open-cluster-management-hub\nNAME
          \                                                      READY   STATUS    RESTARTS
          \  AGE\ncluster-manager-addon-manager-controller-7fcc67bfdf-x7hbz  1/1     Running
          \  0          61m\ncluster-manager-placement-controller-dbcfd4975-26jpw
          \      1/1     Running   0          61m\ncluster-manager-registration-controller-7949d97899-wwtg4
          \  1/1     Running   0          61m\ncluster-manager-registration-webhook-7f675bc4bb-v4rhj
          \     1/1     Running   0          61m\ncluster-manager-work-webhook-6c9d65fb95-jh97b
          \             1/1     Running   0          61m\n```\n\nThere is a ManagedCluster
          created in the Hub cluster when the managed cluster is registered to the
          Hub cluster.\n\n```\n$ kubectl get managedclusters.cluster.open-cluster-management.io\nNAME
          \                  HUB ACCEPTED   MANAGED CLUSTER URLS     JOINED   AVAILABLE
          \  AGE\ncluster-dev            true                                    True
          \    True        7h20m\ncluster-test           false          https://127.0.0.1:8443
          \                       7h25m\n```\n"
        displayName: Cluster Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open-cluster-management
        - ClusterManager
        links:
        - name: Registration Operator
          url: https://github.com/open-cluster-management-io/ocm
        maintainers:
        - email: zyin@redhat.com
          name: Zhiwei Yin
        - email: deads@redhat.com
          name: David Eads
        - email: jqiu@redhat.com
          name: Jian Qiu
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/open-cluster-management/registration-operator:v0.13.1
        version: 0.13.1
      entries:
      - name: cluster-manager.v0.13.1
        version: 0.13.1
      - name: cluster-manager.v0.13.0
        version: 0.13.0
      - name: cluster-manager.v0.12.0
        version: 0.12.0
      - name: cluster-manager.v0.11.0
        version: 0.11.0
      - name: cluster-manager.v0.10.0
        version: 0.10.0
      - name: cluster-manager.v0.9.1
        version: 0.9.1
      - name: cluster-manager.v0.8.0
        version: 0.8.0
      - name: cluster-manager.v0.7.0
        version: 0.7.0
      - name: cluster-manager.v0.6.0
        version: 0.6.0
      - name: cluster-manager.v0.5.0
        version: 0.5.0
      - name: cluster-manager.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: cluster-manager
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: cloudnative-pg
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudnative-pg.v1.27.0-rc1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: ghcr.io/cloudnative-pg/cloudnative-pg@sha256:c2212e806e950ed5b667455f476481aef179c8e467e435352b624dc670000443
          createdAt: "2025-07-31T12:14:29Z"
          description: CloudNativePG is an open source operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/cloudnative-pg/cloudnative-pg
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.cnpg.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.cnpg.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.cnpg.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.cnpg.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.cnpg.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.cnpg.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.cnpg.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnpg` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: CloudNativePG
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnpg
        links:
        - name: CloudNativePG
          url: https://cloudnative-pg.io/
        - name: Documentation
          url: https://cloudnative-pg.io/docs/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: The CloudNativePG Contributors
        relatedImages:
        - ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b
        - ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891
        - ghcr.io/cloudnative-pg/cloudnative-pg@sha256:c2212e806e950ed5b667455f476481aef179c8e467e435352b624dc670000443
        version: 1.27.0-rc1
      entries:
      - name: cloudnative-pg.v1.27.0-rc1
        version: 1.27.0-rc1
      - name: cloudnative-pg.v1.26.0-rc3
        version: 1.26.0-rc3
      - name: cloudnative-pg.v1.26.0-rc2
        version: 1.26.0-rc2
      - name: cloudnative-pg.v1.26.0-rc1
        version: 1.26.0-rc1
      name: candidate
    - currentCSV: cloudnative-pg.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: ghcr.io/cloudnative-pg/cloudnative-pg@sha256:53995e9d6a7760c77d77e5298f78b49029427494f4dd933d5fa7a74f7af3bb44
          createdAt: "2025-08-12T14:08:35Z"
          description: CloudNativePG is an open source operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/cloudnative-pg/cloudnative-pg
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.cnpg.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.cnpg.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.cnpg.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.cnpg.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.cnpg.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.cnpg.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.cnpg.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnpg` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: CloudNativePG
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnpg
        links:
        - name: CloudNativePG
          url: https://cloudnative-pg.io/
        - name: Documentation
          url: https://cloudnative-pg.io/docs/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - ghcr.io/cloudnative-pg/cloudnative-pg@sha256:53995e9d6a7760c77d77e5298f78b49029427494f4dd933d5fa7a74f7af3bb44
        - ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891
        - ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b
        version: 1.27.0
      entries:
      - name: cloudnative-pg.v1.27.0
        version: 1.27.0
      - name: cloudnative-pg.v1.26.1
        version: 1.26.1
      - name: cloudnative-pg.v1.26.0
        version: 1.26.0
      - name: cloudnative-pg.v1.25.1
        version: 1.25.1
      - name: cloudnative-pg.v1.25.0
        version: 1.25.0
      - name: cloudnative-pg.v1.24.1
        version: 1.24.1
      - name: cloudnative-pg.v1.24.0
        version: 1.24.0
      - name: cloudnative-pg.v1.23.3
        version: 1.23.3
      - name: cloudnative-pg.v1.23.2
        version: 1.23.2
      - name: cloudnative-pg.v1.23.1
        version: 1.23.1
      name: stable-v1
    defaultChannel: stable-v1
    packageName: cloudnative-pg
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cincinnati-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: update-service-operator.v5.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "updateservice.operator.openshift.io/v1",
                "kind": "UpdateService",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          description: Creates and maintains an OpenShift Update Service instance
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-update-service
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: UpdateService is the Schema for the updateservices API.
            displayName: Update Service
            kind: UpdateService
            name: updateservices.updateservice.operator.openshift.io
            version: v1
        description: |-
          # Use Case
          Running an Update Service instance in a cluster is appealing for offline OpenShift
          clusters or for admins that want to provide their own graph data instead of
          using [https://github.com/openshift/cincinnati-graph-data](https://github.com/openshift/cincinnati-graph-data/).

          # About Update Service
          Update Service uses **Cincinnati** which is an update protocol designed to facilitate
          automatic updates. It describes a particular method for representing transitions between
          releases of a project and allowing a client to perform automatic updates between these
          releases.

          ## Components
          A **Graph Builder** iterates over the release payloads hosted by the
          storage component and builds a DAG of the releases. It is responsible for
          verifying that the graph described by the releases is acyclic and connected.

          A **Policy Engine** is in charge of altering a client's view of the graph
          by applying a set of filters which are defined within the particular Policy
          Engine instance. Both the input to and the output from Policy Engines is a
          graph, allowing multiple Policy Engines to be chained together. The first
          Policy Engine in a chain will fetch its graph from the Graph Builder and the
          last Policy Engine in a chain will serve the modified graph to the client.

          An **Update Service client** is the end consumer of the release payloads. The
          client periodically queries the Policy Engine for updates and applys them if
          available.

          # Query OpenShift's Update Service Endpoint
          $ curl --silent --header 'Accept:application/json' 'https://api.openshift.com/api/upgrades_info/v1/graph?arch=amd64&channel=stable-4.2' | jq '. as $graph | $graph.nodes | map(.version == "4.2.13") | index(true) as $orig | $graph.edges | map(select(.[0] == $orig)[1]) | map($graph.nodes[.])'
        displayName: OpenShift Update Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Update Service
        - UpdateService
        - update-service
        - OSUS
        - Cincinnati
        - version
        - upgrade
        links:
        - name: Update Service Operator
          url: https://github.com/openshift/cincinnati-operator
        - name: OpenShift Update Service
          url: https://github.com/openshift/cincinnati
        - name: OpenShift Graph Data
          url: https://github.com/openshift/cincinnati-graph-data
        maintainers:
        - email: aos-team-ota@redhat.com
          name: OpenShift Update Service maintainers
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-update-service/openshift-update-service-rhel8-operator@sha256:759f4de107a5b429933c5e44046ffb4ed75c674aea75440513a6df8fa93a83a3
        - registry.redhat.io/openshift-update-service/openshift-update-service-rhel8@sha256:a9468a662babb272917df92bc9c827535b4b37883edb440b6d2d7c32ffb2145e
        version: 5.0.3
      entries:
      - name: update-service-operator.v5.0.3
        version: 5.0.3
      - name: update-service-operator.v5.0.2
        version: 5.0.2
      - name: update-service-operator.v5.0.1
        version: 5.0.1
      - name: update-service-operator.v5.0.0
        version: 5.0.0
      - name: update-service-operator.v4.9.1
        version: 4.9.1
      - name: update-service-operator.v4.9.0
        version: 4.9.0
      name: v1
    defaultChannel: v1
    packageName: cincinnati-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: www.redhat.com
    name: dpu-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dpu-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.openshift.io/v1",
                "kind": "DpuOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dpu-operator",
                    "app.kubernetes.io/instance": "dpuoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dpuoperatorconfig",
                    "app.kubernetes.io/part-of": "dpu-operator"
                  },
                  "name": "dpuoperatorconfig-sample"
                },
                "spec": {
                  "mode": "host"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-13T13:42:44Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: dpu-operator-system
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DpuOperatorConfig is the Schema for the dpuoperatorconfigs
              API
            displayName: Dpu Operator Config
            kind: DpuOperatorConfig
            name: dpuoperatorconfigs.config.openshift.io
            version: v1
        description: This operator will manage and configure data processing unit
          (DPUs) to be used in accelerating/offloading k8s networking functions
        displayName: DPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dpu
        - dpu-operator
        - ipu
        links:
        - name: Dpu Operator
          url: https://github.com/openshift/dpu-operator
        maintainers:
        - email: wizhao@redhat.com
          name: William Zhao
        - email: bnemeth@redhat.com
          name: Balazs Nemeth
        maturity: alpha
        provider:
          name: Red Hat
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-dpu-rhel9-operator@sha256:9d8283238af545c332781b2edef6b3a2857bcf4a264c9bae26720cc48e338a4b
        - registry.redhat.io/openshift4/ose-dpu-daemon-rhel9@sha256:a9a375f70dfb4a730aca0f7e58619fb05ef226261de6980d04eeffb8a043e6a4
        version: 4.16.0-202509240123
      entries:
      - name: dpu-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: dpu-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: dpu-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: dpu-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: dpu-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: dpu-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: dpu-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: dpu-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: dpu-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: dpu-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: dpu-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: dpu-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: dpu-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: dpu-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: dpu-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: dpu-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: dpu-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: dpu-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: dpu-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: dpu-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: dpu-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: dpu-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: dpu-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: dpu-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: dpu-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: dpu-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: dpu-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: dpu-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: dpu-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      name: stable
    defaultChannel: stable
    packageName: dpu-operator
    provider:
      name: Red Hat
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: odr-cluster-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odr-cluster-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "VolumeReplicationGroup",
                "metadata": {
                  "name": "volumereplicationgroup-sample"
                },
                "spec": {
                  "async": {
                    "schedulingInterval": "10m"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ],
                        "name": "cpd"
                      },
                      {
                        "includedResources": [
                          "Deployment"
                        ],
                        "name": "deployments"
                      },
                      {
                        "excludedResources": [
                          ""
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includeClusterResources": true,
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "cpd",
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      },
                      {
                        "backupName": "deployments",
                        "includedResources": [
                          "Deployment"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "excludedResources": [
                          "ConfigMap",
                          "Secret",
                          "Deployment",
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      }
                    ]
                  },
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  },
                  "replicationState": "primary",
                  "s3Profiles": [
                    "s3-profile-of-east",
                    "s3-profile-of-west"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: MaintenanceMode
            name: maintenancemodes.ramendr.openshift.io
            version: v1alpha1
          - kind: ProtectedVolumeReplicationGroupList
            name: protectedvolumereplicationgrouplists.ramendr.openshift.io
            version: v1alpha1
          - description: VolumeReplicationGroup is the Schema for the volumereplicationgroups
              API
            displayName: Volume Replication Group
            kind: VolumeReplicationGroup
            name: volumereplicationgroups.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Cluster is a disaster-recovery orchestrator for
          stateful applications, that operates from an Advanced Cluster Management
          (ACM) managed cluster and is controlled by Openshift DR Hub operator to
          orchestrate the life-cycle of an application, and its state on the managed
          cluster.
        displayName: Openshift DR Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:769acf15d016d4992793c9c527d7c268d2a67502dd7e0fadc240e4896398195b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 4.15.20-rhodf
      entries:
      - name: odr-cluster-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odr-cluster-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odr-cluster-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odr-cluster-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odr-cluster-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odr-cluster-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odr-cluster-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odr-cluster-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odr-cluster-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odr-cluster-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odr-cluster-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odr-cluster-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odr-cluster-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odr-cluster-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odr-cluster-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odr-cluster-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odr-cluster-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odr-cluster-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odr-cluster-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odr-cluster-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odr-cluster-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "VolumeReplicationGroup",
                "metadata": {
                  "name": "volumereplicationgroup-sample"
                },
                "spec": {
                  "async": {
                    "schedulingInterval": "10m"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ],
                        "name": "cpd"
                      },
                      {
                        "includedResources": [
                          "Deployment"
                        ],
                        "name": "deployments"
                      },
                      {
                        "excludedResources": [
                          ""
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includeClusterResources": true,
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "cpd",
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      },
                      {
                        "backupName": "deployments",
                        "includedResources": [
                          "Deployment"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "excludedResources": [
                          "ConfigMap",
                          "Secret",
                          "Deployment",
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      }
                    ]
                  },
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  },
                  "replicationState": "primary",
                  "s3Profiles": [
                    "s3-profile-of-east",
                    "s3-profile-of-west"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: MaintenanceMode
            name: maintenancemodes.ramendr.openshift.io
            version: v1alpha1
          - kind: ProtectedVolumeReplicationGroupList
            name: protectedvolumereplicationgrouplists.ramendr.openshift.io
            version: v1alpha1
          - description: VolumeReplicationGroup is the Schema for the volumereplicationgroups
              API
            displayName: Volume Replication Group
            kind: VolumeReplicationGroup
            name: volumereplicationgroups.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Cluster is a disaster-recovery orchestrator for
          stateful applications, that operates from an Advanced Cluster Management
          (ACM) managed cluster and is controlled by Openshift DR Hub operator to
          orchestrate the life-cycle of an application, and its state on the managed
          cluster.
        displayName: Openshift DR Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:af262b0890300183db1766bc392017b0bda723a9ba841dc25e3d928aa413db9a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        version: 4.16.16-rhodf
      entries:
      - name: odr-cluster-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odr-cluster-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odr-cluster-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odr-cluster-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odr-cluster-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odr-cluster-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odr-cluster-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odr-cluster-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odr-cluster-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odr-cluster-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odr-cluster-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odr-cluster-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odr-cluster-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odr-cluster-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odr-cluster-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odr-cluster-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odr-cluster-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odr-cluster-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Apache Software Foundation
      provider-url: ""
    name: camel-karavan-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: camel-karavan-operator.v3.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Karavan",
                "metadata": {
                  "labels": {
                    "app": "karavan"
                  },
                  "name": "karavan"
                },
                "spec": {
                  "instances": 1,
                  "auth": "public",
                  "environment": "dev",
                  "runtimes": "quarkus,spring-boot",
                  "gitPullInterval": "10s"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/apache/camel-karavan-operator:3.20.1
          repository: https://github.com/apache/camel-karavan
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Karavan
            name: karavans.camel.apache.org
            version: v1
        description: |
          Apache Camel Karavan
          ====================
          Karavan is an Integration Toolkit for Apache Camel, which makes integration easy and fun through the visualization of pipelines,
          integration with runtimes and package, image build and deploy to kubernetes out-of-the-box.

          ## Installation

          1. Install Apache Camel Karavan Operator
          2. Install Tekton operator
          3. Create namespace, ex: `karavan`
          4. Create `Secret` in the namespace, ex:
          ```
          kind: Secret
          apiVersion: v1
          metadata:
            name: karavan
            namespace: karavan
          type: Opaque
          stringData:
            master-password: karavan
            oidc-secret: secret
            oidc-server-url: https://hostname/auth/realms/karavan
            oidc-frontend-url: https://hostname/auth
            git-repository: https://github.com/repository/projects.git
            git-password: password
            git-username: username
            git-branch: main
            image-registry: image-registry.openshift-image-registry.svc:5000
          ```
          5. Create `Karavan` instance in the namespace, ex:
          ```
          apiVersion: camel.apache.org/v1
          kind: Karavan
          metadata:
            name: karavan
            namespace: karavan
          spec:
            instances: 1
            auth: public
            environment: dev
            runtimes: quarkus,spring-boot
            gitPullInterval: 10s
          ```
        displayName: Camel Karavan Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - camel
        - karavan
        - integration
        - microservices
        - low-code
        links:
        - name: Apache Camel Karavan source code repository
          url: https://github.com/apache/camel-karavan
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - ghcr.io/apache/camel-karavan-operator:3.20.1
        version: 3.20.1
      entries:
      - name: camel-karavan-operator.v3.20.1
        version: 3.20.1
      - name: camel-karavan-operator.v3.20.0
        version: 3.20.0
      - name: camel-karavan-operator.v3.18.5
        version: 3.18.5
      name: alpha
    defaultChannel: alpha
    packageName: camel-karavan-operator
    provider:
      name: The Apache Software Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Upbound Inc.
      provider-url: https://upbound.io
    name: universal-crossplane
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: universal-crossplane.1.5.1-up.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"pkg.crossplane.io/v1",
                "kind":"Configuration",
                "metadata":{
                  "name":"xp-getting-started-with-aws"
                },
                "spec":{
                  "package":"registry.upbound.io/xp/getting-started-with-aws:v1.2.2"
                }
              },
              {
                "apiVersion":"apiextensions.crossplane.io/v1",
                "kind":"CompositeResourceDefinition",
                "metadata":{
                  "name":"compositepostgresqlinstances.database.example.org"
                },
                "spec":{
                  "group":"database.example.org",
                  "names":{
                    "kind":"CompositePostgreSQLInstance",
                    "plural":"compositepostgresqlinstances"
                  },
                  "claimNames":{
                    "kind":"PostgreSQLInstance",
                    "plural":"postgresqlinstances"
                  },
                  "connectionSecretKeys":[
                      "username",
                      "password",
                      "endpoint",
                      "port"
                  ],
                  "versions":[
                    {
                      "name":"v1alpha1",
                      "served":true,
                      "referenceable":true,
                      "schema":{
                        "openAPIV3Schema":{
                          "type":"object",
                          "properties":{
                            "spec":{
                              "type":"object",
                              "properties":{
                                "parameters":{
                                  "type":"object",
                                  "properties":{
                                    "storageGB":{
                                      "type":"integer"
                                    }
                                  },
                                  "required":[
                                      "storageGB"
                                  ]
                                }
                              },
                              "required":[
                                  "parameters"
                              ]
                            }
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion":"apiextensions.crossplane.io/v1",
                "kind":"Composition",
                "metadata":{
                  "name":"compositepostgresqlinstances.aws.database.example.org",
                  "labels":{
                    "provider":"aws",
                    "guide":"quickstart",
                    "vpc":"default"
                  }
                },
                "spec":{
                  "writeConnectionSecretsToNamespace":"crossplane-system",
                  "compositeTypeRef":{
                    "apiVersion":"database.example.org/v1alpha1",
                    "kind":"CompositePostgreSQLInstance"
                  },
                  "resources":[
                    {
                      "name":"rdsinstance",
                      "base":{
                        "apiVersion":"database.aws.crossplane.io/v1beta1",
                        "kind":"RDSInstance",
                        "spec":{
                          "forProvider":{
                            "region":"us-east-1",
                            "dbInstanceClass":"db.t2.small",
                            "masterUsername":"masteruser",
                            "engine":"postgres",
                            "engineVersion":"9.6",
                            "skipFinalSnapshotBeforeDeletion":true,
                            "publiclyAccessible":true
                          },
                          "writeConnectionSecretToRef":{
                            "namespace":"crossplane-system"
                          }
                        }
                      },
                      "patches":[
                        {
                          "fromFieldPath":"metadata.uid",
                          "toFieldPath":"spec.writeConnectionSecretToRef.name",
                          "transforms":[
                            {
                              "type":"string",
                              "string":{
                                "fmt":"%s-postgresql"
                              }
                            }
                          ]
                        },
                        {
                          "fromFieldPath":"spec.parameters.storageGB",
                          "toFieldPath":"spec.forProvider.allocatedStorage"
                        }
                      ],
                      "connectionDetails":[
                        {
                          "fromConnectionSecretKey":"username"
                        },
                        {
                          "fromConnectionSecretKey":"password"
                        },
                        {
                          "fromConnectionSecretKey":"endpoint"
                        },
                        {
                          "fromConnectionSecretKey":"port"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion":"pkg.crossplane.io/v1",
                "kind":"Provider",
                "metadata":{
                  "name":"provider-aws"
                },
                "spec":{
                  "package":"crossplane/provider-aws:master"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime, Cloud Provider, Database, Integration &
            Delivery, Networking, Storage
          com.redhat.component: universal-crossplane
          com.redhat.delivery.backport: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.6
          containerImage: upbound/uxp-bootstrapper
          createdAt: 2021-05-18T00:00:00+0000
          description: Upbound Universal Crossplane (UXP) is Upbound's official enterprise-grade
            distribution of Crossplane.
          io.k8s.display-name: universal-crossplane
          io.openshift.tags: uxp,crossplane,upbound
          maintainer: Upbound Inc. <info@upbound.io>
          name: universal-crossplane
          ocs.tags: v4.6
          operatorframework.io/initialization-resource: |
            apiVersion: v1
            kind: Secret
            metadata:
              name: upbound-control-plane-token
              namespace: upbound-system
            type: Opaque
            data:
              token: "Output of `up cloud controlplane attach | base64` command"
          operatorframework.io/suggested-namespace: upbound-system
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io.bundle.channels.v1: stable
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: universal-crossplane
          operators.operatorframework.io/internal-objects: '["locks.pkg.crossplane.io"]'
          repository: https://github.com/upbound/universal-crossplane
          summary: Upbound Universal Crossplane (UXP) is Upbound's official enterprise-grade
            distribution of Crossplane.
          support: Upbound Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: CompositeResourceDefinition
            kind: CompositeResourceDefinition
            name: compositeresourcedefinitions.apiextensions.crossplane.io
            version: v1
          - displayName: CompositionRevision
            kind: CompositionRevision
            name: compositionrevisions.apiextensions.crossplane.io
            version: v1alpha1
          - displayName: Composition
            kind: Composition
            name: compositions.apiextensions.crossplane.io
            version: v1
          - displayName: ConfigurationRevision
            kind: ConfigurationRevision
            name: configurationrevisions.pkg.crossplane.io
            version: v1
          - displayName: Configuration
            kind: Configuration
            name: configurations.pkg.crossplane.io
            version: v1
          - displayName: ControllerConfig
            kind: ControllerConfig
            name: controllerconfigs.pkg.crossplane.io
            version: v1alpha1
          - displayName: Lock
            kind: Lock
            name: locks.pkg.crossplane.io
            version: v1alpha1
          - displayName: ProviderRevision
            kind: ProviderRevision
            name: providerrevisions.pkg.crossplane.io
            version: v1
          - displayName: Provider
            kind: Provider
            name: providers.pkg.crossplane.io
            version: v1
        description: |
          ## Overview

          Upbound Universal Crossplane (UXP) is [Upbound's][upbound] official
          enterprise-grade distribution of [Crossplane][crossplane]. It's fully compatible
          with upstream Crossplane, open source, capable of connecting to Upbound Cloud
          for real-time dashboard visibility, and maintained by Upbound. It's the easiest
          way for both individual community members and enterprises to build their
          production control planes.

          ## Quick Start

          1. Install the [Upbound CLI][upbound-cli].
              ```console
              curl -sL https://cli.upbound.io | sh
              # Install with Homebrew
              brew install upbound/tap/up
              ```

          2. [Create an Upbound account][create-account] for a free dashboard for UXP.

          3. Connect UXP to Upbound Cloud.
              ```console
              # The name of your new UXP control plane.
              UXP_NAME=mycrossplane

              up cloud login
              up cloud controlplane attach ${UXP_NAME} | up uxp connect -
              ```

          4. Manage your UXP control plane by [signing in][login] to your Upbound account.

          ## Additional Resources

          - The [UXP Documentation][uxp-documentation] provides additional information
            about UXP and resources for developers, like examples.
          - The [developer guide][developer-guide] describes how to build and run UXP
            locally from source.
          - [UXP Slack][uxp-slack] is where you can go to get all of your UXP questions
            answered.

          [upbound]: https://upbound.io
          [crossplane]: https://crossplane.io/
          [upbound-cli]: https://github.com/upbound/up
          [create-account]: https://cloud.upbound.io/register
          [login]: https://cloud.upbound.io/login
          [uxp-documentation]: https://cloud.upbound.io/docs/uxp
          [developer-guide]: https://github.com/upbound/universal-crossplane/blob/main/docs/developer-guide.md
          [uxp-slack]: https://crossplane.slack.com/archives/upbound
        displayName: Upbound Universal Crossplane (UXP)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - infrastructure
        - services
        - application
        - database
        - cache
        - bucket
        - infra
        - app
        - ops
        - oam
        - gcp
        - azure
        - aws
        - alibaba
        - cloudsql
        - rds
        - s3
        - azuredatabase
        - asparadb
        - gke
        - aks
        - eks
        links:
        - name: GitHub
          url: https://github.com/upbound/universal-crossplane
        - name: Website
          url: https://upbound.io
        - name: Twitter
          url: https://twitter.com/upbound_io
        - name: Slack
          url: https://slack.crossplane.io/
        maintainers:
        - email: info@upbound.io
          name: Upbound Inc.
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Upbound Inc.
          url: https://upbound.io
        relatedImages:
        - upbound/xgql:v0.1.5
        - upbound/uxp-bootstrapper:v1.5.1-up.1
        - upbound/crossplane:v1.5.1-up.1
        version: 1.5.1-up.1
      entries:
      - name: universal-crossplane.1.5.1-up.1
        version: 1.5.1-up.1
      - name: universal-crossplane.1.4.3-up.1
        version: 1.4.3-up.1
      - name: universal-crossplane.1.3.3-up.1
        version: 1.3.3-up.1
      - name: universal-crossplane.1.3.1-up.1
        version: 1.3.1-up.1
      - name: universal-crossplane.1.2.3-up.1
        version: 1.2.3-up.1
      - name: universal-crossplane.1.2.2-up.1
        version: 1.2.2-up.1
      - name: universal-crossplane.1.2.1-up.4
        version: 1.2.1-up.4
      - name: universal-crossplane.1.2.1-up.3
        version: 1.2.1-up.3
      name: stable
    defaultChannel: stable
    packageName: universal-crossplane
    provider:
      name: Upbound Inc.
      url: https://upbound.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Palo Alto Networks
      provider-url: ""
    name: pcc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pcc-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "name": "pcc-console",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "ConsoleDefender",
                "metadata": {
                  "name": "pcc-consoledefender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "defenderConfig": {
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Defender",
                "metadata": {
                  "name": "pcc-defender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "defenderConfig": {
                    "clusterAddress": "twistlock-console.example.com",
                    "consoleAddress": "https://twistlock-console.example.com:8083",
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: quay.io/prismacloud/pcc-operator@sha256:51a343d5ea0115e3c43ccaa546492e85ab74cf2cf2658d725c1ea7099fc89191
          description: Deploy Prisma Cloud Compute for cloud-native security in your
            clusters
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator
          support: pcc-integrations@paloaltonetworks.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Install the Prisma Cloud Compute Console and Defenders
            displayName: Console and Defenders
            kind: ConsoleDefender
            name: consoledefenders.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install the Prisma Cloud Compute Console
            displayName: Console
            kind: Console
            name: consoles.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install Prisma Cloud Compute Defenders
            displayName: Defenders
            kind: Defender
            name: defenders.pcc.paloaltonetworks.com
            version: v1alpha1
        description: |
          ## Features
          - Deploy Console
              - Create initial local administrator user
              - Add license
          - Deploy Defenders
          - Support for offline deployments
          - Upgrade Console
          - Upgrade Defenders

          ## Installation
          See the [Prisma Cloud Compute Operator documentation](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator) for installation guides.

          ## Support
          This operator is available "as is," and should be seen as community-supported; however, Palo Alto Networks will continue contributing whenever possible.
          Palo Alto Networks does not provide technical support or help with using or troubleshooting the operator through normal support channels.
          We encourage you to open GitHub [issues](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/issues) and [pull requests](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/pulls) to track bugs and feature requests.

          ## Other links
          [Prisma Cloud Compute product documentation](https://docs.paloaltonetworks.com/prisma/prisma-cloud/prisma-cloud-admin-compute.html)

          [Prisma Cloud Compute API documentation](https://prisma.pan.dev/api/cloud/cwpp)
        displayName: Prisma Cloud Compute Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - palo
        - palo alto
        - palo alto networks
        - paloalto
        - twistlock
        - compute
        - prisma
        - prisma cloud
        - prisma cloud compute
        - security
        - image
        - images
        - container
        - containers
        - docker
        - podman
        - vulnerability
        - vulnerabilities
        - compliance
        - runtime
        - scan
        - scanning
        - monitor
        - monitoring
        - alert
        - alerting
        links:
        - name: Prisma Cloud
          url: https://www.paloaltonetworks.com/prisma/cloud
        maintainers:
        - email: pcc-integrations@paloaltonetworks.com
          name: Wyatt Gill
        maturity: alpha
        provider:
          name: Palo Alto Networks
        relatedImages:
        - quay.io/prismacloud/pcc-operator@sha256:51a343d5ea0115e3c43ccaa546492e85ab74cf2cf2658d725c1ea7099fc89191
        version: 0.2.3
      entries:
      - name: pcc-operator.v0.2.3
        version: 0.2.3
      name: stable
    defaultChannel: stable
    packageName: pcc-operator
    provider:
      name: Palo Alto Networks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-iam-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-iam-operator.v4.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "Authentication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-iam-operator",
                    "app.kubernetes.io/managed-by": "ibm-iam-operator",
                    "app.kubernetes.io/name": "ibm-iam-operator"
                  },
                  "name": "example-authentication"
                },
                "spec": {
                  "auditService": {
                    "imageName": "auditService no longer used - ignore",
                    "imageRegistry": "auditService no longer used - ignore",
                    "imageTag": "auditService no longer used - ignore"
                  },
                  "authService": {
                    "imageName": "icp-platform-auth",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "ldapsCACert": "platform-auth-ldaps-ca-cert",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "650Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "400Mi",
                        "memory": "350Mi"
                      }
                    },
                    "routerCertSecret": "saml-auth-secret"
                  },
                  "clientRegistration": {
                    "imageName": "im-initcontainer",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "158Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "config": {
                    "authUniqueHosts": "internal-ip1 internal-ip2 mycluster.icp",
                    "bootstrapUserId": "kubeadmin",
                    "claimsMap": "name=\"givenName\" family_name=\"givenName\" given_name=\"givenName\" preferred_username=\"displayName\" display_name=\"displayName\"",
                    "claimsSupported": "name,family_name,display_name,given_name,preferred_username",
                    "clusterCADomain": "mycluster.icp",
                    "clusterExternalAddress": "10.0.0.1",
                    "clusterInternalAddress": "10.0.0.1",
                    "clusterName": "mycluster",
                    "defaultAdminPassword": "password",
                    "defaultAdminUser": "cpadmin",
                    "defaultLogin": "",
                    "enableImpersonation": false,
                    "fipsEnabled": true,
                    "ibmCloudSaas": false,
                    "icpPort": 8443,
                    "installType": "fresh",
                    "isOpenshiftEnv": true,
                    "nonceEnabled": true,
                    "oidcIssuerURL": "https://127.0.0.1:443/idauth/oidc/endpoint/OP",
                    "openshiftPort": 443,
                    "preferredLogin": "",
                    "providerIssuerURL": "",
                    "roksEnabled": true,
                    "roksURL": "https://roks.domain.name:443",
                    "roksUserPrefix": "changeme",
                    "saasClientRedirectUrl": "",
                    "scimAdminPassword": "scimpassword",
                    "scimAdminUser": "scimadmin",
                    "scopeClaim": "profile=\"name,family_name,display_name,given_name,preferred_username\"",
                    "wlpClientID": "4444be3a738841016ab76d71b650e836",
                    "wlpClientRegistrationSecret": "f1362ca4d20b8389af2d1ea68042c9af",
                    "wlpClientSecret": "aa73bf39752053bf723d1143fb4cf8a2",
                    "xframeDomain": "",
                    "zenFrontDoor": false
                  },
                  "identityManager": {
                    "imageName": "icp-identity-manager",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "masterNodesList": "10.0.0.1",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "550Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "ephemeral-storage": "300Mi",
                        "memory": "150Mi"
                      }
                    }
                  },
                  "identityProvider": {
                    "imageName": "icp-identity-provider",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "550Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "ephemeral-storage": "300Mi",
                        "memory": "150Mi"
                      }
                    }
                  },
                  "initMongodb": {
                    "imageName": "im-initcontainer",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "178Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "operatorVersion": "0.14.1",
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
          createdAt: "2025-02-20T00:03:25Z"
          description: The IAM operator provides a simple Kubernetes CRD-Based API
            to manage the lifecycle of IAM services. With this operator, you can simply
            deploy and upgrade the IAM services
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.11.0
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-iam-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Authentication is the Schema for the authentications API
            displayName: Authentication
            kind: Authentication
            name: authentications.operator.ibm.com
            version: v1alpha1
          - description: Client is the Schema for the clients API
            displayName: Client
            kind: Client
            name: clients.oidc.security.ibm.com
            version: v1
        description: |-
          The `ibm-iam-operator` installs the IBM Cloud Platform Common Services Identity and access management (IAM) service.

          **Important:** Do not install this operator directly. Install this operator only by using the IBM Common Service Operator. For more information about installing this operator and other Common Services operators, see [Installer documentation](http://ibm.biz/cpcs_opinstall). Additionally, you can exit this panel and navigate to the IBM Common Services tile in OperatorHub to learn more about the operator. If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For more information about IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).

          You can use the `ibm-iam-operator` to install the authentication and authorization services for the IBM Cloud Platform Common Services.

          With these services, you can configure security for IBM Cloud Platform Common Services, IBM Certified Containers (IBM products), or IBM Cloud Paks that are installed.

          For more information about the available IBM Cloud Platform Common Services, see the [IBM Knowledge Center](http://ibm.biz/cpcsdocs).

          ## Supported platforms

           - Red Hat OpenShift Container Platform 4.3 or newer installed on one of the following platforms:

              - Linux x86_64
              - Linux on Power (ppc64le)
              - Linux on IBM Z and LinuxONE

          ## Prerequisites

          Before you install this operator, you need to first install the operator dependencies and prerequisites:

          - For the list of operator dependencies, see the IBM Knowledge Center [Common Services dependencies documentation](http://ibm.biz/cpcs_opdependencies).

          - For the list of prerequisites for installing the operator, see the IBM Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).

          ## Documentation

          To install the operator by using the IBM Common Services Operator, follow the installation and configuration instructions that are in the IBM Knowledge Center.

          - If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).
          - If you are using the operator with an IBM Containerized Software, see the IBM Cloud Platform Common Services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall).
        displayName: IBM IM Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:52ce08ac2bce7c02dc30e2c0aa74df4bde8f156debccc59960f02b4b0040ffea
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:97b9ecef8caef156a0eb073b6eb2a34b63209b44511f5694a243f11285d88662
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:0f46b288d060a15a48ecad8f328a8046bf232b8629c3d79da6a82ea712b363f4
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:7cec55c00827b747f23c2f1703d99e1ca5b30a0c445f8ad0471b25b0c9cca57c
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:b8886b9b54e6ee4c6b2981ef092e54348a0286577d9924cf136cee881f5bf1f4
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-bundle@sha256:98cf2a6c7b7af5d7cbfbc61ce140201272b30b12edabaec2b2d4287b0690d89e
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:f289ac0d44803ff12d41c3659e6fdf04c404e0b3b233b6bc44eec671abb1b47e
        version: 4.11.0
      entries:
      - name: ibm-iam-operator.v4.11.0
        version: 4.11.0
      name: v4.10
    defaultChannel: v4.10
    packageName: ibm-iam-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: vfunction-operator
      app.kubernetes.io/managed-by: operator
      app.kubernetes.io/name: vfunction
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      name: vfunction-server-operator
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: vFunction
      provider-url: ""
    name: vfunction-server-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vfunction-server-operator.v4.3.1762
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vfunction.com/v1",
                "kind": "VfunctionServer",
                "metadata": {
                  "name": "vfunction-server",
                  "namespace": "vfunction",
                  "labels": {
                      "name": "vfunction",
                      "app.kubernetes.io/name": "vfunction",
                      "app.kubernetes.io/instance": "vfunction-server",
                      "app.kubernetes.io/managed-by": "operator"
                  }
                },
                "spec": {
                  "admin": {
                    "email": "admin@mycompany.com",
                    "name": "Admin",
                    "password": "Password1!"
                  },
                  "host": "http://my.domain.com",
                  "org_name": "MyCompany",
                  "upgrade": "Never",
                  "measurements": {
                    "min_number_of_services": "2",
                    "max_number_of_services": "10",
                    "max_pod_memory_capacity": "16G",
                    "max_pod_cpu_capacity": "1",
                    "pod_ephemeral_storage": "1G",
                    "S3": {
                      "bucket": "",
                      "key": "",
                      "secret": "",
                      "region": ""
                    }
                  },
                  "smtp": {
                    "identity": "",
                    "password": "",
                    "url": "",
                    "user": ""
                  },
                  "authentication": {
                    "authority": "",
                    "client_id": "",
                    "client_secret": "",
                    "jwks_url": "",
                    "issuer": "",
                    "ca_root_crt": ""
                  },
                  "tls": {
                    "use_letsencrypt": "No",
                    "crt": "",
                    "key": ""
                  },
                  "nginx": {
                    "force_http": "No",
                    "ipv6_disabled": "No",
                    "service_type": ""
                  },
                  "route": {
                    "use_route": "Yes",
                    "tls_termination": "edge",
                    "use_tls_certificate": "No"
                  },
                  "mysql": {
                    "external_mysql_ip": "",
                    "external_mysql_user": "",
                    "external_mysql_password": "",
                    "max_pod_memory_capacity": "4G",
                    "max_pod_cpu_capacity": "1",
                    "pod_ephemeral_storage": "1G"
                  },
                  "storage": {
                    "storage_class": "",
                    "mysql_persistent_volume": "",
                    "mysql_persistent_volume_claim": "",
                    "mysql_persistent_volume_claim_access_mode": "ReadWriteOnce",
                    "storage_persistent_volume": "",
                    "storage_persistent_volume_claim": "",
                    "storage_persistent_volume_claim_access_mode": "ReadWriteOnce",
                    "lets_encrypt_persistent_volume": "",
                    "lets_encrypt_persistent_volume_claim": "",
                    "lets_encrypt_persistent_volume_claim_access_mode": "ReadWriteOnce"
                  },
                  "distributed": {
                    "otlp": {
                      "min_number_of_services": "1",
                      "max_number_of_services": "10",
                      "max_pod_memory_capacity": "8G",
                      "max_pod_cpu_capacity": "1",
                      "pod_ephemeral_storage": "250M",
                      "max_folder_size_in_mb": "200"
                    },
                    "dm": {
                      "min_number_of_services": "1",
                      "max_number_of_services": "10",
                      "max_pod_memory_capacity": "8G",
                      "max_pod_cpu_capacity": "1",
                      "pod_ephemeral_storage": "1G"
                    }
                  },
                  "backup": {
                    "mysql": {
                      "restore_during_installation": "No",
                      "S3": {
                        "bucket": "",
                        "key": "",
                        "secret": "",
                        "region": ""
                      }
                    }
                  },
                  "security": {
                    "image_pull_policy": "",
                    "force_hashed_images": "No",
                    "set_pod_resources": "Yes",
                    "proxy": {
                      "http_proxy": "",
                      "https_proxy": "",
                      "additional_no_proxy": ""
                    },
                    "use_rate_limiting": "No",
                    "requests_per_second": "10"
                  },
                  "affinity": {
                    "node_affinity": [
                      {
                        "key": "",
                        "operator": "",
                        "values": "",
                        "services": ""
                      }
                    ],
                    "node_selectors": [
                      {
                        "key": "",
                        "value": "",
                        "services": ""
                      }
                    ],
                    "additional_labels": [
                      {
                        "key": "",
                        "value": "",
                        "services": ""
                      }
                    ],
                    "additional_annotations": [
                      {
                        "key": "",
                        "value": "",
                        "services": ""
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, AI/Machine Learning
          certified: Red Hat certified
          containerImage: registry.connect.redhat.com/vfunction/vfunction-server-operator@sha256:a5a0f47f7fb9219139f8d749da14950a8c39698b340644011485dfbec54ebe8b
          createdAt: "2025-08-04T16:10:32Z"
          description: vFunction is a cutting-edge code analysis, machine learning,
            and automation to boost your Java modernization projects.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.3.1762
          support: vFunction
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of vFunction running on your cluster.
            displayName: vFunction
            kind: VfunctionServer
            name: vfunctionservers.vfunction.com
            version: v1
        description: |2-

          ## Introduction

          [vFunction](https://www.vfunction.com/) is a cloud-native modernization platform that combines dynamic and static code analysis, machine learning, and automation to automatically identify and extract services from existing applications. vFunction is the only platform purpose-built for modernization of Java applications.

          The vFunction modernization process starts by learning the running monolithic application, and surfacing the interdependencies within it. Using AI, the platform analyzes and identifies services that can be separated from the application. This decomposition can present a range of micro, mini, or even macro services, depending on your application environment, each being an independently deployable and scalable application component.

          vFunction automates the extraction of these services, enabling you to modernize your monolith, quickly and easily.

          ## The vFunction platform

          The platform consists of 3 basic components; the server, the controller package, and a tools package. The server runs as an operand on an OpenShift environment. The controller package is installed on the machine that runs the monolithic application which can be either a Linux or a Windows machine, and the tools are run on a development machine, with access to the code of the monolithic application.

          The controller package consists of three elements: the vFunction agent, that collects data during the dynamic analysis phase; the vFunction Viper application, that performs static analysis on the binaries of the application; and the vFunction controller that handles all the communication between the agent, Viper, and the vFunction server.

          The vFunction agent is a mix of a Java and native agent, and needs to run on the JVM that is currently running your application. [Refer to the vFunction OOB Support Matrix document](https://drive.google.com/file/d/1ccq8LFab1FrYAimDUxwgjiCOdk4QuVzs/view) for a list of supported application servers and JVMs.

          This operator installs a vFunction server instance to be connected later with one or more vFunction controllers installed on your application machine(s).

          ## Before You Start

          The vFunction operator requires a Red Hat OpenShift Kubernetes Application Platform 4.x.

          The cluster recommended configuration (which is the same as the minimal one) should consist of:

          - *1 master node*

          - *1 worker node*

          - *Storage capacity enough for dynamically provision 2 PVs of 50G each*

          - *A default StorageClass configured*

          vFunction doesn’t provide any encryption mechanism for data stored on any attached storage. If required, encryption can be achieved by the client by using its own external means on the storage itself.

          vFunction supports backup of all critical data as part of the application. Nevertheless, for complete backup of entire data it is recommended that the client apply its own storage based backup mechanism.

          ## Install Prerequisites

          ###Cluster:
          The operator was certified and tested on OCP 4.6 - 4.17.
          If you encounter any issue with other OCP versions or cloud providers please contact info@vfunction.com.

          ###Storage:
          The operator creates two new PersistentVolumeClaims (PVCs) during installation time. For their creation, it relies on the default StorageClass to dynamically provision two PersistentVolumes (PVs). The system must have a default StorageClass in place.
          Both PersistentVolumes are defaulted to access with ReadWriteOnce mode, but you can change this by specifying a different access mode in the YAML file.
          Notes:
          - There is an option to specify a specific StorageClass for the PersistentVolumeClaims.
          - There is an option to specify the use of a specific pre-created PersistentVolume for each PersistentVolumeClaim
          - There is an option to specify a pre-created PersistentVolumeClaim for each Deployment.
          - There is an option to change the access mode for each PersistentVolumeClaim.

          ###Project:
          The operator should be installed in a new and dedicated project (namespace). If you intend to install more than one vFunction operands in the same cluster, please use different projects for each one. Installation of more than one operand on the same cluster will allow scaling by load balancing multiple applications to different vFunction servers.

          ## Installation instructions

          1. Prepare the YAML file to use in the installation:

              a. Copy the following YAML template into a text editor:

              ```yaml
                  apiVersion: vfunction.com/v1
                  kind: VfunctionServer
                  metadata:
                    name: vfunction
                    namespace: vfunction
                    labels:
                      name: vfunction
                      app.kubernetes.io/name: vfunction
                      app.kubernetes.io/instance: vfunction-server
                      app.kubernetes.io/managed-by: operator
                  spec:
                    host: "http://my.domain.com"
                    org_name: "MyCompany"
                    upgrade: "Never"
                    admin:
                      email: "admin@mycompany.com"
                      name: "Admin"
                      password: "Password1!"
                    measurements:
                      min_number_of_services: "2"
                      max_number_of_services: "10"
                      max_pod_memory_capacity: "16G"
                      max_pod_cpu_capacity: "1"
                      S3:
                        bucket: ""
                        key: ""
                        secret: ""
                    smtp:
                      password: ""
                      url: ""
                      identity: ""
                      user: ""
                    authentication:
                      authority: ""
                      client_id: ""
                      client_secret: ""
                      jwks_url: ""
                      issuer: ""
                      ca_root_crt: |
                        -----BEGIN CERTIFICATE-----
                        ...
                        -----END CERTIFICATE-----
                    tls:
                      use_letsencrypt: "No"
                      crt: |
                        -----BEGIN CERTIFICATE-----
                        ...
                        -----END CERTIFICATE-----
                      key: |
                        -----BEGIN PRIVATE KEY-----
                        ...
                        -----END PRIVATE KEY-----
                    nginx:
                      force_http: "No"
                      ipv6_disabled: "No"
                      service_type: ""
                    route:
                      use_route: "Yes"
                      tls_termination: "edge"
                      use_tls_certificate: "No"
                    mysql:
                      external_mysql_ip: ""
                      external_mysql_user: ""
                      external_mysql_password: ""
                      max_pod_memory_capacity: "4G"
                      max_pod_cpu_capacity: "1"
                    storage:
                      storage_class: ""
                      mysql_persistent_volume: ""
                      mysql_persistent_volume_claim: ""
                      mysql_persistent_volume_claim_access_mode: "ReadWriteOnce"
                      storage_persistent_volume: ""
                      storage_persistent_volume_claim: ""
                      storage_persistent_volume_claim_access_mode: "ReadWriteOnce"
                      lets_encrypt_persistent_volume: ""
                      lets_encrypt_persistent_volume_claim: ""
                      lets_encrypt_persistent_volume_claim_access_mode: "ReadWriteOnce"
                    distributed:
                        otlp:
                          min_number_of_services: "1"
                          max_number_of_services: "10"
                          max_pod_memory_capacity: "8G"
                          max_pod_cpu_capacity: "1"
                          max_folder_size_in_mb: "200"
                        dm:
                          min_number_of_services: "1"
                          max_number_of_services: "10"
                          max_pod_memory_capacity: "8G"
                          max_pod_cpu_capacity: "1"
                    backup:
                      mysql:
                        restore_during_installation: "No"
                        S3:
                          bucket: ""
                          key: ""
                          secret: ""
                    security:
                      image_pull_policy: ""
                      force_hashed_images: "No"
                      set_pod_resources: "Yes"
                      proxy:
                        http_proxy: ""
                        https_proxy: ""
                        additional_no_proxy: ""
                      use_rate_limiting: "No"
                      requests_per_second: "10"
                    affinity:
                      node_affinity:
                        - key: ""
                          operator: ""
                          values: ""
                          services: ""
                      node_selectors:
                        - key: ""
                          value: ""
                          services: ""
                      additional_labels:
                        - key: ""
                          value: ""
                          services: ""
                      additional_annotations:
                        - key: ""
                          value: ""
                          services: ""
              ```
              b. Customize the template fields by replacing with your information:

              - **host**: Enter the FQDN for accessing the vFunction dashboard.
                          Make sure that you use a domain name and not an IP address. Writing "https://" will mandate TLS while writing "http://" will expose the server through HTTP.
              - **org_name**: Enter your organization name.
              - **upgrade**: Choose auto-upgrade mode.
                  - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically.
                  - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
                  - **Never**: The automatic upgrading mechanism is off (default).
                  - **Maintenance**: Put the operator in a maintenance mode.
              - **admin.email**: Enter the email address of your administrator.
              - **admin.name**: Enter the name of your administrator.
              - **admin.password**: Enter the password you want to use for the vFunction administrator.
                                    The password should be at least 8 characters long, and consist of at least one lowercase letter, at least one uppercase letter, at least one number, and at least one special character.
              - **measurements.min\_num\_of\_services**: Minimum Number of measurement service pods.
              - **measurements.max\_num\_of\_services**: Maximum Number of measurement service pods.
              - **measurements.max\_pod\_memory\_capacity**: Maximum memory capacity limit for measurement service pods. (Default is "8G").
              - **measurements.max\_pod\_cpu\_capacity**: Maximum CPU capacity limit for measurement service pods. (Default is "1").
              - **measurements.pod\_ephemeral\_storage**: Ephemeral Storage Capacity for measurement service pods. (Default is "1G").
              - **measurements.S3**: Save all measurements data to S3, instead of local PV (optional).
                  - **measurements.S3.bucket**: Measurements S3 bucket name.
                  - **measurements.S3.key**: Measurements S3 key.
                  - **measurements.S3.secret**: Measurements S3 secret.
                  - **measurements.S3.region**: Measurements S3 region.
              - **smtp.user**: Enter the email address for a designated user for the SMTP server (optional).
              - **smtp.password**: Enter this user’s password (optional).
              - **smtp.identity**: Enter this user’s password (optional).
              - **smtp.url**: Enter the SMTP server URL (optional).
              - **authentication.authority**: Authority for authentication (Optional).
              - **authentication.client_id**: Client ID for authentication (Optional).
              - **authentication.client_secret**: Client Secret for authentication (Optional).
              - **authentication.jwks_url**: JWKS URL for authentication (Optional).
              - **authentication.issuer**: Issuer for authentication (Optional).
              - **authentication.ca\_root\_crt**: CA root certificate for authentication (Optional).
              - **tls.use_letsencrypt**: If your host FQDN starts with “https”, choose whether to use Let’s Encrypt to obtain certificates.
              - **tls.crt**: If your host FQDN starts with “https”, paste in the certifications you have for using the TLS connection (Applicable only if you don’t use Let’s Encrypt).
              - **tls.key**: If your host FQDN starts with “https”, enter the key you have for using the TLS connection (Applicable only if you don’t use Let’s Encrypt).
              - **nginx.force_http**: Force internal NGINX to listen on http even if the host FQDN starts with “https”.
                                      This is useful if you want to use a TLS termination proxy in front of vFunction, other then the route provided by the operator.
              - **nginx.ipv6_disabled**: Disable IPv6 on internal NGINX.
              - **nginx.service_type**: Set the service type for the internal NGINX service - ClusterIP, NodePort or LoadBalancer.
              - **route.use_route**: Use OpenShift route to expose vFunction. (Default is "Yes").
              - **route.tls_termination**: Set the TLS termination type for the OpenShift route - edge, passthrough or reencrypt. (Default is "edge").
              - **route.use\_tls\_certificate**: Use TLS certificate for OpenShift route (as provided in tls.crt and tls.key) - instead of the default one defined for the route controller in your cluster. (Default is "No").
              - **mysql.external\_mysql\_ip**: If you want to use an external MySQL database, enter its IP address.
              - **mysql.external\_mysql\_user**: If you want to use an external MySQL database, enter its user name.
              - **mysql.external\_mysql\_password**: If you want to use an external MySQL database, enter its password.
              - **mysql.max\_pod\_memory\_capacity**: Maximum memory capacity limit for MySQL pod. (Default is "2G").
              - **mysql.max\_pod\_cpu\_capacity**: Maximum CPU capacity limit for MySQL pod. (Default is "1").
              - **mysql.pod\_ephemeral\_storage**: Ephemeral Storage Capacity for MySQL pod. (Default is "1G").
              - **storage.storage\_class**: If you want to use a specific StorageClass for all PVCs, other than the default one, enter its name.
              - **storage.mysql\_persistent\_volume**: If you want to use a specific PV for mysql PVC, other than the dynamically provisioned one, enter its name.
              - **storage.mysql\_persistent\_volume\_claim**: If you want to use a specific pre-created PVC for mysql deployment, other than the dynamically provisioned one, enter its name.
              - **storage.mysql\_persistent\_volume\_claim\access\_mode**: If you want to use a specific access mode for mysql PVC, other than the default one, enter its name. (Default is "ReadWriteOnce").
              - **storage.storage\_persistent\_volume**: If you want to use a specific PV for storage PVC, other than the dynamically provisioned one, enter its name.
              - **storage.storage\_persistent\_volume\_claim**: If you want to use a specific pre-created PVC for storage deployment, other than the dynamically provisioned one, enter its name.
              - **storage.storage\_persistent\_volume\_claim\access\_mode**: If you want to use a specific access mode for storage PVC, other than the default one, enter its name. (Default is "ReadWriteOnce").
              - **storage.lets\_encrypt\_persistent\_volume**: If you want to use a specific PV for lets encrypt PVC, other than the dynamically provisioned one, enter its name.
              - **storage.lets\_encrypt\_persistent\_volume\_claim**: If you want to use a specific pre-created PVC for lets encrypt deployment, other than the dynamically provisioned one, enter its name.
              - **storage.lets\_encrypt\_persistent\_volume\_claim\access\_mode**: If you want to use a specific access mode for lets encrypt PVC, other than the default one, enter its name. (Default is "ReadWriteOnce").
              - **distributed.otlp.min\_number\_of\_services**: Minimum Number of OTLP pods.
              - **distributed.otlp.max\_number\_of\_services**: Maximum Number of OTLP pods.
              - **distributed.otlp.max\_pod\_memory\_capacity**: Maximum memory capacity for OTLP pods. (Default is "1G").
              - **distributed.otlp.max\_pod\_cpu\_capacity**: Maximum CPU capacity for OTLP pods. (Default is "1").
              - **distributed.otlp.pod\_ephemeral\_storage**: Ephemeral Storage Capacity for OTLP service pods. (Default is "250M").
              - **distributed.otlp.max\_folder\_size\_in\_mb**: Maximum folder size in MB for OTLP pods. (Default is "200").
              - **distributed.dm.min\_number\_of\_services**: Minimum Number of dm pods.
              - **distributed.dm.max\_number\_of\_services**: Maximum Number of dm pods.
              - **distributed.dm.max\_pod\_memory\_capacity**: Maximum memory capacity for dm pods. (Default is "8G").
              - **distributed.dm.max\_pod\_cpu\_capacity**: Maximum CPU capacity for dm pods. (Default is "1").
              - **distributed.dm.pod\_ephemeral\_storage**: Ephemeral Storage Capacity for dm service pods. (Default is "1G").
              - **backup.mysql.restore\_during\_installation**: If you want, during installation, to restore the backup from the latest kept backup with the same major version (if available on attached DB storage or S3 bucket), choose "Yes".
              - **backup.mysql.S3**: Backup MySQL to S3 (optional).
                  - **backup.mysql.S3.bucket**: MySQL backup S3 bucket name.
                  - **backup.mysql.S3.key**: MySQL backup S3 key.
                  - **backup.mysql.S3.secret**: MySQL backup S3 secret.
                  - **backup.mysql.S3.region**: MySQL backup S3 region.
              - **security.image\_pull\_policy**: If you want to use a custom image pull policy for all deployments, enter it here (optional).
              - **security.force\_hashed\_images**: If you want to use hashed images for all deployments, choose "Yes".
              - **security.set\_pod\_resources**: If you want to not set resource requests and limits for all deployments, choose "No".
              - **security.proxy.http\_proxy**: Enter the address of an HTTP proxy (optional).
              - **security.proxy.https\_proxy**: Enter the address of an HTTPS proxy (optional).
              - **security.proxy.additional\_no\_proxy**: Enter additional addresses to exclude from proxying (optional).
              - **security.use\_rate\_limiting**: If you want to use rate limiting in nginx, choose "Yes".
              - **security.requests\_per\_second**: Enter the number of requests per second for rate limiting.
              - **affinity.node\_affinity**: Enter the node affinity for the deployments (optional). There is only one nodeSelectorTerm as requiredDuringSchedulingIgnoredDuringExecution, but it can contain multiple matchExpressions (which are ANDed together).
                  - **key**: Enter the key for the node affinity.
                  - **operator**: Enter the operator for the node affinity (In, NotIn, Exists, DoesNotExist, Gt, Lt).
                  - **values**: Enter the values for the node affinity. A comma-separated list of values.
                  - **services**: Enter the services for the node affinity. A comma-separated list of services to which the node affinity applies. If not specified, the node affinity will apply to all services. List of possible services: mysql, nginx, storage, vfapi-parser, vfapi-users, vfapi-organizations, vfapi-idp, vfapi-mcp, vfapi-measurements, vfapi-otlp, vfapi-dm. (optional).
              - **affinity.node\_selectors**: Enter the node selectors for the deployments (optional).
                  - **key**: Enter the key for the node selector.
                  - **value**: Enter the value for the node selector.
                  - **services**: Enter the services for the node selector. A comma-separated list of services to which the node selector applies. If not specified, the node selector will apply to all services. List of possible services: mysql, nginx, storage, vfapi-parser, vfapi-users, vfapi-organizations, vfapi-idp, vfapi-mcp, vfapi-measurements, vfapi-otlp, vfapi-dm. (optional).
              - **affinity.additional\_labels**: Enter additional labels for the deployments (optional).
                  - **key**: Enter the key for the additional label.
                  - **value**: Enter the value for the additional label.
                  - **services**: Enter the services for the additional label. See above at the node selectors section for more details. (optional).
              - **affinity.additional\_annotations**: Enter additional annotations for the deployments (optional).
                  - **key**: Enter the key for the additional annotation.
                  - **value**: Enter the value for the additional annotation.
                  - **services**: Enter the services for the additional annotation. See above at the node selectors section for more details. (optional).
          1. In the Installed Operators window choose and click on vFunction Operator and then Create Instance.
          1. In the Create vFunctionServer window, paste the YAML file that you created in step 1.
          1. Click Create.
          1. You can now click on your new created vFunction operand in order to see its details and installation progress.

          ## Verifying the installation

          Check that the operand installed successfully:

          1. Check that the "Operand State" property shows “Working”.
          1. Check that the "Successfully Installed" property shows "Yes".

          If "Operand State" shows "Failed", it indicates the operand wasn’t installed correctly. Check the events and logs for all pods (operator and image containers) for any issues. In the event that you cannot troubleshoot, contact vFunction support.

          ## After installation

          The vFunction site is now accessible via the newly created vFunction application custom address, for example, my.domain.com. There are two ways you can access the vFunction dashboard.

          ### Access using the router canonical hostname

          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address, and to the ”vfunction” subdomain of the OpenShift canonical router hostname as the alias.
          For example, ***my.domain.com.   CNAME   vfunction.apps.ocp4.my-openshift.com.***
          2. Find your cluster Router Canonical Hostname address in the newly created vfunction-route-xxx under your vFunction OpenShift project -> **Networking** -> **Routes** -> **vfunction-route-xxx** route -> **Router Canonical Hostname** field.
          You can now access the dashboard using your defined "host" spec property (as above).

          ### Access using the nginx service location

          You may use this access method if your OpenShift is installed on a provider that supports exposing LoadBalancer-type services.
          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address and to the vfunction-nginx-xxx service location.
          For example, ***my.domain.com.   CNAME   a05951ed7cdf-1394239323.us-east-1.elb.amazonaws.com.***
          1. Point your custom domain to the vfunction-nginx-xxx service's external IP location, which you can find in the OpenShift project > **Networking** -> **Services** -> **vfunction-nginx-xxx** service > **Service Address** > **Location** field.

          ## Upgrade and Rollback

          The vFunction operator includes a built-in auto-upgrade mechanism.

          You can choose one of 4 upgrading modes:
          - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically.
          - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
          - **Never**: The automatic upgrading mechanism is off (default).
          - **Maintenance**: Put the operator in a maintenance mode.

          For an on-demand upgrade,  you can change anytime the upgrading mode from “Never” to “Always”, wait for the operand to be upgraded, and change again to “Never”.

          During the upgrade, the "Operand State" property will change to “Upgrading...” and back again to “Working” after a successful upgrade.

          In case of a failure upgrading the operand, an automatic rollback to the last working version will occur. The failed version will be marked as defective and the operator will not try to upgrade to it again.

          Choosing the "Maintenance" option will tell the operator to take down all pods, so you can safely fix storage issues, return volumes from snapshots, etc.
          After finishing the maintenance, return the upgrade mode to your original desired policy.
        displayName: vFunction Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - vfunction
        - analyzing
        - services
        links:
        - name: vFunction
          url: https://vfunction.com
        - name: Terms
          url: https://www.vfunction.com/eula/
        maintainers:
        - email: raviv@vfunction.com
          name: Raviv Halivni
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: vFunction
        relatedImages:
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-mcp@sha256:d50142c1b7f8a52944f9051a13bb821c0a790f137ebef09b0a84256bf7e5d794
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-otlp@sha256:80ca6d64b0d74f2d4f1b267002d5666f7c602956c656fd36a5c430f77eef0672
        - registry.connect.redhat.com/vfunction/vfunction-server-operator@sha256:a5a0f47f7fb9219139f8d749da14950a8c39698b340644011485dfbec54ebe8b
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-parser@sha256:b70fe4340d0de33c03a3b3458ea29e4cbff30cdd43f0e8d94265568e9a8d1ac7
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-idp@sha256:17fe2e3ef706a4bbd8fdf0e0e08bd021c49104b331674aae7b8978b31730a03d
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-organizations@sha256:8dd4d043806da1e85a4565b2f14bb9e24671fc158f8c1c2adc3804fda5213571
        - registry.connect.redhat.com/vfunction/vfunction-storage@sha256:8c36770f1fa5034ba089ae81c36c587cb73a5176f63969023c8ab6f90d8d7abd
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-measurements@sha256:81edae1c7344c13916530f20c8be09512ac8e20542770949783b0025c37079f8
        - registry.connect.redhat.com/vfunction/vfunction-mysql@sha256:b4b6543eb02e396e1339877f61b74cc783977dd607cd3679edce8ac6b627e9e9
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-users@sha256:c471d450d016aa0801ffb3de4648735245f2cf6901576106aed06cacc948d6f6
        - registry.connect.redhat.com/vfunction/vfunction-vfapi-dm@sha256:2f82294b15af873c012094a8984cb0c157bf322c28cbe58b38253adc917b6662
        - registry.connect.redhat.com/vfunction/vfunction-nginx@sha256:8370b871d05a24da9841a93fb255bf47945428b0a192b96b4eb95b9c0efb86a0
        version: 4.3.1762
      entries:
      - name: vfunction-server-operator.v4.3.1762
        version: 4.3.1762
      - name: vfunction-server-operator.v4.0.1576
        version: 4.0.1576
      name: stable
    defaultChannel: stable
    packageName: vfunction-server-operator
    provider:
      name: vFunction
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent
      provider-url: ""
    name: cilium
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cilium.v1.10.15-x084926b
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/hubble-ui@sha256:dec6e7e541cac91ab859245dd2a7fa02de5363b75caa3509a79cf76d4f622550
        - registry.connect.redhat.com/isovalent/cilium-olm:970b1d74dbe2b87c28e8dc1c68936de33123a4df-v1.10.15@sha256:afa3ac140fae24b151ab6ff5a8cdef11566a58f7234d702a6a45a398822e6d34
        - quay.io/cilium/cilium@sha256:adbb60f9d19d641a5496fcfc4769ebb55c1d4cd7699fef752f7c94754194cf83
        - quay.io/cilium/hubble-relay@sha256:67a9183fd1febd9e51f4d02bd35092b81c8612a1716015988f69fffba79d999a
        - quay.io/cilium/clustermesh-apiserver@sha256:e73151c3295589de00dccae75819aabdff8fd907260817020f9e8fe3738e78b2
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        - quay.io/cilium/operator-generic@sha256:7f2f7b59b541e4f660d65ed0a21d25ab69e49022172529771dd1134315e95c82
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/hubble-ui-backend@sha256:a07e13df3e9d7ec0dcc3b158f243cda5a08f6eca38a25a5aa3171f6b3ec491ed
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        version: 1.10.15+x084926b
      entries:
      - name: cilium.v1.10.15-x084926b
        version: 1.10.15+x084926b
      - name: cilium.v1.10.13-xa8bb1e5
        version: 1.10.13+xa8bb1e5
      - name: cilium.v1.10.12-x362ea11
        version: 1.10.12+x362ea11
      - name: cilium.v1.10.10-x6adc6ff
        version: 1.10.10+x6adc6ff
      name: "1.10"
    - currentCSV: cilium.v1.11.20-xdb5ed32
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/cilium@sha256:60df3cb7155886e0b62060c7a4a31e457933c6e35af79febad5fd6e43bab2a99
        - quay.io/cilium/operator-generic@sha256:1439954acf620f048ef663524ae70b4a25693c58527a2f2cee51124496e29f90
        - quay.io/cilium/hubble-ui-backend@sha256:46070270b8ffe7ac3f1794721be0b5a09f906dff3492f8a199a5f995db362d2a
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:3677840d57e1d74e3ac14db0dfceb83088e6ebd70775d775855cdd3bfed68059
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/hubble-ui@sha256:bbf8b041152041d3ef000d6074a8dd5d40ed910b5c0067f365b00e52e449d467
        - quay.io/cilium/hubble-relay@sha256:e2f38b901fd8bd5adc9a765a5e68836364ebd1e7dfb85c2bcd8a5488b23c3470
        - quay.io/cilium/clustermesh-apiserver@sha256:46760182f8c98227cfac27627275616987b71509227775350573d834133a6d49
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        version: 1.11.20+xdb5ed32
      entries:
      - name: cilium.v1.11.20-xdb5ed32
        version: 1.11.20+xdb5ed32
      name: "1.11"
    - currentCSV: cilium.v1.12.19-xef55de8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.12.0 <1.12.19+xef55de8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/hubble-relay@sha256:19f2dee371523888985e6e34b30d07ef3f84f2a9206ad79bdfc8dede96e25d5c
        - quay.io/cilium/hubble-ui-backend@sha256:f88a73a120521eeafbcdbd908f517117f6557bf61e115847853fac371f0d774c
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:d35994cf5cf319b88ff5a7985cc972edd23df5a282360f9ba5516d610390ebf8
        - quay.io/cilium/hubble-ui@sha256:e15af59a2ded739e420be82de6cbdd0ce22f8d3f00e3a10b3d2d2734e074a394
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/cilium/cilium@sha256:b997994ab759443d9fc9807eb39b4784bf03e13730a8132ffe5df3edef22a312
        - quay.io/cilium/operator-generic@sha256:da4d72facecde862c88d08bf99c3f476845beb407052a0cc4c06b68cc93b7570
        - quay.io/cilium/clustermesh-apiserver@sha256:61f8e871760dceaa0c0459bbf44a3049dd0f4ab15e57e063d9b3284434cb3dd8
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        version: 1.12.19+xef55de8
      entries:
      - name: cilium.v1.12.19-xef55de8
        version: 1.12.19+xef55de8
      - name: cilium.v1.12.18-x63f4f4c
        version: 1.12.18+x63f4f4c
      - name: cilium.v1.12.17-x6cc0c18
        version: 1.12.17+x6cc0c18
      - name: cilium.v1.12.16-x14b1d1d
        version: 1.12.16+x14b1d1d
      - name: cilium.v1.12.15-x7d84617
        version: 1.12.15+x7d84617
      - name: cilium.v1.12.14-x17ed60a
        version: 1.12.14+x17ed60a
      - name: cilium.v1.12.13-xb414403
        version: 1.12.13+xb414403
      - name: cilium.v1.12.12-xd280cc3
        version: 1.12.12+xd280cc3
      - name: cilium.v1.12.11-xc8fb7eb
        version: 1.12.11+xc8fb7eb
      - name: cilium.v1.12.10-xe372776
        version: 1.12.10+xe372776
      - name: cilium.v1.12.9-x2ccb2b6
        version: 1.12.9+x2ccb2b6
      - name: cilium.v1.12.8-xa53bac1
        version: 1.12.8+xa53bac1
      - name: cilium.v1.12.7-xba75145
        version: 1.12.7+xba75145
      - name: cilium.v1.12.0-xb42b669
        version: 1.12.0+xb42b669
      name: "1.12"
    - currentCSV: cilium.v1.13.12-x6e56d47
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.13.12+x6e56d47'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/hubble-relay@sha256:01b23ea40bcd81145dde6bfcbfc4d542749d08c2a1c6348954c85123a8d2b1fc
        - quay.io/cilium/clustermesh-apiserver@sha256:f1b1d0a85bab65e7d6adc90d000513a56ac58bdb071aa391a8580d73d20e6b6a
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        - quay.io/cilium/operator-generic@sha256:f83734bbe270f961d545c7929152785507ce04a05d818ebc9776941723736d02
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:b0d22d4376b4c2b2b3b988955227f1ed2ad49a363c5367e53f64edae2cec0ff5
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/cilium/cilium@sha256:d99204aa7b3b7bd2c9ab47fd398cc9f40290799bc0c7a4386c8dc5c1780cd3d3
        version: 1.13.12+x6e56d47
      entries:
      - name: cilium.v1.13.12-x6e56d47
        version: 1.13.12+x6e56d47
      - name: cilium.v1.13.11-xba1ab0b
        version: 1.13.11+xba1ab0b
      - name: cilium.v1.13.10-x412ae64
        version: 1.13.10+x412ae64
      - name: cilium.v1.13.9-x3ae7099
        version: 1.13.9+x3ae7099
      - name: cilium.v1.13.8-xe0355c7
        version: 1.13.8+xe0355c7
      - name: cilium.v1.13.7-x37f9269
        version: 1.13.7+x37f9269
      - name: cilium.v1.13.6-xdb5afcd
        version: 1.13.6+xdb5afcd
      - name: cilium.v1.13.5-xaf5caa6
        version: 1.13.5+xaf5caa6
      - name: cilium.v1.13.4-x129f2a7
        version: 1.13.4+x129f2a7
      - name: cilium.v1.13.3-x5854d64
        version: 1.13.3+x5854d64
      - name: cilium.v1.13.2-x82c79c4
        version: 1.13.2+x82c79c4
      - name: cilium.v1.13.1-x10a5de9
        version: 1.13.1+x10a5de9
      - name: cilium.v1.13.0-x32540df
        version: 1.13.0+x32540df
      name: "1.13"
    - currentCSV: cilium.v1.14.7-x3522df7
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.14.7+x3522df7'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/clustermesh-apiserver@sha256:28f3ffe53365ca79831af600f09a95c0b3e9959f5f891b416dab8cedd90c263d
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:291a05c58de8d7daefeb0689c23cadd92c9368aea09751de70eaa817f7fcc358
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/cilium/hubble-relay@sha256:46762393daf4a0aaef76b106614c2615942f98f10aeacd435ea3fb1a0bdf69e4
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/operator-generic@sha256:37ef0bd85c27c765c637cd58c3ff4a559f8734ae39f9d1839a3ac7803de7b952
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/cilium@sha256:45ce2b87696082ecf7d53ba1c64ceeb4217578033e5ef28ac479ec049a48bc32
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        version: 1.14.7+x3522df7
      entries:
      - name: cilium.v1.14.7-x3522df7
        version: 1.14.7+x3522df7
      - name: cilium.v1.14.6-x11464e1
        version: 1.14.6+x11464e1
      - name: cilium.v1.14.5-x3583390
        version: 1.14.5+x3583390
      - name: cilium.v1.14.4-xb89471d
        version: 1.14.4+xb89471d
      - name: cilium.v1.14.3-x041dc10
        version: 1.14.3+x041dc10
      - name: cilium.v1.14.2-xedd5ce2
        version: 1.14.2+xedd5ce2
      - name: cilium.v1.14.1-xdd859b7
        version: 1.14.1+xdd859b7
      name: "1.14"
    - currentCSV: cilium.v1.15.1-x7095b76
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.15.1+x7095b76'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:9ab6be29447125e886300e9258b9a06bedf0a9d87405832aa8b6565ed1ba4215
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/hubble-relay@sha256:3254aaf85064bc1567e8ce01ad634b6dd269e91858c83be99e47e685d4bb8012
        - quay.io/cilium/operator-generic@sha256:819c7281f5a4f25ee1ce2ec4c76b6fbc69a660c68b7825e9580b1813833fa743
        - quay.io/cilium/clustermesh-apiserver@sha256:b353badd255c2ce47eaa8f394ee4cbf70666773d7294bd887693e0c33503dc37
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        - quay.io/cilium/cilium@sha256:351d6685dc6f6ffbcd5451043167cfa8842c6decf80d8c8e426a417c73fb56d4
        version: 1.15.1+x7095b76
      entries:
      - name: cilium.v1.15.1-x7095b76
        version: 1.15.1+x7095b76
      name: "1.15"
    - currentCSV: cilium.v1.9.14-x70dc065
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Basic Install
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/cilium/cilium@sha256:2c6ce93fa7e625979043a387eb998c17ad57df8768d89facb9b715da42a4c51c
        - quay.io/cilium/operator-generic@sha256:bdcfd2eade99933f2fda55ef79ea697ddfad3512f65b15bcd0ba7702518c1ba3
        - quay.io/cilium/hubble-ui-backend@sha256:2cf1c7b1fde6161f575f3147c2ef885739e60b8c59a1ed444fe74edbdf22c7c1
        - docker.io/envoyproxy/envoy@sha256:55a141c0b8e39497b32074cb758b4e5dc2384bd6d9926144969a4452eaa611fd
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:3a9ca9a4c2f78109f265c4758f5f693d01865f2d0e31bc5f1ecbc065e878628b
        - quay.io/cilium/clustermesh-apiserver@sha256:a0da5edf0372899647da51de1b277f0bab8e676d694aee7f939cddfdd3172010
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        - quay.io/cilium/hubble-relay@sha256:fd6ab1aea260abc5f64eca26c1b1e7009983e4aaa8e5d098e8d442f7659603fb
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/hubble-ui@sha256:c267219946b84f79a5fd2849f5cc71f39453e29d93039e57b9cef2b5802e9c50
        version: 1.9.14+x70dc065
      entries:
      - name: cilium.v1.9.14-x70dc065
        version: 1.9.14+x70dc065
      name: "1.9"
    - currentCSV: cilium.v1.10.4-x5bfd7b3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Basic Install
          categories: Networking,Security
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:b45ff6a6552d71cbc26e358e0b25759d33818fd92301f1656453c5bcf826ff13
        version: 1.10.4+x5bfd7b3
      entries:
      - name: cilium.v1.10.4-x5bfd7b3
        version: 1.10.4+x5bfd7b3
      name: stable
    defaultChannel: "1.15"
    packageName: cilium
    provider:
      name: Isovalent
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Couchbase
      provider-url: ""
    name: couchbase-enterprise-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: couchbase-operator.v2.6.4-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator@sha256:1f0a6865c360a14ba98706fd6fe8dff3bdbdede32e3622f93f19c82e377ca33f
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        version: 2.6.4-7
      entries:
      - name: couchbase-operator.v2.6.4-7
        version: 2.6.4-7
      - name: couchbase-operator.v2.6.4-6
        version: 2.6.4-6
      name: 2.6.4
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator@sha256:707571a7881bc12d5f1b0b1a94eb3c675d5c0ea0114c994a47ceb5aa79df4013
        version: 2.7.0-5
      entries:
      - name: couchbase-operator.v2.7.0-5
        version: 2.7.0-5
      name: 2.7.0
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:a79e1b6781d5f8badf664e7efea6e2889e43fae6e0d37169d766685dcd6ff394
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        version: 2.7.1-1
      entries:
      - name: couchbase-operator.v2.7.1-1
        version: 2.7.1-1
      name: 2.7.1
    - currentCSV: couchbase-operator.v2.8.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:df61928b064014d851c6e7dff6dcd63a433a27fa5623254b6e389c3eea170e36
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        version: 2.8.0-2
      entries:
      - name: couchbase-operator.v2.8.0-2
        version: 2.8.0-2
      - name: couchbase-operator.v2.8.0-1
        version: 2.8.0-1
      name: 2.8.0
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        version: 2.8.1-1
      entries:
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: 2.8.1
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        version: 2.8.1-1
      entries:
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: stable
    defaultChannel: stable
    packageName: couchbase-enterprise-certified-rhmp
    provider:
      name: Couchbase
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: topology-aware-lifecycle-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: topology-aware-lifecycle-manager.v4.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ran.openshift.io/v1alpha1",
                "kind": "ClusterGroupUpgrade",
                "metadata": {
                  "name": "ClusterGroupUpgrade-sample"
                },
                "spec": {
                  "preCaching": true
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:bf24cad2940a227f5bd2f188be5d298ff19176c634fcf609c7225d7e0a6f2fd1
          description: The topology aware lifecycle manager is an operator that facilitates
            platform upgrades of group of clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-operators",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/cluster-group-upgrades-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterGroupUpgrade is the Schema for the ClusterGroupUpgrades
              API
            displayName: Cluster Group Upgrade
            kind: ClusterGroupUpgrade
            name: clustergroupupgrades.ran.openshift.io
            version: v1alpha1
          - description: ImageBasedGroupUpgrade is the schema for upgrading a group
              of clusters using IBU
            displayName: Image-Based Group Upgrade
            kind: ImageBasedGroupUpgrade
            name: imagebasedgroupupgrades.lcm.openshift.io
            version: v1alpha1
          - description: PreCachingConfig is the Schema for the precachingconfigs
              API
            displayName: Pre-caching Config
            kind: PreCachingConfig
            name: precachingconfigs.ran.openshift.io
            version: v1alpha1
        description: |-
          Topology Aware Lifecycle Manager is an operator that facilitates
          platform and operator upgrades of group of clusters
        displayName: Topology Aware Lifecycle Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm lifecyclemanagement upgrades cluster
        links:
        - name: Cluster Group Upgrades Operator
          url: https://cluster-group-upgrades-operator.domain
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:eed8234b0fe47083ab941f513709f778bfcb80e32747b71a65fd5f62328a5ca0
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:bf24cad2940a227f5bd2f188be5d298ff19176c634fcf609c7225d7e0a6f2fd1
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-aztp-rhel9@sha256:f6a37a0de10c847a53e22406990858dbaa4cdd3be075bc047e8dcf5bc6a2f2f7
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-precache-rhel9@sha256:7673a0d7f8944b245124376f2153bf987299f5bd7350b33f0d4bd1ac18543929
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-recovery-rhel9@sha256:37e234f247c841cd12c4617f92bf8bd5a5dbcbf46fa1be042d56b966994efaad
        version: 4.16.5
      entries:
      - name: topology-aware-lifecycle-manager.v4.16.5
        version: 4.16.5
      - name: topology-aware-lifecycle-manager.v4.16.4
        version: 4.16.4
      - name: topology-aware-lifecycle-manager.v4.16.3
        version: 4.16.3
      - name: topology-aware-lifecycle-manager.v4.16.2
        version: 4.16.2
      - name: topology-aware-lifecycle-manager.v4.16.1
        version: 4.16.1
      - name: topology-aware-lifecycle-manager.v4.16.0
        version: 4.16.0
      name: "4.16"
    - currentCSV: topology-aware-lifecycle-manager.v4.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ran.openshift.io/v1alpha1",
                "kind": "ClusterGroupUpgrade",
                "metadata": {
                  "name": "ClusterGroupUpgrade-sample"
                },
                "spec": {
                  "preCaching": true
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:bf24cad2940a227f5bd2f188be5d298ff19176c634fcf609c7225d7e0a6f2fd1
          description: The topology aware lifecycle manager is an operator that facilitates
            platform upgrades of group of clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-operators",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/cluster-group-upgrades-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterGroupUpgrade is the Schema for the ClusterGroupUpgrades
              API
            displayName: Cluster Group Upgrade
            kind: ClusterGroupUpgrade
            name: clustergroupupgrades.ran.openshift.io
            version: v1alpha1
          - description: ImageBasedGroupUpgrade is the schema for upgrading a group
              of clusters using IBU
            displayName: Image-Based Group Upgrade
            kind: ImageBasedGroupUpgrade
            name: imagebasedgroupupgrades.lcm.openshift.io
            version: v1alpha1
          - description: PreCachingConfig is the Schema for the precachingconfigs
              API
            displayName: Pre-caching Config
            kind: PreCachingConfig
            name: precachingconfigs.ran.openshift.io
            version: v1alpha1
        description: |-
          Topology Aware Lifecycle Manager is an operator that facilitates
          platform and operator upgrades of group of clusters
        displayName: Topology Aware Lifecycle Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm lifecyclemanagement upgrades cluster
        links:
        - name: Cluster Group Upgrades Operator
          url: https://cluster-group-upgrades-operator.domain
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:eed8234b0fe47083ab941f513709f778bfcb80e32747b71a65fd5f62328a5ca0
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:bf24cad2940a227f5bd2f188be5d298ff19176c634fcf609c7225d7e0a6f2fd1
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-aztp-rhel9@sha256:f6a37a0de10c847a53e22406990858dbaa4cdd3be075bc047e8dcf5bc6a2f2f7
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-precache-rhel9@sha256:7673a0d7f8944b245124376f2153bf987299f5bd7350b33f0d4bd1ac18543929
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-recovery-rhel9@sha256:37e234f247c841cd12c4617f92bf8bd5a5dbcbf46fa1be042d56b966994efaad
        version: 4.16.5
      entries:
      - name: topology-aware-lifecycle-manager.v4.16.5
        version: 4.16.5
      - name: topology-aware-lifecycle-manager.v4.16.4
        version: 4.16.4
      - name: topology-aware-lifecycle-manager.v4.16.3
        version: 4.16.3
      - name: topology-aware-lifecycle-manager.v4.16.2
        version: 4.16.2
      - name: topology-aware-lifecycle-manager.v4.16.1
        version: 4.16.1
      - name: topology-aware-lifecycle-manager.v4.16.0
        version: 4.16.0
      name: stable
    defaultChannel: stable
    packageName: topology-aware-lifecycle-manager
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Korrel8r
      provider-url: https://github.com/korrel8r/korrel8r
    name: korrel8r
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: korrel8r.v0.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "korrel8r.openshift.io/v1alpha1",
                "kind": "Korrel8r",
                "metadata": {
                  "name": "korrel8r",
                  "namespace": "korrel8r"
                },
                "spec": {
                  "debug": {
                    "verbose": 1
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional,Monitoring,Logging & Tracing
          containerImage: quay.io/korrel8r/operator:0.1.8
          createdAt: "2024-09-27T17:45:00Z"
          description: Correlate cluster resources and observability data.
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "korrel8r.openshift.io/v1alpha1",
              "kind": "Korrel8r",
              "metadata": {
                "name": "korrel8r",
                "namespace": "korrel8r"
              }
            }
          operatorframework.io/suggested-namespace: korrel8r
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/korrel8r/korrel8r
          support: Korrel8r Project
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Korrel8r is a service that correlates observabililty signals
              in the cluster.
            displayName: Korrel8r
            kind: Korrel8r
            name: korrel8rs.korrel8r.openshift.io
            version: v1alpha1
        description: |
          Korrel8r helps navigate from problem symptoms to related resources and signal data that can reveal the cause.
          It can follow relationships between disjointed observability "silos" (logs, metrics, alerts and more) to bring together all the data available to solve a problem.
          This operator allows Korrel8r to be deployed as a cluster service with a REST API.
          It can be used as a back-end for data correlation by the OpenShift console and other tools.
        displayName: Korrel8r
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - signal
        - correlation
        links:
        - name: Documentation
          url: https://korrel8r.github.io/korrel8r
        - name: Source Code and Issues
          url: https://github.com/korrel8r/korrel8r
        maintainers:
        - email: korrel8r@gmail.com
          name: Korrel8r Project
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Korrel8r
          url: https://github.com/korrel8r/korrel8r
        relatedImages:
        - quay.io/korrel8r/operator:0.1.8
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 0.1.8
      entries:
      - name: korrel8r.v0.1.8
        version: 0.1.8
      - name: korrel8r.v0.1.7
        version: 0.1.7
      - name: korrel8r.v0.1.5
        version: 0.1.5
      - name: korrel8r.v0.1.3
        version: 0.1.3
      - name: korrel8r.v0.1.2
        version: 0.1.2
      - name: korrel8r.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: korrel8r
    provider:
      name: Korrel8r
      url: https://github.com/korrel8r/korrel8r
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app: enmasse
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq-online
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-online.1.7.4-0.1657725339.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "StandardInfraConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "broker": {
                    "resources": {
                      "memory": "1Gi",
                      "storage": "5Gi"
                    },
                    "addressFullPolicy": "FAIL"
                  },
                  "router": {
                    "linkCapacity": 50,
                    "resources": {
                      "memory": "512Mi"
                    }
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "BrokeredInfraConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "broker": {
                    "resources": {
                      "memory": "4Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta2",
                "kind": "AddressPlan",
                "metadata": {
                  "name": "standard-small-queue"
                },
                "spec": {
                  "addressType": "queue",
                  "shortDescription": "Small Queue",
                  "resources": {
                    "router": 0.01,
                    "broker": 0.1
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta2",
                "kind": "AddressSpacePlan",
                "metadata": {
                  "name": "standard-small"
                },
                "spec": {
                  "addressSpaceType": "standard",
                  "infraConfigRef": "default",
                  "shortDescription": "Small Address Space Plan",
                  "resourceLimits": {
                    "router": 1.0,
                    "broker": 2.0,
                    "aggregate": 2.0
                  },
                  "addressPlans": [
                    "standard-small-queue"
                  ]
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "AuthenticationService",
                "metadata": {
                  "name": "standard-authservice"
                },
                "spec": {
                  "type": "standard",
                  "standard": {
                    "storage": {
                      "claimName": "standard-authservice",
                      "deleteClaim": true,
                      "size": "1Gi",
                      "type": "persistent-claim"
                    }
                  }
                }
              },
              {
                "apiVersion": "enmasse.io/v1beta1",
                "kind": "AddressSpace",
                "metadata": {
                  "name": "myspace"
                },
                "spec": {
                  "plan": "standard-small",
                  "type": "standard"
                }
              },
              {
                "apiVersion": "enmasse.io/v1beta1",
                "kind": "Address",
                "metadata": {
                  "name": "myspace.myqueue"
                },
                "spec": {
                  "address": "myqueue",
                  "plan": "standard-small-queue",
                  "type": "queue"
                }
              },
              {
                "apiVersion": "user.enmasse.io/v1beta1",
                "kind": "MessagingUser",
                "metadata": {
                  "name": "myspace.user"
                },
                "spec": {
                  "authentication": {
                    "password": "ZW5tYXNzZQ==",
                    "type": "password"
                  },
                  "authorization": [
                    {
                      "addresses": [
                        "myqueue"
                      ],
                      "operations": [
                        "send",
                        "recv"
                      ]
                    }
                  ],
                  "username": "user"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator@sha256:2aaf87224a96aa56dc595e96d8e7ec8d138aaf20240afa25a1e805b405941355
          createdAt: "2019-02-19T00:00:00Z"
          description: Red Hat Integration - AMQ Online provides messaging as a managed
            service on OpenShift
          olm.substitutesFor: amq-online.1.7.4
          operators.openshift.io/capability: '["license-required"]'
          operators.openshift.io/valid-licenses: '["Red Hat AMQ", "Red Hat Integration"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: ""
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A messaging user that can connect to an Address Space
            displayName: Messaging User
            kind: MessagingUser
            name: messagingusers.user.enmasse.io
            version: v1beta1
          - description: A resource representing the available schema of plans and
              authentication services
            displayName: AddressSpaceSchema
            kind: AddressSpaceSchema
            name: addressspaceschemas.enmasse.io
            version: v1beta1
          - description: A group of messaging addresses that can be accessed via the
              same endpoint
            displayName: Address Space
            kind: AddressSpace
            name: addressspaces.enmasse.io
            version: v1beta1
          - description: A messaging address that can be used to send/receive messages
              to/from
            displayName: Address
            kind: Address
            name: addresses.enmasse.io
            version: v1beta1
          - description: Infrastructure configuration template for the standard address
              space type
            displayName: Standard Infra Config
            kind: StandardInfraConfig
            name: standardinfraconfigs.admin.enmasse.io
            version: v1beta1
          - description: Infrastructure configuration template for the brokered address
              space type
            displayName: Brokered Infra Config
            kind: BrokeredInfraConfig
            name: brokeredinfraconfigs.admin.enmasse.io
            version: v1beta1
          - description: Plan describing the resource usage of a given address type
            displayName: Address Plan
            kind: AddressPlan
            name: addressplans.admin.enmasse.io
            version: v1beta2
          - description: Plan describing the capabilities and resource limits of a
              given address space type
            displayName: Address Space Plan
            kind: AddressSpacePlan
            name: addressspaceplans.admin.enmasse.io
            version: v1beta2
          - description: Authentication service configuration available to address
              spaces.
            displayName: Authentication Service
            kind: AuthenticationService
            name: authenticationservices.admin.enmasse.io
            version: v1beta1
          - description: Console Service Singleton for deploying global console.
            displayName: Console Service
            kind: ConsoleService
            name: consoleservices.admin.enmasse.io
            version: v1beta1
        description: "**Red Hat Integration - AMQ Online** provides messaging as a
          managed service on OpenShift.\nWith Red Hat Integration - AMQ Online, administrators
          can configure a cloud-native,\nmulti-tenant messaging service either in
          the cloud or on premise.\nDevelopers can provision messaging using the Red
          Hat Integration - AMQ Online Console.\nMultiple development teams can provision
          the brokers and queues from the\nconsole, without requiring each team to
          install, configure, deploy,\nmaintain, or patch any software. \n\n**The
          core capabilities include:**\n\n  * **Built-in authentication and authorization**
          - Use the built-in authentication service or\n    plug in your own authentication
          service for authentication and\n    authorization of messaging clients.\n\n
          \ * **Self-service messaging for applications** - The service administrator
          deploys\n    and manages the messaging infrastructure, while applications
          can request\n    messaging resources regardless of the messaging infrastructure.\n\n
          \ * **Support for a wide variety of messaging patterns** - Choose between\n
          \   JMS-style messaging with strict guarantees, or messaging that supports\n
          \   a larger number of connections and higher throughput.\n\n## Post-installation
          tasks\n\nTo fully use Red Hat Integration - AMQ Online, you need to create
          a few\ninfrastructure components after the installation, including:\n\n
          \ * An authentication service\n  * Infrastructure configuration for the
          standard and brokered address space\n  * Address and address space plans\n
          \ * (Optional) Create RBAC roles to allow users to discover available plans\n
          \ * (Optional) Create RBAC roles to allow users to self-manage addresses
          and\n    address spaces.\n\nFor a complete overview of how to configure
          Red Hat Integration - AMQ Online, see\n[Configuring Red Hat Integration
          - AMQ Online](https://access.redhat.com/documentation/en-us/red_hat_amq/7.7/html/installing_and_managing_amq_online_on_openshift/configuring-messaging).\n\n###
          Quickstart infrastructure configuration\n\nIf you simply want to get started
          quickly, you can import the following\nYAML by saving the content to a file
          and apply it by running the \n`oc apply -f <file>` command. You can also
          split the content (at the `---` marker)\nand import the single YAML document
          using the Web UI: \n\n~~~yaml\n---\napiVersion: admin.enmasse.io/v1beta1\nkind:
          StandardInfraConfig\nmetadata:\n  name: default\n---\napiVersion: admin.enmasse.io/v1beta2\nkind:
          AddressPlan\nmetadata:\n  name: standard-small-queue\nspec:\n  addressType:
          queue\n  resources:\n    router: 0.01\n    broker: 0.1\n---\napiVersion:
          admin.enmasse.io/v1beta2\nkind: AddressSpacePlan\nmetadata:\n  name: standard-small\nspec:\n
          \ addressSpaceType: standard\n  infraConfigRef: default\n  addressPlans:\n
          \ - standard-small-queue\n  resourceLimits:\n    router: 2.0\n    broker:
          3.0\n    aggregate: 4.0\n---\napiVersion: admin.enmasse.io/v1beta1\nkind:
          AuthenticationService\nmetadata:\n  name: standard-authservice\nspec:\n
          \ type: standard\n  standard:\n    storage:\n      claimName: standard-authservice\n
          \     deleteClaim: true\n      size: 1Gi\n      type: persistent-claim\n~~~\n\n###
          Create RBAC roles to allow users to discover available plans\n\nFor users
          to discover the available plans, cluster-wide roles to read the available\nschema
          can be created.  Import the following YAML by saving the content to a file
          and apply it by running the \n`oc apply -f <file>` command. You can also
          split the content (at the `---` marker)\nand import the single YAML document
          using the Web UI: \n\n~~~yaml\n---\napiVersion: rbac.authorization.k8s.io/v1\nkind:
          ClusterRole\nmetadata:\n  name: enmasse.io:schema\n  labels:\n    app: enmasse\nrules:\n
          \ - apiGroups: [ \"enmasse.io\" ]\n    resources: [ \"addressspaceschemas\"
          ]\n    verbs: [ \"get\", \"list\", \"watch\" ]\n---\napiVersion: rbac.authorization.k8s.io/v1\nkind:
          ClusterRoleBinding\nmetadata:\n  name: \"enmasse.io:schema\"\n  labels:\n
          \   app: enmasse\nroleRef:\n  apiGroup: rbac.authorization.k8s.io\n  kind:
          ClusterRole\n  name: enmasse.io:schema\nsubjects:\n- kind: Group\n  name:
          system:authenticated\n~~~\n\n### Creating infrastructure using the Web UI\n\nYou
          must create a new instance of each of the following custom resources. You
          can\nuse the example data directly, which is provided when using the\nWeb
          UI:\n\n  * *Authentication Service* – Create an authentication service.\n
          \ * *Brokered Infra Config* – Create the broker infrastructure configuration.\n
          \ * *Standard Infra Config* – Create the standard infrastructure\n    configuration.\n\nYou
          must also create at least one address space plan and one address plan.\n\n*Note*:
          The name of the address space plan and address plan is required\nlater when
          creating new addresses. Some examples use specific plan\nnames, which might
          not be available in your environment. You can\ncreate those plans, or edit
          the examples to use different plan names.\n\n## Configuration for messaging
          tenants\n\nWhile service administrators perform the infrastructure configuration,
          the following\nresources are for the actual users of the system, the messaging
          tenants.\n\nYou need to create those resources to satisfy your particular
          use case.\n\n  * *Address space* – A container for addresses\n  * *Address*
          – A messaging address (address, topic, queue, and so on)\n  * *Messaging
          user* – Manages access to an address\n"
        displayName: Red Hat Integration - AMQ Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amq
        - messaging
        - amqp
        - iot
        - mqtt
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/7.7/html/using_amq_online_on_openshift/index
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-online-1-topic-forwarder@sha256:de33a701339d91a4e9275f7cf65f917ff75368a59b8ab36dd66c90bc306875aa
        - registry.redhat.io/amq7/amq-online-1-broker-plugin@sha256:2466a336c0d78b196635de3fad636aeff2d9265839f7271b8ef94f3ba97d9bc0
        - registry.redhat.io/amq7/amq-interconnect@sha256:5ebd4c4a0ed0db2725ae466550c915531b32933c420344cc2083c616da3f326f
        - registry.redhat.io/redhat-sso-7/sso73-openshift@sha256:18d15980f28f0d0c6b8582131a1f062d84463577512875b5391a612c3d7d1096
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6d489a7897bd831ec273034f1ff0e223427a40afd01b9835ebb020a8bf9ddc66
        - registry.redhat.io/amq7/amq-broker@sha256:1ec0437bbd82722148fdbb02dc55682b114c311cd6f521a8bc621a8c1cb4b32b
        - registry.redhat.io/amq7/amq-online-1-address-space-controller@sha256:9acc555c8e869aec7d72f53b60320e1a96cc82d748638e0d50970837229fdbee
        - registry.redhat.io/amq7/amq-online-1-none-auth-service@sha256:a013fb1806c002b408a7753ff943cee24fbcdf8a657eff1e9582c42b5a7b0844
        - registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator@sha256:2aaf87224a96aa56dc595e96d8e7ec8d138aaf20240afa25a1e805b405941355
        - registry.redhat.io/amq7/amq-online-1-standard-controller@sha256:070349af30cec52fa8efc185fc940077a64bb40e3e36310d7cc6e39ce0971436
        - registry.redhat.io/amq7/amq-online-1-console-init@sha256:0970804909b82fdc3d4e7d5d88280e41157d743107de8bfa2d1baa1367543176
        - registry.redhat.io/openshift3/oauth-proxy@sha256:59b20f23c03d0012ec5c95c4725892ee64277c22dedc9d0a90a7b6b0bc80fc72
        - registry.redhat.io/amq7/amq-online-1-console-server-rhel7@sha256:1dcb3d0cc7bc6e84f4f96962ff88849511b7a7686a1f5dbd9300c4c2c8c05e2f
        - registry.redhat.io/amq7/amq-online-1-agent@sha256:f98ed8ee97ac59d8fdd40ec8b2c063a5afd360a3e183057c186e7a77a32801ee
        - registry.redhat.io/amq7/amq-online-1-auth-plugin@sha256:68926f434b2f6ce4d9698938bc3c53d514c21c3f76c73b6b4cdf7fcf51f7cb0d
        version: 1.7.4+0.1657725339.p
      entries:
      - name: amq-online.1.7.4
        version: 1.7.4
      - name: amq-online.1.7.4-0.1657725339.p
        version: 1.7.4+0.1657725339.p
      - name: amq-online.1.7.3
        version: 1.7.3
      - name: amq-online.1.7.2
        version: 1.7.2
      - name: amq-online.1.7.1
        version: 1.7.1
      - name: amq-online.1.7.1-0.1628610187.p
        version: 1.7.1+0.1628610187.p
      - name: amq-online.1.7.0
        version: 1.7.0
      name: stable
    defaultChannel: stable
    packageName: amq-online
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: sandboxed-containers-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sandboxed-containers-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:051bd7f1dad8cc3251430fee32184be8d64077aba78580184cef0255d267bdcf
          createdAt: 2021-06-07-13:38:06
          olm.skipRange: <1.0.2
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          OpenShift sandboxed containers based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift’s Defense-in-Depth strategy. For more information
          [see](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5).

          # Features & benefits
          - **Isolated Developer Environments & Priviliges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as CAP_ADMIN or CAP_BPF. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select preview-1.0 from the list of available Update Channel options.
            This ensures that you install the version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              openshift-sandboxed-containers-operator namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              openshift-sandboxed-containers-operator causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a KataConfig CRD instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.8/sandboxed_containers/understanding-sandboxed-containers.html).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kernel-isolated containers
        links:
        - name: OpenShift sandboxed containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        provider:
          name: Red Hat
          url: https://www.github.com/openshift/sandboxed-containers-operator
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:051bd7f1dad8cc3251430fee32184be8d64077aba78580184cef0255d267bdcf
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:c6c589d5e47ba9564c66c84fc2bc7e5e046dae1d56a3dc99d7343f01e42e4d31
        version: 1.0.2
      entries:
      - name: sandboxed-containers-operator.v1.0.2
        version: 1.0.2
      name: preview-1.0
    - currentCSV: sandboxed-containers-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:b1e824e126c579db0f56d04c3d1796d82ed033110c6bc923de66d95b67099611
          createdAt: "2021-10-08T13:20:20.250789Z"
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift’s Defense-in-Depth strategy. For more information
          [see](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5).

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `preview-1.1` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CRD instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.8/sandboxed_containers/understanding-sandboxed-containers.html).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kernel-isolated containers
        links:
        - name: OpenShift sandboxed containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        provider:
          name: Red Hat
          url: https://www.github.com/openshift/sandboxed-containers-operator
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:b1e824e126c579db0f56d04c3d1796d82ed033110c6bc923de66d95b67099611
        version: 1.1.0
      entries:
      - name: sandboxed-containers-operator.v1.1.0
        version: 1.1.0
      - name: sandboxed-containers-operator.v1.0.2
        version: 1.0.2
      name: preview-1.1
    - currentCSV: sandboxed-containers-operator.v1.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-12T14:30:28Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.10.3'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '["peerpods.confidentialcontainers.org","peerpodconfigs.confidentialcontainers.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/sandboxed-containers-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
          - kind: PeerPod
            name: peerpods.confidentialcontainers.org
            version: v1alpha1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `candidate` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the [OpenShift sandboxed containers documentation](https://docs.redhat.com/en/documentation/openshift_sandboxed_containers/).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-cloud-api-adaptor-rhel9@sha256:7d1fa1e85af0d34c23e75868b1388eff17abe633c04a8dc55fbefcc470195c05
        - registry.redhat.io/openshift-sandboxed-containers/osc-cloud-api-adaptor-webhook-rhel9@sha256:6802f00a939f24ecd6dad150433db5ea3fd72d171f0c410cb894d3af910fc7fd
        - registry.redhat.io/openshift-sandboxed-containers/osc-podvm-builder-rhel9@sha256:c15f36a639fa07fee11f74c6563696ec8ef4da3fb1a99b7bd3f308dd04881339
        - registry.redhat.io/openshift-sandboxed-containers/osc-podvm-payload-rhel9@sha256:8d6b1262c6cedd04e33441264539842d82ed057f64eda5c4bce59566bcb72558
        - registry.redhat.io/openshift-sandboxed-containers/osc-dm-verity-image@sha256:e633bc27853f811d919e64461d0c917e659f5430458abe2304a4c85b0d871c80
        - registry.redhat.io/openshift-sandboxed-containers/osc-monitor-rhel9@sha256:037cbc221ae8fda83cdf1093127795ea806ecda53893deb49cdf981e7863774c
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel9-operator@sha256:1ffb432d9a8833854fa5e5262dcbf42530dfbaadfddf6978f7bc7804b84d2e3f
        version: 1.10.3
      entries:
      - name: sandboxed-containers-operator.v1.10.3
        version: 1.10.3
      - name: sandboxed-containers-operator.v1.10.2
        version: 1.10.2
      - name: sandboxed-containers-operator.v1.10.1
        version: 1.10.1
      - name: sandboxed-containers-operator.v1.10.0
        version: 1.10.0
      - name: sandboxed-containers-operator.v1.9.0
        version: 1.9.0
      - name: sandboxed-containers-operator.v1.8.1
        version: 1.8.1
      - name: sandboxed-containers-operator.v1.7.0
        version: 1.7.0
      - name: sandboxed-containers-operator.v1.6.0
        version: 1.6.0
      - name: sandboxed-containers-operator.v1.5.3
        version: 1.5.3
      - name: sandboxed-containers-operator.v1.5.2
        version: 1.5.2
      - name: sandboxed-containers-operator.v1.5.1
        version: 1.5.1
      - name: sandboxed-containers-operator.v1.5.0
        version: 1.5.0
      name: stable
    - currentCSV: sandboxed-containers-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:8f30a9129d817c3f4e404d2c43fb47e196d8c8da3badba4c48f65d440a4d7584
          createdAt: "2021-11-22T14:44:50.817336Z"
          olm.skipRange: '>=1.1.0 <1.2.2'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `stable-1.2` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.9/sandboxed_containers/understanding-sandboxed-containers.html)
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:a660f0b54b9139bed9a3aeef3408001c0d50ba60648364a98a09059b466fbcc1
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:8f30a9129d817c3f4e404d2c43fb47e196d8c8da3badba4c48f65d440a4d7584
        version: 1.2.2
      entries:
      - name: sandboxed-containers-operator.v1.2.2
        version: 1.2.2
      name: stable-1.2
    - currentCSV: sandboxed-containers-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          olm.skipRange: '>=1.1.0 <1.3.3'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/sandboxed-containers-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `stable-1.3` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/latest/sandboxed_containers/understanding-sandboxed-containers.html)
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:5d2b03721043e5221dfb0cf164cf59eba396ba3aae40a56c53aa3496c625eea0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:cb836456974e510eb4bccbffadbc6d99d5f57c36caec54c767a158ffd8a025d5
        version: 1.3.3
      entries:
      - name: sandboxed-containers-operator.v1.3.3
        version: 1.3.3
      name: stable-1.3
    defaultChannel: stable
    packageName: sandboxed-containers-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/stolostron/multicluster-global-hub
    name: multicluster-global-hub-operator-rh
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-global-hub-operator-rh.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:01e8dc62d5dc48af84b92f5315fab22574f648f2f705ac94f0516f4a7dab9237
          createdAt: 27 Aug 2024, 22:21
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.1.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```


          ### Documentation

          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: release-1.1
        provider:
          name: Red Hat
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:01e8dc62d5dc48af84b92f5315fab22574f648f2f705ac94f0516f4a7dab9237
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:78628c4f7a0286417023410261cc3b36797b7cdf72b38d0f4a2533a0f3db89dc
        - registry.redhat.io/rhel9/postgresql-13@sha256:c67d00643295697d196dd70dea457948ebb9e9d9d61a42019490b350c8286b50
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:5c8cea421be39712503378e32c54c332a7a0bf77a1705781d10fc123224bd60e
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:4ea8d3a5317d3b9968fa93dd68416e07e12e00e492860ff20ed90f88ca6ba225
        version: 1.1.2
      entries:
      - name: multicluster-global-hub-operator-rh.v1.1.2
        version: 1.1.2
      - name: multicluster-global-hub-operator-rh.v1.1.1
        version: 1.1.1
      - name: multicluster-global-hub-operator-rh.v1.1.0
        version: 1.1.0
      name: release-1.1
    - currentCSV: multicluster-global-hub-operator-rh.v1.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:98d470c8683c346107ab943776991554e3173c9a6431202bdacc48a75c4cf18f
          createdAt: 15 May 2025, 21:22
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          kafka-zookeeper-0                                       1/1     Running   0          21m
          kafka-zookeeper-1                                       1/1     Running   0          21m
          kafka-zookeeper-2                                       1/1     Running   0          21m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgres-0                      2/2     Running   0          22m
          ```


          ### Documentation

          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: release-1.2
        provider:
          name: Red Hat
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:c83a7b01bfe5a39d6cdf585180a011448e935d109768cb51560308b29d9ffe14
        - registry.redhat.io/rhel9/postgresql-13@sha256:b1af637cbe0e4041cd2fe6fdbcf665c5183e2991d3cc3e5b7384235f801504d9
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:16bb76ad0c68c0d6b3149b2741126a96e154150c1059621b14c2e6c30a214f16
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:98d470c8683c346107ab943776991554e3173c9a6431202bdacc48a75c4cf18f
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:92f44cff160874efe57a985609e599b13c49c0c4cef846dccc80833af2ae74f3
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:64e5ade4fbcd4cb70d62a6a62d775e051ff49838b891b682574646d4e61d9751
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:9b98fc01b08c3262e5ab99096e74fb55982f2a9c0de61e8fe84e419522e725b0
        version: 1.2.3
      entries:
      - name: multicluster-global-hub-operator-rh.v1.2.3
        version: 1.2.3
      - name: multicluster-global-hub-operator-rh.v1.2.2
        version: 1.2.2
      - name: multicluster-global-hub-operator-rh.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator-rh.v1.2.0
        version: 1.2.0
      name: release-1.2
    - currentCSV: multicluster-global-hub-operator-rh.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:0ef24ded54588aa49e366ba1cafc0971162e9d32024045bbcd7ef16e9f38a2d5
          createdAt: 01 Apr 2025, 20:25
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          kafka-zookeeper-0                                       1/1     Running   0          21m
          kafka-zookeeper-1                                       1/1     Running   0          21m
          kafka-zookeeper-2                                       1/1     Running   0          21m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgres-0                      2/2     Running   0          22m
          ```


          ### Documentation

          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: release-1.3
        provider:
          name: Red Hat
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:a139da58c681de987d847b233dd5c1e6ced5d6e86f3aab0f1b4181a77fc583d4
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:a17e90abc5e23bbc3908613c56c4208ad554eb27ed9679c4ccd06c4a1e9fcbcc
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:0ef24ded54588aa49e366ba1cafc0971162e9d32024045bbcd7ef16e9f38a2d5
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:619d8cb6b3811169af183bd8bc26ed98cdb25486ec428d88b25fee0ef9e6c552
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:e848a90d17e4f00d8021cbde47e488b82c447d8a5f6588f85f3d7cfbd6877f0f
        - registry.redhat.io/rhel9/postgresql-13@sha256:76c49b7dbd7c4ed8b430c4931b32f546db6d07ddc2146232f5d17eb17d3243b4
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-kessel-inventory-api-rhel9@sha256:dbd3924eb24db7ef28301b17851b5b5cd20e044cf259f0c556c67a0721561803
        version: 1.3.3
      entries:
      - name: multicluster-global-hub-operator-rh.v1.3.3
        version: 1.3.3
      - name: multicluster-global-hub-operator-rh.v1.3.2
        version: 1.3.2
      - name: multicluster-global-hub-operator-rh.v1.3.1
        version: 1.3.1
      - name: multicluster-global-hub-operator-rh.v1.3.0
        version: 1.3.0
      name: release-1.3
    - currentCSV: multicluster-global-hub-operator-rh.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:c5b3b3907e8135d3eedba40e27ba18a232ea4f30854d652059fcb14f82a8bf15
          createdAt: "2025-07-23T09:04:08Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: release-1.4
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:047b6c378e2b37339dd498c9025060f20812014bde4d0c98183f1a1201c21e66
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:ea906dcf8d45956251df22e269fe0b8fa8d49734e96a61b10db7b25d626b2e6a
        - registry.redhat.io/rhel9/postgresql-16@sha256:2161182c88022f38d6d9b8b467c0d59156edcededc908d383b80caa4eedf57f9
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:2be8bc79667018386235823e322ae10788e4e7e889c10e63566ae27a65dcabf1
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:0f4683d18404b4a9412e42ed417f6d3ac0024b3c57aa72c9162ff72d6ea8d394
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:c5b3b3907e8135d3eedba40e27ba18a232ea4f30854d652059fcb14f82a8bf15
        version: 1.4.2
      entries:
      - name: multicluster-global-hub-operator-rh.v1.4.2
        version: 1.4.2
      - name: multicluster-global-hub-operator-rh.v1.4.1
        version: 1.4.1
      - name: multicluster-global-hub-operator-rh.v1.4.0
        version: 1.4.0
      name: release-1.4
    - currentCSV: multicluster-global-hub-operator-rh.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:8af7e57b8c6285ad232084e7776fef32d026e080eb0bf14b91986ac28d00bb6a
          createdAt: "2025-08-18T08:30:32Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: release-1.5
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:8af7e57b8c6285ad232084e7776fef32d026e080eb0bf14b91986ac28d00bb6a
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:c95052569312d805b6750477c8bc57f7c72e1efd26143f3899ea0eddc2aed37b
        - registry.redhat.io/rhel9/postgresql-16@sha256:2161182c88022f38d6d9b8b467c0d59156edcededc908d383b80caa4eedf57f9
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:921163d139bca9e47bf95df43d7b408ff6fe2bf9b0902fb6af9f4cfc9d4598cc
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:ab3ed25a4448a7d545648ae4cea52b493918cc4d3b0c5adc56bb2510c0838c80
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:eae390ec42c3582f7896a8346df8782b287d64440794983ae93755490639e21e
        version: 1.5.1
      entries:
      - name: multicluster-global-hub-operator-rh.v1.5.1
        version: 1.5.1
      - name: multicluster-global-hub-operator-rh.v1.5.0
        version: 1.5.0
      name: release-1.5
    defaultChannel: release-1.5
    packageName: multicluster-global-hub-operator-rh
    provider:
      name: Red Hat, Inc
      url: https://github.com/stolostron/multicluster-global-hub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: metallb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: metallb-operator.v4.16.0-202509230119
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "basic",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64512,
                        "neighbors": [
                          {
                            "address": "172.30.0.3",
                            "asn": 64512,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            },
                            "toReceive": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "two-neighs",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64512,
                        "neighbors": [
                          {
                            "address": "172.30.0.3",
                            "asn": 4200000000,
                            "ebgpMultiHop": true,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          },
                          {
                            "address": "172.18.0.6",
                            "asn": 4200000000,
                            "port": 179,
                            "toAdvertise": {
                              "allowed": {
                                "prefixes": [
                                  "192.168.2.0/24"
                                ]
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "two-routers-vrf",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64513,
                        "neighbors": [
                          {
                            "address": "172.32.0.3",
                            "asn": 4200000000,
                            "ebgpMultiHop": true,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ]
                      },
                      {
                        "asn": 64515,
                        "neighbors": [
                          {
                            "address": "172.31.0.3",
                            "asn": 4200000000,
                            "port": 179,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ],
                        "vrf": "red"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRNodeState",
                "metadata": {
                  "name": "frr-k8s-worker"
                },
                "status": {
                  "lastConversionResult": "success",
                  "lastReloadResult": "success",
                  "runningConfig": "Building configuration...\n\nCurrent configuration:\n!\nfrr version 9.0.2_git\nfrr defaults traditional\nhostname frr-k8s-worker\nlog file /etc/frr/frr.log informational\nlog timestamp precision 3\nno ipv6 forwarding\nservice integrated-vtysh-config\n!\nip nht resolve-via-default\n!\nipv6 nht resolve-via-default\n!\nend\n"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BFDProfile",
                "metadata": {
                  "name": "bfd-profile-sample",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "receiveInterval": 380,
                  "transmitInterval": 270
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "peers": [
                    "ebgp-single-hop0"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPoolSelectors": [
                    {
                      "matchLabels": {
                        "test": "ipv6"
                      }
                    },
                    {
                      "matchLabels": {
                        "test": "ipv4"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "communities": [
                    "65535:65282"
                  ],
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "localPref": 50
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample4",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "communities": [
                    "NO_ADVERTISE"
                  ],
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "localPref": 50
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample5",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "nodeSelectors": [
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-control-plane"
                      }
                    },
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-worker"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-beta1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "Community",
                "metadata": {
                  "name": "community1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "communities": [
                    {
                      "name": "NO_ADVERTISE",
                      "value": "65535:65282"
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "name": "ip-addresspool-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "192.168.10.0/24",
                    "192.168.9.1-192.168.9.5",
                    "fc00:f853:0ccd:e799::/124"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "labels": {
                    "test": "ipv4"
                  },
                  "name": "ip-addresspool-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "172.20.0.100/24"
                  ],
                  "autoAssign": false
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "labels": {
                    "test": "ipv6"
                  },
                  "name": "ip-addresspool-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "2002:2:2::1-2002:2:2::100"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "L2Advertisement",
                "metadata": {
                  "name": "l2-adv-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "L2Advertisement",
                "metadata": {
                  "name": "l2-adv-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "nodeSelectors": [
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-control-plane"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "MetalLB",
                "metadata": {
                  "name": "metallb",
                  "namespace": "metallb-system"
                },
                "spec": {}
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "bfdProfile": "bfd-profile-sample",
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "holdTime": "9m0s",
                  "keepaliveTime": "0s",
                  "myASN": 64512,
                  "passwordSecret": {
                    "name": "secretname",
                    "namespace": "metallb-system"
                  },
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3",
                  "peerPort": 180
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:b07c046ba2ae71b93f27d5cdca7cb93f51f3198cb84bb6e65dd6320bcda268c8
          createdAt: "2023-06-06T15:25:00Z"
          description: An operator for deploying MetalLB on a kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.8.0 <4.16.0'
          operatorframework.io/suggested-namespace: metallb-system
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/metallb-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BFDProfile is the Schema for the bfdprofiles API
            kind: BFDProfile
            name: bfdprofiles.metallb.io
            version: v1beta1
          - description: BGPAdvertisement is the Schema for the bgpadvertisements
              API
            kind: BGPAdvertisement
            name: bgpadvertisements.metallb.io
            version: v1beta1
          - description: BGPPeer is the Schema for the bgppeers API
            kind: BGPPeer
            name: bgppeers.metallb.io
            version: v1beta1
          - description: BGPPeer is the Schema for the bgppeers API
            kind: BGPPeer
            name: bgppeers.metallb.io
            version: v1beta2
          - description: Community is the Schema for the communities API
            kind: Community
            name: communities.metallb.io
            version: v1beta1
          - description: FRRConfiguration is the Schema for the frrconfigurations
              API
            kind: FRRConfiguration
            name: frrconfigurations.frrk8s.metallb.io
            version: v1beta1
          - description: FRRNodeState is the Schema for the frrnodestates API
            kind: FRRNodeState
            name: frrnodestates.frrk8s.metallb.io
            version: v1beta1
          - description: IPAddressPool is the Schema for the ipaddresspools API
            kind: IPAddressPool
            name: ipaddresspools.metallb.io
            version: v1beta1
          - description: L2Advertisement is the Schema for the l2advertisements API
            kind: L2Advertisement
            name: l2advertisements.metallb.io
            version: v1beta1
          - description: MetalLB is the Schema for the metallbs API
            displayName: MetalLB
            kind: MetalLB
            name: metallbs.metallb.io
            version: v1beta1
        description: An operator for deploying MetalLB on a kubernetes cluster.
        displayName: MetalLB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - metallb-operator
        links:
        - name: MetalLB Operator
          url: https://github.com/openshift/metallb-operator
        maintainers:
        - email: fpaoline@redhat.com
          name: Federico Paolinelli
        - email: obraunsh@redhat.com
          name: Ori Braunshtein
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/metallb-rhel9@sha256:95abda97d689078bd124ef59c92cc12e661222cac8e8b0e526579470ab54d05e
        - registry.redhat.io/openshift4/frr-rhel9@sha256:48a361644b2fa8fc632f49117172db2e9c12022774e7e0fd409561d884a6c237
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:b07c046ba2ae71b93f27d5cdca7cb93f51f3198cb84bb6e65dd6320bcda268c8
        version: 4.16.0-202509230119
      entries:
      - name: metallb-operator.v4.16.0-202509230119
        version: 4.16.0-202509230119
      - name: metallb-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: metallb-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: metallb-operator.v4.16.0-202507221936
        version: 4.16.0-202507221936
      - name: metallb-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: metallb-operator.v4.16.0-202506090335
        version: 4.16.0-202506090335
      - name: metallb-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: metallb-operator.v4.16.0-202505051339
        version: 4.16.0-202505051339
      - name: metallb-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: metallb-operator.v4.16.0-202503252033
        version: 4.16.0-202503252033
      - name: metallb-operator.v4.16.0-202502260004
        version: 4.16.0-202502260004
      - name: metallb-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: metallb-operator.v4.16.0-202501232204
        version: 4.16.0-202501232204
      - name: metallb-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: metallb-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: metallb-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: metallb-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: metallb-operator.v4.16.0-202411110236
        version: 4.16.0-202411110236
      - name: metallb-operator.v4.16.0-202410292005
        version: 4.16.0-202410292005
      - name: metallb-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: metallb-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: metallb-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: metallb-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: metallb-operator.v4.16.0-202409161407
        version: 4.16.0-202409161407
      - name: metallb-operator.v4.16.0-202409062036
        version: 4.16.0-202409062036
      - name: metallb-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: metallb-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: metallb-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: metallb-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: metallb-operator.v4.16.0-202408051837
        version: 4.16.0-202408051837
      - name: metallb-operator.v4.16.0-202407221107
        version: 4.16.0-202407221107
      - name: metallb-operator.v4.16.0-202407180936
        version: 4.16.0-202407180936
      - name: metallb-operator.v4.16.0-202407091337
        version: 4.16.0-202407091337
      - name: metallb-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: metallb-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: metallb-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: metallb-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: metallb-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ian Boden
      provider-url: ""
    name: eclipse-amlen-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eclipse-amlen-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eclipse.org/v1",
                "kind": "Amlen",
                "metadata": {
                  "name": "amlen-sample"
                },
                "spec": {
                  "_amlen_image_tag": "latest",
                  "_amlen_monitor_image_tag": "latest",
                  "_amlen_name": "amlen-sample",
                  "_amlen_namespace": "osdk-test",
                  "_amlen_persistence": "false",
                  "config": "amlen-config",
                  "device_cert_issuer": {
                    "mode": "automatic",
                    "name": "amlen-selfsigned-issuer"
                  },
                  "ha": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "size": 2
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/amlen/operator:v1.0.1a
          createdAt: "2022-11-24T15:31:00Z"
          description: An operator to run the Eclipse Amlen MQTT Broker
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/eclipse/amlen
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Amlen
            name: amlens.eclipse.org
            version: v1
        description: Operator for the eclipse Amlen MQTT Broker\n\n### Supported Feature\n\nCurrently
          supports:\n* Install - simple install process\n* Upgrade - non-disruptive
          upgrade\n*High Availability - can be enabled/disabled via the CRD\n\n###
          Documentation\nThe main Amlen website has links to product documentation,
          forum, blog and the slack channel\nhttp://www.eclipse.org/amlen\n\n### Getting
          help\nThe quickest way to get help is to use the slack channel but the forum
          is monitored as well\n\nIssues can also be added to the github repository
          https://github.com/eclipse/amlen\n\n### License\nEclipse Amlen Operator
          is licensed under the [Eclipse Public License - v 2.0](https://www.eclipse.org/legal/epl-2.0/)
        displayName: eclipse-amlen-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amlen
        - mqtt
        - broker
        links:
        - name: Eclipse Amlen Operator
          url: https://eclipse.org/amlen
        maintainers:
        - email: ianboden@uk.ibm.com
          name: Ian Boden
        maturity: stable
        provider:
          name: Ian Boden
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/amlen/operator:v1.0.1a
        version: 1.0.0
      entries:
      - name: eclipse-amlen-operator.v1.0.0
        version: 1.0.0
      - name: eclipse-amlen-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: eclipse-amlen-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eclipse.org/v1",
                "kind": "Amlen",
                "metadata": {
                  "name": "amlen-sample"
                },
                "spec": {
                  "config": "amlen-config",
                  "device_cert_issuer": {
                    "mode": "automatic",
                    "name": "amlen-selfsigned-issuer"
                  },
                  "ha": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "ldap": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "size": 2
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/amlen/operator:v1.1.3
          createdAt: "2025-05-22T10:00:00Z"
          description: An operator to run the Eclipse Amlen MQTT Broker
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/eclipse-amlen/amlen
          support: Eclipse
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Amlen
            name: amlens.eclipse.org
            version: v1
        description: |-
          Operator for the eclipse Amlen MQTT Broker

          ### Supported Feature

          Currently supports:
          * Install - simple install process
          * Upgrade - non-disruptive upgrade
          * High Availability - can be enabled/disabled via the CRD

          ### Documentation
          The main Amlen website has links to product documentation, forum, blog and the slack channel
          http://www.eclipse.org/amlen

          ### Getting help
          The quickest way to get help is to use the slack channel but the forum is monitored as well
          Issues can also be added to the github repository https://github.com/eclipse-amlen/amlen

          ### License
          Eclipse Amlen Operator is licensed under [Eclipse Public License - v 2.0](https://www.eclipse.org/legal/epl-2.0/)
        displayName: eclipse-amlen-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amlen
        - mqtt
        - broker
        links:
        - name: Eclipse Amlen Operator
          url: https://eclipse.org/amlen
        maintainers:
        - email: ianboden@uk.ibm.com
          name: Ian Boden
        maturity: stable
        provider:
          name: Eclipse Amlen
          url: https://eclipse.org/amlen
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - controller:latest
        version: 1.1.3
      entries:
      - name: eclipse-amlen-operator.v1.1.3
        version: 1.1.3
      - name: eclipse-amlen-operator.v1.1.0
        version: 1.1.0
      - name: eclipse-amlen-operator.v1.0.2
        version: 1.0.2
      - name: eclipse-amlen-operator.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: alpha
    packageName: eclipse-amlen-operator
    provider:
      name: Ian Boden
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: vault-config-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vault-config-operator.v0.8.34
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AuthEngineMount",
                "metadata": {
                  "name": "authenginemount-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kube-authengine-mount-sample",
                  "type": "kubernetes"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azureauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azureauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azureauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureCredentials": {
                    "passwordKey": "clientsecret",
                    "secret": {
                      "name": "aad-credentials"
                    },
                    "usernameKey": "clientid"
                  },
                  "clientID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "environment": "AzurePublicCloud",
                  "maxRetries": 3,
                  "maxRetryDelay": 60,
                  "path": "azure/test",
                  "resource": "https://management.azure.com/",
                  "retryDelay": 4,
                  "tenantID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azureauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azureauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azureauthenginerole-sample"
                },
                "spec": {
                  "BoundResourceGroups": [
                    "resourcegroup1",
                    "resourcegroup2"
                  ],
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "boundGroupIDs": [
                    "group1",
                    "group2"
                  ],
                  "boundLocations": [
                    "location1",
                    "location2"
                  ],
                  "boundScaleSets": [
                    "scaleset1",
                    "scaleset1"
                  ],
                  "boundServicePrincipalIDs": [
                    "sp1",
                    "sp2"
                  ],
                  "boundSubscriptionIDs": [
                    "subscription1",
                    "subscription2"
                  ],
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "dev-role",
                  "path": "azure",
                  "policies": [
                    "policy1",
                    "policy2"
                  ],
                  "tokenBoundCIDRs": [
                    "10.20.102.13",
                    "10.20.102.15"
                  ],
                  "tokenExplicitMaxTTL": "",
                  "tokenMaxTTL": "",
                  "tokenNoDefaultPolicy": false,
                  "tokenNumUses": 0,
                  "tokenPeriod": 0,
                  "tokenPolicies": [
                    "policy1",
                    "policy2"
                  ],
                  "tokenTTL": "",
                  "tokenType": ""
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureSecretEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azuresecretengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azuresecretengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azuresecretengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureCredentials": {
                    "passwordKey": "clientsecret",
                    "secret": {
                      "name": "aad-credentials"
                    },
                    "usernameKey": "clientid"
                  },
                  "clientID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "environment": "AzurePublicCloud",
                  "passwordPolicy": "",
                  "path": "azure",
                  "rootPasswordTTL": "128d",
                  "subscriptionID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "tenantID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureSecretEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azuresecretenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azuresecretenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azuresecretenginerole-sample"
                },
                "spec": {
                  "TTL": "",
                  "applicationObjectID": "",
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureGroups": "",
                  "azureRoles": "",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "maxTTL": "",
                  "name": "azure-role",
                  "path": "azure",
                  "permanentlyDelete": "",
                  "persistApp": "",
                  "signInAudience": "",
                  "tags": ""
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "CertAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "certauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "certauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "certauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "test",
                  "path": "cert"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "CertAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "certauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "certauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "certauthenginerole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "certificate": "-----BEGIN CERTIFICATE-----\nMIIDJTCCAg2gAwIBAgIUF119kmCNlnMwK6yOOpfRs4JGhAAwDQYJKoZIhvcNAQEL\nBQAwFjEUMBIGA1UEAwwLZXhhbXBsZS5jb20wHhcNMjUwMTIxMTM0MjExWhcNMjYw\nMTIxMTM0MjExWjAWMRQwEgYDVQQDDAtleGFtcGxlLmNvbTCCASIwDQYJKoZIhvcN\nAQEBBQADggEPADCCAQoCggEBALad4Sx97VyFSfOepJoXUGYljg7hL+z20pVL2/M8\n8HXAQl4qf/i45b//Qz8VogMxdg3Y9J7O/M0n9hxbL18i1gm270VjEFyzjJCyUq0R\n72WMM4Sr3OjQPnVtJoQkWnH7wR3zsWK2N4NuIPdE3Ofb48Olyachffv7xCI+MLIa\n/adcyRggM4Z5f2nBCMcXL8D/eY03qCgd/U1aaPJ0WyJTZHLL4euL0ewKlKtB1JnU\n4MrRa5mbPQ3B+OhxcaD/7BNV6hPXAe2zftEv8LcimhbVoj3XmLJ+om58YIaEMadk\nJj61xnXbf+F/iV41onzwvTPNWCSWiLowDl8P3zkabuS583sCAwEAAaNrMGkwHQYD\nVR0OBBYEFH971CwrqkTvUz2qrUncWMS5ULwKMB8GA1UdIwQYMBaAFH971CwrqkTv\nUz2qrUncWMS5ULwKMA8GA1UdEwEB/wQFMAMBAf8wFgYDVR0RBA8wDYILZXhhbXBs\nZS5jb20wDQYJKoZIhvcNAQELBQADggEBAIQzT113TQC9BqitOCOoDFCJqcM+BHmf\nITEKlEhlqubPvA40lD4wkz67RP2SdQbF1glzZW5sDTxMjpHL+8nhdmo65CJpsQGP\nyMPwnOvx7W5JWQJa1Yt+48JNhgIymDMun+xwDsjNnlBqp0Kr7HkOxPimMPWN4R6B\nydG96puDGBYqxB1ujx5seOGfGvxPBBr3h8kNg054rtIIDrUG2iUTfpNIILJrd8Q2\nshop9//3VcIpUJO7Z1brvK/Hp4DrcKBZVo7vd8n+5uCx3vBA0tjqNAXYtFnSRvDZ\n6XV+rCZecDBFkaLuJ5YgYEK15gGe6lQAez6DTVLXMAKhZ52kceefnwM=\n-----END CERTIFICATE-----\n",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "test-role",
                  "path": "cert/test"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineConfig",
                "metadata": {
                  "name": "my-postgresql-database"
                },
                "spec": {
                  "allowedRoles": [
                    "read-only"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "connectionURL": "postgresql://{{username}}:{{password}}@my-postgresql-database.test-vault-config-operator.svc:5432",
                  "path": "test-vault-config-operator/database",
                  "pluginName": "postgresql-database-plugin",
                  "rootCredentials": {
                    "passwordKey": "postgresql-password",
                    "secret": {
                      "name": "postgresql-admin-password"
                    }
                  },
                  "username": "postgres"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineRole",
                "metadata": {
                  "name": "read-only"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "creationStatements": [
                    "CREATE ROLE \"{{name}}\" WITH LOGIN PASSWORD '{{password}}' VALID UNTIL '{{expiration}}'; GRANT SELECT ON ALL TABLES IN SCHEMA public TO \"{{name}}\";"
                  ],
                  "dBName": "my-postgresql-database",
                  "defaultTTL": "1h",
                  "maxTTL": "24h",
                  "path": "test-vault-config-operator/database"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineStaticRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "databasesecretenginestaticrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "databasesecretenginestaticrole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "databasesecretenginestaticrole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "dBName": "db-name",
                  "path": "test-vault-config-operator/database",
                  "rotationPeriod": "24h",
                  "rotationStatements": [
                    "ALTER ROLE \"{{name}}\" WITH PASSWORD '{{password}}';git add"
                  ],
                  "username": "db-user"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GCPAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "gcpauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gcpauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "gcpauthengineconfig-sample"
                },
                "spec": {
                  "GCEalias": "role_id",
                  "GCEmetadata": "default",
                  "GCPCredentials": {
                    "passwordKey": "credentials",
                    "secret": {
                      "name": "gcp-serviceaccount-credentials"
                    },
                    "usernameKey": "serviceaccount"
                  },
                  "IAMalias": "default",
                  "IAMmetadata": "default",
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "customEndpoint": {},
                  "path": "azure",
                  "serviceAccount": "vault-sa"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GCPAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "gcpauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gcpauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "gcpauthenginerole-sample"
                },
                "spec": {
                  "addGroupAliases": false,
                  "allowGCEInference": false,
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "boundInstanceGroups": [],
                  "boundLabels": [],
                  "boundProjects": {},
                  "boundRegions": [],
                  "boundServiceAccounts": {},
                  "boundZones": [],
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "maxJWTExp": "",
                  "name": "gcp-iam-role",
                  "path": "gcp",
                  "policies": [],
                  "tokenBoundCIDRs": [],
                  "tokenExplicitMaxTTL": "",
                  "tokenMaxTTL": "",
                  "tokenNoDefaultPolicy": false,
                  "tokenNumUses": 0,
                  "tokenPeriod": 0,
                  "tokenPolicies": [],
                  "tokenTTL": "",
                  "tokenType": "",
                  "type": "iam"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitHubSecretEngineConfig",
                "metadata": {
                  "name": "githubsecretengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitHubSecretEngineRole",
                "metadata": {
                  "name": "githubsecretenginerole-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Group",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "group-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "group",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "group-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "metadata": {
                    "team": "team-abc"
                  },
                  "policies": [
                    "team-abc-access"
                  ],
                  "type": "external"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupAlias",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "groupalias-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "groupalias",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "groupalias-sample"
                },
                "spec": {
                  "authEngineMountPath": "kubernetes",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "groupName": "group-sample"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "JWTOIDCAuthEngineConfig",
                "metadata": {
                  "name": "azure-oidc-config"
                },
                "spec": {
                  "OIDCClientID": "000000000-0000-0000-0000-000000000",
                  "OIDCCredentials": {
                    "passwordKey": "oidc_client_secret",
                    "secret": {
                      "name": "oidccredentials"
                    },
                    "usernameKey": "oidc_client_id"
                  },
                  "OIDCDiscoveryURL": "https://login.microsoftonline.com/000000000-0000-0000-0000-000000000/v2.0",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "oidc/azuread-oidc/"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "JWTOIDCAuthEngineRole",
                "metadata": {
                  "name": "azure-oidc-dev-role"
                },
                "spec": {
                  "OIDCScopes": [
                    "https://graph.microsoft.com/.default"
                  ],
                  "allowedRedirectURIs": [
                    "http://localhost:8250/oidc/callback",
                    "http://localhost:8200/ui/vault/auth/oidc/azuread-oidc/oidc/callback"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "groupsClaim": "groups",
                  "name": "dev-role",
                  "path": "oidc-aad/azuread-oidc",
                  "roleType": "oidc",
                  "tokenPolicies": [
                    "dev",
                    "prod"
                  ],
                  "userClaim": "email"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesAuthEngineConfig",
                "metadata": {
                  "name": "kubernetesauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kube-authengine-mount-sample"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesAuthEngineRole",
                "metadata": {
                  "name": "database-engine-admin"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kubernetes",
                  "policies": [
                    "database-engine-admin"
                  ],
                  "targetNamespaces": {
                    "targetNamespaceSelector": {
                      "matchLabels": {
                        "database-engine-admin": "true"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesSecretEngineConfig",
                "metadata": {
                  "name": "kubernetessecretengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesSecretEngineRole",
                "metadata": {
                  "name": "kubernetessecretenginerole-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "LDAPAuthEngineConfig",
                "metadata": {
                  "name": "ldapauthengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "LDAPAuthEngineGroup",
                "metadata": {
                  "name": "ldapauthenginegroup-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin",
                    "serviceAccount": {
                      "name": "default"
                    }
                  },
                  "name": "test-group",
                  "path": "ldap/test"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PKISecretEngineConfig",
                "metadata": {
                  "name": "pkisecretengineconfig-sample"
                },
                "spec": {
                  "TTL": "8760h",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "commonName": "pki-vault-demo.internal.io",
                  "crlDistributionPoints": [
                    "https://vault-internal.vault.svc:8200/v1/test-vault-config-operator/pki/crl\""
                  ],
                  "issuingCertificates": [
                    "https://vault-internal.vault.svc:8200/v1/test-vault-config-operator/pki/ca"
                  ],
                  "path": "test-vault-config-operator/pki",
                  "privateKeyType": "internal",
                  "type": "root"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PKISecretEngineRole",
                "metadata": {
                  "name": "pkisecretenginerole-sample"
                },
                "spec": {
                  "allowGlobDomains": true,
                  "allowSubdomains": true,
                  "allowedDomains": [
                    "internal.io",
                    "pki-vault-demo.svc",
                    "example.com"
                  ],
                  "allowedOtherSans": "*",
                  "allowedURISans": [
                    "*-pki-vault-demo.apps.example.com"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "maxTTL": "8760h",
                  "path": "test-vault-config-operator/pki"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PasswordPolicy",
                "metadata": {
                  "name": "simple-password-policy"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "passwordPolicy": "length = 20\nrule \"charset\" {\n  charset = \"abcdefghijklmnopqrstuvwxyz\"\n}"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Policy",
                "metadata": {
                  "name": "database-engine-admin"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "policy": "# mount database secrets engines\npath \"/sys/mounts/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n  allowed_parameters = {\n    \"type\" = [\"database\"]\n    \"*\"   = []\n  }\n}\n\n# tune database secrets engines\npath \"/sys/mounts/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/tune\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n\n# Configure database secrets engines\npath \"/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/config/+\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n\n# Configure database roles\npath \"/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/roles/+\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineConfig",
                "metadata": {
                  "name": "quaysecretengineconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineRole",
                "metadata": {
                  "name": "quaysecretenginerole-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineStaticRole",
                "metadata": {
                  "name": "quaysecretenginestaticrole-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RabbitMQSecretEngineConfig",
                "metadata": {
                  "name": "rabbitmqsecretengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "rabbitMQ-engine-admin"
                  },
                  "connectionURI": "https://test.com",
                  "leaseMaxTTL": 86400,
                  "leaseTTL": 86400,
                  "path": "vault-config-operator/rabbitmq"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RabbitMQSecretEngineRole",
                "metadata": {
                  "name": "rabbitmqsecretenginerole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "rabbitmq-engine-admin"
                  },
                  "path": "vault-config-operator/rabbitmq",
                  "tags": "administrator",
                  "vhostTopics": [
                    {
                      "topics": [
                        {
                          "permissions": {
                            "configure": ".*",
                            "read": ".*",
                            "write": ".*"
                          },
                          "topicName": "my-topic"
                        },
                        {
                          "permissions": {
                            "read": ".*"
                          },
                          "topicName": "my-read-topic"
                        }
                      ],
                      "vhostName": "/"
                    }
                  ],
                  "vhosts": [
                    {
                      "permissions": {
                        "configure": ".*",
                        "read": ".*",
                        "write": ".*"
                      },
                      "vhostName": "/"
                    },
                    {
                      "permissions": {
                        "read": "my-queue",
                        "write": "my-exchange"
                      },
                      "vhostName": "my-vhost"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RandomSecret",
                "metadata": {
                  "name": "randomsecret-password"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "secret-writer"
                  },
                  "path": "test-vault-config-operator/kv",
                  "secretFormat": {
                    "passwordPolicyName": "simple-password-policy"
                  },
                  "secretKey": "password"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "SecretEngineMount",
                "metadata": {
                  "name": "database"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "path": "test-vault-config-operator",
                  "type": "database"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "VaultSecret",
                "metadata": {
                  "name": "vaultsecret-sample"
                },
                "spec": {
                  "output": {
                    "annotations": {
                      "refresh": "every-minute"
                    },
                    "labels": {
                      "app": "test-vault-config-operator"
                    },
                    "name": "randomsecret",
                    "stringData": {
                      "anotherpassword": "{{ .anotherrandomsecret.password }}",
                      "password": "{{ .randomsecret.password }}"
                    },
                    "type": "Opaque"
                  },
                  "refreshPeriod": "3m0s",
                  "vaultSecretDefinitions": [
                    {
                      "authentication": {
                        "path": "kubernetes",
                        "role": "secret-reader",
                        "serviceAccount": {
                          "name": "default"
                        }
                      },
                      "name": "randomsecret",
                      "path": "test-vault-config-operator/kv/randomsecret-password"
                    },
                    {
                      "authentication": {
                        "path": "kubernetes",
                        "role": "secret-reader",
                        "serviceAccount": {
                          "name": "default"
                        }
                      },
                      "name": "anotherrandomsecret",
                      "path": "test-vault-config-operator/kv/another-password"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/vault-config-operator@sha256:490852417fac37f3003ab7c1ee797f9035fd2905c354cc46eca5e33d2ad780ce
          createdAt: "2025-09-18T18:56:36Z"
          description: This operator provides primitives to declaratively configure
            Hashicorp Vault.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: vault-config-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/vault-config-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthEngineMount is the Schema for the authenginemounts API
            displayName: Auth Engine Mount
            kind: AuthEngineMount
            name: authenginemounts.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureAuthEngineConfig is the Schema for the azureauthengineconfigs
              API
            displayName: Azure Auth Engine Config
            kind: AzureAuthEngineConfig
            name: azureauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureAuthEngineRole is the Schema for the azureauthengineroles
              API
            displayName: Azure Auth Engine Role
            kind: AzureAuthEngineRole
            name: azureauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - kind: AzureSecretEngineConfig
            name: azuresecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureSecretEngineRole is the Schema for the azuresecretengineroles
              API
            displayName: Azure Secret Engine Role
            kind: AzureSecretEngineRole
            name: azuresecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: CertAuthEngineConfig is the Schema for the certauthengineconfigs
              API
            displayName: Cert Auth Engine Config
            kind: CertAuthEngineConfig
            name: certauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: CertAuthEngineRole is the Schema for the certauthengineroles
              API
            displayName: Cert Auth Engine Role
            kind: CertAuthEngineRole
            name: certauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineConfig is the Schema for the databasesecretengineconfigs
              API
            displayName: Database Secret Engine Config
            kind: DatabaseSecretEngineConfig
            name: databasesecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineRole is the Schema for the databasesecretengineroles
              API
            displayName: Database Secret Engine Role
            kind: DatabaseSecretEngineRole
            name: databasesecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineStaticRole is the Schema for the databasesecretenginestaticroles
              API
            displayName: Database Secret Engine Static Role
            kind: DatabaseSecretEngineStaticRole
            name: databasesecretenginestaticroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GCPAuthEngineConfig is the Schema for the gcpauthengineconfigs
              API
            displayName: GCPAuth Engine Config
            kind: GCPAuthEngineConfig
            name: gcpauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: GCPAuthEngineRole is the Schema for the gcpauthengineroles
              API
            displayName: GCPAuth Engine Role
            kind: GCPAuthEngineRole
            name: gcpauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GitHubSecretEngineConfig is the Schema for the githubsecretengineconfigs
              API
            displayName: Git Hub Secret Engine Config
            kind: GitHubSecretEngineConfig
            name: githubsecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: GitHubSecretEngineRole is the Schema for the githubsecretengineroles
              API
            displayName: Git Hub Secret Engine Role
            kind: GitHubSecretEngineRole
            name: githubsecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GroupAlias is the Schema for the groupaliases API
            displayName: Group Alias
            kind: GroupAlias
            name: groupaliases.redhatcop.redhat.io
            version: v1alpha1
          - description: Group is the Schema for the groups API
            displayName: Group
            kind: Group
            name: groups.redhatcop.redhat.io
            version: v1alpha1
          - description: JWTOIDCAuthEngineConfig is the Schema for the jwtoidcauthengineconfigs
              API
            displayName: JWTOIDCAuth Engine Config
            kind: JWTOIDCAuthEngineConfig
            name: jwtoidcauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: JWTOIDCAuthEngineRole is the Schema for the jwtoidcauthengineroles
              API
            displayName: JWTOIDCAuth Engine Role
            kind: JWTOIDCAuthEngineRole
            name: jwtoidcauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesAuthEngineConfig is the Schema for the kubernetesauthengineconfigs
              API
            displayName: Kubernetes Auth Engine Config
            kind: KubernetesAuthEngineConfig
            name: kubernetesauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesAuthEngineRole can be used to define a KubernetesAuthEngineRole
              for the kube-auth authentication method
            displayName: Kubernetes Auth Engine Role
            kind: KubernetesAuthEngineRole
            name: kubernetesauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesSecretEngineConfig is the Schema for the kubernetessecretengineconfigs
              API
            displayName: Kubernetes Secret Engine Config
            kind: KubernetesSecretEngineConfig
            name: kubernetessecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesSecretEngineRole is the Schema for the kubernetessecretengineroles
              API
            displayName: Kubernetes Secret Engine Role
            kind: KubernetesSecretEngineRole
            name: kubernetessecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: LDAPAuthEngineConfig is the Schema for the ldapauthengineconfigs
              API
            displayName: LDAPAuth Engine Config
            kind: LDAPAuthEngineConfig
            name: ldapauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: LDAPAuthEngineGroup is the Schema for the ldapauthenginegroups
              API
            displayName: LDAPAuth Engine Group
            kind: LDAPAuthEngineGroup
            name: ldapauthenginegroups.redhatcop.redhat.io
            version: v1alpha1
          - description: PasswordPolicy is the Schema for the passwordpolicies API
            displayName: Password Policy
            kind: PasswordPolicy
            name: passwordpolicies.redhatcop.redhat.io
            version: v1alpha1
          - description: PKISecretEngineConfig is the Schema for the pkisecretengineconfigs
              API
            displayName: PKISecret Engine Config
            kind: PKISecretEngineConfig
            name: pkisecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: PKISecretEngineRole is the Schema for the pkisecretengineroles
              API
            displayName: PKISecret Engine Role
            kind: PKISecretEngineRole
            name: pkisecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: Policy is the Schema for the policies API
            displayName: Policy
            kind: Policy
            name: policies.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineConfig is the Schema for the quaysecretengineconfigs
              API
            displayName: Quay Secret Engine Config
            kind: QuaySecretEngineConfig
            name: quaysecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineRole is the Schema for the quaysecretengineroles
              API
            displayName: Quay Secret Engine Role
            kind: QuaySecretEngineRole
            name: quaysecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineStaticRole is the Schema for the quaysecretenginestaticroles
              API
            displayName: Quay Secret Engine Static Role
            kind: QuaySecretEngineStaticRole
            name: quaysecretenginestaticroles.redhatcop.redhat.io
            version: v1alpha1
          - description: RabbitMQSecretEngineConfig is the Schema for the rabbitmqsecretengineconfigs
              API
            displayName: Rabbit MQSecret Engine Config
            kind: RabbitMQSecretEngineConfig
            name: rabbitmqsecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: RabbitMQSecretEngineRole is the Schema for the rabbitmqsecretengineroles
              API
            displayName: Rabbit MQSecret Engine Role
            kind: RabbitMQSecretEngineRole
            name: rabbitmqsecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: RandomSecret is the Schema for the randomsecrets API
            displayName: Random Secret
            kind: RandomSecret
            name: randomsecrets.redhatcop.redhat.io
            version: v1alpha1
          - description: SecretEngineMount is the Schema for the secretenginemounts
              API
            displayName: Secret Engine Mount
            kind: SecretEngineMount
            name: secretenginemounts.redhatcop.redhat.io
            version: v1alpha1
          - description: VaultSecret is the Schema for the vaultsecrets API
            displayName: Vault Secret
            kind: VaultSecret
            name: vaultsecrets.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator helps set up Vault Configurations. The main intent is to do so such that subsequently pods can consume the secrets made available.
          There are two main principles through all of the capabilities of this operator:

          1. high-fidelity API. The CRD exposed by this operator reflect field by field the Vault APIs. This is because we don't want to make any assumption on the kinds of configuration workflow that user will set up. That being said the Vault API is very extensive and we are starting with enough API coverage to support, we think, some simple and very common configuration workflows.
          2. attention to security (after all we are integrating with a security tool). To prevent credential leaks we give no permissions to the operator itself against Vault. All APIs exposed by this operator contains enough information to authenticate to Vault using a local service account (local to the namespace where the API exist). In other word for a namespace user to be abel to successfully configure Vault, a service account in that namespace must have been previously given the needed Vault permissions.

          Currently this operator supports the following CRDs:

          1. [Policy](https://github.com/redhat-cop/vault-config-operator#policy) Configures Vault [Policies](https://www.vaultproject.io/docs/concepts/policies)
          2. [KubernetesAuthEngineRole](https://github.com/redhat-cop/vault-config-operator#KubernetesAuthEngineRole) Configures a Vault [Kubernetes Authentication](https://www.vaultproject.io/docs/auth/kubernetes) Role
          3. [SecretEngineMount](https://github.com/redhat-cop/vault-config-operator#SecretEngineMount) Configures a Mount point for a [SecretEngine](https://www.vaultproject.io/docs/secrets)
          4. [DatabaseSecretEngineConfig](https://github.com/redhat-cop/vault-config-operator#DatabaseSecretEngineConfig) Configures a [Database Secret Engine](https://www.vaultproject.io/docs/secrets/databases) Connection
          5. [DatabaseSecretEngineRole](https://github.com/redhat-cop/vault-config-operator#DatabaseSecretEngineRole) Configures a [Database Secret Engine](https://www.vaultproject.io/docs/secrets/databases) Role
          6. [RandomSecret](https://github.com/redhat-cop/vault-config-operator#RandomSecret) Creates a random secret in a vault [kv Secret Engine](https://www.vaultproject.io/docs/secrets/kv) with one password field generated using a [PasswordPolicy](https://www.vaultproject.io/docs/concepts/password-policies)vault-config-operator
        displayName: Vault Config Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - secrets
        - gitops
        - policy
        - credentials
        links:
        - name: Vault Config Operator
          url: https://github.com/redhat-cop/vault-config-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/vault-config-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/vault-config-operator@sha256:490852417fac37f3003ab7c1ee797f9035fd2905c354cc46eca5e33d2ad780ce
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 0.8.34
      entries:
      - name: vault-config-operator.v0.8.34
        version: 0.8.34
      - name: vault-config-operator.v0.8.33
        version: 0.8.33
      - name: vault-config-operator.v0.8.32
        version: 0.8.32
      - name: vault-config-operator.v0.8.31
        version: 0.8.31
      - name: vault-config-operator.v0.8.30
        version: 0.8.30
      - name: vault-config-operator.v0.8.29
        version: 0.8.29
      - name: vault-config-operator.v0.8.28
        version: 0.8.28
      - name: vault-config-operator.v0.8.27
        version: 0.8.27
      - name: vault-config-operator.v0.8.25
        version: 0.8.25
      - name: vault-config-operator.v0.8.24
        version: 0.8.24
      - name: vault-config-operator.v0.8.23
        version: 0.8.23
      - name: vault-config-operator.v0.8.22
        version: 0.8.22
      - name: vault-config-operator.v0.8.21
        version: 0.8.21
      - name: vault-config-operator.v0.8.20
        version: 0.8.20
      - name: vault-config-operator.v0.8.19
        version: 0.8.19
      - name: vault-config-operator.v0.8.18
        version: 0.8.18
      - name: vault-config-operator.v0.8.17
        version: 0.8.17
      - name: vault-config-operator.v0.8.16
        version: 0.8.16
      - name: vault-config-operator.v0.8.15
        version: 0.8.15
      - name: vault-config-operator.v0.8.14
        version: 0.8.14
      - name: vault-config-operator.v0.8.13
        version: 0.8.13
      - name: vault-config-operator.v0.8.12
        version: 0.8.12
      - name: vault-config-operator.v0.8.11
        version: 0.8.11
      - name: vault-config-operator.v0.8.10
        version: 0.8.10
      - name: vault-config-operator.v0.8.9
        version: 0.8.9
      - name: vault-config-operator.v0.8.7
        version: 0.8.7
      - name: vault-config-operator.v0.8.5
        version: 0.8.5
      - name: vault-config-operator.v0.8.4
        version: 0.8.4
      - name: vault-config-operator.v0.8.3
        version: 0.8.3
      - name: vault-config-operator.v0.8.2
        version: 0.8.2
      - name: vault-config-operator.v0.8.0
        version: 0.8.0
      - name: vault-config-operator.v0.7.1
        version: 0.7.1
      - name: vault-config-operator.v0.6.5
        version: 0.6.5
      - name: vault-config-operator.v0.6.4
        version: 0.6.4
      - name: vault-config-operator.v0.6.3
        version: 0.6.3
      - name: vault-config-operator.v0.6.2
        version: 0.6.2
      - name: vault-config-operator.v0.6.1
        version: 0.6.1
      - name: vault-config-operator.v0.6.0
        version: 0.6.0
      - name: vault-config-operator.v0.5.1
        version: 0.5.1
      - name: vault-config-operator.v0.5.0
        version: 0.5.0
      - name: vault-config-operator.v0.4.1
        version: 0.4.1
      - name: vault-config-operator.v0.4.0
        version: 0.4.0
      - name: vault-config-operator.v0.3.1
        version: 0.3.1
      - name: vault-config-operator.v0.3.0
        version: 0.3.0
      - name: vault-config-operator.v0.2.0
        version: 0.2.0
      - name: vault-config-operator.v0.1.2
        version: 0.1.2
      - name: vault-config-operator.v0.1.1
        version: 0.1.1
      - name: vault-config-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: vault-config-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-rest-api-browser-v1
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-rest-api-browser-v1.v0.1.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rest.loadcore.keysight.com/v1alpha1",
                "kind": "Apibrowser",
                "metadata": {
                  "name": "apibrowser-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-06-29T14:08:05Z"
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Apibrowser
            name: apibrowsers.rest.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Rest API Browser Operator v1
        displayName: Keysight Rest API Browser Operator v1
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Rest Api Browser V1
          url: https://keysight-rest-api-browser-v1.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5d5061dbc2ec5082cf14b6c600fb5400b83cf91d7ccebfa80680a238d275db
        - prod.harbor.keysight.digital/loadcore/keysight-rest-api-browser-v1@sha256:3ada6a0fa630acb0fa962a11d2c272ff679dded1106c4693d2c8f5ec9ab414a4
        version: 0.1.11
      entries:
      - name: keysight-rest-api-browser-v1.v0.1.11
        version: 0.1.11
      name: alpha
    defaultChannel: alpha
    packageName: keysight-rest-api-browser-v1
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Qualcomm Technologies, Inc.
      provider-url: https://github.com/quic/aic-operator
    name: aic-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aic-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aic.quicinc.com/v1",
                "kind": "AIC",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "aic-operator",
                    "app.kubernetes.io/instance": "aic-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "aic",
                    "app.kubernetes.io/part-of": "aic-operator"
                  },
                  "name": "aic-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "false"
          containerImage: ghcr.io/quic/cloud_ai_openshift_operator@sha256:9cf581e9b5c9cd08819fab8f8b6313836a0a20098ad4c99b80d93dbcf5ab32bf
          description: Automate the management of OpenShift Operator for Qualcomm
            AI 100 accelerators.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: Qualcomm Technologies, Inc.
          repository: https://github.com/quic/aic-operator
          support: Qualcomm Technologies, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AIC is the Schema for the aics API
            displayName: AIC
            kind: AIC
            name: aics.aic.quicinc.com
            version: v1
        description: The Openshift Operator enables the Qualcomm® Cloud AI series
          of hardware on OpenShift clusters by automating the configuration and installation
          of their Linux device drivers and its Kubernetes Device Plugin.
        displayName: AIC Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AIC
        - AIC-Operator
        links:
        - name: AIC Operator
          url: https://github.com/quic/aic-operator
        maintainers:
        - email: quic_vmellemp@quicinc.com
          name: quic_vmellemp
        - email: quic_carls@quicinc.com
          name: quic_carls
        maturity: alpha
        provider:
          name: Qualcomm Technologies, Inc.
          url: https://github.com/quic/aic-operator
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:7a943db9bcb22f817d75341621472f4bb30abc5625c53b9a781d97d431bc4d09
        - ghcr.io/quic/cloud_ai_openshift_operator@sha256:9cf581e9b5c9cd08819fab8f8b6313836a0a20098ad4c99b80d93dbcf5ab32bf
        version: 0.3.1
      entries:
      - name: aic-operator.v0.3.1
        version: 0.3.1
      name: alpha
    defaultChannel: alpha
    packageName: aic-operator
    provider:
      name: Qualcomm Technologies, Inc.
      url: https://github.com/quic/aic-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kong Inc.
      provider-url: https://konghq.com
    name: kong-gateway-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kong-gateway-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "ControlPlane",
                "metadata": {
                  "name": "controlplane-example"
                },
                "spec": {
                  "dataplane": "dataplane-example",
                  "gatewayClass": "kong"
                }
              },
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "DataPlane",
                "metadata": {
                  "name": "dataplane-example"
                },
                "spec": {
                  "env": [
                    {
                      "name": "KONG_LOG_LEVEL",
                      "value": "debug"
                    }
                  ]
                }
              },
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "GatewayConfiguration",
                "metadata": {
                  "name": "kong",
                  "namespace": "default"
                },
                "spec": {
                  "controlPlaneDeploymentOptions": {
                    "env": [
                      {
                        "name": "CONTROLLER_LOG_LEVEL",
                        "value": "debug"
                      }
                    ]
                  },
                  "dataPlaneDeploymentOptions": {
                    "containerImage": "kong",
                    "version": "2.8"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking,Integration & Delivery
          operatorframework.io/suggested-namespace: kong-system
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ControlPlane is the Schema for the controlplanes API
            displayName: Control Plane
            kind: ControlPlane
            name: controlplanes.gateway-operator.konghq.com
            version: v1alpha1
          - description: DataPlane is the Schema for the dataplanes API
            displayName: Data Plane
            kind: DataPlane
            name: dataplanes.gateway-operator.konghq.com
            version: v1alpha1
          - description: GatewayConfiguration is the Schema for the gatewayconfigurations
              API
            displayName: Gateway Configuration
            kind: GatewayConfiguration
            name: gatewayconfigurations.gateway-operator.konghq.com
            version: v1alpha1
        description: Kong Operator for Kubernetes. Technical preview.
        displayName: Kong
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - gateway
        - ingress
        - kong
        - kubernetes
        - go
        links:
        - name: Github repository for the operator
          url: https://github.com/kong/gateway-operator
        maintainers:
        - email: support@konghq.com
          name: Kong Inc.
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Kong Inc.
          url: https://konghq.com
        relatedImages:
        - registry.connect.redhat.com/kong/kong-ingress-controller@sha256:f691db070e7c604da9e163800903d0942ec6fda879eb9dcd5d8e90618ae787da
        - ghcr.io/kong/gateway-operator@sha256:76b227715a3e1915a28f8aaa1e2ff8495349d29d0d61e7b85a72d74b03fbcc7f
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.connect.redhat.com/kong/kong@sha256:54e7c7e57bf93810a7cc36327023b8699f754c4c40c5c289548d54170caa5db2
        version: 0.0.2
      entries:
      - name: kong-gateway-operator.v0.0.2
        version: 0.0.2
      - name: kong-gateway-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: kong-gateway-operator
    provider:
      name: Kong Inc.
      url: https://konghq.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: https://www.turbonomic.com/
    name: t8c-tsc-client-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-tsc-client-certified.v1.4.2-candidate-20250806145822000
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "TurbonomicClient",
                "metadata": {
                  "name": "turbonomicclient-release"
                },
                "spec": {
                  "global": {
                    "version": "8.16.1"
                  },
                  "probes": {
                    "actionScript": {
                      "enabled": false
                    },
                    "actionStreamKafka": {
                      "enabled": false
                    },
                    "appDynamics": {
                      "enabled": false
                    },
                    "compellent": {
                      "enabled": false
                    },
                    "datadog": {
                      "enabled": false
                    },
                    "dynatrace": {
                      "enabled": false
                    },
                    "flexera": {
                      "enabled": false
                    },
                    "hds": {
                      "enabled": false
                    },
                    "hpe3par": {
                      "enabled": false
                    },
                    "hyperFlex": {
                      "enabled": false
                    },
                    "hyperV": {
                      "enabled": false
                    },
                    "ibmStorageFlashSystem": {
                      "enabled": false
                    },
                    "instana": {
                      "enabled": false
                    },
                    "jvm": {
                      "enabled": false
                    },
                    "mssql": {
                      "enabled": false
                    },
                    "mysql": {
                      "enabled": false
                    },
                    "netApp": {
                      "enabled": false
                    },
                    "newRelic": {
                      "enabled": false
                    },
                    "nutanix": {
                      "enabled": false
                    },
                    "oneView": {
                      "enabled": false
                    },
                    "oracle": {
                      "enabled": false
                    },
                    "powerVM": {
                      "enabled": false
                    },
                    "pure": {
                      "enabled": false
                    },
                    "scaleIO": {
                      "enabled": false
                    },
                    "serviceNow": {
                      "enabled": false
                    },
                    "snmp": {
                      "enabled": false
                    },
                    "terraform": {
                      "enabled": false
                    },
                    "tomcat": {
                      "enabled": false
                    },
                    "ucs": {
                      "enabled": false
                    },
                    "vCenter": {
                      "enabled": false
                    },
                    "vmax": {
                      "enabled": false
                    },
                    "vmm": {
                      "enabled": false
                    },
                    "vplex": {
                      "enabled": false
                    },
                    "webLogic": {
                      "enabled": false
                    },
                    "webSphere": {
                      "enabled": false
                    },
                    "wmi": {
                      "enabled": false
                    },
                    "xtremIO": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "VersionManager",
                "metadata": {
                  "name": "versionmanager-release"
                },
                "spec": {
                  "url": "http://remote-nginx-tunnel:9080/cluster-manager/clusterConfiguration"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-client-operator@sha256:8db7396c956e7a04aee92e1e90491cd3eb533cebe5f5616ca43ff16d9f3ed6e2
          createdAt: "2025-08-07T07:15:28Z"
          description: Turbonomic Secure Connect for Hybrid environment management
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.4.2-candidate-20250806145822000'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TurbonomicClient is the Schema for the turbonomicclients
              API
            displayName: Turbonomic Client
            kind: TurbonomicClient
            name: turbonomicclients.clients.turbonomic.ibm.com
            version: v1alpha1
          - description: VersionManager is the Schema for the versionmanagers API
            displayName: Version Manager
            kind: VersionManager
            name: versionmanagers.clients.turbonomic.ibm.com
            version: v1alpha1
        description: "### Turbonomic Secure Connect for Hybrid environment management\nTurbonomic
          Workload Automation for Multicloud simultaneously optimizes performance,
          compliance, and cost in real-time. Workloads are precisely resourced, automatically,
          to perform while satisfying business constraints:\n* Continuous placement
          of workload across multiple clouds both on-prem and public clouds providers.\n*
          Continuous scaling for applications and the underlying infrastructure.\n*
          Capability to connect both public cloud and on-prem targets on single saas
          instance\nIt assures application performance by giving workloads the resources
          they need when they need them.\n\n### How do I start?\nTurbonomic documentation
          can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm.\nTo
          setup Turbonomic Secure Connect client, follow the instructions below:\n1.
          Click Install on the landing page\n2. Upon moving to Install Operator page,
          do not modify the default selection for any of the configuration\n3. Select
          the Installed namespace from the drop down or add a new namespace and click
          on create project to create the entered namespace.\n4. Click Install on
          the Install Operator page.\nAt this point operator is installed.\n\nOn the
          Details tab, under Provided APIs: \n1. Click Create instance for the Turbonomic
          Client API. Enable/Disable the probes based on the requirement, and click
          Create.\n2. Click Create instance for the Version Manager API. Do not modify
          the JSON/YAML file, click on Create.\n\n### How does it work?\nTurbonomic
          uses a public APIs already exposed by application and infrastructure instrumentation
          to discover and monitor your environment.\nTurbonomic determines the right
          actions that drive continuous health, including continuous placement and
          continuous scaling for applications and the underlying cluster.\nTurbonomic
          leverages the built-on orchestration provided by the application and infrastructure
          deployment tools and automates the execution of these actions to continiously
          meet the respective service level objective of each application service."
        displayName: Turbonomic Secure Connect Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - turbonomic
        - turbo secure connect
        links:
        - name: T8c Tsc Client Certified
          url: https://t8c-tsc-certified.domain
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        relatedImages:
        - icr.io/cpopen/t8c-client-operator@sha256:8db7396c956e7a04aee92e1e90491cd3eb533cebe5f5616ca43ff16d9f3ed6e2
        version: 1.4.2-candidate-20250806145822000
      entries:
      - name: t8c-tsc-client-certified.v1.4.2-candidate-20250806145822000
        version: 1.4.2-candidate-20250806145822000
      name: candidate
    - currentCSV: t8c-tsc-client-certified.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "TurbonomicClient",
                "metadata": {
                  "name": "turbonomicclient-release"
                },
                "spec": {
                  "global": {
                    "version": "8.16.1"
                  },
                  "probes": {
                    "actionScript": {
                      "enabled": false
                    },
                    "actionStreamKafka": {
                      "enabled": false
                    },
                    "appDynamics": {
                      "enabled": false
                    },
                    "compellent": {
                      "enabled": false
                    },
                    "datadog": {
                      "enabled": false
                    },
                    "dynatrace": {
                      "enabled": false
                    },
                    "flexera": {
                      "enabled": false
                    },
                    "hds": {
                      "enabled": false
                    },
                    "hpe3par": {
                      "enabled": false
                    },
                    "hyperFlex": {
                      "enabled": false
                    },
                    "hyperV": {
                      "enabled": false
                    },
                    "ibmStorageFlashSystem": {
                      "enabled": false
                    },
                    "instana": {
                      "enabled": false
                    },
                    "jvm": {
                      "enabled": false
                    },
                    "mssql": {
                      "enabled": false
                    },
                    "mysql": {
                      "enabled": false
                    },
                    "netApp": {
                      "enabled": false
                    },
                    "newRelic": {
                      "enabled": false
                    },
                    "nutanix": {
                      "enabled": false
                    },
                    "oneView": {
                      "enabled": false
                    },
                    "oracle": {
                      "enabled": false
                    },
                    "powerVM": {
                      "enabled": false
                    },
                    "pure": {
                      "enabled": false
                    },
                    "scaleIO": {
                      "enabled": false
                    },
                    "serviceNow": {
                      "enabled": false
                    },
                    "snmp": {
                      "enabled": false
                    },
                    "terraform": {
                      "enabled": false
                    },
                    "tomcat": {
                      "enabled": false
                    },
                    "ucs": {
                      "enabled": false
                    },
                    "vCenter": {
                      "enabled": false
                    },
                    "vmax": {
                      "enabled": false
                    },
                    "vmm": {
                      "enabled": false
                    },
                    "vplex": {
                      "enabled": false
                    },
                    "webLogic": {
                      "enabled": false
                    },
                    "webSphere": {
                      "enabled": false
                    },
                    "wmi": {
                      "enabled": false
                    },
                    "xtremIO": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "VersionManager",
                "metadata": {
                  "name": "versionmanager-release"
                },
                "spec": {
                  "url": "http://remote-nginx-tunnel:9080/cluster-manager/clusterConfiguration"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-client-operator@sha256:626e024327388410f7a77bd9b8e5a5df29993fcafdb0176cd0bea08b05713c61
          createdAt: "2025-08-12T23:15:08Z"
          description: Turbonomic Secure Connect for Hybrid environment management
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.4.2'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TurbonomicClient is the Schema for the turbonomicclients
              API
            displayName: Turbonomic Client
            kind: TurbonomicClient
            name: turbonomicclients.clients.turbonomic.ibm.com
            version: v1alpha1
          - description: VersionManager is the Schema for the versionmanagers API
            displayName: Version Manager
            kind: VersionManager
            name: versionmanagers.clients.turbonomic.ibm.com
            version: v1alpha1
        description: "### Turbonomic Secure Connect for Hybrid environment management\nTurbonomic
          Workload Automation for Multicloud simultaneously optimizes performance,
          compliance, and cost in real-time. Workloads are precisely resourced, automatically,
          to perform while satisfying business constraints:\n* Continuous placement
          of workload across multiple clouds both on-prem and public clouds providers.\n*
          Continuous scaling for applications and the underlying infrastructure.\n*
          Capability to connect both public cloud and on-prem targets on single saas
          instance\nIt assures application performance by giving workloads the resources
          they need when they need them.\n\n### How do I start?\nTurbonomic documentation
          can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm.\nTo
          setup Turbonomic Secure Connect client, follow the instructions below:\n1.
          Click Install on the landing page\n2. Upon moving to Install Operator page,
          do not modify the default selection for any of the configuration\n3. Select
          the Installed namespace from the drop down or add a new namespace and click
          on create project to create the entered namespace.\n4. Click Install on
          the Install Operator page.\nAt this point operator is installed.\n\nOn the
          Details tab, under Provided APIs: \n1. Click Create instance for the Turbonomic
          Client API. Enable/Disable the probes based on the requirement, and click
          Create.\n2. Click Create instance for the Version Manager API. Do not modify
          the JSON/YAML file, click on Create.\n\n### How does it work?\nTurbonomic
          uses a public APIs already exposed by application and infrastructure instrumentation
          to discover and monitor your environment.\nTurbonomic determines the right
          actions that drive continuous health, including continuous placement and
          continuous scaling for applications and the underlying cluster.\nTurbonomic
          leverages the built-on orchestration provided by the application and infrastructure
          deployment tools and automates the execution of these actions to continiously
          meet the respective service level objective of each application service."
        displayName: Turbonomic Secure Connect Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - turbonomic
        - turbo secure connect
        links:
        - name: T8c Tsc Client Certified
          url: https://t8c-tsc-certified.domain
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        relatedImages:
        - icr.io/cpopen/t8c-client-operator@sha256:626e024327388410f7a77bd9b8e5a5df29993fcafdb0176cd0bea08b05713c61
        version: 1.4.2
      entries:
      - name: t8c-tsc-client-certified.v1.4.2
        version: 1.4.2
      name: stable
    defaultChannel: stable
    packageName: t8c-tsc-client-certified
    provider:
      name: Turbonomic, Inc.
      url: https://www.turbonomic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-recyclebin-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-recyclebin-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "recyclebin.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/recyclebin-controller:1.1.1
          createdAt: "2025-09-20T01:20:03Z"
          description: AWS Recycle Bin controller is a service controller for managing
            Recycle Bin resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Rule represents the state of an AWS recyclebin Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.recyclebin.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Recycle Bin resources in AWS from within your Kubernetes cluster.

          **About Amazon Recycle Bin**

          Recycle Bin is a data recovery feature that enables you to restore accidentally deleted Amazon EBS snapshots and EBS-backed AMIs. When using Recycle Bin, if your resources are deleted, they are retained in the Recycle Bin for a time period that you specify before being permanently deleted.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Recycle Bin
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - recyclebin
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Recycle Bin Developer Resources
          url: https://docs.aws.amazon.com/ebs/latest/userguide/recycle-bin.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: recyclebin maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/recyclebin-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-recyclebin-controller.v1.1.1
        version: 1.1.1
      - name: ack-recyclebin-controller.v1.1.0
        version: 1.1.0
      - name: ack-recyclebin-controller.v1.0.13
        version: 1.0.13
      - name: ack-recyclebin-controller.v1.0.12
        version: 1.0.12
      - name: ack-recyclebin-controller.v1.0.11
        version: 1.0.11
      - name: ack-recyclebin-controller.v1.0.10
        version: 1.0.10
      - name: ack-recyclebin-controller.v1.0.9
        version: 1.0.9
      - name: ack-recyclebin-controller.v1.0.8
        version: 1.0.8
      - name: ack-recyclebin-controller.v1.0.7
        version: 1.0.7
      - name: ack-recyclebin-controller.v1.0.6
        version: 1.0.6
      - name: ack-recyclebin-controller.v1.0.5
        version: 1.0.5
      - name: ack-recyclebin-controller.v1.0.2
        version: 1.0.2
      - name: ack-recyclebin-controller.v1.0.1
        version: 1.0.1
      - name: ack-recyclebin-controller.v1.0.0
        version: 1.0.0
      - name: ack-recyclebin-controller.v0.0.3
        version: 0.0.3
      - name: ack-recyclebin-controller.v0.0.2
        version: 0.0.2
      - name: ack-recyclebin-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-recyclebin-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicast-community-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicast-community-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/apicast-operator:v0.8.0
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway built on top of NGINX. It is part
            of the Red Hat 3scale API Management Platform
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |
          The APIcast Operator creates and maintains the Red Hat 3scale API Gateway in several deployment configurations.

          APIcast is an API gateway built on top of [NGINX](https://www.nginx.com/). It is part of the [Red Hat 3scale API Management Platform](https://www.redhat.com/en/technologies/jboss-middleware/3scale).

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator/blob/v0.8.0/doc/operator-user-guide.md#upgrading-APIcast) for more information.

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/apicast-operator/tree/v0.8.0).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: APIcast
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://github.com/3scale/apicast-operator/tree/v0.8.0
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/apicast-operator:v0.8.0
        version: 0.8.0
      entries:
      - name: apicast-community-operator.v0.8.0
        version: 0.8.0
      - name: apicast-community-operator.v0.7.1
        version: 0.7.1
      - name: apicast-community-operator.v0.6.0
        version: 0.6.0
      - name: apicast-community-operator.v0.5.1
        version: 0.5.1
      name: alpha
    - currentCSV: apicast-community-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/apicast-operator:v0.8.0
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway built on top of NGINX. It is part
            of the Red Hat 3scale API Management Platform
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |
          The APIcast Operator creates and maintains the Red Hat 3scale API Gateway in several deployment configurations.

          APIcast is an API gateway built on top of [NGINX](https://www.nginx.com/). It is part of the [Red Hat 3scale API Management Platform](https://www.redhat.com/en/technologies/jboss-middleware/3scale).

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator/blob/v0.8.0/doc/operator-user-guide.md#upgrading-APIcast) for more information.

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/apicast-operator/tree/v0.8.0).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: APIcast
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://github.com/3scale/apicast-operator/tree/v0.8.0
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/apicast-operator:v0.8.0
        version: 0.8.0
      entries:
      - name: apicast-community-operator.v0.8.0
        version: 0.8.0
      - name: apicast-community-operator.v0.7.1
        version: 0.7.1
      - name: apicast-community-operator.v0.6.0
        version: 0.6.0
      - name: apicast-community-operator.v0.5.1
        version: 0.5.1
      name: stable
    defaultChannel: stable
    packageName: apicast-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-appsec
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-appsec.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Appsec",
                "metadata": {
                  "name": "appsec-sample"
                },
                "spec": {
                  "appsecDataModel": {
                    "port": 8080
                  },
                  "appsecResourceService": {
                    "port": 8080
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "remoteRepository": "docker.io",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "reportTemplates": {
                    "enabled": false
                  },
                  "systemMonitor": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:39:34Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Appsec
            name: appsecs.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP APPSEC Operator
        displayName: Keysight WAP APPSEC Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Appsec
          url: https://keysight-wap-appsec.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-appsec@sha256:e39ff225c26b750f86553693dda47d0dcc9834a4ab0975aea154361ee3527314
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-appsec.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-appsec
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-operator.v23.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerify",
                "metadata": {
                  "name": "ibmsecurityverify-sample"
                },
                "spec": {
                  "clientSecret": "--secret--",
                  "sessionLifetime": 3600,
                  "ssoPath": "/verify-sso"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isv-saas/verify-operator@sha256:c3bed2e5db48e54884d2278d24ca4864f389a56a667db481c627d01ab5ffaba7
          createdAt: "2023-03-22T02:49:11Z"
          description: The IBM Security Verify operator can consistently enforce policy-driven
            security, including authentication and authorization, using the Ingress
            networking capability of OpenShift.
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.9"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM-Security/verify-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerify is the Schema for the ibmsecurityverifies
              API.
            displayName: IBMSecurity Verify
            kind: IBMSecurityVerify
            name: ibmsecurityverifies.ibm.com
            version: v1
        description: "[IBM Security Verify](https://www.ibm.com/products/verify-saas)
          is an Identity-as-a-Service platform that allows IT, security and business
          leaders to protect their digital users, assets and data in a hybrid, multi-cloud
          world by enabling technical agility and operational efficiency. IBM Security
          Verify SaaS provides single sign-on (SSO), multi-factor authentication (MFA),
          AI-powered context for risk-based authentication for adaptive access decisions,
          user management, access recertification campaigns and identity analytics.\n\nAs
          part of using Red Hat OpenShift, you are entitled to Verify SaaS SSO for
          unlimited apps and users.  If you do not already have a Verify SaaS SSO
          tenant a new one can be [created](https://www.ibm.com/account/reg/signup?formid=urx-51255).\n\nFor
          a detailed description of IBM Security Verify refer to the [Official documentation](https://www.ibm.com/docs/en/security-verify).\n\nThe
          IBM Security Verify operator can consistently enforce policy-driven security
          by using the Ingress networking capability of OpenShift in conjunction with
          the [Nginx Ingress operator](https://www.nginx.com/blog/getting-started-nginx-ingress-operator-red-hat-openshift/).
          With this approach, you can enforce authentication and authorization policies
          for all of the applications in your cluster at the same time, without ever
          changing your application code. You can also dynamically register your application
          to start protecting them centrally from the cloud via Verify SaaS. \n\nSee
          the project [Readme](https://github.com/IBM-Security/verify-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Security Verify Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Product Information
          url: https://www.ibm.com/products/verify-saas
        - name: Verify Documentation
          url: https://www.ibm.com/docs/en/security-verify
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Development Team
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - icr.io/isv-saas/verify-operator@sha256:c3bed2e5db48e54884d2278d24ca4864f389a56a667db481c627d01ab5ffaba7
        version: 23.3.0
      entries:
      - name: ibm-security-verify-operator.v23.3.0
        version: 23.3.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hitachi
      provider-url: ""
    name: hspc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hspc-operator.v1.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.hitachi.com/v1",
                "kind": "HSPC",
                "metadata": {
                  "name": "hspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.hitachivantara.com/hitachicsi-oci-oss/hspc-operator@sha256:4dda4a43cefc19c8b1992756b499b1f9300920c5250dd77379f523997cfd5d54
          createdAt: "2025-06-13T10:30:39Z"
          description: An operator for managing Hitachi Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: Hitachi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HSPC is the Schema for the hspcs API
            displayName: HSPC
            kind: HSPC
            name: hspcs.csi.hitachi.com
            version: v1
        description: |-
          ## About
          Hitachi Storage Plug-in for Containers is a software plugin component for Kubernetes and OpenShift environments that is used to create and manage persistent block storage from Hitachi Virtual Storage Platform One storages.


          For full documentation, go to our [Product Documentation site](https://docs.hitachivantara.com/home) and refer to the reference guide for HSPC v3.16.0.

          ## Requirements

          ### Supported Driver Version

          * HSPC v3.16.0
        displayName: Hitachi Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hitachi
        - Storage
        - Container Storage Interface
        - CSI
        links:
        - name: Product Documentation
          url: https://docs.hitachivantara.com/home
        maintainers:
        - email: Container-info@hitachivantara.com
          name: Hitachi
        maturity: stable
        provider:
          name: Hitachi
        relatedImages:
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:81d32545fb1b2c319bfeb36a70db4b4d3aea9c51c0574b209b8f64750b32d2f0
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.hitachivantara.com/hitachicsi-oci-oss/hspc-csi-driver@sha256:2c580d98be55fffd8f6083379b6f23cbfe2aeb42fc4aec9e890072bc831c87c4
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.hitachivantara.com/hitachicsi-oci-oss/hspc-operator@sha256:4dda4a43cefc19c8b1992756b499b1f9300920c5250dd77379f523997cfd5d54
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        version: 1.16.0
      entries:
      - name: hspc-operator.v1.16.0
        version: 1.16.0
      name: stable
    defaultChannel: stable
    packageName: hspc-operator
    provider:
      name: Hitachi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CyberArk
      provider-url: https://www.cyberark.com/
    name: conjur-follower-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: conjur-follower-operator.v2.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:75fb5de2c111084ffcc3d973e6c63f1adfae78e29f6fd02152419fbca553ce7b",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:bee2cd31b3620a31e98f9d21326a2cdcecb36479099b06cc9660e4e83354d940",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:383776c23556a0a327a0bacc843d22c4b1516082ddd0cfb8a8275500f2eb841a",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:a9b109c7f3bfba5553aa4366338fa211896564f8e33856d75b4918acf8429f1e",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:b3096c17e7f3b9d723c906619dde293ea084e78c0da48c534335a8eee07b23c7",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:680a2044bc0e94fbc9376cd1eed9d067de91657f566c9752654480c939447841",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:ed53aea1866b1e2262386ac38cb9e1e47cc3f9047ae47f49070e2f506af6ca5a"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0 <2.5.3'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:75fb5de2c111084ffcc3d973e6c63f1adfae78e29f6fd02152419fbca553ce7b
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:bee2cd31b3620a31e98f9d21326a2cdcecb36479099b06cc9660e4e83354d940
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:a9b109c7f3bfba5553aa4366338fa211896564f8e33856d75b4918acf8429f1e
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:b3096c17e7f3b9d723c906619dde293ea084e78c0da48c534335a8eee07b23c7
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:cf46bdb4238cdb1dc51329822da55b7836958be2344d26f9e1819c0aa97fece6
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:680a2044bc0e94fbc9376cd1eed9d067de91657f566c9752654480c939447841
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:ed53aea1866b1e2262386ac38cb9e1e47cc3f9047ae47f49070e2f506af6ca5a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:383776c23556a0a327a0bacc843d22c4b1516082ddd0cfb8a8275500f2eb841a
        version: 2.5.3
      entries:
      - name: conjur-follower-operator.v2.5.3
        version: 2.5.3
      name: "2.5"
    - currentCSV: conjur-follower-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:c5337633eb3cc272f989e3e70c2e9bace9c819bba9e817537c7a691a8b6ee596
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3
        version: 2.6.3
      entries:
      - name: conjur-follower-operator.v2.6.3
        version: 2.6.3
      - name: conjur-follower-operator.v2.6.2
        version: 2.6.2
      - name: conjur-follower-operator.v2.6.1
        version: 2.6.1
      - name: conjur-follower-operator.v2.6.0
        version: 2.6.0
      - name: conjur-follower-operator.v2.5.0
        version: 2.5.0
      - name: conjur-follower-operator.v2.4.4
        version: 2.4.4
      - name: conjur-follower-operator.v2.4.3
        version: 2.4.3
      - name: conjur-follower-operator.v2.4.2
        version: 2.4.2
      - name: conjur-follower-operator.v2.3.5
        version: 2.3.5
      - name: conjur-follower-operator.v2.3.4
        version: 2.3.4
      - name: conjur-follower-operator.v2.3.1-290
        version: 2.3.1
      - name: conjur-follower-operator.v2.3.0
        version: 2.3.0
      - name: conjur-follower-operator.v2.2.4
        version: 2.2.4
      - name: conjur-follower-operator.v2.2.2
        version: 2.2.2
      - name: conjur-follower-operator.v2.2.1
        version: 2.2.1
      name: preview
    - currentCSV: conjur-follower-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:c5337633eb3cc272f989e3e70c2e9bace9c819bba9e817537c7a691a8b6ee596
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3
        version: 2.6.3
      entries:
      - name: conjur-follower-operator.v2.6.3
        version: 2.6.3
      - name: conjur-follower-operator.v2.6.2
        version: 2.6.2
      - name: conjur-follower-operator.v2.6.1
        version: 2.6.1
      - name: conjur-follower-operator.v2.6.0
        version: 2.6.0
      - name: conjur-follower-operator.v2.5.0
        version: 2.5.0
      - name: conjur-follower-operator.v2.4.4
        version: 2.4.4
      - name: conjur-follower-operator.v2.4.3
        version: 2.4.3
      - name: conjur-follower-operator.v2.4.2
        version: 2.4.2
      - name: conjur-follower-operator.v2.3.5
        version: 2.3.5
      - name: conjur-follower-operator.v2.3.4
        version: 2.3.4
      - name: conjur-follower-operator.v2.3.1-290
        version: 2.3.1
      - name: conjur-follower-operator.v2.3.0
        version: 2.3.0
      - name: conjur-follower-operator.v2.2.4
        version: 2.2.4
      - name: conjur-follower-operator.v2.2.2
        version: 2.2.2
      - name: conjur-follower-operator.v2.2.1
        version: 2.2.1
      name: stable
    defaultChannel: stable
    packageName: conjur-follower-operator
    provider:
      name: CyberArk
      url: https://www.cyberark.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: ovms-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openvino-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "ModelServer",
                "metadata": {
                  "name": "model-server-sample"
                },
                "spec": {
                  "image_name": "registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0",
                  "deployment_parameters": {
                    "replicas": 1
                  },
                  "service_parameters": {
                    "grpc_port": 8080,
                    "rest_port": 8081
                  },
                  "models_settings": {
                    "single_model_mode": true,
                    "config_configmap_name": "",
                    "config_path": "",
                    "model_name": "resnet",
                    "model_path": "gs://<bucket_name>/<model_dir>",
                    "nireq": 0,
                    "plugin_config": "{\"PERFORMANCE_HINT\":\"LATENCY\"}",
                    "batch_size": "",
                    "shape": "",
                    "model_version_policy": "{\"latest\": { \"num_versions\":1 }}",
                    "layout": "",
                    "target_device": "CPU",
                    "is_stateful": false,
                    "idle_sequence_cleanup": false,
                    "low_latency_transformation": true,
                    "max_sequence_number": 0
                  },
                  "server_settings": {
                    "file_system_poll_wait_seconds": 0,
                    "sequence_cleaner_poll_wait_minutes": 0,
                    "log_level": "INFO",
                    "grpc_workers": 1,
                    "rest_workers": 0
                  },
                  "models_repository": {
                    "https_proxy": "",
                    "http_proxy": "",
                    "storage_type": "google",
                    "models_host_path": "",
                    "models_volume_claim": "",
                    "aws_secret_access_key": "",
                    "aws_access_key_id": "",
                    "aws_region": "",
                    "s3_compat_api_endpoint": "",
                    "gcp_creds_secret_name": "",
                    "azure_storage_connection_string": ""
                  }
                }
              },
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "Notebook",
                "metadata": {
                  "name": "v2024.1"
                },
                "spec": {
                  "git_ref": "2024.1",
                  "auto_update_image": false
                }
              }
            ]
          alm-examples-metadata: '{ "model-server-sample":{"description":"Example
            of the model server configuration"}, "v2024":{"description":"Default notebook
            configuration"} }'
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
          createdAt: "2024-05-14T09:29:35Z"
          description: OpenVINO Toolkit Operator manages OpenVINO components in OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/ovms-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/ovms-operator-rhmp/support?utm_source=openshift_console
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.34.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Provides inference serving through gRPC & REST
            displayName: Model Server
            kind: ModelServer
            name: modelservers.intel.com
            version: v1alpha1
          - description: Adds a collection of ready-to-run Jupyter notebooks for learning
              and experimenting with the OpenVINO™ Toolkit to Red Hat® OpenShift®
              Data Science
            displayName: Notebook
            kind: Notebook
            name: notebooks.intel.com
            version: v1alpha1
        description: |
          OpenVINO Toolkit Operator manages OpenVINO components in OpenShift.

          Currently there available components are ModelServer and Notebook.

          # Model Server
          [OpenVINO™ Model Server](https://github.com/openvinotoolkit/model_server) (OVMS) is a scalable, high-performance solution for serving machine learning models optimized for Intel® architectures. The server provides an inference service via gRPC or REST API for any models trained in a framework that is supported by [OpenVINO](https://software.intel.com/en-us/openvino-toolki://docs.openvino.ai/latest/index.html).
          Model Server configuration parameters is explained [here](https://github.com/openvinotoolkit/operator/blob/main/docs/modelserver_params.md).
          ## Using the cluster
          OpenVINO Model Server can be consumed as a `Service`. It is called like with `ModelServer` resource with `-ovms` suffix.
          The suffix is omitted when `ovms` phrase is included in the name.
          The service exposes gRPC and REST API interfaces to run inference requests.
          ```
          oc get pods
          NAME                                        READY   STATUS    RESTARTS   AGE
          model-server-sample-ovms-586f6f76df-dpps4   1/1     Running   0          8h

          oc get services
          NAME                       TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)             AGE
          model-server-sample-ovms   ClusterIP   172.25.199.210   <none>        8080/TCP,8081/TCP   8h
          ```

          # Notebook
          The Notebook resource integrates JupyterHub from OpenShift Data Science or Open Data Hub with a container image that includes [developer
          tools](https://github.com/openvinotoolkit/openvino/blob/master/docs/install_guides/pypi-openvino-dev.md)
          from the OpenVINO toolkit and a set of Jupyter notebook tutorials. It enables selecting a defined image `OpenVINO™ Toolkit` from
          the Jupyter Spawner choice list.

          Create the `Notebook` resource in the same project with JupyterHub and RedHat OpenShift Data Science operator as seen in [documentation](https://github.com/openvinotoolkit/operator/blob/main/docs/notebook_in_rhods.md).
          It builds the image in the cluster based on Dockerfile from [openvino_notebooks](https://github.com/openvinotoolkit/openvino_notebooks).

          ## References
          OpenVINO Model Server on [Github](https://github.com/openvinotoolkit/model_server)

          OpenVINO Model Server Operator on [Github](https://github.com/openvinotoolkit/operator)
        displayName: OpenVINO Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Inference OpenVINO
        links:
        - name: OpenVINO Toolkit Operator
          url: https://github.com/openvinotoolkit/operator
        - name: OpenVINO Model Server
          url: https://github.com/openvinotoolkit/model_server
        - name: OpenVINO Notebooks
          url: https://github.com/openvinotoolkit/openvino_notebooks
        maintainers:
        - email: openvino-operator-support@intel.com
          name: Operator Support
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5714b93ac39277affdc10c3ba142d8c032ed0ddb2ba61c04cbddac710464c4
        - registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
        - registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0
        version: 1.2.0
      entries:
      - name: openvino-operator.v1.2.0
        version: 1.2.0
      - name: openvino-operator.v1.1.0
        version: 1.1.0
      - name: openvino-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: ovms-operator-rhmp
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Aerospike
      provider-url: ""
    name: aerospike-kubernetes-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aerospike-kubernetes-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "memory-size": 3000000000,
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:6.3.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: quay.io/aerospike/aerospike-kubernetes-operator@sha256:a3033b50ae434b97de4de1ca03013dabcfac580d810a6ee8c0fe05b6a9301952
          createdAt: 2023-05-04T09:32:47UTC
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: <=2.4.0
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Aerospike
        relatedImages:
        - quay.io/aerospike/aerospike-kubernetes-operator@sha256:a3033b50ae434b97de4de1ca03013dabcfac580d810a6ee8c0fe05b6a9301952
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        version: 2.5.0
      entries:
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      name: alpha
    - currentCSV: aerospike-kubernetes-operator-rhmp.v4.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "data-size": 1073741824,
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:8.1.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackup",
                "metadata": {
                  "name": "aerospikebackup",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "aerospike-cluster": {
                      "aerospike-aerospikebackup-test-cluster": {
                        "credentials": {
                          "password": "admin123",
                          "user": "admin"
                        },
                        "seed-nodes": [
                          {
                            "host-name": "aerocluster.aerospike.svc.cluster.local",
                            "port": 3000
                          }
                        ]
                      }
                    },
                    "backup-routines": {
                      "aerospike-aerospikebackup-test-routine": {
                        "backup-policy": "test-policy",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "local"
                      },
                      "aerospike-aerospikebackup-test-routine1": {
                        "backup-policy": "test-policy1",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "s3Storage"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackupService",
                "metadata": {
                  "name": "aerospikebackupservice-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "config": {
                    "backup-policies": {
                      "test-policy": {
                        "parallel": 3
                      },
                      "test-policy1": {
                        "parallel": 3
                      }
                    },
                    "service": {
                      "http": {
                        "port": 8081
                      }
                    },
                    "storage": {
                      "local": {
                        "local-storage": {
                          "path": "/tmp/localStorage"
                        }
                      },
                      "s3Storage": {
                        "s3-storage": {
                          "bucket": "aerospike-kubernetes-operator-test",
                          "s3-profile": "default",
                          "s3-region": "us-east-1"
                        }
                      }
                    }
                  },
                  "image": "aerospike/aerospike-backup-service:3.2.0",
                  "secrets": [
                    {
                      "secretName": "aws-secret",
                      "volumeMount": {
                        "mountPath": "/root/.aws/credentials",
                        "name": "aws-secret",
                        "subPath": "credentials"
                      }
                    }
                  ],
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeRestore",
                "metadata": {
                  "name": "aerospikerestore-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "backup-data-path": "aerospike-aerospikebackup-test-routine/backup/1733506015827/data/test",
                    "destination": {
                      "credentials": {
                        "password": "admin123",
                        "user": "admin"
                      },
                      "label": "destinationCluster",
                      "seed-nodes": [
                        {
                          "host-name": "aerocluster.test.svc.cluster.local",
                          "port": 3000
                        }
                      ]
                    },
                    "policy": {
                      "no-generation": true,
                      "no-indexes": true,
                      "parallel": 3
                    },
                    "source": {
                      "local-storage": {
                        "path": "/tmp/localStorage"
                      }
                    }
                  },
                  "type": "Full"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/aerospike/aerospike-kubernetes-operator@sha256:e6d6bd7c58e036b3728e52475a9b764ca8c46b83cbc4751cac66294495bf2b31
          createdAt: "2025-09-15T10:14:05Z"
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
          support: Aerospike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeBackup is the Schema for the aerospikebackup API
            displayName: Aerospike Backup
            kind: AerospikeBackup
            name: aerospikebackups.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeBackupService is the Schema for the aerospikebackupservices
              API
            displayName: Aerospike Backup Service
            kind: AerospikeBackupService
            name: aerospikebackupservices.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1
          - description: AerospikeRestore is the Schema for the aerospikerestores
              API
            displayName: Aerospike Restore
            kind: AerospikeRestore
            name: aerospikerestores.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
           -   Aerospike cluster monitoring
           -   Backup and restore Aerospike clusters
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Aerospike
        relatedImages:
        - quay.io/aerospike/aerospike-kubernetes-operator@sha256:e6d6bd7c58e036b3728e52475a9b764ca8c46b83cbc4751cac66294495bf2b31
        version: 4.1.1
      entries:
      - name: aerospike-kubernetes-operator-rhmp.v4.1.1
        version: 4.1.1
      - name: aerospike-kubernetes-operator-rhmp.v4.1.0
        version: 4.1.0
      - name: aerospike-kubernetes-operator-rhmp.v4.0.2
        version: 4.0.2
      - name: aerospike-kubernetes-operator-rhmp.v4.0.1
        version: 4.0.1
      - name: aerospike-kubernetes-operator-rhmp.v4.0.0
        version: 4.0.0
      - name: aerospike-kubernetes-operator-rhmp.v3.4.1
        version: 3.4.1
      - name: aerospike-kubernetes-operator-rhmp.v3.4.0
        version: 3.4.0
      - name: aerospike-kubernetes-operator-rhmp.v3.3.1
        version: 3.3.1
      - name: aerospike-kubernetes-operator-rhmp.v3.3.0
        version: 3.3.0
      - name: aerospike-kubernetes-operator-rhmp.v3.2.2
        version: 3.2.2
      - name: aerospike-kubernetes-operator-rhmp.v3.2.1
        version: 3.2.1
      - name: aerospike-kubernetes-operator.v3.2.0
        version: 3.2.0
      - name: aerospike-kubernetes-operator.v3.1.0
        version: 3.1.0
      - name: aerospike-kubernetes-operator.v3.0.0
        version: 3.0.0
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      name: stable
    defaultChannel: stable
    packageName: aerospike-kubernetes-operator-rhmp
    provider:
      name: Aerospike
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: samsung
      provider-url: ""
    name: socmmd
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: socmmd.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cmmd.samsung.com/v1",
                "kind": "CMMD",
                "metadata": {
                  "name": "my-app",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "allocateMode": "auto",
                  "enable": true,
                  "payload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "metadata": {
                      "name": "my-app"
                    },
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "name": "my-app"
                        }
                      },
                      "template": {
                        "metadata": {
                          "labels": {
                            "name": "my-app"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "quay.io/my-account/my-app:latest",
                              "name": "my-app-01"
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "cmmd.samsung.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "machineconfiguration.openshift.io/role": "cmmd"
                  },
                  "name": "cmmd-node-label",
                  "namespace": "socmmd"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          console.openshift.io/plugins: '["socmmd-console-plugin"]'
          containerImage: quay.io/samsung_sag/socmmd@sha256:2ef361fb15f0fec99e0a8f1ef83ad4dd14f701d579c900cce084c4a62cf805f8
          createdAt: "2024-08-01T05:45:38Z"
          description: Samsung Operator For CMM-D
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "cmmd.samsung.com/v1",
              "kind": "Label",
              "metadata": {
                "name": "cmmd-node-label",
                "namespace": "socmmd",
                "labels": {}
              },
              "spec": {}
            }
          operatorframework.io/suggested-namespace: socmmd-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenMPDK/SMDK/tree/main/lib/redhat/socmmd
          support: samsung
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CMMD is the Schema for the cmmds API.
            displayName: CMMD
            kind: CMMD
            name: cmmds.cmmd.samsung.com
            version: v1
          - description: Label is the Schema for the labels API.
            displayName: Node Label
            kind: Label
            name: labels.cmmd.samsung.com
            version: v1
        description: Samsung Operator For CMM-D
        displayName: socmmd
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - samsung
        - memory
        - extend
        - cmm-d
        links:
        - name: Samsung Operator For CMM-D
          url: https://github.com/OpenMPDK/SMDK/tree/main/lib/redhat/socmmd
        maintainers:
        - email: junhyeok.im@samsung.com
          name: junhyeok-im
        maturity: stable
        provider:
          name: samsung
        relatedImages:
        - quay.io/samsung_sag/socmmd@sha256:2ef361fb15f0fec99e0a8f1ef83ad4dd14f701d579c900cce084c4a62cf805f8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        version: 0.0.2
      entries:
      - name: socmmd.v0.0.2
        version: 0.0.2
      - name: socmmd.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: socmmd
    provider:
      name: samsung
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/file-integrity-operator
    name: file-integrity-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: file-integrity-operator.v0.1.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <0.1.13'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator
          url: https://file-integrity-operator.domain
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/openshift4/file-integrity-rhel8-operator@sha256:a65f1244636649734763164a5d6155a6f6a7b79b3950ecc03fb3c58188583599
        version: 0.1.13
      entries:
      - name: file-integrity-operator.v0.1.13
        version: 0.1.13
      name: "4.7"
    - currentCSV: file-integrity-operator.v0.1.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <0.1.32'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:c3e34e6198eadb48c007cd949d8be38e26ef7d15d2f9047a8c3c89c03201e4ea
        version: 0.1.32
      entries:
      - name: file-integrity-operator.v0.1.32
        version: 0.1.32
      name: release-0.1
    - currentCSV: file-integrity-operator.v1.3.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          createdAt: "2024-11-21T18:02:08Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.3.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:f7039c6aaaa3e60dc4a3965cc1341c5217837e4b4030eb943771967e0b206578
        version: 1.3.6
      entries:
      - name: file-integrity-operator.v1.3.6
        version: 1.3.6
      - name: file-integrity-operator.v1.3.5
        version: 1.3.5
      - name: file-integrity-operator.v1.3.4
        version: 1.3.4
      - name: file-integrity-operator.v1.3.3
        version: 1.3.3
      name: stable
    - currentCSV: file-integrity-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <1.0.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:dd1dfeccf59214a0de653c3a4c4dbbc8c41640954318d13386cda73891660e4f
        version: 1.0.0
      entries:
      - name: file-integrity-operator.v1.0.0
        version: 1.0.0
      name: v1
    defaultChannel: stable
    packageName: file-integrity-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/file-integrity-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Jason Skrzypek
      provider-url: https://github.com/shpwrck
    name: openshift-node-upgrade-mutex-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-node-upgrade-mutex-operator.v0.2.66
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexRule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutexrule-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutexrule",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutexrule-sample"
                },
                "spec": {
                  "jobName": "example",
                  "jobNamespace": "node-upgrade-mutex-operator",
                  "jobSpec": {
                    "selector": {},
                    "template": {
                      "metadata": {
                        "name": "mutex"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              "20"
                            ],
                            "image": "ubi9/toolbox",
                            "name": "mutex"
                          }
                        ],
                        "restartPolicy": "Never"
                      }
                    }
                  },
                  "type": "kubernetes"
                }
              },
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexRunConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutexrunconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutexrunconfig",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutexrunconfig-sample"
                },
                "spec": {
                  "canary": {
                    "name": "canary-example"
                  },
                  "mutexRules": [
                    {
                      "name": "mutexrule-sample",
                      "namespace": "node-upgrade-mutex-operator"
                    }
                  ],
                  "mutexTargets": [
                    {
                      "name": "mutextarget-sample",
                      "namespace": "node-upgrade-mutex-operator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexTarget",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutextarget-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutextarget",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutextarget-sample"
                },
                "spec": {
                  "machineConfigPools": [
                    "worker"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/jskrzypek/openshift-node-upgrade-mutex-operator
          createdAt: "2025-05-30T17:49:00Z"
          description: |
            This software injects required operations into the worker node upgrade process.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: node-upgrade-mutex-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          provider: Jason Skrzypek
          repository: https://github.com/shpwrck/openshift-upgrade-rollout-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A node mutex rule definition in the form of a kubernetes
              job spec
            displayName: Mutex Rule
            kind: MutexRule
            name: mutexrules.openshift-optional.skrzypek.io
            version: v1alpha1
          - description: Lists of Targets and Rules to apply to them, as well as an
              upgrade facilitating Machine Config Pool
            displayName: Mutex Run Config
            kind: MutexRunConfig
            name: mutexrunconfigs.openshift-optional.skrzypek.io
            version: v1alpha1
          - description: A list of Machine Config Pools to enforce upgrade procedure.
            displayName: Mutex Target
            kind: MutexTarget
            name: mutextargets.openshift-optional.skrzypek.io
            version: v1alpha1
        description: |
          ## Introduction

          This operator injects required operations in the form of a Kubernetes job into the worker node upgrade process using MachineConfigPools. In short, when provided a list of operations (`MutexRules`) and a list of MachineConfigPools (`MutexTargets`) , the OpenShift Node Upgrade Mutex Operator will execute each operation for a given target, enable updates for the given target by modifying it's MachineConfigPool membership, and wait until the node has been upgraded before proceeding onto subsequent operations and targets.

          ## Dependencies

          ### Release Version Label

          In order to confirm the successful upgrade of each target node, the NFD Operator is used to generate a version feature on every worker node. In order to proceed through given targets, each node must have this label
          `'machineconfiguration.openshift.io/release-image-version'` with a key providing the semantic version of this node (e.g. `'4.16.4'`)

          There are several ways to accomplish this, but an example MachineConfig has been provided [here](https://github.com/shpwrck/openshift-node-upgrade-mutex-operator/blob/main/machineconfig.yaml).

          ### Canary Machine Config Pool

          In addition to node labels, a MachineConfigPool that will be used to accomodate targets subject to upgrades must be provided matching the following criteria:
          * selects only worker nodes
          * is not paused

          Applying the example below will create a compatible canary pool:

          ```yaml
          ---
          apiVersion: machineconfiguration.openshift.io/v1
          kind: MachineConfigPool
          metadata:
            name: canary
          spec:
            machineConfigSelector:
              matchLabels:
                machineconfiguration.openshift.io/role: worker
            nodeSelector:
              matchLabels:
                node-role.kubernetes.io/worker: ''
              matchExpressions:
                - key: 'isCanary'
                  operator: Exists
            paused: false
          ```

          ## Usage / Custom Resources Definitions

          #### MutexRule:

          A resource storing a single rule referring to a template for a Kubernetes job to be created.
          This resource **does not** modify existing resources.

          ```yaml
          ---
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexRule
          metadata:
            name: k8s-example
          spec:
            jobName: "example" # Jobs will be run with this prefix name
            jobNamespace: "node-upgrade-mutex-operator" # Jobs will be run in this namespace
            jobSpec: # Accepts full 'batch/v1' job spec
              selector: {}
              template:
                metadata:
                  name: mutex
                spec:
                  containers:
                  - command:
                    - sleep
                    args:
                    - "30"
                    image: ubi9/toolbox
                    name: mutex
                  restartPolicy: Never
            cleanUpJobSpec: # Accepts full 'batch/v1' job spec (Optional)
              selector: {}
              template:
                metadata:
                  name: mutex
                spec:
                  containers:
                  - command:
                    - sleep
                    args:
                    - "30"
                    image: ubi9/toolbox
                    name: mutex
                  restartPolicy: Never
            type: kubernetes
          ```

          #### MutexTarget:

          A resource storing a set of MachineConfigPools to pause and incrementally upgrade.
          This resource **does not** modify existing resources.

          ```yaml
          ---
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexTarget
          metadata:
            name: worker-example
          spec:
            machineConfigPools:
              - worker # Name of existing machineconfigpool
          ```

          #### MutexRunConfig:

          A resource that references existing MutexRules and MutexTargets as well as an existing MachineConfigPool that will allow members to be upgraded. Every rule defined in the `mutexRules` list will be added as a mandatory task to every node that is a member of the each MachineConfigPool defined in the `mutexTargets` list.
          This resource **does** modify existing resources.

          ```yaml
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexRunConfig
          metadata:
            name: example
          spec:
            skipTLSVerify: false # Defaults to false if omitted
            canary:
              name: canary # Name of unpaused machineconfigpool to modify
            mutexRules: # List of existing mutexRules
              - name: k8s-example
                namespace: node-upgrade-mutex-operator
            mutexTargets: # List of existing mutexTargets
              - name: worker-example
                namespace: node-upgrade-mutex-operator
          ```

          ## Procedure

          #### MutexRule

          1. Upon creation/modification, the schema of each K8s rule is validated.
          2. Status is updated to indicate valid or invalid configuration.

          #### MutexTargets

          1. Upon creation/modification, the existence of each target is validated.
          2. Status is updated to indicate valid or invalid configuration.
          3. The nodes assigned to each target are added to the status field `nodes`.

          #### MutexRunConfigs

          1. Upon creation/modification, the existence as well as validation status of each rule, target, and canary pool is confirmed.
          2. If valid, the existence of an upgrade is queried. An upgrade is available if the following is true:
              * The `status.history` of the `cluster` ClusterVersion equals `Completed`
              * The `release-image-version` of any worker node is less than the `spec.desiredUpdate.version` of the `cluster` ClusterVersion.
          3. If an update exists, each target MachineConfigPool has `spec.paused` set to `true`.
          4. If each pool is paused, each node is updated to include a label for each rule with the following format:
              * `mutexrule.openshift-optional.skrzypek.io/{{MutexRule_Name}}.{{MutexRule_Namespace}}`
          5. If labels have been applied, the provided canary pool is modified to exclude nodes with labels provided by step (4).
          6. Assuming steps (2-5) have completed successfully, then the process for running jobs and updating nodes begins. The process per node is as follows:
              * Launch each K8S job sequentially
              * Remove the current node's rule labels
              * Wait until the `release-image-version` of the node matches the `spec.desiredUpdate.version` of the `cluster` ClusterVersion
              * Launch each K8S cleanup job sequentially
          7. If the update succeeds, all machine config pools are unpaused and all match expressions are removed.

          ## Permissions Granted

          In addition to the permissions required to operate the included CRDs, the following access is granted to the operator's service account.

          | Resource | Permissions |
          | :---: | :---: |
          | Jobs | Create & Read |
          | Nodes | Read & Update |
          | ClusterVersion | Read |
          | MachineConfigPools | Read & Update |

          ## Notes

          * This operator was built using the operator-sdk's ansible plugin and therefore has all of the inherited capabilities and restrictions.
          * This operator has undergone minimal testing and should not be run in production environments without acknowledging the risks.
          * This operator is currently alpha and subject to change based on requirements.

          ## Additional Information
          For additional information refer to the [GitHub Repository](https://github.com/shpwrck/openshift-node-upgrade-mutex-operator)
        displayName: OpenShift Node Upgrade Mutex Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - OpenShift Optional
        - Upgrade
        - Mutex
        links:
        - name: Openshift Node Upgrade Mutex Operator
          url: https://github.com/shpwrck/openshift-node-upgrade-mutex-operator
        maintainers:
        - email: jskrzype@redhat.com
          name: Jason Skrzypek
        maturity: alpha
        minKubeVersion: 1.27.0
        provider:
          name: Jason Skrzypek
          url: https://github.com/shpwrck
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - quay.io/jskrzypek/openshift-node-upgrade-mutex-operator@sha256:d4c28e8b256c00429d9999595c2386f1a25de597ee8e6c33871bb7b22b84c4af
        version: 0.2.66
      entries:
      - name: openshift-node-upgrade-mutex-operator.v0.2.66
        version: 0.2.66
      - name: openshift-node-upgrade-mutex-operator.v0.2.65
        version: 0.2.65
      - name: openshift-node-upgrade-mutex-operator.v0.2.60
        version: 0.2.60
      - name: openshift-node-upgrade-mutex-operator.v0.2.59
        version: 0.2.59
      - name: openshift-node-upgrade-mutex-operator.v0.2.53
        version: 0.2.53
      - name: openshift-node-upgrade-mutex-operator.v0.2.38
        version: 0.2.38
      - name: openshift-node-upgrade-mutex-operator.v0.2.34
        version: 0.2.34
      - name: openshift-node-upgrade-mutex-operator.v0.2.15
        version: 0.2.15
      - name: openshift-node-upgrade-mutex-operator.v0.2.14
        version: 0.2.14
      - name: openshift-node-upgrade-mutex-operator.v0.1.22
        version: 0.1.22
      name: alpha
    defaultChannel: alpha
    packageName: openshift-node-upgrade-mutex-operator
    provider:
      name: Jason Skrzypek
      url: https://github.com/shpwrck
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Digital.ai
      provider-url: https://digital.ai
    name: digitalai-deploy-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: digitalai-deploy-operator.v25.3.0-beta.813
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "quorum",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.0-debian-12-r4"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.0-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.5.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.2-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabsunsupported/deploy-operator@sha256:9ccca8ab999d4b6058c979bbb70568a63cf43a257ae65a787b2ff4d9bf470c39 '
          createdAt: "2025-09-02T05:43:37Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabsunsupported/central-configuration@sha256:3a35c2aada71a7f5ab8fe7d5c68441fe58b0b2c778b6ea85ac6f3f7b351f80ad
        - docker.io/xebialabsunsupported/deploy-operator@sha256:9ccca8ab999d4b6058c979bbb70568a63cf43a257ae65a787b2ff4d9bf470c39
        - docker.io/xebialabsunsupported/xl-deploy@sha256:41feb1d24d70a0c78a560b338d05c96c5383d570e3ef1ff37daff0eb8c92b386
        - docker.io/xebialabsunsupported/deploy-task-engine@sha256:15322f2263c2b9b70336177aec170e59497fa320281c2d73605918b30272b8f1
        version: 25.3.0-beta.813
      entries:
      - name: digitalai-deploy-operator.v25.3.0-beta.813
        version: 25.3.0-beta.813
      - name: digitalai-deploy-operator.v25.3.0-beta.711
        version: 25.3.0-beta.711
      - name: digitalai-deploy-operator.v25.3.0-beta.609
        version: 25.3.0-beta.609
      - name: digitalai-deploy-operator.v25.3.0-beta.512
        version: 25.3.0-beta.512
      name: beta
    - currentCSV: digitalai-deploy-operator.v24.1.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "ingressClass": "haproxy-daid",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "tag": "1.26.0-debian-12-r1"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "tag": "1.10.1-debian-12-r5"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "tag": "15.6.0-debian-12-r7"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "tag": "3.12.13-debian-12-r2"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator without SCC creation and with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:2b3413cd41f2b437e6027c8ae2cff12d1113e0ac1be69fc24840d68a05d3a665 '
          createdAt: "2024-10-16T05:39:21Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind:
          DigitalaiDeploy\nmetadata:\n  name: daid-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  securityContextConstraints:\n
          \   enabled: false\n  centralConfiguration:\n    replicaCount: 1\n    podSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  master:\n    replicaCount: 1\n    persistence:\n
          \     storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n      runAsUser:
          null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  worker:\n    replicaCount: 1\n    persistence:\n
          \     storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n      runAsUser:
          null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  route:\n    enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n        runAsUser: null\n        runAsGroup: null\n
          \       fsGroup: null\n      containerSecurityContext:\n        enabled:
          false\n        runAsUser: null\n        runAsGroup: null\n      securityContextConstraints:\n
          \       enabled: false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n
          \   install: true\n    persistence:\n      size: 1Gi\n      storageClass:
          ''\n    replicaCount: 1\n    podSecurityContext:\n      enabled: false\n
          \     runAsUser: null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     enabled: false\n      runAsUser: null\n      runAsGroup: null\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"daid-doc-digitalai-deploy-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xld-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Embedded PostgreSQL for
          DB management\n- Embedded RabbitMQ for message queue management\n\n### Security
          Configuration Details\n\n- `securityContextConstraints.enabled: false` -
          disables creation of SCCs;\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/central-configuration@sha256:83359e49c66e4fab502cc75663b456eb37d81d7cb512f2c48667f4d6a324e0df
        - docker.io/xebialabs/deploy-task-engine@sha256:8a381fd84483ea919fa7f8bee5cf19d9dec0d1d61cce73ec6b7c56fa504c6c93
        - docker.io/xebialabs/xl-deploy@sha256:77fdf6a3909367b060eba2ef698baddb77e4086d4d977f9c617d1e26a4427472
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:771a9a173e033a3ad8b46f5c00a7036eaa88c8d8d1fbd89217325168998113ea
        - docker.io/xebialabs/deploy-operator@sha256:2b3413cd41f2b437e6027c8ae2cff12d1113e0ac1be69fc24840d68a05d3a665
        version: 24.1.10
      entries:
      - name: digitalai-deploy-operator.v24.1.10
        version: 24.1.10
      name: stable
    - currentCSV: digitalai-deploy-operator.v24.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "image": {
                        "tag": "v0.14.7"
                      },
                      "ingressClass": "haproxy-daid",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.26.2-debian-12-r10"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "15.6.0-debian-12-r20"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.12.14-debian-12-r7"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator without SCC creation and with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:7d505a855728c761433ef7d46f0a64ffe2b4192f4819bad86c23582375099ac9 '
          createdAt: "2025-08-25T04:53:23Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind:
          DigitalaiDeploy\nmetadata:\n  name: daid-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  securityContextConstraints:\n
          \   enabled: false\n  centralConfiguration:\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  master:\n    replicaCount:
          1\n    persistence:\n      storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  worker:\n    replicaCount:
          1\n    persistence:\n      storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  route:\n    enabled: false\n
          \   annotations:\n      haproxy.router.openshift.io/cookie_name: SESSION_XLD\n
          \     haproxy.router.openshift.io/disable_cookies: \"false\"\n      haproxy.router.openshift.io/rewrite-target:
          /\n    hostname: '<mandatory-deploy-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n      securityContextConstraints:\n
          \       enabled: false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n
          \   install: true\n    persistence:\n      size: 1Gi\n      storageClass:
          ''\n    replicaCount: 1\n    podSecurityContext:\n      enabled: false\n
          \   containerSecurityContext:\n      enabled: false\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"daid-doc-digitalai-deploy-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xld-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Embedded PostgreSQL for
          DB management\n- Embedded RabbitMQ for message queue management\n\n### Security
          Configuration Details\n\n- `securityContextConstraints.enabled: false` -
          disables creation of SCCs;\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/central-configuration@sha256:581a0bc71ae3652ceb643d81927fcf8cabea92f1d301f5d354076baff32fbcde
        - docker.io/xebialabs/deploy-task-engine@sha256:a97d67ea53e1b954ffb4d9514fd61426dcc0cab08922009eef02f32162026d6a
        - docker.io/xebialabs/deploy-operator@sha256:7d505a855728c761433ef7d46f0a64ffe2b4192f4819bad86c23582375099ac9
        - docker.io/xebialabs/xl-deploy@sha256:c2fe209f198fb2f42dc78a2e1544c026a6891d60682b4e47634234bbf7346540
        version: 24.1.21
      entries:
      - name: digitalai-deploy-operator.v24.1.21
        version: 24.1.21
      - name: digitalai-deploy-operator.v24.1.20
        version: 24.1.20
      - name: digitalai-deploy-operator.v24.1.19
        version: 24.1.19
      - name: digitalai-deploy-operator.v24.1.18
        version: 24.1.18
      - name: digitalai-deploy-operator.v24.1.17
        version: 24.1.17
      - name: digitalai-deploy-operator.v24.1.14
        version: 24.1.14
      - name: digitalai-deploy-operator.v24.1.10
        version: 24.1.10
      name: stable-v24.1
    - currentCSV: digitalai-deploy-operator.v24.3.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "16.3.0-debian-12-r23"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n{{ if eq .Values.queueType \"quorum\" -}}\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n{{- end }}\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.13.7-debian-12-r5"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "queueType": "classic",
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:41ffc26c1d99a808e537f2366d92473ecf76dbffbac699dfa04785c3016f7470 '
          createdAt: "2025-09-26T18:18:41Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/deploy-operator@sha256:41ffc26c1d99a808e537f2366d92473ecf76dbffbac699dfa04785c3016f7470
        - docker.io/xebialabs/deploy-task-engine@sha256:75fe942305bf3bb403eb243b157b88844363eb702f03227a61e76ad4321bf0ee
        - docker.io/xebialabs/central-configuration@sha256:4b83d4e9a523ae3615579907c65790478a3b22315ca5afce594793424d1d33b1
        - docker.io/xebialabs/xl-deploy@sha256:8536efdf530dde45b84b14191ebfdc7ec5ebb1844f84871032fe2d335b13908d
        version: 24.3.11
      entries:
      - name: digitalai-deploy-operator.v24.3.11
        version: 24.3.11
      - name: digitalai-deploy-operator.v24.3.10
        version: 24.3.10
      - name: digitalai-deploy-operator.v24.3.9
        version: 24.3.9
      - name: digitalai-deploy-operator.v24.3.8
        version: 24.3.8
      - name: digitalai-deploy-operator.v24.3.7
        version: 24.3.7
      - name: digitalai-deploy-operator.v24.3.6
        version: 24.3.6
      - name: digitalai-deploy-operator.v24.3.5
        version: 24.3.5
      name: stable-v24.3
    - currentCSV: digitalai-deploy-operator.v25.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "quorum",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.1-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.4.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.3-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:d192ef9008d72c9792e1b7b988bdae8f1bbce45ed0e44afc48e3e268b657b38b '
          createdAt: "2025-09-26T16:30:53Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/deploy-task-engine@sha256:13bb779d6041836acb95c95f1eed75f4b5b2940d75687306183a9f8a2177cab3
        - docker.io/xebialabs/central-configuration@sha256:0c46275ffa3b251ccb76d2623e0e5a4cd1441bf7f94901845bda743f087bc6e4
        - docker.io/xebialabs/xl-deploy@sha256:e4212d20fad19f09cbe53796371bec7fc368cfe63316f1b7bccecf5faf6ea733
        - docker.io/xebialabs/deploy-operator@sha256:d192ef9008d72c9792e1b7b988bdae8f1bbce45ed0e44afc48e3e268b657b38b
        version: 25.1.5
      entries:
      - name: digitalai-deploy-operator.v25.1.5
        version: 25.1.5
      - name: digitalai-deploy-operator.v25.1.4
        version: 25.1.4
      - name: digitalai-deploy-operator.v25.1.3
        version: 25.1.3
      - name: digitalai-deploy-operator.v25.1.2
        version: 25.1.2
      - name: digitalai-deploy-operator.v25.1.1
        version: 25.1.1
      - name: digitalai-deploy-operator.v25.1.0
        version: 25.1.0
      name: stable-v25.1
    defaultChannel: stable-v24.3
    packageName: digitalai-deploy-operator
    provider:
      name: Digital.ai
      url: https://digital.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://access.redhat.com/products/red-hat-fuse
    name: fuse-online
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-online.v7.10.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:ed4faedcb97bb6a45f568586ff19da19da31ecbc304f12e0ae7978381247744a
          createdAt: 2022-01-31 15:55:46 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:ed4faedcb97bb6a45f568586ff19da19da31ecbc304f12e0ae7978381247744a
        - registry.redhat.io/rhel8/postgresql-12@sha256:44215c5b244b190c82d9d371a8ffd39b0c83576bd85c980cc72782a5cf9e6e1b
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:5048cd4fab2f98e6000bc60835c539cc40b70dbe265e927cfe185595347d8054
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:5dd1bc8f6b7e724736f4aeac86409bda3e0f64f62b736688e9572d14411232f1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:fcdd0292ebfc7e9edc36241f63b54a9d924a7605329c934ff2b10373af26dbbf
        - registry.redhat.io/openshift4/ose-prometheus@sha256:7f679d6c03535e300dac366f4bc6dd8d781d96dc47592836713f8bb76be36fa9
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:4f8849362e9ed6ff95d318a04e287b7ba33254a2cf7abdaf9b90cab7cbba6646
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:df2d55dd8092e9e94d034166890117cf5d1cfd5cf52ef0f90cfa2374cc9e2d06
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:3f0277e2c278740307c9b2d966f539ba12317e306198cb6c4c1c52a2ad9f4f40
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:96810c770f75057b8bba6b78a5ba414c8289315d8af1a387f681ae5e4fb7d6dd
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:ef229a4c8851ec495f28d0eeab83b137cd02698e283f76ea1b32271b7222621b
        version: 7.10.4
      entries:
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.10.x
    - currentCSV: fuse-online.v7.11.1-0.1687357787.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:658af3fc7819efb7cc39cd3a0504dcddc2ce7788c9b1e5178df869f7c9a742a8
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.11.1-0.1684718286.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:a2b1cd0256f30aff8fa37c332419b4483c0bfe34fb4f574ad8f18b0858969342
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:4de91e4527022e1b25df8d3835ddf778f773c20334314be9ba01a885c82287f5
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:8b9bd777bcbd20b1a1bbbc18274a5076a67bea1f928d772cb243c57fb1da7408
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:ac514973ec428d815d7a248d09d27f2a84f762aedfea63b3b48de9e633267ed5
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:658af3fc7819efb7cc39cd3a0504dcddc2ce7788c9b1e5178df869f7c9a742a8
        - registry.redhat.io/rhel8/postgresql-12@sha256:9248c4eaa8aeedacc1c06d7e3141ca1457147eef59e329273eb78e32fcd27e79
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:251c29e81f8e338f980b475b841647a977807908c388f4c86e18da7bff1b198a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:400a6ee03c3200c19b9b15c5d037a849633156a62e13922fb2e62838024ae69c
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:3d98d775bec85704c9a10c2f0d8e4f8116b7a41494f73c0b763e010262e5cd4d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        version: 7.11.1+0.1687357787.p
      entries:
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.11.x
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.12.x
    - currentCSV: fuse-online.v7.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta2",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:c70c5e76dc3db559c722bea2a767eb7dcba1eda786e1ef3a45c4cf32e1aca9ee
          createdAt: 2021-07-21 12:25:15 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta2
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to the product documentation at https://access.redhat.com/documentation/en-us/red_hat_fuse
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:d6df09b15599e776b49e1d46aa01728fd02d5629cee65e77dc277c0166554e0a
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:c668fe1f5d25baddf627acd79df4c8d2ab8b05309d3d41784ddc0a10dd5f34ad
        - registry.redhat.io/openshift4/ose-prometheus@sha256:13e9ea359cbcd3e0e94a6589b86b708dc84197f152701fd4092fffa234d11b58
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:60d5cf00411b03af3f1149a2bb3aeeb196c063d19625f876a094ae6baf46b54b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:f65600506db985689e5b7cd8b838e646d4268a0c4e272e19d7db3577a8dd7ac4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:75d5772c792a7c795281acec3d59e90021c109e9d7d1110868be0b882768aefb
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:c70c5e76dc3db559c722bea2a767eb7dcba1eda786e1ef3a45c4cf32e1aca9ee
        - registry.redhat.io/rhscl/postgresql-12-rhel7@sha256:c3e95e808eecba8ac57632452d3553509131ca783f62c68754a582cf5c55efab
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:eff9f8edf28ccbadcacab2022fff9e6bc93926af97521b49e3968eaf7b6fda65
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:d6f0c804c179fa559d3baed2e2d106c916a0439dc2215c4d34b878bc2e9bc281
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bafd85b0fd1256b8a60a2bb368a7e68f5dbad72718e5059a65660e01a425d8e2
        version: 7.9.3
      entries:
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.9.x
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: candidate
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: latest
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: stable
    defaultChannel: latest
    packageName: fuse-online
    provider:
      name: Red Hat
      url: https://access.redhat.com/products/red-hat-fuse
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: resource-locker-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: resource-locker-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "ResourceLocker",
                "metadata": {
                  "name": "combined"
                },
                "spec": {
                  "patches": [
                    {
                      "id": "patch1",
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 0).metadata.name }}: {{ (index . 1).metadata.name }}\n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "kind": "Namespace",
                          "name": "resource-locker-test"
                        },
                        {
                          "apiVersion": "v1",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "resource-locker-test"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "resource-locker-test"
                      }
                    },
                    {
                      "id": "patch2",
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 0) }}: {{ (index . 1) }}\n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "fieldPath": "$.metadata.uid",
                          "kind": "Namespace",
                          "name": "resource-locker-test"
                        },
                        {
                          "apiVersion": "v1",
                          "fieldPath": "$.metadata.uid",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "resource-locker-test"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "resource-locker-test"
                      }
                    }
                  ],
                  "resources": [
                    {
                      "object": {
                        "apiVersion": "networking.k8s.io/v1",
                        "kind": "NetworkPolicy",
                        "metadata": {
                          "name": "allow-from-same-namespace",
                          "namespace": "resource-locker-test"
                        },
                        "spec": {
                          "ingress": [
                            {
                              "from": [
                                {
                                  "podSelector": {}
                                }
                              ]
                            }
                          ],
                          "podSelector": {}
                        }
                      }
                    },
                    {
                      "object": {
                        "apiVersion": "networking.k8s.io/v1",
                        "kind": "NetworkPolicy",
                        "metadata": {
                          "name": "allow-from-default-namespace",
                          "namespace": "resource-locker-test"
                        },
                        "spec": {
                          "ingress": [
                            {
                              "from": [
                                {
                                  "namespaceSelector": {
                                    "matchLabels": {
                                      "name": "default"
                                    }
                                  }
                                }
                              ]
                            }
                          ],
                          "podSelector": {}
                        }
                      }
                    }
                  ],
                  "serviceAccountRef": {
                    "name": "default"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/resource-locker-operator@sha256:2176db02f561c24012c5179495c89c3819500dff21e89a62052d465eb3314b44
          createdAt: "2021-11-30T19:38:40Z"
          description: This operator provides a facility to lock resources and / or
            patches into the state described by its custom resource preventing any
            drifts.
          operatorframework.io/suggested-namespace: resource-locker-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/resource-locker-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ResourceLocker Represents the intetion to create and enforce
              a set of resources and/or patches.
            displayName: Resource Locker
            kind: ResourceLocker
            name: resourcelockers.redhatcop.redhat.io
            version: v1alpha1
        description: |
          The resource locker operator allows you to specify a set of configurations that the operator will "keep in place" (lock) preventing any drifts.
          Two types of configurations may be specified:
          * Resources. This will instruct the operator to create and enforce the specified resource. In this case the operator "owns" the created resources.
          * Patches to resources. This will instruct the operator to patch- and enforce the change on- a pre-existing resource. In this case the operator does not "own" the resource.
          Locked resources are defined with the `ResourceLocker` CRD. Here is the high-level structure of this CRD:
          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: ResourceLocker
          metadata:
            name: test-simple-resource
          spec:
            resources:
            - object:
                apiVersion: v1
            ...
            patches:
            - targetObjectRef:
              ...
              patchTemplate: |
                metadata:
                  annotations:
                    ciao: hello
            ...
            serviceAccountRef:
              name: default
          ```
          It contains:
          * `resources`: representing an array of resources
          * `patches`: representing an array of patches
          * `serviceAccountRef`: a reference to a service account defined in the same namespace as the ResourceLocker CR, that will be used to create the resources and apply the patches. If not specified the service account will be defaulted to: `default`
          For each ResourceLocker a manager is dynamically allocated. For each resource and patch a controller with the needed watches is created and associated with the previously created manager.
          **Note**: creating a `ResourceLocker` CR will not work without due preparation (of the service account in particular). Follow these [instructions](https://github.com/redhat-cop/resource-locker-operator/blob/master/EXAMPLES.md) for a step by step walkthrough of the needed prerequisites to make this operator work.
        displayName: Resource Locker Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - resource_management
        - enforcing
        - patches
        - gitops
        links:
        - name: Resource Locker Operator
          url: https://github.com/redhat-cop/resource-locker-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/resource-locker-operator:latest
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/resource-locker-operator@sha256:2176db02f561c24012c5179495c89c3819500dff21e89a62052d465eb3314b44
        version: 1.1.3
      entries:
      - name: resource-locker-operator.v1.1.3
        version: 1.1.3
      - name: resource-locker-operator.v1.1.2
        version: 1.1.2
      - name: resource-locker-operator.v1.1.1
        version: 1.1.1
      - name: resource-locker-operator.v1.1.0
        version: 1.1.0
      - name: resource-locker-operator.v1.0.9
        version: 1.0.9
      name: alpha
    defaultChannel: alpha
    packageName: resource-locker-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: APIMatic.io
      provider-url: https://www.apimatic.io/
    name: apimatic-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apimatic-kubernetes-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicodegen.apimatic.io/v1beta1",
                "kind": "APIMatic",
                "metadata": {
                  "name": "apimatic-sample"
                },
                "spec": {
                  "licensespec": {
                    "licenseSourceName": "apimaticlicense",
                    "licenseSourceType": "ConfigMap"
                  },
                  "podspec": {
                    "apimaticContainerSpec": {
                      "image": "apimaticio/apimatic-codegen",
                      "imagePullSecret": "apimaticimagesecret"
                    }
                  },
                  "replicas": 3,
                  "servicespec": {
                    "apimaticserviceport": {
                      "nodePort": 32000,
                      "port": 8070
                    },
                    "servicetype": "NodePort"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: docker.io/apimaticio/apimatic-kubernetes-operator:v0.0.1
          createdAt: "2021-08-24 12:00:00"
          description: Generate client SDKs and interactive Documentation for your
            APIs in minutes
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/apimatic/apimatic-kubernetes-operator
          support: APIMatic.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIMatic is the Schema for the apimatics API
            displayName: APIMatic App
            kind: APIMatic
            name: apimatics.apicodegen.apimatic.io
            version: v1beta1
        description: "##About APIMatic \n\n APIMatic is a Developer Experience   platform
          for Web APIs that empowers users to create multi-language SDKs with API
          reference docs, dynamic code samples, assisted authentication and more -
          all auto-generated to provide a comprehensive developer experience beyond
          any API documentation tool.\n\n## Table of contents\n\n* [Introduction](#introduction)\n*
          [Features](#features)\n* [Running the Sample](#running-the-sample)\n  *
          [Prerequisites](#prerequisites)\n  * [Steps for Direct Deployment](#steps-for-direct-deployment)\n
          \ * [Steps for OLM Deployment](#steps-for-olm-deployment)\n* [Technical
          Support](#technical-support)\n* [Copyrights](#copyrights)\n\n## Introduction\n\nAPIMatic
          Operator simplifies the configuration and lifecycle management of the APIMatic
          code and docs generation solution on different Kubernetes distributions
          and OpenShift. The Operator encapsulates key operational knowledge on how
          to configure and upgrade the APIMatic CodeGen application, making it easy
          to get up and running.\n\n\nMore information about the underlying APIMatic
          CodeGen API that is exposed\nby this operator can be found [here](https://apimatic-core-v3-docs.netlify.app/#/http/getting-started/overview-apimatic-core).\n\n##
          Features\n\nAPIMatic Operator provides the following features:\n- Deploys
          the APIMatic CodeGen Web API service within the Kubernetes or OpenShift
          cluster.\n- Exposes the APIMatic CodeGen API external to the cluster, using
          Service type as [NodePort](https://kubernetes.io/docs/concepts/services-networking/service/#nodeport),
          [LoadBalancer](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          or [ExternalName](https://kubernetes.io/docs/concepts/services-networking/service/#externalname).\n-
          For exposing the service through an ingress resource, create an Ingress
          resource in the namespace of your APIMatic CR and set owned APIMatic service
          created by the operator as a backed service. More information can be found
          [here](https://kubernetes.io/docs/concepts/services-networking/ingress/).\n-
          Manual horizontal scaling of pods.\n  ```sh\n  kubectl scale apm apimatic-sample--replicas=2\n
          \ ```\n\n## Running the Sample \n\n### Prerequisites\n\nPlease contact APIMatic
          at [support@apimatic.io](mailto:support@apimatic.io) to register with the
          APIMatic CodeGen Docker Hub registry and acquire a valid license to run
          the APIMatic CodeGen API.\n\nFurther prerequisites for running the sample
          include:\n\n- [go](https://golang.org/) v1.16.*\n- [git](https://git-scm.com/)\n-
          [make](https://www.gnu.org/software/make/)\n- [Operator SDK](https://sdk.operatorframework.io/docs/overview/)\n-
          A running Kubernetes cluster with [kubectl](https://kubernetes.io/docs/reference/kubectl/overview/)
          on client. For testing purposes, you can use [Minikube](https://minikube.sigs.k8s.io/docs/)
          or [kind](https://kind.sigs.k8s.io/)\n- For checking the service created
          by the APIMatic operator on-prem, you can use [MetalLB](https://metallb.org/)\n\n###
          Steps for Direct Deployment\n\nTo run the sample for checking the APIMatic
          operator:\n\n- Clone the APIMatic repository into your working directory
          using the following command:\n  ```sh\n  git clone https://github.com/apimatic/apimatic-kubernetes-operator.git\n
          \ ```\n- Run `make deploy` to set up the APIMatic operator resources. This
          will deploy the '`apimatic-system`' namespace as well as the CRD and the
          RBAC manifests.\n\n- Create a secret named '`apimaticimagesecret`' to allow
          pulling the APIMatic CodeGen image using the Docker Hub username registered
          with APIMatic. If you have not done so, please contact support@apimatic.io
          for the steps required.\n  ```sh\n  kubectl create secret docker-registry
          apimaticimagesecret --docker-server=https://index.docker.io/v1/ --docker-username=<your-name>
          --docker-password=<your-pword> --docker-email=<your-email>\n  ```\n- Create
          a [configmap](https://kubernetes.io/docs/concepts/configuration/configmap/)
          resource named '`apimaticlicense`' that will serve as the volume storing
          the APIMatic license. An example of this is given below which will create
          the ConfigMap using the *License.lic* file located at */usr/local/apimatic/license/*\n
          \ ```sh\n  kubectl create configmap apimaticlicense --from-file /usr/local/apimatic/license/License.lic\n
          \ ```\n- This will deploy a ConfigMap resource with the following definition:\n
          \ ```sh\n  apiVersion: v1\n  data:\n    License.lic: \\\"<License file contents
          here>\\\"\n  kind: ConfigMap\n  metadata:\n    name: apimaticlicense\n    namespace:
          default\n  ```\n- Now run the sample using the following command:\n  ```sh\n
          \ kubectl apply -f config/samples/apicodegen_v1beta1_apimatic.yaml\n  ```\n-
          You will now see a new [Deployment](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/)
          with replica count of 3 and [Service](https://kubernetes.io/docs/concepts/services-networking/service/)
          of type `NodePort` created, both named ***apimatic-sample***. Accessing
          http://localhost:32000 from your browser (or using curl from within the
          Minikube or Kind cluster, if using that) should now show the APIMatic Titan
          page.\n- You can now use the exposed APIMatic CodeGen API to generate API
          SDKs and Docs using [curl](https://curl.se/), [Postman](https://www.postman.com/)
          or your own custom Web application that consumes the APIMatic CodeGen API
          service.\n\n- Once done, you can remove the APIMatic resources using the
          following command:\n  ```sh\n  make undeploy\n  ```\n\n### Steps for OLM
          Deployment\n\nThe following steps can be used to utilize [Operator LifeCycle
          Manager (OLM)](https://olm.operatorframework.io/docs/) to deploy the operator
          and run the sample. The steps are as follows:\n\n- If not already done so,
          clone the APIMatic repository into your working directory:\n  ```sh\n  git
          clone https://github.com/apimatic/apimatic-kubernetes-operator.git\n  ```
          \n- [Install OLM in your Kubernetes cluster](https://olm.operatorframework.io/docs/getting-started/#installing-olm-in-your-cluster).\n\n-
          Run the following script to install the resources required by OLM to deploy
          the APIMatic operator in the Kubernetes cluster within the `apimatic-system`
          namespace. Information about the different resources required can be found
          using the steps given [here](https://olm.operatorframework.io/docs/tasks/).\n
          \ ```sh\n  kubectl apply -f olm/manifests.yaml\n  ```\n\n- This should spin
          up the `ClusterServiceVersion` of the operator in the `apimatic-system`
          namespace, following which the operator pod will spin up. To ensure the
          that the operator installed successfully, check for the `ClusterServiceVersion`
          and the operator deployment in the namespace it was installed in.\n  ```sh\n
          \ kubectl get csv -n apimatic-system\n  \n  kubectl get deployment -n apimatic-system\n
          \ ```\n\n  - Create a secret named `apimaticimagesecret` to allow pulling
          the APIMatic CodeGen image using the Docker Hub username registered with
          APIMatic. If you have not done so, please contact support@apimatic.io for
          the steps required.\n  ```sh\n  kubectl create secret docker-registry apimaticimagesecret
          --docker-server=https://index.docker.io/v1/ --docker-username=<your-name>
          --docker-password=<your-pword> --docker-email=<your-email>\n  ```\n- Create
          a [configmap](https://kubernetes.io/docs/concepts/configuration/configmap/)
          resource named `apimaticlicense` that will serve as the volume storing the
          APIMatic license. An example of this is given below which will create the
          ConfigMap using the *License.lic* file located at */usr/local/apimatic/license/*\n
          \ ```sh\n  kubectl create configmap apimaticlicense --from-file /usr/local/apimatic/license/License.lic\n
          \ ```\n- This will deploy a ConfigMap resource with the following definition:\n
          \ ```sh\n  apiVersion: v1\n  data:\n    License.lic: \\\"<License file contents
          here>\\\"\n  kind: ConfigMap\n  metadata:\n    name: apimaticlicense\n    namespace:
          default\n  ```\n- Now run the sample using the following command:\n  ```sh\n
          \ kubectl apply -f config/samples/apicodegen_v1beta1_apimatic.yaml\n  ```\n\n-
          You will now see a new [Deployment](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/)
          with replica count of 3 and [Service](https://kubernetes.io/docs/concepts/services-networking/service/)
          of type `NodePort` created, both named ***apimatic-sample***. Accessing
          http://localhost:32000 from your browser (or using curl from within the
          Minikube or Kind cluster, if using that) should now show the APIMatic Titan
          page.\n- You can now use the exposed APIMatic CodeGen API to generate API
          SDKs and Docs using [curl](https://curl.se/), [Postman](https://www.postman.com/)
          or your own custom Web application that consumes the APIMatic CodeGen API
          service.\n- Once done, you can remove the APIMatic operator resources using
          the follow script:\n  ```sh\n  kubectl delete -f olm/manifests.yaml\n  ```\n\n##
          Technical Support\n\n- To request additional features in the future, or
          if you notice any discrepancy regarding this document, please drop an email
          to [support@apimatic.io](mailto:support@apimatic.io).\n\n### Copyrights\n\n&copy;
          2021 APIMatic.io\n"
        displayName: APIMatic Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Rest APIs
        - API SDK Generator
        - API Docs Generator
        - API Developer Portal
        - APIMatic
        - Developer Tools
        links:
        - name: Overview
          url: https://www.apimatic.io/
        - name: Developer Experience Portal
          url: https://www.apimatic.io/developer-experience-portal
        - name: Code-Gen as a Service
          url: https://www.apimatic.io/continuous-code-generation
        - name: API Transformer
          url: https://www.apimatic.io/transformer
        - name: Contact
          url: https://www.apimatic.io/contact
        - name: Documentation
          url: https://docs.apimatic.io/
        - name: Blogs
          url: https://www.apimatic.io/blog/
        maintainers:
        - email: support@apimatic.io
          name: Support
        maturity: alpha
        provider:
          name: APIMatic.io
          url: https://www.apimatic.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - docker.io/apimaticio/apimatic-kubernetes-operator:v0.0.1
        version: 0.0.1
      entries:
      - name: apimatic-kubernetes-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: apimatic-kubernetes-operator
    provider:
      name: APIMatic.io
      url: https://www.apimatic.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-openshift-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.14
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-06-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:154f85cc6721a70909dee755ddd88897ad51038afe2538b6dee4d26e7d8b9afb
        version: 0.0.14
      entries:
      - name: joget-dx8-openshift-operator.v0.0.14
        version: 0.0.14
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-openshift-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sumo Logic
      provider-url: https://www.sumologic.com/
    name: sumologic-kubernetes-collection-helm-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "helm-operator.sumologic.com/v1alpha1",
                    "kind": "SumologicCollection",
                    "metadata": {
                        "name": "collection"
                    },
                    "spec": {
                        "sumologic": {
                            "accessId": "",
                            "accessKey": "",
                            "clusterName": "kubernetes",
                            "scc": {
                                "create": true
                            }
                        },
                        "kube-prometheus-stack": {
                            "prometheus-node-exporter": {
                                "service": {
                                    "port": 9200,
                                    "targetPort": 9200
                                }
                            }
                        },
                        "otellogs": {
                            "daemonset": {
                                "containers": {
                                    "otelcol": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                },
                                "initContainers": {
                                    "changeowner": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                }
                            }
                        }
                    }
                }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing,Monitoring
          containerImage: registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
          createdAt: "2024-07-19T10:48:00Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/sumologic-kubernetes-collection-helm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/sumologic-kubernetes-collection-helm-operator-rhmp/support?utm_source=openshift_console
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator.git
          support: Sumo Logic
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configuration for Sumo Logic Kubernetes Collection
            displayName: SumologicCollection
            kind: SumologicCollection
            name: sumologiccollections.helm-operator.sumologic.com
            version: v1alpha1
        description: Helm Operator for the Sumo Logic Kubernetes Collection Helm Chart
          v4.13.0-2
        displayName: Sumo Logic Kubernetes Collection Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - collection
        - logging
        - metrics
        - monitoring
        links:
        - name: Sumo Logic Kubernetes Collection Helm Operator
          url: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator
        maintainers:
        - email: collection@sumologic.com
          name: Sumo Logic
        maturity: alpha
        provider:
          name: Sumo Logic
          url: https://www.sumologic.com/
        relatedImages:
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
        - registry.connect.redhat.com/sumologic/kube-state-metrics@sha256:3551ededa327a35db798ec8286cd843580e3ae395d2231e041e8c4b7621127ee
        - registry.connect.redhat.com/sumologic/metrics-server@sha256:65d2a3ddc4b4183f438fa18b655079c93c905cb9e3b8c54977336205dc1fff07
        - registry.connect.redhat.com/sumologic/prometheus@sha256:0ef6eaf570a40499dd5de92a79593e950bc771b014e0ee6ff7419b4b82a36fcc
        - registry.connect.redhat.com/sumologic/autoinstrumentation-python@sha256:f0fa45fa45e64c7df88f21181ae18ee96fe83cd3988258663d0806ea6f39e6bb
        - registry.connect.redhat.com/sumologic/kubernetes-setup@sha256:f07eaf68fa11ffc708e4485a83bd3763fc5712f7d73d29edfe5b9c9e08f05df6
        - registry.connect.redhat.com/sumologic/sumologic-otel-collector@sha256:4554d43dba204e3ec8e91b36442a438f10234fa0cb4db7fb7880b257a6485112
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:e1f49fcf5b7e86d9eb49c16dbbf9484cb486f55f999fd51d7c738c3b1f2666cf
        - registry.connect.redhat.com/sumologic/autoinstrumentation-dotnet@sha256:f4f385c4489c74b5c43f2fe6b6620a9eb8d9ecab470567236b3c45f636106640
        - registry.connect.redhat.com/sumologic/tailing-sidecar@sha256:738d653a78dc415243ab3627db9135768175fb7f51d0803566f5d9e824c8744b
        - registry.connect.redhat.com/sumologic/telegraf-operator@sha256:ff958222be01c0c3966474463e43bbd5bcf348790dd34bb761718cb1cea2d793
        - registry.connect.redhat.com/sumologic/autoinstrumentation-nodejs@sha256:c206eef4fdecaa58b4f01e5e7f0dbd74e9606eab2b4debf47461d36191e248d1
        - registry.connect.redhat.com/sumologic/node-exporter@sha256:11e9e604432a66d9b388cb1bf70eac077c7f0372d1264342ef927d76b8ef0272
        - registry.connect.redhat.com/sumologic/kubernetes-tools-kubectl@sha256:d3155b864cc74f327adedf7d013a8f0fcb11c2e73a5655b7cafa848824e80172
        - registry.connect.redhat.com/sumologic/thanos@sha256:16faf7191668e86d8b576bba055dd0cbdeb009635976b9e07bd00f926032f11c
        - registry.connect.redhat.com/sumologic/opentelemetry-operator@sha256:5587fe5b7ebfa57ad320ad2521103f908317388d1741c40e5f7be5ed9e5f9ad7
        - registry.connect.redhat.com/sumologic/nginx-unprivileged@sha256:7260555d16c94c4fe1e016626b4cb8a5c0de6b700e2eea28744331fb298f71ae
        - registry.connect.redhat.com/sumologic/prometheus-operator@sha256:58a329e6582456b32abebb470086191f8366b927af4703741613993465f116b8
        - registry.connect.redhat.com/sumologic/telegraf@sha256:75efd2e509c6017ea0fc5c9103ce068cbc89087b16506a4ef48cb5ff00eab84e
        - registry.connect.redhat.com/sumologic/tailing-sidecar-operator@sha256:c16b8b18e9b263fc3eb24059bba512bcfc3201b32105af6e5518392826ddd66c
        - registry.connect.redhat.com/sumologic/prometheus-config-reloader@sha256:0b605d36782afd8f50fd631d0e9c7b19d39ce962af89b7a737b2f09a8ceff291
        - registry.connect.redhat.com/sumologic/autoinstrumentation-java@sha256:242958b7d8905f4ac02f6ad38897ba58b9badf2a6213a389e67dcc71fee3630c
        - registry.connect.redhat.com/sumologic/busybox@sha256:209bc8e0b1d958fc699f3bb75b1248544e372ad35cdc46c991bd1698f641e1f9
        version: 4.13.0-2
      entries:
      - name: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
        version: 4.13.0-2
      name: alpha
    defaultChannel: alpha
    packageName: sumologic-kubernetes-collection-helm-operator-rhmp
    provider:
      name: Sumo Logic
      url: https://www.sumologic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-bamoe-businessautomation: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: bamoe-businessautomation-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bamoe-businessautomation-operator.8.0.8-1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.kiegroup.org/v2","kind":"KieApp","metadata":{"name":"rhpam-trial"},"spec":{"environment":"rhpam-trial"}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/ibm-bamoe/bamoe-rhel8-operator@sha256:0646da8707d10b3cc2e2344269abd8797db34127a3326bf76bbc1547aa4bbd73
          createdAt: "2025-07-22 17:40:17"
          description: Deploys and manages IBM Business Automation Manager Open Editions
            environment.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM Business Automation Manager
            Open Edition"]'
          repository: https://github.com/kiegroup/kie-cloud-operator
          support: https://www.ibm.com/mysupport
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A project prescription running an IBM BAMOE environment.
            displayName: KieApp
            kind: KieApp
            name: kieapps.app.kiegroup.org
            version: v2
        description: |-
          Deploys and manages IBM Business Automation Manager Open Editions environment.

          * **IBM Process Automation Manager** is a platform for developing containerized microservices and applications that automate business decisions and processes. It includes business process management (BPM), business rules management (BRM), and business resource optimization and complex event processing (CEP) technologies. It also includes a user experience platform to create engaging user interfaces for process and decision services with minimal coding.
        displayName: IBM Business Automation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kieapp
        - pam
        - decision
        - kie
        - cloud
        - bpm
        - process
        - automation
        - operator
        links:
        - name: Product Page
          url: https://ibm.com
        - name: Documentation
          url: https://www.ibm.com/docs/en/ibamoe
        maintainers:
        - email: mysphelp@us.ibm.com
          name: IBM
        maturity: 8.x-stable
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-monitoring-rhel8@sha256:ef1d77bb085172f927ee53d1674c85e2acaf1f580d7cad55bae8bf2c33c49c01
        - registry.redhat.io/openshift4/ose-cli@sha256:b2b93def4eb5edeccfa2d8ba22e493646325842674dff0446b45677bf641a342
        - registry.redhat.io/rhscl/postgresql-13-rhel7@sha256:338dd09681dcf860e6a976a3bc48a1474c2376d698815e4eb611ae9c5aa68966
        - registry.redhat.io/ibm-bamoe/bamoe-process-migration-rhel8@sha256:3fd45f2dabca7220772d823275a5697b42dd35f53c27671b093117933ef2d4f6
        - registry.redhat.io/rhscl/mysql-80-rhel7@sha256:e4f535c78bc2fdb58914c52fb805185035ff31515452b4df714f9ab623ba383a
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-rhel8@sha256:c3553707936cb9eee8a6a622643f72c91e1b0c11add5ec9ad5865242a9e133b4
        - registry.redhat.io/ibm-bamoe/bamoe-kieserver-rhel8@sha256:01bf97beffb3647968aae052cdd401d6d562c35f3e78ac483dbf55c44a71dc2f
        - registry.redhat.io/jboss-datagrid-7/datagrid73-openshift@sha256:4056b1cf8ff5995f56eecfdc087d9ece56a216dba6bd73acc5693f0778e8d573
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-rhel8@sha256:82d8ebde7eb3ecfbf72e5b71e7e392d3a3597fc084b6cf35e8fec23bfe41210d
        - registry.redhat.io/amq7/amq-broker@sha256:792c1cb19e62519cc47d834650fe5a6a6ca589a37d0b53d4f4549e80cabae3f9
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:98e9a6618558bc8a1b9cb91d1c9b3380968d335092f6e953d1abe6af96abf65c
        - registry.redhat.io/ibm-bamoe/bamoe-smartrouter-rhel8@sha256:2f1a7519617b5c0d082f943de0db295370649d968d699d76cfd829f4c9826a0f
        - registry.redhat.io/ibm-bamoe/bamoe-rhel8-operator@sha256:0646da8707d10b3cc2e2344269abd8797db34127a3326bf76bbc1547aa4bbd73
        - registry.redhat.io/ibm-bamoe/bamoe-process-migration-rhel8@sha256:a287a2fbaded371b8fcae5e4124b2d905bb228f94da0fc899ca2600b378dca0d
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-monitoring-rhel8@sha256:25025e489c15108edffe5270d257e1ed9be0a67aa067bcfc184647793aac2617
        - registry.redhat.io/ibm-bamoe/bamoe-dashbuilder-rhel8@sha256:efa55e17190b0ebe1618452f786876ef354b8fbec25c316aed7ed5bb1ce88116
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:457fc73423fa94591bc7e35729a25c58d3465583afccf342fb338b6e11188575
        - registry.redhat.io/ibm-bamoe/bamoe-smartrouter-rhel8@sha256:5ed8821328413ec11859ae317679f0692afff9945f134fc8909e07a3ac387794
        - registry.redhat.io/ibm-bamoe/bamoe-kieserver-rhel8@sha256:c9f08ddc6111cdd74faa0db5721d5730fc7716fccf4dc2a658595f7e0c184c10
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:f714b9c1555e7f27a3c9cf2cf687a74188f36b1283d4e09cec4db3c084d9fe64
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhscl/postgresql-12-rhel7@sha256:191237ccffc476e764714a69e35d9421dcee5ddd7184442d7a8306a06b97e6bb
        - registry.redhat.io/ibm-bamoe/bamoe-dashbuilder-rhel8@sha256:b3b7388b3b116ef37aabbb04b5d2c3e36e58d2fc95f1e59795b423010e071fc3
        - registry.redhat.io/ibm-bamoe/bamoe-controller-rhel8@sha256:3fcd9ae0bc194c198cbb5ee94272e80f302658733e467766bd1789357e9fd015
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:fb326be806b337b0ae1711f974d15e073aebcbad77aaa1123a63cf4f331da5f2
        - registry.redhat.io/ibm-bamoe/bamoe-controller-rhel8@sha256:6cd0f07856d612dca46b0a01a93b796f7a0f112b98cedfd6754f9fd770ee843d
        version: 8.0.8-1
      entries:
      - name: bamoe-businessautomation-operator.8.0.8-1
        version: 8.0.8-1
      - name: bamoe-businessautomation-operator.8.0.7-2
        version: 8.0.7-2
      - name: bamoe-businessautomation-operator.8.0.7-1
        version: 8.0.7-1
      - name: bamoe-businessautomation-operator.8.0.6-3
        version: 8.0.6-3
      - name: bamoe-businessautomation-operator.8.0.6-2
        version: 8.0.6-2
      - name: bamoe-businessautomation-operator.8.0.6-1
        version: 8.0.6-1
      - name: bamoe-businessautomation-operator.8.0.5-3
        version: 8.0.5-3
      - name: bamoe-businessautomation-operator.8.0.5-2
        version: 8.0.5-2
      - name: bamoe-businessautomation-operator.8.0.5-1
        version: 8.0.5-1
      - name: bamoe-businessautomation-operator.8.0.4-3
        version: 8.0.4-3
      - name: bamoe-businessautomation-operator.8.0.4-2
        version: 8.0.4-2
      - name: bamoe-businessautomation-operator.8.0.4-1
        version: 8.0.4-1
      - name: bamoe-businessautomation-operator.8.0.3-4
        version: 8.0.3-4
      - name: bamoe-businessautomation-operator.8.0.3-3
        version: 8.0.3-3
      - name: bamoe-businessautomation-operator.8.0.3-2
        version: 8.0.3-2
      - name: bamoe-businessautomation-operator.8.0.3-1
        version: 8.0.3-1
      - name: bamoe-businessautomation-operator.8.0.2-3
        version: 8.0.2-3
      - name: bamoe-businessautomation-operator.8.0.2-2
        version: 8.0.2-2
      - name: bamoe-businessautomation-operator.8.0.2-1
        version: 8.0.2-1
      - name: bamoe-businessautomation-operator.8.0.1-2
        version: 8.0.1-2
      - name: bamoe-businessautomation-operator.8.0.1-1
        version: 8.0.1-1
      - name: bamoe-businessautomation-operator.8.0.0-1
        version: 8.0.0-1
      name: 8.x-stable
    defaultChannel: 8.x-stable
    packageName: bamoe-businessautomation-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: eG Innovations
      provider-url: https://www.eginnovations.com
    name: eginnovations-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eginnovations-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eginnovations.com/v1beta1",
                "kind": "EgUniversalAgent",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eginnovations-operator"
                  },
                  "name": "eguniversalagent",
                  "namespace": "egagent"
                },
                "spec": {
                  "agent": {
                    "env": [
                      {
                        "name": "EG_MANAGER",
                        "value": "Replace with eG Manager IP"
                      },
                      {
                        "name": "EG_MANAGER_PORT",
                        "value": "Replace with eG Manager Port"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "true"
          containerImage: docker.io/eginnovations/universal-agent-operator:0.0.11
          createdAt: "2025-09-11T20:08:00Z"
          description: eG Innovations provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with eG Universal Agent Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eginnovations/universal-agent-operator
          support: eG Innovations
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgUniversalAgent is the Schema for the eguniversalagents
              API
            displayName: Eg Universal Agent
            kind: EgUniversalAgent
            name: eguniversalagents.eginnovations.com
            version: v1beta1
        description: |
          eG Innovations’ converged application and infrastructure performance monitoring capabilities provide end-to-end visibility and correlated analytics for every layer, every tier of your IT landscape.

          The eG Enterprise Universal Agent Operator automatically configures a host agent on every Kubernetes worker node. The host agent auto-discovers the worker nodes and application containers running as Pods on each node and tracks their performance and utilization levels. In-depth monitoring of applications running on containers is also provided using the same host agent. No additional agents are required for the containers.

          ### Capabilities
            * Rolls out the egagent pod per node to monitor its pods and the node itself
            * Automatic BTM profiler injection using webhook based injection for deep application monitoring and transection tracing.

          ### Installation
            1. Create the namespace "egagent" for installing eG Agent on OpenShift Cluster.

               <br />``` $ oc create ns egagent ```
            2. Add "egagent" serviceaccount to scc( Security Context Constraints ) of OpenShift Cluster as privileged using below command.

               <br />``` $ oc adm policy add-scc-to-user privileged -z egagent -n egagent ```
            2. Install eG Innovations Universal Agent operator from [OpenShift Container Platform](https://www.openshift.com/) OperatorHub.
            3. Create the eG Universal Agent CRD (Custom Resource Definition)
               a. Click "View YAML Example" and copy the yaml contents and save it to (filename).yaml on your host.
               b. To apply the CRD run the following command

               <br />``` $ oc apply -f (filename).yaml ```

          ### Required Parameters
            * EG\_MANAGER, EG\_MANAGER\_PORT - provide the ip and port of your eG Manager Installation.

          ### Advanced Options
            * Disable Certificate Checking - disable any certificate validation that may interact poorly with proxies within your cluster
            * Image Override - use a copy of the eG Iniversal Agent container image from a registry other that Docker's or Red Hat's
            * Environment variables - define environment variables for the eG Universal Agent container of for APM profiler.
            * Namespace Selectors- select a subset of namesapces to enable instrumentation for applications.

          For complete list of supported parameters and the detailed documentation available in [eG Innovations Docs](https://docs.eginnovations.com)
        displayName: eG Innovations Universal Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Tracing
        links:
        - name: eG Innovations Kubernetes Monitoring Overview
          url: https://www.eginnovations.com/documentation/Kubernetes/Introduction-to-Monitoring-Kubernetes.htm
        maintainers:
        - email: devops@eginnovations.com
          name: eG Innovations
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: eG Innovations
          url: https://www.eginnovations.com
        relatedImages:
        - docker.io/eginnovations/universal-agent-operator:0.0.11
        version: 0.0.11
      entries:
      - name: eginnovations-operator.v0.0.11
        version: 0.0.11
      - name: eginnovations-operator.v0.0.10
        version: 0.0.10
      - name: eginnovations-operator.v0.0.9
        version: 0.0.9
      - name: eginnovations-operator.v0.0.8
        version: 0.0.8
      name: beta
    defaultChannel: beta
    packageName: eginnovations-operator
    provider:
      name: eG Innovations
      url: https://www.eginnovations.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fortanix Inc.
      provider-url: https://fortanix.com
    name: ccm-node-agent-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ccm-node-agent-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1",
                "kind": "CcmNodeAgent",
                "metadata": {
                  "name": "ccmnodeagent-sample"
                },
                "spec": {
                  "ccmJoinToken": {
                    "secretName": "ccm-join-token",
                    "token": ""
                  },
                  "fullnameOverride": "",
                  "image": {
                    "ccmNodeAgent": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/ccm-node-agent-ubi7",
                      "tag": "latest"
                    },
                    "ccmNodeAgentInit": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "busybox",
                      "tag": "latest"
                    },
                    "sgxSoftwareInstaller": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/sgx-software-installer-ubi7",
                      "tag": "latest"
                    }
                  },
                  "log": {
                    "debug": false
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "ports": {
                    "containerPort": 9092,
                    "hostPort": 9092
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CCM Node Agent software enables registration of the compute
              nodes to Fortanix CCM when installed in the Openshift Cluster Nodes<U+200B>.
            displayName: CCM Node Agent
            kind: CcmNodeAgent
            name: ccmnodeagents.charts.operatorhub.io
            version: v1
        description: "CCM Node Agent software enables registration of the compute
          nodes to Fortanix CCM when installed in the Openshift Cluster Nodes​. \nIt
          also assists with application attestation and visibility for Fortanix CCM.\n\n##
          About this Operator\nThis Operator is based on a Helm chart for CCM Node
          Agent. It enrolls all the nodes in the openshift cluster to a particular
          CCM account. You have to provide the base64 encode of join token of CCM
          account while creating CcmNodeAgent resource.\n\n## Prerequisites\nAll the
          worker nodes in the cluster must have SGX capabilities.\n\n## Installation\n*
          Choose a namespace to install the Operator. Operator will be available in
          all namespaces.\n* Select Approval Strategy: Available options are Automatic
          or Manual. Automatic approval will automatically upgrade the running instance
          of your Operator without human intervention when a new version is available.
          In case of Manual approval, when a new version is available, OLM will create
          an update request. You must manually approve the request to update the Operator
          to a new version.\n* Generate your CCM Account Join Token: Please log in
          to https://ccm.fortanix.com, and in the Management Console tab, click the
          + ENROLL NODE button. In the ENROLL NODE screen, a Join Token will be generated
          in the text box for \"Get a join token to register an SGX compute node\".
          This Join Token is used by the compute node to authenticate itself.\n* Base64
          encode the Join Token:\n```shell\necho -n <join-token> | openssl base64
          -A\n```\n* Create an instance of CcmNodeAgent by setting the `spec.ccmJoinToken.Token`
          to the base64 encoded CCM account join token.\n* After CcmNodeAgent resource
          is installed successfully you should be able to see all the worker nodes
          of the cluster in your CCM account under Compute Node Tab.\n\n## Getting
          Started\nNow to run your application inside SGX enclaves on your cluster
          follow the [documentation](https://support.fortanix.com/hc/en-us/articles/360043527431-User-s-Guide-Add-and-Edit-an-Application).\n"
        displayName: CCM Node Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dockerfile
        - Helm
        - CCM
        - Confidential Compute
        links:
        - name: CCM Node Agent Operator
          url: https://support.fortanix.com/hc/en-us/categories/360003107511-Confidential-Computing-Manager
        maintainers:
        - email: info@fortanix.com
          name: Fortanix Inc.
        maturity: alpha
        provider:
          name: Fortanix Inc.
          url: https://fortanix.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62575404000cb77fa8d2284c87b31e39e667a55b0fb47821263b23b3d1135a7e
        - docker.io/fortanix/ccm-node-agent-operator@sha256:e2a3d52551198102ddbe7f4ebd9621edc357c57ddeff47b7ed9f2cf56a1bed39
        version: 0.0.2
      entries:
      - name: ccm-node-agent-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ccm-node-agent-operator
    provider:
      name: Fortanix Inc.
      url: https://fortanix.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: https://intel.com
    name: intel-device-plugins-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: intel-device-plugins-operator.v0.29.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "GpuDevicePlugin",
                "metadata": {
                  "name": "gpudeviceplugin-sample"
                },
                "spec": {
                  "enableMonitoring": true,
                  "image": "registry.connect.redhat.com/intel/intel-gpu-plugin@sha256:963caf1e6e16b5e297671c536eec04dc57cf45f5cc5cf458bdcafa56bdb97972",
                  "preferredAllocationPolicy": "none",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/gpu": "true"
                  },
                  "sharedDevNum": 1
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "QatDevicePlugin",
                "metadata": {
                  "name": "qatdeviceplugin-sample"
                },
                "spec": {
                  "dpdkDriver": "vfio-pci",
                  "image": "registry.connect.redhat.com/intel/intel-qat-plugin@sha256:53617c3a72af169031cbaba1380cd6a25c9d2b50c9a2da822cf37b473c6e7a8f",
                  "initImage": "registry.connect.redhat.com/intel/intel-qat-initcontainer@sha256:14f282a0adae43262657f24ce99a620b671c9a2ce12a8216163e6b22ac6e252b",
                  "kernelVfDrivers": [
                    "c6xxvf",
                    "4xxxvf"
                  ],
                  "logLevel": 4,
                  "maxNumDevices": 128,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/qat": "true"
                  }
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "DsaDevicePlugin",
                "metadata": {
                  "name": "dsadeviceplugin-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/intel/intel-dsa-plugin@sha256:092575515739bf9a0f5982c3605b63382bc4d1250128269434c2bfe4a27ec99d",
                  "initImage": "registry.connect.redhat.com/intel/intel-idxd-config-initcontainer@sha256:649e4f4e7f362ba054fb85a67e3baa23ec9649c68df356d1e33721ca3eef2f6b",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/dsa": "true"
                  },
                  "sharedDevNum": 10
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "SgxDevicePlugin",
                "metadata": {
                  "name": "sgxdeviceplugin-sample"
                },
                "spec": {
                  "enclaveLimit": 110,
                  "image": "registry.connect.redhat.com/intel/intel-sgx-plugin@sha256:e54cdf61feb0b67e330ad24d92e7fa4000953c2ec4d4300d052e032d419d5e0a",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/sgx": "true"
                  },
                  "provisionLimit": 110
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Security, AI/Machine Learning
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/intel/intel-technology-enabling-for-openshift
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GpuDevicePlugin is the Schema for the gpudeviceplugins API.
              It represents the GPU device plugin responsible for advertising Intel
              GPU hardware resources to the kubelet.
            displayName: Intel GPU Device Plugin
            kind: GpuDevicePlugin
            name: gpudeviceplugins.deviceplugin.intel.com
            version: v1
          - description: DsaDevicePlugin is the Schema for the dsadeviceplugins API.
              It represents the DSA device plugin responsible for advertising Intel
              DSA hardware resources to the kubelet.
            displayName: Intel DSA Device Plugin
            kind: DsaDevicePlugin
            name: dsadeviceplugins.deviceplugin.intel.com
            version: v1
          - description: QatDevicePlugin is the Schema for the qatdeviceplugins API.
              It represents the QAT device plugin responsible for advertising Intel
              QuickAssist Technology hardware resources to the kubelet.
            displayName: Intel QuickAssist Technology Device Plugin
            kind: QatDevicePlugin
            name: qatdeviceplugins.deviceplugin.intel.com
            version: v1
          - description: SgxDevicePlugin is the Schema for the sgxdeviceplugins API.
              It represents the SGX device plugin responsible for advertising SGX
              device nodes to the kubelet.
            displayName: Intel Software Guard Extensions Device Plugin
            kind: SgxDevicePlugin
            name: sgxdeviceplugins.deviceplugin.intel.com
            version: v1
        description: |
          [Intel Device Plugins for Kubernetes](https://github.com/intel/intel-technology-enabling-for-openshift) is a collection of
          [device plugins](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/) advertising Intel specific hardware resources
          to the kubelet. Currently the operator has basic support for Intel SGX, Intel GPU, Intel QAT and Intel DSA. Intel IAA device plugin will be supported in a future release.
        displayName: Intel Device Plugins Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - qat
        - sgx
        - gpu
        - dsa
        - dlb
        links:
        - name: Intel Device Plugins Operator
          url: https://github.com/intel/intel-technology-enabling-for-openshift
        maintainers:
        - email: martin.xu@intel.com
          name: Martin Xu
        - email: chaitanya.kulkarni@intel.com
          name: Chaitanya Kulkarni
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Intel® Corporation
          url: https://intel.com
        relatedImages:
        - registry.connect.redhat.com/intel/intel-sgx-plugin@sha256:e54cdf61feb0b67e330ad24d92e7fa4000953c2ec4d4300d052e032d419d5e0a
        - registry.connect.redhat.com/intel/intel-gpu-plugin@sha256:963caf1e6e16b5e297671c536eec04dc57cf45f5cc5cf458bdcafa56bdb97972
        - registry.connect.redhat.com/intel/intel-deviceplugin-operator@sha256:283e8b60d7ab8560f8ec8875b905e6446b59a91dd5e0d97da72e3a9aab235cc4
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - registry.connect.redhat.com/intel/intel-qat-plugin@sha256:53617c3a72af169031cbaba1380cd6a25c9d2b50c9a2da822cf37b473c6e7a8f
        - registry.connect.redhat.com/intel/intel-qat-initcontainer@sha256:14f282a0adae43262657f24ce99a620b671c9a2ce12a8216163e6b22ac6e252b
        - registry.connect.redhat.com/intel/intel-dsa-plugin@sha256:092575515739bf9a0f5982c3605b63382bc4d1250128269434c2bfe4a27ec99d
        - registry.connect.redhat.com/intel/intel-idxd-initcontainer@sha256:649e4f4e7f362ba054fb85a67e3baa23ec9649c68df356d1e33721ca3eef2f6b
        version: 0.29.1
      entries:
      - name: intel-device-plugins-operator.v0.29.1
        version: 0.29.1
      - name: intel-device-plugins-operator.v0.29.0
        version: 0.29.0
      - name: intel-device-plugins-operator.v0.28.0
        version: 0.28.0
      name: alpha
    defaultChannel: alpha
    packageName: intel-device-plugins-operator
    provider:
      name: Intel® Corporation
      url: https://intel.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: designate-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: designate-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "designate.openstack.org/v1beta1",
                "kind": "Designate",
                "metadata": {
                  "name": "designate"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "designate",
                  "databaseInstance": "openstack",
                  "designateAPI": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate"
                  },
                  "designateBackendbind9": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "networkAttachments": [
                      "designate"
                    ],
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate",
                    "storageClass": "local-storage",
                    "storageRequest": "1G"
                  },
                  "designateCentral": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate"
                  },
                  "designateMdns": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "networkAttachments": [
                      "designate"
                    ],
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate"
                  },
                  "designateProducer": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate"
                  },
                  "designateUnbound": {
                    "replicas": 1
                  },
                  "designateWorker": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "designate",
                    "networkAttachments": [
                      "designate"
                    ],
                    "rabbitMqClusterName": "rabbitmq",
                    "secret": "osp-secret",
                    "serviceUser": "designate"
                  },
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "osp-secret",
                  "serviceUser": "designate"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:48Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DesignateAPI is the Schema for the designateapis API
            displayName: Designate API
            kind: DesignateAPI
            name: designateapis.designate.openstack.org
            version: v1beta1
          - kind: DesignateBackendbind9
            name: designatebackendbind9s.designate.openstack.org
            version: v1beta1
          - description: DesignateCentral is the Schema for the designatecentral API
            displayName: Designate Central
            kind: DesignateCentral
            name: designatecentrals.designate.openstack.org
            version: v1beta1
          - description: DesignateMdns is the Schema for the designatemdnses API
            displayName: Designate Mdns
            kind: DesignateMdns
            name: designatemdnses.designate.openstack.org
            version: v1beta1
          - description: DesignateProducer is the Schema for the designateproducer
              API
            displayName: Designate Producer
            kind: DesignateProducer
            name: designateproducers.designate.openstack.org
            version: v1beta1
          - description: Designate is the Schema for the designates API
            displayName: Designate
            kind: Designate
            name: designates.designate.openstack.org
            version: v1beta1
          - description: DesignateUnbound is the Schema for the designateunbound API
            displayName: Designate Unbound
            kind: DesignateUnbound
            name: designateunbounds.designate.openstack.org
            version: v1beta1
          - description: DesignateWorker is the Schema for the designateworker API
            displayName: Designate Worker
            kind: DesignateWorker
            name: designateworkers.designate.openstack.org
            version: v1beta1
        description: Designate Operator
        displayName: Designate Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Network
        - Designate
        links:
        - name: Designate Operator
          url: https://github.com/openstack-k8s-operators/designate-operator
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso/openstack-designate-api-rhel9@sha256:1a9baa4a0f99029eb8a367ddcce264581aa37fdf1834ba277c30a3f0ae33f400
        - registry.redhat.io/rhoso/openstack-designate-central-rhel9@sha256:ad610cdc29b31cd82a81cea22793176351574f43c0f7705273c8ffc03ad2fdac
        - registry.redhat.io/rhoso/openstack-designate-producer-rhel9@sha256:4beb652a073a65965f96f125d0fcb454255efa5e53a84bd0dd9120b57e59fcf3
        - registry.redhat.io/rhoso-operators/designate-rhel9-operator@sha256:e6c1ba511cfd0278fc8cd2836c6324d97ae3f6e9affc992600978062fa2d4549
        - registry.redhat.io/rhoso/openstack-designate-mdns-rhel9@sha256:a4109755360f340aadf6bc15d24599f882d82bd59739065f87275f9274d9eeb9
        - registry.redhat.io/rhoso/openstack-designate-worker-rhel9@sha256:94ef291ca85b06dffc93d149b09b412d9190c99f7a1b0b300d50bd70346401ad
        - registry.redhat.io/rhoso/openstack-designate-backend-bind9-rhel9@sha256:89ce30fc1eff339867ce422203e6e0c63296667e102ce85d823d44c30f7556ad
        - registry.redhat.io/rhoso/openstack-unbound-rhel9@sha256:ad95d40227a1e52f739f09a2ce8db647d6a2fa7ef63e58a2444688728fa5bdfd
        version: 1.0.6
      entries:
      - name: designate-operator.v1.0.6
        version: 1.0.6
      - name: designate-operator.v1.0.5
        version: 1.0.5
      - name: designate-operator.v1.0.4
        version: 1.0.4
      - name: designate-operator.v1.0.3
        version: 1.0.3
      - name: designate-operator.v1.0.2
        version: 1.0.2
      - name: designate-operator.v1.0.1
        version: 1.0.1
      - name: designate-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: designate-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-cert-manager-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-manager-operator.v1.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
          createdAt: 2025-08-06T08:21:40
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.16.1 <1.17.0'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.17.4](https://github.com/cert-manager/cert-manager/tree/v1.17.4), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:96d51e3a64bf30cbd92836c7cbd82f06edca16eef78ab1432757d34c16628659
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b
        version: 1.17.0
      entries:
      - name: cert-manager-operator.v1.17.0
        version: 1.17.0
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      - name: cert-manager-operator.v1.14.1
        version: 1.14.1
      - name: cert-manager-operator.v1.14.0
        version: 1.14.0
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1
    - currentCSV: cert-manager-operator.v1.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "\u003cACCESS_KEY_ID\u003e",
                            "hostedZoneID": "\u003cHOSTED_ZONE_ID\u003e",
                            "region": "\u003cAWS_REGION\u003e",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4e854c14fecbd27107b91b76164ce2e1575ce60202d835abba5962fb5d386560
          createdAt: 2024-04-30T13:29:33
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.13.0 <1.13.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/infrastructure-features: '["proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.13.6](https://github.com/cert-manager/cert-manager/tree/v1.13.6), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:afc82895c09d75e3d6a932aea2d3e3ef998f765306da92070de1dab26baf0593
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4e854c14fecbd27107b91b76164ce2e1575ce60202d835abba5962fb5d386560
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:578f74c3e96db7fc2f4e94dd9fbc7d6499d87b0bc55576930ee17de08a95022c
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:7915f8049d2451b425e3674eb60e596b9b11946ffda1497f5d5607799e2bf431
        version: 1.13.1
      entries:
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1.13
    - currentCSV: cert-manager-operator.v1.14.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:5c593bfd70309f87cd05716e233e65aa55394eafcdaf85f178698912cb29d79d
          createdAt: 2025-04-09T05:19:10
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.14.1 <1.14.2'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/infrastructure-features: '["proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.14.7](https://github.com/cert-manager/cert-manager/tree/v1.14.7), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:5c593bfd70309f87cd05716e233e65aa55394eafcdaf85f178698912cb29d79d
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:5df082f8e3787e6ac2edf395113806999ae192425ce18b4ddba9ab20c135fa5b
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:49fb8f710319d6f3479b98d2c5dd7d0c9df8c1678b2c00727314dea1a8e933a1
        version: 1.14.2
      entries:
      - name: cert-manager-operator.v1.14.2
        version: 1.14.2
      - name: cert-manager-operator.v1.14.1
        version: 1.14.1
      - name: cert-manager-operator.v1.14.0
        version: 1.14.0
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1.14
    - currentCSV: cert-manager-operator.v1.15.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:f328263e2d29e34ede65e4501f0447b2d9f84e9445a365c2fa2fbb253939e274
          createdAt: 2025-03-13T18:26:37
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.15.0 <1.15.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/infrastructure-features: '["proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.15.5](https://github.com/cert-manager/cert-manager/tree/v1.15.2), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:f328263e2d29e34ede65e4501f0447b2d9f84e9445a365c2fa2fbb253939e274
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9573d74bd2b926ec94af76f813e6358f14c5b2f4e0eedab7c1ff1070b7279a5c
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:6af3ee8b2a5a87042fb7158bda8d6cf2e6324d1e265974acf77214d4cd0ea0d3
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:20efff60a0caf5eafb38986fd21611697b5bc534c2e789da233983a9739938ed
        version: 1.15.1
      entries:
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      - name: cert-manager-operator.v1.14.1
        version: 1.14.1
      - name: cert-manager-operator.v1.14.0
        version: 1.14.0
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1.15
    - currentCSV: cert-manager-operator.v1.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:d30e755a266bb035d50e2b257e7a26c1fe6a5ccb66125576db99d537aa03ec27
          createdAt: 2025-07-10T08:52:43
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.15.1 <1.16.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.16.5](https://github.com/cert-manager/cert-manager/tree/v1.16.5), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:831e4b1e5aa6eed6fbb22f779ddbcb9c73e4bf6cc079d0c6652737a6a44e6320
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:d30e755a266bb035d50e2b257e7a26c1fe6a5ccb66125576db99d537aa03ec27
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:33308217d2226c2555eb18020710d16175295a8e349263a5435e4b158bbc735f
        version: 1.16.1
      entries:
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      - name: cert-manager-operator.v1.14.1
        version: 1.14.1
      - name: cert-manager-operator.v1.14.0
        version: 1.14.0
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1.16
    - currentCSV: cert-manager-operator.v1.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
          createdAt: 2025-08-06T08:21:40
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.16.1 <1.17.0'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.17.4](https://github.com/cert-manager/cert-manager/tree/v1.17.4), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:96d51e3a64bf30cbd92836c7cbd82f06edca16eef78ab1432757d34c16628659
        version: 1.17.0
      entries:
      - name: cert-manager-operator.v1.17.0
        version: 1.17.0
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      - name: cert-manager-operator.v1.14.1
        version: 1.14.1
      - name: cert-manager-operator.v1.14.0
        version: 1.14.0
      - name: cert-manager-operator.v1.13.1
        version: 1.13.1
      - name: cert-manager-operator.v1.13.0
        version: 1.13.0
      name: stable-v1.17
    defaultChannel: stable-v1
    packageName: openshift-cert-manager-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Odoo Community Association
      provider-url: ""
    name: odoo-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odoo-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"odoo-community.org/v1alpha1","kind":"Odoo","metadata":{"name":"example"},"spec":{"postgresqlVolumeSize":"4Gi","odooVolumeSize":"4Gi"}}]'
          capabilities: Full Lifecycle
          categories: Application Runtime
          certified: "false"
          containerImage: ursa/odoo-operator:0.0.2
          createdAt: "2021-04-26T12:00:00Z"
          description: Operator to manage Odoo instances within an OpenShift cluster
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          repository: https://github.com/ursais/odoo-operator
          support: Open Source Integrators
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Odoo
            displayName: Odoo
            kind: Odoo
            name: odoos.odoo-community.org
            version: v1alpha1
        description: |
          ## About Odoo

          [Odoo](https://odoo-community.org), is a suite of business management software tools including CRM, e-commerce, billing, accounting, manufacturing, warehouse, project management, and inventory management to name a few.
          ## About the Odoo Operator

          This is an Ansible-based operator created with operator-sdk v1.5.0.

          The upgrade is handled by the Odoo container itself using [Marabunta](https://pypi.org/project/marabunta/) and [Anthem](https://pypi.org/project/anthem/).

          ## Roadmap
          * Add horizontal pod autoscaler
          * Add cronjob for backup
          ## Prerequisites for enabling this Operator

          Please refer to the README file on the [GitHub project](https://github.com/ursais/odoo-operator).
        displayName: Odoo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ERP
        - Business Software
        links:
        - name: Odoo Community Association
          url: https://odoo-community.org
        - name: GitHub Repository
          url: https://github.com/ursais/odoo-operator
        - name: Open Source Integrators
          url: https://opensourceintegrators.com
        maintainers:
        - email: mchambreuil@opensourceintegrators.com
          name: Maxime Chambreuil
        maturity: alpha
        provider:
          name: Odoo Community Association
        relatedImages:
        - ursa/odoo-operator:0.0.2
        version: 0.0.2
      entries:
      - name: odoo-operator.v0.0.2
        version: 0.0.2
      - name: odoo-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: odoo-operator
    provider:
      name: Odoo Community Association
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operator-kogitocloud: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kogito-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kogito-operator.v1.44.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "process-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogito-infinispan-infra"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "infinispan.org/v1",
                    "kind": "Infinispan"
                  }
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/kiegroup/kogito-operator:1.44.1
          createdAt: "2019-08-22T13:12:22Z"
          description: Kogito Operator for deployment and management of Kogito services.
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.app.kiegroup.org
            version: v1beta1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.app.kiegroup.org
            version: v1beta1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.app.kiegroup.org
            version: v1beta1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.app.kiegroup.org
            version: v1beta1
        description: |-
          Kogito Operator is designed to deploy Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Creates two kinds of build configuration: Native or JVM. For more information refer to [Native build and deployment](https://docs.jboss.org/kogito/release/latest/html_single/#_native_build_and_deployment) section on our docs
          * Builds a Kogito service from a remote git repository and creates a custom image for the given runtime: Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Installs the [Kogito Data Index Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-data-index-service_kogito-deploying-on-openshift) to enable data index feature for Kogito services.
          * Installs the [Kogito Jobs Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-jobs-service_kogito-deploying-on-openshift) to enable jobs scheduling.

          For Kubernetes deployment of Kogito services, please refer to the section [Deploying Kogito services on Kubernetes](https://docs.jboss.org/kogito/release/latest/html_single/#proc-kogito-deploying-on-kubernetes_kogito-deploying-on-openshift) on our docs.

          **IMPORTANT!** Kogito Operator might need Infinispan Operator and Strimzi to implement persistence and messaging use cases if you do not provide these services beforehand.
          If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the Kogito Operator.
        displayName: Kogito
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://kogito.kie.org/
        - name: Documentation
          url: https://github.com/kiegroup/kogito-operator/blob/main/README.md
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/kiegroup/kogito-operator:1.44.1
        version: 1.44.1
      entries:
      - name: kogito-operator.v1.44.1
        version: 1.44.1
      - name: kogito-operator.v1.43.0
        version: 1.43.0
      - name: kogito-operator.v1.42.0
        version: 1.42.0
      - name: kogito-operator.v1.41.0
        version: 1.41.0
      - name: kogito-operator.v1.40.0
        version: 1.40.0
      - name: kogito-operator.v1.39.0
        version: 1.39.0
      - name: kogito-operator.v1.38.0
        version: 1.38.0
      - name: kogito-operator.v1.37.0
        version: 1.37.0
      - name: kogito-operator.v1.36.0
        version: 1.36.0
      - name: kogito-operator.v1.35.0
        version: 1.35.0
      - name: kogito-operator.v1.34.0
        version: 1.34.0
      - name: kogito-operator.v1.33.0
        version: 1.33.0
      - name: kogito-operator.v1.32.0
        version: 1.32.0
      - name: kogito-operator.v1.31.0
        version: 1.31.0
      - name: kogito-operator.v1.30.0
        version: 1.30.0
      - name: kogito-operator.v1.29.0
        version: 1.29.0
      - name: kogito-operator.v1.28.0
        version: 1.28.0
      - name: kogito-operator.v1.27.0
        version: 1.27.0
      - name: kogito-operator.v1.26.0
        version: 1.26.0
      - name: kogito-operator.v1.25.0
        version: 1.25.0
      - name: kogito-operator.v1.24.0
        version: 1.24.0
      - name: kogito-operator.v1.23.0
        version: 1.23.0
      - name: kogito-operator.v1.22.1
        version: 1.22.1
      - name: kogito-operator.v1.22.0
        version: 1.22.0
      - name: kogito-operator.v1.21.0
        version: 1.21.0
      - name: kogito-operator.v1.20.0
        version: 1.20.0
      - name: kogito-operator.v1.19.0
        version: 1.19.0
      - name: kogito-operator.v1.18.0
        version: 1.18.0
      - name: kogito-operator.v1.17.0
        version: 1.17.0
      - name: kogito-operator.v1.16.1
        version: 1.16.1
      - name: kogito-operator.v1.16.0
        version: 1.16.0
      - name: kogito-operator.v1.15.0
        version: 1.15.0
      - name: kogito-operator.v1.14.0
        version: 1.14.0
      - name: kogito-operator.v1.13.0
        version: 1.13.0
      - name: kogito-operator.v1.12.0
        version: 1.12.0
      - name: kogito-operator.v1.11.1
        version: 1.11.1
      - name: kogito-operator.v1.11.0
        version: 1.11.0
      - name: kogito-operator.v1.10.0
        version: 1.10.0
      - name: kogito-operator.v1.9.1
        version: 1.9.1
      - name: kogito-operator.v1.9.0
        version: 1.9.0
      - name: kogito-operator.v1.8.0
        version: 1.8.0
      - name: kogito-operator.v1.7.0
        version: 1.7.0
      - name: kogito-operator.v1.6.0
        version: 1.6.0
      - name: kogito-operator.v1.5.0
        version: 1.5.0
      - name: kogito-operator.v1.4.0
        version: 1.4.0
      name: 1.x
    - currentCSV: kogito-operator.v1.44.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "process-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogito-infinispan-infra"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "infinispan.org/v1",
                    "kind": "Infinispan"
                  }
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/kiegroup/kogito-operator:1.44.1
          createdAt: "2019-08-22T13:12:22Z"
          description: Kogito Operator for deployment and management of Kogito services.
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.app.kiegroup.org
            version: v1beta1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.app.kiegroup.org
            version: v1beta1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.app.kiegroup.org
            version: v1beta1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.app.kiegroup.org
            version: v1beta1
        description: |-
          Kogito Operator is designed to deploy Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Creates two kinds of build configuration: Native or JVM. For more information refer to [Native build and deployment](https://docs.jboss.org/kogito/release/latest/html_single/#_native_build_and_deployment) section on our docs
          * Builds a Kogito service from a remote git repository and creates a custom image for the given runtime: Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Installs the [Kogito Data Index Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-data-index-service_kogito-deploying-on-openshift) to enable data index feature for Kogito services.
          * Installs the [Kogito Jobs Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-jobs-service_kogito-deploying-on-openshift) to enable jobs scheduling.

          For Kubernetes deployment of Kogito services, please refer to the section [Deploying Kogito services on Kubernetes](https://docs.jboss.org/kogito/release/latest/html_single/#proc-kogito-deploying-on-kubernetes_kogito-deploying-on-openshift) on our docs.

          **IMPORTANT!** Kogito Operator might need Infinispan Operator and Strimzi to implement persistence and messaging use cases if you do not provide these services beforehand.
          If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the Kogito Operator.
        displayName: Kogito
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://kogito.kie.org/
        - name: Documentation
          url: https://github.com/kiegroup/kogito-operator/blob/main/README.md
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/kiegroup/kogito-operator:1.44.1
        version: 1.44.1
      entries:
      - name: kogito-operator.v1.44.1
        version: 1.44.1
      - name: kogito-operator.v1.43.0
        version: 1.43.0
      - name: kogito-operator.v1.42.0
        version: 1.42.0
      - name: kogito-operator.v1.41.0
        version: 1.41.0
      - name: kogito-operator.v1.40.0
        version: 1.40.0
      - name: kogito-operator.v1.39.0
        version: 1.39.0
      - name: kogito-operator.v1.38.0
        version: 1.38.0
      - name: kogito-operator.v1.37.0
        version: 1.37.0
      - name: kogito-operator.v1.36.0
        version: 1.36.0
      - name: kogito-operator.v1.35.0
        version: 1.35.0
      - name: kogito-operator.v1.34.0
        version: 1.34.0
      - name: kogito-operator.v1.33.0
        version: 1.33.0
      - name: kogito-operator.v1.32.0
        version: 1.32.0
      - name: kogito-operator.v1.31.0
        version: 1.31.0
      - name: kogito-operator.v1.30.0
        version: 1.30.0
      - name: kogito-operator.v1.29.0
        version: 1.29.0
      - name: kogito-operator.v1.28.0
        version: 1.28.0
      - name: kogito-operator.v1.27.0
        version: 1.27.0
      - name: kogito-operator.v1.26.0
        version: 1.26.0
      - name: kogito-operator.v1.25.0
        version: 1.25.0
      - name: kogito-operator.v1.24.0
        version: 1.24.0
      - name: kogito-operator.v1.23.0
        version: 1.23.0
      - name: kogito-operator.v1.22.1
        version: 1.22.1
      - name: kogito-operator.v1.22.0
        version: 1.22.0
      - name: kogito-operator.v1.21.0
        version: 1.21.0
      - name: kogito-operator.v1.20.0
        version: 1.20.0
      - name: kogito-operator.v1.19.0
        version: 1.19.0
      - name: kogito-operator.v1.18.0
        version: 1.18.0
      - name: kogito-operator.v1.17.0
        version: 1.17.0
      - name: kogito-operator.v1.16.1
        version: 1.16.1
      - name: kogito-operator.v1.16.0
        version: 1.16.0
      - name: kogito-operator.v1.15.0
        version: 1.15.0
      - name: kogito-operator.v1.14.0
        version: 1.14.0
      - name: kogito-operator.v1.13.0
        version: 1.13.0
      - name: kogito-operator.v1.12.0
        version: 1.12.0
      - name: kogito-operator.v1.11.1
        version: 1.11.1
      - name: kogito-operator.v1.11.0
        version: 1.11.0
      - name: kogito-operator.v1.10.0
        version: 1.10.0
      - name: kogito-operator.v1.9.1
        version: 1.9.1
      - name: kogito-operator.v1.9.0
        version: 1.9.0
      - name: kogito-operator.v1.8.0
        version: 1.8.0
      - name: kogito-operator.v1.7.0
        version: 1.7.0
      - name: kogito-operator.v1.6.0
        version: 1.6.0
      - name: kogito-operator.v1.5.0
        version: 1.5.0
      - name: kogito-operator.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: 1.x
    packageName: kogito-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: neutron-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neutron-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "neutron.openstack.org/v1beta1",
                "kind": "NeutronAPI",
                "metadata": {
                  "name": "neutron",
                  "namespace": "openstack"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "neutron",
                  "databaseInstance": "openstack",
                  "memcachedInstance": "memcached",
                  "preserveJobs": false,
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "neutron-secret",
                  "serviceUser": "neutron"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:57Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NeutronAPI is the Schema for the neutronapis API
            displayName: Neutron API
            kind: NeutronAPI
            name: neutronapis.neutron.openstack.org
            version: v1beta1
        description: Neutron Operator
        displayName: Neutron Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Network
        - Neutron
        links:
        - name: Neutron Operator
          url: https://github.com/openstack-k8s-operators/neutron-operator
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/neutron-rhel9-operator@sha256:16b1a5f83692f3d193754df664992e5b88cb357f158b69261eb802430f3a69f1
        - registry.redhat.io/rhoso/openstack-neutron-server-rhel9@sha256:806e5e009a77a29c82699487e539669e1ec2136514a0332d222b5cb3fed6f452
        version: 1.0.6
      entries:
      - name: neutron-operator.v1.0.6
        version: 1.0.6
      - name: neutron-operator.v1.0.5
        version: 1.0.5
      - name: neutron-operator.v1.0.4
        version: 1.0.4
      - name: neutron-operator.v1.0.3
        version: 1.0.3
      - name: neutron-operator.v1.0.2
        version: 1.0.2
      - name: neutron-operator.v1.0.1
        version: 1.0.1
      - name: neutron-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: neutron-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ram-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ram-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ram-controller:1.1.1
          createdAt: "2025-09-19T22:18:09Z"
          description: AWS RAM controller is a service controller for managing RAM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Permission represents the state of an AWS ram Permission
              resource.
            displayName: Permission
            kind: Permission
            name: permissions.ram.services.k8s.aws
            version: v1alpha1
          - description: ResourceShare represents the state of an AWS ram ResourceShare
              resource.
            displayName: ResourceShare
            kind: ResourceShare
            name: resourceshares.ram.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon RAM resources in AWS from within your Kubernetes cluster.

          **About Amazon RAM**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon RAM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ram
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon RAM Developer Resources
          url: https://aws.amazon.com/RAM/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ram maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ram-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ram-controller.v1.1.1
        version: 1.1.1
      - name: ack-ram-controller.v1.1.0
        version: 1.1.0
      - name: ack-ram-controller.v1.0.11
        version: 1.0.11
      - name: ack-ram-controller.v1.0.10
        version: 1.0.10
      name: alpha
    defaultChannel: alpha
    packageName: ack-ram-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: project-quay
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-operator.v3.10.0-nightly.20231014
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.10.0-nightly.20231014
          createdAt: 2023-10-14 06:26 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.10.0-nightly.20231014'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.10.0-nightly.20231014
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.10.0-nightly.20231014
        version: 3.10.0-nightly.20231014
      entries:
      - name: quay-operator.v3.10.0-nightly.20231014
        version: 3.10.0-nightly.20231014
      - name: quay-operator.v3.10.0-nightly.20231013
        version: 3.10.0-nightly.20231013
      - name: quay-operator.v3.10.0-nightly.20231012
        version: 3.10.0-nightly.20231012
      - name: quay-operator.v3.10.0-nightly.20231011
        version: 3.10.0-nightly.20231011
      - name: quay-operator.v3.10.0-nightly.20231010
        version: 3.10.0-nightly.20231010
      - name: quay-operator.v3.10.0-nightly.20231009
        version: 3.10.0-nightly.20231009
      - name: quay-operator.v3.10.0-nightly.20231008
        version: 3.10.0-nightly.20231008
      - name: quay-operator.v3.10.0-nightly.20231007
        version: 3.10.0-nightly.20231007
      - name: quay-operator.v3.10.0-nightly.20231006
        version: 3.10.0-nightly.20231006
      - name: quay-operator.v3.10.0-nightly.20231005
        version: 3.10.0-nightly.20231005
      - name: quay-operator.v3.10.0-nightly.20231002
        version: 3.10.0-nightly.20231002
      - name: quay-operator.v3.10.0-nightly.20230930
        version: 3.10.0-nightly.20230930
      - name: quay-operator.v3.10.0-nightly.20230929
        version: 3.10.0-nightly.20230929
      - name: quay-operator.v3.10.0-nightly.20230928
        version: 3.10.0-nightly.20230928
      - name: quay-operator.v3.10.0-nightly.20230926
        version: 3.10.0-nightly.20230926
      - name: quay-operator.v3.10.0-nightly.20230925
        version: 3.10.0-nightly.20230925
      - name: quay-operator.v3.10.0-nightly.20230923
        version: 3.10.0-nightly.20230923
      - name: quay-operator.v3.10.0-nightly.20230922
        version: 3.10.0-nightly.20230922
      - name: quay-operator.v3.10.0-nightly.20230921
        version: 3.10.0-nightly.20230921
      - name: quay-operator.v3.10.0-nightly.20230920
        version: 3.10.0-nightly.20230920
      - name: quay-operator.v3.10.0-nightly.20230919
        version: 3.10.0-nightly.20230919
      - name: quay-operator.v3.10.0-nightly.20230918
        version: 3.10.0-nightly.20230918
      - name: quay-operator.v3.10.0-nightly.20230917
        version: 3.10.0-nightly.20230917
      - name: quay-operator.v3.10.0-nightly.20230916
        version: 3.10.0-nightly.20230916
      - name: quay-operator.v3.10.0-nightly.20230915
        version: 3.10.0-nightly.20230915
      - name: quay-operator.v3.10.0-nightly.20230914
        version: 3.10.0-nightly.20230914
      - name: quay-operator.v3.10.0-nightly.20230913
        version: 3.10.0-nightly.20230913
      name: candidate-3.10
    - currentCSV: quay-operator.v3.9.0-nightly.20230812
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.9.0-nightly.20230812
          createdAt: 2023-08-12 06:06 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.9.0-nightly.20230812'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.9.0-nightly.20230812
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.9.0-nightly.20230812
        version: 3.9.0-nightly.20230812
      entries:
      - name: quay-operator.v3.9.0-nightly.20230812
        version: 3.9.0-nightly.20230812
      - name: quay-operator.v3.9.0-nightly.20230724
        version: 3.9.0-nightly.20230724
      - name: quay-operator.v3.9.0-nightly.20230723
        version: 3.9.0-nightly.20230723
      - name: quay-operator.v3.9.0-nightly.20230722
        version: 3.9.0-nightly.20230722
      - name: quay-operator.v3.9.0-nightly.20230721
        version: 3.9.0-nightly.20230721
      - name: quay-operator.v3.9.0-nightly.20230720
        version: 3.9.0-nightly.20230720
      - name: quay-operator.v3.9.0-nightly.20230717
        version: 3.9.0-nightly.20230717
      - name: quay-operator.v3.9.0-nightly.20230716
        version: 3.9.0-nightly.20230716
      - name: quay-operator.v3.9.0-nightly.20230715
        version: 3.9.0-nightly.20230715
      - name: quay-operator.v3.9.0-nightly.20230714
        version: 3.9.0-nightly.20230714
      - name: quay-operator.v3.9.0-nightly.20230712
        version: 3.9.0-nightly.20230712
      - name: quay-operator.v3.9.0-nightly.20230711
        version: 3.9.0-nightly.20230711
      name: candidate-3.9
    - currentCSV: quay-operator.v3.7.0-rc.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.7.0-rc.4
          createdAt: 2022-05-17 14:05 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.3.x <3.7.0-rc.4'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.7.0-rc.4
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.7.0-rc.4
        version: 3.7.0-rc.4
      entries:
      - name: quay-operator.v3.7.0-rc.4
        version: 3.7.0-rc.4
      - name: quay-operator.v3.7.0-rc.3
        version: 3.7.0-rc.3
      - name: quay-operator.v3.7.0-rc.2
        version: 3.7.0-rc.2
      - name: quay-operator.v3.7.0-rc.1
        version: 3.7.0-rc.1
      name: preview-3.7
    - currentCSV: quay-operator.v3.8.0-rc.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.8.0-rc.7
          createdAt: 2022-11-29 15:11 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.8.0-rc.7'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.8.0-rc.7
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.8.0-rc.7
        version: 3.8.0-rc.7
      entries:
      - name: quay-operator.v3.8.0-rc.7
        version: 3.8.0-rc.7
      - name: quay-operator.v3.8.0-rc.6
        version: 3.8.0-rc.6
      - name: quay-operator.v3.8.0-rc.5
        version: 3.8.0-rc.5
      - name: quay-operator.v3.8.0-rc.4
        version: 3.8.0-rc.4
      - name: quay-operator.v3.8.0-rc.3
        version: 3.8.0-rc.3
      - name: quay-operator.v3.8.0-rc.2
        version: 3.8.0-rc.2
      - name: quay-operator.v3.8.0-rc.1
        version: 3.8.0-rc.1
      name: preview-3.8
    - currentCSV: quay-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.10.15
          createdAt: "2025-09-25T14:06:05Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.10.15
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.10.15
        version: 3.10.15
      entries:
      - name: quay-operator.v3.10.15
        version: 3.10.15
      - name: quay-operator.v3.10.14
        version: 3.10.14
      - name: quay-operator.v3.10.13
        version: 3.10.13
      - name: quay-operator.v3.10.12
        version: 3.10.12
      - name: quay-operator.v3.10.11
        version: 3.10.11
      - name: quay-operator.v3.10.10
        version: 3.10.10
      - name: quay-operator.v3.10.9
        version: 3.10.9
      - name: quay-operator.v3.10.8
        version: 3.10.8
      - name: quay-operator.v3.10.7
        version: 3.10.7
      - name: quay-operator.v3.10.6
        version: 3.10.6
      - name: quay-operator.v3.10.5
        version: 3.10.5
      - name: quay-operator.v3.10.4
        version: 3.10.4
      - name: quay-operator.v3.10.3
        version: 3.10.3
      - name: quay-operator.v3.10.2
        version: 3.10.2
      - name: quay-operator.v3.10.1
        version: 3.10.1
      - name: quay-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.11.13
          createdAt: "2025-09-02T06:02:15Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.11.13
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.11.13
        version: 3.11.13
      entries:
      - name: quay-operator.v3.11.13
        version: 3.11.13
      - name: quay-operator.v3.11.12
        version: 3.11.12
      - name: quay-operator.v3.11.11
        version: 3.11.11
      - name: quay-operator.v3.11.10
        version: 3.11.10
      - name: quay-operator.v3.11.9
        version: 3.11.9
      - name: quay-operator.v3.11.8
        version: 3.11.8
      - name: quay-operator.v3.11.7
        version: 3.11.7
      - name: quay-operator.v3.11.6
        version: 3.11.6
      - name: quay-operator.v3.11.5
        version: 3.11.5
      - name: quay-operator.v3.11.4
        version: 3.11.4
      - name: quay-operator.v3.11.3
        version: 3.11.3
      - name: quay-operator.v3.11.2
        version: 3.11.2
      - name: quay-operator.v3.11.1
        version: 3.11.1
      - name: quay-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.12.12
          createdAt: "2025-09-23T09:59:30Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.12.12
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.12.12
        version: 3.12.12
      entries:
      - name: quay-operator.v3.12.12
        version: 3.12.12
      - name: quay-operator.v3.12.11
        version: 3.12.11
      - name: quay-operator.v3.12.10
        version: 3.12.10
      - name: quay-operator.v3.12.9
        version: 3.12.9
      - name: quay-operator.v3.12.8
        version: 3.12.8
      - name: quay-operator.v3.12.7
        version: 3.12.7
      - name: quay-operator.v3.12.6
        version: 3.12.6
      - name: quay-operator.v3.12.5
        version: 3.12.5
      - name: quay-operator.v3.12.4
        version: 3.12.4
      - name: quay-operator.v3.12.3
        version: 3.12.3
      - name: quay-operator.v3.12.2
        version: 3.12.2
      - name: quay-operator.v3.12.1
        version: 3.12.1
      - name: quay-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.13.8
          createdAt: "2025-08-28T03:28:40Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.13.8
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.13.8
        version: 3.13.8
      entries:
      - name: quay-operator.v3.13.8
        version: 3.13.8
      - name: quay-operator.v3.13.7
        version: 3.13.7
      - name: quay-operator.v3.13.6
        version: 3.13.6
      - name: quay-operator.v3.13.5
        version: 3.13.5
      - name: quay-operator.v3.13.4
        version: 3.13.4
      - name: quay-operator.v3.13.3
        version: 3.13.3
      - name: quay-operator.v3.13.2
        version: 3.13.2
      - name: quay-operator.v3.13.1
        version: 3.13.1
      - name: quay-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.14.5
          createdAt: "2025-09-22T14:19:49Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.14.5
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.14.5
        version: 3.14.5
      entries:
      - name: quay-operator.v3.14.5
        version: 3.14.5
      - name: quay-operator.v3.14.4
        version: 3.14.4
      - name: quay-operator.v3.14.3
        version: 3.14.3
      - name: quay-operator.v3.14.2
        version: 3.14.2
      - name: quay-operator.v3.14.1
        version: 3.14.1
      - name: quay-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.15.2
          createdAt: "2025-09-22T13:44:32Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.15.2
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.15.2
        version: 3.15.2
      entries:
      - name: quay-operator.v3.15.2
        version: 3.15.2
      - name: quay-operator.v3.15.1
        version: 3.15.1
      - name: quay-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-operator.v3.6.2
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"quay.redhat.com/v1\",\n    \"kind\":
            \"QuayRegistry\",\n    \"metadata\": {\n      \"name\": \"example-registry\"\n
            \   },\n    \"spec\": {\n      \"components\": [\n        {\"kind\": \"clair\",
            \"managed\": true},\n        {\"kind\": \"postgres\", \"managed\": true},\n
            \       {\"kind\": \"objectstorage\", \"managed\": true},\n        {\"kind\":
            \"redis\", \"managed\": true},\n        {\"kind\": \"horizontalpodautoscaler\",
            \"managed\": true},\n        {\"kind\": \"route\", \"managed\": true},\n
            \       {\"kind\": \"mirror\", \"managed\": true},\n        {\"kind\":
            \"monitoring\", \"managed\": true},\n        {\"kind\": \"tls\", \"managed\":
            true}\n      ]\n    }\n  }, \n  {\n    \"apiVersion\": \"redhatcop.redhat.io/v1alpha1\",\n
            \   \"kind\": \"QuayEcosystem\",\n    \"metadata\": {\n      \"name\":
            \"example-quayecosystem\"\n    },\n    \"spec\": {}\n  }\n]"
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:v3.6.2
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.5.x <3.6.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.6.2
          repository: https://github.com/quay/quay-operator
          support: Quay Team
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Project Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        - quay
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:v3.6.2
        version: 3.6.2
      entries:
      - name: quay-operator.v3.6.2
        version: 3.6.2
      name: stable-3.6
    - currentCSV: quay-operator.v3.7.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.7.11
          createdAt: 2023-02-22 10:02 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.3.x <3.7.11'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.7.11
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.7.11
        version: 3.7.11
      entries:
      - name: quay-operator.v3.7.11
        version: 3.7.11
      - name: quay-operator.v3.7.10
        version: 3.7.10
      - name: quay-operator.v3.7.9
        version: 3.7.9
      - name: quay-operator.v3.7.8
        version: 3.7.8
      - name: quay-operator.v3.7.7
        version: 3.7.7
      - name: quay-operator.v3.7.6
        version: 3.7.6
      - name: quay-operator.v3.7.5
        version: 3.7.5
      - name: quay-operator.v3.7.4
        version: 3.7.4
      - name: quay-operator.v3.7.3
        version: 3.7.3
      - name: quay-operator.v3.7.2
        version: 3.7.2
      - name: quay-operator.v3.7.1
        version: 3.7.1
      - name: quay-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: quay-operator.v3.8.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.8.13
          createdAt: 2023-10-17 16:13 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.8.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.8.13
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.8.13
        version: 3.8.13
      entries:
      - name: quay-operator.v3.8.13
        version: 3.8.13
      - name: quay-operator.v3.8.12
        version: 3.8.12
      - name: quay-operator.v3.8.11
        version: 3.8.11
      - name: quay-operator.v3.8.10
        version: 3.8.10
      - name: quay-operator.v3.8.8
        version: 3.8.8
      - name: quay-operator.v3.8.7
        version: 3.8.7
      - name: quay-operator.v3.8.6
        version: 3.8.6
      - name: quay-operator.v3.8.5
        version: 3.8.5
      - name: quay-operator.v3.8.4
        version: 3.8.4
      - name: quay-operator.v3.8.3
        version: 3.8.3
      - name: quay-operator.v3.8.2
        version: 3.8.2
      - name: quay-operator.v3.8.1
        version: 3.8.1
      - name: quay-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.9.16
          createdAt: "2025-09-25T14:31:58Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.9.16
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.9.16
        version: 3.9.16
      entries:
      - name: quay-operator.v3.9.16
        version: 3.9.16
      - name: quay-operator.v3.9.15
        version: 3.9.15
      - name: quay-operator.v3.9.14
        version: 3.9.14
      - name: quay-operator.v3.9.13
        version: 3.9.13
      - name: quay-operator.v3.9.12
        version: 3.9.12
      - name: quay-operator.v3.9.11
        version: 3.9.11
      - name: quay-operator.v3.9.10
        version: 3.9.10
      - name: quay-operator.v3.9.9
        version: 3.9.9
      - name: quay-operator.v3.9.6
        version: 3.9.6
      - name: quay-operator.v3.9.5
        version: 3.9.5
      - name: quay-operator.v3.9.4
        version: 3.9.4
      - name: quay-operator.v3.9.3
        version: 3.9.3
      - name: quay-operator.v3.9.2
        version: 3.9.2
      - name: quay-operator.v3.9.1
        version: 3.9.1
      - name: quay-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: project-quay
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-ui
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-ui.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Ui",
                "metadata": {
                  "name": "ui-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "volumeMounts": [
                    {
                      "mountPath": "/opt/app-root/src/assets/config.json",
                      "name": "config",
                      "subPath": "config.json"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:37:58Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ui
            name: uis.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP UI Operator
        displayName: Keysight WAP UI Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Ui
          url: https://keysight-wap-ui.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-ui@sha256:876a394f5cc2f745091ad2ee9ca0d08404cfb95284222e1f3a8607e193bd986a
        version: 0.1.21
      entries:
      - name: keysight-wap-ui.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-ui
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator-certified-rhmp.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator-certified-rhmp.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator-certified-rhmp.v2.6.1
        version: 2.6.1
      - name: percona-postgresql-operator-certified-rhmp.v2.6.0
        version: 2.6.0
      - name: percona-postgresql-operator-certified-rhmp.v2.5.0
        version: 2.5.0
      name: stable
    - currentCSV: percona-postgresql-operator-certified-rhmp.v2.7.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        version: 2.7.0-cw
      entries:
      - name: percona-postgresql-operator-certified-rhmp.v2.7.0-cw
        version: 2.7.0-cw
      - name: percona-postgresql-operator-certified-rhmp.v2.6.1-cw
        version: 2.6.1-cw
      - name: percona-postgresql-operator-certified-rhmp.v2.6.0-cw
        version: 2.6.0-cw
      - name: percona-postgresql-operator-certified-rhmp.v2.5.0-cw
        version: 2.5.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-postgresql-operator-certified-rhmp
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Tempo Operator SIG
      provider-url: ""
    name: tempo-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tempo-operator.v0.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoMonolithic",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "jaegerui": {
                    "enabled": true,
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    }
                  },
                  "storage": {
                    "traces": {
                      "backend": "memory"
                    }
                  }
                }
              },
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoStack",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "resources": {
                    "total": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2Gi"
                      }
                    }
                  },
                  "storage": {
                    "secret": {
                      "name": "my-storage-secret",
                      "type": "s3"
                    }
                  },
                  "storageSize": "1Gi",
                  "template": {
                    "queryFrontend": {
                      "jaegerQuery": {
                        "enabled": true,
                        "ingress": {
                          "type": "route"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          containerImage: ghcr.io/grafana/tempo-operator/tempo-operator:v0.18.0
          createdAt: "2025-09-17T14:23:28Z"
          description: Create and manage deployments of Tempo, a high-scale distributed
            tracing backend.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: tempo-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/tempo-operator
          support: Grafana Tempo Operator SIG
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TempoMonolithic manages a Tempo deployment in monolithic
              mode.
            displayName: TempoMonolithic
            kind: TempoMonolithic
            name: tempomonolithics.tempo.grafana.com
            version: v1alpha1
          - description: TempoStack manages a Tempo deployment in microservices mode.
            displayName: TempoStack
            kind: TempoStack
            name: tempostacks.tempo.grafana.com
            version: v1alpha1
        description: |-
          Tempo is an open source, easy-to-use, and high-scale distributed tracing backend.
          It can ingest common open source tracing protocols including Jaeger, Zipkin, and OpenTelemetry and requires only object storage to operate.
          Please refer to the [Tempo documentation](https://grafana.com/docs/tempo/latest/) for more information about Tempo.

          The Community Tempo Operator can manage Tempo deployments in Microservices mode (`TempoStack` CR) and Monolithic mode (`TempoMonolithic` CR).

          ### Operator features
          * **Resource Limits** - Specify overall resource requests and limits in the `TempoStack` CR; the operator assigns fractions of it to each component
          * **AuthN and AuthZ** - Supports OpenID Control (OIDC) and role-based access control (RBAC)
          * **Managed upgrades** - Updating the operator will automatically update all managed Tempo clusters
          * **Multitenancy** - Multiple tenants can send traces to the same Tempo cluster
          * **mTLS** - Communication between the Tempo components can be secured via mTLS
          * **Jaeger UI** - Traces can be visualized in Jaeger UI and exposed via Ingress or OpenShift Route
          * **Observability** - The operator and `TempoStack` operands expose telemetry (metrics, traces) and integrate with Prometheus `ServiceMonitor` and `PrometheusRule`

          ### Prerequisites
          `TempoStack` requires object storage to store its traces. `TempoMonolithic` can store traces in-memory, in a Persistent Volume and in object storage.
          Please ensure that an object storage solution is available and configured.
        displayName: Community Tempo Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tempo
        - tracing
        - observability
        - monitoring
        - database
        links:
        - name: Tempo Operator
          url: https://github.com/grafana/tempo-operator
        maintainers:
        - email: ruben.vp8510@gmail.com
          name: Ruben Vargas
        - email: p.loffay@gmail.com
          name: Pavol Loffay
        - email: bongartz@klimlive.de
          name: Benedikt Bongartz
        - email: andreas@gerstmayr.me
          name: Andreas Gerstmayr
        - email: iblancas@redhat.com
          name: Israel Blancas Alvarez
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Grafana Tempo Operator SIG
        relatedImages:
        - ghcr.io/grafana/tempo-operator/tempo-operator:v0.18.0
        - docker.io/grafana/tempo:2.8.2
        - docker.io/jaegertracing/jaeger-query:1.68.0
        - docker.io/grafana/tempo-query:2.8.2
        - quay.io/observatorium/api:main-2025-08-25-7489eb0
        - quay.io/observatorium/opa-openshift:main-2025-06-16-ecdeca0
        - quay.io/openshift/origin-oauth-proxy:4.14
        version: 0.18.0
      entries:
      - name: tempo-operator.v0.18.0
        version: 0.18.0
      - name: tempo-operator.v0.17.1
        version: 0.17.1
      - name: tempo-operator.v0.15.3
        version: 0.15.3
      - name: tempo-operator.v0.15.2
        version: 0.15.2
      - name: tempo-operator.v0.15.1
        version: 0.15.1
      - name: tempo-operator.v0.15.0
        version: 0.15.0
      - name: tempo-operator.v0.14.2
        version: 0.14.2
      - name: tempo-operator.v0.14.1
        version: 0.14.1
      - name: tempo-operator.v0.14.0
        version: 0.14.0
      - name: tempo-operator.v0.13.0
        version: 0.13.0
      - name: tempo-operator.v0.12.0
        version: 0.12.0
      - name: tempo-operator.v0.11.1
        version: 0.11.1
      - name: tempo-operator.v0.11.0
        version: 0.11.0
      - name: tempo-operator.v0.10.0
        version: 0.10.0
      - name: tempo-operator.v0.9.0
        version: 0.9.0
      - name: tempo-operator.v0.8.0
        version: 0.8.0
      - name: tempo-operator.v0.7.0
        version: 0.7.0
      - name: tempo-operator.v0.6.0
        version: 0.6.0
      - name: tempo-operator.v0.5.0
        version: 0.5.0
      - name: tempo-operator.v0.4.0
        version: 0.4.0
      - name: tempo-operator.v0.3.0
        version: 0.3.0
      - name: tempo-operator.v0.2.0
        version: 0.2.0
      - name: tempo-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: tempo-operator
    provider:
      name: Grafana Tempo Operator SIG
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-application-gateway-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-application-gateway-operator.v25.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMApplicationGateway",
                "metadata": {
                  "name": "iag-instance"
                },
                "spec": {
                  "configuration": [
                    {
                      "dataKey": "config",
                      "name": "test-config",
                      "type": "configmap"
                    },
                    {
                      "discoveryEndpoint": "https://isam.mmfa.ibm.com/mga/sps/oauth/oauth20/metadata/test",
                      "postData": [
                        {
                          "name": "redirect_uris",
                          "values": [
                            "https://isam.mmfa.ibm.com/pkmsoidc"
                          ]
                        },
                        {
                          "name": "client_name",
                          "value": "OperatorTest"
                        },
                        {
                          "name": "enforce_pkce",
                          "value": "false"
                        },
                        {
                          "name": "all_users_entitled",
                          "value": "true"
                        },
                        {
                          "name": "consent_action",
                          "value": "never_prompt"
                        }
                      ],
                      "secret": "oidc-client",
                      "type": "oidc_registration"
                    },
                    {
                      "headers": [
                        {
                          "name": "Authorization",
                          "secretKey": "value",
                          "type": "secret",
                          "value": "githubsecret"
                        }
                      ],
                      "type": "web",
                      "url": "https://raw.github.com/ibm-security/iag-config/master/test/sample1.yaml"
                    },
                    {
                      "type": "literal",
                      "value": "version: \"22.07\"\n\nserver:\n  local_applications:\n    cred_viewer:\n      path_segment: creds\n      enable_html: true\n\nidentity:\n  oidc:\n    discovery_endpoint: \"https://iag-dev.ibmcloudsecurity.com/oidc/endpoint/default/.well-known/openid-configuration\"\n    client_id: 1cbfe647-9e5f-4d99-8e05-8ed1c862eb47\n    client_secret: uPP8rM7N0e\n"
                    }
                  ],
                  "deployment": {
                    "image": "icr.io/ibmappgateway/ibm-application-gateway@sha256:1144b79a2930fe942ce627b0abcc3bed2f0b0770ae8313cb77ffe6772511c1f9",
                    "imagePullPolicy": "Always",
                    "imagePullSecrets": [
                      {
                        "name": "regcred"
                      }
                    ],
                    "lang": "C",
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 8,
                      "periodSeconds": 9,
                      "successThreshold": 7,
                      "timeoutSeconds": 1
                    },
                    "readinessProbe": {
                      "failureThreshold": 2,
                      "initialDelaySeconds": 7,
                      "periodSeconds": 8,
                      "successThreshold": 4,
                      "timeoutSeconds": 5
                    },
                    "serviceAccountName": "iag"
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/ibmappgateway/ibm-application-gateway-operator@sha256:6a16d4ef802a866942f504b11a948565245d303f322944200c1a1269fe7c371c
          createdAt: "2025-02-05T21:15:37Z"
          description: The IBM Application Gateway operator manages the lifecycle
            of IBM Application Gateway containers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM-Security/ibm-application-gateway-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMApplicationGateway is the Schema for the ibmapplicationgateways
              API
            displayName: IBMApplication Gateway
            kind: IBMApplicationGateway
            name: ibmapplicationgateways.ibm.com
            version: v1
        description: "The [IBM Application Gateway (IAG)](https://ibm.biz/ibm-app-gateway)
          image provides a containerized secure Web Reverse proxy which is designed
          to sit in front of your application, seamlessly adding authentication and
          authorization protection to your application.\n\nAn IAG instance deployed
          on Kubernetes can be a complex deployment.  In particular the configuration
          can be defined externally in one or more locations, and changes to this
          configuration may require all instances to be reloaded for the changes to
          take effect. \n\nThe internal Kubernetes deployment controller does not
          have any knowledge of how an IBM Application Gateway instance should behave
          when the configuration changes. As such this IBM Application Gateway specific
          Kubernetes operator is available to be deployed. Once deployed the operator
          can manage the lifecycle of the IBM Application Gateway instances.\n\nSee
          the project [Readme](https://github.com/IBM-Security/ibm-application-gateway-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Application Gateway Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Security
        - Identity and Access
        links:
        - name: IBM Application Gateway product documentation
          url: https://ibm.biz/ibm-app-gateway
        maintainers:
        - email: isamdev@au1.ibm.com
          name: IBM Application Gateway Development Team
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - icr.io/ibmappgateway/ibm-application-gateway-operator@sha256:6a16d4ef802a866942f504b11a948565245d303f322944200c1a1269fe7c371c
        - icr.io/ibmappgateway/ibm-application-gateway@sha256:1144b79a2930fe942ce627b0abcc3bed2f0b0770ae8313cb77ffe6772511c1f9
        version: 25.2.0
      entries:
      - name: ibm-application-gateway-operator.v25.2.0
        version: 25.2.0
      - name: ibm-application-gateway-operator.v24.4.0
        version: 24.4.0
      name: stable
    defaultChannel: stable
    packageName: ibm-application-gateway-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cockroach Labs
      provider-url: ""
    name: cockroachdb-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroach-operator.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v21.1.10",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cockroachdb/cockroachdb-operator@sha256:9fba7cd5fdf589317911925cb69338508bf9de928b1817c6f117c612635bbab3
          createdAt: "2021-10-19T15:47:03Z"
          description: CockroachDB Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:e67299fc0e79cff5b6caf9c0df154180dbaedc57b84f072acf68b7b2c958668d
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ec1672ce8091b677689a31cf09dbde5d1697f4caddae0151040b0fb156722e9d
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:e39bce1d9b08a7d62eb05d53e29fabc668bf6a4e4ae3eb9994eebf5b642ff6cf
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ab47ddd22328e0a1564814f049616c3e3569d24c0a290abee9308392c6a0de23
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:49421968126f9c98499bc0cf0c65729370ab566da3fbd1588c100b20052b972e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:c56f980b87a066d1e65ac8cff3d75040af9bd13093494cf4b4d30b5234d56edc
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:de94036e9cc8618e0493313512d880ae145e94d3bff54ed51b0de66a88da1319
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:3edda46eccce8350440a0ea78dad2e41a3eaa0838fe57945ad5318739e968516
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:0bd22d69db5ac570f30d3379afc67f6c2e46fc5b027c8aab42c3f562085d4672
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:bcf3f976437c4ee52e922aa5d7e382a86a031d780b711c08c6c158dbf2d26cfa
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:0ef0234b6fd42977b8a9eda2c59ca8194cc1c8cf6425b99d7bf8fde8feb826c5
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:87dcdbdc07904b217880e70484582c9d87dde7e4071e01cf9e2e6da43111190e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:2664d78fbe88c368927ddbf80704f430e6ca2ce1f72cb4b7e1906dfe72be7cd0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:fe0c615b4d1ef4ec6e089e85deae2b6bd85410be8ed525e2317c759669417f47
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1b56420753eac1126039b8ad4be82ddb59c89a4ec72fb7b6198cea88a10f3a91
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1ed18af1d3ffc65c24946c817d7d04d5f01025d3fb6c9eb78940e343b989bed0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:162a356cc8b423926634f83a1372b6321104d8de7a2c5b9b7b2681c769a88cc7
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:288ae92ebdfc848540ff80ef682b74e50809e9742cafce22b028112326d66b65
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:867c46bb4d8ca3f2fb8dc94063facc99758344c7b9d04f10e50932023ef5d262
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ba9da9aa9d662c0fa9fd16a40551eedf100869b008cf9c0c09cbb87a09fda697
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:25791b76b0d4b7220dadab3a405b3fdc330ba264ef6c9fbfdfb924c154c4bb5e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:cdab9fc62e07dd349703ae394adb48ab3d2281610df35e081f4b76e208b38c8e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:a5414e54c4c513c7877f925b09eadc4c28eb592d3e120bcf9c65ab10e07cd607
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:98b76cf63aabc39d026317905d6ba72eec3f0d35c0f770ebb8b789c3fa0709ee
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:9f5dfcde260804cd2d6976a6b3544ddabdd899365be2104210dc78ba13755ec3
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:27ef61d9f320a52ee9617f2ab8a1d1a1eacc576a0add2d85d99f9413912b651f
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:38e940144e34b69dfb7d1cf31456e7a0016c5dd40d5626baa346446c7678ceb0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1cea12847a0d236437f7fc28fba476fe512c343cd020d21bb50316e50da6ffd6
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:c1a07a8686bbadcb8c68ffba4e40e1285e581e24d43e5811f38a22d57fb9cc56
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:d290099496d070f135b5cc1e6bd59607512524d399df81698f11cfa09b8dca4a
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:b02549e5c3673b8154441990afa05a478d1ecc2bd0a70af64fd693cb6b469cb0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:5eb59093ee81f8115d82ec6a8a1c35eefad055cbb3a175978cef7afe9196e6a0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:fec6b0f56f2b94f141c8553a63b2e08b9494d4e6c5dc109d73268621ca445c56
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:d9a756df2ed2536b938d7c24fbf3de149cd1a0f44a3a1a0cc5d042fe8980362e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroachdb-operator@sha256:9fba7cd5fdf589317911925cb69338508bf9de928b1817c6f117c612635bbab3
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:6667919a41d304d5d4ade3ded4f11b42d722a995a4283e11e15320529f7f9abf
        version: 2.3.0
      entries:
      - name: cockroach-operator.v2.3.0
        version: 2.3.0
      name: beta
    - currentCSV: cockroach-operator.v2.18.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v25.2.2",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
          createdAt: "2025-07-22T11:30:56Z"
          description: CockroachDB Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9ab968e4ca62f23bf9848e5475ef4076df7318a8560f63d5b2090a32af6fe4e2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:90cf21c36d19d87b97d06edd1c0912a3bb288c769e6159651877db3206aa7355
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8f1f2fd1aca914a1ef007117c85ac944055bbfb873a468a7374d239ef20b065e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9b5a2e0c006eb4be8e1444734a98743551f278a24e81c20bebc96f0f3b631eb0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6371377718325235a2720dba0c97eb4321ee29b31a28cb893f76abd62baa339
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:68fe64fe3e112288ccfc2a7dffd69b9b2786a5da3d3bdedd6addf41659d460c9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5240f9f591263569fad7ee934e81fa8a1617de3c25da94ddd59ee5b2b3407f90
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f29c31bb267451b389608af3e8a4a28d6ca0691e67b77e8f57bb5e461b064893
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8737b67b91983817e3e4ff3f47f0b070bd36162d309b60993529426c6979cb24
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a3ac544d87a9d8e7c92897c4c23f803753c7ab31467db91b5c51244afc2c6be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf001173bbf00cdf443b8d64a3ad724389f1cf83e3628d0b609019c8cddad35a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce3d94a3fbdcde5858beb3c265987759971d991c397c6486e19904bd4706f44d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7d6320ac7fb0925f10cf6f524a13de629cbf13ccc371912e4fe8c777bc15d1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bd75aa918355e77cd671c3e11ee807861378e96245245f4fdf243cba5749f9a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d590946b2f7c7ec0c7ec057a338e81f864d6a9e3d3ac4e0769aa5756a8e13fc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:32fdca575c334822e4356aab36a7ed97b685c065925fe85f1b8ba8425c57159b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc4b75ed71845d5b18e1d9163d480fd9d0d3cb5ebbcfed00a2e4e1b174c0a5de
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f74c8a5dea4560ad59c6f6f6ab1e699f063334ee8db704460f870508dbeaeff0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4b3aa590623a38f435b075de4c09f8a343d8ece8b673e6bf7c6df5d5eb4bdf4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:46d9287ef11e3b21a36abfa31a26da3b93040bb6163b9223972a5dd1930fecff
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e0f21abea34825e55622c432a306f2e8aafd64d913eb78f2f601c51dc878a1da
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2b5c028f293c6f394e58a5495cb3dbfbedc215d0d03778d0217e045586d55a46
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08cec123813304ab03bd66d72a09949667a874a42320bedafda391ebcdb2e56c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6a2ee8d62c586eb74ebbef6bef256e06e1ba5cb51c1ae55b1654adafc2ca8859
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c2a62179e8ca293d7e30b9da599e48795345cf07a72abd3a83063107b53b03cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:72ea3d0ef72972dba93362dc31366c5946a9898f5c16780b877316a97068ef8f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1304d7f871a95c8b087bad1b548217562792edc66c8fe07c621e05242b8cd53e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9492894eaa5f43d00ac80b60029b32a0b0a85ae88ba609b46f98702c9349a5bb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bf36dab9b3257a4bba24d986f1451fc51a3a37023fbb988281f1941fd3faedd
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d467383af41aa80c26172964c34152abeba45121599804e502984655b72179f0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:022c6e1a0de875474bc953718a481a0af526c49dc910094dcdce0b5677aced18
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2d89398f03736f8ffaa82b10fed0f19fba166d3d2145b26f25cb71f3c39acf07
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e23af8f9d32c3f80a52d79ceb4b359dda1d75ba0dd4cb8ed4b66ac2f4e40c69b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:61e602b0e3582bf00460eb9328808ccf66e3ff1b1f6df07ee6c5c76a91846876
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8443c15eb1653dd9e5ec7aad54d4b8b092d41a0977e42b76e372e5e21280746e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0ec1dc184a2199736b4bd7de3241a357be7427caea5da1c931b08bc06f5c3dc0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c9a31bdf19175880d0981e21dd19a1458b6e8cde8d4acbcb554f3bea3578a64e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3f587e7b9f5a611013af2608e2151ed782859a52501b00adbc95739c56a8ff1e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7ad12df8e258b03c4099d375d02d272bd774bf188d108901d5cf6cb865e57ef3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4f753151af4c3d3f5a1df9014461316249a6d0621b8fba5c4430307b3355251
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:58de9d0b2ce005c7ad3e99152100e59b7677398557028a7a59e56e669fabb81c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1f6b1ec92251397dbcf7ed4a82f1f05619d64676aae2717f4eff1c33e4c2c27a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:266165ab87b19a05d6796e8b15ff0eba632d6ac96e6393400464c6136dd500ec
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:086d7b435993fdf21fd59b0093f52105a9028d6b769398b0033f98a65dfb7e79
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc1ab77a9537ccf236ce6dc370940dfedcfaa91b6e3db37bf2d5eade7ff84708
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4d8e804b35b73d7978c112748aea2d2bb68f65081b90bca62c39f2b6e2989378
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:744193037d6b64cb73fecab15ebe022abc2b79ab0335b76ec1415ac6379b7dfa
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5181196fff3bf8a5ec8fcfe4f1958cc69f285f0beb4f0b476b4b5e0912f3db42
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:40f00476b63cb9e370fe85f01407173693213a9242fc117d28b06378ca0d98e0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fd6c9b8ca82ff4f2d0fe29166d1fc9bca6439cb40907be60f4d8f1b9735de2f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:85a0e085901c2a73192556d5d70795d4b66cbcfe01d24c73b6fe6db8d5f2e5ca
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6bb15b36d64eebb6e4c1db5a5466e108b271d53383c58e0b6c78cec214756a9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b83eaff2e217c06098509709cead525b55b00ccd48019d8357e6231892008055
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:858dc82977b3c9e989a3a20626671b0c67c834c97132094a9212bb227ab55929
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da367cf0ac52045002e1c186f8e6964267ad87d5c25f9e72fd2c9b9a98a32702
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b77b80eed7ce149d2a3f228e5f188e76d73f32bc50ecde20bba9cfa7bdc4714c
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:53c7b960a2ed70a0998a4f8e28a4f78f7f676e01cffebe665e91e3a7629d88ed
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1558a1f5eab79f1a8c6c85d81740cc539dc23025ea457c0019d93a72dc05d57a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db4e5c2f27ba23e95a363c8174c01f0f9c6b6a27e0e581c86a38f3a1f68dc43c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:37bf9e1ec7e3eff47755fe7521223f31ec6aefc6e63901db3881b30898cf78f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:35ede65457d845e8c063ad7d9515d3a69f98e0751424ecbac484963678a0b1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:50393ad005fed19618d04f0716c6d55e2f1afce651e785c396638e852cac83b3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cad60044ad6573bd80b65e892a9ec0510dbe4fcbbfc6b51010ecf419f56f1024
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f42fa22aac3cc41ce72e144a72530ffbee765a392b24593dafd06cc098f8410b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:31b161f9bf9028a1a62cd2eef6abda01257d8889bdc3159822642549a75e3429
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9c85c25d5392817d17b68714dca6195ad53255d43e9c65b021d75b12d97e0579
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a6e2fe24ad52913e2160f85b9fe665a4e4c4d79b164b4b899f69dd9d92f999e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cc20270212f01fc231ff788e20b0476b8228c440bfc40430c0836d5eb702a042
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1374cbdc4ecf38e4b4226328f532e830b8932b34d9449bf2fbb36e60d0310521
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bd336715332f776a72aaf5c3ca52e62dd53a9e8521c1feb6e135cc868d694a39
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7481c0cf3b579c8d98d94d4aa0334ca94bf2ccfcd25ad70b57c42b22bbf4bcee
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4635550222775fc75737be5e5c0359303e3a23c4d1e1c296d7408cd15ad94aad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7daaf1368ab2c25694cdea0a598e1de59247b04a4b4b4b8348ed6edbaa194f9d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fe4a40006d1e8439423de1417cc9bc55716d790b882485074bb2a5ec68d7d1ad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5b3de46526d169fcd1318354545b3a18861843fa2022a23ed04d795493825287
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acc083f7dde1fbc557d858d03fc26b68b2056997fc1a457e4695650214d5b303
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0b0e0588c1aadf0c3826cd584bd2b3a7d6781882dcb5c5e037b47ff5cb84509b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3bf017b2b2506724500f5e29017d1b956825dd0f61a7c0049b0ff084c3554b5b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce1209ca6cc8724baee981c97c66e2852e512423bc0c0a16cdd80a6838b54031
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e82684b828256b3c2814930210031b3b70bc818514d4359a25f87ffdbf22be6e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:19f36f53f7da67755eb86da77098c5eeb84b381f9c18aa29ba573793dc498564
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f5a0ccc02dc9e938e484d5b5282ff650d1890d5f754c30a0c0d717989ed5d600
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9d0bfb642090ff55a3cc3da0f72663cae18063622a4ec785ba99bc71ef4db116
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f045fe5037d44aab1a0b2209241b3931ec7cb9897cb3c6a96a9f8d5280e55dd2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:44a0f40a6c46ba08297b482f3d96defe87b181c944803e5381c9acc42cabf97f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db07461824e73410456ef540b0bce3c7fe550aa67a73d6174300ca74b8f75276
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0754f7901feea9017bc5041cb91b7ac1cdbca10c2064f37c3648a2c395f24f83
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:28ae9661b38a72decf413c6a6c3758e61098debf5162197c27a12a070c7a096d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:743cda1eda6e273c798e3405a9a97e07dd6aa5ad4d5a52789be3b0f39a32a039
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:91f7e59f9e35a0395756eafed906963c535e886125294d671d930ce959a8f511
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d78c2e0709214f96794564f13386861f4b541e55b5bd316ad495cf104f7d40b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da3bd2e9ff2b08d5bb86b1f5faa95a865f85c5e8912df00b6b79190c1e518a2f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f3ffebd812881a830ff46c860984814dd61edcd51b478ebd3667759fb7710682
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5d6901b6b6fd6594d42bfc1842d050c7cffcb0867a5591a67469f0f860b1f21b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e8f26fd132affbaeb8ef6b1807a646c83b301e5dc334a592dfb2cf8dc3dbaf10
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c74cf927a2b2c5098ad74ad0880996f8b042186a9911d888a231ef5b9e2a0715
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5a54d564e08baefc2786f4b742b2507d31984664cd00333406aa7a3aa5b9c913
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:310310515625f099a928545865f7096997871ee4a16650a01c76c3799a18b684
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a0faa52a999289b308dc4caf050c6d9fe2f42e112aa30e4af282ed1f21c5c346
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a59cafff8d4ff2c62204ece15880117011c30978823c232902c7182d747ec827
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1e5d85f5db10fa79d09263db90daa44549bcc3b025e9f3f98cf0d4e51766394b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6ca927f137be28481e3a226e6c6bcb7c5ac27664984f17ffc6c1419cd7d08eb7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e50aab39722af22ccbf3d1db19e37972d52b0fc9a40998a64618dc966b2bac57
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d978678b3c254abd52f69f226b0aacc4b02c2aaca328c54ef10f6f9bb453582d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6408f89b0a88ec154f7ee98c314be574971c4c9aaf906b473d10124d822889be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:468b0a40a33be1dd5a2a89ce745d7e9012bc7371f2f01cc0f8f12c0ac887db43
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b06bd7f3802e8fa2af14bb0b07804cad4cc0fbbc97da18d09a9106bf074ed6d8
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f438fbbf9c87ca0d54c170e2c5105b1e265182e1c2ede48adea370e2be9dc895
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d8cfb221f67516675fdfde061f6ec5dfee1224ca7eb3d8be375c27d483f306af
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e151bded670f4cbd165a8ef864324805667d3e84ce3ace0cdf8f9aad087fccf1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cbc74889e9ca5edbd5aaa243d7649bca0c4da2719239f8d0ec9827868112a771
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a6f1c43486d044fed9f59406118f429c8037b389bd8c8d7a1c7aeb89ed2661f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2af60025ecb3bb933b61328be4c2b2bfd0e7d26f53b72430629208505e45c6d2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:79d688bebe0643ef0dd6ba7108cdb34c0442470c0c20ae0df88b5d02a3476cd0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9097e135159f46396ef4649f724d61bb5b9113c488f989d404586fbb7486fd85
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d14f77f70aea0422a906510c1ac2652ff76e131ed705cb002f80e13e887b0cd9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0fea6ed8a78ea624240ee6ffb3573d5cf1115186aa9180fcd7d1273351deaaa3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e18d73af3257c8cdf00245f2f718b56667524c1cc69d3db28e9b868a539ae16f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:dd521d7bad62da054eec54179be75c643a5492d981e11c0e66db41902cd2182c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:363d8feb560324a848c5771edd224dc19020013af7641979ee81ae8ee536d7e3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7b907d03b001b758d8a4b5e898725c73dda2ec2aa4b7425b685320b947154d11
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4b56c0311af14220a466aa11f25fd10fa38280825539bd02215c945bbafb861c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:42c9930d6861f6a2147ed6e01827eddd730279966763ec68868567a19cc164ab
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:740ba10f1ef0a6c5fa6ad31ab18855fc75f8e024ce4c23b0676c5a9c5b62cd75
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8e8c081c231b5b15eca483ceec16ea9190ba73aae6e421865162446eb584519a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:56109e57ee0379cf48644bcf8226a9238e01139cacc5499002c99f973f121911
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:742a6f69a3a209a7080a1e911fd96467d2a91aacb3f91d6be80a6a6e2e54b2ae
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0e2c7023f1f97acc20c46a4944635a9c5f57bb9bb3dd534cf41fea1a9cfdfdbe
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf7d3bea47638b763b312c7c11ca034a7ac62d993d75a13edceec74799b97e52
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:226d7f1c3f0d70ac6532ad9490bebcdef44a40959b6aa7c11b0f2d1383a46217
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6cd9f9b9fcb6e8f9a931a6ad01cb66329a74d40b22f9029fbc50e57248bad7cb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1ce71e80c0cd0e0063eba275ab91f77982f2e92a8aef8e89cab99b93feab0a88
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1b6d934a7b6c42a61e5ad51380e69af341a575b42205299fd58bc0c6ff306f6a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b046aa0a42dae992a03685ed82a20835532d6d4bf048e2366ff7382be4b3138
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acbfee4492327a720fe7da76929a86192d8af2442bbb76f0a089f6a81a890be3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:818f592c50d4ed5599faa053344f7ecd963003352f1faa3b72dc0f5e00a0e78b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4e5f7df1dc1e1db398c36d590431e7e5782897b209972d8e9e4671971c10d1b6
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5458a8fd89aa664b60e59c35bd933975cd37d18e19b11589e4ca0c2844ea0ed4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2a128b570352a5a7f82b27353f9b774caccce83e0e7bb2673e4802d96ad0bd51
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2171e1c75d0c665fe33e1ea334e605604c51de6dbc22d88a3a0ac604259c448f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ec6eb7c28c213cc83b2d7919cd87988f9a07f12276eb7351d0915f1567a5b095
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e5814748c5ab229a1dea5662a007063c0df06fbbfcfe041fd9a6df9ef67e07cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:82d8104a37349c513447e68ea3790efb256370632353176e5fc13c2e8df1bd44
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b78a4d3864e16d270979d9d7756012abf09e6cdb7f1eb4832f6497c26281099
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9f42e929f2cc7e9b6d24dd61e19c5d93dcfeabd34b9c8b935afbe78577efb18b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8fc8de4f3671a73f9aa747e5b23ec973d192fff6f2869c38985cd656063c1e3d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:02e1ef01a023da0eee653f63b314786fd66a78d184a97a314422f6f9c3ef8e7e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:65923c37fecda2d3174212fa9d9a6f34241d065389336ea69a7be99ba16cedc2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bed27f3f2c2d7627d1af4500d7ea0302da5c3fd3d875a6480fb6d98da481312e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:10b9df672260c023dfb37166e156c0d7e0e13144dc3e485af922e8151fd3ab05
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f05824bb8eaa5a93e54059449eaa236dd83d729049ad00065ef3242f5f74c91e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:87db6e079b96fb392d98ef5fc7063836bff10717efd75061016ce688be0ba38e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:29abb169ffefe36f6d69edd438a7135ee23059cedab8f87ad3b2d52361b96b35
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a4f932c3e9ff3aeb70ba1f7a92aa4642bd14cbb7d37d04ff750ed517b916cdb9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08496b84b3cc691909227be17e67bcd6f4d5185a93712f44d8a0985bd2c5fcd7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f53b07212320c6939d3bb4ddd4e6a0bce311ab5b63858eb1a8728531f0dbd510
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b0057de903654ea0315a2c96affcc8f50a92ede8ad121ed30edb3aaf0fb7b554
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d0014b1c19890b9e8ef1a4f3efe7eb5b269be5610ba6ababd4258b52d2f85e8a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:aeac9cd8e8acdbd2b0c339c48b85079c159b04f0ade24df8e3f9cae3f48b4c70
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6366ba29f301a82215c05e60d78e1de66bcee85440ce2fc3c692712d4da32685
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8a70ba12f610ac0e62066e417a0833bc1d01e91ac99d0dd794a2c0cd2a350164
        version: 2.18.2
      entries:
      - name: cockroach-operator.v2.18.2
        version: 2.18.2
      name: stable
    defaultChannel: stable
    packageName: cockroachdb-certified-rhmp
    provider:
      name: Cockroach Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      control-plane: odfmo-controller-manager
      full_version: 4.15.4-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-compose-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-compose-operator.v4.15.4-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "labels": {
                    "control-plane": "odfmo-controller-manager"
                  },
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              }
            ]
          capabilities: Basic Install
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.4-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          This is a CSV used for mirroring the entire product properly. Do not install it please.
        displayName: ODF Compose (DO NOT INSTALL)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: branto@redhat.com
          name: Boris Ranto
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-operator-bundle@sha256:3773158dbd141530cd851f36c3c8a5f54dc0892d003633118e123f5862923666
        - registry.redhat.io/odf4/odr-cluster-operator-bundle@sha256:4c7049b6137b3d49fa9b5e51e0cf7cd430d70c8790e0ed017e6bd811fcc506c8
        - registry.redhat.io/odf4/odf-csi-addons-operator-bundle@sha256:b2b8144c34df667b13987773b11df2fc8ac96d418999292303ef2c03cd3c79d6
        - registry.redhat.io/odf4/mcg-operator-bundle@sha256:9ce0ab75f56ebeeac63aefd0f48cba671da3377ac628830b3c03be9569f8c87a
        - registry.redhat.io/odf4/ocs-operator-bundle@sha256:1e91c5c6c122fdccab52c37ff2cef76317af80970dbdc06aec752115117b2fa0
        - registry.redhat.io/odf4/ocs-client-operator-bundle@sha256:8904c171ea317249e9a19c3a7638f688357a0b238582330b1bddc23a11a9925e
        - registry.redhat.io/odf4/odf-multicluster-operator-bundle@sha256:f7e1c75ee6b1a62fddbcc1e6a718a9cbd5ecf364e4ef10f29da7a15bc82d5b17
        - registry.redhat.io/odf4/odf-multicluster-rhel8-operator@sha256:861e56167698558109197b515aff93691f8d0fdba19311be18a32512f6896139
        - registry.redhat.io/odf4/odr-hub-operator-bundle@sha256:862a9d983013c16b7dfd5194ab5752c26a473ca5a4575bbcd80be59f97f8ac27
        version: 4.15.4-rhodf
      entries:
      - name: odf-compose-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      name: stable-4.15
    defaultChannel: stable-4.15
    packageName: odf-compose-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: unsupported
      operatorframework.io/arch.ppc64le: unsupported
      operatorframework.io/arch.s390x: unsupported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: mta-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mta-operator.v6.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:f5d10cd2528bebf5094c7744e47d9de6fe827fd09024c5ffc0e1b74f1a6cc3b4
          createdAt: "2023-04-26T14:04:20Z"
          description: MTA contains tools that support the modernization and migration
            of applications to Kubernetes and OpenShift
          olm.skipRange: '>=0.0.0 <6.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/tackle2-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is a collection of tools that supports large-scale application modernization and migration projects to Kubernetes and OpenShift.


          Migration Toolkit for Applications allows users to maintain their portfolio of applications with a full set of metadata and to assess their suitability for modernization leveraging a questionnaire based approach.


          ### Install

          By default, MTA requires a RWX capable storage class for caching purposes, this can be disabled via the rwx_supported parameter in Tackle CR.

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-windup-addon-rhel8@sha256:a4200fc0a39df67eae31c53ee9ae0c05769fb410d0312ebd0bbf8bf1567b2a48
        - registry.redhat.io/mta/mta-hub-rhel8@sha256:c75e1d9d3307d95be392a53440d9123601a6ada3693e3f2573e03022565e4339
        - registry.redhat.io/rhel8/postgresql-12@sha256:3a69efa85af6a6f95665eca69085b007fa1fcb645a37f356895aeb4c9527a621
        - registry.redhat.io/mta/mta-pathfinder-rhel8@sha256:a66dd030eb99207b6aaacc6b7c7cb5182901da836ebf50373c14f8e5ecb7b642
        - registry.redhat.io/mta/mta-ui-rhel8@sha256:de6cb95ce3663922ecc725b186876b043a9c7794e95510459ffbcba7d668fc26
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:f5d10cd2528bebf5094c7744e47d9de6fe827fd09024c5ffc0e1b74f1a6cc3b4
        - registry.redhat.io/mta/mta-admin-addon-rhel8@sha256:812b8dc621a06bb99cf2c59abb8257da3ceeb165c563e14d258e69803a557cc8
        version: 6.0.3
      entries:
      - name: mta-operator.v6.0.3
        version: 6.0.3
      - name: mta-operator.v6.0.2
        version: 6.0.2
      - name: mta-operator.v6.0.1
        version: 6.0.1
      - name: mta-operator.v6.0.0
        version: 6.0.0
      name: stable-v6.0
    - currentCSV: mta-operator.v6.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:6da71d4caecfbe049e63576d4325d4919002196da2e875f79703d8ff0d5a1705
          createdAt: "2023-10-26T15:29:26Z"
          description: MTA contains tools that support the modernization and migration
            of applications to Kubernetes and OpenShift
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0 <6.1.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/tackle2-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is a collection of tools that supports large-scale application modernization and migration projects to Kubernetes and OpenShift.


          Migration Toolkit for Applications allows users to maintain their portfolio of applications with a full set of metadata and to assess their suitability for modernization leveraging a questionnaire based approach.


          ### Install

          By default, MTA requires a RWX capable storage class for caching purposes, this can be disabled via the rwx_supported parameter in Tackle CR.

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:6da71d4caecfbe049e63576d4325d4919002196da2e875f79703d8ff0d5a1705
        - registry.redhat.io/mta/mta-hub-rhel8@sha256:32624f2bddfe8b6b9c6c1c3f2be541c8c61be29680d13b6d446e8874bdf8e219
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/mta/mta-pathfinder-rhel8@sha256:1df881da1e20c2501caf53e1cf59b6b366281d842f0948ca5f91694ec1db829d
        - registry.redhat.io/mta/mta-ui-rhel8@sha256:3845d6d246003ebeb49bcf1d95d6be2eb9cd05a280a5bf8320b201e878a41427
        - registry.redhat.io/mta/mta-windup-addon-rhel8@sha256:6bcaf8e8b63e169652d6568d77e91a289084e621a586105a2d429b5726569058
        version: 6.1.4
      entries:
      - name: mta-operator.v6.1.4
        version: 6.1.4
      - name: mta-operator.v6.1.3
        version: 6.1.3
      - name: mta-operator.v6.1.2
        version: 6.1.2
      - name: mta-operator.v6.1.1
        version: 6.1.1
      - name: mta-operator.v6.1.0
        version: 6.1.0
      name: stable-v6.1
    - currentCSV: mta-operator.v6.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:6938c3fc6dd8fb032e1765fc11d9586dbfd1aaba4b423b0d87da4120ee0bb60d
          createdAt: "2024-06-10T22:12:58Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0 <6.2.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-windup-addon-rhel9@sha256:d1f4939b92dd13c15cb7344ec47a571669f6d108dc2e7d795ce29350f2cb3838
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:6938c3fc6dd8fb032e1765fc11d9586dbfd1aaba4b423b0d87da4120ee0bb60d
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:e59725832547e80cab58077a1b6e42c313a4057621defa841f35c0a7627f149d
        - registry.redhat.io/rhel8/postgresql-12@sha256:c4716bee5e2cf2e43698bc99e5aa42f16a0f2d68c1fbdcd3602696bf88e408dc
        - registry.redhat.io/mta/mta-pathfinder-rhel9@sha256:a2a8b6d204d99494a6a005e8cb1fbd92afdc84dc01e4c1de4ad7d8798bd83741
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:c35f3ed429a7d351cb8c2d07600b4ad35615a5ba08e5cc9f8b9bd504aea3f7ee
        version: 6.2.3
      entries:
      - name: mta-operator.v6.2.3
        version: 6.2.3
      - name: mta-operator.v6.2.2
        version: 6.2.2
      - name: mta-operator.v6.2.1
        version: 6.2.1
      - name: mta-operator.v6.2.0
        version: 6.2.0
      name: stable-v6.2
    - currentCSV: mta-operator.v7.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:268ecf5efc51af71f9b2bc2d01b410f8cfcf81a948b82d2d476fef6c0cd3e033
          createdAt: "2024-05-10T19:37:30Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:268ecf5efc51af71f9b2bc2d01b410f8cfcf81a948b82d2d476fef6c0cd3e033
        - registry.redhat.io/mta/mta-windup-shim-rhel9@sha256:e2ea89055936d7ad229b5e890c6ba234c6cb22d8b688515ed4d6a34848efea83
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:d554cc36748fc05e30eb10e6fe816ff6aab2fe07b61cc60e9d93de5c3cf17748
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:0e1ec99af5666a060176adc52612f20209e3398a9a5a2aa1dcff11ef260b2006
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:1b5e016046d93b6a2674d63f08e0a46f8050e785bd58fd94ba7a70487efe7e2a
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:18c611415589a093e16e9c2f2fba9f3c83d9514f79c8e4fde9704ef0679ea4d5
        - registry.redhat.io/mta/mta-analyzer-lsp-rhel9@sha256:0c8f59f35a950ea8b45f16cf71362128ccc335b58a996a099384fd737dd6ff90
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:af879d8b9b5e396a5308fb675a1d6dba6d70151b9bdbe7829362e081a6adab1c
        - registry.redhat.io/rhel8/postgresql-12@sha256:da619f6617d99790b4fe00fdb23ab4aea0cdb0592590162d51b3ef94aca01a7b
        version: 7.0.3
      entries:
      - name: mta-operator.v7.0.3
        version: 7.0.3
      - name: mta-operator.v7.0.2
        version: 7.0.2
      - name: mta-operator.v7.0.1
        version: 7.0.1
      - name: mta-operator.v7.0.0
        version: 7.0.0
      name: stable-v7.0
    - currentCSV: mta-operator.v7.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:08b33b44140397785f8aa1b736796e09cf6e49ac30cf7c386a2fbd8b678d47b1
          createdAt: "2024-09-16T23:45:42Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.1.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:16cb28d683159ba71a5d7c233404ac1d6b7a07a374b46d24effe551a13a596c3
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:23c97ac122eb7cd1e953df8da968da420f6db8708baf772759517aca4929748b
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:e4e70201df23c9e1168646d3c6013a8ffa22481a85b07df491f240310c13ebe1
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:08b33b44140397785f8aa1b736796e09cf6e49ac30cf7c386a2fbd8b678d47b1
        - registry.redhat.io/rhel8/postgresql-15@sha256:16faf1ee82382263408f713ee4002f47ff13c82d4545e8a69d5b0ebd1d289330
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:30a3b1d5a70dec74a9253ca34f18bd8f2cdfaa68b85b0b25dc395158c2596330
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel8@sha256:582f28e609c1a0466d30c5586d6b6023f1b1de3c28f76cabb498322914ab24d4
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:ac38f1c3d34bc718928eb20743835c99783d846f63a3197d13839ea22f1634c8
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:4ad19b50cdd6be21e010155f2026abf0dc865e72aac1353e35bbb5eeb8320a4e
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:ee2c8ce2af0be097353758d9ee7eaa7e6b3fcd5f2bf5e04280e450e2b619d2d8
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        version: 7.1.1
      entries:
      - name: mta-operator.v7.1.1
        version: 7.1.1
      - name: mta-operator.v7.1.0
        version: 7.1.0
      name: stable-v7.1
    - currentCSV: mta-operator.v7.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:917d6cd534c158f6c4a7e9c83f0c66fa85ba81037c352d8a039e198581126a7f
          createdAt: "2025-06-17T16:04:15Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.2.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:917d6cd534c158f6c4a7e9c83f0c66fa85ba81037c352d8a039e198581126a7f
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel8@sha256:c0643c05833dd1b70ed361bf56c46f56ed6536aac3106b92f0d85b83da1b9d15
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:10ef4b218d603432508b5e1f3685fb32a80cbce6c327e35952cdcccb0465468b
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:9d6c02f6463771e747a30126de95fc0728fd95fff4ddfa6c70d828c55e0b6790
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:2dfc51d5398a90e73273b3e332f70d7ed94bac05e12b142fba22104ecc929436
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:b10fdc78c6f453a34a3d306313dc29fcb3152688a6a1d54cd236d00d9fab88af
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:4cde2c2df482293bd4d6b86a344336e13d2f9f81b29323698f603e656b14b190
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:70ff3b5e429a8fe0c68752ee05c47ead885b59bf9c8d7235337801270e7fc1de
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:c9447659195b9eebb8f8a265fd5052ca9b2611d57080017448297a33409863c6
        - registry.redhat.io/rhel8/postgresql-15@sha256:dd4aec2775da1cf187c0839384cb16d69268e9781d4330874bbc865f98153a7a
        version: 7.2.3
      entries:
      - name: mta-operator.v7.2.3
        version: 7.2.3
      - name: mta-operator.v7.2.2
        version: 7.2.2
      - name: mta-operator.v7.2.1
        version: 7.2.1
      - name: mta-operator.v7.2.0
        version: 7.2.0
      name: stable-v7.2
    - currentCSV: mta-operator.v7.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel9-operator@sha256:6abd6541a955543207c1ff4b903928debcc89dd6ea3d7aeaa3e47465849f7b37
          createdAt: "2025-08-15T16:32:16Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.3.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-15@sha256:854710b428c4c1132c448cbe9ad30087a8a4b3c33ae9c0c3a4de21a15c0ae642
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel9@sha256:9748d954ad8654ac203257286497ccf19cfdfa211fee40449cbe732ff48c1176
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:32328fc0459099ef1de0f92e064c15d29a46d88dfe2f54c88f822d44ab82e3f4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        - registry.redhat.io/mta/mta-rhel9-operator@sha256:6abd6541a955543207c1ff4b903928debcc89dd6ea3d7aeaa3e47465849f7b37
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:b2ff154a4b4ead22311005b7115c02346928b5ebf77d776ef909ad4af1f350bc
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:6118c1db15667851f26eb2acf6060e045ed9065459e3e649bbdd0e2625d825bd
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:072f676323e9abcb2ff22eb5e7a70f6681e9d885d8666ddd81060fbad54c443e
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:8ca00460b2fb64583a247f5b906c19743b7732b60ff8215f1baec12a766c61d1
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:80aa13813fe121da49e166d81d09d63749393f3375c3b96d762d0b09c6685ca9
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:ef57a58e609e244333d35681a164a1d3bf382c0eee37d33c993b018439814198
        version: 7.3.2
      entries:
      - name: mta-operator.v7.3.2
        version: 7.3.2
      - name: mta-operator.v7.3.1
        version: 7.3.1
      - name: mta-operator.v7.3.0
        version: 7.3.0
      name: stable-v7.3
    - currentCSV: mta-operator.v8.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel9-operator@sha256:83c2b804667d28f782d64132a0c393194083d10c820b88df49c2503d3351fd8b
          createdAt: "2025-09-26T03:03:31Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.3.0 <8.0.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Schema
            displayName: Schema
            kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-platform-addon-rhel9@sha256:dcb35178810b7eddc2f2713434d377003f8f4fb7022a40f6c1926b74b13fe23a
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:a77ee9b76225b6cca667fa3402492b0845587fe5e400fa8d141b5ab636c56811
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel9@sha256:2a6f0b392b86fe16244d5bf9fcb84ed9770a22329a96e9861f03872b7209fb04
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ef10b2ecbfa0c4ac8568af9ee64ee4a554e3fdc9089ee5683b6c2c3a74caafbe
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:616d04cfbab92341157c42eee0683f31867d81c5a289723da8da8208d1054774
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:aaf7bf10b1af107ad55289caab2cf9dd052e69211968d2e8d05c1d25f21ed2b3
        - registry.redhat.io/mta/mta-rhel9-operator@sha256:83c2b804667d28f782d64132a0c393194083d10c820b88df49c2503d3351fd8b
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:b1b38e218ea37b529f344eaa1f0f3b55b716ad7953e852c19f34f4a25ab6c5a7
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:af2c1790f2b98ebcfadb5d6bdd43cb9e7e96cdabd5b653d48d0fb67901781c49
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:460d1d45ed44ae9328ed93c5fed240e154549bbec662dc1fc55569ff7256aca9
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:8ae0887fa27d6f4281e5bb9b629bf6b64a9c460948a002839ba42da9b25b5521
        - registry.redhat.io/mta/mta-solution-server-rhel9@sha256:e9d48d11ae203108229d67bc96721daa1c3bb10160b9d6d957cd2e0ba68a2ef7
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        version: 8.0.0
      entries:
      - name: mta-operator.v8.0.0
        version: 8.0.0
      name: stable-v8.0
    defaultChannel: stable-v8.0
    packageName: mta-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: aws-efs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-efs-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"aws-efs.managed.openshift.io/v1alpha1","kind":"SharedVolume","metadata":{"name":"sv1"},"spec":{"accessPointID":
            "fsap-0123456789abcdef", "fileSystemID": "fs-0123cdef"}}]'
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/app-sre/aws-efs-operator:9c87267
          createdAt: "2021-12-20T19:51:50Z"
          description: Manage read-write-many access to AWS EFS volumes in an OpenShift
            cluster.
          repository: https://github.com/openshift/aws-efs-operator
          support: https://github.com/openshift/aws-efs-operator/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Shared Volume
            displayName: SharedVolume
            kind: SharedVolume
            name: sharedvolumes.aws-efs.managed.openshift.io
            version: v1alpha1
        description: |-
          ## About the managed application
          AWS provides Elastic File System, an NFS-based shared storage solution. A couple of mechanisms exist to enable mounting such a file system in containers. Curating these options and making them work in a cluster is nontrivial.
          ## About this Operator
          The aws-efs-operator minimizes the barrier to entry for container use of EFS by:
          - Installing the [AWS EFS CSI Driver](https://github.com/kubernetes-sigs/aws-efs-csi-driver) and supporting infrastructure.
          - Reducing the identification of a shared storage resource to a simple Custom Resource Definition with only two fields.
          - Creating the necessary PersistentVolume and PersistentVolumeClaim resources that can simply be referenced from a Pod specification.

          More information can be found in the operator's [source repository](https://github.com/openshift/aws-efs-operator).
          ## Prerequisites for enabling this Operator
          An OpenShift cluster backed by AWS.
        displayName: AWS EFS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AWS
        - EFS
        - NFS
        - Shared storage
        links:
        - name: README
          url: https://github.com/openshift/aws-efs-operator/blob/master/README.md
        - name: Design
          url: https://github.com/openshift/aws-efs-operator/blob/master/DESIGN.md
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/app-sre/aws-efs-operator:9c87267
        version: 0.0.8
      entries:
      - name: aws-efs-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: aws-efs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: opdev.github.io
    name: spark-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spark-helm-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apachespark.apache.org/v1alpha1",
                "kind": "Spark",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/install-disable-crds": "true"
                  },
                  "name": "spark-operator"
                },
                "spec": {
                  "affinity": {},
                  "batchScheduler": {
                    "enable": false
                  },
                  "controllerThreads": 10,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/googlecloudplatform/spark-operator",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "ingressUrlFormat": "",
                  "istio": {
                    "enabled": false
                  },
                  "labelSelectorFilter": "",
                  "leaderElection": {
                    "lockName": "spark-operator-lock",
                    "lockNamespace": ""
                  },
                  "logLevel": 2,
                  "metrics": {
                    "enable": true,
                    "endpoint": "/metrics",
                    "port": 10254,
                    "portName": "metrics",
                    "prefix": ""
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podMonitor": {
                    "enable": false,
                    "jobLabel": "spark-operator-podmonitor",
                    "labels": {},
                    "podMetricsEndpoint": {
                      "interval": "5s",
                      "scheme": "http"
                    }
                  },
                  "podSecurityContext": {},
                  "rbac": {
                    "create": false,
                    "createClusterRole": true,
                    "createRole": true
                  },
                  "replicaCount": 1,
                  "resourceQuotaEnforcement": {
                    "enable": false
                  },
                  "resources": {},
                  "resyncInterval": 30,
                  "securityContext": {},
                  "serviceAccounts": {
                    "spark": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "sparkoperator": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    }
                  },
                  "sparkJobNamespace": "",
                  "tolerations": [],
                  "uiService": {
                    "enable": true
                  },
                  "volumeMounts": [],
                  "volumes": [],
                  "webhook": {
                    "cleanupAnnotations": {
                      "helm.sh/hook": "pre-delete, pre-upgrade",
                      "helm.sh/hook-delete-policy": "hook-succeeded"
                    },
                    "enable": false,
                    "initAnnotations": {
                      "helm.sh/hook": "pre-install, pre-upgrade",
                      "helm.sh/hook-weight": "50"
                    },
                    "namespaceSelector": "",
                    "port": 8080,
                    "timeout": 30
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "ScheduledSparkApplication",
                "metadata": {
                  "name": "spark-pi-scheduled",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "concurrencyPolicy": "Allow",
                  "schedule": "@every 5m",
                  "template": {
                    "driver": {
                      "coreLimit": "1200m",
                      "cores": 1,
                      "labels": {
                        "version": "3.1.1"
                      },
                      "memory": "512m",
                      "serviceAccount": "spark-operator-spark",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "executor": {
                      "cores": 1,
                      "instances": 1,
                      "labels": {
                        "version": "3.1.1"
                      },
                      "memory": "512m",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "image": "gcr.io/spark-operator/spark:v3.1.1",
                    "imagePullPolicy": "Always",
                    "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.12-3.1.1.jar",
                    "mainClass": "org.apache.spark.examples.SparkPi",
                    "mode": "cluster",
                    "restartPolicy": {
                      "type": "Never"
                    },
                    "sparkVersion": "3.1.1",
                    "type": "Scala",
                    "volumes": [
                      {
                        "hostPath": {
                          "path": "/tmp",
                          "type": "Directory"
                        },
                        "name": "test-volume"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "pyspark-pi",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "driver": {
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "serviceAccount": "spark-operator-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "gcr.io/spark-operator/spark-py:v3.1.1",
                  "imagePullPolicy": "Always",
                  "mainApplicationFile": "local:///opt/spark/examples/src/main/python/pi.py",
                  "mode": "cluster",
                  "pythonVersion": "3",
                  "restartPolicy": {
                    "onFailureRetries": 3,
                    "onFailureRetryInterval": 10,
                    "onSubmissionFailureRetries": 5,
                    "onSubmissionFailureRetryInterval": 20,
                    "type": "OnFailure"
                  },
                  "sparkVersion": "3.1.1",
                  "type": "Python",
                  "volumes": [
                    {
                      "hostPath": {
                        "path": "/tmp",
                        "type": "Directory"
                      },
                      "name": "test-volume"
                    }
                  ]
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "spark-pi",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "driver": {
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "serviceAccount": "spark-operator-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "gcr.io/spark-operator/spark:v3.1.1",
                  "imagePullPolicy": "Always",
                  "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.12-3.1.1.jar",
                  "mainClass": "org.apache.spark.examples.SparkPi",
                  "mode": "cluster",
                  "restartPolicy": {
                    "type": "Never"
                  },
                  "sparkVersion": "3.1.1",
                  "type": "Scala",
                  "volumes": [
                    {
                      "hostPath": {
                        "path": "/tmp",
                        "type": "Directory"
                      },
                      "name": "test-volume"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/spark-helm-operator-controller:0.0.2
          createdAt: "2023-07-14T17:57:44Z"
          description: An experimental operator that installs the spark on k8s operator
            from its helm chart.
          operatorframework.io/initialization-resource: "{\n  \"apiVersion\": \"apachespark.apache.org/v1alpha1\",\n
            \ \"kind\": \"Spark\",\n  \"metadata\": {\n    \"name\": \"spark-operator\"\n
            \ },\n  \"spec\": {\n    \"affinity\": {},\n    \"batchScheduler\": {\n
            \     \"enable\": false\n    },\n    \"controllerThreads\": 10,\n    \"fullnameOverride\":
            \"\",\n    \"image\": {\n      \"pullPolicy\": \"IfNotPresent\",\n      \"repository\":
            \"ghcr.io/googlecloudplatform/spark-operator\",\n      \"tag\": \"\"\n
            \   },\n    \"imagePullSecrets\": [],\n    \"ingressUrlFormat\": \"\",\n
            \   \"istio\": {\n      \"enabled\": false\n    },\n    \"labelSelectorFilter\":
            \"\",\n    \"leaderElection\": {\n      \"lockName\": \"spark-operator-lock\",\n
            \     \"lockNamespace\": \"\"\n    },\n    \"logLevel\": 2,\n    \"metrics\":
            {\n      \"enable\": true,\n      \"endpoint\": \"/metrics\",\n      \"port\":
            10254,\n      \"portName\": \"metrics\",\n      \"prefix\": \"\"\n    },\n
            \   \"nameOverride\": \"\",\n    \"nodeSelector\": {},\n    \"podAnnotations\":
            {},\n    \"podLabels\": {},\n    \"podMonitor\": {\n      \"enable\":
            false,\n      \"jobLabel\": \"spark-operator-podmonitor\",\n      \"labels\":
            {},\n      \"podMetricsEndpoint\": {\n        \"interval\": \"5s\",\n
            \       \"scheme\": \"http\"\n      }\n    },\n    \"podSecurityContext\":
            {},\n    \"rbac\": {\n      \"create\": false,\n      \"createClusterRole\":
            true,\n      \"createRole\": true\n    },\n    \"replicaCount\": 1,\n
            \   \"resourceQuotaEnforcement\": {\n      \"enable\": false\n    },\n
            \   \"resources\": {},\n    \"resyncInterval\": 30,\n    \"securityContext\":
            {},\n    \"serviceAccounts\": {\n      \"spark\": {\n        \"annotations\":
            {},\n        \"create\": true,\n        \"name\": \"\"\n      },\n      \"sparkoperator\":
            {\n        \"annotations\": {},\n        \"create\": true,\n        \"name\":
            \"\"\n      }\n    },\n    \"sparkJobNamespace\": \"\",\n    \"tolerations\":
            [],\n    \"uiService\": {\n      \"enable\": true\n    },\n    \"volumeMounts\":
            [],\n    \"volumes\": [],\n    \"webhook\": {\n      \"cleanupAnnotations\":
            {\n        \"helm.sh/hook\": \"pre-delete, pre-upgrade\",\n        \"helm.sh/hook-delete-policy\":
            \"hook-succeeded\"\n      },\n      \"enable\": false,\n      \"initAnnotations\":
            {\n        \"helm.sh/hook\": \"pre-install, pre-upgrade\",\n        \"helm.sh/hook-weight\":
            \"50\"\n      },\n      \"namespaceSelector\": \"\",\n      \"port\":
            8080,\n      \"timeout\": 30\n    }\n  }\n}    "
          operatorframework.io/suggested-namespace: spark-operator
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys a scheduled spark application based on the spec provided
            displayName: Scheduled Spark Application
            kind: ScheduledSparkApplication
            name: scheduledsparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - description: Deploys a spark application based on the spec provided
            displayName: Spark Application
            kind: SparkApplication
            name: sparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - description: Deploys the spark operator based on the google cloud platform
              helm chart
            displayName: Spark Operator
            kind: Spark
            name: sparks.apachespark.apache.org
            version: v1alpha1
        description: An experimental helm based operator that wraps the spark on k8s
          helm chart from google cloud platform enabling installation and workload
          management of Apache Spark
        displayName: Spark Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - big data
        - map reduce
        - data science
        links:
        - name: Spark Helm Operator
          url: https://github.com/skattoju/spark-helm-operator
        maintainers:
        - email: skattoju@redhat.com
          name: Sid Kattoju
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: opdev.github.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/opdev/spark-helm-operator-controller:0.0.2
        version: 0.0.2
      entries:
      - name: spark-helm-operator.v0.0.2
        version: 0.0.2
      - name: spark-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: spark-helm-operator
    provider:
      name: opdev
      url: opdev.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-test-core
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-test-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Testcore",
                "metadata": {
                  "name": "testcore-sample"
                },
                "spec": {
                  "agentController": {
                    "port": 8080
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "dataseqService": {
                    "ingressEnabled": true,
                    "port": 8080,
                    "providers": {
                      "loadCore": {
                        "enabled": true
                      }
                    }
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "ingress": {
                    "backwardsCompatibilityEnabled": true
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "systemMonitor": {
                    "enabled": false
                  },
                  "trafficController": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:35:34Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Testcore
            name: testcores.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Test Core Operator
        displayName: Keysight WAP Test Core Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Test Core
          url: https://keysight-wap-test-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-test-core@sha256:9e2d81061abff49112f908e71dace03a748e283f32597aa22ad58471fff30caf
        version: 0.1.21
      entries:
      - name: keysight-wap-test-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-test-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp
      provider-url: https://www.netapp.com/
    name: netapp-spark-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netapp-spark-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.netapp.com/v1alpha1",
                "kind": "SparkOperator",
                "metadata": {
                  "name": "sparkoperator-sample"
                },
                "spec": {
                  "spark-operator": {
                    "affinity": {},
                    "controllerThreads": 20,
                    "disableExecutorReporting": false,
                    "hostNetwork": false,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "public.ecr.aws/f4k1p1n4/spark-operator",
                      "tag": "108810d-ubi9-rel20240916"
                    },
                    "imagePullSecrets": [],
                    "istio": {
                      "enabled": true
                    },
                    "jobAffinity": {},
                    "podLabels": {},
                    "pprof": {
                      "enable": true,
                      "port": 6060,
                      "portName": "pprof"
                    },
                    "rbac": {
                      "createClusterRole": true,
                      "createRole": true
                    },
                    "resources": {
                      "requests": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      }
                    },
                    "securityContext": {},
                    "serviceAccounts": {
                      "spark": {
                        "create": true
                      }
                    },
                    "sparkJobNamespace": "",
                    "tolerations": [],
                    "uiService": {
                      "enable": false
                    },
                    "webhook": {
                      "enable": true,
                      "jobTTLSecondsAfterFinished": 60,
                      "port": 9443
                    },
                    "workqueue": {
                      "maxDelay": {
                        "enabled": false,
                        "value": "6h"
                      },
                      "tokenBucketSize": 500,
                      "tokenRefillRate": 50
                    }
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "ScheduledSparkApplication",
                "metadata": {
                  "name": "spark-pi-scheduled"
                },
                "spec": {
                  "concurrencyPolicy": "Allow",
                  "schedule": "@every 5m",
                  "template": {
                    "arguments": [
                      "1000"
                    ],
                    "driver": {
                      "annotations": {
                        "openshift.io/required-scc": "sparkoperator-sample-spark"
                      },
                      "coreLimit": "1200m",
                      "cores": 1,
                      "labels": {
                        "version": "3.5.0"
                      },
                      "memory": "512m",
                      "serviceAccount": "sparkoperator-sample-spark",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "executor": {
                      "annotations": {
                        "openshift.io/required-scc": "sparkoperator-sample-spark"
                      },
                      "cores": 1,
                      "instances": 1,
                      "labels": {
                        "version": "3.5.0"
                      },
                      "memory": "512m",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "image": "public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce",
                    "imagePullPolicy": "IfNotPresent",
                    "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.13-3.5.0.jar",
                    "mainClass": "org.apache.spark.examples.SparkPi",
                    "mode": "cluster",
                    "restartPolicy": {
                      "type": "Never"
                    },
                    "sparkVersion": "3.5.0",
                    "type": "Scala",
                    "volumes": [
                      {
                        "emptyDir": {},
                        "name": "test-volume"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "spark-pi"
                },
                "spec": {
                  "arguments": [
                    "1000"
                  ],
                  "driver": {
                    "annotations": {
                      "openshift.io/required-scc": "sparkoperator-sample-spark"
                    },
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.5.0"
                    },
                    "memory": "512m",
                    "serviceAccount": "sparkoperator-sample-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "annotations": {
                      "openshift.io/required-scc": "sparkoperator-sample-spark"
                    },
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.5.0"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce",
                  "imagePullPolicy": "IfNotPresent",
                  "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.13-3.5.0.jar",
                  "mainClass": "org.apache.spark.examples.SparkPi",
                  "mode": "cluster",
                  "restartPolicy": {
                    "type": "Never"
                  },
                  "sparkVersion": "3.5.0",
                  "type": "Scala",
                  "volumes": [
                    {
                      "emptyDir": {},
                      "name": "test-volume"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Big Data, Integration & Delivery, AI/Machine Learning, Streaming
            & Messaging
          com.redhat.openshift.versions: v4.14
          containerImage: public.ecr.aws/f4k1p1n4/netapp-spark-operator-controller@sha256:9fe1d0793f11b1c525a9a92cbcde6f4bc1bf3c887617a8496bed3a5a0008a67e
          createdAt: "2024-10-21T13:47:47Z"
          description: Red Hat and NetApp have partnered to offer enterprise-grade
            supported Apache Spark on OpenShift.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ScheduledSparkApplication
            name: scheduledsparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - kind: SparkApplication
            name: sparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - kind: SparkOperator
            name: sparkoperators.charts.netapp.com
            version: v1alpha1
        description: |-
          # NetApp Supported Apache Spark on OpenShift

          Red Hat and NetApp have partnered to offer enterprise-grade supported Apache Spark on OpenShift.
          Use this certified operator to launch Apache Spark applications on OpenShift.

          ## Documentation

          Please find the official documentation at https://www.instaclustr.com/support/documentation/apache-spark/apache-spark-on-red-hat-openshift/getting-started/

          ## Support

          With questions about Apache Spark, contact NetApp's Instaclustr support team on the [web](https://support.instaclustr.com/) or by email at support@instaclustr.com

          With questions about Operators in the OpenShift OperatorHub, contact Red Hat's support team on the [web](https://www.redhat.com/en/services/support) or by email at customerservice@redhat.com

          With questions about pricing, email ng-netapp-apache-spark-support-redhat-openshift@netapp.com
        displayName: NetApp Supported Apache Spark on OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Spark
        - Big Data
        - Apache Spark
        - AI
        - ML
        - Data Engineering
        links:
        - name: NetApp Supported Apache Spark on OpenShift
          url: https://www.instaclustr.com/support-solutions
        maintainers:
        - email: thorsteinn.sigurdsson@netapp.com
          name: Thorsteinn Sigurdsson
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: NetApp
          url: https://www.netapp.com/
        relatedImages:
        - public.ecr.aws/f4k1p1n4/netapp-spark-operator-controller@sha256:9fe1d0793f11b1c525a9a92cbcde6f4bc1bf3c887617a8496bed3a5a0008a67e
        - public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce
        version: 1.1.3
      entries:
      - name: netapp-spark-operator.v1.1.3
        version: 1.1.3
      name: alpha
    defaultChannel: alpha
    packageName: netapp-spark-operator
    provider:
      name: NetApp
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Trilio
      provider-url: ""
    name: k8s-triliovault
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k8s-triliovault-stable.3.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/trilio/k8s-triliovault-operator@sha256:0e3094945427231ecbb884b903dc4d347dfaa73e5a04eb4c5cbaed6fce1cbbb8
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          olm.skipRange: '>=3.0.0 <3.1.3'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          TrilioVault for Kubernetes is an enterprise-grade, cloud-native platform purpose built for data protection and management of Kubernetes applications for IT managers, administrators and developers. TrilioVault supports upstream Kubernetes and OpenShift environments and offers the following features:
            * Application-Centric - protects both data and metadata for Helm, Operator or custom Label based applications.
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any Storage (CSI, NFS, S3), or any Cloud (Private or Public).
            * Application Deployment & Tooling: Helm, Operators, Labels, Prometheus, Fluentd.
          Tutorials
          ------------
          Please click the link below to access the TrilioVault for Kubernetes "How-To" series for deployment, best practice and use-case videos
          <a href="https://www.trilio.io/triliovault-for-kubernetes">TrilioVault for Kubernetes - Tutorials and How-To </a>”
          Licensing
          --------
          Customers can download a 30-day Free Trial or a 10-node Basic Edition at no cost.  You can also connect with the Trilio team for an Enterprise Edition license with no limitations and Premium Support.
          For more information on license plans please vist:
          <a href="www.trilio.io/plans"> Trilio Vault for Kubernetes licensing and plans </a>
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our TrilioVault technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. www.trilio.io and @triliodata on Twitter.
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://www.trilio.io/
        - name: Product Licensing
          url: https://www.trilio.io/plans
        - name: Tutorials
          url: https://www.trilio.io/triliovault-for-kubernetes/
        - name: Support for Free/Basic Users
          url: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Trilio
        relatedImages:
        - registry.connect.redhat.com/trilio/k8s-triliovault-operator@sha256:0e3094945427231ecbb884b903dc4d347dfaa73e5a04eb4c5cbaed6fce1cbbb8
        - registry.connect.redhat.com/trilio/dex@sha256:b6331cf3cb3a6a66d9b9ea639cda3c2559fbf1f92f7174032352c1c7bb5176d9
        - registry.connect.redhat.com/trilio/event-stack@sha256:37c33ae1adddfe739c98fd2b2183c49ec7cfdafc88dec2970ad7c670016ebaac
        - registry.connect.redhat.com/trilio/minio@sha256:cee0f976edc79ff9e7a3f3574a7115dd960042241f3ed5b4636240ad609f9bca
        - registry.connect.redhat.com/trilio/web@sha256:76865d8afa8059d40810cb94df29420950a7593a40e87d142d771c57b3319e1d
        - registry.connect.redhat.com/trilio/trilio-webhook-init@sha256:c45e9b6b75862451d873d3eba29d6bd11cafc64823855f82ccb8c0b3d779ebb0
        - registry.connect.redhat.com/trilio/datamover@sha256:c7bc15f8217899ed73a812d278eb118633edbee1bf284132054b6f9ca2afd801
        - registry.connect.redhat.com/trilio/control-plane@sha256:01e36ed5b07104865f922d2996d2211b2b6000adc90266197d9bac74d908db97
        - registry.connect.redhat.com/trilio/nats@sha256:47396bfcff3e65efb3410fa8369f11f20d794ab543f9ada87b451448b1813178
        version: 3.1.3
      entries:
      - name: k8s-triliovault-stable.3.1.3
        version: 3.1.3
      - name: k8s-triliovault-stable.3.1.2
        version: 3.1.2
      - name: k8s-triliovault-stable.3.1.1
        version: 3.1.1
      - name: k8s-triliovault-stable.3.1.0
        version: 3.1.0
      name: 3.1.x
    - currentCSV: k8s-triliovault-stable.4.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:69871613ceecefff7f691f21bb0b9147236af5a5dc2c7b36ad5e0cd8551f0e6a
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <4.0.5'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/operator-webhook-init@sha256:aa85b787f7f37a71e909f049d2abdda746cdd0627e2f74aaf7476d5e5a7f4401
        - quay.io/triliodata/datamover@sha256:e6848d94ba9f81843bde882a030b0f2556aae4bf16d79e2c5b199417cf626c85
        - quay.io/triliodata/dex@sha256:b4b2372fd03be586b9ff8cd269b9001797d813a34155ab37de31a44fe3cab1ea
        - quay.io/triliodata/nats@sha256:e6723595bcc5674b6201091d45d711d29a5cd1421c8b97c76cf00e468379fddf
        - quay.io/triliodata/event-stack@sha256:c793418df5a1fc1a49f629d335f4684d9cde1a5f2b462f7c9483ffba904456d2
        - quay.io/triliodata/control-plane@sha256:53368a72a60a38d3ce4293236a841346da6da2380ffd2525e5ffdb8dbcf11c7f
        - quay.io/triliodata/web@sha256:20de1a2f9477ad238bf24f48c8c16a38521c3a86164f2eb6b10d20f09839016e
        - quay.io/triliodata/k8s-triliovault-operator@sha256:69871613ceecefff7f691f21bb0b9147236af5a5dc2c7b36ad5e0cd8551f0e6a
        version: 4.0.5
      entries:
      - name: k8s-triliovault-stable.4.0.5
        version: 4.0.5
      - name: k8s-triliovault-stable.4.0.4
        version: 4.0.4
      - name: k8s-triliovault-stable.4.0.3
        version: 4.0.3
      - name: k8s-triliovault-stable.4.0.2
        version: 4.0.2
      - name: k8s-triliovault-stable.4.0.1
        version: 4.0.1
      - name: k8s-triliovault-stable.4.0.0
        version: 4.0.0
      name: 4.0.x
    - currentCSV: k8s-triliovault-stable.5.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:8d582842dbef1c64bbe3b0ce5eeb64790910bdc0416c5e5561557642dede7022
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <5.0.4'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/datamover@sha256:104bdca9513c78f81c4340e9a3219df0a3bc8a5b026e4a252363a7e2c10977ba
        - quay.io/triliodata/dex@sha256:b4b2372fd03be586b9ff8cd269b9001797d813a34155ab37de31a44fe3cab1ea
        - quay.io/triliodata/k8s-triliovault-operator@sha256:8d582842dbef1c64bbe3b0ce5eeb64790910bdc0416c5e5561557642dede7022
        - quay.io/triliodata/operator-webhook-init@sha256:ffcc55e85f884b884947b73fa7d2fb257704ff70ccd7f2dc37d07e02292abed6
        - quay.io/triliodata/nats@sha256:e6723595bcc5674b6201091d45d711d29a5cd1421c8b97c76cf00e468379fddf
        - quay.io/triliodata/event-stack@sha256:859535e19e0ebebc8261c649c2446f8ca5fe4b4e5b6236e69b780dcc0d0bb5b9
        - quay.io/triliodata/control-plane@sha256:e60ef0a2c7a5b7879c9718e9a14c223f5290b4710b598fae19466a7558484824
        - quay.io/triliodata/web@sha256:04c4baadae9c05f4add098ff62774d00102d56581e261626840d17f74effe438
        version: 5.0.4
      entries:
      - name: k8s-triliovault-stable.5.0.4
        version: 5.0.4
      - name: k8s-triliovault-stable.5.0.3
        version: 5.0.3
      - name: k8s-triliovault-stable.5.0.2
        version: 5.0.2
      - name: k8s-triliovault-stable.5.0.1
        version: 5.0.1
      - name: k8s-triliovault-stable.5.0.0
        version: 5.0.0
      name: 5.0.x
    - currentCSV: k8s-triliovault-stable.5.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "2Gi",
                      "cpu": "1"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:f0cf3f586561414671668210c59ac44071fde6abc954917e6ca04df5777a9a2f
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <5.1.1'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/k8s-triliovault-operator@sha256:f0cf3f586561414671668210c59ac44071fde6abc954917e6ca04df5777a9a2f
        - quay.io/triliodata/filerecovery@sha256:42559ac5d949d5ecf819cd4768a71fa688c72230502ff0b56539a8bb5e2f5838
        - quay.io/triliodata/filerecovery-vm-ssh-server@sha256:09186be228bd0b59e83119b28f3239c6173e410a3cf36957fe2d9b56d25200f8
        - quay.io/triliodata/web@sha256:3908a39eed44d702c6b6804d392d08a570b3f5562590d050f17ddc7a5c87b37f
        - quay.io/triliodata/datamover@sha256:8a1350c8975610f23dd0b40d604d218f59dca79db6beb81ffe78a2682e40864f
        - quay.io/triliodata/control-plane@sha256:2e7e44cf6bc70dd9b978dce2cb2d9a36b8886c0f55286d5a6f5eb0a5182c4350
        - quay.io/triliodata/dex@sha256:6b9d1667916968fb1f187ad5945fcc08db702dcab4c947e309dd9d2881a7b9cf
        - quay.io/triliodata/nats@sha256:72283871b6b9343aeabc3ff090305979422f192d30a6cbc4b0f20b2de7533996
        - quay.io/triliodata/event-stack@sha256:548ae1a748919f518e811774ec87f128aa8857be234837e09308c295de165726
        version: 5.1.1
      entries:
      - name: k8s-triliovault-stable.5.1.1
        version: 5.1.1
      - name: k8s-triliovault-stable.5.1.0
        version: 5.1.0
      name: 5.1.x
    defaultChannel: 5.1.x
    packageName: k8s-triliovault
    provider:
      name: Trilio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Couchbase
      provider-url: ""
    name: couchbase-enterprise-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: couchbase-operator.v2.6.4-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:1f0a6865c360a14ba98706fd6fe8dff3bdbdede32e3622f93f19c82e377ca33f
        version: 2.6.4-7
      entries:
      - name: couchbase-operator.v2.6.4-7
        version: 2.6.4-7
      - name: couchbase-operator.v2.6.4-6
        version: 2.6.4-6
      name: 2.6.4
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator@sha256:707571a7881bc12d5f1b0b1a94eb3c675d5c0ea0114c994a47ceb5aa79df4013
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        version: 2.7.0-5
      entries:
      - name: couchbase-operator.v2.7.0-5
        version: 2.7.0-5
      name: 2.7.0
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/operator@sha256:a79e1b6781d5f8badf664e7efea6e2889e43fae6e0d37169d766685dcd6ff394
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        version: 2.7.1-1
      entries:
      - name: couchbase-operator.v2.7.1-1
        version: 2.7.1-1
      name: 2.7.1
    - currentCSV: couchbase-operator.v2.8.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator@sha256:df61928b064014d851c6e7dff6dcd63a433a27fa5623254b6e389c3eea170e36
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        version: 2.8.0-2
      entries:
      - name: couchbase-operator.v2.8.0-2
        version: 2.8.0-2
      - name: couchbase-operator.v2.8.0-1
        version: 2.8.0-1
      name: 2.8.0
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        version: 2.8.1-2
      entries:
      - name: couchbase-operator.v2.8.1-2
        version: 2.8.1-2
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: 2.8.1
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        version: 2.8.1-2
      entries:
      - name: couchbase-operator.v2.8.1-2
        version: 2.8.1-2
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: stable
    defaultChannel: stable
    packageName: couchbase-enterprise-certified
    provider:
      name: Couchbase
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubemq.io
      provider-url: ""
    name: kubemq-operator-marketplace-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubemq-operator.v1.7.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
              "apiVersion": "core.k8s.kubemq.io/v1beta1",
              "kind": "KubemqCluster",
              "metadata": {
                "name": "kubemq-cluster"
              },
              "spec": {
                "replicas": 3
              }
            },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqDashboard",
                "metadata": {
                  "name": "kubemq-dashboard"
                }
              },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqConnector",
                "metadata": {
                  "name": "kubemq-bridges"
                },
                "spec": {
                  "type": "bridges",
                  "config": "bindings: null"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/kubemq/kubemq-operator:v1.7.2
          createdAt: 2020-03-11 12:00:00+00:00
          description: Kubernetes Messaging Platform
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/kubemq-operator-marketplace-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/kubemq-operator-marketplace-rhmp/support?utm_source=openshift_console
          support: support@kubemq.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kubemq Cluster
            displayName: Kubemq Cluster
            kind: KubemqCluster
            name: kubemqclusters.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Cluster Dashboard
            displayName: Kubemq Dashboard
            kind: KubemqDashboard
            name: kubemqdashboards.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Connector
            displayName: Kubemq Connector
            kind: KubemqConnector
            name: kubemqconnectors.core.k8s.kubemq.io
            version: v1beta1
        description: |2-

          ## Kubernetes Messaging Platform
          A message-based architecture for microservices providing flexible connectivity for hybrid environments: multi-cloud, on-premises, and at the edge. Powered by enterprise-grade message broker and message queue with ready to use connectors, bridges, and control center.

          Messaging platform components:
            - **KubeMQ Cluster** - Enterprise-grade message broker and message queue, scalable, high available, and secured.
            - **KubeMQ Targets Connector** - connects a KubeMQ cluster with external systems and cloud services.
            - **KubeMQ Bridges Connector** - bridge, replicate, aggregate, and transform messages between KubeMQ clusters.
            - **KubeMQ Sources Connector** - transform external systems and cloud services data to KubeMQ cluster messages.
            - **KubeMQ Dashboard** - monitor KubeMQ cluster metrics.

          ### Kubernetes Native
          Innovative and modern message queue and message broker in a lightweight container developed to run in Kubernetes, certified in the CNCF landscape, and connect natively to the cloud-native ecosystem.

          ### Ease of use

          Simple deployment in Kubernetes in less than 1 minute. Developer friendly by simple to use SDKs and elimination of the many developers and DevOps-centered challenges to define exchanges, brokers, channels, routes, and predefined topics

          ### All messaging patterns

          KubeMQ is available with all messaging patterns such as Queue, Stream, Pub/Sub, and RPC, Saving the need to maintain multiple messaging platforms for different use cases.

          ### Hybrid Environments

          KubeMQ enables the microservices from multiple environments to communicate and build one hybrid infrastructure solution across clouds, on-prem, and at the edge.

          ### KubeMQ Cluster key features

          - All-batteries included Messaging Queue Broker for the Kubernetes environment
          - Deployed with Operator for full life cycle operation
          - Blazing fast (written in Go), small and lightweight Docker container
          - Asynchronous and Synchronous messaging with support for  `At Most Once Delivery` and `At Least Once Delivery` models
          - Supports durable FIFO based Queue, Publish-Subscribe Events, Publish-Subscribe with Persistence (Events Store), RPC Command and Query messaging patterns
          - Supports gRPC, Rest and WebSocket Transport protocols with TLS support (both RPC and Stream modes)
          - Supports Access control Authorization and Authentication
          - Supports message masticating and smart routing
          - No Message broker configuration needed (i.e., queues, exchanges)
          - .Net, Java, Python, Go, and NodeJS SDKs

          ### KubeMQ Targets Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major services types (databases, cache, messaging, serverless, HTTP, etc.)
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-targets)

          ### KubeMQ Bridges Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Build Any Topology - connects KubeMQ clusters in 1:1, 1:n , n:1, n:n
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-bridges)

          ### KubeMQ Sources Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major messaging services such as Kafka, ActiveMQ, SQS, and more.
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-sources)

          ## Getting Started

          [Start with OpenShift](https://docs.kubemq.io/getting-started/create-cluster/openshift)

          ## Get Enterprise Trial License

          [Register](https://account.kubemq.io/login/register)

          ## Request a Demo

          [See KubeMQ in Action](https://kubemq.io/request-a-demo/)
        displayName: Kubemq Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - messaging
        - queue
        - broker
        - pubsub
        - rpc
        links:
        - name: Learn more about Kubemq
          url: https://kubemq.io/
        - name: Register for free license
          url: https://account.kubemq.io/login/register
        - name: Contact Sales
          url: https://kubemq.io/contact-sales/
        - name: Documentation
          url: https://docs.kubemq.io/
        - name: Get Help
          url: https://kubemq.io/contact-support/
        maintainers:
        - email: support@kubemq.io
          name: kubemq
        maturity: alpha
        provider:
          name: Kubemq.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-grafana:v4.3
        - registry.marketplace.redhat.com/rhm/kubemq/kubemq-operator:v1.7.2
        - registry.marketplace.redhat.com/rhm/kubemq/kubemq-uni:v2.2.12
        - registry.redhat.io/openshift4/ose-prometheus:v4.3
        version: 1.7.8
      entries:
      - name: kubemq-operator.v1.7.8
        version: 1.7.8
      name: alpha
    defaultChannel: alpha
    packageName: kubemq-operator-marketplace-rhmp
    provider:
      name: Kubemq.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: crunchy-postgres-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/crunchy-postgres-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/crunchy-postgres-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-exporter@sha256:3b231902b2116cd38de05c7378202716c95813b9961fb8cee811b61d36cddf86
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:0e283616c2d32e53258d62d02702bc4e179ef8c03933592154babed1a70bf666
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:a7f62f367817fc08294a0b8b82ac9826a42c0ae668efdfacaa158a4287b8a688
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:c559c896e174674d151f93543ac92a1b5647eb4676313f7dc6cbb8faee786f18
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:56a5e3a5758982e086c9f78201a7137e0d934316f058a31cf30df44639a569a2
        - registry.connect.redhat.com/crunchydata/crunchy-pgadmin4@sha256:0d84acb1ce6a9b046e3f8b222a583ddcc91dda0a117c9f637ef173d0ac992f57
        - registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:b2c3c74fdcc0e08f555e50c527f0d322292333c5d3dce061a0dee3680a1fbc56
        - registry.connect.redhat.com/crunchydata/crunchy-pgbackrest@sha256:b70da0035b74379f2507cfabdd9d1cc1e3c3eecb9066c22f292981c9a27916e1
        - registry.connect.redhat.com/crunchydata/crunchy-pgbouncer@sha256:17cb58dec78447c2f70a414a2aeff6facbf921b42a0f366d0dbf7be95aa8987a
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:4bdd0588ffbb3f628e7df1c72670e7a5087269438343b68779012bc3a13dea31
        - registry.connect.redhat.com/crunchydata/crunchy-upgrade@sha256:68bc93cce61d9b1e44822ae3e3ac8f7a9c2b712eddfaea090eabefacd60c67f6
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: crunchy-postgres-operator-rhmp
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-networkfirewall-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-networkfirewall-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "Firewall",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "FirewallPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "RuleGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/networkfirewall-controller:1.1.1
          createdAt: "2025-09-19T21:03:12Z"
          description: AWS Network Firewall controller is a service controller for
            managing Network Firewall resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FirewallPolicy represents the state of an AWS networkfirewall
              FirewallPolicy resource.
            displayName: FirewallPolicy
            kind: FirewallPolicy
            name: firewallpolicies.networkfirewall.services.k8s.aws
            version: v1alpha1
          - description: Firewall represents the state of an AWS networkfirewall Firewall
              resource.
            displayName: Firewall
            kind: Firewall
            name: firewalls.networkfirewall.services.k8s.aws
            version: v1alpha1
          - description: RuleGroup represents the state of an AWS networkfirewall
              RuleGroup resource.
            displayName: RuleGroup
            kind: RuleGroup
            name: rulegroups.networkfirewall.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Network Firewall resources in AWS from within your Kubernetes cluster.

          **About Amazon Network Firewall**

          AWS Network Firewall is a stateful, managed, network firewall and intrusion detection and prevention service for your virtual private cloud (VPC) that you create in Amazon Virtual Private Cloud (Amazon VPC).
          With Network Firewall, you can filter traffic at the perimeter of your VPC. This includes filtering traffic going to and coming from an internet gateway, NAT gateway, or over VPN or AWS Direct Connect. Network Firewall uses the open source intrusion prevention system (IPS), Suricata, for stateful inspection. Network Firewall supports Suricata compatible rules. For more information, see [Working with stateful rule groups in AWS Network Firewall](https://docs.aws.amazon.com/network-firewall/latest/developerguide/stateful-rule-groups-ips.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Network Firewall
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networkfirewall
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Network Firewall Developer Resources
          url: https://aws.amazon.com/network-firewall/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: network firewall maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/networkfirewall-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-networkfirewall-controller.v1.1.1
        version: 1.1.1
      - name: ack-networkfirewall-controller.v1.1.0
        version: 1.1.0
      - name: ack-networkfirewall-controller.v1.0.13
        version: 1.0.13
      - name: ack-networkfirewall-controller.v1.0.12
        version: 1.0.12
      - name: ack-networkfirewall-controller.v1.0.11
        version: 1.0.11
      - name: ack-networkfirewall-controller.v1.0.10
        version: 1.0.10
      - name: ack-networkfirewall-controller.v1.0.9
        version: 1.0.9
      - name: ack-networkfirewall-controller.v1.0.8
        version: 1.0.8
      - name: ack-networkfirewall-controller.v1.0.7
        version: 1.0.7
      - name: ack-networkfirewall-controller.v1.0.6
        version: 1.0.6
      - name: ack-networkfirewall-controller.v1.0.5
        version: 1.0.5
      - name: ack-networkfirewall-controller.v1.0.4
        version: 1.0.4
      - name: ack-networkfirewall-controller.v1.0.1
        version: 1.0.1
      - name: ack-networkfirewall-controller.v1.0.0
        version: 1.0.0
      - name: ack-networkfirewall-controller.v0.0.10
        version: 0.0.10
      - name: ack-networkfirewall-controller.v0.0.9
        version: 0.0.9
      - name: ack-networkfirewall-controller.v0.0.8
        version: 0.0.8
      - name: ack-networkfirewall-controller.v0.0.7
        version: 0.0.7
      - name: ack-networkfirewall-controller.v0.0.6
        version: 0.0.6
      - name: ack-networkfirewall-controller.v0.0.5
        version: 0.0.5
      - name: ack-networkfirewall-controller.v0.0.2
        version: 0.0.2
      - name: ack-networkfirewall-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-networkfirewall-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: self-node-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: self-node-remediation.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:cb03286b9d9cda4a47e5b1c0058ca315bf082bcc8697506ea88d97ba77452bea
          createdAt: "2023-10-20 16:38:06"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          olm.skipRange: '>=0.4.0 <0.5.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the heath detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-self-node-remediation-operator.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:cb03286b9d9cda4a47e5b1c0058ca315bf082bcc8697506ea88d97ba77452bea
        version: 0.5.1
      entries:
      - name: self-node-remediation.v0.5.1
        version: 0.5.1
      name: 4.12-eus
    - currentCSV: self-node-remediation.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:e2048a8a2ec37e10b57dafeba2e131311e61982242cda4beb989e5e30c1a5537
          createdAt: "2023-10-20 15:32:20"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          olm.skipRange: '>=0.5.0 <0.7.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:e2048a8a2ec37e10b57dafeba2e131311e61982242cda4beb989e5e30c1a5537
        version: 0.7.1
      entries:
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.14-eus
    - currentCSV: self-node-remediation.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:6bd86a47d69f7662e24a93ce205a8179018152a97696a9223d41ed1887c35257
          createdAt: "2025-07-02 16:40:20"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.9.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:6bd86a47d69f7662e24a93ce205a8179018152a97696a9223d41ed1887c35257
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ef91935704bb9f0af4cdad07333e802587e3d458b2cd911e04d557b1defe3412
        version: 0.9.1
      entries:
      - name: self-node-remediation.v0.9.1
        version: 0.9.1
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.16-eus
    - currentCSV: self-node-remediation.v0.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
          createdAt: "2025-07-02 18:34:25"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        version: 0.10.2
      entries:
      - name: self-node-remediation.v0.10.2
        version: 0.10.2
      - name: self-node-remediation.v0.10.1
        version: 0.10.1
      - name: self-node-remediation.v0.10.0
        version: 0.10.0
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.18-eus
    - currentCSV: self-node-remediation.v0.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
          createdAt: "2025-07-02 18:34:25"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        version: 0.10.2
      entries:
      - name: self-node-remediation.v0.10.2
        version: 0.10.2
      - name: self-node-remediation.v0.10.1
        version: 0.10.1
      - name: self-node-remediation.v0.10.0
        version: 0.10.0
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: stable
    defaultChannel: stable
    packageName: self-node-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: aws-efs-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-efs-csi-driver-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-aws-efs-csi-driver-rhel9-operator@sha256:35ef24108df5f24d24fefe915f501965c98edea934fa6c2895e375ed006e5e7b
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure AWS EFS CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0-0 <4.16.0-202509240123'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/aws-efs-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          AWS EFS CSI Driver Operator provides AWS Elastic File Service (EFS) CSI Driver that enables you to create and mount AWS EFS PersistentVolumes.

          Please follow [AWS EFS CSI Driver Operator documentation](https://docs.openshift.com/container-platform/4.16/storage/container_storage_interface/persistent-storage-csi-aws-efs.html) to:

          1. [Install the operator](https://docs.openshift.com/container-platform/4.16/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#persistent-storage-csi-olm-operator-install_persistent-storage-csi-aws-efs). This section also covers details about configuring an AWS role in OCP cluster that uses Secure Token Services (STS).

          2. [Install the AWS EFS CSI driver using the operator](https://docs.openshift.com/container-platform/4.16/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#persistent-storage-csi-efs-driver-install_persistent-storage-csi-aws-efs)

          3. Finally, [create a StorageClass](https://docs.openshift.com/container-platform/4.16/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#storage-create-storage-class_persistent-storage-csi-aws-efs) that enables dynamic provisioning of PersistentVolumes.
        displayName: AWS EFS CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - efs
        links:
        - name: Documentation
          url: https://github.com/openshift/aws-efs-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/aws-efs-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-aws-efs-csi-driver-rhel9-operator@sha256:35ef24108df5f24d24fefe915f501965c98edea934fa6c2895e375ed006e5e7b
        - registry.redhat.io/openshift4/ose-aws-efs-csi-driver-container-rhel9@sha256:c857dda8f399fa104bd4f392a354b74365657acea6b8fc5b023fdb4bf55057c2
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:9c3c3d2b857ce289324aee56df1ba4ab5e5ea1e76cd86ac988233463915834b4
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0627c09fe3f5a015f117fb30af1f4d17340aefa48b23994737d3901a804b63d3
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:a5704d26d0b562dba9633d1c107f34e05c750c0e3a64dbcace4dab7b964fff8d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        version: 4.16.0-202509240123
      entries:
      - name: aws-efs-csi-driver-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: aws-efs-csi-driver-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: aws-efs-csi-driver-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: aws-efs-csi-driver-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: aws-efs-csi-driver-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: aws-efs-csi-driver-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: aws-efs-csi-driver-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: aws-efs-csi-driver-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: aws-efs-csi-driver-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: aws-efs-csi-driver-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: aws-efs-csi-driver-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: aws-efs-csi-driver-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: aws-efs-csi-driver-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: aws-efs-csi-driver-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: aws-efs-csi-driver-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: aws-efs-csi-driver-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: aws-efs-csi-driver-operator.v4.16.0-202410280205
        version: 4.16.0-202410280205
      - name: aws-efs-csi-driver-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: aws-efs-csi-driver-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: aws-efs-csi-driver-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: aws-efs-csi-driver-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: aws-efs-csi-driver-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: aws-efs-csi-driver-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: aws-efs-csi-driver-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: aws-efs-csi-driver-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: aws-efs-csi-driver-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: aws-efs-csi-driver-operator.v4.16.0-202408061508
        version: 4.16.0-202408061508
      - name: aws-efs-csi-driver-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: aws-efs-csi-driver-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: aws-efs-csi-driver-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: aws-efs-csi-driver-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: aws-efs-csi-driver-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: aws-efs-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: proactive-node-scaling-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: proactive-node-scaling-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NodeScalingWatermark",
                "metadata": {
                  "name": "ai-ml"
                },
                "spec": {
                  "nodeSelector": {
                    "machine.openshift.io/cluster-api-machine-type": "ai-ml"
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "workload",
                      "operator": "Equal",
                      "value": "ai-ml"
                    }
                  ],
                  "watermarkPercentage": 20
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NodeScalingWatermark",
                "metadata": {
                  "name": "us-west-2a"
                },
                "spec": {
                  "nodeSelector": {
                    "topology.kubernetes.io/zone": "us-west-2a"
                  },
                  "watermarkPercentage": 20
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "false"
          containerImage: quay.io/redhat-cop/proactive-node-scaling-operator@sha256:892303babb45090a6458cd4b1c61a01364012c43ea0bd6ed4d10bdcb8041e2ab
          createdAt: "2022-01-22T14:14:23Z"
          description: This operator provides automation proactively scaling nodes
            before pods get stuck waiting for the cluster autoscaler.
          operatorframework.io/suggested-namespace: proactive-node-scaling-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/proactive-node-scaling-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeScalingWatermark is the Schema for the nodescalingwatermarks
              API
            displayName: Node Scaling Watermark
            kind: NodeScalingWatermark
            name: nodescalingwatermarks.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator makes the [cluster autoscaler](https://github.com/kubernetes/autoscaler/tree/master/cluster-autoscaler) more proactive. As of now the cluster auto scaler will create new nodes only when a pod is pending because it cannot be allocated due to lack of capacity. This is not a goos user experience as the pending workload has to wait for several minutes as the new node is create and joins the cluster.

          The Proactive Node Scaling Operator improves the user experience by allocating low priority pods that don't do anything. When the cluster is full and a new user pod is created the following happens:

          1. some of the low priority pods are de-scheduled to make room for the user pod, which can then be scheduled. The user workload does not have to wait in this case.

          2. the de-scheduled low priority pods are rescheduled and in doing so the trigger the cluster autoscaler to add new nodes.

          Essentially this operator allows you to trade wasted resources for faster response time.

          In order for this operator to work correctly [pod priorities](https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/) must be defined. The default name for the priority class used by this operator is "proactive-node-autoscaling-pods" and it should have the lowest possible priority, 0. To ensure your regular workloads get a normal priority you should also define a PriorityClass for those with a higher priority than 0 and set globalDefault to true.

          For example:

          ```yaml
          apiVersion: scheduling.k8s.io/v1
          kind: PriorityClass
          metadata:
            name: proactive-node-autoscaling-pods
          value: 0
          globalDefault: false
          description: "This priority class is the priority class used for Proactive Node Scaling Pods."
          ---
          apiVersion: scheduling.k8s.io/v1
          kind: PriorityClass
          metadata:
            name: normal-workload
          value: 1000
          globalDefault: true
          description: "This priority classis the cluster default and should be used for normal workloads."
          ```

          Also for this operator to work the [cluster autoscaler](https://github.com/kubernetes/autoscaler/tree/master/cluster-autoscaler) must be active, see OpenShift instructions [here](https://docs.openshift.com/container-platform/4.6/machine_management/applying-autoscaling.html) on how to turn it on.

          To activate the proactive autoscaling, a CR must be defined, here is an example:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: NodeScalingWatermark
          metadata:
            name: us-west-2a
          spec:
            priorityClassName: proactive-node-autoscaling-pods
            watermarkPercentage: 20
            nodeSelector:
              topology.kubernetes.io/zone: us-west-2a
          ```

          The `nodeSelector` selects the nodes observed by this operator, which are also the nodes on which the low priority pods will be scheduled. The nodes observed by the cluster autoscaler should coincide with the nodes selected by this operator CR.

          The `watermarkPercentage` define the percentage of capacity of user workload that will be allocated to low priority pods. So in this example 20% of the user allocated capacity will be allocated via low priority pods. This also means that when the user workload reaches 80% capacity of the nodes selected by this CR (and the autoscaler), the cluster will start to scale.
        displayName: Proactive Node Scaling Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Capacity Management
        - Node Autoscaling
        - Proactive Autoscaling
        - Elastic Capacity
        links:
        - name: Proactive Node Scaling Operator
          url: https://github.com/redhat-cop/proactive-node-scaling-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/proactive-node-scaling-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/proactive-node-scaling-operator@sha256:892303babb45090a6458cd4b1c61a01364012c43ea0bd6ed4d10bdcb8041e2ab
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 0.4.2
      entries:
      - name: proactive-node-scaling-operator.v0.4.2
        version: 0.4.2
      - name: proactive-node-scaling-operator.v0.4.0
        version: 0.4.0
      - name: proactive-node-scaling-operator.v0.3.10
        version: 0.3.10
      - name: proactive-node-scaling-operator.v0.3.9
        version: 0.3.9
      - name: proactive-node-scaling-operator.v0.3.8
        version: 0.3.8
      - name: proactive-node-scaling-operator.v0.3.7
        version: 0.3.7
      - name: proactive-node-scaling-operator.v0.3.6
        version: 0.3.6
      - name: proactive-node-scaling-operator.v0.3.5
        version: 0.3.5
      - name: proactive-node-scaling-operator.v0.3.4
        version: 0.3.4
      - name: proactive-node-scaling-operator.v0.3.3
        version: 0.3.3
      - name: proactive-node-scaling-operator.v0.3.2
        version: 0.3.2
      - name: proactive-node-scaling-operator.v0.3.1
        version: 0.3.1
      - name: proactive-node-scaling-operator.v0.3.0
        version: 0.3.0
      - name: proactive-node-scaling-operator.v0.2.0
        version: 0.2.0
      - name: proactive-node-scaling-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: proactive-node-scaling-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Debezium Authors
      provider-url: https://debezium.io/
    name: debezium-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: debezium-operator.v2.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.4.2.Final
          createdAt: "2023-12-13T12:20:56Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.4.2.Final
        version: 2.4.2
      entries:
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.4.x
    - currentCSV: debezium-operator.v2.5.3-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.5.3.Final
          createdAt: "2024-03-19T08:17:17Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.5.3.Final
        version: 2.5.3-final
      entries:
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.5.x
    - currentCSV: debezium-operator.v2.6.1-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.6.1.Final
          createdAt: "2024-04-12T17:18:16Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.5.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.6.1.Final
        version: 2.6.1-final
      entries:
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.6.x
    - currentCSV: debezium-operator.v2.7.3-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.7.3.Final
          createdAt: "2024-09-20T10:31:16Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.7.3.Final
        version: 2.7.3-final
      entries:
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.7.x
    - currentCSV: debezium-operator.v3.0.0-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:3.0.0.Final
          createdAt: "2024-10-02T09:55:38Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.7.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:3.0.0.Final
        version: 3.0.0-final
      entries:
      - name: debezium-operator.v3.0.0-final
        version: 3.0.0-final
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-3.0.x
    - currentCSV: debezium-operator.v3.0.0-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:3.0.0.Final
          createdAt: "2024-10-02T09:55:38Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.7.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:3.0.0.Final
        version: 3.0.0-final
      entries:
      - name: debezium-operator.v3.0.0-final
        version: 3.0.0-final
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-latest
    defaultChannel: debezium-latest
    packageName: debezium-operator
    provider:
      name: Debezium Authors
      url: https://debezium.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-notifications
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-notifications.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Notifications",
                "metadata": {
                  "name": "notifications-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configuration": {
                    "port": 8080
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp",
                    "test": "wap-rhel8-toolbox:8.7.0-3747-36-ocp",
                    "testRepository": "prod.harbor.keysight.digital/loadcore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:34:47Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Notifications
            name: notifications.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Notifications Operator
        displayName: Keysight WAP Notifications Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Notifications
          url: https://keysight-wap-notifications.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-notifications@sha256:50c24180b3c16cf3eb82f6f4eeb61b91a62a198c122f7e59aa680a25bc488c2d
        version: 0.1.21
      entries:
      - name: keysight-wap-notifications.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-notifications
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-stats-dashboards
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-stats-dashboards.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Statsdashboard",
                "metadata": {
                  "name": "statsdashboard-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "statsDashboardService": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:41:08Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Statsdashboard
            name: statsdashboards.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Stats Dashboard Operator
        displayName: Keysight WAP Stats Dashboard Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Stats Dashboards
          url: https://keysight-wap-stats-dashboards.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-stats-dashboards@sha256:bcdc68bf0b9e9c22d775d704320829d258a9ebc77dfe41e2414061740f87f236
        version: 0.1.21
      entries:
      - name: keysight-wap-stats-dashboards.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-stats-dashboards
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Saison Technology Co., Ltd.
      provider-url: https://www.hulft.com/
    name: hulft
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hulft.v10.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hulft.saison-technology.com/v1",
                "kind": "Hulft",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "hulft",
                    "app.kubernetes.io/instance": "sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hulft",
                    "app.kubernetes.io/part-of": "hulft"
                  },
                  "name": "sample",
                  "namespace": "user-apps"
                },
                "spec": {
                  "applicationName": "hulft-apps",
                  "control": {
                    "allowExternalAccess": true,
                    "replicas": {
                      "max": 1,
                      "min": 1
                    }
                  },
                  "db": {
                    "authorizationSecret": "secret-hulft-db-authorization",
                    "connectionRetryCount": 30,
                    "hostname": "database.example.com",
                    "name": "hulft",
                    "port": 3306,
                    "tablePrefix": "hulft"
                  },
                  "licenseSecret": "secret-hulft-license",
                  "resourceIdentifier": "sample",
                  "transfer": {
                    "control": {
                      "retryCount": 10,
                      "retryInterval": 10000,
                      "timeout": 60000
                    },
                    "gracefulShutdownWait": 3600,
                    "observe": {
                      "websocket": {
                        "allowExternalAccess": false
                      }
                    },
                    "replicas": {
                      "max": 5,
                      "min": 1
                    },
                    "s3": {
                      "defaultRegion": "ap-northeast-1"
                    },
                    "websocket": {
                      "allowExternalAccess": false
                    }
                  }
                },
                "status": {
                  "applicationName": "hulft-apps",
                  "control": {
                    "allowExternalAccess": true,
                    "apiClientTlsSecret": "secret-hulft-sample-transfer-control-server-tls",
                    "consoleLoginSecret": "secret-hulft-sample-console-login",
                    "consoleTlsSecret": "secret-hulft-sample-control-console-tls",
                    "hostname": "",
                    "httpPort": 30080,
                    "httpsPort": 30443,
                    "replicas": {
                      "max": 1,
                      "min": 1
                    },
                    "samlSecret": ""
                  },
                  "db": {
                    "authorizationSecret": "secret-hulft-db-authorization",
                    "connectionRetryCount": 30,
                    "encryptionKeySecret": "secret-hulft-sample-db-encryption-key",
                    "hostname": "database.example.com",
                    "name": "hulft",
                    "port": 3306,
                    "tablePrefix": "hulft"
                  },
                  "licenseSecret": "secret-hulft-license",
                  "resourceIdentifier": "sample",
                  "transfer": {
                    "control": {
                      "clientTlsSecret": "secret-hulft-sample-control-console-tls",
                      "port": 30443,
                      "retryCount": 10,
                      "retryInterval": 10000,
                      "serverTlsSecret": "secret-hulft-sample-transfer-control-server-tls",
                      "timeout": 60000
                    },
                    "gracefulShutdownWait": 3600,
                    "hostname": "hulft-sample-transfer-wss-ci.apps.openshift.example.com",
                    "hulft": {
                      "port": 30000
                    },
                    "observe": {
                      "port": 31000,
                      "websocket": {
                        "allowExternalAccess": true,
                        "port": 33443,
                        "serverTlsSecret": "secret-hulft-sample-transfer-observe-wss-server-tls"
                      }
                    },
                    "replicas": {
                      "max": 5,
                      "min": 1
                    },
                    "s3": {
                      "authenticationSecret": "",
                      "defaultRegion": "ap-northeast-1"
                    },
                    "websocket": {
                      "allowExternalAccess": true,
                      "clientTlsSecret": "",
                      "port": 31443,
                      "serverTlsSecret": "secret-hulft-sample-transfer-wss-server-tls"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/saison-technology/hulft-operator@sha256:8ee0a29d0ce96f4ec230f387467b241dccbb2bd350cd642b275ef5f5d3b64972
          createdAt: "2025-09-30T07:54:26Z"
          description: HULFT10 for Container Platform is file integration middleware
            that provides transfer functions for the data accumulated in your IT system
            during the daily operations of your business, such as sales data, customer
            information, and various logs. File transfer from/to on-premises HULFT
            is also available.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: To use HULFT10 for Container
            Platform, you must purchase the product and have a key issued. Please
            refer to [Start Up Guide](https://www.hulft.com/help/ja-jp/HULFT-V10/CONP-SUGOS)
            for the deployment procedure.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Saison Technology Co., Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hulft is the Schema for the hulfts API
            displayName: Hulft
            kind: Hulft
            name: hulfts.hulft.saison-technology.com
            version: v1
        description: |
          HULFT10 for Container Platform is file integration middleware that provides transfer functions for the data accumulated in your IT system during the daily operations of your business, such as sales data, customer information, and various logs. File transfer from/to on-premises HULFT is also available.
          To use HULFT10 for Container Platform, you must purchase the product and have a key issued. Please refer to [Start Up Guide](https://www.hulft.com/help/ja-jp/HULFT-V10/CONP-SUGOS) for the deployment procedure.
        displayName: HULFT10 for Container Platform
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backup & Recovery
        - Storage
        - Migration
        links:
        - name: Product page
          url: https://www.hulft.com/software/hulft-10
        - name: Product manual
          url: https://www.hulft.com/help/ja-jp/HULFT-V10/index.htm
        maintainers:
        - email: salessupport_mp@hulft.com
          name: Saison Technology Co., Ltd.
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Saison Technology Co., Ltd.
          url: https://www.hulft.com/
        relatedImages:
        - registry.connect.redhat.com/saison-technology/hulft-operator@sha256:8ee0a29d0ce96f4ec230f387467b241dccbb2bd350cd642b275ef5f5d3b64972
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fa0acefda43e309e7f9f31f5247c933d0b868ff1573ece2f38e0549d7e2ec942
        version: 10.5.0
      entries:
      - name: hulft.v10.5.0
        version: 10.5.0
      - name: hulft.v10.4.0
        version: 10.4.0
      - name: hulft.v10.3.0
        version: 10.3.0
      name: stable
    defaultChannel: stable
    packageName: hulft
    provider:
      name: Saison Technology Co., Ltd.
      url: https://www.hulft.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC Corporation
      provider-url: https://jpn.nec.com/webotx/
    name: webotx-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: webotx-operator.v11.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "webotx.nec.com/v3",
                "kind": "ApplicationServer",
                "metadata": {
                  "name": "webotx-express"
                },
                "spec": {
                  "deployment": {
                    "image": "registry.connect.redhat.com/nec/webotx-express:11.11.00.00@sha256:73a5dc48b4e3340951dbe207d744f5ac68474701dd134fa7aa05e65e3510c8ad",
                    "replicas": 1
                  },
                  "service": {
                    "externalTrafficPolicy": "Cluster",
                    "ports": [
                      {
                        "name": "http",
                        "nodePort": 30080,
                        "port": 8080,
                        "protocol": "TCP",
                        "targetPort": 8080
                      },
                      {
                        "name": "https",
                        "nodePort": 30443,
                        "port": 8443,
                        "protocol": "TCP",
                        "targetPort": 8443
                      }
                    ],
                    "sessionAffinity": "ClientIP",
                    "sessionAffinityConfig": {
                      "clientIP": {
                        "timeoutSeconds": 10800
                      }
                    },
                    "type": "NodePort"
                  },
                  "webotx": {
                    "metering": {
                      "enabled": false,
                      "port": 8443,
                      "scheme": "https"
                    },
                    "sessionReplication": {
                      "enabled": false,
                      "useAllProcessGroups": false,
                      "useJmxAgent": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApplicationServer is the Schema for the applicationservers
              API
            displayName: Application Server
            kind: ApplicationServer
            name: applicationservers.webotx.nec.com
            version: v3
        description: WebOTX Operator is the Operator for WebOTX Application Server.
          WebOTX Operator enables to deploy and administrate WebOTX Application Server
          automatically.
        displayName: WebOTX Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - webotx
        - WebOTX
        - nec
        - NEC
        links:
        - name: WebOTX Operator for Kubernetes
          url: https://github.com/webotx/operator
        maintainers:
        - email: info-webotx@isd.jp.nec.com
          name: NEC Corporation
        maturity: stable
        provider:
          name: NEC Corporation
          url: https://jpn.nec.com/webotx/
        relatedImages:
        - registry.connect.redhat.com/nec/webotx-operator:11.11.00.00@sha256:df0d0315ed88be875ca7a6474500f43f17855f763c5c7d0f4d3ec58a2e46f6fb
        version: 11.11.0
      entries:
      - name: webotx-operator.v11.11.0
        version: 11.11.0
      name: stable
    defaultChannel: stable
    packageName: webotx-operator
    provider:
      name: NEC Corporation
      url: https://jpn.nec.com/webotx/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator-certified-rhmp.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a",
                  "imagePullPolicy": "Always",
                  "initImage": "registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-server-mongodb-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-server-mongodb-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:3c03fb18a5d73876576b80aca2b61e7c84ca052839ccb467d14e15f6d1f9d834
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a57926b53eef862b70a53818a7ccb0e9756f878f30d4128b608090b109bdc585
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator-certified-rhmp.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator-certified-rhmp.v1.20.0
        version: 1.20.0
      - name: percona-server-mongodb-operator-certified-rhmp.v1.19.1
        version: 1.19.1
      - name: percona-server-mongodb-operator-certified-rhmp.v1.19.0
        version: 1.19.0
      - name: percona-server-mongodb-operator-certified-rhmp.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator-certified-rhmp
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kms-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kms-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Alias",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Key",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kms-controller:1.1.1
          createdAt: "2025-09-19T22:02:02Z"
          description: AWS KMS controller is a service controller for managing KMS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alias represents the state of an AWS kms Alias resource.
            displayName: Alias
            kind: Alias
            name: aliases.kms.services.k8s.aws
            version: v1alpha1
          - description: Grant represents the state of an AWS kms Grant resource.
            displayName: Grant
            kind: Grant
            name: grants.kms.services.k8s.aws
            version: v1alpha1
          - description: Key represents the state of an AWS kms Key resource.
            displayName: Key
            kind: Key
            name: keys.kms.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Key Management Service (KMS) resources in AWS from within your Kubernetes cluster.

          **About Amazon KMS**

          AWS Key Management Service (AWS KMS) is a managed service that makes it easy for you to create and control the cryptographic keys that are used to protect your data. AWS KMS uses hardware security modules (HSM) to protect and validate your AWS KMS keys under the [FIPS 140-2 Cryptographic Module Validation Program](https://csrc.nist.gov/projects/cryptographic-module-validation-program/Certificate/3139), except in the China (Beijing) and China (Ningxia) Regions.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon KMS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kms
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon KMS Developer Resources
          url: https://aws.amazon.com/kms/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kms maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/kms-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-kms-controller.v1.1.1
        version: 1.1.1
      - name: ack-kms-controller.v1.1.0
        version: 1.1.0
      - name: ack-kms-controller.v1.0.31
        version: 1.0.31
      - name: ack-kms-controller.v1.0.30
        version: 1.0.30
      - name: ack-kms-controller.v1.0.29
        version: 1.0.29
      - name: ack-kms-controller.v1.0.28
        version: 1.0.28
      - name: ack-kms-controller.v1.0.27
        version: 1.0.27
      - name: ack-kms-controller.v1.0.26
        version: 1.0.26
      - name: ack-kms-controller.v1.0.25
        version: 1.0.25
      - name: ack-kms-controller.v1.0.24
        version: 1.0.24
      - name: ack-kms-controller.v1.0.23
        version: 1.0.23
      - name: ack-kms-controller.v1.0.22
        version: 1.0.22
      - name: ack-kms-controller.v1.0.19
        version: 1.0.19
      - name: ack-kms-controller.v1.0.18
        version: 1.0.18
      - name: ack-kms-controller.v1.0.17
        version: 1.0.17
      - name: ack-kms-controller.v1.0.16
        version: 1.0.16
      - name: ack-kms-controller.v1.0.15
        version: 1.0.15
      - name: ack-kms-controller.v1.0.14
        version: 1.0.14
      - name: ack-kms-controller.v1.0.13
        version: 1.0.13
      - name: ack-kms-controller.v1.0.12
        version: 1.0.12
      - name: ack-kms-controller.v1.0.8
        version: 1.0.8
      - name: ack-kms-controller.v1.0.7
        version: 1.0.7
      - name: ack-kms-controller.v1.0.5
        version: 1.0.5
      - name: ack-kms-controller.v1.0.4
        version: 1.0.4
      - name: ack-kms-controller.v1.0.3
        version: 1.0.3
      - name: ack-kms-controller.v1.0.2
        version: 1.0.2
      - name: ack-kms-controller.v1.0.0
        version: 1.0.0
      - name: ack-kms-controller.v0.1.5
        version: 0.1.5
      - name: ack-kms-controller.v0.1.4
        version: 0.1.4
      - name: ack-kms-controller.v0.1.3
        version: 0.1.3
      - name: ack-kms-controller.v0.1.2
        version: 0.1.2
      - name: ack-kms-controller.v0.1.1
        version: 0.1.1
      - name: ack-kms-controller.v0.1.0
        version: 0.1.0
      - name: ack-kms-controller.v0.0.20
        version: 0.0.20
      - name: ack-kms-controller.v0.0.19
        version: 0.0.19
      - name: ack-kms-controller.v0.0.18
        version: 0.0.18
      - name: ack-kms-controller.v0.0.17
        version: 0.0.17
      - name: ack-kms-controller.v0.0.16
        version: 0.0.16
      - name: ack-kms-controller.v0.0.15
        version: 0.0.15
      - name: ack-kms-controller.v0.0.14
        version: 0.0.14
      - name: ack-kms-controller.v0.0.13
        version: 0.0.13
      - name: ack-kms-controller.v0.0.12
        version: 0.0.12
      - name: ack-kms-controller.v0.0.11
        version: 0.0.11
      - name: ack-kms-controller.v0.0.10
        version: 0.0.10
      - name: ack-kms-controller.v0.0.9
        version: 0.0.9
      - name: ack-kms-controller.v0.0.8
        version: 0.0.8
      - name: ack-kms-controller.v0.0.7
        version: 0.0.7
      - name: ack-kms-controller.v0.0.6
        version: 0.0.6
      - name: ack-kms-controller.v0.0.5
        version: 0.0.5
      - name: ack-kms-controller.v0.0.4
        version: 0.0.4
      - name: ack-kms-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-kms-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: boxboat-github-practice
      provider-url: https://github.com/boxboat-github-practice
    name: github-arc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: github-arc-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "github-practice.boxboat.com/v1alpha1",
                "kind": "ActionsRunnerController",
                "metadata": {
                  "name": "arc-sample"
                },
                "spec": {
                  "authSecret": {
                    "name": "ghauth"
                  },
                  "createRunnerNamespaces": true,
                  "openshift": true,
                  "runnerNamespaces": [
                    "test1",
                    "test2"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery, OpenShift Optional
          containerImage: ghcr.io/boxboat-github-practice/github-arc-operator:1.1.0
          createdAt: "2023-03-03T20:49:16Z"
          description: Operator for deploying the github actions runner controller
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/boxboat-github-practice/github-arc-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActionsRunnerController is the schema for the ActionsRunnerController
              API
            displayName: Actions Runner Controller
            kind: ActionsRunnerController
            name: actionsrunnercontrollers.github-practice.boxboat.com
            version: v1alpha1
        description: helm based operator for github actions runner controller
        displayName: github-arc-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - github
        - actions
        - boxboat
        - ibm
        - self-hosted
        - runners
        links:
        - name: Github Arc Operator
          url: https://github.com/boxboat-github-practice/github-arc-operator
        - name: Actions Runner Controller
          url: https://github.com/actions/actions-runner-controller
        - name: Boxboat
          url: https://boxboat.com/
        maintainers:
        - email: Michael.McMullen@ibm.com
          name: mbmcmullen27
        maturity: alpha
        provider:
          name: boxboat-github-practice
          url: https://github.com/boxboat-github-practice
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - ghcr.io/boxboat-github-practice/github-arc-operator:1.1.0
        version: 1.1.0
      entries:
      - name: github-arc-operator.v1.1.0
        version: 1.1.0
      - name: github-arc-operator.v1.0.4
        version: 1.0.4
      name: alpha
    defaultChannel: alpha
    packageName: github-arc-operator
    provider:
      name: boxboat-github-practice
      url: https://github.com/boxboat-github-practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: special-resource-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: special-resource-operator.v4.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sro.openshift.io/v1beta1",
                "kind": "SpecialResource",
                "metadata": {
                  "name": "simple-kmod"
                },
                "spec": {
                  "chart": {
                    "name": "simple-kmod",
                    "repository": {
                      "name": "example",
                      "url": "file:///charts/example"
                    },
                    "version": "0.0.1"
                  },
                  "driverContainer": {
                    "source": {
                      "git": {
                        "ref": "master",
                        "uri": "https://github.com/openshift-psap/kvc-simple-kmod.git"
                      }
                    }
                  },
                  "namespace": "simple-kmod",
                  "set": {
                    "apiVersion": "sro.openshift.io/v1beta1",
                    "buildArgs": [
                      {
                        "name": "KMODVER",
                        "value": "SRO"
                      }
                    ],
                    "kind": "Values",
                    "kmodNames": [
                      "simple-kmod",
                      "simple-procfs-kmod"
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional, Storage
          containerImage: quay.io/openshift-psap/special-resource-operator:release-4.9
          description: The special resource operator is an orchestrator for resources
            in a cluster required to deploy special hardware and software. Examples
            include hardware accelerators or software defined storage, where extra
            management or kernel modules are needed.
          olm.skipRange: '>=4.6.0 <4.9.0'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpecialResource is the Schema for the specialresources API
            displayName: Special Resource
            kind: SpecialResource
            name: specialresources.sro.openshift.io
            version: v1beta1
          required:
          - description: Discovers special resources and labels the nodes with the
              PCI vendor ID.
            displayName: Node Feature Discovery
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
        description: "## About the managed application \nThe SRO will deploy the complete
          stack for enabling the accelerator in a cluster. \nIt manages the drivers,
          device-plugin, node-exporter, grafana and feature discovery.\n## About this
          Operator\n README: [https://github.com/openshift-psap/special-resource-operator](https://github.com/openshift-psap/special-resource-operator)\n\n##
          Prerequisites for enabling this Operator\nThe SRO has a dependency on NFD
          (Node Feature Discovery)\n"
        displayName: Special Resource Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GPI
        links:
        - name: README
          url: https://github.com/openshift-psap/special-resource-operator
        maintainers:
        - email: zkaiser@redhat.com
          name: Zvonko Kaiser
        - email: dagray@redhat.com
          name: David Gray
        maturity: alpha
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy
        - quay.io/openshift-psap/special-resource-operator:release-4.9
        version: 4.9.0
      entries:
      - name: special-resource-operator.v4.9.0
        version: 4.9.0
      - name: special-resource-operator.v4.8.2
        version: 4.8.2
      - name: special-resource-operator.v4.8.1
        version: 4.8.1
      - name: special-resource-operator.v4.8.0
        version: 4.8.0
      name: alpha
    defaultChannel: alpha
    packageName: special-resource-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Digital.ai
      provider-url: https://digital.ai
    name: digitalai-release-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: digitalai-release-operator.v25.3.0-beta.813
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 5,
                    "metrics": [],
                    "minReplicas": 3
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "cluster": {
                    "management": {
                      "ssl": {
                        "enabled": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    },
                    "manager": "xlr-legacy",
                    "mode": "full",
                    "name": "xl-release_cluster",
                    "node": {
                      "datacenter": "default"
                    },
                    "remote": {
                      "sslEngine": {
                        "enabled": false,
                        "enabledAlgorithms": [
                          "TLS_DHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_DHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_RSA_WITH_AES_128_CBC_SHA"
                        ],
                        "hostnameVerification": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "requireMutualAuthentication": false,
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"${XLR_CLUSTER_NAME}\"\n    manager = \"${XLR_CLUSTER_MANAGER}\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n      {{- if .Values.storage.reportingEngine.downingProviderClass }}\n      cluster {\n        downing-provider-class = {{.Values.storage.reportingEngine.downingProviderClass | quote}}\n      }\n      {{- end }}\n\n      {{- if .Values.cluster.remote.sslEngine.enabled }}\n      remote {\n        artery {  \n          ssl {  \n            config-ssl-engine {  \n              enabled-algorithms = [\n                {{- $last := sub (len .Values.cluster.remote.sslEngine.enabledAlgorithms) 1 }}\n                {{ range $index, $v := .Values.cluster.remote.sslEngine.enabledAlgorithms }}\n                {{ $v | quote }}{{ if ne $index $last }},{{ end }}\n                {{- end }}\n              ]  \n              key-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEY_PASSWORD}\"  \n              key-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.keystore \"default\" (printf \"cluster_remote_keystore.%s\" .Values.cluster.remote.sslEngine.keystoreType) \"context\" $) }}\"  \n              key-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEYSTORE_PASSWORD}\"  \n              trust-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.truststore \"default\" (printf \"cluster_remote_truststore.%s\" .Values.cluster.remote.sslEngine.truststoreType) \"context\" $) }}\"  \n              trust-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_TRUSTSTORE_PASSWORD}\"  \n              hostname-verification = {{ if .Values.cluster.remote.sslEngine.hostnameVerification }}on{{ else }}off{{ end }}\n              require-mutual-authentication = {{ if .Values.cluster.remote.sslEngine.requireMutualAuthentication }}on{{ else }}off{{ end }}  \n            }  \n          }  \n          transport = tls-tcp  \n        }  \n      }\n      {{- end }}\n    }\n    {{- if .Values.cluster.management.ssl.enabled }}\n    xlr-pekko-native {\n      pekko {\n        management {\n          http {  \n            route-providers-read-only = true  \n          }  \n          ssl {  \n            enabled = true\n            keyPassword = \"${?CLUSTER_MANAGEMENT_SSL_KEY_PASSWORD}\"  \n            keyStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.keystore \"default\" (printf \"cluster_management_keystore.%s\" .Values.cluster.management.ssl.keystoreType) \"context\" $) }}\"  \n            keyStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_KEYSTORE_PASSWORD}\"  \n            keystoreType = {{ .Values.cluster.management.ssl.keystoreType }}  \n            trustStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.truststore \"default\" (printf \"cluster_management_truststore.%s\" .Values.cluster.management.ssl.truststoreType) \"context\" $) }}\"  \n            trustStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_TRUSTSTORE_PASSWORD}\"  \n            trustStoreType = {{ .Values.cluster.management.ssl.truststoreType }}  \n          }  \n        }  \n      }\n    }\n    {{- end }}\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\nmanagement.endpoint.prometheus.enabled = {{ .Values.metrics.prometheus.enabled | default false }}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      datacenter = ${XLR_CLUSTER_DATACENTER}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "cd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "livenessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/liveness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "jmx": {
                      "enabled": false
                    },
                    "prometheus": {
                      "enabled": false,
                      "podAnnotations": {
                        "prometheus.io/port": "{{ .Values.service.ports.metrics }}",
                        "prometheus.io/scrape": "true"
                      },
                      "rule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "serviceMonitor": {
                        "annotations": {},
                        "basicAuth": {
                          "password": {
                            "key": "releasePassword",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          },
                          "username": {
                            "key": "releaseUsername",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          }
                        },
                        "enabled": false,
                        "honorLabels": false,
                        "interval": "30s",
                        "jobLabel": "",
                        "labels": {},
                        "metricRelabelings": [],
                        "namespace": "",
                        "params": {},
                        "path": "/s/actuator/prometheus",
                        "podTargetLabels": {},
                        "relabelings": [],
                        "scrapeTimeout": "",
                        "selector": {
                          "service-type": "normal"
                        },
                        "targetLabels": {}
                      }
                    }
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.0-debian-12-r4"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.0-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/extra/lib",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.5.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rbac": {
                    "create": true
                  },
                  "readinessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {
                      "service-type": "normal"
                    },
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "startupProbe": {
                    "enabled": true,
                    "failureThreshold": 12,
                    "initialDelaySeconds": 60,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 450,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnamilegacy/os-shell",
                      "tag": "12-debian-12-r49"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabsunsupported/release-operator@sha256:52914f17dc55dfaef55aed382e7269408fb676a213d12117eafa906d81506824 '
          createdAt: "2025-09-02T05:48:19Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql.\n\n```yaml\napiVersion:
          xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n
          \ global:\n    security:\n      allowInsecureImages: true\n  k8sSetup:\n
          \   platform: Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula:
          true\n  keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n```\n\n#### Configuration Details\n\nThe sample configuration
          uses:\n\n- Autogenerated limited license\n- Simple administrator initial
          password\n- Default keystore setup\n- Only one replica is set for all pods\n-
          All Pods are missing resource limit setup\n- Default storage class and minimal
          size for persistent storage\n- Release pods are using persistence access
          mode `ReadWriteOnce`, for Release pods the requirement is to have `ReadWriteMany`\n-
          Embedded PostgreSQL for DB management\n\n#### Security Configuration Details\n\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabsunsupported/xl-release@sha256:0548a1ba63da5ae09c59b67a30240bf48877ab8950e09b0f0f6fd7c88ba34c13
        - docker.io/xebialabsunsupported/release-operator@sha256:52914f17dc55dfaef55aed382e7269408fb676a213d12117eafa906d81506824
        version: 25.3.0-beta.813
      entries:
      - name: digitalai-release-operator.v25.3.0-beta.813
        version: 25.3.0-beta.813
      - name: digitalai-release-operator.v25.3.0-beta.711
        version: 25.3.0-beta.711
      - name: digitalai-release-operator.v25.3.0-beta.609
        version: 25.3.0-beta.609
      - name: digitalai-release-operator.v25.3.0-beta.512
        version: 25.3.0-beta.512
      name: beta
    - currentCSV: digitalai-release-operator.v24.1.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "fsGroup": null,
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "fsGroup": null,
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  jvm {\n      override-jvm-options = {{ .Values.jvm.overrideOptions }}\n      networkaddress.cache.ttl = {{ index .Values.jvm.options \"networkaddress.cache.ttl\" }}\n  }\n}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "ingressClass": "haproxy-dair",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-client/xl",
                        "kube",
                        "upgrade",
                        "--skip-context-check",
                        "--local-repo",
                        "/opt/xebialabs/xl-op-blueprints",
                        "--answers",
                        "/opt/xebialabs/xl-client/generated_answers.yaml"
                      ],
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    },
                    "overrideOptions": true
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "tag": "1.26.0-debian-12-r1"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "tag": "1.10.1-debian-12-r5"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "tag": "15.6.0-debian-12-r7"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "tag": "3.12.13-debian-12-r2"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnami/os-shell",
                      "tag": "12-debian-12-r16"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator without SCC creation and with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: docker.io/xebialabs/release-operator@sha256:47d6e31f66fce52c6566240b85fb028890409ad3addeea134120d1afa4814883
          createdAt: "2024-10-16T04:25:19Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind:
          DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    runAsUser: null\n    runAsGroup: null\n
          \   fsGroup: null\n  containerSecurityContext:\n    runAsUser: null\n    runAsGroup:
          null\n  securityContextConstraints:\n    enabled: false\n  volumePermissions:\n
          \   enabled: false\n  route:\n    enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n        runAsUser: null\n        runAsGroup: null\n        fsGroup:
          null\n      containerSecurityContext:\n        enabled: false\n        runAsUser:
          null\n        runAsGroup: null\n      securityContextConstraints:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n      runAsUser: null\n
          \     runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     enabled: false\n      runAsUser: null\n      runAsGroup: null\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"dair-doc-digitalai-release-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xlr-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Release pods are using
          persistence access mode `ReadWriteOnce`, for Release pods the requirement
          is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB management\n- Embedded
          RabbitMQ for message queue management\n\n#### Security Configuration Details\n\n-
          `securityContextConstraints.enabled: false` - disables creation of SCCs;\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/release-operator@sha256:47d6e31f66fce52c6566240b85fb028890409ad3addeea134120d1afa4814883
        - docker.io/xebialabs/xl-release@sha256:d1b65493ed524d37c60483f44b8bcc877ff398c81d36c7e767f2254dadec98cc
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:771a9a173e033a3ad8b46f5c00a7036eaa88c8d8d1fbd89217325168998113ea
        version: 24.1.10
      entries:
      - name: digitalai-release-operator.v24.1.10
        version: 24.1.10
      name: stable
    - currentCSV: digitalai-release-operator.v24.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  jvm {\n      override-jvm-options = {{ .Values.jvm.overrideOptions }}\n      networkaddress.cache.ttl = {{ index .Values.jvm.options \"networkaddress.cache.ttl\" }}\n  }\n}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "image": {
                        "tag": "v0.14.7"
                      },
                      "ingressClass": "haproxy-dair",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-client/xl",
                        "kube",
                        "upgrade",
                        "--skip-context-check",
                        "--local-repo",
                        "/opt/xebialabs/xl-op-blueprints",
                        "--answers",
                        "/opt/xebialabs/xl-client/generated_answers.yaml"
                      ],
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    },
                    "overrideOptions": true
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.26.2-debian-12-r10"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "15.6.0-debian-12-r20"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.12.14-debian-12-r7"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnamilegacy/os-shell",
                      "tag": "12-debian-12-r49"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator without SCC creation and with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:aca76f0f5db170ddfc6ad692ea2874b9efb83fe732540946d88a0c078907ec8c '
          createdAt: "2025-08-22T08:54:25Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind:
          DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  securityContextConstraints:\n    enabled: false\n
          \ volumePermissions:\n    enabled: false\n  route:\n    enabled: false\n
          \   annotations:\n      haproxy.router.openshift.io/cookie_name: JSESSIONID\n
          \     haproxy.router.openshift.io/disable_cookies: 'false'\n      haproxy.router.openshift.io/rewrite-target:
          /\n      haproxy.router.openshift.io/timeout: 120s\n    hostname: '<mandatory-release-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n      securityContextConstraints:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   securityContextConstraints:\n      enabled: false\n    volumePermissions:\n
          \     enabled: false\n```\n\nIf SCC is required, it must be managed manually.
          The necessary permissions for accessing SecurityContextConstraints must
          be granted when SCC is enabled; otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"dair-doc-digitalai-release-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xlr-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Release pods are using
          persistence access mode `ReadWriteOnce`, for Release pods the requirement
          is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB management\n- Embedded
          RabbitMQ for message queue management\n\n#### Security Configuration Details\n\n-
          `securityContextConstraints.enabled: false` - disables creation of SCCs;\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/xl-release@sha256:1cadd1d6ad7389ac9fe398ffba246b78238e54c180045b9dd44fdeefa3cf2af5
        - docker.io/xebialabs/release-operator@sha256:aca76f0f5db170ddfc6ad692ea2874b9efb83fe732540946d88a0c078907ec8c
        version: 24.1.21
      entries:
      - name: digitalai-release-operator.v24.1.21
        version: 24.1.21
      - name: digitalai-release-operator.v24.1.20
        version: 24.1.20
      - name: digitalai-release-operator.v24.1.19
        version: 24.1.19
      - name: digitalai-release-operator.v24.1.18
        version: 24.1.18
      - name: digitalai-release-operator.v24.1.17
        version: 24.1.17
      - name: digitalai-release-operator.v24.1.14
        version: 24.1.14
      - name: digitalai-release-operator.v24.1.10
        version: 24.1.10
      name: stable-v24.1
    - currentCSV: digitalai-release-operator.v24.3.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "#!/bin/sh\n\ncd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "16.3.0-debian-12-r23"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n{{ if eq .Values.queueType \"quorum\" -}}\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n{{- end }}\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.13.7-debian-12-r5"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "queueType": "classic",
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-release",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:725fb5f7cfa4098c4110a40f54dc646f7663267bcba53603a7c1303c44051803 '
          createdAt: "2025-09-26T18:24:33Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n
          \ name: dair-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n```\n\n#### Configuration
          Details\n\nThe sample configuration uses:\n\n- Autogenerated limited license\n-
          Simple administrator initial password\n- Default keystore setup\n- Only
          one replica is set for all pods\n- All Pods are missing resource limit setup\n-
          Default storage class and minimal size for persistent storage\n- Release
          pods are using persistence access mode `ReadWriteOnce`, for Release pods
          the requirement is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB
          management\n- Embedded RabbitMQ for message queue management\n\n#### Security
          Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined ranges (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/release-operator@sha256:725fb5f7cfa4098c4110a40f54dc646f7663267bcba53603a7c1303c44051803
        - docker.io/xebialabs/xl-release@sha256:e9e87975888fcd4c8c164db90cb0a5c7abde88bf272c59d97e465af7b49bc6be
        version: 24.3.11
      entries:
      - name: digitalai-release-operator.v24.3.11
        version: 24.3.11
      - name: digitalai-release-operator.v24.3.10
        version: 24.3.10
      - name: digitalai-release-operator.v24.3.9
        version: 24.3.9
      - name: digitalai-release-operator.v24.3.8
        version: 24.3.8
      - name: digitalai-release-operator.v24.3.7
        version: 24.3.7
      - name: digitalai-release-operator.v24.3.6
        version: 24.3.6
      - name: digitalai-release-operator.v24.3.5
        version: 24.3.5
      name: stable-v24.3
    - currentCSV: digitalai-release-operator.v25.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "cluster": {
                    "management": {
                      "ssl": {
                        "enabled": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    },
                    "manager": "xlr-legacy",
                    "mode": "full",
                    "name": "xl-release_cluster",
                    "node": {
                      "datacenter": "default"
                    },
                    "remote": {
                      "sslEngine": {
                        "enabled": false,
                        "enabledAlgorithms": [
                          "TLS_DHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_DHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_RSA_WITH_AES_128_CBC_SHA"
                        ],
                        "hostnameVerification": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "requireMutualAuthentication": false,
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"${XLR_CLUSTER_NAME}\"\n    manager = \"${XLR_CLUSTER_MANAGER}\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n      {{- if .Values.storage.reportingEngine.downingProviderClass }}\n      cluster {\n        downing-provider-class = {{.Values.storage.reportingEngine.downingProviderClass | quote}}\n      }\n      {{- end }}\n\n      {{- if .Values.cluster.remote.sslEngine.enabled }}\n      remote {\n        artery {  \n          ssl {  \n            config-ssl-engine {  \n              enabled-algorithms = [\n                {{- $last := sub (len .Values.cluster.remote.sslEngine.enabledAlgorithms) 1 }}\n                {{ range $index, $v := .Values.cluster.remote.sslEngine.enabledAlgorithms }}\n                {{ $v | quote }}{{ if ne $index $last }},{{ end }}\n                {{- end }}\n              ]  \n              key-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEY_PASSWORD}\"  \n              key-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.keystore \"default\" (printf \"cluster_remote_keystore.%s\" .Values.cluster.remote.sslEngine.keystoreType) \"context\" $) }}\"  \n              key-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEYSTORE_PASSWORD}\"  \n              trust-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.truststore \"default\" (printf \"cluster_remote_truststore.%s\" .Values.cluster.remote.sslEngine.truststoreType) \"context\" $) }}\"  \n              trust-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_TRUSTSTORE_PASSWORD}\"  \n              hostname-verification = {{ if .Values.cluster.remote.sslEngine.hostnameVerification }}on{{ else }}off{{ end }}\n              require-mutual-authentication = {{ if .Values.cluster.remote.sslEngine.requireMutualAuthentication }}on{{ else }}off{{ end }}  \n            }  \n          }  \n          transport = tls-tcp  \n        }  \n      }\n      {{- end }}\n    }\n    {{- if .Values.cluster.management.ssl.enabled }}\n    xlr-pekko-native {\n      pekko {\n        management {\n          http {  \n            route-providers-read-only = true  \n          }  \n          ssl {  \n            enabled = true\n            keyPassword = \"${?CLUSTER_MANAGEMENT_SSL_KEY_PASSWORD}\"  \n            keyStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.keystore \"default\" (printf \"cluster_management_keystore.%s\" .Values.cluster.management.ssl.keystoreType) \"context\" $) }}\"  \n            keyStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_KEYSTORE_PASSWORD}\"  \n            keystoreType = {{ .Values.cluster.management.ssl.keystoreType }}  \n            trustStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.truststore \"default\" (printf \"cluster_management_truststore.%s\" .Values.cluster.management.ssl.truststoreType) \"context\" $) }}\"  \n            trustStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_TRUSTSTORE_PASSWORD}\"  \n            trustStoreType = {{ .Values.cluster.management.ssl.truststoreType }}  \n          }  \n        }  \n      }\n    }\n    {{- end }}\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\nmanagement.endpoint.prometheus.enabled = {{ .Values.metrics.prometheus.enabled | default false }}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      datacenter = ${XLR_CLUSTER_DATACENTER}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "#!/bin/sh\n\ncd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "livenessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/liveness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "jmx": {
                      "enabled": false
                    },
                    "prometheus": {
                      "enabled": false,
                      "podAnnotations": {
                        "prometheus.io/port": "{{ .Values.service.ports.metrics }}",
                        "prometheus.io/scrape": "true"
                      },
                      "rule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "serviceMonitor": {
                        "annotations": {},
                        "basicAuth": {
                          "password": {
                            "key": "releasePassword",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          },
                          "username": {
                            "key": "releaseUsername",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          }
                        },
                        "enabled": false,
                        "honorLabels": false,
                        "interval": "30s",
                        "jobLabel": "",
                        "labels": {},
                        "metricRelabelings": [],
                        "namespace": "",
                        "params": {},
                        "path": "/s/actuator/prometheus",
                        "podTargetLabels": {},
                        "relabelings": [],
                        "scrapeTimeout": "",
                        "selector": {
                          "service-type": "normal"
                        },
                        "targetLabels": {}
                      }
                    }
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.1-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.4.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.3-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "readinessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {
                      "service-type": "normal"
                    },
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "startupProbe": {
                    "enabled": true,
                    "failureThreshold": 12,
                    "initialDelaySeconds": 60,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 450,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-release",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:838443baf39514c5e0ac0a09cab411942a3cc4a13e261c8c60138788fe8d96e4 '
          createdAt: "2025-09-26T17:28:37Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n
          \ name: dair-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n```\n\n#### Configuration
          Details\n\nThe sample configuration uses:\n\n- Autogenerated limited license\n-
          Simple administrator initial password\n- Default keystore setup\n- Only
          one replica is set for all pods\n- All Pods are missing resource limit setup\n-
          Default storage class and minimal size for persistent storage\n- Release
          pods are using persistence access mode `ReadWriteOnce`, for Release pods
          the requirement is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB
          management\n- Embedded RabbitMQ for message queue management\n\n#### Security
          Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined ranges (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/release-operator@sha256:838443baf39514c5e0ac0a09cab411942a3cc4a13e261c8c60138788fe8d96e4
        - docker.io/xebialabs/xl-release@sha256:9b63789cdb87745832f1283d3f0e28ac5a4514d767895e0914a6be6672f2d14f
        version: 25.1.5
      entries:
      - name: digitalai-release-operator.v25.1.5
        version: 25.1.5
      - name: digitalai-release-operator.v25.1.4
        version: 25.1.4
      - name: digitalai-release-operator.v25.1.3
        version: 25.1.3
      - name: digitalai-release-operator.v25.1.2
        version: 25.1.2
      - name: digitalai-release-operator.v25.1.1
        version: 25.1.1
      - name: digitalai-release-operator.v25.1.0
        version: 25.1.0
      name: stable-v25.1
    defaultChannel: stable-v24.3
    packageName: digitalai-release-operator
    provider:
      name: Digital.ai
      url: https://digital.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Confluent, Inc.
      provider-url: https://confluent.io
    name: confluent-for-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: confluent-for-kubernetes.v2.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.8.4",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.8.4",
                    "init": "confluentinc/confluent-init-container:2.10.3"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:0ca8769b04d6134ee231af84dfe98fb65161148d0d32e85f18dad6a3a345ed33
          createdAt: "2025-10-01T13:12:32Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0 <2.10.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - confluentinc/confluent-operator@sha256:0ca8769b04d6134ee231af84dfe98fb65161148d0d32e85f18dad6a3a345ed33
        - docker.io/confluentinc/cp-kafka-rest@sha256:8218e8db4731219ceaf9dbc378ead9c49aa86f98bc5106f62fbf70958dfd1ee4
        - docker.io/confluentinc/cp-server@sha256:45eb6e4f9c1201202d90216fbbdfcd8655eb378c937f09ac0129cd259956337d
        - docker.io/confluentinc/confluent-operator@sha256:0ca8769b04d6134ee231af84dfe98fb65161148d0d32e85f18dad6a3a345ed33
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:2024d117dd413e207e49f37558d79dc882fd29f6b39e2d722708f18076813361
        - docker.io/confluentinc/cp-ksqldb-server@sha256:093e87359c176d13d81317e972888a835cd01092d061bfc03fd60f0d1b4c8428
        - docker.io/confluentinc/cp-zookeeper@sha256:3e0af6ac6d34dcb8b9ed8f7a08109a48b2b2acd331c1dc4ea875222685c9afbd
        - docker.io/confluentinc/cp-schema-registry@sha256:bba639a3f9e2bf5f99b5d056ed520d2c268a125bee4f546d6cbca7a2c58f7dbb
        - docker.io/confluentinc/cp-server-connect@sha256:24aaa1cc8eb0df643f1efa37999b7ac33d844bfd1e4502da0d678996e4ac5ddd
        - docker.io/confluentinc/confluent-init-container@sha256:175ba68c876eca5804597b1591fdf551b2091d2f1c50e4daed2c32958af5d649
        version: 2.10.3
      entries:
      - name: confluent-for-kubernetes.v2.10.3
        version: 2.10.3
      name: "2.10"
    - currentCSV: confluent-for-kubernetes.v2.11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.9.3",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.9.3",
                    "init": "confluentinc/confluent-init-container:2.11.3"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:85b512d3f20dc6e4a4da1759ee5ed306a4da0604592563a9af35be2cdf9ba1b6
          createdAt: "2025-10-01T13:10:39Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.10.0 <2.11.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-schema-registry@sha256:bd974dda5abf35fcad66c25fad92ee684e82305435b291b953d221f74d2bc119
        - docker.io/confluentinc/cp-server@sha256:e61e6c122b50ed345bc3e4e39cda6c93430e669535f8a70957e221563a7af1bd
        - docker.io/confluentinc/cp-server-connect@sha256:0d3b81e511cdf820daac4a2458b3cada40b4a7c3d142062a48880d38186bbcc4
        - docker.io/confluentinc/confluent-init-container@sha256:bccc7b723afdbe4e04c668720cdb1570835f8fd58ab7dd9a06ecf807f1176019
        - docker.io/confluentinc/cp-ksqldb-server@sha256:14a44cbbc13094974e7a5e14b33a829480ec3da75868130fdc7abaeda707dc5e
        - docker.io/confluentinc/cp-zookeeper@sha256:03eadde2afa390f7339707e6baef2ecf7affb898aaa432378f61da54f5636ba3
        - docker.io/confluentinc/cp-kafka-rest@sha256:cc72b3585ef9e7557910ab3efe3082444abe1fee191cdfb8747907fd2643971f
        - confluentinc/confluent-operator@sha256:85b512d3f20dc6e4a4da1759ee5ed306a4da0604592563a9af35be2cdf9ba1b6
        - docker.io/confluentinc/confluent-operator@sha256:85b512d3f20dc6e4a4da1759ee5ed306a4da0604592563a9af35be2cdf9ba1b6
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:351b68b062824c15db7ef999ba04062defdd4ea1ebe9665932147f99cf74879d
        version: 2.11.3
      entries:
      - name: confluent-for-kubernetes.v2.11.3
        version: 2.11.3
      name: "2.11"
    - currentCSV: confluent-for-kubernetes.v2.8.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.6.4",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.6.4",
                    "init": "confluentinc/confluent-init-container:2.8.5"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:551b14777b121b1e083524a0d5cce03cbe2fd2833537751bcd6be6cf9801f95e
          createdAt: "2024-12-02T06:51:43Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.0 <2.8.5'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-schema-registry@sha256:dec88d30c8f21bc7fb38d28603c97c81c666ccaee63ff0136ecff48b3b3ec775
        - docker.io/confluentinc/confluent-init-container@sha256:b1ad1aa4daa81967a93b02323172f96db6a50c7135866f016344b2eb1a736a55
        - docker.io/confluentinc/cp-ksqldb-server@sha256:50e4834c44759ab4a2a0bbab0f20545e124325b7ee12ba1a275af73b57084104
        - docker.io/confluentinc/confluent-operator@sha256:551b14777b121b1e083524a0d5cce03cbe2fd2833537751bcd6be6cf9801f95e
        - docker.io/confluentinc/cp-zookeeper@sha256:7e728feebfc90a693d028216da68a5d79e5e8212855c63bac08e5b10b141d143
        - docker.io/confluentinc/cp-server-connect@sha256:255875bc57d81c322d47d5f6f11bc217e99ae709d1fdc5aec8e11829dd34279a
        - confluentinc/confluent-operator@sha256:551b14777b121b1e083524a0d5cce03cbe2fd2833537751bcd6be6cf9801f95e
        - docker.io/confluentinc/cp-server@sha256:1b2ae325f87047b83d37d506066f073c27b0c42a0fd3c3409aa079ba736d1975
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:01bcde2aa35c0ad062ece49c63d41bf487618ea3c33e42ea4c23fde7194b26c9
        - docker.io/confluentinc/cp-kafka-rest@sha256:2a5bc0d39693682e25d79f5036e18b71376df0941fedc2cb962a4b4d5a5eff82
        version: 2.8.5
      entries:
      - name: confluent-for-kubernetes.v2.8.5
        version: 2.8.5
      name: "2.8"
    - currentCSV: confluent-for-kubernetes.v2.9.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.7.5",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.7.5",
                    "init": "confluentinc/confluent-init-container:2.9.7"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:20648758fdb98d0f8913e75fb919ad0b622de8ecda9ab394912103e5e17d4007
          createdAt: "2025-10-03T06:52:55Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0 <2.9.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-server-connect@sha256:560833472dbdf23cf85b780e8306ae1bfe16ce6e2c0a9b326c64a513117525f1
        - confluentinc/confluent-operator@sha256:20648758fdb98d0f8913e75fb919ad0b622de8ecda9ab394912103e5e17d4007
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:257c38a75b3f4241780e20f16a7adc026dcca590f2e9c8da7c4e86b961e7611f
        - docker.io/confluentinc/confluent-operator@sha256:20648758fdb98d0f8913e75fb919ad0b622de8ecda9ab394912103e5e17d4007
        - docker.io/confluentinc/cp-server@sha256:79243e7e8b721f9fdf26fc57330ebb27d69c3b4686dbdf5319f6f39c2c7be836
        - docker.io/confluentinc/cp-kafka-rest@sha256:b6ac629a880474d81182eaf8459b4d83fe8126db796f1f285798beecda55a50a
        - docker.io/confluentinc/cp-schema-registry@sha256:6506a8c659d4154ae3e2d702f08b4ccac63513392ba454cbc9cf384d070c46eb
        - docker.io/confluentinc/cp-ksqldb-server@sha256:6817a555ae4c041869f9b422a115c22894b2cd3988e82dc601694ce713b3ccd4
        - docker.io/confluentinc/cp-zookeeper@sha256:2bd46d1f5d1eec0675bb5d1c08634d6458bbac7cacfb233aa61c68137fd4a7ce
        - docker.io/confluentinc/confluent-init-container@sha256:b138d954318cc98767c10b4564a1412a22766317a83b9a452285bb0b3d893c63
        version: 2.9.7
      entries:
      - name: confluent-for-kubernetes.v2.9.7
        version: 2.9.7
      name: "2.9"
    - currentCSV: confluent-for-kubernetes.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:8.0.1",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-ng-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "alertManagerClient": {
                      "url": "http://controlcenter-next-gen.confluent.svc.cluster.local:9093"
                    },
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    },
                    "prometheusClient": {
                      "url": "http://controlcenter-next-gen.confluent.svc.cluster.local:9090"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center-next-gen:2.2.0",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "replicas": 1,
                  "services": {
                    "alertmanager": {
                      "image": "confluentinc/cp-enterprise-alertmanager:2.2.0"
                    },
                    "prometheus": {
                      "image": "confluentinc/cp-enterprise-prometheus:2.2.0",
                      "pvc": {
                        "dataVolumeCapacity": "10Gi"
                      }
                    }
                  },
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:8.0.1",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:8.0.1",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:8.0.1",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:8.0.1",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:8.0.1",
                    "init": "confluentinc/confluent-init-container:3.0.1"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:55b61717690fe334b5eebf5e7a55590537abddbd404f2eede81d4dd8cd29eda4
          createdAt: "2025-10-03T07:59:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <3.0.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-schema-registry@sha256:2152975dfb62afbdf1d438dd0eb1f7fc3c929adf5804fe4790eccda0c96b9208
        - docker.io/confluentinc/cp-enterprise-alertmanager@sha256:fcce6f5311c26cd58427a039530a9f8dfd009249ff5a519694aa945d34502252
        - docker.io/confluentinc/cp-ksqldb-server@sha256:df94bb805cff4f92315135a042f650c42f887bca70a40df8ecef2603532f11a7
        - docker.io/confluentinc/cp-server@sha256:207cabc12e0d122199da875b2b0e39d1cf6f7fe435d5f184a9c8d7dc57ea81d9
        - docker.io/confluentinc/cp-kafka-rest@sha256:f45f23e891ae073037ed7670659ecb6c3f72f9eea33e109f16243eec24a287a3
        - docker.io/confluentinc/confluent-operator@sha256:55b61717690fe334b5eebf5e7a55590537abddbd404f2eede81d4dd8cd29eda4
        - confluentinc/confluent-operator@sha256:55b61717690fe334b5eebf5e7a55590537abddbd404f2eede81d4dd8cd29eda4
        - docker.io/confluentinc/cp-server-connect@sha256:4739f3a07d9f8eadd58f448a8e05b1392ef9b1dc9e9085d25a71f32b5b806a6e
        - docker.io/confluentinc/cp-enterprise-control-center-next-gen@sha256:6378a0174433685ee52a7fe2e48424d3f598827ed2a5570ec74b916102485e0d
        - docker.io/confluentinc/confluent-init-container@sha256:bf38b3bd9851bfd3ef0635e93b9e8b9ada570d393f650f60d8ae8943537730dc
        - docker.io/confluentinc/cp-enterprise-prometheus@sha256:db0989ef9859dbc88bc59f9b8ade86f2ba13e9dc36533e654566218328e5ff57
        version: 3.0.1
      entries:
      - name: confluent-for-kubernetes.v3.0.1
        version: 3.0.1
      name: "3.0"
    defaultChannel: "3.0"
    packageName: confluent-for-kubernetes
    provider:
      name: Confluent, Inc.
      url: https://confluent.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pachyderm, Inc.
      provider-url: https://pachyderm.com
    name: pachyderm-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pachyderm-operator.v0.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "Pachyderm",
                "metadata": {
                  "name": "pachyderm-sample"
                },
                "spec": {
                  "console": {
                    "disable": true
                  },
                  "pachd": {
                    "metrics": {
                      "disable": false
                    },
                    "storage": {
                      "amazon": {
                        "credentialSecretName": "pachyderm-aws-secret"
                      },
                      "backend": "AMAZON"
                    }
                  },
                  "version": "v2.1.6"
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermExport",
                "metadata": {
                  "name": "pachydermexport-sample"
                },
                "spec": {
                  "backup": {
                    "target": "pachyderm-sample"
                  }
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermImport",
                "metadata": {
                  "name": "pachydermimport-sample"
                },
                "spec": {
                  "backup": "backups/pachyderm-backup-20220610.tar.gz",
                  "destination": {
                    "name": "pachyderm-restore",
                    "namespace": "pachyderm-workloads"
                  },
                  "storageSecret": "s3-credentials-secret"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
          createdAt: "2021-05-20T08:05:00Z"
          description: Pachyderm provides the data layer that allows data science
            teams to productionize and scale their machine learning lifecycle with
            data driven automation, petabyte scalability and end-to-end reproducibility.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/pachyderm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/pachyderm-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pachyderm/openshift-operator
          support: Pachyderm, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PachydermExport is the Schema for the pachydermexports API
            displayName: Pachyderm Export
            kind: PachydermExport
            name: pachydermexports.aiml.pachyderm.com
            version: v1beta1
          - description: PachydermImport is the Schema for the pachydermimports API
            displayName: Pachyderm Import
            kind: PachydermImport
            name: pachydermimports.aiml.pachyderm.com
            version: v1beta1
          - description: Pachyderm is the Schema for the pachyderms API
            displayName: Pachyderm
            kind: Pachyderm
            name: pachyderms.aiml.pachyderm.com
            version: v1beta1
        description: "The Pachyderm operator aims to make deploying and managing Pachyderm
          instances on Openshift container platform easier.\n\nPachyderm is the data
          foundation for machine learning.  Pachyderm provides industry leading data
          versioning, pipelines and lineage that allow data science teams to automate
          the machine learning lifecycle and optimize their machine learning operations
          (MLOps). \n\n## Getting started\n\n1. Pachyderm uses an object store to
          store your data. Set the pachd storage backend at `pachyderm.spec.pachd.storage.backend`
          to `AMAZON`, `MICROSOFT`, `GOOGLE` or `MINIO` depending on your storage
          provider. Additionally, create a user with read/write access to the object
          store.\n\n2. Setup the user credentials in the section, `pachyderm.spec.pachd.storage`.
          Required fields depend on the backend type set in step 1 above.\n\n3. Finally,
          deploy the pachyderm cluster in its own namespace.\n\nFor more information
          on getting started, reference the documentation.\n"
        displayName: Pachyderm
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pachyderm
        - versioning
        - machine learning
        - artificial intelligence
        links:
        - name: Documentation
          url: https://docs.pachyderm.com/latest/
        maintainers:
        - email: kyle@pachyderm.io
          name: Kyle White
        - email: eochieng@redhat.com
          name: Edmund Ochieng
        maturity: beta
        minKubeVersion: v1.19.0
        provider:
          name: Pachyderm, Inc.
          url: https://pachyderm.com
        relatedImages:
        - registry.connect.redhat.com/pachyderm/worker@sha256:374f83d92899707bc87b0318fda6bffa78d2e9e523b827bb3d40693a2d1bdecd
        - registry.connect.redhat.com/pachyderm/etcd@sha256:7f330b51874379b1d4b71d94e37e0ecfa52b86d628296136970688cf204f53b5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
        - registry.connect.redhat.com/pachyderm/backup-handler@sha256:e2ef6bd8191e8ad58c765c0db8f5bb21ea7123caed0c0f58fe681a024201306d
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/pachd@sha256:116477fefdf413ea02c6ac3d013f1f91d3f6e4b16aed2408c9890832cdb41430
        - registry.redhat.io/rhel8/postgresql-13@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/init-utils@sha256:5c0258bfa0203079925648c0a7dda6d05a29b6f461854821a1eec541777ff359
        - registry.connect.redhat.com/pachyderm/pgbouncer@sha256:c3b6949227df24abb8934d111a34b44e87fb16194ac34a5d0fe284cff02e9ac0
        version: 0.1.4
      entries:
      - name: pachyderm-operator.v0.1.4
        version: 0.1.4
      name: stable
    defaultChannel: stable
    packageName: pachyderm-operator-rhmp
    provider:
      name: Pachyderm, Inc.
      url: https://pachyderm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: local-storage-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: local-storage-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "local.storage.openshift.io/v1",
                "kind": "LocalVolume",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "storageClassDevices": [
                    {
                      "devicePaths": [
                        "/dev/disk/by-id/ata-crucial"
                      ],
                      "fsType": "ext4",
                      "storageClassName": "foobar",
                      "volumeMode": "Filesystem"
                    }
                  ]
                }
              },
              {
                "apiVersion": "local.storage.openshift.io/v1alpha1",
                "kind": "LocalVolumeSet",
                "metadata": {
                  "name": "example-localvolumeset"
                },
                "spec": {
                  "deviceInclusionSpec": {
                    "deviceMechanicalProperties": [
                      "Rotational",
                      "NonRotational"
                    ],
                    "deviceTypes": [
                      "disk"
                    ],
                    "maxSize": "100G",
                    "minSize": "10G"
                  },
                  "maxDeviceCount": 10,
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "storageClassName": "example-storageclass",
                  "volumeMode": "Block"
                }
              },
              {
                "apiVersion": "local.storage.openshift.io/v1alpha1",
                "kind": "LocalVolumeDiscovery",
                "metadata": {
                  "name": "auto-discover-devices"
                },
                "spec": {
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-local-storage-rhel9-operator@sha256:6f6491f023ffa21eff6a9b938cb1869f362d6b4a5b04c27dc5f13f49622ca178
          createdAt: "2019-08-14T00:00:00Z"
          description: Configure and use local storage volumes.
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.17"}]'
          olm.skipRange: '>=4.3.0 <4.16.0-202509240123'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-local-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["localvolumediscoveryresults.local.storage.openshift.io"]'
          repository: https://github.com/openshift/local-storage-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manage local storage volumes for OpenShift
            displayName: Local Volume
            kind: LocalVolume
            name: localvolumes.local.storage.openshift.io
            version: v1
          - description: A Local Volume set allows you to filter a set of storage
              volumes, group them and create a dedicated storage class to consume
              storage from the set of volumes.
            displayName: Local Volume Set
            kind: LocalVolumeSet
            name: localvolumesets.local.storage.openshift.io
            version: v1alpha1
          - description: Discover list of potentially usable disks on the chosen set
              of nodes
            displayName: Local Volume Discovery
            kind: LocalVolumeDiscovery
            name: localvolumediscoveries.local.storage.openshift.io
            version: v1alpha1
          - description: Disc inventory of available disks from selected nodes
            displayName: Local Volume Discovery Result
            kind: LocalVolumeDiscoveryResult
            name: localvolumediscoveryresults.local.storage.openshift.io
            version: v1alpha1
        description: |
          Operator that configures local storage volumes for use in Kubernetes and OpenShift. OpenShift 4.2 and above are the only supported OpenShift versions.
        displayName: Local Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - local storage
        links:
        - name: Documentation
          url: https://github.com/openshift/local-storage-operator/tree/master/docs
        - name: Source Repository
          url: https://github.com/openshift/local-storage-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-local-storage-rhel9-operator@sha256:6f6491f023ffa21eff6a9b938cb1869f362d6b4a5b04c27dc5f13f49622ca178
        - registry.redhat.io/openshift4/ose-local-storage-diskmaker-rhel9@sha256:d8d8b21fce5930a36c9755d82037fce1ab81f5ae2d5fb1bfd643d500c4ec7fba
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel9@sha256:ec5847d5006d66a014315decd579e10104b086a06bfefa33430545b4ffcece71
        version: 4.16.0-202509240123
      entries:
      - name: local-storage-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: local-storage-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: local-storage-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: local-storage-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: local-storage-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: local-storage-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: local-storage-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: local-storage-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: local-storage-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: local-storage-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: local-storage-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: local-storage-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: local-storage-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: local-storage-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: local-storage-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: local-storage-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: local-storage-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: local-storage-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: local-storage-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: local-storage-operator.v4.16.0-202410031007
        version: 4.16.0-202410031007
      - name: local-storage-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: local-storage-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: local-storage-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: local-storage-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: local-storage-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: local-storage-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: local-storage-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: local-storage-operator.v4.16.0-202408021139
        version: 4.16.0-202408021139
      - name: local-storage-operator.v4.16.0-202407190036
        version: 4.16.0-202407190036
      - name: local-storage-operator.v4.16.0-202407111006
        version: 4.16.0-202407111006
      - name: local-storage-operator.v4.16.0-202407081937
        version: 4.16.0-202407081937
      - name: local-storage-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: local-storage-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: local-storage-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: local-storage-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: local-storage-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: security-profiles-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: security-profiles-operator.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "AppArmorProfile",
                "metadata": {
                  "annotations": {
                    "description": "Block writing to any files in the disk."
                  },
                  "name": "test-profile"
                },
                "spec": {
                  "abstract": {
                    "filesystem": {
                      "readOnlyPaths": [],
                      "readWritePaths": [],
                      "writeOnlyPaths": []
                    }
                  }
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "SecurityProfilesOperatorDaemon",
                "metadata": {
                  "name": "spod",
                  "namespace": "security-profiles-operator"
                },
                "spec": {
                  "enableAppArmor": false,
                  "enableLogEnricher": false,
                  "enableSelinux": false
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "ProfileRecording",
                "metadata": {
                  "name": "test-recording"
                },
                "spec": {
                  "kind": "SeccompProfile",
                  "podSelector": {
                    "matchLabels": {
                      "app": "alpine"
                    }
                  },
                  "recorder": "bpf"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "ProfileBinding",
                "metadata": {
                  "name": "profile-binding"
                },
                "spec": {
                  "image": "nginx:1.19.1",
                  "profileRef": {
                    "kind": "SeccompProfile",
                    "name": "profile-allow-unsafe"
                  }
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha2",
                "kind": "RawSelinuxProfile",
                "metadata": {
                  "name": "errorlogger"
                },
                "spec": {
                  "policy": "(blockinherit container)\n(allow process var_log_t ( dir ( open read getattr lock search ioctl add_name remove_name write ))) \n(allow process var_log_t ( file ( getattr read write append ioctl lock map open create  ))) \n(allow process var_log_t ( sock_file ( getattr read write append open  ))) \n"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "Blocks all syscalls."
                  },
                  "name": "profile-block-all"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_ERRNO"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "UNSAFE: Allows all syscalls whilst logging their use. Similar to running as unconfined in terms of enforcement."
                  },
                  "name": "profile-complain-unsafe"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_LOG"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "UNSAFE: Allows all syscalls. Similar to running as unconfined as it provides no enforcement."
                  },
                  "name": "profile-allow-unsafe"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_ALLOW"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "Enables complain mode whilst blocking high-risk syscalls. Some essential syscalls are allowed to decrease log noise."
                  },
                  "name": "profile-complain-block-high-risk"
                },
                "spec": {
                  "architectures": [
                    "SCMP_ARCH_X86_64",
                    "SCMP_ARCH_PPC64LE"
                  ],
                  "defaultAction": "SCMP_ACT_LOG",
                  "syscalls": [
                    {
                      "action": "SCMP_ACT_ALLOW",
                      "names": [
                        "exit",
                        "exit_group",
                        "futex",
                        "nanosleep"
                      ]
                    },
                    {
                      "action": "SCMP_ACT_ERRNO",
                      "names": [
                        "acct",
                        "add_key",
                        "bpf",
                        "clock_adjtime",
                        "clock_settime",
                        "create_module",
                        "delete_module",
                        "finit_module",
                        "get_kernel_syms",
                        "get_mempolicy",
                        "init_module",
                        "ioperm",
                        "iopl",
                        "kcmp",
                        "kexec_file_load",
                        "kexec_load",
                        "keyctl",
                        "lookup_dcookie",
                        "mbind",
                        "mount",
                        "move_pages",
                        "name_to_handle_at",
                        "nfsservctl",
                        "open_by_handle_at",
                        "perf_event_open",
                        "personality",
                        "pivot_root",
                        "process_vm_readv",
                        "process_vm_writev",
                        "ptrace",
                        "query_module",
                        "quotactl",
                        "reboot",
                        "request_key",
                        "set_mempolicy",
                        "setns",
                        "settimeofday",
                        "stime",
                        "swapoff",
                        "swapon",
                        "_sysctl",
                        "sysfs",
                        "umount2",
                        "umount",
                        "unshare",
                        "uselib",
                        "userfaultfd",
                        "ustat",
                        "vm86old",
                        "vm86"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha2",
                "kind": "SelinuxProfile",
                "metadata": {
                  "name": "errorlogger"
                },
                "spec": {
                  "allow": {
                    "var_log_t": {
                      "dir": [
                        "open",
                        "read",
                        "getattr",
                        "lock",
                        "search",
                        "ioctl",
                        "add_name",
                        "remove_name",
                        "write"
                      ],
                      "file": [
                        "getattr",
                        "read",
                        "write",
                        "append",
                        "ioctl",
                        "lock",
                        "map",
                        "open",
                        "create"
                      ],
                      "sock_file": [
                        "getattr",
                        "read",
                        "write",
                        "append",
                        "open"
                      ]
                    }
                  },
                  "inherit": [
                    {
                      "name": "container"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          createdAt: "2025-02-25T13:34:54Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.1 <0.9.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-security-profiles
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ProfileBinding is the Schema for the profilebindings API.
            displayName: Profile Binding
            kind: ProfileBinding
            name: profilebindings.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: ProfileRecording is the Schema for the profilerecordings
              API.
            displayName: Profile Recording
            kind: ProfileRecording
            name: profilerecordings.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: SeccompProfile is a cluster level specification for a seccomp
              profile. See https://github.com/opencontainers/runtime-spec/blob/master/config-linux.md#seccomp
            displayName: Seccomp Profile
            kind: SeccompProfile
            name: seccompprofiles.security-profiles-operator.x-k8s.io
            version: v1beta1
          - description: SecurityProfileNodeStatus is a per-node status of a security
              profile
            displayName: Security Profile Node Status
            kind: SecurityProfileNodeStatus
            name: securityprofilenodestatuses.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: SecurityProfilesOperatorDaemon is the Schema to configure
              the spod deployment.
            displayName: Security Profiles Operator Daemon
            kind: SecurityProfilesOperatorDaemon
            name: securityprofilesoperatordaemons.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - kind: RawSelinuxProfile
            name: rawselinuxprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha2
          - description: SelinuxProfile is the Schema for the selinuxprofiles API.
            displayName: Selinux Profile
            kind: SelinuxProfile
            name: selinuxprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha2
          - description: AppArmorProfile is a cluster level specification for an AppArmor
              profile.
            displayName: App Armor Profile
            kind: AppArmorProfile
            name: apparmorprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha1
        description: SPO is an operator which aims to make it easier for users to
          use SELinux, seccomp and AppArmor in Kubernetes clusters
        displayName: Security Profiles Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - selinux
        - seccomp
        - apparmor
        - ebpf
        links:
        - name: Security Profiles Operator
          url: https://github.com/openshift/security-profiles-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-security-profiles-rhel8-operator@sha256:828274d71c65ef820e6a8b652e3f102624d09a0609fb00bae7b32dc62ff37fd1
        - registry.redhat.io/compliance/openshift-selinuxd-rhel9@sha256:3278d69814a8d0d1e31dcdef454c24828aadbb468a3d5d3e15910c76481765bd
        - registry.redhat.io/compliance/openshift-selinuxd-rhel8@sha256:81273ff149482109417064b40d333edb131481c05547e4ffa6485475460a48ba
        version: 0.9.0
      entries:
      - name: security-profiles-operator.v0.9.0
        version: 0.9.0
      - name: security-profiles-operator.v0.8.6
        version: 0.8.6
      - name: security-profiles-operator.v0.8.5
        version: 0.8.5
      - name: security-profiles-operator.v0.8.4
        version: 0.8.4
      - name: security-profiles-operator.v0.8.2
        version: 0.8.2
      name: release-alpha-rhel-8
    defaultChannel: release-alpha-rhel-8
    packageName: security-profiles-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: prometheus-exporter-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometheus-exporter-operator.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.3scale.net/v1alpha1",
                "kind": "PrometheusExporter",
                "metadata": {
                  "name": "example-memcached"
                },
                "spec": {
                  "dbHost": "your-memcached-host",
                  "dbPort": 11211,
                  "grafanaDashboard": {
                    "label": {
                      "key": "autodiscovery",
                      "value": "enabled"
                    }
                  },
                  "type": "memcached"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/3scale/prometheus-exporter-operator:v0.7.0
          createdAt: "2023-09-07 00:00:00"
          description: Operator to setup 3rd party prometheus exporters, with a collection
            of grafana dashboards
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/3scale-ops/prometheus-exporter-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configures a prometheus exporter to monitor a memcached instance
            displayName: PrometheusExporter
            kind: PrometheusExporter
            name: prometheusexporters.monitoring.3scale.net
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK to centralize the setup of 3rd party prometheus exporters on **Kubernetes/OpenShift**, with a collection of grafana dashboards.

          By just providing a few parameters like *dbHost* or *dbPort* (operator manage the container image, port, argument, command, volumes... and also prometheus `ServiceMonitor` and `GrafanaDashboard` k8s objects), you can setup different prometheus exporters to monitor:
          * The **internals from different databases**
          * **HTTP/TCP endpoints** (availability, latency, SSL/TLS certificate expiration...)
          * Any available **cloudwatch metric from any AWS Service**
          * Sendgrid email statistics (delivered, bounces, errors, spam...)

          Current prometheus exporters types supported, managed by same prometheus-exporter-operator:
          * memcached
          * redis
          * mysql
          * postgresql
          * sphinx
          * es (elasticsearch)
          * cloudwatch
          * probe (blackbox)
          * sendgrid

          The operator manages the lifecycle of the following objects:
          * Deployment (one per CR)
          * Service (one per CR)
          * ServiceMonitor (optional, one per CR)
          * GrafanaDashboard (optional, one per Namespace)

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale-ops/prometheus-exporter-operator#documentation).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale-ops/prometheus-exporter-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:
          * Raising any issues you find using Prometheus Exporter Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale-ops/prometheus-exporter-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving [documentation](https://github.com/3scale-ops/prometheus-exporter-operator)
          * Talking about Prometheus Exporter Operator
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale-ops/prometheus-exporter-operator/issues).

          ### License
          Prometheus Exporter Operator is licensed under the [Apache 2.0 license](https://github.com/3scale-ops/prometheus-exporter-operator/blob/main/LICENSE)
        displayName: Prometheus Exporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - metrics
        - observability
        - prometheus
        - exporter
        - grafana
        - mysql
        - postgresql
        - memcached
        - redis
        - sphinx
        - elasticsearch
        - cloudwatch
        - blackbox
        - sendgrid
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/prometheus-exporter-operator
        maintainers:
        - email: 3scale-operations@redhat.com
          name: 3scale Ops
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/3scale/prometheus-exporter-operator:v0.7.0
        version: 0.7.0
      entries:
      - name: prometheus-exporter-operator.v0.7.0
        version: 0.7.0
      - name: prometheus-exporter-operator.v0.3.4
        version: 0.3.4
      name: alpha
    defaultChannel: alpha
    packageName: prometheus-exporter-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: mariadb-operator
      provider-url: https://github.com/mariadb-operator/mariadb-operator
    name: mariadb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-operator.v25.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.mariadb.mmontes.io/v1alpha1",
                "kind": "MariadbOperator",
                "metadata": {
                  "name": "mariadb-operator"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "caValidity": "35064h",
                    "certValidity": "8766h",
                    "enabled": true,
                    "extrArgs": [],
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "ha": {
                      "enabled": false,
                      "replicas": 3
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "lookaheadValidity": "2160h",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podSecurityContext": {},
                    "requeueDuration": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "automount": true,
                      "enabled": true,
                      "extraLabels": {},
                      "name": ""
                    },
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    },
                    "tolerations": []
                  },
                  "clusterName": "cluster.local",
                  "extrArgs": [],
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "fullnameOverride": "",
                  "ha": {
                    "enabled": false,
                    "replicas": 3
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "logLevel": "INFO",
                  "metrics": {
                    "enabled": false,
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "rbac": {
                    "enabled": true
                  },
                  "resources": {},
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "automount": true,
                    "enabled": true,
                    "extraLabels": {},
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "annotations": {},
                    "cert": {
                      "caPath": "/tmp/k8s-webhook-server/certificate-authority",
                      "certManager": {
                        "duration": "",
                        "enabled": false,
                        "issuerRef": {},
                        "renewBefore": ""
                      },
                      "path": "/tmp/k8s-webhook-server/serving-certs",
                      "secretAnnotations": {}
                    },
                    "extrArgs": [],
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "ha": {
                      "enabled": false,
                      "replicas": 3
                    },
                    "hostNetwork": false,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podSecurityContext": {},
                    "port": 10250,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "automount": true,
                      "enabled": true,
                      "extraLabels": {},
                      "name": ""
                    },
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    },
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup"
                },
                "spec": {
                  "args": [
                    "--single-transaction",
                    "--all-databases"
                  ],
                  "logLevel": "info",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "backups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Connection",
                "metadata": {
                  "name": "connection"
                },
                "spec": {
                  "database": "mariadb",
                  "healthCheck": {
                    "interval": "10s",
                    "retryInterval": "3s"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "params": {
                    "parseTime": "true"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "secretName": "connection",
                  "secretTemplate": {
                    "annotations": {
                      "k8s.mariadb.com/connection": "sample"
                    },
                    "databaseKey": "database",
                    "hostKey": "host",
                    "key": "dsn",
                    "labels": {
                      "k8s.mariadb.com/connection": "sample"
                    },
                    "passwordKey": "password",
                    "portKey": "port",
                    "usernameKey": "username"
                  },
                  "serviceName": "mariadb",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Database",
                "metadata": {
                  "name": "data-test"
                },
                "spec": {
                  "characterSet": "utf8",
                  "collate": "utf8_general_ci",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "retryInterval": "5s"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "ExternalMariaDB",
                "metadata": {
                  "name": "external-mariadb"
                },
                "spec": {
                  "connection": {
                    "healthCheck": {
                      "interval": "5s"
                    },
                    "secretName": "external-mariadb"
                  },
                  "host": "mariadb.default.svc.cluster.local",
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "port": 3306,
                  "username": "root"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "grant"
                },
                "spec": {
                  "database": "*",
                  "grantOption": true,
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "privileges": [
                    "SELECT",
                    "INSERT",
                    "UPDATE"
                  ],
                  "requeueInterval": "30s",
                  "retryInterval": "5s",
                  "table": "*",
                  "username": "user"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "MariaDB",
                "metadata": {
                  "name": "mariadb"
                },
                "spec": {
                  "connection": {
                    "secretName": "mariadb-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "database": "mariadb",
                  "galera": {
                    "enabled": true
                  },
                  "metrics": {
                    "enabled": true,
                    "passwordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "mariadb-metrics"
                    }
                  },
                  "myCnf": "[mariadb]\nbind-address=*\ndefault_storage_engine=InnoDB\nbinlog_format=row\ninnodb_autoinc_lock_mode=2\ninnodb_buffer_pool_size=1024M\nmax_allowed_packet=256M\n",
                  "passwordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-password"
                  },
                  "primaryConnection": {
                    "secretName": "mariadb-conn-primary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "primaryService": {
                    "type": "ClusterIP"
                  },
                  "replicas": 3,
                  "rootPasswordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-root"
                  },
                  "secondaryConnection": {
                    "secretName": "mariadb-conn-secondary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "secondaryService": {
                    "type": "ClusterIP"
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "storage": {
                    "size": "1Gi"
                  },
                  "updateStrategy": {
                    "type": "ReplicasFirstPrimaryLast"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "MaxScale",
                "metadata": {
                  "name": "maxscale-galera"
                },
                "spec": {
                  "admin": {
                    "guiEnabled": true,
                    "port": 8989
                  },
                  "auth": {
                    "generate": true
                  },
                  "config": {
                    "sync": {
                      "database": "mysql",
                      "interval": "5s",
                      "timeout": "10s"
                    }
                  },
                  "connection": {
                    "port": 3306,
                    "secretName": "mxs-galera-conn"
                  },
                  "guiKubernetesService": {
                    "metadata": {
                      "annotations": {
                        "metallb.universe.tf/loadBalancerIPs": "172.18.0.231"
                      }
                    },
                    "type": "LoadBalancer"
                  },
                  "kubernetesService": {
                    "annotations": {
                      "metallb.universe.tf/loadBalancerIPs": "172.18.0.224"
                    },
                    "type": "LoadBalancer"
                  },
                  "mariaDbRef": {
                    "name": "mariadb-galera"
                  },
                  "monitor": {
                    "cooperativeMonitoring": "majority_of_all",
                    "interval": "2s",
                    "params": {
                      "available_when_donor": "false",
                      "disable_master_failback": "false",
                      "disable_master_role_setting": "false"
                    }
                  },
                  "replicas": 3,
                  "requeueInterval": "10s",
                  "services": [
                    {
                      "listener": {
                        "params": {
                          "connection_metadata": "tx_isolation=auto"
                        },
                        "port": 3306,
                        "protocol": "MariaDBProtocol"
                      },
                      "name": "rw-router",
                      "params": {
                        "master_accept_reads": "true",
                        "max_replication_lag": "3s",
                        "max_slave_connections": "255",
                        "transaction_replay": "true",
                        "transaction_replay_attempts": "10",
                        "transaction_replay_timeout": "5s"
                      },
                      "router": "readwritesplit"
                    },
                    {
                      "listener": {
                        "port": 3307
                      },
                      "name": "rconn-master-router",
                      "params": {
                        "master_accept_reads": "true",
                        "max_replication_lag": "3s",
                        "router_options": "master"
                      },
                      "router": "readconnroute"
                    },
                    {
                      "listener": {
                        "port": 3308
                      },
                      "name": "rconn-slave-router",
                      "params": {
                        "max_replication_lag": "3s",
                        "router_options": "slave"
                      },
                      "router": "readconnroute"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "PhysicalBackup",
                "metadata": {
                  "name": "physicalbackup"
                },
                "spec": {
                  "compression": "bzip2",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "podAffinity": true,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "immediate": true,
                    "suspend": false
                  },
                  "serviceAccountName": "backup",
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "physicalbackups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "region": "us-east-1",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  },
                  "timeout": "1h"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore"
                },
                "spec": {
                  "backupRef": {
                    "name": "backup"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "targetRecoveryTime": "2023-12-19T09:00:00Z"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "SqlJob",
                "metadata": {
                  "name": "03-stars"
                },
                "spec": {
                  "database": "mariadb",
                  "dependsOn": [
                    {
                      "name": "01-users"
                    },
                    {
                      "name": "02-repos"
                    }
                  ],
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "sql": "CREATE TABLE IF NOT EXISTS stars (\n  id bigint PRIMARY KEY AUTO_INCREMENT,\n  user_id bigint NOT NULL,\n  repo_id bigint NOT NULL,\n  FOREIGN KEY (user_id) REFERENCES users(id) ON DELETE CASCADE,\n  FOREIGN KEY (repo_id) REFERENCES repos(id) ON DELETE CASCADE,\n  UNIQUE KEY (user_id, repo_id)\n);\nINSERT INTO stars(user_id, repo_id) \n  VALUES((SELECT id FROM users ORDER BY RAND() LIMIT 1), (SELECT id FROM repos ORDER BY RAND() LIMIT 1))\n  ON DUPLICATE KEY UPDATE id=id;\nDELETE FROM stars WHERE id = (SELECT id FROM stars ORDER BY RAND() LIMIT 1);\nSELECT r.name AS repo, COUNT(*) AS stars\nFROM stars s\nJOIN repos r\nON s.repo_id = r.id\nGROUP BY r.id\nORDER BY stars DESC;",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "user"
                },
                "spec": {
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxUserConnections": 20,
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "retryInterval": "5s"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: ghcr.io/mariadb-operator/mariadb-operator-helm:25.8.4
          createdAt: "2025-09-19T13:24:35Z"
          description: Run and operate MariaDB in a cloud native way
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mariadb-operator/mariadb-operator
          support: mariadb-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configures a backup
            displayName: Backup
            kind: Backup
            name: backups.k8s.mariadb.com
            version: v1alpha1
          - description: Configures a connection
            displayName: Connection
            kind: Connection
            name: connections.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a logical database
            displayName: Database
            kind: Database
            name: databases.k8s.mariadb.com
            version: v1alpha1
          - description: Configures an external MariaDB resources
            displayName: ExternalMariaDB
            kind: ExternalMariaDB
            name: externalmariadbs.k8s.mariadb.com
            version: v1alpha1
          - description: Grants permissions to an user in a database
            displayName: Grant
            kind: Grant
            name: grants.k8s.mariadb.com
            version: v1alpha1
          - description: Configures MariaDB helm chart based operator
            displayName: MariadbOperator
            kind: MariadbOperator
            name: mariadboperators.helm.mariadb.mmontes.io
            version: v1alpha1
          - description: Provisions a MariaDB instance
            displayName: MariaDB
            kind: MariaDB
            name: mariadbs.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a MaxScale database proxy
            displayName: MaxScale
            kind: MaxScale
            name: maxscales.k8s.mariadb.com
            version: v1alpha1
          - description: Configures a physical backup
            displayName: PhysicalBackup
            kind: PhysicalBackup
            name: physicalbackups.k8s.mariadb.com
            version: v1alpha1
          - description: Restores a backup
            displayName: Restore
            kind: Restore
            name: restores.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a SQL job
            displayName: SqlJob
            kind: SqlJob
            name: sqljobs.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a user
            displayName: User
            kind: User
            name: users.k8s.mariadb.com
            version: v1alpha1
        description: |
          Install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) via [OLM](https://olm.operatorframework.io/) using the [helm chart](https://artifacthub.io/packages/helm/mariadb-operator/mariadb-operator).

          This helm operator provides provides a 1:1 mapping between the official helm chart and the [`MariadbOperator`](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/config/samples/helm_v1alpha1_mariadboperator.yaml) CRD, allowing to install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) via OLM without having to do any change in the helm chart.

          Normally, you would install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) providing this `values.yaml` to the helm chart:
          ```yaml
          image:
            repository: ghcr.io/mariadb-operator/mariadb-operator
            pullPolicy: IfNotPresent
          logLevel: INFO
          ha:
            enabled: true
          metrics:
            enabled: true
            serviceMonitor:
              enabled: true
          webhook:
            cert:
              certManager:
                enabled: true
          ```

          This helm chart installation is abstracted in the [`MariadbOperator`](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/config/samples/helm_v1alpha1_mariadboperator.yaml) CRD, which will be reconciled by the helm operator:
          ```yaml
          apiVersion: helm.k8s.mariadb.com/v1alpha1
          kind: MariadbOperator
          metadata:
            name: mariadb-operator
          spec:
            image:
              repository: ghcr.io/mariadb-operator/mariadb-operator
              pullPolicy: IfNotPresent
            logLevel: INFO
            ha:
              enabled: true
            metrics:
              enabled: true
              serviceMonitor:
                enabled: true
            webhook:
              cert:
                certManager:
                  enabled: true
          ```

          Once you have installed the operator, you will able to install a [`MariaDB`](https://github.com/mariadb-operator/mariadb-operator/blob/main/examples/manifests/mariadb_v1alpha1_mariadb.yaml) instance. Refer to the main [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) documentation for getting started with the rest of CRDs.

          ## Documentation
          * [mariadb-operator](https://github.com/mariadb-operator/mariadb-operator/blob/main/README.md)
          * [mariadb-operator-helm](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/README.md)

          ## Releases
          This operator is automatically published in the following repositories whenever a new version of the [helm chart](https://artifacthub.io/packages/helm/mariadb-operator/mariadb-operator) is released:
          - [k8s-operatorhub/community-operators](https://github.com/k8s-operatorhub/community-operators)
          - [redhat-openshift-ecosystem/community-operators-prod](https://github.com/redhat-openshift-ecosystem/community-operators-prod)

          ## Roadmap
          Take a look at our [roadmap](https://github.com/mariadb-operator/mariadb-operator/blob/main/ROADMAP.md) and feel free to open an issue to suggest new features.

          ## Contributing
          We welcome and encourage contributions to this project! Please check our [contributing](https://github.com/mariadb-operator/mariadb-operator/blob/main/CONTRIBUTING.md) and [development](https://github.com/mariadb-operator/mariadb-operator/blob/main/docs/DEVELOPMENT.md) guides. PRs welcome!

          ## Get in touch
          Join us on Slack: **[MariaDB Community Slack](https://r.mariadb.com/join-community-slack)**.
        displayName: MariaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mariadb
        - mysql
        - operator
        - mariadb-operator
        - database
        - maxscale
        links:
        - name: GitHub
          url: https://github.com/mariadb-operator/mariadb-operator
        maintainers:
        - email: mariadb-operator@proton.me
          name: mmontes11
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: mariadb-operator
          url: https://github.com/mariadb-operator/mariadb-operator
        relatedImages:
        - ghcr.io/mariadb-operator/mariadb-operator-helm:25.8.4
        version: 25.8.4
      entries:
      - name: mariadb-operator.v25.8.4
        version: 25.8.4
      - name: mariadb-operator.v25.8.3
        version: 25.8.3
      - name: mariadb-operator.v25.8.2
        version: 25.8.2
      - name: mariadb-operator.v25.8.1
        version: 25.8.1
      - name: mariadb-operator.v25.8.0
        version: 25.8.0
      - name: mariadb-operator.v0.38.1
        version: 0.38.1
      - name: mariadb-operator.v0.38.0
        version: 0.38.0
      - name: mariadb-operator.v0.37.1
        version: 0.37.1
      - name: mariadb-operator.v0.37.0
        version: 0.37.0
      - name: mariadb-operator.v0.36.0
        version: 0.36.0
      - name: mariadb-operator.v0.35.1
        version: 0.35.1
      - name: mariadb-operator.v0.35.0
        version: 0.35.0
      - name: mariadb-operator.v0.34.0
        version: 0.34.0
      - name: mariadb-operator.v0.33.0
        version: 0.33.0
      - name: mariadb-operator.v0.32.0
        version: 0.32.0
      - name: mariadb-operator.v0.31.0
        version: 0.31.0
      - name: mariadb-operator.v0.30.0
        version: 0.30.0
      - name: mariadb-operator.v0.29.0
        version: 0.29.0
      - name: mariadb-operator.v0.28.1
        version: 0.28.1
      - name: mariadb-operator.v0.27.0
        version: 0.27.0
      - name: mariadb-operator.v0.26.0
        version: 0.26.0
      - name: mariadb-operator.v0.24.0
        version: 0.24.0
      - name: mariadb-operator.v0.23.1
        version: 0.23.1
      - name: mariadb-operator.v0.23.0
        version: 0.23.0
      - name: mariadb-operator.v0.22.0
        version: 0.22.0
      - name: mariadb-operator.v0.21.0
        version: 0.21.0
      - name: mariadb-operator.v0.20.0
        version: 0.20.0
      - name: mariadb-operator.v0.19.0
        version: 0.19.0
      - name: mariadb-operator.v0.18.0
        version: 0.18.0
      - name: mariadb-operator.v0.17.0
        version: 0.17.0
      - name: mariadb-operator.v0.16.0
        version: 0.16.0
      - name: mariadb-operator.v0.15.1
        version: 0.15.1
      - name: mariadb-operator.v0.15.0
        version: 0.15.0
      - name: mariadb-operator.v0.14.0
        version: 0.14.0
      - name: mariadb-operator.v0.13.0
        version: 0.13.0
      - name: mariadb-operator.v0.11.0
        version: 0.11.0
      - name: mariadb-operator.v0.10.0
        version: 0.10.0
      - name: mariadb-operator.v0.9.1
        version: 0.9.1
      - name: mariadb-operator.v0.9.0
        version: 0.9.0
      - name: mariadb-operator.v0.8.0
        version: 0.8.0
      - name: mariadb-operator.v0.7.1
        version: 0.7.1
      - name: mariadb-operator.v0.7.0
        version: 0.7.0
      - name: mariadb-operator.v0.6.7
        version: 0.6.7
      - name: mariadb-operator.v0.6.6
        version: 0.6.6
      - name: mariadb-operator.v0.6.5
        version: 0.6.5
      - name: mariadb-operator.v0.6.4
        version: 0.6.4
      name: alpha
    defaultChannel: alpha
    packageName: mariadb-operator
    provider:
      name: mariadb-operator
      url: https://github.com/mariadb-operator/mariadb-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Anchore Inc.
      provider-url: ""
    name: anchore-engine
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anchore-engine.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.anchore.io/v1alpha1",
                "kind": "AnchoreEngine",
                "metadata": {
                  "name": "anchoreengine-sample"
                },
                "spec": {
                  "anchoreGlobal": {
                    "defaultAdminEmail": "example@email.com"
                  },
                  "postgresql": {
                    "extraEnv": [
                      {
                        "name": "POSTGRESQL_USER",
                        "value": "anchoreengine"
                      },
                      {
                        "name": "POSTGRESQL_PASSWORD",
                        "value": "anchore-postgres,123"
                      },
                      {
                        "name": "POSTGRESQL_DATABASE",
                        "value": "anchore"
                      },
                      {
                        "name": "PGUSER",
                        "value": "postgres"
                      }
                    ],
                    "postgresDatabase": "anchore",
                    "postgresPassword": "anchore-postgres,123",
                    "postgresUser": "anchoreengine"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/anchore/engine-operator@sha256:baf0f5acd319bc4d0132c88922ee3f357fdcdc278d931f243d0cbec3f6d416c6
          createdAt: "2021-10-06T23:53:51Z"
          description: Anchore Engine - container image scanning service for policy-based
            security, best-practice and compliance enforcement.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/anchore/engine-operator
          support: anchore.slack.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys Anchore Engine
            displayName: Anchore Engine Operator
            kind: AnchoreEngine
            name: anchoreengines.charts.anchore.io
            version: v1alpha1
        description: "Anchore Engine is an open source software system that provides
          a centralized service for analyzing container images, \n scanning for security
          vulnerabilities, and enforcing deployment policies. Anchore Engine is provided
          as a Docker \n container image that can be run standalone or with an orchestration
          platform such as Kubernetes. Anchore Engine allows \n users to perform detailed
          analysis on their container images, run queries, produce reports, and define
          policies that \n can be used in CI/CD pipelines. Users can extend Anchore
          Engine with plugins that add new queries, image analysis, and \n policies.
          Anchore Engine can be accessed directly through a RESTful API or via the
          Anchore CLI.\n\n The Anchore Engine Operator is based on the official stable
          [Helm Chart](https://github.com/anchore/anchore-charts/tree/master/stable/anchore-engine)."
        displayName: Anchore Engine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - analysis
        - docker
        - anchore
        - anchore-engine
        - image
        - security
        - vulnerability
        - scanner
        links:
        - name: Anchore Inc.
          url: https://www.anchore.com
        - name: Anchore Engine
          url: https://github.com/anchore/anchore-engine
        maintainers:
        - email: bradyt@anchore.com
          name: Brady Todhunter
        - email: zach@anchore.com
          name: Zach Hill
        maturity: alpha
        provider:
          name: Anchore Inc.
        relatedImages:
        - registry.connect.redhat.com/anchore/engine-operator@sha256:baf0f5acd319bc4d0132c88922ee3f357fdcdc278d931f243d0cbec3f6d416c6
        - registry.redhat.io/rhel8/postgresql-96@sha256:e8177c5de05ccdd6d12e360b65c63c889b2de725d573ffdedd8914230b118639
        - registry.connect.redhat.com/anchore/engine0@sha256:68e09d5ff695b46f24de3d7610f88dfc14b77b38758bbd68d03350dbae3deb0f
        version: 1.0.1
      entries:
      - name: anchore-engine.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: anchore-engine
    provider:
      name: Anchore Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: https://github.com/Mellanox/maintenance-operator/
    name: nvidia-maintenance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nvidia-maintenance-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "MaintenanceOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "maintenanceoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "maintenanceoperatorconfig",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "maintenanceoperatorconfig-sample"
                },
                "spec": {
                  "logLevel": "info",
                  "maxNodeMaintenanceTimeSeconds": 1600,
                  "maxParallelOperations": 1,
                  "maxUnavailable": "30%"
                }
              },
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "nodemaintenance-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodemaintenance",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "cordon": true,
                  "drainSpec": {
                    "deleteEmptyDir": true,
                    "force": true,
                    "podEvictionFilters": [
                      {
                        "byResourceNameRegex": "nvidia.com/gpu-*"
                      },
                      {
                        "byResourceNameRegex": "nvidia.com/rdma*"
                      }
                    ],
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  },
                  "nodeName": "worker-node-01",
                  "requestorID": "some-cluster-component.vendor.com",
                  "waitForPodCompletion": {
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
          createdAt: "2025-07-13T12:33:06Z"
          description: Node maintenance in K8s cluster in a coordinated manner
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-maintenance-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: NVIDIA
          repository: https://github.com/Mellanox/maintenance-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MaintenanceOperatorConfig is the Schema for the maintenanceoperatorconfigs
              API
            displayName: Maintenance Operator Config
            kind: MaintenanceOperatorConfig
            name: maintenanceoperatorconfigs.maintenance.nvidia.com
            version: v1alpha1
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.maintenance.nvidia.com
            version: v1alpha1
        description: Coordinates And Performs Common Node Preparations For Maintenance
        displayName: NVIDIA Maintenance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node
        - maintenance
        - drain
        links:
        - name: NVIDIA Maintenance Operator
          url: https://github.com/Mellanox/maintenance-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/maintenance-operator/
        relatedImages:
        - ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
        version: 0.2.3
      entries:
      - name: nvidia-maintenance-operator.v0.2.3
        version: 0.2.3
      name: stable
    - currentCSV: nvidia-maintenance-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "MaintenanceOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "maintenanceoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "maintenanceoperatorconfig",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "maintenanceoperatorconfig-sample"
                },
                "spec": {
                  "logLevel": "info",
                  "maxNodeMaintenanceTimeSeconds": 1600,
                  "maxParallelOperations": 1,
                  "maxUnavailable": "30%"
                }
              },
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "nodemaintenance-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodemaintenance",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "cordon": true,
                  "drainSpec": {
                    "deleteEmptyDir": true,
                    "force": true,
                    "podEvictionFilters": [
                      {
                        "byResourceNameRegex": "nvidia.com/gpu-*"
                      },
                      {
                        "byResourceNameRegex": "nvidia.com/rdma*"
                      }
                    ],
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  },
                  "nodeName": "worker-node-01",
                  "requestorID": "some-cluster-component.vendor.com",
                  "waitForPodCompletion": {
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
          createdAt: "2025-07-13T12:33:06Z"
          description: Node maintenance in K8s cluster in a coordinated manner
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-maintenance-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: NVIDIA
          repository: https://github.com/Mellanox/maintenance-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MaintenanceOperatorConfig is the Schema for the maintenanceoperatorconfigs
              API
            displayName: Maintenance Operator Config
            kind: MaintenanceOperatorConfig
            name: maintenanceoperatorconfigs.maintenance.nvidia.com
            version: v1alpha1
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.maintenance.nvidia.com
            version: v1alpha1
        description: Coordinates And Performs Common Node Preparations For Maintenance
        displayName: NVIDIA Maintenance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node
        - maintenance
        - drain
        links:
        - name: NVIDIA Maintenance Operator
          url: https://github.com/Mellanox/maintenance-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/maintenance-operator/
        relatedImages:
        - ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
        version: 0.2.3
      entries:
      - name: nvidia-maintenance-operator.v0.2.3
        version: 0.2.3
      name: v0.2
    defaultChannel: v0.2
    packageName: nvidia-maintenance-operator
    provider:
      name: NVIDIA
      url: https://github.com/Mellanox/maintenance-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: mariadb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mariadb.openstack.org/v1beta1",
                "kind": "Galera",
                "metadata": {
                  "name": "openstack"
                },
                "spec": {
                  "replicas": 3,
                  "secret": "osp-secret",
                  "storageClass": "local-storage",
                  "storageRequest": "500M"
                }
              },
              {
                "apiVersion": "mariadb.openstack.org/v1beta1",
                "kind": "MariaDBAccount",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "mariadb-operator",
                    "app.kubernetes.io/instance": "mariadbaccount-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mariadbaccount",
                    "app.kubernetes.io/part-of": "mariadb-operator",
                    "mariaDBDatabaseName": "neutron"
                  },
                  "name": "neutron1"
                },
                "spec": {
                  "secret": "neutrondb-secret",
                  "userName": "neutron1"
                }
              },
              {
                "apiVersion": "mariadb.openstack.org/v1beta1",
                "kind": "MariaDBDatabase",
                "metadata": {
                  "labels": {
                    "dbName": "mariadb"
                  },
                  "name": "scott"
                },
                "spec": {
                  "name": "scott",
                  "secret": "tiger"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:39Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Galera is the Schema for the galeras API
            displayName: Galera
            kind: Galera
            name: galeras.mariadb.openstack.org
            version: v1beta1
          - description: MariaDBAccount is the Schema for the mariadbaccounts API
            displayName: Maria DBAccount
            kind: MariaDBAccount
            name: mariadbaccounts.mariadb.openstack.org
            version: v1beta1
          - description: MariaDBDatabase is the Schema for the mariadbdatabases API
            displayName: Maria DBDatabase
            kind: MariaDBDatabase
            name: mariadbdatabases.mariadb.openstack.org
            version: v1beta1
        description: MariaDB Operator
        displayName: MariaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Identity
        - MariaDB
        links:
        - name: MariaDB Operator
          url: https://github.com/openstack-k8s-operators/mariadb-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/mariadb-rhel9-operator@sha256:5ae64a3cf9bf5eaf244548c3bf2e8ef89c9436cb676721c1f1d1c2ff8285e7e5
        - registry.redhat.io/rhoso/openstack-mariadb-rhel9@sha256:13119fe0ec56263a2bff3fc9c3892ea6386b837955280d8488f05d8ea6c4e44d
        version: 1.0.6
      entries:
      - name: mariadb-operator.v1.0.6
        version: 1.0.6
      - name: mariadb-operator.v1.0.5
        version: 1.0.5
      - name: mariadb-operator.v1.0.4
        version: 1.0.4
      - name: mariadb-operator.v1.0.3
        version: 1.0.3
      - name: mariadb-operator.v1.0.2
        version: 1.0.2
      - name: mariadb-operator.v1.0.1
        version: 1.0.1
      - name: mariadb-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: mariadb-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-community-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.5.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "registry": "docker.io",
                  "tag": "5.4.6",
                  "oem": "",
                  "imagePullSecrets": "",
                  "psp": false,
                  "rbac": true,
                  "serviceAccount": "default",
                  "leastPrivilege": false,
                  "global": {
                    "cattle": {
                      "url": ""
                    },
                    "azure": {
                      "enabled": false,
                      "identity": {
                        "clientId": "DONOTMODIFY"
                      },
                      "marketplace": {
                        "planId": "DONOTMODIFY"
                      },
                      "extension": {
                        "resourceId": "DONOTMODIFY"
                      },
                      "serviceAccount": "csp",
                      "imagePullSecrets": "",
                      "images": {
                        "neuvector_csp_pod": {
                          "tag": "latest",
                          "image": "neuvector-billing-azure-by-suse-llc",
                          "registry": "registry.suse.de/suse/sle-15-sp5/update/pubclouds/images",
                          "imagePullPolicy": "IfNotPresent"
                        },
                        "controller": {
                          "tag": "5.2.4",
                          "image": "controller",
                          "registry": "docker.io/neuvector"
                        },
                        "manager": {
                          "tag": "5.2.4",
                          "image": "manager",
                          "registry": "docker.io/neuvector"
                        },
                        "enforcer": {
                          "tag": "5.2.4",
                          "image": "enforcer",
                          "registry": "docker.io/neuvector"
                        }
                      }
                    },
                    "aws": {
                      "enabled": false,
                      "accountNumber": "",
                      "roleName": "",
                      "serviceAccount": "csp",
                      "annotations": {},
                      "imagePullSecrets": "",
                      "image": {
                        "digest": "",
                        "repository": "neuvector/neuvector-csp-adapter",
                        "tag": "latest",
                        "imagePullPolicy": "IfNotPresent"
                      }
                    }
                  },
                  "bootstrapPassword": "",
                  "autoGenerateCert": true,
                  "defaultValidityPeriod": 365,
                  "internal": {
                    "certmanager": {
                      "enabled": false,
                      "secretname": "neuvector-internal"
                    },
                    "autoGenerateCert": true,
                    "autoRotateCert": true
                  },
                  "controller": {
                    "enabled": true,
                    "annotations": {},
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": {
                      "repository": "neuvector/controller",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "searchRegistries": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "nodePort": "",
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      }
                    },
                    "ranchersso": {
                      "enabled": false
                    },
                    "pvc": {
                      "enabled": false,
                      "existingClaim": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {
                        "userinitcfg.yaml": {
                          "users": [
                            {
                              "Fullname": "admin",
                              "Password": "",
                              "Role": "admin"
                            }
                          ]
                        }
                      }
                    },
                    "certupgrader": {
                      "env": [],
                      "schedule": "",
                      "imagePullPolicy": "IfNotPresent",
                      "timeout": 3600,
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "prime": {
                      "enabled": false,
                      "image": {
                        "repository": "neuvector/compliance-config",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "1.0.7",
                        "hash": ""
                      }
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": {
                      "repository": "neuvector/enforcer",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "env": [],
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/etcd"
                      }
                    ],
                    "resources": {},
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "manager": {
                    "enabled": true,
                    "image": {
                      "repository": "neuvector/manager",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "priorityClassName": "",
                    "env": {
                      "ssl": true,
                      "envs": []
                    },
                    "svc": {
                      "type": "ClusterIP",
                      "nodePort": "",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": "",
                      "tls": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "topologySpreadConstraints": [],
                    "affinity": {},
                    "podLabels": {},
                    "podAnnotations": {},
                    "tolerations": [],
                    "nodeSelector": {},
                    "runAsUser": "",
                    "probes": {
                      "enabled": false,
                      "timeout": 1,
                      "periodSeconds": 10,
                      "startupFailureThreshold": 30
                    }
                  },
                  "cve": {
                    "adapter": {
                      "enabled": false,
                      "image": {
                        "repository": "neuvector/registry-adapter",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "0.1.9",
                        "hash": ""
                      },
                      "priorityClassName": "",
                      "resources": {},
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt"
                      },
                      "harbor": {
                        "protocol": "https",
                        "secretName": ""
                      },
                      "svc": {
                        "type": "ClusterIP",
                        "loadBalancerIP": "",
                        "annotations": {}
                      },
                      "route": {
                        "enabled": true,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      },
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ingressClassName": "",
                        "path": "/",
                        "annotations": {
                          "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                        },
                        "tls": false,
                        "secretName": ""
                      },
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      }
                    },
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "cacert": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                      "image": {
                        "registry": "",
                        "repository": "neuvector/updater",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "0.0.5",
                        "hash": ""
                      },
                      "schedule": "0 0 * * *",
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": {
                        "registry": "",
                        "repository": "neuvector/scanner",
                        "imagePullPolicy": "Always",
                        "tag": "6",
                        "hash": ""
                      },
                      "priorityClassName": "",
                      "resources": {},
                      "topologySpreadConstraints": [],
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      },
                      "volumes": "",
                      "volumeMounts": ""
                    }
                  },
                  "resources": {},
                  "runtimePath": "",
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhooksvc": {
                    "enabled": true
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  },
                  "lease": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "false"
          containerImage: docker.io/neuvector/neuvector-operator@sha256:6815df6435eafaf7b2bc9d5e3f0c9b65f0a50ad4e05d2862e56b227e19aeca27
          createdAt: "2019-11-18T02:09:59Z"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: NeuVector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart. NeuVector Operator version 1.5.5 deploys NeuVector app version 5.4.6.

          **Complete below steps to create namespace and Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Privileged SCC is added to Service Account specified in the deploment yaml by Operator version 1.3.4 and above in new deployment. In case of upgrading NeuVector Operator from previous version to 1.3.4, please delete Privileged SCC before upgrading.

                   oc delete rolebinding -nneuvector system:openshift:scc:privileged

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s  neuvector/default


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - wavefront
        - wavefront collector
        - wavefront proxy
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - docker.io/neuvector/neuvector-operator@sha256:6815df6435eafaf7b2bc9d5e3f0c9b65f0a50ad4e05d2862e56b227e19aeca27
        version: 1.5.5
      entries:
      - name: neuvector-operator.v1.5.5
        version: 1.5.5
      - name: neuvector-operator.v1.5.4
        version: 1.5.4
      - name: neuvector-operator.v1.5.3
        version: 1.5.3
      - name: neuvector-operator.v1.5.2
        version: 1.5.2
      - name: neuvector-operator.v1.5.1
        version: 1.5.1
      - name: neuvector-operator.v1.5.0
        version: 1.5.0
      - name: neuvector-operator.v1.4.9
        version: 1.4.9
      - name: neuvector-operator.v1.4.8
        version: 1.4.8
      - name: neuvector-operator.v1.4.7
        version: 1.4.7
      - name: neuvector-operator.v1.4.6
        version: 1.4.6
      - name: neuvector-operator.v1.4.5
        version: 1.4.5
      - name: neuvector-operator.v1.4.4
        version: 1.4.4
      - name: neuvector-operator.v1.4.3
        version: 1.4.3
      - name: neuvector-operator.v1.4.2
        version: 1.4.2
      - name: neuvector-operator.v1.4.1
        version: 1.4.1
      - name: neuvector-operator.v1.4.0
        version: 1.4.0
      - name: neuvector-operator.v1.3.9
        version: 1.3.9
      - name: neuvector-operator.v1.3.8
        version: 1.3.8
      - name: neuvector-operator.v1.3.7
        version: 1.3.7
      - name: neuvector-operator.v1.3.6
        version: 1.3.6
      - name: neuvector-operator.v1.3.5
        version: 1.3.5
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-community-operator
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "docker.io/percona/percona-postgresql-operator:2.7.0-ppg17.5.2-postgres",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "docker.io/percona/percona-pgbouncer:1.24.1",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "docker.io/percona/percona-pgbackrest:2.55.0",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:3.2.0",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: docker.io/percona/percona-postgresql-operator:2.7.0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          olm.skipRange: <v2.7.0
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native

          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - docker.io/percona/percona-postgresql-operator:2.7.0
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator.v2.6.1
        version: 2.6.1
      - name: percona-postgresql-operator.v2.6.0
        version: 2.6.0
      - name: percona-postgresql-operator.v2.5.0
        version: 2.5.0
      name: stable
    defaultChannel: stable
    packageName: percona-postgresql-operator
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: GitLab, Inc.
      provider-url: https://gitlab.com
    name: gitlab-runner-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-runner-operator.v1.41.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta2",
                "kind": "Runner",
                "metadata": {
                  "name": "gitlab-runner-example",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "gitlabUrl": "https://gitlab.example.com",
                  "token": "runner-token-secret",
                  "tags": "kubernetes,example",
                  "concurrent": 10,
                  "imagePullPolicy": "IfNotPresent"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery, Developer Tools
          containerImage: registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
          createdAt: "2025-09-25T15:08:40Z"
          description: GitLab Runner operator manages lifecycle of GitLab Runner instances
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.11.0 <1.41.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: none
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
          support: GitLab, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Runner is the open source project used to run your jobs and
              send the results back to GitLab
            displayName: GitLab Runner
            kind: Runner
            name: runners.apps.gitlab.com
            version: v1beta2
        description: |-
          GitLab Runner is the lightweight, highly-scalable agent that runs your build jobs and sends the results back to a GitLab instance. GitLab Runner works in conjunction with GitLab CI/CD, the open-source continuous integration service included with GitLab.

              The GitLab Runner operator manages the lifecycle of GitLab Runner in Kubernetes or Openshift clusters. The operator aims to automate the tasks needed to run your CI/CD jobs in your container orchestration platform.

              ## Prerequisites

                For Kubernetes cluster, install cert-manager:

                ```shell
                kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.7.1/cert-manager.yaml
                ```

              ## GitLab Runner version

              This version of **GitLab Runner Operator** ships with **GitLab Runner v18.4.0**.

              To use a different version of **GitLab Runner** change the [`runnerImage` and `helperImage` properties](https://docs.gitlab.com/runner/configuration/configuring_runner_operator.html#operator-properties).

              ## Usage

               To link a GitLab Runner instance to a self-hosted GitLab instance or the hosted [GitLab](https://gitlab.com), you first need to:

               1. Create a secret containing the `runner-token` from your GitLab project.

                ```
                cat > gitlab-runner-secret.yml << EOF
                apiVersion: v1
                kind: Secret
                metadata:
                  name: gitlab-runner-secret
                type: Opaque
                # Only one of the following fields must be set. The Operator will fail to register Runner if both are provided.
                # # NOTE: runner-registration-token is deprecated and will be removed in GitLab 18.0. You should use runner-token instead.
                stringData:
                  runner-token: REPLACE_ME # your project runner token
                  # runner-registration-token: "" # your project runner secret
                EOF
                ```

                ```
                oc apply -f gitlab-runner-secret.yml
                ```

               2. Create the Custom Resource Definition (CRD) file and include the following information. The tags value must be openshift for the job to run.

                 ```
                 cat > gitlab-runner.yml << EOF
                 apiVersion: apps.gitlab.com/v1beta2
                 kind: Runner
                 metadata:
                   name: gitlab-runner
                 spec:
                   gitlabUrl: https://gitlab.example.com
                   buildImage: alpine
                   token: gitlab-runner-secret
                   tags: openshift
                 EOF
                 ```

                ```
                oc apply -f gitlab-runner.yml
                ```

              ## Full documentation

              Visit [Install GitLab Runner Operator](https://docs.gitlab.com/runner/install/operator.html)
        displayName: GitLab Runner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GitLab
        - CI/CD
        - DevOps
        - SAST
        - DAST
        links:
        - name: Gitlab Runner Operator
          url: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
        - name: GitLab Docs
          url: https://docs.gitlab.com
        - name: GitLab Runner
          url: https://docs.gitlab.com/runner/
        - name: GitLab CI/CD
          url: https://docs.gitlab.com/ee/ci/quick_start/
        maintainers:
        - email: support@gitlab.com
          name: Georgi N. Georgiev
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: GitLab, Inc.
          url: https://gitlab.com
        relatedImages:
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-helper-ocp@sha256:29727fae83a226bcf57607240d4046a0b8c208b5ff42ee1f8ca0f115bb188809
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-ocp@sha256:0f2a5a3b6779d719eecbf079b99cf06b652603d74b563f57d753e32966d253dd
        version: 1.41.1
      entries:
      - name: gitlab-runner-operator.v1.41.1
        version: 1.41.1
      - name: gitlab-runner-operator.v1.41.0
        version: 1.41.0
      - name: gitlab-runner-operator.v1.40.1
        version: 1.40.1
      - name: gitlab-runner-operator.v1.40.0
        version: 1.40.0
      - name: gitlab-runner-operator.v1.39.0
        version: 1.39.0
      name: stable
    defaultChannel: stable
    packageName: gitlab-runner-operator
    provider:
      name: GitLab, Inc.
      url: https://gitlab.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Omar Gaye IBM
      provider-url: omarg.net
    name: memcached-operator-ogaye
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: memcached-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.omarg.net/v1alpha1",
                "kind": "Memcached",
                "metadata": {
                  "name": "memcached-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.6.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Memcached is the Schema for the memcacheds API
            displayName: Memcached
            kind: Memcached
            name: memcacheds.cache.omarg.net
            version: v1alpha1
        description: Omar memcached operator
        displayName: memcached-operator-ogaye
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Omar
        - Memcached
        links:
        - name: Memcached Operator OGaye
          url: https://memcached-operator.domain
        maturity: alpha
        provider:
          name: Omar Gaye IBM
          url: omarg.net
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9211b70edbb93d718428e1c9a992e5fc74a20cb213f8a0592ffe340c362890fe
        - quay.io/ogaye/memcached-operator@sha256:5340ac11ce676dd3204bf3807a4c2333a8dddcbcec918db0aeafbc977163c1c6
        version: 0.0.1
      entries:
      - name: memcached-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: memcached-operator-ogaye
    provider:
      name: Omar Gaye IBM
      url: omarg.net
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Aqua Security, Inc.
      provider-url: ""
    name: aqua-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aqua-operator.2022.4.836
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaCsp",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "namespace": "aqua",
                    "version": "2022.4",
                    "requirements": true
                  },
                  "common": {
                    "dbDiskSize": 10,
                    "databaseSecret": {
                      "key": "db-password",
                      "name": "aqua-database-password"
                    }
                  },
                  "database": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "gateway": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "server": {
                    "replicas": 1,
                    "service": "LoadBalancer"
                  },
                  "route": true,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaDatabase",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "splitDB": false
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "diskSize": 10,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "runAsNonRoot": false,
                  "gateway": {
                    "host": "aqua-gateway",
                    "port": 8443
                  },
                  "token": "<<your-token>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaKubeEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "config": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "aqua-secure",
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "service": "ClusterIP"
                  },
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaGateway",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  }
                },
                "runAsNonRoot": false
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaScanner",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "deploy": {
                    "replicas": 1
                  },
                  "runAsNonRoot": false,
                  "login": {
                    "username": "<<YOUR AQUA USER NAME>>",
                    "password": "<<YOUR AQUA USER PASSWORD>>",
                    "host": "http://aqua-server:8080"
                  }
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaServer",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "LoadBalancer"
                  }
                },
                "runAsNonRoot": false
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaLightning",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "global": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "Default-cluster-name"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "kubeEnforcer": {
                    "infra": {
                      "serviceAccount": "aqua-kube-enforcer-sa"
                    },
                    "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                    "deploy": {
                      "service": "ClusterIP"
                    },
                    "starboard": {
                      "infra": {
                        "serviceAccount": "aqua-kube-enforcer-sa"
                      },
                      "config": {
                        "imagePullSecret": "aqua-registry"
                      },
                      "deploy": {
                        "replicas": 1
                      }
                    },
                    "env": []
                  },
                  "enforcer": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "deploy": null,
                    "token": "<<ENFORCER_GROUP_TOKEN>>",
                    "env": [
                    ],
                    "secret": {
                      "name": null,
                      "key": null
                    },
                    "runAsNonRoot": null,
                    "rhcosVersion": null
                  }
                }
              },
            {
                  "apiVersion": "operator.aquasec.com/v1alpha1",
                  "kind": "AquaCloudConnector",
                  "metadata": {
                    "name": "aqua"
                  },
                  "spec": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "common": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    },
                    "runAsNonRoot": false,
                    "gateway": {
                      "host": "aqua-gateway",
                      "port": "8443"
                    },
                    "login": {
                      "host": "aqua-gateway:8443",
                      "password": "<<YOUR AQUA USER PASSWORD>>",
                      "token": "<<YOUR AQUA CLOUD CONNECTOR TOKEN>>",
                      "username": "<<YOUR AQUA USER NAME>>"
                    },
                    "tunnels": [
                      {
                        "host": "<account>.azurecr.io",
                        "port": "443"
                      },
                      {
                        "host": "gke.gcr.io",
                        "port": "443"
                      },
                      {
                        "host": "<account>.jfrog.io",
                        "port": "443"
                      },
                      {
                        "host": "1.1.1.1",
                        "port": "443"
                      },
                      {
                        "region": "us-east1",
                        "type": "ecr"
                      }
                    ]
                  }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/aquasec/aquasec@sha256:c9f2c770a51e9664006c12d124ca18aab76bea8c32a6a253fe31d44c02633642
          createdAt: ""
          description: The Aqua Security Operator runs within a Openshift cluster
            and provides a means to deploy and manage Aqua Security cluster and components.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2022.4.836'
          operators.operatorframework.io/internal-objects: '["aquastarboards.aquasecurity.github.io","configauditreports.aquasecurity.github.io",
            "clusterconfigauditreports.aquasecurity.github.io"]'
          repository: https://github.com/aquasecurity/aqua-operator
          support: Aqua Security, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aqua Security CSP Deployment with Aqua Operator
            displayName: AquaCsp
            kind: AquaCsp
            name: aquacsps.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Lightning Enforcers Deployment with Aqua Operator
            displayName: AquaLightningEnforcers
            kind: AquaLightning
            name: aqualightnings.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Database Deployment with Aqua Operator
            displayName: AquaDatabase
            kind: AquaDatabase
            name: aquadatabases.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Enforcer Deployment with Aqua Operator
            displayName: AquaEnforcer
            kind: AquaEnforcer
            name: aquaenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security KubeEnforcer Deployment with Aqua Operator
            displayName: AquaKubeEnforcer
            kind: AquaKubeEnforcer
            name: aquakubeenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Gateway Deployment with Aqua Operator
            displayName: AquaGateway
            kind: AquaGateway
            name: aquagateways.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Scanner Deployment with Aqua Operator
            displayName: AquaScanner
            kind: AquaScanner
            name: aquascanners.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Server Deployment with Aqua Operator
            displayName: AquaServer
            kind: AquaServer
            name: aquaservers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Starboard for kube-enforcer.
            displayName: AquaStarboard
            kind: AquaStarboard
            name: aquastarboards.aquasecurity.github.io
            version: v1alpha1
          - description: Aqua Cloud Connector Deployment with Aqua Operator
            displayName: AquaCloudConnector
            kind: AquaCloudConnector
            name: aquacloudconnectors.operator.aquasec.com
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ClusterConfigAuditReport
            kind: ClusterConfigAuditReport
            name: clusterconfigauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ConfigAuditReport
            kind: ConfigAuditReport
            name: configauditreports.aquasecurity.github.io
            version: v1alpha1
        description: |-
          The Aqua Security Operator runs within an OpenShift cluster and provides a means to deploy and manage the Aqua Security cluster and components :
          * Server (aka “console”)
          * Database (for production environments we recommend to use an external database and not the Aqua default database)
          * Gateway
          * Enforcer (aka “agent”)
          * KubeEnforcer
          * Scanner
          * CSP (package that contains the Server, Database, and Gateway)
          * LightningEnforcer (package that contains Enforcer and Kube Enforcer) - Do not deploy using Certified Operator.
          * AquaCloudConnector.

          Use the Aqua-Operator to
          * Deploy Aqua Security components on OpenShift
          * Scale up Aqua Security components with extra replicas
          * Assign metadata tags to Aqua Security components

          The Aqua Operator provides a few [Custom Resources](https://github.com/aquasecurity/aqua-operator/tree/master/config/crd) for managing the Aqua CSP platform.

          ## Prerequisites
          There are only 4 prerequisites:
          1. Verify that the OpenShift version is officially supported [Verify Here](https://docs.aquasec.com/saas/workload-protection/product-information/system-requirements/)
          2. Make sure you have a license and to obtain one, please contact Aqua Security at [Contact Us](mailto:cloudsales@aquasec.com).
          3. Create a new project for aqua

              ```oc new-project aqua```

          4. Create the secret for Aqua Database password. If you are using an External Database, make sure you use those credentials.
          NOTE: This step is optional and you can specify the Database password when creating the CRs.

              ```oc create secret generic aqua-database-password --from-literal=db-password=<password> -n aqua```

          Please note that for the certified operator, the Docker registry secret is NOT needed. For the Red Hat certified operator, the Aqua application images are stored in the Red Hat Connect registry and will be automatically pulled by the Aqua operator.

          ## Choosing the right channel

          Channels are a flexible publishing mechanism that offers you different Aqua versions to deploy, at any point. These channels map to their respective Aqua versions, describing the maturity based on where they are in their lifecycle:
          * 4.6: This maps to the older version of Aqua 4.6.0
          * 5.0: This maps to the Aqua version 5.0.0
          * 5.3: This maps to the Aqua version 5.3.0
          * 6.0: This maps to the Aqua version 6.0.0
          * 6.2: This maps to the Aqua version 6.2.0
          * 6.5: This maps to the Aqua version 6.5.0
          * 2022.4: This is the latest and greatest version of Aqua (default)

          ## Deploying the Aqua Operator

          Aqua Operator follows an easy push button deployment that installs the operator in its own aqua namespace.
          Please refer to the instructions [here](https://github.com/aquasecurity/aqua-operator/blob/2022.4.0/docs/DeployOpenShiftOperator.md).

          ## Configuring the Aqua Operator

          Operators leverage CRDs as a configuration mechanism for the application. Aqua offers a number of CRDs that can be tweaked and configured as per your needs. Please refer to this [link](https://github.com/aquasecurity/aqua-operator/blob/6.2.0/docs/DeployOpenShiftOperator.md#aquacsp-crds) for Custom Resource examples for popular scenarios.

          ## Support

          For support please contact support@aquasec.com.
        displayName: Aqua Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aqua-security
        - scanning
        - security
        - runtime-security
        links:
        - name: Aqua Security
          url: https://www.aquasec.com/
        - name: Aqua Operator Github
          url: https://github.com/aquasecurity/aqua-operator
        maintainers:
        - email: support@aquasec.com
          name: Aqua Support, Aqua Security
        maturity: alpha
        provider:
          name: Aqua Security, Inc.
        relatedImages:
        - registry.connect.redhat.com/aquasec/aquasec@sha256:c9f2c770a51e9664006c12d124ca18aab76bea8c32a6a253fe31d44c02633642
        - registry.connect.redhat.com/aquasec/gateway@sha256:11aa01e681986f6b5bd53b41a07db4456023080a753970d8c872e000f68f3b9b
        - registry.connect.redhat.com/aquasec/kube-enforcer@sha256:c456b65a3f1dc2de7209f3d9cbcf1bb523130021433b8fd740c41d9bd6f7db9f
        - registry.connect.redhat.com/aquasec/cloud-connector@sha256:f9da1936ac5e4b8e4537dca69f920c682fb9fac98d091680e01d2122f2b0f85f
        - registry.connect.redhat.com/aquasec/enforcer@sha256:50128b750e0a094de266c9a2cb42ce06a757f441a68d206a8a9c17f7b7119066
        - registry.connect.redhat.com/aquasec/database@sha256:c91ae82b0179c3ec4e156835a9f23b66767ba15514ff33fb268b35d8f04d6fde
        - registry.connect.redhat.com/aquasec/scanner@sha256:c086d735c7e6ea842ad1aaeb1eb19d3a951e6e88d397a20f2cd559c6bc042d26
        - registry.connect.redhat.com/aquasec/starboard@sha256:61673f7ba1d983ff4bb4bd76036ac55e5b81696c9a3e3aa4ab1a49a27a7a2106
        - registry.connect.redhat.com/aquasec/console@sha256:ed371053e46e89ad68ed48282ddf28abc668a613a9c32e3aa45e5111512b8466
        - registry.connect.redhat.com/aquasec/kube-bench@sha256:bbb4e8a61033b303525af6764f685e62a2f441453c3f0eaaf4cbab6997bf93ee
        version: 2022.4.836
      entries:
      - name: aqua-operator.2022.4.836
        version: 2022.4.836
      - name: aqua-operator.2022.4.829
        version: 2022.4.829
      - name: aqua-operator.2022.4.818
        version: 2022.4.818
      - name: aqua-operator.2022.4.805
        version: 2022.4.805
      - name: aqua-operator.2022.4.760
        version: 2022.4.760
      - name: aqua-operator.2022.4.759
        version: 2022.4.759
      - name: aqua-operator.2022.4.720
        version: 2022.4.720
      - name: aqua-operator.2022.4.683
        version: 2022.4.683
      - name: aqua-operator.2022.4.662
        version: 2022.4.662
      - name: aqua-operator.2022.4.588
        version: 2022.4.588
      - name: aqua-operator.2022.4.587
        version: 2022.4.587
      - name: aqua-operator.2022.4.519
        version: 2022.4.519
      - name: aqua-operator.2022.4.518
        version: 2022.4.518
      name: "2022.4"
    defaultChannel: "2022.4"
    packageName: aqua-operator-certified
    provider:
      name: Aqua Security, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: netobserv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: network-observability-operator.v1.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowMetric",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netobserv-operator",
                    "app.kubernetes.io/instance": "flowmetric-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "flowmetric",
                    "app.kubernetes.io/part-of": "netobserv-operator"
                  },
                  "name": "flowmetric-sample"
                },
                "spec": {
                  "charts": [
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "",
                          "promQL": "sum(rate($METRIC[2m]))"
                        }
                      ],
                      "title": "External ingress traffic",
                      "type": "SingleStat",
                      "unit": "Bps"
                    },
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "{{DstK8S_Namespace}} / {{DstK8S_OwnerName}}",
                          "promQL": "sum(rate($METRIC{DstK8S_Namespace!=\"\"}[2m])) by (DstK8S_Namespace, DstK8S_OwnerName)"
                        }
                      ],
                      "sectionName": "External",
                      "title": "Top external ingress traffic per workload",
                      "type": "StackArea",
                      "unit": "Bps"
                    }
                  ],
                  "direction": "Ingress",
                  "filters": [
                    {
                      "field": "SrcSubnetLabel",
                      "matchType": "Absence"
                    }
                  ],
                  "labels": [
                    "DstK8S_HostName",
                    "DstK8S_Namespace",
                    "DstK8S_OwnerName",
                    "DstK8S_OwnerType"
                  ],
                  "metricName": "cluster_external_ingress_bytes_total",
                  "type": "Counter",
                  "valueField": "Bytes"
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "app"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "infra"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "enable": true,
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "includeList": [
                        "node_ingress_bytes_total",
                        "workload_ingress_bytes_total",
                        "namespace_flows_total"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta2",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "metrics": {
                        "server": {
                          "port": 9400
                        }
                      },
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "eBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "Disabled"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "\"app\""
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "\"infra\""
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "\"Pod\"",
                          "src_kind": "\"Pod\""
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "\"Service\""
                        },
                        "name": "Services network"
                      }
                    ],
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    }
                  },
                  "deploymentModel": "Direct",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "enable": true,
                    "lokiStack": {
                      "name": "loki"
                    },
                    "mode": "Monolithic",
                    "monolithic": {
                      "tenantID": "netobserv",
                      "tls": {
                        "caCert": {
                          "certFile": "service-ca.crt",
                          "name": "loki-gateway-ca-bundle",
                          "type": "configmap"
                        },
                        "enable": false
                      },
                      "url": "http://loki.netobserv.svc:3100/"
                    },
                    "readTimeout": "30s",
                    "writeBatchSize": 10485760,
                    "writeBatchWait": "1s",
                    "writeTimeout": "10s"
                  },
                  "namespace": "netobserv",
                  "networkPolicy": {
                    "additionalNamespaces": [],
                    "enable": false
                  },
                  "processor": {
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "Flows",
                    "metrics": {
                      "disableAlerts": [],
                      "server": {
                        "port": 9401
                      }
                    },
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "prometheus": {
                    "querier": {
                      "enable": true,
                      "mode": "Auto",
                      "timeout": "30s"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking, Observability, OpenShift Optional
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:13b6dd0ddf4bdd2836b0d3ec9c79119a902456dcee100cfac6898768e1f22aab
          createdAt: 2025-09-08T21:48:09
          description: Network flows collector and monitoring solution
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: '{"apiVersion":"flows.netobserv.io/v1beta2",
            "kind":"FlowCollector","metadata":{"name":"cluster"},"spec": {}}'
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
          support: NetObserv team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta2
          - description: '`FlowMetric` is the schema for the custom metrics API, which
              allows to generate more metrics out of flow logs. You can find examples
              here: https://github.com/netobserv/network-observability-operator/tree/main/config/samples/flowmetrics'
            displayName: Flow Metric
            kind: FlowMetric
            name: flowmetrics.flows.netobserv.io
            version: v1alpha1
        description: |
          Network Observability is an OpenShift operator that deploys a monitoring pipeline consisting in:
          - an eBPF agent, that generates network flows from captured packets
          - flowlogs-pipeline, a component that collects, enriches and exports these flows
          - a Console plugin for flows visualization with powerful filtering options, a topology representation and more

          Flow data is then available in multiple ways, each optional:

          - As Cluster Monitoring metrics
          - As raw flow logs stored in Grafana Loki
          - As raw flow logs exported to a collector

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, can optionally be used as the backend to store all collected flows. The Network Observability operator does not install Loki directly, however we provide some guidance to help you there.

          - For a production or production-like environment usage, refer to [the operator documentation](https://docs.openshift.com/container-platform/latest/network_observability/installing-operators.html).

          - For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 10GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          oc create namespace netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/1-storage.yaml) -n netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          If you prefer to not use Loki, you must set `spec.loki.enable` to `false` in `FlowCollector`.
          In that case, you can still get the Prometheus metrics or export raw flows to a custom collector. But be aware that some of the Console plugin features will be disabled. For instance, you will not be able to view raw flows there, and the metrics / topology will have a more limited level of details, missing information such as pods or IPs.

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for example [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/config/samples/flows_v1beta2_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          oc edit flowcollector cluster
          ```

          As it operates cluster-wide on every node, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you might have to configure differently if you used another installation method. Make sure to disable it (`spec.loki.enable`) if you don't want to use Loki.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`) an optional list of exporters to which to send enriched flows. KAFKA and IPFIX exporters are supported. This allows you to define any custom storage or processing that can read from Kafka or use the IPFIX standard.

          - To enable availability zones awareness, set `spec.processor.addZone` to `true`.

          ## Resource considerations

          The following table outlines examples of resource considerations for clusters with certain workload sizes.
          The examples outlined in the table demonstrate scenarios that are tailored to specific workloads. Consider each example only as a baseline from which adjustments can be made to accommodate your workload needs.


          | Resource recommendations                        | Extra small (10 nodes) | Small (25 nodes)       | Medium (65 nodes) **    | Large (120 nodes) **          |
          | ----------------------------------------------- | ---------------------- | ---------------------- | ----------------------- | ----------------------------- |
          | *Worker Node vCPU and memory*                   | 4 vCPUs\| 16GiB mem *  | 16 vCPUs\| 64GiB mem * | 16 vCPUs\| 64GiB mem  * |16 vCPUs\| 64GiB Mem *         |
          | *LokiStack size*                                | `1x.extra-small`       | `1x.small`             | `1x.small`              | `1x.medium`                   |
          | *Network Observability controller memory limit* | 400Mi (default)        | 400Mi (default)        | 400Mi (default)         | 800Mi                         |
          | *eBPF sampling rate*                            | 50 (default)           | 50 (default)           | 50 (default)            | 50 (default)                  |
          | *eBPF memory limit*                             | 800Mi (default)        | 800Mi (default)        | 2000Mi                  | 800Mi (default)               |
          | *FLP memory limit*                              | 800Mi (default)        | 800Mi (default)        | 800Mi (default)         | 800Mi (default)               |
          | *FLP Kafka partitions*                          | N/A                    | 48                     | 48                      | 48                            |
          | *Kafka consumer replicas*                       | N/A                    | 24                     | 24                      | 24                            |
          | *Kafka brokers*                                 | N/A                    | 3 (default)            | 3 (default)             | 3 (default)                   |

          *. Tested with AWS M6i instances.
          **. In addition to this worker and its controller, 3 infra nodes (size `M6i.12xlarge`) and 1 workload node (size `M6i.8xlarge`) were tested.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - More information on [configuring metrics](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/Metrics.md).
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: Network Observability
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/network-observability/network-observability-console-plugin-compat-rhel9@sha256:d130f94cdd3cc863a31c59f395a6c4f663e8e0da23747304353c6674b20d81be
        - registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:13b6dd0ddf4bdd2836b0d3ec9c79119a902456dcee100cfac6898768e1f22aab
        - registry.redhat.io/network-observability/network-observability-ebpf-agent-rhel9@sha256:0b4337f4db1e74a2bea1021a617cb42690251036f583bb771cc5ec888f38c97f
        - registry.redhat.io/network-observability/network-observability-flowlogs-pipeline-rhel9@sha256:09454618709a5c83beb4252f1c14c149716d1c09b6567c0a571d5390e8a639a2
        - registry.redhat.io/network-observability/network-observability-console-plugin-rhel9@sha256:7d92eadb3cd08a08672a65a196a5d3181e764c68e8c20cab2554933cdb10fd0c
        version: 1.9.3
      entries:
      - name: network-observability-operator.v1.9.3
        version: 1.9.3
      - name: network-observability-operator.v1.9.2
        version: 1.9.2
      - name: network-observability-operator.v1.9.1
        version: 1.9.1
      - name: network-observability-operator.v1.9.0
        version: 1.9.0
      - name: network-observability-operator.v1.8.1
        version: 1.8.1
      - name: network-observability-operator.v1.8.0
        version: 1.8.0
      - name: network-observability-operator.v1.7.0
        version: 1.7.0
      - name: network-observability-operator.v1.6.2
        version: 1.6.2
      - name: network-observability-operator.v1.6.1
        version: 1.6.1
      - name: network-observability-operator.v1.6.0
        version: 1.6.0
      - name: network-observability-operator.v1.5.0
        version: 1.5.0
      - name: network-observability-operator.v1.4.2
        version: 1.4.2
      - name: network-observability-operator.v1.4.1
        version: 1.4.1
      - name: network-observability-operator.v1.4.0
        version: 1.4.0
      - name: network-observability-operator.v1.3.0
        version: 1.3.0
      - name: network-observability-operator.v1.2.0
        version: 1.2.0
      - name: network-observability-operator.v1.1.0-202302110050
        version: 1.1.0
      - name: network-observability-operator.v1.0.0-202301070345
        version: 1.0.0
      name: stable
    - currentCSV: network-observability-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:fd943746840cba26996dabb47b1beb3445571fa43cb1b859fb073e4aa114b376
          createdAt: 2023-06-22T09:31:50
          description: Network flows collector and monitoring solution
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
          - description: FlowCollector is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
        description: |
          Network Observability is an OpenShift operator that deploys a monitoring pipeline to collect and enrich network flows that are produced by the Network Observability eBPF agent.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When a `FlowCollector` instance is created, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For normal usage, we recommend two options:

          - Installing the [Loki Operator](https://docs.openshift.com/container-platform/4.11//logging/cluster-logging-loki.html). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          oc create namespace netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.0.2/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.0.2/config/samples/flows_v1beta1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          oc edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.0.2/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: Network Observability
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:fd943746840cba26996dabb47b1beb3445571fa43cb1b859fb073e4aa114b376
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:37b3c9b2eed351d9d1a39c99045bbc7993ccb819f67f7b390deaa34930f45946
        - registry.redhat.io/network-observability/network-observability-console-plugin-rhel9@sha256:880c0812a2cd4b7194f417e31fbdcee951ce76776fa06e33c5c9cd479a854daa
        - registry.redhat.io/network-observability/network-observability-ebpf-agent-rhel9@sha256:e31acc12d8690c339e13b5dd22d33e707701362bdd31086c9b4c0f48cad62cc3
        - registry.redhat.io/network-observability/network-observability-flowlogs-pipeline-rhel9@sha256:7c7c42fff63e0599bb943a9e9617b43181360d5a5254b27f368010af275cf0c3
        version: 1.3.0
      entries:
      - name: network-observability-operator.v1.3.0
        version: 1.3.0
      - name: network-observability-operator.v1.2.0
        version: 1.2.0
      - name: network-observability-operator.v1.1.0-202302110050
        version: 1.1.0
      - name: network-observability-operator.v1.0.0-202301070345
        version: 1.0.0
      name: v1.0.x
    defaultChannel: stable
    packageName: netobserv-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: openshift-builds-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-builds-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "openshift-builds"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-rhel8-operator@sha256:7100a83aa053a976ceb77ace79f090ec8601bc9b2152ff6f7d00f0b1deea7f0b
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n\n## Usage\n\nTo deploy and manage [builds
          for Red Hat OpenShift](https://github.com/shipwright-io/build) in your cluster,\nfirst
          make sure this operator is installed and running on your cluster.\n\nNext,
          create the following:\n\n```yaml\n---\napiVersion: operator.shipwright.io/v1alpha1\nkind:
          ShipwrightBuild\nmetadata:\n  name: openshift-builds-operator\nspec:\n  targetNamespace:
          openshift-builds\n```\n\nThe operator will deploy Shipwright Build in the
          provided `targetNamespace`.\nWhen `.spec.targetNamespace` is not set, the
          namespace will default to ` openshift-builds`.\nRefer to the [ShipwrightBuild
          operator documentation](https://github.com/shipwright-io/operator/blob/main/docs/shipwrightbuild.md)
          for more information about this custom resource.\n"
        displayName: builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://shipwright.io
        - name: Shipwright Builds
          url: https://github.com/shipwright-io/build
        maintainers:
        - email: pipeline-integrations-team@redhat.com
          name: Red Hat Pipelines Integration Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-rhel8-operator@sha256:7100a83aa053a976ceb77ace79f090ec8601bc9b2152ff6f7d00f0b1deea7f0b
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel8@sha256:73efcb6ad82f698d7463718b06c86f2b2b3b5326689a62dc39c9f5d68f8943f2
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel8@sha256:0cf079a24b8b3a4b3e1f30710fea85039d9cf323cfedd06e7d1e33e6336fb848
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel8@sha256:22930726a25b11de2571da0395b53352939d142a16f4db6fd5460abd4cc37ba5
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel8@sha256:fd903804672ec08acf9a2b49aa7ecc5f9b543c12f7462637ef756733171b3817
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel8@sha256:f95348e5fc48676d93ebc919af2c27a041dfd632ec285296246d886daab11576
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel8@sha256:d01c1749ce76dd4bf5fefaa6c553725d33533d6e70e884e916f12217b674c3cf
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:97cade2c1ee468261aec5400728c8d44de387b459134aec7a4c3b5ec5a335d2c
        version: 1.0.2
      entries:
      - name: openshift-builds-operator.v1.0.2
        version: 1.0.2
      - name: openshift-builds-operator.v1.0.1
        version: 1.0.1
      - name: openshift-builds-operator.v1.0.0
        version: 1.0.0
      name: builds-1.0
    - currentCSV: openshift-builds-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          createdAt: "2025-01-23T14:48:43Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.openshift.com/builds/latest/about/overview-openshift-builds.html
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:778ad1946bdad68e1430a2fb366cee25cd98d87e44cf5ad34b87603c2f4171d0
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:e8582857566020c71653228355cbadc2c1cda41e36fc767c204a4edb05e7f42a
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:fa039b62ebd1fc6e86db8ec1c7ae5fe0dd3a6f6f642b63b4dc89af32a78cea6b
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:1ec9a1dcfc08b707df48175bba76f8a6cbfcc8396d9e9940cc3d7100e6731701
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:0e139db4370c37e91ef7eb3e0cf9a2587cb81f37709009a37f2724d4461ed2f3
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:949a1a5150642167b6c7d2b8f6d92a2885afe453319c02495f351b2e79eff3dd
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:7cfffdd43fdd33a20d115f35c2ba492da3c1d0e3ab970587bc888c68c47dfacf
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:97cade2c1ee468261aec5400728c8d44de387b459134aec7a4c3b5ec5a335d2c
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:e9f7b8741f265a332008424a2c5b3dae9ea154e623c469085c2abafb53fc62d4
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:5babdf590730b217deb59acbccf142c0500bc692542131810714cc00724ee02c
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:5ec474ab9e62c92e44c1a6635913032aead6246faec0895dcfbdb28ab0dc8a1a
        version: 1.1.1
      entries:
      - name: openshift-builds-operator.v1.1.1
        version: 1.1.1
      - name: openshift-builds-operator.v1.1.0
        version: 1.1.0
      name: builds-1.1
    - currentCSV: openshift-builds-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          createdAt: "2025-01-23T13:47:38Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.2
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:ee2792972dc52d81814603e26103e0d1117194db5ab3b5577736b0fc3285a8ab
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3fa22124916523b958c67af8ad652e73a2c3d68bb5579da1cba1ade537f3b7ae
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:8ace52981058ed51cec5b893e6d773de963b12ab715258ac1f2cbc6430ad52e7
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:775c15b93b19ba855791d34c4a95ef203dd9f8e51110c67ebf131e38eda49422
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:2b288e36e3606cf8a0d7723772edb8d4ac11772ec450539ef3448ad517155ecb
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:4bee8594d17172e5b9893a865a9922a1c7869027633ad64b3f684ec44755bac6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3fcd8e2bf0bcb8ff8c93a87af2c59a3bcae7be8792f9d3236c9b5bbd9b6db3b2
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:e8582857566020c71653228355cbadc2c1cda41e36fc767c204a4edb05e7f42a
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:78ee35ccc70b78638c667152b7b6247417c0aa1bb6d3f817d59b22eb179899ca
        - registry.redhat.io/ubi8/buildah@sha256:56c71ecd450f284325ba8434cbf5b3006078b11f0625c6258689699fa811c95d
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:ebfe4f8a97f6f0f9afe38953ae8cc2fdf2ec0b76401c819bb12c1df7cd671e9e
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:016c93578908f1dbdc3052333a1d5a9ab0c03104070ac36a6ddf99350d7510f4
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:87c4c0e9fe4d9baf8ae4f191c51d90c46729273e182fb5b3dfd951632e7bfad3
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:a91c2bb1239ef7fbce8c1ab9308da03b533df9cea75c7d978fd76f38087bc912
        version: 1.2.2
      entries:
      - name: openshift-builds-operator.v1.2.2
        version: 1.2.2
      - name: openshift-builds-operator.v1.2.1
        version: 1.2.1
      - name: openshift-builds-operator.v1.2.0
        version: 1.2.0
      name: builds-1.2
    - currentCSV: openshift-builds-operator.v1.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-01-29T08:58:54Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.2
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:6e44758fd4a7b34814ac7c71e52d93633781d39f7897ecab98a090859b487b4f
        - registry.redhat.io/ubi9/buildah@sha256:27d837f9bc69ad3c3651cf3315e2501b137c11baa553d9d46140e5cf7fa7873a
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:0783132b14ed8050664810f37e7ad6388fefbc586c45ecc7ba69dd5041a55c9f
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:e8582857566020c71653228355cbadc2c1cda41e36fc767c204a4edb05e7f42a
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:04b8bfc0dc6f260ef367ca1cce2bcc84032bc8be17b4a595be12a344813a2632
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:1e9a54212690f7930d2b681db87a6f79f5a2acc5c982734d4b283a2472b59db5
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:3523e7c2215e156081e5f05443884d65efad9b7b9ca81a35416695020231916a
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:d4b3f451becb7a42341c98dd5b9fc5d2b582c93cac2728070ef4bb83b64abb4c
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:a4ba591c57bbc815c93023f4a080213be32581ba6a1c2d94b0ecbc5c90b07a1e
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:0b1dd35f55181f34775188bc71b3ff8dff66c70acfcfabad19353a02e19351f7
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:6214708f3b517cb901af9f9bd0502ba54f8e5183d0d909a3b559deee6394026e
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:ed3fdd1e85f5c5434de6ec0feaec8c7fe4e680ca6c64258287f18ec2886bb71f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3fcd8e2bf0bcb8ff8c93a87af2c59a3bcae7be8792f9d3236c9b5bbd9b6db3b2
        version: 1.3.2
      entries:
      - name: openshift-builds-operator.v1.3.2
        version: 1.3.2
      - name: openshift-builds-operator.v1.3.1
        version: 1.3.1
      - name: openshift-builds-operator.v1.3.0
        version: 1.3.0
      name: builds-1.3
    - currentCSV: openshift-builds-operator.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2024-11-15T22:28:01Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:9d12e6cbf26e31a1e4458c99a972ed36125d759c5ec57773f203cd8dd0985de7
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:ed3fdd1e85f5c5434de6ec0feaec8c7fe4e680ca6c64258287f18ec2886bb71f
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:538416bc1384eb987e8a7e80b57eda3feddaf4260b196709a6e9a9125e3af39e
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:1a6ca51c2c68950af605ff367356ab6c0ef7d1b2fd6fb8f00aa665196ff7af84
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:f4b935c16dcbbc0393733437747497c1d24f2cc782c40c49738e5f37c037b3e7
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:c60ca44275297d2ee8c499d9d7fb4056d104f4bd5dec8a02e7388241ac466c44
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:1c7c24111f1709d7648156873ced09462638aa4b3cf7e7aee13297e51bcdfb3e
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:6251cb303774e29d137a37d6527875c371fe1d3f6a5fa0383e80f0e9ef5c02be
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:f87c2069105121927fb0c0a2e37f8f72bc6cb5e035fa8d22f7b41ce0e86c0631
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:4e9c152023b2db04ee5d55f102490294b393bf78baa3aa8917f85f8547fbe3c8
        - registry.redhat.io/ubi9/buildah@sha256:6886997da8c92214e99e3e8da47a0f08239172d2544b4edae058d5cb018effab
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:adc4a31d63b7044db7094fddcb1a4b235d848e842257f5baf22b6ead03fd6c91
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:5e87ba403f9e9fff3002a8275aba0d6c36aed54fb4754d90fa0516ad989f187a
        version: 1.4.1
      entries:
      - name: openshift-builds-operator.v1.4.1
        version: 1.4.1
      - name: openshift-builds-operator.v1.4.0
        version: 1.4.0
      name: builds-1.4
    - currentCSV: openshift-builds-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-08-14T12:12:15Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              OpenShiftBuild describes the desired state of Builds for OpenShift, and the status of
              all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I), Buildah, and Buildpacks. You can create and apply
          build resources, view logs of build runs, \nand manage builds in your OpenShift
          Container Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:38560dea7627632d8b24fa670c1d7ac1218cf86975329908786a460f76de558d
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ce7050192d9850ed4c35cb3295ccdeec1aa593e27e3a3da6d6aca97cde782abe
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c1a84feb88d53e93bdcf2cd5f76e2cbb18541c8b0e2979c132a888d6c280b664
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:ddcfcebd5d788d11f756939f07215b2e3507a1d8fe704a1052e15bb4ebe592e8
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:753e24a1cd57ca22552447363aed48f064b525b49e52f774f3e25f41d8dcb329
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:e2a7a42cd5916c05d95d3fb0decc34d0840420532270422aa7f6b7bdd94de1f2
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:bcae84267f582278591d93aecf94c8d7fb1376521d9b5208201c959a0fe56774
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:f4a75c115b5dc8853845f666f3bd063b2736e01e428a605fb5cd0f02736ecce8
        - registry.redhat.io/ubi9/buildah@sha256:8ea2b81b32f70b9502d88722a5f69de8c0cf5efa6f30df041873546e9f9438cb
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:4dc123fa703006662efcc5569a843a043bb31cf69b6868d6a9b6321c66bfca18
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:1d5a0bfe0013ab7602efe95be8a814be1e5586728fdf0e198c97739cc12f1d87
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:504179e8ee9f88a23f1d4c28c7683ab7395f4c1c4868765bf3c1406669fd52fc
        version: 1.5.1
      entries:
      - name: openshift-builds-operator.v1.5.1
        version: 1.5.1
      - name: openshift-builds-operator.v1.5.0
        version: 1.5.0
      name: builds-1.5
    - currentCSV: openshift-builds-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-09-23T13:38:59Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I), Buildah, and Buildpacks. You can create and apply
          build resources, view logs of build runs, \nand manage builds in your OpenShift
          Container Platform namespaces.\n\n## Prerequisites\n\n* OpenShift Pipelines
          operator must be installed before installing this operator\n\nRead more:
          [https://shipwright.io](https://shipwright.io)\n\n## Features\n\n* Standard
          Kubernetes-native API for building container images from source code and
          Dockerfile\n\n* Support for Source-to-Image (S2I) and Buildah build strategies\n\n*
          Extensibility with your own custom build strategies\n\n* Execution of builds
          from source code in a local directory\n\n* Shipwright CLI for creating and
          viewing logs, and managing builds on the cluster\n\n* Integrated user experience
          with the Developer perspective of the OpenShift Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:c6805e85ad4dc29f3a0582a1fbf99647a90c1af0b3924647839ad881cbb199fa
        - registry.redhat.io/ubi9/buildah@sha256:8ea2b81b32f70b9502d88722a5f69de8c0cf5efa6f30df041873546e9f9438cb
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:f9f43795c7ed1e416b92497cf2fe647beb5e45d755f6a6d89c3915b7a9c76570
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:1a61d4a8f69fa845ee678dd96f934685821828e7575852f97b6e7b1dd79a8297
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c1a84feb88d53e93bdcf2cd5f76e2cbb18541c8b0e2979c132a888d6c280b664
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:3d2e501abe1aa757e4e6ba03fe8e7a45bbfdb4303360ed7130c6344051887ed8
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:9f38cd42591b6a6b65e78a3cc5c165f8498ad3c79460a6c47acfb8744d2dd883
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:0ea2b0153ae07ee15972abbbe034ef7feb0dcaa403ac993afcb171906d9cc796
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:23e3f29ae63546db018701abb29b444b9b825fb3a2bda5084542649d9827d4ba
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:dd96e8c857bc52097d4d194db8fa2cb368bbdcbb6594722660d4e906d0bdc420
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ce7050192d9850ed4c35cb3295ccdeec1aa593e27e3a3da6d6aca97cde782abe
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:fd7c805c6d0638f2814503c6779719279119c7268b1eb0eda9b7b58ac775efa6
        version: 1.6.0
      entries:
      - name: openshift-builds-operator.v1.6.0
        version: 1.6.0
      name: builds-1.6
    - currentCSV: openshift-builds-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-09-23T13:38:59Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I), Buildah, and Buildpacks. You can create and apply
          build resources, view logs of build runs, \nand manage builds in your OpenShift
          Container Platform namespaces.\n\n## Prerequisites\n\n* OpenShift Pipelines
          operator must be installed before installing this operator\n\nRead more:
          [https://shipwright.io](https://shipwright.io)\n\n## Features\n\n* Standard
          Kubernetes-native API for building container images from source code and
          Dockerfile\n\n* Support for Source-to-Image (S2I) and Buildah build strategies\n\n*
          Extensibility with your own custom build strategies\n\n* Execution of builds
          from source code in a local directory\n\n* Shipwright CLI for creating and
          viewing logs, and managing builds on the cluster\n\n* Integrated user experience
          with the Developer perspective of the OpenShift Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:0ea2b0153ae07ee15972abbbe034ef7feb0dcaa403ac993afcb171906d9cc796
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:c6805e85ad4dc29f3a0582a1fbf99647a90c1af0b3924647839ad881cbb199fa
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:3d2e501abe1aa757e4e6ba03fe8e7a45bbfdb4303360ed7130c6344051887ed8
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:fd7c805c6d0638f2814503c6779719279119c7268b1eb0eda9b7b58ac775efa6
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:23e3f29ae63546db018701abb29b444b9b825fb3a2bda5084542649d9827d4ba
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ce7050192d9850ed4c35cb3295ccdeec1aa593e27e3a3da6d6aca97cde782abe
        - registry.redhat.io/ubi9/buildah@sha256:8ea2b81b32f70b9502d88722a5f69de8c0cf5efa6f30df041873546e9f9438cb
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:f9f43795c7ed1e416b92497cf2fe647beb5e45d755f6a6d89c3915b7a9c76570
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:1a61d4a8f69fa845ee678dd96f934685821828e7575852f97b6e7b1dd79a8297
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:9f38cd42591b6a6b65e78a3cc5c165f8498ad3c79460a6c47acfb8744d2dd883
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c1a84feb88d53e93bdcf2cd5f76e2cbb18541c8b0e2979c132a888d6c280b664
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:dd96e8c857bc52097d4d194db8fa2cb368bbdcbb6594722660d4e906d0bdc420
        version: 1.6.0
      entries:
      - name: openshift-builds-operator.v1.6.0
        version: 1.6.0
      - name: openshift-builds-operator.v1.5.1
        version: 1.5.1
      - name: openshift-builds-operator.v1.5.0
        version: 1.5.0
      - name: openshift-builds-operator.v1.4.1
        version: 1.4.1
      - name: openshift-builds-operator.v1.4.0
        version: 1.4.0
      - name: openshift-builds-operator.v1.3.2
        version: 1.3.2
      - name: openshift-builds-operator.v1.3.1
        version: 1.3.1
      - name: openshift-builds-operator.v1.3.0
        version: 1.3.0
      - name: openshift-builds-operator.v1.2.2
        version: 1.2.2
      - name: openshift-builds-operator.v1.2.1
        version: 1.2.1
      - name: openshift-builds-operator.v1.2.0
        version: 1.2.0
      - name: openshift-builds-operator.v1.1.1
        version: 1.1.1
      - name: openshift-builds-operator.v1.1.0
        version: 1.1.0
      - name: openshift-builds-operator.v1.0.2
        version: 1.0.2
      - name: openshift-builds-operator.v1.0.1
        version: 1.0.1
      - name: openshift-builds-operator.v1.0.0
        version: 1.0.0
      name: latest
    defaultChannel: latest
    packageName: openshift-builds-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: VMware Tanzu
      provider-url: ""
    name: rabbitmq-messaging-topology-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-messaging-topology-operator.v1.17.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [{"apiVersion":"rabbitmq.com/v1beta1","kind":"Binding","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Exchange","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Federation","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Permission","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Policy","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"OperatorPolicy","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Queue","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"SchemaReplication","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Shovel","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1alpha1","kind":"SuperStream","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"User","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Vhost","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"TopicPermission","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}}]
          capabilities: Full Lifecycle
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/rabbitmqoperator/messaging-topology-operator:1.17.4
          createdAt: 2025-08-21T11:15:42
          description: Kubernetes operator to allow developers to create and manage
            RabbitMQ messaging topologies within a RabbitMQ cluster using a declarative
            Kubernetes API.
          repository: https://github.com/rabbitmq/messaging-topology-operator
          support: VMware Tanzu
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Binding
            displayName: Binding
            kind: Binding
            name: bindings.rabbitmq.com
            version: v1beta1
          - description: Exchange
            displayName: Exchange
            kind: Exchange
            name: exchanges.rabbitmq.com
            version: v1beta1
          - description: Federation
            displayName: Federation
            kind: Federation
            name: federations.rabbitmq.com
            version: v1beta1
          - description: Permission
            displayName: Permission
            kind: Permission
            name: permissions.rabbitmq.com
            version: v1beta1
          - description: TopicPermission
            displayName: TopicPermission
            kind: TopicPermission
            name: topicpermissions.rabbitmq.com
            version: v1beta1
          - description: Policy
            displayName: Policy
            kind: Policy
            name: policies.rabbitmq.com
            version: v1beta1
          - description: OperatorPolicy
            displayName: OperatorPolicy
            kind: OperatorPolicy
            name: operatorpolicies.rabbitmq.com
            version: v1beta1
          - description: Queue
            displayName: Queue
            kind: Queue
            name: queues.rabbitmq.com
            version: v1beta1
          - description: Schema Replication
            displayName: SchemaReplication
            kind: SchemaReplication
            name: schemareplications.rabbitmq.com
            version: v1beta1
          - description: Shovel
            displayName: Shovel
            kind: Shovel
            name: shovels.rabbitmq.com
            version: v1beta1
          - description: Super Stream
            displayName: SuperStream
            kind: SuperStream
            name: superstreams.rabbitmq.com
            version: v1alpha1
          - description: User
            displayName: User
            kind: User
            name: users.rabbitmq.com
            version: v1beta1
          - description: Vhost
            displayName: Vhost
            kind: Vhost
            name: vhosts.rabbitmq.com
            version: v1beta1
          required:
          - description: Rabbitmq Cluster
            displayName: RabbitmqCluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: |
          ## About the managed application
          RabbitMQ is an open-source message-broker software that originally implemented the Advanced Message Queuing Protocol and has since been extended with a plug-in architecture to support Streaming Text Oriented Messaging Protocol, MQ Telemetry Transport, and other protocols.
          ## About this Operator
          Kubernetes operator to allow developers to create and manage RabbitMQ messaging topologies within a RabbitMQ cluster using a declarative Kubernetes API. A Messaging topology is the collection of objects such as exchanges, queues, bindings and policies that provides specific messaging or streaming scenario. This operator is used with RabbitMQ clusters deployed via the RabbitMQ Cluster Kubernetes Operator. This repository contains custom controllers and custom resource definitions (CRDs) enabling a declarative API for RabbitMQ messaging topologies.
          ## Prerequisites for enabling this Operator
          The operator works on top of the RabbitMQ cluster operator.
          Messaging Topology Operator is tested with the latest release of RabbitMQ Cluster Operator. It uses the generated default user secret from RabbitmqCluster (set in rabbitmqcluster.status.binding) to authenticate with RabbitMQ server.
        displayName: rabbitmq-messaging-topology-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        - messaging
        - amqp
        - amqp10
        - stomp
        - mqtt
        - mqtt5
        - queues
        - streaming
        links:
        - name: GitHub
          url: https://github.com/rabbitmq/messaging-topology-operator
        - name: Documentation
          url: https://www.rabbitmq.com/kubernetes/operator/install-topology-operator.html
        maintainers:
        - email: rabbitmq-users@googlegroups.com
          name: VMware Tanzu
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        provider:
          name: VMware Tanzu
        relatedImages:
        - quay.io/rabbitmqoperator/messaging-topology-operator:1.17.4
        version: 1.17.4
      entries:
      - name: rabbitmq-messaging-topology-operator.v1.17.4
        version: 1.17.4
      - name: rabbitmq-messaging-topology-operator.v1.16.0
        version: 1.16.0
      - name: rabbitmq-messaging-topology-operator.v1.15.0
        version: 1.15.0
      - name: rabbitmq-messaging-topology-operator.v1.14.2
        version: 1.14.2
      - name: rabbitmq-messaging-topology-operator.v1.14.1
        version: 1.14.1
      - name: rabbitmq-messaging-topology-operator.v1.13.0
        version: 1.13.0
      - name: rabbitmq-messaging-topology-operator.v1.12.2
        version: 1.12.2
      - name: rabbitmq-messaging-topology-operator.v1.12.1
        version: 1.12.1
      name: stable
    defaultChannel: stable
    packageName: rabbitmq-messaging-topology-operator
    provider:
      name: VMware Tanzu
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: kaoto.io
      provider-url: https://kaoto.io
    name: kaoto-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kaoto-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "designer.kaoto.io/v1alpha1",
                "kind": "Kaoto",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "kaoto-operator",
                    "app.kubernetes.io/instance": "kaoto-demo",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kaoto",
                    "app.kubernetes.io/part-of": "kaoto-operator"
                  },
                  "name": "kaoto-demo"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          containerImage: quay.io/kaotoio/kaoto-operator@sha256:590d4958f148954ea1edb9825fb7fb4935d041ddcdb4a99302da14a605380071
          createdAt: "2025-09-03T12:48:58Z"
          description: Integration Editor for Apache Camel
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/KaotoIO/kaoto-operator
          support: contact@kaoto.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kaoto is the Schema for the kaotoes API.
            displayName: Kaoto
            kind: Kaoto
            name: kaotoes.designer.kaoto.io
            version: v1alpha1
        description: Kaoto is a low code and no code integration designer to edit
          and deploy based on Apache Camel . It is designed to be extendable and flexible
          and adaptable to different usecases.
        displayName: Kaoto Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - kaoto
        - camel-k
        - designer
        links:
        - name: Kaoto Operator
          url: https://kaoto-operator.domain
        maintainers:
        - email: info@kaoto.io
          name: kaoto
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: kaoto.io
          url: https://kaoto.io
        relatedImages:
        - quay.io/kaotoio/kaoto-operator@sha256:590d4958f148954ea1edb9825fb7fb4935d041ddcdb4a99302da14a605380071
        version: 2.7.0
      entries:
      - name: kaoto-operator.v2.7.0
        version: 2.7.0
      - name: kaoto-operator.v2.6.0
        version: 2.6.0
      - name: kaoto-operator.v2.5.0
        version: 2.5.0
      - name: kaoto-operator.v2.4.0
        version: 2.4.0
      - name: kaoto-operator.v2.3.0
        version: 2.3.0
      - name: kaoto-operator.v2.1.0
        version: 2.1.0
      - name: kaoto-operator.v2.0.1
        version: 2.0.1
      - name: kaoto-operator.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: kaoto-operator
    provider:
      name: kaoto.io
      url: https://kaoto.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: fence-agents-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fence-agents-remediation.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/fence-agents-remediation-operator:v0.4.0
          createdAt: "2024-05-23 19:57:04"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://www.medik8s.io/remediation/fence-agents-remediation/fence-agents-remediation/
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/medik8s/fence-agents-remediation-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      - name: fence-agents-remediation.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: fence-agents-remediation
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: hpe-filex-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-filex-csi-operator.v1.0.0-beta3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.hpe.com/v1",
                "kind": "HPEGreenLakeFileCSIDriver",
                "metadata": {
                  "name": "hpegreenlakefilecsidriver-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "disableNodeConformance": false,
                  "imagePullPolicy": "IfNotPresent",
                  "images": {
                    "csiAttacher": "registry.k8s.io/sig-storage/csi-attacher@sha256:6e54dae32284f60a2de1645c527715e4c934e5ce7899560c0d845bac957118dc",
                    "csiControllerDriver": "quay.io/hpestorage/filex-csi-driver@sha256:8e96df8f21db969dcd507c6fef8ecc82b7f4d9b33a9159f16345ebe7e02d8161",
                    "csiNodeDriver": "quay.io/hpestorage/filex-csi-driver@sha256:8e96df8f21db969dcd507c6fef8ecc82b7f4d9b33a9159f16345ebe7e02d8161",
                    "csiNodeDriverRegistrar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f",
                    "csiNodeInit": "quay.io/hpestorage/filex-csi-init@sha256:65a014b344d262f628100dcd96fecd228c42162bbfa67037952a5a67f16a1437",
                    "csiProvisioner": "registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f",
                    "csiResizer": "registry.k8s.io/sig-storage/csi-resizer@sha256:ab774734705a906561e15b67f6a96538f3ad520335d636f793aaafe87a3b5200",
                    "csiSnapshotter": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:dd788d79cf4c1b8edee6d9b80b8a1ebfc51a38a365c5be656986b129be9ac784"
                  },
                  "kubeletRootDir": "/var/lib/kubelet",
                  "node": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/hpestorage/filex-csi-driver-operator@sha256:e00066fccf2c673547ccafffb0a280130b6825ab6a82a52bd43d752535c63775
          createdAt: "2025-01-10T18:20:35Z"
          description: A Container Storage Interface (CSI) driver for HPE GreenLake
            File Storage.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.0.0-beta3'
          operatorhub.io/ui-metadata-max-k8s-version: "1.32"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/hpe-storage/co-deployments
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE GreenLake File Storage CSI Driver
            displayName: HPEGreenLakeFileCSIDriver
            kind: HPEGreenLakeFileCSIDriver
            name: hpegreenlakefilecsidrivers.storage.hpe.com
            version: v1
        description: |
          The HPE GreenLake File Storage CSI Operator deploys the  HPE GreenLake File Storage CSI Driver on OpenShift for dynamic  provisioning of persistent volumes.
          Please refer to our CSI driver [documentation](https://scod.hpedev.io/filex_csi_driver/index.html) for supported HPE platforms and full list of supported CSI features.
          ## Installation
          Refer to the HPE GreenLake File Storage CSI Operator [documentation](https://scod.hpedev.io/filex_csi_driver/deployment.html#operator).
        displayName: HPE GreenLake File Storage CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE CSI
        - HPE Storage
        - CSI
        links:
        - name: Documentation
          url: https://scod.hpedev.io/filex_csi_driver/index.html
        maintainers:
        - email: dev-hi-containers@hpe.com
          name: HPE Storage Containers Team
        maturity: beta
        minKubeVersion: 1.25.0
        provider:
          name: HPE
        relatedImages:
        - quay.io/hpestorage/filex-csi-driver@sha256:8e96df8f21db969dcd507c6fef8ecc82b7f4d9b33a9159f16345ebe7e02d8161
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:dd788d79cf4c1b8edee6d9b80b8a1ebfc51a38a365c5be656986b129be9ac784
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3fcd8e2bf0bcb8ff8c93a87af2c59a3bcae7be8792f9d3236c9b5bbd9b6db3b2
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - quay.io/hpestorage/filex-csi-init@sha256:65a014b344d262f628100dcd96fecd228c42162bbfa67037952a5a67f16a1437
        - registry.k8s.io/sig-storage/csi-attacher@sha256:6e54dae32284f60a2de1645c527715e4c934e5ce7899560c0d845bac957118dc
        - registry.connect.redhat.com/hpestorage/filex-csi-driver-operator@sha256:e00066fccf2c673547ccafffb0a280130b6825ab6a82a52bd43d752535c63775
        - registry.k8s.io/sig-storage/csi-resizer@sha256:ab774734705a906561e15b67f6a96538f3ad520335d636f793aaafe87a3b5200
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        version: 1.0.0-beta3
      entries:
      - name: hpe-filex-csi-operator.v1.0.0-beta3
        version: 1.0.0-beta3
      name: beta
    defaultChannel: beta
    packageName: hpe-filex-csi-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SmartX
      provider-url: https://www.smartx.com
    name: iomesh-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: iomesh-operator.0.10.1-rc4
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
          createdAt: 2021-08-17 00:00:00+00:00
          repository: https://github.com/iomesh/docs
          support: SmartX, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: iomesh
            displayName: iomesh
            kind: IOMeshCluster
            name: iomeshclusters.iomesh.com
            version: v1alpha1
        description: "IOMesh (https://www.iomesh.com/) is a cloud-native storage product
          specifically designed and developed for Kubernetes. It takes ZBS as its
          core, a “production-ready” distributed storage system developed independently
          by SmartX, to provide production-level high performance and reliable persistent
          storage for business-critical applications such as MySQL, Cassandra, MongoDB,
          etc., contributing to the containerized transformation of stateful applications.
          \     \n## Pre-Requisites\n* According to https://docs.iomesh.com/deploy/prerequisites,
          make sure that your openshift cluster meets the \"Installation Requirements\"
          and set up the worker nodes correctly according to the \"Setup Worker Node\"
          chapter.\n* Run IOMesh Operator pre-installation script in an environment
          where oc or kubectl can be used to access the openshift cluster, the script
          will install the dependencies of IOMesh Operator and specific IOMesh settings
          for the openshift cluster:\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-pre-install-openshift.sh
          | sh -\n```\n## Install IOMesh Operator and IOMesh Cluster\n* Click install
          button on the current web page to install IOMesh Operator\n* On the **Installed
          Operators** -> **IOMesh Operator** -> **Create instance** -> **YAML view**,
          fill in an IOMesh Custom Resources according https://iomesh.run/iomesh.yaml,
          change the `spec.*.dataCIDR` to your own data network CIDR:\n## Post Install\n*
          Run IOMesh Operator post-installation script in an environment where oc
          or kubectl can be used to access the openshift cluster, the script will
          install IOMesh CSI Driver\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-post-install-openshift.sh
          | sh -\n```\n## Getting Started\n* Try to quickstart setup and use IOMesh:
          https://docs.iomesh.com/deploy/setup-iomesh\n"
        displayName: IOMesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iomesh-operator
        links:
        - name: IOMesh Operator
          url: https://iomesh.com
        maintainers:
        - email: na.wang@smartx.com
          name: Na Wang
        maturity: alpha
        provider:
          name: SmartX
          url: https://www.smartx.com
        relatedImages:
        - iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
        version: 0.10.1-rc4
      entries:
      - name: iomesh-operator.0.10.1-rc4
        version: 0.10.1-rc4
      name: stable
    defaultChannel: stable
    packageName: iomesh-operator
    provider:
      name: SmartX
      url: https://www.smartx.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-activemqartemis: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/
    name: amq-broker-rhel9
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-broker-operator.v7.13.1-opr-1-0.1758027574.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 9 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-broker-operator.v7.13.1-opr-1-0.1756381365.p
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 9 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel9-operator@sha256:f34672152017a94ed51056435e218341e22787788ceed52439c1ff066480c715
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9cc48eecf1442ae04b8543fa5d4381a13bc2831390850828834d387006d1342b
        - registry.redhat.io/amq7/amq-broker-rhel9@sha256:755d0f7421efba76f52818bb3a8667cd7d2dcff2c7054dac1a4637b74ecac51d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:47fd4ce2533496828aba37bd1f9715e2164d5c90bd0fc6b25e7e0786d723bf01
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b2402179b1d0ff9c950defe5d8808195ef9a4c3e13e3c469cae31ab51d5f5014
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:bb6fbd68475a7852b4d99eea6c4ab313f9267da7963162f0d75375d7063409e7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:40920687402fd747ca02d384e63692a5e1db625723a2bf3056a607c92ee8eee1
        - registry.redhat.io/amq7/amq-broker-rhel9@sha256:4304a5414911fe55ce4c0133723be234ff288463f4916bd18e743d6359489138
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:7f9b921d89b6d4d91b365659afff127d99f1455b94b48a10a6a7f4c38dd8b700
        - registry.redhat.io/amq7/amq-broker-init-rhel9@sha256:ff80e0a7accda6825bf834d9aec27f48205dd6329991890dd017b4d21d0da5c8
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:16253895bc113f395dc6ab5cd511901e43b6ee676369ec80d598a8741c3c9bf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d42d713da0ce6806fdc6492b6342586783e6865a82a8647d3c4288439b1751ee
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:dc13871dfed52402c40fd192900fc3cb5a49da4fa08e4ec4d0e23a879e83db25
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ffffa9875f0379e9373f89f05eb06e5a193273bb04bc3aa5f85b044357b79098
        - registry.redhat.io/amq7/amq-broker-init-rhel9@sha256:e2cb996e0bf131e71fb8e80c128af18ab2b9330c9fca7a5437842796158e2870
        version: 7.13.1-opr-1+0.1758027574.p
      entries:
      - name: amq-broker-operator.v7.13.1-opr-1
        version: 7.13.1-opr-1
      - name: amq-broker-operator.v7.13.1-opr-1-0.1754564924.p
        version: 7.13.1-opr-1+0.1754564924.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1756374447.p
        version: 7.13.1-opr-1+0.1756374447.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1756381365.p
        version: 7.13.1-opr-1+0.1756381365.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1758024989.p
        version: 7.13.1-opr-1+0.1758024989.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1758027574.p
        version: 7.13.1-opr-1+0.1758027574.p
      - name: amq-broker-operator.v7.13.0-opr-1
        version: 7.13.0-opr-1
      - name: amq-broker-operator.v7.13.0-opr-1-0.1747315188.p
        version: 7.13.0-opr-1+0.1747315188.p
      - name: amq-broker-operator.v7.13.0-opr-1-0.1747658456.p
        version: 7.13.0-opr-1+0.1747658456.p
      name: 7.13.x
    defaultChannel: 7.13.x
    packageName: amq-broker-rhel9
    provider:
      name: Red Hat
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: quay-bridge-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-bridge-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "my-quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecretName": "<NAMESPACE>/<SECRET>",
                  "quayHostname": "https://<QUAY_URL>",
                  "insecureRegistry": false,
                  "whitelistNamespaces": ["default"]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:501fe9b003b5066b3936def0ef4de05abd1a12e795e718d6c9a62d2c6c804c2e
          createdAt: 2021-11-02 11:11 UTC
          description: Enhance OCP using Red Hat Quay container registry
          repository: https://github.com/redhat-cop/quay-openshift-registry-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a connected Quay instance.
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.redhatcop.redhat.io
            version: v1alpha1
        description: "## Overview\n\nThe OpenShift Container Platform contains an
          in cluster image registry for pushing and pulling images that either produced
          within the cluster or are references from external sources. Quay is a highly
          available, enterprise grade image registry with enhanced role based access
          control (RBAC) features. The goal of this operator is to support replicating
          the features found within OpenShift intgrated image registry with Quay.\n\n##
          Functionality\n\nThe following capabilities are enabled through this operator:\n\n*
          Synchronization of OpenShift namespaces as Quay organizations\n    * Creation
          of Robot accounts for each default namespace service account\n    * Creation
          of Secrets for each created Robot Account\n        * Associate Robot Secret
          to Service Account as Mountable and Image Pull Secret\n    * Synchronization
          of OpenShift ImageStreams as Quay Repositories\n* Automatically rewrite
          new Builds making use of ImageStreams to output to Quay\n* Automatically
          import ImageStream tag once build completes\n\n## Prerequisites\n\nThe following
          requirements must be satisfied prior to setting and configuring the integration
          between OpenShift and Quay:\n\n* A Red Hat Quay environment with a user
          with superuser permissions\n* A Red Hat OpenShift Container Platform environment
          (at least version 4.0) with cluster-admin permissions on an OpenShift cluster
          \n*\tOpenShift Command Line tool (CLI)\n\n## Setup and Configuration\n\nThe
          setup process involves steps within both Quay and OpenShift as described
          below:\n\n### Quay Setup\n\nThe integration between OpenShift and Quay makes
          extensive use of the Quay RESTFul API. To allow for other methods aside
          from user credentials to be used for authenticating against the API, an
          application is required to be created. When an application is created, an
          OAuth token is generated which can be used by the operator in OpenShift.
          Applications are created within a Quay organization. While an existing organization
          can be utilized, it is recommended that a dedicated organization be used.\n\n1.\tLogin
          to Quay as a user with superuser access and select the organization for
          which the external application will be configured within.\n2.\tOn the left
          hand side of the page, select **Applications**\n3.\tCreate a new application
          by selecting the **Create New Application** button and entering a name for
          the application\n4.\tAfter the new application has been created, select
          the application that was created\n5.\tOn the lefthand navigation bar, select
          the **Generate Token** button to create a new OAuth2 token\n6.\tSelect all
          of the checkboxes to grant the necessary access that is needed for the integration\n7.\tReview
          the permissions that will be assigned and then select the **Authorize Application**
          button\n8.\tTake note of the generated Access Token as it will be needed
          in the following section\n\n### OpenShift Setup\n\nTo complete the process
          to begin to integrate OpenShift and Quay, several steps are required. Before
          beginning, ensure that you have the OpenShift Command Line tool installed
          and that you are logged into OpenShift as a cluster administrator. \n\n####
          Create A Secret for the Quay OAuth Token\n\nThe Operator will use the previously
          obtained Access Token to communicate with Quay. Store this token within
          OpenShift as a secret.\n\nExecute the following command to create a secret
          called `quay-integration` with a key called `token` containing the access
          token:\n\n```\n$ oc create secret generic quay-integration --from-literal=token=<access_token>\n```\n\nThis
          token will be referenced in the final section.\n\n#### MutatingWebhookConfiguration
          Support\n\nSupport for dynamic interception of API requests that are performed
          as part of OpenShift’s typical build process is facilitated through a [MutatingWebhookConfiguration](https://docs.openshift.com/container-platform/3.11/architecture/additional_concepts/dynamic_admission_controllers.html).
          A MutatingWebhookConfiguration allows for invoking an API running within
          a project on OpenShift when certain API requests are received. In particular,
          to support Quay’s integration with OpenShift, any new Build requests should
          be intercepted so that the output can be modified to target Quay instead
          of OpenShift’s integrated registry. \n\nKubernetes requires that the webhook
          endpoint is secured via SSL using a certificate that makes use of the certificate
          authority for the cluster. Fortunately, OpenShift provides support for generating
          a certificate signed by the cluster. A script is available in the documentation
          called `webhook-create-signed-cert.sh` that automates the process for requesting
          a new certificate.\n\nFirst, create a `Service` for the Operator's webhook
          endpoint:\n```yaml\napiVersion: v1\nkind: Service\nmetadata:\n  labels:\n
          \   name: quay-bridge-operator\n  name: quay-bridge-operator\n  namespace:
          openshift-operators\nspec:\n  ports:\n    - name: https\n      port: 443\n
          \     protocol: TCP\n      targetPort: 8443\n  selector:\n    name: quay-bridge-operator\n
          \ sessionAffinity: None\n  type: ClusterIP\n```\n\nExecute the following
          command to request the certificate:\n\n```\n./webhook-create-signed-cert.sh
          --namespace openshift-operators --secret quay-bridge-operator-webhook-certs
          --service quay-bridge-operator\n```\n\nThe result will place the newly created
          private key and certificate within a secret within the secret specified.
          The secret will be mounted into the appropriate located within the operator
          as declared in the Deployment of the Operator. \n\nOnce the secret has been
          created, focus can shift to the `MutatingWebhookConfiguration` (see below).
          The majority of the content should be in place to apply to the cluster.
          The one section that requires configuration is the caBundle property. This
          refers to the Certificate Authority (CA) for the OpenShift environment.
          The OpenShift CA is available as a _ConfigMap_ within the `kube-system`
          namespace.\n\nExecute the following command to retrieve the CA and format
          the result as a single line so that it can be entered into the MutatingWebhookConfiguration
          resource:\n\n```\noc get configmap -n kube-system extension-apiserver-authentication
          -o=jsonpath='{.data.client-ca-file}' | base64 | tr -d '\\n'\n```\n\nReplace
          the `${CA_BUNDLE}` variable in the following `MutatingWebhookConfiguration`
          YAML and create it:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\nkind:
          MutatingWebhookConfiguration \nmetadata:\n  name: quay-bridge-operator\nwebhooks:\n
          \ - name: quayintegration.redhatcop.redhat.io\n    clientConfig: \n      service:\n
          \       namespace: openshift-operators\n        name: quay-bridge-operator\n
          \       path: \"/admissionwebhook\"\n      caBundle: \"${CA_BUNDLE}\"\n
          \   rules: \n    - operations:  [ \"CREATE\" ]\n      apiGroups: [ \"build.openshift.io\"
          ]\n      apiVersions: [\"v1\" ]\n      resources: [ \"builds\" ]\n    failurePolicy:
          Fail\n```\n\n_Note: Until the operator is running, new requests for builds
          will fail since the webserver the MutatingWebhookConfiguration invokes is
          not available and a proper is response is required in order for the object
          to be persisted in etcd._\n\n#### Create the QuayIntegration Custom Resource\n\nFinally,
          to complete the integration between OpenShift and Quay, a `QuayIntegration`
          custom resource needs to be created.\n\nThe following is an example of a
          basic definition of a `QuayIntegration` resource associated from the associated
          CRD.\n\n```\napiVersion: redhatcop.redhat.io/v1alpha1\nkind: QuayIntegration\nmetadata:\n
          \ name: example-quayintegration\nspec:\n  clusterID: openshift\n  credentialsSecretName:
          <NAMESPACE>/<SECRET>\n  quayHostname: https://<QUAY_URL>\n```\n\nThe _clusterID_
          is a value which should be unique across the entire ecosystem. This value
          is optional and defaults to `openshift`.\n\nThe _credentialsSecretName_
          is a NamespacedName value of the secret containing the token that was previously
          created.\n\nNote: If Quay is using self signed certificates, the property
          `insecureRegistry: true`\n\nOrganizations within Quay should be created
          for the related namespaces in OpenShift."
        displayName: Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Source Code
          url: https://github.com/redhat-cop/quay-openshift-registry-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:501fe9b003b5066b3936def0ef4de05abd1a12e795e718d6c9a62d2c6c804c2e
        version: 3.4.7
      entries:
      - name: quay-bridge-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: quay-bridge-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "my-quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecretName": "<NAMESPACE>/<SECRET>",
                  "quayHostname": "https://<QUAY_URL>",
                  "insecureRegistry": false,
                  "whitelistNamespaces": ["default"]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:9149df5ffc83dce6593072b7f72805e07e700b84527e45b4b7486b9c38f70731
          createdAt: 2021-09-15 23:09 UTC
          description: Enhance OCP using Red Hat Quay container registry
          olm.skipRange: '>=3.4.x <3.5.7'
          repository: https://github.com/redhat-cop/quay-openshift-registry-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a connected Quay instance.
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.redhatcop.redhat.io
            version: v1alpha1
        description: "## Overview\n\nThe OpenShift Container Platform contains an
          in cluster image registry for pushing and pulling images that either produced
          within the cluster or are references from external sources. Quay is a highly
          available, enterprise grade image registry with enhanced role based access
          control (RBAC) features. The goal of this operator is to support replicating
          the features found within OpenShift intgrated image registry with Quay.\n\n##
          Functionality\n\nThe following capabilities are enabled through this operator:\n\n*
          Synchronization of OpenShift namespaces as Quay organizations\n    * Creation
          of Robot accounts for each default namespace service account\n    * Creation
          of Secrets for each created Robot Account\n        * Associate Robot Secret
          to Service Account as Mountable and Image Pull Secret\n    * Synchronization
          of OpenShift ImageStreams as Quay Repositories\n* Automatically rewrite
          new Builds making use of ImageStreams to output to Quay\n* Automatically
          import ImageStream tag once build completes\n\n## Prerequisites\n\nThe following
          requirements must be satisfied prior to setting and configuring the integration
          between OpenShift and Quay:\n\n* A Red Hat Quay environment with a user
          with superuser permissions\n* A Red Hat OpenShift Container Platform environment
          (at least version 4.0) with cluster-admin permissions on an OpenShift cluster
          \n*\tOpenShift Command Line tool (CLI)\n\n## Setup and Configuration\n\nThe
          setup process involves steps within both Quay and OpenShift as described
          below:\n\n### Quay Setup\n\nThe integration between OpenShift and Quay makes
          extensive use of the Quay RESTFul API. To allow for other methods aside
          from user credentials to be used for authenticating against the API, an
          application is required to be created. When an application is created, an
          OAuth token is generated which can be used by the operator in OpenShift.
          Applications are created within a Quay organization. While an existing organization
          can be utilized, it is recommended that a dedicated organization be used.\n\n1.\tLogin
          to Quay as a user with superuser access and select the organization for
          which the external application will be configured within.\n2.\tOn the left
          hand side of the page, select **Applications**\n3.\tCreate a new application
          by selecting the **Create New Application** button and entering a name for
          the application\n4.\tAfter the new application has been created, select
          the application that was created\n5.\tOn the lefthand navigation bar, select
          the **Generate Token** button to create a new OAuth2 token\n6.\tSelect all
          of the checkboxes to grant the necessary access that is needed for the integration\n7.\tReview
          the permissions that will be assigned and then select the **Authorize Application**
          button\n8.\tTake note of the generated Access Token as it will be needed
          in the following section\n\n### OpenShift Setup\n\nTo complete the process
          to begin to integrate OpenShift and Quay, several steps are required. Before
          beginning, ensure that you have the OpenShift Command Line tool installed
          and that you are logged into OpenShift as a cluster administrator. \n\n####
          Create A Secret for the Quay OAuth Token\n\nThe Operator will use the previously
          obtained Access Token to communicate with Quay. Store this token within
          OpenShift as a secret.\n\nExecute the following command to create a secret
          called `quay-integration` with a key called `token` containing the access
          token:\n\n```\n$ oc create secret generic quay-integration --from-literal=token=<access_token>\n```\n\nThis
          token will be referenced in the final section.\n\n#### MutatingWebhookConfiguration
          Support\n\nSupport for dynamic interception of API requests that are performed
          as part of OpenShift’s typical build process is facilitated through a [MutatingWebhookConfiguration](https://docs.openshift.com/container-platform/3.11/architecture/additional_concepts/dynamic_admission_controllers.html).
          A MutatingWebhookConfiguration allows for invoking an API running within
          a project on OpenShift when certain API requests are received. In particular,
          to support Quay’s integration with OpenShift, any new Build requests should
          be intercepted so that the output can be modified to target Quay instead
          of OpenShift’s integrated registry. \n\nKubernetes requires that the webhook
          endpoint is secured via SSL using a certificate that makes use of the certificate
          authority for the cluster. Fortunately, OpenShift provides support for generating
          a certificate signed by the cluster. A script is available in the documentation
          called `webhook-create-signed-cert.sh` that automates the process for requesting
          a new certificate.\n\nFirst, create a `Service` for the Operator's webhook
          endpoint:\n```yaml\napiVersion: v1\nkind: Service\nmetadata:\n  labels:\n
          \   name: quay-bridge-operator\n  name: quay-bridge-operator\n  namespace:
          openshift-operators\nspec:\n  ports:\n    - name: https\n      port: 443\n
          \     protocol: TCP\n      targetPort: 8443\n  selector:\n    name: quay-bridge-operator\n
          \ sessionAffinity: None\n  type: ClusterIP\n```\n\nExecute the following
          command to request the certificate:\n\n```\n./webhook-create-signed-cert.sh
          --namespace openshift-operators --secret quay-bridge-operator-webhook-certs
          --service quay-bridge-operator\n```\n\nThe result will place the newly created
          private key and certificate within a secret within the secret specified.
          The secret will be mounted into the appropriate located within the operator
          as declared in the Deployment of the Operator. \n\nOnce the secret has been
          created, focus can shift to the `MutatingWebhookConfiguration` (see below).
          The majority of the content should be in place to apply to the cluster.
          The one section that requires configuration is the caBundle property. This
          refers to the Certificate Authority (CA) for the OpenShift environment.
          The OpenShift CA is available as a _ConfigMap_ within the `kube-system`
          namespace.\n\nExecute the following command to retrieve the CA and format
          the result as a single line so that it can be entered into the MutatingWebhookConfiguration
          resource:\n\n```\noc get configmap -n kube-system extension-apiserver-authentication
          -o=jsonpath='{.data.client-ca-file}' | base64 | tr -d '\\n'\n```\n\nReplace
          the `${CA_BUNDLE}` variable in the following `MutatingWebhookConfiguration`
          YAML and create it:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\nkind:
          MutatingWebhookConfiguration \nmetadata:\n  name: quay-bridge-operator\nwebhooks:\n
          \ - name: quayintegration.redhatcop.redhat.io\n    clientConfig: \n      service:\n
          \       namespace: openshift-operators\n        name: quay-bridge-operator\n
          \       path: \"/admissionwebhook\"\n      caBundle: \"${CA_BUNDLE}\"\n
          \   rules: \n    - operations:  [ \"CREATE\" ]\n      apiGroups: [ \"build.openshift.io\"
          ]\n      apiVersions: [\"v1\" ]\n      resources: [ \"builds\" ]\n    failurePolicy:
          Fail\n    sideEffects: None\n    admissionReviewVersions: [v1beta1]\n    matchPolicy:
          Exact\n    timeoutSeconds: 30  \n```\n\n_Note: Until the operator is running,
          new requests for builds will fail since the webserver the MutatingWebhookConfiguration
          invokes is not available and a proper is response is required in order for
          the object to be persisted in etcd._\n\n#### Create the QuayIntegration
          Custom Resource\n\nFinally, to complete the integration between OpenShift
          and Quay, a `QuayIntegration` custom resource needs to be created.\n\nThe
          following is an example of a basic definition of a `QuayIntegration` resource
          associated from the associated CRD.\n\n```\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          QuayIntegration\nmetadata:\n  name: example-quayintegration\nspec:\n  clusterID:
          openshift\n  credentialsSecretName: <NAMESPACE>/<SECRET>\n  quayHostname:
          https://<QUAY_URL>\n```\n\nThe _clusterID_ is a value which should be unique
          across the entire ecosystem. This value is optional and defaults to `openshift`.\n\nThe
          _credentialsSecretName_ is a NamespacedName value of the secret containing
          the token that was previously created.\n\nNote: If Quay is using self signed
          certificates, the property `insecureRegistry: true`\n\nOrganizations within
          Quay should be created for the related namespaces in OpenShift."
        displayName: Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Source Code
          url: https://github.com/redhat-cop/quay-openshift-registry-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:9149df5ffc83dce6593072b7f72805e07e700b84527e45b4b7486b9c38f70731
        version: 3.5.7
      entries:
      - name: quay-bridge-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: quay-bridge-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:33dc7673155f01728225221768f6b90d87cc7609dcda46b394dba2421c136a23
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.10.15
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:33dc7673155f01728225221768f6b90d87cc7609dcda46b394dba2421c136a23
        version: 3.10.15
      entries:
      - name: quay-bridge-operator.v3.10.15
        version: 3.10.15
      - name: quay-bridge-operator.v3.10.14
        version: 3.10.14
      - name: quay-bridge-operator.v3.10.13
        version: 3.10.13
      - name: quay-bridge-operator.v3.10.12
        version: 3.10.12
      - name: quay-bridge-operator.v3.10.11
        version: 3.10.11
      - name: quay-bridge-operator.v3.10.10
        version: 3.10.10
      - name: quay-bridge-operator.v3.10.9
        version: 3.10.9
      - name: quay-bridge-operator.v3.10.8
        version: 3.10.8
      - name: quay-bridge-operator.v3.10.7
        version: 3.10.7
      - name: quay-bridge-operator.v3.10.6
        version: 3.10.6
      - name: quay-bridge-operator.v3.10.5
        version: 3.10.5
      - name: quay-bridge-operator.v3.10.4
        version: 3.10.4
      - name: quay-bridge-operator.v3.10.3
        version: 3.10.3
      - name: quay-bridge-operator.v3.10.2
        version: 3.10.2
      - name: quay-bridge-operator.v3.10.1
        version: 3.10.1
      - name: quay-bridge-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-bridge-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:511f8471db70e8d5fd9d0b547c73f1416bbac6f0fe6cdeb6c59a9ef8bd845ac2
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.11.13
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:511f8471db70e8d5fd9d0b547c73f1416bbac6f0fe6cdeb6c59a9ef8bd845ac2
        version: 3.11.13
      entries:
      - name: quay-bridge-operator.v3.11.13
        version: 3.11.13
      - name: quay-bridge-operator.v3.11.12
        version: 3.11.12
      - name: quay-bridge-operator.v3.11.11
        version: 3.11.11
      - name: quay-bridge-operator.v3.11.10
        version: 3.11.10
      - name: quay-bridge-operator.v3.11.9
        version: 3.11.9
      - name: quay-bridge-operator.v3.11.8
        version: 3.11.8
      - name: quay-bridge-operator.v3.11.7
        version: 3.11.7
      - name: quay-bridge-operator.v3.11.6
        version: 3.11.6
      - name: quay-bridge-operator.v3.11.5
        version: 3.11.5
      - name: quay-bridge-operator.v3.11.4
        version: 3.11.4
      - name: quay-bridge-operator.v3.11.3
        version: 3.11.3
      - name: quay-bridge-operator.v3.11.2
        version: 3.11.2
      - name: quay-bridge-operator.v3.11.1
        version: 3.11.1
      - name: quay-bridge-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-bridge-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:c6e04de3ad3b8c0ded95146d090e120bffed7f9dc5e74332a183fe1c2e45d16d
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.12.12
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:c6e04de3ad3b8c0ded95146d090e120bffed7f9dc5e74332a183fe1c2e45d16d
        version: 3.12.12
      entries:
      - name: quay-bridge-operator.v3.12.12
        version: 3.12.12
      - name: quay-bridge-operator.v3.12.11
        version: 3.12.11
      - name: quay-bridge-operator.v3.12.10
        version: 3.12.10
      - name: quay-bridge-operator.v3.12.9
        version: 3.12.9
      - name: quay-bridge-operator.v3.12.8
        version: 3.12.8
      - name: quay-bridge-operator.v3.12.7
        version: 3.12.7
      - name: quay-bridge-operator.v3.12.6
        version: 3.12.6
      - name: quay-bridge-operator.v3.12.5
        version: 3.12.5
      - name: quay-bridge-operator.v3.12.4
        version: 3.12.4
      - name: quay-bridge-operator.v3.12.3
        version: 3.12.3
      - name: quay-bridge-operator.v3.12.2
        version: 3.12.2
      - name: quay-bridge-operator.v3.12.1
        version: 3.12.1
      - name: quay-bridge-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-bridge-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:e8ee1944d8e0e9d515a6e63a3e39901c304abd421fcad2abbf9a3c43c7ca6243
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.13.8
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:e8ee1944d8e0e9d515a6e63a3e39901c304abd421fcad2abbf9a3c43c7ca6243
        version: 3.13.8
      entries:
      - name: quay-bridge-operator.v3.13.8
        version: 3.13.8
      - name: quay-bridge-operator.v3.13.7
        version: 3.13.7
      - name: quay-bridge-operator.v3.13.6
        version: 3.13.6
      - name: quay-bridge-operator.v3.13.5
        version: 3.13.5
      - name: quay-bridge-operator.v3.13.4
        version: 3.13.4
      - name: quay-bridge-operator.v3.13.3
        version: 3.13.3
      - name: quay-bridge-operator.v3.13.2
        version: 3.13.2
      - name: quay-bridge-operator.v3.13.1
        version: 3.13.1
      - name: quay-bridge-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-bridge-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b12f0dfb8fdeae3f521e41fcf60ef118bdfd08ad7f68bd87c5494b067b7c7c58
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.14.5
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b12f0dfb8fdeae3f521e41fcf60ef118bdfd08ad7f68bd87c5494b067b7c7c58
        version: 3.14.5
      entries:
      - name: quay-bridge-operator.v3.14.5
        version: 3.14.5
      - name: quay-bridge-operator.v3.14.4
        version: 3.14.4
      - name: quay-bridge-operator.v3.14.3
        version: 3.14.3
      - name: quay-bridge-operator.v3.14.2
        version: 3.14.2
      - name: quay-bridge-operator.v3.14.1
        version: 3.14.1
      - name: quay-bridge-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-bridge-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b2ac4a48c80209d7bebe8b1b2d066571f99a9f47b78604e47a8300d027941b66
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.15.2
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b2ac4a48c80209d7bebe8b1b2d066571f99a9f47b78604e47a8300d027941b66
        version: 3.15.2
      entries:
      - name: quay-bridge-operator.v3.15.2
        version: 3.15.2
      - name: quay-bridge-operator.v3.15.1
        version: 3.15.1
      - name: quay-bridge-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-bridge-operator.v3.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:0ad7af440c6e00fb578ef40e6d600964324fd94ce7835a28d0ec355b7415093f
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.6.10
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:99dc6b5af811bb33892c17334dd4e1a7eddc6982d49a793b1789f505e00f49b1
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:0ad7af440c6e00fb578ef40e6d600964324fd94ce7835a28d0ec355b7415093f
        version: 3.6.10
      entries:
      - name: quay-bridge-operator.v3.6.10
        version: 3.6.10
      - name: quay-bridge-operator.v3.6.9
        version: 3.6.9
      - name: quay-bridge-operator.v3.6.8
        version: 3.6.8
      - name: quay-bridge-operator.v3.6.7
        version: 3.6.7
      - name: quay-bridge-operator.v3.6.6
        version: 3.6.6
      - name: quay-bridge-operator.v3.6.5
        version: 3.6.5
      - name: quay-bridge-operator.v3.6.4
        version: 3.6.4
      - name: quay-bridge-operator.v3.6.2
        version: 3.6.2
      - name: quay-bridge-operator.v3.6.1
        version: 3.6.1
      - name: quay-bridge-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: quay-bridge-operator.v3.7.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:6a0dd80dc48ad0aaa576b3de78b2b265e29a3c10e4996aa6b8f129592fe20f09
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.7.14
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:975f287513039b2244894b9031e590f665d48f9af480dfdc4b85d4f96dd4bc37
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:6a0dd80dc48ad0aaa576b3de78b2b265e29a3c10e4996aa6b8f129592fe20f09
        version: 3.7.14
      entries:
      - name: quay-bridge-operator.v3.7.14
        version: 3.7.14
      - name: quay-bridge-operator.v3.7.13
        version: 3.7.13
      - name: quay-bridge-operator.v3.7.12
        version: 3.7.12
      - name: quay-bridge-operator.v3.7.11
        version: 3.7.11
      - name: quay-bridge-operator.v3.7.10
        version: 3.7.10
      - name: quay-bridge-operator.v3.7.9
        version: 3.7.9
      - name: quay-bridge-operator.v3.7.8
        version: 3.7.8
      - name: quay-bridge-operator.v3.7.7
        version: 3.7.7
      - name: quay-bridge-operator.v3.7.6
        version: 3.7.6
      - name: quay-bridge-operator.v3.7.5
        version: 3.7.5
      - name: quay-bridge-operator.v3.7.4
        version: 3.7.4
      - name: quay-bridge-operator.v3.7.3
        version: 3.7.3
      - name: quay-bridge-operator.v3.7.2
        version: 3.7.2
      - name: quay-bridge-operator.v3.7.1
        version: 3.7.1
      - name: quay-bridge-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: quay-bridge-operator.v3.8.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:f8544a546dd6d209355c3a1cbe35138f6910491fc4584b401634431da4414ec0
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.8.15
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:975f287513039b2244894b9031e590f665d48f9af480dfdc4b85d4f96dd4bc37
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:f8544a546dd6d209355c3a1cbe35138f6910491fc4584b401634431da4414ec0
        version: 3.8.15
      entries:
      - name: quay-bridge-operator.v3.8.15
        version: 3.8.15
      - name: quay-bridge-operator.v3.8.14
        version: 3.8.14
      - name: quay-bridge-operator.v3.8.13
        version: 3.8.13
      - name: quay-bridge-operator.v3.8.12
        version: 3.8.12
      - name: quay-bridge-operator.v3.8.11
        version: 3.8.11
      - name: quay-bridge-operator.v3.8.10
        version: 3.8.10
      - name: quay-bridge-operator.v3.8.9
        version: 3.8.9
      - name: quay-bridge-operator.v3.8.8
        version: 3.8.8
      - name: quay-bridge-operator.v3.8.7
        version: 3.8.7
      - name: quay-bridge-operator.v3.8.6
        version: 3.8.6
      - name: quay-bridge-operator.v3.8.5
        version: 3.8.5
      - name: quay-bridge-operator.v3.8.4
        version: 3.8.4
      - name: quay-bridge-operator.v3.8.3
        version: 3.8.3
      - name: quay-bridge-operator.v3.8.2
        version: 3.8.2
      - name: quay-bridge-operator.v3.8.1
        version: 3.8.1
      - name: quay-bridge-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-bridge-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:875e0b63fdb7d44879650eb2f788ea705d753520841a5d875428f0b3ca70a36c
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.9.16
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:875e0b63fdb7d44879650eb2f788ea705d753520841a5d875428f0b3ca70a36c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 3.9.16
      entries:
      - name: quay-bridge-operator.v3.9.16
        version: 3.9.16
      - name: quay-bridge-operator.v3.9.15
        version: 3.9.15
      - name: quay-bridge-operator.v3.9.14
        version: 3.9.14
      - name: quay-bridge-operator.v3.9.13
        version: 3.9.13
      - name: quay-bridge-operator.v3.9.12
        version: 3.9.12
      - name: quay-bridge-operator.v3.9.11
        version: 3.9.11
      - name: quay-bridge-operator.v3.9.10
        version: 3.9.10
      - name: quay-bridge-operator.v3.9.9
        version: 3.9.9
      - name: quay-bridge-operator.v3.9.8
        version: 3.9.8
      - name: quay-bridge-operator.v3.9.7
        version: 3.9.7
      - name: quay-bridge-operator.v3.9.6
        version: 3.9.6
      - name: quay-bridge-operator.v3.9.5
        version: 3.9.5
      - name: quay-bridge-operator.v3.9.4
        version: 3.9.4
      - name: quay-bridge-operator.v3.9.3
        version: 3.9.3
      - name: quay-bridge-operator.v3.9.2
        version: 3.9.2
      - name: quay-bridge-operator.v3.9.1
        version: 3.9.1
      - name: quay-bridge-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: quay-bridge-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/securesign/secure-sign-operator
    name: rhtas-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhtas-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": {
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname-intermediate",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname-leaf",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname-root",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname-intermediate",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname-leaf",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname-root",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": {
                  "database": {
                    "create": true
                  },
                  "server": {
                    "replicas": 1
                  },
                  "signer": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:0ebc6b3bd992309a914cc0da8ad8fc628224f583b66f1ca2cd10d2281b781cde
          createdAt: "2025-09-30T11:33:48Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:e7f5a9a2ea76fdfde34f6aac264b51c596cf4ab53f6922b58514c77a1977677e
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:f2798c21dbcb3334b256bf1370a42c013f8f3eefb0641e1c2f250428a82ddb82
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:2dd696771c0188a119d746711551b1643a14a579d0fc66d70918cca9de0c0107
        - registry.redhat.io/rhtas/createtree-rhel9@sha256:244758721aa1c63adfffbae31df58059ff8a06311f6cb60515eeae14b20e78ce
        - registry.redhat.io/rhtas/rekor-monitor-rhel9@sha256:da3aa5c441653f00c4b558ef6ebf21fef518731e70f152a1fd3a750b1145d901
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:b311fb891203fef071d53c9dab67a45961638af98bdfd8e02026dc2dbc25fd00
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:8f8a5d4abf7878ba34151f8632aa7db25218300dbb9f17b121c5f61e0e5c5e03
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:eaae082559c81c5e1253f1b5cf9e298b14235470c67e9111c18ece46078228d5
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:2d91dceb216d8cb8dacefcfe86f357f4ff4eb5e5e998f264d1ad231821a92fad
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:043103f7c888d150053b873276b92b2ee1907c8a7238652ad0770e36d44c10e2
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:0ebc6b3bd992309a914cc0da8ad8fc628224f583b66f1ca2cd10d2281b781cde
        - registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:08302a0af012c1a1f04ef15de86558c9eef55620ab7b179fbda170025f0584be
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:497203991dcbca71275d691ec970ba480cb8e80c9ada6af876f5559abe0d6155
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:110509f5f24240d79afff7cb0d8a4ead0ed038c6100df8e435b01f270b7d83b2
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:abaa247777eb05bf8c97ba1ffc30c00d0d84590fac2fdfbffa06ba374b924888
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:fcb225019f865dd4b7af1f5134a9aed9c5c8a425f3ee43b74e0517ce602e209c
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:e6a5bf7804f1c4371e4de06805d6e1ddfc98e356e355ac3b55a3be25c5636b4d
        version: 1.3.0
      entries:
      - name: rhtas-operator.v1.3.0
        version: 1.3.0
      - name: rhtas-operator.v1.2.1
        version: 1.2.1
      - name: rhtas-operator.v1.2.0
        version: 1.2.0
      - name: rhtas-operator.v1.1.2
        version: 1.1.2
      - name: rhtas-operator.v1.1.1
        version: 1.1.1
      - name: rhtas-operator.v1.1.0
        version: 1.1.0
      - name: rhtas-operator.v1.0.2
        version: 1.0.2
      - name: rhtas-operator.v1.0.1
        version: 1.0.1
      - name: rhtas-operator.v1.0.0
        version: 1.0.0
      name: stable
    - currentCSV: rhtas-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e1a78636cd7cc8e7227b6b0a5ddfd93d002db0464b15e966e88337d38989ea5a
          createdAt: "2024-06-27T09:45:39Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e1a78636cd7cc8e7227b6b0a5ddfd93d002db0464b15e966e88337d38989ea5a
        version: 1.0.2
      entries:
      - name: rhtas-operator.v1.0.2
        version: 1.0.2
      - name: rhtas-operator.v1.0.1
        version: 1.0.1
      - name: rhtas-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: rhtas-operator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e52d07e2cd8cbdc2ccd0336a4fd29cbb2e5b8a1be628350294a54579cc2636ed
          createdAt: "2025-04-08T09:29:59Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e52d07e2cd8cbdc2ccd0336a4fd29cbb2e5b8a1be628350294a54579cc2636ed
        version: 1.1.2
      entries:
      - name: rhtas-operator.v1.1.2
        version: 1.1.2
      - name: rhtas-operator.v1.1.1
        version: 1.1.1
      - name: rhtas-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: rhtas-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
          createdAt: "2025-08-17T21:52:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:02fbd4772c185edf38fcee4e5c5e1c9dad82b77c9da824addaebc445b67d35ac
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
        - registry.redhat.io/ubi9/httpd-24@sha256:badeee0aa3477e61de64f6a2b93e8a3673edd38fc860125f7536506f3d592132
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:c22cdb9a17f85e13c00412cba83aa644e90a0faea0236029d0f1c53fe212e227
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:12b438ea70f83ee6edb12c96b507f92510f4afba824eac34c8f50793c4199db6
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:fc018a45b7eb48690c46c83f52cac55fa40089e6b02686bc21265dc7c2205b8f
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:1d782a127d2f8bdab633f287cb2df2e400ae91bf5d4b913473539d8512dfaab6
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:09b4aeeb607c88c72f69e6f87cb840c82ef5752c64a58fa13551db749fac2530
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:eb38e98dbb9828fe033a5388609132f7246fe77d2f4d258d015a94ea30752b24
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:6270ba29adf779de5d3916d5e82f1cd9491ba89d4ae2042347b6ad6d1819dd26
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:1c2201d50469d70ec6e21546c7c74bd52251d420dc1dcfa5375c1cf61dd3a9fd
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:030986ea26f33db3a192c67a93cd2bde352f23c68a17b062fc062955675c3c51
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:531603563c69aac6905cd6b8157d30781040527404ac2313acba5a9310e0b713
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:fe31830df2d99359749cb86cd1ac2f32fcc3de07aae427d2f2cd3998d31d2833
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:a93df3206409de75613d398e7dcd8b30ff7cc36eb00349c435784f73f00a2335
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:640633823a54c11fe2a8b3d05e571399fda415b6f0b3adf1ba9806882a94bdd4
        version: 1.2.1
      entries:
      - name: rhtas-operator.v1.2.1
        version: 1.2.1
      - name: rhtas-operator.v1.2.0
        version: 1.2.0
      name: stable-v1.2
    - currentCSV: rhtas-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": {
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname-intermediate",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname-leaf",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname-root",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname-intermediate",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname-leaf",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname-root",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": {
                  "database": {
                    "create": true
                  },
                  "server": {
                    "replicas": 1
                  },
                  "signer": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:0ebc6b3bd992309a914cc0da8ad8fc628224f583b66f1ca2cd10d2281b781cde
          createdAt: "2025-09-30T11:33:48Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:abaa247777eb05bf8c97ba1ffc30c00d0d84590fac2fdfbffa06ba374b924888
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:eaae082559c81c5e1253f1b5cf9e298b14235470c67e9111c18ece46078228d5
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:2d91dceb216d8cb8dacefcfe86f357f4ff4eb5e5e998f264d1ad231821a92fad
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:e6a5bf7804f1c4371e4de06805d6e1ddfc98e356e355ac3b55a3be25c5636b4d
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:8f8a5d4abf7878ba34151f8632aa7db25218300dbb9f17b121c5f61e0e5c5e03
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:2dd696771c0188a119d746711551b1643a14a579d0fc66d70918cca9de0c0107
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:110509f5f24240d79afff7cb0d8a4ead0ed038c6100df8e435b01f270b7d83b2
        - registry.redhat.io/rhtas/createtree-rhel9@sha256:244758721aa1c63adfffbae31df58059ff8a06311f6cb60515eeae14b20e78ce
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:b311fb891203fef071d53c9dab67a45961638af98bdfd8e02026dc2dbc25fd00
        - registry.redhat.io/rhtas/rekor-monitor-rhel9@sha256:da3aa5c441653f00c4b558ef6ebf21fef518731e70f152a1fd3a750b1145d901
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:08302a0af012c1a1f04ef15de86558c9eef55620ab7b179fbda170025f0584be
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:043103f7c888d150053b873276b92b2ee1907c8a7238652ad0770e36d44c10e2
        - registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:e7f5a9a2ea76fdfde34f6aac264b51c596cf4ab53f6922b58514c77a1977677e
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:fcb225019f865dd4b7af1f5134a9aed9c5c8a425f3ee43b74e0517ce602e209c
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:0ebc6b3bd992309a914cc0da8ad8fc628224f583b66f1ca2cd10d2281b781cde
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:497203991dcbca71275d691ec970ba480cb8e80c9ada6af876f5559abe0d6155
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:f2798c21dbcb3334b256bf1370a42c013f8f3eefb0641e1c2f250428a82ddb82
        version: 1.3.0
      entries:
      - name: rhtas-operator.v1.3.0
        version: 1.3.0
      name: stable-v1.3
    defaultChannel: stable
    packageName: rhtas-operator
    provider:
      name: Red Hat
      url: https://github.com/securesign/secure-sign-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-lambda-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-lambda-controller.v1.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "Alias",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "CodeSigningConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "EventSourceMapping",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/lambda-controller:1.9.1
          createdAt: "2025-09-19T19:23:07Z"
          description: AWS Lambda controller is a service controller for managing
            Lambda resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alias represents the state of an AWS lambda Alias resource.
            displayName: Alias
            kind: Alias
            name: aliases.lambda.services.k8s.aws
            version: v1alpha1
          - description: CodeSigningConfig represents the state of an AWS lambda CodeSigningConfig
              resource.
            displayName: CodeSigningConfig
            kind: CodeSigningConfig
            name: codesigningconfigs.lambda.services.k8s.aws
            version: v1alpha1
          - description: EventSourceMapping represents the state of an AWS lambda
              EventSourceMapping resource.
            displayName: EventSourceMapping
            kind: EventSourceMapping
            name: eventsourcemappings.lambda.services.k8s.aws
            version: v1alpha1
          - description: Function represents the state of an AWS lambda Function resource.
            displayName: Function
            kind: Function
            name: functions.lambda.services.k8s.aws
            version: v1alpha1
          - description: FunctionURLConfig represents the state of an AWS lambda FunctionURLConfig
              resource.
            displayName: FunctionURLConfig
            kind: FunctionURLConfig
            name: functionurlconfigs.lambda.services.k8s.aws
            version: v1alpha1
          - description: LayerVersion represents the state of an AWS lambda LayerVersion
              resource.
            displayName: LayerVersion
            kind: LayerVersion
            name: layerversions.lambda.services.k8s.aws
            version: v1alpha1
          - description: Version represents the state of an AWS lambda Version resource.
            displayName: Version
            kind: Version
            name: versions.lambda.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Lambda resources in AWS from within your Kubernetes cluster.

          **About Amazon Lambda**

          Lambda is a compute service that lets you run code without provisioning or managing servers. Lambda runs your code on a high-availability compute infrastructure and performs all of the administration of the compute resources, including server and operating system maintenance, capacity provisioning and automatic scaling, code monitoring and logging. With Lambda, you can run code for virtually any type of application or backend service. All you need to do is supply your code in one of the [languages that Lambda supports](https://docs.aws.amazon.com/lambda/latest/dg/lambda-runtimes.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Lambda
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lambda
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Lambda Developer Resources
          url: https://aws.amazon.com/lambda/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: lambda maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/lambda-controller:1.9.1
        version: 1.9.1
      entries:
      - name: ack-lambda-controller.v1.9.1
        version: 1.9.1
      - name: ack-lambda-controller.v1.9.0
        version: 1.9.0
      - name: ack-lambda-controller.v1.8.0
        version: 1.8.0
      - name: ack-lambda-controller.v1.7.7
        version: 1.7.7
      - name: ack-lambda-controller.v1.7.6
        version: 1.7.6
      - name: ack-lambda-controller.v1.7.5
        version: 1.7.5
      - name: ack-lambda-controller.v1.7.4
        version: 1.7.4
      - name: ack-lambda-controller.v1.7.2
        version: 1.7.2
      - name: ack-lambda-controller.v1.7.1
        version: 1.7.1
      - name: ack-lambda-controller.v1.7.0
        version: 1.7.0
      - name: ack-lambda-controller.v1.6.3
        version: 1.6.3
      - name: ack-lambda-controller.v1.6.2
        version: 1.6.2
      - name: ack-lambda-controller.v1.6.1
        version: 1.6.1
      - name: ack-lambda-controller.v1.6.0
        version: 1.6.0
      - name: ack-lambda-controller.v1.5.8
        version: 1.5.8
      - name: ack-lambda-controller.v1.5.7
        version: 1.5.7
      - name: ack-lambda-controller.v1.5.6
        version: 1.5.6
      - name: ack-lambda-controller.v1.5.5
        version: 1.5.5
      - name: ack-lambda-controller.v1.5.4
        version: 1.5.4
      - name: ack-lambda-controller.v1.5.3
        version: 1.5.3
      - name: ack-lambda-controller.v1.5.2
        version: 1.5.2
      - name: ack-lambda-controller.v1.5.1
        version: 1.5.1
      - name: ack-lambda-controller.v1.5.0
        version: 1.5.0
      - name: ack-lambda-controller.v1.4.2
        version: 1.4.2
      - name: ack-lambda-controller.v1.4.0
        version: 1.4.0
      - name: ack-lambda-controller.v1.3.4
        version: 1.3.4
      - name: ack-lambda-controller.v1.3.2
        version: 1.3.2
      - name: ack-lambda-controller.v1.3.1
        version: 1.3.1
      - name: ack-lambda-controller.v1.3.0
        version: 1.3.0
      - name: ack-lambda-controller.v1.2.1
        version: 1.2.1
      - name: ack-lambda-controller.v1.2.0
        version: 1.2.0
      - name: ack-lambda-controller.v1.1.1
        version: 1.1.1
      - name: ack-lambda-controller.v1.1.0
        version: 1.1.0
      - name: ack-lambda-controller.v1.0.1
        version: 1.0.1
      - name: ack-lambda-controller.v1.0.0
        version: 1.0.0
      - name: ack-lambda-controller.v0.1.7
        version: 0.1.7
      - name: ack-lambda-controller.v0.1.6
        version: 0.1.6
      - name: ack-lambda-controller.v0.1.5
        version: 0.1.5
      - name: ack-lambda-controller.v0.1.3
        version: 0.1.3
      - name: ack-lambda-controller.v0.1.2
        version: 0.1.2
      - name: ack-lambda-controller.v0.1.1
        version: 0.1.1
      - name: ack-lambda-controller.v0.1.0
        version: 0.1.0
      - name: ack-lambda-controller.v0.0.17
        version: 0.0.17
      - name: ack-lambda-controller.v0.0.16
        version: 0.0.16
      - name: ack-lambda-controller.v0.0.15
        version: 0.0.15
      - name: ack-lambda-controller.v0.0.14
        version: 0.0.14
      - name: ack-lambda-controller.v0.0.13
        version: 0.0.13
      - name: ack-lambda-controller.v0.0.12
        version: 0.0.12
      - name: ack-lambda-controller.v0.0.11
        version: 0.0.11
      - name: ack-lambda-controller.v0.0.10
        version: 0.0.10
      - name: ack-lambda-controller.v0.0.9
        version: 0.0.9
      - name: ack-lambda-controller.v0.0.8
        version: 0.0.8
      - name: ack-lambda-controller.v0.0.7
        version: 0.0.7
      - name: ack-lambda-controller.v0.0.6
        version: 0.0.6
      - name: ack-lambda-controller.v0.0.5
        version: 0.0.5
      - name: ack-lambda-controller.v0.0.4
        version: 0.0.4
      - name: ack-lambda-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-lambda-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-websocket-service
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-websocket-service.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "websocket.loadcore.keysight.com/v1alpha1",
                "kind": "Websocket",
                "metadata": {
                  "name": "websocket-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "websocket": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:26:39Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Websocket
            name: websockets.websocket.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Websocket Operator
        displayName: Keysight Websocket Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Websocket Service
          url: https://keysight-websocket-service.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-websocket-service@sha256:c91da3557ddcb8f052c9caa514d2dbf6c22b4963c080117a9d1859f1458fc404
        version: 0.1.21
      entries:
      - name: keysight-websocket-service.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-websocket-service
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/in-en/products/cloud-pak-for-data
    name: ibm-cpd-datastage-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-datastage-operator.v7.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ds.cpd.ibm.com/v1",
                "kind": "DataStage",
                "metadata": {
                  "name": "datastage"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "fileStorageClass": "managed-nfs-storage",
                  "scaleConfig": "small"
                }
              }
            ]
          capabilities: Seamless Upgrades
          cloudpakId: 49a42b864bb94569bef0188ead948f11
          cloudpakName: IBM DataStage Enterprise Plus Cartridge for IBM Cloud Pak
            for Data
          containerImage: registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
          description: The IBM DataStage Operator is used to deploy IBM DataStage
            services
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-datastage-operator
          olm.skipRange: '>=1.0.0 <7.3.0'
          operatorframework.io/arch.amd64: supported
          operatorframework.io/arch.s390x: supported
          operatorframework.io/os.linux: supported
          operatorframework.io/suggested-namespace: ibm-common-services
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation: [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)
              License: [IBM DataStage Enterprise Cartridge for Cloud Pak for Data](https://ibm.biz/Bdahic),
              [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)'
            kind: DataStage
            name: datastages.ds.cpd.ibm.com
            version: v1
          - description: Documentation [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)
              License [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)
            kind: PXRuntime
            name: pxruntimes.ds.cpd.ibm.com
            version: v1
        description: |-
          # Introduction
          Industry-leading batch and real-time data integration service for data engineers to build trusted pipelines (ETL/ELT) for AI and operational use cases across on-premises and hybrid cloud environments.
          With DataStage, you can design and run data flows that move and transform data anywhere, at any scale.
          No matter how complex your data landscape, DataStage can streamline your data movement costs and increase productivity. DataStage offers:
          <ul><li>A best-in-breed parallel processing engine that enables you to process your data where it resides</li><li>Automated job design</li><li>Simple integration with cloud data lakes, real-time data sources, relational databases, big data, and NoSQL data stores</li></ul>
          The DataStage service uses Cloud Pak for Data platform connections and integration points, with services like Data Virtualization, to simplify the process of connecting to and accessing your data.

          With DataStage, Data Engineers can use the simple user interface to build no-code/low-code data pipelines. The interface offers hundreds of functions and connectors that reduce development time and inconsistencies across pipelines. The interface also makes it easy to collaborate with your peers and control access to specific analytics projects.
          The service also provides automatic workload balancing to provide high performance pipelines that make efficient use of available compute resources.

          ## Details
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=SSQNUZ_4.0/svc-welcome/dsent.html)

          ## Documentation

          [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)

          ## Installing
          To install the operator follow the the installation and configuration instructions within the [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=datastage-installing).

          ## Prerequisites

          - OpenShift Version >= 4.6
          - IBM Cloud Pak for Data >= 4.0.2
          - IBM Cloud Pak for Data Common Core Services (CCS) >= 2.0.2 (The DataStage service automatically installs the CCS if they are not already installed.)

          ### Resources Required

          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=requirements-hardware)

          ## Configuration

          To scale DataStage services, you can specify spec.scaleConfig in the DataStage custom resource to small, medium, or large.

          ## License
          [IBM DataStage Enterprise Cartridge for Cloud Pak for Data](https://ibm.biz/Bdahic)
          [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)

          ## Limitations
          - Only the `x86-64` platform is supported by this operator.
          - See [Limitations and known issues in DataStage](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=issues-datastage)

          ## SecurityContextConstraints Requirements
          Custom SecurityContextConstraints definition:
          ```
          apiVersion: security.openshift.io/v1
          kind: SecurityContextConstraints
          allowHostDirVolumePlugin: false
          allowHostIPC: false
          allowHostNetwork: false
          allowHostPID: false
          allowHostPorts: false
          allowPrivilegeEscalation: true
          allowPrivilegedContainer: false
          allowedCapabilities: null
          defaultAddCapabilities: null
          fsGroup:
            type: MustRunAs
          metadata:
            annotations:
              include.release.openshift.io/self-managed-high-availability: "true"
              kubernetes.io/description: restrict denies access to all host features and requires pods to be run with a UID, and SELinux context that are allocated to the namespace.  This is the most restrictive SCC and it is used by default for authenticated users.
              release.openshift.io/create-only: "true"
            name: restrict
          priority: null
          readOnlyRootFilesystem: false
          requiredDropCapabilities:
          - KILL
          - MKNOD
          - SETUID
          - SETGID
          runAsUser:
            type: MustRunAsRange
          seLinuxContext:
            type: MustRunAs
          supplementalGroups:
            type: RunAsAny
          users: []
          volumes:
          - configMap
          - downwardAPI
          - emptyDir
          - persistentVolumeClaim
          - projected
          - secret
          ```
        displayName: IBM DataStage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DataStage
        - Analytics
        - Cloud Pak
        - Data
        - Cloud Pak for Data
        - IBM
        maintainers:
        - email: cp4d@ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/in-en/products/cloud-pak-for-data
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:10492010b28f57f8badbc7cadbe15d1447269052b387d878c6d9ecbd2f62923b
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:046eb99c8f1f37dd4cffa1c6c9a93aea50bb5d10a549aa12c3c90134d03c6d1c
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:1b39e1d5d7c13451252250ceeb29b91c1e7ef3739fc17baf9958e1bf07c4c2fd
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:8b72cafe9de355634bc3218cc1dc701ffb376a29aecb5cd12a5719d43a1304bd
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:6c569c4687546a03a5f4e127acc13b219ce3efa057fbdfd55e96868e549c70cb
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:ef03bb6ed647d92b9c84d34810376da7049aad88589eb2e35798d82a22f5b994
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:4604a076c837600cd4b3e29b31e22470915fe5b6b1191ae9401407f8620f2472
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:99fc418fb1a09ee6c45e90dfba47f73ea8e1da1b19f9f570d2a749b68118b7a3
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:3eecc54f26cd87870c201b5ade1d037d2733ccfda40fc1aedd9a2164e582275d
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:7015b410b947338dcafde98b68b95cb4199c08511ae46e733691e8fa835ad988
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:001bf35df08cd397397c3fdd1ddbd7953cdf4bde06965427ae640e5608e60ddf
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:3f8133fcd904befa579b99d3a828fe737faf13ae14e842ed0408eeeee72de5fb
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:5f5faf7d15f46d79eea873c12ff631ed59bc55bc57f1f7aee9fd0a8f35fdd3cf
        version: 7.3.0
      entries:
      - name: ibm-cpd-datastage-operator.v7.3.0
        version: 7.3.0
      name: v7.3
    defaultChannel: v7.3
    packageName: ibm-cpd-datastage-operator
    provider:
      name: IBM
      url: https://www.ibm.com/in-en/products/cloud-pak-for-data
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Advanced Micro Devices, Inc.
      provider-url: ""
    name: amd-gpu-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amd-gpu-operator.v1.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "amd.com/v1alpha1",
                "kind": "DeviceConfig",
                "metadata": {
                  "name": "test-deviceconfig",
                  "namespace": "openshift-amd-gpu"
                },
                "spec": {
                  "devicePlugin": {
                    "devicePluginImage": "rocm/k8s-device-plugin:rhubi-latest",
                    "nodeLabellerImage": "rocm/k8s-node-labeller:rhubi-latest"
                  },
                  "driver": {
                    "enable": "true",
                    "version": "6.4.2"
                  },
                  "selector": {
                    "feature.node.kubernetes.io/amd-gpu": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning,Monitoring
          containerImage: docker.io/rocm/gpu-operator@sha256:1583d4ec172d50d4416f4030eb85910ec54633b7e8e061d4bd79b14dbf98d28c
          createdAt: "2025-08-01T01:55:48Z"
          description: |-
            Operator responsible for deploying AMD GPU kernel drivers, device plugin, device test runner and device metrics exporter
            For more information, visit [documentation](https://instinct.docs.amd.com/projects/gpu-operator/en/latest/)
          deviceConfigManagerImage: docker.io/rocm/device-config-manager@sha256:d2896271eacfc664ac7651425f46a85af4f4ef98d713988c2a4fee4a846548ee
          devicePluginImage: docker.io/rocm/k8s-device-plugin@sha256:840c4ea1bf58df60ef3e80e3bf27bbf620a393854e42f58b2224b288767e3dd4
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          metricsExporterImage: docker.io/rocm/device-metrics-exporter@sha256:6c946c6ca10d5c0f40e67b5759759f12d51c45e3b934cb1af912821d5fd76e6e
          nodelabellerImage: docker.io/rocm/k8s-node-labeller@sha256:75f916c5a407e40e9ffdd26cdfc65b7da855710568937e740a969094e2991065
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-amd-gpu
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ROCm/gpu-operator
          support: Advanced Micro Devices, Inc.
          testRunnerImage: docker.io/rocm/test-runner@sha256:98e698263b4df26959341b14b3ddded9df51e504e7c6819d5c7c56c434855e9d
          utilsContainerImage: docker.io/rocm/gpu-operator-utils@sha256:9cbe60dce5b3a2ffd93f7148a363ed5f072a320377d5e5bcbbfb1e05d4959287
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DeviceConfig describes how to enable AMD GPU device
            displayName: DeviceConfig
            kind: DeviceConfig
            name: deviceconfigs.amd.com
            version: v1alpha1
        description: |-
          Operator responsible for deploying AMD GPU kernel drivers, device plugin, device test runner and device metrics exporter
          For more information, visit [documentation](https://instinct.docs.amd.com/projects/gpu-operator/en/latest/)
        displayName: amd-gpu-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AMD
        - GPU
        - AI
        - Deep Learning
        - Hardware
        - Driver
        - Monitoring
        links:
        - name: AMD GPU Operator
          url: https://github.com/ROCm/gpu-operator
        maintainers:
        - email: Yan.Sun3@amd.com
          name: Yan Sun
        - email: farshad.ghodsian@amd.com
          name: Farshad Ghodsian
        - email: shrey.ajmera@amd.com
          name: Shrey Ajmera
        maturity: stable
        provider:
          name: Advanced Micro Devices, Inc.
        relatedImages:
        - docker.io/rocm/gpu-operator@sha256:1583d4ec172d50d4416f4030eb85910ec54633b7e8e061d4bd79b14dbf98d28c
        - docker.io/rocm/k8s-device-plugin@sha256:840c4ea1bf58df60ef3e80e3bf27bbf620a393854e42f58b2224b288767e3dd4
        - docker.io/rocm/k8s-node-labeller@sha256:75f916c5a407e40e9ffdd26cdfc65b7da855710568937e740a969094e2991065
        - docker.io/rocm/gpu-operator-utils@sha256:9cbe60dce5b3a2ffd93f7148a363ed5f072a320377d5e5bcbbfb1e05d4959287
        - docker.io/rocm/test-runner@sha256:98e698263b4df26959341b14b3ddded9df51e504e7c6819d5c7c56c434855e9d
        - docker.io/rocm/device-config-manager@sha256:d2896271eacfc664ac7651425f46a85af4f4ef98d713988c2a4fee4a846548ee
        - docker.io/rocm/device-metrics-exporter@sha256:6c946c6ca10d5c0f40e67b5759759f12d51c45e3b934cb1af912821d5fd76e6e
        version: 1.3.2
      entries:
      - name: amd-gpu-operator.v1.3.2
        version: 1.3.2
      - name: amd-gpu-operator.v1.3.1
        version: 1.3.1
      - name: amd-gpu-operator.v1.2.1
        version: 1.2.1
      - name: amd-gpu-operator.v1.1.1
        version: 1.1.1
      name: alpha
    defaultChannel: alpha
    packageName: amd-gpu-operator
    provider:
      name: Advanced Micro Devices, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-zookeeper-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zookeeper-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:1d36865fe1cac88c9edc63c0e35e42a80456486668de5a575f10cc4b445d47db
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:1d36865fe1cac88c9edc63c0e35e42a80456486668de5a575f10cc4b445d47db
        version: 24.11.1
      entries:
      - name: zookeeper-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: zookeeper-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:6a4729289275e473907a3792c001dbe7dd2dc0498aabc719cf4175d956e3c87f
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:6a4729289275e473907a3792c001dbe7dd2dc0498aabc719cf4175d956e3c87f
        version: 25.3.0
      entries:
      - name: zookeeper-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: zookeeper-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
        version: 25.7.0
      entries:
      - name: zookeeper-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: zookeeper-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
        version: 25.7.0
      entries:
      - name: zookeeper-operator.v25.7.0
        version: 25.7.0
      - name: zookeeper-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-zookeeper-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cluster-kube-descheduler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clusterkubedescheduleroperator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "KubeDescheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-kube-descheduler-operator"
                },
                "spec": {
                  "deschedulingIntervalSeconds": 3600,
                  "managementState": "Managed",
                  "mode": "Predictive",
                  "profiles": ["AffinityAndTaints"]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-cluster-kube-descheduler-operator@sha256:ba0b71ff2a30a069b4a8a8f3c1e0898aaadc6db112e4cc12aff7c77ced7a0405
          createdAt: 2021/10/13
          description: An operator to run descheduler in Openshift cluster.
          olm.skipRange: '>=4.10.0 <4.13.0-202309181427'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-kube-descheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeDescheduler is the Schema for the deschedulers API
            displayName: Kube Descheduler
            kind: KubeDescheduler
            name: kubedeschedulers.operator.openshift.io
            version: v1
        description: |
          The Kube Descheduler Operator provides the ability to evict a running pod so that the pod can be rescheduled onto a more suitable node.

          There are several situations where descheduling can benefit your cluster:

          * Nodes are underutilized or overutilized.
          * Pod and node affinity requirements, such as taints or labels, have changed and the original scheduling decisions are no longer appropriate for certain nodes.
          * Node failure requires pods to be moved.
          * New nodes are added to clusters.

          ## Descheduler Profiles

          Once the operator is installed, you can configure one or more profiles to identify pods to evict. The scheduler will schedule the replacement of the evicted pods.

          The following profiles are available:

          * AffinityAndTaints
          * TopologyAndDuplicates
          * SoftTopologyAndDuplicates
          * LifecycleAndUtilization
          * EvictPodsWithPVC
          * EvictPodsWithLocalStorage

          These profiles are documented in detail in the [descheduler operator README](https://github.com/openshift/cluster-kube-descheduler-operator#profiles).

          ## Additional Parameters

          In addition to the profiles, the following parameters can be configured:

          * `deschedulingIntervalSeconds` - Set the number of seconds between descheduler runs. A value of `0` in this field runs the descheduler once and exits.
          * `profileCustomizations` - Allows certain profile parameters to be tweaked, such as `podLifetime` (see [README](https://github.com/openshift/cluster-kube-descheduler-operator#profile-customizations) for more info).
        displayName: Kube Descheduler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - scheduling
        - descheduler
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/cluster-kube-descheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-cluster-kube-descheduler-operator@sha256:ba0b71ff2a30a069b4a8a8f3c1e0898aaadc6db112e4cc12aff7c77ced7a0405
        - registry.redhat.io/openshift4/ose-descheduler@sha256:45dc69ad93ab50bdf9ce1bb79f6d98f849e320db68af30475b10b7f5497a1b13
        version: 4.13.0-202309181427
      entries:
      - name: clusterkubedescheduleroperator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: "4.13"
    - currentCSV: clusterkubedescheduleroperator.v5.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "KubeDescheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-kube-descheduler-operator"
                },
                "spec": {
                  "deschedulingIntervalSeconds": 3600,
                  "managementState": "Managed",
                  "mode": "Predictive",
                  "profiles": ["AffinityAndTaints"]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/kube-descheduler-operator/kube-descheduler-rhel9-operator@sha256:03404d74da227fdf5805f61657272af2c915ce0014d6a902ef384b7a70bd10e2
          createdAt: 2024/10/30
          description: An operator to run descheduler in Openshift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.12.0 <5.0.2'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-kube-descheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeDescheduler is the Schema for the deschedulers API
            displayName: Kube Descheduler
            kind: KubeDescheduler
            name: kubedeschedulers.operator.openshift.io
            version: v1
        description: |
          The Kube Descheduler Operator provides the ability to evict a running pod so that the pod can be rescheduled onto a more suitable node.

          There are several situations where descheduling can benefit your cluster:

          * Nodes are underutilized or overutilized.
          * Pod and node affinity requirements, such as taints or labels, have changed and the original scheduling decisions are no longer appropriate for certain nodes.
          * Node failure requires pods to be moved.
          * New nodes are added to clusters.

          ## Descheduler Profiles

          Once the operator is installed, you can configure one or more profiles to identify pods to evict. The scheduler will schedule the replacement of the evicted pods.

          The following profiles are available:

          * AffinityAndTaints
          * TopologyAndDuplicates
          * SoftTopologyAndDuplicates
          * LifecycleAndUtilization
          * EvictPodsWithPVC
          * EvictPodsWithLocalStorage

          These profiles are documented in detail in the [descheduler operator README](https://github.com/openshift/cluster-kube-descheduler-operator#profiles).

          ## Additional Parameters

          In addition to the profiles, the following parameters can be configured:

          * `deschedulingIntervalSeconds` - Set the number of seconds between descheduler runs. A value of `0` in this field runs the descheduler once and exits.
          * `profileCustomizations` - Allows certain profile parameters to be tweaked, such as `podLifetime` (see [README](https://github.com/openshift/cluster-kube-descheduler-operator#profile-customizations) for more info).
        displayName: Kube Descheduler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - scheduling
        - descheduler
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/cluster-kube-descheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/kube-descheduler-operator/kube-descheduler-rhel9-operator@sha256:03404d74da227fdf5805f61657272af2c915ce0014d6a902ef384b7a70bd10e2
        - registry.redhat.io/kube-descheduler-operator/descheduler-rhel9@sha256:7ba3f46cbbec2e47edcc255fae405bdbf2b9df3a232fc6809ce65fb1827a3b45
        version: 5.0.2
      entries:
      - name: clusterkubedescheduleroperator.v5.0.2
        version: 5.0.2
      - name: clusterkubedescheduleroperator.4.14.0-202311021650
        version: 4.14.0-202311021650
      - name: clusterkubedescheduleroperator.4.14.0-202310201027
        version: 4.14.0-202310201027
      name: stable
    defaultChannel: stable
    packageName: cluster-kube-descheduler-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: machine-deletion-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: machine-deletion-remediation.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:420a4d4a8610a3694fd3ef9809c55a4f140ec8cea7f09219647f82a84d1f411b
          createdAt: "2023-10-30 13:59:28"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          olm.skipRange: '>=0.1.0 <0.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:420a4d4a8610a3694fd3ef9809c55a4f140ec8cea7f09219647f82a84d1f411b
        version: 0.2.1
      entries:
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.14-eus
    - currentCSV: machine-deletion-remediation.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:1445e6df8f49672aa729806f0e257043ecc22e19e666bb5a799286c6b878b1dd
          createdAt: "2024-08-11 06:47:20"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.2.0 <0.3.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f4381aaf002b761a2ba614332a2af973f4de5f6a4c84dd96b8808725e561fa89
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:1445e6df8f49672aa729806f0e257043ecc22e19e666bb5a799286c6b878b1dd
        version: 0.3.1
      entries:
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.16-eus
    - currentCSV: machine-deletion-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-29 09:35:10"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.3.1 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.6/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel9-operator@sha256:dbe9095d494ebc7b56982cb1f74329d7a7fd0cbb7544e9f90207eb35de7997dc
        version: 0.4.1
      entries:
      - name: machine-deletion-remediation.v0.4.1
        version: 0.4.1
      - name: machine-deletion-remediation.v0.4.0
        version: 0.4.0
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.18-eus
    - currentCSV: machine-deletion-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:7e1cad02f1e74a1355c168fc53925bd0fbcc55f2535a46f734c8de082eb49400
          createdAt: "2023-04-24 08:22:17"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3af07bfe0fb73b1a7f00487eba302d6da442e742ab4007fc50e52efd4d45dae5
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:7e1cad02f1e74a1355c168fc53925bd0fbcc55f2535a46f734c8de082eb49400
        version: 0.1.0
      entries:
      - name: machine-deletion-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: machine-deletion-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-29 09:35:10"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.3.1 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.6/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel9-operator@sha256:dbe9095d494ebc7b56982cb1f74329d7a7fd0cbb7544e9f90207eb35de7997dc
        version: 0.4.1
      entries:
      - name: machine-deletion-remediation.v0.4.1
        version: 0.4.1
      - name: machine-deletion-remediation.v0.4.0
        version: 0.4.0
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: stable
    defaultChannel: stable
    packageName: machine-deletion-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: sailoperator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sailoperator.v1.28.0-nightly-2025-10-06
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:07ea72de6ff9184f346ee2948228e73cfe4180c8d05fe7ee972c22378ad450ef
          createdAt: "2025-10-06T03:29:39Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0
          - master
          - v1.28-alpha.8de61b9a

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-testing/ztunnel@sha256:38b5902bf36f5163274ce0963dfb1e106000ac639d429f05846448de8454492d
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-testing/proxyv2@sha256:11cf6a3083a3c1ffa539a21cee30ddc0abe47fb44bb16520cdc2f4141dcfee03
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - quay.io/sail-dev/sail-operator@sha256:07ea72de6ff9184f346ee2948228e73cfe4180c8d05fe7ee972c22378ad450ef
        - gcr.io/istio-testing/pilot@sha256:13e2125870f59f752dc5834b702c2dab28f36edc00dccdeeaf872023018e3df0
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-testing/install-cni@sha256:9525a3711201bcca04ce2ca9b1576d4094a03779d60db0a36b15a956020bbec1
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        version: 1.28.0-nightly-2025-10-06
      entries:
      - name: sailoperator.v1.28.0-nightly-2025-10-06
        version: 1.28.0-nightly-2025-10-06
      - name: sailoperator.v1.28.0-nightly-2025-10-05
        version: 1.28.0-nightly-2025-10-05
      - name: sailoperator.v1.28.0-nightly-2025-10-04
        version: 1.28.0-nightly-2025-10-04
      - name: sailoperator.v1.28.0-nightly-2025-10-03
        version: 1.28.0-nightly-2025-10-03
      - name: sailoperator.v1.28.0-nightly-2025-10-02
        version: 1.28.0-nightly-2025-10-02
      - name: sailoperator.v1.28.0-nightly-2025-10-01
        version: 1.28.0-nightly-2025-10-01
      - name: sailoperator.v1.28.0-nightly-2025-09-30
        version: 1.28.0-nightly-2025-09-30
      - name: sailoperator.v1.28.0-nightly-2025-09-29
        version: 1.28.0-nightly-2025-09-29
      - name: sailoperator.v1.28.0-nightly-2025-09-28
        version: 1.28.0-nightly-2025-09-28
      - name: sailoperator.v1.28.0-nightly-2025-09-27
        version: 1.28.0-nightly-2025-09-27
      - name: sailoperator.v1.28.0-nightly-2025-09-26
        version: 1.28.0-nightly-2025-09-26
      - name: sailoperator.v1.28.0-nightly-2025-09-25
        version: 1.28.0-nightly-2025-09-25
      - name: sailoperator.v1.28.0-nightly-2025-09-23
        version: 1.28.0-nightly-2025-09-23
      - name: sailoperator.v1.28.0-nightly-2025-09-22
        version: 1.28.0-nightly-2025-09-22
      name: 1.28-nightly
    - currentCSV: sailoperator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.23.0"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.23.0"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/maistra-dev/sail-operator:0.1-latest
          createdAt: "2024-08-20T11:54:30Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1alpha1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1alpha1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1alpha1
          - description: |-
              RemoteIstio represents a remote Istio Service Mesh deployment consisting of one or more
              remote control plane instances (represented by one or more IstioRevision objects).
            displayName: Remote Istio
            kind: RemoteIstio
            name: remoteistios.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.23.0
          - v1.22.4

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/maistra-dev/sail-operator:0.1-latest
        - docker.io/istio/ztunnel:1.22.4
        - docker.io/istio/install-cni:1.23.0
        - docker.io/istio/pilot:1.23.0
        - docker.io/istio/proxyv2:1.23.0
        - docker.io/istio/ztunnel:1.23.0
        - docker.io/istio/install-cni:1.22.4
        - docker.io/istio/pilot:1.22.4
        - docker.io/istio/proxyv2:1.22.4
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.16.0
        version: 0.1.0
      entries:
      - name: sailoperator.v0.1.0
        version: 0.1.0
      - name: sailoperator.v0.1.0-rc.1
        version: 0.1.0-rc.1
      name: candidates
    - currentCSV: sailoperator.v1.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
          createdAt: "2025-09-05T10:12:19Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        version: 1.27.1
      entries:
      - name: sailoperator.v1.27.1
        version: 1.27.1
      - name: sailoperator.v1.27.0
        version: 1.27.0
      - name: sailoperator.v1.26.3
        version: 1.26.3
      - name: sailoperator.v1.26.2
        version: 1.26.2
      - name: sailoperator.v1.26.1
        version: 1.26.1
      - name: sailoperator.v1.26.0
        version: 1.26.0
      - name: sailoperator.v1.25.2
        version: 1.25.2
      - name: sailoperator.v1.25.1
        version: 1.25.1
      - name: sailoperator.v1.25.0
        version: 1.25.0
      name: stable
    - currentCSV: sailoperator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.3"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.3"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator:1.0-latest
          createdAt: "2025-03-07T10:17:30Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.24-latest
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.23-latest
          - v1.23.5
          - v1.23.4
          - v1.23.3
          - v1.23.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/ztunnel:1.23.5
        - gcr.io/istio-release/proxyv2:1.24.1
        - gcr.io/istio-release/proxyv2:1.23.4
        - gcr.io/istio-release/pilot:1.24.3
        - gcr.io/istio-release/ztunnel:1.24.3
        - gcr.io/istio-release/ztunnel:1.24.1
        - gcr.io/istio-release/proxyv2:1.23.0
        - gcr.io/istio-release/pilot:1.24.2
        - gcr.io/istio-release/install-cni:1.24.1
        - gcr.io/istio-release/ztunnel:1.23.4
        - gcr.io/istio-release/pilot:1.23.0
        - gcr.io/istio-release/install-cni:1.23.0
        - gcr.io/istio-release/pilot:1.23.3
        - gcr.io/istio-release/proxyv2:1.23.5
        - gcr.io/istio-release/pilot:1.23.5
        - gcr.io/istio-release/ztunnel:1.24.2
        - gcr.io/istio-release/pilot:1.24.1
        - gcr.io/istio-release/ztunnel:1.23.0
        - gcr.io/istio-release/ztunnel:1.23.3
        - gcr.io/istio-release/install-cni:1.24.3
        - gcr.io/istio-release/install-cni:1.23.4
        - gcr.io/istio-release/proxyv2:1.23.3
        - gcr.io/istio-release/install-cni:1.23.5
        - gcr.io/istio-release/proxyv2:1.24.3
        - gcr.io/istio-release/install-cni:1.23.3
        - quay.io/sail-dev/sail-operator:1.0-latest
        - gcr.io/istio-release/install-cni:1.24.2
        - gcr.io/istio-release/proxyv2:1.24.2
        - gcr.io/istio-release/pilot:1.23.4
        version: 1.0.0
      entries:
      - name: sailoperator.v1.0.0
        version: 1.0.0
      name: stable-1.0
    - currentCSV: sailoperator.v1.25.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.25.2"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.25.2"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator:1.25.2
          createdAt: "2025-05-07T19:44:51Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.25.2'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.25-latest
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0
          - v1.23-latest
          - v1.23.5
          - v1.23.4
          - v1.23.3
          - v1.23.2
          - v1.22-latest
          - v1.22.8
          - v1.22.7
          - v1.22.6
          - v1.22.5
          - v1.21.6

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/proxyv2:1.25.1
        - gcr.io/istio-release/install-cni:1.24.1
        - gcr.io/istio-release/ztunnel:1.25.1
        - gcr.io/istio-release/pilot:1.25.2
        - gcr.io/istio-release/pilot:1.22.7
        - gcr.io/istio-release/ztunnel:1.23.4
        - gcr.io/istio-release/install-cni:1.23.2
        - quay.io/sail-dev/sail-operator:1.25.2
        - gcr.io/istio-release/install-cni:1.22.8
        - gcr.io/istio-release/ztunnel:1.22.5
        - gcr.io/istio-release/install-cni:1.23.3
        - gcr.io/istio-release/proxyv2:1.24.1
        - gcr.io/istio-release/install-cni:1.23.5
        - gcr.io/istio-release/pilot:1.23.5
        - gcr.io/istio-release/proxyv2:1.24.2
        - gcr.io/istio-release/install-cni:1.24.0
        - gcr.io/istio-release/pilot:1.24.0
        - gcr.io/istio-release/pilot:1.22.8
        - gcr.io/istio-release/install-cni:1.23.4
        - gcr.io/istio-release/pilot:1.24.4
        - gcr.io/istio-release/pilot:1.22.6
        - gcr.io/istio-release/ztunnel:1.21.6
        - gcr.io/istio-release/pilot:1.23.4
        - gcr.io/istio-release/proxyv2:1.23.3
        - gcr.io/istio-release/install-cni:1.25.2
        - gcr.io/istio-release/proxyv2:1.23.4
        - gcr.io/istio-release/ztunnel:1.25.2
        - gcr.io/istio-release/install-cni:1.21.6
        - gcr.io/istio-release/ztunnel:1.23.3
        - gcr.io/istio-release/pilot:1.22.5
        - gcr.io/istio-release/install-cni:1.22.5
        - gcr.io/istio-release/proxyv2:1.22.7
        - gcr.io/istio-release/proxyv2:1.22.8
        - gcr.io/istio-release/ztunnel:1.24.0
        - gcr.io/istio-release/proxyv2:1.24.3
        - gcr.io/istio-release/ztunnel:1.22.6
        - gcr.io/istio-release/pilot:1.24.2
        - gcr.io/istio-release/ztunnel:1.22.7
        - gcr.io/istio-release/install-cni:1.22.7
        - gcr.io/istio-release/ztunnel:1.23.2
        - gcr.io/istio-release/proxyv2:1.23.5
        - gcr.io/istio-release/pilot:1.24.3
        - gcr.io/istio-release/proxyv2:1.21.6
        - gcr.io/istio-release/install-cni:1.24.4
        - gcr.io/istio-release/install-cni:1.24.2
        - gcr.io/istio-release/install-cni:1.24.3
        - gcr.io/istio-release/ztunnel:1.23.5
        - gcr.io/istio-release/proxyv2:1.24.4
        - gcr.io/istio-release/ztunnel:1.24.3
        - gcr.io/istio-release/proxyv2:1.25.2
        - gcr.io/istio-release/pilot:1.23.2
        - gcr.io/istio-release/proxyv2:1.22.5
        - gcr.io/istio-release/install-cni:1.22.6
        - gcr.io/istio-release/pilot:1.23.3
        - gcr.io/istio-release/proxyv2:1.23.2
        - gcr.io/istio-release/ztunnel:1.24.4
        - gcr.io/istio-release/install-cni:1.25.1
        - gcr.io/istio-release/ztunnel:1.22.8
        - gcr.io/istio-release/ztunnel:1.24.2
        - gcr.io/istio-release/pilot:1.24.1
        - gcr.io/istio-release/pilot:1.25.1
        - gcr.io/istio-release/proxyv2:1.22.6
        - gcr.io/istio-release/pilot:1.21.6
        - gcr.io/istio-release/proxyv2:1.24.0
        - gcr.io/istio-release/ztunnel:1.24.1
        version: 1.25.2
      entries:
      - name: sailoperator.v1.25.2
        version: 1.25.2
      - name: sailoperator.v1.25.1
        version: 1.25.1
      - name: sailoperator.v1.25.0
        version: 1.25.0
      name: stable-1.25
    - currentCSV: sailoperator.v1.26.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.3"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.3"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:d37258646d76a9f708a69fbaa081a0a95908a2b0c8837a14d7ecbfcf505f9bd4
          createdAt: "2025-08-04T06:23:45Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.26-latest
          - v1.26.3
          - v1.26.2
          - v1.26.1
          - v1.26.0
          - v1.25-latest
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/ztunnel@sha256:2d567893e5b452d5c0b9c43a43efd4f8e28e639a344e28d90498f13858f13d46
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/proxyv2@sha256:fd734e6031566b4fb92be38f0f6bb02fdba6c199c45c2db5dc988bbc4fdee026
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/install-cni@sha256:8e782f5077f63c0a15fd4b9a4fa75f8543033c28ccf2457e25afbf1aefb6a72a
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/pilot@sha256:c4e33921e030bb22d3b8fabc693851d52c07a7af79a9534addc2f50005fa462a
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - quay.io/sail-dev/sail-operator@sha256:d37258646d76a9f708a69fbaa081a0a95908a2b0c8837a14d7ecbfcf505f9bd4
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        version: 1.26.3
      entries:
      - name: sailoperator.v1.26.3
        version: 1.26.3
      - name: sailoperator.v1.26.2
        version: 1.26.2
      - name: sailoperator.v1.26.1
        version: 1.26.1
      - name: sailoperator.v1.26.0
        version: 1.26.0
      name: stable-1.26
    - currentCSV: sailoperator.v1.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
          createdAt: "2025-09-05T10:12:19Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        version: 1.27.1
      entries:
      - name: sailoperator.v1.27.1
        version: 1.27.1
      - name: sailoperator.v1.27.0
        version: 1.27.0
      name: stable-1.27
    defaultChannel: stable
    packageName: sailoperator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: keepalived-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keepalived-operator.v1.5.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KeepalivedGroup",
                "metadata": {
                  "name": "keepalivedgroup-workers"
                },
                "spec": {
                  "blacklistRouterIDs": [
                    1,
                    2
                  ],
                  "image": "registry.redhat.io/openshift4/ose-keepalived-ipfailover",
                  "interface": "ens3",
                  "nodeSelector": {
                    "node-role.kubernetes.io/loadbalancer": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Networking, Cloud Provider
          certified: "false"
          containerImage: quay.io/redhat-cop/keepalived-operator@sha256:cd0c2f7df82df1178795e4a18a0942da33adb9670b192a7b7a2b1483a5b4ccf8
          createdAt: "2023-03-11T18:48:22Z"
          description: This operator provides a simple way to create self-hosted load
            balancers using keepalived.
          operatorframework.io/suggested-namespace: keepalived-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/keepalived-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KeepalivedGroup
            name: keepalivedgroups.redhatcop.redhat.io
            version: v1alpha1
        description: "The objective of the keepalived operator is to allow for a way
          to create self-hosted load balancers in an automated way. From a user experience
          point of view the behavior is the same as of when creating [`LoadBalancer`](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          services with a cloud provider able to manage them.\n\nThe keepalived operator
          can be used in all environments that allows nodes to advertise additional
          IPs on their NICs (and at least for now, in networks that allow multicast),
          however it's mainly aimed at supporting LoadBalancer services and ExternalIPs
          on bare metal installations (or other installation environments where a
          cloud provider is not available).\n\nOne possible use of the keepalived
          operator is also to support [OpenShift Ingresses](https://docs.openshift.com/container-platform/4.5/networking/configuring_ingress_cluster_traffic/overview-traffic.html)
          in environments where an external load balancer cannot be provisioned.\n\n##
          How it works\n\nThe keepalived operator will create one or more [VIPs](https://en.wikipedia.org/wiki/Virtual_IP_address)
          (an HA IP that floats between multiple nodes), based on the [`LoadBalancer`](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          services and/or services requesting [`ExternalIPs`](https://kubernetes.io/docs/concepts/services-networking/service/#external-ips).\n\nFor
          `LoadBalancer` services the IPs found at `.Status.LoadBalancer.Ingress[].IP`
          will become VIPs.\n\nFor services requesting a `ExternalIPs`, the IPs found
          at `.Spec.ExternalIPs[]` will become VIPs.\n\nNote that a service can be
          of `LoadBalancer` type and also request `ExternalIPs`, it this case both
          sets of IPs will become VIPs.\n\nDue to a [keepalived](https://www.keepalived.org/manpage.html)
          limitation a single keepalived cluster can manage up to 256 VIP configurations.
          Multiple keepalived clusters can coexists in the same network as long as
          they use different multicast ports [TODO verify this statement].\n\nTo address
          this limitation the `KeepalivedGroup` [CRD](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/)
          has been introduced. This CRD is supposed to be configured by an administrator
          and allows you to specify a node selector to pick on which nodes the keepalived
          pods should be deployed. Here is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          KeepalivedGroup\nmetadata:\n  name: keepalivedgroup-router\nspec:\n  image:
          registry.redhat.io/openshift4/ose-keepalived-ipfailover\n  interface: ens3\n
          \ nodeSelector:\n    node-role.kubernetes.io/loadbalancer: \"\"\n  blacklistRouterIDs:\n
          \ - 1\n  - 2  \n```\n\nThis KeepalivedGroup will be deployed on all the
          nodes with role `loadbalancer`. One must also specify the network device
          on which the VIPs will be exposed, it is assumed that all the nodes have
          the same network device configuration.\n\nServices must be annotated to
          opt-in to being observed by the keepalived operator and to specify which
          KeepalivedGroup they refer to. The annotation looks like this:\n\n`keepalived-operator.redhat-cop.io/keepalivedgroup:
          <keepalivedgroup namespace>/<keepalivedgroup-name>`\n\nThe image used for
          the keepalived containers can be specified with `.Spec.Image` it will default
          to `registry.redhat.io/openshift4/ose-keepalived-ipfailover` if undefined.
          \n\n## Requirements\n\n### Security Context Constraints\n\nEach KeepalivedGroup
          deploys a [daemonset](https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/)
          that requires the [privileged scc](https://docs.openshift.com/container-platform/4.5/authentication/managing-security-context-constraints.html),
          this permission must be given to the `default` service account in the namespace
          where the keepalived group is created by and administrator.\n\n```shell\noc
          adm policy add-scc-to-user privileged -z default -n keepalived-operator\n```\n\n###
          Cluster Network Operator\n\nIn Openshift, use of an external IP address
          is governed by the following fields in the `Network.config.openshift.io`
          CR named `cluster`\n\n* `spec.externalIP.autoAssignCIDRs` defines an IP
          address block used by the load balancer when choosing an external IP address
          for the service. OpenShift supports only a single IP address block for automatic
          assignment.\n\n* `spec.externalIP.policy` defines the permissible IP address
          blocks when manually specifying an IP address. OpenShift does not apply
          policy rules to IP address blocks defined by `spec.externalIP.autoAssignCIDRs`\n\nThe
          following patch can be used to configure the Cluster Network Operator:\n\n```yaml\nspec:\n
          \ externalIP:\n    policy:\n      allowedCIDRs:\n      - ${ALLOWED_CIDR}\n
          \   autoAssignCIDRs:\n      - \"${AUTOASSIGNED_CIDR}\"\n```\n\nHere is an
          example of how to apply the patch:\n\n```shell\nexport ALLOWED_CIDR=\"192.168.131.128/26\"\nexport
          AUTOASSIGNED_CIDR=\"192.168.131.192/26\"\noc patch network cluster -p \"$(envsubst
          < ./network-patch.yaml | yq r -j -)\" --type=merge\n```\nAdditionally, the
          fields can be edited manually via `oc edit Network.config.openshift.io cluster`\n\n##
          Blacklisting router IDs\n\nIf the Keepalived pods are deployed on nodes
          which are in the same network (same broadcast domain to be precise) with
          other keepalived the process, it's necessary to ensure that there is no
          collision between the used routers it.\nFor this purpose it is possible
          to provide a `blacklistRouterIDs` field with a list of black-listed IDs
          that will not be used.\n\n## OpenShift RHV, vSphere, OSP and bare metal
          IPI instructions\n\nWhen IPI is used for RHV, vSphere, OSP or bare metal
          platforms, three keepalived VIPs are deployed. To make sure that keepalived-operator
          can work in these environment we need to discover and blacklist the corresponding
          VRRP router IDs.\n\nTo discover the VRRP router IDs being used, run the
          following command, you can run this command from you laptop:\n\n```shell\npodman
          run quay.io/openshift/origin-baremetal-runtimecfg:4.5 vr-ids <cluster_name>\n```\n\nIf
          you don't know your cluster name, run this command:\n\n```shell\npodman
          run quay.io/openshift/origin-baremetal-runtimecfg:4.5 vr-ids $(oc get cm
          cluster-config-v1 -n kube-system -o jsonpath='{.data.install-config}'| yq
          -r .metadata.name)\n```\n\nThen use these [instructions](#Blacklisting-router-IDs)
          to blacklist those VRRP router IDs.\n\n## Verbatim Configurations\n\nKeepalived
          has dozens of [configurations](https://www.keepalived.org/manpage.html).
          At the early stage of this project it's difficult to tell which one should
          be modeled in the API. Yet, users of this project may still need to use
          them. To account for that there is a way to pass verbatim options both at
          the keepalived group level (which maps to the keepalived config `global_defs`
          section) and at the service level (which maps to the keepalived config `vrrp_instance`
          section).\n\nKeepalivedGroup-level verbatim configurations can be passed
          as in the following example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          KeepalivedGroup\nmetadata:\n  name: keepalivedgroup-router\nspec:\n  interface:
          ens3\n  nodeSelector:\n    node-role.kubernetes.io/loadbalancer: \"\"\n
          \ verbatimConfig:  \n    vrrp_iptables: my-keepalived\n```\n\nthis will
          map to the following `global_defs`:\n\n```\n    global_defs {\n        router_id
          keepalivedgroup-router\n        vrrp_iptables my-keepalived\n    }\n```\n\nService-level
          verbatim configurations can be passed as in the following example:\n\n```yaml\napiVersion:
          v1\nkind: Service\nmetadata:\n  annotations:\n    keepalived-operator.redhat-cop.io/keepalivedgroup:
          keepalived-operator/keepalivedgroup-router\n    keepalived-operator.redhat-cop.io/verbatimconfig:
          '{ \"track_src_ip\": \"\" }'\n```\n\nthis will map to the following `vrrp_instance`
          section\n\n```\n    vrrp_instance openshift-ingress/router-default {\n        interface
          ens3\n        virtual_router_id 1  \n        virtual_ipaddress {\n          192.168.131.129\n
          \       }\n        track_src_ip\n    }\n```\n\n## Metrics collection\n\nEach
          keepalived pod exposes a [Prometheus](https://prometheus.io/) metrics port
          at `9650`. Metrics are collected with [keepalived_exporter](github.com/gen2brain/keepalived_exporter),
          the available metrics are described in the project documentation.\n\nWhen
          a keepalived group is created a [`PodMonitor`](https://github.com/coreos/prometheus-operator/blob/master/Documentation/api.md#podmonitor)
          rule to collect those metrics. All PodMonitor resources created that way
          have the label: `metrics: keepalived`. It is up to you to make sure your
          Prometheus instance watches for those `PodMonitor` rules. Here is an example
          of a fragment of a `Prometheus` CR configured to collect the keepalived
          pod metrics:\n\n```yaml\n  podMonitorSelector:\n    matchLabels:\n      metrics:
          keepalived\n```\n"
        displayName: Keepalived Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadbalancer
        - keepalived
        - ipfailover
        - externalIP
        - self-hosted
        links:
        - name: Keepalived Operator
          url: https://github.com/redhat-cop/keepalived-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/keepalived-operator
        maintainers:
        - email: rspazzol@redhat.com'
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/keepalived-operator@sha256:cd0c2f7df82df1178795e4a18a0942da33adb9670b192a7b7a2b1483a5b4ccf8
        version: 1.5.4
      entries:
      - name: keepalived-operator.v1.5.4
        version: 1.5.4
      - name: keepalived-operator.v1.5.3
        version: 1.5.3
      - name: keepalived-operator.v1.5.2
        version: 1.5.2
      - name: keepalived-operator.v1.5.1
        version: 1.5.1
      - name: keepalived-operator.v1.5.0
        version: 1.5.0
      - name: keepalived-operator.v1.4.2
        version: 1.4.2
      - name: keepalived-operator.v1.4.1
        version: 1.4.1
      - name: keepalived-operator.v1.4.0
        version: 1.4.0
      - name: keepalived-operator.v1.3.4
        version: 1.3.4
      - name: keepalived-operator.v1.3.3
        version: 1.3.3
      - name: keepalived-operator.v1.3.2
        version: 1.3.2
      - name: keepalived-operator.v1.3.1
        version: 1.3.1
      - name: keepalived-operator.v1.3.0
        version: 1.3.0
      - name: keepalived-operator.v1.2.2
        version: 1.2.2
      - name: keepalived-operator.v1.2.1
        version: 1.2.1
      - name: keepalived-operator.v1.2.0
        version: 1.2.0
      - name: keepalived-operator.v1.1.0
        version: 1.1.0
      - name: keepalived-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: keepalived-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel Corporation
      provider-url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
    name: sriov-fec
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sriov-fec.v2.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecClusterConfig",
                "metadata": {
                  "name": "acc100SampleConfig",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "acceleratorSelector": {
                    "deviceID": "someDevice",
                    "driver": "someDriver",
                    "maxVirtualFunctions": 2,
                    "pciAddress": "somePciAddress",
                    "vendorID": "someVendor"
                  },
                  "drainSkip": false,
                  "nodeSelector": {
                    "expectedLabel1": "valueOfExpectedLabel1",
                    "expectedLabelN": "valueOfExpectedLabelN"
                  },
                  "physicalFunction": {
                    "bbDevConfig": {
                      "acc100": {
                        "downlink4G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 0
                        },
                        "downlink5G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 4
                        },
                        "maxQueueSize": 1024,
                        "numVfBundles": 16,
                        "pfMode": true,
                        "uplink4G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 0
                        },
                        "uplink5G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 4
                        }
                      }
                    },
                    "pfDriver": "pci-pf-stub",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  },
                  "priority": 100
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecClusterConfig",
                "metadata": {
                  "name": "n3000SampleConfig",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "acceleratorSelector": {
                    "deviceID": "someDevice",
                    "driver": "someDriver",
                    "maxVirtualFunctions": 2,
                    "pciAddress": "somePciAddress",
                    "vendorID": "someVendor"
                  },
                  "drainSkip": false,
                  "nodeSelector": {
                    "expectedLabel1": "valueOfExpectedLabel1",
                    "expectedLabelN": "valueOfExpectedLabelN"
                  },
                  "physicalFunction": {
                    "bbDevConfig": {
                      "n3000": {
                        "downlink": {
                          "bandwidth": 3,
                          "loadBalance": 128,
                          "queues": {
                            "vf0": 16,
                            "vf1": 16,
                            "vf2": 0,
                            "vf3": 0,
                            "vf4": 0,
                            "vf5": 0,
                            "vf6": 0,
                            "vf7": 0
                          }
                        },
                        "flrTimeout": 610,
                        "networkType": "FPGA_5GNR",
                        "pfMode": true,
                        "uplink": {
                          "bandwidth": 3,
                          "loadBalance": 128,
                          "queues": {
                            "vf0": 16,
                            "vf1": 16,
                            "vf2": 0,
                            "vf3": 0,
                            "vf4": 0,
                            "vf5": 0,
                            "vf6": 0,
                            "vf7": 0
                          }
                        }
                      }
                    },
                    "pfDriver": "pci-pf-stub",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  },
                  "priority": 100
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecNodeConfig",
                "metadata": {
                  "name": "acc100-worker",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "drainSkip": false,
                  "physicalFunctions": [
                    {
                      "bbDevConfig": {
                        "acc100": {
                          "downlink4G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 0
                          },
                          "downlink5G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 4
                          },
                          "maxQueueSize": 1024,
                          "numVfBundles": 16,
                          "pfMode": true,
                          "uplink4G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 0
                          },
                          "uplink5G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 4
                          }
                        }
                      },
                      "pci_addr": "somePciAddress",
                      "pf_driver": "pci-pf-stub",
                      "vf_amount": 2,
                      "vf_driver": "vfio-pci"
                    }
                  ]
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecNodeConfig",
                "metadata": {
                  "name": "n3000-worker",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "drainSkip": false,
                  "physicalFunctions": [
                    {
                      "bbDevConfig": {
                        "n3000": {
                          "downlink": {
                            "bandwidth": 3,
                            "loadBalance": 128,
                            "queues": {
                              "vf0": 16,
                              "vf1": 16,
                              "vf2": 0,
                              "vf3": 0,
                              "vf4": 0,
                              "vf5": 0,
                              "vf6": 0,
                              "vf7": 0
                            }
                          },
                          "flrTimeout": 610,
                          "networkType": "FPGA_5GNR",
                          "pfMode": true,
                          "uplink": {
                            "bandwidth": 3,
                            "loadBalance": 128,
                            "queues": {
                              "vf0": 16,
                              "vf1": 16,
                              "vf2": 0,
                              "vf3": 0,
                              "vf4": 0,
                              "vf5": 0,
                              "vf6": 0,
                              "vf7": 0
                            }
                          }
                        }
                      },
                      "pci_addr": "somePciAddress",
                      "pf_driver": "pci-pf-stub",
                      "vf_amount": 2,
                      "vf_driver": "vfio-pci"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/intel/sriov-fec-operator@sha256:adb6a4c2da6399c8ed40c906f790bce13bfcefae3326a047f6a50592b3f5e485
          createdAt: "2025-01-15"
          description: An operator for Intel Wireless FEC Accelerator to orchestrate
            and manage the resources/devices exposed by a range of Intel's vRAN FEC
            acceleration devices/hardware within the OpenShift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/intel/sriov-fec-operator
          support: Intel Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovFecClusterConfig is the Schema for the sriovfecclusterconfigs
              API
            displayName: SriovFecClusterConfig
            kind: SriovFecClusterConfig
            name: sriovfecclusterconfigs.sriovfec.intel.com
            version: v2
          - description: SriovFecNodeConfig is the Schema for the sriovfecnodeconfigs
              API
            displayName: SriovFecNodeConfig
            kind: SriovFecNodeConfig
            name: sriovfecnodeconfigs.sriovfec.intel.com
            version: v2
          - description: SriovVrbClusterConfig is the Schema for the sriovvrbclusterconfigs
              API
            displayName: SriovVrbClusterConfig
            kind: SriovVrbClusterConfig
            name: sriovvrbclusterconfigs.sriovvrb.intel.com
            version: v1
          - description: SriovVrbNodeConfig is the Schema for the sriovvrbnodeconfigs
              API
            displayName: SriovVrbNodeConfig
            kind: SriovVrbNodeConfig
            name: sriovvrbnodeconfigs.sriovvrb.intel.com
            version: v1
        description: The Intel vRAN Boost family of accelerators ACC100/VRB1/VRB2,
          to offload and accelerate the computing-intensive process of forward error
          correction (FEC) and more for 4G/LTE and 5G technology, freeing up processing
          power. It allows the optimization of data plane performance to reduce total
          cost of ownership while maintaining a high degree of flexibility.  The Intel
          vRAN Boost accelerators ACC100/VRB1/VRB2 plays a key role in accelerating
          5G and network functions virtualization (NFV) adoption for ecosystem partners
          such as telecommunications equipment manufacturers (TEMs) virtual network
          functions (VNF) vendors, system integrators and telcos, to bring scalable
          and high-performance solutions to market. The Intel ACC100 includes a variant
          that is design to be Network Equipment Building System (NEBS)-friendly,
          and features a Root-of-Trust device that helps protect systems from FPGA
          host security exploits.
        displayName: SRIOV-FEC Operator for Intel vRAN Boost Accelerators
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ACC100
        - vRAN
        - ORAN
        links:
        - name: SRIOV-FEC source code
          url: https://github.com/intel/sriov-fec-operator
        maturity: alpha
        provider:
          name: Intel Corporation
          url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
        relatedImages:
        - registry.connect.redhat.com/intel/sriov-fec-operator@sha256:adb6a4c2da6399c8ed40c906f790bce13bfcefae3326a047f6a50592b3f5e485
        - registry.connect.redhat.com/intel/sriov-fec-daemon@sha256:5b0c60b67bcf5d7b2797deab0d5e3fa079577179f0acc6b0062533584ea0bc23
        - registry.connect.redhat.com/intel/n3000-labeler@sha256:b417cfcedc998a35491784005cc7eee3d9c647b84eac5ca41cb459076870b8f6
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin@sha256:d0fcb6e9135d9a19ea26851bcdb4ac72c04bd747706fc5123020460e9dbfc21b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1ed3b514bf9af6ac0d386e3739874d1d964e667d54e041c2d67edd769b0009fb
        version: 2.11.1
      entries:
      - name: sriov-fec.v2.11.1
        version: 2.11.1
      - name: sriov-fec.v2.11.0
        version: 2.11.0
      - name: sriov-fec.v2.10.0
        version: 2.10.0
      - name: sriov-fec.v2.9.0
        version: 2.9.0
      - name: sriov-fec.v2.8.0
        version: 2.8.0
      - name: sriov-fec.v2.7.2
        version: 2.7.2
      - name: sriov-fec.v2.7.1
        version: 2.7.1
      name: stable
    defaultChannel: stable
    packageName: sriov-fec
    provider:
      name: Intel Corporation
      url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: spyre-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spyre-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreClusterPolicy",
                "metadata": {
                  "name": "spyreclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2",
                      "spyreFilter": ".",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "latest"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "spyre-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "spyre-device-plugin-init",
                      "repository": "icr.io/ibmaiu_internal",
                      "version": "0.1.1"
                    },
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "spyre-exporter",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.2"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "spyre-webhook-validator",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  },
                  "scheduler": {
                    "image": "spyre-scheduler",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  }
                }
              },
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu_internal/spyre-operator@sha256:0f657c98668a8358c37b3fc2ed3f31ec78b7ef7f9a9678e7cd35d383d22ce2da
          createdAt: "2025-07-31T10:32:48Z"
          description: Operator for IBM Spyre devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM Spyre
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["spyrenodestates.spyre.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpyreClusterPolicy is the Schema for the Spyre API
            displayName: Spyre Cluster Policy
            kind: SpyreClusterPolicy
            name: spyreclusterpolicies.spyre.ibm.com
            version: v1alpha1
          - description: SpyreNodeState is the Schema for the SpyreNodeState API
            displayName: Spyre Node State
            kind: SpyreNodeState
            name: spyrenodestates.spyre.ibm.com
            version: v1alpha1
        description: Operator for IBM Spyre devices
        displayName: IBM Spyre Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: IBM Spyre Operator
          url: https://spyre-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: Satyam.Gupta.3@ibm.com
          name: Satyam Gupta
        - email: Rehan.Khan7@ibm.com
          name: Rehan Khan
        - email: Nishan.Acharya@ibm.com
          name: Nishan Acharya
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu_internal/spyre-operator@sha256:0f657c98668a8358c37b3fc2ed3f31ec78b7ef7f9a9678e7cd35d383d22ce2da
        - icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 0.1.2
      entries:
      - name: spyre-operator.v0.1.2
        version: 0.1.2
      name: stable-v0.1
    - currentCSV: spyre-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreClusterPolicy",
                "metadata": {
                  "name": "spyreclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb",
                      "spyreFilter": ".",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "latest"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "spyre-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "spyre-device-plugin-init",
                      "repository": "quay.io/ibm-aiu",
                      "version": "0.2.0"
                    },
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "spyre-exporter",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "spyre-webhook-validator",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "scheduler": {
                    "image": "spyre-scheduler",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.1"
                  }
                }
              },
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/ibm-aiu/spyre-operator@sha256:d0cdd449f1d8f0dd9281c4d9a809333db6c4ed201bb14187b6c70ae686a30e99
          createdAt: "2025-09-02T13:25:34Z"
          description: Operator for IBM Spyre devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM Spyre
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["spyrenodestates.spyre.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpyreClusterPolicy is the Schema for the Spyre API
            displayName: Spyre Cluster Policy
            kind: SpyreClusterPolicy
            name: spyreclusterpolicies.spyre.ibm.com
            version: v1alpha1
          - description: SpyreNodeState is the Schema for the SpyreNodeState API
            displayName: Spyre Node State
            kind: SpyreNodeState
            name: spyrenodestates.spyre.ibm.com
            version: v1alpha1
        description: Operator for IBM Spyre devices
        displayName: IBM Spyre Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: IBM Spyre Operator
          url: https://spyre-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: Satyam.Gupta.3@ibm.com
          name: Satyam Gupta
        - email: Rehan.Khan7@ibm.com
          name: Rehan Khan
        - email: Nishan.Acharya@ibm.com
          name: Nishan Acharya
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - quay.io/ibm-aiu/spyre-operator@sha256:d0cdd449f1d8f0dd9281c4d9a809333db6c4ed201bb14187b6c70ae686a30e99
        - icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb
        version: 0.2.1
      entries:
      - name: spyre-operator.v0.2.1
        version: 0.2.1
      name: stable-v0.2
    - currentCSV: spyre-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreClusterPolicy",
                "metadata": {
                  "name": "spyreclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "config": {
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb",
                      "spyreFilter": ".",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "latest"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "spyre-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "spyre-device-plugin-init",
                      "repository": "quay.io/ibm-aiu",
                      "version": "1.0.0"
                    },
                    "repository": "quay.io/ibm-aiu",
                    "version": "1.0.0"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "spyre-exporter",
                    "repository": "quay.io/ibm-aiu",
                    "version": "1.0.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "spyre-webhook-validator",
                    "repository": "quay.io/ibm-aiu",
                    "version": "1.0.0"
                  },
                  "scheduler": {
                    "image": "spyre-scheduler",
                    "repository": "quay.io/ibm-aiu",
                    "version": "1.0.0"
                  }
                }
              },
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/ibm-aiu/spyre-operator@sha256:dd262e8c65f35b12c5d79571e8b5ddd54dd0ee2693f28bea05a6fc02ac1fe6f0
          createdAt: "2025-10-01T00:40:23Z"
          description: Operator for IBM Spyre devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: spyre-operator
          operators.openshift.io/valid-subscription: use with IBM Spyre
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["spyrenodestates.spyre.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpyreClusterPolicy is the Schema for the Spyre API
            displayName: Spyre Cluster Policy
            kind: SpyreClusterPolicy
            name: spyreclusterpolicies.spyre.ibm.com
            version: v1alpha1
          - description: SpyreNodeState is the Schema for the SpyreNodeState API
            displayName: Spyre Node State
            kind: SpyreNodeState
            name: spyrenodestates.spyre.ibm.com
            version: v1alpha1
        description: Operator for IBM Spyre devices
        displayName: IBM Spyre Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: IBM Spyre Operator
          url: https://spyre-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: Satyam.Gupta.3@ibm.com
          name: Satyam Gupta
        - email: Rehan.Khan7@ibm.com
          name: Rehan Khan
        - email: Nishan.Acharya@ibm.com
          name: Nishan Acharya
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - quay.io/ibm-aiu/spyre-operator@sha256:dd262e8c65f35b12c5d79571e8b5ddd54dd0ee2693f28bea05a6fc02ac1fe6f0
        - icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb
        version: 1.0.0
      entries:
      - name: spyre-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: spyre-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hive-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hive-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:e71fc728632622f7ea97357ec361dd2bb73e63c0707e8014ca8ecf8ec4ab93f1
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:e71fc728632622f7ea97357ec361dd2bb73e63c0707e8014ca8ecf8ec4ab93f1
        version: 24.11.1
      entries:
      - name: hive-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: hive-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:5530f5034cd83c2e8b860fce737ac491a2adc6d22246356bee0c79768ed64706
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:5530f5034cd83c2e8b860fce737ac491a2adc6d22246356bee0c79768ed64706
        version: 25.3.0
      entries:
      - name: hive-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: hive-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
        version: 25.7.0
      entries:
      - name: hive-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hive-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
        version: 25.7.0
      entries:
      - name: hive-operator.v25.7.0
        version: 25.7.0
      - name: hive-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hive-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CSI Addons Community
      provider-url: ""
    name: odf-csi-addons-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-csi-addons-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          createdAt: "2025-09-12T07:33:01Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIAddonsNode is the Schema for the csiaddonsnode API
            displayName: CSIAddons Node
            kind: CSIAddonsNode
            name: csiaddonsnodes.csiaddons.openshift.io
            version: v1alpha1
          - description: NetworkFence is the Schema for the networkfences API
            displayName: Network Fence
            kind: NetworkFence
            name: networkfences.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceCronJob is the Schema for the reclaimspacecronjobs
              API
            displayName: Reclaim Space CronJob
            kind: ReclaimSpaceCronJob
            name: reclaimspacecronjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceJob is the Schema for the reclaimspacejobs API
            displayName: Reclaim Space Job
            kind: ReclaimSpaceJob
            name: reclaimspacejobs.csiaddons.openshift.io
            version: v1alpha1
          - description: VolumeReplicationClass is the Schema for the volumereplicationclasses
              API
            displayName: Volume Replication Class
            kind: VolumeReplicationClass
            name: volumereplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplication is the Schema for the volumereplications
              API
            displayName: Volume Replication
            kind: VolumeReplication
            name: volumereplications.replication.storage.openshift.io
            version: v1alpha1
        description: CSI Addons provides the CSI Addons Controller that enables advanced
          storage operations for CSI-drivers.
        displayName: CSI Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: CSI Addons
          url: https://csi-addons.github.io
        maintainers:
        - email: csi-addons@redhat.com
          name: CSI Addons Community
        maturity: alpha
        provider:
          name: CSI Addons Community
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/odf4/odf-csi-addons-rhel9-operator@sha256:8b0c16c21fccca9b674c4d980a793b6cde482b86363b793e173126b030bdcc55
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:9a0008f662b6afa9e2677a8a83bd662d806ad6f4d03fb49bccb78ed487ce4c7a
        version: 4.15.20-rhodf
      entries:
      - name: odf-csi-addons-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odf-csi-addons-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odf-csi-addons-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odf-csi-addons-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odf-csi-addons-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odf-csi-addons-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odf-csi-addons-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odf-csi-addons-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odf-csi-addons-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odf-csi-addons-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odf-csi-addons-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odf-csi-addons-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odf-csi-addons-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odf-csi-addons-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odf-csi-addons-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odf-csi-addons-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odf-csi-addons-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odf-csi-addons-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odf-csi-addons-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odf-csi-addons-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odf-csi-addons-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          createdAt: "2025-09-12T06:32:07Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIAddonsNode is the Schema for the csiaddonsnode API
            displayName: CSIAddons Node
            kind: CSIAddonsNode
            name: csiaddonsnodes.csiaddons.openshift.io
            version: v1alpha1
          - description: NetworkFence is the Schema for the networkfences API
            displayName: Network Fence
            kind: NetworkFence
            name: networkfences.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceCronJob is the Schema for the reclaimspacecronjobs
              API
            displayName: Reclaim Space CronJob
            kind: ReclaimSpaceCronJob
            name: reclaimspacecronjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceJob is the Schema for the reclaimspacejobs API
            displayName: Reclaim Space Job
            kind: ReclaimSpaceJob
            name: reclaimspacejobs.csiaddons.openshift.io
            version: v1alpha1
          - description: VolumeReplicationClass is the Schema for the volumereplicationclasses
              API
            displayName: Volume Replication Class
            kind: VolumeReplicationClass
            name: volumereplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplication is the Schema for the volumereplications
              API
            displayName: Volume Replication
            kind: VolumeReplication
            name: volumereplications.replication.storage.openshift.io
            version: v1alpha1
        description: CSI Addons provides the CSI Addons Controller that enables advanced
          storage operations for CSI-drivers.
        displayName: CSI Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: CSI Addons
          url: https://csi-addons.github.io
        maintainers:
        - email: csi-addons@redhat.com
          name: CSI Addons Community
        maturity: alpha
        provider:
          name: CSI Addons Community
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/odf4/odf-csi-addons-rhel9-operator@sha256:d370f19578d1c466d39d00da82f1f3fe3d9f13d7021ed0c39f759b24b0ed7573
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:0afcb516eb528857887c6ee1a9cba081dc8f6b7d47d9f26a0dbb2dca0cefe07c
        version: 4.16.16-rhodf
      entries:
      - name: odf-csi-addons-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odf-csi-addons-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odf-csi-addons-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odf-csi-addons-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odf-csi-addons-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odf-csi-addons-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odf-csi-addons-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odf-csi-addons-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odf-csi-addons-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odf-csi-addons-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odf-csi-addons-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odf-csi-addons-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odf-csi-addons-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odf-csi-addons-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odf-csi-addons-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odf-csi-addons-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odf-csi-addons-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odf-csi-addons-operator
    provider:
      name: CSI Addons Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Weights & Biases
      provider-url: https://wandb.ai
    name: wandb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: wandb-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.wandb.com/v1",
                "kind": "WeightsAndBiases",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "wandb-operator",
                    "app.kubernetes.io/instance": "weightsandbiases-sample",
                    "app.kubernetes.io/managed-by": "olm",
                    "app.kubernetes.io/name": "weightsandbiases",
                    "app.kubernetes.io/part-of": "wandb-operator"
                  },
                  "name": "weightsandbiases-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning,Developer Tools
          certified: "false"
          containerImage: quay.io/wandb_tools/wandb-k8s-operator:1.0.0
          createdAt: "2024-10-10T14:46:54Z"
          description: Operator for Weights and Biases, enabling seamless integration
            and management of machine learning experiments in Kubernetes environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: wandb-system
          repository: https://github.com/wandb/operator
          support: WeightsAndBiases
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: WeightsAndBiases is a custom resource for managing W&B configurations.
            displayName: Weights And Biases
            kind: WeightsAndBiases
            name: weightsandbiases.apps.wandb.com
            version: v1
        description: |-
          The Weights and Biases Operator simplifies the management of machine learning experiments, tracking, and infrastructure. It integrates with Kubernetes to deploy and manage Weights and Biases instances.
          Features:
          - Experiment Tracking
          - Hyperparameter Tuning
          - Model Management
          - Integration with popular ML frameworks
        displayName: Weights & Biases Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - machine learning
        - experiment tracking
        - ML infrastructure
        - weights and biases
        links:
        - name: Weights & Biases Documentation
          url: https://docs.wandb.ai/
        maintainers:
        - email: abhinav.garg@domain.com
          name: Abhinav Garg
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Weights & Biases
          url: https://wandb.ai
        relatedImages:
        - quay.io/wandb_tools/wandb-k8s-operator:1.0.0
        version: 1.0.0
      entries:
      - name: wandb-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: wandb-operator
    provider:
      name: Weights & Biases
      url: https://wandb.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: multicluster-operators-subscription
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-operators-subscription.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.3.0
          createdAt: "2021-07-19T18:00:00Z"
          description: An operator to subscribe resources from object Bucket, helm
            and github types of channels
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-application:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-channel:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-placementrule:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-deployable:2.3.0
        version: 0.3.0
      entries:
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.3
    - currentCSV: multicluster-operators-subscription.v0.3.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.4.0
          createdAt: "2021-09-29T10:23:00Z"
          description: An operator to subscribe resources from object Bucket, helm
            and github types of channels
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-application:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-channel:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-placementrule:2.4.0
        - quay.io/open-cluster-management/multicloud-integrations:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-deployable:2.4.0
        version: 0.3.6
      entries:
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.4
    - currentCSV: multicluster-operators-subscription.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.5.0
          createdAt: "2022-03-23T18:00:00Z"
          description: An operator for deploying resources from object Bucket, helm
            repo or github repo to multiple managed clusters
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management-io/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: subscription status per cluster per app
            displayName: App Subscription Report
            kind: SubscriptionReport
            name: subscriptionreports.apps.open-cluster-management.io
            version: v1alpha1
          - description: subscription status per package
            displayName: App Subscription Status
            kind: SubscriptionStatus
            name: subscriptionstatuses.apps.open-cluster-management.io
            version: v1alpha1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicloud-integrations:2.5.0
        - quay.io/open-cluster-management/multicluster-operators-application:2.5.0
        - quay.io/open-cluster-management/multicluster-operators-channel:2.5.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.5.0
        version: 0.4.0
      entries:
      - name: multicluster-operators-subscription.v0.4.0
        version: 0.4.0
      - name: multicluster-operators-subscription.v0.3.9
        version: 0.3.9
      - name: multicluster-operators-subscription.v0.3.8
        version: 0.3.8
      - name: multicluster-operators-subscription.v0.3.7
        version: 0.3.7
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.5
    - currentCSV: multicluster-operators-subscription.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-operators-subscription:2.6.0
          createdAt: "2022-06-06T18:00:00Z"
          description: An operator for deploying resources from object Bucket, helm
            repo or github repo to multiple managed clusters
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management-io/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: subscription status per cluster per app
            displayName: App Subscription Report
            kind: SubscriptionReport
            name: subscriptionreports.apps.open-cluster-management.io
            version: v1alpha1
          - description: subscription status per package
            displayName: App Subscription Status
            kind: SubscriptionStatus
            name: subscriptionstatuses.apps.open-cluster-management.io
            version: v1alpha1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/multicluster-operators-subscription:2.6.0
        - quay.io/stolostron/multicloud-integrations:2.6.0
        - quay.io/stolostron/multicluster-operators-application:2.6.0
        - quay.io/stolostron/multicluster-operators-channel:2.6.0
        version: 0.4.1
      entries:
      - name: multicluster-operators-subscription.v0.4.1
        version: 0.4.1
      - name: multicluster-operators-subscription.v0.4.0
        version: 0.4.0
      - name: multicluster-operators-subscription.v0.3.9
        version: 0.3.9
      - name: multicluster-operators-subscription.v0.3.8
        version: 0.3.8
      - name: multicluster-operators-subscription.v0.3.7
        version: 0.3.7
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.6
    defaultChannel: release-2.5
    packageName: multicluster-operators-subscription
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.90.0-beta.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"externalDbIP":
            "10.0.2.15","customImageNames": false,"repository": "icr.io/cpopen/turbonomic","tag":"8.17.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
          createdAt: "2023-03-24T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=42.53.0-beta.1 <42.90.0-beta.2'
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints. Turbonomic platform install and configuration
              documentation can be found online here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How do I start?

          Turbonomic platform installation and configuration documentation can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
          Carefully read through the prerequisites before beginning the installation.  Once ready, on the `Details` tab for the Turbonomic Platform Operator, under `Provided APIs`, click `Create instance`.
          For samples of common configurations, refer to sample custom resource YAMLs provided here: https://github.com/turbonomic/t8c-install/tree/master/samples
          Looking to upgrade your current deployment? Refer to documentation here: https://www.ibm.com/docs/en/tarm/latest?topic=version-updating-using-red-hat-openshift-operatorhub

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
        version: 42.90.0-beta.2
      entries:
      - name: t8c-operator.v42.90.0-beta.2
        version: 42.90.0-beta.2
      name: beta
    - currentCSV: t8c-operator.v42.90.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"externalDbIP":
            "10.0.2.15","customImageNames": false,"repository": "icr.io/cpopen/turbonomic","tag":"8.17.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
          createdAt: "2023-03-24T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=42.53.0 <42.90.0'
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints. Turbonomic platform install and configuration
              documentation can be found online here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How do I start?

          Turbonomic platform installation and configuration documentation can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
          Carefully read through the prerequisites before beginning the installation.  Once ready, on the `Details` tab for the Turbonomic Platform Operator, under `Provided APIs`, click `Create instance`.
          For samples of common configurations, refer to sample custom resource YAMLs provided here: https://github.com/turbonomic/t8c-install/tree/master/samples
          Looking to upgrade your current deployment? Refer to documentation here: https://www.ibm.com/docs/en/tarm/latest?topic=version-updating-using-red-hat-openshift-operatorhub

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
        version: 42.90.0
      entries:
      - name: t8c-operator.v42.90.0
        version: 42.90.0
      name: stable
    defaultChannel: stable
    packageName: t8c-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: https://stakater.com
    name: tenant-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tenant-operator.v0.12.66
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: =0.12.65
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:9f806a065b25229741dac8437a32ec45a99acecf4e6bcbe6e7d1fa4192bc560d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a35b96eacb4e46114b39354f66429f8f05df09c55c98c9416b4ffa2768310ef7
        version: 0.12.66
      entries:
      - name: tenant-operator.v0.12.66
        version: 0.12.66
      name: release-0.12
    - currentCSV: tenant-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2024-10-16T08:15:54Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.0
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:feb71a5045e621f8a8cbfe7e41831100b6b85af9812138071b6217cc4396860a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a35b96eacb4e46114b39354f66429f8f05df09c55c98c9416b4ffa2768310ef7
        version: 1.0.0
      entries:
      - name: tenant-operator.v1.0.0
        version: 1.0.0
      name: release-1.0
    - currentCSV: tenant-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-01-27T13:23:27Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.0
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a35b96eacb4e46114b39354f66429f8f05df09c55c98c9416b4ffa2768310ef7
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:f412b6d05038a185f5fe35f9c917388194a5ea66afb538a513423c24390983ce
        version: 1.1.0
      entries:
      - name: tenant-operator.v1.1.0
        version: 1.1.0
      name: release-1.1
    - currentCSV: tenant-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-03-19T13:30:50Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.2.1
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:a899d60b31efd5cf2358efcd48737b1c4f4a0356ab6c0abe7a5f62ea825a52cd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a35b96eacb4e46114b39354f66429f8f05df09c55c98c9416b4ffa2768310ef7
        version: 1.2.1
      entries:
      - name: tenant-operator.v1.2.1
        version: 1.2.1
      name: release-1.2
    - currentCSV: tenant-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-06-11T15:25:04Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.3.0
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:e88aa5b008ece406bca35ec22fdc87c842544049d613d4d81c8bc5dd2fb50812
        version: 1.3.0
      entries:
      - name: tenant-operator.v1.3.0
        version: 1.3.0
      name: release-1.3
    - currentCSV: tenant-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-09-08T08:14:28Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.4.2
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:7ee14c0974d1eda849d2c199a56786b0ae777ef6569085be917449f1998b1158
        version: 1.4.2
      entries:
      - name: tenant-operator.v1.4.2
        version: 1.4.2
      name: release-1.4
    defaultChannel: release-1.4
    packageName: tenant-operator
    provider:
      name: Stakater
      url: https://stakater.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.0.4
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
        version: 1.0.4
      entries:
      - name: datadog-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      name: stable
    defaultChannel: stable
    packageName: datadog-operator-certified-rhmp
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp, Inc.
      provider-url: https://www.netapp.com/
    name: trident-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trident-operator.v25.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "trident.netapp.io/v1",
                "kind": "TridentOrchestrator",
                "metadata": {
                  "name": "trident"
                },
                "spec": {
                  "IPv6": false,
                  "debug": true,
                  "nodePrep": [],
                  "imagePullSecrets": [],
                  "imageRegistry": "",
                  "namespace": "trident",
                  "silenceAutosupport": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: docker.io/netapp/trident-operator:25.06.2
          createdAt: "2025-08-03"
          description: Trident Operator, to manage NetApp Trident installations
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          repository: https://github.com/netapp/trident
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Used to deploy Astra Trident.
            displayName: Trident Orchestrator
            kind: TridentOrchestrator
            name: tridentorchestrators.trident.netapp.io
            version: v1
          - description: Automates AWS FSxN backend configuration
            displayName: Trident Configurator
            kind: TridentConfigurator
            name: tridentconfigurators.trident.netapp.io
            version: v1
        description: |
          NetApp Trident is an open source storage provisioner and orchestrator maintained by NetApp. It enables you to create storage volumes for containerized applications managed by Docker and Kubernetes. For full release information, including patch release changes, see https://docs.netapp.com/us-en/trident/trident-rn.html.
        displayName: NetApp Trident
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - NetApp
        - ONTAP
        - Trident
        - storage
        - performance
        - open source
        - Astra
        - Cloud Native
        - AWS
        - Azure
        - GCP
        - Cloud Volumes
        links:
        - name: GitHub Repository
          url: https://github.com/NetApp/trident
        - name: Trident documentation
          url: https://docs.netapp.com/us-en/trident/index.html
        - name: Support policy
          url: https://mysupport.netapp.com/site/info/version-support
        - name: Release Notes
          url: https://docs.netapp.com/us-en/trident/trident-rn.html
        maintainers:
        - email: ng-astra-product@netapp.com
          name: NetApp
        provider:
          name: NetApp, Inc.
          url: https://www.netapp.com/
        relatedImages:
        - docker.io/netapp/trident-operator:25.06.2
        version: 25.6.2
      entries:
      - name: trident-operator.v25.6.2
        version: 25.6.2
      - name: trident-operator.v25.6.1
        version: 25.6.1
      - name: trident-operator.v25.6.0
        version: 25.6.0
      - name: trident-operator.v25.2.1
        version: 25.2.1
      - name: trident-operator.v25.2.0
        version: 25.2.0
      - name: trident-operator.v24.10.1
        version: 24.10.1
      - name: trident-operator.v24.10.0
        version: 24.10.0
      - name: trident-operator.v24.6.0
        version: 24.6.0
      - name: trident-operator.v24.2.0
        version: 24.2.0
      name: stable
    defaultChannel: stable
    packageName: trident-operator
    provider:
      name: NetApp, Inc.
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeVirt project
      provider-url: ""
    name: community-kubevirt-hyperconverged
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubevirt-hyperconverged-operator.v1.10.7
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployTektonTaskResources":true,"deployVmConsoleProxy":true},"tektonPipelines":{"namespace":"kubevirt"},"tektonTasks":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:0ee30557d5d5a42dec19121e70ed77fbdacd8932f55b11e2c7c1a3d5ad5df4aa
          createdAt: "2024-02-09 14:34:30"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:f7567bea7f24184da06c3ed87b396c804342fcfd89f54989622780ce0b7b0724
        - quay.io/kubevirt/hostpath-csi-driver@sha256:0ca8b50b0adb20afd4e20da9974200f3290b6d51aa7c1129de6e1b953bf26594
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:e79027973b09aac75860f267c7d7f830599978c38a081001f404b9a1c3d2990f
        - quay.io/kubevirt/libguestfs-tools@sha256:b7e248be783691daad1326cf41641d5042be80c5c057ebc5315ca7a1619b0581
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:57573a24923e5588bca6bc337a8b2b08406c5b77583974365d2cf063c0dd5d06
        - quay.io/kubevirt/virt-controller@sha256:0789fafed2913b35a771e3db882748502b3250be04ece86d97f30201779b4e54
        - quay.io/kubevirt/mtq-controller@sha256:96dd73ea0d7847d9f6322192559747bd85655e25ee25482d112e6e3c6fef21b2
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:16ecabb1d733aa17084cd16d88963c5255d04f389e0f4d41941d67e8be973594
        - quay.io/kubevirt/mtq-operator@sha256:ef5c4bb1fa72f9e6e75d5585ca352f7563720389e6c9876aa8b07d12d1be6194
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5ff5392cb762a41ee9bc4767e3a693249ae08b46a4322f623f24ad6b63dd06b5
        - quay.io/kubevirt/virt-exportproxy@sha256:f14444b0200a85efb4b3c176caefe70aabba7d295e33d2af14027a59ce297e24
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4dc0b87ccd69f9865b89234d8555d3a614ab0a16ed94a3016ffd27f8106132ce
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f6717ce72a2615c7fbc746b4068f788e78579c54c43b8716e5ce650d97af2df1
        - quay.io/kubevirt/virt-launcher@sha256:4c5fce3de2e2589197de72fb0c9436490ea318aca952c05a622c43e067023f35
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:f64ab5981d9ddef1662dd0a05888783f5dbadda8ede596585ba80bfc97a0f51e
        - quay.io/kubevirt/virtio-container-disk@sha256:f8e71a3a930ce95d646d80862d5ff6f9d3ed7ef2541dd2e2acec0bce26274734
        - quay.io/kubevirt/cdi-uploadproxy@sha256:551221d79902a5053d1c734b81163d69f087217e2ac13c49bdf6900336ef0786
        - quay.io/kubevirt/cni-default-plugins@sha256:825e3f9fec1996c54a52cec806154945b38f76476b160d554c36e38dfffe5e61
        - quay.io/kubevirt/kubemacpool@sha256:3214020c21204c189cdf2d767beb8efb61ee3ef65dd8de7f0f1edc61e13e8adf
        - quay.io/kubevirt/mtq-lock-server@sha256:9614df0d5af7c998bcceaf73b548f2e3b29ff8ee4ca64c32613bfbdd6e452f6c
        - quay.io/kubevirt/cdi-uploadserver@sha256:30f1827d3696cf996b081c22c3267ca78e7219c872fdb54950198fa54359f6ee
        - quay.io/kubevirt/cdi-controller@sha256:27c47883a08226f83757971d3adafb0cd9bcb26e58fbcf7208236070e0adf37e
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:3fbcc32bd4e4d15bd93c96def784a229cd84cca27942bf4858b581f31c97ee02
        - quay.io/kubevirt/virt-artifacts-server:1.10.7-unstable
        - quay.io/kubevirt/virt-operator@sha256:cbca7f93afbb18a840f7e763e4c03aeeb1fd97967f381c12c443cc179b51da29
        - quay.io/kubevirt/cdi-apiserver@sha256:e9e39408413b1478d2e98eba68913f9e20c93000558b190b47de73bdfd1d9ac4
        - ghcr.io/kubevirt/kubesecondarydns@sha256:e87e829380a1e576384145f78ccaa885ba1d5690d5de7d0b73d40cfb804ea24d
        - quay.io/kubevirt/cdi-importer@sha256:3143bbc67cdc6267eb48b7eaac664b8551ac4c11401dfbf4921efd3f233e6ce9
        - quay.io/kubevirt/kubevirt-template-validator@sha256:a148b6e812b70f326b48d0439b1e44a967a6e1f8cf2f98b1a08e9294a93674eb
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:0ee30557d5d5a42dec19121e70ed77fbdacd8932f55b11e2c7c1a3d5ad5df4aa
        - quay.io/kubevirt/cdi-operator@sha256:6c63521d835578b0fbb77d3145b648f95ab0932d5d1b36ff2068ed8fcd91bc5a
        - quay.io/kubevirt/virt-exportserver@sha256:73311f79a9c71007f8572b3cc40cd6f6da404c7ef0a9c6509fb717d979546582
        - quay.io/kubevirt/bridge-marker@sha256:bba066e3b5ff3fb8c5e20861fe8abe51e3c9b50ad6ce3b2616af9cb5479a06d0
        - quay.io/kubevirt/cdi-cloner@sha256:9d31b14f23259398c5bac636f5ead13ad0afd6fe8eeab4499e8e047b4d85074f
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - quay.io/kubevirt/hostpath-provisioner@sha256:eae30f37b3e837478a4500144b7bc8eb4f5ddf15f49fa7354bc9d5e240ea8aea
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:86c10c122767661e31ee26a9e9de7ea606c2c1168b32793b4ea567ede2ae3b21
        - quay.io/kubevirt/ssp-operator@sha256:49cac00844f091d6e80301573e00235d685415185ec29c89b4bd7361f938711e
        - quay.io/kubevirt/virt-handler@sha256:138dfda5fea8622f3da0d6413fe214fef80c2fd6a6f9533592a0dbfa7e1865b5
        - quay.io/kubevirt/virt-api@sha256:707003b221496b4432da2f507d1e36e528b45888b5d321e06d460f0678da44ae
        - quay.io/kubevirt/macvtap-cni@sha256:434420511e09b2b5ede785a2c9062b6658ffbc26fbdd4629ce06110f9039c600
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:38293a1050bce9c7cb786cbe8719be1edb91a922a6c583eedb61e56729ed4c72
        - quay.io/kubevirt/ovs-cni-plugin@sha256:08f72edf2bef876bba0b0f5513d30225304ad5e7ad6912a61c083664acdb99ff
        version: 1.10.7
      entries:
      - name: kubevirt-hyperconverged-operator.v1.10.7
        version: 1.10.7
      name: 1.10.7
    - currentCSV: kubevirt-hyperconverged-operator.v1.11.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployTektonTaskResources":true,"deployVmConsoleProxy":true},"tektonPipelines":{"namespace":"kubevirt"},"tektonTasks":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:dd6538e720f18b723b0807037ca7cc1014cba09e57b20706e10e6eb641ac6acc
          createdAt: "2024-03-06 15:11:00"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: quay.io/kubevirt/must-gather
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - quay.io/kubevirt/virt-handler@sha256:238550bb263608ba334b44a15647cbf875482ccfb46049713c14af5d5f0cc788
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:dd6538e720f18b723b0807037ca7cc1014cba09e57b20706e10e6eb641ac6acc
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:70efc16ab2bfec04f73ccc292064d080870659cfd47c2e0c9017ff7b6fd78863
        - quay.io/kubevirt/kubevirt-template-validator@sha256:f6db539fe4d4b752644c6bc8a308f5a80ae09a71da5c5d535114dc4c3cf18e0a
        - quay.io/kubevirt/cdi-controller@sha256:606de89e68be2ac911466e6713345932bda496c1e4f20e34ef5efa270456c9fd
        - quay.io/kubevirt/virt-operator@sha256:0e6d56ea0ebdf9bb832f29ab0dc4603f0f619474bb8bc02a01e0a06e9d8d6f4a
        - quay.io/kubevirt/libguestfs-tools@sha256:4971af1ee42959d366a3db6129ab684d11064e00bbd3c2b36324ad14e90cb903
        - quay.io/kubevirt/virt-api@sha256:735662c824da3d6069624d40674163d758c2256f8043cf6f02f505c3a724afa3
        - quay.io/kubevirt/cdi-uploadproxy@sha256:b89104ef5513d63cab43b34ac52f752c3d935cb986d5527d920c2c31d3b6f239
        - quay.io/kubevirt/macvtap-cni@sha256:850b89343ace7c7ea6b18dd8e11964613974e9d1f7377af03854d407fb15230a
        - quay.io/kubevirt/mtq-controller@sha256:ae7dd07e4018c63414b021eba34aa4776a6c7d6ff4fcdf9fd2d7ca30aae98a29
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - quay.io/kubevirt/cdi-uploadserver@sha256:16d257e575b2bd504f148f63b57dc181fa6058d33a11ada64888183c3e4d4863
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:4ed1f37116cbaf883edd8783201d091fbbe121fcb84e7184ee531d3b1abd77ab
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:e2def4213ec0657e72eb790ae8a115511d5b8f164a62d3568d2f1bff189917e8
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:dd2948ca8e62d17eafba40bdcaad8367a97f092fd3b8d9849b2d62aec3152a72
        - quay.io/kubevirt/aaq-controller@sha256:cb2fa52ae3f6a1a99b2de77b58c5db54c43ba8ee86c6540b4844c901180ce611
        - quay.io/kubevirt/hostpath-csi-driver@sha256:8d84b9a33e0ec3749aece7c5f9879d1e0d248e8028ab3f776528a044d387033c
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:f1bd6ee18c4021c1c94f29edfab89b49b6a4d1b800936c19dbef2d75f8202f2d
        - quay.io/kubevirt/virtio-container-disk@sha256:f8e71a3a930ce95d646d80862d5ff6f9d3ed7ef2541dd2e2acec0bce26274734
        - quay.io/kubevirt/virt-artifacts-server@sha256:22fa3218deca38dc98fa0e905df16173a47ccec85707eb91404a367da6480817
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:bf17da674ed6dac95aa4fd7637173fd4e026ebb4523b66f529c0eff394a25bcc
        - quay.io/kubevirt/ssp-operator@sha256:d57ef05fea09c4690dbe728ebe6a39d9df1a3574dfa7d18346bf3fe66afed702
        - quay.io/kubevirt/virt-launcher@sha256:4fbfd76561667e98fafc75c0b804f77df251f96471d1121d69f149cf8fb9a4d8
        - quay.io/kubevirt/bridge-marker@sha256:bba066e3b5ff3fb8c5e20861fe8abe51e3c9b50ad6ce3b2616af9cb5479a06d0
        - quay.io/kubevirt/virt-controller@sha256:4f576e30b83db7b9f96dde77896898bde87db3517963c4227634d0c3e96bb86d
        - quay.io/kubevirt/virt-exportserver@sha256:f8473fc824f8bac5f21c64b71cdb4227de28f9bb25c049932f1259216b874a86
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:38293a1050bce9c7cb786cbe8719be1edb91a922a6c583eedb61e56729ed4c72
        - quay.io/kubevirt/aaq-operator@sha256:590d24742401eeff3fb6a9c8ed977c63442f8f0a0505509fae69aa1bf8aa84ae
        - quay.io/kubevirt/cdi-apiserver@sha256:95f6e71b36883abcb9ec463cfea2332f5490ec9eb3e4868bdf0a4da769279094
        - quay.io/kubevirt/cni-default-plugins@sha256:c884d6d08f8c0db98964f1eb3877b44ade41fa106083802a9914775df17d5291
        - registry.k8s.io/sig-storage/livenessprobe@sha256:5baeb4a6d7d517434292758928bb33efc6397368cbb48c8a4cf29496abf4e987
        - quay.io/kubevirt/ovs-cni-plugin@sha256:e16ac74343da21abb8fb668ce71e728053d00503a992dae2164b9e94a280113e
        - quay.io/kubevirt/mtq-operator@sha256:90c7608b0ba085582a65ccdc11d089049c5e6b0438d018dd9dc76eb3490383c7
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:57573a24923e5588bca6bc337a8b2b08406c5b77583974365d2cf063c0dd5d06
        - quay.io/kubevirt/hostpath-provisioner@sha256:adb1d33e95e8991dc47107926172469b5ecbeace509391ce021ca42d305a41d3
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:c53535af8a7f7e3164609838c4b191b42b2d81238d75c1b2a2b582ada62a9780
        - quay.io/kubevirt/cdi-cloner@sha256:b22985af0ee7b1ce4e1dedded08c853d90d9e7209f98c5d470897bbefce2a063
        - quay.io/kubevirt/kubemacpool@sha256:cf8daa57ae6603b776d3af512331b143fa03bc2f4b72f28420fddcf5e4156d0a
        - quay.io/kubevirt/mtq-lock-server@sha256:8e686107dd9990dab1a3d90ee1068541938f673bf54ea9e1308b2ed7218546b0
        - ghcr.io/kubevirt/kubesecondarydns@sha256:e87e829380a1e576384145f78ccaa885ba1d5690d5de7d0b73d40cfb804ea24d
        - quay.io/kubevirt/cdi-operator@sha256:8a27e1450c3bf8c1d8924e74470a62419d02ed2202924b082a74e2414b1fc21a
        - quay.io/kubevirt/aaq-server@sha256:0c6b01c48030e03e99b3ff2b319658c2bf2b2001f4e3e030ba70d1143b25717a
        - quay.io/kubevirt/cdi-importer@sha256:7320dce01f4628363992997b592200b4804daf15a9f237cd50a4a6acab7798e8
        - quay.io/kubevirt/virt-exportproxy@sha256:942c548d4848a87945eba51683002797eb4e1f5c463a1873115b2ec0e1e6e169
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:3fbcc32bd4e4d15bd93c96def784a229cd84cca27942bf4858b581f31c97ee02
        version: 1.11.0
      entries:
      - name: kubevirt-hyperconverged-operator.v1.11.0
        version: 1.11.0
      name: 1.11.0
    - currentCSV: kubevirt-hyperconverged-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:7afeef450521fb7bbc918e874d9297dc3762cdeb28db0bef96d66d48b8752a9d
          createdAt: "2024-08-07 15:33:27"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: quay.io/kubevirt/must-gather
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - quay.io/kubevirt/virt-controller@sha256:3471b00eb9f4774ae33867c7d0b83e1d26a065a2bd19d305b5d3946ab665b5e9
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - quay.io/kubevirt/aaq-operator@sha256:002513b78dfd795f5195534b651eace136ead3d73c8ac8cae145607e79c8a8ba
        - quay.io/kubevirt/bridge-marker@sha256:bba066e3b5ff3fb8c5e20861fe8abe51e3c9b50ad6ce3b2616af9cb5479a06d0
        - quay.io/kubevirt/libguestfs-tools@sha256:0de3de89720d4e5433a33944f00dac590daed6ab12e4d26a657ef082d0573747
        - quay.io/kubevirt/virt-exportserver@sha256:acac33a19528592cc3ef0cd3a25fa4c72bcf992e4c2c247437007fcb9de0aa6d
        - quay.io/kubevirt/ovs-cni-plugin@sha256:e16ac74343da21abb8fb668ce71e728053d00503a992dae2164b9e94a280113e
        - quay.io/kubevirt/virtio-container-disk@sha256:f8e71a3a930ce95d646d80862d5ff6f9d3ed7ef2541dd2e2acec0bce26274734
        - quay.io/kubevirt/kubevirt-template-validator@sha256:8f7c389a7cd7ef34f2edc84245bc1d0ebcf13e86da5e0b8893f56733c482b10d
        - quay.io/kubevirt/macvtap-cni@sha256:850b89343ace7c7ea6b18dd8e11964613974e9d1f7377af03854d407fb15230a
        - quay.io/kubevirt/ssp-operator@sha256:a15a53d0f3ef530fe469acf3b7673486011605d0021d4dbe29af94096b3b65ee
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:3fbcc32bd4e4d15bd93c96def784a229cd84cca27942bf4858b581f31c97ee02
        - quay.io/kubevirt/virt-handler@sha256:2a7d96da4de3ea51339b3a047abb4b05e088e3d1080110b42a85ea5baff4a0cc
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:935475c2850466aa5ac57e4de627fb177515cb2c402a95842ead095d82b6df5f
        - quay.io/kubevirt/cdi-cloner@sha256:05da8df0b64024f9e2a48a1a54f06c821d79a3efe7828fad2c4fda6a01d0cd86
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - quay.io/kubevirt/cdi-apiserver@sha256:4ba004f57e7d589ddb5372a9729360460a97effecc6e5ecaccd7fbf16c6fcf13
        - quay.io/kubevirt/virt-api@sha256:aeeae94c9dd3b352ff2fbcdba77a6b084fb4ebe0fed35fe7af9481e9e2b1f24c
        - quay.io/kubevirt/cni-default-plugins@sha256:0c354fa9d695b8cab97b459e8afea2f7662407a987e83f6f6f1a8af4b45726be
        - quay.io/kubevirt/cdi-controller@sha256:392ead2fdc5050c584765ebf99706cc5c0ab9c2b424b692ad35005e77d110b58
        - quay.io/kubevirt/aaq-controller@sha256:170abd7a75397ddce07ba0d64886ead25183c49cacbccf0d191a4dd37ca013fd
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:3c70cb833819d8346e5531b10ac516283bab7098e782708768a6fe00ae0ddfac
        - quay.io/kubevirt/mtq-controller@sha256:04b38720b65479046270c8132cff5ab0f2dabc9cae91de1cc95fecf167efc356
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:413a6c229ebb608868e8b2c7c06b251c6b1d43c07af12a76b893292cbed8bef9
        - quay.io/kubevirt/mtq-lock-server@sha256:8b5189518ac6f68161d6a259e2a6f70e7902855ea1510aa260e79093520a7803
        - quay.io/kubevirt/cdi-importer@sha256:de5b89c879a531165482d540deffdc2dec0c188fd0f7b5fbd49c0298cf9a5b1d
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:d51821f001c959a79943cb22f5ea47589d7cf5c032e11ec1d571e8db7bf29b78
        - quay.io/kubevirt/hostpath-csi-driver@sha256:b7ccc469ad7e640fed2e98b50eae57a64e81c7b9862a799b7a9892d14ade8937
        - quay.io/kubevirt/kubemacpool@sha256:20e156be33e6d3692c456081acbb91a4349f94de448f7d1f1cddd0228931b31f
        - quay.io/kubevirt/virt-exportproxy@sha256:dd3b5c12314fa76b88d55d978210de81562fb888351748ea3997551c8c96a877
        - quay.io/kubevirt/cdi-operator@sha256:f68e4f2a082fc72d92bc1df0340285af03cd12d8c7721b38c619d1de0f4c175a
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:246b3819cfa3833f9a8cb01e8cb859aa61a1924a85ea4a112e43e8414f32c429
        - ghcr.io/kubevirt/kubesecondarydns@sha256:6268d84154e2483fbce8c1adacbdaf6f0839117b2d48d9fa4687cc8f76bd5130
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:83b460502671fb4f34116363a1a39b2ddfc9d14a920ee0a6413bfc3bd0580404
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:7afeef450521fb7bbc918e874d9297dc3762cdeb28db0bef96d66d48b8752a9d
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:e2def4213ec0657e72eb790ae8a115511d5b8f164a62d3568d2f1bff189917e8
        - quay.io/kubevirt/mtq-operator@sha256:859524721ae1a3b1e5dd6e29036b4c80acff13f74a4ee6769d867283e9ccbba5
        - quay.io/kubevirt/cdi-uploadserver@sha256:d7fb653bb33dade84583fff6048092346639532d6675aaf9f88070171cb0fe7c
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:c4b6b02737bc24906fcce57fe6626d1a36cb2b91baa971af2a5e5a919093c34e
        - quay.io/kubevirt/cdi-uploadproxy@sha256:86285df5cf52533732ddf4fdb74674cbce877ba69833cb8ed058c2a157cb5573
        - quay.io/kubevirt/aaq-server@sha256:9db6ca9147b1d75ba4d6bb1326c4ec289327eb9f8aadbcd951fcc689e584cd0d
        - quay.io/kubevirt/virt-artifacts-server@sha256:cc3e1e17bc3fa2a895f934210c9106bee01c3f9bf001f988d920eb58628e347b
        - quay.io/kubevirt/virt-launcher@sha256:cd37543ed2bc6baadbcbffa6951d740e4c65dc0ea366dec0a8593193c384e6de
        - quay.io/kubevirt/hostpath-provisioner@sha256:2b93b6c4ce0307a8f0dbdae2226da4acb1bae781788fbde7f0c32d6ad6421aa0
        - registry.k8s.io/sig-storage/livenessprobe@sha256:5baeb4a6d7d517434292758928bb33efc6397368cbb48c8a4cf29496abf4e987
        - quay.io/kubevirt/virt-operator@sha256:1ecc62c25f869e17027c930150d5c9b52b72d7e1fb434376bc0be0d886952b2e
        version: 1.12.0
      entries:
      - name: kubevirt-hyperconverged-operator.v1.12.0
        version: 1.12.0
      - name: kubevirt-hyperconverged-operator.v1.11.1
        version: 1.11.1
      - name: kubevirt-hyperconverged-operator.v1.10.0
        version: 1.10.0
      - name: kubevirt-hyperconverged-operator.v1.9.0
        version: 1.9.0
      - name: kubevirt-hyperconverged-operator.v1.8.2
        version: 1.8.2
      - name: kubevirt-hyperconverged-operator.v1.8.1
        version: 1.8.1
      - name: kubevirt-hyperconverged-operator.v1.8.0
        version: 1.8.0
      - name: kubevirt-hyperconverged-operator.v1.7.0
        version: 1.7.0
      - name: kubevirt-hyperconverged-operator.v1.6.0
        version: 1.6.0
      - name: kubevirt-hyperconverged-operator.v1.5.2
        version: 1.5.2
      - name: kubevirt-hyperconverged-operator.v1.5.1
        version: 1.5.1
      - name: kubevirt-hyperconverged-operator.v1.5.0
        version: 1.5.0
      - name: kubevirt-hyperconverged-operator.v1.4.4
        version: 1.4.4
      - name: kubevirt-hyperconverged-operator.v1.4.3
        version: 1.4.3
      - name: kubevirt-hyperconverged-operator.v1.4.2
        version: 1.4.2
      - name: kubevirt-hyperconverged-operator.v1.4.1
        version: 1.4.1
      - name: kubevirt-hyperconverged-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: community-kubevirt-hyperconverged
    provider:
      name: KubeVirt project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: StreamNative
      provider-url: https://streamnative.io
    name: sn-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
        - docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21
        - docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      name: alpha
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4
        - quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
        - docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      name: beta
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator@sha256:7051e28cbeb1c0058eabd45e07e7a6c2943cb16922177499f03c78fb376656ae
        - docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway@sha256:b4895775adceb874b226a976c5a4c129a684845cf2f4447a9162c441480ffd21
        - docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc@sha256:c6470b2c409f3c5e9e35f51f88303a9242bca5a09fb5d6c65d6358b2d18b33f4
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      name: stable
    defaultChannel: alpha
    packageName: sn-operator
    provider:
      name: StreamNative
      url: https://streamnative.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzo-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzo-operator.v2.0.101
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzo.cambridgesemantics.com/v1",
                "kind": "Anzo",
                "metadata": {
                  "name": "agent01"
                },
                "spec": {
                  "nodeConfig": {
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "app": "anzo"
                        }
                      },
                      "serviceName": "anzo-agent01",
                      "template": {
                        "metadata": {
                          "labels": {
                            "app": "anzo"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo@sha256:d1c2d54ca205d669c5ea1bb6aae105afe2385cc5695cecd78423ffa1d98929f8",
                              "name": "anzo",
                              "resources": {
                                "limits": {
                                  "cpu": "4000m",
                                  "memory": "12Gi"
                                },
                                "requests": {
                                  "cpu": "4000m",
                                  "memory": "12Gi"
                                }
                              }
                            }
                          ],
                          "serviceAccountName": "anzo-operator"
                        }
                      }
                    }
                  },
                  "role": "AnzoAgent"
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-operator@sha256:563e964f5771b9bfd860ea993b6e0936a54504908f5b2d1dff6f99ade74bb7a0
          createdAt: "2021-08-02"
          description: kubernetes operator for Anzo
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzo-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzo-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Anzo is the Schema for the anzos API
            displayName: Anzo Operator
            kind: Anzo
            name: anzos.anzo.cambridgesemantics.com
            version: v1
        description: |-
          The Anzo Operator provides the way to install and configure an anzo agent setup on Red Hat K8S environment.
          Currently, this is supported only through when deployed as an Anzo Agent as part of an Anzo Unstructured deployment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v2.0.0/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers through [ Cambridge Semantics Customer Center ]( https://customercenter.cambridgesemantics.com/ ).
        displayName: Anzo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        links:
        - name: Anzo Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v2.0.0/README_openshift.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-operator@sha256:563e964f5771b9bfd860ea993b6e0936a54504908f5b2d1dff6f99ade74bb7a0
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo@sha256:d1c2d54ca205d669c5ea1bb6aae105afe2385cc5695cecd78423ffa1d98929f8
        version: 2.0.0
      entries:
      - name: anzo-operator.v2.0.101
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anzo-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: odr-hub-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odr-hub-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east"
                },
                "spec": {
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east1"
                },
                "spec": {
                  "cidrs": [
                    "192.168.39.45/32"
                  ],
                  "clusterFence": "ManuallyFenced",
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east1"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPlacementControl",
                "metadata": {
                  "name": "drplacementcontrol-sample",
                  "namespace": "application-namespace"
                },
                "spec": {
                  "drPolicyRef": {
                    "name": "drpolicy-sample"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "excludedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "includedResources": [
                          "*"
                        ]
                      }
                    ]
                  },
                  "placementRef": {
                    "kind": "PlacementRule",
                    "name": "application-placement-rule"
                  },
                  "preferredCluster": "east",
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east",
                    "west"
                  ],
                  "replicationClassSelector": {
                    "matchLabels": {
                      "class": "ramen"
                    }
                  },
                  "schedulingInterval": "1h"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "metrodr-drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east1",
                    "east2"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DRCluster is the Schema for the drclusters API
            displayName: DRCluster
            kind: DRCluster
            name: drclusters.ramendr.openshift.io
            version: v1alpha1
          - description: DRPlacementControl is the Schema for the drplacementcontrols
              API
            displayName: DRPlacement Control
            kind: DRPlacementControl
            name: drplacementcontrols.ramendr.openshift.io
            version: v1alpha1
          - description: DRPolicy is the Schema for the drpolicies API
            displayName: DRPolicy
            kind: DRPolicy
            name: drpolicies.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Hub is a disaster-recovery orchestrator for stateful
          applications. It operates from an Advanced Cluster Management (ACM) hub
          cluster to orchestrate the recovery of application state, and scheduling
          of ACM PlacementRule for disaster recovery operations.
        displayName: Openshift DR Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:769acf15d016d4992793c9c527d7c268d2a67502dd7e0fadc240e4896398195b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 4.15.20-rhodf
      entries:
      - name: odr-hub-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odr-hub-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odr-hub-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odr-hub-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odr-hub-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odr-hub-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odr-hub-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odr-hub-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odr-hub-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odr-hub-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odr-hub-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odr-hub-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odr-hub-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odr-hub-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odr-hub-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odr-hub-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odr-hub-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odr-hub-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odr-hub-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odr-hub-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odr-hub-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east"
                },
                "spec": {
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east1"
                },
                "spec": {
                  "cidrs": [
                    "192.168.39.45/32"
                  ],
                  "clusterFence": "ManuallyFenced",
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east1"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPlacementControl",
                "metadata": {
                  "name": "drplacementcontrol-sample",
                  "namespace": "application-namespace"
                },
                "spec": {
                  "drPolicyRef": {
                    "name": "drpolicy-sample"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "excludedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "includedResources": [
                          "*"
                        ]
                      }
                    ]
                  },
                  "placementRef": {
                    "kind": "PlacementRule",
                    "name": "application-placement-rule"
                  },
                  "preferredCluster": "east",
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east",
                    "west"
                  ],
                  "replicationClassSelector": {
                    "matchLabels": {
                      "class": "ramen-replication-class"
                    }
                  },
                  "schedulingInterval": "1h",
                  "volumeSnapshotClassSelector": {
                    "matchLabels": {
                      "class": "ramen-snapshot-class"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "metrodr-drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east1",
                    "east2"
                  ],
                  "replicationClassSelector": {},
                  "schedulingInterval": {},
                  "volumeSnapshotClassSelector": {}
                }
              }
            ]
          capabilities: Deep Insights
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DRCluster is the Schema for the drclusters API
            displayName: DRCluster
            kind: DRCluster
            name: drclusters.ramendr.openshift.io
            version: v1alpha1
          - description: DRPlacementControl is the Schema for the drplacementcontrols
              API
            displayName: DRPlacement Control
            kind: DRPlacementControl
            name: drplacementcontrols.ramendr.openshift.io
            version: v1alpha1
          - description: DRPolicy is the Schema for the drpolicies API
            displayName: DRPolicy
            kind: DRPolicy
            name: drpolicies.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Hub is a disaster-recovery orchestrator for stateful
          applications. It operates from an Advanced Cluster Management (ACM) hub
          cluster to orchestrate the recovery of application state, and scheduling
          of ACM PlacementRule for disaster recovery operations.
        displayName: Openshift DR Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:af262b0890300183db1766bc392017b0bda723a9ba841dc25e3d928aa413db9a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        version: 4.16.16-rhodf
      entries:
      - name: odr-hub-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odr-hub-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odr-hub-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odr-hub-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odr-hub-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odr-hub-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odr-hub-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odr-hub-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odr-hub-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odr-hub-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odr-hub-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odr-hub-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odr-hub-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odr-hub-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odr-hub-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odr-hub-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odr-hub-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odr-hub-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: container-security-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: container-security-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secscan.quay.redhat.com/v1alpha1",
                "kind": "ImageManifestVuln",
                "metadata": {
                  "selfLink": "/apis/secscan.quay.redhat.com/v1alpha1/namespaces/openshift-cluster-version/imagemanifestvulns/sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18",
                  "resourceVersion": "14565395",
                  "name": "sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18",
                  "uid": "3f16a188-e69a-406c-aa06-df340195409c",
                  "creationTimestamp": "2021-03-17T12:07:16Z",
                  "generation": 2,
                  "managedFields": [
                    {
                      "apiVersion": "secscan.quay.redhat.com/v1alpha1",
                      "fieldsType": "FieldsV1",
                      "fieldsV1": {
                        "f:metadata": {
                          "f:labels": {
                            ".": {},
                            "f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": {}
                          }
                        },
                        "f:spec": {
                          ".": {},
                          "f:features": {},
                          "f:image": {},
                          "f:manifest": {}
                        },
                        "f:status": {
                          ".": {},
                          "f:affectedPods": {
                            ".": {},
                            "f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": {}
                          },
                          "f:fixableCount": {},
                          "f:highCount": {},
                          "f:highestSeverity": {},
                          "f:lastUpdate": {},
                          "f:lowCount": {},
                          "f:mediumCount": {}
                        }
                      },
                      "manager": "security-labeller",
                      "operation": "Update",
                      "time": "2021-03-17T13:07:20Z"
                    }
                  ],
                  "namespace": "openshift-cluster-version",
                  "labels": {
                    "openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": "true"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:447e05859134a0455ca55f8aae16c8a27ce06af4c18c191b41f6339df553f1bf
          createdAt: 2021-11-02 11:11 UTC
          description: Identify image vulnerabilities in Kubernetes pods
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Quay Container Security Operator (CSO) brings Quay and Clair
          metadata to Kubernetes / OpenShift. Starting with vulnerability information
          the scope will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Quay Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Quay Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:447e05859134a0455ca55f8aae16c8a27ce06af4c18c191b41f6339df553f1bf
        version: 3.4.7
      entries:
      - name: container-security-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: container-security-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"secscan.quay.redhat.com/v1alpha1\",\n
            \   \"kind\": \"ImageManifestVuln\",\n    \"metadata\": {\n      \"selfLink\":
            \"/apis/secscan.quay.redhat.com/v1alpha1/namespaces/openshift-cluster-version/imagemanifestvulns/sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18\",\n
            \     \"resourceVersion\": \"14565395\",\n      \"name\": \"sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18\",\n
            \     \"uid\": \"3f16a188-e69a-406c-aa06-df340195409c\",\n      \"creationTimestamp\":
            \"2021-03-17T12:07:16Z\",\n      \"generation\": 2,\n      \"managedFields\":
            [\n        {\n          \"apiVersion\": \"secscan.quay.redhat.com/v1alpha1\",\n
            \         \"fieldsType\": \"FieldsV1\",\n          \"fieldsV1\": {\n            \"f:metadata\":
            {\n              \"f:labels\": {\n                \".\": {},\n                \"f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            {}\n              }\n            },\n            \"f:spec\": {\n              \".\":
            {},\n              \"f:features\": {},\n              \"f:image\": {},\n
            \             \"f:manifest\": {}\n            },\n            \"f:status\":
            {\n              \".\": {},\n              \"f:affectedPods\": {\n                \".\":
            {},\n                \"f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            {}\n              },\n              \"f:fixableCount\": {},\n              \"f:highCount\":
            {},\n              \"f:highestSeverity\": {},\n              \"f:lastUpdate\":
            {},\n              \"f:lowCount\": {},\n              \"f:mediumCount\":
            {}\n            }\n          },\n          \"manager\": \"security-labeller\",\n
            \         \"operation\": \"Update\",\n          \"time\": \"2021-03-17T13:07:20Z\"\n
            \       }\n      ],\n      \"namespace\": \"openshift-cluster-version\",\n
            \     \"labels\": {\n        \"openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            \"true\"\n      }\n    },\n    \"spec\": {\n      \"features\": [\n        {\n
            \         \"name\": \"musl\",\n          \"namespaceName\": \"alpine:v3.14\",\n
            \         \"version\": \"1.2.2-r3\",\n          \"versionformat\": \"dpkg\",\n
            \         \"vulnerabilities\": [\n            {\n              \"fixedby\":
            \"1.2.2_pre2-r0\",\n              \"link\": \"https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28928\",\n
            \             \"metadata\": \"{\\\"NVD\\\": {\\\"CVSSv3\\\": {\\\"Vectors\\\":
            \\\"CVSS:3.0/AV:L/AC:L/PR:L/UI:N/S:U/C:N/I:N/A:H\\\", \\\"ImpactScore\\\":
            3.6, \\\"Score\\\": 5.5, \\\"ExploitabilityScore\\\": 1.8}, \\\"CVSSv2\\\":
            {\\\"Score\\\": 2.1, \\\"PublishedDateTime\\\": \\\"2020-11-24T18:15Z\\\",
            \\\"Vectors\\\": \\\"AV:L/AC:L/Au:N/C:N/I:N/A:P\\\"}}}\",\n              \"name\":
            \"CVE-2020-28928\",\n              \"namespaceName\": \"alpine:v3.14\",\n
            \             \"severity\": \"Low\"\n            }\n          ]\n        }\n
            \     ],\n      \"image\": \"quay.io/harishg/pruned-redhat-operators-index\",\n
            \     \"manifest\": \"sha256:19631c86642609254777ff59fc9e128aa5d0a94a7dbb504ef1fcfb2bbc6e486e\"\n
            \   }\n  }\n]    "
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:6b3581e62550d6a615d36eee6eaf4c89b9c35ba53cab8433bd93c68393a9d115
          createdAt: 2021-09-10 23:09 UTC
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.4.x <3.5.7'
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Quay Container Security Operator (CSO) brings Quay and Clair
          metadata to Kubernetes / OpenShift. Starting with vulnerability information
          the scope will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Quay Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Quay Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:6b3581e62550d6a615d36eee6eaf4c89b9c35ba53cab8433bd93c68393a9d115
        version: 3.5.7
      entries:
      - name: container-security-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: container-security-operator.v3.10.15
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:de7c66a0004f3c0643c09d4c3bbb4a0f1f8f2cbf264250802cdc58028fc72b50
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.10.15
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:de7c66a0004f3c0643c09d4c3bbb4a0f1f8f2cbf264250802cdc58028fc72b50
        version: 3.10.15
      entries:
      - name: container-security-operator.v3.10.15
        version: 3.10.15
      - name: container-security-operator.v3.10.14
        version: 3.10.14
      - name: container-security-operator.v3.10.13
        version: 3.10.13
      - name: container-security-operator.v3.10.12
        version: 3.10.12
      - name: container-security-operator.v3.10.11
        version: 3.10.11
      - name: container-security-operator.v3.10.10
        version: 3.10.10
      - name: container-security-operator.v3.10.9
        version: 3.10.9
      - name: container-security-operator.v3.10.8
        version: 3.10.8
      - name: container-security-operator.v3.10.7
        version: 3.10.7
      - name: container-security-operator.v3.10.6
        version: 3.10.6
      - name: container-security-operator.v3.10.5
        version: 3.10.5
      - name: container-security-operator.v3.10.4
        version: 3.10.4
      - name: container-security-operator.v3.10.3
        version: 3.10.3
      - name: container-security-operator.v3.10.2
        version: 3.10.2
      - name: container-security-operator.v3.10.1
        version: 3.10.1
      - name: container-security-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: container-security-operator.v3.11.13
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:3914389fd53c371627921298c5dc3564999d92bc720f27a240f74c034ebbf5ec
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.11.13
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:3914389fd53c371627921298c5dc3564999d92bc720f27a240f74c034ebbf5ec
        version: 3.11.13
      entries:
      - name: container-security-operator.v3.11.13
        version: 3.11.13
      - name: container-security-operator.v3.11.12
        version: 3.11.12
      - name: container-security-operator.v3.11.11
        version: 3.11.11
      - name: container-security-operator.v3.11.10
        version: 3.11.10
      - name: container-security-operator.v3.11.9
        version: 3.11.9
      - name: container-security-operator.v3.11.8
        version: 3.11.8
      - name: container-security-operator.v3.11.7
        version: 3.11.7
      - name: container-security-operator.v3.11.6
        version: 3.11.6
      - name: container-security-operator.v3.11.5
        version: 3.11.5
      - name: container-security-operator.v3.11.4
        version: 3.11.4
      - name: container-security-operator.v3.11.3
        version: 3.11.3
      - name: container-security-operator.v3.11.2
        version: 3.11.2
      - name: container-security-operator.v3.11.1
        version: 3.11.1
      - name: container-security-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: container-security-operator.v3.12.12
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:1a15328f0b2c64d3a857f2f155ad4def095027e809610de98545e303eec63225
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.12.12
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:1a15328f0b2c64d3a857f2f155ad4def095027e809610de98545e303eec63225
        version: 3.12.12
      entries:
      - name: container-security-operator.v3.12.12
        version: 3.12.12
      - name: container-security-operator.v3.12.11
        version: 3.12.11
      - name: container-security-operator.v3.12.10
        version: 3.12.10
      - name: container-security-operator.v3.12.9
        version: 3.12.9
      - name: container-security-operator.v3.12.8
        version: 3.12.8
      - name: container-security-operator.v3.12.7
        version: 3.12.7
      - name: container-security-operator.v3.12.6
        version: 3.12.6
      - name: container-security-operator.v3.12.5
        version: 3.12.5
      - name: container-security-operator.v3.12.4
        version: 3.12.4
      - name: container-security-operator.v3.12.3
        version: 3.12.3
      - name: container-security-operator.v3.12.2
        version: 3.12.2
      - name: container-security-operator.v3.12.1
        version: 3.12.1
      - name: container-security-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: container-security-operator.v3.13.8
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:ba50621739581da9a661c2c2e355849c328cfa33b7311b06abb148e1183a6321
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.13.8
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:ba50621739581da9a661c2c2e355849c328cfa33b7311b06abb148e1183a6321
        version: 3.13.8
      entries:
      - name: container-security-operator.v3.13.8
        version: 3.13.8
      - name: container-security-operator.v3.13.7
        version: 3.13.7
      - name: container-security-operator.v3.13.6
        version: 3.13.6
      - name: container-security-operator.v3.13.5
        version: 3.13.5
      - name: container-security-operator.v3.13.4
        version: 3.13.4
      - name: container-security-operator.v3.13.3
        version: 3.13.3
      - name: container-security-operator.v3.13.2
        version: 3.13.2
      - name: container-security-operator.v3.13.1
        version: 3.13.1
      - name: container-security-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: container-security-operator.v3.14.5
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:c42487ccf098df3ec6dbeb5d409346bdd30e92781cf3ffea5fdb6dd0b196d0cf
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.14.5
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:c42487ccf098df3ec6dbeb5d409346bdd30e92781cf3ffea5fdb6dd0b196d0cf
        version: 3.14.5
      entries:
      - name: container-security-operator.v3.14.5
        version: 3.14.5
      - name: container-security-operator.v3.14.4
        version: 3.14.4
      - name: container-security-operator.v3.14.3
        version: 3.14.3
      - name: container-security-operator.v3.14.2
        version: 3.14.2
      - name: container-security-operator.v3.14.1
        version: 3.14.1
      - name: container-security-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: container-security-operator.v3.15.2
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:81e929ec095f791a83310ee7fcda239d5c348ba69a3cf5adcf1b3a77c349ce77
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.15.2
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:81e929ec095f791a83310ee7fcda239d5c348ba69a3cf5adcf1b3a77c349ce77
        version: 3.15.2
      entries:
      - name: container-security-operator.v3.15.2
        version: 3.15.2
      - name: container-security-operator.v3.15.1
        version: 3.15.1
      - name: container-security-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: container-security-operator.v3.6.10
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:641b80df8d7102d49a5dfd5b99245669f60393880a6f1bc8ba8450b88a640d96
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.6.10
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:641b80df8d7102d49a5dfd5b99245669f60393880a6f1bc8ba8450b88a640d96
        version: 3.6.10
      entries:
      - name: container-security-operator.v3.6.10
        version: 3.6.10
      - name: container-security-operator.v3.6.9
        version: 3.6.9
      - name: container-security-operator.v3.6.8
        version: 3.6.8
      - name: container-security-operator.v3.6.7
        version: 3.6.7
      - name: container-security-operator.v3.6.6
        version: 3.6.6
      - name: container-security-operator.v3.6.5
        version: 3.6.5
      - name: container-security-operator.v3.6.4
        version: 3.6.4
      - name: container-security-operator.v3.6.2
        version: 3.6.2
      - name: container-security-operator.v3.6.1
        version: 3.6.1
      - name: container-security-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: container-security-operator.v3.7.14
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:10736a06beb503b4eb73260a87a2cde9bd83353374ee0ac9e8dd0a6260cf2665
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.7.14
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:10736a06beb503b4eb73260a87a2cde9bd83353374ee0ac9e8dd0a6260cf2665
        version: 3.7.14
      entries:
      - name: container-security-operator.v3.7.14
        version: 3.7.14
      - name: container-security-operator.v3.7.13
        version: 3.7.13
      - name: container-security-operator.v3.7.12
        version: 3.7.12
      - name: container-security-operator.v3.7.11
        version: 3.7.11
      - name: container-security-operator.v3.7.10
        version: 3.7.10
      - name: container-security-operator.v3.7.9
        version: 3.7.9
      - name: container-security-operator.v3.7.8
        version: 3.7.8
      - name: container-security-operator.v3.7.7
        version: 3.7.7
      - name: container-security-operator.v3.7.6
        version: 3.7.6
      - name: container-security-operator.v3.7.5
        version: 3.7.5
      - name: container-security-operator.v3.7.4
        version: 3.7.4
      - name: container-security-operator.v3.7.3
        version: 3.7.3
      - name: container-security-operator.v3.7.2
        version: 3.7.2
      - name: container-security-operator.v3.7.1
        version: 3.7.1
      - name: container-security-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: container-security-operator.v3.8.15
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:108bd033543beb7cce077130c96f992c202f7a258229c239795f8f19e5fae5e5
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.8.15
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:108bd033543beb7cce077130c96f992c202f7a258229c239795f8f19e5fae5e5
        version: 3.8.15
      entries:
      - name: container-security-operator.v3.8.15
        version: 3.8.15
      - name: container-security-operator.v3.8.14
        version: 3.8.14
      - name: container-security-operator.v3.8.13
        version: 3.8.13
      - name: container-security-operator.v3.8.12
        version: 3.8.12
      - name: container-security-operator.v3.8.11
        version: 3.8.11
      - name: container-security-operator.v3.8.10
        version: 3.8.10
      - name: container-security-operator.v3.8.9
        version: 3.8.9
      - name: container-security-operator.v3.8.8
        version: 3.8.8
      - name: container-security-operator.v3.8.7
        version: 3.8.7
      - name: container-security-operator.v3.8.6
        version: 3.8.6
      - name: container-security-operator.v3.8.5
        version: 3.8.5
      - name: container-security-operator.v3.8.4
        version: 3.8.4
      - name: container-security-operator.v3.8.3
        version: 3.8.3
      - name: container-security-operator.v3.8.2
        version: 3.8.2
      - name: container-security-operator.v3.8.1
        version: 3.8.1
      - name: container-security-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: container-security-operator.v3.9.16
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:df8e1d246273d21049c6bb8169a2e5f3fc8d69c85c564359abea7364969c6acc
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.9.16
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:df8e1d246273d21049c6bb8169a2e5f3fc8d69c85c564359abea7364969c6acc
        version: 3.9.16
      entries:
      - name: container-security-operator.v3.9.16
        version: 3.9.16
      - name: container-security-operator.v3.9.15
        version: 3.9.15
      - name: container-security-operator.v3.9.14
        version: 3.9.14
      - name: container-security-operator.v3.9.13
        version: 3.9.13
      - name: container-security-operator.v3.9.12
        version: 3.9.12
      - name: container-security-operator.v3.9.11
        version: 3.9.11
      - name: container-security-operator.v3.9.10
        version: 3.9.10
      - name: container-security-operator.v3.9.9
        version: 3.9.9
      - name: container-security-operator.v3.9.8
        version: 3.9.8
      - name: container-security-operator.v3.9.7
        version: 3.9.7
      - name: container-security-operator.v3.9.6
        version: 3.9.6
      - name: container-security-operator.v3.9.5
        version: 3.9.5
      - name: container-security-operator.v3.9.4
        version: 3.9.4
      - name: container-security-operator.v3.9.3
        version: 3.9.3
      - name: container-security-operator.v3.9.2
        version: 3.9.2
      - name: container-security-operator.v3.9.1
        version: 3.9.1
      - name: container-security-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: container-security-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sfn-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sfn-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sfn.services.k8s.aws/v1alpha1",
                "kind": "Activity",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sfn.services.k8s.aws/v1alpha1",
                "kind": "StateMachine",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sfn-controller:1.1.1
          createdAt: "2025-09-19T18:41:33Z"
          description: AWS Step Functions controller is a service controller for managing
            Step Functions resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Activity represents the state of an AWS sfn Activity resource.
            displayName: Activity
            kind: Activity
            name: activities.sfn.services.k8s.aws
            version: v1alpha1
          - description: StateMachine represents the state of an AWS sfn StateMachine
              resource.
            displayName: StateMachine
            kind: StateMachine
            name: statemachines.sfn.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Step Functions resources in AWS from within your Kubernetes cluster.

          **About Amazon Step FUnctions**

          Step Functions is a serverless orchestration service that lets you combine AWS Lambda functions and other AWS services to build business-critical applications. Through Step Functions' graphical console, you see your application’s workflow as a series of event-driven steps.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Step Functions
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sfn
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Step Functions Developer Resources
          url: https://aws.amazon.com/step-functions/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: step functions maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sfn-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-sfn-controller.v1.1.1
        version: 1.1.1
      - name: ack-sfn-controller.v1.1.0
        version: 1.1.0
      - name: ack-sfn-controller.v1.0.29
        version: 1.0.29
      - name: ack-sfn-controller.v1.0.28
        version: 1.0.28
      - name: ack-sfn-controller.v1.0.27
        version: 1.0.27
      - name: ack-sfn-controller.v1.0.26
        version: 1.0.26
      - name: ack-sfn-controller.v1.0.25
        version: 1.0.25
      - name: ack-sfn-controller.v1.0.24
        version: 1.0.24
      - name: ack-sfn-controller.v1.0.23
        version: 1.0.23
      - name: ack-sfn-controller.v1.0.22
        version: 1.0.22
      - name: ack-sfn-controller.v1.0.21
        version: 1.0.21
      - name: ack-sfn-controller.v1.0.20
        version: 1.0.20
      - name: ack-sfn-controller.v1.0.19
        version: 1.0.19
      - name: ack-sfn-controller.v1.0.18
        version: 1.0.18
      - name: ack-sfn-controller.v1.0.17
        version: 1.0.17
      - name: ack-sfn-controller.v1.0.16
        version: 1.0.16
      - name: ack-sfn-controller.v1.0.15
        version: 1.0.15
      - name: ack-sfn-controller.v1.0.14
        version: 1.0.14
      - name: ack-sfn-controller.v1.0.13
        version: 1.0.13
      - name: ack-sfn-controller.v1.0.11
        version: 1.0.11
      - name: ack-sfn-controller.v1.0.10
        version: 1.0.10
      - name: ack-sfn-controller.v1.0.9
        version: 1.0.9
      - name: ack-sfn-controller.v1.0.8
        version: 1.0.8
      - name: ack-sfn-controller.v1.0.6
        version: 1.0.6
      - name: ack-sfn-controller.v1.0.5
        version: 1.0.5
      - name: ack-sfn-controller.v1.0.2
        version: 1.0.2
      - name: ack-sfn-controller.v1.0.1
        version: 1.0.1
      - name: ack-sfn-controller.v1.0.0
        version: 1.0.0
      - name: ack-sfn-controller.v0.1.3
        version: 0.1.3
      - name: ack-sfn-controller.v0.1.2
        version: 0.1.2
      - name: ack-sfn-controller.v0.1.1
        version: 0.1.1
      - name: ack-sfn-controller.v0.1.0
        version: 0.1.0
      - name: ack-sfn-controller.v0.0.19
        version: 0.0.19
      - name: ack-sfn-controller.v0.0.18
        version: 0.0.18
      - name: ack-sfn-controller.v0.0.17
        version: 0.0.17
      - name: ack-sfn-controller.v0.0.16
        version: 0.0.16
      - name: ack-sfn-controller.v0.0.15
        version: 0.0.15
      - name: ack-sfn-controller.v0.0.14
        version: 0.0.14
      - name: ack-sfn-controller.v0.0.13
        version: 0.0.13
      - name: ack-sfn-controller.v0.0.12
        version: 0.0.12
      - name: ack-sfn-controller.v0.0.11
        version: 0.0.11
      - name: ack-sfn-controller.v0.0.10
        version: 0.0.10
      - name: ack-sfn-controller.v0.0.9
        version: 0.0.9
      - name: ack-sfn-controller.v0.0.8
        version: 0.0.8
      - name: ack-sfn-controller.v0.0.7
        version: 0.0.7
      - name: ack-sfn-controller.v0.0.6
        version: 0.0.6
      - name: ack-sfn-controller.v0.0.5
        version: 0.0.5
      name: alpha
    defaultChannel: alpha
    packageName: ack-sfn-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: authorino-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: authorino-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/authorino-operator:v0.8.0
          createdAt: 2021-12-08T10-00-00Z
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta1
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: The operator to manage instances of Authorino
        displayName: Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: Authorino Operator
          url: https://github.com/Kuadrant/authorino-operator
        - name: Authorino
          url: https://github.com/Kuadrant/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/authorino-operator:v0.8.0
        version: 0.8.0
      entries:
      - name: authorino-operator.v0.8.0
        version: 0.8.0
      - name: authorino-operator.v0.7.0
        version: 0.7.0
      - name: authorino-operator.v0.6.0
        version: 0.6.0
      - name: authorino-operator.v0.5.0
        version: 0.5.0
      - name: authorino-operator.v0.4.1
        version: 0.4.1
      - name: authorino-operator.v0.4.0
        version: 0.4.0
      - name: authorino-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: authorino-operator.v0.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta3",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/authorino-operator:v0.16.0
          createdAt: "2024-11-25T13:22:24Z"
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta3
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: The operator to manage instances of Authorino
        displayName: Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: Authorino Operator
          url: https://github.com/Kuadrant/authorino-operator
        - name: Authorino
          url: https://github.com/Kuadrant/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/authorino-operator:v0.16.0
        - quay.io/kuadrant/authorino:v0.20.0
        version: 0.16.0
      entries:
      - name: authorino-operator.v0.16.0
        version: 0.16.0
      - name: authorino-operator.v0.15.1
        version: 0.15.1
      - name: authorino-operator.v0.15.0
        version: 0.15.0
      - name: authorino-operator.v0.14.0
        version: 0.14.0
      - name: authorino-operator.v0.13.1
        version: 0.13.1
      - name: authorino-operator.v0.13.0
        version: 0.13.0
      - name: authorino-operator.v0.12.0
        version: 0.12.0
      - name: authorino-operator.v0.11.1
        version: 0.11.1
      - name: authorino-operator.v0.10.0
        version: 0.10.0
      - name: authorino-operator.v0.9.0
        version: 0.9.0
      - name: authorino-operator.v0.8.0
        version: 0.8.0
      - name: authorino-operator.v0.7.0
        version: 0.7.0
      - name: authorino-operator.v0.6.0
        version: 0.6.0
      - name: authorino-operator.v0.5.0
        version: 0.5.0
      - name: authorino-operator.v0.4.1
        version: 0.4.1
      - name: authorino-operator.v0.4.0
        version: 0.4.0
      - name: authorino-operator.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: authorino-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Eclipse Foundation
      provider-url: ""
    name: eclipse-che
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eclipse-che.v7.110.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "org.eclipse.che/v1",
                "kind": "CheCluster",
                "metadata": {
                  "name": "eclipse-che",
                  "namespace": "eclipse-che"
                },
                "spec": {
                  "auth": {
                    "identityProviderURL": "",
                    "oAuthClientName": "",
                    "oAuthSecret": ""
                  },
                  "database": {
                    "externalDb": false
                  },
                  "k8s": {
                    "ingressDomain": null,
                    "tlsSecretName": null
                  },
                  "metrics": {
                    "enable": true
                  },
                  "server": {
                    "externalDevfileRegistries": [
                      {
                        "url": "https://registry.devfile.io"
                      }
                    ],
                    "externalDevfileRegistry": true,
                    "externalPluginRegistry": true,
                    "workspaceNamespaceDefault": "<username>-che"
                  },
                  "storage": {
                    "pvcStrategy": "common"
                  }
                }
              },
              {
                "apiVersion": "org.eclipse.che/v2",
                "kind": "CheCluster",
                "metadata": {
                  "name": "eclipse-che",
                  "namespace": "eclipse-che"
                },
                "spec": {
                  "components": {
                    "devfileRegistry": {
                      "disableInternalRegistry": true,
                      "externalDevfileRegistries": [
                        {
                          "url": "https://registry.devfile.io"
                        }
                      ]
                    },
                    "pluginRegistry": {
                      "disableInternalRegistry": true
                    }
                  },
                  "containerRegistry": {},
                  "devEnvironments": {},
                  "gitServices": {},
                  "networking": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/eclipse/che-operator@sha256:10c4d05c2cb848d04b97caf752c7f4d181e6973683263e45adec417342164e51
          createdAt: "2025-10-02T14:44:20Z"
          description: A Kube-native development solution that delivers portable and
            collaborative developer workspaces.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eclipse-che/che-operator
          support: Eclipse Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              The `CheCluster` custom resource allows defining and managing Eclipse Che server installation.
              Based on these settings, the  Operator automatically creates and maintains several ConfigMaps:
              `che`, `plugin-registry` that will contain the appropriate environment variables
              of the various components of the installation. These generated ConfigMaps must NOT be updated manually.
            displayName: Eclipse Che instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v2
          - description: The `CheCluster` custom resource allows defining and managing
              a Che server installation
            displayName: Eclipse Che instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v1
        description: |
          A collaborative Kubernetes-native development solution that delivers OpenShift workspaces and in-browser IDE for rapid cloud application development.
          This operator installs the Plugin and Devfile registries, Dashboard, Gateway and the Eclipse Che server, and configures these services.
          OpenShift OAuth is used directly for authentication. TLS mode is on.

          ## How to Install
          Press the **Install** button, choose the channel and the upgrade strategy, and wait for the **Installed** Operator status.
          When the operator is installed, create a new Custom Resource (CR) of Kind CheCluster (click the **Create New** button).
          The CR spec contains all defaults. You can start using Eclipse Che when the CR status is set to **Available**, and you see a URL to Eclipse Che.

          ## Defaults
          By default, the operator deploys Eclipse Che with:
          * 10Gi storage
          * Auto-generated passwords
          * Bundled Plugin and Devfile registries

          Use `oc edit checluster/eclipse-che -n eclipse-che` to update Eclipse Che default installation options.
          See more in the [Installation guide](https://www.eclipse.org/che/docs/stable/administration-guide/configuring-che/).

          ### Certificates
          Operator uses a default router certificate to secure Eclipse Che routes.
          Follow the [guide](https://www.eclipse.org/che/docs/stable/administration-guide/importing-untrusted-tls-certificates/)
          to import certificates into Eclipse Che.
        displayName: Eclipse Che
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eclipse che
        - workspaces
        - devtools
        - developer
        - ide
        - java
        links:
        - name: Product Page
          url: https://www.eclipse.org/che
        - name: Documentation
          url: https://www.eclipse.org/che/docs
        - name: Operator GitHub Repo
          url: https://github.com/eclipse-che/che-operator
        maintainers:
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Eclipse Foundation
        relatedImages:
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:354fc75eb7a21a934381e93d03ef9d42bc2c8ae8989fdcacecfb39b863b96ced
        - quay.io/eclipse/che-operator@sha256:10c4d05c2cb848d04b97caf752c7f4d181e6973683263e45adec417342164e51
        - quay.io/che-incubator/che-idea-dev-server@sha256:87c8b53cc295e3d7eb35b326e221a949553be3345135a4b6c1999d66f345a8b2
        - quay.io/oauth2-proxy/oauth2-proxy@sha256:dcb6ff8dd21bf3058f6a22c6fa385fa5b897a9cd3914c88a2cc2bb0a85f8065d
        - quay.io/che-incubator/che-code-server@sha256:357c95297ce2182ef0a792038ea122075fdeb6c8e835099be09e8d4f105eac2a
        - quay.io/che-incubator/che-idea@sha256:8aae69dc4b0c122491a75400639af0fe92b5e214c6e68ac97cda29fb58b44151
        - quay.io/devfile/universal-developer-image@sha256:3cf53fb9eab60ee3156e9edd4ec5af72b70e55d5ffb28f41ecb5b26e01dc9838
        - quay.io/che-incubator/che-code@sha256:cfcfbe4d665b18b11401c3c35b23b04d292c31e7f4be6d78a5e6d60fce776ed0
        - quay.io/che-incubator/che-code-server@sha256:2093baab1a73e9682f5892f80582481abfbef99d744472ee6098e2d82fb2899f
        - quay.io/eclipse/che--traefik@sha256:4cf907247939b5d20bf4eff73abd21cb413c339600dde76dbc94a874b2578a27
        - quay.io/eclipse/che-dashboard@sha256:2db3255752aebd75dceea578cd27b7e8ecb41c2290827ff7104ed77e1b61feb9
        - quay.io/eclipse/che-tls-secret-creator@sha256:cbb2fd8ce903d822377f2093af1aa2543672ed8c516d607a7a9dd857379dac5d
        - quay.io/eclipse/che-server@sha256:b83d536cdd3cb2b86def5599495c40870dde20c194a003736a2b47a5e5e17a06
        - quay.io/openshift/origin-oauth-proxy@sha256:870bfe92a4663720775c0dfe5728ecbb10a17f0644eef5f57276ec135034c6a1
        - quay.io/brancz/kube-rbac-proxy@sha256:738c854322f56d63ebab75de5210abcdd5e0782ce2d30c0ecd4620f63b24694d
        - quay.io/che-incubator/configbump@sha256:396714521307ae9f79453c492fb8064cf940cd6b4e6183a825f1056045623874
        - quay.io/che-incubator/header-rewrite-proxy@sha256:bd7873b8feef35f218f54c6251ea224bea2c8bf202a328230019a0ba2941245d
        - quay.io/che-incubator/che-code-sshd@sha256:7eb5381fba8878b8f252523caf70c9957b9f9e699db4564061017fb0a023cc1a
        - quay.io/che-incubator/che-code@sha256:ac62a2f42ae6d6f6628620a00133afece954fcbed2df309fa3de79e4c28cc029
        - quay.io/eclipse/che-plugin-registry@sha256:a70a0d98505ba5aeed691fc70c3c59aed3304155fba0ab3ad70a6d0453b22093
        version: 7.110.0
      entries:
      - name: eclipse-che.v7.110.0
        version: 7.110.0
      - name: eclipse-che.v7.109.0
        version: 7.109.0
      - name: eclipse-che.v7.108.0
        version: 7.108.0
      - name: eclipse-che.v7.107.0
        version: 7.107.0
      - name: eclipse-che.v7.106.0
        version: 7.106.0
      - name: eclipse-che.v7.105.0
        version: 7.105.0
      - name: eclipse-che.v7.104.0
        version: 7.104.0
      - name: eclipse-che.v7.103.0
        version: 7.103.0
      - name: eclipse-che.v7.102.0
        version: 7.102.0
      - name: eclipse-che.v7.101.0
        version: 7.101.0
      - name: eclipse-che.v7.100.0
        version: 7.100.0
      - name: eclipse-che.v7.99.0
        version: 7.99.0
      - name: eclipse-che.v7.98.0
        version: 7.98.0
      - name: eclipse-che.v7.97.0
        version: 7.97.0
      - name: eclipse-che.v7.96.0
        version: 7.96.0
      - name: eclipse-che.v7.95.0
        version: 7.95.0
      - name: eclipse-che.v7.94.0
        version: 7.94.0
      - name: eclipse-che.v7.93.0
        version: 7.93.0
      - name: eclipse-che.v7.92.0
        version: 7.92.0
      - name: eclipse-che.v7.91.0
        version: 7.91.0
      - name: eclipse-che.v7.90.0
        version: 7.90.0
      - name: eclipse-che.v7.89.0
        version: 7.89.0
      - name: eclipse-che.v7.88.0
        version: 7.88.0
      - name: eclipse-che.v7.87.0
        version: 7.87.0
      - name: eclipse-che.v7.86.0
        version: 7.86.0
      - name: eclipse-che.v7.85.0
        version: 7.85.0
      - name: eclipse-che.v7.84.0
        version: 7.84.0
      - name: eclipse-che.v7.83.0
        version: 7.83.0
      - name: eclipse-che.v7.82.0
        version: 7.82.0
      - name: eclipse-che.v7.81.0
        version: 7.81.0
      - name: eclipse-che.v7.80.0
        version: 7.80.0
      - name: eclipse-che.v7.79.0
        version: 7.79.0
      - name: eclipse-che.v7.78.0
        version: 7.78.0
      - name: eclipse-che.v7.77.0
        version: 7.77.0
      - name: eclipse-che.v7.76.0
        version: 7.76.0
      - name: eclipse-che.v7.75.0
        version: 7.75.0
      - name: eclipse-che.v7.74.0
        version: 7.74.0
      - name: eclipse-che.v7.73.0
        version: 7.73.0
      - name: eclipse-che.v7.72.0
        version: 7.72.0
      - name: eclipse-che.v7.71.0
        version: 7.71.0
      - name: eclipse-che.v7.70.0
        version: 7.70.0
      - name: eclipse-che.v7.69.0
        version: 7.69.0
      - name: eclipse-che.v7.68.0
        version: 7.68.0
      - name: eclipse-che.v7.67.0
        version: 7.67.0
      - name: eclipse-che.v7.66.0
        version: 7.66.0
      - name: eclipse-che.v7.65.0
        version: 7.65.0
      - name: eclipse-che.v7.64.0
        version: 7.64.0
      - name: eclipse-che.v7.63.0
        version: 7.63.0
      - name: eclipse-che.v7.62.0
        version: 7.62.0
      - name: eclipse-che.v7.61.0
        version: 7.61.0
      - name: eclipse-che.v7.60.1
        version: 7.60.1
      - name: eclipse-che.v7.60.0
        version: 7.60.0
      - name: eclipse-che.v7.59.0
        version: 7.59.0
      - name: eclipse-che.v7.58.0
        version: 7.58.0
      - name: eclipse-che.v7.57.0
        version: 7.57.0
      - name: eclipse-che.v7.56.0
        version: 7.56.0
      - name: eclipse-che.v7.55.0
        version: 7.55.0
      - name: eclipse-che.v7.54.0
        version: 7.54.0
      - name: eclipse-che.v7.53.0
        version: 7.53.0
      - name: eclipse-che.v7.52.0
        version: 7.52.0
      - name: eclipse-che.v7.51.0
        version: 7.51.0
      - name: eclipse-che.v7.50.0
        version: 7.50.0
      - name: eclipse-che.v7.49.0
        version: 7.49.0
      - name: eclipse-che.v7.48.1
        version: 7.48.1
      - name: eclipse-che.v7.47.0
        version: 7.47.0
      - name: eclipse-che.v7.46.0
        version: 7.46.0
      - name: eclipse-che.v7.45.0
        version: 7.45.0
      - name: eclipse-che.v7.44.0
        version: 7.44.0
      - name: eclipse-che.v7.42.0
        version: 7.42.0
      - name: eclipse-che.v7.41.2
        version: 7.41.2
      - name: eclipse-che.v7.41.1
        version: 7.41.1
      - name: eclipse-che.v7.41.0
        version: 7.41.0
      - name: eclipse-che.v7.40.2
        version: 7.40.2
      - name: eclipse-che.v7.40.1
        version: 7.40.1
      name: stable
    defaultChannel: stable
    packageName: eclipse-che
    provider:
      name: Eclipse Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solution-Soft
      provider-url: www.solution-soft.com
    name: timemachine-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: timemachine-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tm.solution-soft.com/v1alpha1",
                "kind": "TimeMachine",
                "metadata": {
                  "name": "timemachine-sample"
                },
                "spec": {
                  "cluster_provider": "ibm",
                  "aws": {
                    "shared_volume_name": ""
                  },
                  "ibm": {
                    "persistentvolume": {
                      "storageclass": "ibmc-file-silver-gid"
                    }
                  },
                  "tm_lichost": "52.152.148.40",
                  "tm_licpass": "qwaszx",
                  "tm_licport": "57777",
                  "imagePullSecrets": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
          createdAt: "2024-01-30T10:20:30Z"
          description: Time Machine Operator enables time travel of OpenShift pods,
            without the need to change the system clock on any of the cluster nodes
          olm.skipRange: '>=0.0.4 <1.0.17'
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: SolutionSoft Systems, Inc. - support@solution-soft.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Time Machine
            displayName: Time Machine
            kind: TimeMachine
            name: timemachines.tm.solution-soft.com
            version: v1alpha1
        description: |-
          Time Machine operator enables time travel of OpenShift pods/deployments.

          To time travel a pod where your application is, you just need to install the Time Machine operator and create a Time Machine custom resource in the same namespace, without the need of changing the container images of the application you’re using, or adding containers to the pod with your application.

          Time travelling is done on the namespace level where all the (configured) target pods/deployments see the same virtual time.
        displayName: Time Machine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - time shift
        - time travel
        - virtual clocks
        - time machine
        - virtual time
        - solution-soft
        links:
        - name: Timemachine Operator
          url: https://catalog.redhat.com/software/operators/detail/5e98727a6c5dcb34dfbb19e3
        maintainers:
        - email: dtrajkovic@solution-soft.com
          name: Dragan Trajkovic
        maturity: alpha
        provider:
          name: Solution-Soft
          url: www.solution-soft.com
        relatedImages:
        - registry.connect.redhat.com/rhel-timemachine-container/virtual-clock-testing-tool@sha256:a81a823328b0ee795c40668c89a08bcf545c5192d9eed0146071e3721a0a8372
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0c304c11986790f446604319dfc29db494b15f5f9fa13f2c436688ec29c92dbe
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
        version: 1.0.17
      entries:
      - name: timemachine-operator.v1.0.17
        version: 1.0.17
      - name: timemachine-operator.v1.0.16
        version: 1.0.16
      name: alpha
    defaultChannel: alpha
    packageName: timemachine-operator
    provider:
      name: Solution-Soft
      url: www.solution-soft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: antrea.io
      provider-url: ""
    name: antrea-by-broadcom-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: antrea-by-broadcom-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.antrea.vmware.com/v1",
                "kind": "AntreaInstall",
                "metadata": {
                  "name": "antrea-install",
                  "namespace": "antrea-operator"
                },
                "spec": {
                  "antreaAgentConfig": "# FeatureGates is a map of feature names to bools that enable or disable experimental features.\nfeatureGates:\n# AllAlpha is a global toggle for alpha features. Per-feature key values override the default set by AllAlpha.\n#  AllAlpha: false\n\n# AllBeta is a global toggle for beta features. Per-feature key values override the default set by AllBeta.\n#  AllBeta: false\n\n# Enable AntreaProxy which provides ServiceLB for in-cluster Services in antrea-agent.\n# It should be enabled on Windows, otherwise NetworkPolicy will not take effect on\n# Service traffic.\n#  AntreaProxy: true\n\n# Enable EndpointSlice support in AntreaProxy. Don't enable this feature unless that EndpointSlice\n# API version v1 is supported and set as enabled in Kubernetes. If AntreaProxy is not enabled,\n# this flag will not take effect.\n#  EndpointSlice: false\n\n# Enable TopologyAwareHints in AntreaProxy. This requires AntreaProxy and EndpointSlice to be\n# enabled, otherwise this flag will not take effect.\n#  TopologyAwareHints: false\n\n# Enable traceflow which provides packet tracing feature to diagnose network issue.\n#  Traceflow: true\n\n# Enable NodePortLocal feature to make the Pods reachable externally through NodePort\n#  NodePortLocal: true\n\n# Enable Antrea ClusterNetworkPolicy feature to complement K8s NetworkPolicy for cluster admins\n# to define security policies which apply to the entire cluster, and Antrea NetworkPolicy\n# feature that supports priorities, rule actions and externalEntities in the future.\n#  AntreaPolicy: true\n\n# Enable flowexporter which exports polled conntrack connections as IPFIX flow records from each\n# agent to a configured collector.\n#  FlowExporter: false\n\n# Enable collecting and exposing NetworkPolicy statistics.\n#  NetworkPolicyStats: true\n\n# Enable controlling SNAT IPs of Pod egress traffic.\n#  Egress: true\n\n# Enable AntreaIPAM, which can allocate IP addresses from IPPools. AntreaIPAM is required by the\n# bridging mode and allocates IPs to Pods in bridging mode. It is also required to use Antrea for\n# IPAM when configuring secondary network interfaces with Multus.\n#  AntreaIPAM: false\n\n# Enable multicast traffic.\n#  Multicast: false\n\n# Enable Antrea Multi-cluster Gateway to support cross-cluster traffic.\n# This feature is supported only with encap mode.\n#  Multicluster: false\n\n# Enable support for provisioning secondary network interfaces for Pods (using\n# Pod annotations). At the moment, Antrea can only create secondary network\n# interfaces using SR-IOV VFs on baremetal Nodes.\n#  SecondaryNetwork: false\n\n# Enable managing external IPs of Services of LoadBalancer type.\n#  ServiceExternalIP: false\n\n# Enable mirroring or redirecting the traffic Pods send or receive.\n#  TrafficControl: false\n\n# Enable certificated-based authentication for IPsec.\n#  IPsecCertAuth: false\n\n# Enable collecting support bundle files with SupportBundleCollection CRD.\n#  SupportBundleCollection: false\n\n# Enable users to protect their applications by specifying how they are allowed to communicate with others, taking\n# into account application context.\n#  L7NetworkPolicy: false\n\n# Name of the OpenVSwitch bridge antrea-agent will create and use.\n# Make sure it doesn't conflict with your existing OpenVSwitch bridges.\novsBridge: \"br-int\"\n\n# Datapath type to use for the OpenVSwitch bridge created by Antrea. At the moment, the only\n# supported value is 'system', which corresponds to the kernel datapath.\n#ovsDatapathType: system\n\n# Name of the interface antrea-agent will create and use for host \u003c--\u003e pod communication.\n# Make sure it doesn't conflict with your existing interfaces.\nhostGateway: \"antrea-gw0\"\n\n# Determines how traffic is encapsulated. It has the following options:\n# encap(default):    Inter-node Pod traffic is always encapsulated and Pod to external network\n#                    traffic is SNAT'd.\n# noEncap:           Inter-node Pod traffic is not encapsulated; Pod to external network traffic is\n#                    SNAT'd if noSNAT is not set to true. Underlying network must be capable of\n#                    supporting Pod traffic across IP subnets.\n# hybrid:            noEncap if source and destination Nodes are on the same subnet, otherwise encap.\n# networkPolicyOnly: Antrea enforces NetworkPolicy only, and utilizes CNI chaining and delegates Pod\n#                    IPAM and connectivity to the primary CNI.\n#\ntrafficEncapMode: \"encap\"\n\n# Whether or not to SNAT (using the Node IP) the egress traffic from a Pod to the external network.\n# This option is for the noEncap traffic mode only, and the default value is false. In the noEncap\n# mode, if the cluster's Pod CIDR is reachable from the external network, then the Pod traffic to\n# the external network needs not be SNAT'd. In the networkPolicyOnly mode, antrea-agent never\n# performs SNAT and this option will be ignored; for other modes it must be set to false.\nnoSNAT: false\n\n# Tunnel protocols used for encapsulating traffic across Nodes. If WireGuard is enabled in trafficEncryptionMode,\n# this option will not take effect. Supported values:\n# - geneve (default)\n# - vxlan\n# - gre\n# - stt\n# Note that \"gre\" is not supported for IPv6 clusters (IPv6-only or dual-stack clusters).\ntunnelType: \"geneve\"\n\n# TunnelPort is the destination port for UDP and TCP based tunnel protocols (Geneve, VXLAN, and STT).\n# If zero, it will use the assigned IANA port for the protocol, i.e. 6081 for Geneve, 4789 for VXLAN,\n# and 7471 for STT.\ntunnelPort: 0\n\n# TunnelCsum determines whether to compute UDP encapsulation header (Geneve or VXLAN) checksums on outgoing\n# packets. For Linux kernel before Mar 2021, UDP checksum must be present to trigger GRO on the receiver for better\n# performance of Geneve and VXLAN tunnels. The issue has been fixed by\n# https://github.com/torvalds/linux/commit/89e5c58fc1e2857ccdaae506fb8bc5fed57ee063, thus computing UDP checksum is\n# no longer necessary.\n# It should only be set to true when you are using an unpatched Linux kernel and observing poor transfer performance.\ntunnelCsum: false\n\n# Determines how tunnel traffic is encrypted. Currently encryption only works with encap mode.\n# It has the following options:\n# - none (default):  Inter-node Pod traffic will not be encrypted.\n# - ipsec:           Enable IPsec (ESP) encryption for Pod traffic across Nodes. Antrea uses\n#                    Preshared Key (PSK) for IKE authentication. When IPsec tunnel is enabled,\n#                    the PSK value must be passed to Antrea Agent through an environment\n#                    variable: ANTREA_IPSEC_PSK.\n# - wireGuard:       Enable WireGuard for tunnel traffic encryption.\ntrafficEncryptionMode: \"none\"\n\n# Enable bridging mode of Pod network on Nodes, in which the Node's transport interface is connected\n# to the OVS bridge, and cross-Node/VLAN traffic of AntreaIPAM Pods (Pods whose IP addresses are\n# allocated by AntreaIPAM from IPPools) is sent to the underlay network, and forwarded/routed by the\n# underlay network.\n# This option requires the `AntreaIPAM` feature gate to be enabled. At this moment, it supports only\n# IPv4 and Linux Nodes, and can be enabled only when `ovsDatapathType` is `system`,\n# `trafficEncapMode` is `noEncap`, and `noSNAT` is true.\nenableBridgingMode: false\n\n# Disable TX checksum offloading for container network interfaces. It's supposed to be set to true when the\n# datapath doesn't support TX checksum offloading, which causes packets to be dropped due to bad checksum.\n# It affects Pods running on Linux Nodes only.\ndisableTXChecksumOffload: false\n\n# Default MTU to use for the host gateway interface and the network interface of each Pod.\n# If omitted, antrea-agent will discover the MTU of the Node's primary interface and\n# also adjust MTU to accommodate for tunnel encapsulation overhead (if applicable).\ndefaultMTU: 0\n\n# wireGuard specifies WireGuard related configurations.\nwireGuard:\n  # The port for WireGuard to receive traffic.\n  port: 51820\n\negress:\n  # exceptCIDRs is the CIDR ranges to which outbound Pod traffic will not be SNAT'd by Egresses.\n  exceptCIDRs:\n\n# ClusterIP CIDR range for Services. It's required when AntreaProxy is not enabled, and should be\n# set to the same value as the one specified by --service-cluster-ip-range for kube-apiserver. When\n# AntreaProxy is enabled, this parameter is not needed and will be ignored if provided.\nserviceCIDR: \"\"\n\n# ClusterIP CIDR range for IPv6 Services. It's required when using kube-proxy to provide IPv6 Service in a Dual-Stack\n# cluster or an IPv6 only cluster. The value should be the same as the configuration for kube-apiserver specified by\n# --service-cluster-ip-range. When AntreaProxy is enabled, this parameter is not needed.\n# No default value for this field.\nserviceCIDRv6: \"\"\n\n# The port for the antrea-agent APIServer to serve on.\n# Note that if it's set to another value, the `containerPort` of the `api` port of the\n# `antrea-agent` container must be set to the same value.\napiPort: 10350\n\n# Enable metrics exposure via Prometheus. Initializes Prometheus metrics listener.\nenablePrometheusMetrics: true\n\n# Provide the IPFIX collector address as a string with format \u003cHOST\u003e:[\u003cPORT\u003e][:\u003cPROTO\u003e].\n# HOST can either be the DNS name or the IP of the Flow Collector. For example,\n# \"flow-aggregator.flow-aggregator.svc\" can be provided as DNS name to connect\n# to the Antrea Flow Aggregator service. If IP, it can be either IPv4 or IPv6.\n# However, IPv6 address should be wrapped with [].\n# If PORT is empty, we default to 4739, the standard IPFIX port.\n# If no PROTO is given, we consider \"tls\" as default. We support \"tls\", \"tcp\" and\n# \"udp\" protocols. \"tls\" is used for securing communication between flow exporter and\n# flow aggregator.\nflowCollectorAddr: \"flow-aggregator.flow-aggregator.svc:4739:tls\"\n\n# Provide flow poll interval as a duration string. This determines how often the\n# flow exporter dumps connections from the conntrack module. Flow poll interval\n# should be greater than or equal to 1s (one second).\n# Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\", \"m\", \"h\".\nflowPollInterval: \"5s\"\n\n# Provide the active flow export timeout, which is the timeout after which a flow\n# record is sent to the collector for active flows. Thus, for flows with a continuous\n# stream of packets, a flow record will be exported to the collector once the elapsed\n# time since the last export event is equal to the value of this timeout.\n# Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\", \"m\", \"h\".\nactiveFlowExportTimeout: \"5s\"\n\n# Provide the idle flow export timeout, which is the timeout after which a flow\n# record is sent to the collector for idle flows. A flow is considered idle if no\n# packet matching this flow has been observed since the last export event.\n# Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\", \"m\", \"h\".\nidleFlowExportTimeout: \"15s\"\n\nnodePortLocal:\n# Enable NodePortLocal, a feature used to make Pods reachable using port forwarding on the host. To\n# enable this feature, you need to set \"enable\" to true, and ensure that the NodePortLocal feature\n# gate is also enabled (which is the default).\n  enable: false\n# Provide the port range used by NodePortLocal. When the NodePortLocal feature is enabled, a port\n# from that range will be assigned whenever a Pod's container defines a specific port to be exposed\n# (each container can define a list of ports as pod.spec.containers[].ports), and all Node traffic\n# directed to that port will be forwarded to the Pod.\n  portRange: \"61000-62000\"\n\n# Provide the address of Kubernetes apiserver, to override any value provided in kubeconfig or InClusterConfig.\n# Defaults to \"\". It must be a host string, a host:port pair, or a URL to the base of the apiserver.\nkubeAPIServerOverride: \"\"\n\n# Provide the address of DNS server, to override the kube-dns service. It's used to resolve hostname in FQDN policy.\n# Defaults to \"\". It must be a host string or a host:port pair of the DNS server (e.g. 10.96.0.10, 10.96.0.10:53,\n# [fd00:10:96::a]:53).\ndnsServerOverride: \"\"\n\n# Comma-separated list of Cipher Suites. If omitted, the default Go Cipher Suites will be used.\n# https://golang.org/pkg/crypto/tls/#pkg-constants\n# Note that TLS1.3 Cipher Suites cannot be added to the list. But the apiserver will always\n# prefer TLS1.3 Cipher Suites whenever possible.\ntlsCipherSuites: \"\"\n\n# TLS min version from: VersionTLS10, VersionTLS11, VersionTLS12, VersionTLS13.\ntlsMinVersion: \"\"\n\n# The name of the interface on Node which is used for tunneling or routing the traffic across Nodes.\n# If there are multiple IP addresses configured on the interface, the first one is used. The IP\n# address used for tunneling or routing traffic to remote Nodes is decided in the following order of\n# preference (from highest to lowest):\n# 1. transportInterface\n# 2. transportInterfaceCIDRs\n# 3. The Node IP\ntransportInterface: \"\"\n\nmulticast:\n# The names of the interfaces on Nodes that are used to forward multicast traffic.\n# Defaults to transport interface if not set.\n  multicastInterfaces:\n\n# The interval at which the antrea-agent sends IGMP queries to Pods.\n# Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\", \"m\", \"h\".\n  igmpQueryInterval: \"125s\"\n\n# The network CIDRs of the interface on Node which is used for tunneling or routing the traffic across\n# Nodes. If there are multiple interfaces configured the same network CIDR, the first one is used. The\n# IP address used for tunneling or routing traffic to remote Nodes is decided in the following order of\n# preference (from highest to lowest):\n# 1. transportInterface\n# 2. transportInterfaceCIDRs\n# 3. The Node IP\ntransportInterfaceCIDRs:\n\n# Option antreaProxy contains AntreaProxy related configuration options.\nantreaProxy:\n  # ProxyAll tells antrea-agent to proxy all Service traffic, including NodePort, LoadBalancer, and ClusterIP traffic,\n  # regardless of where they come from. Therefore, running kube-proxy is no longer required. This requires the AntreaProxy\n  # feature to be enabled.\n  # Note that this option is experimental. If kube-proxy is removed, option kubeAPIServerOverride must be used to access\n  # apiserver directly.\n  proxyAll: false\n  # A string array of values which specifies the host IPv4/IPv6 addresses for NodePort. Values can be valid IP blocks.\n  # (e.g. 1.2.3.0/24, 1.2.3.4/32). An empty string slice is meant to select all host IPv4/IPv6 addresses.\n  # Note that the option is only valid when proxyAll is true.\n  nodePortAddresses:\n  # An array of string values to specify a list of Services which should be ignored by AntreaProxy (traffic to these\n  # Services will not be load-balanced). Values can be a valid ClusterIP (e.g. 10.11.1.2) or a Service name\n  # with Namespace (e.g. kube-system/kube-dns)\n  skipServices:\n  # When ProxyLoadBalancerIPs is set to false, AntreaProxy no longer load-balances traffic destined to the\n  # External IPs of LoadBalancer Services. This is useful when the external LoadBalancer provides additional\n  # capabilities (e.g. TLS termination) and it is desirable for Pod-to-ExternalIP traffic to be sent to the\n  # external LoadBalancer instead of being load-balanced to an Endpoint directly by AntreaProxy.\n  # Note that setting ProxyLoadBalancerIPs to false usually only makes sense when ProxyAll is set to true and\n  # kube-proxy is removed from the cluser, otherwise kube-proxy will still load-balance this traffic.\n  proxyLoadBalancerIPs: true\n\n# IPsec tunnel related configurations.\nipsec:\n  # The authentication mode of IPsec tunnel. It has the following options:\n  # - psk (default): Use pre-shared key (PSK) for IKE authentication.\n  # - cert:          Use CA-signed certificates for IKE authentication. This option requires the `IPsecCertAuth`\n  #                  feature gate to be enabled.\n  authenticationMode: \"psk\"\n\nmulticluster:\n# Enable Antrea Multi-cluster Gateway to support cross-cluster traffic.\n# This feature is supported only with encap mode.\n  enable: false\n# The Namespace where Antrea Multi-cluster Controller is running.\n# The default is antrea-agent's Namespace.\n  namespace: \"\"\n# Enable StretchedNetworkPolicy which could be enforced on cross-cluster traffic.\n# This feature is supported only with encap mode.\n  enableStretchedNetworkPolicy: false\n",
                  "antreaCNIConfig": "{\n    \"cniVersion\":\"0.3.0\",\n    \"name\": \"antrea\",\n    \"plugins\": [\n        {\n            \"type\": \"antrea\",\n            \"ipam\": {\n                \"type\": \"host-local\"\n            }\n        }\n        ,\n        {\n            \"type\": \"portmap\",\n            \"capabilities\": {\"portMappings\": true}\n        }\n        ,\n        {\n            \"type\": \"bandwidth\",\n            \"capabilities\": {\"bandwidth\": true}\n        }\n    ]\n}\n",
                  "antreaControllerConfig": "# FeatureGates is a map of feature names to bools that enable or disable experimental features.\nfeatureGates:\n# AllAlpha is a global toggle for alpha features. Per-feature key values override the default set by AllAlpha.\n#  AllAlpha: false\n\n# AllBeta is a global toggle for beta features. Per-feature key values override the default set by AllBeta.\n#  AllBeta: false\n\n# Enable traceflow which provides packet tracing feature to diagnose network issue.\n#  Traceflow: true\n\n# Enable Antrea ClusterNetworkPolicy feature to complement K8s NetworkPolicy for cluster admins\n# to define security policies which apply to the entire cluster, and Antrea NetworkPolicy\n# feature that supports priorities, rule actions and externalEntities in the future.\n#  AntreaPolicy: true\n\n# Enable collecting and exposing NetworkPolicy statistics.\n#  NetworkPolicyStats: true\n\n# Enable multicast traffic.\n#  Multicast: false\n\n# Enable controlling SNAT IPs of Pod egress traffic.\n#  Egress: true\n\n# Run Kubernetes NodeIPAMController with Antrea.\n#  NodeIPAM: false\n\n# Enable AntreaIPAM, which can allocate IP addresses from IPPools. AntreaIPAM is required by the\n# bridging mode and allocates IPs to Pods in bridging mode. It is also required to use Antrea for\n# IPAM when configuring secondary network interfaces with Multus.\n#  AntreaIPAM: false\n\n# Enable managing external IPs of Services of LoadBalancer type.\n#  ServiceExternalIP: false\n\n# Enable certificated-based authentication for IPsec.\n#  IPsecCertAuth: false\n\n# Enable managing ExternalNode for unmanaged VM/BM.\n#  ExternalNode: false\n\n# Enable collecting support bundle files with SupportBundleCollection CRD.\n#  SupportBundleCollection: false\n\n# Enable multi-cluster features.\n#  Multicluster: false\n\n# Enable users to protect their applications by specifying how they are allowed to communicate with others, taking\n# into account application context.\n#  L7NetworkPolicy: false\n\n# The port for the antrea-controller APIServer to serve on.\n# Note that if it's set to another value, the `containerPort` of the `api` port of the\n# `antrea-controller` container must be set to the same value.\napiPort: 10349\n\n# Enable metrics exposure via Prometheus. Initializes Prometheus metrics listener.\nenablePrometheusMetrics: true\n\n# Indicates whether to use auto-generated self-signed TLS certificate.\n# If false, a Secret named \"antrea-controller-tls\" must be provided with the following keys:\n#   ca.crt: \u003cCA certificate\u003e\n#   tls.crt: \u003cTLS certificate\u003e\n#   tls.key: \u003cTLS private key\u003e\nselfSignedCert: true\n\n# Comma-separated list of Cipher Suites. If omitted, the default Go Cipher Suites will be used.\n# https://golang.org/pkg/crypto/tls/#pkg-constants\n# Note that TLS1.3 Cipher Suites cannot be added to the list. But the apiserver will always\n# prefer TLS1.3 Cipher Suites whenever possible.\ntlsCipherSuites: \"\"\n\n# TLS min version from: VersionTLS10, VersionTLS11, VersionTLS12, VersionTLS13.\ntlsMinVersion: \"\"\n\nnodeIPAM:\n  # Enable the integrated Node IPAM controller within the Antrea controller.\n  enableNodeIPAM: false\n  # CIDR ranges for Pods in cluster. String array containing single CIDR range, or multiple ranges.\n  # The CIDRs could be either IPv4 or IPv6. At most one CIDR may be specified for each IP family.\n  # Value ignored when enableNodeIPAM is false.\n  clusterCIDRs:\n  # CIDR ranges for Services in cluster. It is not necessary to specify it when there is no overlap with clusterCIDRs.\n  # Value ignored when enableNodeIPAM is false.\n  serviceCIDR: \"\"\n  serviceCIDRv6: \"\"\n  # Mask size for IPv4 Node CIDR in IPv4 or dual-stack cluster. Value ignored when enableNodeIPAM is false\n  # or when IPv4 Pod CIDR is not configured. Valid range is 16 to 30.\n  nodeCIDRMaskSizeIPv4: 24\n  # Mask size for IPv6 Node CIDR in IPv6 or dual-stack cluster. Value ignored when enableNodeIPAM is false\n  # or when IPv6 Pod CIDR is not configured. Valid range is 64 to 126.\n  nodeCIDRMaskSizeIPv6: 64\n\nipsecCSRSigner:\n  # Determines the auto-approve policy of Antrea CSR signer for IPsec certificates management.\n  # If enabled, Antrea will auto-approve the CertificateSingingRequest (CSR) if its subject and x509 extensions\n  # are permitted, and the requestor can be validated. If K8s `BoundServiceAccountTokenVolume` feature is enabled,\n  # the Pod identity will also be validated to provide maximum security.\n  # If set to false, Antrea will not auto-approve CertificateSingingRequests and they need to be approved\n  # manually by `kubectl certificate approve`.\n  autoApprove: true\n  # Indicates whether to use auto-generated self-signed CA certificate.\n  # If false, a Secret named \"antrea-ipsec-ca\" must be provided with the following keys:\n  #   tls.crt: \u003cCA certificate\u003e\n  #   tls.key: \u003cCA private key\u003e\n  selfSignedCA: true\n\nmulticluster:\n  # Enable StretchedNetworkPolicy which allow Antrea-native policies to select peers\n  # from other clusters in a ClusterSet.\n  enableStretchedNetworkPolicy: false\n",
                  "antreaImage": "antrea/antrea-ubi:v2.1.0",
                  "antreaPlatform": "openshift"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: projects.packages.broadcom.com/antreainterworking/antrea-operator:v2.1.0_vmware.3@sha256:70f129499c801f164adf8772e7ccd5d1fc659df5f298d7851e6bf02ee760f2a9
          createdAt: "2025-04-16T12:00:00Z"
          description: An operator which installs Antrea network CNI plugin on the
            Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          support: projectantrea-maintainers@googlegroups.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AntreaInstall is the Schema for the antreainstalls API
            displayName: Antrea Install
            kind: AntreaInstall
            name: antreainstalls.operator.antrea.vmware.com
            version: v1
        description: An operator which installs Antrea network CNI plugin on the Kubernetes
          cluster.
        displayName: Antrea-by-Broadcom Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        links:
        - name: Antrea Operator For Kubernetes
          url: https://github.com/vmware/antrea-operator-for-kubernetes
        maintainers:
        - email: projectantrea-maintainers@googlegroups.com
          name: Project Antrea Maintainers
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: antrea.io
        relatedImages:
        - projects.packages.broadcom.com/antreainterworking/antrea-operator:v2.1.0_vmware.3@sha256:70f129499c801f164adf8772e7ccd5d1fc659df5f298d7851e6bf02ee760f2a9
        version: 1.10.0
      entries:
      - name: antrea-by-broadcom-operator.v1.10.0
        version: 1.10.0
      name: alpha
    defaultChannel: alpha
    packageName: antrea-by-broadcom-operator
    provider:
      name: antrea.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Nokia
      provider-url: https://docs.eda.dev
    name: eda-k8s-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eda-openshift-operator.v4.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.eda.nokia.com/v1",
                "kind": "ConnectControllerConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "operator"
                  },
                  "name": "connectcontrollerconfig-sample",
                  "namespace": "eda-connect-k8s"
                },
                "spec": {
                  "connectPluginName": "",
                  "controllerImagePullSecret": "",
                  "edaNamespace": "eda",
                  "heartBeat": "",
                  "logLevel": "",
                  "skipTLSVerify": false,
                  "tlsCertificateData": "",
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:c43a7fa7475db83011cc08b728982511fc0279a259e1aacc8ddde9e15f947725
          createdAt: "2025-06-10T15:47:27Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: eda-connect-k8s
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["connectnetworkattachmentstatuses.connect.eda.nokia.com",
            connectnetworkdefinitions.connect.eda.nokia.com]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConnectNetworkAttachmentStatus is the Schema for the connectnetworkattachmentstatuses
              API
            displayName: Connect NetworkAttachment Status
            kind: ConnectNetworkAttachmentStatus
            name: connectnetworkattachmentstatuses.connect.eda.nokia.com
            version: v1
          - description: ConnectNetworkDefinition is the Schema for the connectnetworkdefinitions
              API
            displayName: Connect Network Definition
            kind: ConnectNetworkDefinition
            name: connectnetworkdefinitions.connect.eda.nokia.com
            version: v1
          - description: ConnectControllerConfig is the Schema for the connectcontrollerconfigs
              API
            displayName: Connect Controller Config
            kind: ConnectControllerConfig
            name: connectcontrollerconfigs.config.eda.nokia.com
            version: v1
        description: "K8s operator to deploy eda connect controller to configure EDA
          Connect.\eb"
        displayName: eda-openshift-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - EDA Connect
        - K8s Controller
        - EDA K8s Controller Operator
        links:
        - name: EDA Connect K8s Controller Operator
          url: https://docs.eda.dev/connect/openshift-plugin/
        maintainers:
        - email: support@nokia.com
          name: nokia
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Nokia
          url: https://docs.eda.dev
        relatedImages:
        - ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:c43a7fa7475db83011cc08b728982511fc0279a259e1aacc8ddde9e15f947725
        version: 4.0.0
      entries:
      - name: eda-openshift-operator.v4.0.0
        version: 4.0.0
      name: alpha
    - currentCSV: eda-openshift-operator.v4.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.eda.nokia.com/v1",
                "kind": "ConnectControllerConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "operator"
                  },
                  "name": "connectcontrollerconfig-sample",
                  "namespace": "eda-connect-k8s"
                },
                "spec": {
                  "connectPluginName": "",
                  "controllerImagePullSecret": "",
                  "edaNamespace": "eda",
                  "heartBeat": "",
                  "logLevel": "",
                  "skipTLSVerify": false,
                  "tlsCertificateData": "",
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:ec264b52d581b8fd76e737b9811ab894957a4edf2f627f90bd9fe707890e6bf5
          createdAt: "2025-06-10T15:47:27Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: eda-connect-k8s
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["connectnetworkattachmentstatuses.connect.eda.nokia.com",
            connectnetworkdefinitions.connect.eda.nokia.com]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConnectNetworkAttachmentStatus is the Schema for the connectnetworkattachmentstatuses
              API
            displayName: Connect NetworkAttachment Status
            kind: ConnectNetworkAttachmentStatus
            name: connectnetworkattachmentstatuses.connect.eda.nokia.com
            version: v1
          - description: ConnectNetworkDefinition is the Schema for the connectnetworkdefinitions
              API
            displayName: Connect Network Definition
            kind: ConnectNetworkDefinition
            name: connectnetworkdefinitions.connect.eda.nokia.com
            version: v1
          - description: ConnectControllerConfig is the Schema for the connectcontrollerconfigs
              API
            displayName: Connect Controller Config
            kind: ConnectControllerConfig
            name: connectcontrollerconfigs.config.eda.nokia.com
            version: v1
        description: "K8s operator to deploy eda connect controller to configure EDA
          Connect.\eb"
        displayName: eda-openshift-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - EDA Connect
        - K8s Controller
        - EDA K8s Controller Operator
        links:
        - name: EDA Connect K8s Controller Operator
          url: https://docs.eda.dev/connect/openshift-plugin/
        maintainers:
        - email: support@nokia.com
          name: nokia
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Nokia
          url: https://docs.eda.dev
        relatedImages:
        - ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:ec264b52d581b8fd76e737b9811ab894957a4edf2f627f90bd9fe707890e6bf5
        version: 4.0.1
      entries:
      - name: eda-openshift-operator.v4.0.1
        version: 4.0.1
      name: stable
    defaultChannel: alpha
    packageName: eda-k8s-operator
    provider:
      name: Nokia
      url: https://docs.eda.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: prometurbo-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometurbo-operator.v8.17.40-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5-beta.1 <8.17.40-beta.1'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
          - kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.

          ### Get Started
          Please follow [this](https://www.ibm.com/docs/en/tarm/8.15.x?topic=configuration-prometheus) for an overview of Prometurbo and deployment instructions.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/prometurbo
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/prometurbo-operator@sha256:a1605506884f6bce7471d79c6243728bce991ceb381d7fe0c51f1cdcbeb96e22
        version: 8.17.40-beta.1
      entries:
      - name: prometurbo-operator.v8.17.40-beta.1
        version: 8.17.40-beta.1
      name: beta
    - currentCSV: prometurbo-operator.v8.17.40
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5 <8.17.40'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
          - kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.

          ### Get Started
          Please follow [this](https://www.ibm.com/docs/en/tarm/8.15.x?topic=configuration-prometheus) for an overview of Prometurbo and deployment instructions.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/prometurbo
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/prometurbo-operator@sha256:cbcf90e44ffcedec8e3f161eff4c95b581924a1db195bf55d13da3714346b26f
        version: 8.17.40
      entries:
      - name: prometurbo-operator.v8.17.40
        version: 8.17.40
      name: stable
    defaultChannel: stable
    packageName: prometurbo-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: AppDynamics LLC
      provider-url: https://www.appdynamics.com/
    name: appdynamics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appdynamics-operator.v24.7.0-861
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clusteragent",
                "metadata": {
                  "name": "k8s-cluster-agent",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "appName": "<app-name>",
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/cluster-agent:tag"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "InfraViz",
                "metadata": {
                  "name": "appdynamics-infraviz",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/machine-agent:latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3c420cce8b626d8f5cf424410527c6a7a17ca3348e903dd8cdfcd74819b8de41
          createdAt: "2023-02-06T14:26:07Z"
          description: AppDynamics Operator to install AppDynamics Cluster Agent &
            AppDynamics Machine Agent for K8s Monitoring for various Cloud Providers
            like EKS, AKS etc.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: support@appdynamics.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clusteragent is the Schema for the clusteragents API
            displayName: Clusteragent
            kind: Clusteragent
            name: clusteragents.cluster.appdynamics.com
            version: v1alpha1
          - description: InfraViz is the Schema for the infravizs API
            displayName: Infra Viz
            kind: InfraViz
            name: infravizs.cluster.appdynamics.com
            version: v1alpha1
        description: |-
          AppDynamics Operator to install AppDynamics Cluster Agent & Machine Agent for K8s Monitoring on various Cloud Providers like EKS, AKS, GCP, Openshift etc.
          **About AppDynamics Operator**
          AppDynamics Operator simplifies the installation and lifecycle management of the AppDynamics ClusterAgent and the AppDynamics Machine Agent for K8s Monitoring on different Cloud Providers like EKS, AKS, GCP, OpenShift etc.
          **Installation Steps**
          Please refer below links for installation guide:
          [Install AppDynamics Cluster Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli)
          [Install AppDynamics Machine Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli)
          [Install AppDynamics Cluster Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts)
          [Install Appdynamics Machine Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts)
        displayName: AppDynamics Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AppDynamics
        - Operator
        - AppDynamics Cluster Agent
        - AppDynamics Machine Agent
        - AppDynamics K8s Monitoring
        links:
        - name: AppDynamics Kubernetes Monitoring Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent
        - name: AppDynamics Cluster Agent Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/overview-of-cluster-monitoring
        - name: AppDynamics Cluster Agent Requirements
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/cluster-agent-requirements-and-supported-environments
        - name: Install AppDynamics Cluster Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli
        - name: Install AppDynamics Cluster Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts
        - name: Install AppDynamics Machine Agent & NetViz Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli
        - name: Install AppDynamics Machine Agent & NetViz Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: AppDynamics LLC
          url: https://www.appdynamics.com/
        relatedImages:
        - registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3c420cce8b626d8f5cf424410527c6a7a17ca3348e903dd8cdfcd74819b8de41
        version: 24.7.0-861
      entries:
      - name: appdynamics-operator.v24.7.0-861
        version: 24.7.0-861
      name: alpha
    - currentCSV: appdynamics-operator.v25.9.0-1218
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clusteragent",
                "metadata": {
                  "name": "k8s-cluster-agent",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "appName": "<app-name>",
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/cluster-agent:tag"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "InfraViz",
                "metadata": {
                  "name": "appdynamics-infraviz",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/machine-agent:latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3320c2c8d5f60adc541befd099dac1ef96257b16e0df28b727b0e589f78d981a
          createdAt: "2023-02-06T14:26:07Z"
          description: AppDynamics Operator to install AppDynamics Cluster Agent &
            AppDynamics Machine Agent for K8s Monitoring for various Cloud Providers
            like EKS, AKS etc.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: support@appdynamics.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clusteragent is the Schema for the clusteragents API
            displayName: Clusteragent
            kind: Clusteragent
            name: clusteragents.cluster.appdynamics.com
            version: v1alpha1
          - description: InfraViz is the Schema for the infravizs API
            displayName: Infra Viz
            kind: InfraViz
            name: infravizs.cluster.appdynamics.com
            version: v1alpha1
        description: |-
          AppDynamics Operator to install AppDynamics Cluster Agent & Machine Agent for K8s Monitoring on various Cloud Providers like EKS, AKS, GCP, Openshift etc.
          **About AppDynamics Operator**
          AppDynamics Operator simplifies the installation and lifecycle management of the AppDynamics ClusterAgent and the AppDynamics Machine Agent for K8s Monitoring on different Cloud Providers like EKS, AKS, GCP, OpenShift etc.
          **Installation Steps**
          Please refer below links for installation guide:
          [Install AppDynamics Cluster Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli)
          [Install AppDynamics Machine Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli)
          [Install AppDynamics Cluster Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts)
          [Install Appdynamics Machine Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts)
        displayName: AppDynamics Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - AppDynamics
        - Operator
        - AppDynamics Cluster Agent
        - AppDynamics Machine Agent
        - AppDynamics K8s Monitoring
        links:
        - name: AppDynamics Kubernetes Monitoring Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent
        - name: AppDynamics Cluster Agent Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/overview-of-cluster-monitoring
        - name: AppDynamics Cluster Agent Requirements
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/cluster-agent-requirements-and-supported-environments
        - name: Install AppDynamics Cluster Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli
        - name: Install AppDynamics Cluster Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts
        - name: Install AppDynamics Machine Agent & NetViz Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli
        - name: Install AppDynamics Machine Agent & NetViz Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: AppDynamics LLC
          url: https://www.appdynamics.com/
        relatedImages:
        - registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3320c2c8d5f60adc541befd099dac1ef96257b16e0df28b727b0e589f78d981a
        version: 25.9.0-1218
      entries:
      - name: appdynamics-operator.v25.9.0-1218
        version: 25.9.0-1218
      name: stable
    defaultChannel: alpha
    packageName: appdynamics-operator
    provider:
      name: AppDynamics LLC
      url: https://www.appdynamics.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: www.redhat.com
    name: odf-node-recovery-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-node-recovery-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "NodeRecovery",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "odf-node-recovery-operator",
                    "app.kubernetes.io/instance": "noderecovery-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "noderecovery",
                    "app.kubernetes.io/part-of": "odf-node-recovery-operator"
                  },
                  "name": "noderecovery-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Storage,OpenShift Optional
          description: ODF Node Recovery is an operator that assist in the recovery
            of and ODF cluster that has a device or a node replaced
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/jordigilh/odf-node-recovery-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeRecovery is the Schema for the noderecoveries API
            displayName: Node Recovery
            kind: NodeRecovery
            name: noderecoveries.odf.openshift.io
            version: v1alpha1
        description: |
          ODF Node Recovery is an operator that assist in the recovery of an
          ODF cluster in these 2 scenarios:
          * <a href="https://docs.redhat.com/en/documentation/red_hat_openshift_data_foundation/4.14/html-single/replacing_devices/index"> Device replacement<a/>
          * <a href="https://docs.redhat.com/en/documentation/red_hat_openshift_data_foundation/4.14/html-single/replacing_nodes/index"> Node replacement</a>
        displayName: ODF Node Recovery Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - node recovery
        links:
        - name: GitHub Repository
          url: https://github.com/jordigilh/odf-node-recovery-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
          url: www.redhat.com
        relatedImages:
        - quay.io/jordigilh/odf-node-recovery-controller@sha256:4164fafa9bb1c4d6121f4c5fc700fa2827c3994ca98a65d641471a2a46d4f59c
        version: 1.1.0
      entries:
      - name: odf-node-recovery-operator.v1.1.0
        version: 1.1.0
      - name: odf-node-recovery-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: odf-node-recovery-operator
    provider:
      name: Red Hat, Inc.
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-prometheus-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-prometheus-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:b6071d6a1ec7cdd5f65b6d7296c1e9cd5660972a8ed614da4277e3a4ab5d1df1
          createdAt: "2024-03-01T23:43:00Z"
          description: Manage the full lifecycle of configuring and managing Prometheus
            and Alertmanager servers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/prometheus-operator/prometheus-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Prometheus Operator for Kubernetes provides easy monitoring definitions for Kubernetes services and deployment and management of Prometheus instances.

          Once installed, the Prometheus Operator provides the following features:

          * **Create/Destroy**: Easily launch a Prometheus instance for your Kubernetes namespace, a specific application or team easily using the Operator.
          * **Simple Configuration**: Configure the fundamentals of Prometheus like versions, persistence, retention policies, and replicas from a native Kubernetes resource.
          * **Target Services via Labels**: Automatically generate monitoring target configurations based on familiar Kubernetes label queries; no need to learn a Prometheus specific configuration language.

          ### Other Supported Features

          **High availability**

          Multiple instances are run across failure zones and data is replicated. This keeps your monitoring available during an outage, when you need it most.

          **Updates via automated operations**

          New Prometheus versions are deployed using a rolling update with no downtime, making it easy to stay up to date.

          **Handles the dynamic nature of containers**

          Alerting rules are attached to groups of containers instead of individual instances, which is ideal for the highly dynamic nature of container deployment.
        displayName: Prometheus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - prometheus
        - monitoring
        - tsdb
        - alerting
        - observability
        links:
        - name: Prometheus
          url: https://www.prometheus.io/
        - name: Documentation
          url: https://github.com/prometheus-operator/prometheus-operator/tree/master/Documentation
        - name: Prometheus Operator
          url: https://github.com/prometheus-operator/prometheus-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:b6071d6a1ec7cdd5f65b6d7296c1e9cd5660972a8ed614da4277e3a4ab5d1df1
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:413f7da5cb2e78abf33eb4e43f5fd483c19e791416eec7f5d4c3dbeb54db33a0
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader-rhel9@sha256:eb7614528884f2eefe1ae5e33a09624590c23e0e75510b4d88a274880cf090bc
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager-rhel9@sha256:7d1a3392872fd3990ae3acce79da659ba2238b30421824b7be533dcf8b93e89c
        version: 4.16.16-rhodf
      entries:
      - name: odf-prometheus-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odf-prometheus-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odf-prometheus-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odf-prometheus-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odf-prometheus-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odf-prometheus-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odf-prometheus-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odf-prometheus-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odf-prometheus-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odf-prometheus-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odf-prometheus-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odf-prometheus-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odf-prometheus-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odf-prometheus-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odf-prometheus-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odf-prometheus-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odf-prometheus-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odf-prometheus-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solace Corporation
      provider-url: www.solace.com
    name: pubsubplus-eventbroker-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pubsubplus-eventbroker-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pubsubplus.solace.com/v1beta1",
                "kind": "PubSubPlusEventBroker",
                "metadata": {
                  "name": "non-ha-standalone-example"
                },
                "spec": {
                  "redundancy": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "true"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.10
          containerImage: registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator@sha256:d68db92bcb217d94dc0ad0a1cea625a0a9c9299d886d325a903163f9f887dcd5
          createdAt: "2025-07-31T19:48:36Z"
          description: The Solace PubSub+ Event Broker Operator deploys and manages
            the lifecycle of PubSub+ Event Brokers
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
          support: Solace Products
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PubSub+ Event Broker
            displayName: PubSub+ Event Broker
            kind: PubSubPlusEventBroker
            name: pubsubpluseventbrokers.pubsubplus.solace.com
            version: v1beta1
        description: |
          ## Solace PubSub+ Platform
          Solace [PubSub+ Platform](https://solace.com/products/platform/) is a complete event streaming and management platform
          for real-time enterprises. The [PubSub+ Software Event Broker](https://solace.com/products/event-broker/software/) efficiently streams
          event-driven information between applications, IoT devices, and user interfaces running in the cloud, on-premises,
          and in hybrid environments using open APIs and protocols like AMQP, JMS, MQTT, REST and WebSocket.
          It can be installed into a variety of public and private clouds, PaaS, and on-premises environments.
          Event brokers in multiple locations can be linked together in an [Event Mesh](https://solace.com/what-is-an-event-mesh/)
          to dynamically share events across the distributed enterprise.

          ## Solace PubSub+ Event Broker Operator
          Solace Pubsub+ Event Broker Operator automatically deploys and manages PubSub+ Software Event Brokers on Kubernetes and OpenShift environments.

          Features include:
          * Initial deployment configuration: Production-ready HA, non-HA or for developers
          * Rolling update for configuration changes or broker upgrades
          * Prometheus monitoring support

          ## Getting started
          Refer to the [Solace PubSub+ Event Broker Operator Quick Start](https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/README.md)
        displayName: Solace PubSub+ Event Broker Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - solace
        - pubsubplus
        - pubsub+
        - pubsub
        - messaging
        - advanced event broker
        - event broker
        - event mesh
        - event streaming
        - data streaming
        - event integration
        - middleware
        links:
        - name: Operator QuickStart
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
        - name: Operator Documentation
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/docs/EventBrokerOperatorUserGuide.md
        - name: Solace.com
          url: https://solace.com
        - name: Understanding Solace Technology
          url: https://solace.com/products/platform/
        - name: Solace PubSub+ Technical Documentation
          url: https://docs.solace.com/
        - name: Solace Developer Portal
          url: https://www.solace.dev/
        - name: Solace Community
          url: https://solace.community/
        maintainers:
        - email: support@solace.com
          name: Solace R&D
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Solace Corporation
          url: www.solace.com
        relatedImages:
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator@sha256:d68db92bcb217d94dc0ad0a1cea625a0a9c9299d886d325a903163f9f887dcd5
        version: 1.4.0
      entries:
      - name: pubsubplus-eventbroker-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: pubsubplus-eventbroker-operator
    provider:
      name: Solace Corporation
      url: www.solace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: StreamNative
      provider-url: https://streamnative.io
    name: function-mesh
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: function-mesh.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "java-function-sample",
                  "namespace": "default"
                },
                "spec": {
                  "autoAck": true,
                  "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                  "clusterName": "test-pulsar",
                  "forwardSourceMessageProperty": true,
                  "input": {
                    "topics": [
                      "persistent://public/default/java-function-input-topic"
                    ],
                    "typeClassName": "java.lang.String"
                  },
                  "java": {
                    "extraDependenciesDir": "random-dir/",
                    "jar": "pulsar-functions-api-examples.jar",
                    "jarLocation": "public/default/nlu-test-java-function"
                  },
                  "logTopic": "persistent://public/default/logging-function-logs",
                  "maxPendingAsyncRequests": 1000,
                  "maxReplicas": 5,
                  "output": {
                    "topic": "persistent://public/default/java-function-output-topic",
                    "typeClassName": "java.lang.String"
                  },
                  "pod": {
                    "annotations": {
                      "managed-function": "true"
                    },
                    "env": [
                      {
                        "name": "EXAMPLE_VARIABLE",
                        "value": "exampleValue"
                      }
                    ],
                    "imagePullSecrets": [
                      {
                        "name": "regcred"
                      }
                    ],
                    "initContainers": [
                      {
                        "command": [
                          "sh",
                          "-c",
                          "echo The app is running! \u0026\u0026 sleep 30"
                        ],
                        "image": "busybox:1.28",
                        "name": "init-function"
                      }
                    ],
                    "labels": {
                      "locaction": "mtv"
                    },
                    "sidecars": [
                      {
                        "command": [
                          "sh",
                          "-c",
                          "echo The app is running! \u0026\u0026 sleep 30000"
                        ],
                        "image": "busybox:1.28",
                        "name": "sidecar-function"
                      }
                    ],
                    "volumes": [
                      {
                        "emptyDir": {},
                        "name": "cache-volume"
                      }
                    ]
                  },
                  "pulsar": {
                    "authSecret": "test-auth",
                    "pulsarConfig": "test-pulsar",
                    "tlsSecret": "test-tls"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "secretsMap": {
                    "name": {
                      "key": "username",
                      "path": "test-secret"
                    },
                    "pwd": {
                      "key": "password",
                      "path": "test-secret"
                    }
                  },
                  "volumeMounts": [
                    {
                      "mountPath": "/cache",
                      "name": "cache-volume"
                    }
                  ]
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "FunctionMesh",
                "metadata": {
                  "name": "functionmesh-sample"
                },
                "spec": {
                  "functions": [
                    {
                      "autoAck": true,
                      "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                      "clusterName": "test-pulsar",
                      "forwardSourceMessageProperty": true,
                      "input": {
                        "topics": [
                          "persistent://public/default/functionmesh-input-topic"
                        ],
                        "typeClassName": "java.lang.String"
                      },
                      "java": {
                        "jar": "pulsar-functions-api-examples.jar",
                        "jarLocation": "public/default/nlu-test-functionmesh-ex1"
                      },
                      "logTopic": "persistent://public/default/logging-function-log",
                      "maxReplicas": 1,
                      "name": "ex1",
                      "output": {
                        "topic": "persistent://public/default/mid-topic",
                        "typeClassName": "java.lang.String"
                      },
                      "pulsar": {
                        "pulsarConfig": "mesh-test-pulsar"
                      },
                      "replicas": 1,
                      "resources": {
                        "limits": {
                          "cpu": "0.2",
                          "memory": "1.1G"
                        },
                        "requests": {
                          "cpu": "0.1",
                          "memory": "1G"
                        }
                      }
                    },
                    {
                      "autoAck": true,
                      "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                      "clusterName": "test-pulsar",
                      "forwardSourceMessageProperty": true,
                      "input": {
                        "topics": [
                          "persistent://public/default/mid-topic"
                        ],
                        "typeClassName": "java.lang.String"
                      },
                      "java": {
                        "jar": "pulsar-functions-api-examples.jar",
                        "jarLocation": "public/default/nlu-test-functionmesh-ex2"
                      },
                      "logTopic": "persistent://public/default/logging-function-logs",
                      "maxReplicas": 1,
                      "name": "ex2",
                      "output": {
                        "topic": "persistent://public/default/functionmesh-output-topic",
                        "typeClassName": "java.lang.String"
                      },
                      "pulsar": {
                        "pulsarConfig": "mesh-test-pulsar"
                      },
                      "replicas": 1,
                      "resources": {
                        "limits": {
                          "cpu": "0.2",
                          "memory": "1.1G"
                        },
                        "requests": {
                          "cpu": "0.1",
                          "memory": "1G"
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Sink",
                "metadata": {
                  "name": "sink-sample"
                },
                "spec": {
                  "autoAck": true,
                  "className": "org.apache.pulsar.io.elasticsearch.ElasticSearchSink",
                  "clusterName": "test-pulsar",
                  "image": "streamnative/pulsar-io-elastic-search:2.10.0.0-rc10",
                  "input": {
                    "topics": [
                      "persistent://public/default/input"
                    ],
                    "typeClassName": "[B"
                  },
                  "java": {
                    "jar": "connectors/pulsar-io-elastic-search-2.10.0.0-rc10.nar",
                    "jarLocation": ""
                  },
                  "maxReplicas": 1,
                  "pulsar": {
                    "pulsarConfig": "test-sink"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "sinkConfig": {
                    "elasticSearchUrl": "http://quickstart-es-http.default.svc.cluster.local:9200",
                    "indexName": "my_index",
                    "password": "wJ757TmoXEd941kXm07Z2GW3",
                    "typeName": "doc",
                    "username": "elastic"
                  }
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Source",
                "metadata": {
                  "name": "source-sample"
                },
                "spec": {
                  "className": "org.apache.pulsar.io.debezium.mongodb.DebeziumMongoDbSource",
                  "clusterName": "test-pulsar",
                  "forwardSourceMessageProperty": true,
                  "image": "streamnative/pulsar-io-debezium-mongodb:2.10.0.0-rc10",
                  "java": {
                    "jar": "connectors/pulsar-io-debezium-mongodb-2.10.0.0-rc10.nar",
                    "jarLocation": ""
                  },
                  "maxReplicas": 1,
                  "output": {
                    "producerConf": {
                      "maxPendingMessages": 1000,
                      "maxPendingMessagesAcrossPartitions": 50000,
                      "useThreadLocalProducers": true
                    },
                    "topic": "persistent://public/default/destination",
                    "typeClassName": "org.apache.pulsar.common.schema.KeyValue"
                  },
                  "pulsar": {
                    "pulsarConfig": "test-source"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "sourceConfig": {
                    "database.whitelist": "inventory",
                    "mongodb.hosts": "rs0/mongo-dbz-0.mongo.default.svc.cluster.local:27017,rs0/mongo-dbz-1.mongo.default.svc.cluster.local:27017,rs0/mongo-dbz-2.mongo.default.svc.cluster.local:27017",
                    "mongodb.name": "dbserver1",
                    "mongodb.password": "dbz",
                    "mongodb.task.id": "1",
                    "mongodb.user": "debezium",
                    "pulsar.service.url": "pulsar://test-pulsar-broker.default.svc.cluster.local:6650"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/function-mesh@sha256:3e13e5cc3ac4d7387ee0adc02cb5183aad6cf2301f76ab230482609c4da1c0b7
          createdAt: "2022-10-27T07:49:28Z"
          description: The Function Mesh Operator manages the Pulsar Functions and
            Connectors deployed on a Kubernetes cluster.
          olm.properties: '[{"type":"olm.maxOpenShiftVersion","value":"4.8"}]'
          olm.skipRange: <0.7.0
          operatorhub.io/ui-metadata-max-k8s-version: "1.21"
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/streamnative/function-mesh
          support: StreamNative
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FunctionMesh is the Schema for the functionmeshes API
            displayName: Mesh
            kind: FunctionMesh
            name: functionmeshes.compute.functionmesh.io
            version: v1alpha1
          - description: Function is the Schema for the functions API
            displayName: Function
            kind: Function
            name: functions.compute.functionmesh.io
            version: v1alpha1
          - description: Sink is the Schema for the sinks API
            displayName: Sink
            kind: Sink
            name: sinks.compute.functionmesh.io
            version: v1alpha1
          - description: Source is the Schema for the sources API
            displayName: Source
            kind: Source
            name: sources.compute.functionmesh.io
            version: v1alpha1
        description: |
          [Function Mesh](https://functionmesh.io/) is a serverless framework purpose-built for stream processing applications. It brings powerful event-streaming capabilities to your applications by orchestrating multiple [Pulsar Functions](http://pulsar.apache.org/docs/en/next/functions-overview/) and [Pulsar IO connectors](http://pulsar.apache.org/docs/en/next/io-overview/) for complex stream processing jobs on Kubernetes.
          Prerequisites:
          - Install cert-manager operator first.
        displayName: Function Mesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pulsar
        - function-mesh
        - serverless
        - pulsar-functions
        - pulsar-io-connectors
        - StreamNative
        links:
        - name: Function Mesh
          url: https://function-mesh.io
        maintainers:
        - email: function-mesh@streamnative.io
          name: Function Mesh
        maturity: alpha
        minKubeVersion: v1.17.0
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/function-mesh@sha256:3e13e5cc3ac4d7387ee0adc02cb5183aad6cf2301f76ab230482609c4da1c0b7
        - docker.cloudsmith.io/streamnative/mirrors/gcr.io/kubebuilder/kube-rbac-proxy@sha256:67ecb332573384515406ebd71816781366b70adb0eb66345e5980e92603373e1
        version: 0.7.0
      entries:
      - name: function-mesh.v0.7.0
        version: 0.7.0
      name: alpha
    defaultChannel: alpha
    packageName: function-mesh
    provider:
      name: StreamNative
      url: https://streamnative.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.3.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo-operator@sha256:32178a16dd2115e0a11a2e3dfbfc130ae1d0afd8e293131e76b6fc1a79b36c12
          createdAt: "2019-05-01T00:00:00Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/kubeturbo-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/kubeturbo-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo-operator@sha256:32178a16dd2115e0a11a2e3dfbfc130ae1d0afd8e293131e76b6fc1a79b36c12
        - registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo@sha256:671be694176d9cc31b8a22c11bbfefc2504ae72d27fd010b315b6839f9ab59d5
        version: 8.3.3
      entries:
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: stable
    defaultChannel: stable
    packageName: kubeturbo-certified-rhmp
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: devspaces
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devspacesoperator.v3.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "org.eclipse.che/v1",
                "kind": "CheCluster",
                "metadata": {
                  "name": "devspaces",
                  "namespace": "openshift-devspaces"
                },
                "spec": {
                  "auth": {
                    "identityProviderURL": "",
                    "oAuthClientName": "",
                    "oAuthSecret": ""
                  },
                  "database": {
                    "externalDb": false
                  },
                  "k8s": {
                    "ingressDomain": null,
                    "tlsSecretName": null
                  },
                  "metrics": {
                    "enable": true
                  },
                  "server": {
                    "workspaceNamespaceDefault": "<username>-devspaces"
                  },
                  "storage": {
                    "pvcStrategy": "common"
                  }
                }
              },
              {
                "apiVersion": "org.eclipse.che/v2",
                "kind": "CheCluster",
                "metadata": {
                  "name": "devspaces",
                  "namespace": "openshift-devspaces"
                },
                "spec": {
                  "components": {},
                  "containerRegistry": {},
                  "devEnvironments": {},
                  "gitServices": {},
                  "networking": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/devspaces/devspaces-rhel9-operator@sha256:39c95279bced68928d028ef026f82eafe53eddb3dbe57703c8007cb5a170a70a
          createdAt: "2025-09-11T22:30:49+00:00"
          description: Devfile v2 and v1 development solution, 1 instance per cluster,
            for portable, collaborative k8s workspaces.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/devspaces-images/
          skipRange: '>=3.0.0 <3.23.0'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              The `CheCluster` custom resource allows defining and managing Red Hat OpenShift Dev Spaces server installation.
              Based on these settings, the  Operator automatically creates and maintains several ConfigMaps:
              `che`, `plugin-registry` that will contain the appropriate environment variables
              of the various components of the installation. These generated ConfigMaps must NOT be updated manually.
            displayName: Red Hat OpenShift Dev Spaces instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v2
          - description: The `CheCluster` custom resource allows defining and managing
              a Che server installation
            displayName: Red Hat OpenShift Dev Spaces instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v1
        description: |
          A collaborative Kubernetes-native development solution that delivers OpenShift workspaces and in-browser IDE for rapid cloud application development.
          This operator installs the Plugin and Devfile registries, Dashboard, Gateway and the Red Hat OpenShift Dev Spaces server, and configures these services.
          OpenShift OAuth is used directly for authentication. TLS mode is on.

          ## How to Install
          Press the **Install** button, choose the channel and the upgrade strategy, and wait for the **Installed** Operator status.
          When the operator is installed, create a new Custom Resource (CR) of Kind CheCluster (click the **Create New** button).
          The CR spec contains all defaults. You can start using Red Hat OpenShift Dev Spaces when the CR status is set to **Available**, and you see a URL to Red Hat OpenShift Dev Spaces.

          ## Defaults
          By default, the operator deploys Red Hat OpenShift Dev Spaces with:
          * 10Gi storage
          * Auto-generated passwords
          * Bundled Plugin and Devfile registries

          Use `oc edit checluster/devspaces -n openshift-devspaces` to update Red Hat OpenShift Dev Spaces default installation options.
          See more in the [Installation guide](https://docs.redhat.com/en/documentation/red_hat_openshift_dev_spaces/3.23/html/administration_guide/configuring-devspaces).

          ### Certificates
          Operator uses a default router certificate to secure Red Hat OpenShift Dev Spaces routes.
          Follow the [guide](https://docs.redhat.com/en/documentation/red_hat_openshift_dev_spaces/3.23/html/administration_guide/configuring-devspaces#importing-untrusted-tls-certificates)
          to import certificates into Red Hat OpenShift Dev Spaces.
        displayName: Red Hat OpenShift Dev Spaces
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eclipse che
        - workspaces
        - devtools
        - developer
        - ide
        - java
        - devspaces
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/openshift-dev-spaces/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_openshift_dev_spaces
        - name: Operator GitHub Repo
          url: https://github.com/redhat-developer/devspaces-images/
        maintainers:
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/devspaces/pluginregistry-rhel9@sha256:3ee131041761f5250b29d497043ecf7820e02e03e73ed0f929c17e2247c0d7a3
        - registry.redhat.io/devspaces/traefik-rhel9@sha256:95cf6a28028ec7a546a3d8d78377f204bb559c3d1022a73f08c646e76b85ce85
        - registry.redhat.io/devspaces-tech-preview/jetbrains-ide-rhel9@sha256:46bf7bea716df74f14d2629fb7e0e2ab692564cec348ffb8a6f4fea1739b23c2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5a6edfecf1636805e6f0bcad02df10bceec764f0580f3ac8fe62aba978d50006
        - registry.redhat.io/devspaces/server-rhel9@sha256:c0ec8c792f88aacf6ee31a4b6d2d847d864707a084148b48e4e402388989f868
        - registry.redhat.io/devspaces/devspaces-rhel9-operator@sha256:39c95279bced68928d028ef026f82eafe53eddb3dbe57703c8007cb5a170a70a
        - registry.redhat.io/devspaces/code-rhel9@sha256:abc9b12131c8cdca0719e0404d354310f53769990b089f1a3adea31193798a8e
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6bcfa981a7a75c242a16afa6dab36bf6e3b8a1442fa863875919ad551f48eceb
        - registry.redhat.io/devspaces/configbump-rhel9@sha256:cecbc3796058618da9ad0473a6f937aae5aca25776bb434894286f22baa4677c
        - registry.redhat.io/devspaces/dashboard-rhel9@sha256:c34cf5ac9bf95c1c5082afd5a2941b5e412be8ca43aef6a72e8fe5a279a43271
        - registry.redhat.io/devspaces/udi-rhel9@sha256:f7325ee7259a0585e8017d4442d778a37f4c8a67d81fe70c241a014001e37539
        - registry.redhat.io/devspaces-tech-preview/idea-rhel9@sha256:df24ebb239a2cc0d4543309f51a7268c7d524d428a536e0b3443c51089c3f87e
        version: 3.23.0
      entries:
      - name: devspacesoperator.v3.23.0
        version: 3.23.0
      - name: devspacesoperator.v3.22.0
        version: 3.22.0
      - name: devspacesoperator.v3.21.0
        version: 3.21.0
      - name: devspacesoperator.v3.20.0
        version: 3.20.0
      - name: devspacesoperator.v3.19.0
        version: 3.19.0
      - name: devspacesoperator.v3.18.1-0.1739475915.p
        version: 3.18.1+0.1739475915.p
      - name: devspacesoperator.v3.18.1
        version: 3.18.1
      - name: devspacesoperator.v3.18.0
        version: 3.18.0
      - name: devspacesoperator.v3.17.0
        version: 3.17.0
      - name: devspacesoperator.v3.16.1
        version: 3.16.1
      - name: devspacesoperator.v3.16.0
        version: 3.16.0
      - name: devspacesoperator.v3.15.0
        version: 3.15.0
      - name: devspacesoperator.v3.14.0
        version: 3.14.0
      - name: devspacesoperator.v3.13.0
        version: 3.13.0
      - name: devspacesoperator.v3.12.0-0.1712003563.p
        version: 3.12.0+0.1712003563.p
      - name: devspacesoperator.v3.12.0
        version: 3.12.0
      - name: devspacesoperator.v3.11.0
        version: 3.11.0
      - name: devspacesoperator.v3.10.1
        version: 3.10.1
      - name: devspacesoperator.v3.10.0
        version: 3.10.0
      - name: devspacesoperator.v3.9.1
        version: 3.9.1
      - name: devspacesoperator.v3.9.0
        version: 3.9.0
      - name: devspacesoperator.v3.8.0
        version: 3.8.0
      - name: devspacesoperator.v3.7.2
        version: 3.7.2
      - name: devspacesoperator.v3.7.1-0.1691407091.p
        version: 3.7.1+0.1691407091.p
      - name: devspacesoperator.v3.7.1
        version: 3.7.1
      - name: devspacesoperator.v3.7.1-0.1691548593.p
        version: 3.7.1+0.1691548593.p
      - name: devspacesoperator.v3.7.0
        version: 3.7.0
      - name: devspacesoperator.v3.6.0
        version: 3.6.0
      - name: devspacesoperator.v3.6.0-0.1685320856.p
        version: 3.6.0+0.1685320856.p
      - name: devspacesoperator.v3.6.0-0.1685322634.p
        version: 3.6.0+0.1685322634.p
      - name: devspacesoperator.v3.5.0
        version: 3.5.0
      - name: devspacesoperator.v3.4.0-0.1675914643.p
        version: 3.4.0+0.1675914643.p
      - name: devspacesoperator.v3.4.0
        version: 3.4.0
      - name: devspacesoperator.v3.3.0-0.1673528130.p
        version: 3.3.0+0.1673528130.p
      - name: devspacesoperator.v3.3.0
        version: 3.3.0
      name: stable
    defaultChannel: stable
    packageName: devspaces
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Alvearie
      provider-url: https://alvearie.github.io
    name: alvearie-imaging-ingestion
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: imaging-ingestion-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomEventBridge",
                "metadata": {
                  "name": "events"
                },
                "spec": {
                  "dicomEventDrivenIngestionName": "core",
                  "eventBridge": {},
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "events",
                  "natsTokenSecret": "nats-events-secure-bound-token",
                  "natsUrl": "jetstream.imaging-ingestion.svc.cluster.local:4222",
                  "role": "hub"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomEventDrivenIngestion",
                "metadata": {
                  "name": "core"
                },
                "spec": {
                  "databaseConfigName": "db-config",
                  "databaseSecretName": "db-secret",
                  "eventProcessor": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  },
                  "imagePullPolicy": "Always"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomInstanceBinding",
                "metadata": {
                  "name": "stow"
                },
                "spec": {
                  "bindingConfigName": "instance-binding-config",
                  "bindingSecretName": "instance-binding-secret",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "instanceBinding": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomStudyBinding",
                "metadata": {
                  "name": "fhir"
                },
                "spec": {
                  "bindingConfigName": "study-binding-config",
                  "bindingSecretName": "study-binding-secret",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "studyBinding": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomwebIngestionService",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "bucketConfigName": "imaging-ingestion",
                  "bucketSecretName": "imaging-ingestion",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "providerName": "provider",
                  "stowService": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  },
                  "wadoService": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DimseIngestionService",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "applicationEntityTitle": "DICOM-INGEST",
                  "bucketConfigName": "imaging-ingestion",
                  "bucketSecretName": "imaging-ingestion",
                  "dicomEventDrivenIngestionName": "core",
                  "dimseService": {},
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "DIMSE",
                  "natsTokenSecret": "ingestion-nats-secure-bound-token",
                  "natsUrl": "nats-secure.imaging-ingestion.svc.cluster.local:4222",
                  "providerName": "provider"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DimseProxy",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "applicationEntityTitle": "DCM4CHEE",
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "DIMSE",
                  "natsTokenSecret": "ingestion-nats-secure-bound-token",
                  "natsUrl": "nats-secure.imaging-ingestion.svc.cluster.local:4222",
                  "proxy": {},
                  "targetDimseHost": "arc.dcm4chee.svc.cluster.local",
                  "targetDimsePort": 11112
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: alvearie/imaging-ingestion-operator:0.0.3
          createdAt: "2021-04-08T15:29:55Z"
          description: The Alvearie Imaging Ingestion provides a collection of components
            for extending enterprise medical imaging informatics systems to Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Alvearie/imaging-ingestion
          support: alvearie.slack.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DicomEventBridge is the Schema for the dicomeventbridges
              API
            displayName: Dicom Event Bridge
            kind: DicomEventBridge
            name: dicomeventbridges.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Event driven manifest of all DICOM data across all associated
              S3 buckets
            displayName: Dicom Event Driven Ingestion
            kind: DicomEventDrivenIngestion
            name: dicomeventdriveningestions.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Fan-out notification of DICOM instance data (STOW-RS)
            displayName: Dicom Instance Binding
            kind: DicomInstanceBinding
            name: dicominstancebindings.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Fan-out notification of DICOM studies (FHIRv4 ImagingStudy)
            displayName: Dicom Study Binding
            kind: DicomStudyBinding
            name: dicomstudybindings.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides DICOMweb WADO-RS and STOW-RS access to a S3 bucket
            displayName: Dicomweb Ingestion Service
            kind: DicomwebIngestionService
            name: dicomwebingestionservices.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides a proxied DIMSE Application Entity (AE) in the cluster
              for C-STORE operations to a storage space
            displayName: Dimse Ingestion Service
            kind: DimseIngestionService
            name: dimseingestionservices.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides a bidirectional proxied DIMSE Application Entity
              (AE) in the cluster
            displayName: Dimse Proxy
            kind: DimseProxy
            name: dimseproxies.imaging-ingestion.alvearie.org
            version: v1alpha1
        description: |
          The Alvearie Imaging Ingestion provides a collection of components for extending enterprise medical imaging informatics systems to Kubernetes.

          ## About this Operator
          This operator enables standards-based medical imaging application interfaces such as DICOM Message Service Element (DIMSE) and DICOMweb.
          The API's provided allow for extending enterprise imaging systems to the cloud with custom, fit for purpose deployments.
          This is possible at speed, low cost, and little or zero integration effort.or extending enterprise medical imaging informatics systems to a modern cloud.


          Some potential usages:

          - Perform lightweight secure bridging of DICOM from the enterprise imaging zone to a cloud availability zone.
          - Utilize S3 buckets for segregating DICOM data to be used for clinical or scientific purposes.
          - Fan-out DICOM data to a number of different DICOM subscribers
          - Generate FHIRv4 *ImagingStudy* resources from ingested content and publish to a FHIRv4 based patient logitudinal record.
          - Distribute imaging study arrival and update notifications to imaging study subscribers


          ## Prerequisites

          - [Knative](https://operatorhub.io/operator/knative-operator)
          - ProgresSQL such as [Crunchy](https://operatorhub.io/operator/postgresql)
          - S3 storage such as [MinIO](https://operatorhub.io/operator/minio-operator)
          - [NATS operator](https://github.com/nats-io/nats-operator/) (Optional:  only required for DIMSE Proxy component)



          ## Documentation

          Visit the [official github repository](https://github.com/Alvearie/imaging-ingestion) for documentation and examples.
        displayName: Alvearie Imaging Ingestion Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - dicom
        - medical imaging
        - fhir
        - dimse-c
        - dimse
        - imaging study
        links:
        - name: Alvearie Imaging Ingestion Operator
          url: https://github.com/Alvearie/imaging-ingestion
        maintainers:
        - email: rduggan@ca.ibm.com
          name: Richard Duggan
        - email: jjacob@us.ibm.com
          name: Jeesmon Jacob
        maturity: alpha
        provider:
          name: Alvearie
          url: https://alvearie.github.io
        relatedImages:
        - alvearie/imaging-ingestion-operator:0.0.3
        version: 0.0.3
      entries:
      - name: imaging-ingestion-operator.v0.0.3
        version: 0.0.3
      - name: imaging-ingestion-operator.v0.0.2
        version: 0.0.2
      - name: imaging-ingestion-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: alvearie-imaging-ingestion
    provider:
      name: Alvearie
      url: https://alvearie.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-licensing
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-licensing.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Licensing",
                "metadata": {
                  "name": "licensing-sample"
                },
                "spec": {
                  "environment": {
                    "deploymentType": "openshift"
                  },
                  "hostid": {
                    "enabled": true,
                    "environment": {},
                    "hostMounts": {},
                    "image": "prod.harbor.keysight.digital/loadcore/kcos-licensing-hostid-vmuuid@sha256:076a2c6aeec875470414af7905613a25b56bcb2aca5edcb1c4c53037cd01b332",
                    "imagePullPolicy": "IfNotPresent"
                  },
                  "image": {
                    "dbimage": "prod.harbor.keysight.digital/loadcore/postgresql-ubi-image@sha256:17f6a8f775d06650939d2c73e82b45d73d50190cc735ca70695d082f7790834b",
                    "initOwnership": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf",
                    "repository": "prod.harbor.keysight.digital/loadcore"
                  },
                  "ingress": {
                    "annotations": {
                      "nginx.ingress.kubernetes.io/auth-method": "POST",
                      "nginx.ingress.kubernetes.io/auth-response-headers": "X-AUTH-USER",
                      "nginx.ingress.kubernetes.io/auth-url": "https://127.0.0.1/token/validate"
                    },
                    "enabled": true
                  },
                  "nodeSelector": {
                    "node-role.kubernetes.io/kcos-licensing": "kcos-licensing"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:29:04Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Licensing
            name: licensings.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Licensing Operator
        displayName: Keysight KCOS Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Licensing
          url: https://keysight-kcos-licensing.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-licensing@sha256:a09291923ad8d7533b33e40e1a3c4e735c2a4ae3b2b87be7c9d1a1a9244b3dad
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - prod.harbor.keysight.digital/loadcore/postgresql-ubi-image@sha256:17f6a8f775d06650939d2c73e82b45d73d50190cc735ca70695d082f7790834b
        - prod.harbor.keysight.digital/loadcore/kcos-licensing-hostid-vmuuid@sha256:076a2c6aeec875470414af7905613a25b56bcb2aca5edcb1c4c53037cd01b332
        version: 0.1.21
      entries:
      - name: keysight-kcos-licensing.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-licensing
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: multicluster-engine
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-engine.v2.5.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-08-06T12:21:10Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0 <2.5.10'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '[]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine is the Schema for the multiclusterengines
              API
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:80f33a762535ff69f57ecb04de043b8fe249a63767712f4942ac6685450e8f28
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:1af88c3703b668020137f90657e6262b73e407aa858bbc6d9c12ba3a1cab8043
        - registry.redhat.io/multicluster-engine/agent-service-rhel8@sha256:92464c57700dcb13ccf5431c104a3fba6845f651776aa5beb1ce7171087dc1ea
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:cb2c57c51edc8a2320a10ac217cca6a401db09de55c0c9d0c9a7b4545e2636e9
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:7f32ed997e03dba1746b3627d5a33fc69282b2e831edd21ec74020fe4fa1ca5d
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:90e792c92ef8c75ffb97545394206c9197997d474493d93f09fca273f65133e4
        - registry.redhat.io/multicluster-engine/hive-rhel8@sha256:27eefe6ad7d93e752f4d76b776884129615c2e87ff24116c401a886e98abfcc5
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:1e9eff48ea837a1d19c595986d765d74810d6d1397602ad79e2e820b1cbb5c6e
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:297d8d0d5855db90ad40af702d6ad5f3f32a4aab486686919d86bbd21c367ca9
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:b0a70eee404ba1267c0ae1a8eb140251d054e722305d680685223cfda360aa37
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:53b8d48e540c3d9bb5d948af0375c26377271e595fffdb3d5f921e3a358fa298
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:36abf06619cd3971da8f42f4193621a1384970365fce0f48e86a65e2840bca38
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel8@sha256:360d3dfb823d8e199574cf1c3fbf890fa74c71fc48644e22b47339363e941b3e
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:c8d48e7ccdce43ce72154e464e74c32477efb0094aab327ed0eaab6c1244774a
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:515eda1e4e1ced222709103789af14f7b4e99d4da037deddd8499aad3b45451b
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:3e3f7ee898e0f6b0f74026399753d7fde41013ade76f7963d5c25e578eb1906e
        - registry.redhat.io/rhel8/postgresql-12@sha256:da619f6617d99790b4fe00fdb23ab4aea0cdb0592590162d51b3ef94aca01a7b
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel8@sha256:6bf850d8fb250740113bfb004b19edbd5cf560c8175b81e2f3766280b3482e27
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:43cc171404571bb3c87275a86e9ab4409861c1c01b5f1441c455565d56e1921d
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:d71c1441577b1cd360d074581ad19c2b0cfa99cdfe1869af1777ff7de6821735
        - registry.redhat.io/multicluster-engine/cluster-api-provider-aws-rhel9@sha256:7f65119dc5d81b8ba485b651efe1d9013b62ec57e4725dbf8bdc2e3b7748f677
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:1a5ff852d96bd15c631a6f89e2b0047cd8dde6264646f51106f2b523decc64ba
        - registry.redhat.io/multicluster-engine/aws-encryption-provider-rhel9@sha256:22cefc43e205e4bdd7874c7e1fccba60fe556df4b2bab055d01c834d62f194bd
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:53ad9402a0a13c894b485f418dd1353c8870792893a5ce33f9d6c3cc9839ebc4
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:2497e077a00dfa94b913359be0c696126e932db1a45e066ead65dee76a1db8fd
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel8@sha256:1e755b64358c10309804a3bf44ca640a7d5bb6e0474d33774a8d76095f2b8c3d
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:25b30d4f4ccc23fda707db88969c0ae036ecb4a2baf15b82221e3351c23292f9
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:a4f58b94d45c1ae1c4ba69b12030b9945b37f3ce6e96c27919f256ae887ec28b
        - registry.redhat.io/multicluster-engine/cluster-api-rhel9@sha256:0e329ca39b1f18520da093a5194ef6e1d5527270c74bf260517620577a21b6c7
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:19dba8be2295f54a93853dbe566132011f0120b2c4b48de2816509f5d201e420
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:bd6d0189ea2bbc87b5745f9d7f9c82118f8dcc9ee63fe0a36bb5e7dfc0e56021
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:424ca238d9c8b20dc434a07aa6a27fe6e7c63b3e2480eead284570864d682175
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:3d420ba8bc5e1dbe629c0d4aa65c0a79a7914ac349828046994e245f4f2adb2b
        - registry.redhat.io/multicluster-engine/assisted-installer-reporter-rhel8@sha256:5d41ca13f4ae102760b65375578c2f632aa6ce41092732300d6727f195da3d55
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:460d37a0032813aa85a3b18bd6f152e4ed6b89f802d17d787c3365082420d379
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:d7ccc8ddbeb11ac8a3a0d4946920a0ab87d9e9595d7c61740bd57f3f4a19a711
        version: 2.5.10
      entries:
      - name: multicluster-engine.v2.5.10
        version: 2.5.10
      - name: multicluster-engine.v2.5.9
        version: 2.5.9
      - name: multicluster-engine.v2.5.8
        version: 2.5.8
      - name: multicluster-engine.v2.5.7
        version: 2.5.7
      - name: multicluster-engine.v2.5.6
        version: 2.5.6
      - name: multicluster-engine.v2.5.5
        version: 2.5.5
      - name: multicluster-engine.v2.5.4
        version: 2.5.4
      - name: multicluster-engine.v2.5.3
        version: 2.5.3
      - name: multicluster-engine.v2.5.2
        version: 2.5.2
      - name: multicluster-engine.v2.5.1
        version: 2.5.1
      - name: multicluster-engine.v2.5.0-103
        version: 2.5.0-103
      name: stable-2.5
    - currentCSV: multicluster-engine.v2.6.8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-08-04T11:16:16Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0 <2.6.8'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:69a40339bf3b3dd3c3b6359721895313147a5f682b62faf5151288ba63b65812
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:f6b15f827aeabc33074de1ff9ed20edaf04271b4e4d2eb16abbe786c78e033ca
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:9e1ea96b998b0c7854ca3c7c8a505fc96cbc1b35e6ce9f49b1dbec703fdc8070
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:7dde47f7fe16a26676a6e5b1739954d76cad1cddbefce2d6869ffb2af399e403
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:f563ff0acb4ea9d1bb112765b67bee535063304a98a2aa2cb7f21b0155d49262
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:54b8e17cacae2d355d82677ed334b4981b50e825579ff30406039933833e828b
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:a8117c99b33eedc702baeab32f0347892b587dfc1860f467cb1900565c3fdb3b
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:00ea8fe4945e93d2cbeceda4440acd6a53a48b4dd420dfe2cbd5287e20f0362f
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel8@sha256:42fee1066ff00066a172f82754ad9f08e8c29029e75f76712cb283d9fe5d1d9a
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:1d098eab26421f99ef71ffce82e4fa59ef0f9043de04ee74ca43486cbdb82006
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:b39819aa916974f4a408d23177e99cfb71e21af8ff9b0eacf6c37159c354ba70
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:0737e57ae8af6a62ac580710276d2c1620ac1b1283d3b84369db43e76e518610
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:5f9d5bdfd45f931989d069dcb1dea91a69020d83011fe0846ddfe22bba14b001
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel8@sha256:a630a20a633d057e475b4c39b342512119a71021e199aa0becf80899d8840ad7
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:c3c654be0b20cac501af1e31631464c32aecdbd16cbfb88f865900c70ccf77c2
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:7c5f339e762db1714b3f6cf1fc1bc8f08892cac433ab5a46e73b530fa08486fb
        - registry.redhat.io/multicluster-engine/cluster-api-rhel9@sha256:bd4eb05427ba2ea35f01227600cebd0fe8598d4de456348c24bb5e88d9ed8720
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:aff06db527e4ddc713a2fb8885824f0482647cbdcd844b311ccc3880a0dd2c3f
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:5ef8acc612817d478cede3b1bda55abafac00d0debebd68d948d157ee5b09286
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:294928c58b33ae8e776f86df211dac3b63a2cd9fc00aa995a0de037a29cb2c7d
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:2d807348774b8c6c0255c4bc931a956fc987c961fae6be689835410e8e301fd0
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel8@sha256:4d96379ad99d299d2a89eb1b0eb8a22db8bdada9a16248503cd8a9a5900eab35
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:d2c1d90a4e829b181fdf3d2838764c8f2b61c22b761279f53cb22316ad46ca03
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:93d65bfa0102c7ad7ec6ebfaad9ab062a31098b7ad7026313bcd8c6a495d5ca7
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:cd23f237c91dd8585946b31e17e38193f2e5ac8993d65459bb703a68f6fa8ca0
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:80a6615500949cc8c7a5b2a1e8c8c5f4564e9cdb9f995787a44f462e5bbd62cb
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel8@sha256:a703f2b1f4eb538aa02dbc0f4ff793506076480e94905c367b0d4ae352b87b65
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:5d488c8a644f2bfce97444d770516de10ceecc988cb3e9fbf0072d0b0a2eeb83
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:be1aa4cf5e09914774dcf69ab7ccdaf67504309d40c7e8cf4d91e9c1080e706f
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:82c0af2d285ba943dfa1da9f0856b26ec7e4c2c431577fcd4ab031ae040cfaea
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:87f5db53ad18f6d84501a5c2294cff73a5425276cdbde92225909826091b5ffa
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:46c4ac0cacef587b6e678b6cb0f68c29f5ac63bd59209d42a3ae0b7807e99972
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:d7c8bef4c3dec710ccb299482a10bfdf6efacd4e2ed2e6e323d448ed505000b3
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:d71ef80913d4e34389bce56c4a282cf7d2be705f003575dab87c07ca71f8c2d6
        version: 2.6.8
      entries:
      - name: multicluster-engine.v2.6.8
        version: 2.6.8
      - name: multicluster-engine.v2.6.7
        version: 2.6.7
      - name: multicluster-engine.v2.6.6
        version: 2.6.6
      - name: multicluster-engine.v2.6.5
        version: 2.6.5
      - name: multicluster-engine.v2.6.4
        version: 2.6.4
      - name: multicluster-engine.v2.6.3
        version: 2.6.3
      - name: multicluster-engine.v2.6.2
        version: 2.6.2
      - name: multicluster-engine.v2.6.1
        version: 2.6.1
      - name: multicluster-engine.v2.6.0
        version: 2.6.0
      name: stable-2.6
    - currentCSV: multicluster-engine.v2.7.5
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-21T22:32:21Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0 <2.7.5'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:44b07a18116ad827c23ca5a98b9fe8ef74a99abe4787cd01d848de42df44178d
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:15c2873703afcc74e41ae27dbb4761ad19c4bd2594ca6de82d3b6c02752e0ce3
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:816712bb3ccb58b705a85e5f4a55e519316fd8b9d7d9295d1042420ce44a4bad
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:e6a5e71f0aee8a682179cfd8f85b45a2ee04a1de680fc03dabdb659a2d2c0396
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:2a78ab192a554e78f3fc367c47e571f96b650ee9fb5ba0d06674f340f4320d94
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:9d1be5eb4fa357fe6bd97a52a53b3eeb537c431daf7af8a30efc3e0b6c5d5d66
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:2e3fe5e3b39721cc77b7d32b5682300eb60b1bc6940b8e83662b458d88489140
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:a24e9855a11d879d5308e29af6e3b2947365b7e466e7fa12dc57997140753604
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:b124242a55b1b2a2258cb12094617a0e009cf1cf58a60ab1b42d48d4ea2514ac
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:b66f2f4b5e0d77d501f9cc9da483f976f3c0eaf0ca72a5e9c9435c92d58e1388
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:00d415071f4c10575f9e5068868a7f80b0c0fb9c8614da86bc515923bed136a3
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:e2e16ff6213c0274126c9d64d495c0abbe17379e841ab730de66520908f1a190
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:4a0f1d3e696310f7a5bccbb358c538023a28f5bbc23b5743d10fd61b452d45d0
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:041044c08a1d526b349b5ad5fc4c4457f3632b1353530ed925c65ef2cce8e9c5
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:2025e424c11c6fe4a4395945c2e7d124d07cde43bb8d8cbe70286fe6ae345754
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:8c420720a48ea7c3aa096ae1929dcaa979c39e7865c5d43cba899e294745fc94
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:081151ca906b0f1afbcd3ee9c6d9e5bfe234494223c660ef376aa4004b0508f4
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:3defab1abc7395c9f9c22912ec6825be11b120a1416404eb0599e91f7f9c679a
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:c0fb7985620d5bd0dfadd57cb75664ee6777b8dac5a3463967e9b649e9eb836c
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:9e5c66ffdf39ae923ecccf79be538b883f2bdc8019169d55057565f077997d67
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:4fa34ab0aa8854eb5e942f2ed315041bf2026ce0bed86f57125878af0c73dc8b
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:7369bc79a283ff8b34c3189ea83e55f88a00d619cc78f8dad03d302dcb34a550
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:fedc699c0f786fb303d3956769f9c5fca1473ec4bf014ec87c5a4d40e34dcdb3
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:b82b1837dfc24a367f68741dece6d19c89c576706f006ee8d5123a0c6782c492
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:90c19d2aaa50e7cc9184f2f220b568c34c9fdcfdcba314da555885c999078057
        - registry.redhat.io/multicluster-engine/cluster-api-rhel9@sha256:cd8d6e0056fb0436275c12cc6b42f11c0ccd06729c7108e964b03d6c48d063bd
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:0684bcde2c018ae893bd71be8df3448afa09d199415e83285d2c740cff66192d
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:2216b015943eb8d9f25bf4ac37fbad2b7215a5c0003bd7c3b3d97b3b0a5ba708
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:555fcfce226e3b059f8f911d44a48578b898e085b6ad2575479feee71ec8baaf
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:a6736c7cac00f78c0e8877549ae2b74b7207cc348f24f175133db49fd38f240a
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:8e41d489e1def9b540ddd6387a1cc151b24e91a5fa44a74fbbaa0d1978588d57
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:45571dbb55bd213400eec6d2065b40615185bf3e52efdf495ab73201ec94faff
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:1c176688f13052783662115d5cabe8d614b8d9f79507dc2bdd4be1604e2eecc0
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:2cbd8c470dc4b263fd11952aa792094a7550f4b99e10f22186d51621bc8d3b3f
        version: 2.7.5
      entries:
      - name: multicluster-engine.v2.7.5
        version: 2.7.5
      - name: multicluster-engine.v2.7.4
        version: 2.7.4
      - name: multicluster-engine.v2.7.3
        version: 2.7.3
      - name: multicluster-engine.v2.7.2
        version: 2.7.2
      - name: multicluster-engine.v2.7.1
        version: 2.7.1
      - name: multicluster-engine.v2.7.0
        version: 2.7.0
      name: stable-2.7
    - currentCSV: multicluster-engine.v2.8.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-09-08T13:08:30Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0 <2.8.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:f869555beec0875c74f30a88428ac9641029dfed0365d45b23d5cae469cf7d0f
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:7ca7e87615e12b9d7fa75befd05f4f45bdc4bf8ef1b576c69f36a23c66c90cd6
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:988eb4aa01ba0d7e6728b23d043e4c0babe872ac23b6520f4daac714619a360b
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:f809a8774b0e3f891af46931dd4d2c11b95da4e7f5a8aee1814152d9a468d357
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:c5ac848e3b7eb88cdbdbf0b5f4aa96698d852b3a8118cf7146eb7e1948afaeba
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:a7f63077799a62fe6ece76a12346a87b5cbae966ab69df6ec6d951259dbf0e4a
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:c3b62ab1234257d30f41fa1ab33cb319321777a18de8ed751be1fcd48b204f2e
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:318ca0c7e9cd9a7336178519cadbf21eaeb474cd4a8eb5cf411a640d1c77372e
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:5ea86c20383a6932c593e020fd3b7b1a2a8e666bc374bff8b4bca0ee2d4ecc33
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:aa2754d35c8460a2879d4d523ebf32351d36e8fd7886335b5c0d2ca4befb8e8e
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:bee8f0c047cf8dbfa3f28712ce91bea55e926b2edb26c394f18a2d780e142083
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:7c86c89ff57091d4112783d82d1add47f2d56dba584418d5664c198ba048e0c2
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:f708f3e1007ed2b0e420e51d8b7ff9dd9bccaea9234755f4375f5514018211b4
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:3788b2674dda0b2cb4e9ca21bdce0b9e44702d72458b2396ef0acdad37fa866b
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:18b0e86305afe2eae3c26e637755fa9e8ab2abf30bc15e5428467a9a81ec99b0
        - registry.redhat.io/openshift4/ose-aws-cluster-api-controllers-rhel9@sha256:ada8b480a86d0a9e3ecc58f8b1cb53dbd7452b2b9183b3ba2b15a6a58c4436d1
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:f694b6e9e76941b11cc8a15153d9005cffe9bbbf58d0962f8ff546e90a63d93c
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:806b6c956f7bad824f0fe9acc3b534964e8f5f3399a1813f75511bc0f3904414
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:cf457b5a13cdd54c45a7b88f0f150a04e756bfac727c4365cb12f113d904937d
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:5461643a61850d95409d989e81c304053372f1dc5dbba291f3f3386de9ecab1c
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:261b99b93d98825ade0a0d3b9c4e465a785f35d2a310dedd4bbd6e4aa1f33ef9
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:0392c81a404be343d658dbc64153cbc46a64d0adf69aa8024c3f727351102d75
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:8379827635dfc4b51e04145d8048f7dcf9c6956ea1a22371bd22f948ae01b183
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:ee014c576263e19798df98460ebb5f1a21cb6249f09fe8a267b2dfa98b07d1eb
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:d4d11d0aeb3e1c9ff2a80e769d92dce509a0202a2ef1c9d2430030acdbf85bce
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:83a82f9558c406899e573308947c15b1f36de454006bb0e2ac44709f0c2c28d7
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:ea2ba066abd170e1087558c7b701faeb1be2e1667943dc14440f19c2ee5b0c30
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:fb2b8cef9916f387d26e65d9b53c7567ae1433d43d16e879ddb582ddc248eaa3
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:5ad96dad74c6f4967bd5fd397d6697e01b3b25bb3b8bae123a38e96964ddfa17
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:b9afa4d5328f17c7c3ce1dc105d177ecacf3447092d51a50f250a4b0ea692e2c
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:c6cd578123c1c86fcb3305ecca117543106f943cbcc4baf9065683988b2e10ff
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:e5e459e665afc14a1540e300015222b7c1ac896e2bee889283d67c3e19ca01ac
        - registry.redhat.io/openshift4/ose-cluster-api-rhel9@sha256:ec58463239eaa9bd15b31cbcfea0a34fd650201bac6120aa7ddaf72734451e06
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:ab0898d513b9b5ca2f5ca5879adc4e73da3232d515c0647a2a0ec28792264db1
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:0d1dfcb06eaa26de74994891371a5657aded7a59972f006f8c4a3859369877f3
        version: 2.8.3
      entries:
      - name: multicluster-engine.v2.8.3
        version: 2.8.3
      - name: multicluster-engine.v2.8.2
        version: 2.8.2
      - name: multicluster-engine.v2.8.1
        version: 2.8.1
      - name: multicluster-engine.v2.8.0
        version: 2.8.0
      name: stable-2.8
    - currentCSV: multicluster-engine.v2.9.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-30T19:22:43Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0 <2.9.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/open-cluster-management/backplane-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational components for managing multiple Kubernetes-based clusters. The deployment of the multicluster engine components is
              determined based on the configuration that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:8901e984b3b47f0650cddc1c36bdd89af22390f3d4a562d562fde7774d8ba6db
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:f3d65f68538e94260fad54f145433a36366f8c88f9c9f80b7fc5545ba62e1773
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:0a9c0ccb6a75273474fb04b785aa8fc58aec2f9ffbff8c232d359e14892bbae2
        - registry.redhat.io/multicluster-engine/capoa-control-plane-rhel9@sha256:b510a25a6370ca6b8d2f74165814ebfdae0ce9898cc343e73fab6344eebf2bed
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:8db1386d12d40dfbc80054c023897e5bfc0a987f6d07792b0f9f50f59f1a68d8
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:57699ecb529cb46843de847deef3e3757ca252910249d8deac16fcb00c01b636
        - registry.redhat.io/multicluster-engine/capoa-bootstrap-rhel9@sha256:a8b1278a2df9b17cee66813715995385bb3539af05e4df9b5bf7d9ccd508c313
        - registry.redhat.io/multicluster-engine/mce-capi-webhook-config-rhel9@sha256:7f1f70813e6d753b4acb04c7e9d7acb0047ad91f8a9a10e0fe2fff6c4411c8b4
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:a91276fcf5a589ef11f01b39df7d33ceb58ebcb47f95dfab772114bce7c1b4a8
        - registry.redhat.io/openshift4/ose-cluster-api-rhel9@sha256:9a72b8081a57846d942981ced07da86ad8e197309a1bc3c9d53876528b58f8ec
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:6a75df76fb0d9516b13fa6b079bc6c4d49dc6774f0d4f31343abfb8d70edcc23
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:ce0fd7bdc73cd2c84d1a9c5af528916488ca8116dd05d83dacf72f2af4248e65
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:ac43feaf079fa0d250d53cd9685fb68677d888dd091730faa32757dcdf2c95e4
        - registry.redhat.io/rhel8/postgresql-12@sha256:4773a29755ea826c0d3b58a8b8297c1d7388019516195cae9bd815f28ec4bca0
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:5e6594464bf7ddc2a66f127f645d93ad35d8cd95dd70544fa7fcec791f67f45c
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:36a6000b6482966f8bbf86e45c25335c066736ae1f8695a8b68e6b76b7c491ac
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:1b218f0aa0ea1f494094e3bc783ef40968cbff780e22de186d578a1d64c33bf8
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:3c450cb6a3ba6f2e3028856f14967803c1a691d95b8f54bd9d29d93691d6e795
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:8076db8ccd5578cb5aba2d66f53790563d3f42b93a6cfd14c7094d3a25707982
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:ad382d5eb347c448c46495a0a17503d789db4eb1096ad71a588b316b235e437f
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:6b95560a6a6dc1ddd41c76f275f83baec52ec9d79de04bfe54feaed02db123a3
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:ee75e805f0f9d73f073ce870f5d46962c0033d63158017841d17410febfc2c57
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:25ec9a0cc4ac92d4bed75c38fa37ef8d5ff9478cc475049b303d74eb58687a81
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:f8c3898e29f8c0a20c3be92bc4ee5a9443b9fc8218db95ba541fe3e57a89c40d
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:bf343592bf2f4399b5ac61709f56528435990e903812e264cbef845ff4433231
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:9555fd5f577f32a1d8ebe25f0d03e453c8ea925e4a12dbbe1015c58d3c8528b0
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:1cc6fddaf9e55637e90d5c3e598467df8967b8a362f8b8d46286b256bad6dd3f
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:a41cafc65e30deb073e051542702e258901b552ad00e42eacab05d149d6f0565
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:ea57247cf11f6ab966893ce4e6d2948232a4314588af2ca33c1d9689da7482dd
        - registry.redhat.io/openshift4/ose-baremetal-cluster-api-controllers-rhel9@sha256:f855d5ce17d3ac4560afc73c35d03f045f84551fdccc00312c890288a889b9ce
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:72a4ac73bb5e7797329e1856f785751d48815a64e82350a5d694bb5a8e121956
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:cd9dc89b933b394b08fe05ab5f0bbecea727bfb2c29770f11ec1db4aa5740b3c
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:37087ad8678f3e8ec8e2b2d035d3a12087ce9e0c993194a04a9874fd07e80d91
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:cecec1f7d7f0d37cf17fb25ccf34602728f2a09c61af7e212967b4ee28d2dec3
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:d068d73f41d8f8ecbec36752d0a432913bfbe2580f489abefe513c06233e80c9
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:edbb174cb69dd31339a53542d24c4fd00552e41fa09d7e2553baa7a1ad8dd810
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:1bbd134d44839258172e013fe84bfd174d36d6b7232e2cc8a3d8556998181d9d
        - registry.redhat.io/openshift4/ose-aws-cluster-api-controllers-rhel9@sha256:cf13b0658271cd0dbd1ce6dcc350e64c5459ef02ba4dc88a0031de9d6f26ffd1
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:8d189b565e4c3c434988aa4ff0ecad4c5f86b2a94120325c142ff111b25facde
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:902c5b3b4a4e8cb0a14fa5aba8451ec02830d5167ff1b91323ded62255983cd9
        version: 2.9.0
      entries:
      - name: multicluster-engine.v2.9.0
        version: 2.9.0
      name: stable-2.9
    defaultChannel: stable-2.9
    packageName: multicluster-engine
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Jooho Lee
      provider-url: https://github.com/jooho/nfs-provisioner-operator
    name: nfs-provisioner-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfs-provisioner-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.jhouse.com/v1alpha1",
                "kind": "NFSProvisioner",
                "metadata": {
                  "name": "nfsprovisioner-sample"
                },
                "spec": {
                  "nfsImageConfiguration": {
                    "image": "k8s.gcr.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439",
                    "imagePullPolicy": "IfNotPresent"
                  },
                  "scForNFS": "nfs",
                  "scForNFSPvc": "local-sc",
                  "storageSize": "1G"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: quay.io/jooholee/nfs-provisioner-operator@sha256:d9c013967421ec72644a588a155975bf856a856f1ef38ba71dfea306fdb47acd
          createdAt: "2025-08-08T20:23:50Z"
          description: Create and manage NFS Server and Provisioner
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/jooho/nfs-provisioner-operator
          support: NFS
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NFSProvisioner is the Schema for the nfsprovisioners API
            displayName: NFS Provisioner App
            kind: NFSProvisioner
            name: nfsprovisioners.cache.jhouse.com
            version: v1alpha1
        description: "This operator deploy NFS server with local storage and also
          provide provisioner for storageClass.\n### Core Capabilities\n* **NFS Server:**
          Deployed\n* **NFS Provisioner:** Help customers to create PV using StorageClass\n*
          **StorageClass:** Dynamically create PV for requested PVC\n### NFS Provisioner
          Operator Features\n* NFS Server can use localStorage PVC or HostPath on
          the node\n### Must-Gather \n> isv-cli must-gather --image=quay.io/jooholee/nfs-provisioner-operator-must-gather:0.0.1\n"
        displayName: NFS Provisioner Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nfs
        - storage
        links:
        - name: Nfs Provisioner Operator
          url: https://github.com/jooho/nfs-provisioner-operator
        maintainers:
        - email: ljhiyh@gmail.com
          name: Jooho
        maturity: alpha
        provider:
          name: Jooho Lee
          url: https://github.com/jooho/nfs-provisioner-operator
        relatedImages:
        - quay.io/jooholee/nfs-provisioner-operator@sha256:d9c013967421ec72644a588a155975bf856a856f1ef38ba71dfea306fdb47acd
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        version: 0.0.8
      entries:
      - name: nfs-provisioner-operator.v0.0.8
        version: 0.0.8
      - name: nfs-provisioner-operator.v0.0.7
        version: 0.0.7
      - name: nfs-provisioner-operator.v0.0.6
        version: 0.0.6
      - name: nfs-provisioner-operator.v0.0.5
        version: 0.0.5
      - name: nfs-provisioner-operator.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: nfs-provisioner-operator
    provider:
      name: Jooho Lee
      url: https://github.com/jooho/nfs-provisioner-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator-certified.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator-certified.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator-certified.v2.6.1
        version: 2.6.1
      - name: percona-postgresql-operator-certified.v2.6.0
        version: 2.6.0
      - name: percona-postgresql-operator-certified.v2.5.0
        version: 2.5.0
      name: stable
    - currentCSV: percona-postgresql-operator-certified.v2.7.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        version: 2.7.0-cw
      entries:
      - name: percona-postgresql-operator-certified.v2.7.0-cw
        version: 2.7.0-cw
      - name: percona-postgresql-operator-certified.v2.6.1-cw
        version: 2.6.1-cw
      - name: percona-postgresql-operator-certified.v2.6.0-cw
        version: 2.6.0-cw
      - name: percona-postgresql-operator-certified.v2.5.0-cw
        version: 2.5.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-postgresql-operator-certified
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Starburst Data, Inc.
      provider-url: https://www.starburstdata.com
    name: starburst-enterprise-helm-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: starburst-enterprise-helm-operator.v402.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstEnterprise",
                "metadata": {
                  "name": "starburstenterprise-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "cache": {
                    "diskUsagePercentage": 80,
                    "enabled": false,
                    "ttl": "7d",
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "catalogs": {
                    "tpch": "connector.name=tpch\ntpch.splits-per-node=4"
                  },
                  "commonLabels": {},
                  "coordinator": {
                    "additionalProperties": "",
                    "affinity": {},
                    "deploymentAnnotations": {},
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=true\nnode-scheduler.include-coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://localhost:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "resources": {
                      "cpu": 16,
                      "memory": "60Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "tolerations": []
                  },
                  "environment": null,
                  "expose": {
                    "clusterIp": {
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      }
                    },
                    "ingress": {
                      "annotations": {},
                      "host": null,
                      "ingressClassName": null,
                      "ingressName": "coordinator-ingress",
                      "path": "/",
                      "pathType": "ImplementationSpecific",
                      "serviceName": "starburst",
                      "servicePort": 8080,
                      "tls": {
                        "enabled": true,
                        "secretName": null
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "nodePort": 30080,
                          "port": 8080
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "externalSecrets": {
                    "enabled": false,
                    "goDaddy": {
                      "backendType": "secretsManager"
                    },
                    "secretPrefix": "external0",
                    "type": "goDaddy"
                  },
                  "extraArguments": [],
                  "extraSecret": {
                    "file": null,
                    "name": null
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise",
                    "tag": "402-e.1"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "initImage": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise-init",
                    "tag": "1.3.2"
                  },
                  "internal": {
                    "ports": {
                      "http": {
                        "port": 8080
                      },
                      "https": {
                        "port": 8443
                      }
                    }
                  },
                  "internalTls": false,
                  "licenseManager": {
                    "cronjob": {
                      "affinity": {},
                      "nodeSelector": {},
                      "priorityClassName": null,
                      "securityContext": {},
                      "tolerations": []
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "registry.connect.redhat.com/starburst/license-manager",
                      "tag": "0.0.2"
                    },
                    "url": "https://license-api.dogfood.eng.starburstdata.net/test/license/checkout"
                  },
                  "livenessProbe": null,
                  "prometheus": {
                    "agent": {
                      "config": "/etc/starburst/telemetry/prometheus.yaml",
                      "port": 8081,
                      "version": "0.16.1"
                    },
                    "enabled": true,
                    "rules": [
                      {
                        "attrNameSnakeCase": true,
                        "name": "$1",
                        "pattern": "trino.execution<name=QueryManager><>(running_queries|queued_queries)",
                        "type": "GAUGE"
                      },
                      {
                        "name": "failed_queries",
                        "pattern": "trino.execution<name=QueryManager><>FailedQueries\\.TotalCount",
                        "type": "COUNTER"
                      }
                    ]
                  },
                  "query": {
                    "maxConcurrentQueries": 3
                  },
                  "readinessProbe": null,
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "securityContext": {},
                  "sharedSecret": null,
                  "siac": {
                    "additionalHeapSizePercentage": 15,
                    "enabled": false
                  },
                  "spilling": {
                    "enabled": false,
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "startupProbe": null,
                  "usageMetrics": {
                    "enabled": false,
                    "usageClient": {
                      "initialDelay": "1m",
                      "interval": "1m"
                    }
                  },
                  "userDatabase": {
                    "enabled": false,
                    "users": [
                      {
                        "password": "46991b33f7a75ff79213c0dc0e610610",
                        "username": "admin"
                      }
                    ]
                  },
                  "worker": {
                    "additionalProperties": "",
                    "affinity": {},
                    "autoscaling": {
                      "enabled": false,
                      "maxReplicas": 100,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "deploymentAnnotations": {},
                    "deploymentTerminationGracePeriodSeconds": 300,
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://coordinator:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "kedaScaler": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.connect.redhat.com/starburst/keda-trino-scaler",
                        "tag": "0.1.7"
                      },
                      "port": 8021,
                      "scaledObjectSpec": {
                        "advanced": {},
                        "cooldownPeriod": 300,
                        "idleReplicaCount": 0,
                        "maxReplicaCount": 100,
                        "minReplicaCount": 1,
                        "pollingInterval": 30,
                        "scaleTargetRef": {
                          "name": "worker"
                        },
                        "triggers": [
                          {
                            "metadata": {
                              "numberOfQueriesPerWorker": "10",
                              "scaleInToIdleReplicaCountIfNoQueuedQueriesLeft": "true",
                              "scaleMethod": "query_queue"
                            },
                            "type": "external"
                          }
                        ]
                      }
                    },
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "replicas": 2,
                    "resources": {
                      "cpu": 16,
                      "memory": "100Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "starburstWorkerShutdownGracePeriodSeconds": 120,
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstHive",
                "metadata": {
                  "name": "starbursthive-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "affinity": {},
                  "commonLabels": {},
                  "database": {
                    "external": {
                      "driver": null,
                      "jdbcUrl": null,
                      "password": null,
                      "user": null
                    },
                    "internal": {
                      "databaseName": "hive",
                      "databasePassword": "HivePass1234",
                      "databaseUser": "hive",
                      "driver": "org.postgresql.Driver",
                      "env": [],
                      "envFrom": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.redhat.io/rhel8/postgresql-10",
                        "tag": "1"
                      },
                      "port": 5432,
                      "resources": {
                        "limits": {
                          "cpu": 2,
                          "memory": "1Gi"
                        },
                        "requests": {
                          "cpu": 2,
                          "memory": "1Gi"
                        }
                      },
                      "volume": {
                        "emptyDir": {}
                      }
                    },
                    "type": "internal"
                  },
                  "deploymentAnnotations": {},
                  "env": [],
                  "envFrom": [],
                  "expose": {
                    "clusterIp": {
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "nodePort": 30083,
                          "port": 9083
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "extraArguments": [],
                  "hdfs": {
                    "hadoopUserName": null
                  },
                  "heapSizePercentage": 85,
                  "hiveMetastoreStorageAuthorization": true,
                  "hiveMetastoreWarehouseDir": null,
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/hive",
                    "tag": "3.1.2-e.22"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "nodeSelector": {},
                  "objectStorage": {
                    "adl": {
                      "oauth2": {
                        "clientId": null,
                        "credential": null,
                        "refreshUrl": null
                      }
                    },
                    "awsS3": {
                      "accessKey": null,
                      "endpoint": null,
                      "pathStyleAccess": false,
                      "region": null,
                      "secretKey": null
                    },
                    "azure": {
                      "abfs": {
                        "accessKey": {
                          "accessKey": null,
                          "storageAccount": null
                        },
                        "authType": "accessKey",
                        "oauth": {
                          "clientId": null,
                          "endpoint": null,
                          "secret": null
                        }
                      },
                      "wasb": {
                        "accessKey": null,
                        "storageAccount": null
                      }
                    },
                    "gs": {
                      "cloudKeyFileSecret": null
                    }
                  },
                  "podAnnotations": {},
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "resources": {
                    "limits": {
                      "cpu": 1,
                      "memory": "1Gi"
                    },
                    "requests": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "securityContext": {},
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
          description: Deploys StarburstEnterprise Helm Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Starburst Data
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: StarburstEnterprise
            name: starburstenterprises.charts.starburstdata.com
            version: v1alpha1
          - kind: StarburstHive
            name: starbursthives.charts.starburstdata.com
            version: v1alpha1
        description: CRs for deploying starburst enterprise and hive
        displayName: Starburst Enterprise Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - starburst
        - trino
        - sep
        - bigdata
        - data
        links:
        - name: Starburst Enterprise Helm Operator
          url: https://www.starburstdata.com
        maintainers:
        - email: hello@starburstdata.com
          name: Starburst Data, Inc.
        maturity: alpha
        provider:
          name: Starburst Data, Inc.
          url: https://www.starburstdata.com
        relatedImages:
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
        version: 402.1.0
      entries:
      - name: starburst-enterprise-helm-operator.v402.1.0
        version: 402.1.0
      name: alpha
    defaultChannel: alpha
    packageName: starburst-enterprise-helm-operator-rhmp
    provider:
      name: Starburst Data, Inc.
      url: https://www.starburstdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: keystone-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keystone-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keystone.openstack.org/v1beta1",
                "kind": "KeystoneAPI",
                "metadata": {
                  "name": "keystone"
                },
                "spec": {
                  "adminProject": "admin",
                  "adminUser": "admin",
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "keystone",
                  "databaseInstance": "openstack",
                  "preserveJobs": false,
                  "region": "regionOne",
                  "resources": {
                    "requests": {
                      "cpu": "1.0",
                      "memory": "500Mi"
                    }
                  },
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "keystone.openstack.org/v1beta1",
                "kind": "KeystoneEndpoint",
                "metadata": {
                  "name": "placement"
                },
                "spec": {
                  "endpoints": {
                    "admin": "http://placement-admin-openstack.apps-crc.testing",
                    "internal": "http://placement-internal-openstack.apps-crc.testing",
                    "public": "http://placement-public-openstack.apps-crc.testing"
                  },
                  "serviceName": "placement"
                }
              },
              {
                "apiVersion": "keystone.openstack.org/v1beta1",
                "kind": "KeystoneService",
                "metadata": {
                  "name": "placement"
                },
                "spec": {
                  "enabled": true,
                  "passwordSelector": "PlacementPassword",
                  "secret": "osp-secret",
                  "serviceDescription": "Placement service",
                  "serviceName": "placement",
                  "serviceType": "placement",
                  "serviceUser": "placement"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:01Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KeystoneAPI is the Schema for the keystoneapis API
            displayName: Keystone API
            kind: KeystoneAPI
            name: keystoneapis.keystone.openstack.org
            version: v1beta1
          - description: KeystoneEndpoint is the Schema for the keystoneendpoints
              API
            displayName: Keystone Endpoint
            kind: KeystoneEndpoint
            name: keystoneendpoints.keystone.openstack.org
            version: v1beta1
          - description: KeystoneService is the Schema for the keystoneservices API
            displayName: Keystone Service
            kind: KeystoneService
            name: keystoneservices.keystone.openstack.org
            version: v1beta1
        description: Keystone Operator
        displayName: Keystone Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Identity
        - Keystone
        links:
        - name: Keystone Operator
          url: https://github.com/openstack-k8s-operators/keystone-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-keystone-rhel9@sha256:740e92a78d1e3874d2a40e37c970c6163d9e295dbc236914f1204baed070c595
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/keystone-rhel9-operator@sha256:8b055945a3808f7397265a10e6a6463af0ee3ef80fdff9d66402b488e5b70523
        version: 1.0.6
      entries:
      - name: keystone-operator.v1.0.6
        version: 1.0.6
      - name: keystone-operator.v1.0.5
        version: 1.0.5
      - name: keystone-operator.v1.0.4
        version: 1.0.4
      - name: keystone-operator.v1.0.3
        version: 1.0.3
      - name: keystone-operator.v1.0.2
        version: 1.0.2
      - name: keystone-operator.v1.0.1
        version: 1.0.1
      - name: keystone-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: keystone-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: ""
    name: ingress-node-firewall
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ingress-node-firewall.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ingressnodefirewall.openshift.io/v1alpha1",
                "kind": "IngressNodeFirewall",
                "metadata": {
                  "name": "ingressnodefirewall-demo2"
                },
                "spec": {
                  "ingress": [
                    {
                      "rules": [
                        {
                          "action": "Deny",
                          "order": 10,
                          "protocolConfig": {
                            "icmp": {
                              "icmpType": 8
                            },
                            "protocol": "ICMP"
                          }
                        },
                        {
                          "action": "Deny",
                          "order": 20,
                          "protocolConfig": {
                            "protocol": "TCP",
                            "tcp": {
                              "ports": "8000-9000"
                            }
                          }
                        }
                      ],
                      "sourceCIDRs": [
                        "172.16.0.0/12"
                      ]
                    },
                    {
                      "rules": [
                        {
                          "action": "Deny",
                          "order": 10,
                          "protocolConfig": {
                            "icmpv6": {
                              "icmpType": 128
                            },
                            "protocol": "ICMPv6"
                          }
                        }
                      ],
                      "sourceCIDRs": [
                        "fc00:f853:ccd:e793::0/64"
                      ]
                    }
                  ],
                  "interfaces": [
                    "eth0"
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "do-node-ingress-firewall": "true"
                    }
                  }
                }
              },
              {
                "apiVersion": "ingressnodefirewall.openshift.io/v1alpha1",
                "kind": "IngressNodeFirewallConfig",
                "metadata": {
                  "name": "ingressnodefirewallconfig",
                  "namespace": "ingress-node-firewall-system"
                },
                "spec": {
                  "nodeSelector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ingress-node-firewall-rhel9-operator@sha256:144403b009d1b5f2b2ee25ed3cebc525f3588f21bfa671b25c662cf800798839
          createdAt: "2024-03-13T12:36:01Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.11.0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-ingress-node-firewall
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/ingress-node-firewall
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: IngressNodeFirewallConfig
            name: ingressnodefirewallconfigs.ingressnodefirewall.openshift.io
            version: v1alpha1
          - kind: IngressNodeFirewallNodeState
            name: ingressnodefirewallnodestates.ingressnodefirewall.openshift.io
            version: v1alpha1
          - kind: IngressNodeFirewall
            name: ingressnodefirewalls.ingressnodefirewall.openshift.io
            version: v1alpha1
        description: The Ingress node firewall operator, manages the configuration
          of cluster scoped Ingress Node firewall rules for Kubernetes.
        displayName: Ingress Node Firewall Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ingressnodefirewall
        links:
        - name: Ingress Node Firewall
          url: https://github.com/openshift/ingress-node-firewall
        maintainers:
        - email: mmahmoud@redhat.com
          name: msherif1234
        maturity: alpha
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ingress-node-firewall-rhel9@sha256:4e3d69793f9280f6910a6749e80e8f6fe5d20c7a2c160d5682f310309c497e64
        - registry.redhat.io/openshift4/ingress-node-firewall-rhel9-operator@sha256:144403b009d1b5f2b2ee25ed3cebc525f3588f21bfa671b25c662cf800798839
        version: 4.16.0-202509240123
      entries:
      - name: ingress-node-firewall.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: ingress-node-firewall.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: ingress-node-firewall.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: ingress-node-firewall.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: ingress-node-firewall.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: ingress-node-firewall.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: ingress-node-firewall.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: ingress-node-firewall.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: ingress-node-firewall.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: ingress-node-firewall.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: ingress-node-firewall.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: ingress-node-firewall.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: ingress-node-firewall.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: ingress-node-firewall.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: ingress-node-firewall.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: ingress-node-firewall.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: ingress-node-firewall.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: ingress-node-firewall.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: ingress-node-firewall.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: ingress-node-firewall.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: ingress-node-firewall.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: ingress-node-firewall.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: ingress-node-firewall.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: ingress-node-firewall.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: ingress-node-firewall.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: ingress-node-firewall.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: ingress-node-firewall.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: ingress-node-firewall.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: ingress-node-firewall.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: ingress-node-firewall.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: ingress-node-firewall.v4.16.0-202405281506
        version: 4.16.0-202405281506
      name: stable
    defaultChannel: stable
    packageName: ingress-node-firewall
    provider:
      name: Red Hat, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cluster-observability-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-observability-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e
          createdAt: "2025-06-24T11:30:26Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=..0 <1.2.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-cluster-observability-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs",
              "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs",
              "prometheusagents.monitoring.rhobs",
              "perses.perses.dev"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
          support: Cluster Observability (https://issues.redhat.com/projects/COO/)
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: Perses is the Schema for the perses API
            displayName: Perses
            kind: Perses
            name: perses.perses.dev
            version: v1alpha1
          - description: A Perses Dashboard
            displayName: Perses Dashboard
            kind: PersesDashboard
            name: persesdashboards.perses.dev
            version: v1alpha1
          - description: A Perses Datasource
            displayName: Perses Datasource
            kind: PersesDatasource
            name: persesdatasources.perses.dev
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
          - description: UIPlugin defines a console plugin for observability
            displayName: UIPlugin
            kind: UIPlugin
            name: uiplugins.observability.openshift.io
            version: v1alpha1
        description: |-
          Cluster Observability Operator is a Go based Kubernetes operator to easily setup and manage various observability tools.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          - Setup console plugins
          - Setup korrel8r
          - Setup Perses
          - Setup Cluster Health Analyzer
          ### Documentation
          - **[Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/cluster_observability_operator/index)**
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)
        displayName: Cluster Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad
        - registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf
        - registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd
        - registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a
        - registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552
        - registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:e2681bce57dc9c15701f5591532c2dfe8f19778606661339553a28dc003dbca5
        - registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e
        - registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:e54c1e1301be66933f3ecb01d5a0ca27f58aabfd905b18b7d057bbf23bdb7b0d
        - registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:8597c48fc71fc6ec8e87dbe40dace4dbb7b817c1039db608af76a0d90f7ac2d0
        - registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404
        - registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223
        - registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00
        - registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1
        - registry.redhat.io/cluster-observability-operator/perses-0-1-rhel9-operator@sha256:bfed9f442aea6e8165644f1dc615beea06ec7fd84ea3f8ca393a63d3627c6a7c
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed
        version: 1.2.2
      entries:
      - name: cluster-observability-operator.v1.2.2
        version: 1.2.2
      - name: cluster-observability-operator.v1.2.1
        version: 1.2.1
      - name: cluster-observability-operator.v1.2.0
        version: 1.2.0
      - name: cluster-observability-operator.v1.1.1
        version: 1.1.1
      - name: cluster-observability-operator.v1.1.0
        version: 1.1.0
      - name: cluster-observability-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: cluster-observability-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JohnSnowLabs
      provider-url: ""
    name: nlp-server
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nlp-server.v0.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.johnsnowlabs.com/v1",
                "kind": "Nlpserver",
                "metadata": {
                  "name": "nlpserver-sample"
                },
                "spec": {
                  "ExtraEnv": {
                    "johnsnowlabs_client_id": "sI4MKSmLHOX2Pg7XhM3McJS2oyKG5PHcp0BlANEW",
                    "nb_uid": "1000",
                    "nb_user": "johnsnowlabs",
                    "pyspark_version": "3.0.2",
                    "spark_nlp_version": "3.4.4"
                  },
                  "affinity": {},
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "johnsnowlabs/nlp-server",
                    "tag": "0.8.0rc1"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {
                      "kubernetes.io/ingress.class": "nginx",
                      "nginx.ingress.kubernetes.io/ssl-redirect": "true"
                    },
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "nlp-server.jsl.aiplatform.eks.test.johnsnowlabs.com",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": [
                      {
                        "hosts": [
                          "nlp-server.jsl.aiplatform.eks.test.johnsnowlabs.com"
                        ],
                        "secretName": "annotationlab-tls-secret"
                      }
                    ]
                  },
                  "livenessProbe": "httpGet:\n  path: /\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 3\n",
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "readinessProbe": "httpGet:\n  path: /\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 3\n",
                  "replicaCount": 1,
                  "resources": {},
                  "securityContext": {},
                  "service": {
                    "port": 5000,
                    "portName": "http",
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": "nlp-server-sa"
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.johnsnowlabs.com/nlp-server/nlp-server-operator:0.8.6
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nlpserver
            name: nlpservers.apps.johnsnowlabs.com
            version: v1
        description: NLPServer from JohnSnow labs
        displayName: nlp-server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nlp-server
        links:
        - name: Nlp Server
          url: https://johnsnowlabs.com
        maturity: alpha
        provider:
          name: JohnSnowLabs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - registry.johnsnowlabs.com/nlp-server/nlp-server-operator:0.8.6
        version: 0.8.6
      entries:
      - name: nlp-server.v0.8.6
        version: 0.8.6
      name: alpha
    defaultChannel: alpha
    packageName: nlp-server
    provider:
      name: JohnSnowLabs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Seldon Technologies
      provider-url: ""
    name: seldon-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: seldon-operator.v1.14.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1",
                "kind": "SeldonDeployment",
                "metadata": {
                  "labels": {
                    "app": "seldon",
                    "app.kubernetes.io/instance": "seldon1",
                    "app.kubernetes.io/name": "seldon",
                    "app.kubernetes.io/version": "v0.5"
                  },
                  "name": "seldon-model"
                },
                "spec": {
                  "name": "test-deployment",
                  "predictors": [
                    {
                      "componentSpecs": [
                        {
                          "spec": {
                            "containers": [
                              {
                                "image": "seldonio/mock_classifier:1.6.0",
                                "name": "classifier"
                              }
                            ]
                          }
                        }
                      ],
                      "graph": {
                        "children": [],
                        "name": "classifier",
                        "type": "MODEL"
                      },
                      "name": "example",
                      "replicas": 1
                    }
                  ]
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: docker.io/seldonio/seldon-core-operator:1.14.1
          createdAt: "2019-05-21 15:00:00"
          description: The Seldon operator for management, monitoring and operations
            of machine learning systems through the Seldon Engine. Once installed,
            the Seldon Operator provides multiple functions which facilitate the productisation,
            monitoring and maintenance of machine learning systems at scale.
          olm.skipRange: <1.13.1
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/SeldonIO/seldon-core
          support: Clive Cox
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A seldon engine deployment
            displayName: Seldon Deployment
            kind: SeldonDeployment
            name: seldondeployments.machinelearning.seldon.io
            version: v1
        description: "The Seldon operator enables for native operation of production
          machine learning workloads, including monitoring and operations of language-agnostic
          models with the benefits of real-time metrics and log analysis.\n   \n##
          Overview\nSeldon Core is an open source platform for deploying machine learning
          models on a Kubernetes cluster.\n\n* Deploy machine learning models in the
          cloud or on-premise.\n* Get metrics and ensure proper governance and compliance
          for your running machine learning models.\n* Create powerful inference graphs
          made up of multiple components.\n* Provide a consistent serving layer for
          models built using heterogeneous ML toolkits.\n\nYou can get started by
          following the guides in our documentation at https://docs.seldon.io/projects/seldon-core/en/latest/workflow/README.html\n"
        displayName: Seldon Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - aiops
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://docs.seldon.io/projects/seldon-core/en/latest/
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Seldon Technologies
        relatedImages:
        - docker.io/seldonio/seldon-core-operator:1.14.1
        - ""
        version: 1.14.1
      entries:
      - name: seldon-operator.v1.14.1
        version: 1.14.1
      name: stable
    defaultChannel: stable
    packageName: seldon-operator
    provider:
      name: Seldon Technologies
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: keycloak-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keycloak-operator.v19.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "False"
          containerImage: quay.io/keycloak/keycloak-operator:19.0.3-legacy
          createdAt: "2022-10-07 17:00:00"
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
          * Install Extensions
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/docs/latest/server_installation/index.html#_operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@lists.jboss.org
          name: Keycloak DEV mailing list
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/keycloak/keycloak-operator:19.0.3-legacy
        version: 19.0.3
      entries:
      - name: keycloak-operator.v19.0.3
        version: 19.0.3
      - name: keycloak-operator.v19.0.2
        version: 19.0.2
      - name: keycloak-operator.v19.0.1
        version: 19.0.1
      - name: keycloak-operator.v19.0.0
        version: 19.0.0
      - name: keycloak-operator.v18.0.2
        version: 18.0.2
      - name: keycloak-operator.v18.0.1
        version: 18.0.1
      - name: keycloak-operator.v18.0.0
        version: 18.0.0
      - name: keycloak-operator.v17.0.1
        version: 17.0.1
      - name: keycloak-operator.v17.0.0
        version: 17.0.0
      - name: keycloak-operator.v16.1.0
        version: 16.1.0
      - name: keycloak-operator.v16.0.0
        version: 16.0.0
      - name: keycloak-operator.v15.1.1
        version: 15.1.1
      - name: keycloak-operator.v15.1.0
        version: 15.1.0
      - name: keycloak-operator.v15.0.2
        version: 15.0.2
      - name: keycloak-operator.v15.0.1
        version: 15.0.1
      - name: keycloak-operator.v15.0.0
        version: 15.0.0
      - name: keycloak-operator.v14.0.0
        version: 14.0.0
      name: alpha
    - currentCSV: keycloak-operator.v20.0.0-alpha.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "False"
          containerImage: quay.io/keycloak/keycloak-operator:19.0.3
          createdAt: 10/07/22 17:00:00
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/docs/latest/server_installation/index.html#_operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@lists.jboss.org
          name: Keycloak DEV mailing list
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/keycloak/keycloak-operator:19.0.3
        version: 20.0.0-alpha.7
      entries:
      - name: keycloak-operator.v20.0.0-alpha.7
        version: 20.0.0-alpha.7
      - name: keycloak-operator.v20.0.0-alpha.6
        version: 20.0.0-alpha.6
      - name: keycloak-operator.v20.0.0-alpha.5
        version: 20.0.0-alpha.5
      - name: keycloak-operator.v20.0.0-alpha.4
        version: 20.0.0-alpha.4
      - name: keycloak-operator.v20.0.0-alpha.3
        version: 20.0.0-alpha.3
      - name: keycloak-operator.v20.0.0-alpha.2
        version: 20.0.0-alpha.2
      - name: keycloak-operator.v20.0.0-alpha.1
        version: 20.0.0-alpha.1
      name: candidate
    - currentCSV: keycloak-operator.v26.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/keycloak/keycloak-operator:26.4.0
          createdAt: "2025-09-30T11:58:06Z"
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/keycloak/keycloak-operator:26.4.0
        version: 26.4.0
      entries:
      - name: keycloak-operator.v26.4.0
        version: 26.4.0
      - name: keycloak-operator.v26.3.5
        version: 26.3.5
      - name: keycloak-operator.v26.3.4
        version: 26.3.4
      - name: keycloak-operator.v26.3.3
        version: 26.3.3
      - name: keycloak-operator.v26.3.2
        version: 26.3.2
      - name: keycloak-operator.v26.3.1
        version: 26.3.1
      - name: keycloak-operator.v26.3.0
        version: 26.3.0
      - name: keycloak-operator.v26.2.5
        version: 26.2.5
      - name: keycloak-operator.v26.2.4
        version: 26.2.4
      - name: keycloak-operator.v26.2.3
        version: 26.2.3
      - name: keycloak-operator.v26.2.2
        version: 26.2.2
      - name: keycloak-operator.v26.2.1
        version: 26.2.1
      - name: keycloak-operator.v26.2.0
        version: 26.2.0
      - name: keycloak-operator.v26.1.5
        version: 26.1.5
      - name: keycloak-operator.v26.1.4
        version: 26.1.4
      - name: keycloak-operator.v26.1.3
        version: 26.1.3
      - name: keycloak-operator.v26.1.2
        version: 26.1.2
      - name: keycloak-operator.v26.1.1
        version: 26.1.1
      - name: keycloak-operator.v26.1.0
        version: 26.1.0
      - name: keycloak-operator.v26.0.8
        version: 26.0.8
      - name: keycloak-operator.v26.0.7
        version: 26.0.7
      - name: keycloak-operator.v26.0.6
        version: 26.0.6
      - name: keycloak-operator.v26.0.5
        version: 26.0.5
      - name: keycloak-operator.v26.0.4
        version: 26.0.4
      - name: keycloak-operator.v26.0.2
        version: 26.0.2
      - name: keycloak-operator.v26.0.1
        version: 26.0.1
      - name: keycloak-operator.v26.0.0
        version: 26.0.0
      - name: keycloak-operator.v25.0.6
        version: 25.0.6
      - name: keycloak-operator.v25.0.5
        version: 25.0.5
      - name: keycloak-operator.v25.0.4
        version: 25.0.4
      - name: keycloak-operator.v25.0.2
        version: 25.0.2
      - name: keycloak-operator.v25.0.1
        version: 25.0.1
      - name: keycloak-operator.v25.0.0
        version: 25.0.0
      - name: keycloak-operator.v24.0.5
        version: 24.0.5
      - name: keycloak-operator.v24.0.4
        version: 24.0.4
      - name: keycloak-operator.v24.0.3
        version: 24.0.3
      - name: keycloak-operator.v24.0.2
        version: 24.0.2
      - name: keycloak-operator.v24.0.1
        version: 24.0.1
      - name: keycloak-operator.v24.0.0
        version: 24.0.0
      - name: keycloak-operator.v23.0.7
        version: 23.0.7
      - name: keycloak-operator.v23.0.6
        version: 23.0.6
      - name: keycloak-operator.v23.0.5
        version: 23.0.5
      - name: keycloak-operator.v23.0.4
        version: 23.0.4
      - name: keycloak-operator.v23.0.3
        version: 23.0.3
      - name: keycloak-operator.v23.0.2
        version: 23.0.2
      - name: keycloak-operator.v23.0.1
        version: 23.0.1
      - name: keycloak-operator.v23.0.0
        version: 23.0.0
      - name: keycloak-operator.v22.0.5
        version: 22.0.5
      - name: keycloak-operator.v22.0.4
        version: 22.0.4
      - name: keycloak-operator.v22.0.3
        version: 22.0.3
      - name: keycloak-operator.v22.0.2
        version: 22.0.2
      - name: keycloak-operator.v22.0.1
        version: 22.0.1
      - name: keycloak-operator.v22.0.0
        version: 22.0.0
      - name: keycloak-operator.v21.1.2
        version: 21.1.2
      - name: keycloak-operator.v21.1.1
        version: 21.1.1
      - name: keycloak-operator.v21.1.0
        version: 21.1.0
      - name: keycloak-operator.v21.0.2
        version: 21.0.2
      - name: keycloak-operator.v21.0.1
        version: 21.0.1
      - name: keycloak-operator.v21.0.0
        version: 21.0.0
      - name: keycloak-operator.v20.0.3
        version: 20.0.3
      - name: keycloak-operator.v20.0.2
        version: 20.0.2
      - name: keycloak-operator.v20.0.1
        version: 20.0.1
      - name: keycloak-operator.v20.0.0
        version: 20.0.0
      name: fast
    defaultChannel: fast
    packageName: keycloak-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: redis-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"Redis","metadata":{"name":"redis"},"spec":{"mode":"cluster","size":3,"global":{"image":"quay.io/opstree/redis:v6.2","imagePullPolicy":"IfNotPresent","password":"Opstree@1234","resources":{"requests":{"cpu":"100m","memory":"128Mi"},"limits":{"cpu":"100m","memory":"128Mi"}}},"master":{"service":{"type":"ClusterIP"},"redisConfig":{}},"slave":{"service":{"type":"ClusterIP"},"redisConfig":{}},"service":{"type":"ClusterIP"},"redisConfig":{},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:1.0","imagePullPolicy":"Always","resources":{"requests":{"cpu":"100m","memory":"128Mi"},"limits":{"cpu":"100m","memory":"128Mi"}}},"storage":{"volumeClaimTemplate":{"spec":{"storageClassName":"buildpiper-storage","accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}}}}]'
          capabilities: Seamless Upgrades
          categories: Database, Storage
          certified: "false"
          containerImage: quay.io/opstree/redis-operator:v0.4.0
          createdAt: 06-02-2021
          description: A Golang based redis operator that will make/oversee Redis
            standalone/cluster mode setup on top of the Kubernetes.
          repository: https://github.com/ot-container-kit/redis-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Redis
            displayName: Redis
            kind: Redis
            name: redis.redis.redis.opstreelabs.in
            version: v1beta1
        description: |+
          A Golang based redis operator that will make/oversee Redis standalone/cluster mode setup on top of the Kubernetes. It can create a redis cluster setup with best practices on Cloud as well as the Bare metal environment. Also, it provides an in-built monitoring capability using redis-exporter.

          ### Supported Features

          Here the features which are supported by this operator:-
          - Redis cluster/standalone mode setup
          - Inbuilt monitoring with prometheus exporter
          - Dynamic storage provisioning with pvc template
          - Resources restrictions with k8s requests and limits
          - Password/Password-less setup
          - Node selector and affinity
          - Priority class to manage setup priority
          - SecurityContext to manipulate kernel parameters
          ### Documentation

          [Documentation](https://docs.opstreelabs.in/redis-operator/)

          ### Getting Help

          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.

          ### Contributing

          You can contribute to this project by:-
          - Raising Bugs and Feature Requests
          - Fixing issue by opening a Pull Request
          - Improving Documentation

          ### License

          Logging Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/redis-operator/blob/master/LICENSE)


        displayName: Redis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - Redis
        - Opstree
        - CRD
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://docs.opstreelabs.in/redis-operator/
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: beta
        minKubeVersion: 1.11.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/opstree/redis-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: redis-operator.v0.4.0
        version: 0.4.0
      - name: redis-operator.v0.3.0
        version: 0.3.0
      name: preview
    - currentCSV: redis-operator.v0.15.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisCluster","metadata":{"name":"redis-cluster"},"spec":{"clusterSize":3,"clusterVersion":"v7","securityContext":{"runAsUser":1000,"fsGroup":1000},"persistenceEnabled":true,"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"Redis","metadata":{"name":"redis-standalone"},"spec":{"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"IfNotPresent"},"securityContext":{"runAsUser":1000,"fsGroup":1000}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisSentinel","metadata":{"name":"redis-sentinel"},"spec":{"clusterSize":3,"podSecurityContext":{"runAsUser":1000,"fsGroup":1000},"redisSentinelConfig":{"redisReplicationName":"redis-replication"},"kubernetesConfig":{"image":"quay.io/opstree/redis-sentinel:v7.0.7","imagePullPolicy":"IfNotPresent","resources":{"requests":{"cpu":"101m","memory":"128Mi"},"limits":{"cpu":"101m","memory":"128Mi"}}}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisReplication","metadata":{"name":"redis-replication"},"spec":{"clusterSize":3,"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"Always"},"podSecurityContext":{"runAsUser":1000,"fsGroup":1000}}}]'
          capabilities: Seamless Upgrades
          categories: Database, Storage
          certified: "false"
          containerImage: quay.io/opstree/redis-operator:v0.15.1
          createdAt: 24-09-2023
          description: A Golang based redis operator that will make/oversee Redis
            standalone/cluster mode setup on top of the Kubernetes.
          repository: https://github.com/ot-container-kit/redis-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Redis Cluster
            displayName: RedisCluster
            kind: RedisCluster
            name: redisclusters.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis Replication
            displayName: RedisReplication
            kind: RedisReplication
            name: redisreplications.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis Sentinel
            displayName: RedisSentinel
            kind: RedisSentinel
            name: redissentinels.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis
            displayName: Redis
            kind: Redis
            name: redis.redis.redis.opstreelabs.in
            version: v1beta1
        description: "A Golang based redis operator that will make/oversee Redis\nstandalone/cluster
          mode setup on top of the Kubernetes. It can create a\nredis cluster setup
          with best practices on Cloud as well as the Bare metal\nenvironment. Also,
          it provides an in-built monitoring capability using\nredis-exporter.\n\nOrganizations
          that are using Redis Operator to manage their redis workload \ncan be found
          here. If your organization is also using Redis Operator, please \nfree to
          add by creating a PR https://github.com/OT-CONTAINER-KIT/redis-operator/pulls.\n\nThis
          operator only supports versions of redis =>6.\n\n### Supported Features\n\nHere
          the features which are supported by this operator:-\n- Redis cluster and
          standalone mode setup\n- Redis cluster failover and recovery\n- Redis leader
          and follower setup with Sentinel for High Availability\n- Inbuilt monitoring
          with redis exporter\n- Password and password-less setup of redis\n- TLS
          support for additional security layer\n- Ipv4 and Ipv6 support for redis
          setup\n- External configuration support for leader and follower\n- Detailed
          monitoring grafana dashboard\n\n### Documentation\n\n[Documentation](https://ot-redis-operator.netlify.app/)\n\n###
          Getting Help\nThis project is managed by OpsTree Solutions. If you have
          any queries or\nsuggestions, mail us at opensource@opstree.com.\n\n### Contributing\n\nYou
          can contribute to this project by:-\n- Raising Bugs and Feature Requests\n-
          Fixing issue by opening a Pull Request\n- Improving Documentation\n\n###
          License\n\nLogging Operator is licensed under [Apache License, Version\n2.0](https://github.com/OT-CONTAINER-KIT/redis-operator/blob/master/LICENSE)\n"
        displayName: Redis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - Redis
        - Opstree
        - CRD
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-redis-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/opstree/redis-operator:v0.15.1
        version: 0.15.1
      entries:
      - name: redis-operator.v0.15.1
        version: 0.15.1
      - name: redis-operator.v0.15.0
        version: 0.15.0
      - name: redis-operator.v0.14.0
        version: 0.14.0
      - name: redis-operator.v0.13.0
        version: 0.13.0
      - name: redis-operator.v0.8.0
        version: 0.8.0
      - name: redis-operator.v0.6.0
        version: 0.6.0
      - name: redis-operator.v0.5.0
        version: 0.5.0
      - name: redis-operator.v0.4.0
        version: 0.4.0
      - name: redis-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: redis-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Veeam Kasten
      provider-url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
    name: k10-kasten-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k10-kasten-operator-rhmp.v8.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apik10.kasten.io/v1alpha1",
                "kind": "K10",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/reconcile-period": "2m",
                    "helm.sdk.operatorframework.io/rollback-force": "false"
                  },
                  "name": "k10"
                },
                "spec": {
                  "auth": {
                    "basicAuth": {
                      "enabled": false,
                      "htpasswd": "",
                      "secretName": ""
                    },
                    "tokenAuth": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "persistence": {
                      "catalog": {
                        "size": ""
                      },
                      "storageClass": ""
                    }
                  },
                  "metering": {
                    "mode": ""
                  },
                  "route": {
                    "enabled": false,
                    "host": "",
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "apik10.kasten.io/v1alpha1",
                "kind": "K10restore",
                "metadata": {
                  "name": "k10restore"
                },
                "spec": {
                  "pointInTime": null,
                  "profileName": "",
                  "skipResource": null,
                  "sourceClusterID": ""
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          console.openshift.io/plugins: '["veeam-kasten-console-plugin"]'
          containerImage: registry.connect.redhat.com/kasten/k10-operator@sha256:4e8ccf38e47f92d2a308587591f85fb4f15aea1e7fd5c248803764c94eaecd4a
          createdAt: Wed Sep 24 02:11:59 PM UTC 2025
          description: Kasten data management platform for OpenShift, The Market Leader
            for Kubernetes Backup and DR
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/k10-kasten-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/k10-kasten-operator-rhmp/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: kasten-io
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: contact@kasten.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              k10restore can recover the K10 platform in case of a variety of disasters such as
              the accidental deletion of K10, failure of underlying storage that K10 uses for its catalog,
              or even the accidental destruction of the Kubernetes cluster on which K10 is deployed.

              **Before creating k10restore**: [K10 Disaster Recovery documentation](https://docs.kasten.io/latest/operating/dr.html#recovering-with-the-operator)
            kind: K10restore
            name: k10restores.apik10.kasten.io
            version: v1alpha1
          - description: |-
              [Kasten's k10](https://docs.kasten.io/) is a data lifecycle management system for all your persistence enabled container-based applications.

              **Note: The complete list of values supported by the K10 spec field in the "YAML View" are documented [here](https://docs.kasten.io/latest/install/advanced.html#complete-list-of-k10-helm-options).**
            kind: K10
            name: k10s.apik10.kasten.io
            version: v1alpha1
        description: |-
          Kasten data management platform for OpenShift, The Market Leader for Kubernetes Backup and DR.

          The supported Kubernetes versions are listed here - https://docs.kasten.io/latest/operating/support.html
          Please review release notes before an upgrade - https://docs.kasten.io/latest/releasenotes.html
        displayName: Veeam Kasten (Free)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kubernetes
        - storage
        - data-management
        links:
        - name: Kasten Documentation
          url: https://docs.kasten.io
        - name: Veeam Kasten
          url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
        maintainers:
        - email: contact@kasten.io
          name: Veeam Kasten
        maturity: stable
        provider:
          name: Veeam Kasten
          url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:503d5ee0fbcd5bd3492785f4246e19cbdb92960be0d05f3c6f123a32fe0fe84d
        - registry.connect.redhat.com/kasten/bloblifecyclemanager@sha256:7b209314b96bc8ebaa77e97426bb34176d05844cc3e8d4b462cfba5c4a8d3798
        - registry.connect.redhat.com/kasten/kanister-tools@sha256:7af2e37c9027f46c565901d35a3a7da4d9c7ddd2812628b67275d4acf006b80c
        - registry.connect.redhat.com/kasten/kanister@sha256:1d69ff91cec8e7c131dfbe1ad3e282aab854d3412f5f6e3b808d25eb7b8c38bc
        - registry.connect.redhat.com/kasten/metric-sidecar@sha256:20c142258de56cf8c60d4b0afa5e72b9317023ebf4ab109fb84e0d57fae854a4
        - registry.connect.redhat.com/kasten/controllermanager@sha256:ab3b815bb12cd3c263cbfdba3939a95097450f1864431cca7380827986d704ed
        - registry.connect.redhat.com/kasten/configmap-reload@sha256:01010f2bb580eb131d63d3271efe40eadd8269b84cdb02bdea36750d1359a43f
        - registry.connect.redhat.com/kasten/aggregatedapis@sha256:db868afd5c6d4a8d2e7c8376ed6943963c4be299fca6b26931f907430763d244
        - registry.connect.redhat.com/kasten/jobs@sha256:c6358878e981baee4ac72abcc1269c206ee014af2c16584d1b41fe18e0f0e57f
        - registry.connect.redhat.com/kasten/upgrade@sha256:bf2200697d38f3ee95ba2863dfbae4d982273d2fd8450c9ffbc3a85be7d6d87d
        - registry.connect.redhat.com/kasten/prometheus@sha256:29d913bb06c3245cca5798772dba7d7d5dc9b76ed2e0700f5eb4d20416e24fa2
        - registry.connect.redhat.com/kasten/k10tools@sha256:59caa1920c76a1bc86e7733731602e3f0fc08026d9a16000bb748503a4d9cd9b
        - registry.connect.redhat.com/kasten/garbagecollector@sha256:6f2972bb255f5a2de33ace5ce805b15d5e04d3c259c9c8d018c1257c74562878
        - registry.connect.redhat.com/kasten/events@sha256:7cae1c0909ea7a468e96484d99300677bdad49f34cc57c412ebaf342cb176264
        - registry.connect.redhat.com/kasten/logging@sha256:014bfdfc3620cc2fe6540b9993b9b9e7d70c9b969714477db4de91a90bdb4a38
        - registry.connect.redhat.com/kasten/k10-operator@sha256:4e8ccf38e47f92d2a308587591f85fb4f15aea1e7fd5c248803764c94eaecd4a
        - registry.connect.redhat.com/kasten/auth@sha256:f8543e67de3d5276c597c26f16053bf9bbcbd9cbdad4652214935a1e40a26817
        - registry.connect.redhat.com/kasten/frontend@sha256:a021964ed85133897e4956014b5c5013da3332a36a7e96a0de2e2e1b67888491
        - registry.connect.redhat.com/kasten/init@sha256:cc7fbbda1b6946e93dd083fb7092f38789f626c26ac68c226b61fe57064a8f11
        - registry.connect.redhat.com/kasten/dex@sha256:4c8522a27c0d4f0977baa074800bba96a9bd9aefc57473de283b1d95586f7517
        - registry.connect.redhat.com/kasten/catalog@sha256:11a77ac968634850c72cd10f2bde55afc632c68cdc0881aeae83c100288c6aec
        - registry.connect.redhat.com/kasten/dashboardbff@sha256:e1208c60146faa8c3c7ef73b2290663840276ed8d3b3c990f495cee2967104ee
        - registry.connect.redhat.com/kasten/state@sha256:191383c76c4ff22805c06217e9f32a2f41a1206afab5f16acec4a5dd56e4a6b4
        - registry.connect.redhat.com/kasten/repositories@sha256:924a607874e1807a877cb454e00539d64f4487e7435d4bfb524f3e569ad1f942
        - registry.connect.redhat.com/kasten/metering@sha256:f809696ac1309fc54fde87e5d2030267d266386fb73a56287868bbac433756d8
        - registry.connect.redhat.com/kasten/restorectl@sha256:9dc3c834bd3d4543115ae1ad586ee1248fc2b637dbc2d1e2c66ecc2a459db67e
        - registry.connect.redhat.com/kasten/datamover@sha256:c9e2dc5b646aa0c465a7b2daf011462273a6f5b53fdff772d879293208e2d57c
        - registry.connect.redhat.com/kasten/vbrintegrationapi@sha256:1a20b0da4efb7e802516224675b37751bd56d406e57b16216388d2b5167c648d
        - registry.connect.redhat.com/kasten/executor@sha256:32c4ac1aa2ca802ff77e913991949e6841149963863ad527a67f6bab6c94a5af
        - registry.connect.redhat.com/kasten/gateway@sha256:dc2d7bdeb8abaf9454470afd53af2cd5430505351a97b6501ff65a60dd2a6db1
        - registry.connect.redhat.com/kasten/ocpconsoleplugin@sha256:69f6bfb9421f4264dd7d066d809311022c0db57c41319a6242532f6acfbd2145
        - registry.connect.redhat.com/kasten/crypto@sha256:86bfbbf4c4f06bab1e4b091f9f312d536813e2c11aa899690b3277ff73110a09
        version: 8.0.9
      entries:
      - name: k10-kasten-operator-rhmp.v8.0.9
        version: 8.0.9
      - name: k10-kasten-operator-rhmp.v8.0.8
        version: 8.0.8
      - name: k10-kasten-operator-rhmp.v8.0.7
        version: 8.0.7
      - name: k10-kasten-operator-rhmp.v8.0.6
        version: 8.0.6
      - name: k10-kasten-operator-rhmp.v8.0.5
        version: 8.0.5
      - name: k10-kasten-operator-rhmp.v8.0.4
        version: 8.0.4
      - name: k10-kasten-operator-rhmp.v8.0.3
        version: 8.0.3
      - name: k10-kasten-operator-rhmp.v8.0.2
        version: 8.0.2
      - name: k10-kasten-operator-rhmp.v8.0.1
        version: 8.0.1
      - name: k10-kasten-operator-rhmp.v8.0.0
        version: 8.0.0
      - name: k10-kasten-operator-rhmp.v7.5.10
        version: 7.5.10
      - name: k10-kasten-operator-rhmp.v7.5.9
        version: 7.5.9
      - name: k10-kasten-operator-rhmp.v7.5.8
        version: 7.5.8
      - name: k10-kasten-operator-rhmp.v7.5.7
        version: 7.5.7
      - name: k10-kasten-operator-rhmp.v7.5.4
        version: 7.5.4
      - name: k10-kasten-operator-rhmp.v7.5.3
        version: 7.5.3
      - name: k10-kasten-operator-rhmp.v7.5.2
        version: 7.5.2
      - name: k10-kasten-operator-rhmp.v7.5.1
        version: 7.5.1
      - name: k10-kasten-operator-rhmp.v7.5.0
        version: 7.5.0
      - name: k10-kasten-operator-rhmp.v7.0.14
        version: 7.0.14
      - name: k10-kasten-operator-rhmp.v7.0.13
        version: 7.0.13
      - name: k10-kasten-operator-rhmp.v7.0.12
        version: 7.0.12
      - name: k10-kasten-operator-rhmp.v7.0.11
        version: 7.0.11
      - name: k10-kasten-operator-rhmp.v7.0.9
        version: 7.0.9
      - name: k10-kasten-operator-rhmp.v7.0.8
        version: 7.0.8
      - name: k10-kasten-operator-rhmp.v7.0.7
        version: 7.0.7
      - name: k10-kasten-operator-rhmp.v7.0.6
        version: 7.0.6
      - name: k10-kasten-operator-rhmp.v7.0.5
        version: 7.0.5
      - name: k10-kasten-operator-rhmp.v7.0.4
        version: 7.0.4
      - name: k10-kasten-operator-rhmp.v7.0.3
        version: 7.0.3
      - name: k10-kasten-operator-rhmp.v7.0.1
        version: 7.0.1
      - name: k10-kasten-operator-rhmp.v6.5.14
        version: 6.5.14
      - name: k10-kasten-operator-rhmp.v6.5.13
        version: 6.5.13
      - name: k10-kasten-operator-rhmp.v6.5.12
        version: 6.5.12
      - name: k10-kasten-operator-rhmp.v6.5.11
        version: 6.5.11
      - name: k10-kasten-operator-rhmp.v6.5.10
        version: 6.5.10
      - name: k10-kasten-operator-rhmp.v6.5.8
        version: 6.5.8
      - name: k10-kasten-operator-rhmp.v6.5.6
        version: 6.5.6
      - name: k10-kasten-operator-rhmp.v6.5.5
        version: 6.5.5
      - name: k10-kasten-operator-rhmp.v6.5.4
        version: 6.5.4
      - name: k10-kasten-operator-rhmp.v6.5.3
        version: 6.5.3
      - name: k10-kasten-operator-rhmp.v6.5.1
        version: 6.5.1
      - name: k10-kasten-operator-rhmp.v6.5.0
        version: 6.5.0
      - name: k10-kasten-operator-rhmp.v6.0.12
        version: 6.0.12
      - name: k10-kasten-operator-rhmp.v6.0.11
        version: 6.0.11
      - name: k10-kasten-operator-rhmp.v6.0.9
        version: 6.0.9
      - name: k10-kasten-operator-rhmp.v6.0.8
        version: 6.0.8
      - name: k10-kasten-operator-rhmp.v6.0.7
        version: 6.0.7
      - name: k10-kasten-operator-rhmp.v6.0.6
        version: 6.0.6
      - name: k10-kasten-operator-rhmp.v6.0.5
        version: 6.0.5
      - name: k10-kasten-operator-rhmp.v6.0.4
        version: 6.0.4
      - name: k10-kasten-operator-rhmp.v6.0.3
        version: 6.0.3
      - name: k10-kasten-operator-rhmp.v6.0.2
        version: 6.0.2
      - name: k10-kasten-operator-rhmp.v6.0.1
        version: 6.0.1
      - name: k10-kasten-operator-rhmp.v6.0.0
        version: 6.0.0
      name: stable
    defaultChannel: stable
    packageName: k10-kasten-operator-rhmp
    provider:
      name: Veeam Kasten
      url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kinesis-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kinesis-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kinesis.services.k8s.aws/v1alpha1",
                "kind": "Stream",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kinesis-controller:1.1.1
          createdAt: "2025-09-19T20:56:43Z"
          description: AWS Kinesis controller is a service controller for managing
            Kinesis resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Stream represents the state of an AWS kinesis Stream resource.
            displayName: Stream
            kind: Stream
            name: streams.kinesis.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Kinesis resources in AWS from within your Kubernetes cluster.

          **About Amazon Kinesis**

          You can use Amazon Kinesis Data Streams to collect and process large streams of data records in real time. You can create data-processing applications, known as Kinesis Data Streams applications. A typical Kinesis Data Streams application reads data from a data stream as data records. These applications can use the Kinesis Client Library, and they can run on Amazon EC2 instances.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Kinesis
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kinesis
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Kinesis Developer Resources
          url: https://aws.amazon.com/kinesis/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kinesis maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/kinesis-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-kinesis-controller.v1.1.1
        version: 1.1.1
      - name: ack-kinesis-controller.v1.1.0
        version: 1.1.0
      - name: ack-kinesis-controller.v1.0.16
        version: 1.0.16
      - name: ack-kinesis-controller.v1.0.15
        version: 1.0.15
      - name: ack-kinesis-controller.v1.0.14
        version: 1.0.14
      - name: ack-kinesis-controller.v1.0.13
        version: 1.0.13
      - name: ack-kinesis-controller.v1.0.12
        version: 1.0.12
      - name: ack-kinesis-controller.v1.0.11
        version: 1.0.11
      - name: ack-kinesis-controller.v1.0.10
        version: 1.0.10
      - name: ack-kinesis-controller.v1.0.9
        version: 1.0.9
      - name: ack-kinesis-controller.v1.0.8
        version: 1.0.8
      - name: ack-kinesis-controller.v1.0.7
        version: 1.0.7
      - name: ack-kinesis-controller.v1.0.6
        version: 1.0.6
      - name: ack-kinesis-controller.v1.0.5
        version: 1.0.5
      - name: ack-kinesis-controller.v1.0.4
        version: 1.0.4
      - name: ack-kinesis-controller.v1.0.3
        version: 1.0.3
      - name: ack-kinesis-controller.v1.0.2
        version: 1.0.2
      - name: ack-kinesis-controller.v1.0.1
        version: 1.0.1
      - name: ack-kinesis-controller.v1.0.0
        version: 1.0.0
      - name: ack-kinesis-controller.v0.0.19
        version: 0.0.19
      - name: ack-kinesis-controller.v0.0.18
        version: 0.0.18
      - name: ack-kinesis-controller.v0.0.17
        version: 0.0.17
      - name: ack-kinesis-controller.v0.0.16
        version: 0.0.16
      - name: ack-kinesis-controller.v0.0.15
        version: 0.0.15
      - name: ack-kinesis-controller.v0.0.10
        version: 0.0.10
      - name: ack-kinesis-controller.v0.0.9
        version: 0.0.9
      - name: ack-kinesis-controller.v0.0.8
        version: 0.0.8
      - name: ack-kinesis-controller.v0.0.6
        version: 0.0.6
      - name: ack-kinesis-controller.v0.0.5
        version: 0.0.5
      - name: ack-kinesis-controller.v0.0.4
        version: 0.0.4
      - name: ack-kinesis-controller.v0.0.3
        version: 0.0.3
      - name: ack-kinesis-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-kinesis-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cockroach Labs
      provider-url: ""
    name: cockroachdb-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroach-operator.v2.18.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v25.2.2",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
          createdAt: "2025-07-22T11:30:55Z"
          description: CockroachDB Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:85a0e085901c2a73192556d5d70795d4b66cbcfe01d24c73b6fe6db8d5f2e5ca
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c2a62179e8ca293d7e30b9da599e48795345cf07a72abd3a83063107b53b03cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1b6d934a7b6c42a61e5ad51380e69af341a575b42205299fd58bc0c6ff306f6a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:086d7b435993fdf21fd59b0093f52105a9028d6b769398b0033f98a65dfb7e79
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:363d8feb560324a848c5771edd224dc19020013af7641979ee81ae8ee536d7e3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2b5c028f293c6f394e58a5495cb3dbfbedc215d0d03778d0217e045586d55a46
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d14f77f70aea0422a906510c1ac2652ff76e131ed705cb002f80e13e887b0cd9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5240f9f591263569fad7ee934e81fa8a1617de3c25da94ddd59ee5b2b3407f90
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:56109e57ee0379cf48644bcf8226a9238e01139cacc5499002c99f973f121911
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5181196fff3bf8a5ec8fcfe4f1958cc69f285f0beb4f0b476b4b5e0912f3db42
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:818f592c50d4ed5599faa053344f7ecd963003352f1faa3b72dc0f5e00a0e78b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5b3de46526d169fcd1318354545b3a18861843fa2022a23ed04d795493825287
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:744193037d6b64cb73fecab15ebe022abc2b79ab0335b76ec1415ac6379b7dfa
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8a70ba12f610ac0e62066e417a0833bc1d01e91ac99d0dd794a2c0cd2a350164
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e82684b828256b3c2814930210031b3b70bc818514d4359a25f87ffdbf22be6e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a3ac544d87a9d8e7c92897c4c23f803753c7ab31467db91b5c51244afc2c6be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1ce71e80c0cd0e0063eba275ab91f77982f2e92a8aef8e89cab99b93feab0a88
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f3ffebd812881a830ff46c860984814dd61edcd51b478ebd3667759fb7710682
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:022c6e1a0de875474bc953718a481a0af526c49dc910094dcdce0b5677aced18
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:dd521d7bad62da054eec54179be75c643a5492d981e11c0e66db41902cd2182c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bed27f3f2c2d7627d1af4500d7ea0302da5c3fd3d875a6480fb6d98da481312e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:858dc82977b3c9e989a3a20626671b0c67c834c97132094a9212bb227ab55929
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc1ab77a9537ccf236ce6dc370940dfedcfaa91b6e3db37bf2d5eade7ff84708
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f5a0ccc02dc9e938e484d5b5282ff650d1890d5f754c30a0c0d717989ed5d600
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e18d73af3257c8cdf00245f2f718b56667524c1cc69d3db28e9b868a539ae16f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:310310515625f099a928545865f7096997871ee4a16650a01c76c3799a18b684
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b77b80eed7ce149d2a3f228e5f188e76d73f32bc50ecde20bba9cfa7bdc4714c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2a128b570352a5a7f82b27353f9b774caccce83e0e7bb2673e4802d96ad0bd51
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0fea6ed8a78ea624240ee6ffb3573d5cf1115186aa9180fcd7d1273351deaaa3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d978678b3c254abd52f69f226b0aacc4b02c2aaca328c54ef10f6f9bb453582d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ec6eb7c28c213cc83b2d7919cd87988f9a07f12276eb7351d0915f1567a5b095
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d590946b2f7c7ec0c7ec057a338e81f864d6a9e3d3ac4e0769aa5756a8e13fc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:468b0a40a33be1dd5a2a89ce745d7e9012bc7371f2f01cc0f8f12c0ac887db43
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cc20270212f01fc231ff788e20b0476b8228c440bfc40430c0836d5eb702a042
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1304d7f871a95c8b087bad1b548217562792edc66c8fe07c621e05242b8cd53e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7b907d03b001b758d8a4b5e898725c73dda2ec2aa4b7425b685320b947154d11
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4b56c0311af14220a466aa11f25fd10fa38280825539bd02215c945bbafb861c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7d6320ac7fb0925f10cf6f524a13de629cbf13ccc371912e4fe8c777bc15d1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1374cbdc4ecf38e4b4226328f532e830b8932b34d9449bf2fbb36e60d0310521
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:58de9d0b2ce005c7ad3e99152100e59b7677398557028a7a59e56e669fabb81c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6408f89b0a88ec154f7ee98c314be574971c4c9aaf906b473d10124d822889be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2171e1c75d0c665fe33e1ea334e605604c51de6dbc22d88a3a0ac604259c448f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:29abb169ffefe36f6d69edd438a7135ee23059cedab8f87ad3b2d52361b96b35
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6ca927f137be28481e3a226e6c6bcb7c5ac27664984f17ffc6c1419cd7d08eb7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7481c0cf3b579c8d98d94d4aa0334ca94bf2ccfcd25ad70b57c42b22bbf4bcee
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0b0e0588c1aadf0c3826cd584bd2b3a7d6781882dcb5c5e037b47ff5cb84509b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5d6901b6b6fd6594d42bfc1842d050c7cffcb0867a5591a67469f0f860b1f21b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b78a4d3864e16d270979d9d7756012abf09e6cdb7f1eb4832f6497c26281099
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:50393ad005fed19618d04f0716c6d55e2f1afce651e785c396638e852cac83b3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7ad12df8e258b03c4099d375d02d272bd774bf188d108901d5cf6cb865e57ef3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5a54d564e08baefc2786f4b742b2507d31984664cd00333406aa7a3aa5b9c913
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d467383af41aa80c26172964c34152abeba45121599804e502984655b72179f0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bf36dab9b3257a4bba24d986f1451fc51a3a37023fbb988281f1941fd3faedd
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d78c2e0709214f96794564f13386861f4b541e55b5bd316ad495cf104f7d40b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e5814748c5ab229a1dea5662a007063c0df06fbbfcfe041fd9a6df9ef67e07cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:68fe64fe3e112288ccfc2a7dffd69b9b2786a5da3d3bdedd6addf41659d460c9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da367cf0ac52045002e1c186f8e6964267ad87d5c25f9e72fd2c9b9a98a32702
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:28ae9661b38a72decf413c6a6c3758e61098debf5162197c27a12a070c7a096d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cbc74889e9ca5edbd5aaa243d7649bca0c4da2719239f8d0ec9827868112a771
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:46d9287ef11e3b21a36abfa31a26da3b93040bb6163b9223972a5dd1930fecff
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6a2ee8d62c586eb74ebbef6bef256e06e1ba5cb51c1ae55b1654adafc2ca8859
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:740ba10f1ef0a6c5fa6ad31ab18855fc75f8e024ce4c23b0676c5a9c5b62cd75
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:743cda1eda6e273c798e3405a9a97e07dd6aa5ad4d5a52789be3b0f39a32a039
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:53c7b960a2ed70a0998a4f8e28a4f78f7f676e01cffebe665e91e3a7629d88ed
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:79d688bebe0643ef0dd6ba7108cdb34c0442470c0c20ae0df88b5d02a3476cd0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08496b84b3cc691909227be17e67bcd6f4d5185a93712f44d8a0985bd2c5fcd7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:37bf9e1ec7e3eff47755fe7521223f31ec6aefc6e63901db3881b30898cf78f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:32fdca575c334822e4356aab36a7ed97b685c065925fe85f1b8ba8425c57159b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:742a6f69a3a209a7080a1e911fd96467d2a91aacb3f91d6be80a6a6e2e54b2ae
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1558a1f5eab79f1a8c6c85d81740cc539dc23025ea457c0019d93a72dc05d57a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e50aab39722af22ccbf3d1db19e37972d52b0fc9a40998a64618dc966b2bac57
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e0f21abea34825e55622c432a306f2e8aafd64d913eb78f2f601c51dc878a1da
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08cec123813304ab03bd66d72a09949667a874a42320bedafda391ebcdb2e56c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a6f1c43486d044fed9f59406118f429c8037b389bd8c8d7a1c7aeb89ed2661f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc4b75ed71845d5b18e1d9163d480fd9d0d3cb5ebbcfed00a2e4e1b174c0a5de
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4b3aa590623a38f435b075de4c09f8a343d8ece8b673e6bf7c6df5d5eb4bdf4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e23af8f9d32c3f80a52d79ceb4b359dda1d75ba0dd4cb8ed4b66ac2f4e40c69b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bd75aa918355e77cd671c3e11ee807861378e96245245f4fdf243cba5749f9a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3bf017b2b2506724500f5e29017d1b956825dd0f61a7c0049b0ff084c3554b5b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f438fbbf9c87ca0d54c170e2c5105b1e265182e1c2ede48adea370e2be9dc895
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8f1f2fd1aca914a1ef007117c85ac944055bbfb873a468a7374d239ef20b065e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f53b07212320c6939d3bb4ddd4e6a0bce311ab5b63858eb1a8728531f0dbd510
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da3bd2e9ff2b08d5bb86b1f5faa95a865f85c5e8912df00b6b79190c1e518a2f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0ec1dc184a2199736b4bd7de3241a357be7427caea5da1c931b08bc06f5c3dc0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5458a8fd89aa664b60e59c35bd933975cd37d18e19b11589e4ca0c2844ea0ed4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9ab968e4ca62f23bf9848e5475ef4076df7318a8560f63d5b2090a32af6fe4e2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f74c8a5dea4560ad59c6f6f6ab1e699f063334ee8db704460f870508dbeaeff0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f045fe5037d44aab1a0b2209241b3931ec7cb9897cb3c6a96a9f8d5280e55dd2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f05824bb8eaa5a93e54059449eaa236dd83d729049ad00065ef3242f5f74c91e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:82d8104a37349c513447e68ea3790efb256370632353176e5fc13c2e8df1bd44
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce1209ca6cc8724baee981c97c66e2852e512423bc0c0a16cdd80a6838b54031
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b046aa0a42dae992a03685ed82a20835532d6d4bf048e2366ff7382be4b3138
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9492894eaa5f43d00ac80b60029b32a0b0a85ae88ba609b46f98702c9349a5bb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f42fa22aac3cc41ce72e144a72530ffbee765a392b24593dafd06cc098f8410b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4e5f7df1dc1e1db398c36d590431e7e5782897b209972d8e9e4671971c10d1b6
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db4e5c2f27ba23e95a363c8174c01f0f9c6b6a27e0e581c86a38f3a1f68dc43c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6cd9f9b9fcb6e8f9a931a6ad01cb66329a74d40b22f9029fbc50e57248bad7cb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6371377718325235a2720dba0c97eb4321ee29b31a28cb893f76abd62baa339
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:87db6e079b96fb392d98ef5fc7063836bff10717efd75061016ce688be0ba38e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d8cfb221f67516675fdfde061f6ec5dfee1224ca7eb3d8be375c27d483f306af
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0754f7901feea9017bc5041cb91b7ac1cdbca10c2064f37c3648a2c395f24f83
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c74cf927a2b2c5098ad74ad0880996f8b042186a9911d888a231ef5b9e2a0715
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:65923c37fecda2d3174212fa9d9a6f34241d065389336ea69a7be99ba16cedc2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9c85c25d5392817d17b68714dca6195ad53255d43e9c65b021d75b12d97e0579
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:31b161f9bf9028a1a62cd2eef6abda01257d8889bdc3159822642549a75e3429
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8443c15eb1653dd9e5ec7aad54d4b8b092d41a0977e42b76e372e5e21280746e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6366ba29f301a82215c05e60d78e1de66bcee85440ce2fc3c692712d4da32685
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:02e1ef01a023da0eee653f63b314786fd66a78d184a97a314422f6f9c3ef8e7e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:42c9930d6861f6a2147ed6e01827eddd730279966763ec68868567a19cc164ab
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:35ede65457d845e8c063ad7d9515d3a69f98e0751424ecbac484963678a0b1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8e8c081c231b5b15eca483ceec16ea9190ba73aae6e421865162446eb584519a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a0faa52a999289b308dc4caf050c6d9fe2f42e112aa30e4af282ed1f21c5c346
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6bb15b36d64eebb6e4c1db5a5466e108b271d53383c58e0b6c78cec214756a9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0e2c7023f1f97acc20c46a4944635a9c5f57bb9bb3dd534cf41fea1a9cfdfdbe
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4d8e804b35b73d7978c112748aea2d2bb68f65081b90bca62c39f2b6e2989378
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b83eaff2e217c06098509709cead525b55b00ccd48019d8357e6231892008055
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:226d7f1c3f0d70ac6532ad9490bebcdef44a40959b6aa7c11b0f2d1383a46217
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a59cafff8d4ff2c62204ece15880117011c30978823c232902c7182d747ec827
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a6e2fe24ad52913e2160f85b9fe665a4e4c4d79b164b4b899f69dd9d92f999e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d0014b1c19890b9e8ef1a4f3efe7eb5b269be5610ba6ababd4258b52d2f85e8a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acc083f7dde1fbc557d858d03fc26b68b2056997fc1a457e4695650214d5b303
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf7d3bea47638b763b312c7c11ca034a7ac62d993d75a13edceec74799b97e52
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9f42e929f2cc7e9b6d24dd61e19c5d93dcfeabd34b9c8b935afbe78577efb18b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:44a0f40a6c46ba08297b482f3d96defe87b181c944803e5381c9acc42cabf97f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8737b67b91983817e3e4ff3f47f0b070bd36162d309b60993529426c6979cb24
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3f587e7b9f5a611013af2608e2151ed782859a52501b00adbc95739c56a8ff1e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:aeac9cd8e8acdbd2b0c339c48b85079c159b04f0ade24df8e3f9cae3f48b4c70
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1f6b1ec92251397dbcf7ed4a82f1f05619d64676aae2717f4eff1c33e4c2c27a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cad60044ad6573bd80b65e892a9ec0510dbe4fcbbfc6b51010ecf419f56f1024
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c9a31bdf19175880d0981e21dd19a1458b6e8cde8d4acbcb554f3bea3578a64e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:40f00476b63cb9e370fe85f01407173693213a9242fc117d28b06378ca0d98e0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4635550222775fc75737be5e5c0359303e3a23c4d1e1c296d7408cd15ad94aad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:72ea3d0ef72972dba93362dc31366c5946a9898f5c16780b877316a97068ef8f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7daaf1368ab2c25694cdea0a598e1de59247b04a4b4b4b8348ed6edbaa194f9d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce3d94a3fbdcde5858beb3c265987759971d991c397c6486e19904bd4706f44d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acbfee4492327a720fe7da76929a86192d8af2442bbb76f0a089f6a81a890be3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4f753151af4c3d3f5a1df9014461316249a6d0621b8fba5c4430307b3355251
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a4f932c3e9ff3aeb70ba1f7a92aa4642bd14cbb7d37d04ff750ed517b916cdb9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e151bded670f4cbd165a8ef864324805667d3e84ce3ace0cdf8f9aad087fccf1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db07461824e73410456ef540b0bce3c7fe550aa67a73d6174300ca74b8f75276
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:19f36f53f7da67755eb86da77098c5eeb84b381f9c18aa29ba573793dc498564
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2af60025ecb3bb933b61328be4c2b2bfd0e7d26f53b72430629208505e45c6d2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2d89398f03736f8ffaa82b10fed0f19fba166d3d2145b26f25cb71f3c39acf07
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:61e602b0e3582bf00460eb9328808ccf66e3ff1b1f6df07ee6c5c76a91846876
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:266165ab87b19a05d6796e8b15ff0eba632d6ac96e6393400464c6136dd500ec
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9097e135159f46396ef4649f724d61bb5b9113c488f989d404586fbb7486fd85
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1e5d85f5db10fa79d09263db90daa44549bcc3b025e9f3f98cf0d4e51766394b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9d0bfb642090ff55a3cc3da0f72663cae18063622a4ec785ba99bc71ef4db116
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b06bd7f3802e8fa2af14bb0b07804cad4cc0fbbc97da18d09a9106bf074ed6d8
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fe4a40006d1e8439423de1417cc9bc55716d790b882485074bb2a5ec68d7d1ad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b0057de903654ea0315a2c96affcc8f50a92ede8ad121ed30edb3aaf0fb7b554
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:10b9df672260c023dfb37166e156c0d7e0e13144dc3e485af922e8151fd3ab05
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8fc8de4f3671a73f9aa747e5b23ec973d192fff6f2869c38985cd656063c1e3d
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9b5a2e0c006eb4be8e1444734a98743551f278a24e81c20bebc96f0f3b631eb0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bd336715332f776a72aaf5c3ca52e62dd53a9e8521c1feb6e135cc868d694a39
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f29c31bb267451b389608af3e8a4a28d6ca0691e67b77e8f57bb5e461b064893
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf001173bbf00cdf443b8d64a3ad724389f1cf83e3628d0b609019c8cddad35a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fd6c9b8ca82ff4f2d0fe29166d1fc9bca6439cb40907be60f4d8f1b9735de2f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:91f7e59f9e35a0395756eafed906963c535e886125294d671d930ce959a8f511
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:90cf21c36d19d87b97d06edd1c0912a3bb288c769e6159651877db3206aa7355
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e8f26fd132affbaeb8ef6b1807a646c83b301e5dc334a592dfb2cf8dc3dbaf10
        version: 2.18.2
      entries:
      - name: cockroach-operator.v2.18.2
        version: 2.18.2
      name: stable
    defaultChannel: stable
    packageName: cockroachdb-certified
    provider:
      name: Cockroach Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Elastic
      provider-url: ""
    name: elasticsearch-eck-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: elasticsearch-eck-operator-certified.v3.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "elasticsearch.k8s.elastic.co/v1",
                    "kind": "Elasticsearch",
                    "metadata": {
                        "name": "elasticsearch-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "nodeSets": [
                            {
                                "name": "default",
                                "config": {
                                    "node.roles": ["master", "data"],
                                    "node.attr.attr_name": "attr_value",
                                    "node.store.allow_mmap": false
                                },
                                "podTemplate": {
                                    "metadata": {
                                        "labels": {
                                            "foo": "bar"
                                        }
                                    },
                                    "spec": {
                                        "containers": [
                                            {
                                                "name": "elasticsearch",
                                                "resources": {
                                                    "requests": {
                                                        "memory": "4Gi",
                                                        "cpu": 1
                                                    },
                                                    "limits": {
                                                        "memory": "4Gi",
                                                        "cpu": 2
                                                    }
                                                }
                                            }
                                        ]
                                    }
                                },
                                "count": 3
                            }
                        ]
                    }
                },
                {
                    "apiVersion": "kibana.k8s.elastic.co/v1",
                    "kind": "Kibana",
                    "metadata": {
                        "name": "kibana-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        },
                        "podTemplate": {
                            "metadata": {
                                "labels": {
                                    "foo": "bar"
                                }
                            },
                            "spec": {
                                "containers": [
                                    {
                                        "name": "kibana",
                                        "resources": {
                                            "requests": {
                                                "memory": "1Gi",
                                                "cpu": 0.5
                                            },
                                            "limits": {
                                                "memory": "2Gi",
                                                "cpu": 2
                                            }
                                        }
                                    }
                                ]
                            }
                        }
                    }
                },
                {
                    "apiVersion": "apm.k8s.elastic.co/v1",
                    "kind": "ApmServer",
                    "metadata": {
                        "name": "apmserver-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                    "apiVersion": "enterprisesearch.k8s.elastic.co/v1",
                    "kind": "EnterpriseSearch",
                    "metadata": {
                        "name": "ent-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "config": {
                            "ent_search.external_url": "https://localhost:3002"
                        },
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                  "apiVersion": "beat.k8s.elastic.co/v1beta1",
                  "kind": "Beat",
                  "metadata": {
                    "name": "heartbeat-sample"
                  },
                  "spec": {
                    "type": "heartbeat",
                    "version": "9.1.0",
                    "elasticsearchRef": {
                      "name": "elasticsearch-sample"
                    },
                    "config": {
                      "heartbeat.monitors": [
                        {
                          "type": "tcp",
                          "schedule": "@every 5s",
                          "hosts": [
                            "elasticsearch-sample-es-http.default.svc:9200"
                          ]
                        }
                      ]
                    },
                    "deployment": {
                      "replicas": 1,
                      "podTemplate": {
                        "spec": {
                          "securityContext": {
                            "runAsUser": 0
                          }
                        }
                      }
                    }
                  }
                },
                {
                  "apiVersion": "agent.k8s.elastic.co/v1alpha1",
                  "kind": "Agent",
                  "metadata": {
                    "name": "agent-sample"
                  },
                  "spec": {
                    "version": "9.1.0",
                    "elasticsearchRefs": [
                      {
                        "name": "elasticsearch-sample"
                      }
                    ],
                    "daemonSet": {},
                    "config": {
                      "inputs": [
                        {
                          "name": "system-1",
                          "revision": 1,
                          "type": "system/metrics",
                          "use_output": "default",
                          "meta": {
                            "package": {
                              "name": "system",
                              "version": "0.9.1"
                            }
                          },
                          "data_stream": {
                            "namespace": "default"
                          },
                          "streams": [
                            {
                              "id": "system/metrics-system.cpu",
                              "data_stream": {
                                "dataset": "system.cpu",
                                "type": "metrics"
                              },
                              "metricsets": [
                                "cpu"
                              ],
                              "cpu.metrics": [
                                "percentages",
                                "normalized_percentages"
                              ],
                              "period": "10s"
                            }
                          ]
                        }
                      ]
                    }
                  }
                },
                {
                    "apiVersion": "maps.k8s.elastic.co/v1alpha1",
                    "kind": "ElasticMapsServer",
                    "metadata": {
                        "name": "ems-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                    "apiVersion": "logstash.k8s.elastic.co/v1alpha1",
                    "kind": "Logstash",
                    "metadata" : {
                        "name": "logstash-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.connect.redhat.com/elastic/eck-operator@sha256:1750178c5dd716c031155f2e7058e19d4b7f1d3fdbb3e8a806b177a71be88500
          createdAt: "2025-07-30 11:10:34"
          description: Run Elasticsearch, Kibana, APM Server, Beats, Enterprise Search,
            Elastic Agent, Elastic Maps Server and Logstash on Kubernetes and OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Elastic Basic license
          repository: https://github.com/elastic/cloud-on-k8s
          support: elastic.co
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Elastic Agent instance
            displayName: Elastic Agent
            kind: Agent
            name: agents.agent.k8s.elastic.co
            version: v1alpha1
          - description: APM Server instance
            displayName: APM Server
            kind: ApmServer
            name: apmservers.apm.k8s.elastic.co
            version: v1
          - description: Beats instance
            displayName: Beats
            kind: Beat
            name: beats.beat.k8s.elastic.co
            version: v1beta1
          - description: Elastic Maps Server instance
            displayName: Elastic Maps Server
            kind: ElasticMapsServer
            name: elasticmapsservers.maps.k8s.elastic.co
            version: v1alpha1
          - description: Instance of an Elasticsearch autoscaler
            displayName: Elasticsearch Autoscaler
            kind: ElasticsearchAutoscaler
            name: elasticsearchautoscalers.autoscaling.k8s.elastic.co
            version: v1alpha1
          - description: Instance of an Elasticsearch cluster
            displayName: Elasticsearch Cluster
            kind: Elasticsearch
            name: elasticsearches.elasticsearch.k8s.elastic.co
            version: v1
          - description: Enterprise Search instance
            displayName: Enterprise Search
            kind: EnterpriseSearch
            name: enterprisesearches.enterprisesearch.k8s.elastic.co
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.kibana.k8s.elastic.co
            version: v1
          - description: Logstash instance
            displayName: Logstash
            kind: Logstash
            name: logstashes.logstash.k8s.elastic.co
            version: v1alpha1
          - description: Elastic Stack Config Policy
            displayName: Elastic Stack Config Policy
            kind: StackConfigPolicy
            name: stackconfigpolicies.stackconfigpolicy.k8s.elastic.co
            version: v1alpha1
        description: |-
          Elastic Cloud on Kubernetes (ECK) is the official operator by Elastic for automating the deployment, provisioning, management, and orchestration of Elasticsearch, Kibana, APM Server, Beats, Enterprise Search, Elastic Agent, Elastic Maps Server, and Logstash on Kubernetes.

          Current features:

          *  Elasticsearch, Kibana, APM Server, Enterprise Search, Beats, Elastic Agent, Elastic Maps Server, and Logstash deployments
          *  TLS Certificates management
          *  Safe Elasticsearch cluster configuration and topology changes
          *  Persistent volumes usage
          *  Custom node configuration and attributes
          *  Secure settings keystore updates

          Supported versions:

          * Kubernetes 1.29-1.33
          * OpenShift 4.15-4.19
          * Google Kubernetes Engine (GKE), Azure Kubernetes Service (AKS), and Amazon Elastic Kubernetes Service (EKS)
          * Elasticsearch, Kibana, APM Server: 7.17+, 8+, 9+
          * Enterprise Search: 7.17+, 8+
          * Beats: 7.17+, 8+, 9+
          * Elastic Agent: 7.17+, 8+, 9+
          * Elastic Maps Server: 7.17+, 8+, 9+
          * Logstash 8.12+, 9+

          ECK should work with all conformant installers as listed in these [FAQs](https://github.com/cncf/k8s-conformance/blob/master/faq.md#what-is-a-distribution-hosted-platform-and-an-installer). Distributions include source patches and so may not work as-is with ECK.
          Alpha, beta, and stable API versions follow the same [conventions used by Kubernetes](https://kubernetes.io/docs/concepts/overview/kubernetes-api/#api-versioning).
          See the full [Elastic support matrix](https://www.elastic.co/support/matrix#matrix_kubernetes) for more information.
          See the [Quickstart](https://www.elastic.co/guide/en/cloud-on-k8s/3.1/k8s-quickstart.html) to get started with ECK.
        displayName: Elasticsearch (ECK) Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - analytics
        - search
        - database
        - apm
        - logstash
        links:
        - name: Documentation
          url: https://www.elastic.co/guide/en/cloud-on-k8s/3.1/index.html
        maintainers:
        - email: eck@elastic.co
          name: Elastic
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Elastic
        relatedImages:
        - registry.connect.redhat.com/elastic/eck-operator@sha256:1750178c5dd716c031155f2e7058e19d4b7f1d3fdbb3e8a806b177a71be88500
        version: 3.1.0
      entries:
      - name: elasticsearch-eck-operator-certified.v3.1.0
        version: 3.1.0
      - name: elasticsearch-eck-operator-certified.v3.0.0
        version: 3.0.0
      - name: elasticsearch-eck-operator-certified.v2.16.1
        version: 2.16.1
      - name: elasticsearch-eck-operator-certified.v2.16.0
        version: 2.16.0
      - name: elasticsearch-eck-operator-certified.v2.15.0
        version: 2.15.0
      - name: elasticsearch-eck-operator-certified.v2.14.0
        version: 2.14.0
      - name: elasticsearch-eck-operator-certified.v2.13.0
        version: 2.13.0
      - name: elasticsearch-eck-operator-certified.v2.12.1
        version: 2.12.1
      - name: elasticsearch-eck-operator-certified.v2.12.0
        version: 2.12.0
      - name: elasticsearch-eck-operator-certified.v2.11.1
        version: 2.11.1
      - name: elasticsearch-eck-operator-certified.v2.11.0
        version: 2.11.0
      - name: elasticsearch-eck-operator-certified.v2.10.0
        version: 2.10.0
      - name: elasticsearch-eck-operator-certified.v2.9.0
        version: 2.9.0
      - name: elasticsearch-eck-operator-certified.v2.8.0
        version: 2.8.0
      - name: elasticsearch-eck-operator-certified.v2.7.0
        version: 2.7.0
      - name: elasticsearch-eck-operator-certified.v2.6.2
        version: 2.6.2
      - name: elasticsearch-eck-operator-certified.v2.6.1
        version: 2.6.1
      - name: elasticsearch-eck-operator-certified.v2.5.0
        version: 2.5.0
      - name: elasticsearch-eck-operator-certified.v2.4.0
        version: 2.4.0
      - name: elasticsearch-eck-operator-certified.v2.3.0
        version: 2.3.0
      - name: elasticsearch-eck-operator-certified.v2.2.0
        version: 2.2.0
      - name: elasticsearch-eck-operator-certified.v2.1.0
        version: 2.1.0
      - name: elasticsearch-eck-operator-certified.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: elasticsearch-eck-operator-certified
    provider:
      name: Elastic
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Portworx by PureStorage
      provider-url: ""
    name: portworx-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: portworx-operator.v25.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageCluster",
                "metadata": {
                  "name": "portworx",
                  "namespace": "test-operator",
                  "annotations": {
                    "portworx.io/is-openshift": "true"
                  }
                },
                "spec": {}
              },
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageNode",
                "metadata": {
                  "name": "example",
                  "namespace": "test-operator"
                },
                "spec": {}
              }
            ]
          capabilities: Auto Pilot
          categories: Storage
          certified: "true"
          console.openshift.io/plugins: '["portworx"]'
          containerImage: registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
          createdAt: "2025-09-06T19:38:43Z"
          description: Cloud native storage solution for production workloads
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "core.libopenstorage.org/v1",
              "kind": "StorageCluster",
              "metadata": {
                "name": "portworx",
                "annotations": {
                  "portworx.io/is-openshift": "true"
                }
              }
            }
          support: Portworx, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Storage Cluster installs Portworx in the cluster. It has
              all the necessary configurations to setup and update a Portworx cluster.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.core.libopenstorage.org
            version: v1
          - description: Do not create Storage Node as it is internally created by
              the operator. It represents the status of a Portworx node.
            displayName: Storage Node
            kind: StorageNode
            name: storagenodes.core.libopenstorage.org
            version: v1
        description: |
          Portworx-Enterprise is the most widely-used and reliable cloud-native
          storage solution for production workloads and provides high-availability,
          data protection and security for containerized applications.

          Portworx Enterprise enables you to migrate entire applications, including
          data, between clusters in a single data center or cloud, or between clouds,
          with a single kubectl command.

          The cloud native storage and data management platform that enterprises trust
          to manage data in containers now has an operator which simplifies the install,
          configuration, upgrades and manages the Portworx Enterprise cluster lifecycle.

          Learn more about the Portworx Enterprise
          [the data platform for Kubernetes](https://portworx.com/products/introduction)

          To learn more about the platform features, please visit our
          [product features page](https://portworx.com/products/features)

          ### About Portworx

          Portworx is the solution for running stateful containers in production,
          designed with DevOps in mind. With Portworx, users can manage any database
          or stateful service on any infrastructure using any container scheduler,
          including Kubernetes, Mesosphere DC/OS, and Docker Swarm. Portworx solves
          the five most common problems DevOps teams encounter when running stateful
          services in production: persistence, high availability, data automation,
          security, and support for multiple data stores and infrastructure.

          ### How to install StorageCluster

          To customize your cluster's configuration (specification), use the
          [Spec Generator](https://central.portworx.com/) from PX-Central.

          ### Prerequisite

          Ensure ports 17001-17020 on worker nodes are reachable from master and other worker nodes.

          ### Tutorials

          * [Portworx Enterprise on Openshift](https://portworx.com/openshift)

          * [Stateful applications on Kubernetes](https://docs.portworx.com/portworx-enterprise/concepts/kubernetes-storage-101/applications)

          * [Portworx Enterprise on Kubernetes](https://docs.portworx.com/portworx-install-with-kubernetes)

          * [Kafka on Kubernetes](https://portworx.com/kafka-kubernetes)

          * [Elastisearch on Kubernetes](https://portworx.com/elasticsearch-kubernetes)

          * [PostgreSQL on Kubernetes](https://portworx.com/postgres-kubernetes/)

          * [MongoDB on Kubernetes](https://portworx.com/mongodb-kubernetes/)

          * [Cassandra on Kubernetes](https://portworx.com/cassandra-kubernetes/)

          * [Kubernetes backup and recovery](https://portworx.com/kubernetes-backup/)

          * [Disaster Recovery for Kubernetes](https://portworx.com/kubernetes-disaster-recovery/)

          ### Uninstall

          Deleting the StorageCluster object for Portworx cluster does not stop Portworx
          service running on the nodes, to avoid application downtime.

          To uninstall Portworx completely without wiping the data, you should add the
          following delete strategy to the StorageCluster spec:
          ```
          spec:
            deleteStrategy:
              type: Uninstall
          ```
          **Caution:** To uninstall Portworx and **wipe all the data**, you should use the following
          delete strategy:
          ```
          spec:
            deleteStrategy:
              type: UninstallAndWipe
          ```
        displayName: Portworx Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - portworx
        - persistent storage
        - storage
        - cloud native
        - open source
        links:
        - name: Product Features
          url: https://portworx.com/products/features
        - name: Documentation
          url: https://docs.portworx.com/portworx-enterprise/platform/openshift
        - name: Support
          url: https://docs.portworx.com/portworx-enterprise/support/contact-support
        - name: Source Code
          url: https://github.com/libopenstorage/operator
        maintainers:
        - email: support@portworx.com
          name: Portworx
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Portworx by PureStorage
        relatedImages:
        - registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
        version: 25.3.1
      entries:
      - name: portworx-operator.v25.3.1
        version: 25.3.1
      - name: portworx-operator.v25.3.0
        version: 25.3.0
      - name: portworx-operator.v25.2.2
        version: 25.2.2
      - name: portworx-operator.v25.2.1
        version: 25.2.1
      - name: portworx-operator.v25.2.0
        version: 25.2.0
      - name: portworx-operator.v24.2.4
        version: 24.2.4
      - name: portworx-operator.v24.2.3
        version: 24.2.3
      - name: portworx-operator.v24.2.2
        version: 24.2.2
      - name: portworx-operator.v24.2.1
        version: 24.2.1
      - name: portworx-operator.v24.2.0
        version: 24.2.0
      - name: portworx-operator.v24.1.3
        version: 24.1.3
      - name: portworx-operator.v24.1.2
        version: 24.1.2
      - name: portworx-operator.v24.1.1
        version: 24.1.1
      - name: portworx-operator.v24.1.0
        version: 24.1.0
      - name: portworx-operator.v23.10.5
        version: 23.10.5
      - name: portworx-operator.v23.10.4
        version: 23.10.4
      - name: portworx-operator.v23.10.3
        version: 23.10.3
      - name: portworx-operator.v23.10.2
        version: 23.10.2
      - name: portworx-operator.v23.10.1
        version: 23.10.1
      - name: portworx-operator.v23.7.0
        version: 23.7.0
      - name: portworx-operator.v23.5.1
        version: 23.5.1
      - name: portworx-operator.v23.5.0
        version: 23.5.0
      - name: portworx-operator.v23.4.0
        version: 23.4.0
      - name: portworx-operator.v23.3.1
        version: 23.3.1
      - name: portworx-operator.v23.3.0
        version: 23.3.0
      - name: portworx-operator.v1.10.5
        version: 1.10.5
      - name: portworx-operator.v1.10.4
        version: 1.10.4
      - name: portworx-operator.v1.10.3
        version: 1.10.3
      - name: portworx-operator.v1.10.2
        version: 1.10.2
      - name: portworx-operator.v1.10.1
        version: 1.10.1
      - name: portworx-operator.v1.10.0
        version: 1.10.0
      - name: portworx-operator.v1.9.1
        version: 1.9.1
      - name: portworx-operator.v1.9.0
        version: 1.9.0
      - name: portworx-operator.v1.8.1
        version: 1.8.1
      - name: portworx-operator.v1.8.0
        version: 1.8.0
      - name: portworx-operator.v1.7.1
        version: 1.7.1
      - name: portworx-operator.v1.6.1
        version: 1.6.1
      name: alpha
    - currentCSV: portworx-operator.v25.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageCluster",
                "metadata": {
                  "name": "portworx",
                  "namespace": "test-operator",
                  "annotations": {
                    "portworx.io/is-openshift": "true"
                  }
                },
                "spec": {}
              },
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageNode",
                "metadata": {
                  "name": "example",
                  "namespace": "test-operator"
                },
                "spec": {}
              }
            ]
          capabilities: Auto Pilot
          categories: Storage
          certified: "true"
          console.openshift.io/plugins: '["portworx"]'
          containerImage: registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
          createdAt: "2025-09-06T19:38:43Z"
          description: Cloud native storage solution for production workloads
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "core.libopenstorage.org/v1",
              "kind": "StorageCluster",
              "metadata": {
                "name": "portworx",
                "annotations": {
                  "portworx.io/is-openshift": "true"
                }
              }
            }
          support: Portworx, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Storage Cluster installs Portworx in the cluster. It has
              all the necessary configurations to setup and update a Portworx cluster.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.core.libopenstorage.org
            version: v1
          - description: Do not create Storage Node as it is internally created by
              the operator. It represents the status of a Portworx node.
            displayName: Storage Node
            kind: StorageNode
            name: storagenodes.core.libopenstorage.org
            version: v1
        description: |
          Portworx-Enterprise is the most widely-used and reliable cloud-native
          storage solution for production workloads and provides high-availability,
          data protection and security for containerized applications.

          Portworx Enterprise enables you to migrate entire applications, including
          data, between clusters in a single data center or cloud, or between clouds,
          with a single kubectl command.

          The cloud native storage and data management platform that enterprises trust
          to manage data in containers now has an operator which simplifies the install,
          configuration, upgrades and manages the Portworx Enterprise cluster lifecycle.

          Learn more about the Portworx Enterprise
          [the data platform for Kubernetes](https://portworx.com/products/introduction)

          To learn more about the platform features, please visit our
          [product features page](https://portworx.com/products/features)

          ### About Portworx

          Portworx is the solution for running stateful containers in production,
          designed with DevOps in mind. With Portworx, users can manage any database
          or stateful service on any infrastructure using any container scheduler,
          including Kubernetes, Mesosphere DC/OS, and Docker Swarm. Portworx solves
          the five most common problems DevOps teams encounter when running stateful
          services in production: persistence, high availability, data automation,
          security, and support for multiple data stores and infrastructure.

          ### How to install StorageCluster

          To customize your cluster's configuration (specification), use the
          [Spec Generator](https://central.portworx.com/) from PX-Central.

          ### Prerequisite

          Ensure ports 17001-17020 on worker nodes are reachable from master and other worker nodes.

          ### Tutorials

          * [Portworx Enterprise on Openshift](https://portworx.com/openshift)

          * [Stateful applications on Kubernetes](https://docs.portworx.com/portworx-enterprise/concepts/kubernetes-storage-101/applications)

          * [Portworx Enterprise on Kubernetes](https://docs.portworx.com/portworx-install-with-kubernetes)

          * [Kafka on Kubernetes](https://portworx.com/kafka-kubernetes)

          * [Elastisearch on Kubernetes](https://portworx.com/elasticsearch-kubernetes)

          * [PostgreSQL on Kubernetes](https://portworx.com/postgres-kubernetes/)

          * [MongoDB on Kubernetes](https://portworx.com/mongodb-kubernetes/)

          * [Cassandra on Kubernetes](https://portworx.com/cassandra-kubernetes/)

          * [Kubernetes backup and recovery](https://portworx.com/kubernetes-backup/)

          * [Disaster Recovery for Kubernetes](https://portworx.com/kubernetes-disaster-recovery/)

          ### Uninstall

          Deleting the StorageCluster object for Portworx cluster does not stop Portworx
          service running on the nodes, to avoid application downtime.

          To uninstall Portworx completely without wiping the data, you should add the
          following delete strategy to the StorageCluster spec:
          ```
          spec:
            deleteStrategy:
              type: Uninstall
          ```
          **Caution:** To uninstall Portworx and **wipe all the data**, you should use the following
          delete strategy:
          ```
          spec:
            deleteStrategy:
              type: UninstallAndWipe
          ```
        displayName: Portworx Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - portworx
        - persistent storage
        - storage
        - cloud native
        - open source
        links:
        - name: Product Features
          url: https://portworx.com/products/features
        - name: Documentation
          url: https://docs.portworx.com/portworx-enterprise/platform/openshift
        - name: Support
          url: https://docs.portworx.com/portworx-enterprise/support/contact-support
        - name: Source Code
          url: https://github.com/libopenstorage/operator
        maintainers:
        - email: support@portworx.com
          name: Portworx
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Portworx by PureStorage
        relatedImages:
        - registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
        version: 25.3.1
      entries:
      - name: portworx-operator.v25.3.1
        version: 25.3.1
      - name: portworx-operator.v25.3.0
        version: 25.3.0
      - name: portworx-operator.v25.2.2
        version: 25.2.2
      - name: portworx-operator.v25.2.1
        version: 25.2.1
      - name: portworx-operator.v25.2.0
        version: 25.2.0
      - name: portworx-operator.v24.2.4
        version: 24.2.4
      - name: portworx-operator.v24.2.3
        version: 24.2.3
      - name: portworx-operator.v24.2.2
        version: 24.2.2
      - name: portworx-operator.v24.2.1
        version: 24.2.1
      - name: portworx-operator.v24.2.0
        version: 24.2.0
      - name: portworx-operator.v24.1.3
        version: 24.1.3
      - name: portworx-operator.v24.1.2
        version: 24.1.2
      - name: portworx-operator.v24.1.1
        version: 24.1.1
      - name: portworx-operator.v24.1.0
        version: 24.1.0
      - name: portworx-operator.v23.10.5
        version: 23.10.5
      - name: portworx-operator.v23.10.4
        version: 23.10.4
      - name: portworx-operator.v23.10.3
        version: 23.10.3
      - name: portworx-operator.v23.10.2
        version: 23.10.2
      - name: portworx-operator.v23.10.1
        version: 23.10.1
      - name: portworx-operator.v23.7.0
        version: 23.7.0
      - name: portworx-operator.v23.5.1
        version: 23.5.1
      - name: portworx-operator.v23.5.0
        version: 23.5.0
      - name: portworx-operator.v23.4.0
        version: 23.4.0
      - name: portworx-operator.v23.3.1
        version: 23.3.1
      - name: portworx-operator.v23.3.0
        version: 23.3.0
      - name: portworx-operator.v1.10.5
        version: 1.10.5
      - name: portworx-operator.v1.10.4
        version: 1.10.4
      - name: portworx-operator.v1.10.3
        version: 1.10.3
      - name: portworx-operator.v1.10.2
        version: 1.10.2
      - name: portworx-operator.v1.10.1
        version: 1.10.1
      - name: portworx-operator.v1.10.0
        version: 1.10.0
      - name: portworx-operator.v1.9.1
        version: 1.9.1
      - name: portworx-operator.v1.9.0
        version: 1.9.0
      - name: portworx-operator.v1.8.1
        version: 1.8.1
      - name: portworx-operator.v1.8.0
        version: 1.8.0
      - name: portworx-operator.v1.7.1
        version: 1.7.1
      - name: portworx-operator.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: portworx-certified
    provider:
      name: Portworx by PureStorage
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: clusterresourceoverride
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clusterresourceoverride-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.autoscaling.openshift.io/v1",
                "kind": "ClusterResourceOverride",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "podResourceOverride": {
                    "spec": {
                      "memoryRequestToLimitPercent": 50,
                      "cpuRequestToLimitPercent": 25,
                      "limitCPUToMemoryPercent": 200
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certifiedLevel: Primed
          containerImage: registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9-operator@sha256:40b19c94f688c3a58189fcb404df35a40a8acbbdbf73fadb8818a86fd14a9ab9
          createdAt: 2019/11/15
          description: An operator to manage the OpenShift ClusterResourceOverride
            Mutating Admission Webhook Server
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          healthIndex: B
          olm.skipRange: '>=4.3.0 <4.16.0-202509240123'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-resource-override-admission-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of ClusterResourceOverride Admission
              Webhook
            displayName: ClusterResourceOverride
            kind: ClusterResourceOverride
            name: clusterresourceoverrides.operator.autoscaling.openshift.io
            version: v1
        description: "ClusterResourceOverride\n==============\n\nContainers can specify
          compute resource requests and limits. Requests are used for scheduling your
          container and provide a minimum service guarantee. Limits constrain the
          amount of compute resource that may be consumed on your node.\n\nThe scheduler
          attempts to optimize the compute resource use across all nodes in your cluster.
          It places pods onto specific nodes, taking the pods' compute resource requests
          and nodes' available capacity into consideration.\n\nRequests and limits
          enable administrators to allow and manage the overcommitment of resources
          on a node, which may be desirable in development environments where a trade
          off of guaranteed performance for capacity is acceptable.\n\n### Requests
          and Limits\n\nFor each compute resource, a container may specify a resource
          request and limit. Scheduling decisions are made based on the request to
          ensure that a node has enough capacity available to meet the requested value.
          If a container specifies limits, but omits requests, the requests are defaulted
          to the limits. A container is not able to exceed the specified limit on
          the node.\n\nThe enforcement of limits is dependent upon the compute resource
          type. If a container makes no request or limit, the container is scheduled
          to a node with no resource guarantees. In practice, the container is able
          to consume as much of the specified resource as is available with the lowest
          local priority. In low resource situations, containers that specify no resource
          requests are given the lowest quality of service.\n\n### Compute Resources\n\nThe
          node-enforced behavior for compute resources is specific to the resource
          type.\n\n#### CPU\n\nA container is guaranteed the amount of CPU it requests
          and is additionally able to consume excess CPU available on the node, up
          to any limit specified by the container. If multiple containers are attempting
          to use excess CPU, CPU time is distributed based on the amount of CPU requested
          by each container.\n\nFor example, if one container requested 500m of CPU
          time and another container requested 250m of CPU time, then any extra CPU
          time available on the node is distributed among the containers in a 2:1
          ratio. If a container specified a limit, it will be throttled not to use
          more CPU than the specified limit.\n\nCPU requests are enforced using the
          CFS shares support in the Linux kernel. By default, CPU limits are enforced
          using the CFS quota support in the Linux kernel over a 100ms measuring interval,
          though this can be disabled.\n\n#### Memory\n\nA container is guaranteed
          the amount of memory it requests. A container may use more memory than requested,
          but once it exceeds its requested amount, it could be killed in a low memory
          situation on the node.\n\nIf a container uses less memory than requested,
          it will not be killed unless system tasks or daemons need more memory than
          was accounted for in the node’s resource reservation. If a container specifies
          a limit on memory, it is immediately killed if it exceeds the limit amount.\n\n###
          Configuring the Cluster for Overcommitment\n\nScheduling is based on resources
          requested, while quota and hard limits refer to resource limits, which can
          be set higher than requested resources. The difference between request and
          limit determines the level of overcommit; for instance, if a container is
          given a memory request of 1Gi and a memory limit of 2Gi, it is scheduled
          based on the 1Gi request being available on the node, but could use up to
          2Gi; so it is 200% overcommitted. If OpenShift Container Platform administrators
          would like to control the level of overcommit and manage container density
          on nodes, ClusterResourceOverride Admission Webhook can be configured to
          override the ratio between request and limit set on developer containers.
          In conjunction with a per-project LimitRange specifying limits and defaults,
          this adjusts the container limit and request to achieve the desired level
          of overcommit.\n\nThis requires creating a custom resource of `ClusterResourceOverride`
          type as in the following example:\n\n    \n    - apiVersion: operator.autoscaling.openshift.io/v1\n
          \   - kind: ClusterResourceOverride\n    - metadata:\n    -   name: cluster\n
          \   - spec:\n    -   podResourceOverride:\n    -     spec:\n    -       memoryRequestToLimitPercent:
          25\n    -       cpuRequestToLimitPercent: 25\n    -       limitCPUToMemoryPercent:
          200\n    \n **memoryRequestToLimitPercent**: (optional, 1-100) If a container
          memory limit has been specified or defaulted, the memory request is overridden
          to this percentage of the limit.\n\n **cpuRequestToLimitPercent**: (optional,
          1-100) If a container CPU limit has been specified or defaulted, the CPU
          request is overridden to this percentage of the limit.\n\n **limitCPUToMemoryPercent**:
          (optional, positive integer) If a container memory limit has been specified
          or defaulted, the CPU limit is overridden to a percentage of the memory
          limit, with a 100 percentage scaling 1Gi of RAM to equal 1 CPU core. This
          is processed prior to overriding CPU request (if configured).\n\n Note that
          these overrides have no effect if no limits have been set on containers.
          [Create a LimitRange object] (https://docs.openshift.com/container-platform/3.3/admin_guide/limits.html#admin-guide-limits)
          with default limits (per individual project, or in the [project template](https://docs.openshift.com/container-platform/3.3/admin_guide/managing_projects.html#modifying-the-template-for-new-projects))
          in order to ensure that the overrides apply.\n\nWhen configured, overrides
          can be enabled per-project by applying the following label.\n    ```\n    clusterresourceoverrides.admission.autoscaling.openshift.io/enabled:
          \"true\"\n    ```\n\n"
        displayName: ClusterResourceOverride Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - deschedule
        - scale
        - binpack
        - efficiency
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9-operator@sha256:40b19c94f688c3a58189fcb404df35a40a8acbbdbf73fadb8818a86fd14a9ab9
        - registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9@sha256:e4992c556fe6be50fa30a45b05cffd6029ebd8914eca7c1c683564ff869e0e7d
        version: 4.16.0-202509240123
      entries:
      - name: clusterresourceoverride-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: clusterresourceoverride-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: clusterresourceoverride-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: clusterresourceoverride-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: clusterresourceoverride-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: clusterresourceoverride-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: clusterresourceoverride-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: clusterresourceoverride-operator.v4.16.0-202505092110
        version: 4.16.0-202505092110
      - name: clusterresourceoverride-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: clusterresourceoverride-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: clusterresourceoverride-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: clusterresourceoverride-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: clusterresourceoverride-operator.v4.16.0-202501241104
        version: 4.16.0-202501241104
      - name: clusterresourceoverride-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: clusterresourceoverride-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: clusterresourceoverride-operator.v4.16.0-202412021735
        version: 4.16.0-202412021735
      - name: clusterresourceoverride-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: clusterresourceoverride-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: clusterresourceoverride-operator.v4.16.0-202411081636
        version: 4.16.0-202411081636
      - name: clusterresourceoverride-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: clusterresourceoverride-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: clusterresourceoverride-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: clusterresourceoverride-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: clusterresourceoverride-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: clusterresourceoverride-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: clusterresourceoverride-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: clusterresourceoverride-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: clusterresourceoverride-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: clusterresourceoverride-operator.v4.16.0-202407040206
        version: 4.16.0-202407040206
      - name: clusterresourceoverride-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: clusterresourceoverride-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: clusterresourceoverride-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: clusterresourceoverride-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: clusterresourceoverride
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicurio-registry-3
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-3.v3.0.15-r1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "simple"
                },
                "spec": {
                  "app": {
                    "ingress": {
                      "host": "simple-app.apps.cluster.example"
                    }
                  },
                  "ui": {
                    "ingress": {
                      "host": "simple-ui.apps.cluster.example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/apicurio/apicurio-registry-rhel9-operator@sha256:e6d2ceab8accdb6e1b9766733039e24df620d7128eb1e13e60a61339130d52a7
          createdAt: "2025-09-29"
          description: Deploy and manage Red Hat build of Apicurio Registry 3 on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations",
            "Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Red Hat build
              of Apicurio Registry version 3.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Red Hat build of Apicurio Registry 3

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Allows you manage event schemas and API designs using the Apicurio Registry web console, REST API, Maven plug-in, or Java client
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, MySQL\*, SQL Server\*, in-memory\*
          - Compatible with Confluent APIs
          - Runs on a lightweight Quarkus platform

          _Note: Storage options marked with * have not yet been tested by Red Hat for production deployments._

          ## Red Hat build of Apicurio Registry 3 Operator

          Provides a quick and easy way to deploy and manage Red Hat build of Apicurio Registry 3 on OpenShift.

          **Features**
          - Supports installation and configuration of Red Hat build of Apicurio Registry 3
          - Perform a rolling upgrade of Red Hat build of Apicurio Registry 3

          ## Prerequisites

          This operator does not deploy storage for Red Hat build of Apicurio Registry 3. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Red Hat build of Apicurio Registry 3 and Apicurio Registry 3 Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Red Hat build of Apicurio Registry 3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/apicurio/apicurio-registry-ui-rhel9@sha256:e4791546fd371b34c7580c6aca2faed8527c985c8029fa8087fb46fd94ab8feb
        - registry.redhat.io/apicurio/apicurio-registry-rhel9-operator@sha256:e6d2ceab8accdb6e1b9766733039e24df620d7128eb1e13e60a61339130d52a7
        - registry.redhat.io/apicurio/apicurio-registry-rhel9@sha256:c86d73d5e2b4146f016ce7ec67265fe80bd596f472247eb9f5590cc0c89dba75
        version: 3.0.15-r1
      entries:
      - name: apicurio-registry-3.v3.0.15-r1
        version: 3.0.15-r1
      - name: apicurio-registry-3.v3.0.9-r3
        version: 3.0.9-r3
      - name: apicurio-registry-3.v3.0.9-r2
        version: 3.0.9-r2
      - name: apicurio-registry-3.v3.0.9-r1
        version: 3.0.9-r1
      - name: apicurio-registry-3.v3.0.7-r2
        version: 3.0.7-r2
      - name: apicurio-registry-3.v3.0.7-r1
        version: 3.0.7-r1
      - name: apicurio-registry-3.v3.0.6-beta.r1
        version: 3.0.6-beta.r1
      - name: apicurio-registry-3.v3.0.6-beta.r1-0.1740495889.p
        version: 3.0.6-beta.r1+0.1740495889.p
      - name: apicurio-registry-3.v3.0.6-beta
        version: 3.0.6-beta
      name: 3.x
    defaultChannel: 3.x
    packageName: apicurio-registry-3
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SusQL Operator Contributors
      provider-url: https://github.com/sustainable-computing-io/susql-operator
    name: susql-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: susql-operator.v0.0.35
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "susql.ibm.com/v1",
                "kind": "LabelGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "susql-operator",
                    "app.kubernetes.io/instance": "labelgroup-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "susql-operator",
                    "app.kubernetes.io/part-of": "susql-operator"
                  },
                  "name": "labelgroup-sample"
                },
                "spec": {
                  "labels": [
                    "labelgroup-sample-1",
                    "labelgroup-sample-2"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: quay.io/sustainable_computing_io/susql_operator:0.0.34
          createdAt: "2025-07-10T01:20:37Z"
          description: 'Aggregates energy and CO2 emission data for pods tagged with
            SusQL labels '
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/sustainable-computing-io/susql-operator
          support: https://github.com/sustainable-computing-io/susql-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LabelGroup is the Schema for the LabelGroups API
            displayName: Label Group
            kind: LabelGroup
            name: labelgroups.susql.ibm.com
            version: v1
        description: |-
          ### About this Operator

          SusQL is a Kubernetes operator that aggregates energy and estimated
          carbon dioxide emission data for pods tagged with SusQL specific
          labels. The energy measurements are taken from Kepler which should
          be deployed in the cluster before using SusQL.

          ### Prerequisites

          1. Deployment of [Kepler](https://sustainable-computing.io/) on the cluster
          2. Ensure that [User Project Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          is enabled to monitor energy consumed in user projects.
          3. Creation of a SusQL Service Monitor:
                `oc apply -n <NAMESPACE> -f https://raw.githubusercontent.com/sustainable-computing-io/susql-operator/main/hack/susql-servicemonitor.yaml`

          ### API Backward Compatibility

          Until Operator supports **seamless upgrades**, a new version of the operator may
          introduce a change that is **NOT** backwards compatible with the previous version.
          Thus, to upgrade the operator, uninstall the already installed version completely
          (including CRDs) and install the new version. For example to upgrade from 0.1.0
          to 0.2.0, you must first uninstall 0.1.0.

          ### LabelGroup spec Example

          ```
          apiVersion: susql.ibm.com/v1
          kind: LabelGroup
          metadata:
            name: labelgroup-name
            namespace: susql-operator
          spec:
            labels:
              - my-label-1
              - my-label-2
          ```

          ### Documentation

          Documentation can be found at [SusQL Operator](https://github.com/sustainable-computing-io/susql-operator#readme).

          ### Getting help

          If you encounter any issues while using the SusQL Operator, you can create an
          [issue](https://github.com/sustainable-computing-io/susql-operator/issues)
          on our GitHub repo for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by raising [issues](https://github.com/sustainable-computing-io/susql-operator/issues)
          related to susql-operator.

          ### License
          susql-operator is licensed under the Apache 2.0 license
        displayName: SusQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - aggregation
        - sustainability
        - energy
        - kepler
        - susql
        - carbon dioxide (CO2) emissions
        - green computing
        links:
        - name: SusQL Operator
          url: https://github.com/sustainable-computing-io/susql-operator
        - name: Sustainable Computing
          url: https://github.com/sustainable-computing-io
        maintainers:
        - email: trent@jp.ibm.com
          name: Scott Trent
        - email: mohammed.munir.abdi@ibm.com
          name: Mohammed Abdi
        maturity: alpha
        provider:
          name: SusQL Operator Contributors
          url: https://github.com/sustainable-computing-io/susql-operator
        relatedImages:
        - quay.io/sustainable_computing_io/susql_operator:0.0.34
        version: 0.0.35
      entries:
      - name: susql-operator.v0.0.35
        version: 0.0.35
      - name: susql-operator.v0.0.34
        version: 0.0.34
      - name: susql-operator.v0.0.33
        version: 0.0.33
      - name: susql-operator.v0.0.32
        version: 0.0.32
      - name: susql-operator.v0.0.30
        version: 0.0.30
      - name: susql-operator.v0.0.29
        version: 0.0.29
      - name: susql-operator.v0.0.28
        version: 0.0.28
      - name: susql-operator.v0.0.27
        version: 0.0.27
      name: alpha
    defaultChannel: alpha
    packageName: susql-operator
    provider:
      name: SusQL Operator Contributors
      url: https://github.com/sustainable-computing-io/susql-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CodeFlare
      provider-url: https://github.com/project-codeflare
    name: codeflare-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: codeflare-operator.v1.16.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: quay.io/project-codeflare/codeflare-operator:v1.16.0
          createdAt: "2025-06-16T10:53:51Z"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/project-codeflare/codeflare-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AppWrapper
            name: appwrappers.workload.codeflare.dev
            version: v1beta2
        description: CodeFlare allows you to scale complex pipelines anywhere
        displayName: CodeFlare Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Pipelines
        - Scaling
        - App
        - Jobs
        links:
        - name: Codeflare Operator
          url: https://codeflare.dev/
        maintainers:
        - email: aasthana@redhat.com
          name: Anish Asthana
        - email: kpostlet@redhat.com
          name: Kevin Postlethwait
        - email: meyceoz@redhat.com
          name: Mustafa Eyceoz
        maturity: alpha
        provider:
          name: CodeFlare
          url: https://github.com/project-codeflare
        relatedImages:
        - quay.io/project-codeflare/codeflare-operator:v1.16.0
        version: 1.16.0
      entries:
      - name: codeflare-operator.v1.16.0
        version: 1.16.0
      - name: codeflare-operator.v1.15.0
        version: 1.15.0
      - name: codeflare-operator.v1.14.0
        version: 1.14.0
      - name: codeflare-operator.v1.13.0
        version: 1.13.0
      - name: codeflare-operator.v1.12.0
        version: 1.12.0
      - name: codeflare-operator.v1.11.0
        version: 1.11.0
      - name: codeflare-operator.v1.10.0
        version: 1.10.0
      - name: codeflare-operator.v1.9.0
        version: 1.9.0
      - name: codeflare-operator.v1.8.0
        version: 1.8.0
      - name: codeflare-operator.v1.7.0
        version: 1.7.0
      - name: codeflare-operator.v1.6.0
        version: 1.6.0
      - name: codeflare-operator.v1.5.0
        version: 1.5.0
      - name: codeflare-operator.v1.4.4
        version: 1.4.4
      - name: codeflare-operator.v1.4.3
        version: 1.4.3
      - name: codeflare-operator.v1.4.2
        version: 1.4.2
      - name: codeflare-operator.v1.4.1
        version: 1.4.1
      - name: codeflare-operator.v1.4.0
        version: 1.4.0
      - name: codeflare-operator.v1.3.1
        version: 1.3.1
      - name: codeflare-operator.v1.3.0
        version: 1.3.0
      - name: codeflare-operator.v1.2.0
        version: 1.2.0
      - name: codeflare-operator.v1.1.0
        version: 1.1.0
      - name: codeflare-operator.v1.0.1
        version: 1.0.1
      - name: codeflare-operator.v1.0.0
        version: 1.0.0
      - name: codeflare-operator.v1.0.0-rc.4
        version: 1.0.0-rc.4
      - name: codeflare-operator.v1.0.0-rc.3
        version: 1.0.0-rc.3
      - name: codeflare-operator.v1.0.0-rc.2
        version: 1.0.0-rc.2
      - name: codeflare-operator.v1.0.0-rc.1
        version: 1.0.0-rc.1
      - name: codeflare-operator.v0.2.3
        version: 0.2.3
      - name: codeflare-operator.v0.2.2
        version: 0.2.2
      - name: codeflare-operator.v0.2.1
        version: 0.2.1
      - name: codeflare-operator.v0.2.0
        version: 0.2.0
      - name: codeflare-operator.v0.1.0
        version: 0.1.0
      - name: codeflare-operator.v0.0.6
        version: 0.0.6
      - name: codeflare-operator.v0.0.5
        version: 0.0.5
      - name: codeflare-operator.v0.0.4
        version: 0.0.4
      - name: codeflare-operator.v0.0.3
        version: 0.0.3
      - name: codeflare-operator.v0.0.2
        version: 0.0.2
      - name: codeflare-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: codeflare-operator
    provider:
      name: CodeFlare
      url: https://github.com/project-codeflare
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: opdev.github.io
    name: airflow-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airflow-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "workflow.apache.org/v1alpha1",
                "kind": "AirFlow",
                "metadata": {
                  "name": "airflow-helm"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/airflow-helm-operator:0.0.1
          createdAt: "2023-06-28T16:57:21Z"
          description: An experimental operator that installs Apache Airflow.
          operatorframework.io/suggested-namespace: airflow-helm
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys Apache Airflow based on the official helm chart in
              openshift
            kind: AirFlow
            name: airflows.workflow.apache.org
            version: v1alpha1
        description: Deploys Apache Airflow based on the official helm chart with
          openshift specific modifications applied by default
        displayName: Airflow Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        - data
        - batch
        - scheduler
        - data workflow
        links:
        - name: Airflow Helm Operator
          url: https://airflow-helm-operator.domain
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: opdev.github.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/opdev/airflow-helm-operator:0.0.1
        version: 0.0.1
      entries:
      - name: airflow-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: airflow-helm-operator
    provider:
      name: opdev
      url: opdev.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-sso
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-sso.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Sso",
                "metadata": {
                  "name": "sso-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "kcos-keycloak": {
                    "auth": {
                      "baseURL": "https://keycloak.kcos-sso.svc.cluster.local:8443",
                      "enabled": true,
                      "sessionCookie": "AT"
                    },
                    "authproxy": {
                      "image": "prod.harbor.keysight.digital/loadcore/kcos-auth-proxy@sha256:a777ee5116280fbeb1e1342445fef5135f9ca61e983a3e2821f9ca4b2038b044",
                      "initContainer": {
                        "image": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf"
                      }
                    },
                    "crsCleanupJob": {
                      "clientExec": "oc",
                      "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "realm": {
                      "accessTokenLifespan": 266400,
                      "accessTokenLifespanForImplicitFlow": 266400,
                      "admin": {
                        "password": "admin"
                      },
                      "offlineSessionIdleTimeout": 2592000,
                      "offlineSessionMaxLifespan": 5184000,
                      "registrationAllowed": true,
                      "ssoSessionIdleTimeout": 266400,
                      "ssoSessionMaxLifespan": 266400
                    }
                  },
                  "subcharts": {
                    "enable": {
                      "kcos-keycloak": true,
                      "keycloak-operator": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:29:57Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Sso
            name: ssoes.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS SSO Operator
        displayName: Keysight KCOS SSO Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Sso
          url: https://keysight-kcos-sso.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/kcos-auth-proxy@sha256:a777ee5116280fbeb1e1342445fef5135f9ca61e983a3e2821f9ca4b2038b044
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-sso@sha256:298417fac2fc053bdc352f77b59b14e39eadd66c871cfb547304a517489f294e
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        version: 0.1.21
      entries:
      - name: keysight-kcos-sso.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-sso
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-cpd-wspipelines
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-wspipelines.v10.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wspipelines.cpd.ibm.com/v1beta1",
                "kind": "WSPipelines",
                "metadata": {
                  "name": "wspipelines-cr",
                  "labels": {
                    "app.kubernetes.io/name": "wspipelines",
                    "app.kubernetes.io/instance": "wspipelines.wspipelines.cpd.ibm.com",
                    "app.kubernetes.io/managed-by": "wspipelines-operator",
                    "wspipelinesBuildNumber": "2017"
                }
                },
                "spec": {
                  "version": "5.1.3",
                  "scaleConfig": "small",
                  "license": {
                    "accept": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:f9f616d17f816468957b69bbb09e6c92d678acb4ba7a40f1cb493d08fb98deb8
          createdAt: "2025-02-11"
          description: WS Pipelines CSV file
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-wspipelines
          olm.skipRange: '>=1.0.0 <10.3.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM watsonx.ai"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation For additional details regarding install parameters
              check: https://ibm.biz/BdaaL9 or https://ibm.biz/BdaaLC. License By
              installing this product you accept the license terms https://ibm.biz/BdaaL9
              or https://ibm.biz/BdaaLC.'
            kind: WSPipelines
            name: wspipelines.wspipelines.cpd.ibm.com
            version: v1beta1
        description: |-
          # Name
          IBM&reg; Watson Pipelines
          # Introduction
          ## Summary
          The Watson Pipelines editor provides a graphical interface for orchestrating an end-to-end flow of assets from creation through deployment. It enables to automate an end-to-end flow to prepare data, then create, train, deploy, monitor and update machine learning models and Python scripts. To design a pipeline you drag nodes onto the canvas, specify objects and parameters, then run and monitor the pipeline.
          ## Features
          Watson Pipelines are built off of Kubeflow pipelines on Tekton runtime and are fully integrated into the Watson Studio platform, allowing users to combine tools including:
          - Notebooks
          - Data refinery flows
          - AutoAI experiments
          - Web service / online deployments
          - Batch deployments
          - Import and export of Project / Space assets
          into repeatable and automated pipelines.
          # Details
          ## Prerequisites
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-pre-installation-tasks)
          ### Resources Required
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=planning-system-requirements)
          # Installing
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing)
          ## Configuration
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing)
          ## Storage
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=planning-storage-considerations)
          ## Limitations
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=issues-watson-studio)
          ## Documentation
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=overview)
          ## PodSecurityPolicy Requirements
          Custom PodSecurityPolicy definition:
          ```
          Not Applicable
          ```
          ## SecurityContextConstraints Requirements
          Custom SecurityContextConstraints definition:
          ```
          apiVersion: security.openshift.io/v1
          kind: SecurityContextConstraints
          allowHostDirVolumePlugin: false
          allowHostIPC: false
          allowHostNetwork: false
          allowHostPID: false
          allowHostPorts: false
          allowPrivilegeEscalation: false
          allowPrivilegedContainer: false
          allowedCapabilities: null
          defaultAddCapabilities: null
          fsGroup:
            type: MustRunAs
          metadata:
            annotations:
              include.release.openshift.io/self-managed-high-availability: "true"
              kubernetes.io/description: restrict denies access to all host features and requires pods to be run with a UID, and SELinux context that are allocated to the namespace.  This is the most restrictive SCC and it is used by default for authenticated users.
              release.openshift.io/create-only: "true"
            name: restrict
          priority: null
          readOnlyRootFilesystem: false
          requiredDropCapabilities:
          - ALL
          runAsUser:
            type: MustRunAsRange
          seLinuxContext:
            type: MustRunAs
          supplementalGroups:
            type: RunAsAny
          users: []
          volumes:
          - configMap
          - downwardAPI
          - emptyDir
          - persistentVolumeClaim
          - projected
          - secret
          ```
        displayName: Watson Pipelines
        installModes:
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: OwnNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - wspipelines
        - WSPipelines
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: v1
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:e4888fadfadb3462e3a9ad0395a526a2631872f70caf291dc18b60269ded199f
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:00f8bcee4ba7e0c5cf5f0c52e9e7707df367041c9f4dd5508dbefe0dbdba8877
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:a2949c3868c2c91bf35e50c64b51fcc24372986744b55520e21e9ce31fde59f0
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:e29885c1ee083a73b952b0de8d1cd675b2b2830abf44d69a8577f56972a601fb
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:ad2215adfb8dd2dc600bf5097b7f88c257d5fbaf541409c01b36cdb67c130236
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:ebcfc8941330da6991480dc18ac34d6df4fe879e562ee537e5c36df73f4bc536
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a24f4cb635e8b5759a5e2bda20ddefc6d5ea9b16fb1081592885c54aefb5480c
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:61155acdb8982648ae5748be9f5944ba913db1590ec3e773206d6fbb1d8952cb
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:25a4bb9cea1c3bdbeeeb205622512bcba98163f9629eba35625756346ce9d652
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:8d7625a0ba656425f601b6ec9a398a6d30511fe1de119cf0e29f680c704ea225
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:f9f616d17f816468957b69bbb09e6c92d678acb4ba7a40f1cb493d08fb98deb8
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:48b7cfa31b9a185176c20afc119e37346a6f9380d74000edf6c729b4677ca44e
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:29cd5c9cd60bbcfe211337786616622b911e2319db9870eb9d38d4f630c03e75
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:0164d8fbea591bbd9d67d84603162c3781b08551eddcc2a8a842aafc828af18c
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:19bb32e0389361e8314e627bcad2e9bd93984289485cf632466834353d9c3449
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:f0c68715f78bd20c352c97dd48ac5a912cad58e19a00159b5e6878f6beeb00fb
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:a477d47eaf8f8e569601d3c23c1def66f31c098ac3f5a1ae350618c460917dd8
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:99d0b3192e6e46167423505f48655e7d49af1a461ae00b95c9c7c9d4209e5b65
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:164f6063f3a50a36e28a240ae7c060c4f90266b7ed799bfdc075f90628bea7ff
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-bundle@sha256:842f2ac568db4db79cf0450b413b4b09b5f2180ef3541385b46fb650ac7066c0
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:6fc976bc622c2165dd8bf93c9244e1c29b7c973a4da7a5ce4d0a097a1c1ea4fb
        version: 10.3.0
      entries:
      - name: ibm-cpd-wspipelines.v10.3.0
        version: 10.3.0
      name: v10.3
    defaultChannel: v10.3
    packageName: ibm-cpd-wspipelines
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-enterprise
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-enterprise.v1.33.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "my-replica-set"
                },
                "spec": {
                  "credentials": "my-credentials",
                  "members": 3,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "sample-sharded-cluster"
                },
                "spec": {
                  "configServerCount": 3,
                  "credentials": "my-credentials",
                  "mongodsPerShardCount": 3,
                  "mongosCount": 2,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "shardCount": 1,
                  "type": "ShardedCluster",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBMultiCluster",
                "metadata": {
                  "name": "multi-replica-set"
                },
                "spec": {
                  "clusterSpecList": [
                    {
                      "clusterName": "e2e.cluster1.mongokubernetes.com",
                      "members": 2
                    },
                    {
                      "clusterName": "e2e.cluster2.mongokubernetes.com",
                      "members": 1
                    },
                    {
                      "clusterName": "e2e.cluster3.mongokubernetes.com",
                      "members": 2
                    }
                  ],
                  "credentials": "my-credentials",
                  "duplicateServiceObjects": false,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBOpsManager",
                "metadata": {
                  "name": "ops-manager"
                },
                "spec": {
                  "adminCredentials": "ops-manager-admin",
                  "applicationDatabase": {
                    "members": 3,
                    "podSpec": {
                      "cpu": 1
                    }
                  },
                  "configuration": {
                    "mms.fromEmailAddr": "admin@thecompany.com"
                  },
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "version": "6.0.3"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBUser",
                "metadata": {
                  "name": "my-replica-set-x509-user"
                },
                "spec": {
                  "db": "$external",
                  "mongodbResourceRef": {
                    "name": "my-replica-set"
                  },
                  "roles": [
                    {
                      "db": "admin",
                      "name": "dbOwner"
                    }
                  ],
                  "username": "CN=my-replica-set-x509-user,OU=cloud,O=MongoDB,L=New York,ST=New York,C=US"
                },
                "status": {
                  "phase": "Updated"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/mongodb/mongodb-enterprise-operator-ubi@sha256:6d394a5247cc346c4421bd9b37948c0378970dd3f4fc5d6eaac681f4c9655957
          createdAt: "2025-04-15T11:32:47Z"
          description: The MongoDB Enterprise Kubernetes Operator enables easy deploys
            of MongoDB into Kubernetes clusters, using our management, monitoring
            and backup platforms, Ops Manager and Cloud Manager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mongodb/mongodb-enterprise-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MongoDB Deployment
            displayName: MongoDB Deployment
            kind: MongoDB
            name: mongodb.mongodb.com
            version: v1
          - description: MongoDB Multi Deployment
            displayName: MongoDB Multi Deployment
            kind: MongoDBMultiCluster
            name: mongodbmulticluster.mongodb.com
            version: v1
          - description: MongoDB x509 User
            displayName: MongoDB User
            kind: MongoDBUser
            name: mongodbusers.mongodb.com
            version: v1
          - description: MongoDB Ops Manager
            displayName: MongoDB Ops Manager
            kind: MongoDBOpsManager
            name: opsmanagers.mongodb.com
            version: v1
        description: |
          The MongoDB Enterprise Kubernetes Operator enables easy deploys of MongoDB
          into Kubernetes clusters, using our management, monitoring and backup
          platforms, Ops Manager and Cloud Manager.

          ## Before You Start

          To start using the operator you''ll need an account in MongoDB Cloud Manager or
          a MongoDB Ops Manager deployment.

          * [Create a Secret with your OpsManager API key](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-operator-credentials/#procedure)

          * [Create a ConfigMap with your OpsManager project ID and URL](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-project-using-configmap/)

          By installing this integration, you will be able to deploy MongoDB instances
          with a single simple command.

          ## Required Parameters

          * `opsManager` or `cloudManager` - Enter the name of the ConfigMap containing project information
          * `credentials` - Enter the name of the Secret containing your OpsManager credentials
          * `type` - Enter MongoDB Deployment Types ("Standalone", "ReplicaSet", "ShardedCluster"

          ## Supported MongoDB Deployment Types ##

          * Standalone: An instance of mongod that is running as a single server and
          not as part of a replica set, this is, it does not do any kind of
          replication.

          * Replica Set: A replica set in MongoDB is a group of mongod processes that
          maintain the same data set. Replica sets provide redundancy and high
          availability, and are the basis for all production deployments. This section
          introduces replication in MongoDB as well as the components and architecture
          of replica sets. The section also provides tutorials for common tasks
          related to replica sets.

          * Sharded Cluster: The set of nodes comprising a sharded MongoDB deployment.
          A sharded cluster consists of config servers, shards, and one or more mongos
          routing processes. Sharding is a A database architecture that partitions
          data by key ranges and distributes the data among two or more database
          instances. Sharding enables horizontal scaling.

          ## Requirements for deploying MongoDB OpsManager

          * In order to deploy resources of type MongoDB OpsManager, you will need to
          create a secret containing the [credentials](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/plan-om-resource/#om-rsrc-prereqs)
          for the Global Owner user

          ## Security ##

          The operator can enable TLS for all traffic between servers and also between
          clients and servers. Before enabling `security.tls.enabled` to `true` you
          should create your certificates.  or you can leave the operator to create all
          the certificates for you. The operator ability to create certs is been
          deprecated due to Kubernetes API changes.

          For more information, please read the official MongoDB
          Kubernetes Operator  [docs](https://docs.mongodb.com/kubernetes-operator/stable/).
        displayName: MongoDB Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mongodb
        - databases
        - nosql
        links:
        - name: Documentation
          url: https://docs.opsmanager.mongodb.com/current/tutorial/install-k8s-operator/index.html
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: stable
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-agent-ubi@sha256:c96ad084edbdc6a6d907f57795fa38f39199b730328dbc64997d56ea50d55dfc
        - quay.io/mongodb/mongodb-agent-ubi@sha256:23d962d038888c6aba5540c8f5d81d7e86c4ccfaa07b9e159825d98676609f8e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7e5361c2a1604dec3fc671d629806fe6ddc6df524b69a441b86933a4b30093
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1d418c65d82ccad205b7889be4ba161d3cbfa8b1fb75b21c8fcf02de38fad9ca
        - quay.io/mongodb/mongodb-enterprise-server@sha256:df6081325ea2fe01d1b27d0c70005dec8232aaf51a43d91661b722b8a1761263
        - quay.io/mongodb/mongodb-enterprise-server@sha256:15e31285c8949a4666f7702df53416b0905d76945a28575a6028a876cc8a6f0f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:19925e1297ab652db59c35f53e7b7c8b2b277f8ecdc4215682961aaf2b3e924e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:67238129df09d33ee4746ff1acb5515d699a29d9f42350eae3f3f69ccb7fbf54
        - quay.io/mongodb/mongodb-agent-ubi@sha256:66b49d9dce414dcd657c3550b26f71964b82edfb67a41a3ce8d2386b04ef07e8
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0fde236fe26d5b20210d6f246aadacd7845874712def516cbe2eec11d79d5181
        - quay.io/mongodb/mongodb-agent-ubi@sha256:bd7edb8d0931dd952524f1f0ef99f61b0f69385d9e61efd034a4f45fd315a40b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d49417ea85ee1ddb84429552a5aac87a82c2843904d8c0300a43b657b7a49e0e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:7d93aee462c60cacc4e5e66729f5131c5c0fa434e6e6400c8e2ccac4bd5e20ba
        - quay.io/mongodb/mongodb-agent-ubi@sha256:5119a1ceeae6440dbe88b89550a7ae18de8e85c9a2059f467ad66adf7ec5d75f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:d3248506b6bd6b01c5973ab7bdac51ee41efa0ec8dd6211be4cc0b76a53a40c1
        - quay.io/mongodb/mongodb-agent-ubi@sha256:ee150e969c0bb786969c7e0a45fb1812766b752aee539072e28648817826433b
        - quay.io/mongodb/mongodb-agent-ubi@sha256:3b57a03dc17a301fda86c8afed1917c10d9b98947680116399c975ece944f5de
        - quay.io/mongodb/mongodb-enterprise-server@sha256:dfd4277f0c16ed476ac7de51894739e8c8d13a65ea5024b9a97b3dce0f6c6975
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c752a7f0726d21d55f204a359b85fa29cd727a5fe571b37289ae3ea52d6206b0
        - quay.io/mongodb/mongodb-agent-ubi@sha256:bd868284d9dfcb7c5a20e1160218d0172358384229b556c37e0789af0030eb24
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3add8b68bfbaa42596fe5dbbee4623ba5ff1a826b18bb71301525adf574dcb94
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f2d9a3264a8814f33143b73c56624572fb80afeab184133d3a901713a31c1a87
        - quay.io/mongodb/mongodb-agent-ubi@sha256:03763de5018fd5fa14843444d24adfc5e387605970b531801ef2d0916233a010
        - quay.io/mongodb/mongodb-enterprise-server@sha256:523566f44cbb6bbf0cd12b1adae939a883f2254c78dc4b8c9d0098581337b1db
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:c9a211a2c31b97eed3a25a814198547f2aa3470df93d661ca3700b828100d8de
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:985f770667e6c4c083e0c4bb79a948b83efa0cdc4ed3012b1dbf72fe48b38986
        - quay.io/mongodb/mongodb-enterprise-server@sha256:db63f6d93e03a3e644ef6ef4846ed7d24576c688ffb584226c78e1d1fba38878
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c07f20f8e7a5479e43cac7c74fc9f18cb4d4a1bd8b6b46f7dcc6f7ec7525fe28
        - quay.io/mongodb/mongodb-agent-ubi@sha256:411f85772e7a4b2c76e21d26d320b7743612080bf0cb4c7900d9a626cfeea33b
        - quay.io/mongodb/mongodb-agent-ubi@sha256:8467b1c2924a19929fc101ef2ff8d7e53cb0115a0cdc2e371dc0733bd0262395
        - quay.io/mongodb/mongodb-enterprise-init-ops-manager-ubi@sha256:6ccc8148c266a9037ea744e8d19875146b1c231909f29a8f01c7ce415ca15936
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a48a4d5c275fd669a27469649fb39ea38c728dbcf9f0a9da4e7cf27a2ce912ed
        - quay.io/mongodb/mongodb-agent-ubi@sha256:7144ac5e5038e92a6d142bf39eba46554c1b5fafeb8ea59f0390896a0d35ed17
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8445d59d491a6d3890ea9c8b421214bcd0fdcfbc420ed40497a790b79f0bc89e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d21c65780cadc9d8668d2c28935d492625d6993f13ba562722200a819281cbcc
        - quay.io/mongodb/mongodb-agent-ubi@sha256:2395fe9d38a1a9160f7cb2f400b776ea3fc2e72187d23b6dee7fd1e4fb9ca721
        - quay.io/mongodb/mongodb-agent-ubi@sha256:b175972792aa1be268d5fb3cf3ad0156f4aecffa323e1b473360fc2a3a8d8891
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b00ac773623e1bf422b47e6d783263d2d08737865abe9abbe106d86d1ea25165
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0b2d01d543d38feddcc9d351e711df33a0a5187b22d95cdb926f6015ee1093bd
        - quay.io/mongodb/mongodb-enterprise-server@sha256:35b263861d6ea66b541abc190ce1aaf4d4e36f128e5d8ba2a5e9fa8e25d84446
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f1bae6375f846af6fca7faeeff4aedf9fab41a9d45b515ba74e9fd7336408617
        - quay.io/mongodb/mongodb-enterprise-operator-ubi@sha256:6d394a5247cc346c4421bd9b37948c0378970dd3f4fc5d6eaac681f4c9655957
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5421e00c0935b7a9ced51a11e935773ef5df4d259f3e1b1930fbf2f1b9d46586
        - quay.io/mongodb/mongodb-agent-ubi@sha256:97bb6bb5ecd5a802e8a91e69aa10d8bde0f648b106dc5591d49ee94209822daa
        - quay.io/mongodb/mongodb-agent-ubi@sha256:6ebe232ed0ea87045792e578ecb11dd46411333ad0ea61383c247704b4a0efe9
        - quay.io/mongodb/mongodb-enterprise-database-ubi@sha256:e2ca129f04ad6bf4e0dbb988badb083187725f28bac01bb7080cc80cad9ac316
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:2127d7f2684ae36b123422f55237055aad02fdb71c4c21a5b8f0b0f6d8baedfc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa8683500e8d3632c4588bbcdddb4e3186a01f48aefe1b6d95f8860dae8b6c76
        - quay.io/mongodb/mongodb-enterprise-server@sha256:110a560f5a0ebe1fde1b2c20ae5ff5638682229d1c8656b993bdaf7b71eb94bc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:172e338a43df01ee5408d8d1e3cb038d2c55e77761c1a2ac6ed37959d1525140
        - quay.io/mongodb/mongodb-enterprise-server@sha256:584255582db31d0bb1f9593feb3b675d934d8750766fc98d3a1b3340feae4995
        - quay.io/mongodb/mongodb-enterprise-server@sha256:605dcca06a45e5ea127edb9ff262a1c3ddafdab2f30b6d3a18f73bfb8a6acdb0
        - quay.io/mongodb/mongodb-agent-ubi@sha256:85f883076bc8edc516800704e2290c54b1bee14a9a533df84928c7ce1ae807d2
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8f8735d377d214f64961381b5b1ed7cd3c306e6354dbc6b45c51134a25444b93
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:83af86f8098e3c81b6e0879eabd582592ef6cbb1d6592f5e175c0794b9df2a0e
        - quay.io/mongodb/mongodb-enterprise-init-appdb-ubi@sha256:be4295843d92ce0f325b8daaeb3a5490c11772d0348cbd9ea35a6126d038b7fb
        - quay.io/mongodb/mongodb-agent-ubi@sha256:575a43be4c667f3249f46f1fdcc06dd2e68cacc9eed989da0f6edd7e214d51b7
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0ee80032abe873b330537d4b64511ee76c5efb512af95287f28ddeffab13bb6d
        - quay.io/mongodb/mongodb-agent-ubi@sha256:b23cfa639c5a4d20381245772dfe396ec915af3a3046df5d9b10016b900431f5
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ca2aaf57d70a0b9d743a6533c4e1b88324056197453de540d3352e679a3c9c92
        - quay.io/mongodb/mongodb-agent-ubi@sha256:38175af108178d89e4d3359ee2754534e5e9b24e843f12711fc14fe1dead516a
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:2be0c6c89898b18618932b6a6303aeefdde60f48ec37d5c1fca5b8d04a00fd78
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7aa00dca44559318356627e88440765576dd6706187044edd1f21d0d0dcb0bf6
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:f64d99a77d6a6b9ed10de9a28b1b8b8155414eca2e561f3fdb04bd66914fb3c5
        - quay.io/mongodb/mongodb-agent-ubi@sha256:2dbaaece4dd48087370e43d9bcf6b0a0312f86cb5dfe916e8725323bd5401992
        - quay.io/mongodb/mongodb-agent-ubi@sha256:0f8d102780a9d2ce54b7453f9575d0e24b15f014cd9b74816b1e6a4090b979b2
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7885ab6e88214f9289d8babf5189136dde0dd4c84a15a5481ccd3ef512b2e66f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:37af0913742b6de339c7e664b30f35dcef04506529e4a09a9bbbdd57643f7149
        - quay.io/mongodb/mongodb-agent-ubi@sha256:c03e4a7663206f54c5ef39974502816f66d8d64936c19849f0337ea4547e614e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d876f57ea45582e84748b02459748e440af443a7eb0882585638973409c08f6a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:13186ad45a2ba3ab55ed5c2a008eca494ee3937c8519475379476cc8e6f1d055
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ef732e3eea91cc86caf2eabedc0c9115ed235a8d204f33e74bbf090a0b0259fb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:85c1d669c073caf9cbaac815e4c8ea2ca18b71deefa03bf84cd6d8b8ebd53e12
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d323c1bcd14118cc537ac282996ea5a58756d20b6b73f1eec0efa8710d246077
        - quay.io/mongodb/mongodb-agent-ubi@sha256:36c84115c6c645fd7d65d1004d2507ef3592aa1f07176645b1ee8e8b2a4f5f73
        - quay.io/mongodb/mongodb-agent-ubi@sha256:31cfefd2660da3efe1b6d63e39090464c28ae006d6ed1dffb2189a09183b9c61
        - quay.io/mongodb/mongodb-enterprise-server@sha256:cd0cdf1b2a58f4c3b514e602cc1d96543410f631825045bbf65bd3fb84113f1f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:fb8397cbcdbf7c33f6ad142492792771127f0d9a8c46388f350b6927405ee590
        - quay.io/mongodb/mongodb-agent-ubi@sha256:9d1f21df008b14ca721c9ed0ff905af23e1a56b64cd4eddab4df1e8e9b63df8c
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9a3a6496be0d2d446e6d654eb56386c7726e0cd2d7232f6dcec579923edf0014
        - quay.io/mongodb/mongodb-agent-ubi@sha256:766e6cbeec27f995211aa509ff7f118b6945f3e728141f28f9c4ea0dd2297af5
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:dfec259209995bb040a1eff1fcafb1d6c3fdd145129f12c04a56ec74834b80a6
        - quay.io/mongodb/mongodb-enterprise-init-database-ubi@sha256:ab12a330e6b40735ee516168073887a4f37b58694bd10026a451eb9972e55500
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c54eeb910855594ed4bf5dcbbbcc14aa6f733e97313896a8381fb4be6aeac583
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa300abf739e26350a258f643fa8e57aedacd041055c7b923bd22b21ef554cdb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:11ab324698f919f84a5dab5a964d93fe63f98a48146bb3b86974cba7606d30ec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b2bd8eb7f8b96adb874468ff1b0ebae2db4ebe490b6e1da4c9a6c5108f813ff5
        - quay.io/mongodb/mongodb-agent-ubi@sha256:695074fbbfcadf454230b80fbf1094260e631f6108cefb2050473edaee2a217a
        - quay.io/mongodb/mongodb-agent-ubi@sha256:8c11348eac180f5faa113db2804817302ec78847472e1891b063744adba4aa60
        - quay.io/mongodb/mongodb-enterprise-server@sha256:33a30d11969eb21952681108092fe7f916e25c8e1de43e6a4e40e126025a2bec
        - quay.io/mongodb/mongodb-agent-ubi@sha256:d927afa9d3d569d943518f0a684e19afd4adb0875f55df7c904b9dd019589e5e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:442f5d99d02674099525abaa69daa42cec758ce5554198f978927aa5c49006c8
        - quay.io/mongodb/mongodb-agent-ubi@sha256:f5f3b14ee6533b21d89d9ed6d4af47dbdee9e071f909dc686dfcdf6742179e46
        - quay.io/mongodb/mongodb-agent-ubi@sha256:6d007271faed24bdf06eb64f2b54e5cdc3e11ce71840a7f586dabe2ee02c640e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:655896bac9e82de8e9729d59a36c3a31f66235f22b93449260a2ccb7709de084
        - quay.io/mongodb/mongodb-enterprise-server@sha256:83db493ac6158d654d3ce55619f312d53cfbac258e0faf22e9110ca004e174bd
        version: 1.33.0
      entries:
      - name: mongodb-enterprise.v1.33.0
        version: 1.33.0
      - name: mongodb-enterprise.v1.32.0
        version: 1.32.0
      - name: mongodb-enterprise.v1.31.0
        version: 1.31.0
      - name: mongodb-enterprise.v1.30.0
        version: 1.30.0
      - name: mongodb-enterprise.v1.29.0
        version: 1.29.0
      - name: mongodb-enterprise.v1.28.0
        version: 1.28.0
      - name: mongodb-enterprise.v1.27.0
        version: 1.27.0
      - name: mongodb-enterprise.v1.26.0
        version: 1.26.0
      - name: mongodb-enterprise.v1.25.0
        version: 1.25.0
      - name: mongodb-enterprise.v1.24.0
        version: 1.24.0
      - name: mongodb-enterprise.v1.23.0
        version: 1.23.0
      - name: mongodb-enterprise.v1.22.0
        version: 1.22.0
      - name: mongodb-enterprise.v1.21.0
        version: 1.21.0
      - name: mongodb-enterprise.v1.20.1
        version: 1.20.1
      - name: mongodb-enterprise.v1.19.1
        version: 1.19.1
      - name: mongodb-enterprise.v1.18.0
        version: 1.18.0
      - name: mongodb-enterprise.v1.17.2
        version: 1.17.2
      - name: mongodb-enterprise.v1.17.1
        version: 1.17.1
      - name: mongodb-enterprise.v1.17.0
        version: 1.17.0
      - name: mongodb-enterprise.v1.16.4
        version: 1.16.4
      - name: mongodb-enterprise.v1.16.3
        version: 1.16.3
      - name: mongodb-enterprise.v1.16.2
        version: 1.16.2
      - name: mongodb-enterprise.v1.16.1
        version: 1.16.1
      name: stable
    defaultChannel: stable
    packageName: mongodb-enterprise
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: MariaDB Corporation
      provider-url: ""
    name: mariadb-enterprise-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-enterprise-operator.v25.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup"
                },
                "spec": {
                  "args": [
                    "--single-transaction",
                    "--all-databases"
                  ],
                  "logLevel": "info",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "0 0 * * *",
                    "suspend": false
                  },
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "backups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Connection",
                "metadata": {
                  "name": "connection"
                },
                "spec": {
                  "database": "mariadb",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "secretName": "connection",
                  "secretTemplate": {
                    "key": "dsn"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Database",
                "metadata": {
                  "name": "database"
                },
                "spec": {
                  "characterSet": "utf8",
                  "collate": "utf8_general_ci",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "name": "my-logical-database"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "grant"
                },
                "spec": {
                  "database": "*",
                  "grantOption": true,
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "privileges": [
                    "SELECT",
                    "INSERT",
                    "UPDATE"
                  ],
                  "table": "*",
                  "username": "my-user"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "MariaDB",
                "metadata": {
                  "name": "mariadb"
                },
                "spec": {
                  "connection": {
                    "secretName": "mariadb-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "database": "mariadb",
                  "galera": {
                    "enabled": true
                  },
                  "metrics": {
                    "enabled": true,
                    "passwordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "mariadb-metrics"
                    }
                  },
                  "myCnf": "[mariadb]\nbind-address=*\ndefault_storage_engine=InnoDB\nbinlog_format=row\ninnodb_autoinc_lock_mode=2\ninnodb_buffer_pool_size=1024M\nmax_allowed_packet=256M\n",
                  "passwordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-password"
                  },
                  "primaryConnection": {
                    "secretName": "mariadb-conn-primary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "primaryService": {
                    "type": "ClusterIP"
                  },
                  "replicas": 3,
                  "rootPasswordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-root"
                  },
                  "secondaryConnection": {
                    "secretName": "mariadb-conn-secondary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "secondaryService": {
                    "type": "ClusterIP"
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "storage": {
                    "size": "1Gi"
                  },
                  "updateStrategy": {
                    "type": "ReplicasFirstPrimaryLast"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "MaxScale",
                "metadata": {
                  "name": "maxscale"
                },
                "spec": {
                  "admin": {
                    "guiEnabled": true,
                    "port": 8989
                  },
                  "auth": {
                    "adminPasswordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "maxscale-admin"
                    },
                    "adminUsername": "mariadb-enterprise-operator",
                    "generate": true
                  },
                  "config": {
                    "sync": {
                      "database": "mysql",
                      "interval": "5s",
                      "timeout": "10s"
                    }
                  },
                  "connection": {
                    "port": 3306,
                    "secretName": "maxscale-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "guiKubernetesService": {
                    "type": "ClusterIP"
                  },
                  "kubernetesService": {
                    "type": "ClusterIP"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "metrics": {
                    "enabled": true
                  },
                  "monitor": {
                    "cooperativeMonitoring": "majority_of_all",
                    "interval": "2s",
                    "params": {
                      "available_when_donor": "false",
                      "disable_master_failback": "false",
                      "disable_master_role_setting": "false"
                    }
                  },
                  "replicas": 2,
                  "requeueInterval": "10s",
                  "services": [
                    {
                      "listener": {
                        "port": 3306,
                        "protocol": "MariaDBProtocol"
                      },
                      "name": "rw-router",
                      "params": {
                        "transaction_replay": "true",
                        "transaction_replay_attempts": "10",
                        "transaction_replay_timeout": "5s"
                      },
                      "router": "readwritesplit"
                    }
                  ]
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "PhysicalBackup",
                "metadata": {
                  "name": "physicalbackup"
                },
                "spec": {
                  "compression": "bzip2",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "podAffinity": true,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "0 0 * * *",
                    "immediate": true,
                    "suspend": false
                  },
                  "serviceAccountName": "backup",
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "physicalbackups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "region": "us-east-1",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  },
                  "timeout": "1h"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore"
                },
                "spec": {
                  "backupRef": {
                    "name": "backup"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "targetRecoveryTime": "2023-12-19T09:00:00Z"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "SqlJob",
                "metadata": {
                  "name": "sqljob"
                },
                "spec": {
                  "database": "mariadb",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "sql": "CREATE TABLE IF NOT EXISTS orders (\n  id bigint PRIMARY KEY AUTO_INCREMENT,\n  customer_id INT,\n  order_date DATE,\n  total_amount_cents INT\n);\nINSERT INTO orders (customer_id, order_date, total_amount_cents) VALUES\n  (101, '2024-05-15', 5000),\n  (102, '2024-05-14', 7525),\n  (103, '2024-05-13', 10050),\n  (104, '2024-05-12', 2575);",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "user"
                },
                "spec": {
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxUserConnections": 20,
                  "name": "my-user",
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.mariadb.com/mariadb-enterprise-operator@sha256:1effbbfa7632be92c17ef61c18a25f010b187840beec42c31d2410d0209e6985
          createdAt: "2025-08-28T11:35:03Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["MariaDB Enterprise"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: MariaDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API. It is used to define
              backup jobs and its storage.
            displayName: Backup
            kind: Backup
            name: backups.enterprise.mariadb.com
            version: v1alpha1
          - description: Connection is the Schema for the connections API. It is used
              to configure connection strings for the applications connecting to MariaDB.
            displayName: Connection
            kind: Connection
            name: connections.enterprise.mariadb.com
            version: v1alpha1
          - description: Database is the Schema for the databases API. It is used
              to define a logical database as if you were running a 'CREATE DATABASE'
              statement.
            displayName: Database
            kind: Database
            name: databases.enterprise.mariadb.com
            version: v1alpha1
          - description: Grant is the Schema for the grants API. It is used to define
              grants as if you were running a 'GRANT' statement.
            displayName: Grant
            kind: Grant
            name: grants.enterprise.mariadb.com
            version: v1alpha1
          - description: MariaDB is the Schema for the mariadbs API. It is used to
              define MariaDB clusters.
            displayName: Maria DB
            kind: MariaDB
            name: mariadbs.enterprise.mariadb.com
            version: v1alpha1
          - description: MaxScale is the Schema for the maxscales API. It is used
              to define MaxScale clusters.
            displayName: Max Scale
            kind: MaxScale
            name: maxscales.enterprise.mariadb.com
            version: v1alpha1
          - description: PhysicalBackup is the Schema for the physicalbackups API.
              It is used to define physical backup jobs and its storage.
            displayName: Physical Backup
            kind: PhysicalBackup
            name: physicalbackups.enterprise.mariadb.com
            version: v1alpha1
          - description: Restore is the Schema for the restores API. It is used to
              define restore jobs and its restoration source.
            displayName: Restore
            kind: Restore
            name: restores.enterprise.mariadb.com
            version: v1alpha1
          - description: SqlJob is the Schema for the sqljobs API. It is used to run
              sql scripts as jobs.
            displayName: Sql Job
            kind: SqlJob
            name: sqljobs.enterprise.mariadb.com
            version: v1alpha1
          - description: User is the Schema for the users API.  It is used to define
              grants as if you were running a 'CREATE USER' statement.
            displayName: User
            kind: User
            name: users.enterprise.mariadb.com
            version: v1alpha1
        description: |
          MariaDB Enterprise Operator for Openshift is an operator that manages MariaDB Enterprise operands declaratively via Custom Resources, with support for the following features:

          * Provision and Configure MariaDB and MaxScale Declaratively: Define MariaDB Enterprise Server and MaxScale clusters in YAML manifests and deploy them with ease in Kubernetes.
          * High Availability with Galera: Ensure availability with MariaDB Enterprise Cluster, providing synchronous multi-master replication.
          * Query and Connection-Based Routing with MaxScale: MaxScale provides query routing and connection load balancing for improved application performance.
          * Cluster-Aware Rolling Updates: Perform rolling updates on MariaDB and MaxScale clusters, ensuring zero-downtime upgrades with no disruptions to your applications.
          * Flexible Storage Configuration and Volume Expansion: Easily configure storage for MariaDB instances, including the ability to expand volumes as needed.
          * Physical Backups based on [mariadb-backup](https://mariadb.com/docs/server/server-usage/backup-and-restore/mariadb-backup/full-backup-and-restore-with-mariadb-backup) and [Kubernetes VolumeSnapshots](https://kubernetes.io/docs/concepts/storage/volume-snapshots/). By leveraging the [BACKUP STAGE](https://mariadb.com/docs/server/reference/sql-statements/administrative-sql-statements/backup-commands/backup-stage) feature, backups are taken without long read locks or service interruptions.
          * Logical Backups based on [mariadb-dump](https://mariadb.com/docs/server/clients-and-utilities/backup-restore-and-import-clients/mariadb-dump).
          * Backup Management: Take, restore, and schedule backups with multiple storage types supported: S3, PVCs, Kubernetes volumes and VolumeSnapshots..
          * Policy-Driven Backup Retention: Implement backup retention policies with bzip2 and gzip compression.
          * Target Recovery Time: Restore your database to the closest available backup based on a specified recovery time.
          * Bootstrap New Instances: Initialize new MariaDB instances from backups, S3, PVCs or VolumeSnapshots to quickly spin up new clusters.
          * TLS Certificate Management: Issue, configure, and rotate TLS certificates and Certificate Authorities (CAs) for secure connections.
          * Advanced TLS Support: customize certificate lifetime, private key algorithm and TLS version.
          * Native Integration with cert-manager: Leverage [cert-manager](https://cert-manager.io/docs/), the de-facto standard for managing certificates in Kubernetes, to enable issuance with private CAs, public CAs and HashiCorp Vault.
          * Prometheus Metrics: Expose metrics using the MariaDB and MaxScale Prometheus exporters.
          * Native Integration with prometheus-operator: Leverage [prometheus-operator](https://github.com/prometheus-operator/prometheus-operator) to scrape metrics from MariaDB and MaxScale instances.
          * Declarative User and Database Management: Manage users, grants, and logical databases in a declarative manner using Kubernetes resources.
          * Secure, immutable and lightweight images based on Red Hat UBI, available for multiple architectires (amd64, arm64 and ppc64le).
          * [Operator certified ](https://catalog.redhat.com/en/software/container-stacks/detail/65789bcbe17f1b31944acb1d#overview) by Red Hat.

          ## Installation

          In order to use this product, you will need a [MariaDB Enterprise subscription](https://mariadb.com/products/enterprise/).

          To be able to pull MariaDB Enterprise images, you will need to configure the MariaDB registry credentials in Openshift using the [MariaDB customer download token](https://customers.mariadb.com/downloads/token/) as described below:

          - Extract your [Openshift global pull secret](https://docs.redhat.com/en/documentation/openshift_container_platform/4.18/html/images/managing-images#images-update-global-pull-secret_using-image-pull-secrets):
            ```bash
            oc extract secret/pull-secret -n openshift-config --confirm
            ```

          - Login in the MariaDB registry providing the customer download token as password:
            ```bash
            oc registry login \
              --registry="docker.mariadb.com" \
              --auth-basic="<email>:<customer-download-token>" \
              --to=.dockerconfigjson
            ```

          - Update the global pull secret:
            ```bash
            oc set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson
            ```

          Alternatively, you can also create a `Secret` named `mariadb-enterprise` in the namespace where are you installing the operator by using the following command:
            ```bash
            oc create secret docker-registry mariadb-enterprise \
              --docker-server=docker.mariadb.com \
              --docker-username=<email> \
              --docker-password=<customer-download-token> \
              -n <operator-namespace>
            ```

          At this point, you can proceed to install the MariaDB Enterprise Operator, either via the Openshift console or by creating a `Subscription` object as described in the [Openshift installation docs](https://mariadb.com/kb/en/openshift/).

          Refer to the [quickstart guide](https://mariadb.com/kb/en/quickstart/), the [examples catalog](https://mariadb.com/kb/en/examples-catalog/) and the [API reference](https://mariadb.com/kb/en/api-reference/) for getting started with the Custom Resources offered by this operator.
        displayName: MariaDB Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mariadb
        - mysql
        - database
        - operator
        - mariadb-enterprise-operator
        - database
        links:
        - name: MariaDB Enterprise
          url: https://mariadb.com/products/enterprise/
        - name: Documentation
          url: https://mariadb.com/kb/en/mariadb-enterprise-operator/
        maintainers:
        - email: operator@mariadb.com
          name: MariaDB Corporation
        - email: support@mariadb.com
          name: Support
        - email: sales-EMEA@mariadb.com
          name: Sales EMEA
        - email: sales-APAC@mariadb.com
          name: Sales APAC
        - email: sales-nam@mariadb.com
          name: Sales Americas
        maturity: beta
        minKubeVersion: 1.29.0
        provider:
          name: MariaDB Corporation
        relatedImages:
        - docker.mariadb.com/mariadb-enterprise-operator@sha256:1effbbfa7632be92c17ef61c18a25f010b187840beec42c31d2410d0209e6985
        - docker.mariadb.com/maxscale@sha256:d9e04a7a406353dab2051e190be0c32ed20b780ed71e56e55cedd4ccb61af416
        - mariadb/mariadb-prometheus-exporter-ubi@sha256:f072175198cad050e80617d13e63f415bacdfe47e2dd177fde7283a96ee56c07
        - mariadb/maxscale-prometheus-exporter-ubi@sha256:cc17c1554795ad01b4d8db799f6aa90e00e41ede431f09d30eec64d9dd4479d9
        - docker.mariadb.com/enterprise-server@sha256:9872721d35ebe8d820b1aa9c6dd83b363a7b2df9af11a65bff6b80ff4144460a
        version: 25.8.1-1
      entries:
      - name: mariadb-enterprise-operator.v25.8.1-1
        version: 25.8.1-1
      - name: mariadb-enterprise-operator.v25.8.0
        version: 25.8.0
      - name: mariadb-enterprise-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: mariadb-enterprise-operator
    provider:
      name: MariaDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Instana
      provider-url: ""
    name: instana-agent-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: instana-agent-operator.v2.1.36
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instana.io/v1",
                "kind": "InstanaAgent",
                "metadata": {
                  "name": "instana-agent",
                  "namespace": "instana-agent"
                },
                "spec": {
                  "agent": {
                    "configuration_yaml": "# You can leave this empty, or use this to configure your instana agent.\n# See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.\n",
                    "endpointHost": "ingress-red-saas.instana.io",
                    "endpointPort": "443",
                    "env": {},
                    "key": "replace-key"
                  },
                  "cluster": {
                    "name": "my-cluster"
                  },
                  "opentelemetry": {
                    "grpc": {
                      "enabled": true
                    },
                    "http": {
                      "enabled": true
                    }
                  },
                  "zone": {
                    "name": "edited-zone"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring,Logging & Tracing,OpenShift Optional
          certified: "false"
          containerImage: icr.io/instana/instana-agent-operator@sha256:ab8b35f0b45c68f69210d1b17c5dd609799f6bb0bc9f12d0d4e5302c910607a1
          description: Fully automated Application Performance Monitoring (APM) for
            microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/instana-agent-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/instana-agent-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/instana/instana-agent-operator
          support: Instana
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InstanaAgent is the Schema for the agents API
            displayName: Instana Agent
            kind: InstanaAgent
            name: agents.instana.io
            version: v1
          - description: InstanaAgentRemote is the Schema for the agents API
            displayName: Instana Agent Remote
            kind: InstanaAgentRemote
            name: agentsremote.instana.io
            version: v1
        description: |
          # Instana

          Instana is an [APM solution](https://www.instana.com/product-overview/) built for microservices that enables IT Ops to build applications faster and deliver higher quality services by automating monitoring, tracing and root cause analysis. The solution is optimized for [Kubernetes](https://www.instana.com/automatic-kubernetes-monitoring/) and [OpenShift](https://www.instana.com/blog/automatic-root-cause-analysis-for-openshift-applications/).

          ## Instana Agent Operator

          This is the Kubernetes Operator for installing the Instana Agent on Kubernetes or OpenShift.

          ## Prerequisites for OpenShift

          Before the agent will be able to run in OpenShift, you need to perform a couple of extra configuration steps.

          You need to set up a project for the Instana Agent and configure it's permissions.

          The project you create here needs to be the namespace where you create the Instana Agent custom resource that the operator will use to deploy the agent.

          For example, create the `instana-agent` project:

              oc new-project instana-agent

          Then, ensure the `instana-agent` service account is in the privileged security context:

              oc adm policy add-scc-to-user privileged -z instana-agent

          The `instana-agent-remote` service account must be allowed to use any userid:

              oc adm policy add-scc-to-user anyuid -z instana-agent-remote -n instana-agent

          This service account will be created by the operator.

          Now you can proceed with installing the operator for the Instana agent.

          ## Installation and Configuration

          First, install this operator from [OperatorHub.io](https://operatorhub.io/), [OpenShift Container Platform](https://www.openshift.com/), or [OKD](https://www.okd.io/).

          Second, create a custom resource with the agent configuration in the target namespace (for now, this MUST always be the `instana-agent` namespace). The operator will pick up the custom resource and install the Instana agent accordingly.

          The following is a minimal template of the custom resource:

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgent
          metadata:
            name: instana-agent
            namespace: instana-agent
          spec:
            zone:
              name: my-zone # (optional) name of the zone of the host
            cluster:
              name: replace-me # replace with the name of your Kubernetes cluster
            agent:
              key: replace-me # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io # the monitoring ingress endpoint
              endpointPort: "443" # the monitoring ingress endpoint port, wrapped in quotes
              env:
                INSTANA_AGENT_TAGS: example
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://ibm.biz/monitoring-k8s
          ```

          Save the template in a file `instana-agent.yaml` and edit the following values:

          * If your target namespace is not `instana-agent`, replace the `namespace:` accordingly.
          * `agent.key` must be set with your Instana agent key.
          * `agent.endpointHost` must be set with the monitoring ingress endpoint, generally either `saas-us-west-2.instana.io` or `saas-eu-west-1.instana.io`.
          * `agent.endpointPort` must be set with the monitoring ingress port, generally "443" (wrapped in quotes).
          * `zone.name` should be set with the name of the Kubernetes cluster that is be displayed in Instana.

          For advanced configuration, you can edit the contents of the `configuration.yaml` file. View documentation [here](https://ibm.biz/monitoring-k8s).

          Apply the custom resource with `kubectl apply -f instana-agent.yaml`. After some time, you should see `instana-agent` Pods being created on each node of your cluster, and your cluster should show on the infrastructure map on your Instana Web interface.

          In addition to the InstanaAgent custom resource, a remote agent can be deployed. This agent will not monitor workload running inside the cluster, but can deploy a single replica of an agent which can get an independent configuration for running remote monitoring for certain sensors.

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgentRemote
          metadata:
            name: remote-agent
            namespace: "instana-agent"
          spec:
            zone:
              name: edited-zone # name of the zone of the host
            hostname:
              name: use-custom-hostname-for-remote-agent # (optional) give your remote agent pod a custom hostname
            agent:
              key: replace-key # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io
              endpointPort: "443"
              env: {}
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.
          ```

          ## Uninstalling

          In order to uninstall the Instana agent, simply remove the custom resource with `kubectl delete -f instana-agent.yaml`.

          ## Source Code

          The Instana agent operator is an open source project hosted on [https://github.com/instana/instana-agent-operator](https://github.com/instana/instana-agent-operator/).
        displayName: Instana Agent Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - APM
        - Instana
        links:
        - name: GitHub Repository
          url: https://github.com/instana/instana-agent-operator
        maintainers:
        - email: support@instana.com
          name: Instana
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: Instana
        relatedImages:
        - icr.io/instana/instana-agent-operator@sha256:ab8b35f0b45c68f69210d1b17c5dd609799f6bb0bc9f12d0d4e5302c910607a1
        - icr.io/instana/agent@sha256:ed71e6c3e5bdb155853a2db8c916460e6a867fb02627e3a332186188cbbd98dd
        version: 2.1.36
      entries:
      - name: instana-agent-operator.v2.1.36
        version: 2.1.36
      - name: instana-agent-operator.v2.1.35
        version: 2.1.35
      - name: instana-agent-operator.v2.1.34
        version: 2.1.34
      - name: instana-agent-operator.v2.1.33
        version: 2.1.33
      - name: instana-agent-operator.v2.1.32
        version: 2.1.32
      - name: instana-agent-operator.v2.1.31
        version: 2.1.31
      - name: instana-agent-operator.v2.1.30
        version: 2.1.30
      - name: instana-agent-operator.v2.1.28
        version: 2.1.28
      - name: instana-agent-operator.v2.1.27
        version: 2.1.27
      - name: instana-agent-operator.v2.1.24
        version: 2.1.24
      - name: instana-agent-operator.v2.1.23
        version: 2.1.23
      - name: instana-agent-operator.v2.1.22
        version: 2.1.22
      - name: instana-agent-operator.v2.1.21
        version: 2.1.21
      - name: instana-agent-operator.v2.1.20
        version: 2.1.20
      - name: instana-agent-operator.v2.1.19
        version: 2.1.19
      - name: instana-agent-operator.v2.1.18
        version: 2.1.18
      - name: instana-agent-operator.v2.1.17
        version: 2.1.17
      - name: instana-agent-operator.v2.1.16
        version: 2.1.16
      - name: instana-agent-operator.v2.1.15
        version: 2.1.15
      - name: instana-agent-operator.v2.1.14
        version: 2.1.14
      - name: instana-agent-operator.v2.1.13
        version: 2.1.13
      - name: instana-agent-operator.v2.1.12
        version: 2.1.12
      - name: instana-agent-operator.v2.1.11
        version: 2.1.11
      - name: instana-agent-operator.v2.1.10
        version: 2.1.10
      - name: instana-agent-operator.v2.1.9
        version: 2.1.9
      - name: instana-agent-operator.v2.1.8
        version: 2.1.8
      - name: instana-agent-operator.v2.1.7
        version: 2.1.7
      - name: instana-agent-operator.v2.1.6
        version: 2.1.6
      - name: instana-agent-operator.v2.1.5
        version: 2.1.5
      - name: instana-agent-operator.v2.1.4
        version: 2.1.4
      - name: instana-agent-operator.v2.1.3
        version: 2.1.3
      - name: instana-agent-operator.v2.1.2
        version: 2.1.2
      - name: instana-agent-operator.v2.1.1
        version: 2.1.1
      - name: instana-agent-operator.v2.1.0
        version: 2.1.0
      - name: instana-agent-operator.v2.0.32
        version: 2.0.32
      - name: instana-agent-operator.v2.0.31
        version: 2.0.31
      - name: instana-agent-operator-rhmp.v2.0.30
        version: 2.0.30
      - name: instana-agent-operator.v2.0.27
        version: 2.0.27
      - name: instana-agent-operator.v2.0.19
        version: 2.0.19
      - name: instana-agent-operator.v2.0.18
        version: 2.0.18
      - name: instana-agent-operator.v2.0.17
        version: 2.0.17
      - name: instana-agent-operator.v2.0.13
        version: 2.0.13
      - name: instana-agent-operator.v2.0.12
        version: 2.0.12
      - name: instana-agent-operator.v2.0.11
        version: 2.0.11
      - name: instana-agent-operator.v2.0.10
        version: 2.0.10
      - name: instana-agent-operator.v2.0.9
        version: 2.0.9
      - name: instana-agent-operator.v2.0.8
        version: 2.0.8
      - name: instana-agent-operator.v2.0.7
        version: 2.0.7
      - name: instana-agent-operator.v2.0.5
        version: 2.0.5
      - name: instana-agent-operator.v2.0.4
        version: 2.0.4
      name: stable
    defaultChannel: stable
    packageName: instana-agent-operator-rhmp
    provider:
      name: Instana
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: pelorus-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pelorus-operator.v0.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.pelorus.dora-metrics.io/v1alpha1",
                "kind": "Pelorus",
                "metadata": {
                  "name": "pelorus-sample"
                },
                "spec": {
                  "exporters": {
                    "global": {},
                    "instances": [
                      {
                        "app_name": "deploytime-exporter",
                        "exporter_type": "deploytime"
                      },
                      {
                        "app_name": "failuretime-exporter",
                        "enabled": false,
                        "exporter_type": "failure"
                      },
                      {
                        "app_name": "committime-exporter",
                        "exporter_type": "committime"
                      },
                      {
                        "app_name": "webhook-exporter",
                        "exporter_type": "webhook"
                      }
                    ]
                  },
                  "federate_openshift_monitoring": {
                    "enabled": false
                  },
                  "openshift_prometheus_basic_auth_pass": "changeme",
                  "openshift_prometheus_htpasswd_auth": "internal:{SHA}+pvrmeQCmtWmYVOZ57uuITVghrM=",
                  "prometheus_retention": "1y",
                  "prometheus_retention_size": "1GB",
                  "prometheus_storage": false,
                  "prometheus_storage_pvc_capacity": "2Gi",
                  "prometheus_storage_pvc_storageclass": "gp2"
                }
              }
            ]
          capabilities: Basic Install
          categories: |
            Modernization & Migration,Developer Tools,Monitoring,Integration & Delivery
          containerImage: quay.io/pelorus/pelorus-operator:0.0.9
          createdAt: "2024-05-27T20:04:01Z"
          description: |
            Tool that helps IT organizations measure their impact on the overall performance of their organization
          operatorframework.io/suggested-namespace: pelorus
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/dora-metrics/pelorus/
          support: Pelorus Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pelorus is the Schema for pelorus API to adopt instance to
              the requested workflow
            displayName: Pelorus
            kind: Pelorus
            name: peloruses.charts.pelorus.dora-metrics.io
            version: v1alpha1
        description: |
          Pelorus is a tool that helps IT organizations measure their impact on the overall performance of their organization. It does this by gathering metrics about team and organizational behaviors over time in some key areas of IT that have been shown to impact the value they deliver to the organization as a whole. Some of the key outcomes Pelorus can focus on are:

          - Software Delivery Performance
          - Product Quality and Sustainability
          - Customer experience

          For more background on the project you can read [@trevorquinn](https://github.com/trevorquinn)'s blog post on [Metrics Driven Transformation](https://www.openshift.com/blog/exploring-a-metrics-driven-approach-to-transformation).

          ## Software Delivery Performance as an outcome

          Currently, Pelorus functionality can capture proven metrics that measure Software Delivery Performance -- a significant outcome that IT organizations aim to deliver.

          Pelorus is a Grafana dashboard that can easily be deployed to an OpenShift cluster, and provides an organizational-level view of the [four critical measures of software delivery performance](https://blog.openshift.com/exploring-a-metrics-driven-approach-to-transformation/).

          ## Software Delivery Metrics Dashboard

          A short video describing each of metrics provided by Pelorus is available [here](https://www.youtube.com/watch?v=7-iB_KhUaQg).

          ## Demo

          [YouTube Video](https://www.youtube.com/watch?v=VPCOIfDcgso) with the Pelorus in action recorded during online Konveyor Community event.

          ## Documentation

          Pelorus documentation is available at [pelorus.readthedocs.io](https://pelorus.readthedocs.io/en/latest/).

          ## Contributing to Pelorus

          If you are interested in contributing to the Pelorus project, please review our Contribution guide which can be found in the [contribution guide](https://github.com/dora-metrics/pelorus/blob/master/CONTRIBUTING.md)

          ## Statement of Support

          Our support policy can be found in the [Upstream Support statement](https://github.com/dora-metrics/pelorus/blob/master/docs/UpstreamSupport.md)

          ## Code of Conduct
          Refer to dora-metrics's Code of Conduct [here](https://github.com/dora-metrics/pelorus/blob/master/CODE_OF_CONDUCT.md).

          ## License

          This repository is licensed under the terms of [Apache-2.0 License](https://raw.githubusercontent.com/dora-metrics/pelorus/master/LICENSE).
        displayName: Pelorus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pelorus-operator
        - dora
        - dora metrics
        - pelorus
        - metrics
        - transformation
        - devops
        links:
        - name: Pelorus Docs
          url: https://pelorus.readthedocs.io/en/latest/
        - name: Pelorus Getting Started
          url: https://pelorus.readthedocs.io/en/latest/GettingStarted/Overview/
        - name: Pelorus GIT repository
          url: https://github.com/dora-metrics/pelorus/
        maintainers:
        - email: kgranger@redhat.com
          name: Kevin Granger
        - email: whayutin@redhat.com
          name: Wesley Hayutin
        - email: mpryc@redhat.com
          name: Michal Pryc
        - email: msouzaol@redhat.com
          name: Mateus Souza Oliveira
        - email: esauer@redhat.com
          name: Eric Sauer
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - quay.io/pelorus/pelorus-operator:0.0.9
        version: 0.0.9
      entries:
      - name: pelorus-operator.v0.0.9
        version: 0.0.9
      - name: pelorus-operator.v0.0.8
        version: 0.0.8
      - name: pelorus-operator.v0.0.7
        version: 0.0.7
      - name: pelorus-operator.v0.0.6
        version: 0.0.6
      - name: pelorus-operator.v0.0.5
        version: 0.0.5
      - name: pelorus-operator.v0.0.4
        version: 0.0.4
      - name: pelorus-operator.v0.0.3
        version: 0.0.3
      - name: pelorus-operator.v0.0.2
        version: 0.0.2
      - name: pelorus-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: pelorus-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: horizon-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: horizon-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "horizon.openstack.org/v1beta1",
                "kind": "Horizon",
                "metadata": {
                  "name": "horizon"
                },
                "spec": {
                  "customServiceConfig": "DEBUG = True\nLOGGING['handlers']['console']['level'] = 'DEBUG'\n",
                  "replicas": 1,
                  "secret": "osp-secret"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:00Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Horizon is the Schema for the horizons API
            displayName: Horizon
            kind: Horizon
            name: horizons.horizon.openstack.org
            version: v1beta1
        description: Horizon Operator
        displayName: Horizon Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - UI
        - Horizon
        links:
        - name: Horizon Operator
          url: https://github.com/openstack-k8s-operators/horizon-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-horizon-rhel9@sha256:b02c115d40aacc30e6763e295da90d69681b24fba20ebbc53133be852393b399
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/horizon-rhel9-operator@sha256:b76038b96a064c275721c1d45b95d7b0dfdb9fc626a2b3250d0adb1f97f89319
        version: 1.0.6
      entries:
      - name: horizon-operator.v1.0.6
        version: 1.0.6
      - name: horizon-operator.v1.0.5
        version: 1.0.5
      - name: horizon-operator.v1.0.4
        version: 1.0.4
      - name: horizon-operator.v1.0.3
        version: 1.0.3
      - name: horizon-operator.v1.0.2
        version: 1.0.2
      - name: horizon-operator.v1.0.1
        version: 1.0.1
      - name: horizon-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: horizon-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq-streams
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amqstreams.v2.2.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.3",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.3",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:f908d848b5e67de4b162597b190ef18d449ffd4b6540457b24b491445d5c5ea3
          createdAt: "2022-07-01 10:33:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.1.0-0 <2.2.2-2'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ", "Red Hat Application Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from AMQ Streams 1.7, migration to `v1beta2` needs to be completed for all AMQ Streams CRDs and CRs before the upgrade to 2.2 is done! If upgrading from AMQ Streams version earlier than 1.7, you need to first install the CRDs from AMQ Streams 1.7 and complete the migration to `v1beta2` for all AMQ Streams CRDs and CRs before the upgrade to 2.2 is done! For more details about the CRD upgrades, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.2/html/deploying_and_upgrading_amq_streams_on_openshift/assembly-upgrade-str#assembly-upgrade-resources-str).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.2/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: Red Hat Integration - AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        - registry.redhat.io/amq7/amq-streams-kafka-31-rhel8@sha256:d0607285fae95365b9d34b9474466824fb5ad012871204e8b13e52e438b66bf5
        - registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:f908d848b5e67de4b162597b190ef18d449ffd4b6540457b24b491445d5c5ea3
        - registry.redhat.io/amq7/amq-streams-kafka-32-rhel8@sha256:e3bbfff505f12c0360e175f4640c2dd9a7adcf3dce1ae2f586f2bde805ed8b1f
        - registry.redhat.io/amq7/amq-streams-bridge-rhel8@sha256:5d49bd23da40590e74422424b9457398bad89da8276718ff6a8a088e901c8345
        version: 2.2.2-2
      entries:
      - name: amqstreams.v2.2.2-2
        version: 2.2.2-2
      - name: amqstreams.v2.2.2-1
        version: 2.2.2-1
      - name: amqstreams.v2.2.2-0
        version: 2.2.2-0
      - name: amqstreams.v2.2.1-7
        version: 2.2.1-7
      - name: amqstreams.v2.2.1-6
        version: 2.2.1-6
      - name: amqstreams.v2.2.1-5
        version: 2.2.1-5
      - name: amqstreams.v2.2.1-4
        version: 2.2.1-4
      - name: amqstreams.v2.2.1-3
        version: 2.2.1-3
      - name: amqstreams.v2.2.1-2
        version: 2.2.1-2
      - name: amqstreams.v2.2.1-1
        version: 2.2.1-1
      - name: amqstreams.v2.2.1-0
        version: 2.2.1-0
      - name: amqstreams.v2.2.0-5
        version: 2.2.0-5
      name: amq-streams-2.2.x
    - currentCSV: amqstreams.v2.3.0-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.3.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.3.1",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:2e46a25144d697761f0119deaf528c6f1c2ccd04e2acd95c661b0df4f32e1a9f
          createdAt: "2022-10-31 17:05:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.2.0-0 <2.3.0-4'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.3.0 release, AMQ Streams supports only OCP 4.8 to 4.12!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 2.2 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more details, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.3/html/deploying_and_upgrading_amq_streams_on_openshift/assembly-upgrade-str#assembly-upgrade-resources-str

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.3/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-streams-kafka-32-rhel8@sha256:0152a85f2d0bc85881dacaee62c6e8b0b8b04f57a542b9b82dc93ad5303d6bf9
        - registry.redhat.io/amq7/amq-streams-kafka-33-rhel8@sha256:046c2539fef32bd928fc1f4f4c41e63f9a143b371fcc201d78c66f48fc34ea32
        - registry.redhat.io/amq7/amq-streams-bridge-rhel8@sha256:c822b3d87f7dea233fc523238d8d0516e42f9a3d252f50f4d427c934995f9b22
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        - registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:2e46a25144d697761f0119deaf528c6f1c2ccd04e2acd95c661b0df4f32e1a9f
        version: 2.3.0-4
      entries:
      - name: amqstreams.v2.3.0-4
        version: 2.3.0-4
      name: amq-streams-2.3.x
    - currentCSV: amqstreams.v2.4.0-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:cd1fefcd199400a504b4b9942c24ba6becafcecc952a9e39fa2729228be79e77
          createdAt: "2022-10-31 17:05:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.3.0-0 <2.4.0-3'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.4.0 release, AMQ Streams supports only OCP 4.8 to 4.12!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more details, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.4/html/deploying_and_managing_amq_streams_on_openshift/assembly-upgrade-str)

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.4/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:8d467fc746b78440df06f6d88d266f6ac9b5a4574a2b17fd870c173a107f8d65
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:cd1fefcd199400a504b4b9942c24ba6becafcecc952a9e39fa2729228be79e77
        - registry.redhat.io/amq-streams/kafka-33-rhel8@sha256:aa780a250444c2426be79e95366050a33f4a4c9d68c08b58743413ea67a2d6d7
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:50b3be789993cfeedc4923a8e760c1cf52498b2178be3abfa3719fc2d15ce723
        version: 2.4.0-3
      entries:
      - name: amqstreams.v2.4.0-3
        version: 2.4.0-3
      - name: amqstreams.v2.4.0-2
        version: 2.4.0-2
      - name: amqstreams.v2.4.0-1
        version: 2.4.0-1
      name: amq-streams-2.4.x
    - currentCSV: amqstreams.v2.5.2-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
          createdAt: "2023-07-10 21:47:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.2.0-0 <2.5.2-8'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.5.2 release, AMQ Streams supports only OCP 4.12 to 4.16!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more information, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:8c7e8ccb6c2d5da32b70d9d0c3919c303ad73a25447f6993d964f7b19cca1072
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:4ede5eeed159bddacfe830a768bab4996003d4c57a496fc8ec9b41df916d4acc
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:95bc35951eb545b918114af1e92a7139e2e3323b7d054dcb2a060843c7326d28
        version: 2.5.2-8
      entries:
      - name: amqstreams.v2.5.2-8
        version: 2.5.2-8
      - name: amqstreams.v2.5.2-7
        version: 2.5.2-7
      - name: amqstreams.v2.5.2-6
        version: 2.5.2-6
      - name: amqstreams.v2.5.2-5
        version: 2.5.2-5
      - name: amqstreams.v2.5.2-4
        version: 2.5.2-4
      - name: amqstreams.v2.5.2-3
        version: 2.5.2-3
      - name: amqstreams.v2.5.2-2
        version: 2.5.2-2
      - name: amqstreams.v2.5.2-1
        version: 2.5.2-1
      - name: amqstreams.v2.5.2-0
        version: 2.5.2-0
      - name: amqstreams.v2.5.1-5
        version: 2.5.1-5
      - name: amqstreams.v2.5.1-4
        version: 2.5.1-4
      - name: amqstreams.v2.5.1-3
        version: 2.5.1-3
      - name: amqstreams.v2.5.1-2
        version: 2.5.1-2
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.5.x
    - currentCSV: amqstreams.v2.6.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.6.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:dacf7ddffe7cbe2dc35e7d3290637b3334117e09dde986e6eaeaceded7069fbf
          createdAt: "2023-10-14 18:18:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.6.0-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: "**Red Hat AMQ Streams** is a massively scalable, distributed,
          and high performance data streaming platform based on the Apache Kafka®
          project. AMQ Streams provides an event streaming backbone that allows microservices
          and other application components to exchange data with extremely high throughput
          and low latency.\n**From the 2.6.0 release, AMQ Streams supports only OCP
          4.11 to 4.14! \n**The core capabilities include:** * A pub/sub messaging
          model, similar to a traditional enterprise messaging system, in which application
          components publish and consume events to/from an ordered stream\n* The long
          term, fault-tolerant storage of events\n* The ability for a consumer to
          replay streams of events\n* The ability to partition topics for horizontal
          scalability\n### Upgrades\n**!!! IMPORTANT !!!** **Direct upgrade from AMQ
          Stream 1.7 or earlier is not supported!** You have to upgrade first to one
          of the previous versions of AMQ Streams. You will also need to convert the
          CRD resources. For more information, see the [documentation for AMQ Streams
          on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).\n###
          Supported Features\n* **Manages the Kafka Cluster** - Deploys and manages
          all of the components of this complex application, including dependencies
          like Apache ZooKeeper® that are traditionally hard to administer.\n* **Includes
          Kafka Connect** - Allows for configuration of common data sources and sinks
          to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for mirroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashboards\n### Upgrading your Clusters\nThe
          AMQ Streams Operator understands how to run and upgrade between a set of
          Kafka versions. When specifying a new version in your config, check to make
          sure you aren't using any features that may have been removed. For more
          information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).\n###
          Storage\nAn efficient data storage infrastructure is essential to the optimal
          performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires
          block storage. The use of file storage (for example, NFS) is not recommended.\nThe
          AMQ Streams Operator supports three types of data storage:\n* Ephemeral
          (Recommended for development only!)\n* Persistent\n* JBOD (Just a Bunch
          of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nAMQ Streams
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation for the current release
          can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).\n"
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:dacf7ddffe7cbe2dc35e7d3290637b3334117e09dde986e6eaeaceded7069fbf
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:be7466ffd7f624f769eedca2e4106c597d2bf1d95cc9679f235e6dcdec66cfa9
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:fa8361e5089ab4449261f5a0ad57bd0bd059dc254b94bd9ab68340faae386947
        - registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:8b10bfb697b48ba3ed246fb846384f4cc67f05f670e9521edf0f47f829869404
        version: 2.6.0-2
      entries:
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.6.x
    - currentCSV: amqstreams.v2.7.0-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:671c1a2fcb0856cb208c0ead539fff9571fa2281a1e3d45b3df9ebd3df8c787e
          createdAt: "2024-03-09 21:27:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0-0 <2.7.0-7'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.6.0 release, Streams supports only OCP 4.12 to 4.14!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:671c1a2fcb0856cb208c0ead539fff9571fa2281a1e3d45b3df9ebd3df8c787e
        - registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:66362c800b78003d26df9471b7f21f5c9ad7fca8e8fd2511428dfc549cdf46c2
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:be9b82af13cbc120bb415abf4fcc97420e543172d6450b836c47344e4dcacd00
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:a7620559f3e5a0a4d5ecee44538619f698158847d9d0a06c7168404ed6f7cd36
        - registry.redhat.io/amq-streams/kafka-36-rhel9@sha256:8890c3217343d9aa3dd2ab9c7fa1742f960b5a26e09906581ce6075e61dc2888
        version: 2.7.0-7
      entries:
      - name: amqstreams.v2.7.0-7
        version: 2.7.0-7
      - name: amqstreams.v2.7.0-6
        version: 2.7.0-6
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.7.x
    - currentCSV: amqstreams.v2.8.0-0-0.1738265624.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:675d4acc375fb20e1a3fd5755184a318341c15f9337a11ea41e902fcda6d0bec
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.0-0 <2.8.0-0'
          olm.substitutesFor: amqstreams.v2.8.0-0
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.8.0 release, Streams supports only OCP 4.12 to 4.14!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:675d4acc375fb20e1a3fd5755184a318341c15f9337a11ea41e902fcda6d0bec
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:b46457a0a7f63518bc12b2789f75afb0e44496e62de918b9b584129c7078de86
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:4df906cf1490922a4587d96250f3be6ff605fe2caf87dfba7b812c144c5c5774
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:e771a84a458388c07ecf970ec625e4203c18e5d2bd91e7571d9c44ebb09313de
        - registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:17b2d255439f38c6ef54c0a56929ae636743b12c688d42223ac3fb58729e51ad
        version: 2.8.0-0+0.1738265624.p
      entries:
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.8.x
    - currentCSV: amqstreams.v2.9.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.9.2-1'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.9.0 release, Streams supports only OCP 4.14 to 4.18!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:f5985592b7f05c9b3c64caa0ac90f34f26b9b4277b342b6bd4a6415d2c563f9a
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:86ea2e823cd83c476521f2a1e9da5619c4810661c60d28295d0f5482c8aa519d
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:ddfb50ec2f1cf45ccccc632182bf0ff549c14969da35b7248802eab67d9c2e9d
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:67647bff1a76adb5b51f43c99d5f88a8aa38055c1e87190add8972bb621ddc15
        version: 2.9.2-1
      entries:
      - name: amqstreams.v2.9.2-1
        version: 2.9.2-1
      - name: amqstreams.v2.9.2-0
        version: 2.9.2-0
      - name: amqstreams.v2.9.1-2
        version: 2.9.1-2
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.9.x
    - currentCSV: amqstreams.v2.9.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.9.2-1'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.9.0 release, Streams supports only OCP 4.14 to 4.18!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:86ea2e823cd83c476521f2a1e9da5619c4810661c60d28295d0f5482c8aa519d
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:ddfb50ec2f1cf45ccccc632182bf0ff549c14969da35b7248802eab67d9c2e9d
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:67647bff1a76adb5b51f43c99d5f88a8aa38055c1e87190add8972bb621ddc15
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:f5985592b7f05c9b3c64caa0ac90f34f26b9b4277b342b6bd4a6415d2c563f9a
        version: 2.9.2-1
      entries:
      - name: amqstreams.v2.9.2-1
        version: 2.9.2-1
      - name: amqstreams.v2.9.2-0
        version: 2.9.2-0
      - name: amqstreams.v2.9.1-2
        version: 2.9.1-2
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.x
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-3.0.x
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-3.x
    - currentCSV: amqstreams.v2.5.2-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
          createdAt: "2023-07-10 21:47:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.2.0-0 <2.5.2-8'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.5.2 release, AMQ Streams supports only OCP 4.12 to 4.16!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more information, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:4ede5eeed159bddacfe830a768bab4996003d4c57a496fc8ec9b41df916d4acc
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:95bc35951eb545b918114af1e92a7139e2e3323b7d054dcb2a060843c7326d28
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:8c7e8ccb6c2d5da32b70d9d0c3919c303ad73a25447f6993d964f7b19cca1072
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
        version: 2.5.2-8
      entries:
      - name: amqstreams.v2.5.2-8
        version: 2.5.2-8
      - name: amqstreams.v2.5.2-7
        version: 2.5.2-7
      - name: amqstreams.v2.5.2-6
        version: 2.5.2-6
      - name: amqstreams.v2.5.2-5
        version: 2.5.2-5
      - name: amqstreams.v2.5.2-4
        version: 2.5.2-4
      - name: amqstreams.v2.5.2-3
        version: 2.5.2-3
      - name: amqstreams.v2.5.2-2
        version: 2.5.2-2
      - name: amqstreams.v2.5.2-1
        version: 2.5.2-1
      - name: amqstreams.v2.5.2-0
        version: 2.5.2-0
      - name: amqstreams.v2.5.1-5
        version: 2.5.1-5
      - name: amqstreams.v2.5.1-4
        version: 2.5.1-4
      - name: amqstreams.v2.5.1-3
        version: 2.5.1-3
      - name: amqstreams.v2.5.1-2
        version: 2.5.1-2
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-lts
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: stable
    defaultChannel: stable
    packageName: amq-streams
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: manila-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: manila-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "manila.openstack.org/v1beta1",
                "kind": "Manila",
                "metadata": {
                  "name": "manila",
                  "namespace": "openstack"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "manila",
                  "databaseInstance": "openstack",
                  "manilaAPI": {},
                  "manilaScheduler": {},
                  "manilaShares": {
                    "share1": {}
                  },
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "osp-secret",
                  "serviceUser": "manila"
                }
              },
              {
                "apiVersion": "manila.openstack.org/v1beta1",
                "kind": "ManilaAPI",
                "metadata": {
                  "name": "manilaapi-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "manila.openstack.org/v1beta1",
                "kind": "ManilaScheduler",
                "metadata": {
                  "name": "manilascheduler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "manila.openstack.org/v1beta1",
                "kind": "ManilaShare",
                "metadata": {
                  "name": "manilashare-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:57Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManilaAPI is the Schema for the manilaapis API
            displayName: Manila API
            kind: ManilaAPI
            name: manilaapis.manila.openstack.org
            version: v1beta1
          - description: Manila is the Schema for the manilas API
            displayName: Manila
            kind: Manila
            name: manilas.manila.openstack.org
            version: v1beta1
          - description: ManilaScheduler is the Schema for the manilaschedulers API
            displayName: Manila Scheduler
            kind: ManilaScheduler
            name: manilaschedulers.manila.openstack.org
            version: v1beta1
          - description: ManilaShare is the Schema for the manilashares API
            displayName: Manila Share
            kind: ManilaShare
            name: manilashares.manila.openstack.org
            version: v1beta1
        description: Manila Operator
        displayName: Manila Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Share
        - Manila
        links:
        - name: Manila Operator
          url: https://github.com/openstack-k8s-operators/manila-operator
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/manila-rhel9-operator@sha256:1208676b94d8a726c277088591d898e08a3a8331a0125ba7bff101152afba934
        - registry.redhat.io/rhoso/openstack-manila-scheduler-rhel9@sha256:7a8f645126c073213ce066cdcac2164df51276f8aa4a3f89954664d3010e6fae
        - registry.redhat.io/rhoso/openstack-manila-share-rhel9@sha256:83a6d292785805a0ac2895810ea466cfb6e1cb7aa1a53ae8f4e22671381490df
        - registry.redhat.io/rhoso/openstack-manila-api-rhel9@sha256:4cbf725169614c2974d6d1a683cfc0179bfe06ca29cf30f7dca88caeb368c51a
        version: 1.0.6
      entries:
      - name: manila-operator.v1.0.6
        version: 1.0.6
      - name: manila-operator.v1.0.5
        version: 1.0.5
      - name: manila-operator.v1.0.4
        version: 1.0.4
      - name: manila-operator.v1.0.3
        version: 1.0.3
      - name: manila-operator.v1.0.2
        version: 1.0.2
      - name: manila-operator.v1.0.1
        version: 1.0.1
      - name: manila-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: manila-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JFrog
      provider-url: ""
    name: openshiftxray-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: xray-operator.v2.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftXray",
                "metadata": {
                  "name": "openshiftxray"
                },
                "spec": {
                  "xray": {
                    "common": {
                      "xrayUserId": "1000721035",
                      "xrayGroupId": "1000721035"
                    },
                    "analysis": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-analysis"
                      },
                      "name": "xray-analysis",
                      "podManagementPolicy": "Parallel",
                      "preStartCommand": null,
                      "updateStrategy": "RollingUpdate"
                    },
                    "database": {
                      "password": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "global": {
                      "postgresqlTlsSecret": null
                    },
                    "indexer": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-indexer"
                      },
                      "name": "xray-indexer",
                      "podManagementPolicy": "Parallel",
                      "updateStrategy": "RollingUpdate"
                    },
                    "persist": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-persist"
                      },
                      "name": "xray-persist",
                      "persistence": {
                        "size": "10Gi"
                      },
                      "podManagementPolicy": "Parallel",
                      "preStartCommand": null,
                      "updateStrategy": "RollingUpdate"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "rabbitmq": {
                      "enabled": true,
                      "replicaCount": 1,
                      "rbac": {
                        "create": true
                      },
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-rabbitmq"
                      },
                      "podSecurityContext": {
                        "fsGroup": 1000721001,
                        "runAsUser": 1000721001
                      },
                      "auth": {
                        "username": "xray",
                        "password": "xray",
                        "erlangCookie": "XRAYRABBITMQCLUSTER",
                        "tls": {
                          "enabled": false,
                          "failIfNoPeerCert": true,
                          "sslOptionsVerify": "verify_peer",
                          "caCertificate": "",
                          "serverCertificate": "",
                          "serverKey": ""
                        }
                      }
                    },
                    "rabbitmq-ha": {
                      "enabled": false,
                      "image": {
                        "repository": "registry.connect.redhat.com/jfrog/xray-rabbitmq"
                      },
                      "rabbitmqUsername": "guest",
                      "rabbitmqPassword": "guest",
                      "managementUsername": "guest",
                      "managementPassword": "guest",
                      "initContainer": {
                        "enabled": false
                      },
                      "securityContext": {
                        "fsGroup": 1000721035,
                        "runAsUser": 1000721035,
                        "runAsGroup": 1000721035
                      },
                      "replicaCount": 1
                    },
                    "replicaCount": 1,
                    "unifiedUpgradeAllowed":  "true",
                    "router": {
                      "image": {
                        "imagePullPolicy": "IfNotPresent",
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-router"
                      },
                      "name": "router"
                    },
                    "server": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-server"
                      },
                      "name": "xray-server",
                      "podManagementPolicy": "Parallel",
                      "replicaCount": 1,
                      "updateStrategy": "RollingUpdate"
                    },
                    "xray": {
                      "consoleLog": false,
                      "jfrogUrl": "OVERRIDE",
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/xray-operator@sha256:da7f6977170876b335875f4074ddfc0266b665351249295ce3c53f0e95891cc2
          createdAt: 2020-05-22 00:00:00+00:00
          description: JFrog Xray Enterprise deploys Xray continuous security scanner
            into Openshift (Requires Jfrog Artifactory)
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Xray Instances
            displayName: Xray
            kind: OpenshiftXray
            name: openshiftxrays.charts.helm.k8s.io
            version: v1
        description: "## Overview Openshift Operator to deploy JFrog Xray Continuous
          Security scanner into your Openshift cluster. NOTE: Artifactory is required
          for JFrog Xray to connect with and scan artifacts against.\n### Major version
          v2.0.4 Breaking change notification!!\nThis release has major bug fix for
          Rabbitmq Pod. Please update to version 2.0.4 of the operator.\n### Major
          version v2.0.0 Breaking change notification!! Xray is dependent upon Rabbitmq.
          The version of Rabbitmq that this chart uses has significantly changed from
          version 1.1.8 of the operator. For full details on how to migrate please
          visit: `https://github.com/jfrog/charts/blob/master/stable/xray/RABBITMQ_MIGRATION_NOTES.md`\nIf
          you are using operator version < 2.1.0 there is a issue with upgrade where
          xray version is not updated if you update operator version.As workaround
          you can update the Custom Resource OpenshiftXray which has hardcored image
          tag value.After upgrading to version 2.1.0 you can remove the image tags
          parameters from your CR so that future upgrade works without workaround.\n###
          Security Context Constraints To deploy this helm chart you will need to
          be a cluster admin w/ access to the anyuid scc and add the operator service
          account to the anyuid scc. ``` oc adm policy add-scc-to-user anyuid -z xray-operator
          ``` ``` oc adm policy add-scc-to-user anyuid -z openshiftxray ``` ``` oc
          adm policy add-scc-to-user anyuid -z openshiftxray-rabbitmq-ha ``` ## Usage
          An external DB is required. The operator will not deploy a DB but will require
          you to specify the configuration values to connect to it. Search for JFrog
          and click JFrog Xray Operator to install. Go to the Installed Operators.
          Wait for the JFrog Xray Operator to complete the installation. Open the
          Operator and click on the provided API: Xray Click Create New Instance and
          provide the following parameters for your DB configuration: ``` JFROG_URL
          DATABASE_URL DATABASE_USER DATABASE_PASSWORD ``` JFROG_URL is the external
          ip or DNS of your Artifactory to connect Xray to. Artifactory is required
          to use this operator. You can set your JFROG_URL to the service name of
          your Artifactory Nginx: ``` oc get svc -n my_namespace | grep nginx ```
          DATABASE_URL must be a Postgresql URL in the format: `` postgres://postgres-postgresql:5432/xraydb?sslmode=disable
          `` DATABASE_USER and DATABASE_PASSWORD must supply a valid user on Postgresql.
          Click Create for Xray to deploy into OpenShift. Open Artifactory in a web
          browser to complete the onboarding wizard for Xray! ### Air gap environments
          To use Xray Operator in an air gap environment you will need to download
          the images as image streams into your Openshift air gap cluster manually.
          Use the image overrides to then specify the image stream locations that
          are local to your cluster. Next you will need to setup the database of index
          data for xray to use when scanning artifacts. Follow the link below for
          instructions on setup: https://www.jfrog.com/confluence/display/JFROG/Configuring+Xray#ConfiguringXray-SynchronizingtheDatabase "
        displayName: JFrog Xray Continuous Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Xray Continuous Security
          url: https://jfrog.com/xray/
        - name: JFrog Xray Continuous Security Video
          url: https://youtu.be/wGzyKlQklKQ
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: alpha
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/xray-operator@sha256:da7f6977170876b335875f4074ddfc0266b665351249295ce3c53f0e95891cc2
        - registry.connect.redhat.com/jfrog/xray-persist@sha256:93fae0c124919a57d5162e32e88d3456ef6254889e454f89c21c89e69e6ae323
        - registry.connect.redhat.com/jfrog/xray-indexer@sha256:7f40d2dfe4a35b25d16dfcbfa04b4a58f227c132d4c3c86345e7b2bd19a97311
        - registry.connect.redhat.com/jfrog/xray-router@sha256:019235289e17d0e22dcd23a1505f71a78710827dfab338d8fa980f64c14f162f
        - registry.connect.redhat.com/jfrog/xray-rabbitmq@sha256:459d7d5e78192a6f45a5527e8fbaf1d8b30299b10ad4c42515d3d2bb5fcc75f4
        - registry.connect.redhat.com/jfrog/init@sha256:3f27e44b44c3556c50318110dbd542e34ef1061e28d7f0e0e7b8baa73c5f3e87
        - registry.connect.redhat.com/jfrog/xray-server@sha256:c5584c4e9528f0199e840b2b1ad1144220850ca9595ebf3e270bc6f86de4a71a
        - registry.connect.redhat.com/jfrog/xray-analysis@sha256:bcd8211f1bc753bfd7735165ce39b476e4efff198bffa877883620b41fcb33b5
        version: 2.1.6
      entries:
      - name: xray-operator.v2.1.6
        version: 2.1.6
      name: alpha
    defaultChannel: alpha
    packageName: openshiftxray-operator
    provider:
      name: JFrog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Aerospike
      provider-url: ""
    name: aerospike-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aerospike-kubernetes-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "memory-size": 3000000000,
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:6.3.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: docker.io/aerospike/aerospike-kubernetes-operator:2.5.0
          createdAt: 2023-05-04T09:32:47UTC
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Aerospike
        relatedImages:
        - docker.io/aerospike/aerospike-kubernetes-operator:2.5.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 2.5.0
      entries:
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      - name: aerospike-kubernetes-operator.v2.4.0
        version: 2.4.0
      - name: aerospike-kubernetes-operator.v2.3.0
        version: 2.3.0
      - name: aerospike-kubernetes-operator.v2.2.1
        version: 2.2.1
      - name: aerospike-kubernetes-operator.v2.2.0
        version: 2.2.0
      - name: aerospike-kubernetes-operator.v2.1.0
        version: 2.1.0
      name: alpha
    - currentCSV: aerospike-kubernetes-operator.v4.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "data-size": 1073741824,
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:8.1.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackup",
                "metadata": {
                  "name": "aerospikebackup",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "aerospike-cluster": {
                      "aerospike-aerospikebackup-test-cluster": {
                        "credentials": {
                          "password": "admin123",
                          "user": "admin"
                        },
                        "seed-nodes": [
                          {
                            "host-name": "aerocluster.aerospike.svc.cluster.local",
                            "port": 3000
                          }
                        ]
                      }
                    },
                    "backup-routines": {
                      "aerospike-aerospikebackup-test-routine": {
                        "backup-policy": "test-policy",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "local"
                      },
                      "aerospike-aerospikebackup-test-routine1": {
                        "backup-policy": "test-policy1",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "s3Storage"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackupService",
                "metadata": {
                  "name": "aerospikebackupservice-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "config": {
                    "backup-policies": {
                      "test-policy": {
                        "parallel": 3
                      },
                      "test-policy1": {
                        "parallel": 3
                      }
                    },
                    "service": {
                      "http": {
                        "port": 8081
                      }
                    },
                    "storage": {
                      "local": {
                        "local-storage": {
                          "path": "/tmp/localStorage"
                        }
                      },
                      "s3Storage": {
                        "s3-storage": {
                          "bucket": "aerospike-kubernetes-operator-test",
                          "s3-profile": "default",
                          "s3-region": "us-east-1"
                        }
                      }
                    }
                  },
                  "image": "aerospike/aerospike-backup-service:3.2.0",
                  "secrets": [
                    {
                      "secretName": "aws-secret",
                      "volumeMount": {
                        "mountPath": "/root/.aws/credentials",
                        "name": "aws-secret",
                        "subPath": "credentials"
                      }
                    }
                  ],
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeRestore",
                "metadata": {
                  "name": "aerospikerestore-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "backup-data-path": "aerospike-aerospikebackup-test-routine/backup/1733506015827/data/test",
                    "destination": {
                      "credentials": {
                        "password": "admin123",
                        "user": "admin"
                      },
                      "label": "destinationCluster",
                      "seed-nodes": [
                        {
                          "host-name": "aerocluster.test.svc.cluster.local",
                          "port": 3000
                        }
                      ]
                    },
                    "policy": {
                      "no-generation": true,
                      "no-indexes": true,
                      "parallel": 3
                    },
                    "source": {
                      "local-storage": {
                        "path": "/tmp/localStorage"
                      }
                    }
                  },
                  "type": "Full"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: docker.io/aerospike/aerospike-kubernetes-operator:4.1.1
          createdAt: "2025-09-15T10:14:05Z"
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          olm.skipRange: '>=4.0.0 <4.1.1'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
          support: Aerospike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeBackup is the Schema for the aerospikebackup API
            displayName: Aerospike Backup
            kind: AerospikeBackup
            name: aerospikebackups.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeBackupService is the Schema for the aerospikebackupservices
              API
            displayName: Aerospike Backup Service
            kind: AerospikeBackupService
            name: aerospikebackupservices.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1
          - description: AerospikeRestore is the Schema for the aerospikerestores
              API
            displayName: Aerospike Restore
            kind: AerospikeRestore
            name: aerospikerestores.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
           -   Aerospike cluster monitoring
           -   Backup and restore Aerospike clusters
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Aerospike
        relatedImages:
        - docker.io/aerospike/aerospike-kubernetes-operator:4.1.1
        version: 4.1.1
      entries:
      - name: aerospike-kubernetes-operator.v4.1.1
        version: 4.1.1
      - name: aerospike-kubernetes-operator.v4.1.0
        version: 4.1.0
      - name: aerospike-kubernetes-operator.v4.0.2
        version: 4.0.2
      - name: aerospike-kubernetes-operator.v4.0.1
        version: 4.0.1
      - name: aerospike-kubernetes-operator.v4.0.0
        version: 4.0.0
      - name: aerospike-kubernetes-operator.v3.4.1
        version: 3.4.1
      - name: aerospike-kubernetes-operator.v3.4.0
        version: 3.4.0
      - name: aerospike-kubernetes-operator.v3.3.1
        version: 3.3.1
      - name: aerospike-kubernetes-operator.v3.3.0
        version: 3.3.0
      - name: aerospike-kubernetes-operator.v3.2.2
        version: 3.2.2
      - name: aerospike-kubernetes-operator.v3.2.1
        version: 3.2.1
      - name: aerospike-kubernetes-operator.v3.2.0
        version: 3.2.0
      - name: aerospike-kubernetes-operator.v3.1.0
        version: 3.1.0
      - name: aerospike-kubernetes-operator.v3.0.0
        version: 3.0.0
      name: stable
    defaultChannel: stable
    packageName: aerospike-kubernetes-operator
    provider:
      name: Aerospike
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: 3scale-community-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: 3scale-community-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.8.2
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Improving [documentation](https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.8.2
        version: 0.8.2
      entries:
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.11
    - currentCSV: 3scale-community-operator.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.9.0
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.9.0/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.9.0/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.9.0
        version: 0.9.0
      entries:
      - name: 3scale-community-operator.v0.9.0
        version: 0.9.0
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.12
    - currentCSV: 3scale-community-operator.v0.10.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "ProductCRName": "product1-sample",
                  "Production": true
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.10.1
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.10.1/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Improving [documentation](https://github.com/3scale/3scale-operator/blob/master/doc/operator-user-guide.md)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.10.1/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.10.1
        version: 0.10.1
      entries:
      - name: 3scale-community-operator.v0.10.1
        version: 0.10.1
      - name: 3scale-community-operator.v0.9.0
        version: 0.9.0
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.13
    defaultChannel: threescale-2.13
    packageName: 3scale-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: https://www.intel.com
    name: ipm-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ipm-controller.v25.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ipm.intel.com/v1",
                "kind": "CPUPowerProv",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ipm-controller",
                    "node-label": "ipm-provisioning"
                  },
                  "name": "cpupowerprov-sample"
                },
                "spec": {
                  "cstates": {
                    "cores": {
                      "5": {
                        "C1": true,
                        "C1E": true,
                        "C6": false
                      },
                      "9,12": {
                        "C1": true,
                        "C6": true
                      }
                    },
                    "idleGovernor": "ladder"
                  },
                  "pstates": {
                    "cores": {
                      "5": {
                        "energyPerformanceBias": "normal",
                        "energyPerformancePreference": "performance",
                        "freqGovernor": "performance",
                        "maxFreqMHz": "3500",
                        "minFreqMHz": "800"
                      },
                      "9,12": {
                        "freqGovernor": "powersave",
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    },
                    "global": {
                      "conservative": {
                        "downThreshold": 20,
                        "freqStep": 5,
                        "ignoreNiceLoad": 0,
                        "samplingDownFactor": 1,
                        "samplingRate": 8888,
                        "upThreshold": 88
                      },
                      "maxPerfPct": 95,
                      "minPerfPct": 45,
                      "ondemand": {
                        "ignoreNiceLoad": 0,
                        "ioIsBusy": 1,
                        "powersaveBias": 0,
                        "samplingDownFactor": 1,
                        "samplingRate": 8888,
                        "upThreshold": 88
                      },
                      "schedutil": {
                        "rateLimitUs": 5000
                      },
                      "turbo": true
                    }
                  },
                  "uncoreFreqCPU": {
                    "cpu0": [
                      {
                        "maxFreqMHz": "1600",
                        "minFreqMHz": "800"
                      }
                    ],
                    "cpu1": [
                      {
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    ]
                  },
                  "uncoreFreqIO": {
                    "cpu0": [
                      {
                        "maxFreqMHz": "1600",
                        "minFreqMHz": "800"
                      }
                    ],
                    "cpu1": [
                      {
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/ipm-controller@sha256:32135c888f86a2c725d093ada9a59d222229bee5640d37f6e19a9e9812b4b67a
          createdAt: "2025-08-07T13:13:25Z"
          description: IPM Controller manages resources via REST API interactions.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.operatorframework.io/required-sccs: '["privileged"]'
          olm.securityContextConstraints.required: privileged
          operators.openshift.io/infrastructure-features: RestrictedNetwork
          operators.openshift.io/valid-subscription: Intel OBL Limited Distribution
            (Commercial Use) License Agreement [v2022.12.20]
          operators.operatorframework.io.bundle.package.v1: ipm-controller
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Intel® Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CPUPowerProv is the Schema for the cpupowerprovs API
            displayName: CPUPower Prov
            kind: CPUPowerProv
            name: cpupowerprovs.ipm.intel.com
            version: v1
        description: Infrastructure Power Manager (IPM) governs platform power technologies
          based on input from application workloads, the platform itself and customer
          input
        displayName: Intel® Infrastructure Power Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - intel
        - infrastructure
        - power
        - manager
        - management
        links:
        - name: IPM
          url: https://builders.intel.com/intel-technologies/power-management
        maintainers:
        - email: infrastructure.power.manager.operator@intel.com
          name: Infrastructure Power Manager Operator
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Intel® Corporation
          url: https://www.intel.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.13.0@sha256:471549d9cc07fcec0f285a05745b3d39a48e9a503b01b4047156a86ec1a30c00
        - registry.connect.redhat.com/intel/ipm-controller@sha256:32135c888f86a2c725d093ada9a59d222229bee5640d37f6e19a9e9812b4b67a
        - registry.connect.redhat.com/intel/ipm-agent@sha256:54de45cdc2e34768f6cdbf1cdee219406684143dddc4d236877709dc2536fe9c
        - registry.connect.redhat.com/intel/ipm-watcher@sha256:51eb773d5a4d1f401d4cbf2b42b8df4d0359beb85d5e274b8392dd7c5165e6ff
        version: 25.7.0
      entries:
      - name: ipm-controller.v25.7.0
        version: 25.7.0
      name: stable
    defaultChannel: stable
    packageName: ipm-controller
    provider:
      name: Intel® Corporation
      url: https://www.intel.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: ""
    name: nim-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nim-operator-certified.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMCache",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "source": {
                    "ngc": {
                      "modelPuller": "nvcr.io/nim/meta/llama-3.2-1b-instruct@sha256:1ff21ad32124d2382f8c733879394e7849cc3df5172651a24fcb671ac4fe77c6",
                      "pullSecret": "ngc-secret",
                      "authSecret": "ngc-api-secret",
                      "model": {
                        "lora": false,
                        "precision": "fp16",
                        "engine": "tensorrt_llm",
                        "qosProfile": "throughput",
                        "gpus": [
                          {
                            "product": "a100",
                            "ids": []
                          }
                        ]
                      }
                    }
                  },
                  "storage": {
                    "pvc": {
                      "create": true,
                      "storageClass": "local-path",
                      "size": "50Gi",
                      "volumeAccessMode": "ReadWriteOnce"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMService",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "image": {
                    "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                    "tag": "1.12.0",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [
                      "ngc-secret"
                    ]
                  },
                  "authSecret": "ngc-api-secret",
                  "storage": {
                    "nimCache": {
                      "name": "meta-llama3-1b-instruct",
                      "profile": ""
                    }
                  },
                  "scale": {
                    "enabled": false,
                    "hpa": {
                      "minReplicas": 1,
                      "maxReplicas": 10
                    }
                  },
                  "resources": {
                    "limits": {
                      "nvidia.com/gpu": 1
                    }
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMPipeline",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "k8s-nim-operator",
                    "app.kubernetes.io/managed-by": "k8s-nim-operator"
                  },
                  "name": "nimpipeline-sample"
                },
                "spec": {
                  "services": [
                    {
                      "name": "meta-llama3-1b-instruct",
                      "enabled": true,
                      "spec": {
                        "image": {
                          "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                          "tag": "1.12.0",
                          "pullPolicy": "IfNotPresent",
                          "pullSecrets": [
                            "ngc-secret"
                          ]
                        },
                        "authSecret": "ngc-api-secret",
                        "storage": {
                          "nimCache": {
                            "name": "meta-llama3-1b-instruct",
                            "profile": ""
                          }
                        },
                        "replicas": 1,
                        "resources": {
                          "limits": {
                            "nvidia.com/gpu": 1
                          }
                        },
                        "expose": {
                          "service": {
                            "type": "ClusterIP",
                            "port": 8000
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMBuild",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "nimCache": {
                    "name": "meta-llama3-1b-instruct",
                    "profile": "7cc8597690a35aba19a3636f35e7f1c7e7dbc005fe88ce9394cad4a4adeed414"
                  },
                  "image": {
                    "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                    "tag": "1.12.0",
                    "pullSecrets": [
                     "ngc-secret"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoCustomizer",
                "metadata": {
                  "name": "nemocustomizer-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "scheduler": {
                    "type": "volcano"
                  },
                  "wandb": {
                    "secretName": "wandb-secret",
                    "apiKeyKey": "apiKey",
                    "encryptionKey": "encryptionKey"
                  },
                  "otel": {
                    "enabled": true,
                    "exporterOtlpEndpoint": "http://customizer-otel-opentelemetry-collector.nemo.svc.cluster.local:4317"
                  },
                  "databaseConfig": {
                    "credentials": {
                      "user": "ncsuser",
                      "secretName": "customizer-pg-existing-secret",
                      "passwordKey": "password"
                    },
                    "host": "customizer-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "ncsdb"
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/customizer-api",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "entitystore": {
                    "endpoint": "http://nemoentitystore-sample.nemo.svc.cluster.local:8000"
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000"
                  },
                  "mlflow": {
                    "endpoint": "http://mlflow-tracking.nemo.svc.cluster.local:80"
                  },
                  "nemoDatastoreTools": {
                    "image": "nvcr.io/nvidia/nemo-microservices/nds-v2-huggingface-cli@sha256:7958e5244217de8af0ed723dfe7ea8e291b3afbb2a0aa73f5c27b218a1149b11"
                  },
                  "modelDownloadJobs": {
                    "image": "nvcr.io/nvidia/nemo-microservices/customizer-api@sha256:c590fa19121af2aac0379933deace3991f27c3ef88a8043560d6e3a172273b0b",
                    "ngcAPISecret": {
                      "name": "ngc-api-secret",
                      "key": "NGC_API_KEY"
                    },
                    "securityContext": {
                      "fsGroup": 1000,
                      "runAsNonRoot": true,
                      "runAsUser": 1000,
                      "runAsGroup": 1000
                    },
                    "ttlSecondsAfterFinished": 600,
                    "pollIntervalSeconds": 15
                  },
                  "modelConfig": {
                    "name": "nemo-model-config"
                  },
                  "trainingConfig": {
                    "configMap": {
                      "name": "nemo-training-config"
                    },
                    "modelPVC": {
                      "create": true,
                      "name": "finetuning-ms-models-pvc",
                      "storageClass": "",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": 50
                    },
                    "workspacePVC": {
                      "storageClass": "local-path",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": 10,
                      "mountPath": "/pvc/workspace"
                    },
                    "image": {
                      "repository": "nvcr.io/nvidia/nemo-microservices/customizer",
                      "tag": "25.08"
                    },
                    "env": [
                      { "name": "LOG_LEVEL", "value": "INFO" }
                    ],
                    "networkConfig": [
                      { "name": "NCCL_IB_SL", "value": "0" },
                      { "name": "NCCL_IB_TC", "value": "41" },
                      { "name": "NCCL_IB_QPS_PER_CONNECTION", "value": "4" },
                      { "name": "UCX_TLS", "value": "TCP" },
                      { "name": "UCX_NET_DEVICES", "value": "eth0" },
                      { "name": "HCOLL_ENABLE_MCAST_ALL", "value": "0" },
                      { "name": "NCCL_IB_GID_INDEX", "value": "3" }
                    ],
                    "ttlSecondsAfterFinished": 3600,
                    "timeout": 3600,
                    "tolerations": [
                      {
                        "key": "nvidia.com/gpu",
                        "operator": "Exists",
                        "effect": "NoSchedule"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoGuardrail",
                "metadata": {
                  "name": "nemoguardrails-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "configStore": {
                    "pvc": {
                      "name": "pvc-guardrail-config",
                      "create": true,
                      "storageClass": "",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": "1Gi"
                    }
                  },
                  "nimEndpoint": {
                    "baseURL": "http://meta-llama3-1b-instruct.nemo.svc.cluster.local:8000/v1"
                  },
                  "databaseConfig": {
                    "host": "guardrail-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "guardraildb",
                    "credentials": {
                      "user": "guardrailuser",
                      "secretName": "guardrail-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/guardrails",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [
                      "ngc-secret"
                    ]
                  },
                  "metrics": {
                    "serviceMonitor": {}
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "ephemeral-storage": "10Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoEvaluator",
                "metadata": {
                  "name": "nemoevaluator-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "evaluationImages": {
                    "bigcodeEvalHarness": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-bigcode@sha256:62f74c32d1ed07925ddb906de8975025bf63d73e82e7faefb840339aaf21a9dc",
                    "lmEvalHarness": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-lm-eval-harness@sha256:58e7c3da7f1edfed2ff508fa8abab7660366c2c17bb55d17aa68159d6e16be6d",
                    "similarityMetrics": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-custom-eval@sha256:eb51edf668481d1f5d2b100c850958dadf184fd8779433e45da40fa2a35ec4b9",
                    "llmAsJudge": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055",
                    "mtBench": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055",
                    "retriever": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-retriever@sha256:cb99fb3ea5b4e9989832679d4c6028573c28b99e43744e008ac5c9748b7dd250",
                    "rag": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-rag@sha256:b5e35cf277185e0901a4bbba35780b3673be71ae2c5e8efcdbab7710a8a6f783",
                    "bfcl": "nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-bfcl@sha256:95f45f66140d481d544b36b16bc3de7ad5e090d76225fd73ed70461b9bd9907c",
                    "agenticEval": "nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-agentic-eval@sha256:b38bbc5e61df0930204efa096cf4f40db4655b92d03741190513b019d67fc7dc"
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/evaluator",
                    "tag": "25.06",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "argoWorkflows": {
                    "endpoint": "https://argo-workflows-server.nemo.svc.cluster.local:2746",
                    "serviceAccount": "argo-workflows-executor"
                  },
                  "vectorDB": {
                    "endpoint": "http://milvus.nemo.svc.cluster.local:19530"
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000/v1/hf"
                  },
                  "entitystore": {
                    "endpoint": "http://nemoentitystore-sample.nemo.svc.cluster.local:8000"
                  },
                  "databaseConfig": {
                    "host": "evaluator-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "evaldb",
                    "credentials": {
                      "user": "evaluser",
                      "secretName": "evaluator-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "otel": {
                    "enabled": true,
                    "exporterOtlpEndpoint": "http://evaluator-otel-opentelemetry-collector.nemo.svc.cluster.local:4317"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoEntitystore",
                "metadata": {
                  "name": "nemoentitystore-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/entity-store",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "databaseConfig": {
                    "databaseName": "nesdb",
                    "host": "entity-store-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "credentials": {
                      "user": "nesuser",
                      "secretName": "entity-store-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000"
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoDatastore",
                "metadata": {
                  "name": "nemodatastore-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "secrets": {
                    "datastoreConfigSecret": "nemo-ms-nemo-datastore",
                    "datastoreInitSecret": "nemo-ms-nemo-datastore-init",
                    "datastoreInlineConfigSecret": "nemo-ms-nemo-datastore-inline-config",
                    "giteaAdminSecret": "gitea-admin-credentials",
                    "lfsJwtSecret": "nemo-ms-nemo-datastore--lfs-jwt"
                  },
                  "databaseConfig": {
                    "credentials": {
                      "user": "ndsuser",
                      "secretName": "datastore-pg-existing-secret",
                      "passwordKey": "password"
                    },
                    "host": "datastore-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "ndsdb"
                  },
                  "pvc": {
                    "name": "pvc-shared-data",
                    "create": true,
                    "storageClass": "",
                    "volumeAccessMode": "ReadWriteOnce",
                    "size": "10Gi"
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/datastore",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "replicas": 1,
                  "resources": {
                    "requests": {
                      "memory": "256Mi",
                      "cpu": "500m"
                    },
                    "limits": {
                      "memory": "512Mi",
                      "cpu": "1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: nvcr.io/nvidia/cloud-native/k8s-nim-operator@sha256:f42dd5c3dc269553ee46725ed4b1ab2d2f253ef06117f1d9d52faaa22d8fe60d
          createdAt: Wed Sep 10 05:49:48 PDT 2025
          description: An Operator for deployment and maintenance of NVIDIA NIMs and
            NeMo microservices in a Kubernetes environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          k8sMinVersion: 1.26.0
          repository: github.com/NVIDIA/k8s-nim-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NIM Cache
            displayName: NIMCache
            kind: NIMCache
            name: nimcaches.apps.nvidia.com
            version: v1alpha1
          - description: NIM Service
            displayName: NIMService
            kind: NIMService
            name: nimservices.apps.nvidia.com
            version: v1alpha1
          - description: NIM Pipeline
            displayName: NIMPipeline
            kind: NIMPipeline
            name: nimpipelines.apps.nvidia.com
            version: v1alpha1
          - description: NIM Build
            displayName: NIMBuild
            kind: NIMBuild
            name: nimbuilds.apps.nvidia.com
            version: v1alpha1
          - description: NEMO guardrail Service
            displayName: NemoGuardrail
            kind: NemoGuardrail
            name: nemoguardrails.apps.nvidia.com
            version: v1alpha1
          - description: NEMO datastore Service
            displayName: NemoDatastore
            kind: NemoDatastore
            name: nemodatastores.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Eval Service
            displayName: NemoEvaluator
            kind: NemoEvaluator
            name: nemoevaluators.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Customizer Service
            displayName: NemoCustomizer
            kind: NemoCustomizer
            name: nemocustomizers.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Entitystore Service
            displayName: NemoEntitystore
            kind: NemoEntitystore
            name: nemoentitystores.apps.nvidia.com
            version: v1alpha1
        description: |-
          ## NVIDIA NIM Operator
          The NVIDIA NIM Operator enables Kubernetes cluster administrators to operate the software components and services necessary to run NVIDIA NIMs in various domains such as reasoning, retrieval, speech, and biology. Additionally, it allows the use of NeMo Microservices to fine-tune, evaluate, or apply guardrails to your models.
          The Operator manages the life cycle of the following microservices and the models they use:
          ### NVIDIA NIM models, such as:
          * Reasoning LLMs
          * Retrieval - Embedding, Reranking, etc.
          * Speech
          * Biology
          ### NeMo core microservices:
          * NeMo Customizer
          * NeMo Evaluator
          * NeMo Guardrails
          ### NeMo platform component microservices:
          * NeMo Data Store
          * NeMo Entity Store
        displayName: The NVIDIA NIM Operator for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        links:
        - name: NVIDIA NIM
          url: https://docs.nvidia.com/nim/index.html
        - name: Documentation
          url: https://docs.nvidia.com/nim-operator/latest
        maintainers:
        - email: operator_feedback@nvidia.com
          name: shivamerla
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: NVIDIA
        relatedImages:
        - nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-bfcl@sha256:95f45f66140d481d544b36b16bc3de7ad5e090d76225fd73ed70461b9bd9907c
        - nvcr.io/nvidia/nemo-microservices/customizer-api@sha256:c590fa19121af2aac0379933deace3991f27c3ef88a8043560d6e3a172273b0b
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-bigcode@sha256:62f74c32d1ed07925ddb906de8975025bf63d73e82e7faefb840339aaf21a9dc
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-custom-eval@sha256:eb51edf668481d1f5d2b100c850958dadf184fd8779433e45da40fa2a35ec4b9
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-retriever@sha256:cb99fb3ea5b4e9989832679d4c6028573c28b99e43744e008ac5c9748b7dd250
        - nvcr.io/nvidia/cloud-native/k8s-nim-operator@sha256:f42dd5c3dc269553ee46725ed4b1ab2d2f253ef06117f1d9d52faaa22d8fe60d
        - nvcr.io/nim/meta/llama-3.2-1b-instruct@sha256:1ff21ad32124d2382f8c733879394e7849cc3df5172651a24fcb671ac4fe77c6
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-lm-eval-harness@sha256:58e7c3da7f1edfed2ff508fa8abab7660366c2c17bb55d17aa68159d6e16be6d
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-rag@sha256:b5e35cf277185e0901a4bbba35780b3673be71ae2c5e8efcdbab7710a8a6f783
        - nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-agentic-eval@sha256:b38bbc5e61df0930204efa096cf4f40db4655b92d03741190513b019d67fc7dc
        - nvcr.io/nvidia/nemo-microservices/nds-v2-huggingface-cli@sha256:7958e5244217de8af0ed723dfe7ea8e291b3afbb2a0aa73f5c27b218a1149b11
        version: 3.0.0
      entries:
      - name: nim-operator-certified.v3.0.0
        version: 3.0.0
      - name: nim-operator-certified.v2.0.2
        version: 2.0.2
      - name: nim-operator-certified.v2.0.1
        version: 2.0.1
      - name: nim-operator-certified.v2.0.0
        version: 2.0.0
      - name: nim-operator-certified.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: stable
    packageName: nim-operator-certified
    provider:
      name: NVIDIA
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-druid-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: druid-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:909a03d95f8173c418f6547092421a5531ff8957aa4319ce5fc890f32fe1f7be
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:909a03d95f8173c418f6547092421a5531ff8957aa4319ce5fc890f32fe1f7be
        version: 24.11.1
      entries:
      - name: druid-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: druid-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:253239268b8d32dbea10233fe6ab202bbe8a074c369a929630681dea37935383
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:253239268b8d32dbea10233fe6ab202bbe8a074c369a929630681dea37935383
        version: 25.3.0
      entries:
      - name: druid-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: druid-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
        version: 25.7.0
      entries:
      - name: druid-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: druid-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
        version: 25.7.0
      entries:
      - name: druid-operator.v25.7.0
        version: 25.7.0
      - name: druid-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-druid-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: eap
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eap-operator.v3.2.2
      currentCSVDesc:
        annotations:
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "false"
          containerImage: registry.redhat.io/jboss-eap/eap-rhel9-operator@sha256:2361e73b8c517177e80f60d204996ebf65b166e4e8b60b329323a2b409c54f6c
          createdAt: "2025-07-10T14:38:00Z"
          description: Operator that creates and manages Java applications running
            on JBoss EAP.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat JBoss Enterprise Application
            Platform", "Red Hat Runtimes", "Red Hat Integration", "Red Hat Process
            Automation"]'
          repository: https://github.com/wildfly/wildfly-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An application running on EAP application runtime.
            displayName: WildFlyServer
            kind: WildFlyServer
            name: wildflyservers.wildfly.org
            version: v1alpha1
        description: "# Red Hat JBoss Enterprise Application Platform (JBoss EAP)\n\nRed
          Hat® JBoss® Enterprise Application Platform (JBoss EAP) delivers enterprise-grade
          security, performance, and scalability in any environment.  Some features
          of JBoss EAP are:\n\n**Optimized for cloud and containers**: with simplified
          deployment and full Jakarta EE performance for applications in any environment.\n\n**Lightweight,
          flexible architecture**: which is well-suited for microservices and traditional
          applications.\n\n**Support and standardize microservices development**:
          With the JBoss Enterprise Application Platform expansion pack that supports
          Eclipse MicroProfile application programming interfaces (APIs).\n\n# JBoss
          EAP Operator\nOperators are software components that extend the Kubernetes
          API to customize how to manage applications and their components.\n\nThe
          JBoss EAP Operator facilitates the deployment and management of JBoss EAP
          on container orchestration platforms such as Red Hat OpenShift. The operator
          acts as an automated management tool, simplifying the process of provisioning,
          scaling, and updating JBoss EAP applications.\n\n# JBoss EAP Operator capabilities\nThe
          JBoss EAP Operator deploys and manages multiple instances of the JBoss EAP
          Java application in a Red Hat OpenShift cluster. \n\n**Simplifies management**:
          Creates, configures, manages, and seamlessly upgrades instances of complex
          stateful applications. No more manual configuration, making management a
          breeze.\n\n**Efficient scaling**: Effortlessly scales the instances of your
          JBoss EAP Java application based on demand.\n\n**Faster updates**: Seamlessly
          updates, enabling you to roll out new operator features and bug fixes without
          downtime\n\n**Safe transaction recovery**: Provides safe transaction recovery
          in your application cluster by verifying all transactions are completed
          before scaling down the replicas and marking a pod as clean for termination.\n\n#
          Operator installation and configuration\nLet's walk through the steps to
          install the JBoss EAP Operator on your OpenShift Cluster and deploy your
          first example application:\n\n**Step 1: Install JBoss EAP from OperatorHub**\n1.
          Click Install in the JBoss EAP operator window.\n2. In the Install Operator
          window, for Installation mode, select All namespaces on the cluster The
          operator is available in openshift-operators namespace.\n\n**Step 2: Build
          an application image**\n\nThe JBoss EAP Operator deploys an application
          image. If you do not have your JBoss EAP application image built yet, you
          can use the following example to build the image using Helm Charts.\n\n1.
          In the main navigation, click the dropdown menu (perspective switcher) and
          select Developer.\n2. Click Add > Helm Chart.\n3. Search for \"JBoss EAP
          8\" (Provided by Red Hat) and create a new Helm Chart release.\n4. On the
          Release name text box, type “eap-application”. \n5. Switch to the YAML view
          and replace the existing code with the following:\n\n    ```\n    build:\n
          \     uri: https://github.com/jboss-eap-up-and-running/eap8-getting-started\n
          \   deploy:\n      enabled: false\n    ```\n\n6. Click Create to build the
          application image.\n\n    The application image is created with the name
          eap-application:latest.\n\n**Step 3: Deploy your application with JBoss
          EAP Operator**\n\n1. In the main navigation, click the dropdown menu (perspective
          switcher) and select Developer.\n2. Click Add > Operator Backed > WildFlyServer
          > Create.\n3. Switch to the YAML view and replace the existing code with
          the following:\n\n    ```\n    apiVersion: wildfly.org/v1alpha1\n    kind:
          WildFlyServer\n    metadata:\n      name: eap-application\n    spec:\n      applicationImage:
          eap-application:latest\n      replicas: 1\n    ```\n\n 4. Click Create to
          deploy the application with JBoss EAP Operator.\n"
        displayName: JBoss EAP
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eap
        - java
        - open source
        - application runtime
        links:
        - name: JBoss EAP Homepage
          url: https://www.redhat.com/en/technologies/jboss-middleware/application-platform
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_jboss_enterprise_application_platform/
        - name: Blog
          url: https://developers.redhat.com/blog/
        - name: Quickstart
          url: https://github.com/jboss-developer/jboss-eap-quickstarts
        - name: Red Hat Customer Portal
          url: https://access.redhat.com/support
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat, Inc.
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/jboss-eap/eap-rhel9-operator@sha256:2361e73b8c517177e80f60d204996ebf65b166e4e8b60b329323a2b409c54f6c
        version: 3.2.2
      entries:
      - name: eap-operator.v3.2.2
        version: 3.2.2
      - name: eap-operator.v3.2.1
        version: 3.2.1
      - name: eap-operator.v3.2.0
        version: 3.2.0
      - name: eap-operator.v3.1.9
        version: 3.1.9
      - name: eap-operator.v3.1.8
        version: 3.1.8
      - name: eap-operator.v3.1.7
        version: 3.1.7
      - name: eap-operator.v3.1.6
        version: 3.1.6
      - name: eap-operator.v3.1.5
        version: 3.1.5
      - name: eap-operator.v3.1.4
        version: 3.1.4
      - name: eap-operator.v3.1.3
        version: 3.1.3
      - name: eap-operator.v3.1.2
        version: 3.1.2
      - name: eap-operator.v3.1.1
        version: 3.1.1
      - name: eap-operator.v3.1.0
        version: 3.1.0
      - name: eap-operator.v3.0.2
        version: 3.0.2
      - name: eap-operator.v3.0.1
        version: 3.0.1
      - name: eap-operator.v3.0.0
        version: 3.0.0
      - name: eap-operator.v2.4.3
        version: 2.4.3
      - name: eap-operator.v2.4.2
        version: 2.4.2
      - name: eap-operator.v2.4.1
        version: 2.4.1
      - name: eap-operator.v2.4.0
        version: 2.4.0
      - name: eap-operator.v2.3.10
        version: 2.3.10
      - name: eap-operator.v2.3.9
        version: 2.3.9
      - name: eap-operator.v2.3.8
        version: 2.3.8
      - name: eap-operator.v2.3.7
        version: 2.3.7
      - name: eap-operator.v2.3.6
        version: 2.3.6
      - name: eap-operator.v2.3.5
        version: 2.3.5
      - name: eap-operator.v2.3.4
        version: 2.3.4
      - name: eap-operator.v2.3.3
        version: 2.3.3
      - name: eap-operator.v2.3.2
        version: 2.3.2
      - name: eap-operator.v2.3.1
        version: 2.3.1
      - name: eap-operator.v2.3.0
        version: 2.3.0
      - name: eap-operator.v2.2.2
        version: 2.2.2
      - name: eap-operator.v2.2.1
        version: 2.2.1
      - name: eap-operator.v2.2.0
        version: 2.2.0
      name: stable
    defaultChannel: stable
    packageName: eap
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: m88i Labs
      provider-url: ""
    name: nexus-operator-m88i
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.m88i.io/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "name": "nexus3"
                },
                "spec": {
                  "networking": {
                    "expose": false
                  },
                  "persistence": {
                    "persistent": false
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    },
                    "requests": {
                      "cpu": "1",
                      "memory": "2Gi"
                    }
                  },
                  "useRedHatImage": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/m88i/nexus-operator:0.6.0
          createdAt: "2019-11-16T13:12:22Z"
          description: Nexus Operator to deploy and manage Nexus 3.x servers
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/m88i/nexus-operator
          support: m88i Labs
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: Nexus
            kind: Nexus
            name: nexus.apps.m88i.io
            version: v1alpha1
        description: |-
          Creates a new Nexus 3.x deployment in a Kubernetes cluster. Will help DevOps to have a quick Nexus application exposed to the world that can be used in a CI/CD process:

          * Deploys a new Nexus 3.x server based on either Community or Red Hat images
          * Creates an [Ingress](https://kubernetes.io/docs/concepts/services-networking/ingress/) in Kubernetes (1.14+) environments to expose the application to the world
          * On OpenShift, creates a Route to expose the service outside the cluster
          * Automatically creates Apache, Red Hat and JBoss Maven repositories
          * Automatically updates Nexus within the same minor version

          [See our documentation](https://github.com/m88i/nexus-operator/blob/main/README.md) for more installation and usage scenarios.

          If you experience any issues or have any ideas for new features, please [file an issue in our Github repository](https://github.com/m88i/nexus-operator/issues) or send an email to our maillist: [nexus-operator@googlegroups.com](mailto:nexus-operator@googlegroups.com)

          *Please note that the operator is an individual work and it's not provided nor supported by Sonatype.*
        displayName: Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nexus
        - sonatype
        - maven
        - docker
        - ci
        - continuous integration
        - continuous delivery
        - repository
        - repository manager
        - dev tools
        links:
        - name: Documentation
          url: https://github.com/m88i/nexus-operator/blob/main/README.md
        - name: Nexus Operator source code repository
          url: https://github.com/m88i/nexus-operator
        maintainers:
        - email: nexus-operator@googlegroups.com
          name: m88i Labs
        maturity: alpha
        provider:
          name: m88i Labs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/m88i/nexus-operator:0.6.0
        version: 0.6.0
      entries:
      - name: nexus-operator.v0.6.0
        version: 0.6.0
      name: alpha
    defaultChannel: alpha
    packageName: nexus-operator-m88i
    provider:
      name: m88i Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
    name: flink-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flink-kubernetes-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"flink.apache.org/v1beta1",
                "kind":"FlinkDeployment",
                "metadata":{
                    "name":"basic-example"
                },
                "spec":{
                    "image":"flink:1.16",
                    "flinkVersion":"v1_16",
                    "flinkConfiguration":{
                        "taskmanager.numberOfTaskSlots":"2"
                    },
                    "serviceAccount":"flink",
                    "jobManager":{
                        "resource":{
                            "memory":"2048m",
                            "cpu":1
                        }
                    },
                    "taskManager":{
                        "resource":{
                            "memory":"2048m",
                            "cpu":1
                        }
                    },
                    "job":{
                        "jarURI":"local:///opt/flink/examples/streaming/StateMachineExample.jar",
                        "parallelism":2,
                        "upgradeMode":"stateless"
                    }
                }
              },
              {
                "apiVersion":"flink.apache.org/v1beta1",
                "kind":"FlinkSessionJob",
                "metadata":{
                  "name":"basic-session-job-example2"
                },
                "spec":{
                  "deploymentName":"basic-session-cluster",
                  "job":{
                      "jarURI":"https://repo1.maven.org/maven2/org/apache/flink/flink-examples-streaming_2.12/1.16.1/flink-examples-streaming_2.12-1.16.1.jar",
                      "parallelism":2,
                      "upgradeMode":"stateless",
                      "entryClass":"org.apache.flink.streaming.examples.statemachine.StateMachineExample"
                  }
                }
              },
              {
                "apiVersion": "flink.apache.org/v1beta1",
                "kind": "FlinkStateSnapshot",
                "metadata": {
                  "name": "example-savepoint"
                },
                "spec": {
                  "savepoint": {
                    "alreadyExists": true,
                    "disposeOnDelete": false,
                    "path": "/flink-data/savepoints/savepoint-98682a-21931bf10c39"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "basic-example": {
                "description": "Example flink deployment"
              },
              "basic-session-job-example2": {
                "description": "Example flink session job"
              }
            }
          capabilities: Deep Insights
          categories: Big Data, Streaming & Messaging
          containerImage: ghcr.io/apache/flink-kubernetes-operator:1.12.0
          createdAt: "2025-06-23 18:42:54"
          description: Flink Kubernetes Operator manages the complete deployment lifecycle
            of Apache Flink applications.
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/apache/flink-kubernetes-operator.git
          support: Apache Software Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The FlinkDeployment CR defines Flink Application and Session
              cluster deployments
            displayName: Flink Deployment
            kind: FlinkDeployment
            name: flinkdeployments.flink.apache.org
            version: v1beta1
          - description: The FlinkSessionJob CR defines the session job on the Session
              cluster and each Session cluster can run multiple FlinkSessionJob.
            displayName: Flink Session Job
            kind: FlinkSessionJob
            name: flinksessionjobs.flink.apache.org
            version: v1beta1
          - description: The FlinkStateSnapshots CR is used to describe savepoint
              or checkpoint for a Flink job.
            displayName: Flink State Snapshot
            kind: FlinkStateSnapshot
            name: flinkstatesnapshots.flink.apache.org
            version: v1beta1
        description: |-
          Flink Kubernetes Operator acts as a control plane to manage the complete deployment lifecycle of Apache Flink applications.
          ## Installation
          This operator can be installed from [OperatorHub.io](https://operatorhub.io/operator/flink-kubernetes-operator).

          By default, the Flink operator monitors and run Flink applications in the same namespace as the operator.
          To run Flink jobs in another namespace, users are responsible for setting up the RBAC in that namespace.

          For example, users need to do the following for each additional namespace that runs the Flink applications:

          1. Switch to the namespace by running:

              ```sh
              kubectl config set-context --current --namespace=CHANGEIT
              ```

          2. Create the service account, role, and role binding in the namespace using the commands below:

              ```sh
              kubectl apply -f - <<EOF
              apiVersion: v1
              kind: ServiceAccount
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink
              ---
              apiVersion: rbac.authorization.k8s.io/v1
              kind: Role
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink
              rules:
              - apiGroups:
                - ""
                resources:
                - pods
                - configmaps
                verbs:
                - '*'
              - apiGroups:
                - apps
                resources:
                - deployments
                - deployments/finalizers
                verbs:
                - '*'
              ---
              apiVersion: rbac.authorization.k8s.io/v1
              kind: RoleBinding
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink-role-binding
              roleRef:
                apiGroup: rbac.authorization.k8s.io
                kind: Role
                name: flink
              subjects:
              - kind: ServiceAccount
                name: flink
              EOF
              ```

          3. Proceed deploying Flink applications using the the Custom Resource Definition below.

              ```sh
              kubectl create -f https://raw.githubusercontent.com/apache/flink-kubernetes-operator/release-1.12.0/examples/basic.yaml
              ```

          See [Flink custom resources](https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/custom-resource/overview/#overview) for more detail.
        displayName: Flink Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flink
        - streaming
        links:
        - name: Website
          url: https://flink.apache.org/
        - name: Documentation
          url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/
        - name: Mailing list
          url: https://lists.apache.org/list.html?dev@flink.apache.org
        - name: Slack
          url: https://apache-flink.slack.com/join/shared_invite/zt-1llkzbgyt-K2nNGGg88rfsDGLkT09Qzg#/shared-invite/email
        - name: GitHub
          url: https://github.com/apache/flink-kubernetes-operator
        maintainers:
        - email: mbalassi@apache.org
          name: Marton Balassi
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: jbusche@us.ibm.com
          name: James Busche
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Community
          url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
        relatedImages:
        - ghcr.io/apache/flink-kubernetes-operator:1.12.0
        version: 1.12.0
      entries:
      - name: flink-kubernetes-operator.v1.12.0
        version: 1.12.0
      - name: flink-kubernetes-operator.v1.11.0
        version: 1.11.0
      - name: flink-kubernetes-operator.v1.10.0
        version: 1.10.0
      - name: flink-kubernetes-operator.v1.6.0
        version: 1.6.0
      - name: flink-kubernetes-operator.v1.5.0
        version: 1.5.0
      - name: flink-kubernetes-operator.v1.4.0
        version: 1.4.0
      - name: flink-kubernetes-operator.v1.3.1
        version: 1.3.1
      - name: flink-kubernetes-operator.v1.3.0
        version: 1.3.0
      name: alpha
    defaultChannel: alpha
    packageName: flink-kubernetes-operator
    provider:
      name: Community
      url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeRocketCI
      provider-url: https://docs.kuberocketci.io/
    name: edp-keycloak-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: edp-keycloak-operator.v1.19.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "keycloak-sample"
                },
                "spec": {
                  "secret": "my-keycloak-secret",
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakAuthFlow",
                "metadata": {
                  "name": "keycloakauthflow-sample"
                },
                "spec": {
                  "alias": "MyBrowser",
                  "authenticationExecutions": [
                    {
                      "authenticator": "auth-cookie",
                      "priority": 0,
                      "requirement": "ALTERNATIVE"
                    },
                    {
                      "authenticator": "identity-provider-redirector",
                      "authenticatorConfig": {
                        "alias": "my-alias",
                        "config": {
                          "defaultProvider": "my-alias"
                        }
                      },
                      "priority": 1,
                      "requirement": "REQUIRED"
                    }
                  ],
                  "builtIn": false,
                  "description": "browser with idp",
                  "providerId": "basic-flow",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "topLevel": true
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-sample"
                },
                "spec": {
                  "advancedProtocolMappers": true,
                  "clientId": "agocd",
                  "defaultClientScopes": [
                    "argocd_groups"
                  ],
                  "directAccess": true,
                  "public": false,
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-name:client-secret-key",
                  "webUrl": "https://argocd.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClientScope",
                "metadata": {
                  "name": "keycloakclientscope-sample"
                },
                "spec": {
                  "description": "Group Membership",
                  "name": "groups",
                  "protocol": "openid-connect",
                  "protocolMappers": [
                    {
                      "config": {
                        "access.token.claim": "true",
                        "claim.name": "groups",
                        "full.path": "false",
                        "id.token.claim": "true",
                        "userinfo.token.claim": "true"
                      },
                      "name": "groups",
                      "protocol": "openid-connect",
                      "protocolMapper": "oidc-group-membership-mapper"
                    }
                  ],
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "keycloakrealm-sample"
                },
                "spec": {
                  "id": "d1-id-kc-realm-name",
                  "keycloakRef": {
                    "kind": "Keycloak",
                    "name": "keycloak-sample"
                  },
                  "passwordPolicy": [
                    {
                      "type": "forceExpiredPasswordChange",
                      "value": "365"
                    },
                    {
                      "type": "length",
                      "value": "8"
                    }
                  ],
                  "realmEventConfig": {
                    "adminEventsDetailsEnabled": false,
                    "adminEventsEnabled": true,
                    "enabledEventTypes": [
                      "UPDATE_CONSENT_ERROR",
                      "CLIENT_LOGIN"
                    ],
                    "eventsEnabled": true,
                    "eventsExpiration": 15000,
                    "eventsListeners": [
                      "jboss-logging"
                    ]
                  },
                  "realmName": "d2-id-kc-realm-name",
                  "ssoAutoRedirectEnabled": false,
                  "ssoRealmEnabled": false,
                  "ssoRealmName": "openshift"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmComponent",
                "metadata": {
                  "name": "keycloakrealmcomponent-sample"
                },
                "spec": {
                  "config": {
                    "allowPasswordAuthentication": [
                      "true"
                    ],
                    "cachePolicy": [
                      "EVICT_WEEKLY"
                    ],
                    "debug": [
                      "true"
                    ],
                    "editMode": [
                      "READ_ONLY"
                    ],
                    "enabled": [
                      "true"
                    ],
                    "evictionDay": [
                      "3"
                    ],
                    "evictionHour": [
                      "5"
                    ],
                    "evictionMinute": [
                      "7"
                    ],
                    "kerberosRealm": [
                      "test-realm"
                    ],
                    "keyTab": [
                      "test-key-tab"
                    ],
                    "priority": [
                      "0"
                    ],
                    "serverPrincipal": [
                      "srv-principal-test"
                    ],
                    "updateProfileFirstLogin": [
                      "true"
                    ]
                  },
                  "name": "cr-kerb-test",
                  "providerId": "kerberos",
                  "providerType": "org.keycloak.storage.UserStorageProvider",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmGroup",
                "metadata": {
                  "name": "keycloakrealmgroup-sample"
                },
                "spec": {
                  "name": "ArgoCDAdmins",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "keycloakrealmidentityprovider-sample"
                },
                "spec": {
                  "alias": "instagram",
                  "authenticateByDefault": false,
                  "config": {
                    "clientId": "foo",
                    "clientSecret": "$secretName:secretKey",
                    "hideOnLoginPage": "true",
                    "syncMode": "IMPORT",
                    "useJwksUrl": "true"
                  },
                  "enabled": true,
                  "firstBrokerLoginFlowAlias": "first broker login",
                  "mappers": [
                    {
                      "config": {
                        "role": "role-tr",
                        "syncMode": "INHERIT"
                      },
                      "identityProviderAlias": "instagram",
                      "identityProviderMapper": "oidc-hardcoded-role-idp-mapper",
                      "name": "test3212"
                    },
                    {
                      "config": {
                        "attribute": "foo",
                        "attribute.value": "bar",
                        "syncMode": "IMPORT"
                      },
                      "identityProviderAlias": "instagram",
                      "identityProviderMapper": "hardcoded-attribute-idp-mapper",
                      "name": "test-33221"
                    }
                  ],
                  "providerId": "instagram",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRole",
                "metadata": {
                  "name": "keycloakrealmrole-sample"
                },
                "spec": {
                  "composite": true,
                  "description": "default developer role",
                  "name": "developer",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRoleBatch",
                "metadata": {
                  "name": "keycloakrealmrolebatch-sample"
                },
                "spec": {
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "roles": [
                    {
                      "composite": true,
                      "description": "default developer role",
                      "isDefault": false,
                      "name": "developer"
                    },
                    {
                      "composite": true,
                      "description": "default administrator role",
                      "isDefault": false,
                      "name": "administrator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmUser",
                "metadata": {
                  "name": "keycloakrealmuser-sample"
                },
                "spec": {
                  "attributes": {
                    "baz": "jazz",
                    "foo": "bar"
                  },
                  "email": "john.snow13@example.com",
                  "emailVerified": true,
                  "enabled": true,
                  "firstName": "John",
                  "keepResource": true,
                  "lastName": "Snow",
                  "password": "12345678",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "requiredUserActions": [
                    "UPDATE_PASSWORD"
                  ],
                  "username": "john.snow13"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloak",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloak-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloak",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloak-sample"
                },
                "spec": {
                  "secret": "keycloak-access",
                  "url": "https://keycloak.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloakRealm",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloakrealm-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloakrealm",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloakrealm-sample"
                },
                "spec": {
                  "clusterKeycloakRef": "clusterkeycloak-sample",
                  "realmName": "realm-sample"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          containerImage: docker.io/epamedp/keycloak-operator:1.19.0
          createdAt: "2023-11-15T16:04:45Z"
          description: An Operator for managing Keycloak
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-keycloak-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterKeycloakRealm is the Schema for the clusterkeycloakrealms
              API.
            displayName: Cluster Keycloak Realm
            kind: ClusterKeycloakRealm
            name: clusterkeycloakrealms.v1.edp.epam.com
            version: v1alpha1
          - description: ClusterKeycloak is the Schema for the clusterkeycloaks API.
            displayName: Cluster Keycloak
            kind: ClusterKeycloak
            name: clusterkeycloaks.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakAuthFlow is the Schema for the keycloak authentication
              flow API.
            displayName: Keycloak Auth Flow
            kind: KeycloakAuthFlow
            name: keycloakauthflows.v1.edp.epam.com
            version: v1
          - description: KeycloakClient is the Schema for the keycloak clients API.
            displayName: Keycloak Client
            kind: KeycloakClient
            name: keycloakclients.v1.edp.epam.com
            version: v1
          - description: KeycloakClientScope is the Schema for the keycloakclientscopes
              API.
            displayName: Keycloak Client Scope
            kind: KeycloakClientScope
            name: keycloakclientscopes.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmComponent is the Schema for the keycloak component
              API.
            displayName: Keycloak Realm Component
            kind: KeycloakRealmComponent
            name: keycloakrealmcomponents.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmGroup is the Schema for the keycloak group API.
            displayName: Keycloak Realm Group
            kind: KeycloakRealmGroup
            name: keycloakrealmgroups.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmIdentityProvider is the Schema for the keycloak
              realm identity provider API.
            displayName: Keycloak Realm Identity Provider
            kind: KeycloakRealmIdentityProvider
            name: keycloakrealmidentityproviders.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRoleBatch is the Schema for the keycloak roles
              API.
            displayName: Keycloak Realm Role Batch
            kind: KeycloakRealmRoleBatch
            name: keycloakrealmrolebatches.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRole is the Schema for the keycloak group API.
            displayName: Keycloak Realm Role
            kind: KeycloakRealmRole
            name: keycloakrealmroles.v1.edp.epam.com
            version: v1
          - description: KeycloakRealm is the Schema for the keycloak realms API.
            displayName: Keycloak Realm
            kind: KeycloakRealm
            name: keycloakrealms.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmUser is the Schema for the keycloak user API.
            displayName: Keycloak Realm User
            kind: KeycloakRealmUser
            name: keycloakrealmusers.v1.edp.epam.com
            version: v1
          - description: Keycloak is the Schema for the keycloaks API.
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.v1.edp.epam.com
            version: v1
        description: |
          Keycloak Operator is an operator that is responsible for establishing
          a connection to provided Keycloak Server, reconciling Keycloak entities (realms,
          roles, groups, users, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Keycloak `Master` realm, and assign a `create-realm` role.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  keycloak-access
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Keycloak` with Keycloak instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: Keycloak
              metadata:
                name: keycloak-sample
              spec:
                secret: keycloak-access             # Secret name
                url: https://keycloak.example.com   # Keycloak URL
              ```

              Wait for the `.status` field with  `status.connected: true`

          4. Create Keycloak realm and group using Custom Resources:

            ```yaml
            apiVersion: v1.edp.epam.com/v1
            kind: KeycloakRealm
            metadata:
              name: keycloakrealm-sample
            spec:
              realmName: realm-sample
              keycloakOwner: keycloak-sample   # the name of `kind: Keycloak`
              ```

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: KeycloakRealmGroup
              metadata:
                name: argocd-admins
              spec:
                name: ArgoCDAdmins
                realm: keycloakrealm-sample   # the name of `kind: KeycloakRealm`
              ```
        displayName: EDP Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - authentication
        - authorization
        - edp
        - idp
        - keycloak
        - oauth
        - oidc
        - operator
        - saml
        - sso
        links:
        - name: Edp Keycloak Operator
          url: https://github.com/epam/edp-keycloak-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/keycloak-operator:1.19.0
        version: 1.19.0
      entries:
      - name: edp-keycloak-operator.v1.19.0
        version: 1.19.0
      - name: edp-keycloak-operator.v1.18.2
        version: 1.18.2
      - name: edp-keycloak-operator.v1.17.1
        version: 1.17.1
      - name: edp-keycloak-operator.v1.13.0
        version: 1.13.0
      name: alpha
    - currentCSV: edp-keycloak-operator.v1.29.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "keycloak-sample"
                },
                "spec": {
                  "secret": "my-keycloak-secret",
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakAuthFlow",
                "metadata": {
                  "name": "keycloakauthflow-sample"
                },
                "spec": {
                  "alias": "MyBrowser",
                  "authenticationExecutions": [
                    {
                      "authenticator": "auth-cookie",
                      "priority": 0,
                      "requirement": "ALTERNATIVE"
                    },
                    {
                      "authenticator": "identity-provider-redirector",
                      "authenticatorConfig": {
                        "alias": "my-alias",
                        "config": {
                          "defaultProvider": "my-alias"
                        }
                      },
                      "priority": 1,
                      "requirement": "REQUIRED"
                    }
                  ],
                  "builtIn": false,
                  "description": "browser with idp",
                  "providerId": "basic-flow",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "topLevel": true
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-authorization-sample"
                },
                "spec": {
                  "adminFineGrainedPermissionsEnabled": true,
                  "authorization": {
                    "permissions": [
                      {
                        "decisionStrategy": "AFFIRMATIVE",
                        "description": "Resource permission",
                        "logic": "POSITIVE",
                        "name": "resource-permission",
                        "policies": [
                          "role-policy"
                        ],
                        "resources": [
                          "resource1"
                        ],
                        "type": "resource"
                      },
                      {
                        "decisionStrategy": "CONSENSUS",
                        "description": "Scope permission",
                        "logic": "POSITIVE",
                        "name": "scope-permission",
                        "policies": [
                          "role-policy"
                        ],
                        "scopes": [
                          "scope1"
                        ],
                        "type": "scope"
                      }
                    ],
                    "policies": [
                      {
                        "decisionStrategy": "AFFIRMATIVE",
                        "description": "Role policy",
                        "logic": "POSITIVE",
                        "name": "role-policy",
                        "rolePolicy": {
                          "roles": [
                            {
                              "name": "developer",
                              "required": true
                            }
                          ]
                        },
                        "type": "role"
                      },
                      {
                        "aggregatedPolicy": {
                          "policies": [
                            "policy1",
                            "policy2"
                          ]
                        },
                        "description": "Aggregate policy",
                        "name": "aggregate-policy",
                        "type": "aggregate"
                      },
                      {
                        "clientPolicy": {
                          "clients": [
                            "client1",
                            "client2"
                          ]
                        },
                        "description": "Client policy",
                        "name": "client-policy",
                        "type": "client"
                      },
                      {
                        "description": "Group policy",
                        "groupPolicy": {
                          "groups": [
                            {
                              "extendChildren": true,
                              "name": "group1"
                            }
                          ]
                        },
                        "name": "group-policy",
                        "type": "group"
                      },
                      {
                        "description": "Role policy",
                        "name": "role-policy",
                        "rolePolicy": {
                          "roles": [
                            {
                              "name": "developer",
                              "required": true
                            }
                          ]
                        },
                        "type": "role"
                      },
                      {
                        "description": "Time policy",
                        "name": "time-policy",
                        "timePolicy": {
                          "notBefore": "2021-01-01T00:00:00Z",
                          "notOnOrAfter": "2021-12-31T23:59:59Z"
                        },
                        "type": "time"
                      },
                      {
                        "description": "User policy",
                        "name": "user-policy",
                        "type": "user",
                        "userPolicy": {
                          "users": [
                            "user1",
                            "user2"
                          ]
                        }
                      }
                    ],
                    "resources": [
                      {
                        "displayName": "Resource 1",
                        "iconUri": "https://example.com/icon.png",
                        "name": "resource1",
                        "scopes": [
                          "scope1"
                        ],
                        "type": "test"
                      }
                    ],
                    "scopes": [
                      "scope1"
                    ]
                  },
                  "authorizationServicesEnabled": true,
                  "clientId": "authorization-sample",
                  "directAccess": true,
                  "permission": {
                    "scopePermissions": [
                      {
                        "name": "token-exchange",
                        "policies": [
                          "policy1"
                        ]
                      }
                    ]
                  },
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-authorization-sample:client-secret-key",
                  "serviceAccount": {
                    "enabled": true
                  },
                  "webUrl": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-sample"
                },
                "spec": {
                  "advancedProtocolMappers": true,
                  "authenticationFlowBindingOverrides": {
                    "browser": "browser",
                    "directGrant": "direct grant"
                  },
                  "clientId": "agocd",
                  "defaultClientScopes": [
                    "argocd_groups"
                  ],
                  "directAccess": true,
                  "public": false,
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-name:client-secret-key",
                  "webUrl": "https://argocd.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClientScope",
                "metadata": {
                  "name": "keycloakclientscope-sample"
                },
                "spec": {
                  "description": "Group Membership",
                  "name": "groups",
                  "protocol": "openid-connect",
                  "protocolMappers": [
                    {
                      "config": {
                        "access.token.claim": "true",
                        "claim.name": "groups",
                        "full.path": "false",
                        "id.token.claim": "true",
                        "userinfo.token.claim": "true"
                      },
                      "name": "groups",
                      "protocol": "openid-connect",
                      "protocolMapper": "oidc-group-membership-mapper"
                    }
                  ],
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "keycloakrealm-sample"
                },
                "spec": {
                  "id": "d1-id-kc-realm-name",
                  "keycloakRef": {
                    "kind": "Keycloak",
                    "name": "keycloak-sample"
                  },
                  "organizationsEnabled": true,
                  "passwordPolicy": [
                    {
                      "type": "forceExpiredPasswordChange",
                      "value": "365"
                    },
                    {
                      "type": "length",
                      "value": "8"
                    }
                  ],
                  "realmEventConfig": {
                    "adminEventsDetailsEnabled": false,
                    "adminEventsEnabled": true,
                    "enabledEventTypes": [
                      "UPDATE_CONSENT_ERROR",
                      "CLIENT_LOGIN"
                    ],
                    "eventsEnabled": true,
                    "eventsExpiration": 15000,
                    "eventsListeners": [
                      "jboss-logging"
                    ]
                  },
                  "realmName": "d2-id-kc-realm-name",
                  "userProfileConfig": {
                    "attributes": [
                      {
                        "annotations": {
                          "inputType": "text"
                        },
                        "displayName": "Test Attribute",
                        "group": "test-group",
                        "multivalued": true,
                        "name": "test-attribute",
                        "permissions": {
                          "edit": [
                            "admin"
                          ],
                          "view": [
                            "admin",
                            "user"
                          ]
                        },
                        "required": {
                          "roles": [
                            "admin"
                          ],
                          "scopes": [
                            "profile"
                          ]
                        },
                        "selector": {
                          "scopes": [
                            "profile"
                          ]
                        },
                        "validations": {
                          "email": {
                            "max-local-length": {
                              "intVal": 64
                            }
                          },
                          "local-date": {},
                          "multivalued": {
                            "max": {
                              "stringVal": "10"
                            },
                            "min": {
                              "stringVal": "1"
                            }
                          },
                          "options": {
                            "options": {
                              "sliceVal": [
                                "option1",
                                "option2"
                              ]
                            }
                          }
                        }
                      }
                    ],
                    "groups": [
                      {
                        "annotations": {
                          "groupAnnotation": "groupAnnotation"
                        },
                        "displayDescription": "Test Group",
                        "displayHeader": "Test Group",
                        "name": "test-group"
                      }
                    ],
                    "unmanagedAttributePolicy": "ENABLED"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmComponent",
                "metadata": {
                  "name": "keycloakrealmcomponent-sample"
                },
                "spec": {
                  "config": {
                    "allowPasswordAuthentication": [
                      "true"
                    ],
                    "cachePolicy": [
                      "EVICT_WEEKLY"
                    ],
                    "debug": [
                      "true"
                    ],
                    "editMode": [
                      "READ_ONLY"
                    ],
                    "enabled": [
                      "true"
                    ],
                    "evictionDay": [
                      "3"
                    ],
                    "evictionHour": [
                      "5"
                    ],
                    "evictionMinute": [
                      "7"
                    ],
                    "kerberosRealm": [
                      "test-realm"
                    ],
                    "keyTab": [
                      "test-key-tab"
                    ],
                    "priority": [
                      "0"
                    ],
                    "serverPrincipal": [
                      "srv-principal-test"
                    ],
                    "updateProfileFirstLogin": [
                      "true"
                    ]
                  },
                  "name": "cr-kerb-test",
                  "providerId": "kerberos",
                  "providerType": "org.keycloak.storage.UserStorageProvider",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmGroup",
                "metadata": {
                  "name": "keycloakrealmgroup-sample"
                },
                "spec": {
                  "name": "ArgoCDAdmins",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "test-org-idp",
                  "namespace": "default"
                },
                "spec": {
                  "alias": "test-org-idp",
                  "config": {
                    "clientId": "test-org-client-id",
                    "clientSecret": "test-org-client-secret",
                    "kc.org.broker.redirect.mode.email-matches": "true",
                    "kc.org.domain": "example.com"
                  },
                  "enabled": true,
                  "providerId": "github",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "test-org-realm"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "keycloakrealmidentityprovider-sample"
                },
                "spec": {
                  "alias": "github",
                  "authenticateByDefault": false,
                  "config": {
                    "clientId": "foo",
                    "clientSecret": "$secretName:secretKey",
                    "syncMode": "IMPORT",
                    "useJwksUrl": "true"
                  },
                  "enabled": true,
                  "firstBrokerLoginFlowAlias": "first broker login",
                  "mappers": [
                    {
                      "config": {
                        "role": "role-tr",
                        "syncMode": "INHERIT"
                      },
                      "identityProviderAlias": "github",
                      "identityProviderMapper": "oidc-hardcoded-role-idp-mapper",
                      "name": "test3212"
                    },
                    {
                      "config": {
                        "attribute": "foo",
                        "attribute.value": "bar",
                        "syncMode": "IMPORT"
                      },
                      "identityProviderAlias": "github",
                      "identityProviderMapper": "hardcoded-attribute-idp-mapper",
                      "name": "test-33221"
                    }
                  ],
                  "providerId": "github",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRole",
                "metadata": {
                  "name": "keycloakrealmrole-sample"
                },
                "spec": {
                  "composite": true,
                  "description": "default developer role",
                  "name": "developer",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRoleBatch",
                "metadata": {
                  "name": "keycloakrealmrolebatch-sample"
                },
                "spec": {
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "roles": [
                    {
                      "composite": true,
                      "description": "default developer role",
                      "isDefault": false,
                      "name": "developer"
                    },
                    {
                      "composite": true,
                      "description": "default administrator role",
                      "isDefault": false,
                      "name": "administrator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmUser",
                "metadata": {
                  "name": "keycloakrealmuser-sample"
                },
                "spec": {
                  "attributes": {
                    "baz": "jazz",
                    "foo": "bar"
                  },
                  "email": "john.snow13@example.com",
                  "emailVerified": true,
                  "enabled": true,
                  "firstName": "John",
                  "keepResource": true,
                  "lastName": "Snow",
                  "password": "12345678",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "requiredUserActions": [
                    "UPDATE_PASSWORD"
                  ],
                  "username": "john.snow13"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloak",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloak-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloak",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloak-sample"
                },
                "spec": {
                  "secret": "keycloak-access",
                  "url": "https://keycloak.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloakRealm",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloakrealm-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloakrealm",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloakrealm-sample"
                },
                "spec": {
                  "authenticationFlows": {
                    "browserFlow": "browserFlow-sample"
                  },
                  "clusterKeycloakRef": "clusterkeycloak-sample",
                  "organizationsEnabled": false,
                  "realmName": "realm-sample"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "KeycloakOrganization",
                "metadata": {
                  "name": "test-keycloak-organization",
                  "namespace": "default"
                },
                "spec": {
                  "alias": "test-org",
                  "attributes": {
                    "department": [
                      "engineering",
                      "qa"
                    ],
                    "location": [
                      "us-east"
                    ]
                  },
                  "description": "Test organization",
                  "domains": [
                    "example.com",
                    "test.com"
                  ],
                  "identityProviders": [
                    {
                      "alias": "test-org-idp"
                    }
                  ],
                  "name": "Test Organization",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "test-org-realm"
                  },
                  "redirectUrl": "https://example.com/redirect"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          containerImage: docker.io/epamedp/keycloak-operator:1.29.0
          createdAt: "2025-09-26T12:16:55Z"
          description: An Operator for managing Keycloak
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/epam/edp-keycloak-operator
          support: KubeRocketCI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterKeycloakRealm is the Schema for the clusterkeycloakrealms
              API.
            displayName: Cluster Keycloak Realm
            kind: ClusterKeycloakRealm
            name: clusterkeycloakrealms.v1.edp.epam.com
            version: v1alpha1
          - description: ClusterKeycloak is the Schema for the clusterkeycloaks API.
            displayName: Cluster Keycloak
            kind: ClusterKeycloak
            name: clusterkeycloaks.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakAuthFlow is the Schema for the keycloak authentication
              flow API.
            displayName: Keycloak Auth Flow
            kind: KeycloakAuthFlow
            name: keycloakauthflows.v1.edp.epam.com
            version: v1
          - description: KeycloakClient is the Schema for the keycloak clients API.
            displayName: Keycloak Client
            kind: KeycloakClient
            name: keycloakclients.v1.edp.epam.com
            version: v1
          - description: KeycloakClientScope is the Schema for the keycloakclientscopes
              API.
            displayName: Keycloak Client Scope
            kind: KeycloakClientScope
            name: keycloakclientscopes.v1.edp.epam.com
            version: v1
          - description: KeycloakOrganization is the Schema for the organizations
              API.
            displayName: Keycloak Organization
            kind: KeycloakOrganization
            name: keycloakorganizations.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakRealmComponent is the Schema for the keycloak component
              API.
            displayName: Keycloak Realm Component
            kind: KeycloakRealmComponent
            name: keycloakrealmcomponents.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmGroup is the Schema for the keycloak group API.
            displayName: Keycloak Realm Group
            kind: KeycloakRealmGroup
            name: keycloakrealmgroups.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmIdentityProvider is the Schema for the keycloak
              realm identity provider API.
            displayName: Keycloak Realm Identity Provider
            kind: KeycloakRealmIdentityProvider
            name: keycloakrealmidentityproviders.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRoleBatch is the Schema for the keycloak roles
              API.
            displayName: Keycloak Realm Role Batch
            kind: KeycloakRealmRoleBatch
            name: keycloakrealmrolebatches.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRole is the Schema for the keycloak group API.
            displayName: Keycloak Realm Role
            kind: KeycloakRealmRole
            name: keycloakrealmroles.v1.edp.epam.com
            version: v1
          - description: KeycloakRealm is the Schema for the keycloak realms API.
            displayName: Keycloak Realm
            kind: KeycloakRealm
            name: keycloakrealms.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmUser is the Schema for the keycloak user API.
            displayName: Keycloak Realm User
            kind: KeycloakRealmUser
            name: keycloakrealmusers.v1.edp.epam.com
            version: v1
          - description: Keycloak is the Schema for the keycloaks API.
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.v1.edp.epam.com
            version: v1
        description: |
          Keycloak Operator is an operator that is responsible for establishing
          a connection to provided Keycloak Server, reconciling Keycloak entities (realms,
          roles, groups, users, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Keycloak `Master` realm, and assign a `create-realm` role.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  keycloak-access
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Keycloak` with Keycloak instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: Keycloak
              metadata:
                name: keycloak-sample
              spec:
                secret: keycloak-access             # Secret name
                url: https://keycloak.example.com   # Keycloak URL
              ```

              Wait for the `.status` field with  `status.connected: true`

          4. Create Keycloak realm and group using Custom Resources:

            ```yaml
            apiVersion: v1.edp.epam.com/v1
            kind: KeycloakRealm
            metadata:
              name: keycloakrealm-sample
            spec:
              realmName: realm-sample
              keycloakOwner: keycloak-sample   # the name of `kind: Keycloak`
              ```

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: KeycloakRealmGroup
              metadata:
                name: argocd-admins
              spec:
                name: ArgoCDAdmins
                realm: keycloakrealm-sample   # the name of `kind: KeycloakRealm`
              ```
        displayName: EDP Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - authentication
        - authorization
        - edp
        - idp
        - keycloak
        - kuberocketci
        - oauth
        - oidc
        - operator
        - saml
        - sso
        links:
        - name: Edp Keycloak Operator
          url: https://github.com/epam/edp-keycloak-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: KubeRocketCI
          url: https://docs.kuberocketci.io/
        relatedImages:
        - docker.io/epamedp/keycloak-operator:1.29.0
        version: 1.29.0
      entries:
      - name: edp-keycloak-operator.v1.29.0
        version: 1.29.0
      - name: edp-keycloak-operator.v1.28.0
        version: 1.28.0
      - name: edp-keycloak-operator.v1.27.1
        version: 1.27.1
      - name: edp-keycloak-operator.v1.26.0
        version: 1.26.0
      - name: edp-keycloak-operator.v1.25.0
        version: 1.25.0
      - name: edp-keycloak-operator.v1.24.0
        version: 1.24.0
      - name: edp-keycloak-operator.v1.23.0
        version: 1.23.0
      name: stable
    defaultChannel: stable
    packageName: edp-keycloak-operator
    provider:
      name: KubeRocketCI
      url: https://docs.kuberocketci.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: minhokim
      provider-url: ""
    name: node-discovery-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-discovery-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.wunicorns.com/v1alpha1",
                "kind": "NodeDiscovery",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "node-discovery-operator",
                    "app.kubernetes.io/instance": "nodediscovery-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodediscovery",
                    "app.kubernetes.io/part-of": "node-discovery-operator"
                  },
                  "name": "nodediscovery-sample"
                },
                "spec": {
                  "containers": [
                    {
                      "args": [
                        "tail",
                        "-f",
                        "/dev/null"
                      ],
                      "image": "alicek106/stress",
                      "imagePullPolicy": "IfNotPresent",
                      "name": "stress",
                      "resources": {
                        "limits": {
                          "cpu": "100m",
                          "memory": "1000Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "1000Mi"
                        }
                      },
                      "terminationGracePeriodSeconds": 10
                    }
                  ],
                  "cpus": "0",
                  "enabled": "true",
                  "mems": "0",
                  "nodeName": "worker3"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-18T01:32:51Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeDiscovery is the Schema for the nodediscoveries API
            displayName: Node Discovery
            kind: NodeDiscovery
            name: nodediscoveries.cache.wunicorns.com
            version: v1alpha1
        description: find cgroup infor and set value.
        displayName: node-discovery-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cgroup
        - cpulist
        - mems
        - cpus
        - numa
        links:
        - name: Node Discovery Operator
          url: https://node-discovery-operator.domain
        maintainers:
        - email: wunicorns1436@gmail.com
          name: minhokim
        maturity: alpha
        provider:
          name: minhokim
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/caspell/cgroup-discovery@sha256:b437dcde0107124e43a4e2fb8026ec191f44b5e761bedc2c295f47ff489cf343
        version: 0.0.1
      entries:
      - name: node-discovery-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: node-discovery-operator
    provider:
      name: minhokim
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HashiCorp
      provider-url: https://www.hashicorp.com/
    name: hcp-terraform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hcp-terraform-operator.v2.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "AgentPool",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "agentTokens": [
                    {
                      "name": "token"
                    }
                  ],
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Module",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "module": {
                    "source": "redeux/terraform-cloud-agent/kubernetes",
                    "version": "1.0.1"
                  },
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  },
                  "workspace": {
                    "id": "WORKSPACE_ID"
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Project",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Workspace",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.connect.redhat.com/hashicorp/hcp-terraform-operator@sha256:8885a6a5eae092117a670ec5541a3e4a24c6b23807fd17b0446d91b01eb6e421
          createdAt: "2025-05-28T13:58:52Z"
          description: HCP Terraform Operator for Kubernetes allows managing HCP Terraform
            / Terraform Enterprise resources via Kubernetes Custom Resources.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Valid HCP Terraform / Terraform
            Enterprise subscription is required
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hashicorp/hcp-terraform-operator
          support: HashiCorp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              AgentPool manages HCP Terraform Agent Pools, HCP Terraform Agent Tokens and can perform HCP Terraform Agent scaling.
              More infromation:
                - https://developer.hashicorp.com/terraform/cloud-docs/agents/agent-pools
                - https://developer.hashicorp.com/terraform/cloud-docs/users-teams-organizations/api-tokens#agent-api-tokens
                - https://developer.hashicorp.com/terraform/cloud-docs/agents
            displayName: Agent Pool
            kind: AgentPool
            name: agentpools.app.terraform.io
            version: v1alpha2
          - description: |-
              Module implements API-driven Run Workflows.
              More information:
                - https://developer.hashicorp.com/terraform/cloud-docs/run/api
            displayName: Module
            kind: Module
            name: modules.app.terraform.io
            version: v1alpha2
          - description: |-
              Project manages HCP Terraform Projects.
              More information:
              - https://developer.hashicorp.com/terraform/cloud-docs/projects/manage
            displayName: Project
            kind: Project
            name: projects.app.terraform.io
            version: v1alpha2
          - description: |-
              Workspace manages HCP Terraform Workspaces.
              More information:
                - https://developer.hashicorp.com/terraform/cloud-docs/workspaces
            displayName: Workspace
            kind: Workspace
            name: workspaces.app.terraform.io
            version: v1alpha2
        description: HCP Terraform Operator for Kubernetes allows managing HCP Terraform
          / Terraform Enterprise resources via Kubernetes Custom Resources.
        displayName: HCP Terraform Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - terraform
        links:
        - name: HCP Terraform Operator
          url: https://github.com/hashicorp/hcp-terraform-operator
        maintainers:
        - email: hcp-tf-operator@hashicorp.com
          name: HashiCorp
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: HashiCorp
          url: https://www.hashicorp.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.connect.redhat.com/hashicorp/hcp-terraform-operator@sha256:8885a6a5eae092117a670ec5541a3e4a24c6b23807fd17b0446d91b01eb6e421
        version: 2.9.2
      entries:
      - name: hcp-terraform-operator.v2.9.2
        version: 2.9.2
      - name: hcp-terraform-operator.v2.9.1
        version: 2.9.1
      - name: hcp-terraform-operator.v2.9.0
        version: 2.9.0
      - name: hcp-terraform-operator.v2.8.1
        version: 2.8.1
      - name: hcp-terraform-operator.v2.8.0
        version: 2.8.0
      name: alpha
    defaultChannel: alpha
    packageName: hcp-terraform-operator
    provider:
      name: HashiCorp
      url: https://www.hashicorp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Appian Corporation
      provider-url: https://appian.com/
    name: appian-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appian-operator.v0.184.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: AI/Machine Learning, Application Runtime, Big Data, Integration
            & Delivery, Modernization & Migration
          createdAt: "2025-08-07T15:50:26Z"
          description: The Appian operator facilitates managing instances of Appian
            on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appian-operator
          operators.openshift.io/valid-subscription: '["Appian License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Appian Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Appian
            name: appians.crd.k8s.appian.com
            version: v1beta1
        description: The Appian operator facilitates managing instances of Appian
          on Kubernetes.
        displayName: Appian operator
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - Appian
        links:
        - name: Appian
          url: https://appian.com/
        - name: Documentation
          url: https://docs.appian.com/suite/help/latest/k8s
        maintainers:
        - email: support@appian.com
          name: Appian Corporation
        maturity: planning
        provider:
          name: Appian Corporation
          url: https://appian.com/
        relatedImages:
        - 937620572175.dkr.ecr.us-east-1.amazonaws.com/appian/appian-operator@sha256:26c2ba27f9360226f34f289109cb0d10e9bf44777146ccfd5f367adf2367a400
        version: 0.184.0
      entries:
      - name: appian-operator.v0.184.0
        version: 0.184.0
      - name: appian-operator.v0.171.0
        version: 0.171.0
      name: alpha
    defaultChannel: alpha
    packageName: appian-operator
    provider:
      name: Appian Corporation
      url: https://appian.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: aiu-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aiu-operator.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "labels": {
                    "operator-version": "2.1.0-dev"
                  },
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "aiu": {
                    "strategy": "single"
                  },
                  "devicePlugin": {
                    "aiuSelector": {
                      "deviceID": "06a7",
                      "vendor": "1014"
                    },
                    "image": "aiu-device-plugin",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "deviceType": "vfio-pci",
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation"
                  ],
                  "featureDiscovery": {
                    "image": "aiu-feature-discovery",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "isRdma": false,
                  "loglevel": "info",
                  "metricsExporter": {
                    "enabled": true,
                    "image": "aiu-exporter",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "numVfs": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
          createdAt: "2024-10-07T13:07:51Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 2.1.1
      entries:
      - name: aiu-operator.v2.1.1
        version: 2.1.1
      name: candidate-2.1
    - currentCSV: aiu-operator.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "labels": {
                    "operator-version": "2.1.0-dev"
                  },
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "aiu": {
                    "strategy": "single"
                  },
                  "devicePlugin": {
                    "aiuSelector": {
                      "deviceID": "06a7",
                      "vendor": "1014"
                    },
                    "image": "aiu-device-plugin",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "deviceType": "vfio-pci",
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation"
                  ],
                  "featureDiscovery": {
                    "image": "aiu-feature-discovery",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "isRdma": false,
                  "loglevel": "info",
                  "metricsExporter": {
                    "enabled": true,
                    "image": "aiu-exporter",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "numVfs": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
          createdAt: "2024-10-07T13:07:51Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 2.1.1
      entries:
      - name: aiu-operator.v2.1.1
        version: 2.1.1
      name: fast-2.1
    - currentCSV: aiu-operator.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "labels": {
                    "operator-version": "2.1.0-dev"
                  },
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "aiu": {
                    "strategy": "single"
                  },
                  "devicePlugin": {
                    "aiuSelector": {
                      "deviceID": "06a7",
                      "vendor": "1014"
                    },
                    "image": "aiu-device-plugin",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "deviceType": "vfio-pci",
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation"
                  ],
                  "featureDiscovery": {
                    "image": "aiu-feature-discovery",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "isRdma": false,
                  "loglevel": "info",
                  "metricsExporter": {
                    "enabled": true,
                    "image": "aiu-exporter",
                    "imagePullPolicy": "Always",
                    "repository": "docker-na-public.artifactory.swg-devops.com/res-ai-chip-toolchain-team-docker-local/aiu",
                    "version": "2.1.1"
                  },
                  "numVfs": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
          createdAt: "2024-10-07T13:07:51Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu/release_2024_08/aiu-operator@sha256:761a27b5af685fd3d2148cf2563c12313425f956090df6c812b571472f728a3c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 2.1.1
      entries:
      - name: aiu-operator.v2.1.1
        version: 2.1.1
      name: stable-2.1
    - currentCSV: aiu-operator.v2.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "labels": {
                    "operator-version": "2.3.3"
                  },
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "aiu-device-plugin",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.1"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "featureDiscovery": {
                    "image": "aiu-feature-discovery",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "aiu-exporter",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "aiu-webhook-validator",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "scheduler": {
                    "image": "aiu-kube-scheduler-ocp-4.16",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.2"
                  }
                }
              },
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuNodeState",
                "metadata": {
                  "labels": {
                    "operator-version": "2.3.3"
                  },
                  "name": "aiunodestate"
                },
                "spec": {
                  "nodeName": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2025_03/aiu-operator@sha256:2ea8268e94d44581b3490cd4bbbd5525359073129bae42c7f007c05727afe1c3
          createdAt: "2025-04-03T01:19:38Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["aiunodestates.aiu.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu/release_2025_03/aiu-operator@sha256:2ea8268e94d44581b3490cd4bbbd5525359073129bae42c7f007c05727afe1c3
        version: 2.3.3
      entries:
      - name: aiu-operator.v2.3.3
        version: 2.3.3
      name: stable-v2.3
    - currentCSV: aiu-operator.v2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "aiuFilter": ".",
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "release_2025_05"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "aiu-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "e2e_stable",
                      "repository": "icr.io/ibmaiu_internal/x86_64/dd2",
                      "version": "release_2025_05"
                    },
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "aiu-exporter",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "aiu-webhook-validator",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "scheduler": {
                    "image": "aiu-kube-scheduler-ocp-4.16",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  }
                }
              },
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2025_05/aiu-operator@sha256:9796ab1c39d1dd55f8aa79c9f585be0f988b5026745184bdc890dfd1846909df
          createdAt: "2025-05-28T05:53:45Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["aiunodestates.aiu.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu/release_2025_05/aiu-operator@sha256:9796ab1c39d1dd55f8aa79c9f585be0f988b5026745184bdc890dfd1846909df
        - icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 2.4.0
      entries:
      - name: aiu-operator.v2.4.0
        version: 2.4.0
      name: stable-v2.4
    defaultChannel: stable-v2.4
    packageName: aiu-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lifecycle-agent
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lifecycle-agent.v4.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "ImageBasedUpgrade",
                "metadata": {
                  "name": "upgrade"
                },
                "spec": {
                  "autoRollbackOnFailure": {},
                  "extraManifests": [
                    {
                      "name": "sno-extramanifests",
                      "namespace": "openshift-lifecycle-agent"
                    }
                  ],
                  "oadpContent": [
                    {
                      "name": "oadp-cm-sno-backup",
                      "namespace": "openshift-adp"
                    }
                  ],
                  "seedImageRef": {
                    "image": "quay.io/xyz",
                    "version": "4.16.0"
                  },
                  "stage": "Idle"
                }
              },
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "SeedGenerator",
                "metadata": {
                  "name": "seedimage"
                },
                "spec": {
                  "seedImage": "quay.io/xyz"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:6c603c89401823086d7a02d7ad221da44d8c84000ba63dd30675ae0cbcacc95f
          description: The Lifecycle Agent for OpenShift provides local lifecycle
            management services for Single Node Openshift (SNO) clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.3'
          operatorframework.io/suggested-namespace: openshift-lifecycle-agent
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-lifecycle-agent",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/lifecycle-agent
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ImageBasedUpgrade is the Schema for the ImageBasedUpgrades
              API
            displayName: Image-based Cluster Upgrade
            kind: ImageBasedUpgrade
            name: imagebasedupgrades.lca.openshift.io
            version: v1
          - description: SeedGenerator is the Schema for the seedgenerators API
            displayName: Seed Generator
            kind: SeedGenerator
            name: seedgenerators.lca.openshift.io
            version: v1
        description: "# Lifecycle Agent for OpenShift\nThe Lifecycle Agent for OpenShift
          provides local lifecycle management services \nfor Single Node Openshift
          (SNO) clusters.\n\n## Where to find more information\nYou can find additional
          guidance in the [agent repository](https://github.com/openshift-kni/lifecycle-agent).\n"
        displayName: Lifecycle Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - upgrade
        links:
        - name: Lifecycle Agent
          url: https://lifecycle-agent.domain
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift4/recert-rhel9@sha256:c91def5d3da4e777d2555c354e39bf021d9f499993f86614969dee5ecc00fc49
        - registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:6c603c89401823086d7a02d7ad221da44d8c84000ba63dd30675ae0cbcacc95f
        version: 4.16.3
      entries:
      - name: lifecycle-agent.v4.16.3
        version: 4.16.3
      - name: lifecycle-agent.v4.16.2
        version: 4.16.2
      - name: lifecycle-agent.v4.16.1
        version: 4.16.1
      - name: lifecycle-agent.v4.16.0
        version: 4.16.0
      name: "4.16"
    - currentCSV: lifecycle-agent.v4.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "ImageBasedUpgrade",
                "metadata": {
                  "name": "upgrade"
                },
                "spec": {
                  "autoRollbackOnFailure": {},
                  "extraManifests": [
                    {
                      "name": "sno-extramanifests",
                      "namespace": "openshift-lifecycle-agent"
                    }
                  ],
                  "oadpContent": [
                    {
                      "name": "oadp-cm-sno-backup",
                      "namespace": "openshift-adp"
                    }
                  ],
                  "seedImageRef": {
                    "image": "quay.io/xyz",
                    "version": "4.16.0"
                  },
                  "stage": "Idle"
                }
              },
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "SeedGenerator",
                "metadata": {
                  "name": "seedimage"
                },
                "spec": {
                  "seedImage": "quay.io/xyz"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:6c603c89401823086d7a02d7ad221da44d8c84000ba63dd30675ae0cbcacc95f
          description: The Lifecycle Agent for OpenShift provides local lifecycle
            management services for Single Node Openshift (SNO) clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.3'
          operatorframework.io/suggested-namespace: openshift-lifecycle-agent
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-lifecycle-agent",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/lifecycle-agent
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ImageBasedUpgrade is the Schema for the ImageBasedUpgrades
              API
            displayName: Image-based Cluster Upgrade
            kind: ImageBasedUpgrade
            name: imagebasedupgrades.lca.openshift.io
            version: v1
          - description: SeedGenerator is the Schema for the seedgenerators API
            displayName: Seed Generator
            kind: SeedGenerator
            name: seedgenerators.lca.openshift.io
            version: v1
        description: "# Lifecycle Agent for OpenShift\nThe Lifecycle Agent for OpenShift
          provides local lifecycle management services \nfor Single Node Openshift
          (SNO) clusters.\n\n## Where to find more information\nYou can find additional
          guidance in the [agent repository](https://github.com/openshift-kni/lifecycle-agent).\n"
        displayName: Lifecycle Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - upgrade
        links:
        - name: Lifecycle Agent
          url: https://lifecycle-agent.domain
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:6c603c89401823086d7a02d7ad221da44d8c84000ba63dd30675ae0cbcacc95f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift4/recert-rhel9@sha256:c91def5d3da4e777d2555c354e39bf021d9f499993f86614969dee5ecc00fc49
        version: 4.16.3
      entries:
      - name: lifecycle-agent.v4.16.3
        version: 4.16.3
      - name: lifecycle-agent.v4.16.2
        version: 4.16.2
      - name: lifecycle-agent.v4.16.1
        version: 4.16.1
      - name: lifecycle-agent.v4.16.0
        version: 4.16.0
      name: stable
    defaultChannel: stable
    packageName: lifecycle-agent
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-opensearchservice-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-opensearchservice-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opensearchservice.services.k8s.aws/v1alpha1",
                "kind": "Domain",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/opensearchservice-controller:1.1.1
          createdAt: "2025-09-19T18:05:50Z"
          description: AWS OpenSearch Service controller is a service controller for
            managing OpenSearch Service resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Domain represents the state of an AWS opensearchservice Domain
              resource.
            displayName: Domain
            kind: Domain
            name: domains.opensearchservice.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage OpenSearch Service resources in AWS from within your Kubernetes cluster.

          **About Amazon OpenSearch Service**

          Amazon OpenSearch Service (successor to Amazon Elasticsearch Service) is a managed service that makes it easy to deploy, operate, and scale OpenSearch clusters in the AWS Cloud. Amazon OpenSearch Service supports OpenSearch and legacy Elasticsearch OSS. When you create a cluster, you have the option of which search engine to use. OpenSearch Service offers broad compatibility with Elasticsearch OSS 7.10, the final open source version of the software. For information about what changed with the recent service renaming, and the actions you might need to take, see [Amazon OpenSearch Service - Summary of changes](https://docs.aws.amazon.com/opensearch-service/latest/developerguide/rename.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon OpenSearch Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opensearchservice
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon OpenSearch Service Developer Resources
          url: https://aws.amazon.com/opensearch-service/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: opensearch service maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/opensearchservice-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-opensearchservice-controller.v1.1.1
        version: 1.1.1
      - name: ack-opensearchservice-controller.v1.1.0
        version: 1.1.0
      - name: ack-opensearchservice-controller.v1.0.15
        version: 1.0.15
      - name: ack-opensearchservice-controller.v1.0.12
        version: 1.0.12
      - name: ack-opensearchservice-controller.v1.0.10
        version: 1.0.10
      - name: ack-opensearchservice-controller.v1.0.9
        version: 1.0.9
      - name: ack-opensearchservice-controller.v1.0.8
        version: 1.0.8
      - name: ack-opensearchservice-controller.v1.0.7
        version: 1.0.7
      - name: ack-opensearchservice-controller.v1.0.6
        version: 1.0.6
      - name: ack-opensearchservice-controller.v1.0.5
        version: 1.0.5
      - name: ack-opensearchservice-controller.v1.0.4
        version: 1.0.4
      - name: ack-opensearchservice-controller.v1.0.2
        version: 1.0.2
      - name: ack-opensearchservice-controller.v1.0.1
        version: 1.0.1
      - name: ack-opensearchservice-controller.v1.0.0
        version: 1.0.0
      - name: ack-opensearchservice-controller.v0.0.28
        version: 0.0.28
      - name: ack-opensearchservice-controller.v0.0.27
        version: 0.0.27
      - name: ack-opensearchservice-controller.v0.0.26
        version: 0.0.26
      - name: ack-opensearchservice-controller.v0.0.24
        version: 0.0.24
      - name: ack-opensearchservice-controller.v0.0.23
        version: 0.0.23
      - name: ack-opensearchservice-controller.v0.0.19
        version: 0.0.19
      - name: ack-opensearchservice-controller.v0.0.17
        version: 0.0.17
      - name: ack-opensearchservice-controller.v0.0.15
        version: 0.0.15
      - name: ack-opensearchservice-controller.v0.0.14
        version: 0.0.14
      - name: ack-opensearchservice-controller.v0.0.13
        version: 0.0.13
      - name: ack-opensearchservice-controller.v0.0.12
        version: 0.0.12
      - name: ack-opensearchservice-controller.v0.0.11
        version: 0.0.11
      - name: ack-opensearchservice-controller.v0.0.10
        version: 0.0.10
      - name: ack-opensearchservice-controller.v0.0.9
        version: 0.0.9
      - name: ack-opensearchservice-controller.v0.0.8
        version: 0.0.8
      - name: ack-opensearchservice-controller.v0.0.7
        version: 0.0.7
      - name: ack-opensearchservice-controller.v0.0.6
        version: 0.0.6
      - name: ack-opensearchservice-controller.v0.0.5
        version: 0.0.5
      - name: ack-opensearchservice-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-opensearchservice-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-migration
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-migration.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Migration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {
                  "auth": {
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configuration": {
                    "port": 8080
                  },
                  "database": {
                    "image": {
                      "name": "postgresql-ubi-image",
                      "tag": "14.2.0-3747-32"
                    }
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "esinfo": {
                    "batchSize": 10000,
                    "inputIndex": "wireless_stats"
                  },
                  "image": {
                    "remoteRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "systemMonitor": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:36:21Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Migration
            name: migrations.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Migration Operator
        displayName: Keysight WAP Migration Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Migration
          url: https://keysight-wap-migration.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-migration@sha256:118b466f1c59832a3620fa9b153507cc66ba38ee08b337b3a1a4225b158bebc5
        version: 0.1.21
      entries:
      - name: keysight-wap-migration.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-migration
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-xtradb-cluster-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator-certified.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator-certified.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator-certified.v1.17.0
        version: 1.17.0
      - name: percona-xtradb-cluster-operator-certified.v1.16.1
        version: 1.16.1
      - name: percona-xtradb-cluster-operator-certified.v1.16.0
        version: 1.16.0
      - name: percona-xtradb-cluster-operator-certified.v1.15.1
        version: 1.15.1
      - name: percona-xtradb-cluster-operator-certified.v1.15.0
        version: 1.15.0
      name: stable
    - currentCSV: percona-xtradb-cluster-operator-certified.v1.18.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        version: 1.18.0-cw
      entries:
      - name: percona-xtradb-cluster-operator-certified.v1.18.0-cw
        version: 1.18.0-cw
      - name: percona-xtradb-cluster-operator-certified.v1.17.0-cw
        version: 1.17.0-cw
      - name: percona-xtradb-cluster-operator-certified.v1.16.1-cw
        version: 1.16.1-cw
      - name: percona-xtradb-cluster-operator-certified.v1.16.0-cw
        version: 1.16.0-cw
      - name: percona-xtradb-cluster-operator-certified.v1.15.0-cw
        version: 1.15.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator-certified
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-storage-odf-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-storage-odf-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:962ca15393dd2c48f8a7189c1801b43e9fd1cd073a1971286d485f4b96060426
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.6.0'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:962ca15393dd2c48f8a7189c1801b43e9fd1cd073a1971286d485f4b96060426
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:2f785096aaa63b042e415aa0b64ace2a085cb7c219853f089d803bc37d8295ce
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        version: 1.6.0
      entries:
      - name: ibm-storage-odf-operator.v1.6.0
        version: 1.6.0
      name: stable-v1.6
    - currentCSV: ibm-storage-odf-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:2dd8608645b13f81330466e25664b132d46af480a3ab2b0dd48fb1feba38870a
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.7.1'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:2dd8608645b13f81330466e25664b132d46af480a3ab2b0dd48fb1feba38870a
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:24eae9d1502ce98bb024d3eba154ddec07fdac0d78ecad0c3a64b55f5b3600f7
        version: 1.7.1
      entries:
      - name: ibm-storage-odf-operator.v1.7.1
        version: 1.7.1
      name: stable-v1.7
    - currentCSV: ibm-storage-odf-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:7432340ac156e620240eea73a19321a871fa89cc513c340a91d97d3f3fd4ded2
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.8.0'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:7432340ac156e620240eea73a19321a871fa89cc513c340a91d97d3f3fd4ded2
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:3344c7e987674fb98173bbc6ae1b189613e3ccefb468f1af9866734ddbf0408f
        version: 1.8.0
      entries:
      - name: ibm-storage-odf-operator.v1.8.0
        version: 1.8.0
      name: stable-v1.8
    defaultChannel: stable-v1.8
    packageName: ibm-storage-odf-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-ingress
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-ingress.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Ingress",
                "metadata": {
                  "name": "ingress-sample"
                },
                "spec": {
                  "admissionWebhooks": {
                    "patch": {
                      "image": {
                        "repository": "docker.io/jettech/kube-webhook-certgen"
                      }
                    }
                  },
                  "configMapMerger": {
                    "registry": "prod.harbor.keysight.digital/loadcore",
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "environment": "openshift",
                  "ingress-nginx": {
                    "controller": {
                      "config": {
                        "global-auth-signin": "https://$host/eula/static/",
                        "global-auth-url": "http://kcos-framework-v1-kcos-eula.kcos-framework.svc.cluster.local/v1/check"
                      },
                      "daemonset": {
                        "useHostPort": false
                      },
                      "hostNetwork": false,
                      "image": {
                        "registry": "prod.harbor.keysight.digital/loadcore"
                      },
                      "ingressClassResource": {
                        "default": true
                      },
                      "nodeSelector": "",
                      "service": {
                        "nodePorts": {
                          "http": 80,
                          "https": 443
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "defaultBackend": {
                      "image": {
                        "repository": "k8s.gcr.io/defaultbackend-amd64"
                      }
                    }
                  },
                  "jobDelWebhook": {
                    "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:28:18Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ingress
            name: ingresses.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Ingress Operator
        displayName: Keysight KCOS Ingress Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Ingress
          url: https://keysight-kcos-ingress.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-ingress@sha256:c6f6f8b44276efd8f5ffaa099b393e7e97d9cfa627952c0d049b98d92dc3ab59
        version: 0.1.21
      entries:
      - name: keysight-kcos-ingress.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-ingress
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sns-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sns-controller.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "PlatformApplication",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "PlatformEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sns-controller:1.2.2
          createdAt: "2025-09-30T21:19:34Z"
          description: AWS SNS controller is a service controller for managing SNS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PlatformApplication represents the state of an AWS sns PlatformApplication
              resource.
            displayName: PlatformApplication
            kind: PlatformApplication
            name: platformapplications.sns.services.k8s.aws
            version: v1alpha1
          - description: PlatformEndpoint represents the state of an AWS sns PlatformEndpoint
              resource.
            displayName: PlatformEndpoint
            kind: PlatformEndpoint
            name: platformendpoints.sns.services.k8s.aws
            version: v1alpha1
          - description: Subscription represents the state of an AWS sns Subscription
              resource.
            displayName: Subscription
            kind: Subscription
            name: subscriptions.sns.services.k8s.aws
            version: v1alpha1
          - description: Topic represents the state of an AWS sns Topic resource.
            displayName: Topic
            kind: Topic
            name: topics.sns.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Simple Notification Service (SNS) resources in AWS from within your Kubernetes cluster.

          **About Amazon SNS**

          Amazon Simple Notification Service (Amazon SNS) is a managed service that provides message delivery from publishers to subscribers (also known as producers and consumers). Publishers communicate asynchronously with subscribers by sending messages to a topic, which is a logical access point and communication channel. Clients can subscribe to the SNS topic and receive published messages using a supported endpoint type, such as Amazon Kinesis Data Firehose, Amazon SQS, AWS Lambda, HTTP, email, mobile push notifications, and mobile text messages (SMS).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SNS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sns
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SNS Developer Resources
          url: https://aws.amazon.com/sns/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sns maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sns-controller:1.2.2
        version: 1.2.2
      entries:
      - name: ack-sns-controller.v1.2.2
        version: 1.2.2
      - name: ack-sns-controller.v1.2.1
        version: 1.2.1
      - name: ack-sns-controller.v1.2.0
        version: 1.2.0
      - name: ack-sns-controller.v1.1.16
        version: 1.1.16
      - name: ack-sns-controller.v1.1.15
        version: 1.1.15
      - name: ack-sns-controller.v1.1.14
        version: 1.1.14
      - name: ack-sns-controller.v1.1.13
        version: 1.1.13
      - name: ack-sns-controller.v1.1.12
        version: 1.1.12
      - name: ack-sns-controller.v1.1.11
        version: 1.1.11
      - name: ack-sns-controller.v1.1.10
        version: 1.1.10
      - name: ack-sns-controller.v1.1.9
        version: 1.1.9
      - name: ack-sns-controller.v1.1.8
        version: 1.1.8
      - name: ack-sns-controller.v1.1.7
        version: 1.1.7
      - name: ack-sns-controller.v1.1.6
        version: 1.1.6
      - name: ack-sns-controller.v1.1.5
        version: 1.1.5
      - name: ack-sns-controller.v1.1.4
        version: 1.1.4
      - name: ack-sns-controller.v1.1.3
        version: 1.1.3
      - name: ack-sns-controller.v1.1.2
        version: 1.1.2
      - name: ack-sns-controller.v1.1.1
        version: 1.1.1
      - name: ack-sns-controller.v1.1.0
        version: 1.1.0
      - name: ack-sns-controller.v1.0.14
        version: 1.0.14
      - name: ack-sns-controller.v1.0.13
        version: 1.0.13
      - name: ack-sns-controller.v1.0.12
        version: 1.0.12
      - name: ack-sns-controller.v1.0.11
        version: 1.0.11
      - name: ack-sns-controller.v1.0.9
        version: 1.0.9
      - name: ack-sns-controller.v1.0.6
        version: 1.0.6
      - name: ack-sns-controller.v1.0.5
        version: 1.0.5
      - name: ack-sns-controller.v1.0.4
        version: 1.0.4
      - name: ack-sns-controller.v1.0.3
        version: 1.0.3
      - name: ack-sns-controller.v1.0.2
        version: 1.0.2
      - name: ack-sns-controller.v1.0.1
        version: 1.0.1
      - name: ack-sns-controller.v1.0.0
        version: 1.0.0
      - name: ack-sns-controller.v0.0.12
        version: 0.0.12
      - name: ack-sns-controller.v0.0.11
        version: 0.0.11
      - name: ack-sns-controller.v0.0.10
        version: 0.0.10
      - name: ack-sns-controller.v0.0.9
        version: 0.0.9
      - name: ack-sns-controller.v0.0.8
        version: 0.0.8
      - name: ack-sns-controller.v0.0.7
        version: 0.0.7
      - name: ack-sns-controller.v0.0.5
        version: 0.0.5
      - name: ack-sns-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-sns-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Helm Community
      provider-url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
    name: cockroachdb
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroachdb.v5.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1alpha1",
                "kind": "Cockroachdb",
                "metadata": {
                  "name": "cockroachdb-sample"
                },
                "spec": {
                  "clusterDomain": "cluster.local",
                  "conf": {
                    "attrs": [],
                    "cache": "25%",
                    "cluster-name": "",
                    "disable-cluster-name-verification": false,
                    "http-port": 8080,
                    "join": [],
                    "locality": "",
                    "logtostderr": "INFO",
                    "max-sql-memory": "25%",
                    "port": 26257,
                    "single-node": false,
                    "sql-audit-dir": ""
                  },
                  "image": {
                    "credentials": {},
                    "pullPolicy": "IfNotPresent",
                    "repository": "cockroachdb/cockroach",
                    "tag": "v20.2.4"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [],
                    "labels": {},
                    "paths": [
                      "/"
                    ],
                    "tls": []
                  },
                  "init": {
                    "affinity": {},
                    "annotations": {},
                    "labels": {
                      "app.kubernetes.io/component": "init"
                    },
                    "nodeSelector": {},
                    "resources": {},
                    "tolerations": []
                  },
                  "labels": {},
                  "networkPolicy": {
                    "enabled": false,
                    "ingress": {
                      "grpc": [],
                      "http": []
                    }
                  },
                  "service": {
                    "discovery": {
                      "annotations": {},
                      "labels": {
                        "app.kubernetes.io/component": "cockroachdb"
                      }
                    },
                    "ports": {
                      "grpc": {
                        "external": {
                          "name": "grpc",
                          "port": 26257
                        },
                        "internal": {
                          "name": "grpc-internal",
                          "port": 26257
                        }
                      },
                      "http": {
                        "name": "http",
                        "port": 8080
                      }
                    },
                    "public": {
                      "annotations": {},
                      "labels": {
                        "app.kubernetes.io/component": "cockroachdb"
                      },
                      "type": "ClusterIP"
                    }
                  },
                  "serviceMonitor": {
                    "annotations": {},
                    "enabled": false,
                    "interval": "10s",
                    "labels": {}
                  },
                  "statefulset": {
                    "annotations": {},
                    "args": [],
                    "budget": {
                      "maxUnavailable": 1
                    },
                    "env": [],
                    "labels": {
                      "app.kubernetes.io/component": "cockroachdb"
                    },
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "podAffinity": {},
                    "podAntiAffinity": {
                      "topologyKey": "kubernetes.io/hostname",
                      "type": "soft",
                      "weight": 100
                    },
                    "podManagementPolicy": "Parallel",
                    "priorityClassName": "",
                    "replicas": 3,
                    "resources": {},
                    "secretMounts": [],
                    "tolerations": [],
                    "topologySpreadConstraints": {
                      "maxSkew": 1,
                      "topologyKey": "topology.kubernetes.io/zone",
                      "whenUnsatisfiable": "ScheduleAnyway"
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    }
                  },
                  "storage": {
                    "hostPath": "",
                    "persistentVolume": {
                      "annotations": {},
                      "enabled": true,
                      "labels": {},
                      "size": "100Gi",
                      "storageClass": ""
                    }
                  },
                  "tls": {
                    "certs": {
                      "clientRootSecret": "cockroachdb-root",
                      "nodeSecret": "cockroachdb-node",
                      "provided": false,
                      "tlsSecret": false
                    },
                    "enabled": false,
                    "init": {
                      "image": {
                        "credentials": {},
                        "pullPolicy": "IfNotPresent",
                        "repository": "cockroachdb/cockroach-k8s-request-cert",
                        "tag": "0.4"
                      }
                    },
                    "serviceAccount": {
                      "create": true,
                      "name": ""
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          certified: "false"
          containerImage: quay.io/helmoperators/cockroachdb:v5.0.4
          createdAt: 2021-01-23T10-00-00Z
          description: CockroachDB Operator based on the CockroachDB helm chart
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/dmesser/cockroachdb-operator
          support: dmesser
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a CockroachDB cluster
            displayName: CockroachDB
            kind: Cockroachdb
            name: cockroachdbs.charts.operatorhub.io
            version: v1alpha1
        description: |
          CockroachDB is a scalable, survivable, strongly-consistent SQL database.
          ## About this Operator
          This Operator is based on a Helm chart for CockroachDB. It supports the same parameters like the official [CockroachDB Helm Chart](https://github.com/cockroachdb/helm-charts/tree/master/cockroachdb).

          ## Core capabilities
          * **StatefulSet** - Sets up a dynamically scalable CockroachDB cluster using a Kubernetes StatefulSet
          * **Expand Replicas** - Supports expanding the set of replicas by simply editing your object
          * **Dashboard** - Installs the CockroachDB user interface to administer your cluster. Easily expose it via an Ingress rule.
          Review all of the [confiuguration options](https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration) to best run your database instance. The example configuration is derived from the chart's [`values.yaml`](https://github.com/helm/charts/blob/master/stable/cockroachdb/values.yaml).
          ## Using the cluster
          The resulting cluster endpoint can be consumed from a `Service` that follows the pattern: `<StatefulSet-name>-public`. For example to connect using the command line client, use something like the following to obtain the name of the service:
          ```
          kubectl get service -l chart=cockroachdb-5.0.4
          NAME                                           TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)              AGE
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb          ClusterIP   None             <none>        26257/TCP,8080/TCP   24m
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public   ClusterIP   10.106.249.134   <none>        26257/TCP,8080/TCP   24m
          ```
          Then you can use the CockroachDB command line client to connect to the database cluster:
          ```
          kubectl run -it --rm cockroach-client --image=cockroachdb/cockroach --restart=Never --command -- ./cockroach sql --insecure --host example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public
          ```
          ## Before you start
          This Operator requires a cluster with PV support in order to run correctly.

          Note that for a production cluster, you will likely want to override the following parameters in [`values.yaml`](values.yaml) with your own values.

          - `statefulset.resources.requests.memory` and `statefulset.resources.limits.memory` allocate memory resources to CockroachDB pods in your cluster.
          - `conf.cache` and `conf.max-sql-memory` are memory limits that we recommend setting to 1/4 of the above resource allocation. When running CockroachDB, you must set these limits explicitly to avoid running out of memory.
          - `storage.persistentVolume.size` defaults to `100Gi` of disk space per pod, which you may increase or decrease for your use case.
          - `storage.persistentVolume.storageClass` uses the default storage class for your environment. We strongly recommend that you specify a storage class which uses an SSD.
          - `tls.enabled` must be set to `yes`/`true` to deploy in secure mode.

          For more information on overriding the `values.yaml` parameters, please see:
          > <https://www.cockroachlabs.com/docs/stable/orchestrate-cockroachdb-with-kubernetes.html#step-2-start-cockroachdb>

          If you are running in secure mode (with configuration parameter `tls.enabled` set to `yes`/`true`) and `tls.certs.provided` set to `no`/`false`), then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          When `tls.certs.provided` is set to `yes`/`true`, this chart will use certificates created outside of Kubernetes. You may want to use this if you want to use a different certificate authority from the one being used by Kubernetes or if your Kubernetes cluster doesn't fully support certificate-signing requests. To use this, first set up your certificates and load them into your Kubernetes cluster as Secrets using the commands below:

          ```
          mkdir certs
          mkdir my-safe-directory
          cockroach cert create-ca --certs-dir=certs --ca-key=my-safe-directory/ca.key
          cockroach cert create-client root --certs-dir=certs --ca-key=my-safe-directory/ca.key
          kubectl create secret generic cockroachdb-root --from-file=certs
          cockroach cert create-node --certs-dir=certs --ca-key=my-safe-directory/ca.key localhost 127.0.0.1 eerie-horse-cockroachdb-public eerie-horse-cockroachdb-public.default eerie-horse-cockroachdb-public.default.svc.cluster.local *.eerie-horse-cockroachdb *.eerie-horse-cockroachdb.default *.eerie-horse-cockroachdb.default.svc.cluster.local
          kubectl create secret generic cockroachdb-node --from-file=certs
          ```

          Set `tls.certs.tlsSecret` to `yes/true` if you make use of [cert-manager][3] in your cluster.

          [cert-manager][3] stores generated certificates in dedicated TLS secrets. Thus, they are always named:

          * `ca.crt`
          * `tls.crt`
          * `tls.key`

          On the other hand, CockroachDB also demands dedicated certificate filenames:

          * `ca.crt`
          * `node.crt`
          * `node.key`
          * `client.root.crt`
          * `client.root.key`

          By activating `tls.certs.tlsSecret` we benefit from projected secrets and convert the TLS secret filenames to their according CockroachDB filenames.

          If you are running in secure mode, then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          Confirm that all pods are `Running` successfully and init has been completed:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS      RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running     0          1m
          my-release-cockroachdb-1            1/1       Running     0          1m
          my-release-cockroachdb-2            1/1       Running     0          1m
          my-release-cockroachdb-init-k6jcr   0/1       Completed   0          1m
          ```

          Confirm that persistent volumes are created and claimed for each pod:

          ```shell
          kubectl get pv
          ```

          ```
          NAME                                       CAPACITY   ACCESS MODES   RECLAIM POLICY   STATUS    CLAIM                                      STORAGECLASS   REASON    AGE
          pvc-64878ebf-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-0   standard                 51s
          pvc-64945b4f-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-1   standard                 51s
          pvc-649d920d-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-2   standard                 51s
          ```

          ## Upgrading the cluster

          ### Chart version 3.0.0 and after

          Launch a temporary interactive pod and start the built-in SQL client:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          Set `cluster.preserve_downgrade_option`, where `$current_version` is the CockroachDB version currently running (e.g., `19.2`):

          ```sql
          > SET CLUSTER SETTING cluster.preserve_downgrade_option = '$current_version';
          ```

          Exit the shell and delete the temporary pod:

          ```sql
          > \q
          ```

          Kick off the upgrade process by changing the new Docker image, where `$new_version` is the CockroachDB version to which you are upgrading:

          ```shell
          kubectl delete job my-release-cockroachdb-init
          ```

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"image":{"tag":"<new version tag here>"}}}' --type merge
          ```

          Kubernetes will carry out a safe [rolling upgrade](https://kubernetes.io/docs/tutorials/stateful-application/basic-stateful-set/#updating-statefulsets) of your CockroachDB nodes one-by-one. Monitor the cluster's pods until all have been successfully restarted:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS              RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running             0          2m
          my-release-cockroachdb-1            1/1       Running             0          3m
          my-release-cockroachdb-2            1/1       Running             0          3m
          my-release-cockroachdb-3            0/1       ContainerCreating   0          25s
          my-release-cockroachdb-init-nwjkh   0/1       ContainerCreating   0          6s
          ```

          ```shell
          kubectl get pods \
          -o jsonpath='{range .items[*]}{.metadata.name}{"\t"}{.spec.containers[0].image}{"\n"}'
          ```

          ```
          my-release-cockroachdb-0    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-1    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-2    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-3    cockroachdb/cockroach:v20.2.4
          ```

          Resume normal operations. Once you are comfortable that the stability and performance of the cluster is what you'd expect post-upgrade, finalize the upgrade:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          ```sql
          > RESET CLUSTER SETTING cluster.preserve_downgrade_option;
          > \q
          ```

          ## Configuration

          The following table lists the configurable parameters of the CockroachDB chart and their default values.
          For details see the [`values.yaml`](values.yaml) file.

          | Parameter                                                 | Description                                                     | Default                                          |
          | ---------                                                 | -----------                                                     | -------                                          |
          | `clusterDomain`                                           | Cluster's default DNS domain                                    | `cluster.local`                                  |
          | `conf.attrs`                                              | CockroachDB node attributes                                     | `[]`                                             |
          | `conf.cache`                                              | Size of CockroachDB's in-memory cache                           | `25%`                                            |
          | `conf.cluster-name`                                       | Name of CockroachDB cluster                                     | `""`                                             |
          | `conf.disable-cluster-name-verification`                  | Disable CockroachDB cluster name verification                   | `no`                                             |
          | `conf.join`                                               | List of already-existing CockroachDB instances                  | `[]`                                             |
          | `conf.max-disk-temp-storage`                              | Max storage capacity for temp data                              | `0`                                              |
          | `conf.max-offset`                                         | Max allowed clock offset for CockroachDB cluster                | `500ms`                                          |
          | `conf.max-sql-memory`                                     | Max memory to use processing SQL querie                         | `25%`                                            |
          | `conf.locality`                                           | Locality attribute for this deployment                          | `""`                                             |
          | `conf.single-node`                                        | Disable CockroachDB clustering (standalone mode)                | `no`                                             |
          | `conf.sql-audit-dir`                                      | Directory for SQL audit log                                     | `""`                                             |
          | `conf.port`                                               | CockroachDB primary serving port in Pods                        | `26257`                                          |
          | `conf.http-port`                                          | CockroachDB HTTP port in Pods                                   | `8080`                                           |
          | `image.repository`                                        | Container image name                                            | `cockroachdb/cockroach`                          |
          | `image.tag`                                               | Container image tag                                             | `v20.2.4`                                        |
          | `image.pullPolicy`                                        | Container pull policy                                           | `IfNotPresent`                                   |
          | `image.credentials`                                       | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `statefulset.replicas`                                    | StatefulSet replicas number                                     | `3`                                              |
          | `statefulset.updateStrategy`                              | Update strategy for StatefulSet Pods                            | `{"type": "RollingUpdate"}`                      |
          | `statefulset.podManagementPolicy`                         | `OrderedReady`/`Parallel` Pods creation/deletion order          | `Parallel`                                       |
          | `statefulset.budget.maxUnavailable`                       | k8s PodDisruptionBudget parameter                               | `1`                                              |
          | `statefulset.args`                                        | Extra command-line arguments                                    | `[]`                                             |
          | `statefulset.env`                                         | Extra env vars                                                  | `[]`                                             |
          | `statefulset.secretMounts`                                | Additional Secrets to mount at cluster members                  | `[]`                                             |
          | `statefulset.labels`                                      | Additional labels of StatefulSet and its Pods                   | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `statefulset.annotations`                                 | Additional annotations of StatefulSet Pods                      | `{}`                                             |
          | `statefulset.nodeAffinity`                                | [Node affinity rules][2] of StatefulSet Pods                    | `{}`                                             |
          | `statefulset.podAffinity`                                 | [Inter-Pod affinity rules][1] of StatefulSet Pods               | `{}`                                             |
          | `statefulset.podAntiAffinity`                             | [Anti-affinity rules][1] of StatefulSet Pods                    | auto                                             |
          | `statefulset.podAntiAffinity.topologyKey`                 | The topologyKey for auto [anti-affinity rules][1]               | `kubernetes.io/hostname`                         |
          | `statefulset.podAntiAffinity.type`                        | Type of auto [anti-affinity rules][1]                           | `soft`                                           |
          | `statefulset.podAntiAffinity.weight`                      | Weight for `soft` auto [anti-affinity rules][1]                 | `100`                                            |
          | `statefulset.nodeSelector`                                | Node labels for StatefulSet Pods assignment                     | `{}`                                             |
          | `statefulset.priorityClassName`                           | [PriorityClassName][4] for StatefulSet Pods                     | `""`                                             |
          | `statefulset.tolerations`                                 | Node taints to tolerate by StatefulSet Pods                     | `[]`                                             |
          | `statefulset.topologySpreadConstraints`                   | [Topology Spread Constraints rules][5] of StatefulSet Pods      | auto                                             |
          | `statefulset.topologySpreadConstraints.maxSkew`           | Degree to which Pods may be unevenly distributed                | `1`                                              |
          | `statefulset.topologySpreadConstraints.topologyKey`       | The key of node labels                                          | `topology.kubernetes.io/zone`                    |
          | `statefulset.topologySpreadConstraints.whenUnsatisfiable` | `ScheduleAnyway`/`DoNotSchedule` for unsatisfiable constraints  | `ScheduleAnyway`                                 |
          | `statefulset.resources`                                   | Resource requests and limits for StatefulSet Pods               | `{}`                                             |
          | `service.ports.grpc.external.port`                        | CockroachDB primary serving port in Services                    | `26257`                                          |
          | `service.ports.grpc.external.name`                        | CockroachDB primary serving port name in Services               | `grpc`                                           |
          | `service.ports.grpc.internal.port`                        | CockroachDB inter-communication port in Services                | `26257`                                          |
          | `service.ports.grpc.internal.name`                        | CockroachDB inter-communication port name in Services           | `grpc-internal`                                  |
          | `service.ports.http.port`                                 | CockroachDB HTTP port in Services                               | `8080`                                           |
          | `service.ports.http.name`                                 | CockroachDB HTTP port name in Services                          | `http`                                           |
          | `service.public.type`                                     | Public Service type                                             | `ClusterIP`                                      |
          | `service.public.labels`                                   | Additional labels of public Service                             | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.public.annotations`                              | Additional annotations of public Service                        | `{}`                                             |
          | `service.discovery.labels`                                | Additional labels of discovery Service                          | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.discovery.annotations`                           | Additional annotations of discovery Service                     | `{}`                                             |
          | `ingress.enabled`                                         | Enable ingress resource for CockroachDB                         | `false`                                          |
          | `ingress.labels`                                          | Additional labels of Ingress                                    | `{}`                                             |
          | `ingress.annotations`                                     | Additional annotations of Ingress                               | `{}`                                             |
          | `ingress.paths`                                           | Paths for the default host                                      | `[/]`                                            |
          | `ingress.hosts`                                           | CockroachDB Ingress hostnames                                   | `[]`                                             |
          | `ingress.tls[0].hosts`                                    | CockroachDB Ingress tls hostnames                               | `nil`                                            |
          | `ingress.tls[0].secretName`                               | CockroachDB Ingress tls secret name                             | `nil`                                            |
          | `serviceMonitor.enabled`                                  | Create [ServiceMonitor](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/design.md#servicemonitor) Resource for scraping metrics using [PrometheusOperator](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/user-guides/getting-started.md#prometheus-operator)                     | `false`                                             |
          | `serviceMonitor.labels`                                   | Additional labels of ServiceMonitor                             | `{}`                                             |
          | `serviceMonitor.annotations`                              | Additional annotations of ServiceMonitor                        | `{}`                                             |
          | `serviceMonitor.interval`                                 | ServiceMonitor scrape metrics interval                          | `10s`                                            |
          | `serviceMonitor.scrapeTimeout`                            | ServiceMonitor scrape timeout                                   | `nil`                                            |
          | `storage.hostPath`                                        | Absolute path on host to store data                             | `""`                                             |
          | `storage.persistentVolume.enabled`                        | Whether to use PersistentVolume to store data                   | `yes`                                            |
          | `storage.persistentVolume.size`                           | PersistentVolume size                                           | `100Gi`                                          |
          | `storage.persistentVolume.storageClass`                   | PersistentVolume class                                          | `""`                                             |
          | `storage.persistentVolume.labels`                         | Additional labels of PersistentVolumeClaim                      | `{}`                                             |
          | `storage.persistentVolume.annotations`                    | Additional annotations of PersistentVolumeClaim                 | `{}`                                             |
          | `init.labels`                                             | Additional labels of init Job and its Pod                       | `{"app.kubernetes.io/component": "init"}`        |
          | `init.annotations`                                        | Additional labels of the Pod of init Job                        | `{}`                                             |
          | `init.affinity`                                           | [Affinity rules][2] of init Job Pod                             | `{}`                                             |
          | `init.nodeSelector`                                       | Node labels for init Job Pod assignment                         | `{}`                                             |
          | `init.tolerations`                                        | Node taints to tolerate by init Job Pod                         | `[]`                                             |
          | `init.resources`                                          | Resource requests and limits for the Pod of init Job            | `{}`                                             |
          | `tls.enabled`                                             | Whether to run securely using TLS certificates                  | `no`                                             |
          | `tls.serviceAccount.create`                               | Whether to create a new RBAC service account                    | `yes`                                            |
          | `tls.serviceAccount.name`                                 | Name of RBAC service account to use                             | `""`                                             |
          | `tls.certs.provided`                                      | Bring your own certs scenario, i.e certificates are provided    | `no`                                             |
          | `tls.certs.clientRootSecret`                              | If certs are provided, secret name for client root cert         | `cockroachdb-root`                               |
          | `tls.certs.nodeSecret`                                    | If certs are provided, secret name for node cert                | `cockroachdb-node`                               |
          | `tls.certs.tlsSecret`                                     | Own certs are stored in TLS secret                              | `no`                                             |
          | `tls.init.image.repository`                               | Image to use for requesting TLS certificates                    | `cockroachdb/cockroach-k8s-request-cert`         |
          | `tls.init.image.tag`                                      | Image tag to use for requesting TLS certificates                | `0.4`                                            |
          | `tls.init.image.pullPolicy`                               | Requesting TLS certificates container pull policy               | `IfNotPresent`                                   |
          | `tls.init.image.credentials`                              | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `networkPolicy.enabled`                                   | Enable NetworkPolicy for CockroachDB's Pods                     | `no`                                             |
          | `networkPolicy.ingress.grpc`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |
          | `networkPolicy.ingress.http`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |


          Override the default parameters using the `CockroachDB` custom resource.

          ## Deep dive

          ### Connecting to the CockroachDB cluster

          Once you've created the cluster, you can start talking to it by connecting to its `-public` Service. CockroachDB is PostgreSQL wire protocol compatible, so there's a [wide variety of supported clients](https://www.cockroachlabs.com/docs/install-client-drivers.html). As an example, we'll open up a SQL shell using CockroachDB's built-in shell and play around with it a bit, like this (likely needing to replace `my-release-cockroachdb-public` with the name of the `-public` Service that was created with your installed chart):

          ```shell
          kubectl run cockroach-client --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host my-release-cockroachdb-public
          ```

          ```
          Waiting for pod default/cockroach-client to be running, status is Pending,
          pod ready: false
          If you don't see a command prompt, try pressing enter.
          root@my-release-cockroachdb-public:26257> SHOW DATABASES;
          +--------------------+
          |      Database      |
          +--------------------+
          | information_schema |
          | pg_catalog         |
          | system             |
          +--------------------+
          (3 rows)
          root@my-release-cockroachdb-public:26257> CREATE DATABASE bank;
          CREATE DATABASE
          root@my-release-cockroachdb-public:26257> CREATE TABLE bank.accounts (id INT
          PRIMARY KEY, balance DECIMAL);
          CREATE TABLE
          root@my-release-cockroachdb-public:26257> INSERT INTO bank.accounts VALUES
          (1234, 10000.50);
          INSERT 1
          root@my-release-cockroachdb-public:26257> SELECT * FROM bank.accounts;
          +------+---------+
          |  id  | balance |
          +------+---------+
          | 1234 | 10000.5 |
          +------+---------+
          (1 row)
          root@my-release-cockroachdb-public:26257> \q
          Waiting for pod default/cockroach-client to terminate, status is Running
          pod "cockroach-client" deleted
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          ### Cluster health

          Because our pod spec includes regular health checks of the CockroachDB processes, simply running `kubectl get pods` and looking at the `STATUS` column is sufficient to determine the health of each instance in the cluster.

          If you want more detailed information about the cluster, the best place to look is the Admin UI.

          ### Accessing the Admin UI

          If you want to see information about how the cluster is doing, you can try pulling up the CockroachDB Admin UI by port-forwarding from your local machine to one of the pods (replacing `my-release-cockroachdb-0` with the name of one of your pods:

          ```shell
          kubectl port-forward my-release-cockroachdb-0 8080
          ```

          You should then be able to access the Admin UI by visiting <http://localhost:8080/> in your web browser.

          ### Failover

          If any CockroachDB member fails, it is restarted or recreated automatically by the Kubernetes infrastructure, and will re-join the cluster automatically when it comes back up. You can test this scenario by killing any of the CockroachDB pods:

          ```shell
          kubectl delete pod my-release-cockroachdb-1
          ```

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          After a while:

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-1  1/1       Running       0          20s
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          You can check the state of re-joining from the new pod's logs:

          ```shell
          kubectl logs my-release-cockroachdb-1
          ```

          ```
          [...]
          I161028 19:32:09.754026 1 server/node.go:586  [n1] node connected via gossip and
          verified as part of cluster {"35ecbc27-3f67-4e7d-9b8f-27c31aae17d6"}
          [...]
          cockroachdb-0.my-release-cockroachdb.default.svc.cluster.local:26257
          build:      beta-20161027-55-gd2d3c7f @ 2016/10/28 19:27:25 (go1.7.3)
          admin:      http://0.0.0.0:8080
          sql:
          postgresql://root@my-release-cockroachdb-1.my-release-cockroachdb.default.svc.cluster.local:26257?sslmode=disable
          logs:       cockroach-data/logs
          store[0]:   path=cockroach-data
          status:     restarted pre-existing node
          clusterID:  {35ecbc27-3f67-4e7d-9b8f-27c31aae17d6}
          nodeID:     2
          [...]
          ```

          ### NetworkPolicy

          To enable NetworkPolicy for CockroachDB, install [a networking plugin that implements the Kubernetes NetworkPolicy spec](https://kubernetes.io/docs/tasks/administer-cluster/declare-network-policy#before-you-begin), and set `networkPolicy.enabled` to `yes`/`true`.

          For Kubernetes v1.5 & v1.6, you must also turn on NetworkPolicy by setting the `DefaultDeny` Namespace annotation. Note: this will enforce policy for _all_ pods in the Namespace:

          ```shell
          kubectl annotate namespace default "net.beta.kubernetes.io/network-policy={\"ingress\":{\"isolation\":\"DefaultDeny\"}}"
          ```

          For more precise policy, set `networkPolicy.ingress.grpc` and `networkPolicy.ingress.http` rules. This will only allow pods that match the provided rules to connect to CockroachDB.

          ### Scaling

          Scaling should be managed via the `helm upgrade` command. After resizing your cluster on your cloud environment (e.g., GKE or EKS), run the following command to add a pod. This assumes you scaled from 3 to 4 nodes:

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"statefulset":{"replicas":"4"}}}' --type merge
          ```

          Note, that if you are running in secure mode (`tls.enabled` is `yes`/`true`) and increase the size of your cluster, you will also have to approve the CSR (certificate-signing request) of each new node (using `kubectl get csr` and `kubectl certificate approve`).

          [1]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [2]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#node-affinity
          [3]: https://cert-manager.io/
          [4]: https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass
          [5]: https://kubernetes.io/docs/concepts/workloads/pods/pod-topology-spread-constraints/
        displayName: CockroachDB Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cockroach
        - cockroachdb
        links:
        - name: Helm Chart Source
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb
        - name: Configuration Options
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration
        - name: CockroachDB Source
          url: https://github.com/cockroachdb/cockroach
        maintainers:
        - email: dmesser@redhat.com
          name: Daniel Messer
        maturity: stable
        minKubeVersion: 1.8.0
        provider:
          name: Helm Community
          url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/helmoperators/cockroachdb:v5.0.4
        version: 5.0.4
      entries:
      - name: cockroachdb.v5.0.4
        version: 5.0.4
      - name: cockroachdb.v5.0.3
        version: 5.0.3
      name: stable-5.x
    - currentCSV: cockroachdb.v6.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1alpha1",
                "kind": "Cockroachdb",
                "metadata": {
                  "name": "cockroachdb-sample"
                },
                "spec": {"image":{"repository":"cockroachdb/cockroach","tag":"v22.2.4","pullPolicy":"IfNotPresent","credentials":{}},"labels":{},"clusterDomain":"cluster.local","conf":{"attrs":[],"cache":"25%","cluster-name":"","disable-cluster-name-verification":false,"join":[],"log":{"enabled":false,"config":{}},"logtostderr":"INFO","max-sql-memory":"25%","locality":"","single-node":false,"sql-audit-dir":"","port":26257,"http-port":8080,"path":"cockroach-data","store":{"enabled":false,"type":null,"size":null,"attrs":null}},"statefulset":{"replicas":3,"updateStrategy":{"type":"RollingUpdate"},"podManagementPolicy":"Parallel","budget":{"maxUnavailable":1},"args":[],"env":[],"secretMounts":[],"labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{},"nodeAffinity":{},"podAffinity":{},"podAntiAffinity":{"topologyKey":"kubernetes.io/hostname","type":"soft","weight":100},"nodeSelector":{},"priorityClassName":"","tolerations":[],"topologySpreadConstraints":{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"},"resources":{},"customLivenessProbe":{},"customReadinessProbe":{}},"service":{"ports":{"grpc":{"external":{"port":26257,"name":"grpc"},"internal":{"port":26257,"name":"grpc-internal"}},"http":{"port":8080,"name":"http"}},"public":{"type":"ClusterIP","labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{}},"discovery":{"labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{}}},"ingress":{"enabled":false,"labels":{},"annotations":{},"paths":["/"],"hosts":[],"tls":[]},"prometheus":{"enabled":true},"serviceMonitor":{"enabled":false,"labels":{},"annotations":{},"interval":"10s","namespaced":false},"storage":{"hostPath":"","persistentVolume":{"enabled":true,"size":"100Gi","storageClass":"","labels":{},"annotations":{}}},"init":{"labels":{"app.kubernetes.io/component":"init"},"jobAnnotations":{},"annotations":{},"affinity":{},"nodeSelector":{},"tolerations":[],"resources":{},"provisioning":{"enabled":false,"clusterSettings":null,"users":[],"databases":[]}},"tls":{"enabled":true,"serviceAccount":{"create":true,"name":""},"copyCerts":{"image":"busybox"},"certs":{"provided":false,"clientRootSecret":"cockroachdb-root","nodeSecret":"cockroachdb-node","tlsSecret":false,"selfSigner":{"enabled":true,"caProvided":false,"caSecret":"","minimumCertDuration":"624h","caCertDuration":"43800h","caCertExpiryWindow":"648h","clientCertDuration":"672h","clientCertExpiryWindow":"48h","nodeCertDuration":"8760h","nodeCertExpiryWindow":"168h","rotateCerts":true,"readinessWait":"30s","podUpdateTimeout":"2m"},"certManager":false,"certManagerIssuer":{"group":"cert-manager.io","kind":"Issuer","name":"cockroachdb","clientCertDuration":"672h","clientCertExpiryWindow":"48h","nodeCertDuration":"8760h","nodeCertExpiryWindow":"168h"},"useCertManagerV1CRDs":false},"selfSigner":{"image":{"repository":"cockroachlabs-helm-charts/cockroach-self-signer-cert","tag":"1.3","pullPolicy":"IfNotPresent","credentials":{},"registry":"gcr.io"}}},"networkPolicy":{"enabled":false,"ingress":{"grpc":[],"http":[]}},"iap":{"enabled":false}}
              }
            ]
          capabilities: Basic Install
          categories: Database
          certified: "false"
          containerImage: quay.io/cockroachdb/cockroach-helm-operator:6.0.0
          createdAt: 2021-01-23T10-00-00Z
          description: CockroachDB Operator based on the CockroachDB helm chart
          olm.skipRange: <6.0.0
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/cockroachdb/helm-charts
          support: CockroachDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a CockroachDB cluster
            displayName: CockroachDB
            kind: Cockroachdb
            name: cockroachdbs.charts.operatorhub.io
            version: v1alpha1
        description: |
          CockroachDB is a scalable, survivable, strongly-consistent SQL database.
          ## About this Operator
          This Operator is based on a Helm chart for CockroachDB. It supports reconfiguration for some parameters, but notably does not handle scale down of the replica count in a seamless manner. Scale up works great.

          ## Core capabilities
          * **StatefulSet** - Sets up a dynamically scalable CockroachDB cluster using a Kubernetes StatefulSet
          * **Expand Replicas** - Supports expanding the set of replicas by simply editing your object
          * **Dashboard** - Installs the CockroachDB user interface to administer your cluster. Easily expose it via an Ingress rule.
          Review all of the [confiuguration options](https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration) to best run your database instance. The example configuration is derived from the chart's [`values.yaml`](https://github.com/helm/charts/blob/master/stable/cockroachdb/values.yaml).
          ## Using the cluster
          The resulting cluster endpoint can be consumed from a `Service` that follows the pattern: `<StatefulSet-name>-public`. For example to connect using the command line client, use something like the following to obtain the name of the service:
          ```
          kubectl get service -l chart=cockroachdb-6.0.0
          NAME                                           TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)              AGE
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb          ClusterIP   None             <none>        26257/TCP,8080/TCP   24m
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public   ClusterIP   10.106.249.134   <none>        26257/TCP,8080/TCP   24m
          ```
          Then you can use the CockroachDB command line client to connect to the database cluster:
          ```
          kubectl run -it --rm cockroach-client --image=cockroachdb/cockroach --restart=Never --command -- ./cockroach sql --insecure --host example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public
          ```
          ## Before you start
          This Operator requires a cluster with PV support in order to run correctly.

          Note that for a production cluster, you will likely want to override the following parameters in [`values.yaml`](values.yaml) with your own values.

          - `statefulset.resources.requests.memory` and `statefulset.resources.limits.memory` allocate memory resources to CockroachDB pods in your cluster.
          - `conf.cache` and `conf.max-sql-memory` are memory limits that we recommend setting to 1/4 of the above resource allocation. When running CockroachDB, you must set these limits explicitly to avoid running out of memory.
          - `storage.persistentVolume.size` defaults to `100Gi` of disk space per pod, which you may increase or decrease for your use case.
          - `storage.persistentVolume.storageClass` uses the default storage class for your environment. We strongly recommend that you specify a storage class which uses an SSD.
          - `tls.enabled` must be set to `yes`/`true` to deploy in secure mode.

          For more information on overriding the `values.yaml` parameters, please see:
          > <https://www.cockroachlabs.com/docs/stable/orchestrate-cockroachdb-with-kubernetes.html#step-2-start-cockroachdb>

          If you are running in secure mode (with configuration parameter `tls.enabled` set to `yes`/`true`) and `tls.certs.provided` set to `no`/`false`), then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          When `tls.certs.provided` is set to `yes`/`true`, this chart will use certificates created outside of Kubernetes. You may want to use this if you want to use a different certificate authority from the one being used by Kubernetes or if your Kubernetes cluster doesn't fully support certificate-signing requests. To use this, first set up your certificates and load them into your Kubernetes cluster as Secrets using the commands below:

          ```
          mkdir certs
          mkdir my-safe-directory
          cockroach cert create-ca --certs-dir=certs --ca-key=my-safe-directory/ca.key
          cockroach cert create-client root --certs-dir=certs --ca-key=my-safe-directory/ca.key
          kubectl create secret generic cockroachdb-root --from-file=certs
          cockroach cert create-node --certs-dir=certs --ca-key=my-safe-directory/ca.key localhost 127.0.0.1 eerie-horse-cockroachdb-public eerie-horse-cockroachdb-public.default eerie-horse-cockroachdb-public.default.svc.cluster.local *.eerie-horse-cockroachdb *.eerie-horse-cockroachdb.default *.eerie-horse-cockroachdb.default.svc.cluster.local
          kubectl create secret generic cockroachdb-node --from-file=certs
          ```

          Set `tls.certs.tlsSecret` to `yes/true` if you make use of [cert-manager][3] in your cluster.

          [cert-manager][3] stores generated certificates in dedicated TLS secrets. Thus, they are always named:

          * `ca.crt`
          * `tls.crt`
          * `tls.key`

          On the other hand, CockroachDB also demands dedicated certificate filenames:

          * `ca.crt`
          * `node.crt`
          * `node.key`
          * `client.root.crt`
          * `client.root.key`

          By activating `tls.certs.tlsSecret` we benefit from projected secrets and convert the TLS secret filenames to their according CockroachDB filenames.

          If you are running in secure mode, then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          Confirm that all pods are `Running` successfully and init has been completed:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS      RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running     0          1m
          my-release-cockroachdb-1            1/1       Running     0          1m
          my-release-cockroachdb-2            1/1       Running     0          1m
          my-release-cockroachdb-init-k6jcr   0/1       Completed   0          1m
          ```

          Confirm that persistent volumes are created and claimed for each pod:

          ```shell
          kubectl get pv
          ```

          ```
          NAME                                       CAPACITY   ACCESS MODES   RECLAIM POLICY   STATUS    CLAIM                                      STORAGECLASS   REASON    AGE
          pvc-64878ebf-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-0   standard                 51s
          pvc-64945b4f-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-1   standard                 51s
          pvc-649d920d-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-2   standard                 51s
          ```

          ## Upgrading the cluster

          ### Chart version 3.0.0 and after

          Launch a temporary interactive pod and start the built-in SQL client:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          Set `cluster.preserve_downgrade_option`, where `$current_version` is the CockroachDB version currently running (e.g., `19.2`):

          ```sql
          > SET CLUSTER SETTING cluster.preserve_downgrade_option = '$current_version';
          ```

          Exit the shell and delete the temporary pod:

          ```sql
          > \q
          ```

          Kick off the upgrade process by changing the new Docker image, where `$new_version` is the CockroachDB version to which you are upgrading:

          ```shell
          kubectl delete job my-release-cockroachdb-init
          ```

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"image":{"tag":"<new version tag here>"}}}' --type merge
          ```

          Kubernetes will carry out a safe [rolling upgrade](https://kubernetes.io/docs/tutorials/stateful-application/basic-stateful-set/#updating-statefulsets) of your CockroachDB nodes one-by-one. Monitor the cluster's pods until all have been successfully restarted:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS              RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running             0          2m
          my-release-cockroachdb-1            1/1       Running             0          3m
          my-release-cockroachdb-2            1/1       Running             0          3m
          my-release-cockroachdb-3            0/1       ContainerCreating   0          25s
          my-release-cockroachdb-init-nwjkh   0/1       ContainerCreating   0          6s
          ```

          ```shell
          kubectl get pods \
          -o jsonpath='{range .items[*]}{.metadata.name}{"\t"}{.spec.containers[0].image}{"\n"}'
          ```

          ```
          my-release-cockroachdb-0    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-1    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-2    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-3    cockroachdb/cockroach:v20.2.4
          ```

          Resume normal operations. Once you are comfortable that the stability and performance of the cluster is what you'd expect post-upgrade, finalize the upgrade:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          ```sql
          > RESET CLUSTER SETTING cluster.preserve_downgrade_option;
          > \q
          ```

          ## Configuration

          The following table lists the configurable parameters of the CockroachDB chart and their default values.
          For details see the [`values.yaml`](values.yaml) file.

          | Parameter                                                 | Description                                                     | Default                                          |
          | ---------                                                 | -----------                                                     | -------                                          |
          | `clusterDomain`                                           | Cluster's default DNS domain                                    | `cluster.local`                                  |
          | `conf.attrs`                                              | CockroachDB node attributes                                     | `[]`                                             |
          | `conf.cache`                                              | Size of CockroachDB's in-memory cache                           | `25%`                                            |
          | `conf.cluster-name`                                       | Name of CockroachDB cluster                                     | `""`                                             |
          | `conf.disable-cluster-name-verification`                  | Disable CockroachDB cluster name verification                   | `no`                                             |
          | `conf.join`                                               | List of already-existing CockroachDB instances                  | `[]`                                             |
          | `conf.max-disk-temp-storage`                              | Max storage capacity for temp data                              | `0`                                              |
          | `conf.max-offset`                                         | Max allowed clock offset for CockroachDB cluster                | `500ms`                                          |
          | `conf.max-sql-memory`                                     | Max memory to use processing SQL querie                         | `25%`                                            |
          | `conf.locality`                                           | Locality attribute for this deployment                          | `""`                                             |
          | `conf.single-node`                                        | Disable CockroachDB clustering (standalone mode)                | `no`                                             |
          | `conf.sql-audit-dir`                                      | Directory for SQL audit log                                     | `""`                                             |
          | `conf.port`                                               | CockroachDB primary serving port in Pods                        | `26257`                                          |
          | `conf.http-port`                                          | CockroachDB HTTP port in Pods                                   | `8080`                                           |
          | `image.repository`                                        | Container image name                                            | `cockroachdb/cockroach`                          |
          | `image.tag`                                               | Container image tag                                             | `v20.2.4`                                        |
          | `image.pullPolicy`                                        | Container pull policy                                           | `IfNotPresent`                                   |
          | `image.credentials`                                       | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `statefulset.replicas`                                    | StatefulSet replicas number                                     | `3`                                              |
          | `statefulset.updateStrategy`                              | Update strategy for StatefulSet Pods                            | `{"type": "RollingUpdate"}`                      |
          | `statefulset.podManagementPolicy`                         | `OrderedReady`/`Parallel` Pods creation/deletion order          | `Parallel`                                       |
          | `statefulset.budget.maxUnavailable`                       | k8s PodDisruptionBudget parameter                               | `1`                                              |
          | `statefulset.args`                                        | Extra command-line arguments                                    | `[]`                                             |
          | `statefulset.env`                                         | Extra env vars                                                  | `[]`                                             |
          | `statefulset.secretMounts`                                | Additional Secrets to mount at cluster members                  | `[]`                                             |
          | `statefulset.labels`                                      | Additional labels of StatefulSet and its Pods                   | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `statefulset.annotations`                                 | Additional annotations of StatefulSet Pods                      | `{}`                                             |
          | `statefulset.nodeAffinity`                                | [Node affinity rules][2] of StatefulSet Pods                    | `{}`                                             |
          | `statefulset.podAffinity`                                 | [Inter-Pod affinity rules][1] of StatefulSet Pods               | `{}`                                             |
          | `statefulset.podAntiAffinity`                             | [Anti-affinity rules][1] of StatefulSet Pods                    | auto                                             |
          | `statefulset.podAntiAffinity.topologyKey`                 | The topologyKey for auto [anti-affinity rules][1]               | `kubernetes.io/hostname`                         |
          | `statefulset.podAntiAffinity.type`                        | Type of auto [anti-affinity rules][1]                           | `soft`                                           |
          | `statefulset.podAntiAffinity.weight`                      | Weight for `soft` auto [anti-affinity rules][1]                 | `100`                                            |
          | `statefulset.nodeSelector`                                | Node labels for StatefulSet Pods assignment                     | `{}`                                             |
          | `statefulset.priorityClassName`                           | [PriorityClassName][4] for StatefulSet Pods                     | `""`                                             |
          | `statefulset.tolerations`                                 | Node taints to tolerate by StatefulSet Pods                     | `[]`                                             |
          | `statefulset.topologySpreadConstraints`                   | [Topology Spread Constraints rules][5] of StatefulSet Pods      | auto                                             |
          | `statefulset.topologySpreadConstraints.maxSkew`           | Degree to which Pods may be unevenly distributed                | `1`                                              |
          | `statefulset.topologySpreadConstraints.topologyKey`       | The key of node labels                                          | `topology.kubernetes.io/zone`                    |
          | `statefulset.topologySpreadConstraints.whenUnsatisfiable` | `ScheduleAnyway`/`DoNotSchedule` for unsatisfiable constraints  | `ScheduleAnyway`                                 |
          | `statefulset.resources`                                   | Resource requests and limits for StatefulSet Pods               | `{}`                                             |
          | `service.ports.grpc.external.port`                        | CockroachDB primary serving port in Services                    | `26257`                                          |
          | `service.ports.grpc.external.name`                        | CockroachDB primary serving port name in Services               | `grpc`                                           |
          | `service.ports.grpc.internal.port`                        | CockroachDB inter-communication port in Services                | `26257`                                          |
          | `service.ports.grpc.internal.name`                        | CockroachDB inter-communication port name in Services           | `grpc-internal`                                  |
          | `service.ports.http.port`                                 | CockroachDB HTTP port in Services                               | `8080`                                           |
          | `service.ports.http.name`                                 | CockroachDB HTTP port name in Services                          | `http`                                           |
          | `service.public.type`                                     | Public Service type                                             | `ClusterIP`                                      |
          | `service.public.labels`                                   | Additional labels of public Service                             | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.public.annotations`                              | Additional annotations of public Service                        | `{}`                                             |
          | `service.discovery.labels`                                | Additional labels of discovery Service                          | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.discovery.annotations`                           | Additional annotations of discovery Service                     | `{}`                                             |
          | `ingress.enabled`                                         | Enable ingress resource for CockroachDB                         | `false`                                          |
          | `ingress.labels`                                          | Additional labels of Ingress                                    | `{}`                                             |
          | `ingress.annotations`                                     | Additional annotations of Ingress                               | `{}`                                             |
          | `ingress.paths`                                           | Paths for the default host                                      | `[/]`                                            |
          | `ingress.hosts`                                           | CockroachDB Ingress hostnames                                   | `[]`                                             |
          | `ingress.tls[0].hosts`                                    | CockroachDB Ingress tls hostnames                               | `nil`                                            |
          | `ingress.tls[0].secretName`                               | CockroachDB Ingress tls secret name                             | `nil`                                            |
          | `serviceMonitor.enabled`                                  | Create [ServiceMonitor](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/design.md#servicemonitor) Resource for scraping metrics using [PrometheusOperator](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/user-guides/getting-started.md#prometheus-operator)                     | `false`                                             |
          | `serviceMonitor.labels`                                   | Additional labels of ServiceMonitor                             | `{}`                                             |
          | `serviceMonitor.annotations`                              | Additional annotations of ServiceMonitor                        | `{}`                                             |
          | `serviceMonitor.interval`                                 | ServiceMonitor scrape metrics interval                          | `10s`                                            |
          | `serviceMonitor.scrapeTimeout`                            | ServiceMonitor scrape timeout                                   | `nil`                                            |
          | `storage.hostPath`                                        | Absolute path on host to store data                             | `""`                                             |
          | `storage.persistentVolume.enabled`                        | Whether to use PersistentVolume to store data                   | `yes`                                            |
          | `storage.persistentVolume.size`                           | PersistentVolume size                                           | `100Gi`                                          |
          | `storage.persistentVolume.storageClass`                   | PersistentVolume class                                          | `""`                                             |
          | `storage.persistentVolume.labels`                         | Additional labels of PersistentVolumeClaim                      | `{}`                                             |
          | `storage.persistentVolume.annotations`                    | Additional annotations of PersistentVolumeClaim                 | `{}`                                             |
          | `init.labels`                                             | Additional labels of init Job and its Pod                       | `{"app.kubernetes.io/component": "init"}`        |
          | `init.annotations`                                        | Additional labels of the Pod of init Job                        | `{}`                                             |
          | `init.affinity`                                           | [Affinity rules][2] of init Job Pod                             | `{}`                                             |
          | `init.nodeSelector`                                       | Node labels for init Job Pod assignment                         | `{}`                                             |
          | `init.tolerations`                                        | Node taints to tolerate by init Job Pod                         | `[]`                                             |
          | `init.resources`                                          | Resource requests and limits for the Pod of init Job            | `{}`                                             |
          | `tls.enabled`                                             | Whether to run securely using TLS certificates                  | `no`                                             |
          | `tls.serviceAccount.create`                               | Whether to create a new RBAC service account                    | `yes`                                            |
          | `tls.serviceAccount.name`                                 | Name of RBAC service account to use                             | `""`                                             |
          | `tls.certs.provided`                                      | Bring your own certs scenario, i.e certificates are provided    | `no`                                             |
          | `tls.certs.clientRootSecret`                              | If certs are provided, secret name for client root cert         | `cockroachdb-root`                               |
          | `tls.certs.nodeSecret`                                    | If certs are provided, secret name for node cert                | `cockroachdb-node`                               |
          | `tls.certs.tlsSecret`                                     | Own certs are stored in TLS secret                              | `no`                                             |
          | `tls.init.image.repository`                               | Image to use for requesting TLS certificates                    | `cockroachdb/cockroach-k8s-request-cert`         |
          | `tls.init.image.tag`                                      | Image tag to use for requesting TLS certificates                | `0.4`                                            |
          | `tls.init.image.pullPolicy`                               | Requesting TLS certificates container pull policy               | `IfNotPresent`                                   |
          | `tls.init.image.credentials`                              | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `networkPolicy.enabled`                                   | Enable NetworkPolicy for CockroachDB's Pods                     | `no`                                             |
          | `networkPolicy.ingress.grpc`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |
          | `networkPolicy.ingress.http`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |


          Override the default parameters using the `CockroachDB` custom resource.

          ## Deep dive

          ### Connecting to the CockroachDB cluster

          Once you've created the cluster, you can start talking to it by connecting to its `-public` Service. CockroachDB is PostgreSQL wire protocol compatible, so there's a [wide variety of supported clients](https://www.cockroachlabs.com/docs/install-client-drivers.html). As an example, we'll open up a SQL shell using CockroachDB's built-in shell and play around with it a bit, like this (likely needing to replace `my-release-cockroachdb-public` with the name of the `-public` Service that was created with your installed chart):

          ```shell
          kubectl run cockroach-client --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host my-release-cockroachdb-public
          ```

          ```
          Waiting for pod default/cockroach-client to be running, status is Pending,
          pod ready: false
          If you don't see a command prompt, try pressing enter.
          root@my-release-cockroachdb-public:26257> SHOW DATABASES;
          +--------------------+
          |      Database      |
          +--------------------+
          | information_schema |
          | pg_catalog         |
          | system             |
          +--------------------+
          (3 rows)
          root@my-release-cockroachdb-public:26257> CREATE DATABASE bank;
          CREATE DATABASE
          root@my-release-cockroachdb-public:26257> CREATE TABLE bank.accounts (id INT
          PRIMARY KEY, balance DECIMAL);
          CREATE TABLE
          root@my-release-cockroachdb-public:26257> INSERT INTO bank.accounts VALUES
          (1234, 10000.50);
          INSERT 1
          root@my-release-cockroachdb-public:26257> SELECT * FROM bank.accounts;
          +------+---------+
          |  id  | balance |
          +------+---------+
          | 1234 | 10000.5 |
          +------+---------+
          (1 row)
          root@my-release-cockroachdb-public:26257> \q
          Waiting for pod default/cockroach-client to terminate, status is Running
          pod "cockroach-client" deleted
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          ### Cluster health

          Because our pod spec includes regular health checks of the CockroachDB processes, simply running `kubectl get pods` and looking at the `STATUS` column is sufficient to determine the health of each instance in the cluster.

          If you want more detailed information about the cluster, the best place to look is the Admin UI.

          ### Accessing the Admin UI

          If you want to see information about how the cluster is doing, you can try pulling up the CockroachDB Admin UI by port-forwarding from your local machine to one of the pods (replacing `my-release-cockroachdb-0` with the name of one of your pods:

          ```shell
          kubectl port-forward my-release-cockroachdb-0 8080
          ```

          You should then be able to access the Admin UI by visiting <http://localhost:8080/> in your web browser.

          ### Failover

          If any CockroachDB member fails, it is restarted or recreated automatically by the Kubernetes infrastructure, and will re-join the cluster automatically when it comes back up. You can test this scenario by killing any of the CockroachDB pods:

          ```shell
          kubectl delete pod my-release-cockroachdb-1
          ```

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          After a while:

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-1  1/1       Running       0          20s
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          You can check the state of re-joining from the new pod's logs:

          ```shell
          kubectl logs my-release-cockroachdb-1
          ```

          ```
          [...]
          I161028 19:32:09.754026 1 server/node.go:586  [n1] node connected via gossip and
          verified as part of cluster {"35ecbc27-3f67-4e7d-9b8f-27c31aae17d6"}
          [...]
          cockroachdb-0.my-release-cockroachdb.default.svc.cluster.local:26257
          build:      beta-20161027-55-gd2d3c7f @ 2016/10/28 19:27:25 (go1.7.3)
          admin:      http://0.0.0.0:8080
          sql:
          postgresql://root@my-release-cockroachdb-1.my-release-cockroachdb.default.svc.cluster.local:26257?sslmode=disable
          logs:       cockroach-data/logs
          store[0]:   path=cockroach-data
          status:     restarted pre-existing node
          clusterID:  {35ecbc27-3f67-4e7d-9b8f-27c31aae17d6}
          nodeID:     2
          [...]
          ```

          ### NetworkPolicy

          To enable NetworkPolicy for CockroachDB, install [a networking plugin that implements the Kubernetes NetworkPolicy spec](https://kubernetes.io/docs/tasks/administer-cluster/declare-network-policy#before-you-begin), and set `networkPolicy.enabled` to `yes`/`true`.

          For Kubernetes v1.5 & v1.6, you must also turn on NetworkPolicy by setting the `DefaultDeny` Namespace annotation. Note: this will enforce policy for _all_ pods in the Namespace:

          ```shell
          kubectl annotate namespace default "net.beta.kubernetes.io/network-policy={\"ingress\":{\"isolation\":\"DefaultDeny\"}}"
          ```

          For more precise policy, set `networkPolicy.ingress.grpc` and `networkPolicy.ingress.http` rules. This will only allow pods that match the provided rules to connect to CockroachDB.

          ### Scaling

          Scaling should be managed via the `helm upgrade` command. After resizing your cluster on your cloud environment (e.g., GKE or EKS), run the following command to add a pod. This assumes you scaled from 3 to 4 nodes:

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"statefulset":{"replicas":"4"}}}' --type merge
          ```

          Note, that if you are running in secure mode (`tls.enabled` is `yes`/`true`) and increase the size of your cluster, you will also have to approve the CSR (certificate-signing request) of each new node (using `kubectl get csr` and `kubectl certificate approve`).

          [1]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [2]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#node-affinity
          [3]: https://cert-manager.io/
          [4]: https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass
          [5]: https://kubernetes.io/docs/concepts/workloads/pods/pod-topology-spread-constraints/
        displayName: CockroachDB Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cockroach
        - cockroachdb
        links:
        - name: Helm Chart Source
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb
        - name: Configuration Options
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration
        - name: CockroachDB Source
          url: https://github.com/cockroachdb/cockroach
        maintainers:
        - email: yandu@cockroachlabs.com
          name: Yandu Oppacher
        maturity: stable
        minKubeVersion: 1.8.0
        provider:
          name: Helm Community
          url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/cockroachdb/cockroach-helm-operator:6.0.0
        version: 6.0.0
      entries:
      - name: cockroachdb.v6.0.0
        version: 6.0.0
      name: stable-v6.x
    defaultChannel: stable-v6.x
    packageName: cockroachdb
    provider:
      name: Helm Community
      url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CrowdStrike
      provider-url: https://crowdStrike.com
    name: falcon-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: falcon-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconContainer",
                "metadata": {
                  "name": "falcon-sidecar-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "sidecar"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconNodeSensor",
                "metadata": {
                  "name": "falcon-node-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "daemonset"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security,Monitoring
          containerImage: quay.io/crowdstrike/falcon-operator:0.8.1
          createdAt: "2024-06-06T10:00:00Z"
          description: '[DEPRECATED] Use the CrowdStrike Falcon Operator from the
            certified channel instead'
          operatorframework.io/suggested-namespace: falcon-operator
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/CrowdStrike/falcon-operator
          support: Community Only
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FalconContainer is the Schema for the falconcontainers API
            displayName: Falcon Container
            kind: FalconContainer
            name: falconcontainers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconNodeSensor is the Schema for the falconnodesensors
              API
            displayName: Falcon Node Sensor
            kind: FalconNodeSensor
            name: falconnodesensors.falcon.crowdstrike.com
            version: v1alpha1
        description: |-
          ## Deprecated
          This CrowdStrike community operator is deprecated, use the CrowdStrike Falcon Operator from the certified channel instead.

          ## About the CrowdStrike Falcon Operator
          The CrowdStrike Falcon Operator deploys CrowdStrike Falcon Workload Protection to the cluster. The operator exposes 2 custom resources that allows you to deploy either the Falcon Container Sensor or Falcon Node Sensor.

          ## About Custom Resources

          | Custom Resource                                       | Description                                                      |
          | :--------                                             | :------------                                                    |
          | [FalconContainer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/container/README.md) | Manages installation of Falcon Container Sensor on the cluster   |
          | [FalconNodeSensor](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/node/README.md)     | Manages installation of Falcon Linux Sensor on the cluster nodes |

          ## Installation and Deployment

          For installation and deployment of the CrowdStrike Falcon Operator and its Custom Resources, please read the [Installation and Deployment Guide](./docs/install_guide.md) and choose the deployment method that is right for your target environment.

          ## Getting Help
          If you encounter any issues while using the Falcon Operator, you can create an issue on our [Github repo](https://github.com/CrowdStrike/falcon-operator) for bugs, enhancements, or other requests.

          ## Contributing
          You can contribute by:

          * Raising any issues you find using Falcon Operator
          * Fixing issues by opening [Pull Requests](https://github.com/CrowdStrike/falcon-operator/pulls)
          * Improving documentation
          * Talking about the CrowdStrike Falcon Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/CrowdStrike/falcon-operator/issues).

          ## Additional Resources
           - CrowdStrike Container Security: [Product Page](https://www.crowdstrike.com/products/cloud-security/falcon-cloud-workload-protection/container-security/)
           - So You Think Your Containers Are Secure? Four Steps to Ensure a Secure Container Deployment: [Blog Post](https://www.crowdstrike.com/blog/four-steps-to-ensure-a-secure-containter-deployment/)
           - Container Security With CrowdStrike: [Blog Post](https://www.crowdstrike.com/blog/tech-center/container-security/)
           - To learn more about Falcon Container Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/support/documentation/146/falcon-container-sensor-for-linux), [Release Notes](https://falcon.crowdstrike.com/support/news/release-notes-falcon-container-sensor-for-linux)
           - To learn more about Falcon Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/documentation/20/falcon-sensor-for-linux#kubernetes-support)
           - [Developer Documentation](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/developer_guide.md)
        displayName: '[DEPRECATED] CrowdStrike Operator'
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - runtime-protection
        links:
        - name: Falcon Operator
          url: https://github.com/CrowdStrike/falcon-operator
        maintainers:
        - email: integrations@crowdstrike.com
          name: CrowdStrike Solutions Architecture
        maturity: alpha
        minKubeVersion: 1.17.0
        provider:
          name: CrowdStrike
          url: https://crowdStrike.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/crowdstrike/falcon-operator:0.8.1
        version: 0.8.2
      entries:
      - name: falcon-operator.v0.8.2
        version: 0.8.2
      - name: falcon-operator.v0.8.1
        version: 0.8.1
      name: alpha
    defaultChannel: alpha
    packageName: falcon-operator
    provider:
      name: CrowdStrike
      url: https://crowdStrike.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Foundation Model Stack
      provider-url: ""
    name: multi-nic-cni-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multi-nic-cni-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.3.0",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.3.0",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.3.0
          createdAt: "2024-04-17T04:51:40Z"
          description: Automate container multi-network configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: http://github.com/foundation-model-stack/multi-nic-cni
          support: Foundation Model Stack Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Foundation Model Stack
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.3.0
        version: 1.3.0
      entries:
      - name: multi-nic-cni-operator.v1.3.0
        version: 1.3.0
      - name: multi-nic-cni-operator.v1.2.9
        version: 1.2.9
      - name: multi-nic-cni-operator.v1.2.8
        version: 1.2.8
      - name: multi-nic-cni-operator.v1.2.2
        version: 1.2.2
      - name: multi-nic-cni-operator.v1.2.1
        version: 1.2.1
      - name: multi-nic-cni-operator.v1.0.2
        version: 1.0.2
      name: alpha
    - currentCSV: multi-nic-cni-operator.v1.2.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.7",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.7",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.7
          createdAt: "2025-05-09T12:45:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: minzhang@redhat.com
          name: Min Zhang
        - email: ebendror@redhat.com
          name: Eran Ben Dror
        maturity: alpha
        provider:
          name: Foundation Model Stack
        relatedImages:
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.7
        version: 1.2.7
      entries:
      - name: multi-nic-cni-operator.v1.2.7
        version: 1.2.7
      - name: multi-nic-cni-operator.v1.2.3
        version: 1.2.3
      - name: multi-nic-cni-operator.v1.2.0
        version: 1.2.0
      - name: multi-nic-cni-operator.v1.1.0
        version: 1.1.0
      name: beta
    - currentCSV: multi-nic-cni-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.6",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.6",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.6
          createdAt: "2025-03-24T14:49:54Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: minzhang@redhat.com
          name: Min Zhang
        - email: ebendror@redhat.com
          name: Eran Ben Dror
        maturity: alpha
        provider:
          name: Foundation Model Stack
        relatedImages:
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.6
        version: 1.2.6
      entries:
      - name: multi-nic-cni-operator.v1.2.6
        version: 1.2.6
      - name: multi-nic-cni-operator.v1.2.5
        version: 1.2.5
      - name: multi-nic-cni-operator.v1.2.4
        version: 1.2.4
      - name: multi-nic-cni-operator.v1.0.5
        version: 1.0.5
      - name: multi-nic-cni-operator.v1.0.4
        version: 1.0.4
      - name: multi-nic-cni-operator.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: multi-nic-cni-operator
    provider:
      name: Foundation Model Stack
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: assisted-service-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: assisted-service-operator.v0.7.74
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-x86_64-live.x86_64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-aarch64-live.aarch64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-s390x-live.s390x.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-x86_64-live.x86_64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-aarch64-live.aarch64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-s390x-live.s390x.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-x86_64-live.x86_64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-aarch64-live.aarch64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-ppc64le-live.ppc64le.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-s390x-live.s390x.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-x86_64-live.x86_64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-aarch64-live.aarch64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-ppc64le-live.ppc64le.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-s390x-live.s390x.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-s390x-live.s390x.iso",
                      "version": "418.94.202409162337-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service:latest
          createdAt: "2024-10-16T17:43:34Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.74'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AgentClassification is the Schema for the AgentClassifications
              API
            displayName: Agent Classification
            kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - description: Agent is the Schema for the hosts API
            displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: Infra Env
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service:latest
        version: 0.7.74
      entries:
      - name: assisted-service-operator.v0.7.74
        version: 0.7.74
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: assisted-service-operator.v0.7.71
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-x86_64-live.x86_64.iso",
                      "version": "414.92.202310170514-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-aarch64-live.aarch64.iso",
                      "version": "414.92.202310170514-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202309201615-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-s390x-live.s390x.iso",
                      "version": "414.92.202309201615-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-s390x-live.s390x.iso",
                      "version": "415.92.202310310037-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:cffbdc5fd51c08f16ec2aa484cf226e3b29510919a8aa9668d766007de8f3d79
          createdAt: "2023-12-15T15:26:08Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.71'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:98740646fbfc0520f9d85ac5b483d1a898f22e164c5234901be51970bef7fcf3
        - quay.io/edge-infrastructure/assisted-image-service@sha256:027fa9c5a881848786745aafeac5e48575eefe22417abe29d9e8eb17940b8f80
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/edge-infrastructure/assisted-service@sha256:cffbdc5fd51c08f16ec2aa484cf226e3b29510919a8aa9668d766007de8f3d79
        - quay.io/edge-infrastructure/assisted-installer@sha256:0f4ecf9a75b6755d02a45f9b70357ac45b6b4db66f24f0e83f888269d64f4293
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:40bafed94547a5fe96deb5492f10225501b965198227c9634869931692bda492
        version: 0.7.71
      entries:
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.10
    - currentCSV: assisted-service-operator.v0.7.73
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-x86_64-live.x86_64.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-aarch64-live.aarch64.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-s390x-live.s390x.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-x86_64-live.x86_64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-aarch64-live.aarch64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-s390x-live.s390x.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.16.0-ec.4/rhcos-4.16.0-ec.4-x86_64-live.x86_64.iso",
                      "version": "416.94.202403071059-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-aarch64-live.aarch64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-s390x-live.s390x.iso",
                      "version": "415.92.202402130021-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:641fb65d8f0117cc6007567337ae38d5e912add454a76905617b60db99b60a64
          createdAt: "2024-03-20T10:08:49Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.73'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:d0da07ec81877302fc376e8049dca69c1acdbf56a37aa3e320cd75f2b85a2ed0
        - quay.io/edge-infrastructure/assisted-service@sha256:641fb65d8f0117cc6007567337ae38d5e912add454a76905617b60db99b60a64
        - quay.io/edge-infrastructure/assisted-image-service@sha256:b33068bbe88e38c0444fb158e233e5644fb127416198d52fdc809a3f2e684fde
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/edge-infrastructure/assisted-installer@sha256:234e4680b0ff4376f2c11d6f94a75a5dd6ae16c3218d6b63d2003cf43c65e185
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:2cc14304ba27c51d61b86db05445f3f0acc68593a4d76081da8eb9d6bae41757
        version: 0.7.73
      entries:
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.11
    - currentCSV: assisted-service-operator.v0.7.74
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-x86_64-live.x86_64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-aarch64-live.aarch64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-s390x-live.s390x.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-x86_64-live.x86_64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-aarch64-live.aarch64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-s390x-live.s390x.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-x86_64-live.x86_64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-aarch64-live.aarch64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-ppc64le-live.ppc64le.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-s390x-live.s390x.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-x86_64-live.x86_64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-aarch64-live.aarch64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-ppc64le-live.ppc64le.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-s390x-live.s390x.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-s390x-live.s390x.iso",
                      "version": "418.94.202409162337-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service:latest
          createdAt: "2024-10-16T17:43:34Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.74'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AgentClassification is the Schema for the AgentClassifications
              API
            displayName: Agent Classification
            kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - description: Agent is the Schema for the hosts API
            displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: Infra Env
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service:latest
        version: 0.7.74
      entries:
      - name: assisted-service-operator.v0.7.74
        version: 0.7.74
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.13
    - currentCSV: assisted-service-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "osImages": [
                    {
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202105062123-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/ocpmetal/assisted-service@sha256:64e88c5aa41c606bc2db2b0991acdbb1166559306474e13f5e2810f325616d64
          createdAt: "2021-09-23T11:11:20Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.0 <0.0.8'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled, for
              all other names a "Warning" event will be emitted + message logged and
              it will be ignored forever.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent".
          No other name will be accepted.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/ocpmetal/assisted-installer@sha256:db97f7a565a8d5af61f3876d6465996e90c7f79dad6ec71c5bb5cce837444974
        - quay.io/ocpmetal/assisted-installer-agent@sha256:a391892766d1ce81300bf82943318e7c3be0d2945fc19065527ab9f64fc24cf8
        - quay.io/ocpmetal/assisted-installer-controller@sha256:43b65c310eee985ee0fc20ad8ee684f7557ffe704fd1a2fa53831e3dbc41e849
        - quay.io/ocpmetal/assisted-service@sha256:64e88c5aa41c606bc2db2b0991acdbb1166559306474e13f5e2810f325616d64
        - quay.io/ocpmetal/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        version: 0.0.8
      entries:
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.3
    - currentCSV: assisted-service-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.2/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.2/rhcos-4.8.2-x86_64-live.x86_64.iso",
                      "version": "48.84.202107202156-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-x86_64-live.x86_64.iso",
                      "version": "49.84.202110081407-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:c3a852842b88f47c51cf1c62d3a19b102729fb355df19ced0e03e8265739c39c
          createdAt: "2022-04-07T11:23:06Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.1.3'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled, for
              all other names a "Warning" event will be emitted + message logged and
              it will be ignored forever.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent".
          No other name will be accepted.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer@sha256:461d60b21d1d3ccb15a170f937f4871088cd0f83108b2271e84bdf2883a66870
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:979650f2e03e457d71958e8813285fcb7185fe0035f21a6ba956b4277cb29b70
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:658787d1e6c1f733c5f03863655aff315258c0b38e1972573c3878df46e4e463
        - quay.io/edge-infrastructure/assisted-service@sha256:c3a852842b88f47c51cf1c62d3a19b102729fb355df19ced0e03e8265739c39c
        - quay.io/edge-infrastructure/assisted-image-service@sha256:add19e251cafdfc4f34264d0aec5b660f0ea2dcc13326f4a2c8d50a2e465c0a2
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        version: 0.1.3
      entries:
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.4
    - currentCSV: assisted-service-operator.v0.3.30
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-x86_64-live.x86_64.iso",
                      "version": "49.84.202110081407-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-aarch64-live-rootfs.aarch64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-aarch64-live.aarch64.iso",
                      "version": "49.84.202110080947-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.3/rhcos-4.10.3-x86_64-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.3/rhcos-4.10.3-x86_64-live.x86_64.iso",
                      "version": "410.84.202201251210-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:7c13ce52d82f761aec47bc5a6639d7baabe99d0f4f2e2106c4341018c606b458
          createdAt: "2022-06-13T23:16:28Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.3.30'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:2267d2d7be38f81d530fed977b0c46f6c976369cd89c123c8e036c3b1dd988d5
        - quay.io/edge-infrastructure/assisted-service@sha256:7c13ce52d82f761aec47bc5a6639d7baabe99d0f4f2e2106c4341018c606b458
        - quay.io/edge-infrastructure/assisted-image-service@sha256:d6396f9360b877a4f57139ba5b4969f243278db709be4de3573f2fba75e50018
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/edge-infrastructure/assisted-installer@sha256:57cf8d5fe49067a8c1345bedaf90f2b63aededdd7d615878496182c35a13892e
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:6271d639e4c0647bd28fbc85cea54f31143a876c3237376b9c670ec254a9c2fc
        version: 0.3.30
      entries:
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.5
    - currentCSV: assisted-service-operator.v0.4.41
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.40/rhcos-4.9.40-x86_64-live.x86_64.iso",
                      "version": "49.84.202206171736-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.16/rhcos-4.10.16-x86_64-live.x86_64.iso",
                      "version": "410.84.202205191234-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.16/rhcos-4.10.16-aarch64-live.aarch64.iso",
                      "version": "410.84.202205191023-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.11.0-rc.2/rhcos-4.11.0-rc.2-x86_64-live.x86_64.iso",
                      "version": "411.86.202207090519-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.11.0-rc.2/rhcos-4.11.0-rc.2-aarch64-live.aarch64.iso",
                      "version": "411.86.202207090159-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:0fa869095185759c26c391f0e353f293c903ea739576f160d6668f5f20404bdb
          createdAt: "2022-09-11T14:56:07Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.4.41'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:4ca00219efbc0d9dcce415dadc1039e9495f3a3b00ab3bd575b41554d7934899
        - quay.io/edge-infrastructure/assisted-installer@sha256:803168aeb7e01db77e3b2ff7418ceea3d93daccaceeb95f8fc02027144ebac75
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:56484f99799d22f345215bb65c6b81b6ee14e07d26b625c3f283a2ddfdc37238
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:951b9d625bc300b5b72a3916cb382015527ea786f10b7acbeebf83ae854ca51a
        - quay.io/edge-infrastructure/assisted-image-service@sha256:104d2aae313c724fb10d3ef3bba78b2370fc0597de31335453894b679a450ab8
        - quay.io/edge-infrastructure/assisted-service@sha256:0fa869095185759c26c391f0e353f293c903ea739576f160d6668f5f20404bdb
        version: 0.4.41
      entries:
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.6
    - currentCSV: assisted-service-operator.v0.4.50
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-x86_64-live.x86_64.iso",
                      "version": "411.86.202210041459-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-aarch64-live.aarch64.iso",
                      "version": "411.86.202210032347-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.12.0-ec.2/rhcos-4.12.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "412.86.202208101039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.12.0-ec.2/rhcos-4.12.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "412.86.202208101040-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:0ca1cfe04f088cd4c7318cb91978754d6472ada29e3436d05e22335740bfff49
          createdAt: "2022-11-23T19:17:47Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.4.50'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer@sha256:f225c86242b0bfbf58753123ef33b7c07f367a5099641d19cfda5a7121382586
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:34e0341c0471b1772db11535107db06e511c8d5d9816edc3971e745928164ecd
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:c1c7281b9f06e56330cc55ac8d74185d42ee82dff105ae747db677b8eb059aea
        - quay.io/edge-infrastructure/assisted-image-service@sha256:71e8c9c76caa8fe01255fdb4a1172cda372969f7fe1399a3a54fb93e967e68cf
        - quay.io/edge-infrastructure/assisted-service@sha256:0ca1cfe04f088cd4c7318cb91978754d6472ada29e3436d05e22335740bfff49
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:54b39c6cafefb70d258ec1d491b6a0f1dd888cd85bd442c8f40a4805cf4721ca
        version: 0.4.50
      entries:
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.7
    - currentCSV: assisted-service-operator.v0.6.65
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-x86_64-live.x86_64.iso",
                      "version": "411.86.202210041459-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-aarch64-live.aarch64.iso",
                      "version": "411.86.202210032347-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-s390x-live.s390x.iso",
                      "version": "411.86.202210171842-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202210171844-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-x86_64-live.x86_64.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-aarch64-live.aarch64.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-s390x-live.s390x.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-x86_64-live.x86_64.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-aarch64-live.aarch64.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-s390x-live.s390x.iso",
                      "version": "413.92.202303281804-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:2dfaf2c3ac58b2987b0088df302c7cf717d0b6640cdb8b2327370aa110120784
          createdAt: "2023-05-24T11:15:57Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.6.65'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-image-service@sha256:b1ba1076fe16b2f868ed31a660b16583a957a83634c7ec3cf03574b56e02144c
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:1287d947464fbf5275feb31d106d2feb856c6117f0eaec8d947f2e03bd8618c5
        - quay.io/edge-infrastructure/assisted-installer@sha256:f8a7e7cb6ff1a4a7daa0cb26f0fab8da17f109653af5979c9d53c35b7ca2c1bb
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:1ad91b13c763437863480df68f5b65576e95c558d4537886432a81c0b8d3cedc
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:a6f99547564cc2e047df74003019817b2dbd3dad992a0c8e53aaaba4f5b02ab6
        - quay.io/edge-infrastructure/assisted-service@sha256:2dfaf2c3ac58b2987b0088df302c7cf717d0b6640cdb8b2327370aa110120784
        version: 0.6.65
      entries:
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.8
    - currentCSV: assisted-service-operator.v0.7.67
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-s390x-live.s390x.iso",
                      "version": "414.92.202305090606-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:fb0397a1402d42a64e3128e66a0817aa9466e67eb407cb792dda0d380bd5a189
          createdAt: "2023-10-08T07:14:20Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.67'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-image-service@sha256:69ff087be8e5d8458a97f49d88b8b2025ee341d362dcf5618988c011b66688e0
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/edge-infrastructure/assisted-installer@sha256:d85be59bc379927f46d0a88ef04172c5c6733e31e1e4a8188de2d434d5bedb66
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:f2aeb4f7d193ea2c5702cae0f6039c02991b9f417d091468d46974408c97efe4
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:35c118d8c07dd72fe3307424cc051d9e0426e916412171449a05e41445b9f599
        - quay.io/edge-infrastructure/assisted-service@sha256:fb0397a1402d42a64e3128e66a0817aa9466e67eb407cb792dda0d380bd5a189
        version: 0.7.67
      entries:
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.9
    defaultChannel: alpha
    packageName: assisted-service-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Strimzi
      provider-url: ""
    name: strimzi-kafka-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: strimzi-cluster-operator.v0.48.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.1.0",
                      "metadataVersion": "4.1-IV1",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.1.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
          createdAt: "2025-09-25 10:32:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.48.0 supports only Kubernetes 1.27 and newer.**
            Kubernetes 1.25 and 1.26 are not supported anymore

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.48.0
          * Support for Kafka 4.1.0
          * Add monitoring of custom resources using kubernetes-state-metrics (KSM)
          * Support for type: custom client authentication
          * Added support for Strimzi Metrics Reporter to Kafka Connect, Mirror Maker 2 and Kafka Bridge.
          * Added support for KIP-745 in Kafka Connector and Mirror Maker 2, allowing the usage of includeTasks and onlyFailed arguments in Kafka connectors restart.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
        - quay.io/strimzi/kafka@sha256:34450afc1a3399a9026fc3c6ec90e9f6adfec9faf6520545b4d84e90cda34964
        - quay.io/strimzi/kafka-bridge@sha256:53034f64f0b672f10b5bacea1c7a25132f118df7fd5c9032c4dbf702ed93796a
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:64380efcb8da87fdc0e0de238a20456be9abb830174c2f2331882b7a84657e57
        - quay.io/strimzi/kafka@sha256:48db05ad2e9fa39eb388089e7e25654bcbf4da1ddc3a5635d675056c28bbcb96
        version: 0.48.0
      entries:
      - name: strimzi-cluster-operator.v0.48.0
        version: 0.48.0
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: stable
    - currentCSV: strimzi-cluster-operator.v0.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnectS2I",
                "metadata":{
                  "name":"my-connect-s2i-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "whitelist":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 2,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": 1,
                        "offset.storage.replication.factor": 1,
                        "status.storage.replication.factor": 1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "NetworkOutboundCapacityGoal",
                    "CpuCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:59d939f342f96d0e8a54c96fd187832c4b19cd4137bf55ace53d486728ae6c5f
          createdAt: "2021-05-14 11:37:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster with Source 2 Image support
            displayName: Kafka Connect Source to Image
            kind: KafkaConnectS2I
            name: kafkaconnects2is.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  Starting with Strimzi 0.23, we support only the `v1beta2` API version of all Strimzi custom resources.  This is a required for migration to `apiextensions/v1` which is needed because Kubernetes 1.22 will remove support for `apiextensions/v1beta1`.  Migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs **before upgrading to Strimzi 0.23**. For more details about the CRD upgrades, see the [documentation](https://strimzi.io/docs/operators/0.23.0/deploying.html#assembly-upgrade-resources-str).
          ### New in 0.23
          * Migrate to CRD v1 (required by Kubernetes 1.22+)
          * Strimzi API versions `v1alpha1` and `v1beta1` were removed from all Strimzi custom resources apart from `KafkaTopic` and `KafkaUser` (use `v1beta2` versions instead)
          * Add support for Kafka 2.8.0 and 2.6.2, remove support for Kafka 2.5.x
          * Make it possible to configure maximum number of connections and maximum connection creation rate in listener configuration
          * Add support for configuring finalizers for `loadbalancer` type listeners
          * Use dedicated Service Account for Kafka Connect Build on Kubernetes
          * Remove direct ZooKeeper access for handling user quotas in the User Operator. Add usage of Admin Client API instead.
          * Support for configuring custom Authorizer implementation
          * Changed Reconciliation interval for Topic Operator from 90 to 120 seconds (to keep it the same as for other operators)
          * Changed Zookeeper session timeout default value to 18 seconds for Topic and User Operators (for improved resiliency)
          * Removed requirement for replicas and partitions KafkaTopic spec making these parameters optional
          * Support to configure a custom filter for parent CR's labels propagation into subresources
          * Allow disabling service links (environment variables describing Kubernetes services) in Pod template
          * Add support for separate control plane listener (disabled by default, available via the `ControlPlaneListener` feature gate)
          * Support for Dual Stack networking
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _master_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/master/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/master/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:59d939f342f96d0e8a54c96fd187832c4b19cd4137bf55ace53d486728ae6c5f
        - quay.io/strimzi/kafka@sha256:699bff48ee445d8ba381a791cd0340015e3ecede8d169f41bc151095ab7fb5cc
        - quay.io/strimzi/jmxtrans@sha256:b4ec1844f4cc5eb74cd0aac83d74b81b977e178bf670d1136d1af0b68663d0a3
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/kafka@sha256:dbc53bb374d2cb9513a5bb8f80fea751c91b249cabd36d8b3c04af5e5c961738
        - quay.io/strimzi/kafka@sha256:d2745f48ed262290ce66c3362ffb9b9a8cfeb9c76beac40c995f72f3f06ad10e
        - quay.io/strimzi/kafka@sha256:641f46114c59d61577bcb3f4ef5af16aea3ed42c2a575f8d3641c82848df85e8
        - quay.io/strimzi/kafka@sha256:e90695b35d6dd5db20c9900dd0dab95d1ef908f3145b4f9f436b903303c88549
        - quay.io/strimzi/kafka-bridge@sha256:bed07186355a4c04ee50e90048e3bd3a023264093a9ef28f0278cce44640e93f
        version: 0.23.0
      entries:
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.23.x
    - currentCSV: strimzi-cluster-operator.v0.24.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnectS2I",
                "metadata":{
                  "name":"my-connect-s2i-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 2,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:786ea69c77e0651f009a73a8b3ce0b3d41b3cd71b917dc0d9bdc61c725f494a6
          createdAt: "2021-06-24 13:13:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster with Source 2 Image support
            displayName: Kafka Connect Source to Image
            kind: KafkaConnectS2I
            name: kafkaconnects2is.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: "Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org)
          cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/)
          in various deployment configurations. See our [website](https://strimzi.io)
          for more details about the project.\n### CRD Upgrades\n**!!! IMPORTANT !!!**
          \ This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.
          \ If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed
          for all Strimzi CRDs and CRs before the upgrade to 0.24 is done!  If upgrading
          from Strimzi version earlier than 0.22, you need to first install the CRDs
          from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi
          CRDs and CRs before the upgrade to 0.24 is done!  For more details about
          the CRD upgrades, see the documentation.\n### New in 0.24\n* Add support
          for Kafka 2.7.1 and remove support for 2.6.0, 2.6.1, and 2.6.2\n* Add support
          for [Kubernetes Configuration Provider for Apache Kafka](https://github.com/strimzi/kafka-kubernetes-config-provider)\n*
          Use Red Hat UBI8 base image\n* Support for patching of service accounts
          and configuring their labels and annotations. The feature is disabled by
          default and enabled using the new `ServiceAccountPatching` feature gate.\n*
          Added support for configuring cluster-operator's worker thread pool size
          that is used for various sync and async tasks\n* Add Kafka Quotas plugin
          with produce, consume, and storage quotas\n* Support pausing reconciliation
          of KafkaTopic CR with annotation `strimzi.io/pause-reconciliation`\n* Update
          cruise control to 2.5.57\n* Update to Strimzi Kafka Bridge to 0.20.1\n*
          Support for broker load information added to the rebalance optimization
          proposal. Information on the load difference, before and after a rebalance
          is stored in a ConfigMap\n* Add support for selectively changing the verbosity
          of logging for individual CRs, using markers.\n* Added support for `controller_mutation_rate'
          quota. Creation/Deletion of topics and creation of partitions can be configured
          through this.\n* Use newer version of Kafka Exporter with different bugfixes
          \n### Supported Features\n* **Manages the Kafka Cluster** - Deploys and
          manages all of the components of this complex application, including dependencies
          like Apache ZooKeeper® that are traditionally hard to administer.\n* **Includes
          Kafka Connect** - Allows for configuration of common data sources and sinks
          to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for morroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashabords\n### Upgrading your Clusters\nThe
          Strimzi Operator understands how to run and upgrade between a set of Kafka
          versions. When specifying a new version in your config, check to make sure
          you aren't using any features that may have been removed. See [the upgrade
          guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str)
          for more information.\n### Storage\nAn efficient data storage infrastructure
          is essential to the optimal performance of Apache Kafka®. Apache Kafka®
          deployed via Strimzi requires block storage. The use of file storage (for
          example, NFS) is not recommended.\nThe Strimzi Operator supports three types
          of data storage:\n* Ephemeral (Recommended for development only!)\n* Persistent\n*
          JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nStrimzi
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation to the current _main_
          branch as well as all releases can be found on our [website](https://strimzi.io/documentation).\n###
          Getting help\nIf you encounter any issues while using Strimzi, you can get
          help using:\n* [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          Contributing\nYou can contribute by:\n* Raising any issues you find using
          Strimzi\n* Fixing issues by opening Pull Requests\n* Improving documentation\n*
          Talking about Strimzi\n\nAll bugs, tasks or enhancements are tracked as
          [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues).
          Issues which  might be a good start for new contributors are marked with
          [\"good-start\"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start)
          label.\n\nThe [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md)
          describes how to build Strimzi and how to  test your changes before submitting
          a patch or opening a PR.\n\nThe [Documentation Contributor Guide](https://strimzi.io/contributing/guide/)
          describes how to contribute to Strimzi documentation.\n\nIf you want to
          get in touch with us first before contributing, you can use:\n* [Strimzi
          mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          License\nStrimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).\n"
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/kafka@sha256:95cfe9000afda2f7def269ca46472d3803ee85146c521884884d8505a7187daf
        - quay.io/strimzi/kafka@sha256:8959b7968ab8b3306906cdbff2ebb8d63329af37e58124a601843795c4ef5bd6
        - quay.io/strimzi/kafka@sha256:fbb08410d9595029bc4a02ed859971264e6ce2dc85dd6a9855eaa7bb58b52a25
        - quay.io/strimzi/operator@sha256:786ea69c77e0651f009a73a8b3ce0b3d41b3cd71b917dc0d9bdc61c725f494a6
        - quay.io/strimzi/kafka-bridge@sha256:537ea10e87619251204747ebc1005c632a6c6ed270f5ef06f86c0c3ae4a7553b
        - quay.io/strimzi/jmxtrans@sha256:2c54adabf6f31c087dbb898057f9c5e200aaf4c5b39fb3a47bbaf7f6587df72d
        version: 0.24.0
      entries:
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.24.x
    - currentCSV: strimzi-cluster-operator.v0.25.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:afaf1dd1f2f696ae80674d74159661ea239ae609f78a73a1423c7a17a39e5044
          createdAt: "2021-06-24 13:13:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.25 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.25 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.25
          * Move from Scala 2.12 to Scala 2.13
          * Open Policy Agent authorizer updated to a new version supporting Scala 2.13.
            The new version of the authroizer uses different format of the input data sent to the OPA server and might require you to update your policies.
            See [release notes](https://github.com/strimzi/strimzi-kafka-operator/releases/tag/0.25.0) for more details.

          * Allow a custom password to be set for SCRAM-SHA-512 users by referencing a secret in the `KafkaUser` resource
          * Add support for [EnvVar Configuration Provider for Apache Kafka](https://github.com/strimzi/kafka-env-var-config-provider)
          * Add support for `tls-external` authentication to User Operator to allow management of ACLs and Quotas for TLS users with user certificates generated externally
          * Support for disabling the automatic generation of network policies by the Cluster Operator
          * Configure fixed size limit for `emptyDir` volumes used for temporary files
          * Update Strimzi Kafka Bridge to 0.20.2
          * The `KafkaConnectS2I` resource has been removed and is no longer supported by the operator.
            Please use the [migration guide](https://strimzi.io/docs/operators/0.24.0/full/using.html#proc-migrating-kafka-connect-s2i-str) to migrate your `KafkaConnectS2I` deployments to [`KafkaConnect` Build](https://strimzi.io/docs/operators/latest/full/deploying.html#creating-new-image-using-kafka-connect-build-str) instead.

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/jmxtrans@sha256:af5eb8459ac9a0acbca51e67c6e0217f62ba404cba9f2251968c927dcef75252
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/kafka@sha256:d8991608ec71f17e850c33714ee02ca87ba94bc0f077be62ad1ed127ca244823
        - quay.io/strimzi/operator@sha256:afaf1dd1f2f696ae80674d74159661ea239ae609f78a73a1423c7a17a39e5044
        - quay.io/strimzi/kafka@sha256:ee17eb3611019b0cfa5e9e196610edc1dbd7084471fb20990f055c89da8eef45
        - quay.io/strimzi/kafka@sha256:56a33815f7d50a97ea9bb4cce1e3a6d04d918b841b5aa285f16b792d8930059e
        - quay.io/strimzi/kafka-bridge@sha256:a005e20d7fbe04d6f587f2093e21a00556ec62b850df757b9f7d9d57e3141c40
        version: 0.25.0
      entries:
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.25.x
    - currentCSV: strimzi-cluster-operator.v0.26.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.0.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"3.0",
                        "inter.broker.protocol.version":"3.0"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.0.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:5d90caa66f57939ba62a9a50efa00fce319ba596e719a3f8a42170d7635ee3f4
          createdAt: "2021-12-14 00:16:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.26 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.26 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.26.1
          * **Updated Log4j2 to fix CVE-2021-44228**
          ### New in 0.26.0
          * Add support for Kafka 2.8.1 and 3.0.0; remove Kafka 2.7.0 and 2.7.1
          * Add support for exposing JMX in Zookeeper
          * Allow configuring labels and annotations for JMX authentication secrets
          * Enable Cruise Control anomaly.detection configurations
          * Add support for building connector images from the Maven coordinates
          * Allow Kafka Connect Build artifacts to be downloaded from insecure servers
          * Add option to specify pull secret in Kafka Connect Build on OpenShift
          * Configurable authentication, authorization, and SSL for Cruise Control API
          * Allow to configure `/tmp` volume size via Pod template. By default `1Mi` is used.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:bd22a134b0c351ddfc78a3263850c2f16733c52f57ac92984cbe0ffff477c724
        - quay.io/strimzi/kafka@sha256:6367b0fbba9593e3ac736040ea2d7ef71296b1ec144d3b54d9370352141b90e4
        - quay.io/strimzi/kafka-bridge@sha256:d3899c79747e4cb3a6df5a9fcd80b81f038ca0634a97329c1d160499469876dd
        - quay.io/strimzi/jmxtrans@sha256:bff5d3bab575d0b9ed95dd79d2d22518cc00c4d1c9dbb5614ed16149d4920a1b
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/maven-builder@sha256:663eb81388ae8f824e7920c272f6d2e2274cf6c140d61416607261cdce9d50e2
        - quay.io/strimzi/operator@sha256:5d90caa66f57939ba62a9a50efa00fce319ba596e719a3f8a42170d7635ee3f4
        - quay.io/strimzi/kafka@sha256:8c022a8535742b2eea691737f863af5ddef1e8df08b415b8139ecd02a79052a5
        version: 0.26.1
      entries:
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.26.x
    - currentCSV: strimzi-cluster-operator.v0.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.0.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"3.0",
                        "inter.broker.protocol.version":"3.0"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.0.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:9ea49407a04006d0bddfaaee822adfb95f5857af32ffcee81ba60ce891b07ba8
          createdAt: "2022-01-14 20:00:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.27 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.27 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.27.1
          * Updated Log4j2 to version 2.17.1 and Fabric8 Kubernetes Client to 5.10.2
          ### New in 0.27.0
          * Updated Log4j2 to version 2.17.0 (adresses CVE-2021-44228, CVE-2021-45046, CVE-2021-45105)
          * (Experimental) support for AArch64 / ARM64 architecture
          * `ControlPlaneListener` and `ServiceAccountPatching` feature gates are now in the beta phase and are enabled by default
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:5682a6bfcc4ee13d53a30d587c8f96bc40abe3618880989336c5af274ea3cefa
        - quay.io/strimzi/kafka-bridge@sha256:55a52316aead02efa2004174e11ad4728bb8b6f1c1415f725f597f1d2f2da188
        - quay.io/strimzi/jmxtrans@sha256:d2d2477517eb1ac98b51ad6d64d8c1c578141637afa8d46037076f0013183c0b
        - quay.io/strimzi/kaniko-executor@sha256:a07ebe6600d6baba09980abdde2931b3b6abd2f75aeb2cac5c6a2d22a37aa807
        - quay.io/strimzi/maven-builder@sha256:9acd178cc504be8acaa70a39f3c17830aae1f0e93a656110790f9476b09260d5
        - quay.io/strimzi/kafka@sha256:f4d68bb94447c6612f70de7f6587e9e7ef712c83769ea1f11a8dbda0d241a059
        - quay.io/strimzi/kafka@sha256:a3ec567ca529df2fa41f8986d21e86dde66be94aa4baa56faa66db4dcb3b258e
        - quay.io/strimzi/operator@sha256:9ea49407a04006d0bddfaaee822adfb95f5857af32ffcee81ba60ce891b07ba8
        version: 0.27.1
      entries:
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.27.x
    - currentCSV: strimzi-cluster-operator.v0.28.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.1.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.1"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.1.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.1.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:8b010c5fb8244b742a37f264feefc116d75bb2abd2d6895375cbeea250727b39
          createdAt: "2022-02-12 23:02:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.28 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.28 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.28.0
          * Support for Kafka 3.1.0
          * Support for `StrimziPodSet` resources (disabled by default through the `UseStrimziPodSets` feature gate)
          * Support custom authentication mechanisms in Kafka listeners
          * Intra-broker disk balancing using Cruise Control
          * Add support for disabling the FIPS mode in OpenJDK
          * Fix renewing your own CA certificates
          * Added the option `createBootstrapService` to disable the creation of the bootstrap service for the Load Balancer Type Listener
          * Dependency updates (Strimzi Kafka Bridge, Strimzi OAuth, Open Policy Agent Authroizer plugin and others)
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:06e9bda15495feb2b92d9950e834cdeed6ef20064f021a879610f1b59dd46950
        - quay.io/strimzi/kafka-bridge@sha256:c345c783d8fbbee4edeca2ebc47e7c454fcc2e26686820b8037de97369280ba0
        - quay.io/strimzi/operator@sha256:8b010c5fb8244b742a37f264feefc116d75bb2abd2d6895375cbeea250727b39
        - quay.io/strimzi/jmxtrans@sha256:3a55a81790dc88b90d1bf858e2cfd7e591ad0e79f763c6d1667b2d3eb0d0c79d
        - quay.io/strimzi/kaniko-executor@sha256:a07ebe6600d6baba09980abdde2931b3b6abd2f75aeb2cac5c6a2d22a37aa807
        - quay.io/strimzi/maven-builder@sha256:3d5d664e089651944712361a6d8ba42b578dbd9bb3dfd9972f614520c9cfab75
        - quay.io/strimzi/kafka@sha256:7c04f65e216bb941458aeadffb85e9eeee43d5ca67da566017c2c93a0664c98e
        version: 0.28.0
      entries:
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.28.x
    - currentCSV: strimzi-cluster-operator.v0.29.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:afde94c6d2544207cb7f3c94b54455643558a9d5daaa13b9f736cc8da20e319c
          createdAt: "2022-05-15 00:01:07"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.29 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.29 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.29.0
          * Add support for Apache Kafka 3.0.1, 3.1.1 and 3.2.0
          * Renew user certificates in User Operator only during maintenance windows
          * Remove TLS sidecar from Cruise Control pod. Cruise Control is now configured to not using ZooKeeper, so the TLS sidecar is not needed anymore.
          * Allow Cruise Control topic names to be configured
          * Add support for `spec.rack.topologyKey` property in Mirror Maker 2 to enable "fetch from the closest replica" feature.
          * Experiemntal support for the s390x platform
          * Added new rebalancing modes on the `KafkaRebalance` custom resource to add or remove brokers
          * Experimental KRaft mode (ZooKeeper-less Kafka). Use it for development and testing only!
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/jmxtrans@sha256:f89e3c9618430908a8b390f2490f5aa43687b12889b434472a962ce0170c6463
        - quay.io/strimzi/maven-builder@sha256:0c096a59faf8fd224298c01415a895043db4647534c060cb7e53207e9448bf65
        - quay.io/strimzi/kafka@sha256:ac190141120c1800cc7131d3a1f701e64efdcd4a5d87f0b2e0e2cadfdef8b438
        - quay.io/strimzi/kafka@sha256:95678e843bc87cbe571f3f299045dd075725816b59e1b87821cd6025876ec559
        - quay.io/strimzi/kafka@sha256:284cde40784b6ff7bfb81b6dd28928e1e8c598fe6a72504f025592e5b43fa94c
        - quay.io/strimzi/kafka@sha256:9dfc09273ad8fc4bdc75cbe93b4a061a05a5defb5884a221ae1e178ba602787f
        - quay.io/strimzi/kafka-bridge@sha256:b7bc3dd90e87223766f608a0087b521f0bba79d1c21eaa833bc2708c71148b7f
        - quay.io/strimzi/kaniko-executor@sha256:544c4987f7f3ca558b04e9ef4357be204f088212b7b5b9ed0f909401a8bc5238
        - quay.io/strimzi/operator@sha256:afde94c6d2544207cb7f3c94b54455643558a9d5daaa13b9f736cc8da20e319c
        - quay.io/strimzi/kafka@sha256:935caab03f1d0365b47ddefec0c1d8fa449fc3de715ccdd298aedef5ce968c73
        version: 0.29.0
      entries:
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.29.x
    - currentCSV: strimzi-cluster-operator.v0.30.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:587db6f4389f56209f48853beedbfe249f6d590a4cb0000be2c42b57e43c2adb
          createdAt: "2022-07-16 00:00:10"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: "Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org)
          cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/)
          in various deployment configurations. See our [website](https://strimzi.io)
          for more details about the project.\n### CRD Upgrades\n**!!! IMPORTANT !!!**
          \ This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.
          \ If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed
          for all Strimzi CRDs and CRs before the upgrade to 0.30 is done!  If upgrading
          from Strimzi version earlier than 0.22, you need to first install the CRDs
          from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi
          CRDs and CRs before the upgrade to 0.30 is done!  For more details about
          the CRD upgrades, see the documentation.\n### New in 0.30.0\n* The `UseStrimziPodSets`
          feature gate moves to beta stage. By default, StrimziPodSets are used instead
          of StatefulSets. If needed, `UseStrimziPodSets` can be disabled in the feature
          gates configuration in the Cluster Operator.\n* The `ServiceAccountPatching`
          feature gate moves to GA. It cannot be disabled anymore and will be permanently
          enabled.\n* Remove support for Kafka 3.0.0 and 3.0.1\n* Add support for
          `simple` authorization and for the User Operator to the experimental `UseKRaft`
          feature gate. _(Note: Due to [KAFKA-13909](https://issues.apache.org/jira/browse/KAFKA-13909),
          broker restarts currently don't work when authorization is enabled.)_\n*
          Add CPU and network capacity overrides for Cruise Control capacity config\n*
          Use better encryption and digest algorithms when creating the PKCS12 stores.
          For existing clusters, the certificates will not be updated during upgrade
          but only next time the PKCS12 store is created. \n* Operator emits Kafka
          events to explain why it restarted a Kafka broker\n* Better configurability
          of the Kafka Admin client in the User Operator\n* Update Strimzi Kafka Bridge
          to 0.21.6\n### Supported Features\n* **Manages the Kafka Cluster** - Deploys
          and manages all of the components of this complex application, including
          dependencies like Apache ZooKeeper® that are traditionally hard to administer.\n*
          **Includes Kafka Connect** - Allows for configuration of common data sources
          and sinks to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for mirroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashboards\n### Upgrading your Clusters\nThe
          Strimzi Operator understands how to run and upgrade between a set of Kafka
          versions. When specifying a new version in your config, check to make sure
          you aren't using any features that may have been removed. See [the upgrade
          guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str)
          for more information.\n### Storage\nAn efficient data storage infrastructure
          is essential to the optimal performance of Apache Kafka®. Apache Kafka®
          deployed via Strimzi requires block storage. The use of file storage (for
          example, NFS) is not recommended.\nThe Strimzi Operator supports three types
          of data storage:\n* Ephemeral (Recommended for development only!)\n* Persistent\n*
          JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nStrimzi
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation to the current _main_
          branch as well as all releases can be found on our [website](https://strimzi.io/documentation).\n###
          Getting help\nIf you encounter any issues while using Strimzi, you can get
          help using:\n* [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          Contributing\nYou can contribute by:\n* Raising any issues you find using
          Strimzi\n* Fixing issues by opening Pull Requests\n* Improving documentation\n*
          Talking about Strimzi\nAll bugs, tasks or enhancements are tracked as [GitHub
          issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues
          which  might be a good start for new contributors are marked with [\"good-start\"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start)
          label.\nThe [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md)
          describes how to build Strimzi and how to  test your changes before submitting
          a patch or opening a PR.\nThe [Documentation Contributor Guide](https://strimzi.io/contributing/guide/)
          describes how to contribute to Strimzi documentation.\nIf you want to get
          in touch with us first before contributing, you can use:\n* [Strimzi mailing
          list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n* [Strimzi
          Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          License\nStrimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).\n"
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:d9258ddd453c3a53ca8e737b989d27725818da95f7a4fc879c9f07fe743c0cef
        - quay.io/strimzi/kafka@sha256:5d881605425ab0b37d0b9cac6957d00ba904df0db185ca6181c810151bdab8e7
        - quay.io/strimzi/kafka-bridge@sha256:77fc68da171c74d3a4b75b1030eaaa5eae12e200002f7aa2ee1969385eba1d66
        - quay.io/strimzi/jmxtrans@sha256:de1ca88b713b926b72df535aca2cfad24fb1b2bf40481065d6fa0c421b872320
        - quay.io/strimzi/kaniko-executor@sha256:544c4987f7f3ca558b04e9ef4357be204f088212b7b5b9ed0f909401a8bc5238
        - quay.io/strimzi/maven-builder@sha256:f2cb45234f3a7a010824a370d03032b706271cd0d1d260da929f046f9259d7e0
        - quay.io/strimzi/kafka@sha256:68d6f4cb5cc5ccad44de799bd9cf417c469a09e1024dbe42298942f25a09e004
        - quay.io/strimzi/operator@sha256:587db6f4389f56209f48853beedbfe249f6d590a4cb0000be2c42b57e43c2adb
        version: 0.30.0
      entries:
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.30.x
    - currentCSV: strimzi-cluster-operator.v0.31.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.3",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.3",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:6e5e207c70c25c595d3ce60268c0214c7518e95836c9b1212ef1eb2f008c7a76
          createdAt: "2022-09-20 21:07:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          **This is the last version of Strimzi with support for Kubernetes 1.16, 1.17 and 1.18!**
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.31 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.31 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.31.1
          * Kafka 3.1.2 and 3.2.3 (fixes CVE-2022-34917)
          * Make `sasl.server.max.receive.size` broker option user configurable
          * Update Strimzi Kafka Bridge to 0.22.1
          ### New in 0.31.0
          * Support for Apache Kafka 3.2.1
          * Pluggable Pod Security Profiles with built-in support for _restricted_ Kubernetes Security Profile
          * Support for leader election and running multiple operator replicas (1 active leader replicas and one or more stand-by replicas)
          * Support for IPv6 addresses being used in Strimzi issued certificates
          * `StrimziPodSet` reconciliation metrics
          * Updated Strimzi Kafka Bridge to 0.22.0 and Kafka Exporter to 1.6.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/jmxtrans@sha256:52dd107062d78c29bb58c905ac654c39df72bbe3c25613023aa33e834930606d
        - quay.io/strimzi/kafka@sha256:9553aae22d35448681cbbf7a831a28149fb159f3df58ccf8c8ec4ca7eb9c1fdf
        - quay.io/strimzi/kafka@sha256:b7c059bd72d8b063a96dde788d91d1ad87498bad3e73acba600929d595aeb7d6
        - quay.io/strimzi/kafka@sha256:430b3eb70e47a91d7b5b6f230d5be024c2de853099cc0437e219b443e0ead9f5
        - quay.io/strimzi/kafka@sha256:652924a1dba46aa716caba8381f1e791c09035e31ce62fdcc81f9b69af36cd02
        - quay.io/strimzi/kafka@sha256:03cc48bc5cf367f898338665b253855bed7d6096fe577adfe344da2f2cddcfe4
        - quay.io/strimzi/kaniko-executor@sha256:06581fb615f0b13ddfe27fae26ba5836b12c6e28afd55a81cba7685c9ca4bf79
        - quay.io/strimzi/maven-builder@sha256:670ca70900f2229ef140a592b77f2107934601f8b050b79250eab17ff62099b6
        - quay.io/strimzi/operator@sha256:6e5e207c70c25c595d3ce60268c0214c7518e95836c9b1212ef1eb2f008c7a76
        - quay.io/strimzi/kafka-bridge@sha256:3ddac62a4f8c8a3ebb72d5841d3826f7fd3fbb0e00eb75fb40bced0a576b86a9
        - quay.io/strimzi/kafka@sha256:83acc1e2cc1e64a88bdbb54b9818a6ad3658c3da60e72c4be365f824c5fcf57c
        version: 0.31.1
      entries:
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.31.x
    - currentCSV: strimzi-cluster-operator.v0.32.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.3.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.3.1",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:c5e0e5dca7503a8371b8d366db0f49e3fc68f62ea480262e39b6cc277991c97c
          createdAt: "2022-10-31 17:05:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **From the 0.32.0 release, Strimzi supports only Kubernetes 1.19 and newer! Kubernetes 1.16, 1.17 and 1.18 are not supported anymore.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.32.0/deploying.html#assembly-upgrade-str).
          ### New in 0.32.0
          * Add support for Kafka 3.3.1 and remove support for Kafka 3.1.0, 3.1.1, and 3.1.2
          * Update KafkaConnector CR status so the 'NotReady' condition is added if the connector or any tasks are reporting a 'FAILED' state.
          * Add auto-approval mechanism on KafkaRebalance resource when an optimization proposal is ready
          * The ControlPlaneListener feature gate moves to GA
          * Add client rack-awareness support to Strimzi Bridge pods
          * Add support for OpenTelemetry for distributed tracing
          * ZookeeperRoller considers unready pods
          * Support multiple operations per ACLRule
          * Add new listener type "cluster-ip" which is using per-broker services instead of the pod DNS names
          * Update Strimzi OAuth library to 0.11.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/maven-builder@sha256:593e9980b786ca8ad3223151c3e17f16c7652ee0e86d722af4c34b023038ecc2
        - quay.io/strimzi/kafka-bridge@sha256:aad0cee4381cc1c8c6638ab5732b135acd271b5889486ca9a33312cce67e4a02
        - quay.io/strimzi/operator@sha256:c5e0e5dca7503a8371b8d366db0f49e3fc68f62ea480262e39b6cc277991c97c
        - quay.io/strimzi/kafka@sha256:680ae1958dbcb9da8ee4128a67c1163a6ee2744221f7d29f73d8bcc237fd0173
        - quay.io/strimzi/jmxtrans@sha256:208928133b2854f60bb34912e07561a3511de895b815960e73d565be2d553918
        - quay.io/strimzi/kafka@sha256:bb97008b7f928ed98939b2008a7baf8909c88e726a59aaf607f2eeba55375cc0
        - quay.io/strimzi/kafka@sha256:8c9f38b45dd814e617c9c51cc90fd51e23d7dd23e8efe7e9eff013522bd6a5a0
        - quay.io/strimzi/kafka@sha256:4bacf5b54041bd0ab9199c8c5d51e8058758139db4b7abc380cf0400737716ef
        version: 0.32.0
      entries:
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.32.x
    - currentCSV: strimzi-cluster-operator.v0.33.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:cf0d41fdb99b20e91063e23eb1d664817c18866c9a2f5c4430eac897e78d5d53
          createdAt: "2023-02-18 21:21:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.33 supports only Kubernetes 1.19 and newer! Kubernetes versions 1.16, 1.17 and 1.18 are not supported anymore since Strimzi 0.32.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.33.2/deploying.html#assembly-upgrade-str).
          ### New in 0.33
          _Note: Strimzi 0.33.2 add support for Kafka 3.4.0 which fixes CVE-2023-25194._
          * Add support for Kafka 3.3.2
          * Automatic restarts of failed Connect or Mirror Maker 2 connectors
          * Redesign of Strimzi User Operator to improve its scalability
          * Move to Java 17
          * Improved FIPS support
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:fe9d3f8b029ff32d5d725e57384592914957314d63e92fda3abf4028fe4e66f6
        - quay.io/strimzi/kafka@sha256:116cf543fbadadd0b45007ce24d44a0c79a62280e8faa1fc80df40291fbdf05d
        - quay.io/strimzi/kafka@sha256:301dc52ed5891ed4fa33e79ca3c344b1ac0eec7469dd46a5a7ac22064e053fc2
        - quay.io/strimzi/kafka@sha256:9eca149cfc436580e3cdd6cee0ca318366b8d0ec53c2cf007e4e2c614b988921
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/maven-builder@sha256:de9e51365e7acc63feae745618804e6286fd730f41c3b31582cbb7319a93b483
        - quay.io/strimzi/kafka@sha256:30945aecb9d63cf8863c672222f130469c89aefa07078c95aaed52c40fd16146
        - quay.io/strimzi/kafka@sha256:40a847f8455ec6ea9ff9fe1fff20b1ed01f848326dc13977d13ac911f0cb40b7
        - quay.io/strimzi/operator@sha256:cf0d41fdb99b20e91063e23eb1d664817c18866c9a2f5c4430eac897e78d5d53
        - quay.io/strimzi/kafka-bridge@sha256:3a380d5f3ea509f90cdf95280d83b42f5de4f86ee3a512e02234e7639f50c9eb
        - quay.io/strimzi/jmxtrans@sha256:5ebdafa7a5e3718ed3274d6254c9e655613c272340094b997b1c97fde2598b4a
        version: 0.33.2
      entries:
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.33.x
    - currentCSV: strimzi-cluster-operator.v0.34.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:f4ee3ca482a2dd19785f59634578a0850273444c1173a5699a99cf02f3a018fe
          createdAt: "2023-03-17 13:02:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.34 supports only Kubernetes 1.19 and newer! Kubernetes versions 1.16, 1.17 and 1.18 are not supported anymore since Strimzi 0.32.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.34.0/deploying.html#assembly-upgrade-str).
          ### New in 0.34
          * Add support for Apache Kafka 3.4.0 (fixes CVE-2023-25194) and remove support for Apache Kafka 3.2.x
          * Stable Pod identities for Kafka Connect and MirrorMaker 2 (disabled by default)
          * Use JDK HTTP client in the Kubernetes client instead of the OkHttp client
          * Add truststore configuration for HTTPS connections to Open Policy Agent server
          * Update Strimzi HTTP Bridge to 0.25.0 and strimzi Oauth to 0.12.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/jmxtrans@sha256:62272e8a03da2144340bd7267f7a824d1ed70f4f3e30dc7aadfba9ee1b9baf76
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/maven-builder@sha256:f781b6b857d437c2de62585462e97b9fae357912c9f6711c1df39f431278878d
        - quay.io/strimzi/kafka@sha256:1da35fd7b36b113d1343e254fdc6efd97ca8660cc7977d8985aa3a3547d5ebb1
        - quay.io/strimzi/kafka@sha256:1ace2cc0db9653ac1b388477fa56c3c1cd97c7b3846eba3c7824dc58ea179476
        - quay.io/strimzi/kafka@sha256:37cf8d0fea1d2078511920190a72b49977ff4047b6ecbfed6b394c6f9e1e2ff0
        - quay.io/strimzi/operator@sha256:f4ee3ca482a2dd19785f59634578a0850273444c1173a5699a99cf02f3a018fe
        - quay.io/strimzi/kafka-bridge@sha256:51babca0f7a328f4db005c6a16daa8ac4a4306c46fe3ca89382919134aa09179
        version: 0.34.0
      entries:
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.34.x
    - currentCSV: strimzi-cluster-operator.v0.35.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:06a94a3021cf028ccc1a49271f35f79216029e344536e664f196c1725ff2c663
          createdAt: "2023-06-07 16:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.35 is the last version with support for Kubernetes 1.19 and 1.20. From Strimzi 0.36, only Kubernetes 1.21 and newer will be supported.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.35.1/deploying.html#assembly-upgrade-str).
          ### New in 0.35
          * Move feature gate `UseStrimziPodSets` to GA and remove support for StatefulSets
          * Remove support for JMX Trans
          * Redesigned the Cluster and User Operator configuration to make it more efficient and flexible
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:39778b90c2b2afc30261e4ad5135805e1a10a2b60e2e53108fb9f80487f1208a
        - quay.io/strimzi/maven-builder@sha256:88a79eff3b3a386880a630658964b7754caed9e99dd6e645a4c0d23d0fdb47ee
        - quay.io/strimzi/kafka@sha256:4de4874a7b722ad813f4dcc58acf509527bca0609999b81e70d81e3b38534d9d
        - quay.io/strimzi/kafka@sha256:0d910e7138cb49e1cd8cd84cef88bce35698b93ddd683a3398f1d485a3162693
        - quay.io/strimzi/kafka@sha256:54c6b25b31f51ef401c1b6e2a1b27432911e819bf3e502e05186f01be3f798e5
        - quay.io/strimzi/kafka-bridge@sha256:d6be183e492f8f88157ab9fe0af53950df8b6711a8a8c33da465de6064f6f86e
        - quay.io/strimzi/operator@sha256:06a94a3021cf028ccc1a49271f35f79216029e344536e664f196c1725ff2c663
        version: 0.35.1
      entries:
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.35.x
    - currentCSV: strimzi-cluster-operator.v0.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:e9e03b31007c0f9b67a6a33e6b644e8c1ba35bd9976804ee3d22a04e4bf511b1
          createdAt: "2023-07-10 21:47:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!! Strimzi 0.36.1 supports only Kubernetes 1.21 and newer! Kubernetes versions 1.19 and 1.20 are not supported anymore since Strimzi 0.36.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.35.1/deploying.html#assembly-upgrade-str).
          ### New in 0.36
          * Add support for Apache Kafka 3.4.1, 3.5.0, and 3.5.1
          * Remove support for 3.3.1 and 3.3.2
          * Add support for _Kafka node pools_
          * Add support for _Unidirectional Topic Operator_
          * Enable SCRAM-SHA authentication in KRaft mode (supported in Apache Kafka 3.5.0 and newer)
          * Add support for insecure flag in Maven artifacts in Kafka Connect Build
          * Improve Kafka rolling update to avoid rolling broker in log recovery
          * Added support for Kafka Exporter topic exclude and consumer group exclude parameters
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:e9e03b31007c0f9b67a6a33e6b644e8c1ba35bd9976804ee3d22a04e4bf511b1
        - quay.io/strimzi/kaniko-executor@sha256:5ff9fc7d0b7b694875ad0150fc787c582d42da3b4864d807ee8a6524e47b0e4e
        - quay.io/strimzi/maven-builder@sha256:2a469b40122027b981914db5b5c102be48280e039c3089963422c181174af226
        - quay.io/strimzi/kafka@sha256:a0da6a48b5bcbff355365f07df2d99ef6443d22c5690495f10a6417e761b3799
        - quay.io/strimzi/kafka@sha256:9f35f40be3d47b0cb1f6628f7ffd9c6a687ab9f564829f341dd66fd0759c064f
        - quay.io/strimzi/kafka@sha256:67c2b563a65c85c9393608f9af7eaf12ea48dcc0ef5f726d51f27273a3845186
        - quay.io/strimzi/kafka@sha256:ea35817ff1ef295c682d70562f209ef658fcd87c61f467a4549d7a2f39e71700
        - quay.io/strimzi/kafka-bridge@sha256:fefee252895e9f94756ed5c9ea71749e3fa89281df01bc55878a244ca7d17697
        version: 0.36.1
      entries:
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.36.x
    - currentCSV: strimzi-cluster-operator.v0.37.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:52f376e64b9b31d74162e26248c278e9bd301d2d8b4240ee5bcf780dfe2f8ef8
          createdAt: "2023-09-03 17:45:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.37.0/deploying.html#assembly-upgrade-str).
          ### New in 0.37
          * The StableConnectIdentites feature gate moves to a beta stage.
          * Support for the ppc64le platform
          * Version fields to the Kafka custom resource status to track installation and upgrade state
          * Infinite auto-restarts of Kafka Connect and Kafka Mirror Maker 2 connectors
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:52f376e64b9b31d74162e26248c278e9bd301d2d8b4240ee5bcf780dfe2f8ef8
        - quay.io/strimzi/kafka@sha256:dfeb1eb7b94107c443b6bc82fcfee7fcf782516c83c6fc64451fb6a7e2de74a6
        - quay.io/strimzi/kafka-bridge@sha256:5b821f6f3396002ffa7c5804391b627ff9da3b494b2777dde518c55c2b73cefc
        - quay.io/strimzi/kaniko-executor@sha256:5729f1f0511da95e64e358cfabb3980ad9a660af2968118a87d6755f5d652a67
        - quay.io/strimzi/maven-builder@sha256:995be06a50b86c992fcee9f00e44f6a21c3406975e4e35802f10e88b72ffb9df
        - quay.io/strimzi/kafka@sha256:38bbadcdb196a373e1698be890caa1dc01134ce91763d8f95c71768a117e1932
        - quay.io/strimzi/kafka@sha256:1de6e073e8625a193fcfd11dd1a6717cc17596911c10bcebd792c00b08a7ebaf
        - quay.io/strimzi/kafka@sha256:2997ba1c169eb792a6004d7a41b674560515f3ea6e62ab8b9a1fdb747eab40c1
        version: 0.37.0
      entries:
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.37.x
    - currentCSV: strimzi-cluster-operator.v0.38.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:aee45bbe439335e45a0423e6565bae0d0727d85c7b807812949bfa127d467fe6
          createdAt: "2023-10-14 18:18:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.38.0/deploying.html#assembly-upgrade-str).
          ### New in 0.38
          * Add support for Apache Kafka 3.6.0 and remove support for 3.4.0 and 3.4.1
          * Container signing and SBOMs
          * Stopping Connect and MirrorMaker 2 connectors
          * Manual rolling updates of Kafka Connect and Kafka Mirror Maker 2 pods using annotation
          * Prevent scale-down of brokers that still contain any replicas
          * Run ZooKeeper- and KRaft-based clusters in parallel
          * Unidirectional Topic Operator improvements
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:20d451fb32ec82eab3c16891198ace46bb9b9193f1f5511bb60ce249080dd7d4
        - quay.io/strimzi/kafka-bridge@sha256:13838889cc20a3a382ea64e749cad94f997c0b928840588f2d1befcf19247bea
        - quay.io/strimzi/kaniko-executor@sha256:4fe7a3f3f78e93369284a5ab456873c911230926ca071d1d88167e63d8ce376c
        - quay.io/strimzi/maven-builder@sha256:922d91dab9e2b8e14b6db25d8b48d7f4e3521bedd435468cc22039599ac4f888
        - quay.io/strimzi/kafka@sha256:50e24c797455f9cdbb3cdde8e4f26bb8e1f0503469456da5d179017fd3e03ffe
        - quay.io/strimzi/operator@sha256:aee45bbe439335e45a0423e6565bae0d0727d85c7b807812949bfa127d467fe6
        - quay.io/strimzi/kafka@sha256:0cc01438e8596643407b6b8ff05fafcea7ef54fa9410d4ae2492a9dbf41118b1
        version: 0.38.0
      entries:
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.38.x
    - currentCSV: strimzi-cluster-operator.v0.39.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.6.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:50b49f9b5e554cca517fc2794e11319d5aaba1676494b285baa3a74fc79eb501
          createdAt: "2023-12-14 20:28:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.39.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.39 is the last minor release with support for Kubernetes 1.21 and 1.22.** From Strimzi 0.40 on, only Kubernetes 1.23 and newer will be supported.
          ### New in 0.39
          * Add support for Apache Kafka 3.5.2 and 3.6.1
          * The `StableConnectIdentities` feature gate moves to GA stage and is now permanently enabled without the possibility to disable it.
            All Connect and Mirror Maker 2 operands will now use StrimziPodSets.

          * The `KafkaNodePools` feature gate moves to the beta stage and is enabled by default.
            If needed, `KafkaNodePools` can be disabled in the feature gates configuration in the Cluster Operator.

          * The `UnidirectionalTopicOperator` feature gate moves to the beta stage and is enabled by default.
            If needed, `UnidirectionalTopicOperator` can be disabled in the feature gates configuration in the Cluster Operator.

          * Improved Kafka Connect metrics and dashboard example files
          * Allow specifying and managing KRaft metadata version
          * Add support for KRaft to KRaft upgrades (Apache Kafka upgrades for the KRaft-based clusters)
          * Improved Kafka Mirror Maker 2 dashboard example file
          * Support for rolling updates of KRaft controller nodes
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:13838889cc20a3a382ea64e749cad94f997c0b928840588f2d1befcf19247bea
        - quay.io/strimzi/kaniko-executor@sha256:723b7add2473e73c45c81361e1708af4971c321b0391c1298a0557fb8ca4f7f6
        - quay.io/strimzi/kafka@sha256:84691b3afb67b312400e01572ff08b64c1e1b1d32a4276e0c21bba10c71c36fe
        - quay.io/strimzi/kafka@sha256:dd1706f36031c96dfcd0da7b3f804168e5f076220caa302973455fdcd76bd03e
        - quay.io/strimzi/kafka@sha256:2badc65532b483bba6c1f6a8b0c9895e993ce24e9ebd3cd953d0d17fbe36d27a
        - quay.io/strimzi/kafka@sha256:53b960411e852f5cb182ecc5d395a3562865f83e15e6251f3a9af0eb413feeed
        - quay.io/strimzi/kafka@sha256:49b7530ada1175750ad5a706e545ba3594d85e1e3a90f98b0200a73fa9875057
        - quay.io/strimzi/operator@sha256:50b49f9b5e554cca517fc2794e11319d5aaba1676494b285baa3a74fc79eb501
        - quay.io/strimzi/maven-builder@sha256:6ce3c16284379d5db9e42a6f3e2b765c981091164a28fdd6bf927b6c5d4ab557
        version: 0.39.0
      entries:
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.39.x
    - currentCSV: strimzi-cluster-operator.v0.40.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:2228d89b53e3524e69d1bcf5c2cd373884ce696d9005389a6c78dad23310e837
          createdAt: "2024-03-09 21:27:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.40.0/deploying.html#assembly-upgrade-str).

          **From Strimzi 0.40 on, we support only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.40
          * Add support for Apache Kafka 3.7.0. Remove support for Apache Kafka 3.5.0, 3.5.1, and 3.5.2.
          * The `UseKRaft` feature gate moves to beta stage and is enabled by default. If needed, `UseKRaft` can be disabled in the feature gates configuration in the Cluster Operator.
          * Add support for ZooKeeper to KRaft migration by enabling the users to move from using ZooKeeper to store metadata to KRaft.
          * Add support for moving from dedicated controller-only KRaft nodes to mixed KRaft nodes
          * Added support for Kafka Exporter `offset.show-all` parameter
          * Prevent removal of the `broker` process role from KRaft mixed-nodes that have assigned partition-replicas
          * Improve broker scale-down prevention to continue in reconciliation when scale-down cannot be executed
          * Added support for Tiered Storage by enabling the configuration of custom storage plugins through the Kafka custom resource.
          * Update HTTP bridge to the latest 0.28.0 release
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:2228d89b53e3524e69d1bcf5c2cd373884ce696d9005389a6c78dad23310e837
        - quay.io/strimzi/kaniko-executor@sha256:3fb4ace22589fcf8d3f01b74c37b10b32d87e15d103ea59e0933b61c4e46ff4a
        - quay.io/strimzi/maven-builder@sha256:d02a9b48c77a44eeab6426c9af4adf75405217994bcb01ce0196e616041350d8
        - quay.io/strimzi/kafka@sha256:2c143eb10b66037ef877311eb767457e4b877d5ff2c47d393c76babec155caff
        - quay.io/strimzi/kafka@sha256:e2b9f2835498831bbf30ac2d651de143d57088bcb23943101b066e941db92aca
        - quay.io/strimzi/kafka@sha256:c491a95f9b9b58f406461faf222ef612ad8d2f4d7b53f7f3645b20c99d7d1583
        - quay.io/strimzi/kafka-bridge@sha256:73deb1af0d93e9fa8955a5a412d0e7138384bdc80e7574e815d21796e446b636
        version: 0.40.0
      entries:
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.40.x
    - currentCSV: strimzi-cluster-operator.v0.41.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:4c687a4fac96ffc22433ca6d00fb1bd31722767b44c9d87c465f95b57fb70089
          createdAt: "2024-05-09 22:23:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.41.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.41 supports only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.41
          * Support for Apache Kafka 3.6.2
          * Metrics to monitor CA expiration
          * Support for JBOD storage in KRaft mode. (Note: JBOD support in KRaft mode is considered early access in Apache Kafka 3.7.x)
          * Support for changing topic replication factor by changing the `KafkaTopic` resources.
          * The `KafkaNodePools` feature gate moves to GA stage and is permanently enabled without the possibility to disable it. To use the Kafka Node Pool resources, you still need to use the `strimzi.io/node-pools: enabled` annotation on the `Kafka` custom resources.
          * Support for configuring the `externalIPs` field in node port type services.
          * The `UnidirectionalTopicOperator` feature gate moves to GA stage and is permanently enabled without the possibility to disable it. If the topics whose names start with `strimzi-store-topic` and `strimzi-topic-operator` still exist, you can delete them.
          * Improve validation of `KafkaMirrorMaker2` resource to make sure the Connect cluster is set to the target cluster
          * Continue reconciliation after failed manual rolling update using the `strimzi.io/manual-rolling-update` annotation (when the `ContinueReconciliationOnManualRollingUpdateFailure` feature gate is enabled)
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:9404037bd5431b012c8c633ec3345907667453c5532e5bfaaedc78286de22c62
        - quay.io/strimzi/operator@sha256:4c687a4fac96ffc22433ca6d00fb1bd31722767b44c9d87c465f95b57fb70089
        - quay.io/strimzi/maven-builder@sha256:17f09b07fc41a059e6a048cb766046d43a072cfd3a8266c97388a37543dadee8
        - quay.io/strimzi/kafka@sha256:c3a346235ce4dea42986648ca07f2482d8fc0bf654c1d441b1301a23999304db
        - quay.io/strimzi/kafka@sha256:bf49499980bbe5d4dd399de0dd396089b0fd9dd29bc46e14d39b20f1e826aec1
        - quay.io/strimzi/kafka@sha256:41811bb106d6182978a889eaedffeadffaa5ab1599c4cd8e4f5015b131c7ac30
        - quay.io/strimzi/kafka@sha256:6243d98890c665ef65b5f5fa2ffd046412281ffef31bc7a07cac37bd77b2579d
        - quay.io/strimzi/kafka-bridge@sha256:c833f930ead2d1e9054912a20689ee979d335f5e5ca8769e3c6b8a5514a0df96
        version: 0.41.0
      entries:
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.41.x
    - currentCSV: strimzi-cluster-operator.v0.42.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:ca5c1b70e1196deec4bf519921bfa55a967aa1ea15bd4c182bc56bea08d2e114
          createdAt: "2024-07-03 12:34:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.42.0/deploying.html#assembly-upgrade-str).

          **From version 0.40.0, Strimzi supports only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.42
          * Support for Apache Kafka 3.7.1
          * The UseKRaft feature gate moves to GA stage and is permanently enabled without the possibility of disabling it. To use KRaft (ZooKeeper-less Apache Kafka), you still need to use the strimzi.io/kraft: enabled annotation on the Kafka custom resources or migrate from an existing ZooKeeper-based cluster.
          * Update the base image used by Strimzi containers from UBI8 to UBI9
          * Add support for filename patterns when configuring trusted certificates
          * Enhance KafkaBridge resource with consumer inactivity timeout and HTTP consumer/producer enablement.
          * Add support for setting publishNotReadyAddresses on services for listener types other than internal.
          * Added support for configuring the quotas plugin with type strimzi or kafka in the Kafka custom resource. The Strimzi Quotas plugin version was updated to 0.3.1.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:ca5c1b70e1196deec4bf519921bfa55a967aa1ea15bd4c182bc56bea08d2e114
        - quay.io/strimzi/kafka@sha256:c56c5abf94897f20f976603cba0678a33cdfbbcd6f9030a193f7684cb9894a6c
        - quay.io/strimzi/kafka@sha256:6c519531cdd5f40f326b295d8827344140efe2607a1826f7f1cab8c3427f43f7
        - quay.io/strimzi/kaniko-executor@sha256:11821b0078a270612f1bcc37f5dc716ab273c974648c40f95fa4804352e647a7
        - quay.io/strimzi/kafka@sha256:841a46e9c10481df3ef702412caf362b55ae07c30a3a5927073356cd36dc049f
        - quay.io/strimzi/kafka-bridge@sha256:fdddffb3de59ba48c05964bb0e760593088d24d1ee35b090112fe6cc6b553cd8
        - quay.io/strimzi/kafka@sha256:6527c426f753521c07f3a3df2ee432273a41b109594096b4c6cae2e8e609d669
        - quay.io/strimzi/kafka@sha256:bc49c10b3e839e161e515c3a1e411d1f8df9de5e2c8d71aa7b63ccdbbae9e8d2
        - quay.io/strimzi/maven-builder@sha256:342e8d0086c53cf7334127798e5bd3764b28f5e4909ba67d65f6d937dc3dba05
        version: 0.42.0
      entries:
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.42.x
    - currentCSV: strimzi-cluster-operator.v0.43.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:d3732245f34afaa6b571f1f912e1cd017891b22003d6367ae8ca059b2a5cb13c
          createdAt: "2024-08-22 15:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.43.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.43 is the last version with support for Kubernetes 1.23 and 1.24.** From Strimzi 0.44, only Kubernetes versions 1.25 and newer will be supported.
          ### New in 0.43
          * Support for Apache Kafka 3.8.0
          * Removed support for Apache Kafka 3.6.0, 3.6.1, and 3.6.2
          * Specifying additional volumes and volume mounts in Strimzi custom resources
          * Support for custom Cruise Control API users
          * Unregistration of KRaft nodes after scale-down
          * Kafka Exporter updated to 1.8.0 (and the Grafana dashboard updated to work with it)
          * Service account based authentication
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:6414f8f3985f221b0eda53867c00494140108d17579c4dcbbc93934bb00e5d49
        - quay.io/strimzi/operator@sha256:d3732245f34afaa6b571f1f912e1cd017891b22003d6367ae8ca059b2a5cb13c
        - quay.io/strimzi/kafka@sha256:d27cf5bf3624e18ba705246b14f92d12b8a15eb8551b13abf027cf71b549baf0
        - quay.io/strimzi/kafka@sha256:ad06ead5125bce87498b1582a76e78143a6d21408ac4b12bb7e6007aa4cabd84
        - quay.io/strimzi/kafka@sha256:b2f4e8895e9ed66228512ce14aec685101fa75dc329128d47cb1c15e3aef3c9c
        - quay.io/strimzi/kafka-bridge@sha256:80f329ecd3ba96fac12e75be874dce0debc8395337737581de34238ce11c8fb1
        version: 0.43.0
      entries:
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.43.x
    - currentCSV: strimzi-cluster-operator.v0.44.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:b07b81f7e282dea2e4e29a7c93cfdd911d4715a5c32fe4c4e3d7c71cba3091e8
          createdAt: "2024-08-22 15:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.44.0/deploying.html#assembly-upgrade-str).

          **From version 0.44, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**

          **ZooKeeper support will be soon removed from Apache Kafka and Strimzi.** Currently, the last Strimzi version with ZooKeeper support is expected to be Strimzi 0.45. Please plan your migration to KRaft (ZooKeeper-less Apache Kafka) accordingly. Follow the [documentation](https://strimzi.io/docs/operators/latest/full/deploying.html#assembly-kraft-mode-str) for more details.

          **Kafka Mirror Maker 1 support will be soon removed from Apache Kafka and Strimzi.** Currently, the last Strimzi version with Mirror Maker 1 support is expected to be Strimzi 0.45. Please plan your migration to Mirror Maker 2 or another mirroring tool.

          ### New in 0.44
          * Managing offsets in Connect and Mirror Maker 2 connectors through custom resources
          * Templating `advertisedHost` and `host` fields in listener configuration
          * Configuring environment variables in operands based on Secrets and ConfigMaps
          * Disabling Pod disruption Budget management
          * Auto-rebalancing at scale-up/scale-down
          * `ContinueReconciliationOnManualRollingUpdateFailure` feature gate moves to the beta stage and is enabled by default
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:7b55f971a3b62923a9a61a61c82f542e07deeb6ffdbaede68244ba358a0853f2
        - quay.io/strimzi/operator@sha256:b07b81f7e282dea2e4e29a7c93cfdd911d4715a5c32fe4c4e3d7c71cba3091e8
        - quay.io/strimzi/kafka@sha256:238bb080825b7ecebaa819be09788bc3d056ae8162ebcd41444076dc22f85740
        - quay.io/strimzi/kafka@sha256:98d332b7fa5c9e48f60b516e00ed5812bd9268e014fd2337284e7cc3e664f1eb
        - quay.io/strimzi/kafka@sha256:9a4224635e09ece2c1d3ffe9d715813cf63037461c9f06c1b5f53554aed5ecdc
        - quay.io/strimzi/kafka-bridge@sha256:cc95ce9f496057f3b18d5e8297b40a410cc66f9fd4a2227615971458a26bac25
        version: 0.44.0
      entries:
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.44.x
    - currentCSV: strimzi-cluster-operator.v0.45.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:58e8722216b0b0e8b1dd79109725238fa4016a16cc6a9fca0e327bccb1506810
          createdAt: "2025-06-27 01:38:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.45.1/deploying.html#assembly-upgrade-str).

          **From version 0.45, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**

          **ZooKeeper and Kafka Mirror Maker 1 support will be removed from Apache Kafka and Strimzi.** Strimzi 0.45 is the last minor Strimzi version with support for ZooKeeper-based Apache Kafka clusters and MirrorMaker 1 deployments. Please make sure to [migrate to KRaft](https://strimzi.io/docs/operators/latest/full/deploying.html#assembly-kraft-mode-str) and MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          **Strimzi EnvVar Configuration Provider and Strimzi MirrorMaker 2 Extensions will be removed from Strimzi.** Strimzi 0.45 is the last Strimzi version to include the Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0). Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.
          ### New in 0.45.1
          * Add support for Kafka 3.9.1
          * Update HTTP bridge to 0.31.2
          * Update Strimzi Kafka OAuth library to 0.15.1
          * Fixed bug which may lead to metadata loss within the cluster when restarting a KRaft migration after a previous rollback due to missing `/migration` znode deletion.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:58e8722216b0b0e8b1dd79109725238fa4016a16cc6a9fca0e327bccb1506810
        - quay.io/strimzi/kafka@sha256:e98fbe49cadbc5d3afbd8f17abea218f216c95d52f4fa1fe514a7a988624da24
        - quay.io/strimzi/kafka@sha256:ba52ed046b1dccdbd96f4e68057ce014d862a7c9c1fc670760c023b9aa09f23f
        - quay.io/strimzi/kafka-bridge@sha256:c462441a966f0732855e684b4840563bfad45391345c9511fa49e1630ba9c3e3
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:770e410f1e7e0e92c31bae6c19831ba36d59dd11b44e90c159aef16a965aa998
        - quay.io/strimzi/kafka@sha256:173cb5793756f75739c51a83bbe06d7f0094fd003a2982160a5cc6e216cd6d28
        - quay.io/strimzi/kafka@sha256:6a05924839a2c7da0431408e6d6223a055e75d88b3f057ac5caa7f9097227f56
        version: 0.45.1
      entries:
      - name: strimzi-cluster-operator.v0.45.1
        version: 0.45.1
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.45.x
    - currentCSV: strimzi-cluster-operator.v0.46.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.0.0",
                      "metadataVersion": "4.0-IV3",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:6a22e4968e9d409405423d12c04a0e3fecaea2ee2794bd27de7cf21f59453265
          createdAt: "2025-06-26 11:58:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.46.0/deploying.html#assembly-upgrade-str).

          **From version 0.46, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**
          **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed. Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!
          **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images. Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.
          **Support for MirrorMaker 1 has been removed. Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.
          ### New in 0.46.1
          * Update CSV file for OLM installation to add missing permission for broker Pods to "get" nodes Kubernetes resource.
          * Allow the injection of annotations on the deployment object in Helm charts.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:82273d873ce21a73f8f3118bf61627cb15734b43ea7ca92d43b290340b44e485
        - quay.io/strimzi/kafka-bridge@sha256:3e1815aed3f74b8573cebfced66b2c78ca0cf7c746c26db2002380cd39d09ae6
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:b532234c234fc30fecec992ad1a6154c5ef08ebf9a20ec8bc62e0c88d791ff36
        - quay.io/strimzi/operator@sha256:6a22e4968e9d409405423d12c04a0e3fecaea2ee2794bd27de7cf21f59453265
        - quay.io/strimzi/kafka@sha256:f3837f8cb2a60e906e84486128c41c398424f4085df2348e9e0486bfc12089fd
        version: 0.46.1
      entries:
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.46.x
    - currentCSV: strimzi-cluster-operator.v0.47.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.0.0",
                      "metadataVersion": "4.0-IV3",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:589c4d63641d9a944462dd682f6e5febe8b38ff55b9253949b061aca16feb154
          createdAt: "2025-07-12 17:21:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.47 is the last Strimzi version with support for Kubernetes 1.25 and 1.26.**
            From Strimzi 0.48.0 on, we will support only Kubernetes 1.27 and newer.

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.47.0
          * Support for Kafka 3.9.1
          * Progress tracking for Cruise Control rebalances
          * Support for Kubernetes Image Volumes to mount custom plugins
          * Support for [JsonTemplateLayout](https://logging.apache.org/log4j/2.x/manual/json-template-layout.html) in Operators, Cruise Control, and Kafka 4.0.0
          * Support for Strimzi Metrics Reporter for the Kafka brokers / controllers
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:589c4d63641d9a944462dd682f6e5febe8b38ff55b9253949b061aca16feb154
        - quay.io/strimzi/kafka@sha256:551882f47a6d2dc8300bcfca251fffffc3379acb2724a7efe16f17c0c4f529fb
        - quay.io/strimzi/kafka-bridge@sha256:428c5587f547dabfc995a73745569d113380c7e6b8bf0b2504c6a9a00d6c4608
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:c67a3ad3d92420ac8b7f8435eeec53ebffaea4e894e65414a7c47c20e3faba27
        - quay.io/strimzi/kafka@sha256:978e2ed5f0d315e8bf0ab0503ee839acbd9da01ca427e6581b59ac25d36bc9c9
        - quay.io/strimzi/kafka@sha256:8a793cdf5084cdb72c6663419530985158b16330a518720e9e9d1d96babbaf7e
        version: 0.47.0
      entries:
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.47.x
    - currentCSV: strimzi-cluster-operator.v0.48.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.1.0",
                      "metadataVersion": "4.1-IV1",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.1.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
          createdAt: "2025-09-25 10:32:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.48.0 supports only Kubernetes 1.27 and newer.**
            Kubernetes 1.25 and 1.26 are not supported anymore

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.48.0
          * Support for Kafka 4.1.0
          * Add monitoring of custom resources using kubernetes-state-metrics (KSM)
          * Support for type: custom client authentication
          * Added support for Strimzi Metrics Reporter to Kafka Connect, Mirror Maker 2 and Kafka Bridge.
          * Added support for KIP-745 in Kafka Connector and Mirror Maker 2, allowing the usage of includeTasks and onlyFailed arguments in Kafka connectors restart.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/maven-builder@sha256:64380efcb8da87fdc0e0de238a20456be9abb830174c2f2331882b7a84657e57
        - quay.io/strimzi/kafka@sha256:48db05ad2e9fa39eb388089e7e25654bcbf4da1ddc3a5635d675056c28bbcb96
        - quay.io/strimzi/kafka@sha256:34450afc1a3399a9026fc3c6ec90e9f6adfec9faf6520545b4d84e90cda34964
        - quay.io/strimzi/kafka-bridge@sha256:53034f64f0b672f10b5bacea1c7a25132f118df7fd5c9032c4dbf702ed93796a
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
        version: 0.48.0
      entries:
      - name: strimzi-cluster-operator.v0.48.0
        version: 0.48.0
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.48.x
    defaultChannel: stable
    packageName: strimzi-kafka-operator
    provider:
      name: Strimzi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-nifi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nifi-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:05946a5e8f040e8ee46d56f2d4cca3d97e02d086e6d8224200204de4d76851fa
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:05946a5e8f040e8ee46d56f2d4cca3d97e02d086e6d8224200204de4d76851fa
        version: 24.11.1
      entries:
      - name: nifi-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: nifi-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:e5c3c4e2ad1c455fbb04a60658936bfb44c2bc9d2b616737ce1a0eba201b7a3f
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiClusterSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:e5c3c4e2ad1c455fbb04a60658936bfb44c2bc9d2b616737ce1a0eba201b7a3f
        version: 25.3.0
      entries:
      - name: nifi-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: nifi-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiClusterSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
        version: 25.7.0
      entries:
      - name: nifi-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: nifi-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiClusterSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
        version: 25.7.0
      entries:
      - name: nifi-operator.v25.7.0
        version: 25.7.0
      - name: nifi-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-nifi-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/en/products/edge/validated-patterns
    name: patterns-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: patterns-operator.v0.0.62
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gitops.hybrid-cloud-patterns.io/v1alpha1",
                "kind": "Pattern",
                "metadata": {
                  "name": "pattern-sample"
                },
                "spec": {
                  "clusterGroupName": "hub",
                  "gitSpec": {
                    "targetRepo": "https://github.com/validatedpatterns/multicloud-gitops",
                    "targetRevision": "main"
                  },
                  "multiSourceConfig": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/hybridcloudpatterns/patterns-operator:0.0.62
          createdAt: "2025-09-02 16:14:40"
          description: "An operator to deploy and manage architecture patterns from
            https://validatedpatterns.io. \nThis operator collects some Analytics
            like cluster version, cloud type, etc. To disable this\nadd \"ANALYTICS:
            false\" to the environment variables in the subscription."
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/validatedpatterns/patterns-operator
          support: validatedpatterns@googlegroups.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pattern is the Schema for the patterns API
            displayName: Pattern
            kind: Pattern
            name: patterns.gitops.hybrid-cloud-patterns.io
            version: v1alpha1
        description: Deploys and manages architecture patterns from https://validatedpatterns.io
        displayName: Validated Patterns Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Validated
        - Pattern
        - Blueprint
        - Architecture
        links:
        - name: Hybrid Cloud Patterns
          url: https://validatedpatterns.io
        maintainers:
        - email: validatedpatterns@googlegroups.com
          name: patterns-team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/en/products/edge/validated-patterns
        relatedImages:
        - quay.io/hybridcloudpatterns/kube-rbac-proxy@sha256:409a464692c7db3cf68893376479dca1b53802412237637725927f7e78aa0844
        - quay.io/hybridcloudpatterns/patterns-operator@sha256:15ed44938b112bcf9b3a9a566749b39ba2b9053a6a8b9b499ef247977c18496c
        version: 0.0.62
      entries:
      - name: patterns-operator.v0.0.62
        version: 0.0.62
      - name: patterns-operator.v0.0.61
        version: 0.0.61
      - name: patterns-operator.v0.0.60
        version: 0.0.60
      - name: patterns-operator.v0.0.59
        version: 0.0.59
      - name: patterns-operator.v0.0.58
        version: 0.0.58
      - name: patterns-operator.v0.0.57
        version: 0.0.57
      - name: patterns-operator.v0.0.56
        version: 0.0.56
      - name: patterns-operator.v0.0.55
        version: 0.0.55
      - name: patterns-operator.v0.0.54
        version: 0.0.54
      - name: patterns-operator.v0.0.53
        version: 0.0.53
      - name: patterns-operator.v0.0.52
        version: 0.0.52
      - name: patterns-operator.v0.0.51
        version: 0.0.51
      - name: patterns-operator.v0.0.50
        version: 0.0.50
      - name: patterns-operator.v0.0.49
        version: 0.0.49
      - name: patterns-operator.v0.0.48
        version: 0.0.48
      - name: patterns-operator.v0.0.47
        version: 0.0.47
      - name: patterns-operator.v0.0.46
        version: 0.0.46
      - name: patterns-operator.v0.0.45
        version: 0.0.45
      - name: patterns-operator.v0.0.44
        version: 0.0.44
      - name: patterns-operator.v0.0.43
        version: 0.0.43
      - name: patterns-operator.v0.0.42
        version: 0.0.42
      - name: patterns-operator.v0.0.41
        version: 0.0.41
      - name: patterns-operator.v0.0.40
        version: 0.0.40
      - name: patterns-operator.v0.0.39
        version: 0.0.39
      - name: patterns-operator.v0.0.38
        version: 0.0.38
      - name: patterns-operator.v0.0.37
        version: 0.0.37
      - name: patterns-operator.v0.0.36
        version: 0.0.36
      - name: patterns-operator.v0.0.35
        version: 0.0.35
      - name: patterns-operator.v0.0.34
        version: 0.0.34
      - name: patterns-operator.v0.0.33
        version: 0.0.33
      - name: patterns-operator.v0.0.32
        version: 0.0.32
      - name: patterns-operator.v0.0.31
        version: 0.0.31
      - name: patterns-operator.v0.0.30
        version: 0.0.30
      - name: patterns-operator.v0.0.29
        version: 0.0.29
      - name: patterns-operator.v0.0.28
        version: 0.0.28
      - name: patterns-operator.v0.0.27
        version: 0.0.27
      - name: patterns-operator.v0.0.26
        version: 0.0.26
      - name: patterns-operator.v0.0.25
        version: 0.0.25
      - name: patterns-operator.v0.0.24
        version: 0.0.24
      - name: patterns-operator.v0.0.23
        version: 0.0.23
      - name: patterns-operator.v0.0.22
        version: 0.0.22
      - name: patterns-operator.v0.0.21
        version: 0.0.21
      - name: patterns-operator.v0.0.20
        version: 0.0.20
      - name: patterns-operator.v0.0.19
        version: 0.0.19
      - name: patterns-operator.v0.0.18
        version: 0.0.18
      - name: patterns-operator.v0.0.17
        version: 0.0.17
      - name: patterns-operator.v0.0.16
        version: 0.0.16
      - name: patterns-operator.v0.0.15
        version: 0.0.15
      - name: patterns-operator.v0.0.14
        version: 0.0.14
      - name: patterns-operator.v0.0.13
        version: 0.0.13
      - name: patterns-operator.v0.0.12
        version: 0.0.12
      - name: patterns-operator.v0.0.11
        version: 0.0.11
      - name: patterns-operator.v0.0.10
        version: 0.0.10
      - name: patterns-operator.v0.0.9
        version: 0.0.9
      - name: patterns-operator.v0.0.8
        version: 0.0.8
      - name: patterns-operator.v0.0.7
        version: 0.0.7
      - name: patterns-operator.v0.0.6
        version: 0.0.6
      - name: patterns-operator.v0.0.5
        version: 0.0.5
      - name: patterns-operator.v0.0.4
        version: 0.0.4
      - name: patterns-operator.v0.0.3
        version: 0.0.3
      - name: patterns-operator.v0.0.2
        version: 0.0.2
      - name: patterns-operator.v0.0.1
        version: 0.0.1
      name: fast
    defaultChannel: fast
    packageName: patterns-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com/en/products/edge/validated-patterns
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: ""
    name: hawtio-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hawtio-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.14.0",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/hawtio/operator:0.5.0
          createdAt: "2022-04-08T00:00:00Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Hawtio Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - docker.io/hawtio/operator:0.5.0
        version: 0.5.0
      entries:
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: alpha
    - currentCSV: hawtio-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/hawtio/operator:1.2.0
          createdAt: "2024-12-02T17:22:48Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          olm.skipRange: '>=1.0.0 <1.0.2'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Hawtio console that eases the discovery and management
              of hawtio-enabled applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#jvm-agent).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/hawtio/operator:1.2.0
        version: 1.2.0
      entries:
      - name: hawtio-operator.v1.2.0
        version: 1.2.0
      - name: hawtio-operator.v1.1.1
        version: 1.1.1
      - name: hawtio-operator.v1.1.0
        version: 1.1.0
      - name: hawtio-operator.v1.0.1
        version: 1.0.1
      - name: hawtio-operator.v1.0.0
        version: 1.0.0
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: latest
    - currentCSV: hawtio-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/hawtio/operator:1.2.0
          createdAt: "2024-12-02T17:22:48Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          olm.skipRange: '>=1.0.0 <1.0.2'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Hawtio console that eases the discovery and management
              of hawtio-enabled applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#jvm-agent).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/hawtio/operator:1.2.0
        version: 1.2.0
      entries:
      - name: hawtio-operator.v1.2.0
        version: 1.2.0
      - name: hawtio-operator.v1.1.1
        version: 1.1.1
      - name: hawtio-operator.v1.1.0
        version: 1.1.0
      - name: hawtio-operator.v1.0.1
        version: 1.0.1
      - name: hawtio-operator.v1.0.0
        version: 1.0.0
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: stable-v1
    defaultChannel: stable-v1
    packageName: hawtio-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "docker.io/percona/percona-server-mongodb:7.0.18-11",
                  "imagePullPolicy": "Always",
                  "initImage": "docker.io/percona/percona-server-mongodb-operator:1.20.1",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:2.44.1",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "docker.io/percona/percona-backup-mongodb:2.9.1",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.io/percona/percona-server-mongodb-operator:1.20.1
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.20.1
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - docker.io/percona/percona-server-mongodb-operator:1.20.1
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator.v1.20.0
        version: 1.20.0
      - name: percona-server-mongodb-operator.v1.19.1
        version: 1.19.1
      - name: percona-server-mongodb-operator.v1.19.0
        version: 1.19.0
      - name: percona-server-mongodb-operator.v1.18.0
        version: 1.18.0
      - name: percona-server-mongodb-operator.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: directpv
      provider-url: ""
    name: minio-volume-manager-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-volume-manager-operator.v5.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.quay.io/v1alpha1",
                "kind": "AistorVolumemanager",
                "metadata": {
                  "name": "aistorvolumemanager-sample"
                },
                "spec": {
                  "annotations": {},
                  "controller": {
                    "podAnnotations": {},
                    "podLabels": {},
                    "replicas": 3
                  },
                  "images": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": []
                  },
                  "labels": {},
                  "license": "",
                  "nodeserver": {
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-10-05T23:18:24Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AistorVolumemanager
            name: aistorvolumemanagers.charts.quay.io
            version: v1alpha1
        description: directpv
        displayName: directpv
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - directpv
        links:
        - name: Private Dpv Operator
          url: https://private-dpv-operator.domain
        maintainers:
        - email: cesarcelis@minio.io
          name: celis
        maturity: alpha
        provider:
          name: directpv
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - quay.io/cniackz4/directpv-operator@sha256:af06e57661a7e05809e44c3fac571ed949cc04fd656fdb0c8324b533778eb604
        version: 5.0.3
      entries:
      - name: minio-volume-manager-operator.v5.0.3
        version: 5.0.3
      - name: private-dpv-operator.v5.0.0
        version: 5.0.0
      name: stable
    defaultChannel: stable
    packageName: minio-volume-manager-operator
    provider:
      name: directpv
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-commonui-operator-app
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-commonui-operator.v4.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operators.ibm.com/v1alpha1",
                "kind": "CommonWebUI",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "example-commonwebui",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator"
                  },
                  "name": "example-commonwebui"
                },
                "spec": {
                  "commonWebUIConfig": {
                    "cpuLimits": "300",
                    "cpuMemory": "256",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.8.0",
                    "ingressPath": "/common-nav",
                    "landingPage": "",
                    "requestLimits": "300",
                    "requestMemory": "256",
                    "serviceName": "common-web-ui"
                  },
                  "globalUIConfig": {
                    "cloudPakVersion": "3.6.0",
                    "defaultAdminUser": "admin",
                    "defaultAuth": "",
                    "enterpriseLDAP": "",
                    "enterpriseSAML": "",
                    "osAuth": "",
                    "sessionPollingInterval": 5000
                  },
                  "operatorVersion": "4.8.0",
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "256Mi"
                    },
                    "requests": {
                      "cpu": "300m",
                      "memory": "256Mi",
                      "ephemeral-storage": "251Mi"
                    }
                  },
                  "version": "4.8.0"
                }
              },
              {
                "apiVersion": "foundation.ibm.com/v1",
                "kind": "NavConfiguration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "common-web-ui-config",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator",
                    "default": "true",
                    "name": "common-web-ui-config"
                  },
                  "name": "common-web-ui-config"
                },
                "spec": {
                  "about": {
                    "copyright": "© 2018, 2020 IBM. All rights reserved.",
                    "licenses": [
                      "yq, version 3.3.0, MIT+GPL",
                      "MongoDB, version 4.0.16 Community Edition, SSPL",
                      "Ansible: © 2017 Red Hat, Inc., http://www.redhat.com; © Henry Graham (hzgraham) \u003cHenry.Graham@mail.wvu.edu\u003e",
                      "calico-bird: © 1998–2008, Martin Mares \u003cmj@ucw.cz\u003e; © 1998–2000 Pavel Machek \u003cpavel@ucw.cz\u003e;",
                      "© 1998–2008, Ondrej Filip \u003cfeela@network.cz\u003e; © 2009–2013,  CZ.NIC z.s.p.o.",
                      "chrony: © Richard P. Curnow  1997-2003, GPL v2",
                      "collectd, © 2017-2018, version 5.7.2, GPL v2, \u003chttps://github.com/collectd/collectd/tree/collectd-5.7.2\u003e",
                      "crudini: © Pádraig Brady \u003cP@draigBrady.com\u003e",
                      "Galera-3: © 2007–2014 Codership Oy \u003cinfo@codership.com\u003e",
                      "glusterfs: © 2010–2013+ James Shubin \u003chttps://ttboj.wordpress.com/\u003e",
                      "haproxy: © 2000–2013  Willy Tarreau \u003cw@1wt.eu\u003e",
                      "heketi v6.0.0: © 2015 The heketi Authors, GPL v2",
                      "heketi v8.0.0: © 2015 The heketi Authors, GPL v2",
                      "heketi-master/apps.app.go: © 2015 The heketi Authors",
                      "heketi-master/client/api/go-client/backup.go: © 2016 The heketi Authors",
                      "heketi-master/doc/man/heketi-cli.8: © 2016 The heketi Authors",
                      "heketi-master/extras/docker/gluster/gluster-setup.sh: © 2016 Red Hat, Inc. \u003chttp://www.redhat.com\u003e",
                      "ieee-data: © 2013 Luciano Bello \u003cluciano@debian.org\u003e",
                      "javax.mail: © 2017 Oracle and/or its affiliates. All rights reserved.",
                      "keepalived: © 2001-2017 Alexandre Cassen \u003cacassen@gmail.com\u003e",
                      "libonig2: © 2006–2008 Max Kellermann \u003cmax@duempel.org\u003e; © 2014–2015 Jörg Frings-Fürst \u003cdebian@jff-webhosting.net\u003e",
                      "libtomcrypt: © 2004 Sam Hocevar \u003csam@hocevar.net\u003e, GPL v2",
                      "mariadb-common: © 2018 MariaDB. All rights reserved.",
                      "mariaDB: © 2018 MariaDB. All rights reserved. \u003chttps://mariadb.com/\u003e",
                      "mariadb-server: © 2018 MariaDB. All rights reserved.",
                      "minitar: © 2004 Mauricio Julio Fernandez Pradier and Austin Ziegler",
                      "MongoDB: © 2007 Free Software Foundation, Inc. \u003chttp://fsf.org/\u003e",
                      "nvmi-cli: © 1989, 1991 Free Software Foundation, Inc., GPL v2",
                      "OpenJDK: © 2018 Oracle Corporation and/or its affiliates",
                      "openshift-mariadb-galera: © 2007 Free Software Foundation, Inc. \u003chttp://fsf.org/\u003e",
                      "percona-xtrabackup: © 2006–2018 Percona LLC.",
                      "pwgen: © Christian Thöing \u003cc.thoeing@web.de\u003e",
                      "rdoc: © 2001–2003 Dave Thomas, The Pragmatic Programmers",
                      "readline: © Chet Ramey \u003cchet.ramey@case.edu\u003e",
                      "John the Ripper password cracker: © 1996–2013 by Solar Designer \u003csolar@openwall.com\u003e",
                      "spdx-exceptions: © 2018 SPDX Workgroup a Linux Foundation Project. All rights reserved.",
                      "socat: © 2001–2010 Gerhard Rieger",
                      "sshpass: © 2006, 2008 Lingnu Open Source Consulting Ltd. \u003chttp://www.lingnu.com\u003e",
                      "timelimit: © 2001, 2007 - 2010  Peter Pentchev, GPL v2",
                      "ua-parser-js: © 2012-2018 Faisal Salman \u003cf@faisalman.com\u003e, GPL v2",
                      "ubuntu-cloud-keyring: © 2010 Michael Vogt \u003cmichael.vogt@canonical.com\u003e",
                      "unboundid-ldapsdk: © 2015 UnboundID. The LDAP SDK for Java is developed by UnboundID. \u003cinfo@unboundid.com\u003e",
                      "xmpp4r: © Lucas Nussbaum \u003clucas@lucas-nussbaum.net\u003e, Stephan Maka \u003cstephan@spaceboyz.net\u003e, and others.",
                      "module-assistant: © 2003-2008 Eduard Bloch \u003cblade@debian.org\u003e, version 0.11.8, GPL v2;",
                      "© 2009 Cyril Brulebois \u003ckibi@debian.org\u003e, version 0.11.8, GPL v2; © 2013-2018 Andreas Beckmann \u003canbe@debian.org\u003e, version 0.11.8, GPL v2",
                      "module-init-tools: © 2011 ProFUSION embedded systems, version 22, GPL v2",
                      "thin: © 2017 Marc-Andre Cournoyer \u003cmacournoyer@gmail.com\u003e, version 1.7.2, GPL v2",
                      "gosu, © 1999-2014, version 1.1, GPL v3",
                      "mercurial (Python), © 2006-2018 ,version v4.5.3, GPL v2",
                      "garden-runc, © 2015-Present CloudFoundry.org Foundation, Inc. All Rights Reserved, version 1.17.0, GPLv2",
                      "libtomcrypt0, © 2003-2007 Tom St Denis \u003ctomstdenis@gmail.com\u003e, version 1.17-7, GPLv2",
                      "console-setup-min, © 1999,2000,2001,2002,2003,2006,2007,2008,2009,2010,2011 Anton Zinoviev, \u003canton@lml.bas.bg\u003e,version 1.108, GPLv2",
                      "dracut, © 2009 Harald Hoyer \u003charald@redhat.com\u003e, version 044+3-3, GPLv2",
                      "dracut-core, © 2009 Harald Hoyer \u003charald@redhat.com\u003e, version 044+3-3, GPLv2",
                      "g++, version 5.4.0-6ubuntu, GPL v2",
                      "libstdc++6, version 5.4.0-6ubuntu, GPL v3",
                      "libstdc++-5-dev, version 5.4.0-6ubuntu, GPL v3",
                      "docker-engine-selinux, version 3b5fac4, GPLv2",
                      "unorm, version 1.5.0, GPL v2",
                      "psmisc, version 22.20, GPL v2",
                      "lvm2-devel, version 2.0.2, GPL v2",
                      "nfs-utils, version 1.3, GPL v2",
                      "popt-static, version 1.13, GPL v2",
                      "sysvinit-tools, version 2.88, GPL v2",
                      "stunnel, version 5.53, GPL v2",
                      "stunnel, version 5.39, GPL v2",
                      "LVM2, version 2.02.180-10.el7_6.2, GPL v2",
                      "sysdig, version 2c43237, GPL",
                      "chisels, version 9722dbc, GPL",
                      "MongoDB, version 4.0.12, SSPL",
                      "ffi (Ruby Gem), 1.11.1, GPL",
                      "inotify-tools, v3.14, GPL v2",
                      "logrotate, v3.8.6, GPL v2",
                      "checker-qual, version 2.0.0, GPLv2",
                      "ocrad-bower, v1.0.0, GPL v3",
                      "Grafana, v7.5.12, AGPL",
                      "IBM Semeru Runtime Open Edition binaries, GPL v2",
                      "checker-compat-qual 2.5.5, GPL-2.0",
                      "javax.annotation-api 1.3.1, GPL-2.0"
                    ],
                    "logoUrl": "IBM Cloud Pak | Administration Hub",
                    "logoAltText": "IBM Cloud Pak | Administration Hub"
                  },
                  "header": {
                    "disabledItems": [
                      "createResource",
                      "catalog"
                    ],
                    "docUrlMapping": "http://ibm.biz/cpcs_adminui",
                    "logoAltText": "IBM Cloud Pak Administration Hub",
                    "logoHeight": "47px",
                    "logoUrl": "/common-nav/graphics/ibm-cloudpack-logo.svg",
                    "logoWidth": "190px"
                  },
                  "login": {
                    "loginDialog": {
                      "acceptText": "Your acceptance text here",
                      "dialogText": "You must set your dialog for this environment",
                      "enable": false,
                      "headerText": "Header text here"
                    },
                    "logoAltText": "Cloud Pak",
                    "logoHeight": "47px",
                    "logoUrl": "/common-nav/api/graphics/logincloudpak.svg",
                    "logoWidth": "190px"
                  },
                  "navItems": [
                    {
                      "id": "providers",
                      "label": "Identity providers",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/realms"
                    }
                  ]
                }
              },
              {
                "apiVersion": "foundation.ibm.com/v1",
                "kind": "NavConfiguration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "icp4i",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator",
                    "name": "icp4i"
                  },
                  "name": "icp4i"
                },
                "spec": {
                  "header": {
                    "disabledItems": [
                      "createResource",
                      "catalog",
                      "bookmark"
                    ],
                    "logoAltText": "Cloud Pak for Integration",
                    "logoUrl": "/common-nav/graphics/ibm-cloudpak-integration.svg"
                  },
                  "navItems": [
                    {
                      "detectionServiceName": true,
                      "id": "metering",
                      "label": "Metering",
                      "serviceId": "metering-ui",
                      "serviceName": "metering-ui",
                      "url": "/metering/dashboard?ace_config={ 'showClusterData': false }\u0026dashboard=cpi.icp.main"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "monitoring",
                      "isAuthorized": [
                        "Administrator",
                        "ClusterAdministrator",
                        "Operator"
                      ],
                      "label": "Monitoring",
                      "serviceId": "monitoring-ui",
                      "serviceName": "ibm-monitoring-grafana",
                      "target": "_blank",
                      "url": "/grafana"
                    },
                    {
                      "id": "id-access",
                      "label": "Identity and access",
                      "serviceId": "webui-nav"
                    },
                    {
                      "id": "providers",
                      "parentId": "id-access",
                      "label": "Identity providers",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/realms?useNav=icp4i"
                    },
                    {
                      "id": "teams-ids",
                      "parentId": "id-access",
                      "label": "Teams and service IDs",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/teams?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "logging",
                      "label": "Logging",
                      "serviceId": "kibana",
                      "serviceName": "kibana",
                      "target": "_blank",
                      "url": "/kibana"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "releases",
                      "label": "Helm Releases",
                      "serviceId": "catalog-ui",
                      "serviceName": "catalog-ui",
                      "url": "/catalog/instances?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "repos",
                      "label": "Helm Repositories",
                      "serviceId": "catalog-ui",
                      "serviceName": "catalog-ui",
                      "url": "/catalog/repositories?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "licensing",
                      "label": "Licensing",
                      "serviceId": "ibm-license-service-reporter",
                      "url": "/license-service-reporter",
                      "isAuthorized": [
                        "ClusterAdministrator",
                        "CloudPakAdministrator"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operators.ibm.com/v1alpha1",
                "kind": "SwitcherItem",
                "metadata": {
                  "name": "example-switcheritem"
                },
                "spec": {
                  "cloudPakInfo": {
                    "display": "Administration",
                    "label": "administration-hub",
                    "landingPage": "/common-nav/dashboard",
                    "logoURL": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
          description: The IBM Common Web UI delivers the common header API and the
            identity and access pages for IBM Cloud Platform Common Services.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.8.0
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-commonui-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            displayName: CommonWebUI service
            kind: CommonWebUI
            name: commonwebuis.operators.ibm.com
            version: v1alpha1
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            kind: NavConfiguration
            name: navconfigurations.foundation.ibm.com
            version: v1
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            kind: SwitcherItem
            name: switcheritems.operators.ibm.com
            version: v1alpha1
        description: "**Important:** Do not install this operator directly. Only install
          this operator using the IBM Common Services Operator. For more information
          about installing this operator and other Common Services operators, see
          [Installer documentation](http://ibm.biz/cpcs_opinstall). Additionally,
          you can exit this panel and navigate to the IBM Common Services tile in
          OperatorHub to learn more about the operator.\n\n If you are using this
          operator as part of an IBM Cloud Pak, see the documentation for that IBM
          Cloud Pak to learn more about how to install and use the operator service.
          For more information about IBM Cloud Paks, see [IBM Cloud Paks that use
          Common Services](http://ibm.biz/cpcs_cloudpaks).\n\n You can use the ibm-commonui-operator
          to install the Common Web UI service for the IBM Cloud Platform Common Services
          and access the Common Web UI console. You can use the Common Web UI console
          to access information and features from other IBM Cloud Platform Common
          Services or IBM Cloud Paks that you install. \n\nFor more information about
          the available IBM Cloud Platform Common Services, see the [IBM Knowledge
          Center](http://ibm.biz/cpcsdocs). \n## Supported platforms \n\n Red Hat
          OpenShift Container Platform 4.2 or newer installed on one of the following
          platforms: \n\n- Linux x86_64 \n- Linux on Power (ppc64le) \n- Linux on
          IBM Z and LinuxONE \n## Prerequisites\n\n The Common Web UI service has
          dependencies on other IBM Cloud Platform Common Services. Before you install
          this operator, you need to first install the operator dependencies and prerequisites:
          \n For the list of operator dependencies, see the IBM Knowledge Center [Common
          Services dependencies documentation](http://ibm.biz/cpcs_opdependencies).
          \n For the list of prerequisites for installing the operator, see the IBM
          Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).
          \n## Documentation \n\n To install the operator with the IBM Common Services
          Operator follow the the installation and configuration instructions within
          the IBM Knowledge Center. \n- If you are using the operator as part of an
          IBM Cloud Pak, see the documentation for that IBM Cloud Pak, for a list
          of IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).
          \n- If you are using the operator with an IBM Containerized Software, see
          the IBM Cloud Platform Common Services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall)."
        displayName: Ibm Common UI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        - Web Console
        - Common Services
        links:
        - name: GitHub
          url: https://github.com/IBM/ibm-commonui-operator
        maintainers:
        - email: siddarth.patel@ibm.com
          name: Siddarth Patel
        - email: sgrube@us.ibm.com
          name: Steve Grube
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:e4eb6a75e0d207232e479693d9c99ba863f8e62830f3191ca4d1e7ceaf3586d1
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator-bundle@sha256:e4340bd9481da998a6b89e94c884c46e17cfb50d9d0c1099f39a3c54e500d2b2
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
        version: 4.8.0
      entries:
      - name: ibm-commonui-operator.v4.8.0
        version: 4.8.0
      name: v4.7
    defaultChannel: v4.7
    packageName: ibm-commonui-operator-app
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent
      provider-url: ""
    name: cilium-enterprise
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cilium.v1.10.13-xed86eae
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"prometheus":{"serviceMonitor":{"enabled":false}},"mode":"cluster-pool"},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium Enterprise - eBPF-based Networking, Security, and Observability
        displayName: Cilium Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Enterprise Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Cilium Enterprise
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:cc1f0e6b27ccda8a843dfb7fa4a8fef052ba29716646557815677c5d1bea2a1a
        - quay.io/isovalent/hubble-enterprise-operator@sha256:0d5e6a5d9c07780c18e5aa81d8f1c95c399bff2d36a44fa5f61ac89a788ca83b
        - quay.io/isovalent/hubble-export-fluentd@sha256:05cb13f533f2c89b98a47cd58badda07add54077cd22283ca5028cb39f2e4486
        - quay.io/isovalent/hubble-enterprise-metadata@sha256:72c7aa3961f14df413569cd156820a45553bb24d393b11246a18bf8b6553e022
        - quay.io/isovalent/hubble-enterprise@sha256:1e1d3e6c199546b2398f4166ff64210c04075bb407c8f07c5be8d3db08560d4a
        - quay.io/isovalent/hubble-ui-enterprise@sha256:c9921c65892a5293a4f1bef787636d6ae58cec1af65da06862a47e3450afdc6e
        - quay.io/oauth2-proxy/oauth2-proxy@sha256:ecd26b74a01f2b547ddaed4d32d35f8f5e09c378d5c1fc6cfa63f0adf659ac2b
        - quay.io/isovalent/hubble-relay@sha256:9145b03acc954efde54e1ab9da878462526c5e820b83a92e32cc24291f784d33
        - quay.io/isovalent/operator-generic@sha256:1cc0b0d2e71564ad8759be326c66ea13b4955b9441fb67e906da534e1b38aefa
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/hubble-rbac@sha256:03f3304c157024356a6de13c3d91b4506b8e54d587920d80ebcc76cbcfeaeda6
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/isovalent/hubble-export-s3@sha256:ce4f71f7f13301133b7d471b87fb81108b2c2da7fc29865ef96bc8431b2be72b
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/isovalent/cilium@sha256:6887537789603946543aaa370a3ec592287a235231bf16aa386e091823285515
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        - quay.io/cilium/hubble-export-stdout@sha256:f1b1475a3c89950481264f5f8ca63eea3b911a45dc83fdf15122d695a0fba9b3
        - quay.io/isovalent/clustermesh-apiserver@sha256:eb3ee3d2cb07e6c5966e4bc2e50615949303be5d5128e156d5c0dbe8404dc093
        - quay.io/isovalent/hubble-ui-enterprise-backend@sha256:e892cba6dafcdf6704df3135a0b38d11894a5f276ace40683e74cb62b9c311d8
        version: 1.10.13+xed86eae
      entries:
      - name: cilium.v1.10.13-xed86eae
        version: 1.10.13+xed86eae
      - name: cilium.v1.10.12-xb203fae
        version: 1.10.12+xb203fae
      - name: cilium.v1.10.11-x10ae342
        version: 1.10.11+x10ae342
      name: "1.10"
    - currentCSV: cilium.v1.11.21-xd0d07b1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium Enterprise - eBPF-based Networking, Security, and Observability
        displayName: Cilium Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Enterprise Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Cilium Enterprise
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-enterprise-operator:v1.10.5@sha256:31b18a1ea7002936f04dd070549e0c9a02f8a33394666fc5b0523ab1350c6120
        - quay.io/isovalent/hubble-enterprise:v1.10.5@sha256:340700d4e28ded2aa9c50ff9f938996bb43faa84e0b2ba38ec79452648d7e5d7
        - quay.io/isovalent/hubble-rbac:v1.3.0@sha256:4c106dd133861a678eb24cf98b35d5023cccca0834c954c49b68514c366aedf0
        - quay.io/cilium/certgen:v0.1.5@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - cgr.dev/chainguard/oauth2-proxy:7.4.0@sha256:54857c70321004d3e0b8b8d7208ce5422c1f71c18db5be02d2543f187ea8e8e2
        - quay.io/isovalent/hubble-relay:v1.11.21-cee.1@sha256:5933dd1f8111d6bca0c68db9247385b5d215d5b613095bafab8244098f2c5a26
        - quay.io/isovalent/clustermesh-apiserver:v1.11.21-cee.1@sha256:e16d8375ed8a1c3a872047b3aacf3469bde43552f1a51ee3e159ef2e9fea3e97
        - quay.io/isovalent/cilium:v1.11.21-cee.1@sha256:f41241a15cd96d26727aa1ec413ace643728ddbe42f7fb7b234dbda22bf3950b
        - quay.io/cilium/startup-script:62093c5c233ea914bfa26a10ba41f8780d9b737f@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/isovalent/hubble-ui-enterprise-backend:v0.23.0@sha256:2fbdad86bb5244075f31487262d790f6fa75cdd993f868b952252905c118621c
        - quay.io/isovalent/hubble-ui-enterprise:v0.23.0@sha256:02635d1392cca7e09ab7797f69147880068ba7e3b0d865be97eeab33edd646a0
        - quay.io/isovalent/hubble-export-stdout:v1.0.3@sha256:bef600089c1671fce11bb4bcb98b81f25e65152a1a2f13b64ea637cfc5779c11
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:372511f6fa3523dadfe9128caf77520cf74ce24df4a19e1583ee197375e06014
        - quay.io/isovalent/hubble-export-fluentd:v1.5.1@sha256:eaf42a84821509b69e2ef2b658796062e79a90e63045a98689da3da99a12e492
        - quay.io/isovalent/hubble-export-s3:v1.0.0@sha256:89c178c0a010166c68819ab5de32ab59b64a88aa23a05d8e4f78adb1a4b6c3d3
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:d3148bf3b9765467936ee067b25b5ef3683d3e55260f517c7fe271ac0e19cb47
        - quay.io/isovalent/operator-generic:v1.11.21-cee.1@sha256:25b668c1febe1d1a70b63dfe926be4e55fbf77fd186e9df7581e41d7e9813469
        - quay.io/coreos/etcd:v3.4.13@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        version: 1.11.21+xd0d07b1
      entries:
      - name: cilium.v1.11.21-xd0d07b1
        version: 1.11.21+xd0d07b1
      - name: cilium.v1.11.20-x0f6d966
        version: 1.11.20+x0f6d966
      name: "1.11"
    - currentCSV: cilium.v1.12.21-x3861686
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.12.14 <1.12.21'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-export-fluentd:v1.5.1@sha256:eaf42a84821509b69e2ef2b658796062e79a90e63045a98689da3da99a12e492
        - quay.io/isovalent/operator-generic:v1.12.21-cee.1@sha256:4e19cda799a747b69f53b5e5a860ba256399d18221d5a4bbf243142ba1c69bb8
        - quay.io/isovalent/clustermesh-apiserver:v1.12.21-cee.1@sha256:3f6549409966086a49e338c7829b9141a73ef2ebeebaac91db430aeb3ccd2e22
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:b6e40d76f0925e76f0a5b181be3f6c62728fa3e4aaef979caf9ea245f4ed03da
        - quay.io/isovalent/hubble-enterprise:v1.10.5@sha256:340700d4e28ded2aa9c50ff9f938996bb43faa84e0b2ba38ec79452648d7e5d7
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:f95c9e8002d8e381d604752b1d59b75595badbdb0b11237ac1cac4ecd04f4800
        - quay.io/isovalent/hubble-rbac:v1.3.0@sha256:4c106dd133861a678eb24cf98b35d5023cccca0834c954c49b68514c366aedf0
        - quay.io/cilium/startup-script:62093c5c233ea914bfa26a10ba41f8780d9b737f@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/isovalent/cilium:v1.12.21-cee.1@sha256:3a462308cf12b572db987c03d190ee33691ccb04775d01beb192703a52f8d47b
        - quay.io/isovalent/hubble-relay:v1.12.21-cee.1@sha256:9d8dadc3f330acf0fa169c428d5fdf1d138d0294065eec5c628f6647106fe8b6
        - quay.io/isovalent/hubble-ui-enterprise:v0.28.1@sha256:11dfba0e4fcf43415aa6e8b3b6e98a5c4f468ec67b377328ad48b92037bbfbb0
        - quay.io/isovalent/hubble-export-stdout:v1.0.3@sha256:bef600089c1671fce11bb4bcb98b81f25e65152a1a2f13b64ea637cfc5779c11
        - quay.io/isovalent/hubble-enterprise-operator:v1.10.5@sha256:31b18a1ea7002936f04dd070549e0c9a02f8a33394666fc5b0523ab1350c6120
        - quay.io/coreos/etcd:v3.5.4@sha256:795d8660c48c439a7c3764c2330ed9222ab5db5bb524d8d0607cac76f7ba82a3
        - quay.io/cilium/certgen:v0.1.8@sha256:4a456552a5f192992a6edcec2febb1c54870d665173a33dc7d876129b199ddbd
        - quay.io/isovalent/hubble-ui-enterprise-backend:v0.28.1@sha256:515aeb8166dd6995b04d4cd3333c94a740182943ba4e90a38871ee89c0b64e92
        - quay.io/isovalent/hubble-export-s3:v1.0.0@sha256:89c178c0a010166c68819ab5de32ab59b64a88aa23a05d8e4f78adb1a4b6c3d3
        version: 1.12.21+x3861686
      entries:
      - name: cilium.v1.12.21-x3861686
        version: 1.12.21+x3861686
      - name: cilium.v1.12.19-xca684e0
        version: 1.12.19+xca684e0
      - name: cilium.v1.12.18-xcb1adcf
        version: 1.12.18+xcb1adcf
      - name: cilium.v1.12.17-xae21c14
        version: 1.12.17+xae21c14
      - name: cilium.v1.12.16-x38e0b62
        version: 1.12.16+x38e0b62
      - name: cilium.v1.12.15-x8ca64a3
        version: 1.12.15+x8ca64a3
      - name: cilium.v1.12.14-xc86b260
        version: 1.12.14+xc86b260
      - name: cilium.v1.12.13-x2b5cec9
        version: 1.12.13+x2b5cec9
      - name: cilium.v1.12.12-xbd7b835
        version: 1.12.12+xbd7b835
      - name: cilium.v1.12.11-xffdea76
        version: 1.12.11+xffdea76
      - name: cilium.v1.12.10-x4a0e9e7
        version: 1.12.10+x4a0e9e7
      - name: cilium.v1.12.9-x4be66e4
        version: 1.12.9+x4be66e4
      - name: cilium.v1.12.8-x87b90a9
        version: 1.12.8+x87b90a9
      - name: cilium.v1.12.7-x53a2aab
        version: 1.12.7+x53a2aab
      name: "1.12"
    - currentCSV: cilium-enterprise.v1.13.21-x96449fb
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.7 <1.13.21+x96449fb'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/startup-script:c54c7edeab7fde4da68e59acd319ab24af242c3f@sha256:a9d0f451353f1b2f29ed3868e230ecfae10150eab9fd521de67adb3c8ca9c33a
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/operator-generic:v1.13.21-cee.2@sha256:6f2d0d575e3f74395c959833992cd5fbb6971f4e292df1b7cf45be8b07be5e7a
        - quay.io/isovalent/hubble-rbac:v1.5.6@sha256:739ca45ea09829b0fa8a6198901b34eb5254df55be071bc854bbd37cae12a557
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:b04e75456359b50b18e75464da5d7b467fdfeb058fc08ce3a10f057ba5f60107
        - quay.io/isovalent/hubble-enterprise:v1.11.6@sha256:eb60d59d1fb1b136855829274b31805c4206f8d050c3bf068e63a552576c1b38
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:fc7046dee1a204163e68a6d9a42fec200b02129493a7d166f899dd12e2763364
        - quay.io/isovalent/cilium:v1.13.21-cee.2@sha256:f760cd85fa64eb8a197ec282644a8db630745a8a2854ee08aa2662cf67188db5
        - quay.io/isovalent/clustermesh-apiserver:v1.13.21-cee.2@sha256:980b0235fd16817ecc4457ad4317a4601993b76df2e81b07a53f6f437c2705af
        - quay.io/isovalent/hubble-enterprise-operator:v1.11.6@sha256:38253c2e17a8b2e88a226f0976c9cf4a00445212e2886e6f07cbb3a52757324f
        - quay.io/cilium/certgen:v0.1.13@sha256:01802e6a153a9473b06ebade7ee5730f8f2c6cc8db8768508161da3cdd778641
        - quay.io/isovalent/hubble-export-stdout:v1.0.4@sha256:a0ff77efa890478f18f5deeba2cd1e2dfd6c5c77981adac57b14d7734de71ff1
        - quay.io/isovalent/hubble-export-fluentd:v1.11.0@sha256:50cd2e4c2d04e8befcc2ae1a1463f07cba01edb9b546fb455fe7976311165385
        - quay.io/isovalent/hubble-ui-enterprise-backend:v1.0.3@sha256:6756d2b808c38efce4765f90641d26aa2e0e79329ecc378a123cd3dc2c439640
        - quay.io/isovalent/hubble-relay:v1.13.21-cee.2@sha256:aaf5f959a76d6b517e627d2d0b09dab86ae942616819a9a95f057ed4f45ea480
        - quay.io/isovalent/hubble-ui-enterprise:v1.0.3@sha256:55bf787c96561f4a460e2bfe5966421e2d85f3f9a38ac2487c6f6486c97cd5af
        - quay.io/coreos/etcd:v3.5.4@sha256:795d8660c48c439a7c3764c2330ed9222ab5db5bb524d8d0607cac76f7ba82a3
        - quay.io/isovalent/hubble-export-s3:v1.0.2@sha256:518daea1074d260d824c3e03a93ab6b2a7a5f4c22b8e07938770add8257f4a52
        version: 1.13.21+x96449fb
      entries:
      - name: cilium-enterprise.v1.13.21-x96449fb
        version: 1.13.21+x96449fb
      - name: cilium-enterprise.v1.13.20-x5b70650
        version: 1.13.20+x5b70650
      - name: cilium-enterprise.v1.13.19-xdb27142
        version: 1.13.19+xdb27142
      - name: cilium-enterprise.v1.13.18-x0c21a94
        version: 1.13.18+x0c21a94
      - name: cilium-enterprise.v1.13.17-xc8f905f
        version: 1.13.17+xc8f905f
      - name: cilium-enterprise.v1.13.16-xcfe62ed
        version: 1.13.16+xcfe62ed
      - name: cilium-enterprise.v1.13.15-x1dfd26f
        version: 1.13.15+x1dfd26f
      - name: cilium.v1.13.14-xb70933f
        version: 1.13.14+xb70933f
      - name: cilium.v1.13.12-x453f6ad
        version: 1.13.12+x453f6ad
      - name: cilium.v1.13.11-xf6d6935
        version: 1.13.11+xf6d6935
      - name: cilium.v1.13.10-x9d57fc7
        version: 1.13.10+x9d57fc7
      - name: cilium.v1.13.9-xe8a359b
        version: 1.13.9+xe8a359b
      - name: cilium.v1.13.8-x96e1ca9
        version: 1.13.8+x96e1ca9
      - name: cilium.v1.13.7-x7320e7d
        version: 1.13.7+x7320e7d
      - name: cilium.v1.13.6-xd3714ef
        version: 1.13.6+xd3714ef
      - name: cilium.v1.13.5-x8386cae
        version: 1.13.5+x8386cae
      - name: cilium.v1.13.4-xc8b22ea
        version: 1.13.4+xc8b22ea
      - name: cilium.v1.13.3-x69e9561
        version: 1.13.3+x69e9561
      - name: cilium.v1.13.2-x4b1c6c8
        version: 1.13.2+x4b1c6c8
      name: "1.13"
    - currentCSV: cilium-enterprise.v1.14.22-x6d86c88
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.14.22+x6d86c88'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-ui-enterprise-backend:v1.2.2@sha256:4258f7e0bff24ff1e0b214f62b37097032b6426ba28edd3dfc4f451a34de0110
        - quay.io/cilium/cilium-envoy:v1.32.5-1743506100-0821ef0acdf9f824d47d34e02932be522b3e7233@sha256:305c8b7da56ef1b64e4deb0642acf93a3944f2712e40c99a40bb005e6dc9761b
        - quay.io/isovalent/hubble-rbac:v1.5.2@sha256:29309837e2246a1b6849d7a09d01ceebfdb43fa0ade973d14584f0e52b856b95
        - quay.io/cilium/startup-script:c54c7edeab7fde4da68e59acd319ab24af242c3f@sha256:a9d0f451353f1b2f29ed3868e230ecfae10150eab9fd521de67adb3c8ca9c33a
        - quay.io/isovalent/hubble-export-stdout:v1.0.4@sha256:a0ff77efa890478f18f5deeba2cd1e2dfd6c5c77981adac57b14d7734de71ff1
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/isovalent/operator-generic:v1.14.22-cee.1@sha256:a997fe24ff0b7f7df3ee12748413ae53d6e07c803162b0db4d812b916b0cfc74
        - quay.io/coreos/etcd:v3.5.4@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/isovalent/cilium:v1.14.22-cee.1@sha256:0e664a5a9e0edba38ab1a25eb7fa50861bb15f69ed0bbe3f3c792bd909240da3
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:fc7046dee1a204163e68a6d9a42fec200b02129493a7d166f899dd12e2763364
        - quay.io/isovalent/hubble-enterprise:v1.11.6@sha256:eb60d59d1fb1b136855829274b31805c4206f8d050c3bf068e63a552576c1b38
        - quay.io/isovalent/hubble-ui-enterprise:v1.2.2@sha256:3e03a58f6eea5d2df195f42ee2817f7b28fa20bbe140f5086937eee8da1b8d56
        - quay.io/cilium/certgen:v0.1.17@sha256:07959f3125dae659ab9036c14df046e2536f7a4f6c92dabda837388b0344dc74
        - quay.io/isovalent/hubble-relay:v1.14.22-cee.1@sha256:d7ffc880d2e42e0876417ec26115b0e9155ad9a2e7f47c02c434746afa42b29a
        - quay.io/isovalent/hubble-enterprise-operator:v1.11.6@sha256:38253c2e17a8b2e88a226f0976c9cf4a00445212e2886e6f07cbb3a52757324f
        - quay.io/isovalent/hubble-export-fluentd:v1.8.1@sha256:ce0cce63a8473d5ae73872fc65d82c54f2875511bb03f0bc8ee8480148c43ee1
        - quay.io/isovalent/hubble-export-s3:v1.0.2@sha256:518daea1074d260d824c3e03a93ab6b2a7a5f4c22b8e07938770add8257f4a52
        - quay.io/isovalent/clustermesh-apiserver:v1.14.22-cee.1@sha256:588e2b173ab409288eda2d404571184562d3d9fd12e2d9afe7282a5a7d79113a
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:148b1029eac14049e2ba84d4d75be1ca43ef2d89951adbdeddec392ad7c9fe1e
        version: 1.14.22+x6d86c88
      entries:
      - name: cilium-enterprise.v1.14.22-x6d86c88
        version: 1.14.22+x6d86c88
      - name: cilium-enterprise.v1.14.21-xe0c9553
        version: 1.14.21+xe0c9553
      - name: cilium-enterprise.v1.14.20-xe01e743
        version: 1.14.20+xe01e743
      - name: cilium-enterprise.v1.14.19-xc1b9aff
        version: 1.14.19+xc1b9aff
      name: "1.14"
    - currentCSV: cilium-enterprise.v1.15.21-x2661a01
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.15.21+x2661a01'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/operator-generic-ubi:v1.15.21-cee.1@sha256:97daac94735e497e1e897beaa01eb50866abace1aa9a3919fc57971e45f538fe
        - quay.io/isovalent/oauth2-proxy:7.12.0-20250819@sha256:fe8f76728c36536d498b2a5c11172cb18dc30bbb555df2a6d14c036b29cb9f10
        - quay.io/isovalent/hubble-relay-ubi:v1.15.21-cee.1@sha256:8461f720ca95a9c543a298bee1c1e56241f56d5426030c13afed8262f2f5c5ff
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.15.21-cee.1@sha256:36d00eb9f8ed45a3d8892942cfae097782dd6509889ef7d1a451f9b0334d17e1
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:bd1a4d47ef48a6eb84c49d979e8f5ce867af5932828dfd3233fa9f78c2eac1f4
        - quay.io/isovalent/hubble-ui-backend-ubi:v1.3.7@sha256:a4b5e7a032c1bbcd969a6873bc12935f8b6aa4c2d3a2ed87f5494fcfcdfd02d4
        - quay.io/isovalent/hubble-ui-frontend-ubi:v1.3.7@sha256:f2b4e935abe3eb94f82dab41c9d3d929fca71d82fa39a6c8002683375b5303a4
        - quay.io/isovalent/startup-script-ubi:1755531540-60ee83e@sha256:2e1d5048cbfcd3c77f732da8203a6c792cbf19897b5cf137b9ad23a6a267f6b2
        - quay.io/isovalent/cilium-envoy-ubi:v1.33.4-1752151664-7c2edb0b44cf95f326d628b837fcdd845102ba68@sha256:e81fdde48c3d5cd93111e1d51d9f8554adb32ee937601a9124354e0894933b99
        - quay.io/isovalent/cilium:v1.15.21-cee.1@sha256:39ef22d467783481eccb376fcb5d6bc254ed8f4b3b25032121498db40278f0b8
        - quay.io/isovalent/certgen-ubi:v0.1.19@sha256:c32e96b0972c9bcd3896945c46ea23e330c9dea407afee3a637daae637296f4c
        version: 1.15.21+x2661a01
      entries:
      - name: cilium-enterprise.v1.15.21-x2661a01
        version: 1.15.21+x2661a01
      - name: cilium-enterprise.v1.15.20-x6df37bb
        version: 1.15.20+x6df37bb
      - name: cilium-enterprise.v1.15.19-x0be49ab
        version: 1.15.19+x0be49ab
      - name: cilium-enterprise.v1.15.18-xeeaa70b
        version: 1.15.18+xeeaa70b
      - name: cilium-enterprise.v1.15.17-xdfd9482
        version: 1.15.17+xdfd9482
      - name: cilium-enterprise.v1.15.16-xc7e31ca
        version: 1.15.16+xc7e31ca
      - name: cilium-enterprise.v1.15.15-xf37f22c
        version: 1.15.15+xf37f22c
      - name: cilium-enterprise.v1.15.14-x0e512e3
        version: 1.15.14+x0e512e3
      - name: cilium-enterprise.v1.15.13-x134fa29
        version: 1.15.13+x134fa29
      name: "1.15"
    - currentCSV: cilium-enterprise.v1.16.15-x6419e26
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.16.0 <1.16.15+x6419e26'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-relay-ubi:v1.16.15-cee.1@sha256:c992d549627e8f9180c272e7d96ff3783de184796633ed5d0b6fe2cf7e74055d
        - quay.io/isovalent/operator-generic-ubi:v1.16.15-cee.1@sha256:970aa156b6f5c390b6835a626190e18656f02db6a5cf27aaa1ac9c7872f4b15a
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.16.15-cee.1@sha256:f7573ab4145d306051b5d6fa7ff1af395fc71935f5204839ab1c9f574aee17ad
        - quay.io/isovalent/certgen-ubi:v1.16.15-cee.1@sha256:e3f1fa5ee59887d3a15c43947a853c8a3f4fc650ed72bcd891b277037ea16a9d
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:ec8a17956e75e91b60e2999d10e67ef76f9b4cfd754db415aaf387d21690da87
        - quay.io/isovalent/startup-script-ubi:v1.16.15-cee.1@sha256:5d59a46be44c4d899576e69e0f20e148245e17d905a0769a73fd01f2432d1bab
        - quay.io/isovalent/cilium-envoy-ubi:v1.16.15-cee.1@sha256:843e6fdf6845f3d5f841dbd2dce75224b16c9d34c2b0ce7f329d9666466bfc6b
        - quay.io/isovalent/cilium-ubi:v1.16.15-cee.1@sha256:1d7bbbcb613ef20cd1719bd77473e71c198960ee5d8e563aec9c65cfd3c2efe2
        version: 1.16.15+x6419e26
      entries:
      - name: cilium-enterprise.v1.16.15-x6419e26
        version: 1.16.15+x6419e26
      - name: cilium-enterprise.v1.16.13-xb2de075
        version: 1.16.13+xb2de075
      - name: cilium-enterprise.v1.16.12-xd373871
        version: 1.16.12+xd373871
      - name: cilium-enterprise.v1.16.11-xbd1cd23
        version: 1.16.11+xbd1cd23
      - name: cilium-enterprise.v1.16.10-x89ab4b0
        version: 1.16.10+x89ab4b0
      - name: cilium-enterprise.v1.16.9-x8ec4137
        version: 1.16.9+x8ec4137
      - name: cilium-enterprise.v1.16.8-yd2b50b4
        version: 1.16.8+yd2b50b4
      - name: cilium-enterprise.v1.16.7-x4867421
        version: 1.16.7+x4867421
      - name: cilium-enterprise.v1.16.6-x7e6eb1c
        version: 1.16.6+x7e6eb1c
      name: "1.16"
    defaultChannel: "1.16"
    packageName: cilium-enterprise
    provider:
      name: Isovalent
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-commons-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: commons-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:6768e4189fa8deebf337e4732e5f12bb2ad4acdc112d9d05f5dc1a4ed81f08de
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for S3ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for S3BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/commons-operator@sha256:6768e4189fa8deebf337e4732e5f12bb2ad4acdc112d9d05f5dc1a4ed81f08de
        version: 24.11.1
      entries:
      - name: commons-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: commons-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:adc2b25a676f7357ca3971c13ecf5c9ed3cb1af24ffc902b03733cc814e8946f
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for S3ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for S3BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/commons-operator@sha256:adc2b25a676f7357ca3971c13ecf5c9ed3cb1af24ffc902b03733cc814e8946f
        version: 25.3.0
      entries:
      - name: commons-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: commons-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
        version: 25.7.0
      entries:
      - name: commons-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: commons-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
        version: 25.7.0
      entries:
      - name: commons-operator.v25.7.0
        version: 25.7.0
      - name: commons-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-commons-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: web-terminal
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: web-terminal.v1.11.1-0.1747215995.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: registry.redhat.io/web-terminal/web-terminal-rhel9-operator@sha256:0478d14e92df84fdfbf4584384d34dc9a71427a6487c2564d2eb7815ba1ac12b
          createdAt: "2021-10-26T07:24:32Z"
          description: Start a web terminal in your browser with common CLI tools
            for interacting with the cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.substitutesFor: web-terminal.v1.11.1
          operatorframework.io/suggested-namespace: openshift-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/redhat-developer/web-terminal-operator/
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - kind: DevWorkspaceRouting
            name: devworkspaceroutings.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha1
        description: |
          Start a web terminal in your browser with common CLI tools for interacting with
          the cluster.

          **Note:** The Web Terminal Operator integrates with the OpenShift Console in
          OpenShift 4.5.3 and higher to simplify web terminal instance creation and
          automate OpenShift login. In earlier versions of OpenShift, the operator can
          be installed but web terminals will have to be created and accessed manually.

          ## Description
          The Web Terminal Operator leverages the
          [DevWorkspace Operator](https://github.com/devfile/devworkspace-operator)
          to provision enviroments which support common cloud CLI tools. When this
          operator is installed, the DevWorkspace Operator will be installed as a
          dependency.

          ## How to Install
          Press the **Install** button, choose the upgrade strategy, and wait for the
          **Installed** Operator status.

          When the operator is installed, you will see a terminal button appear on the
          top right of the console after refreshing the OpenShift console window.

          ## How to Uninstall
          The Web Terminal Operator requires manual steps to fully uninstall the operator.
          As the Web Terminal Operator is designed as a way to access the OpenShift
          cluster, web terminal instances store user credentials. To avoid exposing these
          credentials to unwanted parties, the operator deploys webhooks and finalizers
          that aren't removed when the operator is uninstalled. See the
          [uninstall guide](https://docs.openshift.com/container-platform/latest/web_console/web_terminal/uninstalling-web-terminal.html)
          for more details.

          ## Documentation
          Documentation for this Operator is available at https://docs.openshift.com/container-platform/latest/web\_console/web\_terminal/installing-web-terminal.html
        displayName: Web Terminal
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - workspace
        - devtools
        - developer
        - ide
        - terminal
        links:
        - name: Web Terminal Repo
          url: https://github.com/redhat-developer/web-terminal-operator/
        maintainers:
        - email: aobuchow@redhat.com
          name: Andrew Obuchowicz
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/web-terminal/web-terminal-tooling-rhel9@sha256:5c24220f884dcdf1b1e5ac1e20dc6b7c8c4300bb89e8f118c0b11c331a56ab3f
        - registry.redhat.io/web-terminal/web-terminal-exec-rhel9@sha256:cfc8200340655a045f45d02fa327538f87e98d6369bdef2b46cf447053c44426
        - registry.redhat.io/web-terminal/web-terminal-rhel9-operator@sha256:0478d14e92df84fdfbf4584384d34dc9a71427a6487c2564d2eb7815ba1ac12b
        version: 1.11.1+0.1747215995.p
      entries:
      - name: web-terminal.v1.11.1-0.1747215995.p
        version: 1.11.1+0.1747215995.p
      - name: web-terminal.v1.11.1
        version: 1.11.1
      - name: web-terminal.v1.11.0
        version: 1.11.0
      - name: web-terminal.v1.10.1
        version: 1.10.1
      - name: web-terminal.v1.10.1-0.1740684238.p
        version: 1.10.1+0.1740684238.p
      - name: web-terminal.v1.10.0-0.1731481377.p
        version: 1.10.0+0.1731481377.p
      - name: web-terminal.v1.10.0-0.1732652667.p
        version: 1.10.0+0.1732652667.p
      - name: web-terminal.v1.10.0
        version: 1.10.0
      - name: web-terminal.v1.10.0-0.1727169028.p
        version: 1.10.0+0.1727169028.p
      - name: web-terminal.v1.10.0-0.1720435222.p
        version: 1.10.0+0.1720435222.p
      - name: web-terminal.v1.10.0-0.1720402943.p
        version: 1.10.0+0.1720402943.p
      - name: web-terminal.v1.9.0-0.1708477317.p
        version: 1.9.0+0.1708477317.p
      - name: web-terminal.v1.9.0
        version: 1.9.0
      - name: web-terminal.v1.8.0-0.1708477299.p
        version: 1.8.0+0.1708477299.p
      - name: web-terminal.v1.8.0-0.1701199376.p
        version: 1.8.0+0.1701199376.p
      - name: web-terminal.v1.8.0-0.1692219820.p
        version: 1.8.0+0.1692219820.p
      - name: web-terminal.v1.8.0
        version: 1.8.0
      - name: web-terminal.v1.7.0-0.1682321121.p
        version: 1.7.0+0.1682321121.p
      - name: web-terminal.v1.7.0-0.1692219820.p
        version: 1.7.0+0.1692219820.p
      - name: web-terminal.v1.7.0
        version: 1.7.0
      - name: web-terminal.v1.7.0-0.1681197295.p
        version: 1.7.0+0.1681197295.p
      - name: web-terminal.v1.7.0-0.1708477265.p
        version: 1.7.0+0.1708477265.p
      - name: web-terminal.v1.7.0-0.1684429884.p
        version: 1.7.0+0.1684429884.p
      - name: web-terminal.v1.6.0
        version: 1.6.0
      - name: web-terminal.v1.6.0-0.1692219820.p
        version: 1.6.0+0.1692219820.p
      - name: web-terminal.v1.5.1-0.1661829403.p
        version: 1.5.1+0.1661829403.p
      - name: web-terminal.v1.5.1
        version: 1.5.1
      - name: web-terminal.v1.5.0-0.1657220207.p
        version: 1.5.0+0.1657220207.p
      - name: web-terminal.v1.5.0
        version: 1.5.0
      - name: web-terminal.v1.4.0
        version: 1.4.0
      - name: web-terminal.v1.3.0
        version: 1.3.0
      name: fast
    defaultChannel: fast
    packageName: web-terminal
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM Quantum
      provider-url: ""
    name: openshift-qiskit-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-qiskit-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "qiskit.ibm.com/v1alpha1",
                "kind": "QiskitPlayground",
                "metadata": {
                  "name": "qiskitplayground-sample"
                },
                "spec": {
                  "image": "qiskit/jupyter:0.1",
                  "imagePullPolicy": "Always",
                  "resources": {
                    "requests": {
                      "memory": "2G",
                      "cpu": "1"
                    },
                    "limits": {
                      "memory": "3G",
                      "cpu": "1.5"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: docker.io/qiskit/operator:0.1
          createdAt: 2021-004-23 22:25:53
          description: |-
            Launches a developement environment Jupyter notebok development environment for
            developing quantum applications using Qiskit.
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator
          support: https://github.com/qiskit-community/openshift-quantum-operators/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QiskitPlayground is the Schema for the qiskitplaygrounds
              API
            displayName: Qiskit Playground
            kind: QiskitPlayground
            name: qiskitplaygrounds.qiskit.ibm.com
            version: v1alpha1
        description: |-
          ### Introduction
          Launches a developement Jupiter lab environment with Qiskit.

          QiskitPlayground is a Jupyter notebook with the pre installed qiskit libraries for implementing quantum application development.
          It comes with pre installed python packages for visualising results. QiskitPlayground can run quantum
          applications on simulator or on real quantum machines using IBM Q account.

          ### Documentation
          For detailed installation guide [follow](https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator)

          ### Contributing
          You can contribute by
            * Raising any [issues](https://github.com/qiskit-community/openshift-quantum-operators/issues) you find using ibm-quantum-operator community operator
            * Fixing issues by opening [Pull Requests](https://github.com/qiskit-community/openshift-quantum-operators/pulls)
            * Talking about openshift-qiskit-operator

          ### License
          kubernetes-qiskit-operator is licensed under the [Apache 2.0 license](https://github.com/qiskit-community/openshift-quantum-operators/blob/master/LICENSE)

          #### Note
          Any api could be changed any time without any notice. That said, your feedback is very important and appreciated to make this project more stable and useful.
        displayName: QiskitPlayground
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Qiskit
        - Open Source
        - Quantum
        links:
        - name: GitHub
          url: https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator
        maintainers:
        - email: blublinsky@ibm.com
          name: boris
        maturity: alpha
        provider:
          name: IBM Quantum
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - docker.io/qiskit/operator:0.1
        version: 0.2.1
      entries:
      - name: openshift-qiskit-operator.v0.2.1
        version: 0.2.1
      - name: openshift-qiskit-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: openshift-qiskit-operator
    provider:
      name: IBM Quantum
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: servicemeshoperator3
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: servicemeshoperator3.v3.0.0-tp.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh-tech-preview/istio-rhel9-operator@sha256:f5dfe495ba92ecb880485b66cfafbf6451f8142b8e0d5fd7f29db94171c51336
          createdAt: "2024-12-12T10:28:38Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1alpha1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1alpha1
          - description: IstioRevisionTag references a Istio or IstioRevision object
              and serves as an alias for sidecar injection.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1alpha1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-rhel9-operator@sha256:f5dfe495ba92ecb880485b66cfafbf6451f8142b8e0d5fd7f29db94171c51336
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-pilot-rhel9@sha256:376f64eab0ebeb06035983a60d1e7647c5d41354f31a49cb34f4deb7b6708e3e
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-pilot-rhel9@sha256:ea9372fabe06cd8e65ea7b4567ce3c8fbddd9d673dc34a12743fd95c548059d1
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-proxyv2-rhel9@sha256:c21481d6dae58b82dac93b1b47b9d9f53a30617fa9d930a0b5d075620cffa74b
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-must-gather-rhel9@sha256:2bf3704ba8d5945973ccf6bd13418457e929836352acf0ea6a47976ffa246616
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-proxyv2-rhel9@sha256:cbd6bc47341ab8618609f89d063552e17232541c727429855c01deab73e4d06c
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-must-gather-rhel9@sha256:e3978aecaafe2678355e21a7303f8620f53d8b774f6592f10527c15bae957224
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-cni-rhel9@sha256:d0ab36246c556249b6d67346eb0c72ebd22483a6502caa0fc75112a8d327b053
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-cni-rhel9@sha256:f16cd8a29ddcbb3de77cbfcf7e1da78e42cc0c1c502132d477a01c52689cdf23
        version: 3.0.0-tp.2
      entries:
      - name: servicemeshoperator3.v3.0.0-tp.2
        version: 3.0.0-tp.2
      - name: servicemeshoperator3.v3.0.0-tp.1
        version: 3.0.0-tp.1
      name: candidates
    - currentCSV: servicemeshoperator3.v3.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.4"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.4"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
          createdAt: "2025-09-10T06:22:42Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:366266f10e658c4cea86bddf6ee847e1908ea4dd780cb5f7fb0e466bac8995f1
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:0d6930075b8906bee0ab1aeac82f6e2431c16163cf87f7f0c68c10b2c40552a1
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:26747627ae22bbdffcf9de58077454fc0c890cda83659d7834b15dea2b5aaaf2
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2ea112ab90b8540f11e9949d77c3e7e3b3ef57ac3bf23f6cf1e883a88430e1f9
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:028e10651db0d1ddb769a27c9483c6d41be6ac597f253afd9d599f395d9c82d8
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:ecc6a22a471f4f4b836648f3a72ea7f57a4ea960ebcbc7dbe3fe729500902d0b
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d518f3d1539f45e1253c5c9fa22062802804601d4998cd50344e476a3cc388fe
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:33feac0c9c888b375e207e86b3f9a1a289126ca46b64632a7f693e2191cfbbfd
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:2ca96716812d339775d3dfd396c863834dd3810720e98bfcf87bc67c5fbd29b5
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:6bb44c84363a4113c5bce2ae0d6528b345fddd12bc559bbfaacf6222060ddcb0
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:6881fb9f5a4b447be4c468b8585fff86171bd3d82671a48559f3ceabcb9dad24
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:14c5a52faf20267baa43d9a72ee6416f56fbc41dd640adc2aba3c4043802a0e9
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:b757b81d34711cb8abfcb0e8896c023d39d4778f237dd80b637ba4bf613df893
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:f7ab868455b5e887e2d89b7078678531290365f9166cae99ce56119f31fa2ba4
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d757e700a124d7583b9456ec9e08e5e7e249c1de202bc4e973102922b9f69191
        version: 3.1.2
      entries:
      - name: servicemeshoperator3.v3.1.2
        version: 3.1.2
      - name: servicemeshoperator3.v3.1.1
        version: 3.1.1
      - name: servicemeshoperator3.v3.1.0
        version: 3.1.0
      - name: servicemeshoperator3.v3.0.5
        version: 3.0.5
      - name: servicemeshoperator3.v3.0.4
        version: 3.0.4
      - name: servicemeshoperator3.v3.0.3
        version: 3.0.3
      - name: servicemeshoperator3.v3.0.2
        version: 3.0.2
      - name: servicemeshoperator3.v3.0.1
        version: 3.0.1
      - name: servicemeshoperator3.v3.0.0
        version: 3.0.0
      name: stable
    - currentCSV: servicemeshoperator3.v3.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.6"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.6"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:d716f3558fbdf8514f225de708135ba87fdcac4c718ef101f394daf546514afb
          createdAt: "2025-09-03T07:53:34Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:d716f3558fbdf8514f225de708135ba87fdcac4c718ef101f394daf546514afb
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        version: 3.0.5
      entries:
      - name: servicemeshoperator3.v3.0.5
        version: 3.0.5
      - name: servicemeshoperator3.v3.0.4
        version: 3.0.4
      - name: servicemeshoperator3.v3.0.3
        version: 3.0.3
      - name: servicemeshoperator3.v3.0.2
        version: 3.0.2
      - name: servicemeshoperator3.v3.0.1
        version: 3.0.1
      - name: servicemeshoperator3.v3.0.0
        version: 3.0.0
      name: stable-3.0
    - currentCSV: servicemeshoperator3.v3.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.4"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.4"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
          createdAt: "2025-09-10T06:22:42Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:028e10651db0d1ddb769a27c9483c6d41be6ac597f253afd9d599f395d9c82d8
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d518f3d1539f45e1253c5c9fa22062802804601d4998cd50344e476a3cc388fe
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:366266f10e658c4cea86bddf6ee847e1908ea4dd780cb5f7fb0e466bac8995f1
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:33feac0c9c888b375e207e86b3f9a1a289126ca46b64632a7f693e2191cfbbfd
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:ecc6a22a471f4f4b836648f3a72ea7f57a4ea960ebcbc7dbe3fe729500902d0b
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:b757b81d34711cb8abfcb0e8896c023d39d4778f237dd80b637ba4bf613df893
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:14c5a52faf20267baa43d9a72ee6416f56fbc41dd640adc2aba3c4043802a0e9
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:6bb44c84363a4113c5bce2ae0d6528b345fddd12bc559bbfaacf6222060ddcb0
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:0d6930075b8906bee0ab1aeac82f6e2431c16163cf87f7f0c68c10b2c40552a1
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:6881fb9f5a4b447be4c468b8585fff86171bd3d82671a48559f3ceabcb9dad24
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d757e700a124d7583b9456ec9e08e5e7e249c1de202bc4e973102922b9f69191
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2ea112ab90b8540f11e9949d77c3e7e3b3ef57ac3bf23f6cf1e883a88430e1f9
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:26747627ae22bbdffcf9de58077454fc0c890cda83659d7834b15dea2b5aaaf2
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:f7ab868455b5e887e2d89b7078678531290365f9166cae99ce56119f31fa2ba4
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:2ca96716812d339775d3dfd396c863834dd3810720e98bfcf87bc67c5fbd29b5
        version: 3.1.2
      entries:
      - name: servicemeshoperator3.v3.1.2
        version: 3.1.2
      - name: servicemeshoperator3.v3.1.1
        version: 3.1.1
      - name: servicemeshoperator3.v3.1.0
        version: 3.1.0
      name: stable-3.1
    defaultChannel: stable
    packageName: servicemeshoperator3
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: heat-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: heat-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "heat.openstack.org/v1beta1",
                "kind": "Heat",
                "metadata": {
                  "name": "heat"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = True\n",
                  "databaseInstance": "openstack",
                  "heatAPI": {},
                  "heatCfnAPI": {},
                  "heatEngine": {},
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "heat.openstack.org/v1beta1",
                "kind": "HeatAPI",
                "metadata": {
                  "name": "heat-api"
                },
                "spec": {
                  "databaseAccount": "heat",
                  "databaseHostname": "openstack",
                  "passwordSelectors": {
                    "authEncryptionKey": "HeatAuthEncryptionKey",
                    "service": "HeatPassword"
                  },
                  "replicas": 1,
                  "resources": {},
                  "secret": "osp-secret",
                  "serviceAccount": "heat-heat",
                  "serviceUser": "heat",
                  "transportURLSecret": "rabbitmq-transport-url-heat-heat-transport"
                }
              },
              {
                "apiVersion": "heat.openstack.org/v1beta1",
                "kind": "HeatCfnAPI",
                "metadata": {
                  "name": "heat-cfnapi"
                },
                "spec": {
                  "databaseAccount": "heat",
                  "databaseHostname": "openstack",
                  "passwordSelectors": {
                    "authEncryptionKey": "HeatAuthEncryptionKey",
                    "service": "HeatPassword"
                  },
                  "replicas": 1,
                  "resources": {},
                  "secret": "osp-secret",
                  "serviceAccount": "heat-heat",
                  "serviceUser": "heat",
                  "transportURLSecret": "rabbitmq-transport-url-heat-heat-transport"
                }
              },
              {
                "apiVersion": "heat.openstack.org/v1beta1",
                "kind": "HeatEngine",
                "metadata": {
                  "name": "heat-engine"
                },
                "spec": {
                  "databaseAccount": "heat",
                  "databaseHostname": "openstack",
                  "passwordSelectors": {
                    "authEncryptionKey": "HeatAuthEncryptionKey",
                    "service": "HeatPassword"
                  },
                  "replicas": 1,
                  "resources": {},
                  "secret": "osp-secret",
                  "serviceAccount": "heat-heat",
                  "serviceUser": "heat",
                  "transportURLSecret": "rabbitmq-transport-url-heat-heat-transport"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:16Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HeatAPI is the Schema for the heatapis API
            displayName: Heat API
            kind: HeatAPI
            name: heatapis.heat.openstack.org
            version: v1beta1
          - description: HeatCfnAPI is the Schema for the heatcfnapis API
            displayName: Heat Cfn API
            kind: HeatCfnAPI
            name: heatcfnapis.heat.openstack.org
            version: v1beta1
          - description: HeatEngine is the Schema for the heatengines API
            displayName: Heat Engine
            kind: HeatEngine
            name: heatengines.heat.openstack.org
            version: v1beta1
          - description: Heat is the Schema for the heats API
            displayName: Heat
            kind: Heat
            name: heats.heat.openstack.org
            version: v1beta1
        description: Heat Operator
        displayName: Heat Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Orchestration
        - Heat
        links:
        - name: Heat Operator
          url: https://github.com/openstack-k8s-operators/heat-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/heat-rhel9-operator@sha256:fa61aeee8d63087d150a7641d1e8d4b67dd067bcfed6a140401fe396f1e32403
        - registry.redhat.io/rhoso/openstack-heat-api-rhel9@sha256:dba236e5fed89fbe46397b6ab01de5795e66e9194f5cd99fdf137c37a7b96135
        - registry.redhat.io/rhoso/openstack-heat-api-cfn-rhel9@sha256:95548184e68325a5fe1604d5c4d35d1e9cca7f39b84303fce50479e0e35f8280
        - registry.redhat.io/rhoso/openstack-heat-engine-rhel9@sha256:0b6fc782986bf11b26c0839fd81c52257f4a31aa5e5a6728ce251f3670537aba
        version: 1.0.6
      entries:
      - name: heat-operator.v1.0.6
        version: 1.0.6
      - name: heat-operator.v1.0.5
        version: 1.0.5
      - name: heat-operator.v1.0.4
        version: 1.0.4
      - name: heat-operator.v1.0.3
        version: 1.0.3
      - name: heat-operator.v1.0.2
        version: 1.0.2
      - name: heat-operator.v1.0.1
        version: 1.0.1
      - name: heat-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: heat-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CrowdStrike
      provider-url: https://crowdStrike.com
    name: falcon-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: falcon-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconAdmission",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-admission",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconadmission",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-kac"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "admission_controller"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconContainer",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-sidecar-sensor",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconcontainer",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-container-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "sidecar"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconDeployment",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcondeployment-sample",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falcon-deployment",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-deployment"
                },
                "spec": {
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconImageAnalyzer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "falcon-operator",
                    "app.kubernetes.io/instance": "falconimageanalyzer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "falconimageanalyzer",
                    "app.kubernetes.io/part-of": "falcon-operator"
                  },
                  "name": "falcon-image-analyzer"
                },
                "spec": {
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconNodeSensor",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-node-sensor",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconnodesensor",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-node-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "daemonset"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security,Monitoring
          containerImage: registry.connect.redhat.com/crowdstrike/falcon-operator@sha256:ffe00e9e22e209d1e08f29c3d257bbd51937fcbde68a1ed93b542a26dd436dee
          createdAt: "2025-06-03T22:34:20Z"
          description: Falcon Operator installs CrowdStrike Falcon Sensors on the
            cluster
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: falcon-operator
          operators.openshift.io/valid-subscription: Not required to use this operator
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: registry.connect.redhat.com/crowdstrike/falcon-operator
          support: CrowdStrike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FalconAdmission is the Schema for the falconadmissions API
            displayName: Falcon Admission
            kind: FalconAdmission
            name: falconadmissions.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconContainer is the Schema for the falconcontainers API
            displayName: Falcon Container
            kind: FalconContainer
            name: falconcontainers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconDeployment is the Schema for the falcondeployments
              API
            displayName: Falcon Deployment
            kind: FalconDeployment
            name: falcondeployments.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconImageAnalyzer is the Schema for the falconImageAnalyzers
              API
            displayName: Falcon Image Analyzer
            kind: FalconImageAnalyzer
            name: falconimageanalyzers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconNodeSensor is the Schema for the falconnodesensors
              API
            displayName: Falcon Node Sensor
            kind: FalconNodeSensor
            name: falconnodesensors.falcon.crowdstrike.com
            version: v1alpha1
        description: |-
          The CrowdStrike Falcon Operator installs CrowdStrike Falcon custom resources on a Kubernetes cluster.

          ## About the CrowdStrike Falcon Operator
          The CrowdStrike Falcon Operator deploys CrowdStrike Falcon to the cluster. The operator exposes custom resources that allow you to protect your Kubernetes clusters when deployed.

          ## About Custom Resources

          | Custom Resource                                       | Description                                                      |
          | :--                                             | :--                                                    |
          | [FalconAdmission](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/admission/README.md) | Manages installation of Falcon Admission Controller on the cluster |
          | [FalconImageAnalyzer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/imageanalyzer/README.md) | Manages installation of Falcon Image Assessment at Runtime on the cluster |
          | [FalconContainer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/container/README.md) | Manages installation of Falcon Container Sensor on the cluster   |
          | [FalconNodeSensor](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/node/README.md)     | Manages installation of Falcon Linux Sensor on the cluster nodes |
          | [FalconDeployment](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/falcondeployment/README.md)       | Deploys FalconAdmission, FalconImageAnalyzer, FalconContainer, and FalconNodeSensor CRs from a single manifest |

          ## Installation and Deployment

          For installation and deployment of the CrowdStrike Falcon Operator and its Custom Resources, please read the [Installation and Deployment Guide](./docs/install_guide.md) and choose the deployment method that is right for your target environment.

          ## Getting Help
          If you encounter any issues while using the Falcon Operator, you can create an issue on our [Github repo](https://github.com/CrowdStrike/falcon-operator) for bugs, enhancements, or other requests.

          ## Contributing
          You can contribute by:

          * Raising any issues you find using Falcon Operator
          * Fixing issues by opening [Pull Requests](https://github.com/CrowdStrike/falcon-operator/pulls)
          * Improving documentation
          * Talking about the CrowdStrike Falcon Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/CrowdStrike/falcon-operator/issues).

          ## Additional Resources
           - CrowdStrike Container Security: [Product Page](https://www.crowdstrike.com/products/cloud-security/falcon-cloud-workload-protection/container-security/)
           - So You Think Your Containers Are Secure? Four Steps to Ensure a Secure Container Deployment: [Blog Post](https://www.crowdstrike.com/blog/four-steps-to-ensure-a-secure-containter-deployment/)
           - Container Security With CrowdStrike: [Blog Post](https://www.crowdstrike.com/blog/tech-center/container-security/)
           - To learn more about Falcon Container Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/support/documentation/146/falcon-container-sensor-for-linux), [Release Notes](https://falcon.crowdstrike.com/support/news/release-notes-falcon-container-sensor-for-linux)
           - To learn more about Falcon Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/documentation/20/falcon-sensor-for-linux#kubernetes-support)
           - [Developer Documentation](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/developer_guide.md)
        displayName: CrowdStrike Falcon Platform - Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - runtime-protection
        links:
        - name: Falcon Operator
          url: https://github.com/CrowdStrike/falcon-operator
        maintainers:
        - email: integrations@crowdstrike.com
          name: CrowdStrike Solutions Architecture
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: CrowdStrike
          url: https://crowdStrike.com
        relatedImages:
        - registry.connect.redhat.com/crowdstrike/falcon-operator@sha256:ffe00e9e22e209d1e08f29c3d257bbd51937fcbde68a1ed93b542a26dd436dee
        - registry.crowdstrike.com/falcon-sensor/us-1/release/falcon-sensor@sha256:ed4f3fe8d9f5f491600f18d4440dc1d8e3a07d4b9c08a1f17382a23a3a5cf523
        - registry.crowdstrike.com/falcon-container/us-1/release/falcon-sensor@sha256:92d222cd604c9f2c9e9f2c9b402b9fb0858fcb04dac38c15e25125978f43c99b
        - registry.crowdstrike.com/falcon-kac/us-1/release/falcon-kac@sha256:5049128dcfe9565f39fc6e28fb94f5bf05fb7c51cf787fc6fe5c63a2b04f800d
        - registry.crowdstrike.com/falcon-imageanalyzer/us-1/release/falcon-imageanalyzer@sha256:f08603648658041ad52ccf0cfd058da40b385904830030aa82dc50609aed7ceb
        version: 1.8.0
      entries:
      - name: falcon-operator.v1.8.0
        version: 1.8.0
      - name: falcon-operator.v1.7.0
        version: 1.7.0
      - name: falcon-operator.v1.6.0
        version: 1.6.0
      - name: falcon-operator.v1.5.0
        version: 1.5.0
      - name: falcon-operator.v1.4.0
        version: 1.4.0
      - name: falcon-operator.v1.3.1
        version: 1.3.1
      - name: falcon-operator.v1.3.0
        version: 1.3.0
      - name: falcon-operator.v1.2.0
        version: 1.2.0
      - name: falcon-operator.v1.1.0
        version: 1.1.0
      - name: falcon-operator.v1.0.1
        version: 1.0.1
      - name: falcon-operator.v1.0.0
        version: 1.0.0
      name: certified-1.0
    defaultChannel: certified-1.0
    packageName: falcon-operator
    provider:
      name: CrowdStrike
      url: https://crowdStrike.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Airlock
      provider-url: ""
    name: airlock-microgateway
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airlock-microgateway.v4.5.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControl","metadata":{"name":"access-control-example"},"spec":{"policies":[{"requestConditions":{"path":{"matcher":{"exact":"/secure"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}},{"oidc":{"claim":{"name":"roles","value":{"matcher":{"contains":"\"admin\""}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"oidc-example"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"identity-propagation-example"}}],"onFailure":"Pass"}},{"authorization":{}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurity","metadata":{"name":"content-security-example"},"spec":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1"},"spec":{"value":{"name":"ext-authz","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}}}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"lifetime":"16h"},"prefix":"session_prefix_example"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SidecarGateway","metadata":{"name":"sidecar-gateway-example"},"spec":{"podSelector":{"matchLabels":{"app":"example-app"}},"sessionHandlingRef":{"name":"session-handling-example"},"applications":[{"containerPort":8443,"routes":[{"pathPrefix":"/metrics","unsecured":{}},{"pathPrefix":"/","secured":{"accessControlRef":{"name":"access-control-example"},"contentSecurityRef":{"name":"content-security-example"}}}],"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}},"restrictions":{"http":{"headersLength":"80Ki"}},"protocol":{"auto":{}},"tls":{"enable":true,"secretRef":{"name":"example-tls-secret"},"protocol":{"minimum":"TLSv1_3"},"clientCertificate":{"ignored":{}},"xfcc":"AlwaysForwardOnly"}},"upstream":{"tls":{"enable":true,"protocol":{"minimum":"TLSv1_2"}}},"envoyHTTPFilterRefs":{"prepend":[{"name":"envoy-http-filter-example-1"}]},"telemetryRef":{"name":"telemetry-example"}}],"envoyClusterRefs":[{"name":"envoy-cluster-example-1"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-06-13T17:33:31Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: AccessControl
            name: accesscontrols.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurity
            name: contentsecurities.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: SidecarGateway
            name: sidecargateways.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056)
          * Gateway API
          * (Optional) Airlock Microgateway CNI (Required for [data plane mode sidecar](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000137)).

          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ### (Optional) Install Airlock Microgateway CNI
          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          Refer to [our documentation](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000140) for installation instructions

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD_MONITOR\_CREATE | "false" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.5/
        - name: Airlock Microgateway Labs
          url: https://play.instruqt.com/airlock/invite/hyi9fy4b4jzc?icp_referrer=openshift
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:806366a0b3f78844fe14775fc290fbfbde8d2a701d34a6478b019a0e7140b699
        version: 4.5.4
      entries:
      - name: airlock-microgateway.v4.5.4
        version: 4.5.4
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.5"
    - currentCSV: airlock-microgateway.v4.6.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControl","metadata":{"name":"access-control-example"},"spec":{"policies":[{"requestConditions":{"path":{"matcher":{"exact":"/secure"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}},{"oidc":{"claim":{"name":"roles","value":{"matcher":{"contains":"\"admin\""}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"oidc-example"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"token-exchange-example"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"identity-propagation-example"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"jwt-example"}}}}},{"authorization":{}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurity","metadata":{"name":"content-security-example"},"spec":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1"},"spec":{"value":{"name":"ext-authz","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}}}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"lifetime":"16h"},"prefix":"session_prefix_example"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SidecarGateway","metadata":{"name":"sidecar-gateway-example"},"spec":{"podSelector":{"matchLabels":{"app":"example-app"}},"sessionHandlingRef":{"name":"session-handling-example"},"applications":[{"containerPort":8443,"routes":[{"pathPrefix":"/metrics","unsecured":{}},{"pathPrefix":"/","secured":{"accessControlRef":{"name":"access-control-example"},"contentSecurityRef":{"name":"content-security-example"}}}],"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}},"restrictions":{"http":{"headersLength":"80Ki"}},"protocol":{"auto":{}},"tls":{"enable":true,"secretRef":{"name":"example-tls-secret"},"protocol":{"minimum":"TLSv1_3"},"clientCertificate":{"ignored":{}},"xfcc":"AlwaysForwardOnly"}},"upstream":{"tls":{"enable":true,"protocol":{"minimum":"TLSv1_2"}}},"envoyHTTPFilterRefs":{"prepend":[{"name":"envoy-http-filter-example-1"}]},"telemetryRef":{"name":"telemetry-example"}}],"envoyClusterRefs":[{"name":"envoy-cluster-example-1"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-23T06:27:21Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: AccessControl
            name: accesscontrols.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurity
            name: contentsecurities.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: SidecarGateway
            name: sidecargateways.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056)
          * Gateway API
          * (Optional) Airlock Microgateway CNI (Required for [data plane mode sidecar](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000137)).

          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.3.0/standard-install.yaml
          ```

          ### (Optional) Install Airlock Microgateway CNI
          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          Refer to [our documentation](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000140) for installation instructions

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD_MONITOR\_CREATE | "false" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.6/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:8fa300ced5b69acbad824c037d2b344fd1926bcfb39e264e9f0f055ceb9bd68d
        version: 4.6.4
      entries:
      - name: airlock-microgateway.v4.6.4
        version: 4.6.4
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.6"
    - currentCSV: airlock-microgateway.v4.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1","namespace":"default"},"spec":{"value":{"name":"envoycluster/default/envoy-cluster-example-1","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyExtensionPolicy","metadata":{"name":"envoy-extension-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"envoyHTTPFilterRefs":{"prepend":[{"name":"geoip-blocker"}],"append":[{"name":"lua-transform-request"}]},"envoyClusterRefs":[{"name":"geoip-server"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"volumes":[{"name":"vol-1","emptyDir":{}}],"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}},"volumeMounts":[{"name":"vol-1","mountPath":"/tmp/data"}]}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"telemetryRef":{"name":"telemetry-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"mode":"Enforce","persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"idle":"30s","lifetime":"16h"},"prefix":"session_prefix_example","cookie":{"name":"SESS","attributes":{"domain":"example.org","path":"/","sameSite":"Strict","secure":true}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"},"session":{"id":"%SESSION_ID%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-03T13:02:23Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyExtensionPolicy
            name: envoyextensionpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ***Note:*** Sidecar-based data plane mode is *not* supported by this OLM Bundle.
          If you want to use the Sidecar-based data plane mode, use Helm instead. Please note that the Airlock Microgateway CNI and the Helm charts are not certified for RedHat OpenShift.
          Refer to https://github.com/airlock/microgateway for installation instruction with Helm.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056)
          * [Gateway API](https://gateway-api.sigs.k8s.io/)

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ***Note:*** *This step is only required for OpenShift versions below 4.19, as Gateway API is included by default in OpenShift 4.19+.*

          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD\_MONITOR\_CREATE | "true" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gateway API
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.7/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:d4ce35184922953c209a9ebb234ca3370dbfc5054de411b5503c26ac2730b7ae
        version: 4.7.1
      entries:
      - name: airlock-microgateway.v4.7.1
        version: 4.7.1
      - name: airlock-microgateway.v4.7.0
        version: 4.7.0
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.7"
    - currentCSV: airlock-microgateway.v4.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1","namespace":"default"},"spec":{"value":{"name":"envoycluster/default/envoy-cluster-example-1","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyExtensionPolicy","metadata":{"name":"envoy-extension-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"envoyHTTPFilterRefs":{"prepend":[{"name":"geoip-blocker"}],"append":[{"name":"lua-transform-request"}]},"envoyClusterRefs":[{"name":"geoip-server"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"volumes":[{"name":"vol-1","emptyDir":{}}],"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}},"volumeMounts":[{"name":"vol-1","mountPath":"/tmp/data"}]}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"telemetryRef":{"name":"telemetry-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"mode":"Enforce","persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"idle":"30s","lifetime":"16h"},"prefix":"session_prefix_example","cookie":{"name":"SESS","attributes":{"domain":"example.org","path":"/","sameSite":"Strict","secure":true}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"},"session":{"id":"%SESSION_ID%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-03T13:02:23Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyExtensionPolicy
            name: envoyextensionpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ***Note:*** Sidecar-based data plane mode is *not* supported by this OLM Bundle.
          If you want to use the Sidecar-based data plane mode, use Helm instead. Please note that the Airlock Microgateway CNI and the Helm charts are not certified for RedHat OpenShift.
          Refer to https://github.com/airlock/microgateway for installation instruction with Helm.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056)
          * [Gateway API](https://gateway-api.sigs.k8s.io/)

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ***Note:*** *This step is only required for OpenShift versions below 4.19, as Gateway API is included by default in OpenShift 4.19+.*

          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD\_MONITOR\_CREATE | "true" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gateway API
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.7/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:d4ce35184922953c209a9ebb234ca3370dbfc5054de411b5503c26ac2730b7ae
        version: 4.7.1
      entries:
      - name: airlock-microgateway.v4.7.1
        version: 4.7.1
      - name: airlock-microgateway.v4.7.0
        version: 4.7.0
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: stable
    defaultChannel: "4.7"
    packageName: airlock-microgateway
    provider:
      name: Airlock
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: must-gather-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: must-gather-operator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "MustGather",
                "metadata": {
                  "name": "example-mustgather"
                },
                "spec": {
                  "caseID": "02527285",
                  "caseManagementAccountSecretRef": {
                    "name": "case-management-creds"
                  },
                  "serviceAccountRef": {
                    "name": "must-gather-admin"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security, Monitoring
          certified: "false"
          containerImage: quay.io/redhat-cop/must-gather-operator@sha256:aebef122cdb8409f5a5d70d3cdc585e02272704d57e103553a84bc6fc238e646
          createdAt: "2021-12-02T17:13:53Z"
          description: This operator provides a facility to easily upload must-gather
            reports to a Red Hat case.
          operatorframework.io/suggested-namespace: global-load-balancer-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/must-gather-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: MustGather
            name: mustgathers.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The must gather operator helps collecting must gather information on a cluster and uploading it to a case.
          To use the operator a cluster administrator can create the following must gather CR:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: MustGather
          metadata:
            name: example-mustgather
          spec:
            caseID: '02527285'
            caseManagementAccountSecretRef:
              name: case-management-creds
          ```

          this request will collect the standard must gather info and upload it to case `#02527285` using the credentials found in the `caseManagementCreds` secret.

          A more complex example:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: MustGather
          metadata:
            name: full-mustgather
          spec:
            caseID: '02527285'
            caseManagementAccountSecretRef:
              name: case-management-creds
            serviceAccountRef:
              name: must-gather-admin
            mustGatherImages:
            - quay.io/kubevirt/must-gather:latest
            - quay.io/ocs-dev/ocs-must-gather
          ```

          in this example we are using a specific service account (which must have cluster admin permissions as per must-gather requirements) and we are specifying a couple of additional must gather images to be run for the `kubevirt` and `ocs` subsystem. If not specified serviceAccountRef.Name will default to `default`. Also the standard must gather image: `quay.io/openshift/origin-must-gather:latest` is always added by default.
        displayName: Must Gather Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - case_management
        - support
        - must_gather
        links:
        - name: Must Gather Operator
          url: https://github.com/redhat-cop/must-gather-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/must-gather-operator
        maintainers:
        - email: rspazzol@redhat.com'
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/must-gather-operator@sha256:aebef122cdb8409f5a5d70d3cdc585e02272704d57e103553a84bc6fc238e646
        version: 1.1.2
      entries:
      - name: must-gather-operator.v1.1.2
        version: 1.1.2
      - name: must-gather-operator.v1.1.1
        version: 1.1.1
      - name: must-gather-operator.v1.1.0
        version: 1.1.0
      - name: must-gather-operator.v1.0.6
        version: 1.0.6
      - name: must-gather-operator.v1.0.5
        version: 1.0.5
      - name: must-gather-operator.v1.0.4
        version: 1.0.4
      - name: must-gather-operator.v1.0.3
        version: 1.0.3
      - name: must-gather-operator.v1.0.2
        version: 1.0.2
      - name: must-gather-operator.v1.0.1
        version: 1.0.1
      - name: must-gather-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: must-gather-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: machine-deletion-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: machine-deletion-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/machine-deletion-remediation-operator:v0.1.0
          createdAt: "2023-05-11 10:46:12"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/medik8s/machine-deletion-remediation-operator:v0.1.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        version: 0.1.0
      entries:
      - name: machine-deletion-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: machine-deletion-remediation.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/machine-deletion-remediation-operator:v0.3.0
          createdAt: "2024-03-05 18:45:16"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/medik8s/machine-deletion-remediation-operator:v0.3.0
        version: 0.3.0
      entries:
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      - name: machine-deletion-remediation.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: machine-deletion-remediation
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-nimbusmosaic-dusim
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-nimbusmosaic-dusim.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Nimbusmosaicdusim",
                "metadata": {
                  "name": "nimbusmosaicdusim-sample"
                },
                "spec": {
                  "development": false,
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "microservices": {
                    "jvm": "-Xmx300M -Xms64M -Xss256k"
                  },
                  "service": {
                    "type": "ClusterIP"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:41:56Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nimbusmosaicdusim
            name: nimbusmosaicdusims.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Nimbus Mosaic DUSim Operator
        displayName: Keysight Nimbus Mosaic DUSim Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Nimbusmosaic Dusim
          url: https://keysight-nimbusmosaic-dusim.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-nimbusmosaic-dusim@sha256:655dd6595a9c8e82947bed24991ac52f8fa55be99ca410e7d9f0026c51ef4520
        version: 0.1.21
      entries:
      - name: keysight-nimbusmosaic-dusim.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-nimbusmosaic-dusim
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-namespace-scope-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-namespace-scope-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1",
                "kind": "NamespaceScope",
                "metadata": {
                  "name": "namespacescope"
                },
                "spec": {
                  "namespaceMembers": [
                    "ibm-common-services",
                    "default"
                  ],
                  "restartLabels": {
                    "intent": "projected"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
          createdAt: "2025-02-19T00:54:24Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.2.14
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/IBM/ibm-namespace-scope-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NamespaceScope is the Schema for the namespacescopes API
            displayName: Namespace Scope
            kind: NamespaceScope
            name: namespacescopes.operator.ibm.com
            version: v1
        description: This operator automates the extension of operator watch and service
          account permission scope to other namespaces in an openshift cluster.
        displayName: IBM NamespaceScope Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Namespace Scope Operator
          url: https://github.com/IBM/ibm-namespace-scope-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-bundle@sha256:b6cb9e5976b749dd7818ac19d1cfd5e24673f3c052a41d6e2fe1d2d7094694b4
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:7a2cd72736bf9d30341763fd1652ff4b9ce3c785da5123645ee7664ee38c7283
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-catalog@sha256:c016904e8c5cfcf8270241ad5e28b52f528859984ed894cdd5e1283fa90cd6e8
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator-bundle@sha256:c743dd7de2edbd08c6c73e20cf15f3b147491ad77a81677339e88caffb93e84b
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:0774b1d916ebc354050612474800dda3a501c66a7a5aab6d18698c8c4e81ac4e
        version: 4.2.14
      entries:
      - name: ibm-namespace-scope-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-namespace-scope-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hitachi
      provider-url: ""
    name: hnas-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hnas-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.csi.hitachi.com/v1",
                "kind": "HNAS",
                "metadata": {
                  "name": "hnas",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-operator@sha256:9d3f9f10fc7172ee9cebbd9ee318906756d7e96909f2e66e0534e98bb53e0a7b
          createdAt: "2025-07-30T15:12:23Z"
          description: An operator for managing Hitachi NAS CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HNAS is the Schema for the hnas API
            displayName: HNAS
            kind: HNAS
            name: hnas.charts.csi.hitachi.com
            version: v1
        description: |-
          The Hitachi NAS CSI driver operator deploys the Hitachi NAS CSI driver on Kubernetes and OpenShift environments.  The driver can be used to create and manage persistent storage hosted on Hitachi NAS products, including those in the Hitachi VSP One File 3x range and the HNAS range of products.

          For the complete documentation, go to our [Product Documentation site](https://docs.hitachivantara.com/home) and refer to the Hitachi NAS CSI Driver User Guide - MK-92HNAS091.  The guide covers prerequisites for the NAS system and covers the driver functionality and all configuration options available for the driver.
        displayName: Hitachi NAS CSI Driver Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hitachi
        - Storage
        - Container Storage Interface
        - CSI
        - NAS
        links:
        - name: Product Documentation
          url: https://docs.hitachivantara.com/home
        maturity: stable
        minKubeVersion: 1.28.0
        provider:
          name: Hitachi
        relatedImages:
        - registry.k8s.io/sig-storage/csi-attacher@sha256:a399393ff5bd156277c56bae0c08389b1a1b95b7fd6ea44a316ce55e0dd559d7
        - registry.k8s.io/sig-storage/csi-resizer@sha256:ab774734705a906561e15b67f6a96538f3ad520335d636f793aaafe87a3b5200
        - registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-operator@sha256:9d3f9f10fc7172ee9cebbd9ee318906756d7e96909f2e66e0534e98bb53e0a7b
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:dd788d79cf4c1b8edee6d9b80b8a1ebfc51a38a365c5be656986b129be9ac784
        - registry.k8s.io/sig-storage/livenessprobe@sha256:2c5f9dc4ea5ac5509d93c664ae7982d4ecdec40ca7b0638c24e5b16243b8360f
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:d7138bcc3aa5f267403d45ad4292c95397e421ea17a0035888850f424c7de25d
        - registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-driver@sha256:6b6f6816c9f4e608e93e7fbc0347255415426a0ff0c9a53b7750e5ba54d91262
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d5e46da8aff7d73d6f00c761dae94472bcda6e78f4f17b3802dc89d44de0111b
        version: 1.0.0
      entries:
      - name: hnas-csi-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: hnas-csi-operator
    provider:
      name: Hitachi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: marin3r
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marin3r.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyConfig",
                "metadata": {
                  "name": "envoyconfig-example",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "nodeID": "example",
                  "resources": [
                    {
                      "type": "cluster",
                      "value": {
                        "connect_timeout": "0.01s",
                        "dns_lookup_family": "V4_ONLY",
                        "lb_policy": "ROUND_ROBIN",
                        "load_assignment": {
                          "cluster_name": "example",
                          "endpoints": [
                            {
                              "lb_endpoints": [
                                {
                                  "endpoint": {
                                    "address": {
                                      "socket_address": {
                                        "address": "example",
                                        "port_value": 8080
                                      }
                                    }
                                  }
                                }
                              ]
                            }
                          ]
                        },
                        "name": "example",
                        "type": "STRICT_DNS"
                      }
                    },
                    {
                      "type": "route",
                      "value": {
                        "name": "local",
                        "virtual_hosts": [
                          {
                            "domains": [
                              "*"
                            ],
                            "name": "all",
                            "routes": [
                              {
                                "match": {
                                  "prefix": "/"
                                },
                                "route": {
                                  "cluster": "example"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    {
                      "type": "listener",
                      "value": {
                        "address": {
                          "socket_address": {
                            "address": "0.0.0.0",
                            "port_value": 8443
                          }
                        },
                        "filter_chains": [
                          {
                            "filters": [
                              {
                                "name": "envoy.filters.network.http_connection_manager",
                                "typed_config": {
                                  "@type": "type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager",
                                  "access_log": [
                                    {
                                      "name": "envoy.access_loggers.file",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.access_loggers.file.v3.FileAccessLog",
                                        "path": "/dev/stdout"
                                      }
                                    }
                                  ],
                                  "http_filters": [
                                    {
                                      "name": "envoy.filters.http.router",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.filters.http.router.v3.Router"
                                      }
                                    }
                                  ],
                                  "rds": {
                                    "config_source": {
                                      "ads": {},
                                      "resource_api_version": "V3"
                                    },
                                    "route_config_name": "local"
                                  },
                                  "stat_prefix": "https"
                                }
                              }
                            ],
                            "transport_socket": {
                              "name": "envoy.transport_sockets.tls",
                              "typed_config": {
                                "@type": "type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext",
                                "common_tls_context": {
                                  "tls_certificate_sds_secret_configs": [
                                    {
                                      "name": "example.default.svc",
                                      "sds_config": {
                                        "ads": {},
                                        "resource_api_version": "V3"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          }
                        ],
                        "name": "https"
                      }
                    },
                    {
                      "blueprint": "tlsCertificate",
                      "generateFromTlsSecret": "example.default.svc",
                      "type": "secret"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "DiscoveryService",
                "metadata": {
                  "name": "discoveryservice-example",
                  "namespace": "my-namespace"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyDeployment",
                "metadata": {
                  "name": "envoydeployment-sample",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "discoveryServiceRef": "discoveryservice-example",
                  "envoyConfigRef": "envoyconfig-example",
                  "ports": [
                    {
                      "name": "https",
                      "port": 8443
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: quay.io/3scale/marin3r
          createdAt: "2024-08-05T16:08:25Z"
          description: Lighweight, CRD based Envoy control plane for Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["envoyconfigrevisions.marin3r.3scale.net","discoveryservicecertificates.operator.marin3r.3scale.net"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/3scale-ops/marin3r
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DiscoveryServiceCertificate is an internal resource used
              to create certificates. This resource is used by the DiscoveryService
              controller to create the required certificates for the different components.
              Direct use of DiscoveryServiceCertificate objects is discouraged.
            displayName: DiscoveryServiceCertificate
            kind: DiscoveryServiceCertificate
            name: discoveryservicecertificates.operator.marin3r.3scale.net
            version: v1alpha1
          - description: DiscoveryService represents an envoy discovery service server.
              Only one instance per namespace is currently supported.
            displayName: DiscoveryService
            kind: DiscoveryService
            name: discoveryservices.operator.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfigRevision is an internal resource that stores a
              specific version of an EnvoyConfig resource. EnvoyConfigRevisions are
              automatically created and deleted by the EnvoyConfig controller and
              are not intended to be directly used. Use EnvoyConfig objects instead.
            displayName: EnvoyConfigRevision
            kind: EnvoyConfigRevision
            name: envoyconfigrevisions.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfig holds the configuration for a given envoy nodeID.
              The spec of an EnvoyConfig object holds the Envoy resources that conform
              the desired configuration for the given nodeID and that the discovery
              service will send to any envoy client that identifies itself with that
              nodeID.
            displayName: EnvoyConfig
            kind: EnvoyConfig
            name: envoyconfigs.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyDeployment is a resource to deploy and manage a Kubernetes
              Deployment of Envoy Pods.
            displayName: EnvoyDeployment
            kind: EnvoyDeployment
            name: envoydeployments.operator.marin3r.3scale.net
            version: v1alpha1
        description: |
          MARIN3R implements a control plane to deploy, configure and operate a fleet of envoy instances within a Kubernetes cluster.
          It operates a group of gateways and Pod sidecar containers to provide general purpose functionality required to operate production systems.

          Features:

          * Deploy and manage an Envoy xDS server using the DiscoveryService custom resource.
          * Inject Envoy sidecar containers based on Pod annotations.
          * Deploy Envoy as a Kubernetes Deployment using the EnvoyDeployment custom resource.
          * Dynamic Envoy configuration and re-configuration using the EnvoyConfig custom resource.
          * Use Kubernetes Secrets as certificate sources.
          * Syntactic validation of Envoy configurations.
          * Lifecycle management of Envoy containers (graceful shutdown and connection draining).
          * Self-healing.

          Check the [docs](https://github.com/3scale-ops/marin3r#readme) for more information.

          ## License
          MARIN3R is licensed under the [Apache 2.0 license](https://github.com/3scale/prometheus-exporter-operator/blob/master/LICENSE)
        displayName: MARIN3R
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - envoy
        - discovery
        - controlplane
        - proxy
        - sidecar
        - xds
        - network
        - gateway
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/marin3r
        maintainers:
        - email: 3scale-operations+marin3r@redhat.com
          name: 3scale Operations Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/marin3r:v0.13.1
        version: 0.13.1
      entries:
      - name: marin3r.v0.13.1
        version: 0.13.1
      - name: marin3r.v0.12.3
        version: 0.12.3
      - name: marin3r.v0.11.1
        version: 0.11.1
      - name: marin3r.v0.10.0
        version: 0.10.0
      name: alpha
    - currentCSV: marin3r.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyConfig",
                "metadata": {
                  "name": "envoyconfig-example",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "nodeID": "example",
                  "resources": [
                    {
                      "type": "cluster",
                      "value": {
                        "connect_timeout": "0.01s",
                        "dns_lookup_family": "V4_ONLY",
                        "lb_policy": "ROUND_ROBIN",
                        "load_assignment": {
                          "cluster_name": "example",
                          "endpoints": [
                            {
                              "lb_endpoints": [
                                {
                                  "endpoint": {
                                    "address": {
                                      "socket_address": {
                                        "address": "example",
                                        "port_value": 8080
                                      }
                                    }
                                  }
                                }
                              ]
                            }
                          ]
                        },
                        "name": "example",
                        "type": "STRICT_DNS"
                      }
                    },
                    {
                      "type": "route",
                      "value": {
                        "name": "local",
                        "virtual_hosts": [
                          {
                            "domains": [
                              "*"
                            ],
                            "name": "all",
                            "routes": [
                              {
                                "match": {
                                  "prefix": "/"
                                },
                                "route": {
                                  "cluster": "example"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    {
                      "type": "listener",
                      "value": {
                        "address": {
                          "socket_address": {
                            "address": "0.0.0.0",
                            "port_value": 8443
                          }
                        },
                        "filter_chains": [
                          {
                            "filters": [
                              {
                                "name": "envoy.filters.network.http_connection_manager",
                                "typed_config": {
                                  "@type": "type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager",
                                  "access_log": [
                                    {
                                      "name": "envoy.access_loggers.file",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.access_loggers.file.v3.FileAccessLog",
                                        "path": "/dev/stdout"
                                      }
                                    }
                                  ],
                                  "http_filters": [
                                    {
                                      "name": "envoy.filters.http.router",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.filters.http.router.v3.Router"
                                      }
                                    }
                                  ],
                                  "rds": {
                                    "config_source": {
                                      "ads": {},
                                      "resource_api_version": "V3"
                                    },
                                    "route_config_name": "local"
                                  },
                                  "stat_prefix": "https"
                                }
                              }
                            ],
                            "transport_socket": {
                              "name": "envoy.transport_sockets.tls",
                              "typed_config": {
                                "@type": "type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext",
                                "common_tls_context": {
                                  "tls_certificate_sds_secret_configs": [
                                    {
                                      "name": "example.default.svc",
                                      "sds_config": {
                                        "ads": {},
                                        "resource_api_version": "V3"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          }
                        ],
                        "name": "https"
                      }
                    },
                    {
                      "blueprint": "tlsCertificate",
                      "generateFromTlsSecret": "example.default.svc",
                      "type": "secret"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "DiscoveryService",
                "metadata": {
                  "name": "discoveryservice-example",
                  "namespace": "my-namespace"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyDeployment",
                "metadata": {
                  "name": "envoydeployment-sample",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "discoveryServiceRef": "discoveryservice-example",
                  "envoyConfigRef": "envoyconfig-example",
                  "ports": [
                    {
                      "name": "https",
                      "port": 8443
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: quay.io/3scale/marin3r
          createdAt: "2024-08-05T16:08:25Z"
          description: Lighweight, CRD based Envoy control plane for Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["envoyconfigrevisions.marin3r.3scale.net","discoveryservicecertificates.operator.marin3r.3scale.net"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/3scale-ops/marin3r
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DiscoveryServiceCertificate is an internal resource used
              to create certificates. This resource is used by the DiscoveryService
              controller to create the required certificates for the different components.
              Direct use of DiscoveryServiceCertificate objects is discouraged.
            displayName: DiscoveryServiceCertificate
            kind: DiscoveryServiceCertificate
            name: discoveryservicecertificates.operator.marin3r.3scale.net
            version: v1alpha1
          - description: DiscoveryService represents an envoy discovery service server.
              Only one instance per namespace is currently supported.
            displayName: DiscoveryService
            kind: DiscoveryService
            name: discoveryservices.operator.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfigRevision is an internal resource that stores a
              specific version of an EnvoyConfig resource. EnvoyConfigRevisions are
              automatically created and deleted by the EnvoyConfig controller and
              are not intended to be directly used. Use EnvoyConfig objects instead.
            displayName: EnvoyConfigRevision
            kind: EnvoyConfigRevision
            name: envoyconfigrevisions.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfig holds the configuration for a given envoy nodeID.
              The spec of an EnvoyConfig object holds the Envoy resources that conform
              the desired configuration for the given nodeID and that the discovery
              service will send to any envoy client that identifies itself with that
              nodeID.
            displayName: EnvoyConfig
            kind: EnvoyConfig
            name: envoyconfigs.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyDeployment is a resource to deploy and manage a Kubernetes
              Deployment of Envoy Pods.
            displayName: EnvoyDeployment
            kind: EnvoyDeployment
            name: envoydeployments.operator.marin3r.3scale.net
            version: v1alpha1
        description: |
          MARIN3R implements a control plane to deploy, configure and operate a fleet of envoy instances within a Kubernetes cluster.
          It operates a group of gateways and Pod sidecar containers to provide general purpose functionality required to operate production systems.

          Features:

          * Deploy and manage an Envoy xDS server using the DiscoveryService custom resource.
          * Inject Envoy sidecar containers based on Pod annotations.
          * Deploy Envoy as a Kubernetes Deployment using the EnvoyDeployment custom resource.
          * Dynamic Envoy configuration and re-configuration using the EnvoyConfig custom resource.
          * Use Kubernetes Secrets as certificate sources.
          * Syntactic validation of Envoy configurations.
          * Lifecycle management of Envoy containers (graceful shutdown and connection draining).
          * Self-healing.

          Check the [docs](https://github.com/3scale-ops/marin3r#readme) for more information.

          ## License
          MARIN3R is licensed under the [Apache 2.0 license](https://github.com/3scale/prometheus-exporter-operator/blob/master/LICENSE)
        displayName: MARIN3R
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - envoy
        - discovery
        - controlplane
        - proxy
        - sidecar
        - xds
        - network
        - gateway
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/marin3r
        maintainers:
        - email: 3scale-operations+marin3r@redhat.com
          name: 3scale Operations Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/marin3r:v0.13.1
        version: 0.13.1
      entries:
      - name: marin3r.v0.13.1
        version: 0.13.1
      - name: marin3r.v0.12.3
        version: 0.12.3
      - name: marin3r.v0.11.1
        version: 0.11.1
      - name: marin3r.v0.10.0
        version: 0.10.0
      name: stable
    defaultChannel: stable
    packageName: marin3r
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: GitLab, Inc.
      provider-url: https://gitlab.com
    name: gitlab-runner-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-runner-operator.v1.41.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta2",
                "kind": "Runner",
                "metadata": {
                  "name": "gitlab-runner-example",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "gitlabUrl": "https://gitlab.example.com",
                  "token": "runner-token-secret",
                  "tags": "kubernetes,example",
                  "concurrent": 10,
                  "imagePullPolicy": "IfNotPresent"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery, Developer Tools
          containerImage: registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
          createdAt: "2025-09-25T15:08:40Z"
          description: GitLab Runner operator manages lifecycle of GitLab Runner instances
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.11.0 <1.41.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: none
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
          support: GitLab, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Runner is the open source project used to run your jobs and
              send the results back to GitLab
            displayName: GitLab Runner
            kind: Runner
            name: runners.apps.gitlab.com
            version: v1beta2
        description: |-
          GitLab Runner is the lightweight, highly-scalable agent that runs your build jobs and sends the results back to a GitLab instance. GitLab Runner works in conjunction with GitLab CI/CD, the open-source continuous integration service included with GitLab.

              The GitLab Runner operator manages the lifecycle of GitLab Runner in Kubernetes or Openshift clusters. The operator aims to automate the tasks needed to run your CI/CD jobs in your container orchestration platform.

              ## Prerequisites

                For Kubernetes cluster, install cert-manager:

                ```shell
                kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.7.1/cert-manager.yaml
                ```

              ## GitLab Runner version

              This version of **GitLab Runner Operator** ships with **GitLab Runner v18.4.0**.

              To use a different version of **GitLab Runner** change the [`runnerImage` and `helperImage` properties](https://docs.gitlab.com/runner/configuration/configuring_runner_operator.html#operator-properties).

              ## Usage

               To link a GitLab Runner instance to a self-hosted GitLab instance or the hosted [GitLab](https://gitlab.com), you first need to:

               1. Create a secret containing the `runner-token` from your GitLab project.

                ```
                cat > gitlab-runner-secret.yml << EOF
                apiVersion: v1
                kind: Secret
                metadata:
                  name: gitlab-runner-secret
                type: Opaque
                # Only one of the following fields must be set. The Operator will fail to register Runner if both are provided.
                # # NOTE: runner-registration-token is deprecated and will be removed in GitLab 18.0. You should use runner-token instead.
                stringData:
                  runner-token: REPLACE_ME # your project runner token
                  # runner-registration-token: "" # your project runner secret
                EOF
                ```

                ```
                oc apply -f gitlab-runner-secret.yml
                ```

               2. Create the Custom Resource Definition (CRD) file and include the following information. The tags value must be openshift for the job to run.

                 ```
                 cat > gitlab-runner.yml << EOF
                 apiVersion: apps.gitlab.com/v1beta2
                 kind: Runner
                 metadata:
                   name: gitlab-runner
                 spec:
                   gitlabUrl: https://gitlab.example.com
                   buildImage: alpine
                   token: gitlab-runner-secret
                   tags: openshift
                 EOF
                 ```

                ```
                oc apply -f gitlab-runner.yml
                ```

              ## Full documentation

              Visit [Install GitLab Runner Operator](https://docs.gitlab.com/runner/install/operator.html)
        displayName: GitLab Runner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GitLab
        - CI/CD
        - DevOps
        - SAST
        - DAST
        links:
        - name: Gitlab Runner Operator
          url: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
        - name: GitLab Docs
          url: https://docs.gitlab.com
        - name: GitLab Runner
          url: https://docs.gitlab.com/runner/
        - name: GitLab CI/CD
          url: https://docs.gitlab.com/ee/ci/quick_start/
        maintainers:
        - email: support@gitlab.com
          name: Georgi N. Georgiev
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: GitLab, Inc.
          url: https://gitlab.com
        relatedImages:
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-ocp@sha256:0f2a5a3b6779d719eecbf079b99cf06b652603d74b563f57d753e32966d253dd
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-helper-ocp@sha256:29727fae83a226bcf57607240d4046a0b8c208b5ff42ee1f8ca0f115bb188809
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
        version: 1.41.1
      entries:
      - name: gitlab-runner-operator.v1.41.1
        version: 1.41.1
      - name: gitlab-runner-operator.v1.41.0
        version: 1.41.0
      - name: gitlab-runner-operator.v1.40.1
        version: 1.40.1
      - name: gitlab-runner-operator.v1.40.0
        version: 1.40.0
      - name: gitlab-runner-operator.v1.39.0
        version: 1.39.0
      name: stable
    defaultChannel: stable
    packageName: gitlab-runner-operator
    provider:
      name: GitLab, Inc.
      url: https://gitlab.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Anjuna Security
      provider-url: https://www.anjuna.io
    name: anjuna-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anjuna-operator.v2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anjuna.io/v1",
                "kind": "AnjunaConfig",
                "metadata": {
                  "name": "anjuna-config-remote"
                },
                "spec": {
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "runtimeImage": "registry.connect.redhat.com/anjuna/anjuna-k8s-sev-runtime@sha256:1531b1ba955b2d3ea5d9238d2e20617a534bb1bab73f5fadadf8c9d3aee410b4"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-02-19T01:11:18Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          operatorframework.io/suggested-namespace: anjuna-operator
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnjunaConfig operand instructs the Anjuna Seaglass Operator
              to know (1) what needs to be installed into the cluster and (2) what
              nodes will be affected.
            displayName: Anjuna Config
            kind: AnjunaConfig
            name: anjunaconfigs.anjuna.io
            version: v1
        description: "The Anjuna Seaglass Operator for OpenShift allows running any
          Pod in\nOpenShift in a Confidential Pod leveraging the full power of the
          Confidential\nComputing technology.\n\n# Anjuna Seaglass\nAnjuna Seaglass™
          is the Universal Confidential Computing Platform to run applications in
          any cloud, and on-premises, with complete data security and privacy. \nIt
          isolates workloads in a hardware-backed environment that creates a zero-trust
          model for code and data.\n\nThe Anjuna Seaglass platform allows users to
          leverage the power of Confidential Computing:\n\nFor any Confidential Computing
          technology\nIn any Cloud Service Provider, as well as on-premises\nWith
          no changes to the application\n\nRunning a Confidential Pod with Anjuna
          Seaglass protects the Pod's memory with hardware-enforced encryption and
          integrity protection, and enables you to authenticate and trust the workload
          using Remote Attestation.\n\n# Data-in-use protection\nWhen protecting sensitive
          data managed by a Pod in OpenShift, one can protect the data at rest by
          encrypting the data written to the persistent storage, and protecting the
          data in transit when sending data over the network.\n\nHowever, when the
          data is used in the Pod's memory, it cannot be both encrypted and used at
          the same time.\nThis means that if a malicious insider (either human or
          software) had access to the OpenShift Node, they could dump the Pod's memory
          and access all of its sensitive data. This includes accessing the secrets
          used to encrypt the data at rest and in transit.\n\nRunning the Pod in a
          Confidential Pod with the Anjuna Seaglass platform solves this problem.
          Anjuna Seaglass runs the Pod inside a Trusted Execution Environment, where
          the Pod memory is encrypted by the hardware so that no privileged human
          or software on the Pod can access this memory in the clear.\n\n# Enabling
          zero trust for the Pod\nWhen running a Pod in OpenShift, the Pod may need
          to create trust with other Pods or with software running outside of the
          OpenShift cluster.\nThe standard way to create this trust is to identify
          with a token or credentials that could be easily read and stolen from the
          Pod disk or by providing a machine identity, a process name, or an IP that
          could be spoofed or reused.\n\nConfidential Computing technology provides
          a mechanism called Remote Attestation, which provides an identity to the
          software via the Hardware—the Software Measurement.\n\nWhen running a Pod
          as a Confidential Pod with the Anjuna Seaglass platform, a user can leverage
          the Remote Attestation to perform trusted authentication with the Confidential
          Pod, in a way that cannot be spoofed.\nThe Anjuna Policy Manager allows
          you to protect the first secret of a Confidential Pod and remove any secret
          from the Pod image or Kubernetes control plane that could be read by malicious
          insiders.\n\n# Quick start and configuration instructions\nFollow the directions
          in Installing the [Anjuna Kubernetes Toolset to OpenShift on GCP](https://docs.anjuna.io/k8s-sev/latest/installation/openshift_gcp.html)
          \nto learn how to deploy the Operator into your OpenShift cluster.\n\nFollow
          the directions in [Quickstart guides for OpenShift on GCP](https://docs.anjuna.io/k8s-sev/latest/applications/gcp.html)
          \nto learn how to deploy your first Confidential Pod in your OpenShift cluster.\n\nFollow
          the directions in [Attestation with Anjuna Policy Manager](https://docs.anjuna.io/k8s-sev/latest/attestation.html)
          \nto learn how to leverage Remote Attestation to protect your Pod's first
          secret."
        displayName: Anjuna Seaglass™ Operator for OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Security
        links:
        - name: Anjuna Operator Documentation
          url: https://docs.anjuna.io/k8s-sev/latest/index.html
        - name: Anjuna Seaglass
          url: https://www.anjuna.io/essence-of-anjuna-seaglass
        - name: Anjuna Security
          url: https://www.anjuna.io
        maintainers:
        - email: support@anjuna.io
          name: Anjuna Security
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Anjuna Security
          url: https://www.anjuna.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - registry.connect.redhat.com/anjuna/anjuna-k8s-sev-tools@sha256:e5ba31c240aaff8ca8f6addf12577fd6045bf6619a1d2cb2987b2ff5d75f4028
        - registry.connect.redhat.com/anjuna/anjuna-k8s-sev-runtime@sha256:1531b1ba955b2d3ea5d9238d2e20617a534bb1bab73f5fadadf8c9d3aee410b4
        version: 2.0.0
      entries:
      - name: anjuna-operator.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anjuna-operator
    provider:
      name: Anjuna Security
      url: https://www.anjuna.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzograph-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzograph-operator.v3.2.500
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzograph.clusters.cambridgesemantics.com/v2",
                "kind": "AnzoGraph",
                "metadata": {
                  "name": "azg01"
                },
                "spec": {
                  "db": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_data": "anzograph-data-grpc",
                            "app_mgmt": "anzograph-mgmt-grpc"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_data": "anzograph-data-grpc",
                              "app_mgmt": "anzograph-mgmt-grpc"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzograph-db@sha256:30ed2ec337135701ab8b54e75bb48ad8928a2c48818643c23b9dba4484255a2e",
                                "name": "db",
                                "resources": {
                                  "limits": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  },
                                  "requests": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  }
                                }
                              }
                            ],
                            "securityContext": {
                              "fsGroup": 1000,
                              "runAsNonRoot": true,
                              "runAsUser": 1000,
                              "seccompProfile": {
                                "type": "RuntimeDefault"
                              }
                            },
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  },
                  "deployFrontend": false,
                  "frontend": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_mgmt": "anzograph-frontend"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_mgmt": "anzograph-frontend"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzograph-frontend@sha256:e5446b3c04889f2189e955702c4165eb243eb89f47ac797eb417b60b1459647f",
                                "name": "frontend",
                                "resources": {
                                  "limits": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "securityContext": {
                              "fsGroup": 1000,
                              "runAsNonRoot": true,
                              "runAsUser": 1000,
                              "seccompProfile": {
                                "type": "RuntimeDefault"
                              }
                            },
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/anzograph-operator@sha256:2333154a14ab7f3a03a601f00c017feeafb9d9814aff5e59c95cef615ab2abd6
          createdAt: "2025-09-22T12:03:18Z"
          description: kubernetes operator for Altair Graph Lakehouse DB
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Altair Engineering Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoGraph is the Schema for the anzographs API
            displayName: Altair Graph Lakehouse Operator
            kind: AnzoGraph
            name: anzographs.anzograph.clusters.cambridgesemantics.com
            version: v2
        description: |-
          The Altair Graph Lakehouse Operator provides the way to install and configure an Altair Graph Lakehouse
          cluster on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v3.2.5/README_openshift_marketplace.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzograph/userdoc/ )

          ### Support

          We offer Support to our customers with the Altair Graph Lakehouse db Enterprise Edition License [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Lakehouse Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        - Database
        links:
        - name: Anzograph Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v3.2.5/README_openshift_marketplace.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzograph-operator@sha256:2333154a14ab7f3a03a601f00c017feeafb9d9814aff5e59c95cef615ab2abd6
        - registry.connect.redhat.com/cambridgesemantics/anzograph-db@sha256:30ed2ec337135701ab8b54e75bb48ad8928a2c48818643c23b9dba4484255a2e
        - registry.connect.redhat.com/cambridgesemantics/anzograph-frontend@sha256:e5446b3c04889f2189e955702c4165eb243eb89f47ac797eb417b60b1459647f
        version: 3.2.500
      entries:
      - name: anzograph-operator.v3.2.500
        version: 3.2.500
      - name: anzograph-operator.v3.2.300
        version: 3.2.300
      - name: anzograph-operator.v3.2.200
        version: 3.2.200
      - name: anzograph-operator.v3.2.104
        version: 3.2.104
      name: stable
    defaultChannel: stable
    packageName: anzograph-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-memorydb-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-memorydb-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "ACL",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "ParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "Snapshot",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "SubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/memorydb-controller:1.1.1
          createdAt: "2025-09-19T19:36:49Z"
          description: Amazon MemoryDB for Redis controller is a service controller
            for managing MemoryDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ACL represents the state of an AWS memorydb ACL resource.
            displayName: ACL
            kind: ACL
            name: acls.memorydb.services.k8s.aws
            version: v1alpha1
          - description: Cluster represents the state of an AWS memorydb Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.memorydb.services.k8s.aws
            version: v1alpha1
          - description: ParameterGroup represents the state of an AWS memorydb ParameterGroup
              resource.
            displayName: ParameterGroup
            kind: ParameterGroup
            name: parametergroups.memorydb.services.k8s.aws
            version: v1alpha1
          - description: Snapshot represents the state of an AWS memorydb Snapshot
              resource.
            displayName: Snapshot
            kind: Snapshot
            name: snapshots.memorydb.services.k8s.aws
            version: v1alpha1
          - description: SubnetGroup represents the state of an AWS memorydb SubnetGroup
              resource.
            displayName: SubnetGroup
            kind: SubnetGroup
            name: subnetgroups.memorydb.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS memorydb User resource.
            displayName: User
            kind: User
            name: users.memorydb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon MemoryDB for Redis resources in AWS from within your Kubernetes cluster.

          **About Amazon MemoryDB for Redis**

          MemoryDB for Redis is a durable, in-memory database service that delivers ultra-fast performance. It is purpose-built for modern applications with microservices architectures.
          MemoryDB is compatible with Redis, a popular open source data store, enabling you to quickly build applications using the same flexible and friendly Redis data structures, APIs, and commands that they already use today. With MemoryDB, all of your data is stored in memory, which enables you to achieve microsecond read and single-digit millisecond write latency and high throughput. MemoryDB also stores data durably across multiple Availability Zones (AZs) using a Multi-AZ transactional log to enable fast failover, database recovery, and node restarts.
          Delivering both in-memory performance and Multi-AZ durability, MemoryDB can be used as a high-performance primary database for your microservices applications, eliminating the need to separately manage both a cache and durable database.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - MemoryDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - memorydb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon MemoryDB for Redis Developer Resources
          url: https://docs.aws.amazon.com/memorydb/?id=docs_gateway
        maintainers:
        - email: ack-maintainers@amazon.com
          name: memorydb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/memorydb-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-memorydb-controller.v1.1.1
        version: 1.1.1
      - name: ack-memorydb-controller.v1.1.0
        version: 1.1.0
      - name: ack-memorydb-controller.v1.0.15
        version: 1.0.15
      - name: ack-memorydb-controller.v1.0.13
        version: 1.0.13
      - name: ack-memorydb-controller.v1.0.12
        version: 1.0.12
      - name: ack-memorydb-controller.v1.0.11
        version: 1.0.11
      - name: ack-memorydb-controller.v1.0.10
        version: 1.0.10
      - name: ack-memorydb-controller.v1.0.9
        version: 1.0.9
      - name: ack-memorydb-controller.v1.0.8
        version: 1.0.8
      - name: ack-memorydb-controller.v1.0.7
        version: 1.0.7
      - name: ack-memorydb-controller.v1.0.6
        version: 1.0.6
      - name: ack-memorydb-controller.v1.0.5
        version: 1.0.5
      - name: ack-memorydb-controller.v1.0.4
        version: 1.0.4
      - name: ack-memorydb-controller.v1.0.3
        version: 1.0.3
      - name: ack-memorydb-controller.v1.0.2
        version: 1.0.2
      - name: ack-memorydb-controller.v1.0.1
        version: 1.0.1
      - name: ack-memorydb-controller.v1.0.0
        version: 1.0.0
      - name: ack-memorydb-controller.v0.0.12
        version: 0.0.12
      - name: ack-memorydb-controller.v0.0.11
        version: 0.0.11
      - name: ack-memorydb-controller.v0.0.10
        version: 0.0.10
      - name: ack-memorydb-controller.v0.0.9
        version: 0.0.9
      - name: ack-memorydb-controller.v0.0.8
        version: 0.0.8
      - name: ack-memorydb-controller.v0.0.7
        version: 0.0.7
      - name: ack-memorydb-controller.v0.0.6
        version: 0.0.6
      - name: ack-memorydb-controller.v0.0.5
        version: 0.0.5
      - name: ack-memorydb-controller.v0.0.4
        version: 0.0.4
      - name: ack-memorydb-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-memorydb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibmcloud-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibmcloud-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"ibmcloud.ibm.com/v1","kind":"Service","metadata":{"annotations":{"ibmcloud.ibm.com/self-healing":"enabled"},"name":"mytranslator"},"spec":{"plan":"lite","serviceClass":"language-translator"}},{"apiVersion":"ibmcloud.ibm.com/v1","kind":"Binding","metadata":{"name":"binding-translator"},"spec":{"parameters":[{"name":"serviceid_crn","value":"crn:v1:bluemix:public:iam-identity::a/ACCOUNT_ID::serviceid:SERVICE_ID"}],"secretName":"translator-secret","serviceName":"mytranslator"}}]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: cloudoperators/ibmcloud-operator:1.1.0
          createdAt: "2022-03-17T04:17:40Z"
          description: The IBM Cloud Operator provides a Kubernetes CRD-Based API
            to manage the lifecycle of IBM public cloud services.
          repository: https://github.com/IBM/cloud-operators
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Binding is an instance of a service binding resource on IBM
              Cloud. A Binding creates a secret with the service instance credentials.
            displayName: Binding
            kind: Binding
            name: bindings.ibmcloud.ibm.com
            version: v1
          - description: Service is an instance of a Service resource on IBM Cloud.
            displayName: Service
            kind: Service
            name: services.ibmcloud.ibm.com
            version: v1
        description: |-
          With the IBM Cloud Operator, you can provision and bind [IBM public cloud services](https://cloud.ibm.com/catalog#services) to your Kubernetes cluster in a Kubernetes-native way. The IBM Cloud Operator is based on the [Kubernetes custom resource definition (CRD) API](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/) so that your applications can create, update, and delete IBM Cloud services from within the cluster by calling Kubnernetes APIs, instead of needing to use several IBM Cloud APIs in addition to configuring your app for Kubernetes.


          ## Features

          * **Service provisioning**: Create any service with any plan that is available in the [IBM Cloud catalog](https://cloud.ibm.com/catalog#services).

          * **Bindings management**: Automatically create Kubernetes secrets in your Kubernetes cluster with the credentials to bind to any provisioned service to the cluster.


          ## Upgrading the operator


          To upgrade an OperatorHub installation, use the OpenShift web console to perform the upgrade. If upgrading from the `alpha` channel subscription to `stable`, follow the additional upgrade instructions below.

          ### Upgrading to version v0.3 or v1.0
          **IMPORTANT NOTICE:** v0.1 and v0.2 used a different naming scheme for secrets and configmaps. Before you update the IBM Cloud Operator, create secret and configmap resources with these names and copy the contents of your previous resources to the new resources. Then, the upgraded operator recognizes and continues to update the resources.

          Existing `Service` and `Binding` resources **do not** need modification for the upgrade.

          | Previous names (v0.1 and v0.2)         | **Current names (v0.3 and v1.0)**             | Description                                                                           |
          |:---------------------------------------|:--------------------------------------------|:--------------------------------------------------------------------------------------|
          | secret-ibm-cloud-operator              | **ibmcloud-operator-secret**                | Secret with the API key, scoped to the namespace.                                     |
          | config-ibm-cloud-operator              | **ibmcloud-operator-defaults**              | ConfigMap with the default values for new resources.                                  |
          | ibm-cloud-operator                     | **ibmcloud-operator-config**                | ConfigMap with the management namespace configuration.                                |
          | ${namespace}-secret-ibm-cloud-operator | **${namespace}-ibmcloud-operator-secret**   | Management namespace Secret with the API key for ${namespace}.                        |
          | ${namespace}-config-ibm-cloud-operator | **${namespace}-ibmcloud-operator-defaults** | Management namespace ConfigMap with default values for new resources in ${namespace}. |

          **Tip:** Forgot to update the secret and configmap names before upgrading? The operator will not take action on Service and Binding resources until the new secrets and configmaps have been created. Creating these after the upgrade will also work.


          ### OperatorHub stable channel
          Continuous updates and bug fixes are provided in the latest stable channel. Subscribing to the `stable` channel in OperatorHub means you automatically get the latest updates without breaking backward compatibility.

          ### Changelog

          For a list of recent changes, see the [releases page](https://github.com/IBM/cloud-operators/releases).


          ## Prerequisites

          1.  Have an [IBM Cloud account](https://cloud.ibm.com/registration).
          2.  Have a cluster that runs Kubernetes version 1.11 or later (OpenShift 3.11 or later).
          3.  Install the required command line tools.
              *   [IBM Cloud CLI](https://cloud.ibm.com/docs/cli?topic=cloud-cli-getting-started) (`ibmcloud`)
              *   [Kubernetes CLI](https://kubernetes.io/docs/tasks/tools/install-kubectl/) (`kubectl`)
          4.  Log in to your IBM Cloud account from the CLI.

              ```bash
              ibmcloud login
              ```

          5.  Target the appropriate resource group (`-g`) and default region (`-r`) for provisioning services:

              ```bash
              ibmcloud target -g default -r us-south
              ```

          6.  If you also want to provision Cloud Foundry services, target the appropriate `org` and `space` (`--cf`).

              ```bash
              ibmcloud target --cf
              ```

          7.  Set the Kubernetes context of your CLI to your cluster so that you can run `kubectl` commands. For example, if your cluster runs OpenShift, use the `oc login` command.


          ## Setting up the operator

          You can use an installation script to set up the IBM Cloud Operator.
          By default, the installation script stores an API key in a Kubernetes secret in your cluster that can be accessed by the IBM Cloud Operator.
          Next, the script sets default values that are used to provision IBM Cloud services, like the resource group and region to provision the services in. Later, you can override any default value in the `Service` custom resource. Finally, the script deploys the operator in your cluster.

          To use your own API key, set the `IBMCLOUD_API_KEY` environment variable to the key before running the installation script:

              export IBMCLOUD_API_KEY="CUSTOM_API_KEY"

          If installed with Operator Hub, the operator will run in the `openshift-operators` namespace. Otherwise, it will run in the `ibmcloud-operator-system` namespace.

          Prefer to create the secrets and defaults yourself? See the [IBM Cloud Operator installation guide](https://github.com/IBM/cloud-operators/blob/master/docs/install.md).

          ### Using a service ID

          By default, the installation script creates an IBM Cloud API key that impersonates your user credentials, to use to set up the IBM Cloud Operator. However, you might want to create a service ID in IBM Cloud Identity and Access Managment (IAM). By using a service ID, you can control access for the IBM Cloud Operator without having the permissions tied to a particular user, such as if that user leaves the company. For more information, see the [IBM Cloud docs](https://cloud.ibm.com/docs/account?topic=account-serviceids).

          1.  Create a service ID in IBM Cloud IAM. If possible, do not use spaces in the names for your IAM credentials. When you use the operator binding feature, any spaces are replaced with underscores.

              ```bash
              ibmcloud iam service-id-create serviceid-ico -d service-ID-for-ibm-cloud-operator
              ```

          2.  Assign the service ID access to the required permissions to work with the IBM Cloud services that you want the operator to manage. The required permissions vary with each IBM Cloud service. You can also scope an access policy to different regions or resource groups. For example, the following command grants the service ID the **Administrator** platform role in the default resource group in the US South (Dallas) region. For more information, see the [IBM Cloud docs](https://cloud.ibm.com/docs/account?topic=account-userroles).

              ```bash
              ibmcloud iam service-policy-create serviceid-ico --roles Administrator --resource-group-name default --region us-south
              ```

          3.  Create an API key for the service ID.

              ```bash
              ibmcloud iam service-api-key-create apikey-ico serviceid-ico -d api-key-for-ibm-cloud-operator
              ```

          4.  Set the API key of the service ID as your CLI environment variable. Now, when you run the installation script, the script uses the service ID's API key. The following command is an example for macOS.

              ```bash
              export IBMCLOUD_API_KEY=<apikey-ico-value>
              ```

          5.  Confirm that the API key environment variable is set in your CLI.

              ```bash
              echo $IBMCLOUD_API_KEY
              ```

          6.  Follow the [prerequisite steps](https://github.com/IBM/cloud-operators/blob/master/README.md#prerequisites) to log in to the IBM Cloud account that owns the service ID.


          ### Installing the operator for OpenShift clusters

          Before you begin, complete the [prerequisite steps](https://github.com/IBM/cloud-operators/blob/master/README.md#prerequisites) to log in to IBM Cloud and your cluster, and optionally set up a [service ID API key](https://github.com/IBM/cloud-operators/blob/master/README.md#using-a-service-id).

          To configure the latest release for OpenShift before installing via the OperatorHub, run the following script:

          *   **Latest release**:

              ```bash
              curl -sL https://raw.githubusercontent.com/IBM/cloud-operators/master/hack/configure-operator.sh | bash
              ```

          *   **Specific release**: Replace `-v 0.0.0` with the specific version that you want to install.

              ```bash
              curl -sL https://raw.githubusercontent.com/IBM/cloud-operators/master/hack/configure-operator.sh | bash -s -- -v 0.0.0 store-creds
              ```


          ## Using the IBM Cloud Operator

          To use the IBM Cloud Operator, create a service instance and then bind the service to your cluster. For more information, see the [sample configuration files](https://github.com/IBM/cloud-operators/blob/master/config/samples), [user guide](https://github.com/IBM/cloud-operators/blob/master/docs/user-guide.md), and [reference documentation](https://github.com/IBM/cloud-operators/blob/master/README.md#reference).

          #### Step 1: Creating a service instance

          1.  To create an instance of an IBM public cloud service, first create a `Service` custom resource file. For more options, see the [Service properties](https://github.com/IBM/cloud-operators/blob/master/README.md#service-properties) reference doc.
              *   `<SERVICE_CLASS>` is the IBM Cloud service that you want to create. To list IBM Cloud services, run `ibmcloud catalog service-marketplace` and use the **Name** value from the output.
              *   `<PLAN>` is the plan for the IBM Cloud service that you want to create, such as `free` or `standard`. To list available plans, run `ibmcloud catalog service <SERVICE_CLASS> | grep plan`.

              ```yaml
              apiVersion: ibmcloud.ibm.com/v1
              kind: Service
              metadata:
                  name: myservice
              spec:
                  plan: <PLAN>
                  serviceClass: <SERVICE_CLASS>
              ```

          2.  Create the service instance in your cluster.

              ```bash
              kubectl apply -f filepath/myservice.yaml
              ```

          3.  Check that your service status is **Online** in your cluster.

              ```bash
              kubectl get services.ibmcloud
              NAME           STATUS   AGE
              myservice      Online   12s
              ```

          4.  Verify that your service instance is created in IBM Cloud.

              ```bash
              ibmcloud resource service-instances | grep myservice
              ```

          #### Step 2: Binding the service instance

          1.  To bind your service to the cluster so that your apps can use the service, create a `Binding` custom resource, where the `serviceName` field is the name of the `Service` custom resource that you previously created. For more options, see [Binding properties](https://github.com/IBM/cloud-operators/blob/master/README.md#binding-properties).

              ```yaml
              apiVersion: ibmcloud.ibm.com/v1
              kind: Binding
              metadata:
                  name: mybinding
              spec:
                  serviceName: myservice
              ```

          2.  Create the binding in your cluster.

              ```bash
              kubectl apply -f filepath/mybinding.yaml
              ```

          3.  Check that your service status is **Online**.

              ```bash
              kubectl get bindings.ibmcloud
              NAME         STATUS   AGE
              mybinding    Online   25s
              ```

          4.  Check that a secret of the same name as your binding is created. The secret contains the service credentials that apps in your cluster can use to access the service.

              ```bash
              kubectl get secrets
              NAME         TYPE      DATA   AGE
              mybinding    Opaque    6      102s
              ```
        displayName: IBM Cloud Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: Source on GitHub
          url: https://github.com/ibm/cloud-operators
        maintainers:
        - email: johnstarich@johnstarich.com
          name: John Starich
        - email: mvaziri@us.ibm.com
          name: vazirim
        - email: greg.lanthier@ca.ibm.com
          name: Greg Lanthier
        - email: halacs@users.noreply.github.com
          name: Halacs
        - email: artberger@users.noreply.github.com
          name: Art
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: IBM
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - cloudoperators/ibmcloud-operator:1.1.0
        version: 1.1.0
      entries:
      - name: ibmcloud-operator.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: ibmcloud-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Infinispan
      provider-url: ""
    name: infinispan
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infinispan-operator.v2.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.2.5.Final
          description: Create and manage Infinispan clusters.
          olm.skipRange: '>=2.1.x <2.2.1'
          operators.operatorframework.io/builder: operator-sdk-v1.3.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/infinispan/operator:2.2.5.Final
        version: 2.2.5
      entries:
      - name: infinispan-operator.v2.2.5
        version: 2.2.5
      - name: infinispan-operator.v2.2.4
        version: 2.2.4
      - name: infinispan-operator.v2.2.3
        version: 2.2.3
      - name: infinispan-operator.v2.2.2
        version: 2.2.2
      - name: infinispan-operator.v2.2.1
        version: 2.2.1
      - name: infinispan-operator.v2.2.0
        version: 2.2.0
      name: 2.2.x
    - currentCSV: infinispan-operator.v2.3.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.3.7.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/infinispan/operator:2.3.7.Final
        version: 2.3.8
      entries:
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: 2.3.x
    - currentCSV: infinispan-operator.v2.4.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.4.17.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/infinispan/operator:2.4.17.Final
        version: 2.4.17
      entries:
      - name: infinispan-operator.v2.4.17
        version: 2.4.17
      - name: infinispan-operator.v2.4.16
        version: 2.4.16
      - name: infinispan-operator.v2.4.15
        version: 2.4.15
      - name: infinispan-operator.v2.4.14
        version: 2.4.14
      - name: infinispan-operator.v2.4.13
        version: 2.4.13
      - name: infinispan-operator.v2.4.12
        version: 2.4.12
      - name: infinispan-operator.v2.4.11
        version: 2.4.11
      - name: infinispan-operator.v2.4.10
        version: 2.4.10
      - name: infinispan-operator.v2.4.9
        version: 2.4.9
      - name: infinispan-operator.v2.4.8
        version: 2.4.8
      - name: infinispan-operator.v2.4.7
        version: 2.4.7
      - name: infinispan-operator.v2.4.6
        version: 2.4.6
      - name: infinispan-operator.v2.4.5
        version: 2.4.5
      - name: infinispan-operator.v2.4.4
        version: 2.4.4
      - name: infinispan-operator.v2.4.3
        version: 2.4.3
      - name: infinispan-operator.v2.4.2
        version: 2.4.2
      - name: infinispan-operator.v2.4.1
        version: 2.4.1
      - name: infinispan-operator.v2.4.0
        version: 2.4.0
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: 2.4.x
    - currentCSV: infinispan-operator.v2.4.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.4.18.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/infinispan/operator:2.4.18.Final
        version: 2.4.18
      entries:
      - name: infinispan-operator.v2.4.18
        version: 2.4.18
      - name: infinispan-operator.v2.4.16
        version: 2.4.16
      - name: infinispan-operator.v2.4.15
        version: 2.4.15
      - name: infinispan-operator.v2.4.14
        version: 2.4.14
      - name: infinispan-operator.v2.4.13
        version: 2.4.13
      - name: infinispan-operator.v2.4.12
        version: 2.4.12
      - name: infinispan-operator.v2.4.11
        version: 2.4.11
      - name: infinispan-operator.v2.4.10
        version: 2.4.10
      - name: infinispan-operator.v2.4.9
        version: 2.4.9
      - name: infinispan-operator.v2.4.8
        version: 2.4.8
      - name: infinispan-operator.v2.4.7
        version: 2.4.7
      - name: infinispan-operator.v2.4.6
        version: 2.4.6
      - name: infinispan-operator.v2.4.5
        version: 2.4.5
      - name: infinispan-operator.v2.4.4
        version: 2.4.4
      - name: infinispan-operator.v2.4.3
        version: 2.4.3
      - name: infinispan-operator.v2.4.2
        version: 2.4.2
      - name: infinispan-operator.v2.4.1
        version: 2.4.1
      - name: infinispan-operator.v2.4.0
        version: 2.4.0
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: stable
    defaultChannel: stable
    packageName: infinispan
    provider:
      name: Infinispan
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent, part of Cisco
      provider-url: https://isovalent.com/
    name: clife
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clife.v1.17.8-cee.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cilium.io/v1alpha1",
                "kind": "CiliumConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "clife"
                  },
                  "name": "ciliumconfig"
                },
                "spec": {
                  "cluster": {
                    "name": "default"
                  },
                  "clusterHealthPort": 9940,
                  "cni": {
                    "binPath": "/var/lib/cni/bin",
                    "chainingMode": "portmap",
                    "confPath": "/var/run/multus/cni/net.d",
                    "exclusive": false
                  },
                  "enterprise": {
                    "featureGate": {
                      "approved": [
                        "CNIChainingMode"
                      ]
                    }
                  },
                  "hubble": {
                    "enabled": true
                  },
                  "ipam": {
                    "mode": "cluster-pool"
                  },
                  "kubeProxyReplacement": false,
                  "operator": {
                    "prometheus": {
                      "enabled": true,
                      "serviceMonitor": {
                        "enabled": true
                      }
                    }
                  },
                  "prometheus": {
                    "enabled": true,
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "securityContext": {
                    "privileged": true
                  },
                  "sessionAffinity": true,
                  "tunnelPort": 4789
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Networking,Security
          createdAt: "2025-09-23T07:55:19Z"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CiliumConfig defines the configuration of Isovalent Networking
              for Kubernetes and all its components
            displayName: CiliumConfig
            kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Networking for Kubernetes, built on Cilium, provides
          eBPF-based networking, observability, and security with optimal scale and
          performance for platform teams operating Kubernetes environments across
          cloud and on-prem infrastructure.
        displayName: Isovalent Networking for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cilium
        - networking
        - observability
        - security
        - eBPF
        - CNI
        links:
        - name: Isovalent Networking for Kubernetes
          url: https://isovalent.com/product/
        maintainers:
        - email: support@isovalent.com
          name: Isovalent, part of Cisco
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Isovalent, part of Cisco
          url: https://isovalent.com/
        relatedImages:
        - quay.io/isovalent/startup-script-ubi:v1.17.8-cee.1@sha256:9a6a413262ec59026345aefc743de847891d80e31ef2f41db954acdcee11c2ff
        - quay.io/isovalent/certgen-ubi:v1.17.8-cee.1@sha256:fb63bed641a279cc09114e93821c7b7555c61393ab594ea2b823ce34fa73335e
        - quay.io/isovalent/cilium-envoy-ubi:v1.17.8-cee.1@sha256:c3de0334f90f9b2540cdec7e2356d813f2df93f0d7d8701005eab7161c7663b5
        - quay.io/isovalent/operator-generic-ubi:v1.17.8-cee.1@sha256:37e4b9f65b2c67d24afeeebebe5d0cb59c09280179de57cd77da2b68e19c3c1b
        - quay.io/isovalent/clife@sha256:ba9bd4a5c35e1d26e395afbc2a4de89f42df1e5e61c9b23959c23194368b7b0e
        - quay.io/isovalent/cilium-ubi:v1.17.8-cee.1@sha256:e9816096e481fad5d95209f902e4a7018aeb65d1e3be1574e6b289e21229f87a
        - quay.io/isovalent/hubble-relay-ubi:v1.17.8-cee.1@sha256:88964d4836d59420dc7e0abb758654ceaf2abcfb53e19bbe2c7962aa68874e37
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.17.8-cee.1@sha256:b097c823638b4fdbe5bdd2b6040a17fef0c70a020e67a18f8231d60eeac02dd5
        version: 1.17.8-cee.1
      entries:
      - name: clife.v1.17.8-cee.1
        version: 1.17.8-cee.1
      - name: clife.v1.17.7-cee.1
        version: 1.17.7-cee.1
      - name: clife.v1.17.5-cee.1
        version: 1.17.5-cee.1
      - name: clife.v1.17.4-cee.1
        version: 1.17.4-cee.1
      - name: clife.v0.0.1
        version: 0.0.1
      name: "1.17"
    defaultChannel: "1.17"
    packageName: clife
    provider:
      name: Isovalent, part of Cisco
      url: https://isovalent.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microsafe
      provider-url: https://microsafe.com.mx/
    name: operator-aims-mobile
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operator-aims-mobile.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.microsafe.com.mx/v1",
                "kind": "AimsMobile",
                "metadata": {
                  "name": "aimsmobile-sample"
                },
                "spec": {
                  "adminApp": {
                    "autoscaling": {
                      "enabled": false
                    },
                    "config": {
                      "NODE_ENV": "production",
                      "VITE_EXPRESS_HOST": "aims-mob-service",
                      "VITE_EXPRESS_PORT": "8080",
                      "VITE_EXPRESS_PROTOCOL": "http",
                      "VITE_KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "VITE_KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "VITE_KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "VITE_VERSION": "1.0.0"
                    },
                    "image": "quay.io/microsafe_app_ban/aims-mob-admin-private@sha256:8e333f900885b0d3f6442c8ffcb43bd93540b8c38ad5bf17cbbd78929e0b1be0",
                    "ingress": {
                      "enabled": true,
                      "isOpenshiftRoute": true
                    },
                    "name": "aims-mob-admin",
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "secretName": "aims-mob-admin-secrets",
                    "service": {
                      "enabled": true,
                      "port": 8083,
                      "targetPort": 8080
                    }
                  },
                  "dbService": {
                    "autoscaling": {
                      "enabled": false
                    },
                    "config": {
                      "MSSQL_HOST": "mssql-database.prod.host",
                      "MSSQL_PORT": "mssql-database.prod.port"
                    },
                    "image": "quay.io/microsafe_app_ban/aims-db-service-private@sha256:bcb4d8624626436526549427f9f169cff714b39cbcc530dd2b3470c6712abaf9",
                    "ingress": {
                      "enabled": false
                    },
                    "name": "aims-db-service",
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "300m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "128Mi"
                      }
                    },
                    "secretName": "aims-db-service-secrets",
                    "service": {
                      "enabled": true,
                      "port": 3000,
                      "targetPort": 3000
                    }
                  },
                  "mainService": {
                    "annotations": {
                      "instrumentation.opentelemetry.io/inject-nodejs": "true"
                    },
                    "autoscaling": {
                      "enabled": true,
                      "maxReplicas": 2,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "config": {
                      "AIMS1_API_URL": "http://aims-db-service:3000",
                      "EXPRESS_PORT": "4000",
                      "FRONTEND_EXPRESS_HOST": "aims-mob-admin",
                      "FRONTEND_EXPRESS_PORT": "4000",
                      "FRONTEND_EXPRESS_PROTOCOL": "http",
                      "FRONTEND_IP": "aims-mob-admin",
                      "FRONTEND_KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "FRONTEND_KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "FRONTEND_KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "FRONTEND_PORT": "80",
                      "FRONTEND_PROTOCOL": "http",
                      "FRONTEND_VERSION": "1.0.0",
                      "KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "MSSQL_HOST": "mssql-database.prod.host",
                      "MSSQL_PORT": "mssql-database.prod.port",
                      "NODE_ENV": "production",
                      "OTLP_LOG_EXPORTER_URL": "opentelemetry-logs.prod.url",
                      "OTLP_METRIC_EXPORTER_URL": "opentelemetry-metrics.prod.url",
                      "OTLP_TRACE_EXPORTER_URL": "opentelemetry-traces.prod.url",
                      "REMEDY_API_URL": "remedy-aims.prod.svc",
                      "REMEDY_USER": "RMDYUSR"
                    },
                    "ingress": {
                      "enabled": true,
                      "isOpenshiftRoute": true
                    },
                    "name": "aims-mob-service",
                    "proxyContainer": {
                      "image": "quay.io/microsafe_app_ban/aims-mob-proxy-private@sha256:29b82de987b28c23f17d5f2176877c5aecea1a0ef182bb3a63612290ff469949",
                      "name": "aims-mob-service-proxy",
                      "resources": {
                        "limits": {
                          "cpu": "2000m",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "replicas": 2,
                    "secretName": "aims-mob-service-secrets",
                    "service": {
                      "enabled": true,
                      "port": 8085,
                      "targetPort": 8081
                    },
                    "serviceContainer": {
                      "image": "quay.io/microsafe_app_ban/aims-mob-service-private@sha256:f9586200106aa4de259f5cae2d22d13dc147a76dfd06e4f4bea09797071dad17",
                      "name": "aims-mob-service-main",
                      "resources": {
                        "limits": {
                          "cpu": "2500m",
                          "memory": "2048Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "128Mi"
                        }
                      }
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": false,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          createdAt: "2025-09-25T19:23:26Z"
          description: Este operador despliega los servicios AIMS Mobile.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/network-aware: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          instrumentation.opentelemetry.io/inject-nodejs: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: Valid Subscription Microsafe
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://quay.io/repository/microsafe_app_ban/operador-aims-mobile-private
          support: Microsafe
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AimsMobile
            name: aimsmobiles.cache.microsafe.com.mx
            version: v1
        description: Este operador despliega la aplicación AIMS Mobile.
        displayName: AIMS Mobile Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aims mobile
        - aims
        links:
        - name: AIMS Mobile Operator
          url: https://microsafe.com.mx/
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Microsafe
          url: https://microsafe.com.mx/
        relatedImages:
        - quay.io/microsafe_app_ban/aims-mob-service-private@sha256:f9586200106aa4de259f5cae2d22d13dc147a76dfd06e4f4bea09797071dad17
        - quay.io/microsafe_app_ban/aims-db-service-private@sha256:bcb4d8624626436526549427f9f169cff714b39cbcc530dd2b3470c6712abaf9
        - quay.io/microsafe_app_ban/operador-aims-mobile-private@sha256:731f2d731c79ade02f4b6c39b380590833cacddfd795b703cec10ad55f730fad
        - quay.io/microsafe_app_ban/aims-mob-admin-private@sha256:8e333f900885b0d3f6442c8ffcb43bd93540b8c38ad5bf17cbbd78929e0b1be0
        - quay.io/microsafe_app_ban/aims-mob-proxy-private@sha256:29b82de987b28c23f17d5f2176877c5aecea1a0ef182bb3a63612290ff469949
        version: 0.0.10
      entries:
      - name: operator-aims-mobile.v0.0.10
        version: 0.0.10
      - name: operator-aims-mobile.v0.0.9
        version: 0.0.9
      - name: operator-aims-mobile.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: operator-aims-mobile
    provider:
      name: Microsafe
      url: https://microsafe.com.mx/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzograph-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzograph-operator.v2.0.102
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzograph.clusters.cambridgesemantics.com/v2",
                "kind": "AnzoGraph",
                "metadata": {
                  "name": "azg01"
                },
                "spec": {
                  "db": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_data": "anzograph-data-grpc",
                            "app_mgmt": "anzograph-mgmt-grpc"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_data": "anzograph-data-grpc",
                              "app_mgmt": "anzograph-mgmt-grpc"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-db@sha256:8a5bd26551990929fdb776d6b421a9713cb84e03b19c18f81b625582f7f41af5",
                                "name": "db",
                                "resources": {
                                  "limits": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  },
                                  "requests": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  },
                  "deployFrontend": false,
                  "frontend": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_mgmt": "anzograph-frontend"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_mgmt": "anzograph-frontend"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-frontend@sha256:c1443bcdc8e998fc699f61980a665f1a083350adf2d3d1d55a7cb55e31ca50b3",
                                "name": "frontend",
                                "resources": {
                                  "limits": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-operator@sha256:a5b212558f748ddc8156541b546290320b656506e039a18903ad9fa6e638e732
          createdAt: "2021-08-11"
          description: kubernetes operator for AnzoGraph DB
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzograph-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzograph-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoGraph is the Schema for the anzographs API
            displayName: AnzoGraph Operator
            kind: AnzoGraph
            name: anzographs.anzograph.clusters.cambridgesemantics.com
            version: v2
        description: |-
          The AnzoGraph Operator provides the way to install and configure an AnzoGraph
          cluster on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v2.0.1/README_openshift_marketplace.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzograph/userdoc/ )

          ### Support

          We offer Support to our customers with the AnzoGraph db Enterprise Edition License [here]( https://customercenter.cambridgesemantics.com/ ). For AnzoGraph db Free Edition questions, get help from our Anzograph User Community at Stack Overflow. When submitting a question, include the tag 'anzograph'.
        displayName: AnzoGraph Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        - Database
        links:
        - name: Anzograph Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v2.0.1/README_openshift_marketplace.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-operator@sha256:a5b212558f748ddc8156541b546290320b656506e039a18903ad9fa6e638e732
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph@sha256:e3bdddf811b5ba69f388d023aea1ba538edd00e9f415c33ec555321d39129a36
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-db@sha256:8a5bd26551990929fdb776d6b421a9713cb84e03b19c18f81b625582f7f41af5
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-frontend@sha256:c1443bcdc8e998fc699f61980a665f1a083350adf2d3d1d55a7cb55e31ca50b3
        version: 2.0.1
      entries:
      - name: anzograph-operator.v2.0.102
        version: 2.0.1
      name: stable
    defaultChannel: stable
    packageName: anzograph-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: kubernetes-nmstate-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubernetes-nmstate-operator.4.13.0-202309261702
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nmstate.io/v1",
                "kind": "NMState",
                "metadata": {
                  "name": "nmstate"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift4/kubernetes-nmstate-rhel8-operator@sha256:bdcb80591b839ec6b4b94efa838ca41585f3044826534413b91efd2acdca4b4d
          createdAt: "2022-02-21 08:46:16"
          description: |
            Kubernetes NMState is a declaritive means of configuring NetworkManager.
          olm.skipRange: '>=4.3.0 <4.13.0'
          operatorframework.io/suggested-namespace: openshift-nmstate
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/kubernetes-nmstate
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an NMState deployment.
            displayName: NMState
            kind: NMState
            name: nmstates.nmstate.io
            version: v1
        description: A Kubernetes Operator to install Kubernetes NMState
        displayName: Kubernetes NMState Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - nmstate
        - networking
        - NetworkManager
        links:
        - name: Kubernetes Nmstate Operator
          url: https://github.com/nmstate/kubernetes-nmstate
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: GA
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/kubernetes-nmstate-rhel8-operator@sha256:bdcb80591b839ec6b4b94efa838ca41585f3044826534413b91efd2acdca4b4d
        - registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel8@sha256:a96434b959c1f61c1cfb4352284aaae453d2e41ca0824d016975a1216451b665
        version: 4.13.0-202309261702
      entries:
      - name: kubernetes-nmstate-operator.4.13.0-202309261702
        version: 4.13.0-202309261702
      name: "4.13"
    - currentCSV: kubernetes-nmstate-operator.4.16.0-202509241326
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nmstate.io/v1",
                "kind": "NMState",
                "metadata": {
                  "name": "nmstate"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:14750ac40ba2a9a92e2b37351eed94c22f2663747ea09f241a3a80607a9fca09
          createdAt: "2022-02-21 08:46:16"
          description: |
            Kubernetes NMState is a declaritive means of configuring NetworkManager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-nmstate
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/kubernetes-nmstate
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an NMState deployment.
            displayName: NMState
            kind: NMState
            name: nmstates.nmstate.io
            version: v1
        description: A Kubernetes Operator to install Kubernetes NMState
        displayName: Kubernetes NMState Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - nmstate
        - networking
        - NetworkManager
        links:
        - name: Kubernetes Nmstate Operator
          url: https://github.com/nmstate/kubernetes-nmstate
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: GA
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:14750ac40ba2a9a92e2b37351eed94c22f2663747ea09f241a3a80607a9fca09
        - registry.redhat.io/openshift4/nmstate-console-plugin-rhel9@sha256:8d03a6c414da30521e0794d3d61336bfbb1d9d4c3fec5ab5dac4bd24c897a540
        - registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel9@sha256:c592fdb65d99de2c3d274d7e72621408e578bdba39753bb425d0765ac016b11a
        version: 4.16.0-202509241326
      entries:
      - name: kubernetes-nmstate-operator.4.16.0-202509241326
        version: 4.16.0-202509241326
      - name: kubernetes-nmstate-operator.4.16.0-202509101329
        version: 4.16.0-202509101329
      - name: kubernetes-nmstate-operator.4.16.0-202508251030
        version: 4.16.0-202508251030
      - name: kubernetes-nmstate-operator.4.16.0-202507221936
        version: 4.16.0-202507221936
      - name: kubernetes-nmstate-operator.4.16.0-202507100305
        version: 4.16.0-202507100305
      - name: kubernetes-nmstate-operator.4.16.0-202506250035
        version: 4.16.0-202506250035
      - name: kubernetes-nmstate-operator.4.16.0-202506041738
        version: 4.16.0-202506041738
      - name: kubernetes-nmstate-operator.4.16.0-202505150035
        version: 4.16.0-202505150035
      - name: kubernetes-nmstate-operator.4.16.0-202505060805
        version: 4.16.0-202505060805
      - name: kubernetes-nmstate-operator.4.16.0-202504141106
        version: 4.16.0-202504141106
      - name: kubernetes-nmstate-operator.4.16.0-202503260136
        version: 4.16.0-202503260136
      - name: kubernetes-nmstate-operator.4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: kubernetes-nmstate-operator.4.16.0-202502190034
        version: 4.16.0-202502190034
      - name: kubernetes-nmstate-operator.4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: kubernetes-nmstate-operator.4.16.0-202501281135
        version: 4.16.0-202501281135
      - name: kubernetes-nmstate-operator.4.16.0-202501211505
        version: 4.16.0-202501211505
      - name: kubernetes-nmstate-operator.4.16.0-202501080105
        version: 4.16.0-202501080105
      - name: kubernetes-nmstate-operator.4.16.0-202412180105
        version: 4.16.0-202412180105
      - name: kubernetes-nmstate-operator.4.16.0-202412110604
        version: 4.16.0-202412110604
      - name: kubernetes-nmstate-operator.4.16.0-202412040032
        version: 4.16.0-202412040032
      - name: kubernetes-nmstate-operator.4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: kubernetes-nmstate-operator.4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: kubernetes-nmstate-operator.4.16.0-202410300036
        version: 4.16.0-202410300036
      - name: kubernetes-nmstate-operator.4.16.0-202410231737
        version: 4.16.0-202410231737
      - name: kubernetes-nmstate-operator.4.16.0-202410180404
        version: 4.16.0-202410180404
      - name: kubernetes-nmstate-operator.4.16.0-202410091137
        version: 4.16.0-202410091137
      - name: kubernetes-nmstate-operator.4.16.0-202410021935
        version: 4.16.0-202410021935
      - name: kubernetes-nmstate-operator.4.16.0-202409251605
        version: 4.16.0-202409251605
      - name: kubernetes-nmstate-operator.4.16.0-202409111235
        version: 4.16.0-202409111235
      - name: kubernetes-nmstate-operator.4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: kubernetes-nmstate-operator.4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: kubernetes-nmstate-operator.4.16.0-202408211407
        version: 4.16.0-202408211407
      - name: kubernetes-nmstate-operator.4.16.0-202408140143
        version: 4.16.0-202408140143
      - name: kubernetes-nmstate-operator.4.16.0-202408070540
        version: 4.16.0-202408070540
      - name: kubernetes-nmstate-operator.4.16.0-202407251436
        version: 4.16.0-202407251436
      - name: kubernetes-nmstate-operator.4.16.0-202407240806
        version: 4.16.0-202407240806
      - name: kubernetes-nmstate-operator.4.16.0-202407181806
        version: 4.16.0-202407181806
      - name: kubernetes-nmstate-operator.4.16.0-202407101306
        version: 4.16.0-202407101306
      - name: kubernetes-nmstate-operator.4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: kubernetes-nmstate-operator.4.16.0-202406260236
        version: 4.16.0-202406260236
      - name: kubernetes-nmstate-operator.4.16.0-202406140306
        version: 4.16.0-202406140306
      - name: kubernetes-nmstate-operator.4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: kubernetes-nmstate-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: istio-workspace-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: istio-workspace-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "workspace.maistra.io/v1alpha1",
                "kind": "Session",
                "metadata": {
                  "name": "session-sample"
                },
                "spec": {
                  "ref": [
                    {
                      "args": {
                        "image": "quay.io/maistra-dev/istio-workspace-test:latest"
                      },
                      "name": "ratings-v1",
                      "strategy": "prepared-image"
                    }
                  ],
                  "route": {
                    "name": "x-workspace-route",
                    "type": "header",
                    "value": "feature-x"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/maistra/istio-workspace:v0.5.3
          createdAt: "2022-11-16T13:50:35Z"
          description: Safely develop and test on any Kubernetes cluster without affecting
            others.
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/maistra/istio-workspace
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Session controls the creation of the specialized hidden routes.
            displayName: Session
            kind: Session
            name: sessions.workspace.maistra.io
            version: v1alpha1
        description: "Istio Workspace enables developers to:\n\n * run one or more
          services locally during development but make it appear like it would be
          developed in the actual cluster\n   * this way you can use your favourite
          development tools\n   * have instant feedback loop\n   * don't make your
          machine overheating trying to run the whole cluster locally\n * safely preview
          incoming changes in pull requests without affecting regular users \n * have
          confidence in testing or troubleshooting your services directly in the cluster\n\nIstio
          Workspace supports testing on multi-user environment in an unobtrusive way.\nIt
          doesn’t really matter if it is QE cluster or actual production. We give
          you confidence that your changes won’t blow up the cluster, and your users
          won’t even notice a glitch.\n\nYou will need [Kubernetes](https://k8s.io)
          or [Openshift](https://openshift.com) cluster with  [Istio](https://istio.io/)
          installed. \n\nYou can read more about our vision on [Red Hat Developer’s
          blog](https://developers.redhat.com/blog/2020/07/14/developing-and-testing-on-production-with-kubernetes-and-istio-workspace/)
          or ...\n\n## See it in action!\n\n[![https://youtu.be/XTNVadUzMCc](https://img.youtube.com/vi/XTNVadUzMCc/hqdefault.jpg)](https://youtu.be/XTNVadUzMCc)\n\nWatch
          demo: [\"How to develop on production: An introduction to Istio-Workspaces\"](https://youtu.be/XTNVadUzMCc).\n\n##
          Documentation\n\nMore details can be found on our [documentation page](https://istio-workspace-docs.netlify.com/).\n"
        displayName: Istio Workspace
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - telepresence
        links:
        - name: Istio Workspace Documentation
          url: https://istio-workspace-docs.netlify.com/
        - name: How to install CLI tool
          url: https://github.com/maistra/istio-workspace#install-in-two-easy-steps
        maintainers:
        - email: aslak@redhat.com
          name: Aslak Knutsen
        - email: bartosz@redhat.com
          name: Bartosz Majsak
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/maistra/istio-workspace:v0.5.3
        version: 0.5.3
      entries:
      - name: istio-workspace-operator.v0.5.3
        version: 0.5.3
      - name: istio-workspace-operator.v0.5.2
        version: 0.5.2
      - name: istio-workspace-operator.v0.5.1
        version: 0.5.1
      - name: istio-workspace-operator.v0.5.0
        version: 0.5.0
      - name: istio-workspace-operator.v0.4.0
        version: 0.4.0
      - name: istio-workspace-operator.v0.3.0
        version: 0.3.0
      - name: istio-workspace-operator.v0.2.0
        version: 0.2.0
      - name: istio-workspace-operator.v0.1.0
        version: 0.1.0
      - name: istio-workspace-operator.v0.0.10
        version: 0.0.10
      - name: istio-workspace-operator.v0.0.9
        version: 0.0.9
      - name: istio-workspace-operator.v0.0.8
        version: 0.0.8
      - name: istio-workspace-operator.v0.0.7
        version: 0.0.7
      - name: istio-workspace-operator.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: istio-workspace-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ThinkParQ
      provider-url: https://www.thinkparq.com
    name: beegfs-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: beegfs-csi-driver-operator.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "beegfs.csi.netapp.com/v1",
                "kind": "BeegfsDriver",
                "metadata": {
                  "name": "csi-beegfs-cr"
                },
                "spec": {
                  "nodeAffinityControllerService": {
                    "preferredDuringSchedulingIgnoredDuringExecution": [
                      {
                        "preference": {
                          "matchExpressions": [
                            {
                              "key": "node-role.kubernetes.io/master",
                              "operator": "Exists"
                            }
                          ]
                        },
                        "weight": 50
                      }
                    ],
                    "requiredDuringSchedulingIgnoredDuringExecution": {
                      "nodeSelectorTerms": [
                        {
                          "matchExpressions": [
                            {
                              "key": "node.openshift.io/os_id",
                              "operator": "NotIn",
                              "values": [
                                "rhcos"
                              ]
                            }
                          ]
                        }
                      ]
                    }
                  },
                  "nodeAffinityNodeService": {
                    "requiredDuringSchedulingIgnoredDuringExecution": {
                      "nodeSelectorTerms": [
                        {
                          "matchExpressions": [
                            {
                              "key": "node.openshift.io/os_id",
                              "operator": "NotIn",
                              "values": [
                                "rhcos"
                              ]
                            }
                          ]
                        }
                      ]
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: ghcr.io/thinkparq/beegfs-csi-driver-operator:v1.7.0
          createdAt: 2021-06-30:03:11:00
          description: Install and manage the BeeGFS CSI driver
          operatorframework.io/suggested-namespace: beegfs-csi
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ThinkParQ/beegfs-csi-driver
          support: ThinkParQ
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys the BeeGFS CSI driver
            displayName: BeeGFS Driver
            kind: BeegfsDriver
            name: beegfsdrivers.beegfs.csi.netapp.com
            version: v1
        description: |
          The BeeGFS Container Storage Interface (CSI) driver provides high performing and scalable storage for workloads
          running in container orchestrators like Kubernetes. It allows containers to access existing datasets or request
          on-demand ephemeral or persistent high speed storage backed by BeeGFS parallel file systems.

          This operator eases the installation and management of the BeeGFS CSI driver in an OpenShift or other Operator
          Lifecycle Manager enabled cluster.

          ### Documentation

          Compatibility information and a quick start guide for the BeeGFS CSI driver itself are available on our [main GitHub
          page](https://www.github.com/ThinkParQ/beegfs-csi-driver). Additional detailed documentation for the driver is
          available in the [docs directory](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/docs) and for the
          operator is available in the [operator
          README](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/operator/README.md).

          ### Important Notes

          The BeeGFS client packages and kernel module must be pre-installed on all nodes that will run the driver. See the
          driver ["Getting Started" guide](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/README.md#getting-started)
          for additional information.

          The BeeGFS CSI driver is only supported on nodes running a [BeeGFS supported Linux
          distribution](https://doc.beegfs.io/latest/release_notes.html#supported-linux-distributions-and-kernels). Red Hat
          CoreOS (RHCOS), the default distribution in OpenShift environments, is NOT supported. However, Red Hat Enterprise
          Linux (RHEL) nodes can be added to an OpenShift cluster and RHEL is supported for running BeeGFS. By default, the
          operator will install the driver with a node affinity that ensures it does not run on RHCOS nodes. See the driver
          [compatibility
          matrix](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/README.md#interoperability-and-csi-feature-matrix)
          for additional information.

          We have recently published experimental manifests that enable the BeeGFS CSI driver on OpenShift nodes running
          RHCOS. These manifests are not intended for production use. See [this
          README](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/deploy/openshift-beegfs-client/README.md) for more
          details.

          All driver configuration can be specified through the BeegfsDrivers API (in a BeegfsDriver Custom Resource)
          except for BeeGFS connAuth information, which must be placed separately into a Kubernetes Secret. See the [ConnAuth
          Configuration
          section](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/operator/README.md#connauth-configuration) of the
          operator documentation for additional information.

          Starting with version 1.4.0 of the BeeGFS CSI driver the driver and operator images will be signed using the Cosign
          tool. Refer to the [BeeGFS CSI driver operator
          README](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/operator/README.md)
          for details on how to verify the operator and driver image signatures.
        displayName: BeeGFS CSI Driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CSI
        - BeeGFS
        - ThinkParQ
        - HPC
        - AI
        - High-Performance Computing
        links:
        - name: BeeGFS introduction
          url: https://www.beegfs.io/c/
        - name: BeeGFS documentation
          url: https://doc.beegfs.io/latest/index.html
        - name: BeeGFS CSI driver source code
          url: https://github.com/ThinkParQ/beegfs-csi-driver
        - name: BeeGFS blog posts
          url: https://www.beegfs.io/c/blog/
        maintainers:
        - email: csi-driver@beegfs.io
          name: ThinkParQ
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: ThinkParQ
          url: https://www.thinkparq.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - ghcr.io/thinkparq/beegfs-csi-driver-operator:v1.7.0
        version: 1.7.0
      entries:
      - name: beegfs-csi-driver-operator.v1.7.0
        version: 1.7.0
      - name: beegfs-csi-driver-operator.v1.6.0
        version: 1.6.0
      - name: beegfs-csi-driver-operator.v1.5.0
        version: 1.5.0
      - name: beegfs-csi-driver-operator.v1.4.0
        version: 1.4.0
      - name: beegfs-csi-driver-operator.v1.3.0
        version: 1.3.0
      - name: beegfs-csi-driver-operator.v1.2.2
        version: 1.2.2
      - name: beegfs-csi-driver-operator.v1.2.1
        version: 1.2.1
      - name: beegfs-csi-driver-operator.v1.2.0
        version: 1.2.0
      name: stable
    defaultChannel: stable
    packageName: beegfs-csi-driver-operator
    provider:
      name: ThinkParQ
      url: https://www.thinkparq.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stackwatch
      provider-url: www.kubecost.com
    name: kubecost
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubecost.v2.2.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kubecost.com/v1alpha1",
                "kind": "CostAnalyzer",
                "metadata": {
                  "name": "costanalyzer-sample"
                },
                "spec": {
                  "global": {
                    "prometheus": {
                      "enabled": true,
                      "fqdn": "http://cost-analyzer-prometheus-server.default.svc"
                    },
                    "grafana": {
                      "enabled": false,
                      "domainName": "cost-analyzer-grafana.default.svc",
                      "scheme": "http",
                      "proxy": false
                    },
                    "gcpstore": {
                      "enabled": false
                    },
                    "gmp": {
                      "enabled": false,
                      "prometheusServerEndpoint": "http://localhost:8085/",
                      "gmpProxy": {
                        "enabled": false,
                        "image": "gke.gcr.io/prometheus-engine/frontend:v0.4.1-gke.0",
                        "imagePullPolicy": "Always",
                        "name": "gmp-proxy",
                        "port": 8085,
                        "projectId": "YOUR_PROJECT_ID"
                      }
                    },
                    "amp": {
                      "enabled": false,
                      "prometheusServerEndpoint": "http://localhost:8005/workspaces/<workspaceId>/",
                      "remoteWriteService": "https://aps-workspaces.us-west-2.amazonaws.com/workspaces/<workspaceId>/api/v1/remote_write",
                      "sigv4": {
                        "region": "us-west-2"
                      }
                    },
                    "mimirProxy": {
                      "enabled": false,
                      "name": "mimir-proxy",
                      "image": "nginxinc/nginx-unprivileged",
                      "port": 8085,
                      "mimirEndpoint": "$mimir_endpoint",
                      "orgIdentifier": "$your_tenant_ID"
                    },
                    "notifications": {
                      "alertmanager": {
                        "enabled": false,
                        "fqdn": "http://cost-analyzer-prometheus-server.default.svc"
                      }
                    },
                    "savedReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Saved Report 0",
                          "window": "today",
                          "aggregateBy": "namespace",
                          "chartDisplay": "category",
                          "idle": "separate",
                          "rate": "cumulative",
                          "accumulate": false,
                          "filters": [
                            {
                              "key": "cluster",
                              "operator": ":",
                              "value": "dev"
                            }
                          ]
                        },
                        {
                          "title": "Example Saved Report 1",
                          "window": "month",
                          "aggregateBy": "controllerKind",
                          "chartDisplay": "category",
                          "idle": "share",
                          "rate": "monthly",
                          "accumulate": false,
                          "filters": [
                            {
                              "key": "namespace",
                              "operator": "!:",
                              "value": "kubecost"
                            }
                          ]
                        },
                        {
                          "title": "Example Saved Report 2",
                          "window": "2020-11-11T00:00:00Z,2020-12-09T23:59:59Z",
                          "aggregateBy": "service",
                          "chartDisplay": "category",
                          "idle": "hide",
                          "rate": "daily",
                          "accumulate": true,
                          "filters": []
                        }
                      ]
                    },
                    "assetReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Asset Report 0",
                          "window": "today",
                          "aggregateBy": "type",
                          "accumulate": false,
                          "filters": [
                            {
                              "property": "cluster",
                              "value": "cluster-one"
                            }
                          ]
                        }
                      ]
                    },
                    "advancedReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Advanced Report 0",
                          "window": "7d",
                          "aggregateBy": "namespace",
                          "filters": [
                            {
                              "key": "cluster",
                              "operator": ":",
                              "value": "dev"
                            }
                          ],
                          "cloudBreakdown": "service",
                          "cloudJoin": "label:kubernetes_namespace"
                        }
                      ]
                    },
                    "cloudCostReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Cloud Cost Report 0",
                          "window": "today",
                          "aggregateBy": "service",
                          "accumulate": false
                        }
                      ]
                    },
                    "podAnnotations": {},
                    "additionalLabels": {},
                    "securityContext": {
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      },
                      "fsGroup": 1001,
                      "runAsGroup": 1001,
                      "runAsUser": 1001,
                      "fsGroupChangePolicy": "OnRootMismatch"
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "privileged": false,
                      "readOnlyRootFilesystem": true,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "platforms": {
                      "openshift": {
                        "enabled": true,
                        "route": {
                          "enabled": false,
                          "annotations": {}
                        },
                        "scc": {
                          "nodeExporter": false,
                          "networkCosts": false
                        },
                        "securityContext": {
                          "runAsNonRoot": true,
                          "seccompProfile": {
                            "type": "RuntimeDefault"
                          }
                        }
                      },
                      "cicd": {
                        "enabled": true,
                        "skipSanityChecks": true
                      }
                    }
                  },
                  "upgrade": {
                    "toV2": false
                  },
                  "kubecostToken": null,
                  "pricingCsv": {
                    "enabled": false,
                    "location": {
                      "provider": "AWS",
                      "region": "us-east-1",
                      "URI": "s3://kc-csv-test/pricing_schema.csv",
                      "csvAccessCredentials": "pricing-schema-access-secret"
                    }
                  },
                  "saml": {
                    "enabled": false,
                    "rbac": {
                      "enabled": false
                    }
                  },
                  "oidc": {
                    "enabled": false,
                    "clientID": "",
                    "clientSecret": "",
                    "existingCustomSecret": {
                      "enabled": false,
                      "name": ""
                    },
                    "skipOnlineTokenValidation": false,
                    "rbac": {
                      "enabled": false
                    }
                  },
                  "systemProxy": {
                    "enabled": false,
                    "httpProxyUrl": "",
                    "httpsProxyUrl": "",
                    "noProxy": ""
                  },
                  "kubecostFrontend": {
                    "enabled": true,
                    "deployMethod": "singlepod",
                    "haReplicas": 2,
                    "image": "gcr.io/kubecost1/frontend",
                    "imagePullPolicy": "Always",
                    "resources": {
                      "requests": {
                        "cpu": "10m",
                        "memory": "55Mi"
                      }
                    },
                    "deploymentStrategy": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 1,
                      "periodSeconds": 5,
                      "failureThreshold": 6
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 1,
                      "periodSeconds": 5,
                      "failureThreshold": 6
                    },
                    "ipv6": {
                      "enabled": true
                    },
                    "useDefaultFqdn": false
                  },
                  "kubecostMetrics": {
                    "exporter": {
                      "enabled": false,
                      "port": 9005,
                      "prometheusScrape": true,
                      "resources": {},
                      "tolerations": [],
                      "affinity": {},
                      "service": {
                        "annotations": {}
                      },
                      "serviceMonitor": {
                        "enabled": false,
                        "additionalLabels": {},
                        "metricRelabelings": [],
                        "relabelings": []
                      },
                      "priorityClassName": "",
                      "additionalLabels": {},
                      "nodeSelector": {},
                      "extraArgs": []
                    }
                  },
                  "sigV4Proxy": {
                    "image": "public.ecr.aws/aws-observability/aws-sigv4-proxy:latest",
                    "imagePullPolicy": "Always",
                    "name": "aps",
                    "port": 8005,
                    "region": "us-west-2",
                    "host": "aps-workspaces.us-west-2.amazonaws.com",
                    "extraEnv": null,
                    "resources": {}
                  },
                  "kubecostModel": {
                    "image": "gcr.io/kubecost1/cost-model",
                    "imagePullPolicy": "Always",
                    "outOfClusterPromMetricsEnabled": false,
                    "warmCache": false,
                    "etl": true,
                    "etlFileStoreEnabled": true,
                    "etlDailyStoreDurationDays": 91,
                    "etlHourlyStoreDurationHours": 49,
                    "etlWeeklyStoreDurationWeeks": 53,
                    "etlReadOnlyMode": false,
                    "plugins": {
                      "enabled": false,
                      "install": {
                        "enabled": false,
                        "fullImageName": "curlimages/curl:latest",
                        "securityContext": {
                          "allowPrivilegeEscalation": false,
                          "seccompProfile": {
                            "type": "RuntimeDefault"
                          },
                          "capabilities": {
                            "drop": [
                              "ALL"
                            ]
                          },
                          "readOnlyRootFilesystem": true,
                          "runAsNonRoot": true,
                          "runAsUser": 1001
                        }
                      },
                      "folder": "/opt/opencost/plugin",
                      "enabledPlugins": [],
                      "configSecret": "kubecost-plugin-secret"
                    },
                    "cloudCost": {
                      "labelList": {
                        "IsIncludeList": false,
                        "labels": ""
                      },
                      "topNItems": 1000
                    },
                    "allocation": null,
                    "containerStatsEnabled": true,
                    "maxQueryConcurrency": 5,
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "55Mi"
                      }
                    },
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "extraArgs": [],
                    "ingress": {
                      "enabled": false,
                      "labels": null,
                      "annotations": null,
                      "paths": [
                        "/"
                      ],
                      "pathType": "ImplementationSpecific",
                      "hosts": [
                        "cost-analyzer-model.local"
                      ],
                      "tls": []
                    },
                    "utcOffset": "+00:00",
                    "extraPorts": []
                  },
                  "etlUtils": {
                    "enabled": false,
                    "fullImageName": null,
                    "resources": {},
                    "env": {},
                    "nodeSelector": {},
                    "tolerations": {},
                    "affinity": {}
                  },
                  "ingress": {
                    "enabled": false,
                    "labels": null,
                    "annotations": null,
                    "paths": [
                      "/"
                    ],
                    "pathType": "ImplementationSpecific",
                    "hosts": [
                      "cost-analyzer.local"
                    ],
                    "tls": []
                  },
                  "nodeSelector": {},
                  "tolerations": [],
                  "affinity": {},
                  "topologySpreadConstraints": [],
                  "priority": {
                    "enabled": false,
                    "name": ""
                  },
                  "networkPolicy": {
                    "enabled": false,
                    "denyEgress": true,
                    "sameNamespace": true,
                    "costAnalyzer": {
                      "enabled": false,
                      "annotations": {},
                      "additionalLabels": {}
                    }
                  },
                  "extraVolumes": [],
                  "extraVolumeMounts": [],
                  "persistentVolume": {
                    "size": "32Gi",
                    "dbSize": "32.0Gi",
                    "enabled": true,
                    "labels": {},
                    "annotations": {},
                    "dbPVEnabled": false
                  },
                  "service": {
                    "type": "ClusterIP",
                    "port": 9090,
                    "targetPort": 9090,
                    "nodePort": {},
                    "labels": {},
                    "annotations": {},
                    "sessionAffinity": {
                      "enabled": false,
                      "timeoutSeconds": 10800
                    }
                  },
                  "prometheus": {
                    "rbac": {
                      "create": true
                    },
                    "serviceAccounts": {
                      "alertmanager": {
                        "create": true,
                        "name": null
                      },
                      "nodeExporter": {
                        "create": true,
                        "name": null
                      },
                      "pushgateway": {
                        "create": true,
                        "name": null
                      },
                      "server": {
                        "create": true,
                        "name": null,
                        "annotations": {}
                      }
                    },
                    "imagePullSecrets": null,
                    "extraScrapeConfigs": "- job_name: kubecost\n  honor_labels: true\n  scrape_interval: 1m\n  scrape_timeout: 60s\n  metrics_path: /metrics\n  scheme: http\n  dns_sd_configs:\n  - names:\n    - {{ template \"cost-analyzer.serviceName\" . }}\n    type: 'A'\n    port: 9003\n- job_name: kubecost-networking\n  kubernetes_sd_configs:\n    - role: pod\n  relabel_configs:\n  # Scrape only the the targets matching the following metadata\n    - source_labels: [__meta_kubernetes_pod_label_app_kubernetes_io_instance]\n      action: keep\n      regex:  kubecost\n    - source_labels: [__meta_kubernetes_pod_label_app_kubernetes_io_name]\n      action: keep\n      regex:  network-costs\n",
                    "server": {
                      "enabled": true,
                      "name": "server",
                      "sidecarContainers": null,
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "image": {
                        "repository": "quay.io/prometheus/prometheus",
                        "tag": "v2.50.1",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "prefixURL": "",
                      "baseURL": "",
                      "env": [],
                      "extraFlags": [
                        "web.enable-lifecycle"
                      ],
                      "configPath": "/etc/config/prometheus.yml",
                      "global": {
                        "scrape_interval": "1m",
                        "scrape_timeout": "60s",
                        "evaluation_interval": "1m",
                        "external_labels": {
                          "cluster_id": "cluster-one"
                        }
                      },
                      "remoteWrite": {},
                      "remoteRead": {},
                      "extraArgs": {
                        "query.max-concurrency": 1,
                        "query.max-samples": 100000000
                      },
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "extraHostPathMounts": [],
                      "extraConfigmapMounts": [],
                      "extraSecretMounts": [],
                      "configMapOverrideName": "",
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "extraLabels": {},
                        "hosts": [],
                        "pathType": "Prefix",
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "affinity": {},
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "32Gi",
                        "subPath": ""
                      },
                      "emptyDir": {
                        "sizeLimit": ""
                      },
                      "podAnnotations": {},
                      "deploymentAnnotations": {},
                      "podLabels": {},
                      "alertmanagers": [],
                      "replicaCount": 1,
                      "statefulSet": {
                        "enabled": false,
                        "annotations": {},
                        "labels": {},
                        "podManagementPolicy": "OrderedReady",
                        "headless": {
                          "annotations": {},
                          "labels": {},
                          "servicePort": 80
                        }
                      },
                      "readinessProbeInitialDelay": 5,
                      "readinessProbeTimeout": 3,
                      "readinessProbeFailureThreshold": 3,
                      "readinessProbeSuccessThreshold": 1,
                      "livenessProbeInitialDelay": 5,
                      "livenessProbeTimeout": 3,
                      "livenessProbeFailureThreshold": 3,
                      "livenessProbeSuccessThreshold": 1,
                      "resources": {},
                      "verticalAutoscaler": {
                        "enabled": false
                      },
                      "securityContext": {},
                      "containerSecurityContext": {},
                      "service": {
                        "annotations": {},
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 80,
                        "sessionAffinity": "None",
                        "type": "ClusterIP",
                        "gRPC": {
                          "enabled": false,
                          "servicePort": 10901
                        },
                        "statefulsetReplica": {
                          "enabled": false,
                          "replica": 0
                        }
                      },
                      "terminationGracePeriodSeconds": 300,
                      "retention": "97h"
                    },
                    "alertmanager": {
                      "enabled": false,
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "name": "alertmanager",
                      "image": {
                        "repository": "quay.io/prometheus/alertmanager",
                        "tag": "v0.27.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "extraArgs": {},
                      "prefixURL": "",
                      "baseURL": "http://localhost:9093",
                      "extraEnv": {},
                      "extraSecretMounts": [],
                      "configMapOverrideName": "",
                      "configFromSecret": "",
                      "configFileName": "alertmanager.yml",
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "extraLabels": {},
                        "hosts": [],
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "affinity": {},
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "2Gi",
                        "subPath": ""
                      },
                      "podAnnotations": {},
                      "podLabels": {},
                      "replicaCount": 1,
                      "statefulSet": {
                        "enabled": false,
                        "podManagementPolicy": "OrderedReady",
                        "headless": {
                          "annotations": {},
                          "labels": {},
                          "servicePort": 80
                        }
                      },
                      "resources": {},
                      "securityContext": {
                        "runAsUser": 1001,
                        "runAsNonRoot": true,
                        "runAsGroup": 1001,
                        "fsGroup": 1001
                      },
                      "service": {
                        "annotations": {},
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 80,
                        "sessionAffinity": "None",
                        "type": "ClusterIP"
                      }
                    },
                    "alertmanagerFiles": {
                      "alertmanager.yml": {
                        "global": {},
                        "receivers": [
                          {
                            "name": "default-receiver"
                          }
                        ],
                        "route": {
                          "group_wait": "10s",
                          "group_interval": "5m",
                          "receiver": "default-receiver",
                          "repeat_interval": "3h"
                        }
                      }
                    },
                    "configmapReload": {
                      "prometheus": {
                        "enabled": false,
                        "name": "configmap-reload",
                        "image": {
                          "repository": "quay.io/prometheus-operator/prometheus-config-reloader",
                          "tag": "v0.72.0",
                          "pullPolicy": "IfNotPresent"
                        },
                        "extraArgs": {},
                        "extraVolumeDirs": [],
                        "extraConfigmapMounts": [],
                        "resources": {},
                        "containerSecurityContext": {}
                      },
                      "alertmanager": {
                        "enabled": false,
                        "name": "configmap-reload",
                        "image": {
                          "repository": "quay.io/prometheus-operator/prometheus-config-reloader",
                          "tag": "v0.72.0",
                          "pullPolicy": "IfNotPresent"
                        },
                        "extraArgs": {},
                        "extraVolumeDirs": [],
                        "extraConfigmapMounts": [],
                        "resources": {}
                      }
                    },
                    "nodeExporter": {
                      "enabled": false,
                      "hostNetwork": true,
                      "hostPID": true,
                      "dnsPolicy": "ClusterFirstWithHostNet",
                      "name": "node-exporter",
                      "image": {
                        "repository": "prom/node-exporter",
                        "tag": "v1.7.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "extraArgs": {},
                      "extraHostPathMounts": [],
                      "extraConfigmapMounts": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "deploymentAnnotations": {},
                      "pod": {
                        "labels": {}
                      },
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "resources": {},
                      "securityContext": {},
                      "service": {
                        "annotations": {
                          "prometheus.io/scrape": "true"
                        },
                        "labels": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "hostPort": 9100,
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 9100,
                        "type": "ClusterIP"
                      }
                    },
                    "pushgateway": {
                      "enabled": false,
                      "name": "pushgateway",
                      "image": {
                        "repository": "prom/pushgateway",
                        "tag": "v1.6.2",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "extraArgs": {},
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "hosts": [],
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "replicaCount": 1,
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "resources": {},
                      "securityContext": {
                        "runAsUser": 1001,
                        "runAsNonRoot": true
                      },
                      "service": {
                        "annotations": {
                          "prometheus.io/probe": "pushgateway"
                        },
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 9091,
                        "type": "ClusterIP"
                      },
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "2Gi",
                        "subPath": ""
                      }
                    },
                    "serverFiles": {
                      "alerting_rules.yml": {},
                      "alerts": {},
                      "recording_rules.yml": {},
                      "prometheus.yml": {
                        "rule_files": [
                          "/etc/config/recording_rules.yml",
                          "/etc/config/alerting_rules.yml",
                          "/etc/config/rules",
                          "/etc/config/alerts"
                        ],
                        "scrape_configs": [
                          {
                            "job_name": "prometheus",
                            "static_configs": [
                              {
                                "targets": [
                                  "localhost:9090"
                                ]
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-nodes-cadvisor",
                            "scheme": "https",
                            "tls_config": {
                              "ca_file": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                              "insecure_skip_verify": true
                            },
                            "bearer_token_file": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                            "kubernetes_sd_configs": [
                              {
                                "role": "node"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_node_label_(.+)"
                              },
                              {
                                "target_label": "__address__",
                                "replacement": "kubernetes.default.svc:443"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_node_name"
                                ],
                                "regex": "(.+)",
                                "target_label": "__metrics_path__",
                                "replacement": "/api/v1/nodes/$1/proxy/metrics/cadvisor"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(container_cpu_usage_seconds_total|container_memory_working_set_bytes|container_network_receive_errors_total|container_network_transmit_errors_total|container_network_receive_packets_dropped_total|container_network_transmit_packets_dropped_total|container_memory_usage_bytes|container_cpu_cfs_throttled_periods_total|container_cpu_cfs_periods_total|container_fs_usage_bytes|container_fs_limit_bytes|container_cpu_cfs_periods_total|container_fs_inodes_free|container_fs_inodes_total|container_fs_usage_bytes|container_fs_limit_bytes|container_cpu_cfs_throttled_periods_total|container_cpu_cfs_periods_total|container_network_receive_bytes_total|container_network_transmit_bytes_total|container_fs_inodes_free|container_fs_inodes_total|container_fs_usage_bytes|container_fs_limit_bytes|container_spec_cpu_shares|container_spec_memory_limit_bytes|container_network_receive_bytes_total|container_network_transmit_bytes_total|container_fs_reads_bytes_total|container_network_receive_bytes_total|container_fs_writes_bytes_total|container_fs_reads_bytes_total|cadvisor_version_info|kubecost_pv_info)",
                                "action": "keep"
                              },
                              {
                                "source_labels": [
                                  "container"
                                ],
                                "target_label": "container_name",
                                "regex": "(.+)",
                                "action": "replace"
                              },
                              {
                                "source_labels": [
                                  "pod"
                                ],
                                "target_label": "pod_name",
                                "regex": "(.+)",
                                "action": "replace"
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-nodes",
                            "scheme": "https",
                            "tls_config": {
                              "ca_file": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                              "insecure_skip_verify": true
                            },
                            "bearer_token_file": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                            "kubernetes_sd_configs": [
                              {
                                "role": "node"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_node_label_(.+)"
                              },
                              {
                                "target_label": "__address__",
                                "replacement": "kubernetes.default.svc:443"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_node_name"
                                ],
                                "regex": "(.+)",
                                "target_label": "__metrics_path__",
                                "replacement": "/api/v1/nodes/$1/proxy/metrics"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(kubelet_volume_stats_used_bytes)",
                                "action": "keep"
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-service-endpoints",
                            "kubernetes_sd_configs": [
                              {
                                "role": "endpoints"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_scrape"
                                ],
                                "action": "keep",
                                "regex": true
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_endpoints_name"
                                ],
                                "action": "keep",
                                "regex": "(.*node-exporter|kubecost-network-costs)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_scheme"
                                ],
                                "action": "replace",
                                "target_label": "__scheme__",
                                "regex": "(https?)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_path"
                                ],
                                "action": "replace",
                                "target_label": "__metrics_path__",
                                "regex": "(.+)"
                              },
                              {
                                "source_labels": [
                                  "__address__",
                                  "__meta_kubernetes_service_annotation_prometheus_io_port"
                                ],
                                "action": "replace",
                                "target_label": "__address__",
                                "regex": "([^:]+)(?::\\d+)?;(\\d+)",
                                "replacement": "$1:$2"
                              },
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_service_label_(.+)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_namespace"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_namespace"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_name"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_name"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_pod_node_name"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_node"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(container_cpu_allocation|container_cpu_usage_seconds_total|container_fs_limit_bytes|container_fs_writes_bytes_total|container_gpu_allocation|container_memory_allocation_bytes|container_memory_usage_bytes|container_memory_working_set_bytes|container_network_receive_bytes_total|container_network_transmit_bytes_total|DCGM_FI_DEV_GPU_UTIL|deployment_match_labels|kube_daemonset_status_desired_number_scheduled|kube_daemonset_status_number_ready|kube_deployment_spec_replicas|kube_deployment_status_replicas|kube_deployment_status_replicas_available|kube_job_status_failed|kube_namespace_annotations|kube_namespace_labels|kube_node_info|kube_node_labels|kube_node_status_allocatable|kube_node_status_allocatable_cpu_cores|kube_node_status_allocatable_memory_bytes|kube_node_status_capacity|kube_node_status_capacity_cpu_cores|kube_node_status_capacity_memory_bytes|kube_node_status_condition|kube_persistentvolume_capacity_bytes|kube_persistentvolume_status_phase|kube_persistentvolumeclaim_info|kube_persistentvolumeclaim_resource_requests_storage_bytes|kube_pod_container_info|kube_pod_container_resource_limits|kube_pod_container_resource_limits_cpu_cores|kube_pod_container_resource_limits_memory_bytes|kube_pod_container_resource_requests|kube_pod_container_resource_requests_cpu_cores|kube_pod_container_resource_requests_memory_bytes|kube_pod_container_status_restarts_total|kube_pod_container_status_running|kube_pod_container_status_terminated_reason|kube_pod_labels|kube_pod_owner|kube_pod_status_phase|kube_replicaset_owner|kube_statefulset_replicas|kube_statefulset_status_replicas|kubecost_cluster_info|kubecost_cluster_management_cost|kubecost_cluster_memory_working_set_bytes|kubecost_load_balancer_cost|kubecost_network_internet_egress_cost|kubecost_network_region_egress_cost|kubecost_network_zone_egress_cost|kubecost_node_is_spot|kubecost_pod_network_egress_bytes_total|node_cpu_hourly_cost|node_cpu_seconds_total|node_disk_reads_completed|node_disk_reads_completed_total|node_disk_writes_completed|node_disk_writes_completed_total|node_filesystem_device_error|node_gpu_count|node_gpu_hourly_cost|node_memory_Buffers_bytes|node_memory_Cached_bytes|node_memory_MemAvailable_bytes|node_memory_MemFree_bytes|node_memory_MemTotal_bytes|node_network_transmit_bytes_total|node_ram_hourly_cost|node_total_hourly_cost|pod_pvc_allocation|pv_hourly_cost|service_selector_labels|statefulSet_match_labels|kubecost_pv_info|up)",
                                "action": "keep"
                              }
                            ]
                          }
                        ]
                      },
                      "rules": {
                        "groups": [
                          {
                            "name": "CPU",
                            "rules": [
                              {
                                "expr": "sum(rate(container_cpu_usage_seconds_total{container!=\"\"}[5m]))",
                                "record": "cluster:cpu_usage:rate5m"
                              },
                              {
                                "expr": "rate(container_cpu_usage_seconds_total{container!=\"\"}[5m])",
                                "record": "cluster:cpu_usage_nosum:rate5m"
                              },
                              {
                                "expr": "avg(irate(container_cpu_usage_seconds_total{container!=\"POD\", container!=\"\"}[5m])) by (container,pod,namespace)",
                                "record": "kubecost_container_cpu_usage_irate"
                              },
                              {
                                "expr": "sum(container_memory_working_set_bytes{container!=\"POD\",container!=\"\"}) by (container,pod,namespace)",
                                "record": "kubecost_container_memory_working_set_bytes"
                              },
                              {
                                "expr": "sum(container_memory_working_set_bytes{container!=\"POD\",container!=\"\"})",
                                "record": "kubecost_cluster_memory_working_set_bytes"
                              }
                            ]
                          },
                          {
                            "name": "Savings",
                            "rules": [
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind!=\"DaemonSet\"}) by (pod) * sum(container_cpu_allocation) by (pod))",
                                "record": "kubecost_savings_cpu_allocation",
                                "labels": {
                                  "daemonset": "false"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind=\"DaemonSet\"}) by (pod) * sum(container_cpu_allocation) by (pod)) / sum(kube_node_info)",
                                "record": "kubecost_savings_cpu_allocation",
                                "labels": {
                                  "daemonset": "true"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind!=\"DaemonSet\"}) by (pod) * sum(container_memory_allocation_bytes) by (pod))",
                                "record": "kubecost_savings_memory_allocation_bytes",
                                "labels": {
                                  "daemonset": "false"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind=\"DaemonSet\"}) by (pod) * sum(container_memory_allocation_bytes) by (pod)) / sum(kube_node_info)",
                                "record": "kubecost_savings_memory_allocation_bytes",
                                "labels": {
                                  "daemonset": "true"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    "alertRelabelConfigs": null,
                    "networkPolicy": {
                      "enabled": false
                    }
                  },
                  "networkCosts": {
                    "enabled": false,
                    "image": {
                      "repository": "gcr.io/kubecost1/kubecost-network-costs",
                      "tag": "v0.17.3"
                    },
                    "imagePullPolicy": "Always",
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "prometheusScrape": false,
                    "trafficLogging": true,
                    "logLevel": "info",
                    "port": 3001,
                    "resources": {
                      "limits": {
                        "cpu": "500m"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "20Mi"
                      }
                    },
                    "extraArgs": [],
                    "config": {
                      "destinations": {
                        "in-zone": [
                          "127.0.0.0/8",
                          "169.254.0.0/16",
                          "10.0.0.0/8",
                          "172.16.0.0/12",
                          "192.168.0.0/16"
                        ],
                        "in-region": [],
                        "cross-region": [],
                        "internet": [],
                        "direct-classification": []
                      },
                      "services": {
                        "google-cloud-services": false,
                        "amazon-web-services": false,
                        "azure-cloud-services": false
                      }
                    },
                    "tolerations": [],
                    "affinity": {},
                    "service": {
                      "annotations": {},
                      "labels": {}
                    },
                    "priorityClassName": "",
                    "podMonitor": {
                      "enabled": false,
                      "additionalLabels": {}
                    },
                    "additionalLabels": {},
                    "nodeSelector": {},
                    "annotations": {},
                    "healthCheckProbes": {},
                    "additionalSecurityContext": {}
                  },
                  "kubecostDeployment": {
                    "replicas": 1,
                    "labels": {},
                    "annotations": {}
                  },
                  "forecasting": {
                    "enabled": true,
                    "fullImageName": "gcr.io/kubecost1/kubecost-modeling:v0.1.6",
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "limits": {
                        "cpu": "1500m",
                        "memory": "1Gi"
                      }
                    },
                    "env": {
                      "GUNICORN_CMD_ARGS": "--log-level info -t 1200"
                    },
                    "priority": {
                      "enabled": false,
                      "name": ""
                    },
                    "nodeSelector": {},
                    "tolerations": {},
                    "affinity": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    }
                  },
                  "kubecostAggregator": {
                    "deployMethod": "singlepod",
                    "enabled": false,
                    "replicas": 1,
                    "stagingEmptyDirSizeLimit": "2Gi",
                    "numDBCopyPartitions": 25,
                    "env": {
                      "LOG_LEVEL": "info",
                      "DB_READ_THREADS": "1",
                      "DB_WRITE_THREADS": "1",
                      "DB_CONCURRENT_INGESTION_COUNT": "3"
                    },
                    "persistentConfigsStorage": {
                      "storageClass": "",
                      "storageRequest": "1Gi"
                    },
                    "aggregatorDbStorage": {
                      "storageClass": "",
                      "storageRequest": "128Gi"
                    },
                    "resources": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "cloudCost": {
                      "enabled": false,
                      "resources": {},
                      "readinessProbe": {
                        "enabled": true,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 10,
                        "failureThreshold": 200
                      }
                    },
                    "jaeger": {
                      "enabled": false,
                      "image": "jaegertracing/all-in-one",
                      "imageVersion": "latest"
                    }
                  },
                  "diagnostics": {
                    "enabled": true,
                    "primary": {
                      "enabled": false,
                      "retention": "7d",
                      "readonly": false
                    },
                    "pollingInterval": "300s",
                    "keepDiagnosticHistory": false,
                    "collectHelmValues": false,
                    "deployment": {
                      "enabled": false,
                      "resources": {
                        "requests": {
                          "cpu": "10m",
                          "memory": "20Mi"
                        }
                      },
                      "env": {},
                      "labels": {},
                      "securityContext": {},
                      "containerSecurityContext": {},
                      "nodeSelector": {},
                      "tolerations": {},
                      "affinity": {}
                    }
                  },
                  "clusterController": {
                    "enabled": false,
                    "image": {
                      "repository": "gcr.io/kubecost1/cluster-controller",
                      "tag": "v0.16.0"
                    },
                    "imagePullPolicy": "Always",
                    "priorityClassName": "",
                    "tolerations": [],
                    "actionConfigs": {
                      "clusterTurndown": [],
                      "namespaceTurndown": null,
                      "clusterRightsize": null,
                      "containerRightsize": null
                    },
                    "kubescaler": {
                      "defaultResizeAll": false
                    },
                    "namespaceTurndown": {
                      "rbac": {
                        "enabled": true
                      }
                    }
                  },
                  "reporting": {
                    "logCollection": true,
                    "productAnalytics": true,
                    "errorReporting": true,
                    "valuesReporting": true
                  },
                  "serviceMonitor": {
                    "enabled": false,
                    "additionalLabels": {},
                    "metricRelabelings": [],
                    "relabelings": [],
                    "networkCosts": {
                      "enabled": false,
                      "scrapeTimeout": "10s",
                      "additionalLabels": {},
                      "metricRelabelings": [],
                      "relabelings": []
                    }
                  },
                  "prometheusRule": {
                    "enabled": false,
                    "additionalLabels": {}
                  },
                  "supportNFS": false,
                  "initChownDataImage": "busybox",
                  "initChownData": {
                    "resources": {}
                  },
                  "grafana": {
                    "rbac": {
                      "create": true
                    },
                    "serviceAccount": {
                      "create": true,
                      "name": ""
                    },
                    "replicas": 1,
                    "deploymentStrategy": "RollingUpdate",
                    "readinessProbe": {
                      "httpGet": {
                        "path": "/api/health",
                        "port": 3000
                      }
                    },
                    "livenessProbe": {
                      "httpGet": {
                        "path": "/api/health",
                        "port": 3000
                      },
                      "initialDelaySeconds": 60,
                      "timeoutSeconds": 30,
                      "failureThreshold": 10
                    },
                    "image": {
                      "repository": "grafana/grafana",
                      "tag": "10.3.4",
                      "pullPolicy": "IfNotPresent"
                    },
                    "securityContext": {},
                    "priorityClassName": "",
                    "downloadDashboardsImage": {
                      "repository": "curlimages/curl",
                      "tag": "latest",
                      "pullPolicy": "IfNotPresent"
                    },
                    "podAnnotations": {},
                    "annotations": {},
                    "service": {
                      "type": "ClusterIP",
                      "port": 80,
                      "annotations": {},
                      "labels": {}
                    },
                    "ingress": {
                      "enabled": false,
                      "annotations": {},
                      "labels": {},
                      "path": "/",
                      "pathType": "Prefix",
                      "hosts": [
                        "chart-example.local"
                      ],
                      "tls": []
                    },
                    "resources": {},
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "persistence": {
                      "enabled": false
                    },
                    "adminUser": "admin",
                    "adminPassword": "strongpassword",
                    "env": {},
                    "envFromSecret": "",
                    "extraSecretMounts": [],
                    "plugins": [],
                    "dashboardProviders": {},
                    "dashboards": {},
                    "dashboardsConfigMaps": {},
                    "ldap": {
                      "existingSecret": "",
                      "config": ""
                    },
                    "smtp": {
                      "existingSecret": ""
                    },
                    "sidecar": {
                      "image": {
                        "repository": "kiwigrid/k8s-sidecar",
                        "tag": "1.26.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "resources": {},
                      "dashboards": {
                        "enabled": true,
                        "label": "grafana_dashboard",
                        "labelValue": "1",
                        "annotations": {},
                        "error_throttle_sleep": 0,
                        "folder": "/tmp/dashboards"
                      },
                      "datasources": {
                        "enabled": false,
                        "error_throttle_sleep": 0,
                        "label": "grafana_datasource"
                      }
                    },
                    "grafana.ini": {
                      "server": {
                        "serve_from_sub_path": false,
                        "root_url": "%(protocol)s://%(domain)s:%(http_port)s/grafana"
                      },
                      "paths": {
                        "data": "/var/lib/grafana/data",
                        "logs": "/var/log/grafana",
                        "plugins": "/var/lib/grafana/plugins",
                        "provisioning": "/etc/grafana/provisioning"
                      },
                      "analytics": {
                        "check_for_updates": true
                      },
                      "log": {
                        "mode": "console"
                      },
                      "grafana_net": {
                        "url": "https://grafana.net"
                      },
                      "auth.anonymous": {
                        "enabled": true,
                        "org_role": "Editor",
                        "org_name": "Main Org."
                      }
                    }
                  },
                  "serviceAccount": {
                    "create": true,
                    "annotations": {}
                  },
                  "awsstore": {
                    "useAwsStore": false,
                    "imageNameAndVersion": "gcr.io/kubecost1/awsstore:latest",
                    "createServiceAccount": false,
                    "priorityClassName": "",
                    "nodeSelector": {},
                    "annotations": {}
                  },
                  "federatedETL": {
                    "agentOnly": false,
                    "federatedCluster": false,
                    "redirectS3Backup": false,
                    "useMultiClusterDB": false
                  },
                  "kubecostAdmissionController": {
                    "enabled": false,
                    "secretName": "webhook-server-tls",
                    "caBundle": "${CA_BUNDLE}"
                  },
                  "costEventsAudit": {
                    "enabled": false
                  },
                  "extraObjects": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: gcr.io/kubecost1/operator/kubecost-operator-controller:v2.2.4
          createdAt: "2024-05-19T20:46:35Z"
          description: An operator that deploys Kubecost 2.2.4 via the Helm chart.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Stackwatch
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The CostAnalyzer API describes a deployment of Kubecost and
              its dependencies.
            displayName: CostAnalyzer
            kind: CostAnalyzer
            name: costanalyzers.charts.kubecost.com
            version: v1alpha1
        description: |-
          Kubecost provides real-time cost visibility and insights for teams using Kubernetes, helping you continuously reduce your cloud costs.

          This OpenShift operator deploys Kubecost from the official Helm chart.

          Kubecost allows you to:

          * See allocated spend across all native Kubernetes concepts, so you can provide your teams with transparent, accurate cost data reconciled with your actual cloud bill.
          * Join in-cluster costs like CPU and memory with out-of-cluster spend from cloud infrastructure services for complete cost visibility—all services from AWS, GCP, and Microsoft Azure are supported.
          * Gain insights you can use to save 30-50% or more on your infrastructure spend, without exposing your private information—your data never gets shared externally, even with us.
          * Have real-time alerts and recurring reports, empowering teams to take control of their Kubernetes-enabled infrastructure, stay within budgeted limits, and address monitoring interruptions immediately.

          Visit the Kubecost website for more details including features, pricing, documentation, and more at www.kubecost.com.
        displayName: Kubecost
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubecost
        - cost
        - kubernetes
        links:
        - name: Kubecost
          url: https://www.kubecost.com
        maintainers:
        - email: czoller@kubecost.com
          name: Chip Zoller
        maturity: alpha
        provider:
          name: Stackwatch
          url: www.kubecost.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - gcr.io/kubecost1/operator/kubecost-operator-controller:v2.2.4
        version: 2.2.4
      entries:
      - name: kubecost.v2.2.4
        version: 2.2.4
      - name: kubecost.v2.1.0
        version: 2.1.0
      - name: kubecost-operator.v1.107.1
        version: 1.107.1
      - name: kubecost-operator.v1.106.2
        version: 1.106.2
      - name: kubecost-operator.v1.106.0
        version: 1.106.0
      - name: kubecost.v1.99.0
        version: 1.99.0
      - name: kubecost.v1.97.0
        version: 1.97.0
      name: alpha
    defaultChannel: alpha
    packageName: kubecost
    provider:
      name: Stackwatch
      url: www.kubecost.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: prometurbo
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometurbo-operator.v8.7.1-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.7.1-SNAPSHOT
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.7.1-SNAPSHOT
        version: 8.7.1-beta.1
      entries:
      - name: prometurbo-operator.v8.7.1-beta.1
        version: 8.7.1-beta.1
      name: beta
    - currentCSV: prometurbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.6
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.6
        version: 8.6.0
      entries:
      - name: prometurbo-operator.v8.6.0
        version: 8.6.0
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: stable
    - currentCSV: prometurbo-operator.v8.5.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.5
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.5
        version: 8.5.0
      entries:
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: v8.5.0
    - currentCSV: prometurbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.6
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.6
        version: 8.6.0
      entries:
      - name: prometurbo-operator.v8.6.0
        version: 8.6.0
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: v8.6.0
    defaultChannel: stable
    packageName: prometurbo
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apache KIE
      provider-url: ""
    name: sonataflow-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sonataflow-operator.v10.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "docker.io/apache",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: docker.io/apache/incubator-kie-sonataflow-operator:10.0.0
          description: SonataFlow Kubernetes Operator for deploying workflow applications
            based on the CNCF Serverless Workflow specification
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/apache/incubator-kie-tools/packages/sonataflow-operator
          support: Apache Software Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: Sonata Flow Build
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: Sonata Flow Cluster Platform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: Sonata Flow Platform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: Sonata Flow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          SonataFlow Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification](https://serverlessworkflow.io/):

          * Deploy workflow applications using the [dev profile](https://sonataflow.org/serverlessworkflow/latest/cloud/operator/developing-workflows.html), suited for the your development cycle
          * Build workflow applications based on the platform you're currently working on.
        displayName: SonataFlow Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        links:
        - name: Product Page
          url: https://github.com/apache/incubator-kie-tools
        maintainers:
        - email: dev@kie.apache.org
          name: Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Apache Software Foundation
        relatedImages:
        - docker.io/apache/incubator-kie-sonataflow-operator:10.0.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 10.0.0
      entries:
      - name: sonataflow-operator.v10.0.0
        version: 10.0.0
      - name: sonataflow-operator.v1.40.1
        version: 1.40.1
      - name: sonataflow-operator.v1.40.0
        version: 1.40.0
      name: alpha
    - currentCSV: sonataflow-operator.v10.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "host/namespace",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: docker.io/apache/incubator-kie-sonataflow-operator:10.1.0
          createdAt: "2025-09-11T14:44:56Z"
          description: SonataFlow Kubernetes Operator for deploying workflow applications
            based on the CNCF Serverless Workflow specification
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/apache/incubator-kie-tools/packages/sonataflow-operator
          support: Apache KIE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: SonataFlowBuild
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: SonataFlowClusterPlatform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: SonataFlowPlatform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: SonataFlow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          SonataFlow Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification](https://serverlessworkflow.io/):

          * Deploy workflow applications using the [dev profile](https://sonataflow.org/serverlessworkflow/latest/cloud/operator/developing-workflows.html), suited for the your development cycle
          * Build workflow applications based on the platform you're currently working on.
        displayName: SonataFlow Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        links:
        - name: Product Page
          url: https://sonataflow.org/serverlessworkflow/latest/index.html
        maintainers:
        - email: dev@kie.apache.org
          name: Apache KIE
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Apache KIE
        relatedImages:
        - docker.io/apache/incubator-kie-sonataflow-operator:10.1.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        version: 10.1.0
      entries:
      - name: sonataflow-operator.v10.1.0
        version: 10.1.0
      - name: sonataflow-operator.v10.0.0
        version: 10.0.0
      - name: sonataflow-operator.v1.40.1
        version: 1.40.1
      - name: sonataflow-operator.v1.40.0
        version: 1.40.0
      name: stable
    defaultChannel: stable
    packageName: sonataflow-operator
    provider:
      name: Apache KIE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: directpv
      provider-url: ""
    name: private-dpv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: private-dpv-operator.v5.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.quay.io/v1alpha1",
                "kind": "AistorVolumemanager",
                "metadata": {
                  "name": "aistorvolumemanager-sample"
                },
                "spec": {
                  "annotations": {},
                  "controller": {
                    "podAnnotations": {},
                    "podLabels": {},
                    "replicas": 3
                  },
                  "images": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": []
                  },
                  "labels": {},
                  "license": "",
                  "nodeserver": {
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-10-05T15:47:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AistorVolumemanager
            name: aistorvolumemanagers.charts.quay.io
            version: v1alpha1
        description: directpv
        displayName: directpv
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - directpv
        links:
        - name: Private Dpv Operator
          url: https://private-dpv-operator.domain
        maintainers:
        - email: cesarcelis@minio.io
          name: celis
        maturity: alpha
        provider:
          name: directpv
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - quay.io/cniackz4/directpv-operator@sha256:af06e57661a7e05809e44c3fac571ed949cc04fd656fdb0c8324b533778eb604
        version: 5.0.3
      entries:
      - name: private-dpv-operator.v5.0.3
        version: 5.0.3
      - name: private-dpv-operator.v5.0.2
        version: 5.0.2
      name: alpha
    defaultChannel: alpha
    packageName: private-dpv-operator
    provider:
      name: directpv
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fortanix Inc.
      provider-url: https://fortanix.com
    name: ccm-node-agent-dcap-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ccm-node-agent-dcap-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1",
                "kind": "CcmNodeAgent",
                "metadata": {
                  "name": "ccmnodeagent-sample"
                },
                "spec": {
                  "ccmJoinToken": {
                    "secretName": "ccm-join-token",
                    "token": ""
                  },
                  "fullnameOverride": "",
                  "image": {
                    "aesmd": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/aesmd-dcap",
                      "tag": "latest"
                    },
                    "ccmNodeAgent": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/ccm-node-agent-ubi7",
                      "tag": "latest"
                    },
                    "ccmNodeAgentInit": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "busybox",
                      "tag": "latest"
                    },
                    "multiPlatformRegistration": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/multi-platform-registration-ubi8",
                      "tag": "latest"
                    }
                  },
                  "log": {
                    "debug": false
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "ports": {
                    "containerPort": 9092,
                    "hostPort": 9092
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CCM Node Agent software enables registration of the compute
              nodes to Fortanix CCM when installed in the Openshift Cluster Nodes<U+200B>.
            displayName: CCM Node Agent
            kind: CcmNodeAgent
            name: ccmnodeagents.charts.operatorhub.io
            version: v1
        description: "CCM Node Agent software enables registration of the compute
          nodes to Fortanix CCM when installed in the Openshift Cluster Nodes​. \nIt
          also assists with application attestation and visibility for Fortanix CCM.\n\n##
          About this Operator\nThis Operator is based on a Helm chart for CCM Node
          Agent. It enrolls all the nodes in the openshift cluster to a particular
          CCM account. You have to provide the base64 encode of join token of CCM
          account while creating CcmNodeAgent resource.\n\n## Prerequisites\nAll the
          worker nodes in the cluster must have SGX capabilities, in-kernel driver
          installed and fsgsbase support.\n\n## Installation\n * Choose a namespace
          to install the Operator. \n Operator will be available in all namespaces.\n*
          Select Approval Strategy: Available options are Automatic or Manual. Automatic
          approval will automatically upgrade the running instance of your Operator
          without human intervention when a new version is available. In case of Manual
          approval, when a new version is available, OLM will create an update request.
          You must manually approve the request to update the Operator to a new version.\n*
          Generate your CCM Account Join Token: Please log in to https://ccm.fortanix.com,
          and in the Management Console tab, click the + ENROLL NODE button. In the
          ENROLL NODE screen, a Join Token will be generated in the text box for \"Get
          a join token to register an SGX compute node\". This Join Token is used
          by the compute node to authenticate itself.\n* Base64 encode the Join Token:\n```shell\necho
          -n <join-token> | openssl base64 -A\n```\n* Create an instance of CcmNodeAgent
          by setting the `spec.ccmJoinToken.Token` to the base64 encoded CCM account
          join token.\n* After CcmNodeAgent resource is installed successfully you
          should be able to see all the worker nodes of the cluster in your CCM account
          under Compute Node Tab.\n\n## Getting Started\nNow to run your application
          inside SGX enclaves on your cluster follow the [documentation](https://support.fortanix.com/hc/en-us/articles/360043527431-User-s-Guide-Add-and-Edit-an-Application).\n"
        displayName: CCM Node Agent DCAP Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dockerfile
        - Helm
        - CCM
        - Confidential Compute
        links:
        - name: CCM Node Agent DCAP Operator
          url: https://support.fortanix.com/hc/en-us/categories/360003107511-Confidential-Computing-Manager
        maintainers:
        - email: info@fortanix.com
          name: Fortanix Inc.
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Fortanix Inc.
          url: https://fortanix.com
        relatedImages:
        - docker.io/fortanix/ccm-node-agent-dcap-operator@sha256:4fd28bea26aee461eb9c1800cdf5f041b64ae3ea6652203f8ad9feb680aaefee
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62575404000cb77fa8d2284c87b31e39e667a55b0fb47821263b23b3d1135a7e
        version: 0.0.2
      entries:
      - name: ccm-node-agent-dcap-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ccm-node-agent-dcap-operator
    provider:
      name: Fortanix Inc.
      url: https://fortanix.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Infinidat
      provider-url: https://infinidat.com
    name: infinibox-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infinibox-operator-certified.v2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csidriver.infinidat.com/v1alpha1",
                "kind": "InfiniboxCsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infinibox-operator-certified",
                    "app.kubernetes.io/instance": "infiniboxcsidriver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infiniboxcsidriver",
                    "app.kubernetes.io/part-of": "infinibox-operator-certified"
                  },
                  "name": "infiniboxcsidriver-sample"
                },
                "spec": {
                  "Ibox_Creds": [
                    {
                      "SecretName": "infinibox-creds",
                      "hostname": "ibox0000",
                      "inbound_secret": "0.0000000000000",
                      "inbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonout",
                      "outbound_secret": "0.0000000000001",
                      "outbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonin",
                      "password": "my-pool-admin-passwd",
                      "username": "my-pool-admin"
                    }
                  ],
                  "Infinibox_Protocol_Secrets": [
                    {
                      "SecretName": "infinibox-protocol-secret",
                      "iscsi": {
                        "network_space": "default_iscsi_space",
                        "useCHAP": "none"
                      },
                      "nfs": {
                        "network_space": "default_nas_space",
                        "nfs_export_permissions": ""
                      },
                      "nvme": {
                        "network_space": "default_nvme_space"
                      },
                      "preferred_auto_order": "fc,nvme,iscsi",
                      "storage_protocol": "fc"
                    }
                  ],
                  "autoUpdate": false,
                  "cleanupNFSPerms": false,
                  "controllerserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "controllerservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "createEvents": true,
                  "csiDriverName": "infinibox-csi-driver",
                  "csiDriverVersion": "v2.23.0",
                  "e2etesting": false,
                  "images": {
                    "attachersidecar": "registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901",
                    "attachersidecar_pull_policy": "IfNotPresent",
                    "csidriver": "registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61",
                    "csidriver_pull_policy": "Always",
                    "livenesssidecar": "registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc",
                    "livenesssidecar_pull_policy": "IfNotPresent",
                    "provisionersidecar": "registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2",
                    "provisionersidecar_pull_policy": "IfNotPresent",
                    "registrarsidecar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24",
                    "registrarsidecar_pull_policy": "IfNotPresent",
                    "resizersidecar": "registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d",
                    "resizersidecar_pull_policy": "IfNotPresent",
                    "snapshottersidecar": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a",
                    "snapshottersidecar_pull_policy": "IfNotPresent"
                  },
                  "instanceCount": 1,
                  "logLevel": "debug",
                  "nodeSelector": {
                    "kubernetes.io/os": "linux"
                  },
                  "nodeserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "nodeservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "protocolSecret": "",
                  "removeDomainName": false,
                  "replicaCount": 1,
                  "skipCredentialsCreation": false,
                  "volumeNamePrefix": "ibox"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
          createdAt: "2025-09-27T16:12:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2.23.0
          operatorframework.io/suggested-namespace: infinidat-csi
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Infinidat/infinibox-csi-driver
          support: Infinidat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Iboxreplica
            name: iboxreplicas.csidriver.infinidat.com
            version: v1
          - kind: InfiniboxCsiDriver
            name: infiniboxcsidrivers.csidriver.infinidat.com
            version: v1alpha1
        description: "Infinidat InfiniBox Container Storage Interface (CSI) Driver
          is a\n    CNCF-compliant Kubernetes integration for InfiniBox storage systems,
          offering\n    advanced enterprise functionality for petabyte-scale Kubernetes
          deployments including\n    Red Hat OpenShift.\n \n## Features and Benefits\n
          \ \n* **Multi-protocol flexibility** - manage\n    Kubernetes Persistent
          Volumes attached via block and file protocols,\n    including Fibre Channel,
          iSCSI, and NFS, with all Kubernetes PV access modes\n    \n* **Multi-petabyte
          scalability** - support hundreds of thousands of PVs per InfiniBox\n    system
          and control multiple InfiniBox arrays within a single Kubernetes cluster\n
          \   \\ \n* **Advanced enterprise features** - manage native InfiniBox snapshots
          and\n    clones, including restoring from snapshots, and import PVs created
          outside of\n    InfiniBox CSI Driver\n  \n## Required Parameters\n  \n*
          `hostname` - IP address\n    or hostname of the InfiniBox management interface\n*
          `username` / `password` -\n    InfiniBox credentials\n* `SecretName` - secret
          name, to be used in the StorageClass\n    to define a specific InfiniBox
          for persistent volumes\n\n## Optional Parameters\n*\n    `inbound_user`
          / `inbound_secret` / `outbound_user` / `outbound_secret` - credentials \n
          \   for iSCSI CHAP authentication\n\n    \n## Installation Instructions\n
          \n1. Create *infinidat-csi* namespace to install operator into.\n    \n2.
          Delete the following *clusterrolebindings*, if they exist, before installing
          operator\n     * infinidat-csi-operator-infinidat-csi-attacher\n     * infinidat-csi-operator-infinidat-csi-controller
          \                    \n     * infinidat-csi-operator-infinidat-csi-driver
          \                    \n     * infinidat-csi-operator-infinidat-csi-node
          \                              \n     * infinidat-csi-operator-infinidat-csi-provisioner
          \                       \n     * infinidat-csi-operator-infinidat-csi-resizer
          \          \n     * infinidat-csi-operator-infinidat-csi-snapshotter \n
          \   \n3. Install Infinidat Infinibox CSI Driver - Operator (this) into the
          *infinidat-csi* namespace.\n    \n4. Apply the following Security Context
          Constraints to the created service accounts in the *infinidat-csi* \n    namespace\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    \n5. Recreate or edit any existing storage classes, pointing to the
          new namespace\n\n     \n6. If you choose a different namespace other than
          *infinidat-csi* you will need to edit the clusterrolebindings listed above
          to specify your custom namespace\n\n     \n7. Install an instance of Infinidat
          Driver into the *infinidat-csi* namespace"
        displayName: InfiniBox CSI Driver - Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - csi
        - nas
        - san
        - storage
        - infinidat
        - infinibox
        links:
        - name: Infinidat
          url: https://infinidat.com
        maintainers:
        - email: info@infinidat.com
          name: Infinidat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Infinidat
          url: https://infinidat.com
        relatedImages:
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified@sha256:42ff34a019bc67076515e89735b4bb9921fbd4610c127ac8b61e1fa6a6a08194
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        version: 2.23.0
      entries:
      - name: infinibox-operator-certified.v2.23.0
        version: 2.23.0
      - name: infinibox-operator-certified.v2.22.0
        version: 2.22.0
      - name: infinibox-operator-certified.v2.21.0
        version: 2.21.0
      name: alpha
    - currentCSV: infinibox-operator-certified.v2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csidriver.infinidat.com/v1alpha1",
                "kind": "InfiniboxCsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infinibox-operator-certified",
                    "app.kubernetes.io/instance": "infiniboxcsidriver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infiniboxcsidriver",
                    "app.kubernetes.io/part-of": "infinibox-operator-certified"
                  },
                  "name": "infiniboxcsidriver-sample"
                },
                "spec": {
                  "Ibox_Creds": [
                    {
                      "SecretName": "infinibox-creds",
                      "hostname": "ibox0000",
                      "inbound_secret": "0.0000000000000",
                      "inbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonout",
                      "outbound_secret": "0.0000000000001",
                      "outbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonin",
                      "password": "my-pool-admin-passwd",
                      "username": "my-pool-admin"
                    }
                  ],
                  "Infinibox_Protocol_Secrets": [
                    {
                      "SecretName": "infinibox-protocol-secret",
                      "iscsi": {
                        "network_space": "default_iscsi_space",
                        "useCHAP": "none"
                      },
                      "nfs": {
                        "network_space": "default_nas_space",
                        "nfs_export_permissions": ""
                      },
                      "nvme": {
                        "network_space": "default_nvme_space"
                      },
                      "preferred_auto_order": "fc,nvme,iscsi",
                      "storage_protocol": "fc"
                    }
                  ],
                  "autoUpdate": false,
                  "cleanupNFSPerms": false,
                  "controllerserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "controllerservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "createEvents": true,
                  "csiDriverName": "infinibox-csi-driver",
                  "csiDriverVersion": "v2.23.0",
                  "e2etesting": false,
                  "images": {
                    "attachersidecar": "registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901",
                    "attachersidecar_pull_policy": "IfNotPresent",
                    "csidriver": "registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61",
                    "csidriver_pull_policy": "Always",
                    "livenesssidecar": "registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc",
                    "livenesssidecar_pull_policy": "IfNotPresent",
                    "provisionersidecar": "registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2",
                    "provisionersidecar_pull_policy": "IfNotPresent",
                    "registrarsidecar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24",
                    "registrarsidecar_pull_policy": "IfNotPresent",
                    "resizersidecar": "registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d",
                    "resizersidecar_pull_policy": "IfNotPresent",
                    "snapshottersidecar": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a",
                    "snapshottersidecar_pull_policy": "IfNotPresent"
                  },
                  "instanceCount": 1,
                  "logLevel": "debug",
                  "nodeSelector": {
                    "kubernetes.io/os": "linux"
                  },
                  "nodeserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "nodeservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "protocolSecret": "",
                  "removeDomainName": false,
                  "replicaCount": 1,
                  "skipCredentialsCreation": false,
                  "volumeNamePrefix": "ibox"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
          createdAt: "2025-09-27T16:12:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2.23.0
          operatorframework.io/suggested-namespace: infinidat-csi
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Infinidat/infinibox-csi-driver
          support: Infinidat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Iboxreplica
            name: iboxreplicas.csidriver.infinidat.com
            version: v1
          - kind: InfiniboxCsiDriver
            name: infiniboxcsidrivers.csidriver.infinidat.com
            version: v1alpha1
        description: "Infinidat InfiniBox Container Storage Interface (CSI) Driver
          is a\n    CNCF-compliant Kubernetes integration for InfiniBox storage systems,
          offering\n    advanced enterprise functionality for petabyte-scale Kubernetes
          deployments including\n    Red Hat OpenShift.\n \n## Features and Benefits\n
          \ \n* **Multi-protocol flexibility** - manage\n    Kubernetes Persistent
          Volumes attached via block and file protocols,\n    including Fibre Channel,
          iSCSI, and NFS, with all Kubernetes PV access modes\n    \n* **Multi-petabyte
          scalability** - support hundreds of thousands of PVs per InfiniBox\n    system
          and control multiple InfiniBox arrays within a single Kubernetes cluster\n
          \   \\ \n* **Advanced enterprise features** - manage native InfiniBox snapshots
          and\n    clones, including restoring from snapshots, and import PVs created
          outside of\n    InfiniBox CSI Driver\n  \n## Required Parameters\n  \n*
          `hostname` - IP address\n    or hostname of the InfiniBox management interface\n*
          `username` / `password` -\n    InfiniBox credentials\n* `SecretName` - secret
          name, to be used in the StorageClass\n    to define a specific InfiniBox
          for persistent volumes\n\n## Optional Parameters\n*\n    `inbound_user`
          / `inbound_secret` / `outbound_user` / `outbound_secret` - credentials \n
          \   for iSCSI CHAP authentication\n\n    \n## Installation Instructions\n
          \n1. Create *infinidat-csi* namespace to install operator into.\n    \n2.
          Delete the following *clusterrolebindings*, if they exist, before installing
          operator\n     * infinidat-csi-operator-infinidat-csi-attacher\n     * infinidat-csi-operator-infinidat-csi-controller
          \                    \n     * infinidat-csi-operator-infinidat-csi-driver
          \                    \n     * infinidat-csi-operator-infinidat-csi-node
          \                              \n     * infinidat-csi-operator-infinidat-csi-provisioner
          \                       \n     * infinidat-csi-operator-infinidat-csi-resizer
          \          \n     * infinidat-csi-operator-infinidat-csi-snapshotter \n
          \   \n3. Install Infinidat Infinibox CSI Driver - Operator (this) into the
          *infinidat-csi* namespace.\n    \n4. Apply the following Security Context
          Constraints to the created service accounts in the *infinidat-csi* \n    namespace\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    \n5. Recreate or edit any existing storage classes, pointing to the
          new namespace\n\n     \n6. If you choose a different namespace other than
          *infinidat-csi* you will need to edit the clusterrolebindings listed above
          to specify your custom namespace\n\n     \n7. Install an instance of Infinidat
          Driver into the *infinidat-csi* namespace"
        displayName: InfiniBox CSI Driver - Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - csi
        - nas
        - san
        - storage
        - infinidat
        - infinibox
        links:
        - name: Infinidat
          url: https://infinidat.com
        maintainers:
        - email: info@infinidat.com
          name: Infinidat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Infinidat
          url: https://infinidat.com
        relatedImages:
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified@sha256:42ff34a019bc67076515e89735b4bb9921fbd4610c127ac8b61e1fa6a6a08194
        version: 2.23.0
      entries:
      - name: infinibox-operator-certified.v2.23.0
        version: 2.23.0
      - name: infinibox-operator-certified.v2.22.0
        version: 2.22.0
      - name: infinibox-operator-certified.v2.21.0
        version: 2.21.0
      name: stable
    defaultChannel: stable
    packageName: infinibox-operator-certified
    provider:
      name: Infinidat
      url: https://infinidat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Shipwright Contributors
      provider-url: https://shipwright.io
    name: shipwright-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: shipwright-operator.v0.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "shipwright-build"
                },
                "spec": {
                  "targetNamespace": "shipwright-build"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: ghcr.io/shipwright-io/operator/operator:0.13.0@sha256:12cd5f936043c96f31cf571d7a4f8b6f2e240852cd9ad1a5423dbab0c208b542
          description: Shipwright is a framework for building container images on
            Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/shipwright-io/operator
          support: The Shipwright Contributors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
        description: |
          Shipwright is a framework for building container images on Kubernetes.

          Read more: [https://shipwright.io](https://shipwright.io)

          ## Usage

          To deploy and manage [Shipwright Builds](https://github.com/shipwright-io/build) in your cluster,
          first make sure this operator is installed and running on your cluster.

          Next, create the following:

          ```yaml
          ---
          apiVersion: operator.shipwright.io/v1alpha1
          kind: ShipwrightBuild
          metadata:
            name: shipwright-operator
          spec:
            targetNamespace: shipwright-build
          ```

          The operator will deploy Shipwright Builds in the provided `targetNamespace`.
          When `.spec.targetNamespace` is not set, the namespace will default to `shipwright-build`.
          Refer to the [ShipwrightBuild documentation](https://github.com/shipwright-io/operator/blob/main/docs/shipwrightbuild.md) for more information about this custom resource.
        displayName: Shipwright Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://shipwright.io
        - name: Shipwright Builds
          url: https://github.com/shipwright-io/build
        maintainers:
        - email: shipwright-dev@lists.shipwright.io
          name: The Shipwright Contributors
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Shipwright Contributors
          url: https://shipwright.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - ghcr.io/shipwright-io/operator/operator:0.13.0@sha256:12cd5f936043c96f31cf571d7a4f8b6f2e240852cd9ad1a5423dbab0c208b542
        version: 0.13.0
      entries:
      - name: shipwright-operator.v0.13.0
        version: 0.13.0
      - name: shipwright-operator.v0.9.0
        version: 0.9.0
      name: alpha
    defaultChannel: alpha
    packageName: shipwright-operator
    provider:
      name: The Shipwright Contributors
      url: https://shipwright.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: barbican-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: barbican-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "barbican.openstack.org/v1beta1",
                "kind": "Barbican",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "barbican-operator",
                    "app.kubernetes.io/instance": "barbican",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "barbican",
                    "app.kubernetes.io/part-of": "barbican-operator"
                  },
                  "name": "barbican"
                },
                "spec": {
                  "barbicanAPI": {
                    "customServiceConfig": "[optional]\noverrides = True\n",
                    "defautlConfigOverwrite": {
                      "optional_policy.json": "{\"some\": \"custom policy\"}\n"
                    },
                    "passwordSelectors": {
                      "database": "BarbicanDatabasePassword",
                      "service": "BarbicanPassword"
                    }
                  },
                  "barbicanKeystoneListener": {
                    "customServiceConfig": "[optional]\noverrides = True\n",
                    "defautlConfigOverwrite": {
                      "optional_policy.json": "{\"some\": \"custom policy\"}\n"
                    }
                  },
                  "barbicanWorker": {
                    "customServiceConfig": "[optional]\noverrides = True\n",
                    "defautlConfigOverwrite": {
                      "optional_policy.json": "{\"some\": \"custom policy\"}\n"
                    }
                  },
                  "customServiceConfig": "[DEFAULT]\ndebug = True\n",
                  "databaseAccount": "barbican",
                  "databaseInstance": "openstack",
                  "defaultConfigOverwrite": {
                    "policy.json": "{\"some\": \"custom policy\"}\n"
                  },
                  "passwordSelectors": {
                    "database": "BarbicanDatabasePassword",
                    "service": "BarbicanPassword",
                    "simplecryptokek": "BarbicanSimpleCryptoKEK"
                  },
                  "preserveJobs": true,
                  "rabbitMqCusterName": "barbican_rabbit",
                  "secret": "osp-secret",
                  "serviceAccount": "barbican",
                  "serviceUser": "barbican"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:32Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BarbicanAPI is the Schema for the barbicanapis API
            displayName: Barbican API
            kind: BarbicanAPI
            name: barbicanapis.barbican.openstack.org
            version: v1beta1
          - kind: BarbicanKeystoneListener
            name: barbicankeystonelisteners.barbican.openstack.org
            version: v1beta1
          - description: Barbican is the Schema for the barbicans API
            displayName: Barbican
            kind: Barbican
            name: barbicans.barbican.openstack.org
            version: v1beta1
          - description: BarbicanWorker is the Schema for the barbicanworkers API
            displayName: Barbican Worker
            kind: BarbicanWorker
            name: barbicanworkers.barbican.openstack.org
            version: v1beta1
        description: Barbican Operator
        displayName: Barbican Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - KeyManager
        - Barbican
        links:
        - name: Barbican Operator
          url: https://github.com/openstack-k8s-operators/barbican-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-barbican-worker-rhel9@sha256:b5133f539658ed5d532e0c3cf63d90bc0f76c7f0368963a8ebec49123f27f7a7
        - registry.redhat.io/rhoso/openstack-barbican-keystone-listener-rhel9@sha256:ac112444c0a12664e8a7900e72cd97fa795df167c0d95437a029644f0819f30d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/barbican-rhel9-operator@sha256:1275a3f7f09999e9dc25bc38f917a3b5d780d9d309442cddd0e7adf313d6f576
        - registry.redhat.io/rhoso/openstack-barbican-api-rhel9@sha256:0db6c1ab5b4ca302e38338a3ab9cee9bd4ae7c8e1e7bd3419e47b65df9ef1f4f
        version: 1.0.6
      entries:
      - name: barbican-operator.v1.0.6
        version: 1.0.6
      - name: barbican-operator.v1.0.5
        version: 1.0.5
      - name: barbican-operator.v1.0.4
        version: 1.0.4
      - name: barbican-operator.v1.0.3
        version: 1.0.3
      - name: barbican-operator.v1.0.2
        version: 1.0.2
      - name: barbican-operator.v1.0.1
        version: 1.0.1
      - name: barbican-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: barbican-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: ansible-cloud-addons-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aca-operator.v2.4.0-0.1737660894
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Cloud Addons Operator manages everything Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.1.1-0.1645500829 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy an AAP AWS Billing instance
            displayName: AnsibleAutomationPlatformAwsBilling
            kind: AnsibleAutomationPlatformAwsBilling
            name: ansibleautomationplatformawsbillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Azure Billing instance
            displayName: AnsibleAutomationPlatformAzureBilling
            kind: AnsibleAutomationPlatformAzureBilling
            name: ansibleautomationplatformazurebillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Metrics instance
            displayName: AnsibleAutomationPlatformMetrics
            kind: AnsibleAutomationPlatformMetrics
            name: ansibleautomationplatformmetrics.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP UI instance
            displayName: AutomationPlatform
            kind: AutomationPlatform
            name: automationplatforms.platform.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Cloud Addons is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Cloud Addons Operator provides for push-button deployments of Cloud Add-ons into your OpenShift environment. Ansible Cloud Addons deployments include:
            * Cloud Billing - Cloud billing connector for Ansible Automation Platform
            * Cloud UI - Unified welcome page UI for Ansible Automation Platform

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Cloud Addons operator by following instructions presented when you click the Install button.
        displayName: Ansible Cloud Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-platform
        - ansible-automation-platform
        - aap
        - cloud
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:38ee3484ba60bb474e7231e9d0718157386716cea7ffd6da6f32ddd551a5243d
        - registry.redhat.io/ansible-automation-platform-24/aap-cloud-billing-rhel8-operator@sha256:c461b0c4781e95d2b2e31e7bbd6d6d192806e8a8dd2e8b9225b5735da93dd6a2
        - registry.redhat.io/ansible-automation-platform-24/aap-cloud-ui-rhel8-operator@sha256:06bf313d4daf080ce6f2b1ba859fefd84431c7a2a0618feae55b92c9f5fff594
        - registry.redhat.io/ansible-automation-platform-24/aap-cloud-billing-rhel8@sha256:84fa86d312663a74502516f2ba838ac80e500c7bc17dc518940984a603da2548
        - registry.redhat.io/ansible-automation-platform-24/aap-cloud-metrics-collector-rhel8@sha256:6bd27ad06d0a65c7fa1cd2419bfc30976d53a5b89698d309cabbbd65b531a1ab
        - registry.redhat.io/ansible-automation-platform-24/aap-cloud-ui-rhel8@sha256:8489c253072186db064289f5ecc0a08ed3654a73be3c3f2130be479098e59047
        version: 2.4.0+0.1737660894
      entries:
      - name: aca-operator.v2.4.0-0.1708974547
        version: 2.4.0+0.1708974547
      - name: aca-operator.v2.4.0-0.1711587287
        version: 2.4.0+0.1711587287
      - name: aca-operator.v2.4.0-0.1713463331
        version: 2.4.0+0.1713463331
      - name: aca-operator.v2.4.0-0.1715891874
        version: 2.4.0+0.1715891874
      - name: aca-operator.v2.4.0-0.1717108266
        version: 2.4.0+0.1717108266
      - name: aca-operator.v2.4.0-0.1717558193
        version: 2.4.0+0.1717558193
      - name: aca-operator.v2.4.0-0.1718149348
        version: 2.4.0+0.1718149348
      - name: aca-operator.v2.4.0-0.1720478479
        version: 2.4.0+0.1720478479
      - name: aca-operator.v2.4.0-0.1725253934
        version: 2.4.0+0.1725253934
      - name: aca-operator.v2.4.0-0.1726108417
        version: 2.4.0+0.1726108417
      - name: aca-operator.v2.4.0-0.1727255735
        version: 2.4.0+0.1727255735
      - name: aca-operator.v2.4.0-0.1730152463
        version: 2.4.0+0.1730152463
      - name: aca-operator.v2.4.0-0.1733181569
        version: 2.4.0+0.1733181569
      - name: aca-operator.v2.4.0-0.1733937635
        version: 2.4.0+0.1733937635
      - name: aca-operator.v2.4.0-0.1737660894
        version: 2.4.0+0.1737660894
      name: stable-2.4-cluster-scoped
    - currentCSV: aca-operator.v2.5.0-0.1758143647
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Cloud Addons Operator manages everything Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.1.1-0.1645500829 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy an AAP AWS Billing instance
            displayName: AnsibleAutomationPlatformAwsBilling
            kind: AnsibleAutomationPlatformAwsBilling
            name: ansibleautomationplatformawsbillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Azure Billing instance
            displayName: AnsibleAutomationPlatformAzureBilling
            kind: AnsibleAutomationPlatformAzureBilling
            name: ansibleautomationplatformazurebillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Metrics instance
            displayName: AnsibleAutomationPlatformMetrics
            kind: AnsibleAutomationPlatformMetrics
            name: ansibleautomationplatformmetrics.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP UI instance
            displayName: AutomationPlatform
            kind: AutomationPlatform
            name: automationplatforms.platform.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Cloud Addons is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Cloud Addons Operator provides for push-button deployments of Cloud Add-ons into your OpenShift environment. Ansible Cloud Addons deployments include:
            * Cloud Billing - Cloud billing connector for Ansible Automation Platform
            * Cloud UI - Unified welcome page UI for Ansible Automation Platform

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Cloud Addons operator by following instructions presented when you click the Install button.
        displayName: Ansible Cloud Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-platform
        - ansible-automation-platform
        - aap
        - cloud
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-billing-rhel8-operator@sha256:f4b29ec59498bb6c9fdb2790ffc9d8c878efa3fd9da3416f7fe9f2a253171508
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-ui-rhel8-operator@sha256:47e2c40cdcfe218d5bf7abbf6fc35340bacdc2d02dac7f81deee7876b5570c53
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-billing-rhel8@sha256:0e50f539bfeaec91f0b18409cbc1888297a26af5114252a8ca07e1b97436001d
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-metrics-collector-rhel8@sha256:c6ab25b3839f8c3134e7d5e1f0b7716ea5fd66fe8d131261ff08aabd97c3cd2d
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-ui-rhel8@sha256:869efe830cdf23f9344263f0e9b5a209cb13b84d9d7588a1744fd3ea4cd3f83b
        version: 2.5.0+0.1758143647
      entries:
      - name: aca-operator.v2.5.0-0.1727240231
        version: 2.5.0+0.1727240231
      - name: aca-operator.v2.5.0-0.1727874815
        version: 2.5.0+0.1727874815
      - name: aca-operator.v2.5.0-0.1728516369
        version: 2.5.0+0.1728516369
      - name: aca-operator.v2.5.0-0.1729737962
        version: 2.5.0+0.1729737962
      - name: aca-operator.v2.5.0-0.1731452082
        version: 2.5.0+0.1731452082
      - name: aca-operator.v2.5.0-0.1733188778
        version: 2.5.0+0.1733188778
      - name: aca-operator.v2.5.0-0.1733964232
        version: 2.5.0+0.1733964232
      - name: aca-operator.v2.5.0-0.1736892777
        version: 2.5.0+0.1736892777
      - name: aca-operator.v2.5.0-0.1737095650
        version: 2.5.0+0.1737095650
      - name: aca-operator.v2.5.0-0.1737671171
        version: 2.5.0+0.1737671171
      - name: aca-operator.v2.5.0-0.1739244024
        version: 2.5.0+0.1739244024
      - name: aca-operator.v2.5.0-0.1740088272
        version: 2.5.0+0.1740088272
      - name: aca-operator.v2.5.0-0.1740769078
        version: 2.5.0+0.1740769078
      - name: aca-operator.v2.5.0-0.1741367998
        version: 2.5.0+0.1741367998
      - name: aca-operator.v2.5.0-0.1742427652
        version: 2.5.0+0.1742427652
      - name: aca-operator.v2.5.0-0.1743655743
        version: 2.5.0+0.1743655743
      - name: aca-operator.v2.5.0-0.1746134365
        version: 2.5.0+0.1746134365
      - name: aca-operator.v2.5.0-0.1747339736
        version: 2.5.0+0.1747339736
      - name: aca-operator.v2.5.0-0.1749072064
        version: 2.5.0+0.1749072064
      - name: aca-operator.v2.5.0-0.1750895274
        version: 2.5.0+0.1750895274
      - name: aca-operator.v2.5.0-0.1753899647
        version: 2.5.0+0.1753899647
      - name: aca-operator.v2.5.0-0.1755834478
        version: 2.5.0+0.1755834478
      - name: aca-operator.v2.5.0-0.1758143647
        version: 2.5.0+0.1758143647
      name: stable-2.5-cluster-scoped
    defaultChannel: stable-2.5-cluster-scoped
    packageName: ansible-cloud-addons-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: service-telemetry-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-telemetry-operator.v1.5.1757650583
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infra.watch/v1beta1",
                "kind": "ServiceTelemetry",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "alerting": {
                    "alertmanager": {
                      "receivers": {
                        "snmpTraps": {
                          "alertOidLabel": "oid",
                          "community": "public",
                          "enabled": false,
                          "port": 162,
                          "retries": 5,
                          "target": "192.168.24.254",
                          "timeout": 1,
                          "trapDefaultOid": "1.3.6.1.4.1.50495.15.1.2.1",
                          "trapDefaultSeverity": "",
                          "trapOidPrefix": "1.3.6.1.4.1.50495.15"
                        }
                      },
                      "storage": {
                        "persistent": {
                          "pvcStorageRequest": "20G"
                        },
                        "strategy": "persistent"
                      }
                    },
                    "enabled": true
                  },
                  "backends": {
                    "events": {
                      "elasticsearch": {
                        "certificates": {
                          "caCertDuration": "70080h",
                          "endpointCertDuration": "70080h"
                        },
                        "enabled": false,
                        "forwarding": {
                          "hostUrl": "https://elasticsearch-es-http:9200",
                          "tlsSecretName": "elasticsearch-es-cert",
                          "tlsServerName": "",
                          "useBasicAuth": true,
                          "useTls": true,
                          "userSecretName": "elasticsearch-es-elastic-user"
                        },
                        "storage": {
                          "persistent": {
                            "pvcStorageRequest": "20Gi"
                          },
                          "strategy": "persistent"
                        },
                        "version": "7.16.1"
                      }
                    },
                    "metrics": {
                      "prometheus": {
                        "enabled": true,
                        "scrapeInterval": "30s",
                        "storage": {
                          "persistent": {
                            "pvcStorageRequest": "20G"
                          },
                          "retention": "24h",
                          "strategy": "persistent"
                        }
                      }
                    }
                  },
                  "clouds": [
                    {
                      "metrics": {
                        "collectors": [
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 16384,
                              "verbose": false
                            },
                            "collectorType": "collectd",
                            "debugEnabled": false,
                            "subscriptionAddress": "collectd/cloud1-telemetry"
                          },
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 16384,
                              "verbose": false
                            },
                            "collectorType": "ceilometer",
                            "debugEnabled": false,
                            "subscriptionAddress": "anycast/ceilometer/cloud1-metering.sample"
                          },
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 65535,
                              "verbose": false
                            },
                            "collectorType": "sensubility",
                            "debugEnabled": false,
                            "subscriptionAddress": "sensubility/cloud1-telemetry"
                          }
                        ]
                      },
                      "name": "cloud1"
                    }
                  ],
                  "graphing": {
                    "enabled": false,
                    "grafana": {
                      "baseImage": "registry.redhat.io/rhel8/grafana:9",
                      "dashboards": {
                        "enabled": true
                      },
                      "disableSignoutMenu": false,
                      "ingressEnabled": true
                    }
                  },
                  "highAvailability": {
                    "enabled": false
                  },
                  "observabilityStrategy": "use_redhat",
                  "transports": {
                    "qdr": {
                      "auth": "basic",
                      "certificates": {
                        "caCertDuration": "70080h",
                        "endpointCertDuration": "70080h"
                      },
                      "enabled": true,
                      "web": {
                        "enabled": false
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/stf/service-telemetry-rhel9-operator@sha256:822becd77fa8835e753c88b3e307f62e42e090bdedc6e1fbcbe9fd6110310846
          createdAt: "2025-09-12T04:17:40Z"
          description: Service Telemetry Framework. Umbrella Operator for instantiating
            the required dependencies and configuration of various components to build
            a Service Telemetry platform for telco grade monitoring.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: service-telemetry
          operators.openshift.io/valid-subscription: '["OpenStack Platform", "Cloud
            Infrastructure", "Cloud Suite"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/infrawatch/service-telemetry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of the Service Telemetry Framework
            displayName: STF Cluster
            kind: ServiceTelemetry
            name: servicetelemetrys.infra.watch
            version: v1beta1
        description: Service Telemetry Operator for monitoring clouds
        displayName: Service Telemetry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - servicetelemetry
        - monitoring
        - telemetry
        - notifications
        - telecommunications
        links:
        - name: Source Code
          url: https://github.com/infrawatch/service-telemetry-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_openstack_platform/17.1/html/service_telemetry_framework_1.5/index
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy:latest@sha256:dc24aa382e02c4ac058d4b824a21f0e90a6259d6d817d0341b2a41809b5daa78
        - registry.redhat.io/cluster-observability-operator/coo-prometheus-rhel8:0.2.0@sha256:6bbb1aebb1cfdc3310e398ec7720ca1d4a9a89cecf7406412071bfaceb0d3cb3
        - registry.redhat.io/cluster-observability-operator/coo-prometheus-alertmanager-rhel8:0.2.0@sha256:cfd18e83d64a55ac310937eefe83fc5b9e98660962a127cd4f8d9a382764c171
        - registry.redhat.io/stf/service-telemetry-rhel9-operator@sha256:822becd77fa8835e753c88b3e307f62e42e090bdedc6e1fbcbe9fd6110310846
        - registry.redhat.io/stf/prometheus-webhook-snmp-rhel9@sha256:03199f8a3b6b22c82e01ede3adab6e98b276369d642029ae03641aac3b25581a
        version: 1.5.1757650583
      entries:
      - name: service-telemetry-operator.v1.5.1757650583
        version: 1.5.1757650583
      - name: service-telemetry-operator.v1.5.1743671601
        version: 1.5.1743671601
      - name: service-telemetry-operator.v1.5.1740051851
        version: 1.5.1740051851
      - name: service-telemetry-operator.v1.5.1731336911
        version: 1.5.1731336911
      - name: service-telemetry-operator.v1.5.1725952694
        version: 1.5.1725952694
      name: stable-1.5
    defaultChannel: stable-1.5
    packageName: service-telemetry-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: node-healthcheck-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-healthcheck-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "maxUnhealthy": "49%",
                  "remediationTemplate": {
                    "apiVersion": "poison-pill.medik8s.io/v1alpha1",
                    "kind": "PoisonPillRemediationTemplate",
                    "name": "group-x",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-healthcheck-operator:0.1.0-latest
          createdAt: "2021-04-20 12:00:00"
          description: An operator to monitor node healthcheck and remdiate via 3rd
            party providers like poison-pill.
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: This operator deploys the Node Health Check controller and its
          artifacts
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://node-healthcheck-operator.domain
        maintainers:
        - email: medik8s@googlegroups.com
          name: medik8s team
        maturity: alpha
        provider:
          name: medik8s
          url: https://github.com/medik8s
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/node-healthcheck-operator:0.1.0-latest
        version: 0.1.1
      entries:
      - name: node-healthcheck-operator.v0.1.1
        version: 0.1.1
      name: alpha
    - currentCSV: node-healthcheck-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-healthcheck-operator:v0.3.0
          createdAt: "2022-08-24 11:40:38"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.2.0 <0.3.0'
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: medik8s team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: medik8s
          url: https://github.com/medik8s
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/node-healthcheck-operator:v0.3.0
        version: 0.3.0
      entries:
      - name: node-healthcheck-operator.v0.3.0
        version: 0.3.0
      name: candidate
    - currentCSV: node-healthcheck-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: quay.io/medik8s/node-healthcheck-operator:v0.6.1
          createdAt: "2023-11-14T15:45:34Z"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.1.0 <0.6.1'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/medik8s/node-healthcheck-operator:v0.6.1
        - quay.io/medik8s/node-remediation-console:v0.6.1
        version: 0.6.1
      entries:
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      - name: node-healthcheck-operator.v0.6.0
        version: 0.6.0
      - name: node-healthcheck-operator.v0.5.0
        version: 0.5.0
      - name: node-healthcheck-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: node-healthcheck-operator
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: GitLab Inc
      provider-url: https://about.gitlab.com/
    name: gitlab-operator-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "gitlab": {
                        "gitlab-shell": {
                          "maxReplicas": 2,
                          "minReplicas": 1
                        },
                        "sidekiq": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1000M"
                            }
                          }
                        },
                        "webservice": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1500M"
                            }
                          }
                        }
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "installCertmanager": false,
                      "minio": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      },
                      "nginx-ingress": {
                        "controller": {
                          "electionID": "ingress-controller-leader",
                          "replicaCount": 1
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      },
                      "redis": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-29T12:42:23Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      name: stable
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "gitlab": {
                        "gitlab-shell": {
                          "maxReplicas": 2,
                          "minReplicas": 1
                        },
                        "sidekiq": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1000M"
                            }
                          }
                        },
                        "webservice": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1500M"
                            }
                          }
                        }
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "installCertmanager": false,
                      "minio": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      },
                      "nginx-ingress": {
                        "controller": {
                          "electionID": "ingress-controller-leader",
                          "replicaCount": 1
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      },
                      "redis": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-29T12:42:23Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      name: unstable
    defaultChannel: stable
    packageName: gitlab-operator-kubernetes
    provider:
      name: GitLab Inc
      url: https://about.gitlab.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: edb-hcp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: edb-hcp-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edbpgai.edb.com/v1alpha1",
                "kind": "HybridControlPlane",
                "metadata": {
                  "name": "placeholder"
                },
                "spec": {
                  "beaconAgent": {
                    "applianceProviderEnabled": true,
                    "beaconDBConnected": true,
                    "beaconServerAddress": "upm-beacon-server-ingestion.upm-beacon.svc.cluster.local:8080",
                    "inCluster": true,
                    "location": "devspatcher",
                    "logLevel": "debug",
                    "plaintext": true,
                    "providers": [
                      "appliance"
                    ],
                    "provisioning": {
                      "imageDiscovery": true,
                      "imagesetDiscoveryAuthenticationType": "token",
                      "imagesetDiscoveryContainerRegistryURL": "placeholder",
                      "loadBalancersEnabled": true,
                      "nodePortDomain": "nodeportdomain.com",
                      "openshift": true,
                      "provider": "aws"
                    },
                    "tlsInsecure": false,
                    "transparentDataEncryptionMethods": [
                      "passphrase",
                      "hashicorp_vault"
                    ]
                  },
                  "beaconServer": {
                    "logLevel": "debug"
                  },
                  "componentsParameters": {
                    "edb-migration-portal": {
                      "enable_pdb": "false"
                    },
                    "transporter-data-operator": {
                      "fips_enabled": "false"
                    },
                    "transporter-dp-agent": {
                      "rw_service_url": "placeholder"
                    },
                    "transporter-rw-service": {
                      "domain_name": "placeholder"
                    },
                    "upm-api-pnxt-metric": {
                      "log_level": "info",
                      "tsdb_host": "thanos-query.monitoring.svc.cluster.local",
                      "tsdb_port": "9090"
                    },
                    "upm-api-pnxt-pd": {
                      "log_level": "info"
                    },
                    "upm-beaco-ff-base": {
                      "usage_generator_interval": "24h"
                    },
                    "upm-beaco-usages": {
                      "usage_generator_interval": "24h"
                    },
                    "upm-beacon": {
                      "beacon_db_hibernation": "\"off\"",
                      "beacon_server_replicas": "1",
                      "external_metrics_replicas": "1",
                      "image_discovery_credentials_secret_name": "placeholder",
                      "server_host": "placeholder"
                    },
                    "upm-istio": {
                      "ingress_grpc_tls_node_port": "30290",
                      "ingress_http_node_port": "32542",
                      "ingress_https_node_port": "30288"
                    },
                    "upm-istio-gateway": {
                      "cookie_aeskey": "rzkutHl8NJNztPMEJYykZouHslNiA7xmIXH+58ISUVo=",
                      "ingress_grpc_tls_node_port": "30290",
                      "ingress_http_node_port": "32542",
                      "ingress_https_node_port": "30288",
                      "openshift_console_domain_name": "placeholder"
                    },
                    "upm-prometheus": {
                      "cpu": "200m",
                      "memory": "6Gi",
                      "prometheus_retention_size": "15GB",
                      "prometheus_storage_size": "20Gi"
                    },
                    "upm-thanos": {
                      "memcached_cpu": "500m",
                      "memcached_memory": "2Gi",
                      "query_frontend_cpu": "200m",
                      "query_frontend_memory": "128Mi"
                    },
                    "upm-trust-manager": {
                      "cert_manager_namespace": "cert-manager"
                    }
                  },
                  "flavour": "placeholder",
                  "globalParameters": {
                    "default_ingress_domain": "placeholder",
                    "internal_backup_folder": "a7462dbc7106",
                    "portal_certificate_secret": "beaco-istio-gw-cert",
                    "portal_domain_name": "placeholder",
                    "portal_port": "443",
                    "storage_class": "placeholder"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "placeholder",
                      "namespace": "default"
                    }
                  ],
                  "imageRegistry": "placeholder",
                  "pgai": {
                    "portal": {
                      "authentication": {
                        "clientSecret": "edb-postgres-ai-secret",
                        "idpConnectors": [],
                        "staticPasswords": [
                          {
                            "email": "owner@mycompany.com",
                            "hash": "$2y$10/JLTqiuGblH.RemIutWxkztN5m",
                            "userID": "c5998173-a605-449a-a9a5-4a9c33e26df7",
                            "username": "Owner MyCompany"
                          }
                        ]
                      }
                    }
                  },
                  "source": {
                    "useLocalKustomizations": true
                  },
                  "version": "placeholder"
                }
              }
            ]
          capabilities: Basic Install
          categories: Database,AI/Machine Learning
          containerImage: docker.enterprisedb.com/pgai-platform/edb-hcp-operator/manager@sha256:82610c624e956a67df4858929e1aabd561a7537ba68d40e6512b68f55e2649e0
          createdAt: "2025-06-17T11:00:42Z"
          description: Unify transactional, analytical, and AI workloads and bring
            cloud agility to your hybrid data infrastructure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/EnterpriseDB/edb-hcp-operator
          support: techsupport@enterprisedb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HybridControlPlane is the Schema for the hybridcontrolplanes
              API
            displayName: Hybrid Control Plane
            kind: HybridControlPlane
            name: hybridcontrolplanes.edbpgai.edb.com
            version: v1alpha1
        description: |
          EDB Postgres AI is a sovereign data and AI platform that brings analytics and AI closer to core operational data and enables cloud agility in private environments.
          Unify your structured and unstructured data in one multi-model data platform with centralized management and observability to:
          - Accelerate enterprise-grade transactional databases
          - Enable rapid analytics and powerful lakehouse integrations
          - Automate AI pipelines and semantic search for faster GenAI development
          Using EDB’s unmatched Postgres expertise, EDB Postgres AI automates database operations, ensuring expert-level execution across deployments – boosting operational efficiency by up to 30% and accelerating innovation.
        displayName: EDB Postgres AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Postgres
        - Database
        - DBaaS
        - EDB Postgres AI
        links:
        - name: EDB Postgres AI Software Deployment
          url: https://www.enterprisedb.com/products/software-deployment
        maintainers:
        - email: techsupport@enterprisedb.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - docker.enterprisedb.com/pgai-platform/edb-hcp-operator/manager@sha256:82610c624e956a67df4858929e1aabd561a7537ba68d40e6512b68f55e2649e0
        version: 1.2.0
      entries:
      - name: edb-hcp-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: edb-hcp-operator
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SAP LinuxLab - Red Hat
      provider-url: https://github.com/sap-linuxlab
    name: sap-data-intelligence-observer-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: observer-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sdi.sap-redhat.io/v1alpha1",
                "kind": "SDIObserver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "observer-operator",
                    "app.kubernetes.io/instance": "sdiobserver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sdiobserver",
                    "app.kubernetes.io/part-of": "observer-operator"
                  },
                  "name": "sdiobserver-sample"
                },
                "spec": {
                  "manageSDINodeConfig": true,
                  "sdiNamespace": "sdi",
                  "sdiVSystemRoute": {
                    "managementState": "Managed"
                  },
                  "slcbNamespace": "sap-slcbridge",
                  "slcbRoute": {
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/redhat-sap-cop/sdi-observer-operator:0.1.5
          description: Configure and enable SAP Data Intelligence to run on OpenShift
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SDIObserver is the Schema for the sdiobservers API
            displayName: SDIObserver
            kind: SDIObserver
            name: sdiobservers.sdi.sap-redhat.io
            version: v1alpha1
        description: Operator for monitoring SAP Data Intelligence (SDI) namespace
          and modifying objects in there that enable running of SDI on top of OpenShift.
          The observer shall be run in a dedicated namespace. It must be deployed
          before the SDI installation is started. This is the prerequisite for the
          successful installation of the SDI on OpenShift. Please refer to https://access.redhat.com/articles/5100521
          for detailed information.
        displayName: SAP Data Intelligence 3 - Observer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sdi
        - sap data intelligence
        links:
        - name: SAP Data Intelligence 3 - Observer Operator
          url: https://github.com/redhat-sap/sap-data-intelligence/tree/master/observer-operator
        maintainers:
        - email: mjiao@redhat.com
          name: Manjun Jiao
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/redhat-sap-cop/sdi-observer-operator:0.1.5
        version: 0.0.3
      entries:
      - name: observer-operator.v0.0.3
        version: 0.0.3
      name: alpha
    - currentCSV: observer-operator.v1.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sdi.sap-redhat.io/v1alpha1",
                "kind": "SDIObserver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/part-of": "observer-operator"
                  },
                  "name": "sdiobserver-sample",
                  "namespace": "sdi-observer"
                },
                "spec": {
                  "manageSDINodeConfig": true,
                  "sdiNamespace": "sdi",
                  "sdiVSystemRoute": {
                    "managementState": "Managed"
                  },
                  "slcbNamespace": "sap-slcbridge",
                  "slcbRoute": {
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/redhat-sap-cop/sdi-observer-operator:0.1.32
          createdAt: "20240726"
          description: Configure and enable SAP Data Intelligence 3 to run on OpenShift
            4
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mjiao/sap-data-intelligence/tree/master/observer-operator
          support: RedHat-SAPLinuxLab
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SDIObserver is the Schema for the sdiobservers API
            displayName: SDIObserver
            kind: SDIObserver
            name: sdiobservers.sdi.sap-redhat.io
            version: v1alpha1
        description: Operator for monitoring SAP Data Intelligence (SDI) namespace
          and modifying objects in there that enable running of SDI on top of OpenShift.
          It must be deployed before the SDI installation is started. This is the
          prerequisite for the successful installation of the SDI on OpenShift. Please
          refer to https://access.redhat.com/articles/7042265#ocp-install-sdi-observer-operator
          for detailed information.
        displayName: SAP Data Intelligence 3 - Observer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sdi
        - sap data intelligence
        links:
        - name: SAP Data Intelligence 3 - Observer Operator
          url: https://github.com/redhat-sap/sap-data-intelligence/tree/master/observer-operator
        maintainers:
        - email: mjiao@redhat.com
          name: Manjun Jiao
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/redhat-sap-cop/sdi-observer-operator:0.1.32
        version: 1.0.9
      entries:
      - name: observer-operator.v1.0.9
        version: 1.0.9
      - name: observer-operator.v1.0.8
        version: 1.0.8
      - name: observer-operator.v1.0.7
        version: 1.0.7
      - name: observer-operator.v1.0.6
        version: 1.0.6
      - name: observer-operator.v1.0.5
        version: 1.0.5
      - name: observer-operator.v1.0.4
        version: 1.0.4
      - name: observer-operator.v1.0.3
        version: 1.0.3
      - name: observer-operator.v1.0.2
        version: 1.0.2
      - name: observer-operator.v1.0.1
        version: 1.0.1
      - name: observer-operator.v0.0.3
        version: 0.0.3
      name: stable
    defaultChannel: stable
    packageName: sap-data-intelligence-observer-operator
    provider:
      name: SAP LinuxLab - Red Hat
      url: https://github.com/sap-linuxlab
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: k8gb.io
      provider-url: https://github.com/k8gb-io/k8gb
    name: k8gb
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k8gb.v0.11.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"k8gb.absa.oss/v1beta1","kind":"Gslb","metadata":{"name":"test-gslb-failover","namespace":"test-gslb"},"spec":{"ingress":{"rules":[{"host":"failover.test.k8gb.io","http":{"paths":[{"backend":{"service":{"name":"frontend-podinfo","port":{"name":"http"}}},"path":"/"}]}}]},"strategy":{"primaryGeoTag":"eu-west-1","type":"failover"}}}]'
          capabilities: Seamless Upgrades
          categories: Networking
          certified: "false"
          containerImage: docker.io/absaoss/k8gb:v0.11.4
          createdAt: "2021-09-24 12:00:00"
          description: A cloud native Kubernetes Global Balancer
          operatorframework.io/suggested-namespace: k8gb
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: k8gb
          operators.operatorframework.io/builder: operator-sdk-v1.12.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/k8gb-io/k8gb
          support: cncf-k8gb-maintainers@lists.cncf.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gslb resource for global load balancing strategy configuration
            displayName: Gslb
            kind: Gslb
            name: gslbs.k8gb.absa.oss
            version: v1beta1
          - description: Using ExternalDNS it synchronizes exposed Kubernetes Services
              and Ingresses with DNS providers
            displayName: DNSEndpoint
            kind: DNSEndpoint
            name: dnsendpoints.externaldns.k8s.io
            version: v1alpha1
        description: |
          **A cloud native Kubernetes Global Balancer**

          A Global Service Load Balancing solution with a focus on having cloud native qualities and work natively in a Kubernetes context.

          Key Differentiators:
           - Load balancing is based on timeproof DNS protocol which is perfect for global scope and extremely reliable
           - No dedicated management cluster and no single point of failure
           - Kubernetes native application health checks utilizing status of Liveness and Readiness probes for load balancing decisions
           - Configuration with a single Kubernetes CRD of `Gslb` kind

          Operator needs to be configured by setting couple of environment variables. This can be done by modifying
          the subscription yaml in the web ui or via kubectl. Here is an example of changing the `EDGE_DNS_ZONE`:

          ```yaml
          kind: Subscription
            ...
          spec:
            ...
            config:
              env:
              - name: CLUSTER_GEO_TAG
                value: us
              - name: EDGE_DNS_ZONE
                value: mycloud.example.com
                ...
           ```
           For more deployment variables see [helm code](https://github.com/k8gb-io/k8gb/blob/v0.8.3/chart/k8gb/templates/operator.yaml#L53:L68).

           For more information see [k8gb.io](https://k8gb.io).
        displayName: k8gb
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gslb
        - dns-lb
        - kuberneters-global-balancer
        - kubernetes-operator
        - balancer
        - multi-cluster
        links:
        - name: Home
          url: https://www.k8gb.io/
        - name: Source
          url: https://github.com/k8gb-io/k8gb
        maintainers:
        - email: dinar.valeev@absa.africa
          name: Dinar Valeev
        - email: jiri.kremser@gmail.com
          name: Jiri Kremser
        - email: kuritka@gmail.com
          name: Michal Kuritka
        - email: timofey.ilinykh@absa.africa
          name: Timofey Ilinykh
        - email: yury.tsarev@absa.africa
          name: Yury Tsarev
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: k8gb.io
          url: https://github.com/k8gb-io/k8gb
        relatedImages:
        - absaoss/k8s_crd:v0.0.11
        - docker.io/absaoss/k8gb:v0.11.4
        version: 0.11.4
      entries:
      - name: k8gb.v0.11.4
        version: 0.11.4
      name: alpha
    defaultChannel: alpha
    packageName: k8gb
    provider:
      name: k8gb.io
      url: https://github.com/k8gb-io/k8gb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Internal
      provider-url: redhat.com
    name: l5-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: l5-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pets.bestie.com/v1",
                "kind": "Bestie",
                "metadata": {
                  "name": "bestie"
                },
                "spec": {
                  "image": "quay.io/opdev/bestie",
                  "maxReplicas": 10,
                  "size": 3,
                  "version": "1.3"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Application Runtime
          certified: "false"
          containerImage: quay.io/opdev/l5-operator:v0.0.3
          createdAt: "2023-06-27T16:56:10Z"
          description: L5 Operator is a demo operator that aims to demonstrate level
            5 capabilities for web applications.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          relatedImages: quay.io/opdev/bestie@sha256:0e99ada2b1de19043845dd9076789584aeddc2646402e5f0987dd00424be1ee7
          repository: https://github.com/opdev/l5-operator-demo
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Bestie is the Schema for the besties API.
            displayName: Bestie
            kind: Bestie
            name: besties.pets.bestie.com
            version: v1
        description: |
          ## Usage
          The L5 operator is meant to demonstrate a minimum viable prototype of all capability levels outilned in the operator framework. The L5 operator is compatible with both Kubernetes and OpenShift clusters. When running on vanilla kubernetes a few prerequisites need to be kept in mind which may not be available out of the box. These include an ingress controller and a prometheus operator for level 4 capability. The L5 operator also depends on the Crunchy Data PostgreSQL operator. On openshift when installing via the Operator Lifecycle Manager, this will automatically be installed when installing the operator via the embdeded operator hub. The Operator Lifecycle Manager is also available for vanilla kubernetes however it may not be installed out of the box. In this case the Crunchy Data PostgreSQL operator must be installed manually. The Crunchy Data PostgreSQL operator can be installed on vanilla kuberentes using the [instructions provided by Crunchy Data](https://access.crunchydata.com/documentation/postgres-operator/4.7.5/installation/postgres-operator/) More documentation and examples can be found on [github](https://github.com/opdev/l5-operator-demo).
        displayName: L5 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - l5-operator
        links:
        - name: L5 Operator
          url: https://opdev.github.io/l5-operator-demo/
        maintainers:
        - email: partnereng@redhat.com
          name: OpDev Team
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat Internal
          url: redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/opdev/l5-operator:v0.0.3
        version: 0.0.3
      entries:
      - name: l5-operator.v0.0.3
        version: 0.0.3
      - name: l5-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: l5-operator
    provider:
      name: Red Hat Internal
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
    name: aikit-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aikit-operator.v2021.2.102120
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aikit.intel/v1alpha1",
                "kind": "AIKitContainer",
                "metadata": {
                  "name": "intel-aikit-container",
                  "namespace": "redhat-ods-applications"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagestream": {
                    "name": "oneapi-aikit",
                    "namespace": "redhat-ods-applications",
                    "registry": {
                      "root": "registry.connect.redhat.com",
                      "repo": "intel",
                      "name": "oneapi-aikit",
                      "version": "latest"
                    }
                  },
                  "nameOverride": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AIKitContainer
            name: aikitcontainers.aikit.intel
            version: v1alpha1
        description: |
          The Intel® oneAPI AI Analytics Toolkit gives data scientists, AI developers, and researchers familiar Python* tools and frameworks to accelerate end-to-end data science and machine learning pipelines on Intel® architectures. The components are built using oneAPI libraries for low-level compute optimizations. This toolkit maximizes performance from preprocessing through machine and deep learning phases and provides interoperability for efficient model development and deployment across single and multinodes.
        displayName: Intel® oneAPI AI Analytics Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aikit-operator
        links:
        - name: Aikit Operator
          url: https://aikit-operator.domain
        maintainers:
        - email: kam.d.kasravi@intel.com
          name: Kam Kasravi
        - email: abolfazl.shahbazi@intel.com
          name: Abolfazl Shahbazi
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Intel
          url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
        relatedImages:
        - registry.connect.redhat.com/intel/aikit-operator@sha256:7abb5e6164ffb3e4b87c39061f45d09848218fcfb74b48bf5e1035d56b41ee0c
        version: 2021.2.102120
      entries:
      - name: aikit-operator.v2021.2.102120
        version: 2021.2.102120
      - name: aikit-operator.v2021.2.100410
        version: 2021.2.100410
      - name: aikit-operator.v2021.2.92911
        version: 2021.2.92911
      - name: aikit-operator.v2021.2.92814
        version: 2021.2.92814
      - name: aikit-operator.v2021.2.92808-dev-a60e5ba
        version: 2021.2.92808-dev-a60e5ba
      name: alpha
    defaultChannel: alpha
    packageName: aikit-operator
    provider:
      name: Intel
      url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-airflow-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airflow-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:d8f8c439f33e5a80e16c7867a008d71693fd4928357ac6919c77bd0ff0bb1339
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:d8f8c439f33e5a80e16c7867a008d71693fd4928357ac6919c77bd0ff0bb1339
        version: 24.11.1
      entries:
      - name: airflow-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: airflow-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:173363db3cb0997aff09fe4b7f3a54c04f22bc56db83d2b30757e3cc745c4587
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:173363db3cb0997aff09fe4b7f3a54c04f22bc56db83d2b30757e3cc745c4587
        version: 25.3.0
      entries:
      - name: airflow-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: airflow-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
        version: 25.7.0
      entries:
      - name: airflow-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: airflow-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
        version: 25.7.0
      entries:
      - name: airflow-operator.v25.7.0
        version: 25.7.0
      - name: airflow-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-airflow-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: windows-machine-config-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: windows-machine-config-operator.v10.16.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          createdAt: "2025-06-06T22:08:07Z"
          description: An operator that enables Windows container workloads on OCP
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=10.15.0 <10.16.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.openshift.io/valid-subscription: '["Red Hat OpenShift support
            for Windows Containers"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ### Introduction
          The Windows Machine Config Operator configures Windows Machines into nodes, enabling Windows container workloads to
          be run on OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/4.16/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://docs.openshift.com/container-platform/4.16/windows_containers/byoh-windows-instance.html)
          The operator completes all the necessary steps to configure the Windows instance so that it can join the cluster as a worker node.

          Usage of this operator requires an OpenShift Container Platform for Windows Containers subscription. Users looking to
          deploy Windows containers workloads in production clusters should acquire a subscription before attempting to
          install this operator. Users without a subscription can try the community operator, a distribution which lacks
          official support.

          ### Pre-requisites
          * A Red Hat OpenShift Container Platform for Windows Containers [subscription](https://access.redhat.com/support/policy/updates/openshift#windows)
          * OCP 4.16 cluster running on Azure, AWS, GCP or vSphere configured with hybrid OVN Kubernetes networking
          * [WMCO prerequisites](https://docs.openshift.com/container-platform/4.16/windows_containers/understanding-windows-container-workloads.html#wmco-prerequisites__understanding-windows-container-workloads)

          ### Usage
          Once the `openshift-windows-machine-config-operator` namespace has been created, a secret must be created containing
          the private key that will be used to access the Windows VMs:
          ```
          # Create secret containing the private key in the openshift-windows-machine-config-operator namespace
          oc create secret generic cloud-private-key --from-file=private-key.pem=/path/to/key -n openshift-windows-machine-config-operator
          ```
          We strongly recommend not using the same
          [private key](https://docs.openshift.com/container-platform/4.16/installing/installing_azure/installing-azure-default.html#ssh-agent-using_installing-azure-default)
          used when installing the cluster

          Below is an example of a vSphere Windows MachineSet which can create Windows Machines that the WMCO can react upon.
          Please note that the windows-user-data secret will be created by the WMCO lazily when it is configuring the first
          Windows Machine. After that, the windows-user-data will be available for the subsequent MachineSets to be consumed.
          It might take around 10 minutes for the Windows VM to be configured so that it joins the cluster. Please note that
          the MachineSet should have following labels:
          * `machine.openshift.io/os-id: Windows`
          * `machine.openshift.io/cluster-api-machine-role: worker`
          * `machine.openshift.io/cluster-api-machine-type: worker`

          The following label must be added to the Machine spec within the MachineSet spec:
          * `node-role.kubernetes.io/worker: ""`

          Not having these labels will result in the Windows node not being marked as a worker.

          `<infrastructureID>` should be replaced with the output of:
          ```
          oc get -o jsonpath='{.status.infrastructureName}{"\n"}' infrastructure cluster
          ```

          The following template variables need to be replaced as follows with values from your vSphere environment:
          * `<Windows_VM_template>`: template name
          * `<VM Network Name>`: network name
          * `<vCenter DataCenter Name>`: datacenter name
          * `<Path to VM Folder in vCenter>`: path where your OpenShift cluster is running
          * `<vCenter Datastore Name>`: datastore name
          * `<vCenter Server FQDN/IP>`: IP address or FQDN of the vCenter server

          Please note that on vSphere, Windows Machine names cannot be more than 15 characters long. The MachineSet name therefore
          cannot be more than 9 characters long, due to the way Machine names are generated from it.
          ```
          apiVersion: machine.openshift.io/v1beta1
          kind: MachineSet
          metadata:
            labels:
              machine.openshift.io/cluster-api-cluster: <infrastructureID>
            name: winworker
            namespace: openshift-machine-api
          spec:
            replicas: 1
            selector:
              matchLabels:
                machine.openshift.io/cluster-api-cluster: <infrastructureID>
                machine.openshift.io/cluster-api-machineset: winworker
            template:
              metadata:
                labels:
                  machine.openshift.io/cluster-api-cluster: <infrastructureID>
                  machine.openshift.io/cluster-api-machine-role: worker
                  machine.openshift.io/cluster-api-machine-type: worker
                  machine.openshift.io/cluster-api-machineset: winworker
                  machine.openshift.io/os-id: Windows
              spec:
                metadata:
                  labels:
                    node-role.kubernetes.io/worker: ""
                providerSpec:
                  value:
                    apiVersion: vsphereprovider.openshift.io/v1beta1
                    credentialsSecret:
                      name: vsphere-cloud-credentials
                      namespace: openshift-machine-api
                    diskGiB: 128
                    kind: VSphereMachineProviderSpec
                    memoryMiB: 16384
                    metadata:
                      creationTimestamp: null
                    network:
                      devices:
                      - networkName: "<VM Network Name>"
                    numCPUs: 4
                    numCoresPerSocket: 1
                    snapshot: ""
                    template: <Windows_VM_template
                    userDataSecret:
                      name: windows-user-data
                    workspace:
                      datacenter: <vCenter DataCenter Name>
                      datastore: <vCenter Datastore Name>
                      folder: <Path to VM Folder in vCenter> # e.g. /DC/vm/ocp45-2tdrm
                      server: <vCenter Server FQDN/IP>
          ```
          Example MachineSet for other cloud providers:
          - [AWS](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-aws.html)
          - [Azure](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-azure.html)
          - [GCP](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-gcp.html)

          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.

          ### Reporting issues
          Support for this distribution of WMCO requires a Red Hat OpenShfit subscription. Support should be requested through the Red Hat Customer Portal.
          Please read through the [troubleshooting document](https://docs.openshift.com/container-platform/4.16/support/troubleshooting/troubleshooting-windows-container-workload-issues.html)
          before opening a support case.
        displayName: Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4-wincw/windows-machine-config-rhel9-operator@sha256:c9a6166158e1ccb1764007d79c6e0f879505c3546eb45cbbe79ebeefcc78a728
        version: 10.16.2
      entries:
      - name: windows-machine-config-operator.v10.16.2
        version: 10.16.2
      - name: windows-machine-config-operator.v10.16.1
        version: 10.16.1
      - name: windows-machine-config-operator.v10.16.0
        version: 10.16.0
      name: preview
    - currentCSV: windows-machine-config-operator.v10.16.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          createdAt: "2025-06-06T22:08:07Z"
          description: An operator that enables Windows container workloads on OCP
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=10.15.0 <10.16.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.openshift.io/valid-subscription: '["Red Hat OpenShift support
            for Windows Containers"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ### Introduction
          The Windows Machine Config Operator configures Windows Machines into nodes, enabling Windows container workloads to
          be run on OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/4.16/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://docs.openshift.com/container-platform/4.16/windows_containers/byoh-windows-instance.html)
          The operator completes all the necessary steps to configure the Windows instance so that it can join the cluster as a worker node.

          Usage of this operator requires an OpenShift Container Platform for Windows Containers subscription. Users looking to
          deploy Windows containers workloads in production clusters should acquire a subscription before attempting to
          install this operator. Users without a subscription can try the community operator, a distribution which lacks
          official support.

          ### Pre-requisites
          * A Red Hat OpenShift Container Platform for Windows Containers [subscription](https://access.redhat.com/support/policy/updates/openshift#windows)
          * OCP 4.16 cluster running on Azure, AWS, GCP or vSphere configured with hybrid OVN Kubernetes networking
          * [WMCO prerequisites](https://docs.openshift.com/container-platform/4.16/windows_containers/understanding-windows-container-workloads.html#wmco-prerequisites__understanding-windows-container-workloads)

          ### Usage
          Once the `openshift-windows-machine-config-operator` namespace has been created, a secret must be created containing
          the private key that will be used to access the Windows VMs:
          ```
          # Create secret containing the private key in the openshift-windows-machine-config-operator namespace
          oc create secret generic cloud-private-key --from-file=private-key.pem=/path/to/key -n openshift-windows-machine-config-operator
          ```
          We strongly recommend not using the same
          [private key](https://docs.openshift.com/container-platform/4.16/installing/installing_azure/installing-azure-default.html#ssh-agent-using_installing-azure-default)
          used when installing the cluster

          Below is an example of a vSphere Windows MachineSet which can create Windows Machines that the WMCO can react upon.
          Please note that the windows-user-data secret will be created by the WMCO lazily when it is configuring the first
          Windows Machine. After that, the windows-user-data will be available for the subsequent MachineSets to be consumed.
          It might take around 10 minutes for the Windows VM to be configured so that it joins the cluster. Please note that
          the MachineSet should have following labels:
          * `machine.openshift.io/os-id: Windows`
          * `machine.openshift.io/cluster-api-machine-role: worker`
          * `machine.openshift.io/cluster-api-machine-type: worker`

          The following label must be added to the Machine spec within the MachineSet spec:
          * `node-role.kubernetes.io/worker: ""`

          Not having these labels will result in the Windows node not being marked as a worker.

          `<infrastructureID>` should be replaced with the output of:
          ```
          oc get -o jsonpath='{.status.infrastructureName}{"\n"}' infrastructure cluster
          ```

          The following template variables need to be replaced as follows with values from your vSphere environment:
          * `<Windows_VM_template>`: template name
          * `<VM Network Name>`: network name
          * `<vCenter DataCenter Name>`: datacenter name
          * `<Path to VM Folder in vCenter>`: path where your OpenShift cluster is running
          * `<vCenter Datastore Name>`: datastore name
          * `<vCenter Server FQDN/IP>`: IP address or FQDN of the vCenter server

          Please note that on vSphere, Windows Machine names cannot be more than 15 characters long. The MachineSet name therefore
          cannot be more than 9 characters long, due to the way Machine names are generated from it.
          ```
          apiVersion: machine.openshift.io/v1beta1
          kind: MachineSet
          metadata:
            labels:
              machine.openshift.io/cluster-api-cluster: <infrastructureID>
            name: winworker
            namespace: openshift-machine-api
          spec:
            replicas: 1
            selector:
              matchLabels:
                machine.openshift.io/cluster-api-cluster: <infrastructureID>
                machine.openshift.io/cluster-api-machineset: winworker
            template:
              metadata:
                labels:
                  machine.openshift.io/cluster-api-cluster: <infrastructureID>
                  machine.openshift.io/cluster-api-machine-role: worker
                  machine.openshift.io/cluster-api-machine-type: worker
                  machine.openshift.io/cluster-api-machineset: winworker
                  machine.openshift.io/os-id: Windows
              spec:
                metadata:
                  labels:
                    node-role.kubernetes.io/worker: ""
                providerSpec:
                  value:
                    apiVersion: vsphereprovider.openshift.io/v1beta1
                    credentialsSecret:
                      name: vsphere-cloud-credentials
                      namespace: openshift-machine-api
                    diskGiB: 128
                    kind: VSphereMachineProviderSpec
                    memoryMiB: 16384
                    metadata:
                      creationTimestamp: null
                    network:
                      devices:
                      - networkName: "<VM Network Name>"
                    numCPUs: 4
                    numCoresPerSocket: 1
                    snapshot: ""
                    template: <Windows_VM_template
                    userDataSecret:
                      name: windows-user-data
                    workspace:
                      datacenter: <vCenter DataCenter Name>
                      datastore: <vCenter Datastore Name>
                      folder: <Path to VM Folder in vCenter> # e.g. /DC/vm/ocp45-2tdrm
                      server: <vCenter Server FQDN/IP>
          ```
          Example MachineSet for other cloud providers:
          - [AWS](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-aws.html)
          - [Azure](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-azure.html)
          - [GCP](https://docs.openshift.com/container-platform/4.16/windows_containers/creating_windows_machinesets/creating-windows-machineset-gcp.html)

          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.

          ### Reporting issues
          Support for this distribution of WMCO requires a Red Hat OpenShfit subscription. Support should be requested through the Red Hat Customer Portal.
          Please read through the [troubleshooting document](https://docs.openshift.com/container-platform/4.16/support/troubleshooting/troubleshooting-windows-container-workload-issues.html)
          before opening a support case.
        displayName: Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4-wincw/windows-machine-config-rhel9-operator@sha256:c9a6166158e1ccb1764007d79c6e0f879505c3546eb45cbbe79ebeefcc78a728
        version: 10.16.2
      entries:
      - name: windows-machine-config-operator.v10.16.2
        version: 10.16.2
      - name: windows-machine-config-operator.v10.16.1
        version: 10.16.1
      - name: windows-machine-config-operator.v10.16.0
        version: 10.16.0
      name: stable
    defaultChannel: stable
    packageName: windows-machine-config-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-documentdb-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-documentdb-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBInstance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/documentdb-controller:1.1.1
          createdAt: "2025-09-19T22:13:37Z"
          description: AWS DocumentDB controller is a service controller for managing
            DocumentDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DBCluster represents the state of an AWS documentdb DBCluster
              resource.
            displayName: DBCluster
            kind: DBCluster
            name: dbclusters.documentdb.services.k8s.aws
            version: v1alpha1
          - description: DBInstance represents the state of an AWS documentdb DBInstance
              resource.
            displayName: DBInstance
            kind: DBInstance
            name: dbinstances.documentdb.services.k8s.aws
            version: v1alpha1
          - description: DBSubnetGroup represents the state of an AWS documentdb DBSubnetGroup
              resource.
            displayName: DBSubnetGroup
            kind: DBSubnetGroup
            name: dbsubnetgroups.documentdb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon DocumentDB resources in AWS from within your Kubernetes cluster.

          **About Amazon DocumentDB**

          Amazon DocumentDB (with MongoDB compatibility) is a fast, reliable, and fully managed database service. Amazon DocumentDB makes it easy to set up, operate, and scale MongoDB-compatible databases in the cloud. With Amazon DocumentDB, you can run the same application code and use the same drivers and tools that you use with MongoDB.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon DocumentDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - documentdb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon DocumentDB Developer Resources
          url: https://aws.amazon.com/documentdb/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: documentdb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/documentdb-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-documentdb-controller.v1.1.1
        version: 1.1.1
      - name: ack-documentdb-controller.v1.1.0
        version: 1.1.0
      - name: ack-documentdb-controller.v1.0.14
        version: 1.0.14
      - name: ack-documentdb-controller.v1.0.13
        version: 1.0.13
      - name: ack-documentdb-controller.v1.0.12
        version: 1.0.12
      - name: ack-documentdb-controller.v1.0.11
        version: 1.0.11
      - name: ack-documentdb-controller.v1.0.10
        version: 1.0.10
      - name: ack-documentdb-controller.v1.0.9
        version: 1.0.9
      - name: ack-documentdb-controller.v1.0.7
        version: 1.0.7
      - name: ack-documentdb-controller.v1.0.6
        version: 1.0.6
      - name: ack-documentdb-controller.v1.0.5
        version: 1.0.5
      - name: ack-documentdb-controller.v1.0.2
        version: 1.0.2
      - name: ack-documentdb-controller.v1.0.1
        version: 1.0.1
      - name: ack-documentdb-controller.v1.0.0
        version: 1.0.0
      - name: ack-documentdb-controller.v0.0.9
        version: 0.0.9
      - name: ack-documentdb-controller.v0.0.8
        version: 0.0.8
      - name: ack-documentdb-controller.v0.0.7
        version: 0.0.7
      - name: ack-documentdb-controller.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: ack-documentdb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NGINX Inc
      provider-url: ""
    name: nginx-ingress-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nginx-ingress-operator.v3.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.nginx.org/v1alpha1",
                "kind": "NginxIngress",
                "metadata": {
                  "name": "nginxingress-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "annotations": {},
                    "appprotect": {
                      "enable": false
                    },
                    "appprotectdos": {
                      "debug": false,
                      "enable": false,
                      "maxDaemons": 0,
                      "maxWorkers": 0,
                      "memory": 0
                    },
                    "autoscaling": {
                      "annotations": {},
                      "behavior": {},
                      "enabled": false,
                      "maxReplicas": 3,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 50,
                      "targetMemoryUtilizationPercentage": 50
                    },
                    "config": {
                      "annotations": {},
                      "entries": {}
                    },
                    "containerPort": {
                      "http": 80,
                      "https": 443
                    },
                    "customConfigMap": "",
                    "customPorts": [],
                    "defaultHTTPListenerPort": 80,
                    "defaultHTTPSListenerPort": 443,
                    "defaultTLS": {
                      "cert": "",
                      "key": "",
                      "secret": ""
                    },
                    "disableIPV6": false,
                    "dnsPolicy": "ClusterFirst",
                    "enableCertManager": false,
                    "enableCustomResources": true,
                    "enableExternalDNS": false,
                    "enableLatencyMetrics": false,
                    "enableOIDC": false,
                    "enableSSLDynamicReload": true,
                    "enableSnippets": false,
                    "enableTLSPassthrough": false,
                    "env": [],
                    "extraContainers": [],
                    "globalConfiguration": {
                      "create": false,
                      "spec": {}
                    },
                    "healthStatus": false,
                    "healthStatusURI": "/nginx-health",
                    "hostNetwork": false,
                    "hostPort": {
                      "enable": false,
                      "http": 80,
                      "https": 443
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "nginx/nginx-ingress",
                      "tag": "5.2.0-ubi"
                    },
                    "ingressClass": {
                      "create": true,
                      "name": "nginx",
                      "setAsDefaultIngress": false
                    },
                    "initContainerResources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "initContainers": [],
                    "kind": "deployment",
                    "lifecycle": {},
                    "logFormat": "glog",
                    "logLevel": "info",
                    "mgmt": {
                      "licenseTokenSecretName": "license-token"
                    },
                    "minReadySeconds": 0,
                    "name": "controller",
                    "nginxDebug": false,
                    "nginxReloadTimeout": 60000,
                    "nginxStatus": {
                      "allowCidrs": "127.0.0.1",
                      "enable": true,
                      "port": 8080
                    },
                    "nginxplus": false,
                    "pod": {
                      "annotations": {},
                      "extraLabels": {}
                    },
                    "podDisruptionBudget": {
                      "annotations": {},
                      "enabled": false
                    },
                    "readOnlyRootFilesystem": false,
                    "readyStatus": {
                      "enable": true,
                      "initialDelaySeconds": 0,
                      "port": 8081
                    },
                    "replicaCount": 1,
                    "reportIngressStatus": {
                      "annotations": {},
                      "enable": true,
                      "enableLeaderElection": true,
                      "ingressLink": "",
                      "leaderElectionLockName": "nginx-ingress-leader"
                    },
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "selectorLabels": {},
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "create": true,
                      "customPorts": [],
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraLabels": {},
                      "httpPort": {
                        "enable": true,
                        "port": 80,
                        "targetPort": 80
                      },
                      "httpsPort": {
                        "enable": true,
                        "port": 443,
                        "targetPort": 443
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "type": "LoadBalancer"
                    },
                    "serviceAccount": {
                      "annotations": {},
                      "imagePullSecretName": "",
                      "imagePullSecretsNames": []
                    },
                    "shareProcessNamespace": false,
                    "strategy": {},
                    "terminationGracePeriodSeconds": 30,
                    "tlsPassthroughPort": 443,
                    "tolerations": [],
                    "volumeMounts": [],
                    "volumes": [],
                    "watchNamespace": "",
                    "watchNamespaceLabel": "",
                    "watchSecretNamespace": "",
                    "wildcardTLS": {
                      "cert": "",
                      "key": "",
                      "secret": ""
                    }
                  },
                  "nginxServiceMesh": {
                    "enable": false,
                    "enableEgress": false
                  },
                  "prometheus": {
                    "create": true,
                    "port": 9113,
                    "scheme": "http",
                    "secret": "",
                    "service": {
                      "create": false,
                      "labels": {
                        "service": "nginx-ingress-prometheus-service"
                      }
                    },
                    "serviceMonitor": {
                      "create": false,
                      "endpoints": [
                        {
                          "port": "prometheus"
                        }
                      ],
                      "labels": {},
                      "selectorMatchLabels": {
                        "service": "nginx-ingress-prometheus-service"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "serviceInsight": {
                    "create": false,
                    "port": 9114,
                    "scheme": "http",
                    "secret": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking
          certified: "true"
          containerImage: quay.io/nginx/nginx-ingress-operator@sha256:fb372438634b4a667218e725091bd448a253e7a9c40e7dda1044a318780d995f
          createdAt: "2025-09-15T16:08:18Z"
          description: The NGINX Ingress Operator is a Kubernetes/OpenShift component
            which deploys and manages one or more NGINX/NGINX Plus Ingress Controllers
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nginx-ingress
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/nginx/nginx-ingress-helm-operator
          support: NGINX Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The `NginxIngress` Custom Resource is the definition of a
              deployment of the Ingress Controller.
            displayName: Nginx Ingress Controller
            kind: NginxIngress
            name: nginxingresses.charts.nginx.org
            version: v1alpha1
        description: The NGINX Ingress Operator is a Kubernetes/OpenShift component
          which deploys and manages one or more NGINX/NGINX Plus Ingress Controllers
        displayName: Nginx Ingress Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nginx
        - ingress-controller
        - ingress
        - controller
        - kubernetes
        - openshift
        links:
        - name: Nginx Ingress Operator
          url: https://github.com/nginx/nginx-ingress-helm-operator
        maintainers:
        - email: kubernetes@nginx.com
          name: NGINX Inc
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: NGINX Inc
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy@sha256:754ab2a723c8fc7ea8e09ae1224f8af5be91849b8b2f2a3735c6237f1c534187
        - quay.io/nginx/nginx-ingress-operator@sha256:fb372438634b4a667218e725091bd448a253e7a9c40e7dda1044a318780d995f
        version: 3.3.0
      entries:
      - name: nginx-ingress-operator.v3.3.0
        version: 3.3.0
      name: alpha
    defaultChannel: alpha
    packageName: nginx-ingress-operator
    provider:
      name: NGINX Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: QuestLabs
      provider-url: https://www.questlabs.in
    name: alloy
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: alloy.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.alloys.com/v1alpha1",
                "kind": "Alloy",
                "metadata": {
                  "name": "alloy"
                },
                "spec": {
                  "acmeEmail": "",
                  "dnsFQDN": "",
                  "sslTermination": "true"
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Alloy
            name: alloys.charts.alloys.com
            version: v1alpha1
        description: |-
          Alloy is platform to build trusted data in OSDU, available on-premise or in the cloud. It is equipped with built-in data quality rules derived from PPDM and OSDU standards that control the ingestion of data.

              • Applicable for conventional and new energy streams
              • Highly Scalable

              Benefits
              • Helps to build robust applications/solutions on stable data foundation using full data lifecycle APIs & built-in data/business rules
              • Help data governance team to monitor/oversea data across organisation and intervene when required
              • Substantially low implementation time for any new data entity or new domain (solar, wind etc.) because of modular rules packages
              • Capability to deploy workflows for complex corporate specific data quality and enrichment requirements
              • Spend less time & resources to deploy fit-for-purpose data
              • Schema based checks out of the box for conventional and new energy streams.
              • Sustainable and future proof


              Prerequisites
              1. Please create alloy project before installing alloy operator

          Getting Help

          Once Alloy is successfully installed you can also browse documentation at https://your.domain.com/alloy/docs/introduction
          Please visit https://www.questlabs.in/alloy for more details and if you have any queries or suggestions, mail us at info@questlabs.in
        displayName: Alloy
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - alloy
        links:
        - name: alloy
          url: https://www.questlabs.in
        maintainers:
        - email: info@questlabs.in
          name: QuestLabs
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: QuestLabs
          url: https://www.questlabs.in
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d291e0d38bd95eb2731a10bd5ff06bfa07c6621528d3c2bbe888397d87b3feaa
        - alloy.azurecr.io/alloy@sha256:bf29f69889d408507cf1a9a940beb80d95ba9faff1da2d70a4aea82cc235d96d
        version: 1.0.4
      entries:
      - name: alloy.v1.0.4
        version: 1.0.4
      - name: alloy.v1.0.3
        version: 1.0.3
      - name: alloy.v1.0.2
        version: 1.0.2
      - name: alloy.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: alloy
    provider:
      name: QuestLabs
      url: https://www.questlabs.in
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: SynaXG Engineering Team
      provider-url: ""
    name: synaxg-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: synaxg-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "card.synaxg.app.com/v1",
                "kind": "SG3Reset",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "synaxg-operator",
                    "app.kubernetes.io/instance": "sg3reset-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sg3reset",
                    "app.kubernetes.io/part-of": "synaxg-operator"
                  },
                  "name": "sg3reset-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/synaxg/synaxg-reset-utility@sha256:4050d5731a21bbec6e26165b8dc2b7632577ffa376e7eb056aa3245cf9574753",
                  "sg3ResetList": [
                    {
                      "nodeName": "test",
                      "pciList": [
                        {
                          "enable": "false",
                          "pciAddr": "000d:01:00.0"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "card.synaxg.app.com/v1",
                "kind": "SG3Update",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "synaxg-operator",
                    "app.kubernetes.io/instance": "sg3update-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sg3update",
                    "app.kubernetes.io/part-of": "synaxg-operator"
                  },
                  "name": "sg3update-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/synaxg/synaxg-upgrade-utility@sha256:7ffec3249edd2467f2414c2ce4232e8e436278fd1b43a51b66f54b6ea10a60d1",
                  "sg3UpdateList": [
                    {
                      "nodeName": "test",
                      "pciList": [
                        {
                          "enable": "false",
                          "pciAddr": "000d:01:00.0",
                          "sdkImageAddr": "quay.io/synaxgcom/sdk-img@sha256:1ed57013f278e5a858f71f343649652dbd52b6f3cf451374e45b6ac427fa76f9"
                        }
                      ]
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/synaxg/synaxg-operator-utility@sha256:3b22a8022a97cb1ffdb47ffcc9d9d562601f48669aa14fd58cd8ba6128b39bd7
          createdAt: "2025-07-01T03:42:07Z"
          description: SynaXG Operator will work with the SuperMicro Server + SG3
            card  and Openshift Container Platform.The Marvell kernel module “octeon_ep”
            and “octeon_ep_vf” are necessary to the Operator. The Operator can send
            command of “reset” and “update” to SG3 board via the VF channel.As a result,
            users can realize the process of “reset card” and “update firmware” on
            the web GUI, and monitor the executing result.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: free text
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: SynaXG
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SG3Reset is the Schema for the sg3resets API
            displayName: SG3 Reset
            kind: SG3Reset
            name: sg3resets.card.synaxg.app.com
            version: v1
          - description: SG3Update is the Schema for the sg3updates API
            displayName: SG3 Update
            kind: SG3Update
            name: sg3updates.card.synaxg.app.com
            version: v1
        description: SynaXG Operator will work with the SuperMicro Server + SG3 card  and
          Openshift Container Platform.The Marvell kernel module “octeon_ep” and “octeon_ep_vf”
          are necessary to the Operator. The Operator can send command of “reset”
          and “update” to SG3 board via the VF channel.As a result, users can realize
          the process of “reset card” and “update firmware” on the web GUI, and monitor
          the executing result.
        displayName: SynaXG vDU card 102X Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vDU
        - 102X
        - SynaXG
        links:
        - name: Synaxg Operator
          url: https://synaxg-operator.domain
        maintainers:
        - email: notice@synaxg.com
          name: Mantosh Malhotra
        maturity: alpha
        provider:
          name: SynaXG Engineering Team
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:38ee3484ba60bb474e7231e9d0718157386716cea7ffd6da6f32ddd551a5243d
        - registry.connect.redhat.com/synaxg/synaxg-operator-utility@sha256:3b22a8022a97cb1ffdb47ffcc9d9d562601f48669aa14fd58cd8ba6128b39bd7
        - quay.io/synaxgcom/sdk-img@sha256:1ed57013f278e5a858f71f343649652dbd52b6f3cf451374e45b6ac427fa76f9
        - registry.connect.redhat.com/synaxg/synaxg-reset-utility@sha256:4050d5731a21bbec6e26165b8dc2b7632577ffa376e7eb056aa3245cf9574753
        - registry.connect.redhat.com/synaxg/synaxg-upgrade-utility@sha256:7ffec3249edd2467f2414c2ce4232e8e436278fd1b43a51b66f54b6ea10a60d1
        version: 1.2.0
      entries:
      - name: synaxg-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: synaxg-operator
    provider:
      name: SynaXG Engineering Team
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-opa-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opa-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:08f995a39db5c2c6a17008cffa5216bb9bebd784d13d2b94154474d727fc796f
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/opa-operator@sha256:08f995a39db5c2c6a17008cffa5216bb9bebd784d13d2b94154474d727fc796f
        version: 24.11.1
      entries:
      - name: opa-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: opa-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:826aa12df4531d93a337ba70bd0c0d1e2e8062efe6954fc22797f6474c7732ca
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/opa-operator@sha256:826aa12df4531d93a337ba70bd0c0d1e2e8062efe6954fc22797f6474c7732ca
        version: 25.3.0
      entries:
      - name: opa-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: opa-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaClusterSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
        version: 25.7.0
      entries:
      - name: opa-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: opa-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaClusterSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
        version: 25.7.0
      entries:
      - name: opa-operator.v25.7.0
        version: 25.7.0
      - name: opa-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-opa-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: datagrid
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datagrid-operator.v8.2.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "adminAuth": {
                    "secretName": "basic-auth"
                  },
                  "clusterName": "example-infinispan",
                  "name": "mycache"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC batch-cache"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:1435aa60e8ab4d25c86daae68dcb93d14999f50d7c1e11a3c5726a86d3d30453
          createdAt: "2021-08-24T17:39:44Z"
          description: Create and manage Red Hat Data Grid clusters.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Restore API schema.
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Interoperability:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Data Grid Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
          Red Hat Data Grid is a distributed, in-memory data store that increases application performance and delivers open-source capabilities to handle demanding use cases.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.2/html/data_grid_operator_8.2_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.2
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:1435aa60e8ab4d25c86daae68dcb93d14999f50d7c1e11a3c5726a86d3d30453
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4d8057709958ef60bcabdcec52f05eae8df121fd3f4762c3218abca38130f638
        - registry.redhat.io/datagrid/datagrid-8-openj9-11-rhel8@sha256:c8607d6538a7528e27e6ecb138e0a93a03b3c9fa1f8a9a5aad3fd9f4a6814448
        version: 8.2.8
      entries:
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.2.x
    - currentCSV: datagrid-operator.v8.3.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:4d9d7221b029866837add5f3cb52dd0b7956f57996ce2d8b76344ce2de678716
          createdAt: "2022-11-03T08:40:57Z"
          description: Create and manage Red Hat Data Grid clusters.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.3/html/data_grid_operator_8.3_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.3
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:4d9d7221b029866837add5f3cb52dd0b7956f57996ce2d8b76344ce2de678716
        version: 8.3.9
      entries:
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.3.x
    - currentCSV: datagrid-operator.v8.4.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:e5315682869de72514e3e027bfb02bf0bb64919453ca738c7ecfa1e34634ac2b
          createdAt: 2024-05-30T08:05:19 EDT
          description: Create and manage Red Hat Data Grid clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.4/html/data_grid_operator_8.4_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.4
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:985193caf9bc3698c4d85aeedd1cfea09bccda881636d67885387b2a83c65ebd
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9d9e466b7c367d0406406550c74bc790aa68f3f875ea4f652eb28f2e25a05458
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:b56269cb849afca34e7f82cc8e01a44bb32af6126f24db28019a58e556d4cb3c
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:d522a20ca6909bd464f85967d8b86fe42b83f773478a80785f9a85a0cb87dfed
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bf7760520f5bea151a531d538d525c4591c47166995aefb4d77fef138de11544
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bcb040bf640691568a93a4c71c8072c37ffb76b4296ade83186f279862b7f197
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:24cf79efa89bbfa495665b1711a27531b7ab0813695a89b2c38c96c2e792eb24
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:532dca6facb584c1261c0480c499af9f857c70dc390e0d46e46afec2561468af
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:78a6bed6fcb6962e5689b92806789b72a66db21b7c607b392a573c98593f2c09
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:ae852e4679978725bb7d24c1bdfe25b40bd883130ed04a36d4fca49dbea024ef
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:424b30dccd773f69c079b35a42e8c3e339246c5d490e14e3fcfb8bd8144a9160
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9223b4b5c1c866a91033dac40f30a4caae0f6946cdb64935a3fdfffdb9ede660
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:c15d4fd827f554291c14851f187947f4109c0e20257fa2f1e141b32669569358
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:54bc26a0fddc160a4e0d1a813aec78abb3c9cb77db5039b74c23de759faa67f5
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:554b831a0803d55a4e3b7022011b5a5f6d2b4b65009065a9b28bc8e55fc3ad46
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:e5315682869de72514e3e027bfb02bf0bb64919453ca738c7ecfa1e34634ac2b
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4dbe1b7a1a36990c0a31d5a66f3cc4ca2954c9e1ec3b87431113a4111530277d
        version: 8.4.16
      entries:
      - name: datagrid-operator.v8.4.16
        version: 8.4.16
      - name: datagrid-operator.v8.4.15
        version: 8.4.15
      - name: datagrid-operator.v8.4.14
        version: 8.4.14
      - name: datagrid-operator.v8.4.13
        version: 8.4.13
      - name: datagrid-operator.v8.4.12
        version: 8.4.12
      - name: datagrid-operator.v8.4.11
        version: 8.4.11
      - name: datagrid-operator.v8.4.10
        version: 8.4.9
      - name: datagrid-operator.v8.4.8
        version: 8.4.8
      - name: datagrid-operator.v8.4.7
        version: 8.4.7
      - name: datagrid-operator.v8.4.6
        version: 8.4.6
      - name: datagrid-operator.v8.4.5
        version: 8.4.5
      - name: datagrid-operator.v8.4.4
        version: 8.4.4
      - name: datagrid-operator.v8.4.3
        version: 8.4.3
      - name: datagrid-operator.v8.4.2
        version: 8.4.2
      - name: datagrid-operator.v8.4.1
        version: 8.4.1
      - name: datagrid-operator.v8.4.0
        version: 8.4.0
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.4.x
    - currentCSV: datagrid-operator.v8.5.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel9-operator@sha256:d719d778b0848588a7e61d3c3fa689763f51a6533af4cf9e1c738c4ed8fac760
          createdAt: 2025-08-29T05:44:37 UTC
          description: Create and manage Red Hat Data Grid clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.5/html/data_grid_operator_8.5_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.5
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:a711c50e489995fb60e5dec23ce286360743feda5066e4a0861d448aa4f2e903
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bf7760520f5bea151a531d538d525c4591c47166995aefb4d77fef138de11544
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bcb040bf640691568a93a4c71c8072c37ffb76b4296ade83186f279862b7f197
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:4b9bf39d2ff14bcfbd4628a399093703d87a2b2b9c8e74c3d7cfdff4b5167fa4
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:ae852e4679978725bb7d24c1bdfe25b40bd883130ed04a36d4fca49dbea024ef
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:d522a20ca6909bd464f85967d8b86fe42b83f773478a80785f9a85a0cb87dfed
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9223b4b5c1c866a91033dac40f30a4caae0f6946cdb64935a3fdfffdb9ede660
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:11bfe5c8109d26164cff428232b2001ab8df464f14a382661b936543336e2960
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:ad4f91f984b4b898012963d52e6ebcd166ec03e76e26eba3191ecb87f51fb18d
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:c15d4fd827f554291c14851f187947f4109c0e20257fa2f1e141b32669569358
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:5656cd188d054551489d45be996266d03e3faa8814b8bfc0f360cf0dceb83cbe
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:424b30dccd773f69c079b35a42e8c3e339246c5d490e14e3fcfb8bd8144a9160
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:879f3dcc24ffe795472779f777d955666ba4041c17e717d5e165cd7f12d7955a
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:985193caf9bc3698c4d85aeedd1cfea09bccda881636d67885387b2a83c65ebd
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:b56269cb849afca34e7f82cc8e01a44bb32af6126f24db28019a58e556d4cb3c
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:9bc0b730570ab55fbeca83075949f4caa96ed5c37637420a59cd4a7db01ea840
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:532dca6facb584c1261c0480c499af9f857c70dc390e0d46e46afec2561468af
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9d9e466b7c367d0406406550c74bc790aa68f3f875ea4f652eb28f2e25a05458
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:54bc26a0fddc160a4e0d1a813aec78abb3c9cb77db5039b74c23de759faa67f5
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:78a6bed6fcb6962e5689b92806789b72a66db21b7c607b392a573c98593f2c09
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4dbe1b7a1a36990c0a31d5a66f3cc4ca2954c9e1ec3b87431113a4111530277d
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:48cf13e71953cea6f62354289f0d8ece7a8a481ee0bb2c76eb9352d438db9ca3
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:0d368459c98c5788dd55d2add5b4b7d438b9a8c814ea8aee2222ea956699cbb6
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:c25e3152d6a6b6d5cd86d311bf567e594b9948440785209cd8d7986076a5a10d
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:f59ae21773dfc207454a709ac5f7aad2be6a6a20ac88da26c379edeed54537b8
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:24cf79efa89bbfa495665b1711a27531b7ab0813695a89b2c38c96c2e792eb24
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:6e1766784b684561666134d41dca5331b80d50e63695fe7b1a1d6d4cdb73f0bd
        - registry.redhat.io/datagrid/datagrid-8-rhel9-operator@sha256:d719d778b0848588a7e61d3c3fa689763f51a6533af4cf9e1c738c4ed8fac760
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:554b831a0803d55a4e3b7022011b5a5f6d2b4b65009065a9b28bc8e55fc3ad46
        version: 8.5.11
      entries:
      - name: datagrid-operator.v8.5.11
        version: 8.5.11
      - name: datagrid-operator.v8.5.10
        version: 8.5.10
      - name: datagrid-operator.v8.5.9
        version: 8.5.9
      - name: datagrid-operator.v8.5.8
        version: 8.5.8
      - name: datagrid-operator.v8.5.7
        version: 8.5.7
      - name: datagrid-operator.v8.5.6
        version: 8.5.6
      - name: datagrid-operator.v8.5.5
        version: 8.5.5
      - name: datagrid-operator.v8.5.4
        version: 8.5.4
      - name: datagrid-operator.v8.5.3
        version: 8.5.3
      - name: datagrid-operator.v8.5.2
        version: 8.5.2
      - name: datagrid-operator.v8.5.1
        version: 8.5.1
      - name: datagrid-operator.v8.5.0
        version: 8.5.0
      - name: datagrid-operator.v8.4.16
        version: 8.4.16
      - name: datagrid-operator.v8.4.15
        version: 8.4.15
      - name: datagrid-operator.v8.4.14
        version: 8.4.14
      - name: datagrid-operator.v8.4.13
        version: 8.4.13
      - name: datagrid-operator.v8.4.12
        version: 8.4.12
      - name: datagrid-operator.v8.4.11
        version: 8.4.11
      - name: datagrid-operator.v8.4.10
        version: 8.4.9
      - name: datagrid-operator.v8.4.8
        version: 8.4.8
      - name: datagrid-operator.v8.4.7
        version: 8.4.7
      - name: datagrid-operator.v8.4.6
        version: 8.4.6
      - name: datagrid-operator.v8.4.5
        version: 8.4.5
      - name: datagrid-operator.v8.4.4
        version: 8.4.4
      - name: datagrid-operator.v8.4.3
        version: 8.4.3
      - name: datagrid-operator.v8.4.2
        version: 8.4.2
      - name: datagrid-operator.v8.4.1
        version: 8.4.1
      - name: datagrid-operator.v8.4.0
        version: 8.4.0
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: stable
    defaultChannel: stable
    packageName: datagrid
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JohnSnowLABS
      provider-url: nlp.johnsnowlabs.com
    name: annotationlab
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: annotationlab.v5.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.johnsnowlabs.com/v1",
                "kind": "Annotationlab",
                "metadata": {
                  "name": "annotationlab"
                },
                "spec": {
                  "activelearning": {
                    "extraEnv": "",
                    "extraVolumeMounts": "",
                    "extraVolumes": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/johnsnowlabs/annotationlab",
                      "tag": "active-learning-5.7.0"
                    },
                    "lifecycleHooks": ""
                  },
                  "affinity": "podAntiAffinity:\n  requiredDuringSchedulingIgnoredDuringExecution:\n    - labelSelector:\n        matchLabels:\n          {{- include \"annotationlab.selectorLabels\" . | nindent 10 }}\n        matchExpressions:\n          - key: role\n            operator: NotIn\n            values:\n              - test\n      topologyKey: kubernetes.io/hostname\n  preferredDuringSchedulingIgnoredDuringExecution:\n    - weight: 100\n      podAffinityTerm:\n        labelSelector:\n          matchLabels:\n            {{- include \"annotationlab.selectorLabels\" . | nindent 12 }}\n          matchExpressions:\n            - key: role\n              operator: NotIn\n              values:\n                - test\n        topologyKey: failure-domain.beta.kubernetes.io/zone\n",
                  "airflow": {
                    "activelearning": {
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/johnsnowlabs/annotationlab",
                        "tag": "active-learning-5.7.0"
                      }
                    },
                    "affinity": {},
                    "airflowConfigAnnotations": {},
                    "airflowHome": "/opt/airflow",
                    "airflowLocalSettings": "{{- if semverCompare \"\u003e=2.2.0\" .Values.airflowVersion }}\n{{- if not (or .Values.webserverSecretKey .Values.webserverSecretKeySecretName) }}\nfrom airflow.www.utils import UIAlert\n\nDASHBOARD_UIALERTS = [\n  UIAlert(\n    'Usage of a dynamic webserver secret key detected. We recommend a static webserver secret key instead.'\n    ' See the \u003ca href='\n    '\"https://airflow.apache.org/docs/helm-chart/stable/production-guide.html#webserver-secret-key\"\u003e'\n    'Helm Chart Production Guide\u003c/a\u003e for more details.',\n    category=\"warning\",\n    roles=[\"Admin\"],\n    html=True,\n  )\n]\n{{- end }}\n{{- end }}",
                    "airflowPodAnnotations": {},
                    "airflowVersion": "2.5.1",
                    "allowPodLaunching": true,
                    "cleanup": {
                      "affinity": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow kubernetes cleanup-pods --namespace={{ .Release.Namespace }}"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "failedJobsHistoryLimit": null,
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "resources": {},
                      "schedule": "*/15 * * * *",
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "successfulJobsHistoryLimit": null,
                      "tolerations": [],
                      "topologySpreadConstraints": []
                    },
                    "config": {
                      "api": {
                        "auth_backend": "airflow.api.auth.backend.default"
                      },
                      "celery": {
                        "flower_url_prefix": "{{ .Values.ingress.flower.path }}",
                        "worker_concurrency": 16
                      },
                      "celery_kubernetes_executor": {
                        "kubernetes_queue": "kubernetes"
                      },
                      "core": {
                        "colored_console_log": "False",
                        "dags_are_paused_at_creation": "False",
                        "dags_folder": "{{ include \"airflow_dags\" . }}",
                        "executor": "{{ .Values.executor }}",
                        "load_examples": "False",
                        "remote_logging": "{{- ternary \"True\" \"False\" .Values.elasticsearch.enabled }}"
                      },
                      "elasticsearch": {
                        "json_format": "True",
                        "log_id_template": "{dag_id}_{task_id}_{execution_date}_{try_number}"
                      },
                      "elasticsearch_configs": {
                        "max_retries": 3,
                        "retry_timeout": "True",
                        "timeout": 30
                      },
                      "kerberos": {
                        "ccache": "{{ .Values.kerberos.ccacheMountPath }}/{{ .Values.kerberos.ccacheFileName }}",
                        "keytab": "{{ .Values.kerberos.keytabPath }}",
                        "principal": "{{ .Values.kerberos.principal }}",
                        "reinit_frequency": "{{ .Values.kerberos.reinitFrequency }}"
                      },
                      "kubernetes": {
                        "airflow_configmap": "{{ include \"airflow_config\" . }}",
                        "airflow_local_settings_configmap": "{{ include \"airflow_config\" . }}",
                        "multi_namespace_mode": "{{ ternary \"True\" \"False\" .Values.multiNamespaceMode }}",
                        "namespace": "{{ .Release.Namespace }}",
                        "pod_template_file": "{{ include \"airflow_pod_template_file\" . }}/pod_template_file.yaml",
                        "worker_container_repository": "{{ .Values.images.airflow.repository | default .Values.defaultAirflowRepository }}",
                        "worker_container_tag": "{{ .Values.images.airflow.tag | default .Values.defaultAirflowTag }}"
                      },
                      "logging": {
                        "colored_console_log": "False",
                        "remote_logging": "{{- ternary \"True\" \"False\" .Values.elasticsearch.enabled }}"
                      },
                      "metrics": {
                        "statsd_host": "{{ printf \"%s-statsd\" .Release.Name }}",
                        "statsd_on": "{{ ternary \"True\" \"False\" .Values.statsd.enabled }}",
                        "statsd_port": 9125,
                        "statsd_prefix": "airflow"
                      },
                      "scheduler": {
                        "run_duration": 41460,
                        "standalone_dag_processor": "{{ ternary \"True\" \"False\" .Values.dagProcessor.enabled }}",
                        "statsd_host": "{{ printf \"%s-statsd\" .Release.Name }}",
                        "statsd_on": "{{ ternary \"True\" \"False\" .Values.statsd.enabled }}",
                        "statsd_port": 9125,
                        "statsd_prefix": "airflow"
                      },
                      "webserver": {
                        "authenticate": "False",
                        "base_url": "http://0.0.0.0:8080",
                        "enable_proxy_fix": "True",
                        "rbac": "False"
                      }
                    },
                    "createUserJob": {
                      "affinity": {},
                      "annotations": {},
                      "applyCustomEnv": true,
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"users create\" \"create_user\" }} \"$@\"",
                        "--",
                        "-r",
                        "{{ .Values.webserver.defaultUser.role }}",
                        "-u",
                        "{{ .Values.webserver.defaultUser.username }}",
                        "-e",
                        "{{ .Values.webserver.defaultUser.email }}",
                        "-f",
                        "{{ .Values.webserver.defaultUser.firstName }}",
                        "-l",
                        "{{ .Values.webserver.defaultUser.lastName }}",
                        "-p",
                        "{{ .Values.webserver.defaultUser.password }}"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "jobAnnotations": {},
                      "labels": {},
                      "nodeSelector": {},
                      "resources": {},
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-create-user-job"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "useHelmHooks": true
                    },
                    "dagProcessor": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow dag-processor"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 60,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "dags": {
                      "gitSync": {
                        "branch": "v2-2-stable",
                        "containerName": "git-sync",
                        "depth": 1,
                        "enabled": false,
                        "env": [],
                        "extraVolumeMounts": [],
                        "maxFailures": 0,
                        "repo": "https://github.com/apache/airflow.git",
                        "resources": {},
                        "rev": "HEAD",
                        "securityContext": {},
                        "subPath": "tests/dags",
                        "uid": 65533,
                        "wait": 5
                      },
                      "persistence": {
                        "accessMode": "ReadWriteOnce",
                        "enabled": false,
                        "existingClaim": null,
                        "size": "1Gi",
                        "storageClassName": null,
                        "subPath": null
                      }
                    },
                    "data": {
                      "brokerUrl": null,
                      "brokerUrlSecretName": null,
                      "metadataConnection": {
                        "db": "airflow",
                        "pass": "airflow",
                        "port": 5432,
                        "protocol": "postgresql",
                        "sslmode": "disable",
                        "user": "airflow"
                      },
                      "metadataSecretName": null,
                      "resultBackendConnection": null,
                      "resultBackendSecretName": null
                    },
                    "defaultAirflowRepository": "apache/airflow",
                    "defaultAirflowTag": "2.5.1",
                    "elasticsearch": {
                      "connection": {},
                      "enabled": false,
                      "secretName": null
                    },
                    "enableBuiltInSecretEnvVars": {
                      "AIRFLOW_CONN_AIRFLOW_DB": true,
                      "AIRFLOW__CELERY__BROKER_URL": true,
                      "AIRFLOW__CELERY__CELERY_RESULT_BACKEND": true,
                      "AIRFLOW__CELERY__RESULT_BACKEND": true,
                      "AIRFLOW__CORE__FERNET_KEY": true,
                      "AIRFLOW__CORE__SQL_ALCHEMY_CONN": true,
                      "AIRFLOW__DATABASE__SQL_ALCHEMY_CONN": true,
                      "AIRFLOW__ELASTICSEARCH__ELASTICSEARCH_HOST": true,
                      "AIRFLOW__ELASTICSEARCH__HOST": true,
                      "AIRFLOW__WEBSERVER__SECRET_KEY": true
                    },
                    "env": [],
                    "executor": "CeleryExecutor",
                    "extraConfigMaps": {},
                    "extraEnv": "- name: ACTIVE_LEARNING_IMAGE_PULL_POLICY\n  value: \"{{ .Values.activelearning.image.pullPolicy }}\"\n- name: ACTIVE_LEARNING_IMAGE\n  value: \"{{ .Values.activelearning.image.repository }}:{{ .Values.activelearning.image.tag }}\"\n- name: NUMBER_OF_PREANNOTATOR_DAG\n  value: {{ .Values.preannotator_queues.count | quote }}\n- name: AIRFLOW__LOGGING__LOGGING_LEVEL\n  value: INFO\n- name: ANNOTATIONLAB_FULLNAME\n  value: {{ include \"annotationlab.annotationlab.fullname\" . }}\n- name: ANNOTATIONLAB_URL\n  value: http://{{ include \"annotationlab.annotationlab.fullname\" . }}:8200\n- name: AIRFLOW_NAMESPACE\n  value: {{ .Release.Namespace }}\n- name: DATABASE_CONNECTION_STRING\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"annotationlab.annotationlab.fullname\" . }}-secret\n      key: DATABASE_CONNECTION_STRING\n",
                    "extraEnvFrom": null,
                    "extraSecrets": {},
                    "fernetKey": null,
                    "fernetKeySecretName": null,
                    "flower": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"celery flower\" \"flower\" }}"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "networkPolicy": {
                        "ingress": {
                          "from": [],
                          "ports": [
                            {
                              "port": "{{ .Values.ports.flowerUI }}"
                            }
                          ]
                        }
                      },
                      "nodeSelector": {},
                      "password": null,
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "secretName": null,
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "loadBalancerIP": null,
                        "loadBalancerSourceRanges": [],
                        "ports": [
                          {
                            "name": "flower-ui",
                            "port": "{{ .Values.ports.flowerUI }}"
                          }
                        ],
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "username": null
                    },
                    "fullnameOverride": "",
                    "gid": 0,
                    "global": {},
                    "image": {
                      "repository": "docker.io/johnsnowlabs/annotationlab"
                    },
                    "images": {
                      "airflow": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": "dataflows-5.7.0"
                      },
                      "flower": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": null
                      },
                      "gitSync": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "k8s.gcr.io/git-sync/git-sync",
                        "tag": "v3.6.3"
                      },
                      "migrationsWaitTimeout": 60,
                      "pgbouncer": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "apache/airflow",
                        "tag": "airflow-pgbouncer-2021.04.28-1.14.0"
                      },
                      "pgbouncerExporter": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "apache/airflow",
                        "tag": "airflow-pgbouncer-exporter-2021.09.22-0.12.0"
                      },
                      "pod_template": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": null
                      },
                      "redis": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/redis",
                        "tag": "alpine3.16"
                      },
                      "statsd": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "quay.io/prometheus/statsd-exporter",
                        "tag": "v0.22.8"
                      },
                      "useDefaultImageForMigration": false
                    },
                    "ingress": {
                      "enabled": false,
                      "flower": {
                        "annotations": {},
                        "enabled": false,
                        "host": "",
                        "hosts": [],
                        "ingressClassName": "",
                        "path": "/",
                        "pathType": "ImplementationSpecific",
                        "tls": {
                          "enabled": false,
                          "secretName": ""
                        }
                      },
                      "web": {
                        "annotations": {},
                        "enabled": false,
                        "host": "",
                        "hosts": [],
                        "ingressClassName": "",
                        "path": "/",
                        "pathType": "ImplementationSpecific",
                        "precedingPaths": [],
                        "succeedingPaths": [],
                        "tls": {
                          "enabled": false,
                          "secretName": ""
                        }
                      }
                    },
                    "kerberos": {
                      "ccacheFileName": "cache",
                      "ccacheMountPath": "/var/kerberos-ccache",
                      "config": "# This is an example config showing how you can use templating and how \"example\" config\n# might look like. It works with the test kerberos server that we are using during integration\n# testing at Apache Airflow (see `scripts/ci/docker-compose/integration-kerberos.yml` but in\n# order to make it production-ready you must replace it with your own configuration that\n# Matches your kerberos deployment. Administrators of your Kerberos instance should\n# provide the right configuration.\n\n[logging]\ndefault = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kerberos_libs.log\"\nkdc = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kerberos_kdc.log\"\nadmin_server = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kadmind.log\"\n\n[libdefaults]\ndefault_realm = FOO.COM\nticket_lifetime = 10h\nrenew_lifetime = 7d\nforwardable = true\n\n[realms]\nFOO.COM = {\n  kdc = kdc-server.foo.com\n  admin_server = admin_server.foo.com\n}\n",
                      "configPath": "/etc/krb5.conf",
                      "enabled": false,
                      "keytabBase64Content": null,
                      "keytabPath": "/etc/airflow.keytab",
                      "principal": "airflow@FOO.COM",
                      "reinitFrequency": 3600
                    },
                    "kubeVersionOverride": "",
                    "labels": {},
                    "limits": [],
                    "logs": {
                      "persistence": {
                        "enabled": false,
                        "existingClaim": null,
                        "size": "100Gi",
                        "storageClassName": null
                      }
                    },
                    "migrateDatabaseJob": {
                      "affinity": {},
                      "annotations": {},
                      "applyCustomEnv": true,
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"db upgrade\" \"upgradedb\" }}"
                      ],
                      "command": null,
                      "enabled": true,
                      "extraContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "jobAnnotations": {},
                      "nodeSelector": {},
                      "resources": {},
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-migrate-database-job"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "useHelmHooks": true
                    },
                    "multiNamespaceMode": false,
                    "nameOverride": "",
                    "networkPolicies": {
                      "enabled": false
                    },
                    "nodeSelector": {},
                    "pgbouncer": {
                      "affinity": {},
                      "annotations": {},
                      "args": null,
                      "auth_file": "/etc/pgbouncer/users.txt",
                      "auth_type": "md5",
                      "ciphers": "normal",
                      "command": [
                        "pgbouncer",
                        "-u",
                        "nobody",
                        "/etc/pgbouncer/pgbouncer.ini"
                      ],
                      "configSecretName": null,
                      "enabled": false,
                      "extraIni": null,
                      "extraIniMetadata": null,
                      "extraIniResultBackend": null,
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "logConnections": 0,
                      "logDisconnections": 0,
                      "maxClientConn": 100,
                      "metadataPoolSize": 10,
                      "metricsExporterSidecar": {
                        "resources": {},
                        "sslmode": "disable"
                      },
                      "nodeSelector": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "resultBackendPoolSize": 5,
                      "revisionHistoryLimit": null,
                      "service": {
                        "extraAnnotations": {}
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "ssl": {
                        "ca": null,
                        "cert": null,
                        "key": null
                      },
                      "sslmode": "prefer",
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 65534,
                      "verbose": 0
                    },
                    "podTemplate": null,
                    "ports": {
                      "airflowUI": 8080,
                      "flowerUI": 5555,
                      "pgbouncer": 6543,
                      "pgbouncerScrape": 9127,
                      "redisDB": 6379,
                      "statsdIngest": 9125,
                      "statsdScrape": 9102,
                      "workerLogs": 8793
                    },
                    "postgresql": {
                      "auth": {
                        "enablePostgresUser": true,
                        "password": null,
                        "postgresPassword": "postgres",
                        "username": null
                      },
                      "common": {
                        "exampleValue": "common-chart",
                        "global": {
                          "postgresql": {}
                        }
                      },
                      "commonAnnotations": {},
                      "enabled": true,
                      "extraEnv": [],
                      "global": {
                        "postgresql": {}
                      },
                      "image": {
                        "debug": false,
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgresql",
                        "tag": "11"
                      },
                      "ldap": {
                        "baseDN": "",
                        "bindDN": "",
                        "enabled": false,
                        "port": "",
                        "prefix": "",
                        "scheme": "",
                        "search_attr": "",
                        "search_filter": "",
                        "server": "",
                        "suffix": "",
                        "tls": false,
                        "url": ""
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 30,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "master": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": [],
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "metrics": {
                        "enabled": false,
                        "extraEnvVars": {},
                        "image": {
                          "pullPolicy": "IfNotPresent",
                          "registry": "docker.io",
                          "repository": "bitnami/postgres-exporter",
                          "tag": "0.8.0-debian-10-r166"
                        },
                        "livenessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "prometheusRule": {
                          "additionalLabels": {},
                          "enabled": false,
                          "namespace": "",
                          "rules": []
                        },
                        "readinessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "securityContext": {},
                        "service": {
                          "annotations": {
                            "prometheus.io/port": "9187",
                            "prometheus.io/scrape": "true"
                          },
                          "type": "ClusterIP"
                        },
                        "serviceMonitor": {
                          "additionalLabels": {},
                          "enabled": false
                        }
                      },
                      "nameOverride": "airflow-postgresql",
                      "networkPolicy": {
                        "allowExternal": true,
                        "enabled": false,
                        "explicitNamespacesSelector": {}
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "enabled": true,
                        "mountPath": "/bitnami/postgresql",
                        "size": "8Gi",
                        "subPath": ""
                      },
                      "postgresqlDataDir": "/bitnami/postgresql/data",
                      "postgresqlDatabase": "airflow",
                      "postgresqlPassword": "airflow",
                      "postgresqlUsername": "airflow",
                      "psp": {
                        "create": false
                      },
                      "rbac": {
                        "create": false
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "replication": {
                        "applicationName": "my_application",
                        "enabled": false,
                        "numSynchronousReplicas": 0,
                        "password": "repl_password",
                        "slaveReplicas": 1,
                        "synchronousCommit": "off",
                        "user": "repl_user"
                      },
                      "resources": {
                        "requests": {
                          "cpu": "250m",
                          "memory": "256Mi"
                        }
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "port": 5432,
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "enabled": true,
                        "name": "annotationlab-airflow-postgresql"
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": true
                        },
                        "enabled": true
                      },
                      "slave": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "tls": {
                        "certFilename": "",
                        "certKeyFilename": "",
                        "certificatesSecret": "",
                        "enabled": false,
                        "preferServerCiphers": true
                      },
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "volumePermissions": {
                        "enabled": false,
                        "image": {
                          "pullPolicy": "Always",
                          "registry": "docker.io",
                          "repository": "bitnami/minideb",
                          "tag": "buster"
                        },
                        "securityContext": {}
                      }
                    },
                    "preannotator_queues": {
                      "count": 5
                    },
                    "quotas": {},
                    "rbac": {
                      "create": true,
                      "createSCCRoleBinding": false
                    },
                    "redis": {
                      "affinity": {},
                      "enabled": true,
                      "nodeSelector": {},
                      "password": "DefaulTRediS-P0ss",
                      "passwordSecretName": null,
                      "persistence": {
                        "annotations": {},
                        "enabled": true,
                        "size": "1Gi",
                        "storageClassName": null
                      },
                      "podAnnotations": {},
                      "resources": {},
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "annotationlab-airflow-redis"
                      },
                      "terminationGracePeriodSeconds": 600,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 0
                    },
                    "registry": {
                      "connection": {},
                      "secretName": "kubernetesdockerhub"
                    },
                    "registryCredentials": {
                      "pullPolicy": "IfNotPresent",
                      "registryUrl": "docker.io/annotationlab"
                    },
                    "revisionHistoryLimit": null,
                    "scheduler": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow scheduler"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "logGroomerSidecar": {
                        "args": [
                          "bash",
                          "/clean-logs"
                        ],
                        "command": null,
                        "enabled": true,
                        "resources": {},
                        "retentionDays": 15
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-scheduler"
                      },
                      "strategy": null,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "updateStrategy": null,
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "secret": [],
                    "securityContext": {},
                    "statsd": {
                      "affinity": {},
                      "args": [
                        "--statsd.mapping-config=/etc/statsd-exporter/mappings.yml"
                      ],
                      "enabled": false,
                      "extraMappings": [],
                      "extraNetworkPolicies": [],
                      "nodeSelector": {},
                      "overrideMappings": [],
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "securityContext": {},
                      "service": {
                        "extraAnnotations": {}
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-statsd"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 65534
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "triggerer": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow triggerer"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "annotationlab-airflow-triggerer"
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 60,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "uid": 50000,
                    "volumeMounts": [],
                    "volumes": [],
                    "webserver": {
                      "affinity": {},
                      "allowPodLogReading": true,
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow webserver"
                      ],
                      "command": null,
                      "defaultUser": {
                        "email": "admin@example.com",
                        "enabled": false,
                        "firstName": "admin",
                        "lastName": "user",
                        "password": "admin",
                        "role": "Admin",
                        "username": "admin"
                      },
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "failureThreshold": 20,
                        "initialDelaySeconds": 15,
                        "periodSeconds": 5,
                        "scheme": "HTTP",
                        "timeoutSeconds": 30
                      },
                      "networkPolicy": {
                        "ingress": {
                          "from": [],
                          "ports": [
                            {
                              "port": "{{ .Values.ports.airflowUI }}"
                            }
                          ]
                        }
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "readinessProbe": {
                        "failureThreshold": 20,
                        "initialDelaySeconds": 15,
                        "periodSeconds": 5,
                        "scheme": "HTTP",
                        "timeoutSeconds": 30
                      },
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "loadBalancerIP": null,
                        "loadBalancerSourceRanges": [],
                        "ports": [
                          {
                            "name": "airflow-ui",
                            "port": "{{ .Values.ports.airflowUI }}"
                          }
                        ],
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-webserver"
                      },
                      "strategy": null,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      },
                      "webserverConfig": null,
                      "webserverConfigConfigMapName": null
                    },
                    "webserverSecretKey": "f565cd64401775f93c820dc63fc3d4f6",
                    "webserverSecretKeySecretName": null,
                    "workers": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"celery worker\" \"worker\" }}"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "hostAliases": [],
                      "keda": {
                        "advanced": {},
                        "cooldownPeriod": 30,
                        "enabled": false,
                        "maxReplicaCount": 10,
                        "minReplicaCount": 0,
                        "namespaceLabels": {},
                        "pollingInterval": 5
                      },
                      "kerberosSidecar": {
                        "enabled": false,
                        "resources": {}
                      },
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "enabled": true,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "logGroomerSidecar": {
                        "args": [
                          "bash",
                          "/clean-logs"
                        ],
                        "command": null,
                        "enabled": true,
                        "resources": {},
                        "retentionDays": 15
                      },
                      "nodeSelector": {},
                      "persistence": {
                        "annotations": {},
                        "enabled": true,
                        "fixPermissions": false,
                        "size": "5Gi",
                        "storageClassName": null
                      },
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-worker"
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 600,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "updateStrategy": null,
                      "waitForMigrations": {
                        "env": []
                      }
                    }
                  },
                  "backup": {
                    "enable": false,
                    "files": true,
                    "image": {
                      "repository": null,
                      "tag": null
                    },
                    "s3_access_key": "",
                    "s3_bucket_fullpath": "",
                    "s3_secret_key": ""
                  },
                  "configuration": {
                    "FLASK_SECRET_KEY": "rpaQghA-9lmpLwB8qxW0FsW1TJyZZYuB60ppzswpTzM=",
                    "KEYCLOAK_CLIENT_ID": "annotationlab",
                    "KEYCLOAK_CLIENT_SECRET_KEY": "09a71c59-0351-4ce6-bc8f-8fd3feb9d2ff",
                    "KEYCLOAK_REALM_NAME": "master",
                    "PERSISTENT_STORAGE": "/projects"
                  },
                  "extraContainers": "",
                  "extraEnv": "",
                  "extraNetworkPolicies": "- namespaceSelector: {}\n",
                  "extraVolumeMounts": "",
                  "extraVolumes": "",
                  "github_oauth_origin": "https://github.oauth.johnsnowlabs.com",
                  "google_analytics": {
                    "api_secret": "WENfd00yeUxRRWl4N0Utbml5SWt3Zw==",
                    "measurement_id": "Ry1XODBIMDhTTFBZ"
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "docker.io/johnsnowlabs/annotationlab",
                    "tag": "5.7.0"
                  },
                  "ingress": {
                    "annotations": {},
                    "defaultBackend": false,
                    "enabled": true,
                    "hosts": [
                      {
                        "host": "annotationlab",
                        "path": "/"
                      }
                    ],
                    "ingressroute": {
                      "enabled": false
                    },
                    "servicePort": "http",
                    "tls": [],
                    "uploadLimitInMegabytes": 16
                  },
                  "installAirflow": true,
                  "installKeycloak": true,
                  "installPostgresql": true,
                  "keycloak": {
                    "affinity": "podAntiAffinity:\n  requiredDuringSchedulingIgnoredDuringExecution:\n    - labelSelector:\n        matchLabels:\n          {{- include \"keycloak.selectorLabels\" . | nindent 10 }}\n        matchExpressions:\n          - key: app.kubernetes.io/component\n            operator: NotIn\n            values:\n              - test\n      topologyKey: kubernetes.io/hostname\n  preferredDuringSchedulingIgnoredDuringExecution:\n    - weight: 100\n      podAffinityTerm:\n        labelSelector:\n          matchLabels:\n            {{- include \"keycloak.selectorLabels\" . | nindent 12 }}\n          matchExpressions:\n            - key: app.kubernetes.io/component\n              operator: NotIn\n              values:\n                - test\n        topologyKey: topology.kubernetes.io/zone\n",
                    "args": [],
                    "auththeme": {
                      "image": {
                        "repository": "docker.io/johnsnowlabs/annotationlab",
                        "tag": "auth-theme-5.7.0"
                      }
                    },
                    "autoscaling": {
                      "behavior": {
                        "scaleDown": {
                          "policies": [
                            {
                              "periodSeconds": 300,
                              "type": "Pods",
                              "value": 1
                            }
                          ],
                          "stabilizationWindowSeconds": 300
                        }
                      },
                      "enabled": false,
                      "labels": {},
                      "maxReplicas": 10,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 80,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 3
                    },
                    "cache": {
                      "stack": "default"
                    },
                    "clusterDomain": "cluster.local",
                    "command": [
                      "/opt/keycloak/bin/kc.sh",
                      "--verbose",
                      "start",
                      "--http-enabled=true",
                      "--http-port=8080",
                      "--hostname-strict=false",
                      "--hostname-strict-https=false",
                      "--spi-events-listener-jboss-logging-success-level=info",
                      "--spi-events-listener-jboss-logging-error-level=warn"
                    ],
                    "database": {
                      "enabled": true,
                      "postgresqlDatabase": "keycloak",
                      "postgresqlPassword": "keycloak",
                      "postgresqlUsername": "keycloak"
                    },
                    "dbchecker": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/busybox",
                        "tag": 1.32
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "32Mi"
                        },
                        "requests": {
                          "cpu": "20m",
                          "memory": "32Mi"
                        }
                      },
                      "securityContext": {}
                    },
                    "enableServiceLinks": true,
                    "extraContainers": "",
                    "extraEnv": "- name: KC_FEATURES\n  value: \"authorization,account-api,admin-api,admin-fine-grained-authz,ciba,impersonation,kerberos,preview,step-up-authentication,token-exchange,web-authn\"\n- name: KEYCLOAK_ADMIN\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"keycloak.fullname\" . }}-admincreds\n      key: user\n- name: KEYCLOAK_ADMIN_PASSWORD\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"keycloak.fullname\" . }}-admincreds\n      key: password\n- name: JAVA_OPTS_APPEND\n  value: \u003e-\n    -XX:+UseContainerSupport\n    -XX:MaxRAMPercentage=50.0\n    -Djava.awt.headless=true\n    -Djgroups.dns.query={{ include \"keycloak.fullname\" . }}-headless\n",
                    "extraEnvFrom": "",
                    "extraInitContainers": "- name: theme-provider\n  image: {{ .Values.auththeme.image.repository }}:{{ .Values.auththeme.image.tag }}\n  imagePullPolicy: IfNotPresent\n  command:\n    - sh\n  args:\n    - -c\n    - |\n      echo \"Copying theme...\"\n      cp -R /johnsnowlabs/* /theme\n  volumeMounts:\n    - name: theme\n      mountPath: /theme\n",
                    "extraPorts": [],
                    "extraServiceMonitor": {
                      "annotations": {},
                      "enabled": false,
                      "interval": "10s",
                      "labels": {},
                      "namespace": "",
                      "namespaceSelector": {},
                      "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/realms/master/metrics",
                      "port": "http",
                      "scrapeTimeout": "10s"
                    },
                    "extraVolumeMounts": "- name: admincreds\n  mountPath: /secrets/admincreds\n  readOnly: true\n- name: theme\n  mountPath: /opt/keycloak/themes/johnsnowlabs\n",
                    "extraVolumes": "- name: admincreds\n  secret:\n    secretName: '{{ include \"annotationlab.keycloak.fullname\" . }}-admincreds'\n- name: theme\n  emptyDir: {}\n",
                    "fullnameOverride": "",
                    "global": {},
                    "health": {
                      "enabled": true
                    },
                    "hostAliases": [],
                    "http": {
                      "relativePath": "/auth"
                    },
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/keycloak/keycloak",
                      "tag": "20.0.3"
                    },
                    "imagePullSecrets": [
                      {
                        "name": "kubernetesdockerhub"
                      }
                    ],
                    "ingress": {
                      "annotations": {
                        "nginx.ingress.kubernetes.io/proxy-buffer-size": "16k"
                      },
                      "console": {
                        "annotations": {},
                        "enabled": false,
                        "ingressClassName": "",
                        "rules": [
                          {
                            "host": "{{ .Release.Name }}.keycloak.example.com",
                            "paths": [
                              {
                                "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/admin",
                                "pathType": "Prefix"
                              }
                            ]
                          }
                        ],
                        "tls": []
                      },
                      "enabled": false,
                      "ingressClassName": "",
                      "labels": {},
                      "rules": [
                        {
                          "host": "keycloak.example.com",
                          "paths": [
                            {
                              "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/",
                              "pathType": "Prefix"
                            }
                          ]
                        }
                      ],
                      "servicePort": "http",
                      "tls": []
                    },
                    "lifecycleHooks": "",
                    "livenessProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health/live'\n  port: http\ninitialDelaySeconds: 0\ntimeoutSeconds: 5\n",
                    "metrics": {
                      "enabled": true
                    },
                    "nameOverride": "",
                    "networkPolicy": {
                      "egress": [],
                      "enabled": false,
                      "extraFrom": [],
                      "labels": {}
                    },
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podDisruptionBudget": {},
                    "podLabels": {},
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {},
                    "postgresql": {
                      "common": {
                        "exampleValue": "common-chart",
                        "global": {
                          "postgresql": {}
                        }
                      },
                      "commonAnnotations": {},
                      "enabled": true,
                      "extraEnv": [],
                      "global": {
                        "postgresql": {}
                      },
                      "image": {
                        "debug": false,
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgresql",
                        "tag": "11.20.0-debian-11-r12"
                      },
                      "ldap": {
                        "baseDN": "",
                        "bindDN": "",
                        "bind_password": null,
                        "enabled": false,
                        "port": "",
                        "prefix": "",
                        "scheme": "",
                        "search_attr": "",
                        "search_filter": "",
                        "server": "",
                        "suffix": "",
                        "tls": false,
                        "url": ""
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 30,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "master": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": [],
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "metrics": {
                        "enabled": false,
                        "extraEnvVars": {},
                        "image": {
                          "pullPolicy": "IfNotPresent",
                          "registry": "docker.io",
                          "repository": "bitnami/postgres-exporter",
                          "tag": "0.8.0-debian-10-r166"
                        },
                        "livenessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "prometheusRule": {
                          "additionalLabels": {},
                          "enabled": false,
                          "namespace": "",
                          "rules": []
                        },
                        "readinessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "securityContext": {},
                        "service": {
                          "annotations": {
                            "prometheus.io/port": "9187",
                            "prometheus.io/scrape": "true"
                          },
                          "loadBalancerIP": null,
                          "type": "ClusterIP"
                        },
                        "serviceMonitor": {
                          "additionalLabels": {},
                          "enabled": false
                        }
                      },
                      "nameOverride": "keycloak-postgres",
                      "networkPolicy": {
                        "allowExternal": true,
                        "enabled": false,
                        "explicitNamespacesSelector": {}
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "enabled": true,
                        "mountPath": "/bitnami/postgresql",
                        "size": "8Gi",
                        "subPath": ""
                      },
                      "postgresqlDataDir": "/bitnami/postgresql/data",
                      "postgresqlDatabase": "keycloak",
                      "postgresqlPassword": "keycloak",
                      "postgresqlUsername": "keycloak",
                      "psp": {
                        "create": false
                      },
                      "rbac": {
                        "create": false
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "replication": {
                        "applicationName": "my_application",
                        "enabled": false,
                        "numSynchronousReplicas": 0,
                        "password": "repl_password",
                        "slaveReplicas": 1,
                        "synchronousCommit": "off",
                        "user": "repl_user"
                      },
                      "resources": {
                        "requests": {
                          "cpu": "250m",
                          "memory": "256Mi"
                        }
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "port": 5432,
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "enabled": true,
                        "name": "annotationlab-keycloak-postgresql"
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": true
                        },
                        "enabled": true
                      },
                      "slave": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "tls": {
                        "certCAFilename": null,
                        "certFilename": "",
                        "certKeyFilename": "",
                        "certificatesSecret": "",
                        "crlFilename": null,
                        "enabled": false,
                        "preferServerCiphers": true
                      },
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "volumePermissions": {
                        "enabled": false,
                        "image": {
                          "pullPolicy": "Always",
                          "registry": "docker.io",
                          "repository": "bitnami/minideb",
                          "tag": "buster"
                        },
                        "securityContext": {}
                      }
                    },
                    "priorityClassName": "",
                    "prometheusRule": {
                      "annotations": {},
                      "enabled": false,
                      "labels": {},
                      "namespace": "",
                      "rules": []
                    },
                    "proxy": {
                      "enabled": true,
                      "mode": "edge"
                    },
                    "rbac": {
                      "create": false,
                      "rules": []
                    },
                    "readinessProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health/ready'\n  port: http\ninitialDelaySeconds: 10\ntimeoutSeconds: 1\n",
                    "replicas": 1,
                    "resources": {},
                    "restartPolicy": "Always",
                    "route": {
                      "annotations": {},
                      "enabled": false,
                      "host": "",
                      "labels": {},
                      "path": "/",
                      "tls": {
                        "enabled": true,
                        "insecureEdgeTerminationPolicy": "Redirect",
                        "termination": "edge"
                      }
                    },
                    "secrets": {
                      "admincreds": {
                        "stringData": {
                          "password": "secret",
                          "user": "admin"
                        }
                      }
                    },
                    "securityContext": {},
                    "service": {
                      "annotations": {},
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "httpNodePort": null,
                      "httpPort": 80,
                      "httpsNodePort": null,
                      "httpsPort": 8443,
                      "labels": {},
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "sessionAffinity": "",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "serviceAccount": {
                      "allowReadPods": false,
                      "annotations": {},
                      "automountServiceAccountToken": true,
                      "create": false,
                      "imagePullSecrets": [],
                      "labels": {},
                      "name": "annotationlab-keyclo"
                    },
                    "serviceHeadless": {
                      "annotations": {}
                    },
                    "serviceMonitor": {
                      "annotations": {},
                      "enabled": false,
                      "interval": "10s",
                      "labels": {},
                      "namespace": "",
                      "namespaceSelector": {},
                      "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/metrics",
                      "port": "http",
                      "scrapeTimeout": "10s"
                    },
                    "skipInitContainers": false,
                    "startupProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health'\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 1\nfailureThreshold: 60\nperiodSeconds: 5\n",
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {},
                    "terminationGracePeriodSeconds": 60,
                    "test": {
                      "deletionPolicy": "before-hook-creation",
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/seleniarm/standalone-chromium",
                        "tag": "111.0"
                      },
                      "podSecurityContext": {},
                      "securityContext": {}
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": null,
                    "updateStrategy": "RollingUpdate"
                  },
                  "licenseserver": {
                    "origin": "https://license.johnsnowlabs.com"
                  },
                  "lifecycleHooks": "# postStart:\n#   exec:\n#     command: [\"/bin/sh\", \"-c\", \"ls\"]\n",
                  "modelshub_origin": "https://modelshub.johnsnowlabs.com",
                  "myjohnsnowlabs": {
                    "client_id": "UIGgoiIL9szXW8Nnik0O7tPT4to87oVtWvYWfc6r",
                    "origin": "https://my.johnsnowlabs.com"
                  },
                  "networkPolicy": {
                    "enabled": false
                  },
                  "nodeSelector": {},
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "annotations": {},
                    "enabled": true,
                    "size": "8Gi"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "postgresql": {
                    "common": {
                      "exampleValue": "common-chart",
                      "global": {
                        "postgresql": {}
                      }
                    },
                    "commonAnnotations": {},
                    "extraEnv": [],
                    "global": {
                      "postgresql": {}
                    },
                    "image": {
                      "debug": false,
                      "pullPolicy": "IfNotPresent",
                      "registry": "docker.io",
                      "repository": "bitnami/postgresql",
                      "tag": "11.20.0-debian-11-r12"
                    },
                    "ldap": {
                      "baseDN": "",
                      "bindDN": "",
                      "bind_password": null,
                      "enabled": false,
                      "port": "",
                      "prefix": "",
                      "scheme": "",
                      "search_attr": "",
                      "search_filter": "",
                      "server": "",
                      "suffix": "",
                      "tls": false,
                      "url": ""
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "master": {
                      "affinity": {},
                      "annotations": {},
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podLabels": {},
                      "priorityClassName": "",
                      "service": {},
                      "sidecars": [],
                      "tolerations": []
                    },
                    "metrics": {
                      "enabled": false,
                      "extraEnvVars": {},
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgres-exporter",
                        "tag": "0.8.0-debian-10-r166"
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "prometheusRule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {
                          "prometheus.io/port": "9187",
                          "prometheus.io/scrape": "true"
                        },
                        "loadBalancerIP": null,
                        "type": "ClusterIP"
                      },
                      "serviceMonitor": {
                        "additionalLabels": {},
                        "enabled": false
                      }
                    },
                    "networkPolicy": {
                      "allowExternal": true,
                      "enabled": false,
                      "explicitNamespacesSelector": {}
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {},
                      "enabled": true,
                      "mountPath": "/bitnami/postgresql",
                      "size": "8Gi",
                      "subPath": ""
                    },
                    "postgresqlDataDir": "/bitnami/postgresql/data",
                    "postgresqlDatabase": "annotationlab",
                    "postgresqlPassword": "annotationlab",
                    "postgresqlUsername": "annotationlab",
                    "psp": {
                      "create": false
                    },
                    "rbac": {
                      "create": false
                    },
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 6,
                      "initialDelaySeconds": 5,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "replication": {
                      "applicationName": "my_application",
                      "enabled": false,
                      "numSynchronousReplicas": 0,
                      "password": "repl_password",
                      "slaveReplicas": 1,
                      "synchronousCommit": "off",
                      "user": "repl_user"
                    },
                    "resources": {
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {},
                    "service": {
                      "annotations": {},
                      "port": 5432,
                      "type": "ClusterIP"
                    },
                    "serviceAccount": {
                      "enabled": false
                    },
                    "shmVolume": {
                      "chmod": {
                        "enabled": true
                      },
                      "enabled": true
                    },
                    "slave": {
                      "affinity": {},
                      "annotations": {},
                      "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podLabels": {},
                      "priorityClassName": "",
                      "service": {},
                      "sidecars": [],
                      "tolerations": []
                    },
                    "tls": {
                      "certCAFilename": null,
                      "certFilename": "",
                      "certKeyFilename": "",
                      "certificatesSecret": "",
                      "crlFilename": null,
                      "enabled": false,
                      "preferServerCiphers": true
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "volumePermissions": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "Always",
                        "registry": "docker.io",
                        "repository": "bitnami/minideb",
                        "tag": "buster"
                      },
                      "securityContext": {}
                    }
                  },
                  "preannotator_queues": {
                    "count": 5,
                    "service": {
                      "port": 5000,
                      "portName": "http",
                      "type": "ClusterIP"
                    }
                  },
                  "priorityClassName": "",
                  "readinessProbe": "httpGet:\n  path: /healthz\n  port: http\ninitialDelaySeconds: 10\ntimeoutSeconds: 1\n",
                  "registryCredentials": {
                    "create": false,
                    "password": "XXXXX",
                    "secretName": "kubernetesdockerhub",
                    "username": "XXXXX"
                  },
                  "replicas": 1,
                  "resources": {},
                  "restartPolicy": "Always",
                  "securityContext": {
                    "enabled": false,
                    "fsGroup": 65534,
                    "fsGroupChangePolicy": "OnRootMismatch",
                    "runAsUser": 65534
                  },
                  "service": {
                    "port": 8200,
                    "portName": "http",
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": false,
                    "name": "annotationlab"
                  },
                  "sharedData": {
                    "enabled": true,
                    "images": {
                      "storageSize": "5Gi"
                    },
                    "storageSize": "100Gi"
                  },
                  "sqlalchemy_pool_size": 10,
                  "telemetry": {
                    "enabled": true
                  },
                  "terminationGracePeriodSeconds": 60,
                  "tolerations": [],
                  "updateStrategy": {
                    "type": "Recreate"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.johnsnowlabs.com/annotationlab/annotationlab-operator:v5.7.0
          createdAt: "2023-12-19T10:26:54Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/annotationlab/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/annotationlab/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: johnsnowlabs-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: NLP Lab
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Annotationlab
            name: annotationlabs.apps.johnsnowlabs.com
            version: v1
        description: "NLP Lab is a data labeling and annotation tool that helps you\nlabel
          different types of data using a simple interface with a standardized output\nformat.\n\nIt
          contains enterprise features like Project Management and sharing, SSO\nintegration,
          OpenAPI Specification 3.0 compliant API, task management, and active\nlearning/preannotation
          using Spark NLP.\n\nFurthermore, the production deployment of the application
          is done using a helm chart.\nThe same helm chart is also used to create
          an AWS AMI for an EC2 based single server \ndeployment using packer.\n\nMake
          sure you have defined shared storage such as efs/nfs/cephfs prior to installing
          the NLP Lab Operator.\nUpdate the host property in the YAML configuration
          to define the required domain name to use instead of the \ndefault hostname
          annotationlab.\n\nNLP Lab Operator Installtion guide [NLP Lab Operator][]:\n[NLP
          Lab Operator]: https://nlp.johnsnowlabs.com/docs/en/alab/install    \n"
        displayName: NLPLab
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nlplab
        links:
        - name: Annotationlab
          url: https://annotationlab.domain
        maintainers:
        - email: bhandari@johnsnowlabs.com
          name: bikash
        maturity: alpha
        provider:
          name: JohnSnowLABS
          url: nlp.johnsnowlabs.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - registry.johnsnowlabs.com/annotationlab/annotationlab-operator:v5.7.0
        version: 5.7.0
      entries:
      - name: annotationlab.v5.7.0
        version: 5.7.0
      - name: annotationlab.v5.4.1
        version: 5.4.1
      - name: annotationlab.v4.9.2
        version: 4.9.2
      name: alpha
    defaultChannel: alpha
    packageName: annotationlab
    provider:
      name: JohnSnowLABS
      url: nlp.johnsnowlabs.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.6.5-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6.5
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/kubeturbo-operator:8.6.5
        version: 8.6.5-beta.1
      entries:
      - name: kubeturbo-operator.v8.6.5-beta.1
        version: 8.6.5-beta.1
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: beta
    - currentCSV: kubeturbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/kubeturbo-operator:8.6
        version: 8.6.0
      entries:
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: stable
    - currentCSV: kubeturbo-operator.v8.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.4
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/kubeturbo-operator:8.4
        version: 8.4.0
      entries:
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.4.0
    - currentCSV: kubeturbo-operator.v8.5.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.5
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/kubeturbo-operator:8.5
        version: 8.5.0
      entries:
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.5.0
    - currentCSV: kubeturbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/kubeturbo-operator:8.6
        version: 8.6.0
      entries:
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.6.0
    defaultChannel: stable
    packageName: kubeturbo
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Silicom Ltd.
      provider-url: http://www.silicom-usa.com
    name: silicom-sts-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: silicom-sts-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 3,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-BC-8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/mode": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "hp-ocp",
                  "namespase": "openshift-operators"
                },
                "spec": {
                  "dummy": 0
                },
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicom-sts/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicom-sts/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/rmr/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - quay.io/silicom/ice-driver-src@sha256:569c8333e21c7c022ea71fa5b418404899040d5b3a3a4128eeac21f62aacaf3e
        - quay.io/silicom/sts-operator@sha256:7d474d0aa91ac57d8f5817b033b67b511c8601e0dbd83d808a69004bd24b4e60
        - quay.io/silicom/gpsd@sha256:e479c542e061b5422865f2617ed0563794f265c6e6660e5a9d204c9e1cb2f346
        - quay.io/silicom/phc2sys@sha256:b067167a1a5d93e9fe715c7774f6cdbbcd93cac27780f022e44dcb8a4d4d7bb9
        - quay.io/silicom/grpc-tsyncd@sha256:631bb754742e1a24376913baed60e480a6e6ae988ad9a3892254ff96996c15d1
        - quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - quay.io/silicom/sts-plugin@sha256:233bc95f8617a06bd452548ec116d91551dd4933970860d8d5ccb81f75ad2220
        - quay.io/silicom/tsyncd@sha256:3f1cbd4a8b5d62d7d3e525af5f26d9f27b8c08e886b0639a1e368163e166b537
        version: 0.1.3
      entries:
      - name: silicom-sts-operator.v0.1.3
        version: 0.1.3
      - name: silicom-sts-operator.v0.0.9
        version: 0.0.9
      - name: silicom-sts-operator.v0.0.8
        version: 0.0.8
      name: alpha
    - currentCSV: silicom-sts-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 3,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-BC-8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/mode": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "hp-ocp",
                  "namespase": "openshift-operators"
                },
                "spec": {
                  "dummy": 0
                },
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicom-sts/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicom-sts/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/rmr/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - quay.io/silicom/tsyncd@sha256:3f1cbd4a8b5d62d7d3e525af5f26d9f27b8c08e886b0639a1e368163e166b537
        - quay.io/silicom/gpsd@sha256:e479c542e061b5422865f2617ed0563794f265c6e6660e5a9d204c9e1cb2f346
        - quay.io/silicom/sts-plugin@sha256:233bc95f8617a06bd452548ec116d91551dd4933970860d8d5ccb81f75ad2220
        - quay.io/silicom/ice-driver-src@sha256:569c8333e21c7c022ea71fa5b418404899040d5b3a3a4128eeac21f62aacaf3e
        - quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
        - quay.io/silicom/sts-operator@sha256:7d474d0aa91ac57d8f5817b033b67b511c8601e0dbd83d808a69004bd24b4e60
        - quay.io/silicom/grpc-tsyncd@sha256:631bb754742e1a24376913baed60e480a6e6ae988ad9a3892254ff96996c15d1
        - quay.io/silicom/phc2sys@sha256:b067167a1a5d93e9fe715c7774f6cdbbcd93cac27780f022e44dcb8a4d4d7bb9
        version: 0.1.3
      entries:
      - name: silicom-sts-operator.v0.1.3
        version: 0.1.3
      - name: silicom-sts-operator.v0.0.9
        version: 0.0.9
      - name: silicom-sts-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: silicom-sts-operator
    provider:
      name: Silicom Ltd.
      url: http://www.silicom-usa.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-operator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-flashsystemcluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "flashsystemcluster.odf.ibm.com/v1alpha1",
                  "name": "my-flashsystemcluster",
                  "namespace": "openshift-storage"
                }
              },
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-storagecluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "storagecluster.ocs.openshift.io/v1",
                  "name": "my-storagecluster",
                  "namespace": "openshift-storage"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-console"]'
          containerImage: registry.redhat.io/odf4/odf-rhel9-operator@sha256:8d377d2d8a94e461a6eca8d1c3b4f5cf1f9ec1ba6a03044304403b58bbbabaa4
          createdAt: "2025-09-12T07:34:03Z"
          description: OpenShift Data Foundation provides a common control plane for
            storage solutions on OpenShift Container Platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.16"}]'
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operator.openshift.io/uninstall-message: 'Warning: deleting the operand
            (StorageSystem) will result in data loss'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "odf.openshift.io/v1alpha1",
              "kind": "StorageSystem",
              "metadata": {
                "name": "ocs-storagecluster-storagesystem",
                "namespace": "openshift-storage"
              },
              "spec": {
                "kind": "storagecluster.ocs.openshift.io/v1",
                "name": "ocs-storagecluster",
                "namespace": "openshift-storage"
              }
            }
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/odf-operator
          support: Red Hat
          vendors.odf.openshift.io/kind: '["storagecluster.ocs.openshift.io/v1", "flashsystemcluster.odf.ibm.com/v1alpha1"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageSystem is the Schema for the storagesystems API
            displayName: Storage System
            kind: StorageSystem
            name: storagesystems.odf.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Data Foundation** deploys three operators.

          ### OpenShift Data Foundation
          The OpenShift Data Foundation operator is the primary operator for
          OpenShift Data Foundation. It serves to facilitate the other operators in
          OpenShift Data Foundation by performing administrative tasks outside their
          scope as well as watching and configuring their CustomResources.

          ### Rook
          [Rook][1] deploys and manages Ceph on OpenShift, which provides block,
          file and object storage.

          ### NooBaa operator
          The NooBaa operator deploys and manages the [NooBaa][2] Multi-Cloud Gateway
          on OpenShift, which provides object storage.

          # Core Capabilities
          * **Self-managing service:** No matter which supported storage technologies
          you choose, OpenShift Data Foundation ensures that resources can be
          deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Data Foundation you
          can either build dedicated storage clusters or hyper-converged clusters
          where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Data Foundation:**
          OpenShift Data Foundation integrates Ceph with multiple storage
          presentations including object storage (compatible with S3), block storage,
          and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Data Foundation efficiently
          distributes and replicates your data across your cluster to minimize the
          risk of data loss. With snapshots, cloning, and versioning, no more losing
          sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your
          datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data
          storage for your workloads, using a single namespace.

          # Subscriptions
          OpenShift Data Foundation requires a matching 1:1 ratio with the OpenShift
          subscription used. The Essential subscription is included in the OpenShift
          Platform Plus subscription. Disaster Recovery features, external mode, and
          KMS integration for encryption at rest require a bundle of OpenShift
          Platform Plus with OpenShift Data Foundation Advanced subscription.

          [1]: https://rook.io
          [2]: https://www.noobaa.io
        displayName: OpenShift Data Foundation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/odf4/odf-rhel9-operator@sha256:8d377d2d8a94e461a6eca8d1c3b4f5cf1f9ec1ba6a03044304403b58bbbabaa4
        - registry.redhat.io/odf4/odf-console-rhel9@sha256:5d0d27a091a74e65985d2f047d3f6322043f2ae80fa5ff3192ec84cce92f143d
        version: 4.15.20-rhodf
      entries:
      - name: odf-operator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odf-operator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odf-operator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odf-operator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odf-operator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odf-operator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odf-operator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odf-operator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odf-operator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odf-operator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odf-operator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odf-operator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odf-operator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odf-operator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odf-operator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odf-operator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odf-operator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odf-operator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odf-operator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odf-operator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odf-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-flashsystemcluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "flashsystemcluster.odf.ibm.com/v1alpha1",
                  "name": "my-flashsystemcluster",
                  "namespace": "openshift-storage"
                }
              },
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-storagecluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "storagecluster.ocs.openshift.io/v1",
                  "name": "my-storagecluster",
                  "namespace": "openshift-storage"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-console"]'
          containerImage: registry.redhat.io/odf4/odf-rhel9-operator@sha256:20cb12259667539a0244cfde37297b975e2682833fbe537caecb5107ebecf15f
          createdAt: "2025-09-12T06:33:04Z"
          description: OpenShift Data Foundation provides a common control plane for
            storage solutions on OpenShift Container Platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.17"}]'
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operator.openshift.io/uninstall-message: 'Warning: deleting the operand
            (StorageSystem) will result in data loss'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "odf.openshift.io/v1alpha1",
              "kind": "StorageSystem",
              "metadata": {
                "name": "ocs-storagecluster-storagesystem",
                "namespace": "openshift-storage"
              },
              "spec": {
                "kind": "storagecluster.ocs.openshift.io/v1",
                "name": "ocs-storagecluster",
                "namespace": "openshift-storage"
              }
            }
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/odf-operator
          support: Red Hat
          vendors.odf.openshift.io/kind: '["storagecluster.ocs.openshift.io/v1", "flashsystemcluster.odf.ibm.com/v1alpha1"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageSystem is the Schema for the storagesystems API
            displayName: Storage System
            kind: StorageSystem
            name: storagesystems.odf.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Data Foundation** deploys three operators.

          ### OpenShift Data Foundation
          The OpenShift Data Foundation operator is the primary operator for
          OpenShift Data Foundation. It serves to facilitate the other operators in
          OpenShift Data Foundation by performing administrative tasks outside their
          scope as well as watching and configuring their CustomResources.

          ### Rook
          [Rook][1] deploys and manages Ceph on OpenShift, which provides block,
          file and object storage.

          ### NooBaa operator
          The NooBaa operator deploys and manages the [NooBaa][2] Multi-Cloud Gateway
          on OpenShift, which provides object storage.

          # Core Capabilities
          * **Self-managing service:** No matter which supported storage technologies
          you choose, OpenShift Data Foundation ensures that resources can be
          deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Data Foundation you
          can either build dedicated storage clusters or hyper-converged clusters
          where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Data Foundation:**
          OpenShift Data Foundation integrates Ceph with multiple storage
          presentations including object storage (compatible with S3), block storage,
          and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Data Foundation efficiently
          distributes and replicates your data across your cluster to minimize the
          risk of data loss. With snapshots, cloning, and versioning, no more losing
          sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your
          datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data
          storage for your workloads, using a single namespace.

          # Subscriptions
          OpenShift Data Foundation requires a matching 1:1 ratio with the OpenShift
          subscription used. The Essential subscription is included in the OpenShift
          Platform Plus subscription. Disaster Recovery features, external mode, and
          KMS integration for encryption at rest require a bundle of OpenShift
          Platform Plus with OpenShift Data Foundation Advanced subscription.

          [1]: https://rook.io
          [2]: https://www.noobaa.io
        displayName: OpenShift Data Foundation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/odf4/odf-rhel9-operator@sha256:20cb12259667539a0244cfde37297b975e2682833fbe537caecb5107ebecf15f
        - registry.redhat.io/odf4/odf-console-rhel9@sha256:8bd822a3410e520fc57ac69c7599831173003e2c34b980c517f79aca97210257
        version: 4.16.16-rhodf
      entries:
      - name: odf-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odf-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odf-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odf-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odf-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odf-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odf-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odf-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odf-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odf-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odf-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odf-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odf-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odf-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odf-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odf-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odf-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odf-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: jaeger-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jaeger-operator.v1.65.0-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "jaegertracing.io/v1",
                "kind": "Jaeger",
                "metadata": {
                  "name": "my-jaeger"
                },
                "spec": {
                  "strategy": "allInOne"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/rhosdt/jaeger-rhel8-operator@sha256:51a8685e4dc70888c991996f149cd42233049d18f702ec72b066d6dad77a07c9
          createdAt: 25 Aug 2025, 13:30
          description: Provides tracing, monitoring and troubleshooting for microservices-based
            distributed systems
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.65.0-4'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-distributed-tracing
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/os-observability/konflux-jaeger
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Jaeger is the Schema for the jaegers API
            displayName: Jaeger
            kind: Jaeger
            name: jaegers.jaegertracing.io
            version: v1
        description: |2-

          ## Operator deprecation warning

          This Operator is deprecated and not recommended for new deployments. This Operator will be removed from the `redhat-operators` catalog in a future release.

          If this is not the first time you are installing it, migrate to the Red Hat build of OpenTelemetry Operator and the Tempo Operator for distributed tracing collection and storage. For more information, see "Migrating" in the Red Hat build of OpenTelemetry documentation, "Installing" in the Red Hat build of OpenTelemetry documentation, and "Installing" in the distributed tracing platform (Tempo) documentation.

          ## About

          Red Hat OpenShift distributed tracing platform based on Jaeger. Jaeger is project inspired by [Dapper](https://research.google.com/pubs/pub36356.html) and [OpenZipkin](http://zipkin.io/). It is a distributed tracing system released as open source by Uber Technologies. It is used for monitoring and troubleshooting microservices-based distributed systems.

          ### Core capabilities
          Jaeger is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed context propagation
          * Distributed transaction monitoring
          * Root cause analysis
          * Service dependency analysis
          * Performance / latency optimization
          * OpenTracing compatible data model
          * Multiple storage backends: Elasticsearch, Memory.

          ### Operator features

          * **Multiple modes** - Provides `allInOne`, `production` and `streaming` [modes of deployment](https://www.jaegertracing.io/docs/latest/operator/#deployment-strategies).

          * **Configuration** - The Operator manages [configuration information](https://www.jaegertracing.io/docs/latest/operator/#configuring-the-custom-resource) when installing Jaeger instances.

          * **Storage** - [Configure storage](https://www.jaegertracing.io/docs/latest/operator/#storage-options) used by Jaeger. By default, `memory` is used. Other options include `elasticsearch`. The operator can delegate creation of an Elasticsearch cluster to the Elasticsearch Operator if deployed.

          * **Agent** - can be deployed as [sidecar](https://www.jaegertracing.io/docs/latest/operator/#auto-injecting-jaeger-agent-sidecars) (default) and/or [daemonset](https://www.jaegertracing.io/docs/latest/operator/#installing-the-agent-as-daemonset).

          * **UI** - Optionally setup secure route to provide [access to the Jaeger UI](https://www.jaegertracing.io/docs/latest/operator/#accessing-the-jaeger-console-ui).

          ### Before you start
          1. Ensure that the appropriate storage solution, that will be used by the Jaeger instance, is available and configured.
          2. If intending to deploy an Elasticsearch cluster via the Jaeger custom resource, then the Elasticsearch Operator must first be installed.

          ### Support & Troubleshooting

          Red Hat OpenShift distributed tracing Jaeger is available and supported as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/4.17/html/distributed_tracing/distributed-tracing-platform-jaeger) is recommended when installing, configuring, and managing the Operator and instances.
          Support is provided to Red Hat OpenShift entitled customers subject to the [Production Scope for Coverage](https://access.redhat.com/support/offerings/production/soc) and the [Red Hat OpenShift distributed tracing Life Cycle](https://access.redhat.com/support/policy/updates/openshift#jaeger).
        displayName: Red Hat OpenShift distributed tracing platform
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tracing
        - monitoring
        - troubleshooting
        links:
        - name: Jaeger Operator Source Code
          url: https://github.com/jaegertracing/jaeger-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:667d6545b92aa8381f65369de3b5a6781d1de5b4bcfa58ccde4721b6b0406f3d
        - registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:d92003d6cf0b4f3dae11f849355c9179e78e768f3b7dbeaf3224a4a019ec5d5d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e77efd450e757772d0753a73062753dabd08e53f21315e9b87f4cc1efa332661
        - registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:1d586f9aafd1cf847a14f509692ce54956b0e754a9a04beaa7a75c480f86717c
        - registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:7b021ed467d35b2bcae87765dfc64f639b03bf71f7e34bbea1dfad16626847b9
        - registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:1f4a9daf152420d9ca7cf89c38ae5550b12dd274bada389c7b85ea4cbcbc4fb7
        - registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:fd8110529daa8cf3e7f9fcf71346adc5f48cb05717ebbad1c25bc61c94f31656
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        - registry.redhat.io/rhosdt/jaeger-rhel8-operator@sha256:51a8685e4dc70888c991996f149cd42233049d18f702ec72b066d6dad77a07c9
        version: 1.65.0-4
      entries:
      - name: jaeger-operator.v1.65.0-4
        version: 1.65.0-4
      - name: jaeger-operator.v1.65.0-3
        version: 1.65.0-3
      - name: jaeger-operator.v1.65.0-2
        version: 1.65.0-2
      - name: jaeger-operator.v1.65.0-1
        version: 1.65.0-1
      - name: jaeger-operator.v1.62.0-2
        version: 1.62.0-2
      - name: jaeger-operator.v1.62.0-1
        version: 1.62.0-1
      - name: jaeger-operator.v1.57.0-10
        version: 1.57.0-10
      - name: jaeger-operator.v1.57.0-10-0.1730817420.p
        version: 1.57.0-10+0.1730817420.p
      - name: jaeger-operator.v1.57.0-7
        version: 1.57.0-7
      - name: jaeger-operator.v1.57.0-6
        version: 1.57.0-6
      - name: jaeger-operator.v1.57.0-5
        version: 1.57.0-5
      - name: jaeger-operator.v1.53.0-4
        version: 1.53.0-4
      - name: jaeger-operator.v1.53.0-3
        version: 1.53.0-3
      - name: jaeger-operator.v1.51.0-1
        version: 1.51.0-1
      - name: jaeger-operator.v1.47.1-5
        version: 1.47.1-5
      - name: jaeger-operator.v1.42.0-5-0.1687199951.p
        version: 1.42.0-5+0.1687199951.p
      - name: jaeger-operator.v1.42.0-5
        version: 1.42.0-5
      - name: jaeger-operator.v1.34.1-5
        version: 1.34.1-5
      - name: jaeger-operator.v1.30.2
        version: 1.30.2
      name: stable
    defaultChannel: stable
    packageName: jaeger-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeRocketCI
      provider-url: https://docs.kuberocketci.io/
    name: sonar-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sonar-operator.v3.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Sonar",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonar-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonar",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonar-sample"
                },
                "spec": {
                  "defaultPermissionTemplate": "edp-default",
                  "secret": "sonar-secret",
                  "settings": [
                    {
                      "key": "sonar.global.exclusions",
                      "values": [
                        "**/*.jsx",
                        "**/*.tsx"
                      ]
                    },
                    {
                      "fieldValues": {
                        "beginBlockRegexp": ".*",
                        "endBlockRegexp": ".*"
                      },
                      "key": "sonar.issue.ignore.block"
                    },
                    {
                      "key": "sonar.dbcleaner.hoursBeforeKeepingOnlyOneSnapshotByDay",
                      "value": "20"
                    }
                  ],
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonargroup-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonargroup",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonargroup-sample"
                },
                "spec": {
                  "description": "Developers",
                  "name": "sonar-developers",
                  "permissions": [
                    "admin"
                  ],
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarPermissionTemplate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarpermissiontemplate-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarpermissiontemplate",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarpermissiontemplate-sample"
                },
                "spec": {
                  "default": true,
                  "description": "EDP permission templates",
                  "groupsPermissions": {
                    "sonar-developers": [
                      "scan"
                    ],
                    "sonar-users": [
                      "issueadmin"
                    ]
                  },
                  "name": "edp-default",
                  "projectKeyPattern": ".+",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityGate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarqualitygate-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualitygate",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarqualitygate-sample"
                },
                "spec": {
                  "conditions": {
                    "new_duplicated_lines_density": {
                      "error": "1",
                      "op": "GT"
                    }
                  },
                  "default": true,
                  "name": "sonarqualitygate-sample",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityProfile",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarqualityprofile-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualityprofile",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarqualityprofile-sample"
                },
                "spec": {
                  "default": true,
                  "language": "go",
                  "name": "sonarqualityprofile-sample",
                  "rules": {
                    "go:S1151": {
                      "params": "max=\"6\"",
                      "severity": "MAJOR"
                    }
                  },
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonaruser-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonaruser",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonaruser-sample"
                },
                "spec": {
                  "email": "sonaruser@mail.com",
                  "groups": [
                    "sonar-administrators"
                  ],
                  "login": "sonaruser",
                  "name": "sonar user",
                  "permissions": [
                    "scan"
                  ],
                  "secret": "sonaruser-secret",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/sonar-operator:3.1.0
          createdAt: "2023-09-08T12:53:57Z"
          description: An Operator for managing Sonar entities.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-sonar-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonarGroup is the Schema for the sonar group API.
            displayName: Sonar Group
            kind: SonarGroup
            name: sonargroups.edp.epam.com
            version: v1alpha1
          - description: SonarPermissionTemplate is the Schema for the sonar permission
              template API.
            displayName: Sonar Permission Template
            kind: SonarPermissionTemplate
            name: sonarpermissiontemplates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityGate is the Schema for the sonarqualitygates
              API
            displayName: Sonar Quality Gate
            kind: SonarQualityGate
            name: sonarqualitygates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityProfile is the Schema for the sonarqualityprofiles
              API
            displayName: Sonar Quality Profile
            kind: SonarQualityProfile
            name: sonarqualityprofiles.edp.epam.com
            version: v1alpha1
          - description: Sonar is the Schema for the sonars API.
            displayName: Sonar
            kind: Sonar
            name: sonars.edp.epam.com
            version: v1alpha1
          - description: SonarUser is the Schema for the sonarusers API
            displayName: Sonar User
            kind: SonarUser
            name: sonarusers.edp.epam.com
            version: v1alpha1
        description: |
          Sonar Operator is an operator that is responsible for establishing
          a connection to provided SonarQube server, reconciling Sonar entities (users,
          groups, permission templates, etc) according to the created CRs.
        displayName: EDP Sonar Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - analysis
        - automation
        - code
        - continuous
        - devops
        - edp
        - inspection
        - metrics
        - metrics
        - monitoring
        - operator
        - quality
        - security
        - software
        - static
        - technical
        - vulnerabilities
        links:
        - name: Sonar Operator
          url: https://github.com/epam/edp-sonar-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/sonar-operator:3.1.0
        version: 3.1.0
      entries:
      - name: sonar-operator.v3.1.0
        version: 3.1.0
      name: alpha
    - currentCSV: sonar-operator.v3.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Sonar",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonar"
                  },
                  "name": "sonar-sample"
                },
                "spec": {
                  "defaultPermissionTemplate": "edp-default",
                  "secret": "sonar-secret",
                  "settings": [
                    {
                      "key": "sonar.global.exclusions",
                      "values": [
                        "**/*.jsx",
                        "**/*.tsx"
                      ]
                    },
                    {
                      "fieldValues": {
                        "beginBlockRegexp": ".*",
                        "endBlockRegexp": ".*"
                      },
                      "key": "sonar.issue.ignore.block"
                    },
                    {
                      "key": "sonar.dbcleaner.hoursBeforeKeepingOnlyOneSnapshotByDay",
                      "value": "20"
                    }
                  ],
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonargroup"
                  },
                  "name": "sonargroup-sample"
                },
                "spec": {
                  "description": "Developers",
                  "name": "sonar-developers",
                  "permissions": [
                    "admin"
                  ],
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarPermissionTemplate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarpermissiontemplate"
                  },
                  "name": "sonarpermissiontemplate-sample"
                },
                "spec": {
                  "default": true,
                  "description": "EDP permission templates",
                  "groupsPermissions": {
                    "sonar-developers": [
                      "scan"
                    ],
                    "sonar-users": [
                      "issueadmin"
                    ]
                  },
                  "name": "edp-default",
                  "projectKeyPattern": ".+",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityGate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualitygate"
                  },
                  "name": "sonarqualitygate-sample"
                },
                "spec": {
                  "conditions": {
                    "new_duplicated_lines_density": {
                      "error": "1",
                      "op": "GT"
                    }
                  },
                  "default": true,
                  "name": "sonarqualitygate-sample",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityProfile",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualityprofile"
                  },
                  "name": "sonarqualityprofile-sample"
                },
                "spec": {
                  "default": true,
                  "language": "go",
                  "name": "sonarqualityprofile-sample",
                  "rules": {
                    "go:S1151": {
                      "params": "max=\"6\"",
                      "severity": "MAJOR"
                    }
                  },
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonaruser"
                  },
                  "name": "sonaruser-sample"
                },
                "spec": {
                  "email": "sonaruser@mail.com",
                  "groups": [
                    "sonar-administrators"
                  ],
                  "login": "sonaruser",
                  "name": "sonar user",
                  "permissions": [
                    "scan"
                  ],
                  "secret": "sonaruser-secret",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/sonar-operator:3.3.0
          createdAt: "2025-04-22T11:39:32Z"
          description: An Operator for managing Sonar entities.
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/epam/edp-sonar-operator
          support: KubeRocketCI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonarGroup is the Schema for the sonar group API.
            displayName: Sonar Group
            kind: SonarGroup
            name: sonargroups.edp.epam.com
            version: v1alpha1
          - description: SonarPermissionTemplate is the Schema for the sonar permission
              template API.
            displayName: Sonar Permission Template
            kind: SonarPermissionTemplate
            name: sonarpermissiontemplates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityGate is the Schema for the sonarqualitygates
              API
            displayName: Sonar Quality Gate
            kind: SonarQualityGate
            name: sonarqualitygates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityProfile is the Schema for the sonarqualityprofiles
              API
            displayName: Sonar Quality Profile
            kind: SonarQualityProfile
            name: sonarqualityprofiles.edp.epam.com
            version: v1alpha1
          - description: Sonar is the Schema for the sonars API.
            displayName: Sonar
            kind: Sonar
            name: sonars.edp.epam.com
            version: v1alpha1
          - description: SonarUser is the Schema for the sonarusers API.
            displayName: Sonar User
            kind: SonarUser
            name: sonarusers.edp.epam.com
            version: v1alpha1
        description: |
          Sonar Operator is an operator that is responsible for establishing
          a connection to provided SonarQube server, reconciling Sonar entities (users,
          groups, permission templates, etc) according to the created CRs.
        displayName: EDP Sonar Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - analysis
        - automation
        - code
        - continuous
        - devops
        - edp
        - inspection
        - metrics
        - metrics
        - monitoring
        - operator
        - quality
        - security
        - software
        - static
        - technical
        - vulnerabilities
        links:
        - name: Sonar Operator
          url: https://github.com/epam/edp-sonar-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: KubeRocketCI
          url: https://docs.kuberocketci.io/
        relatedImages:
        - docker.io/epamedp/sonar-operator:3.3.0
        version: 3.3.0
      entries:
      - name: sonar-operator.v3.3.0
        version: 3.3.0
      - name: sonar-operator.v3.2.0
        version: 3.2.0
      name: stable
    defaultChannel: stable
    packageName: sonar-operator
    provider:
      name: KubeRocketCI
      url: https://docs.kuberocketci.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco
      provider-url: ""
    name: aci-containers-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aci-containers-operator.v1.0.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"aci.ctrl/v1alpha1","kind":"AciContainersOperator","metadata":{"name":"acicnioperator","namespace":"aci-containers-system"},"spec":{}}]'
          capabilities: Seamless Upgrades
          categories: Networking
          certified: "false"
          containerImage: quay.io/noiro/aci-containers-operator@sha256:2b2d685e254c206be5371285852a8f95d0318af69cb85b786427969e035c125c
          createdAt: "2025-04-10T12:00:00Z"
          description: The Cisco Application Centric Infrastructure (ACI) Containers
            Operator brings the ACI CNI Plugin and Policy model to Kubernetes clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://github.com/noironetworks/aci-containers/blob/master/README.md
          repository: https://github.com/noironetworks/aci-containers
          support: acisupport@cisco.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aci Containers Operator
            displayName: AciContainersOperator
            kind: AciContainersOperator
            name: acicontainersoperators.aci.ctrl
            version: v1alpha1
        description: |
          ## ACI CNI Plugin

          The Cisco Application Centric Infrastructure (ACI) CNI plugin brings the ACI Networking and Policy model to Kubernetes clusters that reside on-prem or in the cloud. It is fully open source and relies on the Opflex Protocol to program Open vSwitch instances running on the Kubernetes nodes. It provides IP address management, L2/L3 networking, load balancing, and security functions for container workloads.

          The Cisco ACI CNI plugin for the Red Hat OpenShift Container Platform provides a single, programmable network infrastructure that eliminates bottlenecks between OpenShift workloads and the rest of the infrastructure while providing enterprise-grade security and offers flexible micro segmentation possibilities.

          Unified Networking - By extending the integrated overlay of the ACI fabric to OpenShift workloads, Pod and Service endpoints are managed as native ACI network resources at the same level as Bare Metal or Virtual Machines.

          No Egress Router Required - Applications running in OpenShift Pods that need high-bandwidth, low-latency traffic for data transfers external to the cluster suffer the bottleneck imposed by the egress router implementation. Pods can easily connect with other fabric or non-fabric endpoints leveraging the combined hardware and software distributed routing and switching provided by ACI and the ACI CNI plugin.

          Simplified Operations and Enhanced Visibility - Whether policies and networking are defined by the devops or fabric or security admins, the APIC becomes the single source of truth for operations and security governance. Kubernetes workload visibility is provided to Network Operators within their familiar Cisco tools.

          Hardware-assisted Load Balancing - Ingress connections to LoadBalancer-type services using ACI’s Policy Based Redirect technology moves load balancing into the fabric's hardware.

          Granular Security - Security can be implemented by using native NetworkPolicy or by using ACI EPGs and contracts, or both models complementing each other. Network policy of container workloads becomes a shared responsibility while ensuring global compliance and tenant constraints are enforced.
        displayName: Cisco ACI Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cisco
        - aci
        links:
        - name: Cisco ACI Homepage
          url: https://www.cisco.com/c/en/us/td/docs/switches/datacenter/aci/apic/sw/kb/b_Kubernetes_Integration_with_ACI.html
        maintainers:
        - email: abhishs9@cisco.com
          name: 'Abhishek Sharma '
        - email: sumitnaiksatam@gmail.com
          name: Sumit Naiksatam
        - email: vlella@cisco.com
          name: Vishal Lella
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Cisco
        relatedImages:
        - quay.io/noiro/aci-containers-operator@sha256:2b2d685e254c206be5371285852a8f95d0318af69cb85b786427969e035c125c
        - quay.io/noiro/acc-provision-operator@sha256:9c8fa6d705602af13d539fb919e72e7ee5064c6998fc9ca95d10108e986c68fe
        version: 1.0.10
      entries:
      - name: aci-containers-operator.v1.0.10
        version: 1.0.10
      name: stable
    defaultChannel: stable
    packageName: aci-containers-operator
    provider:
      name: Cisco
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Accuknox
      provider-url: https://accuknox.com
    name: kubearmor-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubearmor-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          com.redhat.openshift.versions: v4.12
          createdAt: "2023-05-01T05:17:07Z"
          description: The KubeArmor Operator deloys and manages KubeArmor daemonset
            and its componenets within a cluster.
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.28.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/kubearmor/KubeArmor
          support: Accuknox
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          The KubeArmor Operator deploys and manages KubeArmor and its components:
          - KubeArmor Daemonset
          - KubeArmor Relay Server
          - KubeArmor Controller to manage CRDs (KubeArmorPolicy, KubeArmorHostPolicy)

          KubeArmor is a cloud-native runtime security enforcement system that restricts the behavior (such as process
          execution, file access, and networking operations) of pods, containers, and nodes (VMs) at the system level.

          KubeArmor leverages Linux security modules (LSMs) such as AppArmor, SELinux, or BPF-LSM to enforce the user-specified policies. KubeArmor generates rich alerts/telemetry events with container/pod/namespace identities by leveraging eBPF.

          ## Documentation

          github: [https://github.com/kubearmor/KubeArmor](https://github.com/kubearmor/KubeArmor)
          website: [https://kubearmor.io/](https://kubearmor.io/)
        displayName: Kubearmor Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubearmor-operator
        - runtime-security
        - security
        - cloud-native
        links:
        - name: Kubearmor
          url: https://kubearmor.io
        maintainers:
        - email: ramakant@accuknox.com
          name: Ramakant Sharma
        maturity: alpha
        provider:
          name: Accuknox
          url: https://www.accuknox.com/
        relatedImages:
        - docker.io/kubearmor/kubearmor-operator@sha256:8c2d56781a2adf78ee58f956e01417f38b6ad2f319feada6470e93f4f9bf0808
        - docker.io/kubearmor/kubearmor-controller@sha256:ccec0fb8bb1e91bd48631cc89baf752b842f19c33b50edab08f8b33e57af64d4
        - docker.io/kubearmor/kubearmor@sha256:52f7049ecdbfa779663d2cd44ddd7f36aecd8ccd2ca91e69709b628600029f13
        - docker.io/kubearmor/kubearmor-init@sha256:44fa534587972c5f85e3e6f362dd8268e9088a33c5657a21b72a7e8563164fb5
        - docker.io/kubearmor/kubearmor-relay-server@sha256:695f9ade3382bcf858077e2a7b9b01e0713071517603778d72f7aca5a6d71323
        version: 1.0.0
      entries:
      - name: kubearmor-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: kubearmor-operator.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.kubearmor.com/v1",
                "kind": "KubeArmorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kubearmorconfig",
                    "app.kubernetes.io/part-of": "kubearmoroperator"
                  },
                    "name": "kubearmorconfig-default",
                    "namespace": "placeholder"
                },
                "spec": {
                  "adapters": {
                    "elasticsearch": {
                      "url": "",
                      "enabled": false,
                      "auth": {
                        "secretName": "elastic-secret",
                        "usernameKey": "username",
                        "passwordKey": "password",
                        "allowInsecureTLS": false,
                        "caCertSecretName": "",
                        "caCertKey": "ca.crt"
                      },
                      "alertsIndex": "kubearmor-alerts"
                    }
                  },
                  "recommendedPolicies": {
                      "enable": true,
                      "matchExpressions": [
                          {
                              "key": "namespace",
                              "operator": "NotIn",
                              "values": [
                                  "kube-system",
                                  "kubearmor",
                                  "openshift-infra",
                                  "openshift-ingress",
                                  "openshift-ingress-canary",
                                  "openshift-ingress-operator",
                                  "openshift-insights",
                                  "openshift-kni-infra",
                                  "openshift-kube-apiserver",
                                  "openshift-kube-apiserver-operator",
                                  "openshift-kube-controller-manager",
                                  "openshift-kube-controller-manager-operator",
                                  "openshift-kube-scheduler",
                                  "openshift-kube-scheduler-operator",
                                  "openshift-kube-storage-version-migrator",
                                  "openshift-kube-storage-version-migrator-operator",
                                  "openshift-logging",
                                  "openshift-machine-api",
                                  "openshift-machine-config-operator",
                                  "openshift-managed-node-metadata-operator",
                                  "openshift-managed-upgrade-operator",
                                  "openshift-marketplace",
                                  "openshift-monitoring",
                                  "openshift-multus",
                                  "openshift-must-gather-operator",
                                  "openshift-network-diagnostics",
                                  "openshift-network-node-identity",
                                  "openshift-network-operator",
                                  "openshift-node",
                                  "openshift-nutanix-infra",
                                  "openshift-oauth-apiserver",
                                  "openshift-observability-operator",
                                  "openshift-ocm-agent-operator",
                                  "openshift-openstack-infra",
                                  "openshift-operator-lifecycle-manager",
                                  "openshift-operators",
                                  "openshift-operators-redhat",
                                  "openshift-osd-metrics",
                                  "openshift-ovirt-infra",
                                  "openshift-ovn-kubernetes",
                                  "openshift-package-operator",
                                  "openshift-rbac-permissions",
                                  "openshift-route-controller-manager",
                                  "openshift-route-monitor-operator",
                                  "openshift-security",
                                  "openshift-service-ca",
                                  "openshift-service-ca-operator",
                                  "openshift-splunk-forwarder-operator",
                                  "openshift-sre-pruning",
                                  "openshift-user-workload-monitoring",
                                  "openshift-validation-webhook",
                                  "openshift-velero",
                                  "openshift-vsphere-infra",
                                  "openshift",
                                  "openshift-addon-operator",
                                  "openshift-apiserver",
                                  "openshift-apiserver-operator",
                                  "openshift-aqua",
                                  "openshift-authentication",
                                  "openshift-authentication-operator",
                                  "openshift-backplane",
                                  "openshift-backplane-cee",
                                  "openshift-backplane-csa",
                                  "openshift-backplane-cse",
                                  "openshift-backplane-csm",
                                  "openshift-backplane-managed-scripts",
                                  "openshift-backplane-mcs-tier-two",
                                  "openshift-backplane-mobb",
                                  "openshift-backplane-srep",
                                  "openshift-backplane-tam",
                                  "openshift-cloud-controller-manager",
                                  "openshift-cloud-controller-manager-operator",
                                  "openshift-cloud-credential-operator",
                                  "openshift-cloud-ingress-operator",
                                  "openshift-cloud-network-config-controller",
                                  "openshift-cloud-platform-infra",
                                  "openshift-cluster-csi-drivers",
                                  "openshift-cluster-machine-approver",
                                  "openshift-cluster-node-tuning-operator",
                                  "openshift-cluster-samples-operator",
                                  "openshift-cluster-storage-operator",
                                  "openshift-cluster-version",
                                  "openshift-codeready-workspaces",
                                  "openshift-config",
                                  "openshift-config-managed",
                                  "openshift-config-operator",
                                  "openshift-console",
                                  "openshift-console-operator",
                                  "openshift-console-user-settings",
                                  "openshift-controller-manager",
                                  "openshift-controller-manager-operator",
                                  "openshift-custom-domains-operator",
                                  "openshift-customer-monitoring",
                                  "openshift-deployment-validation-operator",
                                  "openshift-dns",
                                  "openshift-dns-operator",
                                  "openshift-etcd",
                                  "openshift-etcd-operator",
                                  "openshift-host-network",
                                  "openshift-image-registry"
                              ]
                          }
                      ]
                  },
                  "defaultCapabilitiesPosture": "audit",
                  "defaultFilePosture": "audit",
                  "defaultNetworkPosture": "audit",
                  "defaultVisibility": "process,network",
                  "enableStdOutLogs": false,
                  "enableStdOutAlerts": false,
                  "enableStdOutMsgs": false,
                  "seccompEnabled": false
                }
              }
            ]
          alm-examples-metadata: '{"kubearmorconfig-default":{"description":"Default
            KubeArmorConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Security, Monitoring
          certified: "true"
          com.redhat.openshift.versions: v4.9
          containerImage: docker.io/kubearmor/kubearmor-operator@sha256:d6c6b16e636474568835ded97cecb5ed17bc6004501227f6b6d63522ed57b01c
          createdAt: "2024-03-20T17:43:15Z"
          description: KubeArmor Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: kubearmor
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kubearmor/KubeArmor/blob/main/pkg/KubeArmorOperator/
          support: Accuknox
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeArmorConfig is the Schema for the kubearmorconfigs API
            kind: KubeArmorConfig
            name: kubearmorconfigs.operator.kubearmor.com
            version: v1
        description: |2-

          ## Prerequisites

          1. Create a new project for KubeArmor (Recommended)

             ```
             oc new-project kubearmor
             ```

           2. Create [KubeArmor SCC](https://github.com/kubearmor/KubeArmor/blob/main/pkg/KubeArmorOperator/config/rbac/kubearmor-scc.yaml)

             ```
             oc create -f https://raw.githubusercontent.com/kubearmor/KubeArmor/main/pkg/KubeArmorOperator/config/rbac/kubearmor-scc.yaml
             ```

             Note: please change the serviceaccount namespace if you are using some other namespace than the kubearmor.

          KubeArmor is a cloud-native runtime security enforcement system that restricts the behavior
          (such as process execution, file access, and networking operations) of pods, containers,
          and nodes (VMs) at the system level.
          KubeArmor leverages Linux security modules (LSMs) such as AppArmor, SELinux, or BPF-LSM to enforce the
          user-specified policies. KubeArmor generates rich alerts/telemetry events with container/pod/namespace
          identities by leveraging eBPF.

          ## Features and Benefits

          ### Harden Infrastructure

          safeguards vital paths like certificate bundles, enforces security rules based on MITRE, STIGs, and CIS guidelines,
          and restricts access to raw database tables, enhancing overall cluster security.

          ### Least Permissive Access

          controls sensitive asset access through process and network whitelisting,
          reducing security risks and maintaining system integrity.

          ### Application Behavior

          in-depth monitoring of process executions, file system access, service bindings,
          ingress and egress connections, and sensitive system call profiling

          ## Documentation

          github: [https://github.com/kubearmor/KubeArmor](https://github.com/kubearmor/KubeArmor)

          website: [https://kubearmor.io/](https://kubearmor.io/)

          ## Installing the KubeArmor Operator

          The KubeArmor Operator, available through Red Hat's Operator catalog, offers stable update channel.
          **Stable Channel**: This channel provides the latest stable version of KubeArmor along with its patches.

          The KubeArmor operator is responsible for deploying various components of the KubeArmor application, including the KubeArmor
          Daemonset, KubeArmor Controller, KubeArmor Relay Server, and KubeArmor Snitch. The role of KubeArmor Snitch is to gather information
          about the enabled Linux security modules on the host, which is utilized by KubeArmor to enforce runtime security policies.
          Additionally, KubeArmor Snitch collects details about the container runtime socket and storage volume cluster configurations.
          This deployment is executed as a short-term job and necessitates root user privileges. KubeArmor Snitch is certified by RedHat
          as a container with root privileges.

          KubeArmor Operator comes with a custom resource:

          - **KubeArmorConfig** - User will provide the initial configurations of the KubeArmor using KubeArmorConfig CR.

          ### KubeArmorConfig Customization Options:

          | Parameter                                 | Description                                          | Default                                 |
          | ---                                       | ---                                                  | ---                                     |
          | `alertThrottling`                         | Enable throttling for the KubeArmor generated alerts | False                                   |
          | `defaultFilePosture`                      | Default file posture for KubeArmor                   | audit                                   |
          | `defaultNetworkPosture`                   | Default network posture for KubeArmor                | audit                                   |
          | `defaultCapabilitiesPosture`              | Default capabilities posture for KubeArmor           | audit                                   |
          | `defaultVisibility`                       | Default visibility for KubeArmor                     | process,file,network                    |
          | `enableStdOutLogs` |  show stdout logs for relay server | false |
          | `enableStdOutAlerts` |  show stdout alerts for relay server | false |
          | `enableStdOutMsgs` |  show stdout messages for relay server | false |
          | `kubearmorImage.image`                    | Default image for KubeArmor Daemonset                | kubearmor/kubearmor-ubi:stable          |
          | `kubearmorImage.imagePullPolicy`          | Default imagePullPolicy for KubeArmor Daemonset      | Always                                  |
          | `kubearmorInitImage.image`                | Default image for KubeArmor Init Container           | kubearmor/kubearmor-init:stable         |
          | `kubearmorInitImage.imagePullPolicy`      | Default imagePullPolicy for KubeArmor Init Container | Always                                  |
          | `kubearmorRelayImage.image`               | Default image for KubeArmor Relay Server             | kubearmor/kubearmor-relay-server:latest |
          | `kubearmorRelayImage.imagePullPolicy`     | Default imagePullPolicy for KubeArmor Relay Server   | Always                                  |
          | `kubearmorControllerImage.image`          | Default image for KubeArmor Controller               | kubearmor/kubearmor-controller:latest   |
          | `kubearmorControllerImage.imagePullPolicy`| Default imagePullPolicy for KubeArmor Controller     | Always                                  |
          | `maxAlertPerSec`                          | Maximum no of alerts per second to trigger throttling| 10                                      |
          | `recommendedPolicies`                     | Recommended Policies Spec                            | -                                       |
          | `recommendedPolicies.matchExpressions`    | Target Namespaces for recommended policies           | -                                       |
          | `recommendedPolicies.excludePolicy`       | Exclude policies from recommended policies           | -                                       |
          | `seccompEnabled`                          | Deploy KubeArmor Daemon with default seccomp profile | True                                    |
          | `throttleSec`                             | Thottling duration in seconds                        | 30                                      |

          Users are not required to specify the images in the KubearmorConfig custom resource (CR). By default, the operator will utilize the images published with the operator release.

          For detailed information on Default Posture refer [this guide](https://github.com/kubearmor/KubeArmor/blob/main/getting-started/default_posture.md)

          similarly for default visibility refer [this guide](https://github.com/kubearmor/KubeArmor/blob/main/getting-started/kubearmor_visibility.md)

          ## Support

          For Support please join KubeArmor Slack [Here](https://join.slack.com/t/kubearmor/shared_invite/zt-1ltmqdbc6-rSHw~LM6MesZZasmP2hAcA)
        displayName: KubeArmor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - security
        - alerting
        - runtime-security
        - cloud native
        - container security
        links:
        - name: Website
          url: https://kubearmor.io
        - name: Github
          url: https://github.com/kubearmor/KubeArmor.git
        - name: Slack
          url: https://join.slack.com/t/kubearmor/shared_invite/zt-1ltmqdbc6-rSHw~LM6MesZZasmP2hAcA
        maintainers:
        - email: ramakant@accuknox.com
          name: Ramakant Sharma
        maturity: stable
        provider:
          name: Accuknox
          url: https://accuknox.com
        relatedImages:
        - docker.io/kubearmor/kubearmor-snitch@sha256:c3d0f45590f9a0ebb50254047f12a7468844bcae7961eccb11f5ef578832c0eb
        - docker.io/kubearmor/kubearmor-ubi@sha256:629e295a2e64953bf95f3f3e842c42ff665ed6e4db32db8e43dcde4befdd3fc2
        - docker.io/kubearmor/kubearmor-operator@sha256:d6c6b16e636474568835ded97cecb5ed17bc6004501227f6b6d63522ed57b01c
        - docker.io/kubearmor/kubearmor-init@sha256:e30ee508aa95d273f74a543e1b412b98c1714d04f093b59991c0ec2abaa450be
        - docker.io/kubearmor/kubearmor-relay-server@sha256:0a1a20ef02bd05eaa4a900948cadee6ae7650b0dde70abe1935608eaa5f931b7
        - docker.io/kubearmor/kubearmor-controller@sha256:2d369138cdbb31468c41e9ce70eda986a16bbeb9dd1192cd8a011c9df8a83857
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:1153a4592746b05e019bde4d818d176ff9350c013f84d49829032540de882841
        version: 1.5.2
      entries:
      - name: kubearmor-operator.v1.5.2
        version: 1.5.2
      name: stable
    defaultChannel: stable
    packageName: kubearmor-operator-certified
    provider:
      name: Accuknox
      url: https://accuknox.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Tigera
      provider-url: https://tigera.io/
    name: tigera-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tigera-operator.v1.20.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:bd93e66be5cfbc123fe4e8c3f6e8ddf7654217bf26136f06dcf23f28b53063d2
          createdAt: "2021-10-05T20:02:14.564507Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.20.4
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:bd93e66be5cfbc123fe4e8c3f6e8ddf7654217bf26136f06dcf23f28b53063d2
        version: 1.20.4
      entries:
      - name: tigera-operator.v1.20.4
        version: 1.20.4
      - name: tigera-operator.v1.20.3
        version: 1.20.3
      - name: tigera-operator.v1.20.2
        version: 1.20.2
      - name: tigera-operator.v1.20.1
        version: 1.20.1
      name: release-v1.20
    - currentCSV: tigera-operator.v1.23.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:f258275fe3d48c2ae16346a57f628523a6f2fb7b854b6d43b169056e158537f1
          createdAt: "2022-01-12T22:55:57.409244Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.23.4
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:f258275fe3d48c2ae16346a57f628523a6f2fb7b854b6d43b169056e158537f1
        version: 1.23.4
      entries:
      - name: tigera-operator.v1.23.4
        version: 1.23.4
      - name: tigera-operator.v1.23.3
        version: 1.23.3
      - name: tigera-operator.v1.23.2
        version: 1.23.2
      - name: tigera-operator.v1.23.1
        version: 1.23.1
      - name: tigera-operator.v1.23.0
        version: 1.23.0
      name: release-v1.23
    - currentCSV: tigera-operator.v1.25.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:e862708c700df55f99a9c9bb3c453e31f22aef1993d6079d80433eb8da831c94
          createdAt: "2022-02-17T05:39:13.720885Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.25.1
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:e862708c700df55f99a9c9bb3c453e31f22aef1993d6079d80433eb8da831c94
        version: 1.25.1
      entries:
      - name: tigera-operator.v1.25.1
        version: 1.25.1
      name: release-v1.25
    - currentCSV: tigera-operator.v1.28.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:9e8e47045fc67a2b00f9f2fd575fa526d7ddcac50af0b8b6d583b7518ccce120
          createdAt: "2023-05-17T16:13:52.148441156Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.28.13
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:9e8e47045fc67a2b00f9f2fd575fa526d7ddcac50af0b8b6d583b7518ccce120
        version: 1.28.13
      entries:
      - name: tigera-operator.v1.28.13
        version: 1.28.13
      name: release-v1.28
    - currentCSV: tigera-operator.v1.29.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:2a5acd29d8dca6adb318af04dbb15d93d899ed2b28b2ee9aaac09147395efb58
          createdAt: "2023-03-31T20:49:52.518230079Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.29.3
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:2a5acd29d8dca6adb318af04dbb15d93d899ed2b28b2ee9aaac09147395efb58
        version: 1.29.3
      entries:
      - name: tigera-operator.v1.29.3
        version: 1.29.3
      name: release-v1.29
    - currentCSV: tigera-operator.v1.30.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:431f037ff18b5c867d01312e42671effc55602421aeed25dd3f6109f70596b4a
          createdAt: "2023-11-17T00:52:31.263564294Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.30.9
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: APIServer
            name: apiservers.operator.tigera.io
            version: v1
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:431f037ff18b5c867d01312e42671effc55602421aeed25dd3f6109f70596b4a
        version: 1.30.9
      entries:
      - name: tigera-operator.v1.30.9
        version: 1.30.9
      name: release-v1.30
    - currentCSV: tigera-operator.v1.32.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:715ac9a30f8a9579e44258af20de354715429e11836b493918e9e1a696e9b028
          createdAt: "2023-12-15T23:45:03.964812504Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.32.3
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:715ac9a30f8a9579e44258af20de354715429e11836b493918e9e1a696e9b028
        version: 1.32.3
      entries:
      - name: tigera-operator.v1.32.3
        version: 1.32.3
      name: release-v1.32
    - currentCSV: tigera-operator.v1.36.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:3341fa9475c0325b86228c8726389f9bae9fd6c430c66fe5cd5dc39d7bb6ad4b
          createdAt: "2025-02-06T02:39:01.007218336Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.36.5
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:3341fa9475c0325b86228c8726389f9bae9fd6c430c66fe5cd5dc39d7bb6ad4b
        version: 1.36.5
      entries:
      - name: tigera-operator.v1.36.5
        version: 1.36.5
      name: release-v1.36
    - currentCSV: tigera-operator.v1.38.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "calicoNetwork": {
                    "linuxDataplane": "BPF"
                  },
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing, Monitoring, Networking, Security
          containerImage: quay.io/tigera/operator@sha256:e0a34b265aebce1a2db906d8dad99190706e8bf3910cae626b9c2eb6bbb21775
          createdAt: "2025-05-06T13:27:07.802897329Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          features.operators.openshift.io/valid-subscription: No subscription required
          olm.skipRange: <1.38.0
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift.
        displayName: Tigera Operator v1.38
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tigera-operator
        - networking
        - security
        - monitoring
        links:
        - name: Tigera
          url: https://tigera.io/
        - name: Calico Introduction
          url: https://docs.tigera.io/calico/latest/about/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.tigera.io/calico/latest/getting-started/kubernetes/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
          url: https://tigera.io/
        relatedImages:
        - quay.io/tigera/operator@sha256:e0a34b265aebce1a2db906d8dad99190706e8bf3910cae626b9c2eb6bbb21775
        version: 1.38.0
      entries:
      - name: tigera-operator.v1.38.0
        version: 1.38.0
      name: release-v1.38
    defaultChannel: release-v1.38
    packageName: tigera-operator
    provider:
      name: Tigera
      url: https://tigera.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solution-Soft
      provider-url: www.solution-soft.com
    name: timemachine-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: timemachine-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tm.solution-soft.com/v1alpha1",
                "kind": "TimeMachine",
                "metadata": {
                  "name": "timemachine-sample"
                },
                "spec": {
                  "cluster_provider": "ibm",
                  "aws": {
                    "shared_volume_name": ""
                  },
                  "ibm": {
                    "persistentvolume": {
                      "storageclass": "ibmc-file-silver-gid"
                    }
                  },
                  "tm_lichost": "52.152.148.40",
                  "tm_licpass": "qwaszx",
                  "tm_licport": "57777",
                  "imagePullSecrets": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
          createdAt: "2024-02-05T09:46:17Z"
          description: Time Machine Operator enables time travel of OpenShift pods,
            without the need to change the system clock on any of the cluster nodes
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/timemachine-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/timemachine-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=0.0.4 <1.0.17'
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: SolutionSoft Systems, Inc. - support@solution-soft.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Time Machine
            displayName: Time Machine
            kind: TimeMachine
            name: timemachines.tm.solution-soft.com
            version: v1alpha1
        description: |-
          Time Machine operator enables time travel of OpenShift pods/deployments.

          To time travel a pod where your application is, you just need to install Time Machine operator and create a Time Machine custom resource in the same namespace, without the need of changing the container images of the application you’re using, or adding containers to the pod with your application.

          Time travelling is done on the namespace level where all the (configured) target pods/deployments see the same virtual time.
        displayName: Time Machine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - time shift
        - time travel
        - virtual clocks
        - time machine
        - virtual time
        - solution-soft
        links:
        - name: Timemachine Operator
          url: https://catalog.redhat.com/software/operators/detail/5e98727a6c5dcb34dfbb19e3
        maintainers:
        - email: dtrajkovic@solution-soft.com
          name: Dragan Trajkovic
        maturity: alpha
        provider:
          name: Solution-Soft
          url: www.solution-soft.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0c304c11986790f446604319dfc29db494b15f5f9fa13f2c436688ec29c92dbe
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
        - registry.connect.redhat.com/rhel-timemachine-container/virtual-clock-testing-tool@sha256:a81a823328b0ee795c40668c89a08bcf545c5192d9eed0146071e3721a0a8372
        version: 1.0.17
      entries:
      - name: timemachine-operator.v1.0.17
        version: 1.0.17
      - name: timemachine-operator.v1.0.16
        version: 1.0.16
      name: alpha
    defaultChannel: alpha
    packageName: timemachine-operator-rhmp
    provider:
      name: Solution-Soft
      url: www.solution-soft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: ironic-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ironic-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "Ironic",
                "metadata": {
                  "name": "ironic"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseInstance": "openstack",
                  "ironicAPI": {},
                  "ironicConductors": [
                    {
                      "storageRequest": "10G"
                    }
                  ],
                  "ironicInspector": {},
                  "ironicNeutronAgent": {},
                  "secret": "osp-secret",
                  "serviceUser": "ironic",
                  "storageClass": "local-storage"
                }
              },
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "IronicAPI",
                "metadata": {
                  "name": "ironicapi-sample"
                },
                "spec": {
                  "databaseHostname": "openstack.ironic-kuttl-tests.svc"
                }
              },
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "IronicConductor",
                "metadata": {
                  "name": "ironicconductor-sample"
                },
                "spec": {
                  "databaseHostname": "openstack.ironic-kuttl-tests.svc",
                  "storageRequest": "10G"
                }
              },
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "IronicInspector",
                "metadata": {
                  "name": "ironicinspector-sample"
                },
                "spec": {
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "IronicNeutronAgent",
                "metadata": {
                  "name": "ironic-neutron-agent-sample",
                  "namespace": "openstack"
                },
                "spec": {
                  "secret": "osp-secret"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:02Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IronicAPI is the Schema for the ironicapis API
            displayName: Ironic API
            kind: IronicAPI
            name: ironicapis.ironic.openstack.org
            version: v1beta1
          - description: IronicConductor is the Schema for the ironicconductors Conductor
            displayName: Ironic Conductor
            kind: IronicConductor
            name: ironicconductors.ironic.openstack.org
            version: v1beta1
          - description: IronicInspector is the Schema for the ironicinspectors
            displayName: Ironic Inspector
            kind: IronicInspector
            name: ironicinspectors.ironic.openstack.org
            version: v1beta1
          - description: IronicNeutronAgent is the Schema for the ML2 networking-baremetal's
              agent
            displayName: Ironic Neutron Agent
            kind: IronicNeutronAgent
            name: ironicneutronagents.ironic.openstack.org
            version: v1beta1
          - description: Ironic is the Schema for the ironics API
            displayName: Ironic
            kind: Ironic
            name: ironics.ironic.openstack.org
            version: v1beta1
        description: Ironic Operator
        displayName: Ironic Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Identity
        - Ironic
        links:
        - name: Ironic Operator
          url: https://github.com/openstack-k8s-operators/ironic-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/ironic-rhel9-operator@sha256:184a1783a3b2a38bdb23f9811243414a2b21c16d78ca79ed83e2bcaa720d199b
        - registry.redhat.io/rhoso/openstack-ironic-api-rhel9@sha256:86c5330a3065a11f251e8c333bd855d37294c4e9d36cc68208fd16b3f95ef15a
        - registry.redhat.io/rhoso/openstack-ironic-conductor-rhel9@sha256:4950d74335eb48ed04f1a4f602730f1dfd42f139233a4edfebef0d8403c00b2d
        - registry.redhat.io/rhoso/openstack-ironic-inspector-rhel9@sha256:34ff13efc254d4390fbef0a5c088c569241d44d063ec16aed42e36d8179e0c85
        - registry.redhat.io/rhoso/openstack-ironic-pxe-rhel9@sha256:d49af96baa8d90dad3b244ed8c006a9945c41d05e5fa197cc35cc662b49cdd7c
        - registry.redhat.io/rhoso/openstack-ironic-neutron-agent-rhel9@sha256:132627eaf483353c5daefbbc84c6b06f51c50d0813a229ddc558e9df18bb5578
        - registry.redhat.io/rhoso/ironic-python-agent-rhel9@sha256:f41166809d2e7d76446c01f81d325a320df21e64f4b776d0f83cadcb9387eb7f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: ironic-operator.v1.0.6
        version: 1.0.6
      - name: ironic-operator.v1.0.5
        version: 1.0.5
      - name: ironic-operator.v1.0.4
        version: 1.0.4
      - name: ironic-operator.v1.0.3
        version: 1.0.3
      - name: ironic-operator.v1.0.2
        version: 1.0.2
      - name: ironic-operator.v1.0.1
        version: 1.0.1
      - name: ironic-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: ironic-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator.v1.6.3
        version: 1.6.3
      - name: dynatrace-operator.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator.v1.4.0
        version: 1.4.0
      - name: dynatrace-operator.v1.3.2
        version: 1.3.2
      - name: dynatrace-operator.v1.3.1
        version: 1.3.1
      - name: dynatrace-operator.v1.3.0
        version: 1.3.0
      - name: dynatrace-operator.v0.6.0
        version: 0.6.0
      - name: dynatrace-operator.v0.5.1
        version: 0.5.1
      - name: dynatrace-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Fujitsu
      provider-url: https://www.postgresql.fastware.com/
    name: fujitsu-enterprise-postgres-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fujitsu-enterprise-postgres-operator.v7.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPAction",
                "metadata": {
                  "name": "new-fep-action"
                },
                "spec": {
                  "fepAction": {
                    "args": [
                      "new-fep-sts-0",
                      "new-fep-sts-1"
                    ],
                    "type": "reload"
                  },
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "targetClusterName": "new-fep"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPExporter",
                "metadata": {
                  "name": "new-fep-exporter"
                },
                "spec": {
                  "fepExporter": {
                    "exporterLogLevel": "error",
                    "fepClusterList": [
                      "new-fep1"
                    ],
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "userCustomQueries": "usr_example:\n  query: \"SELECT EXTRACT(EPOCH FROM (now() - pg_last_xact_replay_timestamp())) as lag\"\n  master: true\n  metrics:\n    - lag:\n        usage: \"GAUGE\"\n        description: \"Replication lag behind master in seconds\""
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPLogging",
                "metadata": {
                  "name": "new-fep-logging"
                },
                "spec": {
                  "fepLogging": {
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPPgpool2",
                "metadata": {
                  "name": "new-fep-pgpool2"
                },
                "spec": {
                  "clientAuthMethod": "scram",
                  "count": 2,
                  "customhba": "local   all         all                               trust\nhost    all         all         127.0.0.1/32          trust\nhost    all         all         ::1/128               trust\n",
                  "customlogsize": "128Mi",
                  "customparams": "listen_addresses = '*'\npcp_listen_addresses = '*'\nnum_init_children = 32\nreserved_connections = 0\nenable_pool_hba = off\nallow_clear_text_frontend_auth = off\nauthentication_timeout = 80\nbackend_weight0 = 1\nbackend_weight1 = 1\nbackend_flag0 = 'ALWAYS_PRIMARY'\nbackend_flag1 = 'DISALLOW_TO_FAILOVER'\nconnection_cache = on\nmax_pool = 4\nlisten_backlog_multiplier = 2\nserialize_accept = off\nchild_life_time = 300\nclient_idle_limit = 0\nchild_max_connections = 0\nconnection_life_time = 0\nreset_query_list = 'ABORT; DISCARD ALL'\nclient_min_messages = info\nlog_min_messages = debug1\nlog_statement = on\nlog_per_node_statement = on\nlog_client_messages = on\nlog_hostname = on\nlog_connections = on\nlog_line_prefix = '%t: pid %p: '\nload_balance_mode = on\nignore_leading_white_space = on\nwhite_function_list = ''\nblack_function_list = 'currval,lastval,nextval,setval'\nblack_query_pattern_list = ''\ndatabase_redirect_preference_list = ''\napp_name_redirect_preference_list = ''\nallow_sql_comments = off\ndisable_load_balance_on_write = 'transaction'\nstatement_level_load_balance = on\nsr_check_period = 0\nsr_check_user = 'postgres'\ndelay_threshold = 0\nlog_standby_delay = 'none'\nssl = on\nssl_ciphers = 'HIGH:MEDIUM:+3DES:!aNULL'\nssl_prefer_server_ciphers = off\nssl_ecdh_curve = 'prime256v1'\nssl_dh_params_file = ''\nrelcache_expire = 0\nrelcache_size = 256\ncheck_temp_table = catalog\ncheck_unlogged_table = on\nenable_shared_relcache = off\nrelcache_query_target = primary\nwd_port0 = 9000\nfailover_on_backend_error = off\n",
                  "custompcp": "none",
                  "customsslcacert": "none",
                  "customsslcert": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                  "customsslkey": "none",
                  "fepclustername": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "limits": {
                    "cpu": "400m",
                    "memory": "512Mi"
                  },
                  "requests": {
                    "cpu": "200m",
                    "memory": "256Mi"
                  },
                  "serviceport": 9999,
                  "statusport": 9898
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPRestore",
                "metadata": {
                  "name": "new-fep-restore"
                },
                "spec": {
                  "fromFEPcluster": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "mcSpec": {
                    "limits": {
                      "cpu": "200m",
                      "memory": "300Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "restoretype": "latest",
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "toFEPcluster": "new-fep-2"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v2",
                "kind": "FEPCluster",
                "metadata": {
                  "name": "new-fep"
                },
                "spec": {
                  "fep": {
                    "customAnnotations": {
                      "allDeployments": {}
                    },
                    "forceSsl": true,
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "instances": 1,
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "podAntiAffinity": false,
                    "podDisruptionBudget": false,
                    "servicePort": 27500,
                    "syncMode": "off",
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "usePodName": true
                  },
                  "fepChildCrVal": {
                    "backup": {
                      "image": {
                        "pullPolicy": "IfNotPresent"
                      },
                      "mcSpec": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      },
                      "pgbackrestParams": "# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[global]\nrepo1-retention-full=7\nrepo1-retention-full-type=time\nlog-path=/database/log/backup\n",
                      "postScript": " ",
                      "preScript": " ",
                      "schedule": {
                        "num": 2
                      },
                      "schedule1": {
                        "schedule": "15 0 * * 0",
                        "type": "full"
                      },
                      "schedule2": {
                        "schedule": "15 0 * * 1-6",
                        "type": "incr"
                      },
                      "schedule3": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule4": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule5": {
                        "schedule": " ",
                        "type": " "
                      }
                    },
                    "customPgAudit": "# define pg audit custom params here to override defaults.\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[output]\nlogger = 'auditlog'\nlog_directory = '/database/log/audit'\nlog_truncate_on_rotation = on\nlog_filename = 'pgaudit-%a.log'\nlog_rotation_age = 1d\nlog_rotation_size = 0\n[rule]\n",
                    "customPgHba": "# define pg_hba custom rules here to be merged with default rules.\n# TYPE     DATABASE        USER        ADDRESS        METHOD\n",
                    "customPgParams": "# define custom postgresql.conf parameters below to override defaults.\n# Current values are as per default FEP deployment\n# If you add a library to shared_preload_libraries, add it after the default value.\n# fsep_operator_security is available in FEP15 and later container images.\nshared_preload_libraries='pgx_datamasking,pg_prewarm,pg_stat_statements,fsep_operator_security'\nsession_preload_libraries='pg_prewarm'\nmax_prepared_transactions = 100\nmax_worker_processes = 30\nmax_connections = 100\nwork_mem = 1MB\nmaintenance_work_mem = 12MB\nshared_buffers = 128MB\neffective_cache_size = 384MB\ncheckpoint_completion_target = 0.8\n# tcp parameters\ntcp_keepalives_idle = 30\ntcp_keepalives_interval = 10\ntcp_keepalives_count = 3\n# logging parameters in default fep installation\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\nlog_directory = '/database/log'\nlog_filename = 'logfile-%a.log'\nlog_file_mode = 0600\nlog_truncate_on_rotation = on\nlog_rotation_age = 1d\nlog_rotation_size = 0\nlog_checkpoints = on\nlog_line_prefix = '%e %t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h'\nlog_lock_waits = on\nlog_autovacuum_min_duration = 60s\nlogging_collector = on\npgaudit.config_file='/opt/app-root/src/pgaudit-cfg/pgaudit.conf'\nlog_replication_commands = on\nlog_min_messages = WARNING\nlog_destination = csvlog\n# wal_archive parameters in default fep installation\narchive_mode = on\narchive_command = 'pgbackrest --stanza=backupstanza --config=/database/userdata/pgbackrest.conf archive-push %p'\nwal_level = replica\nmax_wal_senders = 12\nwal_keep_segments = 64\ntrack_activities = on\ntrack_counts = on\npassword_encryption = 'scram-sha-256'\n",
                    "storage": {
                      "archivewalVol": {
                        "size": "1Gi"
                      },
                      "backupVol": {
                        "size": "2Gi"
                      },
                      "dataVol": {
                        "size": "2Gi"
                      },
                      "logVol": {
                        "size": "1Gi"
                      },
                      "tablespaceVol": {
                        "size": "512Mi"
                      },
                      "walVol": {
                        "size": "1200Mi"
                      }
                    },
                    "sysUsers": {
                      "pgRewindPassword": "rewind_password",
                      "pgRewindUser": "rewind_user",
                      "pgdb": "mydb",
                      "pgpassword": "mydbpassword",
                      "pgreplpassword": "repluserpwd",
                      "pgrepluser": "repluser",
                      "pguser": "mydbuser",
                      "tdepassphrase": "tde-passphrase"
                    },
                    "systemCertificates": {
                      "cacrt": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                      "crt": "-----BEGIN CERTIFICATE-----\nMIIDUTCCAjmgAwIBAgIRAMocW3qMoHrD6qRvMPppMkMwDQYJKoZIhvcNAQELBQAw\nNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9yIEt1\nYmVybmV0ZXMwHhcNMjEwMjA2MDQzMjM2WhcNMjYwMjA1MDQzMjM2WjA/MRAwDgYD\nVQQKEwdGdWppdHN1MSswKQYDVQQDEyJGVUpJVFNVIEVudGVycHJpc2UgUG9zdGdy\nZXMgU2VydmVyMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA4AI33yvH\nZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I2e4SceTKi6O3C/I1XuvWlpng\n5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4ANB5JWWqDOjrRT3o7nRPGXfila\nbP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYpmjdbfxabTz07ig0+6/cwKoRR\nxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTFYvmAH7gcdssSFBt8NPlUATHE\nsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6Wdgmu5H2pDml8CDNLDv98Aj7i\n+I5SRKKcVPlnuQIDAQABo1AwTjAdBgNVHSUEFjAUBggrBgEFBQcDAQYIKwYBBQUH\nAwIwDAYDVR0TAQH/BAIwADAfBgNVHSMEGDAWgBQcwrrUO0u+FhIUuVdrDRCQRsi6\nZjANBgkqhkiG9w0BAQsFAAOCAQEAm5dxBoI9pScOCvRAchg4CprdRDSJb9K6yB3O\nnCAxnM47iHeXnY3WlnI388kHu8DU7O4ba1tJbGs3KY9KzioPk43pU12jWkO1onoF\n+mTDjx/Ef1cYWA9r5q/LtgTa6Q2sxV4O2x67QW82aAnaxO34dV5zWCPIvAoovZBV\nHRT+BgCg3r2vD1RGKK2nl1aYJtWhO1SZubam+VttdZ/vbM9oOJctxmImsEtBXjkY\nKteePdQtLL5o03JhyXWyRshCq+HMmKf2KgyY8gvydGcP4eLQdBWcW40LcnVq6UjT\n0kJycJEKngMVademq1ZWHGaiYB7hyT6GhgIcHUJ2cKrPgbEh1Q==\n-----END CERTIFICATE-----",
                      "key": "-----BEGIN RSA PRIVATE KEY-----\nMIIEowIBAAKCAQEA4AI33yvHZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I\n2e4SceTKi6O3C/I1XuvWlpng5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4AN\nB5JWWqDOjrRT3o7nRPGXfilabP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYp\nmjdbfxabTz07ig0+6/cwKoRRxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTF\nYvmAH7gcdssSFBt8NPlUATHEsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6W\ndgmu5H2pDml8CDNLDv98Aj7i+I5SRKKcVPlnuQIDAQABAoIBAFPQYKlOzw/+BA0b\nyMIUpdctIMb/54CR/xR0mVw1DbSjigNVPjHUQvB8Y1B2FAITQObgJO06bAv0QdWN\nRb0/v/yYiNJDFjaLjaIAHlO/2+oWrXbFaZqgpVDJhB+e1xaZr2x7XGxm+p925k30\nl6pvIRY+I8JRKvZiV1VZHwL/R3JOtPr++xMZtLVjVOI+f+ySqJ+TZHuAjm49EKxj\ncEmmJ28b7QcziXsvKy00f+zbqLIBKXQdZAFU5eEr1BsDRXdRW+Kf0XIvftuy4BJZ\nvoKT+VGhEvF/qysswL4+6IAO6tpuYnnM0Y2d3sOGoWPkTcQK0MekYKzL/WmtCjNs\n9hodJtECgYEA5EWyhEOf4uOKe5TDp697UCUvXLoOR58FDe/S8XNvScn29jjOkqIg\nOMoqo9xAkJTNTzqn5UUdt1x/pgM2NxlPLFijrc0zQlX3SoOO2ryDd9WNi7YKtN16\nKJqa536WeZu2OEbuAZ+S3GALVy1RPeTNPnUOmKnF06DjDUGzLNCZy10CgYEA+zfw\n952DWuz1U0Z4wvAEqqcgUKXPKrkTXV/iUnjkDkrLYVr0ZofDNTXrdHl+UedFmaOC\ncieZn6DNhcdz5tKtyysGMH3g/qs9PfoGUngvcXsy0Egk04l3x1jc8TTCLqXZXYaQ\nHMsx51n+R58oncPtzYSUOr9qQ6PbC2CstTbFJA0CgYEAjGEsUliAB/jknfEzjXjG\nPdhQUxb8VyE864Az2lah9t/kJzFyIAziAeqZ5GE7t247AGFTBRTHHI8e1Qoemi3P\nWbc9GVIbFs1lIYbcIDpUIyrKPEP8O5QEXtoNLxXTFgAjRGKiVY87spjCAJ+W2ZhO\ne/1it5GYXfgQCYQA2yuBmOUCgYANRkR2YR1axaCk+NlSu6oTdmdPu6M5x7PNQE7O\nOtMaKjua9lppvIzFGAdMDUtueoEEAE7ZR1xnwfB6PDLUpJdIYAqgr1YfPt8qkjaZ\nTv56yZ7CwL0pbF8m6nwqRrZoDp1wwraEvvvxFKFKGY/k3kCHlpTakdjEoDjn3gDi\nRnWeVQKBgCEneMSzucei5LRppRtRaJw/Btll8qlPMlX3W7dxQ3cLwpmLOn0m51Fp\nPIZ44zYK8R6fu4+/sSrlfaIg86Ugeufp6YNxyNROKxUGza5vDIu5OftwWtBeg+UK\nZ8lLWNdX6pp7WMujmF3H1DrkBbauYMUKZ4UxUYtelgHERMePIxwb\n-----END RSA PRIVATE KEY-----"
                    }
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "false"
          containerImage: quay.io/fujitsu/fujitsu-enterprise-postgres-operator@sha256:b7dd55c0bfda84e398cbfbb87345fcec218cb866e286e18a111330321ab6ad54
          createdAt: "2025-09-10T11:50:37Z"
          description: OpenShift Operator for Fujitsu Enterprise Postgres
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.1.0 <7.3.0'
          operators.openshift.io/valid-subscription: '["Fujitsu Enterprise Postgres
            Advanced Edition Annual Subscription License per Core 17 Operator for
            Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/internal-objects: '["fepautoscales.fep.fujitsu.io","fepconfigs.fep.fujitsu.io","fepusers.fep.fujitsu.io","fepcerts.fep.fujitsu.io","fepvolumes.fep.fujitsu.io","fepbackups.fep.fujitsu.io"]'
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: fj-dbaas-tec@dl.jp.fujitsu.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: FEPAction
            name: fepactions.fep.fujitsu.io
            version: v1
          - kind: FEPAutoscale
            name: fepautoscales.fep.fujitsu.io
            version: v1
          - kind: FEPBackup
            name: fepbackups.fep.fujitsu.io
            version: v1
          - kind: FEPCert
            name: fepcerts.fep.fujitsu.io
            version: v1
          - kind: FEPCluster
            name: fepclusters.fep.fujitsu.io
            version: v2
          - kind: FEPConfig
            name: fepconfigs.fep.fujitsu.io
            version: v1
          - kind: FEPExporter
            name: fepexporters.fep.fujitsu.io
            version: v1
          - kind: FEPLogging
            name: feploggings.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2Cert
            name: feppgpool2certs.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2
            name: feppgpool2s.fep.fujitsu.io
            version: v1
          - kind: FEPRestore
            name: feprestores.fep.fujitsu.io
            version: v1
          - kind: FEPUpgrade
            name: fepupgrades.fep.fujitsu.io
            version: v1
          - kind: FEPUser
            name: fepusers.fep.fujitsu.io
            version: v1
          - kind: FEPVolume
            name: fepvolumes.fep.fujitsu.io
            version: v1
        description: "Fujitsu Enterprise Postgres 17 delivers an enterprise-grade
          PostgreSQL on OpenShift Container Platform.\n\nThis solution provides the
          flexibility of a hybrid cloud solution while delivering an enhanced distribution
          of PostgreSQL to support enterprise-level workloads and provide improved
          deployment and management, availability, performance, data governance and
          security.  \n\nAvailable as a multi-architecture container built for both
          amd64, s390x and ppc64le.\n\nThe download and Use of the Product is strictly
          subject to the terms of the End User License Agreement with Fujitsu Limited
          found at https://www.fast.fujitsu.com/fujitsu-enterprise-postgres-license-agreements.
          Where the Product that has been embedded as a whole or part into a third
          party program, only Authorised Customers may download and use the Product.\n"
        displayName: Fujitsu Enterprise Postgres Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - Fujitsu
        - fep
        links:
        - name: Fujitsu Enterprise Postgres
          url: https://www.postgresql.fastware.com/
        maintainers:
        - email: fj-dbaas-tec@dl.jp.fujitsu.com
          name: Fujitsu
        maturity: stable
        provider:
          name: Fujitsu
          url: https://www.postgresql.fastware.com/
        relatedImages:
        - quay.io/fujitsu/fujitsu-enterprise-postgres-operator@sha256:b7dd55c0bfda84e398cbfbb87345fcec218cb866e286e18a111330321ab6ad54
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-pgpool2@sha256:8eaace2f9c8e5e844992fe5e008cd1061af33cc79b4d442c4a83a65610bcc502
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-backup@sha256:ac268e2af652bdabb8ee5d764a2b335dc73c48b9022eee8c9f3cd422227a4d98
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-backup@sha256:78a3ad639d0c63858b9db0a5de36ce11a9ea121f4535e9ca07ac7defb689bc0d
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-upgrade@sha256:04af997b4f548ffec711b63a06c35715b4e7863537c83697b3ee0c1d95a6aec3
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentd@sha256:3c5534b62695867c5973c9106fc46895ed73a4b9136ad83b2586bf7e71c427d0
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentbit@sha256:9622d2191525da39e5100e1882dd2cfa3eda740d8094f3fbb71ca8a5734d2843
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-server@sha256:1c5823afc7e77d46940214693c12c0ad6ffd6c7d9a156e26399ab55cd5ea1dfa
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-pgpool2@sha256:c878935ffc796c405d811a6e8a9194aaaaa8652ea7d7b879c82919ce09794b35
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-upgrade@sha256:8a0f7ca1f59ba162424a308e8a70c4717135e2847e61210753b71876c9b0a75f
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-backup@sha256:dd8b84e200c9d0fb32cd567a33968e15e15efc115ff120f54fdd8862c7ea3111
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-upgrade@sha256:c0e44447349f02eb9f6105001136402f849d876a288d61997c8c05625facff45
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-utils@sha256:b7ae535f5e5fa5eb51c6f4cdb406c7b7a9dabe8573939b0431dada6f7d27b87e
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-restore@sha256:a6e90dd77634033f2009942b635078cceb0f4c6b15b031624c5dce820713b14c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-restore@sha256:17301c2191c0d75be82d12c5a35c0f74b8d65304bcfb6d4ab312a10ffa286eda
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-server@sha256:fa4e8849d75daf331ba7d488bcb52c083cd92bb5f288e4b62c980754e299f618
        - quay.io/fujitsu/fujitsu-enterprise-postgres-cronjob@sha256:7582964eef14f89f2ca9046298357afd0132fb236f30f44f84cee14692227b0c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-server@sha256:5c65f8e76f70341c512100760de6792a880ad0d6eebc9e68154b5bbd1332a896
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-restore@sha256:09be02274427e3022849fb635a4c9abdc34696838cae39f3afafc8009567db9d
        - quay.io/fujitsu/fujitsu-enterprise-postgres-exporter@sha256:1c0e22aeab65ff3848772424fc3ea6baa6b58b0d88db5413bda160a08435bcff
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-upgrade@sha256:bf3b069a71202ea78c7c0c7b36315c8bef8e6a1c11cd0668dcd888329e26ecfb
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-pgpool2@sha256:38aac8108dd3590abcfd854a6ec2c025a68d2170a1ba68d359b45fea6c83cec1
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-server@sha256:3e93a368a6198818e1b7c5525324c8130c5ed8d7b4cb499a47d4bec45ef0347b
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-pgpool2@sha256:185af2e4fd1701aded1ff5e44d4336562f5a219c30ba535c2a3097b914c24efe
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-backup@sha256:42252444e2dcfe5435532f318cce041990d2b71efa26007924717924d30a46ca
        - quay.io/fujitsu/fujitsu-enterprise-postgres-utils@sha256:57a3fb792480a0d3a8f075c48dee2d02f1d79cf459761e3305cf62799575a885
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-restore@sha256:a81a0ce926572b75aa3151e2d2cc45a0ca411cc0158d69b89247ef70418f3740
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-utils@sha256:edeec4d00fd8282aaa7ec6fe3ec4cd84070a6402611002a18addcac87fb8ee94
        version: 7.3.0
      entries:
      - name: fujitsu-enterprise-postgres-operator.v7.3.0
        version: 7.3.0
      - name: fujitsu-enterprise-postgres-operator.v7.2.7
        version: 7.2.7
      name: stable
    defaultChannel: stable
    packageName: fujitsu-enterprise-postgres-operator
    provider:
      name: Fujitsu
      url: https://www.postgresql.fastware.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-block-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12
    - currentCSV: ibm-block-csi-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "k8s.gcr.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
        version: 1.12.0
      entries:
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.0
    - currentCSV: ibm-block-csi-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
        version: 1.12.1
      entries:
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.1
    - currentCSV: ibm-block-csi-operator.v1.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:af9215670675fe9c4fa2dac7a14d547c249f3aa0c0e72f1e5390d237e411735b
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:af9215670675fe9c4fa2dac7a14d547c249f3aa0c0e72f1e5390d237e411735b
        version: 1.12.2
      entries:
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.2
    - currentCSV: ibm-block-csi-operator.v1.12.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:6e7e13d259f4d743524ebcd546af5c73d1a3ba90edaf4b5edd2005a3f8adcdfe
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:6e7e13d259f4d743524ebcd546af5c73d1a3ba90edaf4b5edd2005a3f8adcdfe
        version: 1.12.3
      entries:
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.3
    - currentCSV: ibm-block-csi-operator.v1.12.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:42728e7fd1d2302511ab08f51335ab188eef125870fa27f322df83ebb8e1f447
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:42728e7fd1d2302511ab08f51335ab188eef125870fa27f322df83ebb8e1f447
        version: 1.12.4
      entries:
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.4
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.5
    defaultChannel: stable
    packageName: ibm-block-csi-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cli-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cli-manager-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "CliManager",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-cli-manager-operator"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed",
                  "operatorLogLevel": "Normal"
                }
              }
            ]
          capabilities: Basic Install
          categories: Openshift Optional
          createdAt: "2024-06-04"
          description: "The CLI Manager is a comprehensive tool designed to simplify
            the management of \nOpenShift CLI plugins within the OpenShift environment.
            \nModeled after the popular krew plugin manager, it offers seamless integration,
            \neasy installation, and efficient handling of a wide array of plugins,
            \nenhancing your OpenShift command-line experience.\n"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          repository: https://github.com/openshift/cli-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: CLI Manager
            kind: CliManager
            name: climanagers.operator.openshift.io
            version: v1
          - displayName: CLI Plugin
            kind: Plugin
            name: plugins.config.openshift.io
            version: v1alpha1
        description: "The CLI Manager is a comprehensive tool designed to simplify
          the management of \nOpenShift CLI plugins within the OpenShift environment.
          \nModeled after the popular krew plugin manager, it offers seamless integration,
          \neasy installation, and efficient handling of a wide array of plugins,
          \nenhancing your OpenShift command-line experience.\n"
        displayName: CLI Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cli-manager-operator
        - cli-manager
        links:
        - name: Source Code
          url: https://github.com/openshift/cli-manager
        - name: Source Code
          url: https://github.com/openshift/cli-manager-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/cli-manager-operator/cli-manager-rhel9-operator@sha256:31973d7f43d052ddfb9554dd623e37fb541a8c1f08ea981558eaf4bb5ab22d5a
        - registry.redhat.io/cli-manager-operator/cli-manager-rhel9@sha256:07853bc33a0ce783d8d45c08c161e454a3abe09a8a8dadeca70c3a199466b58b
        version: 0.1.0
      entries:
      - name: cli-manager-operator.v0.1.0
        version: 0.1.0
      name: tech-preview
    defaultChannel: tech-preview
    packageName: cli-manager
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cryostat-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cryostat-operator.v4.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.cryostat.io/v1beta2",
                "kind": "Cryostat",
                "metadata": {
                  "name": "cryostat-sample"
                },
                "spec": {
                  "enableCertManager": true,
                  "eventTemplates": [],
                  "networkPolicies": {},
                  "reportOptions": {
                    "replicas": 0
                  },
                  "trustedCertSecrets": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Developer Tools
          containerImage: registry.redhat.io/cryostat/cryostat-rhel9-operator@sha256:867a649b54ad144295c7aceda6b90631841aa31a62062df62d4980313723f359
          createdAt: "2025-09-30T15:10:08Z"
          description: JVM monitoring and profiling tool
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.0 <4.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "operator.cryostat.io/v1beta2",
              "kind": "Cryostat",
              "metadata": {
                "name": "cryostat-sample"
              },
              "spec": {
                "enableCertManager": true,
                "reportOptions": {
                  "replicas": 0
                }
              }
            }
          operators.openshift.io/valid-subscription: '["Red Hat build of OpenJDK"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/cryostatio/cryostat-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cryostat allows you to install Cryostat for a single namespace,
              or multiple namespaces. It contains configuration options for controlling
              the Deployment of the Cryostat application and its related components.
              A Cryostat instance must be created to instruct the operator to deploy
              the Cryostat application.
            displayName: Cryostat
            kind: Cryostat
            name: cryostats.operator.cryostat.io
            version: v1beta2
          - description: Cryostat allows you to install Cryostat for a single namespace.
              It contains configuration options for controlling the Deployment of
              the Cryostat application and its related components. A Cryostat instance
              must be created to instruct the operator to deploy the Cryostat application.
            displayName: Cryostat
            kind: Cryostat
            name: cryostats.operator.cryostat.io
            version: v1beta1
        description: |
          The Red Hat build of Cryostat provides a cloud-based solution for interacting with the JDK Flight Recorder already present in OpenJDK 11+ JVMs. With Cryostat, users can remotely start, stop, retrieve, and even analyze JFR event data, providing the capability to easily take advantage of Flight Recorder's extremely low runtime cost and overhead and the flexibility to monitor applications and analyze recording data without transferring data outside of the cluster the application runs within.
          ##Prerequisites
          Cryostat requires [cert-manager](https://cert-manager.io/) to run. If not already installed in your cluster, please [install](https://cert-manager.io/docs/installation/) it using your preferred method.
          When running on Kubernetes, the operator also requires [Ingress configurations](https://github.com/cryostatio/cryostat-operator/blob/v4.0.0/docs/config.md#network-options) for each of its services to make them available outside of the cluster. The user is responsible for providing the hostnames for each Ingress.
        displayName: Red Hat build of Cryostat
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flightrecorder
        - java
        - jdk
        - jfr
        - jmc
        - missioncontrol
        - monitoring
        - profiling
        - diagnostic
        links:
        - name: Upstream Project
          url: https://github.com/cryostatio/cryostat
        - name: Website
          url: https://cryostat.io/
        maintainers:
        - email: cryostat-development@googlegroups.com
          name: The Cryostat Authors
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi9/nginx-124@sha256:7c1d7172880a3649591a0dc1ceb670c133ef88eed3d8192d8a2e80e3878d04c3
        - registry.redhat.io/cryostat/cryostat-reports-rhel9@sha256:5535a539b185fa2781621741ccd7f10a6b9947adc6170a95e930f2d9d7a9f608
        - registry.redhat.io/insights-runtimes-tech-preview/runtimes-inventory-rhel8-operator@sha256:58e01d11c3d30abc5b0ee3795511bd04e39aea21e2853b4f4cc6d1c8a2bd7bb0
        - registry.redhat.io/cryostat/cryostat-rhel9@sha256:fadae8713daf82c329573191381f8f9e4ebf3fc852ac5f1a68ce0ea1da376469
        - registry.redhat.io/cryostat/jfr-datasource-rhel9@sha256:1e5f29aa096d783fd0a18b3ba31c391dc203761ba3f4af8e5aeeef1f7cc86cee
        - registry.redhat.io/cryostat/cryostat-db-rhel9@sha256:23d61204ecaf037cf17bf9a5cc4428461a60fecce077884d9458802fb071daab
        - registry.redhat.io/cryostat/cryostat-grafana-dashboard-rhel9@sha256:5715e897f9a8969aedd0066b2ba597264495b4c85e4ae8ae7d8871f93d13f4c1
        - registry.redhat.io/cryostat/cryostat-agent-init-rhel9@sha256:273d6c3d3194a4e9ec50c1d8e147cd9bc6d277d8d66878d885a18528b9ba7dc6
        - registry.redhat.io/cryostat/cryostat-rhel9-operator@sha256:867a649b54ad144295c7aceda6b90631841aa31a62062df62d4980313723f359
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        - registry.redhat.io/cryostat/cryostat-ose-oauth-proxy-rhel9@sha256:d38ce4dcac5b073b001d886f0494e0fc9b750f8bd68b855fc103767a1ed8b7df
        - registry.redhat.io/cryostat/cryostat-storage-rhel9@sha256:f642ca8a0ca0aa130462c0f423ce4c775cb296f8a40c9270e5369176e87c7ee6
        - registry.redhat.io/cryostat/cryostat-openshift-console-plugin-rhel9@sha256:ca748f96414ad648eb1e3953aeeeb828b187f3c5c30e014513adc3619a698683
        version: 4.0.3
      entries:
      - name: cryostat-operator.v4.0.3
        version: 4.0.3
      - name: cryostat-operator.v4.0.2-4
        version: 4.0.2+4
      - name: cryostat-operator.v4.0.2
        version: 4.0.2
      - name: cryostat-operator.v4.0.1
        version: 4.0.1
      - name: cryostat-operator.v4.0.1-5
        version: 4.0.1+5
      - name: cryostat-operator.v4.0.1-4
        version: 4.0.1+4
      - name: cryostat-operator.v4.0.1-3
        version: 4.0.1+3
      - name: cryostat-operator.v4.0.1-2
        version: 4.0.1+2
      - name: cryostat-operator.v4.0.0-11
        version: 4.0.0+11
      - name: cryostat-operator.v4.0.0
        version: 4.0.0
      - name: cryostat-operator.v4.0.0-9
        version: 4.0.0+9
      - name: cryostat-operator.v4.0.0-10
        version: 4.0.0+10
      - name: cryostat-operator.v3.0.1-7
        version: 3.0.1+7
      - name: cryostat-operator.v3.0.1-9
        version: 3.0.1+9
      - name: cryostat-operator.v3.0.1-8
        version: 3.0.1+8
      - name: cryostat-operator.v3.0.1
        version: 3.0.1
      - name: cryostat-operator.v3.0.1-6
        version: 3.0.1+6
      - name: cryostat-operator.v3.0.1-5
        version: 3.0.1+5
      - name: cryostat-operator.v3.0.1-3
        version: 3.0.1+3
      - name: cryostat-operator.v3.0.1-13
        version: 3.0.1+13
      - name: cryostat-operator.v3.0.1-12
        version: 3.0.1+12
      - name: cryostat-operator.v3.0.1-11
        version: 3.0.1+11
      - name: cryostat-operator.v3.0.1-10
        version: 3.0.1+10
      - name: cryostat-operator.v3.0.0-12
        version: 3.0.0+12
      - name: cryostat-operator.v3.0.0
        version: 3.0.0
      - name: cryostat-operator.v3.0.0-6
        version: 3.0.0+6
      - name: cryostat-operator.v3.0.0-8
        version: 3.0.0+8
      - name: cryostat-operator.v2.4.0
        version: 2.4.0
      - name: cryostat-operator.v2.4.0-8
        version: 2.4.0+8
      - name: cryostat-operator.v2.4.0-7
        version: 2.4.0+7
      - name: cryostat-operator.v2.4.0-6
        version: 2.4.0+6
      - name: cryostat-operator.v2.4.0-4
        version: 2.4.0+4
      - name: cryostat-operator.v2.4.0-3
        version: 2.4.0+3
      - name: cryostat-operator.v2.3.1-6
        version: 2.3.1+6
      - name: cryostat-operator.v2.3.1-8
        version: 2.3.1+8
      - name: cryostat-operator.v2.3.1-9
        version: 2.3.1+9
      - name: cryostat-operator.v2.3.1
        version: 2.3.1
      - name: cryostat-operator.v2.3.0
        version: 2.3.0
      - name: cryostat-operator.v2.3.0-6
        version: 2.3.0+6
      - name: cryostat-operator.v2.2.1-14
        version: 2.2.1+14
      - name: cryostat-operator.v2.2.1-8
        version: 2.2.1+8
      - name: cryostat-operator.v2.2.1-5
        version: 2.2.1+5
      - name: cryostat-operator.v2.2.1-13
        version: 2.2.1+13
      - name: cryostat-operator.v2.2.1-12
        version: 2.2.1+12
      - name: cryostat-operator.v2.2.1-10
        version: 2.2.1+10
      - name: cryostat-operator.v2.2.1
        version: 2.2.1
      - name: cryostat-operator.v2.2.0
        version: 2.2.0
      - name: cryostat-operator.v2.1.1-2.1659469835.p
        version: 2.1.1+2.1659469835.p
      - name: cryostat-operator.v2.1.1-2.1655334614.p
        version: 2.1.1+2.1655334614.p
      - name: cryostat-operator.v2.1.1
        version: 2.1.1
      - name: cryostat-operator.v2.1.0
        version: 2.1.0
      name: stable
    defaultChannel: stable
    packageName: cryostat-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kove
      provider-url: https://kove.com
    name: kove-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kove-operator.v3.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "containers.kove.com/v1",
                "kind": "KoveMemoryConfig",
                "metadata": {
                  "labels": {
                    "KoveMemoryConfig": ""
                  },
                  "name": "example-memory-config"
                },
                "spec": {
                  "container": "",
                  "namespace": "",
                  "size": ""
                }
              },
              {
                "apiVersion": "containers.kove.com/v1",
                "kind": "KoveSystem",
                "metadata": {
                  "labels": {
                    "KoveSystem": ""
                  },
                  "name": "kove-system"
                },
                "spec": {
                  "autogrowInitialIncrement": "1 GiB",
                  "autogrowMaxIncrement": "1 TiB",
                  "l4MaxMemoryPercent": 75,
                  "managementConsoleAddress": "",
                  "memoryPlusLocalMemoryMaxSize": "0 MiB"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: quay.io/kove/kove-operator@sha256:0fb5fa2e6702d5e5bf692c25f190f65d306fc72cdceebf0814ff64adcb5f6994
          createdAt: "2025-09-25T16:31:36Z"
          description: Configure containers with Kove® pooled memory
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: kove-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Kove
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configure containers with Kove® pooled memory
            displayName: Kove Memory Configuration
            kind: KoveMemoryConfig
            name: kovememoryconfigs.containers.kove.com
            version: v1
          - description: Configure cluster with Kove:SDM™
            displayName: Kove System Configuration
            kind: KoveSystem
            name: kovesystems.containers.kove.com
            version: v1
        description: |
          ## About the Operator
          The Kove:SDM™ operator integrates Kove's software-defined memory
          solution with Kubernetes and OpenShift clusters. The operator enables
          containers to transparently provision any amount of memory, enabling total
          container memory capacity to exceed physical hardware limitations.

          Use the Kove:SDM™ operator to:

          * Create custom resources to automatically allocate memory to specific
            containers
          * Increase container flexibility with dynamic memory capacity that
            auto-grows as needed
          * Scale up container density by removing memory as an orchestration
            bottleneck

          ## Using the Operator
          Using the operator is as simple as creating a custom resource which
          specifies an initial Kove memory allocation size, a namespace, and a
          container name. Containers in the specified namespace with a matching
          container name will automatically be provided with a Kove memory allocation
          which will auto-grow as needed.

          ## About Kove
          Kove makes memory a provisionable resource, just like storage, CPUs, and
          networking. Kove:SDM™ provides a transparent, linearly scalable
          memory resource accessible to any application without code changes, from
          bare-metal apps to virtual machines and containers.
        displayName: Kove:SDM™ Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kove
        - Pooled Memory
        - Memory Pool
        - Memory Target
        - External Memory
        - Memory
        links:
        - name: Kove
          url: https://kove.com
        maintainers:
        - email: contact-kove@kove.com
          name: Kove
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Kove
          url: https://kove.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fa0acefda43e309e7f9f31f5247c933d0b868ff1573ece2f38e0549d7e2ec942
        - registry.access.redhat.com/ubi8/pause@sha256:eacbaaadc932633330e799eda24e9b731b5a4ba7ed823ee70791d21029724a9a
        - quay.io/kove/kove-client-drivers@sha256:1c1786e1bbf16609218154126b4340b2f7c4b71cbd9045589402f75c032ea38d
        - quay.io/kove/kove-xmem@sha256:41bfc536a278153b0a4460d2836b56762b56cf50b94d17e123a63be066c28fe7
        - quay.io/kove/kove-operator@sha256:0fb5fa2e6702d5e5bf692c25f190f65d306fc72cdceebf0814ff64adcb5f6994
        - quay.io/kove/kove-container-hook@sha256:66ee33d59ced63406d9b40f3a3226fab7f52071f8bed9aa1ba7d859837038528
        - quay.io/kove/kove-agent@sha256:9d026b925e1d414a531530a30ed8b5f10b421ee12c7c13949ea0ee9907d36b75
        - quay.io/kove/kove-client-drivers@sha256:716ccda2d8f082a954f2042fd189fec257bd2dcb2c50e497bad42b9a9e5f0e3d
        - quay.io/kove/kove-compatibility-detection@sha256:0303830a1f785b5e4ce6824db7c50f8feee2678fd5e87ded983686ab3f60b232
        version: 3.0.2
      entries:
      - name: kove-operator.v3.0.2
        version: 3.0.2
      - name: kove-operator.v3.0.1
        version: 3.0.1
      - name: kove-operator.v3.0.0
        version: 3.0.0
      name: stable
    defaultChannel: stable
    packageName: kove-operator
    provider:
      name: Kove
      url: https://kove.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom
      provider-url: https://www.broadcom.com/
    name: layer7-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: layer7-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "security.brcmlabs.com/v1",
                "kind": "Gateway",
                "metadata": {
                  "name": "ssg"
                },
                "spec": {
                  "app": {
                    "image": "docker.io/caapim/gateway:11.1.2",
                    "management": {
                      "cluster": {
                        "hostname": "gateway.brcmlabs.com",
                        "password": "7layer"
                      },
                      "password": "7layer",
                      "username": "admin"
                    },
                    "replicas": 1,
                    "service": {
                      "ports": [
                        {
                          "name": "https",
                          "port": 8443,
                          "protocol": "TCP",
                          "targetPort": 8443
                        },
                        {
                          "name": "management",
                          "port": 9443,
                          "protocol": "TCP",
                          "targetPort": 9443
                        }
                      ],
                      "type": "LoadBalancer"
                    }
                  },
                  "license": {
                    "accept": false,
                    "secretName": "gateway-license"
                  },
                  "version": "11.1.2"
                },
                "status": {}
              },
              {
                "apiVersion": "security.brcmlabs.com/v1",
                "kind": "Repository",
                "metadata": {
                  "name": "my-repository"
                },
                "spec": {
                  "auth": {},
                  "branch": "main",
                  "enabled": true,
                  "endpoint": "https://github.com/\u003cusername\u003e/\u003crepository\u003e",
                  "type": "git"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7Api",
                "metadata": {
                  "name": "l7api-sample"
                },
                "spec": {
                  "deploymentTags": [
                    "ssg",
                    "ssg1"
                  ],
                  "graphmanBundle": "jsongz",
                  "portalPublished": false,
                  "serviceUrl": "/v1/api"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7Portal",
                "metadata": {
                  "name": "dev-portal"
                },
                "spec": {
                  "auth": {},
                  "deploymentTags": [
                    "ssg"
                  ],
                  "enabled": true,
                  "endpoint": "dev-ssg.brcmlabs.com",
                  "enrollmentBundle": "",
                  "portalTenant": "portal"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7StateStore",
                "metadata": {
                  "name": "state-store-example"
                },
                "spec": {
                  "redis": {
                    "database": 0,
                    "groupName": "l7gw",
                    "masterPassword": "7layer",
                    "standalone": {
                      "host": "redis-standalone",
                      "port": 6379
                    },
                    "storeId": "test",
                    "type": "standalone"
                  },
                  "type": "redis"
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: docker.io/caapim/layer7-operator:v1.2.1
          createdAt: "2025-07-04T17:22:57Z"
          description: The Layer7 Operator covers all aspects of deploying, maintaining
            and upgrading Layer7 API Gateways in Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: layer7-operator-system
          operators.openshift.io/valid-subscription: Layer7 (Broadcom) API Gateway
            license
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/caapim/layer7-operator
          support: Broadcom Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gateway is the Schema for the Gateway Custom Resource
            displayName: Gateway
            kind: Gateway
            name: gateways.security.brcmlabs.com
            version: v1
          - description: L7Api is the Schema for the l7apis API
            displayName: L7 Api
            kind: L7Api
            name: l7apis.security.brcmlabs.com
            version: v1alpha1
          - description: L7Portal is the Schema for the l7portals API
            displayName: L7 Portal
            kind: L7Portal
            name: l7portals.security.brcmlabs.com
            version: v1alpha1
          - description: L7StateStore is the Schema for the l7statestores API
            displayName: L7 State Store
            kind: L7StateStore
            name: l7statestores.security.brcmlabs.com
            version: v1alpha1
          - description: Repository is the Schema for the repositories API
            displayName: Repository
            kind: Repository
            name: repositories.security.brcmlabs.com
            version: v1
        description: |
          The [Layer7 Operator](https://github.com/CAAPIM/layer7-operator) covers all aspects of deploying, maintaining and upgrading Layer7 API Gateways in Kubernetes. The Layer7 Operator is exclusive to Kubernetes and utilizes the Container Gateway with specific emphasis on running in Ephemeral mode (without an external MySQL policy database).

          ## Getting Started
          This [Getting Started](https://github.com/CAAPIM/layer7-operator/wiki/Getting-Started) guide covers a simple introduction to the Layer7 Operator. You can find more documentation [here](https://github.com/CAAPIM/layer7-operator/wiki).
        displayName: Layer7 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - layer7-operator
        links:
        - name: Layer7 Operator
          url: https://github.com/caapim/layer7-operator
        maintainers:
        - email: gary.vermeulen@broadcom.com
          name: Gary Vermeulen
        maturity: alpha
        provider:
          name: Broadcom
          url: https://www.broadcom.com/
        relatedImages:
        - docker.io/caapim/layer7-operator:v1.2.1
        - docker.io/caapim/gateway:11.1.2
        - docker.io/caapim/graphman-static-init:1.0.3
        version: 1.2.1
      entries:
      - name: layer7-operator.v1.2.1
        version: 1.2.1
      - name: layer7-operator.v1.2.0
        version: 1.2.0
      name: preview
    defaultChannel: preview
    packageName: layer7-operator
    provider:
      name: Broadcom
      url: https://www.broadcom.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: eG Innovations
      provider-url: https://www.eginnovations.com
    name: eginnovations-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eginnovations-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eginnovations.com/v1beta1",
                "kind": "EgUniversalAgent",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eginnovations-operator"
                  },
                  "name": "eguniversalagent",
                  "namespace": "egagent"
                },
                "spec": {
                  "agent": {
                    "env": [
                      {
                        "name": "EG_MANAGER",
                        "value": "Replace with eG Manager IP"
                      },
                      {
                        "name": "EG_MANAGER_PORT",
                        "value": "Replace with eG Manager Port"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "true"
          containerImage: docker.io/eginnovations/universal-agent-operator@sha256:b8e3e26dca1b94fabfd79496feb36f3f2de4434331ef9ca810125892a936dca7
          createdAt: "2025-09-11T20:08:00Z"
          description: eG Innovations provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with eG Universal Agent Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eginnovations/universal-agent-operator
          support: eG Innovations
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgUniversalAgent is the Schema for the eguniversalagents
              API
            displayName: Eg Universal Agent
            kind: EgUniversalAgent
            name: eguniversalagents.eginnovations.com
            version: v1beta1
        description: |
          eG Innovations’ converged application and infrastructure performance monitoring capabilities provide end-to-end visibility and correlated analytics for every layer, every tier of your IT landscape.

          The eG Enterprise Universal Agent Operator automatically configures a host agent on every Kubernetes worker node. The host agent auto-discovers the worker nodes and application containers running as Pods on each node and tracks their performance and utilization levels. In-depth monitoring of applications running on containers is also provided using the same host agent. No additional agents are required for the containers.

          ### Capabilities
            * Rolls out the egagent pod per node to monitor its pods and the node itself
            * Automatic BTM profiler injection using webhook based injection for deep application monitoring and transection tracing.

          ### Installation
            1. Create the namespace "egagent" for installing eG Agent on OpenShift Cluster.

               <br />``` $ oc create ns egagent ```
            2. Add "egagent" serviceaccount to scc( Security Context Constraints ) of OpenShift Cluster as privileged using below command.

               <br />``` $ oc adm policy add-scc-to-user privileged -z egagent -n egagent ```
            2. Install eG Innovations Universal Agent operator from [OpenShift Container Platform](https://www.openshift.com/) OperatorHub.
            3. Create the eG Universal Agent CRD (Custom Resource Definition)
               a. Click "View YAML Example" and copy the yaml contents and save it to (filename).yaml on your host.
               b. To apply the CRD run the following command

               <br />``` $ oc apply -f (filename).yaml ```

          ### Required Parameters
            * EG\_MANAGER, EG\_MANAGER\_PORT - provide the ip and port of your eG Manager Installation.

          ### Advanced Options
            * Disable Certificate Checking - disable any certificate validation that may interact poorly with proxies within your cluster
            * Image Override - use a copy of the eG Iniversal Agent container image from a registry other that Docker's or Red Hat's
            * Environment variables - define environment variables for the eG Universal Agent container of for APM profiler.
            * Namespace Selectors- select a subset of namesapces to enable instrumentation for applications.

          For complete list of supported parameters and the detailed documentation available in [eG Innovations Docs](https://docs.eginnovations.com)
        displayName: eG Innovations Universal Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Tracing
        links:
        - name: eG Innovations Kubernetes Monitoring Overview
          url: https://www.eginnovations.com/documentation/Kubernetes/Introduction-to-Monitoring-Kubernetes.htm
        maintainers:
        - email: devops@eginnovations.com
          name: eG Innovations
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: eG Innovations
          url: https://www.eginnovations.com
        relatedImages:
        - docker.io/eginnovations/universal-agent-operator@sha256:b8e3e26dca1b94fabfd79496feb36f3f2de4434331ef9ca810125892a936dca7
        version: 0.0.11
      entries:
      - name: eginnovations-operator.v0.0.11
        version: 0.0.11
      name: beta
    defaultChannel: beta
    packageName: eginnovations-operator
    provider:
      name: eG Innovations
      url: https://www.eginnovations.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-trino-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trino-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:904eb8ac055171c75c8c86e4923df85461b81e6d5d9269d9cddfb03fc615f9f4
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:904eb8ac055171c75c8c86e4923df85461b81e6d5d9269d9cddfb03fc615f9f4
        version: 24.11.1
      entries:
      - name: trino-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: trino-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b50b247ed8f0ad262ed59e8aafd099919e98156472e5663957f9922616140bd9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b50b247ed8f0ad262ed59e8aafd099919e98156472e5663957f9922616140bd9
        version: 25.3.0
      entries:
      - name: trino-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: trino-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
        version: 25.7.0
      entries:
      - name: trino-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: trino-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
        version: 25.7.0
      entries:
      - name: trino-operator.v25.7.0
        version: 25.7.0
      - name: trino-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-trino-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-kafka-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kafka-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:f76b5558088bead829be8018a89e3efc1bb7a84d40429a0a1df1c88552db169b
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:f76b5558088bead829be8018a89e3efc1bb7a84d40429a0a1df1c88552db169b
        version: 24.11.1
      entries:
      - name: kafka-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: kafka-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:72b390c2d382af0a9cbe2d9ae394fc61c62b9a3f6c7310cf5f325d0f17257f1f
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:72b390c2d382af0a9cbe2d9ae394fc61c62b9a3f6c7310cf5f325d0f17257f1f
        version: 25.3.0
      entries:
      - name: kafka-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: kafka-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
        version: 25.7.0
      entries:
      - name: kafka-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: kafka-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
        version: 25.7.0
      entries:
      - name: kafka-operator.v25.7.0
        version: 25.7.0
      - name: kafka-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-kafka-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: vfunction-operator
      app.kubernetes.io/managed-by: operator
      app.kubernetes.io/name: vfunction
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      name: vfunction-server-operator
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: vFunction
      provider-url: ""
    name: vfunction-server-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vfunction-server-operator.v2.2.469
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vfunction.com/v1",
                "kind": "VfunctionServer",
                "metadata": {
                  "name": "example-vfunction-server",
                  "namespace": "vfunction",
                  "labels": {
                      "name": "vfunction",
                      "app.kubernetes.io/name": "vfunction",
                      "app.kubernetes.io/instance": "vfunction-server",
                      "app.kubernetes.io/managed-by": "operator"
                  }
                },
                "spec": {
                  "admin": {
                    "email": "admin@mycompany.com",
                    "name": "Admin",
                    "password": "Password1!"
                  },
                  "host": "http://my.domain.com",
                  "org_name": "MyCompany",
                  "upgrade": "Daily",
                  "measurements": {
                    "auto_scaling": "No",
                    "min_number_of_services": "2",
                    "max_number_of_services": "10"
                  },
                  "smtp": {
                    "identity": "",
                    "password": "",
                    "url": "",
                    "user": ""
                  },
                  "tls": {
                    "crt": "",
                    "key": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, AI/Machine Learning
          certified: Red Hat certified
          containerImage: registry.marketplace.redhat.com/rhm/vfunction/vfunction-server-operator@sha256:60d1965460101d0fd90ecc9499d0a6d4c9684a54ada61a43a253d30661fab22f
          createdAt: "2021-08-03T19:08:54Z"
          description: vFunction is a cutting-edge code analysis, machine learning,
            and automation to boost your Java modernization projects.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/vfunction-server-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/vfunction-server-operator-rhmp/support?utm_source=openshift_console
          support: vFunction
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of vFunction running on your cluster.
            displayName: vFunction
            kind: VfunctionServer
            name: vfunctionservers.vfunction.com
            version: v1
        description: |2-

          ## Introduction

          [vFunction](https://www.vfunction.com/) is a cloud-native modernization platform that combines dynamic and static code analysis, machine learning, and automation to automatically identify and extract services from existing applications. vFunction is the only platform purpose-built for modernization of Java applications.

          The vFunction modernization process starts by learning the running monolithic application, and surfacing the interdependencies within it. Using AI, the platform analyzes and identifies services that can be separated from the application. This decomposition can present a range of micro, mini, or even macro services, depending on your application environment, each being an independently deployable and scalable application component.

          vFunction automates the extraction of these services, enabling you to modernize your monolith, quickly and easily.

          ## The vFunction platform

          The platform consists of 3 basic components; the server, the controller package, and a tools package. The server runs as an operand on an OpenShift environment. The controller package is installed on the machine that runs the monolithic application which can be either a Linux or a Windows machine, and the tools are run on a development machine, with access to the code of the monolithic application.

          The controller package consists of three elements: the vFunction agent, that collects data during the dynamic analysis phase; the vFunction Viper application, that performs static analysis on the binaries of the application; and the vFunction controller that handles all the communication between the agent, Viper, and the vFunction server.

          The vFunction agent is a mix of a Java and native agent, and needs to run on the JVM that is currently running your application. [Refer to the vFunction OOB Support Matrix document](https://drive.google.com/file/d/1ccq8LFab1FrYAimDUxwgjiCOdk4QuVzs/view) for a list of supported application servers and JVMs.

          This operator installs a vFunction server instance to be connected later with one or more vFunction controllers installed on your application machine(s).

          ## Before You Start

          The vFunction operator requires a Red Hat OpenShift Kubernetes Application Platform 4.x.

          The cluster recommended configuration (which is the same as the minimal one) should consist of:

          - *1 master node*

          - *1 worker node*

          - *Storage capacity enough for dynamically provision 2 PVs of 50G each*

          - *A default StorageClass configured*

          vFunction doesn’t provide any encryption mechanism for data stored on any attached storage. If required, encryption can be achieved by the client by using its own external means on the storage itself.

          vFunction supports backup of all critical data as part of the application. Nevertheless, for complete backup of entire data it is recommended that the client apply its own storage based backup mechanism.

          ## Install Prerequisites

          ###Cluster:
          The operator was certified and tested on OCP 4.6 and 4.7.
          If you encounter any issue with other OCP versions or cloud providers please contact info@vfunction.com.

          ###Storage:
          The operator creates two new PersistentVolumeClaims (PVCs) during installation time. For their creation, it relies on the default StorageClass to dynamically provision two PersistentVolumes (PVs). The system must have a default StorageClass in place.

          Both PersistentVolumes are accessed with ReadWriteOnce mode.

          ###Project:
          The operator should be installed in a new and dedicated project (namespace). If you intend to install more than one vFunction operands in the same cluster, please use different projects for each one. Installation of more than one operand on the same cluster will allow scaling by load balancing multiple applications to different vFunction servers.

          ## Installation instructions

          1. Prepare the YAML file to use in the installation:

              a. Copy the following YAML template into a text editor:

              ```yaml
                  apiVersion: vfunction.com/v1
                  kind: VfunctionServer
                  metadata:
                    name: vfunction
                    namespace: vfunction
                    labels:
                      name: vfunction
                      app.kubernetes.io/name: vfunction
                      app.kubernetes.io/instance: vfunction-server
                      app.kubernetes.io/managed-by: operator
                  spec:
                    host: "http://my.domain.com"
                    org_name: "MyCompany"
                    upgrade: "Daily"
                    admin:
                      email: "admin@mycompany.com"
                      name: "Admin"
                      password: "Password1!"
                    smtp:
                      password: ""
                      url: ""
                      identity: ""
                      user: ""
                    measurements:
                      auto_scaling: "No"
                      min_number_of_services: "2"
                      max_number_of_services: "10"
                    tls:
                      crt: |
                        -----BEGIN CERTIFICATE-----
                        ...
                        -----END CERTIFICATE-----
                      key: |
                        -----BEGIN PRIVATE KEY-----
                        ...
                        -----END PRIVATE KEY-----
              ```
              b. Customize the template fields by replacing with your information:

              - **host**: Enter the FQDN for accessing the vFunction dashboard.
                          Make sure that you use a domain name and not an IP address. Writing "https://" will mandate TLS while writing "http://" will expose the server through HTTP.
              - **org_name**: Enter your organization name.
              - **upgrade**: Choose auto-upgrade mode
                  - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically (default).
                  - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
                  - **Never**: The automatic upgrading mechanism is off.
                  - **Maintenance**: Put the operator in a maintenance mode.

              - **admin.email**: Enter the email address of your administrator.
              - **admin.name**: Enter the name of your administrator.
              - **admin.password**: Enter the password you want to use for the vFunction administrator.
                                    The password should be at least 8 characters long, and consist of at least one lowercase letter, at least one uppercase letter, at least one number, and at least one special character.
              - **smtp.user**: Enter the email address for a designated user for the SMTP server (optional).
              - **smtp.password**: Enter this user’s password (optional).
              - **smtp.identity**: Enter this user’s password (optional).
              - **smtp.url**: Enter the SMTP server URL (optional).
              - **tls.crt**: If your host FQDN starts with “https”, paste in the certifications you have for using the TLS connection.
              - **tls.key**: If your host FQDN starts with “https”, enter the key you have for using the TLS connection.
              - **measurements.auto_scaling**: Indicates if the measurement services auto scalling is active.
              - **measurements.min_num_of_services**: Minimum Number of measurement service pods.
              - **measurements.max_num_of_services**: Maximum Number of measurement service pods.
              - **measurements.S3**: Save all measurements data to S3, instead of local PV (optional).
                  - **measurements.S3.bucket**: Measurements S3 bucket name.
                  - **measurements.S3.key**: Measurements S3 key.
                  - **measurements.S3.secret**: Measurements S3 secret.
                  - **measurements.S3.region**: Measurements S3 region.
              - **backup.mysql.S3**: Backup MySQL to S3 (optional).
                  - **backup.mysql.S3.bucket**: MySQL backup S3 bucket name.
                  - **backup.mysql.S3.key**: MySQL backup S3 key.
                  - **backup.mysql.S3.secret**: MySQL backup S3 secret.
                  - **backup.mysql.S3.region**: MySQL backup S3 region.

          1. In the Installed Operators window choose and click on vFunction Operator and then Create Instance.
          1. In the Create vFunctionServer window, paste the YAML file that you created in step 1.
          1. Click Create.
          1. You can now click on your new created vFunction operand in order to see its details and installation progress.

          ## Verifying the installation

          Check that the operand installed successfully:

          1. Check that the "Operand State" property shows “Working”.
          1. Check that the "Successfully Installed" property shows "Yes".

          If "Operand State" shows "Failed", it indicates the operand wasn’t installed correctly. Check the events and logs for all pods (operator and image containers) for any issues. In the event that you cannot troubleshoot, contact vFunction support.

          ## After installation

          The vFunction site is now accessible via the newly created vFunction application custom address, for example, my.domain.com. There are two ways you can access the vFunction dashboard.

          ### Access using the router canonical hostname

          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address, and to the ”vfunction” subdomain of the  OpenShift canonical router hostname as the alias.
          For example, ***my.domain.com.   CNAME   vfunction.apps.ocp4.my-openshift.com.***
          2. Find your cluster Router Canonical Hostname address in the newly created vfunction-route-xxx under your vFunction OpenShift project -> **Networking** -> **Routes** -> **vfunction-route-xxx** route -> **Router Canonical Hostname** field.
          You can now access the dashboard using your defined "host" spec property (as above).

          ### Access using the nginx service location

          You may use this access method if your OpenShift is installed on a provider that supports exposing LoadBalancer-type services.
          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address and to the vfunction-nginx-xxx service location.
          For example, ***my.domain.com.   CNAME   a05951ed7cdf-1394239323.us-east-1.elb.amazonaws.com.***
          1. Point your custom domain to the vfunction-nginx-xxx service's external IP location, which you can find in the OpenShift project > **Networking** -> **Services** -> **vfunction-nginx-xxx** service > **Service Address** > **Location** field.

          ## Upgrade and Rollback

          The vFunction operator includes a built-in auto-upgrade mechanism.

          You can choose one of 4 upgrading modes:
          - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically (default).
          - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
          - **Never**: The automatic upgrading mechanism is off.
          - **Maintenance**: Put the operator in a maintenance mode.

          For an on-demand upgrade,  you can change anytime the upgrading mode from “Never” to “Always”, wait for the operand to be upgraded, and change again to “Never”.

          During the upgrade, the "Operand State" property will change to “Upgrading...” and back again to “Working” after a successful upgrade.

          In case of a failure upgrading the operand, an automatic rollback to the last working version will occur. The failed version will be marked as defective and the operator will not try to upgrade to it again.

          Choosing the "Maintenance" option will tell the operator to take down all pods, so you can safely fix storage issues, return volumes from snapshots, etc.
          After finishing the maintenance, return the upgrade mode to your original desired policy.
        displayName: vFunction Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - vfunction
        - analyzing
        - services
        links:
        - name: vFunction
          url: https://vfunction.com
        - name: Terms
          url: https://www.vfunction.com/eula/
        maintainers:
        - email: raviv@vfunction.com
          name: Raviv Halivni
        maturity: alpha
        provider:
          name: vFunction
        relatedImages:
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-organizations@sha256:bcc2c3d6222af95e614318daea3108fffc6bc77aa17b23031cce296a3811b79f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-users@sha256:e47cc051953976a0e075e8cfb457cd4749e9b9e4394f2fbecab2458c2ade92ca
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-mysql@sha256:a6e06ea08fffa33b709d4d1240587bb3ab5b4cc6a28f493a227555e88e645798
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-nginx@sha256:435d1489aaf29a811081ddb51f40ae692e5833f42c31dbced6bc189c0ddc40ab
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-storage@sha256:b778cb9c724a59e51b4cb9088047be471cdf338f1138c31b81f79e50a491203f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-idp@sha256:21ca95a0209f3c557a45658452d154777f928539a93129148470562c79e68d70
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-measurements@sha256:35ffbee4b0eead58706b23cf5a702b1d55d2a7913cd0019cefbff94957d4e5d6
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-server-operator@sha256:60d1965460101d0fd90ecc9499d0a6d4c9684a54ada61a43a253d30661fab22f
        version: 2.2.469
      entries:
      - name: vfunction-server-operator.v2.2.469
        version: 2.2.469
      name: stable
    defaultChannel: stable
    packageName: vfunction-server-operator-rhmp
    provider:
      name: vFunction
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-xtradb-cluster-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.17.0
        version: 1.17.0
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.16.1
        version: 1.16.1
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.16.0
        version: 1.16.0
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.15.1
        version: 1.15.1
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.15.0
        version: 1.15.0
      name: stable
    - currentCSV: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        version: 1.18.0-cw
      entries:
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0-cw
        version: 1.18.0-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.17.0-cw
        version: 1.17.0-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.16.1-cw
        version: 1.16.1-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.16.0-cw
        version: 1.16.0-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.15.0-cw
        version: 1.15.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator-certified-rhmp
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: community-windows-machine-config-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: community-windows-machine-config-operator.v6.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-windows/community-windows-machine-config-operator:community-4.11-6c3601f
          createdAt: "2022-08-31T18:57:22Z"
          description: An operator that enables Windows container workloads on OCP
          olm.skipRange: '>=5.0.0 <6.0.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ## WARNING
          Community distribution of the Windows Machine Config Operator.
          This is a preview build, and is not meant for production workloads.
          Issues with this distribution of WMCO can be opened against the [WMCO repository](https://github.com/openshift/windows-machine-config-operator).
          Please read through the [troubleshooting doc](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/docs/TROUBLESHOOTING.md)
          before opening an issue.
          Please ensure that when installing this operator the starting CSV you subscribe to is supported on the
          version of OKD/OCP you are using. This CSV is meant for OKD/OCP community-4.11.
          ## Documentation
          ### Introduction
          The Windows Machine Config Operator configures Windows instances into nodes, enabling Windows container workloads
          to be ran within OKD/OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/latest/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/README.md#configuring-byoh-bring-your-own-host-windows-instances).
          The operator will do all the necessary steps to configure the instance so that it can join the cluster as a worker node.
          ### Pre-requisites
          - [Cluster and OS pre-requisites](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/docs/wmco-prerequisites.md)
          ### Usage
          Please see the usage section of [README.md](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/README.md#usage).
          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.
        displayName: Community Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: preview
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-windows/community-windows-machine-config-operator:community-4.11-6c3601f
        version: 6.0.0
      entries:
      - name: community-windows-machine-config-operator.v6.0.0
        version: 6.0.0
      name: preview
    defaultChannel: preview
    packageName: community-windows-machine-config-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stolostron
      provider-url: ""
    name: cluster-aas-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-aas-operator.v0.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplate",
                "metadata": {
                  "name": "clustertemplate-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplateInstance",
                "metadata": {
                  "name": "clustertemplateinstance-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplateQuota",
                "metadata": {
                  "name": "clustertemplatequota-sample"
                },
                "spec": {}
              }
            ]
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/cluster-templates-operator:2.11.0-0c3131154065397d37e3b66ae635d7ea88650bf9
          createdAt: "2024-06-06T19:34:44Z"
          description: Easily install fully configured clusters with guard-rails.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/cluster-templates-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents instance of a cluster
            displayName: Cluster template instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.clustertemplate.openshift.io
            version: v1alpha1
          - description: Defines which ClusterTemplates can be used in a given namespace
            displayName: Cluster template quota
            kind: ClusterTemplateQuota
            name: clustertemplatequotas.clustertemplate.openshift.io
            version: v1alpha1
          - description: Template of a cluster - both installation and post-install
              setup are defined as ArgoCD application spec. Any application source
              is supported - typically a Helm chart
            displayName: Cluster template
            kind: ClusterTemplate
            name: clustertemplates.clustertemplate.openshift.io
            version: v1alpha1
          - kind: ClusterTemplateSetup
            name: clustertemplatesetup.clustertemplate.openshift.io
            version: v1alpha1
          - kind: Config
            name: config.clustertemplate.openshift.io
            version: v1alpha1
        description: |
          **Self-service clusters with guardrails!**
          Cluster as a service operator provides an easy way to define clusters as templates and allows creating instances of those templates even for non-privileged developer/devops engineers. Cluster templates operator also allows specifing quotas for the developer/devops engineers.

          ### User Interface
          The User interface is installed but is not enabled by default. To enable go to `Home` -> `Overview` -> in the status card click `Dynamic Plugins` -> click `View all` -> click the `pencil` icon -> pick `Enable` and hit `Save`. You can access the UI in the `All Clusters` perspective under `Infrastructure` -> `Cluster Templates`.

          ### Documentation
          Documentation can be found on our [website](https://github.com/stolostron/cluster-templates-operator).

          ### Getting help
          To report any issues or ask questions, create an issue on our [Github repo](https://github.com/stolostron/cluster-templates-operator/issues)
        displayName: Cluster as a service operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cluster
        - templates
        - hive
        - hypershift
        - selfservice
        - openshift
        - multicluster
        - quota
        - helm
        - argocd
        links:
        - name: Cluster as a service operator
          url: https://github.com/stolostron/cluster-templates-operator
        maintainers:
        - email: rawagner@redhat.com
          name: Rastislav Wagner
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Stolostron
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/stolostron/cluster-templates-operator:2.11.0-0c3131154065397d37e3b66ae635d7ea88650bf9
        version: 0.1.5
      entries:
      - name: cluster-aas-operator.v0.1.5
        version: 0.1.5
      - name: cluster-aas-operator.v0.1.4
        version: 0.1.4
      - name: cluster-aas-operator.v0.1.3
        version: 0.1.3
      - name: cluster-aas-operator.v0.1.2
        version: 0.1.2
      - name: cluster-aas-operator.v0.1.1
        version: 0.1.1
      - name: cluster-aas-operator.v0.1.0
        version: 0.1.0
      - name: cluster-aas-operator.v0.0.6
        version: 0.0.6
      - name: cluster-aas-operator.v0.0.5
        version: 0.0.5
      - name: cluster-aas-operator.v0.0.4
        version: 0.0.4
      - name: cluster-aas-operator.v0.0.3
        version: 0.0.3
      - name: cluster-aas-operator.v0.0.2
        version: 0.0.2
      - name: cluster-aas-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cluster-aas-operator
    provider:
      name: Stolostron
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: ibm-spectrum-scale-csi-operator
      app.kubernetes.io/managed-by: ibm-spectrum-scale-csi-operator
      app.kubernetes.io/name: ibm-spectrum-scale-csi-operator
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-spectrum-scale-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-spectrum-scale-csi-operator.v2.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "CSIScaleOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-spectrum-scale-csi-operator",
                    "app.kubernetes.io/managed-by": "ibm-spectrum-scale-csi-operator",
                    "app.kubernetes.io/name": "ibm-spectrum-scale-csi-operator",
                    "release": "ibm-spectrum-scale-csi-operator"
                  },
                  "name": "ibm-spectrum-scale-csi",
                  "namespace": "ibm-spectrum-scale-csi-driver"
                },
                "spec": {
                  "attacherNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "clusters": [
                    {
                      "id": "\u003cPrimary Cluster ID\u003e",
                      "primary": {
                        "primaryFs": "\u003c Primary Filesystem \u003e"
                      },
                      "restApi": [
                        {
                          "guiHost": "\u003c Primary cluster GUI IP/Hostname \u003e"
                        }
                      ],
                      "secrets": "secret1",
                      "secureSslMode": false
                    }
                  ],
                  "pluginNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "provisionerNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "resizerNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "snapshotterNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ]
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/ibm-spectrum-scale/ibm-spectrum-scale-csi-operator@sha256:b62ad588fbc338090f405740fd8bceaf59b3267cdb532c60ae965643ccae4d5d
          createdAt: Thu May 5 04:42:15 IST 2022
          description: An operator for deploying and managing the IBM Spectrum Scale
            CSI Driver.
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-spectrum-scale-csi/
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIScaleOperator is the Schema for the csiscaleoperators
              API
            displayName: IBM CSI Spectrum Scale Driver
            kind: CSIScaleOperator
            name: csiscaleoperators.csi.ibm.com
            version: v1
        description: |
          The IBM Spectrum Scale CSI Operator for Kubernetes installs, manages,
          upgrades the IBM Spectrum Scale CSI Driver on OpenShift and Kubernetes
          for provisioning of persistent volumes on IBM Spectrum Scale.

          ### Documentation Documentation for IBM Spectrum Scale CSI Driver and Operator can be found at [IBM Documentation](https://www.ibm.com/docs/en/spectrum-scale-csi).

          ### Getting help
          IBM Spectrum Scale CSI driver is part of the IBM Spectrum Scale offering.
          Please follow the [IBM support procedure](https://www.ibm.com/mysupport/s/?language=en_US) for any issues with the driver.

          For non-urgent issues, suggestions, recommendations feel free to open an
          issue in [github](https://github.com/IBM/ibm-spectrum-scale-csi). Issues will be addressed as team availability permits.

          **Note:** Custom Resource example represents the minimum settings needed to run
          the operator. If your environment needs more advanced settings (e.g. remote clusters,
          node mapping, image pull secret, Taint and Toleration etc.) please refer to  the sample [Custom Resource](https://raw.githubusercontent.com/IBM/ibm-spectrum-scale-csi/v2.6.0/operator/config/samples/csiscaleoperators.csi.ibm.com_cr.yaml).
        displayName: IBM Spectrum Scale CSI Plugin Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - GPFS
        - Spectrum Scale
        - CSI Driver
        links:
        - name: IBM CSI Spectrum Scale Documentation
          url: https://www.ibm.com/docs/en/spectrum-scale-csi
        - name: CSI Developer Documentation
          url: https://kubernetes-csi.github.io/docs/
        maintainers:
        - email: madhu.punjabi@in.ibm.com
          name: Madhu Thorat
        - email: amdabhad@in.ibm.com
          name: Amol Dabhade
        - email: Nitish.Kumar4@ibm.com
          name: Nitish Kumar
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibm-spectrum-scale/ibm-spectrum-scale-csi-operator@sha256:b62ad588fbc338090f405740fd8bceaf59b3267cdb532c60ae965643ccae4d5d
        version: 2.6.0
      entries:
      - name: ibm-spectrum-scale-csi-operator.v2.6.0
        version: 2.6.0
      - name: ibm-spectrum-scale-csi-operator.v2.5.1
        version: 2.5.1
      - name: ibm-spectrum-scale-csi-operator.v2.5.0
        version: 2.5.0
      - name: ibm-spectrum-scale-csi-operator.v2.4.0
        version: 2.4.0
      - name: ibm-spectrum-scale-csi-operator.v2.3.1
        version: 2.3.1
      - name: ibm-spectrum-scale-csi-operator.v2.3.0
        version: 2.3.0
      name: stable
    defaultChannel: stable
    packageName: ibm-spectrum-scale-csi-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Managed Openshift Black Belt
      provider-url: ""
    name: ecr-secret-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ecr-secret-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "ecr-secret-sample",
                  "namespace": "test-ecr-secret-operator"
                },
                "spec": {
                  "ecr_registry": "ACCOUNT_ID.dkr.ecr.us-east-2.amazonaws.com",
                  "frequency": "10h",
                  "generated_secret_name": "ecr-docker-secret",
                  "region": "us-east-2"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ecr-secret-operator:v0.3.2
          operatorframework.io/suggested-namespace: ecr-secret-operator
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ecr-secret-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret is the Schema for the secrets API
            displayName: Secret
            kind: Secret
            name: secrets.ecr.mobb.redhat.com
            version: v1alpha1
        description: "# ECR Secret Operator\n\nAmazon Elastic Container Registry [Private
          Registry Authentication](https://docs.aws.amazon.com/AmazonECR/latest/userguide/registry_auth.html)
          provides a temporary token that is valid only for 12 hours. It is a challenge
          for automatic container image build process to refresh the token or secret
          in a timely manner.\n\nThis operators frequently talks with AWS ECR GetAuthroization
          Token and create/update the secret, so that the service account can perform
          docker image build.\n\n\n## How to use this operator\n\n### Prerequisites\n\n*
          [Create an ECR private repository](https://docs.aws.amazon.com/AmazonECR/latest/userguide/repository-create.html)\n*
          Create An Openshift Cluster\n* Provide AWS Authentication to the operator.
          Two Options:\n  **Note: You have to provide a secret aws-ecr-cloud-credentials
          in the installed namespace**\n  * [IAM User](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_user.md)\n
          \ * [STS Assume Role](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_assume_role.md)\n\n###
          Install the operator\n\n### Create the ECR Secret CRD\n\n```\napiVersion:
          ecr.mobb.redhat.com/v1alpha1\nkind: Secret\nmetadata:\n  name: ecr-secret\n
          \ namespace: test-ecr-secret-operator\nspec:\n  generated_secret_name: ecr-docker-secret\n
          \ ecr_registry: [ACCOUNT_ID].dkr.ecr.us-east-2.amazonaws.com\n  frequency:
          10h\n  region: us-east-2\n```\n\n```\noc create -f CRD_FILE\n```\n\nA docker
          registry secret is created by the operator momentally and the token is patched
          every 10 hours\n\n```\noc get secret ecr-docker-secret   \nNAME                TYPE
          \                            DATA   AGE\necr-docker-secret   kubernetes.io/dockerconfigjson
          \  1      16h\n```\n\n### A sample build process with generated secret\n\n\nLink
          the secret to builder\n\n```\noc secrets link builder ecr-docker-secret
          \n```\n\nConfigure [build config](https://github.com/rh-mobb/ecr-secret-operator/tree/main/samples/build-config.yaml)
          to point to your ECR Container repository\n\n```\noc create imagestream
          ruby\noc tag openshift/ruby:2.5-ubi8 ruby:2.5\noc create -f samples/build-config.yaml\noc
          start-build ruby-sample-build --wait\n```\n\nBuild should succeed and push
          the image to the the private ECR Container repository\n\n"
        displayName: ECR Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - redhat
        links:
        - name: Ecr Secret Operator
          url: https://github.com/rh-mobb/ecr-secret-operator
        maintainers:
        - email: shading@redhat.com
          name: Shaozhen Ding
        maturity: alpha
        provider:
          name: Managed Openshift Black Belt
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/mobb/ecr-secret-operator:v0.3.2
        version: 0.3.2
      entries:
      - name: ecr-secret-operator.v0.3.2
        version: 0.3.2
      name: v0.3
    - currentCSV: ecr-secret-operator.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "ecr-secret-sample",
                  "namespace": "test-ecr-secret-operator"
                },
                "spec": {
                  "ecr_registry": "ACCOUNT_ID.dkr.ecr.us-east-2.amazonaws.com",
                  "frequency": "10h",
                  "generated_secret_name": "ecr-docker-secret",
                  "region": "us-east-2"
                }
              },
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "ArgoHelmRepoSecret",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ecr-secret-operator",
                    "app.kubernetes.io/instance": "argohelmreposecret-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "argohelmreposecret",
                    "app.kubernetes.io/part-of": "ecr-secret-operator"
                  },
                  "name": "argohelmreposecret-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ecr-secret-operator:v0.4.1
          createdAt: "2023-04-14T14:52:13Z"
          operatorframework.io/suggested-namespace: ecr-secret-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ecr-secret-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ArgoHelmRepoSecret is the Schema for the argohelmreposecrets
              API
            displayName: Argo Helm Repo Secret
            kind: ArgoHelmRepoSecret
            name: argohelmreposecrets.ecr.mobb.redhat.com
            version: v1alpha1
          - description: Secret is the Schema for the secrets API
            displayName: Secret
            kind: Secret
            name: secrets.ecr.mobb.redhat.com
            version: v1alpha1
        description: "# ECR Secret Operator\n\nAmazon Elastic Container Registry [Private
          Registry Authentication](https://docs.aws.amazon.com/AmazonECR/latest/userguide/registry_auth.html)
          provides a temporary token that is valid only for 12 hours. It is a challenge
          for automatic container image build process to refresh the token or secret
          in a timely manner.\n\nThis operators frequently talks with AWS ECR GetAuthroization
          Token and create/update the secret, so that the service account can perform
          docker image build.\n\n\n## How to use this operator\n\n### Prerequisites\n\n*
          [Create an ECR private repository](https://docs.aws.amazon.com/AmazonECR/latest/userguide/repository-create.html)\n*
          Create An Openshift Cluster\n* Provide AWS Authentication to the operator.
          Two Options:\n  **Note: You have to provide a secret aws-ecr-cloud-credentials
          in the installed namespace**\n  * [IAM User](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_user.md)\n
          \ * [STS Assume Role](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_assume_role.md)\n\n###
          Install the operator\n\n### Create the ECR Secret CRD\n\n```\napiVersion:
          ecr.mobb.redhat.com/v1alpha1\nkind: Secret\nmetadata:\n  name: ecr-secret\n
          \ namespace: test-ecr-secret-operator\nspec:\n  generated_secret_name: ecr-docker-secret\n
          \ ecr_registry: [ACCOUNT_ID].dkr.ecr.us-east-2.amazonaws.com\n  frequency:
          10h\n  region: us-east-2\n```\n\n```\noc create -f CRD_FILE\n```\n\nA docker
          registry secret is created by the operator momentally and the token is patched
          every 10 hours\n\n```\noc get secret ecr-docker-secret   \nNAME                TYPE
          \                            DATA   AGE\necr-docker-secret   kubernetes.io/dockerconfigjson
          \  1      16h\n```\n\n### A sample build process with generated secret\n\n\nLink
          the secret to builder\n\n```\noc secrets link builder ecr-docker-secret
          \n```\n\nConfigure [build config](https://github.com/rh-mobb/ecr-secret-operator/tree/main/samples/build-config.yaml)
          to point to your ECR Container repository\n\n```\noc create imagestream
          ruby\noc tag openshift/ruby:2.5-ubi8 ruby:2.5\noc create -f samples/build-config.yaml\noc
          start-build ruby-sample-build --wait\n```\n\nBuild should succeed and push
          the image to the the private ECR Container repository\n\n"
        displayName: ECR Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - redhat
        links:
        - name: Ecr Secret Operator
          url: https://github.com/rh-mobb/ecr-secret-operator
        maintainers:
        - email: shading@redhat.com
          name: Shaozhen Ding
        maturity: alpha
        provider:
          name: Managed Openshift Black Belt
        relatedImages:
        - quay.io/mobb/ecr-secret-operator:v0.4.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        version: 0.4.1
      entries:
      - name: ecr-secret-operator.v0.4.1
        version: 0.4.1
      - name: ecr-secret-operator.v0.4.0
        version: 0.4.0
      name: v0.4
    defaultChannel: v0.4
    packageName: ecr-secret-operator
    provider:
      name: Managed Openshift Black Belt
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-apigatewayv2-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-apigatewayv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Deployment",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "API",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Authorizer",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Route",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Stage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "VPCLink",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/apigatewayv2-controller:1.1.1
          createdAt: "2025-09-19T21:58:07Z"
          description: AWS API Gateway v2 controller is a service controller for managing
            API Gateway v2 resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIMapping represents the state of an AWS apigatewayv2 APIMapping
              resource.
            displayName: APIMapping
            kind: APIMapping
            name: apimappings.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: API represents the state of an AWS apigatewayv2 API resource.
            displayName: API
            kind: API
            name: apis.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Authorizer represents the state of an AWS apigatewayv2 Authorizer
              resource.
            displayName: Authorizer
            kind: Authorizer
            name: authorizers.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Deployment represents the state of an AWS apigatewayv2 Deployment
              resource.
            displayName: Deployment
            kind: Deployment
            name: deployments.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: DomainName represents the state of an AWS apigatewayv2 DomainName
              resource.
            displayName: DomainName
            kind: DomainName
            name: domainnames.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Integration represents the state of an AWS apigatewayv2 Integration
              resource.
            displayName: Integration
            kind: Integration
            name: integrations.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Route represents the state of an AWS apigatewayv2 Route resource.
            displayName: Route
            kind: Route
            name: routes.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Stage represents the state of an AWS apigatewayv2 Stage resource.
            displayName: Stage
            kind: Stage
            name: stages.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: VPCLink represents the state of an AWS apigatewayv2 VPCLink
              resource.
            displayName: VPCLink
            kind: VPCLink
            name: vpclinks.apigatewayv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon API Gateway v2 resources in AWS from within your Kubernetes cluster. Currently only HTTP APIs are supported. WebSocket API support will be added soon. For HTTP APIs, DomainName and APIMapping resource support will be added soon.

          **About Amazon API Gateway v2**

          Amazon API Gateway is a fully managed service that makes it easy for developers to create, publish, maintain, monitor, and secure APIs at any scale. APIs act as the "front door" for applications to access data, business logic, or functionality from your backend services. Using API Gateway, you can create RESTful APIs and WebSocket APIs that enable real-time two-way communication applications. API Gateway supports containerized and serverless workloads, as well as web applications.

          API Gateway handles all the tasks involved in accepting and processing up to hundreds of thousands of concurrent API calls, including traffic management, CORS support, authorization and access control, throttling, monitoring, and API version management. API Gateway has no minimum fees or startup costs. You pay for the API calls you receive and the amount of data transferred out and, with the API Gateway tiered pricing model, you can reduce your cost as your API usage scales.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon API Gateway v2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apigatewayv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon API Gateway v2 Developer Resources
          url: https://aws.amazon.com/api-gateway/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: api gateway maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/apigatewayv2-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-apigatewayv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-apigatewayv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-apigatewayv2-controller.v1.0.30
        version: 1.0.30
      - name: ack-apigatewayv2-controller.v1.0.29
        version: 1.0.29
      - name: ack-apigatewayv2-controller.v1.0.28
        version: 1.0.28
      - name: ack-apigatewayv2-controller.v1.0.26
        version: 1.0.26
      - name: ack-apigatewayv2-controller.v1.0.25
        version: 1.0.25
      - name: ack-apigatewayv2-controller.v1.0.23
        version: 1.0.23
      - name: ack-apigatewayv2-controller.v1.0.21
        version: 1.0.21
      - name: ack-apigatewayv2-controller.v1.0.17
        version: 1.0.17
      - name: ack-apigatewayv2-controller.v1.0.16
        version: 1.0.16
      - name: ack-apigatewayv2-controller.v1.0.15
        version: 1.0.15
      - name: ack-apigatewayv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-apigatewayv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-apigatewayv2-controller.v1.0.7
        version: 1.0.7
      - name: ack-apigatewayv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-apigatewayv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-apigatewayv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-apigatewayv2-controller.v1.0.3
        version: 1.0.3
      - name: ack-apigatewayv2-controller.v1.0.2
        version: 1.0.2
      - name: ack-apigatewayv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-apigatewayv2-controller.v0.1.6
        version: 0.1.6
      - name: ack-apigatewayv2-controller.v0.1.4
        version: 0.1.4
      - name: ack-apigatewayv2-controller.v0.1.3
        version: 0.1.3
      - name: ack-apigatewayv2-controller.v0.1.2
        version: 0.1.2
      - name: ack-apigatewayv2-controller.v0.1.1
        version: 0.1.1
      - name: ack-apigatewayv2-controller.v0.1.0
        version: 0.1.0
      - name: ack-apigatewayv2-controller.v0.0.22
        version: 0.0.22
      - name: ack-apigatewayv2-controller.v0.0.21
        version: 0.0.21
      - name: ack-apigatewayv2-controller.v0.0.20
        version: 0.0.20
      - name: ack-apigatewayv2-controller.v0.0.19
        version: 0.0.19
      - name: ack-apigatewayv2-controller.v0.0.18
        version: 0.0.18
      - name: ack-apigatewayv2-controller.v0.0.17
        version: 0.0.17
      - name: ack-apigatewayv2-controller.v0.0.16
        version: 0.0.16
      - name: ack-apigatewayv2-controller.v0.0.15
        version: 0.0.15
      - name: ack-apigatewayv2-controller.v0.0.14
        version: 0.0.14
      - name: ack-apigatewayv2-controller.v0.0.13
        version: 0.0.13
      - name: ack-apigatewayv2-controller.v0.0.12
        version: 0.0.12
      - name: ack-apigatewayv2-controller.v0.0.11
        version: 0.0.11
      - name: ack-apigatewayv2-controller.v0.0.10
        version: 0.0.10
      - name: ack-apigatewayv2-controller.v0.0.9
        version: 0.0.9
      name: alpha
    defaultChannel: alpha
    packageName: ack-apigatewayv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Backube
      provider-url: https://github.com/backube
    name: snapscheduler
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: snapscheduler.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "snapscheduler.backube/v1",
                "kind": "SnapshotSchedule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "snapscheduler",
                    "app.kubernetes.io/instance": "daily",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "snapshotschedule",
                    "app.kubernetes.io/part-of": "snapscheduler"
                  },
                  "name": "daily"
                },
                "spec": {
                  "retention": {
                    "maxCount": 7
                  },
                  "schedule": "0 0 * * *"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/backube/snapscheduler:3.5.0
          createdAt: "2025-05-14T14:46:43Z"
          description: The SnapScheduler operator takes snapshots of Kubernetes CSI-based
            persistent volumes according to user-supplied schedules.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/snapscheduler
          support: https://github.com/backube/snapscheduler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SnapshotSchedule defines a schedule for taking automated
              snapshots of PVC(s)
            displayName: Snapshot Schedule
            kind: SnapshotSchedule
            name: snapshotschedules.snapscheduler.backube
            version: v1
        description: |-
          ## About this operator

          The SnapScheduler operator takes snapshots of CSI-based PersistentVolumes
          according to a configurable
          [Cron-like](https://en.wikipedia.org/wiki/Cron#Overview) schedule. The
          schedules include configurable retention policies for snapshots as well as
          selectors to limit the volumes that are snapshotted. An example schedule
          could be:

          > *Snapshot **all volumes** in a namespace **daily at midnight**, retaining
          > the most recent **7** snapshots for each volume.*

          Multiple schedules can be combined to provide more elaborate protection
          schemes. For example, a given volume (or collection of volumes) could be
          protected with:

          - 6 hourly snapshots
          - 7 daily snapshots
          - 4 weekly snapshots
          - 12 monthly snapshots

          ### How it works

          The operator watches for `SnapshotSchedule` CRs in each namespace. When the
          current time matches the schedule's cronspec, the operator creates a
          `VolumeSnapshot` object for each `PersistentVolumeClaim` in the namespace
          (or subset thereof if a label selector is provided). The `VolumeSnapshot`
          objects are named according to the template:
          `<pvcname>-<schedulename>-<timestamp>`. After creating the new snapshots,
          the oldest snapshots are removed if necessary, according to the retention
          policy of the schedule.

          Please see the [full
          documentation](https://backube.github.io/snapscheduler/) for more
          information.

          ## Requirements

          - OpenShift >= 4.7 (Kubernetes >= 1.20)
          - CSI-based storage driver that supports snapshots (i.e. has the
            `CREATE_DELETE_SNAPSHOT` capability)

          ## Examples

          The schedule for snapshotting is controlled by the
          `snapshotschedules.snapscheduler.backube` Custom Resource. This is a
          namespaced resource that applies only to the PersistentVolumeClaims in its
          namespace.

          Below is a simple example that keeps 7 daily (taken at midnight) snapshots
          of all PVCs in a given namespace:

          ```yaml
          ---
          apiVersion: snapscheduler.backube/v1
          kind: SnapshotSchedule
          metadata:
            name: daily
          spec:
            retention:
              maxCount: 7
            schedule: "0 0 * * *"
          ```

          See the [usage
          documentation](https://backube.github.io/snapscheduler/usage.html) for full
          details, including how to:

          - add label selectors to restrict which PVCs this schedule applies to
          - set the VolumeSnapshotClass used by the schedule
          - apply custom labels to the automatically created VolumeSnapshot objects
        displayName: SnapScheduler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - snapshot
        - storage
        - scheduler
        links:
        - name: Documentation
          url: https://backube.github.io/snapscheduler/
        - name: Source
          url: https://github.com/backube/snapscheduler
        - name: Blog
          url: https://backube.github.io/
        maintainers:
        - email: jstrunk@redhat.com
          name: John Strunk
        maturity: stable
        minKubeVersion: 1.20.0-0
        provider:
          name: Backube
          url: https://github.com/backube
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.19.1
        - quay.io/backube/snapscheduler:3.5.0
        version: 3.5.0
      entries:
      - name: snapscheduler.v3.5.0
        version: 3.5.0
      - name: snapscheduler.v3.4.0
        version: 3.4.0
      - name: snapscheduler.v3.3.0
        version: 3.3.0
      - name: snapscheduler.v3.2.0
        version: 3.2.0
      - name: snapscheduler.v3.1.0
        version: 3.1.0
      - name: snapscheduler.v3.0.0
        version: 3.0.0
      - name: snapscheduler.v2.1.0
        version: 2.1.0
      - name: snapscheduler.v2.0.0
        version: 2.0.0
      name: candidate
    - currentCSV: snapscheduler.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "snapscheduler.backube/v1",
                "kind": "SnapshotSchedule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "snapscheduler",
                    "app.kubernetes.io/instance": "daily",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "snapshotschedule",
                    "app.kubernetes.io/part-of": "snapscheduler"
                  },
                  "name": "daily"
                },
                "spec": {
                  "retention": {
                    "maxCount": 7
                  },
                  "schedule": "0 0 * * *"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/backube/snapscheduler:3.5.0
          createdAt: "2025-05-14T14:46:43Z"
          description: The SnapScheduler operator takes snapshots of Kubernetes CSI-based
            persistent volumes according to user-supplied schedules.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/snapscheduler
          support: https://github.com/backube/snapscheduler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SnapshotSchedule defines a schedule for taking automated
              snapshots of PVC(s)
            displayName: Snapshot Schedule
            kind: SnapshotSchedule
            name: snapshotschedules.snapscheduler.backube
            version: v1
        description: |-
          ## About this operator

          The SnapScheduler operator takes snapshots of CSI-based PersistentVolumes
          according to a configurable
          [Cron-like](https://en.wikipedia.org/wiki/Cron#Overview) schedule. The
          schedules include configurable retention policies for snapshots as well as
          selectors to limit the volumes that are snapshotted. An example schedule
          could be:

          > *Snapshot **all volumes** in a namespace **daily at midnight**, retaining
          > the most recent **7** snapshots for each volume.*

          Multiple schedules can be combined to provide more elaborate protection
          schemes. For example, a given volume (or collection of volumes) could be
          protected with:

          - 6 hourly snapshots
          - 7 daily snapshots
          - 4 weekly snapshots
          - 12 monthly snapshots

          ### How it works

          The operator watches for `SnapshotSchedule` CRs in each namespace. When the
          current time matches the schedule's cronspec, the operator creates a
          `VolumeSnapshot` object for each `PersistentVolumeClaim` in the namespace
          (or subset thereof if a label selector is provided). The `VolumeSnapshot`
          objects are named according to the template:
          `<pvcname>-<schedulename>-<timestamp>`. After creating the new snapshots,
          the oldest snapshots are removed if necessary, according to the retention
          policy of the schedule.

          Please see the [full
          documentation](https://backube.github.io/snapscheduler/) for more
          information.

          ## Requirements

          - OpenShift >= 4.7 (Kubernetes >= 1.20)
          - CSI-based storage driver that supports snapshots (i.e. has the
            `CREATE_DELETE_SNAPSHOT` capability)

          ## Examples

          The schedule for snapshotting is controlled by the
          `snapshotschedules.snapscheduler.backube` Custom Resource. This is a
          namespaced resource that applies only to the PersistentVolumeClaims in its
          namespace.

          Below is a simple example that keeps 7 daily (taken at midnight) snapshots
          of all PVCs in a given namespace:

          ```yaml
          ---
          apiVersion: snapscheduler.backube/v1
          kind: SnapshotSchedule
          metadata:
            name: daily
          spec:
            retention:
              maxCount: 7
            schedule: "0 0 * * *"
          ```

          See the [usage
          documentation](https://backube.github.io/snapscheduler/usage.html) for full
          details, including how to:

          - add label selectors to restrict which PVCs this schedule applies to
          - set the VolumeSnapshotClass used by the schedule
          - apply custom labels to the automatically created VolumeSnapshot objects
        displayName: SnapScheduler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - snapshot
        - storage
        - scheduler
        links:
        - name: Documentation
          url: https://backube.github.io/snapscheduler/
        - name: Source
          url: https://github.com/backube/snapscheduler
        - name: Blog
          url: https://backube.github.io/
        maintainers:
        - email: jstrunk@redhat.com
          name: John Strunk
        maturity: stable
        minKubeVersion: 1.20.0-0
        provider:
          name: Backube
          url: https://github.com/backube
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.19.1
        - quay.io/backube/snapscheduler:3.5.0
        version: 3.5.0
      entries:
      - name: snapscheduler.v3.5.0
        version: 3.5.0
      - name: snapscheduler.v3.4.0
        version: 3.4.0
      - name: snapscheduler.v3.3.0
        version: 3.3.0
      - name: snapscheduler.v3.2.0
        version: 3.2.0
      - name: snapscheduler.v3.1.0
        version: 3.1.0
      - name: snapscheduler.v3.0.0
        version: 3.0.0
      - name: snapscheduler.v2.1.0
        version: 2.1.0
      - name: snapscheduler.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: snapscheduler
    provider:
      name: Backube
      url: https://github.com/backube
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CNCF
      provider-url: ""
    name: jaeger
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jaeger-operator.v1.65.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "jaegertracing.io/v1",
                "kind": "Jaeger",
                "metadata": {
                  "name": "my-jaeger"
                },
                "spec": {
                  "strategy": "allInOne"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing
          certified: "false"
          containerImage: quay.io/jaegertracing/jaeger-operator:1.65.0
          createdAt: "2025-01-22T20:40:19Z"
          description: Provides tracing, monitoring and troubleshooting for microservices-based
            distributed systems
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/jaegertracing/jaeger-operator
          support: Jaeger Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Jaeger is the Schema for the jaegers API
            displayName: Jaeger
            kind: Jaeger
            name: jaegers.jaegertracing.io
            version: v1
        description: |-
          Jaeger, inspired by [Dapper](https://research.google.com/pubs/pub36356.html) and [OpenZipkin](http://zipkin.io/), is a distributed tracing system released as open source by Uber Technologies. It is used for monitoring and troubleshooting microservices-based distributed systems.
          ### Core capabilities
          Jaeger is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed context propagation
          * Distributed transaction monitoring
          * Root cause analysis
          * Service dependency analysis
          * Performance / latency optimization
          * OpenTracing compatible data model
          * Multiple storage backends: Badger, Cassandra, Elasticsearch, Memory.
          ### Operator features
          * **Multiple modes** - Supports `allInOne`, `production` and `streaming` [modes of deployment](https://www.jaegertracing.io/docs/latest/operator/#deployment-strategies).
          * **Configuration** - The Operator manages [configuration information](https://www.jaegertracing.io/docs/latest/operator/#configuring-the-custom-resource) when installing Jaeger instances.
          * **Storage** - [Configure storage](https://www.jaegertracing.io/docs/latest/operator/#storage-options) used by Jaeger. By default, `memory` is used. Other options include `badger`, `cassandra` or `elasticsearch`. On OpenShift, the operator can delegate creation of an Elasticsearch cluster to the Elasticsearch Operator if deployed.
          * **Agent** - can be deployed as [sidecar](https://www.jaegertracing.io/docs/latest/operator/#auto-injecting-jaeger-agent-sidecars) (default) and/or [daemonset](https://www.jaegertracing.io/docs/latest/operator/#installing-the-agent-as-daemonset).
          * **UI** - Optionally setup ingress (Kubernetes) or secure route (OpenShift) to provide [access to the Jaeger UI](https://www.jaegertracing.io/docs/latest/operator/#accessing-the-jaeger-console-ui).
          ### Before you start
          1. Ensure that the appropriate storage solution, that will be used by the Jaeger instance, is available and configured.
          2. If intending to deploy an Elasticsearch cluster via the Jaeger custom resource, then the Elasticsearch Operator must first be installed.
          ### Troubleshooting
          * https://www.jaegertracing.io/docs/latest/troubleshooting/
        displayName: Community Jaeger Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tracing
        - monitoring
        - troubleshooting
        links:
        - name: Jaeger Operator Source Code
          url: https://github.com/jaegertracing/jaeger-operator
        maintainers:
        - email: jaeger-tracing@googlegroups.com
          name: Jaeger Google Group
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: CNCF
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/jaegertracing/jaeger-operator:1.65.0
        version: 1.65.0
      entries:
      - name: jaeger-operator.v1.65.0
        version: 1.65.0
      - name: jaeger-operator.v1.62.0
        version: 1.62.0
      - name: jaeger-operator.v1.61.0
        version: 1.61.0
      - name: jaeger-operator.v1.60.0
        version: 1.60.0
      - name: jaeger-operator.v1.59.0
        version: 1.59.0
      - name: jaeger-operator.v1.57.0
        version: 1.57.0
      - name: jaeger-operator.v1.56.0
        version: 1.56.0
      - name: jaeger-operator.v1.55.0
        version: 1.55.0
      - name: jaeger-operator.v1.53.0
        version: 1.53.0
      - name: jaeger-operator.v1.52.0
        version: 1.52.0
      - name: jaeger-operator.v1.51.0
        version: 1.51.0
      - name: jaeger-operator.v1.49.0
        version: 1.49.0
      - name: jaeger-operator.v1.48.1
        version: 1.48.1
      - name: jaeger-operator.v1.47.0
        version: 1.47.0
      - name: jaeger-operator.v1.46.0
        version: 1.46.0
      - name: jaeger-operator.v1.45.0
        version: 1.45.0
      - name: jaeger-operator.v1.44.0
        version: 1.44.0
      - name: jaeger-operator.v1.43.0
        version: 1.43.0
      - name: jaeger-operator.v1.42.0
        version: 1.42.0
      - name: jaeger-operator.v1.41.1
        version: 1.41.1
      - name: jaeger-operator.v1.41.0
        version: 1.41.0
      - name: jaeger-operator.v1.40.0
        version: 1.40.0
      - name: jaeger-operator.v1.39.0
        version: 1.39.0
      - name: jaeger-operator.v1.38.1
        version: 1.38.1
      - name: jaeger-operator.v1.38.0
        version: 1.38.0
      - name: jaeger-operator.v1.37.0
        version: 1.37.0
      - name: jaeger-operator.v1.36.0
        version: 1.36.0
      - name: jaeger-operator.v1.35.0
        version: 1.35.0
      - name: jaeger-operator.v1.34.1
        version: 1.34.1
      - name: jaeger-operator.v1.34.0
        version: 1.34.0
      - name: jaeger-operator.v1.33.0
        version: 1.33.0
      - name: jaeger-operator.v1.32.0
        version: 1.32.0
      - name: jaeger-operator.v1.31.0
        version: 1.31.0
      - name: jaeger-operator.v1.30.0
        version: 1.30.0
      - name: jaeger-operator.v1.29.1
        version: 1.29.1
      - name: jaeger-operator.v1.25.0
        version: 1.25.0
      - name: jaeger-operator.v1.24.0
        version: 1.24.0
      name: stable
    defaultChannel: stable
    packageName: jaeger
    provider:
      name: CNCF
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: JWS team
      provider-url: ""
    name: jws-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jws-operator.v2.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "web.servers.org/v1alpha1",
                "kind": "WebServer",
                "metadata": {
                  "name": "webserver-example"
                },
                "spec": {
                  "applicationName": "webapp",
                  "replicas": 1,
                  "useInsightsClient": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/jboss-webserver-5/jws5-rhel8-operator@sha256:ea042277bdcdb70b2dc32083ea868efd4bf474d5a8ec0f9b47f5bea35aaeae1e
          createdAt: "2025-09-12T08:31:52.558719Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat JBoss Web Server",
            "Red Hat Runtimes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/web-servers/jws-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Web Server is the schema for the webservers API
            displayName: Web Server
            kind: WebServer
            name: webservers.web.servers.org
            version: v1alpha1
        description: WebServer operator can manage JBoss Web Server images.
        displayName: WebServer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WebServer
        - JBoss
        links:
        - name: Jws Operator
          url: https://jws-operator.domain
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: JWS team
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:021ff671b0fb72eed635b05489ec86342fcb7ada3ba0353591fdb9328871d067
        - registry.redhat.io/jboss-webserver-5/jws5-rhel8-operator@sha256:ea042277bdcdb70b2dc32083ea868efd4bf474d5a8ec0f9b47f5bea35aaeae1e
        - registry.redhat.io/insights-runtimes-tech-preview/runtimes-inventory-rhel8-operator@sha256:58e01d11c3d30abc5b0ee3795511bd04e39aea21e2853b4f4cc6d1c8a2bd7bb0
        version: 2.2.0
      entries:
      - name: jws-operator.v2.2.0
        version: 2.2.0
      - name: jws-operator.v2.1.3
        version: 2.1.3
      - name: jws-operator.v2.1.2
        version: 2.1.2
      - name: jws-operator.v2.1.1
        version: 2.1.1
      - name: jws-operator.v2.1.0
        version: 2.1.0
      - name: jws-operator.v2.0.16
        version: 2.0.16
      - name: jws-operator.v2.0.15
        version: 2.0.15
      - name: jws-operator.v2.0.14
        version: 2.0.14
      - name: jws-operator.v2.0.13
        version: 2.0.13
      - name: jws-operator.v2.0.12
        version: 2.0.12
      - name: jws-operator.v2.0.11
        version: 2.0.11
      - name: jws-operator.v2.0.10
        version: 2.0.10
      - name: jws-operator.v2.0.9
        version: 2.0.9
      - name: jws-operator.v2.0.8
        version: 2.0.8
      - name: jws-operator.v2.0.7
        version: 2.0.7
      - name: jws-operator.v2.0.6
        version: 2.0.6
      - name: jws-operator.v2.0.5
        version: 2.0.5
      - name: jws-operator.v2.0.4
        version: 2.0.4
      - name: jws-operator.v2.0.3
        version: 2.0.3
      - name: jws-operator.v2.0.2
        version: 2.0.2
      - name: jws-operator.v2.0.1
        version: 2.0.1
      name: alpha
    defaultChannel: alpha
    packageName: jws-operator
    provider:
      name: JWS team
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Windup
      provider-url: ""
    name: windup-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: windup-operator.v0.0.25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
             {
               "apiVersion": "windup.jboss.org/v1",
               "kind": "Windup",
               "metadata": {
                 "name": "windup",
                 "labels": {
                   "application": "windup"
                 }
               },
               "spec": {}
             }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/windupeng/windup-operator-native:0.0.25
          createdAt: 01/22/24
          description: Windup is an analysis tool that supports the modernization
            and migration of Java applications.
          imagePullPolicy: Always
          operatorframework.io/suggested-namespace: windup
          repository: https://github.com/windup/windup-operator/
          support: https://issues.redhat.com/projects/WINDUP
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "Create a Windup Web Console application.  \n\nTo access
              the Windup Web Console application from the \"Developer\" perspective,
              please go to Topology and click on the \"Open URL\" icon on the web-console
              pod.  \n\nAlternatively from the \"Administrator\" perspective, please
              go to Networking->Routes and click on the Location hyperlink.  \n\nKnown
              issue  \nIf you want to customize the Windup Web Console instance's
              parameters and you can not see any in the `Form view`, please switch
              to the `YAML view` and change them as needed.\n"
            displayName: Windup
            kind: Windup
            name: windups.windup.jboss.org
            version: v1
        description: |
          Windup is a web console application that supports large-scale Java application modernization and migration projects across a broad range of transformations and use cases.

          It analyzes application code, supports effort estimation, accelerates code migration, and enables users to move applications to containers.

          For more information please refer to the [https://windup.github.io/](https://windup.github.io/) page.
        displayName: Windup Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - modernization
        - windup
        links:
        - name: Website
          url: https://windup.github.io
        - name: GitHub
          url: https://github.com/windup/windup-operator
        maintainers:
        - email: migrate@redhat.com
          name: Windup
        maturity: alpha
        provider:
          name: Windup
        relatedImages:
        - quay.io/windupeng/windup-operator-native:0.0.25
        version: 0.0.25
      entries:
      - name: windup-operator.v0.0.25
        version: 0.0.25
      - name: windup-operator.v0.0.24
        version: 0.0.24
      - name: windup-operator.v0.0.23
        version: 0.0.23
      - name: windup-operator.v0.0.22
        version: 0.0.22
      - name: windup-operator.v0.0.21
        version: 0.0.21
      - name: windup-operator.v0.0.20
        version: 0.0.20
      - name: windup-operator.v0.0.19
        version: 0.0.19
      - name: windup-operator.v0.0.18
        version: 0.0.18
      - name: windup-operator.v0.0.17
        version: 0.0.17
      - name: windup-operator.v0.0.16
        version: 0.0.16
      - name: windup-operator.v0.0.15
        version: 0.0.15
      - name: windup-operator.v0.0.14
        version: 0.0.14
      - name: windup-operator.v0.0.13
        version: 0.0.13
      - name: windup-operator.v0.0.12
        version: 0.0.12
      - name: windup-operator.v0.0.11
        version: 0.0.11
      name: alpha
    defaultChannel: alpha
    packageName: windup-operator
    provider:
      name: Windup
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: watcher-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: watcher-operator.v1.0.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "Watcher",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcher",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcher",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcher"
                },
                "spec": {
                  "apiServiceTemplate": {
                    "tls": {
                      "caBundleSecretName": "combined-ca-bundle"
                    }
                  },
                  "databaseInstance": "openstack"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherAPI",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcher-operator"
                  },
                  "name": "watcherapi-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherApplier",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcherapplier-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcherapplier",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcherapplier-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherDecisionEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcherdecisionengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcherdecisionengine",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcherdecisionengine-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-09-02T13:32:31Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.14
          operatorframework.io/suggested-namespace: openstack-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: WatcherAPI is the Schema for the watcherapis API
            displayName: Watcher API
            kind: WatcherAPI
            name: watcherapis.watcher.openstack.org
            version: v1beta1
          - description: WatcherApplier is the Schema for the watcherappliers API
            displayName: Watcher Applier
            kind: WatcherApplier
            name: watcherappliers.watcher.openstack.org
            version: v1beta1
          - description: WatcherDecisionEngine is the Schema for the watcherdecisionengines
              API
            displayName: Watcher Decision Engine
            kind: WatcherDecisionEngine
            name: watcherdecisionengines.watcher.openstack.org
            version: v1beta1
          - description: Watcher is the Schema for the watchers API
            displayName: Watcher
            kind: Watcher
            name: watchers.watcher.openstack.org
            version: v1beta1
        description: The Watcher Operator project
        displayName: Watcher Operator (Technology Preview)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        links:
        - name: Watcher Operator
          url: https://github.com/openstack-k8s-operators/watcher-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/watcher-rhel9-operator@sha256:63c341a3e20a522c601098502a8fd8f1164d23c773ae8f55782c43ff4b6c8c18
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:eed8234b0fe47083ab941f513709f778bfcb80e32747b71a65fd5f62328a5ca0
        - registry.redhat.io/rhoso/openstack-watcher-api-rhel9@sha256:e97369e3a47bb882f1920c6af3f6345a842af45c58f0f5c5df9ac2933cf3098a
        - registry.redhat.io/rhoso/openstack-watcher-decision-engine-rhel9@sha256:c759d66cf0986c0036be91c008cbd059b4412c94160ed82dd12cd4e00d0685a9
        - registry.redhat.io/rhoso/openstack-watcher-applier-rhel9@sha256:c92a82f901613991926367bb8aab4f875c27a3e4a5282d92847a8463d98cd9e3
        version: 1.0.14
      entries:
      - name: watcher-operator.v1.0.14
        version: 1.0.14
      - name: watcher-operator.v1.0.13
        version: 1.0.13
      - name: watcher-operator.v1.0.12
        version: 1.0.12
      - name: watcher-operator.v1.0.11
        version: 1.0.11
      - name: watcher-operator.v1.0.10
        version: 1.0.10
      - name: watcher-operator.v1.0.9
        version: 1.0.9
      - name: watcher-operator.v1.0.8
        version: 1.0.8
      - name: watcher-operator.v1.0.7
        version: 1.0.7
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: watcher-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: VMware Tanzu
      provider-url: ""
    name: rabbitmq-cluster-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-cluster-operator.v2.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "rabbitmq.com/v1beta1",
                "kind": "RabbitmqCluster",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "override": {
                    "statefulset": {
                      "spec": {
                        "template": {
                          "spec": {
                            "containers": [],
                            "securityContext": {}
                          }
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/rabbitmqoperator/cluster-operator:2.16.0
          createdAt: 2025-08-12T09:34:33
          repository: https://github.com/rabbitmq/cluster-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Rabbitmq Cluster
            displayName: RabbitmqCluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: "## About the managed application\nRabbitMQ is an open-source
          message broker that originally implemented the\nAdvanced Message Queuing
          Protocol (AMQP). Nowadays features a plug-in\narchitecture supporting protocols
          such as Streaming Text Oriented Messaging\nProtocol (STOMP), MQ Telemetry
          Transport (MQTT), and others. RabbitMQ\nsupports native streaming and AMQP
          1.0.\n\nVisit [rabbitmq.com](rabbitmq.com) for more information.\n\n## About
          this Operator\nRabbitMQ Cluster Kubernetes Operator is developed and maintened
          by the VMware \nRabbitMQ team and provides a consistent and easy way to
          deploy RabbitMQ\nclusters to Kubernetes and run them, including \"day two\"
          (continuous)\noperations. RabbitMQ clusters deployed using the Operator
          can be used by\napplications running on Kubernetes or outside of Kubernetes.\n\nRead
          [RabbitMQ operator docs](https://www.rabbitmq.com/kubernetes/operator/operator-overview)\nfor
          more information.\n"
        displayName: RabbitMQ-cluster-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        - messaging
        - amqp
        - amqp10
        - stomp
        - mqtt
        - mqtt5
        - queues
        - streaming
        links:
        - name: Documentation
          url: https://www.rabbitmq.com/kubernetes/operator/operator-overview.html
        - name: GitHub
          url: https://github.com/rabbitmq/cluster-operator
        maintainers:
        - email: rabbitmq-users@googlegroups.com
          name: VMware Tanzu
        maturity: stable
        minKubeVersion: 1.26.0
        nativeApis:
        - group: apps
          kind: Deployment
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: ReplicaSet
          version: v1
        - group: ""
          kind: Pod
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: ConfigMap
          version: v1
        provider:
          name: VMware Tanzu
        relatedImages:
        - quay.io/rabbitmqoperator/cluster-operator:2.16.0
        version: 2.16.0
      entries:
      - name: rabbitmq-cluster-operator.v2.16.0
        version: 2.16.0
      - name: rabbitmq-cluster-operator.v2.15.0
        version: 2.15.0
      - name: rabbitmq-cluster-operator.v2.14.0
        version: 2.14.0
      - name: rabbitmq-cluster-operator.v2.13.0
        version: 2.13.0
      - name: rabbitmq-cluster-operator.v2.12.1
        version: 2.12.1
      - name: rabbitmq-cluster-operator.v2.10.0
        version: 2.10.0
      - name: rabbitmq-cluster-operator.v2.9.0
        version: 2.9.0
      - name: rabbitmq-cluster-operator.v2.8.0
        version: 2.8.0
      - name: rabbitmq-cluster-operator.v2.7.0
        version: 2.7.0
      - name: rabbitmq-cluster-operator.v2.6.0
        version: 2.6.0
      - name: rabbitmq-cluster-operator.v2.5.0
        version: 2.5.0
      - name: rabbitmq-cluster-operator.v2.4.0
        version: 2.4.0
      - name: rabbitmq-cluster-operator.v2.3.0
        version: 2.3.0
      - name: rabbitmq-cluster-operator.v2.2.0
        version: 2.2.0
      - name: rabbitmq-cluster-operator.v2.1.0
        version: 2.1.0
      - name: rabbitmq-cluster-operator.v2.0.0
        version: 2.0.0
      - name: rabbitmq-cluster-operator.v1.14.0
        version: 1.14.0
      name: stable
    defaultChannel: stable
    packageName: rabbitmq-cluster-operator
    provider:
      name: VMware Tanzu
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: ""
    name: ibm-metrics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-metrics-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MarketplaceConfig",
                "metadata": {
                  "name": "marketplaceconfig"
                },
                "spec": {
                  "features": {
                    "deployment": true,
                    "meterDefinitionCatalogServer": false,
                    "registration": true
                  },
                  "installIBMCatalogSource": true,
                  "isDisconnected": false,
                  "license": {
                    "accept": false
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterBase",
                "metadata": {
                  "name": "meterbase"
                },
                "spec": {
                  "enabled": true,
                  "prometheus": {
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "1",
                        "memory": "1G"
                      }
                    },
                    "storage": {
                      "size": "20Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterReport",
                "metadata": {
                  "name": "example-meterreport"
                },
                "spec": {
                  "endTime": "2020-08-05T00:00:00Z",
                  "prometheusService": {
                    "name": "rhm-prometheus-meterbase",
                    "namespace": "openshift-redhat-marketplace",
                    "targetPort": "rbac"
                  },
                  "startTime": "2020-08-04T00:00:00Z"
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "example-meterdefinition-4"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "meterdefinition-sample"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
          createdAt: "2025-08-28T14:18:55Z"
          description: The IBM Metrics Operator provides workload metering and reporting
            for IBM and Red Hat Marketplace customers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
          - description: MeterBase is the resource that sets up Metering for Red Hat
              Marketplace. This is an internal resource not meant to be modified directly.
            displayName: Meter Base
            kind: MeterBase
            name: meterbases.marketplace.redhat.com
            version: v1alpha1
          - description: MeterDefinition defines the meter workloads used to enable
              pay for use billing.
            displayName: Meter Definition
            kind: MeterDefinition
            name: meterdefinitions.marketplace.redhat.com
            version: v1beta1
          - kind: MeterReport
            name: meterreports.marketplace.redhat.com
            version: v1alpha1
        description: "The IBM Metrics Operator provides workload metering and reporting
          for IBM Software Central and Red Hat Marketplace customers.\n### **Important
          Note**\nA set of instructions for onboarding is provided here. For more
          detailed onboarding instructions or information about what is installed
          please visit [swc.saas.ibm.com](https://swc.saas.ibm.com).\n\nUsage metrics
          may be monitored through [https://swc.saas.ibm.com](https://swc.saas.ibm.com)
          with only IBM Metrics Operator and a Red Hat Marketplace account.\n\n###
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n### Prerequisites\n1. Installations are required
          to [enable monitoring for user-defined projects](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          as the Prometheus provider.\n2. Edit the cluster-monitoring-config ConfigMap
          object:\n\n   ```sh\n   $ oc -n openshift-monitoring edit configmap cluster-monitoring-config\n
          \   ```\n\n3. Add enableUserWorkload: true under data/config.yaml:\n  \n
          \   ```sh\n    apiVersion: v1\n    kind: ConfigMap\n    metadata:\n      name:
          cluster-monitoring-config\n        namespace: openshift-monitoring\n    data:\n
          \     config.yaml: |\n        enableUserWorkload: true\n    ```\n\n4. Configure
          the user-workload-monitoring-config ConfigMap object:\n\n    ```sh\n    $
          oc -n openshift-user-workload-monitoring edit configmap user-workload-monitoring-config\n
          \   ```\n\n5. Configure a minimum retention time of 168h and minimum storage
          capacity of 40Gi\n  \n    ```sh\n    apiVersion: v1\n    kind: ConfigMap\n
          \   metadata:\n      name: user-workload-monitoring-config\n      namespace:
          openshift-user-workload-monitoring\n\n    data:\n      config.yaml: |\n
          \       prometheus:\n          retention: 168h\n          volumeClaimTemplate:\n
          \           spec:\n              resources:\n                requests:\n
          \                 storage: 40Gi\n    ```\n\n### Resources Required\n\nMinimum
          system resources required:\n\n| Operator                | Memory (MB) |
          CPU (cores) | Disk (GB) | Nodes |\n| ----------------------- | -----------
          | ----------- | --------- | ----- |\n| **Metrics**   |        750  |     0.25
          \   | 3x1       |    3  |\n| **Deployment** |        250  |     0.25    |
          -         |    1  |\n\n| Prometheus Provider  | Memory (GB) | CPU (cores)
          | Disk (GB) | Nodes |\n| --------- | ----------- | ----------- | ---------
          | ----- |\n| **[Openshift User Workload Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)**
          |          1  |     0.1       | 2x40        |   2    |\n\nMultiple nodes
          are required to provide pod scheduling for high availability for IBM Metrics
          Operator Data Service and Prometheus.\n\nThe IBM Metrics Operator automatically
          creates 3 x 1Gi PersistentVolumeClaims to store reports as part of the data
          service, with _ReadWriteOnce_ access mode. Te PersistentVolumeClaims are
          automatically created by the ibm-metrics-operator after creating a `redhat-marketplace-pull-secret`
          and accepting the license in `marketplaceconfig`.\n\n| NAME                                |
          CAPACITY | ACCESS MODES |\n| ----------------------------------- | --------
          | ------------ |\n| rhm-data-service-rhm-data-service-0 | 1Gi | RWO |\n|
          rhm-data-service-rhm-data-service-1 | 1Gi | RWO |\n| rhm-data-service-rhm-data-service-2
          | 1Gi | RWO |\n\n### Supported Storage Providers\n\n- OpenShift Container
          Storage / OpenShift Data Foundation version 4.x, from version 4.2 or higher\n-
          IBM Cloud Block storage and IBM Cloud File storage\n- IBM Storage Suite
          for IBM Cloud Paks:\n  - File storage from IBM Spectrum Fusion/Scale \n
          \ - Block storage from IBM Spectrum Virtualize, FlashSystem or DS8K\n- Portworx
          Storage, version 2.5.5 or above\n- Amazon Elastic File Storage\n\n### Access
          Modes required\n\n - ReadWriteOnce (RWO)\n\n### Provisioning Options supported\n\nChoose
          one of the following options to provision storage for the ibm-metrics-operator
          data-service\n\n#### Dynamic provisioning using a default StorageClass\n
          \  - A StorageClass is defined with a `metadata.annotations: storageclass.kubernetes.io/is-default-class:
          \"true\"`\n   - PersistentVolumes will be provisioned automatically for
          the generated PersistentVolumeClaims\n--- \n#### Manually create each PersistentVolumeClaim
          with a specific StorageClass\n   - Must be performed before creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`. Otherwise, the automatically
          generated PersistentVolumeClaims are immutable.\n```\napiVersion: v1\nkind:
          PersistentVolumeClaim\nmetadata:\n  labels:\n    app: rhm-data-service\n
          \ name: rhm-data-service-rhm-data-service-0\n  namespace: ibm-software-central\nspec:\n
          \ storageClassName: rook-cephfs\n  accessModes:\n  - ReadWriteOnce\n  resources:\n
          \   requests:\n      storage: 1Gi\n```\n---\n#### Manually provision each
          PersistentVolume for the generated PersistentVolumeClaims with a specific
          StorageClass\n  - May be performed before or after creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`.  \n```\napiVersion: v1\nkind:
          PersistentVolume\nmetadata:\n  name: rhm-data-service-rhm-data-service-0\nspec:\n
          \ csi:\n    driver: rook-ceph.cephfs.csi.ceph.com\n    volumeHandle: rhm-data-service-rhm-data-service-0\n
          \ capacity:\n    storage: 1Gi\n  accessModes:\n    - ReadWriteOnce\n  persistentVolumeReclaimPolicy:
          Delete\n  storageClassName: rook-cephfs\n  volumeMode: Filesystem\n  claimRef:\n
          \   kind: PersistentVolumeClaim\n    namespace: ibm-software-central\n    name:
          rhm-data-service-rhm-data-service-0\n```\n\n### Installation\n1. Create
          or get your [pull secret](https://swc.saas.ibm.com/en-us/documentation/clusters#get-pull-secret).\n2.
          Install the IBM Metrics Operator\n3. Create a Kubernetes secret in the installed
          namespace with the name `redhat-marketplace-pull-secret` and key `PULL_SECRET`
          with the value of the pull secret.\n    ```sh\n    # Replace ${PULL_SECRET}
          with your pull secret\n    oc create secret generic redhat-marketplace-pull-secret
          -n ibm-software-central --from-literal=PULL_SECRET=${PULL_SECRET}\n    ```\n4.
          Use of the platform is governed by the:\n\n    [IBM Cloud Services Agreement](https://www.ibm.com/support/customer/csol/terms/?id=Z126-6304_WS&_ga=2.116312197.2046730452.1684328846-812467790.1684328846)
          (or other base agreement between you and IBM such as a [Passport Advantage
          Agreement](https://www.ibm.com/software/passportadvantage/pa_agreements.html?_ga=2.116312197.2046730452.1684328846-812467790.1684328846))
          and the [Service Description for the Red Hat Marketplace](https://www.ibm.com/support/customer/csol/terms/?id=i126-8719&_ga=2.83289621.2046730452.1684328846-812467790.1684328846).\n
          \   \n5. Update MarketplaceConfig to accept the license.\n    ```\n    oc
          patch marketplaceconfig marketplaceconfig -n ibm-software-central --type='merge'
          -p '{\"spec\": {\"license\": {\"accept\": true}}}'\n    ```\n6. Install
          the pull secret as a global pull secret on the cluster.\n\n    These steps
          require `oc`, `jq`, and `base64` to be available on your machine.\n\n    ```sh\n
          \   # Create the docker pull secret file using your pull secret.\n    #
          Store it in a file called entitledregistryconfigjson.\n    oc create secret
          docker-registry entitled-registry --docker-server=registry.marketplace.redhat.com
          --docker-username \"cp\" --docker-password \"${PULL_SECRET}\" --dry-run=client
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > entitledregistryconfigjson\n
          \   # Get the current global secrets on the cluster and store it as a file
          named dockerconfigjson\n    oc get secret pull-secret -n openshift-config
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > dockerconfigjson\n
          \   # Merge the two dockerconfigs together into a file called dockerconfigjson-merged.\n
          \   jq -s '.[0] * .[1]' dockerconfigjson entitledregistryconfigjson > dockerconfigjson-merged\n
          \   # Set the cluster's dockerconfig file to the new merged version.\n    oc
          set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson=dockerconfigjson-merged\n
          \   ```\n  7. (Optional) Your IBM product may have installed IBM Metrics
          Operator using an `ibm-entitlement-key` for authentication to IBM Software
          Central. To authorize the automatic creation of an IBM Software Central
          account for the IBM account owning the `ibm-entitlement-key`\n      ```\n
          \     oc patch marketplaceconfig marketplaceconfig -n ibm-software-central
          --type='merge' -p '{\"spec\": {\"authorizeAccountCreation\": true}}'\n      ```\n\n###
          Why is a global pull secret required?\nIn order to successfully install
          the products hosted by the container image registry, you will need to make
          the pull secret available across the cluster. This can be achieved by applying
          the pull as a [global pull secret](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html#images-update-global-pull-secret_using-image-pull-secrets).
          For alternative approaches, please see the official OpenShift [documentation](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html).\n\n\n###
          SecurityContextConstraints requirements\n\nThe Operators and their components
          support running under the OpenShift Container Platform default restricted
          and restricted-v2 security context constraints.\n\n### Installation Namespace
          and ClusterRoleBinding requirements\n\nThe IBM Metrics Operator components
          require specific ClusterRoleBindings.\n- The metric-state component requires
          a ClusterRoleBinding for the the `view` ClusterRole.\n  - Ability to read
          non-sensitive CustomResources. \n  - The `view` ClusterRole dynamically
          adds new CustomResourceDefinitions.\n- The operator & reporter component
          requires a ClusterRoleBinding for the the `cluster-monitoring-view` ClusterRole.\n
          \ - Ability to view Prometheus metrics. \n  - The underlying ClusterRole
          RBAC often updated between OpenShift versions.\n\nDue to limitations of
          Operator Lifecycle Manager (OLM), this ClusterRoleBinding can not be provided
          dynamically for arbitrary installation target namespaces.\n\nA static ClusterRoleBinding
          is included for installation to the default namespace of `ibm-software-central`,
          and namespaces `openshift-redhat-marketplace`,  `redhat-marketplace`, `ibm-common-services`.\n\nTo
          create the ClusterRoleBindings for installation to an alternate namespace\n```\noc
          project INSTALL-NAMESPACE\noc adm policy add-cluster-role-to-user view -z
          ibm-metrics-operator-metric-state\noc adm policy add-cluster-role-to-user
          cluster-monitoring-view -z ibm-metrics-operator-controller-manager,ibm-metrics-operator-reporter\n```\n\n###
          Metric State scoping requirements\nThe metric-state Deployment obtains `get/list/watch`
          access to metered resources via the `view` ClusterRole. For operators deployed
          using Operator Lifecycle Manager (OLM), permissions are added to `clusterrole/view`
          dynamically via a generated and annotated `-view` ClusterRole. If you wish
          to meter an operator, and its Custom Resource Definitions (CRDs) are not
          deployed through OLM, one of two options are required\n1. Add the following
          label to a clusterrole that has get/list/watch access to your CRD: `rbac.authorization.k8s.io/aggregate-to-view:
          \"true\"`, thereby dynamically adding it to `clusterrole/view`\n2. Create
          a ClusterRole that has get/list/watch access to your CRD, and create a ClusterRoleBinding
          for the metric-state ServiceAccount\n\nAttempting to meter a resource with
          a MeterDefinition without the required permissions will log an `AccessDeniedError`
          in metric-state.\n\n### Disaster Recovery\n\nTo plan for disaster recovery,
          note the PhysicalVolumeClaims `rhm-data-service-rhm-data-service-N`. \n-
          In connected environments, MeterReport data upload attempts occur hourly,
          and are then removed from data-service. There is a low risk of losing much
          unreported data.\n- In an airgap environment, MeterReport data must be pulled
          from data-service and uploaded manually using [datactl](https://github.com/redhat-marketplace/datactl/).
          To prevent data loss in a disaster scenario, the data-service volumes should
          be considered in a recovery plan.\n\n### Subscription Config\n\nIt is possible
          to [configure](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md)
          how OLM deploys an Operator via the `config` field in the [Subscription](https://github.com/operator-framework/olm-book/blob/master/docs/subscriptions.md)
          object.\n\nThe IBM Metrics Operator will also read the `config` and append
          it to the operands. The primary use case is to control scheduling using
          [Tolerations](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/)
          and [NodeSelectors](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/).\n\nA
          limitation is that the `config` elements are only appended to the operands.
          The elements in the operands are not removed if the `config` is removed
          from the `Subscripion`. The operand must be modified manually, or deleted
          and recreated by the controller.\n\n### Cluster permission requirements\n\n|API
          group             |Resources              |Verbs                                     |\n|----------------------|-----------------------|------------------------------------------|\n|apps
          \                 |statefulsets           |get;list;watch                            |\n|authentication.k8s.io
          |tokenreviews           |create                                    |\n|authorization.k8s.io
          \ |subjectaccessreviews   |create                                    |\n|config.openshift.io
          \  |clusterversions        |get;list;watch                            |\n|marketplace.redhat.com|meterdefinitions
          \      |get;list;watch;create;update;patch;delete |\n|operators.coreos.com
          \ |clusterserviceversions |get;list;watch                            |\n|operators.coreos.com
          \ |subscriptions          |get;list;watch                            |\n|operators.coreos.com
          \ |operatorgroups         |get;list                                  |\n|storage.k8s.io
          \       |storageclasses         |get;list;watch                            |\n|
          \                     |namespaces             |get;list;watch                            |\n|
          \                     |configmaps             |get;list;watch                            |\n|
          \                     |services               |get;list;watch                            |\n|
          \                     |**clusterrole/view**   |                                          |\n\n\n###
          Documentation\nYou can find our documentation [here.](https://swc.saas.ibm.com/en-us/documentation/)\n\n###
          Getting help\nIf you encounter any issues while using the operators, you
          can create an issue on our [Github\nrepo](https://github.com/redhat-marketplace/redhat-marketplace-operator)
          for bugs, enhancements, or other requests. You can also visit our main page
          and\nreview our [support](https://swc.saas.ibm.com/en-us/support) and [documentation](https://swc.saas.ibm.com/en-us/documentation/).\n\n###
          Readme\nYou can find our readme [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/README.md)\n\n###
          License information\nYou can find our license information [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/LICENSE)\n"
        displayName: IBM Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        - buy
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/redhat-marketplace-data-service@sha256:30fd8d2996b921cfce51f697ee4490a6d9a836021952429b877336f5edf2749d
        - quay.io/rh-marketplace/redhat-marketplace-reporter@sha256:2bd098f057ca07f2b14be9b60ea6ad8ca6b0cae1d226b1ca3b27e23d04ccdf20
        - quay.io/rh-marketplace/redhat-marketplace-metric-state@sha256:e4014c9a546c10e84b13a09db6a18431ee9489ad1296808635e48d9ef4963cf1
        - quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
        - quay.io/rh-marketplace/redhat-marketplace-authcheck@sha256:13a81b8aca79ae8c89ee24dcc30f3322a4cfdedf929e7bb984295c1e7e631fa9
        version: 2.20.0
      entries:
      - name: ibm-metrics-operator.v2.20.0
        version: 2.20.0
      name: beta
    - currentCSV: ibm-metrics-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MarketplaceConfig",
                "metadata": {
                  "name": "marketplaceconfig"
                },
                "spec": {
                  "features": {
                    "deployment": true,
                    "meterDefinitionCatalogServer": false,
                    "registration": true
                  },
                  "installIBMCatalogSource": true,
                  "isDisconnected": false,
                  "license": {
                    "accept": false
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterBase",
                "metadata": {
                  "name": "meterbase"
                },
                "spec": {
                  "enabled": true,
                  "prometheus": {
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "1",
                        "memory": "1G"
                      }
                    },
                    "storage": {
                      "size": "20Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterReport",
                "metadata": {
                  "name": "example-meterreport"
                },
                "spec": {
                  "endTime": "2020-08-05T00:00:00Z",
                  "prometheusService": {
                    "name": "rhm-prometheus-meterbase",
                    "namespace": "openshift-redhat-marketplace",
                    "targetPort": "rbac"
                  },
                  "startTime": "2020-08-04T00:00:00Z"
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "example-meterdefinition-4"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "meterdefinition-sample"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
          createdAt: "2025-08-28T14:18:55Z"
          description: The IBM Metrics Operator provides workload metering and reporting
            for IBM and Red Hat Marketplace customers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
          - description: MeterBase is the resource that sets up Metering for Red Hat
              Marketplace. This is an internal resource not meant to be modified directly.
            displayName: Meter Base
            kind: MeterBase
            name: meterbases.marketplace.redhat.com
            version: v1alpha1
          - description: MeterDefinition defines the meter workloads used to enable
              pay for use billing.
            displayName: Meter Definition
            kind: MeterDefinition
            name: meterdefinitions.marketplace.redhat.com
            version: v1beta1
          - kind: MeterReport
            name: meterreports.marketplace.redhat.com
            version: v1alpha1
        description: "The IBM Metrics Operator provides workload metering and reporting
          for IBM Software Central and Red Hat Marketplace customers.\n### **Important
          Note**\nA set of instructions for onboarding is provided here. For more
          detailed onboarding instructions or information about what is installed
          please visit [swc.saas.ibm.com](https://swc.saas.ibm.com).\n\nUsage metrics
          may be monitored through [https://swc.saas.ibm.com](https://swc.saas.ibm.com)
          with only IBM Metrics Operator and a Red Hat Marketplace account.\n\n###
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n### Prerequisites\n1. Installations are required
          to [enable monitoring for user-defined projects](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          as the Prometheus provider.\n2. Edit the cluster-monitoring-config ConfigMap
          object:\n\n   ```sh\n   $ oc -n openshift-monitoring edit configmap cluster-monitoring-config\n
          \   ```\n\n3. Add enableUserWorkload: true under data/config.yaml:\n  \n
          \   ```sh\n    apiVersion: v1\n    kind: ConfigMap\n    metadata:\n      name:
          cluster-monitoring-config\n        namespace: openshift-monitoring\n    data:\n
          \     config.yaml: |\n        enableUserWorkload: true\n    ```\n\n4. Configure
          the user-workload-monitoring-config ConfigMap object:\n\n    ```sh\n    $
          oc -n openshift-user-workload-monitoring edit configmap user-workload-monitoring-config\n
          \   ```\n\n5. Configure a minimum retention time of 168h and minimum storage
          capacity of 40Gi\n  \n    ```sh\n    apiVersion: v1\n    kind: ConfigMap\n
          \   metadata:\n      name: user-workload-monitoring-config\n      namespace:
          openshift-user-workload-monitoring\n\n    data:\n      config.yaml: |\n
          \       prometheus:\n          retention: 168h\n          volumeClaimTemplate:\n
          \           spec:\n              resources:\n                requests:\n
          \                 storage: 40Gi\n    ```\n\n### Resources Required\n\nMinimum
          system resources required:\n\n| Operator                | Memory (MB) |
          CPU (cores) | Disk (GB) | Nodes |\n| ----------------------- | -----------
          | ----------- | --------- | ----- |\n| **Metrics**   |        750  |     0.25
          \   | 3x1       |    3  |\n| **Deployment** |        250  |     0.25    |
          -         |    1  |\n\n| Prometheus Provider  | Memory (GB) | CPU (cores)
          | Disk (GB) | Nodes |\n| --------- | ----------- | ----------- | ---------
          | ----- |\n| **[Openshift User Workload Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)**
          |          1  |     0.1       | 2x40        |   2    |\n\nMultiple nodes
          are required to provide pod scheduling for high availability for IBM Metrics
          Operator Data Service and Prometheus.\n\nThe IBM Metrics Operator automatically
          creates 3 x 1Gi PersistentVolumeClaims to store reports as part of the data
          service, with _ReadWriteOnce_ access mode. Te PersistentVolumeClaims are
          automatically created by the ibm-metrics-operator after creating a `redhat-marketplace-pull-secret`
          and accepting the license in `marketplaceconfig`.\n\n| NAME                                |
          CAPACITY | ACCESS MODES |\n| ----------------------------------- | --------
          | ------------ |\n| rhm-data-service-rhm-data-service-0 | 1Gi | RWO |\n|
          rhm-data-service-rhm-data-service-1 | 1Gi | RWO |\n| rhm-data-service-rhm-data-service-2
          | 1Gi | RWO |\n\n### Supported Storage Providers\n\n- OpenShift Container
          Storage / OpenShift Data Foundation version 4.x, from version 4.2 or higher\n-
          IBM Cloud Block storage and IBM Cloud File storage\n- IBM Storage Suite
          for IBM Cloud Paks:\n  - File storage from IBM Spectrum Fusion/Scale \n
          \ - Block storage from IBM Spectrum Virtualize, FlashSystem or DS8K\n- Portworx
          Storage, version 2.5.5 or above\n- Amazon Elastic File Storage\n\n### Access
          Modes required\n\n - ReadWriteOnce (RWO)\n\n### Provisioning Options supported\n\nChoose
          one of the following options to provision storage for the ibm-metrics-operator
          data-service\n\n#### Dynamic provisioning using a default StorageClass\n
          \  - A StorageClass is defined with a `metadata.annotations: storageclass.kubernetes.io/is-default-class:
          \"true\"`\n   - PersistentVolumes will be provisioned automatically for
          the generated PersistentVolumeClaims\n--- \n#### Manually create each PersistentVolumeClaim
          with a specific StorageClass\n   - Must be performed before creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`. Otherwise, the automatically
          generated PersistentVolumeClaims are immutable.\n```\napiVersion: v1\nkind:
          PersistentVolumeClaim\nmetadata:\n  labels:\n    app: rhm-data-service\n
          \ name: rhm-data-service-rhm-data-service-0\n  namespace: ibm-software-central\nspec:\n
          \ storageClassName: rook-cephfs\n  accessModes:\n  - ReadWriteOnce\n  resources:\n
          \   requests:\n      storage: 1Gi\n```\n---\n#### Manually provision each
          PersistentVolume for the generated PersistentVolumeClaims with a specific
          StorageClass\n  - May be performed before or after creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`.  \n```\napiVersion: v1\nkind:
          PersistentVolume\nmetadata:\n  name: rhm-data-service-rhm-data-service-0\nspec:\n
          \ csi:\n    driver: rook-ceph.cephfs.csi.ceph.com\n    volumeHandle: rhm-data-service-rhm-data-service-0\n
          \ capacity:\n    storage: 1Gi\n  accessModes:\n    - ReadWriteOnce\n  persistentVolumeReclaimPolicy:
          Delete\n  storageClassName: rook-cephfs\n  volumeMode: Filesystem\n  claimRef:\n
          \   kind: PersistentVolumeClaim\n    namespace: ibm-software-central\n    name:
          rhm-data-service-rhm-data-service-0\n```\n\n### Installation\n1. Create
          or get your [pull secret](https://swc.saas.ibm.com/en-us/documentation/clusters#get-pull-secret).\n2.
          Install the IBM Metrics Operator\n3. Create a Kubernetes secret in the installed
          namespace with the name `redhat-marketplace-pull-secret` and key `PULL_SECRET`
          with the value of the pull secret.\n    ```sh\n    # Replace ${PULL_SECRET}
          with your pull secret\n    oc create secret generic redhat-marketplace-pull-secret
          -n ibm-software-central --from-literal=PULL_SECRET=${PULL_SECRET}\n    ```\n4.
          Use of the platform is governed by the:\n\n    [IBM Cloud Services Agreement](https://www.ibm.com/support/customer/csol/terms/?id=Z126-6304_WS&_ga=2.116312197.2046730452.1684328846-812467790.1684328846)
          (or other base agreement between you and IBM such as a [Passport Advantage
          Agreement](https://www.ibm.com/software/passportadvantage/pa_agreements.html?_ga=2.116312197.2046730452.1684328846-812467790.1684328846))
          and the [Service Description for the Red Hat Marketplace](https://www.ibm.com/support/customer/csol/terms/?id=i126-8719&_ga=2.83289621.2046730452.1684328846-812467790.1684328846).\n
          \   \n5. Update MarketplaceConfig to accept the license.\n    ```\n    oc
          patch marketplaceconfig marketplaceconfig -n ibm-software-central --type='merge'
          -p '{\"spec\": {\"license\": {\"accept\": true}}}'\n    ```\n6. Install
          the pull secret as a global pull secret on the cluster.\n\n    These steps
          require `oc`, `jq`, and `base64` to be available on your machine.\n\n    ```sh\n
          \   # Create the docker pull secret file using your pull secret.\n    #
          Store it in a file called entitledregistryconfigjson.\n    oc create secret
          docker-registry entitled-registry --docker-server=registry.marketplace.redhat.com
          --docker-username \"cp\" --docker-password \"${PULL_SECRET}\" --dry-run=client
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > entitledregistryconfigjson\n
          \   # Get the current global secrets on the cluster and store it as a file
          named dockerconfigjson\n    oc get secret pull-secret -n openshift-config
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > dockerconfigjson\n
          \   # Merge the two dockerconfigs together into a file called dockerconfigjson-merged.\n
          \   jq -s '.[0] * .[1]' dockerconfigjson entitledregistryconfigjson > dockerconfigjson-merged\n
          \   # Set the cluster's dockerconfig file to the new merged version.\n    oc
          set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson=dockerconfigjson-merged\n
          \   ```\n  7. (Optional) Your IBM product may have installed IBM Metrics
          Operator using an `ibm-entitlement-key` for authentication to IBM Software
          Central. To authorize the automatic creation of an IBM Software Central
          account for the IBM account owning the `ibm-entitlement-key`\n      ```\n
          \     oc patch marketplaceconfig marketplaceconfig -n ibm-software-central
          --type='merge' -p '{\"spec\": {\"authorizeAccountCreation\": true}}'\n      ```\n\n###
          Why is a global pull secret required?\nIn order to successfully install
          the products hosted by the container image registry, you will need to make
          the pull secret available across the cluster. This can be achieved by applying
          the pull as a [global pull secret](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html#images-update-global-pull-secret_using-image-pull-secrets).
          For alternative approaches, please see the official OpenShift [documentation](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html).\n\n\n###
          SecurityContextConstraints requirements\n\nThe Operators and their components
          support running under the OpenShift Container Platform default restricted
          and restricted-v2 security context constraints.\n\n### Installation Namespace
          and ClusterRoleBinding requirements\n\nThe IBM Metrics Operator components
          require specific ClusterRoleBindings.\n- The metric-state component requires
          a ClusterRoleBinding for the the `view` ClusterRole.\n  - Ability to read
          non-sensitive CustomResources. \n  - The `view` ClusterRole dynamically
          adds new CustomResourceDefinitions.\n- The operator & reporter component
          requires a ClusterRoleBinding for the the `cluster-monitoring-view` ClusterRole.\n
          \ - Ability to view Prometheus metrics. \n  - The underlying ClusterRole
          RBAC often updated between OpenShift versions.\n\nDue to limitations of
          Operator Lifecycle Manager (OLM), this ClusterRoleBinding can not be provided
          dynamically for arbitrary installation target namespaces.\n\nA static ClusterRoleBinding
          is included for installation to the default namespace of `ibm-software-central`,
          and namespaces `openshift-redhat-marketplace`,  `redhat-marketplace`, `ibm-common-services`.\n\nTo
          create the ClusterRoleBindings for installation to an alternate namespace\n```\noc
          project INSTALL-NAMESPACE\noc adm policy add-cluster-role-to-user view -z
          ibm-metrics-operator-metric-state\noc adm policy add-cluster-role-to-user
          cluster-monitoring-view -z ibm-metrics-operator-controller-manager,ibm-metrics-operator-reporter\n```\n\n###
          Metric State scoping requirements\nThe metric-state Deployment obtains `get/list/watch`
          access to metered resources via the `view` ClusterRole. For operators deployed
          using Operator Lifecycle Manager (OLM), permissions are added to `clusterrole/view`
          dynamically via a generated and annotated `-view` ClusterRole. If you wish
          to meter an operator, and its Custom Resource Definitions (CRDs) are not
          deployed through OLM, one of two options are required\n1. Add the following
          label to a clusterrole that has get/list/watch access to your CRD: `rbac.authorization.k8s.io/aggregate-to-view:
          \"true\"`, thereby dynamically adding it to `clusterrole/view`\n2. Create
          a ClusterRole that has get/list/watch access to your CRD, and create a ClusterRoleBinding
          for the metric-state ServiceAccount\n\nAttempting to meter a resource with
          a MeterDefinition without the required permissions will log an `AccessDeniedError`
          in metric-state.\n\n### Disaster Recovery\n\nTo plan for disaster recovery,
          note the PhysicalVolumeClaims `rhm-data-service-rhm-data-service-N`. \n-
          In connected environments, MeterReport data upload attempts occur hourly,
          and are then removed from data-service. There is a low risk of losing much
          unreported data.\n- In an airgap environment, MeterReport data must be pulled
          from data-service and uploaded manually using [datactl](https://github.com/redhat-marketplace/datactl/).
          To prevent data loss in a disaster scenario, the data-service volumes should
          be considered in a recovery plan.\n\n### Subscription Config\n\nIt is possible
          to [configure](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md)
          how OLM deploys an Operator via the `config` field in the [Subscription](https://github.com/operator-framework/olm-book/blob/master/docs/subscriptions.md)
          object.\n\nThe IBM Metrics Operator will also read the `config` and append
          it to the operands. The primary use case is to control scheduling using
          [Tolerations](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/)
          and [NodeSelectors](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/).\n\nA
          limitation is that the `config` elements are only appended to the operands.
          The elements in the operands are not removed if the `config` is removed
          from the `Subscripion`. The operand must be modified manually, or deleted
          and recreated by the controller.\n\n### Cluster permission requirements\n\n|API
          group             |Resources              |Verbs                                     |\n|----------------------|-----------------------|------------------------------------------|\n|apps
          \                 |statefulsets           |get;list;watch                            |\n|authentication.k8s.io
          |tokenreviews           |create                                    |\n|authorization.k8s.io
          \ |subjectaccessreviews   |create                                    |\n|config.openshift.io
          \  |clusterversions        |get;list;watch                            |\n|marketplace.redhat.com|meterdefinitions
          \      |get;list;watch;create;update;patch;delete |\n|operators.coreos.com
          \ |clusterserviceversions |get;list;watch                            |\n|operators.coreos.com
          \ |subscriptions          |get;list;watch                            |\n|operators.coreos.com
          \ |operatorgroups         |get;list                                  |\n|storage.k8s.io
          \       |storageclasses         |get;list;watch                            |\n|
          \                     |namespaces             |get;list;watch                            |\n|
          \                     |configmaps             |get;list;watch                            |\n|
          \                     |services               |get;list;watch                            |\n|
          \                     |**clusterrole/view**   |                                          |\n\n\n###
          Documentation\nYou can find our documentation [here.](https://swc.saas.ibm.com/en-us/documentation/)\n\n###
          Getting help\nIf you encounter any issues while using the operators, you
          can create an issue on our [Github\nrepo](https://github.com/redhat-marketplace/redhat-marketplace-operator)
          for bugs, enhancements, or other requests. You can also visit our main page
          and\nreview our [support](https://swc.saas.ibm.com/en-us/support) and [documentation](https://swc.saas.ibm.com/en-us/documentation/).\n\n###
          Readme\nYou can find our readme [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/README.md)\n\n###
          License information\nYou can find our license information [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/LICENSE)\n"
        displayName: IBM Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        - buy
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/redhat-marketplace-data-service@sha256:30fd8d2996b921cfce51f697ee4490a6d9a836021952429b877336f5edf2749d
        - quay.io/rh-marketplace/redhat-marketplace-reporter@sha256:2bd098f057ca07f2b14be9b60ea6ad8ca6b0cae1d226b1ca3b27e23d04ccdf20
        - quay.io/rh-marketplace/redhat-marketplace-metric-state@sha256:e4014c9a546c10e84b13a09db6a18431ee9489ad1296808635e48d9ef4963cf1
        - quay.io/rh-marketplace/redhat-marketplace-authcheck@sha256:13a81b8aca79ae8c89ee24dcc30f3322a4cfdedf929e7bb984295c1e7e631fa9
        - quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 2.20.0
      entries:
      - name: ibm-metrics-operator.v2.20.0
        version: 2.20.0
      name: stable
    defaultChannel: stable
    packageName: ibm-metrics-operator
    provider:
      name: Red Hat Marketplace
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-dr-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-dr-manager.v8.0.220
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com",
                  "resRecord": {
                    "HostName_1": "100.200.30.41",
                    "HostName_2": "100.200.30.43",
                    "abc": "HostName_2",
                    "www": "HostName_1",
                    "xyz": "HostName_3"
                  },
                  "stealthMasters": [
                    "1.2.3.4"
                  ],
                  "tsigKeyFile": "/Kvcslnxdns.com.+157+10641.private"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "dataReplicationPointer": "test-datareplication",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 10,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          olm.skipRange: <8.0.220
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
            - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:a98f9cc51f2c7baeec6dbb4933a42ae8ae20ff1a42c9d48a37174cf5211bbcf8
        version: 8.0.220
      entries:
      - name: infoscale-dr-manager.v8.0.220
        version: 8.0.220
      - name: infoscale-dr-manager.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-dr-manager.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          createdAt: "2024-06-17T10:14:13Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:c46cbf78631fc790c77b6c13c1d1d4b5ab35f49164f7c2e36c7ec59220cbdce2
        version: 8.0.330
      entries:
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-dr-manager.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
          createdAt: "2024-02-22T14:14:42Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721648-168743109)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721654-168743109)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
        version: 8.0.410
      entries:
      - name: infoscale-dr-manager.v8.0.410
        version: 8.0.410
      - name: infoscale-dr-manager.v8.0.409
        version: 8.0.409
      - name: infoscale-dr-manager.v8.0.400
        version: 8.0.400
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-dr-manager.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
          createdAt: "2024-02-22T14:14:42Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721648-168743109)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721654-168743109)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
        version: 8.0.410
      entries:
      - name: infoscale-dr-manager.v8.0.410
        version: 8.0.410
      - name: infoscale-dr-manager.v8.0.409
        version: 8.0.409
      - name: infoscale-dr-manager.v8.0.400
        version: 8.0.400
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-dr-manager.v8.0.100
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com",
                  "resRecord": {
                    "abc": "r7515-054-vm9",
                    "r7515-054-vm10": "10.221.85.83",
                    "r7515-054-vm8": "10.221.85.81",
                    "r7515-054-vm9": "10.221.85.82",
                    "www": "r7515-054-vm8",
                    "xyz": "r7515-054-vm10"
                  },
                  "stealthMasters": [
                    "1.2.3.4"
                  ],
                  "tsigKeyFile": "/Kvcslnxdns.com.+157+10641.private"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "currentPrimary": "Clus1",
                  "force": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "fail",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "dataReplicationPointer": "test-datareplication",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 10,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:eb999187a1f9a9ce5566089e24adf598b1c229d31c5e57ddcfe4439ba25e97dc
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
            - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membershrip
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membershrip
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponsding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:eb999187a1f9a9ce5566089e24adf598b1c229d31c5e57ddcfe4439ba25e97dc
        version: 8.0.100
      entries:
      - name: infoscale-dr-manager.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-dr-manager
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NdmSpc project
      provider-url: ""
    name: ndmspc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ndmspc-operator.v0.20250209.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.ndmspc.io/v1alpha1",
                "kind": "NdmSpcConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ndmspc-operator",
                    "app.kubernetes.io/instance": "ndmspcconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ndmspcconfig",
                    "app.kubernetes.io/part-of": "ndmspc-operator"
                  },
                  "name": "ndmspc-sample"
                },
                "spec": {
                  "executor": {
                    "image": "registry.gitlab.com/ndmspc/api:v0.20240215.2",
                    "name": "executor"
                  },
                  "executors": 1,
                  "ingressclasstype": "nginx",
                  "ingressenabled": true,
                  "ingresshostdomain": "127.0.0.1.sslip.io",
                  "ingressnostname": "ndmspc",
                  "ingressssltraefik": false,
                  "naf": {
                    "image": "registry.gitlab.com/ndmspc/naf:v0.20250129.0",
                    "name": "naf"
                  },
                  "ndmvr": {
                    "image": "registry.gitlab.com/ndmspc/ndmvr/shared:v0.20230320.0",
                    "name": "ndmvr"
                  },
                  "salsa": {
                    "image": "registry.gitlab.com/ndmspc/user/ndmspc:v0.20240412.0"
                  },
                  "storage": {
                    "image": "registry.gitlab.com/ndmspc/storage-server:v0.20250205.0",
                    "mongodb": {
                      "image": "docker.io/library/mongo:latest",
                      "password": "admin",
                      "port": 27017,
                      "username": "admin"
                    },
                    "name": "storage",
                    "port": 8080
                  },
                  "webapp": {
                    "image": "registry.gitlab.com/ndmspc/react-ndmspc:v0.20240412.0",
                    "name": ""
                  },
                  "webappcontainerport": 8080,
                  "webapps": 1,
                  "worker": {
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "1Gi"
                      }
                    }
                  },
                  "workers": 2,
                  "workerslots": 1,
                  "zmq2ws": {
                    "image": "registry.gitlab.com/ndmspc/zmq2ws:v1.20231012.0",
                    "name": "zmq2ws"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: registry.gitlab.com/ndmspc/ndmspc-operator:0.20250209.0
          createdAt: "2025-02-09T12:49:06Z"
          description: The NdmSpc operator automates the deployment and management
            of Ndmspc suite on Kubernetes cluster.
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://gitlab.com/ndmspc/ndmspc-operator
          support: NDMSPC Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NdmSpc configuration
            displayName: NdmSpc Config
            kind: NdmSpcConfig
            name: ndmspcconfigs.apps.ndmspc.io
            version: v1alpha1
        description: "The NdmSpc operator automates the deployment and management
          of Ndmspc suite on Kubernetes cluster. This operator allows user to deploy
          SALSA cluster with ZMQ2WS and job executor service and specify custom webapp
          page.\n\n### Capabilities\n\n- SALSA cluster\n- ZMQ2WS api service\n- Job
          executor service \n- Custom WebApp page\n"
        displayName: NdmSpc operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ndmspc
        links:
        - name: Ndmspc Operator
          url: https://ndmspc.gitlab.io/docs
        maintainers:
        - email: ndmspc@gmail.com
          name: NdmSpc Support
        maturity: alpha
        provider:
          name: NdmSpc project
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - registry.gitlab.com/ndmspc/ndmspc-operator:0.20250209.0
        version: 0.20250209.0
      entries:
      - name: ndmspc-operator.v0.20250209.0
        version: 0.20250209.0
      - name: ndmspc-operator.v0.20250131.0
        version: 0.20250131.0
      - name: ndmspc-operator.v0.20240412.0
        version: 0.20240412.0
      - name: ndmspc-operator.v0.11.6
        version: 0.11.6
      - name: ndmspc-operator.v0.11.4
        version: 0.11.4
      - name: ndmspc-operator.v0.11.3
        version: 0.11.3
      - name: ndmspc-operator.v0.11.2
        version: 0.11.2
      - name: ndmspc-operator.v0.11.1
        version: 0.11.1
      - name: ndmspc-operator.v0.10.2
        version: 0.10.2
      - name: ndmspc-operator.v0.10.1
        version: 0.10.1
      - name: ndmspc-operator.v0.10.0
        version: 0.10.0
      - name: ndmspc-operator.v0.9.0
        version: 0.9.0
      - name: ndmspc-operator.v0.7.4
        version: 0.7.4
      - name: ndmspc-operator.v0.7.1
        version: 0.7.1
      - name: ndmspc-operator.v0.6.0
        version: 0.6.0
      - name: ndmspc-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: ndmspc-operator
    provider:
      name: NdmSpc project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: sosivio
      provider-url: https://www.sosiv.io
    name: sosivio
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sosivio-operator.v1.5.1-97
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sosivio.sosiv.io/v1alpha1",
                "kind": "Sosivio",
                "metadata": {
                  "name": "sosivio"
                },
                "spec": {
                  "platform": "ocp4",
                  "useSha": false
                }
              },
              {
                "apiVersion": "sosivio.sosiv.io/v1alpha1",
                "kind": "Sosivio",
                "metadata": {
                  "name": "sosivio"
                },
                "spec": {
                  "platform": "ocp4",
                  "useSha": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Monitoring,Integration & Delivery,Logging & Tracing,Developer
            Tools
          com.redhat.openshift.versions: v4.1-v4.1
          containerImage: quay.io/sosivio/sosivio-operator:v1.5.1-97
          createdAt: "2021-10-28 01:01:01"
          description: Sosivio | Predictive Troubleshooting for Kubernetes
          operatorhub.io/ui-metadata-max-k8s-version: "1.25"
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Sosivio/operator
          support: Sosivio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Sosivio is the Schema for the sosivios API
            displayName: Sosivio
            kind: Sosivio
            name: sosivios.sosivio.sosiv.io
            version: v1alpha1
        description: "# Sosivio Operator\nThe Sosivio Operator for RedHat’s Openshift
          Container Platform allows users to quickly and easily deploy Sosivio’s Kubernetes
          Predictive Troubleshooting Platform on your cluster.\n\n# Sosivio Overview\n\nSosivio
          is the first and only non-intrusive and predictive autonomous cloud-native
          platform built specifically for Kubernetes from the ground up. Sosivio provides
          predictive troubleshooting tools, root cause determination and instant remediation
          for any failures in Kubernetes environments.\n\nPredict and Prevent!\n\nThe
          Operator will deploy and manage Sosivio on Kubernetes and OpenShift. Included
          are the following features;\n-   **Root Cause Analysis and Remediation**
          - predictive troubleshooting to prevent failures before they occur. Instantly
          understand where the root cause is and how to fix it.\n-   **Application
          Profiling** - analyze container resource consumption and recommendations
          for requests and limits.\n-   **Real-time Metrics** - in-depth, granular,
          and accurate real-time data from the process level inside containers (and
          pods) with failure correlation.\n-   **Cluster Health Checks** - proactive
          health checks for cluster infrastructure and applications running on the
          cluster.\n-   **Multi-Cluster View** - management, monitoring, and maintenance
          capabilities for all clusters from one location. \n-   **Kubernetes Components
          Alerting** - notifications sent through Slack, email and other 3rd party
          platforms.\n\n# Installation\nSosivio is completely **non-intrusive** and
          **does not require any persistent storage**. There is **no data offloading**
          outside of your Kubernetes cluster. Applying our Data Swirling methodology
          enables the product to perform at a **remarkably low resource footprint**
          and requires **no persistent volume configurations**.\n\n\nSosivio is installed
          under one namespace (labeled: \"sosivio\") which creates all the required
          components for the product.\n\nAll users will start with 4 weeks of Sosivio
          Premium after which Sosivio will automatically convert to the Community
          Version (free forever).\n\n# Help\nMore information about Sosivio and our
          product usage instructions can be found in our documentation [here](https://docs.sosiv.io/)
          \ \n"
        displayName: Sosivio | Predictive Troubleshooting for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sosivio
        - kubernetes
        - observanility
        - monitoring
        - root cause
        links:
        - name: Sosivio
          url: https://www.sosiv.io
        - name: Documentation
          url: https://docs.sosiv.io
        maintainers:
        - email: info@sosiv.io
          name: Sosivio
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: sosivio
          url: https://www.sosiv.io
        relatedImages:
        - quay.io/sosivio/classifier@sha256:dd8bf145524db23ff333efa1bcdff258dab116d3534f95f73775d8dd339e6963
        - quay.io/sosivio/notifier@sha256:35f5f3e8e5ccad914fcfe34fb4b4a1b044ee87f5a6141284cd5273f0b2b3bf22
        - quay.io/sosivio/kube-rbac-proxy@sha256:34e8724e0f47e31eb2ec3279ac398b657db5f60f167426ee73138e2e84af6486
        - quay.io/sosivio/nsq@sha256:781aac23ef0a1ed0af2b8e93d625ae49bd11dbdd98787d3c79dde8d128f3810d
        - quay.io/sosivio/network-aggregator@sha256:e59d937f4d9c52acad0adadab39c6849d20b4ee72a83e239a33ea137f08323f4
        - quay.io/sosivio/sosivio-dashboard@sha256:fdfca5bb2123d1008cec2b656cde73c7e9364f84c5ff4f4a6fee5751b1bfb2f8
        - quay.io/sosivio/kube-rbac-proxy:v0.8.0
        - quay.io/sosivio/sequence-recognition@sha256:7335cf8f3e40bc46e8ca0de525d99f0655833b7c52d36dc273aa7b096eb70fa2
        - quay.io/sosivio/sosivio-node-pinger@sha256:937de52ad7b76e6cff4ad5197de9705988d61f285ebadac5ebadb5856f59ee58
        - quay.io/sosivio/actuator@sha256:a9abe9730800ca2b0a1bbd7a361b2ab3a818cbf127798314e5477fb1cb16ba4c
        - quay.io/sosivio/authentication@sha256:0ea6025f333576792f5b1fc4d96d9ab1a404749f185830410e71248683004da2
        - quay.io/sosivio/sosiviodb@sha256:a3c6c3975bc2b0316135ee5c547f1eb215f3a1f9c8b0a9c8eceff23415f93597
        - quay.io/sosivio/sosivio-node-collector@sha256:61ed8dd52de5edc09d16a31b18f2f65ad9344c0b6c9cfab323c3eb3d7e7ed060
        - quay.io/sosivio/sosivio-operator@sha256:11bc627773bb5e08b629e0b013e7f600fcb4aaefbd97f2de337d339475616763
        - quay.io/sosivio/contract-testing@sha256:cad6fac6ed3fdf2597ceb09ba5dbb7d81f70cf45f483a7064e3968215dd982e3
        - quay.io/sosivio/crud-manager@sha256:38289a28829fa737917d76dd4810ba6140e123061ead0a4b79fa79b2e246a293
        - quay.io/sosivio/communicator@sha256:fffec9a6c66594fde5f05e2dfb84361544ca9d53a345188c197a0ea6f53bc4b9
        - quay.io/sosivio/analyzer@sha256:3825608fe7c43ef059108e76665acc99fe72bea77d4d9e3852f6ad45e74379b1
        - quay.io/sosivio/discovery-engine@sha256:b87b3f7c9ed3aed80d0a77788fd8925789450d9dc05d25017e89dff611cd82da
        - quay.io/sosivio/sosivio-operator:v1.5.1-97
        - quay.io/sosivio/correlation-entities@sha256:c00af09b2589bad4740fd44d7a81c69b2e29138c360939b2a45f65edd44b6901
        - quay.io/sosivio/sosivio-node-netpacket@sha256:84dd159b1dfdb52dee3eb1c4c93f323bae49ce517ff44663b69978e14bdd8390
        version: 1.5.1
      entries:
      - name: sosivio-operator.v1.5.1-97
        version: 1.5.1
      - name: sosivio-operator.v1.5.0-95
        version: 1.5.0
      - name: sosivio-operator.v1.4.4-85
        version: 1.4.4
      - name: sosivio-operator.v1.4.3-83
        version: 1.4.3
      - name: sosivio-operator.v1.4.2-82
        version: 1.4.2
      - name: sosivio-operator.v1.4.1-81
        version: 1.4.1-81
      - name: sosivio-operator.v1.4.1.1-80
        version: 1.4.1-80
      - name: sosivio-operator.v1.4.1-78
        version: 1.4.1-79
      - name: sosivio-operator.v1.3.1-76
        version: 1.3.1-76
      - name: sosivio-operator.v1.3.0-75
        version: 1.3.0-75
      - name: sosivio-operator.v1.2.1-1
        version: 1.2.1-1
      - name: sosivio-operator.v1.2.0-1
        version: 1.2.0-1
      name: stable
    defaultChannel: stable
    packageName: sosivio
    provider:
      name: sosivio
      url: https://www.sosiv.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-keycloak-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-keycloak-operator.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2023-09-19T12:43:29Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KCOSClientRoleMapping
            name: kcosclientrolemappings.keycloak.org
            version: v1alpha1
          - kind: KCOSRealmRole
            name: kcosrealmroles.keycloak.org
            version: v1alpha1
          - kind: KCOSUserRoleMapping
            name: kcosuserrolemappings.keycloak.org
            version: v1alpha1
          - kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: Keysight Keycloak Operator
        displayName: Keysight Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Keycloak Operator
          url: https://keysight-keycloak-operator.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak@sha256:772f73cce6a77b9c46f8035faac99a28602219657d18b75b19b4d905d91942c9
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak-postgresql@sha256:b636955199065e18f9d4679835ae78214843258e5abb4ae758288d6f00df2f8e
        - prod.harbor.keysight.digital/loadcore/keysight-keycloak-operator@sha256:d74e7e98a114768a6f3327edabce9336743a078844fffee985053b6a90100da8
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak-init-container@sha256:90965beaf41c760cb8c8bf906acf08f8fe4770d313de8dfd92d18d7c25348a38
        version: 0.1.21
      entries:
      - name: keysight-keycloak-operator.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-keycloak-operator
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: fence-agents-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fence-agents-remediation.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  }
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:00cc420250d8e25d9da7c43d9c4cff35c7edeb7d390af754c48c0ce00e9e7770
          createdAt: "2023-10-30 14:06:57"
          description: Fence Agents Remediation Operator for remediating nodes using
            upstream fence-agents.
          olm.skipRange: '>=0.2.0 <0.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: "Fence Agents Remediation Operator (FAR)\n\nThe operator will
          remediate/fence a node when a FenceAgentsRemediation (far) custom resource
          exists with the node's name.\nFAR uses an existing set of upstream fencing
          agents for environments with a \ntraditional API end-point (e.g., IPMI)
          for power cycling cluster nodes.\n\nThe operator is recommended to be used
          with Node Healthcheck Operator for easier and smoother experience,\nbut
          it can also work as a standalonde remediator for the more advanced user.\n"
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:00cc420250d8e25d9da7c43d9c4cff35c7edeb7d390af754c48c0ce00e9e7770
        version: 0.2.1
      entries:
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.14-eus
    - currentCSV: fence-agents-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:a367e8fdd80938e97ac195631a996e0c537bf3666aef0199147534762dcf739f
          createdAt: "2024-08-10 22:49:23"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.2.0 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:a367e8fdd80938e97ac195631a996e0c537bf3666aef0199147534762dcf739f
        version: 0.4.1
      entries:
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.16-eus
    - currentCSV: fence-agents-remediation.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-16 10:25:57"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.5.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.5/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel9-operator@sha256:da5975557078fbfa6882df87cff5bda5671b9838b5fb0a738fea724469d03576
        version: 0.5.1
      entries:
      - name: fence-agents-remediation.v0.5.1
        version: 0.5.1
      - name: fence-agents-remediation.v0.5.0
        version: 0.5.0
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.18-eus
    - currentCSV: fence-agents-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  }
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:5c3778c7aae45f54e949fbae3d133777f7aa78fe1251ac3d5510aa28b71e2ba4
          createdAt: "2023-04-20 15:36:05"
          description: Fence Agents Remediation Operator for remediating nodes using
            upstream fence-agents.
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: "Fence Agents Remediation Operator (FAR)\n\nThe operator will
          remediate/fence a node when a FenceAgentsRemediation (far) custom resource
          exists with the node's name.\nFAR uses an existing set of upstream fencing
          agents for environments with a \ntraditional API end-point (e.g., IPMI)
          for power cycling cluster nodes.\n\nThe operator is recommended to be used
          with Node Healthcheck Operator for easier and smoother experience,\nbut
          it can also work as a standalonde remediator for the more advanced user.\n"
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3af07bfe0fb73b1a7f00487eba302d6da442e742ab4007fc50e52efd4d45dae5
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:5c3778c7aae45f54e949fbae3d133777f7aa78fe1251ac3d5510aa28b71e2ba4
        version: 0.1.0
      entries:
      - name: fence-agents-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: fence-agents-remediation.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeSecretNames": {
                    "master-0": "fence-agents-credentials-master0",
                    "master-1": "fence-agents-credentials-master1",
                    "master-2": "fence-agents-credentials-master2",
                    "worker-0": "fence-agents-credentials-worker0",
                    "worker-1": "fence-agents-credentials-worker1",
                    "worker-2": "fence-agents-credentials-worker2"
                  },
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedSecretName": "fence-agents-credentials-shared",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {
                    "spec": {
                      "agent": "fence_ipmilan",
                      "nodeSecretNames": {
                        "master-0": "fence-agents-credentials-master0",
                        "master-1": "fence-agents-credentials-master1",
                        "master-2": "fence-agents-credentials-master2",
                        "worker-0": "fence-agents-credentials-worker0",
                        "worker-1": "fence-agents-credentials-worker1",
                        "worker-2": "fence-agents-credentials-worker2"
                      },
                      "nodeparameters": {
                        "--ipport": {
                          "master-0": "6230",
                          "master-1": "6231",
                          "master-2": "6232",
                          "worker-0": "6233",
                          "worker-1": "6234",
                          "worker-2": "6235"
                        }
                      },
                      "remediationStrategy": "ResourceDeletion",
                      "retrycount": 5,
                      "retryinterval": "5s",
                      "sharedSecretName": "fence-agents-credentials-shared",
                      "sharedparameters": {
                        "--action": "reboot",
                        "--ip": "192.168.111.1",
                        "--lanplus": "",
                        "--username": "admin"
                      },
                      "timeout": "60s"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-09-08 09:19:00"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.5.0 <0.6.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.8/html/remediation_fencing_and_maintenance/about-remediation-fencing-maintenance
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel9-operator@sha256:eef1a298718650a22a3da4a07c140ce21b6f287d87af79801610317fa8f63d52
        version: 0.6.0
      entries:
      - name: fence-agents-remediation.v0.6.0
        version: 0.6.0
      - name: fence-agents-remediation.v0.5.1
        version: 0.5.1
      - name: fence-agents-remediation.v0.5.0
        version: 0.5.0
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: stable
    defaultChannel: stable
    packageName: fence-agents-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-s3control-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-s3control-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/s3control-controller:1.1.1
          createdAt: "2025-09-19T20:52:02Z"
          description: AWS S3 Control controller is a service controller for managing
            S3 Control resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessPoint represents the state of an AWS s3control AccessPoint
              resource.
            displayName: AccessPoint
            kind: AccessPoint
            name: accesspoints.s3control.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon S3 Control resources in AWS from within your Kubernetes cluster.

          **About Amazon S3 Control**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon S3 Control
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - s3control
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon S3 Control Developer Resources
          url: https://aws.amazon.com/S3 Control/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: s3control maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/s3control-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-s3control-controller.v1.1.1
        version: 1.1.1
      - name: ack-s3control-controller.v1.1.0
        version: 1.1.0
      - name: ack-s3control-controller.v1.0.11
        version: 1.0.11
      - name: ack-s3control-controller.v1.0.10
        version: 1.0.10
      name: alpha
    defaultChannel: alpha
    packageName: ack-s3control-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-efs-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-efs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "AccessPoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "FileSystem",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "MountTarget",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/efs-controller:1.1.1
          createdAt: "2025-09-19T18:38:13Z"
          description: AWS EFS controller is a service controller for managing EFS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessPoint represents the state of an AWS efs AccessPoint
              resource.
            displayName: AccessPoint
            kind: AccessPoint
            name: accesspoints.efs.services.k8s.aws
            version: v1alpha1
          - description: FileSystem represents the state of an AWS efs FileSystem
              resource.
            displayName: FileSystem
            kind: FileSystem
            name: filesystems.efs.services.k8s.aws
            version: v1alpha1
          - description: MountTarget represents the state of an AWS efs MountTarget
              resource.
            displayName: MountTarget
            kind: MountTarget
            name: mounttargets.efs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EFS resources in AWS from within your Kubernetes cluster.

          **About Amazon Elastic File System (Amazon EFS)**

          Amazon Elastic File System (Amazon EFS) provides serverless, fully elastic file storage so that you can share file data without provisioning or managing storage capacity and performance. Amazon EFS is built to scale on demand to petabytes without disrupting applications, growing and shrinking automatically as you add and remove files. Because Amazon EFS has a simple web services interface, you can create and configure file systems quickly and easily. The service manages all the file storage infrastructure for you, meaning that you can avoid the complexity of deploying, patching, and maintaining complex file system configurations.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Elastic File System (Amazon
          EFS)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - efs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EFS Developer Resources
          url: https://aws.amazon.com/efs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: efs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/efs-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-efs-controller.v1.1.1
        version: 1.1.1
      - name: ack-efs-controller.v1.1.0
        version: 1.1.0
      - name: ack-efs-controller.v1.0.19
        version: 1.0.19
      - name: ack-efs-controller.v1.0.18
        version: 1.0.18
      - name: ack-efs-controller.v1.0.17
        version: 1.0.17
      - name: ack-efs-controller.v1.0.16
        version: 1.0.16
      - name: ack-efs-controller.v1.0.15
        version: 1.0.15
      - name: ack-efs-controller.v1.0.14
        version: 1.0.14
      - name: ack-efs-controller.v1.0.13
        version: 1.0.13
      - name: ack-efs-controller.v1.0.12
        version: 1.0.12
      - name: ack-efs-controller.v1.0.11
        version: 1.0.11
      - name: ack-efs-controller.v1.0.10
        version: 1.0.10
      - name: ack-efs-controller.v1.0.8
        version: 1.0.8
      - name: ack-efs-controller.v1.0.6
        version: 1.0.6
      - name: ack-efs-controller.v1.0.5
        version: 1.0.5
      - name: ack-efs-controller.v1.0.3
        version: 1.0.3
      - name: ack-efs-controller.v1.0.2
        version: 1.0.2
      - name: ack-efs-controller.v1.0.0
        version: 1.0.0
      - name: ack-efs-controller.v0.0.9
        version: 0.0.9
      - name: ack-efs-controller.v0.0.8
        version: 0.0.8
      - name: ack-efs-controller.v0.0.7
        version: 0.0.7
      - name: ack-efs-controller.v0.0.6
        version: 0.0.6
      - name: ack-efs-controller.v0.0.5
        version: 0.0.5
      - name: ack-efs-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-efs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solace Corporation
      provider-url: www.solace.com
    name: pubsubplus-eventbroker-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pubsubplus-eventbroker-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pubsubplus.solace.com/v1beta1",
                "kind": "PubSubPlusEventBroker",
                "metadata": {
                  "name": "non-ha-standalone-example"
                },
                "spec": {
                  "redundancy": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "true"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.10
          containerImage: registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator:1.4.0
          createdAt: "2025-07-31T19:48:36Z"
          description: The Solace PubSub+ Event Broker Operator deploys and manages
            the lifecycle of PubSub+ Event Brokers
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
          support: Solace Products
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PubSub+ Event Broker
            displayName: PubSub+ Event Broker
            kind: PubSubPlusEventBroker
            name: pubsubpluseventbrokers.pubsubplus.solace.com
            version: v1beta1
        description: |
          ## Solace PubSub+ Platform
          Solace [PubSub+ Platform](https://solace.com/products/platform/) is a complete event streaming and management platform
          for real-time enterprises. The [PubSub+ Software Event Broker](https://solace.com/products/event-broker/software/) efficiently streams
          event-driven information between applications, IoT devices, and user interfaces running in the cloud, on-premises,
          and in hybrid environments using open APIs and protocols like AMQP, JMS, MQTT, REST and WebSocket.
          It can be installed into a variety of public and private clouds, PaaS, and on-premises environments.
          Event brokers in multiple locations can be linked together in an [Event Mesh](https://solace.com/what-is-an-event-mesh/)
          to dynamically share events across the distributed enterprise.

          ## Solace PubSub+ Event Broker Operator
          Solace Pubsub+ Event Broker Operator automatically deploys and manages PubSub+ Software Event Brokers on Kubernetes and OpenShift environments.

          Features include:
          * Initial deployment configuration: Production-ready HA, non-HA or for developers
          * Rolling update for configuration changes or broker upgrades
          * Prometheus monitoring support

          ## Getting started
          Refer to the [Solace PubSub+ Event Broker Operator Quick Start](https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/README.md)
        displayName: Solace PubSub+ Event Broker Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - solace
        - pubsubplus
        - pubsub+
        - pubsub
        - messaging
        - advanced event broker
        - event broker
        - event mesh
        - event streaming
        - data streaming
        - event integration
        - middleware
        links:
        - name: Operator QuickStart
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
        - name: Operator Documentation
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/docs/EventBrokerOperatorUserGuide.md
        - name: Solace.com
          url: https://solace.com
        - name: Understanding Solace Technology
          url: https://solace.com/products/platform/
        - name: Solace PubSub+ Technical Documentation
          url: https://docs.solace.com/
        - name: Solace Developer Portal
          url: https://www.solace.dev/
        - name: Solace Community
          url: https://solace.community/
        maintainers:
        - email: support@solace.com
          name: Solace R&D
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Solace Corporation
          url: www.solace.com
        relatedImages:
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator:1.4.0
        version: 1.4.0
      entries:
      - name: pubsubplus-eventbroker-operator.v1.4.0
        version: 1.4.0
      - name: pubsubplus-eventbroker-operator.v1.3.0
        version: 1.3.0
      - name: pubsubplus-eventbroker-operator.v1.2.0
        version: 1.2.0
      - name: pubsubplus-eventbroker-operator.v1.1.2
        version: 1.1.2
      - name: pubsubplus-eventbroker-operator.v1.1.1
        version: 1.1.1
      - name: pubsubplus-eventbroker-operator.v1.1.0
        version: 1.1.0
      - name: pubsubplus-eventbroker-operator.v1.0.3
        version: 1.0.3
      - name: pubsubplus-eventbroker-operator.v1.0.2
        version: 1.0.2
      - name: pubsubplus-eventbroker-operator.v1.0.1
        version: 1.0.1
      - name: pubsubplus-eventbroker-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pubsubplus-eventbroker-operator
    provider:
      name: Solace Corporation
      url: www.solace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: ibm-data-reporter-operator
      app.kubernetes.io/managed-by: ibm-data-reporter-operator
      app.kubernetes.io/name: ibm-data-reporter-operator
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: ""
    name: ibm-data-reporter-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-data-reporter-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "DataReporterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-data-reporter-operator",
                    "app.kubernetes.io/instance": "datareporterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datareporterconfig",
                    "app.kubernetes.io/part-of": "ibm-data-reporter-operator"
                  },
                  "name": "datareporterconfig"
                },
                "spec": {
                  "userConfig": [
                    {
                      "metadata": {
                        "ameta1": "ametadata1",
                        "bmeta1": "bmetadata1",
                        "cmeta1": "cmetadata1",
                        "dmeta1": "dmetadata1"
                      },
                      "userName": "system:serviceaccount:redhat-marketplace:ibm-data-reporter-operator-api"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
          createdAt: ""
          description: The IBM Data Reporter Operator accepts events and transforms
            them into reports submitted to the Data Service of the IBM Metrics Operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReporterConfig is the Schema for the datareporterconfigs
              API
            displayName: Data Reporter Config
            kind: DataReporterConfig
            name: datareporterconfigs.marketplace.redhat.com
            version: v1alpha1
          required:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
        description: "# IBM Data Reporter Operator\n\n# Introduction\n\nThe IBM Data
          Reporter Operator accepts events and transforms them into reports submitted
          to the IBM Metrics Operator Data Service.\n\n# Details\n\nThe IBM Data Reporter
          Operator deploys a service that provides an endpoint to which callers can
          send raw json event data. The event data is transformed into a report and
          is sent to the IBM Metrics Operator's Data Service. The IBM Metrics Operator
          Data Service periodically uploads the reports to IBM Software Central. It
          can also filter, transform and forward events to alternate endpoints.\n\n##
          Prerequisites\n\n- OpenShift Container Platform, major version 4 with any
          available supported minor version\n- Install IBM Metrics Operator\n  - IBM
          Data Reporter Operator prerequisties the IBM Metrics Operator's Data Service\n
          \ - Register the Cluster by creating a `redhat-marketplace-pull-secret`,
          as per the instructions\n  - Once `rhm-data-service` has started IBM Data
          Reporter Operator will accept requests\n\n## SecurityContextConstraints
          Requirements\n\n- The operator runs under Red Hat restricted SCC\n\n## Resources
          Required\n\n- The operator requires at least 85 Mi memory.\n\n## Limitations\n\n-
          Only runs on amd64, s390x and ppc64le architectures\n\n## Installing\n\n-
          A user with the cluster administrator role.\n- Install this operator in
          the same namespace as the IBM Metrics Operator\n  - default namespace: `ibm-software-central`\n\n##
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n# Configuration\n\nOptional:\n- Configure the
          Custom Resource Definition `DataReporterConfig` named `datareporterconfig`
          as per the following sample\n  - Configuration that can not be reconciled
          successfully will be reported on the DataReporterConfig Status\n  - jsonPath
          expressions are handled by: https://github.com/ohler55/ojg\n    - jsonPath
          comparison: https://cburgmer.github.io/json-path-comparison/\n  - Transformer
          type currently supported is `kazaam`\n    - https://github.com/willie68/kazaam\n\nSample
          DataReporterConfig:\n```YAML\napiVersion: marketplace.redhat.com/v1alpha1\nkind:
          DataReporterConfig\nmetadata:\n  name: datareporterconfig\nspec:\n  # true:
          report generated for 1 event and provide return code from data-service,
          confirming delivery\n  # false: events will be accumulated before building
          report, and immediately return 200, delivery unfconfirmed\n  confirmDelivery:
          false\n  dataFilters:   # dataFilters matches the first dataFilter selector,
          applies transforms and forwards to data-service and destinations\n  - altDestinations:
          \  # List of desintations to transform and foward events to\n    - authorization:
          \  # optional endpoint if an authorization token must first be requested
          \n        authDestHeader: Authorization   # http header key to be appended
          to destination request\n        authDestHeaderPrefix: 'Bearer '   # optional
          prefix for constructing destination request http header value\n        bodyData:
          \  # body data to send to authorization request, often an apiKey\n          secretKeyRef:\n
          \           name: auth-body-data-secret\n            key: bodydata\n        header:
          \  # secret map to use for authorization request headers\n          secret:\n
          \           name: auth-header-map-secret\n        tokenExpr: $.token   #
          optionally extract the returned body data using a jsonPath expression\n
          \       url: https://127.0.0.1:8443/api/2.0/accounts/my_account_id/apikeys/token
          \  # the url to make an authorization request to\n      header:   # secret
          map to use for destination request headers\n        secret:\n          name:
          dest-header-map-secret\n      transformer:   # The transformer to apply
          to event data for this destination\n        configMapKeyRef:\n          key:
          kazaam.json\n          name: kazaam-configmap\n        type: kazaam\n      url:
          https://127.0.0.1:8443/metering/api/1.0/usage   # the url to send the event
          to\n      urlSuffixExpr: $.properties.productId   # optional jsonPath expression
          on the event to extract and set a suffix to the url path\n    manifestType:
          dataReporter   # override manifestType of report sent to data-service\n
          \   selector:   # Matches against event data and a user. Empty selector
          matches all.\n      matchExpressions:   # jsonPath expression must return
          a result for all expressions to match (AND)\n      - $[?($.event == \"Account
          Contractual Usage\")]\n      - $.properties.productId\n      - $[?($.properties.source
          != null)]\n      - $[?($.properties.unit == \"AppPoints\")]\n      - $[?($.properties.quantity
          >= 0)]\n      - $[?($.timestamp != null)]\n      matchUsers:   # must match
          one of these users (OR). Omitting users matches any user\n      - system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n
          \   transformer:   # The transformer to apply to the event for data-service\n
          \     configMapKeyRef:\n        key: kazaam.json\n        name: kazaam-configmap\n
          \     type: kazaam\n  tlsConfig:   # TLS configuration for requests outbound
          to destinations\n    caCerts:\n    - name: tls-ca-certificates\n      key:
          ca.crt\n    certificates:\n    - clientCert:\n        secretKeyRef:\n          name:
          tls-ca-certificates\n          key: tls.crt\n      clientKey:\n        secretKeyRef:\n
          \         name: tls-ca-certificates\n          key: tls.key\n    cipherSuites:
          \n    - TLS_AES_128_GCM_SHA256\n    - TLS_AES_256_GCM_SHA384\n    - TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\n    - TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\n    insecureSkipVerify: false\n
          \   minVersion: VersionTLS12\n  userConfig:   # Optional metadata to apply
          to report sent to data-service for a specific user\n  - metadata:\n      ameta1:
          ametadata1\n      bmeta1: bmetadata1\n      cmeta1: cmetadata1\n      dmeta1:
          dmetadata1\n    userName: system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n```\n\nSample
          Header Secrets:\n```YAML\napiVersion: v1\nstringData:\n  accept: | \n    */*\nkind:
          Secret\nmetadata:\n  name: dest-header-map-secret\ntype: Opaque\n---\napiVersion:
          v1\nstringData:\n  accept: application/json\n  Content-Type: application/json\nkind:
          Secret\nmetadata:\n  name: auth-header-map-secret\ntype: Opaque\n```\n\nSample
          Authentication Body Secret:\n```YAML\napiVersion: v1\nstringData:\n  bodydata:
          |\n    {\"apikey\": \"myapikey\"}\nkind: Secret\nmetadata:\n  name: auth-body-data-secret\ntype:
          Opaque\n```\n\nSample TLS Secret:\n```YAML\napiVersion: v1\ndata:\n  ca.crt:
          LS0t...\n  tls.key: LS0t...\n  tls.crt: LS0t...\nkind: Secret\nmetadata:\n
          \ name: tls-ca-certificates\n```\n\nSample Kazaam ConfigMap:\n```YAML\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: kazaam-configmap\ndata:\n  kazaam.json:
          |\n    [\n        {\n            \"operation\": \"timestamp\",\n            \"spec\":
          {\n                \"timestamp\": {\n                \"inputFormat\": \"2006-01-02T15:04:05.999999+00:00\",\n
          \               \"outputFormat\": \"$unixmilli\"\n                }\n            }\n
          \       },\n        {\n            \"operation\": \"shift\",\n            \"spec\":
          {\n                \"instances[0].instanceId\": \"properties.source\",\n
          \               \"instances[0].startTime\": \"timestamp\",\n                \"instances[0].endTime\":
          \"timestamp\",\n                \"instances[0].metricUsage[0].metricId\":
          \"properties.unit\",\n                \"instances[0].metricUsage[0].quantity\":
          \"properties.quantity\"\n            }\n        },\n        {\n            \"operation\":
          \"default\",\n            \"spec\": {\n                \"meteringModel\":
          \"point-in-time\",\n                \"meteringPlan\": \"contract\"\n            }\n
          \       }\n    ]\n```\n\n### API Service User Configuration\n\n- The ClusterRole
          for api access is `clusterrole/ibm-data-reporter-operator-api`\n- The default
          ServiceAccount provided as an api user is `system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api`\n
          \ - The default ClusterRoleBinding for this user is `clusterrolebinding/ibm-data-reporter-operator-api`\n\nOptional:\n\n-
          To create an additional ServiceAccount and ClusterRoleBinding for api access\n\n```SHELL\nNAMESPACE=$(oc
          config view --minify -o jsonpath='{..namespace}') && \\\noc create serviceaccount
          my-api-service-account && \\\noc create clusterrolebinding ibm-data-reporter-operator-api
          --clusterrole=ibm-data-reporter-operator-api --serviceaccount=${NAMESPACE}:my-api-service-account\n```\n\n-
          Update datareporterconfig to attach metadata to reports associated with
          this user\n\n## API Service Usage\n\n- Get Token & Host\n\n```SHELL\noc
          project ibm-software-central\nexport DRTOKEN=$(oc create token ibm-data-reporter-operator-api
          --namespace ibm-software-central --duration 1h)\nexport DRHOST=$(oc get
          route ibm-data-reporter --template='{{ .spec.host }}')\n```\n\n- Get the
          Status\n\n```SHELL\ncurl -k -H \"Authorization: Bearer ${DRTOKEN}\" https://${DRHOST}/v1/status
          \n```\n\n- Post an Event\n\n```SHELL\ncurl -k -H \"Authorization: Bearer
          ${DRTOKEN}\" -X POST -d '{\"event\":\"myevent\"}' https://${DRHOST}/v1/event\n```\n\n##
          Storage\n\n- The operator temporarily stores events in pod memory, and writes
          to the IBM Metrics Operator data-service, which requires a PersistentVolume\n\n##
          License\n\nCopyright IBM Corporation 2023. All Rights Reserved.\n\nLicensed
          under the Apache License, Version 2.0 (the \"License\");\nyou may not use
          this file except in compliance with the License.\nYou may obtain a copy
          of the License at\n\n    http://www.apache.org/licenses/LICENSE-2.0\n\nUnless
          required by applicable law or agreed to in writing, software\ndistributed
          under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES
          OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for
          the specific language governing permissions and\nlimitations under the License.\n"
        displayName: IBM Data Reporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 2.20.0
      entries:
      - name: ibm-data-reporter-operator.v2.20.0
        version: 2.20.0
      name: beta
    - currentCSV: ibm-data-reporter-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "DataReporterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-data-reporter-operator",
                    "app.kubernetes.io/instance": "datareporterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datareporterconfig",
                    "app.kubernetes.io/part-of": "ibm-data-reporter-operator"
                  },
                  "name": "datareporterconfig"
                },
                "spec": {
                  "userConfig": [
                    {
                      "metadata": {
                        "ameta1": "ametadata1",
                        "bmeta1": "bmetadata1",
                        "cmeta1": "cmetadata1",
                        "dmeta1": "dmetadata1"
                      },
                      "userName": "system:serviceaccount:redhat-marketplace:ibm-data-reporter-operator-api"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
          createdAt: ""
          description: The IBM Data Reporter Operator accepts events and transforms
            them into reports submitted to the Data Service of the IBM Metrics Operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReporterConfig is the Schema for the datareporterconfigs
              API
            displayName: Data Reporter Config
            kind: DataReporterConfig
            name: datareporterconfigs.marketplace.redhat.com
            version: v1alpha1
          required:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
        description: "# IBM Data Reporter Operator\n\n# Introduction\n\nThe IBM Data
          Reporter Operator accepts events and transforms them into reports submitted
          to the IBM Metrics Operator Data Service.\n\n# Details\n\nThe IBM Data Reporter
          Operator deploys a service that provides an endpoint to which callers can
          send raw json event data. The event data is transformed into a report and
          is sent to the IBM Metrics Operator's Data Service. The IBM Metrics Operator
          Data Service periodically uploads the reports to IBM Software Central. It
          can also filter, transform and forward events to alternate endpoints.\n\n##
          Prerequisites\n\n- OpenShift Container Platform, major version 4 with any
          available supported minor version\n- Install IBM Metrics Operator\n  - IBM
          Data Reporter Operator prerequisties the IBM Metrics Operator's Data Service\n
          \ - Register the Cluster by creating a `redhat-marketplace-pull-secret`,
          as per the instructions\n  - Once `rhm-data-service` has started IBM Data
          Reporter Operator will accept requests\n\n## SecurityContextConstraints
          Requirements\n\n- The operator runs under Red Hat restricted SCC\n\n## Resources
          Required\n\n- The operator requires at least 85 Mi memory.\n\n## Limitations\n\n-
          Only runs on amd64, s390x and ppc64le architectures\n\n## Installing\n\n-
          A user with the cluster administrator role.\n- Install this operator in
          the same namespace as the IBM Metrics Operator\n  - default namespace: `ibm-software-central`\n\n##
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n# Configuration\n\nOptional:\n- Configure the
          Custom Resource Definition `DataReporterConfig` named `datareporterconfig`
          as per the following sample\n  - Configuration that can not be reconciled
          successfully will be reported on the DataReporterConfig Status\n  - jsonPath
          expressions are handled by: https://github.com/ohler55/ojg\n    - jsonPath
          comparison: https://cburgmer.github.io/json-path-comparison/\n  - Transformer
          type currently supported is `kazaam`\n    - https://github.com/willie68/kazaam\n\nSample
          DataReporterConfig:\n```YAML\napiVersion: marketplace.redhat.com/v1alpha1\nkind:
          DataReporterConfig\nmetadata:\n  name: datareporterconfig\nspec:\n  # true:
          report generated for 1 event and provide return code from data-service,
          confirming delivery\n  # false: events will be accumulated before building
          report, and immediately return 200, delivery unfconfirmed\n  confirmDelivery:
          false\n  dataFilters:   # dataFilters matches the first dataFilter selector,
          applies transforms and forwards to data-service and destinations\n  - altDestinations:
          \  # List of desintations to transform and foward events to\n    - authorization:
          \  # optional endpoint if an authorization token must first be requested
          \n        authDestHeader: Authorization   # http header key to be appended
          to destination request\n        authDestHeaderPrefix: 'Bearer '   # optional
          prefix for constructing destination request http header value\n        bodyData:
          \  # body data to send to authorization request, often an apiKey\n          secretKeyRef:\n
          \           name: auth-body-data-secret\n            key: bodydata\n        header:
          \  # secret map to use for authorization request headers\n          secret:\n
          \           name: auth-header-map-secret\n        tokenExpr: $.token   #
          optionally extract the returned body data using a jsonPath expression\n
          \       url: https://127.0.0.1:8443/api/2.0/accounts/my_account_id/apikeys/token
          \  # the url to make an authorization request to\n      header:   # secret
          map to use for destination request headers\n        secret:\n          name:
          dest-header-map-secret\n      transformer:   # The transformer to apply
          to event data for this destination\n        configMapKeyRef:\n          key:
          kazaam.json\n          name: kazaam-configmap\n        type: kazaam\n      url:
          https://127.0.0.1:8443/metering/api/1.0/usage   # the url to send the event
          to\n      urlSuffixExpr: $.properties.productId   # optional jsonPath expression
          on the event to extract and set a suffix to the url path\n    manifestType:
          dataReporter   # override manifestType of report sent to data-service\n
          \   selector:   # Matches against event data and a user. Empty selector
          matches all.\n      matchExpressions:   # jsonPath expression must return
          a result for all expressions to match (AND)\n      - $[?($.event == \"Account
          Contractual Usage\")]\n      - $.properties.productId\n      - $[?($.properties.source
          != null)]\n      - $[?($.properties.unit == \"AppPoints\")]\n      - $[?($.properties.quantity
          >= 0)]\n      - $[?($.timestamp != null)]\n      matchUsers:   # must match
          one of these users (OR). Omitting users matches any user\n      - system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n
          \   transformer:   # The transformer to apply to the event for data-service\n
          \     configMapKeyRef:\n        key: kazaam.json\n        name: kazaam-configmap\n
          \     type: kazaam\n  tlsConfig:   # TLS configuration for requests outbound
          to destinations\n    caCerts:\n    - name: tls-ca-certificates\n      key:
          ca.crt\n    certificates:\n    - clientCert:\n        secretKeyRef:\n          name:
          tls-ca-certificates\n          key: tls.crt\n      clientKey:\n        secretKeyRef:\n
          \         name: tls-ca-certificates\n          key: tls.key\n    cipherSuites:
          \n    - TLS_AES_128_GCM_SHA256\n    - TLS_AES_256_GCM_SHA384\n    - TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\n    - TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\n    insecureSkipVerify: false\n
          \   minVersion: VersionTLS12\n  userConfig:   # Optional metadata to apply
          to report sent to data-service for a specific user\n  - metadata:\n      ameta1:
          ametadata1\n      bmeta1: bmetadata1\n      cmeta1: cmetadata1\n      dmeta1:
          dmetadata1\n    userName: system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n```\n\nSample
          Header Secrets:\n```YAML\napiVersion: v1\nstringData:\n  accept: | \n    */*\nkind:
          Secret\nmetadata:\n  name: dest-header-map-secret\ntype: Opaque\n---\napiVersion:
          v1\nstringData:\n  accept: application/json\n  Content-Type: application/json\nkind:
          Secret\nmetadata:\n  name: auth-header-map-secret\ntype: Opaque\n```\n\nSample
          Authentication Body Secret:\n```YAML\napiVersion: v1\nstringData:\n  bodydata:
          |\n    {\"apikey\": \"myapikey\"}\nkind: Secret\nmetadata:\n  name: auth-body-data-secret\ntype:
          Opaque\n```\n\nSample TLS Secret:\n```YAML\napiVersion: v1\ndata:\n  ca.crt:
          LS0t...\n  tls.key: LS0t...\n  tls.crt: LS0t...\nkind: Secret\nmetadata:\n
          \ name: tls-ca-certificates\n```\n\nSample Kazaam ConfigMap:\n```YAML\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: kazaam-configmap\ndata:\n  kazaam.json:
          |\n    [\n        {\n            \"operation\": \"timestamp\",\n            \"spec\":
          {\n                \"timestamp\": {\n                \"inputFormat\": \"2006-01-02T15:04:05.999999+00:00\",\n
          \               \"outputFormat\": \"$unixmilli\"\n                }\n            }\n
          \       },\n        {\n            \"operation\": \"shift\",\n            \"spec\":
          {\n                \"instances[0].instanceId\": \"properties.source\",\n
          \               \"instances[0].startTime\": \"timestamp\",\n                \"instances[0].endTime\":
          \"timestamp\",\n                \"instances[0].metricUsage[0].metricId\":
          \"properties.unit\",\n                \"instances[0].metricUsage[0].quantity\":
          \"properties.quantity\"\n            }\n        },\n        {\n            \"operation\":
          \"default\",\n            \"spec\": {\n                \"meteringModel\":
          \"point-in-time\",\n                \"meteringPlan\": \"contract\"\n            }\n
          \       }\n    ]\n```\n\n### API Service User Configuration\n\n- The ClusterRole
          for api access is `clusterrole/ibm-data-reporter-operator-api`\n- The default
          ServiceAccount provided as an api user is `system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api`\n
          \ - The default ClusterRoleBinding for this user is `clusterrolebinding/ibm-data-reporter-operator-api`\n\nOptional:\n\n-
          To create an additional ServiceAccount and ClusterRoleBinding for api access\n\n```SHELL\nNAMESPACE=$(oc
          config view --minify -o jsonpath='{..namespace}') && \\\noc create serviceaccount
          my-api-service-account && \\\noc create clusterrolebinding ibm-data-reporter-operator-api
          --clusterrole=ibm-data-reporter-operator-api --serviceaccount=${NAMESPACE}:my-api-service-account\n```\n\n-
          Update datareporterconfig to attach metadata to reports associated with
          this user\n\n## API Service Usage\n\n- Get Token & Host\n\n```SHELL\noc
          project ibm-software-central\nexport DRTOKEN=$(oc create token ibm-data-reporter-operator-api
          --namespace ibm-software-central --duration 1h)\nexport DRHOST=$(oc get
          route ibm-data-reporter --template='{{ .spec.host }}')\n```\n\n- Get the
          Status\n\n```SHELL\ncurl -k -H \"Authorization: Bearer ${DRTOKEN}\" https://${DRHOST}/v1/status
          \n```\n\n- Post an Event\n\n```SHELL\ncurl -k -H \"Authorization: Bearer
          ${DRTOKEN}\" -X POST -d '{\"event\":\"myevent\"}' https://${DRHOST}/v1/event\n```\n\n##
          Storage\n\n- The operator temporarily stores events in pod memory, and writes
          to the IBM Metrics Operator data-service, which requires a PersistentVolume\n\n##
          License\n\nCopyright IBM Corporation 2023. All Rights Reserved.\n\nLicensed
          under the Apache License, Version 2.0 (the \"License\");\nyou may not use
          this file except in compliance with the License.\nYou may obtain a copy
          of the License at\n\n    http://www.apache.org/licenses/LICENSE-2.0\n\nUnless
          required by applicable law or agreed to in writing, software\ndistributed
          under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES
          OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for
          the specific language governing permissions and\nlimitations under the License.\n"
        displayName: IBM Data Reporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 2.20.0
      entries:
      - name: ibm-data-reporter-operator.v2.20.0
        version: 2.20.0
      name: stable
    defaultChannel: stable
    packageName: ibm-data-reporter-operator
    provider:
      name: Red Hat Marketplace
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ssm-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ssm-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ssm-controller:1.1.1
          createdAt: "2025-09-19T20:57:07Z"
          description: AWS SSM controller is a service controller for managing SSM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Document represents the state of an AWS ssm Document resource.
            displayName: Document
            kind: Document
            name: documents.ssm.services.k8s.aws
            version: v1alpha1
          - description: PatchBaseline represents the state of an AWS ssm PatchBaseline
              resource.
            displayName: PatchBaseline
            kind: PatchBaseline
            name: patchbaselines.ssm.services.k8s.aws
            version: v1alpha1
          - description: ResourceDataSync represents the state of an AWS ssm ResourceDataSync
              resource.
            displayName: ResourceDataSync
            kind: ResourceDataSync
            name: resourcedatasyncs.ssm.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon SSM resources in AWS from within your Kubernetes cluster.

          **About Amazon SSM**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon SSM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ssm
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SSM Developer Resources
          url: https://aws.amazon.com/SSM/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ssm maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ssm-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ssm-controller.v1.1.1
        version: 1.1.1
      - name: ack-ssm-controller.v1.1.0
        version: 1.1.0
      - name: ack-ssm-controller.v1.0.12
        version: 1.0.12
      - name: ack-ssm-controller.v1.0.10
        version: 1.0.10
      - name: ack-ssm-controller.v1.0.9
        version: 1.0.9
      - name: ack-ssm-controller.v1.0.8
        version: 1.0.8
      - name: ack-ssm-controller.v1.0.7
        version: 1.0.7
      - name: ack-ssm-controller.v1.0.6
        version: 1.0.6
      - name: ack-ssm-controller.v1.0.5
        version: 1.0.5
      - name: ack-ssm-controller.v1.0.4
        version: 1.0.4
      - name: ack-ssm-controller.v1.0.3
        version: 1.0.3
      - name: ack-ssm-controller.v1.0.2
        version: 1.0.2
      - name: ack-ssm-controller.v1.0.1
        version: 1.0.1
      - name: ack-ssm-controller.v1.0.0
        version: 1.0.0
      - name: ack-ssm-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-ssm-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: authorino-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: authorino-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta3",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/authorino-rhel9-operator@sha256:2d81613856df9725ad20928a7a9895d14482c42effce0a3a0de88f40ec35d411
          createdAt: "2025-03-25T22:39:23Z"
          description: Enables authentication and authorization for Gateways and applications
            in a Gateway API network.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta3
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: |-
          [Authorino](https://docs.kuadrant.io/authorino/) is Red Hat's Kubernetes-native lightweight external authorization service for tailor-made Zero Trust API security.

          Install this Red Hat official distribution of [Authorino Operator](https://docs.kuadrant.io/authorino-operator/) to manage instances of Authorino in this cluster.

          The current state of this distribution of the operator is: **Generally Available**.

          The Community version of Authorino Operator, based on upstream public images, is available in [OperatorHub.io](https://operatorhub.io/operator/authorino-operator).

          ## Getting started

          After installing the Operator, request an instance of the external authorization service by creating an `Authorino` custom resource.

          **Minimal example (namespaced)**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            listener:
              tls:
                enabled: false
            oidcServer:
              tls:
                enabled: false
          ```

          **Extended example**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            clusterWide: true
            authConfigLabelSelectors: environment=production
            secretLabelSelectors: authorino.kuadrant.io/component=authorino,environment=production

            replicas: 2

            evaluatorCacheSize: 2 # mb

            logLevel: info
            logMode: production

            listener:
              ports:
                grpc: 50001
                http: 5001
              tls:
                certSecretRef:
                  name: authorino-server-cert # secret must contain `tls.crt` and `tls.key` entries
              timeout: 2

            oidcServer:
              port: 8083
              tls:
                certSecretRef:
                  name: authorino-oidc-server-cert # secret must contain `tls.crt` and `tls.key` entries

            metrics:
              port: 8080
              deep: true

            healthz:
              port: 8081

            tracing:
              endpoint: rpc://otel-collector.observability.svc.cluster.local:4317
              insecure: true

            volumes:
              items:
                - name: keycloak-tls-cert
                  mountPath: /etc/ssl/certs
                  configMaps:
                    - keycloak-tls-cert
                  items: # details to mount the k8s configmap in the authorino pods
                    - key: keycloak.crt
                      path: keycloak.crt
              defaultMode: 420
          ```

          ### Cluster-wide vs Namespaced

          Namespaced instances only watch auth resources (`AuthConfig` and `Secrets`) created in the same namespace as the Authorino service. Use this mode for dedicated instances that do not require elevated privileges.

          Cluster-wide instances watch resources across the entire cluster (all namespaces.) Deploying and running Authorino in this mode requires elevated privileges.

          ### Multi-tenancy

          Use the `authConfigLabelSelectors` field of the `Authorino` custom resource to narrow the scope of the Authorino instance.

          Only `AuthConfig` custom resources whose labels match the value of this field will be handled by the Authorino instance.

          ## Protect a host

          To protect a host, create an `AuthConfig` custom resource for the host. E.g.:

          ```yaml
          apiVersion: authorino.kuadrant.io/v1beta3
          kind: AuthConfig
          metadata:
            name: my-api-protection
          spec:
            hosts:
            - my-api.io

            authentication:
              "keycloak":
                jwt:
                  issuerUrl: https://keycloak.keycloak.svc.cluster.local:8080/realms/my-realm

            authorization:
              "k8s-rbac":
                kubernetesSubjectAccessReview:
                  user:
                    selector: auth.identity.user.username
                  resourceAttributes:
                    resource:
                      value: my-api
                    verb:
                      selector: request.method
                cache:
                  key:
                    selector: auth.identity.user.username
                  ttl: 30
              "after-2am-only":
                rego: |
                  allow {
                    [hour, _, _] := time.clock(time.now_ns())
                    hour >= 2
                  }
          ```

          Make sure all requests to the host are fisrt checked with the Authorino instance, by configuring an Envoy proxy for external authz:

          ```yaml
          clusters:
          - name: my-api
            …
          - name: authorino
            connect_timeout: 0.25s
            type: STRICT_DNS
            lb_policy: ROUND_ROBIN
            http2_protocol_options: {}
            load_assignment:
              cluster_name: authorino
              endpoints:
              - lb_endpoints:
                - endpoint:
                    address:
                      socket_address:
                        address: authorino-authorino-authorization
                        port_value: 50051
          listeners:
          - filter_chains:
            - filters:
              name: envoy.http_connection_manager
              typed_config:
                "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
                use_remote_address: true
                route_config:
                  name: my-api-config
                  virtual_hosts:
                  - name: my-api-vs
                    domains:
                    - my-api.io
                    routes:
                    - match:
                        prefix: /
                      route:
                        cluster: my-api
                http_filters:
                - name: envoy.filters.http.ext_authz
                  typed_config:
                    "@type": type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz
                    transport_api_version: V3
                    failure_mode_allow: false
                    include_peer_certificate: true
                    grpc_service:
                      envoy_grpc:
                        cluster_name: authorino
                      timeout: 1s
          ```

          ...or, if using Istio, by creating an [`AuthorizationPolicy`](https://istio.io/latest/docs/reference/config/security/authorization-policy/#AuthorizationPolicy) custom resource. Use `action: CUSTOM` in the resource and the Authorino authorization service configured in the mesh extension provider settings.

          ## Features

          **Authentication**

          * JWT validation (with OpenID Connect Discovery)
          * OAuth 2.0 Token Introspection (opaque tokens)
          * Kubernetes TokenReview (ServiceAccount tokens)
          * API key authentication
          * X.509 client certificate authentication
          * Anonymous access
          * Proxy-handled (authentication performed by the proxy)

          **Authorization**

          * Built-in simple pattern matching (e.g. JWT claims, request attributes checking)
          * OPA policies (inline Rego and fetch from external registry)
          * Kubernetes SubjectAccessReview (resource and non-resource attributes)
          * Authzed SpiceDB

          **External metadata**

          * HTTP request
          * OpenID Connect User Info
          * UMA-protected resource attributes

          **Custom responses**

          * Header injection (Festival Wristbands tokens, JSON, plain text)
          * Envoy Dynamic Metadata
          * Custom HTTP response (status code, headers, messages, body, etc)

          **Callbacks**

          * HTTP webhooks

          **Caching**

          * OpenID Connect and User-Managed Access configs
          * JSON Web Keys (JWKs) and JSON Web Key Sets (JWKS)
          * Access tokens
          * External metadata
          * Precompiled Rego policies
          * Policy evaluation

          Check out the full [Feature Specification](https://docs.kuadrant.io/authorino/docs/features/) and how-to guides in the [Kuadrant Docs](https://docs.kuadrant.io) website.
        displayName: Red Hat - Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: GitHub
          url: https://github.com/kuadrant/authorino-operator
        - name: Documentation
          url: https://docs.kuadrant.io/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhcl-1/authorino-rhel9-operator@sha256:2d81613856df9725ad20928a7a9895d14482c42effce0a3a0de88f40ec35d411
        - registry.redhat.io/rhcl-1/authorino-rhel9@sha256:413d3796ad0101bf14d3341d31d9367e67fea1e73ebb3fb79124ef4c5dde0c09
        version: 1.2.2
      entries:
      - name: authorino-operator.v1.2.2
        version: 1.2.2
      - name: authorino-operator.v1.2.1
        version: 1.2.1
      - name: authorino-operator.v1.2.0
        version: 1.2.0
      - name: authorino-operator.v1.1.3
        version: 1.1.3
      - name: authorino-operator.v1.1.2
        version: 1.1.2
      - name: authorino-operator.v1.1.1
        version: 1.1.1
      - name: authorino-operator.v1.1.0
        version: 1.1.0
      - name: authorino-operator.v1.0.2
        version: 1.0.2
      - name: authorino-operator.v0.16.1
        version: 0.16.1
      - name: authorino-operator.v0.16.0
        version: 0.16.0
      name: stable
    - currentCSV: authorino-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta1",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "hosts": [
                    "my-api.io"
                  ],
                  "identity": [
                    {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "in": "authorization_header",
                        "keySelector": "APIKEY"
                      },
                      "name": "api-key-users"
                    }
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/3scale-tech-preview/authorino-rhel9-operator@sha256:eb574781cf39917f8f5d9e5c806f520534601f819ee1ebcbb751832da8938008
          createdAt: "2024-11-08T11:06:33Z"
          description: Authorino
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta1
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: |-
          [Authorino](https://docs.kuadrant.io/authorino/) is Red Hat's Kubernetes-native lightweight external authorization service for tailor-made Zero Trust API security.

          Install this Red Hat official distribution of [Authorino Operator](https://docs.kuadrant.io/authorino-operator/) to manage instances of Authorino in this cluster.

          The current state of this distribution of the operator is: **_Technical Preview_**.

          The Community version of Authorino Operator, based on upstream public images, is available in [OperatorHub.io](https://operatorhub.io/operator/authorino-operator).

          ## Getting started

          After installing the Operator, request an instance of the external authorization service by creating an `Authorino` custom resource.

          **Minimal example (namespaced)**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            listener:
              tls:
                enabled: false
            oidcServer:
              tls:
                enabled: false
          ```

          **Extended example**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            clusterWide: true
            authConfigLabelSelectors: environment=production
            secretLabelSelectors: authorino.kuadrant.io/component=authorino,environment=production

            replicas: 2

            evaluatorCacheSize: 2 # mb

            logLevel: info
            logMode: production

            listener:
              ports:
                grpc: 50001
                http: 5001
              tls:
                certSecretRef:
                  name: authorino-server-cert # secret must contain `tls.crt` and `tls.key` entries
              timeout: 2

            oidcServer:
              port: 8083
              tls:
                certSecretRef:
                  name: authorino-oidc-server-cert # secret must contain `tls.crt` and `tls.key` entries

            metrics:
              port: 8080
              deep: true

            healthz:
              port: 8081

            tracing:
              endpoint: rpc://otel-collector.observability.svc.cluster.local:4317
              insecure: true

            volumes:
              items:
                - name: keycloak-tls-cert
                  mountPath: /etc/ssl/certs
                  configMaps:
                    - keycloak-tls-cert
                  items: # details to mount the k8s configmap in the authorino pods
                    - key: keycloak.crt
                      path: keycloak.crt
              defaultMode: 420
          ```

          ### Cluster-wide vs Namespaced

          Namespaced instances only watch auth resources (`AuthConfig` and `Secrets`) created in the same namespace as the Authorino service. Use this mode for dedicated instances that do not require elevated privileges.

          Cluster-wide instances watch resources across the entire cluster (all namespaces.) Deploying and running Authorino in this mode requires elevated privileges.

          ### Multi-tenancy

          Use the `authConfigLabelSelectors` field of the `Authorino` custom resource to narrow the scope of the Authorino instance.

          Only `AuthConfig` custom resources whose labels match the value of this field will be handled by the Authorino instance.

          ## Protect a host

          To protect a host, create an `AuthConfig` custom resource for the host. E.g.:

          ```yaml
          apiVersion: authorino.kuadrant.io/v1beta2
          kind: AuthConfig
          metadata:
            name: my-api-protection
          spec:
            hosts:
            - my-api.io

            authentication:
              "keycloak":
                jwt:
                  issuerUrl: https://keycloak.keycloak.svc.cluster.local:8080/realms/my-realm

            authorization:
              "k8s-rbac":
                kubernetesSubjectAccessReview:
                  user:
                    selector: auth.identity.user.username
                  resourceAttributes:
                    resource:
                      value: my-api
                    verb:
                      selector: request.method
                cache:
                  key:
                    selector: auth.identity.user.username
                  ttl: 30
              "after-2am-only":
                rego: |
                  allow {
                    [hour, _, _] := time.clock(time.now_ns())
                    hour >= 2
                  }
          ```

          Make sure all requests to the host are fisrt checked with the Authorino instance, by configuring an Envoy proxy for external authz:

          ```yaml
          clusters:
          - name: my-api
            …
          - name: authorino
            connect_timeout: 0.25s
            type: STRICT_DNS
            lb_policy: ROUND_ROBIN
            http2_protocol_options: {}
            load_assignment:
              cluster_name: authorino
              endpoints:
              - lb_endpoints:
                - endpoint:
                    address:
                      socket_address:
                        address: authorino-authorino-authorization
                        port_value: 50051
          listeners:
          - filter_chains:
            - filters:
              name: envoy.http_connection_manager
              typed_config:
                "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
                use_remote_address: true
                route_config:
                  name: my-api-config
                  virtual_hosts:
                  - name: my-api-vs
                    domains:
                    - my-api.io
                    routes:
                    - match:
                        prefix: /
                      route:
                        cluster: my-api
                http_filters:
                - name: envoy.filters.http.ext_authz
                  typed_config:
                    "@type": type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz
                    transport_api_version: V3
                    failure_mode_allow: false
                    include_peer_certificate: true
                    grpc_service:
                      envoy_grpc:
                        cluster_name: authorino
                      timeout: 1s
          ```

          ...or, if using Istio, by creating an [`AuthorizationPolicy`](https://istio.io/latest/docs/reference/config/security/authorization-policy/#AuthorizationPolicy) custom resource. Use `action: CUSTOM` in the resource and the Authorino authorization service configured in the mesh extension provider settings.

          ## Features

          **Authentication**

          * JWT validation (with OpenID Connect Discovery)
          * OAuth 2.0 Token Introspection (opaque tokens)
          * Kubernetes TokenReview (ServiceAccount tokens)
          * API key authentication
          * X.509 client certificate authentication
          * Anonymous access
          * Proxy-handled (authentication performed by the proxy)

          **Authorization**

          * Built-in simple pattern matching (e.g. JWT claims, request attributes checking)
          * OPA policies (inline Rego and fetch from external registry)
          * Kubernetes SubjectAccessReview (resource and non-resource attributes)
          * Authzed SpiceDB

          **External metadata**

          * HTTP request
          * OpenID Connect User Info
          * UMA-protected resource attributes

          **Custom responses**

          * Header injection (Festival Wristbands tokens, JSON, plain text)
          * Envoy Dynamic Metadata
          * Custom HTTP response (status code, headers, messages, body, etc)

          **Callbacks**

          * HTTP webhooks

          **Caching**

          * OpenID Connect and User-Managed Access configs
          * JSON Web Keys (JWKs) and JSON Web Key Sets (JWKS)
          * Access tokens
          * External metadata
          * Precompiled Rego policies
          * Policy evaluation

          Check out the full [Feature Specification](https://docs.kuadrant.io/authorino/docs/features/) and how-to guides in the [Kuadrant Docs](https://docs.kuadrant.io) website.
        displayName: Red Hat - Authorino (Technical Preview)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: GitHub
          url: https://github.com/kuadrant/authorino-operator
        - name: Documentation
          url: https://docs.kuadrant.io/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-tech-preview/authorino-rhel9@sha256:a473dae20e71e3e813ac30ba978f2ab3c5e19d7d39b501ae9103dca892107c87
        - registry.redhat.io/3scale-tech-preview/authorino-rhel9-operator@sha256:eb574781cf39917f8f5d9e5c806f520534601f819ee1ebcbb751832da8938008
        - registry.redhat.io/openshift4/ose-cli-rhel9@sha256:3a398158d10018bfac5df291d6998b513cf9a30e5477bcaefddcf66d229627c7
        version: 1.1.3
      entries:
      - name: authorino-operator.v1.1.3
        version: 1.1.3
      - name: authorino-operator.v1.1.2
        version: 1.1.2
      - name: authorino-operator.v1.1.1
        version: 1.1.1
      - name: authorino-operator.v1.1.0
        version: 1.1.0
      - name: authorino-operator.v1.0.2
        version: 1.0.2
      name: tech-preview-v1
    defaultChannel: stable
    packageName: authorino-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-xtradb-cluster-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "docker.io/percona/percona-xtradb-cluster:8.0.42-33.1",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "docker.io/percona/haproxy:2.8.15",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "docker.io/percona/proxysql2:2.7.3",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "docker.io/percona/fluentbit:4.0.1",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:3.3.1",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "docker.io/percona/percona-xtrabackup:8.0.35-34.1",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.io/percona/percona-xtradb-cluster-operator:1.18.0
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <v1.18.0
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            clustercheck: Y2x1c3RlcmNoZWNrcGFzc3dvcmQ=
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            clustercheck: Y2x1c3RlcmNoZWNrcGFzc3dvcmQ=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ```
          ###  Release Highlights
          * **PMM3 support**

          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**

          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**

          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**

          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.

        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - database
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - docker.io/percona/percona-xtradb-cluster-operator:1.18.0
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator.v1.17.0
        version: 1.17.0
      - name: percona-xtradb-cluster-operator.v1.16.1
        version: 1.16.1
      - name: percona-xtradb-cluster-operator.v1.16.0
        version: 1.16.0
      - name: percona-xtradb-cluster-operator.v1.15.1
        version: 1.15.1
      - name: percona-xtradb-cluster-operator.v1.15.0
        version: 1.15.0
      - name: percona-xtradb-cluster-operator.v1.9.0
        version: 1.9.0
      name: stable
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Flux project
      provider-url: ""
    name: flux
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.toolkit.fluxcd.io/v2beta1",
                "kind": "HelmRelease",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "5m",
                    "chart": {
                      "spec": {
                          "chart": "podinfo",
                          "version": "4.0.x",
                          "sourceRef": {
                            "kind": "HelmRepository",
                            "name": "podinfo",
                            "namespace": "flux-system"
                          },
                          "interval": "1m"
                      }
                    },
                    "values": {
                      "replicaCount": 2
                    }
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "GitRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "1m",
                    "ref": {
                      "branch": "master"
                    },
                    "url": "https://github.com/stefanprodan/podinfo"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmRepository",
                "metadata": {
                    "name": "stable",
                    "namespace": "default"
                },
                "spec": {
                    "url": "https://kubernetes-charts.storage.googleapis.com/",
                    "interval": "10m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmChart",
                "metadata": {
                    "name": "redis",
                    "namespace": "default"
                },
                "spec": {
                    "chart": "redis",
                    "version": "10.5.7",
                    "sourceRef": {
                      "name": "stable",
                      "kind": "HelmRepository"
                    },
                    "interval": "5m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "Bucket",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "1m",
                    "provider": "generic",
                    "bucketName": "podinfo",
                    "endpoint": "minio.minio.svc.cluster.local:9000",
                    "insecure": true,
                    "secretRef": {
                      "name": "minio-credentials"
                    }
                }
              },
              {
                "apiVersion": "kustomize.toolkit.fluxcd.io/v1beta2",
                "kind": "Kustomization",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "5m",
                    "path": "./deploy/overlays/dev",
                    "prune": true,
                    "sourceRef": {
                      "kind": "GitRepository",
                      "name": "podinfo"
                    },
                    "timeout": "2m",
                    "wait": true
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "address": "https://github.com/stefanprodan/podinfo",
                    "secretRef": {
                      "name": "api-token"
                    }
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Alert",
                "metadata": {
                    "name": "webapp",
                    "namespace": "default"
                },
                "spec": {
                    "providerRef": {
                      "name": "on-call-slack"
                    },
                    "eventSeverity": "info",
                    "eventSources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "Bucket",
                          "name": "secrets"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-backend"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-frontend"
                      }
                    ]
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Receiver",
                "metadata": {
                    "name": "github-receiver",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "events": [
                      "ping",
                      "push"
                    ],
                    "secretRef": {
                      "name": "webhook-token"
                    },
                    "resources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "HelmRepository",
                          "name": "webapp"
                      }
                    ]
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "image": "ghcr.io/stefanprodan/podinfo",
                    "interval": "1m0s"
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImagePolicy",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "imageRepositoryRef": {
                      "name": "podinfo"
                    },
                    "policy": {
                      "semver": {
                          "range": "5.0.x"
                      }
                    }
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageUpdateAutomation",
                "metadata": {
                  "name": "flux-system",
                  "namespace": "flux-system"
                },
                "spec": {
                  "interval": "1m0s",
                  "sourceRef": {
                    "kind": "GitRepository",
                    "name": "flux-system"
                  },
                  "git": {
                    "checkout": {
                      "ref": {
                        "branch": "main"
                      }
                    },
                    "commit": {
                      "author": {
                        "email": "fluxcdbot@users.noreply.github.com",
                        "name": "fluxcdbot"
                      },
                      "messageTemplate": "{{range .Updated.Images}}{{println .}}{{end}}"
                    },
                    "push": {
                      "branch": "main"
                    }
                  },
                  "update": {
                    "path": "./clusters/my-cluster",
                    "strategy": "Setters"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/fluxcd/source-controller:v1.3.0
          description: Flux is a Continuous Delivery solution for Kubernetes.
          operatorframework.io/suggested-namespace: flux-system
          support: The Flux community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta3
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta1
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta2
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta1
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta2
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta2
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta2
          - description: OCIRepository
            displayName: OCIRepository
            kind: OCIRepository
            name: ocirepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta3
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta2
        description: |-
          [Flux](https://fluxcd.io/) is a Continuous Delivery solution for Kubernetes.
          Flux is constructed with the [GitOps Toolkit](https://toolkit.fluxcd.io/components/), a set of composable APIs and specialized tools for keeping Kubernetes clusters in sync with sources of configuration (like Git & Helm repositories), and automating updates to configuration when there is new code to deploy.

          Flux is a CNCF incubation project made for:

          - **cluster operators** who automate provision and configuration of clusters
          - **platform engineers** who build continuous delivery for developer teams
          - **app developers** who rely on continuous delivery to get their code live

          Features:

          - Source configuration from Git and Helm repositories, and S3-compatible buckets (e.g., Minio)
          - Kustomize and Helm support
          - Event-triggered and periodic reconciliation
          - Integration with Kubernetes RBAC
          - Health assessment (clusters and workloads)
          - Dependency management (infrastructure and workloads)
          - Secrets management (builtin support for Mozilla SOPS and Cloud KMS)
          - Alerting to external systems (webhook senders)
          - External events handling (webhook receivers)
          - Automated container image updates to Git (image scanning and patching)
          - Policy-driven validation (OPA, Kyverno, admission controllers)
          - Seamless integration with Git providers (GitHub, GitLab, Bitbucket)
          - Interoperability with Cluster API (CAPI) providers

          ### Steps to support bootstraping with Flux CLI

          SecComp Profile should be removed as it requires a special privilege to get the access.
          Also, it requires to enforce `runUserAs` to the same UID provided by the images to prevent OpenShift
          to alter the user id expected by our controllers.

          To do so, please specifiy a patch using the following Kustomization file, for example.

          ```yaml
          apiVersion: kustomize.config.k8s.io/v1beta1
          kind: Kustomization
          resources:
            - gotk-components.yaml
            - gotk-sync.yaml
          patches:
            - patch: |
                apiVersion: apps/v1
                kind: Deployment
                metadata:
                  name: all
                spec:
                  template:
                    spec:
                      containers:
                        - name: manager
                          securityContext:
                            runAsUser: 65534
                            seccompProfile:
                              $patch: delete
              target:
                kind: Deployment
                labelSelector: app.kubernetes.io/part-of=flux
          ```

          Then before installing Flux, please also perform the following steps to bind Flux to the `nonroot` SecurityContextConstraint (SCC). This applies to installation via OC's OLM or via Flux CLI.

          ```sh
          # replace default "flux-system" with the project/namespace name for your flux installation
          namespace="flux-system"
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:source-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:kustomize-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:helm-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:notification-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-automation-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-reflector-controller

          ```

          To see complete steps please refer to this [documentation](https://github.com/fluxcd/website/blob/main/content/en/docs/use-cases/openshift.md).

          If you've already installed Flux, you might need to uninstall it after applying the above pre-requisite steps.

          ### Network Policies

          This version of Flux is shipped without the normally provided network Policies.
          By default all traffic is enabled, and anyone can access the source-controller
          artifacts.

          It is not recommended to run it in production as is.

          Please visit [Flux's policies](https://github.com/fluxcd/flux2/tree/main/manifests/policies)
          to deploy the needed network policies and secure your Flux setup.

          ### Breaking Changes

          - Flux 0.18.x ships with v1beta2 of the kustomize.toolkit.fluxcd.io/Kustomization CRD.
          - Flux 0.25.x is the last Flux release where Kubernetes 1.19 is supported.

          ### Flux on OpenShift Use Cases

          - [The GitOps with Flux + OpenShift](https://www.youtube.com/watch?v=8ybCtNVP5To) talk by Andrew Block (Red Hat) showing how to setup multi-tenacy on OpenShift using Flux.
          - [Flux OpenShift Use Case](https://fluxcd.io/docs/use-cases/openshift/) on FluxCD.io
        displayName: Flux
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - gitops
        - cd
        - continuous delivery
        - fluxcd
        - helm
        - kustomize
        links:
        - name: Homepage
          url: https://fluxcd.io/
        - name: Documentation
          url: https://fluxcd.io/docs/
        - name: Slack
          url: https://cloud-native.slack.com/messages/flux
        - name: Mailing List
          url: https://lists.cncf.io/g/cncf-flux-dev
        - name: Source Code
          url: https://github.com/fluxcd/flux2
        maintainers:
        - email: support@weave.works
          name: The Flux maintainers
        maturity: stable
        provider:
          name: The Flux project
        relatedImages:
        - ghcr.io/fluxcd/kustomize-controller:v1.3.0
        - ghcr.io/fluxcd/notification-controller:v1.3.0
        - ghcr.io/fluxcd/source-controller:v1.3.0
        - ghcr.io/fluxcd/helm-controller:v1.0.1
        - ghcr.io/fluxcd/image-automation-controller:v0.38.0
        - ghcr.io/fluxcd/image-reflector-controller:v0.32.0
        version: 2.3.0
      entries:
      - name: flux.v2.3.0
        version: 2.3.0
      - name: flux.v2.2.2
        version: 2.2.2
      - name: flux.v2.2.0
        version: 2.2.0
      - name: flux.v2.0.1
        version: 2.0.1
      - name: flux.v0.41.2
        version: 0.41.2
      - name: flux.v0.40.0
        version: 0.40.0
      - name: flux.v0.39.0
        version: 0.39.0
      - name: flux.v0.38.3
        version: 0.38.3
      - name: flux.v0.37.0
        version: 0.37.0
      - name: flux.v0.36.0
        version: 0.36.0
      - name: flux.v0.35.0
        version: 0.35.0
      - name: flux.v0.31.4
        version: 0.31.4
      - name: flux.v0.31.3
        version: 0.31.3
      - name: flux.v0.30.2
        version: 0.30.2
      - name: flux.v0.28.5
        version: 0.28.5
      - name: flux.v0.28.4
        version: 0.28.4
      - name: flux.v0.28.3
        version: 0.28.3
      - name: flux.v0.27.4
        version: 0.27.4
      - name: flux.v0.27.3
        version: 0.27.3
      - name: flux.v0.27.2
        version: 0.27.2
      - name: flux.v0.27.0
        version: 0.27.0
      - name: flux.v0.26.1
        version: 0.26.1
      - name: flux.v0.26.0
        version: 0.26.0
      - name: flux.v0.25.3
        version: 0.25.3
      - name: flux.v0.25.2
        version: 0.25.2
      - name: flux.v0.24.1
        version: 0.24.1
      - name: flux.v0.24.0
        version: 0.24.0
      - name: flux.v0.23.0
        version: 0.23.0
      - name: flux.v0.22.1
        version: 0.22.1
      - name: flux.v0.22.0
        version: 0.22.0
      - name: flux.v0.21.1
        version: 0.21.1
      - name: flux.v0.21.0
        version: 0.21.0
      - name: flux.v0.20.1
        version: 0.20.1
      - name: flux.v0.20.0
        version: 0.20.0
      - name: flux.v0.19.1
        version: 0.19.1
      - name: flux.v0.18.3
        version: 0.18.3
      - name: flux.v0.18.1
        version: 0.18.1
      - name: flux.v0.17.2
        version: 0.17.2
      - name: flux.v0.17.1
        version: 0.17.1
      - name: flux.v0.17.0
        version: 0.17.0
      - name: flux.v0.16.2
        version: 0.16.2
      - name: flux.v0.16.1
        version: 0.16.1
      - name: flux.v0.16.0
        version: 0.16.0
      - name: flux.v0.15.3
        version: 0.15.3
      - name: flux.v0.15.2
        version: 0.15.2
      - name: flux.v0.14.2
        version: 0.14.2
      - name: flux.v0.14.1
        version: 0.14.1
      - name: flux.v0.14.0
        version: 0.14.0
      - name: flux.v0.13.4
        version: 0.13.4
      name: stable
    defaultChannel: stable
    packageName: flux
    provider:
      name: The Flux project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          olm.skipRange: <1.0.4
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
        version: 1.0.4
      entries:
      - name: datadog-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      name: stable
    defaultChannel: stable
    packageName: datadog-operator-certified
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kiali-ossm
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kiali-operator.v2.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "Kiali",
                "metadata": {
                  "name": "kiali"
                },
                "spec": {
                  "version": "default",
                  "istio_namespace": "istio-system",
                  "deployment": {
                    "logger": {
                      "log_level": "info"
                    },
                    "view_only_mode": false
                  }
                }
              },
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "OSSMConsole",
                "metadata": {
                  "name": "ossmconsole"
                },
                "spec": {
                  "version": "default",
                  "kiali": {
                    "serviceName": "",
                    "serviceNamespace": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9-operator@sha256:0612dfa52f2bf39a1d87e640409313585fedc7cfcff6d00e77bb14809d27d646
          createdAt: "2024-12-12T10:44:15Z"
          description: This productized operator provides Kiali and OSSMC. Kiali is
            the Istio observability and management Console. OSSMC is the OpenShift
            Service Mesh Console plugin, powered by Kiali.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.1.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kiali/kiali
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A configuration file for a Kiali installation.
            displayName: Kiali
            kind: Kiali
            name: kialis.kiali.io
            version: v1alpha1
          - description: A configuration file for a OpenShift Service Mesh Console
              installation.
            displayName: OpenShift Service Mesh Console
            kind: OSSMConsole
            name: ossmconsoles.kiali.io
            version: v1alpha1
        description: |-
          ## About the managed application

          A Microservice Architecture breaks up the monolith into many smaller pieces
          that are composed together. Patterns to secure the communication between
          services like fault tolerance (via timeout, retry, circuit breaking, etc.)
          have come up as well as distributed tracing to be able to see where calls
          are going.

          A service mesh can now provide these services on a platform level and frees
          the application writers from those tasks. Routing decisions are done at the
          mesh level.

          Kiali works with OpenShift Service Mesh to visualize the service
          mesh topology, to provide visibility into features like circuit breakers,
          request rates and more. It offers insights about the mesh components at
          different levels, from abstract Applications to Services and Workloads.

          See [https://www.kiali.io](https://www.kiali.io) to read more.

          ### Accessing the UI

          By default, the Kiali operator exposes the Kiali UI as an OpenShift Route.

          If on OpenShift, you can create an OSSMConsole CR to have the operator
          install the OpenShift ServiceMesh Console plugin to the OpenShift Console
          thus providing an interface directly integrated with the OpenShift Console.

          ## About this Operator

          ### Kiali Custom Resource Configuration Settings

          For quick descriptions of all the settings you can configure in the Kiali
          Custom Resource (CR), see
          [the kiali.io docs](https://kiali.io/docs/configuration/kialis.kiali.io/).

          ## Prerequisites for enabling this Operator

          Kiali is a companion tool for OpenShift Service Mesh. So before you install Kiali, you must have
          already installed Service Mesh.
        displayName: Kiali Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - service-mesh
        - observability
        - monitoring
        - istio
        links:
        - name: Installation Guide
          url: https://www.kiali.io/docs/installation
        - name: Features
          url: https://www.kiali.io/docs/features
        - name: Documentation Home
          url: https://www.kiali.io
        - name: Blogs and Articles
          url: https://medium.com/kialiproject
        - name: Server and UI Source Code
          url: https://github.com/kiali/kiali
        - name: Operator Source Code
          url: https://github.com/kiali/kiali-operator
        - name: OpenShift Service Mesh Console Code
          url: https://github.com/kiali/openshift-servicemesh-plugin
        maintainers:
        - email: kiali-dev@googlegroups.com
          name: Kiali Developers Google Group
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel8@sha256:e37d76dd7c7a2db79ff1e96afef19b07c3beaa8ebf0c7bd85b808908b8f0ce77
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9-operator@sha256:0612dfa52f2bf39a1d87e640409313585fedc7cfcff6d00e77bb14809d27d646
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9@sha256:06fde6c696a33d17b0d2f7620ec46422c82277fd17a3fe6d7319de05c732aace
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:976a31ca8cf17ac0887b8a586edfcc29eb30175802461b3114758af525e7b800
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:66e4d933a9d69158bf87dd45835183bf2308b77c48908ab64496c5d22c816293
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:27a6aebffa3d049ee48050126e6107db2e07c49e9ce9297175facb6368e1e792
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-ossmc-rhel9@sha256:99643f294c3f92a6fc82f321cab7c756e4879fefb48fa824f1bf49ff7182658d
        version: 2.1.2
      entries:
      - name: kiali-operator.v2.1.2
        version: 2.1.2
      - name: kiali-operator.v1.89.8
        version: 1.89.8
      - name: kiali-operator.v1.89.7
        version: 1.89.7
      - name: kiali-operator.v1.89.6
        version: 1.89.6
      - name: kiali-operator.v1.89.1
        version: 1.89.1
      - name: kiali-operator.v1.73.10
        version: 1.73.10
      - name: kiali-operator.v1.73.8
        version: 1.73.8
      - name: kiali-operator.v1.73.7
        version: 1.73.7
      - name: kiali-operator.v1.73.4
        version: 1.73.4
      - name: kiali-operator.v1.65.11
        version: 1.65.11
      - name: kiali-operator.v1.65.10
        version: 1.65.10
      - name: kiali-operator.v1.65.9
        version: 1.65.9
      - name: kiali-operator.v1.65.8
        version: 1.65.8
      - name: kiali-operator.v1.65.7
        version: 1.65.7
      - name: kiali-operator.v1.65.6
        version: 1.65.6
      - name: kiali-operator.v1.57.7
        version: 1.57.7
      name: candidate
    - currentCSV: kiali-operator.v2.11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "Kiali",
                "metadata": {
                  "name": "kiali"
                },
                "spec": {
                  "version": "default",
                  "istio_namespace": "istio-system",
                  "deployment": {
                    "logger": {
                      "log_level": "info"
                    },
                    "view_only_mode": false
                  }
                }
              },
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "OSSMConsole",
                "metadata": {
                  "name": "ossmconsole"
                },
                "spec": {
                  "version": "default",
                  "kiali": {
                    "serviceName": "",
                    "serviceNamespace": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/kiali-rhel9-operator@sha256:2970861bd8e65e36d5080dc4d56340d08a52dbb1217be352998d135eb854732f
          createdAt: "2025-09-10T00:28:49Z"
          description: This productized operator provides Kiali and OSSMC. Kiali is
            the Istio observability and management Console. OSSMC is the OpenShift
            Service Mesh Console plugin, powered by Kiali.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kiali/kiali
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A configuration file for a Kiali installation.
            displayName: Kiali
            kind: Kiali
            name: kialis.kiali.io
            version: v1alpha1
          - description: A configuration file for a OpenShift Service Mesh Console
              installation.
            displayName: OpenShift Service Mesh Console
            kind: OSSMConsole
            name: ossmconsoles.kiali.io
            version: v1alpha1
        description: |-
          ## About the managed application

          A Microservice Architecture breaks up the monolith into many smaller pieces
          that are composed together. Patterns to secure the communication between
          services like fault tolerance (via timeout, retry, circuit breaking, etc.)
          have come up as well as distributed tracing to be able to see where calls
          are going.

          A service mesh can now provide these services on a platform level and frees
          the application writers from those tasks. Routing decisions are done at the
          mesh level.

          Kiali works with OpenShift Service Mesh to visualize the service
          mesh topology, to provide visibility into features like circuit breakers,
          request rates and more. It offers insights about the mesh components at
          different levels, from abstract Applications to Services and Workloads.

          See [https://www.kiali.io](https://www.kiali.io) to read more.

          ### Accessing the UI

          By default, the Kiali operator exposes the Kiali UI as an OpenShift Route.

          If on OpenShift, you can create an OSSMConsole CR to have the operator
          install the OpenShift ServiceMesh Console plugin to the OpenShift Console
          thus providing an interface directly integrated with the OpenShift Console.

          ## About this Operator

          ### Kiali Custom Resource Configuration Settings

          For quick descriptions of all the settings you can configure in the Kiali
          Custom Resource (CR), see
          [the kiali.io docs](https://kiali.io/docs/configuration/kialis.kiali.io/).

          ## Prerequisites for enabling this Operator

          Kiali is a companion tool for OpenShift Service Mesh. So before you install Kiali, you must have
          already installed Service Mesh.
        displayName: Kiali Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - service-mesh
        - observability
        - monitoring
        - istio
        links:
        - name: Installation Guide
          url: https://www.kiali.io/docs/installation
        - name: Features
          url: https://www.kiali.io/docs/features
        - name: Documentation Home
          url: https://www.kiali.io
        - name: Blogs and Articles
          url: https://medium.com/kialiproject
        - name: Server and UI Source Code
          url: https://github.com/kiali/kiali
        - name: Operator Source Code
          url: https://github.com/kiali/kiali-operator
        - name: OpenShift Service Mesh Console Code
          url: https://github.com/kiali/openshift-servicemesh-plugin
        maintainers:
        - email: kiali-dev@googlegroups.com
          name: Kiali Developers Google Group
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9-operator@sha256:2970861bd8e65e36d5080dc4d56340d08a52dbb1217be352998d135eb854732f
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:7487fbb4b776dfca28ad09311ba857e07ad98bc91ca5ee6fcaaa5b7783d04fbd
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel9@sha256:9a0d4862a25c2999a3f85b9a7a0332dab4ec1744548191d7a00c838b06ba26d2
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel9@sha256:fe25be563d50e5ad4e5475f153e03d4b24c7db094d7e824c51ebaae79418b1f7
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel8@sha256:81db94df331799f4087596378914b712d85d117dde3df2d4902079a82d4a9c55
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9@sha256:a00ec00e466e080592dca5c1cc153698b1412a54e1979647f74281f79b5691cf
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9@sha256:85eeb972e46f26ff30e04991e76ec3050d98b334e7b7f201a094f8d50b06854e
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:d729ef8d74c2f6e02d8c68ca761b1fb3c7c3706f1290b684468f0acb0ef313e0
        version: 2.11.3
      entries:
      - name: kiali-operator.v2.11.3
        version: 2.11.3
      - name: kiali-operator.v2.11.2
        version: 2.11.2
      - name: kiali-operator.v2.11.1
        version: 2.11.1
      - name: kiali-operator.v2.4.7
        version: 2.4.7
      - name: kiali-operator.v2.4.6
        version: 2.4.6
      - name: kiali-operator.v2.4.5
        version: 2.4.5
      - name: kiali-operator.v2.4.4
        version: 2.4.4
      - name: kiali-operator.v2.4.3
        version: 2.4.3
      - name: kiali-operator.v1.89.10
        version: 1.89.10
      - name: kiali-operator.v1.89.9
        version: 1.89.9
      - name: kiali-operator.v1.89.8
        version: 1.89.8
      - name: kiali-operator.v1.89.7
        version: 1.89.7
      - name: kiali-operator.v1.89.6
        version: 1.89.6
      - name: kiali-operator.v1.89.1
        version: 1.89.1
      - name: kiali-operator.v1.73.10
        version: 1.73.10
      - name: kiali-operator.v1.73.8
        version: 1.73.8
      - name: kiali-operator.v1.73.7
        version: 1.73.7
      - name: kiali-operator.v1.73.4
        version: 1.73.4
      - name: kiali-operator.v1.65.11
        version: 1.65.11
      - name: kiali-operator.v1.65.10
        version: 1.65.10
      - name: kiali-operator.v1.65.9
        version: 1.65.9
      - name: kiali-operator.v1.65.8
        version: 1.65.8
      - name: kiali-operator.v1.65.7
        version: 1.65.7
      - name: kiali-operator.v1.65.6
        version: 1.65.6
      - name: kiali-operator.v1.57.7
        version: 1.57.7
      name: stable
    defaultChannel: stable
    packageName: kiali-ossm
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: test-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: test-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "test.openstack.org/v1beta1",
                "kind": "HorizonTest",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "test-operator",
                    "app.kubernetes.io/instance": "horizontest-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "horizontest",
                    "app.kubernetes.io/part-of": "test-operator"
                  },
                  "name": "horizontest-sample"
                },
                "spec": {
                  "adminPassword": "12345678",
                  "adminUsername": "admin",
                  "authUrl": "https://keystone-public-openstack.apps.ocp.openstack.lab",
                  "backoffLimit": 0,
                  "containerImage": "",
                  "dashboardUrl": "https://horizon-openstack.apps.ocp.openstack.lab/",
                  "flavorName": "m1.tiny",
                  "horizonRepoBranch": "master",
                  "horizonTestDir": "/var/lib/horizontest",
                  "imageUrl": "http://download.cirros-cloud.net/0.6.2/cirros-0.6.2-x86_64-disk.img",
                  "logsDirectoryName": "horizon",
                  "parallel": false,
                  "password": "horizontest",
                  "projectName": "horizontest",
                  "repoUrl": "https://review.opendev.org/openstack/horizon",
                  "storageClass": "local-storage",
                  "user": "horizontest"
                }
              },
              {
                "apiVersion": "test.openstack.org/v1beta1",
                "kind": "Tempest",
                "metadata": {
                  "name": "tempest-tests",
                  "namespace": "openstack"
                },
                "spec": {
                  "containerImage": "",
                  "tempestRun": {
                    "concurrency": 8,
                    "includeList": "tempest.api.identity.v3.*\n"
                  },
                  "tempestconfRun": null
                }
              },
              {
                "apiVersion": "test.openstack.org/v1beta1",
                "kind": "Tobiko",
                "metadata": {
                  "name": "tobiko-tests",
                  "namespace": "openstack"
                },
                "spec": {
                  "config": "[DEFAULT]\nlog_file = value\n\n[testcase]\ntimeout = value\ntest_runner_timeout = value\n\n[ubuntu]\ninterface_name = value\n\n[keystone]\ninterface_name = value\n",
                  "containerImage": "",
                  "testenv": "py3",
                  "version": "master",
                  "workflow": [
                    {
                      "stepName": "unit-tests",
                      "testenv": "py3"
                    },
                    {
                      "stepName": "functional-tests",
                      "testenv": "functional"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:43Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: Ansible Test
            kind: AnsibleTest
            name: ansibletests.test.openstack.org
            version: v1beta1
          - displayName: Horizon Test
            kind: HorizonTest
            name: horizontests.test.openstack.org
            version: v1beta1
          - displayName: Tempest
            kind: Tempest
            name: tempests.test.openstack.org
            version: v1beta1
          - kind: Tobiko
            name: tobikoes.test.openstack.org
            version: v1beta1
        description: Test Operator
        displayName: OpenStack Test Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Tempest
        - Tobiko
        - Tests
        links:
        - name: Test Operator
          url: https://github.com/openstack-k8s-operators/test-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/test-rhel9-operator@sha256:17949ad6a69a14ea38870d471fd4f5a51122927f795976ac7e2b0189e2d0da5a
        - registry.redhat.io/rhoso/openstack-tempest-all-rhel9@sha256:2f07239d932a67a6589844f4380f613cc16614ff4de7699f64c17359bbdc5534
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: test-operator.v1.0.6
        version: 1.0.6
      - name: test-operator.v1.0.5
        version: 1.0.5
      - name: test-operator.v1.0.4
        version: 1.0.4
      - name: test-operator.v1.0.3
        version: 1.0.3
      - name: test-operator.v1.0.2
        version: 1.0.2
      - name: test-operator.v1.0.1
        version: 1.0.1
      - name: test-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: test-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhsso-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhsso-operator.7.5.1-opr-005
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm"
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "False"
          containerImage: registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:66577ec32a9a975f304e52b8ab49b9f7602158dcab1ef5d8c28a5d3455491336
          createdAt: "2021-08-09 00:00:00"
          description: An Operator for installing and managing Red Hat Single Sign-On
          olm.skipRange: <7.5.0
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Red Hat Single Sign-On Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Red Hat Single Sign-On.

          Red Hat Single Sign-On lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
        displayName: Red Hat Single Sign-On Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://access.redhat.com/products/red-hat-single-sign-on
        - name: Keycloak
          url: https://www.keycloak.org/
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat Customer Service
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso7-rhel8-init-container@sha256:614a9b1d7975769e8559e0b8c1578d16f6e3f7fee0778283797f4d62f5fc2100
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:5f0840a69642d13016430151892ec987fd67e711563b1dfb1a57353037c3bf79
        - registry.redhat.io/rh-sso-7/sso75-openj9-openshift-rhel8@sha256:2d4ea7226f43e7e8e8d3fcef6d425eb5104a15acf07791355033300b80db2901
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:66577ec32a9a975f304e52b8ab49b9f7602158dcab1ef5d8c28a5d3455491336
        version: 7.5.1-opr-005
      entries:
      - name: rhsso-operator.7.5.1-opr-005
        version: 7.5.1-opr-005
      - name: rhsso-operator.7.5.1-opr-004
        version: 7.5.1-opr-004
      - name: rhsso-operator.7.5.1-opr-003
        version: 7.5.1-opr-003
      - name: rhsso-operator.7.5.1-opr-002
        version: 7.5.1-opr-002
      - name: rhsso-operator.7.5.1-opr-001
        version: 7.5.1-opr-001
      - name: rhsso-operator.7.5.0
        version: 7.5.0
      name: alpha
    - currentCSV: rhsso-operator.7.6.12-opr-004
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm"
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "False"
          containerImage: registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:229265579ec77f2bd18c936ba19e1f9dbc6b2f0a13d963d5eae5b52c260010d1
          createdAt: "2021-08-09 00:00:00"
          description: An Operator for installing and managing Red Hat Single Sign-On
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Red Hat Single Sign-On Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Red Hat Single Sign-On.

          Red Hat Single Sign-On lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
        displayName: Red Hat Single Sign-On Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://access.redhat.com/products/red-hat-single-sign-on
        - name: Keycloak
          url: https://www.keycloak.org/
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat Customer Service
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:229265579ec77f2bd18c936ba19e1f9dbc6b2f0a13d963d5eae5b52c260010d1
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:923206112ed0873ee58edd0817dbce7bb4c9a89dc75d9b3ba2affe9aed4811de
        - registry.redhat.io/rh-sso-7/sso7-rhel8-init-container@sha256:282daf78934ca495ad4051371a58e3af0917bb9241d043a66e23fc8b12fbd1ce
        version: 7.6.12-opr-004
      entries:
      - name: rhsso-operator.7.6.12-opr-004
        version: 7.6.12-opr-004
      - name: rhsso-operator.7.6.12-opr-003
        version: 7.6.12-opr-003
      - name: rhsso-operator.7.6.12-opr-002
        version: 7.6.12-opr-002
      - name: rhsso-operator.7.6.12-opr-001
        version: 7.6.12-opr-001
      - name: rhsso-operator.7.6.11-opr-006
        version: 7.6.11-opr-006
      - name: rhsso-operator.7.6.11-opr-005
        version: 7.6.11-opr-005
      - name: rhsso-operator.7.6.11-opr-004
        version: 7.6.11-opr-004
      - name: rhsso-operator.7.6.11-opr-003
        version: 7.6.11-opr-003
      - name: rhsso-operator.7.6.11-opr-002
        version: 7.6.11-opr-002
      - name: rhsso-operator.7.6.11-opr-001
        version: 7.6.11-opr-001
      - name: rhsso-operator.7.6.10-opr-001
        version: 7.6.10-opr-001
      - name: rhsso-operator.7.6.9-opr-003
        version: 7.6.9-opr-003
      - name: rhsso-operator.7.6.9-opr-002
        version: 7.6.9-opr-002
      - name: rhsso-operator.7.6.9-opr-001
        version: 7.6.9-opr-001
      - name: rhsso-operator.7.6.8-opr-002
        version: 7.6.8-opr-002
      - name: rhsso-operator.7.6.8-opr-001
        version: 7.6.8-opr-001
      - name: rhsso-operator.7.6.7-opr-002
        version: 7.6.7-opr-002
      - name: rhsso-operator.7.6.7-opr-001
        version: 7.6.7-opr-001
      - name: rhsso-operator.7.6.6-opr-004
        version: 7.6.6-opr-004
      - name: rhsso-operator.7.6.6-opr-003
        version: 7.6.6-opr-003
      - name: rhsso-operator.7.6.6-opr-002
        version: 7.6.6-opr-002
      - name: rhsso-operator.7.6.6-opr-001
        version: 7.6.6-opr-001
      - name: rhsso-operator.7.6.5-opr-004
        version: 7.6.5-opr-004
      - name: rhsso-operator.7.6.5-opr-003
        version: 7.6.5-opr-003
      - name: rhsso-operator.7.6.5-opr-002
        version: 7.6.5-opr-002
      - name: rhsso-operator.7.6.5-opr-001
        version: 7.6.5-opr-001
      - name: rhsso-operator.7.6.4-opr-003
        version: 7.6.4-opr-003
      - name: rhsso-operator.7.6.4-opr-002
        version: 7.6.4-opr-002
      - name: rhsso-operator.7.6.4-opr-001
        version: 7.6.4-opr-001
      - name: rhsso-operator.7.6.3-opr-002
        version: 7.6.3-opr-002
      - name: rhsso-operator.7.6.3-opr-001
        version: 7.6.3-opr-001
      - name: rhsso-operator.7.6.2-opr-003
        version: 7.6.2-opr-003
      - name: rhsso-operator.7.6.2-opr-002
        version: 7.6.2-opr-002
      - name: rhsso-operator.7.6.2-opr-001
        version: 7.6.2-opr-001
      - name: rhsso-operator.7.6.1-opr-005
        version: 7.6.1-opr-005
      - name: rhsso-operator.7.6.1-opr-004
        version: 7.6.1-opr-004
      - name: rhsso-operator.7.6.1-opr-003
        version: 7.6.1-opr-003
      - name: rhsso-operator.7.6.1-opr-002
        version: 7.6.1-opr-002
      - name: rhsso-operator.7.6.1-opr-001
        version: 7.6.1-opr-001
      - name: rhsso-operator.7.6.0-opr-005
        version: 7.6.0-opr-005
      - name: rhsso-operator.7.6.0-opr-004
        version: 7.6.0-opr-004
      - name: rhsso-operator.7.6.0-opr-003
        version: 7.6.0-opr-003
      - name: rhsso-operator.7.6.0-opr-002
        version: 7.6.0-opr-002
      - name: rhsso-operator.7.6.0-opr-001
        version: 7.6.0-opr-001
      - name: rhsso-operator.7.5.2-opr-003
        version: 7.5.2-opr-003
      - name: rhsso-operator.7.5.2-opr-002
        version: 7.5.2-opr-002
      - name: rhsso-operator.7.5.2-opr-001
        version: 7.5.2-opr-001
      - name: rhsso-operator.7.5.1-opr-011
        version: 7.5.1-opr-011
      - name: rhsso-operator.7.5.1-opr-010
        version: 7.5.1-opr-010
      - name: rhsso-operator.7.5.1-opr-009
        version: 7.5.1-opr-009
      - name: rhsso-operator.7.5.1-opr-009-0.1646836600.p
        version: 7.5.1-opr-009+0.1646836600.p
      - name: rhsso-operator.7.5.1-opr-008
        version: 7.5.1-opr-008
      - name: rhsso-operator.7.5.1-opr-007
        version: 7.5.1-opr-007
      - name: rhsso-operator.7.5.1-opr-006
        version: 7.5.1-opr-006
      - name: rhsso-operator.7.5.1-opr-005
        version: 7.5.1-opr-005
      - name: rhsso-operator.7.5.1-opr-004
        version: 7.5.1-opr-004
      - name: rhsso-operator.7.5.1-opr-003
        version: 7.5.1-opr-003
      - name: rhsso-operator.7.5.1-opr-002
        version: 7.5.1-opr-002
      - name: rhsso-operator.7.5.1-opr-001
        version: 7.5.1-opr-001
      - name: rhsso-operator.7.5.0
        version: 7.5.0
      name: stable
    defaultChannel: stable
    packageName: rhsso-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-athena-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-athena-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "athena.services.k8s.aws/v1alpha1",
                "kind": "WorkGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/athena-controller:1.1.1
          createdAt: "2025-09-19T21:06:02Z"
          description: AWS Athena controller is a service controller for managing
            Athena resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PreparedStatement represents the state of an AWS athena PreparedStatement
              resource.
            displayName: PreparedStatement
            kind: PreparedStatement
            name: preparedstatements.athena.services.k8s.aws
            version: v1alpha1
          - description: WorkGroup represents the state of an AWS athena WorkGroup
              resource.
            displayName: WorkGroup
            kind: WorkGroup
            name: workgroups.athena.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Athena resources in AWS from within your Kubernetes cluster.

          **About Amazon Athena**

          Amazon Athena is an interactive query service that makes it easy to analyze data directly in Amazon Simple Storage Service (Amazon S3) using standard SQL. With a few actions in the AWS Management Console, you can point Athena at your data stored in Amazon S3 and begin using standard SQL to run ad-hoc queries and get results in seconds.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Athena
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - athena
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Athena Developer Resources
          url: https://aws.amazon.com/athena/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: athena maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/athena-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-athena-controller.v1.1.1
        version: 1.1.1
      - name: ack-athena-controller.v1.1.0
        version: 1.1.0
      - name: ack-athena-controller.v1.0.14
        version: 1.0.14
      - name: ack-athena-controller.v1.0.13
        version: 1.0.13
      - name: ack-athena-controller.v1.0.12
        version: 1.0.12
      - name: ack-athena-controller.v1.0.11
        version: 1.0.11
      - name: ack-athena-controller.v1.0.10
        version: 1.0.10
      - name: ack-athena-controller.v1.0.9
        version: 1.0.9
      - name: ack-athena-controller.v1.0.8
        version: 1.0.8
      - name: ack-athena-controller.v1.0.7
        version: 1.0.7
      - name: ack-athena-controller.v1.0.6
        version: 1.0.6
      - name: ack-athena-controller.v1.0.5
        version: 1.0.5
      - name: ack-athena-controller.v1.0.4
        version: 1.0.4
      - name: ack-athena-controller.v1.0.1
        version: 1.0.1
      - name: ack-athena-controller.v1.0.0
        version: 1.0.0
      - name: ack-athena-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-athena-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Protocol Labs
      provider-url: ""
    name: ipfs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ipfs-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.ipfs.io/v1alpha1",
                "kind": "CircuitRelay",
                "metadata": {
                  "name": "circuitrelay-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cluster.ipfs.io/v1alpha1",
                "kind": "IpfsCluster",
                "metadata": {
                  "name": "ipfscluster-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Storage, Networking, Big Data
          containerImage: quay.io/redhat-et-ipfs/ipfs-operator:v0.0.1
          createdAt: 2022-04-05T15:50:37-0400
          operatorframework.io/suggested-namespace: ipfs-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ipfs-cluster/ipfs-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CircuitRelay is the Schema for the circuitrelays API.
            displayName: Circuit Relay
            kind: CircuitRelay
            name: circuitrelays.cluster.ipfs.io
            version: v1alpha1
          - description: IpfsCluster is the Schema for the ipfs API.
            displayName: Ipfs Cluster
            kind: IpfsCluster
            name: ipfsclusters.cluster.ipfs.io
            version: v1alpha1
        description: "The IPFS Operator enables users to easily create IPFS Cluster
          instances\nwithin their Kubernetes or OpenShift clusters. \n\n### Supported
          Features\n* Horizontal scaling\n* Joining collaborative clusters\n* Controlling
          IPFS node resource usage\n* Specifying reprovider settings  \n\n### Documentation
          \nDocumentation can be found on our [read-the-docs page](https://ipfs-operator.readthedocs.io/en/latest/),
          as well as our [GitHub repository](https://github.com/ipfs-cluster/ipfs-operator).\n\n###
          Getting Help\nIn case you run into issues while using the operator, please
          create an issue in our [GitHub\nrepository](https://github.com/ipfs-cluster/ipfs-operator).\n\n###
          Contributing\nWe welcome contributions of all types and sizes! You can find
          open issues\nin our [GitHub Repository](https://github.com/ipfs-cluster/ipfs-operator/issues),\nas
          well as make your own issue if you feel that something is missing.\n\n###
          License\nIPFS Operator is licensed under the [Apache 2.0 License](https://github.com/ipfs-cluster/ipfs-operator/blob/main/LICENSE)"
        displayName: IPFS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipfs
        - ipfs cluster
        - decentralized storage
        - peer-to-peer
        - distributed hash table
        - content-addressed
        links:
        - name: IPFS Operator GitHub page
          url: https://github.com/ipfs-cluster/ipfs-operator
        - name: IPFS Cluster
          url: https://ipfscluster.io/
        - name: IPFS
          url: https://ipfs.io/
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Protocol Labs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/redhat-et-ipfs/ipfs-operator:v0.0.1
        version: 0.0.2
      entries:
      - name: ipfs-operator.v0.0.2
        version: 0.0.2
      - name: ipfs-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ipfs-operator
    provider:
      name: Protocol Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: gaudi-ai-sw-tools-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gaudi-ai-sw-tools-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aitools.intel/v1",
                "kind": "GaudiAIToolsContainer",
                "metadata": {
                  "name": "intel-gaudi-1.17.0-495"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagespec": {
                    "registry": "registry.connect.redhat.com",
                    "repo": "intel/gaudi-notebooks",
                    "tags": [
                      {
                        "gaudi_software": "1.17.0-495",
                        "rhel_os": "9.2"
                      }
                    ]
                  },
                  "nameOverride": "",
                  "namespace": "redhat-ods-applications"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: registry.connect.redhat.com/intel/gaudi-ai-sw-tools-operator@sha256:0c286f1e4cec57a189d749501f82345afbe0a1dbbbf92b906ad3c549e47b89cd
          createdAt: "2024-09-11T21:27:02Z"
          description: Operator to enable Gaudi AI SW Tools into OpenShift AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: GaudiAIToolsContainer
            name: gaudiaitoolscontainers.aitools.intel
            version: v1
        description: Operator to enable Gaudi AI SW Tools into OpenShift AI
        displayName: Intel® Gaudi AI SW Tools Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gaudi
        - opendatahub
        - openshiftai
        - notebooks
        links:
        - name: Intel® Gaudi AI SW Tools Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/66d7aa630eb66a02febc8103
        maintainers:
        - email: sharvil.shah@intel.com
          name: sharvil
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.connect.redhat.com/intel/gaudi-ai-sw-tools-operator@sha256:0c286f1e4cec57a189d749501f82345afbe0a1dbbbf92b906ad3c549e47b89cd
        version: 0.0.2
      entries:
      - name: gaudi-ai-sw-tools-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: gaudi-ai-sw-tools-operator
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-certified-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "psp": false,
                  "serviceAccount": "default",
                  "controller": {
                    "enabled": true,
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": "registry.connect.redhat.com/neuvector/controller",
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": ""
                      }
                    },
                    "pvc": {
                      "enabled": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "path": "/",
                          "annotations": {
                            "ingress.kubernetes.io/protocol": "https"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "path": "/",
                          "annotations": {
                            "ingress.kubernetes.io/protocol": "https"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "path": "/",
                      "annotations": {
                        "ingress.kubernetes.io/protocol": "https"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {}
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/enforcer",
                    "priorityClassName": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      }
                    ],
                    "resources": {}
                  },
                  "manager": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/manager",
                    "priorityClassName": "",
                    "env": {
                      "ssl": true
                    },
                    "svc": {
                      "type": "NodePort",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "path": "/",
                      "annotations": {},
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "affinity": {},
                    "tolerations": [],
                    "nodeSelector": {}
                  },
                  "cve": {
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "image": "registry.access.redhat.com/ubi8@sha256:228824aa581f3b31bf79411f8448b798291c667a37155bdea61cfa128b2833f2",
                      "schedule": "0 0 * * *",
                      "priorityClassName": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": "registry.connect.redhat.com/neuvector/scanner@sha256:42004e13b348bb43b58ae8538c983a70b237e3aca870758ac629fb4fc43e8130",
                      "priorityClassName": "",
                      "resources": {},
                      "affinity": {},
                      "tolerations": [],
                      "nodeSelector": {}
                    }
                  },
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "resources": {},
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "true"
          containerImage: registry.connect.redhat.com/neuvector/neuvector-operator@sha256:6155ac3c95aa6471c9857e0a6f5e5d19e6fc1d8cc7d014108d7d4a10d14f96bb
          createdAt: "2020-12-02 14:19:21"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/neuvector-certified-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/neuvector-certified-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: Neuvector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm/tree/master/charts/core) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart.  NeuVector Operator versions are tied to NeuVector product versions. Version 1.3.4 of the NeuVector Certified Operator deploys version 4.4.3 of NeuVector.

          **Complete below steps to Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Grant Service Account Access to the Privileged SCC

                   oc -n neuvector adm policy add-scc-to-user privileged -z default

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s                    neuvector/default


          **Add NeuVector license from NeuVector WebUI->setting**


          #Deploying the NeuVector Operator#


          Please refer to the instructions [here](https://github.com/neuvector/neuvector-operator/blob/master/README.md)


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Networking
        - Security
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - registry.connect.redhat.com/neuvector/scanner@sha256:42004e13b348bb43b58ae8538c983a70b237e3aca870758ac629fb4fc43e8130
        - registry.access.redhat.com/ubi8@sha256:228824aa581f3b31bf79411f8448b798291c667a37155bdea61cfa128b2833f2
        - registry.connect.redhat.com/neuvector/neuvector-operator@sha256:6155ac3c95aa6471c9857e0a6f5e5d19e6fc1d8cc7d014108d7d4a10d14f96bb
        - registry.connect.redhat.com/neuvector/controller@sha256:acc943568520282061547b310c50080dc32924a531bbb543d959b281eb6664f4
        - registry.connect.redhat.com/neuvector/enforcer@sha256:be1d299ec3532d55444681a882e9fd1238949da6598987308e2800ed26291b7a
        - registry.connect.redhat.com/neuvector/manager@sha256:8245cd179d11fff279690c5a422f46b4ff252ed72b9b998f0c9b5453bc7d8fc0
        version: 1.3.4
      entries:
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-certified-operator-rhmp
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dpu-network-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dpu-network-operator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dpu.openshift.io/v1alpha1",
                "kind": "OVNKubeConfig",
                "metadata": {
                  "name": "ovnkubeconfig-sample"
                },
                "spec": {
                  "kubeConfigFile": "tenant-cluster-1-kubeconf",
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/dpu-worker": ""
                    }
                  },
                  "poolName": "dpu"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional, Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/dpu-network-rhel8-operator@sha256:d07b8ecb5d1625599878b5c1e34b0de94188b81d90c829787a2d13969f6d5971
          createdAt: "2023-07-13T18:03:11Z"
          description: The operator is responsible for the life-cycle management of
            the ovn-kube components and the necessary host network initialization
            on DPU cards.
          olm.skipRange: '>=4.10.0-0 <4.13.0-202309181427'
          operatorframework.io/suggested-namespace: openshift-dpu-network-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/dpu-network-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OVNKubeConfig is the Schema for the ovnkubeconfigs API
            displayName: OVNKube Config
            kind: OVNKubeConfig
            name: ovnkubeconfigs.dpu.openshift.io
            version: v1alpha1
        description: The operator to be responsible for the life-cycle management
          of the ovn-kube components and the necessary host network initialization
          on DPU cards.
        displayName: DPU Network Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dpu
        - network
        links:
        - name: Dpu Network Operator
          url: https://dpu-network-operator.domain
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:767682dd3bd65651a8204c9fb732b9c48b99127189992b43abc6ecce027f4589
        - registry.redhat.io/openshift4/dpu-network-rhel8-operator@sha256:d07b8ecb5d1625599878b5c1e34b0de94188b81d90c829787a2d13969f6d5971
        version: 4.13.0-202309181427
      entries:
      - name: dpu-network-operator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: stable
    defaultChannel: stable
    packageName: dpu-network-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-glue-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-glue-controller.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/glue-controller:0.2.1
          createdAt: "2025-09-19T20:55:49Z"
          description: AWS Glue controller is a service controller for managing Glue
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Job represents the state of an AWS glue Job resource.
            displayName: Job
            kind: Job
            name: jobs.glue.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Glue resources in AWS from within your Kubernetes cluster.

          **About Amazon Glue**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Glue
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - glue
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Glue Developer Resources
          url: https://aws.amazon.com/Glue/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: glue maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/glue-controller:0.2.1
        version: 0.2.1
      entries:
      - name: ack-glue-controller.v0.2.1
        version: 0.2.1
      - name: ack-glue-controller.v0.2.0
        version: 0.2.0
      - name: ack-glue-controller.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ack-glue-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fujitsu
      provider-url: https://www.postgresql.fastware.com/
    name: fujitsu-enterprise-postgres-operator-with-cg-module
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPAction",
                "metadata": {
                  "name": "new-fep-action"
                },
                "spec": {
                  "fepAction": {
                    "args": [
                      "new-fep-sts-0",
                      "new-fep-sts-1"
                    ],
                    "type": "reload"
                  },
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "targetClusterName": "new-fep"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPExporter",
                "metadata": {
                  "name": "new-fep-exporter"
                },
                "spec": {
                  "fepExporter": {
                    "exporterLogLevel": "error",
                    "fepClusterList": [
                      "new-fep1"
                    ],
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "userCustomQueries": "usr_example:\n  query: \"SELECT EXTRACT(EPOCH FROM (now() - pg_last_xact_replay_timestamp())) as lag\"\n  master: true\n  metrics:\n    - lag:\n        usage: \"GAUGE\"\n        description: \"Replication lag behind master in seconds\""
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPLogging",
                "metadata": {
                  "name": "new-fep-logging"
                },
                "spec": {
                  "fepLogging": {
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPPgpool2",
                "metadata": {
                  "name": "new-fep-pgpool2"
                },
                "spec": {
                  "clientAuthMethod": "scram",
                  "count": 2,
                  "customhba": "local   all         all                               trust\nhost    all         all         127.0.0.1/32          trust\nhost    all         all         ::1/128               trust\n",
                  "customlogsize": "128Mi",
                  "customparams": "listen_addresses = '*'\npcp_listen_addresses = '*'\nnum_init_children = 32\nreserved_connections = 0\nenable_pool_hba = off\nallow_clear_text_frontend_auth = off\nauthentication_timeout = 80\nbackend_weight0 = 1\nbackend_weight1 = 1\nbackend_flag0 = 'ALWAYS_PRIMARY'\nbackend_flag1 = 'DISALLOW_TO_FAILOVER'\nconnection_cache = on\nmax_pool = 4\nlisten_backlog_multiplier = 2\nserialize_accept = off\nchild_life_time = 300\nclient_idle_limit = 0\nchild_max_connections = 0\nconnection_life_time = 0\nreset_query_list = 'ABORT; DISCARD ALL'\nclient_min_messages = info\nlog_min_messages = debug1\nlog_statement = on\nlog_per_node_statement = on\nlog_client_messages = on\nlog_hostname = on\nlog_connections = on\nlog_line_prefix = '%t: pid %p: '\nload_balance_mode = on\nignore_leading_white_space = on\nwhite_function_list = ''\nblack_function_list = 'currval,lastval,nextval,setval'\nblack_query_pattern_list = ''\ndatabase_redirect_preference_list = ''\napp_name_redirect_preference_list = ''\nallow_sql_comments = off\ndisable_load_balance_on_write = 'transaction'\nstatement_level_load_balance = on\nsr_check_period = 0\nsr_check_user = 'postgres'\ndelay_threshold = 0\nlog_standby_delay = 'none'\nssl = on\nssl_ciphers = 'HIGH:MEDIUM:+3DES:!aNULL'\nssl_prefer_server_ciphers = off\nssl_ecdh_curve = 'prime256v1'\nssl_dh_params_file = ''\nrelcache_expire = 0\nrelcache_size = 256\ncheck_temp_table = catalog\ncheck_unlogged_table = on\nenable_shared_relcache = off\nrelcache_query_target = primary\nwd_port0 = 9000\nfailover_on_backend_error = off\n",
                  "custompcp": "none",
                  "customsslcacert": "none",
                  "customsslcert": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                  "customsslkey": "none",
                  "fepclustername": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "limits": {
                    "cpu": "400m",
                    "memory": "512Mi"
                  },
                  "requests": {
                    "cpu": "200m",
                    "memory": "256Mi"
                  },
                  "serviceport": 9999,
                  "statusport": 9898
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPRestore",
                "metadata": {
                  "name": "new-fep-restore"
                },
                "spec": {
                  "fromFEPcluster": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "mcSpec": {
                    "limits": {
                      "cpu": "200m",
                      "memory": "300Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "restoretype": "latest",
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "toFEPcluster": "new-fep-2"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v2",
                "kind": "FEPCluster",
                "metadata": {
                  "name": "new-fep"
                },
                "spec": {
                  "fep": {
                    "customAnnotations": {
                      "allDeployments": {}
                    },
                    "forceSsl": true,
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "instances": 1,
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "podAntiAffinity": false,
                    "podDisruptionBudget": false,
                    "servicePort": 27500,
                    "syncMode": "off",
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "usePodName": true
                  },
                  "fepChildCrVal": {
                    "backup": {
                      "image": {
                        "pullPolicy": "IfNotPresent"
                      },
                      "mcSpec": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      },
                      "pgbackrestParams": "# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[global]\nrepo1-retention-full=7\nrepo1-retention-full-type=time\nlog-path=/database/log/backup\n",
                      "postScript": " ",
                      "preScript": " ",
                      "schedule": {
                        "num": 2
                      },
                      "schedule1": {
                        "schedule": "15 0 * * 0",
                        "type": "full"
                      },
                      "schedule2": {
                        "schedule": "15 0 * * 1-6",
                        "type": "incr"
                      },
                      "schedule3": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule4": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule5": {
                        "schedule": " ",
                        "type": " "
                      }
                    },
                    "customPgAudit": "# define pg audit custom params here to override defaults.\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[output]\nlogger = 'auditlog'\nlog_directory = '/database/log/audit'\nlog_truncate_on_rotation = on\nlog_filename = 'pgaudit-%a.log'\nlog_rotation_age = 1d\nlog_rotation_size = 0\n[rule]\n",
                    "customPgHba": "# define pg_hba custom rules here to be merged with default rules.\n# TYPE     DATABASE        USER        ADDRESS        METHOD\n",
                    "customPgParams": "# define custom postgresql.conf parameters below to override defaults.\n# Current values are as per default FEP deployment\n# If you add a library to shared_preload_libraries, add it after the default value.\n# fsep_operator_security is available in FEP15 and later container images.\nshared_preload_libraries='pgx_datamasking,pg_prewarm,pg_stat_statements,fsep_operator_security'\nsession_preload_libraries='pg_prewarm'\nmax_prepared_transactions = 100\nmax_worker_processes = 30\nmax_connections = 100\nwork_mem = 1MB\nmaintenance_work_mem = 12MB\nshared_buffers = 128MB\neffective_cache_size = 384MB\ncheckpoint_completion_target = 0.8\n# tcp parameters\ntcp_keepalives_idle = 30\ntcp_keepalives_interval = 10\ntcp_keepalives_count = 3\n# logging parameters in default fep installation\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\nlog_directory = '/database/log'\nlog_filename = 'logfile-%a.log'\nlog_file_mode = 0600\nlog_truncate_on_rotation = on\nlog_rotation_age = 1d\nlog_rotation_size = 0\nlog_checkpoints = on\nlog_line_prefix = '%e %t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h'\nlog_lock_waits = on\nlog_autovacuum_min_duration = 60s\nlogging_collector = on\npgaudit.config_file='/opt/app-root/src/pgaudit-cfg/pgaudit.conf'\nlog_replication_commands = on\nlog_min_messages = WARNING\nlog_destination = csvlog\n# wal_archive parameters in default fep installation\narchive_mode = on\narchive_command = 'pgbackrest --stanza=backupstanza --config=/database/userdata/pgbackrest.conf archive-push %p'\nwal_level = replica\nmax_wal_senders = 12\nwal_keep_segments = 64\ntrack_activities = on\ntrack_counts = on\npassword_encryption = 'scram-sha-256'\n",
                    "storage": {
                      "archivewalVol": {
                        "size": "1Gi"
                      },
                      "backupVol": {
                        "size": "2Gi"
                      },
                      "dataVol": {
                        "size": "2Gi"
                      },
                      "logVol": {
                        "size": "1Gi"
                      },
                      "tablespaceVol": {
                        "size": "512Mi"
                      },
                      "walVol": {
                        "size": "1200Mi"
                      }
                    },
                    "sysUsers": {
                      "pgRewindPassword": "rewind_password",
                      "pgRewindUser": "rewind_user",
                      "pgdb": "mydb",
                      "pgpassword": "mydbpassword",
                      "pgreplpassword": "repluserpwd",
                      "pgrepluser": "repluser",
                      "pguser": "mydbuser",
                      "tdepassphrase": "tde-passphrase"
                    },
                    "systemCertificates": {
                      "cacrt": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                      "crt": "-----BEGIN CERTIFICATE-----\nMIIDUTCCAjmgAwIBAgIRAMocW3qMoHrD6qRvMPppMkMwDQYJKoZIhvcNAQELBQAw\nNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9yIEt1\nYmVybmV0ZXMwHhcNMjEwMjA2MDQzMjM2WhcNMjYwMjA1MDQzMjM2WjA/MRAwDgYD\nVQQKEwdGdWppdHN1MSswKQYDVQQDEyJGVUpJVFNVIEVudGVycHJpc2UgUG9zdGdy\nZXMgU2VydmVyMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA4AI33yvH\nZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I2e4SceTKi6O3C/I1XuvWlpng\n5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4ANB5JWWqDOjrRT3o7nRPGXfila\nbP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYpmjdbfxabTz07ig0+6/cwKoRR\nxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTFYvmAH7gcdssSFBt8NPlUATHE\nsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6Wdgmu5H2pDml8CDNLDv98Aj7i\n+I5SRKKcVPlnuQIDAQABo1AwTjAdBgNVHSUEFjAUBggrBgEFBQcDAQYIKwYBBQUH\nAwIwDAYDVR0TAQH/BAIwADAfBgNVHSMEGDAWgBQcwrrUO0u+FhIUuVdrDRCQRsi6\nZjANBgkqhkiG9w0BAQsFAAOCAQEAm5dxBoI9pScOCvRAchg4CprdRDSJb9K6yB3O\nnCAxnM47iHeXnY3WlnI388kHu8DU7O4ba1tJbGs3KY9KzioPk43pU12jWkO1onoF\n+mTDjx/Ef1cYWA9r5q/LtgTa6Q2sxV4O2x67QW82aAnaxO34dV5zWCPIvAoovZBV\nHRT+BgCg3r2vD1RGKK2nl1aYJtWhO1SZubam+VttdZ/vbM9oOJctxmImsEtBXjkY\nKteePdQtLL5o03JhyXWyRshCq+HMmKf2KgyY8gvydGcP4eLQdBWcW40LcnVq6UjT\n0kJycJEKngMVademq1ZWHGaiYB7hyT6GhgIcHUJ2cKrPgbEh1Q==\n-----END CERTIFICATE-----",
                      "key": "-----BEGIN RSA PRIVATE KEY-----\nMIIEowIBAAKCAQEA4AI33yvHZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I\n2e4SceTKi6O3C/I1XuvWlpng5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4AN\nB5JWWqDOjrRT3o7nRPGXfilabP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYp\nmjdbfxabTz07ig0+6/cwKoRRxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTF\nYvmAH7gcdssSFBt8NPlUATHEsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6W\ndgmu5H2pDml8CDNLDv98Aj7i+I5SRKKcVPlnuQIDAQABAoIBAFPQYKlOzw/+BA0b\nyMIUpdctIMb/54CR/xR0mVw1DbSjigNVPjHUQvB8Y1B2FAITQObgJO06bAv0QdWN\nRb0/v/yYiNJDFjaLjaIAHlO/2+oWrXbFaZqgpVDJhB+e1xaZr2x7XGxm+p925k30\nl6pvIRY+I8JRKvZiV1VZHwL/R3JOtPr++xMZtLVjVOI+f+ySqJ+TZHuAjm49EKxj\ncEmmJ28b7QcziXsvKy00f+zbqLIBKXQdZAFU5eEr1BsDRXdRW+Kf0XIvftuy4BJZ\nvoKT+VGhEvF/qysswL4+6IAO6tpuYnnM0Y2d3sOGoWPkTcQK0MekYKzL/WmtCjNs\n9hodJtECgYEA5EWyhEOf4uOKe5TDp697UCUvXLoOR58FDe/S8XNvScn29jjOkqIg\nOMoqo9xAkJTNTzqn5UUdt1x/pgM2NxlPLFijrc0zQlX3SoOO2ryDd9WNi7YKtN16\nKJqa536WeZu2OEbuAZ+S3GALVy1RPeTNPnUOmKnF06DjDUGzLNCZy10CgYEA+zfw\n952DWuz1U0Z4wvAEqqcgUKXPKrkTXV/iUnjkDkrLYVr0ZofDNTXrdHl+UedFmaOC\ncieZn6DNhcdz5tKtyysGMH3g/qs9PfoGUngvcXsy0Egk04l3x1jc8TTCLqXZXYaQ\nHMsx51n+R58oncPtzYSUOr9qQ6PbC2CstTbFJA0CgYEAjGEsUliAB/jknfEzjXjG\nPdhQUxb8VyE864Az2lah9t/kJzFyIAziAeqZ5GE7t247AGFTBRTHHI8e1Qoemi3P\nWbc9GVIbFs1lIYbcIDpUIyrKPEP8O5QEXtoNLxXTFgAjRGKiVY87spjCAJ+W2ZhO\ne/1it5GYXfgQCYQA2yuBmOUCgYANRkR2YR1axaCk+NlSu6oTdmdPu6M5x7PNQE7O\nOtMaKjua9lppvIzFGAdMDUtueoEEAE7ZR1xnwfB6PDLUpJdIYAqgr1YfPt8qkjaZ\nTv56yZ7CwL0pbF8m6nwqRrZoDp1wwraEvvvxFKFKGY/k3kCHlpTakdjEoDjn3gDi\nRnWeVQKBgCEneMSzucei5LRppRtRaJw/Btll8qlPMlX3W7dxQ3cLwpmLOn0m51Fp\nPIZ44zYK8R6fu4+/sSrlfaIg86Ugeufp6YNxyNROKxUGza5vDIu5OftwWtBeg+UK\nZ8lLWNdX6pp7WMujmF3H1DrkBbauYMUKZ4UxUYtelgHERMePIxwb\n-----END RSA PRIVATE KEY-----"
                    }
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "false"
          containerImage: quay.io/fujitsu/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:eaf601e7aef525ebcd4b769c859ed74c7a8077f843b3a75d76713eabf0295745
          createdAt: "2025-09-11T09:32:52Z"
          description: OpenShift Operator for Fujitsu Enterprise Postgres
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.1.0 <7.1.6'
          operators.openshift.io/valid-subscription: '["Fujitsu Enterprise Postgres
            Advanced Edition with Cryptographic Module Annual Subscription License
            per Core 17 Operator for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/internal-objects: '["fepautoscales.fep.fujitsu.io","fepconfigs.fep.fujitsu.io","fepusers.fep.fujitsu.io","fepcerts.fep.fujitsu.io","fepvolumes.fep.fujitsu.io","fepbackups.fep.fujitsu.io"]'
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: fj-dbaas-tec@dl.jp.fujitsu.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: FEPAction
            name: fepactions.fep.fujitsu.io
            version: v1
          - kind: FEPAutoscale
            name: fepautoscales.fep.fujitsu.io
            version: v1
          - kind: FEPBackup
            name: fepbackups.fep.fujitsu.io
            version: v1
          - kind: FEPCert
            name: fepcerts.fep.fujitsu.io
            version: v1
          - kind: FEPCluster
            name: fepclusters.fep.fujitsu.io
            version: v2
          - kind: FEPConfig
            name: fepconfigs.fep.fujitsu.io
            version: v1
          - kind: FEPExporter
            name: fepexporters.fep.fujitsu.io
            version: v1
          - kind: FEPLogging
            name: feploggings.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2Cert
            name: feppgpool2certs.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2
            name: feppgpool2s.fep.fujitsu.io
            version: v1
          - kind: FEPRestore
            name: feprestores.fep.fujitsu.io
            version: v1
          - kind: FEPUpgrade
            name: fepupgrades.fep.fujitsu.io
            version: v1
          - kind: FEPUser
            name: fepusers.fep.fujitsu.io
            version: v1
          - kind: FEPVolume
            name: fepvolumes.fep.fujitsu.io
            version: v1
        description: "Fujitsu Enterprise Postgres 17 delivers an enterprise-grade
          PostgreSQL on OpenShift Container Platform.\n\nThis solution provides the
          flexibility of a hybrid cloud solution while delivering an enhanced distribution
          of PostgreSQL to support enterprise-level workloads and provide improved
          deployment and management, availability, performance, data governance and
          security.  \n\nAvailable as a multi-architecture container built for both
          amd64.\n\nThe download and Use of the Product is strictly subject to the
          terms of the End User License Agreement with Fujitsu Limited found at https://www.fast.fujitsu.com/fujitsu-enterprise-postgres-license-agreements.
          Where the Product that has been embedded as a whole or part into a third
          party program, only Authorised Customers may download and use the Product.\n"
        displayName: Fujitsu Enterprise Postgres Operator with Cryptographic Module
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - Fujitsu
        - fep
        links:
        - name: Fujitsu Enterprise Postgres
          url: https://www.postgresql.fastware.com/
        maintainers:
        - email: fj-dbaas-tec@dl.jp.fujitsu.com
          name: Fujitsu
        maturity: stable
        provider:
          name: Fujitsu
          url: https://www.postgresql.fastware.com/
        relatedImages:
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-backup-with-cryptographic-module@sha256:dd15cf3b6b4a5c959efef8b602abe5d64b4f5a424fb6a3e55079a2040e67cd7a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-utils-with-cryptographic-module@sha256:8b2b6a981c863f7075ced1235c614317b6526626b42405295a9c664736f0a80f
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-restore-with-cryptographic-module@sha256:6c0d6e10cfe1868be2986fdde14c6f55f86b12cea01a82199b1e2e0c3154ef75
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-restore-with-cryptographic-module@sha256:98a3b0bfa6ec7a9a4a08fb524c331c3de6a9b580d6386a7b5430955ed84b2fc5
        - quay.io/fujitsu/fujitsu-enterprise-postgres-exporter-with-cryptographic-module@sha256:71e1d1e715e2d89757e80d90b8d5b11baa729110cd5f7898df920a4c85f2bf8a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-pgpool2-with-cryptographic-module@sha256:a4c5a59ca18f6fa295b78d570f03991f1a4b4bd6df52f8526a1254ac1a567990
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-upgrade-with-cryptographic-module@sha256:1fe690a9c04a55213062ad3bef8c0cff73849f8777c3945b9a8ae2dcff6330fa
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-backup-with-cryptographic-module@sha256:03b7f3c76e3eb775e4ec7f66f7506ee1feeab56dca96e868ac3f38b9c1ca5567
        - quay.io/fujitsu/fujitsu-enterprise-postgres-cronjob-with-cryptographic-module@sha256:809a64600396378ff8c825595d40f72aedbc2b57f394420548262b085a3f2ea5
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentbit-with-cryptographic-module@sha256:bb661b3aad0b825d56a9696d6d2d9f5c04add2a83e793fc3d4a73980f4a7a0ac
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-server-with-cryptographic-module@sha256:cad4e36dc2b1e5ca6bf9308d9f93958d4c8f2bbdc14e000ca0b5700fa335df44
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-utils-with-cryptographic-module@sha256:4caf1dc5211ac03893ce9a62b417f412ffd411f13569f3ac75e12ca5128af9e4
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-server-with-cryptographic-module@sha256:a012f21f63729ba667af097375e08874156e2cf1b94ae203985d45f11031fea9
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-upgrade-with-cryptographic-module@sha256:caf6e1d2cc31d01958e4adcb4c3c307d761f350d5b0963ce19214f041c14184a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-upgrade-with-cryptographic-module@sha256:b8be0344b6cf99b2c5de2b16f04f6f2b05bde861d9b38afd319bbbc4dd72a620
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-server-with-cryptographic-module@sha256:7778f4b03861b56371d18f7726b68ad21d0faa7c1a25cb38fe7cac483d81c095
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-pgpool2-with-cryptographic-module@sha256:e510bc2d30df33e4c9d3babb898ff9cbd13704fa3fa69a284bf672765fb03083
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-pgpool2-with-cryptographic-module@sha256:e8aa27cbb57503ec5f4f37beeb2e833ace08d4c583d4361b7ce945c7461a904c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-restore-with-cryptographic-module@sha256:edeb7b72a89e05995661e6a17e75b5b477a3ca351e0db146a939a33e88e16c97
        - quay.io/fujitsu/fujitsu-enterprise-postgres-utils-with-cryptographic-module@sha256:7d169df0b8fe47000891afde79be8dd7978da3f893d1e92f01f57fd8c0b3b4c6
        - quay.io/fujitsu/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:eaf601e7aef525ebcd4b769c859ed74c7a8077f843b3a75d76713eabf0295745
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-backup-with-cryptographic-module@sha256:5b49aa46386ca954ebcbea28fbabd15929cb64fc2a4d2dd2796c9d2e15b212c6
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentd-with-cryptographic-module@sha256:f743f35012e6fd3d54a6b54baeeee98bd7c65fe52923acd87546066fa9b54d7e
        version: 7.1.6
      entries:
      - name: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.6
        version: 7.1.6
      - name: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.5
        version: 7.1.5
      name: stable
    defaultChannel: stable
    packageName: fujitsu-enterprise-postgres-operator-with-cg-module
    provider:
      name: Fujitsu
      url: https://www.postgresql.fastware.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-05-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx8-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx8-operator-rhmp/support?utm_source=openshift_console
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:154f85cc6721a70909dee755ddd88897ad51038afe2538b6dee4d26e7d8b9afb
        version: 0.0.10
      entries:
      - name: joget-dx8-openshift-operator.v0.0.10
        version: 0.0.10
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-operator-rhmp
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://www.redhat.com/
    name: rhdh
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhdh-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
          createdAt: "2025-09-12T06:25:11Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.7.1'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.7/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:62f32e50727c5006766a34daa5b6472b649cd9894f3fe9543b8ecc67e6760e8e
        version: 1.7.1
      entries:
      - name: rhdh-operator.v1.7.1
        version: 1.7.1
      - name: rhdh-operator.v1.7.0
        version: 1.7.0
      - name: rhdh-operator.v1.6.5
        version: 1.6.5
      - name: rhdh-operator.v1.6.4
        version: 1.6.4
      - name: rhdh-operator.v1.6.3
        version: 1.6.3
      - name: rhdh-operator.v1.6.2
        version: 1.6.2
      - name: rhdh-operator.v1.6.1
        version: 1.6.1
      - name: rhdh-operator.v1.6.0
        version: 1.6.0
      - name: rhdh-operator.v1.5.3
        version: 1.5.3
      - name: rhdh-operator.v1.5.2
        version: 1.5.2
      - name: rhdh-operator.v1.5.1
        version: 1.5.1
      - name: rhdh-operator.v1.4.3
        version: 1.4.3
      - name: rhdh-operator.v1.4.2
        version: 1.4.2
      - name: rhdh-operator.v1.4.1
        version: 1.4.1
      - name: rhdh-operator.v1.4.0
        version: 1.4.0
      - name: rhdh-operator.v1.3.5
        version: 1.3.5
      - name: rhdh-operator.v1.3.4
        version: 1.3.4
      - name: rhdh-operator.v1.3.3
        version: 1.3.3
      - name: rhdh-operator.v1.3.1
        version: 1.3.1
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast
    - currentCSV: rhdh-operator.v1.1.2-0.1714688890.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:9bea7eabdea44342248dfd6091a8f5c3b6e65884c916b3c2a073a4a64481aa6f
          createdAt: "2023-03-20T16:11:34Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins, configuration settings, and deployment
            mechanisms, which can help streamline the process of setting up a self-managed
            internal developer portal for adopters who are just starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: rhdh-operator.v1.1.2
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.1.0 <1.2.0'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/janus-idp/operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: asoro@redhat.com
          name: Armel Soro
        - email: gazarenk@redhat.com
          name: Gennady Azarenkov
        - email: jianrzha@redhat.com
          name: Jianrong Zhang
        - email: nboldt@redhat.com
          name: Nick Boldt
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:95ed239eb427c7b1e422ece28c5277b5018efb751c2b472bf502b8216219c047
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:9bea7eabdea44342248dfd6091a8f5c3b6e65884c916b3c2a073a4a64481aa6f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:d7ae2456857cab09a9e9acf69c9990ef18deb7f920046b8c2483fbfd0d42bbfc
        - registry.redhat.io/rhel9/postgresql-15@sha256:d5f3d496ecd55688e384cc0627700a97af6924582123bcf00749c375724af3de
        version: 1.1.2+0.1714688890.p
      entries:
      - name: rhdh-operator.v1.1.2
        version: 1.1.2
      - name: rhdh-operator.v1.1.2-0.1714608881.p
        version: 1.1.2+0.1714608881.p
      - name: rhdh-operator.v1.1.2-0.1714688618.p
        version: 1.1.2+0.1714688618.p
      - name: rhdh-operator.v1.1.2-0.1714688890.p
        version: 1.1.2+0.1714688890.p
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.1
    - currentCSV: rhdh-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:c9051a6c339cd40cbc7474a33b32a6b94913af964e12ca5da0c0b7e1c6f444d5
          createdAt: "2024-11-28T17:55:19Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.2.6'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.2/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.2/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: asoro@redhat.com
          name: Armel Soro
        - email: cdaley@redhat.com
          name: Corey Daley
        - email: gazarenk@redhat.com
          name: Gennady Azarenkov
        - email: nboldt@redhat.com
          name: Nick Boldt
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:ef29f2b340a6dc52c5014239876d54a54f7c0f0bf4ea917391438a69c94a8532
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:c9051a6c339cd40cbc7474a33b32a6b94913af964e12ca5da0c0b7e1c6f444d5
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:025eb58c69b9cf5992ecf2fccb29a3d36061cbcf62652a1c6878f0c4d767c1fb
        - registry.redhat.io/rhel9/postgresql-15@sha256:24fb4e7914a6e1464d015be9e5582cc4b9da224137408bd429e7ea4f391aa198
        version: 1.2.6
      entries:
      - name: rhdh-operator.v1.2.6
        version: 1.2.6
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.2
    - currentCSV: rhdh-operator.v1.3.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:5ce579edc10bde5c1c400104f7e0d9f86ca277cd565aaf3dd84b8636f5b949a8
          createdAt: "2025-02-10T18:12:52Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.3.5'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.3/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:5ce579edc10bde5c1c400104f7e0d9f86ca277cd565aaf3dd84b8636f5b949a8
        - registry.redhat.io/rhel9/postgresql-15@sha256:24fb4e7914a6e1464d015be9e5582cc4b9da224137408bd429e7ea4f391aa198
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:c044fed8485a62a5051ac726525da546936be22f268db05491bf81582e0e6194
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b4934c45ba595b36e65c73b734c58d1517e7f43b7cfc309ca8bc853733582f76
        version: 1.3.5
      entries:
      - name: rhdh-operator.v1.3.5
        version: 1.3.5
      - name: rhdh-operator.v1.3.4
        version: 1.3.4
      - name: rhdh-operator.v1.3.3
        version: 1.3.3
      - name: rhdh-operator.v1.3.1
        version: 1.3.1
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.3
    - currentCSV: rhdh-operator.v1.4.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:ae5b52c1b35e82f3c564402e367a2a2028e16806867bba0868643089ba7ed223
          createdAt: "2025-04-03T05:42:32Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.4.3'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.4/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:ae5b52c1b35e82f3c564402e367a2a2028e16806867bba0868643089ba7ed223
        version: 1.4.3
      entries:
      - name: rhdh-operator.v1.4.3
        version: 1.4.3
      - name: rhdh-operator.v1.4.2
        version: 1.4.2
      - name: rhdh-operator.v1.4.1
        version: 1.4.1
      - name: rhdh-operator.v1.4.0
        version: 1.4.0
      name: fast-1.4
    - currentCSV: rhdh-operator.v1.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:8fe167cc649a1c53994c376f3cf3eed8c6e318741a0c08e5aaf07cb0b74e7f8a
          createdAt: "2025-06-05T14:31:21Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.5.3'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.5/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:8fe167cc649a1c53994c376f3cf3eed8c6e318741a0c08e5aaf07cb0b74e7f8a
        - registry.redhat.io/rhel9/postgresql-15@sha256:74377f5a61dd7e3e0ba5d2fa880b7107025433a026396567c79eac927b4212bc
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:2ad8c87e82df27c2b6ceba94179d1f0a40fe775368663c1b5572611888e0e02b
        version: 1.5.3
      entries:
      - name: rhdh-operator.v1.5.3
        version: 1.5.3
      - name: rhdh-operator.v1.5.2
        version: 1.5.2
      - name: rhdh-operator.v1.5.1
        version: 1.5.1
      name: fast-1.5
    - currentCSV: rhdh-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:50f20300c81621221260d03d22b1442750e3f1d2e62f3e511d63925d84c4d1f9
          createdAt: "2025-08-27T18:43:12Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.6.5'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.6/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:50f20300c81621221260d03d22b1442750e3f1d2e62f3e511d63925d84c4d1f9
        - registry.redhat.io/rhel9/postgresql-15@sha256:5e225c1646251d2509b3a6e45fad23abfb11dd5c45d3224b59178f90deb38a70
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:177a7308f61a6b1ba21bc0f7bf550a93d4478e52479e2c94a2437f2152e3d8ff
        version: 1.6.5
      entries:
      - name: rhdh-operator.v1.6.5
        version: 1.6.5
      - name: rhdh-operator.v1.6.4
        version: 1.6.4
      - name: rhdh-operator.v1.6.3
        version: 1.6.3
      - name: rhdh-operator.v1.6.2
        version: 1.6.2
      - name: rhdh-operator.v1.6.1
        version: 1.6.1
      - name: rhdh-operator.v1.6.0
        version: 1.6.0
      name: fast-1.6
    - currentCSV: rhdh-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
          createdAt: "2025-09-12T06:25:11Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.7.1'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.7/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:62f32e50727c5006766a34daa5b6472b649cd9894f3fe9543b8ecc67e6760e8e
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
        version: 1.7.1
      entries:
      - name: rhdh-operator.v1.7.1
        version: 1.7.1
      - name: rhdh-operator.v1.7.0
        version: 1.7.0
      name: fast-1.7
    defaultChannel: fast
    packageName: rhdh
    provider:
      name: Red Hat Inc.
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-odlm
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operand-deployment-lifecycle-manager.v4.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandBindInfo",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "bindings": {
                    "public": {
                      "configmap": "mongodb-configmap",
                      "secret": "mongodb-secret"
                    }
                  },
                  "description": "Binding information that should be accessible to MongoDB adopters",
                  "operand": "mongodb-atlas-kubernetes",
                  "registry": "example-service"
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "services": [
                    {
                      "name": "jaeger",
                      "spec": {
                        "jaeger": {
                          "strategy": "allinone"
                        }
                      }
                    },
                    {
                      "name": "mongodb-atlas-kubernetes",
                      "spec": {
                        "atlasDeployment": {
                          "deploymentSpec": {
                            "name": "test-deployment"
                          },
                          "projectRef": {
                            "name": "my-project"
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "operators": [
                    {
                      "channel": "stable",
                      "installMode": "cluster",
                      "name": "jaeger",
                      "namespace": "default",
                      "packageName": "jaeger",
                      "sourceName": "community-operators",
                      "sourceNamespace": "openshift-marketplace"
                    },
                    {
                      "channel": "stable",
                      "name": "mongodb-atlas-kubernetes",
                      "namespace": "default",
                      "packageName": "mongodb-atlas-kubernetes",
                      "sourceName": "community-operators",
                      "sourceNamespace": "openshift-marketplace"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandRequest",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "requests": [
                    {
                      "operands": [
                        {
                          "name": "jaeger"
                        },
                        {
                          "name": "mongodb-atlas-kubernetes"
                        }
                      ],
                      "registry": "example-service"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, Monitoring, Logging & Tracing, Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
          createdAt: "2025-02-14T21:16:49Z"
          description: The Operand Deployment Lifecycle Manager provides a Kubernetes
            CRD-based API to manage the lifecycle of operands.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-odlm
          olm.skipRange: '>=1.2.0 <4.5.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/operand-deployment-lifecycle-manager
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OperandBindInfo is the Schema for the operandbindinfoes API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandBindInfo
            kind: OperandBindInfo
            name: operandbindinfos.operator.ibm.com
            version: v1alpha1
          - description: OperandConfig is the Schema for the operandconfigs API. Documentation
              For additional details regarding install parameters check https://ibm.biz/icpfs39install.
              License By installing this product you accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandConfig
            kind: OperandConfig
            name: operandconfigs.operator.ibm.com
            version: v1alpha1
          - description: OperandRegistry is the Schema for the operandregistries API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandRegistry
            kind: OperandRegistry
            name: operandregistries.operator.ibm.com
            version: v1alpha1
          - description: OperandRequest is the Schema for the operandrequests API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandRequest
            kind: OperandRequest
            name: operandrequests.operator.ibm.com
            version: v1alpha1
          - description: OperatorConfig is the Schema for the operatorconfigs API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.ibm.com
            version: v1alpha1
        description: |-
          # Introduction

          The operand is the instance managed by the operator. ODLM is used to manage the lifecycle of for a group of operands, compared with operator lifecycle manager, ODLM focus on the management of operands but not operators.

          # Details

          ### What are operands?

          Operator is a method of packaging, deploying and managing a Kubernetes application.
          Operands are the services and applications that Operator manage.

          For example, cert-manager operator deploys a cert-manager deployment, then the cert-manager-operator is an operator and the cert-manager deployment is an operand.

          The ODLM will have four CRDs:

          | Resource                 | Short Name | Description                                                                                |
          |--------------------------|------------|--------------------------------------------------------------------------------------------|
          | OperandRequest | opreq | It defines which operator/operand want to be installed in the cluster |
          | OperandRegistry | opreg | It defines the OLM information, like channel and catalog source, for each operator |
          | OperandConfig | opcon | It defines the parameters that should be used to install the operator's operand |
          | OperandBindInfo | opbi | It identifies secrets and/or configmaps that should be shared with requests |

          ### Goal

          1. A single entry point to manage a group of operands
          1. User can select a set of operands to install
          1. The install can be invoked through either OCP UI or CLI

          ## Prerequisites
          - [operator-sdk] version v1.3.0.
          - [go] version 1.15.7+
          - [oc] version v3.11+ or [kubectl] v1.11.3+
          - Access to an Openshift v4.3+ cluster

          ### Resources Required

          #### Minimum scheduling capacity

          | Software                   | Memory (MB) | CPU (cores) |
          | -------------------------- | ----------- | ----------- |
          | Operand Deployment Lifecycle Manager | 200          | 0.2        |
          | **Total**                  | 200         | 0.2         |

          ## Installing

          ### Create Operator NS, Group, Subscription

          ```yaml
          apiVersion: v1
          kind: Namespace
          metadata:
            name: odlm

          ---
          apiVersion: operators.coreos.com/v1alpha2
          kind: OperatorGroup
          metadata:
            name: operatorgroup
            namespace: odlm
          spec:
            targetNamespaces:
            - odlm

          ---
          apiVersion: v1
          data:
            namespaces: odlm
          kind: ConfigMap
          metadata:
            name: namespace-scope
            namespace: odlm

          ---
          apiVersion: operators.coreos.com/v1alpha1
          kind: Subscription
          metadata:
            name: operand-deployment-lifecycle-manager
            namespace: odlm
          spec:
            channel: v3.20
            name: ibm-odlm
            source: opencloud-operators
            sourceNamespace: olm
            config:
              env:
              - name: INSTALL_SCOPE
                value: namespaced
          END
          ```

          ## Configuration

          OperandRegistry defines the OLM information used for installation, like package name and catalog source, for each operator.

          Following is an example of the OperandRegistry CR:

          **NOTE:** The "spec.operators[*].name" parameter must be unique for each entry.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRegistry
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            operators:
            - name: jenkins [3]
              namespace: default [4]
              channel: alpha [5]
              packageName: jenkins-operator [6]
              scope: public [7]
              sourceName: community-operators [8]
              sourceNamespace: openshift-marketplace [9]
              installMode: cluster [10]
              installPlanApproval: Manual [11]
          ```

          The OperandRegistry Custom Resource (CR) lists OLM Operator information for operands that may be requested for installation and/or access by an application that runs in a namespace. The registry CR specifies:

          1. `name` of the OperandRegistry
          2. `namespace` of the OperandRegistry
          3. `name` is the name of the operator, which should be the same as the services name in the OperandConfig and OperandRequest.
          4. `namespace` defines the namespace where the operator and its CR will be deployed. (1) When InstallMode is `cluster`, the operator will be deployed into the `openshift-operators` namespace and the operator CRs will be deployed into the namespace this parameter defines. (2) When InstallMode is empty or set to `namespace`, it is the namespace where both operator and operator CR will be deployed.
          5. `channel` is the name of OLM channel that is subscribed for the operator.
          6. `packageName` is the name of the package in CatalogSource that is subscribed for the operator.
          7. (optional) `scope` is an indicator, either public or private, that dictates whether deployment can be requested from other namespaces (public) or only from the namespace of this OperandRegistry (private). The default value is private.
          8. `sourceName` is the name of the CatalogSource.
          9. `sourceNamespace` is the namespace of the CatalogSource.
          10. (optional) `installMode` is the install mode of the operator, can be either `namespace` (OLM one namespace) or `cluster` (OLM all namespaces). The default value is `namespace`. Operator is deployed in `openshift-operators` namespace when InstallMode is set to `cluster`.
          11. (optional) `installPlanApproval` is the approval mode for emitted installplan. The default value is `Automatic`.

          ### OperandConfig Spec

          OperandConfig defines the individual operand configuration. The OperandConfig Custom Resource (CR) defines the parameters for each operator that is listed in the OperandRegistry that should be used to install the operator instance by specifying an installation CR.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          Kind: OperandConfigs
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            services:
            - name: jenkins [3]
              spec: [4]
                jenkins:
                  port: 8081
          ```

          OperandConfig defines the individual operand deployment config:

          1. `name` of the OperandConfig
          2. `namespace` of the OperandConfig
          3. `name` is the name of the operator, which should be the same as the services name in the OperandRegistry and OperandRequest.
          4. `spec` defines a map. Its key is the kind name of the custom resource. Its value is merged to the spec field of custom resource. For more details, you can check the following topic **How does ODLM create the individual operator CR?**

          #### How does Operator create the individual operator CR

          Jenkins Operator has one CRD: Jenkins:

          The OperandConfig CR has

          ```yaml
          - name: jenkins
            spec:
              jenkins:
                service:
                  port: 8081
          ```

          The Jenkins Operator CSV has

          ```yaml
          apiVersion: operators.coreos.com/v1alpha1
          kind: ClusterServiceVersion
          metadata:
            annotations:
            alm-examples: |-
              [
              {
                "apiVersion":"jenkins.io/v1alpha2",
                "kind":"Jenkins",
                "metadata": {
                  "name":"example"
                },
                "spec":{
                  ...
                  "service":{"port":8080,"type":"ClusterIP"},
                  ...
                }
              }
            ]
          ```

          The ODLM will deep merge the OperandConfig CR spec and Jenkins Operator CSV alm-examples to create the Jenkins CR.

          ```yaml
          apiVersion: jenkins.io/v1alpha2
          kind: Jenkins
          metadata:
            name: example
          spec:
            ...
            service:
              port: 8081
              type: ClusterIP
            ...
          ```

          For day2 operations, the ODLM will patch the OperandConfigs CR spec to the existing Jenkins CR.

          ### OperandRequest Spec

          OperandRequest defines which operator/operand you want to install in the cluster.

          **NOTE:** OperandRequest custom resource is used to trigger a deployment for Operators and Operands.

          #### OperandRequest sample to create custom resource via OperandConfig

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRequest
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            requests:
            - registry: example-service [3]
              registryNamespace: example-service-ns [4]
              operands: [5]
              - name: jenkins [6]
                bindings: [7]
                  public:
                    secret: jenkins-operator-credential [8]
                    configmap: jenkins-operator-base-configuration [9]
          ```

          1. `name` of the OperandRequest
          2. `namespace` of the OperandRequest
          3. `registry` identifies the name of the OperandRegistry CR from which this operand deployment is being requested.
          4. (optional) `registryNamespace` identifies the namespace in which the OperandRegistry CR is defined. **Note:** If the `registryNamespace` is not specified then it is assumed that the OperandRegistry CR is in the current (OperandRequest's) namespace.
          5. `operands` in the CR is a list of operands.
          6. `name` of operands in the CR must match a name specification in an OperandRegistry's CR.
          7. (optional) The `bindings` of the operands is a map to get and rename the secret and/or configmap from the provider and create them in the requester's namespace. If the requester wants to rename the secret and/or configmap, they need to know the key of the binding in the OperandBindInfo. If the key of the bindings map is prefixed with public, it means the secret and/or configmap can be shared with the requester in the other namespace. If the key of the bindings map is prefixed with private, it means the secret and/or configmap can only be shared within its own namespace.
          8. (optional) `secret` names a secret that should be created in the requester's namespace with formatted data that can be used to interact with the service.
          9. (optional) `configmap` names a configmap that should be created in the requester's namespace with formatted data that can be used to interact with the service.

          #### OperandRequest sample to create custom resource via OperandRequest

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRequest
          metadata:
            name: example-service
            namespace: example-service-ns
          spec:
            requests:
            - registry: example-service
              registryNamespace: example-service-ns
              operands:
              - name: jenkins
                kind: Jenkins [1]
                apiVersion: "jenkins.io/v1alpha2" [2]
                instanceName: "example" [3]
                spec: [4]
                  service:
                    port: 8081
          ```

          1. `kind` of the target custom resource. If it is set in the operand item, ODLM will create CR according to OperandRequest only and get rid of the alm-example and OperandConfig.
          2. `apiVersion` of the target custom resource.
          3. `instanceName` is the name of the custom resource. If `instanceName` is not set, the name of the custom resource will be created with the name of the OperandRequest as a prefix.
          4. `spec` is the spec field of the target CR.

          ### OperandBindInfo Spec

          The ODLM will use the OperandBindInfo to copy the generated secret and/or configmap to a requester's namespace when a service is requested with the OperandRequest CR. An example specification for an OperandBindInfo CR is shown below.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandBindInfo
          metadata:
            name: publicjenkinsbinding [1]
            namespace: example-service-ns [2]
          spec:
            operand: jenkins [3]
            registry: example-service [4]
            description: "Binding information that should be accessible to jenkins adopters" [5]
            bindings: [6]
              public:
                secret: jenkins-operator-credentials-example [7]
                configmap: jenkins-operator-base-configuration-example [8]
          ```

          Fields in this CR are described below.

          1. `name` of the OperandBindInfo
          2. `namespace` of the OperandBindInfo
          3. The `operand` should be the the individual operator name.
          4. The `registry` section must match the name in the OperandRegistry CR in the current namespace.
          5. `description` is used to add a detailed description of a service.
          6. The `bindings` section is used to specify information about the access/configuration data that is to be shared. If the key of the bindings map is prefixed with public, it means the secret and/or configmap can be shared with the requester in the other namespace. If the key of the bindings map is prefixed with private, it means the secret and/or configmap can only be shared within its own namespace. If the key of the bindings map is prefixed with protected, it means the secret and/or configmap can only be shared if it is explicitly declared in the OperandRequest.
          7. The `secret` field names an existing secret, if any, that has been created and holds information that is to be shared with the requester.
          8. The `configmap` field identifies a configmap object, if any, that should be shared with the requester

          ODLM will use the OperandBindInfo CR to pass information to an adopter when they create a OperandRequest to access the service, assuming that both have compatible scopes. ODLM will copy the information from the shared service's "OperandBindInfo.bindinfo[].secret" and/or "OperandBindInfo.bindinfo[].configmap" to the requester namespace.

          **NOTE:** If in the OperandRequest, there is no secret and/or configmap name specified in the bindings or no bindings field in the element of operands, ODLM will copy the secret and/or configmap to the requester's namespace and rename them to the name of the OperandBindInfo + secret/configmap name.

          ### E2E Use Case

          1. User installs ODLM from OLM

              The ODLM will automatically generate two default CRD CRs, since OperandRegistry and OperandConfigs CRs don't define the state, so it should be fine.

          2. **Optional:** Users can update the OperandConfigs CR with their own parameter values

              Users can skip this step if the default configurations match the requirement.

          3. User creates the OperandRequest CR from OLM

              This tells ODLM that users want to install some of the individual operator/operand.

          4. The rest works will be done by OLM and ODLM

              Finally, users will get what they want.

          ### Operator/Operand Upgrade

          - For operator/operand upgrade, you only need to publish your operator OLM to your operator channel, and OLM will handle the upgrade automatically.
          - If there are major version, then you may want to update `channel` in `OperandRegistry` to trigger upgrade.

          ### Limitations

          ODLM depends on NamespaceScope Operator

          ### SecurityContextConstraints Requirements

          The Operand Deployment Lifecycle Manager supports running under the OpenShift Container Platform default restricted security context constraints.
        displayName: Operand Deployment Lifecycle Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Operand Deployment Lifecycle Manager Project
          url: https://github.com/IBM/operand-deployment-lifecycle-manager
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bb074bdf89cbc7c703fadb17639608599ab4f66db9be5617e5f79a0550676790
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d4dff1a8968c18d984e16c4c9f1517c3a10e23a11bf166c64d734773324d4c9d
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
        - registry.marketplace.redhat.com/cpopen/operand-deployment-lifecycle-manager-bundle@sha256:68f71f2767a40ca9333f0a670e07493e8a8c13a5c7ca877ea5bd8a544b352d9a
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:27195e9758afdf8521971866fd60ea3dafed91a146c3c4165876e5ea476a4622
        version: 4.5.0
      entries:
      - name: operand-deployment-lifecycle-manager.v4.5.0
        version: 4.5.0
      name: v4.5
    defaultChannel: v4.5
    packageName: ibm-odlm
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: egressip-ipam-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: egressip-ipam-operator.v1.2.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "EgressIPAM",
                "metadata": {
                  "name": "egressipam-aws"
                },
                "spec": {
                  "cidrAssignments": [
                    {
                      "CIDR": "10.0.128.0/19",
                      "labelValue": "us-west-2a"
                    },
                    {
                      "CIDR": "10.0.160.0/19",
                      "labelValue": "us-west-2b",
                      "reservedIPs": [
                        "10.0.160.126"
                      ]
                    },
                    {
                      "CIDR": "10.0.192.0/19",
                      "labelValue": "us-west-2c"
                    }
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "topologyLabel": "failure-domain.beta.kubernetes.io/zone"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "EgressIPAM",
                "metadata": {
                  "name": "egressipam-baremetal"
                },
                "spec": {
                  "cidrAssignments": [
                    {
                      "CIDR": "192.169.0.0/24",
                      "labelValue": "true"
                    }
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "topologyLabel": "egressGateway"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Networking, Cloud Provider, Security
          certified: "false"
          containerImage: quay.io/redhat-cop/egressip-ipam-operator@sha256:d6129bd287b87696a23e0da64dcd598958f779dc88ae33b46b3c909dbab80adf
          createdAt: "2022-07-25T15:07:22Z"
          description: This operator provides automation around managing egressIPs.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: egressip-ipam-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/egressip-ipam-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgressIPAM is the Schema for the egressipams API
            displayName: Egress IPAM
            kind: EgressIPAM
            name: egressipams.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator automates the assignment of egressIPs to namespaces.
          Namespaces can opt in to receiving one or more egressIPs with the following annotation `egressip-ipam-operator.redhat-cop.io/egressipam:<egressIPAM>`, where `egressIPAM` is the CRD that controls how egressIPs are assigned.
          IPs assigned to the namespace can be looked up in the following annotation: `egressip-ipam-operator.redhat-cop.io/egressips`.

          EgressIP assignments is managed by the EgressIPAM CRD. Here is an example of it:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: EgressIPAM
          metadata:
            name: example-egressipam
          spec:
            cidrAssignments:
              - labelValue: "true"
                CIDR: 192.169.0.0/24
                reservedIPs:
                - "192.159.0.5"
            topologyLabel: egressGateway
            nodeSelector:
              node-role.kubernetes.io/worker: ""
          ```

          This EgressIPAM specifies that all nodes that comply with the specified node selector and that also have labels `egressGateway: "true"` will be assigned egressIP from the specified CIDR.

          Note that the `cidrAssigments` field is an array and therefore, multiple groups of nodes can be identified with the labelValue and different CIDRs can be assigned to them. This is usually not necessary on a bare metal deployment.

          In the bare metal scenario, when this egressCRD is created, all the `hostsubnet` relative to the nodes selected by this EgressIPAM will be update to have the EgressCIDRs field equal to the specified CIDR (see below for cloud provider scenarios).

          When a namespace is created with the opt-in annotation: `egressip-ipam-operator.redhat-cop.io/egressipam=<egressIPAM>`, an available egressIP is selected from the CIDR and assigned to the namespace.The `netnamespace` associated with this namespace is updated to use that egressIP.

          It is possible to specify a set of reserved IPs. These IPs must belong to the CIDR and will never be assigned.

          ## Passing EgressIPs as input

          The normal mode of operation of this operator is to pick a random IP from the configured CIDR. However, it also supports a scenario where egressIPs are picked by an external process and passed as input.

          In this case IPs must me passed as an annotation to the namespace, like this: `egressip-ipam-operator.redhat-cop.io/egressips=IP1,IP2...`. The value of the annotation is a comma separated array of ip with no spaces.

          There must be exactly one IP per CIDR defined in the referenced egressIPAM. Moreover, each IP must belong to the corresponding CIDR. Because this situation can lead to inconsistencies, failing to have correspondence between IPs in the namespace annotation and CIDRs in the egressIPAM CR will cause the operator to error out and stop processing all the namespaces associated with the given EgressIPAM CR.

          It also also a responsibility of the progress picking the IPs to ensure that those IPs are actually available.

          ## Assumptions

          The following assumptions apply when using this operator:

          1. If multiple EgressIPAMs are defined, their selected nodes MUST NOT overlap.
          2. When using a cloud provider the topology label MUST be `failure-domain.beta.kubernetes.io/zone`.

          ## Support for AWS

          In AWS as well as other cloud providers, one cannot freely assign IPs to machines. Additional steps need to be performed in this case. Considering this EgressIPAM

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: EgressIPAM
          metadata:
            name: egressipam-aws
          spec:
            cidrAssignments:
              - labelValue: "eu-central-1a"
                CIDR: 10.0.128.0/20
              - labelValue: "eu-central-1b"
                CIDR: 10.0.144.0/20
              - labelValue: "eu-central-1c"
                CIDR: 10.0.160.0/20
            topologyLabel: failure-domain.beta.kubernetes.io/zone
            nodeSelector:
              matchLabels:
                node-role.kubernetes.io/worker: ""
          ```

          When a namespace with the opt-in annotation is created, the following happens:

          1. for each of the CIDRs, an available IP is selected and assigned to the namespace
          2. the relative `netnamespace` is update to reflect the assignment (multiple IPs will be assigned in this case).
          3. one node per zone is selected to carry the egressIP
          4. the relative aws machines are assigned the additional IP on the main interface (support for secondary interfaces in not available)
          5. the relative `hostsubnets` are updated to reflect the assigned IP, the `egressIP` field is updated.
        displayName: Egressip Ipam Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - egressIPs
        - high availability
        - firewall support
        - traffic management
        - ipam
        links:
        - name: Egressip Ipam Operator
          url: https://github.com/redhat-cop/egressip-ipam-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/egressip-ipam-operator:v0.1.2
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/egressip-ipam-operator@sha256:d6129bd287b87696a23e0da64dcd598958f779dc88ae33b46b3c909dbab80adf
        version: 1.2.4
      entries:
      - name: egressip-ipam-operator.v1.2.4
        version: 1.2.4
      - name: egressip-ipam-operator.v1.2.3
        version: 1.2.3
      - name: egressip-ipam-operator.v1.2.2
        version: 1.2.2
      - name: egressip-ipam-operator.v1.2.1
        version: 1.2.1
      - name: egressip-ipam-operator.v1.2.0
        version: 1.2.0
      - name: egressip-ipam-operator.v1.1.0
        version: 1.1.0
      - name: egressip-ipam-operator.v1.0.8
        version: 1.0.8
      - name: egressip-ipam-operator.v1.0.7
        version: 1.0.7
      - name: egressip-ipam-operator.v1.0.6
        version: 1.0.6
      - name: egressip-ipam-operator.v1.0.5
        version: 1.0.5
      - name: egressip-ipam-operator.v1.0.4
        version: 1.0.4
      - name: egressip-ipam-operator.v1.0.3
        version: 1.0.3
      - name: egressip-ipam-operator.v1.0.2
        version: 1.0.2
      - name: egressip-ipam-operator.v1.0.1
        version: 1.0.1
      - name: egressip-ipam-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: egressip-ipam-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: hive-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hive-operator.v1.2.4953-c248a51
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:c248a51ae1
          createdAt: "2025-09-22T13:41:19Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeploymentCustomization is the Schema for clusterdeploymentcustomizations
              API.
            displayName: ClusterDeploymentCustomization
            kind: ClusterDeploymentCustomization
            name: clusterdeploymentcustomizations.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: |-
              ClusterPool represents a pool of clusters that should be kept ready to be given out to users. Clusters are removed
              from the pool once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: |-
              MachinePoolNameLease is the Schema for the MachinePoolNameLeases API. This resource is mostly empty
              as we're primarily relying on the name to determine if a lease is available.
              Note that not all cloud providers require the use of a lease for naming, at present this
              is only required for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * IBM Cloud
          * Red Hat OpenStack
          * vSphere

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Troubleshooting](https://github.com/openshift/hive/blob/master/docs/troubleshooting.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
              logLevel: debug
              targetNamespace: hive
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:c248a51ae1
        version: 1.2.4953-c248a51
      entries:
      - name: hive-operator.v1.2.4953-c248a51
        version: 1.2.4953-c248a51
      - name: hive-operator.v1.2.4928-76a65b6
        version: 1.2.4928-76a65b6
      - name: hive-operator.v1.2.4902-4929f16
        version: 1.2.4902-4929f16
      - name: hive-operator.v1.2.4896-a29a238
        version: 1.2.4896-a29a238
      - name: hive-operator.v1.2.4868-45ae912
        version: 1.2.4868-45ae912
      - name: hive-operator.v1.2.4863-e9ad0fa
        version: 1.2.4863-e9ad0fa
      - name: hive-operator.v1.2.4774-87bff59
        version: 1.2.4774-87bff59
      - name: hive-operator.v1.2.4766-6a82857
        version: 1.2.4766-6a82857
      - name: hive-operator.v1.2.4746-d25e0db
        version: 1.2.4746-d25e0db
      - name: hive-operator.v1.2.4738-b19773e
        version: 1.2.4738-b19773e
      - name: hive-operator.v1.2.4731-41f153f
        version: 1.2.4731-41f153f
      - name: hive-operator.v1.2.4703-02f3f00
        version: 1.2.4703-02f3f00
      - name: hive-operator.v1.2.4693-ec3bc72
        version: 1.2.4693-ec3bc72
      - name: hive-operator.v1.2.4686-553c6c2
        version: 1.2.4686-553c6c2
      - name: hive-operator.v1.2.4682-406d6ca
        version: 1.2.4682-406d6ca
      - name: hive-operator.v1.2.4672-af54e2f
        version: 1.2.4672-af54e2f
      - name: hive-operator.v1.2.4668-3ccd44d
        version: 1.2.4668-3ccd44d
      - name: hive-operator.v1.2.4664-00f84cc
        version: 1.2.4664-00f84cc
      - name: hive-operator.v1.2.4658-c13e207
        version: 1.2.4658-c13e207
      - name: hive-operator.v1.2.4654-3268a18
        version: 1.2.4654-3268a18
      - name: hive-operator.v1.2.4650-531d549
        version: 1.2.4650-531d549
      - name: hive-operator.v1.2.4644-365f074
        version: 1.2.4644-365f074
      - name: hive-operator.v1.2.4638-56d6d64
        version: 1.2.4638-56d6d64
      - name: hive-operator.v1.2.4625-389d717
        version: 1.2.4625-389d717
      - name: hive-operator.v1.2.4615-52100dd
        version: 1.2.4615-52100dd
      - name: hive-operator.v1.2.4605-b41ca3f
        version: 1.2.4605-b41ca3f
      - name: hive-operator.v1.2.4584-95efcb7
        version: 1.2.4584-95efcb7
      - name: hive-operator.v1.2.4574-1ec27ad
        version: 1.2.4574-1ec27ad
      - name: hive-operator.v1.2.4568-b3eca52
        version: 1.2.4568-b3eca52
      - name: hive-operator.v1.2.4564-bdf9d08
        version: 1.2.4564-bdf9d08
      - name: hive-operator.v1.2.4552-bf2d689
        version: 1.2.4552-bf2d689
      - name: hive-operator.v1.2.4532-cddd4f9
        version: 1.2.4532-cddd4f9
      - name: hive-operator.v1.2.4511-1fce770
        version: 1.2.4511-1fce770
      - name: hive-operator.v1.2.4500-b1ac27b
        version: 1.2.4500-b1ac27b
      - name: hive-operator.v1.2.4485-47df0d9
        version: 1.2.4485-47df0d9
      - name: hive-operator.v1.2.4469-7be6433
        version: 1.2.4469-7be6433
      - name: hive-operator.v1.2.4463-d7ead60
        version: 1.2.4463-d7ead60
      - name: hive-operator.v1.2.4461-97686bb
        version: 1.2.4461-97686bb
      - name: hive-operator.v1.2.4459-f910c9b
        version: 1.2.4459-f910c9b
      - name: hive-operator.v1.2.4451-83aedb9
        version: 1.2.4451-83aedb9
      - name: hive-operator.v1.2.4446-e11a23b
        version: 1.2.4446-e11a23b
      - name: hive-operator.v1.2.4441-fdbfaf0
        version: 1.2.4441-fdbfaf0
      - name: hive-operator.v1.2.4436-44323bc
        version: 1.2.4436-44323bc
      - name: hive-operator.v1.2.4428-ceaafa8
        version: 1.2.4428-ceaafa8
      - name: hive-operator.v1.2.4426-fb27114
        version: 1.2.4426-fb27114
      - name: hive-operator.v1.2.4420-fe0e302
        version: 1.2.4420-fe0e302
      - name: hive-operator.v1.2.4418-eb5403f
        version: 1.2.4418-eb5403f
      - name: hive-operator.v1.2.4416-fba8d52
        version: 1.2.4416-fba8d52
      - name: hive-operator.v1.2.4408-9c5806a
        version: 1.2.4408-9c5806a
      - name: hive-operator.v1.2.4404-410fed9
        version: 1.2.4404-410fed9
      - name: hive-operator.v1.2.4400-8c54fc9
        version: 1.2.4400-8c54fc9
      - name: hive-operator.v1.2.4386-6f1fd3e
        version: 1.2.4386-6f1fd3e
      - name: hive-operator.v1.2.4382-0e95d5b
        version: 1.2.4382-0e95d5b
      - name: hive-operator.v1.2.4366-df65e7e
        version: 1.2.4366-df65e7e
      - name: hive-operator.v1.2.4360-de169dd
        version: 1.2.4360-de169dd
      - name: hive-operator.v1.2.4355-e5f809f
        version: 1.2.4355-e5f809f
      name: alpha
    - currentCSV: hive-operator.v2.5.3516-a2ed9b3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.5.3516-a2ed9b3
          createdAt: "2022-06-03T13:24:12Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:v2.5.3516-a2ed9b3
        version: 2.5.3516-a2ed9b3
      entries:
      - name: hive-operator.v2.5.3516-a2ed9b3
        version: 2.5.3516-a2ed9b3
      - name: hive-operator.v2.5.3514-4052426
        version: 2.5.3514-4052426
      - name: hive-operator.v2.5.3508-6cb94c6
        version: 2.5.3508-6cb94c6
      - name: hive-operator.v2.5.3506-f42617c
        version: 2.5.3506-f42617c
      - name: hive-operator.v2.5.3495-5cfdd44
        version: 2.5.3495-5cfdd44
      - name: hive-operator.v2.5.3491-6feded1
        version: 2.5.3491-6feded1
      - name: hive-operator.v2.5.3489-6dec9c3
        version: 2.5.3489-6dec9c3
      name: mce-2.0
    - currentCSV: hive-operator.v2.3.3039-60b8a9a
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.3.3039-60b8a9a
          createdAt: "2022-08-15T17:23:39Z"
          description: OpenShift cluster provisioning and management at scale.
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:v2.3.3039-60b8a9a
        version: 2.3.3039-60b8a9a
      entries:
      - name: hive-operator.v2.3.3039-60b8a9a
        version: 2.3.3039-60b8a9a
      name: ocm-2.3
    - currentCSV: hive-operator.v2.4.3231-ba51985
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.4.3231-ba51985
          createdAt: "2022-08-15T11:53:28Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:v2.4.3231-ba51985
        version: 2.4.3231-ba51985
      entries:
      - name: hive-operator.v2.4.3231-ba51985
        version: 2.4.3231-ba51985
      - name: hive-operator.v2.4.3227-74831db
        version: 2.4.3227-74831db
      - name: hive-operator.v2.4.3222-d68346c
        version: 2.4.3222-d68346c
      - name: hive-operator.v2.4.3206-b0714a7
        version: 2.4.3206-b0714a7
      - name: hive-operator.v2.4.3203-672cfe7
        version: 2.4.3203-672cfe7
      - name: hive-operator.v2.4.3201-8c50440
        version: 2.4.3201-8c50440
      - name: hive-operator.v2.4.3199-cce4cb9
        version: 2.4.3199-cce4cb9
      - name: hive-operator.v2.4.3197-6c9c8e4
        version: 2.4.3197-6c9c8e4
      - name: hive-operator.v2.4.3194-f82b7d8
        version: 2.4.3194-f82b7d8
      - name: hive-operator.v2.4.3192-fae0111
        version: 2.4.3192-fae0111
      - name: hive-operator.v2.4.3189-026033b
        version: 2.4.3189-026033b
      - name: hive-operator.v2.4.3180-7299056
        version: 2.4.3180-7299056
      - name: hive-operator.v1.2.3179-0756b70e6
        version: 1.2.3179-0756b70e6
      - name: hive-operator.v1.1.16
        version: 1.1.16
      - name: hive-operator.v1.1.15
        version: 1.1.15
      - name: hive-operator.v1.1.14
        version: 1.1.14
      - name: hive-operator.v1.1.13
        version: 1.1.13
      - name: hive-operator.v1.1.12
        version: 1.1.12
      - name: hive-operator.v1.1.11
        version: 1.1.11
      - name: hive-operator.v1.1.10
        version: 1.1.10
      - name: hive-operator.v1.1.9
        version: 1.1.9
      - name: hive-operator.v1.1.8
        version: 1.1.8
      - name: hive-operator.v1.1.7
        version: 1.1.7
      - name: hive-operator.v1.1.6
        version: 1.1.6
      name: ocm-2.4
    defaultChannel: alpha
    packageName: hive-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-admin-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-admin-operator.v2024.10.2-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": null
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:ed279ff5416cce855118f66a73d55385aba26af19735eed052010ba653f64b75
          createdAt: "2024-10-25T01:47:18Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2024.10.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2023.9.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:7c4868db021beaffa762eb0de43739bbc865a4bf5900f1d17f78caed8612c073
        - registry.connect.redhat.com/ca/uma-operator@sha256:ed279ff5416cce855118f66a73d55385aba26af19735eed052010ba653f64b75
        version: 2024.10.2-12
      entries:
      - name: uma-admin-operator.v2024.10.2-12
        version: 2024.10.2-12
      name: "2024.10"
    - currentCSV: uma-admin-operator.v2025.1.3-33
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.1.1.33,990033",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": null
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:6a932a065add9f36f9615a93037315603f2d9cd8a2e3c64b1b1fc390bb4f1dcf
          createdAt: "2025-02-18T05:34:34Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.1.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2023.9.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.1.1.26,990026"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:6a932a065add9f36f9615a93037315603f2d9cd8a2e3c64b1b1fc390bb4f1dcf
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:ae91a28d57c8a359f81757e4bf79472f2ac189ac0ae3ee0d87e8ff9c433f1ccc
        version: 2025.1.3-33
      entries:
      - name: uma-admin-operator.v2025.1.3-33
        version: 2025.1.3-33
      name: "2025.1"
    - currentCSV: uma-admin-operator.v2025.2.1-25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.2.1.25,990025",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverCollectorCert": "",
                    "serverCollectorKey": "",
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:d201790c2d9ded6bc7a1c5f68a267574e7b5d39c39ce2f98e20ef630b0f7365f
          createdAt: "2025-03-05T05:51:02Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.2.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2023.9.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.1.1.26,990026"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:d201790c2d9ded6bc7a1c5f68a267574e7b5d39c39ce2f98e20ef630b0f7365f
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:8bcf0dac889f38a0bf77eb9babe3f726e3b2564d5988c05bdc5c2d91ae189c61
        version: 2025.2.1-25
      entries:
      - name: uma-admin-operator.v2025.2.1-25
        version: 2025.2.1-25
      name: "2025.2"
    - currentCSV: uma-admin-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:4c2a7f68a47428f3e386107254f74fa8c34e477eb40fc17c173df92f3c66c559
          createdAt: "2025-04-16T06:55:39Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.4.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.4.2.5,990005"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:4c2a7f68a47428f3e386107254f74fa8c34e477eb40fc17c173df92f3c66c559
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        version: 2025.4.5-5
      entries:
      - name: uma-admin-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-admin-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:514b953eaf0959bc120f80bdc0a2e19d8f059d3fe2883a4f6843a398470a56b1
          createdAt: "2025-07-24T14:54:21Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.5.2.12,990012"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:514b953eaf0959bc120f80bdc0a2e19d8f059d3fe2883a4f6843a398470a56b1
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        version: 2025.5.4-12
      entries:
      - name: uma-admin-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-admin-operator.v2025.7.3-21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:188a2b8d2255ded6141a4b795f1b6d7bd98017b4fd57c9c1a1d38fc701358057
          createdAt: "2025-08-23T09:00:37Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.3
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.7.1.23,990023"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:188a2b8d2255ded6141a4b795f1b6d7bd98017b4fd57c9c1a1d38fc701358057
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.3-21
      entries:
      - name: uma-admin-operator.v2025.7.3-21
        version: 2025.7.3-21
      name: "2025.7"
    - currentCSV: uma-admin-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:0cfb63c62258e54d67e7baca874003a04477f30bcf7f56861f565ad068070580
          createdAt: "2025-08-15T03:20:28Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-admin-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.8.1.9,990009"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:0cfb63c62258e54d67e7baca874003a04477f30bcf7f56861f565ad068070580
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        version: 2025.8.2-9
      entries:
      - name: uma-admin-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-admin-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
          createdAt: "2025-09-24T12:23:56Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.12,990012"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-admin-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-admin-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:7b93de818fa5160e72f878b0350d13fc1d5943eac95b0794d3d4f930cf21b59a
          createdAt: "2025-07-30T08:07:39Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-admin-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "99.99.0.serp_F152813,19"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:7b93de818fa5160e72f878b0350d13fc1d5943eac95b0794d3d4f930cf21b59a
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        version: 2025.7.2-19
      entries:
      - name: uma-admin-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-admin-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1cf3909dd2caeabc53209d6f8565dcb6c3562b9c6eb715dda8ba1953f264668e
          createdAt: "2025-09-11T11:16:06Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.8,990008"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:1cf3909dd2caeabc53209d6f8565dcb6c3562b9c6eb715dda8ba1953f264668e
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-admin-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-admin-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
          createdAt: "2025-09-24T12:23:56Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.12,990012"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-admin-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-admin-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: octavia-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: octavia-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "octavia.openstack.org/v1beta1",
                "kind": "Octavia",
                "metadata": {
                  "name": "octavia"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "octavia",
                  "databaseInstance": "openstack",
                  "lbMgmtNetwork": {
                    "availabilityZones": [
                      "zone-1"
                    ]
                  },
                  "octaviaAPI": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "octavia",
                    "databaseInstance": "openstack",
                    "persistenceDatabaseAccount": "octavia-persistence",
                    "preserveJobs": false,
                    "secret": "osp-secret",
                    "serviceAccount": "octavia",
                    "serviceUser": "octavia"
                  },
                  "octaviaHealthManager": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "octavia",
                    "databaseInstance": "openstack",
                    "networkAttachments": [
                      "octavia"
                    ],
                    "persistenceDatabaseAccount": "octavia-persistence",
                    "preserveJobs": false,
                    "role": "healthmanager",
                    "secret": "osp-secret",
                    "serviceAccount": "octavia",
                    "serviceUser": "octavia"
                  },
                  "octaviaHousekeeping": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "octavia",
                    "databaseInstance": "openstack",
                    "networkAttachments": [
                      "octavia"
                    ],
                    "persistenceDatabaseAccount": "octavia-persistence",
                    "preserveJobs": false,
                    "role": "housekeeping",
                    "secret": "osp-secret",
                    "serviceAccount": "octavia",
                    "serviceUser": "octavia"
                  },
                  "octaviaWorker": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "octavia",
                    "databaseInstance": "openstack",
                    "networkAttachments": [
                      "octavia"
                    ],
                    "persistenceDatabaseAccount": "octavia-persistence",
                    "preserveJobs": false,
                    "role": "worker",
                    "secret": "osp-secret",
                    "serviceAccount": "octavia",
                    "serviceUser": "octavia"
                  },
                  "persistenceDatabaseAccount": "octavia-persistence",
                  "preserveJobs": false,
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "osp-secret",
                  "serviceUser": "octavia"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:01Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: Octavia Amphora Controller
            kind: OctaviaAmphoraController
            name: octaviaamphoracontrollers.octavia.openstack.org
            version: v1beta1
          - description: OctaviaAPI is the Schema for the octaviaapis API
            displayName: Octavia API
            kind: OctaviaAPI
            name: octaviaapis.octavia.openstack.org
            version: v1beta1
          - kind: OctaviaRsyslog
            name: octaviarsyslogs.octavia.openstack.org
            version: v1beta1
          - description: Octavia is the Schema for the octavia API
            displayName: Octavia
            kind: Octavia
            name: octavias.octavia.openstack.org
            version: v1beta1
        description: Octavia Operator
        displayName: Octavia Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Load-Balancer
        - Octavia
        links:
        - name: Octavia Operator
          url: https://github.com/openstack-k8s-operators/octavia-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/octavia-rhel9-operator@sha256:ddd280ec099cde63e38bd964d7f0608c1c34e5c9f6b140f86dde2126b0357ef5
        - registry.redhat.io/rhoso/openstack-octavia-health-manager-rhel9@sha256:7d305af6e9e6c636510186dd80037849d19bb1e6d2b2507c025d1d0d9254874a
        - registry.redhat.io/rhoso/openstack-octavia-worker-rhel9@sha256:0919ddff6421430f8ad24cb21a89c65f02f98df424267f7e436c97d555112e2a
        - registry.redhat.io/ubi9/httpd-24@sha256:dbfbe3383e12d63df7178e4cdd5b1e6e51a84c781041b9192026ea6edb72fbf3
        - registry.redhat.io/rhoso/openstack-octavia-api-rhel9@sha256:ae9aae94c2523933a346bdade16158b4302b3a0bee93dec89bae5d15606c8dfb
        - registry.redhat.io/rhoso/openstack-octavia-housekeeping-rhel9@sha256:d54868077feb205199b76e83fc9220d9b4215dbeba037cf13fb68dc6577a07dc
        version: 1.0.6
      entries:
      - name: octavia-operator.v1.0.6
        version: 1.0.6
      - name: octavia-operator.v1.0.5
        version: 1.0.5
      - name: octavia-operator.v1.0.4
        version: 1.0.4
      - name: octavia-operator.v1.0.3
        version: 1.0.3
      - name: octavia-operator.v1.0.2
        version: 1.0.2
      - name: octavia-operator.v1.0.1
        version: 1.0.1
      - name: octavia-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: octavia-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: gatekeeper-operator-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gatekeeper-operator-product.v3.11.2-0.1725401426.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel8-operator@sha256:c8894e4d02fb2c822ab2e02e489dd9b67c953d96bd3b96a8305219a2245dfd99
          createdAt: 21 Mar 2024, 01:59
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.11.0
          olm.substitutesFor: gatekeeper-operator-product.v3.11.2-0.1721233953.p
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/gatekeeper/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the Schema for the gatekeepers API
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |-
          Gatekeeper allows administrators to detect and reject non-compliant commits to an infrastructure-as-code system's source-of-truth. This strengthens compliance efforts and prevents a bad state from slowing down the organization.

          ## Support & Troubleshooting
          Gatekeeper is available and supported as part of a valid Red Hat software subscription when it is used with Red Hat Advanced Cluster Management for Kubernetes. Troubleshooting information is available in the Gatekeeper documentation.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/gatekeeper/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - registry.redhat.io/gatekeeper/gatekeeper-rhel8-operator@sha256:c8894e4d02fb2c822ab2e02e489dd9b67c953d96bd3b96a8305219a2245dfd99
        - registry.redhat.io/gatekeeper/gatekeeper-rhel8@sha256:701b165def9e4b92d91b8eac4917721e76e634d2159b8efc69bef23e10b51b9d
        version: 3.11.2+0.1725401426.p
      entries:
      - name: gatekeeper-operator-product.v3.11.2-0.1725401426.p
        version: 3.11.2+0.1725401426.p
      - name: gatekeeper-operator-product.v3.11.2-0.1721233953.p
        version: 3.11.2+0.1721233953.p
      - name: gatekeeper-operator-product.v3.11.2-0.1718224960.p
        version: 3.11.2+0.1718224960.p
      - name: gatekeeper-operator-product.v3.11.2
        version: 3.11.2
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.11"
    - currentCSV: gatekeeper-operator-product.v3.14.3-0.1746550072.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e47f0e9da800eb0cf4a0da6af98986508a35798a41581158ee7dbc977cb651dc
          createdAt: 16 Jan 2025, 06:57
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.14.3
          olm.substitutesFor: gatekeeper-operator-product.v3.14.3-0.1744033158.p
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the Schema for the gatekeepers API
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |-
          Gatekeeper allows administrators to detect and reject non-compliant commits to an infrastructure-as-code system's source-of-truth. This strengthens compliance efforts and prevents a bad state from slowing down the organization.

          ## Support & Troubleshooting
          Gatekeeper is available and supported as part of a valid Red Hat software subscription when it is used with Red Hat Advanced Cluster Management for Kubernetes. Troubleshooting information is available in the Gatekeeper documentation.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e47f0e9da800eb0cf4a0da6af98986508a35798a41581158ee7dbc977cb651dc
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:6255f7a673e6989617176842d8f0cc897e4630b5283e812a7a2d52dbb2a1f335
        version: 3.14.3+0.1746550072.p
      entries:
      - name: gatekeeper-operator-product.v3.14.3-0.1746550072.p
        version: 3.14.3+0.1746550072.p
      - name: gatekeeper-operator-product.v3.14.3-0.1744033158.p
        version: 3.14.3+0.1744033158.p
      - name: gatekeeper-operator-product.v3.14.3-0.1742934403.p
        version: 3.14.3+0.1742934403.p
      - name: gatekeeper-operator-product.v3.14.3-0.1740676608.p
        version: 3.14.3+0.1740676608.p
      - name: gatekeeper-operator-product.v3.14.3
        version: 3.14.3
      - name: gatekeeper-operator-product.v3.14.2
        version: 3.14.2
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.14"
    - currentCSV: gatekeeper-operator-product.v3.15.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:8b43248aed06563e06f7a0347ff1c45fa5892e3a9adb2b5c363962a0b152eff0
          createdAt: 13 Mar 2025, 17:07
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.15.4
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: Gatekeeper allows administrators to detect and reject non-compliant
          commits to an infrastructure-as-code system's source-of-truth. This strengthens
          compliance efforts and prevents a bad state from slowing down the organization.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:8b43248aed06563e06f7a0347ff1c45fa5892e3a9adb2b5c363962a0b152eff0
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:29690144667b3753a7e76095e136d6fc42325270e7b9f22f97dfd411439e9f54
        version: 3.15.4
      entries:
      - name: gatekeeper-operator-product.v3.15.4
        version: 3.15.4
      - name: gatekeeper-operator-product.v3.15.3
        version: 3.15.3
      - name: gatekeeper-operator-product.v3.15.2
        version: 3.15.2
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.15"
    - currentCSV: gatekeeper-operator-product.v3.17.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:3c027ec78398df0a8fc09eb79e222a9508b4f18154d1e1b477d4f156181ae51a
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.17.3
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e44078fc6073a79168ed52922b370773b451d69bf335aeb91d26a1462df21a9b
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:3c027ec78398df0a8fc09eb79e222a9508b4f18154d1e1b477d4f156181ae51a
        version: 3.17.3
      entries:
      - name: gatekeeper-operator-product.v3.17.3
        version: 3.17.3
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.17"
    - currentCSV: gatekeeper-operator-product.v3.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:0803b7219ceff7aa676488eeba58607a19c9c18a2bb8eae8b3b16cfec411cb6e
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.18.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:49726cb2702441305e8047b24054575170fb595cfe5ebf815cdbfdaa957b02f4
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:0803b7219ceff7aa676488eeba58607a19c9c18a2bb8eae8b3b16cfec411cb6e
        version: 3.18.1
      entries:
      - name: gatekeeper-operator-product.v3.18.1
        version: 3.18.1
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.18"
    - currentCSV: gatekeeper-operator-product.v3.19.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.19.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:60ee347425c6cfb36f0e1e834b33cbb3d77aa2d6c89ac513216d0b9fc55d4b52
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
        version: 3.19.1
      entries:
      - name: gatekeeper-operator-product.v3.19.1
        version: 3.19.1
      - name: gatekeeper-operator-product.v3.19.0
        version: 3.19.0
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.19"
    - currentCSV: gatekeeper-operator-product.v3.19.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.19.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:60ee347425c6cfb36f0e1e834b33cbb3d77aa2d6c89ac513216d0b9fc55d4b52
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
        version: 3.19.1
      entries:
      - name: gatekeeper-operator-product.v3.19.1
        version: 3.19.1
      - name: gatekeeper-operator-product.v3.19.0
        version: 3.19.0
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: stable
    defaultChannel: stable
    packageName: gatekeeper-operator-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-activemqartemis: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/
    name: amq-broker-rhel8
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-broker-operator.v7.10.7-opr-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                    "name": "ex-aao",
                    "application": "ex-aao-app"
                },
                "spec": {
                    "deploymentPlan": {
                        "image": "placeholder",
                        "size": 2,
                        "requireLogin": false,
                        "persistenceEnabled": false,
                        "journalType": "nio",
                        "messageMigration": false,
                        "jolokiaAgentEnabled": false,
                        "managementRBACEnabled": true
                    }
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                    "name": "ex-aaoaddress"
                },
                "spec": {
                    "addressName": "myAddress0",
                    "queueName": "myQueue0",
                    "routingType": "anycast"
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                    "name": "ex-aaoscaledown"
                },
                "spec": {
                    "localOnly": "true"
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                    "name": "ex-keycloak"
                },
                "spec": {
                    "loginModules": {
                        "propertiesLoginModules": [
                            {
                                "name": "prop-module",
                                "users": [
                                    {
                                        "name": "superman",
                                        "roles": [
                                            "root"
                                        ]
                                    }
                                ]
                            }
                        ]
                    },
                    "securityDomains": {
                        "brokerDomain": {
                            "name": "activemq",
                            "loginModules": [
                                {
                                    "name": "prop-module",
                                    "flag": "sufficient"
                                }
                            ]
                        }
                    }
                }
            }]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:1a7aa54d2799d238eb5f49f7a95a78a896f6bf8d222567e9118e0e3963cc9aad
          createdAt: "2021-06-18T00:02:04Z"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A stateful deployment of one or more brokers
            displayName: AMQ Broker
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: Adding and removing addresses via custom resource definitions
            displayName: AMQ Broker Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: Provides message migration on clustered broker scaledown
            displayName: AMQ Broker Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: Security configuration for the broker
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
          to deploy and manage stateful AMQ Broker broker clusters
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: rkieley@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:1a7aa54d2799d238eb5f49f7a95a78a896f6bf8d222567e9118e0e3963cc9aad
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ba21779c29b431a22c824e91820705c9501fb0b7d6f7e1ae22fcd3becc8c8a2f
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:d4010af53429502f41d722c39d69a7bb147bb06d19e3d4a69de70dbe88e3990e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b575d6eebf44dc9049445e357488efb17a690953852419b534df30e0105b9ca4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ff6fefad518a6c997d4c5a6e475ba89640260167f0bc27715daf3cc30116fad1
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:56282c6a461e945a83bffb13d3f3d8d9621800f0fc9cd755a75d21eacd8f5fbf
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:979b59325aa0f34eb05625201beba53fccbb83bd5eb80a89dcb5261ae358138f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:83faefa67b33d93836a1be3cb438bb147242d602565dcbe59b4ed42b26072191
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:967d840b831b310abd3a468a8652d1fa48b2ce28b342bfab53999245deb004f1
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:2cb14c917246ffcbd5d861ca6f3d58cb1146a4e55715ee08160e115d3dfc1c3e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f37f98c809c6f29a83e3d5a3ac4494e28efe9b25d33c54f533c6a08662244622
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:4e09ee3bd8ff3479fdd00e221b7b1505ae04ca6c64fd6bc9e2f1a34b8e98d137
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0d857b95084e39c011651be89482d4dcc77c2d1c3ac3c6a819f3c9d5a57a425f
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:9465e398a55c89c121b3ae50374414190a991839fa5e1018fc02551f6aa3b9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3c00b524722b500ff0922fe7a8722868557512183a8f69248f37a68c45c6a770
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a94fa8135d3765363e442bda45df6e02a3e22bee7ea198bcb0d3584331a4320c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c5f4c08e068b9721967cf7c7cbd9a9e93fb5e39b264dd13b653e99d8f3fa9e0e
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:aa7af7e038a12b5202108b362c8af243954603e84034a16c0192684c86d5d28b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b74d03ed852a3731467ffda95266ce49f2065972f1c37bf254f3d52b34c11991
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:7abfdcfd7b2d3d12b9234f7b057bd199bceefa47bdf02f62236f82a13fdbf5b7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:869cda365447b0b7b53ed21d2496884c15dc9c07295a9dfa0337c2c9974fe018
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:5ff28e3062d3eabb9f52d8271b837fdab06cafbd8b8f732b1bc00de49ad69aad
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f58678f03a250dbb296ebeb112c9150340155799ba4dad9e90bf347b9ed0ca7b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3219f90789690a3c5cfcc86e5f602c443e75f1fdb3a19660f43ee3996bfa8f62
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:f2651cb097eb631fc6af90e4a8404d36516aa2205c546125bf0c37739ab525fa
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:35b7d96d5065f280a528626abd2c3df84e79ccea93ab6bd8b8837af66c330f30
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:45a5e054546a2c1ee338518f8477f728f5f7108c9dd8b516b4c79d4a7d6f5ba7
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:1a9f13fd565113055b2aa2022bd0709a63c1dc058a8570e0fd7eba5c7b101334
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:97c3aa5673610875a53571449f7d5210ae8d675072d044f17b416725ed0dba88
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:ba0f5df7da08bc030ff5e0e68ecb52a931c25f6b67c6556aa8a9661c82d69d42
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:83ed89fa942ccbf4451954956db42138c83f9b015e7de07fc159ddc4bb6dc843
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9a02decbc460ddfc501d8a4a9b41cd084776090951906b027838c0829e1fc15a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ff819e913406911735da39c51699405f747bd697116a273a00897ff176bf05ff
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:5d39f26f06769fe5c989ea0e0ee8f9166120e152ce2c9bc3336fc8d7cab2ff92
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:e6f71001dccfbbe1cd7cf4fded69e181c10f46cc5c8bd3a8d3045a3f3ac41e2f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:61e5a8c1e46f1c18648e0406e2a57c52f911979079020cc57c7a4b06a56940bb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:e03945756d86bd189e8c33027fa668d2065c168510085dfd06defd9f40d3e0ab
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:34739366153b825dcb113e0ed6115ef5656085cf8d47492789ff17464042b576
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:34ab1f66d8daa544e3823464383e9788e307665c875499838fbed95f3a74a4dd
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0b2126cfb6054fdf428c1f43b69e36e93a09a49ce15350e9273c98cc08c6598b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a8b2a9364fd06c8e1c29b8f8f28d3e4844bf7840a62b67137b663c558982d4a3
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:6e9e0e3fddc44fbf87591d8904176407f1ec409d578a7d108c13cdd48d85131b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:67ae74e72927557c55611d0248347b2099b117ac1aa245dac05856fe782fb59c
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:c3ac4602b48867f3d4f7cc0efd5c9f133098e409626cfb402f80386ff460dbf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c9eba0d1f9fba588711774f4db75221f057578407065127ed60c33eef564f499
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:86ff145bd68e797a0507495570063b54cbe4186fcb4acd83c900406619ef9524
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:d4a9176f49ffd9db6051fbda15d47a1aa32c282fb5010d0caf5b9243e80bf055
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:4440b20349032bb5c0bb2a8deb89832b37767af6f0b3208b067867a85641913e
        version: 7.10.7-opr-1
      entries:
      - name: amq-broker-operator.v7.10.7-opr-1
        version: 7.10.7-opr-1
      - name: amq-broker-operator.v7.10.6-opr-1
        version: 7.10.6-opr-1
      - name: amq-broker-operator.v7.10.5-opr-1
        version: 7.10.5-opr-1
      - name: amq-broker-operator.v7.10.4-opr-1
        version: 7.10.4-opr-1
      - name: amq-broker-operator.v7.10.3-opr-3
        version: 7.10.3-opr-3
      - name: amq-broker-operator.v7.10.3-opr-2
        version: 7.10.3-opr-2
      - name: amq-broker-operator.v7.10.3-opr-1
        version: 7.10.3-opr-1
      - name: amq-broker-operator.v7.10.2-opr-2-0.1676475747.p
        version: 7.10.2-opr-2+0.1676475747.p
      - name: amq-broker-operator.v7.10.2-opr-2-0.1680622941.p
        version: 7.10.2-opr-2+0.1680622941.p
      - name: amq-broker-operator.v7.10.2-opr-2
        version: 7.10.2-opr-2
      - name: amq-broker-operator.v7.10.2-opr-1
        version: 7.10.2-opr-1
      - name: amq-broker-operator.v7.10.1-opr-2
        version: 7.10.1-opr-2
      - name: amq-broker-operator.v7.10.1-opr-1
        version: 7.10.1-opr-1
      - name: amq-broker-operator.v7.10.0-opr-4
        version: 7.10.0-opr-4
      - name: amq-broker-operator.v7.10.0-opr-3
        version: 7.10.0-opr-3
      - name: amq-broker-operator.v7.10.0-opr-2
        version: 7.10.0-opr-2
      - name: amq-broker-operator.v7.10.0-opr-1
        version: 7.10.0-opr-1
      name: 7.10.x
    - currentCSV: amq-broker-operator.v7.11.8-opr-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisSecurity is the Schema for the activemqartemissecurities
              API
            displayName: Active MQArtemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:b9aecae6b0ce0b241c7a5b9e6e096395940bb6a384373a4d121822cf10c26d99
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b74d03ed852a3731467ffda95266ce49f2065972f1c37bf254f3d52b34c11991
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:442b6638133d42fa7faa41d86e47be95cdca56c2e16efcd6e7e22991c3812a36
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f37f98c809c6f29a83e3d5a3ac4494e28efe9b25d33c54f533c6a08662244622
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0b2126cfb6054fdf428c1f43b69e36e93a09a49ce15350e9273c98cc08c6598b
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:4440b20349032bb5c0bb2a8deb89832b37767af6f0b3208b067867a85641913e
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:7abfdcfd7b2d3d12b9234f7b057bd199bceefa47bdf02f62236f82a13fdbf5b7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b9f93c243ead118651083880786a81c624f390fcc3596c7ec5ebd55f34f95ea3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:35b7d96d5065f280a528626abd2c3df84e79ccea93ab6bd8b8837af66c330f30
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b575d6eebf44dc9049445e357488efb17a690953852419b534df30e0105b9ca4
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:f2651cb097eb631fc6af90e4a8404d36516aa2205c546125bf0c37739ab525fa
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:83ed89fa942ccbf4451954956db42138c83f9b015e7de07fc159ddc4bb6dc843
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:967d840b831b310abd3a468a8652d1fa48b2ce28b342bfab53999245deb004f1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0d857b95084e39c011651be89482d4dcc77c2d1c3ac3c6a819f3c9d5a57a425f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:e03945756d86bd189e8c33027fa668d2065c168510085dfd06defd9f40d3e0ab
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:56282c6a461e945a83bffb13d3f3d8d9621800f0fc9cd755a75d21eacd8f5fbf
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5e5d8b551ab8ca377e0bc939037970a2f31ee14e9f555853ebea2759d673895a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9a02decbc460ddfc501d8a4a9b41cd084776090951906b027838c0829e1fc15a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0f443f5aa818dd85bfa93e51e37bdac3909628a283aa52a89125b527dd75d9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:5d39f26f06769fe5c989ea0e0ee8f9166120e152ce2c9bc3336fc8d7cab2ff92
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:9465e398a55c89c121b3ae50374414190a991839fa5e1018fc02551f6aa3b9aa
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:1a9f13fd565113055b2aa2022bd0709a63c1dc058a8570e0fd7eba5c7b101334
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:45a5e054546a2c1ee338518f8477f728f5f7108c9dd8b516b4c79d4a7d6f5ba7
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c9eba0d1f9fba588711774f4db75221f057578407065127ed60c33eef564f499
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a94fa8135d3765363e442bda45df6e02a3e22bee7ea198bcb0d3584331a4320c
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:34ab1f66d8daa544e3823464383e9788e307665c875499838fbed95f3a74a4dd
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:86ff145bd68e797a0507495570063b54cbe4186fcb4acd83c900406619ef9524
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c5f4c08e068b9721967cf7c7cbd9a9e93fb5e39b264dd13b653e99d8f3fa9e0e
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:d4a9176f49ffd9db6051fbda15d47a1aa32c282fb5010d0caf5b9243e80bf055
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:29f07006ad62023c20d34971cb3ae900349efae5ab2918aaa8e5647d788032be
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ff819e913406911735da39c51699405f747bd697116a273a00897ff176bf05ff
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b0b3684ef1ebf8b44d3a4c46391984173792dde1bd7363ea06d43a8f95a717d0
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:cc438b4bc2b14f6bff4e5d32e7dc0fad6d2ca391c4e84b5ae925b7b40af76729
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ba21779c29b431a22c824e91820705c9501fb0b7d6f7e1ae22fcd3becc8c8a2f
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:ba0f5df7da08bc030ff5e0e68ecb52a931c25f6b67c6556aa8a9661c82d69d42
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:e6f71001dccfbbe1cd7cf4fded69e181c10f46cc5c8bd3a8d3045a3f3ac41e2f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:67ae74e72927557c55611d0248347b2099b117ac1aa245dac05856fe782fb59c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:979b59325aa0f34eb05625201beba53fccbb83bd5eb80a89dcb5261ae358138f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ade37271426ab8f543d72f8da3e9f663d665f9427379a2539fc9323f6897cc44
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:34739366153b825dcb113e0ed6115ef5656085cf8d47492789ff17464042b576
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:869cda365447b0b7b53ed21d2496884c15dc9c07295a9dfa0337c2c9974fe018
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ff6fefad518a6c997d4c5a6e475ba89640260167f0bc27715daf3cc30116fad1
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:61e5a8c1e46f1c18648e0406e2a57c52f911979079020cc57c7a4b06a56940bb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0dde839c2dce7cb684094bf26523c8e16677de03149a0fff468b8c3f106e1f4f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5f842ebe9a6a4294a056ad10189a5f277cfd3fb4e6e0b3225e194bd5c131151d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3c00b524722b500ff0922fe7a8722868557512183a8f69248f37a68c45c6a770
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:4e09ee3bd8ff3479fdd00e221b7b1505ae04ca6c64fd6bc9e2f1a34b8e98d137
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:d4010af53429502f41d722c39d69a7bb147bb06d19e3d4a69de70dbe88e3990e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a8b2a9364fd06c8e1c29b8f8f28d3e4844bf7840a62b67137b663c558982d4a3
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:5ff28e3062d3eabb9f52d8271b837fdab06cafbd8b8f732b1bc00de49ad69aad
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:d2ae3defac36f00112c6b773e13626336ba15d6f50f905ad1c33f122bfe1a6a9
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:97c3aa5673610875a53571449f7d5210ae8d675072d044f17b416725ed0dba88
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f58678f03a250dbb296ebeb112c9150340155799ba4dad9e90bf347b9ed0ca7b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:40ccaf2bb2e0f8132f2541b39cb835f17b38050b3ca9232f0861fd5fcaf4324a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:6e9e0e3fddc44fbf87591d8904176407f1ec409d578a7d108c13cdd48d85131b
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:c3ac4602b48867f3d4f7cc0efd5c9f133098e409626cfb402f80386ff460dbf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:137b893ad15e0cd5353bae134c8fae833257a12641b12d0f774f1565ce66f63c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:89bb769b4c33eb931cc7866eeabec9a315b913851cfa84bc6e11ae8d1e8b2cbb
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:2cb14c917246ffcbd5d861ca6f3d58cb1146a4e55715ee08160e115d3dfc1c3e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:25c3732e945b4bc3e9aca721ce709d0217339b065d9dac9afdb6202e9fd8a799
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:8b4c4400a02c6cacfd978147b158e2ccbe5da955578b4b970a5ee7edac865326
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3219f90789690a3c5cfcc86e5f602c443e75f1fdb3a19660f43ee3996bfa8f62
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:7d53744ae8c77aea09c59165486c9ed0eaa131140f6b346042293bce0bd8a745
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:aa7af7e038a12b5202108b362c8af243954603e84034a16c0192684c86d5d28b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9c1ac38bf8eec67a3da57c3cc61105513776a749408cb287fe6448a17889de4e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:83faefa67b33d93836a1be3cb438bb147242d602565dcbe59b4ed42b26072191
        version: 7.11.8-opr-1
      entries:
      - name: amq-broker-operator.v7.11.8-opr-1
        version: 7.11.8-opr-1
      - name: amq-broker-operator.v7.11.7-opr-2-1742200979
        version: 7.11.7-opr-2-1742200979
      - name: amq-broker-operator.v7.11.7-opr-2-1740583644
        version: 7.11.7-opr-2-1740583644
      - name: amq-broker-operator.v7.11.7-opr-2-1736956309
        version: 7.11.7-opr-2-1736956309
      - name: amq-broker-operator.v7.11.7-opr-2-1731330819
        version: 7.11.7-opr-2-1731330819
      - name: amq-broker-operator.v7.11.7-opr-2
        version: 7.11.7-opr-2
      - name: amq-broker-operator.v7.11.7-opr-1-1726157430
        version: 7.11.7-opr-1-1726157430
      - name: amq-broker-operator.v7.11.7-opr-1-1723204800
        version: 7.11.7-opr-1-1723204800
      - name: amq-broker-operator.v7.11.7-opr-1
        version: 7.11.7-opr-1
      - name: amq-broker-operator.v7.11.6-opr-2
        version: 7.11.6-opr-2
      - name: amq-broker-operator.v7.11.5-opr-1
        version: 7.11.5-opr-1
      - name: amq-broker-operator.v7.11.4-opr-1
        version: 7.11.4-opr-1
      - name: amq-broker-operator.v7.11.3-opr-1
        version: 7.11.3-opr-1
      - name: amq-broker-operator.v7.11.2-opr-1
        version: 7.11.2-opr-1
      - name: amq-broker-operator.v7.11.1-opr-2
        version: 7.11.1-opr-2
      - name: amq-broker-operator.v7.11.1-opr-1
        version: 7.11.1-opr-1
      - name: amq-broker-operator.v7.11.0-opr-3
        version: 7.11.0-opr-3
      - name: amq-broker-operator.v7.11.0-opr-2
        version: 7.11.0-opr-2
      name: 7.11.x
    - currentCSV: amq-broker-operator.v7.12.5-opr-1-0.1757928550.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-broker-operator.v7.12.5-opr-1-0.1756216162.p
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:7d53744ae8c77aea09c59165486c9ed0eaa131140f6b346042293bce0bd8a745
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9c1ac38bf8eec67a3da57c3cc61105513776a749408cb287fe6448a17889de4e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:29f07006ad62023c20d34971cb3ae900349efae5ab2918aaa8e5647d788032be
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:dc13871dfed52402c40fd192900fc3cb5a49da4fa08e4ec4d0e23a879e83db25
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5e5d8b551ab8ca377e0bc939037970a2f31ee14e9f555853ebea2759d673895a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:cc438b4bc2b14f6bff4e5d32e7dc0fad6d2ca391c4e84b5ae925b7b40af76729
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d71506472636d940b5dbd3e708d3e0448d777eeecfd38de61ba49e281d2979c6
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ffffa9875f0379e9373f89f05eb06e5a193273bb04bc3aa5f85b044357b79098
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:b9aecae6b0ce0b241c7a5b9e6e096395940bb6a384373a4d121822cf10c26d99
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:442b6638133d42fa7faa41d86e47be95cdca56c2e16efcd6e7e22991c3812a36
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:25c3732e945b4bc3e9aca721ce709d0217339b065d9dac9afdb6202e9fd8a799
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ade37271426ab8f543d72f8da3e9f663d665f9427379a2539fc9323f6897cc44
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:16253895bc113f395dc6ab5cd511901e43b6ee676369ec80d598a8741c3c9bf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:89bb769b4c33eb931cc7866eeabec9a315b913851cfa84bc6e11ae8d1e8b2cbb
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:bb6fbd68475a7852b4d99eea6c4ab313f9267da7963162f0d75375d7063409e7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b9f93c243ead118651083880786a81c624f390fcc3596c7ec5ebd55f34f95ea3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9cc48eecf1442ae04b8543fa5d4381a13bc2831390850828834d387006d1342b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:47fd4ce2533496828aba37bd1f9715e2164d5c90bd0fc6b25e7e0786d723bf01
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:023478b773a0c42739b86e822a4d435e01ef341b8136566ea52463f1b7e447ed
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:eacc4fb55e8c4c30894a51944960d536508001a49fda62782c6cb2ca8767961d
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0dde839c2dce7cb684094bf26523c8e16677de03149a0fff468b8c3f106e1f4f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:8b4c4400a02c6cacfd978147b158e2ccbe5da955578b4b970a5ee7edac865326
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d42d713da0ce6806fdc6492b6342586783e6865a82a8647d3c4288439b1751ee
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:7f9b921d89b6d4d91b365659afff127d99f1455b94b48a10a6a7f4c38dd8b700
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:90f1678c19c74454b77aed14fa6b3074a1c6f80fe884f1c6136e7e907f402aec
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b2402179b1d0ff9c950defe5d8808195ef9a4c3e13e3c469cae31ab51d5f5014
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:8329acdff170c5a8ae6ae16f6c37d3f7ab68d106a6e8be50c3fe131ec5ecceeb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5f842ebe9a6a4294a056ad10189a5f277cfd3fb4e6e0b3225e194bd5c131151d
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:40920687402fd747ca02d384e63692a5e1db625723a2bf3056a607c92ee8eee1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0f443f5aa818dd85bfa93e51e37bdac3909628a283aa52a89125b527dd75d9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:40ccaf2bb2e0f8132f2541b39cb835f17b38050b3ca9232f0861fd5fcaf4324a
        version: 7.12.5-opr-1+0.1757928550.p
      entries:
      - name: amq-broker-operator.v7.12.5-opr-1-0.1757928550.p
        version: 7.12.5-opr-1+0.1757928550.p
      - name: amq-broker-operator.v7.12.5-opr-1-0.1756216162.p
        version: 7.12.5-opr-1+0.1756216162.p
      - name: amq-broker-operator.v7.12.5-opr-1-0.1754272113.p
        version: 7.12.5-opr-1+0.1754272113.p
      - name: amq-broker-operator.v7.12.5-opr-1
        version: 7.12.5-opr-1
      - name: amq-broker-operator.v7.12.4-opr-1
        version: 7.12.4-opr-1
      - name: amq-broker-operator.v7.12.4-opr-1-0.1747217191.p
        version: 7.12.4-opr-1+0.1747217191.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741881887.p
        version: 7.12.3-opr-1+0.1741881887.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739991235.p
        version: 7.12.3-opr-1+0.1739991235.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1730142550.p
        version: 7.12.3-opr-1+0.1730142550.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1730813578.p
        version: 7.12.3-opr-1+0.1730813578.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1731483798.p
        version: 7.12.3-opr-1+0.1731483798.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1732654016.p
        version: 7.12.3-opr-1+0.1732654016.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1733315503.p
        version: 7.12.3-opr-1+0.1733315503.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1737643631.p
        version: 7.12.3-opr-1+0.1737643631.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1738596517.p
        version: 7.12.3-opr-1+0.1738596517.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739390944.p
        version: 7.12.3-opr-1+0.1739390944.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739889884.p
        version: 7.12.3-opr-1+0.1739889884.p
      - name: amq-broker-operator.v7.12.3-opr-1
        version: 7.12.3-opr-1
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741884937.p
        version: 7.12.3-opr-1+0.1741884937.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741883943.p
        version: 7.12.3-opr-1+0.1741883943.p
      - name: amq-broker-operator.v7.12.2-opr-1
        version: 7.12.2-opr-1
      - name: amq-broker-operator.v7.12.1-opr-1-0.1724054375.p
        version: 7.12.1-opr-1+0.1724054375.p
      - name: amq-broker-operator.v7.12.1-opr-1-0.1723669176.p
        version: 7.12.1-opr-1+0.1723669176.p
      - name: amq-broker-operator.v7.12.1-opr-1-0.1721247907.p
        version: 7.12.1-opr-1+0.1721247907.p
      - name: amq-broker-operator.v7.12.1-opr-1
        version: 7.12.1-opr-1
      - name: amq-broker-operator.v7.12.0-opr-2
        version: 7.12.0-opr-2
      - name: amq-broker-operator.v7.12.0-opr-1
        version: 7.12.0-opr-1
      name: 7.12.x
    defaultChannel: 7.12.x
    packageName: amq-broker-rhel8
    provider:
      name: Red Hat
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecrpublic-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecrpublic-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecrpublic-controller:1.1.1
          createdAt: "2025-09-20T01:21:52Z"
          description: AWS ECR PUBLIC controller is a service controller for managing
            ECR PUBLIC resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Repository represents the state of an AWS ecrpublic Repository
              resource.
            displayName: Repository
            kind: Repository
            name: repositories.ecrpublic.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ECR PUBLIC resources in AWS from within your Kubernetes cluster.

          **About Amazon ECR PUBLIC**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECR PUBLIC
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecrpublic
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECR PUBLIC Developer Resources
          url: https://aws.amazon.com/ECR PUBLIC/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecrpublic maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecrpublic-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ecrpublic-controller.v1.1.1
        version: 1.1.1
      - name: ack-ecrpublic-controller.v1.1.0
        version: 1.1.0
      - name: ack-ecrpublic-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecrpublic-controller.v1.0.12
        version: 1.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecrpublic-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-s3-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-s3-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "s3.services.k8s.aws/v1alpha1",
                "kind": "Bucket",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/s3-controller:1.1.1
          createdAt: "2025-09-19T19:23:45Z"
          description: AWS S3 controller is a service controller for managing S3 resources
            in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Bucket represents the state of an AWS s3 Bucket resource.
            displayName: Bucket
            kind: Bucket
            name: buckets.s3.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Simple Storage Service (S3) resources in AWS from within your Kubernetes cluster.

          **About Amazon S3**

          Amazon Simple Storage Service (Amazon S3) is an object storage service that offers industry-leading scalability, data availability, security, and performance. This means customers of all sizes and industries can use it to store and protect any amount of data for a range of use cases, such as data lakes, websites, mobile applications, backup and restore, archive, enterprise applications, IoT devices, and big data analytics. Amazon S3 provides easy-to-use management features so you can organize your data and configure finely-tuned access controls to meet your specific business, organizational, and compliance requirements. Amazon S3 is designed for 99.999999999% (11 9s) of durability, and stores data for millions of applications for companies all around the world.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon S3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - s3
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon S3 Developer Resources
          url: https://aws.amazon.com/s3/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: s3 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/s3-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-s3-controller.v1.1.1
        version: 1.1.1
      - name: ack-s3-controller.v1.1.0
        version: 1.1.0
      - name: ack-s3-controller.v1.0.35
        version: 1.0.35
      - name: ack-s3-controller.v1.0.33
        version: 1.0.33
      - name: ack-s3-controller.v1.0.32
        version: 1.0.32
      - name: ack-s3-controller.v1.0.31
        version: 1.0.31
      - name: ack-s3-controller.v1.0.30
        version: 1.0.30
      - name: ack-s3-controller.v1.0.29
        version: 1.0.29
      - name: ack-s3-controller.v1.0.28
        version: 1.0.28
      - name: ack-s3-controller.v1.0.27
        version: 1.0.27
      - name: ack-s3-controller.v1.0.26
        version: 1.0.26
      - name: ack-s3-controller.v1.0.25
        version: 1.0.25
      - name: ack-s3-controller.v1.0.24
        version: 1.0.24
      - name: ack-s3-controller.v1.0.23
        version: 1.0.23
      - name: ack-s3-controller.v1.0.21
        version: 1.0.21
      - name: ack-s3-controller.v1.0.20
        version: 1.0.20
      - name: ack-s3-controller.v1.0.19
        version: 1.0.19
      - name: ack-s3-controller.v1.0.18
        version: 1.0.18
      - name: ack-s3-controller.v1.0.17
        version: 1.0.17
      - name: ack-s3-controller.v1.0.16
        version: 1.0.16
      - name: ack-s3-controller.v1.0.15
        version: 1.0.15
      - name: ack-s3-controller.v1.0.14
        version: 1.0.14
      - name: ack-s3-controller.v1.0.13
        version: 1.0.13
      - name: ack-s3-controller.v1.0.12
        version: 1.0.12
      - name: ack-s3-controller.v1.0.11
        version: 1.0.11
      - name: ack-s3-controller.v1.0.9
        version: 1.0.9
      - name: ack-s3-controller.v1.0.7
        version: 1.0.7
      - name: ack-s3-controller.v1.0.6
        version: 1.0.6
      - name: ack-s3-controller.v1.0.5
        version: 1.0.5
      - name: ack-s3-controller.v1.0.4
        version: 1.0.4
      - name: ack-s3-controller.v1.0.3
        version: 1.0.3
      - name: ack-s3-controller.v1.0.2
        version: 1.0.2
      - name: ack-s3-controller.v1.0.1
        version: 1.0.1
      - name: ack-s3-controller.v1.0.0
        version: 1.0.0
      - name: ack-s3-controller.v0.1.8
        version: 0.1.8
      - name: ack-s3-controller.v0.1.7
        version: 0.1.7
      - name: ack-s3-controller.v0.1.6
        version: 0.1.6
      - name: ack-s3-controller.v0.1.5
        version: 0.1.5
      - name: ack-s3-controller.v0.1.4
        version: 0.1.4
      - name: ack-s3-controller.v0.1.3
        version: 0.1.3
      - name: ack-s3-controller.v0.1.2
        version: 0.1.2
      - name: ack-s3-controller.v0.1.1
        version: 0.1.1
      - name: ack-s3-controller.v0.1.0
        version: 0.1.0
      - name: ack-s3-controller.v0.0.20
        version: 0.0.20
      - name: ack-s3-controller.v0.0.19
        version: 0.0.19
      - name: ack-s3-controller.v0.0.18
        version: 0.0.18
      - name: ack-s3-controller.v0.0.17
        version: 0.0.17
      - name: ack-s3-controller.v0.0.16
        version: 0.0.16
      - name: ack-s3-controller.v0.0.15
        version: 0.0.15
      - name: ack-s3-controller.v0.0.14
        version: 0.0.14
      - name: ack-s3-controller.v0.0.13
        version: 0.0.13
      - name: ack-s3-controller.v0.0.12
        version: 0.0.12
      - name: ack-s3-controller.v0.0.11
        version: 0.0.11
      - name: ack-s3-controller.v0.0.10
        version: 0.0.10
      - name: ack-s3-controller.v0.0.9
        version: 0.0.9
      - name: ack-s3-controller.v0.0.8
        version: 0.0.8
      - name: ack-s3-controller.v0.0.7
        version: 0.0.7
      name: alpha
    defaultChannel: alpha
    packageName: ack-s3-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: numaresources-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: numaresources-operator.v4.16.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodetopology.openshift.io/v1alpha1",
                "kind": "NUMAResourcesOperator",
                "metadata": {
                  "name": "numaresourcesoperator"
                },
                "spec": {
                  "nodeGroups": [
                    {
                      "machineConfigPoolSelector": {
                        "matchLabels": {
                          "pools.operator.machineconfiguration.openshift.io/worker": ""
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "nodetopology.openshift.io/v1alpha1",
                "kind": "NUMAResourcesScheduler",
                "metadata": {
                  "name": "numaresourcesscheduler"
                },
                "spec": {
                  "imageSpec": "URL_OF_SCHEDULER_IMAGE_FROM_REDHAT_REGISTRY",
                  "logLevel": "Normal",
                  "schedulerName": "topo-aware-scheduler"
                }
              }
            ]
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "true"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.15.0 <4.16.6'
          operators.openshift.io/valid-subscription: "[\n\t\"OpenShift Kubernetes
            Engine\",\n\t\"OpenShift Container Platform\",\n\t\"OpenShift Platform
            Plus\"\n]"
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NUMAResourcesOperator is the Schema for the numaresourcesoperators
              API
            displayName: NUMA Resources Operator
            kind: NUMAResourcesOperator
            name: numaresourcesoperators.nodetopology.openshift.io
            version: v1
          - description: NUMAResourcesOperator is the Schema for the numaresourcesoperators
              API
            displayName: NUMA Resources Operator
            kind: NUMAResourcesOperator
            name: numaresourcesoperators.nodetopology.openshift.io
            version: v1alpha1
          - description: NUMAResourcesScheduler is the Schema for the numaresourcesschedulers
              API
            displayName: NUMA Aware Scheduler
            kind: NUMAResourcesScheduler
            name: numaresourcesschedulers.nodetopology.openshift.io
            version: v1
          - description: NUMAResourcesScheduler is the Schema for the numaresourcesschedulers
              API
            displayName: NUMA Aware Scheduler
            kind: NUMAResourcesScheduler
            name: numaresourcesschedulers.nodetopology.openshift.io
            version: v1alpha1
        description: NUMA resources exporter operator
        displayName: numaresources-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - RTE
        - NUMA
        links:
        - name: Numaresources Operator
          url: https://github.com/openshift-kni/numaresources-operator
        maintainers:
        - email: fromani@redhat.com
          name: fromani
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/numaresources-rhel9-operator@sha256:aca8232e89172ca9fd9ce4a66cae9e8bb027b6d16482ce6b8e27fb7779c398e2
        version: 4.16.6
      entries:
      - name: numaresources-operator.v4.16.6
        version: 4.16.6
      - name: numaresources-operator.v4.16.5
        version: 4.16.5
      - name: numaresources-operator.v4.16.4
        version: 4.16.4
      - name: numaresources-operator.v4.16.3
        version: 4.16.3
      - name: numaresources-operator.v4.16.2
        version: 4.16.2
      - name: numaresources-operator.v4.16.1
        version: 4.16.1
      - name: numaresources-operator.v4.16.0
        version: 4.16.0
      name: "4.16"
    defaultChannel: "4.16"
    packageName: numaresources-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.9-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lvms-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lvms-operator.v4.15.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "my-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "default": true,
                        "fstype": "xfs",
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:696edeb4c168646b49b8b0ca5b3d06071ab66fdb010e7cd46888d8b2f3a0c7a8
          description: Logical volume manager storage provides dynamically provisioned
            local storage for container workloads
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.11'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "test-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
          operatorframework.io/suggested-namespace: openshift-storage
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-storage",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/internal-objects: '["logicalvolumes.topolvm.io",
            "lvmvolumegroups.lvm.topolvm.io", "lvmvolumegroupnodestatuses.lvm.topolvm.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/lvm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LogicalVolume
            name: logicalvolumes.topolvm.io
            version: v1
          - description: LVMCluster is the Schema for the lvmclusters API
            displayName: LVMCluster
            kind: LVMCluster
            name: lvmclusters.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroupNodeStatus
            name: lvmvolumegroupnodestatuses.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroup
            name: lvmvolumegroups.lvm.topolvm.io
            version: v1alpha1
        description: Logical volume manager storage provides dynamically provisioned
          local storage.
        displayName: LVM Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - local storage
        - operator
        - LVM
        links:
        - name: Source Repository
          url: https://github.com/openshift/lvm-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:771d513c4d4e5d77d797b4c18f86e23162a20c6d39955e93d24f62b94e62e785
        - registry.redhat.io/lvms4/lvms-must-gather-rhel9@sha256:382ceea1151da56d0931403207c9f85721d58a66b39b1cd79790eca6219807d0
        - registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:696edeb4c168646b49b8b0ca5b3d06071ab66fdb010e7cd46888d8b2f3a0c7a8
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:82c01635b32821f04a7cfcf80da8b3f7e097786f220e0c85aab47bda4536a1d6
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:2db4f32824a8f9f6780dc3d236f73684cb20db41466ffa8f9a89fd3f5eb62fdd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:7ce50398cd29df6e07ccd89790a53a711a7e5b484698c48f6581f5d4e9efb05a
        - registry.redhat.io/lvms4/topolvm-rhel9@sha256:074a206d36c9e4198d90baed86f3046abd2ec12f9b36a715b60900011366b3d4
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:05dceaff7f0b2b769ddfd9261cc66d42aab9158ce99ed3dbd6c229e784072635
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:479454d0d4de6609d8a4696b9c7f0097115afb45db3bdd46631d58afd77dcc58
        version: 4.15.11
      entries:
      - name: lvms-operator.v4.15.11
        version: 4.15.11
      - name: lvms-operator.v4.15.10
        version: 4.15.10
      - name: lvms-operator.v4.15.9
        version: 4.15.9
      - name: lvms-operator.v4.15.8
        version: 4.15.8
      - name: lvms-operator.v4.15.7
        version: 4.15.7
      - name: lvms-operator.v4.15.6
        version: 4.15.6
      - name: lvms-operator.v4.15.5
        version: 4.15.5
      - name: lvms-operator.v4.15.4
        version: 4.15.4
      - name: lvms-operator.v4.15.3
        version: 4.15.3
      - name: lvms-operator.v4.15.2
        version: 4.15.2
      - name: lvms-operator.v4.15.1
        version: 4.15.1
      name: stable-4.15
    - currentCSV: lvms-operator.v4.16.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "my-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "default": true,
                        "fstype": "xfs",
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:b9d80bb263a54192501d5ba9366a4f0c19b34875b51bad4fa75ef22c60bb3ee7
          description: Logical volume manager storage provides dynamically provisioned
            local storage for container workloads
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.9'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "test-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
          operatorframework.io/suggested-namespace: openshift-storage
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-storage",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["logicalvolumes.topolvm.io",
            "lvmvolumegroups.lvm.topolvm.io", "lvmvolumegroupnodestatuses.lvm.topolvm.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lvm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LogicalVolume
            name: logicalvolumes.topolvm.io
            version: v1
          - description: LVMCluster is the Schema for the lvmclusters API
            displayName: LVMCluster
            kind: LVMCluster
            name: lvmclusters.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroupNodeStatus
            name: lvmvolumegroupnodestatuses.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroup
            name: lvmvolumegroups.lvm.topolvm.io
            version: v1alpha1
        description: Logical volume manager storage provides dynamically provisioned
          local storage.
        displayName: LVM Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - local storage
        - operator
        - LVM
        links:
        - name: Source Repository
          url: https://github.com/openshift/lvm-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/lvms4/topolvm-rhel9@sha256:227748103049d2cced921ba9fc4e2d3671f08c897cc2a5627d755fe5f9d2767d
        - registry.redhat.io/lvms4/lvms-must-gather-rhel9@sha256:20b28b87b3fe6d6bf739181376231d6f8b7454c34b59eaf2e06c04a9fca8f41d
        - registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:b9d80bb263a54192501d5ba9366a4f0c19b34875b51bad4fa75ef22c60bb3ee7
        version: 4.16.9
      entries:
      - name: lvms-operator.v4.16.9
        version: 4.16.9
      - name: lvms-operator.v4.16.8
        version: 4.16.8
      - name: lvms-operator.v4.16.7
        version: 4.16.7
      - name: lvms-operator.v4.16.6
        version: 4.16.6
      - name: lvms-operator.v4.16.5
        version: 4.16.5
      - name: lvms-operator.v4.16.4
        version: 4.16.4
      - name: lvms-operator.v4.16.3
        version: 4.16.3
      - name: lvms-operator.v4.16.2
        version: 4.16.2
      - name: lvms-operator.v4.16.1
        version: 4.16.1
      - name: lvms-operator.v4.16.0
        version: 4.16.0
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: lvms-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-custom-metrics-autoscaler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: custom-metrics-autoscaler.v2.17.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eventing.keda.sh/v1alpha1",
                "kind": "CloudEventSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keda",
                    "app.kubernetes.io/instance": "cloudeventsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cloudeventsource",
                    "app.kubernetes.io/part-of": "keda"
                  },
                  "name": "cloudeventsource-sample"
                },
                "spec": {
                  "clusterName": "cluster-sample",
                  "destination": {
                    "http": {
                      "uri": "http://foo.bar"
                    }
                  }
                }
              },
              {
                "apiVersion": "eventing.keda.sh/v1alpha1",
                "kind": "ClusterCloudEventSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keda",
                    "app.kubernetes.io/instance": "clustercloudeventsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clustercloudeventsource",
                    "app.kubernetes.io/part-of": "keda"
                  },
                  "name": "clustercloudeventsource-sample"
                },
                "spec": {
                  "clusterName": "clustercluster-sample",
                  "destination": {
                    "http": {
                      "uri": "http://foo.bar"
                    }
                  }
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ClusterTriggerAuthentication",
                "metadata": {
                  "name": "example-clustertriggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "example-secret-parameter"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "KedaController",
                "metadata": {
                  "name": "keda",
                  "namespace": "openshift-keda"
                },
                "spec": {
                  "admissionWebhooks": {
                    "logEncoder": "console",
                    "logLevel": "info"
                  },
                  "metricsServer": {
                    "logLevel": "0"
                  },
                  "operator": {
                    "logEncoder": "console",
                    "logLevel": "info"
                  },
                  "serviceAccount": null,
                  "watchNamespace": ""
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledJob",
                "metadata": {
                  "name": "scaledjob-sample"
                },
                "spec": {
                  "failedJobsHistoryLimit": 5,
                  "jobTargetRef": {
                    "activeDeadlineSeconds": 600,
                    "backoffLimit": 6,
                    "completions": 1,
                    "parallelism": 1,
                    "template": null
                  },
                  "maxReplicaCount": 100,
                  "pollingInterval": 30,
                  "successfulJobsHistoryLimit": 5,
                  "triggers": [
                    {
                      "metadata": {
                        "property": "example-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledObject",
                "metadata": {
                  "name": "example-scaledobject"
                },
                "spec": {
                  "cooldownPeriod": 300,
                  "maxReplicaCount": 100,
                  "minReplicaCount": 0,
                  "pollingInterval": 30,
                  "scaleTargetRef": {
                    "name": "example-deployment"
                  },
                  "triggers": [
                    {
                      "metadata": {
                        "property": "example-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "TriggerAuthentication",
                "metadata": {
                  "name": "example-triggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "example-secret-parameter"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Cloud Provider
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9-operator@sha256:9a359ae6612d2f22fe983653418143873cb9939f83dca6f008ec3cb8f3d53990
          createdAt: 20 Aug 2025, 21:31
          description: Custom Metrics Autoscaler Operator, an event-driven autoscaler
            based upon KEDA
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.1 <2.17.2-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-keda
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/custom-metrics-autoscaler-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudEventSource
            name: cloudeventsources.eventing.keda.sh
            version: v1alpha1
          - kind: ClusterCloudEventSource
            name: clustercloudeventsources.eventing.keda.sh
            version: v1alpha1
          - description: Describes cluster wide authentication parameters
            displayName: ClusterTriggerAuthentication
            kind: ClusterTriggerAuthentication
            name: clustertriggerauthentications.keda.sh
            version: v1alpha1
          - description: |
              Represents an installation of a particular version of KEDA Controller.
            displayName: KedaController
            kind: KedaController
            name: kedacontrollers.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your jobs and what the triggers
            displayName: ScaledJob
            kind: ScaledJob
            name: scaledjobs.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your application and what the
              triggers are
            displayName: ScaledObject
            kind: ScaledObject
            name: scaledobjects.keda.sh
            version: v1alpha1
          - description: Describes authentication parameters
            displayName: TriggerAuthentication
            kind: TriggerAuthentication
            name: triggerauthentications.keda.sh
            version: v1alpha1
        description: |
          ## About the managed application
          Custom Metrics Autoscaler for OpenShift is an event driven autoscaler based upon KEDA.  Custom Metrics Autoscaler can monitor event sources like Kafka, RabbitMQ, or cloud event sources and feed the metrics from those sources into the Kubernetes horizontal pod autoscaler.  With Custom Metrics Autoscaler, you can have event driven and serverless scale of deployments within any Kubernetes cluster.
          ## About this Operator
          The Custom Metrics Autoscaler Operator deploys and manages installation of KEDA Controller in the cluster. Install this operator and follow installation instructions on how to install Custom Metrics Autoscaler in your cluster.

          ## Prerequisites for enabling this Operator
          ## How to install Custom Metrics Autoscaler in the cluster
          The installation of Custom Metrics Autoscaler is triggered by the creation of `KedaController` resource. Please refer to the [KedaController Spec](https://github.com/openshift/custom-metrics-autoscaler-operator/blob/main/README.md#the-kedacontroller-custom-resource) for more details on available options.

          Only a resource named `keda` in the namespace where the operator is installed will trigger the installation, reconfiguration or removal of the KEDA Controller resource.

          There should be only one KEDA Controller in the cluster.

          The operator will behave in this manner whether it is installed with the `AllNamespaces` or `OwnNamespace` install mode. While the operator more closely matches the `OwnNamespace` semantics, `AllNamespaces` is a supported installation mode to allow it to be installed to namespaces with existing `OperatorGroups` which require that installation mode, such as `openshift-operators`.
        displayName: Custom Metrics Autoscaler
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keda
        - scaling
        - event
        links:
        - name: Custom Metrics Autoscaler Documentation
          url: https://docs.openshift.com/container-platform/latest/nodes/cma/nodes-cma-autoscaling-custom.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9-operator@sha256:9a359ae6612d2f22fe983653418143873cb9939f83dca6f008ec3cb8f3d53990
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9@sha256:ace9358fbd2d81a6e16bfa049c763ca16acf6cc2a59cdc0b6282e5e0dfc902cc
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-adapter-rhel9@sha256:0c8ae9323f927346e168ba829f6c7679af6b2dd359c8d0a020dd0ae3954481ce
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-admission-webhooks-rhel9@sha256:75631a21de057844e9ae45ae2496a22e16f3a32a6fd5f6cb5f90e173e3be2180
        version: 2.17.2-1
      entries:
      - name: custom-metrics-autoscaler.v2.17.2-1
        version: 2.17.2-1
      - name: custom-metrics-autoscaler.v2.15.1-6
        version: 2.15.1-6
      - name: custom-metrics-autoscaler.v2.15.1-4
        version: 2.15.1-4
      - name: custom-metrics-autoscaler.v2.14.1-467
        version: 2.14.1-467
      - name: custom-metrics-autoscaler.v2.14.1-454
        version: 2.14.1-454
      - name: custom-metrics-autoscaler.v2.13.1-427
        version: 2.13.1-427
      - name: custom-metrics-autoscaler.v2.12.1-394
        version: 2.12.1-394
      - name: custom-metrics-autoscaler.v2.11.2-322
        version: 2.11.2-322
      - name: custom-metrics-autoscaler.v2.11.2-311
        version: 2.11.2-311
      - name: custom-metrics-autoscaler.v2.10.1-267
        version: 2.10.1-267
      name: stable
    defaultChannel: stable
    packageName: openshift-custom-metrics-autoscaler-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com
    name: aws-load-balancer-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-load-balancer-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
          createdAt: 2022-06-28T17:56:41
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:d7bc364512178c36671d8a4b5a76cf7cb10f8e56997106187b0fe1f032670ece
        version: 0.0.1
      entries:
      - name: aws-load-balancer-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: aws-load-balancer-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
          createdAt: 2022-12-09T15:32:32
          olm.skipRange: <0.2.0
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:cfd210a2cb705156883fda126d052e05eb84345cc27ba04688067f8ed9a802ea
        version: 0.2.0
      entries:
      - name: aws-load-balancer-operator.v0.2.0
        version: 0.2.0
      name: stable-v0
    - currentCSV: aws-load-balancer-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
          createdAt: 2022-06-28T17:56:41
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:d7bc364512178c36671d8a4b5a76cf7cb10f8e56997106187b0fe1f032670ece
        version: 0.0.1
      entries:
      - name: aws-load-balancer-operator.v0.0.1
        version: 0.0.1
      name: stable-v0.1
    - currentCSV: aws-load-balancer-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
          createdAt: 2022-12-09T15:32:32
          olm.skipRange: <0.2.0
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:cfd210a2cb705156883fda126d052e05eb84345cc27ba04688067f8ed9a802ea
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        version: 0.2.0
      entries:
      - name: aws-load-balancer-operator.v0.2.0
        version: 0.2.0
      name: stable-v0.2
    - currentCSV: aws-load-balancer-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
          createdAt: 2024-01-23T11:21:00
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:2e0b9332a44d8d9c23e19c7accab0813a651f39210257820db508cac28876595
        version: 1.1.1
      entries:
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1
    - currentCSV: aws-load-balancer-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:6e15df39a3d25755a79888f739e1af8b052c3fba57bf8f6b3434ef58fd4288cf
          createdAt: 2024-01-24T09:13:21
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat <<EOF | oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
          EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note about UPI:__ For UPI based installs, additional documentation can be found in
          [VPC and subnets](https://github.com/openshift/aws-load-balancer-operator/blob/release-1.0/docs/prerequisites.md#vpc-and-subnets)

          __Note about STS:__ For clusters using STS, use the [STS documentation](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#for-sts-clusters) instead.
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9f9f8f69ab98c1f25cc301572339e0cc5c5b76f2e29ed26e10ac91ad20e0090c
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:6e15df39a3d25755a79888f739e1af8b052c3fba57bf8f6b3434ef58fd4288cf
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:73db97d20ee9415e9a8941cc6a36ed81d901d760d9949e95e10c90866ef8bf64
        version: 1.0.1
      entries:
      - name: aws-load-balancer-operator.v1.0.1
        version: 1.0.1
      - name: aws-load-balancer-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: aws-load-balancer-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
          createdAt: 2024-01-23T11:21:00
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:2e0b9332a44d8d9c23e19c7accab0813a651f39210257820db508cac28876595
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
        version: 1.1.1
      entries:
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    defaultChannel: stable-v1
    packageName: aws-load-balancer-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://github.com/redhat-developer/app-services-operator/
    name: rhoas-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhoas-operator.0.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "ServiceRegistryConnection",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace",
                  "labels": {
                    "app.kubernetes.io/component": "external-service",
                    "app.kubernetes.io/managed-by": "rhoas"
                  }
                },
                "spec": {
                  "accessTokenSecretName": "rh-managed-services-api-accesstoken",
                  "serviceRegistryId": "exampleId",
                  "credentials": {
                    "serviceAccountSecretName": "service-account-secret"
                  }
                }
              },
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "CloudServiceAccountRequest",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace"
                },
                "spec": {
                  "serviceAccountName": "rhoas-sa",
                  "serviceAccountDescription": "Operator created service account",
                  "serviceAccountSecretName": "service-account-credentials",
                  "accessTokenSecretName": "rh-managed-services-api-accesstoken"
                }
              },
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "CloudServicesRequest",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace",
                  "labels": {
                    "app.kubernetes.io/component": "external-service",
                    "app.kubernetes.io/managed-by": "rhoas"
                  }
                },
                "spec": {
                  "accessTokenSecretName": "rh-cloud-services-api-accesstoken"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: quay.io/rhoas/service-operator:0.13.0
          createdAt: "2023-01-20 16:09:37"
          description: Red Hat OpenShift Application Services Operator
          olm.skipRange: '>=0.0.1 <0.13.0'
          repository: https://github.com/redhat-developer/app-services-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Creates RHOAS Service Accounts
            displayName: Cloud Service Account Request
            kind: CloudServiceAccountRequest
            name: cloudserviceaccountrequests.rhoas.redhat.com
            version: v1alpha1
          - description: Cloud Services Request
            displayName: Cloud Services Request
            kind: CloudServicesRequest
            name: cloudservicesrequests.rhoas.redhat.com
            version: v1alpha1
          - description: Binds Kafka Instance
            displayName: Kafka Connection
            kind: KafkaConnection
            name: kafkaconnections.rhoas.redhat.com
            version: v1alpha1
          - description: Binds Service Registry Instance
            displayName: Service Registry Connection
            kind: ServiceRegistryConnection
            name: serviceregistryconnections.rhoas.redhat.com
            version: v1alpha1
        description: "**OpenShift Application Services Operator (RHOAS-Operator)**
          creates instances of Red Hat OpenShift Application Services in your cluster.
          Operator is currently available as limited Beta and it cannot be consumed
          without invitation.\n\nRed Hat OpenShift Application Services, when added
          to Red Hat OpenShift Dedicated,  provide a streamlined developer experience
          for building, deploying,  and scaling cloud-native applications in open
          hybrid-cloud environments.   \n\nRHOAS-Operator enables developers to connect
          with their Application Services directly in their own cluster by using [RHOAS
          cli](https://github.com/redhat-developer/app-services-cli) and OpenShift
          Console UI.  \n\nRHOAS-Operator requires developers to connect with their
          Red Hat OpenShift Streams for Apache Kafka (RHOSAK) instances.  "
        displayName: OpenShift Application Services (RHOAS)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - rhoas
        - kafka
        - serviceregistry
        - app-services
        - application
        - services
        links:
        - name: RHOAS Operator Documentation
          url: https://github.com/redhat-developer/app-services-operator/tree/main/docs
        maintainers:
        - email: rhosak-eval-support@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://github.com/redhat-developer/app-services-operator/
        relatedImages:
        - quay.io/rhoas/service-operator:0.13.0
        version: 0.13.0
      entries:
      - name: rhoas-operator.0.13.0
        version: 0.13.0
      - name: rhoas-operator.0.12.0
        version: 0.12.0
      - name: rhoas-operator.0.11.0
        version: 0.11.0
      - name: rhoas-operator.0.9.6
        version: 0.9.6
      - name: rhoas-operator.0.9.5
        version: 0.9.5
      - name: rhoas-operator.0.9.3
        version: 0.9.3
      - name: rhoas-operator.0.9.2
        version: 0.9.2
      - name: rhoas-operator.0.9.1
        version: 0.9.1
      - name: rhoas-operator.0.9.0
        version: 0.9.0
      - name: rhoas-operator.0.8.2
        version: 0.8.2
      - name: rhoas-operator.0.7.10
        version: 0.7.10
      - name: rhoas-operator.0.7.9
        version: 0.7.9
      - name: rhoas-operator.0.7.8
        version: 0.7.8
      - name: rhoas-operator.0.7.6
        version: 0.7.6
      - name: rhoas-operator.0.7.2
        version: 0.7.2
      - name: rhoas-operator.0.7.1
        version: 0.7.1
      - name: rhoas-operator.0.6.8
        version: 0.6.8
      name: beta
    defaultChannel: beta
    packageName: rhoas-operator
    provider:
      name: Red Hat Inc.
      url: https://github.com/redhat-developer/app-services-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Rebaca Technologies Pvt Ltd
      provider-url: https://www.rebaca.com/telecom/solutions
    name: abot-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: abot-operator-rhmp.3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "abot.com/v3",
                "kind": "Abot",
                "metadata": {
                  "name": "abot"
                },
                "spec": {
                  "abotanalytics": {
                    "image": "registry.connect.redhat.com/rebaca/abot-analytics@sha256:c854d9eafbe26f834e10f5a97a587b2a2bf05558554897817bb8e2556124b75d",
                    "resources": {
                      "limits": {
                        "cpu": "4000m",
                        "ephemeral-storage": "8Gi",
                        "memory": "6000Mi"
                      },
                      "requests": {
                        "cpu": "2000m",
                        "ephemeral-storage": "4Gi",
                        "memory": "4000Mi"
                      }
                    }
                  },
                  "abotanalyticsbase": {
                    "image": "registry.connect.redhat.com/rebaca/abot-analyticsbase@sha256:dcc2a417a889acd6ec63538a6dfa393d4c1942dac178a04e5d65189a1ae054dc",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotbaseimage": "registry.connect.redhat.com/rebaca/abot-baseimage@sha256:2cbe38298d26be5896e0dc071aae111dd921d0dbd05c3525290f6ff31c73e926",
                  "abotbdd": {
                    "image": "registry.connect.redhat.com/rebaca/abot-bdd@sha256:d5c6bfe12a548cbdb4e0d83d5f18e5ab0c17c3af5340e437b2d0c23740c9c9e6",
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "ephemeral-storage": "8Gi",
                        "memory": "2000Mi"
                      },
                      "requests": {
                        "cpu": "1000m",
                        "ephemeral-storage": "4Gi",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotgrafana": {
                    "image": "registry.redhat.io/rhel8/grafana@sha256:d894cd34f2d8cac30e0fab69bd869f4b0f258b5086a8fd7ead0c5886ce2c5f7a",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "1500Mi"
                      },
                      "requests": {
                        "cpu": "300m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotgui": {
                    "image": "registry.connect.redhat.com/rebaca/abot-gui@sha256:78b512d6bd223a1f5d35f4cbd221f6637e2c9267726ce8899ee5d86aaac82b6f",
                    "resources": {
                      "limits": {
                        "cpu": "4000m",
                        "ephemeral-storage": "8Gi",
                        "memory": "6000Mi"
                      },
                      "requests": {
                        "cpu": "2000m",
                        "ephemeral-storage": "4Gi",
                        "memory": "4000Mi"
                      }
                    }
                  },
                  "abotkibana": {
                    "image": "registry.connect.redhat.com/elastic/kibana@sha256:2be3542d50f51ad786ead958a399f26852339e9417279688aad194c6794336ce",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "abotlogstash": {
                    "image": "registry.connect.redhat.com/elastic/logstash@sha256:dbc25dc5c31dfee08e86221af9f549b50075b2ccf7f255779c5f022c54cd892e",
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "2000Mi"
                      }
                    }
                  },
                  "abotmongo": {
                    "image": "registry.connect.redhat.com/rebaca/abot-mongodb@sha256:be9059e7e8bc1636a84575416b823c20463228785598662ee81c27e4c3c987dc",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "abotnginx": {
                    "image": "registry.connect.redhat.com/rebaca/abot-nginx@sha256:3edc22b40cd30230768115efdceec3d054d71f78214b4100b7425d8484154378",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "abotrabbitmq": {
                    "image": "registry.connect.redhat.com/rebaca/abot-rabbitmq@sha256:e609bec067fe7740590bb272db27d1a7770db47a0fc802e2288c022e101c9e25",
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "abotsidecar": {
                    "alertmanagerimage": "registry.redhat.io/rhacm2/prometheus-alertmanager-rhel8@sha256:ecaee8ad24a9c7cb3b945ee6d2c1275b9de2a676e1868abc2687e687385b3f6a",
                    "kubestatemetricsimage": "registry.redhat.io/rhacm2/kube-state-metrics-rhel8@sha256:e6bb6fa0c155bebf64749368dd951c15f1dbfa3559e2ba6734d3661d3a95c00c",
                    "prometheusimage": "registry.connect.redhat.com/kasten/prometheus@sha256:1e490fcccdd50c1026150ba4cad7c29297d2b73b3d5c80240ff31509e666363d",
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "8000Mi"
                      },
                      "requests": {
                        "cpu": "1000m",
                        "memory": "6000Mi"
                      }
                    }
                  },
                  "elasticsearch": {
                    "image": "registry.connect.redhat.com/elastic/elasticsearch@sha256:f47fd7df80ab05635e31d733f6969d38205f9f81cf95da8594a047f3d07c59ae",
                    "replicas": 3,
                    "resources": {
                      "limits": {
                        "cpu": "1200m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "800m",
                        "memory": "1000Mi"
                      }
                    },
                    "volumeClaim": {
                      "storage": "50Gi"
                    }
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "initContainers": {
                    "image": "registry.access.redhat.com/ubi8/ubi@sha256:a965f33ee4ee57dc8e40a1f9350ddf28ed0727b6cf80db46cdad0486a7580f9d",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "protocol": "https",
                  "pvc": {
                    "abotvol": "100Gi",
                    "analyticsvol": "100Gi",
                    "databak": "10Gi",
                    "mongovol": "10Gi",
                    "prometheusvol": "10Gi"
                  },
                  "replicas": 1,
                  "security": {
                    "fsGroup": 1018450000,
                    "seLinuxOptions": {
                      "level": "s0:c123,c456"
                    }
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "name": "abot"
                  },
                  "storageclass": "ocs-storagecluster-cephfs"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/rebaca/abot-operator@sha256:d3e8e1b6988827b3c636bdfcb57cda26870089d90cf2a732be522f5474c0a734
          createdAt: "2024-10-03T20:32:24Z"
          description: The ABot Test Bench is a test automation tool for 4G, 5G and
            ORAN Telecom networks that follow specifications and use standards driven
            design and compliance as specified by the 3GPP consortium.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/abot-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/abot-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=1.0.0 <3.0.0'
          operatorframework.io/suggested-namespace: abot
          operators.openshift.io/valid-subscription: '["ABot Basic License for 15
            days"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: https://abot.rebaca.com/support/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Abot
            name: abots.abot.com
            version: v3
        description: A Helm based operator for deploying Abot application in Openshift
          cluster. The ABot Test Bench is a test automation tool for 4G, 5G and ORAN
          Telecom networks that follow specifications and use standards driven design
          and compliance as specified by the 3GPP consortium. It provides a platform
          for a user to quickly check for protocol compliance and test for load characteristics
          of a particular system under test.
        displayName: Abot Operator-v3.0.0
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 5G
        - 4G
        - ORAN
        - test tool
        - compliance
        - interop
        links:
        - name: Abot Operator
          url: https://abot-operator.domain
        maintainers:
        - email: suchismita.nag@rebaca.com
          name: Suchismita Nag
        - email: samir@rebaca.com
          name: Samir Chatterjee
        maturity: alpha
        provider:
          name: Rebaca Technologies Pvt Ltd
          url: https://www.rebaca.com/telecom/solutions
        relatedImages:
        - registry.connect.redhat.com/rebaca/abot-rabbitmq@sha256:e609bec067fe7740590bb272db27d1a7770db47a0fc802e2288c022e101c9e25
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel8@sha256:e6bb6fa0c155bebf64749368dd951c15f1dbfa3559e2ba6734d3661d3a95c00c
        - registry.connect.redhat.com/rebaca/abot-bdd@sha256:d5c6bfe12a548cbdb4e0d83d5f18e5ab0c17c3af5340e437b2d0c23740c9c9e6
        - registry.connect.redhat.com/rebaca/abot-baseimage@sha256:2cbe38298d26be5896e0dc071aae111dd921d0dbd05c3525290f6ff31c73e926
        - registry.connect.redhat.com/elastic/elasticsearch@sha256:f47fd7df80ab05635e31d733f6969d38205f9f81cf95da8594a047f3d07c59ae
        - registry.connect.redhat.com/rebaca/abot-analyticsbase@sha256:dcc2a417a889acd6ec63538a6dfa393d4c1942dac178a04e5d65189a1ae054dc
        - registry.connect.redhat.com/rebaca/abot-mongodb@sha256:be9059e7e8bc1636a84575416b823c20463228785598662ee81c27e4c3c987dc
        - registry.connect.redhat.com/rebaca/abot-nginx@sha256:3edc22b40cd30230768115efdceec3d054d71f78214b4100b7425d8484154378
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e34f912158985525f6efb3c4b675f20854fc677c867d12fe3cd2e971309428dc
        - registry.connect.redhat.com/rebaca/abot-operator@sha256:d3e8e1b6988827b3c636bdfcb57cda26870089d90cf2a732be522f5474c0a734
        - registry.redhat.io/rhel8/grafana@sha256:d894cd34f2d8cac30e0fab69bd869f4b0f258b5086a8fd7ead0c5886ce2c5f7a
        - registry.connect.redhat.com/elastic/logstash@sha256:dbc25dc5c31dfee08e86221af9f549b50075b2ccf7f255779c5f022c54cd892e
        - registry.connect.redhat.com/kasten/prometheus@sha256:1e490fcccdd50c1026150ba4cad7c29297d2b73b3d5c80240ff31509e666363d
        - registry.connect.redhat.com/rebaca/abot-analytics@sha256:c854d9eafbe26f834e10f5a97a587b2a2bf05558554897817bb8e2556124b75d
        - registry.connect.redhat.com/rebaca/abot-gui@sha256:78b512d6bd223a1f5d35f4cbd221f6637e2c9267726ce8899ee5d86aaac82b6f
        - registry.connect.redhat.com/elastic/kibana@sha256:2be3542d50f51ad786ead958a399f26852339e9417279688aad194c6794336ce
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel8@sha256:ecaee8ad24a9c7cb3b945ee6d2c1275b9de2a676e1868abc2687e687385b3f6a
        - registry.access.redhat.com/ubi8/ubi@sha256:a965f33ee4ee57dc8e40a1f9350ddf28ed0727b6cf80db46cdad0486a7580f9d
        version: 3.0.0
      entries:
      - name: abot-operator-rhmp.3.0.0
        version: 3.0.0
      - name: abot-operator.2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: abot-operator-rhmp
    provider:
      name: Rebaca Technologies Pvt Ltd
      url: https://www.rebaca.com/telecom/solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: secrets-store-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secrets-store-csi-driver-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9-operator@sha256:96f3a8df9ac5ff3828b0add052af00f7d9df0ebb622a6adfd9772658d55baeb8
          createdAt: "2023-06-12T00:00:00Z"
          description: Install and configure Secrets Store CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.13.0-0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/secrets-store-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SecretProviderClass is the Schema for the secretproviderclasses
              API
            displayName: Secret Provider Class
            kind: SecretProviderClass
            name: secretproviderclasses.secrets-store.csi.x-k8s.io
            version: v1
          - description: SecretProviderClassPodStatus is the Schema for the secretproviderclassespodstatus
              API
            displayName: Secret Provider Class Pod Status
            kind: SecretProviderClassPodStatus
            name: secretproviderclasspodstatuses.secrets-store.csi.x-k8s.io
            version: v1
        description: |
          Operator that installs and configures the CSI driver for Secrets Store. The Secrets Store CSI Driver Operator is a Technology Preview feature. For more information on Red Hat Technology Preview features support scope, see https://access.redhat.com/support/offerings/techpreview/.
        displayName: Secrets Store CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: Documentation
          url: https://github.com/openshift/secrets-store-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/secrets-store-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: preview
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0627c09fe3f5a015f117fb30af1f4d17340aefa48b23994737d3901a804b63d3
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:a5704d26d0b562dba9633d1c107f34e05c750c0e3a64dbcace4dab7b964fff8d
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9-operator@sha256:96f3a8df9ac5ff3828b0add052af00f7d9df0ebb622a6adfd9772658d55baeb8
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9@sha256:148672ccd882a43c5c1d52763da465558b5ed54310418ec40c7ca0b56e408b1a
        version: 4.16.0-202509240123
      entries:
      - name: secrets-store-csi-driver-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: secrets-store-csi-driver-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: secrets-store-csi-driver-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: secrets-store-csi-driver-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: secrets-store-csi-driver-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: secrets-store-csi-driver-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: secrets-store-csi-driver-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: secrets-store-csi-driver-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: secrets-store-csi-driver-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: secrets-store-csi-driver-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: secrets-store-csi-driver-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: secrets-store-csi-driver-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: secrets-store-csi-driver-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: secrets-store-csi-driver-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: secrets-store-csi-driver-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: secrets-store-csi-driver-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: secrets-store-csi-driver-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: secrets-store-csi-driver-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: secrets-store-csi-driver-operator.v4.16.0-202410031007
        version: 4.16.0-202410031007
      - name: secrets-store-csi-driver-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: secrets-store-csi-driver-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: secrets-store-csi-driver-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: secrets-store-csi-driver-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: secrets-store-csi-driver-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: secrets-store-csi-driver-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: secrets-store-csi-driver-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: secrets-store-csi-driver-operator.v4.16.0-202408021139
        version: 4.16.0-202408021139
      - name: secrets-store-csi-driver-operator.v4.16.0-202407190036
        version: 4.16.0-202407190036
      - name: secrets-store-csi-driver-operator.v4.16.0-202407111006
        version: 4.16.0-202407111006
      - name: secrets-store-csi-driver-operator.v4.16.0-202407081937
        version: 4.16.0-202407081937
      - name: secrets-store-csi-driver-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: secrets-store-csi-driver-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: secrets-store-csi-driver-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: secrets-store-csi-driver-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: preview
    defaultChannel: preview
    packageName: secrets-store-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/limitador-operator
    name: limitador-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: limitador-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "name": "toy_get_route",
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/limitador-rhel9-operator@sha256:edb28f112a0d3536f2f52467f38df685d953f46f1999cbd61f4fd4fa4805cd34
          createdAt: "2025-04-07T15:52:04Z"
          description: Enables rate limiting for Gateways and applications in a Gateway
            API network
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Limitador is the Schema for the limitadors API
            displayName: Limitador
            kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: Enables rate limiting for Gateways and applications in a Gateway
          API network. - Part of Red Hat Connectivity Link
        displayName: Red Hat - Limitador Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        - kuadrant
        links:
        - name: GitHub
          url: https://github.com/kuadrant/limitador-operator
        - name: Documentation
          url: https://github.com/Kuadrant/limitador/tree/main/docs
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - registry.redhat.io/rhcl-1/limitador-rhel9-operator@sha256:edb28f112a0d3536f2f52467f38df685d953f46f1999cbd61f4fd4fa4805cd34
        - registry.redhat.io/rhcl-1/limitador-rhel9@sha256:98f85a05da53fbf7f20845d1969441620407cc13a185fd52714308679f116425
        version: 1.1.0
      entries:
      - name: limitador-operator.v1.1.0
        version: 1.1.0
      - name: limitador-operator.v1.0.2
        version: 1.0.2
      - name: limitador-operator.v1.0.1
        version: 1.0.1
      - name: limitador-operator.v0.12.1
        version: 0.12.1
      name: stable
    defaultChannel: stable
    packageName: limitador-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/limitador-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres-community
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.17.0-rc1
        version: 1.17.0-rc1
      - name: stackgres-community.v1.16.3
        version: 1.16.3
      - name: stackgres-community.v1.16.2
        version: 1.16.2
      - name: stackgres-community.v1.16.1
        version: 1.16.1
      - name: stackgres-community.v1.16.0
        version: 1.16.0
      - name: stackgres-community.v1.16.0-rc3
        version: 1.16.0-rc3
      - name: stackgres-community.v1.16.0-rc2
        version: 1.16.0-rc2
      - name: stackgres-community.v1.16.0-rc1
        version: 1.16.0-rc1
      - name: stackgres-community.v1.15.2
        version: 1.15.2
      - name: stackgres-community.v1.15.1
        version: 1.15.1
      - name: stackgres-community.v1.15.0
        version: 1.15.0
      - name: stackgres-community.v1.15.0-rc4
        version: 1.15.0-rc4
      - name: stackgres-community.v1.15.0-rc3
        version: 1.15.0-rc3
      - name: stackgres-community.v1.15.0-rc2
        version: 1.15.0-rc2
      - name: stackgres-community.v1.15.0-rc1
        version: 1.15.0-rc1
      - name: stackgres-community.v1.14.3
        version: 1.14.3
      - name: stackgres-community.v1.14.2
        version: 1.14.2
      - name: stackgres-community.v1.14.1
        version: 1.14.1
      - name: stackgres-community.v1.14.0
        version: 1.14.0
      - name: stackgres-community.v1.14.0-rc2
        version: 1.14.0-rc2
      - name: stackgres-community.v1.14.0-rc1
        version: 1.14.0-rc1
      - name: stackgres-community.v1.13.0
        version: 1.13.0
      - name: stackgres-community.v1.13.0-rc1
        version: 1.13.0-rc1
      - name: stackgres-community.v1.12.0
        version: 1.12.0
      - name: stackgres-community.v1.12.0-rc1
        version: 1.12.0-rc1
      name: candidate
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.17.0-rc1
        version: 1.17.0-rc1
      - name: stackgres-community.v1.16.3
        version: 1.16.3
      - name: stackgres-community.v1.16.2
        version: 1.16.2
      - name: stackgres-community.v1.16.1
        version: 1.16.1
      - name: stackgres-community.v1.16.0
        version: 1.16.0
      - name: stackgres-community.v1.16.0-rc3
        version: 1.16.0-rc3
      - name: stackgres-community.v1.16.0-rc2
        version: 1.16.0-rc2
      - name: stackgres-community.v1.16.0-rc1
        version: 1.16.0-rc1
      - name: stackgres-community.v1.15.2
        version: 1.15.2
      - name: stackgres-community.v1.15.1
        version: 1.15.1
      - name: stackgres-community.v1.15.0
        version: 1.15.0
      - name: stackgres-community.v1.15.0-rc4
        version: 1.15.0-rc4
      - name: stackgres-community.v1.15.0-rc3
        version: 1.15.0-rc3
      - name: stackgres-community.v1.15.0-rc2
        version: 1.15.0-rc2
      - name: stackgres-community.v1.15.0-rc1
        version: 1.15.0-rc1
      - name: stackgres-community.v1.14.3
        version: 1.14.3
      - name: stackgres-community.v1.14.2
        version: 1.14.2
      - name: stackgres-community.v1.14.1
        version: 1.14.1
      - name: stackgres-community.v1.14.0
        version: 1.14.0
      - name: stackgres-community.v1.14.0-rc2
        version: 1.14.0-rc2
      - name: stackgres-community.v1.14.0-rc1
        version: 1.14.0-rc1
      - name: stackgres-community.v1.13.0
        version: 1.13.0
      - name: stackgres-community.v1.13.0-rc1
        version: 1.13.0-rc1
      - name: stackgres-community.v1.12.0
        version: 1.12.0
      - name: stackgres-community.v1.12.0-rc1
        version: 1.12.0-rc1
      name: fast
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.16.3
        version: 1.16.3
      - name: stackgres-community.v1.16.2
        version: 1.16.2
      - name: stackgres-community.v1.16.1
        version: 1.16.1
      - name: stackgres-community.v1.16.0
        version: 1.16.0
      - name: stackgres-community.v1.15.2
        version: 1.15.2
      - name: stackgres-community.v1.15.1
        version: 1.15.1
      - name: stackgres-community.v1.15.0
        version: 1.15.0
      - name: stackgres-community.v1.14.3
        version: 1.14.3
      - name: stackgres-community.v1.14.2
        version: 1.14.2
      - name: stackgres-community.v1.14.1
        version: 1.14.1
      - name: stackgres-community.v1.14.0
        version: 1.14.0
      - name: stackgres-community.v1.13.0
        version: 1.13.0
      - name: stackgres-community.v1.12.0
        version: 1.12.0
      name: stable
    defaultChannel: stable
    packageName: stackgres-community
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: stolostron-engine
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stolostron-engine.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:b7d0c8413a80c251779fe3c1c9200f9fcfbbe085ccee3ab968ebd4a39209946c
          createdAt: "2024-06-28T02:42:26Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/cluster-proxy-addon@sha256:f70a26b276c67ad791d337e35cd102afa82f1005bc6ecbc60309932d2d163a6b
        - quay.io/stolostron/backplane-operator@sha256:b7d0c8413a80c251779fe3c1c9200f9fcfbbe085ccee3ab968ebd4a39209946c
        - quay.io/stolostron/registration@sha256:8dca354dd63739d7578c24d57b2a72c0676c979344e05c6f2634a8a9fae3218d
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/assisted-service@sha256:653728f209070c6adad846a4780449f56985b19006024cfe8137a3a1854fdc85
        - quay.io/stolostron/backplane-must-gather@sha256:2676ebb7438ca42f0d852100612715c429fe9fbe072297fc178d51c34378d1af
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/hypershift-operator@sha256:7dd5eaf195e6b1ca4cecd50514f5b7045948184f9ee617341bc90edc5f9914f2
        - quay.io/stolostron/discovery-operator@sha256:acb2c4033da62055feb49ff3023b2850a2768996e1fa80f52588b3e8e458e8f5
        - quay.io/stolostron/addon-manager@sha256:87c829911438061a53aff2b297416bf43080fdd8d93558caefc198e6f9cb382f
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:b269d88e30551636e024aa1472ea4dd71b873b5a2778a199b32732488fd776fb
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/assisted-image-service@sha256:0c29a74e0f82d1782d5170ff674f5bfcd89ecc5d180fe5b583a5e0f388b9504b
        - quay.io/stolostron/assisted-installer-controller@sha256:213dc6fde247a5b9862f5076d40394164dd596cde7ca2061e58b73f4d2dbc89e
        - quay.io/stolostron/console-mce@sha256:4e49f2b2fdb0491c225e4d5dc143a6d7c0a25cd855057b3792c7ee94a868ac85
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:f0d6d46d3a4d65fba6b74bdc9eab99e8149133f83a5c8486f22a73ab92f78ed6
        - quay.io/stolostron/image-based-install-operator@sha256:b9af0ddb7ae8e6fd29b953dfb917546dfd676206e2ad143e892172b6e1cea5a1
        - quay.io/stolostron/work@sha256:4ef2518aa107ebd28d5fb1a98d63d82a8e0610544600169e0dd5f345599e518b
        - quay.io/stolostron/cluster-api-provider-agent@sha256:c074ca1f28d03e86efe4f09200a38af4ac947247a19c46c5300eefcd4f337006
        - quay.io/stolostron/discovery-operator-tests@sha256:c98c42b0c16306de86d7b365096cb809a0c797b0a4b7dc04a7de877d0ba5733a
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/helloworld@sha256:abc0d8002c5fe6d95a1f24fce1b7899ca24fac576c894ea7443cf7e15e6a42c2
        - quay.io/stolostron/cluster-curator-controller@sha256:e3a91a4b06dd73fc4e401ddf034e61d4d5b2d324d2c5cea92941e1a56c4023c1
        - quay.io/stolostron/assisted-service-el8@sha256:31422ffce8b94d662131986bd084fcba503809cb939389096f9c9dbb27cb58db
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/discovery-testserver@sha256:83d7d6cfaef629a4e3d5d0e9fe35456aa220f1eb68f6fe611d85232acf46bc73
        - quay.io/stolostron/placement@sha256:0dbcc7c62569920afcad1443529f10e38d651a5fe5ae14ade89d6e42fb522e60
        - quay.io/stolostron/provider-credential-controller@sha256:d3bd513c06591786216f2baab46c9f57e373d1939f038edacecb2180f673d888
        - quay.io/stolostron/managedcluster-import-controller@sha256:925bbc6173501bdb60d329a1f5a134ea3e9230034509320f93382f1c570be275
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/multicloud-manager@sha256:05da692b2780453115ef6c195658cadb2f28eb0f3a86dc0a586e459dd7c55afe
        - quay.io/stolostron/cluster-proxy@sha256:6cd53c588bb9c5064ba0eb8eab03cd12c86976b353209102059f7d76ec59d7c3
        - quay.io/stolostron/clusterclaims-controller@sha256:a0613147b0ec010b8a4442a3c515f7ffc48ec7d8137436a0e8b935443cac2f96
        - quay.io/stolostron/cluster-image-set-controller@sha256:632661c6cbfce7df0545bc2fccf1ef87257abb3cc92dcede054806613a3540e3
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:26f8bb3c9b531e96e6538a5bff96569628f45c06eb3b991ebb421d4e90d7c7e3
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:a39c52097faf7a21cc717768401851466f8c9341f08103e4d1fe8459ae7f5ea2
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/assisted-installer@sha256:a31ed7b8e4ff2f9e5ade2e61191cc759e1d47ddbba95a2af41bfa53fa6ddf855
        - quay.io/stolostron/backplane-operator-test@sha256:961d8c40aa3526ca368fd6833202800ba2eb344c97788a646d6fb350a5338397
        - quay.io/stolostron/hive@sha256:fd39b8340cfaeb9849e83bd910e8a3575791fcc58dbbdeb3e66528f009b33d65
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:b76a201edcad4805e32c40d3f0ad3f61584a879aa00c9e49aae3b2985bfb5fbc
        - quay.io/stolostron/managed-serviceaccount@sha256:4c3f1e0b9c4f2937ceb46e6b3f63e3a07061f9b29fdda7dcfdf9ebd74c3bb30b
        - quay.io/stolostron/registration-operator@sha256:4b6c31592c405db11375f4694a7f15bb13c5c9a770786a1607ea0be615396435
        - quay.io/stolostron/assisted-installer-agent@sha256:a2efc62ca0080779acfea2e413e524608bc0e8f72b2b917e6c7e76864d08b2b6
        - quay.io/stolostron/hypershift-addon-operator@sha256:47ed6a7c00036599051718d6015bee35818ea5bbaa14f1c01bedb338dce90668
        version: 0.4.0
      entries:
      - name: stolostron-engine.v0.4.0
        version: 0.4.0
      name: community-0.4
    - currentCSV: stolostron-engine.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:30d5bcccc542988ac788dbca305e6f29fb678470b19e48adada9ca6f5803f04a
          createdAt: "2024-10-02T05:51:59Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.5.0 <0.5.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/discovery-operator-tests@sha256:e8d3d1675242272cc139b0dba15605d00e25b17b4b1cc1b5b4f3ea8913ffb2da
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:2fe53994fa5bde93112a2cc75758e0f68be013a34c8206066d0b315c2e6a6f47
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:4b5b2bea2aad6c7e16fe1520a2fb8454b0312f2749d2f398ca54e644e813055c
        - quay.io/stolostron/assisted-service@sha256:6383ba508a0cac030a46df96ac37daaeddc27ad1fb41d8b4dca5a87cf4ab9c31
        - quay.io/stolostron/cluster-image-set-controller@sha256:654e218ecf0a3f67520c0521b544eea9d4552e92f54bf424e88ccab24e7c151e
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:14a0648b76301de182154a8d7d6b7e7eca0bfbed226e6dcebddb75af5152451f
        - quay.io/stolostron/backplane-operator-test@sha256:048f21cf89fb5ed2557ffc90ae5d7104cdd5ce0e6a2a865ce6f4230d09400115
        - quay.io/stolostron/assisted-installer@sha256:dcc4e28c7eea55dd2ac22dd75c2b7b96ee3552fcbeba8128ad8528fda235d879
        - quay.io/stolostron/cluster-proxy-addon@sha256:7aa09b0d53fdc0f5cec9c8a13c160277a9274466025f11b5b04da95c9a10e59c
        - quay.io/stolostron/provider-credential-controller@sha256:d240ad43c23de822f91a33849448ae70a4cf27570da3387bdecd101ca2f0b65d
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/assisted-service-el8@sha256:c8248f9a7ac02f47373d097953cee38b66125e98b98922f18c578a591f04bcca
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/multicloud-manager@sha256:6dd212840ea7fadd427b503ce69e6e32464881a0047e2af86f9f2c7562bd4703
        - quay.io/stolostron/hypershift-operator@sha256:a30fb6cf4d5e103820813019ce452d8134b0fee8ecd3c2166d5681199d4a0cad
        - quay.io/stolostron/helloworld@sha256:96a9b8855fb8800a85aed6fdc4911295dc2c12629bbdcc34ed8dc8061827b17e
        - quay.io/stolostron/console-mce@sha256:10a54c5012ef6439539297919c71d9b53aaaddc6342d90677113d6ce7b42e3fe
        - quay.io/stolostron/hypershift-addon-operator@sha256:ba199ebfe0eea3ef6b07123f51116831e7b0a73860b0410ece723bcc9a3230af
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:b76a201edcad4805e32c40d3f0ad3f61584a879aa00c9e49aae3b2985bfb5fbc
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/cluster-curator-controller@sha256:fc2ad8e0b6701638ab89bb62a49700fdaadfe1149228cd0b3cd4a41810020c6d
        - quay.io/stolostron/cluster-api-provider-agent@sha256:1458de75311675a5acec7361e507edb24b970ffa3403e916de3413f7061bfe36
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        - quay.io/stolostron/assisted-image-service@sha256:703c12a918d3c66b62c29a5abd0e8c5032299289cfed459671d6ca82c4f35c73
        - quay.io/stolostron/registration@sha256:4779cf430d479806efd2f3cb7e2a9ed3dce749287609c39ad45f1b06e220d023
        - quay.io/stolostron/assisted-installer-agent@sha256:16b040dd0690b62bc051aa6369272d004d2157a8e5230d318ba827adba8e8d63
        - quay.io/stolostron/managedcluster-import-controller@sha256:01ab0c8b664692d69c2a9b3363e007a5611c332ac0e3491234dd0cbfba31c88a
        - quay.io/stolostron/image-based-install-operator@sha256:bd76503406e6d20693e636461411e158c8dde3d03c4904ecb8a1850deeefa4f5
        - quay.io/stolostron/placement@sha256:dc0ad451b085b3a747ee648475f8e8102b645df69b898fc88f1cb190d1b95a83
        - quay.io/stolostron/clusterclaims-controller@sha256:1c0b3c4730ef1ab628bf2c1b7cf12b1b93b149abdefc4d908dbdc90e4b6c1c19
        - quay.io/stolostron/backplane-operator@sha256:30d5bcccc542988ac788dbca305e6f29fb678470b19e48adada9ca6f5803f04a
        - quay.io/stolostron/addon-manager@sha256:9384ae0e9bcc9f85d3fe36e44ab7b3a61777197846559b5add0c56f513e35333
        - quay.io/stolostron/discovery-operator@sha256:db56e2a78ed27a46b47b519e29db66636b9abb499fd507d99451bae47438e32b
        - quay.io/stolostron/cluster-proxy@sha256:be09bb293a990600ec48aba2f9bf8b457d5bdbc6b136be319e5feebadb232a8c
        - quay.io/stolostron/work@sha256:4467a4bd00586ab837cb7e8bf56ea1faff82ed5ce2fffdeb7e51ee5fafb433b0
        - quay.io/stolostron/assisted-installer-controller@sha256:01dd7c7fb9cb335215afefde204309d3127847b900658ee9744b9d31df847565
        - quay.io/stolostron/hive@sha256:707ac9a2ad18dbc4c4686f1b72946077c7232afd05505d5e50e13f806a452392
        - quay.io/stolostron/backplane-must-gather@sha256:92e992e23855375a8bae542480d789d1bb25a48959896bfb75f659df2a37dd98
        - quay.io/stolostron/managed-serviceaccount@sha256:667f823799984404c0c1589df7b857a6a35218c2653254905dc22ca90ce1542b
        - quay.io/stolostron/discovery-testserver@sha256:0b23021d6c0c7d6c7aa97a92fd0c40f31fdff6363a5b10a1edc864b5d8d54108
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/registration-operator@sha256:b42341143b1e1693748f4b7d69e97c56c8a5f9408abc5db1de96ba246b7b98a9
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        version: 0.5.3
      entries:
      - name: stolostron-engine.v0.5.3
        version: 0.5.3
      - name: stolostron-engine.v0.5.2
        version: 0.5.2
      - name: stolostron-engine.v0.5.1
        version: 0.5.1
      - name: stolostron-engine.v0.5.0
        version: 0.5.0
      name: community-0.5
    - currentCSV: stolostron-engine.v0.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:fa1fa6ef3958fd9d50bfb38197f033a635a05b47c32e653822246e708d8598cc
          createdAt: "2025-02-24T01:37:53Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.6.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/console-mce@sha256:bfaf9180cd1e5af4d34daf4f1dcdee89951067a605f21bfe6e1ce6bda85b0334
        - quay.io/stolostron/registration@sha256:9c221a40ffa119ac154d76f578e0919b5a63137c908afb0583519a9b8a399b2b
        - quay.io/stolostron/cluster-proxy@sha256:b341dd7058cdffccb00a026908031fa123fec5ba58ba9307dbb873b75d250e4d
        - quay.io/stolostron/clusterclaims-controller@sha256:b796cb2edd3d4c22b29fc6d2b45e5328a59b7ec1bdc52d5500e1a4ee74725d6c
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:2781d87c1568f2e40409f5431bfd1a9a4c9ca35655537e7d4db65d9b881edd08
        - quay.io/stolostron/backplane-operator-test@sha256:b04661ad658423c9c7557b722adb2159d988c10445de1c59f39c44583262cb42
        - quay.io/stolostron/registration-operator@sha256:52289e6cf5f159d344717b49e6226ff9d4ee33942f63cc543d1b820d7ee24fc5
        - quay.io/stolostron/work@sha256:6ec081826f32467ea1e1a91f6624df349a382fb8a1cbbda906141281ac2bad5c
        - quay.io/stolostron/discovery-operator@sha256:43f40b9c20b112248fcdb42d114fe4af73fdd45b05d9ff7855d827cdd5e971b4
        - quay.io/stolostron/multicloud-manager@sha256:0293f0852206562b891375b167cfe17843e2e17585f04c9bb6482d1fe78089d0
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/image-based-install-operator@sha256:fecbff73eb97b28bedd7c17d5adbef4826f181a6838ad4cc72e7478acf9b4254
        - quay.io/stolostron/placement@sha256:543f09446652bff8c4149945d94d715eff7a76d82783f2bad89e712798542c73
        - quay.io/stolostron/assisted-installer@sha256:c88976ab4b35d458a39044e522d2d5ab1916d8485af07bbef373af1bcbbd6af0
        - quay.io/stolostron/assisted-service-el8@sha256:f50f4dece91100f45ef9e20b1e52557a55db45b17a380f32d365ee013d695d5c
        - quay.io/stolostron/assisted-service@sha256:58271e54eda979c43fb474c3bb255f7328e380a7af9ba833e5f7558f323bd239
        - quay.io/stolostron/ose_cluster_api_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/hypershift-operator@sha256:2d1809223ce5e75b4133a39b78a269703c6b5234f7ec2a7c0072124c345fae0a
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:65180432218d035ffc7996a7643ddb5300eb8b09e47e09cb270cb9b961aecf7a
        - quay.io/stolostron/helloworld@sha256:4838f699ea2461f8358dcda32b3ac96ed05b55d4e1958eac2bdacae4807501d1
        - quay.io/stolostron/hive@sha256:1e1f60492c6a3b43b69fe8a0640f9b81f5a1cfafc0afcd6854c3cfe362ed063a
        - quay.io/stolostron/assisted-installer-controller@sha256:3b0b43fdbbbbd5c1e7588fc449b002e1ed9ad0ea6daa774925a962bd499a4195
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        - quay.io/stolostron/cluster-image-set-controller@sha256:629361c1c0ec50fe6410fc6fe6417cba89957c49ab89201c6470c0f97c8fc91d
        - quay.io/stolostron/addon-manager@sha256:7e347a177d8c5b80a78411a3e8dbf04e49d157a4ca67e7fd5170970c2ba62da9
        - quay.io/stolostron/discovery-operator-tests@sha256:21449ecd291cf7203cc8e27a2a00ac60e74f5993708f9e7465e8efdfdca1c89f
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/backplane-operator@sha256:fa1fa6ef3958fd9d50bfb38197f033a635a05b47c32e653822246e708d8598cc
        - quay.io/stolostron/assisted-image-service@sha256:6846719f91dd954ce4b4c6e7c9654c7d79db72990270acd3c34bdd2a89af34fc
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:44f8bb72b975456d4151a88a560fc525ad6f4da7968a38e94a86b6dd77d3c543
        - quay.io/stolostron/cluster-proxy-addon@sha256:7da9da3e4ecd1e0df636c9f72f9f30cc7ff24b81e247ba9825cef39f18e86182
        - quay.io/stolostron/ose_aws_cluster_api_controllers_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:b016a001a53bc37ae534b74cc832552759800746c27e16adcf0626ff9ad2e4c0
        - quay.io/stolostron/discovery-testserver@sha256:f757a7a11108a88de4bc0ecc5e16f74b3bd9b5110ec94bfeef2074121ec1e47f
        - quay.io/stolostron/hypershift-addon-operator@sha256:65f2b7a356c22413307ce7b7de38ce00c31e4c43b6ee3b4b484c0bc61dc483e9
        - quay.io/stolostron/managed-serviceaccount@sha256:0e4edd9ab95c1558fba40a0c41c5cc326996e9e19e296324ea10b5a442a7ef63
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/cluster-curator-controller@sha256:5d05462c2c9d93b21656a436c01fbda4256757bd057e5de7014eb6f320310a1e
        - quay.io/stolostron/assisted-installer-agent@sha256:cd74c8b8888c100bf567dd07cfff40097c458d18fcc29a11ed3830f697156267
        - quay.io/stolostron/cluster-api-provider-agent@sha256:dda3e9aef611c744c4c89c2689a12d0649a173c1a64954c3037de55ec8eacec9
        - quay.io/stolostron/managedcluster-import-controller@sha256:ecfe7c1f2d24a75bfb8db2a0a2765fe5e9136f81f341882de896431570fd10e9
        - quay.io/stolostron/backplane-must-gather@sha256:99b558d687e1c17036b164d819cb1c9b96a814be79e3c87960ef7cdb521eee7d
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/provider-credential-controller@sha256:a5ac3999d0aefb1dd9a239f9394485f93e4252a45eb4a43a43e603aaff80b3bb
        version: 0.6.3
      entries:
      - name: stolostron-engine.v0.6.3
        version: 0.6.3
      - name: stolostron-engine.v0.6.2
        version: 0.6.2
      - name: stolostron-engine.v0.6.1
        version: 0.6.1
      - name: stolostron-engine.v0.6.0
        version: 0.6.0
      name: community-0.6
    - currentCSV: stolostron-engine.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:9b39a5b6a502778d7884e276b276a9b44e6dbc88e7e82470e26d91ded3cb3839
          createdAt: "2025-06-09T15:04:58Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.7.1'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational components for managing multiple Kubernetes-based clusters. The deployment of the multicluster engine components is
              determined based on the configuration that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/registration-operator@sha256:a9fdf837e149c8659fc99deab6a284975c8eb8785549381fa082303cdb172e44
        - quay.io/stolostron/discovery-testserver@sha256:faef24208a65283b4157521079a7bf23bc4ebd553d70a83a78c324b5cb0e24e4
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/hypershift-addon-operator@sha256:ed1c1580fee28310fc1b137c5c70271448e58ed29af4371515fe9c74512a2c7e
        - quay.io/stolostron/hypershift-operator@sha256:c9a8fc943ef35126434cc767f152abd09943a4d2d194a9966495fdc0870b417c
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/provider-credential-controller@sha256:3484cb0c2102f4a522efe31bc4c77910a202b42e1ad082a0a8709cc1a7698649
        - quay.io/stolostron/assisted-installer@sha256:17b78d73775a93b31ca872db79520ddc49d2be042fdf1599f419fbcbfd9a98ac
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/hive@sha256:86ff94d3e7c174ac4d6818183c66791edc7638965dd09e2d41172609c81f451f
        - quay.io/stolostron/ose_cluster_api_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/work@sha256:709c41d1a7a1eaf10b334dd0c0105a261d6db811693d5ae67eb55b9a26daed87
        - quay.io/stolostron/backplane-operator@sha256:9b39a5b6a502778d7884e276b276a9b44e6dbc88e7e82470e26d91ded3cb3839
        - quay.io/stolostron/cluster-image-set-controller@sha256:7f05ad13ffaa11daec066c5889b87df1b631e6eb3dc1c2913a2766c8e65218aa
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/cluster-proxy@sha256:1528e36dfa4fa2c9063d84c0f2b9aa8d2b549a750e812d4a2e2b8c7515bfe5cb
        - quay.io/stolostron/backplane-operator-test@sha256:8d6514fc0db5ee79c346ddc31bb094ad87fa375519da67e32398300d70d72f6e
        - quay.io/stolostron/cluster-api-provider-agent@sha256:2b22890b948ad1d597c424d73d4fe34c7c4d265ad6e54fa0ed87a7dec920ac57
        - quay.io/stolostron/assisted-service@sha256:af3532d487e03a077ce23e4178abbcb41cf45726b646d7660840c036677715e8
        - quay.io/stolostron/backplane-must-gather@sha256:f4f9596eade78333628199f62f8e71e4c2b3957c20da5bad807c12a2a02560db
        - quay.io/stolostron/addon-manager@sha256:a245847f135f96c4b4fcaa673fc32013c8e407252b3259175c796e3c16111c71
        - quay.io/stolostron/cluster-api-provider-openshift-assisted-control-plane@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/assisted-image-service@sha256:b6df9c9fc8bbfdc58e2169b9e199e1eed188d3d847511eba01eb4920fabe3992
        - quay.io/stolostron/ose_aws_cluster_api_controllers_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:44f8bb72b975456d4151a88a560fc525ad6f4da7968a38e94a86b6dd77d3c543
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/assisted-installer-agent@sha256:ffceb06ce038013291e092324d3c49091cd1b2ff1569abe77c29484075e28107
        - quay.io/stolostron/discovery-operator-tests@sha256:d96fe1f4f49ecfd4a27bedd87d7070a3c7f3f82b6ee8bb59e98a262d8cedcfc0
        - quay.io/stolostron/discovery-operator@sha256:a6a37237bc06f0635a936b803d3c53f0f0caf5426c76b37951c4e157b181ecb3
        - quay.io/stolostron/placement@sha256:6276e411c9f7cf7072d3d364569b6c3cd472742ab4ad48912f18cb4091d27a6f
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:1757b9ad635839087c4bef198d5e81f4abb7aee38bc74690021f9e8887366bcf
        - quay.io/stolostron/ip-address-manager@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        - quay.io/stolostron/multicloud-manager@sha256:5a95a5d1dbb17456d7ce62f1a517ae49ca780e25f3ebb74b32346de1df23b4c9
        - quay.io/stolostron/assisted-service-el8@sha256:f52c55691850a1f2f254b0e9377ba97493be337e012223e5a2b3190f03e1a349
        - quay.io/stolostron/ose-baremetal-cluster-api-controllers-rhel9@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/image-based-install-operator@sha256:9d6796007b616de60560f55f14546295c4e48b3768d059098fce9367674a3368
        - quay.io/stolostron/helloworld@sha256:a790795409f6e781f506b6a63f39f7d344f01efa30c3de87b1727dc2010d5bbf
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:1d4221cfc364743b58e15bad3afe9548f4151f06b0dbf0aec7c00715ba75abba
        - quay.io/stolostron/assisted-installer-controller@sha256:8ea9d12fa30825fe4f965a7d506af687e8de709f8d07dd75204747da2a6e09a8
        - quay.io/stolostron/cluster-proxy-addon@sha256:c776405c4b1b623500c69a12b7d76563646b27cf23bab74dc8f1361534cfb81e
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/cluster-curator-controller@sha256:d667bd0892812eb4e1f724679759469db335db4ab801f9d57ea8818d8cd13188
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:1e95f55d6918263e79de1c1e2be8a96cd3d35a346dde6d3e2fb484f48e567811
        - quay.io/stolostron/console-mce@sha256:7f538a21eb22ccd57765f47692da23be9cecfd3b7b65986f1413234bdf709530
        - quay.io/stolostron/managed-serviceaccount@sha256:4bcd5c9b2044254c28215fd78bf1e3fcd919578238986831fcea8055b2764696
        - quay.io/stolostron/clusterclaims-controller@sha256:afab145b12a52f81e1f2eaa189731e3a6287b54f2f1f1ad2976ad378fa7dfa3a
        - quay.io/stolostron/managedcluster-import-controller@sha256:4d0be693db0576eb2e9e3d409bf1479cc60784f5747ed1b25b6a2443c5e91e36
        - quay.io/stolostron/cluster-api-provider-openshift-assisted-bootstrap@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/registration@sha256:2db2c28af25007c77979cc456d929f5f077cd671cc4e0562c43efce2c42c4cd8
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        version: 0.7.1
      entries:
      - name: stolostron-engine.v0.7.1
        version: 0.7.1
      - name: stolostron-engine.v0.7.0
        version: 0.7.0
      name: community-0.7
    defaultChannel: community-0.7
    packageName: stolostron-engine
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pulp Community
      provider-url: https://github.com/pulp/pulp-operator
    name: pulp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pulp-operator.v1.0.0-alpha.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache_enabled": true,
                  "content": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "ingress_type": "Route",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/"
                  },
                  "storage_type": "File",
                  "worker": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example-pulp"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true,
                    "redis_storage_class": "standard"
                  },
                  "content": {
                    "replicas": 1
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "pulp",
                  "file_storage_access_mode": "ReadWriteOnce",
                  "file_storage_size": "2Gi",
                  "file_storage_storage_class": "standard",
                  "image_version": "nightly",
                  "image_web_version": "nightly",
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/",
                    "telemetry": false
                  },
                  "web": {
                    "replicas": 1
                  },
                  "worker": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "galaxy-example"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "api": {
                    "replicas": 1
                  },
                  "content": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "galaxy",
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "file_storage_storage_class": "standard",
                  "image": "quay.io/pulp/galaxy-minimal",
                  "image_version": "nightly",
                  "image_web": "quay.io/pulp/galaxy-web",
                  "image_web_version": "nightly",
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/api/galaxy/pulp/",
                    "telemetry": false
                  },
                  "signing_scripts_configmap": "signing-scripts",
                  "signing_secret": "signing-galaxy",
                  "storage_type": "File",
                  "web": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "worker": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "galaxy-example",
                  "deployment_type": "galaxy",
                  "instance_name": "pulp",
                  "postgres_configuration_secret": "galaxy-example-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "galaxy-example"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.0.0-alpha.1
          createdAt: "2021-03-26 16:57:40"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1alpha1
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1alpha1
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1alpha1
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://docs.pulpproject.org/](https://docs.pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          Note that Pulp operator works with three different types of service containers (the operator itself, the main service and the web service):

          |           | Operator | Main | Web |
          | --------- | -------- | ---- | --- |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[pulp-minimal](https://quay.io/repository/pulp/pulp-minimal?tab=tags) | [pulp-web](https://quay.io/repository/pulp/pulp-web?tab=tags) |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[galaxy-minimal](https://quay.io/repository/pulp/galaxy-minimal?tab=tags) | [galaxy-web](https://quay.io/repository/pulp/galaxy-web?tab=tags) |

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://docs.pulpproject.org/pulp_operator/container/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://docs.pulpproject.org/pulp_operator/](https://docs.pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://docs.pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: beta
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/pulp/pulp-operator:v1.0.0-alpha.1
        - quay.io/pulp/pulp-minimal:stable
        - quay.io/pulp/pulp-web:stable
        - docker.io/library/redis:latest
        - docker.io/library/postgres:13
        version: 1.0.0-alpha.1
      entries:
      - name: pulp-operator.v1.0.0-alpha.1
        version: 1.0.0-alpha.1
      name: alpha
    - currentCSV: pulp-operator.v1.0.0-beta.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "Pulp",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true
                  },
                  "content": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/"
                  },
                  "worker": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "Pulp",
                "metadata": {
                  "name": "galaxy-example"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "api": {
                    "replicas": 1
                  },
                  "content": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "galaxy",
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "file_storage_storage_class": "standard",
                  "image": "quay.io/pulp/galaxy-minimal",
                  "image_version": "4.6.3",
                  "image_web": "quay.io/pulp/galaxy-web",
                  "image_web_version": "4.6.3",
                  "ingress_type": "nodeport",
                  "nodeport_port": 30000,
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "analytics": false,
                    "ansible_api_hostname": "http://nodeport.local:30000",
                    "api_root": "/api/galaxy/pulp/",
                    "content_origin": "http://nodeport.local:30000",
                    "galaxy_feature_flags": {
                      "ai_deny_index": true,
                      "execution_environments": true
                    },
                    "token_server": "http://nodeport.local:30000/token/"
                  },
                  "signing_scripts": "signing-scripts",
                  "signing_secret": "signing-galaxy",
                  "web": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "worker": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "galaxy-example",
                  "deployment_type": "galaxy",
                  "postgres_configuration_secret": "galaxy-example-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "galaxy-example"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.0.0-beta.4
          createdAt: "2024-01-18T20:33:37Z"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1beta2
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1beta2
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1beta2
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://docs.pulpproject.org/](https://docs.pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          Note that Pulp operator works with three different types of service containers (the operator itself, the main service and the web service):

          |           | Operator | Main | Web |
          | --------- | -------- | ---- | --- |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[pulp-minimal](https://quay.io/repository/pulp/pulp-minimal?tab=tags) | [pulp-web](https://quay.io/repository/pulp/pulp-web?tab=tags) |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[galaxy-minimal](https://quay.io/repository/pulp/galaxy-minimal?tab=tags) | [galaxy-web](https://quay.io/repository/pulp/galaxy-web?tab=tags) |

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://docs.pulpproject.org/pulp_operator/container/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://docs.pulpproject.org/pulp_operator/](https://docs.pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://docs.pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: beta
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - docker.io/library/postgres:13
        - quay.io/pulp/pulp-minimal:stable
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.4
        - quay.io/pulp/pulp-operator:v1.0.0-beta.4
        - quay.io/pulp/pulp-web:stable
        - docker.io/library/redis:latest
        version: 1.0.0-beta.4
      entries:
      - name: pulp-operator.v1.0.0-beta.4
        version: 1.0.0-beta.4
      - name: pulp-operator.v1.0.0-beta.3
        version: 1.0.0-beta.3
      - name: pulp-operator.v1.0.0-beta.2
        version: 1.0.0-beta.2
      - name: pulp-operator.v1.0.0-beta.1
        version: 1.0.0-beta.1
      - name: pulp-operator.v1.0.0-alpha.9
        version: 1.0.0-alpha.9
      - name: pulp-operator.v1.0.0-alpha.8
        version: 1.0.0-alpha.8
      - name: pulp-operator.v1.0.0-alpha.7
        version: 1.0.0-alpha.7
      - name: pulp-operator.v1.0.0-alpha.6
        version: 1.0.0-alpha.6
      - name: pulp-operator.v1.0.0-alpha.5
        version: 1.0.0-alpha.5
      - name: pulp-operator.v1.0.0-alpha.4
        version: 1.0.0-alpha.4
      - name: pulp-operator.v1.0.0-alpha.3
        version: 1.0.0-alpha.3
      - name: pulp-operator.v1.0.0-alpha.2
        version: 1.0.0-alpha.2
      - name: pulp-operator.v1.0.0-alpha.1
        version: 1.0.0-alpha.1
      name: beta
    defaultChannel: beta
    packageName: pulp-operator
    provider:
      name: Pulp Community
      url: https://github.com/pulp/pulp-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-common-service-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-common-service-operator.v4.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v3",
                "kind": "CommonService",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-common-service-operator",
                    "app.kubernetes.io/managed-by": "ibm-common-service-operator",
                    "app.kubernetes.io/name": "ibm-common-service-operator"
                  },
                  "name": "example-commonservice"
                },
                "spec": {
                  "size": "starterset"
                }
              }
            ]
          capabilities: Seamless Upgrades
          cloudPakThemesVersion: styles4100.css
          containerImage: registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
          createdAt: "2025-03-18T21:27:31Z"
          description: The IBM Cloud Pak foundational services operator is used to
            deploy IBM foundational services.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-common-service-operator
          nss.operator.ibm.com/managed-webhooks: ""
          olm.skipRange: '>=3.3.0 <4.12.0'
          operatorChannel: v4.12
          operatorVersion: 4.12.0
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/ibm-common-service-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CommonService is the Schema for the commonservices API. Documentation
              For additional details regarding install parameters check https://ibm.biz/icpfs39install.
              License By installing this product you accept the license terms https://ibm.biz/icpfs39license
            displayName: CommonService
            kind: CommonService
            name: commonservices.operator.ibm.com
            version: v3
        description: |-
          # Introduction

          The `IBM Cloud Pak foundational services` is a bridge to connect IBM Cloud Paks and Operand Deployment Lifecycle Manager (ODLM) with IBM Cloud Pak foundational services. You can also install the `IBM Cloud Pak foundational services` in stand-alone mode.

          When you install this operator, the operator completes the following tasks:

            - Installs ODLM operator namespace
            - Creates the Common Services `OperandRegistry` and `OperandConfig` in the service namespace

          # Details
          For more information about installing this operator and other Common Services operators, see [Installer documentation](http://ibm.biz/cpcs_opinstall). If you are using this operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For more information about IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).

          For more information about the available IBM Cloud Pak foundational services, see the [IBM Knowledge Center](http://ibm.biz/cpcsdocs).

          ## Supported platforms

          Red Hat OpenShift Container Platform 4.5 or newer installed on one of the following platforms:

           - Linux x86_64
           - Linux on Power (ppc64le)
           - Linux on IBM Z and LinuxONE

          ## Operator versions

           - 4.12.0

          ## Prerequisites

          Before you install this operator, you need to first install the operator prerequisites:

          - For the list of prerequisites for installing the operator, see the IBM Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).

          ### Resources Required

          #### Minimum scheduling capacity

          | Software                   | Memory (MB) | CPU (cores) | Disk (GB) | Nodes  |
          | -------------------------- | ----------- | ----------- | --------- | ------ |
          | ibm common service operator | 200          | 0.2        | 1          | worker |
          | **Total**                  | 200         | 0.2         | 1         |        |

          ## Documentation

          - If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak. For a list of IBM Cloud Paks, see [IBM Cloud Paks that use IBM Cloud Pak foundational services](http://ibm.biz/cpcs_cloudpaks).
          - If you are using the operator in stand-alone mode or with an IBM Containerized Software, see the IBM Cloud Pak foundational services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall).

          ## Installing

          For installation, see the [IBM Cloud Platform Common Services documentation](http://ibm.biz/cpcsdocs).

          ## Configuration

          For configuration, see the [IBM Cloud Platform Common Services documentation](http://ibm.biz/cpcsdocs).

          ### SecurityContextConstraints Requirements

          The IBM Common Service Operator supports running with the OpenShift Container Platform 4.3 default restricted Security Context Constraints (SCCs) and IBM Cloud Pak Security Context Constraints (SCCs).

          For more information about the OpenShift Container Platform Security Context Constraints, see [Managing Security Context Constraints](https://docs.openshift.com/container-platform/4.3/authentication/managing-security-context-constraints.html).

          For more information about the IBM Cloud Pak Security Context Constraints, see [Managing Security Context Constraints](https://ibm.biz/cpkspec-scc).

          ### Limitations

          No
        displayName: IBM Cloud Pak foundational services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Common Service Operator
          url: https://github.com/IBM/ibm-common-service-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:4209e78f12120b9c9f320496b780c3993a9317825ea17dc7d554452b28c99014
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:9a537eae7e581322e2f0764ed38a797ba7cd260724a285f24d5095058b7b47a0
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-operator-bundle@sha256:086eaeaaaf440a1d4a254984afacd8d04583aafb9ec11ac072feb42a87b986f7
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90046eb6170e12439ccc5aa343f19a4d4fc99a10d98b383cf1dc0e020ebeed55
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:18f44ab15289625fa2c6b72ece6e06afb51058dcba30b8f38b4fdeedf8c9e51b
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:0fdf1beeb4888f648f6dbd17a29dcc53a715d2c4301bc714a4172b2daf8d13be
        version: 4.12.0
      entries:
      - name: ibm-common-service-operator.v4.12.0
        version: 4.12.0
      name: v4.12
    defaultChannel: v4.12
    packageName: ibm-common-service-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: rebellions
      provider-url: https://rebellions.ai/
    name: rbln-npu-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: candidate
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: fast
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: rbln-npu-operator
    provider:
      name: rebellions
      url: https://rebellions.ai/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/component: operator
      app.kubernetes.io/managed-by: kustomize
      app.kubernetes.io/name: zabbix-operator-certified
      app.kubernetes.io/part-of: zabbix-operator-certified
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zabbix LLC
      provider-url: ""
    name: zabbix-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zabbix-operator-certified-rhmp.v6.0.40
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent",
                "metadata": {
                  "labels": {
                    "app": "agent",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "log_remote_commands": true,
                  "max_lines_per_second": 20,
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent2",
                "metadata": {
                  "labels": {
                    "app": "agent2",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent2"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "enable_persistent_buffer": false,
                  "enable_status_port": false,
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "persistent_buffer_claim": "",
                  "persistent_buffer_period": "1h",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixFull",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-full"
                },
                "spec": {
                  "builtin_report_service": true,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysql_volumeclaim": "mysql-full-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxyMysql",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-mysql"
                },
                "spec": {
                  "db_server_port": 3306,
                  "enable_snmp_traps": false,
                  "internal_db": true,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix_proxy",
                  "proxy": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "zabbix_mysql_volumeclaim": "mysql-proxy-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxySqlite",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-sqlite"
                },
                "spec": {
                  "db_data_volumeclaim": "zabbix-proxy-sqlite",
                  "enable_snmp_traps": false,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "proxy": {
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "node_name_as_db_name": true,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "proxy_size": 1,
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  }
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixServer",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-server"
                },
                "spec": {
                  "builtin_report_service": true,
                  "db_server_host": "mysql-server",
                  "db_server_port": 3306,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix",
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              }
            ]
          ansible.operator-sdk/reconcile-period: 5s
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          certified: "true"
          containerImage: registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
          createdAt: "2025-05-26T07:41:22Z"
          description: Zabbix operator with multiple deployment variants and different
            components
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/zabbix-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/zabbix-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <6.0.40
          operatorframework.io/suggested-namespace: zabbix
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: ""
          support: Zabbix
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Zabbix agent 2 is a new generation of Zabbix agent and may
              be used in place of Zabbix agent
            displayName: Zabbix agent 2
            kind: ZabbixAgent2
            name: zabbixagent2s.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix agent is deployed on a monitoring nodes to actively
              monitor local resources and applications
            displayName: Zabbix agent
            kind: ZabbixAgent
            name: zabbixagents.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database server, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Full
            kind: ZabbixFull
            name: zabbixfulls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with MySQL database server
            displayName: Zabbix proxy (MySQL)
            kind: ZabbixProxyMysql
            name: zabbixproxymysqls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with SQLite database
            displayName: Zabbix proxy (SQLite3)
            kind: ZabbixProxySqlite
            name: zabbixproxysqlites.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database support, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Server
            kind: ZabbixServer
            name: zabbixservers.kubernetes.zabbix.com
            version: v1alpha1
        description: "## About this Operator\n\nZabbix helps you to real-time monitoring
          of millions of metrics collected from tens of thousands of servers, virtual
          machines and network devices.\nThe Zabbix Operator allows users to easily
          deploy, manage, and maintain Zabbix deployments on OpenShift. By installing
          this integration you will be able to deploy Zabbix server / proxies and
          other components with a single command.\n## Supported Features\n* **Zabbix
          Server** - Simple Zabbix installation with included Zabbix server, Zabbix
          web-interface and Zabbix Java Gateway with MySQL database support. The feature
          does not provide MySQL service and requires an external MySQL database.
          \n* **Zabbix Server (Full)** - Zabbix installation with included Zabbix
          server, Zabbix web-interface, Zabbix Java Gateway and MySQL server instance.\n*
          **Zabbix proxy (SQLite3)** - Very simple way to gain power of Zabbix proxy.
          The feature has  SQLite3 support for Zabbix proxies and allow to specify
          amount of proxies. \n* **Zabbix proxy (MySQL)** - Another option of Zabbix
          proxy. The option support and deliver MySQL database.\n* **Zabbix agent**
          - Zabbix agent can be deployed on each available node for stability and
          performance monitoring on remote nodes. It allows to gather metrics with
          full automation!* \n## Prerequisites\nAll deployment options are require
          additional information during deployment. Please, check the following instructions
          and provide required configuration:\n* **Zabbix Server** - MySQL database
          host information and MySQL database credentials in specially formatted *Secret*.
          Additionally it is possible to specify SSL certificates for HTTPS support
          in *Secret*.\n* **Zabbix Server (Full)** - MySQL database credentials in
          specially formatted *Secret*. MySQL database volume name information. Additionally
          it is possible to specify SSL certificates for HTTPS support in *Secret*.\n*
          **Zabbix proxy (SQLite3)** - Zabbix server host information only.\n* **Zabbix
          proxy (MySQL)** - MySQL database credentials in specially formatted *Secrets*
          and Zabbix server host information.\n* **Zabbix agent** - Zabbix server
          host information only for outgoing and incoming connections. Zabbix agent
          uses \"privileged\" mode to monitor node resources! For example, running
          processes.* \n### MySQL credentials\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-full-secrets\ndata:\n  mysql_root_password: emFiYml4X3N1cGVyX3Jvb3Q=
          [1]\n  mysql_zabbix_password: emFiYml4X3VzZXJfcGFzc3dvcmQ= [2]\n  mysql_zabbix_username:
          emFiYml4 [3]\ntype: Opaque\n```\nUsing MySQL root password (*mysqlrootpassword*
          [1]) Zabbix server / proxy will try to create MySQL database schema with
          grant permissions to *mysqlzabbixusername* [2] and *mysqlzabbixpassword*
          [3].\n### SSL certificates for HTTPS\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-web-sslsecret\ndata:\n  ssl.crt: >-\n   < ssl.crt data>\n
          \ ssl.key: >-\n\t < ssl.key data >\n  dhparam.pem: >-\n   <  dhparam.pem
          data >\n```\nFiles *ssl.crt*, *ssl.key* and *dhparam.perm* are required
          for Zabbix web-interface for SSL support.\n"
        displayName: Zabbix Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - zabbix
        - appliance
        - agent
        - metrics
        links:
        - name: Zabbix
          url: https://www.zabbix.com
        - name: Zabbix Official Documentation
          url: https://www.zabbix.com/documentation/6.0/en/manual/quickstart
        - name: Zabbix Operator Installation
          url: https://www.zabbix.com/documentation/6.0/en/manual/installation/containers/openshift
        - name: Downloads
          url: https://www.zabbix.com/download
        maintainers:
        - email: alexey.pustovalov@zabbix.com
          name: Alexey Pustovalov
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Zabbix LLC
        relatedImages:
        - registry.connect.redhat.com/zabbix/zabbix-agent2-60@sha256:424003d52f3cb4360adf2b3619d8b9f8de9196c630ee7e4ceb5d177c1c5a36e4
        - registry.connect.redhat.com/zabbix/zabbix-server-mysql-60@sha256:637484faec6e9a55049c13dbd1f100c03d757c3a21749388c0605040f02d05ce
        - registry.connect.redhat.com/zabbix/zabbix-proxy-mysql-60@sha256:c2ae74d97c083ff348dc9c55932a33da4779cd9efdd9d7c5a79acfffae5b9ed0
        - registry.connect.redhat.com/zabbix/zabbix-java-gateway-60@sha256:a4feb725c12c4ac90eb9244c6fafdb6708679f2e0e7f6aa683c765127306412e
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
        - registry.connect.redhat.com/zabbix/zabbix-web-mysql-60@sha256:611d2670e1bf95b81226976c0a1f64434b2a7766f196dadbd646a7d8319f4cb1
        - registry.connect.redhat.com/zabbix/zabbix-web-service-60@sha256:4a333094e110713470a897b6406c5e340e962caf32a429dddf19865041fcc448
        - registry.redhat.io/rhel9/mysql-80@sha256:c4b5246d0c0d089a766ea07768c22d9d6be8049ac1379bd345db99390ffe2a36
        - registry.connect.redhat.com/zabbix/zabbix-snmptraps-60@sha256:e61e6b1c551dec8488cce0998f94d83d6d10b5d89a465d7119b088313032cb81
        - registry.connect.redhat.com/zabbix/zabbix-proxy-sqlite-60@sha256:62e62ebc8c1bc13fb610f7a83412e65d852f74b3c2c212baa75ea27fc7fa03eb
        - registry.connect.redhat.com/zabbix/zabbix-agent-60@sha256:2895b5e45785aaf528d4a6f78c8af0a0d494dbe9362403fdd470673d911d5004
        - quay.io/zabbix/zabbix-operator@sha256:43d26488a19967bdb2a8aafde5f43820be95011811da9cc5ec623c9029801675
        version: 6.0.40
      entries:
      - name: zabbix-operator-certified-rhmp.v6.0.40
        version: 6.0.40
      - name: zabbix-operator-certified-rhmp.v6.0.38
        version: 6.0.38
      - name: zabbix-operator-certified-rhmp.v6.0.36
        version: 6.0.36
      - name: zabbix-operator-certified-rhmp.v6.0.33
        version: 6.0.33
      - name: zabbix-operator-certified-rhmp.v6.0.29
        version: 6.0.29
      - name: zabbix-operator-certified-rhmp.v6.0.27
        version: 6.0.27
      - name: zabbix-operator-certified-rhmp.v6.0.26
        version: 6.0.26
      - name: zabbix-operator-certified-rhmp.v6.0.25
        version: 6.0.25
      - name: zabbix-operator-certified-rhmp.v6.0.17
        version: 6.0.17
      - name: zabbix-operator-certified-rhmp.v6.0.14
        version: 6.0.14
      - name: zabbix-operator-certified-rhmp.v6.0.13
        version: 6.0.13
      name: lts
    defaultChannel: lts
    packageName: zabbix-operator-certified-rhmp
    provider:
      name: Zabbix LLC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: run-once-duration-override-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runoncedurationoverrideoperator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "RunOnceDurationOverride",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "runOnceDurationOverride": {
                    "spec": {
                      "activeDeadlineSeconds": 3600
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certifiedLevel: "false"
          containerImage: registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9-operator@sha256:54993506a3a2fc1e88c9b5aac97ae90d2a04a06ae6e3a97bf487b5e8996b95fb
          createdAt: 2024/07/01
          description: An operator to manage the OpenShift RunOnceDurationOverride
            Mutating Admission Webhook Server
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.1.2'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/run-once-duration-override-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RunOnceDurationOverride is the schema for the runoncedurationoverrides
              API
            displayName: Run Once Duration Override
            kind: RunOnceDurationOverride
            name: runoncedurationoverrides.operator.openshift.io
            version: v1
        displayName: Run Once Duration Override Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mutating
        - webhook
        - workload
        - run-once
        links:
        - name: Source Code
          url: https://github.com/openshift/run-once-duration-override
        - name: Source Code
          url: https://github.com/openshift/run-once-duration-override-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9-operator@sha256:54993506a3a2fc1e88c9b5aac97ae90d2a04a06ae6e3a97bf487b5e8996b95fb
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9@sha256:3dc6e2a5900472e3e083abcb659a13fea3d08f7a9ff167df524b6f890a16dc8e
        version: 1.1.2
      entries:
      - name: runoncedurationoverrideoperator.v1.1.2
        version: 1.1.2
      - name: runoncedurationoverrideoperator.v1.1.1
        version: 1.1.1
      - name: runoncedurationoverrideoperator.v1.1.0
        version: 1.1.0
      - name: runoncedurationoverrideoperator.v1.0.2
        version: 1.0.2
      - name: runoncedurationoverrideoperator.v1.0.1
        version: 1.0.1
      - name: runoncedurationoverrideoperator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: run-once-duration-override-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicast-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicast-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:ccd64090b85536ddaf1b81ca07d40279c9f277b1c5911628f11d678b98dc3bad
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:ccd64090b85536ddaf1b81ca07d40279c9f277b1c5911628f11d678b98dc3bad
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:f0e9f64c353fdbee3bf7ec0eaf070c3b75330ef659a881ec4ebc874ba584d085
        version: 0.5.3
      entries:
      - name: apicast-operator.v0.5.3
        version: 0.5.3
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      name: threescale-2.11
    - currentCSV: apicast-operator.v0.6.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:8c7fe66b8d4044ad965e7b3d149e9d8e3846ac4e55465873f7095a25b297b3df
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:8c7fe66b8d4044ad965e7b3d149e9d8e3846ac4e55465873f7095a25b297b3df
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:12f7a8e69a1369a51bbcd2b34d4ec8c66dd1fc9fb515349a04e7b202953423e3
        version: 0.6.4
      entries:
      - name: apicast-operator.v0.6.4
        version: 0.6.4
      - name: apicast-operator.v0.6.3
        version: 0.6.3
      - name: apicast-operator.v0.6.2
        version: 0.6.2
      - name: apicast-operator.v0.6.1
        version: 0.6.1
      - name: apicast-operator.v0.6.0
        version: 0.6.0
      - name: apicast-operator.v0.5.3
        version: 0.5.3
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      name: threescale-2.12
    - currentCSV: apicast-operator.v0.7.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:bde44c0fb9f13f212184b24e0bef45db734a8439a92179a2089e5aebd6b87b63
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          olm.skipRange: '>=0.6.2 <0.7.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.13/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:bde44c0fb9f13f212184b24e0bef45db734a8439a92179a2089e5aebd6b87b63
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:88dc7a289c803248efbe151999707325b1bc0aac2d26827ea1707e46e5c71652
        version: 0.7.5
      entries:
      - name: apicast-operator.v0.7.5
        version: 0.7.5
      name: threescale-2.13
    - currentCSV: apicast-operator.v0.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:c5e49be7b43a0058f974a5496976255ab3fbd0677bb7bcf9c3438341cf55965a
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.2 <0.8.3'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.14/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:c5e49be7b43a0058f974a5496976255ab3fbd0677bb7bcf9c3438341cf55965a
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:f8cd25658fa6a154e7ea6e5adf0f9dae3fc596ec9989021de6a021ec94eef226
        version: 0.8.3
      entries:
      - name: apicast-operator.v0.8.3
        version: 0.8.3
      - name: apicast-operator.v0.8.2
        version: 0.8.2
      - name: apicast-operator.v0.8.1
        version: 0.8.1
      name: threescale-2.14
    - currentCSV: apicast-operator.v0.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel9-operator@sha256:c4d6acae4e4cc3236c5b8d8dcfb339b072688ab114785d8598bbe68b1f3571a9
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=0.8.0 <0.12.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.14/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel9-operator@sha256:c4d6acae4e4cc3236c5b8d8dcfb339b072688ab114785d8598bbe68b1f3571a9
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        version: 0.12.5
      entries:
      - name: apicast-operator.v0.12.5
        version: 0.12.5
      - name: apicast-operator.v0.12.4
        version: 0.12.4
      - name: apicast-operator.v0.12.3
        version: 0.12.3
      - name: apicast-operator.v0.12.2
        version: 0.12.2
      - name: apicast-operator.v0.12.1
        version: 0.12.1
      - name: apicast-operator.v0.12.0
        version: 0.12.0
      name: threescale-2.15
    - currentCSV: apicast-operator.v0.12.1-mas
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-mas/apicast-rhel9-operator@sha256:e9f22a8af53f881f1948f5d6a033c35a530895633f2eae522b2fb67f9dc6088b
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.17"}]'
          olm.skipRange: '>=0.7.9 <0.12.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-mas/apicast-gateway-rhel8@sha256:f174636a9ad43a2419fff5a9e6797e3185c540a9509188fd1606aef4acea7bdf
        - registry.redhat.io/3scale-mas/apicast-rhel9-operator@sha256:e9f22a8af53f881f1948f5d6a033c35a530895633f2eae522b2fb67f9dc6088b
        version: 0.12.1-mas
      entries:
      - name: apicast-operator.v0.12.1-mas
        version: 0.12.1-mas
      - name: apicast-operator.v0.7.9-mas
        version: 0.7.9-mas
      - name: apicast-operator.v0.7.8-mas
        version: 0.7.8-mas
      - name: apicast-operator.v0.7.7-mas
        version: 0.7.7-mas
      - name: apicast-operator.v0.7.6-mas
        version: 0.7.6-mas
      - name: apicast-operator.v0.7.5-mas
        version: 0.7.5-mas
      - name: apicast-operator.v0.7.4-mas
        version: 0.7.4-mas
      - name: apicast-operator.v0.7.3-mas
        version: 0.7.3-mas
      - name: apicast-operator.v0.7.2-mas
        version: 0.7.2-mas
      - name: apicast-operator.v0.7.1-mas
        version: 0.7.1-mas
      - name: apicast-operator.v0.7.0-mas
        version: 0.7.0-mas
      - name: apicast-operator.v0.6.0-mas
        version: 0.6.0-mas
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      name: threescale-mas
    defaultChannel: threescale-2.15
    packageName: apicast-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: cloud-bulldozer
      provider-url: ""
    name: ripsaw
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ripsaw.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery, Cloud Provider
          containerImage: quay.io/cloud-bulldozer/benchmark-operator:latest
          description: '[Benchmark-operator](https://github.com/cloud-bulldozer/ripsaw/)
            is a benchmarking tool to benchmark your kubernetes cluster. The intent
            of this Operator is to deploy common workloads to establish a performance
            baseline of Kubernetes cluster'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/cloud-bulldozer/benchmark-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Benchmark
            name: benchmarks.ripsaw.cloudbulldozer.io
            version: v1alpha1
        description: '[Benchmark-operator](https://github.com/cloud-bulldozer/benchmark-operator/)
          is a benchmarking tool to benchmark your kubernetes cluster. The intent
          of this Operator is to deploy common workloads to establish a performance
          baseline of Kubernetes cluster'
        displayName: benchmark-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - performance
        - benchmark
        links:
        - name: Site
          url: https://github.com/cloud-bulldozer/benchmark-operator
        - name: Documentation
          url: https://github.com/cloud-bulldozer/benchmark-operator/tree/master/docs
        maintainers:
        - email: rsevilla@redhat.com
          name: rsevilla87
        - email: jtaleric@redhat.com
          name: jtaleric
        - email: rzaleski@redhat.com
          name: dry939
        - email: kwhitley@redhat.com
          name: whitleykeith
        maturity: alpha
        provider:
          name: cloud-bulldozer
        relatedImages:
        - quay.io/cloud-bulldozer/benchmark-operator:latest
        - k8s.gcr.io/redis:v1
        version: 0.1.0
      entries:
      - name: ripsaw.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ripsaw
    provider:
      name: cloud-bulldozer
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: 3scale-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: 3scale-operator.v0.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "ProductCRName": "product1-sample",
                  "Production": true
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/3scale-rhel7-operator@sha256:0751f9f5343a1264224c36db3db292e5a9e28c40cd96649ef48b5399c34668dd
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.16"}]'
          olm.skipRange: '>=0.10.0 <0.11.13'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/3scale-rhel7-operator@sha256:0751f9f5343a1264224c36db3db292e5a9e28c40cd96649ef48b5399c34668dd
        - registry.redhat.io/3scale-amp2/backend-rhel8@sha256:cdc53799ff9d022462a9a4afcbcd8043497cd40ccdc6fafdcb9ccf7ce69f43d1
        - registry.redhat.io/3scale-amp2/system-rhel7@sha256:61dcc91e5d3c91a0ea85c02f2599cadcb5496b1c09a8838b4e1c8e4efe126e8f
        - registry.redhat.io/3scale-amp2/memcached-rhel7@sha256:7e2ab1e24646fbfef878eb2b5a74ed865ae812b60aa0ba505045f838f26ebbfc
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:a95d09fc3b224f550dec3de3d23fd2dbfc0a220dc869b4ad9559ee2f85327daa
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:f8cd25658fa6a154e7ea6e5adf0f9dae3fc596ec9989021de6a021ec94eef226
        - registry.redhat.io/rhel8/redis-6@sha256:5250d644e80aea794b0f732e9285883a4348dace83557f177f64d78fe4dcabd0
        - registry.redhat.io/openshift4/ose-cli@sha256:1a8caae40be21f23ba3cb48b75809a73333907033a6041927aa568b5c9290f3c
        - registry.redhat.io/3scale-amp2/searchd-rhel7@sha256:8e2001c021fad554cb8a96bb34e9973010fe4206526f19aa040fa196709fb688
        - registry.redhat.io/3scale-amp2/zync-rhel9@sha256:0d2205d9f61a58414c4ba472ce8f19b12d2df722772847d5fb7442e664759627
        - registry.redhat.io/rhel8/mysql-80@sha256:08547a93659ad804d637b91e8e1eb3464b75783d494e6c40e1c921bd74f31645
        version: 0.11.13
      entries:
      - name: 3scale-operator.v0.11.13
        version: 0.11.13
      - name: 3scale-operator.v0.11.12
        version: 0.11.12
      name: threescale-2.14
    - currentCSV: 3scale-operator.v0.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                },
                "status": {
                  "deployments": {
                    "ready": [
                      "deployment x ready"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                },
                "status": {
                  "adminId": 1,
                  "tenantId": 2
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ApplicationAuth",
                "metadata": {
                  "name": "applicationauth-sample"
                },
                "spec": {
                  "applicationCRName": "application-cr-name",
                  "authSecretRef": {
                    "name": "auth-secret-reference"
                  },
                  "generateSecret": true
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "productCRName": "product1-sample",
                  "production": true
                },
                "status": {}
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/3scale-rhel9-operator@sha256:fcea6a77d0a48f35f4c3ce63e72dde104aac6b580f340950b311eb418fff5ec9
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=0.11.0 <0.12.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          rht_backend_redis_requirements: 6.2.0
          rht_mysql_requirements: 8.0.0
          rht_postgres_requirements: 10.0.0
          rht_system_redis_requirements: 6.2.0
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: ApplicationAuth is the Schema for the applicationauths API
            displayName: Application Auth
            kind: ApplicationAuth
            name: applicationauths.capabilities.3scale.net
            version: v1beta1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/3scale-rhel9-operator@sha256:fcea6a77d0a48f35f4c3ce63e72dde104aac6b580f340950b311eb418fff5ec9
        - registry.redhat.io/3scale-amp2/manticore-rhel9@sha256:ccf6baf5f2e9494a34739a79abe0f46e36d8ab9dc81c426f7a3a8964e8f39859
        - registry.redhat.io/3scale-amp2/backend-rhel8@sha256:eb2288de900dc42ead755b28cdb2509b25ecc8dd256781beaa4a2a86e7cbe532
        - registry.redhat.io/rhel9/memcached@sha256:8738110ef16c9e35deeb07c0577ffe09d6ef029761b504a8d20253fa715da394
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:a95d09fc3b224f550dec3de3d23fd2dbfc0a220dc869b4ad9559ee2f85327daa
        - registry.redhat.io/3scale-amp2/zync-rhel9@sha256:df9d62897ff286165772150a8b5e6952d2ce3b52a8afa07bd693a2d95c2c67ff
        - registry.redhat.io/openshift4/ose-cli@sha256:1a8caae40be21f23ba3cb48b75809a73333907033a6041927aa568b5c9290f3c
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/rhel8/mysql-80@sha256:ee9aca9096473da004f9140305742eb7b9a77ed81ce55d5a6503388cccab030f
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        - registry.redhat.io/3scale-amp2/system-rhel8@sha256:4c72dfeb1919a1de4fbc47f4611d0c830ef89cf9f17255ad55a9b370618e0e8f
        version: 0.12.5
      entries:
      - name: 3scale-operator.v0.12.5
        version: 0.12.5
      - name: 3scale-operator.v0.12.4
        version: 0.12.4
      - name: 3scale-operator.v0.12.3
        version: 0.12.3
      - name: 3scale-operator.v0.12.2
        version: 0.12.2
      - name: 3scale-operator.v0.12.1
        version: 0.12.1
      - name: 3scale-operator.v0.12.0
        version: 0.12.0
      name: threescale-2.15
    - currentCSV: 3scale-operator.v0.12.1-mas
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                },
                "status": {
                  "deployments": {
                    "ready": [
                      "deployment x ready"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                },
                "status": {
                  "adminId": 1,
                  "tenantId": 2
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ApplicationAuth",
                "metadata": {
                  "name": "applicationauth-sample"
                },
                "spec": {
                  "applicationCRName": "application-cr-name",
                  "authSecretRef": {
                    "name": "auth-secret-reference"
                  },
                  "generateSecret": true
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "productCRName": "product1-sample",
                  "production": true
                },
                "status": {}
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-mas/3scale-rhel9-operator@sha256:8f41ab6027616e4bb65b1fddbcf44cc8a25a79cdda19d5cc354ac4228a15141e
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.17"}]'
          olm.skipRange: '>=0.11.8 <0.12.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          rht_backend_redis_requirements: 6.2.0
          rht_mysql_requirements: 8.0.0
          rht_postgres_requirements: 10.0.0
          rht_system_redis_requirements: 6.2.0
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: ApplicationAuth is the Schema for the applicationauths API
            displayName: Application Auth
            kind: ApplicationAuth
            name: applicationauths.capabilities.3scale.net
            version: v1beta1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-mas/3scale-manticore-search-rhel9@sha256:7379186175c9c0f0a6d922cd2a0754b429c9728dfbe8979a9d6025789ebae0da
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:a95d09fc3b224f550dec3de3d23fd2dbfc0a220dc869b4ad9559ee2f85327daa
        - registry.redhat.io/3scale-mas/backend-rhel8@sha256:d0c561c6b9dfe9741722ba4a257fbd2df4d2a709457a0b0657ef7bcca1f2a5ca
        - registry.redhat.io/3scale-mas/zync-rhel9@sha256:78e5223a1a878f6527bd88d382948c125b7cb4a81540648244553af426d1a944
        - registry.redhat.io/3scale-mas/memcached-rhel7@sha256:7a1abee11b4120f399485b438bda1e5262f04de28d3e788b207eecc3a9de9ed2
        - registry.redhat.io/3scale-mas/3scale-rhel9-operator@sha256:8f41ab6027616e4bb65b1fddbcf44cc8a25a79cdda19d5cc354ac4228a15141e
        - registry.redhat.io/rhel8/mysql-80@sha256:754f6830e0bc4aadb973b1f613f2bc7278ec13ad0f0e1e87655a2bb2f76cb9b3
        - registry.redhat.io/openshift4/ose-cli@sha256:a688994ea404ce6ce4bb5ec1249f0e779de0d6c187b5078e2ff067e60b50937d
        - registry.redhat.io/3scale-mas/apicast-gateway-rhel8@sha256:f174636a9ad43a2419fff5a9e6797e3185c540a9509188fd1606aef4acea7bdf
        - registry.redhat.io/rhel8/redis-6@sha256:218f7fa2458c274cc34f79190423e0aba03027f55ef6b3334d916951a668b7b7
        - registry.redhat.io/3scale-mas/system-rhel8@sha256:0fa28aaf3467fbd22fc1335dc6a712043e4dead694c3f14b60e89740bb7eb13e
        version: 0.12.1-mas
      entries:
      - name: 3scale-operator.v0.12.1-mas
        version: 0.12.1-mas
      - name: 3scale-operator.v0.11.8-mas
        version: 0.11.8-mas
      - name: 3scale-operator.v0.11.7-mas
        version: 0.11.7-mas
      - name: 3scale-operator.v0.11.6-mas
        version: 0.11.6-mas
      - name: 3scale-operator.v0.11.5-mas
        version: 0.11.5-mas
      - name: 3scale-operator.v0.11.3-mas
        version: 0.11.3-mas
      - name: 3scale-operator.v0.11.2-mas
        version: 0.11.2-mas
      - name: 3scale-operator.v0.11.1-mas
        version: 0.11.1-mas
      - name: 3scale-operator.v0.11.0-mas
        version: 0.11.0-mas
      - name: 3scale-operator.v0.10.0-mas
        version: 0.10.0-mas
      name: threescale-mas
    defaultChannel: threescale-2.15
    packageName: 3scale-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/streamshub
    name: amq-streams-console
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: alpha
    - currentCSV: amq-streams-console.v2.8.0-13-0.1738265629.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:1b1b80710885b43600e22ff2a24930b5cab06cd91fad439ad8a282c00b4564b9
          createdAt: "2024-11-12T15:34:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-streams-console.v2.8.0-13
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.8
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.23.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:1b1b80710885b43600e22ff2a24930b5cab06cd91fad439ad8a282c00b4564b9
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:38b80d81e54d59667c1f7dfa4da4125d5c7430816da130ddc84678a4e3cff6cf
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:366d60ff48fb9ba32dc73d92a775046d7b1eb1df21c93e31605372c7858f16ca
        version: 2.8.0-13+0.1738265629.p
      entries:
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.8.x
    - currentCSV: amq-streams-console.v2.9.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
          createdAt: "2025-09-22T10:41:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0-0 <2.9.2-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.9
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:839368767bec735cd12f002874917c23f58550fe21df976858bdc92946985ec6
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:b7444b70a5dc4543a9976c08a619ff6978c9ff86ac7085d300b4870b7c6b7851
        version: 2.9.2-2
      entries:
      - name: amq-streams-console.v2.9.2-2
        version: 2.9.2-2
      - name: amq-streams-console.v2.9.2-1
        version: 2.9.2-1
      - name: amq-streams-console.v2.9.1-7
        version: 2.9.1-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.9.x
    - currentCSV: amq-streams-console.v2.9.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
          createdAt: "2025-09-22T10:41:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0-0 <2.9.2-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.9
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:839368767bec735cd12f002874917c23f58550fe21df976858bdc92946985ec6
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:b7444b70a5dc4543a9976c08a619ff6978c9ff86ac7085d300b4870b7c6b7851
        version: 2.9.2-2
      entries:
      - name: amq-streams-console.v2.9.2-2
        version: 2.9.2-2
      - name: amq-streams-console.v2.9.2-1
        version: 2.9.2-1
      - name: amq-streams-console.v2.9.1-7
        version: 2.9.1-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.x
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-3.0.x
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-3.x
    defaultChannel: alpha
    packageName: amq-streams-console
    provider:
      name: Red Hat
      url: https://github.com/streamshub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat CCIT
      provider-url: ""
    name: reportportal-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: reportportal-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ReportPortal",
                "metadata": {
                  "name": "reportportal-sample"
                },
                "spec": {
                  "accesskey": "AKIAIOSFODNN7EXAMPLE",
                  "analyzer_image": "quay.io/waynesun09/service-auto-analyzer:5.3.5",
                  "api_image": "quay.io/waynesun09/service-api:5.3.5-rootless",
                  "api_replicas": 1,
                  "app_domain": "apps.test-example.com",
                  "bucket_prefix": null,
                  "default_bucket_name": null,
                  "enable_pg_restic_backup": "yes",
                  "es_backup_schedule": "@daily",
                  "es_s3_backup_dir": "s3_backup",
                  "es_snapshot_bucket": "es-snapshot-123123",
                  "gateway_replicas": 1,
                  "index_image": "quay.io/waynesun09/service-index:5.0.10",
                  "index_replicas": 1,
                  "integration_salt_path": null,
                  "migration_image": "quay.io/waynesun09/migrations:5.3.5",
                  "pg_restic_password": "rp_user",
                  "pg_restic_s3_bucket_init": "yes",
                  "pg_s3_bucket_name": "pgbackup-123123",
                  "s3_enabled": false,
                  "s3_endpoint": null,
                  "secretkey": "wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY",
                  "uat_image": "quay.io/waynesun09/service-authorization:5.3.5",
                  "ui_image": "quay.io/waynesun09/service-ui:5.3.5",
                  "ui_replicas": 1
                }
              },
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ReportPortalRestore",
                "metadata": {
                  "name": "reportportalrestore-sample"
                },
                "spec": {
                  "accesskey": "AKIAIOSFODNN7EXAMPLE",
                  "bucket_prefix": null,
                  "es_restore": "yes",
                  "es_restore_date": "2020-08-07",
                  "es_s3_backup_dir": "s3_backup",
                  "es_snapshot_bucket": "es-snapshot-123123",
                  "pg_recovery": "yes",
                  "pg_recovery_target_time": "2020-08-07 00:00:00",
                  "pg_restic_password": "rp_user",
                  "pg_s3_bucket_name": "pgbackup-123123",
                  "s3_endpoint": null,
                  "secretkey": "wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY"
                }
              },
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ServiceApiServiceMonitor",
                "metadata": {
                  "name": "serviceapiservicemonitor-sample"
                },
                "spec": {
                  "bearer_token_secret_key": "token",
                  "bearer_token_secret_name": "rp-user-auth"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/waynesun09/rp5-operator:v0.0.8
          createdAt: "2021-04-22 14:39:00"
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/waynesun09/reportportal-operator
          support: waynesun09
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ReportPortal Restore from s3 backup
            kind: ReportPortalRestore
            name: reportportalrestores.rp5.reportportal.io
            version: v1
          - description: ReportPortal v5 Deployment with ReportPortal Operator
            kind: ReportPortal
            name: reportportals.rp5.reportportal.io
            version: v1
          - description: ReportPortal service-api ServiceMonitor for metrics
            kind: ServiceApiServiceMonitor
            name: serviceapiservicemonitors.rp5.reportportal.io
            version: v1
        description: |-
          ReportPortal is a test automation analytics platform. ReportPortal Operator support instantiating the ReportPortal services and required dependencies.

          What services are included:

              - Report Portal v5.3
              - PostgreSQL
              - Elasticsearch
              - RabbitMQ
              - Minio

          The Operator also support data backup & restore and metrics monitoring.

          ## Install

          Install the Operator to a namespace, then create ReportPortal v5 instance under Provided APIs on the UI, all Report Portal v5 services and dependencies will be installed.

          ## Data backup and restore

          The Report Portal v5 data backup and restore in the operator include two parts, PostgreSQL database data and Elasticsearch index data.

          The Operator support automatically data backup and restore for both.

          ### PostgreSQL database WAL archiving and recovery

          The Operator supports PostgreSQL WAL archiving and recovery point-in-time via Minio S3 using restic.

          #### WAL archiving

          When deploy Report Portal v5 instance, make sure `enable_pg_restic_backup` option is set as `yes`, then the WAL archiving will be enabled when initiate the database, the WAL archive data will be automatically backup to S3 storage via Minio using restic command.

          Parameter `pg_restic_s3_bucket_init` is for init the bucket if not exist, `pg_s3_bucket_name` is the bucket name and `pg_restic_password` is the restic password for encryption.

          #### WAL recovery

          WAL archive recovery will download all WAL archive files to the Postgres container and restore point-in-time.

          To restore PostgreSQL database point-in-time, on Operator UI select `ReportPortalRestore` under Provided APIs to create ReportPortalRestore operand.

          After the ReportPortalRestore operand created, the PG WAL restore will get started, as restore time will base on the database size, the process will last in minutes, hours or days.

          **Note:** Current Postgres container liveless probe timeout is set as one hour in the postgresql statefulset, it might cause restore fail if restore time is longer than one hour, so it might need be adjusted to be longer.

          ### Elasticsearch snapshot and restore

          The operator supports elasticsearch snapshot and restore with data saved in S3 bucket via Minio, the elasticsearch container have installed the s3 repository plugin.

          #### Snapshot

          When deploy Report Portal v5 instance, make sure following elasticsearch parameters are set:

              es_s3_backup_dir: s3_backup
              es_snapshot_bucket: es-snapshot-123123
              es_backup_schedule: '@daily'

          #### Restore

          Elasticsearch restore support restore to certain date as default snapshot schedule is daily.

          To restore elasticsearch from snapshot, on Operator UI select `ReportPortalRestore` under Provided APIs to create ReportPortalRestore operand.

          The backup dir and bucket name should be same as when create the RP instance, `es_restore` need set to `yes` to enable the restore and `es_restore_date` is the restore target date.

          The restore time will also be depend on the snapshot size.

          ## Metrics

          The operator by default have enabled metrics, the default ansible operator port 8080 is used for the operator container.

          All Report Portal v5 depended services, rabbitmq, minio, postgresql and elasticsearch have metrics enabled in this operator.

          The Report Portal gatway service have metrics enabled.

          By default several ServiceMonitors have been defined when create ReportPortal v5 operand:

              $ oc get servicemonitor
              NAME                    AGE
              elasticsearch-metrics   1d21h
              gateway-metrics         1d21h
              minio-metrics           1d21h
              postgres-metrics        1d21h
              rabbitmq-metrics        1d21h

          To export Report Portal v5 service-api metrics with ServiceMonitor, create a `secret` with RP user api token. The api access token could be found with login to Report Portal -> Click user icon -> Profile.

          Then select `ServiceApiServiceMonitor` under provided APIs to Create Instance on operator UI.

          Input the secret name and key name that have been created, and click create, a api-metrics servicemonitor will be created.

          To check the metrics, if [user workload monitoring](https://docs.openshift.com/container-platform/4.5/monitoring/monitoring-your-own-services.html) is enabled on the cluster, click to Monitoring on OpenShift UI and navigate to Metrics, where you could query the metrics.

          You could also deploy Promethues Operator in the current project and query via it.

          ## Doc

          More detail with installation, data backup & restore and metrics could be found at https://github.com/waynesun09/reportportal-operator#reportportal-operator
        displayName: reportportal-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - reportportal
        - failure analysis
        - test automation
        links:
        - name: Documentation
          url: https://github.com/waynesun09/reportportal-operator/blob/master/README.md
        - name: ReportPortal.io
          url: https://reportportal.io
        maintainers:
        - email: gsun@redhat.com
          name: Wayne Sun
        maturity: alpha
        provider:
          name: Red Hat CCIT
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.7.0
        - quay.io/waynesun09/rp5-operator:v0.0.8
        version: 0.0.8
      entries:
      - name: reportportal-operator.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: reportportal-operator
    provider:
      name: Red Hat CCIT
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://www.redhat.com/
    name: analytics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: analytics-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyData",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalydata-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalydata",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalydata-sample"
                },
                "spec": {
                  "anomalyname": "etcd_object_namespaces",
                  "config": {
                    "max": 100,
                    "min": 60,
                    "query": "max(apiserver_storage_objects{resource=~\"namespaces\"}) by (resource)"
                  },
                  "method": "min_max",
                  "metricdata": {
                    "groupeddata": "{ \"resource\": \"namespaces\" }",
                    "latestvalue": 121,
                    "timestamp": 1695804362
                  }
                }
              },
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalyengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalyengine",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalyengine-sample"
                },
                "spec": {
                  "anomalyqueryconfiguration": "etcd_objects_all:\n  method: percentage_change\n  query: max(apiserver_storage_objects{resource!~\"certificatesigningrequests.certificates.k8s.io|installplans.operators.coreos.com|operators.operators.coreos.com|subscriptions.operators.coreos.com|anomalydata.observability-analytics.redhat.com|anomalyengines.observability-analytics.redhat.com\"}) by (resource)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: True\nfailed_pods:\n  method: min_max\n  query: count(kube_pod_status_phase{phase!~\"Succeeded|Pending|Running|Unknown\"}) by (phase, namespace)\n  min: 0\n  max: 10\n  have_multi_result_data: True\nworkload_memory_usage_bytes:\n  method: percentage_change\n  query: sum(container_memory_working_set_bytes{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nworkload_cpu_usage_cores:\n  method: percentage_change\n  query: sum(container_cpu_usage_seconds_total{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nrest_client_requests_total:\n  method: percentage_change\n  query: sum(rest_client_requests_total)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\n",
                  "cronjobconfig": {
                    "anomalyqueries": "etcd_objects_all,failed_pods,workload_memory_usage_bytes,workload_cpu_usage_cores,rest_client_requests_total",
                    "name": "osa-anomaly-detection",
                    "resource": {
                      "cpulimit": "256m",
                      "cpurequest": "128m",
                      "memorylimit": "512Mi",
                      "memoryrequest": "256Mi"
                    },
                    "schedule": "*/5 * * * *"
                  },
                  "namespace": "osa-anomaly-detection",
                  "serviceaccountrolebinding": {
                    "clusterrolename": "cluster-monitoring-view",
                    "name": "osa-in-cluster-thanos-querier"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
          createdAt: "2024-03-12T08:02:25Z"
          description: Deploys and Manages Observability Analytics Operator
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4-alpha
          repository: https://github.com/openshift/analytics-operator
          support: RedHat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnomalyData is the Schema for the anomalydata API
            displayName: Anomaly Data
            kind: AnomalyData
            name: anomalydata.observability-analytics.redhat.com
            version: v1alpha1
          - description: AnomalyEngine is the Schema for the anomalyengines API
            displayName: Anomaly Engine
            kind: AnomalyEngine
            name: anomalyengines.observability-analytics.redhat.com
            version: v1alpha1
        description: "### Introduction\nObservability Analytics for Red Hat OpenShift
          runs OpenShift analytics processes like anomaly detection.\n\n### API Backward
          Compatibility\n**NOTE:** Until the Operator supports **seamless upgrades**,
          a new version of the operator may introduce a change that is **NOT** backward-compatible
          with the previous version. Thus, to upgrade the operator, uninstall the
          already installed version completely (including CRDs) and install the new
          version.\n\n### Installtion Steps for the Anomaly Engine\n Create namespace
          if you would like to install this operator into separate namespace (You
          can utilize existing namespace as well). ```bash kubectl create ns <Namespace>
          ``` Install the Observability Analytics operator to specific `<Namespace>`,
          wait until you see the operator in the **Installed Operator** list. Click
          on the Operator name and go to Operator details.\n\nHere you should see
          a tab called **Anomaly Engine**. Click on it, and then click on the **Create
          AnomalyEngine** button, which will redirect you to a form where you need
          to provide the required inputs for the Anomaly Engine. Once successfully
          created, you should see a cronjob in the given namespace. This cronjob will
          identify anomalies and store them into **anomalydata**.\n\n### How to See
          Anomaly Data Captured by Anomaly Engine?\nWe are storing anomaly data into
          CR.\n ####Option 1:Through `kubectl`` command\nYou can check detected anomalies
          using the following command from your terminal. \n\n```bash\nkubectl get
          anomalydata -n <Namespace>\nkubectl describe anomalydata <NAME of the anomalydata>
          -n <Namespace> \n```\n #### Option 2: From Installed Operator UI Go to installed
          operator and select **Observability Analytics** operator, now go to **Anomaly
          Data** tab where you can able to see the detected anomaly, click on specific
          anomaly and you should be able to see details of the same. ### Documentation\nDocumentation
          around how to test Operator codebase as a developer can be found [here](https://github.com/openshift/analytics-operator#readme)\n\n###
          Contributing\nYou can contribute by:\n  * Raising [issues](https://github.com/openshift/analytics-operator/issues)\n
          \   related to observability-analytics-operator\n  * Fixing issues by opening
          [Pull Requests](https://github.com/openshift/analytics-operator/pulls)\n
          \ * Discussing observability-analytics-operator\n\n### License\nobservability-analytics-operator
          is licensed under the Apache 2.0 license.\n"
        displayName: Observability Analytics
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - observability-analytics-operator
        - analytics-operator
        - analytics
        - anomaly
        links:
        - name: Analytics Operator
          url: https://analytics-operator.domain
        maintainers:
        - email: rzalavad@redhat.com
          name: Raj
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
        version: 0.1.1
      entries:
      - name: analytics-operator.v0.1.1
        version: 0.1.1
      - name: analytics-operator.v0.1.0
        version: 0.1.0
      name: alpha
    - currentCSV: analytics-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyData",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalydata-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalydata",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalydata-sample"
                },
                "spec": {
                  "anomalyname": "etcd_object_namespaces",
                  "config": {
                    "max": 100,
                    "min": 60,
                    "query": "max(apiserver_storage_objects{resource=~\"namespaces\"}) by (resource)"
                  },
                  "method": "min_max",
                  "metricdata": {
                    "groupeddata": "{ \"resource\": \"namespaces\" }",
                    "latestvalue": 121,
                    "timestamp": 1695804362
                  }
                }
              },
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalyengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalyengine",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalyengine-sample"
                },
                "spec": {
                  "anomalyqueryconfiguration": "etcd_objects_all:\n  method: percentage_change\n  query: max(apiserver_storage_objects{resource!~\"certificatesigningrequests.certificates.k8s.io|installplans.operators.coreos.com|operators.operators.coreos.com|subscriptions.operators.coreos.com|anomalydata.observability-analytics.redhat.com|anomalyengines.observability-analytics.redhat.com\"}) by (resource)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: True\nfailed_pods:\n  method: min_max\n  query: count(kube_pod_status_phase{phase!~\"Succeeded|Pending|Running|Unknown\"}) by (phase, namespace)\n  min: 0\n  max: 10\n  have_multi_result_data: True\nworkload_memory_usage_bytes:\n  method: percentage_change\n  query: sum(container_memory_working_set_bytes{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nworkload_cpu_usage_cores:\n  method: percentage_change\n  query: sum(container_cpu_usage_seconds_total{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nrest_client_requests_total:\n  method: percentage_change\n  query: sum(rest_client_requests_total)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\n",
                  "cronjobconfig": {
                    "anomalyqueries": "etcd_objects_all,failed_pods,workload_memory_usage_bytes,workload_cpu_usage_cores,rest_client_requests_total",
                    "name": "osa-anomaly-detection",
                    "resource": {
                      "cpulimit": "256m",
                      "cpurequest": "128m",
                      "memorylimit": "512Mi",
                      "memoryrequest": "256Mi"
                    },
                    "schedule": "*/5 * * * *"
                  },
                  "namespace": "osa-anomaly-detection",
                  "serviceaccountrolebinding": {
                    "clusterrolename": "cluster-monitoring-view",
                    "name": "osa-in-cluster-thanos-querier"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
          createdAt: "2024-03-12T08:02:25Z"
          description: Deploys and Manages Observability Analytics Operator
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4-alpha
          repository: https://github.com/openshift/analytics-operator
          support: RedHat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnomalyData is the Schema for the anomalydata API
            displayName: Anomaly Data
            kind: AnomalyData
            name: anomalydata.observability-analytics.redhat.com
            version: v1alpha1
          - description: AnomalyEngine is the Schema for the anomalyengines API
            displayName: Anomaly Engine
            kind: AnomalyEngine
            name: anomalyengines.observability-analytics.redhat.com
            version: v1alpha1
        description: "### Introduction\nObservability Analytics for Red Hat OpenShift
          runs OpenShift analytics processes like anomaly detection.\n\n### API Backward
          Compatibility\n**NOTE:** Until the Operator supports **seamless upgrades**,
          a new version of the operator may introduce a change that is **NOT** backward-compatible
          with the previous version. Thus, to upgrade the operator, uninstall the
          already installed version completely (including CRDs) and install the new
          version.\n\n### Installtion Steps for the Anomaly Engine\n Create namespace
          if you would like to install this operator into separate namespace (You
          can utilize existing namespace as well). ```bash kubectl create ns <Namespace>
          ``` Install the Observability Analytics operator to specific `<Namespace>`,
          wait until you see the operator in the **Installed Operator** list. Click
          on the Operator name and go to Operator details.\n\nHere you should see
          a tab called **Anomaly Engine**. Click on it, and then click on the **Create
          AnomalyEngine** button, which will redirect you to a form where you need
          to provide the required inputs for the Anomaly Engine. Once successfully
          created, you should see a cronjob in the given namespace. This cronjob will
          identify anomalies and store them into **anomalydata**.\n\n### How to See
          Anomaly Data Captured by Anomaly Engine?\nWe are storing anomaly data into
          CR.\n ####Option 1:Through `kubectl`` command\nYou can check detected anomalies
          using the following command from your terminal. \n\n```bash\nkubectl get
          anomalydata -n <Namespace>\nkubectl describe anomalydata <NAME of the anomalydata>
          -n <Namespace> \n```\n #### Option 2: From Installed Operator UI Go to installed
          operator and select **Observability Analytics** operator, now go to **Anomaly
          Data** tab where you can able to see the detected anomaly, click on specific
          anomaly and you should be able to see details of the same. ### Documentation\nDocumentation
          around how to test Operator codebase as a developer can be found [here](https://github.com/openshift/analytics-operator#readme)\n\n###
          Contributing\nYou can contribute by:\n  * Raising [issues](https://github.com/openshift/analytics-operator/issues)\n
          \   related to observability-analytics-operator\n  * Fixing issues by opening
          [Pull Requests](https://github.com/openshift/analytics-operator/pulls)\n
          \ * Discussing observability-analytics-operator\n\n### License\nobservability-analytics-operator
          is licensed under the Apache 2.0 license.\n"
        displayName: Observability Analytics
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - observability-analytics-operator
        - analytics-operator
        - analytics
        - anomaly
        links:
        - name: Analytics Operator
          url: https://analytics-operator.domain
        maintainers:
        - email: rzalavad@redhat.com
          name: Raj
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
        version: 0.1.1
      entries:
      - name: analytics-operator.v0.1.1
        version: 0.1.1
      - name: analytics-operator.v0.1.0
        version: 0.1.0
      name: dev-preview
    defaultChannel: alpha
    packageName: analytics-operator
    provider:
      name: Red Hat Inc.
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-atlas-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-atlas-kubernetes.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupCompliancePolicy",
                "metadata": {
                  "name": "my-backup-compliance-policy"
                },
                "spec": {
                  "authorizedEmail": "john.doe@example.com",
                  "authorizedUserFirstName": "John",
                  "authorizedUserLastName": "Doe",
                  "copyProtectionEnabled": false,
                  "encryptionAtRestEnabled": false,
                  "onDemandPolicy": {
                    "retentionUnit": "weeks",
                    "retentionValue": 3
                  },
                  "overwriteBackupPolicies": false,
                  "pointInTimeEnabled": true,
                  "restoreWindowDays": 42,
                  "scheduledPolicyItems": [
                    {
                      "frequencyInterval": 2,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "hourly",
                      "retentionUnit": "days",
                      "retentionValue": 2
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "weekly",
                      "retentionUnit": "weeks",
                      "retentionValue": 4
                    },
                    {
                      "frequencyInterval": 15,
                      "frequencyType": "monthly",
                      "retentionUnit": "months",
                      "retentionValue": 12
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": false,
                  "copySettings": [
                    {
                      "cloudProvider": "AWS",
                      "frequencies": [
                        "WEEKLY"
                      ],
                      "regionName": "US_WEST_1",
                      "shouldCopyOplogs": false
                    }
                  ],
                  "policy": {
                    "name": "atlasbackuppolicy-sample"
                  },
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2,
                  "useOrgAndGroupNamesInExportPrefix": true
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasCustomRole",
                "metadata": {
                  "name": "shard-operator-role"
                },
                "spec": {
                  "projectRef": {
                    "name": "my-project"
                  },
                  "role": {
                    "actions": [
                      {
                        "name": "getShardMap",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "shardingState",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "connPoolStats",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "getLog",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      }
                    ],
                    "inheritedRoles": [
                      {
                        "database": "admin",
                        "name": "operator-role-1"
                      }
                    ],
                    "name": "my-role"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDataFederation",
                "metadata": {
                  "name": "my-data-federation",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "name": "my-data-federation",
                  "privateEndpoints": [
                    {
                      "endpointId": "vpce-03f9eeaa764e32454",
                      "provider": "AWS",
                      "type": "DATA_LAKE"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project",
                    "namespace": "mongodb-atlas-system"
                  },
                  "storage": {
                    "databases": [
                      {
                        "collections": [
                          {
                            "dataSources": [
                              {
                                "storeName": "http-test",
                                "urls": [
                                  "https://data.cityofnewyork.us/api/views/vfnx-vebw/rows.csv"
                                ]
                              }
                            ],
                            "name": "test-collection-1"
                          }
                        ],
                        "name": "test-db-1"
                      }
                    ],
                    "stores": [
                      {
                        "name": "http-test",
                        "provider": "http"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "the-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "theuser"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-deployment"
                },
                "spec": {
                  "backupRef": {
                    "name": "atlasbackupschedule-sample"
                  },
                  "deploymentSpec": {
                    "backupEnabled": true,
                    "clusterType": "REPLICASET",
                    "mongoDBMajorVersion": "7.0",
                    "name": "test-deployment",
                    "replicationSpecs": [
                      {
                        "regionConfigs": [
                          {
                            "electableSpecs": {
                              "instanceSize": "M10",
                              "nodeCount": 3
                            },
                            "priority": 7,
                            "providerName": "AWS",
                            "regionName": "US_EAST_1"
                          }
                        ],
                        "zoneName": "Zone 1"
                      }
                    ],
                    "tags": [
                      {
                        "key": "team",
                        "value": "atlas-operator"
                      }
                    ]
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasFederatedAuth",
                "metadata": {
                  "name": "atlas-default-federated-auth",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "connectionSecretRef": {
                    "name": "my-org-secret",
                    "namespace": "mongodb-atlas-system"
                  },
                  "domainAllowList": [
                    "my-org-domain.com"
                  ],
                  "domainRestrictionEnabled": true,
                  "enabled": true,
                  "postAuthRoleGrants": [
                    "ORG_MEMBER"
                  ],
                  "roleMappings": [
                    {
                      "externalGroupName": "org-admin",
                      "roleAssignments": [
                        {
                          "role": "ORG_OWNER"
                        }
                      ]
                    },
                    {
                      "externalGroupName": "dev-team",
                      "roleAssignments": [
                        {
                          "role": "ORG_GROUP_CREATOR"
                        },
                        {
                          "projectName": "dev-project",
                          "role": "GROUP_OWNER"
                        }
                      ]
                    }
                  ],
                  "ssoDebugEnabled": false
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasIPAccessList",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasipaccesslist-sample"
                },
                "spec": {
                  "entries": [
                    {
                      "cidrBlock": "192.168.1.0/24"
                    },
                    {
                      "deleteAfterDate": "2025-01-20T23:59:59+02:00",
                      "ipAddress": "10.0.0.1"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkContainer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasnetworkcontainer-sample"
                },
                "spec": {
                  "cidrBlock": "10.11.0.0/16",
                  "connectionSecret": {
                    "name": "atlas-connection-secret"
                  },
                  "externalProjectRef": {
                    "projectId": "66e2f2b621571b7e69a89b66"
                  },
                  "provider": "AWS",
                  "region": "US_EAST_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkPeering",
                "metadata": {
                  "name": "atlasnetworkpeering-sample"
                },
                "spec": {
                  "atlasCidrBlock": "10.12.0.0/16",
                  "awsConfiguration": {
                    "accepterRegionName": "US_EAST_1",
                    "awsAccountId": "23214235145",
                    "routeTableCidrBlock": "10.11.0.0/16",
                    "vpcId": "vpcid-7123782134"
                  },
                  "containerId": "623412394512350",
                  "containerRegion": "US_EAST_1",
                  "projectRef": {
                    "name": "atlas-project",
                    "namespace": "namespace"
                  },
                  "provider": "AWS"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasPrivateEndpoint",
                "metadata": {
                  "name": "atlasprivateendpoint-sample"
                },
                "spec": {
                  "awsConfiguration": {
                    "id": "vpce-f4k34w51d"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "provider": "AWS",
                  "region": "EU_CENTRAL_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    },
                    {
                      "cidrBlock": "203.0.113.0/24",
                      "comment": "CIDR block for Application Server B - D"
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasSearchIndexConfig",
                "metadata": {
                  "name": "atlassearchindexconfig-sample"
                },
                "spec": {
                  "analyzer": "lucene.standard",
                  "analyzers": [
                    {
                      "name": "test-analyzer",
                      "tokenFilters": [
                        {
                          "asciiFolding": {
                            "originalTokens": "omit"
                          },
                          "type": "asciiFolding"
                        }
                      ],
                      "tokenizer": {
                        "type": "whiteSpace",
                        "whitespace": {
                          "maxTokenLenght": 200
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamConnection",
                "metadata": {
                  "name": "my-streamconnection-kafka"
                },
                "spec": {
                  "kafkaConfig": {
                    "authentication": {
                      "credentials": {
                        "name": "kafka-userpass"
                      },
                      "mechanism": "SCRAM-512"
                    },
                    "bootstrapServers": "kafka.server1:9001,kafka.server2:9002,kafka.server3:9003",
                    "security": {
                      "certificate": {
                        "name": "kafka-cert"
                      },
                      "protocol": "SSL"
                    }
                  },
                  "name": "kafka-config",
                  "type": "Kafka"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamInstance",
                "metadata": {
                  "name": "my-streaminstance-sample"
                },
                "spec": {
                  "clusterConfig": {
                    "provider": "AWS",
                    "region": "VIRGINIA_USA",
                    "tier": "SP30"
                  },
                  "connectionRegistry": [
                    {
                      "name": "my-streamconnection-sample"
                    }
                  ],
                  "name": "my-instance",
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasTeam",
                "metadata": {
                  "name": "atlasteam-sample"
                },
                "spec": {
                  "name": "my-team",
                  "usernames": [
                    "username@mongodb.com"
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasThirdPartyIntegration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasintegration-sample"
                },
                "spec": {
                  "datadog": {
                    "apiKeySecretRef": {
                      "name": "secret-ref"
                    },
                    "region": "string",
                    "sendCollectionLatencyMetrics": "string",
                    "sendDatabaseMetrics": "string"
                  },
                  "projectRef": {
                    "name": "my-atlas-project"
                  },
                  "type": "DATADOG"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/mongodb/mongodb-atlas-kubernetes-operator:2.11.0
          createdAt: "2025-09-25T14:34:34Z"
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/mongodb/mongodb-atlas-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The AtlasBackupCompliancePolicy is a configuration that enforces
              specific backup and retention requirements
            displayName: Atlas Backup Compliance Policy
            kind: AtlasBackupCompliancePolicy
            name: atlasbackupcompliancepolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasCustomRole is the Schema for the AtlasCustomRole API
            displayName: Atlas Custom Role
            kind: AtlasCustomRole
            name: atlascustomroles.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDataFederation is the Schema for the Atlas Data Federation
              API
            displayName: Atlas Data Federation
            kind: AtlasDataFederation
            name: atlasdatafederations.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasdeployments API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasFederatedAuth is the Schema for the Atlasfederatedauth
              API
            displayName: Atlas Federated Auth
            kind: AtlasFederatedAuth
            name: atlasfederatedauths.atlas.mongodb.com
            version: v1
          - description: AtlasIPAccessList is the Schema for the atlasipaccesslists
              API.
            displayName: Atlas IPAccess List
            kind: AtlasIPAccessList
            name: atlasipaccesslists.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkContainer is the Schema for the AtlasNetworkContainer
              API
            displayName: Atlas Network Container
            kind: AtlasNetworkContainer
            name: atlasnetworkcontainers.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkPeering is the Schema for the AtlasNetworkPeering
              API
            displayName: Atlas Network Peering
            kind: AtlasNetworkPeering
            name: atlasnetworkpeerings.atlas.mongodb.com
            version: v1
          - displayName: Atlas Org Settings
            kind: AtlasOrgSettings
            name: atlasorgsettings.atlas.mongodb.com
            version: v1
          - description: "The AtlasPrivateEndpoint custom resource definition (CRD)
              defines a desired [Private Endpoint](https://www.mongodb.com/docs/atlas/security-private-endpoint/#std-label-private-endpoint-overview)
              configuration for an Atlas project. It allows a private connection between
              your cloud provider and Atlas that doesn't send information through
              a public network. \n You can use private endpoints to create a unidirectional
              connection to Atlas clusters from your virtual network."
            displayName: Atlas Private Endpoint
            kind: AtlasPrivateEndpoint
            name: atlasprivateendpoints.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
          - description: Atlas Search Index Config is the Schema for the Atlas Search
              API
            displayName: Atlas Search Index Config
            kind: AtlasSearchIndexConfig
            name: atlassearchindexconfigs.atlas.mongodb.com
            version: v1
          - description: AtlasStreamConnection is the Schema for the atlasstreamconnections
              API
            displayName: Atlas Stream Connection
            kind: AtlasStreamConnection
            name: atlasstreamconnections.atlas.mongodb.com
            version: v1
          - description: AtlasStreamInstance is the Schema for the atlasstreaminstances
              API
            displayName: Atlas Stream Instance
            kind: AtlasStreamInstance
            name: atlasstreaminstances.atlas.mongodb.com
            version: v1
          - description: AtlasTeam is the Schema for the Atlas Teams API
            displayName: Atlas Team
            kind: AtlasTeam
            name: atlasteams.atlas.mongodb.com
            version: v1
          - description: AtlasThirdPartyIntegration is the Schema for the atlas 3rd
              party integrations API.
            displayName: Atlas Third Party Integration
            kind: AtlasThirdPartyIntegration
            name: atlasthirdpartyintegrations.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB
          Atlas — the only multi-cloud document database service that gives you the versatility you need to build sophisticated
          and resilient applications that can adapt to changing customer demands and market trends.

          The full documentation for the Operator can be found [here](https://docs.atlas.mongodb.com/atlas-operator/)

          ## Getting Started

          ### Supported features

          * Create and configure a project, or connect to an existing one.
          * Deploy, manage, scale, and tear down clusters.
          * Support for serverless instances.
          * Create and edit database users.
          * Manage IP Access Lists, network peering and private endpoints.
          * Configure and control Atlas’s fully managed cloud backup.
          * Configure federated authentication for your Atlas organization
          * Integrate Atlas monitoring with Prometheus.

          ... and more.

          To view the list of custom resources and each CR's schema, visit our [reference](https://www.mongodb.com/docs/atlas/operator/stable/custom-resources/)
          documentation. See the [Quickstart](https://www.mongodb.com/docs/atlas/operator/stable/ak8so-quick-start/) to get
          started with Atlas Kubernetes Operator.
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator:2.11.0
        version: 2.11.0
      entries:
      - name: mongodb-atlas-kubernetes.v2.11.0
        version: 2.11.0
      - name: mongodb-atlas-kubernetes.v2.10.0
        version: 2.10.0
      - name: mongodb-atlas-kubernetes.v2.9.1
        version: 2.9.1
      - name: mongodb-atlas-kubernetes.v2.9.0
        version: 2.9.0
      - name: mongodb-atlas-kubernetes.v2.8.2
        version: 2.8.2
      - name: mongodb-atlas-kubernetes.v2.8.1
        version: 2.8.1
      - name: mongodb-atlas-kubernetes.v2.8.0
        version: 2.8.0
      - name: mongodb-atlas-kubernetes.v2.7.1
        version: 2.7.1
      - name: mongodb-atlas-kubernetes.v2.7.0
        version: 2.7.0
      - name: mongodb-atlas-kubernetes.v2.6.1
        version: 2.6.1
      - name: mongodb-atlas-kubernetes.v2.6.0
        version: 2.6.0
      - name: mongodb-atlas-kubernetes.v2.5.0
        version: 2.5.0
      - name: mongodb-atlas-kubernetes.v2.4.1
        version: 2.4.1
      - name: mongodb-atlas-kubernetes.v2.4.0
        version: 2.4.0
      - name: mongodb-atlas-kubernetes.v2.3.1
        version: 2.3.1
      - name: mongodb-atlas-kubernetes.v2.3.0
        version: 2.3.0
      - name: mongodb-atlas-kubernetes.v2.2.2
        version: 2.2.2
      - name: mongodb-atlas-kubernetes.v2.2.1
        version: 2.2.1
      - name: mongodb-atlas-kubernetes.v2.2.0
        version: 2.2.0
      - name: mongodb-atlas-kubernetes.v2.1.0
        version: 2.1.0
      - name: mongodb-atlas-kubernetes.v2.0.1
        version: 2.0.1
      - name: mongodb-atlas-kubernetes.v2.0.0
        version: 2.0.0
      - name: mongodb-atlas-kubernetes.v1.9.1
        version: 1.9.1
      - name: mongodb-atlas-kubernetes.v1.9.0
        version: 1.9.0
      - name: mongodb-atlas-kubernetes.v1.8.2
        version: 1.8.2
      - name: mongodb-atlas-kubernetes.v1.8.1
        version: 1.8.1
      - name: mongodb-atlas-kubernetes.v1.8.0
        version: 1.8.0
      - name: mongodb-atlas-kubernetes.v1.7.3
        version: 1.7.3
      - name: mongodb-atlas-kubernetes.v1.7.2
        version: 1.7.2
      - name: mongodb-atlas-kubernetes.v1.7.1
        version: 1.7.1
      - name: mongodb-atlas-kubernetes.v1.7.0
        version: 1.7.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-atlas-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OpenTelemetry Community
      provider-url: ""
    name: opentelemetry-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opentelemetry-operator.v0.136.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "Instrumentation",
                "metadata": {
                  "name": "instrumentation"
                },
                "spec": {
                  "exporter": {
                    "endpoint": "http://otel-collector-headless:4317"
                  }
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpAMPBridge",
                "metadata": {
                  "name": "opampbridge-sample"
                },
                "spec": {
                  "capabilities": {
                    "AcceptsOpAMPConnectionSettings": true,
                    "AcceptsOtherConnectionSettings": true,
                    "AcceptsRemoteConfig": true,
                    "AcceptsRestartCommand": true,
                    "ReportsEffectiveConfig": true,
                    "ReportsHealth": true,
                    "ReportsOwnLogs": true,
                    "ReportsOwnMetrics": true,
                    "ReportsOwnTraces": true,
                    "ReportsRemoteConfig": true,
                    "ReportsStatus": true
                  },
                  "componentsAllowed": {
                    "exporters": [
                      "debug"
                    ],
                    "processors": [
                      "memory_limiter"
                    ],
                    "receivers": [
                      "otlp"
                    ]
                  },
                  "endpoint": "ws://opamp-server:4320/v1/opamp"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": "receivers:\n  otlp:\n    protocols: \n      grpc: {}\n      http: {}\n\nexporters:\n  debug: {}\n\nservice:\n  pipelines:\n    traces:\n      receivers: [otlp]\n      exporters: [debug]\n"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1beta1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": {
                    "exporters": {
                      "debug": {}
                    },
                    "receivers": {
                      "otlp": {
                        "protocols": {
                          "grpc": {},
                          "http": {}
                        }
                      }
                    },
                    "service": {
                      "pipelines": {
                        "traces": {
                          "exporters": [
                            "debug"
                          ],
                          "receivers": [
                            "otlp"
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          certified: "false"
          containerImage: ghcr.io/open-telemetry/opentelemetry-operator/opentelemetry-operator
          createdAt: "2025-09-29T13:00:19Z"
          description: Provides the OpenTelemetry components, including the Collector
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/open-telemetry/opentelemetry-operator
          support: OpenTelemetry Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instrumentation is the spec for OpenTelemetry instrumentation.
            displayName: OpenTelemetry Instrumentation
            kind: Instrumentation
            name: instrumentations.opentelemetry.io
            version: v1alpha1
          - description: OpAMPBridge is the Schema for the opampbridges API.
            displayName: OpAMP Bridge
            kind: OpAMPBridge
            name: opampbridges.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1beta1
          - description: TargetAllocator is the Schema for the targetallocators API.
            displayName: Target Allocator
            kind: TargetAllocator
            name: targetallocators.opentelemetry.io
            version: v1alpha1
        description: |-
          OpenTelemetry is a collection of tools, APIs, and SDKs. You use it to instrument, generate, collect, and export telemetry data (metrics, logs, and traces) for analysis in order to understand your software's performance and behavior.

          ### Operator features

          * **Sidecar injection** - annotate your pods and let the operator inject a sidecar.
          * **Managed upgrades** - updating the operator will automatically update your OpenTelemetry collectors.
          * **Deployment modes** - your collector can be deployed as sidecar, daemon set, or regular deployment.
          * **Service port management** - the operator detects which ports need to be exposed based on the provided configuration.
        displayName: Community OpenTelemetry Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opentelemetry
        - tracing
        - logging
        - metrics
        - monitoring
        - troubleshooting
        links:
        - name: OpenTelemetry Operator
          url: https://github.com/open-telemetry/opentelemetry-operator
        maintainers:
        - email: jpkroehling@redhat.com
          name: Juraci Paixão Kröhling
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: OpenTelemetry Community
        relatedImages:
        - ghcr.io/open-telemetry/opentelemetry-operator/opentelemetry-operator:0.136.0
        - quay.io/brancz/kube-rbac-proxy:v0.13.1
        version: 0.136.0
      entries:
      - name: opentelemetry-operator.v0.136.0
        version: 0.136.0
      - name: opentelemetry-operator.v0.135.0
        version: 0.135.0
      - name: opentelemetry-operator.v0.134.0
        version: 0.134.0
      - name: opentelemetry-operator.v0.132.0
        version: 0.132.0
      - name: opentelemetry-operator.v0.131.0
        version: 0.131.0
      - name: opentelemetry-operator.v0.129.1
        version: 0.129.1
      - name: opentelemetry-operator.v0.127.0
        version: 0.127.0
      - name: opentelemetry-operator.v0.126.0
        version: 0.126.0
      - name: opentelemetry-operator.v0.125.0
        version: 0.125.0
      - name: opentelemetry-operator.v0.124.0
        version: 0.124.0
      - name: opentelemetry-operator.v0.123.0
        version: 0.123.0
      - name: opentelemetry-operator.v0.122.0
        version: 0.122.0
      - name: opentelemetry-operator.v0.121.0
        version: 0.121.0
      - name: opentelemetry-operator.v0.120.0
        version: 0.120.0
      - name: opentelemetry-operator.v0.119.0
        version: 0.119.0
      - name: opentelemetry-operator.v0.118.0
        version: 0.118.0
      - name: opentelemetry-operator.v0.117.0
        version: 0.117.0
      - name: opentelemetry-operator.v0.116.0
        version: 0.116.0
      - name: opentelemetry-operator.v0.115.0
        version: 0.115.0
      - name: opentelemetry-operator.v0.114.1
        version: 0.114.1
      - name: opentelemetry-operator.v0.112.0
        version: 0.112.0
      - name: opentelemetry-operator.v0.111.0
        version: 0.111.0
      - name: opentelemetry-operator.v0.110.0
        version: 0.110.0
      - name: opentelemetry-operator.v0.109.0
        version: 0.109.0
      - name: opentelemetry-operator.v0.108.0
        version: 0.108.0
      - name: opentelemetry-operator.v0.107.0
        version: 0.107.0
      - name: opentelemetry-operator.v0.106.0
        version: 0.106.0
      - name: opentelemetry-operator.v0.105.0
        version: 0.105.0
      name: alpha
    defaultChannel: alpha
    packageName: opentelemetry-operator
    provider:
      name: OpenTelemetry Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: StreamNative
      provider-url: https://streamnative.io
    name: sn-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc:latest",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway:0.10.0"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator:v0.10.0
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator:v0.10.0
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      - name: sn-operator.v0.9.0
        version: 0.9.0
      - name: sn-operator.v0.8.5
        version: 0.8.5
      - name: sn-operator.v0.7.10
        version: 0.7.10
      name: alpha
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc:latest",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway:0.10.0"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator:v0.10.0
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator:v0.10.0
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      - name: sn-operator.v0.9.0
        version: 0.9.0
      - name: sn-operator.v0.8.5
        version: 0.8.5
      - name: sn-operator.v0.7.10
        version: 0.7.10
      name: beta
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc:latest",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway:0.10.0"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator:v0.10.0
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator:v0.10.0
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      - name: sn-operator.v0.9.0
        version: 0.9.0
      - name: sn-operator.v0.8.5
        version: 0.8.5
      - name: sn-operator.v0.7.10
        version: 0.7.10
      name: stable
    defaultChannel: alpha
    packageName: sn-operator
    provider:
      name: StreamNative
      url: https://streamnative.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: operator-certification-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operator-certification-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "certification.redhat.com/v1alpha1",
                "kind": "OperatorPipeline",
                "metadata": {
                  "name": "operatorpipeline-sample"
                },
                "spec": {
                  "applyCIPipeline": true,
                  "applyHostedPipeline": false,
                  "applyReleasePipeline": false,
                  "gitHubSecretName": "github-api-token",
                  "kubeconfigSecretName": "kubeconfig",
                  "operatorPipelinesRelease": "main",
                  "pyxisSecretName": "pyxis-api-secret"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-06-23T16:02:00Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io.bundle.channel.default.v1: alpha
          operators.operatorframework.io.bundle.channels: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-openshift-ecosystem/operator-certification-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OperatorPipeline is the Schema for the operatorpipelines
              API
            displayName: Operator Pipeline
            kind: OperatorPipeline
            name: operatorpipelines.certification.redhat.com
            version: v1alpha1
        description: |-
          A Kubernetes operator to provision resources for the operator certification pipeline. This operator is installed in all namespaces which can support multi-tenant scenarios. **Note:** This operator should only be used by Red Hat partners attempting to certify their operator(s).


          **Requirements**


          The certification operator requires that you have the following tools installed, functional, and in your path.
          - [Install](https://docs.openshift.com/container-platform/4.8/cli_reference/openshift_cli/getting-started-cli.html#installing-openshift-cli) oc, the OpenShift CLI tool (tested with version 4.7.13)
          - [Install](https://tekton.dev/docs/cli/) tkn, the Tekton CLI tool (tested with version 0.19.1)
          - [Install](https://git-scm.com/downloads) git, the Git CLI tool (tested with 2.32.0)
          - The certification pipeline expects you to have the source files of your Operator bundle


          **Pre - Installation**


          The below steps exist for simplicity and to make the installation clearer. The operator watches all namespaces, so if secrets/configs/etc already exist in another namespace, feel free to use the existing namespace when following the operator installation steps.


          **Create a new namespace where the following secrets will be applied.**
          ```
          oc new-project oco
          ```


          **Add the kubeconfig secret which will be used to deploy the operator under test and run the certification checks.**
          * Open a terminal window
          * Set the `KUBECONFIG` environment variable
          ```
          export KUBECONFIG=/path/to/your/cluster/kubeconfig
          ```
          > *This kubeconfig will be used to deploy the Operator under test and run the certification checks.*
          ```
          oc create secret generic kubeconfig --from-file=kubeconfig=$KUBECONFIG
          ```


          **Configuring steps for submitting the results**
          - Add the github API token to the repo where the PR will be created
          ```
          oc create secret generic github-api-token --from-literal GITHUB_TOKEN=<github token>
          ```
          - Add RedHat Container API access key

            This API access key is specifically related to your unique partner account for Red Hat Connect portal. Instructions to obtain your API key can be found: [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/operator-cert-workflow.md#step-b---get-api-key)
          ```
          oc create secret generic pyxis-api-secret --from-literal pyxis_api_key=< API KEY >
          ```

          - Optional pipeline configurations can be found [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md#optional-configuration)


          **Execute the Pipeline (Development Iterations)**


          A pre-requisite to running a pipeline is that a `workspace-template.yaml` exists in the directory you want to execute the `tkn` commands from.

          To create a workspace-template.yaml
          ```
          cat <<EOF > workspace-template.yaml
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 5Gi
          EOF
          ```

          There are multiple ways to execute the Pipeline which can be found [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md#execute-the-pipeline-development-iterations)
        displayName: Operator Certification Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - certification
        - pipeline
        - openshift
        links:
        - name: Operator Certification CI Pipeline
          url: https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md
        - name: Red Hat OpenShift Pipelines
          url: https://github.com/openshift/tektoncd-operator
        maintainers:
        - email: certification-operator-devel@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/opdev/operator-certification-operator@sha256:7c84772f96650ff757c2552ba421355d54da41a5ed2155f72a14911b8a3f18ff
        version: 1.3.1
      entries:
      - name: operator-certification-operator.v1.3.1
        version: 1.3.1
      name: alpha
    defaultChannel: alpha
    packageName: operator-certification-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: LINBIT
      provider-url: https://linbit.com
    name: linstor-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: linstor-operator.v2.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorCluster",
                "metadata": {
                  "name": "linstorcluster"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorNodeConnection",
                "metadata": {
                  "name": "example-connection"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatellite",
                "metadata": {
                  "name": "example-satellite"
                },
                "spec": {
                  "clusterRef": {
                    "name": "linstorcluster"
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatelliteConfiguration",
                "metadata": {
                  "name": "all-satellites"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          com.redhat.openshift.versions: v4.7
          containerImage: drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
          createdAt: "2025-09-25T13:44:22Z"
          description: LINSTOR Operator deploys LINBIT SDS in the cluster.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: linbit-sds
          operators.openshift.io/valid-subscription: '[''LINBIT SDS Subscription'']'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/internal-objects: '["linstorsatellites.piraeus.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster-wide configuration for LINSTOR.
            displayName: Linstor Cluster
            kind: LinstorCluster
            name: linstorclusters.piraeus.io
            version: v1
          - description: Configuration for connections between specific nodes.
            displayName: Linstor Node Connection
            kind: LinstorNodeConnection
            name: linstornodeconnections.piraeus.io
            version: v1
          - description: Configuration for specific Satellites/Nodes in the LINSTOR
              Cluster.
            displayName: Linstor Satellite Configuration
            kind: LinstorSatelliteConfiguration
            name: linstorsatelliteconfigurations.piraeus.io
            version: v1
          - description: Resource to configure a single satellite, managed by the
              Operator itself.
            displayName: Linstor Satellite
            kind: LinstorSatellite
            name: linstorsatellites.piraeus.io
            version: v1
          required:
          - kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
        description: |
          The LINSTOR Operator manages [LINSTOR](https://github.com/LINBIT/linstor-server) clusters in Kubernetes.

          All components of the LINSTOR software stack can be managed by the operator:
          * DRBD
          * LINSTOR
          * LINSTOR CSI driver
          * LINSTOR HA Controller
        displayName: LINSTOR Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: User Guide
          url: https://linbit.com/drbd-user-guide/linstor-guide-1_0-en/
        - name: LINSTOR Operator
          url: https://charts.linstor.io/
        - name: Source Code
          url: https://github.com/LINBIT/linstor-operator-builder
        maintainers:
        - email: support@linbit.com
          name: LINBIT
        maturity: stable
        minKubeVersion: 1.20.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: DaemonSet
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: storage.k8s.io
          kind: CSIDriver
          version: v1
        provider:
          name: LINBIT
          url: https://linbit.com
        relatedImages:
        - drbd.io/drbd9-rhel8@sha256:7cddfa770de66cefcb42c2b33da57d0f1adfcc5f1cc12522977aa6993722bb2a
        - drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ac3963bfd4246da386b82d810a348e9d42a9e4b746acaf4dbcbe0766643cc2fa
        - drbd.io/drbd9-rhel10@sha256:da0a7015cf15724a7c73100d505891e62262a726141347df7f2d51ec79a0a1e1
        - drbd.io/linstor-k8s-ha-controller@sha256:ed0f0f86dad88feadf8c3c32d7d143eaa52b1130af3576315a562909b39d2afa
        - drbd.io/linstor-controller@sha256:841b3aa84cdf13aaaf1243bc5da8ac2205ca3fe0186fdfc7a04b69397f0020d6
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:2669cde3db2268d37bc671f494eaa06480478a90b8943054af14cbffc050debf
        - drbd.io/drbd9-rhel9@sha256:4eb553336ed64e127068a4566370b040166bf11301e21ac47ac5b92237ec9d2a
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:624f61b1ae9d32f5a074ee8c12839d975fba9efcd3bfa7981cfe38febf5c4495
        - drbd.io/linstor-csi@sha256:9c7ef452b0997d3373803201614c3b8075687e75c397359738a9e4ef5530cab3
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:700be620923fc99ed03f9f1c4a911e4084beb30685361270c49165b27ad1d374
        - drbd.io/linstor-affinity-controller@sha256:78e5b299ef1702114528678387b5e74821d7658bc02f05f0bd7edd9a176fba39
        - drbd.io/drbd-shutdown-guard@sha256:7ba049ecb205a0086e425d2ff37285b00449549d89e179b4b943ace6e59a9156
        - drbd.io/linstor-satellite@sha256:ce0a0b05494c6e6d5f16121e5762bf58e7b78f40c00b84780e2abba2a690af1b
        - drbd.io/drbd-reactor@sha256:77c51aeeefdcd788cb1583738f93778032fcc03c36fb289b692d3cf195c8b184
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:5a650dc21a6a038de6267f117d01adfd904ff3f5c94f428f6a13735f18bf94d6
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:cd58f1789027e32eae08f00dac7b2275496c41020fb7cdebc10e9986c625d4f4
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:f8cde5b5c289f7b3d233773f4e79990ba392fe6e14e0177de80714e7d3e98862
        - drbd.io/drbd9-rhel7@sha256:0c96b216e55a128a9944276af9080aced8bd72fdf1fd424f024ed9ace2c1feb5
        - drbd.io/ktls-utils@sha256:3351a6e0fc048fe1544499095cdb8fdd221e832cecfa8a9a01f046c271991138
        version: 2.9.1
      entries:
      - name: linstor-operator.v2.9.1
        version: 2.9.1
      name: dev
    - currentCSV: linstor-operator.v2.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorCluster",
                "metadata": {
                  "name": "linstorcluster"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorNodeConnection",
                "metadata": {
                  "name": "example-connection"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatellite",
                "metadata": {
                  "name": "example-satellite"
                },
                "spec": {
                  "clusterRef": {
                    "name": "linstorcluster"
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatelliteConfiguration",
                "metadata": {
                  "name": "all-satellites"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          com.redhat.openshift.versions: v4.7
          containerImage: drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
          createdAt: "2025-09-25T13:44:22Z"
          description: LINSTOR Operator deploys LINBIT SDS in the cluster.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: linbit-sds
          operators.openshift.io/valid-subscription: '[''LINBIT SDS Subscription'']'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/internal-objects: '["linstorsatellites.piraeus.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster-wide configuration for LINSTOR.
            displayName: Linstor Cluster
            kind: LinstorCluster
            name: linstorclusters.piraeus.io
            version: v1
          - description: Configuration for connections between specific nodes.
            displayName: Linstor Node Connection
            kind: LinstorNodeConnection
            name: linstornodeconnections.piraeus.io
            version: v1
          - description: Configuration for specific Satellites/Nodes in the LINSTOR
              Cluster.
            displayName: Linstor Satellite Configuration
            kind: LinstorSatelliteConfiguration
            name: linstorsatelliteconfigurations.piraeus.io
            version: v1
          - description: Resource to configure a single satellite, managed by the
              Operator itself.
            displayName: Linstor Satellite
            kind: LinstorSatellite
            name: linstorsatellites.piraeus.io
            version: v1
          required:
          - kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
        description: |
          The LINSTOR Operator manages [LINSTOR](https://github.com/LINBIT/linstor-server) clusters in Kubernetes.

          All components of the LINSTOR software stack can be managed by the operator:
          * DRBD
          * LINSTOR
          * LINSTOR CSI driver
          * LINSTOR HA Controller
        displayName: LINSTOR Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: User Guide
          url: https://linbit.com/drbd-user-guide/linstor-guide-1_0-en/
        - name: LINSTOR Operator
          url: https://charts.linstor.io/
        - name: Source Code
          url: https://github.com/LINBIT/linstor-operator-builder
        maintainers:
        - email: support@linbit.com
          name: LINBIT
        maturity: stable
        minKubeVersion: 1.20.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: DaemonSet
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: storage.k8s.io
          kind: CSIDriver
          version: v1
        provider:
          name: LINBIT
          url: https://linbit.com
        relatedImages:
        - drbd.io/drbd-shutdown-guard@sha256:7ba049ecb205a0086e425d2ff37285b00449549d89e179b4b943ace6e59a9156
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ac3963bfd4246da386b82d810a348e9d42a9e4b746acaf4dbcbe0766643cc2fa
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:624f61b1ae9d32f5a074ee8c12839d975fba9efcd3bfa7981cfe38febf5c4495
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:cd58f1789027e32eae08f00dac7b2275496c41020fb7cdebc10e9986c625d4f4
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:f8cde5b5c289f7b3d233773f4e79990ba392fe6e14e0177de80714e7d3e98862
        - drbd.io/drbd9-rhel10@sha256:da0a7015cf15724a7c73100d505891e62262a726141347df7f2d51ec79a0a1e1
        - drbd.io/linstor-affinity-controller@sha256:78e5b299ef1702114528678387b5e74821d7658bc02f05f0bd7edd9a176fba39
        - drbd.io/linstor-csi@sha256:9c7ef452b0997d3373803201614c3b8075687e75c397359738a9e4ef5530cab3
        - drbd.io/drbd9-rhel9@sha256:4eb553336ed64e127068a4566370b040166bf11301e21ac47ac5b92237ec9d2a
        - drbd.io/drbd9-rhel7@sha256:0c96b216e55a128a9944276af9080aced8bd72fdf1fd424f024ed9ace2c1feb5
        - drbd.io/linstor-satellite@sha256:ce0a0b05494c6e6d5f16121e5762bf58e7b78f40c00b84780e2abba2a690af1b
        - drbd.io/drbd-reactor@sha256:77c51aeeefdcd788cb1583738f93778032fcc03c36fb289b692d3cf195c8b184
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:2669cde3db2268d37bc671f494eaa06480478a90b8943054af14cbffc050debf
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:5a650dc21a6a038de6267f117d01adfd904ff3f5c94f428f6a13735f18bf94d6
        - drbd.io/ktls-utils@sha256:3351a6e0fc048fe1544499095cdb8fdd221e832cecfa8a9a01f046c271991138
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:700be620923fc99ed03f9f1c4a911e4084beb30685361270c49165b27ad1d374
        - drbd.io/drbd9-rhel8@sha256:7cddfa770de66cefcb42c2b33da57d0f1adfcc5f1cc12522977aa6993722bb2a
        - drbd.io/linstor-controller@sha256:841b3aa84cdf13aaaf1243bc5da8ac2205ca3fe0186fdfc7a04b69397f0020d6
        - drbd.io/linstor-k8s-ha-controller@sha256:ed0f0f86dad88feadf8c3c32d7d143eaa52b1130af3576315a562909b39d2afa
        - drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
        version: 2.9.1
      entries:
      - name: linstor-operator.v2.9.1
        version: 2.9.1
      name: stable
    defaultChannel: dev
    packageName: linstor-operator
    provider:
      name: LINBIT
      url: https://linbit.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: https://opdev.github.io/
    name: loki-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.example.com/v1alpha1",
                "kind": "Loki",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/install-disable-crds": "true"
                  },
                  "name": "loki-sample"
                },
                "spec": {
                  "global": {
                    "clusterDomain": "cluster.local",
                    "dnsNamespace": "openshift-dns",
                    "dnsService": "dns-default"
                  },
                  "loki": {
                    "auth_enabled": false,
                    "commonConfig": {
                      "replication_factor": 1
                    },
                    "storage": {
                      "type": "filesystem"
                    }
                  },
                  "monitoring": {
                    "dashboards": {
                      "enabled": false
                    },
                    "lokiCanary": {
                      "enabled": false
                    },
                    "rules": {
                      "alerting": false,
                      "enabled": false
                    },
                    "selfMonitoring": {
                      "enabled": false,
                      "grafanaAgent": {
                        "installOperator": false
                      }
                    },
                    "serviceMonitor": {
                      "enabled": false,
                      "metricsInstance": {
                        "enabled": false
                      }
                    }
                  },
                  "rbac": {
                    "sccEnabled": true
                  },
                  "singleBinary": {
                    "replicas": 1
                  },
                  "test": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/loki-helm-operator-controller:0.0.1
          createdAt: "2023-07-14T15:06:17Z"
          description: An experimental operator that installs loki from its official
            helmchart.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys the loki operator helm chart
            displayName: Loki
            kind: Loki
            name: lokis.charts.example.com
            version: v1alpha1
        description: Loki operator based on the helm chart with single binary mode
          deployment optionality
        displayName: Loki Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loki
        - single-binary
        - demo
        - helm
        links:
        - name: Loki Helm Operator
          url: https://github.com/skattoju/loki-helm-operator
        maintainers:
        - email: skattoju@redhat.com
          name: Sid Kattoju
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: https://opdev.github.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/opdev/loki-helm-operator-controller:0.0.1
        version: 0.0.1
      entries:
      - name: loki-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: loki-helm-operator
    provider:
      name: opdev
      url: https://opdev.github.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hdfs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hdfs-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:e8b4ad2c3b127537239b003ae19a3f58b541d599fe3d5a74a6718612fe645d50
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:e8b4ad2c3b127537239b003ae19a3f58b541d599fe3d5a74a6718612fe645d50
        version: 24.11.1
      entries:
      - name: hdfs-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: hdfs-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:00c79036eed6dd6749736b964787acb638f1a6e0156c0bba5651b93738534377
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:00c79036eed6dd6749736b964787acb638f1a6e0156c0bba5651b93738534377
        version: 25.3.0
      entries:
      - name: hdfs-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: hdfs-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
        version: 25.7.0
      entries:
      - name: hdfs-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hdfs-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
        version: 25.7.0
      entries:
      - name: hdfs-operator.v25.7.0
        version: 25.7.0
      - name: hdfs-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hdfs-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: seldon
      provider-url: ""
    name: seldon-deploy-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: seldon-deploy-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1alpha1",
                "kind": "SeldonDeploy",
                "metadata": {
                  "name": "seldondeploy-sample"
                },
                "spec": {
                  "affinity": {},
                  "alibidetect": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306"
                  },
                  "batchjobs": {
                    "mc": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48"
                    },
                    "processor": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d"
                    },
                    "pvc": {
                      "defaultSize": "1Gi"
                    },
                    "serviceAccount": "workflow"
                  },
                  "defaultUserID": "",
                  "docker": {
                    "user": "unknown"
                  },
                  "elasticsearch": {
                    "basicAuth": true,
                    "secret": {
                      "name": "elastic-credentials",
                      "passwordKey": "password",
                      "userKey": "username"
                    },
                    "url": "https://elasticsearch-seldon-es-http.seldon-logs:9200"
                  },
                  "enableAppAnalytics": false,
                  "enableAppAuth": false,
                  "env": {
                    "USERID_CLAIM_KEY": "name"
                  },
                  "external": {
                    "protocol": "http"
                  },
                  "fullnameOverride": "",
                  "gitops": {
                    "argocd": {
                      "enabled": false,
                      "namespace": "argocd"
                    },
                    "fileFormat": "json",
                    "git": {
                      "email": "",
                      "secret": "git-creds",
                      "skipVerifyGit": true,
                      "token": "",
                      "user": "",
                      "webhook": {
                        "service": {
                          "create": true,
                          "loadBalancerSourceRanges": {}
                        }
                      }
                    }
                  },
                  "image": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e",
                    "pullPolicy": "Always"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": []
                      }
                    ],
                    "tls": []
                  },
                  "ingressGateway": {
                    "ingressNamespace": "istio-system",
                    "kfServingIngressService": "istio-ingressgateway",
                    "seldonIngressService": "istio-ingressgateway"
                  },
                  "kfserving": {
                    "curlForm": "MODEL_NAME={{ .ModelName }}\u003cbr\u003e\nCLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\nSERVICE_HOSTNAME=$(kubectl get inferenceservice {{ .ModelName }} -o jsonpath='{.status.url}' | cut -d \"/\" -f 3)\u003cbr\u003e\ncurl -v -H \"Host: ${SERVICE_HOSTNAME}\" {{ .KfServingProtocol }}://$CLUSTER_IP/v1/models/$MODEL_NAME:predict -d '{{ .Payload }}'\n",
                    "enabled": false,
                    "explainForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:explain",
                    "protocol": "http",
                    "requestForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:predict"
                  },
                  "loadtest": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c"
                  },
                  "metadata": {
                    "pg": {
                      "enabled": false,
                      "secret": "metadata-postgres"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "openshiftMarketplace": {
                    "cleanupClusterServiceVersions": false,
                    "kubectlCleanupImage": "registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382",
                    "prometheus": {
                      "monitorSpecs": {
                        "create": true
                      }
                    },
                    "seldonCore": {
                      "istioGateway": {
                        "create": true,
                        "name": "seldon-gateway",
                        "namespace": "istio-system"
                      },
                      "subscription": {
                        "apiVersion": "operators.coreos.com/v1alpha1",
                        "channel": "stable",
                        "create": true,
                        "istioEnabled": true,
                        "metricsPath": "/metrics",
                        "requestLoggerEndpoint": "http://broker-ingress.knative-eventing.svc.cluster.local/seldon-logs/default"
                      }
                    }
                  },
                  "prometheus": {
                    "knative": {
                      "url": "http://prometheus-system-np.knative-monitoring.svc.cluster.local:8080/api/v1/"
                    },
                    "seldon": {
                      "jwtSecretKey": "jwt-seldon.txt",
                      "jwtSecretName": "jwt-seldon",
                      "namespaceMetricName": "namespace",
                      "resourceMetricsUrl": "https://prometheus-k8s.openshift-monitoring:9091/api/v1/",
                      "serviceMetricName": "exported_service",
                      "url": "https://thanos-querier.openshift-monitoring.svc:9091/api/v1/"
                    }
                  },
                  "rbac": {
                    "clusterWide": true,
                    "create": true,
                    "readNamespaces": true
                  },
                  "replicaCount": 1,
                  "requestLogger": {
                    "create": true,
                    "elasticsearch": {
                      "host": "elasticsearch-seldon-es-http.seldon-logs",
                      "port": "9200",
                      "protocol": "https"
                    },
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398",
                    "imagePullPolicy": "IfNotPresent",
                    "namespace": {
                      "create": false,
                      "name": "seldon-logs"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "600m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "200Mi"
                      }
                    },
                    "trigger": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "broker": "default"
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "800m",
                      "memory": "800Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "seldon": {
                    "curlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions -d '{{ .Payload }}'\n",
                    "enabled": true,
                    "explainForm": "http://{{ .ModelName }}-{{ .Predictor }}-explainer.{{ .Namespace }}:9000/v1/models/{{ .ModelName }}:explain",
                    "kfservingV2CurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .GraphModelName }}/infer -d '{{ .Payload }}'\n",
                    "protocol": "http",
                    "seldonRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions",
                    "tensorFlowCurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict -d '{{ .Payload }}'\n",
                    "tensorflowRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict",
                    "v2RequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .ModelName }}/infer"
                  },
                  "service": {
                    "port": 80,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "create": true
                  },
                  "serviceAccountName": "seldon-deploy",
                  "skipVerifyHttpCalls": true,
                  "tolerations": [],
                  "virtualService": {
                    "create": true,
                    "gateways": [
                      "istio-system/seldon-gateway"
                    ],
                    "prefix": "/seldon-deploy/"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
          createdAt: "2020-12-11 15:00:00"
          description: The Seldon deploy operator allows the creation of Seldon Deploy
            applications to provide management, monitoring and updating of machine
            learning models from a range of technologies. It includes the ability
            to run explanations as well as outlier, drift and adversarial detection
            against running models. There is a full metrics and logging integration
            via Prometheus and ELK stacks as well as full GitOps operations to allow
            full auditing and recovery of updated models.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SeldonIO/seldon-deploy-operator
          support: Seldon Dev
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: SeldonDeploy
            name: seldondeploys.machinelearning.seldon.io
            version: v1alpha1
        description: The Seldon deploy operator allows the creation of Seldon Deploy
          applications to provide management, monitoring and updating of machine learning
          models from a range of technologies. It includes the ability to run explanations
          as well as outlier, drift and adversarial detection against running models.
          There is a full metrics and logging integration via Prometheus and ELK stacks
          as well as full GitOps operations to allow full auditing and recovery of
          updated models.
        displayName: Seldon Deploy Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - machine learning
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://deploy.seldon.io
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        provider:
          name: seldon
        relatedImages:
        - registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398
        - registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
        - registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c
        version: 1.2.1
      entries:
      - name: seldon-deploy-operator.v1.2.1
        version: 1.2.1
      - name: seldon-deploy-operator.v1.2.0
        version: 1.2.0
      - name: seldon-deploy-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: seldon-deploy-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1alpha1",
                "kind": "SeldonDeploy",
                "metadata": {
                  "name": "seldondeploy-sample"
                },
                "spec": {
                  "affinity": {},
                  "alibidetect": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306"
                  },
                  "batchjobs": {
                    "mc": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48"
                    },
                    "processor": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d"
                    },
                    "pvc": {
                      "defaultSize": "1Gi"
                    },
                    "serviceAccount": "workflow"
                  },
                  "defaultUserID": "",
                  "docker": {
                    "user": "unknown"
                  },
                  "elasticsearch": {
                    "basicAuth": true,
                    "secret": {
                      "name": "elastic-credentials",
                      "passwordKey": "password",
                      "userKey": "username"
                    },
                    "url": "https://elasticsearch-seldon-es-http.seldon-logs:9200"
                  },
                  "enableAppAnalytics": false,
                  "enableAppAuth": false,
                  "env": {
                    "USERID_CLAIM_KEY": "name"
                  },
                  "external": {
                    "protocol": "http"
                  },
                  "fullnameOverride": "",
                  "gitops": {
                    "argocd": {
                      "enabled": false,
                      "namespace": "argocd"
                    },
                    "fileFormat": "json",
                    "git": {
                      "email": "",
                      "secret": "git-creds",
                      "skipVerifyGit": true,
                      "token": "",
                      "user": "",
                      "webhook": {
                        "service": {
                          "create": true,
                          "loadBalancerSourceRanges": {}
                        }
                      }
                    }
                  },
                  "image": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e",
                    "pullPolicy": "Always"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": []
                      }
                    ],
                    "tls": []
                  },
                  "ingressGateway": {
                    "ingressNamespace": "istio-system",
                    "kfServingIngressService": "istio-ingressgateway",
                    "seldonIngressService": "istio-ingressgateway"
                  },
                  "kfserving": {
                    "curlForm": "MODEL_NAME={{ .ModelName }}\u003cbr\u003e\nCLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\nSERVICE_HOSTNAME=$(kubectl get inferenceservice {{ .ModelName }} -o jsonpath='{.status.url}' | cut -d \"/\" -f 3)\u003cbr\u003e\ncurl -v -H \"Host: ${SERVICE_HOSTNAME}\" {{ .KfServingProtocol }}://$CLUSTER_IP/v1/models/$MODEL_NAME:predict -d '{{ .Payload }}'\n",
                    "enabled": false,
                    "explainForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:explain",
                    "protocol": "http",
                    "requestForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:predict"
                  },
                  "loadtest": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c"
                  },
                  "metadata": {
                    "pg": {
                      "enabled": false,
                      "secret": "metadata-postgres"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "openshiftMarketplace": {
                    "cleanupClusterServiceVersions": false,
                    "kubectlCleanupImage": "registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382",
                    "prometheus": {
                      "monitorSpecs": {
                        "create": true
                      }
                    },
                    "seldonCore": {
                      "istioGateway": {
                        "create": true,
                        "name": "seldon-gateway",
                        "namespace": "istio-system"
                      },
                      "subscription": {
                        "apiVersion": "operators.coreos.com/v1alpha1",
                        "channel": "stable",
                        "create": true,
                        "istioEnabled": true,
                        "metricsPath": "/metrics",
                        "requestLoggerEndpoint": "http://broker-ingress.knative-eventing.svc.cluster.local/seldon-logs/default"
                      }
                    }
                  },
                  "prometheus": {
                    "knative": {
                      "url": "http://prometheus-system-np.knative-monitoring.svc.cluster.local:8080/api/v1/"
                    },
                    "seldon": {
                      "jwtSecretKey": "jwt-seldon.txt",
                      "jwtSecretName": "jwt-seldon",
                      "namespaceMetricName": "namespace",
                      "resourceMetricsUrl": "https://prometheus-k8s.openshift-monitoring:9091/api/v1/",
                      "serviceMetricName": "exported_service",
                      "url": "https://thanos-querier.openshift-monitoring.svc:9091/api/v1/"
                    }
                  },
                  "rbac": {
                    "clusterWide": true,
                    "create": true,
                    "readNamespaces": true
                  },
                  "replicaCount": 1,
                  "requestLogger": {
                    "create": true,
                    "elasticsearch": {
                      "host": "elasticsearch-seldon-es-http.seldon-logs",
                      "port": "9200",
                      "protocol": "https"
                    },
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398",
                    "imagePullPolicy": "IfNotPresent",
                    "namespace": {
                      "create": false,
                      "name": "seldon-logs"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "600m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "200Mi"
                      }
                    },
                    "trigger": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "broker": "default"
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "800m",
                      "memory": "800Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "seldon": {
                    "curlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions -d '{{ .Payload }}'\n",
                    "enabled": true,
                    "explainForm": "http://{{ .ModelName }}-{{ .Predictor }}-explainer.{{ .Namespace }}:9000/v1/models/{{ .ModelName }}:explain",
                    "kfservingV2CurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .GraphModelName }}/infer -d '{{ .Payload }}'\n",
                    "protocol": "http",
                    "seldonRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions",
                    "tensorFlowCurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict -d '{{ .Payload }}'\n",
                    "tensorflowRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict",
                    "v2RequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .ModelName }}/infer"
                  },
                  "service": {
                    "port": 80,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "create": true
                  },
                  "serviceAccountName": "seldon-deploy",
                  "skipVerifyHttpCalls": true,
                  "tolerations": [],
                  "virtualService": {
                    "create": true,
                    "gateways": [
                      "istio-system/seldon-gateway"
                    ],
                    "prefix": "/seldon-deploy/"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
          createdAt: "2020-12-11 15:00:00"
          description: The Seldon deploy operator allows the creation of Seldon Deploy
            applications to provide management, monitoring and updating of machine
            learning models from a range of technologies. It includes the ability
            to run explanations as well as outlier, drift and adversarial detection
            against running models. There is a full metrics and logging integration
            via Prometheus and ELK stacks as well as full GitOps operations to allow
            full auditing and recovery of updated models.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SeldonIO/seldon-deploy-operator
          support: Seldon Dev
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: SeldonDeploy
            name: seldondeploys.machinelearning.seldon.io
            version: v1alpha1
        description: The Seldon deploy operator allows the creation of Seldon Deploy
          applications to provide management, monitoring and updating of machine learning
          models from a range of technologies. It includes the ability to run explanations
          as well as outlier, drift and adversarial detection against running models.
          There is a full metrics and logging integration via Prometheus and ELK stacks
          as well as full GitOps operations to allow full auditing and recovery of
          updated models.
        displayName: Seldon Deploy Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - machine learning
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://deploy.seldon.io
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        provider:
          name: seldon
        relatedImages:
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d
        - registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48
        - registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c
        - registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e
        version: 1.2.1
      entries:
      - name: seldon-deploy-operator.v1.2.1
        version: 1.2.1
      - name: seldon-deploy-operator.v1.2.0
        version: 1.2.0
      - name: seldon-deploy-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: seldon-deploy-operator-rhmp
    provider:
      name: seldon
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: rhods-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhods-operator.2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
          createdAt: "2025-08-12T18:04:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "llamastackoperator": {
                      "managementState": "Removed"
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (including PyTorch, TensorFlow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack – unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.23
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fbe10282ada2fa6bf07ae97b497f46ebe735de130b4ce0182bf34301429999cb
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:2d2e7803b2b1e15017ca04961f9dcd25e89c47b07b101ca54cb9e793e6027202
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:e50adae13e9d773e7be8bf81b7525dac3f5e3864c9cdf4c287342bd568ee1d6b
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:bb7c352ccc10a8c60ef8d573368d0ada59e73cd131653d11bcebdf9f564acc10
        - registry.redhat.io/rhel9/mariadb-105@sha256:050dd5a7a32395b73b8680570e967e55050b152727412fdd73a25d8816e62d53
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:ccfc84028a95f1cc6a16e06ff6e9ed2ace74f74e5e9963a4fb33544c2ce005c2
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:2886e40f94fde72eb5959ad564bbdd466657208c09bac731b3860b0b54f104c6
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:b038d20def504c140e4365f2eb117a979534629332bec75a2ba750ebafdefe3d
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5fb7e6c3ee1e3b170cd146ea9cd172fa953326b08163897605ca74a3644f2819
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:795ab1842ef675e3d8b47deb8a7485a69064428e823dfeb930a6a7e7017cde0f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:80d284cac2ef6638fe028102b71a70f26041576c11a49414e77430954a3c1b99
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f23f8051886d0f0240937507471c818f440bca8703b8209a836d5dbe3ea2f7c8
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:16ff53f4f1a1e1e66826c8355c654038da83df545427a931485ec7a2d3764b16
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:23c098f7e791441cfcfa6bf37de51763511bd8bf4d543747b7d37ed591d68886
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:edcce7cddfa4c250108532e56541ddb95099612c180b2297e5934bc89d2bb46e
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:ff740fc3615edaa90dae976da1aa6455d64811e239eb66b422836a3aae791b41
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:240692b86bc8a4fe0b545e6e4cd7cf4ff0ca990406bc55f0446d565fd67e37f6
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:a4649e8ad4970bc3d5e967811c9694fa82d78ee9f8f025b8b2fbd37237eb14fc
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:68e74bdf518fc624cc25a59a18b2a87d0aa2a9b8aad30750ace7115d297a21ca
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:ca21e218e26c46e3c63d926241846f8f307fd4a586cc4b04147da49af6018ef5
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:857560240bb45bfc2db6d7743abf11a47f929a0d3035940256a6527afd313412
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:84c3213e11299ec032aa0e5a4dfa9b3055c4246b04b029e888d8d9f1caf5c30c
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:4847ac1be55284815670c2048c1edfc56c7227494280d3c6094aa23f1ed1a9d2
        - registry.redhat.io/ubi9/toolbox@sha256:4213516b6fe7957a01072720efd9ecbc7d4443238e151b5fe0263bd0c36792b1
        - registry.redhat.io/rhoai/odh-llama-stack-core-rhel9@sha256:43b60b1ee6f66fec38fe2ffbbe08dca8541ef162332e4bd8e422ecd24ee02646
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:467b04f4ee4118bbe0934d1ab8b80f72df8703dcdac61c2b53fdb5bd1b404c14
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:a1497e64349634b2567d28f685052de118f64f6b32e7d3ea469212e2004d7eb3
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:e0c9ac73891781ef421e0a020cfd5c962f7632d38020897183cf048b1819326a
        - registry.redhat.io/rhoai/odh-llama-stack-k8s-operator-rhel9@sha256:626b9ef09ee82c8d425ddb7ed3f2ffbb3d1019f541e31604db891787c881d5c9
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:3243caa4c8aac75069d19225dcee361bf9c012782328a6266b7016b3026f6d56
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:756b7bdaa7f6e7c4a18488e690188dd91153a33913d225bab001b0085382e514
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:da8debeb2b84a42aba0b98785ccc20e02ed520e125428fe012df3ca473cbb755
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:208005da389b7d4176ad8bffc703312011a1299954df151e2abbe8f3dec01d19
        version: 2.23.0
      entries:
      - name: rhods-operator.2.23.0
        version: 2.23.0
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.20.0
        version: 2.20.0
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.18.0
        version: 2.18.0
      - name: rhods-operator.2.17.0
        version: 2.17.0
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.15.0
        version: 2.15.0
      - name: rhods-operator.2.14.0
        version: 2.14.0
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: alpha
    - currentCSV: rhods-operator.2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhods/odh-rhel8-operator@sha256:a4d7f020509b9720063e87e1c2f19af04e4c9b4986b199be4dc617831b971434
          createdAt: "2023-11-21T08:37:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            Data Science
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:08c90a8f133b4cdb1477a742bc44be4d8fc4e88153b8bc2a2a148cbdf19f4e7d
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:6a56010de933f172b195a1a575855d37b70a4968be8edb35157f6ca193969ad2
          olm.skipRange: '>=1.33.0 <2.4.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "rhods",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "rhods",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Data Science"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:eb96f4b753606528750b3c28f7f9256582f81fa030450b05afdef684634903be
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift Data Science
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift Data Science (RHODS) is a complete platform for the entire lifecycle of your AI/ML projects. It is the flagship product of OpenShift AI.

          When using RHODS, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. RHODS can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * RHODS dashboard
          * Curated Notebook Images (incl CUDA, PyTorch, Tensorflow)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh (and Kserve) with a provided OpenVino Model Serving Runtime
          * Ability to use other runtimes for serving
          * Model Monitoring
        displayName: Red Hat OpenShift Data Science
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift Data Science
        - RHODS
        - rhods
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift Data Science
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-data-science
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhods/odh-ml-pipelines-api-server-rhel8@sha256:be57f237df156474a709667fa4fb4aa142118e3459963fed56648c1cb2e38cd6
        - registry.redhat.io/rhods/odh-kuberay-operator-controller-rhel8@sha256:bfc805aa1689a4bd6f7e617acf9eb1c2c307624dd49956e9e7442afac328151a
        - registry.redhat.io/rhods/odh-data-science-pipelines-operator-controller-rhel8@sha256:250db47cb873207cfb75c7e8a2f0de08b75b056a72db73362d2d79c08cb7b185
        - registry.redhat.io/rhods/odh-trustyai-service-operator-rhel8@sha256:68471555d0b2bb8150ebfae8d4f3d1dbdff9c5b961ff2305ef23ab7514cbbea1
        - registry.redhat.io/rhods/odh-codeflare-operator-rhel8@sha256:2841fd3051a50a1b0d9e3caafa20a8084b5c472d49393b372828960dbf156674
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhods/odh-model-controller-rhel8@sha256:fde989817cd1a6ec8eb408a7f6c73e28ba043ab2d88035586b1b7e55fb3405cb
        - registry.redhat.io/rhods/odh-modelmesh-runtime-adapter-rhel8@sha256:08478a1647d0413783bcbcd966a73b444b580464de46efedb39504420f22053e
        - registry.redhat.io/rhel8/mariadb-103@sha256:08c90a8f133b4cdb1477a742bc44be4d8fc4e88153b8bc2a2a148cbdf19f4e7d
        - registry.redhat.io/rhods/odh-trustyai-service-rhel8@sha256:fe17d6e07ebaa088f274dd8b5e83892dacc11cb336065b06582a3c055095dba5
        - registry.redhat.io/rhods/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:c5758102609821a6fdabec0277e03c4c728f82176093d1ade0570aba5f3cd8f9
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhods/odh-ml-pipelines-persistenceagent-rhel8@sha256:857a8eb444375505ccb34a11cce69c1794ceabbfb00df06933d7f12cd187319b
        - registry.redhat.io/ubi8/ubi-micro@sha256:6a56010de933f172b195a1a575855d37b70a4968be8edb35157f6ca193969ad2
        - registry.redhat.io/rhods/odh-rhel8-operator@sha256:a4d7f020509b9720063e87e1c2f19af04e4c9b4986b199be4dc617831b971434
        - registry.redhat.io/rhods/odh-mm-rest-proxy-rhel8@sha256:8d989f54009484d7f43fe79d7c37757418fc3c15cd4d7e80a2076677a9645bdf
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhods/odh-ml-pipelines-cache-rhel8@sha256:33aae32e3d348d382906f6e926281852f54006a94d7bdac188bb33debadec059
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhods/odh-modelmesh-rhel8@sha256:3827221f84d3aa22440d6b364b990a4ec0666dfb5963f174f7ed518b3650d223
        - registry.redhat.io/rhods/odh-kf-notebook-controller-rhel8@sha256:cd0ab8d70559be6074fbaf1caa1d50f8eb75e1bff1a22d6d69380f19190df0b6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:eb96f4b753606528750b3c28f7f9256582f81fa030450b05afdef684634903be
        - registry.redhat.io/rhods/odh-ml-pipelines-artifact-manager-rhel8@sha256:76c67d7015a2db6965219db6533e0cddb5b7cf699c88df3ef81155f9c192a3cc
        - registry.redhat.io/rhods/odh-modelmesh-serving-controller-rhel8@sha256:d94395fb251e45aaa73d09bc6106a2947c16b90d87fc3b64bfe4f2f600be5841
        - registry.redhat.io/rhods/odh-dashboard-rhel8@sha256:b98171c5543815abf7c095fba92796ceed9ea44bd8b78db143d2815c6bcfdcf6
        - registry.redhat.io/rhods/odh-notebook-controller-rhel8@sha256:753994fce63802f1e1121990a5c1dd722216446abee54cc1a7c50ab195134414
        version: 2.4.0
      entries:
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: beta
    - currentCSV: rhods-operator.2.21.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:7d0b8d69b6defb5f232da066573b0067ec810a8ccfb2f3cb91d472d2979ab7a7
          createdAt: "2025-09-24T14:42:33Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.21
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:687c8eeed55f021ecaab1307f0e88b5b16d91f72d63b3d7168d7bbee90e8947b
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:04859c878ffcc8c42550a98be625ac52212f9b79a6fd5b3baeafb3592926cbe9
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:2a16c353fd13dc1f3f5c4266f1b9786a8e7a100dc05f2d554bdaef082271063f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:239f226328e5ac77d6dee498217d55103a6d562f4829e38abf65ba55fda0aca2
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:7da004e158006dcbdcf0c3b78c40411b82c9d56ebf54b1c30c245b6550b8ee7d
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:2b4136569e10ad7d47c82bedf55c1705a833dc67c8f833656996bfe1fddbef3e
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:52e1ba8ebf4d55c64c94c9076185c5612b537e6e5ed4944fcd6ea87ccb9cd182
        - registry.redhat.io/ubi9/toolbox@sha256:7656f815a0d13ed37f56d11635559a87a3d00840253755c71b644b858f67d556
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:7d0b8d69b6defb5f232da066573b0067ec810a8ccfb2f3cb91d472d2979ab7a7
        - registry.redhat.io/rhel9/mariadb-105@sha256:c390b12f5c2c3ea7cd897b59eba7127ce22f474130cd8c1bbbbaadae870baca4
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:1b5452b87b36de2b62febc777efdc6233a168ffcc09e2067e482cc444463a57d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:cb47e165935f0d9d3c88088689c37e68a69519357a20d3c52ab6f71240dab507
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:003fdc6cd1821d620e1ce4050ee608e2d2a32f2b08c8fa9c147e2e14418a8d91
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:80fd53a580ce7bc4d02023d38a04912e84c2cd50679fc5ea034afba935623157
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:fe6593bc73e77d315d82b7921bb1975c19fda1e7630ef5b30b69f5901ce30178
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:e68492a6a75cb83be0a968b832c1478b97f14e661376328085fd5416a08cbc31
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:c3e283c0536266195dd461042ca0036ed98cc33650d806be132db4fba45239e4
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:152b425725228953f43b57a4a034a6725993a5ca556af328ae5e06aa3f1acbfc
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:a88696cff8cd2cc2e295c9ec082b7f25f61dfc9ac12f1c208529ff7036d5af29
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fc98ed817a4e9ab84e2a66063987304f0cdbe6e5620a58a7f25cccd61b34ebf9
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:5c0d56f434298f6e8bf6cf7061cd1ae6df9915225cbf04b010b6b59a64775cc9
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:cfaf4eb86ddb9c717340ab17fa4b83323e5b40977ac70265376c87329a3f6f20
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:4c57dcf4df09fe0d03578cc8b9a3ca2ea485b6af0ca0203bfb85e640058b7daf
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:4c192208c618b49f31cf31c9de1a33c622ce3f120204c2eec8ff7c439ea6785c
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:a936d87646843edea2a84eb429bb44eb730bf3e04e8900fecef66721426dba89
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:9c2d6ce05775ac34ee5695b3367d12bf2a9c14e318a9544d00a56653ea5f2007
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:dc0db686b7688f98b5f8595661022cdffa5d26a078289cc7aed528f0ee1d08a4
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:6a388ee3cbe3813ac955d319f0ab65a81da321d982c9881c6c81fe135311fe1d
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5f581957e4f8deebf6104a394272e6fbb1c54d813bf069e95875d865eac209ad
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:2638e3a90bbb242dd2483fefe9fc8b472f334c510d0139b9b7c875c30dab7984
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:85202cb75250a1ba2691d65b52b306fd77566660dd29cdf56bf8962e8bb4939c
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:1b7befcd29c9c8f783bdd3c5654b0412ea45cd88cd18eb5f8f5a411f1323c6bd
        version: 2.21.1
      entries:
      - name: rhods-operator.2.21.1
        version: 2.21.1
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: embedded
    - currentCSV: rhods-operator.2.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
          createdAt: "2025-09-24T13:27:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.16
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:22557a6fa52d2f311750a9ba253860f423ba697d26efa02ef8524a8258d2a909
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:d7d4fa406e0fcf0507894a7676532b27f45be742467e603a86f98ea5d2615df8
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:2ba2b2c4db8bb334c50f4dfb54059f060361186900a44c06eed00b7a3c43977e
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:7b70af8847b0806d8b43c399e2b3109f016fd864f5e9d30c44e2baca5d1359dc
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a0f72ffefb2b74b488dd949493f5d295a39bb9c97f578bf219d0138601f65468
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:6cf74044ae8d5308a2dfe03fa5d81086c89302db7f4cdbe2f4174a1c48b77869
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:de5f91180ead7d73a1825fe8b032fde9b8b01392569c9789f5ce1b4b9c08a98f
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:ee01e89f98feb185f6cd59c564e590a13e4d8d9ea760cca8de51426eb71b83a1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:04f305c44413ae7dcb7017e53570ee49a509701792c5f50efadd64f47395730b
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:c8456d98e90b6505957ab3686e9fd2f156e29f123c5558e581c206daf1e7d93a
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:f738aab1eab25854c93e1d8d4d98100a8ae7bb45a6b83f0326774e4220b1183b
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:5dcdcc2424602a69451f16d31dbfa1d43cb72c095ba561eb9076f0cd1e8182ed
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:efd039012559786deb3c521a130886c265e88d635e08baace2e573a9df270134
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:27b113d96453c2054d4c965a963d2badba6daac235eddaf23234c76e87bcb069
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:4c5ff3496b2a2a739939d94ee9dafc02b682100785d228dc2fde480fb597b7a5
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:4a7599f8a866eb169c9a62885906adbf6df0417c0be15857df1eef20cd9b1be2
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:633984b538f027c93c5886ea1045dcbe81eeda74acf80001d5fc5f765bdbe0be
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:23d307a36b69e0df04f72a7d3b35e28d8417a8bbe23dba31e8e977569785c078
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:11a301728573adf2b64ea72d0cb2d83ea5d4dbebea759f346e99f18c3d368c6e
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:5192f1dbd2a9ab92ae390c4ae506efbed0970545b6122e95b014728ac937e777
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:2eedd1e2465d2e78fafb39e234dbb413a82713f187c64d70a80340ec94807d30
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f37e4048f3a152798286793f3abfc6ed814453fcbe2667255a7e78eee483c5a3
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:97e2bd9b587f08e135a9aeb9b3e0dc6eafa1a9bdacbb5ecb681ce9bd5aa37fc9
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:65482864055021272a18b5b26792ab00cfa5fc9cc005d8d3a884cc82991506f3
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:4a212ec634225c14beac09be24ddf336e562f2aa9a13555fb1196f366ddae23c
        version: 2.16.3
      entries:
      - name: rhods-operator.2.16.3
        version: 2.16.3
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.8.5
        version: 2.8.5
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: eus-2.16
    - currentCSV: rhods-operator.2.8.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:3b9e8a71ba795a2ea91f4129253ccb4be3a79c7c1c6a87082f96ccc67eabbbac
          createdAt: "2024-11-07T07:17:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.8.0 <2.8.5'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:eeb8bdc94e7fa2b454fedf72eb75756052f8bfb9d92b116a8267a5584e2528c1
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:f9c474d454d0b006e06859601ec9a64b5e5c91e9a47bfab6ad6d8cce5aa6b89d
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:45498198fadfbc1457bf2a6084ae7ea515d454d86d148783aae3abdaca6e999d
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:60507a3dddbd88560fd85204699147facd2b6371fb25e0bf6701dad02c0b4dd4
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:df34e9d2d121a3def2dbb86f108b3fb6948bc7ba7cb44c70801dfd2e81d3fbb1
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:13dff890ec67fa56864c244a40fc404e9eac05f33573dd16d088b37f9340ec41
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:28ec01e514c15cae145499840597d1be1b750123bb6014de2ecc3ef0d60fa134
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:9aa5a93831992588766ac8594ae3c4dc7111d08a90fdbc8751c40484339e46d4
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:038ce76571217d6c624d614652e8b96ef6484562cf52ec1be88392b8ef8a3bbd
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:1a1074263ae26e4c8b0e3863228c35e50133169f377ece6c47b58a12e0fcf453
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:81ae2b95f2da740cb5264b9ba8b85f0f58bb4ab80d6431f66d52848ed0c21afb
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:1dbf09cb1624c9f961f8ac8e371c97aa7c3fbfcad4c4d02ab4fad0eb86d2aa95
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:61a506fbf263cd72951ac8bd151127e0c74fc74f0a05f7aaecaa54fde47ee490
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:3b9e8a71ba795a2ea91f4129253ccb4be3a79c7c1c6a87082f96ccc67eabbbac
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:13b850a5a3f8776c6c8db887c666d88df72901895b5ef1be9a8bd372cc02fcb8
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:8ffc5da798e783c65d5be7f344219b2951d8c2233d6494a4ae2bb5c86eccd699
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:d87e4b1cba72eb4f50b3f2d59863799929c22ca37b01a1ff4828a9ce0be31f25
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:2f6db789ceb3f90f9c9911c6392e62d0a8e4d61be68a7f7cbd781f47faab6389
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:4b1b9f05a94713292a7a2f6782fe1d3eea6f36214c36b38f402aa91a554ab18a
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:5c48474563d98df32e7243e33302428ea8836be5e1a2e30d20b4d796a62cbf6b
        version: 2.8.5
      entries:
      - name: rhods-operator.2.8.5
        version: 2.8.5
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: eus-2.8
    - currentCSV: rhods-operator.2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
          createdAt: "2025-08-12T18:04:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "llamastackoperator": {
                      "managementState": "Removed"
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (including PyTorch, TensorFlow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack – unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.23
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:edcce7cddfa4c250108532e56541ddb95099612c180b2297e5934bc89d2bb46e
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-llama-stack-k8s-operator-rhel9@sha256:626b9ef09ee82c8d425ddb7ed3f2ffbb3d1019f541e31604db891787c881d5c9
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:ff740fc3615edaa90dae976da1aa6455d64811e239eb66b422836a3aae791b41
        - registry.redhat.io/ubi9/toolbox@sha256:4213516b6fe7957a01072720efd9ecbc7d4443238e151b5fe0263bd0c36792b1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhel9/mariadb-105@sha256:050dd5a7a32395b73b8680570e967e55050b152727412fdd73a25d8816e62d53
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:68e74bdf518fc624cc25a59a18b2a87d0aa2a9b8aad30750ace7115d297a21ca
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:a1497e64349634b2567d28f685052de118f64f6b32e7d3ea469212e2004d7eb3
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:e50adae13e9d773e7be8bf81b7525dac3f5e3864c9cdf4c287342bd568ee1d6b
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:84c3213e11299ec032aa0e5a4dfa9b3055c4246b04b029e888d8d9f1caf5c30c
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:ccfc84028a95f1cc6a16e06ff6e9ed2ace74f74e5e9963a4fb33544c2ce005c2
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:bb7c352ccc10a8c60ef8d573368d0ada59e73cd131653d11bcebdf9f564acc10
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:2d2e7803b2b1e15017ca04961f9dcd25e89c47b07b101ca54cb9e793e6027202
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:b038d20def504c140e4365f2eb117a979534629332bec75a2ba750ebafdefe3d
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:3243caa4c8aac75069d19225dcee361bf9c012782328a6266b7016b3026f6d56
        - registry.redhat.io/rhoai/odh-llama-stack-core-rhel9@sha256:43b60b1ee6f66fec38fe2ffbbe08dca8541ef162332e4bd8e422ecd24ee02646
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:23c098f7e791441cfcfa6bf37de51763511bd8bf4d543747b7d37ed591d68886
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5fb7e6c3ee1e3b170cd146ea9cd172fa953326b08163897605ca74a3644f2819
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:857560240bb45bfc2db6d7743abf11a47f929a0d3035940256a6527afd313412
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:795ab1842ef675e3d8b47deb8a7485a69064428e823dfeb930a6a7e7017cde0f
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f23f8051886d0f0240937507471c818f440bca8703b8209a836d5dbe3ea2f7c8
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fbe10282ada2fa6bf07ae97b497f46ebe735de130b4ce0182bf34301429999cb
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:80d284cac2ef6638fe028102b71a70f26041576c11a49414e77430954a3c1b99
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:467b04f4ee4118bbe0934d1ab8b80f72df8703dcdac61c2b53fdb5bd1b404c14
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:ca21e218e26c46e3c63d926241846f8f307fd4a586cc4b04147da49af6018ef5
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:e0c9ac73891781ef421e0a020cfd5c962f7632d38020897183cf048b1819326a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:a4649e8ad4970bc3d5e967811c9694fa82d78ee9f8f025b8b2fbd37237eb14fc
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:208005da389b7d4176ad8bffc703312011a1299954df151e2abbe8f3dec01d19
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:240692b86bc8a4fe0b545e6e4cd7cf4ff0ca990406bc55f0446d565fd67e37f6
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:756b7bdaa7f6e7c4a18488e690188dd91153a33913d225bab001b0085382e514
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:2886e40f94fde72eb5959ad564bbdd466657208c09bac731b3860b0b54f104c6
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:16ff53f4f1a1e1e66826c8355c654038da83df545427a931485ec7a2d3764b16
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:4847ac1be55284815670c2048c1edfc56c7227494280d3c6094aa23f1ed1a9d2
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:da8debeb2b84a42aba0b98785ccc20e02ed520e125428fe012df3ca473cbb755
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        version: 2.23.0
      entries:
      - name: rhods-operator.2.23.0
        version: 2.23.0
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.20.0
        version: 2.20.0
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.18.0
        version: 2.18.0
      - name: rhods-operator.2.17.0
        version: 2.17.0
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.15.0
        version: 2.15.0
      - name: rhods-operator.2.14.0
        version: 2.14.0
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: fast
    - currentCSV: rhods-operator.2.22.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
          createdAt: "2025-09-25T11:28:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.22
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:7b3d7787ba95156de9c284ad78f2d63a0d3bcb525f4d8d11ebf0203d4c5c1fe0
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:7fd540586a1ca111370100a473923b3dc6cec4255e9e4f3654bfbabd9158a2bb
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:a44d32a18b1e906c1411cafc82102881573fcd04e040a4e2ba6bd1bd617b70fa
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:6234e15f001122ab222481321153fb0eba7f4d86bf3dfd64edf3f6cecbe7d62f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:abe1f9dd11c7313fdd55b678be511a5cdc3546654a68e2b53d42c1013fec8d00
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:b2ca229bb4c4dbf04b96766b8af8d9f55f2edef59ec4110de0d20f8c857e7d27
        - registry.redhat.io/rhel9/mariadb-105@sha256:5e77b33ce04c1eafc2e8840586ad6a5b51148dea8a46dda151071e28858331f0
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:2f3f305c97e50057781553416fb4026cde306ead0fbe0213627b9b5bbcc8a0c3
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:e6fbe6723b4801f6854303b3fd7d4a4a36727bdd28dc28002b7de4929739ccc0
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:2f1a0a0803b2f5118b167e9d9c2e86660add5f79ee6f47044f63b0e8abe0b75f
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:cb5be3dd7098f4c8d66f9e85ad04656460df6102652deddc6cccb884a12be733
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:9104cbc2d595933f0a2acb7313e41ed69905e76a9fae6bd9820e7717b56a632b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e3b90530c6e38f6940d7f9408e9fa2c81eb06a857ec2b0bab5d006b1e9a63a4f
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:32c9511846077ef6035eb19be7653f6a78e0f207eb420ab93b18f742ab5779b6
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:27415b1c763cf89b79391f0b039936c3e078a078bda4420f13bf10ef9be31d5e
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:463e9be9c4409ba1e5a42713bd65844a023e7a3e9032d4e862edbaa5fe35300a
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:1709fa3c79aad4ba7eb9be8299949396092c8e20210124e0c0936385bc04e839
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:b2c2ea9d35edbcd16e1ae559c88a453129b877fc049ab7a7cd6c2ebb8f0bcb59
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:bd6d1599b4ff9eb890be9808014715b28ca19247222d65f40df024db4dd3525a
        - registry.redhat.io/ubi9/toolbox@sha256:ea4ce8cb8f59cf1b0709f1606d7cab074c3806d4016964eedc0555410cb5dfeb
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:d59ed40de694bc0e0564bab159a5b586cffa0721b6d761fb2d81dd42e6225d1c
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:74d4e0174655cf6f6273857325a1e714a1531c5a06ad38bf139bee9c7d8f5f19
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:5d79f0bcca0a213b53aee6d88834e503d343821596b857b13d3a6362de8ebeab
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:5914f71e37c815ce2567826896e051303bdef52fae1acf8d02e672fc24c7eb9d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:2065b96bd6f0a830eab66e36217a927448cfc5a7dabf2a5e7d907aa6c40657a4
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:2b3aace0d80474470e71c2c8291b25738e096109229fe6d5d3e06e9885ee58b6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:13f3679ab289eea9de54fe129a7c83f6c8ec5d26b61b430637a7c4cfcd31a74a
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:ccba67a6a5aa3403e0018f29dcdfef483220dea8065d8b572e80c653ca992add
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:2fe9473cd1d6c445d2ca7cc357ee4bb009aa1838f9cb69ea418296b2a044816b
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:d7c4ac937add511460a36375d9b6b3c475a3a9a7cc96c00df2fbad6219aee0e9
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f9965177de24f3165ec4c001c14cf1e2c2c01ffe847839af837c442e3c96beae
        version: 2.22.2
      entries:
      - name: rhods-operator.2.22.2
        version: 2.22.2
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable
    - currentCSV: rhods-operator.2.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:8643f7ee8afacfe6ae08cbbf81da76ca97032dc3ce3da4fddf16e330e9e33a08
          createdAt: "2024-11-14T21:18:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.10.0 <2.10.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:a1d590d0e2b3fae9ecbd843a664dab083061e797f66a9f0a021b0c2c487d7e45
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:0af6a91553f40ea40e86d85edc204658af3e5a45e8ab94cdb6b767e545df59b9
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:214ca0608fcbc0fd585bf01768d28c7ba17082d1043349ec8445afd663095383
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:d2514bf4b4b2d9abc19facac3f1a55405eb3ad0d9e9203e8816bfcfa08c9a6c5
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:4876f14ca9cff0e2b998a56f2f56d0775edf3963de610f4b810b4d94fb12d719
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:24238abd46633709feb75f2a3e41cf6c9e8528d2c0e4be2fac6f56de44e3fb9b
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:d9b55e8f25b81050380b4aab040243dc18a7432c48912c9b660c97dcc5eea008
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:8643f7ee8afacfe6ae08cbbf81da76ca97032dc3ce3da4fddf16e330e9e33a08
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:9f7b1cb8623de8b0550b0ba124d3b50f5e960ed011b33a3262c6346dbe2dff95
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:807ec05e9e107afc969e718d0c2c980959e2fd462e777a5ff63e55156ea150a4
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:fdc2bc06b0dfa79ebc6451ac3793b5e8d9f165171633831dafe46ebc18c10c35
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:4da77afafde99b15fd357a4ab6609644bf6330a1d9c0c03e64726ed1b9633230
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:cba9f25ac6bb2734bb990a622b515c838b881edb688bf44ff196ebe7684ab74b
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:8911afbb65c8c4f10c3597624485be54b9103e87a5020b718b35ec3df2c9bf00
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:d709f65e1194bc877fa4f9ec681b598467703d94139d57b58c24490d3a59fb82
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:220cad7a22d40998063c6b4fb2f63e69330e64ca7118490a6680b05fc99b7733
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:8fed8f58732bff512289fff738d0a94be67286c41b24eaf5928bbcb30d88c88a
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:232ebac25084db056fdde49477b5c1e6a03a0228fea3758bcf6234b3d0eaa582
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:2007da5abf840712cd2edb50462b2d8d571a0193c3ae1550d4a279eddf2b335f
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:a4b9553fde02d9c67a9c98f063eaf82fb1d2689f4b1269e57bf5f12c8d4adecf
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:5d15175f934017c04d6a6057200ff55b2172a1f4bcbeb0f2cd9e9e545631d93f
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:f4d7d3ba4a0d3322bc9f327e095e9ec56073758f6dcd8da8c7bb185cf952ead5
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:7df2f8bc2c5c0f4f347d009b62a058b5107413a2abd789f6010eb4d6c9e4f8ee
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:a9c6e3afdf605465766fa93334cc0f6bb5e46a5fdfd7ec64d648e2ed72b05ccf
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:267d6795bf53f9a1955d73445776a428dd042af0cd3cb8bb5b91692a1bc1e5d2
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:65db5e8cb2fa46d020ed4837d4479d133b52167c373e65f81cb590eec5391af2
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:0ff5664c2eaadf3d3073b3d7fb711da47fb78d21f67710d8b8aad5452cb0355d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:12741379ab24d2286b5fae77f4383fb65713af7b3356fb43d8254ea85d4b0ee3
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:52e8deeeeaa21b17370a5d5cd644745da745b6e1cee8c6295d6df75c8d2aec04
        version: 2.10.2
      entries:
      - name: rhods-operator.2.10.2
        version: 2.10.2
      - name: rhods-operator.2.10.1
        version: 2.10.1
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.10
    - currentCSV: rhods-operator.2.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:28b5fff602ad6fb36d462e1905f453509ec22fc251a1e0cf64fedba86ff420d7
          createdAt: "2024-10-17T10:53:18Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.13.0 <2.13.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:cd6b535d95eb1b44fe261a5336270fd898b966f431c14e173944484381c26d25
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:c2bd493e9345eccd124bbe0440e45cce8cf57588827d8eb9533f8bbadaa3ebbe
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:74e6cfa0bc481588cb8dc61c206ad567fbd2b28f43c2a86bee4a8e0eff55447c
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:28b5fff602ad6fb36d462e1905f453509ec22fc251a1e0cf64fedba86ff420d7
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:77bca19180f30ef08f4d1598ac6c365e190bb7909ac4a1fb35d9ba9ba8d6ff7a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a37529eff3656a570d796083404e5a3f63603afecfe9e738da1a243c0cb3d7c2
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:5d6354e65106460ab9285b42124faf288298387a8b16317565bd591f009c39a6
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:fb630d6f5817512733b8ea05f7147121928917a24ac8d98239f43d920880c359
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:6b2c0d1dd40e2c276684e17818b14375bc778e6c88df0bdc3cc24bfc9294804f
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:8883bf71abf39b4a6985940c287eb6ac7dc137d491264805d48bb112b65bd008
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:9ccdd7fe8223057a2234a3a2eb3b492e7b69e6ce714cc26a805fe2dd82935195
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:850901511bde065cf73e4c18d787dc10b6e91ca6139fc3ba68da28c41dd28236
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:688389195aebf3f43e74e40d591d2fd5f9ae41994d8ac2db661f292f6eda012a
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:2566b73e98b783871fb500bb1a8cc4e3d58a513c1bb18abfb525da0818e6138b
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:fd4ce524acc1769475079c7231f5ea79ac29551c513855ef34e5694d2c9fbbd3
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:a55f73d72ad2b07f01c49949cdb00969bfe4789375b3d7cdf9814b969fcc808d
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:7d085942e4da73c2ff4885c62f10b904dc70fd08e819e62f925eb1b4ddfa1554
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:8faab013568ee023968a5ec5b3a79977d7ddae3deafff6328e6fd56a27198e24
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:71abac4fb3d613edef7f3ba117a73bbeea61f341d6182e8310360a76576c789e
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:030c53690fe25e15bc9907f51c3f04fab39b4d085ff2098d26316ba32cc8b09f
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:8c19c7fa535f36bdce3b73cf3300e321909d44954cbc63a8aef6321c5f2fa693
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:59e9cc9cd57ba48ca7e87ae1e797a902f826bb890cb198ccf549f76ea581ba6e
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:bf25fce160dd7b8cce11e68a501006e4fa81a7858645038cdefdbba03df7ecdb
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:a3048eaa8924ce3571c6e78014438183fae70d9e8c833a7c07774c675104e3d4
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:5c2ee4136ee0014d1ef395b5e3bcab125c846f5e258ece3631d912f431c8d208
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:7c895490d49beb6bd724f47ad1b17ce501650557b0889314799c4072632f2b99
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f5b17f44c5ab72d6cf658f72f24f15e06a5ac9be1790336f7228ab8e7cfc15c3
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:cf3779c64bbaef1988886ed629c516e7d6ef7bdc73b9c95ad4c087b62e75750d
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:363a35c332fc80b068157951d5e50e6b7209947649cc64803ffc02726e7581ab
        version: 2.13.1
      entries:
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.13
    - currentCSV: rhods-operator.2.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
          createdAt: "2025-09-24T13:27:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.16
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:97e2bd9b587f08e135a9aeb9b3e0dc6eafa1a9bdacbb5ecb681ce9bd5aa37fc9
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:efd039012559786deb3c521a130886c265e88d635e08baace2e573a9df270134
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:d7d4fa406e0fcf0507894a7676532b27f45be742467e603a86f98ea5d2615df8
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:7b70af8847b0806d8b43c399e2b3109f016fd864f5e9d30c44e2baca5d1359dc
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f37e4048f3a152798286793f3abfc6ed814453fcbe2667255a7e78eee483c5a3
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:4a7599f8a866eb169c9a62885906adbf6df0417c0be15857df1eef20cd9b1be2
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:4c5ff3496b2a2a739939d94ee9dafc02b682100785d228dc2fde480fb597b7a5
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:4a212ec634225c14beac09be24ddf336e562f2aa9a13555fb1196f366ddae23c
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:2ba2b2c4db8bb334c50f4dfb54059f060361186900a44c06eed00b7a3c43977e
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:f738aab1eab25854c93e1d8d4d98100a8ae7bb45a6b83f0326774e4220b1183b
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:23d307a36b69e0df04f72a7d3b35e28d8417a8bbe23dba31e8e977569785c078
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:04f305c44413ae7dcb7017e53570ee49a509701792c5f50efadd64f47395730b
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:2eedd1e2465d2e78fafb39e234dbb413a82713f187c64d70a80340ec94807d30
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:5dcdcc2424602a69451f16d31dbfa1d43cb72c095ba561eb9076f0cd1e8182ed
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:11a301728573adf2b64ea72d0cb2d83ea5d4dbebea759f346e99f18c3d368c6e
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:27b113d96453c2054d4c965a963d2badba6daac235eddaf23234c76e87bcb069
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:c8456d98e90b6505957ab3686e9fd2f156e29f123c5558e581c206daf1e7d93a
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:de5f91180ead7d73a1825fe8b032fde9b8b01392569c9789f5ce1b4b9c08a98f
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a0f72ffefb2b74b488dd949493f5d295a39bb9c97f578bf219d0138601f65468
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:6cf74044ae8d5308a2dfe03fa5d81086c89302db7f4cdbe2f4174a1c48b77869
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:633984b538f027c93c5886ea1045dcbe81eeda74acf80001d5fc5f765bdbe0be
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:5192f1dbd2a9ab92ae390c4ae506efbed0970545b6122e95b014728ac937e777
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:65482864055021272a18b5b26792ab00cfa5fc9cc005d8d3a884cc82991506f3
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:22557a6fa52d2f311750a9ba253860f423ba697d26efa02ef8524a8258d2a909
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:ee01e89f98feb185f6cd59c564e590a13e4d8d9ea760cca8de51426eb71b83a1
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        version: 2.16.3
      entries:
      - name: rhods-operator.2.16.3
        version: 2.16.3
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.16
    - currentCSV: rhods-operator.2.19.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:c605e5fdb3d66302781146347df6da5a91dee97e4f8f5d9ac33f105b8cdee383
          createdAt: "2025-09-24T13:27:42Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:b30d60cd458133430d4c92bf84911e03cecd02f60e88a58d1c6c003543cf833a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:f0ee0d27bb784e289f7d88cc8ee0e085ca70e88a5d126562105542f259a1ac01
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:8ce44de8c683f198bf24ba36cd17e89708153d11f5b42c0a27e77f8fdb233551
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.19
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:8a6732ac3ca60f13dcb1fd8f186d2fc5dc601b9b5981a58f039af9e237f20044
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:b20ee16b776871d7c49b23f2bc91726812998561129f5dbd14eddbfe57964c93
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:c605e5fdb3d66302781146347df6da5a91dee97e4f8f5d9ac33f105b8cdee383
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:b30d60cd458133430d4c92bf84911e03cecd02f60e88a58d1c6c003543cf833a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:f49483df1d342d8e39f9a8085920aad79313a475df02a2292a136ecb784402b5
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:c8c23c07038f59db848dea0d4247b62840dbe58e1c581dc0722fb9d7f9be3b75
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:4e8f6c6ca0ed516b8310af0332637683b5bfc8f2946eb935c647698dffbbcc79
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhel8/mariadb-103@sha256:f0ee0d27bb784e289f7d88cc8ee0e085ca70e88a5d126562105542f259a1ac01
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:d24e277f468b0a48f38801e8e0bcfeb8264c3f7c128073ddaa17934b8572cc3c
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:520ed21b80e5487b3a0a13b7e5c7d791d4dd28e0b6799d57abb65d53fca5a03f
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:6275a19ef06114b33d79833a94ef10cc8203368eee98e5d415e73d20cae1db85
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:87685c5cab105d876ff0d14b0e24d8dc2a1ff5b55bebe36028c5f3376b2a5271
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:6ad35cf5c4c62ed5ad4e192afe8cebead9796cb24a964401c76ff8656c47742f
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:b6039c3c33a5dc345ceb8e757437bb13023d2bf85cc2b4c10b9ab3887e78e0c0
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:6f01ac7cc62ab961b9c8a630a558338a61f6d7902301c4a1ad337495ac7ce85a
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:4313cae03c041d461bc276abc23d706b582b5d16feab316796ca43478b9dc705
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:b24c25dfdcefb4ec26e7f311c00798df108f7685def2b1762b7624fdfb160ebf
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:bbef4839f24facce111d21aca43898dcf4e6cf57adb16200e3f1590bd5f89eae
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:49c02285fcef6949892cecca20eeee49d03a9bb4d17bb9ceff7c7a0dbbb8ab40
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel8@sha256:6166f525948a018c41d735014d97fdfa52024f7be31e18976d85d1a0d88901a0
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:baefbe4e7dd5aa74c71cb81da4a7a2b888a02d3eb1398ec4419bd8003b77f0c2
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:8ce44de8c683f198bf24ba36cd17e89708153d11f5b42c0a27e77f8fdb233551
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:af4721d61dfb61c75488a1bd6a5f62b23e20d7b7586a9cf51ba062ece2837715
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:bf6231bfdbe52b597d86e03e4730ec105ebe7a639fcd529268b7c0423f2a890c
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:71cd15f0b1acba3a48ebedff037bf8a846b426c18363a4de8896a376b6b7b3d5
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:53ac36baa374159b9065c718a9ede821bbb61d9ebe9502b2243e0a9f7aca0d16
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:d47514cbceb3cc96e9d33b7db6fc1df60893e5c4118bf6086cdd27f04c472773
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:876ca7ce7b0b975688fb4a77d6cd1a6f87cf9eb40c2a9a5b84d8cffdc80eeebd
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:de0e4b7fb822101198817e83cd7e9f4f7406e75ec91d07a833a27a417cffe66b
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:c1c97b6beb759d3f6f2dd53c4258e4ccbb3bf35cf11873bd446b135defaa69d6
        version: 2.19.3
      entries:
      - name: rhods-operator.2.19.3
        version: 2.19.3
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.19
    - currentCSV: rhods-operator.2.22.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
          createdAt: "2025-09-25T11:28:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.22
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:13f3679ab289eea9de54fe129a7c83f6c8ec5d26b61b430637a7c4cfcd31a74a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:d59ed40de694bc0e0564bab159a5b586cffa0721b6d761fb2d81dd42e6225d1c
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:d7c4ac937add511460a36375d9b6b3c475a3a9a7cc96c00df2fbad6219aee0e9
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:b2ca229bb4c4dbf04b96766b8af8d9f55f2edef59ec4110de0d20f8c857e7d27
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:9104cbc2d595933f0a2acb7313e41ed69905e76a9fae6bd9820e7717b56a632b
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f9965177de24f3165ec4c001c14cf1e2c2c01ffe847839af837c442e3c96beae
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:2f3f305c97e50057781553416fb4026cde306ead0fbe0213627b9b5bbcc8a0c3
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:b2c2ea9d35edbcd16e1ae559c88a453129b877fc049ab7a7cd6c2ebb8f0bcb59
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:6234e15f001122ab222481321153fb0eba7f4d86bf3dfd64edf3f6cecbe7d62f
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:5d79f0bcca0a213b53aee6d88834e503d343821596b857b13d3a6362de8ebeab
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:a44d32a18b1e906c1411cafc82102881573fcd04e040a4e2ba6bd1bd617b70fa
        - registry.redhat.io/rhel9/mariadb-105@sha256:5e77b33ce04c1eafc2e8840586ad6a5b51148dea8a46dda151071e28858331f0
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:7fd540586a1ca111370100a473923b3dc6cec4255e9e4f3654bfbabd9158a2bb
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:2fe9473cd1d6c445d2ca7cc357ee4bb009aa1838f9cb69ea418296b2a044816b
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:27415b1c763cf89b79391f0b039936c3e078a078bda4420f13bf10ef9be31d5e
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:7b3d7787ba95156de9c284ad78f2d63a0d3bcb525f4d8d11ebf0203d4c5c1fe0
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:2065b96bd6f0a830eab66e36217a927448cfc5a7dabf2a5e7d907aa6c40657a4
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e3b90530c6e38f6940d7f9408e9fa2c81eb06a857ec2b0bab5d006b1e9a63a4f
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:e6fbe6723b4801f6854303b3fd7d4a4a36727bdd28dc28002b7de4929739ccc0
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/ubi9/toolbox@sha256:ea4ce8cb8f59cf1b0709f1606d7cab074c3806d4016964eedc0555410cb5dfeb
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:5914f71e37c815ce2567826896e051303bdef52fae1acf8d02e672fc24c7eb9d
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:1709fa3c79aad4ba7eb9be8299949396092c8e20210124e0c0936385bc04e839
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:cb5be3dd7098f4c8d66f9e85ad04656460df6102652deddc6cccb884a12be733
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:2b3aace0d80474470e71c2c8291b25738e096109229fe6d5d3e06e9885ee58b6
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:2f1a0a0803b2f5118b167e9d9c2e86660add5f79ee6f47044f63b0e8abe0b75f
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:ccba67a6a5aa3403e0018f29dcdfef483220dea8065d8b572e80c653ca992add
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:463e9be9c4409ba1e5a42713bd65844a023e7a3e9032d4e862edbaa5fe35300a
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:abe1f9dd11c7313fdd55b678be511a5cdc3546654a68e2b53d42c1013fec8d00
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:bd6d1599b4ff9eb890be9808014715b28ca19247222d65f40df024db4dd3525a
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:74d4e0174655cf6f6273857325a1e714a1531c5a06ad38bf139bee9c7d8f5f19
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:32c9511846077ef6035eb19be7653f6a78e0f207eb420ab93b18f742ab5779b6
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        version: 2.22.2
      entries:
      - name: rhods-operator.2.22.2
        version: 2.22.2
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.22
    - currentCSV: rhods-operator.2.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:b3563e201bf2720f60e73c0b04cb70566ea52eda7d09a28753f3a1487993c2e8
          createdAt: "2024-08-01T06:23:03Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.8.0 <2.8.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:eb4da3a645b53acfd6ce8758c043ae19b4126e62bcca09165114a141c2c8d014
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:92d9e4b53d580df6cc1f14a849b477749a1306941a2e8f9d1b6160f4911c3cff
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:b3563e201bf2720f60e73c0b04cb70566ea52eda7d09a28753f3a1487993c2e8
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:5f39b11f0cd6db2d7d6e51ee666c030e2643aee92a090b49656c1e7dfd8a5430
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:90833e85e0de3d7de870074c8b04319cd11b913b9652af6a5727aae41a91458c
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:91525fac5fadece75fbdafc40b94445fb92c88cfc8c2f06b596f8c6377e59c9e
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:ab8792fcd16b1aa0065444ba54335cb3e747379740c3ec2d56beb46ff80cb044
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:987c768c0688431a53e48a80591f2a41e074994169f9e691b28b89528d8e736f
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:1a6e29214e06a2665baadbed598ba758d6141bfa522984eb05f05a80558a403c
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:850b78a201effcafe33705dcd2670f1c470d8b46ba6e307032403627b55a7cfe
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:080377d42d782bf525b4072384d0cf8409b48019b3748ca7c855f340bcb94251
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:b6ecb3858c201498024e3bfc4a93e630699be4b1fdc3a65fd2d70e4b375eb9c2
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:df2edcaf4402b11ce24439263956e3b0cc43efbac2fdf152afacb1a398dcc59a
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:889f046b255b13d99801108588d3abf577611b828657b5240a18ea21f0e3cbbe
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:1c49888ba2e9449c04d5588040c099290fce4793c2ba8874c1037316e550d6e1
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:0b8665e8978deb73be2cc65830f08a74a5a48f3e3ee7b2b817f7bbacba257d22
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:68e1331f13dfaf6641a1e7d54f4cf29c0ef42663abc98b6dc69167892d2d8adb
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:06a6d5cbbe80d49bca9e4db9929571a69e35ae7951fcae84dcd6460f7ddf622c
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:9799003449bfb61ae8f59efd2a611b0da3bccadd9b20cc1519f6c9279d085337
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:9c90873fbd5af8fac8efc3c7c27fba742a06fad2d47f336319dd05d2b116867b
        version: 2.8.4
      entries:
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.8
    defaultChannel: stable
    packageName: rhods-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: ansible-automation-platform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aap-operator.v2.4.0-0.1758079821
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.1-0.1635279521 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the hub, including all hosted Ansible
              content, secrets and the database. By default, a persistent volume claim
              will be created using the default StorageClass on your cluster to store
              the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous hub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Automation Hub
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-24

          ## How to Upgrade

          Starting with AAP 2.4 in the May 2025 async release, the managed database was upgraded to PostgreSQL 15. It is recommended to take a backup of your deployment before upgrading.

          To take a backup, create AutomationControllerBackup and AutomationHubBackup objects for your deployments.
          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel.

          Note: Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8-operator@sha256:5b957031609c8aabf7575e675bc636c1152babc9564d503802301aff4642a404
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-runner-rhel8@sha256:e14286dc9605be3a8c6cb144d0b2e3f3bb3211129005009e7566ea6c16c91f52
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8@sha256:7ae8db2b5d12b619f40db1e581db1e982a695bc1a29a2628222fff09e7c25249
        - registry.redhat.io/ansible-automation-platform-24/ee-minimal-rhel8@sha256:7670e41037f023e3fc6ee95d2342787a80077273ab8d30ba6804029ff05b4edc
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-ui-rhel8@sha256:d229d4e09b2b788d1ba3413ea7ec15374bd1613d507c285ac1ddc10570b9473f
        - registry.redhat.io/ansible-automation-platform-24/hub-web-rhel8@sha256:6ff1f602f0d9cad201b3addb65acc570a9d527de005ca85fc861d3cb98138710
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8@sha256:bfd13d6b39c5ec3a67adbafb05a0482328f0b9b97a517b45c093f6c35a3a92d3
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8@sha256:04fb5cb6b0d7c79cf5d4b754fe917224ddfcb2b9a4994462db404c833ed69483
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-rhel8-operator@sha256:b120e6a5ef376662b4324f0458eae31627087dc91fc613788143bc3954a15318
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-24/ee-supported-rhel8@sha256:f1384eb558f7da1e4a35d51408b3cf5a5a9be08f4d50503f473b26909bc246fc
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8@sha256:cccb4db08735551c697a7b2cbf75812f34500e691f9cdbaea02b30edea292e47
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8-operator@sha256:5253c9fd587ff2ca94d251f2b5fc7f284080adc5eb64643df5299beb2015592f
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8-operator@sha256:3c28a1326ecb7495952a5bb8be13b0972a4378af82611b0b9d990c5fa9fa82c3
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8-operator@sha256:e20106a8a04abd568a4c56861cea39ae444c0bd204bc807ac01f1705e12937d3
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        version: 2.4.0+0.1758079821
      entries:
      - name: aap-operator.v2.4.0-0.1702352195
        version: 2.4.0+0.1702352195
      - name: aap-operator.v2.4.0-0.1707506910
        version: 2.4.0+0.1707506910
      - name: aap-operator.v2.4.0-0.1708974895
        version: 2.4.0+0.1708974895
      - name: aap-operator.v2.4.0-0.1711590540
        version: 2.4.0+0.1711590540
      - name: aap-operator.v2.4.0-0.1713466670
        version: 2.4.0+0.1713466670
      - name: aap-operator.v2.4.0-0.1715898994
        version: 2.4.0+0.1715898994
      - name: aap-operator.v2.4.0-0.1717558474
        version: 2.4.0+0.1717558474
      - name: aap-operator.v2.4.0-0.1718152210
        version: 2.4.0+0.1718152210
      - name: aap-operator.v2.4.0-0.1720482179
        version: 2.4.0+0.1720482179
      - name: aap-operator.v2.4.0-0.1725256739
        version: 2.4.0+0.1725256739
      - name: aap-operator.v2.4.0-0.1726111413
        version: 2.4.0+0.1726111413
      - name: aap-operator.v2.4.0-0.1727257740
        version: 2.4.0+0.1727257740
      - name: aap-operator.v2.4.0-0.1730152858
        version: 2.4.0+0.1730152858
      - name: aap-operator.v2.4.0-0.1733185647
        version: 2.4.0+0.1733185647
      - name: aap-operator.v2.4.0-0.1733943951
        version: 2.4.0+0.1733943951
      - name: aap-operator.v2.4.0-0.1737661496
        version: 2.4.0+0.1737661496
      - name: aap-operator.v2.4.0-0.1741194076
        version: 2.4.0+0.1741194076
      - name: aap-operator.v2.4.0-0.1742444794
        version: 2.4.0+0.1742444794
      - name: aap-operator.v2.4.0-0.1743629049
        version: 2.4.0+0.1743629049
      - name: aap-operator.v2.4.0-0.1746132523
        version: 2.4.0+0.1746132523
      - name: aap-operator.v2.4.0-0.1749069319
        version: 2.4.0+0.1749069319
      - name: aap-operator.v2.4.0-0.1750877268
        version: 2.4.0+0.1750877268
      - name: aap-operator.v2.4.0-0.1753232084
        version: 2.4.0+0.1753232084
      - name: aap-operator.v2.4.0-0.1755833431
        version: 2.4.0+0.1755833431
      - name: aap-operator.v2.4.0-0.1758079821
        version: 2.4.0+0.1758079821
      name: stable-2.4
    - currentCSV: aap-operator.v2.4.0-0.1758080293
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.1-0.1635283332 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the hub, including all hosted Ansible
              content, secrets and the database. By default, a persistent volume claim
              will be created using the default StorageClass on your cluster to store
              the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous hub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Automation Hub
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-24

          ## How to Upgrade

          Starting with AAP 2.4 in the May 2025 async release, the managed database was upgraded to PostgreSQL 15. It is recommended to take a backup of your deployment before upgrading.

          To take a backup, create AutomationControllerBackup and AutomationHubBackup objects for your deployments.
          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel.

          Note: Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8@sha256:bfd13d6b39c5ec3a67adbafb05a0482328f0b9b97a517b45c093f6c35a3a92d3
        - registry.redhat.io/ansible-automation-platform-24/ee-minimal-rhel8@sha256:7670e41037f023e3fc6ee95d2342787a80077273ab8d30ba6804029ff05b4edc
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8@sha256:04fb5cb6b0d7c79cf5d4b754fe917224ddfcb2b9a4994462db404c833ed69483
        - registry.redhat.io/ansible-automation-platform-24/ee-supported-rhel8@sha256:f1384eb558f7da1e4a35d51408b3cf5a5a9be08f4d50503f473b26909bc246fc
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-rhel8-operator@sha256:b120e6a5ef376662b4324f0458eae31627087dc91fc613788143bc3954a15318
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8-operator@sha256:5253c9fd587ff2ca94d251f2b5fc7f284080adc5eb64643df5299beb2015592f
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8-operator@sha256:3c28a1326ecb7495952a5bb8be13b0972a4378af82611b0b9d990c5fa9fa82c3
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8@sha256:7ae8db2b5d12b619f40db1e581db1e982a695bc1a29a2628222fff09e7c25249
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8@sha256:cccb4db08735551c697a7b2cbf75812f34500e691f9cdbaea02b30edea292e47
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8-operator@sha256:5b957031609c8aabf7575e675bc636c1152babc9564d503802301aff4642a404
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8-operator@sha256:e20106a8a04abd568a4c56861cea39ae444c0bd204bc807ac01f1705e12937d3
        - registry.redhat.io/ansible-automation-platform-24/hub-web-rhel8@sha256:6ff1f602f0d9cad201b3addb65acc570a9d527de005ca85fc861d3cb98138710
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-runner-rhel8@sha256:e14286dc9605be3a8c6cb144d0b2e3f3bb3211129005009e7566ea6c16c91f52
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-ui-rhel8@sha256:d229d4e09b2b788d1ba3413ea7ec15374bd1613d507c285ac1ddc10570b9473f
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        version: 2.4.0+0.1758080293
      entries:
      - name: aap-operator.v2.4.0-0.1702352614
        version: 2.4.0+0.1702352614
      - name: aap-operator.v2.4.0-0.1707507304
        version: 2.4.0+0.1707507304
      - name: aap-operator.v2.4.0-0.1708975279
        version: 2.4.0+0.1708975279
      - name: aap-operator.v2.4.0-0.1711591330
        version: 2.4.0+0.1711591330
      - name: aap-operator.v2.4.0-0.1713467078
        version: 2.4.0+0.1713467078
      - name: aap-operator.v2.4.0-0.1715899379
        version: 2.4.0+0.1715899379
      - name: aap-operator.v2.4.0-0.1717558885
        version: 2.4.0+0.1717558885
      - name: aap-operator.v2.4.0-0.1718152680
        version: 2.4.0+0.1718152680
      - name: aap-operator.v2.4.0-0.1720482740
        version: 2.4.0+0.1720482740
      - name: aap-operator.v2.4.0-0.1725257213
        version: 2.4.0+0.1725257213
      - name: aap-operator.v2.4.0-0.1726111804
        version: 2.4.0+0.1726111804
      - name: aap-operator.v2.4.0-0.1727259094
        version: 2.4.0+0.1727259094
      - name: aap-operator.v2.4.0-0.1730153340
        version: 2.4.0+0.1730153340
      - name: aap-operator.v2.4.0-0.1733186325
        version: 2.4.0+0.1733186325
      - name: aap-operator.v2.4.0-0.1733945743
        version: 2.4.0+0.1733945743
      - name: aap-operator.v2.4.0-0.1737662040
        version: 2.4.0+0.1737662040
      - name: aap-operator.v2.4.0-0.1741194557
        version: 2.4.0+0.1741194557
      - name: aap-operator.v2.4.0-0.1742445409
        version: 2.4.0+0.1742445409
      - name: aap-operator.v2.4.0-0.1743630487
        version: 2.4.0+0.1743630487
      - name: aap-operator.v2.4.0-0.1746133932
        version: 2.4.0+0.1746133932
      - name: aap-operator.v2.4.0-0.1749069693
        version: 2.4.0+0.1749069693
      - name: aap-operator.v2.4.0-0.1750878254
        version: 2.4.0+0.1750878254
      - name: aap-operator.v2.4.0-0.1753232791
        version: 2.4.0+0.1753232791
      - name: aap-operator.v2.4.0-0.1755833968
        version: 2.4.0+0.1755833968
      - name: aap-operator.v2.4.0-0.1758080293
        version: 2.4.0+0.1758080293
      name: stable-2.4-cluster-scoped
    - currentCSV: aap-operator.v2.5.0-0.1758147230
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727257740 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation gateway - Unify the management of AAP from a central location.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner


          ## New in AAP 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-25

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.5

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console. For example, change the channel from `stable-2.4` to `stable-2.5`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.

          After changing to the new 2.5 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place.

          #### Create the AnsibleAutomationPlatform custom resource

          Once that is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: myaap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8@sha256:3a989f148648a5a62e687058ec3ca772fa22aa10faec89453626d2b17b3fd9c2
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8-operator@sha256:43e731b9a435b3a412108170b4b39e6b6215be51d362a9f03c33fa412e97da53
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8-operator@sha256:8d8182e3a87c55eacc222e71df4a9bf7a2e51102d913a5824cc91e36056d4249
        - registry.redhat.io/ansible-automation-platform-25/hub-web-rhel8@sha256:e96832efcc03c6754031cb70f56c808135f62d1b80a882dd6d6fc9e1a4cfc69b
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8@sha256:697343aaea153ab22d7efa8f1bb5aed37a55cbb353fc206f4a355e8b878ae418
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8@sha256:ac0de2b7d354eb0c15a6c12861a4de3577a982c84af661fbab25517d8105c085
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8@sha256:aed3973eb953dda3571260dcf1b33bc7dcbc18c41e3350b8c361b7e635fb9527
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-runner-rhel8@sha256:ace1d2ab49ba5ff9f11ab18e8edce27c9e73ae7298bf7335569fa1740a44899b
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8-operator@sha256:acc0fcd509eeac46db307ab744dd0a0c26770a93e495cf2ec0ccb61a91c6fbc5
        - registry.redhat.io/ansible-automation-platform-25/ee-minimal-rhel8@sha256:41f944eb1450db56c3b14fc7baece8f148d5f10886a57bad730bec60fb28ac04
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8-operator@sha256:07aa7e257d04ae0eda6fcaa4af9121f1aece439231fce4f924e6467da91f0022
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-ui-rhel8@sha256:9ab35e80a6172018ed3f5cd179e0bcb6befc1378583d9e73c3a64a7cf9e8050f
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8@sha256:9f2867603b9747c2fe37862c2c5efac4a4530902f323a8bf7cc0717de4cdb5aa
        - registry.redhat.io/ansible-automation-platform-25/gateway-proxy-rhel8@sha256:789d55f355fa0d90f9509a912d199123bfc5a4041930a72a8fd4ed1d819f0d84
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8-operator@sha256:503fc7a20e2dd7de2cb55b9be5f97efa0d2b31f054ce326f23f3a11b4d683e64
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-rhel8-operator@sha256:74be9182723b41f87b3d13b3d86b6a9049bba88e2902cc9453264aa8a0f584e4
        - registry.redhat.io/ansible-automation-platform-25/ee-supported-rhel8@sha256:52d2c73fa2015af5f1b9a88ee48085c7c7bab44330fc9f95914f48b58e4b48bc
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-chatbot-rhel8@sha256:9a37c9739c488ccf211ec5efc6debb045223d508c9cb3fa5a8256eca98ff2f0c
        version: 2.5.0+0.1758147230
      entries:
      - name: aap-operator.v2.5.0-0.1727240614
        version: 2.5.0+0.1727240614
      - name: aap-operator.v2.5.0-0.1727875185
        version: 2.5.0+0.1727875185
      - name: aap-operator.v2.5.0-0.1728520175
        version: 2.5.0+0.1728520175
      - name: aap-operator.v2.5.0-0.1729741547
        version: 2.5.0+0.1729741547
      - name: aap-operator.v2.5.0-0.1731456196
        version: 2.5.0+0.1731456196
      - name: aap-operator.v2.5.0-0.1733193164
        version: 2.5.0+0.1733193164
      - name: aap-operator.v2.5.0-0.1733969526
        version: 2.5.0+0.1733969526
      - name: aap-operator.v2.5.0-0.1736893446
        version: 2.5.0+0.1736893446
      - name: aap-operator.v2.5.0-0.1737101025
        version: 2.5.0+0.1737101025
      - name: aap-operator.v2.5.0-0.1737675968
        version: 2.5.0+0.1737675968
      - name: aap-operator.v2.5.0-0.1739248042
        version: 2.5.0+0.1739248042
      - name: aap-operator.v2.5.0-0.1740093573
        version: 2.5.0+0.1740093573
      - name: aap-operator.v2.5.0-0.1740773472
        version: 2.5.0+0.1740773472
      - name: aap-operator.v2.5.0-0.1741368605
        version: 2.5.0+0.1741368605
      - name: aap-operator.v2.5.0-0.1742434024
        version: 2.5.0+0.1742434024
      - name: aap-operator.v2.5.0-0.1743660124
        version: 2.5.0+0.1743660124
      - name: aap-operator.v2.5.0-0.1746137767
        version: 2.5.0+0.1746137767
      - name: aap-operator.v2.5.0-0.1747343762
        version: 2.5.0+0.1747343762
      - name: aap-operator.v2.5.0-0.1749074128
        version: 2.5.0+0.1749074128
      - name: aap-operator.v2.5.0-0.1749604727
        version: 2.5.0+0.1749604727
      - name: aap-operator.v2.5.0-0.1750901111
        version: 2.5.0+0.1750901111
      - name: aap-operator.v2.5.0-0.1753900156
        version: 2.5.0+0.1753900156
      - name: aap-operator.v2.5.0-0.1755835086
        version: 2.5.0+0.1755835086
      - name: aap-operator.v2.5.0-0.1758147230
        version: 2.5.0+0.1758147230
      name: stable-2.5
    - currentCSV: aap-operator.v2.5.0-0.1758147817
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727259094 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation gateway - Unify the management of AAP from a central location.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner


          ## New in AAP 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-25

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.5

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console. For example, change the channel from `stable-2.4` to `stable-2.5`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.

          After changing to the new 2.5 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place.

          #### Create the AnsibleAutomationPlatform custom resource

          Once that is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: myaap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8-operator@sha256:acc0fcd509eeac46db307ab744dd0a0c26770a93e495cf2ec0ccb61a91c6fbc5
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-ui-rhel8@sha256:9ab35e80a6172018ed3f5cd179e0bcb6befc1378583d9e73c3a64a7cf9e8050f
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8-operator@sha256:43e731b9a435b3a412108170b4b39e6b6215be51d362a9f03c33fa412e97da53
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8-operator@sha256:8d8182e3a87c55eacc222e71df4a9bf7a2e51102d913a5824cc91e36056d4249
        - registry.redhat.io/ansible-automation-platform-25/hub-web-rhel8@sha256:e96832efcc03c6754031cb70f56c808135f62d1b80a882dd6d6fc9e1a4cfc69b
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-chatbot-rhel8@sha256:9a37c9739c488ccf211ec5efc6debb045223d508c9cb3fa5a8256eca98ff2f0c
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8@sha256:3a989f148648a5a62e687058ec3ca772fa22aa10faec89453626d2b17b3fd9c2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/ee-minimal-rhel8@sha256:41f944eb1450db56c3b14fc7baece8f148d5f10886a57bad730bec60fb28ac04
        - registry.redhat.io/ansible-automation-platform-25/gateway-proxy-rhel8@sha256:789d55f355fa0d90f9509a912d199123bfc5a4041930a72a8fd4ed1d819f0d84
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8@sha256:aed3973eb953dda3571260dcf1b33bc7dcbc18c41e3350b8c361b7e635fb9527
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8@sha256:697343aaea153ab22d7efa8f1bb5aed37a55cbb353fc206f4a355e8b878ae418
        - registry.redhat.io/ansible-automation-platform-25/ee-supported-rhel8@sha256:52d2c73fa2015af5f1b9a88ee48085c7c7bab44330fc9f95914f48b58e4b48bc
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8@sha256:ac0de2b7d354eb0c15a6c12861a4de3577a982c84af661fbab25517d8105c085
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-rhel8-operator@sha256:74be9182723b41f87b3d13b3d86b6a9049bba88e2902cc9453264aa8a0f584e4
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-runner-rhel8@sha256:ace1d2ab49ba5ff9f11ab18e8edce27c9e73ae7298bf7335569fa1740a44899b
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8@sha256:9f2867603b9747c2fe37862c2c5efac4a4530902f323a8bf7cc0717de4cdb5aa
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8-operator@sha256:07aa7e257d04ae0eda6fcaa4af9121f1aece439231fce4f924e6467da91f0022
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8-operator@sha256:503fc7a20e2dd7de2cb55b9be5f97efa0d2b31f054ce326f23f3a11b4d683e64
        version: 2.5.0+0.1758147817
      entries:
      - name: aap-operator.v2.5.0-0.1727241223
        version: 2.5.0+0.1727241223
      - name: aap-operator.v2.5.0-0.1727875896
        version: 2.5.0+0.1727875896
      - name: aap-operator.v2.5.0-0.1728520687
        version: 2.5.0+0.1728520687
      - name: aap-operator.v2.5.0-0.1729742145
        version: 2.5.0+0.1729742145
      - name: aap-operator.v2.5.0-0.1731456854
        version: 2.5.0+0.1731456854
      - name: aap-operator.v2.5.0-0.1733193761
        version: 2.5.0+0.1733193761
      - name: aap-operator.v2.5.0-0.1733970210
        version: 2.5.0+0.1733970210
      - name: aap-operator.v2.5.0-0.1736894220
        version: 2.5.0+0.1736894220
      - name: aap-operator.v2.5.0-0.1737101781
        version: 2.5.0+0.1737101781
      - name: aap-operator.v2.5.0-0.1737676792
        version: 2.5.0+0.1737676792
      - name: aap-operator.v2.5.0-0.1739248649
        version: 2.5.0+0.1739248649
      - name: aap-operator.v2.5.0-0.1740094176
        version: 2.5.0+0.1740094176
      - name: aap-operator.v2.5.0-0.1740774104
        version: 2.5.0+0.1740774104
      - name: aap-operator.v2.5.0-0.1741369251
        version: 2.5.0+0.1741369251
      - name: aap-operator.v2.5.0-0.1742434756
        version: 2.5.0+0.1742434756
      - name: aap-operator.v2.5.0-0.1743660958
        version: 2.5.0+0.1743660958
      - name: aap-operator.v2.5.0-0.1746138413
        version: 2.5.0+0.1746138413
      - name: aap-operator.v2.5.0-0.1747345055
        version: 2.5.0+0.1747345055
      - name: aap-operator.v2.5.0-0.1749074612
        version: 2.5.0+0.1749074612
      - name: aap-operator.v2.5.0-0.1749607543
        version: 2.5.0+0.1749607543
      - name: aap-operator.v2.5.0-0.1750901870
        version: 2.5.0+0.1750901870
      - name: aap-operator.v2.5.0-0.1753900674
        version: 2.5.0+0.1753900674
      - name: aap-operator.v2.5.0-0.1755835623
        version: 2.5.0+0.1755835623
      - name: aap-operator.v2.5.0-0.1758147817
        version: 2.5.0+0.1758147817
      name: stable-2.5-cluster-scoped
    - currentCSV: aap-operator.v2.6.0-0.1758684783
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727257740 <2.6.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Automation Gateway - Unify the management of AAP from a central location.
            * Automation Controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private Automation Hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Ansible Lightspeed: Deploy and manage the lifecycle of an Ansible Lightspeed application.
            * Platform Resource Operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## New in AAP 2.6 and 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.


          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-26

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5 and above, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.6

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console.

          **For customers on AAP 2.4:** Change the channel directly from `stable-2.4` to `stable-2.6` to upgrade directly to AAP 2.6, which will be released in October.

          **For customers on AAP 2.5:** Change the channel from `stable-2.5` to `stable-2.6`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.

          After changing to the new 2.6 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place. When upgrading from AAP 2.4 to 2.6, your external authentication configuration in Automation Controller will automatically be migrated to the new unified Platform UI.

          **Database:** AAP 2.6 continues to use PostgreSQL 15.

          #### Create the AnsibleAutomationPlatform custom resource (Required for upgrades from 2.4 only)

          **If upgrading from AAP 2.4:** Once the operator upgrade is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created.

          **If upgrading from AAP 2.5:** This step is not required as your AnsibleAutomationPlatform custom resource already exists.

          Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          Next, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda (lightspeed is disabled by default):


          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: aap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-26/gateway-rhel9-operator@sha256:fa1b102a0690767a1e08935a0e164c1eabb2d4cf9afccd310262ee9185544de7
        - registry.redhat.io/ansible-automation-platform-26/hub-rhel9@sha256:8041b141cd8921aaaa20cce15eab4c6189da9f449d99057189284891c071ac3b
        - registry.redhat.io/ansible-automation-platform-26/gateway-proxy-rhel9@sha256:13284d20c6fb4f2de24ec1715e7abdb5c4b0e4298c5034a7795f494b93634a02
        - registry.redhat.io/rhel9/postgresql-15@sha256:5e225c1646251d2509b3a6e45fad23abfb11dd5c45d3224b59178f90deb38a70
        - registry.redhat.io/ansible-automation-platform-26/controller-rhel9-operator@sha256:863e77e52ab01cec2a71672a224080a59d963d4fc08b9ed6f3cc7950729b1746
        - registry.redhat.io/ansible-automation-platform-26/hub-rhel9-operator@sha256:18114aab6db0d00780fdeaf9efa74ee44cab94824c2fd35f4b9998b6378ad8b3
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-rhel9-operator@sha256:cfbb98e17297134c187673f355342ac5ad60a56d73f4b0c918be2ff244286cc0
        - registry.redhat.io/ansible-automation-platform-26/platform-resource-rhel9-operator@sha256:969bbd033fbba33624329ee9347b3b89fb5a1f64e71892f5c6c7d74ed14a6d39
        - registry.redhat.io/ansible-automation-platform-26/controller-rhel9@sha256:64f5e77a7842965e9c6f7bc7990354b71bb19ffb06ab8ef7793dde956cb6c0f0
        - registry.redhat.io/ansible-automation-platform-26/ee-minimal-rhel9@sha256:261c324cac081e47d559c5fccbd48e7ec8f5bf64e3f27287fbe4fb8aff6b7d99
        - registry.redhat.io/ansible-automation-platform-26/hub-web-rhel9@sha256:aa4041672542a7f828d8a51c79e9c4dcea2a185388e4f20f828104db37e03a9b
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-rhel9@sha256:01e56bee772d15b120d6e4b02a59024b21eb8b266127b1a9b7f2c535ecdb2855
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-ui-rhel9@sha256:5b5f38a5f76781a042eec9cbf70b8be026878e4fa61c7fb90d1b0ad64bd21592
        - registry.redhat.io/ubi9/ubi-minimal@sha256:7c5495d5fad59aaee12abc3cbbd2b283818ee1e814b00dbc7f25bf2d14fa4f0c
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-chatbot-rhel9@sha256:8901fba14b21569b02108d2ab9050d558dee29f8177fae9d1bfcd5de3f4ce7c1
        - registry.redhat.io/ansible-automation-platform-26/mcp-tools-rhel9@sha256:0c1f6f6958f2570c8cd073287555d28c70ef9dded702ec014d3ead8ce572463e
        - registry.redhat.io/ansible-automation-platform-26/platform-resource-runner-rhel9@sha256:9983043386df38a17ff6fffef64dc187691e0b1fefec01eef3af39aa08e532ef
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-rhel9-operator@sha256:bfd55bffd8239e224d4ebba500d8517e8e66915078d6b59244f5e910fc5e10c8
        - registry.redhat.io/ansible-automation-platform-26/gateway-rhel9@sha256:c85dbbf6beaa264afec480dc1f06576aada531cdfd15f247f4db37dc9702be33
        - registry.redhat.io/rhel9/redis-6@sha256:0962d18bb5292c0e524354f319b63287c2298a4a2fd34556c68fa285df381427
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-rhel9@sha256:c5d4e2e76944e0991729e5c3b99f43b7f29ea08ad112e5e6513feb8fb1d899a1
        - registry.redhat.io/ansible-automation-platform-26/ee-supported-rhel9@sha256:2534d07f1ae41af32adb681e283cbaea465cbe0f9d234ed5478a7bfad7676e7f
        version: 2.6.0+0.1758684783
      entries:
      - name: aap-operator.v2.6.0-0.1758684783
        version: 2.6.0+0.1758684783
      name: stable-2.6
    - currentCSV: aap-operator.v2.6.0-0.1758685241
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727259094 <2.6.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Automation Gateway - Unify the management of AAP from a central location.
            * Automation Controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private Automation Hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Ansible Lightspeed: Deploy and manage the lifecycle of an Ansible Lightspeed application.
            * Platform Resource Operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## New in AAP 2.6 and 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.


          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-26

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5 and above, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.6

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console.

          **For customers on AAP 2.4:** Change the channel directly from `stable-2.4` to `stable-2.6` to upgrade directly to AAP 2.6, which will be released in October.

          **For customers on AAP 2.5:** Change the channel from `stable-2.5` to `stable-2.6`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.

          After changing to the new 2.6 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place. When upgrading from AAP 2.4 to 2.6, your external authentication configuration in Automation Controller will automatically be migrated to the new unified Platform UI.

          **Database:** AAP 2.6 continues to use PostgreSQL 15.

          #### Create the AnsibleAutomationPlatform custom resource (Required for upgrades from 2.4 only)

          **If upgrading from AAP 2.4:** Once the operator upgrade is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created.

          **If upgrading from AAP 2.5:** This step is not required as your AnsibleAutomationPlatform custom resource already exists.

          Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          Next, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda (lightspeed is disabled by default):


          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: aap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-26/gateway-rhel9@sha256:c85dbbf6beaa264afec480dc1f06576aada531cdfd15f247f4db37dc9702be33
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-ui-rhel9@sha256:5b5f38a5f76781a042eec9cbf70b8be026878e4fa61c7fb90d1b0ad64bd21592
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-rhel9-operator@sha256:cfbb98e17297134c187673f355342ac5ad60a56d73f4b0c918be2ff244286cc0
        - registry.redhat.io/ansible-automation-platform-26/controller-rhel9@sha256:64f5e77a7842965e9c6f7bc7990354b71bb19ffb06ab8ef7793dde956cb6c0f0
        - registry.redhat.io/ansible-automation-platform-26/ee-supported-rhel9@sha256:2534d07f1ae41af32adb681e283cbaea465cbe0f9d234ed5478a7bfad7676e7f
        - registry.redhat.io/ansible-automation-platform-26/ee-minimal-rhel9@sha256:261c324cac081e47d559c5fccbd48e7ec8f5bf64e3f27287fbe4fb8aff6b7d99
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-rhel9@sha256:c5d4e2e76944e0991729e5c3b99f43b7f29ea08ad112e5e6513feb8fb1d899a1
        - registry.redhat.io/ansible-automation-platform-26/hub-web-rhel9@sha256:aa4041672542a7f828d8a51c79e9c4dcea2a185388e4f20f828104db37e03a9b
        - registry.redhat.io/ansible-automation-platform-26/platform-resource-runner-rhel9@sha256:9983043386df38a17ff6fffef64dc187691e0b1fefec01eef3af39aa08e532ef
        - registry.redhat.io/ansible-automation-platform-26/gateway-rhel9-operator@sha256:fa1b102a0690767a1e08935a0e164c1eabb2d4cf9afccd310262ee9185544de7
        - registry.redhat.io/ansible-automation-platform-26/platform-resource-rhel9-operator@sha256:969bbd033fbba33624329ee9347b3b89fb5a1f64e71892f5c6c7d74ed14a6d39
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-chatbot-rhel9@sha256:8901fba14b21569b02108d2ab9050d558dee29f8177fae9d1bfcd5de3f4ce7c1
        - registry.redhat.io/ansible-automation-platform-26/mcp-tools-rhel9@sha256:0c1f6f6958f2570c8cd073287555d28c70ef9dded702ec014d3ead8ce572463e
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-rhel9-operator@sha256:bfd55bffd8239e224d4ebba500d8517e8e66915078d6b59244f5e910fc5e10c8
        - registry.redhat.io/ansible-automation-platform-26/hub-rhel9-operator@sha256:18114aab6db0d00780fdeaf9efa74ee44cab94824c2fd35f4b9998b6378ad8b3
        - registry.redhat.io/rhel9/postgresql-15@sha256:5e225c1646251d2509b3a6e45fad23abfb11dd5c45d3224b59178f90deb38a70
        - registry.redhat.io/ubi9/ubi-minimal@sha256:7c5495d5fad59aaee12abc3cbbd2b283818ee1e814b00dbc7f25bf2d14fa4f0c
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-rhel9@sha256:01e56bee772d15b120d6e4b02a59024b21eb8b266127b1a9b7f2c535ecdb2855
        - registry.redhat.io/ansible-automation-platform-26/gateway-proxy-rhel9@sha256:13284d20c6fb4f2de24ec1715e7abdb5c4b0e4298c5034a7795f494b93634a02
        - registry.redhat.io/ansible-automation-platform-26/hub-rhel9@sha256:8041b141cd8921aaaa20cce15eab4c6189da9f449d99057189284891c071ac3b
        - registry.redhat.io/rhel9/redis-6@sha256:0962d18bb5292c0e524354f319b63287c2298a4a2fd34556c68fa285df381427
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/ansible-automation-platform-26/controller-rhel9-operator@sha256:863e77e52ab01cec2a71672a224080a59d963d4fc08b9ed6f3cc7950729b1746
        version: 2.6.0+0.1758685241
      entries:
      - name: aap-operator.v2.6.0-0.1758685241
        version: 2.6.0+0.1758685241
      name: stable-2.6-cluster-scoped
    defaultChannel: stable-2.6-cluster-scoped
    packageName: ansible-automation-platform-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Instana
      provider-url: ""
    name: instana-agent-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: instana-agent-operator.v2.1.35
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instana.io/v1",
                "kind": "InstanaAgent",
                "metadata": {
                  "name": "instana-agent",
                  "namespace": "instana-agent"
                },
                "spec": {
                  "agent": {
                    "configuration_yaml": "# You can leave this empty, or use this to configure your instana agent.\n# See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.\n",
                    "endpointHost": "ingress-red-saas.instana.io",
                    "endpointPort": "443",
                    "env": {},
                    "key": "replace-key"
                  },
                  "cluster": {
                    "name": "my-cluster"
                  },
                  "opentelemetry": {
                    "grpc": {
                      "enabled": true
                    },
                    "http": {
                      "enabled": true
                    }
                  },
                  "zone": {
                    "name": "edited-zone"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring,Logging & Tracing,OpenShift Optional
          certified: "false"
          containerImage: icr.io/instana/instana-agent-operator@sha256:fe949ae0ca9cc05ffe3569cd7a6f4fde48f5e3ee64f27e20e0b3102733a6e809
          description: Fully automated Application Performance Monitoring (APM) for
            microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/instana/instana-agent-operator
          support: Instana
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InstanaAgent is the Schema for the agents API
            displayName: Instana Agent
            kind: InstanaAgent
            name: agents.instana.io
            version: v1
          - description: InstanaAgentRemote is the Schema for the agents API
            displayName: Instana Agent Remote
            kind: InstanaAgentRemote
            name: agentsremote.instana.io
            version: v1
        description: |
          # Instana

          Instana is an [APM solution](https://www.instana.com/product-overview/) built for microservices that enables IT Ops to build applications faster and deliver higher quality services by automating monitoring, tracing and root cause analysis. The solution is optimized for [Kubernetes](https://www.instana.com/automatic-kubernetes-monitoring/) and [OpenShift](https://www.instana.com/blog/automatic-root-cause-analysis-for-openshift-applications/).

          ## Instana Agent Operator

          This is the Kubernetes Operator for installing the Instana Agent on Kubernetes or OpenShift.

          ## Prerequisites for OpenShift

          Before the agent will be able to run in OpenShift, you need to perform a couple of extra configuration steps.

          You need to set up a project for the Instana Agent and configure it's permissions.

          The project you create here needs to be the namespace where you create the Instana Agent custom resource that the operator will use to deploy the agent.

          For example, create the `instana-agent` project:

              oc new-project instana-agent

          Then, ensure the `instana-agent` service account is in the privileged security context:

              oc adm policy add-scc-to-user privileged -z instana-agent

          The `instana-agent-remote` service account must be allowed to use any userid:

              oc adm policy add-scc-to-user anyuid -z instana-agent-remote -n instana-agent

          This service account will be created by the operator.

          Now you can proceed with installing the operator for the Instana agent.

          ## Installation and Configuration

          First, install this operator from [OperatorHub.io](https://operatorhub.io/), [OpenShift Container Platform](https://www.openshift.com/), or [OKD](https://www.okd.io/).

          Second, create a custom resource with the agent configuration in the target namespace (for now, this MUST always be the `instana-agent` namespace). The operator will pick up the custom resource and install the Instana agent accordingly.

          The following is a minimal template of the custom resource:

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgent
          metadata:
            name: instana-agent
            namespace: instana-agent
          spec:
            zone:
              name: my-zone # (optional) name of the zone of the host
            cluster:
              name: replace-me # replace with the name of your Kubernetes cluster
            agent:
              key: replace-me # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io # the monitoring ingress endpoint
              endpointPort: "443" # the monitoring ingress endpoint port, wrapped in quotes
              env:
                INSTANA_AGENT_TAGS: example
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://ibm.biz/monitoring-k8s
          ```

          Save the template in a file `instana-agent.yaml` and edit the following values:

          * If your target namespace is not `instana-agent`, replace the `namespace:` accordingly.
          * `agent.key` must be set with your Instana agent key.
          * `agent.endpointHost` must be set with the monitoring ingress endpoint, generally either `saas-us-west-2.instana.io` or `saas-eu-west-1.instana.io`.
          * `agent.endpointPort` must be set with the monitoring ingress port, generally "443" (wrapped in quotes).
          * `zone.name` should be set with the name of the Kubernetes cluster that is be displayed in Instana.

          For advanced configuration, you can edit the contents of the `configuration.yaml` file. View documentation [here](https://ibm.biz/monitoring-k8s).

          Apply the custom resource with `kubectl apply -f instana-agent.yaml`. After some time, you should see `instana-agent` Pods being created on each node of your cluster, and your cluster should show on the infrastructure map on your Instana Web interface.

          In addition to the InstanaAgent custom resource, a remote agent can be deployed. This agent will not monitor workload running inside the cluster, but can deploy a single replica of an agent which can get an independent configuration for running remote monitoring for certain sensors.

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgentRemote
          metadata:
            name: remote-agent
            namespace: "instana-agent"
          spec:
            zone:
              name: edited-zone # name of the zone of the host
            hostname:
              name: use-custom-hostname-for-remote-agent # (optional) give your remote agent pod a custom hostname
            agent:
              key: replace-key # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io
              endpointPort: "443"
              env: {}
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.
          ```

          ## Uninstalling

          In order to uninstall the Instana agent, simply remove the custom resource with `kubectl delete -f instana-agent.yaml`.

          ## Source Code

          The Instana agent operator is an open source project hosted on [https://github.com/instana/instana-agent-operator](https://github.com/instana/instana-agent-operator/).
        displayName: Instana Agent Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - APM
        - Instana
        links:
        - name: GitHub Repository
          url: https://github.com/instana/instana-agent-operator
        maintainers:
        - email: support@instana.com
          name: Instana
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: Instana
        relatedImages:
        - icr.io/instana/instana-agent-operator@sha256:fe949ae0ca9cc05ffe3569cd7a6f4fde48f5e3ee64f27e20e0b3102733a6e809
        - icr.io/instana/agent@sha256:ed71e6c3e5bdb155853a2db8c916460e6a867fb02627e3a332186188cbbd98dd
        version: 2.1.35
      entries:
      - name: instana-agent-operator.v2.1.35
        version: 2.1.35
      - name: instana-agent-operator.v2.1.34
        version: 2.1.34
      - name: instana-agent-operator.v2.1.33
        version: 2.1.33
      - name: instana-agent-operator.v2.1.32
        version: 2.1.32
      - name: instana-agent-operator.v2.1.31
        version: 2.1.31
      - name: instana-agent-operator.v2.1.30
        version: 2.1.30
      - name: instana-agent-operator.v2.1.29
        version: 2.1.29
      - name: instana-agent-operator.v2.1.28
        version: 2.1.28
      - name: instana-agent-operator.v2.1.27
        version: 2.1.27
      - name: instana-agent-operator.v2.1.25
        version: 2.1.25
      - name: instana-agent-operator.v2.1.24
        version: 2.1.24
      - name: instana-agent-operator.v2.1.23
        version: 2.1.23
      - name: instana-agent-operator.v2.1.22
        version: 2.1.22
      - name: instana-agent-operator.v2.1.21
        version: 2.1.21
      - name: instana-agent-operator.v2.1.20
        version: 2.1.20
      - name: instana-agent-operator.v2.1.19
        version: 2.1.19
      - name: instana-agent-operator.v2.1.18
        version: 2.1.18
      - name: instana-agent-operator.v2.1.17
        version: 2.1.17
      - name: instana-agent-operator.v2.1.16
        version: 2.1.16
      - name: instana-agent-operator.v2.1.15
        version: 2.1.15
      - name: instana-agent-operator.v2.1.14
        version: 2.1.14
      - name: instana-agent-operator.v2.1.13
        version: 2.1.13
      - name: instana-agent-operator.v2.1.12
        version: 2.1.12
      - name: instana-agent-operator.v2.1.11
        version: 2.1.11
      - name: instana-agent-operator.v2.1.10
        version: 2.1.10
      - name: instana-agent-operator.v2.1.9
        version: 2.1.9
      - name: instana-agent-operator.v2.1.8
        version: 2.1.8
      - name: instana-agent-operator.v2.1.7
        version: 2.1.7
      - name: instana-agent-operator.v2.1.6
        version: 2.1.6
      - name: instana-agent-operator.v2.1.5
        version: 2.1.5
      - name: instana-agent-operator.v2.1.4
        version: 2.1.4
      - name: instana-agent-operator.v2.1.3
        version: 2.1.3
      - name: instana-agent-operator.v2.1.2
        version: 2.1.2
      - name: instana-agent-operator.v2.1.1
        version: 2.1.1
      - name: instana-agent-operator.v2.1.0
        version: 2.1.0
      - name: instana-agent-operator.v2.0.32
        version: 2.0.32
      - name: instana-agent-operator.v2.0.31
        version: 2.0.31
      - name: instana-agent-operator.v2.0.30
        version: 2.0.30
      - name: instana-agent-operator.v2.0.27
        version: 2.0.27
      - name: instana-agent-operator.v2.0.19
        version: 2.0.19
      - name: instana-agent-operator.v2.0.18
        version: 2.0.18
      - name: instana-agent-operator.v2.0.17
        version: 2.0.17
      - name: instana-agent-operator.v2.0.13
        version: 2.0.13
      - name: instana-agent-operator.v2.0.12
        version: 2.0.12
      - name: instana-agent-operator.v2.0.11
        version: 2.0.11
      - name: instana-agent-operator.v2.0.10
        version: 2.0.10
      - name: instana-agent-operator.v2.0.9
        version: 2.0.9
      - name: instana-agent-operator.v2.0.8
        version: 2.0.8
      - name: instana-agent-operator.v2.0.7
        version: 2.0.7
      - name: instana-agent-operator.v2.0.5
        version: 2.0.5
      - name: instana-agent-operator.v2.0.4
        version: 2.0.4
      name: stable
    defaultChannel: stable
    packageName: instana-agent-operator
    provider:
      name: Instana
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ses-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ses-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ses-controller:1.1.1
          createdAt: "2025-09-19T19:37:25Z"
          description: AWS SES controller is a service controller for managing SES
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConfigurationSet represents the state of an AWS ses ConfigurationSet
              resource.
            displayName: ConfigurationSet
            kind: ConfigurationSet
            name: configurationsets.ses.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Simple Email Service (SES) resources in AWS from within your Kubernetes cluster.

          **About Amazon SES**

          Amazon Simple Email Service (SES) is an email platform that provides an easy, cost-effective way for you to send and receive email using your own email addresses and domains.
          For example, you can send marketing emails such as special offers, transactional emails such as order confirmations, and other types of correspondence such as newsletters. When you use Amazon SES to receive mail, you can develop software solutions such as email autoresponders, email unsubscribe systems, and applications that generate customer support tickets from incoming emails.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SES
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ses
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SES Developer Resources
          url: https://aws.amazon.com/ses/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ses maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ses-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ses-controller.v1.1.1
        version: 1.1.1
      - name: ack-ses-controller.v1.1.0
        version: 1.1.0
      - name: ack-ses-controller.v1.0.13
        version: 1.0.13
      - name: ack-ses-controller.v1.0.12
        version: 1.0.12
      - name: ack-ses-controller.v1.0.11
        version: 1.0.11
      - name: ack-ses-controller.v1.0.10
        version: 1.0.10
      - name: ack-ses-controller.v1.0.9
        version: 1.0.9
      - name: ack-ses-controller.v1.0.8
        version: 1.0.8
      - name: ack-ses-controller.v1.0.7
        version: 1.0.7
      - name: ack-ses-controller.v1.0.6
        version: 1.0.6
      - name: ack-ses-controller.v1.0.5
        version: 1.0.5
      - name: ack-ses-controller.v1.0.3
        version: 1.0.3
      - name: ack-ses-controller.v1.0.2
        version: 1.0.2
      - name: ack-ses-controller.v1.0.1
        version: 1.0.1
      - name: ack-ses-controller.v1.0.0
        version: 1.0.0
      - name: ack-ses-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-ses-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco Networking - Compute Solutions
      provider-url: ""
    name: ucs-ci-solutions-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ucs-ci-solutions-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intersight.cisco.com/v1alpha1",
                "kind": "UCSCISolutions",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ucs-ci-solutions-operator",
                    "app.kubernetes.io/instance": "ucscisolutions-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ucscisolutions",
                    "app.kubernetes.io/part-of": "ucs-ci-solutions-operator"
                  },
                  "name": "ucscisolutions-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: docker.io/pkoppa/ucs-ci-solutions-operator:latest
          createdAt: "2023-07-15T15:19:39Z"
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: UCSCISolutions
            name: ucscisolutions.intersight.cisco.com
            version: v1alpha1
        description: The operator is used to automate the creation of  Pools, Policies
          and Server Profile Templates required to deploy Cisco UCS Converged Infrastructure
          Solution in Intersight.
        displayName: UCS-CI-Solutions-Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ucscisolutions
        - intersight
        - ucs
        - ci
        - flexpod
        - flashstack
        links:
        - name: Ucs Ci Solutions Operator
          url: https://ucs-ci-solutions-operator.domain
        maintainers:
        - email: pkoppa@cisco.com
          name: Paniraj Koppa
        - email: chrobrie@cisco.com
          name: Chris O'Brien
        maturity: alpha
        provider:
          name: Cisco Networking - Compute Solutions
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - docker.io/pkoppa/ucs-ci-solutions-operator:latest
        version: 0.0.1
      entries:
      - name: ucs-ci-solutions-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ucs-ci-solutions-operator
    provider:
      name: Cisco Networking - Compute Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-diagnostics
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-diagnostics.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Diagnostics",
                "metadata": {
                  "name": "diagnostics-sample"
                },
                "spec": {
                  "agentDiagnosticsService": {
                    "port": 8080
                  },
                  "auth": {
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "wapdbDiagnosticsService": {
                    "dumpedTables": "[ \"--regexpr cluster_logs_week_*\", \"current_week_cluster_logs\", \"config_metadata_dbs\", \"config_metadata_dbs_id_seq\", \"config_metadata_tags\", \"config_tags\", \"csv_reports_meta\", \"custom_import_handler_dbs\", \"data_point_db_models\", \"db_test_entries\", \"license_server_metadata\", \"operations\", \"result_file_metadata_dbs\", \"result_metadata_dbs\", \"result_metadata_tags\", \"result_tags\", \"session_meta\", \"session_types\", \"sessions\", \"stat_view_query_data\", \"test_run_configs\", \"test_run_infos\", \"test_runs\", \"wireless_globalsettings\", \"wireless_sessionconfigs\", \"wireless_templates\"]",
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:37:10Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Diagnostics
            name: diagnostics.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Diagnostics Operator
        displayName: Keysight WAP Diagnostics Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Diagnostics
          url: https://keysight-wap-diagnostics.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-diagnostics@sha256:97bf8affa7bf1f3a391f5f036313e7cd8196be2eb8e95debeab06c1b6c964c5e
        version: 0.1.21
      entries:
      - name: keysight-wap-diagnostics.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-diagnostics
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JFrog
      provider-url: ""
    name: openshiftartifactoryha-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: artifactory-ha-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftArtifactoryHa",
                "metadata": {
                  "name": "openshiftartifactoryha"
                },
                "spec": {
                  "artifactory-ha": {
                    "artifactory": {
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE",
                      "node": {
                        "replicaCount": 2,
                        "waitForPrimaryStartup": {
                          "enabled": false
                        }
                      }
                    },
                    "databaseUpgradeReady": true,
                    "database": {
                      "driver": "OVERRIDE",
                      "password": "OVERRIDE",
                      "type": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "nginx": {
                      "http": {
                        "externalPort": 80,
                        "internalPort": 8080
                      },
                      "https": {
                        "externalPort": 443,
                        "internalPort": 8443
                      },
                      "service": {
                        "ssloffload": false
                      },
                      "tlsSecretName": "OVERRIDE"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "waitForDatabase": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/artifactory-operator@sha256:de69bb16052d5ac6fed9882eb65d49446b6fa1cc5ba830dd2a39e882712b1347
          createdAt: 2024-03-23 00:00:00+00:00
          description: JFrog Artifactory Enterprise deploys Artifactory in a high
            availability environment across multiple pods
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Artifactory HA Instances
            displayName: Artifactory HA
            kind: OpenshiftArtifactoryHa
            name: openshiftartifactoryhas.charts.helm.k8s.io
            version: v1
        description: |2+
           ## Breaking change
          Please update to version 1.1.10 of the operator.
          Version 1.1.8 and 1.1.9 have issues. Please skip these versions.
          If you are using operator version < 1.2.0 there is a issue with upgrade where artifactory version is not updated if you update operator version. As workaround you can update the Custom Resource OpenshiftArtifactoryHa which has hardcored image value.After upgrading to version 1.2.0 you can remove the below parameters from your CR so that future upgrade works without workaround. ``` artifactory-ha.artifactory.image.registry artifactory-ha.artifactory.image.repository artifactory-ha.artifactory.image.tag ```
          ## Overview
          Openshift Operator to deploy JFrog Artifactory Enterprise into your Openshift cluster.
          ## Security Context Constraints
          To deploy this helm chart you will need to be a cluster admin w/ access to the anyuid scc and add the operator service account to the anyuid scc.
          ```
          oc adm policy add-scc-to-user anyuid -z artifactory-ha-operator -n openshift-operators
          ```
          Add the service account for the Artifactory chart to deploy successfully:
          ``` oc adm policy add-scc-to-user anyuid -z openshiftartifactoryha-artifactory-ha -n artifactory ```
          ### Usage

          An external DB is required. The operator will not deploy a DB but will require you to specify the configuration values to connect to it.

          Search for JFrog and click JFrog Artifactory Enterprise Operator to install.

          Go to the Installed Operators.

          Wait for the JFrog Artifactory Enterprise Operator to complete the installation.

          Open the Operator and click on the provided API: Artifactory HA.

          Click Create New Instance and provide the following parameters for your DB configuration:

          ```
          DATABASE_TYPE
          DATABASE_DRIVER
          DATABASE_URL
          DATABASE_USER
          DATABASE_PASSWORD
          ```
          Master key and Join key must be supplied. To generate a new key for each run the command below:

          ```
          # Create a key
          export JOIN_KEY=$(openssl rand -hex 32)
          echo ${JOIN_KEY}
          ```

          To use TLS you will need to first create a k8s tls secret to store your .crt and .key file into.
          Then supply the value of this k8s secret into the TLS_SECRET field.
          ``` oc create secret tls my_tls_secret --cert=tls.crt --key=tls.key --namespace=my_namespace ```
          Click Create for Artifactory Enterprise to deploy into OpenShift and connect to it on the external IP exposed by the load balancer.
          ### Create a route
          To expose Artifactory from Openshift we recommend you create a new route in Openshift.
          You can either use the oc command line tool or the Openshift web console to generate a new route.
          Depending upon where you terminate TLS you will need to either specify pass through or edge.
          Command Line (Edge):
          ``` oc create route edge --service=openshiftartifactory-ha --cert=tls.crt --key=tls.key --ca-cert=ca.crt --hostname=www.example.com ```
          Or for more information visit the official Openshift documentation on routes here:
          https://docs.openshift.com/container-platform/4.6/networking/routes/route-configuration.html

        displayName: JFrog Artifactory Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Artifact Repository
          url: https://jfrog.com/artifactory/
        - name: Artifactory Video
          url: https://www.youtube.com/watch?v=r2_A5CPo43U
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: alpha
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/artifactory-operator@sha256:de69bb16052d5ac6fed9882eb65d49446b6fa1cc5ba830dd2a39e882712b1347
        - registry.connect.redhat.com/jfrog/artifactory-pro@sha256:0986ca71419ae628892db11e7ca7da5ac1c60219c144ce9aeb2de6e683a45262
        - registry.redhat.io/rhel8/nginx-120@sha256:87f658744af695273c434fac8246b716728b691a17d5d0c5b89247362e61e45c
        version: 1.3.3
      entries:
      - name: artifactory-ha-operator.v1.3.3
        version: 1.3.3
      name: alpha
    - currentCSV: artifactory-ha-operator.v1.1.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftArtifactoryHa",
                "metadata": {
                  "name": "openshiftartifactoryha"
                },
                "spec": {
                  "artifactory-ha": {
                    "artifactory": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/artifactory-pro",
                        "tag": "7.27.3-1"
                      },
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE",
                      "uid": "1000721030",
                      "node": {
                        "replicaCount": 2,
                        "waitForPrimaryStartup": {
                          "enabled": false
                        }
                      }
                    },
                    "databaseUpgradeReady": true,
                    "database": {
                      "driver": "OVERRIDE",
                      "password": "OVERRIDE",
                      "type": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "initContainerImage": "registry.connect.redhat.com/jfrog/init@sha256:164b41720a37f75c6129c8a28d05cc56f821022313d44ff918cce4a2146d93e8",
                    "nginx": {
                      "uid": "1000720104",
                      "gid": "1000720107",
                      "http": {
                        "externalPort": 80,
                        "internalPort": 8080
                      },
                      "https": {
                        "externalPort": 443,
                        "internalPort": 8443
                      },
                      "image": {
                        "registry": "registry.redhat.io",
                        "repository": "rhel8/nginx-116",
                        "tag": "latest"
                      },
                      "service": {
                        "ssloffload": false
                      },
                      "tlsSecretName": "OVERRIDE"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "waitForDatabase": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/artifactory-operator@sha256:e292553e1b6e792ff19276c7f6565d7ed61dd95289eb176066329954bc93e371
          createdAt: 2020-03-25 00:00:00+00:00
          description: JFrog Artifactory Enterprise deploys Artifactory in a high
            availability environment across multiple pods
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Artifactory HA Instances
            displayName: Artifactory HA
            kind: OpenshiftArtifactoryHa
            name: openshiftartifactoryhas.charts.helm.k8s.io
            version: v1
        description: |2+
           ## Breaking change
          Please update to version 1.1.10 of the operator.
          Version 1.1.8 and 1.1.9 have issues. Please skip these versions.

          ## Overview
          Openshift Operator to deploy JFrog Artifactory Enterprise into your Openshift cluster.
          ## Security Context Constraints
          To deploy this helm chart you will need to be a cluster admin w/ access to the anyuid scc and add the operator service account to the anyuid scc.
          ```
          oc adm policy add-scc-to-user anyuid -z artifactory-ha-operator -n openshift-operators
          ```
          Add the service account for the Artifactory chart to deploy successfully:
          ``` oc adm policy add-scc-to-user anyuid -z openshiftartifactoryha-artifactory-ha -n artifactory ```
          ### Usage

          An external DB is required. The operator will not deploy a DB but will require you to specify the configuration values to connect to it.

          Search for JFrog and click JFrog Artifactory Enterprise Operator to install.

          Go to the Installed Operators.

          Wait for the JFrog Artifactory Enterprise Operator to complete the installation.

          Open the Operator and click on the provided API: Artifactory HA.

          Click Create New Instance and provide the following parameters for your DB configuration:

          ```
          DATABASE_TYPE
          DATABASE_DRIVER
          DATABASE_URL
          DATABASE_USER
          DATABASE_PASSWORD
          ```
          Master key and Join key must be supplied. To generate a new key for each run the command below:

          ```
          # Create a key
          export JOIN_KEY=$(openssl rand -hex 32)
          echo ${JOIN_KEY}
          ```

          To use TLS you will need to first create a k8s tls secret to store your .crt and .key file into.
          Then supply the value of this k8s secret into the TLS_SECRET field.
          ``` oc create secret tls my_tls_secret --cert=tls.crt --key=tls.key --namespace=my_namespace ```
          Click Create for Artifactory Enterprise to deploy into OpenShift and connect to it on the external IP exposed by the load balancer.
          ### Create a route
          To expose Artifactory from Openshift we recommend you create a new route in Openshift.
          You can either use the oc command line tool or the Openshift web console to generate a new route.
          Depending upon where you terminate TLS you will need to either specify pass through or edge.
          Command Line (Edge):
          ``` oc create route edge --service=openshiftartifactory-ha --cert=tls.crt --key=tls.key --ca-cert=ca.crt --hostname=www.example.com ```
          Or for more information visit the official Openshift documentation on routes here:
          https://docs.openshift.com/container-platform/4.6/networking/routes/route-configuration.html

        displayName: JFrog Artifactory Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Artifact Repository
          url: https://jfrog.com/artifactory/
        - name: Artifactory Video
          url: https://www.youtube.com/watch?v=r2_A5CPo43U
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: stable
        provider:
          name: JFrog
        relatedImages:
        - registry.redhat.io/rhel8/nginx-116@sha256:d71d368d6094a5777000a8d99164959ccb97fca355e855e24bd92c13a8df04f6
        - registry.connect.redhat.com/jfrog/init@sha256:164b41720a37f75c6129c8a28d05cc56f821022313d44ff918cce4a2146d93e8
        - registry.connect.redhat.com/jfrog/artifactory-operator@sha256:e292553e1b6e792ff19276c7f6565d7ed61dd95289eb176066329954bc93e371
        - registry.connect.redhat.com/jfrog/artifactory-pro@sha256:7289dde95d1e4d8e161a28cffd9c4a0cc96f92f7536d069ab63ec304a92254fd
        version: 1.1.18
      entries:
      - name: artifactory-ha-operator.v1.1.18
        version: 1.1.18
      name: stable
    defaultChannel: alpha
    packageName: openshiftartifactoryha-operator
    provider:
      name: JFrog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: advanced-cluster-management
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: advanced-cluster-management.v2.10.9
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-08-06T03:17:11Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0 <2.10.9'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/internal-objects: '[]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterHub defines the configuration for an instance
              of the MultiCluster Hub
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:95df18f4479b7f555a41508486624bde714778730921c20c665e9f23baadbabf
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:8f1bb3a8422d29960d8b90dd24a813a56875abf1f0393f78b862eaa48ddda783
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:50331cdd61332c8f4128b3dab237255463b307c5c40b0607730a7d3b33b9c6b5
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:5b162aeab4aa51185c880696d1472e152a20ada926a3baea0fb8494793ad238e
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:0ba2471b7a67532bdb90e02904671f0e724cdb39f74a45392ae6c693c25b354d
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:4b16da3070e30d879421aaa0d4a73aa2a1b0b852f62b8faabd801c03f22d5319
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:ced03229f208739867a43b9ed9bd158f19f4242dd1f5f045d3e5db81b2b1d171
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:4ad5ccf9ce682024b6e3025a9bcadd6f32692aa18d892afd670eacb7b4031860
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:1c2dbcf179508ea1c602f7817f85b0d3f58d6d85ae93ccc523e88a3936c9c690
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:23713ebb12d510d37de7e060d5ce4ff2198990f83cebbd2302e74bd83a55cecd
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:1aaba9ca65e363c7f9c07ef2f66f23cb8fff677204a56373634d319113d2c84b
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:e3c80e8e3af5c02c202dfb288c63b20c44c08f2a0474d2568cfb0b878f7fe888
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:9e78db48ff0b0fc0ac0347ca055a4b66e2f917073499adc6e2db2a9c16a70e01
        - registry.redhat.io/rhacm2/console-rhel9@sha256:1538a3a935aa01d8800d8145c45037eaab6987ad6e6e927017ca1cfea4bb111c
        - registry.redhat.io/rhel9/postgresql-13@sha256:31c77c19480ab82cb33525d32419b68cf59daccf0b9936e88e8238e48a2d47e5
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:fbc5b1456a37951c9f70f64357919222739f4a7e533fb6dde5b93672535fc87e
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:95890bb1aad544d97ae1b80e801a800c95ef51e1034b60e680b67bdb2f5aeb69
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:9e931cab5fcbf3df0ffe6fff8d12303800423ace3d749ee0e33b55300c816b19
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:dcd6a420467929b9af8c52c77c8717d22f3977e6b6b392b5dc90cbb04005b9f1
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:14094f09dfb166a6b342bd91e2db9162f0de88d55655ccce44d345047c700187
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:23b4d7fbc776c63dcd7007442a557124581939ce110257822e8233261d31fb7d
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:bbe57064cb1c40c35929526a082f11b2da6fd3746b6eef8af51ca63a4fd7d98b
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:46e3b8ccf3e7069b44eb3ad407329c03048b5219ef2886e237e1ab0dc3f7bae4
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:bb4b72c1510da84aad2d67516c58f37495d7682c51c8f2088e2c4a350fa36c50
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:dcbc4c5e2d46ec2b4911d8cd9f96a4c141bf7dd18b79aa4c25b77141e4f60b48
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:69a8faa2f4da2281c6c124508634628510d91ad40f725a9b41790e0821d64c14
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:500718f834e05c7a4531eb4f9d8ce560053c5125f76f6921c38c9cf150046a23
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:9db39230c528bfc675bc422a7acd40c394169b9632b088ca16358751994c9cac
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:335c903dde6f21e9c54f025c45666f6b4b6e7d455dfd5f85a6373b3947069f0a
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:3ccc45b0e9483f33be0d3b32a6d46fa794976131ad2f9861d96e319c8e35b9be
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:2f68a16a14f192b2cceaa07ed4840c08fbf413718a39bd584cacde866be7f2b1
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:9c4eb0ef8758d943796e13f65bc92b190db614eabd0fe7375ebfecc90e96e32c
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:da09087ba7d558611d733551b6ebc27ccf441581fc264efe7a5933db3816ef89
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:e02992cec8edcc9374c4cd30dfd7bccbda77d6d5cdd309862ab5265494f1278e
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:f4474ed21a71ad490778a6a378a856bec1884b4b9dc2c0e6c6140cc6c9d03b61
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:805e443fbfca4f816dc2dcea71f4cf723fd753ccea6d13f04d25945f29cb47dc
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:d8b8d130f75647cfeb9036db50d66fc4f1fca56fead6add4ba24c52f9213543e
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:13c33a997be0ddef59d1fb31835751367d4f901a1a6eac41284931bf7988035c
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:814f34f9e55b59d96021d3160b1dad7a68ffe208b3f4a999840b228aa7b3300b
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:47a8fa9541275f849a24e2ef1df87b7e9051dedd0ee8bd833d60defbd8724b46
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:3803d82d2b829c0e0883c0504176a96d2012cb7096cfcace316831a2b55bca3f
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:de68d598587a967d4b3f5e0e28c4675f74fd79cae7a5bb197c1b4b44bb4b31d9
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:35ff861165fccd848b8d328fae46cb8b3ac0c853c63062ec8672a6d41a23f333
        - registry.redhat.io/rhacm2/iam-policy-controller-rhel9@sha256:8c2b356bd9ff4ebfba9059a5f20cad8bd336af237d84cdc21f5c0aa35fabeabe
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:d989e9d2078f8a1e4b18de64f917e6d5e290c5cac769899f829b5f6bbfe6f0f2
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:87093f85c6235c99d4cca9d503b0e4198e3731e5fc3171506485091570e70bb5
        version: 2.10.9
      entries:
      - name: advanced-cluster-management.v2.10.9
        version: 2.10.9
      - name: advanced-cluster-management.v2.10.8
        version: 2.10.8
      - name: advanced-cluster-management.v2.10.7
        version: 2.10.7
      - name: advanced-cluster-management.v2.10.6
        version: 2.10.6
      - name: advanced-cluster-management.v2.10.5
        version: 2.10.5
      - name: advanced-cluster-management.v2.10.4
        version: 2.10.4
      - name: advanced-cluster-management.v2.10.3
        version: 2.10.3
      - name: advanced-cluster-management.v2.10.2
        version: 2.10.2
      - name: advanced-cluster-management.v2.10.1
        version: 2.10.1
      - name: advanced-cluster-management.v2.10.0
        version: 2.10.0
      name: release-2.10
    - currentCSV: advanced-cluster-management.v2.11.8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-08-01T05:25:37Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.10.0 <2.11.8'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:956d549e64c412f86f2083c8ddf97fea269d7806d8826aab93178bec9ed5d7af
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:9d821f4373722d9cb2b679cf5793ef22cb0b9ebcbb5a38060504c93c05c492c4
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:aef5ffbaa578d006bc7caec8fb35446dc263a51813f84c756aa57903b428329b
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:d248a5fad4ae3ee3210eb6dc4f58406d2f1e8f5c390f3c491040b437fe1adeba
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:447b1f6067a82476e14dfbb74a8f324fb86956420e55575f0017b91784f303a5
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:a081f531efb5b54d7d14d9ba0519552ecfc87c3a80d9d3ec4e598d2b445fc689
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:ffd3d9c363d8e15232bd6d12c5f1a77f7cb1ca18f94b7aea79d226568be4586e
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:95890bb1aad544d97ae1b80e801a800c95ef51e1034b60e680b67bdb2f5aeb69
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:f5c56a7381eb80c02a7c140f7c3f0ed08bb93697cd51732ee57498c080d8d3e5
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:8d15be5f40d135fd7069ac768dcf9188d8d8299c65db56df04e66f8a28941966
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:94c89ebd9257797ddc264511399689feb70d56e53332d9894c6d95fbdc51de40
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:6d4e3114494e051b3b875eeeae4bad4d48665fbcbbd3574a98f87b5a5742b41d
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:9be2d97fa3c1140db7d86157503a96ca2d3bdfb57125ab2230c8cf0649ea1e90
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:d4feeb3d3322210765d916ef1f49b47a1527f1c61041f94834c4395319a9ec32
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:49f0ec1aced72e6a2f0c862b0c6c0cd430d997a09ceb9eb0a7d25aaa1d9798b7
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:a9e0533af92ebadca27c9fe93fe4db9c68a1a8bc0060f4d13f67f26eb5691cb3
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:b3e0427341e3b1cbb1fe9166c56184eb0b0ec526d8a3992306e4861f65e4f0c0
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:ded47519fcf30e60cf38d44a101494b0318757969675782365de88b51cab24f0
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:968eb25c5cfcbdcea87b92dbff734d01b5954c13c13f5e794ad1db7bd886e2b4
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:d7552b2a3c864a3fc0d279a2519535e170fd8608195fe1f9abb4b538dfe2a3df
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:d6666c572294c7c85990304d21eaae3040b9c4e47fcb7b446289cb09aabc2d26
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:4a62d8706772ede443cffa005b984c26d137c35a33aafedaa7f01050b618c2a2
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:d4bfef313ff9044ae36aa5f832651a4dd09228d4b9b186803009aafe9774ee2c
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:1e8a26ebfcc33e60f60da4ba8d58cf7c152a99752eac6ebc9e114187e4cc2267
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:4dce3cebeb4ea0f1d3fc1c7e036dc6a3fd09e02f8f2ae4f3a753d9f4f97defe7
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:b713264826124df2c2b3a99dc268a3f3bd7eb24a9010b93c9876d83656e70e1a
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:e36832aa4852ea807d44f2b1f2c402bdaac6b7e2246bf9dd811b41f0b5e8e647
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:eaa719f9d9db8f91eced81e0e164bbc12bb502bf4760fc8c673cc542b4414771
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:45d4404abe02596611b29316bd10985f7a0684b1d305a57d3b24068ac776cf6e
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:251d3ee881c3e902e87f1752648ab084e8780aef79eb88a253ba79f984a01bcf
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:e347cd0f57e0691c61263e6554495c76ce56c7443bce3fd010cf76d068083fee
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:ad10d15efb6833b2daaf153210a080420395a7debfcb873dd757bb924197a654
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:36c90a52b12af24faaf28301d0d1a0c522723bd6cadd7391e5ee1f43f3019eef
        - registry.redhat.io/rhacm2/console-rhel9@sha256:2ed2f4add387caf790b2b41a65a3986e0046e08d0b8406ad17e4f0440ba0ac4a
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:5ba07381111651906ac3c93344e705152eb006826cb36bf5c05bd88723c4c999
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:7d39d0c29b02c399ffdcd83278693a8dd9c5ca3931e660af056030b126d10055
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:eec0e9bd6ea6f146939516da6c7ca9c3755888d96e81954f2f282fae63c1847c
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:827ca1beb707c02908914c517c8ab14b77b1c65ca7690dbbe1bcff600b7551d5
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:4e0c2cf5229fd68102a9e6d0f3155c651b1159b7af7b057151ce52990f747b08
        - registry.redhat.io/rhel9/postgresql-13@sha256:31c77c19480ab82cb33525d32419b68cf59daccf0b9936e88e8238e48a2d47e5
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:ef66846a3f178234618fd1e265067f863bbb56f14cbca0dfc96b0d741885dc9b
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:4b16da3070e30d879421aaa0d4a73aa2a1b0b852f62b8faabd801c03f22d5319
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:028b7258ec88d04397798552ed19a459110660ddbdf3ab375a7a1f74049c455a
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:147c567ce95673ba7151bcd290474756e13018ebc9f7275d61677ffe586a8a76
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:2878b5c41f8d58e986e86803bfd7eb3a05ccf23f7e49ebba56ad30245609b23d
        version: 2.11.8
      entries:
      - name: advanced-cluster-management.v2.11.8
        version: 2.11.8
      - name: advanced-cluster-management.v2.11.7
        version: 2.11.7
      - name: advanced-cluster-management.v2.11.6
        version: 2.11.6
      - name: advanced-cluster-management.v2.11.5
        version: 2.11.5
      - name: advanced-cluster-management.v2.11.4
        version: 2.11.4
      - name: advanced-cluster-management.v2.11.3
        version: 2.11.3
      - name: advanced-cluster-management.v2.11.2
        version: 2.11.2
      - name: advanced-cluster-management.v2.11.1
        version: 2.11.1
      - name: advanced-cluster-management.v2.11.0
        version: 2.11.0
      name: release-2.11
    - currentCSV: advanced-cluster-management.v2.12.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-21T21:16:17Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <2.12.4'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:b83e630e77bd011c374d73d801f10a3230f6798ede10d7f39bf2e8ff6437cf2c
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:3d3260ff206c5ee249443b1f8da49562584f7ee1abd284691b42f1728b6ed5dc
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:b85f3de5231f2f1bc3b4e9c95917f562f39e09dc9b7db35690024163e3a5afe5
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:6f2c4f1abc4388e76414544e65f08238c96fc4d6fbf71fa79330d9dca117d221
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:ec65a5744945ca20f804d94999a4002efab9ddad360a58f9498800bffa9fa0f4
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:187879345174e515863bc1b29a961a0a0d9ad44006a6e12de715b4508dec38a8
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:93df0586f9281fa2879eed0fdf2fb115be355d572cc24dc5c5c9a92137044877
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:58f08a93cd35bd2ac19f94a1779205c81a2fc3a552d1e77b7c659fd51c9dd34f
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:bd24cfabb81b464a31c3d8c21fa6371e46b83094d5c750acc33346203fe15e84
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:c2259b90decf24f8fb1cb058d4751c75fd1558e3c9fd1079857ad9c47ca9e8d2
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:22e262c0353c6f0d5d1b916074836c53437e22697f67c85a43e6d717daf651fa
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:c402a4ac61e11dc9e45fdc7bc7fcb5bf5ca3475a66d782aea60d9f6524e51a24
        - registry.redhat.io/rhel8/postgresql-13@sha256:826a1a75f4186f36217b5c1f1aa270838c4b55b1021d33a4876566bfc3b8e629
        - registry.redhat.io/rhacm2/console-rhel9@sha256:d239df44dafac3bde3eb7ae8c80cb342fa4bb4eabe22f27d2c92d283324e6e10
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:9c297b4c1f9ba17908594f1843ea6de6c5a5e0f3500258c23fbb8b5684249c9e
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:cdaf7be69a5d66cfa4777feb8266421dafd55353d6703e7e591bb9fb23017abc
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:7c548dd869f3458fb70a536708306cb615b9166c7e866cc1215e82f183b39529
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:1419e274b4e2e4c37985208c3b1b9ecd0cdef493427ace3c71ff35ed8fcc394b
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:a5c65dee1f7e4177050244109c10e384899e01d1924b41a540373aa7daf4f274
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:c46657831915ed5f42f4a55e725bc9d17e0671191f4509e610468fda160710c0
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:1e5c12ed0de090198e097cc5deb133ad075fba224f239b8b892e6e1672ec9883
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:fff4fb0924f866b305ad0ecc5d2682da6cb20de15448a2a8c640df182719b4e2
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:28730c8f2344e51b0b9ca603ac63b60885b39be86582943bd1cc200779eee0c5
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:1ccd17b482df844ec42936993ff162f665b5a473ad344270ff2425cd63d428fc
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:6a7c54cb7e6c3cc506ffb8a4f3ec7443ef80798e2c917e20b5f06d39bde02bc4
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:7375991ce3b9e64cb7542dc176414f0974047667d08beac184e2adadd9f83532
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:3ac71b363aef8db127b88e0a63ee616d0d06342142f83522e34c9cac4dd77b2e
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:c4c3bc4f45ae194b884d89319625f2a0970dada823877a35feecd8c733cb2c8d
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:7050f71c3a038fec16db8d0eee69d7d5a4294c043c13e9cf69c3cb2e28897f4c
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:70ddf96707d398d5444bb7bbde6c684275cff62b8a83c7cf6143ff529daf0dee
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:f3a00f55dee152e0345e2616ac14f652dcb4f37f3b9dfca6f049ac5522874c6e
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:f1c371fef66e96199c4a96e21192b47bca423b282d6f99e927644840aaedfd80
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:8ccbbc058d324429aef0c31c99de8c7b3a5abc02ec6f7a14cc60eef63a8f5c69
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:7b64803ddd3b7886ad3ec3de943a38cec5aeb9604732ab7a5ea5061748bf1411
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:5156129064dbe4e9989b41018b1bb250b2eb3f02e0c0afd75a874da9b54ad3c2
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:875e4bf42ad192cfb954e38c8474a529aa938855d69dc0d83c4ee4a4777c1028
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:3faa2318f159d264a9194fc01239d3a2098c184b365b153e3c56e555d5d4992e
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:7968e96b48b2ed63e7657793cf6cc2564055809df19cb43b7e284ac03f2cc26c
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:dca845ad05b02619b81f231eb57dd826720c221d22282fe7422ceebe57191371
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:794ef15591f217be629d6c680c56e84962923993f6dd77514ca5d2bdbd96217b
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:c10a72bf8c0d76b9b4d9d7c0bfca4d83a3f7dd7f6a37b4cc24ee7591638a2482
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:08dd3f6805fa02478671927982a2c77bdec0a5b7e7fd1c9faf20c8af2bb7b428
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:fbfade039739179995828b2ce4f4411b1315ac9a0e4a98e96a115012676a859d
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:b782190b3a285003d319321be09a25ee1376c951cb0cd146cbb634c9b1ca97b3
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:8e62fa4decf01a3beae7080f19469f33822f5010a688ada5db8d2d23b008e498
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:3f286e3c24fe73b4b42e6cf4b6769f91e333014450e1ab135b1e5b4a271cbaec
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:c2ef90b4704251dc9f7ef5288f5bad7aeb4cd2b3706ec531c60f6f4346073a77
        version: 2.12.4
      entries:
      - name: advanced-cluster-management.v2.12.4
        version: 2.12.4
      - name: advanced-cluster-management.v2.12.3
        version: 2.12.3
      - name: advanced-cluster-management.v2.12.2
        version: 2.12.2
      - name: advanced-cluster-management.v2.12.1
        version: 2.12.1
      - name: advanced-cluster-management.v2.12.0
        version: 2.12.0
      name: release-2.12
    - currentCSV: advanced-cluster-management.v2.13.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-09-08T18:37:07Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <2.13.4'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:dc271885425ab74e0554b531a2078133ce02006d0bb9688f3b54a850cfefd8f8
        - registry.redhat.io/rhel8/postgresql-16@sha256:89dfeaefc7e9e974e47a3365e22f354314f8a5e906e93a0152afe4e42f6a82d0
        - registry.redhat.io/rhacm2/console-rhel9@sha256:5526edf70b784bddfc2b7e1c154a44201c346fc20117635500bb54d325921c25
        - registry.redhat.io/rhacm2/acm-flightctl-worker-rhel9@sha256:57f69874b2df77c185374ddcea530af6783410a0f35528ee1a44cc2dee6f7fa2
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:0040791500337f88af7d922895487549915faa3b69dacba1a23540e7d6f2e580
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:ba51f7d077858a0abe32a254970390b50cd7d75433b5661e17f887c48c56a30f
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:65e1f0eb4e6c797c6dc6518674b28370db3fdaa6d7e17f4233c07207224dcb48
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:3028d011e7b88dd9597821128275f606fdba639e5b83e4e6a8783fa2616cba6f
        - registry.redhat.io/rhacm2/acm-flightctl-api-rhel9@sha256:8302b536c896290533f3f50084cb9047cb55af3a4404887b70704d23862619fa
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:d29ceda561f36faa581d76d8b889f7e01d253d33c88709078855ae26cada4917
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:0272215ab43b5688295d94cce0de4490a2fbabc1dd09ed41645be7acef1e3c9b
        - registry.redhat.io/rhacm2/acm-flightctl-ocp-ui-rhel9@sha256:89863ce0eb5fb73272bda8762d74461552f8502fbe262b16cbde01064cc533b3
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:84109be1814719c343b8b4676981cdbcea0356f012b6c854221c271bdb518fea
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:fc16203de469dd1ba16e428a9d7efbb16a5f511d866ef6fcc253530de1f032a6
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:ab1e4f8685896e2016a0b4c4479378058314f365e4f09ef42cba67f3649bb306
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:b3535378919524df259b85587ca83f5620433914418c9941374951c36e0530ae
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:1b7e6e6a4ae36209f98b1539c37c5975a82363b33fe5ac1e3a320e58b99be239
        - registry.redhat.io/rhacm2/acm-flightctl-periodic-rhel9@sha256:2c9b3c3d6cd7e435da337447580dfbfde8c4932308405eeae5d97ed5ec57fbfc
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:588678bea97237d87cd2e55a0e8000e73461cfc421cc91d554a2f1e352b68d14
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:7f742f024c787932932b8a4100a5329e70aad39537e0c20c2a4e245fc6245e25
        - registry.redhat.io/rhel9/redis-7@sha256:848f4298a9465dafb7ce9790e991bd8a11de2558e3a6685e1d7c4a6e0fc5f371
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:ccc9c0524149820d701c39355af89287ad749832053ea55d1585e77347fa0ec0
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:7b15c6b492e5cb2725b9b345aa3bfae4108e2815bc233ac063b81a67c6550fb6
        - registry.redhat.io/rhel8/postgresql-13@sha256:826a1a75f4186f36217b5c1f1aa270838c4b55b1021d33a4876566bfc3b8e629
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:b1620cebc62240f93b12e6937d15760440be119c421c127214974bdfa9d08ebd
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:0d8202cd10170bb0a138d52a1d71e630ff74df466541897ba86ed324a75ea1d8
        - registry.redhat.io/openshift4/ose-cli-artifacts-rhel9@sha256:67ff9f19edac8d623ce9080a8c05cde3da73afd68e7352e4ee5b22a9ccb596bc
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:3490daccc203fade26e7df348c74a907ab0ffc8901f4965606c8227849087c44
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:f74fb6b4d42f1d7431bfc9d2424f953732fed06d2abd8d1fc27b878b4d1322d5
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:ce342242c50c103fc5e4b0126551231977b35bd477eb627fb8f5d2a6510ddaa0
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:e23115b10a92c323b4cdaf0412919c273dbca33f19910120498bd49cb5810ff5
        - registry.redhat.io/rhacm2/acm-flightctl-ui-rhel9@sha256:86800d67ec8ed63e5408947ba6a441e5afc1050b72a86487b9357674ed495a7e
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:f1db9b10481319dc142ca6046e4e3141addb4c0afab482eeaf80643fad9ad039
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:96bdb2e25ae0f381de0c94a09a3cd11f3c0bdc1d8cfb4fd5ced2cd7b24fbe2c2
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:4d068813e3eb704ea7b2d342d5a2ce8fa2ac71793b1b37634c18704ef0a5e8d7
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:5fc35a944895cf7ea6afa5929c45304aaa9f17397b4f1e42cb8c799d9aead9d5
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:f9ba60c2a415a863c6bd6030fb4663913368a9c271bb40b8acfdc294dcf3a7eb
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:8b152051ef2c4c077379e72aa510c0bd613b219254cad66fab0cd3ec0d1d4bf1
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:04bcddae590e2debf35090055dd52ae4bcaac4b369727266be53ac64670203d0
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:49e99f41cbea3faefc3ae93385a10b544ce91df0d6639cfee65760be6ae0df00
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:81cc71faaf20a986d233fee799a7d2b954cd3f43875a2689f76c53920ccf4df0
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:67939d6e4342372820d7af6777f0ffd123e86e4e8186132c3d4892cb717a5e7c
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:6f95308a508499d2d995653d3fd2def94ac5170b4c3f391d32a1e0594250c33a
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:669e9e2178476bc7a4688f3ebf4c20402608bb4b10e4df6863ec821c7e572299
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:ae1f175ffbd3698f2b3036bf9b62b5c9b24729cfb8248345b8bdd1525ccc5bb4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:ab0192a51ee37b7508f3e4db66ffa53b904dcb1f76d9aa1ac0e80447d8ff9e6a
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:5aec572fc18c8ed533620553aba2c676453297ad2440336a5228a5a4f50f5782
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:a0218b9d5f1803f60d9b295d1504a870780895cffe8f1c5b3ce4c3553da5e4f9
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:a2ab651af1c97e54d2c14de78ff554c0f25e8d000af93a74fc7b7ed913f4e338
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:72c6896a5a4dbb7717d76d9eeade96130d97909eee07e8d3cc17af0658b306ab
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:9ef9f1e2010d77211126482c03e8c72dfb7458940dc95ac952ee8dd1dedd2991
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:6ed6f62f5c05a24938335287b785b80c253424e1ee20d83311cd87c6a2090895
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:2da264ca59278f63990e579939371d45afec3bc7d9922a608520e7647a0377ac
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:1393744757874e79d1fb058cb5d3fe7dcbbcd2f187bdcdcdf34e67227c61fbb9
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:296a31e1aad9f6288bdf80fb3191642e3dd404e9e8fcad198016bedbd3abe9b4
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:3cbb9d271c2f23d430a623c83606bbc8a80379c24d13e610886ad51fd21ce1e4
        version: 2.13.4
      entries:
      - name: advanced-cluster-management.v2.13.4
        version: 2.13.4
      - name: advanced-cluster-management.v2.13.3
        version: 2.13.3
      - name: advanced-cluster-management.v2.13.2
        version: 2.13.2
      - name: advanced-cluster-management.v2.13.1
        version: 2.13.1
      - name: advanced-cluster-management.v2.13.0
        version: 2.13.0
      name: release-2.13
    - currentCSV: advanced-cluster-management.v2.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-25T02:10:33Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.13.0 <2.14.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          repository: https://github.com/open-cluster-management/multiclusterhub-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MulticlusterHub defines the configuration
              for an instance of a multicluster hub, a central point for managing multiple
              Kubernetes-based clusters. The deployment of multicluster hub components
              is determined based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n### Where to find more installation
          information\nYou can find additional installation guidance in the [install
          documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:364c7416a9642dfb37f05e2eac4427918959b12b159ac90a9ca310c584d6434a
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:69f0b9a7fb494572adf2795be865d6d6ba4b67db0eed0e9d61f5dcb5e9bf6efe
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:c2ef90b4704251dc9f7ef5288f5bad7aeb4cd2b3706ec531c60f6f4346073a77
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:0e1eba23e0e8454f2dec240704e7ccc922351d58880f72514d3b2bc30c9e995e
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:c6ff6cae5bc6447d0ed24e458fc5d5fc709b2106fe93b143f9546bd962ea7f87
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:ad7f30c61bb11c0d32686e75b01f3b286a1fb06edd8d2ef6a9e2d924fba2b565
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:b489229190f9a3da66bd154e65490c3a35d1999d2db6707759f06eadb02b968d
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:8e6ee423acfafe28c002911e832d47dffc00a5227325bd6a5b1ab359c7eee5f4
        - registry.redhat.io/rhacm2/console-rhel9@sha256:d4c42a203110050a3834be8d5a0eee384ad7e9fbb991da589d75acd15055b8b1
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:0cecc3d87afcd0d63bdeae3ecae37f8a38534fbdc8bad5c548e0d4f4fb0156a4
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:8f4f59e237f366f08367cabbea3eb51e766ad52db0d35c5efc56182c13fc5c09
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:8253a9225f30b0f0a405e2c3293924acbfbf81f3d9c50bc4431fe4052a9fe908
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:55ef2c163b0c0deb3402e39acf37ff6bb05e2a5d268dcae5f8a400128d67c584
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:b7b7581e7743148d19283250d3300d292b96c48ecae96f6e6202a7ca8204e2ab
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:c329cfd004c58a7fcc40f4fbb986399dc2fe08d7922d2b4f210664f58751b1a6
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:143f49d2ce1f58282e276396722ba37165cd1bf2051d537bb4e6208786e5aed6
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:53deacec64238607afda29c27463d72ac3b582c9670e2e77a9c7e48851214795
        - registry.redhat.io/openshift4/ose-cli-rhel9@sha256:a48758b69b9a7451b7e0caea465a181f854b41f1e74ce86c15216ee1bb31fcb9
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:f9a9e10fde9065d650a7dc8eba83349e5db0a21727027464cf144724d0de8cfa
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:935baf67627d5e76c0766d5d2a9e36b398a12535f097749a987dd0c5de788570
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:0252cf95019717f03c7ac72fb034b5bd9ad7aef2da0a92dd21c56cf47f03b812
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:d56b997a9084d56c63f539147052e64ba30258afe91a9ff832f3673cbccbe467
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ee742e7baa8bcbbfccf79abcf15d22347d14f280eafaf22c42cef71051b10fba
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:0ab21d4189669901fe085c7f3a4b4c9778c853f8abd706bbfffcb59942fbc8e6
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:9c79b1153fc927e7582c5346ea0408e03f794eb983597f02a5e816ae961ba86b
        - registry.redhat.io/rhem/flightctl-periodic-rhel9@sha256:ee59f307e468e2b7b38f7c040898d587d441f8fb3d59cb022ae23e4717739e3f
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:2b1b1c707e0b5217c66a96fb82070de47fc0342747081d4be0049c4542944c8a
        - registry.redhat.io/rhel8/postgresql-16@sha256:d4abbd9492347e0aa7e1b0ff967e74d4b6a4edd6fd20d896f2f2ae03c623181c
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:46c5835a4bfc0e4767d0959bc47e939c195be197e9715aae17788d4d0e799d56
        - registry.redhat.io/rhel9/redis-7@sha256:ff31822cba556e40b6b00720839cc56f474fdb9257fb39b7604fafc11761d379
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:0ea80a1039d217e899ea012b77e99a7c5832e5b7e292dca473759c449bf2d1ea
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:e063c54c7182b70d78d24b5026f174436885383980c9b1e3e7664fa2adfad883
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:d550013d385552d6242b56c5b60f15b3ebb488c2033f808de54bcc639abc2f30
        - registry.redhat.io/rhem/flightctl-ui-ocp-rhel9@sha256:973eb3ca3da32c87ba308928a21fbd25514b2f6a6662300d4928479cb554cff3
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:bcb5949ebc31bbf6b45dfd34aae9c80c82a08531e8a3d9f6eef3354f942a5a5d
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:cee29bdb03927dc1866c267232d391ea5ee762b23c6b04a777e7d61d14a61bba
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:0cc23b5063c0d3264ed345dad8453901e4db18cc9bed1656f2799f72de75bd53
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:8151a89ebfebf2e6d336849fdd329b123b7f486938808ee80bf6bd8d44f6985c
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:7ee0aea84bf00b7b42027fa3e970c1f4b279e345a698b4050c9f413eb0383bb1
        - registry.redhat.io/rhem/flightctl-ui-rhel9@sha256:940c0564ced75b7912d3ffc923a1f6bab9aed0f4a52962f7f894d08937b7d37e
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:ce5ff191023b5e14c339b2a03aa1be05a12d024efa5302652065597e630996b6
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:90d399736ca81bb244415efaf05738beff806912a6ac7444a7d4839089b7a081
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:556d737195a4878877c650a93619af1528e52694f2a8e47e13463a603d2f4900
        - registry.redhat.io/rhem/flightctl-cli-artifacts-rhel9@sha256:3491ec932d90112e89f3a4cebe5bdd9ed6a7811f116df43bab8bb5f827129361
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:d9fa60b5b1d6b6c94c5d8b610c9a7b979202de8b46a344e0a65d06f6ef21a7ef
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:40b3e063435eea6e526a57d6bd1ae489ca703e81eeca82dda1c2e6aedec1262c
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:95085fc33a0150a39b6ef4e616fd23f0bddd9dcad0e75c4a89ed27113626e6eb
        - registry.redhat.io/rhel9/memcached@sha256:955138db9a7a23ee31198aabebd0811438fee11bb0b624c38ddb90efe52bd07c
        - registry.redhat.io/rhem/flightctl-worker-rhel9@sha256:1b1f6f85af0910f1acd029f54a608b25d3a2977c3d8b323266391543324305b4
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:f3e34b200e9480363c56da1cd88b2de37fc6a490691e4596a132217f8ea286d8
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:c70242a037e36690d8dd1030254dde778917356b4f1f61db2cecb03b6c2d7380
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:cb5346afe1c7b58eb758cdced94f2c1161c1621033b52f29757efe3a50320a65
        - registry.redhat.io/rhem/flightctl-api-rhel9@sha256:83c6c7819e04aa58d961518367ce6ad3159a28300ca097967f630044d181be34
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:8e57af77cc5dee3aaa83f6ae6e62e232c5fc4fbe4dca8eb755226351963058e3
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:44e305ca15e7c9b50bd7fa8bb391f5ee66906aa26ecc5f612d029aaee5bb4cb8
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:d65aef370a7ed01c0c7cc0ad7f0800359db7863db4ec8845de5b450992f26933
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:ad534f7b0efea45b9f4a0a026c76a2a030cb608acdb835dfea3852332acfcae1
        version: 2.14.0
      entries:
      - name: advanced-cluster-management.v2.14.0
        version: 2.14.0
      name: release-2.14
    defaultChannel: release-2.14
    packageName: advanced-cluster-management
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/component: kmm-hub
      app.kubernetes.io/name: kmm-hub
      app.kubernetes.io/part-of: kmm
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: kernel-module-management-hub
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kernel-module-management-hub.v2.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:888a3acad339cf38a324c718e0cbe1237ca5bf2f4636b01d001752658e4cca1b
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.0.2'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:888a3acad339cf38a324c718e0cbe1237ca5bf2f4636b01d001752658e4cca1b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62c44057a27abd41cf3daf60b8736da59a82dadb65fa6c32bef65d19bd49ea49
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:192e08b121a04ae69aea60c35199e8f5e3c42305e1ebd024a605a371de4ab472
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:94b4ddc45cb34d4ffe794a51b99c868cb709f90cc4eb8b28284d70c526a0a226
        version: 2.0.2
      entries:
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.0
    - currentCSV: kernel-module-management-hub.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:d1c1490d4c20dc2ec31effa1af067676c4d264e3b7d4670fc11c2baf8a400001
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.1.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:14e100e5853444114029116255ebfb7a54e5ed51503ba969740ad48b5ca80f70
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:1a9a19df17b464d1708b2905d03b8726e4480ca1f11aa97a15ff4c049a24ea4b
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:d1c1490d4c20dc2ec31effa1af067676c4d264e3b7d4670fc11c2baf8a400001
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:611556d999ea97f9dbb78fec5963baa02de7a26626d7096ba63b177e3bcf9707
        version: 2.1.1
      entries:
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.1
    - currentCSV: kernel-module-management-hub.v2.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:90c1946b3f577cde0435402b04baa52dc54e5aab3bb84bd834ac026a57453296
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.2.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:b1d98c0f14448ae8eb80821cbec236150c338674aae2a2b37f8d0ab0d5f1430c
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:53fddf0b226eb579c71c86112a8d9b6b1d90a1498d6f6985932b16b8be532f35
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:c3dc9ad4ebc68cd85a074b520903a9310e72a90a73bdf4aa7c42b7ed66b7bcba
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:90c1946b3f577cde0435402b04baa52dc54e5aab3bb84bd834ac026a57453296
        version: 2.2.1
      entries:
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.2
    - currentCSV: kernel-module-management-hub.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:f1d8f3c95f621c3faf41867c5dd4bc13fa2b9b291422d6184e6e42f1d1460259
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.3.0'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:f1d8f3c95f621c3faf41867c5dd4bc13fa2b9b291422d6184e6e42f1d1460259
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:c5005b74402f680d11c9ea682a9a0d3f74b6ab0d01d25298d232f910f1a29c34
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:f9fde334af68220943017a7491130e6a82f15e63522f5bb3931aa7d9b00ccc40
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:06ac089cc89b303e525fd71d6439b958f34ab944a72209f1cc8cb562ecfd1112
        version: 2.3.0
      entries:
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.3
    - currentCSV: kernel-module-management-hub.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm-hub",
                    "app.kubernetes.io/name": "kmm-hub",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:2efb81137ad2981b994c62d460c917749308133a7dcad5b9a2b251e18f3052ec
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        version: 2.4.1
      entries:
      - name: kernel-module-management-hub.v2.4.1
        version: 2.4.1
      - name: kernel-module-management-hub.v2.4.0
        version: 2.4.0
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.4
    - currentCSV: kernel-module-management-hub.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm-hub",
                    "app.kubernetes.io/name": "kmm-hub",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:2efb81137ad2981b994c62d460c917749308133a7dcad5b9a2b251e18f3052ec
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        version: 2.4.1
      entries:
      - name: kernel-module-management-hub.v2.4.1
        version: 2.4.1
      - name: kernel-module-management-hub.v2.4.0
        version: 2.4.0
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: kernel-module-management-hub
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: cert-utils-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-utils-operator.v1.3.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/cert-utils-operator@sha256:0290e7b2800a5aa8eb57b0268908f2a20a93b16403c76df4c61021048161d6ff
          createdAt: "2024-01-23T12:29:04Z"
          description: Set of utilities for TLS certificates
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-utils-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/cert-utils-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Cert utils operator is a set of functionalities around certificates packaged in a [Kubernetes operator](https://github.com/operator-framework/operator-sdk).

          Certificates are assumed to be available in a [secret](https://kubernetes.io/docs/concepts/configuration/secret/) of type `kubernetes.io/tls` (other types of secrets are *ignored* by this operator).
          By convention this type of secrets have three optional entries:

          1. `tls.key`: the private key of the certificate.
          2. `tls.crt`: the actual certificate.
          3. `ca.crt`: the CA bundle that validates the certificate.

          The functionalities are the following:

          1. [Ability to populate route certificates](https://github.com/redhat-cop/cert-utils-operator#Populating-route-certificates)
          2. [Ability to create java keystore and truststore from the certificates](https://github.com/redhat-cop/cert-utils-operator#Creating-java-keystore-and-truststore)
          3. [Ability to show info regarding the certificates](https://github.com/redhat-cop/cert-utils-operator#Showing-info-on-the-certificates)
          4. [Ability to alert when a certificate is about to expire](https://github.com/redhat-cop/cert-utils-operator#Alerting-when-a-certificate-is-about-to-expire)
          5. [Ability to inject ca bundles in ValidatingWebhookConfiguration, MutatingWebhookConfiguration, CustomResourceDefinition object](https://github.com/redhat-cop/cert-utils-operator#CA-injection)

          All these feature are activated via opt-in annotations.
        displayName: Cert Utils Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Cert Utils Operator
          url: https://github.com/redhat-cop/cert-utils-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/cert-utils-operator
        maintainers:
        - email: rspazzoli@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/cert-utils-operator@sha256:0290e7b2800a5aa8eb57b0268908f2a20a93b16403c76df4c61021048161d6ff
        version: 1.3.12
      entries:
      - name: cert-utils-operator.v1.3.12
        version: 1.3.12
      - name: cert-utils-operator.v1.3.11
        version: 1.3.11
      - name: cert-utils-operator.v1.3.10
        version: 1.3.10
      - name: cert-utils-operator.v1.3.9
        version: 1.3.9
      - name: cert-utils-operator.v1.3.8
        version: 1.3.8
      - name: cert-utils-operator.v1.3.7
        version: 1.3.7
      - name: cert-utils-operator.v1.3.6
        version: 1.3.6
      - name: cert-utils-operator.v1.3.5
        version: 1.3.5
      - name: cert-utils-operator.v1.3.4
        version: 1.3.4
      - name: cert-utils-operator.v1.3.3
        version: 1.3.3
      - name: cert-utils-operator.v1.3.2
        version: 1.3.2
      - name: cert-utils-operator.v1.3.1
        version: 1.3.1
      - name: cert-utils-operator.v1.3.0
        version: 1.3.0
      - name: cert-utils-operator.v1.2.0
        version: 1.2.0
      - name: cert-utils-operator.v1.1.0
        version: 1.1.0
      - name: cert-utils-operator.v1.0.7
        version: 1.0.7
      - name: cert-utils-operator.v1.0.6
        version: 1.0.6
      - name: cert-utils-operator.v1.0.5
        version: 1.0.5
      - name: cert-utils-operator.v1.0.4
        version: 1.0.4
      - name: cert-utils-operator.v1.0.3
        version: 1.0.3
      - name: cert-utils-operator.v1.0.2
        version: 1.0.2
      - name: cert-utils-operator.v1.0.1
        version: 1.0.1
      - name: cert-utils-operator.v1.0.0
        version: 1.0.0
      - name: cert-utils-operator.v0.2.2
        version: 0.2.2
      - name: cert-utils-operator.v0.2.1
        version: 0.2.1
      - name: cert-utils-operator.v0.2.0
        version: 0.2.0
      - name: cert-utils-operator.v0.1.1
        version: 0.1.1
      - name: cert-utils-operator.v0.1.0
        version: 0.1.0
      - name: cert-utils-operator.v0.0.3
        version: 0.0.3
      - name: cert-utils-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cert-utils-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://access.redhat.com/products/red-hat-integration
    name: red-hat-camel-k
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: red-hat-camel-k-operator.v1.10.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
          createdAt: "2025-01-08T17:00:43Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
        version: 1.10.10
      entries:
      - name: red-hat-camel-k-operator.v1.10.10
        version: 1.10.10
      - name: red-hat-camel-k-operator.v1.10.9
        version: 1.10.9
      - name: red-hat-camel-k-operator.v1.10.9-0.1743522834.p
        version: 1.10.9+0.1743522834.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1742828111.p
        version: 1.10.9+0.1742828111.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741884247.p
        version: 1.10.9+0.1741884247.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741875993.p
        version: 1.10.9+0.1741875993.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741580860.p
        version: 1.10.9+0.1741580860.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739988175.p
        version: 1.10.9+0.1739988175.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739388437.p
        version: 1.10.9+0.1739388437.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739387844.p
        version: 1.10.9+0.1739387844.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.10.x
    - currentCSV: red-hat-camel-k-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1beta1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:9e4fb5369e0479d4a68f60725e7d12abdc66c552babe59f1358f6e6b0f26357d
          createdAt: "2020-10-06T02:45:00Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:9e4fb5369e0479d4a68f60725e7d12abdc66c552babe59f1358f6e6b0f26357d
        version: 1.4.2
      entries:
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.4.x
    - currentCSV: red-hat-camel-k-operator.v1.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
          createdAt: "2022-08-15T20:17:25Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <=1.6.8'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
        version: 1.6.10
      entries:
      - name: red-hat-camel-k-operator.v1.6.10
        version: 1.6.10
      - name: red-hat-camel-k-operator.v1.6.9
        version: 1.6.9
      name: 1.6.x
    - currentCSV: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:e916f9040d3f4feb26cc1534d3ab2722890b34c7fe146cdce8b6085b47206e22
          createdAt: "2022-11-28T10:43:18Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <1.8.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:e916f9040d3f4feb26cc1534d3ab2722890b34c7fe146cdce8b6085b47206e22
        version: 1.8.2+0.1675913507.p
      entries:
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.8.x
    - currentCSV: red-hat-camel-k-operator.v1.10.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
          createdAt: "2025-01-08T17:00:43Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
        version: 1.10.10
      entries:
      - name: red-hat-camel-k-operator.v1.10.10
        version: 1.10.10
      - name: red-hat-camel-k-operator.v1.10.9
        version: 1.10.9
      - name: red-hat-camel-k-operator.v1.10.9-0.1743522834.p
        version: 1.10.9+0.1743522834.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1742828111.p
        version: 1.10.9+0.1742828111.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741884247.p
        version: 1.10.9+0.1741884247.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741875993.p
        version: 1.10.9+0.1741875993.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741580860.p
        version: 1.10.9+0.1741580860.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739988175.p
        version: 1.10.9+0.1739988175.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739388437.p
        version: 1.10.9+0.1739388437.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739387844.p
        version: 1.10.9+0.1739387844.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: candidate
    - currentCSV: red-hat-camel-k-operator.v1.10.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
          createdAt: "2025-01-08T17:00:43Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
        version: 1.10.10
      entries:
      - name: red-hat-camel-k-operator.v1.10.10
        version: 1.10.10
      - name: red-hat-camel-k-operator.v1.10.9
        version: 1.10.9
      - name: red-hat-camel-k-operator.v1.10.9-0.1743522834.p
        version: 1.10.9+0.1743522834.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1742828111.p
        version: 1.10.9+0.1742828111.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741884247.p
        version: 1.10.9+0.1741884247.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741875993.p
        version: 1.10.9+0.1741875993.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741580860.p
        version: 1.10.9+0.1741580860.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739988175.p
        version: 1.10.9+0.1739988175.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739388437.p
        version: 1.10.9+0.1739388437.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739387844.p
        version: 1.10.9+0.1739387844.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: latest
    - currentCSV: red-hat-camel-k-operator.v1.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
          createdAt: "2022-08-15T20:17:25Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <=1.6.8'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
        version: 1.6.10
      entries:
      - name: red-hat-camel-k-operator.v1.6.10
        version: 1.6.10
      - name: red-hat-camel-k-operator.v1.6.9
        version: 1.6.9
      name: stable
    defaultChannel: 1.10.x
    packageName: red-hat-camel-k
    provider:
      name: Red Hat
      url: https://access.redhat.com/products/red-hat-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.17.40-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5-beta.1 <8.17.40-beta.1'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBM Turbonomic software automates and simultaneously optimizes
              performance, compliance, and cost in real time. Turbonomic discovers
              and automates resource optimization for the complete application stack.
              Applications are continually resourced to perform while satisfying technical
              and business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
          - kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Optimization for Kubernetes
          IBM Turbonomic software automates and simultaneously optimizes performance, compliance, and cost in real time.
          Turbonomic discovers and automates resource optimization for the complete application stack. Applications are continually resourced to perform while satisfying technical and business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (proactive rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          Turbonomic assures application performance by giving workloads the resources they need when they need them.

          ### How Does It Work?
          Turbonomic uses a client agent called — KubeTurbo — that runs in your cluster to discover, monitor and optimize resource usage within your environment.
          KubeTurbo runs together with the default Kubernetes scheduler and sends data back to the Turbonomic for analysis.
          Turbonomic determines the right actions that drive continuous health for your workloads and your cluster.

          Turbonomic discovers Red Hat OpenShift Virtualization resources through a new or existing target configuration with Red Hat OpenShift. For more information refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=targets-red-hat-openshift-virtualization)

          ### Get Started
          Please follow the [Kubeturbo Documentation Page](https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets) for all deployment options and an overview of typical use cases.

          ### Known Issues
          For known issues refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=notes-known-issues)
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Kubeturbo Documentation Page
          url: https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets
        - name: Kubeturbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
        - name: IBM Turbonomic Application Resource Management
          url: https://www.ibm.com/products/turbonomic
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/kubeturbo-operator@sha256:7111c7c4002a447dc6977dbb73e96507bae56766fc0552fb2d036b20583722ba
        version: 8.17.40-beta.1
      entries:
      - name: kubeturbo-operator.v8.17.40-beta.1
        version: 8.17.40-beta.1
      name: beta
    - currentCSV: kubeturbo-operator.v8.17.40
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5 <8.17.40'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBM Turbonomic software automates and simultaneously optimizes
              performance, compliance, and cost in real time. Turbonomic discovers
              and automates resource optimization for the complete application stack.
              Applications are continually resourced to perform while satisfying technical
              and business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
          - kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Optimization for Kubernetes
          IBM Turbonomic software automates and simultaneously optimizes performance, compliance, and cost in real time.
          Turbonomic discovers and automates resource optimization for the complete application stack. Applications are continually resourced to perform while satisfying technical and business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (proactive rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          Turbonomic assures application performance by giving workloads the resources they need when they need them.

          ### How Does It Work?
          Turbonomic uses a client agent called — KubeTurbo — that runs in your cluster to discover, monitor and optimize resource usage within your environment.
          KubeTurbo runs together with the default Kubernetes scheduler and sends data back to the Turbonomic for analysis.
          Turbonomic determines the right actions that drive continuous health for your workloads and your cluster.

          Turbonomic discovers Red Hat OpenShift Virtualization resources through a new or existing target configuration with Red Hat OpenShift. For more information refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=targets-red-hat-openshift-virtualization)

          ### Get Started
          Please follow the [Kubeturbo Documentation Page](https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets) for all deployment options and an overview of typical use cases.

          ### Known Issues
          For known issues refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=notes-known-issues)
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Kubeturbo Documentation Page
          url: https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets
        - name: Kubeturbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
        - name: IBM Turbonomic Application Resource Management
          url: https://www.ibm.com/products/turbonomic
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/kubeturbo-operator@sha256:829dc6b9800efd3fd77d07f955a53610c7b2888b93495b2db32f1f48e93a05a6
        version: 8.17.40
      entries:
      - name: kubeturbo-operator.v8.17.40
        version: 8.17.40
      name: stable
    defaultChannel: stable
    packageName: kubeturbo-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: VASTData
      provider-url: https://www.vastdata.com
    name: vast-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vast-csi-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"storage.vastdata.com/v1","kind":"VastCSIDriver","metadata":{"name":"csi.vastdata.com"},"spec":{"applySecurityContextConstraints":true,"attachRequired":true,"controller":{"dnsPolicy":"Default","extraArgs":{"csiAttacher":[],"csiProvisioner":[],"csiResizer":[],"csiSnapshotter":[]},"nodeAffinity":{},"nodeSelector":{},"podAffinity":{},"podAntiAffinity":{},"resources":{"csiAttacher":{"limits":{"memory":"200Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiProvisioner":{"limits":{"memory":"400Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiResizer":{"limits":{"memory":"300Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiSnapshotter":{"limits":{"memory":"200Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiVastPlugin":{"limits":{"memory":"400Mi"},"requests":{"cpu":"100m","memory":"50Mi"}}},"runOnControlPlane":false,"runOnMaster":false,"tolerations":[]},"deletionViewPolicy":null,"deletionVipPool":null,"dontUseTrashApi":false,"driverType":"nfs","endpoint":null,"image":{"csiAttacher":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-attacher@sha256:b4d611100ece2f9bc980d1cb19c2285b8868da261e3b1ee8f45448ab5512ab94"},"csiNodeDriverRegistrar":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac"},"csiProvisioner":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-provisioner@sha256:405a14e1aa702f7ea133cea459e8395fe40a6125c088c55569e696d48e1bd385"},"csiResizer":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-resizer@sha256:a541e6cc2d8b011bb21b1d4ffec6b090e85270cce6276ee302d86153eec0af43"},"csiSnapshotter":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-snapshotter@sha256:2e04046334baf9be425bb0fa1d04c2d1720d770825eedbdbcdb10d430da4ad8c"},"csiVastPlugin":{"imagePullPolicy":"IfNotPresent","repository":"registry.connect.redhat.com/vastdataorg/csi@sha256:7169f279d0dd7130535d74df5b938ed6b7944bfd4fe52d242ceee5178f2ea156"}},"imagePullSecrets":[],"kubeletPath":"/var/lib/kubelet","logLevel":5,"node":{"dnsPolicy":"Default","nodeAffinity":{},"nodeSelector":{},"podAffinity":{},"podAntiAffinity":{},"propagateHostMountOptions":true,"resources":{"csiVastPlugin":{"limits":{"memory":"400Mi"},"requests":{"cpu":"100m","memory":"50Mi"}},"nodeDriverRegistrar":{"limits":{"memory":"100Mi"},"requests":{"cpu":"10m","memory":"20Mi"}}},"tolerations":[]},"numWorkers":10,"operationRetryIntervalMax":60,"operationRetryIntervalStart":10,"operationTimeout":15,"secretName":null,"sslCertsSecretName":null,"truncateVolumeName":64,"useLocalIpForMount":null,"verifySsl":false}},{"apiVersion":"storage.vastdata.com/v1","kind":"VastStorage","metadata":{"name":"vastdata-filesystem"},"spec":{"allowVolumeExpansion":true,"blockingClones":false,"clusterName":"","createSnapshotClass":true,"driverType":"nfs","ephemeralVolumeNameFormat":"csi:{namespace}:{name}:{id}","fsType":"","hostEncryption":null,"mountOptions":[],"provisioner":"","qosPolicy":"","reclaimPolicy":"Delete","secretName":"","secretNamespace":"","setDefaultStorageClass":false,"snapshotClass":{"deletionPolicy":"Delete","setDefaultSnapshotClass":false,"snapshotNameFormat":"csi:{namespace}:{name}:{id}"},"storagePath":"","subsystem":"","tenantName":"","viewPolicy":"","vipPool":"","vipPoolFQDN":null,"volumeGroup":"","volumeNameFormat":"csi:{namespace}:{name}:{id}"}},{"apiVersion":"storage.vastdata.com/v1","kind":"VastCluster","metadata":{"name":"vastcluster-sample"},"spec":{"endpoint":null,"passphrase":null,"password":null,"sslCert":null,"tenant":null,"token":null,"username":null}}]'
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/vastdataorg/csi-operator@sha256:ba7f30a42e327004ecbc81ff5ccc453bf3046a29bab5785b9f8a9746e5c6a3fd
          createdAt: "2024-07-01T11:59:59Z"
          description: |-
            VAST CSI Operator supports seamless installation and upgrade of VAST Data's CSI Driver.
            The VAST CSI Driver allows container orchestration frameworks such as Kubernetes to dynamically provision, manage and scale volumes on the [VAST Data Platform](https://www.vastdata.com/platform/overview). VAST provides scalable, reliable, and fast persistent storage that can be accessed remotely by any Kubernetes application container.
            ### Installation
            Refer to the VASTData Operator for Kubernetes [official documentation](https://support.vastdata.com/s/document-item?bundleId=vast-csi-driver-2.6-administrator-s-guide&topicId=deploying-vast-csi-driver/deploying-vast-csi-driver-on-openshift-with-vast-csi-operator.html&_LANG=enus).
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: vast-csi
          operators.operatorframework.io/builder: operator-sdk-v1.3.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/vast-data/vast-csi
          support: VastData Infra Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VastCSIDriver is a common specification for VAST CSI Controller
              and VAST CSI Node plugins, typically intended for creation in a single
              instance.
            displayName: VastCSIDriver
            kind: VastCSIDriver
            name: vastcsidrivers.storage.vastdata.com
            version: v1
          - description: VastCluster encapsulates the connection details required
              for accessing a VAST cluster. This information is ultimately stored
              in a Kubernetes Secret.
            displayName: VastCluster
            kind: VastCluster
            name: vastclusters.storage.vastdata.com
            version: v1
          - description: Represents VAST storage. Each VastStorage instance generates
              a corresponding storage class that can be used in PVCs to create workloads.
            displayName: VastStorage
            kind: VastStorage
            name: vaststorages.storage.vastdata.com
            version: v1
        description: |-
          VAST CSI Operator supports seamless installation and upgrade of VAST Data's CSI Driver.
          The VAST CSI Driver allows container orchestration frameworks such as Kubernetes to dynamically provision, manage and scale volumes on the [VAST Data Platform](https://www.vastdata.com/platform/overview). VAST provides scalable, reliable, and fast persistent storage that can be accessed remotely by any Kubernetes application container.
          ### Installation
          Refer to the VASTData Operator for Kubernetes [official documentation](https://support.vastdata.com/s/document-item?bundleId=vast-csi-driver-2.6-administrator-s-guide&topicId=deploying-vast-csi-driver/deploying-vast-csi-driver-on-openshift-with-vast-csi-operator.html&_LANG=enus).
        displayName: VAST CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - VAST
        - VASTData
        - CSI
        - Multi Cluster
        - AI
        - High-Performance Computing
        links:
        - name: VAST CSI Plugin
          url: https://github.com/vast-data/vast-csi
        maintainers:
        - email: volodymyr.boiko@vastdata.com
          name: VAST Infrastructure Team
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: VASTData
          url: https://www.vastdata.com
        relatedImages:
        - registry.k8s.io/sig-storage/csi-attacher@sha256:b4d611100ece2f9bc980d1cb19c2285b8868da261e3b1ee8f45448ab5512ab94
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:405a14e1aa702f7ea133cea459e8395fe40a6125c088c55569e696d48e1bd385
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac
        - registry.connect.redhat.com/vastdataorg/csi@sha256:7169f279d0dd7130535d74df5b938ed6b7944bfd4fe52d242ceee5178f2ea156
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.connect.redhat.com/vastdataorg/csi-operator@sha256:ba7f30a42e327004ecbc81ff5ccc453bf3046a29bab5785b9f8a9746e5c6a3fd
        - registry.k8s.io/sig-storage/csi-resizer@sha256:a541e6cc2d8b011bb21b1d4ffec6b090e85270cce6276ee302d86153eec0af43
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:2e04046334baf9be425bb0fa1d04c2d1720d770825eedbdbcdb10d430da4ad8c
        version: 2.6.3
      entries:
      - name: vast-csi-operator.v2.6.3
        version: 2.6.3
      name: stable
    defaultChannel: stable
    packageName: vast-csi-operator
    provider:
      name: VASTData
      url: https://www.vastdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MySQL
      provider-url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
    name: ndb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ndb-operator.v1.6.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "mysql.oracle.com/v1","kind": "NdbCluster","metadata":
            {"name": "example-ndb"},"spec": {"redundancyLevel": 2,"dataNode": {"nodeCount":
            2},"mysqlNode": {"nodeCount":2}}}]'
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: container-registry.oracle.com/mysql/community-ndb-operator@sha256:b562dc3303ffae9f4dab1d73fd0fe05833bdaa8b731af4227da81af617c75c6b
          createdAt: "2024-12-18T15:05:46.000Z"
          description: The MySQL NDB Operator is a Kubernetes operator for managing
            a MySQL NDB Cluster setup inside a Kubernetes Cluster.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Ndb Cluster
            displayName: NdbCluster
            kind: NdbCluster
            name: ndbclusters.mysql.oracle.com
            version: v1
        description: |
          ## NdbCluster Custom Resource
          The NdbCluster custom resource is defined with custom fields that map to the configuration of a MySQL Cluster. To deploy a MySQL Cluster inside the K8s Cluster, an NdbCluster object with the desired configuration of the MySQL Cluster has to be created in the K8s Cluster. The NDB Operator watches for any changes made to any NdbCluster objects and will receive an event when the NdbCluster object is created in the K8s Cluster. It will then start deploying the MySQL Cluster, with the specified configuration, inside the K8s Cluster using native Kubernetes workloads and resources. Once the MySQL Cluster is ready, the NdbCluster resource can be further modified to make changes to the MySQL Cluster configuration. The Operator will receive this event as well and will update the configuration of the already running MySQL Cluster.

          The documentation at [NdbCluster-CRD.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md) explains the various custom fields of the NdbCluster CRD and the effect they have on the MySQL Cluster configuration. The [examples](https://github.com/mysql/mysql-ndb-operator/tree/main/docs/examples) folder has a few examples of the NdbCluster resource objects which can be used to deploy MySQL Cluster inside the K8s Cluster.
          ## Deploy a simple MySQL Cluster

          The [examples/example-ndb.yaml](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/examples/example-ndb.yaml) file has a NdbCluster object example-ndb that defines a simple MySQL Cluster with 2 data nodes and 2 MySQL Servers. To create this object in the default namespace of the K8s Cluster, run this command from the root of the repo :
          ```
          kubectl apply -f docs/examples/example-ndb.yaml
          ```

          Various NdbCluster status fields report back the actual status of the MySQL Cluster running inside the K8s Cluster. Wait for the MySQL Cluster nodes to be started and become ready using the [UpToDate](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md#ndbclusterconditiontypestring-alias) Condition defined in the NdbCluster object status :
          ```
          kubectl wait --for=condition=UpToDate ndb example-ndb --timeout=10m
          ```
          The MySQL Cluster is ready for transactions once the NdbCluster is UpToDate, i.e. the UpToDate condition becomes true.

          At any point, a brief status of an NdbCluster object can be viewed using kubectl get ndb command.
          ```
          kubectl get ndb example-ndb
          ```

          The output of the command gives a short summary the MySQL Cluster nodes controlled by the respective NdbCluster object.
          ```
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE   UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:2/2    Ready:2/2 3m    True
          ```

          To list all the pods created by the NDB Operator, run :
          ```
          kubectl get pods -l mysql.oracle.com/v1=example-ndb
          ```

          For example-ndb NdbCluster object, two management nodes (*-mgmd-* pods), two multi-threaded data nodes (*-ndbmtd-* pods) and two MySQL Servers (*-mysqld-*) would have been created by the NDB operator :
          ```
          NAME                   READY   STATUS    RESTARTS   AGE
          example-ndb-mgmd-0     1/1     Running   0          3m53s
          example-ndb-mgmd-1     1/1     Running   0          2m5s
          example-ndb-mysqld-0   1/1     Running   0          78s
          example-ndb-mysqld-1   1/1     Running   0          78s
          example-ndb-ndbmtd-0   1/1     Running   0          3m53s
          example-ndb-ndbmtd-1   1/1     Running   0          3m53s
          ```
          If running inside a openshift cluster, pods from statefulset "example-ndb-mysqld" will not start up. By default openshift runs all the pod with "restricted" scc, but the statefulset "example-ndb-mysqld" required higher privilege. Use the below command to check the exact scc required
          ```
          oc get pods example-ndb-mysqld-0 -o yaml | oc adm policy scc-subject-review -f -
          ```
          Create a user account with the scc and set the user to the statefulset "example-ndb-mysqld". Below commands creates an user with "anyuid" and assign it to the statefulset.
          ```
          oc create sa any-uid-user
          oc adm policy add-scc-to-user anyuid -z any-uid-user
          oc set sa sts example-ndb-mysqld any-uid-user
          ```
          ## Connect to the MySQL Cluster

          The NDB Operator, by default, creates few Services to expose the services offered by the MySQL Cluster nodes within the Kubernetes Cluster.
          ```
          kubectl get services -l mysql.oracle.com/v1=example-ndb
          ```

          The above command will generate an output similar to :
          ```
          NAME                 TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S) AGE
          example-ndb-mgmd     ClusterIP   10.100.207.255   <none>        1186/TCP 5m
          example-ndb-mysqld   ClusterIP   10.103.182.157   <none>        3306/TCP 5m
          example-ndb-ndbmtd   ClusterIP   None             <none>        1186/TCP 5m
          ```

          The NDB Operator will also create a root user with a random password on all the MySQL Servers. The root password will be stored in a K8s Secret which can be retrieved from the NdbCluster object status :
          ```
          kubectl get ndb example-ndb -o jsonpath={.status.generatedRootPasswordSecretName}
          ```

          To extract and decode the password from the Secret :
          ```
          base64 -d <<< \
          $(kubectl get secret example-ndb-mysqld-root-password \
          -o jsonpath={.data.password})
          ```

          One can also specify the password and host to be used with the root account via the NdbCluster spec. See the [CRD documentation](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md#ndbmysqldspec) for more details on this.
          ## Access MySQL Cluster from inside K8s

          To connect to the Management and MySQL Servers from within the Kubernetes Cluster, you can use the example-ndb-mgmd and example-ndb-mysqld services as the MySQL Cluster connectstring and MySQL host respectively. You can also use the mgmd and mysqld pods' hostnames to directly connect to that particular pod. The mysql, ndb_mgm clients and any other ndb tools will work as expected.

          A demonstration is available at [connect-from-inside-k8s-demo.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/connect-from-inside-k8s-demo.md).
          ## Access MySQL Cluster from outside K8s

          By default, the Management and MySQL services created by the NDB Operator are of type [ClusterIP](https://kubernetes.io/docs/concepts/services-networking/service/#publishing-services-service-types) and are accessible only from within the K8s Cluster. To expose them outside the K8s Cluster, the services have to be upgraded into a LoadBalancer type. This can be done by setting the spec.managementNode.enableLoadBalancer and spec.mysqlNode.enableLoadBalancer fields of the NdbCluster resource to true. These options can be enabled when the NdbCluster resource object is created or via an update to the object when the MySQL Cluster is already running.

          Once enabled, the previously ClusterIP type example-ndb-mgmd and example-ndb-mysqld services will be upgraded to LoadBalancer type, and they will be available at the external IP assigned to them by the cloud provider.

          Another way to access these services without enabling the LoadBalancers support is to use the kubectl port-forward command.

          In both ways, only the mysql and ndb_mgm clients work. Any NDB tool which uses the NDBAPI to connect to the MySQL data nodes will not work as expected from outside the K8s Cluster.

          A demonstration is available at [connect-from-outside-k8s-demo.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/connect-from-outside-k8s-demo.md).
          ## Updating MySQL Cluster Configuration

          Once a MySQL Cluster has been deployed by the NDB Operator inside the K8s Cluster, its configuration can be updated by updating the spec of the NdbCluster resource object that represents it. The NDB Operator will pick up any changes made to the NdbCluster resource object and will apply the updated configuration to the respective MySQL Cluster. The NDB Operator takes care of updating the MySQL Cluster config file and restarting all the management/data nodes if required. When the update is being handled by the NdbOperator, it will set the UpToDate condition of the NdbCluster resource object to false and will not accept any further updates until it completes the current one.

          The kubectl get ndb example-ndb command will report the UpToDate status of an NdbCluster. It can be used to check if the current update has been completed.
          ```
          $ kubectl get ndb example-ndb
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE   UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:1/2    Ready:2/2 10m   False
          ```

          The UpToDate condition can also be used to wait for the operator to complete the update.
          ```
          kubectl wait --for=condition=UpToDate ndb example-ndb --timeout=10m
          ```

          Once the update is complete, the UpToDate condition will be set back to true by the operator.
          ```
          $ kubectl get ndb example-ndb
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE      UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:2/2    Ready:2/2 10m50s   True
          ```

          ## Delete a MySQL Cluster

          To stop and remove the MySQL Cluster running inside the K8s Cluster, delete the NdbCluster resource object.
          ```
          kubectl delete -f docs/examples/example-ndb.yaml
          ```

          (or)
          ```
          kubectl delete ndb example-ndb
          ```

          This will delete all the data, the pods running the MySQL Cluster nodes and also delete all other associated K8s resources created by the NDB Operator.
        displayName: Ndb Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - ndb
        - in-memory
        links:
        - name: Document
          url: https://dev.mysql.com/doc/ndb-operator/en/
        maintainers:
        - email: paulo.p.silva@oracle.com
          name: paulofasilva
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: MySQL
          url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
        relatedImages:
        - container-registry.oracle.com/mysql/community-ndb-operator@sha256:b562dc3303ffae9f4dab1d73fd0fe05833bdaa8b731af4227da81af617c75c6b
        version: 1.6.2
      entries:
      - name: ndb-operator.v1.6.2
        version: 1.6.2
      - name: ndb-operator.v1.6.1
        version: 1.6.1
      - name: ndb-operator.v1.6.0
        version: 1.6.0
      name: alpha
    defaultChannel: alpha
    packageName: ndb-operator
    provider:
      name: MySQL
      url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: mercury-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mercury-operator.1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:1.0.1",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CustomerCreditRating",
                  "expose": ["http"]
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:1.0.1
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:1.0.1
        version: 1.0.1
      entries:
      - name: mercury-operator.1.0.1
        version: 1.0.1
      name: 1.0.x
    - currentCSV: mercury-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_CREDIT_RATING",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-offer",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-offer-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_OFFER",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-party-routing-profile",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/party-routing-profile-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "PARTY_ROUTING_PROFILE",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-example-sdi": {
                "description": "Example Service Domain Infrastructure"
            },
            {
             "example-customer-credit-rating": {
               "description": "Example Customer Credit Rating Service Domain"
            },
            {
             "example-customer-offer": {
               "description": "Example Customer Offer Service Domain"
            },
            {
             "example-party-routing-profile": {
               "description": "Example Party Routing Profile Service Domain"
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:v1.0.2
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - description: Represents a Kafka cluster
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Integration is the Schema for the Camel K integrations API
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:v1.0.2
        version: 1.0.2
      entries:
      - name: mercury-operator.v1.0.2
        version: 1.0.2
      name: dev
    - currentCSV: mercury-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_CREDIT_RATING",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-offer",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-offer-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_OFFER",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-party-routing-profile",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/party-routing-profile-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "PARTY_ROUTING_PROFILE",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-example-sdi": {
                "description": "Example Service Domain Infrastructure"
            },
            {
             "example-customer-credit-rating": {
               "description": "Example Customer Credit Rating Service Domain"
            },
            {
             "example-customer-offer": {
               "description": "Example Customer Offer Service Domain"
            },
            {
             "example-party-routing-profile": {
               "description": "Example Party Routing Profile Service Domain"
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:v1.0.2
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - description: Represents a Kafka cluster
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Integration is the Schema for the Camel K integrations API
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:v1.0.2
        version: 1.0.2
      entries:
      - name: mercury-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: 1.0.x
    packageName: mercury-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-ati-updates
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-ati-updates.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ati.loadcore.keysight.com/v1alpha1",
                "kind": "Updates",
                "metadata": {
                  "name": "updates-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "installWait": {
                    "enabled": true,
                    "image": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:40:21Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Updates
            name: updates.ati.loadcore.keysight.com
            version: v1alpha1
        description: Keysight ATI Updates Operator
        displayName: Keysight ATI Updates Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Ati Updates
          url: https://keysight-ati-updates.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-ati-updates@sha256:abda76eb50d77e5a95191c9dbb21eefe64af5b794511ceafdfda007c15449c6e
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        version: 0.1.21
      entries:
      - name: keysight-ati-updates.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-ati-updates
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: CyberArk
      provider-url: https://www.cyberark.com/
    name: vcp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vcp-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "installer.venafi.com/v1alpha1",
                "kind": "VenafiInstall",
                "metadata": {
                  "name": "install-venafi-enhanced-issuer"
                },
                "spec": {
                  "certManager": {
                    "install": true
                  },
                  "globals": {
                    "imagePullSecretNames": [
                      "venafi"
                    ]
                  },
                  "venafiConnection": {
                    "install": true
                  },
                  "venafiEnhancedIssuer": {
                    "install": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:9da79fd0941c21e433f22aaaccc43ebbd4646e685aea0050194954d8b4019c05
          createdAt: "2024-05-30T11:35:41Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              VenafiInstall is an object for installing Venafi components onto a Kubernetes
              cluster
            displayName: VenafiInstall
            kind: VenafiInstall
            name: venafiinstalls.installer.venafi.com
            version: v1alpha1
        description: |
          Venafi Control Plane Operator for Red Hat OpenShift simplifies deployment of Venafi's enterprise components and dependencies.

          The enterprise components include:
          - Enterprise cert-manager: An enhanced version of cert-manager designed for application certificate management within your clusters.
          - Venafi Enhanced Issuer: Enables seamless certificate enrollment for your clusters directly from the Venafi Control Plane.
          - Venafi Firefly: A high-performance, lightweight issuer for certificates, centrally managed by the Venafi Control Plane.
          - Venafi Kubernetes Agent: Provides real-time visibility into certificates and Kubernetes resources directly within the Venafi Control Plane.
        displayName: Venafi Control Plane Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - venafi
        - vcp-operator
        links:
        - name: Documentation
          url: https://docs.venafi.cloud/
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Venafi
          url: https://venafi.com
        relatedImages:
        - registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:9da79fd0941c21e433f22aaaccc43ebbd4646e685aea0050194954d8b4019c05
        version: 1.0.0
      entries:
      - name: vcp-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: vcp-operator.v1.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "installer.venafi.com/v1alpha1",
                "kind": "VenafiInstall",
                "metadata": {
                  "name": "install-venafi-enhanced-issuer"
                },
                "spec": {
                  "certManager": {
                    "install": true
                  },
                  "globals": {
                    "imagePullSecretNames": [
                      "venafi"
                    ]
                  },
                  "venafiConnection": {
                    "install": true
                  },
                  "venafiEnhancedIssuer": {
                    "install": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:eb2b06d191dbfbb02332375a7bca6bc010d5410e53b3bc5fb8c320101292408c
          createdAt: "2025-07-22T06:44:30Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              VenafiInstall is an object for installing Venafi components onto a Kubernetes
              cluster
            displayName: VenafiInstall
            kind: VenafiInstall
            name: venafiinstalls.installer.venafi.com
            version: v1alpha1
        description: |
          The operator, formerly known as the Venafi Control Plane Operator, simplifies the deployment of enterprise components part of the CyberArk Certificate Manager for Kubernetes.

          The components include:
          - cert-manager: An enterprise distribution of cert-manager designed for application certificate management within your clusters.
          - Enterprise Issuer: Enables seamless certificate enrollment directly from the Certificate Manager.
          - Workload Identity Manager: A high-performance issuer for certificates, centrally managed by the Certificate Manager.
          - Agent: Provides real-time visibility into cluster resources directly within the Certificate Manager.
        displayName: CyberArk Certificate Manager Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cyberark
        - vcp-operator
        links:
        - name: Documentation
          url: https://docs.venafi.cloud/vaas/k8s-components/c-vcpo-overview/
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:eb2b06d191dbfbb02332375a7bca6bc010d5410e53b3bc5fb8c320101292408c
        version: 1.8.1
      entries:
      - name: vcp-operator.v1.8.1
        version: 1.8.1
      - name: vcp-operator.v1.8.0
        version: 1.8.0
      - name: vcp-operator.v1.7.0
        version: 1.7.0
      - name: vcp-operator.v1.6.0
        version: 1.6.0
      - name: vcp-operator.v1.5.0
        version: 1.5.0
      - name: vcp-operator.v1.4.0
        version: 1.4.0
      - name: vcp-operator.v1.3.0
        version: 1.3.0
      - name: vcp-operator.v1.2.1
        version: 1.2.1
      - name: vcp-operator.v1.2.0
        version: 1.2.0
      - name: vcp-operator.v1.1.1
        version: 1.1.1
      name: stable
    defaultChannel: stable
    packageName: vcp-operator
    provider:
      name: CyberArk
      url: https://www.cyberark.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE Ezmeral
      provider-url: ""
    name: hpe-ezmeral-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-ezmeral-csi-operator.v1.0.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ezmeral.hpe.com/v1",
                "kind": "HPEEzmeralCSIDriver",
                "metadata": {
                  "name": "hpeezmeralcsidriver"
                },
                "spec": {
                  "controllerImage": "registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03",
                  "nodeImage": "registry.connect.redhat.com/maprtech/csi-kdfplugin@sha256:ffd273ee2e608d2ef64fd3b2bcbb02d5f0a6c173ed96a15eaf5549ea45fea8d1",
                  "pullPolicy": "Always"
                }
              },
              {
                "apiVersion": "ezmeral.hpe.com/v1",
                "kind": "HPEEzmeralNFSCSIDriver",
                "metadata": {
                  "name": "hpeezmeralnfscsidriver"
                },
                "spec": {
                  "controllerImage": "registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03",
                  "nodeImage": "registry.connect.redhat.com/maprtech/csi-nfsplugin@sha256:a5c954f62f1395bafe2d68454ac604bca4be97573a3482fcbd8e76b6361cc08a",
                  "pullPolicy": "Always"
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator@sha256:c9c243c51cd760c2c2a66b489807a5e22053fcbeb94fa29afd10d9c91e2d3908
          createdAt: "2025-06-25T11:51:05Z"
          description: A Container Storage Interface (CSI) driver for HPE Ezmeral
            Data Fabric platform. The CSI driver allows you to use HPE Ezmeral Data
            Fabric with your preferred container orchestrator.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mapr/mapr-csi
          support: HPE Ezmeral Data Fabric
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE Ezmeral CSI Driver (FUSE)
            displayName: HPEEzmeralCSIDriver
            kind: HPEEzmeralCSIDriver
            name: hpeezmeralcsidrivers.ezmeral.hpe.com
            version: v1
          - description: HPE Ezmeral CSI Driver (NFS)
            displayName: HPEEzmeralNFSCSIDriver
            kind: HPEEzmeralNFSCSIDriver
            name: hpeezmeralnfscsidrivers.ezmeral.hpe.com
            version: v1
        description: |
          The HPE Ezmeral CSI Operator for Kubernetes packages, deploys, manages, upgrades the HPE Ezmeral CSI Driver on Kubernetes and OpenShift for dynamic provisioning of persistent volumes on HPE Ezmeral Data Fabric platform.
          The HPE Ezmeral CSI Driver for Kubernetes leverages HPE Ezmeral Data Fabric platform to provide scalable and persistent storage for stateful applications. Please refer to our CSI driver [documentation](https://docs.datafabric.hpe.com/home/CSIdriver/csi_overview.html) for full list of supported CSI features.
          ## Installation
          Refer to the HPE Ezmeral CSI Operator for Kubernetes [documentation](https://docs.datafabric.hpe.com/home/CSIdriver/csi_installing_operator.html)
        displayName: HPE Ezmeral Data Fabric CSI Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE Ezmeral
        - HPE Ezmeral Data Fabric
        - Storage
        - FileSystem
        - BlockVolume
        links:
        - name: Documentation
          url: https://docs.datafabric.hpe.com/home/CSIdriver/csi_overview.html
        maintainers:
        - email: ezmeral-csi-support@hpe.com
          name: HPE Ezmeral Data Fabric
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: HPE Ezmeral
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6ba961c2c2a29750c0132fe6dd6fa9f6001010afbc5f19b98add87b31b54bcf6
        - registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator@sha256:c9c243c51cd760c2c2a66b489807a5e22053fcbeb94fa29afd10d9c91e2d3908
        - registry.connect.redhat.com/maprtech/csi-nfsplugin@sha256:a5c954f62f1395bafe2d68454ac604bca4be97573a3482fcbd8e76b6361cc08a
        - registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03
        - registry.connect.redhat.com/maprtech/csi-kdfplugin@sha256:ffd273ee2e608d2ef64fd3b2bcbb02d5f0a6c173ed96a15eaf5549ea45fea8d1
        version: 1.0.18
      entries:
      - name: hpe-ezmeral-csi-operator.v1.0.18
        version: 1.0.18
      name: stable
    defaultChannel: stable
    packageName: hpe-ezmeral-csi-operator
    provider:
      name: HPE Ezmeral
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzo-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzo-operator.v3.1.200
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzo.cambridgesemantics.com/v2",
                "kind": "Anzo",
                "metadata": {
                  "name": "no-persistence"
                },
                "spec": {
                  "jsonActivation": true,
                  "jvmMemory": "4096m",
                  "nodeConfig": {
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "app": "anzo"
                        }
                      },
                      "serviceName": "anzo-no-persistence",
                      "template": {
                        "metadata": {
                          "annotations": {
                            "cluster-autoscaler.kubernetes.io/safe-to-evict": "false"
                          },
                          "labels": {
                            "app": "anzo"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "registry.connect.redhat.com/cambridgesemantics/anzo@sha256:5b192604338f602fdabe55fecefd9eb82400a77c5aa7f04bd8e54aead8644c6d",
                              "imagePullPolicy": "Always",
                              "name": "anzo",
                              "resources": {
                                "limits": {
                                  "cpu": "2.0",
                                  "memory": "8.0Gi"
                                },
                                "requests": {
                                  "cpu": "2.0",
                                  "memory": "8.0Gi"
                                }
                              }
                            }
                          ],
                          "initContainers": [
                            {
                              "name": "anzo-init",
                              "resources": {
                                "limits": {
                                  "cpu": "0.5",
                                  "memory": "2Gi"
                                },
                                "requests": {
                                  "cpu": "0.5",
                                  "memory": "2Gi"
                                }
                              }
                            }
                          ],
                          "securityContext": {
                            "fsGroup": 1001
                          },
                          "serviceAccountName": "anzo-operator"
                        }
                      }
                    }
                  },
                  "role": "AnzoServer"
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/anzo-operator@sha256:4c0c8bf7acbfbb96f15459a20364608e5d0a0933c6326a89d219666ee0d88ca2
          createdAt: "2025-07-09T06:52:12Z"
          description: kubernetes operator for Altair Graph Studio
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Altair Customer Support
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoBackup is the Schema for the anzobackups API
            displayName: Altair Graph Studio Backup
            kind: AnzoBackup
            name: anzobackups.anzobackup.cambridgesemantics.com
            version: v1beta1
          - description: Altair Graph Studio is the Schema for the anzos API
            displayName: Altair Graph Studio Operator
            kind: Anzo
            name: anzos.anzo.cambridgesemantics.com
            version: v2
        description: |-
          The Altair Graph Studio Operator provides the way to install and configure an altair graph studio server and agent setup on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v3.1.2/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/deploy-k8s.htm )

          ### Support

          We offer Support to our customers [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Studio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        links:
        - name: Altair Graph Studio Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v3.1.2/README_openshift.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzo-operator@sha256:4c0c8bf7acbfbb96f15459a20364608e5d0a0933c6326a89d219666ee0d88ca2
        - registry.connect.redhat.com/cambridgesemantics/anzo@sha256:5b192604338f602fdabe55fecefd9eb82400a77c5aa7f04bd8e54aead8644c6d
        version: 3.1.200
      entries:
      - name: anzo-operator.v3.1.200
        version: 3.1.200
      - name: anzo-operator.v3.1.106
        version: 3.1.106
      - name: anzo-operator.v3.1.105
        version: 3.1.105
      name: stable
    defaultChannel: stable
    packageName: anzo-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: ""
    name: power-monitoring-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: power-monitoring-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "Kepler",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "kepler",
                    "app.kubernetes.io/name": "kepler",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "kepler"
                },
                "spec": {
                  "exporter": {
                    "deployment": {
                      "port": 9103
                    },
                    "redfish": null
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: registry.redhat.io/openshift-power-monitoring/power-monitoring-rhel9-operator@sha256:5eec15f6fe9bd4bdb1f5fb8057dc67c448c3711bcc7a0a4760f0d90c07688d1e
          createdAt: "2025-02-14T06:23:56Z"
          description: Deploys and Manages Kepler on OpenShift
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/internal-objects: |-
            [
              "keplerinternals.kepler.system.sustainable.computing.io"
            ]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift-power-monitoring/power-monitoring-operator
          support: Team Power Monitoring (team-powermon@redhat.com)
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KeplerInternal is the Schema for internal/unsupported API
            displayName: KeplerInternal
            kind: KeplerInternal
            name: keplerinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: Kepler is the Schema for the keplers API
            displayName: Kepler
            kind: Kepler
            name: keplers.kepler.system.sustainable.computing.io
            version: v1alpha1
        description: |
          ### Introduction
          Kepler (Kubernetes Efficient Power Level Exporter) uses eBPF to probe
          energy-related system stats and exports them as Prometheus metrics.

          ### NOTE
          To view the metrics exposed by Kepler, ensure that
          [User Project Monitoring](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/monitoring/configuring-user-workload-monitoring)
          is enabled in your cluster.

          ### Documentation
          Documentation and installation guide can be found below:
            * [Installation Guide](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/power_monitoring/installing-power-monitoring)
        displayName: Power monitoring for Red Hat OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kepler
        - power
        - energy
        - monitoring
        links:
        - name: Installation Guide
          url: https://docs.redhat.com/documentation/openshift_container_platform/latest/html/power_monitoring/installing-power-monitoring
        maintainers:
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: vimalkum@redhat.com
          name: Vimal Kumar
        - email: vprashar@redhat.com
          name: Vibhu Prashar
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift-power-monitoring/power-monitoring-rhel9-operator@sha256:5eec15f6fe9bd4bdb1f5fb8057dc67c448c3711bcc7a0a4760f0d90c07688d1e
        - registry.redhat.io/openshift-power-monitoring/kepler-rhel9@sha256:27d3cb0a98f1cffd649745e3886ad92995d83a0de6823ddc2d770bf036e2eeaf
        version: 0.4.0
      entries:
      - name: power-monitoring-operator.v0.4.0
        version: 0.4.0
      - name: power-monitoring-operator.v0.3.0
        version: 0.3.0
      - name: power-monitoring-operator.v0.2.0
        version: 0.2.0
      - name: power-monitoring-operator.v0.1.0
        version: 0.1.0
      name: tech-preview
    defaultChannel: tech-preview
    packageName: power-monitoring-operator
    provider:
      name: Red Hat, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtc-operator.v1.7.19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigrationController",
                  "metadata": {
                    "name": "migration-controller",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "azure_resource_group": "",
                    "cluster_name": "host",
                    "migration_velero": true,
                    "migration_controller": true,
                    "migration_ui": true,
                    "migration_log_reader": true,
                    "olm_managed": true,
                    "restic_timeout": "1h",
                    "mig_pv_limit": "100",
                    "mig_pod_limit": "100",
                    "mig_namespace_limit": "10"
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageMigration",
                  "metadata": {
                    "name": "directimagemigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageStreamMigration",
                  "metadata": {
                    "name": "directimagestreammigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigAnalytic",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "plan": {
                      "name": "foo",
                      "namespace": "openshift-migration"
                    },
                    "analyzeK8SResources": true,
                    "analyzeImageCount": true,
                    "analyzePVCapacity": true
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigCluster",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectVolumeMigration",
                  "metadata": {
                    "name": "direct",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                      "kind": "DirectVolumeMigrationProgress",
                      "metadata": {
                        "name": "direct-progress",
                        "namespace": "openshift-migration"
                      },
                      "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigPlan",
                  "metadata": {
                    "name": "migplan",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigMigration",
                  "metadata": {
                    "name": "migmigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigStorage",
                  "metadata": {
                    "name": "migstorage",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:89ba5ccb1cd9c44b8c8d2189f1fc6b04e780d46b19ee8a22e20c0b8c9d94d91f
          createdAt: "2025-04-28T19:55:38Z"
          description: Facilitates migration of container workloads from OpenShift
            3.x to OpenShift 4.x
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.5.0 <1.7.19'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-migration
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:3b6168c6df0d86d76ae4de91755c631b7f6a1f84286257d18c97d5a6af312291
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/mig-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: a request to migrate PVCs from a set of namespaces directly
              to a target
            displayName: DirectVolumeMigration
            kind: DirectVolumeMigration
            name: directvolumemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate watch the progress of rsync pod migrating
              a PVC to a target
            displayName: DirectVolumeMigrationProgress
            kind: DirectVolumeMigrationProgress
            name: directvolumemigrationprogresses.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a set of namespaces
              directly to a target
            displayName: DirectImageMigration
            kind: DirectImageMigration
            name: directimagemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a single imagestream
              directly to a target
            displayName: DirectImageStreamMigration
            kind: DirectImageStreamMigration
            name: directimagestreammigrations.migration.openshift.io
            version: v1alpha1
          - description: A request to Analyze a MigPlan and its namespaces
            displayName: MigAnalytic
            kind: MigAnalytic
            name: miganalytics.migration.openshift.io
            version: v1alpha1
          - description: A cluster defined for migration
            displayName: MigCluster
            kind: MigCluster
            name: migclusters.migration.openshift.io
            version: v1alpha1
          - description: A migration process
            displayName: MigMigration
            kind: MigMigration
            name: migmigrations.migration.openshift.io
            version: v1alpha1
          - description: A migration plan
            displayName: MigPlan
            kind: MigPlan
            name: migplans.migration.openshift.io
            version: v1alpha1
          - description: A migration contoller deployment
            displayName: MigrationController
            kind: MigrationController
            name: migrationcontrollers.migration.openshift.io
            version: v1alpha1
          - description: A migration hook
            displayName: MigHook
            kind: MigHook
            name: mighooks.migration.openshift.io
            version: v1alpha1
          - description: A migration storage location
            displayName: MigStorage
            kind: MigStorage
            name: migstorages.migration.openshift.io
            version: v1alpha1
        description: |-
          The Migration Toolkit for Containers Operator enables installation of the application migration tool components.

          After you have installed the Operator, you can install components by creating a MigrationController CR.

          By default, the Operator installs the following components on an OpenShift Container Platform 4 target cluster:

          * velero and restic, to back up data on the source cluster and restore it on the target cluster
          * mig-controller, to coordinate the backup and restore processes
          * mig-ui, the web console for managing migrations

          If you experience any issues or have feature requests, please file an [issue](https://github.com/konveyor/mig-operator/issues)

          For more information, see the [OpenShift Container Platform Migration Guide](https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html).
        displayName: Migration Toolkit for Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - crane
        - mtc
        links:
        - name: Migrations
          url: https://github.com/konveyor/mig-controller
        - name: Documentation
          url: https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html
        - name: Crane Operator
          url: https://github.com/konveyor/mig-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhmtc/openshift-migration-hook-runner-rhel8@sha256:c8c932a486d9b83e5a93290058d9e84691b87cd4161f8132a1bb6de9a360605e
        - registry.redhat.io/rhmtc/openshift-migration-velero-restic-restore-helper-rhel8@sha256:932216fd08d8ad97baf48ac7b8c989806d30a191337c650df37da537c153dc17
        - registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:3b6168c6df0d86d76ae4de91755c631b7f6a1f84286257d18c97d5a6af312291
        - registry.redhat.io/rhmtc/openshift-migration-velero-rhel8@sha256:bc010bc31e8b553f86aacbca39e28a95e74994c56e85c9fc1b8b80d0791ecb16
        - registry.redhat.io/rhmtc/openshift-migration-log-reader-rhel8@sha256:a0c8ab3a5bbc78c158c2ccccb1d06430a20f40358649196aa4e7311e54f9bdf8
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-aws-rhel8@sha256:023465788745625ad3b4dc0574e35b93d6ac8eea17be2ad1fc88959149a0df2c
        - registry.redhat.io/rhmtc/openshift-migration-ui-rhel8@sha256:66b288170dca8007382d13dff3f2e1868e409fb8002e04ec1fc5c480f868d9b6
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-gcp-rhel8@sha256:cc70c5a63af77948760bc42bf695a5e4eb0bce6285d6b350bba89a53f4b4e7e9
        - registry.redhat.io/rhmtc/openshift-migration-registry-rhel8@sha256:49485dfe708fe1294dec3f0bf34fedc38845a22f91915bba523d728a586b0c30
        - registry.redhat.io/rhmtc/openshift-migration-controller-rhel8@sha256:9b1490b04fc235191515fcd07d7af2e81f846d865016831e17872bb82bf4eaf5
        - registry.redhat.io/rhmtc/openshift-migration-rsync-transfer-rhel8@sha256:23a2bc5ad0f99fc0e7517a0172fdf3597fa619543b0192757983219a578f4ee2
        - registry.redhat.io/rhmtc/openshift-velero-plugin-rhel8@sha256:02698c655a01df2237aee6c570db96f13c0018cee2ae58b96b9570d93c354617
        - registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:89ba5ccb1cd9c44b8c8d2189f1fc6b04e780d46b19ee8a22e20c0b8c9d94d91f
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-microsoft-azure-rhel8@sha256:dad8fa4dbce3b4562bf06756fd74c2007812ec4edbc308e4517ff123c5519767
        version: 1.7.19
      entries:
      - name: mtc-operator.v1.7.19
        version: 1.7.19
      - name: mtc-operator.v1.7.18
        version: 1.7.18
      - name: mtc-operator.v1.7.17
        version: 1.7.17
      - name: mtc-operator.v1.7.16
        version: 1.7.16
      - name: mtc-operator.v1.7.15
        version: 1.7.15
      - name: mtc-operator.v1.7.14
        version: 1.7.14
      - name: mtc-operator.v1.7.13
        version: 1.7.13
      - name: mtc-operator.v1.7.12
        version: 1.7.12
      - name: mtc-operator.v1.7.11
        version: 1.7.11
      - name: mtc-operator.v1.7.10
        version: 1.7.10
      - name: mtc-operator.v1.7.9
        version: 1.7.9
      - name: mtc-operator.v1.7.8
        version: 1.7.8
      - name: mtc-operator.v1.7.7
        version: 1.7.7
      - name: mtc-operator.v1.7.6
        version: 1.7.6
      - name: mtc-operator.v1.7.5
        version: 1.7.5
      - name: mtc-operator.v1.7.4
        version: 1.7.4
      - name: mtc-operator.v1.7.3
        version: 1.7.3
      - name: mtc-operator.v1.7.2
        version: 1.7.2
      - name: mtc-operator.v1.7.1
        version: 1.7.1
      - name: mtc-operator.v1.7.0
        version: 1.7.0
      name: release-v1.7
    - currentCSV: mtc-operator.v1.8.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigrationController",
                  "metadata": {
                    "name": "migration-controller",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "azure_resource_group": "",
                    "cluster_name": "host",
                    "migration_velero": true,
                    "migration_controller": true,
                    "migration_ui": true,
                    "migration_log_reader": true,
                    "olm_managed": true,
                    "restic_timeout": "1h",
                    "mig_pv_limit": "100",
                    "mig_pod_limit": "100",
                    "mig_namespace_limit": "10"
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageMigration",
                  "metadata": {
                    "name": "directimagemigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageStreamMigration",
                  "metadata": {
                    "name": "directimagestreammigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigAnalytic",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "plan": {
                      "name": "foo",
                      "namespace": "openshift-migration"
                    },
                    "analyzeK8SResources": true,
                    "analyzeImageCount": true,
                    "analyzePVCapacity": true
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigCluster",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectVolumeMigration",
                  "metadata": {
                    "name": "direct",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                      "kind": "DirectVolumeMigrationProgress",
                      "metadata": {
                        "name": "direct-progress",
                        "namespace": "openshift-migration"
                      },
                      "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigPlan",
                  "metadata": {
                    "name": "migplan",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigMigration",
                  "metadata": {
                    "name": "migmigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigStorage",
                  "metadata": {
                    "name": "migstorage",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:8a44fe2aea3e842dcccb1e20df4818f4ca573fe13f51427ae8b603d0000a9411
          createdAt: "2025-08-22T15:27:29Z"
          description: Facilitates migration of container workloads from OpenShift
            3.x to OpenShift 4.x
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.5.0 <1.8.10'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-migration
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:f9492269b75608057fbc5edd53603807f1f80cd8581dea12032c88e04cf8f4cb
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/mig-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: a request to migrate PVCs from a set of namespaces directly
              to a target
            displayName: DirectVolumeMigration
            kind: DirectVolumeMigration
            name: directvolumemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate watch the progress of rsync pod migrating
              a PVC to a target
            displayName: DirectVolumeMigrationProgress
            kind: DirectVolumeMigrationProgress
            name: directvolumemigrationprogresses.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a set of namespaces
              directly to a target
            displayName: DirectImageMigration
            kind: DirectImageMigration
            name: directimagemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a single imagestream
              directly to a target
            displayName: DirectImageStreamMigration
            kind: DirectImageStreamMigration
            name: directimagestreammigrations.migration.openshift.io
            version: v1alpha1
          - description: A request to Analyze a MigPlan and its namespaces
            displayName: MigAnalytic
            kind: MigAnalytic
            name: miganalytics.migration.openshift.io
            version: v1alpha1
          - description: A cluster defined for migration
            displayName: MigCluster
            kind: MigCluster
            name: migclusters.migration.openshift.io
            version: v1alpha1
          - description: A migration process
            displayName: MigMigration
            kind: MigMigration
            name: migmigrations.migration.openshift.io
            version: v1alpha1
          - description: A migration plan
            displayName: MigPlan
            kind: MigPlan
            name: migplans.migration.openshift.io
            version: v1alpha1
          - description: A migration contoller deployment
            displayName: MigrationController
            kind: MigrationController
            name: migrationcontrollers.migration.openshift.io
            version: v1alpha1
          - description: A migration hook
            displayName: MigHook
            kind: MigHook
            name: mighooks.migration.openshift.io
            version: v1alpha1
          - description: A migration storage location
            displayName: MigStorage
            kind: MigStorage
            name: migstorages.migration.openshift.io
            version: v1alpha1
        description: |-
          The Migration Toolkit for Containers Operator enables installation of the application migration tool components.

          After you have installed the Operator, you can install components by creating a MigrationController CR.

          By default, the Operator installs the following components on an OpenShift Container Platform 4 target cluster:

          * velero and restic, to back up data on the source cluster and restore it on the target cluster
          * mig-controller, to coordinate the backup and restore processes
          * mig-ui, the web console for managing migrations

          If you experience any issues or have feature requests, please file an [issue](https://github.com/konveyor/mig-operator/issues)

          For more information, see the [OpenShift Container Platform Migration Guide](https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html).
        displayName: Migration Toolkit for Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - crane
        - mtc
        links:
        - name: Migrations
          url: https://github.com/konveyor/mig-controller
        - name: Documentation
          url: https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html
        - name: Crane Operator
          url: https://github.com/konveyor/mig-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhmtc/openshift-migration-registry-rhel8@sha256:accef2a0dd69b254bd120232e5dc86a56aa8f25bcfdba3f9bab7e4127c157762
        - registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:f9492269b75608057fbc5edd53603807f1f80cd8581dea12032c88e04cf8f4cb
        - registry.redhat.io/rhmtc/openshift-migration-controller-rhel8@sha256:c776387c662e21aa61378d0136b2495826e1c157548f9ca9c95d044af38ee169
        - registry.redhat.io/rhmtc/openshift-migration-hook-runner-rhel8@sha256:adaa4f1afefb4d5853713821cb6e7de0d9f5a640f0243f20b41a44be84707caf
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-mtc-rhel8@sha256:5f46ba04146438f2abd6905468d2ecbeb19374a83771d364904ac2ee6afdbefa
        - registry.redhat.io/rhmtc/openshift-migration-openvpn-rhel8@sha256:d21b99d8edefa7a1bab30c85969ec7d7560d06be308bd787e698e409d54cbc19
        - registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:8a44fe2aea3e842dcccb1e20df4818f4ca573fe13f51427ae8b603d0000a9411
        - registry.redhat.io/rhmtc/openshift-migration-log-reader-rhel8@sha256:4f76bcf3df3bb7073b8aabd3da072cacfec05ecf4bfb47e23340074756f7fc30
        - registry.redhat.io/rhmtc/openshift-migration-rsync-transfer-rhel8@sha256:442e7c7f2cc075bf468403ace8846c96ad2fe43eb7f6562d566caa24e4f86873
        - registry.redhat.io/rhmtc/openshift-migration-ui-rhel8@sha256:86e747e0330b0156379d273d0c29c13a65dafbf4f536656723e2bc746872286c
        version: 1.8.10
      entries:
      - name: mtc-operator.v1.8.10
        version: 1.8.10
      - name: mtc-operator.v1.8.9
        version: 1.8.9
      - name: mtc-operator.v1.8.8
        version: 1.8.8
      - name: mtc-operator.v1.8.7
        version: 1.8.7
      - name: mtc-operator.v1.8.6
        version: 1.8.6
      - name: mtc-operator.v1.8.5
        version: 1.8.5
      - name: mtc-operator.v1.8.4
        version: 1.8.4
      - name: mtc-operator.v1.8.3
        version: 1.8.3
      - name: mtc-operator.v1.8.2
        version: 1.8.2
      - name: mtc-operator.v1.8.1
        version: 1.8.1
      - name: mtc-operator.v1.8.0
        version: 1.8.0
      name: release-v1.8
    defaultChannel: release-v1.8
    packageName: mtc-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Performance
      provider-url: ""
    name: cluster-impairment-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-impairment-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "all-impairments"
                },
                "spec": {
                  "duration": 480,
                  "egress": {
                    "bandwidth": 10000,
                    "latency": 50,
                    "loss": 0.02
                  },
                  "ingress": {
                    "bandwidth": 10000,
                    "latency": 50,
                    "loss": 0.02
                  },
                  "interfaces": [
                    "ens2f0"
                  ],
                  "linkFlapping": {
                    "downTime": 30,
                    "enable": true,
                    "upTime": 120
                  },
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "duplication"
                },
                "spec": {
                  "duration": 60,
                  "egress": {
                    "duplication": 25
                  },
                  "ingress": {
                    "duplication": 25
                  },
                  "interfaces": [
                    "ens3"
                  ],
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "two-min-flap"
                },
                "spec": {
                  "duration": 480,
                  "interfaces": [
                    "ens2f0"
                  ],
                  "linkFlapping": {
                    "downTime": 120,
                    "enable": true,
                    "upTime": 120
                  },
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "typical-scenario"
                },
                "spec": {
                  "duration": 30,
                  "egress": {
                    "latency": 50
                  },
                  "ingress": {
                    "latency": 50
                  },
                  "interfaces": [
                    "ens2f0",
                    "eno1"
                  ],
                  "loss": 0.02,
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "uneven-latency"
                },
                "spec": {
                  "duration": 60,
                  "egress": {
                    "latency": 100
                  },
                  "ingress": {
                    "latency": 10
                  },
                  "interfaces": [
                    "ens2f0"
                  ],
                  "nodeSelector": {
                    "key": "node-role.kubernetes.io/worker",
                    "value": ""
                  },
                  "startDelay": 5
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Developer Tools
          containerImage: quay.io/redhat-performance/cluster-impairment-operator:1.0.4
          createdAt: "2021-08-11 12:15:00"
          description: Run Impairments on Kubernetes/Openshift Nodes
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/redhat-performance/cluster-impairment-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterImpairment
            name: clusterimpairments.apps.redhat.com
            version: v1alpha1
        description: "cluster-impairment-operator is an operator designed to automate
          the application of impairments to the nodes in a cluster.\n\n## Features\n\nTraffic
          Direction:\n\n* Egress impairments affect the traffic going out.    \n*
          Ingress impairments affect the traffic coming in. (Ingress impairments require
          the kernel module IFB.)\n\n### Impairments\n\n| Impairment    | Description
          \                               | Unit    | Uses Netem | Correlation Supported
          |\n|---------------|--------------------------------------------|---------|------------|-----------------------|\n|
          Bandwidth     | The bandwidth limit                        | kbit/s  | Yes
          \       | No                    |\n| Latency       | The delay of the packets
          \                  | ms      | Yes        | Yes                   |\n| Packet
          Loss   | The percent of packets that are dropped    | percent | Yes        |
          Yes                   |\n| Corruption    | The percent of packets that are
          corrupted  | percent | Yes        | Yes                   |\n| Duplication
          \  | The percent of packets that are duplicated | percent | Yes        |
          Yes                   |\n| Link Flapping | Turns the interface on and off
          \            | bool    | No         | No                    |\n\n\nOn the
          tested environment (RHEL CoreOS 48.84), the impairments can be used alongside
          link flapping.\n\n\n#### Latency Options\n\nIn addition to simply delaying
          the packets, there are advanced latency options.\n\n##### Jitter\n\nSemi-randomly
          adds or subtracts from the latency according to the distribution up to the
          amount specified.\nIf latency is 100ms, and jitter is 10ms, the actual latency
          will vary 100±10ms.\n\n\n##### Distribution\n\nThe distribution of the jitter.
          The options are:\n* Normal\n* Uniform\n* Pareto\n* Paretonormal\n\n#####
          Jitter correlation\n\nThe percent chance that the next latency's jitter
          value will correlate with the preceeding latency.\n\n##### Reorder\n\nThe
          percentage of packets that are not delayed, causing a reorder due to them
          being sent before the delayed ones.\nApplying jitter itself has the potential
          to also cause reordering.\n\n##### Reorder correlation\n\nThe percet chance
          that the value for the next reorder will correlate with the preceeding value.\n\n####
          Loss options\n\n##### Correlation\n\nThe percent chance that the previous
          loss value for a packet correlates with the loss value for the next packet.\n\n####
          Corruption options\n\n###### Corruption\n\nThe percent chance that each
          packet will be corrupted.\n\n##### Corruption correlation\n\nThe percet
          chance that the value for the next corruption will correlate with the preceeding
          value.\n\n#### Duplication options\n\n##### Duplication\n\nThe percent chance
          that each packet will be duplicated.\n\n##### Duplication correlation\n\nThe
          percet chance that the value for the next duplication will correlate with
          the preceeding value.\n\n\n\n## Configuration\n\nHere is an example of the
          ClusterImpairment custom resource.\n\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: test-impairment-cr\nspec:\n  duration:
          30 # seconds\n  startDelay: 5 # seconds. It typically takes about 2-3 seconds
          for the Daemonset to run\n  interfaces:\n  - \"ens2f0\"\n  ingress: # uses
          ifb\n    bandwidth: 0 # kbit\n    latency: 10 # ms\n    latencyOptions:\n
          \     jitter: 5 # ms\n      jitterCorrelation: 25 # percent\n      distribution:
          normal\n      reorder: 25 # percent of packets that will skip the delay\n
          \     reorderCorrelation: 25 # percent\n    loss: 0 # percent\n    lossOptions:\n
          \     correlation: 25 # percent\n    corruption: 0.1 # percent\n    corruptionOptions:\n
          \     correlation: 25 # percent\n    duplication: 1 # percent\n    duplicationOptions:\n
          \     correlation: 25 # percent\n  egress:\n    bandwidth: 0 # kbit\n    latency:
          100 # ms\n    loss: 0 # percent\n  linkFlapping:\n    enable: false\n    downTime:
          3 # Seconds\n    upTime: 3 # Seconds\n  nodeSelector:\n    key: \"node-role.kubernetes.io/worker\"\n
          \   value: \"\"\n    \n```\n\n#### Interfaces\n\nThe interfaces configuration
          option is a list of all interfaces that the impairments should be applied
          to. The current implementation will error out once it hits an invalid interface.\n\nIf
          an invalid interface is found, it will print out the list of interfaces.\n\n####
          Node Selector\n\nThe default node selector is all worker nodes, but you
          can change it to whatever node selector you want by setting the key and
          value.\nThere is a toleration to allow it to run on master nodes if the
          node selector includes master nodes.\n\n#### Duration\n\nThe number of seconds
          the impairments should run. It will try to sync the start and end time between
          all pods.\nIf set to -1, it will run indefinitely (a year), until you remove
          the ClusterImpairment custom resource. This is good for testing that requires
          steady impairments.\n\nIf the script is link flapping, set this value to
          be short enough so that if the link flap interrupts communication between
          the nodes, the daemonset will remove itself.\n\n#### Start Delay\n\nThe
          delay before starting the script. If you want the pods to be in sync, a
          minimum of a few seconds should be used because the pods take time to start
          up.\n\nYou can also utilize this feature to run an separate impairment after
          another. Just apply two resources at the same time, but add the duration
          and start delay of the first to the start delay of the second.\n\n#### Ingress
          and Egress\n\nThe configuration sections \"ingress\" and \"egress\" apply
          to each direction.\n\n##### Examples:\n\n##### Example 1\n\n\nIn this example,
          egress latency is set to 100ms, but the ingress latency is set to 10ms.
          So the latency to the interface will end up being 10ms, but 100ms going
          out. When pinging, this will result in 110ms of latency.\n\n```\n\napiVersion:
          apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n  name: uneven-latency\nspec:\n
          \ duration: 60\n  startDelay: 5\n  interfaces:\n  - \"ens2f0\"\n  ingress:\n
          \   latency: 10 # ms\n  egress:\n    latency: 100 # ms\n\n```\n\n##### Example
          2\n\n\nIn this example, link flapping will be enabled, and it will turn
          the interface on and off every 2 minutes. That is long enough for kubernetes
          to determine that a node is unavailable.\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: two-min-flap\nspec:\n  duration: 480\n
          \ startDelay: 5\n  interfaces:\n  - \"ens2f0\"\n  linkFlapping:\n    enable:
          true\n    downTime: 120 # Seconds\n    upTime: 120 # Seconds\n\n```\n\n#####
          Example 3\n\n\nIn this example, a realistic set of impairments is applied
          to `ens2f0` and `eno1` for 30 seconds:\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: typical-scenario\nspec:\n  duration:
          30 # seconds\n  startDelay: 5 # seconds\n  interfaces:\n  - \"ens2f0\"\n
          \ - \"eno1\"\n  egress:\n    latency: 50 # ms. Bidirectional, so total of
          100ms\n  ingress:\n    latency: 50 # ms. Bidirectional, so total of 100ms\n
          \ loss: 0.02 # percent\n\n```\n\n##### Example 4\n\n\nIn this example, a
          combination of latency, packet loss, bandwidth, and link flapping is applied.\n\n```\n\napiVersion:
          apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n  name: all-impairments\nspec:\n
          \ duration: 480 # seconds\n  startDelay: 5 # seconds\n  interfaces:\n  -
          \"ens2f0\"\n  egress:\n    latency: 50 # ms. Bidirectional, so total of
          100ms\n    loss: 0.02 # percent\n    bandwidth: 1000 # 1000 kbit/s, about
          1 mbit/s\n  ingress:\n    latency: 50 # ms. Bidirectional, so total of 100ms\n
          \   loss: 0.02 # percent\n    bandwidth: 1000 # 1000 kbit/s, about 1 mbit/s\n
          \ linkFlapping:\n    enable: true\n    downTime: 30 # Seconds\n    upTime:
          120 # Seconds\n\n```\n\n##### Example 5\n\n\nIn this example, packet duplicaiton
          is applied.\n\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n
          \ name: duplication\nspec:\n  duration: 60\n  startDelay: 5\n  interfaces:\n
          \ - \"ens3\"\n  ingress:\n    duplication: 25 # percent\n  egress:\n    duplication:
          25 # percent\n\n```\n\n## Setup\n\n### Requirements\n\n1. You need `make`
          installed.\n2. You need access to the kubernetes cluster with a kubeconfig.\n\n###
          Installation\n\nTo run using the current latest image:\n1. Clone the repository.\n2.
          Run `make deploy` with KUBECONFIG environment variable set to the location
          of the kubeconfig.\n\nTo run with your own code, there are more steps.\n1.
          Fork the repository.\n2. Clone to a machine that has access to the Kubernetes
          cluster and the kubeconfig.\n3. Modify the makefile to change the `IMG`
          variable to your image repository. If you do not have podman installed,
          also change podman to docker.\n4. Run `make docker-build` then `make docker-push`.\n5.
          Then run `make deploy`\n\n### Deploying from operator-hub\n\n*Important*
          You must manually apply the serviceaccount!\n\nFirst, Follow the instructions
          on this site to deploy from here.\n\nNext, apply the manifest \"serviceaccount.yaml\"
          from the manifests folder.\n\nThen you can apply your ClusterImpairment
          resources.\n\n## Running impairments\n\nFirst, configure a ClusterImpairment
          type's spec with the information for the impairment you want to run. Most
          importantly, set the interface(s). You can list the interfaces with `ifconfig`.
          If an invalid interface is set, the pod's logs will show `ifconfig` output.\n\n\nOnce
          the clusterimpairment type is set, apply it and it will work.\n\n\nTo re-run
          an impairment, delete the impairment with `oc delete -f path/to/manifest.yaml`,
          then apply it again.\n\n\n## Limitations\n\n\n### Multiple Impairments\n\n\nYou
          should avoid any configurations that apply impairments to the same interface
          on the same node. There are potential conflicts. The worker pod will attempt
          to remove all impairments on that interface before applying new ones.\n\n\nInstead,
          take advantage of the full control of both ingress and egress impairments
          from within the same ClusterImpairment resource.\n\n\nThere is currently
          a configured limit of 20 ClusterImpairment resources working at the same
          time. If that is a limitation for your use case, you can change it in `config/manager/manager.yaml`
          or you can try to find a way to make each ClusterImpairment do more.\n\n###
          Traffic Control (TC)\n\nTraffic control is how cluster-impairment-operator
          applies the latency, bandwidth, and packet loss impairments. The limitation
          is due to Linux not being realtime operating system, so the impairment will
          not be perfectly consistent.\n\n### Link Flapping\n\nWhen link flapping,
          if you flap the link that Kubernetes uses to communicate with the pods,
          you may be unable to remove the pod until the link is on long enough for
          Kubernetes to communicate with the impaired node.\n\n\nIn this case, it
          is helpful to set the duration properly instead of running for an indefinite
          or large amount of time, because the node will properly go back to the unimpaired
          state at that time.\n\n\nIf the cluster becomes offline due to the link
          flapping when you do not want it to be offline, soft restarting the nodes
          after removing the custom resource should remove all impairments.\n"
        displayName: cluster-impairment-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network
        - impairment
        - kubernetes
        - openshift
        - tc
        - trafficcontrol
        links:
        - name: Cluster Impairment Operator
          url: https://github.com/redhat-performance/cluster-impairment-operator
        maintainers:
        - email: joconnel@redhat.com
          name: Jared O'Connell
        - email: akrzos@redhat.com
          name: Alex Krzos
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat Performance
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/redhat-performance/cluster-impairment-operator:1.0.4
        version: 1.0.4
      entries:
      - name: cluster-impairment-operator.v1.0.4
        version: 1.0.4
      - name: cluster-impairment-operator.v1.0.3
        version: 1.0.3
      - name: cluster-impairment-operator.v1.0.2
        version: 1.0.2
      - name: cluster-impairment-operator.v1.0.1
        version: 1.0.1
      name: beta
    defaultChannel: beta
    packageName: cluster-impairment-operator
    provider:
      name: Red Hat Performance
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zettaset
      provider-url: ""
    name: xcrypt-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zts-xcrypt-operator.v0.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xcrypt.zettaset.com/v1alpha1",
                "kind": "XCrypt",
                "metadata": {
                  "name": "zts-masterset"
                },
                "spec": {
                  "status": {
                    "podNames": "zts-ca",
                    "replicas": "1,"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.connect.redhat.com/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
          createdAt: "2021-06-16T12:29:13.97Z"
          description: XCrypt Operator deploys Zettaset XCrypt Container Encryption  -
            the only software-defined solution that offers transparent high performance
            data-at-rest encryption of critical data in container environments.
          support: Zettaset
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XCrypt is the Schema for the xcrypts API
            kind: XCrypt
            name: xcrypts.xcrypt.zettaset.com
            version: v1alpha1
        description: "Zettaset’s RedHat Certified XCrypt Operator allows users to
          deploy multiple XCrypt components on OpenShift with just a few simple commands.
          The XCrypt Operator deployment provides software-defined encryption that
          transparently protects container data throughout your OpenShift environment.\n\n##
          Core Features\n* **Automated XCrypt provisioning** - Deploying a Zettaset
          XCrypt has never been easier. Adjust cluster specific configuration and
          let the Zettaset Operator take care of provisioning XCrypt services and
          setting up the XCrypt encryption to your exact specification. \n* **CSI
          Driver deployment**: Zettaset XCrypt Operator automatically deploys Zettaset
          XCrypt CSI Driver in order to provide automated and transparent data-at-rest
          \ encryption functionality compatible with Openshift and any other container
          orchestrator. Once the Operator is provisioned, the driver runs fine tuned
          and ready to work with other XCrypt components.\n* **Multi-node topology**:
          Encrypt as many nodes and volumes as you need to. XCrypt components can
          be installed on multiple nodes across the cluster. This is defined by two
          labels: zts-master and zts-worker. Normally zts-master node is a single
          node that runs Zettaset XCrypt major services required for the product operation.
          Multiple nodes can be marked as zts-worker and those are the nodes that
          can have encrypted volumes.\n* **Seamless Encrypted Volume mount**: Volumes
          are not only being encrypted, but also are mounted/unmounted to containers
          with Persistent Volume Claims. This happens as seamlessly as regular volumes
          provision and mount/unmount.\n* **Robust failover**: A node failure doesn’t
          affect XCrypt operations on other nodes of the cluster. Encryption on a
          restarted/recovered node gets back to normal functioning if the node stays
          healthy and data hasn’t been corrupted during the restart\n\n## About Zettaset
          XCrypt Kubernetes Encryption for OpenShift\nXCrypt Kubernetes Encryption
          for OpenShift is a software-only solution that enables encryption of data
          at rest stored in Red Hat OpenShift container environments without slowing
          down the DevSecOps process. In addition to providing a transparent and high-performance
          layer of security that flexibly protects container data across on-premise,
          in hybrid, and multi-cloud environments, it also helps your organization
          efficiently transition from DevOps to DevSecOps.  \n\nMore information about
          Zettaset XCrypt Kubernetes Encryption for OpenShift can be found in the
          [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)\n\n##
          RedHat OpenShift + Zettaset: The next generation of secure enterprise infrastructure
          for DevSecOps  \nBy combining RedHat’s Container Platform with Zettaset’s
          certified software-defined encryption, you can create the next generation
          of secure enterprise infrastructure that DevSecOps requires. Together they
          create the stable, consistent, and supported base that your organization
          needs to develop applications faster, simplify updates, drive innovation,
          and reduce the risk of potential security breaches and data theft.\n###
          Core Features\n* Software-only for simple and scalable deployments \n* Negligible
          performance impact\n* Transparent to developers and administrators \n* Works
          with AWS, Google, and Azure \n* Unique encryption key per container volume
          \n* Encrypted volumes are only available when in use\n* Direct integration
          with OpenShift \n* Automated encryption policy management Secure erase of
          volumes\n* Ability to securely decommission a node\n\n## How does XCrypt
          Kubernetes Encryption fit into OpenShift? \nXCrypt XCrypt Kubernetes Encryption
          for OpenShift Encryption makes it incredibly easy to protect container data
          in your OpenShift environment. \n### Use cases include:\n* Transitioning
          from DevOps to DevSecOps \n* Ensuring data protection in OpenShift container
          environments with negligible performance impact \n* Achieving compliance
          in regulated industries like healthcare and finance\n\n## Other Information\nRole-based
          access control needs to be configured prior to the Operator installation.
          Please find details in the [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)
          \   "
        displayName: Zettaset XCrypt Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - encryption
        - storage
        links:
        - name: Zettaset
          url: https://www.zettaset.com/
        maintainers:
        - email: info@zettaset.com
          name: Zettaset
        maturity: alpha
        provider:
          name: Zettaset
        relatedImages:
        - registry.connect.redhat.com/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
        - registry.connect.redhat.com/zts/xcrypt-ca-1@sha256:9f1015dfb459b69f5dbf449efa0517452095a6cf4f20826ebae278f9f800ffe7
        - registry.connect.redhat.com/zts/xcrypt-kmip-server-1@sha256:9af5d21dd748dfbf4e107f27a5f10fa3c0ba928ade9b989a744b620ce34ec757
        - registry.connect.redhat.com/zts/xcrypt-license-server-1@sha256:bc5486892a41e70f5dc4c4e1d82fca8db2149598fb8b5950ba5da1ee2d9d245d
        - registry.connect.redhat.com/zts/xcrypt-host-manager-1@sha256:9faa4b597eb74df412761dffa057b64fb69dbf9fa016414962eb2099a343e734
        - registry.connect.redhat.com/zts/zts-csi-driver-1@sha256:c15fa9a7cc40891103f426206663dcebe1210c8ee3ab6037cbf64e1f86b9c7e5
        version: 0.0.17
      entries:
      - name: zts-xcrypt-operator.v0.0.17
        version: 0.0.17
      name: alpha
    defaultChannel: alpha
    packageName: xcrypt-operator
    provider:
      name: Zettaset
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-spark-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spark-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:a6c82a6d6ebaa0d2572d644ac9013451e020f134c6fe46222fef37b1e9aa3cbb
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:a6c82a6d6ebaa0d2572d644ac9013451e020f134c6fe46222fef37b1e9aa3cbb
        version: 24.11.1
      entries:
      - name: spark-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: spark-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:8ad9948bf8264924a0eedd92c2e58269117487b9d57816384447bb6c999be681
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:8ad9948bf8264924a0eedd92c2e58269117487b9d57816384447bb6c999be681
        version: 25.3.0
      entries:
      - name: spark-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: spark-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkConnectServerSpec via
              `CustomResource`
            displayName: sparkconnectservers.spark.stackable.tech
            kind: SparkConnectServer
            name: sparkconnectservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
        version: 25.7.0
      entries:
      - name: spark-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: spark-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkConnectServerSpec via
              `CustomResource`
            displayName: sparkconnectservers.spark.stackable.tech
            kind: SparkConnectServer
            name: sparkconnectservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
        version: 25.7.0
      entries:
      - name: spark-operator.v25.7.0
        version: 25.7.0
      - name: spark-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-spark-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://skupper.io
    name: skupper-netobs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-netobs-operator.v2.1.1-rh-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability.skupper.io/v2alpha1",
                "kind": "NetworkObserver",
                "metadata": {
                  "name": "networkobserver-sample"
                },
                "spec": {
                  "auth": {
                    "basic": {
                      "create": true,
                      "secretName": ""
                    },
                    "openshift": {
                      "cookieSecretName": "",
                      "createCookieSecret": true,
                      "serviceAccount": {
                        "create": true,
                        "nameOverride": ""
                      }
                    },
                    "strategy": "openshift"
                  },
                  "extraArgs": null,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "registry.redhat.io/service-interconnect/skupper-network-observer-rhel9",
                    "tag": "2.1.1"
                  },
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nginx": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/ubi9/nginx-124",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": 9.6
                  },
                  "openshiftOauthProxy": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-oauth-proxy-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.19.0"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "prometheus": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-prometheus-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.16.0"
                  },
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "router": {
                    "certificate": {
                      "create": true,
                      "nameOverride": ""
                    },
                    "endpoint": "amqps://skupper-router-local"
                  },
                  "securityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    }
                  },
                  "service": {
                    "port": 443,
                    "type": "ClusterIP"
                  },
                  "skipManagementLabels": false,
                  "tls": {
                    "openshiftIssued": true,
                    "secretName": "",
                    "skupperIssued": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
          createdAt: "2025-08-27T20:21:18Z"
          description: Red Hat Service Interconnect Network Observer Operator provides
            the ability to monitor and observe a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkObserver
            name: networkobservers.observability.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect Network Observer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/ubi9/nginx-124@sha256:cbcca4743015b4ef167383a080f4711a6049e275e0b66f80a74f8b087b0dfc7f
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:d3056b35d9a205b9f2c48d924f199c5ac23904eb18d526e4bff229e7c7181415
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
        version: 2.1.1-rh-8
      entries:
      - name: skupper-netobs-operator.v2.1.1-rh-8
        version: 2.1.1-rh-8
      name: stable-2
    - currentCSV: skupper-netobs-operator.v2.1.1-rh-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability.skupper.io/v2alpha1",
                "kind": "NetworkObserver",
                "metadata": {
                  "name": "networkobserver-sample"
                },
                "spec": {
                  "auth": {
                    "basic": {
                      "create": true,
                      "secretName": ""
                    },
                    "openshift": {
                      "cookieSecretName": "",
                      "createCookieSecret": true,
                      "serviceAccount": {
                        "create": true,
                        "nameOverride": ""
                      }
                    },
                    "strategy": "openshift"
                  },
                  "extraArgs": null,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "registry.redhat.io/service-interconnect/skupper-network-observer-rhel9",
                    "tag": "2.1.1"
                  },
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nginx": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/ubi9/nginx-124",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": 9.6
                  },
                  "openshiftOauthProxy": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-oauth-proxy-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.19.0"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "prometheus": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-prometheus-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.16.0"
                  },
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "router": {
                    "certificate": {
                      "create": true,
                      "nameOverride": ""
                    },
                    "endpoint": "amqps://skupper-router-local"
                  },
                  "securityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    }
                  },
                  "service": {
                    "port": 443,
                    "type": "ClusterIP"
                  },
                  "skipManagementLabels": false,
                  "tls": {
                    "openshiftIssued": true,
                    "secretName": "",
                    "skupperIssued": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
          createdAt: "2025-08-27T20:21:18Z"
          description: Red Hat Service Interconnect Network Observer Operator provides
            the ability to monitor and observe a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkObserver
            name: networkobservers.observability.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect Network Observer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:d3056b35d9a205b9f2c48d924f199c5ac23904eb18d526e4bff229e7c7181415
        - registry.redhat.io/ubi9/nginx-124@sha256:cbcca4743015b4ef167383a080f4711a6049e275e0b66f80a74f8b087b0dfc7f
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        version: 2.1.1-rh-8
      entries:
      - name: skupper-netobs-operator.v2.1.1-rh-8
        version: 2.1.1-rh-8
      name: stable-2.1
    defaultChannel: stable-2
    packageName: skupper-netobs-operator
    provider:
      name: Red Hat
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/trustification/trusted-profile-analyzer-operator
    name: rhtpa-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhtpa-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtpa.io/v1",
                "kind": "TrustedProfileAnalyzer",
                "metadata": {
                  "name": "trustedprofileanalyzer-sample"
                },
                "spec": {
                  "appDomain": "change-me",
                  "collector": {},
                  "database": {},
                  "image": {
                    "fullName": "registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e",
                    "pullPolicy": "IfNotPresent"
                  },
                  "infrastructure": {
                    "port": 9010
                  },
                  "ingress": {},
                  "metrics": {
                    "enabled": false
                  },
                  "modules": {
                    "createDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "createImporters": {
                      "enabled": false,
                      "image": {},
                      "importers": {}
                    },
                    "importer": {
                      "concurrency": 1,
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {},
                      "workingDirectory": {
                        "size": "32Gi"
                      }
                    },
                    "migrateDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "server": {
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "ingress": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {}
                    }
                  },
                  "oidc": {
                    "clients": {
                      "cli": {
                        "clientSecret": null
                      },
                      "frontend": {}
                    }
                  },
                  "openshift": {
                    "useServiceCa": true
                  },
                  "partOf": "trustify",
                  "replicas": 1,
                  "rust": {},
                  "storage": {},
                  "tls": {},
                  "tracing": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-09-19T15:04:16Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Profile Analyzer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trustification/trusted-profile-analyzer-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: TrustedProfileAnalyzer
            name: trustedprofileanalyzers.rhtpa.io
            version: v1
        description: |
          [Tech Preview]
          Red Hat Trusted Profile Analyzer (RHTPA) provides a single source of truth for your software inventory across all your environments. You can verify the integrity of your in-house and third-party applications against their respective Software Bill of Materials (SBOM) document.  Trusted Profile Analyzer provides vulnerability, component and license transparency, strengthens your software supply chain security, and enables an accurate, up-to-date understanding of your software inventory.

          Features and Benefits:

          Enhanced Software Supply Chain Security: Offers critical transparency and integrity verification across the entire software lifecycle.

          Reduced Operational Risk: Proactively identifies and helps remediate vulnerabilities minimizing risk and compliance violations.

          Improved Compliance and Auditability: Simplifies the process of meeting regulatory requirements for software transparency and integrity by providing verifiable documentation and continuous monitoring.

          Faster Remediation and Incident Response: Provides targeted, actionable insights to prioritize and accelerate the patching and remediation of vulnerabilities.

          Increased Developer Productivity & Trust: Integrates security into existing workflows without significant overhead, allowing developers to build and deploy with greater confidence in the integrity of their code and components.

          Centralized Visibility and Control: Consolidates disparate security information into a single pane of glass, offering a holistic view of your application security posture.

          Red Hat Trusted Profile Analyzer is a crucial tool for any organization running applications on OpenShift that needs to confidently understand, verify, and secure the software they deploy, especially in an era of complex software supply chains and increasing regulatory scrutiny.

          [Documentation](https://docs.redhat.com/en/documentation/red_hat_trusted_profile_analyzer/2.1).
        displayName: Red Hat Trusted Profile Analyzer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rhtpa
        - trusted-profile-analyzer
        links:
        - name: Trusted Profile Analyzer Operator
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        maintainers:
        - email: rhtpa-support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        relatedImages:
        - registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
        - registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e
        version: 1.0.2
      entries:
      - name: rhtpa-operator.v1.0.2
        version: 1.0.2
      - name: rhtpa-operator.v1.0.1
        version: 1.0.1
      - name: rhtpa-operator.v1.0.0
        version: 1.0.0
      name: stable
    - currentCSV: rhtpa-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtpa.io/v1",
                "kind": "TrustedProfileAnalyzer",
                "metadata": {
                  "name": "trustedprofileanalyzer-sample"
                },
                "spec": {
                  "appDomain": "change-me",
                  "collector": {},
                  "database": {},
                  "image": {
                    "fullName": "registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e",
                    "pullPolicy": "IfNotPresent"
                  },
                  "infrastructure": {
                    "port": 9010
                  },
                  "ingress": {},
                  "metrics": {
                    "enabled": false
                  },
                  "modules": {
                    "createDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "createImporters": {
                      "enabled": false,
                      "image": {},
                      "importers": {}
                    },
                    "importer": {
                      "concurrency": 1,
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {},
                      "workingDirectory": {
                        "size": "32Gi"
                      }
                    },
                    "migrateDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "server": {
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "ingress": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {}
                    }
                  },
                  "oidc": {
                    "clients": {
                      "cli": {
                        "clientSecret": null
                      },
                      "frontend": {}
                    }
                  },
                  "openshift": {
                    "useServiceCa": true
                  },
                  "partOf": "trustify",
                  "replicas": 1,
                  "rust": {},
                  "storage": {},
                  "tls": {},
                  "tracing": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-09-19T15:04:16Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Profile Analyzer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trustification/trusted-profile-analyzer-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: TrustedProfileAnalyzer
            name: trustedprofileanalyzers.rhtpa.io
            version: v1
        description: |
          [Tech Preview]
          Red Hat Trusted Profile Analyzer (RHTPA) provides a single source of truth for your software inventory across all your environments. You can verify the integrity of your in-house and third-party applications against their respective Software Bill of Materials (SBOM) document.  Trusted Profile Analyzer provides vulnerability, component and license transparency, strengthens your software supply chain security, and enables an accurate, up-to-date understanding of your software inventory.

          Features and Benefits:

          Enhanced Software Supply Chain Security: Offers critical transparency and integrity verification across the entire software lifecycle.

          Reduced Operational Risk: Proactively identifies and helps remediate vulnerabilities minimizing risk and compliance violations.

          Improved Compliance and Auditability: Simplifies the process of meeting regulatory requirements for software transparency and integrity by providing verifiable documentation and continuous monitoring.

          Faster Remediation and Incident Response: Provides targeted, actionable insights to prioritize and accelerate the patching and remediation of vulnerabilities.

          Increased Developer Productivity & Trust: Integrates security into existing workflows without significant overhead, allowing developers to build and deploy with greater confidence in the integrity of their code and components.

          Centralized Visibility and Control: Consolidates disparate security information into a single pane of glass, offering a holistic view of your application security posture.

          Red Hat Trusted Profile Analyzer is a crucial tool for any organization running applications on OpenShift that needs to confidently understand, verify, and secure the software they deploy, especially in an era of complex software supply chains and increasing regulatory scrutiny.

          [Documentation](https://docs.redhat.com/en/documentation/red_hat_trusted_profile_analyzer/2.1).
        displayName: Red Hat Trusted Profile Analyzer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rhtpa
        - trusted-profile-analyzer
        links:
        - name: Trusted Profile Analyzer Operator
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        maintainers:
        - email: rhtpa-support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        relatedImages:
        - registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
        - registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e
        version: 1.0.2
      entries:
      - name: rhtpa-operator.v1.0.2
        version: 1.0.2
      - name: rhtpa-operator.v1.0.1
        version: 1.0.1
      - name: rhtpa-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable
    packageName: rhtpa-operator
    provider:
      name: Red Hat
      url: https://github.com/trustification/trusted-profile-analyzer-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: nova-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nova-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "Nova",
                "metadata": {
                  "name": "nova"
                },
                "spec": {
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaAPI",
                "metadata": {
                  "name": "novaapi-sample"
                },
                "spec": {
                  "apiDatabaseHostname": "openstack",
                  "cell0DatabaseHostname": "openstack",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "registeredCells": {},
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaCell",
                "metadata": {
                  "name": "novacell1-with-upcall-sample"
                },
                "spec": {
                  "apiDatabaseHostname": "openstack",
                  "cellDatabaseHostname": "openstack",
                  "cellName": "cell1",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaCompute",
                "metadata": {
                  "name": "nova-compute-ironic"
                },
                "spec": {
                  "novaComputeTemplates": {
                    "compute-ironic": {
                      "computeDriver": "ironic.IronicDriver",
                      "customServiceConfig": "",
                      "networkAttachments": [
                        "internalapi"
                      ],
                      "replicas": 1,
                      "resources": {}
                    }
                  }
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaConductor",
                "metadata": {
                  "name": "novaconductor-sample"
                },
                "spec": {
                  "cellName": "cell1",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaMetadata",
                "metadata": {
                  "name": "novametadata-sample"
                },
                "spec": {
                  "apiDatabaseAccount": "nova-api",
                  "apiDatabaseHostname": "openstack",
                  "cellDatabaseAccount": "nova-cell0",
                  "cellDatabaseHostname": "openstack",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "registeredCells": {},
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaNoVNCProxy",
                "metadata": {
                  "name": "novanovncproxy-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaScheduler",
                "metadata": {
                  "name": "novascheduler-sample"
                },
                "spec": {
                  "apiDatabaseHostname": "openstack",
                  "cell0DatabaseHostname": "openstack",
                  "customServiceConfig": "[DEFAULT]\ndebug=true\n",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "registeredCells": {},
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:21Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nova
            name: nova.nova.openstack.org
            version: v1beta1
          - description: NovaAPI is the Schema for the novaapis API
            displayName: Nova API
            kind: NovaAPI
            name: novaapis.nova.openstack.org
            version: v1beta1
          - description: NovaCell is the Schema for the novacells API
            displayName: Nova Cell
            kind: NovaCell
            name: novacells.nova.openstack.org
            version: v1beta1
          - description: NovaCompute is the Schema for the NovaCompute
            displayName: Nova Compute
            kind: NovaCompute
            name: novacomputes.nova.openstack.org
            version: v1beta1
          - description: NovaConductor is the Schema for the novaconductors API
            displayName: Nova Conductor
            kind: NovaConductor
            name: novaconductors.nova.openstack.org
            version: v1beta1
          - description: NovaMetadata is the Schema for the novametadata API
            displayName: Nova Metadata
            kind: NovaMetadata
            name: novametadata.nova.openstack.org
            version: v1beta1
          - description: NovaNoVNCProxy is the Schema for the novanovncproxies API
            displayName: Nova No VNCProxy
            kind: NovaNoVNCProxy
            name: novanovncproxies.nova.openstack.org
            version: v1beta1
          - description: NovaScheduler is the Schema for the novaschedulers API
            displayName: Nova Scheduler
            kind: NovaScheduler
            name: novaschedulers.nova.openstack.org
            version: v1beta1
        description: Nova Operator
        displayName: Nova Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Nova
        links:
        - name: Nova Operator
          url: https://github.com/openstack-k8s-operators/nova-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-nova-scheduler-rhel9@sha256:ba91196cd4f9b6e6f49868c91343b8eca7cc0d5016c36e83e2cc76f45cbe777b
        - registry.redhat.io/rhoso/openstack-nova-compute-rhel9@sha256:43cbe46ab263403fc303b6bc2388863029aad33465b237db18da7cd6d42078e1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/nova-rhel9-operator@sha256:f058c56df02a108505d51b9d19973894c2ee59055776c9f3fd7f7afe23a5106a
        - registry.redhat.io/rhoso/openstack-nova-api-rhel9@sha256:bbb8ebceb6f30b6c8beb120429e69f323bd231c09aaaddc5be9f15a527e928ed
        - registry.redhat.io/rhoso/openstack-nova-conductor-rhel9@sha256:500c36fd519611be8662e97ca31b1e59269c14d8aa94286dad9925e957462155
        - registry.redhat.io/rhoso/openstack-nova-novncproxy-rhel9@sha256:bf534f0c72f188d5f51e40b0392ace182db2055dff9ad14b01eeef13056febbf
        version: 1.0.6
      entries:
      - name: nova-operator.v1.0.6
        version: 1.0.6
      - name: nova-operator.v1.0.5
        version: 1.0.5
      - name: nova-operator.v1.0.4
        version: 1.0.4
      - name: nova-operator.v1.0.3
        version: 1.0.3
      - name: nova-operator.v1.0.2
        version: 1.0.2
      - name: nova-operator.v1.0.1
        version: 1.0.1
      - name: nova-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: nova-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq7-interconnect-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: 1.10.x
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: 1.2.0
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: latest
    defaultChannel: 1.10.x
    packageName: amq7-interconnect-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-registry
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-operator.v1.1.3-v2.6.4.final
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-mem"
                },
                "spec": {
                  "configuration": {
                    "persistence": "mem"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-operator:1.1.3
          createdAt: "2024-09-17"
          description: Deploy and manage Apicurio Registry on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents an Apicurio Registry instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Apicurio Registry

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Provides a REST API and web UI to manage the artifacts and artifact meta-data
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, in-memory
          - Compatible with Confluent and IBM APIs
          - Runs on a lightweight Quarkus platform
          - Includes Maven plugin to integrate with Maven based builds

          ## Apicurio Registry Operator

          Provides a quick and easy way to deploy and manage Apicurio Registry on Kubernetes.

          **Features**
          - Supports installation and configuration of the Registry for all storage options
          - Easily perform a rolling upgrade of Apicurio Registry

          ## Prerequisites

          This operator does not deploy storage for Apicurio Registry. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/main/LICENSE)
        displayName: Apicurio Registry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-operator:1.1.3
        version: 1.1.3-v2.6.4.final
      entries:
      - name: apicurio-registry-operator.v1.1.3-v2.6.4.final
        version: 1.1.3-v2.6.4.final
      - name: apicurio-registry-operator.v1.1.2-v2.5.11.final
        version: 1.1.2-v2.5.11.final
      - name: apicurio-registry-operator.v1.1.1-v2.5.9.final
        version: 1.1.1-v2.5.9.final
      - name: apicurio-registry-operator.v1.1.0-v2.4.12.final
        version: 1.1.0-v2.4.12.final
      - name: apicurio-registry-operator.v1.0.0-v2.0.0.final
        version: 1.0.0-v2.0.0.final
      name: 2.x
    defaultChannel: 2.x
    packageName: apicurio-registry
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: volume-expander-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: volume-expander-operator.v0.3.6
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage,Monitoring
          certified: "false"
          containerImage: quay.io/redhat-cop/volume-expander-operator@sha256:e4059a66a9af7c786c8c651abb8a57a55a0e528342e3a59f8c264b40f87d6e25
          createdAt: "2022-05-05T13:25:53Z"
          description: This operator provides automation expanding pod's volumes when
            they are about to run out of space.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: volume-expander-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/volume-expander-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The purpose of the volume-expander-operator is to expand volumes when they are running out of space.
          This is achieved by using the [volume expansion feature](https://kubernetes.io/docs/concepts/storage/persistent-volumes/#expanding-persistent-volumes-claims).

          The operator periodically checks the `kubelet_volume_stats_used_bytes` and `kubelet_volume_stats_capacity_bytes published` by the kubelets to decide when to expand a volume.
          Notice that these metrics are generated only when a volume is mounted to a pod. Also the kubelet takes a minute or two to start generating accurate values for these metrics. The operator accounts for that.

          This operator works based on the following annotations to PersistentVolumeClaim resources:

          | Annotation | Default &nbsp;  &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;  | Description  |
          | :--- | :--- | :--- |
          | `volume-expander-operator.redhat-cop.io/autoexpand`  | N/A  | if set to "true" enables the volume-expander-operator to watch on this PVC  |
          | `volume-expander-operator.redhat-cop.io/polling-frequency`  | `"30s"` | How frequently to poll the volume metrics. Express this value as a valid golang [Duration](https://golang.org/pkg/time/#ParseDuration)  |
          | `volume-expander-operator.redhat-cop.io/expand-threshold-percent` | `"80"` | the percentage of used storage after which the volume will be expanded. This must be a positive integer. |
          | `volume-expander-operator.redhat-cop.io/expand-by-percent` | `"25"` | the percentage by which the volume will be expanded, relative to the current size. This must be an integer between 0 and 100 |
          | `volume-expander-operator.redhat-cop.io/expand-up-to` | MaxInt64 | the upper bound for this volume to be expanded to. The default value is the largest quantity representable and is intended to be interpreted as infinite. If the default is used it is recommend to ensure the namespace has a quota on the used storage class. |

          Note that not all of the storage driver implementations support volume expansion. It is a responsibility of the user/platform administrator to ensure that storage class and the persistent volume claim meet all the requirements needed for the volume expansion feature to work properly.

          This operator was tested with [OCS](https://www.redhat.com/en/technologies/cloud-computing/openshift-container-storage), but should work with any other storage driver that supports volume expansion.
        displayName: Volume Expander Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - volume
        - storage
        - csi
        - expansion
        links:
        - name: Volume Expander Operator
          url: https://github.com/redhat-cop/volume-expander-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/volume-expander-operator@sha256:e4059a66a9af7c786c8c651abb8a57a55a0e528342e3a59f8c264b40f87d6e25
        version: 0.3.6
      entries:
      - name: volume-expander-operator.v0.3.6
        version: 0.3.6
      - name: volume-expander-operator.v0.3.5
        version: 0.3.5
      - name: volume-expander-operator.v0.3.4
        version: 0.3.4
      - name: volume-expander-operator.v0.3.3
        version: 0.3.3
      - name: volume-expander-operator.v0.3.2
        version: 0.3.2
      - name: volume-expander-operator.v0.3.1
        version: 0.3.1
      - name: volume-expander-operator.v0.3.0
        version: 0.3.0
      - name: volume-expander-operator.v0.2.0
        version: 0.2.0
      - name: volume-expander-operator.v0.1.8
        version: 0.1.8
      - name: volume-expander-operator.v0.1.7
        version: 0.1.7
      - name: volume-expander-operator.v0.1.6
        version: 0.1.6
      - name: volume-expander-operator.v0.1.5
        version: 0.1.5
      - name: volume-expander-operator.v0.1.4
        version: 0.1.4
      - name: volume-expander-operator.v0.1.3
        version: 0.1.3
      - name: volume-expander-operator.v0.1.2
        version: 0.1.2
      - name: volume-expander-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: volume-expander-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-eventbridge-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-eventbridge-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Archive",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Endpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "EventBus",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/eventbridge-controller:1.1.1
          createdAt: "2025-09-19T20:51:55Z"
          description: AWS EventBridge controller is a service controller for managing
            EventBridge resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Archive represents the state of an AWS eventbridge Archive
              resource.
            displayName: Archive
            kind: Archive
            name: archives.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: Endpoint represents the state of an AWS eventbridge Endpoint
              resource.
            displayName: Endpoint
            kind: Endpoint
            name: endpoints.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: EventBus represents the state of an AWS eventbridge EventBus
              resource.
            displayName: EventBus
            kind: EventBus
            name: eventbuses.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: Rule represents the state of an AWS eventbridge Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.eventbridge.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EventBridge resources in AWS from within your Kubernetes cluster.

          **About Amazon EventBridge**

          EventBridge is a serverless service that uses events to connect application components together, making it easier for you to build scalable event-driven applications. Use it to route events from sources such as home-grown applications, AWS services, and third- party software to consumer applications across your organization. EventBridge provides a simple and consistent way to ingest, filter, transform, and deliver events so you can build new applications quickly.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EventBridge
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eventbridge
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EventBridge Developer Resources
          url: https://aws.amazon.com/eventbridge/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eventbridge maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/eventbridge-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-eventbridge-controller.v1.1.1
        version: 1.1.1
      - name: ack-eventbridge-controller.v1.1.0
        version: 1.1.0
      - name: ack-eventbridge-controller.v1.0.28
        version: 1.0.28
      - name: ack-eventbridge-controller.v1.0.27
        version: 1.0.27
      - name: ack-eventbridge-controller.v1.0.26
        version: 1.0.26
      - name: ack-eventbridge-controller.v1.0.25
        version: 1.0.25
      - name: ack-eventbridge-controller.v1.0.24
        version: 1.0.24
      - name: ack-eventbridge-controller.v1.0.23
        version: 1.0.23
      - name: ack-eventbridge-controller.v1.0.22
        version: 1.0.22
      - name: ack-eventbridge-controller.v1.0.21
        version: 1.0.21
      - name: ack-eventbridge-controller.v1.0.20
        version: 1.0.20
      - name: ack-eventbridge-controller.v1.0.18
        version: 1.0.18
      - name: ack-eventbridge-controller.v1.0.17
        version: 1.0.17
      - name: ack-eventbridge-controller.v1.0.16
        version: 1.0.16
      - name: ack-eventbridge-controller.v1.0.15
        version: 1.0.15
      - name: ack-eventbridge-controller.v1.0.14
        version: 1.0.14
      - name: ack-eventbridge-controller.v1.0.13
        version: 1.0.13
      - name: ack-eventbridge-controller.v1.0.12
        version: 1.0.12
      - name: ack-eventbridge-controller.v1.0.11
        version: 1.0.11
      - name: ack-eventbridge-controller.v1.0.6
        version: 1.0.6
      - name: ack-eventbridge-controller.v1.0.5
        version: 1.0.5
      - name: ack-eventbridge-controller.v1.0.4
        version: 1.0.4
      - name: ack-eventbridge-controller.v1.0.3
        version: 1.0.3
      - name: ack-eventbridge-controller.v1.0.1
        version: 1.0.1
      - name: ack-eventbridge-controller.v1.0.0
        version: 1.0.0
      - name: ack-eventbridge-controller.v0.0.7
        version: 0.0.7
      - name: ack-eventbridge-controller.v0.0.6
        version: 0.0.6
      - name: ack-eventbridge-controller.v0.0.5
        version: 0.0.5
      - name: ack-eventbridge-controller.v0.0.4
        version: 0.0.4
      - name: ack-eventbridge-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-eventbridge-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Oracle Corporation
      provider-url: https://github.com/oracle/coherence-operator
    name: coherence-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: coherence-operator.v3.5.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "coherence.oracle.com/v1",
                "kind": "Coherence",
                "metadata": {
                  "name": "coherence-sample"
                },
                "spec": {
                  "replicas": 3
                }
              },
              {
                "apiVersion": "coherence.oracle.com/v1",
                "kind": "CoherenceJob",
                "metadata": {
                  "name": "coherence-job-sample"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "false"
          containerImage: container-registry.oracle.com/middleware/coherence-operator:3.5.4
          createdAt: "2025-08-28T16:15:09Z"
          description: The Oracle Coherence Kubernetes Operator enables easy management
            of Coherence clusters in a Kubernetes environment.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/arch.amd64: supported
          operatorframework.io/arch.arm64: supported
          operatorframework.io/os.linux: supported
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/oracle/coherence-operator
          support: Oracle Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Coherence is the top level schema for the Coherence API and
              custom resource definition (CRD).
            displayName: Coherence
            kind: Coherence
            name: coherence.coherence.oracle.com
            version: v1
          - description: |-
              CoherenceJob is the top level schema for the CoherenceJob API and custom resource definition (CRD)
              for configuring Coherence Job workloads.
            displayName: Coherence Job
            kind: CoherenceJob
            name: coherencejob.coherence.oracle.com
            version: v1
        description: |
          The Oracle Coherence Kubernetes Operator enables easy management of Coherence clusters in a Kubernetes environment.

          ### Installation Guide & Configuration

          For the complete installation guide and all configuration options please refer to
          [Coherence Operator Documentation](https://oracle.github.io/coherence-operator/docs/latest/#/about/01_overview).
        displayName: Coherence Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Coherence
        - Oracle
        - In Memory Data
        - Caching
        links:
        - name: Coherence Kubernetes Operator
          url: https://oracle.github.io/coherence-operator/docs/latest
        - name: Oracle Coherence
          url: https://coherence.community
        maintainers:
        - email: jonathan.knight@oracle.com
          name: Jonathan Knight
        maturity: mature
        minKubeVersion: 1.26.0
        provider:
          name: Oracle Corporation
          url: https://github.com/oracle/coherence-operator
        relatedImages:
        - container-registry.oracle.com/middleware/coherence-operator:3.5.4
        version: 3.5.4
      entries:
      - name: coherence-operator.v3.5.4
        version: 3.5.4
      - name: coherence-operator.v3.5.2
        version: 3.5.2
      - name: coherence-operator.v3.5.0
        version: 3.5.0
      - name: coherence-operator.v3.4.3
        version: 3.4.3
      name: stable
    defaultChannel: stable
    packageName: coherence-operator
    provider:
      name: Oracle Corporation
      url: https://github.com/oracle/coherence-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nxrm-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nxrm-operator-certified.v3.58.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusRepo",
                "metadata": {
                  "name": "example-nexusrepo"
                },
                "spec": {
                  "config": {
                    "data": null,
                    "enabled": false,
                    "mountPath": "/sonatype-nexus-conf"
                  },
                  "deployment": {
                    "additionalContainers": null,
                    "additionalVolumeMounts": null,
                    "additionalVolumes": null,
                    "annotations": {},
                    "initContainers": null,
                    "postStart": {
                      "command": null
                    },
                    "preStart": {
                      "command": null
                    },
                    "terminationGracePeriodSeconds": 120
                  },
                  "deploymentStrategy": {
                    "type": "Recreate"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "path": "/",
                    "rules": null,
                    "tls": {
                      "enabled": true,
                      "secretName": "nexus-tls"
                    }
                  },
                  "nexus": {
                    "dockerPort": 5003,
                    "env": [
                      {
                        "name": "INSTALL4J_ADD_VM_PARAMS",
                        "value": "-Xms2703M -Xmx2703M -XX:MaxDirectMemorySize=2703M -XX:+UnlockExperimentalVMOptions -XX:+UseCGroupMemoryLimitForHeap"
                      }
                    ],
                    "hostAliases": [],
                    "imageName": "registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:06fe31f40ff3fd7556fc7b8f6b2efe80ada1d9079ec9aaad60d1fdc0ee0677f0",
                    "imagePullPolicy": "IfNotPresent",
                    "imagePullSecret": "",
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "path": "/",
                      "periodSeconds": 30
                    },
                    "nexusPort": 8081,
                    "podAnnotations": {},
                    "readinessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "path": "/",
                      "periodSeconds": 30
                    },
                    "resources": {},
                    "securityContext": {
                    },
                    "service": {
                      "type": "NodePort"
                    }
                  },
                  "nexusProxyRoute": {
                    "annotations": null,
                    "enabled": false,
                    "labels": null
                  },
                  "persistence": {
                    "accessMode": "ReadWriteOnce",
                    "enabled": true,
                    "storageSize": "8Gi"
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": null,
                    "enabled": false,
                    "labels": null,
                    "name": "docker",
                    "portName": "docker"
                  },
                  "secret": {
                    "data": null,
                    "enabled": false,
                    "mountPath": "/etc/secret-volume",
                    "readOnly": true
                  },
                  "service": {
                    "annotations": {},
                    "enabled": false,
                    "labels": {},
                    "ports": [
                      {
                        "name": "nexus-service",
                        "port": 80,
                        "targetPort": 80
                      }
                    ]
                  },
                  "statefulset": {
                    "enabled": false
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nxrm-operator-certified@sha256:4945ad993a10e27b358b2d9869d05f23ca761162ec088e5a833939022bc23192
          createdAt: "2023-07-14"
          description: |-
            Nexus Repository is the central source of control to efficiently manage all binaries
            and build artifacts across your DevOps pipeline.
          olm.skipRange: <3.58.1-2
          repository: https://github.com/sonatype/operator-nxrm3
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus Repository
            displayName: NexusRepo
            kind: NexusRepo
            name: nexusrepos.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Repository is the central source of control to efficiently manage all binaries
          and build artifacts across your DevOps pipeline.
          The flow of open source and third-party components into and through an organization
          creates a complex software supply chain.
          Nexus Repository delivers speed, efficiency, and quality to the governance
          and management of all dependencies, libraries, and applications for your DevOps teams.

          ## Core Capabilities

          * **Dependency Management**:
            Improves reliability with repeatable, fast access to secure dependencies
          * **Developer Productivity**:
            Streamline developer workflows by enabling the sharing of components and applications across teams
          * **Supply Chain Performance**:
            Improve speed-to-market and reduced build times with release advanced staging and component tagging
          * **CI/CD Integrations**:
            Increase DevOps scalability with integrations to the most popular build and deployment tools

          Version control systems and package registries do not scale when managing proprietary,
          open source, and third-party components.
          Organizations need a central binary and build artifact repository to manage dependencies
          across the entire software supply chain.

          ## Limitations

          High Availability Clustering (HA-C) is not supported for Nexus Repository Pro for OpenShift.

          This operator will be released on a quarterly basis.

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus Repository Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage

          Once the server instance is created by the operator and running,
          you'll want to expose the service as you see fit:
          1. Create a Route to that service for nexus.port (8081).

          By default, the Nexus Repository starts up in OSS mode until a license is installed.

          The Nexus Repository can be further configured via the NexusRepo custom resource definition:

          | Parameter                                   | Description                         | Default                                 |
          | ------------------------------------------  | ----------------------------------  | ----------------------------------------|
          | `statefulset.enabled`                       | Use statefulset instead of deployment | `false` |
          | `deploymentStrategy.type`                   | Deployment Strategy     |  `Recreate` |
          | `nexus.env`                                 | Nexus environment variables         | `See example.` |
          | `nexus.resources`                           | Nexus resource requests and limits  | `{}`                                    |
          | `nexus.dockerPort`                          | Port to access docker               | `5003`                                  |
          | `nexus.nexusPort`                           | Internal port for Nexus service     | `8081`                                  |
          | `nexus.service.type`                        | Service for Nexus                   |`NodePort`                                |
          | `nexus.service.clusterIp`                   | Specific cluster IP when service type is cluster IP. Use None for headless service |`nil`   |
          | `nexus.securityContext`                     | Security Context |
          | `nexus.labels`                              | Service labels                      | `{}`                                    |
          | `nexus.podAnnotations`                      | Pod Annotations                     | `{}`
          | `nexus.livenessProbe.initialDelaySeconds`   | LivenessProbe initial delay         | 30                                      |
          | `nexus.livenessProbe.periodSeconds`         | Seconds between polls               | 30                                      |
          | `nexus.livenessProbe.failureThreshold`      | Number of attempts before failure   | 6                                       |
          | `nexus.livenessProbe.timeoutSeconds`        | Time in seconds after liveness probe times out    | `nil`                     |
          | `nexus.livenessProbe.path`                  | Path for LivenessProbe              | /                                       |
          | `nexus.readinessProbe.initialDelaySeconds`  | ReadinessProbe initial delay        | 30                                      |
          | `nexus.readinessProbe.periodSeconds`        | Seconds between polls               | 30                                      |
          | `nexus.readinessProbe.failureThreshold`     | Number of attempts before failure   | 6                                       |
          | `nexus.readinessProbe.timeoutSeconds`       | Time in seconds after readiness probe times out    | `nil`                    |
          | `nexus.readinessProbe.path`                 | Path for ReadinessProbe             | /                                       |
          | `nexus.hostAliases`                         | Aliases for IPs in /etc/hosts       | []                                      |
          | `ingress.enabled`                           | Create an ingress for Nexus         | `true`                                  |
          | `ingress.annotations`                       | Annotations to enhance ingress configuration  | `{}`                          |
          | `ingress.tls.enabled`                       | Enable TLS                          | `true`                                 |
          | `ingress.tls.secretName`                    | Name of the secret storing TLS cert, `false` to use the Ingress' default certificate | `nexus-tls`                             |
          | `ingress.path`                              | Path for ingress rules. GCP users should set to `/*` | `/`                    |
          | `tolerations`                               | tolerations list                    | `[]`                                    |
          | `config.enabled`                            | Enable configmap                    | `false`                                 |
          | `config.mountPath`                          | Path to mount the config            | `/sonatype-nexus-conf`                  |
          | `config.data`                               | Configmap data                      | `nil`                                   |
          | `deployment.terminationGracePeriodSeconds`  | Time to allow for clean shutdown    | 120                                     |
          | `deployment.annotations`                    | Annotations to enhance deployment configuration  | `{}`                       |
          | `deployment.initContainers`                 | Init containers to run before main containers  | `nil`                        |
          | `deployment.postStart.command`              | Command to run after starting the nexus container  | `nil`                    |
          | `deployment.preStart.command`               | Command to run before starting the nexus container  | `nil`                   |
          | `deployment.additionalContainers`           | Add additional Container         | `nil`                                      |
          | `deployment.additionalVolumes`              | Add additional Volumes           | `nil`                                      |
          | `deployment.additionalVolumeMounts`         | Add additional Volume mounts     | `nil`                                      |
          | `secret.enabled`                            | Enable secret                    | `false`                                    |
          | `secret.mountPath`                          | Path to mount the secret         | `/etc/secret-volume`                       |
          | `secret.readOnly`                           | Secret readonly state            | `true`                                     |
          | `secret.data`                               | Secret data                      | `nil`                                      |
          | `service.enabled`                           | Enable additional service        | `nil`                                      |
          | `service.name`                              | Service name                     | `nil`                                      |
          | `service.portName`                          | Service port name                | `nil`                                      |
          | `service.labels`                            | Service labels                   | `nil`                                      |
          | `service.annotations`                       | Service annotations              | `nil`                                      |
          | `service.loadBalancerSourceRanges`          | Service LoadBalancer source IP whitelist | `nil`                              |
          | `service.targetPort`                        | Service port                     | `nil`                                      |
          | `service.port`                              | Port for exposing service        | `nil`                                      |
          | `route.enabled`         | Set to true to create route for additional service | `false` |
          | `route.name`            | Name of route                                      | `docker` |
          | `route.portName`        | Target port name of service                        | `docker` |
          | `route.labels`          | Labels to be added to route                        | `{}` |
          | `route.annotations`     | Annotations to be added to route                   | `{}` |
          | `route.path`            | Host name of Route e.g jenkins.example.com         | nil |
        displayName: Nexus Repository Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - repository
        - sonatype
        links:
        - name: Nexus Repository Manager
          url: https://www.sonatype.com/product-nexus-repository
        maintainers:
        - email: support@sonatype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nxrm-operator-certified@sha256:4945ad993a10e27b358b2d9869d05f23ca761162ec088e5a833939022bc23192
        - registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:06fe31f40ff3fd7556fc7b8f6b2efe80ada1d9079ec9aaad60d1fdc0ee0677f0
        version: 3.58.1-2
      entries:
      - name: nxrm-operator-certified.v3.58.1-2
        version: 3.58.1-2
      name: stable
    defaultChannel: stable
    packageName: nxrm-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: AMD
      provider-url: ""
    name: amd-t2-january
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amd-t2-january.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovfect2.amd.com/v1",
                "kind": "SriovT2Card",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "amd-t2-january",
                    "app.kubernetes.io/instance": "sriovt2card-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sriovt2card",
                    "app.kubernetes.io/part-of": "amd-t2-january"
                  },
                  "name": "sriovt2card",
                  "namespace": "amd-t2"
                },
                "spec": {
                  "acceleratorSelector": {
                    "pciAddress": "0000:01:00.0"
                  },
                  "nodeSelector": {
                    "kubernetes.io/hostname": "7c-c2-55-85-ba-fe.amd.com"
                  },
                  "physicalFunction": {
                    "pfDriver": "qdma-pf",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-07-14T09:46:54Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovT2Card is the Schema for the sriovt2cards API
            displayName: Sriov T2 Card
            kind: SriovT2Card
            name: sriovt2cards.sriovfect2.amd.com
            version: v1
        description: The AMD-T2 Operator for accelerator enables the seamless integration
          and management of AMD T2 accelerator within Openshift/Kubernetes envirnoment.
        displayName: AMD-T2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - SRIOV
        - Xilinx
        - FPGA
        - Telco
        - T2
        - MultiArch
        - AMD
        - TelcoAccelerator
        links:
        - name: Amd T2 January
          url: https://amd-t2-january.domain
        maintainers:
        - email: dhirajs.lakhane@amd.com
          name: DhirajLakhane
        maturity: alpha
        provider:
          name: AMD
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:928e64203edad8f1bba23593c7be04f0f8410c6e4feb98d9e9c2d00a8ff59048
        - quay.io/amdt2operator/amd-t2-january@sha256:318f56e2670582c217b8973f69f0675e7fff7dc19d07875294f2ff2fc5403902
        version: 1.0.1
      entries:
      - name: amd-t2-january.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: amd-t2-january
    provider:
      name: AMD
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Cloudera
      provider-url: https://cloudera.com
    name: cfm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cfm-operator.v2.9.0-b96
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "Nifi",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifi-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifi",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifi-sample"
                },
                "spec": {
                  "hostName": "nifi.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifi-k8s",
                    "tag": "2.8.0-b94-nifi_1.25.0.2.3.13.0-36"
                  },
                  "replicas": 1,
                  "stateManagement": {
                    "clusterProvider": {
                      "class": "org.apache.nifi.kubernetes.state.provider.KubernetesConfigMapStateProvider",
                      "id": "kubernetes-provider"
                    }
                  },
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.8.0-b94"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              },
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "NifiRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifiregistry-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifiregistry",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifiregistry-sample"
                },
                "spec": {
                  "hostName": "registry.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s",
                    "tag": "2.8.0-b94-nifi_1.25.0.2.3.13.0-36"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.8.0-b94"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning,Big Data
          com.redhat.openshift.versions: v4.10
          containerImage: container.repository.cloudera.com/cloudera/cfm-operator@sha256:890d81188143ccf9ebe2c2fe5e039cdff3a2df53f7512555673f0468147e991e
          createdAt: "2024-12-06T19:03:24Z"
          description: Cloudera Flow Management - Kubernetes Operator for Apache NiFideploys
            and operates Apache NiFi
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: A valid Cloudera license is required.
            Contact a Cloudera account team to learn more.
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NifiRegistry is the Schema for the nifiregistries API.
            displayName: Nifi Registry
            kind: NifiRegistry
            name: nifiregistries.cfm.cloudera.com
            version: v1alpha1
          - description: Nifi is the Schema for the nifis API.
            displayName: Nifi
            kind: Nifi
            name: nifis.cfm.cloudera.com
            version: v1alpha1
        description: |-
          Cloudera Flow Management - Kubernetes Operator for Apache NiFi brings enterprise-grade NiFi and NiFi Registry deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of NiFi and NiFi Registry for enterprise deployments.
          To learn more about Cloudera Flow Management - Kubernetes Operator for Apache NiFi read this
          [deep dive blog post](https://community.cloudera.com/t5/Community-Articles/Cloudera-Flow-Management-Operator-A-technical-deep-dive/ta-p/390473).

          Cloudera Flow Management - Kubernetes Operator for Apache NiFi consists of and ships multiple components including Apache NiFi, Apache NiFi Registry,
          and a custom sidecar image for logging aggregation.

          ## Installation prerequisites

          Installing Cloudera Flow Management - Kubernetes Operator for Apache NiFi requires the a valid license as well as Cloudera credentials
          (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly.
          The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) from which installation artifacts are pulled.

          ### Obtaining a license
          Contact [Cloudera sales](https://cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets
          Before installation, create two `Secrets`. One containing your Cloudera credentials and one containing your license. Cloudera credentials are provided to you
          together with your license.

          1. Create a `Secret` containing your Cloudera credentials.
              `oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***]
               --docker-server=container.repository.cloudera.com`

          2. Create a `Secret` containing your license.
              `oc create secret generic cfm-operator-license --from-file=license.txt=[***PATH TO LICENSE FILE***]`

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces in which you
            plan to deploy NiFi clusters or NiFi Registry deployments. Cloudera recommends that you create the `Secret` in all required namespaces before installation
            if you know which namespaces you will be using to deploy NiFi clusters or NiFi Registry deployments.

          ## Learn more
          To learn more, visit the [Cloudera Flow Management - Kubernetes Operator for Apache NiFi documentation library](https://docs.cloudera.com/cfm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Flow Management - Kubernetes Operator for Apache NiFi, visit the following resources:
          * [Apache NiFi](https://nifi.apache.org/)
          * [Apache NiFi Registry](https://nifi.apache.org/registry.html)
        displayName: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        - nifiregistry
        - cloudera
        - cfm
        links:
        - name: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
          url: https://docs.cloudera.com/cfm-operator
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s@sha256:914dcde9ee97894fa6f5deec285e53bf81d196e071a997cabb40fedd52deccc3
        - container.repository.cloudera.com/cloudera/cfm-tini@sha256:6493d23e20fca8daf31823403fca3dd9b42da983464fbe788a6863e141e8b66b
        - container.repository.cloudera.com/cloudera/cfm-operator@sha256:890d81188143ccf9ebe2c2fe5e039cdff3a2df53f7512555673f0468147e991e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86af06fb3b6c17b22be47118b0a8c797fa7fdd3c00dbe99a2d2a593ca8fc3d63
        - container.repository.cloudera.com/cloudera/cfm-nifi-k8s@sha256:0d700b7eff096dfc362868345826c3a6460fa43a700273b1422e95015b7727ee
        version: 2.9.0-b96
      entries:
      - name: cfm-operator.v2.9.0-b96
        version: 2.9.0-b96
      name: alpha
    - currentCSV: cfm-operator.v2.10.0-b135
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "Nifi",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifi-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifi",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifi-sample"
                },
                "spec": {
                  "hostName": "nifi.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifi-k8s",
                    "tag": "2.9.0-b96-nifi_1.27.0.2.3.14.0-14"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.9.0-b96"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              },
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "NifiRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifiregistry-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifiregistry",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifiregistry-sample"
                },
                "spec": {
                  "hostName": "registry.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s",
                    "tag": "2.9.0-b96-nifi_1.27.0.2.3.14.0-14"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.9.0-b96"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning,Big Data
          com.redhat.openshift.versions: v4.10
          containerImage: container.repository.cloudera.com/cloudera/cfm-operator@sha256:0312b6dadd5904ea9507bb857ab89d7266bf244c93b86855e0528e72813b7a85
          createdAt: "2025-05-05T14:39:51Z"
          description: Cloudera Flow Management - Kubernetes Operator for Apache NiFideploys
            and operates Apache NiFi
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: A valid Cloudera license is required.
            Contact a Cloudera account team to learn more.
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NifiRegistry is the Schema for the nifiregistries API.
            displayName: Nifi Registry
            kind: NifiRegistry
            name: nifiregistries.cfm.cloudera.com
            version: v1alpha1
          - description: Nifi is the Schema for the nifis API.
            displayName: Nifi
            kind: Nifi
            name: nifis.cfm.cloudera.com
            version: v1alpha1
        description: |-
          Cloudera Flow Management - Kubernetes Operator for Apache NiFi brings enterprise-grade NiFi and NiFi Registry deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of NiFi and NiFi Registry for enterprise deployments.
          To learn more about Cloudera Flow Management - Kubernetes Operator for Apache NiFi read this
          [deep dive blog post](https://community.cloudera.com/t5/Community-Articles/Cloudera-Flow-Management-Operator-A-technical-deep-dive/ta-p/390473).

          Cloudera Flow Management - Kubernetes Operator for Apache NiFi consists of and ships multiple components including Apache NiFi, Apache NiFi Registry,
          and a custom sidecar image for logging aggregation.

          ## Installation prerequisites

          Installing Cloudera Flow Management - Kubernetes Operator for Apache NiFi requires the a valid license as well as Cloudera credentials
          (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly.
          The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) from which installation artifacts are pulled.

          ### Obtaining a license
          Contact [Cloudera sales](https://cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets
          Before installation, create two `Secrets`. One containing your Cloudera credentials and one containing your license. Cloudera credentials are provided to you
          together with your license.

          1. Create a `Secret` containing your Cloudera credentials.
              `oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***]
               --docker-server=container.repository.cloudera.com --namespace [***OPERATOR INSTALL NAMESPACE***]`

          2. Create a `Secret` containing your license.
              `oc create secret generic cfm-operator-license --from-file=license.txt=[***PATH TO LICENSE FILE***] --namespace [***OPERATOR INSTALL NAMESPACE***]`

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces in which you
            plan to deploy NiFi clusters or NiFi Registry deployments. Cloudera recommends that you create the `Secret` in all required namespaces before installation
            if you know which namespaces you will be using to deploy NiFi clusters or NiFi Registry deployments.

          ## Learn more
          To learn more, visit the [Cloudera Flow Management - Kubernetes Operator for Apache NiFi documentation library](https://docs.cloudera.com/cfm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Flow Management - Kubernetes Operator for Apache NiFi, visit the following resources:
          * [Apache NiFi](https://nifi.apache.org/)
          * [Apache NiFi Registry](https://nifi.apache.org/registry.html)
        displayName: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        - nifiregistry
        - cloudera
        - cfm
        links:
        - name: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
          url: https://docs.cloudera.com/cfm-operator
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s@sha256:52e00f102579b972500eb187e64a7662a6a46caa883df743e454b1d5111b34e5
        - container.repository.cloudera.com/cloudera/cfm-tini@sha256:d42ad5f05a9bb283fed6dbbd820855714a31662f5b851e691f13703153e90d13
        - container.repository.cloudera.com/cloudera/cfm-nifi-k8s@sha256:8607c81906422d65784c6dca74a37af3b7c87012bca65ce28a995366014aec86
        - container.repository.cloudera.com/cloudera/cfm-operator@sha256:0312b6dadd5904ea9507bb857ab89d7266bf244c93b86855e0528e72813b7a85
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86af06fb3b6c17b22be47118b0a8c797fa7fdd3c00dbe99a2d2a593ca8fc3d63
        version: 2.10.0-b135
      entries:
      - name: cfm-operator.v2.10.0-b135
        version: 2.10.0-b135
      name: stable
    defaultChannel: stable
    packageName: cfm-operator
    provider:
      name: Cloudera
      url: https://cloudera.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Splunk Inc.
      provider-url: www.splunk.com
    name: splunk-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: splunk-operator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "IndexerCluster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "indexercluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "LicenseMaster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "licensemaster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "MonitoringConsole",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "monitoringconsole-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "SearchHeadCluster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "searchheadcluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "Standalone",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "standalone-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "clustermanager-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "IndexerCluster",
                "metadata": {
                  "name": "indexercluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "LicenseManager",
                "metadata": {
                  "name": "licensemanager-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "MonitoringConsole",
                "metadata": {
                  "name": "monitoringconsole-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "SearchHeadCluster",
                "metadata": {
                  "name": "searchheadcluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "Standalone",
                "metadata": {
                  "name": "standalone-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Big Data, Logging & Tracing, Monitoring, Security, AI/Machine
            Learning
          containerImage: splunk/splunk-operator@sha256:ed02561995d46ff8eea143fa78f5da118ee6fdb500832e9a33b42d059a67e229
          createdAt: "2025-09-08T10:17:59Z"
          description: The Splunk Operator for Kubernetes enables you to quickly and
            easily deploy Splunk Enterprise on your choice of private or public cloud
            provider. The Operator simplifies scaling and management of Splunk Enterprise
            by automating administrative workflows using Kubernetes best practices.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=2.4.0 <3.0.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/splunk/splunk-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterManager is the Schema for the cluster manager API
            displayName: Cluster Manager
            kind: ClusterManager
            name: clustermanagers.enterprise.splunk.com
            version: v4
          - description: ClusterMaster is the Schema for the cluster manager API
            displayName: Cluster Master
            kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v3
          - kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v1
          - kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v2
          - description: IndexerCluster is the Schema for a Splunk Enterprise indexer
              cluster
            displayName: Indexer Cluster
            kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v3
          - description: IndexerCluster is the Schema for a Splunk Enterprise indexer
              cluster
            displayName: Indexer Cluster
            kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v4
          - kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v1
          - kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v2
          - description: LicenseManager is the Schema for a Splunk Enterprise license
              manager.
            displayName: License Manager
            kind: LicenseManager
            name: licensemanagers.enterprise.splunk.com
            version: v4
          - description: LicenseMaster is the Schema for a Splunk Enterprise license
              master.
            displayName: License Master
            kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v3
          - kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v1
          - kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v2
          - description: MonitoringConsole is the Schema for the monitoringconsole
              API
            displayName: Monitoring Console
            kind: MonitoringConsole
            name: monitoringconsoles.enterprise.splunk.com
            version: v3
          - description: MonitoringConsole is the Schema for the monitoringconsole
              API
            displayName: Monitoring Console
            kind: MonitoringConsole
            name: monitoringconsoles.enterprise.splunk.com
            version: v4
          - description: SearchHeadCluster is the Schema for a Splunk Enterprise search
              head cluster
            displayName: Search Head Cluster
            kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v3
          - description: SearchHeadCluster is the Schema for a Splunk Enterprise search
              head cluster
            displayName: Search Head Cluster
            kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v4
          - kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v1
          - kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v2
          - description: Standalone is the Schema for a Splunk Enterprise standalone
              instances.
            displayName: Standalone
            kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v3
          - description: Standalone is the Schema for a Splunk Enterprise standalone
              instances.
            displayName: Standalone
            kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v4
          - kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v1
          - kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v2
        description: The Splunk Operator for Kubernetes enables you to quickly and
          easily deploy Splunk Enterprise on your choice of private or public cloud
          provider. The Operator simplifies scaling and management of Splunk Enterprise
          by automating administrative workflows using Kubernetes best practices.
        displayName: Splunk Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Splunk
        - Enterprise
        - Operator
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://github.com/splunk/splunk-operator/tree/main/docs
        maintainers:
        - email: support@splunk.com
          name: Splunk Operator Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Splunk Inc.
          url: www.splunk.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - splunk/splunk-operator@sha256:ed02561995d46ff8eea143fa78f5da118ee6fdb500832e9a33b42d059a67e229
        - splunk/splunk@sha256:4a5ce5a7b4e84a87396f0664895095c6510a0f2bf52e8747a373c7357f3ee313
        version: 3.0.0
      entries:
      - name: splunk-operator.v3.0.0
        version: 3.0.0
      - name: splunk-operator.v2.8.1
        version: 2.8.1
      - name: splunk-operator.v2.8.0
        version: 2.8.0
      - name: splunk-operator.v2.7.1
        version: 2.7.1
      - name: splunk-operator.v2.7.0
        version: 2.7.0
      name: stable
    defaultChannel: stable
    packageName: splunk-operator
    provider:
      name: Splunk Inc.
      url: www.splunk.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-emrcontainers-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-emrcontainers-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "emrcontainers.services.k8s.aws/v1alpha1",
                "kind": "VirtualCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/emrcontainers-controller:1.1.1
          createdAt: "2025-09-19T19:36:26Z"
          description: AWS EMR on EKS controller is a service controller for managing
            EMR on EKS resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: JobRun represents the state of an AWS emrcontainers JobRun
              resource.
            displayName: JobRun
            kind: JobRun
            name: jobruns.emrcontainers.services.k8s.aws
            version: v1alpha1
          - description: VirtualCluster represents the state of an AWS emrcontainers
              VirtualCluster resource.
            displayName: VirtualCluster
            kind: VirtualCluster
            name: virtualclusters.emrcontainers.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EMR on EKS resources in AWS from within your Kubernetes cluster.

          **About Amazon EMR on EKS**

          Amazon EMR on EKS provides a deployment option for Amazon EMR that allows you to run open-source big data frameworks on Amazon Elastic Kubernetes Service (Amazon EKS). With this deployment option, you can focus on running analytics workloads while Amazon EMR on EKS builds, configures, and manages containers for open-source applications.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EMR on EKS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - emrcontainers
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EMR on EKS Developer Resources
          url: https://docs.aws.amazon.com/emr/latest/EMR-on-EKS-DevelopmentGuide/emr-eks.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: emr on eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/emrcontainers-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-emrcontainers-controller.v1.1.1
        version: 1.1.1
      - name: ack-emrcontainers-controller.v1.1.0
        version: 1.1.0
      - name: ack-emrcontainers-controller.v1.0.27
        version: 1.0.27
      - name: ack-emrcontainers-controller.v1.0.25
        version: 1.0.25
      - name: ack-emrcontainers-controller.v1.0.24
        version: 1.0.24
      - name: ack-emrcontainers-controller.v1.0.23
        version: 1.0.23
      - name: ack-emrcontainers-controller.v1.0.22
        version: 1.0.22
      - name: ack-emrcontainers-controller.v1.0.21
        version: 1.0.21
      - name: ack-emrcontainers-controller.v1.0.18
        version: 1.0.18
      - name: ack-emrcontainers-controller.v1.0.16
        version: 1.0.16
      - name: ack-emrcontainers-controller.v1.0.15
        version: 1.0.15
      - name: ack-emrcontainers-controller.v1.0.14
        version: 1.0.14
      - name: ack-emrcontainers-controller.v1.0.13
        version: 1.0.13
      - name: ack-emrcontainers-controller.v1.0.12
        version: 1.0.12
      - name: ack-emrcontainers-controller.v1.0.11
        version: 1.0.11
      - name: ack-emrcontainers-controller.v1.0.9
        version: 1.0.9
      - name: ack-emrcontainers-controller.v1.0.8
        version: 1.0.8
      - name: ack-emrcontainers-controller.v1.0.4
        version: 1.0.4
      - name: ack-emrcontainers-controller.v1.0.3
        version: 1.0.3
      - name: ack-emrcontainers-controller.v1.0.1
        version: 1.0.1
      - name: ack-emrcontainers-controller.v1.0.0
        version: 1.0.0
      - name: ack-emrcontainers-controller.v0.1.2
        version: 0.1.2
      - name: ack-emrcontainers-controller.v0.1.0
        version: 0.1.0
      - name: ack-emrcontainers-controller.v0.0.10
        version: 0.0.10
      - name: ack-emrcontainers-controller.v0.0.9
        version: 0.0.9
      - name: ack-emrcontainers-controller.v0.0.8
        version: 0.0.8
      - name: ack-emrcontainers-controller.v0.0.7
        version: 0.0.7
      - name: ack-emrcontainers-controller.v0.0.6
        version: 0.0.6
      - name: ack-emrcontainers-controller.v0.0.5
        version: 0.0.5
      - name: ack-emrcontainers-controller.v0.0.4
        version: 0.0.4
      - name: ack-emrcontainers-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-emrcontainers-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/cluster-nfd-operator
    name: nfd
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfd.4.16.0-202507211806
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nfd.openshift.io/v1",
                "kind": "NodeFeatureDiscovery",
                "metadata": {
                  "name": "nfd-instance",
                  "namespace": "openshift-nfd"
                },
                "spec": {
                  "customConfig": {
                    "configData": "#    - name: \"more.kernel.features\"\n#      matchOn:\n#      - loadedKMod: [\"example_kmod3\"]\n#    - name: \"more.features.by.nodename\"\n#      value: customValue\n#      matchOn:\n#      - nodename: [\"special-.*-node-.*\"]\n"
                  },
                  "operand": {
                    "image": "registry.redhat.io/openshift4/ose-node-feature-discovery-rhel9@sha256:bb65ec6a6f88688630305e081cb4bcb88aa538b5f6a3385ee28c61522dcd1662",
                    "imagePullPolicy": "IfNotPresent",
                    "servicePort": 12000
                  },
                  "workerConfig": {
                    "configData": "core:\n#  labelWhiteList:\n#  noPublish: false\n  sleepInterval: 60s\n#  sources: [all]\n#  klog:\n#    addDirHeader: false\n#    alsologtostderr: false\n#    logBacktraceAt:\n#    logtostderr: true\n#    skipHeaders: false\n#    stderrthreshold: 2\n#    v: 0\n#    vmodule:\n##   NOTE: the following options are not dynamically run-time \n##          configurable and require a nfd-worker restart to take effect\n##          after being changed\n#    logDir:\n#    logFile:\n#    logFileMaxSize: 1800\n#    skipLogHeaders: false\nsources:\n#  cpu:\n#    cpuid:\n##     NOTE: whitelist has priority over blacklist\n#      attributeBlacklist:\n#        - \"BMI1\"\n#        - \"BMI2\"\n#        - \"CLMUL\"\n#        - \"CMOV\"\n#        - \"CX16\"\n#        - \"ERMS\"\n#        - \"F16C\"\n#        - \"HTT\"\n#        - \"LZCNT\"\n#        - \"MMX\"\n#        - \"MMXEXT\"\n#        - \"NX\"\n#        - \"POPCNT\"\n#        - \"RDRAND\"\n#        - \"RDSEED\"\n#        - \"RDTSCP\"\n#        - \"SGX\"\n#        - \"SSE\"\n#        - \"SSE2\"\n#        - \"SSE3\"\n#        - \"SSE4.1\"\n#        - \"SSE4.2\"\n#        - \"SSSE3\"\n#      attributeWhitelist:\n#  kernel:\n#    kconfigFile: \"/path/to/kconfig\"\n#    configOpts:\n#      - \"NO_HZ\"\n#      - \"X86\"\n#      - \"DMI\"\n  pci:\n    deviceClassWhitelist:\n      - \"0200\"\n      - \"03\"\n      - \"12\"\n    deviceLabelFields:\n#      - \"class\"\n      - \"vendor\"\n#      - \"device\"\n#      - \"subsystem_vendor\"\n#      - \"subsystem_device\"\n#  usb:\n#    deviceClassWhitelist:\n#      - \"0e\"\n#      - \"ef\"\n#      - \"fe\"\n#      - \"ff\"\n#    deviceLabelFields:\n#      - \"class\"\n#      - \"vendor\"\n#      - \"device\"\n#  custom:\n#    - name: \"my.kernel.feature\"\n#      matchOn:\n#        - loadedKMod: [\"example_kmod1\", \"example_kmod2\"]\n#    - name: \"my.pci.feature\"\n#      matchOn:\n#        - pciId:\n#            class: [\"0200\"]\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#        - pciId :\n#            vendor: [\"8086\"]\n#            device: [\"1000\", \"1100\"]\n#    - name: \"my.usb.feature\"\n#      matchOn:\n#        - usbId:\n#          class: [\"ff\"]\n#          vendor: [\"03e7\"]\n#          device: [\"2485\"]\n#        - usbId:\n#          class: [\"fe\"]\n#          vendor: [\"1a6e\"]\n#          device: [\"089a\"]\n#    - name: \"my.combined.feature\"\n#      matchOn:\n#        - pciId:\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#          loadedKMod : [\"vendor_kmod1\", \"vendor_kmod2\"]\n"
                  }
                }
              },
              {
                "apiVersion": "nfd.openshift.io/v1alpha1",
                "kind": "NodeFeatureRule",
                "metadata": {
                  "name": "my-sample-rule-object"
                },
                "spec": {
                  "rules": [
                    {
                      "labels": {
                        "my-sample-feature": "true"
                      },
                      "matchFeatures": [
                        {
                          "feature": "kernel.loadedmodule",
                          "matchExpressions": {
                            "dummy": {
                              "op": "Exists"
                            }
                          }
                        },
                        {
                          "feature": "kernel.config",
                          "matchExpressions": {
                            "X86": {
                              "op": "In",
                              "value": [
                                "y"
                              ]
                            }
                          }
                        }
                      ],
                      "name": "my sample rule"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery,OpenShift Optional
          containerImage: registry.redhat.io/openshift4/ose-cluster-nfd-rhel9-operator@sha256:c1b6a5b8a59c08458d57c0fc2bb2bc9b056f006d15edb4d3161a3d1c0bc97ae6
          description: |-
            The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

            NFD consists  of the following software components:

            The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

            ##NFD-Master
            NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

            ##NFD-Worker
            NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

            ##NFD-Topology-Updater
            NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.0-202507211806'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-nfd
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/internal-objects: '["noderesourcetopologies.topology.node.k8s.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/cluster-nfd-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              The NodeFeatureDiscovery instance is the CustomResource being watched by the NFD-Operator, and holds all the needed information to setup the behaviour of the master and worker pods
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
          - description: |
              NodeFeatureRule resource specifies a configuration for feature-based customization of node objects, such as node labeling.
            kind: NodeFeatureRule
            name: nodefeaturerules.nfd.openshift.io
            version: v1alpha1
          - kind: NodeFeature
            name: nodefeatures.nfd.openshift.io
            version: v1alpha1
        description: |-
          The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

          NFD consists  of the following software components:

          The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

          ##NFD-Master
          NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

          ##NFD-Worker
          NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

          ##NFD-Topology-Updater
          NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
        displayName: Node Feature Discovery Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - feature-discovery
        - feature-detection
        - node-labels
        links:
        - name: Node Feature Discovery Operator
          url: https://docs.openshift.com/container-platform/4.16/hardware_enablement/psap-node-feature-discovery-operator.html
        - name: Node Feature Discovery Documentation
          url: https://kubernetes-sigs.github.io/node-feature-discovery/stable/get-started/index.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/cluster-nfd-operator
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:bcca89fb000b0fe8722ad7807b62a43ef3ae8fa76db1d790c285a975bfecb1aa
        - registry.redhat.io/openshift4/ose-cluster-nfd-rhel9-operator@sha256:c1b6a5b8a59c08458d57c0fc2bb2bc9b056f006d15edb4d3161a3d1c0bc97ae6
        - registry.redhat.io/openshift4/ose-node-feature-discovery-rhel9@sha256:bb65ec6a6f88688630305e081cb4bcb88aa538b5f6a3385ee28c61522dcd1662
        version: 4.16.0-202507211806
      entries:
      - name: nfd.4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: nfd.4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: nfd.4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: nfd.4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: nfd.4.16.0-202504271139
        version: 4.16.0-202504271139
      - name: nfd.4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: nfd.4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: nfd.4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: nfd.4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: nfd.4.16.0-202501271512
        version: 4.16.0-202501271512
      - name: nfd.4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: nfd.4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: nfd.4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: nfd.4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: nfd.4.16.0-202410251436
        version: 4.16.0-202410251436
      - name: nfd.4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: nfd.4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: nfd.4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: nfd.4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: nfd.4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: nfd.4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: nfd.4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: nfd.4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: nfd.4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: nfd.4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: nfd.4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: nfd.4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: nfd.4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: nfd.4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: nfd
    provider:
      name: Red Hat
      url: https://github.com/openshift/cluster-nfd-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhpam-kogito-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhpam-kogito-operator.7.13.5-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "ruleunit-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogitoinfra-sample"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "kafka.strimzi.io/v1beta2",
                    "kind": "Kafka",
                    "name": "kogito-kafka"
                  }
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator@sha256:19497d4aff0ea3ea114dccb0d0e265458bb14ba6d325353eb82b554547dbd2e4
          description: RHPAM Kogito Operator for deployment and management of Kogito
            services.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Process Automation
            Manager", "Red Hat Decision Manager"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/rhpam-kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.rhpam.kiegroup.org
            version: v1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.rhpam.kiegroup.org
            version: v1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.rhpam.kiegroup.org
            version: v1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.rhpam.kiegroup.org
            version: v1
        description: |-
          RHPAM Kogito Operator is designed for deployment and management of Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Builds a RHPAM Kogito service from a remote git repository and creates a custom image for the given runtime: Red Hat build of Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Deploy RHPAM Kogito service from user provided image.
          * Setup Grafana Dashboard when Grafana Operator is available in cluster.
          * Create Prometheus monitor object when Prometheus operator is available in cluster.

          **IMPORTANT!** RHPAM Kogito Operator may need Grafana and Prometheus Operator to configure monitoring in cases if you do not provide these services beforehand. If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the RHPAM Kogito Operator.
        displayName: RHPAM Kogito Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-process-automation-manager
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_process_automation_manager/7.13/html/getting_started_with_red_hat_build_of_kogito_in_red_hat_process_automation_manager/assembly-deploying-kogito-microservices-on-openshift
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator@sha256:19497d4aff0ea3ea114dccb0d0e265458bb14ba6d325353eb82b554547dbd2e4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9492c67eaa7ab37df6f5fa623c4aac955b7d7ffaab9ae2e57dea745e82907e40
        version: 7.13.5-4
      entries:
      - name: rhpam-kogito-operator.7.13.5-4
        version: 7.13.5-4
      - name: rhpam-kogito-operator.7.13.5-3
        version: 7.13.5-3
      - name: rhpam-kogito-operator.7.13.5-2
        version: 7.13.5-2
      - name: rhpam-kogito-operator.7.13.5-1
        version: 7.13.5-1
      - name: rhpam-kogito-operator.7.13.4-3
        version: 7.13.4-3
      - name: rhpam-kogito-operator.7.13.4-2
        version: 7.13.4-2
      - name: rhpam-kogito-operator.7.13.4-1
        version: 7.13.4-1
      - name: rhpam-kogito-operator.7.13.3-1
        version: 7.13.3-1
      - name: rhpam-kogito-operator.7.13.2-3
        version: 7.13.2-3
      - name: rhpam-kogito-operator.7.13.2-2
        version: 7.13.2-2
      - name: rhpam-kogito-operator.7.13.2-1
        version: 7.13.2-1
      - name: rhpam-kogito-operator.7.13.1-2
        version: 7.13.1-2
      - name: rhpam-kogito-operator.7.13.1-1
        version: 7.13.1-1
      - name: rhpam-kogito-operator.7.13.0-1
        version: 7.13.0-1
      - name: rhpam-kogito-operator.v7.12.0-4
        version: 7.12.0-4
      - name: rhpam-kogito-operator.v7.12.0-3
        version: 7.12.0-3
      - name: rhpam-kogito-operator.v7.12.0-2
        version: 7.12.0-2
      name: 7.x
    defaultChannel: 7.x
    packageName: rhpam-kogito-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://skupper.io
    name: skupper-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-operator.v1.4.8-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus@sha256:65602221cdff7f718d4149f4dead167d05dbbe0cea1eaa8bf4b98be588135583
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:c665677c639699d6a5ef24aed5a1a3d9c0a5ee25f077a6db42a9fc6247d93132
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:eedde9f2019d4e919612bcc18f427c0d88b4846e773751af231ba0e560d48869
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:1fd908e80ef21fe7d7051c400cb099f2f3ca51c8cdd698705ef9784d34688911
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:4c9cc9df83baa1ef37cfb429babe5b86a962b8bb096209b3fb74291c67c94000
        version: 1.4.8-rh-1
      entries:
      - name: skupper-operator.v1.4.8-rh-1
        version: 1.4.8-rh-1
      - name: skupper-operator.v1.4.7-rh-3
        version: 1.4.7-rh-3
      - name: skupper-operator.v1.4.7-rh-2
        version: 1.4.7-rh-2
      - name: skupper-operator.v1.4.7-rh-1
        version: 1.4.7-rh-1
      - name: skupper-operator.v1.4.5-rh-4
        version: 1.4.5-rh-4
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1
    - currentCSV: skupper-operator.v1.4.8-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:eedde9f2019d4e919612bcc18f427c0d88b4846e773751af231ba0e560d48869
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:1fd908e80ef21fe7d7051c400cb099f2f3ca51c8cdd698705ef9784d34688911
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:4c9cc9df83baa1ef37cfb429babe5b86a962b8bb096209b3fb74291c67c94000
        - registry.redhat.io/openshift4/ose-prometheus@sha256:65602221cdff7f718d4149f4dead167d05dbbe0cea1eaa8bf4b98be588135583
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:c665677c639699d6a5ef24aed5a1a3d9c0a5ee25f077a6db42a9fc6247d93132
        version: 1.4.8-rh-1
      entries:
      - name: skupper-operator.v1.4.8-rh-1
        version: 1.4.8-rh-1
      - name: skupper-operator.v1.4.7-rh-3
        version: 1.4.7-rh-3
      - name: skupper-operator.v1.4.7-rh-2
        version: 1.4.7-rh-2
      - name: skupper-operator.v1.4.7-rh-1
        version: 1.4.7-rh-1
      - name: skupper-operator.v1.4.5-rh-4
        version: 1.4.5-rh-4
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.4
    - currentCSV: skupper-operator.v1.5.5-rh-6
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:9e9e1f2288f0218e3ba2abf4bebac9731d7cf9f9c567e0f6de6767b0b66cbee4
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:a5d7c1b2b9db672548126c0873d062a55667d93291fd3e32809d3c9c900c548e
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:9e9e1f2288f0218e3ba2abf4bebac9731d7cf9f9c567e0f6de6767b0b66cbee4
        - registry.redhat.io/openshift4/ose-prometheus@sha256:53afaf53975807765fc3777314edd218c789c17680c1f57a8e33a8f9a76f38bf
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:3e78ec8cdb729a704d5a26e54697f285ba7c0b9afee6160b7ecc703876120dea
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:be9dbcd2a3a1e1e536921390dcfd283fc45dc4423d2705376ef37a9fb53f7cf1
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:1de071dabf620265252f3922c632f17f676ba716ce4c3df931a0ccb3871dbb6c
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:5f8621d0013febcecbed4d7c82db11a4dded99ad8ce6b6d83858b9bdbe79ee5a
        version: 1.5.5-rh-6
      entries:
      - name: skupper-operator.v1.5.5-rh-6
        version: 1.5.5-rh-6
      - name: skupper-operator.v1.5.5-rh-5
        version: 1.5.5-rh-5
      - name: skupper-operator.v1.5.5-rh-4
        version: 1.5.5-rh-4
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.5
    - currentCSV: skupper-operator.v1.8.7-rh-2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:b0d43628b779a101a24e4b271318a34c21f3b71bcba2bf20804f941a93d27455
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.8.7-rh-2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:56b8fde75a6e7df235c560a0f4b9114398da5ba8b20f75f1ea7338f48ce19506
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:d7d252cff08ba1589afd8707ed349f87d3b6214906a4fb49932bda0a4d19fee4
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:5ea36e96f878577104ab7601e57e7a2a233f11dbdc533454f95e2e14a066d21f
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:3adeef827eca0382cd410f7c3e99a0fee9b50791cb0e0c6fc130b72a2550684f
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:b0d43628b779a101a24e4b271318a34c21f3b71bcba2bf20804f941a93d27455
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:f798745e799515065af099a816057b9c98a741e951d30a0f45ed47ca50e6724e
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:61542c5f0989f51ac5d6c96f5d0e29aca75614b79e7626299c77c5a74636b564
        version: 1.8.7-rh-2
      entries:
      - name: skupper-operator.v1.8.7-rh-2
        version: 1.8.7-rh-2
      - name: skupper-operator.v1.8.7-rh-1
        version: 1.8.7-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.8
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.9
    - currentCSV: skupper-operator.v2.1.1-rh-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
          createdAt: "2025-08-19T20:43:19Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:36872037066fae806cbf88cab673229f25e592295972aa18227827233362eec5
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:c01d0eab0841b1a5fddb58a0ae110beb464b91390f02662590bebc7b2a7213b6
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e8df53454cf12a14338790af6ed209798010990ca3320096e8b44acf901c4123
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:1a290a5ac8b9d1b082e740054675fe432aa424c2f55dd662caf67c740044f229
        version: 2.1.1-rh-3
      entries:
      - name: skupper-operator.v2.1.1-rh-3
        version: 2.1.1-rh-3
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2
    - currentCSV: skupper-operator.v2.0.1-rh-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:b5c73dc59a9184cd120c2ee35256cc097c790b8ef488af0c732d5efe3192b768
          createdAt: "2025-05-29T13:35:40Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.0.1-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:92f44cff160874efe57a985609e599b13c49c0c4cef846dccc80833af2ae74f3
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:e239c859a1f49816ae3068455e703b8a12919f11158b0f6ee0e6bfb31e8a8082
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:f2a41674ed196c089748d6386c18c14580e502de33235ef23c37eddeb6350f41
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:95062ff1256135cfa17907e82b09b2a211371ea0e76ccc711b7c737e040d87e9
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:498b5bb648e5bbaad07663941600173abac7d60a10744305f619509ddf7563e0
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:bd3f72544841c91c2d573a65ac3c988756f3523e17e0988027793b0f5f3ec3d1
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:b5c73dc59a9184cd120c2ee35256cc097c790b8ef488af0c732d5efe3192b768
        version: 2.0.1-rh-2
      entries:
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2.0
    - currentCSV: skupper-operator.v2.1.1-rh-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
          createdAt: "2025-08-19T20:43:19Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:c01d0eab0841b1a5fddb58a0ae110beb464b91390f02662590bebc7b2a7213b6
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e8df53454cf12a14338790af6ed209798010990ca3320096e8b44acf901c4123
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:1a290a5ac8b9d1b082e740054675fe432aa424c2f55dd662caf67c740044f229
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:36872037066fae806cbf88cab673229f25e592295972aa18227827233362eec5
        version: 2.1.1-rh-3
      entries:
      - name: skupper-operator.v2.1.1-rh-3
        version: 2.1.1-rh-3
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2.1
    defaultChannel: stable-2
    packageName: skupper-operator
    provider:
      name: Red Hat
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: placement-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: placement-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "placement.openstack.org/v1beta1",
                "kind": "PlacementAPI",
                "metadata": {
                  "name": "placement"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "placement",
                  "databaseInstance": "openstack",
                  "preserveJobs": false,
                  "replicas": 1,
                  "secret": "placement-secret",
                  "serviceUser": "placement"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:43Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PlacementAPI is the Schema for the placementapis API
            displayName: Placement API
            kind: PlacementAPI
            name: placementapis.placement.openstack.org
            version: v1beta1
        description: Placement Operator
        displayName: Placement Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Placement
        links:
        - name: Placement Operator
          url: https://github.com/openstack-k8s-operators/placement-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/placement-rhel9-operator@sha256:d3be48618b1f0f717aea162ae41876137fee8b35329d2e3e66e9d4eaccf62376
        - registry.redhat.io/rhoso/openstack-placement-api-rhel9@sha256:3308f2eccfbe7416a85443696273417bfb816df36b810ebbefa108dda12a5b19
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: placement-operator.v1.0.6
        version: 1.0.6
      - name: placement-operator.v1.0.5
        version: 1.0.5
      - name: placement-operator.v1.0.4
        version: 1.0.4
      - name: placement-operator.v1.0.3
        version: 1.0.3
      - name: placement-operator.v1.0.2
        version: 1.0.2
      - name: placement-operator.v1.0.1
        version: 1.0.1
      - name: placement-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: placement-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/stolostron/multicluster-global-hub
    name: multicluster-global-hub-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-global-hub-operator.v1.4.0-dev.20250302
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2025-02-14T08:24:47Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: planning
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/grafana:2.12.0-SNAPSHOT-2024-09-03-21-11-25
        - quay.io/stolostron/multicluster-global-hub-operator:v1.4.0
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-relations/spicedb-operator:latest
        - quay.io/stolostron/multicluster-global-hub-manager:v1.4.0
        - quay.io/stolostron/postgresql-16:9.5-1732622748
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-inventory/inventory-api:latest
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-relations/spicedb:latest
        - quay.io/stolostron/multicluster-global-hub-agent:v1.4.0
        - quay.io/prometheuscommunity/postgres-exporter:v0.15.0
        version: 1.4.0-dev.20250302
      entries:
      - name: multicluster-global-hub-operator.v1.4.0-dev.20250302
        version: 1.4.0-dev.20250302
      - name: multicluster-global-hub-operator.v1.4.0-dev.20250225
        version: 1.4.0-dev.20250225
      - name: multicluster-global-hub-operator.v1.3.0-dev.20240911
        version: 1.3.0-dev.20240911
      - name: multicluster-global-hub-operator.v1.3.0-dev.20240724
        version: 1.3.0-dev.20240724
      - name: multicluster-global-hub-operator.v1.2.0-dev.20240516
        version: 1.2.0-dev.20240516
      - name: multicluster-global-hub-operator.v1.2.0-dev
        version: 1.2.0-dev
      - name: multicluster-global-hub-operator.v1.1.0-dev
        version: 1.1.0-dev
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: multicluster-global-hub-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-operator-globalhub-1-5@sha256:4db1251323bd64930461cb5025da41baffcf5c00c51ea4e49940b7130971b510
          createdAt: "2025-06-10T06:58:35Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/postgres-exporter-globalhub-1-5@sha256:1108572a8bdafde926254f1ff94c2b22e048706be660b07c793e9f9e038a0f4a
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-operator-globalhub-1-5@sha256:4db1251323bd64930461cb5025da41baffcf5c00c51ea4e49940b7130971b510
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-manager-globalhub-1-5@sha256:01253426d65f1297a1955a8393051e6a4b1df5e5249774fc5189a75f35e7d644
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-agent-globalhub-1-5@sha256:6f7520eac83abec7780f831d8473544d0b8db43fc6e77d59e59317b09eb1f6c1
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/glo-grafana-globalhub-1-5@sha256:73fdf00538e63595b7cc92d484e087ff815579a95d7259c579c6e7dc1beacff6
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        version: 1.5.0
      entries:
      - name: multicluster-global-hub-operator.v1.5.0
        version: 1.5.0
      - name: multicluster-global-hub-operator.v1.4.0
        version: 1.4.0
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-1.5
    - currentCSV: multicluster-global-hub-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2024-03-05T02:13:54Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.32.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:latest
        version: 1.1.0
      entries:
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.10
    - currentCSV: multicluster-global-hub-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2024-07-03T06:52:02Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v1.2.0
        version: 1.2.1
      entries:
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.11
    - currentCSV: multicluster-global-hub-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator@sha256:636aea861da80fc36fb27a419a18c5b3fbb1bc072e8ae33c75e1cc47e2c05f75
          createdAt: "2024-10-10T05:42:48Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          kafka-zookeeper-0                                       1/1     Running   0          21m
          kafka-zookeeper-1                                       1/1     Running   0          21m
          kafka-zookeeper-2                                       1/1     Running   0          21m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgres-0                      2/2     Running   0          22m
          ```
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:636aea861da80fc36fb27a419a18c5b3fbb1bc072e8ae33c75e1cc47e2c05f75
        - quay.io/prometheuscommunity/postgres-exporter@sha256:386b12d19eab2a37d7cd8ca8b4c7491cc7a830d9581f49af6c98a393da9605e6
        - quay.io/stolostron/inventory-api@sha256:a15e62ee070dba5d74cf336b3b89b467593c5a33cfe3623d922f0af4fd7df392
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/release-globalhub-1-3/multicluster-global-hub-manager-globalhub-1-3@sha256:7656d01a4d7b39c61c3249695d68011d9e0df308627b9efd117686e901281ab1
        - quay.io/stolostron/multicluster-global-hub-manager@sha256:715cb5d0c6c407ec5c0303fc7e6d83fc77ce7e06775e169a475f7c075f1e7000
        - quay.io/stolostron/grafana@sha256:719b320a05d10158fadce5ae2fc6e4254a18957a6608b6f81e876cf46d3a6897
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        version: 1.3.0
      entries:
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.12
    - currentCSV: multicluster-global-hub-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator@sha256:085e9fc6cdc6a42df10a38856c8ccff25e5acffaba67ef95da72027e90e507fc
          createdAt: "2025-02-14T08:24:47Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-inventory/inventory-api@sha256:6a55a1919ecd92d94e05c12be1b6450956f5fda2ce17af7027cac3806c4970a4
        - quay.io/prometheuscommunity/postgres-exporter@sha256:386b12d19eab2a37d7cd8ca8b4c7491cc7a830d9581f49af6c98a393da9605e6
        - quay.io/stolostron/multicluster-global-hub-manager@sha256:4cd602ff6f8dd951848dc8a95c856004295c9f4710f77d4dafcc0735c1492395
        - quay.io/stolostron/multicluster-global-hub-agent@sha256:8dfc33dfed33a85733ee5e02dcd4f05662c75263791e0961ba17cbb3d0c340c9
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:085e9fc6cdc6a42df10a38856c8ccff25e5acffaba67ef95da72027e90e507fc
        - quay.io/stolostron/grafana@sha256:719b320a05d10158fadce5ae2fc6e4254a18957a6608b6f81e876cf46d3a6897
        version: 1.4.0
      entries:
      - name: multicluster-global-hub-operator.v1.4.0
        version: 1.4.0
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.13
    - currentCSV: multicluster-global-hub-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "storage": {
                    "name": "storage-secret"
                  },
                  "transport": {
                    "name": "transport-secret"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.5.0
          createdAt: "2022-08-23T20:00:31Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha1
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.5 or later needs to be installed.\n- PostgreSQL needs to be installed
          and you must create database for the multicluster global hub. You must create
          a secret named postgresql-secret that contains the database access credential
          in `open-cluster-management` namespace. You can run the following command
          to create the secret: \n\n```\nkubectl create secret generic storage-secret
          -n open-cluster-management \\\n  --from-literal=database_uri=<postgresql-uri>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)
          to install postgres in `hoh-postgres` namespace and create the secret `storage-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n- Kafka needs to be installed,
          and you must create two sections, `spec` and `status`. You need to create
          a secret named kafka-secret that contains the kafka access information should
          be created in `open-cluster-management` namespace. You can run the following
          command to create the secret: \n\n```\nkubectl create secret generic transport-secret
          -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n  --from-literal=CA=<CA-for-kafka-server>\n```\n >_Note:_ There is a
          sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `transport-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n## How to Install\n\nInstall the
          Multicluster Global Hub Operator by following the instructions that are
          displayed after you select the Install button. The operator must be installed
          in the same namespace as Red Hat Advanced Cluster Management for Kubernetes
          (RHACM), by default, in `open-cluster-management` namespace.\n\nA pod will
          be created in `open-cluster-management` namespace\n\n```\n$ kubectl get
          pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          19s\n```\n\nThe operator is
          now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installing the operator,
          create an instance of the MulticlusterGlobalHub resource to install the
          Multicluster Global Hub.\n\nThe pods of Multicluster Global Hub are created
          in the `open-cluster-management` namespace:\n\n```\n$ kubectl get pods -n
          open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\nmulticluster-global-hub-rbac-77888b7dc4-z9vbp
          \                   1/1     Running   0          10s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.5.0
        version: 0.5.0
      entries:
      - name: multicluster-global-hub-operator.v0.5.0
        version: 0.5.0
      name: release-2.6
    - currentCSV: multicluster-global-hub-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha2",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "largeScale": {
                      "kafka": {
                        "name": "transport-secret"
                      },
                      "postgres": {
                        "name": "storage-secret"
                      }
                    },
                    "type": "largeScale"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.6.0
          createdAt: "2023-02-09T20:00:31Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha2
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.7 or later needs to be installed.\n- PostgreSQL needs to be installed
          and you must create database for the multicluster global hub. You must create
          a secret named postgresql-secret that contains the database access credential
          in `open-cluster-management` namespace. You can run the following command
          to create the secret: \n\n```\nkubectl create secret generic storage-secret
          -n open-cluster-management \\\n  --from-literal=database_uri=<postgresql-uri>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)
          to install postgres in `hoh-postgres` namespace and create the secret `storage-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n- Kafka needs to be installed,
          and you must create two sections, `spec` and `status`. You need to create
          a secret named kafka-secret that contains the kafka access information should
          be created in `open-cluster-management` namespace. You can run the following
          command to create the secret: \n\n```\nkubectl create secret generic transport-secret
          -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n  --from-literal=CA=<CA-for-kafka-server>\n```\n >_Note:_ There is a
          sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `transport-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n## How to Install\n\nInstall the
          Multicluster Global Hub Operator by following the instructions that are
          displayed after you select the Install button. The operator must be installed
          in the same namespace as Red Hat Advanced Cluster Management for Kubernetes
          (RHACM), by default, in `open-cluster-management` namespace.\n\nA pod will
          be created in `open-cluster-management` namespace\n\n```\n$ kubectl get
          pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          19s\n```\n\nThe operator is
          now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installing the operator,
          create an instance of the MulticlusterGlobalHub resource to install the
          Multicluster Global Hub.\n\nThe pods of Multicluster Global Hub are created
          in the `open-cluster-management` namespace:\n\n```\n$ kubectl get pods -n
          open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: multicluster-global-hub-operator.v0.6.0
        version: 0.6.0
      name: release-2.7
    - currentCSV: multicluster-global-hub-operator.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha3",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "largeScale": {
                      "kafka": {
                        "transportFormat": "cloudEvents"
                      }
                    },
                    "type": "largeScale"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.7.0
          createdAt: "2023-07-19T07:53:07Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha3
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.7 or later needs to be installed.\n## How to Install\n\nInstall the Multicluster
          Global Hub Operator by following the instructions that are displayed after
          you select the Install button. The operator must be installed in the same
          namespace as Red Hat Advanced Cluster Management for Kubernetes (RHACM),
          by default, in `open-cluster-management` namespace.\n\nA pod will be created
          in `open-cluster-management` namespace\n\n```\n$ kubectl get pods -n open-cluster-management\nNAME
          \                                                            READY   STATUS
          \   RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp                1/1
          \    Running   0          19s\n```\n\nThe operator is now providing new
          Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installed the operator,
          create an instance of the MulticlusterGlobalHub resource to instantiate
          the Multicluster Global Hub.\n\nYou need to prepare:\n\n- PostgreSQL needs
          to be installed and you must create database for the multicluster global
          hub. You must create a secret named multicluster-global-hub-storage that
          contains the database access credential in `open-cluster-management` namespace.
          You can run the following command to create the secret: \n\n```\nkubectl
          create secret generic multicluster-global-hub-storage -n open-cluster-management
          \\\n  --from-literal=database_uri=<postgresql-uri> \\\n --from-file=ca.crt=<CA-for-postgres-server>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)(Note:
          the client version of kubectl must be v1.21+) to install postgres in `hoh-postgres`
          namespace and create the secret `multicluster-global-hub-storage` in namespace
          `open-cluster-management` automatically. To override the secret namespace,
          set `TARGET_NAMESPACE` environment variable to the RHACM installation namespace
          before executing the script.\n\n- Kafka needs to be installed, and you must
          create three topics, `spec` `status` and `event`. You need to create a secret
          named multicluster-global-hub-transport that contains the kafka access information
          should be created in `open-cluster-management` namespace. You can run the
          following command to create the secret: \n\n```\nkubectl create secret generic
          multicluster-global-hub-transport -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n      --from-file=ca.crt=<CA-cert-for-kafka-server> --from-file=client.crt=<Client-cert-for-kafka-server>
          --from-file=client.key=<Client-key-for-kafka-server> \n```\n >_Note:_ There
          is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `multicluster-global-hub-transport`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\nOnce an instance of the MulticlusterGlobalHub
          is created, the following pods are created in the `open-cluster-management`
          namespace:\n\n```\n$ kubectl get pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-grafana-79f8d59f58-qwcjl
          \                1/1     Running   0          22s\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: beta
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.7.0
        version: 0.7.0
      entries:
      - name: multicluster-global-hub-operator.v0.7.0
        version: 0.7.0
      name: release-2.8
    - currentCSV: multicluster-global-hub-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v1.0.0
          createdAt: "2023-10-16T03:07:33Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the MultiCluster Global Hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.7 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```

          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v1.0.0
        version: 1.0.0
      entries:
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.9
    defaultChannel: release-1.5
    packageName: multicluster-global-hub-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/stolostron/multicluster-global-hub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhbk-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhbk-operator.v22.0.13-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
          createdAt: "2024-09-17T11:20:22Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:884afe0ca1698d85a02aa47b6d60b8222e50d70ad280db6f98e1a4bbdbe13c40
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
        version: 22.0.13-opr.1
      entries:
      - name: rhbk-operator.v22.0.13-opr.1
        version: 22.0.13-opr.1
      - name: rhbk-operator.v22.0.12-opr.1
        version: 22.0.12-opr.1
      - name: rhbk-operator.v22.0.11-opr.2
        version: 22.0.11-opr.2
      - name: rhbk-operator.v22.0.11-opr.1
        version: 22.0.11-opr.1
      - name: rhbk-operator.v22.0.10-opr.1
        version: 22.0.10-opr.1
      - name: rhbk-operator.v22.0.9-opr.1
        version: 22.0.9-opr.1
      - name: rhbk-operator.v22.0.8-opr.2
        version: 22.0.8-opr.2
      - name: rhbk-operator.v22.0.8-opr.1
        version: 22.0.8-opr.1
      - name: rhbk-operator.v22.0.7-opr.1
        version: 22.0.7-opr.1
      - name: rhbk-operator.v22.0.6-opr.1
        version: 22.0.6-opr.1
      name: stable-v22
    - currentCSV: rhbk-operator.v22.0.13-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
          createdAt: "2024-09-17T11:20:22Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:884afe0ca1698d85a02aa47b6d60b8222e50d70ad280db6f98e1a4bbdbe13c40
        version: 22.0.13-opr.1
      entries:
      - name: rhbk-operator.v22.0.13-opr.1
        version: 22.0.13-opr.1
      - name: rhbk-operator.v22.0.12-opr.1
        version: 22.0.12-opr.1
      - name: rhbk-operator.v22.0.11-opr.2
        version: 22.0.11-opr.2
      - name: rhbk-operator.v22.0.11-opr.1
        version: 22.0.11-opr.1
      - name: rhbk-operator.v22.0.10-opr.1
        version: 22.0.10-opr.1
      - name: rhbk-operator.v22.0.9-opr.1
        version: 22.0.9-opr.1
      - name: rhbk-operator.v22.0.8-opr.2
        version: 22.0.8-opr.2
      - name: rhbk-operator.v22.0.8-opr.1
        version: 22.0.8-opr.1
      - name: rhbk-operator.v22.0.7-opr.1
        version: 22.0.7-opr.1
      - name: rhbk-operator.v22.0.6-opr.1
        version: 22.0.6-opr.1
      name: stable-v22.0
    - currentCSV: rhbk-operator.v24.0.11-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
          createdAt: "2025-03-26T02:27:51Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:d339ecf788d1681ba9945a2ea63ca675caec863ec75f497ef916103f4c898003
        version: 24.0.11-opr.1
      entries:
      - name: rhbk-operator.v24.0.11-opr.1
        version: 24.0.11-opr.1
      - name: rhbk-operator.v24.0.10-opr.1
        version: 24.0.10-opr.1
      - name: rhbk-operator.v24.0.9-opr.1
        version: 24.0.9-opr.1
      - name: rhbk-operator.v24.0.8-opr.1
        version: 24.0.8-opr.1
      - name: rhbk-operator.v24.0.7-opr.1
        version: 24.0.7-opr.1
      - name: rhbk-operator.v24.0.6-opr.1
        version: 24.0.6-opr.1
      - name: rhbk-operator.v24.0.5-opr.1
        version: 24.0.5-opr.1
      - name: rhbk-operator.v24.0.4-opr.1
        version: 24.0.4-opr.1
      - name: rhbk-operator.v24.0.3-opr.1
        version: 24.0.3-opr.1
      name: stable-v24
    - currentCSV: rhbk-operator.v24.0.11-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
          createdAt: "2025-03-26T02:27:51Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:d339ecf788d1681ba9945a2ea63ca675caec863ec75f497ef916103f4c898003
        version: 24.0.11-opr.1
      entries:
      - name: rhbk-operator.v24.0.11-opr.1
        version: 24.0.11-opr.1
      - name: rhbk-operator.v24.0.10-opr.1
        version: 24.0.10-opr.1
      - name: rhbk-operator.v24.0.9-opr.1
        version: 24.0.9-opr.1
      - name: rhbk-operator.v24.0.8-opr.1
        version: 24.0.8-opr.1
      - name: rhbk-operator.v24.0.7-opr.1
        version: 24.0.7-opr.1
      - name: rhbk-operator.v24.0.6-opr.1
        version: 24.0.6-opr.1
      - name: rhbk-operator.v24.0.5-opr.1
        version: 24.0.5-opr.1
      - name: rhbk-operator.v24.0.4-opr.1
        version: 24.0.4-opr.1
      - name: rhbk-operator.v24.0.3-opr.1
        version: 24.0.3-opr.1
      name: stable-v24.0
    - currentCSV: rhbk-operator.v26.2.9-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
          createdAt: "2025-09-18T15:03:55Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:50a62b46af44ce0ff7e6c3a15e1bfb2a9611ba4a8b65b8d0ab65817064692c55
        version: 26.2.9-opr.1
      entries:
      - name: rhbk-operator.v26.2.9-opr.1
        version: 26.2.9-opr.1
      - name: rhbk-operator.v26.2.8-opr.1
        version: 26.2.8-opr.1
      - name: rhbk-operator.v26.2.7-opr.1
        version: 26.2.7-opr.1
      - name: rhbk-operator.v26.2.6-opr.1
        version: 26.2.6-opr.1
      - name: rhbk-operator.v26.2.5-opr.1
        version: 26.2.5-opr.1
      - name: rhbk-operator.v26.2.4-opr.2
        version: 26.2.4-opr.2
      - name: rhbk-operator.v26.2.4-opr.1
        version: 26.2.4-opr.1
      - name: rhbk-operator.v26.0.15-opr.1
        version: 26.0.15-opr.1
      - name: rhbk-operator.v26.0.14-opr.1
        version: 26.0.14-opr.1
      - name: rhbk-operator.v26.0.13-opr.1
        version: 26.0.13-opr.1
      - name: rhbk-operator.v26.0.12-opr.1
        version: 26.0.12-opr.1
      - name: rhbk-operator.v26.0.11-opr.1
        version: 26.0.11-opr.1
      - name: rhbk-operator.v26.0.10-opr.1
        version: 26.0.10-opr.1
      - name: rhbk-operator.v26.0.9-opr.1
        version: 26.0.9-opr.1
      - name: rhbk-operator.v26.0.8-opr.1
        version: 26.0.8-opr.1
      - name: rhbk-operator.v26.0.7-opr.1
        version: 26.0.7-opr.1
      - name: rhbk-operator.v26.0.6-opr.1
        version: 26.0.6-opr.1
      - name: rhbk-operator.v26.0.5-opr.1
        version: 26.0.5-opr.1
      name: stable-v26
    - currentCSV: rhbk-operator.v26.0.15-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:fc5fc8cfb2c100c5b3a33fc11a78dfca569f5f34f45d88439340392d2c60afb2
          createdAt: "2025-08-28T19:34:30Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:fc5fc8cfb2c100c5b3a33fc11a78dfca569f5f34f45d88439340392d2c60afb2
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:4bb4d3070fe118bdd7de24e8b68133b8b6e158cea48c9f1abad2b4297be0dfdb
        version: 26.0.15-opr.1
      entries:
      - name: rhbk-operator.v26.0.15-opr.1
        version: 26.0.15-opr.1
      - name: rhbk-operator.v26.0.14-opr.1
        version: 26.0.14-opr.1
      - name: rhbk-operator.v26.0.13-opr.1
        version: 26.0.13-opr.1
      - name: rhbk-operator.v26.0.12-opr.1
        version: 26.0.12-opr.1
      - name: rhbk-operator.v26.0.11-opr.1
        version: 26.0.11-opr.1
      - name: rhbk-operator.v26.0.10-opr.1
        version: 26.0.10-opr.1
      - name: rhbk-operator.v26.0.9-opr.1
        version: 26.0.9-opr.1
      - name: rhbk-operator.v26.0.8-opr.1
        version: 26.0.8-opr.1
      - name: rhbk-operator.v26.0.7-opr.1
        version: 26.0.7-opr.1
      - name: rhbk-operator.v26.0.6-opr.1
        version: 26.0.6-opr.1
      - name: rhbk-operator.v26.0.5-opr.1
        version: 26.0.5-opr.1
      name: stable-v26.0
    - currentCSV: rhbk-operator.v26.2.9-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
          createdAt: "2025-09-18T15:03:55Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:50a62b46af44ce0ff7e6c3a15e1bfb2a9611ba4a8b65b8d0ab65817064692c55
        version: 26.2.9-opr.1
      entries:
      - name: rhbk-operator.v26.2.9-opr.1
        version: 26.2.9-opr.1
      - name: rhbk-operator.v26.2.8-opr.1
        version: 26.2.8-opr.1
      - name: rhbk-operator.v26.2.7-opr.1
        version: 26.2.7-opr.1
      - name: rhbk-operator.v26.2.6-opr.1
        version: 26.2.6-opr.1
      - name: rhbk-operator.v26.2.5-opr.1
        version: 26.2.5-opr.1
      - name: rhbk-operator.v26.2.4-opr.2
        version: 26.2.4-opr.2
      - name: rhbk-operator.v26.2.4-opr.1
        version: 26.2.4-opr.1
      name: stable-v26.2
    defaultChannel: stable-v26.2
    packageName: rhbk-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://horreum.hyperfoil.io
    name: horreum-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: horreum-operator.v0.7.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hyperfoil.io/v1alpha1",
                "kind": "Horreum",
                "metadata": {
                  "name": "horreum"
                },
                "spec": {
                  "nodeHost": "127.0.0.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/hyperfoil/horreum-operator:0.7.9
          createdAt: "2023-06-13T10:42:39Z"
          description: Performance results repository
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/hyperfoil/horreum-operator
          support: Red Hat Application Runtimes Performance Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Horreum is the object configuring Horreum performance results
              repository
            displayName: Horreum
            kind: Horreum
            name: horreums.hyperfoil.io
            version: v1alpha1
        description: Performance results repository
        displayName: Horreum
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - performance
        - repository
        - database
        - results
        - hyperfoil
        - benchmark
        links:
        - name: Horreum Operator
          url: https://horreum-operator.domain
        maintainers:
        - email: rvansa@redhat.com
          name: Radim Vansa
        - email: spederse@redhat.com
          name: Ståle W. Pedersen
        - email: wreicher@redhat.com
          name: Will Reichert
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat, Inc.
          url: https://horreum.hyperfoil.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/hyperfoil/horreum-operator:0.7.9
        version: 0.7.9
      entries:
      - name: horreum-operator.v0.7.9
        version: 0.7.9
      - name: horreum-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: horreum-operator
    provider:
      name: Red Hat, Inc.
      url: https://horreum.hyperfoil.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Automotive SIG CentOS
      provider-url: ""
    name: automotive-infra
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: automotive-infra.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "ContainerBuild",
                "metadata": {
                  "name": "cb-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "ImageBuild",
                "metadata": {
                  "name": "imagebuild-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "OperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "automotive-infra"
                  },
                  "name": "operatorconfig-sample",
                  "namespace": "automotive-infra-operator"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/fedora-sig-robotics/infra-ci-operator:v0.0.10
          createdAt: "2025-03-31T14:06:33Z"
          description: Streamline CI/CD for RHIVOS and its ecosystem using OpenShift
            Pipelines
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: automotive-infra-operator
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator
          support: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Client
            name: clients.jumpstarter.dev
            version: v1alpha1
          - kind: ContainerBuild
            name: containerbuilds.automotive.sig.centos.org
            version: v1
          - kind: ExporterAccessPolicy
            name: exporteraccesspolicies.jumpstarter.dev
            version: v1alpha1
          - kind: Exporter
            name: exporters.jumpstarter.dev
            version: v1alpha1
          - kind: ImageBuild
            name: imagebuilds.automotive.sig.centos.org
            version: v1
          - kind: Lease
            name: leases.jumpstarter.dev
            version: v1alpha1
          - kind: OperatorConfig
            name: operatorconfigs.automotive.sig.centos.org
            version: v1
        description: This operator is designed to streamline CI/CD for RHIVOS and
          its ecosystem using OpenShift Pipelines
        displayName: Automotive-CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Automotive
        - RHIVOS
        - Pipelines
        - CI
        links:
        - name: Automotive infra Operator
          url: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator
        maintainers:
        - email: automotive-sig@lists.centos.org
          name: Automotive SIG CentOS Community
        maturity: alpha
        provider:
          name: Automotive SIG CentOS
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/fedora-sig-robotics/infra-ci-operator:v0.0.10
        version: 0.0.10
      entries:
      - name: automotive-infra.v0.0.10
        version: 0.0.10
      - name: automotive-infra.v0.0.9
        version: 0.0.9
      - name: automotive-infra.v0.0.8
        version: 0.0.8
      - name: automotive-infra.v0.0.7
        version: 0.0.7
      - name: automotive-infra.v0.0.6
        version: 0.0.6
      - name: automotive-infra.v0.0.5
        version: 0.0.5
      - name: automotive-infra.v0.0.4
        version: 0.0.4
      - name: automotive-infra.v0.0.3
        version: 0.0.3
      - name: automotive-infra.v0.0.2
        version: 0.0.2
      - name: automotive-infra.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: automotive-infra
    provider:
      name: Automotive SIG CentOS
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: ibm-licensing-operator
      app.kubernetes.io/managed-by: ibm-licensing-operator
      app.kubernetes.io/name: ibm-licensing
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      intent: projected
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-licensing-operator-app
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-licensing-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "IBMLicensing",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-licensing-operator",
                    "app.kubernetes.io/managed-by": "ibm-licensing-operator",
                    "app.kubernetes.io/name": "ibm-licensing"
                  },
                  "name": "instance"
                },
                "spec": {
                  "version": "4.2.14",
                  "apiSecretToken": "ibm-licensing-token",
                  "datasource": "datacollector",
                  "httpsEnable": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
          createdAt: "2025-02-25T12:16:11Z"
          description: The IBM Licensing Operator provides a Kubernetes CRD-Based
            API to monitor the license usage of products.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-licensing-operator-app
          olm.skipRange: '>=1.0.0 <4.2.14'
          operatorframework.io/suggested-namespace: ibm-licensing
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/internal-objects: '["ibmlicensingmetadatas.operator.ibm.com",
            "ibmlicensingdefinitions.operator.ibm.com", "ibmlicensingquerysources.operator.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'IBMLicensing custom resource is used to create an instance
              of the License Service, used to collect information about license usage
              of IBM containerized products and IBM Cloud Paks per cluster. You can
              retrieve license usage data through a dedicated API call and generate
              an audit snapshot on demand. Documentation: For additional details regarding
              install parameters check: https://ibm.biz/icpfs39install. License: Please
              refer to the IBM Terms website (ibm.biz/lsvc-lic) to find the license
              terms for the particular IBM product for which you are deploying this
              component.'
            displayName: IBM License Service
            kind: IBMLicensing
            name: ibmlicensings.operator.ibm.com
            version: v1alpha1
          - kind: IBMLicensingDefinition
            name: ibmlicensingdefinitions.operator.ibm.com
            version: v1
          - description: 'IBMLicensingMetadata is the schema for IBM License Service.
              Thanks to IBMLicensingMetadata, you can track the license usage of Virtual
              Processor Core (VPC) metric by Pods that are managed by automation,
              for which licensing annotations are not defined at the time of deployment,
              such as the open source OpenLiberty. / For more information, see documentation:
              https://ibm.biz/icpfs39install. By installing the operator for IBM License
              Service, you accept the license terms: https://ibm.biz/icpfs39license.'
            displayName: IBMLicensing Metadata
            kind: IBMLicensingMetadata
            name: ibmlicensingmetadatas.operator.ibm.com
            version: v1alpha1
          - kind: IBMLicensingQuerySource
            name: ibmlicensingquerysources.operator.ibm.com
            version: v1
        description: |-
          **Important:**
          - If you are using the IBM Licensing Operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For the link to your IBM Cloud Pak documentation, see [IBM Cloud Paks that use Common Services](https://ibm.biz/BdyGwb).
          - If you are using a stand-alone IBM Container Software, you can use the IBM Licensing Operator directly. For more information, see [ibm-licensing-operator for stand-alone IBM Containerized Software](https://ibm.biz/BdyGwh).

          **IBM Licensing Operator overview**

          IBM Licensing Operator installs License Service. You can use License Service to collect information about license usage of IBM Containerized products and IBM Cloud Paks per cluster. You can retrieve license usage data through a dedicated API call and generate an audit snapshot on demand.

          **Supported platforms**

           Red Hat OpenShift Container Platform 4.10 or newer installed on Linux x86_64, Linux on Power (ppc64le), Linux on IBM Z and LinuxONE.

          **Prerequisites**

           Prerequisites depend on the integration of the License Service with an IBM Cloud Pak or IBM Containerized Software. For more information, see the applicable IBM Cloud Pak documentation or [ibm-licensing-operator for stand-alone IBM Containerized Software](https://ibm.biz/BdyGwh).
        displayName: IBM Licensing
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBMLicensing
        - IBM
        - Cloud
        - License
        - Service
        - License Service
        links:
        - name: GitHub
          url: https://github.com/IBM/ibm-licensing-operator
        maintainers:
        - email: talk2sam@us.ibm.com
          name: talk2sam
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:ec137115210dae9b69cb2899ed8a5437d94c8a658ae1ae5eb295232b34b0925e
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:96e6db93b2d16ba590252d6568b68f254548aaed73ec17e4bdd8d530ea668be6
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9186c350fa54e8b792d72e47be6b54a67b610f726175c9641795010d45ad51f9
        version: 4.2.14
      entries:
      - name: ibm-licensing-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-licensing-operator-app
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator-rhmp.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/dynatrace-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/dynatrace-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator-rhmp.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator-rhmp.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator-rhmp.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator-rhmp.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator-rhmp.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator-rhmp.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator-rhmp.v1.5.0
        version: 1.5.0
      - name: dynatrace-operator-rhmp.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator-rhmp.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator-rhmp.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator-rhmp
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-opensearchserverless-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-opensearchserverless-controller.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/opensearchserverless-controller:0.2.1
          createdAt: "2025-09-20T21:43:17Z"
          description: AWS OpenSearchServerless controller is a service controller
            for managing OpenSearchServerless resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Collection represents the state of an AWS opensearchserverless
              Collection resource.
            displayName: Collection
            kind: Collection
            name: collections.opensearchserverless.services.k8s.aws
            version: v1alpha1
          - description: SecurityPolicy represents the state of an AWS opensearchserverless
              SecurityPolicy resource.
            displayName: SecurityPolicy
            kind: SecurityPolicy
            name: securitypolicies.opensearchserverless.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon OpenSearchServerless resources in AWS from within your Kubernetes cluster.

          **About Amazon OpenSearchServerless**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon OpenSearchServerless
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opensearchserverless
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon OpenSearchServerless Developer Resources
          url: https://aws.amazon.com/OpenSearchServerless/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: opensearchserverless maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/opensearchserverless-controller:0.2.1
        version: 0.2.1
      entries:
      - name: ack-opensearchserverless-controller.v0.2.1
        version: 0.2.1
      - name: ack-opensearchserverless-controller.v0.2.0
        version: 0.2.0
      - name: ack-opensearchserverless-controller.v0.1.2
        version: 0.1.2
      - name: ack-opensearchserverless-controller.v0.1.1
        version: 0.1.1
      - name: ack-opensearchserverless-controller.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ack-opensearchserverless-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-framework
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-framework.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Framework",
                "metadata": {
                  "name": "framework-sample"
                },
                "spec": {
                  "kcos-eula": {
                    "image": {
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    },
                    "service": {
                      "nodePort": 0,
                      "type": "ClusterIP"
                    }
                  },
                  "kcos-logging-framework": {
                    "image": {
                      "remoteRepository": "docker.io",
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    },
                    "ingress": {
                      "annotations": {
                        "nginx.ingress.kubernetes.io/auth-method": "POST",
                        "nginx.ingress.kubernetes.io/auth-response-headers": "X-AUTH-USER",
                        "nginx.ingress.kubernetes.io/auth-url": "https://127.0.0.1/token/validate",
                        "nginx.ingress.kubernetes.io/cors-allow-credentials": "true",
                        "nginx.ingress.kubernetes.io/cors-allow-methods": "PUT, GET, POST, OPTIONS, PATCH",
                        "nginx.ingress.kubernetes.io/cors-allow-origin": "https://admin.example.com,https://localhost:4200",
                        "nginx.ingress.kubernetes.io/enable-cors": "true",
                        "nginx.ingress.kubernetes.io/proxy-body-size": "0",
                        "nginx.ingress.kubernetes.io/ssl-redirect": "true"
                      },
                      "enabled": true
                    }
                  },
                  "kcos-postgresql": {
                    "jobCopySecret": {
                      "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                    },
                    "postgresql": {
                      "image": {
                        "registry": "prod.harbor.keysight.digital/loadcore"
                      },
                      "initdbScripts": {
                        "init_script.sh": "#!/bin/sh\nexport PGUSER=postgres\npsql -c \"SELECT 1 FROM pg_database WHERE datname = 'db'\" | grep -q 1 || psql -c \"CREATE DATABASE db\"\npsql -c \"SELECT 1 FROM pg_database WHERE datname = 'kcoslicensingdb'\" | grep -q 1 || psql -c \"CREATE DATABASE kcoslicensingdb\"\n"
                      },
                      "metrics": {
                        "image": {
                          "registry": "docker.io"
                        }
                      },
                      "persistence": {
                        "storageClass": "managed-loadcore-storage"
                      },
                      "pgHbaConfiguration": "local all all trust\nhost all all 0.0.0.0/0 trust",
                      "securityContext": {
                        "enabled": false
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": false
                        }
                      },
                      "volumePermissions": {
                        "image": {
                          "registry": "docker.io"
                        },
                        "securityContext": {
                          "runAsUser": "auto"
                        }
                      }
                    }
                  },
                  "kcos-system-diagnostics": {
                    "environment": {
                      "ELASTICSEARCH_ENDPOINT": "elasticsearch.keysight-wap.svc.cluster.local.",
                      "IGNORE_DATA_PREFIX": "wireless_stats"
                    },
                    "image": {
                      "initContainer": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf",
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    }
                  },
                  "subcharts": {
                    "enable": {
                      "kcos-coredns-fixup": false,
                      "kcos-eula": true,
                      "kcos-local-storage": false,
                      "kcos-logging-framework": true,
                      "kcos-logging-rotate": false,
                      "kcos-naas": false,
                      "kcos-postgresql": true,
                      "kcos-shell": false,
                      "kcos-system-diagnostics": true,
                      "kcos-vital-service": false,
                      "nfs-server-provisioner": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:27:28Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Framework
            name: frameworks.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Framework Operator
        displayName: Keysight KCOS Framework Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Framework
          url: https://keysight-kcos-framework.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-framework@sha256:531f278def85e33d453479fce74127e12e3ed62a251872eab421c6561bedd0c5
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        version: 0.1.21
      entries:
      - name: keysight-kcos-framework.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-framework
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx-openshift-operator.v0.0.36
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
          createdAt: "2019-07-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX on JBoss EAP 7
            displayName: Joget DX on JBoss EAP 7
            kind: JogetDX
            name: jogetdx.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX7/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
        - registry.connect.redhat.com/joget/joget-dx7-eap7@sha256:7245501f5fccf011d75c034b7d55542cfe3bd2c64cc7e48b3224a48b623af282
        version: 0.0.36
      entries:
      - name: joget-dx-openshift-operator.v0.0.36
        version: 0.0.36
      - name: joget-dx-openshift-operator.v0.0.35
        version: 0.0.35
      - name: joget-dx-openshift-operator.v0.0.34
        version: 0.0.34
      - name: joget-dx-openshift-operator.v0.0.33
        version: 0.0.33
      - name: joget-dx-openshift-operator.v0.0.32
        version: 0.0.32
      - name: joget-dx-openshift-operator.v0.0.31
        version: 0.0.31
      - name: joget-dx-openshift-operator.v0.0.30
        version: 0.0.30
      - name: joget-dx-openshift-operator.v0.0.29
        version: 0.0.29
      - name: joget-dx-openshift-operator.v0.0.28
        version: 0.0.28
      - name: joget-dx-openshift-operator.v0.0.27
        version: 0.0.27
      - name: joget-dx-openshift-operator.v0.0.26
        version: 0.0.26
      - name: joget-dx-openshift-operator.v0.0.25
        version: 0.0.25
      - name: joget-dx-openshift-operator.v0.0.24
        version: 0.0.24
      - name: joget-dx-openshift-operator.v0.0.23
        version: 0.0.23
      - name: joget-dx-openshift-operator.v0.0.22
        version: 0.0.22
      - name: joget-dx-openshift-operator.v0.0.21
        version: 0.0.21
      - name: joget-dx-openshift-operator.v0.0.20
        version: 0.0.20
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Habana Labs Ltd.
      provider-url: https://habana.ai
    name: habana-ai-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: habana-ai-operator.v1.19.2-32-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.19.2-32"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.19.2"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.19.2-32"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.19.2-32"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.19.2-32"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.19.2-32"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.19.2-32"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.19.2-32"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.19.2-32"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.19.2-32"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.19.2-32"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:975ed2135061d3bdc145c83abaa15f99e4c9f6efb15c708de41aa00fbcb72333
          createdAt: "2025-02-10T07:33:30Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.19.2-32-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi AI accelerator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:d728a91a36503f07dfacd49fdbfefe8083607ca33babb5f2dfc1e9de05386454
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:f05cd78d7593b877ec8a75175a9faa72d9b991d81c74af9ecd92208fde296eca
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:170a15041fc3e3901b4e8e0b9a32d53fde11026776c17a6560d500dacb976c56
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:71d4c7181930c71c0ca4e6b696ea416a687eec6d798a43e854fcd1bfdc224a0c
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:373b9caa27a96bb7979f7a6aec246795142ce3a28c6b97a8902035ec20814fa2
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:975ed2135061d3bdc145c83abaa15f99e4c9f6efb15c708de41aa00fbcb72333
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:72cc75353615369d12a07f5488bcdc041a8435906ca6e16c53b001016108e099
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:eeb4573de20d3341a085c75e669b7d37c424e348421d7916440fbbfff12ea482
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:ac16752b53df69573da257d3a3a717a214b2eb1656779a7ef94084d74363fb46
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:b12802e7ef05273fceaef3a9987c3c4ae760714e6f333363f665eb26be9175e9
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:e667695b845cd2be887f117dac66861d2c5e74def009d2939559ae7c30574dc0
        version: 1.19.2-32-1
      entries:
      - name: habana-ai-operator.v1.19.2-32-1
        version: 1.19.2-32-1
      name: "1.19"
    - currentCSV: habana-ai-operator.v1.20.0-543-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.20.0-543"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.20.0-4"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.20.0-543-2"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.20.0-543"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.20.0-543"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.20.0-543-4"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.20.0-543"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.20.0-543-4"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:d5b37eafe48446e1ffa492e195f4519b77d4a11f6750e5c71f27c1536bdfd2cc
          createdAt: "2025-02-26T09:03:44Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.20.0-543-5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:957702b58ee66d7339a20b2a5ece783606d6ccebde627c58951d5ecd6bb1b13f
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:d14c457106b6133fc41ef727ac40a6bf2f0e5c97f21659969d1daebdc1310b43
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:d5b37eafe48446e1ffa492e195f4519b77d4a11f6750e5c71f27c1536bdfd2cc
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:317765eac4e0332a0c4f6bead5563b343a8c22e6e0eaa62276ff66be08608c30
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:68d0f44170cbf3ea15b793e0503fb07ce7e32488011fc85ccb6c9234b3b9302f
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:c5beac6209d0671d3a49b63f1cf95401055b042430eafaefc3859b5c633f0e84
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:c662de244c5900c89b918b16982298676d9a6d553360a5716174b28d4d64a635
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:a673900f08c38c9bb40c90b3cddea580afcc6a99195d816cd0b5cae9831ca5c9
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:3ea80aa8994e8b19fbbdcd18ba2c35800774da3ca4e27a2514d0f5f9676786fa
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:9ab54a84a00e8afc078eec2a7d643f09b1bd674e0e230733fbce7108e392e009
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:cba03259b33c4cb9bcd22622ab0dda962846ea0a80c727779365cf55129ad308
        version: 1.20.0-543-5
      entries:
      - name: habana-ai-operator.v1.20.0-543-5
        version: 1.20.0-543-5
      name: "1.20"
    - currentCSV: habana-ai-operator.v1.21.5-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.21.5-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.21.5-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.21.5-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:0741da99140d137474b9e7edba921672eab8ff6277e1dd8bf3a4179e87f01dd4
          createdAt: "2025-09-17T11:55:32Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.21.5-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:a216ec6a6e7df50e067c557e9ed5871a3308b2ba42a4e41c9642f0c2c4503d88
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:d7f64d15c379524597aa2b2c5b4586c890442fa32356692013cf665b01b0f092
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:256cd8bebec3ed835e187e925a14a6b605ce9b3a17fe7d8886e0b32cde163d90
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:c95d3ffd6c91be3418415740258d33a8dbcedad6fe7d6bfe68d4dd122027447b
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:9692d079ef18a364ed235c187482f329b3a5764a773e363acb1238044ad4a884
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:c84e997c7d8c22c4d4e902a29e45e5364f0ad306f17d03a7d8ddb4391142e167
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:0741da99140d137474b9e7edba921672eab8ff6277e1dd8bf3a4179e87f01dd4
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:c7147359dcd6a2e5adee95259d37743ff302b1deaeb13b184044f4029d33f44a
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:a0b45ccfb6af54b1043ccc2336ad3d7990cc447af5ba3621f3eb1953cd40ffc5
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:bc3749ff153fff71e252820cbf22fcd0753b195b2204f9731e42af29fb7ddaa0
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:1e488c8ad31784dab0fe70144f5146d5809e2fc7c86424d1b512ab887787e28f
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:bd57e7059467f0ca1667d19a70bd9a6ad96b83bbcb53c09fdad5a262fee76af9
        version: 1.21.5-6
      entries:
      - name: habana-ai-operator.v1.21.5-6
        version: 1.21.5-6
      name: "1.21"
    - currentCSV: habana-ai-operator.v1.22.1-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.22.1-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
          createdAt: "2025-09-17T10:15:21Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.22.1-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:d1974d72774c626e77707e8f44bcc631ac142bfe2a5395714969a964c963e36e
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:53c1844b5e8eb08d0cc29865a732100208c5dfc0fe55df8b32e589e319bcab6c
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:e34a6820a58e7e8090b521781c76acc3867c94f8291634de7b138f928e0bdb74
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:1925586fda34c6fa444a59b1fa7be20b3cf16e845bf56566ad2a41c276b0cfe9
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:7140877b23291a5db3e5e28def3c9d5e7cb7be62f046e2fb7ac45eff369cea84
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:c9ff6f3fe24ef87c7bfc71bdb1cd7900941df52f9ad1293badd7e14783ce7ef6
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:1e15f3de21c2b1e20f54d1f8a973290c6d940624506880c5bc3eb880011bcc02
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:8f413edfff1978b35ba3998ed5f1c6293e6764b1d2cd9faa08f0bd81dc6e3e10
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:392fcd2a4730299e6e3bab9f7e96fc3b710de8adb374a2c64b7316e097ed05f9
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:38c21eae6249e5571921998660418e59870bc0880e7b8c9b95a4a8db30ef2e7d
        - vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer@sha256:2cf151c2a518e3e1c7c09efb8301e469d266d1f030fff2e6180b086cc2e04ee7
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:704b4bebdfe6b85693df4b213f67862859a554c21ce4e614736172f64b541ea7
        version: 1.22.1-6
      entries:
      - name: habana-ai-operator.v1.22.1-6
        version: 1.22.1-6
      name: "1.22"
    - currentCSV: habana-ai-operator.v1.22.1-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.22.1-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
          createdAt: "2025-09-17T10:15:21Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.22.1-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:704b4bebdfe6b85693df4b213f67862859a554c21ce4e614736172f64b541ea7
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:8f413edfff1978b35ba3998ed5f1c6293e6764b1d2cd9faa08f0bd81dc6e3e10
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:7140877b23291a5db3e5e28def3c9d5e7cb7be62f046e2fb7ac45eff369cea84
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:38c21eae6249e5571921998660418e59870bc0880e7b8c9b95a4a8db30ef2e7d
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:d1974d72774c626e77707e8f44bcc631ac142bfe2a5395714969a964c963e36e
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:e34a6820a58e7e8090b521781c76acc3867c94f8291634de7b138f928e0bdb74
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:1925586fda34c6fa444a59b1fa7be20b3cf16e845bf56566ad2a41c276b0cfe9
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:c9ff6f3fe24ef87c7bfc71bdb1cd7900941df52f9ad1293badd7e14783ce7ef6
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:53c1844b5e8eb08d0cc29865a732100208c5dfc0fe55df8b32e589e319bcab6c
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:392fcd2a4730299e6e3bab9f7e96fc3b710de8adb374a2c64b7316e097ed05f9
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:1e15f3de21c2b1e20f54d1f8a973290c6d940624506880c5bc3eb880011bcc02
        - vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer@sha256:2cf151c2a518e3e1c7c09efb8301e469d266d1f030fff2e6180b086cc2e04ee7
        version: 1.22.1-6
      entries:
      - name: habana-ai-operator.v1.22.1-6
        version: 1.22.1-6
      name: stable
    defaultChannel: stable
    packageName: habana-ai-operator
    provider:
      name: Habana Labs Ltd.
      url: https://habana.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-multicluster-orchestrator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-multicluster-orchestrator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "items": [
                    {
                      "clusterName": "local-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    },
                    {
                      "clusterName": "spoke-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    }
                  ],
                  "manageS3": false,
                  "schedulingIntervals": [
                    "5m"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["odf-multicluster-console"]'
          createdAt: "2025-09-12T07:33:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          Orchestrator for OpenShift Data Foundation clusters running across multiple OpenShift clusters.
          It uses Red Hat Advanced Cluster Management for Kubernetes as the multicluster control plane.
        displayName: ODF Multicluster Orchestrator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: ODF Multicluster Orchestrator
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: uchapaga@redhat.com
          name: Umanga Chapagain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-multicluster-rhel9-operator@sha256:09e2c7a20ab92ddc5d589072bb0ab9ce2ed8706ed8364f1b11192d7a7396e138
        - registry.redhat.io/odf4/odf-multicluster-console-rhel9@sha256:7527f4a14aed0386b8f160fa8a92b7ff1f58114970e07c3ec6c433d2a53459b5
        version: 4.15.20-rhodf
      entries:
      - name: odf-multicluster-orchestrator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odf-multicluster-orchestrator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odf-multicluster-orchestrator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odf-multicluster-orchestrator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odf-multicluster-orchestrator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odf-multicluster-orchestrator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odf-multicluster-orchestrator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odf-multicluster-orchestrator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odf-multicluster-orchestrator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odf-multicluster-orchestrator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odf-multicluster-orchestrator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odf-multicluster-orchestrator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odf-multicluster-orchestrator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odf-multicluster-orchestrator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odf-multicluster-orchestrator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odf-multicluster-orchestrator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odf-multicluster-orchestrator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odf-multicluster-orchestrator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odf-multicluster-orchestrator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odf-multicluster-orchestrator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odf-multicluster-orchestrator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "items": [
                    {
                      "clusterName": "local-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    },
                    {
                      "clusterName": "spoke-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    }
                  ],
                  "manageS3": false,
                  "schedulingIntervals": [
                    "5m"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["odf-multicluster-console"]'
          createdAt: "2025-09-12T06:32:35Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          Orchestrator for OpenShift Data Foundation clusters running across multiple OpenShift clusters.
          It uses Red Hat Advanced Cluster Management for Kubernetes as the multicluster control plane.
        displayName: ODF Multicluster Orchestrator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: ODF Multicluster Orchestrator
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: uchapaga@redhat.com
          name: Umanga Chapagain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-multicluster-rhel9-operator@sha256:78d6bc14b95d1ee91ed529422062ec191afc489d6ab4e75906ae3c7fe58749cb
        - registry.redhat.io/odf4/odf-multicluster-console-rhel9@sha256:97d3645560388f6ae08fc12cafb58b6c766e38d83de04b645b3e6fb03fbba231
        version: 4.16.16-rhodf
      entries:
      - name: odf-multicluster-orchestrator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odf-multicluster-orchestrator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odf-multicluster-orchestrator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odf-multicluster-orchestrator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odf-multicluster-orchestrator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odf-multicluster-orchestrator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odf-multicluster-orchestrator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odf-multicluster-orchestrator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odf-multicluster-orchestrator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odf-multicluster-orchestrator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odf-multicluster-orchestrator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odf-multicluster-orchestrator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odf-multicluster-orchestrator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odf-multicluster-orchestrator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odf-multicluster-orchestrator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odf-multicluster-orchestrator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odf-multicluster-orchestrator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odf-multicluster-orchestrator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Argo CD Community
      provider-url: ""
    name: argocd-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: argocd-operator.v0.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd-sample"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCDExport",
                "metadata": {
                  "name": "argocdexport-sample"
                },
                "spec": {
                  "argocd": "argocd-sample"
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd-sample"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/argoprojlabs/argocd-operator@sha256:223ad82fb697471ede79201c4a042115c9ca5c00a172d5bdcd5a780e58c46c79
          createdAt: "2025-09-11T06:29:33Z"
          description: Argo CD is a declarative, GitOps continuous delivery tool for
            Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/argoproj-labs/argocd-operator
          support: Argo CD
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: An ApplicationSet is a group or set of Application resources.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: ArgoCDExport is the Schema for the argocdexports API
            displayName: Argo CDExport
            kind: ArgoCDExport
            name: argocdexports.argoproj.io
            version: v1alpha1
          - description: ArgoCD is the Schema for the argocds API
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: ArgoCD is the Schema for the argocds API
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: NotificationsConfiguration is the Schema for the notificationsconfiguration
              API
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
        description: |
          ## Overview

          The Argo CD Operator manages the full lifecycle for [Argo CD](https://argoproj.github.io/argo-cd/) and it's
          components. The operator's goal is to automate the tasks required when operating an Argo CD cluster.

          Beyond installation, the operator helps to automate the process of upgrading, backing up and restoring as needed and
          remove the human as much as possible. In addition, the operator aims to provide deep insights into the Argo CD
          environment by configuring Prometheus and Grafana to aggregate, visualize and expose the metrics already exported by
          Argo CD.

          The operator aims to provide the following, and is a work in progress.

          * Easy configuration and installation of the Argo CD components with sane defaults to get up and running quickly.
          * Provide seamless upgrades to the Argo CD components.
          * Ability to back up and restore an Argo CD cluster from a point in time or on a recurring schedule.
          * Aggregate and expose the metrics for Argo CD and the operator itself using Prometheus and Grafana.
          * Autoscale the Argo CD components as necessary to handle variability in demand.

          ## Usage

          Deploy a basic Argo CD cluster by creating a new ArgoCD resource in the namespace where the operator is installed.

          ```
          apiVersion: argoproj.io/v1alpha1
          kind: ArgoCD
          metadata:
            name: example-argocd
          spec: {}
          ```

          ## Backup

          Backup the cluster above by creating a new ArgoCDExport resource in the namespace where the operator is installed.

          ```
          apiVersion: argoproj.io/v1alpha1
          kind: ArgoCDExport
          metadata:
            name: example-argocdexport
          spec:
            argocd: example-argocd
          ```

          See the [documentation](https://argocd-operator.readthedocs.io) and examples on
          [GitHub](https://github.com/argoproj-labs/argocd-operator) for more information.
        displayName: Argo CD
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - kubernetes
        links:
        - name: Argo CD Project
          url: https://argoproj.github.io/argo-cd/
        - name: Operator Documentation
          url: https://argocd-operator.readthedocs.io
        - name: Operator Source Code
          url: https://github.com/argoproj-labs/argocd-operator
        maintainers:
        - email: aveerama@redhat.com
          name: Abhishek Veeramalla
        maturity: alpha
        provider:
          name: Argo CD Community
        relatedImages:
        - quay.io/argoprojlabs/argocd-operator@sha256:223ad82fb697471ede79201c4a042115c9ca5c00a172d5bdcd5a780e58c46c79
        version: 0.15.0
      entries:
      - name: argocd-operator.v0.15.0
        version: 0.15.0
      - name: argocd-operator.v0.14.1
        version: 0.14.1
      - name: argocd-operator.v0.14.0
        version: 0.14.0
      - name: argocd-operator.v0.13.0
        version: 0.13.0
      - name: argocd-operator.v0.12.0
        version: 0.12.0
      - name: argocd-operator.v0.11.0
        version: 0.11.0
      - name: argocd-operator.v0.10.1
        version: 0.10.1
      - name: argocd-operator.v0.10.0
        version: 0.10.0
      - name: argocd-operator.v0.9.1
        version: 0.9.1
      - name: argocd-operator.v0.9.0
        version: 0.9.0
      - name: argocd-operator.v0.8.0
        version: 0.8.0
      - name: argocd-operator.v0.7.0
        version: 0.7.0
      - name: argocd-operator.v0.5.0
        version: 0.5.0
      - name: argocd-operator.v0.4.0
        version: 0.4.0
      - name: argocd-operator.v0.3.0
        version: 0.3.0
      - name: argocd-operator.v0.2.1
        version: 0.2.1
      - name: argocd-operator.v0.2.0
        version: 0.2.0
      - name: argocd-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: argocd-operator
    provider:
      name: Argo CD Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: oadp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oadp-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "Velero",
                "metadata": {
                  "name": "example-velero",
                  "namespace":"openshift-adp"
                },
                "spec": {
                  "backupStorageLocations": [
                    {
                      "config": {
                        "profile": "default",
                        "region": "us-east-1"
                      },
                      "name": "default",
                      "objectStorage": {
                        "bucket": "myBucket",
                        "prefix": "velero"
                      },
                      "provider": "aws"
                    }
                  ],
                  "defaultVeleroPlugins": [
                    "aws",
                    "csi",
                    "openshift"
                  ],
                  "veleroFeatureFlags": [
                    "EnableCSI"
                  ],
                  "enableRestic": true,
                  "volumeSnapshotLocations": [
                    {
                      "config": {
                        "profile": "default",
                        "region": "us-west-1"
                      },
                      "name": "default",
                      "provider": "aws"
                    }
                  ]
                }
               },
                {
                  "apiVersion": "velero.io/v1",
                  "kind": "Backup",
                  "metadata": {
                    "name": "backup",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "BackupStorageLocation",
                  "metadata": {
                    "name": "backupstoragelocation",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "DeleteBackupRequest",
                  "metadata": {
                    "name": "deletebackuprequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "DownloadRequest",
                  "metadata": {
                    "name": "downloadrequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "PodVolumeBackup",
                  "metadata": {
                    "name": "podvolumebackup",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "PodVolumeRestore",
                  "metadata": {
                    "name": "podvolumerestore",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "ResticRepository",
                  "metadata": {
                    "name": "resticrepository",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "Restore",
                  "metadata": {
                    "name": "restore",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "Schedule",
                  "metadata": {
                    "name": "schedule",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "ServerStatusRequest",
                  "metadata": {
                    "name": "serverstatusrequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "VolumeSnapshotLocation",
                  "metadata": {
                    "name": "volumesnapshotlocation",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/konveyor/oadp-operator@sha256:ecb8534fbc245e890b9c7766946d01e1a8b4ce9be359c4ada4c95fd908b679b3
          createdAt: "2020-09-08T12:21:00Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          olm.skipRange: '>=0.0.0 <0.4.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Velero is a tool to safely backup and restore, perform disaster
              recovery and migrate Kubernetes cluster resources and persistent volumes.
            displayName: Velero
            kind: Velero
            name: veleroes.oadp.openshift.io
            version: v1alpha1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - description: Backup is a Velero resource that respresents the capture
              of Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Velero on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://oadp.konveyor.io/
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/konveyor/velero-restic-restore-helper@sha256:c4d12829e42cabe70b662a659875900677dfb415f912f931c3bf4c09a26aed5b
        - quay.io/konveyor/oadp-operator@sha256:ecb8534fbc245e890b9c7766946d01e1a8b4ce9be359c4ada4c95fd908b679b3
        - quay.io/konveyor/velero-plugin-for-aws@sha256:2197c6fef32910747526d33c1462b75cb7940aeace65791d0a5ddaa038480e52
        - quay.io/konveyor/velero-plugin-for-microsoft-azure@sha256:62a344e5cf196e55dae92e537648987d758ccc75985bc45e9be6d0646a9e64d0
        - quay.io/konveyor/velero-plugin-for-gcp@sha256:3300458b1d6c1514dca7d24353ad03bd92db396b2a29f7d3522563160a5f5219
        - quay.io/konveyor/velero-plugin-for-csi@sha256:5c2261effc902d8baa16e38265f9ad1fbe5284a3e6abd2bbc97761c161fa92cf
        - quay.io/konveyor/velero@sha256:e7865cc707664eb6ee767e77ccd933420604c2562736b53e16c8043a51df9d86
        - quay.io/konveyor/openshift-velero-plugin@sha256:aec8b4fc66cfcebd5a35ce99d8c985453ebf25c11453d30256089a8b425990b5
        - quay.io/konveyor/registry@sha256:b74ab3278b79eb89f99b6970c6aa0f5306221a8495b5496faad6cab2f74ff2cd
        version: 0.4.2
      entries:
      - name: oadp-operator.v0.4.2
        version: 0.4.2
      - name: oadp-operator.v0.4.1
        version: 0.4.1
      - name: oadp-operator.v0.4.0
        version: 0.4.0
      - name: oadp-operator.v0.3.0
        version: 0.3.0
      name: beta
    - currentCSV: oadp-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "restic": {
                      "enable": true
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ResticRepository",
                "metadata": {
                  "name": "resticrepository",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/konveyor/oadp-operator@sha256:90609044d52173619ea1e63ddef0ef3e7c1c939ac737b64bd427cbed5374dd27
          createdAt: "2020-09-08T12:21:00Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          olm.skipRange: '>=0.0.0 <0.5.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is a Velero resource that respresents the capture
              of Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Data Protection Applications on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://oadp.konveyor.io/
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/konveyor/oadp-operator@sha256:90609044d52173619ea1e63ddef0ef3e7c1c939ac737b64bd427cbed5374dd27
        - quay.io/konveyor/velero-plugin-for-aws@sha256:1e8bc27d6b60000e1406085d582910cb530af3fb901ef9bc780628df6c766288
        - quay.io/konveyor/velero-plugin-for-csi@sha256:8c36ccfd294dafc415a7f113770cbbd66dbf6ecf8fef3bc4e73adc9878c606b0
        - quay.io/konveyor/registry@sha256:1be970879206258335357e3cf630138c24b8a5ccab35aae05eccb260e5e8ffb4
        - quay.io/konveyor/velero-restic-restore-helper@sha256:c16248d027c09cd5b7360ae38484df88ca1cc01f9e470e3eade95076ee1a6b84
        - quay.io/konveyor/openshift-velero-plugin@sha256:221ab0bcd1d95fc166ef7319d187e196b59d2fbf9eefd29d91bd6435d45b5c3c
        - quay.io/konveyor/velero-plugin-for-microsoft-azure@sha256:2d0089cbe715b482fb42e1acd9add70b975a8575636d8b143e6f90cef317ea13
        - quay.io/konveyor/velero-plugin-for-gcp@sha256:ee463ed2a8019c560961b118f8a87b587ed302f24fd25480c5d4ddad2557e92f
        - quay.io/konveyor/velero@sha256:0ea5336dcbf0a1fab561697b24da01a0e5b533ee23553fbfd5f8e002f5a31c3e
        version: 0.5.6
      entries:
      - name: oadp-operator.v0.5.6
        version: 0.5.6
      - name: oadp-operator.v0.5.5
        version: 0.5.5
      - name: oadp-operator.v0.5.4
        version: 0.5.4
      - name: oadp-operator.v0.5.3
        version: 0.5.3
      - name: oadp-operator.v0.5.2
        version: 0.5.2
      - name: oadp-operator.v0.5.1
        version: 0.5.1
      - name: oadp-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: oadp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-access-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-access-operator.v23.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyAccess",
                "metadata": {
                  "name": "isva-sample"
                },
                "spec": {
                  "image": "icr.io/isva/verify-access-wrp:10.0.5.0",
                  "instance": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isva/verify-access-operator:23.3.0
          createdAt: Mon Mar  20 00:59:34 UTC 2023
          description: The IBM Security Verify Access Operator manages the lifecycle
            of IBM Security Verify Access worker containers.
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM-Security/verify-access-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyAccess is the Schema for the ibmsecurityverifyaccesses
              API.
            displayName: IBMSecurity Verify Access
            kind: IBMSecurityVerifyAccess
            name: ibmsecurityverifyaccesses.ibm.com
            version: v1
        description: |+
          In a world of highly fragmented access management environments, [IBM Security Verify Access](https://www.ibm.com/au-en/products/verify-access) helps you simplify your users' access while more securely adopting web, mobile and cloud technologies. This solution helps you strike a balance between usability and security through the use of risk-based access, single sign-on, integrated access management control, identity federation and its mobile multi-factor authentication capability, IBM Verify. Take back control of your access management with IBM Security Verify Access.

          The IBM Security Verify Access operator provides lifecycle management of the lightweight containers which are used to protect an environment, namely:
          * [Web Reverse Proxy](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-web-reverse-proxy)
          * [Runtime](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-runtime)
          * [Distributed Session Cache](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-distributed-session-cache)

          The operator will manage the deployment of these lightweight IBM Security Verify Access worker containers, and also control the rolling restart of these containers when a configuration snapshot is updated.

          See the project [Readme](https://github.com/IBM-Security/verify-access-operator/blob/master/README.md) for further information and details.

        displayName: IBM Security Verify Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Access Product Information
          url: https://www.ibm.com/au-en/products/verify-access
        - name: Verify Access Documentation
          url: https://www.ibm.com/docs/en/sva
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Access Development Team
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - icr.io/isva/verify-access-operator:23.3.0
        version: 23.3.0
      entries:
      - name: ibm-security-verify-access-operator.v23.3.0
        version: 23.3.0
      - name: ibm-security-verify-access-operator.v22.10.0
        version: 22.10.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-access-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-directory-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-directory-operator.v23.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyDirectory",
                "metadata": {
                  "name": "ibmsecurityverifydirectory-sample"
                },
                "spec": {
                  "pods": {
                    "configMap": {
                      "proxy": {
                        "key": "config.yaml",
                        "name": "isvd-proxy-config"
                      },
                      "server": {
                        "key": "config.yaml",
                        "name": "isvd-server-config"
                      }
                    },
                    "image": {
                      "label": "latest",
                      "repo": "icr.io/isvd"
                    },
                    "proxy": {
                      "replicas": 1
                    },
                    "serviceAccountName": "isvd"
                  },
                  "replicas": {
                    "pvcs": [
                      "replica-1",
                      "replica-2"
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: icr.io/isvd/verify-directory-operator@sha256:d1dd7037584c6818dcf8d5aa69472278833bb8e37fba2aec5a69185ba88dcc34
          createdAt: "2023-04-18T22:10:32Z"
          description: The IBM Security Verify Directory operator is used to scale
            an IBM Security Verify Directory Server environment.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ibm-security/verify-directory-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyDirectory is the Schema for the ibmsecurityverifydirectories
              API
            displayName: IBMSecurity Verify Directory
            kind: IBMSecurityVerifyDirectory
            name: ibmsecurityverifydirectories.ibm.com
            version: v1
        description: "IBM Security Verify Directory is a scalable, standards-based
          identity directory that helps simplify identity and directory management.
          Verify Directory helps consolidate identity silos into a single identity
          source. Verify Directory is purpose-built to provide a directory foundation
          that can help provide a trusted identity data infrastructure that assists
          in enabling mission-critical security and authentication. It is designed
          to deliver a reliable, scalable, standards-based identity data platform
          that interoperates with a broad range of operating systems and applications.
          Verify Directory supports Lightweight Directory Access Protocol (LDAP) V3,
          offering a flexible and highly scalable LDAP infrastructure. \n\nThe operator
          can be used to manage deployed instances of the IBM Security Verify Directory
          server, providing an easy mechanism to scale the number of servers which
          are used to handle the workload.\n\nSee the project [Readme](https://github.com/ibm-security/verify-directory-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Security Verify Directory
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Security
        - Identity and Access
        links:
        - name: IBM Security Verify Directory Documentation
          url: https://www.ibm.com/docs/en/svd
        maintainers:
        - email: iamdev@us.ibm.com
          name: IBM Security Verify Directory Development Team
        maturity: stable
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - icr.io/isvd/verify-directory-operator@sha256:d1dd7037584c6818dcf8d5aa69472278833bb8e37fba2aec5a69185ba88dcc34
        version: 23.4.1
      entries:
      - name: ibm-security-verify-directory-operator.v23.4.1
        version: 23.4.1
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-directory-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Flux project
      provider-url: ""
    name: flux
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.toolkit.fluxcd.io/v2beta1",
                "kind": "HelmRelease",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "5m",
                    "chart": {
                      "spec": {
                          "chart": "podinfo",
                          "version": "4.0.x",
                          "sourceRef": {
                            "kind": "HelmRepository",
                            "name": "podinfo",
                            "namespace": "flux-system"
                          },
                          "interval": "1m"
                      }
                    },
                    "values": {
                      "replicaCount": 2
                    }
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "GitRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "1m",
                    "ref": {
                      "branch": "master"
                    },
                    "url": "https://github.com/stefanprodan/podinfo"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmRepository",
                "metadata": {
                    "name": "stable",
                    "namespace": "default"
                },
                "spec": {
                    "url": "https://kubernetes-charts.storage.googleapis.com/",
                    "interval": "10m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmChart",
                "metadata": {
                    "name": "redis",
                    "namespace": "default"
                },
                "spec": {
                    "chart": "redis",
                    "version": "10.5.7",
                    "sourceRef": {
                      "name": "stable",
                      "kind": "HelmRepository"
                    },
                    "interval": "5m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "Bucket",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "1m",
                    "provider": "generic",
                    "bucketName": "podinfo",
                    "endpoint": "minio.minio.svc.cluster.local:9000",
                    "insecure": true,
                    "secretRef": {
                      "name": "minio-credentials"
                    }
                }
              },
              {
                "apiVersion": "kustomize.toolkit.fluxcd.io/v1beta2",
                "kind": "Kustomization",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "5m",
                    "path": "./deploy/overlays/dev",
                    "prune": true,
                    "sourceRef": {
                      "kind": "GitRepository",
                      "name": "podinfo"
                    },
                    "timeout": "2m",
                    "wait": true
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "address": "https://github.com/stefanprodan/podinfo",
                    "secretRef": {
                      "name": "api-token"
                    }
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Alert",
                "metadata": {
                    "name": "webapp",
                    "namespace": "default"
                },
                "spec": {
                    "providerRef": {
                      "name": "on-call-slack"
                    },
                    "eventSeverity": "info",
                    "eventSources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "Bucket",
                          "name": "secrets"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-backend"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-frontend"
                      }
                    ]
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Receiver",
                "metadata": {
                    "name": "github-receiver",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "events": [
                      "ping",
                      "push"
                    ],
                    "secretRef": {
                      "name": "webhook-token"
                    },
                    "resources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "HelmRepository",
                          "name": "webapp"
                      }
                    ]
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "image": "ghcr.io/stefanprodan/podinfo",
                    "interval": "1m0s"
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImagePolicy",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "imageRepositoryRef": {
                      "name": "podinfo"
                    },
                    "policy": {
                      "semver": {
                          "range": "5.0.x"
                      }
                    }
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageUpdateAutomation",
                "metadata": {
                  "name": "flux-system",
                  "namespace": "flux-system"
                },
                "spec": {
                  "interval": "1m0s",
                  "sourceRef": {
                    "kind": "GitRepository",
                    "name": "flux-system"
                  },
                  "git": {
                    "checkout": {
                      "ref": {
                        "branch": "main"
                      }
                    },
                    "commit": {
                      "author": {
                        "email": "fluxcdbot@users.noreply.github.com",
                        "name": "fluxcdbot"
                      },
                      "messageTemplate": "{{range .Updated.Images}}{{println .}}{{end}}"
                    },
                    "push": {
                      "branch": "main"
                    }
                  },
                  "update": {
                    "path": "./clusters/my-cluster",
                    "strategy": "Setters"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.connect.redhat.com/flux/source-controller@sha256:b576e3f010e3ff07e716064ab0dad042aba93f0ea5c250535dd18bd678c45337
          description: Flux is a Continuous Delivery solution for Kubernetes.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/flux/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/flux/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: flux-system
          support: The Flux community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta1
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta2
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta2
          - description: OCIRepository
            displayName: OCIRepository
            kind: OCIRepository
            name: ocirepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta2
        description: |-
          [Flux](https://fluxcd.io/) is a Continuous Delivery solution for Kubernetes.
          Flux is constructed with the [GitOps Toolkit](https://toolkit.fluxcd.io/components/), a set of composable APIs and specialized tools for keeping Kubernetes clusters in sync with sources of configuration (like Git & Helm repositories), and automating updates to configuration when there is new code to deploy.

          Flux is a CNCF incubation project made for:

          - **cluster operators** who automate provision and configuration of clusters
          - **platform engineers** who build continuous delivery for developer teams
          - **app developers** who rely on continuous delivery to get their code live

          Features:

          - Source configuration from Git and Helm repositories, and S3-compatible buckets (e.g., Minio)
          - Kustomize and Helm support
          - Event-triggered and periodic reconciliation
          - Integration with Kubernetes RBAC
          - Health assessment (clusters and workloads)
          - Dependency management (infrastructure and workloads)
          - Secrets management (builtin support for Mozilla SOPS and Cloud KMS)
          - Alerting to external systems (webhook senders)
          - External events handling (webhook receivers)
          - Automated container image updates to Git (image scanning and patching)
          - Policy-driven validation (OPA, Kyverno, admission controllers)
          - Seamless integration with Git providers (GitHub, GitLab, Bitbucket)
          - Interoperability with Cluster API (CAPI) providers

          ### Steps to support bootstraping with Flux CLI

          SecComp Profile should be removed as it requires a special privilege to get the access.
          Also, it requires to enforce `runUserAs` to the same UID provided by the images to prevent OpenShift
          to alter the user id expected by our controllers.

          To do so, please specifiy a patch using the following Kustomization file, for example.

          ```yaml
          apiVersion: kustomize.config.k8s.io/v1beta1
          kind: Kustomization
          resources:
            - gotk-components.yaml
            - gotk-sync.yaml
          patches:
            - patch: |
                apiVersion: apps/v1
                kind: Deployment
                metadata:
                  name: all
                spec:
                  template:
                    spec:
                      containers:
                        - name: manager
                          securityContext:
                            runAsUser: 65534
                            seccompProfile:
                              $patch: delete
              target:
                kind: Deployment
                labelSelector: app.kubernetes.io/part-of=flux
          ```

          Then before installing Flux, please also perform the following steps to bind Flux to the `nonroot` SecurityContextConstraint (SCC). This applies to installation via OC's OLM or via Flux CLI.

          ```sh
          # replace default "flux-system" with the project/namespace name for your flux installation
          namespace="flux-system"
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:source-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:kustomize-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:helm-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:notification-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-automation-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-reflector-controller

          ```

          To see complete steps please refer to this [documentation](https://github.com/fluxcd/website/blob/main/content/en/docs/use-cases/openshift.md).

          If you've already installed Flux, you might need to uninstall it after applying the above pre-requisite steps.

          ### Network Policies

          This version of Flux is shipped without the normally provided network Policies.
          By default all traffic is enabled, and anyone can access the source-controller
          artifacts.

          It is not recommended to run it in production as is.

          Please visit [Flux's policies](https://github.com/fluxcd/flux2/tree/main/manifests/policies)
          to deploy the needed network policies and secure your Flux setup.

          ### Breaking Changes

          - Flux 0.18.x ships with v1beta2 of the kustomize.toolkit.fluxcd.io/Kustomization CRD.
          - Flux 0.25.x is the last Flux release where Kubernetes 1.19 is supported.

          ### Flux on OpenShift Use Cases

          - [The GitOps with Flux + OpenShift](https://www.youtube.com/watch?v=8ybCtNVP5To) talk by Andrew Block (Red Hat) showing how to setup multi-tenacy on OpenShift using Flux.
          - [Flux OpenShift Use Case](https://fluxcd.io/docs/use-cases/openshift/) on FluxCD.io
        displayName: Flux
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - gitops
        - cd
        - continuous delivery
        - fluxcd
        - helm
        - kustomize
        links:
        - name: Homepage
          url: https://fluxcd.io/
        - name: Documentation
          url: https://fluxcd.io/docs/
        - name: Slack
          url: https://cloud-native.slack.com/messages/flux
        - name: Mailing List
          url: https://lists.cncf.io/g/cncf-flux-dev
        - name: Source Code
          url: https://github.com/fluxcd/flux2
        maintainers:
        - email: support@weave.works
          name: The Flux maintainers
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: The Flux project
        relatedImages:
        - registry.connect.redhat.com/flux/helm-controller@sha256:cc24a7b73b023ad86a6c3a845088d153ea290dda7cab1ec03eb85aaf814e6049
        - registry.connect.redhat.com/flux/image-automation-controller@sha256:0c51595307e14acc15f22d50b414a87513ebb03530ef4a8f7d6cd9f8b757243e
        - registry.connect.redhat.com/flux/image-reflector-controller@sha256:3bef7fde22be641304e5ca69f77150e416eba5141f8aa13f68be7f810e55aa59
        - registry.connect.redhat.com/flux/kustomize-controller@sha256:246290908d899b9f6feb920f88913bd2d4673cf852a3020714fd4152c7f78237
        - registry.connect.redhat.com/flux/notification-controller@sha256:b3d81234ba3312e0106f2f7cabc73fbe1d7bf683eda36b9ed5bbdaa43789be40
        - registry.connect.redhat.com/flux/source-controller@sha256:b576e3f010e3ff07e716064ab0dad042aba93f0ea5c250535dd18bd678c45337
        version: 2.0.1
      entries:
      - name: flux.v2.0.1
        version: 2.0.1
      name: stable
    defaultChannel: stable
    packageName: flux
    provider:
      name: The Flux project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: telemetry-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: telemetry-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "Autoscaling",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "telemetry-operator",
                    "app.kubernetes.io/instance": "autoscaling-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "autoscaling",
                    "app.kubernetes.io/part-of": "telemetry-operator"
                  },
                  "name": "autoscaling"
                },
                "spec": {
                  "aodh": {
                    "databaseAccount": "aodh",
                    "databaseInstance": "openstack",
                    "memcachedInstance": "memcached",
                    "passwordSelectors": null,
                    "secret": "osp-secret"
                  }
                }
              },
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "Ceilometer",
                "metadata": {
                  "name": "ceilometer",
                  "namespace": "openstack"
                },
                "spec": {
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "Logging",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "telemetry-operator",
                    "app.kubernetes.io/instance": "logging-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "logging",
                    "app.kubernetes.io/part-of": "telemetry-operator"
                  },
                  "name": "logging"
                },
                "spec": {
                  "cloNamespace": "openshift-logging",
                  "port": 10514
                }
              },
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "MetricStorage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "telemetry-operator",
                    "app.kubernetes.io/instance": "metricstorage-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "metricstorage",
                    "app.kubernetes.io/part-of": "telemetry-operator"
                  },
                  "name": "metricstorage-sample"
                },
                "spec": {
                  "monitoringStack": {
                    "alertingEnabled": true,
                    "scrapeInterval": "30s",
                    "storage": {
                      "persistent": {
                        "pvcStorageRequest": "20G"
                      },
                      "retention": "24h",
                      "strategy": "persistent"
                    }
                  }
                }
              },
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "Telemetry",
                "metadata": {
                  "name": "telemetry"
                },
                "spec": {
                  "autoscaling": {
                    "aodh": {
                      "databaseAccount": "aodh",
                      "databaseInstance": "openstack",
                      "memcachedInstance": "memcached",
                      "passwordSelectors": null,
                      "secret": "osp-secret"
                    },
                    "enabled": false,
                    "heatInstance": "heat"
                  },
                  "ceilometer": {
                    "enabled": true,
                    "secret": "osp-secret"
                  },
                  "logging": {
                    "enabled": false,
                    "port": 10514
                  },
                  "metricStorage": {
                    "enabled": false,
                    "monitoringStack": {
                      "alertingEnabled": true,
                      "scrapeInterval": "30s",
                      "storage": {
                        "persistent": {
                          "pvcStorageRequest": "20G"
                        },
                        "retention": "24h",
                        "strategy": "persistent"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:47Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Autoscaling is the Schema for the autoscalings API
            displayName: Autoscaling
            kind: Autoscaling
            name: autoscalings.telemetry.openstack.org
            version: v1beta1
          - description: Ceilometer is the Schema for the ceilometers API
            displayName: Ceilometer
            kind: Ceilometer
            name: ceilometers.telemetry.openstack.org
            version: v1beta1
          - description: Logging is the Schema for the loggings API
            displayName: Logging
            kind: Logging
            name: loggings.telemetry.openstack.org
            version: v1beta1
          - description: MetricStorage is the Schema for the metricstorages API
            displayName: Metric Storage
            kind: MetricStorage
            name: metricstorages.telemetry.openstack.org
            version: v1beta1
          - description: Telemetry is the Schema for the telemetry API
            displayName: Telemetry
            kind: Telemetry
            name: telemetries.telemetry.openstack.org
            version: v1beta1
        description: Telemetry Operator
        displayName: Telemetry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Telemetry
        - Ceilometer
        links:
        - name: Telemetry Operator
          url: https://github.com/openstack-k8s-operators/telemetry-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-ceilometer-notification-rhel9@sha256:239e5376f65682fd644e8d9b10b27ad879747a056c8fd26ff19882e34c5bb056
        - registry.redhat.io/rhoso/openstack-ceilometer-ipmi-rhel9@sha256:86618d8f54df09210349e382188f3b6563ab96ee511fb4da6eb6dc80e4027aae
        - registry.redhat.io/rhoso/openstack-aodh-listener-rhel9@sha256:3599a25515a96b3ff4768f85fecb4bf955789fb9b5d8abc3181678062020ae64
        - registry.redhat.io/rhoso/openstack-ceilometer-compute-rhel9@sha256:b8a0412041c50e312bef39c17a9c2600e9dbb42ea380d55e58243cda641cef4f
        - registry.redhat.io/rhoso/openstack-aodh-evaluator-rhel9@sha256:16027b7d4f07385c87f97e1581d4454d4ce7ed654499720d9ba5bc6fc74fc246
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/telemetry-rhel9-operator@sha256:83962f19193f98b87c54136f02d18797a8047f65e0d75c11ce9518e50aabc050
        - registry.redhat.io/rhoso/openstack-ceilometer-central-rhel9@sha256:1c79d1edb2404a8614ea29f24e2a869695d75b1695ea0937eebc9c57250fc4ab
        - registry.redhat.io/rhoso/openstack-aodh-api-rhel9@sha256:5ab39a468ab014f2fec4cb9cd210b430ae1c1262f52b9e0f22881d6d8e9def97
        - registry.redhat.io/rhoso/openstack-aodh-notifier-rhel9@sha256:f729548ac3d718f2f5b6c85045e7c676e1656207ee54f43540d246bd97426894
        - registry.redhat.io/rhoso-operators/sg-core-rhel9@sha256:a6e4edc24409361cb770bb44fe5b475dc58748426ccc4003abad7e70f01f1b97
        - registry.redhat.io/ubi9/httpd-24@sha256:dbfbe3383e12d63df7178e4cdd5b1e6e51a84c781041b9192026ea6edb72fbf3
        version: 1.0.6
      entries:
      - name: telemetry-operator.v1.0.6
        version: 1.0.6
      - name: telemetry-operator.v1.0.5
        version: 1.0.5
      - name: telemetry-operator.v1.0.4
        version: 1.0.4
      - name: telemetry-operator.v1.0.3
        version: 1.0.3
      - name: telemetry-operator.v1.0.2
        version: 1.0.2
      - name: telemetry-operator.v1.0.1
        version: 1.0.1
      - name: telemetry-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: telemetry-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Trustification
      provider-url: ""
    name: trustify-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trustify-operator.v0.1.0-alpha.9
      currentCSVDesc:
        annotations:
          alm-examples: |
            [{
              "apiVersion": "org.trustify/v1alpha1",
              "kind": "Trustify",
              "metadata": {
                "name": "myapp"
              },
              "spec": { }
            }]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          containerImage: ghcr.io/trustification/trustify-operator:v0.1.0-alpha.9
          createdAt: "2025-03-07T11:57:13+00:00"
          description: An Operator for installing and managing Trustify
          repository: https://github.com/trustification/trustify-operator
          support: https://github.com/trustification/trustify-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Trustify instance
            displayName: Trustify
            kind: Trustify
            name: trustifies.org.trustify
            version: v1alpha1
        description: |+
          Trustify is vendor-neutral, thought-leadering, mostly informational collection of resources devoted to making Software Supply Chains easier to create, manage, consume and ultimately... to trust!


          You can use Trustify for:

          - Store and Manage all the SBOM (Software Bill of Materials) files of your company.
          - Understand which are the Vulnerabilities that affect each of your SBOM files
          - Understand exactly which Packages are included/shipped within each SBOM

          Trustify is a project within the [Trustification community](https://trustification.io/).


          ### Install
          Once you have successfully installed the Operator, proceed to deploy components by creating the required CR.

          By default, the Operator installs the following components on a target cluster:

          * Server
          * UI

          ### Documentation
          Documentation can be found on our [website](https://trustification.io/).

          ### Getting help
          If you encounter any issues while using Trustify, you can create an issue on our [Github repo](https://github.com/trustification/trustify/issues), for bugs, enhancements or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find
          * Fixing issues by opening [Pull Requests](https://github.com/trustification/trustify/pulls)

        displayName: Trustify Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trust
        links:
        - name: Website
          url: https://trustification.io/
        - name: Github
          url: https://github.com/trustification/trustify
        maintainers:
        - email: trustification@googlegroups.com
          name: Trustification
        maturity: alpha
        minKubeVersion: 1.23.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Trustification
        relatedImages:
        - ghcr.io/trustification/trustify-operator:v0.1.0-alpha.9
        version: 0.1.0-alpha.9
      entries:
      - name: trustify-operator.v0.1.0-alpha.9
        version: 0.1.0-alpha.9
      - name: trustify-operator.v0.1.0-alpha.8
        version: 0.1.0-alpha.8
      - name: trustify-operator.v0.1.0-alpha.7
        version: 0.1.0-alpha.7
      - name: trustify-operator.v0.1.0-alpha.6
        version: 0.1.0-alpha.6
      - name: trustify-operator.v0.1.0-alpha.5
        version: 0.1.0-alpha.5
      - name: trustify-operator.v0.1.0-alpha.4
        version: 0.1.0-alpha.4
      - name: trustify-operator.v0.1.0-alpha.3
        version: 0.1.0-alpha.3
      - name: trustify-operator.v0.1.0-alpha.2
        version: 0.1.0-alpha.2
      name: alpha
    defaultChannel: alpha
    packageName: trustify-operator
    provider:
      name: Trustification
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Aqua Security, Inc.
      provider-url: ""
    name: aqua
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aqua-operator.v2022.4.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaCsp",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "namespace": "aqua",
                    "version": "2022.4",
                    "requirements": true
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry",
                    "dbDiskSize": 10,
                    "databaseSecret": {
                      "key": "db-password",
                      "name": "aqua-database-password"
                    }
                  },
                  "database": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "database",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "gateway": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "gateway",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "server": {
                    "replicas": 1,
                    "service": "LoadBalancer",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "console",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "route": true,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaDatabase",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "database",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "diskSize": 10,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "image": {
                      "repository": "enforcer",
                      "registry": "registry.aquasec.com",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "IfNotPresent"
                    }
                  },
                  "runAsNonRoot": false,
                  "gateway": {
                    "host": "aqua-gateway",
                    "port": 8443
                  },
                  "token": "<<your-token>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaKubeEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "config": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "aqua-secure",
                    "imagePullSecret": "aqua-registry",
                    "kubebench_image": "docker.io/aquasec/kube-bench:v0.7.1"
                  },
                  "deploy": {
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "tag": "<<KUBE_ENFORCER_TAG>>",
                      "repository": "kube-enforcer",
                      "pullPolicy": "Always"
                    }
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "starboard-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  }
                }
              },
              {
              "apiVersion": "operator.aquasec.com/v1alpha1",
              "kind": "AquaLightning",
              "metadata": {
                "name": "aqua"
              },
              "spec": {
                "global": {
                  "gateway_address": "aqua-gateway.aqua:8443",
                  "cluster_name": "Default-cluster-name"
                },
                "common": {
                  "imagePullSecret": "aqua-registry"
                },
                "kubeEnforcer": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                  "allowAnyVersion": null,
                  "deploy": {
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "tag": "2022.4",
                      "repository": "kube-enforcer",
                      "pullPolicy": "Always"
                    }
                  },
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "starboard-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  },
                  "env": []
                },
                "enforcer": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "deploy": {
                    "image": {
                      "repository": "enforcer",
                      "registry": "registry.aquasec.com",
                      "tag": "2022.4",
                      "pullPolicy": "IfNotPresent"
                    }
                  },
                  "token": "<<ENFORCER_GROUP_TOKEN>>",
                  "env": [],
                  "secret": {
                    "name": null,
                    "key": null
                  },
                  "runAsNonRoot": null,
                  "rhcosVersion": null
                }
              }
            },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaGateway",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "gateway",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaScanner",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "replicas": 1,
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "scanner",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "runAsNonRoot": false,
                  "login": {
                    "username": "<<YOUR AQUA USER NAME>>",
                    "password": "<<YOUR AQUA USER PASSWORD>>",
                    "host": "http://aqua-server:8080",
                    "token": ""
                  }
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaServer",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "LoadBalancer",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "console",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "runAsNonRoot": false
                }
              },
              {
                  "apiVersion": "operator.aquasec.com/v1alpha1",
                  "kind": "AquaCloudConnector",
                  "metadata": {
                    "name": "aqua"
                  },
                  "spec": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "common": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "image": {
                        "repository": "aqua-cloud-connector",
                        "registry": "registry.aquasec.com",
                        "tag": "<<IMAGE TAG>>"
                      },
                      "replicas": 1
                    },
                    "runAsNonRoot": false,
                    "gateway": {
                      "host": "aqua-gateway",
                      "port": "8443"
                    },
                    "login": {
                      "host": "aqua-gateway:8443",
                      "password": "<<YOUR AQUA USER PASSWORD>>",
                      "token": "<<YOUR AQUA CLOUD CONNECTOR TOKEN>>",
                      "username": "<<YOUR AQUA USER NAME>>"
                    },
                    "tunnels": [
                      {
                        "host": "<account>.azurecr.io",
                        "port": "443"
                      },
                      {
                        "host": "gke.gcr.io",
                        "port": "443"
                      },
                      {
                        "host": "<account>.jfrog.io",
                        "port": "443"
                      },
                      {
                        "host": "1.1.1.1",
                        "port": "443"
                      },
                      {
                        "region": "us-east1",
                        "type": "ecr"
                      }
                    ]
                  }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: docker.io/aquasec/aqua-operator:2022.4.18
          createdAt: "2020-03-25T08:00:00Z"
          description: The Aqua Security Operator runs within a Openshift cluster
            and provides a means to deploy and manage Aqua Security cluster and components.
          operators.operatorframework.io/internal-objects: '["aquastarboards.aquasecurity.github.io","configauditreports.aquasecurity.github.io",
            "clusterconfigauditreports.aquasecurity.github.io"]'
          repository: https://github.com/aquasecurity/aqua-operator
          support: Aqua Security, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aqua Security CSP Deployment with Aqua Operator
            displayName: AquaCsp
            kind: AquaCsp
            name: aquacsps.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Lightning Enforcers Deployment with Aqua Operator
            displayName: AquaLightningEnforcers
            kind: AquaLightning
            name: aqualightnings.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Database Deployment with Aqua Operator
            displayName: AquaDatabase
            kind: AquaDatabase
            name: aquadatabases.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Enforcer Deployment with Aqua Operator
            displayName: AquaEnforcer
            kind: AquaEnforcer
            name: aquaenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security KubeEnforcer Deployment with Aqua Operator
            displayName: AquaKubeEnforcer
            kind: AquaKubeEnforcer
            name: aquakubeenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Gateway Deployment with Aqua Operator
            displayName: AquaGateway
            kind: AquaGateway
            name: aquagateways.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Scanner Deployment with Aqua Operator
            displayName: AquaScanner
            kind: AquaScanner
            name: aquascanners.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Server Deployment with Aqua Operator
            displayName: AquaServer
            kind: AquaServer
            name: aquaservers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Starboard for kube-enforcer.
            displayName: AquaStarboard
            kind: AquaStarboard
            name: aquastarboards.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ClusterConfigAuditReport
            kind: ClusterConfigAuditReport
            name: clusterconfigauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ConfigAuditReport
            kind: ConfigAuditReport
            name: configauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: Aqua Security Cloud Connector Deployment with Aqua Operator
            displayName: AquaCloudConnector
            kind: AquaCloudConnector
            name: aquacloudconnector.operator.aquasec.com
            version: v1alpha1
        description: |-
          The Aqua Security Operator runs within an OpenShift cluster and provides a means to deploy and manage the Aqua Security cluster and components :
          * Server (aka “console”)
          * Database (for production environments we recommend to use an external database and not the Aqua default database)
          * Gateway
          * Enforcer (aka “agent”)
          * KubeEnforcer
          * Scanner
          * CSP (package that contains the Server, Database, and Gateway)
          * LightningEnforcer (package that contains Enforcer and Kube Enforcer)

          Use the Aqua-Operator to
          * Deploy Aqua Security components on OpenShift
          * Scale up Aqua Security components with extra replicas
          * Assign metadata tags to Aqua Security components

          ## Before You Begin Using the Operator CRDs

          Obtain access to the Aqua registry - https://www.aquasec.com/about-us/contact-us/


          ## Instructions and exmaples

          Follow instructions and exmaples
          [here](https://github.com/aquasecurity/aqua-operator/blob/2022.4.0/docs/DeployOpenShiftOperator.md)
        displayName: Aqua Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aqua-security
        - scanning
        - security
        - runtime-security
        links:
        - name: Aqua Security
          url: https://www.aquasec.com/
        - name: Aqua Operator Github
          url: https://github.com/aquasecurity/aqua-operator
        maintainers:
        - email: support@aquasec.com
          name: Aqua Support, Aqua Security
        maturity: alpha
        provider:
          name: Aqua Security, Inc.
        relatedImages:
        - docker.io/aquasec/aqua-operator:2022.4.18
        version: 2022.4.18
      entries:
      - name: aqua-operator.v2022.4.18
        version: 2022.4.18
      - name: aqua-operator.v2022.4.17
        version: 2022.4.17
      name: 2022.4.0
    defaultChannel: 2022.4.0
    packageName: aqua
    provider:
      name: Aqua Security, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: The bpfman Community
      provider-url: https://bpfman.io/
    name: bpfman-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bpfman-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfapplication"
                  },
                  "name": "bpfapplication-sample"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:v0.5.6"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "kprobe": {
                        "bpffunctionname": "kprobe_counter",
                        "func_name": "try_to_wake_up",
                        "offset": 0,
                        "retprobe": false
                      },
                      "type": "Kprobe"
                    },
                    {
                      "tracepoint": {
                        "bpffunctionname": "tracepoint_kill_recorder",
                        "names": [
                          "syscalls/sys_enter_kill"
                        ]
                      },
                      "type": "Tracepoint"
                    },
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "containernames": [
                            "bpfman",
                            "bpfman-agent"
                          ],
                          "namespace": "bpfman",
                          "pods": {
                            "matchLabels": {
                              "name": "bpfman-daemon"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfNsApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfnsapplication"
                  },
                  "name": "bpfapplication-sample",
                  "namespace": "acme"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:v0.5.6"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FentryProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fentryprogram"
                  },
                  "name": "fentry-example"
                },
                "spec": {
                  "bpffunctionname": "test_fentry",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fentry:v0.5.6"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FexitProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fexitprogram"
                  },
                  "name": "fexit-example"
                },
                "spec": {
                  "bpffunctionname": "test_fexit",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fexit:v0.5.6"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "KprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kprobeprogram"
                  },
                  "name": "kprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_kprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/kprobe:v0.5.6"
                    }
                  },
                  "func_name": "try_to_wake_up",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "offset": 0,
                  "retprobe": false
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcnsprogram"
                  },
                  "name": "tc-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcprogram"
                  },
                  "name": "tc-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:v0.5.6"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxnsprogram"
                  },
                  "name": "tcx-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "tcx_next",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxprogram"
                  },
                  "name": "tcx-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "tcx_pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:v0.5.6"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TracepointProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tracepointprogram"
                  },
                  "name": "tracepoint-example"
                },
                "spec": {
                  "bpffunctionname": "enter_openat",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tracepoint:v0.5.6"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "names": [
                    "syscalls/sys_enter_openat"
                  ],
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobensprogram"
                  },
                  "name": "uprobe-example",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobeprogram"
                  },
                  "name": "uprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:v0.5.6"
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpnsprogram"
                  },
                  "name": "xdp-ns-pass-all-nodes",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpprogram"
                  },
                  "name": "xdp-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:v0.5.6"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/bpfman/bpfman-operator:v0.5.6
          createdAt: "2025-02-18T13:44:53Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: bpfman
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "bpfman",
                "labels": {
                  "pod-security.kubernetes.io/enforce": "privileged",
                  "pod-security.kubernetes.io/audit": "privileged",
                  "pod-security.kubernetes.io/warn": "privileged",
                },
                "annotations": {
                  "openshift.io/node-selector": ""
                },
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/bpfman/bpfman
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BpfApplication is the Schema for the bpfapplications API
            displayName: Bpf Application
            kind: BpfApplication
            name: bpfapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsApplication is the Schema for the BpfNsApplications
              API
            displayName: Bpf Namespaced Application
            kind: BpfNsApplication
            name: bpfnsapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsProgram is the Schema for the BpfNsProgram API
            displayName: Bpf Namespaced Program
            kind: BpfNsProgram
            name: bpfnsprograms.bpfman.io
            version: v1alpha1
          - description: BpfProgram is the Schema for the BpfProgram API
            displayName: Bpf Program
            kind: BpfProgram
            name: bpfprograms.bpfman.io
            version: v1alpha1
          - description: FentryProgram is the Schema for the Fentryprograms API
            displayName: Fentry Program
            kind: FentryProgram
            name: fentryprograms.bpfman.io
            version: v1alpha1
          - description: FexitProgram is the Schema for the Fexitprograms API
            displayName: Fexit Program
            kind: FexitProgram
            name: fexitprograms.bpfman.io
            version: v1alpha1
          - description: KprobeProgram is the Schema for the Kprobeprograms API
            displayName: Kprobe Program
            kind: KprobeProgram
            name: kprobeprograms.bpfman.io
            version: v1alpha1
          - description: TcNsProgram is the Schema for the Tcnsprograms API
            displayName: TC Namespaced Program
            kind: TcNsProgram
            name: tcnsprograms.bpfman.io
            version: v1alpha1
          - description: TcProgram is the Schema for the Tcprograms API
            displayName: TC Program
            kind: TcProgram
            name: tcprograms.bpfman.io
            version: v1alpha1
          - description: TcxNsProgram is the Schema for the Tcxnsprograms API
            displayName: TCX Namespaced Program
            kind: TcxNsProgram
            name: tcxnsprograms.bpfman.io
            version: v1alpha1
          - description: TcxProgram is the Schema for the Tcxprograms API
            displayName: TCX Program
            kind: TcxProgram
            name: tcxprograms.bpfman.io
            version: v1alpha1
          - description: TracepointProgram is the Schema for the Tracepointprograms
              API
            displayName: Tracepoint Program
            kind: TracepointProgram
            name: tracepointprograms.bpfman.io
            version: v1alpha1
          - description: UprobeNsProgram is the Schema for the Uprobensprograms API
            displayName: Uprobe Namespaced Program
            kind: UprobeNsProgram
            name: uprobensprograms.bpfman.io
            version: v1alpha1
          - description: UprobeProgram is the Schema for the Uprobeprograms API
            displayName: Uprobe Program
            kind: UprobeProgram
            name: uprobeprograms.bpfman.io
            version: v1alpha1
          - description: XdpNsProgram is the Schema for the Xdpnsprograms API
            displayName: Xdp Namespaced Program
            kind: XdpNsProgram
            name: xdpnsprograms.bpfman.io
            version: v1alpha1
          - description: XdpProgram is the Schema for the Xdpprograms API
            displayName: Xdp Program
            kind: XdpProgram
            name: xdpprograms.bpfman.io
            version: v1alpha1
        description: "The bpfman Operator is a Kubernetes Operator for deploying [bpfman](https://bpfman.netlify.app/),
          a system daemon\nfor managing eBPF programs. It deploys bpfman itself along
          with CRDs to make deploying\neBPF programs in Kubernetes much easier.\n\n##
          Quick Start\n\nTo get bpfman up and running quickly simply click 'install'
          to deploy the bpfman-operator in the bpfman namespace via operator-hub.\n##
          Configuration\n\nThe `bpfman-config` configmap is automatically created
          in the `bpfman` namespace and used to configure the bpfman deployment.\n\nTo
          edit the config simply run\n\n```bash\nkubectl edit cm bpfman-config\n```\n\nThe
          following fields are adjustable\n\n- `bpfman.agent.image`: The image used
          for the bpfman-agent`\n- `bpfman.image`: The image used for bpfman`\n- `bpfman.log.level`:
          the log level for bpfman, currently supports `debug`, `info`, `warn`, `error`,
          and `fatal`, defaults to `info`\n- `bpfman.agent.log.level`: the log level
          for the bpfman-agent currently supports `info`, `debug`, and `trace` \n\nThe
          bpfman operator deploys eBPF programs via CRDs. The following CRDs are currently
          available, \n\n- BpfApplication\n- FentryProgram\n- FexitProgram\n - KprobeProgram\n-
          TcProgram\n- TcxProgram\n- TracepointProgram\n- UprobeProgram\n- XdpProgram\n
          - BpfNsApplication\n- TcNsProgram\n- TcxNsProgram\n- UprobeNsProgram\n-
          XdpNsProgram\n- BpfProgram\n - BpfNsProgram\n\n ## More information\n\nPlease
          checkout the [bpfman community website](https://bpfman.io/) for more information."
        displayName: Bpfman Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ebpf
        - kubernetes
        links:
        - name: bpfman website
          url: https://bpfman.io/
        maintainers:
        - email: astoycos@redhat.com
          name: Andrew Stoycos
        - email: afredette@redhat.com
          name: Andre Fredette
        maturity: alpha
        provider:
          name: The bpfman Community
          url: https://bpfman.io/
        relatedImages:
        - quay.io/bpfman/bpfman-operator:v0.5.6
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 0.5.6
      entries:
      - name: bpfman-operator.v0.5.6
        version: 0.5.6
      - name: bpfman-operator.v0.5.5
        version: 0.5.5
      - name: bpfman-operator.v0.5.4
        version: 0.5.4
      - name: bpfman-operator.v0.5.2
        version: 0.5.2
      - name: bpfman-operator.v0.5.1
        version: 0.5.1
      - name: bpfman-operator.v0.5.0
        version: 0.5.0
      - name: bpfman-operator.v0.4.2
        version: 0.4.2
      - name: bpfman-operator.v0.4.1
        version: 0.4.1
      name: alpha
    defaultChannel: alpha
    packageName: bpfman-operator
    provider:
      name: The bpfman Community
      url: https://bpfman.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Snyk Ltd.
      provider-url: ""
    name: snyk-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: snyk-operator.v1.92.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1alpha1",
                "kind": "SnykMonitor",
                "metadata": {
                  "name": "snyk-monitor"
                },
                "spec": {
                  "clusterName": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "snyk/kubernetes-monitor",
                    "tag": "1.92.0-ubi8"
                  },
                  "integrationApi": "",
                  "monitorSecrets": "snyk-monitor",
                  "scope": "Cluster",
                  "temporaryStorageSize": "50Gi",
                  "pvc": {
                    "enabled": false,
                    "create": false,
                    "name": "snyk-monitor-pvc",
                    "storageClassName": null
                  },
                  "initContainerImage": {
                    "repository": "busybox",
                    "tag": "latest"
                  },
                  "requests": {
                    "memory": "400Mi"
                  },
                  "limits": {
                    "memory": "2Gi"
                  },
                  "nodeAffinity": {
                    "disableBetaArchNodeSelector": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Security
          certified: "false"
          containerImage: docker.io/snyk/kubernetes-operator:1.92.0
          createdAt: "2022-06-09T17:49:36Z"
          description: A Kubernetes Operator for creating and managing Snyk controller
            instances.
          olm.skipRange: <1.68.2
          repository: https://github.com/snyk/kubernetes-monitor
          support: Snyk Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a deployment of the Snyk controller, which scans
              container images in the cluster for vulnerabilities.
            displayName: Snyk Monitor
            kind: SnykMonitor
            name: snykmonitors.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          A Kubernetes Operator for creating and managing Snyk Kubernetes controller instances.

          Snyk integrates with Kubernetes, enabling you to import and test your running workloads and identify vulnerabilities in their associated images and configurations that might make those workloads less secure. Once imported, Snyk continues to monitor those workloads, identifying additional security issues as new images are deployed and the workload configuration changes.

          ## Prerequisites

          To start using the Operator and the Snyk controller that it manages, you will need to create a secret containing your Snyk integrationId and Docker config file:

          ```
          kubectl create namespace snyk-monitor
          kubectl create secret generic snyk-monitor -n snyk-monitor --from-literal=dockercfg.json={} --from-literal=integrationId=${SNYK_INTEGRATION_ID}
          ```

          Refer to the [Snyk documentation](https://support.snyk.io/hc/en-us/articles/360003916138-Kubernetes-integration-overview) for more details.
        displayName: Snyk Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - snyk
        - controller
        - monitor
        - operator
        - security
        - cluster
        - container
        - image
        - security
        - vulnerabilities
        - advisories
        - audit
        - scan
        - scanning
        links:
        - name: Website
          url: https://snyk.io
        - name: Documentation
          url: https://support.snyk.io/hc/en-us/articles/360003916138-Kubernetes-integration-overview
        maintainers:
        - email: support@snyk.io
          name: Snyk Ltd.
        maturity: stable
        provider:
          name: Snyk Ltd.
        relatedImages:
        - docker.io/snyk/kubernetes-operator:1.92.0
        version: 1.92.0
      entries:
      - name: snyk-operator.v1.92.0
        version: 1.92.0
      - name: snyk-operator.v1.90.2
        version: 1.90.2
      - name: snyk-operator.v1.90.1
        version: 1.90.1
      - name: snyk-operator.v1.88.4
        version: 1.88.4
      - name: snyk-operator.v1.85.2
        version: 1.85.2
      - name: snyk-operator.v1.82.1
        version: 1.82.1
      - name: snyk-operator.v1.81.1
        version: 1.81.1
      - name: snyk-operator.v1.80.0
        version: 1.80.0
      - name: snyk-operator.v1.73.0
        version: 1.73.0
      - name: snyk-operator.v1.72.0
        version: 1.72.0
      - name: snyk-operator.v1.70.4
        version: 1.70.4
      - name: snyk-operator.v1.70.1
        version: 1.70.1
      name: stable
    defaultChannel: stable
    packageName: snyk-operator
    provider:
      name: Snyk Ltd.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-sds-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-sds-operator.v8.0.230
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.230"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleStoragePool",
                "metadata": {
                  "name": "infoscalestoragepool-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:f4947eab2b31fccc86b5d5d0d3fd4eb1f7513fd8325e379cc259f5ccd1dfc3e4
          createdAt: "2023-10-26T20:36:59Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          olm.skipRange: <8.0.230
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleStoragePool is the Schema for the infoscalestoragepools
              API
            displayName: InfoScale Storage Pool
            kind: InfoScaleStoragePool
            name: infoscalestoragepools.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.230"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:f4947eab2b31fccc86b5d5d0d3fd4eb1f7513fd8325e379cc259f5ccd1dfc3e4
        version: 8.0.230
      entries:
      - name: infoscale-sds-operator.v8.0.230
        version: 8.0.230
      - name: infoscale-sds-operator.v8.0.220
        version: 8.0.220
      - name: infoscale-sds-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-sds-operator.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-sds-operator.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.330"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:dab501b99373c840c7e8e3ab479fa273b39f9bf8a896f3f1e4e945f4ab81f64e
          createdAt: "2024-06-24T09:03:32Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.330"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:dab501b99373c840c7e8e3ab479fa273b39f9bf8a896f3f1e4e945f4ab81f64e
        version: 8.0.330
      entries:
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-sds-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.410"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
          createdAt: "2024-07-30T09:00:57Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Arctera InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/168721626-168743109-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.410"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Arctera recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
        version: 8.0.410
      entries:
      - name: infoscale-sds-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-sds-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-sds-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-sds-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.410"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
          createdAt: "2024-07-30T09:00:57Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Arctera InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/168721626-168743109-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.410"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Arctera recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
        version: 8.0.410
      entries:
      - name: infoscale-sds-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-sds-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-sds-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-sds-operator.v8.0.210
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.210"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:26573b98743245c0b607abbe0d981abce6ce9fb1c5c0cd9fe735351b7136f4e8
          createdAt: "2022-11-26T16:36:59Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleStoragePool is the Schema for the infoscalestoragepools
              API
            displayName: InfoScale Storage Pool
            kind: InfoScaleStoragePool
            name: infoscalestoragepools.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.210"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:1667227b5111fe52ed304ab19ff9fadfc677239e56b2c93a3b89b1bfbc31354d
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:26573b98743245c0b607abbe0d981abce6ce9fb1c5c0cd9fe735351b7136f4e8
        version: 8.0.210
      entries:
      - name: infoscale-sds-operator.v8.0.210
        version: 8.0.210
      - name: infoscale-sds-operator.v8.0.201
        version: 8.0.201
      - name: infoscale-sds-operator.v8.0.200
        version: 8.0.200
      - name: infoscale-sds-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-sds-operator.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-sds-operator
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: ovn-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ovn-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ovn.openstack.org/v1beta1",
                "kind": "OVNController",
                "metadata": {
                  "name": "ovncontroller-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "ovn.openstack.org/v1beta1",
                "kind": "OVNDBCluster",
                "metadata": {
                  "name": "ovndbcluster-nb-sample"
                },
                "spec": {
                  "dbType": "NB",
                  "logLevel": "info",
                  "storageClass": "local-storage",
                  "storageRequest": "10G"
                }
              },
              {
                "apiVersion": "ovn.openstack.org/v1beta1",
                "kind": "OVNDBCluster",
                "metadata": {
                  "name": "ovndbcluster-sb-sample"
                },
                "spec": {
                  "dbType": "SB",
                  "logLevel": "info",
                  "storageClass": "local-storage",
                  "storageRequest": "10G"
                }
              },
              {
                "apiVersion": "ovn.openstack.org/v1beta1",
                "kind": "OVNNorthd",
                "metadata": {
                  "name": "ovnnorthd-sample"
                },
                "spec": {
                  "logLevel": "info"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:45Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OVNController is the Schema for the ovncontrollers API
            displayName: OVNController
            kind: OVNController
            name: ovncontrollers.ovn.openstack.org
            version: v1beta1
          - description: OVNDBCluster is the Schema for the ovndbclusters API
            displayName: OVNDBCluster
            kind: OVNDBCluster
            name: ovndbclusters.ovn.openstack.org
            version: v1beta1
          - description: OVNNorthd is the Schema for the ovnnorthds API
            displayName: OVNNorthd
            kind: OVNNorthd
            name: ovnnorthds.ovn.openstack.org
            version: v1beta1
        description: OVN Operator
        displayName: OVN Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Network
        - Ovn
        links:
        - name: Ovn Operator
          url: https://github.com/openstack-k8s-operators/ovn-operator
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/ovn-rhel9-operator@sha256:bce19011b3623c66a59efe6ea874e2403440813bdc970cd6170853a8093076f3
        - registry.redhat.io/rhoso/openstack-ovn-base-rhel9@sha256:b0d424f6ea6de8192382541a95aa56fa32b9b95e8d433c41c4a6dadbe4127fae
        - registry.redhat.io/rhoso/openstack-ovn-nb-db-server-rhel9@sha256:e71c8aac00a78e3d786332aaada3657276f69483df1e441b70589601ca038a5b
        - registry.redhat.io/rhoso/openstack-ovn-sb-db-server-rhel9@sha256:7aa2832f80846d5ac71733a1f186d39dce3654ebb661d32b8e4580d92be8e338
        - registry.redhat.io/rhoso/openstack-ovn-northd-rhel9@sha256:f1e4ce04b67d888fa950b83d1f00459c67c602397336da81de3ff71d4052a5d8
        - registry.redhat.io/rhoso/openstack-ovn-controller-rhel9@sha256:6c7827dbe25c168d28250f197d5ea6f3f60cfbab4c974267ec5ea0f8d631c3f5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: ovn-operator.v1.0.6
        version: 1.0.6
      - name: ovn-operator.v1.0.5
        version: 1.0.5
      - name: ovn-operator.v1.0.4
        version: 1.0.4
      - name: ovn-operator.v1.0.3
        version: 1.0.3
      - name: ovn-operator.v1.0.2
        version: 1.0.2
      - name: ovn-operator.v1.0.1
        version: 1.0.1
      - name: ovn-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: ovn-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: fuse-apicurito
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-apicurito.v7.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1alpha1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:8d19ad579f874d6b9a08f127c8900906913a5a7c1e69cbe4b5cb0333b8d86152
          createdAt: 2022-02-28 09:35:38 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.10.3'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:8d19ad579f874d6b9a08f127c8900906913a5a7c1e69cbe4b5cb0333b8d86152
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:8dbeeb8b04c81f41d896f71db90813de57e67493fa85389b63ac23f390785043
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:c02c218ccfbb385ea29af98409995a3d9c26cb0b4c76caf2e0779fa233b5ff8c
        version: 7.10.3
      entries:
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.10.x
    - currentCSV: fuse-apicurito.v7.11.1-0.1687357785.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:2f8d4450187839f0e0c761e5ddcf61b03f5996acfe3d75b60120e76e5b0167c1
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.11.2'
          olm.substitutesFor: fuse-apicurito.v7.11.1-0.1682509045.p
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:2f8d4450187839f0e0c761e5ddcf61b03f5996acfe3d75b60120e76e5b0167c1
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:df61aafc4023b9068a8636c35eabb0d5fda400f523b64cb1fc5d8c480272c435
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:4c493a2d690d0d7c1684c53d15c06b47d691aeffdde6dd471b6e75fed2fec6a9
        version: 7.11.1+0.1687357785.p
      entries:
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.11.x
    - currentCSV: fuse-apicurito.v7.12.1-0.1706187670.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:7ccefee63165e7eaf9a609e8a9f978a06f27deff69a318fe399a29292ad99106
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.12.1'
          olm.substitutesFor: fuse-apicurito.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:7ccefee63165e7eaf9a609e8a9f978a06f27deff69a318fe399a29292ad99106
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:4cd612283e8136839f1f48a3a220dac7c6c38adf14b61bc03a981e514baefe9b
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:bee0eeaa7898af835f80c5217059d73361daa42197fa6cbbaf416e1ba373eb54
        version: 7.12.1+0.1706187670.p
      entries:
      - name: fuse-apicurito.v7.12.1-0.1706187670.p
        version: 7.12.1+0.1706187670.p
      - name: fuse-apicurito.v7.12.1
        version: 7.12.1
      - name: fuse-apicurito.v7.12.0
        version: 7.12.0
      - name: fuse-apicurito.v7.12.0-0.1697748407.p
        version: 7.12.0+0.1697748407.p
      - name: fuse-apicurito.v7.12.0-0.1697419531.p
        version: 7.12.0+0.1697419531.p
      - name: fuse-apicurito.v7.12.0-0.1694087055.p
        version: 7.12.0+0.1694087055.p
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.12.x
    - currentCSV: fuse-apicurito.v7.13.0-0.1753785079.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:d58d345d74059242d57d5c1a75f31517769230d3014a42243fda20aa93ebd5c8
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.7.0 <7.13.0'
          olm.substitutesFor: fuse-apicurito.v7.13.0-0.1751291353.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:d58d345d74059242d57d5c1a75f31517769230d3014a42243fda20aa93ebd5c8
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:f5bf9240b28b3664bdb11b1f57f611e6194ff8643ae939b6ab3b2b176705f487
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:4e64a5930a10b76147ac128a216cd7c0346da2ff3e9f31ed86311941ed193cc1
        version: 7.13.0+0.1753785079.p
      entries:
      - name: fuse-apicurito.v7.13.0-0.1739885035.p
        version: 7.13.0+0.1739885035.p
      - name: fuse-apicurito.v7.13.0-0.1741884705.p
        version: 7.13.0+0.1741884705.p
      - name: fuse-apicurito.v7.13.0-0.1753785079.p
        version: 7.13.0+0.1753785079.p
      - name: fuse-apicurito.v7.13.0-0.1721224330.p
        version: 7.13.0+0.1721224330.p
      - name: fuse-apicurito.v7.13.0-0.1753778625.p
        version: 7.13.0+0.1753778625.p
      - name: fuse-apicurito.v7.13.0-0.1753702605.p
        version: 7.13.0+0.1753702605.p
      - name: fuse-apicurito.v7.13.0-0.1721607497.p
        version: 7.13.0+0.1721607497.p
      - name: fuse-apicurito.v7.13.0-0.1751291353.p
        version: 7.13.0+0.1751291353.p
      - name: fuse-apicurito.v7.13.0-0.1748960166.p
        version: 7.13.0+0.1748960166.p
      - name: fuse-apicurito.v7.13.0-0.1745523192.p
        version: 7.13.0+0.1745523192.p
      - name: fuse-apicurito.v7.13.0-0.1744737027.p
        version: 7.13.0+0.1744737027.p
      - name: fuse-apicurito.v7.13.0-0.1744280419.p
        version: 7.13.0+0.1744280419.p
      - name: fuse-apicurito.v7.13.0-0.1741875975.p
        version: 7.13.0+0.1741875975.p
      - name: fuse-apicurito.v7.13.0-0.1739473863.p
        version: 7.13.0+0.1739473863.p
      - name: fuse-apicurito.v7.13.0-0.1738056248.p
        version: 7.13.0+0.1738056248.p
      - name: fuse-apicurito.v7.13.0-0.1731460815.p
        version: 7.13.0+0.1731460815.p
      - name: fuse-apicurito.v7.13.0-0.1729189341.p
        version: 7.13.0+0.1729189341.p
      - name: fuse-apicurito.v7.13.0-0.1727957847.p
        version: 7.13.0+0.1727957847.p
      - name: fuse-apicurito.v7.13.0-0.1727193594.p
        version: 7.13.0+0.1727193594.p
      - name: fuse-apicurito.v7.13.0-0.1725022690.p
        version: 7.13.0+0.1725022690.p
      - name: fuse-apicurito.v7.13.0-0.1724659732.p
        version: 7.13.0+0.1724659732.p
      - name: fuse-apicurito.v7.13.0-0.1724053718.p
        version: 7.13.0+0.1724053718.p
      - name: fuse-apicurito.v7.13.0-0.1723709434.p
        version: 7.13.0+0.1723709434.p
      - name: fuse-apicurito.v7.13.0
        version: 7.13.0
      - name: fuse-apicurito.v7.13.0-0.1720707249.p
        version: 7.13.0+0.1720707249.p
      - name: fuse-apicurito.v7.13.0-0.1718285973.p
        version: 7.13.0+0.1718285973.p
      - name: fuse-apicurito.v7.13.0-0.1717672390.p
        version: 7.13.0+0.1717672390.p
      - name: fuse-apicurito.v7.13.0-0.1753778936.p
        version: 7.13.0+0.1753778936.p
      - name: fuse-apicurito.v7.13.0-0.1753701210.p
        version: 7.13.0+0.1753701210.p
      - name: fuse-apicurito.v7.12.1-0.1706187670.p
        version: 7.12.1+0.1706187670.p
      - name: fuse-apicurito.v7.12.1
        version: 7.12.1
      - name: fuse-apicurito.v7.12.0-0.1697419531.p
        version: 7.12.0+0.1697419531.p
      - name: fuse-apicurito.v7.12.0-0.1697748407.p
        version: 7.12.0+0.1697748407.p
      - name: fuse-apicurito.v7.12.0-0.1694087055.p
        version: 7.12.0+0.1694087055.p
      - name: fuse-apicurito.v7.12.0
        version: 7.12.0
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.13.x
    - currentCSV: fuse-apicurito.v7.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
                    "apiVersion": "apicur.io/v1alpha1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
            }]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:05c9676aada2d34a985a59abcf55105a3765d75b0d3764089d2b32e9ca5d367c
          createdAt: "2020-10-06 12:48:47"
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.9.3'
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito Project
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CRD for Apicurito
            displayName: Apicurito CRD
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:05c9676aada2d34a985a59abcf55105a3765d75b0d3764089d2b32e9ca5d367c
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:5535864ce393ae0ae5ee52ed0c7661e0351fb7df2f9235da70608245519530b2
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:dc39b79c360511ff999eba5b2c299db3e08c9c244a0a1f3da8c90c02695e506d
        version: 7.9.3
      entries:
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.9.x
    defaultChannel: fuse-apicurito-7.13.x
    packageName: fuse-apicurito
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-organizations-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-organizations-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "organizations.services.k8s.aws/v1alpha1",
                "kind": "OrganizationalUnit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/organizations-controller:1.1.1
          createdAt: "2025-09-19T19:35:20Z"
          description: AWS Organizations controller is a service controller for managing
            Organizations resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Account represents the state of an AWS organizations Account
              resource.
            displayName: Account
            kind: Account
            name: accounts.organizations.services.k8s.aws
            version: v1alpha1
          - description: OrganizationalUnit represents the state of an AWS organizations
              OrganizationalUnit resource.
            displayName: OrganizationalUnit
            kind: OrganizationalUnit
            name: organizationalunits.organizations.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Organizations resources in AWS from within your Kubernetes cluster.

          **About Amazon Organizations**

          With AWS Organizations you can perform account management activities at scale by consolidating multiple AWS accounts into a single organization. Consolidating accounts simplifies how you use other AWS services. You can leverage the multi-account management services available in AWS Organizations with select AWS services to perform tasks on all accounts that are members of your organization.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Organizations
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - organizations
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Organizations Developer Resources
          url: https://aws.amazon.com/organizations/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: organizations maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/organizations-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-organizations-controller.v1.1.1
        version: 1.1.1
      - name: ack-organizations-controller.v1.0.13
        version: 1.0.13
      - name: ack-organizations-controller.v1.0.11
        version: 1.0.11
      - name: ack-organizations-controller.v1.0.10
        version: 1.0.10
      - name: ack-organizations-controller.v1.0.9
        version: 1.0.9
      - name: ack-organizations-controller.v1.0.8
        version: 1.0.8
      - name: ack-organizations-controller.v1.0.7
        version: 1.0.7
      - name: ack-organizations-controller.v1.0.6
        version: 1.0.6
      - name: ack-organizations-controller.v1.0.5
        version: 1.0.5
      - name: ack-organizations-controller.v1.0.4
        version: 1.0.4
      - name: ack-organizations-controller.v1.0.1
        version: 1.0.1
      - name: ack-organizations-controller.v1.0.0
        version: 1.0.0
      - name: ack-organizations-controller.v0.0.10
        version: 0.0.10
      - name: ack-organizations-controller.v0.0.9
        version: 0.0.9
      - name: ack-organizations-controller.v0.0.8
        version: 0.0.8
      - name: ack-organizations-controller.v0.0.7
        version: 0.0.7
      - name: ack-organizations-controller.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: ack-organizations-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeRocketCI
      provider-url: https://docs.kuberocketci.io
    name: nexus-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-operator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexus-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexus",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexus-sample"
                },
                "spec": {
                  "secret": "nexus-secret",
                  "url": "https://nexus-example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRepository",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrepository-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrepository",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrepository-sample"
                },
                "spec": {
                  "go": {
                    "proxy": {
                      "cleanup": {
                        "policyNames": [
                          "cleanup-policy"
                        ]
                      },
                      "httpClient": {
                        "autoBlock": true,
                        "blocked": true,
                        "connection": {
                          "retries": 1440,
                          "timeout": 1440,
                          "userAgentSuffix": "user-agent-suffix"
                        }
                      },
                      "name": "go-proxy",
                      "negativeCache": {
                        "enabled": true,
                        "timeToLive": 1440
                      },
                      "online": true,
                      "proxy": {
                        "contentMaxAge": 1440,
                        "metadataMaxAge": 1440,
                        "remoteUrl": "https://test.com"
                      },
                      "routingRule": "routing-rule",
                      "storage": {
                        "blobStoreName": "blob-store-name",
                        "strictContentTypeValidation": true
                      }
                    }
                  },
                  "nexusRef": {
                    "name": "nexus-sample"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrole",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrole-sample"
                },
                "spec": {
                  "description": "test-role",
                  "id": "test-role",
                  "name": "test-role",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "privileges": [
                    "nx-blobstores-all"
                  ]
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexususer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexususer",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexususer-sample"
                },
                "spec": {
                  "email": "test@gmail.com",
                  "firstName": "John1",
                  "id": "test-user",
                  "lastName": "Doe",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "roles": [
                    "nx-admin"
                  ],
                  "secret": "$user-secret:password"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/nexus-operator:3.0.0
          createdAt: "2023-12-11T12:54:50Z"
          description: An Operator for managing Nexus entities.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-nexus-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus is the Schema for the nexus API.
            displayName: Nexus
            kind: Nexus
            name: nexuses.edp.epam.com
            version: v1alpha1
          - description: NexusRepository is the Schema for the nexusrepositories API.
            displayName: Nexus Repository
            kind: NexusRepository
            name: nexusrepositories.edp.epam.com
            version: v1alpha1
          - description: NexusRole is the Schema for the nexusroles API.
            displayName: Nexus Role
            kind: NexusRole
            name: nexusroles.edp.epam.com
            version: v1alpha1
          - description: NexusUser is the Schema for the nexususers API.
            displayName: Nexus User
            kind: NexusUser
            name: nexususers.edp.epam.com
            version: v1alpha1
        description: |
          Nexus Operator is an operator that is responsible for establishing
          a connection to provided NexusRepository Server, reconciling Nexus entities (groups,
          users, repositories, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Nexus Repository.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  nexus-secret
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Nexus` with Nexus instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1alpha1
              kind: Nexus
              metadata:
                name: nexus-sample
              spec:
                secret: nexus-secret                # Secret name
                url: https://nexus.example.com      # Nexus URL
              ```

              Wait for the `.status` field with  `status.connected: true`
        displayName: EDP Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - edp
        - nexus
        - operator
        - artifact
        - storage
        links:
        - name: Nexus Operator
          url: https://github.com/epam/edp-nexus-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/nexus-operator:3.0.0
        version: 3.0.0
      entries:
      - name: nexus-operator.v3.0.0
        version: 3.0.0
      name: alpha
    - currentCSV: nexus-operator.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexus-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexus",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexus-sample"
                },
                "spec": {
                  "secret": "nexus-secret",
                  "url": "https://nexus-example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusBlobStore",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusblobstore-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusblobstore",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusblobstore-sample"
                },
                "spec": {
                  "file": {
                    "path": "storage-blobstore/nexusblobstore-sample"
                  },
                  "name": "nexusblobstore-sample",
                  "nexusRef": {
                    "name": "nexus-sample"
                  },
                  "softQuota": {
                    "limit": 1000,
                    "type": "spaceUsedQuota"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusCleanupPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexuscleanuppolicy-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexuscleanuppolicy",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexuscleanuppolicy-sample"
                },
                "spec": {
                  "criteria": {
                    "lastBlobUpdated": 30
                  },
                  "description": "Cleanup policy for go",
                  "format": "go",
                  "name": "go-cleanup-policy",
                  "nexusRef": {
                    "name": "nexus-sample"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRepository",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrepository-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrepository",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrepository-sample"
                },
                "spec": {
                  "go": {
                    "proxy": {
                      "cleanup": {
                        "policyNames": [
                          "cleanup-policy"
                        ]
                      },
                      "httpClient": {
                        "autoBlock": true,
                        "blocked": true,
                        "connection": {
                          "retries": 1440,
                          "timeout": 1440,
                          "userAgentSuffix": "user-agent-suffix"
                        }
                      },
                      "name": "go-proxy",
                      "negativeCache": {
                        "enabled": true,
                        "timeToLive": 1440
                      },
                      "online": true,
                      "proxy": {
                        "contentMaxAge": 1440,
                        "metadataMaxAge": 1440,
                        "remoteUrl": "https://test.com"
                      },
                      "routingRule": "routing-rule",
                      "storage": {
                        "blobStoreName": "blob-store-name",
                        "strictContentTypeValidation": true
                      }
                    }
                  },
                  "nexusRef": {
                    "name": "nexus-sample"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrole",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrole-sample"
                },
                "spec": {
                  "description": "test-role",
                  "id": "test-role",
                  "name": "test-role",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "privileges": [
                    "nx-blobstores-all"
                  ]
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusScript",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusscript-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusscript",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusscript-sample"
                },
                "spec": {
                  "content": "security.setAnonymousAccess(Boolean.valueOf(args))\n",
                  "name": "anonymous",
                  "nexusRef": {
                    "name": "nexus-sample"
                  },
                  "payload": "true"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexususer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexususer",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexususer-sample"
                },
                "spec": {
                  "email": "test@gmail.com",
                  "firstName": "John1",
                  "id": "test-user",
                  "lastName": "Doe",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "roles": [
                    "nx-admin"
                  ],
                  "secret": "$user-secret:password"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/nexus-operator:3.5.0
          createdAt: "2025-04-22T11:21:24Z"
          description: An Operator for managing Nexus entities.
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/epam/edp-nexus-operator
          support: KubeRocketCI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NexusBlobStore is the Schema for the nexusblobstores API.
            displayName: Nexus Blob Store
            kind: NexusBlobStore
            name: nexusblobstores.edp.epam.com
            version: v1alpha1
          - description: NexusCleanupPolicy is the Schema for the cleanuppolicies
              API.
            displayName: Nexus Cleanup Policy
            kind: NexusCleanupPolicy
            name: nexuscleanuppolicies.edp.epam.com
            version: v1alpha1
          - description: Nexus is the Schema for the nexus API.
            displayName: Nexus
            kind: Nexus
            name: nexuses.edp.epam.com
            version: v1alpha1
          - description: NexusRepository is the Schema for the nexusrepositories API.
            displayName: Nexus Repository
            kind: NexusRepository
            name: nexusrepositories.edp.epam.com
            version: v1alpha1
          - description: NexusRole is the Schema for the nexusroles API.
            displayName: Nexus Role
            kind: NexusRole
            name: nexusroles.edp.epam.com
            version: v1alpha1
          - description: NexusScript is the Schema for the nexusscripts API.
            displayName: Nexus Script
            kind: NexusScript
            name: nexusscripts.edp.epam.com
            version: v1alpha1
          - description: NexusUser is the Schema for the nexususers API.
            displayName: Nexus User
            kind: NexusUser
            name: nexususers.edp.epam.com
            version: v1alpha1
        description: |
          Nexus Operator is an operator that is responsible for establishing
          a connection to provided NexusRepository Server, reconciling Nexus entities (groups,
          users, repositories, etc) according to the created CRs.

          ## Quick Start

          1. Login into Nexus and create user. Attach permissions to user such as scripts, rules, blobs etc. Insert user credentials into Kubernetes secret.

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name: nexus-admin-password
              data:
                password: cGFzcw==  # base64-encoded value of "pass"
                user:     dXNlcg==  # base64-encoded value of "user"
              ```

          2. Create Custom Resource `kind: Nexus` with Nexus instance URL and secret created on the previous step:

              ```yaml
              apiVersion: edp.epam.com/v1alpha1
              kind: Nexus
              metadata:
                name: nexus
              spec:
                secret: nexus-admin-password
                url: http://nexus.example.com
              ```

              Wait for the `.status` field with  `status.connected: true`

          3. Create Role using Custom Resources NexusRole:

              ```yaml
              apiVersion: edp.epam.com/v1alpha1
              kind: NexusRole
              metadata:
                name: edp-admin
              spec:
                description: Read and write access to all repos and scripts
                id: edp-admin
                name: edp-admin
                nexusRef:
                  kind: Nexus
                  name: nexus
                privileges:
                  - nx-apikey-all
                  - nx-repository-view-*-*-add
                  - nx-repository-view-*-*-browse
                  - nx-repository-view-*-*-edit
                  - nx-repository-view-*-*-read
                  - nx-script-*-add
                  - nx-script-*-delete
                  - nx-script-*-run
                  - nx-search-read
              ```
        displayName: EDP Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - artifact
        - edp
        - kuberocketci
        - nexus
        - operator
        - storage
        links:
        - name: Nexus Operator
          url: https://github.com/epam/edp-nexus-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: KubeRocketCI
          url: https://docs.kuberocketci.io
        relatedImages:
        - docker.io/epamedp/nexus-operator:3.5.0
        version: 3.5.0
      entries:
      - name: nexus-operator.v3.5.0
        version: 3.5.0
      - name: nexus-operator.v3.4.0
        version: 3.4.0
      - name: nexus-operator.v3.3.0
        version: 3.3.0
      name: stable
    defaultChannel: stable
    packageName: nexus-operator
    provider:
      name: KubeRocketCI
      url: https://docs.kuberocketci.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp, Inc.
      provider-url: https://www.netapp.com/
    name: trident-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trident-operator.v25.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "trident.netapp.io/v1",
                "kind": "TridentOrchestrator",
                "metadata": {
                  "name": "trident"
                },
                "spec": {
                  "IPv6": false,
                  "debug": true,
                  "nodePrep": [],
                  "imagePullSecrets": [],
                  "imageRegistry": "",
                  "namespace": "trident",
                  "silenceAutosupport": false
                }
              }
            ]
          categories: Storage
          containerImage: docker.io/netapp/trident-operator@sha256:c3448b6847550f45849a5cdbb9e0cedb35290d21131bc513d833d95eb42a43e0
          createdAt: "2025-09-03"
          description: Trident Operator, to manage NetApp Trident installations
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          operators.openshift.io/valid-subscription: No subscriptions are required
          repository: https://github.com/netapp/trident
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Used to deploy NetApp Trident.
            displayName: Trident Orchestrator
            kind: TridentOrchestrator
            name: tridentorchestrators.trident.netapp.io
            version: v1
          - description: Automates AWS FSxN backend configuration
            displayName: Trident Configurator
            kind: TridentConfigurator
            name: tridentconfigurators.trident.netapp.io
            version: v1
        description: |
          NetApp Trident is an open source storage provisioner and orchestrator maintained by NetApp. It enables you to create storage volumes for containerized applications managed by Docker and Kubernetes. For full release information, including patch release changes, see https://docs.netapp.com/us-en/trident/trident-rn.html.
        displayName: NetApp Trident
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - NetApp
        - ONTAP
        - Trident
        - storage
        - performance
        - open source
        - Astra
        - Cloud Native
        - AWS
        - Azure
        - GCP
        - Cloud Volumes
        links:
        - name: GitHub Repository
          url: https://github.com/NetApp/trident
        - name: Trident documentation
          url: https://docs.netapp.com/us-en/trident/index.html
        - name: Support policy
          url: https://mysupport.netapp.com/site/info/version-support
        - name: Release Notes
          url: https://docs.netapp.com/us-en/trident/trident-rn.html
        maintainers:
        - email: ng-trident-openshift-operator@netapp.com
          name: NetApp
        provider:
          name: NetApp, Inc.
          url: https://www.netapp.com/
        relatedImages:
        - docker.io/netapp/trident-operator@sha256:c3448b6847550f45849a5cdbb9e0cedb35290d21131bc513d833d95eb42a43e0
        - docker.io/netapp/trident@sha256:7925717979a84f93b62a2592682cde3443d6350ec4e88f8fd1a7d3ea812e8900
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:bdee8df520467b30312e7b14d7b6ef3963557ccd38269913de3fa51bb1bceec9
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:145ab76f9606e9d51dcd1e4e159c67021ecf53c24d8d80ed7bbeb2cc7fb5c42b
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6781d64bfb01e252595130eec814b9d2fe6426720a003cc61e82e9963767bfd4
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:b48bd5bc612c1bbe73d203c70e6caaae2b5360bd31eba20eb06cbd2102cc34fc
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:b88958c38883f5b1e3d307da90aac34a5ceafdb6bd3d989b062cb24ecafd1fbd
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:0ef3bf2a4dcbb5aa829ce9cc5f12efaccaf99b246b4b15abbb66b0135bec4cab
        - docker.io/netapp/trident-autosupport@sha256:21b3afbebfac3cc28c4eb319c2d385187bd8ebb728d8cfd82dbc437cc9b83b33
        version: 25.6.2
      entries:
      - name: trident-operator.v25.6.2
        version: 25.6.2
      - name: trident-operator.v25.6.1
        version: 25.6.1
      - name: trident-operator.v25.6.0
        version: 25.6.0
      - name: trident-operator.v25.2.1
        version: 25.2.1
      - name: trident-operator.v25.2.0
        version: 25.2.0
      name: stable
    defaultChannel: stable
    packageName: trident-operator
    provider:
      name: NetApp, Inc.
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: xspc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: xspc-operator.v1.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.hpe.com/v1",
                "kind": "XSPC",
                "metadata": {
                  "name": "xspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.connect.redhat.com/hpestorage/xspc-csi-driver-operator@sha256:8b8724998a8301057a3c4187c7b94bf4484aeb807065a12f84142cf844ea5534
          createdAt: "2024-12-17T03:53:30Z"
          description: An operator for managing HPE XP Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XSPC is the Schema for the xspcs API
            displayName: XSPC
            kind: XSPC
            name: xspcs.csi.hpe.com
            version: v1
        description: |-
          ## About
          HPE XP Storage Plug-in for Containers (XSPC) is a plugin that integrates HPE XP storage into Kubernetes based clusters.
          XSPC provides persistent volume provisioning capabilities from HPE's block storage arrays.
        displayName: HPE XP Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - XP7
        - XP8
        - HPE
        - Storage
        - Container Storage Interface
        - CSI
        links:
        - name: Knowledge Base
          url: https://www.hpe.com/us/en/storage/enterprise-xp-storage.html
        maintainers:
        - email: hpexpci@hpe.com
          name: HPE
        maturity: stable
        provider:
          name: HPE
        relatedImages:
        - registry.connect.redhat.com/hpestorage/xspc-csi-driver-operator@sha256:8b8724998a8301057a3c4187c7b94bf4484aeb807065a12f84142cf844ea5534
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.connect.redhat.com/hpestorage/xspc-csi-driver@sha256:bdeac71c0b54c2779910a8d8faea03c1122000127704f17ceadba4938b685353
        version: 1.15.0
      entries:
      - name: xspc-operator.v1.15.0
        version: 1.15.0
      name: stable
    defaultChannel: stable
    packageName: xspc-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Entando, Inc
      provider-url: ""
    name: entando-k8s-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: entando-k8s-operator.v6.3.2-pr3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: beta
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - docker.io/entando/entando-k8s-service@sha256:775391f0005e64764288bfe007d9ea576f13618c02ea400d030a7502cfb551bd
        - docker.io/entando/entando-k8s-database-service-controller@sha256:58ca4ad2fca83a715aba9b68609f9056f633d556eb9845b3aff5c05769369539
        - docker.io/entando/app-builder@sha256:1ca90c09302b9cadf8950278d92ad4d7f3a14f02dea8704320794aa77ada8631
        - docker.io/entando/entando-k8s-app-controller@sha256:c68d3b129fc625f5d1ee10ed2a107e28c145049d4b513dcdb2c736555025e339
        - docker.io/entando/entando-k8s-plugin-controller@sha256:a20c753dda86e0f54f210e446b404733045ba426fc7186dc3936daed8224fbcb
        - docker.io/entando/entando-component-manager@sha256:668191fc86b090452baaba18ec7bbc3d07df5a99750e0c5b98759599d565cfbe
        - docker.io/entando/entando-redhat-sso@sha256:3692adb3694b3bf278bed7a678c9d0eaa0aba8e3e3ed4117ca04156525bcac72
        - registry.redhat.io/rhel8/mysql-80@sha256:570bce7dc649ac7ac85e7ecb196aabc281c77a94590aa7df7d9746869ce757ea
        - docker.io/entando/entando-k8s-dbjob@sha256:8ae47dbf93201965106663299f06daad544982fa244d4a0f365a1b717727e5ed
        - docker.io/entando/entando-k8s-composite-app-controller@sha256:4bc6797bb41b3e407b64530c3cff7a78f3fc2c22863caad9c606ccd650310b61
        - docker.io/entando/entando-k8s-app-plugin-link-controller@sha256:aa7ba0861705213434e610530c5bf5da1bdb33f1834d3c346174f5a28145c0bd
        - docker.io/entando/entando-de-app-eap@sha256:1f0e35d7f7ee8123721cf8e80ac96ac2f594f21a2b68cf9e720917d090c8095e
        - docker.io/entando/entando-k8s-keycloak-controller@sha256:f1ed42001260802353197433cea7df4c2d0e5fd9bcd926bb5ae7c654dfad2ffe
        - registry.redhat.io/rhel8/postgresql-12@sha256:f4e5c728b644bf1888ec8086424852ed74b5596a511be29e636fb10218fc9b6f
        - entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        - docker.io/entando/entando-k8s-cluster-infrastructure-controller@sha256:86dbea6696349d7b2393e12bacef4d1acec2dd3f8551e397fa26c92706a18dd3
        version: 6.3.2-pr3
      entries:
      - name: entando-k8s-operator.v6.3.2-pr3
        version: 6.3.2-pr3
      name: alpha
    - currentCSV: entando-k8s-operator.v6.3.2-pr3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: beta
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - docker.io/entando/entando-redhat-sso@sha256:3692adb3694b3bf278bed7a678c9d0eaa0aba8e3e3ed4117ca04156525bcac72
        - entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        - docker.io/entando/entando-de-app-eap@sha256:1f0e35d7f7ee8123721cf8e80ac96ac2f594f21a2b68cf9e720917d090c8095e
        - docker.io/entando/entando-k8s-composite-app-controller@sha256:4bc6797bb41b3e407b64530c3cff7a78f3fc2c22863caad9c606ccd650310b61
        - docker.io/entando/entando-k8s-app-controller@sha256:c68d3b129fc625f5d1ee10ed2a107e28c145049d4b513dcdb2c736555025e339
        - docker.io/entando/entando-k8s-plugin-controller@sha256:a20c753dda86e0f54f210e446b404733045ba426fc7186dc3936daed8224fbcb
        - docker.io/entando/entando-component-manager@sha256:668191fc86b090452baaba18ec7bbc3d07df5a99750e0c5b98759599d565cfbe
        - docker.io/entando/entando-k8s-app-plugin-link-controller@sha256:aa7ba0861705213434e610530c5bf5da1bdb33f1834d3c346174f5a28145c0bd
        - docker.io/entando/entando-k8s-database-service-controller@sha256:58ca4ad2fca83a715aba9b68609f9056f633d556eb9845b3aff5c05769369539
        - docker.io/entando/app-builder@sha256:1ca90c09302b9cadf8950278d92ad4d7f3a14f02dea8704320794aa77ada8631
        - docker.io/entando/entando-k8s-keycloak-controller@sha256:f1ed42001260802353197433cea7df4c2d0e5fd9bcd926bb5ae7c654dfad2ffe
        - registry.redhat.io/rhel8/mysql-80@sha256:570bce7dc649ac7ac85e7ecb196aabc281c77a94590aa7df7d9746869ce757ea
        - registry.redhat.io/rhel8/postgresql-12@sha256:f4e5c728b644bf1888ec8086424852ed74b5596a511be29e636fb10218fc9b6f
        - docker.io/entando/entando-k8s-cluster-infrastructure-controller@sha256:86dbea6696349d7b2393e12bacef4d1acec2dd3f8551e397fa26c92706a18dd3
        - docker.io/entando/entando-k8s-dbjob@sha256:8ae47dbf93201965106663299f06daad544982fa244d4a0f365a1b717727e5ed
        - docker.io/entando/entando-k8s-service@sha256:775391f0005e64764288bfe007d9ea576f13618c02ea400d030a7502cfb551bd
        version: 6.3.2-pr3
      entries:
      - name: entando-k8s-operator.v6.3.2-pr3
        version: 6.3.2-pr3
      name: beta
    - currentCSV: entando-k8s-operator.v6.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: registry.hub.docker.com/entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - registry.hub.docker.com/entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        version: 6.3.2
      entries:
      - name: entando-k8s-operator.v6.3.2
        version: 6.3.2
      name: stable
    defaultChannel: stable
    packageName: entando-k8s-operator-rhmp
    provider:
      name: Entando, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: submariner.io
      provider-url: ""
    name: submariner
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: submariner.v0.17.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:067ba6e58a6fe150a212881a29c90aa7cbd22f53a6877eebc413072f73f9179a
          createdAt: "2025-06-18 01:35:57"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.16.0 <0.17.6'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.17 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:d58214791a60cd9e4366c6ff534584eedbd37386da71b868867e5e212df270f9
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:a565d8c094f3755b6c9303c5b8ae03a5883796e0fea754b0202e5792f1e472c3
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:067ba6e58a6fe150a212881a29c90aa7cbd22f53a6877eebc413072f73f9179a
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:b76998e8bbef1d06aa8dbc49af4f42a3c5603d480a67e46baf11168514c7759a
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:56848aefe1d156147ab73e7353017b2663f63e2e01fe43ede57abc72cfc0c50f
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:e6dbc4bf3e8b1a3c1a7ed10eb463d134f9ed28a03c7bb523c82c5431c5c9ae90
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:5510cb1bb3e0ca5f2365f4a5fc147ca693d3c220f1a8e7835f8b13b66b957ae8
        version: 0.17.6
      entries:
      - name: submariner.v0.17.6
        version: 0.17.6
      - name: submariner.v0.17.5
        version: 0.17.5
      - name: submariner.v0.17.4
        version: 0.17.4
      - name: submariner.v0.17.2
        version: 0.17.2
      - name: submariner.v0.17.2-0.1725901206.p
        version: 0.17.2+0.1725901206.p
      - name: submariner.v0.17.1
        version: 0.17.1
      - name: submariner.v0.17.0
        version: 0.17.0
      name: stable-0.17
    - currentCSV: submariner.v0.18.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:d8d7f0147c7b65b41903e101240905703992a7bbadb33e302aa86b23000d2095
          createdAt: "2025-05-07 20:54:56"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.17.0 <0.18.5'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.18 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:e5c16f0d720e15d54781870a5e63c85540256355009584f7d1bd03ef20a2eae4
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:00072fea3fdb5e294a01da0ee44165fc52219a965a27bdf7d792276f6c3a36c3
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:e8366433d1f50e2062b86d7f749f17cd6566197560256e77f57b2c165f277733
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:a07ef1bfec7549b32ad94f2c44a168860d5f3b46fd13593c068c58bb137c77fc
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:d8d7f0147c7b65b41903e101240905703992a7bbadb33e302aa86b23000d2095
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:27f18015b8896b04457c356d03276d50c34b8c1d84142ec07342ff2c1c4e81fa
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:e913be7a440b6cc0c1dddbda93a5979fe28b99e63f12a7949966b69126fcee90
        version: 0.18.5
      entries:
      - name: submariner.v0.18.5
        version: 0.18.5
      - name: submariner.v0.18.4
        version: 0.18.4
      - name: submariner.v0.18.3
        version: 0.18.3
      - name: submariner.v0.18.2
        version: 0.18.2
      - name: submariner.v0.18.0
        version: 0.18.0
      name: stable-0.18
    - currentCSV: submariner.v0.19.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:4afaf6689e2225c5c7f6fbe002202a1cdf9444e23fd210cc2d86da71c9f78626
          createdAt: "2025-04-22 21:12:45"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.18.0 <0.19.4'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.19 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:d9d4d03013d65d11a4e8ecda411015f359e4215a3a53059a37a810acf2e3ea6c
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:4afaf6689e2225c5c7f6fbe002202a1cdf9444e23fd210cc2d86da71c9f78626
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:5b83c5889aa0d8f7838dd5afb8c901dfb0ce0fbe01175e8468ec50ddedef9223
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:70008af49a9d1ee516204715f66bd415c9caa7462f6b178dc1dc91e513605fcd
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:ef73b8964cd66ffb496191056594e0886afa9d17c03976b4159413c89ed569bf
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:5938d67a0a2653e92f0995440700512992ed82269820679f68483dcc1b2d98d9
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:1bbb624bbefbb75407071449d0144258732e219e3a7fc67ac709c3bcee560576
        version: 0.19.4
      entries:
      - name: submariner.v0.19.4
        version: 0.19.4
      - name: submariner.v0.19.3
        version: 0.19.3
      - name: submariner.v0.19.2
        version: 0.19.2
      - name: submariner.v0.19.0
        version: 0.19.0
      name: stable-0.19
    - currentCSV: submariner.v0.20.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "quay.io/submariner",
                  "version": "0.20.2"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "quay.io/submariner",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "0.20.2"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: $(SUBMARINER_OPERATOR_IMAGE)
          createdAt: "2025-09-30 14:21:54"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/submariner-io/submariner-operator
          support: Submariner
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.20 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: submariner.io
        relatedImages:
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:fd6c46b4a0570b5e7b6457c465e5a1b843ba6f9c209c1856dbebfee1068e35ea
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:70d563a6e6526b9a146a2973210b2abdae09d5acab403fa3f0b465ed5e3c9b11
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:77e0da2db6d5bb2831e6c51214e08685c9f7f3b8b1e6ac53b4fd1418644ad6a1
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:0eef6521c96b29722076c1e220770124c8ed214379abc4654d9567cff8b52d04
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:8ea02aba8f9dd901f1f0ccff97ec8d157bebe5a90a6c924affe5e01cf3454e72
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:ed521b241da60efabfe2e2b80c51639d1a82a43945945879decbfb192bbc6d23
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:4a4ac198e9d0f4a4aaed8f0ec0e1d79bd8fa712b671ee697f7ff9d5616a02d84
        version: 0.20.2
      entries:
      - name: submariner.v0.20.2
        version: 0.20.2
      - name: submariner.v0.20.1
        version: 0.20.1
      - name: submariner.v0.20.0
        version: 0.20.0
      name: stable-0.20
    - currentCSV: submariner.v0.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "quay.io/submariner",
                  "version": "v0.21.0"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "quay.io/submariner",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v0.21.0"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: $(SUBMARINER_OPERATOR_IMAGE)
          createdAt: "2025-08-13 21:57:14"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/submariner-io/submariner-operator
          support: Submariner
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.21 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: submariner.io
        relatedImages:
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:cd851eab4a2214ea7a699bd01d2abac1322d01562f27b35f58f3c0b8716f88fb
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:bcdeaa67664568c5e2a6928236ff69b175ce65ddf7a58c1664d091b926000052
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:b0ad338aed82367048cf024767b0b09e6bb09bbf7dd5128e6711e1feec6f696e
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:65765c7f08c5454b557fbda91c49efbd2cc0d4d973cf8acc7ff40ec8a381839d
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:f5909b1802986622ab169a82d5dd7877af72660903e961b3097cc29e30eb9799
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:b307f748125fd459c3474fdab645688675a06515a9924b47cedba80e038458db
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:0b74f058400f24a63cf9b57a388914cc6dda38e03d639196871f3fd76fb7f2ae
        version: 0.21.0
      entries:
      - name: submariner.v0.21.0
        version: 0.21.0
      name: stable-0.21
    defaultChannel: stable-0.21
    packageName: submariner
    provider:
      name: submariner.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app: kube-green
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: kube-green Community
      provider-url: https://kube-green.dev
    name: kube-green
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kube-green.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kube-green.com/v1alpha1",
                "kind": "SleepInfo",
                "metadata": {
                  "labels": {
                    "app": "kube-green"
                  },
                  "name": "sleepinfo-sample"
                },
                "spec": {
                  "excludeRef": [
                    {
                      "apiVersion": "apps/v1",
                      "kind": "Deployment",
                      "name": "api-gateway"
                    }
                  ],
                  "sleepAt": "20:00",
                  "suspendCronJobs": true,
                  "timeZone": "Europe/Rome",
                  "wakeUpAt": "08:00",
                  "weekdays": "1-5"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          containerImage: docker.io/kubegreen/kube-green:0.7.1
          createdAt: "2025-06-29T15:45:17Z"
          description: Suspend your pods when no-one's using them to save energy and
            restore when necessary
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kube-green/kube-green
          support: kube-green Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SleepInfo is the Schema for the sleepinfos API
            displayName: SleepInfo
            kind: SleepInfo
            name: sleepinfos.kube-green.com
            version: v1alpha1
        description: |
          ## About this Operator

          [kube-green](https://kube-green.dev) shuts down (some of) your resources when you don't need them, reducing the consumption of resources of the cluster.

          Install this operator and apply the SleepInfo CRD to stop your dev namespaces from running during weekends, nights or when it is not necessary.
          See the [docs](https://kube-green.dev/docs/getting-started/) for more information.
          This operator is responsible to suspend and restart pods at specific times, defined by the SleepInfo CRD.

          An example of SleepInfo:

          ```yaml
          apiVersion: kube-green.com/v1alpha1
          kind: SleepInfo
          metadata:
            name: working-hours
          spec:
            weekdays: "1-5"
            sleepAt: "20:00"
            wakeUpAt: "08:00"
            timeZone: "Europe/Rome"
          ```

          This configuration stop pods everyday from monday to friday at 20:00 and restore them at 8:00.
          In this way, your cluster nodes could downscale, producing less CO2 and saving money.
        displayName: kube-green
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - green
        - green software
        - resources
        - downscale
        - autoscale
        - climate change
        links:
        - name: Homepage
          url: https://kube-green.dev
        - name: Source Code
          url: https://github.com/kube-green/kube-green
        maintainers:
        - email: bianchidavide12+kubegreen@gmail.com
          name: Davide
        maturity: alpha
        minKubeVersion: 1.27.0
        provider:
          name: kube-green Community
          url: https://kube-green.dev
        relatedImages:
        - docker.io/kubegreen/kube-green:0.7.1
        version: 0.7.1
      entries:
      - name: kube-green.v0.7.1
        version: 0.7.1
      - name: kube-green.v0.7.0
        version: 0.7.0
      - name: kube-green.v0.6.0
        version: 0.6.0
      - name: kube-green.v0.5.2
        version: 0.5.2
      - name: kube-green.v0.5.1
        version: 0.5.1
      - name: kube-green.v0.5.0
        version: 0.5.0
      - name: kube-green.v0.4.1
        version: 0.4.1
      - name: kube-green.v0.4.0
        version: 0.4.0
      - name: kube-green.v0.3.1
        version: 0.3.1
      - name: kube-green.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: kube-green
    provider:
      name: kube-green Community
      url: https://kube-green.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-wafv2-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-wafv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wafv2.services.k8s.aws/v1alpha1",
                "kind": "IPSet",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/wafv2-controller:1.1.1
          createdAt: "2025-09-20T21:41:21Z"
          description: AWS WAFV2 controller is a service controller for managing WAFV2
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IPSet represents the state of an AWS wafv2 IPSet resource.
            displayName: IPSet
            kind: IPSet
            name: ipsets.wafv2.services.k8s.aws
            version: v1alpha1
          - description: RuleGroup represents the state of an AWS wafv2 RuleGroup
              resource.
            displayName: RuleGroup
            kind: RuleGroup
            name: rulegroups.wafv2.services.k8s.aws
            version: v1alpha1
          - description: WebACL represents the state of an AWS wafv2 WebACL resource.
            displayName: WebACL
            kind: WebACL
            name: webacls.wafv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Web Application Firewall (Amazon WAFV2) resources in AWS from within your Kubernetes cluster.

          **About Amazon WAFV2**

          AWS WAF is a web application firewall that lets you monitor the HTTP and HTTPS requests that are forwarded to your protected web application resources. You can protect the following resource types:
          - Amazon CloudFront distribution
          - Amazon API Gateway REST API
          - Application Load Balancer
          - AWS AppSync GraphQL API
          - Amazon Cognito user pool
          - AWS App Runner service
          - AWS Verified Access instance

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon WAFV2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - wafv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon WAFV2 Developer Resources
          url: https://aws.amazon.com/waf/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: wafv2 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/wafv2-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-wafv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-wafv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-wafv2-controller.v1.0.14
        version: 1.0.14
      - name: ack-wafv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-wafv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-wafv2-controller.v1.0.10
        version: 1.0.10
      - name: ack-wafv2-controller.v1.0.9
        version: 1.0.9
      - name: ack-wafv2-controller.v1.0.8
        version: 1.0.8
      - name: ack-wafv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-wafv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-wafv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-wafv2-controller.v1.0.1
        version: 1.0.1
      - name: ack-wafv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-wafv2-controller.v0.0.4
        version: 0.0.4
      - name: ack-wafv2-controller.v0.0.3
        version: 0.0.3
      - name: ack-wafv2-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-wafv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/in-en/products/cloud-pak-for-data
    name: ibm-zen-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-zen-operator.v6.1.3
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.3 <6.1.3'
          operators.operatorframework.io/builder: operator-sdk-v1.6.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: represents zen instance
            kind: ZenService
            name: zenservices.zen.cpd.ibm.com
            version: v1
          - description: represents zen extension instance
            kind: ZenExtension
            name: zenextensions.zen.cpd.ibm.com
            version: v1
        description: |
          The Zen operator is a component in IBM® Cloud Pak foundational services. It serves a framework for user access management, defining user roles and for powering persona driven user experiences for Cloud Paks and Service offerings
        displayName: IBM Zen Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cloud Pak
        - IBM
        - Access management
        - Platform UI
        - Zen framework
        maintainers:
        - email: cp4d@ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/in-en/products/cloud-pak-for-data
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:deb66a64288a370c5da5c7cc0ba4e2ba4ab816d7c5a080a0c505b05f55fa6072
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:a7c050ae29cb93754f59d8f83676275ee9a91ae9a9956395b97abf8c5e20e96e
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:abd96661823553d623eb0ea1fac8271d64391243dd0f13eb4a4fcc858dfe10a5
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:1df782d0d923544767696d719bd0ac10802b8e06508a104af5e023a519f32dee
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:82afac013ef9b6b9d0eacb4575c979b65abd6f2821971182b90c71bd488ef366
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:719b0871f0d32da894ec80dc2a23483623b7802a5c6f935cf3672468f99faa40
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:a99b1f5a4eb0306b83562de0184fc826c34835ffbf46d929ceb9e70e0d325361
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:b5a8fb14c3bb7c89ff970ca85f06dfcdca832e2b45ad86e8bb1b6c0a0e7e0764
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:b87fa8374976ccd3dad90ddbf6b65cd6ae5a028ba896562f284f089380ebe44a
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:525924934f74b28783dacf630a6401ccd4c88d253aafa2d312b7993d8b9a14c1
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:a8d2e42757c6ccbfe4372465b1827c633aad012c5194bf8b08afcd5691caa6f5
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:10ede3c4e3c7e0cbc26e7b6e797cad35e39face1a1b9d4a4e2440ab0feeeb814
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:8444703b2a7822f82141afcf4f36bf255650d755321b9e9fe2cbd1bdea41d5c7
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:ecbc65705de76804a3c7642339ecfb4cc0f5a61891058e07c20bf2b63f1b82b8
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:3d5c6573faf4ae00cbb21cbfe9b2ad5cd8ffb658df0c5ba09bd98b3bf54f9979
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:e9173d81ef2a75839c0e1d677b72bb00b8e6d7ef0511afab103f08a884a03738
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:c8baecad9c97ddbbb3f9f02c90abbf483d46974b7bc9605622a302cbeba1de74
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8da8b240176c00f8f6eadf32e5614c663cc1d56d546c73996cca614b715acf4a
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:ad027c91c719fd1d931cbac8c76735474f676b228f9ab23d1f243201246b3238
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:c4ff5dae4560e5a943c7218df3a5aa9ca1a9ee2d1f7dfac73db05892725c34a9
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:183ee860bf48076ea155db08032eed7004c1204a363625f54218734b00175441
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:45194b56f9fcb3af83158eb0297248364db40e8164fb4fe4f5656c58aeac3fa5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:c6862158a2a5ae0a4e4206bfdc7f73e1c2359544966c72cf4fc3a617acb13697
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:443b51f8b10acc85bdefde7193e0f45b1c423fa7fbdcaa28342805815c43db3d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:ea57c7cb1ccd5bce17f6cb35e326ab974718be8cd19b0cd38ea8112e91c51939
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:5f48081ffa453ec740c2562c24603d78b481b5e236d6bcc73e96203cabe153f2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:f06a1f50cd18d6f48762fdd0394307b44c0573d0cc91f53724c5b3c8c5e2b8e4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:45ca5370b726e31c55f9d7d249374189cb48dc05ab89d675c35225d53b90e7c9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:6bbcb64cf9adb4d3b6a4c259eb910e7da5c3c58732643e60d416399b2083cc0d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:fa01c7d11c23a47bb84375389e69619455c2879b5ed6bb5b9b753b945dcb02c3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:0522fc5482404fa9be179c7225a5692c8d71d755fda0b2d3928e452adcf030fb
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-bundle@sha256:552fc53664722903eccaa7d3782a3e84413509fd357725fe0b0f4a077c48ab2c
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:a72c2b70a9f0a93047d47247cddc1616138cbc87337289ddb208565af6e4369d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:7dbff355db7739152961cb6a97887d2e43bc960ac58837c126e03ed1a4480a3a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ee161954766211651d7cb13f3f5a3c025ac04094ba742be453ce789c6f0e24a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:aaee5240471872a8f1a77d4440a696d45041c5ce06849d516daccb90f5341b6d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:24dff64a80b910f0b4c5442d4f611f9df7602a16c5231eff2c9d08f07b2a9f9a
        version: 6.1.3
      entries:
      - name: ibm-zen-operator.v6.1.3
        version: 6.1.3
      name: v6.1
    defaultChannel: v6.1
    packageName: ibm-zen-operator
    provider:
      name: IBM
      url: https://www.ibm.com/in-en/products/cloud-pak-for-data
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: hpe-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-csi-operator.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.hpe.com/v1",
                "kind": "HPECSIDriver",
                "metadata": {
                  "name": "hpecsidriver-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "csp": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "disable": {
                    "alletra6000": false,
                    "alletra9000": false,
                    "alletraStorageMP": false,
                    "b10000FileService": false,
                    "nimble": false,
                    "primera": false
                  },
                  "disableHostDeletion": false,
                  "disableNodeConfiguration": false,
                  "disableNodeConformance": false,
                  "disableNodeGetVolumeStats": false,
                  "disableNodeMonitor": false,
                  "disablePreInstallHooks": false,
                  "imagePullPolicy": "IfNotPresent",
                  "images": {
                    "b10000FileServiceCSP": "quay.io/hpestorage/alletrastoragemp-b10000-nfs-csp@sha256:6a1855a2ef1a83b886bb1f45b596e8e420e5e13c84bbf446c8c202eaf27a2013",
                    "csiAttacher": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                    "csiControllerDriver": "quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676",
                    "csiExtensions": "quay.io/hpestorage/csi-extensions@sha256:27074aaa8df3b867cfa3d12318612c45f8e541d750c27db5e080734592160c89",
                    "csiNodeDriver": "quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676",
                    "csiNodeDriverRegistrar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                    "csiProvisioner": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                    "csiResizer": "registry.k8s.io/sig-storage/csi-resizer@sha256:8ddd178ba5d08973f1607f9b84619b58320948de494b31c9d7cd5375b316d6d4",
                    "csiSnapshotter": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769",
                    "csiVolumeGroupProvisioner": "quay.io/hpestorage/volume-group-provisioner@sha256:0a5f39f52113b963978a695bb8e5a5a3c4d387a02eb4d4a9647b4bc804de7399",
                    "csiVolumeGroupSnapshotter": "quay.io/hpestorage/volume-group-snapshotter@sha256:177d9abf8fbd3ac26ca4c12ca4b4adde20ebab308c79f10b04632f0171746fa8",
                    "csiVolumeMutator": "quay.io/hpestorage/volume-mutator@sha256:bf6beac1af5edcd3c9774f0dcc9579b79ef9cc8e4b74a4238082ab9aa443b5ed",
                    "nfsProvisioner": "quay.io/hpestorage/nfs-provisioner@sha256:5bf5952453ecca28e26724df85ebd5c8f0b7479ae4cb683d4a09c415da00e002",
                    "nimbleCSP": "quay.io/hpestorage/alletra-6000-and-nimble-csp@sha256:5480ffdebd74fdc547bc4a11b4434abac7bbfee3cee349f1ac893fbad84bb8d1",
                    "primera3parCSP": "quay.io/hpestorage/alletra-9000-primera-and-3par-csp@sha256:8af6e87075c47ba8d9714d55848ab2a0fef62b0768681e30d9e63ed7f396b0bf"
                  },
                  "iscsi": {
                    "chapSecretName": ""
                  },
                  "kubeletRootDir": "/var/lib/kubelet",
                  "logLevel": "info",
                  "maxVolumesPerNode": 100,
                  "node": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/hpestorage/csi-driver-operator@sha256:29e92f68cf0d7691a6c56876e7dde0d69a8cb3ff6ae6c9371c604e61cba16898
          createdAt: "2025-09-25T16:28:16Z"
          description: A Container Storage Interface (CSI) driver for HPE storage
            platforms. The CSI driver allows you to use HPE storage with your preferred
            container orchestrator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <3.0.1'
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/hpe-storage/co-deployments
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE CSI Driver for Kubernetes
            displayName: HPECSIDriver
            kind: HPECSIDriver
            name: hpecsidrivers.storage.hpe.com
            version: v1
        description: |
          The HPE CSI Operator for OpenShift deploys the HPE CSI Driver on Kubernetes and OpenShift for dynamic provisioning of persistent volumes on HPE storage systems.
          The HPE CSI Driver for Kubernetes leverages HPE storage platforms to provide scalable and persistent storage for stateful applications.
          Please refer to our CSI driver [documentation](https://scod.hpedev.io/csi_driver/index.html) for supported HPE platforms and full list of supported CSI features.
          ## Installation
          Refer to the HPE CSI Operator for OpenShift [documentation](https://scod.hpedev.io/csi_driver/deployment.html#operator).
        displayName: HPE CSI Operator for OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE CSI
        - HPE Storage
        - CSI
        links:
        - name: Documentation
          url: https://scod.hpedev.io/csi_driver/index.html
        maintainers:
        - email: dev-hi-containers@hpe.com
          name: HPE Storage Containers Team
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: HPE
        relatedImages:
        - quay.io/hpestorage/volume-mutator@sha256:bf6beac1af5edcd3c9774f0dcc9579b79ef9cc8e4b74a4238082ab9aa443b5ed
        - quay.io/hpestorage/alletrastoragemp-b10000-nfs-csp@sha256:6a1855a2ef1a83b886bb1f45b596e8e420e5e13c84bbf446c8c202eaf27a2013
        - quay.io/hpestorage/nfs-provisioner@sha256:5bf5952453ecca28e26724df85ebd5c8f0b7479ae4cb683d4a09c415da00e002
        - quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8ddd178ba5d08973f1607f9b84619b58320948de494b31c9d7cd5375b316d6d4
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769
        - quay.io/hpestorage/alletra-6000-and-nimble-csp@sha256:5480ffdebd74fdc547bc4a11b4434abac7bbfee3cee349f1ac893fbad84bb8d1
        - quay.io/hpestorage/volume-group-snapshotter@sha256:177d9abf8fbd3ac26ca4c12ca4b4adde20ebab308c79f10b04632f0171746fa8
        - quay.io/hpestorage/volume-group-provisioner@sha256:0a5f39f52113b963978a695bb8e5a5a3c4d387a02eb4d4a9647b4bc804de7399
        - quay.io/hpestorage/alletra-9000-primera-and-3par-csp@sha256:8af6e87075c47ba8d9714d55848ab2a0fef62b0768681e30d9e63ed7f396b0bf
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.connect.redhat.com/hpestorage/csi-driver-operator@sha256:29e92f68cf0d7691a6c56876e7dde0d69a8cb3ff6ae6c9371c604e61cba16898
        - quay.io/hpestorage/csi-extensions@sha256:27074aaa8df3b867cfa3d12318612c45f8e541d750c27db5e080734592160c89
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        version: 3.0.1
      entries:
      - name: hpe-csi-operator.v3.0.1
        version: 3.0.1
      name: stable
    defaultChannel: stable
    packageName: hpe-csi-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SmartX
      provider-url: https://www.smartx.com
    name: iomesh-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: iomesh-operator.0.10.1-rc4
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
          createdAt: 2021-08-17 00:00:00+00:00
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/iomesh-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/iomesh-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/iomesh/docs
          support: SmartX, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: iomesh
            displayName: iomesh
            kind: IOMeshCluster
            name: iomeshclusters.iomesh.com
            version: v1alpha1
        description: "IOMesh (https://www.iomesh.com/) is a cloud-native storage product
          specifically designed and developed for Kubernetes. It takes ZBS as its
          core, a “production-ready” distributed storage system developed independently
          by SmartX, to provide production-level high performance and reliable persistent
          storage for business-critical applications such as MySQL, Cassandra, MongoDB,
          etc., contributing to the containerized transformation of stateful applications.
          \     \n## Pre-Requisites\n* According to https://docs.iomesh.com/deploy/prerequisites,
          make sure that your openshift cluster meets the \"Installation Requirements\"
          and set up the worker nodes correctly according to the \"Setup Worker Node\"
          chapter.\n* Run IOMesh Operator pre-installation script in an environment
          where oc or kubectl can be used to access the openshift cluster, the script
          will install the dependencies of IOMesh Operator and specific IOMesh settings
          for the openshift cluster:\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-pre-install-openshift.sh
          | sh -\n```\n## Install IOMesh Operator and IOMesh Cluster\n* Click install
          button on the current web page to install IOMesh Operator\n* On the **Installed
          Operators** -> **IOMesh Operator** -> **Create instance** -> **YAML view**,
          fill in an IOMesh Custom Resources according https://iomesh.run/iomesh.yaml,
          change the `spec.*.dataCIDR` to your own data network CIDR:\n## Post Install\n*
          Run IOMesh Operator post-installation script in an environment where oc
          or kubectl can be used to access the openshift cluster, the script will
          install IOMesh CSI Driver\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-post-install-openshift.sh
          | sh -\n```\n## Getting Started\n* Try to quickstart setup and use IOMesh:
          https://docs.iomesh.com/deploy/setup-iomesh\n"
        displayName: IOMesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iomesh-operator
        links:
        - name: IOMesh Operator
          url: https://iomesh.com
        maintainers:
        - email: na.wang@smartx.com
          name: Na Wang
        maturity: alpha
        provider:
          name: SmartX
          url: https://www.smartx.com
        relatedImages:
        - iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
        version: 0.10.1-rc4
      entries:
      - name: iomesh-operator.0.10.1-rc4
        version: 0.10.1-rc4
      name: stable
    defaultChannel: stable
    packageName: iomesh-operator-rhmp
    provider:
      name: SmartX
      url: https://www.smartx.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: openstack-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openstack-operator.v1.0.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openstack.org/v1beta1",
                "kind": "OpenStack",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openstack-operator",
                    "app.kubernetes.io/instance": "openstack",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "openstack",
                    "app.kubernetes.io/part-of": "openstack-operator"
                  },
                  "name": "openstack",
                  "namespace": "openstack-operators"
                },
                "spec": {
                  "operatorOverrides": [
                    {
                      "controllerManager": {
                        "resources": {
                          "limits": {
                            "cpu": "600m",
                            "memory": "2Gi"
                          }
                        }
                      },
                      "name": "infra",
                      "replicas": 1
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-02T13:33:09Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.14
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.openstack.org/v1beta1","kind":"OpenStack","metadata":{"name":"openstack","namespace":"openstack-operators"},"spec":{}}'
          operatorframework.io/suggested-namespace: openstack-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenStack is the Schema for the openstacks API
            displayName: Open Stack
            kind: OpenStack
            name: openstacks.operator.openstack.org
            version: v1beta1
        description: Install and configure OpenStack
        displayName: OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        links:
        - name: Openstack Operator
          url: https://github.com/openstack-k8s-operators/
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/glance-rhel9-operator@sha256:ee22cf417b7771930a442c2302277e6d37f368f5ae551fee9a50dbbe41c8fe3a
        - registry.redhat.io/rhoso-operators/mysqld-exporter-rhel9@sha256:b49e966cdd588071e0e1f2791ec6c62310078a7d39ea77f83843d3708a07c2b5
        - registry.redhat.io/rhoso/openstack-swift-proxy-server-rhel9@sha256:858551fd1c519797cae0e146e9c05412ed6e66dac01d0e68e62dd2d4bd335391
        - registry.redhat.io/rhoso/openstack-netutils-rhel9@sha256:b02ab1617922a2560bf739172bc2ce0f6b1fb3ef55054bd17e4b16bdf9512b4b
        - registry.redhat.io/rhoso/openstack-cinder-backup-rhel9@sha256:910693233ee3a1f3832ebe744f67a5139249523456de4abbd731c9d2f544b7c3
        - registry.redhat.io/rhoso/ironic-python-agent-rhel9@sha256:2ccd8e22a242059ca5c4e2bce853c36a71c36fdf40e8e36ab183a8218b0cc87b
        - registry.redhat.io/rhoso/openstack-frr-rhel9@sha256:f6516277149d1560c0bf2ac74a99cfb4ac2c84f9f090ab231b6d12bdb824de3b
        - registry.redhat.io/rhoso/openstack-octavia-housekeeping-rhel9@sha256:0a79259e398af746d0e8a6dda1aae5ee976c6bf8991b17f4a5da0d07dca1e36c
        - registry.redhat.io/rhoso/openstack-swift-object-rhel9@sha256:8fc2ab8caa9c7c35d94ef51c303763ea71e40b9c795c4555c17e6194344419b3
        - registry.redhat.io/ubi9/httpd-24@sha256:29382518403a4361c48fccc057163dc4a9f43ee159432754fa63c76485b223f5
        - registry.redhat.io/rhoso-operators/heat-rhel9-operator@sha256:583abb81a49db85949f7a4b846591ad8532afa79440ad7e2937ee00eacd972b3
        - registry.redhat.io/rhoso/openstack-barbican-keystone-listener-rhel9@sha256:cd8d4f0a94b4fe080ebf8e459f091db6637ceec6d83ac23cf04d66bd03c71e76
        - registry.redhat.io/rhoso/openstack-neutron-metadata-agent-ovn-rhel9@sha256:20323b151474703df610b6c24ccb9c748265196bbc70a7b5ece1b801ff540cd9
        - registry.redhat.io/rhoso-operators/test-rhel9-operator@sha256:df7566483abb18aa3c45d8bc4cd91531c7eb5b63345e511151a7cdca2dbcef17
        - registry.redhat.io/rhoso/openstack-iscsid-rhel9@sha256:508bd67021c271a10bb55ab4af033dc4e931708674b9c5ba94f72c07cd0c8ee2
        - registry.redhat.io/rhoso-operators/openstack-baremetal-agent-rhel9@sha256:08675b16d89d7931b1ef7b96e1e884c88e8e7fbfbe79a944f9213c52aba3ba83
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/rhoso/openstack-ironic-neutron-agent-rhel9@sha256:0abaa8f437025d2e7006c59c546ff6935556d707b57993719faef94142e1c43d
        - registry.redhat.io/rhoso/openstack-aodh-evaluator-rhel9@sha256:48c9192acf157df5b8f0827e35f0380d50cd4562676aba3c0d7791f92328af02
        - registry.redhat.io/rhoso/openstack-manila-scheduler-rhel9@sha256:8ce47f7fd8660167b3b292659d4fc58cf34ed671d5779693d24b71a9fddc1e7c
        - registry.redhat.io/rhoso/openstack-neutron-dhcp-agent-rhel9@sha256:cc9223ef705bcd5d0dcabbf71ad9f6e540ae7f003a6afd13161301f842dfa5fa
        - registry.redhat.io/rhoso/edpm-hardened-uefi-rhel9@sha256:c18ec7023fbcc4ab5183f689adc1f01a626a4923c15c61a37388f99904db7bc3
        - registry.redhat.io/rhoso/openstack-ceilometer-central-rhel9@sha256:e6d9f47406b7e3ff2a7f2256088cc64f043b8e497382b9f569eb90182598bbe9
        - registry.redhat.io/rhoso/openstack-cinder-api-rhel9@sha256:72a674caeeeecce3b60217f3607ad5b00651fa3d52f2f5d905f6e6395638dbc8
        - registry.redhat.io/rhoso/openstack-glance-api-rhel9@sha256:b1c0b09bb9aafe3bc31f43f8a99a92dde8d78cf90e781dd14300580986d8c980
        - registry.redhat.io/rhoso-operators/designate-rhel9-operator@sha256:d446af622bd32666eb4eb5f9b7f38f0c33ac81251a2230db9d5baf71081b9d6a
        - registry.redhat.io/rhoso/openstack-designate-producer-rhel9@sha256:04af4230f815a6e3fde22003cf382eebd9cbc3c3ec4056795f4474abc1e65fea
        - registry.redhat.io/rhoso-operators/mariadb-rhel9-operator@sha256:907e3a9d15941440d5289d7e448d9f0351053fc20fcb8cb3b9db0a84debd876f
        - registry.redhat.io/rhoso/openstack-manila-api-rhel9@sha256:e4dd3d5b66c5216b45a1515a975b85fd2ac4ef698a9350dd16e01417da462b32
        - registry.redhat.io/rhoso/openstack-cinder-volume-rhel9@sha256:8a09e80dac94fb4b3d35033e91812f95f0acac66532d9abc88aa1fb715d21822
        - registry.redhat.io/rhoso/openstack-designate-worker-rhel9@sha256:a6324a8a72722822df7692b8df0420388a1b4aecfff3da7e91da4afe4d9e2260
        - registry.redhat.io/rhoso/openstack-ceilometer-compute-rhel9@sha256:296f7f6c6bdcd7f5d07b8470308a4c71055f29a17c48d2ebb03b7f54a0e927ff
        - registry.redhat.io/rhoso/openstack-heat-api-cfn-rhel9@sha256:2789929ae3edd7c68579f34457d3071273e49a66f462043cd9c46e75bdda282a
        - registry.redhat.io/openshift4/ose-kube-state-metrics-rhel9@sha256:db6074b8aeedfd7603ae3adbe67071115f54e7f01a24997f1cd8fa8958aeddce
        - registry.redhat.io/rhoso-operators/cinder-rhel9-operator@sha256:12b8dd8fd3683250365eff6f07339cb976ee29da80642e591c10e9fa5ba627c9
        - registry.redhat.io/rhoso/openstack-heat-engine-rhel9@sha256:151e04b3e5a408aadef3a17954255acc3733d09961878a135e177188457447f0
        - registry.redhat.io/rhoso/openstack-rsyslog-rhel9@sha256:272e729accc2b64b8fcc0ece4ae5aa9ce6d454e34a25a8a8e2b99ac56b73361e
        - registry.redhat.io/rhoso/openstack-tempest-all-rhel9@sha256:16b1fbfd36ce9a27106bf09c414eaa93f14c9a6b11839068976c2fdb537d2ba7
        - registry.redhat.io/rhoso/openstack-octavia-api-rhel9@sha256:7b8d5927162ddab512096beb02e837ba51f6353c5187f3de72eb3db6e56ab6e7
        - registry.redhat.io/rhoso/openstack-keystone-rhel9@sha256:ac076c81247d774a9aeee1b63540ca6e18058213cf6aad10a6fd1f535a394ef4
        - registry.redhat.io/rhoso/openstack-cinder-scheduler-rhel9@sha256:2079ffe9b620e9277573ac796bf7e393b520631130e5bbb6f5ab51d872c943f3
        - registry.redhat.io/rhoso/openstack-aodh-notifier-rhel9@sha256:d2e89f27aac9afa504411ae6a6163442cd717f8c560e505a7da8e3fd9e2cff95
        - registry.redhat.io/rhoso/openstack-manila-share-rhel9@sha256:277174c885998e6f5b867d099664f6fa28f974bffb524903ff75283035cd8dc9
        - registry.redhat.io/rhoso/openstack-neutron-server-rhel9@sha256:2d3afde3201aa18b192ad0fb079c1e4f4f673565604f1af2b79aba8a7bc88bf2
        - registry.redhat.io/rhoso/openstack-ovn-northd-rhel9@sha256:668f742b2a18579bc8cfc610a862e43c5b877413cf65eabbe0ee36b3c7869446
        - registry.redhat.io/rhoso-operators/openstack-baremetal-rhel9-operator@sha256:03916e507b88943c0e00cedeb3cf20fa0787ccc762e47f38d305ff4d1b125b12
        - registry.redhat.io/rhoso/openstack-ovn-nb-db-server-rhel9@sha256:8c2ce1027779dc66d67a0e0d03e48b226d3278208cae0b2b81b35ad1b1f3ae72
        - registry.redhat.io/rhoso/openstack-neutron-ovn-agent-rhel9@sha256:bed1956c66b0e70064b51e34faac64036f77c7ecb14f46ceac4bc9ef52f0b9e0
        - registry.redhat.io/rhoso/openstack-nova-scheduler-rhel9@sha256:f0e5d3a693c9998ca3839b3ec7dfd5da0f9cf139e0b5b6913e92a10d5a13e6bc
        - registry.redhat.io/rhoso/openstack-designate-central-rhel9@sha256:d28d8d91d8bf75a1b77a5d650c7f2650a1791e2a1b0341450fa46978c955b92f
        - registry.redhat.io/rhoso/openstack-barbican-api-rhel9@sha256:bc942204843c45c6cb7ea841af8611f04efb003de7eeccc6813934aad556da73
        - registry.redhat.io/rhoso/openstack-octavia-health-manager-rhel9@sha256:8065851058019757b56f86ca4f030d44f169100eb2e37e08e0587a7b1e4d2e72
        - registry.redhat.io/rhoso/openstack-nova-api-rhel9@sha256:4116206f1fbe60fdd9458a5edee848571c94e2b0512d9c2cfcfd6890d60d4395
        - registry.redhat.io/rhoso-operators/keystone-rhel9-operator@sha256:879d4b1f9e720f8581b3a55c788a97bbe7b52f9108df3d5dce2e9bcb4200df6d
        - registry.redhat.io/rhoso/openstack-horizon-rhel9@sha256:82d0e6d24733afd305996e0f4a118df3e48621c40c9669b01f48669dbf40d868
        - registry.redhat.io/rhoso/openstack-designate-api-rhel9@sha256:b76ca7820553285bb4ee2bfb08e39aedbfe8c16b4edb478c06418a1e335a64ab
        - registry.redhat.io/rhoso/openstack-swift-container-rhel9@sha256:95fec49643fc18f402b8ba8057f2c35f214bf6138c5f0d7bbe1b96cbd78ef8d6
        - registry.redhat.io/rhoso-operators/swift-rhel9-operator@sha256:c99ef14338290d63e2bcbf129e953b434aea382e47072812bae31b1ea838e7df
        - registry.redhat.io/rhoso/openstack-ovn-base-rhel9@sha256:a3b8ec7ad1073facb8876acc787549a1a68b9bf81bebdc127f13eb6385e2d1da
        - registry.redhat.io/rhoso/openstack-ovn-bgp-agent-rhel9@sha256:a404c3416ab65e7134f438362e887529be474639e725e18f3542e86973edac87
        - registry.redhat.io/rhoso/openstack-nova-novncproxy-rhel9@sha256:04881867f9b24403bf6ef051eee3902a17e8476eee894ef579c0d5d130d2510d
        - registry.redhat.io/rhoso/openstack-ironic-inspector-rhel9@sha256:88357e83e8a4bbc3c5a37f0420ad75556821834a975cd8a1cf5fee56de2a4f99
        - registry.redhat.io/rhoso/openstack-ovn-controller-rhel9@sha256:3c42b7e0305c508d74602b9378143deb07a569443bba35850cdb7b6e3f5202e2
        - registry.redhat.io/rhoso/openstack-nova-conductor-rhel9@sha256:c67ca0e6c8126f56c6a83632b6a96bd4a35006f5f414a8779467b81864b31d12
        - registry.redhat.io/rhoso-operators/ee-openstack-ansible-ee-rhel9@sha256:61e729fb89c4872837cbe9cb1e3cea402b61f0da240751302605df487721f0aa
        - registry.redhat.io/rhoso-operators/ovn-rhel9-operator@sha256:66cd8add6078b63a36879c974db9e9f358593bb0651aa39fe85153d25404ae27
        - registry.redhat.io/rhoso/openstack-designate-backend-bind9-rhel9@sha256:82cca999357b0ab25bcc081bf802fab831587cf661db806af4ca453e0a626743
        - registry.redhat.io/rhoso/openstack-rabbitmq-rhel9@sha256:56bc2d2b201225fb98db43ea9ca68cceb4842f8f0f1a96ab3ab62ab785651d12
        - registry.redhat.io/rhoso-operators/prometheus-podman-exporter-rhel9@sha256:f33e7506db6dfdf530d3e2442baaf360d212bb1a47a7db887dba967ed0655cc2
        - registry.redhat.io/rhoso/openstack-nova-compute-rhel9@sha256:6264075310231af358803b53d5a09250dab32d63fc5966aa8fb957e5f56096a8
        - registry.redhat.io/rhoso-operators/neutron-rhel9-operator@sha256:30b6a812c47db9e6691211185c304c3e8fc6a8b9ba7ca79acf59c5ec07fb9fb0
        - registry.redhat.io/rhoso-operators/openstack-must-gather-rhel9@sha256:60b4355f4dbef3ad64103357086b10a5c0af126dd23798d6a9d725646f1e13b4
        - registry.redhat.io/rhoso/openstack-ironic-api-rhel9@sha256:50ccfe7d9ba22a50e0c1c8159cade6f4b1ac13ad8f5770f20c7be18fba58c8c3
        - registry.redhat.io/rhoso/openstack-ironic-pxe-rhel9@sha256:42e35ded74033f1f07287441917829cc934930a5df696c33cd8957be3c10fa41
        - registry.redhat.io/rhoso/openstack-openstackclient-rhel9@sha256:2952521d4b1799fd32ff773ef2d457a3abc354b639cf7f1e0f7f6615ca35ac8f
        - registry.redhat.io/rhoso-operators/manila-rhel9-operator@sha256:d9e69eb140b8ff76d5bf49deab8d700639ccdfd1ab5c8c7667bc2bbb50f3d60b
        - registry.redhat.io/rhoso-operators/horizon-rhel9-operator@sha256:9a2a13635588c6cc9949b272e1a1708c5fb53f19b05b8310ace703df5958ac63
        - registry.redhat.io/rhoso-operators/nova-rhel9-operator@sha256:3156dde249e49b40bed71092fb5213b112d43b663ba8c556e355aed524bec822
        - registry.redhat.io/rhoso-operators/ironic-rhel9-operator@sha256:dd67ffe3578ce619748735c1a5cd2a57d704d0c36c1b841d513ea36479aa4682
        - registry.redhat.io/rhoso/openstack-aodh-api-rhel9@sha256:4adc09cc1d922b387320da467de89d3f41a9c49b46d9b728458fec36505ec940
        - registry.redhat.io/rhoso/openstack-cron-rhel9@sha256:f09e724c69cec2c3b685783f9263b26c280a57b5909a723341416b342d18f9a4
        - registry.redhat.io/rhoso/openstack-ironic-conductor-rhel9@sha256:538da9fdeec546d794d192361d0a76f8b4e7682b6eb6d9cea371d8a52d603800
        - registry.redhat.io/rhoso/openstack-swift-account-rhel9@sha256:31e727f37db92fcc21319c44396576a03a118ba25d403715650162f7c0ead31e
        - registry.redhat.io/rhoso/openstack-redis-rhel9@sha256:c6dbeacbbd5aed9d13e540f1c987a574ea45747983c280446e5afadcf7c33b37
        - registry.redhat.io/rhoso/openstack-ovn-sb-db-server-rhel9@sha256:8ba4f8ed49e214f421e94133da0f17b54ce99a52bbbd367d0aef357de844ea01
        - registry.redhat.io/openshift4/ose-prometheus-node-exporter-rhel9@sha256:479c51de2fc8a3b3858d71bbe4eeafbd5b43263472346f9dde348115082894cc
        - registry.redhat.io/rhoso-operators/octavia-rhel9-operator@sha256:aa0f00f8e03dfdb547e53cf611581e9cc091eaf049558121c2bc5d702e556bcb
        - registry.redhat.io/rhoso/openstack-mariadb-rhel9@sha256:6115a7458cfa412c1659242fcec64b63f0a5ca414ecf3c5f890f0afec0bca069
        - registry.redhat.io/rhoso/openstack-multipathd-rhel9@sha256:6ff8ae176f8a633bcd78b1f8407924c74bb2f317eba16b90e7f0f9e53af74e8e
        - registry.redhat.io/rhoso/openstack-neutron-sriov-agent-rhel9@sha256:118d5f9430192ab48cf159ee2f7912f08e935ad56ef3364f18b3680210bb3e52
        - registry.redhat.io/rhoso-operators/barbican-rhel9-operator@sha256:b3b769e0dc32c12c5cc9578c2cac2622b973177ca14697bc1042d21d473e4e84
        - registry.redhat.io/rhoso/openstack-heat-api-rhel9@sha256:ccbc41b552bb34d223c9615a4e3b56bb516bc5bd8a694f048c15063ba45fca9f
        - registry.redhat.io/rhoso-operators/telemetry-rhel9-operator@sha256:0635515a5fed12f17de4cc5ac5f9567e3b3546fc2f69730c7beab25cf722f538
        - registry.redhat.io/rhoso/openstack-aodh-listener-rhel9@sha256:037845be6787f610e479345206b0fcfb2f1dc430540816bc5a54a1d26da2acbd
        - registry.redhat.io/rhoso/openstack-ceilometer-notification-rhel9@sha256:bd2b5f50a2bef4e77e3daa89aca51df6e0e55c3784886296d92c97616cd566d1
        - registry.redhat.io/rhoso-operators/sg-core-rhel9@sha256:b56923f869ff57396e6822133dfc3b21007c3df652cfe962c7ad3028214d988c
        - registry.redhat.io/rhoso/openstack-barbican-worker-rhel9@sha256:485525116fb3c22d8c60330dd87ebd3854cd71ded36bf88bcf18cee5957e22a3
        - registry.redhat.io/rhoso-operators/openstack-network-exporter-rhel9@sha256:8e8e775e37c28cb11b7aae9199405b85d5862dce195d165ca598fa1dc89bf914
        - registry.redhat.io/rhoso/openstack-placement-api-rhel9@sha256:0aa5e138947bf236f31c81bfb6f641d95af63812c06dca43141721b383c7e08e
        - registry.redhat.io/openshift-power-monitoring/kepler-rhel9@sha256:8e9e5bfeed61aaeee85bf5eefc3390556b8d4cbfc797a82cabe95bc0367126eb
        - registry.redhat.io/rhoso/openstack-ceilometer-ipmi-rhel9@sha256:2fb5723c68d0f2c3ae28b86e606ccae4b80f3bd9b16c01dd6c9adcae629db7f1
        - registry.redhat.io/rhoso-operators/openstack-rhel9-operator@sha256:64e2ad312f757101f083eedcb17e0b6c05cf6ebf6ed9e7db2e33e84c6aca9898
        - registry.redhat.io/rhoso/openstack-unbound-rhel9@sha256:7f17a719e8e49d457ff16437709c2b4eca5b9c2f58972d1e351c5c1c1ac38962
        - registry.redhat.io/rhoso-operators/infra-rhel9-operator@sha256:a2dc031a6ad2726d155e77cafb676f8e35a86e963b97ccdceed0da137e67ba0d
        - registry.redhat.io/rhoso/openstack-designate-mdns-rhel9@sha256:ce43dd1a30afd4753aa11d32df2b24ed4bf3d67847044aaae8f3e5c8901caf05
        - registry.redhat.io/rhoso/openstack-octavia-worker-rhel9@sha256:303e35926e43fe403673e18a25c2e815547c28a5f8c0f09d470cd4148a21c528
        - registry.redhat.io/rhoso-operators/placement-rhel9-operator@sha256:a16bd4fb929eb5bcf96f2aefcb5a7abf4143f4f9ed24e53de93f4a3caa89a507
        - registry.redhat.io/rhoso-operators/rabbitmq-cluster-rhel9-operator@sha256:e095e6cbda841b9c726c1ec0839a98964b894cf25ae9228052d314b1ee2f161d
        - registry.redhat.io/rhoso/openstack-memcached-rhel9@sha256:3e704b9664c65be39e83c2ed802722ffeaf969d8e571147a2be128d5c56331c1
        version: 1.0.14
      entries:
      - name: openstack-operator.v1.0.14
        version: 1.0.14
      - name: openstack-operator.v1.0.13
        version: 1.0.13
      - name: openstack-operator.v1.0.12
        version: 1.0.12
      - name: openstack-operator.v1.0.11
        version: 1.0.11
      - name: openstack-operator.v1.0.10
        version: 1.0.10
      - name: openstack-operator.v1.0.9
        version: 1.0.9
      - name: openstack-operator.v1.0.8
        version: 1.0.8
      - name: openstack-operator.v1.0.7
        version: 1.0.7
      - name: openstack-operator.v1.0.6
        version: 1.0.6
      - name: openstack-operator.v1.0.5
        version: 1.0.5
      - name: openstack-operator.v1.0.4
        version: 1.0.4
      - name: openstack-operator.v1.0.3
        version: 1.0.3
      - name: openstack-operator.v1.0.2
        version: 1.0.2
      - name: openstack-operator.v1.0.1
        version: 1.0.1
      - name: openstack-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: openstack-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Devfile
      provider-url: https://devfile.io
    name: devworkspace-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devworkspace-operator.v0.36.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/internal-objects: '["devworkspaceroutings.controller.devfile.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DevWorkspaceOperatorConfig
            name: devworkspaceoperatorconfigs.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspaceRouting
            name: devworkspaceroutings.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha2
          - kind: DevWorkspaceTemplate
            name: devworkspacetemplates.workspace.devfile.io
            version: v1alpha1
          - kind: DevWorkspaceTemplate
            name: devworkspacetemplates.workspace.devfile.io
            version: v1alpha2
        description: |
          The DevWorkspace Operator enables cluster-level support for the
          [Devfile 2.0 spec](https://docs.devfile.io), enabling static, reproducible
          configurations for launching cloud-based editors and IDEs in OpenShift and
          Kubernetes clusters.

          Leveraging the language server protocol and the Eclipse Theia web IDE, the
          DevWorkspace operator provides easy configuration of full development
          environments on the cloud with support for a wide variety of languages and
          technologies, including Go, Java, Typescript/Javascript, Python, and more.

          The DevWorkspace Operator is also used in the Web Terminal Operator to
          automatically provision Web Terminal environments.

          ## Installing the operator
          The DevWorkspace Operator can be installed directly from the OperatorHub UI and
          will be available in all namespaces on the cluster. DevWorkspace creation is
          driven by the DevWorkspace custom resource, which can be created in any
          namespace to provision a full development environment. To get started, browse
          the DevWorkspace [spec](https://devfile.io/docs/2.3.0/devfile-schema).

          Once a DevWorkspace is started, it can be accessed via the URL in its
          `.status.mainUrl` field.

          It's recommended to install the DevWorkspace Operator to the
          `openshift-operators` namespace for compatibility.

          ## Uninstalling the operator
          The DevWorkspace Operator utilizes finalizers on resources it creates and
          webhooks to restrict access to development resources. As a result, manual steps
          are required in uninstalling the operator. See the
          [documentation](https://github.com/devfile/devworkspace-operator/blob/main/docs/uninstall.md)
          for details.
        displayName: DevWorkspace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Devworkspace Operator
        - DevWorkspaces
        - Devfile
        links:
        - name: Devworkspace Operator
          url: https://github.com/devfile/devworkspace-operator
        maintainers:
        - email: dakwon@redhat.com
          name: David Kwon
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        maturity: alpha
        provider:
          name: Devfile
          url: https://devfile.io
        relatedImages:
        - registry.redhat.io/devworkspace/devworkspace-rhel9-operator@sha256:83b6abf6e57703da1125e037ea9353c224602bddc3146e5b00df47f1b1976e9a
        - registry.redhat.io/ubi9/ubi-minimal@sha256:a50731d3397a4ee28583f1699842183d4d24fadcc565c4688487af9ee4e13a44
        - registry.redhat.io/devworkspace/devworkspace-project-clone-rhel9@sha256:312a013201877cea03956e3db2ba19a3d7f842de2c3a65927c65c3e8d22a57fe
        version: 0.36.0
      entries:
      - name: devworkspace-operator.v0.36.0
        version: 0.36.0
      - name: devworkspace-operator.v0.35.1
        version: 0.35.1
      - name: devworkspace-operator.v0.34.0
        version: 0.34.0
      - name: devworkspace-operator.v0.33.0
        version: 0.33.0
      - name: devworkspace-operator.v0.32.1
        version: 0.32.1
      - name: devworkspace-operator.v0.31.2
        version: 0.31.2
      - name: devworkspace-operator.v0.31.1
        version: 0.31.1
      - name: devworkspace-operator.v0.31.0
        version: 0.31.0
      - name: devworkspace-operator.v0.30.2
        version: 0.30.2
      - name: devworkspace-operator.v0.30.1
        version: 0.30.1
      - name: devworkspace-operator.v0.30.0
        version: 0.30.0
      - name: devworkspace-operator.v0.29.0
        version: 0.29.0
      - name: devworkspace-operator.v0.28.0-0.1719314743.p
        version: 0.28.0+0.1719314743.p
      - name: devworkspace-operator.v0.28.0
        version: 0.28.0
      - name: devworkspace-operator.v0.27.0-0.1714987663.p
        version: 0.27.0+0.1714987663.p
      - name: devworkspace-operator.v0.27.0
        version: 0.27.0
      - name: devworkspace-operator.v0.26.0
        version: 0.26.0
      - name: devworkspace-operator.v0.25.0
        version: 0.25.0
      - name: devworkspace-operator.v0.24.0
        version: 0.24.0
      - name: devworkspace-operator.v0.23.0
        version: 0.23.0
      - name: devworkspace-operator.v0.22.0
        version: 0.22.0
      - name: devworkspace-operator.v0.21.2
        version: 0.21.2
      - name: devworkspace-operator.v0.21.1
        version: 0.21.1
      - name: devworkspace-operator.v0.20.0
        version: 0.20.0
      - name: devworkspace-operator.v0.19.1
        version: 0.19.1
      - name: devworkspace-operator.v0.19.1-0.1682321189.p
        version: 0.19.1+0.1682321189.p
      - name: devworkspace-operator.v0.19.1-0.1679521112.p
        version: 0.19.1+0.1679521112.p
      - name: devworkspace-operator.v0.18.1
        version: 0.18.1
      - name: devworkspace-operator.v0.18.1-0.1675929565.p
        version: 0.18.1+0.1675929565.p
      - name: devworkspace-operator.v0.17.0
        version: 0.17.0
      - name: devworkspace-operator.v0.16.0-0.1666668361.p
        version: 0.16.0+0.1666668361.p
      - name: devworkspace-operator.v0.16.0
        version: 0.16.0
      - name: devworkspace-operator.v0.15.2-0.1661828401.p
        version: 0.15.2+0.1661828401.p
      - name: devworkspace-operator.v0.15.2
        version: 0.15.2
      - name: devworkspace-operator.v0.14.1
        version: 0.14.1
      - name: devworkspace-operator.v0.13.0
        version: 0.13.0
      - name: devworkspace-operator.v0.12.0
        version: 0.12.0
      - name: devworkspace-operator.v0.11.0
        version: 0.11.0
      - name: devworkspace-operator.v0.10.0
        version: 0.10.0
      - name: devworkspace-operator.v0.9.0
        version: 0.9.0
      name: fast
    defaultChannel: fast
    packageName: devworkspace-operator
    provider:
      name: Devfile
      url: https://devfile.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microcks.io
      provider-url: ""
    name: microcks
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: microcks-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "microcks.github.io/v1alpha1",
                "kind": "MicrocksInstall",
                "metadata": {
                  "name": "my-microcksinstall"
                },
                "spec": {
                  "name": "my-microcksinstall",
                  "version": "1.10.0",
                  "microcks": {
                    "replicas": 1
                  },
                  "postman": {
                    "replicas": 1
                  },
                  "keycloak": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "1Gi"
                  },
                  "mongodb": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "2Gi"
                  }
                }
              },
              {
                "apiVersion": "microcks.github.io/v1alpha1",
                "kind": "MicrocksInstall",
                "metadata": {
                  "name": "my-microcksinstall-minikube"
                },
                "spec": {
                  "name": "my-microcksinstall-minikube",
                  "version": "1.10.0",
                  "microcks": {
                    "replicas": 2,
                    "url": "microcks.192.168.99.100.nip.io"
                  },
                  "postman": {
                    "replicas": 2
                  },
                  "keycloak": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "1Gi",
                    "url": "keycloak.192.168.99.100.nip.io"
                  },
                  "mongodb": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "2Gi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/microcks/microcks-ansible-operator:1.10.0
          createdAt: "2024-08-01T09:14:55Z"
          description: Open Source mocking and testing platform for API and microservices
          repository: https://github.com/microcks/microcks-ansible-operator
          support: microcks.github.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Microcks installation
            displayName: MicrocksInstall
            kind: MicrocksInstall
            name: microcksinstalls.microcks.github.io
            version: v1alpha1
        description: |
          Microcks is an Open source tool for API and microservices mocking and testing. It provides a platform for referencing, deploying mocks and allow contract-testing of your APIs. It can also be considered as a Service Virtualization solution because it will allow you to provide fake API or Service implementation before development being actually done. It supports Async APIs, REST APIs, gRPC, GraphQL and SOAP WebServices. It perfectly integrates into an iterative, contract-first delivery process.
          Microcks tries not reinventing the wheel and let you capitalize on already known tools of standards. It supports [AsyncAPI Specifications](https://asyncapi.com), [OpenAPI Specification](https://www.openapis.org/) `3.x`, [gRPC Protobuf](https://grpc.io/), [GraphQL Schema](https://graphql.org) out-of-the-box and can also relies on popular tools like [SoapUI](http://www.soapui.com) or [Postman](http://www.getpostman.io) to edit your API request / response samples as well as your test scripts.
          ### Operator features
          * **Create/Destroy**: Easily launch Microcks installations (`MicrocksInstall` custom resource) for your Kubernetes namespace
          ### MicrocksInstall options
          * **keycloak** component can be deployed as part of the install or you can reuse an existing instance. You can reference an existing `Secret` for providing TLS certificates for securing the `Ingress`,
          * **mongodb** component can be deployed as part of the install or you can reuse an existing instance, You can reference an existing `Secret` for providing connection credentials,
          * **features** component allow to activate optional features like the mocking of AsyncAPIs. In case you turn `features.async.enabled` to `true` and `features.async.kafka.install` to `true`, you must have the [Strimzi Operator](https://operatorhub.io/operator/strimzi-kafka-operator) already installed on your cluster/namespace.
          ### Documentation
          Documentation to the current _master_ branch as well as all releases can be found on our [GitHub](https://github.com/microcks/microcks-ansible-operator).
          ### License
          Microcks is licensed under the [Apache License, Version 2.0](https://github.com/microcks/microcks/blob/master/LICENSE).
        displayName: Microcks Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mock
        - test
        - API
        - microservice
        links:
        - name: Product Page
          url: https://microcks.io
        - name: Microcks Operator Source Code
          url: https://github.com/microcks/microcks-ansible-operator
        maintainers:
        - email: laurent.broudoux@gmail.com
          name: lbroudoux
        maturity: alpha
        provider:
          name: Microcks.io
        relatedImages:
        - quay.io/microcks/microcks-ansible-operator:1.10.0
        version: 1.10.0
      entries:
      - name: microcks-operator.v1.10.0
        version: 1.10.0
      - name: microcks-operator.v1.9.0
        version: 1.9.0
      - name: microcks-operator.v1.8.1
        version: 1.8.1
      - name: microcks-operator.v1.8.0
        version: 1.8.0
      - name: microcks-operator.v1.7.1
        version: 1.7.1
      - name: microcks-operator.v1.7.0
        version: 1.7.0
      - name: microcks-operator.v1.6.1
        version: 1.6.1
      - name: microcks-operator.v1.6.0
        version: 1.6.0
      - name: microcks-operator.v1.5.2
        version: 1.5.2
      - name: microcks-operator.v1.5.1
        version: 1.5.1
      - name: microcks-operator.v1.5.0
        version: 1.5.0
      - name: microcks-operator.v1.4.1
        version: 1.4.1
      name: stable
    defaultChannel: stable
    packageName: microcks
    provider:
      name: Microcks.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: ""
    name: pmem-csi-operator-os
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pmem-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pmem-csi.intel.com/v1beta1",
                "kind": "PmemCSIDeployment",
                "metadata": {
                  "name": "pmem-csi.intel.com"
                },
                "spec": {
                  "deviceMode": "lvm",
                  "nodeSelector": {
                    "feature.node.kubernetes.io/memory-nv.dax": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
          createdAt: "2021-08-26T10:45:23Z"
          description: An operator for deploying and managing the PMEM-CSI driver
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: ""
          repository: https://github.com/intel/pmem-csi/tree/release-1.0
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a PMEM-CSI driver deployment in the cluster.
            displayName: PmemCSIDeployment
            kind: PmemCSIDeployment
            name: pmemcsideployments.pmem-csi.intel.com
            version: v1beta1
        description: |2

          Intel PMEM-CSI is a [CSI](https://github.com/container-storage-interface/spec)
          storage driver for container orchestrators like Kubernetes. It makes local
          persistent memory ([PMEM](https://pmem.io/)) available as a filesystem volume to
          container applications. This is the operator to deploy and manage the
          [PMEM-CSI](https://intel.github.io/pmem-csi/1.0/README.html)
          driver on a Kubernetes cluster.

          Refer to the [PMEM-CSI documentation](https://intel.github.io/pmem-csi/1.0/README.html)
          for more details on deploying and using the PMEM-CSI driver.
        displayName: Operator for PMEM-CSI driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - PMEM-CSI
        - PMEM-CSI operator
        - Persistent Memory
        - PMEM
        links:
        - name: PMEM-CSI
          url: https://intel.github.io/pmem-csi/1.0
        maintainers:
        - email: olev.kartau@intel.com
          name: Olev Kartau
        - email: patrick.ohly@intel.com
          name: Patrick Ohly
        - email: amarnath.valluri@intel.com
          name: Amarnath Valluri
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Intel® Corporation
        relatedImages:
        - registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
        version: 1.0.0
      entries:
      - name: pmem-csi-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: pmem-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pmem-csi.intel.com/v1beta1",
                "kind": "PmemCSIDeployment",
                "metadata": {
                  "name": "pmem-csi.intel.com"
                },
                "spec": {
                  "deviceMode": "lvm",
                  "nodeSelector": {
                    "feature.node.kubernetes.io/memory-nv.dax": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
          createdAt: "2021-08-26T10:45:23Z"
          description: An operator for deploying and managing the PMEM-CSI driver
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: ""
          repository: https://github.com/intel/pmem-csi/tree/release-1.0
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a PMEM-CSI driver deployment in the cluster.
            displayName: PmemCSIDeployment
            kind: PmemCSIDeployment
            name: pmemcsideployments.pmem-csi.intel.com
            version: v1beta1
        description: |2

          Intel PMEM-CSI is a [CSI](https://github.com/container-storage-interface/spec)
          storage driver for container orchestrators like Kubernetes. It makes local
          persistent memory ([PMEM](https://pmem.io/)) available as a filesystem volume to
          container applications. This is the operator to deploy and manage the
          [PMEM-CSI](https://intel.github.io/pmem-csi/1.0/README.html)
          driver on a Kubernetes cluster.

          Refer to the [PMEM-CSI documentation](https://intel.github.io/pmem-csi/1.0/README.html)
          for more details on deploying and using the PMEM-CSI driver.
        displayName: Operator for PMEM-CSI driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - PMEM-CSI
        - PMEM-CSI operator
        - Persistent Memory
        - PMEM
        links:
        - name: PMEM-CSI
          url: https://intel.github.io/pmem-csi/1.0
        maintainers:
        - email: olev.kartau@intel.com
          name: Olev Kartau
        - email: patrick.ohly@intel.com
          name: Patrick Ohly
        - email: amarnath.valluri@intel.com
          name: Amarnath Valluri
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Intel® Corporation
        relatedImages:
        - registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
        version: 1.0.0
      entries:
      - name: pmem-csi-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pmem-csi-operator-os
    provider:
      name: Intel® Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OceanBase
      provider-url: https://oceanbase.com
    name: oceanbase-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oceanbase-operator.v2.2.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Full Lifecycle
          createdAt: "2024-08-14T16:21:06Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: OBClusterOperation
            name: obclusteroperations.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBCluster
            name: obclusters.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBParameter
            name: obparameters.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBResourceRescue
            name: obresourcerescues.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBServer
            name: observers.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantBackupPolicy
            name: obtenantbackuppolicies.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantBackup
            name: obtenantbackups.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantOperation
            name: obtenantoperations.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantRestore
            name: obtenantrestores.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenant
            name: obtenants.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBZone
            name: obzones.oceanbase.oceanbase.com
            version: v1alpha1
        description: Ob Operator description. TODO.
        displayName: Ob Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - oceanbase-operator
        - ob-operator
        links:
        - name: OceanBase Operator
          url: https://oceanbase.github.io/ob-operator
        maintainers:
        - email: hongxin.sun.vip@gmail.com
          name: Hongxin Sun
        maturity: alpha
        provider:
          name: OceanBase
          url: https://oceanbase.com
        relatedImages:
        - quay.io/oceanbasedev/oceanbase-operator@sha256:b11fe60bb8de545d3ed105acc50f2d6c5d6aa8093f3ae4ca7ebce58b1e5e7b48
        - quay.io/brancz/kube-rbac-proxy@sha256:18f12fc3a0085bbc303cfbd2ad1151127c1c959c4641e36027547b6b66eb4c92
        version: 2.2.2
      entries:
      - name: oceanbase-operator.v2.2.2
        version: 2.2.2
      name: alpha
    - currentCSV: oceanbase-operator.v2.2.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Full Lifecycle
          createdAt: "2024-08-14T16:21:06Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: OBClusterOperation
            name: obclusteroperations.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBCluster
            name: obclusters.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBParameter
            name: obparameters.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBResourceRescue
            name: obresourcerescues.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBServer
            name: observers.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantBackupPolicy
            name: obtenantbackuppolicies.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantBackup
            name: obtenantbackups.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantOperation
            name: obtenantoperations.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantRestore
            name: obtenantrestores.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenant
            name: obtenants.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBZone
            name: obzones.oceanbase.oceanbase.com
            version: v1alpha1
        description: Ob Operator description. TODO.
        displayName: Ob Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - oceanbase-operator
        - ob-operator
        links:
        - name: OceanBase Operator
          url: https://oceanbase.github.io/ob-operator
        maintainers:
        - email: hongxin.sun.vip@gmail.com
          name: Hongxin Sun
        maturity: alpha
        provider:
          name: OceanBase
          url: https://oceanbase.com
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy@sha256:18f12fc3a0085bbc303cfbd2ad1151127c1c959c4641e36027547b6b66eb4c92
        - quay.io/oceanbasedev/oceanbase-operator@sha256:b11fe60bb8de545d3ed105acc50f2d6c5d6aa8093f3ae4ca7ebce58b1e5e7b48
        version: 2.2.2
      entries:
      - name: oceanbase-operator.v2.2.2
        version: 2.2.2
      name: stable
    defaultChannel: stable
    packageName: oceanbase-operator
    provider:
      name: OceanBase
      url: https://oceanbase.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.2.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"customImageNames":
            false,"externalArangoDBName": "arangodb","repository": "registry.connect.redhat.com/turbonomic","tag":"8.3.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/turbonomic/t8c-operator@sha256:d5164e6dd24a0feff10cab62530cafff1bc92d0e8ec2d16c6f4d880c6e2f50bd
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/t8c-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/t8c-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/turbonomic/t8c-operator@sha256:d5164e6dd24a0feff10cab62530cafff1bc92d0e8ec2d16c6f4d880c6e2f50bd
        version: 42.2.0
      entries:
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: stable
    defaultChannel: stable
    packageName: t8c-certified-rhmp
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ptp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ptp-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "NodePtpDevice",
                "metadata": {
                  "name": "node.example.com"
                },
                "spec": {}
              },
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "PtpConfig",
                "metadata": {
                  "name": "example-ptpconfig"
                },
                "spec": {
                  "profile": [
                    {
                      "interface": "ens787f1",
                      "name": "profile1",
                      "phc2sysOpts": "-a -r",
                      "ptp4lOpts": "-s -2",
                      "ptpClockThreshold": {}
                    }
                  ],
                  "recommend": [
                    {
                      "match": [
                        {
                          "nodeLabel": "node-role.kubernetes.io/worker=",
                          "nodeName": "node.example.com"
                        }
                      ],
                      "priority": 4,
                      "profile": "profile1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "PtpOperatorConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "daemonNodeSelector": {
                    "feature.node.kubernetes.io/ptp-capable": "yes",
                    "node-role.kubernetes.io/worker": ""
                  },
                  "ptpEventConfig": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-ptp-rhel9-operator@sha256:012334e6624d5440439824454bda1d1dc79804938174232a3b470176bf462b5e
          createdAt: "2024-03-13"
          description: This software enables configuration of Precision Time Protocol(PTP)
            on Kubernetes. It detects hardware capable PTP devices on each node, and
            configures linuxptp processes such as ptp4l, phc2sys and timemaster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.16.0-202509240123'
          operatorframework.io/suggested-namespace: openshift-ptp
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/ptp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodePtpDevice is the Schema for the nodeptpdevices API
            displayName: Node Ptp Device
            kind: NodePtpDevice
            name: nodeptpdevices.ptp.openshift.io
            version: v1
          - description: PtpConfig is the Schema for the ptpconfigs API
            displayName: Ptp Config
            kind: PtpConfig
            name: ptpconfigs.ptp.openshift.io
            version: v1
          - description: PtpOperatorConfig is the Schema for the ptpoperatorconfigs
              API
            displayName: Ptp Operator Config
            kind: PtpOperatorConfig
            name: ptpoperatorconfigs.ptp.openshift.io
            version: v1
        description: |
          # Precision Time Protocol (PTP) Operator for Openshift

          ## Introdution
          PTP Operator manages cluster wide PTP configuration.
          This operator has to run in namespace 'openshift-ptp'. An Operator Group is also required to install this operator:
          ```
          $ oc create -f - <<EOF
          apiVersion: v1
          kind: Namespace
          metadata:
            name: openshift-ptp
            labels:
              name: openshift-ptp
              openshift.io/cluster-monitoring: "true"
          EOF

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: ptp-operators
            namespace: openshift-ptp
          spec:
            targetNamespaces:
            - openshift-ptp
          EOF
          ```
        displayName: PTP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ptp
        links:
        - name: Documentation operator
          url: https://github.com/openshift/ptp-operator
        - name: Documentation operand
          url: https://github.com/openshift/linuxptp-daemon
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-cloud-event-proxy-rhel9@sha256:3fa2aad8cf026eba586645c851f07a1d752c85b30795fdcf037b138ef3bbb6aa
        - registry.redhat.io/openshift4/ose-ptp-rhel9-operator@sha256:012334e6624d5440439824454bda1d1dc79804938174232a3b470176bf462b5e
        - registry.redhat.io/openshift4/ose-ptp-rhel9@sha256:93b4717337b551603cbb6147d66d7b70a5580332953ddf42304f7092d8739c83
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        version: 4.16.0-202509240123
      entries:
      - name: ptp-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: ptp-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: ptp-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: ptp-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: ptp-operator.v4.16.0-202506201205
        version: 4.16.0-202506201205
      - name: ptp-operator.v4.16.0-202505291140
        version: 4.16.0-202505291140
      - name: ptp-operator.v4.16.0-202505161704
        version: 4.16.0-202505161704
      - name: ptp-operator.v4.16.0-202505020146
        version: 4.16.0-202505020146
      - name: ptp-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: ptp-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: ptp-operator.v4.16.0-202502260004
        version: 4.16.0-202502260004
      - name: ptp-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: ptp-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: ptp-operator.v4.16.0-202501201505
        version: 4.16.0-202501201505
      - name: ptp-operator.v4.16.0-202501072135
        version: 4.16.0-202501072135
      - name: ptp-operator.v4.16.0-202412311307
        version: 4.16.0-202412311307
      - name: ptp-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: ptp-operator.v4.16.0-202412102004
        version: 4.16.0-202412102004
      - name: ptp-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: ptp-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: ptp-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: ptp-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: ptp-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: ptp-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: ptp-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: ptp-operator.v4.16.0-202409131735
        version: 4.16.0-202409131735
      - name: ptp-operator.v4.16.0-202409062036
        version: 4.16.0-202409062036
      - name: ptp-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: ptp-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: ptp-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: ptp-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: ptp-operator.v4.16.0-202407311708
        version: 4.16.0-202407311708
      - name: ptp-operator.v4.16.0-202407301837
        version: 4.16.0-202407301837
      - name: ptp-operator.v4.16.0-202407160236
        version: 4.16.0-202407160236
      - name: ptp-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: ptp-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: ptp-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: ptp-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: ptp-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: ptp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: IBM
      provider-url: ""
    name: ibm-block-csi-operator-community
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12
    - currentCSV: ibm-block-csi-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
        version: 1.12.0
      entries:
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.0
    - currentCSV: ibm-block-csi-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
        version: 1.12.1
      entries:
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.1
    - currentCSV: ibm-block-csi-operator.v1.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:3fc8df547138a2009225d63c2a3111a672340bdf9f3cb24a889cbb69f2c33c2d
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:3fc8df547138a2009225d63c2a3111a672340bdf9f3cb24a889cbb69f2c33c2d
        version: 1.12.2
      entries:
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.2
    - currentCSV: ibm-block-csi-operator.v1.12.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:2bd461a1e40a1366c871585aa7a4e8039e686a66bf7192efdeb1c215209a071d
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:2bd461a1e40a1366c871585aa7a4e8039e686a66bf7192efdeb1c215209a071d
        version: 1.12.3
      entries:
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.3
    - currentCSV: ibm-block-csi-operator.v1.12.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e29db476de3778222bf049926223dee20f6bc34e55a04df5f0f3df38741ff23e
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e29db476de3778222bf049926223dee20f6bc34e55a04df5f0f3df38741ff23e
        version: 1.12.4
      entries:
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.4
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.5
    defaultChannel: stable
    packageName: ibm-block-csi-operator-community
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Citrus Framework
      provider-url: ""
    name: yaks
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: yaks-operator.v0.17.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "yaks.citrusframework.org/v1alpha1",
                "kind": "Instance",
                "metadata": {
                  "labels": {
                    "app": "yaks"
                  },
                  "name": "example-instance"
                },
                "spec": {
                  "operator": {
                    "global": true
                  }
                }
              },
              {
                "apiVersion": "yaks.citrusframework.org/v1alpha1",
                "kind": "Test",
                "metadata": {
                  "labels": {
                    "app": "yaks"
                  },
                  "name": "example-test"
                },
                "spec": {
                  "source": {
                    "content": "Feature: hello world\n\n  Scenario: print slogan\n    Given YAKS does Cloud-Native BDD testing\n    Then YAKS rocks!",
                    "language": "feature",
                    "name": "example.feature"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/citrusframework/yaks:0.17.1
          createdAt: "2023-04-26T20:37:31Z"
          description: YAKS is a platform to enable Cloud Native BDD testing on Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["instances.yaks.citrusframework.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/citrusframework/yaks
          support: Citrus Framework
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance is the Schema for the yaks instance
            displayName: Instance
            kind: Instance
            name: instances.yaks.citrusframework.org
            version: v1alpha1
          - description: Test is the Schema for the tests API
            displayName: Test
            kind: Test
            name: tests.yaks.citrusframework.org
            version: v1alpha1
        description: |
          YAKS
          ====

          YAKS is a platform to enable Cloud Native BDD testing on Kubernetes.

          ## Running a YAKS test

          With the YAKS operator installed, you can run tests by creating a `Test` custom resource on the cluster.

          Tests are defined using [Gherkin](https://cucumber.io/docs/gherkin/) syntax. YAKS provides a set of predefined steps which
          help to connect with different messaging transports (Http REST, JMS, Kafka, Knative eventing) and verify responses with
          assertions on message header and body content.

          The example below defines a simple test resource.

          ```
          apiVersion: yaks.citrusframework.org/v1alpha1
          kind: Test
          metadata:
            name: example
          spec:
            source:
              name: example.feature
              content: |-
                Feature: hello world

                  Scenario: print slogan
                    Given YAKS does Cloud-Native BDD testing
                    Then YAKS rocks!
          ```

          Refer to the [YAKS repository](https://github.com/citrusframework/yaks) for more complex examples.
        displayName: YAKS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - yaks
        - testing
        - microservices
        - knative
        - kafka
        - serverless
        - camel
        - camel-k
        links:
        - name: Project page
          url: https://citrusframework.org/
        - name: YAKS source code repository
          url: https://github.com/citrusframework/yaks
        maintainers:
        - email: cdeppisch@redhat.com
          name: Christoph Deppisch
        - email: nferraro@redhat.com
          name: Nicola Ferraro
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Citrus Framework
        relatedImages:
        - docker.io/citrusframework/yaks:0.17.1
        version: 0.17.1
      entries:
      - name: yaks-operator.v0.17.1
        version: 0.17.1
      - name: yaks-operator.v0.17.0
        version: 0.17.0
      - name: yaks-operator.v0.16.0
        version: 0.16.0
      - name: yaks-operator.v0.15.1
        version: 0.15.1
      - name: yaks-operator.v0.15.0
        version: 0.15.0
      - name: yaks-operator.v0.14.3
        version: 0.14.3
      - name: yaks-operator.v0.14.2
        version: 0.14.2
      - name: yaks-operator.v0.14.1
        version: 0.14.1
      - name: yaks-operator.v0.13.0
        version: 0.13.0
      - name: yaks-operator.v0.12.2
        version: 0.12.2
      - name: yaks-operator.v0.11.0
        version: 0.11.0
      - name: yaks-operator.v0.10.0
        version: 0.10.0
      - name: yaks-operator.v0.9.0
        version: 0.9.0
      - name: yaks-operator.v0.8.0
        version: 0.8.0
      - name: yaks-operator.v0.7.0
        version: 0.7.0
      - name: yaks-operator.v0.6.0
        version: 0.6.0
      - name: yaks-operator.v0.5.0
        version: 0.5.0
      - name: yaks-operator.v0.4.0
        version: 0.4.0
      name: alpha
    defaultChannel: alpha
    packageName: yaks
    provider:
      name: Citrus Framework
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator.v1.5.0
        version: 1.5.0
      - name: dynatrace-operator.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetScaler
      provider-url: https://netscaler.com
    name: netscaler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netscaler-operator.v3.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCloudController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalercloudcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "azure": {
                    "clientsecret": "",
                    "image": "quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb",
                    "location": "",
                    "pullPolicy": "IfNotPresent",
                    "resourcegroupname": "",
                    "resources": {},
                    "subnetname": "",
                    "subscriptionid": "",
                    "tenantid": "",
                    "vnetname": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCpxWithIngressController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalercpxwithingresscontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "admSettings": {
                    "admIP": "",
                    "bandWidth": 1000,
                    "bandWidthLicense": false,
                    "cpxCores": 1,
                    "licenseEdition": "PLATINUM",
                    "licenseServerIP": "",
                    "licenseServerPort": 27000,
                    "loginSecret": "",
                    "platform": false,
                    "vCPULicense": false
                  },
                  "affinity": {},
                  "analyticsConfig": {
                    "distributedTracing": {
                      "enable": false,
                      "samplingRate": 100
                    },
                    "endpoint": {
                      "metrics": {
                        "service": ""
                      },
                      "transactions": {
                        "service": ""
                      }
                    },
                    "required": false,
                    "timeseries": {
                      "auditLogs": {
                        "enable": false
                      },
                      "events": {
                        "enable": false
                      },
                      "metrics": {
                        "enable": false,
                        "enableNativeScrape": false,
                        "exportFrequency": 30,
                        "mode": "avro",
                        "schemaFile": "schema.json"
                      },
                      "port": 5563
                    },
                    "transactions": {
                      "enable": false,
                      "port": 5557
                    }
                  },
                  "exporter": {
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4",
                    "ports": {
                      "containerPort": 8888
                    },
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {}
                  },
                  "extraVolumes": [],
                  "imagePullSecrets": [],
                  "ingressController": {
                    "defaultSSLCertSecret": "",
                    "defaultSSLSNICertSecret": "",
                    "disableAPIServerCertVerify": false,
                    "disableOpenshiftRoutes": true,
                    "enableLivenessProbe": true,
                    "enableReadinessProbe": true,
                    "entityPrefix": "k8s",
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "ingressClass": [],
                    "ipam": false,
                    "jsonLog": false,
                    "kubernetesURL": "",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "logLevel": "INFO",
                    "logProxy": "",
                    "namespaceLabels": "",
                    "openshift": true,
                    "optimizeEndpointBinding": false,
                    "profileHttpFrontend": {},
                    "profileSslFrontend": {},
                    "profileTcpFrontend": {},
                    "pullPolicy": "IfNotPresent",
                    "rbacRole": false,
                    "readinessProbe": {
                      "failureThreshold": 3,
                      "initialDelaySeconds": 50,
                      "periodSeconds": 60,
                      "successThreshold": 1
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "128m",
                        "memory": "500Mi"
                      }
                    },
                    "routeLabels": "",
                    "setAsDefaultIngressClass": false,
                    "updateIngressStatus": true
                  },
                  "netscalerCpx": {
                    "bgpSettings": {
                      "bgpConfig": [
                        {
                          "bgpRouter": {
                            "localAS": 100,
                            "neighbor": [
                              {
                                "ASOriginationInterval": 10,
                                "address": "",
                                "advertisementInterval": 10,
                                "remoteAS": 100
                              }
                            ]
                          }
                        }
                      ],
                      "bgpPort": 179
                    },
                    "cpxBgpRouter": false,
                    "cpxCommands": [],
                    "cpxLicenseAggregator": "",
                    "cpxShellCommands": [],
                    "daemonSet": false,
                    "enableLivenessProbe": true,
                    "enableStartupProbe": true,
                    "extraVolumeMounts": [],
                    "hostname": "",
                    "image": "quay.io/netscaler/netscaler-cpx@sha256:232596eaacda68c0e2e210f26b64cedc8d70ae12c5e8ae2e1a66896a479ae4f8",
                    "ingressIP": "",
                    "livenessProbe": {
                      "failureThreshold": 3,
                      "periodSeconds": 5,
                      "successThreshold": 1
                    },
                    "mgmtHttpPort": 9080,
                    "mgmtHttpsPort": 9443,
                    "nitroReadTimeout": 20,
                    "nsConfigDnsRec": false,
                    "nsCookieVersion": 0,
                    "nsDnsNameserver": "",
                    "nsEnableLabels": true,
                    "nsGateway": "192.168.1.1",
                    "nsHttp2ServerSide": "OFF",
                    "nsIP": "192.168.1.2",
                    "nsLbHashAlgo": {
                      "hashAlgorithm": "default",
                      "hashFingers": 500,
                      "required": false
                    },
                    "nsProtocol": "HTTP",
                    "nsSvcLbDnsRec": false,
                    "prometheusCredentialSecret": "",
                    "pullPolicy": "IfNotPresent",
                    "service": {
                      "annotations": {},
                      "spec": {}
                    },
                    "startupProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1
                    }
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "replicaCount": 1,
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerGslbController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalergslbcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "gslbController": {
                    "disableAPIServerCertVerify": false,
                    "entityPrefix": "",
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "kubernetesURL": "",
                    "localCluster": "",
                    "localRegion": "",
                    "logLevel": "Debug",
                    "openshift": true,
                    "pullPolicy": "IfNotPresent",
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "32m",
                        "memory": "128Mi"
                      }
                    },
                    "siteData": [
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteMask": "",
                        "siteName": "",
                        "sitePublicIp": "",
                        "siteRegion": ""
                      },
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteMask": "",
                        "siteName": "",
                        "sitePublicIp": "",
                        "siteRegion": ""
                      }
                    ]
                  },
                  "imagePullSecrets": [],
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsIP": "",
                    "nsPort": 443,
                    "nsProtocol": "HTTPS"
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerIngressController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscaleringresscontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "analyticsConfig": {
                    "distributedTracing": {
                      "enable": false,
                      "samplingRate": 100
                    },
                    "endpoint": {
                      "metrics": {
                        "service": ""
                      },
                      "transactions": {
                        "service": ""
                      }
                    },
                    "required": false,
                    "timeseries": {
                      "auditLogs": {
                        "enable": false
                      },
                      "events": {
                        "enable": false
                      },
                      "metrics": {
                        "enable": false,
                        "enableNativeScrape": false,
                        "exportFrequency": 50,
                        "mode": "avro",
                        "schemaFile": "schema.json"
                      },
                      "port": 50002
                    },
                    "transactions": {
                      "enable": false,
                      "port": 50005
                    }
                  },
                  "exporter": {
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4",
                    "ports": {
                      "containerPort": 8888
                    },
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {},
                    "serviceMonitorExtraLabels": {}
                  },
                  "extraVolumes": [],
                  "hostAliases": [],
                  "imagePullSecrets": [],
                  "ingressController": {
                    "clusterName": "",
                    "defaultSSLCertSecret": "",
                    "defaultSSLSNICertSecret": "",
                    "disableAPIServerCertVerify": false,
                    "disableOpenshiftRoutes": true,
                    "enableLivenessProbe": true,
                    "enableReadinessProbe": true,
                    "entityPrefix": "",
                    "extraVolumeMounts": [],
                    "ignoreNodeExternalIP": false,
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "ingressClass": [],
                    "ipam": false,
                    "jsonLog": false,
                    "kubernetesURL": "",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "logLevel": "INFO",
                    "logProxy": "",
                    "multiClusterPrefix": "",
                    "namespaceLabels": "",
                    "nodeLabels": "",
                    "nodeWatch": false,
                    "nsValidateCert": {},
                    "nsncPbr": false,
                    "openshift": true,
                    "optimizeEndpointBinding": false,
                    "podIPsforServiceGroupMembers": false,
                    "profileHttpFrontend": {},
                    "profileSslFrontend": {},
                    "profileTcpFrontend": {},
                    "pullPolicy": "IfNotPresent",
                    "rbacRole": false,
                    "readinessProbe": {
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 60,
                      "successThreshold": 1
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "32m",
                        "memory": "128Mi"
                      }
                    },
                    "routeLabels": "",
                    "serviceClass": [],
                    "setAsDefaultIngressClass": false,
                    "updateIngressStatus": true
                  },
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsConfigDnsRec": false,
                    "nsCookieVersion": 0,
                    "nsDnsNameserver": "",
                    "nsHttp2ServerSide": "OFF",
                    "nsIP": "",
                    "nsLbHashAlgo": {
                      "hashAlgorithm": "default",
                      "hashFingers": 256,
                      "required": false
                    },
                    "nsPort": 443,
                    "nsProtocol": "HTTPS",
                    "nsSNIPs": [],
                    "nsSvcLbDnsRec": false,
                    "nsVIP": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerIpamController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscaleripamcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "clusterName": "",
                  "image": "quay.io/netscaler/netscaler-ipam-controller@sha256:7c25b59a2e1df0e4f5615a779a9bea939c60a3eca065cb6c2295daf0d03831d8",
                  "localIPAM": {
                    "vipRange": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "reuseIngressVIP": true,
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerObservabilityExporter",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalerobservabilityexporter-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "elasticsearch": {
                    "enabled": false,
                    "server": "http://elasticsearch:9200"
                  },
                  "headless": false,
                  "image": "quay.io/netscaler/netscaler-observability-exporter@sha256:626e3a954c67bbb910477352dc740aaab21ff1c1cfb03354becf50beab17d4a4",
                  "jsonTransRateLimiting": {
                    "enabled": false,
                    "limit": 100,
                    "queueLimit": 1000,
                    "window": 5
                  },
                  "kafka": {
                    "auditLogs": "no",
                    "broker": "",
                    "dataFormat": "AVRO",
                    "enabled": false,
                    "events": "no",
                    "topic": "HTTP"
                  },
                  "nodePortRequired": false,
                  "nodeSelector": {},
                  "nsoeLogLevel": "INFO",
                  "nstracing": {
                    "enabled": false,
                    "server": "zipkin:9411/api/v1/spans"
                  },
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "replicas": 1,
                  "resources": {},
                  "splunk": {
                    "authToken": "",
                    "enabled": false,
                    "indexPrefix": "adc_nsoe",
                    "server": "http://splunk:8088"
                  },
                  "timeseries": {
                    "enabled": false
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-operator@sha256:e367c8c8ba6171e9ef5dd496592a5c62f4aa764f93d0cb78835aa29024c063d0
          createdAt: "2025-04-15T08:51:39Z"
          description: This is an operator to install Cloud Native portfolio provided
            by NetScaler
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler Cloud Controller for NetScaler VPX/MPX/SDX
              with Azure Redhat OpenShift in Azure.
            displayName: Netscaler Cloud Controller
            kind: NetscalerCloudController
            name: netscalercloudcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler CPX with Ingress Controller.
            displayName: Netscaler CPX With Ingress Controller
            kind: NetscalerCpxWithIngressController
            name: netscalercpxwithingresscontrollers.netscaler.com
            version: v1
          - description: To install NetScaler GSLB Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler GSLB Controller
            kind: NetscalerGslbController
            name: netscalergslbcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler Ingress Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler Ingress Controller
            kind: NetscalerIngressController
            name: netscaleringresscontrollers.netscaler.com
            version: v1
          - description: To install NetScaler IPAM Controller for NetScaler CPX/VPX/MPX/SDX.
            displayName: Netscaler IPAM Controller
            kind: NetscalerIpamController
            name: netscaleripamcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler Observability Exporter for NetScaler.
            displayName: Netscaler Observability Exporter
            kind: NetscalerObservabilityExporter
            name: netscalerobservabilityexporters.netscaler.com
            version: v1
        description: |
          NetScaler provides various products to empower Developer/DevOps managing your applications in OpenShift Cluster. Using this operator you can deploy below controllers for NetScaler appliance:
          1. [NetScaler Cloud Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. [NetScaler Ingress Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/unified-ingress-deployment)
          3. [NetScaler CPX with Ingress Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/two-tier-deployment)
          4. [NetScaler GSLB Controller ](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/gslb/gslb)
          5. [NetScaler Observability Exporter](https://docs.netscaler.com/en-us/netscaler-observability-exporter/deploy-coe)
          6. [NetScaler IPAM Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/ipam-installation.html)

          ### Installation
          Refer [installation instructions](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-operator-openshift).

          This operator version contains:
          1. NetScaler Ingress Controller version [2.2.10](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.2.10)
          2. NetScaler CPX version [14.1-25.111](https://docs.netscaler.com/en-us/updates?product=NetScaler-ADC-%28includes-NetScaler-Gateway%29&version=14.1&build=25.56)
          3. NetScaler GSLB Controller version [2.2.10](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.2.10)
          4. NetScaler Observability Exporter version [1.10.001](https://github.com/citrix/citrix-helm-charts/releases/tag/nsoe-v1.10.001)
          5. NetScaler Metrics Exporter [1.4.9](https://github.com/citrix/citrix-adc-metrics-exporter/releases/tag/1.4.9)
          6. NetScaler IPAM Controller [2.1.2](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.3.15)

          ### Support

          For debugging issues, please use [NetScaler Kubernetes Plugin](https://github.com/netscaler/modern-apps-toolkit/tree/main/netscaler-plugin).

          For questions and support, the following channels are available:
          - [NetScaler community portal](https://community.netscaler.com/s/)
          - [Contact us](https://www.netscaler.com/contact)
        displayName: NetScaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        - ADC
        - Kubernetes
        - OpenShift
        - Citrix
        - Controller
        - Observability
        - Metrices
        links:
        - name: Netscaler Operator
          url: https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-operator-openshift
        maintainers:
        - email: priyanka.sharma@cloud.com
          name: Priyanka Sharma
        - email: arijit.ray@cloud.com
          name: Arijit Ray
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: NetScaler
          url: https://netscaler.com
        relatedImages:
        - registry.connect.redhat.com/citrix/netscaler-operator@sha256:e367c8c8ba6171e9ef5dd496592a5c62f4aa764f93d0cb78835aa29024c063d0
        - quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f
        - quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb
        - quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4
        - quay.io/netscaler/netscaler-cpx@sha256:232596eaacda68c0e2e210f26b64cedc8d70ae12c5e8ae2e1a66896a479ae4f8
        - quay.io/netscaler/netscaler-observability-exporter@sha256:626e3a954c67bbb910477352dc740aaab21ff1c1cfb03354becf50beab17d4a4
        - quay.io/netscaler/netscaler-ipam-controller@sha256:7c25b59a2e1df0e4f5615a779a9bea939c60a3eca065cb6c2295daf0d03831d8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:89f44c45373da73a51b0b9a02de7efe519dc2d5343b65fe382526cccc92f8f5c
        version: 3.2.0
      entries:
      - name: netscaler-operator.v3.2.0
        version: 3.2.0
      - name: netscaler-operator.v3.1.2
        version: 3.1.2
      - name: netscaler-operator.v3.1.1
        version: 3.1.1
      - name: netscaler-operator.v3.0.1
        version: 3.0.1
      name: stable
    - currentCSV: netscaler-operator.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCloudController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netscaler-operator",
                    "app.kubernetes.io/instance": "nscc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscalercloudcontroller",
                    "app.kubernetes.io/part-of": "netscaler-operator"
                  },
                  "name": "nscc"
                },
                "spec": {
                  "azure": {
                    "clientsecret": "",
                    "image": "quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb",
                    "location": "",
                    "pullPolicy": "IfNotPresent",
                    "resourcegroupname": "",
                    "resources": {},
                    "subnetname": "",
                    "subscriptionid": "",
                    "tenantid": "",
                    "vnetname": ""
                  }
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerGslbController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netscaler-operator",
                    "app.kubernetes.io/instance": "nsgc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscalergslbcontroller",
                    "app.kubernetes.io/part-of": "netscaler-operator"
                  },
                  "name": "nsgc"
                },
                "spec": {
                  "gslbController": {
                    "acceptLicense": "Yes",
                    "cleanupGSLBSiteConfig": false,
                    "disableAPIServerCertVerify": false,
                    "entityPrefix": "",
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:88114c518ecceb54b000f38bfc330efe184bfeb5b5df0ee308950cf29e7a777f",
                    "imagePullSecrets": [],
                    "kubernetesURL": "",
                    "localCluster": "",
                    "localRegion": "",
                    "logLevel": "Debug",
                    "openshift": true,
                    "pullPolicy": "IfNotPresent",
                    "siteData": [
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteName": "",
                        "siteRegion": ""
                      },
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteName": "",
                        "siteRegion": ""
                      }
                    ]
                  },
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsIP": "",
                    "nsPort": 443,
                    "nsProtocol": "HTTPS"
                  },
                  "resources": {
                    "limits": {},
                    "requests": {
                      "cpu": "32m",
                      "memory": "128Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-operator@sha256:3f71f9a46d7b29e3da98a301965f635c03a022b5535b586d7d5c699341f68193
          createdAt: "2024-07-31T06:33:49Z"
          description: This is an operator to install Cloud Native portfolio provided
            by NetScaler
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler Cloud Controller for NetScaler VPX/MPX/SDX
              with Azure Redhat OpenShift in Azure.
            displayName: Netscaler Cloud Controller
            kind: NetscalerCloudController
            name: netscalercloudcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler GSLB Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler GSLB Controller
            kind: NetscalerGslbController
            name: netscalergslbcontrollers.netscaler.com
            version: v1
        description: |-
          NetScaler provides various products to empower Developer/DevOps managing your applications in OpenShift Cluster. Using this operator you can deploy below controllers for NetScaler appliance:
          1. [NetScaler Cloud Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. [NetScaler GSLB Controller ](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/gslb-openshift-operator)

          ### Installation
          Refer [installation instructions](https://github.com/netscaler/netscaler-k8s-ingress-controller/blob/master/docs/deploy/deploy-ns-operator.md).

          This operator version contains:
          1. NetScaler Cloud Controller version [1.2.0](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. NetScaler GSLB Controller version [1.43.7](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v1.43.7)

          ### Support

          For debugging issues, please use [NetScaler Kubernetes Plugin](https://github.com/netscaler/modern-apps-toolkit/tree/main/netscaler-plugin).

          For questions and support, the following channels are available:
          - [NetScaler community portal](https://community.netscaler.com/s/)
          - [Contact us](https://www.netscaler.com/contact)
        displayName: NetScaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        links:
        - name: Netscaler Operator
          url: https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/gslb-openshift-operator
        maintainers:
        - email: priyanka.sharma@cloud.com
          name: Priyanka Sharma
        - email: arijit.ray@cloud.com
          name: Arijit Ray
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: NetScaler
          url: https://netscaler.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:23405618050a4ceea8ea0c748cc7d9d16dbb80f10e4c39f7e82014a3ac80cec9
        - registry.connect.redhat.com/citrix/netscaler-operator@sha256:3f71f9a46d7b29e3da98a301965f635c03a022b5535b586d7d5c699341f68193
        - quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:88114c518ecceb54b000f38bfc330efe184bfeb5b5df0ee308950cf29e7a777f
        - quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb
        version: 3.0.1
      entries:
      - name: netscaler-operator.v3.0.1
        version: 3.0.1
      name: tech-preview
    defaultChannel: stable
    packageName: netscaler-operator
    provider:
      name: NetScaler
      url: https://netscaler.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: open-liberty-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: open-liberty-operator.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4",
                  "expose": true,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/open-liberty-operator-controller@sha256:18f188e48725c785557c8edea46f69c464d6f81ab47860a3bb0ec978c20a26db
          createdAt: "2022-03-25T14:00:00Z"
          description: Deploy and manage applications running on Liberty
          olm.skipRange: <0.8.1
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          This advanced Operator can be used to deploy and manage Open Liberty applications with consistent, production-grade QoS. This operator is based on the [Runtime Component Operator](https://github.com/application-stacks/runtime-component-operator) and provides all of its capabilities in addition to Open Liberty specific features such as gathering traces and dumps (Day-2 operations) and easily configuring and managing the single sign-on information for your Open Liberty applications.

          Open Liberty Operator enables enterprise architects to govern the way their applications get deployed & managed in the cluster, while dramatically reducing the learning curve for developers to deploy into Kubernetes - allowing them to focus on writing the code!

          Here are some key features:

          #### Application Lifecyle
          You can deploy your Open Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the Operator will watch for any updates and will re-deploy the modified image.

          #### Custom RBAC
          This Operator is capable of using a custom ServiceAccount from the caller, allowing it to follow RBAC restrictions. By default it creates a ServiceAccount if one is not specified, which can also be bound with specific roles.

          #### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as: labels, annotations, and environment variables from a ConfigMap, a Secret or a value.

          #### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configuration, such as TLS settings, are also easily enabled.

          #### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto scaling to create (and delete) instances based on resource consumption.

          #### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you.
          This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart.
          We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          You can also easily configure and use a single storage for serviceability related Day-2 operations, such as gatherig server traces and dumps.

          #### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating kubernetes Secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to its required services without any intervention or interruption.

          #### Single Sign-On (SSO)
          Open Liberty provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows to easily configure and manage the single sign-on information for your applications.

          #### Exposing metrics to Prometheus
          Expose the Open Liberty application's metrics via the Prometheus Operator.
          You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `spec.monitoring` field to control configurations such as poll interval and security credentials.

          #### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply add the following configuration (to specify the volume size and the location to persist) to your OpenLibertyApplication CR:
          ``` storage: size: 2Gi mountPath: "/logs" ```

          #### Integration with OpenShift Serverless
          Deploy your serverless runtime component using a single toggle.  The Operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing your pod to automatically scale to 0 when it is idle.

          #### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and how they are connected.

          See our [**documentation**](https://github.com/OpenLiberty/open-liberty-operator/tree/main/doc/) for more information.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://github.com/OpenLiberty/open-liberty-operator/tree/main/doc
        - name: Open Liberty
          url: https://openliberty.io/
        - name: Open Liberty Container Images
          url: https://github.com/OpenLiberty/ci.docker
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: beta
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/open-liberty-operator-controller@sha256:18f188e48725c785557c8edea46f69c464d6f81ab47860a3bb0ec978c20a26db
        - registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4
        version: 0.8.1
      entries:
      - name: open-liberty-operator.v0.8.1
        version: 0.8.1
      - name: open-liberty-operator.v0.8.0
        version: 0.8.0
      name: beta2
    - currentCSV: open-liberty-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:d45c296d43dc4c4100f38bb736f073e04c8b4d72b68ddf1189e311aff8ad66e3
          createdAt: "2023-08-28T16:39:10Z"
          description: Deploy and manage containerized Liberty applications
          olm.skipRange: '>=0.8.0 <1.2.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:d45c296d43dc4c4100f38bb736f073e04c8b4d72b68ddf1189e311aff8ad66e3
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02
        version: 1.2.2
      entries:
      - name: open-liberty-operator.v1.2.2
        version: 1.2.2
      name: v1.2
    - currentCSV: open-liberty-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:d68bb66a8133702a198f7be655d7cff5f500bb620c460a81547971e00124ef73
          createdAt: "2024-08-01T20:19:13Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.3.3'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:d68bb66a8133702a198f7be655d7cff5f500bb620c460a81547971e00124ef73
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91
        version: 1.3.3
      entries:
      - name: open-liberty-operator.v1.3.3
        version: 1.3.3
      name: v1.3
    - currentCSV: open-liberty-operator.v1.4.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:67b5b46dfed008dab4ea7fd176d3bc31dae8940d72194a10aebb98b1dea9043e
          createdAt: "2025-07-18T20:45:06Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.4.4'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:67b5b46dfed008dab4ea7fd176d3bc31dae8940d72194a10aebb98b1dea9043e
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3
        version: 1.4.4
      entries:
      - name: open-liberty-operator.v1.4.4
        version: 1.4.4
      name: v1.4
    - currentCSV: open-liberty-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyPerformanceData",
                "metadata": {
                  "name": "openliberty-performance-data-sample"
                },
                "spec": {
                  "interval": 30,
                  "podName": "Specify_Pod_Name_Here",
                  "timespan": 240
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:c3b031fe14054f65d6612cc505b7681336873dc32ce1091451d8c7799034bc83
          createdAt: "2025-09-16T03:08:31Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.5.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server performance data
            displayName: OpenLibertyPerformanceData
            kind: OpenLibertyPerformanceData
            name: openlibertyperformancedata.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:c3b031fe14054f65d6612cc505b7681336873dc32ce1091451d8c7799034bc83
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8
        version: 1.5.0
      entries:
      - name: open-liberty-operator.v1.5.0
        version: 1.5.0
      name: v1.5
    defaultChannel: v1.5
    packageName: open-liberty-certified
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: RabbitMQ
      provider-url: https://rabbitmq.com
    name: rabbitmq-cluster-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-cluster-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rabbitmq.com/v1beta1",
                "kind": "RabbitmqCluster",
                "metadata": {
                  "name": "sample",
                  "namespace": "rabbitmq-system"
                },
                "spec": {
                  "replicas": 1,
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-07-10T12:52:23Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RabbitmqCluster is the Schema for the RabbitmqCluster API.
              Each instance of this object corresponds to a single RabbitMQ cluster.
            displayName: Rabbitmq Cluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: RabbitMQ Cluster Operator
        displayName: RabbitMQ Cluster Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        links:
        - name: Cluster Operator
          url: https://github.com/rabbitmq/cluster-operator
        maturity: alpha
        provider:
          name: RabbitMQ
          url: https://rabbitmq.com
        relatedImages:
        - registry.redhat.io/rhoso-operators/rabbitmq-cluster-rhel9-operator@sha256:e4b198236d70b7dfdd29d8ed45058dec3335b3530e0b676fcc862bfec2efce79
        version: 1.0.6
      entries:
      - name: rabbitmq-cluster-operator.v1.0.6
        version: 1.0.6
      - name: rabbitmq-cluster-operator.v1.0.5
        version: 1.0.5
      - name: rabbitmq-cluster-operator.v1.0.4
        version: 1.0.4
      - name: rabbitmq-cluster-operator.v1.0.3
        version: 1.0.3
      - name: rabbitmq-cluster-operator.v1.0.2
        version: 1.0.2
      - name: rabbitmq-cluster-operator.v1.0.1
        version: 1.0.1
      - name: rabbitmq-cluster-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: rabbitmq-cluster-operator
    provider:
      name: RabbitMQ
      url: https://rabbitmq.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fossul Framework
      provider-url: https://github.com/fossul/fossul
    name: fossul-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fossul-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n    \"kind\":
            \"Backup\",\n    \"metadata\": {\n      \"name\": \"mariadb-sample\"\n
            \   },\n    \"spec\": {\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\"\n    }\n  },\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n    \"kind\":
            \"Restore\",\n    \"metadata\": {\n      \"name\": \"mariadb-sample\"\n
            \   },\n    \"spec\": {\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\",\n      \"workflow_id\": \"xxxx\"\n    }\n  },        \n  {\n
            \   \"apiVersion\": \"fossul.io/v1\",\n    \"kind\": \"BackupConfig\",\n
            \   \"metadata\": {\n      \"name\": \"mariadb-sample\"\n    },\n    \"spec\":
            {\n      \"container_platform\": \"openshift\",\n      \"operator_controlled\":
            true,\n      \"auto_discovery\": false,\n      \"deployment_name\": \"mariadb\",\n
            \     \"deployment_type\": \"DeploymentConfig\",\n      \"job_retention\":
            50,\n      \"overwrite_pcv_on_restore\": true,\n      \"policies\": [\n
            \       {\n          \"policy\": \"hourly\",\n          \"retentionNumber\":
            3\n        },\n        {\n          \"policy\": \"daily\",\n          \"retentionNumber\":
            10\n        }\n      ],\n      \"pvc_deletion_timeout\": 300,\n      \"restore_to_new_pvc\":
            false,\n      \"snapshot_timeout\": 180,\n      \"storage_plugin\": \"csi.so\",\n
            \     \"app_plugin\": \"mariadb.so\"\n    }\n  },\n  {\n    \"apiVersion\":
            \"fossul.io/v1\",\n    \"kind\": \"BackupSchedule\",\n    \"metadata\":
            {\n      \"name\": \"mariadb-sample\"\n    },\n    \"spec\": {\n      \"cron_schedule\":
            \"59 23 * * *\",\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\"\n\n    }\n  },\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n
            \   \"kind\": \"Fossul\",\n    \"metadata\": {\n      \"name\": \"fossul-sample\"\n
            \   },\n    \"spec\": {\n      \"enable_routes\": false,\n      \"enable_virtualmachines\":
            false,\n      \"container_platform\": \"openshift\"\n    }\n  }\n]"
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/fossul/fossul-operator:v0.6.0
          createdAt: "2022-01-25T20:09:00Z"
          description: Fossul is a application centric container native backup and
            restore framework for databases or applications
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/fossul/fossul
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages application backup configurations
            kind: BackupConfig
            name: backupconfigs.fossul.io
            version: v1
          - description: Manages application backups
            kind: Backup
            name: backups.fossul.io
            version: v1
          - description: Manages application restores
            kind: Restore
            name: restores.fossul.io
            version: v1
          - description: Manages application backup scheduling
            kind: BackupSchedule
            name: backupschedules.fossul.io
            version: v1
          - description: Manages deployment of the Fossul Framework
            kind: Fossul
            name: fossuls.fossul.io
            version: v1
        description: "## Fossul Framework\nApplication backup and restore framework
          for container native applications on Kubernetes and OpenShift. Fossul provides
          application centric backup and recovery enabling integration with the application
          creating an application consistent backup. Since there are many applications
          and providers for storage or archive Fossul enables plugins that allow proper
          integration. Fossul provides plugins for postgresql, mariadb, mysql and
          mongo as well as plugin to integrate with CSI snapshots. Plugins can be
          written in any language. The Fossul operator deploys Fossul via the Fossul
          custom resource and additionally provides BackupConfig, Backup, BackupSchedule
          and Restore custom resources. Each user can configure and manage their own
          backups with simple custom resources that can be part of CI/CD or source
          control. Users can also opt to manage backups through Fossul CLI/API instead
          of the custom resources provided by the operator.\n\n## Installation\nIt
          is recommended to install the operator into the openshift-operators namespace
          which is the default location for operators. Once the operator is installed
          create a new project called ```fossul``` and create the Fossul custom resource
          under the fossul namespace. This will deploy the Fossul Framework. You can
          choose another namespace for the Fossul Framework but then also need to
          provide the optional 'fossul_namespace' parameter with the spec for all
          the additional custom resources. This is required so the Fossul operator
          can communicate with Fossul.\n\nOnce Fossul is deployed you can manage backups
          either through custom resources provided by the operator or the Fossul CLI/API.
          If using the customer resources, first create a BackupConfig custom resource
          for every application you want to backup. Once that is done, you can optionally
          create a backup by simply creating a Backup custom resource. You can also
          create a BackupSchedule custom resource using cron syntax which will schedule
          your backup to occur on a regular interval or do so via a Kubernetes job.
          Finally if a backup has been created you can perform a restore through a
          custom resource as well, providing the workflow id from the completed backup.
          If you are creating backups through the custom resource the workflow id
          will be appended to the spec once the backup completes. If you are using
          the Fossul CLI/API to create backups you will need to get the workflow id
          through the Fossul CLI/API directly using the backupList action.\n\nFor
          users of the cli, a cli pod is deployed with credentials and if desired
          you can create or manage backups from within that pod. To use the API you
          need the credentials stored in the fossul secret. The cli has already been
          pre-configured with these credentials. The API can also be optionally exposed
          via routes and API documentation is under ```/api/v1/index.html```.\n\n##
          Quick Setup\nWhen creating Fossul custom resources generally you will only
          need to update the ```deployment_name``` and ```deployment_type``` (DeploymentConfig,
          Deployment or VirtualMachine). All custom resources must be created in the
          namespace where the database or application exists. \n\n### Deploy Fossul
          in fossul namespace\n```$ oc new-project fossul```\n<pre>\n$ vi fossul.yaml\nkind:
          Fossul\napiVersion: fossul.io/v1\nmetadata:\n  name: fossul-sample\n  namespace:
          fossul\nspec: {}\n</pre>\n```$ oc create -f fossul.yaml```\n\n### Deploy
          Mariadb Database\nMake sure you create all Fossul custom resources in the
          same namespace as the database, application or virtual machine\n\n### Create
          MariaDB Fossul BackupConfig\nCreates a backup configuration for MariaDB
          or MySQL databases\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true\n  auto_discovery:
          false\n  deployment_name: mariadb\n  deployment_type: DeploymentConfig\n
          \ job_retention: 50\n  overwrite_pcv_on_restore: true\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc: false\n  snapshot_timeout:
          180\n  storage_plugin: csi.so\n  app_plugin: mariadb.so \n</pre>\n```$ oc
          create -f backupconfig.yaml```\n\n### Create PostgreSQL Fossul BackupConfig\nCreates
          a backup config for PostgreSQL databases, you need to ensure the user defined
          in secret has backup permissions\n\n```$ vi backupconfig.yaml```\n<pre>\nkind:
          BackupConfig\napiVersion: fossul.io/v1\nmetadata:\n  name: postgresql-sample\n
          \ namespace: databases\nspec:\n  container_platform: openshift\n  operator_controlled:
          true    \n  auto_discovery: false\n  deployment_name: postgresql \n  deployment_type:
          DeploymentConfig\n  job_retention: 50\n  overwrite_pcv_on_restore: true\n
          \ policies:\n    - policy: hourly\n      retentionNumber: 3\n    - policy:
          daily\n      retentionNumber: 10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc:
          false\n  snapshot_timeout: 180\n  storage_plugin: csi.so\n  app_plugin:
          postgres.so \n</pre>\n```$ oc create -f backupconfig.yaml```\n\n### Create
          Kubevirt Fossul BackupConfig\nCreates a backup configuration for virtual
          machines\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: rhel8-vm-sample\n  namespace: virtualmachines\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true    \n  auto_discovery:
          false\n  deployment_name: rhel8-vm\n  deployment_type: VirtualMachine\n
          \ job_retention: 50\n  overwrite_pcv_on_restore: true\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc: false\n  snapshot_timeout:
          180\n  storage_plugin: csi.so\n  app_plugin: kubevirt.so \n</pre>\n```$
          oc create -f backupconfig.yaml```\n\n### Create Tar Fossul BackupConfig\nCreates
          a backup config using tar to backup folders and files where csi snapshots
          are not desired\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: tar-sample\n  namespace: databases\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true    \n  auto_discovery:
          false\n  deployment_name: anyname\n  job_retention: 50\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  backup_source_paths: /opt/app-root/src/test\n  storage_plugin: tar.so\n</pre>\n```$
          oc create -f backupconfig.yaml```\n\n### Create BackupSchedule\nA backup
          can be scheduled per policy, defined by backup configuration using cron
          syntax\n\n```$ vi backupschedule.yaml```\n<pre>\nkind: BackupSchedule\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ cron_schedule: '59 23 * * *'\n  deployment_name: mariadb\n  policy: daily\n</pre>\n```$
          oc create -f backupschedule.yaml```\n\n### Create Backup\nA backup will
          be created according to policy and deleted either manually or as defined
          in retention policy\n\n```$ vi backup.yaml```\n<pre>\nkind: Backup\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ deployment_name: mariadb\n  policy: daily\n</pre>\n```$ oc create -f backup.yaml```\n\n###
          Perform Restore\nA restore requires a workflow_id located in the backup
          spec, upon a successful restore the custom resource will deleted automatically\n\n```$
          vi restore.yaml```\n<pre>\nkind: Restore\napiVersion: fossul.io/v1\nmetadata:\n
          \ name: mariadb-sample\n  namespace: databases\nspec:\n  deployment_name:
          mariadb\n  policy: daily\n  workflow_id: xxxx\n</pre>\n```$ oc create -f
          restore.yaml```\n"
        displayName: Fossul Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - application
        - consistent
        links:
        - name: Fossul Framework Community
          url: https://github.com/fossul/fossul
        maintainers:
        - email: ktenzer@redhat.com
          name: Keith Tenzer
        maturity: alpha
        provider:
          name: Fossul Framework
          url: https://github.com/fossul/fossul
        relatedImages:
        - quay.io/fossul/fossul-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: fossul-operator.v0.6.0
        version: 0.6.0
      - name: fossul-operator.v0.5.0
        version: 0.5.0
      - name: fossul-operator.v0.4.0
        version: 0.4.0
      - name: fossul-operator.v0.3.0
        version: 0.3.0
      - name: fossul-operator.v0.2.0
        version: 0.2.0
      - name: fossul-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: fossul-operator
    provider:
      name: Fossul Framework
      url: https://github.com/fossul/fossul
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Weaveworks
      provider-url: ""
    name: tf-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tf-controller.v0.9.0-rc.8
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"infra.contrib.fluxcd.io/v1alpha1\",\n
            \   \"kind\": \"Terraform\",\n    \"metadata\": {\n        \"name\": \"helloworld-tf\",\n
            \       \"namespace\": \"flux-system\"\n    },\n    \"spec\": {\n        \"interval\":
            \"5m\",\n        \"approvePlan\": \"auto\",\n        \"path\": \"./\",\n
            \       \"sourceRef\": {\n          \"kind\": \"GitRepository\",\n          \"name\":
            \"helloworld\",\n          \"namespace\": \"flux-system\"\n        }      \n
            \   }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/weaveworks/tf-controller:v0.9.0-rc.8
          description: Weave TF-controller is a Flux controller that brings GitOps
            to Terraform resources.
          operatorframework.io/suggested-namespace: flux-system
          support: Weaveworks
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Terraform
            displayName: Terraform
            kind: Terraform
            name: terraforms.infra.contrib.fluxcd.io
            version: v1alpha1
        description: "**Weave GitOps Terraform Controller** (aka Weave TF-controller)
          is a Flux controller for reconciling Terraform resources in the GitOps way.
          \nWith the power of Flux together with Terraform, TF-controller allows you
          to GitOps-ify infrastructure, \nand application resources, in the Kubernetes
          and Terraform universe, together at your own pace.\n\nTF-controller offers
          many GitOps models:\n  1. **GitOps Automation Model:** GitOps your Terraform
          resources from the provision steps to the enforcement steps, like a whole
          EKS cluster.\n  2. **Hybrid GitOps Automation Model:** GitOps parts of your
          existing infrastructure resources. For example, you have an existing EKS
          cluster. You can choose to GitOps only its nodegroup, or its security group.\n
          \ 3. **State Enforcement Model:** You have a TFSTATE file, and you'd like
          to use GitOps enforce it, without changing anything else.\n  4. **Drift
          Detection Model:** You have a TFSTATE file, and you'd like to use GitOps
          just for drift detection, so you can decide to do things later when a drift
          occurs.\n\nFeatures\n  * GitOps Automation for Terraform\n  * Drift detection\n
          \ * Plan and Manual Approve"
        displayName: Weave GitOps Terraform Controller
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - weave
        - flux
        - cd
        - continuous delivery
        - terraform
        - gitops
        links:
        - name: Weave GitOps Terraform Controller
          url: https://github.com/weaveworks/tf-controller
        - name: Documentation
          url: https://weaveworks.github.io/tf-controller/
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Weaveworks
        relatedImages:
        - ghcr.io/weaveworks/tf-controller:v0.9.0-rc.8
        version: 0.9.0-rc.8
      entries:
      - name: tf-controller.v0.9.0-rc.8
        version: 0.9.0-rc.8
      - name: tf-controller.v0.9.0-rc.3
        version: 0.9.0-rc.3
      name: stable
    defaultChannel: stable
    packageName: tf-controller
    provider:
      name: Weaveworks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Randoli
      provider-url: https://www.randoli.io
    name: app-insights-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: app-insights-operator.v1.17.5-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "appinsights.randoli/v1beta1",
                "kind": "RandoliAgent",
                "metadata": {
                  "name": "randoli",
                  "namespace": "randoli-agents"
                },
                "spec": {
                  "costManagement": {
                    "opencost": {
                      "exporter": {
                        "image": {
                          "tag": "1.112.1"
                        }
                      }
                    },
                    "vpaOperator": {
                      "enabled": false
                    }
                  },
                  "logs": {
                    "vector": {
                      "agent": {
                        "containerSecurityContext": {
                          "privileged": true
                        }
                      }
                    }
                  },
                  "network": {
                    "standaloneConsole": {
                      "enable": true
                    }
                  },
                  "observability": {
                    "netobserv": {
                      "enabled": false
                    },
                    "otel": {
                      "operator": {
                        "enabled": false
                      }
                    }
                  },
                  "resources": {
                    "limits": {
                      "memory": "2Gi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "512Mi"
                    }
                  },
                  "securityContext": {
                    "fsGroup": 1001
                  },
                  "tags": {
                    "costManagement": false,
                    "observability": false
                  }
                }
              }
            ]
          categories: Developer Tools
          containerImage: docker.io/randoli/app-insights-operator:1.17.5-4@sha256:22e7150c29557f2a71186389e5d7d60bfe82a4ea022e1403033c6ad61bc95f7a
          createdAt: "2025-09-23T19:45:43Z"
          description: Randoli Agent connects OpenShift clusters to Randoli Observability
            & Cost Management.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[]'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Randoli
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: RandoliAgent
            name: randoliagents.appinsights.randoli
            version: v1beta1
        description: |-
          Randoli Agent connects OpenShift clusters to Randoli Observability & Cost Management platforms.

          This operator provides a container-based agent to be installed in your OpenShift clusters for connecting to Randoli Observability & Cost Management platforms

          **Prerequisites**

          If you are using the Randoli Observability platform, you will need to install the following Operators:
          - Red Hat build of OpenTelemetry Operator
          - Network Observability

          If you are using the Randoli Cost Management platform, you will need to install the following Operator:
          - VerticalPodAutoscaler

          For additional information please use our Openshift installation guide here: [https://docs.insights.randoli.io/agent/openshift](https://docs.insights.randoli.io/agent/openshift)

           **Randoli Observability for Kubernetes**

          [Product Page](https://www.randoli.io/product/kubernetes-observability)

          **Randoli Cost Management for Kubernetes**

          [Product Page](https://www.randoli.io/product/kubernetes-cost-management)
        displayName: Randoli Agent
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - developer tools
        - insights
        - monitoring
        links:
        - name: Randoli Observability for Kubernetes
          url: https://www.randoli.io/product/kubernetes-observability
        - name: Randoli Cost Management for Kubernetes
          url: https://www.randoli.io/product/kubernetes-cost-management
        maintainers:
        - email: contact@randoli.ca
          name: Randoli Inc.
        maturity: beta
        minKubeVersion: 1.24.0
        provider:
          name: Randoli
          url: https://www.randoli.io
        relatedImages:
        - docker.io/randoli/app-insights-operator:1.17.5-4@sha256:22e7150c29557f2a71186389e5d7d60bfe82a4ea022e1403033c6ad61bc95f7a
        version: 1.17.5-4
      entries:
      - name: app-insights-operator.v1.17.5-4
        version: 1.17.5-4
      - name: app-insights-operator.v1.17.5-3
        version: 1.17.5-3
      - name: app-insights-operator.v1.17.5-2
        version: 1.17.5-2
      - name: app-insights-operator.v1.17.5-1
        version: 1.17.5-1
      - name: app-insights-operator.v1.17.3-1
        version: 1.17.3-1
      - name: app-insights-operator.v1.17.2-1
        version: 1.17.2-1
      - name: app-insights-operator.v1.17.1-1
        version: 1.17.1-1
      - name: app-insights-operator.v1.17.0-2
        version: 1.17.0-2
      - name: app-insights-operator.v1.17.0-1
        version: 1.17.0-1
      name: stable
    defaultChannel: stable
    packageName: app-insights-operator
    provider:
      name: Randoli
      url: https://www.randoli.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ProphetStor Data Services, Inc.
      provider-url: ""
    name: federatorai-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: federatorai.v4.7.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "federatorai.containers.ai/v1alpha1",
                "kind": "AlamedaService",
                "metadata": {
                  "name": "my-alamedaservice"
                },
                "spec": {
                  "alamedaAi": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "10Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "alamedaInfluxdb": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "100Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "fedemeterInfluxdb": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "10Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "imageLocation": "registry.connect.redhat.com/prophetstor",
                  "storages": [
                    {
                      "accessMode": ["ReadWriteOnce"],
                      "size": "2Gi",
                      "type": "pvc",
                      "usage": "log"
                    }
                  ],
                  "version": "v4.7.1-b1760"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: registry.connect.redhat.com/prophetstor/federatorai-operator@sha256:bdd87f6b7d7ba9d12ed8ddb9c5151e5a89781e88d7a02b73814b16a766968047
          createdAt: "2021-10-12T01:14:00Z"
          description: Federator.ai Operator provides easy configuration and management
            of AI-based Kubernetes resource orchestrator
          operatorframework.io/suggested-namespace: federatorai
          repository: https://access.redhat.com/containers/#/registry.connect.redhat.com/prophetstor/federatorai-operator
          support: ProphetStor Data Services, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of Federator.ai.
            displayName: AlamedaService
            kind: AlamedaService
            name: alamedaservices.federatorai.containers.ai
            version: v1alpha1
        description: |-
          Federator.ai helps enterprises optimize cloud resources, maximize application performance, and save significant cost without excessive over-provisioning or under-provisioning of resources, meeting the service-level requirements of their applications.

          Enterprises often lack understanding of the resources needed to support their applications. This leads to either excessive over-provisioning or under-provisioning of resources (CPU, memory, storage). Using machine learning, Federator.ai determines the optimal cloud resources needed to support any workload on OpenShift and helps users find the best-cost instances from cloud providers for their applications.


          **Multi-layer workload prediction**

          Using machine learning and math-based algorithms, Federator.ai predicts containerized application and cluster node resource usage as the basis for resource recommendations at application level as well as at cluster node level. Federator.ai supports prediction for both physical/virtual CPUs and memories.


          **Auto-scaling via resource recommendation**

          Federator.ai utilizes the predicted resource usage to recommend the right number and size of pods for applications. Integrated with Datadog's WPA, applications are automatically scaled to meet the predicted resource usage.


          **Application-aware recommendation execution**

          Optimizing the resource usage and performance goals, Federator.ai uses application specific metrics for workload prediction and pod capacity estimation to auto-scale the right number of pods for best performance without overprovisioning.


          **Multi-cloud Cost Analysis**

          With resource usage prediction, Federator.ai analyzes potential cost of a cluster on different public cloud providers. It also recommend appropriate cluster nodes and instance types based on resource usage.


          **Custom Datadog/Sysdig Dashboards**

          Predefined custom Datadog/Sysdig Dashboards for workload prediction/recommendation visualization for cluster nodes and applications.


          **Additional resources**

          Want more product information? Explore detailed information about using this product and where to find additional help.

          * [Federator.ai Datasheet](https://prophetstor.com/wp-content/uploads/datasheets/Federator.ai.pdf)
          * [Quick Start Guide](https://prophetstor.com/wp-content/uploads/documentation/Federator.ai/Latest%20Version/ProphetStor%20Federator.ai%20Quick%20Installation%20Guide.pdf)
          * [Installation Guide](https://prophetstor.com/wp-content/uploads/2021/08/ProphetStor-Federator.ai-v4.7.0-Installation-Guide-v1.0.pdf)
          * [User Guide](https://prophetstor.com/wp-content/uploads/2021/08/Federator.ai-4.7.0-User-Guide.pdf)
          * [Release Notes](https://prophetstor.com/wp-content/uploads/2021/08/Federator.ai-4.7.0-Release-Notes.pdf)
          * [Company Information](https://www.prophetstor.com/)


          ### Prerequisite
          *  The [Openshift](https://www.openshift.com) version 4.4.x or later if using OpenShift platform.
          *  The [Kubernetes](https://kubernetes.io/) version 1.16 or later if using Kubernetes environment.
        displayName: Federator.ai
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - AI
        - Resource Orchestration
        - NoOps
        links:
        - name: Website
          url: https://prophetstor.com/federator-ai-2/
        - name: Quickstart guide
          url: https://prophetstor.com/wp-content/uploads/documentation/Federator.ai/Latest%20Version/ProphetStor%20Federator.ai%20Quick%20Installation%20Guide.pdf
        maintainers:
        - email: support@prophetstor.com
          name: ProphetStor Data Services, Inc.
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: ProphetStor Data Services, Inc.
        relatedImages:
        - registry.connect.redhat.com/prophetstor/alameda-recommender-ubi@sha256:13e941df9252cea3024ecb5e2247a696b64928015c1a3a4a04ee4e362388b1e4
        - registry.connect.redhat.com/prophetstor/federatorai-agent-preloader@sha256:66daea6193e399f8363df5ba42e1fcd643db6df3a9075dbcd8c3464fdc91e776
        - registry.connect.redhat.com/prophetstor/alameda-executor-ubi@sha256:4512e3be7a89d0ca9f120b68f6ab9766d000bf2b01b279cd55b2dbc433aacfb6
        - registry.connect.redhat.com/prophetstor/alameda-influxdb@sha256:485c2119a6a56807e28ad03b2087129950118b55531126a66bb414063052ae76
        - registry.connect.redhat.com/prophetstor/alameda-analyzer-ubi@sha256:6117ac4692793d49bc0204f13091dec119bdbc7b169e9deb86c243aca00ded27
        - registry.connect.redhat.com/prophetstor/fedemeter-api-ubi@sha256:451456207fe8d16b9824b3f64d00645766dcccbc8d685b6942e6c00d0d0c627e
        - registry.connect.redhat.com/prophetstor/alameda-ai-dispatcher@sha256:6e0dc1136b0f8e5971d92b95ce0d6dfac475044c073bada50f23b315f4478ae4
        - registry.connect.redhat.com/prophetstor/alameda-ai@sha256:9293298bb2a72d0b0764050b6f6d8da707b46a83363a3b6aa7be60f512e2ec61
        - registry.connect.redhat.com/prophetstor/federatorai-dashboard-backend@sha256:022dbf6696edb8eaf389fdbd1dff14c40a51c237c01f8a02a7ea65c60cd59f38
        - registry.connect.redhat.com/prophetstor/federatorai-rest-ubi@sha256:3e92528615543ea900bf38387c4eb8439b60759be5be2ac5311a9dbaafc7026b
        - registry.connect.redhat.com/prophetstor/federatorai-agent-ubi@sha256:cfadc84f31f3776ded15d4d8dda92e00b7d9b1bce0ff5440277176292f9304d3
        - registry.connect.redhat.com/prophetstor/federatorai-operator@sha256:bdd87f6b7d7ba9d12ed8ddb9c5151e5a89781e88d7a02b73814b16a766968047
        - registry.connect.redhat.com/prophetstor/federatorai-dashboard-frontend@sha256:1c1147425ca6302103363a1a3bcf27baeef13b10ec6780ef266b7f52fe27ffeb
        - registry.connect.redhat.com/prophetstor/alameda-notifier-ubi@sha256:1f75034fb78d62924a87cb4a5435661b4afa990d6960431e784029742fbae0da
        - registry.connect.redhat.com/prophetstor/alameda-rabbitmq@sha256:b396d9aa0470e593a36f5799c04fc3c9e59b5723f703f4b019f7ef92528e6741
        - registry.connect.redhat.com/prophetstor/fedemeter-influxdb@sha256:36904f003e8e29d4a9fd738655f401fbe840d60158d7d820cda79a09858dc2c2
        - registry.connect.redhat.com/prophetstor/federatorai-data-adapter@sha256:adf5d655fa030dc05133516d45642d7282fdd66e0aa8e7a7331ed614a41d3fe0
        - registry.connect.redhat.com/prophetstor/alameda-operator-ubi@sha256:9a863394cb1c89228223b26addb0d12bb91a97233fe2d5d19ecd5998e0cfe0d6
        - registry.connect.redhat.com/prophetstor/alameda-datahub-ubi@sha256:3819d6acce8d376dbe3b12469be2528d0ce9ac95c2a86507aeb76b6e3a5f5f38
        version: 4.7.1-1
      entries:
      - name: federatorai.v4.7.1-1
        version: 4.7.1-1
      - name: federatorai.v4.6.1-4
        version: 4.6.1-4
      name: stable
    defaultChannel: stable
    packageName: federatorai-certified
    provider:
      name: ProphetStor Data Services, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-route53resolver-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-route53resolver-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "route53resolver.services.k8s.aws/v1alpha1",
                "kind": "ResolverEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "route53resolver.services.k8s.aws/v1alpha1",
                "kind": "ResolverRule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/route53resolver-controller:1.1.1
          createdAt: "2025-09-20T01:21:08Z"
          description: AWS Route53Resolver controller is a service controller for
            managing Route53Resolver resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ResolverEndpoint represents the state of an AWS route53resolver
              ResolverEndpoint resource.
            displayName: ResolverEndpoint
            kind: ResolverEndpoint
            name: resolverendpoints.route53resolver.services.k8s.aws
            version: v1alpha1
          - description: ResolverRule represents the state of an AWS route53resolver
              ResolverRule resource.
            displayName: ResolverRule
            kind: ResolverRule
            name: resolverrules.route53resolver.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Route53Resolver resources in AWS from within your Kubernetes cluster.

          **About Amazon Route53Resolver**

          Amazon Route 53 Resolver responds recursively to DNS queries from AWS resources for public records, Amazon VPC-specific DNS names, and Amazon Route 53 private hosted zones, and is available by default in all VPCs.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Route53Resolver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - route53resolver
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Route53Resolver Developer Resources
          url: https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/resolver.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: route53resolver maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/route53resolver-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-route53resolver-controller.v1.1.1
        version: 1.1.1
      - name: ack-route53resolver-controller.v1.0.15
        version: 1.0.15
      - name: ack-route53resolver-controller.v1.0.14
        version: 1.0.14
      - name: ack-route53resolver-controller.v1.0.13
        version: 1.0.13
      - name: ack-route53resolver-controller.v1.0.11
        version: 1.0.11
      - name: ack-route53resolver-controller.v1.0.10
        version: 1.0.10
      - name: ack-route53resolver-controller.v1.0.9
        version: 1.0.9
      - name: ack-route53resolver-controller.v1.0.7
        version: 1.0.7
      - name: ack-route53resolver-controller.v1.0.6
        version: 1.0.6
      - name: ack-route53resolver-controller.v1.0.5
        version: 1.0.5
      - name: ack-route53resolver-controller.v1.0.3
        version: 1.0.3
      - name: ack-route53resolver-controller.v1.0.2
        version: 1.0.2
      - name: ack-route53resolver-controller.v1.0.0
        version: 1.0.0
      - name: ack-route53resolver-controller.v0.0.11
        version: 0.0.11
      - name: ack-route53resolver-controller.v0.0.10
        version: 0.0.10
      - name: ack-route53resolver-controller.v0.0.9
        version: 0.0.9
      - name: ack-route53resolver-controller.v0.0.8
        version: 0.0.8
      - name: ack-route53resolver-controller.v0.0.7
        version: 0.0.7
      - name: ack-route53resolver-controller.v0.0.4
        version: 0.0.4
      - name: ack-route53resolver-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-route53resolver-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: eventing-kogito
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eventing-kogito.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kogito.knative.dev/v1alpha1",
                "kind": "KogitoSource",
                "metadata": {
                  "name": "kogito-order-processing"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "kind": "Broker",
                      "name": "default"
                    }
                  },
                  "subject": {
                    "apiVersion": "serving.knative.dev/v1",
                    "kind": "Service",
                    "name": "serverless-workflow-order-processing"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: gcr.io/knative-releases/knative.dev/eventing-kogito/cmd/controller:v1.2.0
          createdAt: "2021-08-25T12:00:00Z"
          olm.targetNamespaces: knative-kogito
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/knative-sandbox/eventing-kogito
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoSource is the representation of a Knative Eventing
              Source for a Kogito custom Service.
            displayName: Kogito Source
            kind: KogitoSource
            name: kogitosources.kogito.knative.dev
            version: v1alpha1
        description: |-
          Knative Eventing Source for Kogito Services. This controller will enable easy configuration of any Kogito custom Service
          with the Knative Eventing platform. Please find more information about using this source in the [Official Kogito Documentation](https://docs.jboss.org/kogito/release/latest/html_single/#_kogito_knative_eventing_add_on)

          **Note:** The admission webhooks are not installed via OLM at this time. Because of that, please ensure that you create the `KogitoSource` objects with the subbject and sink information with the appropriate namespace.
        displayName: Kogito Knative Eventing Source
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless workflow
        - kogito
        - knative
        - dmn
        - bpmn
        - eventing
        - cloudevents
        - messaging
        links:
        - name: Source Repository
          url: https://github.com/knative-sandbox/eventing-kogito
        - name: Kogito Knative Add-On Documentation
          url: https://docs.jboss.org/kogito/release/latest/html_single/#_kogito_knative_eventing_add_on
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - gcr.io/knative-releases/knative.dev/eventing-kogito/cmd/controller:v1.2.0
        version: 1.2.0
      entries:
      - name: eventing-kogito.v1.2.0
        version: 1.2.0
      - name: eventing-kogito.v1.1.0
        version: 1.1.0
      name: alpha
    defaultChannel: alpha
    packageName: eventing-kogito
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KEDA Community
      provider-url: ""
    name: keda
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keda.v2.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "KedaController",
                "metadata": {
                  "name": "keda",
                  "namespace": "keda"
                },
                "spec": {
                  "watchNamespace": "",
                  "operator": {
                    "logLevel": "info",
                    "logEncoder": "console"
                  },
                  "metricsServer": {
                    "logLevel": "0"
                  },
                  "serviceAccount": {}
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledJob",
                "metadata": {
                  "name": "scaledjob-sample"
                },
                "spec": {
                  "failedJobsHistoryLimit": 5,
                  "jobTargetRef": {
                    "activeDeadlineSeconds": 600,
                    "backoffLimit": 6,
                    "completions": 1,
                    "parallelism": 1,
                    "template": null
                  },
                  "maxReplicaCount": 100,
                  "pollingInterval": 30,
                  "successfulJobsHistoryLimit": 5,
                  "triggers": [
                    {
                      "metadata": {
                        "property": "examle-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledObject",
                "metadata": {
                  "name": "example-scaledobject"
                },
                "spec": {
                  "cooldownPeriod": 300,
                  "maxReplicaCount": 100,
                  "minReplicaCount": 0,
                  "pollingInterval": 30,
                  "scaleTargetRef": {
                    "name": "example-deployment"
                  },
                  "triggers": [
                    {
                      "metadata": {
                        "property": "examle-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "TriggerAuthentication",
                "metadata": {
                  "name": "example-triggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "exmaple-secret-parameter"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ClusterTriggerAuthentication",
                "metadata": {
                  "name": "example-clustertriggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "exmaple-secret-parameter"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Cloud Provider
          certified: "false"
          containerImage: ghcr.io/kedacore/keda-olm-operator:2.7.1
          createdAt: "2022-05-11T00:00:00.000Z"
          description: '[DEPRECATED] Use Custom Metrics Autoscaler Operator instead'
          operatorframework.io/suggested-namespace: keda
          operators.operatorframework.io/builder: operator-sdk-v1.1.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/kedacore/keda-olm-operator
          support: KEDA Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              Represents an installation of a particular version of KEDA Controller.
            displayName: KedaController
            kind: KedaController
            name: kedacontrollers.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your jobs and what the triggers
            displayName: ScaledJob
            kind: ScaledJob
            name: scaledjobs.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your application and what the
              triggers are
            displayName: ScaledObject
            kind: ScaledObject
            name: scaledobjects.keda.sh
            version: v1alpha1
          - description: Describes authentication parameters
            displayName: TriggerAuthentication
            kind: TriggerAuthentication
            name: triggerauthentications.keda.sh
            version: v1alpha1
          - description: Describes cluster wide authentication parameters
            displayName: ClusterTriggerAuthentication
            kind: ClusterTriggerAuthentication
            name: clustertriggerauthentications.keda.sh
            version: v1alpha1
        description: "## Deprecated\n This KEDA community operator is deprecated,
          please use `Custom Metrics Autoscaler` Operator instead.\n## About the managed
          application\nKEDA is a Kubernetes-based event driven autoscaler.  KEDA can
          monitor event sources like Kafka, RabbitMQ, or cloud event sources and feed
          the metrics from those sources into the Kubernetes horizontal pod autoscaler.
          \ With KEDA, you can have event driven and serverless scale of deployments
          within any Kubernetes cluster.\n## About this Operator\nThe KEDA Operator
          deploys and manages installation of KEDA Controller in the cluster. Install
          this operator and follow installation instructions on how to install KEDA
          in you cluster.\n\n## Prerequisites for enabling this Operator\nPlease note
          that you can not run both KEDA 1.x and 2.x on the same Kubernetes cluster.
          You need to uninstall KEDA 1.x first, in order to install and use KEDA 2.x.\n\nDon't
          forget to uninstall KEDA v1 CRDs as well, to ensure that, please run:\n\n```
          oc delete crd scaledobjects.keda.k8s.io ```\n\n``` oc delete crd triggerauthentications.keda.k8s.io
          \ ```\n## How to install KEDA in the cluster\nThe installation of KEDA is
          triggered by the creation of `KedaController` resource, please refer to
          [KedaController Example](https://github.com/kedacore/keda-olm-operator#kedacontroller-spec)
          for more deatils on available options.\n\nOnly resource named `keda` in
          namespace `keda` will trigger the installation, reconfiguration or removal
          of the KEDA Controller resource.\n\nThere could be only one KEDA Controller
          in the cluster. \n"
        displayName: '[DEPRECATED] KEDA'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keda
        - scaling
        - event
        links:
        - name: Homepage
          url: https://keda.sh
        - name: Source Code
          url: https://github.com/kedacore/keda
        - name: Slack Channel
          url: https://kubernetes.slack.com/messages/CKZJ36A5D
        maintainers:
        - email: cncf-keda-maintainers@lists.cncf.io
          name: KEDA Community
        maturity: deprecated
        minKubeVersion: 1.17.0
        provider:
          name: KEDA Community
        relatedImages:
        - ghcr.io/kedacore/keda-olm-operator:2.7.1
        version: 2.7.1
      entries:
      - name: keda.v2.7.1
        version: 2.7.1
      - name: keda.v2.6.1
        version: 2.6.1
      - name: keda.v2.6.0
        version: 2.6.0
      - name: keda.v2.5.0
        version: 2.5.0
      - name: keda.v2.4.0
        version: 2.4.0
      - name: keda.v2.3.0
        version: 2.3.0
      - name: keda.v2.2.0
        version: 2.2.0
      - name: keda.v2.1.0
        version: 2.1.0
      - name: keda.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: keda
    provider:
      name: KEDA Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: stolostron
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stolostron.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:3443da607b276d39ba28a8d6561d296f30e9bac5ab06d5c70715bddbde0a58b2
          createdAt: "2024-06-28T04:52:57Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/volsync-addon-controller@sha256:7774767ed20263bba4857a0d4be012e12215f640fce01fbd3a085bf860d087e7
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:9d1d5e2b5313b898188d2c845dcda5d604b73f16e89c71915ea308281c1f56ff
        - quay.io/stolostron/helloworld@sha256:81a0b06c7fcf915b1bb40d33d29d0bfd4d501404463c1f46bffc503e7d1cc45b
        - quay.io/stolostron/volsync@sha256:ba7399b316359118b1c1b915d667fca9a4a0d12a10beae69e6dee9c7236753be
        - quay.io/stolostron/grafana@sha256:e40d66da749108ae1eecf909c95c87bf7a289328e8b373ae4402368c474f2da5
        - quay.io/stolostron/must-gather@sha256:1f7c129859bdaf6bd93dda77aec01f7cd03000fbfaccd809f26d33b9e3f66e51
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/config-policy-controller@sha256:197d8491e7c889db8f2acb726e0146d384c979acc7f634c24a9d4b21f7d444f3
        - quay.io/stolostron/multicluster-operators-channel@sha256:d791aa2c6032c1de50cf4033c426646613b182e98c406a091b88fb9ffefc0c98
        - quay.io/stolostron/volsync-mover-rsync@sha256:f7affcc16ae9bf6d12f69511b7a269f5fd065d1d87392c4c38cb5d7af5b89e23
        - quay.io/stolostron/search-e2e@sha256:8dfbc899ba1ffaa8fb6eb2861f6db87a13d2565cf4551fd47df7fe8253217fb2
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/search-collector@sha256:136ad44ff41a29dec48e87d10af5b608f335e4242c41ac17cba2c3c4c8205bd6
        - quay.io/stolostron/observatorium-operator@sha256:5da7f478ecf3609a3b86eb563e4dd22588f3d13efdc2e33f5b672fe95583ce68
        - quay.io/stolostron/multicluster-operators-application@sha256:199b696dbd7589b77e2391b71504fbcb7d3ae785972dac3d001c4c1bffa342dd
        - quay.io/stolostron/multicluster-operators-subscription@sha256:8350df0a7d4f19c8e52eb58eecf5223bb2a0ccbe4d4aa00fcf8dff832cb43e5d
        - quay.io/stolostron/insights-client@sha256:aac5b9b66800c0b862e8bfa1fea533a4aec854d8c52baddbfa03977e1ed520cb
        - quay.io/stolostron/search-indexer@sha256:068c38489201316dc12d475de03c8abb11575338bd75e25c1262eca8aac49cb2
        - quay.io/stolostron/rbac-query-proxy@sha256:24aafcba20ded0892209cfcc196bff88f877ed51eb928d5d350b17041d170893
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/governance-policy-framework-addon@sha256:a1c0d40561befaf5e849060beaee4b201f42d77c5090823f23ef43a75106b00b
        - quay.io/stolostron/grc-policy-framework-tests@sha256:f2b01192fd900860428942c7d5b900604ffcab0d42731676a1acdeebd3716557
        - quay.io/stolostron/volsync-mover-restic@sha256:284c3358f0f2158d599f27487d76a7fdb21842fc0e95bc639726dfdad14b7e7d
        - quay.io/stolostron/volsync-mover-rclone@sha256:b247b23944e20104f801c172cbbb620740a5b51d1777ecdbf98b4c128f1ef171
        - quay.io/stolostron/volsync-mover-syncthing@sha256:b5fad266ca8d3cd8efcbf4e8241c446004747b551b2b8e06acae10900d7a0f1b
        - quay.io/stolostron/cert-policy-controller@sha256:27673358610dc1ca308de472dbd794dcbb38250bb299b53c736549b1e8441a0f
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:7f6b6e45f0be0a14d7d9d79828572b9400f4d559f084028e3254cfd40948780d
        - quay.io/stolostron/klusterlet-addon-controller@sha256:01e5fca5e5c6caca6878bf8025471b7c77164ca7b4a2d764e20101fc79290df0
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/observability-e2e-test@sha256:8d9341b1814cd085a20f9feceab181a8989d6684fd44d6c8e54a0c6df4dd7216
        - quay.io/stolostron/observatorium@sha256:af4e315e4c6da3607f84e24f31e0f37f8572d161736b4748b2e57e0a2c2dd36c
        - quay.io/stolostron/multicluster-observability-operator@sha256:245e1815a73b5c48f790ae5a5d85fac6642cce8a5a3bf122275809e97a4d519e
        - quay.io/stolostron/multicloud-integrations@sha256:60be5b945ead8f845a29a8be18bd26922b83f009fde1412017f1ca985eb79b55
        - quay.io/stolostron/thanos-receive-controller@sha256:b1a191714cead24f66e8de7217db3c4d4a253ecf0238a912c8b88ba8b47c972d
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/prometheus-alertmanager@sha256:8222bbdd37c246703de54ed3dc109fa8c2f6ca17c5bb90106948683858afdc65
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/governance-policy-propagator@sha256:200558b5aa2366383bd298e159d9f7966a2c9b981038c79cde536edbfbe0092c
        - quay.io/stolostron/prometheus-operator@sha256:490b0056b9a23a1dc514a3036eac6593c7c8961de27d08ff24817d280325138c
        - quay.io/stolostron/submariner-addon@sha256:0f1b4ae59ff5c933542f9affed20bfa6694deb804f2e97f49ebd0f2e1b925eca
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:bf6bec399bc2072186e9d12ed865167fe940f3f619d69ffe6c178c14a3bdadbe
        - quay.io/stolostron/kube-rbac-proxy@sha256:9626cb9b2ee8f00d2abdb14cd2ce56d77435859071c10d8394bc7ee2754ce87f
        - quay.io/stolostron/cluster-permission@sha256:716d2cbcb7f69b5640d04669b15884e0adf053597f278a0bb1d8ee77e4a7af4d
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:02e53b8a6529e3209809dada4113054e3398de18738948cb01ed1d24dcb8e8be
        - quay.io/stolostron/search-v2-api@sha256:6b97b2c7e64de7cd154fd83da3bc690a57b67f01c83170238fe9c2e764df91a8
        - quay.io/stolostron/kube-state-metrics@sha256:ef1d671ffa098ed95186de6f15ea91fe648d103c1c7a57871ed46af07c6f7881
        - quay.io/stolostron/multiclusterhub-operator@sha256:3443da607b276d39ba28a8d6561d296f30e9bac5ab06d5c70715bddbde0a58b2
        - quay.io/stolostron/console@sha256:c7900403ef8dba3aac52f999532d876bbfe2f66f00a255047850594b64021830
        - quay.io/stolostron/prometheus@sha256:8250af047c03c6b0c07b76420e0a5c2338f053d73b2dca1ecba4bb2f211715fe
        - quay.io/stolostron/governance-policy-addon-controller@sha256:593dc0a689091ce7c916e6be764c222c1020629ccff7eafbf17524936e4f8262
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/thanos@sha256:cc78cd786643d27ed85eb3670852722e915a20b20a21479e1477fb859dec6c5e
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/prometheus-config-reloader@sha256:bc93ca9f1b6d95e8b963fe1af4c17f87c8c7d938a348627c29cc3cf42b70ca3e
        - quay.io/stolostron/search-v2-operator@sha256:c41517f372fc1874895b39e70cd1c07391110697a5e49ebcf56f4a5df499d810
        - quay.io/stolostron/grafana-dashboard-loader@sha256:cff97fe3e32dd14ed2f3b7b02cb205715cfee6cfd6dc7e7a630101ea5147e566
        - quay.io/stolostron/cluster-backup-controller@sha256:9c77444bd42e6e99163f4dce9f9975276aea082100f53fd938cddf74ad3b1574
        - quay.io/stolostron/insights-metrics@sha256:49a177357e865d5f8347840fe695531d539d17e28d2b9971821e41ebbee014bd
        - quay.io/stolostron/metrics-collector@sha256:4101a2f2c4230bdc1420a58b546a241dfc5414cf92b57f3b73619c1be14dbc32
        version: 0.4.0
      entries:
      - name: stolostron.v0.4.0
        version: 0.4.0
      name: community-0.4
    - currentCSV: stolostron.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:781f6712369cd4b9bdc8ba0ee2f35a12db198ca190621e199dc10cfea3f5efc9
          createdAt: "2024-10-02T08:01:20Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.5.0 <0.5.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/multicluster-operators-application@sha256:329c82b92fe2448fe26923fbd82d2eadacb7f96c75689d96393c573f95e6d650
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/observatorium-operator@sha256:38f22e30db7713e9f2e239463803cdb5cdfe7108a524ffa5d7e95201bae5300b
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/grc-policy-framework-tests@sha256:57794e2b15949cedf3462df2d38b95e86ca1e140a5ae81e4a9b3e7e06072847d
        - quay.io/stolostron/cert-policy-controller@sha256:809cae1aef02dd3fb68143ebdba818b20c661bb628a7922ee1d705efe4947f35
        - quay.io/stolostron/multicluster-operators-subscription@sha256:30b1b5bf6a7a12e525450905a474babdd7e38235d9be60f3cb04d60a745a1b96
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/governance-policy-framework-addon@sha256:5f2b326a279b7c42c233ea5f11b89ecaf4cfd7c549d021ddcc4e7b0c8926fc6d
        - quay.io/stolostron/search-collector@sha256:4082f7a37aa3f10c725bb9bebf94ee8dab97d22302cd5e59b35d0e3bcc888cef
        - quay.io/stolostron/volsync-mover-restic@sha256:284c3358f0f2158d599f27487d76a7fdb21842fc0e95bc639726dfdad14b7e7d
        - quay.io/stolostron/prometheus-alertmanager@sha256:19f2e277cbdd2bd4814b8d8034167c1a7cfe9f1d698fb16f7604fa65cb8bcb1c
        - quay.io/stolostron/observability-e2e-test@sha256:c361f5a0f806498c6c1150c0c0d5a42222efff13a18cb28dbfd5e4bd835cb4b7
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/prometheus@sha256:147f3a0cee8fa7b11c82e21ca95c28278e1ab53ccb5b70f3af0425bb8a74f61b
        - quay.io/stolostron/metrics-collector@sha256:c580106b2c7173251afa6714b710fd109811023d88a8009fb741f3d00e486e76
        - quay.io/stolostron/siteconfig-operator@sha256:7f47f0c61d5b53411ed78046c5a02664c6693d940ef37d6db1853670196be334
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:a8d812a527902bd4d9dfd23f90ce90f415f5a6fabb007d1c41b4c998e1ca098a
        - quay.io/stolostron/must-gather@sha256:362bbfd54059975369ce6d3869d34e346a0e76159163714cfcb3597022b7c88f
        - quay.io/stolostron/kube-state-metrics@sha256:dc534373feda042afa385026c7da674912ca310df2a4598d83451d4474bb3f44
        - quay.io/stolostron/volsync-addon-controller@sha256:d0b15da92e0ff9792d750b7b48d72b38d6abb91488c20578c7be3382c561877f
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/multicluster-operators-channel@sha256:260a11aae954299ce8fbb0f9c3102afd3b5cf3f1b972b5a71e56c273ba1d532f
        - quay.io/stolostron/volsync-mover-rclone@sha256:b247b23944e20104f801c172cbbb620740a5b51d1777ecdbf98b4c128f1ef171
        - quay.io/stolostron/volsync@sha256:ba7399b316359118b1c1b915d667fca9a4a0d12a10beae69e6dee9c7236753be
        - quay.io/stolostron/klusterlet-addon-controller@sha256:cf4cad3599c4ce538dc007cda9f5b06538799e8198912d907058ecdd4bbb2451
        - quay.io/stolostron/volsync-mover-syncthing@sha256:b5fad266ca8d3cd8efcbf4e8241c446004747b551b2b8e06acae10900d7a0f1b
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:d42ddd30570ccdf388ccd65cd3f7c72de2d857d328e05aef8e855508ccf91898
        - quay.io/stolostron/search-e2e@sha256:ddaa50ecfafcd7501903420a47495f733dce36fcb91c8ce9b8d94f4448ed501d
        - quay.io/stolostron/console@sha256:e42d2f7e39a93f13003193142a3d1f3f9e55966cbf55b94eac93c726373a2516
        - quay.io/stolostron/prometheus-config-reloader@sha256:337f7d2036c17459a41884e38c569a3601364b02660946b49d39ce40ef97bd48
        - quay.io/stolostron/prometheus-operator@sha256:5ff1f7a4be0216f582f5cb40de3412cbe8126433144c8201750f6c3c0fee920e
        - quay.io/stolostron/grafana@sha256:c9869c7f590bc4c54151c7e4fa9ed76dfe35c24a0ea3c19a6a1b26e61924b6e1
        - quay.io/stolostron/cluster-permission@sha256:716d2cbcb7f69b5640d04669b15884e0adf053597f278a0bb1d8ee77e4a7af4d
        - quay.io/stolostron/governance-policy-addon-controller@sha256:9f387eb8b54d1c86fc20d205dbce199738849cb0ce15a47052ddbee8781d2d9b
        - quay.io/stolostron/search-v2-operator@sha256:6df11a36ea19a1b3da191964f851dee95b87fd993e8f0ef8906b1e63dd450d17
        - quay.io/stolostron/thanos@sha256:dd4d4f8fbb9a8f51134260fc728d19f61a93b511fd7889636089944ac46163bd
        - quay.io/stolostron/search-indexer@sha256:7471f2808686d42c2eae9f0a537f9b520648f31e87072a376243f309fd044caa
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/insights-client@sha256:aac5b9b66800c0b862e8bfa1fea533a4aec854d8c52baddbfa03977e1ed520cb
        - quay.io/stolostron/acm-cli@sha256:85a5c55dfa4708435807d1266b5607525e33f55c5b4672fd8ba78987a1c48ef3
        - quay.io/stolostron/multicluster-observability-operator@sha256:fb0519952ef6436cc66496409f00e89cd6e58a41776df96859a86176a7322d05
        - quay.io/stolostron/rbac-query-proxy@sha256:2f1c78176d0ca0b21c237a9f7a3f77f7d857dd848f3a42254cee1e4e469b4c3c
        - quay.io/stolostron/governance-policy-propagator@sha256:69ab8f492fa418596db6da578b3b579fc2f1506c5bf780e6aba3ccd05d8b02b9
        - quay.io/stolostron/cluster-backup-controller@sha256:688844c722db78f383d90986150bde93bb421068e7f5b4128d19043f3e01f93e
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/observatorium@sha256:8d0f02996b5bf02be02ff3743821909b6c2396f0f0519652dde0c5d1d76f8a05
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/kube-rbac-proxy@sha256:1b708bff7fb46391f2460d34be28d30add38501cdb084306df01faae75637232
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/submariner-addon@sha256:07287987fb0041efc16557cb9d7403dc15b5959a5d0dd0d81cd72bc2658cba10
        - quay.io/stolostron/config-policy-controller@sha256:5af6a0e17da7b7261ac71bcb1e2d991856bbb5450686d188026798e05f0617ba
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/volsync-mover-rsync@sha256:f7affcc16ae9bf6d12f69511b7a269f5fd065d1d87392c4c38cb5d7af5b89e23
        - quay.io/stolostron/thanos-receive-controller@sha256:58c97fe5ec2131f9dcfb70563b59ed0bcae94d8bddefa51cb4784e73b1d90e4f
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/insights-metrics@sha256:49a177357e865d5f8347840fe695531d539d17e28d2b9971821e41ebbee014bd
        - quay.io/stolostron/search-v2-api@sha256:338634d767383716c7a88ed2bc9c58275e9d5953932293443725883d7efcffd5
        - quay.io/stolostron/grafana-dashboard-loader@sha256:85031d1a56514be239c585487aec05a8cbb08c17aeb60c0c84c685eb9d9f08b5
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:dbdbec1600d57c440173aecf19c57709d69e945dcecb348bcf4a7c37b3700b70
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:358ac0b0a3afdfac06f784f2e02a305cbef4e6f648e31a00981d3b097746412a
        - quay.io/stolostron/helloworld@sha256:48724b53542a3fcaf342d6ae2056bdf846cbe2ee50879613953b045bfcbe1c0a
        - quay.io/stolostron/multicloud-integrations@sha256:d4cb3a0a50bc5d22f48c96ed3a10ea9843c6ef2f83daeae411a7ddb7ea332ea6
        - quay.io/stolostron/multiclusterhub-operator@sha256:781f6712369cd4b9bdc8ba0ee2f35a12db198ca190621e199dc10cfea3f5efc9
        version: 0.5.3
      entries:
      - name: stolostron.v0.5.3
        version: 0.5.3
      - name: stolostron.v0.5.2
        version: 0.5.2
      - name: stolostron.v0.5.1
        version: 0.5.1
      - name: stolostron.v0.5.0
        version: 0.5.0
      name: community-0.5
    - currentCSV: stolostron.v0.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:09f1f4266938fba88376a50c3e2837c5517b4f3a947cf86638a17e52eb17c1f6
          createdAt: "2025-02-24T01:42:20Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.6.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/prometheus-config-reloader@sha256:2b26d2fb5c69e54aecfbcaab46e652d136d726f0604844031c8771874f1fd769
        - quay.io/stolostron/thanos-receive-controller@sha256:d73f98e93a258adffc1a4bd875c79a9629166599e68ef777a790a1f3146a2929
        - quay.io/stolostron/observatorium-operator@sha256:5eb4c0da92c62b664aaf3bc187924870275e687440c3b0a3ad4f1c760c8b35e9
        - quay.io/stolostron/insights-client@sha256:978d8b7c3e692272024b71b8c005b752c08dc643dde8258ad75961ddcdcce67f
        - quay.io/stolostron/observability-e2e-test@sha256:03403aeb59b364f578761ae4bcb302bd3589baff500d733775e8cb4f2f961236
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/insights-metrics@sha256:f2f4cf14090252af32832ff6df8a9f896e5314751199156c4e5ae0755e4515e6
        - quay.io/stolostron/cluster-backup-controller@sha256:c85f88d42b42e4366f8ff7ad6c3e6b375e98c5b6a12a9c24a6333929e48f1811
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/helloworld@sha256:56f0d641d4f07b922f1c28c536eacad747c56a564e0a65f084106c2ffc9bd99d
        - quay.io/stolostron/multicluster-operators-channel@sha256:501a5053bd3fd5cc9de34e7e020890d731321b090a0e5bf11b8b3350d002f573
        - quay.io/stolostron/thanos@sha256:53eee303836070a7c7621bfb2a094e3809146497d1653dc65d16055aacea332c
        - quay.io/stolostron/flightctl-api@sha256:64fb1b655bef6aeff03cf64b6450155d3cb570ed4713f084590fe515f4fbad02
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/metrics-collector@sha256:c7433ee4ccdb56717802acf258530e2d6d237787d760a8fbcfd642790254d5ab
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/kube-state-metrics@sha256:cd24916e55ad905e386fa3df8a648272114353a0d75283224a4dddb1955b2687
        - quay.io/stolostron/multicluster-operators-application@sha256:59802fcfae5b6fd0bebe58e98412885e71f32464b8dbbf85364ddaf090aa3e00
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/multicluster-observability-operator@sha256:bed0dd30646d7fce894a625e6cf52e7829c272f47dde0baf1f788e9f85f66e42
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/search-v2-api@sha256:694523242aabb8db1f27b5afcd8ab2f63a9cdb6056c33a79badf8e68a337738d
        - quay.io/stolostron/config-policy-controller@sha256:8cf54eb6f5b49a60ed165fa4b70e4984a1362f2f31e8304c3a22e8a6a953a606
        - quay.io/stolostron/governance-policy-framework-addon@sha256:047124271d0b30176a94454b293c82c2eb67a6f06757fc22c54c455fd0913157
        - quay.io/stolostron/cert-policy-controller@sha256:98dd74d7375ca2b546075d5c8f06a9e2c4f2ce3b77247dccf36bcaa46bf564c7
        - quay.io/stolostron/flightctl-worker@sha256:4dddded7d2d4cce63972c8ca81f213c2cedb551cc645ec85495a9692dc36bb4c
        - quay.io/stolostron/multicluster-operators-subscription@sha256:a056ce7d438f68d33ca3dfd42ec3d9a34b347cdf05499066412784109726e7ad
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:92369b6deb21ab3cb0a12ae5c74b455cfefb8698826bca39e26648e608e48a2b
        - quay.io/stolostron/grafana-dashboard-loader@sha256:6c6c2aa528373536d5db66fffeb87a85b79f3cf723e5d9eeaefcb74f91c1a8a1
        - quay.io/stolostron/search-indexer@sha256:4f9be902fe976fd88a77b737cc1f56f00d48740a6a62661468328cf3bb9da31a
        - quay.io/stolostron/must-gather@sha256:2bf9d3c15f32a6bfe0a7a8d39b9dc62ac6cbb6fe47b7cf1f6bc5e30b17e5e849
        - quay.io/stolostron/flightctl-ocp-ui@sha256:6cff6425850d2b4a8b885a684afb01244d7cd9de7035d7ad71b1e10e5bab25e9
        - quay.io/stolostron/console@sha256:eed52f3ab860a3a224ce178d92765ca50d72d6261f04df040a39577881738252
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/governance-policy-propagator@sha256:c21574a2421fefe3fcabeef14640d6fe699d950106c9ed056b3ba9be3cbeabe0
        - quay.io/stolostron/observatorium@sha256:398ecee4d3f12e84e48b12009a4b02fc8b360dfc0c1607ffecc26e6bc04d0ea2
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/prometheus@sha256:3d1810567f67ae1a7bb024a1c4e494ee510cfb14c7ed21a8b5e63e41fa7bcd68
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/search-collector@sha256:54ece9741641e1bb6e30f15d97a65c92b9c6408ba2572a2f6658f537134adff1
        - quay.io/stolostron/rbac-query-proxy@sha256:096ad23c9b9c7e90451c09b48ded67a82575e75c8cc19f22c459d63ae8b5a786
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:05da11b343938866c5afcfe844bcd15d9c513fdb10fd7ea0d3473b455b427c2a
        - quay.io/stolostron/grc-policy-framework-tests@sha256:9b84dbafef35bd568b2bf79283b6850ba1ff57d70239a27cc0cc815fb1a55515
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/grafana@sha256:51092f6754106a90aff28cdd80c2f0447c85c1bbc72db15878d0b65b93e1ead6
        - quay.io/stolostron/prometheus-alertmanager@sha256:b98480f8ba6ffd19fcee3b412757ad34428e6feab7e4196bb13f2f940306e2c9
        - quay.io/stolostron/search-v2-operator@sha256:85e53a97240f6fd1c4251101146cc2bcc51eb91d9b171efb5b58edade9e0244e
        - quay.io/stolostron/volsync-addon-controller@sha256:d9dad837f2f735674e5711c7fff1cfe2c0c92160fceb7d1548a62348dd43b81c
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/prometheus-operator@sha256:f80e79ea87cde66955e1e46d264e414c072085da9db6876ee2ae68c87d2452e5
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:9c85fb91c4be3287cac2a376465836689df885902099096f0f76ffca50a6d311
        - quay.io/stolostron/siteconfig-operator@sha256:97dc128666780ec9a4c955f392ab98f81f0d6038ec2159e8b1d55aab393cf118
        - quay.io/stolostron/acm-cli@sha256:8c8c059ce2f6d131c410972bf8fd0b7e399cf17cd044c7ddc89a9668c0ceac4c
        - quay.io/stolostron/flightctl-ui@sha256:fea717f3b8b4c343ec60ce7d74a691f1124b6847d55e30a73dc0d50ea610d370
        - quay.io/stolostron/cluster-permission@sha256:443d8b30e3bb83c139c64ddfc38b0d5e6d7a25a4fae783303ab2f2c27d015b26
        - quay.io/stolostron/search-e2e@sha256:43e1bbfb7259612ffb408108c71b825f6e61a268c7e405b8d855f6878b6f96dd
        - quay.io/stolostron/multicloud-integrations@sha256:0081065f9e59c5ed41087344668ac86012f796cdbbf3575fd8491c7dfd24fbf8
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/klusterlet-addon-controller@sha256:39e6940eb956fe9640342e11ecbcbc7d6df0169e1b340a60afc19aff7dd713a5
        - quay.io/stolostron/flightctl-periodic@sha256:a3ed745dc65a5886c4cc05348cf4822bfbf2c6d28b85a39fa0b8f3e1f53fdf8a
        - quay.io/stolostron/submariner-addon@sha256:9c72b5177de8ddc56dd3544997e9bd222895545b7c578e43e4051596fb1fa873
        - quay.io/stolostron/multiclusterhub-operator@sha256:09f1f4266938fba88376a50c3e2837c5517b4f3a947cf86638a17e52eb17c1f6
        - quay.io/stolostron/kube-rbac-proxy@sha256:5b521e4f0e5e54c53cea714020069aa796da23d1482d14e4b7c908098ac69ae9
        - quay.io/stolostron/governance-policy-addon-controller@sha256:fa0738abc0368f9dadd8b0845d868512d26be3525a1a07b9de8a9cefafaa42b5
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:51a8acc47098175db148ebc19a10cdba2803d4a9f5884dd38b1c839702398789
        version: 0.6.3
      entries:
      - name: stolostron.v0.6.3
        version: 0.6.3
      - name: stolostron.v0.6.2
        version: 0.6.2
      - name: stolostron.v0.6.1
        version: 0.6.1
      - name: stolostron.v0.6.0
        version: 0.6.0
      name: community-0.6
    - currentCSV: stolostron.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:ab6660283a25bdf2ea9afe8c76bfbc32f4610d69af2bb13a7bca37a1f5961bdd
          createdAt: "2025-06-09T15:10:05Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.7.1'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MulticlusterHub defines the configuration
              for an instance of a multicluster hub, a central point for managing multiple
              Kubernetes-based clusters. The deployment of multicluster hub components
              is determined based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/governance-policy-framework-addon@sha256:b1b24c40d68415525e2bae21810fc439d3e96fcc5fa75c92e7f2eee284422bfd
        - quay.io/stolostron/thanos@sha256:9d504e2aa1001135eb6e454c58c620d102f2b0a85016b7587452dd1595794c93
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:4b29cc08c9e35054a4a820472ea19cd51e043ab1a90495418a36922df74b9540
        - quay.io/stolostron/insights-client@sha256:2cbd20252f10a5a64f93af50f117f05a865aa9d578e5f6e55e4249470b761604
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/multicluster-operators-application@sha256:9f804b4f2da910e0d7f122176f72c0b0ef4b30785fde0369a66f95f4bac58a82
        - quay.io/stolostron/multicluster-operators-subscription@sha256:d7c137379781ce6b697f73a464b89d04fbc20c3da00ce21a195ac79ab57a6dc7
        - quay.io/stolostron/cert-policy-controller@sha256:6deb41867f8781f4e34f69a54d4aeca190d588ba2991279d39cb5a82c2a5c5be
        - quay.io/stolostron/cluster-permission@sha256:e45139d459951b2e751018fe7ecabe1ff22b90ac87be81fb8f17a12f7b7f6a65
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/governance-policy-addon-controller@sha256:6ecc7557168177474423cea60eadc07d8382adb3e97dee669a8590859a41b08c
        - quay.io/stolostron/must-gather@sha256:ad8db72d559d06f4d467f78213d4f8f2cb00f8e76e942d8e75394eb7dadd29d2
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/multiclusterhub-operator@sha256:ab6660283a25bdf2ea9afe8c76bfbc32f4610d69af2bb13a7bca37a1f5961bdd
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/flightctl-periodic@sha256:a3ed745dc65a5886c4cc05348cf4822bfbf2c6d28b85a39fa0b8f3e1f53fdf8a
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/redis-7-c9s@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/kube-state-metrics@sha256:3503df61f3d9ebf3cbf66deef865450b32c7403839fdaca42504febaac00687c
        - quay.io/stolostron/volsync-addon-controller@sha256:224a0a03a25e5ce61f4c50f6252370876ac0cbf4f9711e997e98a00d6a124fe9
        - quay.io/stolostron/thanos-receive-controller@sha256:f445bbfc18bbc4646e0432842022d3be5305809e52ddcb9daf0ec000844d5854
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:54c6a2eec497eafdbf861fb95d78d961fb24906b197fc69b4e96d3e8b53e6496
        - quay.io/stolostron/search-v2-api@sha256:786866ca8c71a8263ddef45ca9bd9e48cfd1b3ad6215725fbf6f79967bda20ad
        - quay.io/stolostron/flightctl-worker@sha256:4dddded7d2d4cce63972c8ca81f213c2cedb551cc645ec85495a9692dc36bb4c
        - quay.io/stolostron/multicloud-integrations@sha256:665e9385531674a5849aa661eb207a0c6e39ea563fc891c57678cddbf7878d8e
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:1efd8e470b88a992e2f35ce18bdf4970f91af7ba917b8d92627b47ce8c10eb40
        - quay.io/stolostron/siteconfig-operator@sha256:16b00f6c4d81eea26e663ca5596526d85bf7777728d1f0ca7e4dd765c64d2bb5
        - quay.io/stolostron/console@sha256:ba89196efe30c812a767d58a41170b56e2b04167a72e7f13180bf40474bdc824
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/insights-metrics@sha256:861dc03fff55c1d9aff96b5ed15b9d38d5932df1c00a4423e35ead19da6eda03
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/stolostron/flightctl-api@sha256:64fb1b655bef6aeff03cf64b6450155d3cb570ed4713f084590fe515f4fbad02
        - quay.io/stolostron/origin-cli@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/prometheus-config-reloader@sha256:024ee6dda10c6d22ddeab57d7d6503b7aa9ac78981cfcec8ffae94b4914500e3
        - quay.io/stolostron/search-indexer@sha256:58320b54847e3bb6baa558bda65e9d0ae2198df480ba27f46f96ba1454614508
        - quay.io/stolostron/observability-e2e-test@sha256:4e60d3e3e1b5dbab3bcc1c4d37a4940a2a42a065a9df539fb243fb174dde6716
        - quay.io/stolostron/search-v2-operator@sha256:2ae9ab48615966ccf91b2022ccda12913a3db3cfbad5351dd8a20648b01137d8
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/search-e2e@sha256:1c7dcc9f3588620b3570a493c2d744de059f425b572b9cb5725142b89414a904
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:05da11b343938866c5afcfe844bcd15d9c513fdb10fd7ea0d3473b455b427c2a
        - quay.io/stolostron/submariner-addon@sha256:51359ccdc0ea29bcaf7ec2b190878e9f3333b1530cb1128e142d893fb87eba8b
        - quay.io/stolostron/grafana@sha256:cebaed7e5abe0064c65ea69bd4e14794d153de9172c4c3b6de4730f23fed7d24
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/rbac-query-proxy@sha256:1c198b6388a72ccfe2c5b68d57a2b22893fb65d977992ec44ee8cad6ee1ca976
        - quay.io/stolostron/multicluster-observability-operator@sha256:3d95297297a301fca1915268fec71d7a6ecef98c6c2ab8494d9befacb2e169fa
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/grc-policy-framework-tests@sha256:327122b13ec79ea488d6f424806549e36ebfbabd055f9b40217ba541f4a118a9
        - quay.io/stolostron/multicluster-operators-channel@sha256:318d885ab33d5afe17334e831c955f1c3cdf804281b7cc8523330820ac22eee2
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/observatorium-operator@sha256:2f11444703addec163a00c99529c228594fb52df13de47f08faac2e446bec483
        - quay.io/stolostron/prometheus-alertmanager@sha256:a9da6f01cfa61fef803f09ef3f6b27e33a07b97fc1f1e11b6a6c8e13b8a4b90e
        - quay.io/stolostron/klusterlet-addon-controller@sha256:4cb2f4f4e14cd67a4b2f9549cba714b635d9e1c2d8b869b0e0573fd7a41bc13e
        - quay.io/stolostron/prometheus-operator@sha256:8c1c12562ec00da8eaedb7d428694ebdd3a749c0b1ce03fe2b5663ae40e1c701
        - quay.io/stolostron/observatorium@sha256:18a06117ef114a1e68474fd9832adacd30b03f60ec1aa55374285a20e5311f13
        - quay.io/stolostron/governance-policy-propagator@sha256:2b77422fb03177f647c04716e7242bc08f241a1ef9c8968e492ae57816f4c4fb
        - quay.io/stolostron/config-policy-controller@sha256:b1718d5facb95c5572ac658c899f9fce0a7e8c08de93804de3327cd50cdb073b
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/grafana-dashboard-loader@sha256:6c6c2aa528373536d5db66fffeb87a85b79f3cf723e5d9eeaefcb74f91c1a8a1
        - quay.io/stolostron/prometheus@sha256:b959035ff754050ca092863bb74c35ba9622d9adc9e8eb7caf3b430cfe211730
        - quay.io/stolostron/acm-cli@sha256:2f0cec3bed4ad77f573e8cd51a7e2c62d68a59e4e261ea801d5ac203fd2724fc
        - quay.io/stolostron/flightctl-ocp-ui@sha256:6cff6425850d2b4a8b885a684afb01244d7cd9de7035d7ad71b1e10e5bab25e9
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/kube-rbac-proxy@sha256:e5555e256a518ca34bea01c959bbb5497582b53dbfaaf326e1852430c175637a
        - quay.io/stolostron/flightctl-ui@sha256:fea717f3b8b4c343ec60ce7d74a691f1124b6847d55e30a73dc0d50ea610d370
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/cluster-backup-controller@sha256:e576c2de1f931b7d33da2dbc03d71ce323766bd9158bf4b128221ddd1ab3fb9f
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/search-collector@sha256:2b3dfd21c0ba7b7b84fac390002e47fae927cd5fdf13e5d1d960703aa6c5bede
        - quay.io/stolostron/helloworld@sha256:4465550664bd210cf0df8edfd706de2fc51cacf01b66a962b5607935f9666ebc
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/metrics-collector@sha256:c7433ee4ccdb56717802acf258530e2d6d237787d760a8fbcfd642790254d5ab
        version: 0.7.1
      entries:
      - name: stolostron.v0.7.1
        version: 0.7.1
      - name: stolostron.v0.7.0
        version: 0.7.0
      name: community-0.7
    defaultChannel: community-0.7
    packageName: stolostron
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MinIO
      provider-url: https://min.io
    name: minio-object-store-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-object-store-operator.v2025.8.19175300
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aistor.min.io/v1",
                "kind": "ObjectStore",
                "metadata": {
                  "annotations": {
                    "prometheus.io/path": "/minio/v2/metrics/cluster",
                    "prometheus.io/port": "9000",
                    "prometheus.io/scrape": "true"
                  },
                  "labels": {
                    "app": "minio"
                  },
                  "name": "my-objectstore"
                },
                "spec": {
                  "cache": {
                    "enabled": false
                  },
                  "certificates": {
                    "disableAutoCert": false
                  },
                  "configuration": {
                    "name": "storage-configuration"
                  },
                  "firewall": {
                    "enabled": false
                  },
                  "mountPath": "/export",
                  "podManagementPolicy": "Parallel",
                  "pools": [
                    {
                      "name": "pool-0",
                      "securityContext": {
                        "fsGroupChangePolicy": "OnRootMismatch"
                      },
                      "servers": 4,
                      "volumeClaimTemplate": {
                        "apiVersion": "v1",
                        "kind": "persistentvolumeclaims",
                        "spec": {
                          "accessModes": [
                            "ReadWriteOnce"
                          ],
                          "resources": {
                            "requests": {
                              "storage": "1Ti"
                            }
                          },
                          "storageClassName": "standard"
                        }
                      },
                      "volumesPerServer": 4
                    }
                  ]
                }
              },
              {
                "apiVersion": "aistor.min.io/v1alpha1",
                "kind": "AdminJob",
                "metadata": {
                  "name": "minio-job"
                },
                "spec": {
                  "commands": [
                    {
                      "args": {
                        "name": "iac-bucket"
                      },
                      "op": "make-bucket"
                    }
                  ],
                  "containerSecurityContext": {},
                  "objectStore": {
                    "name": "my-objectstore",
                    "namespace": "my-objectstore"
                  },
                  "securityContext": {},
                  "serviceAccountName": "my-objectstore-sa"
                }
              },
              {
                "apiVersion": "sts.min.io/v1beta1",
                "kind": "PolicyBinding",
                "metadata": {
                  "name": "example-binding"
                },
                "spec": {
                  "application": {
                    "namespace": "my-objectstore",
                    "serviceaccount": "my-objectstore-sa"
                  },
                  "policies": [
                    "consoleAdmin"
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,AI/Machine Learning
          createdAt: "2025-08-27T14:50:56Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AdminJob
            name: adminjobs.aistor.min.io
            version: v1alpha1
          - kind: ObjectStore
            name: objectstores.aistor.min.io
            version: v1
          - kind: PolicyBinding
            name: policybindings.sts.min.io
            version: v1alpha1
          - kind: PolicyBinding
            name: policybindings.sts.min.io
            version: v1beta1
        description: "MinIO AIStor extends the most sophisticated object store on
          the market. It is designed for the exascale data \ninfrastructure challenges
          presented by modern AI workloads. AIStor provides new features, along with
          performance and\nscalability improvements, to enable enterprises to store
          all AI data in one infrastructure.\n\nMinIO is the standard for building
          <a href=\"https://min.io/solutions/object-storage-for-ai\">large scale AI
          data infrastructure</a>.\n\nThe MinIO AIStor is a software-defined, Amazon
          S3 compatible object store that will run anywhere - from the public cloud
          to the edge.<br><br>\n\nEnterprises use MinIO to deliver against artificial
          intelligence, machine learning, analytics, application, backup and archival
          workloads - all from a single platform.\n\nThe use cases include storage
          for HDFS replacements, AI datalakes, SIEM datalakes, analytics datalakes,
          object storage as a service, streaming and database warm/hot storage.<br><br>\n\nMinIO’s
          AIStor is the fastest object storage server, with production READ/WRITE
          speeds in excess of 2.2 TiB/s on commodity hardware.\n\nBuilt for the cloud
          operating model it is native to the technologies and architectures that
          define the cloud. These include containerization, orchestration with Kubernetes,
          microservices and multi-tenancy - making it the most widely integrated object
          store in the market today.<br><br>\n\nThe AIStor offers a host of enterprise
          features including inline erasure coding, bit-rot detection, state-of-the-art
          encryption, active-active replication, object locking, lifecycle management,
          observability, key-management, caching, firewall/load balancing, global
          console, search and identity + access management.<br><br>\n\nThe AIStor
          is priced on usable capacity and includes support through the MinIO Subscription
          Network.\n\nThe Subscription Network is designed to optimize deployments
          of MinIO and reduce the risk associated with data loss and data breach while
          providing 24/7/365 direct-to-engineering support.<br><br>\n\n<a href=\"https://resources.min.io/request-a-demo-marketplaces/\">Request
          a demo today</a>"
        displayName: Minio AIStor Object Store
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Storage
        - S3
        - MinIO
        - Object Storage
        links:
        - name: Minio AIStor
          url: https://min.io
        - name: Support
          url: https://subnet.min.io
        - name: Github
          url: https://github.com/minio/aistor
        maintainers:
        - email: dev@min.io
          name: MinIO
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: MinIO
          url: https://min.io
        relatedImages:
        - quay.io/minio/aistor/operator@sha256:f80139efa062fdd83f78318b20f371a827b68e854b03dd9706b9eba4c493aeb2
        version: 2025.8.19175300
      entries:
      - name: minio-object-store-operator.v2025.8.19175300
        version: 2025.8.19175300
      - name: minio-object-store-operator.v2025.8.1184927
        version: 2025.8.1184927
      - name: minio-object-store-operator.v2025.7.1011319
        version: 2025.7.1011319
      - name: minio-object-store-operator.v2025.6.3065135
        version: 2025.6.3065135
      - name: minio-object-store-operator.v2025.5.12190907
        version: 2025.5.12190907
      name: stable
    defaultChannel: stable
    packageName: minio-object-store-operator
    provider:
      name: MinIO
      url: https://min.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: https://stakater.com
    name: template-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: template-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-07-04T00:40:36Z"
          description: Operator to distribute plain kubernetes manifests, helm charts
            and, copy secrets and configmaps to multiple namespaces
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.0.1
          operatorframework.io/suggested-namespace: template-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterTemplateInstance is the Schema for the clustertemplateinstances
              API
            displayName: Cluster Template Instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateGroupInstance is the Schema for the TemplateGroupInstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.templates.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
        description: '"Operator to distribute plain kubernetes manifests, helm charts
          and, copy secrets and configmaps to multiple namespaces"'
        displayName: Template Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - templates
        - secrets
        - configmaps
        - distribution
        - stakater
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/template-operator@sha256:ad8312417e4e9621b74d3ad03020516840feef9c4b52c4f0d894dbb6dc6f906a
        version: 0.0.1
      entries:
      - name: template-operator.v0.0.1
        version: 0.0.1
      name: release-0.0
    - currentCSV: template-operator.v0.1.5
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-08-29T06:41:42Z"
          description: Operator to distribute plain kubernetes manifests, helm charts
            and, copy secrets and configmaps to multiple namespaces
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.1.5
          operatorframework.io/suggested-namespace: template-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterTemplateInstance is the Schema for the clustertemplateinstances
              API
            displayName: Cluster Template Instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateGroupInstance is the Schema for the TemplateGroupInstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.templates.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
        description: '"Operator to distribute plain kubernetes manifests, helm charts
          and, copy secrets and configmaps to multiple namespaces"'
        displayName: Template Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - templates
        - secrets
        - configmaps
        - distribution
        - stakater
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/template-operator@sha256:88f722baf22ac07704503f098f2b0ea7d76815a1ab0552474b1c9ba6639d8142
        version: 0.1.5
      entries:
      - name: template-operator.v0.1.5
        version: 0.1.5
      name: release-0.1
    defaultChannel: release-0.1
    packageName: template-operator
    provider:
      name: Stakater
      url: https://stakater.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Starburst Data, Inc.
      provider-url: https://www.starburstdata.com
    name: starburst-enterprise-helm-operator-paygo-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: starburst-enterprise-helm-operator-paygo.v402.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstEnterprise",
                "metadata": {
                  "name": "starburstenterprise-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "cache": {
                    "diskUsagePercentage": 80,
                    "enabled": false,
                    "ttl": "7d",
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "catalogs": {
                    "tpch": "connector.name=tpch\ntpch.splits-per-node=4"
                  },
                  "commonLabels": {},
                  "coordinator": {
                    "additionalProperties": "",
                    "affinity": {},
                    "deploymentAnnotations": {},
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=true\nnode-scheduler.include-coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://localhost:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "resources": {
                      "cpu": 16,
                      "memory": "60Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "tolerations": []
                  },
                  "environment": null,
                  "expose": {
                    "clusterIp": {
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      }
                    },
                    "ingress": {
                      "annotations": {},
                      "host": null,
                      "ingressClassName": null,
                      "ingressName": "coordinator-ingress",
                      "path": "/",
                      "pathType": "ImplementationSpecific",
                      "serviceName": "starburst",
                      "servicePort": 8080,
                      "tls": {
                        "enabled": true,
                        "secretName": null
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "nodePort": 30080,
                          "port": 8080
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "externalSecrets": {
                    "enabled": false,
                    "goDaddy": {
                      "backendType": "secretsManager"
                    },
                    "secretPrefix": "external0",
                    "type": "goDaddy"
                  },
                  "extraArguments": [],
                  "extraSecret": {
                    "file": null,
                    "name": null
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise",
                    "tag": "402-e.1"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "initImage": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise-init",
                    "tag": "1.3.2"
                  },
                  "internal": {
                    "ports": {
                      "http": {
                        "port": 8080
                      },
                      "https": {
                        "port": 8443
                      }
                    }
                  },
                  "internalTls": false,
                  "licenseManager": {
                    "cronjob": {
                      "affinity": {},
                      "nodeSelector": {},
                      "priorityClassName": null,
                      "securityContext": {},
                      "tolerations": []
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "registry.connect.redhat.com/starburst/license-manager",
                      "tag": "0.0.2"
                    },
                    "url": "https://license-api.dogfood.eng.starburstdata.net/test/license/checkout"
                  },
                  "livenessProbe": null,
                  "prometheus": {
                    "agent": {
                      "config": "/etc/starburst/telemetry/prometheus.yaml",
                      "port": 8081,
                      "version": "0.16.1"
                    },
                    "enabled": true,
                    "rules": [
                      {
                        "attrNameSnakeCase": true,
                        "name": "$1",
                        "pattern": "trino.execution<name=QueryManager><>(running_queries|queued_queries)",
                        "type": "GAUGE"
                      },
                      {
                        "name": "failed_queries",
                        "pattern": "trino.execution<name=QueryManager><>FailedQueries\\.TotalCount",
                        "type": "COUNTER"
                      }
                    ]
                  },
                  "query": {
                    "maxConcurrentQueries": 3
                  },
                  "readinessProbe": null,
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "securityContext": {},
                  "sharedSecret": null,
                  "siac": {
                    "additionalHeapSizePercentage": 15,
                    "enabled": false
                  },
                  "spilling": {
                    "enabled": false,
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "startupProbe": null,
                  "usageMetrics": {
                    "enabled": false,
                    "usageClient": {
                      "initialDelay": "1m",
                      "interval": "1m"
                    }
                  },
                  "userDatabase": {
                    "enabled": false,
                    "users": [
                      {
                        "password": "46991b33f7a75ff79213c0dc0e610610",
                        "username": "admin"
                      }
                    ]
                  },
                  "worker": {
                    "additionalProperties": "",
                    "affinity": {},
                    "autoscaling": {
                      "enabled": false,
                      "maxReplicas": 100,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "deploymentAnnotations": {},
                    "deploymentTerminationGracePeriodSeconds": 300,
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://coordinator:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "kedaScaler": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.connect.redhat.com/starburst/keda-trino-scaler",
                        "tag": "0.1.7"
                      },
                      "port": 8021,
                      "scaledObjectSpec": {
                        "advanced": {},
                        "cooldownPeriod": 300,
                        "idleReplicaCount": 0,
                        "maxReplicaCount": 100,
                        "minReplicaCount": 1,
                        "pollingInterval": 30,
                        "scaleTargetRef": {
                          "name": "worker"
                        },
                        "triggers": [
                          {
                            "metadata": {
                              "numberOfQueriesPerWorker": "10",
                              "scaleInToIdleReplicaCountIfNoQueuedQueriesLeft": "true",
                              "scaleMethod": "query_queue"
                            },
                            "type": "external"
                          }
                        ]
                      }
                    },
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "replicas": 2,
                    "resources": {
                      "cpu": 16,
                      "memory": "100Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "starburstWorkerShutdownGracePeriodSeconds": 120,
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstHive",
                "metadata": {
                  "name": "starbursthive-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "affinity": {},
                  "commonLabels": {},
                  "database": {
                    "external": {
                      "driver": null,
                      "jdbcUrl": null,
                      "password": null,
                      "user": null
                    },
                    "internal": {
                      "databaseName": "hive",
                      "databasePassword": "HivePass1234",
                      "databaseUser": "hive",
                      "driver": "org.postgresql.Driver",
                      "env": [],
                      "envFrom": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.redhat.io/rhel8/postgresql-10",
                        "tag": "1"
                      },
                      "port": 5432,
                      "resources": {
                        "limits": {
                          "cpu": 2,
                          "memory": "1Gi"
                        },
                        "requests": {
                          "cpu": 2,
                          "memory": "1Gi"
                        }
                      },
                      "volume": {
                        "emptyDir": {}
                      }
                    },
                    "type": "internal"
                  },
                  "deploymentAnnotations": {},
                  "env": [],
                  "envFrom": [],
                  "expose": {
                    "clusterIp": {
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "nodePort": 30083,
                          "port": 9083
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "extraArguments": [],
                  "hdfs": {
                    "hadoopUserName": null
                  },
                  "heapSizePercentage": 85,
                  "hiveMetastoreStorageAuthorization": true,
                  "hiveMetastoreWarehouseDir": null,
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/hive",
                    "tag": "3.1.2-e.22"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "nodeSelector": {},
                  "objectStorage": {
                    "adl": {
                      "oauth2": {
                        "clientId": null,
                        "credential": null,
                        "refreshUrl": null
                      }
                    },
                    "awsS3": {
                      "accessKey": null,
                      "endpoint": null,
                      "pathStyleAccess": false,
                      "region": null,
                      "secretKey": null
                    },
                    "azure": {
                      "abfs": {
                        "accessKey": {
                          "accessKey": null,
                          "storageAccount": null
                        },
                        "authType": "accessKey",
                        "oauth": {
                          "clientId": null,
                          "endpoint": null,
                          "secret": null
                        }
                      },
                      "wasb": {
                        "accessKey": null,
                        "storageAccount": null
                      }
                    },
                    "gs": {
                      "cloudKeyFileSecret": null
                    }
                  },
                  "podAnnotations": {},
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "resources": {
                    "limits": {
                      "cpu": 1,
                      "memory": "1Gi"
                    },
                    "requests": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "securityContext": {},
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
          description: Deploys StarburstEnterprise Helm Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-paygo-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-paygo-rhmp/support?utm_source=openshift_console
          marketplace.redhat.com/meterDefinition: |-
            {
              "apiVersion": "marketplace.redhat.com/v1beta1",
              "kind": "MeterDefinition",
              "metadata": {
                "name": "sep-vcpu-usage"
              },
              "spec": {
                "group": "charts.starburstdata.com",
                "kind": "StarburstEnterprise",
                "resourceFilters": [
                  {
                    "namespace": {
                      "useOperatorGroup": true
                    },
                    "ownerCRD": {
                      "apiVersion": "charts.starburstdata.com/v1alpha1",
                      "kind": "StarburstEnterprise"
                    },
                    "workloadType": "Pod"
                  }
                ],
                "meters": [
                  {
                    "aggregation": "sum",
                    "period": "1h",
                    "metricId": "container_cpu_usage_core_avg",
                    "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=~\"coordinator|worker\"}[5m])*100",
                    "workloadType": "Pod"
                  }
                ]
              }
            }
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Starburst Data
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: StarburstEnterprise
            name: starburstenterprises.charts.starburstdata.com
            version: v1alpha1
          - kind: StarburstHive
            name: starbursthives.charts.starburstdata.com
            version: v1alpha1
        description: CRs for deploying starburst enterprise and hive
        displayName: Starburst Enterprise Pay as you go Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - starburst
        - trino
        - sep
        - bigdata
        - data
        links:
        - name: Starburst Enterprise Helm Operator
          url: https://www.starburstdata.com
        maintainers:
        - email: hello@starburstdata.com
          name: Starburst Data, Inc.
        maturity: alpha
        provider:
          name: Starburst Data, Inc.
          url: https://www.starburstdata.com
        relatedImages:
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
        version: 402.1.0
      entries:
      - name: starburst-enterprise-helm-operator-paygo.v402.1.0
        version: 402.1.0
      name: alpha
    defaultChannel: alpha
    packageName: starburst-enterprise-helm-operator-paygo-rhmp
    provider:
      name: Starburst Data, Inc.
      url: https://www.starburstdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: redhat-oadp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oadp-operator.v1.0.19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "restic": {
                      "enable": true
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws",
                        "kubevirt"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ResticRepository",
                "metadata": {
                  "name": "resticrepository",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/oadp/oadp-rhel8-operator@sha256:caf87a9832c153833feb34affc2b797ad0e643362448742b0405d457e5f0ba41
          createdAt: "2025-04-14T13:25:43Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <1.0.19'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/oadp/oadp-mustgather-rhel8@sha256:af4c6af48f38510c7d52f678798a9affea2211dc488ed18f83a5d005b7ef4b6b
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is a Velero resource that represents the capture of
              Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Data Protection Applications on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://github.com/openshift/oadp-operator
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/oadp/oadp-velero-plugin-for-microsoft-azure-rhel8@sha256:3e81b5e55f96504ba76d9d47bbeec50c7b1fd90b7702fc31b1dba248050ce9b4
        - registry.redhat.io/oadp/oadp-velero-plugin-for-gcp-rhel8@sha256:05a164c27da94b374ce03451c851a607a3cd436b73b9e5b70723c6dc8cacb3d8
        - registry.redhat.io/oadp/oadp-mustgather-rhel8@sha256:af4c6af48f38510c7d52f678798a9affea2211dc488ed18f83a5d005b7ef4b6b
        - registry.redhat.io/oadp/oadp-rhel8-operator@sha256:caf87a9832c153833feb34affc2b797ad0e643362448742b0405d457e5f0ba41
        - registry.redhat.io/oadp/oadp-registry-rhel8@sha256:0886b3b4ce0295080bf2c5ee95f2b7b15a4483bcd94cb739e7e7154d21150397
        - registry.redhat.io/oadp/oadp-velero-restic-restore-helper-rhel8@sha256:5d288335b5ad114b9eb61410befb4a9d1f84f0ba0c01558098bf4567fda9a8c5
        - registry.redhat.io/oadp/oadp-velero-plugin-rhel8@sha256:694603b9c1fd5eec939b28f8727fd7c9c0d380620c5081006f2655e98358464f
        - registry.redhat.io/oadp/oadp-velero-plugin-for-csi-rhel8@sha256:05ec16bdf1f0fadc4ddad244b86934f73569d2e66cc19f03b3553e257441a662
        - registry.redhat.io/oadp/oadp-velero-rhel8@sha256:39dec8b9ccdcabeaab3e56e14b79cfed4d8574cb172a886d017813a437e4ffff
        - registry.redhat.io/oadp/oadp-velero-plugin-for-aws-rhel8@sha256:28045538fff9e5051cd01e56876971cf9401a448e1dd63bff17fbb900629f22f
        - registry.redhat.io/oadp/oadp-kubevirt-velero-plugin-rhel8@sha256:d862a1b0420fcfb91403f90047b00035efba112af6bc5c50adb723014a620767
        version: 1.0.19
      entries:
      - name: oadp-operator.v1.0.19
        version: 1.0.19
      - name: oadp-operator.v1.0.18
        version: 1.0.18
      - name: oadp-operator.v1.0.17
        version: 1.0.17
      - name: oadp-operator.v1.0.16
        version: 1.0.16
      - name: oadp-operator.v1.0.15
        version: 1.0.15
      - name: oadp-operator.v1.0.14
        version: 1.0.14
      - name: oadp-operator.v1.0.13
        version: 1.0.13
      - name: oadp-operator.v1.0.12
        version: 1.0.12
      - name: oadp-operator.v1.0.11
        version: 1.0.11
      - name: oadp-operator.v1.0.10
        version: 1.0.10
      - name: oadp-operator.v1.0.9
        version: 1.0.9
      - name: oadp-operator.v1.0.8
        version: 1.0.8
      - name: oadp-operator.v1.0.7
        version: 1.0.7
      - name: oadp-operator.v1.0.6
        version: 1.0.6
      - name: oadp-operator.v1.0.5
        version: 1.0.5
      - name: oadp-operator.v1.0.4
        version: 1.0.4
      - name: oadp-operator.v1.0.3
        version: 1.0.3
      - name: oadp-operator.v1.0.2
        version: 1.0.2
      - name: oadp-operator.v1.0.1
        version: 1.0.1
      - name: oadp-operator.v1.0.0
        version: 1.0.0
      name: stable-1.0
    - currentCSV: oadp-operator.v1.4.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "nodeAgent": {
                      "enable": true,
                      "uploaderType": "restic"
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws",
                        "kubevirt"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: |
            Cloud Provider,Developer Tools,Modernization & Migration,OpenShift Optional,Storage
          certified: "false"
          containerImage: registry.redhat.io/oadp/oadp-rhel9-operator@sha256:ee010fb34448b772690a8a98808aa25aa1578a81061618f948b1d16b5d63d333
          createdAt: "2025-07-10T04:59:11Z"
          description: OADP (OpenShift API for Data Protection) operator sets up and
            installs Data Protection Applications on the OpenShift platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <1.4.5'
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/oadp/oadp-mustgather-rhel9@sha256:0cee6fecf052bae3ede6bb6f032b25c0c7cd63ced1e26328f2b02144f24fbb12
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A backup repository is an indicator of a connection from
              the restic/kopia server to the backupstoragelocation.
            displayName: BackupRepository
            kind: BackupRepository
            name: backuprepositories.velero.io
            version: v1
          - description: Backup is a Velero resource that represents the capture of
              Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - description: The CloudStorage API automates the creation of a bucket for
              object storage.
            displayName: CloudStorage
            kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataDownload represents a data download of a volume snapshot.
              There is one DataDownload created per volume to be restored.
            displayName: DataDownload
            kind: DataDownload
            name: datadownloads.velero.io
            version: v2alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DataUpload acts as the protocol between data mover plugins
              and data mover controller for the datamover backup operation
            displayName: DataUpload
            kind: DataUpload
            name: datauploads.velero.io
            version: v2alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: |
          **OpenShift API for Data Protection (OADP)** operator sets up and installs
          Velero on the OpenShift platform, allowing users to backup and restore
          applications.

          Backup and restore Kubernetes resources and internal images, at the
          granularity of a namespace, using a version of Velero appropriate for
          the installed version of OADP.

          OADP backs up Kubernetes objects and internal images by saving them as
          an archive file on object storage. OADP backs up persistent volumes
          (PVs) by creating snapshots with the native cloud snapshot API or with
          the Container Storage Interface (CSI). For cloud providers that do not
          support snapshots, OADP backs up resources and PV data with Restic or Kopia.

          * [Installing OADP for application backup and restore][install-doc]

          * [Installing OADP on a ROSA cluster and using STS, please follow the Getting Started Steps 1-3 in order to obtain the role ARN needed for using the standardized STS configuration flow via OLM][rosa-sts-doc]

          * [Frequently Asked Questions][faq-doc]

          [install-doc]: https://docs.openshift.com/container-platform/latest/backup_and_restore/application_backup_and_restore/installing/about-installing-oadp.html
          [rosa-sts-doc]: https://mobb.ninja/docs/misc/oadp/rosa-sts/
          [faq-doc]: https://access.redhat.com/articles/5456281
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - disaster backup
        - disaster recovery
        - disaster restore
        - disaster
        - backup
        - restore
        - backup automation
        - recovery
        - data protection
        - data management
        - data
        - protection
        - management
        - application resilience
        - resilience
        - velero
        - openshift
        - oadp
        - cloud-native
        - replication
        - kopia
        - restic
        - network file system
        - nfs
        - simple cloud storage
        - s3
        - minio
        - cloud storage
        - data foundation
        - ceph
        - csi
        - container storage interface
        links:
        - name: OADP Operator
          url: https://github.com/openshift/oadp-operator
        maintainers:
        - email: oadp-maintainers@redhat.com
          name: OADP Operator maintainers
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/oadp/oadp-velero-plugin-for-legacy-aws-rhel9@sha256:ab5c8d33684a361fa3bb10a006ef489be92cc955ad3d173bbfa77949dcc42aa8
        - registry.redhat.io/oadp/oadp-kubevirt-velero-plugin-rhel9@sha256:c3b32cb9f5b82021c08cd70467e68aad66746ef211d73612ecfe0af625706267
        - registry.redhat.io/oadp/oadp-velero-plugin-for-aws-rhel9@sha256:79d39186cae5696949813a0ea6c46eeededad29e6bde171d7a9b0011c993bb26
        - registry.redhat.io/oadp/oadp-velero-plugin-for-microsoft-azure-rhel9@sha256:9f05592bf0e5b76322ca582bdede387f9f4aabbbdc436549156c05d652c951d0
        - registry.redhat.io/oadp/oadp-mustgather-rhel9@sha256:0cee6fecf052bae3ede6bb6f032b25c0c7cd63ced1e26328f2b02144f24fbb12
        - registry.redhat.io/oadp/oadp-rhel9-operator@sha256:ee010fb34448b772690a8a98808aa25aa1578a81061618f948b1d16b5d63d333
        - registry.redhat.io/oadp/oadp-velero-rhel9@sha256:17160f99940aa355dc91ab3d16ba6bf91faade6423434008cf8f6e5023b1e413
        - registry.redhat.io/oadp/oadp-velero-plugin-for-gcp-rhel9@sha256:00800ead11ea308794b1536d964f8a043e13ede9657cabb23afa42cba7cf2ce6
        - registry.redhat.io/oadp/oadp-velero-plugin-rhel9@sha256:3ee9d3e799cedc28c400d364b3a8790c7e88acd5f603b09ff512a22d91e265b3
        - registry.redhat.io/oadp/oadp-velero-restic-restore-helper-rhel9@sha256:e0647defc7c428f9a04a5f6d4532dc86e57b92f31b4f907253ae14af05c5993f
        version: 1.4.5
      entries:
      - name: oadp-operator.v1.4.5
        version: 1.4.5
      - name: oadp-operator.v1.4.4
        version: 1.4.4
      - name: oadp-operator.v1.4.3
        version: 1.4.3
      - name: oadp-operator.v1.4.2
        version: 1.4.2
      - name: oadp-operator.v1.4.1
        version: 1.4.1
      - name: oadp-operator.v1.4.0
        version: 1.4.0
      name: stable-1.4
    defaultChannel: stable-1.4
    packageName: redhat-oadp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Oracle
      provider-url: https://www.oracle.com/cloud/
    name: oci-ccm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oci-ccm-operator.v0.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ccm.oraclecloud.com/v1alpha1",
                "kind": "OCICloudManager",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "oci-ccm-operator",
                    "app.kubernetes.io/instance": "ocicloudmanager",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ocicloudmanager",
                    "app.kubernetes.io/part-of": "oci-ccm-operator"
                  },
                  "name": "ocicloudmanager"
                },
                "spec": {
                  "compartment": "CompartmentId",
                  "driverNamespace": "kube-system",
                  "loadBalancer": {
                    "securityList": "SecurityListId",
                    "securityListManagementMode": "Frontend",
                    "subnet": "SubnetId"
                  },
                  "vcn": "VcnId"
                }
              },
              {
                "apiVersion": "ccm.oraclecloud.com/v1alpha1",
                "kind": "OCICsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "oci-ccm-operator",
                    "app.kubernetes.io/instance": "ocicsidriver",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ocicsidriver",
                    "app.kubernetes.io/part-of": "oci-ccm-operator"
                  },
                  "name": "ocicsidriver"
                },
                "spec": {
                  "compartment": "CompartmentId",
                  "driverNamespace": "kube-system",
                  "vcn": "VcnId"
                }
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider, Storage
          containerImage: phx.ocir.io/ax1eoqvpvbtg/oci-ccm-operator:0.0.9
          createdAt: "2023-07-28T23:42:18Z"
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Oracle
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: OCICloudManager
            name: ocicloudmanagers.ccm.oraclecloud.com
            version: v1alpha1
          - kind: OCICsiDriver
            name: ocicsidrivers.ccm.oraclecloud.com
            version: v1alpha1
        description: OCI Kubernetes Cloud Controller Manager implementation (or out-of-tree
          cloud-provider) for Oracle Cloud Infrastructure (OCI).
        displayName: oci-ccm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OCI
        - CCM
        - OpenShift
        - Kubernetes
        links:
        - name: Oracle
          url: https://www.oracle.com/cloud/
        - name: OCI Cloud Controller Manager (CCM)
          url: https://github.com/oracle/oci-cloud-controller-manager
        maintainers:
        - email: david.d.campbell@oracle.com
          name: David Campbell
        - email: alex.blinov@oracle.com
          name: Alex Blinov
        maturity: alpha
        provider:
          name: Oracle
          url: https://www.oracle.com/cloud/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - phx.ocir.io/ax1eoqvpvbtg/oci-ccm-operator:0.0.9
        version: 0.0.9
      entries:
      - name: oci-ccm-operator.v0.0.9
        version: 0.0.9
      name: alpha
    defaultChannel: alpha
    packageName: oci-ccm-operator
    provider:
      name: Oracle
      url: https://www.oracle.com/cloud/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: loki-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-operator.v5.9.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:6728046841367066216b304ed06249b80c57c74151c6a75c9df04a6f7f76e03d
          createdAt: "2025-01-30T17:34:53Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.7.0-0 <5.9.15'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:0c3943fa009f0c649f6d592b0ac410653c8c44b4a90e41fabdfa5f3f7db1e43d
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:6728046841367066216b304ed06249b80c57c74151c6a75c9df04a6f7f76e03d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:2e164e377d4fba98e7f63f43f25b6b9dad441d2e2914c2bf02f3525d5c9aa483
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:542b28fd51194b46efc4e5af10c3241b52db4c71f0d5d1b1f184eb39764e6c53
        version: 5.9.15
      entries:
      - name: loki-operator.v5.9.15
        version: 5.9.15
      - name: loki-operator.v5.9.14
        version: 5.9.14
      - name: loki-operator.v5.9.13
        version: 5.9.13
      - name: loki-operator.v5.9.12
        version: 5.9.12
      - name: loki-operator.v5.9.11
        version: 5.9.11
      - name: loki-operator.v5.9.10
        version: 5.9.10
      - name: loki-operator.v5.9.9
        version: 5.9.9
      - name: loki-operator.v5.9.8
        version: 5.9.8
      - name: loki-operator.v5.9.7
        version: 5.9.7
      - name: loki-operator.v5.9.6
        version: 5.9.6
      - name: loki-operator.v5.9.5
        version: 5.9.5
      - name: loki-operator.v5.9.4
        version: 5.9.4
      - name: loki-operator.v5.9.3
        version: 5.9.3
      - name: loki-operator.v5.9.2
        version: 5.9.2
      - name: loki-operator.v5.9.1
        version: 5.9.1
      - name: loki-operator.v5.9.0
        version: 5.9.0
      name: stable
    - currentCSV: loki-operator.v5.9.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:6728046841367066216b304ed06249b80c57c74151c6a75c9df04a6f7f76e03d
          createdAt: "2025-01-30T17:34:53Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.7.0-0 <5.9.15'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:2e164e377d4fba98e7f63f43f25b6b9dad441d2e2914c2bf02f3525d5c9aa483
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:542b28fd51194b46efc4e5af10c3241b52db4c71f0d5d1b1f184eb39764e6c53
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:0c3943fa009f0c649f6d592b0ac410653c8c44b4a90e41fabdfa5f3f7db1e43d
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:6728046841367066216b304ed06249b80c57c74151c6a75c9df04a6f7f76e03d
        version: 5.9.15
      entries:
      - name: loki-operator.v5.9.15
        version: 5.9.15
      - name: loki-operator.v5.9.14
        version: 5.9.14
      - name: loki-operator.v5.9.13
        version: 5.9.13
      - name: loki-operator.v5.9.12
        version: 5.9.12
      - name: loki-operator.v5.9.11
        version: 5.9.11
      - name: loki-operator.v5.9.10
        version: 5.9.10
      - name: loki-operator.v5.9.9
        version: 5.9.9
      - name: loki-operator.v5.9.8
        version: 5.9.8
      - name: loki-operator.v5.9.7
        version: 5.9.7
      - name: loki-operator.v5.9.6
        version: 5.9.6
      - name: loki-operator.v5.9.5
        version: 5.9.5
      - name: loki-operator.v5.9.4
        version: 5.9.4
      - name: loki-operator.v5.9.3
        version: 5.9.3
      - name: loki-operator.v5.9.2
        version: 5.9.2
      - name: loki-operator.v5.9.1
        version: 5.9.1
      - name: loki-operator.v5.9.0
        version: 5.9.0
      name: stable-5.9
    - currentCSV: loki-operator.v6.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.14-v4.17
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:1cb1d2cad8632791529eb754938612f5e5ed373284e42f309d603059ab7ec7c5
          createdAt: "2025-03-12T12:05:37Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=5.8.0-0 <6.0.10'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.0
          operators.operatorframework.io.bundle.channels.v1: stable-6.0
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:9aa8d8e4a7fd7ddad5d842a22b09aecdbdfa31ca5eb8c4b17f0818636af5fbf8
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:69e69ab2628742774d12dd9dde50607625efdff50ea4671f2859692fbee5ecb4
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:98d2cf3437faa2400c01edb3c694816e0214c8f9dca0a0b0cdf5882a720248f3
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:1cb1d2cad8632791529eb754938612f5e5ed373284e42f309d603059ab7ec7c5
        version: 6.0.10
      entries:
      - name: loki-operator.v6.0.10
        version: 6.0.10
      - name: loki-operator.v6.0.9
        version: 6.0.9
      - name: loki-operator.v6.0.8
        version: 6.0.8
      - name: loki-operator.v6.0.7
        version: 6.0.7
      - name: loki-operator.v6.0.6
        version: 6.0.6
      - name: loki-operator.v6.0.5
        version: 6.0.5
      - name: loki-operator.v6.0.4
        version: 6.0.4
      - name: loki-operator.v6.0.3
        version: 6.0.3
      - name: loki-operator.v6.0.2
        version: 6.0.2
      - name: loki-operator.v6.0.1
        version: 6.0.1
      - name: loki-operator.v6.0.0
        version: 6.0.0
      name: stable-6.0
    - currentCSV: loki-operator.v6.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.15-v4.18
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:0bde3c1a0b667a693f443688ed3926eef1705c45ecf4cce3a3dedbddca7a2784
          createdAt: "2025-03-12T11:19:10Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=5.9.0-0 <6.1.8'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.1
          operators.operatorframework.io.bundle.channels.v1: stable-6.1
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:0bde3c1a0b667a693f443688ed3926eef1705c45ecf4cce3a3dedbddca7a2784
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:731f92c146f299bc3662b309a3068bde18fad51f65a7d29063e582b61e07042f
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:c34676932ddd14db213dbf42a1d341f51692f3f47f1b751d06187e292c99a170
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b94833f55644c289fe2c1297b68a4f9931baf977587c6a2c4c7a3e484c704f64
        version: 6.1.8
      entries:
      - name: loki-operator.v6.1.8
        version: 6.1.8
      - name: loki-operator.v6.1.7
        version: 6.1.7
      - name: loki-operator.v6.1.6
        version: 6.1.6
      - name: loki-operator.v6.1.5
        version: 6.1.5
      - name: loki-operator.v6.1.4
        version: 6.1.4
      - name: loki-operator.v6.1.3
        version: 6.1.3
      - name: loki-operator.v6.1.2
        version: 6.1.2
      - name: loki-operator.v6.1.1
        version: 6.1.1
      - name: loki-operator.v6.1.0
        version: 6.1.0
      name: stable-6.1
    - currentCSV: loki-operator.v6.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.16-v4.19
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f401be018c5ba35e8ea1e68831cc20067f0f4d1be031d05b222548c1bdfb5f7e
          createdAt: "2025-02-18T18:01:39Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=6.0.0-0 <6.2.5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.2
          operators.operatorframework.io.bundle.channels.v1: stable-6.2
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f401be018c5ba35e8ea1e68831cc20067f0f4d1be031d05b222548c1bdfb5f7e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:99d1e7c5c1f0609b33de576884782550f969b28dc622810e6661d4c60a664215
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:8955d8186ae98095623f2083b6d5389d3ac9f2612fcfcf695835219c3e1216c4
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b8c5b46a94a6c85c9d382eeeb65a1d2b4014a7dc765e9abfba706c92857462dc
        version: 6.2.5
      entries:
      - name: loki-operator.v6.2.5
        version: 6.2.5
      - name: loki-operator.v6.2.4
        version: 6.2.4
      - name: loki-operator.v6.2.3
        version: 6.2.3
      - name: loki-operator.v6.2.2
        version: 6.2.2
      - name: loki-operator.v6.2.1
        version: 6.2.1
      - name: loki-operator.v6.2.0
        version: 6.2.0
      name: stable-6.2
    defaultChannel: stable-6.2
    packageName: loki-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-route53-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-route53-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/route53-controller:1.1.1
          createdAt: "2025-09-19T19:15:31Z"
          description: AWS Route53 controller is a service controller for managing
            Route53 resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HealthCheck represents the state of an AWS route53 HealthCheck
              resource.
            displayName: HealthCheck
            kind: HealthCheck
            name: healthchecks.route53.services.k8s.aws
            version: v1alpha1
          - description: HostedZone represents the state of an AWS route53 HostedZone
              resource.
            displayName: HostedZone
            kind: HostedZone
            name: hostedzones.route53.services.k8s.aws
            version: v1alpha1
          - description: RecordSet represents the state of an AWS route53 RecordSet
              resource.
            displayName: RecordSet
            kind: RecordSet
            name: recordsets.route53.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Route53 resources in AWS from within your Kubernetes cluster.

          **About Amazon Route53**

          Amazon Route 53 is a highly available and scalable Domain Name System (DNS) web service. You can use Route 53 to perform three main functions in any combination: domain registration, DNS routing, and health checking.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Route53
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - route53
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Route53 Developer Resources
          url: https://docs.aws.amazon.com/route53/index.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/route53-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-route53-controller.v1.1.1
        version: 1.1.1
      - name: ack-route53-controller.v1.1.0
        version: 1.1.0
      - name: ack-route53-controller.v1.0.6
        version: 1.0.6
      - name: ack-route53-controller.v1.0.5
        version: 1.0.5
      - name: ack-route53-controller.v1.0.3
        version: 1.0.3
      - name: ack-route53-controller.v1.0.2
        version: 1.0.2
      - name: ack-route53-controller.v0.1.0
        version: 0.1.0
      - name: ack-route53-controller.v0.0.28
        version: 0.0.28
      - name: ack-route53-controller.v0.0.27
        version: 0.0.27
      - name: ack-route53-controller.v0.0.26
        version: 0.0.26
      - name: ack-route53-controller.v0.0.25
        version: 0.0.25
      - name: ack-route53-controller.v0.0.24
        version: 0.0.24
      - name: ack-route53-controller.v0.0.23
        version: 0.0.23
      - name: ack-route53-controller.v0.0.22
        version: 0.0.22
      - name: ack-route53-controller.v0.0.20
        version: 0.0.20
      - name: ack-route53-controller.v0.0.19
        version: 0.0.19
      - name: ack-route53-controller.v0.0.18
        version: 0.0.18
      - name: ack-route53-controller.v0.0.17
        version: 0.0.17
      - name: ack-route53-controller.v0.0.16
        version: 0.0.16
      - name: ack-route53-controller.v0.0.15
        version: 0.0.15
      - name: ack-route53-controller.v0.0.14
        version: 0.0.14
      - name: ack-route53-controller.v0.0.13
        version: 0.0.13
      - name: ack-route53-controller.v0.0.9
        version: 0.0.9
      - name: ack-route53-controller.v0.0.6
        version: 0.0.6
      - name: ack-route53-controller.v0.0.4
        version: 0.0.4
      - name: ack-route53-controller.v0.0.3
        version: 0.0.3
      - name: ack-route53-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-route53-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://www.medik8s.io/
    name: self-node-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: self-node-remediation.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/self-node-remediation-operator:v0.9.0
          createdAt: "2024-07-25 14:40:52"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.9.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Medik8s
          url: https://www.medik8s.io/
        relatedImages:
        - quay.io/medik8s/self-node-remediation-operator:v0.9.0
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        version: 0.9.0
      entries:
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      - name: self-node-remediation.v0.7.0
        version: 0.7.0
      - name: self-node-remediation.v0.6.1
        version: 0.6.1
      - name: self-node-remediation.v0.6.0
        version: 0.6.0
      - name: self-node-remediation.v0.5.0
        version: 0.5.0
      - name: self-node-remediation.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: self-node-remediation
    provider:
      name: Medik8s
      url: https://www.medik8s.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudfront-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudfront-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "CachePolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "Distribution",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudfront-controller:1.2.1
          createdAt: "2025-09-19T21:12:14Z"
          description: AWS CloudFront controller is a service controller for managing
            CloudFront resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CachePolicy represents the state of an AWS cloudfront CachePolicy
              resource.
            displayName: CachePolicy
            kind: CachePolicy
            name: cachepolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: Distribution represents the state of an AWS cloudfront Distribution
              resource.
            displayName: Distribution
            kind: Distribution
            name: distributions.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: Function represents the state of an AWS cloudfront Function
              resource.
            displayName: Function
            kind: Function
            name: functions.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: OriginAccessControl represents the state of an AWS cloudfront
              OriginAccessControl resource.
            displayName: OriginAccessControl
            kind: OriginAccessControl
            name: originaccesscontrols.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: OriginRequestPolicy represents the state of an AWS cloudfront
              OriginRequestPolicy resource.
            displayName: OriginRequestPolicy
            kind: OriginRequestPolicy
            name: originrequestpolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: ResponseHeadersPolicy represents the state of an AWS cloudfront
              ResponseHeadersPolicy resource.
            displayName: ResponseHeadersPolicy
            kind: ResponseHeadersPolicy
            name: responseheaderspolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: VPCOrigin represents the state of an AWS cloudfront VPCOrigin
              resource.
            displayName: VPCOrigin
            kind: VPCOrigin
            name: vpcorigins.cloudfront.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudFront resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudFront**

          Amazon CloudFront is a web service that speeds up distribution of your static and dynamic web content, such as .html, .css, .js, and image files, to your users. CloudFront delivers your content through a worldwide network of data centers called edge locations.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon CloudFront
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudfront
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudFront Developer Resources
          url: https://aws.amazon.com/cloudfront/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudfront maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudfront-controller:1.2.1
        version: 1.2.1
      entries:
      - name: ack-cloudfront-controller.v1.2.1
        version: 1.2.1
      - name: ack-cloudfront-controller.v1.2.0
        version: 1.2.0
      - name: ack-cloudfront-controller.v1.1.6
        version: 1.1.6
      - name: ack-cloudfront-controller.v1.1.4
        version: 1.1.4
      - name: ack-cloudfront-controller.v1.1.3
        version: 1.1.3
      - name: ack-cloudfront-controller.v1.1.2
        version: 1.1.2
      - name: ack-cloudfront-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudfront-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudfront-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudfront-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudfront-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudfront-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudfront-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudfront-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudfront-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudfront-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudfront-controller.v0.0.15
        version: 0.0.15
      - name: ack-cloudfront-controller.v0.0.14
        version: 0.0.14
      - name: ack-cloudfront-controller.v0.0.13
        version: 0.0.13
      - name: ack-cloudfront-controller.v0.0.12
        version: 0.0.12
      - name: ack-cloudfront-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudfront-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudfront-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudfront-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudfront-controller.v0.0.6
        version: 0.0.6
      - name: ack-cloudfront-controller.v0.0.5
        version: 0.0.5
      - name: ack-cloudfront-controller.v0.0.4
        version: 0.0.4
      - name: ack-cloudfront-controller.v0.0.3
        version: 0.0.3
      - name: ack-cloudfront-controller.v0.0.2
        version: 0.0.2
      - name: ack-cloudfront-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudfront-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/openshift/lightspeed-service
    name: lightspeed-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lightspeed-operator.v1.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ols.openshift.io/v1alpha1",
                "kind": "OLSConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightspeed-operator",
                    "app.kubernetes.io/instance": "olsconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "olsconfig",
                    "app.kubernetes.io/part-of": "lightspeed-operator"
                  },
                  "name": "cluster"
                },
                "spec": {
                  "llm": {
                    "providers": [
                      {
                        "credentialsSecretRef": {
                          "name": "credentials"
                        },
                        "models": [
                          {
                            "name": "gpt-3.5-turbo-1106"
                          }
                        ],
                        "name": "OpenAI"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-08-28T16:37:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lightspeed
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Kubernetes Engine", "OpenShift Virtualization
            Engine"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lightspeed-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Red Hat OpenShift Lightspeed instance. OLSConfig is the Schema
              for the olsconfigs API
            displayName: OLSConfig
            kind: OLSConfig
            name: olsconfigs.ols.openshift.io
            version: v1alpha1
        description: |-
          OpenShift Lightspeed Operator provides generative AI-based virtual assistant which integrates into the OpenShift web console. OpenShift Lightspeed can answer natural language questions related to OpenShift Container Platform.

          To configure the OpenShift Lightspeed Operator, you need a Large Language Model (LLM) . You can host the LLM using RHELAI or RHOAI or use one of the models from the following Software-as-a-Service providers:

          - OpenAI
          - Microsoft Azure OpenAI
          - IBM WatsonX

          For more information, see [About OpenShift Lightspeed](https://docs.redhat.com/en/documentation/red_hat_openshift_lightspeed/1.0tp1/html/about/ols-about-openshift-lightspeed) in the official product documentation.

          Questions, comments, concerns, or just want to talk to the OpenShift Lightspeed team? [Email us](mailto:openshift-lightspeed-contact-requests@redhat.com)!
        displayName: OpenShift Lightspeed Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ai
        - assistant
        - openshift
        - llm
        links:
        - name: Lightspeed Operator
          url: https://github.com/openshift/lightspeed-operator
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/lightspeed-service
        relatedImages:
        - registry.redhat.io/openshift-lightspeed/lightspeed-rhel9-operator@sha256:b12803642df4d9b3ad0cf800315021483f543c6a8827b3901e57bdb65f10dc48
        - registry.redhat.io/openshift-lightspeed/openshift-mcp-server-rhel9@sha256:f1e4521ff2cc076f361c04d8c9aeac09ffcedb89a2f1dbe71f85caa27472adab
        - registry.redhat.io/openshift-lightspeed/lightspeed-service-api-rhel9@sha256:ec359967a87bc7ee2c1d6a61e553acbdb0dc85d6f240a1865bf57d3ba5dc7193
        - registry.redhat.io/openshift-lightspeed/lightspeed-console-plugin-rhel9@sha256:dfc42c2651ac8793f61ff0e54334ecb006dc4a77487bc7c6878cdf47bd4ac080
        version: 1.0.5
      entries:
      - name: lightspeed-operator.v1.0.5
        version: 1.0.5
      name: alpha
    - currentCSV: lightspeed-operator.v1.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ols.openshift.io/v1alpha1",
                "kind": "OLSConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightspeed-operator",
                    "app.kubernetes.io/instance": "olsconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "olsconfig",
                    "app.kubernetes.io/part-of": "lightspeed-operator"
                  },
                  "name": "cluster"
                },
                "spec": {
                  "llm": {
                    "providers": [
                      {
                        "credentialsSecretRef": {
                          "name": "credentials"
                        },
                        "models": [
                          {
                            "name": "gpt-3.5-turbo-1106"
                          }
                        ],
                        "name": "OpenAI"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-08-28T16:37:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lightspeed
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Kubernetes Engine", "OpenShift Virtualization
            Engine"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lightspeed-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Red Hat OpenShift Lightspeed instance. OLSConfig is the Schema
              for the olsconfigs API
            displayName: OLSConfig
            kind: OLSConfig
            name: olsconfigs.ols.openshift.io
            version: v1alpha1
        description: |-
          OpenShift Lightspeed Operator provides generative AI-based virtual assistant which integrates into the OpenShift web console. OpenShift Lightspeed can answer natural language questions related to OpenShift Container Platform.

          To configure the OpenShift Lightspeed Operator, you need a Large Language Model (LLM) . You can host the LLM using RHELAI or RHOAI or use one of the models from the following Software-as-a-Service providers:

          - OpenAI
          - Microsoft Azure OpenAI
          - IBM WatsonX

          For more information, see [About OpenShift Lightspeed](https://docs.redhat.com/en/documentation/red_hat_openshift_lightspeed/1.0tp1/html/about/ols-about-openshift-lightspeed) in the official product documentation.

          Questions, comments, concerns, or just want to talk to the OpenShift Lightspeed team? [Email us](mailto:openshift-lightspeed-contact-requests@redhat.com)!
        displayName: OpenShift Lightspeed Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ai
        - assistant
        - openshift
        - llm
        links:
        - name: Lightspeed Operator
          url: https://github.com/openshift/lightspeed-operator
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/lightspeed-service
        relatedImages:
        - registry.redhat.io/openshift-lightspeed/lightspeed-service-api-rhel9@sha256:ec359967a87bc7ee2c1d6a61e553acbdb0dc85d6f240a1865bf57d3ba5dc7193
        - registry.redhat.io/openshift-lightspeed/lightspeed-console-plugin-rhel9@sha256:dfc42c2651ac8793f61ff0e54334ecb006dc4a77487bc7c6878cdf47bd4ac080
        - registry.redhat.io/openshift-lightspeed/openshift-mcp-server-rhel9@sha256:f1e4521ff2cc076f361c04d8c9aeac09ffcedb89a2f1dbe71f85caa27472adab
        - registry.redhat.io/openshift-lightspeed/lightspeed-rhel9-operator@sha256:b12803642df4d9b3ad0cf800315021483f543c6a8827b3901e57bdb65f10dc48
        version: 1.0.5
      entries:
      - name: lightspeed-operator.v1.0.5
        version: 1.0.5
      name: stable
    defaultChannel: stable
    packageName: lightspeed-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/openshift/lightspeed-service
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kubevirt-hyperconverged
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubevirt-hyperconverged-operator.v4.16.23
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:eaacb0d177fd64e2eefbcb2dc47633dadf4d54fcbe0f0f750b078ab3f4f189e1
          createdAt: "2025-09-29 21:23:00"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.14.1 <4.16.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:3e1c68d5695bab96c382661a33581e60597b59afd21660ec760d8fe5e20ff9e3
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.16.23-1,hyperconverged-cluster-operator-test-rhel9:v4.16.23-1,kubevirt-common-instancetypes-test-rhel9:v4.16.23-1,kubevirt-tekton-tasks-test-rhel9:v4.16.23-1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:e24191be577575fa9d462c87c6f51775a0cf0f5d1812d12c4c95a68ec34b34ff
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:6e4aa75bece3c2fe076e282b3f1ad8296299ec0551e379c7c88830a0ce8ed5bf
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:1be66eee321bc47dcc0e7ff4674a485b9a4552243080d796d5f3b0d417721dc4
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:28c0bdf60404fb435cfa0cfdc6d314f789417c44c3206ce8cd1e7b99031fb529
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:659abe1d740607b2cb3850c9ab29101fa5d5c821d2a09aab5994849697a504f3
        - registry.redhat.io/container-native-virtualization/mtq-operator-rhel9@sha256:9de5ded35edfbe98121bdd4404c84dcc72c8cba1742dc27df6382ef882db2183
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:0f0478e9357f8b6ace6d844e6e424db192398195a8cea2fa6398be8996d05be6
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:2843d199dc2748749d5f48fb9f90656962e011533d79dc7e0a4f3592b12aa69a
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:6b0c7cc17abd95fce45b6ec83ec291e843eedbc90dce0ecc449aec6e64562057
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:97d37927164f0ea986f209912e6247fdab37b680c214afc4bdbb7051caaf58b1
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:cc4ff488aa97e179c4df39ac5ae8085070017106adf1c5f2222b35d8f2234ea8
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:f8e2ea3988d4a77a4e809e1fd073ffa5f99393ca3eb69a2c1cb54dd4ab60e482
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:885670f2e9ad43224cbc96f34b23dd2642de00f688bade0f6b89dd3d3fb83dbd
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:251d6fb3197a7642d86193191aa69c85a45c5521d7fc509704edd46d5d4e90a7
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:fcbadecf01b81625d55f9219428baabc33ee7231b4ccf801c90dee3095907187
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:a0af0e5d4b142f28dc72c5636df5040c4ce92c6b12ab183d3a9bc7a9d2d2f418
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:c033ddd0ed30c0d25f4c2e9b30fca4e40959bb6d74029a1f452ff048b97de88a
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:7206ac7959c07f474f454ab48bcf3522ade95efb27fd9fbd3938d1387a1420af
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:468f117a1b55af59e72a9ff46499a8d0bb00884d450d57b692e2f7b557c9bf3d
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:e01871006b9e4cb1b6f477319de63b758eac672a70a298caed2032ec19986680
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:7587832b877f32621eb26e67a5c516728f293ecd98f74dd70d5f03474a5429d6
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:68b650d6779fc4d61f05d97e6d62b80fcd638721994153205eec68f8c1d999ce
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:19ca70dceb79b10ff458f0ea41faa6e5552a5f08828620ad7177efa34166a90e
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:84dbfa51755d60e34b194c1acf679ac083feb49038112212828fe8101dcec2a5
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:8462188ec7533d2105c0dc9154d8dcf1425b96d6945eef16ea774a5dff9bba1d
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:11030a8d8cd03efdf27fe3cfc9ebeadeb20851b6de1e9d8a93ba7f2bc6a77a96
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:8cf7993326ba220f650ae2fa5383d0d6cf3e2c7f3346f0a339a9d212b8df02f6
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:c18b6e315f5577cc18d46cd76595d48d4a49026a6b2f5763c3efb26919bde5eb
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:52b4f7c7a4f5c5c5d07a720c5eefa70709f520b1d939aec4cd971d8f3c9b12bc
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:ef8260d72239ad2126ca023250bfd976f5c050fe9c0727a0f7ef007446f5363b
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:18be4ef8b6c2cedab4a2e031939fc1f9cc23020bde7fccb7997e4222b673f57f
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:def80efdcc58d545aaf4b69b784a348cb36e97afd8e0124480521a2b980fa7c3
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:b031a12e73cb75e207203fd3785619e3a5a22d7d5a8549488c035f70374235be
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:14ea56d4149b72b98b6e7f5b33ba5bdd9115e7c643fd103c9df7a26b2a06700a
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:9e358366fd11dd167255abd930016aa4fc35b1b9dc69ce27316635488efc8ec4
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:6bf45534310768308958428692b29afa25f2740d0c62165d7182b0b3e9e944cf
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:eaacb0d177fd64e2eefbcb2dc47633dadf4d54fcbe0f0f750b078ab3f4f189e1
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:91b86228495f38d5528be1a8ffec96695ad72fafb76d2ead87f7bee25540d8b6
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:19b56bbb081db7198d570c3c691e142991aaa81857ee417d47903291ed0e7a1c
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:372652529b6a8705df581e7ab8acc2ec5c3cdf8fb273448052c59963abe4d984
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:7f89f2ba59309b26cf83f58ff3ba35569182097258fa543d5d88f3a8cd64ca00
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:bf1bb2fe09e028506fe7cf383368d76e358b8c59ecc74d3618bc250c1e97271c
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:a79579b18a7df9f23c584303593a9404c64df023458d55c715aeb71dcacc92a0
        - registry.redhat.io/container-native-virtualization/mtq-lock-server-rhel9@sha256:bb1dfc204e48a13985d6c2ec9cdf17f47330617725ee6dd854f0dae0f1901ba4
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:4005d165f1485eec98ca24ce41bef79cd8088002d43ffb709ec65f842a176bdc
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:3f3554ef043c80494db536b28eb03fa7043e53bb85ce426820c252fb999ce04e
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:531702869a34f76850d332efb84ccf571fafa281014afd1787497ca9de8aed37
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:a081e7e68def13b7de369d3f66469035d61fd1feef1bee4f95ba5f7b4041d960
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:e7672121156e050f05af771f75cfe26558699da8a074bf7051f710705a27d743
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:bd532e59edae5d2edac78b9249cd6a9f0779f44e25b019d824a1b9803f078bde
        - registry.redhat.io/container-native-virtualization/mtq-controller-rhel9@sha256:a774f0bd125c66b45e35e327c9c7e0afd7947d3c5ed222b3d5d223941366ba2c
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:52e76b4208ee4e7da8a0729b9bdb294729454d0f62580d64dd104b6921f464ed
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:3b57db7aaa3c61335044ec78cbd061f78ce99cdb4b75b3ce342e1a603da4bcef
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:7e19b42a648cee71c60cb69fbcf6ebca05c938942e16d23f1cd61a619cffdcb6
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:4046e8e862e0c55f09e9031b95e6a77af4d491593e39770c644287d44888b0bb
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:517c88a6be8842d4eb71ad30f99ce1afca954af1df2c1bf4a9584f4ec69fcf12
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:3e1c68d5695bab96c382661a33581e60597b59afd21660ec760d8fe5e20ff9e3
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:cca121312777d2727fe2c50a14583243ce69e6e356b499f133736787522095b2
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:ad24eb52bd65495224661d4894336d88fde342aeaab16b1084ceff6337f92059
        version: 4.16.23
      entries:
      - name: kubevirt-hyperconverged-operator.v4.16.23
        version: 4.16.23
      - name: kubevirt-hyperconverged-operator.v4.16.22
        version: 4.16.22
      - name: kubevirt-hyperconverged-operator.v4.16.21
        version: 4.16.21
      - name: kubevirt-hyperconverged-operator.v4.16.20
        version: 4.16.20
      - name: kubevirt-hyperconverged-operator.v4.16.19
        version: 4.16.19
      - name: kubevirt-hyperconverged-operator.v4.16.18
        version: 4.16.18
      - name: kubevirt-hyperconverged-operator.v4.16.17
        version: 4.16.17
      - name: kubevirt-hyperconverged-operator.v4.16.16
        version: 4.16.16
      - name: kubevirt-hyperconverged-operator.v4.16.15
        version: 4.16.15
      - name: kubevirt-hyperconverged-operator.v4.16.14
        version: 4.16.14
      - name: kubevirt-hyperconverged-operator.v4.16.13
        version: 4.16.13
      - name: kubevirt-hyperconverged-operator.v4.16.12
        version: 4.16.12
      - name: kubevirt-hyperconverged-operator.v4.16.11
        version: 4.16.11
      - name: kubevirt-hyperconverged-operator.v4.16.10
        version: 4.16.10
      - name: kubevirt-hyperconverged-operator.v4.16.9
        version: 4.16.9
      - name: kubevirt-hyperconverged-operator.v4.16.8
        version: 4.16.8
      name: candidate
    - currentCSV: kubevirt-hyperconverged-operator.v4.99.0-0.1723448771
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:5b2c64218342a5744182fc06f550bcdcf954e931275a9d1505ef0bbfbd0f8044
          createdAt: "2024-08-12 07:46:45"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.14.1 <4.16.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:d74c629e415d177069a0570fb8db9252d2b193d2ec7972c844686ffd14033f62
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.99.0-73,kubevirt-common-instancetypes-test-rhel9:v4.99.0-32,kubevirt-tekton-tasks-test-rhel9:v4.99.0-31
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:cc190d7c44b20f10ae61709fd2f5f7c0b248df73b768eb696d5a6cec90a34e59
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:41ee360ac43ee69ccee62d4ac3d2f07c8a8589e5b85c659b47018dd36c2b932f
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:b410508ecc71b0a3bf4825793efde4b07085e2243d8b3585da852fd1940c4be1
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:820a9e6bd0329df5c541ef0b67b3c7cdd9f81be33de87c5b5e5e43777bea197e
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:811e8e7e9872408794ef36229c42c6c437b1a4ace76c7134d91bb2331fe7815b
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:0769512df0fa1325d76a5f10a1a704cf60fc65f62f57b157ea53bc9ba682851f
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:8c64b526fd42a17c909bf45eb795121a81d502424a9cf11f8087118751333894
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:5f23fc8e9bc386622460c23b9644f58648d9f748f5670927a02888cd36c99a21
        - registry.redhat.io/container-native-virtualization/mtq-controller-rhel9@sha256:ad5938ebd5f07196ce5c35232a1c3b0a653301a83929ce84c3d1cb4dd1b91816
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:37e6b4fa5ff436572412ec0bdb1edc4a49e62c994e056c93578605f6b74adbcd
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:07435335fdbf6178bc00cc02b93cc0adb42784339151b2d9c31dc8da98b8bc6a
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:cffec516ab9ddc50cab77339d0cb3239ef4e7fd43bd2afdc73370aa487ea5704
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:dec430b74887e9b42d639f14c07d9f58a4881cac24e949dab98f96af150f1958
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:c61732be8024b4ede45eec8f786e51b198632ee5bfb0c2f5af84f37544584283
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:5ba79dd6bdf9e3778d3e301919b37e45cfc1eaf72fda77eba6588788829ba29b
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:c14cd037d14489321629908fc6b30248f7a20d866619177a165d72cfc1b6b7ef
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:83c03929cf11d7202e6c6cb9dd767bd4d7473f4fc1d562b79719ceb3d8382e68
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:6b7beed3cba9093b3aeb3109f23ec38c20f5568915dff774ad5e9e6bb5db3a58
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:e39732ebd3a4452fa7a678aeea5e64a7973fd72802eff193b16680c2f0129cb1
        - registry.redhat.io/container-native-virtualization/mtq-lock-server-rhel9@sha256:01cc6797800ccfb99453858ab36c0100c5ca343f1f71f975914b2437d1f757ba
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:9822f7a1590d830ccd64cc0729d2e30a2e04fb2b022f54498e1f1abff3748c2c
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:bf2ec130d1f3f32314d740037313f49f874a271e8b9a7f6cd2b93698a8423e8d
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:42a34caa18b5f5452b9d5603602ea74b1ee59a8bbb52bddf18901fecc27aed8a
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:18dfb38223647e11020e3ef1093a695454b90cc50922f80afb612212b7d13309
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:4c66a26b9f74d83a9bc60edb40ffd9d6ff2382d20500677d4567f8e141c3ddf0
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:758a9f6309a367d3f31cc03dd89356a8232a0fbd277570c01ebfa1fa30089a8d
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:2e5d5a2a70cda1f414f22b1c713d77d068c5fb16b1487a572019a9aa7130e6be
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:84b199e98e283c863d52a7b00044c20cb56ad563494adbec2229af265bf3f7c5
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:6baf67f3f410c071cc26f675d86edbbae5eaeb3410888cf622fad384b43b80ea
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:3a45cd2f3ea25f356167ce4fa20894a2a0706b4bcffa42c806f6017812053466
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:84679ffe8ce86c674818531897d122f5568d10c7e1e2ff919185d5771e5b4c39
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:ac33de246ee42bbfe735d7349d85c6df7e583e7599dea7dd7a2235bd0e59845f
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:e7145f016642b02bd651e21aa19edc7f93d0d1c2da3c6c7eed5d1174cf23619c
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:2564af054f52a8bb340e0270379c72da18658662a2a5bd26ac7bd979ddfccb56
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:4cff43cc145361c74c357055d51060ccc27e6bca58ccc3b5f5b983c7b8545114
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:09a30b59c906ddc4604d70f18210bcebef1f90f8bf965422b0caba363cd88989
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:63389195cf584a07ddbc2efec90cf6deaf737b91314b67db2b9f53dcc7acb57c
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:6193e569d4ab43806ac7b4870ef3634a5de3687527d6d190e59a9f735fa4778c
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:f4e7876fa8b27a8082efa99455d9bf7b0c479cff1e4c8a26136bf63092699c0b
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:d7e8e075a08b148ddfa91b44bb9cae1ef823756760279c4bb6d8489bc7e0eb2c
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:bf30b6c7173ebb6a4592246af16db87d7d391182f52286b559f8bad8f4b6cc5b
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:15179f0826831f051fdb867067be7c5528f01f442ae421b8782ae0cd1ed1b73d
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:4b3ac5ec2fafef0b7702a02847996f0903ec8d4262ccd03e8fa74885265ea4e9
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:d6dfd062df3d0d49f7efb928b835b9c8e0bcfb514a653ba2896a0152f24a0a21
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:185a83c994aced74546fe630cd71656f1e4112cb15be75f41dfcb1b9ed0966bd
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:b59e34cdb6214eacb9dc176a7b33125d7c1871f6b81696d59646294bfaf749e7
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:8e19d99232be4c37e857a4931627fb0aed721893e27b0d0e522fcf13bf67c1c9
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:06cc71822684c3ed094f327553451daf076d6bb97065c314da1a7d2adbbef3d2
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:4bdafc4404d5cfd844d05f32d88c262fba56dc931bfd5522b78c25d4ff253883
        - registry.redhat.io/container-native-virtualization/kubevirt-ipam-controller-rhel9@sha256:f33ac4d7477a730f273a462f171aed35dea38374d5ea7ef0bae2d7f74413a757
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:af5d2a2d4f0b6f0e1a13f1029073bb56980d8a70503b21628f18e8986e9b531c
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:5b2c64218342a5744182fc06f550bcdcf954e931275a9d1505ef0bbfbd0f8044
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:88f02cbebd6c3a2add6e0f8ca744e42bbd182f5a07d4fac59cf5cf07a4edfd45
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:fa0478fbf91248166f0eed8ec5f8dbb5b31d26031438376d9301836b6a069643
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:9bfdaf1cca6eac15b63a4ea6b7de28f2328516c5cf8bd5727935dd729e45657d
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:30c669770b0b992a3c84bb81c892e70a24fb14af187e8a4ee19cb3e0dd56ddbb
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:c5e8f44f647f593b00d3427172932d7aa694379d124209b9c66b4c1ffe3d4116
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:e6d9a6c41433360abd42d1118f5af3df9592d6d36af5f65082ae214cc8513c1b
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:d74c629e415d177069a0570fb8db9252d2b193d2ec7972c844686ffd14033f62
        - registry.redhat.io/container-native-virtualization/mtq-operator-rhel9@sha256:033bf7a6ef8ced1ac7275a9ddea8b16a10340a0fb0fb39619d89354d661f05b6
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:442e5cbcc63961642b662544032240bffb3863227324128a67f0bb6a05a1c364
        version: 4.99.0-0.1723448771
      entries:
      - name: kubevirt-hyperconverged-operator.v4.99.0-0.1723448771
        version: 4.99.0-0.1723448771
      name: dev-preview
    - currentCSV: kubevirt-hyperconverged-operator.v4.16.21
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:08f15657a1eee02927c5a6ef661e6e703d0064e091751bd2d8fbad5f118d9672
          createdAt: "2025-09-20 19:53:21"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.14.1 <4.16.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:88863a8dd2145bf17f7c5378c81e7c0be7a01dc47ee75bd7e5203a8fed8895b9
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.16.21-3,hyperconverged-cluster-operator-test-rhel9:v4.16.21-1,kubevirt-common-instancetypes-test-rhel9:v4.16.21-1,kubevirt-tekton-tasks-test-rhel9:v4.16.21-1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:9afe1ec9f776854caae38b89198a82e8e47a45d6c08f8dbd65a06c85f741c6a0
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:611d60b6ea4285fbc8a9f945fa5d06c09679a8f7ec1d70ec18edae7eec1a6bb3
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:299825fa33ab3f491cf954dd2e4ba23fdba2a96723b9ac3d98b3264b0b9fbb28
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:f6b075a56380143a7892e1fe5fdb212f8aa1c974f6a2d7e7e9b940789e44c010
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:d346688e02d03ddd9918f7a344d6eb3d5ddac3f1802dc82daf72bc56e7a9d496
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:406681aff7ea0a17941449593f839418edffdb073e2afffd53174f0df6292f94
        - registry.redhat.io/container-native-virtualization/mtq-controller-rhel9@sha256:253ef56b2867d33b3a475a456878b4a60f56473122eeed8c169568a9724ed30c
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:e54d87e6735151ed3c5295b23ca2f7ec666afb6421923636ef836bca1cf883d6
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:6385b346b3dd8f2bb3284871d326c851cf743a5e3fc3c49a1e51a1d43b0eafb4
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:8cd1052b9f69d0a0d8db41e90414756e83cbd07d7bd03c2faa85c41393d6539c
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:b80e68822fb0b5e78a995a97e02fe2ec448d67a5aa2224497fdf12b812182a0b
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:51a1fb0eb77e4f5c8bdfaf3b5074369b93ef63f5a7af62e22ed0789731f8dc73
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:361a48df213d52ced2a9963c2a8bde72a0981f1c497235fb918ea8c065580eeb
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:355e0946cfa963aeb5b6c3e8d476a9c3aee202baf77b95ce3f69c8e4b9e73836
        - registry.redhat.io/container-native-virtualization/mtq-lock-server-rhel9@sha256:c447251fea804340047bbaff885ad56be7a0d2612af1de26765f6cfc37da25e6
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:3daa460ed347e6eaab290b0a7fb0a08da6f210fb8be88af33f6e256511af80b6
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:dcc73f518ca7557cbd84b236367b3ecb7baab405268d54543783b4b772e14568
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:f7e5e04adb0467eb0aa4691a68b0d6485db9a52f99494097bf5b598e08f2edfb
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:2a109f94282b76ca0ff929adc35c93c0a98e34b43aba248d2975957fa149133f
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:cca121312777d2727fe2c50a14583243ce69e6e356b499f133736787522095b2
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:a0af0e5d4b142f28dc72c5636df5040c4ce92c6b12ab183d3a9bc7a9d2d2f418
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:88863a8dd2145bf17f7c5378c81e7c0be7a01dc47ee75bd7e5203a8fed8895b9
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:8cf7993326ba220f650ae2fa5383d0d6cf3e2c7f3346f0a339a9d212b8df02f6
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:edf12760e7fe4c8d37c22e172c0da270c219b7ff16d4ebb6b62729545309f6fa
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:a040696505e266e8f06a5f06b108f2db5ad3a6ed803d547d736455c1928a6a0e
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:d9560ed2758d3793139d3e13b7e5ce9d0f4ac0f8a6a8767449822c033cb2a456
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:9d997314bc7c6b30c2c3ff54b27bcbc1cdeac6dba300cab8a898b97fbce05ff7
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:dbff3d1592b192253a288fb3230a0ff125564540a30a6dd5ac881e79ed93b60a
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:9ae64f17191b98bb75f613969a9415103dad304e4fd20fb444f4fc4e57a57e0e
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:14952f38ab1d7f0abe4d54883e2c93d01ec1240ae809aefa9ca8cb4aaef080e1
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:15a971cef680f6f43e6b8ad7e0f5dab572ebacebc35cfbc6f7949a4ccd7789c2
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:d41ece8df9482ff860706edcdf27b53ac97d7c87407b3acfe8b1d9adfbf40e38
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:7581cd2330c20be2b77621e0e14da183d2b94f40fa5d3f448dede5f8c182759e
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:6fbdfdf3b9206058d5efc6653194bf6c42c07655817e3c97044da678e84b310a
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:3aa57a7ee9b342844ea4aa20bbbdf7a43b5d41f7866baa619f25473a230011f3
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:91b86228495f38d5528be1a8ffec96695ad72fafb76d2ead87f7bee25540d8b6
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:2e240514425d70039f5b48fe1908c56fae6efcc3eff93c63ec67f4aa3318a5fb
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:69388282df18b68c1d937ac98a5026b34221a54e5ce4b9cf4b89379b66623488
        - registry.redhat.io/container-native-virtualization/mtq-operator-rhel9@sha256:e29ffbe332d5da66ab4612a50ed73813ccbde92622474273b617cccd3bc24cd7
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:d253f1dbb54c70cf5874c2177bcd1cc3697381916cd4cf825bda934bfdb3d53c
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:7752f41a0f380fcd0d2652d286d4aac1ebb94a8a5fbb655b2e7058054f2eca4d
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:b20230f9a74652cf676bcf05dad81fc1ff64f9cff7232782c62cf585db48b115
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:0c5e1ee8ae0f19237a2c8b93466a7fc217a112cf1fb6dd55e2be27f95c9b66a3
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:bf7562a3bbc7655565b69fc96f2deae15729bbdef158cd90b97fc5ecc58d5bd8
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:fd2b4a337a8e45061cb6f455de3eefe27b004844c724fd821efc3b47215b74d3
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:77211068f3a8d331f1b21c64918861398f2cdbb2a3b760a5e8df51c89efea5b9
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:cddcdbf1d02b6e55cd93787991b5cf8704d0fb3a4a0bef5f58c918b57e26ee16
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:387fba10dc8f16fbfbc8c84274c5f1f5f6d01cf11f17f1b54fa0775a3300c6de
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:53ae4db672daad95d99f2955545ba335095c96630efc2a2ff19c5019a9413e87
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:63262574fa7d51d3b8c26b79e22d78f462c70239746aec6b8b3b951a02919e81
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:4046e8e862e0c55f09e9031b95e6a77af4d491593e39770c644287d44888b0bb
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:42fc0d1d112b30f65d6a48fa0442e2f41872658d2431ac8618dadf0aa7a41208
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:9e87e7d8679211a317d11e588868b40235da8c668bfb2ee10b564f6850fb91d1
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:38e767770ea0a74c4d594a382eed72bf9efa527e45ac6438acb79a1039f57fbd
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:75d76d7274f5169a414d8024125d4f0d8709888ac2b5d6121360635c7967d878
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:96092abe951f495f4f7dd0bdcb371b7efcb03a78a15e3ed8f1b37502203a6316
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:08f15657a1eee02927c5a6ef661e6e703d0064e091751bd2d8fbad5f118d9672
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:7e19b42a648cee71c60cb69fbcf6ebca05c938942e16d23f1cd61a619cffdcb6
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:878e1c8bd72cb3002457b6fbcddca99c0f4b00f6231fea52387d540b12e6329f
        version: 4.16.21
      entries:
      - name: kubevirt-hyperconverged-operator.v4.16.21
        version: 4.16.21
      - name: kubevirt-hyperconverged-operator.v4.16.17
        version: 4.16.17
      - name: kubevirt-hyperconverged-operator.v4.16.13
        version: 4.16.13
      - name: kubevirt-hyperconverged-operator.v4.16.7
        version: 4.16.7
      - name: kubevirt-hyperconverged-operator.v4.16.6
        version: 4.16.6
      - name: kubevirt-hyperconverged-operator.v4.16.5
        version: 4.16.5
      - name: kubevirt-hyperconverged-operator.v4.16.4
        version: 4.16.4
      - name: kubevirt-hyperconverged-operator.v4.16.3
        version: 4.16.3
      - name: kubevirt-hyperconverged-operator.v4.16.2
        version: 4.16.2
      - name: kubevirt-hyperconverged-operator.v4.16.1
        version: 4.16.1
      - name: kubevirt-hyperconverged-operator.v4.16.0
        version: 4.16.0
      - name: kubevirt-hyperconverged-operator.v4.15.2
        version: 4.15.2
      - name: kubevirt-hyperconverged-operator.v4.15.1
        version: 4.15.1
      - name: kubevirt-hyperconverged-operator.v4.15.0
        version: 4.15.0
      - name: kubevirt-hyperconverged-operator.v4.14.3
        version: 4.14.3
      - name: kubevirt-hyperconverged-operator.v4.14.2
        version: 4.14.2
      - name: kubevirt-hyperconverged-operator.v4.14.1
        version: 4.14.1
      - name: kubevirt-hyperconverged-operator.v4.14.0
        version: 4.14.0
      - name: kubevirt-hyperconverged-operator.v4.13.4
        version: 4.13.4
      - name: kubevirt-hyperconverged-operator.v4.13.3
        version: 4.13.3
      - name: kubevirt-hyperconverged-operator.v4.13.2
        version: 4.13.2
      - name: kubevirt-hyperconverged-operator.v4.13.1
        version: 4.13.1
      - name: kubevirt-hyperconverged-operator.v4.13.0
        version: 4.13.0
      - name: kubevirt-hyperconverged-operator.v4.12.2
        version: 4.12.2
      - name: kubevirt-hyperconverged-operator.v4.12.1
        version: 4.12.1
      - name: kubevirt-hyperconverged-operator.v4.12.0
        version: 4.12.0
      name: stable
    defaultChannel: stable
    packageName: kubevirt-hyperconverged
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudtrail-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudtrail-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudtrail.services.k8s.aws/v1alpha1",
                "kind": "EventDataStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudtrail.services.k8s.aws/v1alpha1",
                "kind": "Trail",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudtrail-controller:1.1.1
          createdAt: "2025-09-19T21:16:25Z"
          description: AWS CloudTrail controller is a service controller for managing
            CloudTrail resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventDataStore represents the state of an AWS cloudtrail
              EventDataStore resource.
            displayName: EventDataStore
            kind: EventDataStore
            name: eventdatastores.cloudtrail.services.k8s.aws
            version: v1alpha1
          - description: Trail represents the state of an AWS cloudtrail Trail resource.
            displayName: Trail
            kind: Trail
            name: trails.cloudtrail.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Compute Cloud (CloudTrail) resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudTrail**

          AWS CloudTrail is an AWS service that helps you enable operational and risk auditing, governance, and compliance of your AWS account. Actions taken by a user, role, or an AWS service are recorded as events in CloudTrail. Events include actions taken in the AWS Management Console, AWS Command Line Interface, and AWS SDKs and APIs.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudTrail
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudtrail
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudTrail Developer Resources
          url: https://aws.amazon.com/cloudtrail/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudtrail maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudtrail-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-cloudtrail-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudtrail-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudtrail-controller.v1.0.28
        version: 1.0.28
      - name: ack-cloudtrail-controller.v1.0.26
        version: 1.0.26
      - name: ack-cloudtrail-controller.v1.0.25
        version: 1.0.25
      - name: ack-cloudtrail-controller.v1.0.24
        version: 1.0.24
      - name: ack-cloudtrail-controller.v1.0.23
        version: 1.0.23
      - name: ack-cloudtrail-controller.v1.0.22
        version: 1.0.22
      - name: ack-cloudtrail-controller.v1.0.21
        version: 1.0.21
      - name: ack-cloudtrail-controller.v1.0.20
        version: 1.0.20
      - name: ack-cloudtrail-controller.v1.0.19
        version: 1.0.19
      - name: ack-cloudtrail-controller.v1.0.16
        version: 1.0.16
      - name: ack-cloudtrail-controller.v1.0.15
        version: 1.0.15
      - name: ack-cloudtrail-controller.v1.0.14
        version: 1.0.14
      - name: ack-cloudtrail-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudtrail-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudtrail-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudtrail-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudtrail-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudtrail-controller.v1.0.5
        version: 1.0.5
      - name: ack-cloudtrail-controller.v1.0.4
        version: 1.0.4
      - name: ack-cloudtrail-controller.v1.0.3
        version: 1.0.3
      - name: ack-cloudtrail-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudtrail-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudtrail-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudtrail-controller.v0.1.3
        version: 0.1.3
      - name: ack-cloudtrail-controller.v0.1.2
        version: 0.1.2
      - name: ack-cloudtrail-controller.v0.1.1
        version: 0.1.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudtrail-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Provider Name
      provider-url: https://your.domain
    name: smilecdr-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smilecdr-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.smilecdr.com/v1alpha1",
                "kind": "Smilecdr",
                "metadata": {
                  "name": "smilecdr"
                },
                "spec": {
                  "labels": {
                    "application": "smilecdr",
                    "client": "smilecdr",
                    "env": "dev",
                    "version": "one"
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Smilecdr
            name: smilecdrs.operator.smilecdr.com
            version: v1alpha1
        description: Smilecdr Operator description. TODO.
        displayName: Smilecdr Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - smilecdr-operator
        links:
        - name: Smilecdr Operator
          url: https://smilecdr-operator.domain
        maintainers:
        - email: your@email.com
          name: Maintainer Name
        maturity: alpha
        provider:
          name: Provider Name
          url: https://your.domain
        relatedImages:
        - quay.io/jibran_shaukat/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - quay.io/jibran_shaukat/smilecdr@sha256:cd6b2cec6f5b9ad9e0f0434532f8fb00b6919fcd840e686d96ce08874f2154db
        version: 0.0.1
      entries:
      - name: smilecdr-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: smilecdr-operator
    provider:
      name: Provider Name
      url: https://your.domain
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: model-builder-for-vision-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-modelbuilder-for-vision-rhmp.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "Modelbuilder",
                "metadata": {
                  "name": "modelbuilder-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "backup_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "metadata_storage": {
                    "storage_class": "ibmc-block-bronze",
                    "storage_size": "20G"
                  },
                  "photo_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "500G"
                  },
                  "in_memory_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              },
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "ModelbuilderAWS",
                "metadata": {
                  "name": "modelbuilder-aws-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "postgres": {
                    "storage_size": "20G",
                    "storage_class": "gp2",
                    "backup_storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
          createdAt: "2021-08-28T12:00:00Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/model-builder-for-vision-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/model-builder-for-vision-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.0.3
          operators.operatorframework.io/builder: operator-sdk-v1.11.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/ibm-modelbuilder-for-vision/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Model Builder for Vision on AWS
            displayName: Deploy-Modelbuilder on AWS
            kind: ModelbuilderAWS
            name: modelbuildersaws.modelbuilder.com
            version: v1alpha1
          - description: Deploy Model Builder for Vision on IBM
            displayName: Deploy-Modelbuilder on IBM
            kind: Modelbuilder
            name: modelbuilders.modelbuilder.com
            version: v1alpha1
        description: IBM Services Software Model Builder for Vision is an AI training
          platform that can use the capabilities of IBM Cloud and Amazon Web Services
          and their GPUs to quickly train computer vision models for compatible mobile
          apps. IBM Services Software Inspection Workbench iOS and iPadOS app will
          be used as the exclusive UI for labeling and training. Specifically, IBM
          Model Builder for Vision will train computer vision models that can then
          be deployed for use on the IBM Services Software Inspector Portable , IBM
          Services Software Inspector Wearable and the IBM Maximo Visual Inspection
          Mobile apps.
        displayName: IBM Model Builder for Vision
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Model Builder
        - AI
        - Computer Vision
        - Inspection
        - Mobile
        - Core ML
        - TensorFlow Lite
        links:
        - name: IBM Model Builder for Vision
          url: http://ibm.biz/model_builder
        maintainers:
        - email: isshelp@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.connect.redhat.com/ibm/modelbuilder-augmentor@sha256:ef5d0dde3eb334d919c2773af63007f1ef5e735ae040ef77fffe3440f854f068
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/rhel8/redis-6@sha256:2f65f46b25d84a3cb41a6a10f93786bcab47b7354ca9ac13a8e602ea833eba60
        - registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
        - registry.connect.redhat.com/ibm/modelbuilder-image-server@sha256:ecdb665af651417c50335e7259b89ef8096521abec431976bd7abe4c1edc49d6
        - registry.connect.redhat.com/ibm/modelbuilder-auth-service@sha256:f96b444fdcf7781775c9eba942bcc5e5ebdcf5f50820447f5eb26061c4c674cd
        - registry.connect.redhat.com/ibm/modelbuilder-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/modelbuilder-cv-training-service@sha256:c3c9080b08b87022c784fd98ac3e2774ec89ce29a2dd19d91dee039e3fdb2b36
        version: 1.0.3
      entries:
      - name: ibm-modelbuilder-for-vision-rhmp.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: model-builder-for-vision-certified-rhmp
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: fuse-console
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-console.v7.10.3-0.1655306919.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "replicas": 1,
                  "version": "1.10",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "100Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:37dc38375f5d553ac1d9f20f66ce005ed9495354b81d7915ae3bd7dd6af132f3
          createdAt: "2020-09-14T10:59:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.10.3'
          olm.substitutesFor: fuse-console.v7.10.3-0.1652362834.p
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.10/html/managing_fuse
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:37dc38375f5d553ac1d9f20f66ce005ed9495354b81d7915ae3bd7dd6af132f3
        version: 7.10.3+0.1655306919.p
      entries:
      - name: fuse-console.v7.10.3
        version: 7.10.3
      - name: fuse-console.v7.10.3-0.1652362834.p
        version: 7.10.3+0.1652362834.p
      - name: fuse-console.v7.10.3-0.1655306919.p
        version: 7.10.3+0.1655306919.p
      name: 7.10.x
    - currentCSV: fuse-console.v7.11.1-0.1681224695.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.11",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:518b5ace9178e2aafd7b6f247fd41d9392abd1a7bccb4cb53783531820611193
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.11.1'
          olm.substitutesFor: fuse-console.v7.11.1-0.1675972857.p
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.11/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:518b5ace9178e2aafd7b6f247fd41d9392abd1a7bccb4cb53783531820611193
        version: 7.11.1+0.1681224695.p
      entries:
      - name: fuse-console.v7.11.1
        version: 7.11.1
      - name: fuse-console.v7.11.1-0.1669922044.p
        version: 7.11.1+0.1669922044.p
      - name: fuse-console.v7.11.1-0.1675972857.p
        version: 7.11.1+0.1675972857.p
      - name: fuse-console.v7.11.1-0.1681224695.p
        version: 7.11.1+0.1681224695.p
      name: 7.11.x
    - currentCSV: fuse-console.v7.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.12",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:4e6889bf76d323956e0b594db7381e8162fea6b108f53f9e6dd0b38cf3596c8d
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.12.1'
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.12/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:4e6889bf76d323956e0b594db7381e8162fea6b108f53f9e6dd0b38cf3596c8d
        version: 7.12.1
      entries:
      - name: fuse-console.v7.12.1
        version: 7.12.1
      name: 7.12.x
    - currentCSV: fuse-console.v7.13.0-0.1757604404.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.13",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:72d98254a821bce237b3dde896f2825628cb7b6df564e23949337590586e8bea
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.7.0 <7.13.0'
          olm.substitutesFor: fuse-console.v7.13.0-0.1754559584.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.13/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:72d98254a821bce237b3dde896f2825628cb7b6df564e23949337590586e8bea
        version: 7.13.0+0.1757604404.p
      entries:
      - name: fuse-console.v7.13.0-0.1744281421.p
        version: 7.13.0+0.1744281421.p
      - name: fuse-console.v7.13.0-0.1723709432.p
        version: 7.13.0+0.1723709432.p
      - name: fuse-console.v7.13.0-0.1757604404.p
        version: 7.13.0+0.1757604404.p
      - name: fuse-console.v7.13.0-0.1721224332.p
        version: 7.13.0+0.1721224332.p
      - name: fuse-console.v7.13.0-0.1744737026.p
        version: 7.13.0+0.1744737026.p
      - name: fuse-console.v7.13.0-0.1724659726.p
        version: 7.13.0+0.1724659726.p
      - name: fuse-console.v7.13.0-0.1725026290.p
        version: 7.13.0+0.1725026290.p
      - name: fuse-console.v7.13.0-0.1748960166.p
        version: 7.13.0+0.1748960166.p
      - name: fuse-console.v7.13.0-0.1738056254.p
        version: 7.13.0+0.1738056254.p
      - name: fuse-console.v7.13.0-0.1739885034.p
        version: 7.13.0+0.1739885034.p
      - name: fuse-console.v7.13.0-0.1741875974.p
        version: 7.13.0+0.1741875974.p
      - name: fuse-console.v7.13.0-0.1741884704.p
        version: 7.13.0+0.1741884704.p
      - name: fuse-console.v7.13.0-0.1719822976.p
        version: 7.13.0+0.1719822976.p
      - name: fuse-console.v7.13.0
        version: 7.13.0
      - name: fuse-console.v7.13.0-0.1731460809.p
        version: 7.13.0+0.1731460809.p
      - name: fuse-console.v7.13.0-0.1749640277.p
        version: 7.13.0+0.1749640277.p
      - name: fuse-console.v7.13.0-0.1751291353.p
        version: 7.13.0+0.1751291353.p
      - name: fuse-console.v7.13.0-0.1753702606.p
        version: 7.13.0+0.1753702606.p
      - name: fuse-console.v7.13.0-0.1753778626.p
        version: 7.13.0+0.1753778626.p
      - name: fuse-console.v7.13.0-0.1753785080.p
        version: 7.13.0+0.1753785080.p
      - name: fuse-console.v7.13.0-0.1753860418.p
        version: 7.13.0+0.1753860418.p
      - name: fuse-console.v7.13.0-0.1754303647.p
        version: 7.13.0+0.1754303647.p
      - name: fuse-console.v7.13.0-0.1754559584.p
        version: 7.13.0+0.1754559584.p
      - name: fuse-console.v7.13.0-0.1754567932.p
        version: 7.13.0+0.1754567932.p
      - name: fuse-console.v7.12.1
        version: 7.12.1
      name: 7.13.x
    - currentCSV: fuse-console.v7.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "replicas": 1,
                  "version": "1.9",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "100Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:846d57f3fd9f307a98b0169b6733c26299209d99ed189f219c48f36ba771b198
          createdAt: "2020-09-14T10:59:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.9.0'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.9/html/managing_fuse
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:846d57f3fd9f307a98b0169b6733c26299209d99ed189f219c48f36ba771b198
        version: 7.9.0
      entries:
      - name: fuse-console.v7.9.0
        version: 7.9.0
      name: 7.9.x
    defaultChannel: 7.13.x
    packageName: fuse-console
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: opentelemetry-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opentelemetry-operator.v0.135.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "Instrumentation",
                "metadata": {
                  "name": "instrumentation"
                },
                "spec": {
                  "exporter": {
                    "endpoint": "http://otel-collector-headless:4317"
                  }
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpAMPBridge",
                "metadata": {
                  "name": "opampbridge-sample"
                },
                "spec": {
                  "capabilities": {
                    "AcceptsOpAMPConnectionSettings": true,
                    "AcceptsOtherConnectionSettings": true,
                    "AcceptsRemoteConfig": true,
                    "AcceptsRestartCommand": true,
                    "ReportsEffectiveConfig": true,
                    "ReportsHealth": true,
                    "ReportsOwnLogs": true,
                    "ReportsOwnMetrics": true,
                    "ReportsOwnTraces": true,
                    "ReportsRemoteConfig": true,
                    "ReportsStatus": true
                  },
                  "componentsAllowed": {
                    "exporters": [
                      "debug"
                    ],
                    "processors": [
                      "memory_limiter"
                    ],
                    "receivers": [
                      "otlp"
                    ]
                  },
                  "endpoint": "ws://opamp-server:4320/v1/opamp"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": "receivers:\n  otlp:\n    protocols: \n      grpc: {}\n      http: {}\n\nexporters:\n  debug: {}\n\nservice:\n  pipelines:\n    traces:\n      receivers: [otlp]\n      exporters: [debug]\n"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1beta1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": {
                    "exporters": {
                      "debug": {}
                    },
                    "receivers": {
                      "otlp": {
                        "protocols": {
                          "grpc": {},
                          "http": {}
                        }
                      }
                    },
                    "service": {
                      "pipelines": {
                        "traces": {
                          "exporters": [
                            "debug"
                          ],
                          "receivers": [
                            "otlp"
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/rhosdt/opentelemetry-rhel8-operator@sha256:45c6c52744e9a17ffd883a2864cb7722ce3f9580e522c976d4f55623118f3504
          createdAt: 24 Sep 2025, 10:19
          description: Provides the OpenTelemetry components, including the Collector
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.33.0 <0.135.0-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-opentelemetry-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-telemetry/opentelemetry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instrumentation is the spec for OpenTelemetry instrumentation.
            displayName: OpenTelemetry Instrumentation
            kind: Instrumentation
            name: instrumentations.opentelemetry.io
            version: v1alpha1
          - description: OpAMPBridge is the Schema for the opampbridges API.
            displayName: OpAMP Bridge
            kind: OpAMPBridge
            name: opampbridges.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1beta1
          - description: TargetAllocator is the Schema for the targetallocators API.
            displayName: Target Allocator
            kind: TargetAllocator
            name: targetallocators.opentelemetry.io
            version: v1alpha1
        description: |-
          Red Hat build of OpenTelemetry is a collection of tools, APIs, and SDKs. You use it to instrument, generate, collect, and export telemetry data (metrics, logs, and traces) for analysis in order to understand your software's performance and behavior.
          This operator was previously called Red Hat OpenShift distributed tracing data collection.

          ### Operator features

          * **Sidecar injection** - annotate your pods and let the operator inject a sidecar.
          * **Managed upgrades** - updating the operator will automatically update your OpenTelemetry collectors.
          * **Deployment modes** - your collector can be deployed as sidecar, daemon set, or regular deployment.
          * **Service port management** - the operator detects which ports need to be exposed based on the provided configuration.

          ## Examples

          There is a list of examples to help you deploy different configurations in [this GitHub repository](https://github.com/os-observability/redhat-rhosdt-samples)

          ### Support & Troubleshooting

          Red Hat build of OpenTelemetry is available as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/4.17/html/red_hat_build_of_opentelemetry/index) is recommended when installing, configuring, and managing the Operator and instances.
        displayName: Red Hat build of OpenTelemetry
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opentelemetry
        - tracing
        - logging
        - metrics
        - monitoring
        - troubleshooting
        links:
        - name: OpenTelemetry Operator
          url: https://github.com/open-telemetry/opentelemetry-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/opentelemetry-rhel8-operator@sha256:45c6c52744e9a17ffd883a2864cb7722ce3f9580e522c976d4f55623118f3504
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5a6edfecf1636805e6f0bcad02df10bceec764f0580f3ac8fe62aba978d50006
        - registry.redhat.io/rhosdt/opentelemetry-collector-rhel8@sha256:5f31caf1f4a12cf2832ac3e84877f4fb0c856073820498c8a5be13698348c38a
        - registry.redhat.io/rhosdt/opentelemetry-target-allocator-rhel8@sha256:f77a26edafba06a4fca1a07083e06a908664852cf228b1814455ffa8155cc979
        version: 0.135.0-1
      entries:
      - name: opentelemetry-operator.v0.135.0-1
        version: 0.135.0-1
      - name: opentelemetry-operator.v0.127.0-2
        version: 0.127.0-2
      - name: opentelemetry-operator.v0.127.0-1
        version: 0.127.0-1
      - name: opentelemetry-operator.v0.119.0-2
        version: 0.119.0-2
      - name: opentelemetry-operator.v0.119.0-1
        version: 0.119.0-1
      - name: opentelemetry-operator.v0.113.0-2
        version: 0.113.0-2
      - name: opentelemetry-operator.v0.113.0-1
        version: 0.113.0-1
      - name: opentelemetry-operator.v0.107.0-5
        version: 0.107.0-5
      - name: opentelemetry-operator.v0.107.0-4-0.1727168137.p
        version: 0.107.0-4+0.1727168137.p
      - name: opentelemetry-operator.v0.107.0-4
        version: 0.107.0-4
      - name: opentelemetry-operator.v0.102.0-3
        version: 0.102.0-3
      - name: opentelemetry-operator.v0.102.0-2
        version: 0.102.0-2
      - name: opentelemetry-operator.v0.100.1-3
        version: 0.100.1-3
      - name: opentelemetry-operator.v0.93.0-3
        version: 0.93.0-3
      - name: opentelemetry-operator.v0.93.0-2
        version: 0.93.0-2
      - name: opentelemetry-operator.v0.89.0-3
        version: 0.89.0-3
      - name: opentelemetry-operator.v0.81.1-5
        version: 0.81.1-5
      - name: opentelemetry-operator.v0.74.0-5-0.1687199949.p
        version: 0.74.0-5+0.1687199949.p
      - name: opentelemetry-operator.v0.74.0-5
        version: 0.74.0-5
      - name: opentelemetry-operator.v0.60.0-2
        version: 0.60.0-2
      - name: opentelemetry-operator.v0.56.0-1
        version: 0.56.0-1
      name: stable
    defaultChannel: stable
    packageName: opentelemetry-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: http://marketplace.redhat.com
    name: marketplace-games-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marketplace-games-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "game.marketplace.redhat.com/v1",
                "kind": "Arcade",
                "metadata": {
                  "name": "arcade-sample"
                },
                "spec": {
                  "size": 1
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Arcade is the Schema for the arcades API
            displayName: Arcade Instance
            kind: Arcade
            name: arcades.game.marketplace.redhat.com
            version: v1
        description: RHM Arcade Operator - a simple operator used to showcase how
          to build an operator around an existing web application that can can run
          on Openshift, Vanilla k8s, or on-boarded to Red Hat Marketplace.
        displayName: Arcade Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - rhm
        - red hat
        - marketplace
        - games
        - arcade
        links:
        - name: Marketplace Games Operator
          url: https://github.com/redhat-marketplace/marketplace-games-operator
        maturity: alpha
        provider:
          name: Red Hat Marketplace
          url: http://marketplace.redhat.com
        relatedImages:
        - registry.connect.redhat.com/rhm-labs/arcade-operator@sha256:d6959be6693728d2dc4cd38b69251551d23c5002cb52db2684df31a0a8f7b0ab
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        version: 1.0.0
      entries:
      - name: marketplace-games-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: marketplace-games-operator
    provider:
      name: Red Hat Marketplace
      url: http://marketplace.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-loadcore-agents
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-loadcore-agents.v0.2.25
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2024-11-20T18:10:53Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <=10.0.0'
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Agents
            name: agents.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight LoadCore Agents Operator
        displayName: Keysight LoadCore Agents Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Loadcore Agents
          url: https://keysight-loadcore-agents.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:7efeeb8b29872a6f0271f651d7ae02c91daea16d853c50e374c310f044d8c76c
        - prod.harbor.keysight.digital/loadcore/keysight-loadcore-agents@sha256:96e18b6270097c23bad5445a15613e467d06c6fbdbaf39545edb79915aacc04f
        version: 0.2.25
      entries:
      - name: keysight-loadcore-agents.v0.2.25
        version: 0.2.25
      name: alpha
    defaultChannel: alpha
    packageName: keysight-loadcore-agents
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-secret-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secret-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/tools@sha256:47204b2697dfbab5eaf9d6dcb4c160dd427a148d6b840727239b773810e74fc9
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:161e754437428c68490333337303fb30b695a9ac677c3a00a1dc27565b855ab8
        - quay.io/stackable/secret-operator@sha256:1267311830649f928b528512a3cdc843efad6dd32b851d99c84639fbe70e35a3
        - quay.io/stackable/tools@sha256:47204b2697dfbab5eaf9d6dcb4c160dd427a148d6b840727239b773810e74fc9
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:a6d6631d35c32b874c995b6af889636ac8dd1f39dcb68810d9f9b10a4e2c1987
        version: 24.11.1
      entries:
      - name: secret-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: secret-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:6eb08b99259d62514d87844e1b7470382d500696fecc9a1232f6bb536ed897fa
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/secret-operator@sha256:6eb08b99259d62514d87844e1b7470382d500696fecc9a1232f6bb536ed897fa
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:dd730457133f619d8759269abcfa79d7aeb817e01ba7af2d3aa1417df406ea56
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        version: 25.3.0
      entries:
      - name: secret-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: secret-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
          - description: Trust stores managed by the operator.
            displayName: Trust stores managed by the operator.
            kind: TrustStore
            name: truststores.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        version: 25.7.0
      entries:
      - name: secret-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: secret-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
          - description: Trust stores managed by the operator.
            displayName: Trust stores managed by the operator.
            kind: TrustStore
            name: truststores.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        version: 25.7.0
      entries:
      - name: secret-operator.v25.7.0
        version: 25.7.0
      - name: secret-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-secret-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zettaset
      provider-url: ""
    name: xcrypt-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zts-xcrypt-operator.v0.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xcrypt.zettaset.com/v1alpha1",
                "kind": "XCrypt",
                "metadata": {
                  "name": "zts-masterset"
                },
                "spec": {
                  "status": {
                    "podNames": "zts-ca",
                    "replicas": "1,"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
          createdAt: "2021-06-16T12:29:13.97Z"
          description: XCrypt Operator deploys Zettaset XCrypt Container Encryption  -
            the only software-defined solution that offers transparent high performance
            data-at-rest encryption of critical data in container environments.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/xcrypt-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/xcrypt-operator-rhmp/support?utm_source=openshift_console
          support: Zettaset
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XCrypt is the Schema for the xcrypts API
            kind: XCrypt
            name: xcrypts.xcrypt.zettaset.com
            version: v1alpha1
        description: "Zettaset’s RedHat Certified XCrypt Operator allows users to
          deploy multiple XCrypt components on OpenShift with just a few simple commands.
          The XCrypt Operator deployment provides software-defined encryption that
          transparently protects container data throughout your OpenShift environment.\n\n##
          Core Features\n* **Automated XCrypt provisioning** - Deploying a Zettaset
          XCrypt has never been easier. Adjust cluster specific configuration and
          let the Zettaset Operator take care of provisioning XCrypt services and
          setting up the XCrypt encryption to your exact specification. \n* **CSI
          Driver deployment**: Zettaset XCrypt Operator automatically deploys Zettaset
          XCrypt CSI Driver in order to provide automated and transparent data-at-rest
          \ encryption functionality compatible with Openshift and any other container
          orchestrator. Once the Operator is provisioned, the driver runs fine tuned
          and ready to work with other XCrypt components.\n* **Multi-node topology**:
          Encrypt as many nodes and volumes as you need to. XCrypt components can
          be installed on multiple nodes across the cluster. This is defined by two
          labels: zts-master and zts-worker. Normally zts-master node is a single
          node that runs Zettaset XCrypt major services required for the product operation.
          Multiple nodes can be marked as zts-worker and those are the nodes that
          can have encrypted volumes.\n* **Seamless Encrypted Volume mount**: Volumes
          are not only being encrypted, but also are mounted/unmounted to containers
          with Persistent Volume Claims. This happens as seamlessly as regular volumes
          provision and mount/unmount.\n* **Robust failover**: A node failure doesn’t
          affect XCrypt operations on other nodes of the cluster. Encryption on a
          restarted/recovered node gets back to normal functioning if the node stays
          healthy and data hasn’t been corrupted during the restart\n\n## About Zettaset
          XCrypt Kubernetes Encryption for OpenShift\nXCrypt Kubernetes Encryption
          for OpenShift is a software-only solution that enables encryption of data
          at rest stored in Red Hat OpenShift container environments without slowing
          down the DevSecOps process. In addition to providing a transparent and high-performance
          layer of security that flexibly protects container data across on-premise,
          in hybrid, and multi-cloud environments, it also helps your organization
          efficiently transition from DevOps to DevSecOps.  \n\nMore information about
          Zettaset XCrypt Kubernetes Encryption for OpenShift can be found in the
          [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)\n\n##
          RedHat OpenShift + Zettaset: The next generation of secure enterprise infrastructure
          for DevSecOps  \nBy combining RedHat’s Container Platform with Zettaset’s
          certified software-defined encryption, you can create the next generation
          of secure enterprise infrastructure that DevSecOps requires. Together they
          create the stable, consistent, and supported base that your organization
          needs to develop applications faster, simplify updates, drive innovation,
          and reduce the risk of potential security breaches and data theft.\n###
          Core Features\n* Software-only for simple and scalable deployments \n* Negligible
          performance impact\n* Transparent to developers and administrators \n* Works
          with AWS, Google, and Azure \n* Unique encryption key per container volume
          \n* Encrypted volumes are only available when in use\n* Direct integration
          with OpenShift \n* Automated encryption policy management Secure erase of
          volumes\n* Ability to securely decommission a node\n\n## How does XCrypt
          Kubernetes Encryption fit into OpenShift? \nXCrypt XCrypt Kubernetes Encryption
          for OpenShift Encryption makes it incredibly easy to protect container data
          in your OpenShift environment. \n### Use cases include:\n* Transitioning
          from DevOps to DevSecOps \n* Ensuring data protection in OpenShift container
          environments with negligible performance impact \n* Achieving compliance
          in regulated industries like healthcare and finance\n\n## Other Information\nRole-based
          access control needs to be configured prior to the Operator installation.
          Please find details in the [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)
          \   "
        displayName: Zettaset XCrypt Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - encryption
        - storage
        links:
        - name: Zettaset
          url: https://www.zettaset.com/
        maintainers:
        - email: info@zettaset.com
          name: Zettaset
        maturity: alpha
        provider:
          name: Zettaset
        relatedImages:
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-ca-1@sha256:9f1015dfb459b69f5dbf449efa0517452095a6cf4f20826ebae278f9f800ffe7
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-kmip-server-1@sha256:9af5d21dd748dfbf4e107f27a5f10fa3c0ba928ade9b989a744b620ce34ec757
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-license-server-1@sha256:bc5486892a41e70f5dc4c4e1d82fca8db2149598fb8b5950ba5da1ee2d9d245d
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-host-manager-1@sha256:9faa4b597eb74df412761dffa057b64fb69dbf9fa016414962eb2099a343e734
        - registry.marketplace.redhat.com/rhm/zts/zts-csi-driver-1@sha256:c15fa9a7cc40891103f426206663dcebe1210c8ee3ab6037cbf64e1f86b9c7e5
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
        version: 0.0.17
      entries:
      - name: zts-xcrypt-operator.v0.0.17
        version: 0.0.17
      name: alpha
    defaultChannel: alpha
    packageName: xcrypt-operator-rhmp
    provider:
      name: Zettaset
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://www.medik8s.io/
    name: customized-user-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: customized-user-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediation",
                "metadata": {
                  "name": "customizeduserremediation-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediationConfig",
                "metadata": {
                  "name": "customizeduserremediationconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediationTemplate",
                "metadata": {
                  "name": "customizeduserremediationtemplate-sample",
                  "namespace": "openshift-workload-availability"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/customized-user-remediation-operator:v0.1.0
          createdAt: "2024-03-25 16:17:57"
          description: Customized User Remediation Operator will remediate unhealthy
            nodes by running a pre defined user script on the unhealthy node. Since
            the user has full control over the script content, is may be quite risky
            so apply discretion.
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/customized-user-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CustomizedUserRemediationConfig is the Schema for the customizeduserremediationconfigs
              API
            displayName: Customized User Remediation Config
            kind: CustomizedUserRemediationConfig
            name: customizeduserremediationconfigs.customized-user-remediation.medik8s.io
            version: v1alpha1
          - description: CustomizedUserRemediation is the Schema for the customizeduserremediations
              API
            displayName: Customized User Remediation
            kind: CustomizedUserRemediation
            name: customizeduserremediations.customized-user-remediation.medik8s.io
            version: v1alpha1
          - description: CustomizedUserRemediationTemplate is the Schema for the customizeduserremediationtemplates
              API
            displayName: Customized User Remediation Template
            kind: CustomizedUserRemediationTemplate
            name: customizeduserremediationtemplates.customized-user-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Customized User Remediation (CUR) works with Node Health Check to remediate unhealthy nodes by running a user customized script on the unhealthy node.
          This operator does not preform any fencing on the remediated node so users should be careful with the scripts that are being used.

          ### Compatibility
          CUR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Flexibility - CUR is quite flexible sine it runs in the remediation process a user customized script.
        displayName: Customized User Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - script
        - CUR
        links:
        - name: Customized User Remediation
          url: https://customized-user-remediation.domain
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Medik8s
          url: https://www.medik8s.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/medik8s/customized-user-remediation-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: customized-user-remediation.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: customized-user-remediation
    provider:
      name: Medik8s
      url: https://www.medik8s.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-acm-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-acm-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acm.services.k8s.aws/v1alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/acm-controller:1.1.1
          createdAt: "2025-09-20T01:19:56Z"
          description: AWS ACM controller is a service controller for managing ACM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Certificate represents the state of an AWS acm Certificate
              resource.
            displayName: Certificate
            kind: Certificate
            name: certificates.acm.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage AWS Certificate Manager (ACM) resources in AWS from within your Kubernetes cluster.

          **About Amazon ACM**

          AWS Certificate Manager (ACM) handles the complexity of creating, storing, and renewing public and private SSL/TLS X.509 certificates and keys that protect your AWS websites and applications. You can provide certificates for your [integrated AWS services](https://docs.aws.amazon.com/acm/latest/userguide/acm-services.html) either by issuing them directly with ACM or by [importing](https://docs.aws.amazon.com/acm/latest/userguide/import-certificate.html) third-party certificates into the ACM management system. ACM certificates can secure singular domain names, multiple specific domain names, wildcard domains, or combinations of these. ACM wildcard certificates can protect an unlimited number of subdomains. You can also [export](https://docs.aws.amazon.com/acm/latest/userguide/export-private.html) ACM certificates signed by AWS Private CA for use anywhere in your internal PKI.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon ACM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ACM Developer Resources
          url: https://aws.amazon.com/certificate-manager/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: acm maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/acm-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-acm-controller.v1.1.1
        version: 1.1.1
      - name: ack-acm-controller.v1.1.0
        version: 1.1.0
      - name: ack-acm-controller.v1.0.13
        version: 1.0.13
      - name: ack-acm-controller.v1.0.12
        version: 1.0.12
      - name: ack-acm-controller.v1.0.11
        version: 1.0.11
      - name: ack-acm-controller.v1.0.10
        version: 1.0.10
      - name: ack-acm-controller.v1.0.9
        version: 1.0.9
      - name: ack-acm-controller.v1.0.8
        version: 1.0.8
      - name: ack-acm-controller.v1.0.7
        version: 1.0.7
      - name: ack-acm-controller.v1.0.6
        version: 1.0.6
      - name: ack-acm-controller.v1.0.5
        version: 1.0.5
      - name: ack-acm-controller.v1.0.4
        version: 1.0.4
      - name: ack-acm-controller.v1.0.1
        version: 1.0.1
      - name: ack-acm-controller.v1.0.0
        version: 1.0.0
      - name: ack-acm-controller.v0.0.20
        version: 0.0.20
      - name: ack-acm-controller.v0.0.19
        version: 0.0.19
      - name: ack-acm-controller.v0.0.18
        version: 0.0.18
      - name: ack-acm-controller.v0.0.17
        version: 0.0.17
      - name: ack-acm-controller.v0.0.16
        version: 0.0.16
      - name: ack-acm-controller.v0.0.12
        version: 0.0.12
      - name: ack-acm-controller.v0.0.10
        version: 0.0.10
      - name: ack-acm-controller.v0.0.9
        version: 0.0.9
      - name: ack-acm-controller.v0.0.6
        version: 0.0.6
      - name: ack-acm-controller.v0.0.5
        version: 0.0.5
      - name: ack-acm-controller.v0.0.4
        version: 0.0.4
      - name: ack-acm-controller.v0.0.2
        version: 0.0.2
      - name: ack-acm-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-acm-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ricardo Maraschini
      provider-url: https://github.com/ricardomaraschini/tagger
    name: tagger
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tagger.v2.1.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tagger.dev/v1beta1",
                "kind": "Tag",
                "metadata": {
                  "name": "nginx"
                },
                "spec": {
                  "from": "docker.io/library/nginx:latest",
                  "generation": 0,
                  "mirror": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          containerImage: quay.io/tagger/operator:v2.1.17
          createdAt: "2021-08-02T00:00:00Z"
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/ricardomaraschini/tagger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tag is a Container Image representation as a Custom Resource.
            displayName: Tag
            kind: Tag
            name: tags.tagger.dev
            version: v1beta1
        description: |-
          A Container Image Orchestrator. Adds mirroring and Container Image manipulation capabilities
          to a Kubernetes cluster.

          Keeping track of all Container Images in use in a Kubernetes cluster is a complicated task.
          Container Images may come from numerous different Image Registries. In some cases, controlling
          how a stable version of a given Container Image looks escapes the user's authority. To add to
          this, Container Runtimes rely on remote registries (from the cluster's point of view) when
          obtaining Container Images, potentially making the process of pulling their blobs (manifests,
          config, and layers) slower.

          The notion of indexing Container Image versions by Tags is helpful. Still, it does not provide
          users with the right confidence to always use the intended Container Image – today's "latest"
          tag might not be tomorrow's "latest" tag. In addition to that, these Image Registries allow
          access to Container Images by their Manifest content's hash (i.e., usually sha256), which gives
          users the confidence at a cost in semantics.

          When releasing a new version of an application to Push and to Deploy are split into two
          distinct steps. Both the pusher and the puller need access to the same Image Registry, adding
          complexity.  Credentials are one example of the concerns. Other factors may pop up when
          running, for instance, in an air-gapped environment, where the cluster may not reach external
          Image Registries.

          Tagger aims to overcome these caveats by providing an image management abstraction layer. For
          instance, by providing a direct mapping between a Container Image tag (e.g., "latest") and its
          correspondent Manifest content's hash, users can then refer to the Container Image by its tag
          – and be sure to use that specific version. More than that, when allied with an Internal Image
          Registry, Tagger can also automatically mirror Container Images into the cluster.

          While using Tagger, Deployments can refer to Container Images by an arbitrarily defined name,
          such as "my-app", and Tagger will make sure that they use the right Container Image through its
          internal "tag to Manifest content's hash" mapping.

          For each new "release" of a given Container Image, Tagger creates a new Generation for it,
          making it easy to roll back to previously pushed "releases" of the same Container Image in case
          of problems.

          When integrated with an Internal Registry, Tagger allows users to push or pull Images directly
          without requiring an external Image Registry. It works as a layer between the user and the
          Internal Registry. Every time a new "release" of Container Images is pushed, all Deployments
          are updated automatically. Users don't need to know about the Internal Registry existence, if
          they are logged-in to the Kubernetes cluster, they can obtain old or provide new "Generations"
          for a Container Image.

          In summary, Tagger mirrors remote Container Images into a Kubernetes cluster indexing them in
          different Generations (allowing easy navigation through these multiple Generations), provides
          an interface allowing users to pull and push images directly to the Kubernetes cluster and
          provides full integration with Kubernetes Deployments (automatic triggers new rollouts on
          Container Image changes).

          By default Tagger won't be able to mirror until you provide it with a mirror registry config.
          After install you can configure the mirror by editing the Secret mirror-registry-config in the
          operator namespace. Follow below an example of a valid mirror-registry-config secret (you gonna
          have to provide your own address, username, password, etc).

          ```yaml
          apiVersion: v1
          kind: Secret
          metadata:
            name: mirror-registry-config
          data:
            address: cmVnaXN0cnkuaW8=
            username: YWRtaW4=
            password: d2hhdCB3ZXJlIHlvdSB0aGlua2luZz8K
            token: YW4gb3B0aW9uYWwgdG9rZW4gZm9yIHRva2VuIGJhc2VkIGF1dGg=
            insecure: dHJ1ZQ==
          ```

          To make pushing an pulling images easier you should go ahead and grab the `kubectl` plugin
          available [here](https://github.com/ricardomaraschini/tagger/releases). You can download a
          full version of the README with usage examples and configuration options
          [here](https://github.com/ricardomaraschini/tagger/releases/download/v2.1.17/README.pdf).
        displayName: Tagger
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tagger
        - mirroring
        - image
        links:
        - name: Documentation
          url: https://github.com/ricardomaraschini/tagger/releases/download/v2.1.17/README.pdf
        - name: Plugin (kubectl)
          url: https://github.com/ricardomaraschini/tagger/releases/tag/v2.1.17
        maintainers:
        - email: rmarasch@redhat.com
          name: Ricardo Maraschini
        maturity: alpha
        provider:
          name: Ricardo Maraschini
          url: https://github.com/ricardomaraschini/tagger
        relatedImages:
        - quay.io/tagger/operator:v2.1.17
        version: 2.1.17
      entries:
      - name: tagger.v2.1.17
        version: 2.1.17
      name: alpha
    defaultChannel: alpha
    packageName: tagger
    provider:
      name: Ricardo Maraschini
      url: https://github.com/ricardomaraschini/tagger
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecs-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "Service",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "TaskDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecs-controller:1.1.1
          createdAt: "2025-09-19T21:02:44Z"
          description: AWS ECS controller is a service controller for managing ECS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster represents the state of an AWS ecs Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.ecs.services.k8s.aws
            version: v1alpha1
          - description: Service represents the state of an AWS ecs Service resource.
            displayName: Service
            kind: Service
            name: services.ecs.services.k8s.aws
            version: v1alpha1
          - description: TaskDefinition represents the state of an AWS ecs TaskDefinition
              resource.
            displayName: TaskDefinition
            kind: TaskDefinition
            name: taskdefinitions.ecs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Container Service (ECS) resources in AWS from within your Kubernetes cluster.

          **About Amazon ECS**

          Amazon Elastic Container Service (Amazon ECS) is a fully managed container orchestration service that helps you easily deploy, manage, and scale containerized applications. As a fully managed service, Amazon ECS comes with AWS configuration and operational best practices built-in. It's integrated with both AWS and third-party tools, such as Amazon Elastic Container Registry and Docker. This integration makes it easier for teams to focus on building the applications, not the environment. You can run and scale your container workloads across AWS Regions in the cloud, and on-premises, without the complexity of managing a control plane.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECS Developer Resources
          url: https://aws.amazon.com/ecs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecs-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ecs-controller.v1.1.1
        version: 1.1.1
      - name: ack-ecs-controller.v1.0.15
        version: 1.0.15
      - name: ack-ecs-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecs-controller.v1.0.13
        version: 1.0.13
      - name: ack-ecs-controller.v1.0.12
        version: 1.0.12
      - name: ack-ecs-controller.v1.0.11
        version: 1.0.11
      - name: ack-ecs-controller.v1.0.10
        version: 1.0.10
      - name: ack-ecs-controller.v1.0.9
        version: 1.0.9
      - name: ack-ecs-controller.v1.0.8
        version: 1.0.8
      - name: ack-ecs-controller.v1.0.7
        version: 1.0.7
      - name: ack-ecs-controller.v1.0.6
        version: 1.0.6
      - name: ack-ecs-controller.v1.0.2
        version: 1.0.2
      - name: ack-ecs-controller.v1.0.1
        version: 1.0.1
      - name: ack-ecs-controller.v1.0.0
        version: 1.0.0
      - name: ack-ecs-controller.v0.0.8
        version: 0.0.8
      - name: ack-ecs-controller.v0.0.7
        version: 0.0.7
      - name: ack-ecs-controller.v0.0.6
        version: 0.0.6
      - name: ack-ecs-controller.v0.0.5
        version: 0.0.5
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseWeb
      provider-url: ""
    name: cnfv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cnfv-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eweb.enterpriseweb.com/v1",
                "kind": "CloudNFV",
                "metadata": {
                  "name": "cloudnfv-sample"
                },
                "spec": {
                  "size": 1
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CloudNFV is the Schema for the cloudnfvs API
            displayName: Cloud NFV
            kind: CloudNFV
            name: cloudnfvs.eweb.enterpriseweb.com
            version: v1
        description: Manages a cluster of CloudNFV Runtime Modules
        displayName: CloudNFV
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - automation
        - NFV
        - controller
        links:
        - name: Cnfv Operator
          url: https://cnfv-operator.domain
        maintainers:
        - email: bill@enterpriseweb.com
          name: Bill Malyk
        maturity: alpha
        provider:
          name: EnterpriseWeb
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61a164ad711478ac603242338f7b3c55986319479878b092be128581a823598e
        - quay.io/wmalyk28/cnfv-op@sha256:e5913ad5dd20cb1b2299d2387237ef15e645c9a8c9d4f5fe0044b86a6b3dc1df
        - quay.io/wmalyk28/cloudnfv-runtime@sha256:fcde8375576c9bff602d3a16ef1ba8ae88b780166649ba6ae355737b1d35a03f
        version: 0.0.1
      entries:
      - name: cnfv-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cnfv-operator
    provider:
      name: EnterpriseWeb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzounstructured-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzounstructured-operator.v3.1.200
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzounstructured.clusters.cambridgesemantics.com/v1",
                "kind": "AnzoUnstructured",
                "metadata": {
                  "name": "au01"
                },
                "spec": {
                  "auWorker": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-w",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-worker@sha256:964633d6b16ddc517a09ac86f3e3698f5a161a72cfa50ea9ba4e301fa631e347",
                                "name": "w",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  },
                  "msLeader": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-ms",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzo-microservices-leader@sha256:0637603c00a99533d57e16009516ec73d40703ae151e631fc281f51f1f0ca93b",
                                "name": "ms",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/unstructured-operator@sha256:633a8992e35601d99c0d690dd71af314605c223b57c2038fa7090be06977b2f9
          createdAt: "2025-07-08T09:47:55Z"
          description: Kubernetes Operator for Altair Graph Studio Unstructured
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoUnstructured is the Schema for the anzounstructureds
              API
            displayName: Altair Graph Studio Unstructured Operator
            kind: AnzoUnstructured
            name: anzounstructureds.anzounstructured.clusters.cambridgesemantics.com
            version: v1
        description: |-
          The Altair Graph Studio Unstructured Operator provides the way to install and configure an anzo unstructured setup on Red Hat K8S environment.
          Currently, this is possible only through existing Anzo installation.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v3.1.2/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Studio Unstructured Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Database
        - Big Data
        links:
        - name: Altair Graph Studio Unstructured Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v3.1.2/README.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/unstructured-operator@sha256:633a8992e35601d99c0d690dd71af314605c223b57c2038fa7090be06977b2f9
        - registry.connect.redhat.com/cambridgesemantics/anzo-microservices-leader@sha256:0637603c00a99533d57e16009516ec73d40703ae151e631fc281f51f1f0ca93b
        - registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-worker@sha256:964633d6b16ddc517a09ac86f3e3698f5a161a72cfa50ea9ba4e301fa631e347
        version: 3.1.200
      entries:
      - name: anzounstructured-operator.v3.1.200
        version: 3.1.200
      - name: anzounstructured-operator.v3.1.104
        version: 3.1.104
      - name: anzounstructured-operator.v3.1.103
        version: 3.1.103
      name: stable
    defaultChannel: stable
    packageName: anzounstructured-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: openshift-secondary-scheduler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secondaryscheduleroperator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "SecondaryScheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-secondary-scheduler-operator"
                },
                "spec": {
                  "managementState": "Managed",
                  "schedulerConfig": "secondary-scheduler-config"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-rhel9-operator@sha256:210cc51760c8d3a9e1e5ba2b41f575b667e0d541b995362426e5a28f3f666600
          createdAt: 2024/10/01
          description: Runs a secondary scheduler in an OpenShift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.2.0 <1.3.1'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/secondary-scheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SecondaryScheduler is the schema for the secondaryschedulers
              API
            displayName: Secondary Scheduler
            kind: SecondaryScheduler
            name: secondaryschedulers.operator.openshift.io
            version: v1
        description: |
          The Secondary Scheduler Operator provides the ability to use a customized scheduler image that was developed using the scheduler plugin framework as a secondary scheduler in OpenShift.

          ## Secondary Scheduler Configuration

          Once the Secondary Scheduler Operator is installed, you can configure the plugins to run for your secondary scheduler.

          The configuration of the secondary scheduler is defined through a config map that wraps the KubeSchedulerConfiguration YAML file in "config.yaml".

          ## Additional Parameters

          Additionally, the following parameters can be configured:

          * `schedulerConfig` - Set the config map configuration for the secondary scheduler.
          * `schedulerImage` - Set the default operand image for the secondary scheduler.
        displayName: Secondary Scheduler Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - secondary
        - scheduling
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/secondary-scheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-rhel9-operator@sha256:210cc51760c8d3a9e1e5ba2b41f575b667e0d541b995362426e5a28f3f666600
        version: 1.3.1
      entries:
      - name: secondaryscheduleroperator.v1.3.1
        version: 1.3.1
      - name: secondaryscheduleroperator.v1.3.0
        version: 1.3.0
      - name: secondaryscheduleroperator.v1.2.2
        version: 1.2.2
      - name: secondaryscheduleroperator.v1.2.1
        version: 1.2.1
      - name: secondaryscheduleroperator.v1.2.0
        version: 1.2.0
      - name: secondaryscheduleroperator.v1.1.4
        version: 1.1.4
      - name: secondaryscheduleroperator.v1.1.2
        version: 1.1.2
      name: stable
    defaultChannel: stable
    packageName: openshift-secondary-scheduler-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: costmanagement-metrics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: costmanagement-metrics-operator.4.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "true"
          containerImage: registry.redhat.io/costmanagement/costmanagement-metrics-rhel9-operator@sha256:c5bb5291052d57390b9f4bdc660c1926827072b6dcbe09d85f5b6fc375562182
          createdAt: "2025-08-18T12:10:15Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: costmanagement-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Cost Management Metrics Operator
          ## Introduction
          The `costmanagement-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Cost Management Metrics Operator (`costmanagement-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created CostManagementMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The CostManagementMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **FIPS Compliance:** Supports deployment in high-security environments when run on an OpenShift cluster with FIPS mode enabled.

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created CostManagementMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the CostManagementMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `CostManagementMetricsConfig`, creating a new `CostManagementMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `costmanagement-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: costmanagement-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the CostManagementMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the CostManagementMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon CostManagementMetricsConfig creation** (This parameter will start to appear in CostManagementMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the CostManagementMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the costmanagement-metrics-operator
          **Note** There are separate instructions for configuring the `costmanagement-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `costmanagement-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the CostManagementMetricsConfig
          Configure the costmanagement-metrics-operator by creating a `CostManagementMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `costmanagement-metrics-operator` -> `CostManagementMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the costmanagement-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `costmanagement-metrics-operator-data` with 10Gi of storage. To configure the costmanagement-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `costmanagement-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `costmanagement-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the costmanagement-metrics-operator for a restricted network
          ##### Create the CostManagementMetricsConfig
          Configure the costmanagement-metrics-operator by creating a `CostManagementMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `costmanagement-metrics-operator` -> `CostManagementMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `costmanagement-metrics-operator-data` with 10Gi of storage. To configure the costmanagement-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `costmanagement-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: costmanagement-metrics-operator
                labels:
                  app: costmanagement-metrics-operator
              spec:
                volumes:
                - name: costmanagement-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: costmanagement-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: costmanagement-metrics-operator-reports
                    mountPath: /tmp/costmanagement-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/costmanagement-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/costmanagement-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `costmanagement-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the CostManagementMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Cost Management Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Cost Management Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/costmanagement/costmanagement-metrics-rhel9-operator@sha256:c5bb5291052d57390b9f4bdc660c1926827072b6dcbe09d85f5b6fc375562182
        version: 4.1.0
      entries:
      - name: costmanagement-metrics-operator.4.1.0
        version: 4.1.0
      - name: costmanagement-metrics-operator.4.0.0
        version: 4.0.0
      - name: costmanagement-metrics-operator.3.3.2
        version: 3.3.2
      - name: costmanagement-metrics-operator.3.3.1
        version: 3.3.1
      - name: costmanagement-metrics-operator.3.3.0
        version: 3.3.0
      - name: costmanagement-metrics-operator.3.2.1
        version: 3.2.1
      - name: costmanagement-metrics-operator.3.2.0
        version: 3.2.0
      - name: costmanagement-metrics-operator.3.1.1
        version: 3.1.1
      - name: costmanagement-metrics-operator.3.1.0
        version: 3.1.0
      - name: costmanagement-metrics-operator.3.0.1
        version: 3.0.1
      - name: costmanagement-metrics-operator.3.0.0
        version: 3.0.0
      - name: costmanagement-metrics-operator.2.0.0
        version: 2.0.0
      - name: costmanagement-metrics-operator.1.2.0
        version: 1.2.0
      - name: costmanagement-metrics-operator.1.1.9
        version: 1.1.9
      - name: costmanagement-metrics-operator.1.1.8
        version: 1.1.8
      - name: costmanagement-metrics-operator.1.1.7
        version: 1.1.7
      - name: costmanagement-metrics-operator.1.1.6
        version: 1.1.6
      - name: costmanagement-metrics-operator.1.1.5
        version: 1.1.5
      - name: costmanagement-metrics-operator.1.1.4
        version: 1.1.4
      - name: costmanagement-metrics-operator.1.1.3
        version: 1.1.3
      - name: costmanagement-metrics-operator.1.1.0
        version: 1.1.0
      - name: costmanagement-metrics-operator.1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: costmanagement-metrics-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-applicationautoscaling-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-applicationautoscaling-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "applicationautoscaling.services.k8s.aws/v1alpha1",
                "kind": "ScalableTarget",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "applicationautoscaling.services.k8s.aws/v1alpha1",
                "kind": "ScalingPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/applicationautoscaling-controller:1.1.1
          createdAt: "2025-09-20T21:43:01Z"
          description: AWS Application Auto Scaling controller is a service controller
            for managing Application Auto Scaling resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ScalableTarget represents the state of an AWS applicationautoscaling
              ScalableTarget resource.
            displayName: ScalableTarget
            kind: ScalableTarget
            name: scalabletargets.applicationautoscaling.services.k8s.aws
            version: v1alpha1
          - description: ScalingPolicy represents the state of an AWS applicationautoscaling
              ScalingPolicy resource.
            displayName: ScalingPolicy
            kind: ScalingPolicy
            name: scalingpolicies.applicationautoscaling.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Application Auto Scaling resources in AWS from within your Kubernetes cluster.
          **About Amazon Application Auto Scaling**
          Application Auto Scaling is a web service for developers and system administrators who need a solution for automatically scaling their scalable resources for individual AWS services beyond Amazon EC2. Application Auto Scaling allows you to configure automatic scaling for the following resources:
          - AppStream 2.0 fleets
          - Aurora replicas
          - Amazon Comprehend document classification and entity recognizer endpoints
          - DynamoDB tables and global secondary indexes
          - Amazon Elastic Container Service (ECS) services
          - Amazon EMR clusters
          - Amazon Keyspaces (for Apache Cassandra) tables
          - Lambda function provisioned concurrency
          - Amazon Managed Streaming for Apache Kafka (MSK) broker storage
          - SageMaker endpoint variants
          - Spot Fleet requests
          - Custom resources provided by your own applications or services. For more information, see the [GitHub repository](https://github.com/aws/aws-auto-scaling-custom-resource)

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Application Auto Scaling
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - applicationautoscaling
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Application Auto Scaling User Guide
          url: https://docs.aws.amazon.com/autoscaling/application/userguide/what-is-application-auto-scaling.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: application auto scaling maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/applicationautoscaling-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-applicationautoscaling-controller.v1.1.1
        version: 1.1.1
      - name: ack-applicationautoscaling-controller.v1.1.0
        version: 1.1.0
      - name: ack-applicationautoscaling-controller.v1.0.30
        version: 1.0.30
      - name: ack-applicationautoscaling-controller.v1.0.29
        version: 1.0.29
      - name: ack-applicationautoscaling-controller.v1.0.27
        version: 1.0.27
      - name: ack-applicationautoscaling-controller.v1.0.26
        version: 1.0.26
      - name: ack-applicationautoscaling-controller.v1.0.25
        version: 1.0.25
      - name: ack-applicationautoscaling-controller.v1.0.24
        version: 1.0.24
      - name: ack-applicationautoscaling-controller.v1.0.23
        version: 1.0.23
      - name: ack-applicationautoscaling-controller.v1.0.22
        version: 1.0.22
      - name: ack-applicationautoscaling-controller.v1.0.21
        version: 1.0.21
      - name: ack-applicationautoscaling-controller.v1.0.19
        version: 1.0.19
      - name: ack-applicationautoscaling-controller.v1.0.18
        version: 1.0.18
      - name: ack-applicationautoscaling-controller.v1.0.17
        version: 1.0.17
      - name: ack-applicationautoscaling-controller.v1.0.16
        version: 1.0.16
      - name: ack-applicationautoscaling-controller.v1.0.15
        version: 1.0.15
      - name: ack-applicationautoscaling-controller.v1.0.11
        version: 1.0.11
      - name: ack-applicationautoscaling-controller.v1.0.9
        version: 1.0.9
      - name: ack-applicationautoscaling-controller.v1.0.8
        version: 1.0.8
      - name: ack-applicationautoscaling-controller.v1.0.7
        version: 1.0.7
      - name: ack-applicationautoscaling-controller.v1.0.6
        version: 1.0.6
      - name: ack-applicationautoscaling-controller.v1.0.5
        version: 1.0.5
      - name: ack-applicationautoscaling-controller.v1.0.4
        version: 1.0.4
      - name: ack-applicationautoscaling-controller.v1.0.3
        version: 1.0.3
      - name: ack-applicationautoscaling-controller.v1.0.2
        version: 1.0.2
      - name: ack-applicationautoscaling-controller.v1.0.0
        version: 1.0.0
      - name: ack-applicationautoscaling-controller.v0.2.16
        version: 0.2.16
      - name: ack-applicationautoscaling-controller.v0.2.15
        version: 0.2.15
      - name: ack-applicationautoscaling-controller.v0.2.14
        version: 0.2.14
      - name: ack-applicationautoscaling-controller.v0.2.13
        version: 0.2.13
      - name: ack-applicationautoscaling-controller.v0.2.12
        version: 0.2.12
      - name: ack-applicationautoscaling-controller.v0.2.10
        version: 0.2.10
      - name: ack-applicationautoscaling-controller.v0.2.9
        version: 0.2.9
      - name: ack-applicationautoscaling-controller.v0.2.8
        version: 0.2.8
      - name: ack-applicationautoscaling-controller.v0.2.7
        version: 0.2.7
      - name: ack-applicationautoscaling-controller.v0.2.6
        version: 0.2.6
      - name: ack-applicationautoscaling-controller.v0.2.5
        version: 0.2.5
      - name: ack-applicationautoscaling-controller.v0.2.4
        version: 0.2.4
      - name: ack-applicationautoscaling-controller.v0.2.3
        version: 0.2.3
      - name: ack-applicationautoscaling-controller.v0.2.2
        version: 0.2.2
      - name: ack-applicationautoscaling-controller.v0.2.1
        version: 0.2.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-applicationautoscaling-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-api-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-api-controller.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "example-registry-with-studio"
                },
                "spec": {
                  "app": {
                    "host": "example-registry-with-studio-app.apps.yourcluster.example"
                  },
                  "studioUi": {
                    "enabled": true,
                    "host": "example-registry-with-studio-studio-ui.apps.yourcluster.example"
                  },
                  "ui": {
                    "host": "example-registry-with-studio-ui.apps.yourcluster.example"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-operator:3.0.5
          createdAt: "2024-12-03T11:30:52Z"
          description: Deploy and manage Apicurio API Controller on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Apicurio API
              Controller.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Apicurio API Controller

          Apicurio API Controller consists of 2 components:
           - Apicurio Registry 3
           - Apicurio Studio

          Storage options: PostgreSQL, Kafka, MySQL, SQL Server, in-memory.

          ## Apicurio API Controller Operator

          Provides a quick and easy way to deploy and manage Apicurio API Controller on Kubernetes.

          **Features**
          - Supports installation and configuration of Apicurio API Controller
          - Perform a rolling upgrade of Apicurio API Controller

          ## Prerequisites

          This operator does not deploy storage for Apicurio API Controller. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio API Controller and Apicurio API Controller Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Apicurio API Controller
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - API Management
        - Apicurio Registry
        - Apicurio Studio
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-operator:3.0.5
        version: 0.0.1
      entries:
      - name: apicurio-api-controller.v0.0.1
        version: 0.0.1
      name: 0.x
    defaultChannel: 0.x
    packageName: apicurio-api-controller
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: unsupported
      operatorframework.io/arch.s390x: unsupported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: https://www.konveyor.io
    name: konveyor-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: konveyor-operator.v0.8.0-alpha.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:c1d1942b3bb3c4e99d9b5484493e27a432455608fee4548f1d352be34a8d6ded
          createdAt: "2025-08-13T23:43:34Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0-alpha.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/kantra@sha256:06ae932d9c2f51035bc5a488dd61e8a31c344e72615b013419199168701ea3fb
        - quay.io/konveyor/generic-external-provider@sha256:d58e867b51d5e76cd3597cfb7f2478d221b966f84509dbdd44c609ff5d9169be
        - quay.io/konveyor/kai-solution-server@sha256:83987d12d8290e085b6b66de56d347a8bf262815a5f99796177cb9120513ee3a
        - quay.io/konveyor/tackle-keycloak-init@sha256:07c2ca07922d273bd2a3d9156ad8af0e5bf973aeb316496ee65d7201c381c1b1
        - quay.io/konveyor/tackle2-ui@sha256:867bf24dd1687a79f2a911c22c1059ca7bd025a307d290290265653a696b8e98
        - quay.io/konveyor/tackle2-addon-discovery@sha256:ba865a74287ff94375ed316f348ee2c1d962d8d350c8afae5dbb0aee9b64fdea
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/origin-oauth-proxy@sha256:2c811da843057119b246784f21dc7eec490669a8569e8295d8e914cdce555b24
        - quay.io/konveyor/postgresql-15-c9s@sha256:d75c5a1a0fa686ccf701e5e55633aa73d0cbb2cd9b5d8d4d4e80d7c46d78c121
        - quay.io/konveyor/tackle2-operator@sha256:c1d1942b3bb3c4e99d9b5484493e27a432455608fee4548f1d352be34a8d6ded
        - quay.io/konveyor/tackle2-hub@sha256:6be8fe8307e6686b178ace42f6db5cab8f04c5e2e29e02d2edcb5dc28ee337e9
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:7a5750fb76935e55b7b59e8363f21f5f9b48b39492d7f3e58e2e6a84d07ab027
        - quay.io/konveyor/java-external-provider@sha256:42a156f1d81352f5e09fa0ffe2b51a4d93a1f476755eb0bac9a1d24bd3a89631
        - quay.io/konveyor/tackle2-addon-platform@sha256:12efa5b7fbf6d829879b19e2820dcf782e648059f51544ebd25251dfc67fc7ca
        version: 0.8.0-alpha.4
      entries:
      - name: konveyor-operator.v0.8.0-alpha.4
        version: 0.8.0-alpha.4
      - name: konveyor-operator.v0.8.0-alpha.2
        version: 0.8.0-alpha.2
      - name: konveyor-operator.v0.8.0-alpha.1
        version: 0.8.0-alpha.1
      - name: konveyor-operator.v0.7.0-alpha.2
        version: 0.7.0-alpha.2
      - name: konveyor-operator.v0.7.0-alpha.1
        version: 0.7.0-alpha.1
      - name: konveyor-operator.v0.6.0-alpha.2
        version: 0.6.0-alpha.2
      - name: konveyor-operator.v0.6.0-alpha.1
        version: 0.6.0-alpha.1
      - name: konveyor-operator.v0.5.0-alpha.1
        version: 0.5.0-alpha.1
      - name: konveyor-operator.v0.4.0-alpha.1
        version: 0.4.0-alpha.1
      - name: konveyor-operator.v0.3.0-alpha.6
        version: 0.3.0-alpha.6
      - name: konveyor-operator.v0.3.0-alpha.5
        version: 0.3.0-alpha.5
      - name: konveyor-operator.v0.3.0-alpha.4
        version: 0.3.0-alpha.4
      - name: konveyor-operator.v0.3.0-alpha.3
        version: 0.3.0-alpha.3
      - name: konveyor-operator.v0.3.0-alpha.2
        version: 0.3.0-alpha.2
      - name: konveyor-operator.v0.3.0-alpha.1
        version: 0.3.0-alpha.1
      - name: konveyor-operator.v0.2.0-alpha.3
        version: 0.2.0-alpha.3
      - name: konveyor-operator.v0.2.0-alpha.2
        version: 0.2.0-alpha.2
      - name: konveyor-operator.v0.2.0-alpha.1
        version: 0.2.0-alpha.1
      name: alpha
    - currentCSV: konveyor-operator.v0.8.0-beta.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:019a79dec40445b6289948e173e5bf2ebfa9b495a7bb077abd634c0667953a54
          createdAt: "2025-09-17T19:50:24Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0-beta.5'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/java-external-provider@sha256:0cbdd91cc5057889b8daa2f7611fa0c944fab951c73bdf88a3e28795e9a5fc33
        - quay.io/konveyor/origin-oauth-proxy@sha256:8988f14db7dde3edcaed0541cecb0b4fd58fab9dbaffb79a6c1c4aee13b9bd73
        - quay.io/konveyor/tackle-keycloak-init@sha256:0b4bfc1fea756f10a0657861c4c422303a2349d8d9b04a23238a32ceaf106ac5
        - quay.io/konveyor/tackle2-addon-platform@sha256:d0280ade44ba5e15210c07342608576b38e6993af2e479740675a4706cd2c875
        - quay.io/konveyor/kantra@sha256:ad6f5c05e5e950e33a745accd03ad933fb7bb4907f44455f88a2b5c27e6dd15e
        - quay.io/konveyor/tackle2-ui@sha256:78398832e092557639531230cf6d023bb07b589f3587afeaacf8c68640abc454
        - quay.io/konveyor/generic-external-provider@sha256:a6c550bad7a09a1c51e226247f11c00d8453987698832cb566582842219f1f3c
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:fed49c5ff14bb232e63d2dc180d3b530da63f11bf97055fc9d86cbb9e47104a4
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/postgresql-15-c9s@sha256:157a2941ae971d8f72b2f8af1d4a29deeb5548012a978fa986c1c6e52395da4d
        - quay.io/konveyor/tackle2-addon-discovery@sha256:a29a0248e222d9ac3225e9d3df603f1383fa8844856bf68ea618d303f0b79883
        - quay.io/konveyor/tackle2-operator@sha256:019a79dec40445b6289948e173e5bf2ebfa9b495a7bb077abd634c0667953a54
        - quay.io/konveyor/kai-solution-server@sha256:1e69cf8a342090a03a9dddb1b28ae3627052b6d636ffa290a11761b94fd10648
        - quay.io/konveyor/tackle2-hub@sha256:1611913558c9f1203a844b43969d7b1539dab3578b83460e8383ee6757c3c34c
        version: 0.8.0-beta.5
      entries:
      - name: konveyor-operator.v0.8.0-beta.5
        version: 0.8.0-beta.5
      - name: konveyor-operator.v0.8.0-beta.4
        version: 0.8.0-beta.4
      - name: konveyor-operator.v0.8.0-beta.2
        version: 0.8.0-beta.2
      - name: konveyor-operator.v0.8.0-beta.1
        version: 0.8.0-beta.1
      - name: konveyor-operator.v0.6.0-beta.2
        version: 0.6.0-beta.2
      - name: konveyor-operator.v0.6.0-beta.1
        version: 0.6.0-beta.1
      - name: konveyor-operator.v0.5.0-beta.3
        version: 0.5.0-beta.3
      - name: konveyor-operator.v0.5.0-beta.2
        version: 0.5.0-beta.2
      - name: konveyor-operator.v0.5.0-beta.1
        version: 0.5.0-beta.1
      - name: konveyor-operator.v0.3.0-beta.3
        version: 0.3.0-beta.3
      - name: konveyor-operator.v0.3.0-beta.2
        version: 0.3.0-beta.2
      - name: konveyor-operator.v0.3.0-beta.1.2
        version: 0.3.0-beta.1.2
      - name: konveyor-operator.v0.3.0-beta.1.1
        version: 0.3.0-beta.1.1
      - name: konveyor-operator.v0.3.0-beta.1
        version: 0.3.0-beta.1
      name: beta
    - currentCSV: konveyor-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "move2kube-plan",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "image": "quay.io/konveyor/tackle-addon-move2kube-plan:latest"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:9688d0763d8c7812beee715a061c3c198006448855d7b891d45833f62cf869fb
          createdAt: "2023-05-25T15:23:45Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.1.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)

        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:9688d0763d8c7812beee715a061c3c198006448855d7b891d45833f62cf869fb
        - quay.io/konveyor/tackle2-addon-windup@sha256:81f0625cb82eb24ded51c7258f229f57778820c62e3be6ec05fb2b12d2970825
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle2-hub@sha256:3d15d158ffb6f1fc4ca5aaf29a12173fe344c289773902c9161056bdf69acc92
        - quay.io/konveyor/tackle2-addon@sha256:35be344bbd3393f5634f52aafd21e9c3964d1e0ba5011ecad1561600d476bbe9
        - quay.io/konveyor/tackle-pathfinder@sha256:f9c5aa49f1d27f477cdd82c21441ae0287acdcdc9b2b3fcf197884007dc5f7dd
        - quay.io/konveyor/tackle-keycloak-init@sha256:b0347e5a608352c4b7bb796b65a67d417e98ba4896db71a487fa26243214cfcf
        - quay.io/konveyor/postgresql-12-centos7@sha256:1287d947464fbf5275feb31d106d2feb856c6117f0eaec8d947f2e03bd8618c5
        - quay.io/konveyor/tackle2-ui@sha256:99650804e65e2b6cb15a37c00189d7f463e4d19a0f866a79e9c09c8420a4d028
        version: 0.1.2
      entries:
      - name: konveyor-operator.v0.1.2
        version: 0.1.2
      - name: konveyor-operator.v0.1.1
        version: 0.1.1
      - name: konveyor-operator.v0.1.0
        version: 0.1.0
      name: konveyor-0.1
    - currentCSV: konveyor-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "move2kube-plan",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "image": "quay.io/konveyor/tackle-addon-move2kube-plan:latest"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:1860bb099ac64d9a2af27692f0671b7240e4d19746fffa9f9ae2356ff2818e10
          createdAt: "2023-07-06T06:17:40Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.2.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)

        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:1860bb099ac64d9a2af27692f0671b7240e4d19746fffa9f9ae2356ff2818e10
        - quay.io/konveyor/postgresql-12-centos7@sha256:6016530ab1fce71d29421ce9a97b3a89fbbb126b87b1b079c6aadebb7ff5664c
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle-keycloak-init@sha256:ca6c85bdf8417d4be116c2098b8d8ad5ae8b7ab0b636c4cbe4a392aa424e2d47
        - quay.io/konveyor/tackle2-hub@sha256:d9c4f21f44098a2526474968b6d88094dca8b5eaeb9042716332c9cb67df47cd
        - quay.io/konveyor/tackle2-ui@sha256:d10c1e717d160c5eb2581649a8dd7cdf4922a74d92db014f72f42a48a3134341
        - quay.io/konveyor/tackle2-addon@sha256:11693e16363bcd98249209d63a29f0c6cfffab3a4071649cbeea2fd276bc2440
        - quay.io/konveyor/tackle2-addon-windup@sha256:538dfdb134d53a876d126dbd16679638584a606ef2b9e8fbe69ed1f29733c03a
        - quay.io/konveyor/tackle-pathfinder@sha256:190269b061b2065ca8ba9357bf9b959a2f587a547a44769ae0afd1ac86b4b197
        version: 0.2.1
      entries:
      - name: konveyor-operator.v0.2.1
        version: 0.2.1
      - name: konveyor-operator.v0.2.0
        version: 0.2.0
      name: konveyor-0.2
    - currentCSV: konveyor-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:cd2305373f50ce927393e43f5735915e46c32d6f8a76091ee2808218885bba72
          createdAt: "2024-05-02T22:12:42Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.3.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:51bda0c6577c15684583d607a4f0cebef760814102423f4b2025016093a19fc1
        - quay.io/konveyor/origin-oauth-proxy@sha256:05f6bd1216b921b1f47a3821f0e22b84ed6b141bb2ada474c12bd9fde613108c
        - quay.io/konveyor/tackle2-hub@sha256:bf8803721d06d1a39251e22892343a013d5ef0fb4c1883c95e1eb55aa683f486
        - quay.io/konveyor/tackle2-operator@sha256:cd2305373f50ce927393e43f5735915e46c32d6f8a76091ee2808218885bba72
        - quay.io/konveyor/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle-keycloak-init@sha256:a63702cc0cb36830a7830ed70ca344a9aacc201ba6f3c94806ca2e93160df4dd
        - quay.io/konveyor/tackle2-ui@sha256:bbf58d2ea613489f91efae450421afafb9052416b0d7971606e2729e89fd7de6
        version: 0.3.2
      entries:
      - name: konveyor-operator.v0.3.2
        version: 0.3.2
      - name: konveyor-operator.v0.3.1
        version: 0.3.1
      - name: konveyor-operator.v0.3.0
        version: 0.3.0
      name: konveyor-0.3
    - currentCSV: konveyor-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:496db8ca859608d04008ac36cf7bdc9cafd7214294d33a9c5646083c361ffcb6
          createdAt: "2024-06-05T03:08:51Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.4.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle-keycloak-init@sha256:4c7e07b1fce05f7dca47edecc0ba0ce373cc38e48879cc7eed72510d2865a18e
        - quay.io/konveyor/tackle2-ui@sha256:cb1255fc5fa4608fc659c3455f3164664b960cc2d9a3b2208967798e3532df78
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:e84ea36d206a8c143794e3986940377e77c7618225cad594bdf9a4047d7823dc
        - quay.io/konveyor/origin-oauth-proxy@sha256:05f6bd1216b921b1f47a3821f0e22b84ed6b141bb2ada474c12bd9fde613108c
        - quay.io/konveyor/tackle2-operator@sha256:496db8ca859608d04008ac36cf7bdc9cafd7214294d33a9c5646083c361ffcb6
        - quay.io/konveyor/tackle2-hub@sha256:b87acd8e2ee4a7edc4369bc375f64e2b98bc479d09556439e2ee4d74b74970a9
        - quay.io/konveyor/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        version: 0.4.0
      entries:
      - name: konveyor-operator.v0.4.0
        version: 0.4.0
      name: konveyor-0.4
    - currentCSV: konveyor-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:474d739ed470701b628f1f9e971925ac7fbe0b543f3e36305d932b963b1225fd
          createdAt: "2024-08-27T15:48:02Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.5.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:474d739ed470701b628f1f9e971925ac7fbe0b543f3e36305d932b963b1225fd
        - quay.io/konveyor/java-external-provider@sha256:33fb5d24b998f01c481a0a03e3f1bf3f7a6c93f5b662885102a0c966026e3375
        - quay.io/konveyor/generic-external-provider@sha256:334a631538d31efc8d82adae4f52303336800fa3037b5055c72400918d81955c
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/tackle2-ui@sha256:36f29fe01d3078f0e100b24d1ed126850c5a773f4b0ae922fd4c4487de1ac29d
        - quay.io/konveyor/tackle2-hub@sha256:faf91ca68e24979642a44fa013b1fb9edfee415eee835ccef9dce2eb418e2909
        - quay.io/konveyor/postgresql-15-c9s@sha256:039ce764cadd03b4cfea954700139b8ceff415a9a0970baa7f2cf60f7f929174
        - quay.io/konveyor/tackle-keycloak-init@sha256:5f652be5914ae6e23537b9fe02ccc9cd638c08de5462d9f58135a405c81f5287
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:2afbe8bea09904e084e026cd0a294fbe1c7b05d0e387f6633b09c5eaf4665e02
        - quay.io/konveyor/tackle2-addon-discovery@sha256:461f960bb1d29e719fb472fe9be9cf3deca17d984687957f54546a09047b0d3c
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        version: 0.5.3
      entries:
      - name: konveyor-operator.v0.5.3
        version: 0.5.3
      - name: konveyor-operator.v0.5.1
        version: 0.5.1
      - name: konveyor-operator.v0.5.0
        version: 0.5.0
      name: konveyor-0.5
    - currentCSV: konveyor-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:8ee143c473570efd27b6f93332dc272871e11c66cc20f381c3a1e6e5688abf68
          createdAt: "2025-03-05T21:14:52Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.6.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:8ee143c473570efd27b6f93332dc272871e11c66cc20f381c3a1e6e5688abf68
        - quay.io/konveyor/tackle2-hub@sha256:52a8b9be038fbfbc4e22172a4a8c7191d2b06fd8670e5df2ec442e3ce00d8c8a
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:0b5cbb322fbe96335522410376c2e8505a8302088c0e454f3e5e7e28ba7c96a6
        - quay.io/konveyor/java-external-provider@sha256:15148cde9a6fedef2207fe36f0ac3bd1ecdc6e3ac76f1cbec9eb97ffcf247231
        - quay.io/konveyor/tackle-keycloak-init@sha256:a3741269aa15edd2f9fb7df853a03a942d16a88eec114f3e54f82ef63b82adc4
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/postgresql-15-c9s@sha256:55f8014c1edafab8f9085228a0d207c4e98820c615be34b935ce98e942d388b1
        - quay.io/konveyor/tackle2-ui@sha256:12b0eda13a63508b9c13d4f9309bbfbe19d564627cf0de672307af1a6ec993fb
        - quay.io/konveyor/tackle2-addon-discovery@sha256:3eb6f395c451e2c9a867871b3984a25f9f8e6ddcfe388f37fbd6d46a61a70b1e
        - quay.io/konveyor/generic-external-provider@sha256:27c1731859a4a639d2750cff8d1246281f4b0466bb5b069169c3e170c3c19b30
        version: 0.6.1
      entries:
      - name: konveyor-operator.v0.6.1
        version: 0.6.1
      - name: konveyor-operator.v0.6.0
        version: 0.6.0
      name: konveyor-0.6
    - currentCSV: konveyor-operator.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:f5870bc26b2a07ddacf00ed5b5f7df7d1068489010a80bcc543686c6c9576fbb
          createdAt: "2025-07-10T15:16:17Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-ui@sha256:f87b04762fd163f40c49f3bcffd473f0c3f80a643576c6a3e2c2b4a14395c09f
        - quay.io/konveyor/java-external-provider@sha256:a0ef733e9fce3845d02b5360440add85427acad697d2157ef92242dec06bd73b
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/tackle-keycloak-init@sha256:dab4cb1229c8a729d14b4cecb4aa3b5c4d48e94c36e84d785c922fc08d120617
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:7ed52ea9b95650241b5c2c01914775e29be23f7ed75fabb6e7c71969f3f0728f
        - quay.io/konveyor/tackle2-addon-discovery@sha256:60951801aa539ee59cd43f3c5f42a2ca751ca70e995cbb258b7d55ebb24d8450
        - quay.io/konveyor/origin-oauth-proxy@sha256:2c811da843057119b246784f21dc7eec490669a8569e8295d8e914cdce555b24
        - quay.io/konveyor/generic-external-provider@sha256:facca8861559a658710ad3aef8b325731595bf944cce1907715d6efd33887072
        - quay.io/konveyor/tackle2-operator@sha256:f5870bc26b2a07ddacf00ed5b5f7df7d1068489010a80bcc543686c6c9576fbb
        - quay.io/konveyor/tackle2-hub@sha256:52f66607f3134a4165ce691c6e9ce992d93927e9591055daafdc1bfc1dd09dd9
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/postgresql-15-c9s@sha256:41460103a0bab189cd77acb6e741d9c5a6d3506479f7e56df40eb156d9b2cadd
        version: 0.7.1
      entries:
      - name: konveyor-operator.v0.7.1
        version: 0.7.1
      - name: konveyor-operator.v0.7.0
        version: 0.7.0
      name: konveyor-0.7
    - currentCSV: konveyor-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:68b6672fccdf54c65cc83f4c82db023053d6b61409bea880f58b53ccfea7b1ed
          createdAt: "2025-09-29T20:41:09Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-addon-discovery@sha256:96a5b37834fbed9506014bcbd909258081fd833ef9e16daa104effd0e3a3c6ea
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-addon-platform@sha256:d0280ade44ba5e15210c07342608576b38e6993af2e479740675a4706cd2c875
        - quay.io/konveyor/tackle2-ui@sha256:f3f7380be65ff1fb3459c37a3d4865354bab685dc836893f78221f16ba579f70
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:8c82ad880ca2a2d5cd4b08125feca16d13d014d15f4f77677306086d1067f9aa
        - quay.io/konveyor/tackle-keycloak-init@sha256:c0051b8c3a3c891fe12012623aee775080ecb47af8a4e75b8be8b0b4855d9f2c
        - quay.io/konveyor/postgresql-15-c9s@sha256:af275f35a307367dcee9a339a6903cf79e3a8a41cabf79fb7cea7015895f0b58
        - quay.io/konveyor/origin-oauth-proxy@sha256:8988f14db7dde3edcaed0541cecb0b4fd58fab9dbaffb79a6c1c4aee13b9bd73
        - quay.io/konveyor/java-external-provider@sha256:8eab08a04e98051f89352250d63523a5b591f35a6497f5d69a5a9d556a94abf9
        - quay.io/konveyor/tackle2-hub@sha256:b3e6db1ffc195143b8e5d54b60949256e154b0f8b667736512a084e015cdf2a7
        - quay.io/konveyor/kantra@sha256:a5b5591a51143574a6ce6678b141b9cd3e14328c6fad773412dde21b442fe364
        - quay.io/konveyor/generic-external-provider@sha256:c7b6618bfff81d73f1b455963107ffdaf19334e6552ce076c009476da3f37f21
        - quay.io/konveyor/kai-solution-server@sha256:bab6e224c50de5b44cc37ce0297a3b91255efc08c1ad0fccf14fef9e5182775f
        - quay.io/konveyor/tackle2-operator@sha256:68b6672fccdf54c65cc83f4c82db023053d6b61409bea880f58b53ccfea7b1ed
        version: 0.8.0
      entries:
      - name: konveyor-operator.v0.8.0
        version: 0.8.0
      name: konveyor-0.8
    - currentCSV: konveyor-operator.v0.7.0-rc.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
          createdAt: "2025-05-21T17:12:48Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.0-rc.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
        - quay.io/konveyor/tackle2-addon-discovery@sha256:89d2b0f8636897eb4dc1957a74a5b5041c0f3267a747adcf2bc6ca56b17b9937
        - quay.io/konveyor/tackle2-hub@sha256:bed834eb5488eef23b0d1e8af194b0646cf74c39c197e0d57acc5aed2299d9f5
        - quay.io/konveyor/tackle-keycloak-init@sha256:c3124971dc31be9201de2a6a1f6b2f886adf4c96731188dd5e750798cde3cfe7
        - quay.io/konveyor/tackle2-ui@sha256:afe9141366a23c072286517e9af5436d14fd1df775aaa2fa13c3911764515777
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/java-external-provider@sha256:61ab20e37916abfeeadde685336660f0e8bbf519403d1f7100d43c1e42a72fbd
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:46150725a2adbde8dfdc2250614d437a50e64bbf0b4562dabc73715a287de67b
        - quay.io/konveyor/generic-external-provider@sha256:de55f78939576c110eb6f084a7ca51338fedba699fbb5135f23392783a5d7b95
        - quay.io/konveyor/origin-oauth-proxy@sha256:6c2847fd9667938b1d1294d988038e885e43cd9fb378862dae3419ce4f21631f
        - quay.io/konveyor/postgresql-15-c9s@sha256:012cd6611aa9d3f4e0a82636d6817f381c19e3b0809199216fbe7fa3396a2ef5
        version: 0.7.0-rc.1
      entries:
      - name: konveyor-operator.v0.7.0-rc.1
        version: 0.7.0-rc.1
      - name: konveyor-operator.v0.6.0-rc.2
        version: 0.6.0-rc.2
      - name: konveyor-operator.v0.6.0-rc.1
        version: 0.6.0-rc.1
      - name: konveyor-operator.v0.5.0-rc.1
        version: 0.5.0-rc.1
      - name: konveyor-operator.v0.3.0-rc.4
        version: 0.3.0-rc.4
      - name: konveyor-operator.v0.3.0-rc.3
        version: 0.3.0-rc.3
      - name: konveyor-operator.v0.3.0-rc.2
        version: 0.3.0-rc.2
      - name: konveyor-operator.v0.3.0-rc.1
        version: 0.3.0-rc.1
      name: rc
    - currentCSV: konveyor-operator.v0.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:33cf01f19a7043bcc5bdb12a3ed6bde7a9094981aa2383554111ebec9b4ba498
          createdAt: "2025-03-26T20:25:41Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.6.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-ui@sha256:3c568c6047a044095080b6013b643a4041deae7b45936a443d4c4d708a81b945
        - quay.io/konveyor/tackle2-hub@sha256:0df49087d0c9341c2120a7c3aa2ba7907c056f94e1602ffd0496ca7113c75382
        - quay.io/konveyor/tackle-keycloak-init@sha256:e32fa48351f410361e8bc100303deeb40d8a66d7eefb0cf7b1bc16e067a9a795
        - quay.io/konveyor/tackle2-operator@sha256:33cf01f19a7043bcc5bdb12a3ed6bde7a9094981aa2383554111ebec9b4ba498
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/tackle2-addon-discovery@sha256:1b89266d6bdfa497ba8137e0d33ff53785ac476dd96bf4fd9a185ef0ddf513ce
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:734752be111a97e4b7e0030d563c32883f419ad031b2f350cb8e7333f2f37de3
        - quay.io/konveyor/generic-external-provider@sha256:bc1112352f1a2c3172c5e7d5e637b5c5a69bf7d98b78f494fa593e8872e5cfd1
        - quay.io/konveyor/java-external-provider@sha256:fe85689b57125d678df685012b93c8a567fcb9251777cd711e9b82bdb22a734c
        - quay.io/konveyor/postgresql-15-c9s@sha256:7957d3b272b9f4e469bdc0c514ec9e70e8fbdc34a7250b4f4c62f9bb73cbf08d
        version: 0.6.2
      entries:
      - name: konveyor-operator.v0.6.2
        version: 0.6.2
      name: release-0.6
    - currentCSV: konveyor-operator.v0.7.0-rc.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
          createdAt: "2025-05-21T17:12:48Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.0-rc.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:46150725a2adbde8dfdc2250614d437a50e64bbf0b4562dabc73715a287de67b
        - quay.io/konveyor/generic-external-provider@sha256:de55f78939576c110eb6f084a7ca51338fedba699fbb5135f23392783a5d7b95
        - quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
        - quay.io/konveyor/postgresql-15-c9s@sha256:012cd6611aa9d3f4e0a82636d6817f381c19e3b0809199216fbe7fa3396a2ef5
        - quay.io/konveyor/tackle-keycloak-init@sha256:c3124971dc31be9201de2a6a1f6b2f886adf4c96731188dd5e750798cde3cfe7
        - quay.io/konveyor/java-external-provider@sha256:61ab20e37916abfeeadde685336660f0e8bbf519403d1f7100d43c1e42a72fbd
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/tackle2-ui@sha256:afe9141366a23c072286517e9af5436d14fd1df775aaa2fa13c3911764515777
        - quay.io/konveyor/tackle2-hub@sha256:bed834eb5488eef23b0d1e8af194b0646cf74c39c197e0d57acc5aed2299d9f5
        - quay.io/konveyor/origin-oauth-proxy@sha256:6c2847fd9667938b1d1294d988038e885e43cd9fb378862dae3419ce4f21631f
        - quay.io/konveyor/tackle2-addon-discovery@sha256:89d2b0f8636897eb4dc1957a74a5b5041c0f3267a747adcf2bc6ca56b17b9937
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        version: 0.7.0-rc.1
      entries:
      - name: konveyor-operator.v0.7.0-rc.1
        version: 0.7.0-rc.1
      name: release-0.7
    defaultChannel: konveyor-0.8
    packageName: konveyor-operator
    provider:
      name: Konveyor
      url: https://www.konveyor.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: sriov-network-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sriov-network-operator.v4.13.0-202309190544
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovIBNetwork",
                "metadata": {
                  "name": "example-sriovibnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "linkState": "auto",
                  "networkNamespace": "default",
                  "resourceName": "mellanoxnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetwork",
                "metadata": {
                  "name": "example-sriovnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "intelnics",
                  "spoofChk": "on",
                  "trust": "off",
                  "vlan": 0
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodePolicy",
                "metadata": {
                  "name": "policy-1",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "deviceType": "vfio-pci",
                  "mtu": 9000,
                  "nicSelector": {
                    "pfNames": [
                      "eth1"
                    ],
                    "rootDevices": [
                      "0000:01:00.1"
                    ],
                    "vendor": "8086"
                  },
                  "nodeSelector": {
                    "feature.node.kubernetes.io/network-sriov.capable": "true"
                  },
                  "numVfs": 6,
                  "priority": 99,
                  "resourceName": "intelnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodeState",
                "metadata": {
                  "name": "minikube",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "interfaces": []
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkPoolConfig",
                "metadata": {
                  "name": "sriovnetworkpoolconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovOperatorConfig",
                "metadata": {
                  "name": "default",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "enableInjector": true,
                  "enableOperatorWebhook": true,
                  "logLevel": 2
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: quay.io/pliurh/sriov-network-operator-manager:bd1
          createdAt: "2023-08-02T09:22:34Z"
          description: An operator for configuring SR-IOV components and initializing
            SRIOV network devices in Openshift cluster.
          olm.skipRange: '>=4.3.0-0 <4.13.0-202309190544'
          operatorframework.io/suggested-namespace: openshift-sriov-network-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "cni"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/k8snetworkplumbingwg/sriov-network-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovIBNetwork is the Schema for the sriovibnetworks API
            displayName: Sriov IBNetwork
            kind: SriovIBNetwork
            name: sriovibnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodePolicy is the Schema for the sriovnetworknodepolicies
              API
            displayName: Sriov Network Node Policy
            kind: SriovNetworkNodePolicy
            name: sriovnetworknodepolicies.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodeState is the Schema for the sriovnetworknodestates
              API
            displayName: Sriov Network Node State
            kind: SriovNetworkNodeState
            name: sriovnetworknodestates.sriovnetwork.openshift.io
            version: v1
          - kind: SriovNetworkPoolConfig
            name: sriovnetworkpoolconfigs.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetwork is the Schema for the sriovnetworks API
            displayName: Sriov Network
            kind: SriovNetwork
            name: sriovnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovOperatorConfig is the Schema for the sriovoperatorconfigs
              API
            displayName: Sriov Operator Config
            kind: SriovOperatorConfig
            name: sriovoperatorconfigs.sriovnetwork.openshift.io
            version: v1
        description: |
          # SR-IOV Network Operator for Openshift

          ## Introduction
          The sriov-network-operator is generally responsible for configuring the sriov components in a openshift cluster.

          ### Supported Features
          * Initialize the SR-IOV NICs on nodes.
          * provision SR-IOV device plugin on selected node.
          * provision SR-IOV CNI plugin on selected nodes.
          * manage configuration of SR-IOV device plugin.
          * generate net-att-def CRs for SR-IOV CNI plugin.

          This operator has to run in namespace 'openshift-sriov-network-operator'. An Operator Group is also required to install this operator:

          ```
          $ oc create namespace openshift-sriov-network-operator

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: sriov-network-operators
            namespace: openshift-sriov-network-operator
          spec:
            targetNamespaces:
            - openshift-sriov-network-operator
          EOF
          ```
        displayName: SR-IOV Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - network
        - sr-iov
        - sriov
        links:
        - name: Source Code
          url: https://github.com/k8snetworkplumbingwg/sriov-network-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-sriov-network-config-daemon@sha256:a2053c5366eb89112dfbc7a9994685fc198cce6fa87fdd6d1c0a07727e6eee03
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin@sha256:82d4846f8dfb9e0d02e7ff3f532f8f2bf44c9aba63c37d5ce2efb36969e5c453
        - registry.redhat.io/openshift4/ose-sriov-network-operator@sha256:3314ee580113b985e275583585b1f2adf381c5a0885b51100b3a01872a77a555
        - registry.redhat.io/openshift4/ose-sriov-network-webhook@sha256:bfb84d00dbb6b18d596ca2720b52d922c589249d2f988679189ea0eea985e58a
        - registry.redhat.io/openshift4/sriov-cni-rhel9@sha256:a4f0ec210cd065cb7f77dd25257aadf8345f47819cc728c5b6229acb79a09bb3
        - registry.redhat.io/openshift4/ose-sriov-dp-admission-controller@sha256:8dfaa07a5775a7156d303019879e41b17833e5643794f6e1087881d2a4dd1781
        - registry.redhat.io/openshift4/ose-sriov-infiniband-cni@sha256:716aa0b9ecfddfea0a62615f424f4abeab5362899c695e9ecc9df4df05f1add5
        version: 4.13.0-202309190544
      entries:
      - name: sriov-network-operator.v4.13.0-202309190544
        version: 4.13.0-202309190544
      name: "4.13"
    - currentCSV: sriov-network-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovIBNetwork",
                "metadata": {
                  "name": "example-sriovibnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "linkState": "auto",
                  "networkNamespace": "default",
                  "resourceName": "mellanoxnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetwork",
                "metadata": {
                  "name": "example-sriovnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "intelnics",
                  "spoofChk": "on",
                  "trust": "off",
                  "vlan": 0
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodePolicy",
                "metadata": {
                  "name": "policy-1",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "deviceType": "vfio-pci",
                  "mtu": 9000,
                  "nicSelector": {
                    "pfNames": [
                      "eth1"
                    ],
                    "rootDevices": [
                      "0000:01:00.1"
                    ],
                    "vendor": "8086"
                  },
                  "nodeSelector": {
                    "feature.node.kubernetes.io/network-sriov.capable": "true"
                  },
                  "numVfs": 6,
                  "priority": 99,
                  "resourceName": "intelnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodeState",
                "metadata": {
                  "name": "minikube",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "interfaces": []
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkPoolConfig",
                "metadata": {
                  "name": "sriovnetworkpoolconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovOperatorConfig",
                "metadata": {
                  "name": "default",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "enableInjector": true,
                  "enableOperatorWebhook": true,
                  "logLevel": 2
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-sriov-network-rhel9-operator@sha256:8dcae2a77e904713fba9c5907789f9b22ffb1f941f793adc4b604798d29a44f2
          createdAt: "2024-05-07T23:50:52Z"
          description: An operator for configuring SR-IOV components and initializing
            SRIOV network devices in Openshift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.16.0-202509240123'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "sriovnetwork.openshift.io/v1",
              "kind": "SriovOperatorConfig",
              "metadata": {
                "name": "default"
              },
              "spec": {
                "configurationMode": "daemon",
                "disableDrain": false,
                "enableInjector": true,
                "enableOperatorWebhook": true,
                "logLevel": 2
              }
            }
          operatorframework.io/suggested-namespace: openshift-sriov-network-operator
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-sriov-network-operator",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["sriovnetworknodestates.sriovnetwork.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/k8snetworkplumbingwg/sriov-network-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovIBNetwork is the Schema for the sriovibnetworks API
            displayName: Sriov IBNetwork
            kind: SriovIBNetwork
            name: sriovibnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodePolicy is the Schema for the sriovnetworknodepolicies
              API
            displayName: Sriov Network Node Policy
            kind: SriovNetworkNodePolicy
            name: sriovnetworknodepolicies.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodeState is the Schema for the sriovnetworknodestates
              API
            displayName: Sriov Network Node State
            kind: SriovNetworkNodeState
            name: sriovnetworknodestates.sriovnetwork.openshift.io
            version: v1
          - kind: SriovNetworkPoolConfig
            name: sriovnetworkpoolconfigs.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetwork is the Schema for the sriovnetworks API
            displayName: Sriov Network
            kind: SriovNetwork
            name: sriovnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovOperatorConfig is the Schema for the sriovoperatorconfigs
              API
            displayName: Sriov Operator Config
            kind: SriovOperatorConfig
            name: sriovoperatorconfigs.sriovnetwork.openshift.io
            version: v1
        description: |
          # SR-IOV Network Operator for Openshift

          ## Introduction
          The sriov-network-operator is generally responsible for configuring the sriov components in a openshift cluster.

          ### Supported Features
          * Initialize the SR-IOV NICs on nodes.
          * provision SR-IOV device plugin on selected node.
          * provision SR-IOV CNI plugin on selected nodes.
          * manage configuration of SR-IOV device plugin.
          * generate net-att-def CRs for SR-IOV CNI plugin.

          This operator has to run in namespace 'openshift-sriov-network-operator'. An Operator Group is also required to install this operator:

          ```
          $ oc create namespace openshift-sriov-network-operator

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: sriov-network-operators
            namespace: openshift-sriov-network-operator
          spec:
            targetNamespaces:
            - openshift-sriov-network-operator
          EOF
          ```
        displayName: SR-IOV Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - network
        - sr-iov
        - sriov
        links:
        - name: Source Code
          url: https://github.com/k8snetworkplumbingwg/sriov-network-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-sriov-network-config-daemon-rhel9@sha256:61afbb271a08dc3ab3b5509df535d9456f9b2c1b267554386b4bf859b6bccbe4
        - registry.redhat.io/openshift4/ose-sriov-infiniband-cni-rhel9@sha256:f828e415db73bf2df3fa7b26238a2f5c8a53cda6b396c11a0337d688108e6647
        - registry.redhat.io/openshift4/sriov-cni-rhel9@sha256:cac67ac25a0d319b5ea68cb3f63eac058339f1981811c7a3466b9daa2534a38d
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin-rhel9@sha256:5b3bc02e31eecd0d3468db729b03239d518ef73112b0fdb5ec3f05a788baa0be
        - registry.redhat.io/openshift4/ose-sriov-network-rhel9-operator@sha256:8dcae2a77e904713fba9c5907789f9b22ffb1f941f793adc4b604798d29a44f2
        - registry.redhat.io/openshift4/ose-sriov-network-webhook-rhel9@sha256:a6a5cc76d88b6b7aa86ee24b882ce43be33c8c6d4bcfae5f1b5490ead71ce8b6
        - registry.redhat.io/openshift4/ose-sriov-network-metrics-exporter-rhel9@sha256:a0580d8156de0ca1e5a119441196e9babbeb4d7edc502659c7ac66a49b1928d5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-sriov-dp-admission-controller-rhel9@sha256:20fd8a0081ffdea05796c8f6a0539a0e246da24f739f7bbf0722c90acfdeed15
        version: 4.16.0-202509240123
      entries:
      - name: sriov-network-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: sriov-network-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: sriov-network-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: sriov-network-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: sriov-network-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: sriov-network-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: sriov-network-operator.v4.16.0-202504290837
        version: 4.16.0-202504290837
      - name: sriov-network-operator.v4.16.0-202504151638
        version: 4.16.0-202504151638
      - name: sriov-network-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: sriov-network-operator.v4.16.0-202502262005
        version: 4.16.0-202502262005
      - name: sriov-network-operator.v4.16.0-202502171404
        version: 4.16.0-202502171404
      - name: sriov-network-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: sriov-network-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: sriov-network-operator.v4.16.0-202501081506
        version: 4.16.0-202501081506
      - name: sriov-network-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: sriov-network-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: sriov-network-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: sriov-network-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: sriov-network-operator.v4.16.0-202410300306
        version: 4.16.0-202410300306
      - name: sriov-network-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: sriov-network-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: sriov-network-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: sriov-network-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: sriov-network-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: sriov-network-operator.v4.16.0-202409111535
        version: 4.16.0-202409111535
      - name: sriov-network-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: sriov-network-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: sriov-network-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: sriov-network-operator.v4.16.0-202408120337
        version: 4.16.0-202408120337
      - name: sriov-network-operator.v4.16.0-202407100107
        version: 4.16.0-202407100107
      - name: sriov-network-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: sriov-network-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: sriov-network-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: sriov-network-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: sriov-network-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: klusterlet
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: klusterlet.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "Klusterlet",
                "metadata": {
                  "name": "klusterlet"
                },
                "spec": {
                  "clusterName": "cluster1",
                  "deployOption": {
                    "mode": "Singleton"
                  },
                  "externalServerURLs": [
                    {
                      "url": "https://localhost"
                    }
                  ],
                  "imagePullSpec": "quay.io/open-cluster-management/registration-operator:v0.13.1",
                  "namespace": "open-cluster-management-agent",
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/registration-operator:v0.13.1
          createdAt: "2024-04-10T15:27:49Z"
          description: Manages the installation and upgrade of the Klusterlet.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management-io/ocm
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Klusterlet defines the configurations of agents running on
              the managed cluster
            displayName: Klusterlet
            kind: Klusterlet
            name: klusterlets.operator.open-cluster-management.io
            version: v1
        description: "Note: This operator will stop updating, please use the latest
          operator in OperatorHub.io https://operatorhub.io/operator/klusterlet. \n\nThe
          Klusterlet proivdes the registation to the Hub clusters as a managed cluster.
          This operator supports the installation and upgrade of Klusterlet.\n\n##
          Prerequisites\nYou need a Hub cluster which has installed ClusterManager.\n\n##
          How to Install\n* Install Klusterlet operator by following instructions
          in top right button `Install`.\n* Optional: Create an agent namespace used
          to deploy Klusterlet. The default namespace is `open-cluster-management-agent`.\n*
          Create a secret `bootstrap-hub-kubeconfig` with the kubeconfig file of hub
          cluster in the agent namespace.\n\n  ```\n  $ kubectl create secret generic
          bootstrap-hub-kubeconfig --from-file=kubeconfig=<kubeconfig file of hub
          clsuter> -n <agent namespace>\n  ```\n\n* Create an instance of the Klusterlet
          resource to install the Klusterlet.\n\nThe pods of Klusterlet are created
          in agent namespace.\n\n```\n$ kubectl get pods -n open-cluster-management-agent\n
          \ NAME                                READY   STATUS    RESTARTS   AGE\n
          \ klusterlet-agent-55c6574876-nfxm9   1/1     Running   0          8m38s\n```\n\nThere
          is a ManagedCluster created in the Hub cluster.\n\n```\n$ kubectl get managedclusters.cluster.open-cluster-management.io\nNAME
          \                  HUB ACCEPTED   MANAGED CLUSTER URLS     JOINED   AVAILABLE
          \  AGE\ncluster1               false                                                        27s\n```\n"
        displayName: Klusterlet
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open-cluster-management
        - Klusterlet
        links:
        - name: Registration Operator
          url: https://github.com/open-cluster-management-io/ocm
        maintainers:
        - email: zyin@redhat.com
          name: Zhiwei Yin
        - email: deads@redhat.com
          name: David Eads
        - email: jqiu@redhat.com
          name: Jian Qiu
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/open-cluster-management/registration-operator:v0.13.1
        version: 0.13.1
      entries:
      - name: klusterlet.v0.13.1
        version: 0.13.1
      - name: klusterlet.v0.13.0
        version: 0.13.0
      - name: klusterlet.v0.12.0
        version: 0.12.0
      - name: klusterlet.v0.11.0
        version: 0.11.0
      - name: klusterlet.v0.10.0
        version: 0.10.0
      - name: klusterlet.v0.9.1
        version: 0.9.1
      - name: klusterlet.v0.8.0
        version: 0.8.0
      - name: klusterlet.v0.7.0
        version: 0.7.0
      - name: klusterlet.v0.6.0
        version: 0.6.0
      - name: klusterlet.v0.5.0
        version: 0.5.0
      - name: klusterlet.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: klusterlet
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.12
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-06-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:9aed98a061b2c76f3c512c1d6453df5920f0c779b33d7168c1e7ab9d8bc928a4
        version: 0.0.12
      entries:
      - name: joget-dx8-openshift-operator.v0.0.12
        version: 0.0.12
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Dell Technologies
      provider-url: https://github.com/dell/csm-operator
    name: dell-csm-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dell-csm-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "authorization",
                  "namespace": "authorization"
                },
                "spec": {
                  "modules": [
                    {
                      "components": [
                        {
                          "enabled": true,
                          "name": "nginx"
                        },
                        {
                          "enabled": true,
                          "name": "cert-manager"
                        },
                        {
                          "authorizationController": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231",
                          "authorizationControllerReplicas": 1,
                          "certificate": "",
                          "controllerReconcileInterval": "5m",
                          "enabled": true,
                          "hostname": "csm-authorization.com",
                          "leaderElection": true,
                          "name": "proxy-server",
                          "opa": "docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095",
                          "opaKubeMgmt": "docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517",
                          "openTelemetryCollectorAddress": "",
                          "privateKey": "",
                          "proxyServerIngress": [
                            {
                              "annotations": {},
                              "hosts": [],
                              "ingressClassName": "nginx"
                            }
                          ],
                          "proxyService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682",
                          "proxyServiceReplicas": 1,
                          "roleService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c",
                          "roleServiceReplicas": 1,
                          "storageService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5",
                          "storageServiceReplicas": 1,
                          "tenantService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358",
                          "tenantServiceReplicas": 1
                        },
                        {
                          "commander": "docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a",
                          "name": "redis",
                          "redis": "docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f",
                          "redisCommander": "rediscommander",
                          "redisName": "redis-csm",
                          "redisReplicas": 5,
                          "redisSecretProviderClass": [
                            {
                              "redisPasswordKey": "",
                              "redisSecretName": "",
                              "redisUsernameKey": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "sentinel": "sentinel"
                        },
                        {
                          "configSecretProviderClass": [
                            {
                              "configSecretName": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "name": "config"
                        },
                        {
                          "name": "storage-system-credentials",
                          "secretProviderClasses": {
                            "vault": [
                              "secret-provider-class-1",
                              "secret-provider-class-2"
                            ]
                          }
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": true,
                      "forceRemoveModule": true,
                      "name": "authorization-proxy-server"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "isilon",
                  "namespace": "isilon"
                },
                "spec": {
                  "driver": {
                    "authSecret": "isilon-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VERBOSE",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_ISI_PORT",
                          "value": "8080"
                        },
                        {
                          "name": "X_CSI_ISI_PATH",
                          "value": "/ifs/data/csi"
                        },
                        {
                          "name": "X_CSI_ISI_NO_PROBE_ON_START",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_AUTOPROBE",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_AUTH_TYPE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_CUSTOM_TOPOLOGY_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOISILON_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "AZ_RECONCILE_INTERVAL",
                          "value": "1h"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_ISI_QUOTA_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_ACCESS_ZONE",
                          "value": "System"
                        },
                        {
                          "name": "X_CSI_ISI_VOLUME_PATH_PERMISSIONS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_IGNORE_UNRESOLVABLE_HOSTS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "isilon",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerscale"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSCALE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_PERFORMANCE_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSCALE_QUOTA_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "ISICLIENT_INSECURE",
                              "value": "true"
                            },
                            {
                              "name": "ISICLIENT_AUTH_TYPE",
                              "value": "1"
                            },
                            {
                              "name": "ISICLIENT_VERBOSE",
                              "value": "0"
                            },
                            {
                              "name": "POWERSCALE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSCALE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0",
                          "name": "metrics-powerscale"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-isilon/csi_sock",
                            "--mode=node",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powermax",
                  "namespace": "powermax"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powermax-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_MANAGED_ARRAYS",
                          "value": "000000000001,000000000002"
                        },
                        {
                          "name": "X_CSI_POWERMAX_PORTGROUPS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_TRANSPORT_PROTOCOL",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_K8S_CLUSTER_PREFIX",
                          "value": "CSM"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "X_CSI_VSPHERE_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VSPHERE_PORTGROUP",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VSPHERE_HOSTNAME",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VCENTER_HOST",
                          "value": ""
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "X_CSI_POWERMAX_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "X_CSI_POWERMAX_DEBUG",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_REVPROXY_USE_SECRET",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powermax",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERMAX_ISCSI_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_TOPOLOGY_CONTROL_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": [
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/memory-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/disk-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/network-unavailable",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REVPROXY_TLS_SECRET",
                              "value": "csirevproxy-tls-secret"
                            },
                            {
                              "name": "X_CSI_REVPROXY_PORT",
                              "value": "2222"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            },
                            {
                              "name": "DeployAsSidecar",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265",
                          "name": "csipowermax-reverseproxy"
                        }
                      ],
                      "configVersion": "v2.14.0",
                      "name": "csireverseproxy"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powermax"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERMAX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERMAX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414",
                          "name": "metrics-powermax"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/powermax.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powerstore",
                  "namespace": "powerstore"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powerstore-config",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_NODE_NAME_PREFIX",
                          "value": "csi-node"
                        },
                        {
                          "name": "X_CSI_FC_PORTS_FILTER_FILE_PATH",
                          "value": "/etc/fc-ports-filter"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOPOWERSTORE_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSM_AUTH_ENABLED",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_MULTI_NAS_FAILURE_THRESHOLD",
                          "value": "5"
                        },
                        {
                          "name": "X_CSI_MULTI_NAS_COOLDOWN_PERIOD",
                          "value": "5m"
                        },
                        {
                          "name": "X_CSI_NFS_ACLS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerstore",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-powerstore.dellemc.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerstore"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSTORE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_SPACE_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_ARRAY_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_FILE_SYSTEM_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSTORE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSTORE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "ZIPKIN_URI",
                              "value": ""
                            },
                            {
                              "name": "ZIPKIN_SERVICE_NAME",
                              "value": "metrics-powerstore"
                            },
                            {
                              "name": "ZIPKIN_PROBABILITY",
                              "value": "0.0"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60",
                          "name": "metrics-powerstore"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "unity",
                  "namespace": "unity"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_UNITY_ALLOW_MULTI_POD_ACCESS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_EPHEMERAL_STAGING_PATH",
                          "value": "/var/lib/kubelet/plugins/kubernetes.io/csi/pv/"
                        },
                        {
                          "name": "X_CSI_ISCSI_CHROOT",
                          "value": "/noderoot"
                        },
                        {
                          "name": "X_CSI_UNITY_SYNC_NODEINFO_INTERVAL",
                          "value": "15"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOUNITY_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOUNITY_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "TENANT_NAME",
                          "value": ""
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_UNITY_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "unity",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "vxflexos",
                  "namespace": "vxflexos"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLELISTVOLUMESNAPSHOT",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLESNAPSHOTCGDELETE",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "INFO"
                        },
                        {
                          "name": "GOSCALEIO_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOSCALEIO_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_QUOTA_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "INTERFACE_NAMES",
                          "value": null
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERFLEX_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "File",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerflex",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "initContainers": [
                      {
                        "envs": [
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          },
                          {
                            "name": "REPO_ADDRESS",
                            "value": "sftp://0.0.0.0"
                          },
                          {
                            "name": "REPO_USER",
                            "value": "sdcRepoUser"
                          },
                          {
                            "name": "MODULE_SIGCHECK",
                            "value": "0"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "imagePullPolicy": "IfNotPresent",
                        "name": "sdc"
                      }
                    ],
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_SDC_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_APPROVE_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_PREFIX",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_SDC_SFTP_REPO_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 1,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "enabled": false,
                        "envs": [
                          {
                            "name": "HOST_PID",
                            "value": "1"
                          },
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "name": "sdc-monitor"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "csi-external-health-monitor-controller"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERFLEX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_SDC_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_SDC_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERFLEX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d",
                          "name": "metrics-powerflex"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerflex"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--csisock=unix:/var/lib/kubelet/plugins/vxflexos.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
          createdAt: "2025-09-05T13:18:42Z"
          description: Easily install and manage Dell’s CSI Drivers and CSM
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/dell/csm-operator
          support: Dell Technologies
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ContainerStorageModule is the Schema for the containerstoragemodules
              API
            displayName: Container Storage Module
            kind: ContainerStorageModule
            name: containerstoragemodules.storage.dell.com
            version: v1
        description: "Dell Container Storage Modules (CSM) Operator is a Kubernetes
          Operator which can be \nused to install and manage Dell’s CSI drivers and
          CSM modules. \nBy using Dell CSM Operator, enterprises can quickly and easily
          deploy the \nCSM modules for CSI drivers making it easy for DevOps \nteams
          to build and optimize their Kubernetes workloads using enterprise grade
          storage \nwhile consistently aligning with the desired outcomes of the organization.\nThis
          operator is available as both community and certified operator for Openshift
          \nand as a community operator for upstream Kubernetes.\nOperator can be
          deployed via OperatorHub.io. using Operator Lifecycle Manager (OLM)\n\n###
          About Dell CSM\n\nDell CSM brings powerful enterprise storage features and
          functionality to Kubernetes for \neasier adoption of cloud native workloads,
          improved productivity and scalable operations. \n\nCSM delivers software
          modules for storage management that provide developers access to \nbuild
          automation for enhanced IT needs and other critical enterprise storage features.
          \n\nCurrently, the following modules are available as part of Dell CSM.
          \n*\tObservability - Delivers a high-level view of the storage capacity
          \n  and performance usage via Grafana dashboards to the Kubernetes users\n*\tReplication
          - Enables array replication capabilities for Kubernetes users\n*\tAuthorization
          - Provides both storage and Kubernetes administrators the ability \n  to
          apply RBAC for Dell CSI Drivers\n*\tResiliency - Designed to make Kubernetes
          applications more resilient to \n  node failures\n\nThese modules are available
          for various storage platforms in the Dell portfolio, \nsuch as Dell PowerStore,
          Dell PowerMax, Dell PowerFlex, Dell PowerScale, and Dell Unity XT.\n\nFor
          more information, visit [Dell’s Container Storage Module web page](https://www.dell.com/en-us/dt/solutions/kubernetes-containers/kubernetes-storage.htm)\n\nAdditional
          Materials\n*\t[Documentation](https://dell.github.io/csm-docs)\n*\t[CSM
          Solution Brief](https://www.delltechnologies.com/asset/en-us/products/storage/briefs-summaries/csm-solution-brief.pdf)\n"
        displayName: Dell Container Storage Modules
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dell Container Storage Modules
        - Dell CSI Driver
        - Dell CSM Modules
        - Powerflex
        - Powerscale
        - Powerstore
        - Unity
        - Authorization
        - Observability
        - Replication
        links:
        - name: Documentation
          url: https://dell.github.io/csm-docs
        maintainers:
        - email: container.storage.modules@dell.com
          name: Dell Container Storage Modules
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Dell Technologies
          url: https://github.com/dell/csm-operator
        relatedImages:
        - docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a
        - registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
        - registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3
        - registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231
        - registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0
        - registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:59335acaca515fceed5f0350c46cfcfd6037475f0cf35ae8ab41d37333039e2e
        - ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60
        - docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682
        - registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16
        - registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265
        - registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c
        - quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358
        - quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403
        - docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517
        version: 1.10.0
      entries:
      - name: dell-csm-operator-certified.v1.10.0
        version: 1.10.0
      - name: dell-csm-operator-certified.v1.9.1
        version: 1.9.1
      - name: dell-csm-operator-certified.v1.9.0
        version: 1.9.0
      - name: dell-csm-operator-certified.v1.8.1
        version: 1.8.1
      - name: dell-csm-operator-certified.v1.8.0
        version: 1.8.0
      - name: dell-csm-operator-certified.v1.7.0
        version: 1.7.0
      - name: dell-csm-operator-certified.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: dell-csm-operator-certified
    provider:
      name: Dell Technologies
      url: https://github.com/dell/csm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: external-dns-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: external-dns-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "provider": {
                    "type": "AWS",
                    "aws": {
                      "credentials": {
                        "name": "aws-access-key"
                      }
                    }
                  },
                  "source": {
                    "type": "Service"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:6be1403e418f9b63ec8ebcc8858e602409bc41fdee4586acc627623a78b50eb5
          createdAt: 2022-03-01T21:00:46
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## Prerequisites and Requirements
           The ExternalDNS Operator has minimal cluster level permissions. Unfortunately OLM doesn't allow the local role creation in the operand namespace yet.
           Therefore make sure the following commands are run against your cluster before the installation of the operator:
           * Create the operand namespace: `oc create ns external-dns`
           * Grant the operator access to manage operand resources: `oc apply -f https://raw.githubusercontent.com/openshift/external-dns-operator/release-0.1/config/rbac/extra-roles.yaml`
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b6fde43491c2e0733c8fe6ee8660f7f6435fe20c0b6d435cfc7f86eeab8b4748
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:6be1403e418f9b63ec8ebcc8858e602409bc41fdee4586acc627623a78b50eb5
        - registry.redhat.io/edo/external-dns-rhel8@sha256:c1134bb46172997ef7278b6cefbb0da44e72a9f808a7cd67b3c65d464754cab9
        version: 0.1.2
      entries:
      - name: external-dns-operator.v0.1.2
        version: 0.1.2
      name: alpha
    - currentCSV: external-dns-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
          createdAt: 2023-11-15T09:24:04
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e33874a51971a90917cf437fc6cbeea1da569ba07217f66f1bba96d594f58aed
        - registry.redhat.io/edo/external-dns-rhel8@sha256:6263a78b2a170e56dfca2173f14dbf14f0a5bddb6ea600bfdef248c7d5f20e4b
        version: 1.2.0
      entries:
      - name: external-dns-operator.v1.2.0
        version: 1.2.0
      name: stable-v1
    - currentCSV: external-dns-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:0877eb0a111efe0bff63173a2dfeb7a06a61afd0efbcdba053322a539e0dc667
          createdAt: 2024-01-22T09:26:37
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/edo/external-dns-rhel8@sha256:17f4127cbb9a9304769e9655c19cbea2260107c7e9371a99a9aa3e85d759c00b
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:0877eb0a111efe0bff63173a2dfeb7a06a61afd0efbcdba053322a539e0dc667
        version: 1.0.1
      entries:
      - name: external-dns-operator.v1.0.1
        version: 1.0.1
      - name: external-dns-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: external-dns-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:9ca751e84b8843cc62dc60550e5c4af741bf09a381ef62220e7eb3760ffe8524
          createdAt: 2024-01-22T09:56:14
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:9ca751e84b8843cc62dc60550e5c4af741bf09a381ef62220e7eb3760ffe8524
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/edo/external-dns-rhel8@sha256:ffcc3f2e9b9feb889e7882465f7f034aba6b9ee0699a1b6c24e83f501a63ecca
        version: 1.1.1
      entries:
      - name: external-dns-operator.v1.1.1
        version: 1.1.1
      - name: external-dns-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: external-dns-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
          createdAt: 2023-11-15T09:24:04
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e33874a51971a90917cf437fc6cbeea1da569ba07217f66f1bba96d594f58aed
        - registry.redhat.io/edo/external-dns-rhel8@sha256:6263a78b2a170e56dfca2173f14dbf14f0a5bddb6ea600bfdef248c7d5f20e4b
        version: 1.2.0
      entries:
      - name: external-dns-operator.v1.2.0
        version: 1.2.0
      name: stable-v1.2
    defaultChannel: stable-v1
    packageName: external-dns-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: rook-ceph-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rook-ceph-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephCluster",
                "metadata": {
                  "name": "my-rook-ceph",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "cephVersion": {
                    "image": "registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d"
                  },
                  "dataDirHostPath": "/var/lib/rook",
                  "mon": {
                    "count": 3
                  },
                  "dashboard": {
                    "enabled": true
                  },
                  "network": {
                    "hostNetwork": false
                  },
                  "rbdMirroring": {
                    "workers": 0
                  },
                  "storage": {
                    "useAllNodes": true,
                    "useAllDevices": true
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephBlockPool",
                "metadata": {
                  "name": "replicapool",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "failureDomain": "host",
                  "replicated": {
                    "size": 3
                  },
                  "annotations": null
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStore",
                "metadata": {
                  "name": "my-store",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "metadataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "dataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "gateway": {
                    "type": "s3",
                    "sslCertificateRef": null,
                    "port": 8080,
                    "securePort": null,
                    "instances": 1,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStoreUser",
                "metadata": {
                  "name": "my-user",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "store": "my-store",
                  "displayName": "my display name"
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephNFS",
                "metadata": {
                  "name": "my-nfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "rados": {
                    "pool": "myfs-data0",
                    "namespace": "nfs-ns"
                  },
                  "server": {
                    "active": 3,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephClient",
                "metadata": {
                  "name": "cinder",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "caps": {
                    "mon": "profile rbd",
                    "osd": "profile rbd pool=volumes, profile rbd pool=vms, profile rbd-read-only pool=images"
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephFilesystem",
                "metadata": {
                  "name": "myfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "dataPools": [
                    {
                      "compressionMode": "",
                      "crushRoot": "",
                      "deviceClass": "",
                      "erasureCoded": {
                        "algorithm": "",
                        "codingChunks": 0,
                        "dataChunks": 0
                      },
                      "failureDomain": "host",
                      "replicated": {
                        "requireSafeReplicaSize": false,
                        "size": 1,
                        "targetSizeRatio": 0.5
                      }
                    }
                  ],
                  "metadataPool": {
                    "compressionMode": "",
                    "crushRoot": "",
                    "deviceClass": "",
                    "erasureCoded": {
                      "algorithm": "",
                      "codingChunks": 0,
                      "dataChunks": 0
                    },
                    "failureDomain": "",
                    "replicated": {
                      "requireSafeReplicaSize": false,
                      "size": 1,
                      "targetSizeRatio": 0
                    }
                  },
                  "metadataServer": {
                    "activeCount": 1,
                    "activeStandby": true,
                    "placement": {},
                    "resources": {}
                  },
                  "preservePoolsOnDelete": false,
                  "preserveFilesystemOnDelete": false
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephRBDMirror",
                "metadata": {
                  "name": "my-rbd-mirror",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "annotations": null,
                  "count": 1,
                  "placement": {
                    "topologyKey": "kubernetes.io/hostname"
                  },
                  "resources": null
                }
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:cf7f72469d25fb37301a8159406613673f8968c82af1035fd45a470e0ae8acfa
          externalClusterScript: |-
            IiIiCkNvcHlyaWdodCAyMDIwIFRoZSBSb29rIEF1dGhvcnMuIEFsbCByaWdodHMgcmVzZXJ2ZWQu
            CgpMaWNlbnNlZCB1bmRlciB0aGUgQXBhY2hlIExpY2Vuc2UsIFZlcnNpb24gMi4wICh0aGUgIkxp
            Y2Vuc2UiKTsKeW91IG1heSBub3QgdXNlIHRoaXMgZmlsZSBleGNlcHQgaW4gY29tcGxpYW5jZSB3
            aXRoIHRoZSBMaWNlbnNlLgpZb3UgbWF5IG9idGFpbiBhIGNvcHkgb2YgdGhlIExpY2Vuc2UgYXQK
            CglodHRwOi8vd3d3LmFwYWNoZS5vcmcvbGljZW5zZXMvTElDRU5TRS0yLjAKClVubGVzcyByZXF1
            aXJlZCBieSBhcHBsaWNhYmxlIGxhdyBvciBhZ3JlZWQgdG8gaW4gd3JpdGluZywgc29mdHdhcmUK
            ZGlzdHJpYnV0ZWQgdW5kZXIgdGhlIExpY2Vuc2UgaXMgZGlzdHJpYnV0ZWQgb24gYW4gIkFTIElT
            IiBCQVNJUywKV0lUSE9VVCBXQVJSQU5USUVTIE9SIENPTkRJVElPTlMgT0YgQU5ZIEtJTkQsIGVp
            dGhlciBleHByZXNzIG9yIGltcGxpZWQuClNlZSB0aGUgTGljZW5zZSBmb3IgdGhlIHNwZWNpZmlj
            IGxhbmd1YWdlIGdvdmVybmluZyBwZXJtaXNzaW9ucyBhbmQKbGltaXRhdGlvbnMgdW5kZXIgdGhl
            IExpY2Vuc2UuCiIiIgoKaW1wb3J0IGVycm5vCmltcG9ydCBzeXMKaW1wb3J0IGpzb24KaW1wb3J0
            IGFyZ3BhcnNlCmltcG9ydCByZQppbXBvcnQgc3VicHJvY2VzcwppbXBvcnQgaG1hYwpmcm9tIGhh
            c2hsaWIgaW1wb3J0IHNoYTEgYXMgc2hhCmZyb20gb3MgaW1wb3J0IGxpbmVzZXAgYXMgTElORVNF
            UApmcm9tIG9zIGltcG9ydCBwYXRoCmZyb20gZW1haWwudXRpbHMgaW1wb3J0IGZvcm1hdGRhdGUK
            aW1wb3J0IHJlcXVlc3RzCmZyb20gcmVxdWVzdHMuYXV0aCBpbXBvcnQgQXV0aEJhc2UKCnB5M2sg
            PSBGYWxzZQppZiBzeXMudmVyc2lvbl9pbmZvLm1ham9yID49IDM6CiAgICBweTNrID0gVHJ1ZQog
            ICAgaW1wb3J0IHVybGxpYi5wYXJzZQogICAgZnJvbSBpcGFkZHJlc3MgaW1wb3J0IGlwX2FkZHJl
            c3MsIElQdjRBZGRyZXNzCgpNb2R1bGVOb3RGb3VuZEVycm9yID0gSW1wb3J0RXJyb3IKCnRyeToK
            ICAgIGltcG9ydCByYWRvcwpleGNlcHQgTW9kdWxlTm90Rm91bmRFcnJvciBhcyBub01vZEVycjoK
            ICAgIHByaW50KGYiRXJyb3I6IHtub01vZEVycn1cbkV4aXRpbmcgdGhlIHNjcmlwdC4uLiIpCiAg
            ICBzeXMuZXhpdCgxKQoKdHJ5OgogICAgaW1wb3J0IHJiZApleGNlcHQgTW9kdWxlTm90Rm91bmRF
            cnJvciBhcyBub01vZEVycjoKICAgIHByaW50KGYiRXJyb3I6IHtub01vZEVycn1cbkV4aXRpbmcg
            dGhlIHNjcmlwdC4uLiIpCiAgICBzeXMuZXhpdCgxKQoKdHJ5OgogICAgIyBmb3IgMi43LngKICAg
            IGZyb20gU3RyaW5nSU8gaW1wb3J0IFN0cmluZ0lPCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9y
            OgogICAgIyBmb3IgMy54CiAgICBmcm9tIGlvIGltcG9ydCBTdHJpbmdJTwoKdHJ5OgogICAgIyBm
            b3IgMi43LngKICAgIGZyb20gdXJscGFyc2UgaW1wb3J0IHVybHBhcnNlCiAgICBmcm9tIHVybGxp
            YiBpbXBvcnQgdXJsZW5jb2RlIGFzIHVybGVuY29kZQpleGNlcHQgTW9kdWxlTm90Rm91bmRFcnJv
            cjoKICAgICMgZm9yIDMueAogICAgZnJvbSB1cmxsaWIucGFyc2UgaW1wb3J0IHVybHBhcnNlCiAg
            ICBmcm9tIHVybGxpYi5wYXJzZSBpbXBvcnQgdXJsZW5jb2RlIGFzIHVybGVuY29kZQoKdHJ5Ogog
            ICAgZnJvbSBiYXNlNjQgaW1wb3J0IGVuY29kZXN0cmluZwpleGNlcHQ6CiAgICBmcm9tIGJhc2U2
            NCBpbXBvcnQgZW5jb2RlYnl0ZXMgYXMgZW5jb2Rlc3RyaW5nCgoKY2xhc3MgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbihFeGNlcHRpb24pOgogICAgcGFzcwoKCiMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMgRHVtbXlSYWRv
            cyAjIyMjIyMjIyMjIyMjIyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjCiMgdGhpcyBpcyBtYWlubHkgZm9yIHRlc3RpbmcgYW5kIGNvdWxkIGJlIHVz
            ZWQgd2hlcmUgJ3JhZG9zJyBpcyBub3QgYXZhaWxhYmxlCgoKY2xhc3MgRHVtbXlSYWRvcyhvYmpl
            Y3QpOgogICAgZGVmIF9faW5pdF9fKHNlbGYpOgogICAgICAgIHNlbGYucmV0dXJuX3ZhbCA9IDAK
            ICAgICAgICBzZWxmLmVycl9tZXNzYWdlID0gIiIKICAgICAgICBzZWxmLnN0YXRlID0gImNvbm5l
            Y3RlZCIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwID0ge30KICAgICAgICBzZWxmLmNtZF9u
            YW1lcyA9IHt9CiAgICAgICAgc2VsZi5faW5pdF9jbWRfb3V0cHV0X21hcCgpCiAgICAgICAgc2Vs
            Zi5kdW1teV9ob3N0X2lwX21hcCA9IHt9CgogICAgZGVmIF9pbml0X2NtZF9vdXRwdXRfbWFwKHNl
            bGYpOgogICAgICAgIGpzb25fZmlsZV9uYW1lID0gInRlc3QtZGF0YS9jZXBoLXN0YXR1cy1vdXQi
            CiAgICAgICAgc2NyaXB0X2RpciA9IHBhdGguYWJzcGF0aChwYXRoLmRpcm5hbWUoX19maWxlX18p
            KQogICAgICAgIGNlcGhfc3RhdHVzX3N0ciA9ICIiCiAgICAgICAgd2l0aCBvcGVuKAogICAgICAg
            ICAgICBwYXRoLmpvaW4oc2NyaXB0X2RpciwganNvbl9maWxlX25hbWUpLCBtb2RlPSJyIiwgZW5j
            b2Rpbmc9IlVURi04IgogICAgICAgICkgYXMganNvbl9maWxlOgogICAgICAgICAgICBjZXBoX3N0
            YXR1c19zdHIgPSBqc29uX2ZpbGUucmVhZCgpCiAgICAgICAgc2VsZi5jbWRfbmFtZXNbImZzIGxz
            Il0gPSAiIiJ7ImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJmcyBscyJ9IiIiCiAgICAgICAg
            c2VsZi5jbWRfbmFtZXNbInF1b3J1bV9zdGF0dXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3Jt
            YXQiOiAianNvbiIsICJwcmVmaXgiOiAicXVvcnVtX3N0YXR1cyJ9IiIiCiAgICAgICAgKQogICAg
            ICAgIHNlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXSA9ICgKICAgICAgICAgICAgIiIieyJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAibWdyIHNlcnZpY2VzIn0iIiIKICAgICAgICApCiAg
            ICAgICAgIyBhbGwgdGhlIGNvbW1hbmRzIGFuZCB0aGVpciBvdXRwdXQKICAgICAgICBzZWxmLmNt
            ZF9vdXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJmcyBscyJdXSA9ICgKICAgICAgICAgICAgIiIi
            W3sibmFtZSI6Im15ZnMiLCJtZXRhZGF0YV9wb29sIjoibXlmcy1tZXRhZGF0YSIsIm1ldGFkYXRh
            X3Bvb2xfaWQiOjIsImRhdGFfcG9vbF9pZHMiOlszXSwiZGF0YV9wb29scyI6WyJteWZzLXJlcGxp
            Y2F0ZWQiXX1dIiIiCiAgICAgICAgKQogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbc2VsZi5j
            bWRfbmFtZXNbInF1b3J1bV9zdGF0dXMiXV0gPSAoCiAgICAgICAgICAgICIiInsiZWxlY3Rpb25f
            ZXBvY2giOjMsInF1b3J1bSI6WzBdLCJxdW9ydW1fbmFtZXMiOlsiYSJdLCJxdW9ydW1fbGVhZGVy
            X25hbWUiOiJhIiwicXVvcnVtX2FnZSI6MTQzODUsImZlYXR1cmVzIjp7InF1b3J1bV9jb24iOiI0
            NTQwMTM4MjkyODM2Njk2MDYzIiwicXVvcnVtX21vbiI6WyJrcmFrZW4iLCJsdW1pbm91cyIsIm1p
            bWljIiwib3NkbWFwLXBydW5lIiwibmF1dGlsdXMiLCJvY3RvcHVzIl19LCJtb25tYXAiOnsiZXBv
            Y2giOjEsImZzaWQiOiJhZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5ZDBmMWMiLCJtb2Rp
            ZmllZCI6IjIwMjAtMDUtMDdUMDM6MzY6MzkuOTE4MDM1WiIsImNyZWF0ZWQiOiIyMDIwLTA1LTA3
            VDAzOjM2OjM5LjkxODAzNVoiLCJtaW5fbW9uX3JlbGVhc2UiOjE1LCJtaW5fbW9uX3JlbGVhc2Vf
            bmFtZSI6Im9jdG9wdXMiLCJmZWF0dXJlcyI6eyJwZXJzaXN0ZW50IjpbImtyYWtlbiIsImx1bWlu
            b3VzIiwibWltaWMiLCJvc2RtYXAtcHJ1bmUiLCJuYXV0aWx1cyIsIm9jdG9wdXMiXSwib3B0aW9u
            YWwiOltdfSwibW9ucyI6W3sicmFuayI6MCwibmFtZSI6ImEiLCJwdWJsaWNfYWRkcnMiOnsiYWRk
            cnZlYyI6W3sidHlwZSI6InYyIiwiYWRkciI6IjEwLjExMC4yMDUuMTc0OjMzMDAiLCJub25jZSI6
            MH0seyJ0eXBlIjoidjEiLCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OSIsIm5vbmNlIjowfV19
            LCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OS8wIiwicHVibGljX2FkZHIiOiIxMC4xMTAuMjA1
            LjE3NDo2Nzg5LzAiLCJwcmlvcml0eSI6MCwid2VpZ2h0IjowfV19fSIiIgogICAgICAgICkKICAg
            ICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXV0g
            PSAoCiAgICAgICAgICAgICIiInsiZGFzaGJvYXJkIjoiaHR0cHM6Ly9jZXBoLWRhc2hib2FyZDo4
            NDQzLyIsInByb21ldGhldXMiOiJodHRwOi8vY2VwaC1kYXNoYm9hcmQtZGI6OTI4My8ifSIiIgog
            ICAgICAgICkKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNh
            cHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMiLCAib3Nk
            IiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0
            YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250
            cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJdLCAiZW50aXR5Ijog
            ImNsaWVudC5oZWFsdGhjaGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRo
            IGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuaGVh
            bHRoY2hlY2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRB
            PT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cyIs
            Im9zZCI6InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5j
            b250cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJ9fV0iIiIKICAg
            ICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIs
            ICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAib3NkIiwgInBy
            b2ZpbGUgcmJkIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtbm9kZSIsICJmb3JtYXQiOiAi
            anNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIi
            W3siZW50aXR5IjoiY2xpZW50LmNzaS1yYmQtbm9kZSIsImtleSI6IkFRQk9nck5lSGJLMUF4QUF1
            YllCZVY4UzFVL0dQenE1U1ZlcTZnPT0iLCJjYXBzIjp7Im1vbiI6InByb2ZpbGUgcmJkLCBhbGxv
            dyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsIm9zZCI6InByb2ZpbGUgcmJkIn19XSIiIgogICAg
            ICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsibW9uIiwg
            InByb2ZpbGUgcmJkLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxs
            b3cgcnciLCAib3NkIiwgInByb2ZpbGUgcmJkIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQt
            cHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNy
            ZWF0ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktcmJkLXByb3Zp
            c2lvbmVyIiwia2V5IjoiQVFCTmdyTmUxZ2V5S3hBQThla1ZpUmRFK2hzczVPd2VZQmt3Tmc9PSIs
            ImNhcHMiOnsibWdyIjoiYWxsb3cgcnciLCJtb24iOiJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFu
            ZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJwcm9maWxlIHJiZCJ9fV0iIiIKICAgICAgICBzZWxm
            LmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyBy
            LCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3Nk
            IiwgImFsbG93IHJ3IHRhZyBjZXBoZnMgKj0qIiwgIm1kcyIsICJhbGxvdyBydyJdLCAiZW50aXR5
            IjogImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4Ijog
            ImF1dGggZ2V0LW9yLWNyZWF0ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVu
            dC5jc2ktY2VwaGZzLW5vZGUiLCJrZXkiOiJBUUJPZ3JOZUVOdW5LeEFBUENtZ0U3UjZHOERjWG5h
            SjFGMzJxZz09IiwiY2FwcyI6eyJtZHMiOiJhbGxvdyBydyIsIm1nciI6ImFsbG93IHJ3IiwibW9u
            IjoiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBy
            dyB0YWcgY2VwaGZzICo9KiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAg
            ICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kICdvc2Qg
            YmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3NkIiwgImFsbG93IHJ3IHRhZyBjZXBo
            ZnMgbWV0YWRhdGE9KiJdLCAiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVy
            IiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgog
            ICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIs
            ImtleSI6IkFRQk9nck5lQUZnY0dCQUF2R3FLT0FEMEQzeHhtVlkwUjkxMmRnPT0iLCJjYXBzIjp7
            Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2Nr
            bGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoifX1dIiIiCiAgICAg
            ICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAi
            YWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdyIiwgImFsbG93IHJ3
            IiwgIm9zZCIsICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiXSwgImVudGl0eSI6ICJj
            bGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVuc2hpZnQtc3RvcmFnZSIsICJmb3JtYXQi
            OiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0g
            IiIiW3siZW50aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0
            b3JhZ2UiLCJrZXkiOiJCUUJPZ3JOZUFGZ2NHQkFBdkdxS09BRDBEM3h4bVZZMFI5MTJkZz09Iiwi
            Y2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29z
            ZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIn19XSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsi
            bW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciIsICJh
            bGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT1teWZzIl0sICJl
            bnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0b3JhZ2Ut
            bXlmcyIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0i
            IiIKICAgICAgICBdID0gIiIiW3siZW50aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9u
            ZXItb3BlbnNoaWZ0LXN0b3JhZ2UtbXlmcyIsImtleSI6IkNRQk9nck5lQUZnY0dCQUF2R3FLT0FE
            MEQzeHhtVlkwUjkxMmRnPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cg
            ciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2Vw
            aGZzIG1ldGFkYXRhPW15ZnMifX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAg
            ICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCBxdW9y
            dW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIiwgImFsbG93IGNvbW1hbmQg
            Y29uZmlnIiwgIm9zZCIsICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFsbG93IHJ3eCBwb29sPWRl
            ZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9ZGVm
            YXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJndy5sb2csIGFsbG93IHgg
            cG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4Il0sICJlbnRpdHkiOiAiY2xpZW50LmhlYWx0
            aGNoZWNrZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0
            ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5oZWFsdGhjaGVja2VyIiwi
            a2V5IjoiQVFERmtiTmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsi
            bW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFu
            ZCB2ZXJzaW9uIiwgIm1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsICJvc2QiOiAicHJvZmls
            ZSByYmQtcmVhZC1vbmx5LCBhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhLCBhbGxvdyBy
            IHBvb2w9LnJndy5yb290LCBhbGxvdyBydyBwb29sPWRlZmF1bHQucmd3LmNvbnRyb2wsIGFsbG93
            IHJ4IHBvb2w9ZGVmYXVsdC5yZ3cubG9nLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0
            cy5pbmRleCJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAi
            IiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMs
            IGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cgY29tbWFuZCBjb25maWciLCAi
            b3NkIiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5j
            b250cm9sLCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1
            bHQucmd3LmJ1Y2tldHMuaW5kZXgiXSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIs
            ICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBjYXBzIn0iIiIKICAgICAgICBdID0g
            IiIiW3siZW50aXR5IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0NWJG
            UkFBVG5kTE5VU1JLcnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIjog
            ImFsbG93IGNvbW1hbmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFs
            bG93IHJ3eCBwb29sPWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFs
            bG93IHJ3IHBvb2w9ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJn
            dy5sb2csIGFsbG93IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4In19XSIiIgogICAg
            ICAgIHNlbGYuY21kX291dHB1dF9tYXBbIiIieyJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            bWdyIHNlcnZpY2VzIn0iIiJdID0gKAogICAgICAgICAgICAiIiJ7ImRhc2hib2FyZCI6ICJodHRw
            Oi8vcm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6NzAwMC8iLCAicHJvbWV0aGV1cyI6
            ICJodHRwOi8vcm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6OTI4My8ifSIiIgogICAg
            ICAgICkKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0
            eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            YXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJ7ImRhc2hib2FyZCI6ICJodHRwOi8vcm9vay1j
            ZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6NzAwMC8iLCAicHJvbWV0aGV1cyI6ICJodHRwOi8v
            cm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6OTI4My8ifSIiIgogICAgICAgIHNlbGYu
            Y21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNsaWVudC5oZWFsdGhj
            aGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldCJ9IiIiCiAgICAg
            ICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5oZWFsdGhjaGVja2VyIiwia2V5IjoiQVFERmti
            TmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjogImFsbG93
            IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFuZCB2ZXJzaW9uIiwg
            Im1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsICJvc2QiOiAicHJvZmlsZSByYmQtcmVhZC1v
            bmx5LCBhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhLCBhbGxvdyByIHBvb2w9LnJndy5y
            b290LCBhbGxvdyBydyBwb29sPWRlZmF1bHQucmd3LmNvbnRyb2wsIGFsbG93IHJ4IHBvb2w9ZGVm
            YXVsdC5yZ3cubG9nLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJ9fV0i
            IiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0eSI6
            ICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJh
            dXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0
            X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtbm9kZSIsICJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJb
            XSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5
            IjogImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRf
            b3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMt
            cHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIK
            ICAgICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAg
            ICAgICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVuc2hp
            ZnQtc3RvcmFnZSIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgog
            ICAgICAgIF0gPSAiIiJbXSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAg
            ICAgICIiInsiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlm
            dC1zdG9yYWdlLW15ZnMiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0i
            IiIKICAgICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAg
            ICAgICAgICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJb
            eyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsImtleSI6IkFRREZrYk5l
            ZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRBPT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIi
            LCAibWdyIjoiYWxsb3cgcnciLCAib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0q
            In19XSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2Fw
            cyI6IFsibW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1n
            ciIsICJhbGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIl0s
            ICJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpz
            b24iLCAicHJlZml4IjogImF1dGggY2FwcyJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmtiTmVmdDViRlJBQVRu
            ZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cgciwgIGFsbG93IGNv
            bW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciI6ImFsbG93IHJ3IiwgIm9zZCI6ImFsbG93IHJ3
            IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFw
            Wyd7ImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJzdGF0dXMifSddID0gY2VwaF9zdGF0dXNf
            c3RyCgogICAgZGVmIHNodXRkb3duKHNlbGYpOgogICAgICAgIHBhc3MKCiAgICBkZWYgZ2V0X2Zz
            aWQoc2VsZik6CiAgICAgICAgcmV0dXJuICJhZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5
            ZDBmMWMiCgogICAgZGVmIGNvbmZfcmVhZF9maWxlKHNlbGYpOgogICAgICAgIHBhc3MKCiAgICBk
            ZWYgY29ubmVjdChzZWxmKToKICAgICAgICBwYXNzCgogICAgZGVmIHBvb2xfZXhpc3RzKHNlbGYs
            IHBvb2xfbmFtZSk6CiAgICAgICAgcmV0dXJuIFRydWUKCiAgICBkZWYgbW9uX2NvbW1hbmQoc2Vs
            ZiwgY21kLCBvdXQpOgogICAgICAgIGpzb25fY21kID0ganNvbi5sb2FkcyhjbWQpCiAgICAgICAg
            anNvbl9jbWRfc3RyID0ganNvbi5kdW1wcyhqc29uX2NtZCwgc29ydF9rZXlzPVRydWUpCiAgICAg
            ICAgY21kX291dHB1dCA9IHNlbGYuY21kX291dHB1dF9tYXBbanNvbl9jbWRfc3RyXQogICAgICAg
            IHJldHVybiBzZWxmLnJldHVybl92YWwsIGNtZF9vdXRwdXQsIHN0cihzZWxmLmVycl9tZXNzYWdl
            LmVuY29kZSgidXRmLTgiKSkKCiAgICBkZWYgX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoc2VsZiwg
            aG9zdF9uYW1lKToKICAgICAgICBpcF9yZWdfeCA9IHJlLmNvbXBpbGUociJcZHsxLDN9LlxkezEs
            M30uXGR7MSwzfS5cZHsxLDN9IikKICAgICAgICAjIGlmIHByb3ZpZGVkIGhvc3QgaXMgZGlyZWN0
            bHkgYW4gSVAgYWRkcmVzcywgcmV0dXJuIHRoZSBzYW1lCiAgICAgICAgaWYgaXBfcmVnX3gubWF0
            Y2goaG9zdF9uYW1lKToKICAgICAgICAgICAgcmV0dXJuIGhvc3RfbmFtZQogICAgICAgIGltcG9y
            dCByYW5kb20KCiAgICAgICAgaG9zdF9pcCA9IHNlbGYuZHVtbXlfaG9zdF9pcF9tYXAuZ2V0KGhv
            c3RfbmFtZSwgIiIpCiAgICAgICAgaWYgbm90IGhvc3RfaXA6CiAgICAgICAgICAgIGhvc3RfaXAg
            PSBmIjE3Mi45LntyYW5kb20ucmFuZGludCgwLCAyNTQpfS57cmFuZG9tLnJhbmRpbnQoMCwgMjU0
            KX0iCiAgICAgICAgICAgIHNlbGYuZHVtbXlfaG9zdF9pcF9tYXBbaG9zdF9uYW1lXSA9IGhvc3Rf
            aXAKICAgICAgICBkZWwgcmFuZG9tCiAgICAgICAgcmV0dXJuIGhvc3RfaXAKCiAgICBAY2xhc3Nt
            ZXRob2QKICAgIGRlZiBSYWRvcyhjb25mZmlsZT1Ob25lKToKICAgICAgICByZXR1cm4gRHVtbXlS
            YWRvcygpCgoKY2xhc3MgUzNBdXRoKEF1dGhCYXNlKToKICAgICIiIkF0dGFjaGVzIEFXUyBBdXRo
            ZW50aWNhdGlvbiB0byB0aGUgZ2l2ZW4gUmVxdWVzdCBvYmplY3QuIiIiCgogICAgc2VydmljZV9i
            YXNlX3VybCA9ICJzMy5hbWF6b25hd3MuY29tIgoKICAgIGRlZiBfX2luaXRfXyhzZWxmLCBhY2Nl
            c3Nfa2V5LCBzZWNyZXRfa2V5LCBzZXJ2aWNlX3VybD1Ob25lKToKICAgICAgICBpZiBzZXJ2aWNl
            X3VybDoKICAgICAgICAgICAgc2VsZi5zZXJ2aWNlX2Jhc2VfdXJsID0gc2VydmljZV91cmwKICAg
            ICAgICBzZWxmLmFjY2Vzc19rZXkgPSBzdHIoYWNjZXNzX2tleSkKICAgICAgICBzZWxmLnNlY3Jl
            dF9rZXkgPSBzdHIoc2VjcmV0X2tleSkKCiAgICBkZWYgX19jYWxsX18oc2VsZiwgcik6CiAgICAg
            ICAgIyBDcmVhdGUgZGF0ZSBoZWFkZXIgaWYgaXQgaXMgbm90IGNyZWF0ZWQgeWV0LgogICAgICAg
            IGlmICJkYXRlIiBub3QgaW4gci5oZWFkZXJzIGFuZCAieC1hbXotZGF0ZSIgbm90IGluIHIuaGVh
            ZGVyczoKICAgICAgICAgICAgci5oZWFkZXJzWyJkYXRlIl0gPSBmb3JtYXRkYXRlKHRpbWV2YWw9
            Tm9uZSwgbG9jYWx0aW1lPUZhbHNlLCB1c2VnbXQ9VHJ1ZSkKICAgICAgICBzaWduYXR1cmUgPSBz
            ZWxmLmdldF9zaWduYXR1cmUocikKICAgICAgICBpZiBweTNrOgogICAgICAgICAgICBzaWduYXR1
            cmUgPSBzaWduYXR1cmUuZGVjb2RlKCJ1dGYtOCIpCiAgICAgICAgci5oZWFkZXJzWyJBdXRob3Jp
            emF0aW9uIl0gPSBmIkFXUyB7c2VsZi5hY2Nlc3Nfa2V5fTp7c2lnbmF0dXJlfSIKICAgICAgICBy
            ZXR1cm4gcgoKICAgIGRlZiBnZXRfc2lnbmF0dXJlKHNlbGYsIHIpOgogICAgICAgIGNhbm9uaWNh
            bF9zdHJpbmcgPSBzZWxmLmdldF9jYW5vbmljYWxfc3RyaW5nKHIudXJsLCByLmhlYWRlcnMsIHIu
            bWV0aG9kKQogICAgICAgIGlmIHB5M2s6CiAgICAgICAgICAgIGtleSA9IHNlbGYuc2VjcmV0X2tl
            eS5lbmNvZGUoInV0Zi04IikKICAgICAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZy5lbmNv
            ZGUoInV0Zi04IikKICAgICAgICBlbHNlOgogICAgICAgICAgICBrZXkgPSBzZWxmLnNlY3JldF9r
            ZXkKICAgICAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZwogICAgICAgIGggPSBobWFjLm5l
            dyhrZXksIG1zZywgZGlnZXN0bW9kPXNoYSkKICAgICAgICByZXR1cm4gZW5jb2Rlc3RyaW5nKGgu
            ZGlnZXN0KCkpLnN0cmlwKCkKCiAgICBkZWYgZ2V0X2Nhbm9uaWNhbF9zdHJpbmcoc2VsZiwgdXJs
            LCBoZWFkZXJzLCBtZXRob2QpOgogICAgICAgIHBhcnNlZHVybCA9IHVybHBhcnNlKHVybCkKICAg
            ICAgICBvYmplY3RrZXkgPSBwYXJzZWR1cmwucGF0aFsxOl0KCiAgICAgICAgYnVja2V0ID0gcGFy
            c2VkdXJsLm5ldGxvY1s6IC1sZW4oc2VsZi5zZXJ2aWNlX2Jhc2VfdXJsKV0KICAgICAgICBpZiBs
            ZW4oYnVja2V0KSA+IDE6CiAgICAgICAgICAgICMgcmVtb3ZlIGxhc3QgZG90CiAgICAgICAgICAg
            IGJ1Y2tldCA9IGJ1Y2tldFs6LTFdCgogICAgICAgIGludGVyZXN0aW5nX2hlYWRlcnMgPSB7ImNv
            bnRlbnQtbWQ1IjogIiIsICJjb250ZW50LXR5cGUiOiAiIiwgImRhdGUiOiAiIn0KICAgICAgICBm
            b3Iga2V5IGluIGhlYWRlcnM6CiAgICAgICAgICAgIGxrID0ga2V5Lmxvd2VyKCkKICAgICAgICAg
            ICAgdHJ5OgogICAgICAgICAgICAgICAgbGsgPSBsay5kZWNvZGUoInV0Zi04IikKICAgICAgICAg
            ICAgZXhjZXB0OgogICAgICAgICAgICAgICAgcGFzcwogICAgICAgICAgICBpZiBoZWFkZXJzW2tl
            eV0gYW5kICgKICAgICAgICAgICAgICAgIGxrIGluIGludGVyZXN0aW5nX2hlYWRlcnMua2V5cygp
            IG9yIGxrLnN0YXJ0c3dpdGgoIngtYW16LSIpCiAgICAgICAgICAgICk6CiAgICAgICAgICAgICAg
            ICBpbnRlcmVzdGluZ19oZWFkZXJzW2xrXSA9IGhlYWRlcnNba2V5XS5zdHJpcCgpCgogICAgICAg
            ICMgSWYgeC1hbXotZGF0ZSBpcyB1c2VkIGl0IHN1cGVyc2VkZXMgdGhlIGRhdGUgaGVhZGVyLgog
            ICAgICAgIGlmIG5vdCBweTNrOgogICAgICAgICAgICBpZiAieC1hbXotZGF0ZSIgaW4gaW50ZXJl
            c3RpbmdfaGVhZGVyczoKICAgICAgICAgICAgICAgIGludGVyZXN0aW5nX2hlYWRlcnNbImRhdGUi
            XSA9ICIiCiAgICAgICAgZWxzZToKICAgICAgICAgICAgaWYgIngtYW16LWRhdGUiIGluIGludGVy
            ZXN0aW5nX2hlYWRlcnM6CiAgICAgICAgICAgICAgICBpbnRlcmVzdGluZ19oZWFkZXJzWyJkYXRl
            Il0gPSAiIgoKICAgICAgICBidWYgPSBmInttZXRob2R9XG4iCiAgICAgICAgZm9yIGtleSBpbiBz
            b3J0ZWQoaW50ZXJlc3RpbmdfaGVhZGVycy5rZXlzKCkpOgogICAgICAgICAgICB2YWwgPSBpbnRl
            cmVzdGluZ19oZWFkZXJzW2tleV0KICAgICAgICAgICAgaWYga2V5LnN0YXJ0c3dpdGgoIngtYW16
            LSIpOgogICAgICAgICAgICAgICAgYnVmICs9IGYie2tleX06e3ZhbH1cbiIKICAgICAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgICAgIGJ1ZiArPSBmInt2YWx9XG4iCgogICAgICAgICMgYXBwZW5k
            IHRoZSBidWNrZXQgaWYgaXQgZXhpc3RzCiAgICAgICAgaWYgYnVja2V0ICE9ICIiOgogICAgICAg
            ICAgICBidWYgKz0gZiIve2J1Y2tldH0iCgogICAgICAgICMgYWRkIHRoZSBvYmplY3RrZXkuIGV2
            ZW4gaWYgaXQgZG9lc24ndCBleGlzdCwgYWRkIHRoZSBzbGFzaAogICAgICAgIGJ1ZiArPSBmIi97
            b2JqZWN0a2V5fSIKCiAgICAgICAgcmV0dXJuIGJ1ZgoKCmNsYXNzIFJhZG9zSlNPTjoKICAgIEVY
            VEVSTkFMX1VTRVJfTkFNRSA9ICJjbGllbnQuaGVhbHRoY2hlY2tlciIKICAgIEVYVEVSTkFMX1JH
            V19BRE1JTl9PUFNfVVNFUl9OQU1FID0gInJndy1hZG1pbi1vcHMtdXNlciIKICAgIEVNUFRZX09V
            VFBVVF9MSVNUID0gIkVtcHR5IG91dHB1dCBsaXN0IgogICAgREVGQVVMVF9SR1dfUE9PTF9QUkVG
            SVggPSAiZGVmYXVsdCIKICAgIERFRkFVTFRfTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUID0gIjky
            ODMiCgogICAgQGNsYXNzbWV0aG9kCiAgICBkZWYgZ2VuX2FyZ19wYXJzZXIoY2xzLCBhcmdzX3Rv
            X3BhcnNlPU5vbmUpOgogICAgICAgIGFyZ1AgPSBhcmdwYXJzZS5Bcmd1bWVudFBhcnNlcigpCgog
            ICAgICAgIGNvbW1vbl9ncm91cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJjb21tb24iKQog
            ICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoIi0tdmVyYm9zZSIsICItdiIsIGFjdGlv
            bj0ic3RvcmVfdHJ1ZSIsIGRlZmF1bHQ9RmFsc2UpCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9h
            cmd1bWVudCgKICAgICAgICAgICAgIi0tY2VwaC1jb25mIiwgIi1jIiwgaGVscD0iUHJvdmlkZSBh
            IGNlcGggY29uZiBmaWxlLiIsIHR5cGU9c3RyCiAgICAgICAgKQogICAgICAgIGNvbW1vbl9ncm91
            cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLWtleXJpbmciLCAiLWsiLCBoZWxwPSJQYXRo
            IHRvIGNlcGgga2V5cmluZyBmaWxlLiIsIHR5cGU9c3RyCiAgICAgICAgKQogICAgICAgIGNvbW1v
            bl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJ1bi1hcy11c2VyIiwKICAgICAg
            ICAgICAgIi11IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgdHlwZT1zdHIs
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIGEgdXNlciBuYW1lIHRvIGNoZWNrIHRoZSBjbHVz
            dGVyJ3MgaGVhbHRoIHN0YXR1cywgbXVzdCBiZSBwcmVmaXhlZCBieSAnY2xpZW50LiciLAogICAg
            ICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1j
            bHVzdGVyLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJL
            dWJlcm5ldGVzIGNsdXN0ZXIgbmFtZShsZWdhY3kgZmxhZyksIE5vdGU6IEVpdGhlciB1c2UgdGhp
            cyBvciAtLWs4cy1jbHVzdGVyLW5hbWUiLAogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAu
            YWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1rOHMtY2x1c3Rlci1uYW1lIiwgZGVmYXVsdD0i
            IiwgaGVscD0iS3ViZXJuZXRlcyBjbHVzdGVyIG5hbWUiCiAgICAgICAgKQogICAgICAgIGNvbW1v
            bl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLW5hbWVzcGFjZSIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIGhlbHA9Ik5hbWVzcGFjZSB3aGVyZSBDZXBoQ2x1
            c3RlciBpcyBydW5uaW5nIiwKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tcmd3LXBvb2wtcHJlZml4IiwgZGVmYXVsdD0iIiwgaGVscD0i
            UkdXIFBvb2wgcHJlZml4IgogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3Vt
            ZW50KAogICAgICAgICAgICAiLS1yZXN0cmljdGVkLWF1dGgtcGVybWlzc2lvbiIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlJlc3RyaWN0IGNlcGhDU0lLZXly
            aW5ncyBhdXRoIHBlcm1pc3Npb25zIHRvIHNwZWNpZmljIHBvb2xzLCBjbHVzdGVyLiIKICAgICAg
            ICAgICAgKyAiTWFuZGF0b3J5IGZsYWdzIHRoYXQgbmVlZCB0byBiZSBzZXQgYXJlIC0tcmJkLWRh
            dGEtcG9vbC1uYW1lLCBhbmQgLS1rOHMtY2x1c3Rlci1uYW1lLiIKICAgICAgICAgICAgKyAiLS1j
            ZXBoZnMtZmlsZXN5c3RlbS1uYW1lIGZsYWcgY2FuIGFsc28gYmUgcGFzc2VkIGluIGNhc2Ugb2Yg
            Y2VwaGZzIHVzZXIgcmVzdHJpY3Rpb24sIHNvIGl0IGNhbiByZXN0cmljdCB1c2VyIHRvIHBhcnRp
            Y3VsYXIgY2VwaGZzIGZpbGVzeXN0ZW0iCiAgICAgICAgICAgICsgInNhbXBsZSBydW46IGBweXRo
            b24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwtY2x1c3Rlci1yZXNvdXJjZXMucHkgLS1jZXBo
            ZnMtZmlsZXN5c3RlbS1uYW1lIG15ZnMgLS1yYmQtZGF0YS1wb29sLW5hbWUgcmVwbGljYXBvb2wg
            LS1rOHMtY2x1c3Rlci1uYW1lIHJvb2tzdG9yYWdlIC0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Np
            b24gdHJ1ZWAiCiAgICAgICAgICAgICsgIk5vdGU6IFJlc3RyaWN0aW5nIHRoZSBjc2ktdXNlcnMg
            cGVyIHBvb2wsIGFuZCBwZXIgY2x1c3RlciB3aWxsIHJlcXVpcmUgY3JlYXRpbmcgbmV3IGNzaS11
            c2VycyBhbmQgbmV3IHNlY3JldHMgZm9yIHRoYXQgY3NpLXVzZXJzLiIKICAgICAgICAgICAgKyAi
            U28gYXBwbHkgdGhlc2Ugc2VjcmV0cyBvbmx5IHRvIG5ldyBgQ29uc3VtZXIgY2x1c3RlcmAgZGVw
            bG95bWVudCB3aGlsZSB1c2luZyB0aGUgc2FtZSBgU291cmNlIGNsdXN0ZXJgLiIsCiAgICAgICAg
            KQogICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXYyLXBv
            cnQtZW5hYmxlIiwKICAgICAgICAgICAgYWN0aW9uPSJzdG9yZV90cnVlIiwKICAgICAgICAgICAg
            ZGVmYXVsdD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iRW5hYmxlIHYyIG1vbiBwb3J0KDMzMDAp
            IGZvciBtb25zIiwKICAgICAgICApCgogICAgICAgIG91dHB1dF9ncm91cCA9IGFyZ1AuYWRkX2Fy
            Z3VtZW50X2dyb3VwKCJvdXRwdXQiKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQo
            CiAgICAgICAgICAgICItLWZvcm1hdCIsCiAgICAgICAgICAgICItdCIsCiAgICAgICAgICAgIGNo
            b2ljZXM9WyJqc29uIiwgImJhc2giXSwKICAgICAgICAgICAgZGVmYXVsdD0ianNvbiIsCiAgICAg
            ICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBvdXRwdXQgZm9ybWF0IChqc29uIHwgYmFzaCkiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1vdXRwdXQiLAogICAgICAgICAgICAiLW8iLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAg
            ICAgICAgICBoZWxwPSJPdXRwdXQgd2lsbCBiZSBzdG9yZWQgaW50byB0aGUgcHJvdmlkZWQgZmls
            ZSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAog
            ICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgQ2VwaCBmaWxlc3lzdGVt
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tY2VwaGZzLW1ldGFkYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIs
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBjZXBoZnMgbWV0YWRh
            dGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAg
            ICAgICAgICAgICItLWNlcGhmcy1kYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9
            IiIsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBjZXBoZnMgZGF0
            YSBwb29sIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAg
            ICAgICAgICAgIi0tcmJkLWRhdGEtcG9vbC1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwK
            ICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRo
            ZSBuYW1lIG9mIHRoZSBSQkQgZGF0YXBvb2wiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3Jv
            dXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1hbGlhcy1yYmQtZGF0YS1wb29sLW5hbWUi
            LAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAg
            ICAgICAgICAgaGVscD0iUHJvdmlkZXMgYW4gYWxpYXMgZm9yIHRoZSAgUkJEIGRhdGEgcG9vbCBu
            YW1lLCBuZWNlc3NhcnkgaWYgYSBzcGVjaWFsIGNoYXJhY3RlciBpcyBwcmVzZW50IGluIHRoZSBw
            b29sIG5hbWUgc3VjaCBhcyBhIHBlcmlvZCBvciB1bmRlcnNjb3JlIiwKICAgICAgICApCiAgICAg
            ICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmd3LWVuZHBvaW50
            IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAg
            ICAgICAgICAgIGhlbHA9IlJBRE9TIEdhdGV3YXkgZW5kcG9pbnQgKGluIGA8SVB2ND46PFBPUlQ+
            YCBvciBgPFtJUHY2XT46PFBPUlQ+YCBvciBgPEZRRE4+OjxQT1JUPmAgZm9ybWF0KSIsCiAgICAg
            ICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJn
            dy10bHMtY2VydC1wYXRoIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVx
            dWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlJBRE9TIEdhdGV3YXkgZW5kcG9pbnQgVExT
            IGNlcnRpZmljYXRlIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVu
            dCgKICAgICAgICAgICAgIi0tcmd3LXNraXAtdGxzIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFs
            c2UsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9Iklnbm9yZSBU
            TFMgY2VydGlmaWNhdGlvbiB2YWxpZGF0aW9uIHdoZW4gYSBzZWxmLXNpZ25lZCBjZXJ0aWZpY2F0
            ZSBpcyBwcm92aWRlZCAoTk9UIFJFQ09NTUVOREVEIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tbW9uaXRvcmluZy1lbmRwb2ludCIs
            CiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAg
            ICAgICAgICBoZWxwPSJDZXBoIE1hbmFnZXIgcHJvbWV0aGV1cyBleHBvcnRlciBlbmRwb2ludHMg
            KGNvbW1hIHNlcGFyYXRlZCBsaXN0IG9mIChmb3JtYXQgYDxJUHY0PmAgb3IgYDxbSVB2Nl0+YCBv
            ciBgPEZRRE4+YCkgZW50cmllcyBvZiBhY3RpdmUgYW5kIHN0YW5kYnkgbWdycykiLAogICAgICAg
            ICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1tb25p
            dG9yaW5nLWVuZHBvaW50LXBvcnQiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAg
            ICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iQ2VwaCBNYW5hZ2VyIHByb21ldGhl
            dXMgZXhwb3J0ZXIgcG9ydCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJn
            dW1lbnQoCiAgICAgICAgICAgICItLXNraXAtbW9uaXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAg
            ICAgICAgIGhlbHA9IkRvIG5vdCBjaGVjayBmb3IgYSBtb25pdG9yaW5nIGVuZHBvaW50IGZvciB0
            aGUgQ2VwaCBjbHVzdGVyIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tcmJkLW1ldGFkYXRhLWVjLXBvb2wtbmFtZSIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBo
            ZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiBlcmFzdXJlIGNvZGVkIFJCRCBtZXRhZGF0YSBwb29s
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tZHJ5LXJ1biIsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFj
            dGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGhlbHA9IkRyeSBydW4gcHJpbnRzIHRoZSBl
            eGVjdXRlZCBjb21tYW5kcyB3aXRob3V0IHJ1bm5pbmcgdGhlbSIsCiAgICAgICAgKQogICAgICAg
            IG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJhZG9zLW5hbWVzcGFj
            ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAog
            ICAgICAgICAgICBoZWxwPSJEaXZpZGVzIGEgcG9vbCBpbnRvIHNlcGFyYXRlIGxvZ2ljYWwgbmFt
            ZXNwYWNlcywgdXNlZCBmb3IgY3JlYXRpbmcgUkJEIFBWQyBpbiBhIENlcGhCbG9ja1Bvb2xSYWRv
            c05hbWVzcGFjZSAoc2hvdWxkIGJlIGxvd2VyIGNhc2UpIiwKICAgICAgICApCiAgICAgICAgb3V0
            cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tc3Vidm9sdW1lLWdyb3VwIiwK
            ICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAg
            ICAgICAgIGhlbHA9InByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBzdWJ2b2x1bWUgZ3JvdXAiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1yZ3ctcmVhbG0tbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJl
            cXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUg
            cmd3LXJlYWxtIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgK
            ICAgICAgICAgICAgIi0tcmd3LXpvbmUtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAg
            ICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUg
            bmFtZSBvZiB0aGUgcmd3LXpvbmUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS1yZ3ctem9uZWdyb3VwLW5hbWUiLAogICAgICAgICAg
            ICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVs
            cD0icHJvdmlkZXMgdGhlIG5hbWUgb2YgdGhlIHJndy16b25lZ3JvdXAiLAogICAgICAgICkKICAg
            ICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS10b3BvbG9neS1w
            b29scyIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNl
            LAogICAgICAgICAgICBoZWxwPSJjb21tYS1zZXBhcmF0ZWQgbGlzdCBvZiB0b3BvbG9neS1jb25z
            dHJhaW5lZCByYmQgcG9vbHMiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2Fy
            Z3VtZW50KAogICAgICAgICAgICAiLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi1sYWJlbCIsCiAg
            ICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAg
            ICAgICBoZWxwPSJrOHMgY2x1c3RlciBmYWlsdXJlIGRvbWFpbiBsYWJlbCAoZXhhbXBsZTogem9u
            ZSwgcmFjaywgb3IgaG9zdCkgZm9yIHRoZSB0b3BvbG9neS1wb29scyB0aGF0IG1hdGNoIHRoZSBj
            ZXBoIGRvbWFpbiIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQo
            CiAgICAgICAgICAgICItLXRvcG9sb2d5LWZhaWx1cmUtZG9tYWluLXZhbHVlcyIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBo
            ZWxwPSJjb21tYS1zZXBhcmF0ZWQgbGlzdCBvZiB0aGUgazhzIGNsdXN0ZXIgZmFpbHVyZSBkb21h
            aW4gdmFsdWVzIGNvcnJlc3BvbmRpbmcgdG8gZWFjaCBvZiB0aGUgcG9vbHMgaW4gdGhlIGB0b3Bv
            bG9neS1wb29sc2AgbGlzdCIsCiAgICAgICAgKQoKICAgICAgICB1cGdyYWRlX2dyb3VwID0gYXJn
            UC5hZGRfYXJndW1lbnRfZ3JvdXAoInVwZ3JhZGUiKQogICAgICAgIHVwZ3JhZGVfZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS11cGdyYWRlIiwKICAgICAgICAgICAgYWN0aW9uPSJz
            dG9yZV90cnVlIiwKICAgICAgICAgICAgZGVmYXVsdD1GYWxzZSwKICAgICAgICAgICAgaGVscD0i
            VXBncmFkZXMgdGhlIGNlcGhDU0lLZXlyaW5ncyhGb3IgZXhhbXBsZTogY2xpZW50LmNzaS1jZXBo
            ZnMtcHJvdmlzaW9uZXIpIGFuZCBjbGllbnQuaGVhbHRoY2hlY2tlciBjZXBoIHVzZXJzIHdpdGgg
            bmV3IHBlcm1pc3Npb25zIG5lZWRlZCBmb3IgdGhlIG5ldyBjbHVzdGVyIHZlcnNpb24gYW5kIG9s
            ZGVyIHBlcm1pc3Npb24gd2lsbCBzdGlsbCBiZSBhcHBsaWVkLiIKICAgICAgICAgICAgKyAiU2Ft
            cGxlIHJ1bjogYHB5dGhvbjMgL2V0Yy9jZXBoL2NyZWF0ZS1leHRlcm5hbC1jbHVzdGVyLXJlc291
            cmNlcy5weSAtLXVwZ3JhZGVgLCB0aGlzIHdpbGwgdXBncmFkZSBhbGwgdGhlIGRlZmF1bHQgY3Np
            IHVzZXJzKG5vbi1yZXN0cmljdGVkKSIKICAgICAgICAgICAgKyAiRm9yIHJlc3RyaWN0ZWQgdXNl
            cnMoRm9yIGV4YW1wbGU6IGNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1z
            dG9yYWdlLW15ZnMpLCB1c2VycyBjcmVhdGVkIHVzaW5nIC0tcmVzdHJpY3RlZC1hdXRoLXBlcm1p
            c3Npb24gZmxhZyBuZWVkIHRvIHBhc3MgbWFuZGF0b3J5IGZsYWdzIgogICAgICAgICAgICArICJt
            YW5kYXRvcnkgZmxhZ3M6ICctLXJiZC1kYXRhLXBvb2wtbmFtZSwgLS1rOHMtY2x1c3Rlci1uYW1l
            IGFuZCAtLXJ1bi1hcy11c2VyJyBmbGFncyB3aGlsZSB1cGdyYWRpbmciCiAgICAgICAgICAgICsg
            ImluIGNhc2Ugb2YgY2VwaGZzIHVzZXJzIGlmIHlvdSBoYXZlIHBhc3NlZCAtLWNlcGhmcy1maWxl
            c3lzdGVtLW5hbWUgZmxhZyB3aGlsZSBjcmVhdGluZyB1c2VyIHRoZW4gd2hpbGUgdXBncmFkaW5n
            IGl0IHdpbGwgYmUgbWFuZGF0b3J5IHRvbyIKICAgICAgICAgICAgKyAiU2FtcGxlIHJ1bjogYHB5
            dGhvbjMgL2V0Yy9jZXBoL2NyZWF0ZS1leHRlcm5hbC1jbHVzdGVyLXJlc291cmNlcy5weSAtLXVw
            Z3JhZGUgLS1yYmQtZGF0YS1wb29sLW5hbWUgcmVwbGljYXBvb2wgLS1rOHMtY2x1c3Rlci1uYW1l
            IHJvb2tzdG9yYWdlICAtLXJ1bi1hcy11c2VyIGNsaWVudC5jc2ktcmJkLW5vZGUtcm9va3N0b3Jh
            Z2UtcmVwbGljYXBvb2xgIgogICAgICAgICAgICArICJQUzogQW4gZXhpc3Rpbmcgbm9uLXJlc3Ry
            aWN0ZWQgdXNlciBjYW5ub3QgYmUgY29udmVydGVkIHRvIGEgcmVzdHJpY3RlZCB1c2VyIGJ5IHVw
            Z3JhZGluZy4iCiAgICAgICAgICAgICsgIlVwZ3JhZGUgZmxhZyBzaG91bGQgb25seSBiZSB1c2Vk
            IHRvIGFwcGVuZCBuZXcgcGVybWlzc2lvbnMgdG8gdXNlcnMsIGl0IHNob3VsZG4ndCBiZSB1c2Vk
            IGZvciBjaGFuZ2luZyB1c2VyIGFscmVhZHkgYXBwbGllZCBwZXJtaXNzaW9uLCBmb3IgZXhhbXBs
            ZSB5b3Ugc2hvdWxkbid0IGNoYW5nZSBpbiB3aGljaCBwb29sIHVzZXIgaGFzIGFjY2VzcyIsCiAg
            ICAgICAgKQoKICAgICAgICBpZiBhcmdzX3RvX3BhcnNlOgogICAgICAgICAgICBhc3NlcnQgKAog
            ICAgICAgICAgICAgICAgdHlwZShhcmdzX3RvX3BhcnNlKSA9PSBsaXN0CiAgICAgICAgICAgICks
            ICJBcmd1bWVudCB0byAnZ2VuX2FyZ19wYXJzZXInIHNob3VsZCBiZSBhIGxpc3QiCiAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgYXJnc190b19wYXJzZSA9IHN5cy5hcmd2WzE6XQogICAgICAgIHJl
            dHVybiBhcmdQLnBhcnNlX2FyZ3MoYXJnc190b19wYXJzZSkKCiAgICBkZWYgdmFsaWRhdGVfcmJk
            X21ldGFkYXRhX2VjX3Bvb2xfbmFtZShzZWxmKToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2Vy
            LnJiZF9tZXRhZGF0YV9lY19wb29sX25hbWU6CiAgICAgICAgICAgIHJiZF9tZXRhZGF0YV9lY19w
            b29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUKICAg
            ICAgICAgICAgcmJkX3Bvb2xfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9u
            YW1lCgogICAgICAgICAgICBpZiByYmRfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiRmxh
            ZyAnLS1yYmQtZGF0YS1wb29sLW5hbWUnIHNob3VsZCBub3QgYmUgZW1wdHkiCiAgICAgICAgICAg
            ICAgICApCgogICAgICAgICAgICBpZiByYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lID09ICIiOgog
            ICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAg
            ICAgICAgICAgICAiRmxhZyAnLS1yYmQtbWV0YWRhdGEtZWMtcG9vbC1uYW1lJyBzaG91bGQgbm90
            IGJlIGVtcHR5IgogICAgICAgICAgICAgICAgKQoKICAgICAgICAgICAgY21kX2pzb24gPSB7InBy
            ZWZpeCI6ICJvc2QgZHVtcCIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgICAgIHJldF92YWws
            IGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikK
            ICAgICAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAg
            ICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAg
            ICAgICAgZiJ7Y21kX2pzb25bJ3ByZWZpeCddfSBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAg
            ICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVN
            UFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgbWV0YWRhdGFf
            cG9vbF9leGlzdCwgcG9vbF9leGlzdCA9IEZhbHNlLCBGYWxzZQoKICAgICAgICAgICAgZm9yIGtl
            eSBpbiBqc29uX291dFsicG9vbHMiXToKICAgICAgICAgICAgICAgICMgaWYgZXJhc3VyZV9jb2Rl
            X3Byb2ZpbGUgaXMgZW1wdHkgYW5kIHBvb2wgbmFtZSBleGlzdHMgdGhlbiBpdCByZXBsaWNhIHBv
            b2wKICAgICAgICAgICAgICAgIGlmICgKICAgICAgICAgICAgICAgICAgICBrZXlbImVyYXN1cmVf
            Y29kZV9wcm9maWxlIl0gPT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQga2V5WyJwb29sX25h
            bWUiXSA9PSByYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lCiAgICAgICAgICAgICAgICApOgogICAg
            ICAgICAgICAgICAgICAgIG1ldGFkYXRhX3Bvb2xfZXhpc3QgPSBUcnVlCiAgICAgICAgICAgICAg
            ICAjIGlmIGVyYXN1cmVfY29kZV9wcm9maWxlIGlzIG5vdCBlbXB0eSBhbmQgcG9vbCBuYW1lIGV4
            aXN0cyB0aGVuIGl0IGlzIGVjIHBvb2wKICAgICAgICAgICAgICAgIGlmIGtleVsiZXJhc3VyZV9j
            b2RlX3Byb2ZpbGUiXSBhbmQga2V5WyJwb29sX25hbWUiXSA9PSByYmRfcG9vbF9uYW1lOgogICAg
            ICAgICAgICAgICAgICAgIHBvb2xfZXhpc3QgPSBUcnVlCgogICAgICAgICAgICBpZiBub3QgbWV0
            YWRhdGFfcG9vbF9leGlzdDoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVF
            eGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIlByb3ZpZGVkIHJiZF9lY19tZXRhZGF0YV9w
            b29sIG5hbWUsIgogICAgICAgICAgICAgICAgICAgIGYiIHtyYmRfbWV0YWRhdGFfZWNfcG9vbF9u
            YW1lfSwgZG9lcyBub3QgZXhpc3QiCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmIG5v
            dCBwb29sX2V4aXN0OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICBmIlByb3ZpZGVkIHJiZF9kYXRhX3Bvb2wgbmFtZSwg
            e3JiZF9wb29sX25hbWV9LCBkb2VzIG5vdCBleGlzdCIKICAgICAgICAgICAgICAgICkKICAgICAg
            ICAgICAgcmV0dXJuIHJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUKCiAgICBkZWYgZHJ5X3J1bihz
            ZWxmLCBtc2cpOgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAg
            ICAgcHJpbnQoIkV4ZWN1dGU6ICIgKyAiJyIgKyBtc2cgKyAiJyIpCgogICAgZGVmIHZhbGlkYXRl
            X3Jnd19lbmRwb2ludF90bHNfY2VydChzZWxmKToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd190bHNfY2VydF9wYXRoOgogICAgICAgICAgICB3aXRoIG9wZW4oc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfdGxzX2NlcnRfcGF0aCwgZW5jb2Rpbmc9InV0ZjgiKSBhcyBmOgogICAgICAgICAgICAg
            ICAgY29udGVudHMgPSBmLnJlYWQoKQogICAgICAgICAgICAgICAgcmV0dXJuIGNvbnRlbnRzLnJz
            dHJpcCgpCgogICAgZGVmIF9jaGVja19jb25mbGljdGluZ19vcHRpb25zKHNlbGYpOgogICAgICAg
            IGlmIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnVwZ3JhZGUgYW5kIG5vdCBzZWxmLl9hcmdfcGFyc2Vy
            LnJiZF9kYXRhX3Bvb2xfbmFtZToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICJFaXRoZXIgJy0tdXBncmFkZScgb3IgJy0tcmJkLWRh
            dGEtcG9vbC1uYW1lIDxwb29sX25hbWU+JyBzaG91bGQgYmUgc3BlY2lmaWVkIgogICAgICAgICAg
            ICApCgogICAgZGVmIF9pbnZhbGlkX2VuZHBvaW50KHNlbGYsIGVuZHBvaW50X3N0cik6CiAgICAg
            ICAgIyBleHRyYWN0IHRoZSBwb3J0IGJ5IGdldHRpbmcgdGhlIGxhc3Qgc3BsaXQgb24gYDpgIGRl
            bGltaXRlcgogICAgICAgIHRyeToKICAgICAgICAgICAgZW5kcG9pbnRfc3RyX2lwLCBwb3J0ID0g
            ZW5kcG9pbnRfc3RyLnJzcGxpdCgiOiIsIDEpCiAgICAgICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAg
            ICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oZiJOb3QgYSBwcm9wZXIg
            ZW5kcG9pbnQ6IHtlbmRwb2ludF9zdHJ9IikKCiAgICAgICAgdHJ5OgogICAgICAgICAgICBpZiBl
            bmRwb2ludF9zdHJfaXBbMF0gPT0gIlsiOgogICAgICAgICAgICAgICAgZW5kcG9pbnRfc3RyX2lw
            ID0gZW5kcG9pbnRfc3RyX2lwWzEgOiBsZW4oZW5kcG9pbnRfc3RyX2lwKSAtIDFdCiAgICAgICAg
            ICAgIGlwX3R5cGUgPSAoCiAgICAgICAgICAgICAgICAiSVB2NCIgaWYgdHlwZShpcF9hZGRyZXNz
            KGVuZHBvaW50X3N0cl9pcCkpIGlzIElQdjRBZGRyZXNzIGVsc2UgIklQdjYiCiAgICAgICAgICAg
            ICkKICAgICAgICBleGNlcHQgVmFsdWVFcnJvcjoKICAgICAgICAgICAgaXBfdHlwZSA9ICJGUURO
            IgogICAgICAgIGlmIG5vdCBwb3J0LmlzZGlnaXQoKToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0
            aW9uRmFpbHVyZUV4Y2VwdGlvbihmIlBvcnQgbm90IHZhbGlkOiB7cG9ydH0iKQogICAgICAgIGlu
            dFBvcnQgPSBpbnQocG9ydCkKICAgICAgICBpZiBpbnRQb3J0IDwgMSBvciBpbnRQb3J0ID4gMioq
            MTYgLSAxOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKGYiT3V0
            IG9mIHJhbmdlIHBvcnQgbnVtYmVyOiB7cG9ydH0iKQoKICAgICAgICByZXR1cm4gaXBfdHlwZQoK
            ICAgIGRlZiBlbmRwb2ludF9kaWFsKHNlbGYsIGVuZHBvaW50X3N0ciwgaXBfdHlwZSwgdGltZW91
            dD0zLCBjZXJ0PU5vbmUpOgogICAgICAgICMgaWYgdGhlICdjbHVzdGVyJyBpbnN0YW5jZSBpcyBh
            IGR1bW15IG9uZSwKICAgICAgICAjIGRvbid0IHRyeSB0byByZWFjaCBvdXQgdG8gdGhlIGVuZHBv
            aW50CiAgICAgICAgaWYgaXNpbnN0YW5jZShzZWxmLmNsdXN0ZXIsIER1bW15UmFkb3MpOgogICAg
            ICAgICAgICByZXR1cm4gIiIsICIiLCAiIgogICAgICAgIGlmIGlwX3R5cGUgPT0gIklQdjYiOgog
            ICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBlbmRwb2ludF9zdHJfaXAsIGVuZHBvaW50
            X3N0cl9wb3J0ID0gZW5kcG9pbnRfc3RyLnJzcGxpdCgiOiIsIDEpCiAgICAgICAgICAgIGV4Y2Vw
            dCBWYWx1ZUVycm9yOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICBmIk5vdCBhIHByb3BlciBlbmRwb2ludDoge2VuZHBv
            aW50X3N0cn0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmIGVuZHBvaW50X3N0cl9p
            cFswXSAhPSAiWyI6CiAgICAgICAgICAgICAgICBlbmRwb2ludF9zdHJfaXAgPSAiWyIgKyBlbmRw
            b2ludF9zdHJfaXAgKyAiXSIKICAgICAgICAgICAgZW5kcG9pbnRfc3RyID0gIjoiLmpvaW4oW2Vu
            ZHBvaW50X3N0cl9pcCwgZW5kcG9pbnRfc3RyX3BvcnRdKQoKICAgICAgICBwcm90b2NvbHMgPSBb
            Imh0dHAiLCAiaHR0cHMiXQogICAgICAgIHJlc3BvbnNlX2Vycm9yID0gTm9uZQogICAgICAgIGZv
            ciBwcmVmaXggaW4gcHJvdG9jb2xzOgogICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBl
            cCA9IGYie3ByZWZpeH06Ly97ZW5kcG9pbnRfc3RyfSIKICAgICAgICAgICAgICAgIHZlcmlmeSA9
            IE5vbmUKICAgICAgICAgICAgICAgICMgSWYgdmVyaWZ5IGlzIHNldCB0byBhIHBhdGggdG8gYSBk
            aXJlY3RvcnksCiAgICAgICAgICAgICAgICAjIHRoZSBkaXJlY3RvcnkgbXVzdCBoYXZlIGJlZW4g
            cHJvY2Vzc2VkIHVzaW5nIHRoZSBjX3JlaGFzaCB1dGlsaXR5IHN1cHBsaWVkIHdpdGggT3BlblNT
            TC4KICAgICAgICAgICAgICAgIGlmIHByZWZpeCA9PSAiaHR0cHMiIGFuZCBzZWxmLl9hcmdfcGFy
            c2VyLnJnd19za2lwX3RsczoKICAgICAgICAgICAgICAgICAgICB2ZXJpZnkgPSBGYWxzZQogICAg
            ICAgICAgICAgICAgICAgIHIgPSByZXF1ZXN0cy5oZWFkKGVwLCB0aW1lb3V0PXRpbWVvdXQsIHZl
            cmlmeT1GYWxzZSkKICAgICAgICAgICAgICAgIGVsaWYgcHJlZml4ID09ICJodHRwcyIgYW5kIGNl
            cnQ6CiAgICAgICAgICAgICAgICAgICAgdmVyaWZ5ID0gY2VydAogICAgICAgICAgICAgICAgICAg
            IHIgPSByZXF1ZXN0cy5oZWFkKGVwLCB0aW1lb3V0PXRpbWVvdXQsIHZlcmlmeT1jZXJ0KQogICAg
            ICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgICAgICByID0gcmVxdWVzdHMuaGVhZChl
            cCwgdGltZW91dD10aW1lb3V0KQogICAgICAgICAgICAgICAgaWYgci5zdGF0dXNfY29kZSA9PSAy
            MDA6CiAgICAgICAgICAgICAgICAgICAgcmV0dXJuIHByZWZpeCwgdmVyaWZ5LCAiIgogICAgICAg
            ICAgICBleGNlcHQgRXhjZXB0aW9uIGFzIGVycjoKICAgICAgICAgICAgICAgIHJlc3BvbnNlX2Vy
            cm9yID0gZXJyCiAgICAgICAgICAgICAgICBjb250aW51ZQogICAgICAgIHN5cy5zdGRlcnIud3Jp
            dGUoCiAgICAgICAgICAgIGYidW5hYmxlIHRvIGNvbm5lY3QgdG8gZW5kcG9pbnQ6IHtlbmRwb2lu
            dF9zdHJ9LCBmYWlsZWQgZXJyb3I6IHtyZXNwb25zZV9lcnJvcn0iCiAgICAgICAgKQogICAgICAg
            IHJldHVybiAoCiAgICAgICAgICAgICIiLAogICAgICAgICAgICAiIiwKICAgICAgICAgICAgKCIt
            MSIpLAogICAgICAgICkKCiAgICBkZWYgX19pbml0X18oc2VsZiwgYXJnX2xpc3Q9Tm9uZSk6CiAg
            ICAgICAgc2VsZi5vdXRfbWFwID0ge30KICAgICAgICBzZWxmLl9leGNsdWRlZF9rZXlzID0gc2V0
            KCkKICAgICAgICBzZWxmLl9hcmdfcGFyc2VyID0gc2VsZi5nZW5fYXJnX3BhcnNlcihhcmdzX3Rv
            X3BhcnNlPWFyZ19saXN0KQogICAgICAgIHNlbGYuX2NoZWNrX2NvbmZsaWN0aW5nX29wdGlvbnMo
            KQogICAgICAgIHNlbGYucnVuX2FzX3VzZXIgPSBzZWxmLl9hcmdfcGFyc2VyLnJ1bl9hc191c2Vy
            CiAgICAgICAgc2VsZi5vdXRwdXRfZmlsZSA9IHNlbGYuX2FyZ19wYXJzZXIub3V0cHV0CiAgICAg
            ICAgc2VsZi5jZXBoX2NvbmYgPSBzZWxmLl9hcmdfcGFyc2VyLmNlcGhfY29uZgogICAgICAgIHNl
            bGYuY2VwaF9rZXlyaW5nID0gc2VsZi5fYXJnX3BhcnNlci5rZXlyaW5nCiAgICAgICAgIyBpZiB1
            c2VyIG5vdCBwcm92aWRlZCwgZ2l2ZSBhIGRlZmF1bHQgdXNlcgogICAgICAgIGlmIG5vdCBzZWxm
            LnJ1bl9hc191c2VyIGFuZCBub3Qgc2VsZi5fYXJnX3BhcnNlci51cGdyYWRlOgogICAgICAgICAg
            ICBzZWxmLnJ1bl9hc191c2VyID0gc2VsZi5FWFRFUk5BTF9VU0VSX05BTUUKICAgICAgICBpZiBu
            b3Qgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXggYW5kIG5vdCBzZWxmLl9hcmdfcGFy
            c2VyLnVwZ3JhZGU6CiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4
            ID0gc2VsZi5ERUZBVUxUX1JHV19QT09MX1BSRUZJWAogICAgICAgIGlmIHNlbGYuY2VwaF9jb25m
            OgogICAgICAgICAgICBrd2FyZ3MgPSB7fQogICAgICAgICAgICBpZiBzZWxmLmNlcGhfa2V5cmlu
            ZzoKICAgICAgICAgICAgICAgIGt3YXJnc1siY29uZiJdID0geyJrZXlyaW5nIjogc2VsZi5jZXBo
            X2tleXJpbmd9CiAgICAgICAgICAgIHNlbGYuY2x1c3RlciA9IHJhZG9zLlJhZG9zKGNvbmZmaWxl
            PXNlbGYuY2VwaF9jb25mLCAqKmt3YXJncykKICAgICAgICBlbHNlOgogICAgICAgICAgICBzZWxm
            LmNsdXN0ZXIgPSByYWRvcy5SYWRvcygpCiAgICAgICAgICAgIHNlbGYuY2x1c3Rlci5jb25mX3Jl
            YWRfZmlsZSgpCiAgICAgICAgc2VsZi5jbHVzdGVyLmNvbm5lY3QoKQoKICAgIGRlZiBzaHV0ZG93
            bihzZWxmKToKICAgICAgICBpZiBzZWxmLmNsdXN0ZXIuc3RhdGUgPT0gImNvbm5lY3RlZCI6CiAg
            ICAgICAgICAgIHNlbGYuY2x1c3Rlci5zaHV0ZG93bigpCgogICAgZGVmIGdldF9mc2lkKHNlbGYp
            OgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJu
            IHNlbGYuZHJ5X3J1bigiY2VwaCBmc2lkIikKICAgICAgICByZXR1cm4gc3RyKHNlbGYuY2x1c3Rl
            ci5nZXRfZnNpZCgpKQoKICAgIGRlZiBfY29tbW9uX2NtZF9qc29uX2dlbihzZWxmLCBjbWRfanNv
            bik6CiAgICAgICAgY21kID0ganNvbi5kdW1wcyhjbWRfanNvbiwgc29ydF9rZXlzPVRydWUpCiAg
            ICAgICAgcmV0X3ZhbCwgY21kX291dCwgZXJyX21zZyA9IHNlbGYuY2x1c3Rlci5tb25fY29tbWFu
            ZChjbWQsIGIiIikKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnZlcmJvc2U6CiAgICAgICAg
            ICAgIHByaW50KGYiQ29tbWFuZCBJbnB1dDoge2NtZH0iKQogICAgICAgICAgICBwcmludCgKICAg
            ICAgICAgICAgICAgIGYiUmV0dXJuIFZhbDoge3JldF92YWx9XG5Db21tYW5kIE91dHB1dDoge2Nt
            ZF9vdXR9XG4iCiAgICAgICAgICAgICAgICBmIkVycm9yIE1lc3NhZ2U6IHtlcnJfbXNnfVxuLS0t
            LS0tLS0tLVxuIgogICAgICAgICAgICApCiAgICAgICAganNvbl9vdXQgPSB7fQogICAgICAgICMg
            aWYgdGhlcmUgaXMgbm8gZXJyb3IgKGkuZTsgcmV0X3ZhbCBpcyBaRVJPKSBhbmQgJ2NtZF9vdXQn
            IGlzIG5vdCBlbXB0eQogICAgICAgICMgdGhlbiBjb252ZXJ0ICdjbWRfb3V0JyB0byBhIGpzb24g
            b3V0cHV0CiAgICAgICAgaWYgcmV0X3ZhbCA9PSAwIGFuZCBjbWRfb3V0OgogICAgICAgICAgICBq
            c29uX291dCA9IGpzb24ubG9hZHMoY21kX291dCkKICAgICAgICByZXR1cm4gcmV0X3ZhbCwganNv
            bl9vdXQsIGVycl9tc2cKCiAgICBkZWYgZ2V0X2NlcGhfZXh0ZXJuYWxfbW9uX2RhdGEoc2VsZik6
            CiAgICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJxdW9ydW1fc3RhdHVzIiwgImZvcm1hdCI6
            ICJqc29uIn0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAg
            IHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGggIiArIGNtZF9qc29uWyJwcmVmaXgiXSkKICAgICAg
            ICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4o
            Y21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwK
            ICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgIidxdW9y
            dW1fc3RhdHVzJyBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAgICAgIGYiRXJyb3I6IHtl
            cnJfbXNnIGlmIHJldF92YWwgIT0gMCBlbHNlIHNlbGYuRU1QVFlfT1VUUFVUX0xJU1R9IgogICAg
            ICAgICAgICApCiAgICAgICAgcV9sZWFkZXJfbmFtZSA9IGpzb25fb3V0WyJxdW9ydW1fbGVhZGVy
            X25hbWUiXQogICAgICAgIHFfbGVhZGVyX2RldGFpbHMgPSB7fQogICAgICAgIHFfbGVhZGVyX21h
            dGNoaW5nX2xpc3QgPSBbCiAgICAgICAgICAgIGwgZm9yIGwgaW4ganNvbl9vdXRbIm1vbm1hcCJd
            WyJtb25zIl0gaWYgbFsibmFtZSJdID09IHFfbGVhZGVyX25hbWUKICAgICAgICBdCiAgICAgICAg
            aWYgbGVuKHFfbGVhZGVyX21hdGNoaW5nX2xpc3QpID09IDA6CiAgICAgICAgICAgIHJhaXNlIEV4
            ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oIk5vIG1hdGNoaW5nICdtb24nIGRldGFpbHMgZm91bmQi
            KQogICAgICAgIHFfbGVhZGVyX2RldGFpbHMgPSBxX2xlYWRlcl9tYXRjaGluZ19saXN0WzBdCiAg
            ICAgICAgIyBnZXQgdGhlIGFkZHJlc3MgdmVjdG9yIG9mIHRoZSBxdW9ydW0tbGVhZGVyCiAgICAg
            ICAgcV9sZWFkZXJfYWRkcnZlYyA9IHFfbGVhZGVyX2RldGFpbHMuZ2V0KCJwdWJsaWNfYWRkcnMi
            LCB7fSkuZ2V0KCJhZGRydmVjIiwgW10pCiAgICAgICAgaXBfYWRkciA9IHN0cihxX2xlYWRlcl9k
            ZXRhaWxzWyJwdWJsaWNfYWRkciJdLnNwbGl0KCIvIilbMF0pCgogICAgICAgIGlmIHNlbGYuX2Fy
            Z19wYXJzZXIudjJfcG9ydF9lbmFibGU6CiAgICAgICAgICAgIGlmIHFfbGVhZGVyX2FkZHJ2ZWNb
            MF1bInR5cGUiXSA9PSAidjIiOgogICAgICAgICAgICAgICAgaXBfYWRkciA9IHFfbGVhZGVyX2Fk
            ZHJ2ZWNbMF1bImFkZHIiXQogICAgICAgICAgICBlbGlmIGxlbihxX2xlYWRlcl9hZGRydmVjKSA+
            IDEgYW5kIHFfbGVhZGVyX2FkZHJ2ZWNbMV1bInR5cGUiXSA9PSAidjIiOgogICAgICAgICAgICAg
            ICAgaXBfYWRkciA9IHFfbGVhZGVyX2FkZHJ2ZWNbMV1bImFkZHIiXQogICAgICAgICAgICBlbHNl
            OgogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZSgKICAgICAgICAgICAgICAgICAgICAi
            J3YyJyBhZGRyZXNzIHR5cGUgbm90IHByZXNlbnQsIGFuZCAndjItcG9ydC1lbmFibGUnIGZsYWcg
            aXMgcHJvdmlkZWQiCiAgICAgICAgICAgICAgICApCgogICAgICAgIHJldHVybiBmIntzdHIocV9s
            ZWFkZXJfbmFtZSl9PXtpcF9hZGRyfSIKCiAgICBkZWYgX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAo
            c2VsZiwgaG9zdF9uYW1lLCBwb3J0LCBpcF90eXBlKToKICAgICAgICAjIGlmICdjbHVzdGVyJyBp
            bnN0YW5jZSBpcyBhIGR1bW15IHR5cGUsCiAgICAgICAgIyBjYWxsIHRoZSBkdW1teSBpbnN0YW5j
            ZSdzICJjb252ZXJ0IiBtZXRob2QKICAgICAgICBpZiBub3QgaG9zdF9uYW1lOgogICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKCJFbXB0eSBob3N0bmFtZSBwcm92aWRl
            ZCIpCiAgICAgICAgaWYgaXNpbnN0YW5jZShzZWxmLmNsdXN0ZXIsIER1bW15UmFkb3MpOgogICAg
            ICAgICAgICByZXR1cm4gc2VsZi5jbHVzdGVyLl9jb252ZXJ0X2hvc3RuYW1lX3RvX2lwKGhvc3Rf
            bmFtZSkKCiAgICAgICAgaWYgaXBfdHlwZSA9PSAiRlFETiI6CiAgICAgICAgICAgICMgY2hlY2sg
            d2hpY2ggaXAgRlFETiBzaG91bGQgYmUgY29udmVydGVkIHRvLCBJUHY0IG9yIElQdjYKICAgICAg
            ICAgICAgIyBjaGVjayB0aGUgaG9zdCBpcCwgdGhlIGVuZHBvaW50IGlwIHR5cGUgd291bGQgYmUg
            c2ltaWxhciB0byBob3N0IGlwCiAgICAgICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAib3Jj
            aCBob3N0IGxzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICAgICAgcmV0X3ZhbCwganNvbl9v
            dXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAg
            ICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAogICAgICAgICAgICBp
            ZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICAgICAgcmFp
            c2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJ29yY2gg
            aG9zdCBscycgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAgICAgICAgICAgICAgICAgZiJFcnJvcjog
            e2Vycl9tc2cgaWYgcmV0X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAg
            ICAgICAgICAgICAgICApCiAgICAgICAgICAgIGhvc3RfYWRkciA9IGpzb25fb3V0WzBdWyJhZGRy
            Il0KICAgICAgICAgICAgIyBhZGQgOjgwIHNhbXBsZSBwb3J0IGluIGlwX3R5cGUsIGFzIF9pbnZh
            bGlkX2VuZHBvaW50IGFsc28gdmVyaWZ5IHBvcnQKICAgICAgICAgICAgaG9zdF9pcF90eXBlID0g
            c2VsZi5faW52YWxpZF9lbmRwb2ludChob3N0X2FkZHIgKyAiOjgwIikKICAgICAgICAgICAgaW1w
            b3J0IHNvY2tldAoKICAgICAgICAgICAgIyBleGFtcGxlIG91dHB1dCBbKDxBZGRyZXNzRmFtaWx5
            LkFGX0lORVQ6IDI+LCA8U29ja2V0S2luZC5TT0NLX1NUUkVBTTogMT4sIDYsICcnLCAoJzkzLjE4
            NC4yMTYuMzQnLCA4MCkpLCAuLi5dCiAgICAgICAgICAgICMgd2UgbmVlZCB0byBnZXQgOTMuMTg0
            LjIxNi4zNCBzbyBpdCB3b3VsZCBiZSBpcFswXVs0XVswXQogICAgICAgICAgICBpZiBob3N0X2lw
            X3R5cGUgPT0gIklQdjYiOgogICAgICAgICAgICAgICAgaXAgPSBzb2NrZXQuZ2V0YWRkcmluZm8o
            CiAgICAgICAgICAgICAgICAgICAgaG9zdF9uYW1lLCBwb3J0LCBmYW1pbHk9c29ja2V0LkFGX0lO
            RVQ2LCBwcm90bz1zb2NrZXQuSVBQUk9UT19UQ1AKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgZWxpZiBob3N0X2lwX3R5cGUgPT0gIklQdjQiOgogICAgICAgICAgICAgICAgaXAgPSBzb2Nr
            ZXQuZ2V0YWRkcmluZm8oCiAgICAgICAgICAgICAgICAgICAgaG9zdF9uYW1lLCBwb3J0LCBmYW1p
            bHk9c29ja2V0LkFGX0lORVQsIHByb3RvPXNvY2tldC5JUFBST1RPX1RDUAogICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICBkZWwgc29ja2V0CiAgICAgICAgICAgIHJldHVybiBpcFswXVs0XVsw
            XQogICAgICAgIHJldHVybiBob3N0X25hbWUKCiAgICBkZWYgZ2V0X2FjdGl2ZV9hbmRfc3RhbmRi
            eV9tZ3JzKHNlbGYpOgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAg
            ICAgICAgcmV0dXJuICIiLCBzZWxmLmRyeV9ydW4oImNlcGggc3RhdHVzIikKICAgICAgICBtb25p
            dG9yaW5nX2VuZHBvaW50X3BvcnQgPSBzZWxmLl9hcmdfcGFyc2VyLm1vbml0b3JpbmdfZW5kcG9p
            bnRfcG9ydAogICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdCA9IHNlbGYuX2FyZ19w
            YXJzZXIubW9uaXRvcmluZ19lbmRwb2ludAogICAgICAgIHN0YW5kYnlfbWdycyA9IFtdCiAgICAg
            ICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdDoKICAgICAgICAgICAgY21kX2pz
            b24gPSB7InByZWZpeCI6ICJzdGF0dXMiLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgICAgICBy
            ZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21k
            X2pzb24pCiAgICAgICAgICAgICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQs
            CiAgICAgICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAgICAg
            ICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAg
            ICAgICAgICInbWdyIHNlcnZpY2VzJyBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAgICAg
            ICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09V
            VFBVVF9MSVNUfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRw
            b2ludCA9ICgKICAgICAgICAgICAgICAgIGpzb25fb3V0LmdldCgibWdybWFwIiwge30pLmdldCgi
            c2VydmljZXMiLCB7fSkuZ2V0KCJwcm9tZXRoZXVzIiwgIiIpCiAgICAgICAgICAgICkKICAgICAg
            ICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnQ6CiAgICAgICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJjYW4ndCBmaW5k
            IG1vbml0b3JpbmdfZW5kcG9pbnQsIHByb21ldGhldXMgbW9kdWxlIG1pZ2h0IG5vdCBiZSBlbmFi
            bGVkLCAiCiAgICAgICAgICAgICAgICAgICAgImVuYWJsZSB0aGUgbW9kdWxlIGJ5IHJ1bm5pbmcg
            J2NlcGggbWdyIG1vZHVsZSBlbmFibGUgcHJvbWV0aGV1cyciCiAgICAgICAgICAgICAgICApCiAg
            ICAgICAgICAgICMgbm93IGNoZWNrIHRoZSBzdGFuZC1ieSBtZ3ItcwogICAgICAgICAgICBzdGFu
            ZGJ5X2FyciA9IGpzb25fb3V0LmdldCgibWdybWFwIiwge30pLmdldCgic3RhbmRieXMiLCBbXSkK
            ICAgICAgICAgICAgZm9yIGVhY2hfc3RhbmRieSBpbiBzdGFuZGJ5X2FycjoKICAgICAgICAgICAg
            ICAgIGlmICJuYW1lIiBpbiBlYWNoX3N0YW5kYnkua2V5cygpOgogICAgICAgICAgICAgICAgICAg
            IHN0YW5kYnlfbWdycy5hcHBlbmQoZWFjaF9zdGFuZGJ5WyJuYW1lIl0pCiAgICAgICAgICAgIHRy
            eToKICAgICAgICAgICAgICAgIHBhcnNlZF9lbmRwb2ludCA9IHVybHBhcnNlKG1vbml0b3Jpbmdf
            ZW5kcG9pbnQpCiAgICAgICAgICAgIGV4Y2VwdCBWYWx1ZUVycm9yOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBmImlu
            dmFsaWQgZW5kcG9pbnQ6IHttb25pdG9yaW5nX2VuZHBvaW50fSIKICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0ID0gcGFyc2VkX2VuZHBvaW50
            Lmhvc3RuYW1lCiAgICAgICAgICAgIGlmIG5vdCBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQ6CiAg
            ICAgICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQgPSBzdHIocGFyc2VkX2VuZHBv
            aW50LnBvcnQpCgogICAgICAgICMgaWYgbW9uaXRvcmluZyBlbmRwb2ludCBwb3J0IGlzIG5vdCBz
            ZXQsIHB1dCBhIGRlZmF1bHQgbW9uIHBvcnQKICAgICAgICBpZiBub3QgbW9uaXRvcmluZ19lbmRw
            b2ludF9wb3J0OgogICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQgPSBzZWxmLkRF
            RkFVTFRfTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUCgogICAgICAgICMgdXNlciBjb3VsZCBnaXZl
            IGNvbW1hIGFuZCBzcGFjZSBzZXBhcmF0ZWQgaW5wdXRzIChsaWtlIC0tbW9uaXRvcmluZy1lbmRw
            b2ludD0iPGlwMT4sIDxpcDI+IikKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Qg
            PSBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3QucmVwbGFjZSgiLCIsICIgIikKICAgICAgICBt
            b25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXQgPSBtb25pdG9yaW5nX2VuZHBvaW50X2lw
            X2xpc3Quc3BsaXQoKQogICAgICAgICMgaWYgbW9uaXRvcmluZy1lbmRwb2ludCBjb3VsZCBub3Qg
            YmUgZm91bmQsIHJhaXNlIGFuIGVycm9yCiAgICAgICAgaWYgbGVuKG1vbml0b3JpbmdfZW5kcG9p
            bnRfaXBfbGlzdF9zcGxpdCkgPT0gMDoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVy
            ZUV4Y2VwdGlvbigiTm8gJ21vbml0b3JpbmctZW5kcG9pbnQnIGZvdW5kIikKICAgICAgICAjIGZp
            cnN0IGlwIGlzIHRyZWF0ZWQgYXMgdGhlIG1haW4gbW9uaXRvcmluZy1lbmRwb2ludAogICAgICAg
            IG1vbml0b3JpbmdfZW5kcG9pbnRfaXAgPSBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3Bs
            aXRbMF0KICAgICAgICAjIHJlc3Qgb2YgdGhlIGlwLXMgYXJlIGFkZGVkIHRvIHRoZSAnc3RhbmRi
            eV9tZ3JzJyBsaXN0CiAgICAgICAgc3RhbmRieV9tZ3JzLmV4dGVuZChtb25pdG9yaW5nX2VuZHBv
            aW50X2lwX2xpc3Rfc3BsaXRbMTpdKQogICAgICAgIGZhaWxlZF9pcCA9IG1vbml0b3JpbmdfZW5k
            cG9pbnRfaXAKCiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludCA9ICI6Ii5qb2luKAogICAgICAg
            ICAgICBbbW9uaXRvcmluZ19lbmRwb2ludF9pcCwgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0XQog
            ICAgICAgICkKICAgICAgICBpcF90eXBlID0gc2VsZi5faW52YWxpZF9lbmRwb2ludChtb25pdG9y
            aW5nX2VuZHBvaW50KQogICAgICAgIHRyeToKICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2lu
            dF9pcCA9IHNlbGYuX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoCiAgICAgICAgICAgICAgICBtb25p
            dG9yaW5nX2VuZHBvaW50X2lwLCBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQsIGlwX3R5cGUKICAg
            ICAgICAgICAgKQogICAgICAgICAgICAjIGNvbGxlY3QgYWxsIHRoZSAnc3RhbmQtYnknIG1nciBp
            cHMKICAgICAgICAgICAgbWdyX2lwcyA9IFtdCiAgICAgICAgICAgIGZvciBlYWNoX3N0YW5kYnlf
            bWdyIGluIHN0YW5kYnlfbWdyczoKICAgICAgICAgICAgICAgIGZhaWxlZF9pcCA9IGVhY2hfc3Rh
            bmRieV9tZ3IKICAgICAgICAgICAgICAgIG1ncl9pcHMuYXBwZW5kKAogICAgICAgICAgICAgICAg
            ICAgIHNlbGYuX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoCiAgICAgICAgICAgICAgICAgICAgICAg
            IGVhY2hfc3RhbmRieV9tZ3IsIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydCwgaXBfdHlwZQogICAg
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgICkKICAgICAgICBleGNlcHQ6CiAgICAg
            ICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBm
            IkNvbnZlcnNpb24gb2YgaG9zdDoge2ZhaWxlZF9pcH0gdG8gSVAgZmFpbGVkLiAiCiAgICAgICAg
            ICAgICAgICAiUGxlYXNlIGVudGVyIHRoZSBJUCBhZGRyZXNzZXMgb2YgYWxsIHRoZSBjZXBoLW1n
            cnMgd2l0aCB0aGUgJy0tbW9uaXRvcmluZy1lbmRwb2ludCcgZmxhZyIKICAgICAgICAgICAgKQoK
            ICAgICAgICBfLCBfLCBlcnIgPSBzZWxmLmVuZHBvaW50X2RpYWwobW9uaXRvcmluZ19lbmRwb2lu
            dCwgaXBfdHlwZSkKICAgICAgICBpZiBlcnIgPT0gIi0xIjoKICAgICAgICAgICAgcmFpc2UgRXhl
            Y3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihlcnIpCiAgICAgICAgIyBhZGQgdGhlIHZhbGlkYXRlZCBh
            Y3RpdmUgbWdyIElQIGludG8gdGhlIGZpcnN0IGluZGV4CiAgICAgICAgbWdyX2lwcy5pbnNlcnQo
            MCwgbW9uaXRvcmluZ19lbmRwb2ludF9pcCkKICAgICAgICBhbGxfbWdyX2lwc19zdHIgPSAiLCIu
            am9pbihtZ3JfaXBzKQogICAgICAgIHJldHVybiBhbGxfbWdyX2lwc19zdHIsIG1vbml0b3Jpbmdf
            ZW5kcG9pbnRfcG9ydAoKICAgIGRlZiBjaGVja191c2VyX2V4aXN0KHNlbGYsIHVzZXIpOgogICAg
            ICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAiYXV0aCBnZXQiLCAiZW50aXR5IjogZiJ7dXNlcn0i
            LCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBfID0gc2VsZi5f
            Y29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3Ig
            bGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICByZXR1cm4gIiIKICAgICAgICByZXR1cm4g
            c3RyKGpzb25fb3V0WzBdWyJrZXkiXSkKCiAgICBkZWYgZ2V0X2NlcGhmc19wcm92aXNpb25lcl9j
            YXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gImNsaWVudC5jc2ktY2VwaGZz
            LXByb3Zpc2lvbmVyIgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAgICJtb24iOiAiYWxsb3cg
            ciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLAogICAgICAgICAgICAibWdyIjogImFs
            bG93IHJ3IiwKICAgICAgICAgICAgIm9zZCI6ICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRh
            PSoiLAogICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0
            aF9wZXJtaXNzaW9uOgogICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3Bh
            cnNlci5rOHNfY2x1c3Rlcl9uYW1lCiAgICAgICAgICAgIGlmIGs4c19jbHVzdGVyX25hbWUgPT0g
            IiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAg
            ICAgICAgICAgICAgICAgICJrOHNfY2x1c3Rlcl9uYW1lIG5vdCBmb3VuZCwgcGxlYXNlIHNldCB0
            aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgY2VwaGZzX2ZpbGVzeXN0ZW0gPSBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVt
            X25hbWUKICAgICAgICAgICAgaWYgY2VwaGZzX2ZpbGVzeXN0ZW0gPT0gIiI6CiAgICAgICAgICAg
            ICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9uYW1lfSIKICAgICAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0te2s4c19jbHVzdGVyX25h
            bWV9LXtjZXBoZnNfZmlsZXN5c3RlbX0iCiAgICAgICAgICAgICAgICBjYXBzWyJvc2QiXSA9IGYi
            YWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT17Y2VwaGZzX2ZpbGVzeXN0ZW19IgoKICAgICAg
            ICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVmIGdldF9jZXBoZnNfbm9kZV9jYXBzX2FuZF9l
            bnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiCiAg
            ICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJhbGxvdyByLCBhbGxvdyBjb21tYW5k
            ICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJtZ3IiOiAiYWxsb3cgcnciLAogICAgICAg
            ICAgICAib3NkIjogImFsbG93IHJ3IHRhZyBjZXBoZnMgKj0qIiwKICAgICAgICAgICAgIm1kcyI6
            ICJhbGxvdyBydyIsCiAgICAgICAgfQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJp
            Y3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAgIGs4c19jbHVzdGVyX25hbWUgPSBzZWxm
            Ll9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAgICAgICAgICAgaWYgazhzX2NsdXN0ZXJf
            bmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRp
            b24oCiAgICAgICAgICAgICAgICAgICAgIms4c19jbHVzdGVyX25hbWUgbm90IGZvdW5kLCBwbGVh
            c2Ugc2V0IHRoZSAnLS1rOHMtY2x1c3Rlci1uYW1lJyBmbGFnIgogICAgICAgICAgICAgICAgKQog
            ICAgICAgICAgICBjZXBoZnNfZmlsZXN5c3RlbSA9IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2Zp
            bGVzeXN0ZW1fbmFtZQogICAgICAgICAgICBpZiBjZXBoZnNfZmlsZXN5c3RlbSA9PSAiIjoKICAg
            ICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0te2s4c19jbHVzdGVyX25hbWV9IgogICAg
            ICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17azhzX2Ns
            dXN0ZXJfbmFtZX0te2NlcGhmc19maWxlc3lzdGVtfSIKICAgICAgICAgICAgICAgIGNhcHNbIm9z
            ZCJdID0gZiJhbGxvdyBydyB0YWcgY2VwaGZzICo9e2NlcGhmc19maWxlc3lzdGVtfSIKCiAgICAg
            ICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfZW50aXR5KAogICAgICAgIHNlbGYs
            CiAgICAgICAgZW50aXR5LAogICAgICAgIHJiZF9wb29sX25hbWUsCiAgICAgICAgYWxpYXNfcmJk
            X3Bvb2xfbmFtZSwKICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lLAogICAgICAgIHJhZG9zX25hbWVz
            cGFjZSwKICAgICk6CiAgICAgICAgaWYgKAogICAgICAgICAgICByYmRfcG9vbF9uYW1lLmNvdW50
            KCIuIikgIT0gMAogICAgICAgICAgICBvciByYmRfcG9vbF9uYW1lLmNvdW50KCJfIikgIT0gMAog
            ICAgICAgICAgICBvciBhbGlhc19yYmRfcG9vbF9uYW1lICE9ICIiCiAgICAgICAgICAgICMgY2hl
            Y2tpbmcgYWxpYXNfcmJkX3Bvb2xfbmFtZSBpcyBub3QgZW1wdHkgYXMgdGhlcmUgbWF5YmUgYSBz
            cGVjaWFsIGNoYXJhY3RlciB1c2VkIG90aGVyIHRoYW4gLiBvciBfCiAgICAgICAgKToKICAgICAg
            ICAgICAgaWYgYWxpYXNfcmJkX3Bvb2xfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgInBsZWFzZSBz
            ZXQgdGhlICctLWFsaWFzLXJiZC1kYXRhLXBvb2wtbmFtZScgZmxhZyBhcyB0aGUgcmJkIGRhdGEg
            cG9vbCBuYW1lIGNvbnRhaW5zICcuJyBvciAnXyciCiAgICAgICAgICAgICAgICApCiAgICAgICAg
            ICAgIGlmICgKICAgICAgICAgICAgICAgIGFsaWFzX3JiZF9wb29sX25hbWUuY291bnQoIi4iKSAh
            PSAwCiAgICAgICAgICAgICAgICBvciBhbGlhc19yYmRfcG9vbF9uYW1lLmNvdW50KCJfIikgIT0g
            MAogICAgICAgICAgICApOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJy0tYWxpYXMtcmJkLWRhdGEtcG9vbC1uYW1l
            JyBmbGFnIHZhbHVlIHNob3VsZCBub3QgY29udGFpbiAnLicgb3IgJ18nIgogICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9uYW1lfS17
            YWxpYXNfcmJkX3Bvb2xfbmFtZX0iCiAgICAgICAgZWxzZToKICAgICAgICAgICAgZW50aXR5ID0g
            ZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0te3JiZF9wb29sX25hbWV9IgoKICAgICAgICBp
            ZiByYWRvc19uYW1lc3BhY2U6CiAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0te3JhZG9z
            X25hbWVzcGFjZX0iCiAgICAgICAgcmV0dXJuIGVudGl0eQoKICAgIGRlZiBnZXRfcmJkX3Byb3Zp
            c2lvbmVyX2NhcHNfYW5kX2VudGl0eShzZWxmKToKICAgICAgICBlbnRpdHkgPSAiY2xpZW50LmNz
            aS1yYmQtcHJvdmlzaW9uZXIiCiAgICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJw
            cm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLAogICAgICAgICAgICAi
            bWdyIjogImFsbG93IHJ3IiwKICAgICAgICAgICAgIm9zZCI6ICJwcm9maWxlIHJiZCIsCiAgICAg
            ICAgfQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Np
            b246CiAgICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRh
            X3Bvb2xfbmFtZQogICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3Bh
            cnNlci5hbGlhc19yYmRfZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFt
            ZSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQogICAgICAgICAgICByYWRvc19u
            YW1lc3BhY2UgPSBzZWxmLl9hcmdfcGFyc2VyLnJhZG9zX25hbWVzcGFjZQogICAgICAgICAgICBp
            ZiByYmRfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAibWFuZGF0b3J5IGZsYWcgbm90IGZv
            dW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1yYmQtZGF0YS1wb29sLW5hbWUnIGZsYWciCiAgICAgICAg
            ICAgICAgICApCiAgICAgICAgICAgIGlmIGs4c19jbHVzdGVyX25hbWUgPT0gIiI6CiAgICAgICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLWs4cy1jbHVz
            dGVyLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVudGl0eSA9IHNl
            bGYuZ2V0X2VudGl0eSgKICAgICAgICAgICAgICAgIGVudGl0eSwKICAgICAgICAgICAgICAgIHJi
            ZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lLAogICAgICAg
            ICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAgICAgICAgICAgIHJhZG9zX25hbWVzcGFj
            ZSwKICAgICAgICAgICAgKQogICAgICAgICAgICBpZiByYWRvc19uYW1lc3BhY2UgIT0gIiI6CiAg
            ICAgICAgICAgICAgICBjYXBzWyJvc2QiXSA9ICgKICAgICAgICAgICAgICAgICAgICBmInByb2Zp
            bGUgcmJkIHBvb2w9e3JiZF9wb29sX25hbWV9IG5hbWVzcGFjZT17cmFkb3NfbmFtZXNwYWNlfSIK
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGNhcHNb
            Im9zZCJdID0gZiJwcm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9uYW1lfSIKCiAgICAgICAgcmV0
            dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfcmJkX25vZGVfY2Fwc19hbmRfZW50aXR5KHNl
            bGYpOgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3NpLXJiZC1ub2RlIgogICAgICAgIGNhcHMg
            PSB7CiAgICAgICAgICAgICJtb24iOiAicHJvZmlsZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBi
            bG9ja2xpc3QnIiwKICAgICAgICAgICAgIm9zZCI6ICJwcm9maWxlIHJiZCIsCiAgICAgICAgfQog
            ICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb246CiAg
            ICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xf
            bmFtZQogICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5h
            bGlhc19yYmRfZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNl
            bGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQogICAgICAgICAgICByYWRvc19uYW1lc3Bh
            Y2UgPSBzZWxmLl9hcmdfcGFyc2VyLnJhZG9zX25hbWVzcGFjZQogICAgICAgICAgICBpZiByYmRf
            cG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAibWFuZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBw
            bGVhc2Ugc2V0IHRoZSAnLS1yYmQtZGF0YS1wb29sLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAg
            ICApCiAgICAgICAgICAgIGlmIGs4c19jbHVzdGVyX25hbWUgPT0gIiI6CiAgICAgICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJt
            YW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLWs4cy1jbHVzdGVyLW5h
            bWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVudGl0eSA9IHNlbGYuZ2V0
            X2VudGl0eSgKICAgICAgICAgICAgICAgIGVudGl0eSwKICAgICAgICAgICAgICAgIHJiZF9wb29s
            X25hbWUsCiAgICAgICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lLAogICAgICAgICAgICAg
            ICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAgICAgICAgICAgIHJhZG9zX25hbWVzcGFjZSwKICAg
            ICAgICAgICAgKQogICAgICAgICAgICBpZiByYWRvc19uYW1lc3BhY2UgIT0gIiI6CiAgICAgICAg
            ICAgICAgICBjYXBzWyJvc2QiXSA9ICgKICAgICAgICAgICAgICAgICAgICBmInByb2ZpbGUgcmJk
            IHBvb2w9e3JiZF9wb29sX25hbWV9IG5hbWVzcGFjZT17cmFkb3NfbmFtZXNwYWNlfSIKICAgICAg
            ICAgICAgICAgICkKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJd
            ID0gZiJwcm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9uYW1lfSIKCiAgICAgICAgcmV0dXJuIGNh
            cHMsIGVudGl0eQoKICAgIGRlZiBnZXRfZGVmYXVsdFVzZXJfY2Fwc19hbmRfZW50aXR5KHNlbGYp
            OgogICAgICAgIGVudGl0eSA9IHNlbGYucnVuX2FzX3VzZXIKICAgICAgICBjYXBzID0gewogICAg
            ICAgICAgICAibW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxs
            b3cgY29tbWFuZCB2ZXJzaW9uIiwKICAgICAgICAgICAgIm1nciI6ICJhbGxvdyBjb21tYW5kIGNv
            bmZpZyIsCiAgICAgICAgICAgICJvc2QiOiBmInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cg
            cnd4IHBvb2w9e3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cubWV0YSwgYWxs
            b3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD17c2VsZi5fYXJnX3BhcnNlci5yZ3df
            cG9vbF9wcmVmaXh9LnJndy5jb250cm9sLCBhbGxvdyByeCBwb29sPXtzZWxmLl9hcmdfcGFyc2Vy
            LnJnd19wb29sX3ByZWZpeH0ucmd3LmxvZywgYWxsb3cgeCBwb29sPXtzZWxmLl9hcmdfcGFyc2Vy
            LnJnd19wb29sX3ByZWZpeH0ucmd3LmJ1Y2tldHMuaW5kZXgiLAogICAgICAgIH0KCiAgICAgICAg
            cmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfY2Fwc19hbmRfZW50aXR5KHNlbGYsIHVz
            ZXJfbmFtZSk6CiAgICAgICAgaWYgImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiBpbiB1
            c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIg
            IT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVk
            X2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZ2V0X2NlcGhm
            c19wcm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoKQogICAgICAgIGlmICJjbGllbnQuY3NpLWNl
            cGhmcy1ub2RlIiBpbiB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQuY3NpLWNlcGhm
            cy1ub2RlIiAhPSB1c2VyX25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJl
            c3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5n
            ZXRfY2VwaGZzX25vZGVfY2Fwc19hbmRfZW50aXR5KCkKICAgICAgICBpZiAiY2xpZW50LmNzaS1y
            YmQtcHJvdmlzaW9uZXIiIGluIHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVudC5jc2kt
            cmJkLXByb3Zpc2lvbmVyIiAhPSB1c2VyX25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdf
            cGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1
            cm4gc2VsZi5nZXRfcmJkX3Byb3Zpc2lvbmVyX2NhcHNfYW5kX2VudGl0eSgpCiAgICAgICAgaWYg
            ImNsaWVudC5jc2ktcmJkLW5vZGUiIGluIHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVu
            dC5jc2ktcmJkLW5vZGUiICE9IHVzZXJfbmFtZToKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19w
            YXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVlCiAgICAgICAgICAgIHJldHVy
            biBzZWxmLmdldF9yYmRfbm9kZV9jYXBzX2FuZF9lbnRpdHkoKQogICAgICAgIGlmICJjbGllbnQu
            aGVhbHRoY2hlY2tlciIgaW4gdXNlcl9uYW1lOgogICAgICAgICAgICBpZiAiY2xpZW50LmhlYWx0
            aGNoZWNrZXIiICE9IHVzZXJfbmFtZToKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIu
            cmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVlCiAgICAgICAgICAgIHJldHVybiBzZWxm
            LmdldF9kZWZhdWx0VXNlcl9jYXBzX2FuZF9lbnRpdHkoKQoKICAgICAgICByYWlzZSBFeGVjdXRp
            b25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICBmIm5vIHVzZXIgZm91bmQgd2l0aCB1c2Vy
            X25hbWU6IHt1c2VyX25hbWV9LCAiCiAgICAgICAgICAgICJnZXRfY2Fwc19hbmRfZW50aXR5IGNv
            bW1hbmQgZmFpbGVkLlxuIgogICAgICAgICkKCiAgICBkZWYgY3JlYXRlX2NlcGhDU0lLZXlyaW5n
            X3VzZXIoc2VsZiwgdXNlcik6CiAgICAgICAgIiIiCiAgICAgICAgY29tbWFuZDogY2VwaCBhdXRo
            IGdldC1vci1jcmVhdGUgY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIgbW9uICdhbGxvdyBy
            JyBtZ3IgJ2FsbG93IHJ3JyBvc2QgJ2FsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KicKICAg
            ICAgICAiIiIKICAgICAgICBjYXBzLCBlbnRpdHkgPSBzZWxmLmdldF9jYXBzX2FuZF9lbnRpdHko
            dXNlcikKICAgICAgICBjbWRfanNvbiA9IHsKICAgICAgICAgICAgInByZWZpeCI6ICJhdXRoIGdl
            dC1vci1jcmVhdGUiLAogICAgICAgICAgICAiZW50aXR5IjogZW50aXR5LAogICAgICAgICAgICAi
            Y2FwcyI6IFtjYXAgZm9yIGNhcF9saXN0IGluIGxpc3QoY2Fwcy5pdGVtcygpKSBmb3IgY2FwIGlu
            IGNhcF9saXN0XSwKICAgICAgICAgICAgImZvcm1hdCI6ICJqc29uIiwKICAgICAgICB9CgogICAg
            ICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuICgKICAg
            ICAgICAgICAgICAgIHNlbGYuZHJ5X3J1bigKICAgICAgICAgICAgICAgICAgICAiY2VwaCAiCiAg
            ICAgICAgICAgICAgICAgICAgKyBjbWRfanNvblsicHJlZml4Il0KICAgICAgICAgICAgICAgICAg
            ICArICIgIgogICAgICAgICAgICAgICAgICAgICsgY21kX2pzb25bImVudGl0eSJdCiAgICAgICAg
            ICAgICAgICAgICAgKyAiICIKICAgICAgICAgICAgICAgICAgICArICIgIi5qb2luKGNtZF9qc29u
            WyJjYXBzIl0pCiAgICAgICAgICAgICAgICApLAogICAgICAgICAgICAgICAgIiIsCiAgICAgICAg
            ICAgICkKICAgICAgICAjIGNoZWNrIGlmIHVzZXIgYWxyZWFkeSBleGlzdAogICAgICAgIHVzZXJf
            a2V5ID0gc2VsZi5jaGVja191c2VyX2V4aXN0KGVudGl0eSkKICAgICAgICBpZiB1c2VyX2tleSAh
            PSAiIjoKICAgICAgICAgICAgcmV0dXJuIHVzZXJfa2V5LCBmIntlbnRpdHkuc3BsaXQoJy4nLCAx
            KVsxXX0iCiAgICAgICAgICAgICMgZW50aXR5LnNwbGl0KCcuJywxKVsxXSB0byByZW5hbWUgZW50
            aXR5KGNsaWVudC5jc2ktcmJkLW5vZGUpIGFzIGNzaS1yYmQtbm9kZQoKICAgICAgICByZXRfdmFs
            LCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24p
            CiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwKICAgICAgICBp
            ZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiInYXV0aCBnZXQtb3It
            Y3JlYXRlIHt1c2VyfScgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAgICAgICAgICAgICBmIkVycm9y
            OiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIK
            ICAgICAgICAgICAgKQogICAgICAgIHJldHVybiBzdHIoanNvbl9vdXRbMF1bImtleSJdKSwgZiJ7
            ZW50aXR5LnNwbGl0KCcuJywgMSlbMV19IgogICAgICAgICMgZW50aXR5LnNwbGl0KCcuJywxKVsx
            XSB0byByZW5hbWUgZW50aXR5KGNsaWVudC5jc2ktcmJkLW5vZGUpIGFzIGNzaS1yYmQtbm9kZQoK
            ICAgIGRlZiBnZXRfY2VwaGZzX2RhdGFfcG9vbF9kZXRhaWxzKHNlbGYpOgogICAgICAgIGNtZF9q
            c29uID0geyJwcmVmaXgiOiAiZnMgbHMiLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIGlmIHNl
            bGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigi
            Y2VwaCAiICsgY21kX2pzb25bInByZWZpeCJdKQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBl
            cnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICAjIGlm
            IHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LCByZXBvcnQgYW4gZXJyb3IKICAgICAg
            ICBpZiByZXRfdmFsICE9IDA6CiAgICAgICAgICAgICMgaWYgZnMgYW5kIGRhdGFfcG9vbCBhcmd1
            bWVudHMgYXJlIG5vdCBzZXQsIHNpbGVudGx5IHJldHVybgogICAgICAgICAgICBpZiAoCiAgICAg
            ICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUgPT0gIiIK
            ICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFt
            ZSA9PSAiIgogICAgICAgICAgICApOgogICAgICAgICAgICAgICAgcmV0dXJuCiAgICAgICAgICAg
            ICMgaWYgdXNlciBoYXMgcHJvdmlkZWQgYW55IG9mIHRoZQogICAgICAgICAgICAjICctLWNlcGhm
            cy1maWxlc3lzdGVtLW5hbWUnIG9yICctLWNlcGhmcy1kYXRhLXBvb2wtbmFtZScgYXJndW1lbnRz
            LAogICAgICAgICAgICAjIHJhaXNlIGFuIGV4Y2VwdGlvbiBhcyB3ZSBhcmUgdW5hYmxlIHRvIHZl
            cmlmeSB0aGUgYXJncwogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KAogICAgICAgICAgICAgICAgZiInZnMgbHMnIGNlcGggY2FsbCBmYWlsZWQgd2l0aCBlcnJvcjog
            e2Vycl9tc2d9IgogICAgICAgICAgICApCgogICAgICAgIG1hdGNoaW5nX2pzb25fb3V0ID0ge30K
            ICAgICAgICAjIGlmICctLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUnIGFyZ3VtZW50IGlzIHByb3Zp
            ZGVkLAogICAgICAgICMgY2hlY2sgd2hldGhlciB0aGUgcHJvdmlkZWQgZmlsZXN5c3RlbS1uYW1l
            IGV4aXN0cyBvciBub3QKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lz
            dGVtX25hbWU6CiAgICAgICAgICAgICMgZ2V0IHRoZSBtYXRjaGluZyBsaXN0CiAgICAgICAgICAg
            IG1hdGNoaW5nX2pzb25fb3V0X2xpc3QgPSBbCiAgICAgICAgICAgICAgICBtYXRjaGVkCiAgICAg
            ICAgICAgICAgICBmb3IgbWF0Y2hlZCBpbiBqc29uX291dAogICAgICAgICAgICAgICAgaWYgc3Ry
            KG1hdGNoZWRbIm5hbWUiXSkgPT0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9u
            YW1lCiAgICAgICAgICAgIF0KICAgICAgICAgICAgIyB1bmFibGUgdG8gZmluZCBhIG1hdGNoaW5n
            IGZzLW5hbWUsIHJhaXNlIGFuIGVycm9yCiAgICAgICAgICAgIGlmIGxlbihtYXRjaGluZ19qc29u
            X291dF9saXN0KSA9PSAwOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBmIkZpbGVzeXN0ZW0gcHJvdmlkZWQsICd7c2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lfScsICIKICAgICAgICAgICAgICAg
            ICAgICBmImlzIG5vdCBmb3VuZCBpbiB0aGUgZnMtbGlzdDoge1tzdHIoeFsnbmFtZSddKSBmb3Ig
            eCBpbiBqc29uX291dF19IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBtYXRjaGluZ19q
            c29uX291dCA9IG1hdGNoaW5nX2pzb25fb3V0X2xpc3RbMF0KICAgICAgICAjIGlmIGNlcGhmcyBm
            aWxlc3lzdGVtIG5hbWUgaXMgbm90IHByb3ZpZGVkLAogICAgICAgICMgdHJ5IHRvIGdldCBhIGRl
            ZmF1bHQgZnMgbmFtZSBieSBkb2luZyB0aGUgZm9sbG93aW5nCiAgICAgICAgZWxzZToKICAgICAg
            ICAgICAgIyBhLiBjaGVjayBpZiB0aGVyZSBpcyBvbmx5IG9uZSBmaWxlc3lzdGVtIGlzIHByZXNl
            bnQKICAgICAgICAgICAgaWYgbGVuKGpzb25fb3V0KSA9PSAxOgogICAgICAgICAgICAgICAgbWF0
            Y2hpbmdfanNvbl9vdXQgPSBqc29uX291dFswXQogICAgICAgICAgICAjIGIuIG9yIGVsc2UsIGNo
            ZWNrIGlmIGRhdGFfcG9vbCBuYW1lIGlzIHByb3ZpZGVkCiAgICAgICAgICAgIGVsaWYgc2VsZi5f
            YXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWU6CiAgICAgICAgICAgICAgICAjIGFuZCBp
            ZiBwcmVzZW50LCBjaGVjayB3aGV0aGVyIHRoZXJlIGV4aXN0cyBhIGZzIHdoaWNoIGhhcyB0aGUg
            ZGF0YV9wb29sCiAgICAgICAgICAgICAgICBmb3IgZWFjaEogaW4ganNvbl9vdXQ6CiAgICAgICAg
            ICAgICAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUgaW4g
            ZWFjaEpbImRhdGFfcG9vbHMiXToKICAgICAgICAgICAgICAgICAgICAgICAgbWF0Y2hpbmdfanNv
            bl9vdXQgPSBlYWNoSgogICAgICAgICAgICAgICAgICAgICAgICBicmVhawogICAgICAgICAgICAg
            ICAgIyBpZiB0aGVyZSBpcyBubyBtYXRjaGluZyBmcyBleGlzdHMsIHRoYXQgbWVhbnMgcHJvdmlk
            ZWQgZGF0YV9wb29sIG5hbWUgaXMgaW52YWxpZAogICAgICAgICAgICAgICAgaWYgbm90IG1hdGNo
            aW5nX2pzb25fb3V0OgogICAgICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVF
            eGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgICAgIGYiUHJvdmlkZWQgZGF0YV9wb29sIG5h
            bWUsIHtzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZX0sIgogICAgICAgICAg
            ICAgICAgICAgICAgICAiIGRvZXMgbm90IGV4aXN0cyIKICAgICAgICAgICAgICAgICAgICApCiAg
            ICAgICAgICAgICMgYy4gaWYgbm90aGluZyBpcyBzZXQgYW5kIGNvdWxkbid0IGZpbmQgYSBkZWZh
            dWx0LAogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgIyBqdXN0IHJldHVybiBzaWxl
            bnRseQogICAgICAgICAgICAgICAgcmV0dXJuCgogICAgICAgIGlmIG1hdGNoaW5nX2pzb25fb3V0
            OgogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUgPSBz
            dHIobWF0Y2hpbmdfanNvbl9vdXRbIm5hbWUiXSkKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5jZXBoZnNfbWV0YWRhdGFfcG9vbF9uYW1lID0gc3RyKAogICAgICAgICAgICAgICAgbWF0Y2hp
            bmdfanNvbl9vdXRbIm1ldGFkYXRhX3Bvb2wiXQogICAgICAgICAgICApCgogICAgICAgIGlmIGlz
            aW5zdGFuY2UobWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXSwgbGlzdCk6CiAgICAgICAg
            ICAgICMgaWYgdGhlIHVzZXIgaGFzIGFscmVhZHkgcHJvdmlkZWQgZGF0YS1wb29sLW5hbWUsCiAg
            ICAgICAgICAgICMgdGhyb3VnaCAtLWNlcGhmcy1kYXRhLXBvb2wtbmFtZQogICAgICAgICAgICBp
            ZiBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZToKICAgICAgICAgICAgICAg
            ICMgaWYgdGhlIHByb3ZpZGVkIG5hbWUgaXMgbm90IG1hdGNoaW5nIHdpdGggdGhlIG9uZSBpbiB0
            aGUgbGlzdAogICAgICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2Fy
            Z19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgICAgICAgICAgbm90IGlu
            IG1hdGNoaW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0KICAgICAgICAgICAgICAgICk6CiAgICAg
            ICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAg
            ICAgICAgICAgICAgICAgZiJQcm92aWRlZCBkYXRhLXBvb2wtbmFtZTogJ3tzZWxmLl9hcmdfcGFy
            c2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZX0nLCAiCiAgICAgICAgICAgICAgICAgICAgICAgICJk
            b2Vzbid0IG1hdGNoIGZyb20gdGhlIGRhdGEtcG9vbHMgbGlzdDogIgogICAgICAgICAgICAgICAg
            ICAgICAgICBmIntbc3RyKHgpIGZvciB4IGluIG1hdGNoaW5nX2pzb25fb3V0WydkYXRhX3Bvb2xz
            J11dfSIKICAgICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICMgaWYgZGF0YV9wb29sIG5h
            bWUgaXMgbm90IHByb3ZpZGVkLAogICAgICAgICAgICAjIHRoZW4gdHJ5IHRvIGZpbmQgYSBkZWZh
            dWx0IGRhdGEgcG9vbCBuYW1lCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICAjIGlm
            IG5vIGRhdGFfcG9vbHMgZXhpc3QsIHNpbGVudGx5IHJldHVybgogICAgICAgICAgICAgICAgaWYg
            bGVuKG1hdGNoaW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0pID09IDA6CiAgICAgICAgICAgICAg
            ICAgICAgcmV0dXJuCiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRh
            X3Bvb2xfbmFtZSA9IHN0cigKICAgICAgICAgICAgICAgICAgICBtYXRjaGluZ19qc29uX291dFsi
            ZGF0YV9wb29scyJdWzBdCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICMgaWYgdGhlcmUg
            YXJlIG1vcmUgdGhhbiBvbmUgJ2RhdGFfcG9vbHMnIGV4aXN0LAogICAgICAgICAgICAjIHRoZW4g
            d2FybiB0aGUgdXNlciB0aGF0IHdlIGFyZSB1c2luZyB0aGUgc2VsZWN0ZWQgbmFtZQogICAgICAg
            ICAgICBpZiBsZW4obWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXSkgPiAxOgogICAgICAg
            ICAgICAgICAgcHJpbnQoCiAgICAgICAgICAgICAgICAgICAgIldBUk5JTkc6IE11bHRpcGxlIGRh
            dGEgcG9vbHMgZGV0ZWN0ZWQ6ICIKICAgICAgICAgICAgICAgICAgICBmIntbc3RyKHgpIGZvciB4
            IGluIG1hdGNoaW5nX2pzb25fb3V0WydkYXRhX3Bvb2xzJ11dfVxuIgogICAgICAgICAgICAgICAg
            ICAgIGYiVXNpbmcgdGhlIGRhdGEtcG9vbDogJ3tzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRh
            X3Bvb2xfbmFtZX0nXG4iCiAgICAgICAgICAgICAgICApCgogICAgZGVmIGNyZWF0ZV9jaGVja2Vy
            S2V5KHNlbGYsIHVzZXIpOgogICAgICAgIGNhcHMsIGVudGl0eSA9IHNlbGYuZ2V0X2NhcHNfYW5k
            X2VudGl0eSh1c2VyKQogICAgICAgIGNtZF9qc29uID0gewogICAgICAgICAgICAicHJlZml4Ijog
            ImF1dGggZ2V0LW9yLWNyZWF0ZSIsCiAgICAgICAgICAgICJlbnRpdHkiOiBlbnRpdHksCiAgICAg
            ICAgICAgICJjYXBzIjogW2NhcCBmb3IgY2FwX2xpc3QgaW4gbGlzdChjYXBzLml0ZW1zKCkpIGZv
            ciBjYXAgaW4gY2FwX2xpc3RdLAogICAgICAgICAgICAiZm9ybWF0IjogImpzb24iLAogICAgICAg
            IH0KCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1
            cm4gc2VsZi5kcnlfcnVuKAogICAgICAgICAgICAgICAgImNlcGggIgogICAgICAgICAgICAgICAg
            KyBjbWRfanNvblsicHJlZml4Il0KICAgICAgICAgICAgICAgICsgIiAiCiAgICAgICAgICAgICAg
            ICArIGNtZF9qc29uWyJlbnRpdHkiXQogICAgICAgICAgICAgICAgKyAiICIKICAgICAgICAgICAg
            ICAgICsgIiAiLmpvaW4oY21kX2pzb25bImNhcHMiXSkKICAgICAgICAgICAgKQogICAgICAgICMg
            Y2hlY2sgaWYgdXNlciBhbHJlYWR5IGV4aXN0CiAgICAgICAgdXNlcl9rZXkgPSBzZWxmLmNoZWNr
            X3VzZXJfZXhpc3QoZW50aXR5KQogICAgICAgIGlmIHVzZXJfa2V5ICE9ICIiOgogICAgICAgICAg
            ICByZXR1cm4gdXNlcl9rZXkKCiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBz
            ZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICMgaWYgdGhlcmUgaXMg
            YW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihq
            c29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgICAgIGYiJ2F1dGggZ2V0LW9yLWNyZWF0ZSB7c2VsZi5ydW5fYXNfdXNl
            cn0nIGNvbW1hbmQgZmFpbGVkXG4iCiAgICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBp
            ZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAg
            KQogICAgICAgIHJldHVybiBzdHIoanNvbl9vdXRbMF1bImtleSJdKQoKICAgIGRlZiBnZXRfY2Vw
            aF9kYXNoYm9hcmRfbGluayhzZWxmKToKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4IjogIm1n
            ciBzZXJ2aWNlcyIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNl
            ci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKCJjZXBoICIgKyBjbWRf
            anNvblsicHJlZml4Il0pCiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIF8gPSBzZWxmLl9jb21t
            b25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNj
            ZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkg
            PT0gMDoKICAgICAgICAgICAgcmV0dXJuIE5vbmUKICAgICAgICBpZiAiZGFzaGJvYXJkIiBub3Qg
            aW4ganNvbl9vdXQ6CiAgICAgICAgICAgIHJldHVybiBOb25lCiAgICAgICAgcmV0dXJuIGpzb25f
            b3V0WyJkYXNoYm9hcmQiXQoKICAgIGRlZiBjcmVhdGVfcmd3X2FkbWluX29wc191c2VyKHNlbGYp
            OgogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAgICAgICAg
            ICAidXNlciIsCiAgICAgICAgICAgICJjcmVhdGUiLAogICAgICAgICAgICAiLS11aWQiLAogICAg
            ICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9PUFNfVVNFUl9OQU1FLAogICAgICAgICAg
            ICAiLS1kaXNwbGF5LW5hbWUiLAogICAgICAgICAgICAiUm9vayBSR1cgQWRtaW4gT3BzIHVzZXIi
            LAogICAgICAgICAgICAiLS1jYXBzIiwKICAgICAgICAgICAgImJ1Y2tldHM9Kjt1c2Vycz0qO3Vz
            YWdlPXJlYWQ7bWV0YWRhdGE9cmVhZDt6b25lPXJlYWQiLAogICAgICAgICAgICAiLS1yZ3ctcmVh
            bG0iLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lLAogICAgICAg
            ICAgICAiLS1yZ3ctem9uZWdyb3VwIiwKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3df
            em9uZWdyb3VwX25hbWUsCiAgICAgICAgICAgICItLXJndy16b25lIiwKICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1lLAogICAgICAgIF0KICAgICAgICBpZiBzZWxmLl9h
            cmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGgg
            IiArICIgIi5qb2luKGNtZCkpCiAgICAgICAgdHJ5OgogICAgICAgICAgICBvdXRwdXQgPSBzdWJw
            cm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAg
            ZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yIGFzIGV4ZWNFcnI6CiAgICAgICAg
            ICAgICMgaWYgdGhlIHVzZXIgYWxyZWFkeSBleGlzdHMsIHdlIGp1c3QgcXVlcnkgaXQKICAgICAg
            ICAgICAgaWYgZXhlY0Vyci5yZXR1cm5jb2RlID09IGVycm5vLkVFWElTVDoKICAgICAgICAgICAg
            ICAgIGNtZCA9IFsKICAgICAgICAgICAgICAgICAgICAicmFkb3Nndy1hZG1pbiIsCiAgICAgICAg
            ICAgICAgICAgICAgInVzZXIiLAogICAgICAgICAgICAgICAgICAgICJpbmZvIiwKICAgICAgICAg
            ICAgICAgICAgICAiLS11aWQiLAogICAgICAgICAgICAgICAgICAgIHNlbGYuRVhURVJOQUxfUkdX
            X0FETUlOX09QU19VU0VSX05BTUUsCiAgICAgICAgICAgICAgICAgICAgIi0tcmd3LXJlYWxtIiwK
            ICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lLAogICAg
            ICAgICAgICAgICAgICAgICItLXJndy16b25lZ3JvdXAiLAogICAgICAgICAgICAgICAgICAgIHNl
            bGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lLAogICAgICAgICAgICAgICAgICAgICIt
            LXJndy16b25lIiwKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25l
            X25hbWUsCiAgICAgICAgICAgICAgICBdCiAgICAgICAgICAgICAgICB0cnk6CiAgICAgICAgICAg
            ICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3Vi
            cHJvY2Vzcy5QSVBFKQogICAgICAgICAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJv
            Y2Vzc0Vycm9yIGFzIGV4ZWNFcnI6CiAgICAgICAgICAgICAgICAgICAgZXJyX21zZyA9ICgKICAg
            ICAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5kIHtjbWR9LiBP
            dXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICAgICAgZiJDb2Rl
            OiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtleGVjRXJyLnN0ZGVycn0iCiAgICAgICAg
            ICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21z
            ZykKICAgICAgICAgICAgICAgICAgICByZXR1cm4gTm9uZSwgTm9uZSwgRmFsc2UsICItMSIKICAg
            ICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVycl9tc2cgPSAoCiAgICAgICAgICAgICAg
            ICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5kIHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJy
            Lm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICBmIkNvZGU6IHtleGVjRXJyLnJldHVybmNv
            ZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykKICAgICAgICAgICAgICAgIHJldHVybiBO
            b25lLCBOb25lLCBGYWxzZSwgIi0xIgoKICAgICAgICAjIGlmIGl0IGlzIHB5dGhvbjIsIGRvbid0
            IGNoZWNrIGZvciBjZXBoIHZlcnNpb24gZm9yIGFkZGluZyBgaW5mbz1yZWFkYCBjYXAocmd3X3Zh
            bGlkYXRpb24pCiAgICAgICAgaWYgc3lzLnZlcnNpb25faW5mby5tYWpvciA8IDM6CiAgICAgICAg
            ICAgIGpzb25vdXRwdXQgPSBqc29uLmxvYWRzKG91dHB1dCkKICAgICAgICAgICAgcmV0dXJuICgK
            ICAgICAgICAgICAgICAgIGpzb25vdXRwdXRbImtleXMiXVswXVsiYWNjZXNzX2tleSJdLAogICAg
            ICAgICAgICAgICAganNvbm91dHB1dFsia2V5cyJdWzBdWyJzZWNyZXRfa2V5Il0sCiAgICAgICAg
            ICAgICAgICBGYWxzZSwKICAgICAgICAgICAgICAgICIiLAogICAgICAgICAgICApCgogICAgICAg
            ICMgc2VwYXJhdGVseSBhZGQgaW5mbz1yZWFkIGNhcHMgZm9yIHJndy1lbmRwb2ludCBpcCB2YWxp
            ZGF0aW9uCiAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkID0gVHJ1ZQogICAgICAgIGNtZCA9IFsK
            ICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAgICAgICAgICAiY2FwcyIsCiAgICAgICAg
            ICAgICJhZGQiLAogICAgICAgICAgICAiLS11aWQiLAogICAgICAgICAgICBzZWxmLkVYVEVSTkFM
            X1JHV19BRE1JTl9PUFNfVVNFUl9OQU1FLAogICAgICAgICAgICAiLS1jYXBzIiwKICAgICAgICAg
            ICAgImluZm89cmVhZCIsCiAgICAgICAgICAgICItLXJndy1yZWFsbSIsCiAgICAgICAgICAgIHNl
            bGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUsCiAgICAgICAgICAgICItLXJndy16b25lZ3Jv
            dXAiLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSwKICAg
            ICAgICAgICAgIi0tcmd3LXpvbmUiLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196
            b25lX25hbWUsCiAgICAgICAgXQogICAgICAgIHRyeToKICAgICAgICAgICAgb3V0cHV0ID0gc3Vi
            cHJvY2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAg
            IGV4Y2VwdCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJvciBhcyBleGVjRXJyOgogICAgICAg
            ICAgICAjIGlmIHRoZSBjZXBoIHZlcnNpb24gbm90IHN1cHBvcnRlZCBmb3IgYWRkaW5nIGBpbmZv
            PXJlYWRgIGNhcChyZ3dfdmFsaWRhdGlvbikKICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAg
            ICAgImNvdWxkIG5vdCBhZGQgY2FwczogdW5hYmxlIHRvIGFkZCBjYXBzOiBpbmZvPXJlYWRcbiIK
            ICAgICAgICAgICAgICAgIGluIGV4ZWNFcnIuc3RkZXJyLmRlY29kZSgidXRmLTgiKQogICAgICAg
            ICAgICAgICAgYW5kIGV4ZWNFcnIucmV0dXJuY29kZSA9PSAyNDQKICAgICAgICAgICAgKToKICAg
            ICAgICAgICAgICAgIGluZm9fY2FwX3N1cHBvcnRlZCA9IEZhbHNlCiAgICAgICAgICAgIGVsc2U6
            CiAgICAgICAgICAgICAgICBlcnJfbXNnID0gKAogICAgICAgICAgICAgICAgICAgIGYiZmFpbGVk
            IHRvIGV4ZWN1dGUgY29tbWFuZCB7Y21kfS4gT3V0cHV0OiB7ZXhlY0Vyci5vdXRwdXR9LiAiCiAg
            ICAgICAgICAgICAgICAgICAgZiJDb2RlOiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtl
            eGVjRXJyLnN0ZGVycn0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICBzeXMuc3Rk
            ZXJyLndyaXRlKGVycl9tc2cpCiAgICAgICAgICAgICAgICByZXR1cm4gTm9uZSwgTm9uZSwgRmFs
            c2UsICItMSIKCiAgICAgICAganNvbm91dHB1dCA9IGpzb24ubG9hZHMob3V0cHV0KQogICAgICAg
            IHJldHVybiAoCiAgICAgICAgICAgIGpzb25vdXRwdXRbImtleXMiXVswXVsiYWNjZXNzX2tleSJd
            LAogICAgICAgICAgICBqc29ub3V0cHV0WyJrZXlzIl1bMF1bInNlY3JldF9rZXkiXSwKICAgICAg
            ICAgICAgaW5mb19jYXBfc3VwcG9ydGVkLAogICAgICAgICAgICAiIiwKICAgICAgICApCgogICAg
            ZGVmIHZhbGlkYXRlX3JiZF9wb29sKHNlbGYsIHBvb2xfbmFtZSk6CiAgICAgICAgaWYgbm90IHNl
            bGYuY2x1c3Rlci5wb29sX2V4aXN0cyhwb29sX25hbWUpOgogICAgICAgICAgICByYWlzZSBFeGVj
            dXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcG9v
            bCwgJ3twb29sX25hbWV9JywgZG9lcyBub3QgZXhpc3QiCiAgICAgICAgICAgICkKCiAgICBkZWYg
            aW5pdF9yYmRfcG9vbChzZWxmLCByYmRfcG9vbF9uYW1lKToKICAgICAgICBpZiBpc2luc3RhbmNl
            KHNlbGYuY2x1c3RlciwgRHVtbXlSYWRvcyk6CiAgICAgICAgICAgIHJldHVybgogICAgICAgIGlv
            Y3R4ID0gc2VsZi5jbHVzdGVyLm9wZW5faW9jdHgocmJkX3Bvb2xfbmFtZSkKICAgICAgICByYmRf
            aW5zdCA9IHJiZC5SQkQoKQogICAgICAgIHJiZF9pbnN0LnBvb2xfaW5pdChpb2N0eCwgVHJ1ZSkK
            CiAgICBkZWYgdmFsaWRhdGVfcmFkb3NfbmFtZXNwYWNlKHNlbGYpOgogICAgICAgIHJiZF9wb29s
            X25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgIHJhZG9z
            X25hbWVzcGFjZSA9IHNlbGYuX2FyZ19wYXJzZXIucmFkb3NfbmFtZXNwYWNlCiAgICAgICAgaWYg
            cmFkb3NfbmFtZXNwYWNlID09ICIiOgogICAgICAgICAgICByZXR1cm4KICAgICAgICBpZiByYWRv
            c19uYW1lc3BhY2UuaXNsb3dlcigpID09IEZhbHNlOgogICAgICAgICAgICByYWlzZSBFeGVjdXRp
            b25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmFkb3Mg
            TmFtZXNwYWNlLCAne3JhZG9zX25hbWVzcGFjZX0nLCAiCiAgICAgICAgICAgICAgICBmImNvbnRh
            aW5zIHVwcGVyIGNhc2UiCiAgICAgICAgICAgICkKICAgICAgICByYmRfaW5zdCA9IHJiZC5SQkQo
            KQogICAgICAgIGlvY3R4ID0gc2VsZi5jbHVzdGVyLm9wZW5faW9jdHgocmJkX3Bvb2xfbmFtZSkK
            ICAgICAgICBpZiByYmRfaW5zdC5uYW1lc3BhY2VfZXhpc3RzKGlvY3R4LCByYWRvc19uYW1lc3Bh
            Y2UpIGlzIEZhbHNlOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmFkb3MgTmFtZXNwYWNlLCAne3JhZG9z
            X25hbWVzcGFjZX0nLCAiCiAgICAgICAgICAgICAgICBmImlzIG5vdCBmb3VuZCBpbiB0aGUgcG9v
            bCAne3JiZF9wb29sX25hbWV9JyIKICAgICAgICAgICAgKQoKICAgIGRlZiBnZXRfb3JfY3JlYXRl
            X3N1YnZvbHVtZV9ncm91cChzZWxmLCBzdWJ2b2x1bWVfZ3JvdXAsIGNlcGhmc19maWxlc3lzdGVt
            X25hbWUpOgogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgImNlcGgiLAogICAgICAgICAgICAi
            ZnMiLAogICAgICAgICAgICAic3Vidm9sdW1lZ3JvdXAiLAogICAgICAgICAgICAiZ2V0cGF0aCIs
            CiAgICAgICAgICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgIHN1YnZvbHVt
            ZV9ncm91cCwKICAgICAgICBdCiAgICAgICAgdHJ5OgogICAgICAgICAgICBfID0gc3VicHJvY2Vz
            cy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2Vw
            dCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJvcjoKICAgICAgICAgICAgY21kID0gWwogICAg
            ICAgICAgICAgICAgImNlcGgiLAogICAgICAgICAgICAgICAgImZzIiwKICAgICAgICAgICAgICAg
            ICJzdWJ2b2x1bWVncm91cCIsCiAgICAgICAgICAgICAgICAiY3JlYXRlIiwKICAgICAgICAgICAg
            ICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgICAgICBzdWJ2b2x1bWVfZ3Jv
            dXAsCiAgICAgICAgICAgIF0KICAgICAgICAgICAgdHJ5OgogICAgICAgICAgICAgICAgXyA9IHN1
            YnByb2Nlc3MuY2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAgICAg
            ICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBm
            InN1YnZvbHVtZSBncm91cCB7c3Vidm9sdW1lX2dyb3VwfSBpcyBub3QgYWJsZSB0byBnZXQgY3Jl
            YXRlZCIKICAgICAgICAgICAgICAgICkKCiAgICBkZWYgcGluX3N1YnZvbHVtZSgKICAgICAgICBz
            ZWxmLCBzdWJ2b2x1bWVfZ3JvdXAsIGNlcGhmc19maWxlc3lzdGVtX25hbWUsIHBpbl90eXBlLCBw
            aW5fc2V0dGluZwogICAgKToKICAgICAgICBjbWQgPSBbCiAgICAgICAgICAgICJjZXBoIiwKICAg
            ICAgICAgICAgImZzIiwKICAgICAgICAgICAgInN1YnZvbHVtZWdyb3VwIiwKICAgICAgICAgICAg
            InBpbiIsCiAgICAgICAgICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgIHN1
            YnZvbHVtZV9ncm91cCwKICAgICAgICAgICAgcGluX3R5cGUsCiAgICAgICAgICAgIHBpbl9zZXR0
            aW5nLAogICAgICAgIF0KICAgICAgICB0cnk6CiAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNo
            ZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1
            YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJzdWJ2b2x1bWUgZ3JvdXAge3N1YnZv
            bHVtZV9ncm91cH0gaXMgbm90IGFibGUgdG8gZ2V0IHBpbm5lZCIKICAgICAgICAgICAgKQoKICAg
            IGRlZiBnZXRfcmd3X2ZzaWQoc2VsZiwgYmFzZV91cmwsIHZlcmlmeSk6CiAgICAgICAgYWNjZXNz
            X2tleSA9IHNlbGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX0FDQ0VTU19LRVkiXQogICAg
            ICAgIHNlY3JldF9rZXkgPSBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9TRUNSRVRf
            S0VZIl0KICAgICAgICByZ3dfZW5kcG9pbnQgPSBzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2lu
            dAogICAgICAgIGJhc2VfdXJsID0gYmFzZV91cmwgKyAiOi8vIiArIHJnd19lbmRwb2ludCArICIv
            YWRtaW4vaW5mbz8iCiAgICAgICAgcGFyYW1zID0geyJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAg
            cmVxdWVzdF91cmwgPSBiYXNlX3VybCArIHVybGVuY29kZShwYXJhbXMpCgogICAgICAgIHRyeToK
            ICAgICAgICAgICAgciA9IHJlcXVlc3RzLmdldCgKICAgICAgICAgICAgICAgIHJlcXVlc3RfdXJs
            LAogICAgICAgICAgICAgICAgYXV0aD1TM0F1dGgoYWNjZXNzX2tleSwgc2VjcmV0X2tleSwgcmd3
            X2VuZHBvaW50KSwKICAgICAgICAgICAgICAgIHZlcmlmeT12ZXJpZnksCiAgICAgICAgICAgICkK
            ICAgICAgICBleGNlcHQgcmVxdWVzdHMuZXhjZXB0aW9ucy5UaW1lb3V0OgogICAgICAgICAgICBz
            eXMuc3RkZXJyLndyaXRlKAogICAgICAgICAgICAgICAgZiJpbnZhbGlkIGVuZHBvaW50Oiwgbm90
            IGFibGUgdG8gY2FsbCBhZG1pbi1vcHMgYXBpe3Jnd19lbmRwb2ludH0iCiAgICAgICAgICAgICkK
            ICAgICAgICAgICAgcmV0dXJuICIiLCAiLTEiCiAgICAgICAgcjEgPSByLmpzb24oKQogICAgICAg
            IGlmIHIxIGlzIE5vbmUgb3IgcjEuZ2V0KCJpbmZvIikgaXMgTm9uZToKICAgICAgICAgICAgc3lz
            LnN0ZGVyci53cml0ZSgKICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHJndyBFbmRwb2lu
            dCwgJ3tzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludH0nLCBpcyBpbnZhbGlkLiIKICAgICAg
            ICAgICAgKQogICAgICAgICAgICByZXR1cm4gKAogICAgICAgICAgICAgICAgIiIsCiAgICAgICAg
            ICAgICAgICAiLTEiLAogICAgICAgICAgICApCgogICAgICAgIHJldHVybiByMVsiaW5mbyJdWyJz
            dG9yYWdlX2JhY2tlbmRzIl1bMF1bImNsdXN0ZXJfaWQiXSwgIiIKCiAgICBkZWYgdmFsaWRhdGVf
            cmd3X2VuZHBvaW50KHNlbGYsIGluZm9fY2FwX3N1cHBvcnRlZCk6CiAgICAgICAgIyBpZiB0aGUg
            J2NsdXN0ZXInIGluc3RhbmNlIGlzIGEgZHVtbXkgb25lLAogICAgICAgICMgZG9uJ3QgdHJ5IHRv
            IHJlYWNoIG91dCB0byB0aGUgZW5kcG9pbnQKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1
            c3RlciwgRHVtbXlSYWRvcyk6CiAgICAgICAgICAgIHJldHVybgoKICAgICAgICByZ3dfZW5kcG9p
            bnQgPSBzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludAoKICAgICAgICAjIHZhbGlkYXRlIHJn
            dyBlbmRwb2ludCBvbmx5IGlmIGlwIGFkZHJlc3MgaXMgcGFzc2VkCiAgICAgICAgaXBfdHlwZSA9
            IHNlbGYuX2ludmFsaWRfZW5kcG9pbnQocmd3X2VuZHBvaW50KQoKICAgICAgICAjIGNoZWNrIGlm
            IHRoZSByZ3cgZW5kcG9pbnQgaXMgcmVhY2hhYmxlCiAgICAgICAgY2VydCA9IE5vbmUKICAgICAg
            ICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNlci5yZ3dfc2tpcF90bHMgYW5kIHNlbGYudmFsaWRhdGVf
            cmd3X2VuZHBvaW50X3Rsc19jZXJ0KCk6CiAgICAgICAgICAgIGNlcnQgPSBzZWxmLl9hcmdfcGFy
            c2VyLnJnd190bHNfY2VydF9wYXRoCiAgICAgICAgYmFzZV91cmwsIHZlcmlmeSwgZXJyID0gc2Vs
            Zi5lbmRwb2ludF9kaWFsKHJnd19lbmRwb2ludCwgaXBfdHlwZSwgY2VydD1jZXJ0KQogICAgICAg
            IGlmIGVyciAhPSAiIjoKICAgICAgICAgICAgcmV0dXJuICItMSIKCiAgICAgICAgIyBjaGVjayBp
            ZiB0aGUgcmd3IGVuZHBvaW50IGJlbG9uZ3MgdG8gdGhlIHNhbWUgY2x1c3RlcgogICAgICAgICMg
            b25seSBjaGVjayBpZiBgaW5mb2AgY2FwIGlzIHN1cHBvcnRlZAogICAgICAgIGlmIGluZm9fY2Fw
            X3N1cHBvcnRlZDoKICAgICAgICAgICAgZnNpZCA9IHNlbGYuZ2V0X2ZzaWQoKQogICAgICAgICAg
            ICByZ3dfZnNpZCwgZXJyID0gc2VsZi5nZXRfcmd3X2ZzaWQoYmFzZV91cmwsIHZlcmlmeSkKICAg
            ICAgICAgICAgaWYgZXJyID09ICItMSI6CiAgICAgICAgICAgICAgICByZXR1cm4gIi0xIgogICAg
            ICAgICAgICBpZiBmc2lkICE9IHJnd19mc2lkOgogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53
            cml0ZSgKICAgICAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCByZ3cgRW5kcG9pbnQsICd7
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfZW5kcG9pbnR9JywgaXMgaW52YWxpZC4gV2UgYXJlIHZhbGlk
            YXRpbmcgYnkgY2FsbGluZyB0aGUgYWRtaW5vcHMgYXBpIHRocm91Z2ggcmd3LWVuZHBvaW50IGFu
            ZCB2YWxpZGF0aW5nIHRoZSBjbHVzdGVyX2lkICd7cmd3X2ZzaWR9JyBpcyBlcXVhbCB0byB0aGUg
            Y2VwaCBjbHVzdGVyIGZzaWQgJ3tmc2lkfSciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAg
            ICAgICByZXR1cm4gIi0xIgoKICAgICAgICAjIGNoZWNrIGlmIHRoZSByZ3cgZW5kcG9pbnQgcG9v
            bCBleGlzdAogICAgICAgICMgb25seSB2YWxpZGF0ZSBpZiByZ3dfcG9vbF9wcmVmaXggaXMgcGFz
            c2VkIGVsc2UgaXQgd2lsbCB0YWtlIGRlZmF1bHQgdmFsdWUgYW5kIHdlIGRvbid0IGNyZWF0ZSB0
            aGVzZSBkZWZhdWx0IHBvb2xzCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9w
            cmVmaXggIT0gImRlZmF1bHQiOgogICAgICAgICAgICByZ3dfcG9vbHNfdG9fdmFsaWRhdGUgPSBb
            CiAgICAgICAgICAgICAgICBmIntzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3
            Lm1ldGEiLAogICAgICAgICAgICAgICAgIi5yZ3cucm9vdCIsCiAgICAgICAgICAgICAgICBmIntz
            ZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmNvbnRyb2wiLAogICAgICAgICAg
            ICAgICAgZiJ7c2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5sb2ciLAogICAg
            ICAgICAgICBdCiAgICAgICAgICAgIGZvciBfcmd3X3Bvb2xfdG9fdmFsaWRhdGUgaW4gcmd3X3Bv
            b2xzX3RvX3ZhbGlkYXRlOgogICAgICAgICAgICAgICAgaWYgbm90IHNlbGYuY2x1c3Rlci5wb29s
            X2V4aXN0cyhfcmd3X3Bvb2xfdG9fdmFsaWRhdGUpOgogICAgICAgICAgICAgICAgICAgIHN5cy5z
            dGRlcnIud3JpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHBvb2ws
            ICd7X3Jnd19wb29sX3RvX3ZhbGlkYXRlfScsIGRvZXMgbm90IGV4aXN0IgogICAgICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgICAgICAgICByZXR1cm4gIi0xIgoKICAgICAgICByZXR1cm4g
            IiIKCiAgICBkZWYgdmFsaWRhdGVfcmd3X211bHRpc2l0ZShzZWxmLCByZ3dfbXVsdGlzaXRlX2Nv
            bmZpZ19uYW1lLCByZ3dfbXVsdGlzaXRlX2NvbmZpZyk6CiAgICAgICAgaWYgcmd3X211bHRpc2l0
            ZV9jb25maWcgIT0gIiI6CiAgICAgICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgICAgICJyYWRv
            c2d3LWFkbWluIiwKICAgICAgICAgICAgICAgIHJnd19tdWx0aXNpdGVfY29uZmlnLAogICAgICAg
            ICAgICAgICAgImdldCIsCiAgICAgICAgICAgICAgICAiLS1yZ3ctIiArIHJnd19tdWx0aXNpdGVf
            Y29uZmlnLAogICAgICAgICAgICAgICAgcmd3X211bHRpc2l0ZV9jb25maWdfbmFtZSwKICAgICAg
            ICAgICAgXQogICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBfID0gc3VicHJvY2Vzcy5j
            aGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAgICBleGNl
            cHQgc3VicHJvY2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3IgYXMgZXhlY0VycjoKICAgICAgICAgICAg
            ICAgIGVycl9tc2cgPSAoCiAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBj
            b21tYW5kIHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAgICAgICAgICAg
            ICAgICBmIkNvZGU6IHtleGVjRXJyLnJldHVybmNvZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJy
            fSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJy
            X21zZykKICAgICAgICAgICAgICAgIHJldHVybiAiLTEiCiAgICAgICAgcmV0dXJuICIiCgogICAg
            ZGVmIGNvbnZlcnRfY29tbWFfc2VwYXJhdGVkX3RvX2FycmF5KHNlbGYsIHZhbHVlKToKICAgICAg
            ICByZXR1cm4gdmFsdWUuc3BsaXQoIiwiKQoKICAgIGRlZiByYWlzZV9leGNlcHRpb25faWZfYW55
            X3RvcG9sb2d5X2ZsYWdfaXNfbWlzc2luZyhzZWxmKToKICAgICAgICBpZiAoCiAgICAgICAgICAg
            ICgKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMgIT0gIiIK
            ICAgICAgICAgICAgICAgIGFuZCAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCA9PSAiIgogICAgICAgICAgICAgICAgICAg
            IG9yIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfZmFpbHVyZV9kb21haW5fdmFsdWVzID09ICIi
            CiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICkKICAgICAgICAgICAgb3IgKAogICAgICAg
            ICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCAh
            PSAiIgogICAgICAgICAgICAgICAgYW5kICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdf
            cGFyc2VyLnRvcG9sb2d5X3Bvb2xzID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5f
            YXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMgPT0gIiIKICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgKQogICAgICAgICAgICBvciAoCiAgICAgICAgICAgICAgICBz
            ZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX3ZhbHVlcyAhPSAiIgogICAg
            ICAgICAgICAgICAgYW5kICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRv
            cG9sb2d5X3Bvb2xzID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5fYXJnX3BhcnNl
            ci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCA9PSAiIgogICAgICAgICAgICAgICAgKQog
            ICAgICAgICAgICApCiAgICAgICAgKToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVy
            ZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICJwcm92aWRlIGFsbCB0aGUgdG9wb2xvZ3kgZmxh
            Z3MgLS10b3BvbG9neS1wb29scywgLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi1sYWJlbCwgLS10
            b3BvbG9neS1mYWlsdXJlLWRvbWFpbi12YWx1ZXMiCiAgICAgICAgICAgICkKCiAgICBkZWYgdmFs
            aWRhdGVfdG9wb2xvZ3lfdmFsdWVzKHNlbGYsIHRvcG9sb2d5X3Bvb2xzLCB0b3BvbG9neV9mZCk6
            CiAgICAgICAgaWYgbGVuKHRvcG9sb2d5X3Bvb2xzKSAhPSBsZW4odG9wb2xvZ3lfZmQpOgogICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ZiJUaGUgcHJvdmlkZWQgdG9wb2xvZ3kgcG9vbHMsICd7dG9wb2xvZ3lfcG9vbHN9JywgYW5kICIK
            ICAgICAgICAgICAgICAgIGYidG9wb2xvZ3kgZmFpbHVyZSBkb21haW4sICd7dG9wb2xvZ3lfZmR9
            JywiCiAgICAgICAgICAgICAgICBmImFyZSBvZiBkaWZmZXJlbnQgbGVuZ3RoLCAne2xlbih0b3Bv
            bG9neV9wb29scyl9JyBhbmQgJ3tsZW4odG9wb2xvZ3lfZmQpfScgcmVzcGN0aXZlbHkiCiAgICAg
            ICAgICAgICkKICAgICAgICByZXR1cm4KCiAgICBkZWYgdmFsaWRhdGVfdG9wb2xvZ3lfcmJkX3Bv
            b2xzKHNlbGYsIHRvcG9sb2d5X3JiZF9wb29scyk6CiAgICAgICAgZm9yIHBvb2wgaW4gdG9wb2xv
            Z3lfcmJkX3Bvb2xzOgogICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3JiZF9wb29sKHBvb2wpCgog
            ICAgZGVmIGluaXRfdG9wb2xvZ3lfcmJkX3Bvb2xzKHNlbGYsIHRvcG9sb2d5X3JiZF9wb29scyk6
            CiAgICAgICAgZm9yIHBvb2wgaW4gdG9wb2xvZ3lfcmJkX3Bvb2xzOgogICAgICAgICAgICBzZWxm
            LmluaXRfcmJkX3Bvb2wocG9vbCkKCiAgICBkZWYgX2dlbl9vdXRwdXRfbWFwKHNlbGYpOgogICAg
            ICAgIGlmIHNlbGYub3V0X21hcDoKICAgICAgICAgICAgcmV0dXJuCiAgICAgICAgIyBzdXBwb3J0
            IGxlZ2FjeSBmbGFnIHdpdGggdXBncmFkZXMKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNs
            dXN0ZXJfbmFtZToKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1l
            ID0gc2VsZi5fYXJnX3BhcnNlci5jbHVzdGVyX25hbWUKICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            Lms4c19jbHVzdGVyX25hbWUgPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuazhzX2Ns
            dXN0ZXJfbmFtZS5sb3dlcigpCiAgICAgICAgKSAgIyBhbHdheXMgY29udmVydCBjbHVzdGVyIG5h
            bWUgdG8gbG93ZXJjYXNlIGNoYXJhY3RlcnMKICAgICAgICBzZWxmLnZhbGlkYXRlX3JiZF9wb29s
            KHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lKQogICAgICAgIHNlbGYuaW5pdF9y
            YmRfcG9vbChzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZSkKICAgICAgICBzZWxm
            LnZhbGlkYXRlX3JhZG9zX25hbWVzcGFjZSgpCiAgICAgICAgc2VsZi5fZXhjbHVkZWRfa2V5cy5h
            ZGQoIks4U19DTFVTVEVSX05BTUUiKQogICAgICAgIHNlbGYuZ2V0X2NlcGhmc19kYXRhX3Bvb2xf
            ZGV0YWlscygpCiAgICAgICAgc2VsZi5vdXRfbWFwWyJOQU1FU1BBQ0UiXSA9IHNlbGYuX2FyZ19w
            YXJzZXIubmFtZXNwYWNlCiAgICAgICAgc2VsZi5vdXRfbWFwWyJLOFNfQ0xVU1RFUl9OQU1FIl0g
            PSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBb
            IlJPT0tfRVhURVJOQUxfRlNJRCJdID0gc2VsZi5nZXRfZnNpZCgpCiAgICAgICAgc2VsZi5vdXRf
            bWFwWyJST09LX0VYVEVSTkFMX1VTRVJOQU1FIl0gPSBzZWxmLnJ1bl9hc191c2VyCiAgICAgICAg
            c2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0NFUEhfTU9OX0RBVEEiXSA9IHNlbGYuZ2V0X2Nl
            cGhfZXh0ZXJuYWxfbW9uX2RhdGEoKQogICAgICAgIHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5B
            TF9VU0VSX1NFQ1JFVCJdID0gc2VsZi5jcmVhdGVfY2hlY2tlcktleSgKICAgICAgICAgICAgImNs
            aWVudC5oZWFsdGhjaGVja2VyIgogICAgICAgICkKICAgICAgICBzZWxmLm91dF9tYXBbIlJPT0tf
            RVhURVJOQUxfREFTSEJPQVJEX0xJTksiXSA9IHNlbGYuZ2V0X2NlcGhfZGFzaGJvYXJkX2xpbmso
            KQogICAgICAgICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VDUkVU
            Il0sCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FIl0s
            CiAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3VzZXIoImNsaWVudC5jc2kt
            cmJkLW5vZGUiKQogICAgICAgICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BS
            T1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9SQkRfUFJP
            VklTSU9ORVJfU0VDUkVUX05BTUUiXSwKICAgICAgICApID0gc2VsZi5jcmVhdGVfY2VwaENTSUtl
            eXJpbmdfdXNlcigiY2xpZW50LmNzaS1yYmQtcHJvdmlzaW9uZXIiKQogICAgICAgIHNlbGYub3V0
            X21hcFsiQ0VQSEZTX1BPT0xfTkFNRSJdID0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9w
            b29sX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBbIkNFUEhGU19NRVRBREFUQV9QT09MX05BTUUi
            XSA9ICgKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfbWV0YWRhdGFfcG9vbF9u
            YW1lCiAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSA9IHNl
            bGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgIHNlbGYub3V0X21h
            cFsiUkVTVFJJQ1RFRF9BVVRIX1BFUk1JU1NJT04iXSA9ICgKICAgICAgICAgICAgc2VsZi5fYXJn
            X3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbgogICAgICAgICkKICAgICAgICBzZWxm
            Lm91dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdID0gc2VsZi5fYXJnX3BhcnNlci5yYWRvc19uYW1l
            c3BhY2UKICAgICAgICBzZWxmLm91dF9tYXBbIlNVQlZPTFVNRV9HUk9VUCJdID0gc2VsZi5fYXJn
            X3BhcnNlci5zdWJ2b2x1bWVfZ3JvdXAKICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNf
            Tk9ERV9TRUNSRVQiXSA9ICIiCiAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJ
            U0lPTkVSX1NFQ1JFVCJdID0gIiIKICAgICAgICAjIGNyZWF0ZSBDZXBoRlMgbm9kZSBhbmQgcHJv
            dmlzaW9uZXIga2V5cmluZyBvbmx5IHdoZW4gTURTIGV4aXN0cwogICAgICAgIGlmIHNlbGYub3V0
            X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSBhbmQgc2VsZi5vdXRfbWFwWyJDRVBIRlNfUE9PTF9OQU1F
            Il06CiAgICAgICAgICAgICgKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhG
            U19OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZT
            X05PREVfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lL
            ZXlyaW5nX3VzZXIoImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiKQogICAgICAgICAgICAoCiAgICAg
            ICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUIl0s
            CiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VD
            UkVUX05BTUUiXSwKICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3Vz
            ZXIoImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIikKICAgICAgICAgICAgIyBjcmVhdGUg
            dGhlIGRlZmF1bHQgImNzaSIgc3Vidm9sdW1lZ3JvdXAKICAgICAgICAgICAgc2VsZi5nZXRfb3Jf
            Y3JlYXRlX3N1YnZvbHVtZV9ncm91cCgKICAgICAgICAgICAgICAgICJjc2kiLCBzZWxmLl9hcmdf
            cGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICAgICAgKQogICAgICAgICAgICAj
            IHBpbiB0aGUgZGVmYXVsdCAiY3NpIiBzdWJ2b2x1bWVncm91cAogICAgICAgICAgICBzZWxmLnBp
            bl9zdWJ2b2x1bWUoCiAgICAgICAgICAgICAgICAiY3NpIiwgc2VsZi5fYXJnX3BhcnNlci5jZXBo
            ZnNfZmlsZXN5c3RlbV9uYW1lLCAiZGlzdHJpYnV0ZWQiLCAiMSIKICAgICAgICAgICAgKQogICAg
            ICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlNVQlZPTFVNRV9HUk9VUCJdOgogICAgICAgICAgICAg
            ICAgc2VsZi5nZXRfb3JfY3JlYXRlX3N1YnZvbHVtZV9ncm91cCgKICAgICAgICAgICAgICAgICAg
            ICBzZWxmLl9hcmdfcGFyc2VyLnN1YnZvbHVtZV9ncm91cCwKICAgICAgICAgICAgICAgICAgICBz
            ZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgICAgICBzZWxmLnBpbl9zdWJ2b2x1bWUoCiAgICAgICAgICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci5zdWJ2b2x1bWVfZ3JvdXAsCiAgICAgICAgICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lLAogICAgICAgICAgICAgICAgICAg
            ICJkaXN0cmlidXRlZCIsCiAgICAgICAgICAgICAgICAgICAgIjEiLAogICAgICAgICAgICAgICAg
            KQogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1RMU19DRVJUIl0gPSAiIgogICAgICAgIHNlbGYu
            b3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdID0gIiIKICAgICAgICBzZWxmLm91dF9tYXBb
            Ik1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJdID0gIiIKICAgICAgICBpZiBub3Qgc2VsZi5fYXJn
            X3BhcnNlci5za2lwX21vbml0b3JpbmdfZW5kcG9pbnQ6CiAgICAgICAgICAgICgKICAgICAgICAg
            ICAgICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdLAogICAgICAgICAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UX1BPUlQiXSwKICAgICAgICAgICAg
            KSA9IHNlbGYuZ2V0X2FjdGl2ZV9hbmRfc3RhbmRieV9tZ3JzKCkKICAgICAgICBzZWxmLm91dF9t
            YXBbIlJCRF9QT09MX05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1l
            CiAgICAgICAgc2VsZi5vdXRfbWFwWyJSQkRfTUVUQURBVEFfRUNfUE9PTF9OQU1FIl0gPSAoCiAg
            ICAgICAgICAgIHNlbGYudmFsaWRhdGVfcmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZSgpCiAgICAg
            ICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9PTFMiXSA9IHNlbGYuX2FyZ19w
            YXJzZXIudG9wb2xvZ3lfcG9vbHMKICAgICAgICBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxV
            UkVfRE9NQUlOX0xBQkVMIl0gPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xv
            Z3lfZmFpbHVyZV9kb21haW5fbGFiZWwKICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJU
            T1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiXSA9ICgKICAgICAgICAgICAgc2VsZi5fYXJn
            X3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMKICAgICAgICApCiAgICAgICAg
            aWYgKAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X3Bvb2xzICE9ICIiCiAg
            ICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX2xh
            YmVsICE9ICIiCiAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1
            cmVfZG9tYWluX3ZhbHVlcyAhPSAiIgogICAgICAgICk6CiAgICAgICAgICAgIHNlbGYudmFsaWRh
            dGVfdG9wb2xvZ3lfdmFsdWVzKAogICAgICAgICAgICAgICAgc2VsZi5jb252ZXJ0X2NvbW1hX3Nl
            cGFyYXRlZF90b19hcnJheShzZWxmLm91dF9tYXBbIlRPUE9MT0dZX1BPT0xTIl0pLAogICAgICAg
            ICAgICAgICAgc2VsZi5jb252ZXJ0X2NvbW1hX3NlcGFyYXRlZF90b19hcnJheSgKICAgICAgICAg
            ICAgICAgICAgICBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX1ZBTFVFUyJd
            CiAgICAgICAgICAgICAgICApLAogICAgICAgICAgICApCiAgICAgICAgICAgIHNlbGYudmFsaWRh
            dGVfdG9wb2xvZ3lfcmJkX3Bvb2xzKAogICAgICAgICAgICAgICAgc2VsZi5jb252ZXJ0X2NvbW1h
            X3NlcGFyYXRlZF90b19hcnJheShzZWxmLm91dF9tYXBbIlRPUE9MT0dZX1BPT0xTIl0pCiAgICAg
            ICAgICAgICkKICAgICAgICAgICAgc2VsZi5pbml0X3RvcG9sb2d5X3JiZF9wb29scygKICAgICAg
            ICAgICAgICAgIHNlbGYuY29udmVydF9jb21tYV9zZXBhcmF0ZWRfdG9fYXJyYXkoc2VsZi5vdXRf
            bWFwWyJUT1BPTE9HWV9QT09MUyJdKQogICAgICAgICAgICApCiAgICAgICAgZWxzZToKICAgICAg
            ICAgICAgc2VsZi5yYWlzZV9leGNlcHRpb25faWZfYW55X3RvcG9sb2d5X2ZsYWdfaXNfbWlzc2lu
            ZygpCgogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1BPT0xfUFJFRklYIl0gPSBzZWxmLl9hcmdf
            cGFyc2VyLnJnd19wb29sX3ByZWZpeAogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX0VORFBPSU5U
            Il0gPSAiIgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50OgogICAgICAg
            ICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgICAgICBzZWxmLmNy
            ZWF0ZV9yZ3dfYWRtaW5fb3BzX3VzZXIoKQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAg
            ICAgaWYgKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25h
            bWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9u
            ZWdyb3VwX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfem9uZV9uYW1lICE9ICIiCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAg
            ICAgIGVyciA9IHNlbGYudmFsaWRhdGVfcmd3X211bHRpc2l0ZSgKICAgICAgICAgICAgICAgICAg
            ICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwgInJlYWxtIgogICAgICAgICAg
            ICAgICAgICAgICkKICAgICAgICAgICAgICAgICAgICBlcnIgPSBzZWxmLnZhbGlkYXRlX3Jnd19t
            dWx0aXNpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pv
            bmVncm91cF9uYW1lLCAiem9uZWdyb3VwIgogICAgICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgICAgICAgICBlcnIgPSBzZWxmLnZhbGlkYXRlX3Jnd19tdWx0aXNpdGUoCiAgICAgICAgICAg
            ICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVfbmFtZSwgInpvbmUiCiAgICAg
            ICAgICAgICAgICAgICAgKQoKICAgICAgICAgICAgICAgIGlmICgKICAgICAgICAgICAgICAgICAg
            ICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lID09ICIiCiAgICAgICAgICAgICAgICAg
            ICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lID09ICIiCiAgICAgICAg
            ICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVfbmFtZSA9PSAiIgogICAg
            ICAgICAgICAgICAgKSBvciAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5y
            Z3dfcmVhbG1fbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFy
            c2VyLnJnd196b25lZ3JvdXBfbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd196b25lX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICk6CiAgICAg
            ICAgICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJH
            V19BRE1JTl9PUFNfVVNFUl9BQ0NFU1NfS0VZIl0sCiAgICAgICAgICAgICAgICAgICAgICAgIHNl
            bGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX1NFQ1JFVF9LRVkiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkLAogICAgICAgICAgICAgICAgICAgICAgICBl
            cnIsCiAgICAgICAgICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRlX3Jnd19hZG1pbl9vcHNfdXNl
            cigpCiAgICAgICAgICAgICAgICAgICAgZXJyID0gc2VsZi52YWxpZGF0ZV9yZ3dfZW5kcG9pbnQo
            aW5mb19jYXBfc3VwcG9ydGVkKQogICAgICAgICAgICAgICAgICAgIGlmIHNlbGYuX2FyZ19wYXJz
            ZXIucmd3X3Rsc19jZXJ0X3BhdGg6CiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYub3V0X21h
            cFsiUkdXX1RMU19DRVJUIl0gPSAoCiAgICAgICAgICAgICAgICAgICAgICAgICAgICBzZWxmLnZh
            bGlkYXRlX3Jnd19lbmRwb2ludF90bHNfY2VydCgpCiAgICAgICAgICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgICAgICAgICAjIGlmIHRoZXJlIGlzIG5vIGVycm9yLCBzZXQgdGhlIFJHV19F
            TkRQT0lOVAogICAgICAgICAgICAgICAgICAgIGlmIGVyciAhPSAiLTEiOgogICAgICAgICAgICAg
            ICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdID0gc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfZW5kcG9pbnQKICAgICAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICAgICAg
            ZXJyID0gIlBsZWFzZSBwcm92aWRlIGFsbCB0aGUgUkdXIG11bHRpc2l0ZSBwYXJhbWV0ZXJzIG9y
            IG5vbmUgb2YgdGhlbSIKICAgICAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKGVycikK
            CiAgICBkZWYgZ2VuX3NoZWxsX291dChzZWxmKToKICAgICAgICBzZWxmLl9nZW5fb3V0cHV0X21h
            cCgpCiAgICAgICAgc2hPdXRJTyA9IFN0cmluZ0lPKCkKICAgICAgICBmb3IgaywgdiBpbiBzZWxm
            Lm91dF9tYXAuaXRlbXMoKToKICAgICAgICAgICAgaWYgdiBhbmQgayBub3QgaW4gc2VsZi5fZXhj
            bHVkZWRfa2V5czoKICAgICAgICAgICAgICAgIHNoT3V0SU8ud3JpdGUoZiJleHBvcnQge2t9PXt2
            fXtMSU5FU0VQfSIpCiAgICAgICAgc2hPdXQgPSBzaE91dElPLmdldHZhbHVlKCkKICAgICAgICBz
            aE91dElPLmNsb3NlKCkKICAgICAgICByZXR1cm4gc2hPdXQKCiAgICBkZWYgZ2VuX2pzb25fb3V0
            KHNlbGYpOgogICAgICAgIHNlbGYuX2dlbl9vdXRwdXRfbWFwKCkKICAgICAgICBpZiBzZWxmLl9h
            cmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiAiIgogICAgICAgIGpzb25fb3V0
            ID0gWwogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFtZSI6ICJyb29rLWNlcGgtbW9u
            LWVuZHBvaW50cyIsCiAgICAgICAgICAgICAgICAia2luZCI6ICJDb25maWdNYXAiLAogICAgICAg
            ICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiBzZWxmLm91dF9t
            YXBbIlJPT0tfRVhURVJOQUxfQ0VQSF9NT05fREFUQSJdLAogICAgICAgICAgICAgICAgICAgICJt
            YXhNb25JZCI6ICIwIiwKICAgICAgICAgICAgICAgICAgICAibWFwcGluZyI6ICJ7fSIsCiAgICAg
            ICAgICAgICAgICB9LAogICAgICAgICAgICB9LAogICAgICAgICAgICB7CiAgICAgICAgICAgICAg
            ICAibmFtZSI6ICJyb29rLWNlcGgtbW9uIiwKICAgICAgICAgICAgICAgICJraW5kIjogIlNlY3Jl
            dCIsCiAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAiYWRtaW4t
            c2VjcmV0IjogImFkbWluLXNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImZzaWQiOiBzZWxm
            Lm91dF9tYXBbIlJPT0tfRVhURVJOQUxfRlNJRCJdLAogICAgICAgICAgICAgICAgICAgICJtb24t
            c2VjcmV0IjogIm1vbi1zZWNyZXQiLAogICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgfSwK
            ICAgICAgICAgICAgewogICAgICAgICAgICAgICAgIm5hbWUiOiAicm9vay1jZXBoLW9wZXJhdG9y
            LWNyZWRzIiwKICAgICAgICAgICAgICAgICJraW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAg
            ICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJS
            T09LX0VYVEVSTkFMX1VTRVJOQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgInVzZXJLZXkiOiBz
            ZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAg
            IH0sCiAgICAgICAgICAgIH0sCiAgICAgICAgXQoKICAgICAgICAjIGlmICdNT05JVE9SSU5HX0VO
            RFBPSU5UJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ21vbml0b3JpbmctZW5kcG9pbnQnIHRvIENs
            dXN0ZXIKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19F
            TkRQT0lOVCJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9J
            TlRfUE9SVCJdCiAgICAgICAgKToKICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAg
            ICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjogIm1vbml0b3JpbmctZW5kcG9p
            bnQiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIkNlcGhDbHVzdGVyIiwKICAgICAgICAg
            ICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgIk1vbml0b3JpbmdF
            bmRwb2ludCI6IHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdLAogICAgICAgICAg
            ICAgICAgICAgICAgICAiTW9uaXRvcmluZ1BvcnQiOiBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdf
            RU5EUE9JTlRfUE9SVCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9
            CiAgICAgICAgICAgICkKCiAgICAgICAgIyBpZiAnQ1NJX1JCRF9OT0RFX1NFQ1JFVCcgZXhpc3Rz
            LCB0aGVuIG9ubHkgYWRkICdyb29rLWNzaS1yYmQtcHJvdmlzaW9uZXInIFNlY3JldAogICAgICAg
            IGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VDUkVUIl0KICAg
            ICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FIl0KICAg
            ICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAg
            ICAgICAgICAgICAgICAgICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9O
            T0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQi
            LAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAi
            dXNlcklEIjogc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VDUkVUX05BTUUiXSwKICAgICAg
            ICAgICAgICAgICAgICAgICAgInVzZXJLZXkiOiBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9T
            RUNSRVQiXSwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAg
            ICAgICApCiAgICAgICAgIyBpZiAnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQnIGV4aXN0cywg
            dGhlbiBvbmx5IGFkZCAncm9vay1jc2ktcmJkLXByb3Zpc2lvbmVyJyBTZWNyZXQKICAgICAgICBp
            ZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQi
            XQogICAgICAgICAgICBhbmQgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JF
            VF9OQU1FIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAg
            ICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21h
            cFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAg
            ICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAg
            ICAgICAgICAgICAgICAgICAgInVzZXJJRCI6IHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJ
            T05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNl
            bGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAg
            ICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBpZiAnQ1NJ
            X0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAncm9vay1j
            c2ktY2VwaGZzLXByb3Zpc2lvbmVyJyBTZWNyZXQKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNl
            bGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVQiXQogICAgICAgICAgICBh
            bmQgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0KICAg
            ICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAg
            ICAgICAgICAgICAgICAgICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhG
            U19QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQi
            OiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAg
            ICAgICAgICAgImFkbWluSUQiOiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJf
            U0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgImFkbWluS2V5Ijogc2VsZi5v
            dXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAg
            ICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdDU0lf
            Q0VQSEZTX05PREVfU0VDUkVUJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ3Jvb2stY3NpLWNlcGhm
            cy1ub2RlJyBTZWNyZXQKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJ
            X0NFUEhGU19OT0RFX1NFQ1JFVCJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9D
            RVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJdCiAgICAgICAgKToKICAgICAgICAgICAganNvbl9vdXQu
            YXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjogZiJy
            b29rLXtzZWxmLm91dF9tYXBbJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAg
            ICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0
            YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImFkbWluSUQiOiBzZWxmLm91dF9tYXBbIkNT
            SV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAiYWRt
            aW5LZXkiOiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVQiXSwKICAgICAgICAg
            ICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBp
            ZiAnUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSycgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdy
            b29rLWNlcGgtZGFzaGJvYXJkLWxpbmsnIFNlY3JldAogICAgICAgIGlmIHNlbGYub3V0X21hcFsi
            Uk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSyJdOgogICAgICAgICAgICBqc29uX291dC5hcHBl
            bmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAicm9vay1j
            ZXBoLWRhc2hib2FyZC1saW5rIiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQi
            LAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAi
            dXNlcklEIjogImNlcGgtZGFzaGJvYXJkLWxpbmsiLAogICAgICAgICAgICAgICAgICAgICAgICAi
            dXNlcktleSI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSyJdLAog
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAg
            ICAgICAjIGlmICdSQURPU19OQU1FU1BBQ0UnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCB0aGUgIlJB
            RE9TX05BTUVTUEFDRSIgbmFtZXNwYWNlCiAgICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLm91
            dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIlJF
            U1RSSUNURURfQVVUSF9QRVJNSVNTSU9OIl0KICAgICAgICAgICAgYW5kIG5vdCBzZWxmLm91dF9t
            YXBbIlJCRF9NRVRBREFUQV9FQ19QT09MX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpz
            b25fb3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFt
            ZSI6ICJyYWRvcy1uYW1lc3BhY2UiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIkNlcGhC
            bG9ja1Bvb2xSYWRvc05hbWVzcGFjZSIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAg
            ICAgICAgICAgICAgICAgICAgICAgICJyYWRvc05hbWVzcGFjZU5hbWUiOiBzZWxmLm91dF9tYXBb
            IlJBRE9TX05BTUVTUEFDRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYu
            b3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAg
            ICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAg
            ICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjogImNlcGgtcmJkLXJhZG9z
            LW5hbWVzcGFjZSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwK
            ICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgInBv
            b2wiOiBzZWxmLm91dF9tYXBbIlJCRF9QT09MX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAg
            ICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNyZXQtbmFtZSI6IGYicm9vay17
            c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAg
            ICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250cm9sbGVyLWV4cGFuZC1z
            ZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NF
            Q1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5p
            by9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRf
            Tk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAg
            ICAgIH0KICAgICAgICAgICAgKQogICAgICAgIGVsc2U6CiAgICAgICAgICAgIGlmIHNlbGYub3V0
            X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdOgogICAgICAgICAgICAgICAganNvbl9v
            dXQuYXBwZW5kKAogICAgICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAgICAg
            Im5hbWUiOiAiY2VwaC1yYmQiLAogICAgICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9y
            YWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAg
            ICAgICAgICAgICAgICAgICJkYXRhUG9vbCI6IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJd
            LAogICAgICAgICAgICAgICAgICAgICAgICAgICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIlJCRF9N
            RVRBREFUQV9FQ19QT09MX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJjc2ku
            c3RvcmFnZS5rOHMuaW8vcHJvdmlzaW9uZXItc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0
            X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNlY3Jl
            dC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVU
            X05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5p
            by9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRf
            Tk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAg
            ICAgICAgICAgICAgfQogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBlbHNlOgogICAgICAg
            ICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgICAgIHsKICAgICAgICAg
            ICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaC1yYmQiLAogICAgICAgICAgICAgICAgICAgICAg
            ICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAgICAgICAiZGF0YSI6
            IHsKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJwb29sIjogc2VsZi5vdXRfbWFwWyJSQkRf
            UE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2Uuazhz
            LmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9S
            QkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAg
            ICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYi
            cm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwK
            ICAgICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vbm9kZS1zdGFn
            ZS1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX05PREVfU0VDUkVU
            X05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgICAg
            IH0KICAgICAgICAgICAgICAgICkKCiAgICAgICAgIyBpZiAnVE9QT0xPR1lfUE9PTFMnLCAnVE9Q
            T0xPR1lfRkFJTFVSRV9ET01BSU5fTEFCRUwnLCAnVE9QT0xPR1lfRkFJTFVSRV9ET01BSU5fVkFM
            VUVTJyAgZXhpc3RzLAogICAgICAgICMgdGhlbiBvbmx5IGFkZCAndG9wb2xvZ3knIFN0b3JhZ2VD
            bGFzcwogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09M
            UyJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlO
            X0xBQkVMIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfRkFJTFVSRV9E
            T01BSU5fVkFMVUVTIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAg
            ICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaC1yYmQtdG9w
            b2xvZ3kiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAg
            ICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJ0b3BvbG9n
            eUZhaWx1cmVEb21haW5MYWJlbCI6IHNlbGYub3V0X21hcFsKICAgICAgICAgICAgICAgICAgICAg
            ICAgICAgICJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9MQUJFTCIKICAgICAgICAgICAgICAgICAg
            ICAgICAgXSwKICAgICAgICAgICAgICAgICAgICAgICAgInRvcG9sb2d5RmFpbHVyZURvbWFpblZh
            bHVlcyI6IHNlbGYub3V0X21hcFsKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJUT1BPTE9H
            WV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiCiAgICAgICAgICAgICAgICAgICAgICAgIF0sCiAgICAg
            ICAgICAgICAgICAgICAgICAgICJ0b3BvbG9neVBvb2xzIjogc2VsZi5vdXRfbWFwWyJUT1BPTE9H
            WV9QT09MUyJdLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3By
            b3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJP
            VklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0
            b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxm
            Lm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUi
            OiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAg
            ICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCgogICAg
            ICAgICMgaWYgJ0NFUEhGU19GU19OQU1FJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ2NlcGhmcycg
            U3RvcmFnZUNsYXNzCiAgICAgICAgaWYgc2VsZi5vdXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdOgog
            ICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAg
            ICAgICAgICAgIm5hbWUiOiAiY2VwaGZzIiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJT
            dG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAg
            ICAgICAgICAgICAiZnNOYW1lIjogc2VsZi5vdXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdLAogICAg
            ICAgICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYub3V0X21hcFsiQ0VQSEZTX1BPT0xfTkFN
            RSJdLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lv
            bmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9DRVBIRlNfUFJPVklT
            SU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3Jh
            Z2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91
            dF9tYXBbJ0NTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUi
            OiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FJ119IiwK
            ICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAg
            ICAgICAgIyBpZiAnUkdXX0VORFBPSU5UJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ2NlcGgtcmd3
            JyBTdG9yYWdlQ2xhc3MKICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdOgog
            ICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAg
            ICAgICAgICAgIm5hbWUiOiAiY2VwaC1yZ3ciLAogICAgICAgICAgICAgICAgICAgICJraW5kIjog
            IlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAg
            ICAgICAgICAgICAgICJlbmRwb2ludCI6IHNlbGYub3V0X21hcFsiUkdXX0VORFBPSU5UIl0sCiAg
            ICAgICAgICAgICAgICAgICAgICAgICJwb29sUHJlZml4Ijogc2VsZi5vdXRfbWFwWyJSR1dfUE9P
            TF9QUkVGSVgiXSwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAg
            ICAgICAgICApCiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsK
            ICAgICAgICAgICAgICAgICAgICAibmFtZSI6ICJyZ3ctYWRtaW4tb3BzLXVzZXIiLAogICAgICAg
            ICAgICAgICAgICAgICJraW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEi
            OiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJhY2Nlc3NLZXkiOiBzZWxmLm91dF9tYXBbIlJH
            V19BRE1JTl9PUFNfVVNFUl9BQ0NFU1NfS0VZIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJz
            ZWNyZXRLZXkiOiBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9TRUNSRVRfS0VZIl0s
            CiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQog
            ICAgICAgICMgaWYgJ1JHV19UTFNfQ0VSVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkIHRoZSAiY2Vw
            aC1yZ3ctdGxzLWNlcnQiIHNlY3JldAogICAgICAgIGlmIHNlbGYub3V0X21hcFsiUkdXX1RMU19D
            RVJUIl06CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAg
            ICAgICAgICAgICAgICAgICAibmFtZSI6ICJjZXBoLXJndy10bHMtY2VydCIsCiAgICAgICAgICAg
            ICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsK
            ICAgICAgICAgICAgICAgICAgICAgICAgImNlcnQiOiBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VS
            VCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAg
            ICkKCiAgICAgICAgcmV0dXJuIGpzb24uZHVtcHMoanNvbl9vdXQpICsgTElORVNFUAoKICAgIGRl
            ZiB1cGdyYWRlX3VzZXJzX3Blcm1pc3Npb25zKHNlbGYpOgogICAgICAgIHVzZXJzID0gWwogICAg
            ICAgICAgICAiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIsCiAgICAgICAgICAgICJjbGllbnQuY3Np
            LWNlcGhmcy1wcm92aXNpb25lciIsCiAgICAgICAgICAgICJjbGllbnQuY3NpLXJiZC1ub2RlIiwK
            ICAgICAgICAgICAgImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwKICAgICAgICAgICAgImNs
            aWVudC5oZWFsdGhjaGVja2VyIiwKICAgICAgICBdCiAgICAgICAgaWYgc2VsZi5ydW5fYXNfdXNl
            ciAhPSAiIiBhbmQgc2VsZi5ydW5fYXNfdXNlciBub3QgaW4gdXNlcnM6CiAgICAgICAgICAgIHVz
            ZXJzLmFwcGVuZChzZWxmLnJ1bl9hc191c2VyKQogICAgICAgIGZvciB1c2VyIGluIHVzZXJzOgog
            ICAgICAgICAgICBzZWxmLnVwZ3JhZGVfdXNlcl9wZXJtaXNzaW9ucyh1c2VyKQoKICAgIGRlZiBn
            ZXRfcmd3X3Bvb2xfbmFtZV9kdXJpbmdfdXBncmFkZShzZWxmLCB1c2VyLCBjYXBzKToKICAgICAg
            ICBpZiB1c2VyID09ICJjbGllbnQuaGVhbHRoY2hlY2tlciI6CiAgICAgICAgICAgICMgd2hlbiBh
            ZG1pbiBoYXMgbm90IHByb3ZpZGVkIHJndyBwb29sIG5hbWUgZHVyaW5nIHVwZ3JhZGUsCiAgICAg
            ICAgICAgICMgZ2V0IHRoZSByZ3cgcG9vbCBuYW1lIGZyb20gY2xpZW50LmhlYWx0aGNoZWNrZXIg
            dXNlciB3aGljaCB3YXMgdXNlZCBkdXJpbmcgY29ubmVjdGlvbgogICAgICAgICAgICBpZiBub3Qg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXg6CiAgICAgICAgICAgICAgICAjIFRvIGdl
            dCB2YWx1ZSAnZGVmYXVsdCcgd2hpY2ggaXMgcmd3IHBvb2wgbmFtZSBmcm9tICdhbGxvdyByd3gg
            cG9vbD1kZWZhdWx0LnJndy5tZXRhJwogICAgICAgICAgICAgICAgcGF0dGVybiA9IHIicG9vbD0o
            Lio/KVwucmd3XC5tZXRhIgogICAgICAgICAgICAgICAgbWF0Y2ggPSByZS5zZWFyY2gocGF0dGVy
            biwgY2FwcykKICAgICAgICAgICAgICAgIGlmIG1hdGNoOgogICAgICAgICAgICAgICAgICAgIHNl
            bGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4ID0gbWF0Y2guZ3JvdXAoMSkKICAgICAgICAg
            ICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAgICAgImZhaWxlZCB0byBnZXQgcmd3IHBvb2wg
            bmFtZSBmb3IgdXBncmFkZSIKICAgICAgICAgICAgICAgICAgICApCgogICAgZGVmIHVwZ3JhZGVf
            dXNlcl9wZXJtaXNzaW9ucyhzZWxmLCB1c2VyKToKICAgICAgICAjIGNoZWNrIHdoZXRoZXIgdGhl
            IGdpdmVuIHVzZXIgZXhpc3RzIG9yIG5vdAogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAi
            YXV0aCBnZXQiLCAiZW50aXR5IjogZiJ7dXNlcn0iLCAiZm9ybWF0IjogImpzb24ifQogICAgICAg
            IHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihj
            bWRfanNvbikKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgog
            ICAgICAgICAgICBwcmludChmInVzZXIge3VzZXJ9IG5vdCBmb3VuZCBmb3IgdXBncmFkaW5nLiIp
            CiAgICAgICAgICAgIHJldHVybgogICAgICAgIGV4aXN0aW5nX2NhcHMgPSBqc29uX291dFswXVsi
            Y2FwcyJdCiAgICAgICAgc2VsZi5nZXRfcmd3X3Bvb2xfbmFtZV9kdXJpbmdfdXBncmFkZSh1c2Vy
            LCBzdHIoZXhpc3RpbmdfY2FwcykpCiAgICAgICAgbmV3X2NhcCwgXyA9IHNlbGYuZ2V0X2NhcHNf
            YW5kX2VudGl0eSh1c2VyKQogICAgICAgIGNhcF9rZXlzID0gWyJtb24iLCAibWdyIiwgIm9zZCIs
            ICJtZHMiXQogICAgICAgIGNhcHMgPSBbXQogICAgICAgIGZvciBlYWNoQ2FwIGluIGNhcF9rZXlz
            OgogICAgICAgICAgICBjdXJfY2FwX3ZhbHVlcyA9IGV4aXN0aW5nX2NhcHMuZ2V0KGVhY2hDYXAs
            ICIiKQogICAgICAgICAgICBuZXdfY2FwX3ZhbHVlcyA9IG5ld19jYXAuZ2V0KGVhY2hDYXAsICIi
            KQogICAgICAgICAgICBjdXJfY2FwX3Blcm1fbGlzdCA9IFsKICAgICAgICAgICAgICAgIHguc3Ry
            aXAoKSBmb3IgeCBpbiBjdXJfY2FwX3ZhbHVlcy5zcGxpdCgiLCIpIGlmIHguc3RyaXAoKQogICAg
            ICAgICAgICBdCiAgICAgICAgICAgIG5ld19jYXBfcGVybV9saXN0ID0gWwogICAgICAgICAgICAg
            ICAgeC5zdHJpcCgpIGZvciB4IGluIG5ld19jYXBfdmFsdWVzLnNwbGl0KCIsIikgaWYgeC5zdHJp
            cCgpCiAgICAgICAgICAgIF0KICAgICAgICAgICAgIyBhcHBlbmQgbmV3X2NhcF9saXN0IHRvIGN1
            cl9jYXBfbGlzdCB0byBtYWludGFpbiB0aGUgb3JkZXIgb2YgY2FwcwogICAgICAgICAgICBjdXJf
            Y2FwX3Blcm1fbGlzdC5leHRlbmQobmV3X2NhcF9wZXJtX2xpc3QpCiAgICAgICAgICAgICMgZWxp
            bWluYXRlIGR1cGxpY2F0ZXMgd2l0aG91dCB1c2luZyAnc2V0JwogICAgICAgICAgICAjIHNldCBy
            ZS1vcmRlcnMgaXRlbXMgaW4gdGhlIGxpc3QgYW5kIHdlIGhhdmUgdG8ga2VlcCB0aGUgb3JkZXIK
            ICAgICAgICAgICAgbmV3X2NhcF9saXN0ID0gW10KICAgICAgICAgICAgW25ld19jYXBfbGlzdC5h
            cHBlbmQoeCkgZm9yIHggaW4gY3VyX2NhcF9wZXJtX2xpc3QgaWYgeCBub3QgaW4gbmV3X2NhcF9s
            aXN0XQogICAgICAgICAgICBleGlzdGluZ19jYXBzW2VhY2hDYXBdID0gIiwgIi5qb2luKG5ld19j
            YXBfbGlzdCkKICAgICAgICAgICAgaWYgZXhpc3RpbmdfY2Fwc1tlYWNoQ2FwXToKICAgICAgICAg
            ICAgICAgIGNhcHMuYXBwZW5kKGVhY2hDYXApCiAgICAgICAgICAgICAgICBjYXBzLmFwcGVuZChl
            eGlzdGluZ19jYXBzW2VhY2hDYXBdKQogICAgICAgIGNtZF9qc29uID0gewogICAgICAgICAgICAi
            cHJlZml4IjogImF1dGggY2FwcyIsCiAgICAgICAgICAgICJlbnRpdHkiOiB1c2VyLAogICAgICAg
            ICAgICAiY2FwcyI6IGNhcHMsCiAgICAgICAgICAgICJmb3JtYXQiOiAianNvbiIsCiAgICAgICAg
            fQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9q
            c29uX2dlbihjbWRfanNvbikKICAgICAgICBpZiByZXRfdmFsICE9IDA6CiAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIidhdXRoIGNh
            cHMge3VzZXJ9JyBjb21tYW5kIGZhaWxlZC5cbiBFcnJvcjoge2Vycl9tc2d9IgogICAgICAgICAg
            ICApCiAgICAgICAgcHJpbnQoZiJVcGRhdGVkIHVzZXIge3VzZXJ9IHN1Y2Nlc3NmdWxseS4iKQoK
            ICAgIGRlZiBtYWluKHNlbGYpOgogICAgICAgIGdlbmVyYXRlZF9vdXRwdXQgPSAiIgogICAgICAg
            IGlmIHNlbGYuX2FyZ19wYXJzZXIudXBncmFkZToKICAgICAgICAgICAgc2VsZi51cGdyYWRlX3Vz
            ZXJzX3Blcm1pc3Npb25zKCkKICAgICAgICBlbGlmIHNlbGYuX2FyZ19wYXJzZXIuZm9ybWF0ID09
            ICJqc29uIjoKICAgICAgICAgICAgZ2VuZXJhdGVkX291dHB1dCA9IHNlbGYuZ2VuX2pzb25fb3V0
            KCkKICAgICAgICBlbGlmIHNlbGYuX2FyZ19wYXJzZXIuZm9ybWF0ID09ICJiYXNoIjoKICAgICAg
            ICAgICAgZ2VuZXJhdGVkX291dHB1dCA9IHNlbGYuZ2VuX3NoZWxsX291dCgpCiAgICAgICAgZWxz
            ZToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAg
            ICAgICAgIGYiVW5zdXBwb3J0ZWQgZm9ybWF0OiB7c2VsZi5fYXJnX3BhcnNlci5mb3JtYXR9Igog
            ICAgICAgICAgICApCiAgICAgICAgcHJpbnQoZ2VuZXJhdGVkX291dHB1dCkKICAgICAgICBpZiBz
            ZWxmLm91dHB1dF9maWxlIGFuZCBnZW5lcmF0ZWRfb3V0cHV0OgogICAgICAgICAgICBmT3V0ID0g
            b3BlbihzZWxmLm91dHB1dF9maWxlLCBtb2RlPSJ3IiwgZW5jb2Rpbmc9IlVURi04IikKICAgICAg
            ICAgICAgZk91dC53cml0ZShnZW5lcmF0ZWRfb3V0cHV0KQogICAgICAgICAgICBmT3V0LmNsb3Nl
            KCkKCgojIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMKIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIE1BSU4gIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIwppZiBfX25hbWVfXyA9PSAiX19t
            YWluX18iOgogICAgcmpPYmogPSBSYWRvc0pTT04oKQogICAgdHJ5OgogICAgICAgIHJqT2JqLm1h
            aW4oKQogICAgZXhjZXB0IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24gYXMgZXJyOgogICAgICAg
            IHByaW50KGYiRXhlY3V0aW9uIEZhaWxlZDoge2Vycn0iKQogICAgICAgIHJhaXNlIGVycgogICAg
            ZXhjZXB0IEtleUVycm9yIGFzIGtFcnI6CiAgICAgICAgcHJpbnQoZiJLZXlFcnJvcjoge2tFcnJ9
            IikKICAgIGV4Y2VwdCBPU0Vycm9yIGFzIG9zRXJyOgogICAgICAgIHByaW50KGYiRXJyb3Igd2hp
            bGUgdHJ5aW5nIHRvIG91dHB1dCB0aGUgZGF0YToge29zRXJyfSIpCiAgICBmaW5hbGx5OgogICAg
            ICAgIHJqT2JqLnNodXRkb3duKCkK
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/red-hat-storage/rook
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Ceph cluster.
            displayName: Ceph Cluster
            kind: CephCluster
            name: cephclusters.ceph.rook.io
            version: v1
          - description: Represents a Ceph Block Pool.
            displayName: Ceph Block Pool
            kind: CephBlockPool
            name: cephblockpools.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store.
            displayName: Ceph Object Store
            kind: CephObjectStore
            name: cephobjectstores.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store User.
            displayName: Ceph Object Store User
            kind: CephObjectStoreUser
            name: cephobjectstoreusers.ceph.rook.io
            version: v1
          - description: Represents a cluster of Ceph NFS ganesha gateways.
            displayName: Ceph NFS
            kind: CephNFS
            name: cephnfses.ceph.rook.io
            version: v1
          - description: Represents a Ceph User.
            displayName: Ceph Client
            kind: CephClient
            name: cephclients.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem.
            displayName: Ceph Filesystem
            kind: CephFilesystem
            name: cephfilesystems.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem Mirror.
            displayName: Ceph Filesystem Mirror
            kind: CephFilesystemMirror
            name: cephfilesystemmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph RBD Mirror.
            displayName: Ceph RBD Mirror
            kind: CephRBDMirror
            name: cephrbdmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Realm.
            displayName: Ceph Object Store Realm
            kind: CephObjectRealm
            name: cephobjectrealms.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone Group.
            displayName: Ceph Object Store Zone Group
            kind: CephObjectZoneGroup
            name: cephobjectzonegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone.
            displayName: Ceph Object Store Zone
            kind: CephObjectZone
            name: cephobjectzones.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Notification.
            displayName: Ceph Bucket Notification
            kind: CephBucketNotification
            name: cephbucketnotifications.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Topic.
            displayName: Ceph Bucket Topic
            kind: CephBucketTopic
            name: cephbuckettopics.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem SubVolumeGroup.
            displayName: Ceph Filesystem SubVolumeGroup
            kind: CephFilesystemSubVolumeGroup
            name: cephfilesystemsubvolumegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph BlockPool Rados Namespace.
            displayName: Ceph BlockPool Rados Namespace
            kind: CephBlockPoolRadosNamespace
            name: cephblockpoolradosnamespaces.ceph.rook.io
            version: v1
          - description: Represents a Ceph COSI Driver.
            displayName: Ceph COSI Driver
            kind: CephCOSIDriver
            name: cephcosidrivers.ceph.rook.io
            version: v1
        description: |2

          The Rook-Ceph storage operator packages, deploys, manages, upgrades and scales Ceph storage for providing persistent storage to infrastructure services (Logging, Metrics, Registry) as well as stateful applications in Kubernetes clusters.

          ## Rook-Ceph Storage Operator

          Rook runs as a cloud-native service in Kubernetes clusters for optimal integration with applications in need of storage, and handles the heavy-lifting behind the scenes such as provisioning and management.
          Rook orchestrates battle-tested open-source storage technology Ceph, which has years of production deployments and runs some of the worlds largest clusters.

          Ceph is a massively scalable, software-defined, cloud native storage platform that offers block, file and object storage services.
          Ceph can be used to back a wide variety of applications including relational databases, NoSQL databases, CI/CD tool-sets, messaging, AI/ML and analytics applications.
          Ceph is a proven storage platform that backs some of the world's largest storage deployments and has a large vibrant open source community backing the project.

          ## Supported features
          * **High Availability and resiliency** - Ceph has no single point of failures (SPOF) and all its components work natively in a highly available fashion
          * **Data Protection** - Ceph periodically scrub for inconsistent objects and repair them if necessary, making sure your replicas are always coherent
          * **Consistent storage platform across hybrid cloud** - Ceph can be deployed anywhere (on-premise or bare metal) and thus offers a similar experience regardless
          * **Block, File & Object storage service** - Ceph can expose your data through several storage interfaces, solving all the application use cases
          * **Scale up/down** - addition and removal of storage is fully covered by the operator.
          * **Dashboard** - The Operator deploys a dashboard for monitoring and introspecting your cluster.

          ## Before you start
          https://rook.io/docs/rook/v1.0/k8s-pre-reqs.html
        displayName: Rook-Ceph
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rook
        - ceph
        - storage
        - object storage
        - open source
        - block storage
        - shared filesystem
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/rook
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:cf7f72469d25fb37301a8159406613673f8968c82af1035fd45a470e0ae8acfa
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:80c01633a6bfca44df37ac38f764035be6998fce965c8709e4f6d54d17eeaa7b
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6fee56851da3a87b9f7f477fc59e3385e118dc0b398861a8f13459c998a5af77
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c11a599bfd09b9b22888e03eabdbe1054d4049718a591d2a45bf0ed89800111e
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:82a4b481028195cfd75471656fe40c1cbbade4c7b2d170d6bf9422d9d3b84fa7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:693a60d6a1162659213ae554b1c6ee200b0300a7da72c01415044cf507acab49
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:0afcb516eb528857887c6ee1a9cba081dc8f6b7d47d9f26a0dbb2dca0cefe07c
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:8d8326c74795c9c6bd8f33a9f9e81d2304a70b1f9be3c2f8d50e76be4b0292ff
        version: 4.16.16-rhodf
      entries:
      - name: rook-ceph-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: rook-ceph-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: rook-ceph-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: rook-ceph-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: rook-ceph-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: rook-ceph-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: rook-ceph-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: rook-ceph-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: rook-ceph-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: rook-ceph-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: rook-ceph-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: rook-ceph-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: rook-ceph-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: rook-ceph-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: rook-ceph-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: rook-ceph-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: rook-ceph-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: rook-ceph-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: logic-operator-rhel8
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: logic-operator-rhel8.v1.36.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "host/namespace",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          createdAt: "2025-05-26T14:10:15Z"
          description: OpenShift Serverless Logic Kubernetes Operator for deploying
            workflow applications based on the CNCF Serverless Workflow specification
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-serverless-logic
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kubesmarts/kie-tools
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: Sonata Flow Build
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: Sonata Flow Cluster Platform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: Sonata Flow Platform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: Sonata Flow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          OpenShift Serverless Logic Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification v0.8](https://serverlessworkflow.io/).
        displayName: OpenShift Serverless Logic Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        - workflows
        links:
        - name: Documentation
          url: https://docs.openshift.com/serverless/
        - name: Source Repository
          url: https://github.com/kiegroup/kogito-serverless-operator/
        maintainers:
        - email: serverless-logic-team@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/logic-rhel8-operator@sha256:8d3682448ebdac3aeabb2d23842b7e67a252b95f959c408af805037f9728fd3c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4564ca3dc5bac80d6faddaf94c817fbbc270698a9399d8a21ee1005d85ceda56
        version: 1.36.0
      entries:
      - name: logic-operator-rhel8.v1.36.0
        version: 1.36.0
      - name: logic-operator-rhel8.v1.35.0
        version: 1.35.0
      - name: logic-operator-rhel8.v1.34.0
        version: 1.34.0
      - name: logic-operator-rhel8.v1.33.0
        version: 1.33.0
      name: alpha
    defaultChannel: alpha
    packageName: logic-operator-rhel8
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: The cert-manager maintainers
      provider-url: https://cert-manager.io/
    name: cert-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-manager.v1.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "example-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "example-clusterissuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example-certificate"
                },
                "spec": {
                  "dnsNames": [
                    "example.com"
                  ],
                  "issuerRef": {
                    "name": "example-issuer"
                  },
                  "secretName": "example-certificate-tls"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "name": "example-certificaterequest"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "name": "example-order"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "example-challenge"
                },
                "spec": {}
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: quay.io/jetstack/cert-manager-controller:v1.16.5
          createdAt: 2025-05-02T09:11:31
          olm.skipRange: '>=1.16.0 <1.16.5'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/internal-objects: |-
            [
              "challenges.acme.cert-manager.io",
              "orders.acme.cert-manager.io"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/cert-manager/cert-manager
          support: The cert-manager maintainers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A CertificateRequest is used to request a signed certificate from one of the
              configured issuers.

              All fields within the CertificateRequest's `spec` are immutable after creation.
              A CertificateRequest will either succeed or fail, as denoted by its `Ready` status
              condition and its `status.failureTime` field.

              A CertificateRequest is a one-shot resource, meaning it represents a single
              point in time request for a certificate and cannot be re-used.
            displayName: CertificateRequest
            kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - description: |-
              A Certificate resource should be created to ensure an up to date and signed
              X.509 certificate is stored in the Kubernetes Secret resource named in `spec.secretName`.

              The stored certificate will be renewed before it expires (as configured by `spec.renewBefore`).
            displayName: Certificate
            kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: Challenge is a type to represent a Challenge request with
              an ACME server
            displayName: (Internal) Challenge
            kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - description: |-
              A ClusterIssuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is similar to an Issuer, however it is cluster-scoped and therefore can
              be referenced by resources that exist in *any* namespace, not just the same
              namespace as the referent.
            displayName: ClusterIssuer
            kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - description: |-
              An Issuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is scoped to a single namespace and can therefore only be referenced by
              resources within the same namespace.
            displayName: Issuer
            kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - description: Order is a type to represent an Order with an ACME server
            displayName: (Internal) Order
            kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          cert-manager is a Kubernetes add-on to automate the management and issuance of TLS certificates from various issuing sources.
          It will ensure certificates are valid and up to date periodically, and attempt to renew certificates at an appropriate time before expiry.

          cert-manager has the following [built in issuers][]:

          * ACME: sign certificates using Let's Encrypt (or any other [ACME][] compliant signer).
          * Vault: sign certificates using HashiCorp Vault.
          * Venafi: sign certificates using Venafi Cloud or Venafi Trust Protection Platform.
          * SelfSigned: the private key of the certificate will be used to sign the certificate.
          * CA: a private key stored as a Kubernetes Secret will be used to sign the certificate.

          [built in issuers]: https://cert-manager.io/docs/configuration/#supported-issuer-types
          [ACME]: https://datatracker.ietf.org/doc/html/rfc8555

          cert-manager also supports [external issuers][],
          which are out-of-tree components that provide support for other certificate authorities, such as:
          [Google Cloud Certificate Authority](https://github.com/jetstack/google-cas-issuer),
          [AWS Private Certificate Authority](https://github.com/cert-manager/aws-privateca-issuer), and
          [Cloudflare Origin CA](https://github.com/cloudflare/origin-ca-issuer)

          [external issuers]: https://cert-manager.io/docs/configuration/external/

          In order to configure cert-manager to issue certificates, you must first create an Issuer or ClusterIssuer.
          These resources represent a particular signing authority and tell cert-manager how to handle your certificate requests.
          The [Issuer Documentation][] has more detail.

          [configure cert-manager]: https://cert-manager.io/docs/configuration/
          [Issuer Documentation]: https://cert-manager.io/docs/concepts/issuer/

          Once an Issuer has been configured, you can create a Certificate resource,
          which is a definition of a certificate request that is to be honored by an issuer
          and which is to be kept up-to-date.
          This is the usual way that you will interact with cert-manager to request signed certificates.
          The [Certificate Documentation][] has more detail.

          [Certificate Documentation]: https://cert-manager.io/docs/usage/certificate/
        displayName: cert-manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Documentation
          url: https://cert-manager.io/
        - name: Release Notes
          url: https://cert-manager.io/docs/release-notes
        maintainers:
        - email: cert-manager-maintainers@googlegroups.com
          name: The cert-manager maintainers
        maturity: stable
        minKubeVersion: 1.19.0-0
        provider:
          name: The cert-manager maintainers
          url: https://cert-manager.io/
        relatedImages:
        - quay.io/jetstack/cert-manager-controller:v1.16.5
        - quay.io/jetstack/cert-manager-cainjector:v1.16.5
        - quay.io/jetstack/cert-manager-webhook:v1.16.5
        version: 1.16.5
      entries:
      - name: cert-manager.v1.16.5
        version: 1.16.5
      - name: cert-manager.v1.16.1
        version: 1.16.1
      - name: cert-manager.v1.16.0-beta.0
        version: 1.16.0-beta.0
      - name: cert-manager.v1.15.2
        version: 1.15.2
      - name: cert-manager.v1.15.0
        version: 1.15.0
      - name: cert-manager.v1.14.2
        version: 1.14.2
      - name: cert-manager.v1.14.2-rc1
        version: 1.14.2-rc1
      - name: cert-manager.v1.14.1-rc1
        version: 1.14.1-rc1
      - name: cert-manager.v1.13.3
        version: 1.13.3
      - name: cert-manager.v1.13.3-rc1
        version: 1.13.3-rc1
      - name: cert-manager.v1.13.1
        version: 1.13.1
      - name: cert-manager.v1.13.1-rc1
        version: 1.13.1-rc1
      - name: cert-manager.v1.12.2
        version: 1.12.2
      - name: cert-manager.v1.12.2-rc1
        version: 1.12.2-rc1
      - name: cert-manager.v1.11.4
        version: 1.11.4
      - name: cert-manager.v1.11.4-rc1
        version: 1.11.4-rc1
      - name: cert-manager.v1.11.3-rc1
        version: 1.11.3-rc1
      - name: cert-manager.v1.11.0
        version: 1.11.0
      - name: cert-manager.v1.11.0-rc1
        version: 1.11.0-rc1
      - name: cert-manager.v1.10.2
        version: 1.10.2
      - name: cert-manager.v1.10.2-rc1
        version: 1.10.2-rc1
      - name: cert-manager.v1.10.1
        version: 1.10.1
      - name: cert-manager.v1.10.1-rc1
        version: 1.10.1-rc1
      - name: cert-manager.v1.10.0
        version: 1.10.0
      - name: cert-manager.v1.10.0-rc1
        version: 1.10.0-rc1
      - name: cert-manager.v1.9.1
        version: 1.9.1
      - name: cert-manager.v1.9.1-rc1
        version: 1.9.1-rc1
      - name: cert-manager.v1.8.2
        version: 1.8.2
      - name: cert-manager.v1.8.2-rc1
        version: 1.8.2-rc1
      - name: cert-manager.v1.8.0
        version: 1.8.0
      - name: cert-manager.v1.8.0-rc1
        version: 1.8.0-rc1
      - name: cert-manager.v1.7.2
        version: 1.7.2
      - name: cert-manager.v1.7.1
        version: 1.7.1
      - name: cert-manager.v1.7.1-rc1
        version: 1.7.1-rc1
      - name: cert-manager.v1.6.2
        version: 1.6.2
      - name: cert-manager.v1.6.2-rc1
        version: 1.6.2-rc1
      name: candidate
    - currentCSV: cert-manager.v1.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "example-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "example-clusterissuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example-certificate"
                },
                "spec": {
                  "dnsNames": [
                    "example.com"
                  ],
                  "issuerRef": {
                    "name": "example-issuer"
                  },
                  "secretName": "example-certificate-tls"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "name": "example-certificaterequest"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "name": "example-order"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "example-challenge"
                },
                "spec": {}
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: quay.io/jetstack/cert-manager-controller:v1.16.5
          createdAt: 2025-05-02T09:11:31
          olm.skipRange: '>=1.16.0 <1.16.5'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/internal-objects: |-
            [
              "challenges.acme.cert-manager.io",
              "orders.acme.cert-manager.io"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/cert-manager/cert-manager
          support: The cert-manager maintainers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A CertificateRequest is used to request a signed certificate from one of the
              configured issuers.

              All fields within the CertificateRequest's `spec` are immutable after creation.
              A CertificateRequest will either succeed or fail, as denoted by its `Ready` status
              condition and its `status.failureTime` field.

              A CertificateRequest is a one-shot resource, meaning it represents a single
              point in time request for a certificate and cannot be re-used.
            displayName: CertificateRequest
            kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - description: |-
              A Certificate resource should be created to ensure an up to date and signed
              X.509 certificate is stored in the Kubernetes Secret resource named in `spec.secretName`.

              The stored certificate will be renewed before it expires (as configured by `spec.renewBefore`).
            displayName: Certificate
            kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: Challenge is a type to represent a Challenge request with
              an ACME server
            displayName: (Internal) Challenge
            kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - description: |-
              A ClusterIssuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is similar to an Issuer, however it is cluster-scoped and therefore can
              be referenced by resources that exist in *any* namespace, not just the same
              namespace as the referent.
            displayName: ClusterIssuer
            kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - description: |-
              An Issuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is scoped to a single namespace and can therefore only be referenced by
              resources within the same namespace.
            displayName: Issuer
            kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - description: Order is a type to represent an Order with an ACME server
            displayName: (Internal) Order
            kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          cert-manager is a Kubernetes add-on to automate the management and issuance of TLS certificates from various issuing sources.
          It will ensure certificates are valid and up to date periodically, and attempt to renew certificates at an appropriate time before expiry.

          cert-manager has the following [built in issuers][]:

          * ACME: sign certificates using Let's Encrypt (or any other [ACME][] compliant signer).
          * Vault: sign certificates using HashiCorp Vault.
          * Venafi: sign certificates using Venafi Cloud or Venafi Trust Protection Platform.
          * SelfSigned: the private key of the certificate will be used to sign the certificate.
          * CA: a private key stored as a Kubernetes Secret will be used to sign the certificate.

          [built in issuers]: https://cert-manager.io/docs/configuration/#supported-issuer-types
          [ACME]: https://datatracker.ietf.org/doc/html/rfc8555

          cert-manager also supports [external issuers][],
          which are out-of-tree components that provide support for other certificate authorities, such as:
          [Google Cloud Certificate Authority](https://github.com/jetstack/google-cas-issuer),
          [AWS Private Certificate Authority](https://github.com/cert-manager/aws-privateca-issuer), and
          [Cloudflare Origin CA](https://github.com/cloudflare/origin-ca-issuer)

          [external issuers]: https://cert-manager.io/docs/configuration/external/

          In order to configure cert-manager to issue certificates, you must first create an Issuer or ClusterIssuer.
          These resources represent a particular signing authority and tell cert-manager how to handle your certificate requests.
          The [Issuer Documentation][] has more detail.

          [configure cert-manager]: https://cert-manager.io/docs/configuration/
          [Issuer Documentation]: https://cert-manager.io/docs/concepts/issuer/

          Once an Issuer has been configured, you can create a Certificate resource,
          which is a definition of a certificate request that is to be honored by an issuer
          and which is to be kept up-to-date.
          This is the usual way that you will interact with cert-manager to request signed certificates.
          The [Certificate Documentation][] has more detail.

          [Certificate Documentation]: https://cert-manager.io/docs/usage/certificate/
        displayName: cert-manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Documentation
          url: https://cert-manager.io/
        - name: Release Notes
          url: https://cert-manager.io/docs/release-notes
        maintainers:
        - email: cert-manager-maintainers@googlegroups.com
          name: The cert-manager maintainers
        maturity: stable
        minKubeVersion: 1.19.0-0
        provider:
          name: The cert-manager maintainers
          url: https://cert-manager.io/
        relatedImages:
        - quay.io/jetstack/cert-manager-controller:v1.16.5
        - quay.io/jetstack/cert-manager-cainjector:v1.16.5
        - quay.io/jetstack/cert-manager-webhook:v1.16.5
        version: 1.16.5
      entries:
      - name: cert-manager.v1.16.5
        version: 1.16.5
      - name: cert-manager.v1.16.1
        version: 1.16.1
      - name: cert-manager.v1.15.2
        version: 1.15.2
      - name: cert-manager.v1.15.0
        version: 1.15.0
      - name: cert-manager.v1.14.2
        version: 1.14.2
      - name: cert-manager.v1.13.3
        version: 1.13.3
      - name: cert-manager.v1.13.1
        version: 1.13.1
      - name: cert-manager.v1.13.1-rc1
        version: 1.13.1-rc1
      - name: cert-manager.v1.12.2
        version: 1.12.2
      - name: cert-manager.v1.11.4
        version: 1.11.4
      - name: cert-manager.v1.11.0
        version: 1.11.0
      - name: cert-manager.v1.10.2
        version: 1.10.2
      - name: cert-manager.v1.10.1
        version: 1.10.1
      - name: cert-manager.v1.10.0
        version: 1.10.0
      - name: cert-manager.v1.9.1
        version: 1.9.1
      - name: cert-manager.v1.8.2
        version: 1.8.2
      - name: cert-manager.v1.8.0
        version: 1.8.0
      - name: cert-manager.v1.7.2
        version: 1.7.2
      - name: cert-manager.v1.7.1
        version: 1.7.1
      - name: cert-manager.v1.6.2
        version: 1.6.2
      - name: cert-manager.v1.6.1
        version: 1.6.1
      - name: cert-manager.v1.6.0
        version: 1.6.0
      - name: cert-manager.v1.5.4
        version: 1.5.4
      - name: cert-manager.v1.5.3
        version: 1.5.3
      - name: cert-manager.v1.4.4
        version: 1.4.4
      - name: cert-manager.v1.4.3
        version: 1.4.3
      - name: cert-manager.v1.4.2
        version: 1.4.2
      - name: cert-manager.v1.4.1
        version: 1.4.1
      - name: cert-manager.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: cert-manager
    provider:
      name: The cert-manager maintainers
      url: https://cert-manager.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: TrendMicro
      provider-url: trendmicro.com
    name: visionone-containersecurity
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: visionone-containersecurity.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "visionone-containersecurity-sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          com.redhat.openshift.versions: v4.12-v4.20
          containerImage: public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
          createdAt: "2025-09-18T22:43:41Z"
          description: Trend Vision One™ Container Security delivers full lifecycle
            protection for containers with real-time threat detection, policy enforcement,
            compliance assurance, and rapid incident response.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trendmicro-system
          operatorhub.io/ui-metadata-max-k8s-version: "1.26"
          operators.openshift.io/valid-subscription: VisionOneContainerSecurity Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trendmicro/visionone-container-security-helm
          support: TrendMicro Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trend Vision One™ Container Security delivers full lifecycle
              protection for containers with real-time threat detection, policy enforcement,
              compliance assurance, and rapid incident response.
            displayName: VisionOneContainerSecurity
            kind: VisionOneContainerSecurity
            name: visiononecontainersecurities.container-security.trendmicro.com
            version: v1alpha1
        description: "[Trend Vision One™ Container Security](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-container-security)
          delivers comprehensive, end-to-end\n protection for containerized workloads
          across OpenShift and Kubernetes environments.\n It continuously monitors
          for threats, vulnerabilities, policy violations, and\n exposed secrets ensuring
          compliance and enabling rapid, automated response. Purpose-built\n for hybrid
          cloud, it instantly isolates compromised containers to prevent lateral\n
          movement and minimize impact. Seamlessly integrating into OpenShift DevSecOps\n
          pipelines, it empowers security teams to enforce policies while enabling
          developers\n to innovate and deploy with confidence without disrupting workflows
          or compromising\n agility.\n\n **Installation**\n \n - Use the recommended
          `trendmicro-system` namespace.\n - Refer to the [Trend Vision One™ Documentation](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator)
          for registration and installation steps.\n"
        displayName: Trend Vision One™ Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trendmicro
        - visionone
        - container-security
        links:
        - name: Documentation
          url: https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator
        maintainers:
        - email: support@trendmicro.com
          name: TrendMicro VisionOne Support Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: TrendMicro
          url: trendmicro.com
        relatedImages:
        - public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
        version: 0.0.2
      entries:
      - name: visionone-containersecurity.v0.0.2
        version: 0.0.2
      - name: visionone-containersecurity.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: visionone-containersecurity.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "visionone-containersecurity-sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          com.redhat.openshift.versions: v4.12-v4.20
          containerImage: public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
          createdAt: "2025-09-18T22:43:41Z"
          description: Trend Vision One™ Container Security delivers full lifecycle
            protection for containers with real-time threat detection, policy enforcement,
            compliance assurance, and rapid incident response.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trendmicro-system
          operatorhub.io/ui-metadata-max-k8s-version: "1.26"
          operators.openshift.io/valid-subscription: VisionOneContainerSecurity Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trendmicro/visionone-container-security-helm
          support: TrendMicro Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trend Vision One™ Container Security delivers full lifecycle
              protection for containers with real-time threat detection, policy enforcement,
              compliance assurance, and rapid incident response.
            displayName: VisionOneContainerSecurity
            kind: VisionOneContainerSecurity
            name: visiononecontainersecurities.container-security.trendmicro.com
            version: v1alpha1
        description: "[Trend Vision One™ Container Security](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-container-security)
          delivers comprehensive, end-to-end\n protection for containerized workloads
          across OpenShift and Kubernetes environments.\n It continuously monitors
          for threats, vulnerabilities, policy violations, and\n exposed secrets ensuring
          compliance and enabling rapid, automated response. Purpose-built\n for hybrid
          cloud, it instantly isolates compromised containers to prevent lateral\n
          movement and minimize impact. Seamlessly integrating into OpenShift DevSecOps\n
          pipelines, it empowers security teams to enforce policies while enabling
          developers\n to innovate and deploy with confidence without disrupting workflows
          or compromising\n agility.\n\n **Installation**\n \n - Use the recommended
          `trendmicro-system` namespace.\n - Refer to the [Trend Vision One™ Documentation](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator)
          for registration and installation steps.\n"
        displayName: Trend Vision One™ Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trendmicro
        - visionone
        - container-security
        links:
        - name: Documentation
          url: https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator
        maintainers:
        - email: support@trendmicro.com
          name: TrendMicro VisionOne Support Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: TrendMicro
          url: trendmicro.com
        relatedImages:
        - public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
        version: 0.0.2
      entries:
      - name: visionone-containersecurity.v0.0.2
        version: 0.0.2
      name: stable
    defaultChannel: stable
    packageName: visionone-containersecurity
    provider:
      name: TrendMicro
      url: trendmicro.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Docling Maintainers
      provider-url: https://docling-project.github.io/docling/
    name: docling-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: docling-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2025-05-19T20:20:14Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DoclingServe is the Schema for the doclingserves API
            displayName: Docling Serve
            kind: DoclingServe
            name: doclingserves.docling.github.io
            version: v1alpha1
        description: |-
          **Overview**

          The Docling Operator is responsible for managing the full lifecycle of Docling Serve instances in your Kubernetes or OpenShift container platforms.


          **Docling Serve**

          The Docling Serve application deploys a REST API for consuming and integrating Docling in your workflows.

          The API offers endpoints for multiple types of interaction, i.e. both synchronous and asynchronous endpoints, as well as status subscriptions via websockets.

          The options for the Docling conversion can be customized at runtime by the users (e.g. which OCR engine, which enrichment models, etc). Users can also choose which of the many DoclingDocument export formats (JSON, Markdown, HTML, DocTags, etc) to receive in the result response.

          Docling Serve allows to choose between a local compute engine, i.e. running Docling in the same container as the API server, or a distributed compute engine which leverages Kubeflow Pipelines to parallelize large-scale ingestion jobs.

          **Docling**

          Docling is a powerful library which simplifies document processing, parsing diverse formats — including advanced PDF understanding — and providing seamless integrations with the gen AI ecosystem.

          Learn more about the Docling Project by visiting the GitHub organization at [https://github.com/docling-project](https://github.com/docling-project).


          **Sample DoclingServe Custom Resources**


          To deploy with the `local` compute engine with the UI enabled and Route exposed:
          ```yaml
          apiVersion: docling.github.io/v1alpha1
          kind: DoclingServe
          metadata:
            labels:
              app.kubernetes.io/name: docling-operator
              app.kubernetes.io/managed-by: kustomize
            name: doclingserve-local-sample
          spec:
            apiServer:
              image: "quay.io/docling-project/docling-serve:latest"
              enableUI: true
              instances: 1
            engine:
              local:
                numWorkers: 2
            route:
              enabled: true
          ```

          To deploy with the `KFP` compute engine with the UI enabled and Route exposed. **Note: Kubeflow Pipeline must be installed and setup in the cluster to use this option.**:
          ```yaml
          apiVersion: docling.github.io/v1alpha1
          kind: DoclingServe
          metadata:
            labels:
              app.kubernetes.io/name: docling-operator
              app.kubernetes.io/managed-by: kustomize
            name: doclingserve-kfp-sample
          spec:
            apiServer:
              image: "quay.io/docling-project/docling-serve:latest"
              enableUI: true
              instances: 1
            engine:
              kfp:
                endpoint: "https://NAME.NAMESPACE.svc.cluster.local:8888"
            route:
              enabled: true
          ```

          **Additional Configurations**


          The CRD exposes `apiServer.configMapName` to enable additional configurations of the `docling-serve` workload. For a full list of configurations can be found on docling-serve's
          [Configuration](https://github.com/docling-project/docling-serve/blob/main/docs/configuration.md) page.
        displayName: Docling Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - document
        - processing
        links:
        - name: Docling Operator
          url: https://docling-project.github.io/docling/
        maintainers:
        - email: docling-technical-discuss@lists.lfaidata.foundation
          name: Docling Maintainers
        maturity: alpha
        provider:
          name: Docling Maintainers
          url: https://docling-project.github.io/docling/
        relatedImages:
        - quay.io/docling-project/docling-operator@sha256:200d8db51874c4b1e3d20d8d1423e9a4807843fbb29f514cd5d8ba872e47f4af
        version: 0.0.1
      entries:
      - name: docling-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: docling-operator
    provider:
      name: Docling Maintainers
      url: https://docling-project.github.io/docling/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubemq.io
      provider-url: ""
    name: kubemq-operator-marketplace
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubemq-operator-marketplace.1.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
              "apiVersion": "core.k8s.kubemq.io/v1beta1",
              "kind": "KubemqCluster",
              "metadata": {
                "name": "kubemq-cluster"
              },
              "spec": {
                "key": "<Your-KubeMQ-License-Key>",
                "replicas": 3
              }
            },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqConnector",
                "metadata": {
                  "name": "kubemq-bridges"
                },
                "spec": {
                  "type": "bridges",
                  "config": "bindings: null"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.connect.redhat.com/kubemq/kubemq-operator@sha256:3a9c87e462eb1caa81925dee72407f73fca41f19a3f9bcaf4e76e569db1dfbab
          createdAt: 2020-03-11 12:00:00+00:00
          description: Kubernetes Messaging Platform
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          support: support@kubemq.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kubemq Cluster
            displayName: Kubemq Cluster
            kind: KubemqCluster
            name: kubemqclusters.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Connector
            displayName: Kubemq Connector
            kind: KubemqConnector
            name: kubemqconnectors.core.k8s.kubemq.io
            version: v1beta1
        description: |2-

          ## Kubernetes Messaging Platform
          A message-based architecture for microservices providing flexible connectivity for hybrid environments: multi-cloud, on-premises, and at the edge. Powered by enterprise-grade message broker and message queue with ready to use connectors, bridges, and control center.

          Messaging platform components:
            - **KubeMQ Cluster** - Enterprise-grade message broker and message queue, scalable, high available, and secured.
            - **KubeMQ Targets Connector** - connects a KubeMQ cluster with external systems and cloud services.
            - **KubeMQ Bridges Connector** - bridge, replicate, aggregate, and transform messages between KubeMQ clusters.
            - **KubeMQ Sources Connector** - transform external systems and cloud services data to KubeMQ cluster messages.
          \n### Kubernetes Native
          Innovative and modern message queue and message broker in a lightweight container developed to run in Kubernetes, certified in the CNCF landscape, and connect natively to the cloud-native ecosystem.

          ### Ease of use

          Simple deployment in Kubernetes in less than 1 minute. Developer friendly by simple to use SDKs and elimination of the many developers and DevOps-centered challenges to define exchanges, brokers, channels, routes, and predefined topics

          ### All messaging patterns

          KubeMQ is available with all messaging patterns such as Queue, Stream, Pub/Sub, and RPC, Saving the need to maintain multiple messaging platforms for different use cases.

          ### Hybrid Environments

          KubeMQ enables the microservices from multiple environments to communicate and build one hybrid infrastructure solution across clouds, on-prem, and at the edge.

          ### KubeMQ Cluster key features

          - All-batteries included Messaging Queue Broker for the Kubernetes environment
          - Deployed with Operator for full life cycle operation
          - Blazing fast (written in Go), small and lightweight Docker container
          - Asynchronous and Synchronous messaging with support for  `At Most Once Delivery` and `At Least Once Delivery` models
          - Supports durable FIFO based Queue, Publish-Subscribe Events, Publish-Subscribe with Persistence (Events Store), RPC Command and Query messaging patterns
          - Supports gRPC, Rest and WebSocket Transport protocols with TLS support (both RPC and Stream modes)
          - Supports Access control Authorization and Authentication
          - Supports message masticating and smart routing
          - No Message broker configuration needed (i.e., queues, exchanges)
          - .Net, Java, Python, Go, and NodeJS SDKs

          ### KubeMQ Targets Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major services types (databases, cache, messaging, serverless, HTTP, etc.)
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-targets)

          ### KubeMQ Bridges Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Build Any Topology - connects KubeMQ clusters in 1:1, 1:n , n:1, n:n
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-bridges)

          ### KubeMQ Sources Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major messaging services such as Kafka, ActiveMQ, SQS, and more.
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-sources)

          ## Getting Started

          [Start with OpenShift](https://docs.kubemq.io/getting-started/create-cluster/openshift)

          ## Get Enterprise Trial License

          [Register](https://account.kubemq.io/login/register)

          ## Request a Demo

          [See KubeMQ in Action](https://kubemq.io/request-a-demo/)
        displayName: Kubemq Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - messaging
        - queue
        - broker
        - pubsub
        - rpc
        links:
        - name: Learn more about Kubemq
          url: https://kubemq.io/
        - name: Register for free license
          url: https://account.kubemq.io/login/register
        - name: Contact Sales
          url: https://kubemq.io/contact-sales/
        - name: Documentation
          url: https://docs.kubemq.io/
        - name: Get Help
          url: https://kubemq.io/contact-support/
        maintainers:
        - email: support@kubemq.io
          name: kubemq
        maturity: alpha
        provider:
          name: Kubemq.io
        relatedImages:
        - registry.connect.redhat.com/kubemq/kubemq-sources@sha256:5736575b6cae39d129346f83fe2d8ab9cdbd1c1d2a99a66b4b9c01631fb5db21
        - registry.connect.redhat.com/kubemq/kubemq-sources@sha256:8ed531e58c60760888ffb18165883386bda9cc5624645353406c94750f5c3008
        - registry.connect.redhat.com/kubemq/kubemq-operator@sha256:3a9c87e462eb1caa81925dee72407f73fca41f19a3f9bcaf4e76e569db1dfbab
        - registry.connect.redhat.com/kubemq/kubemq-uni@sha256:51154b2a112c5cdc17ee8d338524c02ccd995f9386180b1cc232b7a875a23e73
        - registry.connect.redhat.com/kubemq/kubemq-targets@sha256:2a09e1718a03729a121728e4e3683e7e1f0690789d7f9857d76bebf1cc01d436
        version: 1.8.6
      entries:
      - name: kubemq-operator-marketplace.1.8.6
        version: 1.8.6
      - name: kubemq-operator-marketplace.1.8.5
        version: 1.8.5
      - name: kubemq-operator-marketplace.v1.8.4
        version: 1.8.4
      - name: kubemq-operator-marketplace.v1.8.3
        version: 1.8.3
      name: alpha
    defaultChannel: alpha
    packageName: kubemq-operator-marketplace
    provider:
      name: Kubemq.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: candidate
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: fast
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: stable
    defaultChannel: stable
    packageName: stackgres-certified
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzounstructured-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzounstructured-operator.v2.0.101
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzounstructured.clusters.cambridgesemantics.com/v1",
                "kind": "AnzoUnstructured",
                "metadata": {
                  "name": "au01"
                },
                "spec": {
                  "auWorker": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-w",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-unstructured-worker@sha256:4db61bdf6dc8b42824a06d2f8d8aeecddde765d123b40a63a7bbf87845719fb2",
                                "name": "w",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  },
                  "msLeader": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-ms",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-microservices-leader@sha256:0579fd97e81e436eaff3b58c5395f13aa924940a3d51f0bdea5163d327620286",
                                "name": "ms",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/unstructured-operator@sha256:a47af068a7aa55c12be84c80d263feeed5597ecc434986b4f2fff782b568bb04
          createdAt: "2021-08-02"
          description: Kubernetes Operator for Anzo Unstructured
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzounstructured-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzounstructured-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoUnstructured is the Schema for the anzounstructureds
              API
            displayName: AnzoUnstructured Operator
            kind: AnzoUnstructured
            name: anzounstructureds.anzounstructured.clusters.cambridgesemantics.com
            version: v1
        description: |-
          The Anzo Unstructured Operator provides the way to install and configure an anzo unstructured setup on Red Hat K8S environment.
          Currently, this is possible only through existing Anzo installation.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v2.0.0/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers through [ Cambridge Semantics Customer Center ]( https://customercenter.cambridgesemantics.com/ ).
        displayName: Anzo Unstructured Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Database
        - Big Data
        links:
        - name: Anzo Unstructured Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v2.0.0/README.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/unstructured-operator@sha256:a47af068a7aa55c12be84c80d263feeed5597ecc434986b4f2fff782b568bb04
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-microservices-leader@sha256:0579fd97e81e436eaff3b58c5395f13aa924940a3d51f0bdea5163d327620286
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-unstructured-worker@sha256:4db61bdf6dc8b42824a06d2f8d8aeecddde765d123b40a63a7bbf87845719fb2
        version: 2.0.0
      entries:
      - name: anzounstructured-operator.v2.0.101
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anzounstructured-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudwatchlogs-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudwatchlogs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudwatchlogs.services.k8s.aws/v1alpha1",
                "kind": "LogGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudwatchlogs-controller:1.1.1
          createdAt: "2025-09-19T20:51:16Z"
          description: AWS CloudWatch Logs controller is a service controller for
            managing CloudWatch Logs resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LogGroup represents the state of an AWS cloudwatchlogs LogGroup
              resource.
            displayName: LogGroup
            kind: LogGroup
            name: loggroups.cloudwatchlogs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudWatch Logs resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudWatch Logs**

          You can use Amazon CloudWatch Logs to monitor, store, and access your log files from Amazon Elastic Compute Cloud (Amazon EC2) instances, AWS CloudTrail, Route 53, and other sources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudWatch Logs
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudwatchlogs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudWatch Logs Developer Resources
          url: https://aws.amazon.com/cloudwatch/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudwatch maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudwatchlogs-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-cloudwatchlogs-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudwatchlogs-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudwatchlogs-controller.v1.0.14
        version: 1.0.14
      - name: ack-cloudwatchlogs-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudwatchlogs-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudwatchlogs-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudwatchlogs-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudwatchlogs-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudwatchlogs-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudwatchlogs-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudwatchlogs-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudwatchlogs-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudwatchlogs-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudwatchlogs-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudwatchlogs-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudwatchlogs-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudwatchlogs-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudwatchlogs-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudwatchlogs-controller.v0.0.7
        version: 0.0.7
      - name: ack-cloudwatchlogs-controller.v0.0.4
        version: 0.0.4
      - name: ack-cloudwatchlogs-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudwatchlogs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Palo Alto Networks
      provider-url: ""
    name: pcc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pcc-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "name": "pcc-console",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "ConsoleDefender",
                "metadata": {
                  "name": "pcc-consoledefender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "defenderConfig": {
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Defender",
                "metadata": {
                  "name": "pcc-defender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "defenderConfig": {
                    "clusterAddress": "twistlock-console.example.com",
                    "consoleAddress": "https://twistlock-console.example.com:8083",
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: quay.io/prismacloud/pcc-operator@sha256:b8fcfbd6c51286c874e00db1bd35523386cec406fa4050ef44c0a887730cf9b8
          createdAt: "2021-11-15"
          description: Deploy Prisma Cloud Compute for cloud-native security in your
            clusters
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator
          support: pcc-integrations@paloaltonetworks.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Install the Prisma Cloud Compute Console and Defenders
            displayName: Console and Defenders
            kind: ConsoleDefender
            name: consoledefenders.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install the Prisma Cloud Compute Console
            displayName: Console
            kind: Console
            name: consoles.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install Prisma Cloud Compute Defenders
            displayName: Defenders
            kind: Defender
            name: defenders.pcc.paloaltonetworks.com
            version: v1alpha1
        description: |
          ## Features
          - Deploy Console
              - Create initial local administrator user
              - Add license
          - Deploy Defenders
          - Support for offline deployments
          - Upgrade Console
          - Upgrade Defenders

          ## Installation
          See the [Prisma Cloud Compute Operator documentation](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator) for installation guides.

          ## Support
          This operator is available "as is," and should be seen as community-supported; however, Palo Alto Networks will continue contributing whenever possible.
          Palo Alto Networks does not provide technical support or help with using or troubleshooting the operator through normal support channels.
          We encourage you to open GitHub [issues](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/issues) and [pull requests](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/pulls) to track bugs and feature requests.

          ## Other links
          [Prisma Cloud Compute product documentation](https://docs.paloaltonetworks.com/prisma/prisma-cloud/prisma-cloud-admin-compute.html)

          [Prisma Cloud Compute API documentation](https://prisma.pan.dev/api/cloud/cwpp)
        displayName: Prisma Cloud Compute Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - palo
        - palo alto
        - palo alto networks
        - paloalto
        - twistlock
        - compute
        - prisma
        - prisma cloud
        - prisma cloud compute
        - security
        - image
        - images
        - container
        - containers
        - docker
        - podman
        - vulnerability
        - vulnerabilities
        - compliance
        - runtime
        - scan
        - scanning
        - monitor
        - monitoring
        - alert
        - alerting
        links:
        - name: Prisma Cloud
          url: https://www.paloaltonetworks.com/prisma/cloud
        maintainers:
        - email: pcc-integrations@paloaltonetworks.com
          name: Wyatt Gill
        maturity: alpha
        provider:
          name: Palo Alto Networks
        relatedImages:
        - quay.io/prismacloud/pcc-operator@sha256:b8fcfbd6c51286c874e00db1bd35523386cec406fa4050ef44c0a887730cf9b8
        version: 0.2.0
      entries:
      - name: pcc-operator.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: pcc-operator
    provider:
      name: Palo Alto Networks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: koku-metrics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: koku-metrics-operator.v4.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
          createdAt: "2025-09-25T19:04:35Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: koku-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Cost Management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Koku Metrics Operator
          ## Introduction
          The `koku-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Koku Metrics Operator (`koku-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created KokuMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The KokuMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.2.0:
          * Enabled gathering and reporting NVIDIA GPU metrics for resource optimization.
          * (Bugfix) Fixed custom CA certificate validation and configuration issues.
          * (Bugfix) Fixed issue where ROS queries were missing workload metadata for namespaces labeled only with the legacy `insights_cost_management_optimizations` label.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **Progress towards FIPS 140 Compliance:** The operator is using the Go Cryptographic Module v1.0.0 to progress toward FIPS 140 compliance. Although this module is not validated at the time of this release, the operator aims to meet stricter security standards when the module does successfully achieve FIPS validation.
          * **API Name Change:** The Custom Resource Definition (CRD) for configuring the upstream `Koku Metrics Operator` was renamed from `KokuMetricsConfig` to `CostManagementMetricsConfig`.

          ### NOTE:
            * The API name change impacts only users of the upstream (community) Koku Metrics Operator.
            * The downstream (Red Hat-supported) Cost Management Metrics Operator is not impacted by this change and users should continue using their existing configurations.

          **Important Upgrade Instructions:**

            If you are upgrading the upstream Koku Metrics Operator to version 4.0.0 or higher, you must manually migrate your configuration. The operator will no longer recognize existing `KokuMetricsConfig` resources.

            To successfully upgrade and retain your operator's configuration, complete the following steps. You can use the provided `oc` commands replacing the angle brackets (`< >`) with your specific values:

            1. Backup existing configuration (recommended):

               Before you upgrade, retrieve the details of your current `KokuMetricsConfig` instance to help you recreate it with the new API name.

               ```
               oc get kokumetricsconfig -n koku-metrics-operator <your-config-name> -o yaml > koku-metrics-config-backup.yaml
               ```

            2. Delete the previous `KokuMetricsConfig` instance (recommended):

                After the operator is upgraded to `version 4.0.0`, delete any existing `KokuMetricsConfig` resources.

                ```
                oc delete kokumetricsconfig -n koku-metrics-operator <your-config-name>
                ```

            3. Create the new `CostManagementMetricsConfig` instance:

                Using the information from your backup (or a new configuration), create a new `CostManagementMetricsConfig` instance. Adjust the spec section based on your specific needs. For more information about configurations, [refer to the configurable parameters](#configurable-parameters).

            4. Apply your new configuration:

                ```
                oc apply -f <your-costmanagementmetricsconfig-file>.yaml -n koku-metrics-operator
                ```

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created KokuMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the KokuMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `KokuMetricsConfig`, creating a new `KokuMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `koku-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: koku-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the KokuMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the KokuMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon KokuMetricsConfig creation** (This parameter will start to appear in KokuMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the KokuMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the koku-metrics-operator
          **Note** There are separate instructions for configuring the `koku-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `koku-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the koku-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `koku-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `koku-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the koku-metrics-operator for a restricted network
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `koku-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: koku-metrics-operator
                labels:
                  app: koku-metrics-operator
              spec:
                volumes:
                - name: koku-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: koku-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: koku-metrics-operator-reports
                    mountPath: /tmp/koku-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/koku-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/koku-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `koku-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the KokuMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Koku Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Koku Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
        version: 4.2.0
      entries:
      - name: koku-metrics-operator.v4.2.0
        version: 4.2.0
      - name: koku-metrics-operator.v4.1.0
        version: 4.1.0
      - name: koku-metrics-operator.v4.0.0
        version: 4.0.0
      - name: koku-metrics-operator.v3.3.2
        version: 3.3.2
      - name: koku-metrics-operator.v3.3.1
        version: 3.3.1
      - name: koku-metrics-operator.v3.3.0
        version: 3.3.0
      - name: koku-metrics-operator.v3.2.1
        version: 3.2.1
      - name: koku-metrics-operator.v3.2.0
        version: 3.2.0
      - name: koku-metrics-operator.v3.1.0
        version: 3.1.0
      - name: koku-metrics-operator.v3.0.1
        version: 3.0.1
      - name: koku-metrics-operator.v3.0.0
        version: 3.0.0
      - name: koku-metrics-operator.v2.0.0
        version: 2.0.0
      - name: koku-metrics-operator.v1.2.0
        version: 1.2.0
      - name: koku-metrics-operator.v1.1.9
        version: 1.1.9
      - name: koku-metrics-operator.v1.1.8
        version: 1.1.8
      - name: koku-metrics-operator.v1.1.7
        version: 1.1.7
      - name: koku-metrics-operator.v1.1.6
        version: 1.1.6
      - name: koku-metrics-operator.v1.1.5
        version: 1.1.5
      - name: koku-metrics-operator.v1.1.4
        version: 1.1.4
      - name: koku-metrics-operator.v1.1.3
        version: 1.1.3
      - name: koku-metrics-operator.v1.1.2
        version: 1.1.2
      - name: koku-metrics-operator.v1.1.1
        version: 1.1.1
      - name: koku-metrics-operator.v0.9.8
        version: 0.9.8
      - name: koku-metrics-operator.v0.9.7
        version: 0.9.7
      - name: koku-metrics-operator.v0.9.6
        version: 0.9.6
      - name: koku-metrics-operator.v0.9.5
        version: 0.9.5
      - name: koku-metrics-operator.v0.9.4
        version: 0.9.4
      - name: koku-metrics-operator.v0.9.3
        version: 0.9.3
      - name: koku-metrics-operator.v0.9.2
        version: 0.9.2
      - name: koku-metrics-operator.v0.9.1
        version: 0.9.1
      - name: koku-metrics-operator.v0.9.0
        version: 0.9.0
      name: alpha
    - currentCSV: koku-metrics-operator.v4.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
          createdAt: "2025-09-25T19:04:35Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: koku-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Cost Management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Koku Metrics Operator
          ## Introduction
          The `koku-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Koku Metrics Operator (`koku-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created KokuMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The KokuMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.2.0:
          * Enabled gathering and reporting NVIDIA GPU metrics for resource optimization.
          * (Bugfix) Fixed custom CA certificate validation and configuration issues.
          * (Bugfix) Fixed issue where ROS queries were missing workload metadata for namespaces labeled only with the legacy `insights_cost_management_optimizations` label.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **Progress towards FIPS 140 Compliance:** The operator is using the Go Cryptographic Module v1.0.0 to progress toward FIPS 140 compliance. Although this module is not validated at the time of this release, the operator aims to meet stricter security standards when the module does successfully achieve FIPS validation.
          * **API Name Change:** The Custom Resource Definition (CRD) for configuring the upstream `Koku Metrics Operator` was renamed from `KokuMetricsConfig` to `CostManagementMetricsConfig`.

          ### NOTE:
            * The API name change impacts only users of the upstream (community) Koku Metrics Operator.
            * The downstream (Red Hat-supported) Cost Management Metrics Operator is not impacted by this change and users should continue using their existing configurations.

          **Important Upgrade Instructions:**

            If you are upgrading the upstream Koku Metrics Operator to version 4.0.0 or higher, you must manually migrate your configuration. The operator will no longer recognize existing `KokuMetricsConfig` resources.

            To successfully upgrade and retain your operator's configuration, complete the following steps. You can use the provided `oc` commands replacing the angle brackets (`< >`) with your specific values:

            1. Backup existing configuration (recommended):

               Before you upgrade, retrieve the details of your current `KokuMetricsConfig` instance to help you recreate it with the new API name.

               ```
               oc get kokumetricsconfig -n koku-metrics-operator <your-config-name> -o yaml > koku-metrics-config-backup.yaml
               ```

            2. Delete the previous `KokuMetricsConfig` instance (recommended):

                After the operator is upgraded to `version 4.0.0`, delete any existing `KokuMetricsConfig` resources.

                ```
                oc delete kokumetricsconfig -n koku-metrics-operator <your-config-name>
                ```

            3. Create the new `CostManagementMetricsConfig` instance:

                Using the information from your backup (or a new configuration), create a new `CostManagementMetricsConfig` instance. Adjust the spec section based on your specific needs. For more information about configurations, [refer to the configurable parameters](#configurable-parameters).

            4. Apply your new configuration:

                ```
                oc apply -f <your-costmanagementmetricsconfig-file>.yaml -n koku-metrics-operator
                ```

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created KokuMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the KokuMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `KokuMetricsConfig`, creating a new `KokuMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `koku-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: koku-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the KokuMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the KokuMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon KokuMetricsConfig creation** (This parameter will start to appear in KokuMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the KokuMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the koku-metrics-operator
          **Note** There are separate instructions for configuring the `koku-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `koku-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the koku-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `koku-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `koku-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the koku-metrics-operator for a restricted network
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `koku-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: koku-metrics-operator
                labels:
                  app: koku-metrics-operator
              spec:
                volumes:
                - name: koku-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: koku-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: koku-metrics-operator-reports
                    mountPath: /tmp/koku-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/koku-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/koku-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `koku-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the KokuMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Koku Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Koku Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
        version: 4.2.0
      entries:
      - name: koku-metrics-operator.v4.2.0
        version: 4.2.0
      - name: koku-metrics-operator.v4.1.0
        version: 4.1.0
      - name: koku-metrics-operator.v4.0.0
        version: 4.0.0
      - name: koku-metrics-operator.v3.3.2
        version: 3.3.2
      - name: koku-metrics-operator.v3.3.1
        version: 3.3.1
      - name: koku-metrics-operator.v3.3.0
        version: 3.3.0
      - name: koku-metrics-operator.v3.2.1
        version: 3.2.1
      - name: koku-metrics-operator.v3.2.0
        version: 3.2.0
      - name: koku-metrics-operator.v3.1.0
        version: 3.1.0
      - name: koku-metrics-operator.v3.0.1
        version: 3.0.1
      - name: koku-metrics-operator.v3.0.0
        version: 3.0.0
      - name: koku-metrics-operator.v2.0.0
        version: 2.0.0
      - name: koku-metrics-operator.v1.2.0
        version: 1.2.0
      - name: koku-metrics-operator.v1.1.9
        version: 1.1.9
      - name: koku-metrics-operator.v1.1.8
        version: 1.1.8
      - name: koku-metrics-operator.v1.1.7
        version: 1.1.7
      - name: koku-metrics-operator.v1.1.6
        version: 1.1.6
      - name: koku-metrics-operator.v1.1.5
        version: 1.1.5
      - name: koku-metrics-operator.v1.1.4
        version: 1.1.4
      - name: koku-metrics-operator.v1.1.3
        version: 1.1.3
      - name: koku-metrics-operator.v1.1.2
        version: 1.1.2
      - name: koku-metrics-operator.v1.1.1
        version: 1.1.1
      - name: koku-metrics-operator.v0.9.8
        version: 0.9.8
      - name: koku-metrics-operator.v0.9.7
        version: 0.9.7
      - name: koku-metrics-operator.v0.9.6
        version: 0.9.6
      - name: koku-metrics-operator.v0.9.5
        version: 0.9.5
      - name: koku-metrics-operator.v0.9.4
        version: 0.9.4
      - name: koku-metrics-operator.v0.9.3
        version: 0.9.3
      - name: koku-metrics-operator.v0.9.2
        version: 0.9.2
      - name: koku-metrics-operator.v0.9.1
        version: 0.9.1
      - name: koku-metrics-operator.v0.9.0
        version: 0.9.0
      name: beta
    defaultChannel: beta
    packageName: koku-metrics-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Lenovo (Shanghai) Computer Technology Co., Ltd.
      provider-url: ""
    name: lenovo-locd-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lenovo-locd-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "locd.lenovo.com/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample",
                  "namespace": "locd-operator-system"
                },
                "spec": {
                  "affinity": {},
                  "argocd": {
                    "password": "xxxxxxxxxxx",
                    "server": "argocd.svc.cluster.local",
                    "user": "admin"
                  },
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "core": {
                    "image": "registry.connect.redhat.com/loc-devops/core@sha256:f771217b3f7849e1b63cc6abb93b8b0fc9636fd4a201e31807a15f773138c609",
                    "pullPolicy": "Always"
                  },
                  "fullnameOverride": "",
                  "global": {
                    "managementNamespace": "loc-devops",
                    "serviceEndpoint": "loc-devops.domain",
                    "serviceaccount": {
                      "annotations": {
                        "managedBy": "LOC DevOps"
                      },
                      "create": true,
                      "name": "loc-devops-admin"
                    }
                  },
                  "gui": {
                    "image": "registry.connect.redhat.com/loc-devops/gui@sha256:65142a36095188e879467038936b7b6b6ee39dcea6d42a437fa7066fb2bc0f44",
                    "pullPolicy": "Always"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "className": "nginx",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "loc-devops.domain",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "Prefix",
                            "servicePort": 8080
                          },
                          {
                            "path": "/api",
                            "pathType": "Prefix",
                            "servicePort": 8000
                          },
                          {
                            "path": "/swagger",
                            "pathType": "Prefix",
                            "servicePort": 8000
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "replicaCount": 1,
                  "resources": {},
                  "securityContext": {},
                  "service": {
                    "port": {
                      "core": 8000,
                      "gui": 8080
                    },
                    "type": "ClusterIP"
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Cluster
            name: clusters.locd.lenovo.com
            version: v1alpha1
        description: This is operator used to manage components of Lenovo Open Cloud
          DevOps
        displayName: lenovo-locd-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Application management
        links:
        - name: LOCD Operator
          url: https://locd-operator.domain
        maintainers:
        - email: xujt5@lenovo.com
          name: jintaoxu
        maturity: alpha
        provider:
          name: Lenovo (Shanghai) Computer Technology Co., Ltd.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b5786bbbef725badf3dfcc2c2c7a86ead5ebb584c978c47aae8b9a62e241b80d
        - registry.connect.redhat.com/loc-devops/operator@sha256:4c17a3688ae3098e4b5c2c0530f49b1da8dedcc6fc1d3cd848513a1f189812a7
        - registry.connect.redhat.com/loc-devops/core@sha256:f771217b3f7849e1b63cc6abb93b8b0fc9636fd4a201e31807a15f773138c609
        - registry.connect.redhat.com/loc-devops/gui@sha256:65142a36095188e879467038936b7b6b6ee39dcea6d42a437fa7066fb2bc0f44
        version: 0.1.0
      entries:
      - name: lenovo-locd-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: lenovo-locd-operator
    provider:
      name: Lenovo (Shanghai) Computer Technology Co., Ltd.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtr-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtr-operator.1.2.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
             {
               "apiVersion": "windup.jboss.org/v1",
               "kind": "Windup",
               "metadata": {
                 "namespace": "mtr",
                 "name": "mtr",
                 "labels": {
                   "application": "windup"
                 }
               },
               "spec": {}
             }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mtr/mtr-rhel8-operator@sha256:4df7c4ad353339e8b958ffc651e811667767c8a0dcdd61f89352c926da52a513
          createdAt: "2024-09-03T17:22:23Z"
          description: Migration Toolkit for Runtimes is an analysis tool that supports
            the modernization and migration of Java applications.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          imagePullPolicy: Always
          operatorframework.io/suggested-namespace: mtr
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/windup/windup-operator/
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "Create an MTR Web Console application.  \n\nTo access the
              MTR Web Console application from the \"Developer\" perspective, please
              go to Topology and click on the \"Open URL\" icon on the web-console
              pod.  \n\nAlternatively from the \"Administrator\" perspective, please
              go to Networking->Routes and click on the Location hyperlink.  \n\nIn
              order to connect with default login credentials, please use \"migration\"
              as username and \"password\" as password.  \n\nKnown issue  \nIf you
              want to customize the MTR Web Console instance's parameters and you
              can not see any in the `Form view`, please switch to the `YAML view`
              and change them as needed.\n"
            displayName: Migration Toolkit for Runtimes
            kind: Windup
            name: windups.windup.jboss.org
            version: v1
        description: "The Migration Toolkit for Runtimes (MTR) is a web console application
          that supports large-scale Java application modernization and migration projects
          across a broad range of transformations and use cases.  \n\nIt analyzes
          application code, supports effort estimation, accelerates code migration,
          and enables users to move applications to containers.  \n\nFor more information
          please refer to the [https://developers.redhat.com/products/mtr/overview](https://developers.redhat.com/products/mtr/overview)
          page.\n"
        displayName: Migration Toolkit for Runtimes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - security
        - alerting
        - metric
        - troubleshooting
        - run-time
        - migration
        - modernization
        links:
        - name: Website
          url: https://developers.redhat.com/products/mtr/overview
        - name: GitHub
          url: https://github.com/windup/windup-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_runtimes
        maintainers:
        - email: migrate@redhat.com
          name: Migration Toolkit for Runtimes
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mtr/mtr-rhel8-operator@sha256:4df7c4ad353339e8b958ffc651e811667767c8a0dcdd61f89352c926da52a513
        - registry.redhat.io/mtr/mtr-web-container-rhel8@sha256:cb88a5453b506ca70563537ec6db2f44ff8cb4050c9652ec86eb5128aa313003
        - registry.redhat.io/mtr/mtr-web-executor-container-rhel8@sha256:2f0622774367f33efc321926196681322ab1535144e6a0785608c8050fc59f2e
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        version: 1.2.7
      entries:
      - name: mtr-operator.1.2.7
        version: 1.2.7
      - name: mtr-operator.1.2.6
        version: 1.2.6
      - name: mtr-operator.1.2.5
        version: 1.2.5
      - name: mtr-operator.1.2.4
        version: 1.2.4
      - name: mtr-operator.1.2.3
        version: 1.2.3
      - name: mtr-operator.1.2.2
        version: 1.2.2
      - name: mtr-operator.1.2.1
        version: 1.2.1
      - name: mtr-operator.1.2.0
        version: 1.2.0
      - name: mtr-operator.1.1.1
        version: 1.1.1
      - name: mtr-operator.1.1.0
        version: 1.1.0
      - name: mtr-operator.1.0.3
        version: 1.0.3
      - name: mtr-operator.1.0.2
        version: 1.0.2
      - name: mtr-operator.1.0.1
        version: 1.0.1
      - name: mtr-operator.1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: mtr-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://github.com/che-incubator/kubernetes-image-puller-operator
    name: kubernetes-imagepuller-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubernetes-imagepuller-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "che.eclipse.org/v1alpha1",
                "kind": "KubernetesImagePuller",
                "metadata": {
                  "name": "image-puller"
                },
                "spec": {
                  "daemonsetName": "k8s-image-puller",
                  "images": "java11-maven=quay.io/eclipse/che-java11-maven:next;che-theia=quay.io/eclipse/che-theia:next"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/eclipse/kubernetes-image-puller-operator:1.1.0
          createdAt: "2024-01-15T16:23:04Z"
          description: Create and manage kubernetes-image-puller instances.
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/che-incubator/kubernetes-image-puller-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubernetesImagePuller is the Schema for the kubernetesimagepullers
              API
            displayName: Kubernetes Image Puller
            kind: KubernetesImagePuller
            name: kubernetesimagepullers.che.eclipse.org
            version: v1alpha1
        description: "## About the managed application\n\nCreate a `KubernetesImagePuller`
          custom resource to automatically configure and run an instance of the [kubernetes-image-puller.](https://github.com/che-incubator/kubernetes-image-puller)\n##
          About this Operator\n\nThe `KubernetesImagePuller` custom resource understands
          the following fields in the `spec`:\n\n1. `configMapName` - The name of
          the `ConfigMap` to create\t\n2. `daemonsetName` - The name of the `DaemonSet`
          to be create.\n3. `deploymentName` - The name of the `kubernetes-image-puller`
          `Deployment` to create.\n4. `images` - A list of key-value pairs separated
          by semicolons of images to pull.  For example: `java=quay.io/eclipse/che-java8-maven:latest;theia=quay.io/eclipse/che-theia:next`\n5.
          `cachingIntervalHours` - The amount of time, in hours, between `DaemonSet`
          health checks.\n6. `cachingMemoryRequest` - The memory request for each
          cached image when the puller is running.\n7. `cachingMemoryLimit` - The
          memory limit for each cached image when the puller is running.\n8. `nodeSelector`
          - Node selector applied to pods created by the `DaemonSet`.\n## Prerequisites
          for enabling this Operator\n\nThe operator requires an existing namespace
          to be installed in.\nCheck the [README](https://github.com/che-incubator/kubernetes-image-puller-operator#installing-the-operator)
          for more details.\n"
        displayName: Kubernetes Image Puller Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - che
        - kubernetes-image-puller
        links:
        - name: Repository
          url: https://github.com/che-incubator/kubernetes-image-puller-operator
        maintainers:
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        provider:
          name: Red Hat, Inc.
          url: https://github.com/che-incubator/kubernetes-image-puller-operator
        relatedImages:
        - quay.io/eclipse/kubernetes-image-puller-operator:1.1.0
        - quay.io/brancz/kube-rbac-proxy:v0.11.0
        version: 1.1.0
      entries:
      - name: kubernetes-imagepuller-operator.v1.1.0
        version: 1.1.0
      - name: kubernetes-imagepuller-operator.v1.0.6
        version: 1.0.6
      - name: kubernetes-imagepuller-operator.v1.0.5
        version: 1.0.5
      - name: kubernetes-imagepuller-operator.v1.0.4
        version: 1.0.4
      - name: kubernetes-imagepuller-operator.v1.0.3
        version: 1.0.3
      - name: kubernetes-imagepuller-operator.v1.0.2
        version: 1.0.2
      - name: kubernetes-imagepuller-operator.v1.0.1
        version: 1.0.1
      - name: kubernetes-imagepuller-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: kubernetes-imagepuller-operator
    provider:
      name: Red Hat, Inc.
      url: https://github.com/che-incubator/kubernetes-image-puller-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Silicom Ltd.
      provider-url: http://www.silicom-usa.com
    name: silicom-sts-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: silicom-sts-operator.v0.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "namespace": "openshift-operators",
                  "nodeSelector": {
                    "sts.silicom.com/config": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1"
                },
                "spec": {
                  "aprLevel": 0,
                  "domainNumber": 24,
                  "esmcMode": 1,
                  "forwardable": 0,
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-TBC.8275.1",
                  "mode10MHz": 2,
                  "modePPS": 2,
                  "namespace": "openshift-operators",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  },
                  "phyLedsCtl": 0,
                  "priority2": 128,
                  "src10MHz": 1,
                  "srcPPS": 1,
                  "syncOption": 1,
                  "synceCpu": 10,
                  "synceRecClkPort": 0,
                  "tracePtpMsg": -1,
                  "twoStep": 0
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "worker2"
                },
                "spec": null,
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "images": null,
                  "sro": null
                }
              }
            ]
          capabilities: Basic Install
          categories: Drivers and plugins
          certified: "false"
          containerImage: quay.io/silicom/sts-operator:0.0.6
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicomdk/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicomDK/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/silicomDK/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - quay.io/silicom/tsync_extts@sha256:be6a5950a5fdc6fdf8015767901856bf173d01bdc803be7141e4f9795ab1478a
        - quay.io/silicom/phc2sys@sha256:0f3b63a3059b7d6158a916814281097f0d418b99a5fac9e8e5fd9bae58fb0b74
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/silicom/sts-operator:0.0.6
        - quay.io/silicom/tsyncd@sha256:78be78ec8939f025f9a8b899693ab77b3990116d1c7ffb2115c620b7c3382d9d
        - quay.io/silicom/gpsd@sha256:649e8a6c524920e1c564d8f797b8d64235b2527aa923f94bc9c7a0b7aba8ae7f
        - quay.io/silicom/grpc-tsyncd@sha256:aac7233423439e352d8eb39cef63e479c30c2f80b063c84ffc8797cc5a045117
        - quay.io/silicom/sts-plugin@sha256:44339fd400d46d9dddb02c2da47c6f9f96ad2605a5c6de8a7dce2f8c3da3bb57
        - quay.io/silicom/ice-driver-src@sha256:ca728f648769d31501cdedda92554bcfa890754405b0b06d41629cea47707da8
        version: 0.0.6
      entries:
      - name: silicom-sts-operator.v0.0.6
        version: 0.0.6
      - name: silicom-sts-operator.v0.0.5
        version: 0.0.5
      - name: silicom-sts-operator.v0.0.4
        version: 0.0.4
      - name: silicom-sts-operator.v0.0.3
        version: 0.0.3
      - name: silicom-sts-operator.v0.0.2
        version: 0.0.2
      - name: silicom-sts-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: silicom-sts-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "sts-silicom"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "namespace": "sts-silicom",
                  "nodeSelector": {
                    "sts.silicom.com/config": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1"
                },
                "spec": {
                  "aprLevel": 0,
                  "domainNumber": 24,
                  "esmcMode": 1,
                  "forwardable": 0,
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-TBC.8275.1",
                  "mode10MHz": 2,
                  "modePPS": 2,
                  "namespace": "sts-silicom",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  },
                  "phyLedsCtl": 0,
                  "priority2": 128,
                  "src10MHz": 1,
                  "srcPPS": 1,
                  "syncOption": 1,
                  "synceCpu": 10,
                  "synceRecClkPort": 0,
                  "tracePtpMsg": -1,
                  "twoStep": 0
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "worker2"
                },
                "spec": null,
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config"
                },
                "spec": {
                  "gpsSvcPort": 2947,
                  "grpcSvcPort": 50051,
                  "images": {
                    "gpsd": "quay.io/silicom/gpsd:3.23.1",
                    "grpcTsyncd": "quay.io/silicom/grpc-tsyncd:2.1.0.0",
                    "phcs2Sys": "quay.io/silicom/phcs2sys:3.1.1",
                    "stsPlugin": "quay.io/silicom/sts-plugin:0.0.2",
                    "ts2Phcs": "quay.io/silicom/ts2phcs:1.0.0",
                    "tsyncd": "quay.io/silicom/tsyncd:2.1.0.0"
                  },
                  "namespace": "sts-silicom",
                  "sro": {
                    "chart": {
                      "name": "ice-special-resource",
                      "repository": {
                        "name": "ice-special-resource",
                        "url": "http://ice-driver-src"
                      },
                      "version": "0.0.1"
                    },
                    "iceVersion": "1.7.16",
                    "namespace": "openshift-operators",
                    "srcImage": "quay.io/silicom/ice-driver-src:1.7.16",
                    "srcSvcPort": 3000
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Drivers and plugins
          certified: "false"
          containerImage: quay.io/silicom/sts-operator:0.0.2
          createdAt: 2021-11-17T10:21:24 CET
          description: This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of Silicom STS device
            plugins for Kubernetes.
          operators.operatorframework.io.bundle.channels.v1: stable,development
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicomdk/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicomDK/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/silicomDK/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/silicom/sts-operator:0.0.2
        version: 0.0.2
      entries:
      - name: silicom-sts-operator.v0.0.2
        version: 0.0.2
      - name: silicom-sts-operator.v0.0.1
        version: 0.0.1
      name: beta
    defaultChannel: alpha
    packageName: silicom-sts-operator
    provider:
      name: Silicom Ltd.
      url: http://www.silicom-usa.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-logging
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-logging.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Logging",
                "metadata": {
                  "name": "logging-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "elasticsearch": {
                    "storage": "16Gi",
                    "tag": "7.17.6-3747-35"
                  },
                  "environment": "openshift",
                  "fluentbit": {
                    "runOnMaster": false,
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "grafana": {
                    "dashboards": [
                      "wap",
                      "load-core"
                    ],
                    "dataSources": [
                      "wap",
                      "load-core"
                    ]
                  },
                  "image": {
                    "elasticInitImage": "rhel8/toolbox@sha256:a753c24cb8b03066d494f010ad654d23fb0559f14d7121542298da3e0d48340d",
                    "elasticInitRepository": "registry.access.redhat.com",
                    "fbImage": "fluent-bit:1.8.15-3747-20",
                    "fbRepository": "prod.harbor.keysight.digital/loadcore",
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "createClusterRoles": true,
                    "serviceAccountName": "load-core-service-account"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:33:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Logging
            name: loggings.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Logging Operator
        displayName: Keysight WAP Logging Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Logging
          url: https://keysight-wap-logging.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-logging@sha256:f1e3b3ad84cd0f96c92cdaf4609e29257035cb982a761e832dae2095e636c721
        version: 0.1.21
      entries:
      - name: keysight-wap-logging.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-logging
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hazelcast, Inc
      provider-url: ""
    name: hazelcast-platform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hazelcast-platform-operator.v5.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "cache-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "CronHotBackup",
                "metadata": {
                  "name": "cronhotbackup-sample"
                },
                "spec": {
                  "hotBackupTemplate": {
                    "metadata": {
                      "labels": {
                        "cron-hotbackup": "true"
                      }
                    },
                    "spec": {
                      "hazelcastResourceName": "hazelcast"
                    }
                  },
                  "schedule": "*/30 * * * *"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Flow",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hazelcast-platform-operator"
                  },
                  "name": "flow"
                },
                "spec": {
                  "database": {
                    "host": "flow-postgresql",
                    "secretName": "flow-db-secret"
                  },
                  "env": [
                    {
                      "name": "OPTIONS",
                      "value": "--flow.analytics.persistRemoteCallResponses=true --flow.analytics.persistResults=false --flow.config.custom.managementCenterUrl=http://34.140.160.158/mc\n"
                    }
                  ],
                  "externalConnectivity": {
                    "ingress": {
                      "annotations": {
                        "kubernetes.io/ingress.class": "nginx",
                        "nginx.ingress.kubernetes.io/enable-opentelemetry": "true",
                        "nginx.ingress.kubernetes.io/ssl-redirect": "false"
                      },
                      "hostname": "",
                      "ingressClassName": "nginx",
                      "path": "/"
                    }
                  },
                  "imagePullSecrets": [
                    {
                      "name": "hazelcast-cloud-hazelcast-connect-demo-pull-secret"
                    }
                  ],
                  "licenseKeySecretName": "flow-mc-license-key",
                  "repository": "quay.io/hazelcast_cloud/hazelcast-flow-internal",
                  "size": 3,
                  "version": "next"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast"
                },
                "spec": {
                  "clusterSize": 3,
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast-sample"
                },
                "spec": {
                  "clusterSize": 3,
                  "jet": {
                    "bucketConfig": {
                      "bucketURI": "gs://operator-user-code/jetJobs",
                      "secretName": "br-secret-gcp"
                    },
                    "enabled": true,
                    "resourceUploadEnabled": true
                  },
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HazelcastEndpoint",
                "metadata": {
                  "name": "hazelcastendpoint-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "port": 5701,
                  "type": "Discovery"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HotBackup",
                "metadata": {
                  "name": "hot-backup"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-longrun"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-longrun-2.0.0.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-1.0.2.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJobSnapshot",
                "metadata": {
                  "name": "jetjobsnapshot-sample"
                },
                "spec": {
                  "cancelJob": false,
                  "jetJobResourceName": "jet-job"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ManagementCenter",
                "metadata": {
                  "name": "managementcenter"
                },
                "spec": {
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "hazelcastClusters": [
                    {
                      "address": "hazelcast",
                      "name": "dev"
                    }
                  ],
                  "licenseKeySecretName": "hazelcast-license-key",
                  "persistence": {
                    "enabled": true,
                    "size": "10Gi"
                  },
                  "repository": "hazelcast/management-center"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Map",
                "metadata": {
                  "name": "map"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "MultiMap",
                "metadata": {
                  "name": "multimap-sample"
                },
                "spec": {
                  "backupCount": 3,
                  "binary": true,
                  "collectionType": "LIST",
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "queue-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ReplicatedMap",
                "metadata": {
                  "name": "replicatedmap-sample"
                },
                "spec": {
                  "asyncFillup": true,
                  "hazelcastResourceName": "hazelcast",
                  "inMemoryFormat": "OBJECT"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "topic-sample"
                },
                "spec": {
                  "globalOrderingEnabled": true,
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "UserCodeNamespace",
                "metadata": {
                  "name": "ucn-sample"
                },
                "spec": {
                  "bucketConfig": {
                    "bucketURI": "gs://operator-user-code/userCodeNamespace",
                    "secretName": "br-secret-gcp"
                  },
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "VectorCollection",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hazelcast-platform-operator"
                  },
                  "name": "vectorcollection-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "indexes": [
                    {
                      "dimension": 10,
                      "efConstruction": 256,
                      "maxDegree": 32,
                      "metric": "Dot",
                      "useDeduplication": false
                    }
                  ]
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanReplication",
                "metadata": {
                  "name": "wanreplication-sample"
                },
                "spec": {
                  "endpoints": "35.192.33.252",
                  "resources": [
                    {
                      "kind": "Map",
                      "name": "map-sample"
                    },
                    {
                      "kind": "Hazelcast",
                      "name": "hz-sample"
                    }
                  ],
                  "targetClusterName": "dev"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanSync",
                "metadata": {
                  "name": "wansync-sample"
                },
                "spec": {
                  "wanReplicationResourceName": "wan-replication-sample"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: docker.io/hazelcast/hazelcast-platform-operator@sha256:59328a4ea7a4e2110e0750364db792cf266f1e7231e3bb321e3de09d535dab23
          createdAt: "2025-03-13T11:59:59Z"
          description: Install Hazelcast clusters in Kubernetes environments.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hazelcast/hazelcast-platform-operator/
          support: Hazelcast, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.hazelcast.com
            version: v1alpha1
          - description: CronHotBackup is the Schema for the cronhotbackups API
            displayName: Cron Hot Backup
            kind: CronHotBackup
            name: cronhotbackups.hazelcast.com
            version: v1alpha1
          - description: Flow is the Schema for the flows API
            displayName: Flow
            kind: Flow
            name: flows.hazelcast.com
            version: v1alpha1
          - description: HazelcastEndpoint is the Schema for the hazelcastendpoints
              API
            displayName: Hazelcast Endpoint
            kind: HazelcastEndpoint
            name: hazelcastendpoints.hazelcast.com
            version: v1alpha1
          - description: Hazelcast is the Schema for the hazelcasts API
            displayName: Hazelcast
            kind: Hazelcast
            name: hazelcasts.hazelcast.com
            version: v1alpha1
          - description: HotBackup is the Schema for the hot backup API
            displayName: Hot Backup
            kind: HotBackup
            name: hotbackups.hazelcast.com
            version: v1alpha1
          - description: JetJob is the Schema for the jetjobs API
            displayName: Jet Job
            kind: JetJob
            name: jetjobs.hazelcast.com
            version: v1alpha1
          - description: JetJobSnapshot is the Schema for the jetjobsnapshots API
            displayName: Jet Job Snapshot
            kind: JetJobSnapshot
            name: jetjobsnapshots.hazelcast.com
            version: v1alpha1
          - description: ManagementCenter is the Schema for the managementcenters
              API
            displayName: Management Center
            kind: ManagementCenter
            name: managementcenters.hazelcast.com
            version: v1alpha1
          - description: Map is the Schema for the maps API
            displayName: Map
            kind: Map
            name: maps.hazelcast.com
            version: v1alpha1
          - description: MultiMap is the Schema for the multimaps API
            displayName: Multi Map
            kind: MultiMap
            name: multimaps.hazelcast.com
            version: v1alpha1
          - description: Queue is the Schema for the queues API
            displayName: Queue
            kind: Queue
            name: queues.hazelcast.com
            version: v1alpha1
          - description: ReplicatedMap is the Schema for the replicatedmaps API
            displayName: Replicated Map
            kind: ReplicatedMap
            name: replicatedmaps.hazelcast.com
            version: v1alpha1
          - description: Topic is the Schema for the topics API
            displayName: Topic
            kind: Topic
            name: topics.hazelcast.com
            version: v1alpha1
          - description: UserCodeNamespace is the Schema for the usercodenamespaces
              API
            displayName: User Code Namespace
            kind: UserCodeNamespace
            name: usercodenamespaces.hazelcast.com
            version: v1alpha1
          - description: VectorCollection is the Schema for the vectorcollections
              API
            displayName: Vector Collection
            kind: VectorCollection
            name: vectorcollections.hazelcast.com
            version: v1alpha1
          - description: WanReplication is the Schema for the wanreplications API
            displayName: Wan Replication
            kind: WanReplication
            name: wanreplications.hazelcast.com
            version: v1alpha1
          - description: WanSync is the Schema for the wansyncs API
            displayName: Wan Sync
            kind: WanSync
            name: wansyncs.hazelcast.com
            version: v1alpha1
        description: |
          Easily deploy Hazelcast clusters and Management Center into Kubernetes environments and manage their lifecycles.

          ## Before You Start

          If you are planning to create Hazelcast Platform Enterprise clusters, you need to [create a secret](https://docs.hazelcast.com/operator/latest/get-started#step-2-start-the-hazelcast-cluster) for the license. You can request a trial license key from [here](https://trialrequest.hazelcast.com).

          For Hazelcast Platform clusters, you can simply continue.

          ## Documentation

          1. [Get started](https://docs.hazelcast.com/operator/latest/get-started) with the Operator
          2. [Connect to the cluster from outside Kubernetes](https://guides.hazelcast.org/hazelcast-platform-operator-expose-externally)
          3. [Restore a Cluster from Cloud Storage with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-external-backup-restore)
          4. [Replicate Data between Two Hazelcast Clusters with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-wan-replication)

          ## Features

          Hazelcast Platform Operator supports the features below:

          * Custom resource for Hazelcast Platform (Enterprise) and Management Center
          * Observe status of Hazelcast clusters and Management Center instances
          * High Availability Mode configuration to create clusters that are resilient to node and zone failures
          * Support for TLS connections between members using self-signed certificates
          * Scale up and down Hazelcast clusters
          * Expose Hazelcast cluster to external clients ([Smart & Unisocket](https://docs.hazelcast.com/hazelcast/latest/clients/java#java-client-operation-modes))
          * Backup Hazelcast persistence data to cloud storage with the possibility of scheduling it and restoring the data accordingly
          * WAN Replication feature when you need to synchronize multiple Hazelcast clusters, which are connected by WANs
          * Full and Delta WAN Sync support
          * Tiered Storage support for Hazelcast and Map CRs
          * CP Subsystem configuration support for Hazelcast CR
          * User Code Deployment feature, which allows you to deploy custom and domain classes from cloud storages and URLs to Hazelcast members
          * The User Code Namespaces feature allows you to deploy custom and domain classes at runtime without a members restart
          * Support the configuration of advanced networking options
          * Support Multi-namespace configuration
          * ExecutorService and EntryProcessor support
          * Support several data structures like Map, Topic, MultiMap, ReplicatedMap, Queue and Cache which can be created dynamically via specific Custom Resources
          * MapStore, Near Cache and off-heap memory (HD memory and native memory) support for the Map CR
          * Native Memory support for the Cache CR
          * Support Jet configuration and Jet Job submission using the JetJob CR
          * Support for exporting the snapshots of JetJob CRs using JetJobSnapshot CR
          * Support for custom configurations using ConfigMap
        displayName: Hazelcast Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hazelcast
        - keyvalue
        - in-memory
        - database
        - caching
        links:
        - name: Documentation
          url: https://docs.hazelcast.com/operator/latest/
        - name: Source Code
          url: https://github.com/hazelcast/hazelcast-platform-operator
        maintainers:
        - email: cloudnative@hazelcast.com
          name: Hazelcast Cloud Native Team
        maturity: alpha
        minKubeVersion: 1.22.7
        provider:
          name: Hazelcast, Inc
        relatedImages:
        - docker.io/hazelcast/hazelcast-platform-operator@sha256:59328a4ea7a4e2110e0750364db792cf266f1e7231e3bb321e3de09d535dab23
        version: 5.15.0
      entries:
      - name: hazelcast-platform-operator.v5.15.0
        version: 5.15.0
      name: alpha
    defaultChannel: alpha
    packageName: hazelcast-platform-operator
    provider:
      name: Hazelcast, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: tang-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tang-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "daemons.redhat.com/v1alpha1",
                "kind": "TangServer",
                "metadata": {
                  "finalizers": [
                    "finalizer.daemons.tangserver.redhat.com"
                  ],
                  "name": "tangserver",
                  "namespace": "nbde"
                },
                "spec": {
                  "healthScript": "/usr/bin/tangd-health-check",
                  "image": "registry.redhat.io/rhel9/tang",
                  "keyPath": "/var/db/tang",
                  "persistentVolumeClaim": "tangserver-pvc",
                  "replicas": 1,
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:f43c96035f43040b3644e37119a3b35fc433f366d4b636459d1b9b02a05b5d3d
          createdAt: 2023-11-23T12:49:40
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TangServer is the Schema for the tangservers API
            displayName: Tang Server
            kind: TangServer
            name: tangservers.daemons.redhat.com
            version: v1alpha1
        description: NBDE Tang Server operator allows Tang Server deployment on OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nbde
        - tang
        links:
        - name: Tang Operator
          url: https://github.com/latchset/tang-operator
        maintainers:
        - email: sarroutb@redhat.com
          name: Sergio Arroutbi Braojos
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:97cade2c1ee468261aec5400728c8d44de387b459134aec7a4c3b5ec5a335d2c
        - registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:f43c96035f43040b3644e37119a3b35fc433f366d4b636459d1b9b02a05b5d3d
        version: 1.0.0
      entries:
      - name: tang-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: tang-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "daemons.redhat.com/v1alpha1",
                "kind": "TangServer",
                "metadata": {
                  "finalizers": [
                    "finalizer.daemons.tangserver.redhat.com"
                  ],
                  "name": "tangserver",
                  "namespace": "nbde"
                },
                "spec": {
                  "healthScript": "/usr/bin/tangd-health-check",
                  "image": "registry.redhat.io/rhel9/tang",
                  "keyPath": "/var/db/tang",
                  "persistentVolumeClaim": "tangserver-pvc",
                  "replicas": 1,
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:1cb2f794689817282ac570cacb106d75a2751a1926d89d3b83b619e02d869e5e
          createdAt: 2024-12-04T19:01:27
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TangServer is the Schema for the tangservers API
            displayName: Tang Server
            kind: TangServer
            name: tangservers.daemons.redhat.com
            version: v1alpha1
        description: NBDE Tang Server operator allows Tang Server deployment on OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nbde
        - tang
        links:
        - name: Tang Operator
          url: https://github.com/latchset/tang-operator
        maintainers:
        - email: sarroutb@redhat.com
          name: Sergio Arroutbi Braojos
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1172e150fff22c5eeab572f26961f3f53fbf896ee76d08c7503cfe2777c55458
        - registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:1cb2f794689817282ac570cacb106d75a2751a1926d89d3b83b619e02d869e5e
        version: 1.0.3
      entries:
      - name: tang-operator.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: alpha
    packageName: tang-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: www.redhat.com
    name: compliance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: compliance-operator.v0.1.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.17 <0.1.32'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          repository: https://github.com/openshift/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repo
          url: https://github.com/openshift/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/4.7/security/compliance_operator/compliance-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/compliance-rhel8-operator@sha256:ed68a0574ddeaf7d676403cd843f28cb586894f35dbb3ca5e06a5264bd42f030
        - registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c
        - registry.redhat.io/openshift4/compliance-openscap-rhel8@sha256:867c6d4205653d8a954d234cbaf470313009eb983d4f455fb45f49c4a7764422
        version: 0.1.32
      entries:
      - name: compliance-operator.v0.1.32
        version: 0.1.32
      name: "4.7"
    - currentCSV: compliance-operator.v0.1.61
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-node"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-platform"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "description": "Example of a tailoredProfile that extends OCP4 FedRAMP Moderate",
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.17 <0.1.61'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            displayName: Compliance Remediation
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            displayName: Compliance Scan
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            displayName: Compliance Suite
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            displayName: Profile Bundle
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            displayName: Scan Setting Binding
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            displayName: Tailored Profile
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repo
          url: https://github.com/openshift/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/compliance_operator/compliance-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b
        - registry.redhat.io/compliance/openshift-compliance-openscap-rhel8@sha256:83ad3113964ec48c5cf28315d923adb6edb95e45960e2756665d617dcd9b2e47
        - registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:5578e2225365ac1c15e4c189c5c7a8d14fd3fd26189414b36028f7f0879bb1dc
        version: 0.1.61
      entries:
      - name: compliance-operator.v0.1.61
        version: 0.1.61
      name: release-0.1
    - currentCSV: compliance-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-node"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-platform"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "description": "Example of a tailoredProfile that extends OCP4 FedRAMP Moderate",
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          must-gather-image: registry.redhat.io/compliance/openshift-compliance-must-gather-rhel8@sha256:e6468ccc55d72539b7fb980d0bce5c072e85879d4e9db39974d53eaab4d32b25
          olm.skipRange: '>=0.1.17 <1.7.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ComplianceAsCode/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            displayName: Compliance Remediation
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            displayName: Compliance Scan
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            displayName: Compliance Suite
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            displayName: Profile Bundle
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            displayName: Scan Setting Binding
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            displayName: Tailored Profile
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repository
          url: https://github.com/ComplianceAsCode/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/compliance_operator/co-overview.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4
        - registry.redhat.io/compliance/openshift-compliance-openscap-rhel8@sha256:d0f89d82845c6245a18e7db345ac3be92b2863d57f55f63e21652086128920c9
        - registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:8d3257bd6a020df53f8d7a400ae877ae029c78ffcd8ec9e35270ee3ff5c8e509
        - registry.redhat.io/compliance/openshift-compliance-must-gather-rhel8@sha256:e6468ccc55d72539b7fb980d0bce5c072e85879d4e9db39974d53eaab4d32b25
        version: 1.7.1
      entries:
      - name: compliance-operator.v1.7.1
        version: 1.7.1
      - name: compliance-operator.v1.7.0
        version: 1.7.0
      - name: compliance-operator.v1.6.2
        version: 1.6.2
      - name: compliance-operator.v1.6.1
        version: 1.6.1
      - name: compliance-operator.v1.6.0
        version: 1.6.0
      - name: compliance-operator.v1.5.1
        version: 1.5.1
      - name: compliance-operator.v1.5.0
        version: 1.5.0
      - name: compliance-operator.v1.4.1
        version: 1.4.1
      - name: compliance-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: compliance-operator
    provider:
      name: Red Hat Inc.
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: GitLab Inc
      provider-url: https://about.gitlab.com/
    name: gitlab-operator-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "certmanager": {
                        "install": false
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-26T07:16:35Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3281fe12aaae428fd7102332e7c0d7033515042d7830eba3953891ab0c3e3a29
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      - name: gitlab-operator-kubernetes.v2.1.1
        version: 2.1.1
      - name: gitlab-operator-kubernetes.v2.1.0
        version: 2.1.0
      - name: gitlab-operator-kubernetes.v2.0.2
        version: 2.0.2
      - name: gitlab-operator-kubernetes.v2.0.1
        version: 2.0.1
      - name: gitlab-operator-kubernetes.v2.0.0
        version: 2.0.0
      - name: gitlab-operator-kubernetes.v1.12.2
        version: 1.12.2
      - name: gitlab-operator-kubernetes.v1.12.1
        version: 1.12.1
      - name: gitlab-operator-kubernetes.v1.12.0
        version: 1.12.0
      - name: gitlab-operator-kubernetes.v1.11.3
        version: 1.11.3
      - name: gitlab-operator-kubernetes.v1.11.2
        version: 1.11.2
      - name: gitlab-operator-kubernetes.v1.11.1
        version: 1.11.1
      - name: gitlab-operator-kubernetes.v1.11.0
        version: 1.11.0
      - name: gitlab-operator-kubernetes.v1.10.2
        version: 1.10.2
      - name: gitlab-operator-kubernetes.v1.10.1
        version: 1.10.1
      - name: gitlab-operator-kubernetes.v1.10.0
        version: 1.10.0
      - name: gitlab-operator-kubernetes.v1.9.2
        version: 1.9.2
      - name: gitlab-operator-kubernetes.v1.9.1
        version: 1.9.1
      - name: gitlab-operator-kubernetes.v1.9.0
        version: 1.9.0
      - name: gitlab-operator-kubernetes.v1.8.3
        version: 1.8.3
      - name: gitlab-operator-kubernetes.v1.8.2
        version: 1.8.2
      - name: gitlab-operator-kubernetes.v1.8.1
        version: 1.8.1
      - name: gitlab-operator-kubernetes.v1.8.0
        version: 1.8.0
      - name: gitlab-operator-kubernetes.v1.7.2
        version: 1.7.2
      - name: gitlab-operator-kubernetes.v1.7.1
        version: 1.7.1
      - name: gitlab-operator-kubernetes.v1.7.0
        version: 1.7.0
      - name: gitlab-operator-kubernetes.v1.6.1
        version: 1.6.1
      - name: gitlab-operator-kubernetes.v1.6.0
        version: 1.6.0
      - name: gitlab-operator-kubernetes.v1.5.1
        version: 1.5.1
      - name: gitlab-operator-kubernetes.v1.4.2
        version: 1.4.2
      - name: gitlab-operator-kubernetes.v1.4.1
        version: 1.4.1
      - name: gitlab-operator-kubernetes.v1.4.0
        version: 1.4.0
      - name: gitlab-operator-kubernetes.v1.3.3
        version: 1.3.3
      - name: gitlab-operator-kubernetes.v1.3.2
        version: 1.3.2
      - name: gitlab-operator-kubernetes.v1.3.1
        version: 1.3.1
      - name: gitlab-operator-kubernetes.v1.3.0
        version: 1.3.0
      name: stable
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "certmanager": {
                        "install": false
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-26T07:16:35Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3281fe12aaae428fd7102332e7c0d7033515042d7830eba3953891ab0c3e3a29
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      - name: gitlab-operator-kubernetes.v2.1.1
        version: 2.1.1
      - name: gitlab-operator-kubernetes.v2.1.0
        version: 2.1.0
      - name: gitlab-operator-kubernetes.v2.0.2
        version: 2.0.2
      - name: gitlab-operator-kubernetes.v2.0.1
        version: 2.0.1
      - name: gitlab-operator-kubernetes.v2.0.0
        version: 2.0.0
      - name: gitlab-operator-kubernetes.v1.12.2
        version: 1.12.2
      - name: gitlab-operator-kubernetes.v1.12.1
        version: 1.12.1
      - name: gitlab-operator-kubernetes.v1.12.0
        version: 1.12.0
      - name: gitlab-operator-kubernetes.v1.11.3
        version: 1.11.3
      - name: gitlab-operator-kubernetes.v1.11.2
        version: 1.11.2
      - name: gitlab-operator-kubernetes.v1.11.1
        version: 1.11.1
      - name: gitlab-operator-kubernetes.v1.11.0
        version: 1.11.0
      - name: gitlab-operator-kubernetes.v1.10.2
        version: 1.10.2
      - name: gitlab-operator-kubernetes.v1.10.1
        version: 1.10.1
      - name: gitlab-operator-kubernetes.v1.10.0
        version: 1.10.0
      - name: gitlab-operator-kubernetes.v1.9.2
        version: 1.9.2
      - name: gitlab-operator-kubernetes.v1.9.1
        version: 1.9.1
      - name: gitlab-operator-kubernetes.v1.9.0
        version: 1.9.0
      - name: gitlab-operator-kubernetes.v1.8.3
        version: 1.8.3
      - name: gitlab-operator-kubernetes.v1.8.2
        version: 1.8.2
      - name: gitlab-operator-kubernetes.v1.8.1
        version: 1.8.1
      - name: gitlab-operator-kubernetes.v1.8.0
        version: 1.8.0
      - name: gitlab-operator-kubernetes.v1.7.2
        version: 1.7.2
      - name: gitlab-operator-kubernetes.v1.7.1
        version: 1.7.1
      - name: gitlab-operator-kubernetes.v1.7.0
        version: 1.7.0
      - name: gitlab-operator-kubernetes.v1.6.1
        version: 1.6.1
      - name: gitlab-operator-kubernetes.v1.6.0
        version: 1.6.0
      - name: gitlab-operator-kubernetes.v1.5.1
        version: 1.5.1
      - name: gitlab-operator-kubernetes.v1.4.2
        version: 1.4.2
      - name: gitlab-operator-kubernetes.v1.4.1
        version: 1.4.1
      - name: gitlab-operator-kubernetes.v1.4.0
        version: 1.4.0
      - name: gitlab-operator-kubernetes.v1.3.3
        version: 1.3.3
      - name: gitlab-operator-kubernetes.v1.3.2
        version: 1.3.2
      - name: gitlab-operator-kubernetes.v1.3.1
        version: 1.3.1
      - name: gitlab-operator-kubernetes.v1.3.0
        version: 1.3.0
      name: unstable
    defaultChannel: stable
    packageName: gitlab-operator-kubernetes
    provider:
      name: GitLab Inc
      url: https://about.gitlab.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Antline-Soft
      provider-url: ""
    name: cxl-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cxl-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-05-08T08:36:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
        description: memory extend operator
        displayName: extend-community-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - memory
        - extend
        - community
        - operator
        links:
        - name: Cxl Operator
          url: https://cxl-operator.domain
        maintainers:
        - email: tomboy@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        provider:
          name: Antline-Soft
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - rasen708/cxl-operator:latest
        version: 0.0.1
      entries:
      - name: cxl-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cxl-operator
    provider:
      name: Antline-Soft
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: hyperfoil-bundle
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hyperfoil-operator.v0.26.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hyperfoil.io/v1alpha2",
                "kind": "Hyperfoil",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "hyperfoil-operator",
                    "app.kubernetes.io/instance": "hyperfoil-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hyperfoil",
                    "app.kubernetes.io/part-of": "hyperfoil-operator"
                  },
                  "name": "hyperfoil-sample"
                },
                "spec": {
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/hyperfoil/hyperfoil-operator:0.26.0
          createdAt: "2024-06-11T17:39:20Z"
          description: Microservice-oriented distributed benchmark framework.
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Hyperfoil/hyperfoil-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hyperfoil is the Schema for the hyperfoils API
            displayName: Hyperfoil
            kind: Hyperfoil
            name: hyperfoils.hyperfoil.io
            version: v1alpha2
        description: |
          Hyperfoil is a modern benchmark framework oriented promising
          accurate results with more flexibility and distributed architecture.

          See more on [hyperfoil.io](http://hyperfoil.io).

          Hyperfoil is licensed under Apache License 2.0
        displayName: Hyperfoil
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hyperfoil
        - benchmark
        - test
        - load
        - driver
        links:
        - name: Website
          url: http://hyperfoil.io
        - name: GitHub
          url: http://github.com/Hyperfoil
        maintainers:
        - email: rvansa@redhat.com
          name: Radim Vansa
        - email: spederse@redhat.com
          name: Ståle W. Pedersen
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/hyperfoil/hyperfoil-operator:0.26.0
        version: 0.26.0
      entries:
      - name: hyperfoil-operator.v0.26.0
        version: 0.26.0
      - name: hyperfoil-operator.v0.24.2
        version: 0.24.2
      - name: hyperfoil-operator.v0.21.0
        version: 0.21.0
      - name: hyperfoil-operator.v0.15.1
        version: 0.15.1
      - name: hyperfoil-operator.v0.15.0
        version: 0.15.0
      name: alpha
    defaultChannel: alpha
    packageName: hyperfoil-bundle
    provider:
      name: Red Hat, Inc.
kind: List
metadata:
  resourceVersion: ""
apiVersion: v1
items:
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator.v1.6.3
        version: 1.6.3
      - name: dynatrace-operator.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator.v1.4.0
        version: 1.4.0
      - name: dynatrace-operator.v1.3.2
        version: 1.3.2
      - name: dynatrace-operator.v1.3.1
        version: 1.3.1
      - name: dynatrace-operator.v1.3.0
        version: 1.3.0
      - name: dynatrace-operator.v0.6.0
        version: 0.6.0
      - name: dynatrace-operator.v0.5.1
        version: 0.5.1
      - name: dynatrace-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco
      provider-url: ""
    name: aci-containers-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aci-containers-operator.v1.0.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"aci.ctrl/v1alpha1","kind":"AciContainersOperator","metadata":{"name":"acicnioperator","namespace":"aci-containers-system"},"spec":{}}]'
          capabilities: Seamless Upgrades
          categories: Networking
          certified: "false"
          containerImage: quay.io/noiro/aci-containers-operator@sha256:2b2d685e254c206be5371285852a8f95d0318af69cb85b786427969e035c125c
          createdAt: "2025-04-10T12:00:00Z"
          description: The Cisco Application Centric Infrastructure (ACI) Containers
            Operator brings the ACI CNI Plugin and Policy model to Kubernetes clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://github.com/noironetworks/aci-containers/blob/master/README.md
          repository: https://github.com/noironetworks/aci-containers
          support: acisupport@cisco.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aci Containers Operator
            displayName: AciContainersOperator
            kind: AciContainersOperator
            name: acicontainersoperators.aci.ctrl
            version: v1alpha1
        description: |
          ## ACI CNI Plugin

          The Cisco Application Centric Infrastructure (ACI) CNI plugin brings the ACI Networking and Policy model to Kubernetes clusters that reside on-prem or in the cloud. It is fully open source and relies on the Opflex Protocol to program Open vSwitch instances running on the Kubernetes nodes. It provides IP address management, L2/L3 networking, load balancing, and security functions for container workloads.

          The Cisco ACI CNI plugin for the Red Hat OpenShift Container Platform provides a single, programmable network infrastructure that eliminates bottlenecks between OpenShift workloads and the rest of the infrastructure while providing enterprise-grade security and offers flexible micro segmentation possibilities.

          Unified Networking - By extending the integrated overlay of the ACI fabric to OpenShift workloads, Pod and Service endpoints are managed as native ACI network resources at the same level as Bare Metal or Virtual Machines.

          No Egress Router Required - Applications running in OpenShift Pods that need high-bandwidth, low-latency traffic for data transfers external to the cluster suffer the bottleneck imposed by the egress router implementation. Pods can easily connect with other fabric or non-fabric endpoints leveraging the combined hardware and software distributed routing and switching provided by ACI and the ACI CNI plugin.

          Simplified Operations and Enhanced Visibility - Whether policies and networking are defined by the devops or fabric or security admins, the APIC becomes the single source of truth for operations and security governance. Kubernetes workload visibility is provided to Network Operators within their familiar Cisco tools.

          Hardware-assisted Load Balancing - Ingress connections to LoadBalancer-type services using ACI’s Policy Based Redirect technology moves load balancing into the fabric's hardware.

          Granular Security - Security can be implemented by using native NetworkPolicy or by using ACI EPGs and contracts, or both models complementing each other. Network policy of container workloads becomes a shared responsibility while ensuring global compliance and tenant constraints are enforced.
        displayName: Cisco ACI Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cisco
        - aci
        links:
        - name: Cisco ACI Homepage
          url: https://www.cisco.com/c/en/us/td/docs/switches/datacenter/aci/apic/sw/kb/b_Kubernetes_Integration_with_ACI.html
        maintainers:
        - email: abhishs9@cisco.com
          name: 'Abhishek Sharma '
        - email: sumitnaiksatam@gmail.com
          name: Sumit Naiksatam
        - email: vlella@cisco.com
          name: Vishal Lella
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Cisco
        relatedImages:
        - quay.io/noiro/aci-containers-operator@sha256:2b2d685e254c206be5371285852a8f95d0318af69cb85b786427969e035c125c
        - quay.io/noiro/acc-provision-operator@sha256:9c8fa6d705602af13d539fb919e72e7ee5064c6998fc9ca95d10108e986c68fe
        version: 1.0.10
      entries:
      - name: aci-containers-operator.v1.0.10
        version: 1.0.10
      name: stable
    defaultChannel: stable
    packageName: aci-containers-operator
    provider:
      name: Cisco
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Accuknox
      provider-url: https://accuknox.com
    name: kubearmor-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubearmor-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          com.redhat.openshift.versions: v4.12
          createdAt: "2023-05-01T05:17:07Z"
          description: The KubeArmor Operator deloys and manages KubeArmor daemonset
            and its componenets within a cluster.
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.28.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/kubearmor/KubeArmor
          support: Accuknox
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          The KubeArmor Operator deploys and manages KubeArmor and its components:
          - KubeArmor Daemonset
          - KubeArmor Relay Server
          - KubeArmor Controller to manage CRDs (KubeArmorPolicy, KubeArmorHostPolicy)

          KubeArmor is a cloud-native runtime security enforcement system that restricts the behavior (such as process
          execution, file access, and networking operations) of pods, containers, and nodes (VMs) at the system level.

          KubeArmor leverages Linux security modules (LSMs) such as AppArmor, SELinux, or BPF-LSM to enforce the user-specified policies. KubeArmor generates rich alerts/telemetry events with container/pod/namespace identities by leveraging eBPF.

          ## Documentation

          github: [https://github.com/kubearmor/KubeArmor](https://github.com/kubearmor/KubeArmor)
          website: [https://kubearmor.io/](https://kubearmor.io/)
        displayName: Kubearmor Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubearmor-operator
        - runtime-security
        - security
        - cloud-native
        links:
        - name: Kubearmor
          url: https://kubearmor.io
        maintainers:
        - email: ramakant@accuknox.com
          name: Ramakant Sharma
        maturity: alpha
        provider:
          name: Accuknox
          url: https://www.accuknox.com/
        relatedImages:
        - docker.io/kubearmor/kubearmor-operator@sha256:8c2d56781a2adf78ee58f956e01417f38b6ad2f319feada6470e93f4f9bf0808
        - docker.io/kubearmor/kubearmor-controller@sha256:ccec0fb8bb1e91bd48631cc89baf752b842f19c33b50edab08f8b33e57af64d4
        - docker.io/kubearmor/kubearmor@sha256:52f7049ecdbfa779663d2cd44ddd7f36aecd8ccd2ca91e69709b628600029f13
        - docker.io/kubearmor/kubearmor-init@sha256:44fa534587972c5f85e3e6f362dd8268e9088a33c5657a21b72a7e8563164fb5
        - docker.io/kubearmor/kubearmor-relay-server@sha256:695f9ade3382bcf858077e2a7b9b01e0713071517603778d72f7aca5a6d71323
        version: 1.0.0
      entries:
      - name: kubearmor-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: kubearmor-operator.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.kubearmor.com/v1",
                "kind": "KubeArmorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kubearmorconfig",
                    "app.kubernetes.io/part-of": "kubearmoroperator"
                  },
                    "name": "kubearmorconfig-default",
                    "namespace": "placeholder"
                },
                "spec": {
                  "adapters": {
                    "elasticsearch": {
                      "url": "",
                      "enabled": false,
                      "auth": {
                        "secretName": "elastic-secret",
                        "usernameKey": "username",
                        "passwordKey": "password",
                        "allowInsecureTLS": false,
                        "caCertSecretName": "",
                        "caCertKey": "ca.crt"
                      },
                      "alertsIndex": "kubearmor-alerts"
                    }
                  },
                  "recommendedPolicies": {
                      "enable": true,
                      "matchExpressions": [
                          {
                              "key": "namespace",
                              "operator": "NotIn",
                              "values": [
                                  "kube-system",
                                  "kubearmor",
                                  "openshift-infra",
                                  "openshift-ingress",
                                  "openshift-ingress-canary",
                                  "openshift-ingress-operator",
                                  "openshift-insights",
                                  "openshift-kni-infra",
                                  "openshift-kube-apiserver",
                                  "openshift-kube-apiserver-operator",
                                  "openshift-kube-controller-manager",
                                  "openshift-kube-controller-manager-operator",
                                  "openshift-kube-scheduler",
                                  "openshift-kube-scheduler-operator",
                                  "openshift-kube-storage-version-migrator",
                                  "openshift-kube-storage-version-migrator-operator",
                                  "openshift-logging",
                                  "openshift-machine-api",
                                  "openshift-machine-config-operator",
                                  "openshift-managed-node-metadata-operator",
                                  "openshift-managed-upgrade-operator",
                                  "openshift-marketplace",
                                  "openshift-monitoring",
                                  "openshift-multus",
                                  "openshift-must-gather-operator",
                                  "openshift-network-diagnostics",
                                  "openshift-network-node-identity",
                                  "openshift-network-operator",
                                  "openshift-node",
                                  "openshift-nutanix-infra",
                                  "openshift-oauth-apiserver",
                                  "openshift-observability-operator",
                                  "openshift-ocm-agent-operator",
                                  "openshift-openstack-infra",
                                  "openshift-operator-lifecycle-manager",
                                  "openshift-operators",
                                  "openshift-operators-redhat",
                                  "openshift-osd-metrics",
                                  "openshift-ovirt-infra",
                                  "openshift-ovn-kubernetes",
                                  "openshift-package-operator",
                                  "openshift-rbac-permissions",
                                  "openshift-route-controller-manager",
                                  "openshift-route-monitor-operator",
                                  "openshift-security",
                                  "openshift-service-ca",
                                  "openshift-service-ca-operator",
                                  "openshift-splunk-forwarder-operator",
                                  "openshift-sre-pruning",
                                  "openshift-user-workload-monitoring",
                                  "openshift-validation-webhook",
                                  "openshift-velero",
                                  "openshift-vsphere-infra",
                                  "openshift",
                                  "openshift-addon-operator",
                                  "openshift-apiserver",
                                  "openshift-apiserver-operator",
                                  "openshift-aqua",
                                  "openshift-authentication",
                                  "openshift-authentication-operator",
                                  "openshift-backplane",
                                  "openshift-backplane-cee",
                                  "openshift-backplane-csa",
                                  "openshift-backplane-cse",
                                  "openshift-backplane-csm",
                                  "openshift-backplane-managed-scripts",
                                  "openshift-backplane-mcs-tier-two",
                                  "openshift-backplane-mobb",
                                  "openshift-backplane-srep",
                                  "openshift-backplane-tam",
                                  "openshift-cloud-controller-manager",
                                  "openshift-cloud-controller-manager-operator",
                                  "openshift-cloud-credential-operator",
                                  "openshift-cloud-ingress-operator",
                                  "openshift-cloud-network-config-controller",
                                  "openshift-cloud-platform-infra",
                                  "openshift-cluster-csi-drivers",
                                  "openshift-cluster-machine-approver",
                                  "openshift-cluster-node-tuning-operator",
                                  "openshift-cluster-samples-operator",
                                  "openshift-cluster-storage-operator",
                                  "openshift-cluster-version",
                                  "openshift-codeready-workspaces",
                                  "openshift-config",
                                  "openshift-config-managed",
                                  "openshift-config-operator",
                                  "openshift-console",
                                  "openshift-console-operator",
                                  "openshift-console-user-settings",
                                  "openshift-controller-manager",
                                  "openshift-controller-manager-operator",
                                  "openshift-custom-domains-operator",
                                  "openshift-customer-monitoring",
                                  "openshift-deployment-validation-operator",
                                  "openshift-dns",
                                  "openshift-dns-operator",
                                  "openshift-etcd",
                                  "openshift-etcd-operator",
                                  "openshift-host-network",
                                  "openshift-image-registry"
                              ]
                          }
                      ]
                  },
                  "defaultCapabilitiesPosture": "audit",
                  "defaultFilePosture": "audit",
                  "defaultNetworkPosture": "audit",
                  "defaultVisibility": "process,network",
                  "enableStdOutLogs": false,
                  "enableStdOutAlerts": false,
                  "enableStdOutMsgs": false,
                  "seccompEnabled": false
                }
              }
            ]
          alm-examples-metadata: '{"kubearmorconfig-default":{"description":"Default
            KubeArmorConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Security, Monitoring
          certified: "true"
          com.redhat.openshift.versions: v4.9
          containerImage: docker.io/kubearmor/kubearmor-operator@sha256:d6c6b16e636474568835ded97cecb5ed17bc6004501227f6b6d63522ed57b01c
          createdAt: "2024-03-20T17:43:15Z"
          description: KubeArmor Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: kubearmor
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kubearmor/KubeArmor/blob/main/pkg/KubeArmorOperator/
          support: Accuknox
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeArmorConfig is the Schema for the kubearmorconfigs API
            kind: KubeArmorConfig
            name: kubearmorconfigs.operator.kubearmor.com
            version: v1
        description: |2-

          ## Prerequisites

          1. Create a new project for KubeArmor (Recommended)

             ```
             oc new-project kubearmor
             ```

           2. Create [KubeArmor SCC](https://github.com/kubearmor/KubeArmor/blob/main/pkg/KubeArmorOperator/config/rbac/kubearmor-scc.yaml)

             ```
             oc create -f https://raw.githubusercontent.com/kubearmor/KubeArmor/main/pkg/KubeArmorOperator/config/rbac/kubearmor-scc.yaml
             ```

             Note: please change the serviceaccount namespace if you are using some other namespace than the kubearmor.

          KubeArmor is a cloud-native runtime security enforcement system that restricts the behavior
          (such as process execution, file access, and networking operations) of pods, containers,
          and nodes (VMs) at the system level.
          KubeArmor leverages Linux security modules (LSMs) such as AppArmor, SELinux, or BPF-LSM to enforce the
          user-specified policies. KubeArmor generates rich alerts/telemetry events with container/pod/namespace
          identities by leveraging eBPF.

          ## Features and Benefits

          ### Harden Infrastructure

          safeguards vital paths like certificate bundles, enforces security rules based on MITRE, STIGs, and CIS guidelines,
          and restricts access to raw database tables, enhancing overall cluster security.

          ### Least Permissive Access

          controls sensitive asset access through process and network whitelisting,
          reducing security risks and maintaining system integrity.

          ### Application Behavior

          in-depth monitoring of process executions, file system access, service bindings,
          ingress and egress connections, and sensitive system call profiling

          ## Documentation

          github: [https://github.com/kubearmor/KubeArmor](https://github.com/kubearmor/KubeArmor)

          website: [https://kubearmor.io/](https://kubearmor.io/)

          ## Installing the KubeArmor Operator

          The KubeArmor Operator, available through Red Hat's Operator catalog, offers stable update channel.
          **Stable Channel**: This channel provides the latest stable version of KubeArmor along with its patches.

          The KubeArmor operator is responsible for deploying various components of the KubeArmor application, including the KubeArmor
          Daemonset, KubeArmor Controller, KubeArmor Relay Server, and KubeArmor Snitch. The role of KubeArmor Snitch is to gather information
          about the enabled Linux security modules on the host, which is utilized by KubeArmor to enforce runtime security policies.
          Additionally, KubeArmor Snitch collects details about the container runtime socket and storage volume cluster configurations.
          This deployment is executed as a short-term job and necessitates root user privileges. KubeArmor Snitch is certified by RedHat
          as a container with root privileges.

          KubeArmor Operator comes with a custom resource:

          - **KubeArmorConfig** - User will provide the initial configurations of the KubeArmor using KubeArmorConfig CR.

          ### KubeArmorConfig Customization Options:

          | Parameter                                 | Description                                          | Default                                 |
          | ---                                       | ---                                                  | ---                                     |
          | `alertThrottling`                         | Enable throttling for the KubeArmor generated alerts | False                                   |
          | `defaultFilePosture`                      | Default file posture for KubeArmor                   | audit                                   |
          | `defaultNetworkPosture`                   | Default network posture for KubeArmor                | audit                                   |
          | `defaultCapabilitiesPosture`              | Default capabilities posture for KubeArmor           | audit                                   |
          | `defaultVisibility`                       | Default visibility for KubeArmor                     | process,file,network                    |
          | `enableStdOutLogs` |  show stdout logs for relay server | false |
          | `enableStdOutAlerts` |  show stdout alerts for relay server | false |
          | `enableStdOutMsgs` |  show stdout messages for relay server | false |
          | `kubearmorImage.image`                    | Default image for KubeArmor Daemonset                | kubearmor/kubearmor-ubi:stable          |
          | `kubearmorImage.imagePullPolicy`          | Default imagePullPolicy for KubeArmor Daemonset      | Always                                  |
          | `kubearmorInitImage.image`                | Default image for KubeArmor Init Container           | kubearmor/kubearmor-init:stable         |
          | `kubearmorInitImage.imagePullPolicy`      | Default imagePullPolicy for KubeArmor Init Container | Always                                  |
          | `kubearmorRelayImage.image`               | Default image for KubeArmor Relay Server             | kubearmor/kubearmor-relay-server:latest |
          | `kubearmorRelayImage.imagePullPolicy`     | Default imagePullPolicy for KubeArmor Relay Server   | Always                                  |
          | `kubearmorControllerImage.image`          | Default image for KubeArmor Controller               | kubearmor/kubearmor-controller:latest   |
          | `kubearmorControllerImage.imagePullPolicy`| Default imagePullPolicy for KubeArmor Controller     | Always                                  |
          | `maxAlertPerSec`                          | Maximum no of alerts per second to trigger throttling| 10                                      |
          | `recommendedPolicies`                     | Recommended Policies Spec                            | -                                       |
          | `recommendedPolicies.matchExpressions`    | Target Namespaces for recommended policies           | -                                       |
          | `recommendedPolicies.excludePolicy`       | Exclude policies from recommended policies           | -                                       |
          | `seccompEnabled`                          | Deploy KubeArmor Daemon with default seccomp profile | True                                    |
          | `throttleSec`                             | Thottling duration in seconds                        | 30                                      |

          Users are not required to specify the images in the KubearmorConfig custom resource (CR). By default, the operator will utilize the images published with the operator release.

          For detailed information on Default Posture refer [this guide](https://github.com/kubearmor/KubeArmor/blob/main/getting-started/default_posture.md)

          similarly for default visibility refer [this guide](https://github.com/kubearmor/KubeArmor/blob/main/getting-started/kubearmor_visibility.md)

          ## Support

          For Support please join KubeArmor Slack [Here](https://join.slack.com/t/kubearmor/shared_invite/zt-1ltmqdbc6-rSHw~LM6MesZZasmP2hAcA)
        displayName: KubeArmor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - security
        - alerting
        - runtime-security
        - cloud native
        - container security
        links:
        - name: Website
          url: https://kubearmor.io
        - name: Github
          url: https://github.com/kubearmor/KubeArmor.git
        - name: Slack
          url: https://join.slack.com/t/kubearmor/shared_invite/zt-1ltmqdbc6-rSHw~LM6MesZZasmP2hAcA
        maintainers:
        - email: ramakant@accuknox.com
          name: Ramakant Sharma
        maturity: stable
        provider:
          name: Accuknox
          url: https://accuknox.com
        relatedImages:
        - docker.io/kubearmor/kubearmor-snitch@sha256:c3d0f45590f9a0ebb50254047f12a7468844bcae7961eccb11f5ef578832c0eb
        - docker.io/kubearmor/kubearmor-ubi@sha256:629e295a2e64953bf95f3f3e842c42ff665ed6e4db32db8e43dcde4befdd3fc2
        - docker.io/kubearmor/kubearmor-operator@sha256:d6c6b16e636474568835ded97cecb5ed17bc6004501227f6b6d63522ed57b01c
        - docker.io/kubearmor/kubearmor-init@sha256:e30ee508aa95d273f74a543e1b412b98c1714d04f093b59991c0ec2abaa450be
        - docker.io/kubearmor/kubearmor-relay-server@sha256:0a1a20ef02bd05eaa4a900948cadee6ae7650b0dde70abe1935608eaa5f931b7
        - docker.io/kubearmor/kubearmor-controller@sha256:2d369138cdbb31468c41e9ce70eda986a16bbeb9dd1192cd8a011c9df8a83857
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:1153a4592746b05e019bde4d818d176ff9350c013f84d49829032540de882841
        version: 1.5.2
      entries:
      - name: kubearmor-operator.v1.5.2
        version: 1.5.2
      name: stable
    defaultChannel: stable
    packageName: kubearmor-operator-certified
    provider:
      name: Accuknox
      url: https://accuknox.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Tigera
      provider-url: https://tigera.io/
    name: tigera-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tigera-operator.v1.20.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:bd93e66be5cfbc123fe4e8c3f6e8ddf7654217bf26136f06dcf23f28b53063d2
          createdAt: "2021-10-05T20:02:14.564507Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.20.4
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:bd93e66be5cfbc123fe4e8c3f6e8ddf7654217bf26136f06dcf23f28b53063d2
        version: 1.20.4
      entries:
      - name: tigera-operator.v1.20.4
        version: 1.20.4
      - name: tigera-operator.v1.20.3
        version: 1.20.3
      - name: tigera-operator.v1.20.2
        version: 1.20.2
      - name: tigera-operator.v1.20.1
        version: 1.20.1
      name: release-v1.20
    - currentCSV: tigera-operator.v1.23.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:f258275fe3d48c2ae16346a57f628523a6f2fb7b854b6d43b169056e158537f1
          createdAt: "2022-01-12T22:55:57.409244Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.23.4
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:f258275fe3d48c2ae16346a57f628523a6f2fb7b854b6d43b169056e158537f1
        version: 1.23.4
      entries:
      - name: tigera-operator.v1.23.4
        version: 1.23.4
      - name: tigera-operator.v1.23.3
        version: 1.23.3
      - name: tigera-operator.v1.23.2
        version: 1.23.2
      - name: tigera-operator.v1.23.1
        version: 1.23.1
      - name: tigera-operator.v1.23.0
        version: 1.23.0
      name: release-v1.23
    - currentCSV: tigera-operator.v1.25.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:e862708c700df55f99a9c9bb3c453e31f22aef1993d6079d80433eb8da831c94
          createdAt: "2022-02-17T05:39:13.720885Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.25.1
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:e862708c700df55f99a9c9bb3c453e31f22aef1993d6079d80433eb8da831c94
        version: 1.25.1
      entries:
      - name: tigera-operator.v1.25.1
        version: 1.25.1
      name: release-v1.25
    - currentCSV: tigera-operator.v1.28.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:9e8e47045fc67a2b00f9f2fd575fa526d7ddcac50af0b8b6d583b7518ccce120
          createdAt: "2023-05-17T16:13:52.148441156Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.28.13
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:9e8e47045fc67a2b00f9f2fd575fa526d7ddcac50af0b8b6d583b7518ccce120
        version: 1.28.13
      entries:
      - name: tigera-operator.v1.28.13
        version: 1.28.13
      name: release-v1.28
    - currentCSV: tigera-operator.v1.29.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:2a5acd29d8dca6adb318af04dbb15d93d899ed2b28b2ee9aaac09147395efb58
          createdAt: "2023-03-31T20:49:52.518230079Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.29.3
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:2a5acd29d8dca6adb318af04dbb15d93d899ed2b28b2ee9aaac09147395efb58
        version: 1.29.3
      entries:
      - name: tigera-operator.v1.29.3
        version: 1.29.3
      name: release-v1.29
    - currentCSV: tigera-operator.v1.30.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:431f037ff18b5c867d01312e42671effc55602421aeed25dd3f6109f70596b4a
          createdAt: "2023-11-17T00:52:31.263564294Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.30.9
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: APIServer
            name: apiservers.operator.tigera.io
            version: v1
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:431f037ff18b5c867d01312e42671effc55602421aeed25dd3f6109f70596b4a
        version: 1.30.9
      entries:
      - name: tigera-operator.v1.30.9
        version: 1.30.9
      name: release-v1.30
    - currentCSV: tigera-operator.v1.32.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:715ac9a30f8a9579e44258af20de354715429e11836b493918e9e1a696e9b028
          createdAt: "2023-12-15T23:45:03.964812504Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.32.3
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:715ac9a30f8a9579e44258af20de354715429e11836b493918e9e1a696e9b028
        version: 1.32.3
      entries:
      - name: tigera-operator.v1.32.3
        version: 1.32.3
      name: release-v1.32
    - currentCSV: tigera-operator.v1.36.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:3341fa9475c0325b86228c8726389f9bae9fd6c430c66fe5cd5dc39d7bb6ad4b
          createdAt: "2025-02-06T02:39:01.007218336Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.36.5
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:3341fa9475c0325b86228c8726389f9bae9fd6c430c66fe5cd5dc39d7bb6ad4b
        version: 1.36.5
      entries:
      - name: tigera-operator.v1.36.5
        version: 1.36.5
      name: release-v1.36
    - currentCSV: tigera-operator.v1.38.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "calicoNetwork": {
                    "linuxDataplane": "BPF"
                  },
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing, Monitoring, Networking, Security
          containerImage: quay.io/tigera/operator@sha256:e0a34b265aebce1a2db906d8dad99190706e8bf3910cae626b9c2eb6bbb21775
          createdAt: "2025-05-06T13:27:07.802897329Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          features.operators.openshift.io/valid-subscription: No subscription required
          olm.skipRange: <1.38.0
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift.
        displayName: Tigera Operator v1.38
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tigera-operator
        - networking
        - security
        - monitoring
        links:
        - name: Tigera
          url: https://tigera.io/
        - name: Calico Introduction
          url: https://docs.tigera.io/calico/latest/about/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.tigera.io/calico/latest/getting-started/kubernetes/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
          url: https://tigera.io/
        relatedImages:
        - quay.io/tigera/operator@sha256:e0a34b265aebce1a2db906d8dad99190706e8bf3910cae626b9c2eb6bbb21775
        version: 1.38.0
      entries:
      - name: tigera-operator.v1.38.0
        version: 1.38.0
      name: release-v1.38
    defaultChannel: release-v1.38
    packageName: tigera-operator
    provider:
      name: Tigera
      url: https://tigera.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solution-Soft
      provider-url: www.solution-soft.com
    name: timemachine-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: timemachine-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tm.solution-soft.com/v1alpha1",
                "kind": "TimeMachine",
                "metadata": {
                  "name": "timemachine-sample"
                },
                "spec": {
                  "cluster_provider": "ibm",
                  "aws": {
                    "shared_volume_name": ""
                  },
                  "ibm": {
                    "persistentvolume": {
                      "storageclass": "ibmc-file-silver-gid"
                    }
                  },
                  "tm_lichost": "52.152.148.40",
                  "tm_licpass": "qwaszx",
                  "tm_licport": "57777",
                  "imagePullSecrets": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
          createdAt: "2024-02-05T09:46:17Z"
          description: Time Machine Operator enables time travel of OpenShift pods,
            without the need to change the system clock on any of the cluster nodes
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/timemachine-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/timemachine-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=0.0.4 <1.0.17'
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: SolutionSoft Systems, Inc. - support@solution-soft.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Time Machine
            displayName: Time Machine
            kind: TimeMachine
            name: timemachines.tm.solution-soft.com
            version: v1alpha1
        description: |-
          Time Machine operator enables time travel of OpenShift pods/deployments.

          To time travel a pod where your application is, you just need to install Time Machine operator and create a Time Machine custom resource in the same namespace, without the need of changing the container images of the application you’re using, or adding containers to the pod with your application.

          Time travelling is done on the namespace level where all the (configured) target pods/deployments see the same virtual time.
        displayName: Time Machine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - time shift
        - time travel
        - virtual clocks
        - time machine
        - virtual time
        - solution-soft
        links:
        - name: Timemachine Operator
          url: https://catalog.redhat.com/software/operators/detail/5e98727a6c5dcb34dfbb19e3
        maintainers:
        - email: dtrajkovic@solution-soft.com
          name: Dragan Trajkovic
        maturity: alpha
        provider:
          name: Solution-Soft
          url: www.solution-soft.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0c304c11986790f446604319dfc29db494b15f5f9fa13f2c436688ec29c92dbe
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
        - registry.connect.redhat.com/rhel-timemachine-container/virtual-clock-testing-tool@sha256:a81a823328b0ee795c40668c89a08bcf545c5192d9eed0146071e3721a0a8372
        version: 1.0.17
      entries:
      - name: timemachine-operator.v1.0.17
        version: 1.0.17
      - name: timemachine-operator.v1.0.16
        version: 1.0.16
      name: alpha
    defaultChannel: alpha
    packageName: timemachine-operator-rhmp
    provider:
      name: Solution-Soft
      url: www.solution-soft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: ironic-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ironic-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "Ironic",
                "metadata": {
                  "name": "ironic"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseInstance": "openstack",
                  "ironicAPI": {},
                  "ironicConductors": [
                    {
                      "storageRequest": "10G"
                    }
                  ],
                  "ironicInspector": {},
                  "ironicNeutronAgent": {},
                  "secret": "osp-secret",
                  "serviceUser": "ironic",
                  "storageClass": "local-storage"
                }
              },
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "IronicAPI",
                "metadata": {
                  "name": "ironicapi-sample"
                },
                "spec": {
                  "databaseHostname": "openstack.ironic-kuttl-tests.svc"
                }
              },
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "IronicConductor",
                "metadata": {
                  "name": "ironicconductor-sample"
                },
                "spec": {
                  "databaseHostname": "openstack.ironic-kuttl-tests.svc",
                  "storageRequest": "10G"
                }
              },
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "IronicInspector",
                "metadata": {
                  "name": "ironicinspector-sample"
                },
                "spec": {
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "ironic.openstack.org/v1beta1",
                "kind": "IronicNeutronAgent",
                "metadata": {
                  "name": "ironic-neutron-agent-sample",
                  "namespace": "openstack"
                },
                "spec": {
                  "secret": "osp-secret"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:02Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IronicAPI is the Schema for the ironicapis API
            displayName: Ironic API
            kind: IronicAPI
            name: ironicapis.ironic.openstack.org
            version: v1beta1
          - description: IronicConductor is the Schema for the ironicconductors Conductor
            displayName: Ironic Conductor
            kind: IronicConductor
            name: ironicconductors.ironic.openstack.org
            version: v1beta1
          - description: IronicInspector is the Schema for the ironicinspectors
            displayName: Ironic Inspector
            kind: IronicInspector
            name: ironicinspectors.ironic.openstack.org
            version: v1beta1
          - description: IronicNeutronAgent is the Schema for the ML2 networking-baremetal's
              agent
            displayName: Ironic Neutron Agent
            kind: IronicNeutronAgent
            name: ironicneutronagents.ironic.openstack.org
            version: v1beta1
          - description: Ironic is the Schema for the ironics API
            displayName: Ironic
            kind: Ironic
            name: ironics.ironic.openstack.org
            version: v1beta1
        description: Ironic Operator
        displayName: Ironic Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Identity
        - Ironic
        links:
        - name: Ironic Operator
          url: https://github.com/openstack-k8s-operators/ironic-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/ironic-rhel9-operator@sha256:184a1783a3b2a38bdb23f9811243414a2b21c16d78ca79ed83e2bcaa720d199b
        - registry.redhat.io/rhoso/openstack-ironic-api-rhel9@sha256:86c5330a3065a11f251e8c333bd855d37294c4e9d36cc68208fd16b3f95ef15a
        - registry.redhat.io/rhoso/openstack-ironic-conductor-rhel9@sha256:4950d74335eb48ed04f1a4f602730f1dfd42f139233a4edfebef0d8403c00b2d
        - registry.redhat.io/rhoso/openstack-ironic-inspector-rhel9@sha256:34ff13efc254d4390fbef0a5c088c569241d44d063ec16aed42e36d8179e0c85
        - registry.redhat.io/rhoso/openstack-ironic-pxe-rhel9@sha256:d49af96baa8d90dad3b244ed8c006a9945c41d05e5fa197cc35cc662b49cdd7c
        - registry.redhat.io/rhoso/openstack-ironic-neutron-agent-rhel9@sha256:132627eaf483353c5daefbbc84c6b06f51c50d0813a229ddc558e9df18bb5578
        - registry.redhat.io/rhoso/ironic-python-agent-rhel9@sha256:f41166809d2e7d76446c01f81d325a320df21e64f4b776d0f83cadcb9387eb7f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: ironic-operator.v1.0.6
        version: 1.0.6
      - name: ironic-operator.v1.0.5
        version: 1.0.5
      - name: ironic-operator.v1.0.4
        version: 1.0.4
      - name: ironic-operator.v1.0.3
        version: 1.0.3
      - name: ironic-operator.v1.0.2
        version: 1.0.2
      - name: ironic-operator.v1.0.1
        version: 1.0.1
      - name: ironic-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: ironic-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kove
      provider-url: https://kove.com
    name: kove-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kove-operator.v3.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "containers.kove.com/v1",
                "kind": "KoveMemoryConfig",
                "metadata": {
                  "labels": {
                    "KoveMemoryConfig": ""
                  },
                  "name": "example-memory-config"
                },
                "spec": {
                  "container": "",
                  "namespace": "",
                  "size": ""
                }
              },
              {
                "apiVersion": "containers.kove.com/v1",
                "kind": "KoveSystem",
                "metadata": {
                  "labels": {
                    "KoveSystem": ""
                  },
                  "name": "kove-system"
                },
                "spec": {
                  "autogrowInitialIncrement": "1 GiB",
                  "autogrowMaxIncrement": "1 TiB",
                  "l4MaxMemoryPercent": 75,
                  "managementConsoleAddress": "",
                  "memoryPlusLocalMemoryMaxSize": "0 MiB"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: quay.io/kove/kove-operator@sha256:0fb5fa2e6702d5e5bf692c25f190f65d306fc72cdceebf0814ff64adcb5f6994
          createdAt: "2025-09-25T16:31:36Z"
          description: Configure containers with Kove® pooled memory
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: kove-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Kove
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configure containers with Kove® pooled memory
            displayName: Kove Memory Configuration
            kind: KoveMemoryConfig
            name: kovememoryconfigs.containers.kove.com
            version: v1
          - description: Configure cluster with Kove:SDM™
            displayName: Kove System Configuration
            kind: KoveSystem
            name: kovesystems.containers.kove.com
            version: v1
        description: |
          ## About the Operator
          The Kove:SDM™ operator integrates Kove's software-defined memory
          solution with Kubernetes and OpenShift clusters. The operator enables
          containers to transparently provision any amount of memory, enabling total
          container memory capacity to exceed physical hardware limitations.

          Use the Kove:SDM™ operator to:

          * Create custom resources to automatically allocate memory to specific
            containers
          * Increase container flexibility with dynamic memory capacity that
            auto-grows as needed
          * Scale up container density by removing memory as an orchestration
            bottleneck

          ## Using the Operator
          Using the operator is as simple as creating a custom resource which
          specifies an initial Kove memory allocation size, a namespace, and a
          container name. Containers in the specified namespace with a matching
          container name will automatically be provided with a Kove memory allocation
          which will auto-grow as needed.

          ## About Kove
          Kove makes memory a provisionable resource, just like storage, CPUs, and
          networking. Kove:SDM™ provides a transparent, linearly scalable
          memory resource accessible to any application without code changes, from
          bare-metal apps to virtual machines and containers.
        displayName: Kove:SDM™ Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kove
        - Pooled Memory
        - Memory Pool
        - Memory Target
        - External Memory
        - Memory
        links:
        - name: Kove
          url: https://kove.com
        maintainers:
        - email: contact-kove@kove.com
          name: Kove
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Kove
          url: https://kove.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fa0acefda43e309e7f9f31f5247c933d0b868ff1573ece2f38e0549d7e2ec942
        - registry.access.redhat.com/ubi8/pause@sha256:eacbaaadc932633330e799eda24e9b731b5a4ba7ed823ee70791d21029724a9a
        - quay.io/kove/kove-client-drivers@sha256:1c1786e1bbf16609218154126b4340b2f7c4b71cbd9045589402f75c032ea38d
        - quay.io/kove/kove-xmem@sha256:41bfc536a278153b0a4460d2836b56762b56cf50b94d17e123a63be066c28fe7
        - quay.io/kove/kove-operator@sha256:0fb5fa2e6702d5e5bf692c25f190f65d306fc72cdceebf0814ff64adcb5f6994
        - quay.io/kove/kove-container-hook@sha256:66ee33d59ced63406d9b40f3a3226fab7f52071f8bed9aa1ba7d859837038528
        - quay.io/kove/kove-agent@sha256:9d026b925e1d414a531530a30ed8b5f10b421ee12c7c13949ea0ee9907d36b75
        - quay.io/kove/kove-client-drivers@sha256:716ccda2d8f082a954f2042fd189fec257bd2dcb2c50e497bad42b9a9e5f0e3d
        - quay.io/kove/kove-compatibility-detection@sha256:0303830a1f785b5e4ce6824db7c50f8feee2678fd5e87ded983686ab3f60b232
        version: 3.0.2
      entries:
      - name: kove-operator.v3.0.2
        version: 3.0.2
      - name: kove-operator.v3.0.1
        version: 3.0.1
      - name: kove-operator.v3.0.0
        version: 3.0.0
      name: stable
    defaultChannel: stable
    packageName: kove-operator
    provider:
      name: Kove
      url: https://kove.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Fujitsu
      provider-url: https://www.postgresql.fastware.com/
    name: fujitsu-enterprise-postgres-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fujitsu-enterprise-postgres-operator.v7.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPAction",
                "metadata": {
                  "name": "new-fep-action"
                },
                "spec": {
                  "fepAction": {
                    "args": [
                      "new-fep-sts-0",
                      "new-fep-sts-1"
                    ],
                    "type": "reload"
                  },
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "targetClusterName": "new-fep"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPExporter",
                "metadata": {
                  "name": "new-fep-exporter"
                },
                "spec": {
                  "fepExporter": {
                    "exporterLogLevel": "error",
                    "fepClusterList": [
                      "new-fep1"
                    ],
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "userCustomQueries": "usr_example:\n  query: \"SELECT EXTRACT(EPOCH FROM (now() - pg_last_xact_replay_timestamp())) as lag\"\n  master: true\n  metrics:\n    - lag:\n        usage: \"GAUGE\"\n        description: \"Replication lag behind master in seconds\""
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPLogging",
                "metadata": {
                  "name": "new-fep-logging"
                },
                "spec": {
                  "fepLogging": {
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPPgpool2",
                "metadata": {
                  "name": "new-fep-pgpool2"
                },
                "spec": {
                  "clientAuthMethod": "scram",
                  "count": 2,
                  "customhba": "local   all         all                               trust\nhost    all         all         127.0.0.1/32          trust\nhost    all         all         ::1/128               trust\n",
                  "customlogsize": "128Mi",
                  "customparams": "listen_addresses = '*'\npcp_listen_addresses = '*'\nnum_init_children = 32\nreserved_connections = 0\nenable_pool_hba = off\nallow_clear_text_frontend_auth = off\nauthentication_timeout = 80\nbackend_weight0 = 1\nbackend_weight1 = 1\nbackend_flag0 = 'ALWAYS_PRIMARY'\nbackend_flag1 = 'DISALLOW_TO_FAILOVER'\nconnection_cache = on\nmax_pool = 4\nlisten_backlog_multiplier = 2\nserialize_accept = off\nchild_life_time = 300\nclient_idle_limit = 0\nchild_max_connections = 0\nconnection_life_time = 0\nreset_query_list = 'ABORT; DISCARD ALL'\nclient_min_messages = info\nlog_min_messages = debug1\nlog_statement = on\nlog_per_node_statement = on\nlog_client_messages = on\nlog_hostname = on\nlog_connections = on\nlog_line_prefix = '%t: pid %p: '\nload_balance_mode = on\nignore_leading_white_space = on\nwhite_function_list = ''\nblack_function_list = 'currval,lastval,nextval,setval'\nblack_query_pattern_list = ''\ndatabase_redirect_preference_list = ''\napp_name_redirect_preference_list = ''\nallow_sql_comments = off\ndisable_load_balance_on_write = 'transaction'\nstatement_level_load_balance = on\nsr_check_period = 0\nsr_check_user = 'postgres'\ndelay_threshold = 0\nlog_standby_delay = 'none'\nssl = on\nssl_ciphers = 'HIGH:MEDIUM:+3DES:!aNULL'\nssl_prefer_server_ciphers = off\nssl_ecdh_curve = 'prime256v1'\nssl_dh_params_file = ''\nrelcache_expire = 0\nrelcache_size = 256\ncheck_temp_table = catalog\ncheck_unlogged_table = on\nenable_shared_relcache = off\nrelcache_query_target = primary\nwd_port0 = 9000\nfailover_on_backend_error = off\n",
                  "custompcp": "none",
                  "customsslcacert": "none",
                  "customsslcert": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                  "customsslkey": "none",
                  "fepclustername": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "limits": {
                    "cpu": "400m",
                    "memory": "512Mi"
                  },
                  "requests": {
                    "cpu": "200m",
                    "memory": "256Mi"
                  },
                  "serviceport": 9999,
                  "statusport": 9898
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPRestore",
                "metadata": {
                  "name": "new-fep-restore"
                },
                "spec": {
                  "fromFEPcluster": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "mcSpec": {
                    "limits": {
                      "cpu": "200m",
                      "memory": "300Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "restoretype": "latest",
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "toFEPcluster": "new-fep-2"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v2",
                "kind": "FEPCluster",
                "metadata": {
                  "name": "new-fep"
                },
                "spec": {
                  "fep": {
                    "customAnnotations": {
                      "allDeployments": {}
                    },
                    "forceSsl": true,
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "instances": 1,
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "podAntiAffinity": false,
                    "podDisruptionBudget": false,
                    "servicePort": 27500,
                    "syncMode": "off",
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "usePodName": true
                  },
                  "fepChildCrVal": {
                    "backup": {
                      "image": {
                        "pullPolicy": "IfNotPresent"
                      },
                      "mcSpec": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      },
                      "pgbackrestParams": "# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[global]\nrepo1-retention-full=7\nrepo1-retention-full-type=time\nlog-path=/database/log/backup\n",
                      "postScript": " ",
                      "preScript": " ",
                      "schedule": {
                        "num": 2
                      },
                      "schedule1": {
                        "schedule": "15 0 * * 0",
                        "type": "full"
                      },
                      "schedule2": {
                        "schedule": "15 0 * * 1-6",
                        "type": "incr"
                      },
                      "schedule3": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule4": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule5": {
                        "schedule": " ",
                        "type": " "
                      }
                    },
                    "customPgAudit": "# define pg audit custom params here to override defaults.\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[output]\nlogger = 'auditlog'\nlog_directory = '/database/log/audit'\nlog_truncate_on_rotation = on\nlog_filename = 'pgaudit-%a.log'\nlog_rotation_age = 1d\nlog_rotation_size = 0\n[rule]\n",
                    "customPgHba": "# define pg_hba custom rules here to be merged with default rules.\n# TYPE     DATABASE        USER        ADDRESS        METHOD\n",
                    "customPgParams": "# define custom postgresql.conf parameters below to override defaults.\n# Current values are as per default FEP deployment\n# If you add a library to shared_preload_libraries, add it after the default value.\n# fsep_operator_security is available in FEP15 and later container images.\nshared_preload_libraries='pgx_datamasking,pg_prewarm,pg_stat_statements,fsep_operator_security'\nsession_preload_libraries='pg_prewarm'\nmax_prepared_transactions = 100\nmax_worker_processes = 30\nmax_connections = 100\nwork_mem = 1MB\nmaintenance_work_mem = 12MB\nshared_buffers = 128MB\neffective_cache_size = 384MB\ncheckpoint_completion_target = 0.8\n# tcp parameters\ntcp_keepalives_idle = 30\ntcp_keepalives_interval = 10\ntcp_keepalives_count = 3\n# logging parameters in default fep installation\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\nlog_directory = '/database/log'\nlog_filename = 'logfile-%a.log'\nlog_file_mode = 0600\nlog_truncate_on_rotation = on\nlog_rotation_age = 1d\nlog_rotation_size = 0\nlog_checkpoints = on\nlog_line_prefix = '%e %t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h'\nlog_lock_waits = on\nlog_autovacuum_min_duration = 60s\nlogging_collector = on\npgaudit.config_file='/opt/app-root/src/pgaudit-cfg/pgaudit.conf'\nlog_replication_commands = on\nlog_min_messages = WARNING\nlog_destination = csvlog\n# wal_archive parameters in default fep installation\narchive_mode = on\narchive_command = 'pgbackrest --stanza=backupstanza --config=/database/userdata/pgbackrest.conf archive-push %p'\nwal_level = replica\nmax_wal_senders = 12\nwal_keep_segments = 64\ntrack_activities = on\ntrack_counts = on\npassword_encryption = 'scram-sha-256'\n",
                    "storage": {
                      "archivewalVol": {
                        "size": "1Gi"
                      },
                      "backupVol": {
                        "size": "2Gi"
                      },
                      "dataVol": {
                        "size": "2Gi"
                      },
                      "logVol": {
                        "size": "1Gi"
                      },
                      "tablespaceVol": {
                        "size": "512Mi"
                      },
                      "walVol": {
                        "size": "1200Mi"
                      }
                    },
                    "sysUsers": {
                      "pgRewindPassword": "rewind_password",
                      "pgRewindUser": "rewind_user",
                      "pgdb": "mydb",
                      "pgpassword": "mydbpassword",
                      "pgreplpassword": "repluserpwd",
                      "pgrepluser": "repluser",
                      "pguser": "mydbuser",
                      "tdepassphrase": "tde-passphrase"
                    },
                    "systemCertificates": {
                      "cacrt": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                      "crt": "-----BEGIN CERTIFICATE-----\nMIIDUTCCAjmgAwIBAgIRAMocW3qMoHrD6qRvMPppMkMwDQYJKoZIhvcNAQELBQAw\nNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9yIEt1\nYmVybmV0ZXMwHhcNMjEwMjA2MDQzMjM2WhcNMjYwMjA1MDQzMjM2WjA/MRAwDgYD\nVQQKEwdGdWppdHN1MSswKQYDVQQDEyJGVUpJVFNVIEVudGVycHJpc2UgUG9zdGdy\nZXMgU2VydmVyMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA4AI33yvH\nZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I2e4SceTKi6O3C/I1XuvWlpng\n5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4ANB5JWWqDOjrRT3o7nRPGXfila\nbP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYpmjdbfxabTz07ig0+6/cwKoRR\nxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTFYvmAH7gcdssSFBt8NPlUATHE\nsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6Wdgmu5H2pDml8CDNLDv98Aj7i\n+I5SRKKcVPlnuQIDAQABo1AwTjAdBgNVHSUEFjAUBggrBgEFBQcDAQYIKwYBBQUH\nAwIwDAYDVR0TAQH/BAIwADAfBgNVHSMEGDAWgBQcwrrUO0u+FhIUuVdrDRCQRsi6\nZjANBgkqhkiG9w0BAQsFAAOCAQEAm5dxBoI9pScOCvRAchg4CprdRDSJb9K6yB3O\nnCAxnM47iHeXnY3WlnI388kHu8DU7O4ba1tJbGs3KY9KzioPk43pU12jWkO1onoF\n+mTDjx/Ef1cYWA9r5q/LtgTa6Q2sxV4O2x67QW82aAnaxO34dV5zWCPIvAoovZBV\nHRT+BgCg3r2vD1RGKK2nl1aYJtWhO1SZubam+VttdZ/vbM9oOJctxmImsEtBXjkY\nKteePdQtLL5o03JhyXWyRshCq+HMmKf2KgyY8gvydGcP4eLQdBWcW40LcnVq6UjT\n0kJycJEKngMVademq1ZWHGaiYB7hyT6GhgIcHUJ2cKrPgbEh1Q==\n-----END CERTIFICATE-----",
                      "key": "-----BEGIN RSA PRIVATE KEY-----\nMIIEowIBAAKCAQEA4AI33yvHZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I\n2e4SceTKi6O3C/I1XuvWlpng5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4AN\nB5JWWqDOjrRT3o7nRPGXfilabP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYp\nmjdbfxabTz07ig0+6/cwKoRRxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTF\nYvmAH7gcdssSFBt8NPlUATHEsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6W\ndgmu5H2pDml8CDNLDv98Aj7i+I5SRKKcVPlnuQIDAQABAoIBAFPQYKlOzw/+BA0b\nyMIUpdctIMb/54CR/xR0mVw1DbSjigNVPjHUQvB8Y1B2FAITQObgJO06bAv0QdWN\nRb0/v/yYiNJDFjaLjaIAHlO/2+oWrXbFaZqgpVDJhB+e1xaZr2x7XGxm+p925k30\nl6pvIRY+I8JRKvZiV1VZHwL/R3JOtPr++xMZtLVjVOI+f+ySqJ+TZHuAjm49EKxj\ncEmmJ28b7QcziXsvKy00f+zbqLIBKXQdZAFU5eEr1BsDRXdRW+Kf0XIvftuy4BJZ\nvoKT+VGhEvF/qysswL4+6IAO6tpuYnnM0Y2d3sOGoWPkTcQK0MekYKzL/WmtCjNs\n9hodJtECgYEA5EWyhEOf4uOKe5TDp697UCUvXLoOR58FDe/S8XNvScn29jjOkqIg\nOMoqo9xAkJTNTzqn5UUdt1x/pgM2NxlPLFijrc0zQlX3SoOO2ryDd9WNi7YKtN16\nKJqa536WeZu2OEbuAZ+S3GALVy1RPeTNPnUOmKnF06DjDUGzLNCZy10CgYEA+zfw\n952DWuz1U0Z4wvAEqqcgUKXPKrkTXV/iUnjkDkrLYVr0ZofDNTXrdHl+UedFmaOC\ncieZn6DNhcdz5tKtyysGMH3g/qs9PfoGUngvcXsy0Egk04l3x1jc8TTCLqXZXYaQ\nHMsx51n+R58oncPtzYSUOr9qQ6PbC2CstTbFJA0CgYEAjGEsUliAB/jknfEzjXjG\nPdhQUxb8VyE864Az2lah9t/kJzFyIAziAeqZ5GE7t247AGFTBRTHHI8e1Qoemi3P\nWbc9GVIbFs1lIYbcIDpUIyrKPEP8O5QEXtoNLxXTFgAjRGKiVY87spjCAJ+W2ZhO\ne/1it5GYXfgQCYQA2yuBmOUCgYANRkR2YR1axaCk+NlSu6oTdmdPu6M5x7PNQE7O\nOtMaKjua9lppvIzFGAdMDUtueoEEAE7ZR1xnwfB6PDLUpJdIYAqgr1YfPt8qkjaZ\nTv56yZ7CwL0pbF8m6nwqRrZoDp1wwraEvvvxFKFKGY/k3kCHlpTakdjEoDjn3gDi\nRnWeVQKBgCEneMSzucei5LRppRtRaJw/Btll8qlPMlX3W7dxQ3cLwpmLOn0m51Fp\nPIZ44zYK8R6fu4+/sSrlfaIg86Ugeufp6YNxyNROKxUGza5vDIu5OftwWtBeg+UK\nZ8lLWNdX6pp7WMujmF3H1DrkBbauYMUKZ4UxUYtelgHERMePIxwb\n-----END RSA PRIVATE KEY-----"
                    }
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "false"
          containerImage: quay.io/fujitsu/fujitsu-enterprise-postgres-operator@sha256:b7dd55c0bfda84e398cbfbb87345fcec218cb866e286e18a111330321ab6ad54
          createdAt: "2025-09-10T11:50:37Z"
          description: OpenShift Operator for Fujitsu Enterprise Postgres
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.1.0 <7.3.0'
          operators.openshift.io/valid-subscription: '["Fujitsu Enterprise Postgres
            Advanced Edition Annual Subscription License per Core 17 Operator for
            Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/internal-objects: '["fepautoscales.fep.fujitsu.io","fepconfigs.fep.fujitsu.io","fepusers.fep.fujitsu.io","fepcerts.fep.fujitsu.io","fepvolumes.fep.fujitsu.io","fepbackups.fep.fujitsu.io"]'
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: fj-dbaas-tec@dl.jp.fujitsu.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: FEPAction
            name: fepactions.fep.fujitsu.io
            version: v1
          - kind: FEPAutoscale
            name: fepautoscales.fep.fujitsu.io
            version: v1
          - kind: FEPBackup
            name: fepbackups.fep.fujitsu.io
            version: v1
          - kind: FEPCert
            name: fepcerts.fep.fujitsu.io
            version: v1
          - kind: FEPCluster
            name: fepclusters.fep.fujitsu.io
            version: v2
          - kind: FEPConfig
            name: fepconfigs.fep.fujitsu.io
            version: v1
          - kind: FEPExporter
            name: fepexporters.fep.fujitsu.io
            version: v1
          - kind: FEPLogging
            name: feploggings.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2Cert
            name: feppgpool2certs.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2
            name: feppgpool2s.fep.fujitsu.io
            version: v1
          - kind: FEPRestore
            name: feprestores.fep.fujitsu.io
            version: v1
          - kind: FEPUpgrade
            name: fepupgrades.fep.fujitsu.io
            version: v1
          - kind: FEPUser
            name: fepusers.fep.fujitsu.io
            version: v1
          - kind: FEPVolume
            name: fepvolumes.fep.fujitsu.io
            version: v1
        description: "Fujitsu Enterprise Postgres 17 delivers an enterprise-grade
          PostgreSQL on OpenShift Container Platform.\n\nThis solution provides the
          flexibility of a hybrid cloud solution while delivering an enhanced distribution
          of PostgreSQL to support enterprise-level workloads and provide improved
          deployment and management, availability, performance, data governance and
          security.  \n\nAvailable as a multi-architecture container built for both
          amd64, s390x and ppc64le.\n\nThe download and Use of the Product is strictly
          subject to the terms of the End User License Agreement with Fujitsu Limited
          found at https://www.fast.fujitsu.com/fujitsu-enterprise-postgres-license-agreements.
          Where the Product that has been embedded as a whole or part into a third
          party program, only Authorised Customers may download and use the Product.\n"
        displayName: Fujitsu Enterprise Postgres Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - Fujitsu
        - fep
        links:
        - name: Fujitsu Enterprise Postgres
          url: https://www.postgresql.fastware.com/
        maintainers:
        - email: fj-dbaas-tec@dl.jp.fujitsu.com
          name: Fujitsu
        maturity: stable
        provider:
          name: Fujitsu
          url: https://www.postgresql.fastware.com/
        relatedImages:
        - quay.io/fujitsu/fujitsu-enterprise-postgres-operator@sha256:b7dd55c0bfda84e398cbfbb87345fcec218cb866e286e18a111330321ab6ad54
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-pgpool2@sha256:8eaace2f9c8e5e844992fe5e008cd1061af33cc79b4d442c4a83a65610bcc502
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-backup@sha256:ac268e2af652bdabb8ee5d764a2b335dc73c48b9022eee8c9f3cd422227a4d98
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-backup@sha256:78a3ad639d0c63858b9db0a5de36ce11a9ea121f4535e9ca07ac7defb689bc0d
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-upgrade@sha256:04af997b4f548ffec711b63a06c35715b4e7863537c83697b3ee0c1d95a6aec3
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentd@sha256:3c5534b62695867c5973c9106fc46895ed73a4b9136ad83b2586bf7e71c427d0
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentbit@sha256:9622d2191525da39e5100e1882dd2cfa3eda740d8094f3fbb71ca8a5734d2843
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-server@sha256:1c5823afc7e77d46940214693c12c0ad6ffd6c7d9a156e26399ab55cd5ea1dfa
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-pgpool2@sha256:c878935ffc796c405d811a6e8a9194aaaaa8652ea7d7b879c82919ce09794b35
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-upgrade@sha256:8a0f7ca1f59ba162424a308e8a70c4717135e2847e61210753b71876c9b0a75f
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-backup@sha256:dd8b84e200c9d0fb32cd567a33968e15e15efc115ff120f54fdd8862c7ea3111
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-upgrade@sha256:c0e44447349f02eb9f6105001136402f849d876a288d61997c8c05625facff45
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-utils@sha256:b7ae535f5e5fa5eb51c6f4cdb406c7b7a9dabe8573939b0431dada6f7d27b87e
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-restore@sha256:a6e90dd77634033f2009942b635078cceb0f4c6b15b031624c5dce820713b14c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-restore@sha256:17301c2191c0d75be82d12c5a35c0f74b8d65304bcfb6d4ab312a10ffa286eda
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-server@sha256:fa4e8849d75daf331ba7d488bcb52c083cd92bb5f288e4b62c980754e299f618
        - quay.io/fujitsu/fujitsu-enterprise-postgres-cronjob@sha256:7582964eef14f89f2ca9046298357afd0132fb236f30f44f84cee14692227b0c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-server@sha256:5c65f8e76f70341c512100760de6792a880ad0d6eebc9e68154b5bbd1332a896
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-restore@sha256:09be02274427e3022849fb635a4c9abdc34696838cae39f3afafc8009567db9d
        - quay.io/fujitsu/fujitsu-enterprise-postgres-exporter@sha256:1c0e22aeab65ff3848772424fc3ea6baa6b58b0d88db5413bda160a08435bcff
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-upgrade@sha256:bf3b069a71202ea78c7c0c7b36315c8bef8e6a1c11cd0668dcd888329e26ecfb
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-pgpool2@sha256:38aac8108dd3590abcfd854a6ec2c025a68d2170a1ba68d359b45fea6c83cec1
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-server@sha256:3e93a368a6198818e1b7c5525324c8130c5ed8d7b4cb499a47d4bec45ef0347b
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-pgpool2@sha256:185af2e4fd1701aded1ff5e44d4336562f5a219c30ba535c2a3097b914c24efe
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-backup@sha256:42252444e2dcfe5435532f318cce041990d2b71efa26007924717924d30a46ca
        - quay.io/fujitsu/fujitsu-enterprise-postgres-utils@sha256:57a3fb792480a0d3a8f075c48dee2d02f1d79cf459761e3305cf62799575a885
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-restore@sha256:a81a0ce926572b75aa3151e2d2cc45a0ca411cc0158d69b89247ef70418f3740
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-utils@sha256:edeec4d00fd8282aaa7ec6fe3ec4cd84070a6402611002a18addcac87fb8ee94
        version: 7.3.0
      entries:
      - name: fujitsu-enterprise-postgres-operator.v7.3.0
        version: 7.3.0
      - name: fujitsu-enterprise-postgres-operator.v7.2.7
        version: 7.2.7
      name: stable
    defaultChannel: stable
    packageName: fujitsu-enterprise-postgres-operator
    provider:
      name: Fujitsu
      url: https://www.postgresql.fastware.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-block-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12
    - currentCSV: ibm-block-csi-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "k8s.gcr.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
        version: 1.12.0
      entries:
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.0
    - currentCSV: ibm-block-csi-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
        version: 1.12.1
      entries:
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.1
    - currentCSV: ibm-block-csi-operator.v1.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:af9215670675fe9c4fa2dac7a14d547c249f3aa0c0e72f1e5390d237e411735b
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:af9215670675fe9c4fa2dac7a14d547c249f3aa0c0e72f1e5390d237e411735b
        version: 1.12.2
      entries:
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.2
    - currentCSV: ibm-block-csi-operator.v1.12.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:6e7e13d259f4d743524ebcd546af5c73d1a3ba90edaf4b5edd2005a3f8adcdfe
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:6e7e13d259f4d743524ebcd546af5c73d1a3ba90edaf4b5edd2005a3f8adcdfe
        version: 1.12.3
      entries:
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.3
    - currentCSV: ibm-block-csi-operator.v1.12.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:42728e7fd1d2302511ab08f51335ab188eef125870fa27f322df83ebb8e1f447
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:42728e7fd1d2302511ab08f51335ab188eef125870fa27f322df83ebb8e1f447
        version: 1.12.4
      entries:
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.4
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.5
    defaultChannel: stable
    packageName: ibm-block-csi-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cli-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cli-manager-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "CliManager",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-cli-manager-operator"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed",
                  "operatorLogLevel": "Normal"
                }
              }
            ]
          capabilities: Basic Install
          categories: Openshift Optional
          createdAt: "2024-06-04"
          description: "The CLI Manager is a comprehensive tool designed to simplify
            the management of \nOpenShift CLI plugins within the OpenShift environment.
            \nModeled after the popular krew plugin manager, it offers seamless integration,
            \neasy installation, and efficient handling of a wide array of plugins,
            \nenhancing your OpenShift command-line experience.\n"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          repository: https://github.com/openshift/cli-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: CLI Manager
            kind: CliManager
            name: climanagers.operator.openshift.io
            version: v1
          - displayName: CLI Plugin
            kind: Plugin
            name: plugins.config.openshift.io
            version: v1alpha1
        description: "The CLI Manager is a comprehensive tool designed to simplify
          the management of \nOpenShift CLI plugins within the OpenShift environment.
          \nModeled after the popular krew plugin manager, it offers seamless integration,
          \neasy installation, and efficient handling of a wide array of plugins,
          \nenhancing your OpenShift command-line experience.\n"
        displayName: CLI Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cli-manager-operator
        - cli-manager
        links:
        - name: Source Code
          url: https://github.com/openshift/cli-manager
        - name: Source Code
          url: https://github.com/openshift/cli-manager-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/cli-manager-operator/cli-manager-rhel9-operator@sha256:31973d7f43d052ddfb9554dd623e37fb541a8c1f08ea981558eaf4bb5ab22d5a
        - registry.redhat.io/cli-manager-operator/cli-manager-rhel9@sha256:07853bc33a0ce783d8d45c08c161e454a3abe09a8a8dadeca70c3a199466b58b
        version: 0.1.0
      entries:
      - name: cli-manager-operator.v0.1.0
        version: 0.1.0
      name: tech-preview
    defaultChannel: tech-preview
    packageName: cli-manager
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cryostat-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cryostat-operator.v4.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.cryostat.io/v1beta2",
                "kind": "Cryostat",
                "metadata": {
                  "name": "cryostat-sample"
                },
                "spec": {
                  "enableCertManager": true,
                  "eventTemplates": [],
                  "networkPolicies": {},
                  "reportOptions": {
                    "replicas": 0
                  },
                  "trustedCertSecrets": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Developer Tools
          containerImage: registry.redhat.io/cryostat/cryostat-rhel9-operator@sha256:867a649b54ad144295c7aceda6b90631841aa31a62062df62d4980313723f359
          createdAt: "2025-09-30T15:10:08Z"
          description: JVM monitoring and profiling tool
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.0 <4.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "operator.cryostat.io/v1beta2",
              "kind": "Cryostat",
              "metadata": {
                "name": "cryostat-sample"
              },
              "spec": {
                "enableCertManager": true,
                "reportOptions": {
                  "replicas": 0
                }
              }
            }
          operators.openshift.io/valid-subscription: '["Red Hat build of OpenJDK"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/cryostatio/cryostat-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cryostat allows you to install Cryostat for a single namespace,
              or multiple namespaces. It contains configuration options for controlling
              the Deployment of the Cryostat application and its related components.
              A Cryostat instance must be created to instruct the operator to deploy
              the Cryostat application.
            displayName: Cryostat
            kind: Cryostat
            name: cryostats.operator.cryostat.io
            version: v1beta2
          - description: Cryostat allows you to install Cryostat for a single namespace.
              It contains configuration options for controlling the Deployment of
              the Cryostat application and its related components. A Cryostat instance
              must be created to instruct the operator to deploy the Cryostat application.
            displayName: Cryostat
            kind: Cryostat
            name: cryostats.operator.cryostat.io
            version: v1beta1
        description: |
          The Red Hat build of Cryostat provides a cloud-based solution for interacting with the JDK Flight Recorder already present in OpenJDK 11+ JVMs. With Cryostat, users can remotely start, stop, retrieve, and even analyze JFR event data, providing the capability to easily take advantage of Flight Recorder's extremely low runtime cost and overhead and the flexibility to monitor applications and analyze recording data without transferring data outside of the cluster the application runs within.
          ##Prerequisites
          Cryostat requires [cert-manager](https://cert-manager.io/) to run. If not already installed in your cluster, please [install](https://cert-manager.io/docs/installation/) it using your preferred method.
          When running on Kubernetes, the operator also requires [Ingress configurations](https://github.com/cryostatio/cryostat-operator/blob/v4.0.0/docs/config.md#network-options) for each of its services to make them available outside of the cluster. The user is responsible for providing the hostnames for each Ingress.
        displayName: Red Hat build of Cryostat
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flightrecorder
        - java
        - jdk
        - jfr
        - jmc
        - missioncontrol
        - monitoring
        - profiling
        - diagnostic
        links:
        - name: Upstream Project
          url: https://github.com/cryostatio/cryostat
        - name: Website
          url: https://cryostat.io/
        maintainers:
        - email: cryostat-development@googlegroups.com
          name: The Cryostat Authors
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi9/nginx-124@sha256:7c1d7172880a3649591a0dc1ceb670c133ef88eed3d8192d8a2e80e3878d04c3
        - registry.redhat.io/cryostat/cryostat-reports-rhel9@sha256:5535a539b185fa2781621741ccd7f10a6b9947adc6170a95e930f2d9d7a9f608
        - registry.redhat.io/insights-runtimes-tech-preview/runtimes-inventory-rhel8-operator@sha256:58e01d11c3d30abc5b0ee3795511bd04e39aea21e2853b4f4cc6d1c8a2bd7bb0
        - registry.redhat.io/cryostat/cryostat-rhel9@sha256:fadae8713daf82c329573191381f8f9e4ebf3fc852ac5f1a68ce0ea1da376469
        - registry.redhat.io/cryostat/jfr-datasource-rhel9@sha256:1e5f29aa096d783fd0a18b3ba31c391dc203761ba3f4af8e5aeeef1f7cc86cee
        - registry.redhat.io/cryostat/cryostat-db-rhel9@sha256:23d61204ecaf037cf17bf9a5cc4428461a60fecce077884d9458802fb071daab
        - registry.redhat.io/cryostat/cryostat-grafana-dashboard-rhel9@sha256:5715e897f9a8969aedd0066b2ba597264495b4c85e4ae8ae7d8871f93d13f4c1
        - registry.redhat.io/cryostat/cryostat-agent-init-rhel9@sha256:273d6c3d3194a4e9ec50c1d8e147cd9bc6d277d8d66878d885a18528b9ba7dc6
        - registry.redhat.io/cryostat/cryostat-rhel9-operator@sha256:867a649b54ad144295c7aceda6b90631841aa31a62062df62d4980313723f359
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        - registry.redhat.io/cryostat/cryostat-ose-oauth-proxy-rhel9@sha256:d38ce4dcac5b073b001d886f0494e0fc9b750f8bd68b855fc103767a1ed8b7df
        - registry.redhat.io/cryostat/cryostat-storage-rhel9@sha256:f642ca8a0ca0aa130462c0f423ce4c775cb296f8a40c9270e5369176e87c7ee6
        - registry.redhat.io/cryostat/cryostat-openshift-console-plugin-rhel9@sha256:ca748f96414ad648eb1e3953aeeeb828b187f3c5c30e014513adc3619a698683
        version: 4.0.3
      entries:
      - name: cryostat-operator.v4.0.3
        version: 4.0.3
      - name: cryostat-operator.v4.0.2-4
        version: 4.0.2+4
      - name: cryostat-operator.v4.0.2
        version: 4.0.2
      - name: cryostat-operator.v4.0.1
        version: 4.0.1
      - name: cryostat-operator.v4.0.1-5
        version: 4.0.1+5
      - name: cryostat-operator.v4.0.1-4
        version: 4.0.1+4
      - name: cryostat-operator.v4.0.1-3
        version: 4.0.1+3
      - name: cryostat-operator.v4.0.1-2
        version: 4.0.1+2
      - name: cryostat-operator.v4.0.0-11
        version: 4.0.0+11
      - name: cryostat-operator.v4.0.0
        version: 4.0.0
      - name: cryostat-operator.v4.0.0-9
        version: 4.0.0+9
      - name: cryostat-operator.v4.0.0-10
        version: 4.0.0+10
      - name: cryostat-operator.v3.0.1-7
        version: 3.0.1+7
      - name: cryostat-operator.v3.0.1-9
        version: 3.0.1+9
      - name: cryostat-operator.v3.0.1-8
        version: 3.0.1+8
      - name: cryostat-operator.v3.0.1
        version: 3.0.1
      - name: cryostat-operator.v3.0.1-6
        version: 3.0.1+6
      - name: cryostat-operator.v3.0.1-5
        version: 3.0.1+5
      - name: cryostat-operator.v3.0.1-3
        version: 3.0.1+3
      - name: cryostat-operator.v3.0.1-13
        version: 3.0.1+13
      - name: cryostat-operator.v3.0.1-12
        version: 3.0.1+12
      - name: cryostat-operator.v3.0.1-11
        version: 3.0.1+11
      - name: cryostat-operator.v3.0.1-10
        version: 3.0.1+10
      - name: cryostat-operator.v3.0.0-12
        version: 3.0.0+12
      - name: cryostat-operator.v3.0.0
        version: 3.0.0
      - name: cryostat-operator.v3.0.0-6
        version: 3.0.0+6
      - name: cryostat-operator.v3.0.0-8
        version: 3.0.0+8
      - name: cryostat-operator.v2.4.0
        version: 2.4.0
      - name: cryostat-operator.v2.4.0-8
        version: 2.4.0+8
      - name: cryostat-operator.v2.4.0-7
        version: 2.4.0+7
      - name: cryostat-operator.v2.4.0-6
        version: 2.4.0+6
      - name: cryostat-operator.v2.4.0-4
        version: 2.4.0+4
      - name: cryostat-operator.v2.4.0-3
        version: 2.4.0+3
      - name: cryostat-operator.v2.3.1-6
        version: 2.3.1+6
      - name: cryostat-operator.v2.3.1-8
        version: 2.3.1+8
      - name: cryostat-operator.v2.3.1-9
        version: 2.3.1+9
      - name: cryostat-operator.v2.3.1
        version: 2.3.1
      - name: cryostat-operator.v2.3.0
        version: 2.3.0
      - name: cryostat-operator.v2.3.0-6
        version: 2.3.0+6
      - name: cryostat-operator.v2.2.1-14
        version: 2.2.1+14
      - name: cryostat-operator.v2.2.1-8
        version: 2.2.1+8
      - name: cryostat-operator.v2.2.1-5
        version: 2.2.1+5
      - name: cryostat-operator.v2.2.1-13
        version: 2.2.1+13
      - name: cryostat-operator.v2.2.1-12
        version: 2.2.1+12
      - name: cryostat-operator.v2.2.1-10
        version: 2.2.1+10
      - name: cryostat-operator.v2.2.1
        version: 2.2.1
      - name: cryostat-operator.v2.2.0
        version: 2.2.0
      - name: cryostat-operator.v2.1.1-2.1659469835.p
        version: 2.1.1+2.1659469835.p
      - name: cryostat-operator.v2.1.1-2.1655334614.p
        version: 2.1.1+2.1655334614.p
      - name: cryostat-operator.v2.1.1
        version: 2.1.1
      - name: cryostat-operator.v2.1.0
        version: 2.1.0
      name: stable
    defaultChannel: stable
    packageName: cryostat-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-xtradb-cluster-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.17.0
        version: 1.17.0
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.16.1
        version: 1.16.1
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.16.0
        version: 1.16.0
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.15.1
        version: 1.15.1
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.15.0
        version: 1.15.0
      name: stable
    - currentCSV: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        version: 1.18.0-cw
      entries:
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0-cw
        version: 1.18.0-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.17.0-cw
        version: 1.17.0-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.16.1-cw
        version: 1.16.1-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.16.0-cw
        version: 1.16.0-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.15.0-cw
        version: 1.15.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator-certified-rhmp
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: community-windows-machine-config-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: community-windows-machine-config-operator.v6.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-windows/community-windows-machine-config-operator:community-4.11-6c3601f
          createdAt: "2022-08-31T18:57:22Z"
          description: An operator that enables Windows container workloads on OCP
          olm.skipRange: '>=5.0.0 <6.0.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ## WARNING
          Community distribution of the Windows Machine Config Operator.
          This is a preview build, and is not meant for production workloads.
          Issues with this distribution of WMCO can be opened against the [WMCO repository](https://github.com/openshift/windows-machine-config-operator).
          Please read through the [troubleshooting doc](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/docs/TROUBLESHOOTING.md)
          before opening an issue.
          Please ensure that when installing this operator the starting CSV you subscribe to is supported on the
          version of OKD/OCP you are using. This CSV is meant for OKD/OCP community-4.11.
          ## Documentation
          ### Introduction
          The Windows Machine Config Operator configures Windows instances into nodes, enabling Windows container workloads
          to be ran within OKD/OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/latest/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/README.md#configuring-byoh-bring-your-own-host-windows-instances).
          The operator will do all the necessary steps to configure the instance so that it can join the cluster as a worker node.
          ### Pre-requisites
          - [Cluster and OS pre-requisites](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/docs/wmco-prerequisites.md)
          ### Usage
          Please see the usage section of [README.md](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/README.md#usage).
          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.
        displayName: Community Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: preview
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-windows/community-windows-machine-config-operator:community-4.11-6c3601f
        version: 6.0.0
      entries:
      - name: community-windows-machine-config-operator.v6.0.0
        version: 6.0.0
      name: preview
    defaultChannel: preview
    packageName: community-windows-machine-config-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom
      provider-url: https://www.broadcom.com/
    name: layer7-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: layer7-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "security.brcmlabs.com/v1",
                "kind": "Gateway",
                "metadata": {
                  "name": "ssg"
                },
                "spec": {
                  "app": {
                    "image": "docker.io/caapim/gateway:11.1.2",
                    "management": {
                      "cluster": {
                        "hostname": "gateway.brcmlabs.com",
                        "password": "7layer"
                      },
                      "password": "7layer",
                      "username": "admin"
                    },
                    "replicas": 1,
                    "service": {
                      "ports": [
                        {
                          "name": "https",
                          "port": 8443,
                          "protocol": "TCP",
                          "targetPort": 8443
                        },
                        {
                          "name": "management",
                          "port": 9443,
                          "protocol": "TCP",
                          "targetPort": 9443
                        }
                      ],
                      "type": "LoadBalancer"
                    }
                  },
                  "license": {
                    "accept": false,
                    "secretName": "gateway-license"
                  },
                  "version": "11.1.2"
                },
                "status": {}
              },
              {
                "apiVersion": "security.brcmlabs.com/v1",
                "kind": "Repository",
                "metadata": {
                  "name": "my-repository"
                },
                "spec": {
                  "auth": {},
                  "branch": "main",
                  "enabled": true,
                  "endpoint": "https://github.com/\u003cusername\u003e/\u003crepository\u003e",
                  "type": "git"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7Api",
                "metadata": {
                  "name": "l7api-sample"
                },
                "spec": {
                  "deploymentTags": [
                    "ssg",
                    "ssg1"
                  ],
                  "graphmanBundle": "jsongz",
                  "portalPublished": false,
                  "serviceUrl": "/v1/api"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7Portal",
                "metadata": {
                  "name": "dev-portal"
                },
                "spec": {
                  "auth": {},
                  "deploymentTags": [
                    "ssg"
                  ],
                  "enabled": true,
                  "endpoint": "dev-ssg.brcmlabs.com",
                  "enrollmentBundle": "",
                  "portalTenant": "portal"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7StateStore",
                "metadata": {
                  "name": "state-store-example"
                },
                "spec": {
                  "redis": {
                    "database": 0,
                    "groupName": "l7gw",
                    "masterPassword": "7layer",
                    "standalone": {
                      "host": "redis-standalone",
                      "port": 6379
                    },
                    "storeId": "test",
                    "type": "standalone"
                  },
                  "type": "redis"
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: docker.io/caapim/layer7-operator:v1.2.1
          createdAt: "2025-07-04T17:22:57Z"
          description: The Layer7 Operator covers all aspects of deploying, maintaining
            and upgrading Layer7 API Gateways in Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: layer7-operator-system
          operators.openshift.io/valid-subscription: Layer7 (Broadcom) API Gateway
            license
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/caapim/layer7-operator
          support: Broadcom Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gateway is the Schema for the Gateway Custom Resource
            displayName: Gateway
            kind: Gateway
            name: gateways.security.brcmlabs.com
            version: v1
          - description: L7Api is the Schema for the l7apis API
            displayName: L7 Api
            kind: L7Api
            name: l7apis.security.brcmlabs.com
            version: v1alpha1
          - description: L7Portal is the Schema for the l7portals API
            displayName: L7 Portal
            kind: L7Portal
            name: l7portals.security.brcmlabs.com
            version: v1alpha1
          - description: L7StateStore is the Schema for the l7statestores API
            displayName: L7 State Store
            kind: L7StateStore
            name: l7statestores.security.brcmlabs.com
            version: v1alpha1
          - description: Repository is the Schema for the repositories API
            displayName: Repository
            kind: Repository
            name: repositories.security.brcmlabs.com
            version: v1
        description: |
          The [Layer7 Operator](https://github.com/CAAPIM/layer7-operator) covers all aspects of deploying, maintaining and upgrading Layer7 API Gateways in Kubernetes. The Layer7 Operator is exclusive to Kubernetes and utilizes the Container Gateway with specific emphasis on running in Ephemeral mode (without an external MySQL policy database).

          ## Getting Started
          This [Getting Started](https://github.com/CAAPIM/layer7-operator/wiki/Getting-Started) guide covers a simple introduction to the Layer7 Operator. You can find more documentation [here](https://github.com/CAAPIM/layer7-operator/wiki).
        displayName: Layer7 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - layer7-operator
        links:
        - name: Layer7 Operator
          url: https://github.com/caapim/layer7-operator
        maintainers:
        - email: gary.vermeulen@broadcom.com
          name: Gary Vermeulen
        maturity: alpha
        provider:
          name: Broadcom
          url: https://www.broadcom.com/
        relatedImages:
        - docker.io/caapim/layer7-operator:v1.2.1
        - docker.io/caapim/gateway:11.1.2
        - docker.io/caapim/graphman-static-init:1.0.3
        version: 1.2.1
      entries:
      - name: layer7-operator.v1.2.1
        version: 1.2.1
      - name: layer7-operator.v1.2.0
        version: 1.2.0
      name: preview
    defaultChannel: preview
    packageName: layer7-operator
    provider:
      name: Broadcom
      url: https://www.broadcom.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: eG Innovations
      provider-url: https://www.eginnovations.com
    name: eginnovations-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eginnovations-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eginnovations.com/v1beta1",
                "kind": "EgUniversalAgent",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eginnovations-operator"
                  },
                  "name": "eguniversalagent",
                  "namespace": "egagent"
                },
                "spec": {
                  "agent": {
                    "env": [
                      {
                        "name": "EG_MANAGER",
                        "value": "Replace with eG Manager IP"
                      },
                      {
                        "name": "EG_MANAGER_PORT",
                        "value": "Replace with eG Manager Port"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "true"
          containerImage: docker.io/eginnovations/universal-agent-operator@sha256:b8e3e26dca1b94fabfd79496feb36f3f2de4434331ef9ca810125892a936dca7
          createdAt: "2025-09-11T20:08:00Z"
          description: eG Innovations provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with eG Universal Agent Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eginnovations/universal-agent-operator
          support: eG Innovations
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgUniversalAgent is the Schema for the eguniversalagents
              API
            displayName: Eg Universal Agent
            kind: EgUniversalAgent
            name: eguniversalagents.eginnovations.com
            version: v1beta1
        description: |
          eG Innovations’ converged application and infrastructure performance monitoring capabilities provide end-to-end visibility and correlated analytics for every layer, every tier of your IT landscape.

          The eG Enterprise Universal Agent Operator automatically configures a host agent on every Kubernetes worker node. The host agent auto-discovers the worker nodes and application containers running as Pods on each node and tracks their performance and utilization levels. In-depth monitoring of applications running on containers is also provided using the same host agent. No additional agents are required for the containers.

          ### Capabilities
            * Rolls out the egagent pod per node to monitor its pods and the node itself
            * Automatic BTM profiler injection using webhook based injection for deep application monitoring and transection tracing.

          ### Installation
            1. Create the namespace "egagent" for installing eG Agent on OpenShift Cluster.

               <br />``` $ oc create ns egagent ```
            2. Add "egagent" serviceaccount to scc( Security Context Constraints ) of OpenShift Cluster as privileged using below command.

               <br />``` $ oc adm policy add-scc-to-user privileged -z egagent -n egagent ```
            2. Install eG Innovations Universal Agent operator from [OpenShift Container Platform](https://www.openshift.com/) OperatorHub.
            3. Create the eG Universal Agent CRD (Custom Resource Definition)
               a. Click "View YAML Example" and copy the yaml contents and save it to (filename).yaml on your host.
               b. To apply the CRD run the following command

               <br />``` $ oc apply -f (filename).yaml ```

          ### Required Parameters
            * EG\_MANAGER, EG\_MANAGER\_PORT - provide the ip and port of your eG Manager Installation.

          ### Advanced Options
            * Disable Certificate Checking - disable any certificate validation that may interact poorly with proxies within your cluster
            * Image Override - use a copy of the eG Iniversal Agent container image from a registry other that Docker's or Red Hat's
            * Environment variables - define environment variables for the eG Universal Agent container of for APM profiler.
            * Namespace Selectors- select a subset of namesapces to enable instrumentation for applications.

          For complete list of supported parameters and the detailed documentation available in [eG Innovations Docs](https://docs.eginnovations.com)
        displayName: eG Innovations Universal Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Tracing
        links:
        - name: eG Innovations Kubernetes Monitoring Overview
          url: https://www.eginnovations.com/documentation/Kubernetes/Introduction-to-Monitoring-Kubernetes.htm
        maintainers:
        - email: devops@eginnovations.com
          name: eG Innovations
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: eG Innovations
          url: https://www.eginnovations.com
        relatedImages:
        - docker.io/eginnovations/universal-agent-operator@sha256:b8e3e26dca1b94fabfd79496feb36f3f2de4434331ef9ca810125892a936dca7
        version: 0.0.11
      entries:
      - name: eginnovations-operator.v0.0.11
        version: 0.0.11
      name: beta
    defaultChannel: beta
    packageName: eginnovations-operator
    provider:
      name: eG Innovations
      url: https://www.eginnovations.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-trino-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trino-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:904eb8ac055171c75c8c86e4923df85461b81e6d5d9269d9cddfb03fc615f9f4
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:904eb8ac055171c75c8c86e4923df85461b81e6d5d9269d9cddfb03fc615f9f4
        version: 24.11.1
      entries:
      - name: trino-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: trino-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b50b247ed8f0ad262ed59e8aafd099919e98156472e5663957f9922616140bd9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b50b247ed8f0ad262ed59e8aafd099919e98156472e5663957f9922616140bd9
        version: 25.3.0
      entries:
      - name: trino-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: trino-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
        version: 25.7.0
      entries:
      - name: trino-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: trino-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
        version: 25.7.0
      entries:
      - name: trino-operator.v25.7.0
        version: 25.7.0
      - name: trino-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-trino-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-kafka-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kafka-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:f76b5558088bead829be8018a89e3efc1bb7a84d40429a0a1df1c88552db169b
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:f76b5558088bead829be8018a89e3efc1bb7a84d40429a0a1df1c88552db169b
        version: 24.11.1
      entries:
      - name: kafka-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: kafka-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:72b390c2d382af0a9cbe2d9ae394fc61c62b9a3f6c7310cf5f325d0f17257f1f
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:72b390c2d382af0a9cbe2d9ae394fc61c62b9a3f6c7310cf5f325d0f17257f1f
        version: 25.3.0
      entries:
      - name: kafka-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: kafka-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
        version: 25.7.0
      entries:
      - name: kafka-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: kafka-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
        version: 25.7.0
      entries:
      - name: kafka-operator.v25.7.0
        version: 25.7.0
      - name: kafka-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-kafka-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: vfunction-operator
      app.kubernetes.io/managed-by: operator
      app.kubernetes.io/name: vfunction
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      name: vfunction-server-operator
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: vFunction
      provider-url: ""
    name: vfunction-server-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vfunction-server-operator.v2.2.469
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vfunction.com/v1",
                "kind": "VfunctionServer",
                "metadata": {
                  "name": "example-vfunction-server",
                  "namespace": "vfunction",
                  "labels": {
                      "name": "vfunction",
                      "app.kubernetes.io/name": "vfunction",
                      "app.kubernetes.io/instance": "vfunction-server",
                      "app.kubernetes.io/managed-by": "operator"
                  }
                },
                "spec": {
                  "admin": {
                    "email": "admin@mycompany.com",
                    "name": "Admin",
                    "password": "Password1!"
                  },
                  "host": "http://my.domain.com",
                  "org_name": "MyCompany",
                  "upgrade": "Daily",
                  "measurements": {
                    "auto_scaling": "No",
                    "min_number_of_services": "2",
                    "max_number_of_services": "10"
                  },
                  "smtp": {
                    "identity": "",
                    "password": "",
                    "url": "",
                    "user": ""
                  },
                  "tls": {
                    "crt": "",
                    "key": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, AI/Machine Learning
          certified: Red Hat certified
          containerImage: registry.marketplace.redhat.com/rhm/vfunction/vfunction-server-operator@sha256:60d1965460101d0fd90ecc9499d0a6d4c9684a54ada61a43a253d30661fab22f
          createdAt: "2021-08-03T19:08:54Z"
          description: vFunction is a cutting-edge code analysis, machine learning,
            and automation to boost your Java modernization projects.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/vfunction-server-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/vfunction-server-operator-rhmp/support?utm_source=openshift_console
          support: vFunction
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of vFunction running on your cluster.
            displayName: vFunction
            kind: VfunctionServer
            name: vfunctionservers.vfunction.com
            version: v1
        description: |2-

          ## Introduction

          [vFunction](https://www.vfunction.com/) is a cloud-native modernization platform that combines dynamic and static code analysis, machine learning, and automation to automatically identify and extract services from existing applications. vFunction is the only platform purpose-built for modernization of Java applications.

          The vFunction modernization process starts by learning the running monolithic application, and surfacing the interdependencies within it. Using AI, the platform analyzes and identifies services that can be separated from the application. This decomposition can present a range of micro, mini, or even macro services, depending on your application environment, each being an independently deployable and scalable application component.

          vFunction automates the extraction of these services, enabling you to modernize your monolith, quickly and easily.

          ## The vFunction platform

          The platform consists of 3 basic components; the server, the controller package, and a tools package. The server runs as an operand on an OpenShift environment. The controller package is installed on the machine that runs the monolithic application which can be either a Linux or a Windows machine, and the tools are run on a development machine, with access to the code of the monolithic application.

          The controller package consists of three elements: the vFunction agent, that collects data during the dynamic analysis phase; the vFunction Viper application, that performs static analysis on the binaries of the application; and the vFunction controller that handles all the communication between the agent, Viper, and the vFunction server.

          The vFunction agent is a mix of a Java and native agent, and needs to run on the JVM that is currently running your application. [Refer to the vFunction OOB Support Matrix document](https://drive.google.com/file/d/1ccq8LFab1FrYAimDUxwgjiCOdk4QuVzs/view) for a list of supported application servers and JVMs.

          This operator installs a vFunction server instance to be connected later with one or more vFunction controllers installed on your application machine(s).

          ## Before You Start

          The vFunction operator requires a Red Hat OpenShift Kubernetes Application Platform 4.x.

          The cluster recommended configuration (which is the same as the minimal one) should consist of:

          - *1 master node*

          - *1 worker node*

          - *Storage capacity enough for dynamically provision 2 PVs of 50G each*

          - *A default StorageClass configured*

          vFunction doesn’t provide any encryption mechanism for data stored on any attached storage. If required, encryption can be achieved by the client by using its own external means on the storage itself.

          vFunction supports backup of all critical data as part of the application. Nevertheless, for complete backup of entire data it is recommended that the client apply its own storage based backup mechanism.

          ## Install Prerequisites

          ###Cluster:
          The operator was certified and tested on OCP 4.6 and 4.7.
          If you encounter any issue with other OCP versions or cloud providers please contact info@vfunction.com.

          ###Storage:
          The operator creates two new PersistentVolumeClaims (PVCs) during installation time. For their creation, it relies on the default StorageClass to dynamically provision two PersistentVolumes (PVs). The system must have a default StorageClass in place.

          Both PersistentVolumes are accessed with ReadWriteOnce mode.

          ###Project:
          The operator should be installed in a new and dedicated project (namespace). If you intend to install more than one vFunction operands in the same cluster, please use different projects for each one. Installation of more than one operand on the same cluster will allow scaling by load balancing multiple applications to different vFunction servers.

          ## Installation instructions

          1. Prepare the YAML file to use in the installation:

              a. Copy the following YAML template into a text editor:

              ```yaml
                  apiVersion: vfunction.com/v1
                  kind: VfunctionServer
                  metadata:
                    name: vfunction
                    namespace: vfunction
                    labels:
                      name: vfunction
                      app.kubernetes.io/name: vfunction
                      app.kubernetes.io/instance: vfunction-server
                      app.kubernetes.io/managed-by: operator
                  spec:
                    host: "http://my.domain.com"
                    org_name: "MyCompany"
                    upgrade: "Daily"
                    admin:
                      email: "admin@mycompany.com"
                      name: "Admin"
                      password: "Password1!"
                    smtp:
                      password: ""
                      url: ""
                      identity: ""
                      user: ""
                    measurements:
                      auto_scaling: "No"
                      min_number_of_services: "2"
                      max_number_of_services: "10"
                    tls:
                      crt: |
                        -----BEGIN CERTIFICATE-----
                        ...
                        -----END CERTIFICATE-----
                      key: |
                        -----BEGIN PRIVATE KEY-----
                        ...
                        -----END PRIVATE KEY-----
              ```
              b. Customize the template fields by replacing with your information:

              - **host**: Enter the FQDN for accessing the vFunction dashboard.
                          Make sure that you use a domain name and not an IP address. Writing "https://" will mandate TLS while writing "http://" will expose the server through HTTP.
              - **org_name**: Enter your organization name.
              - **upgrade**: Choose auto-upgrade mode
                  - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically (default).
                  - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
                  - **Never**: The automatic upgrading mechanism is off.
                  - **Maintenance**: Put the operator in a maintenance mode.

              - **admin.email**: Enter the email address of your administrator.
              - **admin.name**: Enter the name of your administrator.
              - **admin.password**: Enter the password you want to use for the vFunction administrator.
                                    The password should be at least 8 characters long, and consist of at least one lowercase letter, at least one uppercase letter, at least one number, and at least one special character.
              - **smtp.user**: Enter the email address for a designated user for the SMTP server (optional).
              - **smtp.password**: Enter this user’s password (optional).
              - **smtp.identity**: Enter this user’s password (optional).
              - **smtp.url**: Enter the SMTP server URL (optional).
              - **tls.crt**: If your host FQDN starts with “https”, paste in the certifications you have for using the TLS connection.
              - **tls.key**: If your host FQDN starts with “https”, enter the key you have for using the TLS connection.
              - **measurements.auto_scaling**: Indicates if the measurement services auto scalling is active.
              - **measurements.min_num_of_services**: Minimum Number of measurement service pods.
              - **measurements.max_num_of_services**: Maximum Number of measurement service pods.
              - **measurements.S3**: Save all measurements data to S3, instead of local PV (optional).
                  - **measurements.S3.bucket**: Measurements S3 bucket name.
                  - **measurements.S3.key**: Measurements S3 key.
                  - **measurements.S3.secret**: Measurements S3 secret.
                  - **measurements.S3.region**: Measurements S3 region.
              - **backup.mysql.S3**: Backup MySQL to S3 (optional).
                  - **backup.mysql.S3.bucket**: MySQL backup S3 bucket name.
                  - **backup.mysql.S3.key**: MySQL backup S3 key.
                  - **backup.mysql.S3.secret**: MySQL backup S3 secret.
                  - **backup.mysql.S3.region**: MySQL backup S3 region.

          1. In the Installed Operators window choose and click on vFunction Operator and then Create Instance.
          1. In the Create vFunctionServer window, paste the YAML file that you created in step 1.
          1. Click Create.
          1. You can now click on your new created vFunction operand in order to see its details and installation progress.

          ## Verifying the installation

          Check that the operand installed successfully:

          1. Check that the "Operand State" property shows “Working”.
          1. Check that the "Successfully Installed" property shows "Yes".

          If "Operand State" shows "Failed", it indicates the operand wasn’t installed correctly. Check the events and logs for all pods (operator and image containers) for any issues. In the event that you cannot troubleshoot, contact vFunction support.

          ## After installation

          The vFunction site is now accessible via the newly created vFunction application custom address, for example, my.domain.com. There are two ways you can access the vFunction dashboard.

          ### Access using the router canonical hostname

          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address, and to the ”vfunction” subdomain of the  OpenShift canonical router hostname as the alias.
          For example, ***my.domain.com.   CNAME   vfunction.apps.ocp4.my-openshift.com.***
          2. Find your cluster Router Canonical Hostname address in the newly created vfunction-route-xxx under your vFunction OpenShift project -> **Networking** -> **Routes** -> **vfunction-route-xxx** route -> **Router Canonical Hostname** field.
          You can now access the dashboard using your defined "host" spec property (as above).

          ### Access using the nginx service location

          You may use this access method if your OpenShift is installed on a provider that supports exposing LoadBalancer-type services.
          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address and to the vfunction-nginx-xxx service location.
          For example, ***my.domain.com.   CNAME   a05951ed7cdf-1394239323.us-east-1.elb.amazonaws.com.***
          1. Point your custom domain to the vfunction-nginx-xxx service's external IP location, which you can find in the OpenShift project > **Networking** -> **Services** -> **vfunction-nginx-xxx** service > **Service Address** > **Location** field.

          ## Upgrade and Rollback

          The vFunction operator includes a built-in auto-upgrade mechanism.

          You can choose one of 4 upgrading modes:
          - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically (default).
          - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
          - **Never**: The automatic upgrading mechanism is off.
          - **Maintenance**: Put the operator in a maintenance mode.

          For an on-demand upgrade,  you can change anytime the upgrading mode from “Never” to “Always”, wait for the operand to be upgraded, and change again to “Never”.

          During the upgrade, the "Operand State" property will change to “Upgrading...” and back again to “Working” after a successful upgrade.

          In case of a failure upgrading the operand, an automatic rollback to the last working version will occur. The failed version will be marked as defective and the operator will not try to upgrade to it again.

          Choosing the "Maintenance" option will tell the operator to take down all pods, so you can safely fix storage issues, return volumes from snapshots, etc.
          After finishing the maintenance, return the upgrade mode to your original desired policy.
        displayName: vFunction Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - vfunction
        - analyzing
        - services
        links:
        - name: vFunction
          url: https://vfunction.com
        - name: Terms
          url: https://www.vfunction.com/eula/
        maintainers:
        - email: raviv@vfunction.com
          name: Raviv Halivni
        maturity: alpha
        provider:
          name: vFunction
        relatedImages:
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-organizations@sha256:bcc2c3d6222af95e614318daea3108fffc6bc77aa17b23031cce296a3811b79f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-users@sha256:e47cc051953976a0e075e8cfb457cd4749e9b9e4394f2fbecab2458c2ade92ca
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-mysql@sha256:a6e06ea08fffa33b709d4d1240587bb3ab5b4cc6a28f493a227555e88e645798
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-nginx@sha256:435d1489aaf29a811081ddb51f40ae692e5833f42c31dbced6bc189c0ddc40ab
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-storage@sha256:b778cb9c724a59e51b4cb9088047be471cdf338f1138c31b81f79e50a491203f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-idp@sha256:21ca95a0209f3c557a45658452d154777f928539a93129148470562c79e68d70
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-measurements@sha256:35ffbee4b0eead58706b23cf5a702b1d55d2a7913cd0019cefbff94957d4e5d6
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-server-operator@sha256:60d1965460101d0fd90ecc9499d0a6d4c9684a54ada61a43a253d30661fab22f
        version: 2.2.469
      entries:
      - name: vfunction-server-operator.v2.2.469
        version: 2.2.469
      name: stable
    defaultChannel: stable
    packageName: vfunction-server-operator-rhmp
    provider:
      name: vFunction
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CNCF
      provider-url: ""
    name: jaeger
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jaeger-operator.v1.65.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "jaegertracing.io/v1",
                "kind": "Jaeger",
                "metadata": {
                  "name": "my-jaeger"
                },
                "spec": {
                  "strategy": "allInOne"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing
          certified: "false"
          containerImage: quay.io/jaegertracing/jaeger-operator:1.65.0
          createdAt: "2025-01-22T20:40:19Z"
          description: Provides tracing, monitoring and troubleshooting for microservices-based
            distributed systems
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/jaegertracing/jaeger-operator
          support: Jaeger Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Jaeger is the Schema for the jaegers API
            displayName: Jaeger
            kind: Jaeger
            name: jaegers.jaegertracing.io
            version: v1
        description: |-
          Jaeger, inspired by [Dapper](https://research.google.com/pubs/pub36356.html) and [OpenZipkin](http://zipkin.io/), is a distributed tracing system released as open source by Uber Technologies. It is used for monitoring and troubleshooting microservices-based distributed systems.
          ### Core capabilities
          Jaeger is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed context propagation
          * Distributed transaction monitoring
          * Root cause analysis
          * Service dependency analysis
          * Performance / latency optimization
          * OpenTracing compatible data model
          * Multiple storage backends: Badger, Cassandra, Elasticsearch, Memory.
          ### Operator features
          * **Multiple modes** - Supports `allInOne`, `production` and `streaming` [modes of deployment](https://www.jaegertracing.io/docs/latest/operator/#deployment-strategies).
          * **Configuration** - The Operator manages [configuration information](https://www.jaegertracing.io/docs/latest/operator/#configuring-the-custom-resource) when installing Jaeger instances.
          * **Storage** - [Configure storage](https://www.jaegertracing.io/docs/latest/operator/#storage-options) used by Jaeger. By default, `memory` is used. Other options include `badger`, `cassandra` or `elasticsearch`. On OpenShift, the operator can delegate creation of an Elasticsearch cluster to the Elasticsearch Operator if deployed.
          * **Agent** - can be deployed as [sidecar](https://www.jaegertracing.io/docs/latest/operator/#auto-injecting-jaeger-agent-sidecars) (default) and/or [daemonset](https://www.jaegertracing.io/docs/latest/operator/#installing-the-agent-as-daemonset).
          * **UI** - Optionally setup ingress (Kubernetes) or secure route (OpenShift) to provide [access to the Jaeger UI](https://www.jaegertracing.io/docs/latest/operator/#accessing-the-jaeger-console-ui).
          ### Before you start
          1. Ensure that the appropriate storage solution, that will be used by the Jaeger instance, is available and configured.
          2. If intending to deploy an Elasticsearch cluster via the Jaeger custom resource, then the Elasticsearch Operator must first be installed.
          ### Troubleshooting
          * https://www.jaegertracing.io/docs/latest/troubleshooting/
        displayName: Community Jaeger Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tracing
        - monitoring
        - troubleshooting
        links:
        - name: Jaeger Operator Source Code
          url: https://github.com/jaegertracing/jaeger-operator
        maintainers:
        - email: jaeger-tracing@googlegroups.com
          name: Jaeger Google Group
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: CNCF
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/jaegertracing/jaeger-operator:1.65.0
        version: 1.65.0
      entries:
      - name: jaeger-operator.v1.65.0
        version: 1.65.0
      - name: jaeger-operator.v1.62.0
        version: 1.62.0
      - name: jaeger-operator.v1.61.0
        version: 1.61.0
      - name: jaeger-operator.v1.60.0
        version: 1.60.0
      - name: jaeger-operator.v1.59.0
        version: 1.59.0
      - name: jaeger-operator.v1.57.0
        version: 1.57.0
      - name: jaeger-operator.v1.56.0
        version: 1.56.0
      - name: jaeger-operator.v1.55.0
        version: 1.55.0
      - name: jaeger-operator.v1.53.0
        version: 1.53.0
      - name: jaeger-operator.v1.52.0
        version: 1.52.0
      - name: jaeger-operator.v1.51.0
        version: 1.51.0
      - name: jaeger-operator.v1.49.0
        version: 1.49.0
      - name: jaeger-operator.v1.48.1
        version: 1.48.1
      - name: jaeger-operator.v1.47.0
        version: 1.47.0
      - name: jaeger-operator.v1.46.0
        version: 1.46.0
      - name: jaeger-operator.v1.45.0
        version: 1.45.0
      - name: jaeger-operator.v1.44.0
        version: 1.44.0
      - name: jaeger-operator.v1.43.0
        version: 1.43.0
      - name: jaeger-operator.v1.42.0
        version: 1.42.0
      - name: jaeger-operator.v1.41.1
        version: 1.41.1
      - name: jaeger-operator.v1.41.0
        version: 1.41.0
      - name: jaeger-operator.v1.40.0
        version: 1.40.0
      - name: jaeger-operator.v1.39.0
        version: 1.39.0
      - name: jaeger-operator.v1.38.1
        version: 1.38.1
      - name: jaeger-operator.v1.38.0
        version: 1.38.0
      - name: jaeger-operator.v1.37.0
        version: 1.37.0
      - name: jaeger-operator.v1.36.0
        version: 1.36.0
      - name: jaeger-operator.v1.35.0
        version: 1.35.0
      - name: jaeger-operator.v1.34.1
        version: 1.34.1
      - name: jaeger-operator.v1.34.0
        version: 1.34.0
      - name: jaeger-operator.v1.33.0
        version: 1.33.0
      - name: jaeger-operator.v1.32.0
        version: 1.32.0
      - name: jaeger-operator.v1.31.0
        version: 1.31.0
      - name: jaeger-operator.v1.30.0
        version: 1.30.0
      - name: jaeger-operator.v1.29.1
        version: 1.29.1
      - name: jaeger-operator.v1.25.0
        version: 1.25.0
      - name: jaeger-operator.v1.24.0
        version: 1.24.0
      name: stable
    defaultChannel: stable
    packageName: jaeger
    provider:
      name: CNCF
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: JWS team
      provider-url: ""
    name: jws-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jws-operator.v2.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "web.servers.org/v1alpha1",
                "kind": "WebServer",
                "metadata": {
                  "name": "webserver-example"
                },
                "spec": {
                  "applicationName": "webapp",
                  "replicas": 1,
                  "useInsightsClient": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/jboss-webserver-5/jws5-rhel8-operator@sha256:ea042277bdcdb70b2dc32083ea868efd4bf474d5a8ec0f9b47f5bea35aaeae1e
          createdAt: "2025-09-12T08:31:52.558719Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat JBoss Web Server",
            "Red Hat Runtimes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/web-servers/jws-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Web Server is the schema for the webservers API
            displayName: Web Server
            kind: WebServer
            name: webservers.web.servers.org
            version: v1alpha1
        description: WebServer operator can manage JBoss Web Server images.
        displayName: WebServer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WebServer
        - JBoss
        links:
        - name: Jws Operator
          url: https://jws-operator.domain
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: JWS team
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:021ff671b0fb72eed635b05489ec86342fcb7ada3ba0353591fdb9328871d067
        - registry.redhat.io/jboss-webserver-5/jws5-rhel8-operator@sha256:ea042277bdcdb70b2dc32083ea868efd4bf474d5a8ec0f9b47f5bea35aaeae1e
        - registry.redhat.io/insights-runtimes-tech-preview/runtimes-inventory-rhel8-operator@sha256:58e01d11c3d30abc5b0ee3795511bd04e39aea21e2853b4f4cc6d1c8a2bd7bb0
        version: 2.2.0
      entries:
      - name: jws-operator.v2.2.0
        version: 2.2.0
      - name: jws-operator.v2.1.3
        version: 2.1.3
      - name: jws-operator.v2.1.2
        version: 2.1.2
      - name: jws-operator.v2.1.1
        version: 2.1.1
      - name: jws-operator.v2.1.0
        version: 2.1.0
      - name: jws-operator.v2.0.16
        version: 2.0.16
      - name: jws-operator.v2.0.15
        version: 2.0.15
      - name: jws-operator.v2.0.14
        version: 2.0.14
      - name: jws-operator.v2.0.13
        version: 2.0.13
      - name: jws-operator.v2.0.12
        version: 2.0.12
      - name: jws-operator.v2.0.11
        version: 2.0.11
      - name: jws-operator.v2.0.10
        version: 2.0.10
      - name: jws-operator.v2.0.9
        version: 2.0.9
      - name: jws-operator.v2.0.8
        version: 2.0.8
      - name: jws-operator.v2.0.7
        version: 2.0.7
      - name: jws-operator.v2.0.6
        version: 2.0.6
      - name: jws-operator.v2.0.5
        version: 2.0.5
      - name: jws-operator.v2.0.4
        version: 2.0.4
      - name: jws-operator.v2.0.3
        version: 2.0.3
      - name: jws-operator.v2.0.2
        version: 2.0.2
      - name: jws-operator.v2.0.1
        version: 2.0.1
      name: alpha
    defaultChannel: alpha
    packageName: jws-operator
    provider:
      name: JWS team
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Windup
      provider-url: ""
    name: windup-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: windup-operator.v0.0.25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
             {
               "apiVersion": "windup.jboss.org/v1",
               "kind": "Windup",
               "metadata": {
                 "name": "windup",
                 "labels": {
                   "application": "windup"
                 }
               },
               "spec": {}
             }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/windupeng/windup-operator-native:0.0.25
          createdAt: 01/22/24
          description: Windup is an analysis tool that supports the modernization
            and migration of Java applications.
          imagePullPolicy: Always
          operatorframework.io/suggested-namespace: windup
          repository: https://github.com/windup/windup-operator/
          support: https://issues.redhat.com/projects/WINDUP
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "Create a Windup Web Console application.  \n\nTo access
              the Windup Web Console application from the \"Developer\" perspective,
              please go to Topology and click on the \"Open URL\" icon on the web-console
              pod.  \n\nAlternatively from the \"Administrator\" perspective, please
              go to Networking->Routes and click on the Location hyperlink.  \n\nKnown
              issue  \nIf you want to customize the Windup Web Console instance's
              parameters and you can not see any in the `Form view`, please switch
              to the `YAML view` and change them as needed.\n"
            displayName: Windup
            kind: Windup
            name: windups.windup.jboss.org
            version: v1
        description: |
          Windup is a web console application that supports large-scale Java application modernization and migration projects across a broad range of transformations and use cases.

          It analyzes application code, supports effort estimation, accelerates code migration, and enables users to move applications to containers.

          For more information please refer to the [https://windup.github.io/](https://windup.github.io/) page.
        displayName: Windup Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - modernization
        - windup
        links:
        - name: Website
          url: https://windup.github.io
        - name: GitHub
          url: https://github.com/windup/windup-operator
        maintainers:
        - email: migrate@redhat.com
          name: Windup
        maturity: alpha
        provider:
          name: Windup
        relatedImages:
        - quay.io/windupeng/windup-operator-native:0.0.25
        version: 0.0.25
      entries:
      - name: windup-operator.v0.0.25
        version: 0.0.25
      - name: windup-operator.v0.0.24
        version: 0.0.24
      - name: windup-operator.v0.0.23
        version: 0.0.23
      - name: windup-operator.v0.0.22
        version: 0.0.22
      - name: windup-operator.v0.0.21
        version: 0.0.21
      - name: windup-operator.v0.0.20
        version: 0.0.20
      - name: windup-operator.v0.0.19
        version: 0.0.19
      - name: windup-operator.v0.0.18
        version: 0.0.18
      - name: windup-operator.v0.0.17
        version: 0.0.17
      - name: windup-operator.v0.0.16
        version: 0.0.16
      - name: windup-operator.v0.0.15
        version: 0.0.15
      - name: windup-operator.v0.0.14
        version: 0.0.14
      - name: windup-operator.v0.0.13
        version: 0.0.13
      - name: windup-operator.v0.0.12
        version: 0.0.12
      - name: windup-operator.v0.0.11
        version: 0.0.11
      name: alpha
    defaultChannel: alpha
    packageName: windup-operator
    provider:
      name: Windup
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stolostron
      provider-url: ""
    name: cluster-aas-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-aas-operator.v0.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplate",
                "metadata": {
                  "name": "clustertemplate-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplateInstance",
                "metadata": {
                  "name": "clustertemplateinstance-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplateQuota",
                "metadata": {
                  "name": "clustertemplatequota-sample"
                },
                "spec": {}
              }
            ]
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/cluster-templates-operator:2.11.0-0c3131154065397d37e3b66ae635d7ea88650bf9
          createdAt: "2024-06-06T19:34:44Z"
          description: Easily install fully configured clusters with guard-rails.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/cluster-templates-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents instance of a cluster
            displayName: Cluster template instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.clustertemplate.openshift.io
            version: v1alpha1
          - description: Defines which ClusterTemplates can be used in a given namespace
            displayName: Cluster template quota
            kind: ClusterTemplateQuota
            name: clustertemplatequotas.clustertemplate.openshift.io
            version: v1alpha1
          - description: Template of a cluster - both installation and post-install
              setup are defined as ArgoCD application spec. Any application source
              is supported - typically a Helm chart
            displayName: Cluster template
            kind: ClusterTemplate
            name: clustertemplates.clustertemplate.openshift.io
            version: v1alpha1
          - kind: ClusterTemplateSetup
            name: clustertemplatesetup.clustertemplate.openshift.io
            version: v1alpha1
          - kind: Config
            name: config.clustertemplate.openshift.io
            version: v1alpha1
        description: |
          **Self-service clusters with guardrails!**
          Cluster as a service operator provides an easy way to define clusters as templates and allows creating instances of those templates even for non-privileged developer/devops engineers. Cluster templates operator also allows specifing quotas for the developer/devops engineers.

          ### User Interface
          The User interface is installed but is not enabled by default. To enable go to `Home` -> `Overview` -> in the status card click `Dynamic Plugins` -> click `View all` -> click the `pencil` icon -> pick `Enable` and hit `Save`. You can access the UI in the `All Clusters` perspective under `Infrastructure` -> `Cluster Templates`.

          ### Documentation
          Documentation can be found on our [website](https://github.com/stolostron/cluster-templates-operator).

          ### Getting help
          To report any issues or ask questions, create an issue on our [Github repo](https://github.com/stolostron/cluster-templates-operator/issues)
        displayName: Cluster as a service operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cluster
        - templates
        - hive
        - hypershift
        - selfservice
        - openshift
        - multicluster
        - quota
        - helm
        - argocd
        links:
        - name: Cluster as a service operator
          url: https://github.com/stolostron/cluster-templates-operator
        maintainers:
        - email: rawagner@redhat.com
          name: Rastislav Wagner
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Stolostron
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/stolostron/cluster-templates-operator:2.11.0-0c3131154065397d37e3b66ae635d7ea88650bf9
        version: 0.1.5
      entries:
      - name: cluster-aas-operator.v0.1.5
        version: 0.1.5
      - name: cluster-aas-operator.v0.1.4
        version: 0.1.4
      - name: cluster-aas-operator.v0.1.3
        version: 0.1.3
      - name: cluster-aas-operator.v0.1.2
        version: 0.1.2
      - name: cluster-aas-operator.v0.1.1
        version: 0.1.1
      - name: cluster-aas-operator.v0.1.0
        version: 0.1.0
      - name: cluster-aas-operator.v0.0.6
        version: 0.0.6
      - name: cluster-aas-operator.v0.0.5
        version: 0.0.5
      - name: cluster-aas-operator.v0.0.4
        version: 0.0.4
      - name: cluster-aas-operator.v0.0.3
        version: 0.0.3
      - name: cluster-aas-operator.v0.0.2
        version: 0.0.2
      - name: cluster-aas-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cluster-aas-operator
    provider:
      name: Stolostron
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: ibm-spectrum-scale-csi-operator
      app.kubernetes.io/managed-by: ibm-spectrum-scale-csi-operator
      app.kubernetes.io/name: ibm-spectrum-scale-csi-operator
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-spectrum-scale-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-spectrum-scale-csi-operator.v2.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "CSIScaleOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-spectrum-scale-csi-operator",
                    "app.kubernetes.io/managed-by": "ibm-spectrum-scale-csi-operator",
                    "app.kubernetes.io/name": "ibm-spectrum-scale-csi-operator",
                    "release": "ibm-spectrum-scale-csi-operator"
                  },
                  "name": "ibm-spectrum-scale-csi",
                  "namespace": "ibm-spectrum-scale-csi-driver"
                },
                "spec": {
                  "attacherNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "clusters": [
                    {
                      "id": "\u003cPrimary Cluster ID\u003e",
                      "primary": {
                        "primaryFs": "\u003c Primary Filesystem \u003e"
                      },
                      "restApi": [
                        {
                          "guiHost": "\u003c Primary cluster GUI IP/Hostname \u003e"
                        }
                      ],
                      "secrets": "secret1",
                      "secureSslMode": false
                    }
                  ],
                  "pluginNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "provisionerNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "resizerNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "snapshotterNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ]
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/ibm-spectrum-scale/ibm-spectrum-scale-csi-operator@sha256:b62ad588fbc338090f405740fd8bceaf59b3267cdb532c60ae965643ccae4d5d
          createdAt: Thu May 5 04:42:15 IST 2022
          description: An operator for deploying and managing the IBM Spectrum Scale
            CSI Driver.
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-spectrum-scale-csi/
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIScaleOperator is the Schema for the csiscaleoperators
              API
            displayName: IBM CSI Spectrum Scale Driver
            kind: CSIScaleOperator
            name: csiscaleoperators.csi.ibm.com
            version: v1
        description: |
          The IBM Spectrum Scale CSI Operator for Kubernetes installs, manages,
          upgrades the IBM Spectrum Scale CSI Driver on OpenShift and Kubernetes
          for provisioning of persistent volumes on IBM Spectrum Scale.

          ### Documentation Documentation for IBM Spectrum Scale CSI Driver and Operator can be found at [IBM Documentation](https://www.ibm.com/docs/en/spectrum-scale-csi).

          ### Getting help
          IBM Spectrum Scale CSI driver is part of the IBM Spectrum Scale offering.
          Please follow the [IBM support procedure](https://www.ibm.com/mysupport/s/?language=en_US) for any issues with the driver.

          For non-urgent issues, suggestions, recommendations feel free to open an
          issue in [github](https://github.com/IBM/ibm-spectrum-scale-csi). Issues will be addressed as team availability permits.

          **Note:** Custom Resource example represents the minimum settings needed to run
          the operator. If your environment needs more advanced settings (e.g. remote clusters,
          node mapping, image pull secret, Taint and Toleration etc.) please refer to  the sample [Custom Resource](https://raw.githubusercontent.com/IBM/ibm-spectrum-scale-csi/v2.6.0/operator/config/samples/csiscaleoperators.csi.ibm.com_cr.yaml).
        displayName: IBM Spectrum Scale CSI Plugin Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - GPFS
        - Spectrum Scale
        - CSI Driver
        links:
        - name: IBM CSI Spectrum Scale Documentation
          url: https://www.ibm.com/docs/en/spectrum-scale-csi
        - name: CSI Developer Documentation
          url: https://kubernetes-csi.github.io/docs/
        maintainers:
        - email: madhu.punjabi@in.ibm.com
          name: Madhu Thorat
        - email: amdabhad@in.ibm.com
          name: Amol Dabhade
        - email: Nitish.Kumar4@ibm.com
          name: Nitish Kumar
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibm-spectrum-scale/ibm-spectrum-scale-csi-operator@sha256:b62ad588fbc338090f405740fd8bceaf59b3267cdb532c60ae965643ccae4d5d
        version: 2.6.0
      entries:
      - name: ibm-spectrum-scale-csi-operator.v2.6.0
        version: 2.6.0
      - name: ibm-spectrum-scale-csi-operator.v2.5.1
        version: 2.5.1
      - name: ibm-spectrum-scale-csi-operator.v2.5.0
        version: 2.5.0
      - name: ibm-spectrum-scale-csi-operator.v2.4.0
        version: 2.4.0
      - name: ibm-spectrum-scale-csi-operator.v2.3.1
        version: 2.3.1
      - name: ibm-spectrum-scale-csi-operator.v2.3.0
        version: 2.3.0
      name: stable
    defaultChannel: stable
    packageName: ibm-spectrum-scale-csi-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Managed Openshift Black Belt
      provider-url: ""
    name: ecr-secret-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ecr-secret-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "ecr-secret-sample",
                  "namespace": "test-ecr-secret-operator"
                },
                "spec": {
                  "ecr_registry": "ACCOUNT_ID.dkr.ecr.us-east-2.amazonaws.com",
                  "frequency": "10h",
                  "generated_secret_name": "ecr-docker-secret",
                  "region": "us-east-2"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ecr-secret-operator:v0.3.2
          operatorframework.io/suggested-namespace: ecr-secret-operator
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ecr-secret-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret is the Schema for the secrets API
            displayName: Secret
            kind: Secret
            name: secrets.ecr.mobb.redhat.com
            version: v1alpha1
        description: "# ECR Secret Operator\n\nAmazon Elastic Container Registry [Private
          Registry Authentication](https://docs.aws.amazon.com/AmazonECR/latest/userguide/registry_auth.html)
          provides a temporary token that is valid only for 12 hours. It is a challenge
          for automatic container image build process to refresh the token or secret
          in a timely manner.\n\nThis operators frequently talks with AWS ECR GetAuthroization
          Token and create/update the secret, so that the service account can perform
          docker image build.\n\n\n## How to use this operator\n\n### Prerequisites\n\n*
          [Create an ECR private repository](https://docs.aws.amazon.com/AmazonECR/latest/userguide/repository-create.html)\n*
          Create An Openshift Cluster\n* Provide AWS Authentication to the operator.
          Two Options:\n  **Note: You have to provide a secret aws-ecr-cloud-credentials
          in the installed namespace**\n  * [IAM User](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_user.md)\n
          \ * [STS Assume Role](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_assume_role.md)\n\n###
          Install the operator\n\n### Create the ECR Secret CRD\n\n```\napiVersion:
          ecr.mobb.redhat.com/v1alpha1\nkind: Secret\nmetadata:\n  name: ecr-secret\n
          \ namespace: test-ecr-secret-operator\nspec:\n  generated_secret_name: ecr-docker-secret\n
          \ ecr_registry: [ACCOUNT_ID].dkr.ecr.us-east-2.amazonaws.com\n  frequency:
          10h\n  region: us-east-2\n```\n\n```\noc create -f CRD_FILE\n```\n\nA docker
          registry secret is created by the operator momentally and the token is patched
          every 10 hours\n\n```\noc get secret ecr-docker-secret   \nNAME                TYPE
          \                            DATA   AGE\necr-docker-secret   kubernetes.io/dockerconfigjson
          \  1      16h\n```\n\n### A sample build process with generated secret\n\n\nLink
          the secret to builder\n\n```\noc secrets link builder ecr-docker-secret
          \n```\n\nConfigure [build config](https://github.com/rh-mobb/ecr-secret-operator/tree/main/samples/build-config.yaml)
          to point to your ECR Container repository\n\n```\noc create imagestream
          ruby\noc tag openshift/ruby:2.5-ubi8 ruby:2.5\noc create -f samples/build-config.yaml\noc
          start-build ruby-sample-build --wait\n```\n\nBuild should succeed and push
          the image to the the private ECR Container repository\n\n"
        displayName: ECR Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - redhat
        links:
        - name: Ecr Secret Operator
          url: https://github.com/rh-mobb/ecr-secret-operator
        maintainers:
        - email: shading@redhat.com
          name: Shaozhen Ding
        maturity: alpha
        provider:
          name: Managed Openshift Black Belt
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/mobb/ecr-secret-operator:v0.3.2
        version: 0.3.2
      entries:
      - name: ecr-secret-operator.v0.3.2
        version: 0.3.2
      name: v0.3
    - currentCSV: ecr-secret-operator.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "ecr-secret-sample",
                  "namespace": "test-ecr-secret-operator"
                },
                "spec": {
                  "ecr_registry": "ACCOUNT_ID.dkr.ecr.us-east-2.amazonaws.com",
                  "frequency": "10h",
                  "generated_secret_name": "ecr-docker-secret",
                  "region": "us-east-2"
                }
              },
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "ArgoHelmRepoSecret",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ecr-secret-operator",
                    "app.kubernetes.io/instance": "argohelmreposecret-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "argohelmreposecret",
                    "app.kubernetes.io/part-of": "ecr-secret-operator"
                  },
                  "name": "argohelmreposecret-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ecr-secret-operator:v0.4.1
          createdAt: "2023-04-14T14:52:13Z"
          operatorframework.io/suggested-namespace: ecr-secret-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ecr-secret-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ArgoHelmRepoSecret is the Schema for the argohelmreposecrets
              API
            displayName: Argo Helm Repo Secret
            kind: ArgoHelmRepoSecret
            name: argohelmreposecrets.ecr.mobb.redhat.com
            version: v1alpha1
          - description: Secret is the Schema for the secrets API
            displayName: Secret
            kind: Secret
            name: secrets.ecr.mobb.redhat.com
            version: v1alpha1
        description: "# ECR Secret Operator\n\nAmazon Elastic Container Registry [Private
          Registry Authentication](https://docs.aws.amazon.com/AmazonECR/latest/userguide/registry_auth.html)
          provides a temporary token that is valid only for 12 hours. It is a challenge
          for automatic container image build process to refresh the token or secret
          in a timely manner.\n\nThis operators frequently talks with AWS ECR GetAuthroization
          Token and create/update the secret, so that the service account can perform
          docker image build.\n\n\n## How to use this operator\n\n### Prerequisites\n\n*
          [Create an ECR private repository](https://docs.aws.amazon.com/AmazonECR/latest/userguide/repository-create.html)\n*
          Create An Openshift Cluster\n* Provide AWS Authentication to the operator.
          Two Options:\n  **Note: You have to provide a secret aws-ecr-cloud-credentials
          in the installed namespace**\n  * [IAM User](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_user.md)\n
          \ * [STS Assume Role](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_assume_role.md)\n\n###
          Install the operator\n\n### Create the ECR Secret CRD\n\n```\napiVersion:
          ecr.mobb.redhat.com/v1alpha1\nkind: Secret\nmetadata:\n  name: ecr-secret\n
          \ namespace: test-ecr-secret-operator\nspec:\n  generated_secret_name: ecr-docker-secret\n
          \ ecr_registry: [ACCOUNT_ID].dkr.ecr.us-east-2.amazonaws.com\n  frequency:
          10h\n  region: us-east-2\n```\n\n```\noc create -f CRD_FILE\n```\n\nA docker
          registry secret is created by the operator momentally and the token is patched
          every 10 hours\n\n```\noc get secret ecr-docker-secret   \nNAME                TYPE
          \                            DATA   AGE\necr-docker-secret   kubernetes.io/dockerconfigjson
          \  1      16h\n```\n\n### A sample build process with generated secret\n\n\nLink
          the secret to builder\n\n```\noc secrets link builder ecr-docker-secret
          \n```\n\nConfigure [build config](https://github.com/rh-mobb/ecr-secret-operator/tree/main/samples/build-config.yaml)
          to point to your ECR Container repository\n\n```\noc create imagestream
          ruby\noc tag openshift/ruby:2.5-ubi8 ruby:2.5\noc create -f samples/build-config.yaml\noc
          start-build ruby-sample-build --wait\n```\n\nBuild should succeed and push
          the image to the the private ECR Container repository\n\n"
        displayName: ECR Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - redhat
        links:
        - name: Ecr Secret Operator
          url: https://github.com/rh-mobb/ecr-secret-operator
        maintainers:
        - email: shading@redhat.com
          name: Shaozhen Ding
        maturity: alpha
        provider:
          name: Managed Openshift Black Belt
        relatedImages:
        - quay.io/mobb/ecr-secret-operator:v0.4.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        version: 0.4.1
      entries:
      - name: ecr-secret-operator.v0.4.1
        version: 0.4.1
      - name: ecr-secret-operator.v0.4.0
        version: 0.4.0
      name: v0.4
    defaultChannel: v0.4
    packageName: ecr-secret-operator
    provider:
      name: Managed Openshift Black Belt
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-apigatewayv2-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-apigatewayv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Deployment",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "API",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Authorizer",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Route",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Stage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "VPCLink",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/apigatewayv2-controller:1.1.1
          createdAt: "2025-09-19T21:58:07Z"
          description: AWS API Gateway v2 controller is a service controller for managing
            API Gateway v2 resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIMapping represents the state of an AWS apigatewayv2 APIMapping
              resource.
            displayName: APIMapping
            kind: APIMapping
            name: apimappings.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: API represents the state of an AWS apigatewayv2 API resource.
            displayName: API
            kind: API
            name: apis.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Authorizer represents the state of an AWS apigatewayv2 Authorizer
              resource.
            displayName: Authorizer
            kind: Authorizer
            name: authorizers.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Deployment represents the state of an AWS apigatewayv2 Deployment
              resource.
            displayName: Deployment
            kind: Deployment
            name: deployments.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: DomainName represents the state of an AWS apigatewayv2 DomainName
              resource.
            displayName: DomainName
            kind: DomainName
            name: domainnames.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Integration represents the state of an AWS apigatewayv2 Integration
              resource.
            displayName: Integration
            kind: Integration
            name: integrations.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Route represents the state of an AWS apigatewayv2 Route resource.
            displayName: Route
            kind: Route
            name: routes.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Stage represents the state of an AWS apigatewayv2 Stage resource.
            displayName: Stage
            kind: Stage
            name: stages.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: VPCLink represents the state of an AWS apigatewayv2 VPCLink
              resource.
            displayName: VPCLink
            kind: VPCLink
            name: vpclinks.apigatewayv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon API Gateway v2 resources in AWS from within your Kubernetes cluster. Currently only HTTP APIs are supported. WebSocket API support will be added soon. For HTTP APIs, DomainName and APIMapping resource support will be added soon.

          **About Amazon API Gateway v2**

          Amazon API Gateway is a fully managed service that makes it easy for developers to create, publish, maintain, monitor, and secure APIs at any scale. APIs act as the "front door" for applications to access data, business logic, or functionality from your backend services. Using API Gateway, you can create RESTful APIs and WebSocket APIs that enable real-time two-way communication applications. API Gateway supports containerized and serverless workloads, as well as web applications.

          API Gateway handles all the tasks involved in accepting and processing up to hundreds of thousands of concurrent API calls, including traffic management, CORS support, authorization and access control, throttling, monitoring, and API version management. API Gateway has no minimum fees or startup costs. You pay for the API calls you receive and the amount of data transferred out and, with the API Gateway tiered pricing model, you can reduce your cost as your API usage scales.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon API Gateway v2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apigatewayv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon API Gateway v2 Developer Resources
          url: https://aws.amazon.com/api-gateway/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: api gateway maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/apigatewayv2-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-apigatewayv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-apigatewayv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-apigatewayv2-controller.v1.0.30
        version: 1.0.30
      - name: ack-apigatewayv2-controller.v1.0.29
        version: 1.0.29
      - name: ack-apigatewayv2-controller.v1.0.28
        version: 1.0.28
      - name: ack-apigatewayv2-controller.v1.0.26
        version: 1.0.26
      - name: ack-apigatewayv2-controller.v1.0.25
        version: 1.0.25
      - name: ack-apigatewayv2-controller.v1.0.23
        version: 1.0.23
      - name: ack-apigatewayv2-controller.v1.0.21
        version: 1.0.21
      - name: ack-apigatewayv2-controller.v1.0.17
        version: 1.0.17
      - name: ack-apigatewayv2-controller.v1.0.16
        version: 1.0.16
      - name: ack-apigatewayv2-controller.v1.0.15
        version: 1.0.15
      - name: ack-apigatewayv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-apigatewayv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-apigatewayv2-controller.v1.0.7
        version: 1.0.7
      - name: ack-apigatewayv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-apigatewayv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-apigatewayv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-apigatewayv2-controller.v1.0.3
        version: 1.0.3
      - name: ack-apigatewayv2-controller.v1.0.2
        version: 1.0.2
      - name: ack-apigatewayv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-apigatewayv2-controller.v0.1.6
        version: 0.1.6
      - name: ack-apigatewayv2-controller.v0.1.4
        version: 0.1.4
      - name: ack-apigatewayv2-controller.v0.1.3
        version: 0.1.3
      - name: ack-apigatewayv2-controller.v0.1.2
        version: 0.1.2
      - name: ack-apigatewayv2-controller.v0.1.1
        version: 0.1.1
      - name: ack-apigatewayv2-controller.v0.1.0
        version: 0.1.0
      - name: ack-apigatewayv2-controller.v0.0.22
        version: 0.0.22
      - name: ack-apigatewayv2-controller.v0.0.21
        version: 0.0.21
      - name: ack-apigatewayv2-controller.v0.0.20
        version: 0.0.20
      - name: ack-apigatewayv2-controller.v0.0.19
        version: 0.0.19
      - name: ack-apigatewayv2-controller.v0.0.18
        version: 0.0.18
      - name: ack-apigatewayv2-controller.v0.0.17
        version: 0.0.17
      - name: ack-apigatewayv2-controller.v0.0.16
        version: 0.0.16
      - name: ack-apigatewayv2-controller.v0.0.15
        version: 0.0.15
      - name: ack-apigatewayv2-controller.v0.0.14
        version: 0.0.14
      - name: ack-apigatewayv2-controller.v0.0.13
        version: 0.0.13
      - name: ack-apigatewayv2-controller.v0.0.12
        version: 0.0.12
      - name: ack-apigatewayv2-controller.v0.0.11
        version: 0.0.11
      - name: ack-apigatewayv2-controller.v0.0.10
        version: 0.0.10
      - name: ack-apigatewayv2-controller.v0.0.9
        version: 0.0.9
      name: alpha
    defaultChannel: alpha
    packageName: ack-apigatewayv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Backube
      provider-url: https://github.com/backube
    name: snapscheduler
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: snapscheduler.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "snapscheduler.backube/v1",
                "kind": "SnapshotSchedule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "snapscheduler",
                    "app.kubernetes.io/instance": "daily",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "snapshotschedule",
                    "app.kubernetes.io/part-of": "snapscheduler"
                  },
                  "name": "daily"
                },
                "spec": {
                  "retention": {
                    "maxCount": 7
                  },
                  "schedule": "0 0 * * *"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/backube/snapscheduler:3.5.0
          createdAt: "2025-05-14T14:46:43Z"
          description: The SnapScheduler operator takes snapshots of Kubernetes CSI-based
            persistent volumes according to user-supplied schedules.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/snapscheduler
          support: https://github.com/backube/snapscheduler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SnapshotSchedule defines a schedule for taking automated
              snapshots of PVC(s)
            displayName: Snapshot Schedule
            kind: SnapshotSchedule
            name: snapshotschedules.snapscheduler.backube
            version: v1
        description: |-
          ## About this operator

          The SnapScheduler operator takes snapshots of CSI-based PersistentVolumes
          according to a configurable
          [Cron-like](https://en.wikipedia.org/wiki/Cron#Overview) schedule. The
          schedules include configurable retention policies for snapshots as well as
          selectors to limit the volumes that are snapshotted. An example schedule
          could be:

          > *Snapshot **all volumes** in a namespace **daily at midnight**, retaining
          > the most recent **7** snapshots for each volume.*

          Multiple schedules can be combined to provide more elaborate protection
          schemes. For example, a given volume (or collection of volumes) could be
          protected with:

          - 6 hourly snapshots
          - 7 daily snapshots
          - 4 weekly snapshots
          - 12 monthly snapshots

          ### How it works

          The operator watches for `SnapshotSchedule` CRs in each namespace. When the
          current time matches the schedule's cronspec, the operator creates a
          `VolumeSnapshot` object for each `PersistentVolumeClaim` in the namespace
          (or subset thereof if a label selector is provided). The `VolumeSnapshot`
          objects are named according to the template:
          `<pvcname>-<schedulename>-<timestamp>`. After creating the new snapshots,
          the oldest snapshots are removed if necessary, according to the retention
          policy of the schedule.

          Please see the [full
          documentation](https://backube.github.io/snapscheduler/) for more
          information.

          ## Requirements

          - OpenShift >= 4.7 (Kubernetes >= 1.20)
          - CSI-based storage driver that supports snapshots (i.e. has the
            `CREATE_DELETE_SNAPSHOT` capability)

          ## Examples

          The schedule for snapshotting is controlled by the
          `snapshotschedules.snapscheduler.backube` Custom Resource. This is a
          namespaced resource that applies only to the PersistentVolumeClaims in its
          namespace.

          Below is a simple example that keeps 7 daily (taken at midnight) snapshots
          of all PVCs in a given namespace:

          ```yaml
          ---
          apiVersion: snapscheduler.backube/v1
          kind: SnapshotSchedule
          metadata:
            name: daily
          spec:
            retention:
              maxCount: 7
            schedule: "0 0 * * *"
          ```

          See the [usage
          documentation](https://backube.github.io/snapscheduler/usage.html) for full
          details, including how to:

          - add label selectors to restrict which PVCs this schedule applies to
          - set the VolumeSnapshotClass used by the schedule
          - apply custom labels to the automatically created VolumeSnapshot objects
        displayName: SnapScheduler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - snapshot
        - storage
        - scheduler
        links:
        - name: Documentation
          url: https://backube.github.io/snapscheduler/
        - name: Source
          url: https://github.com/backube/snapscheduler
        - name: Blog
          url: https://backube.github.io/
        maintainers:
        - email: jstrunk@redhat.com
          name: John Strunk
        maturity: stable
        minKubeVersion: 1.20.0-0
        provider:
          name: Backube
          url: https://github.com/backube
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.19.1
        - quay.io/backube/snapscheduler:3.5.0
        version: 3.5.0
      entries:
      - name: snapscheduler.v3.5.0
        version: 3.5.0
      - name: snapscheduler.v3.4.0
        version: 3.4.0
      - name: snapscheduler.v3.3.0
        version: 3.3.0
      - name: snapscheduler.v3.2.0
        version: 3.2.0
      - name: snapscheduler.v3.1.0
        version: 3.1.0
      - name: snapscheduler.v3.0.0
        version: 3.0.0
      - name: snapscheduler.v2.1.0
        version: 2.1.0
      - name: snapscheduler.v2.0.0
        version: 2.0.0
      name: candidate
    - currentCSV: snapscheduler.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "snapscheduler.backube/v1",
                "kind": "SnapshotSchedule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "snapscheduler",
                    "app.kubernetes.io/instance": "daily",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "snapshotschedule",
                    "app.kubernetes.io/part-of": "snapscheduler"
                  },
                  "name": "daily"
                },
                "spec": {
                  "retention": {
                    "maxCount": 7
                  },
                  "schedule": "0 0 * * *"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/backube/snapscheduler:3.5.0
          createdAt: "2025-05-14T14:46:43Z"
          description: The SnapScheduler operator takes snapshots of Kubernetes CSI-based
            persistent volumes according to user-supplied schedules.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/snapscheduler
          support: https://github.com/backube/snapscheduler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SnapshotSchedule defines a schedule for taking automated
              snapshots of PVC(s)
            displayName: Snapshot Schedule
            kind: SnapshotSchedule
            name: snapshotschedules.snapscheduler.backube
            version: v1
        description: |-
          ## About this operator

          The SnapScheduler operator takes snapshots of CSI-based PersistentVolumes
          according to a configurable
          [Cron-like](https://en.wikipedia.org/wiki/Cron#Overview) schedule. The
          schedules include configurable retention policies for snapshots as well as
          selectors to limit the volumes that are snapshotted. An example schedule
          could be:

          > *Snapshot **all volumes** in a namespace **daily at midnight**, retaining
          > the most recent **7** snapshots for each volume.*

          Multiple schedules can be combined to provide more elaborate protection
          schemes. For example, a given volume (or collection of volumes) could be
          protected with:

          - 6 hourly snapshots
          - 7 daily snapshots
          - 4 weekly snapshots
          - 12 monthly snapshots

          ### How it works

          The operator watches for `SnapshotSchedule` CRs in each namespace. When the
          current time matches the schedule's cronspec, the operator creates a
          `VolumeSnapshot` object for each `PersistentVolumeClaim` in the namespace
          (or subset thereof if a label selector is provided). The `VolumeSnapshot`
          objects are named according to the template:
          `<pvcname>-<schedulename>-<timestamp>`. After creating the new snapshots,
          the oldest snapshots are removed if necessary, according to the retention
          policy of the schedule.

          Please see the [full
          documentation](https://backube.github.io/snapscheduler/) for more
          information.

          ## Requirements

          - OpenShift >= 4.7 (Kubernetes >= 1.20)
          - CSI-based storage driver that supports snapshots (i.e. has the
            `CREATE_DELETE_SNAPSHOT` capability)

          ## Examples

          The schedule for snapshotting is controlled by the
          `snapshotschedules.snapscheduler.backube` Custom Resource. This is a
          namespaced resource that applies only to the PersistentVolumeClaims in its
          namespace.

          Below is a simple example that keeps 7 daily (taken at midnight) snapshots
          of all PVCs in a given namespace:

          ```yaml
          ---
          apiVersion: snapscheduler.backube/v1
          kind: SnapshotSchedule
          metadata:
            name: daily
          spec:
            retention:
              maxCount: 7
            schedule: "0 0 * * *"
          ```

          See the [usage
          documentation](https://backube.github.io/snapscheduler/usage.html) for full
          details, including how to:

          - add label selectors to restrict which PVCs this schedule applies to
          - set the VolumeSnapshotClass used by the schedule
          - apply custom labels to the automatically created VolumeSnapshot objects
        displayName: SnapScheduler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - snapshot
        - storage
        - scheduler
        links:
        - name: Documentation
          url: https://backube.github.io/snapscheduler/
        - name: Source
          url: https://github.com/backube/snapscheduler
        - name: Blog
          url: https://backube.github.io/
        maintainers:
        - email: jstrunk@redhat.com
          name: John Strunk
        maturity: stable
        minKubeVersion: 1.20.0-0
        provider:
          name: Backube
          url: https://github.com/backube
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.19.1
        - quay.io/backube/snapscheduler:3.5.0
        version: 3.5.0
      entries:
      - name: snapscheduler.v3.5.0
        version: 3.5.0
      - name: snapscheduler.v3.4.0
        version: 3.4.0
      - name: snapscheduler.v3.3.0
        version: 3.3.0
      - name: snapscheduler.v3.2.0
        version: 3.2.0
      - name: snapscheduler.v3.1.0
        version: 3.1.0
      - name: snapscheduler.v3.0.0
        version: 3.0.0
      - name: snapscheduler.v2.1.0
        version: 2.1.0
      - name: snapscheduler.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: snapscheduler
    provider:
      name: Backube
      url: https://github.com/backube
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: watcher-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: watcher-operator.v1.0.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "Watcher",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcher",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcher",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcher"
                },
                "spec": {
                  "apiServiceTemplate": {
                    "tls": {
                      "caBundleSecretName": "combined-ca-bundle"
                    }
                  },
                  "databaseInstance": "openstack"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherAPI",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcher-operator"
                  },
                  "name": "watcherapi-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherApplier",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcherapplier-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcherapplier",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcherapplier-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherDecisionEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcherdecisionengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcherdecisionengine",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcherdecisionengine-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-09-02T13:32:31Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.14
          operatorframework.io/suggested-namespace: openstack-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: WatcherAPI is the Schema for the watcherapis API
            displayName: Watcher API
            kind: WatcherAPI
            name: watcherapis.watcher.openstack.org
            version: v1beta1
          - description: WatcherApplier is the Schema for the watcherappliers API
            displayName: Watcher Applier
            kind: WatcherApplier
            name: watcherappliers.watcher.openstack.org
            version: v1beta1
          - description: WatcherDecisionEngine is the Schema for the watcherdecisionengines
              API
            displayName: Watcher Decision Engine
            kind: WatcherDecisionEngine
            name: watcherdecisionengines.watcher.openstack.org
            version: v1beta1
          - description: Watcher is the Schema for the watchers API
            displayName: Watcher
            kind: Watcher
            name: watchers.watcher.openstack.org
            version: v1beta1
        description: The Watcher Operator project
        displayName: Watcher Operator (Technology Preview)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        links:
        - name: Watcher Operator
          url: https://github.com/openstack-k8s-operators/watcher-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/watcher-rhel9-operator@sha256:63c341a3e20a522c601098502a8fd8f1164d23c773ae8f55782c43ff4b6c8c18
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:eed8234b0fe47083ab941f513709f778bfcb80e32747b71a65fd5f62328a5ca0
        - registry.redhat.io/rhoso/openstack-watcher-api-rhel9@sha256:e97369e3a47bb882f1920c6af3f6345a842af45c58f0f5c5df9ac2933cf3098a
        - registry.redhat.io/rhoso/openstack-watcher-decision-engine-rhel9@sha256:c759d66cf0986c0036be91c008cbd059b4412c94160ed82dd12cd4e00d0685a9
        - registry.redhat.io/rhoso/openstack-watcher-applier-rhel9@sha256:c92a82f901613991926367bb8aab4f875c27a3e4a5282d92847a8463d98cd9e3
        version: 1.0.14
      entries:
      - name: watcher-operator.v1.0.14
        version: 1.0.14
      - name: watcher-operator.v1.0.13
        version: 1.0.13
      - name: watcher-operator.v1.0.12
        version: 1.0.12
      - name: watcher-operator.v1.0.11
        version: 1.0.11
      - name: watcher-operator.v1.0.10
        version: 1.0.10
      - name: watcher-operator.v1.0.9
        version: 1.0.9
      - name: watcher-operator.v1.0.8
        version: 1.0.8
      - name: watcher-operator.v1.0.7
        version: 1.0.7
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: watcher-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NdmSpc project
      provider-url: ""
    name: ndmspc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ndmspc-operator.v0.20250209.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.ndmspc.io/v1alpha1",
                "kind": "NdmSpcConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ndmspc-operator",
                    "app.kubernetes.io/instance": "ndmspcconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ndmspcconfig",
                    "app.kubernetes.io/part-of": "ndmspc-operator"
                  },
                  "name": "ndmspc-sample"
                },
                "spec": {
                  "executor": {
                    "image": "registry.gitlab.com/ndmspc/api:v0.20240215.2",
                    "name": "executor"
                  },
                  "executors": 1,
                  "ingressclasstype": "nginx",
                  "ingressenabled": true,
                  "ingresshostdomain": "127.0.0.1.sslip.io",
                  "ingressnostname": "ndmspc",
                  "ingressssltraefik": false,
                  "naf": {
                    "image": "registry.gitlab.com/ndmspc/naf:v0.20250129.0",
                    "name": "naf"
                  },
                  "ndmvr": {
                    "image": "registry.gitlab.com/ndmspc/ndmvr/shared:v0.20230320.0",
                    "name": "ndmvr"
                  },
                  "salsa": {
                    "image": "registry.gitlab.com/ndmspc/user/ndmspc:v0.20240412.0"
                  },
                  "storage": {
                    "image": "registry.gitlab.com/ndmspc/storage-server:v0.20250205.0",
                    "mongodb": {
                      "image": "docker.io/library/mongo:latest",
                      "password": "admin",
                      "port": 27017,
                      "username": "admin"
                    },
                    "name": "storage",
                    "port": 8080
                  },
                  "webapp": {
                    "image": "registry.gitlab.com/ndmspc/react-ndmspc:v0.20240412.0",
                    "name": ""
                  },
                  "webappcontainerport": 8080,
                  "webapps": 1,
                  "worker": {
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "1Gi"
                      }
                    }
                  },
                  "workers": 2,
                  "workerslots": 1,
                  "zmq2ws": {
                    "image": "registry.gitlab.com/ndmspc/zmq2ws:v1.20231012.0",
                    "name": "zmq2ws"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: registry.gitlab.com/ndmspc/ndmspc-operator:0.20250209.0
          createdAt: "2025-02-09T12:49:06Z"
          description: The NdmSpc operator automates the deployment and management
            of Ndmspc suite on Kubernetes cluster.
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://gitlab.com/ndmspc/ndmspc-operator
          support: NDMSPC Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NdmSpc configuration
            displayName: NdmSpc Config
            kind: NdmSpcConfig
            name: ndmspcconfigs.apps.ndmspc.io
            version: v1alpha1
        description: "The NdmSpc operator automates the deployment and management
          of Ndmspc suite on Kubernetes cluster. This operator allows user to deploy
          SALSA cluster with ZMQ2WS and job executor service and specify custom webapp
          page.\n\n### Capabilities\n\n- SALSA cluster\n- ZMQ2WS api service\n- Job
          executor service \n- Custom WebApp page\n"
        displayName: NdmSpc operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ndmspc
        links:
        - name: Ndmspc Operator
          url: https://ndmspc.gitlab.io/docs
        maintainers:
        - email: ndmspc@gmail.com
          name: NdmSpc Support
        maturity: alpha
        provider:
          name: NdmSpc project
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - registry.gitlab.com/ndmspc/ndmspc-operator:0.20250209.0
        version: 0.20250209.0
      entries:
      - name: ndmspc-operator.v0.20250209.0
        version: 0.20250209.0
      - name: ndmspc-operator.v0.20250131.0
        version: 0.20250131.0
      - name: ndmspc-operator.v0.20240412.0
        version: 0.20240412.0
      - name: ndmspc-operator.v0.11.6
        version: 0.11.6
      - name: ndmspc-operator.v0.11.4
        version: 0.11.4
      - name: ndmspc-operator.v0.11.3
        version: 0.11.3
      - name: ndmspc-operator.v0.11.2
        version: 0.11.2
      - name: ndmspc-operator.v0.11.1
        version: 0.11.1
      - name: ndmspc-operator.v0.10.2
        version: 0.10.2
      - name: ndmspc-operator.v0.10.1
        version: 0.10.1
      - name: ndmspc-operator.v0.10.0
        version: 0.10.0
      - name: ndmspc-operator.v0.9.0
        version: 0.9.0
      - name: ndmspc-operator.v0.7.4
        version: 0.7.4
      - name: ndmspc-operator.v0.7.1
        version: 0.7.1
      - name: ndmspc-operator.v0.6.0
        version: 0.6.0
      - name: ndmspc-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: ndmspc-operator
    provider:
      name: NdmSpc project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: VMware Tanzu
      provider-url: ""
    name: rabbitmq-cluster-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-cluster-operator.v2.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "rabbitmq.com/v1beta1",
                "kind": "RabbitmqCluster",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "override": {
                    "statefulset": {
                      "spec": {
                        "template": {
                          "spec": {
                            "containers": [],
                            "securityContext": {}
                          }
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/rabbitmqoperator/cluster-operator:2.16.0
          createdAt: 2025-08-12T09:34:33
          repository: https://github.com/rabbitmq/cluster-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Rabbitmq Cluster
            displayName: RabbitmqCluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: "## About the managed application\nRabbitMQ is an open-source
          message broker that originally implemented the\nAdvanced Message Queuing
          Protocol (AMQP). Nowadays features a plug-in\narchitecture supporting protocols
          such as Streaming Text Oriented Messaging\nProtocol (STOMP), MQ Telemetry
          Transport (MQTT), and others. RabbitMQ\nsupports native streaming and AMQP
          1.0.\n\nVisit [rabbitmq.com](rabbitmq.com) for more information.\n\n## About
          this Operator\nRabbitMQ Cluster Kubernetes Operator is developed and maintened
          by the VMware \nRabbitMQ team and provides a consistent and easy way to
          deploy RabbitMQ\nclusters to Kubernetes and run them, including \"day two\"
          (continuous)\noperations. RabbitMQ clusters deployed using the Operator
          can be used by\napplications running on Kubernetes or outside of Kubernetes.\n\nRead
          [RabbitMQ operator docs](https://www.rabbitmq.com/kubernetes/operator/operator-overview)\nfor
          more information.\n"
        displayName: RabbitMQ-cluster-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        - messaging
        - amqp
        - amqp10
        - stomp
        - mqtt
        - mqtt5
        - queues
        - streaming
        links:
        - name: Documentation
          url: https://www.rabbitmq.com/kubernetes/operator/operator-overview.html
        - name: GitHub
          url: https://github.com/rabbitmq/cluster-operator
        maintainers:
        - email: rabbitmq-users@googlegroups.com
          name: VMware Tanzu
        maturity: stable
        minKubeVersion: 1.26.0
        nativeApis:
        - group: apps
          kind: Deployment
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: ReplicaSet
          version: v1
        - group: ""
          kind: Pod
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: ConfigMap
          version: v1
        provider:
          name: VMware Tanzu
        relatedImages:
        - quay.io/rabbitmqoperator/cluster-operator:2.16.0
        version: 2.16.0
      entries:
      - name: rabbitmq-cluster-operator.v2.16.0
        version: 2.16.0
      - name: rabbitmq-cluster-operator.v2.15.0
        version: 2.15.0
      - name: rabbitmq-cluster-operator.v2.14.0
        version: 2.14.0
      - name: rabbitmq-cluster-operator.v2.13.0
        version: 2.13.0
      - name: rabbitmq-cluster-operator.v2.12.1
        version: 2.12.1
      - name: rabbitmq-cluster-operator.v2.10.0
        version: 2.10.0
      - name: rabbitmq-cluster-operator.v2.9.0
        version: 2.9.0
      - name: rabbitmq-cluster-operator.v2.8.0
        version: 2.8.0
      - name: rabbitmq-cluster-operator.v2.7.0
        version: 2.7.0
      - name: rabbitmq-cluster-operator.v2.6.0
        version: 2.6.0
      - name: rabbitmq-cluster-operator.v2.5.0
        version: 2.5.0
      - name: rabbitmq-cluster-operator.v2.4.0
        version: 2.4.0
      - name: rabbitmq-cluster-operator.v2.3.0
        version: 2.3.0
      - name: rabbitmq-cluster-operator.v2.2.0
        version: 2.2.0
      - name: rabbitmq-cluster-operator.v2.1.0
        version: 2.1.0
      - name: rabbitmq-cluster-operator.v2.0.0
        version: 2.0.0
      - name: rabbitmq-cluster-operator.v1.14.0
        version: 1.14.0
      name: stable
    defaultChannel: stable
    packageName: rabbitmq-cluster-operator
    provider:
      name: VMware Tanzu
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: ""
    name: ibm-metrics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-metrics-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MarketplaceConfig",
                "metadata": {
                  "name": "marketplaceconfig"
                },
                "spec": {
                  "features": {
                    "deployment": true,
                    "meterDefinitionCatalogServer": false,
                    "registration": true
                  },
                  "installIBMCatalogSource": true,
                  "isDisconnected": false,
                  "license": {
                    "accept": false
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterBase",
                "metadata": {
                  "name": "meterbase"
                },
                "spec": {
                  "enabled": true,
                  "prometheus": {
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "1",
                        "memory": "1G"
                      }
                    },
                    "storage": {
                      "size": "20Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterReport",
                "metadata": {
                  "name": "example-meterreport"
                },
                "spec": {
                  "endTime": "2020-08-05T00:00:00Z",
                  "prometheusService": {
                    "name": "rhm-prometheus-meterbase",
                    "namespace": "openshift-redhat-marketplace",
                    "targetPort": "rbac"
                  },
                  "startTime": "2020-08-04T00:00:00Z"
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "example-meterdefinition-4"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "meterdefinition-sample"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
          createdAt: "2025-08-28T14:18:55Z"
          description: The IBM Metrics Operator provides workload metering and reporting
            for IBM and Red Hat Marketplace customers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
          - description: MeterBase is the resource that sets up Metering for Red Hat
              Marketplace. This is an internal resource not meant to be modified directly.
            displayName: Meter Base
            kind: MeterBase
            name: meterbases.marketplace.redhat.com
            version: v1alpha1
          - description: MeterDefinition defines the meter workloads used to enable
              pay for use billing.
            displayName: Meter Definition
            kind: MeterDefinition
            name: meterdefinitions.marketplace.redhat.com
            version: v1beta1
          - kind: MeterReport
            name: meterreports.marketplace.redhat.com
            version: v1alpha1
        description: "The IBM Metrics Operator provides workload metering and reporting
          for IBM Software Central and Red Hat Marketplace customers.\n### **Important
          Note**\nA set of instructions for onboarding is provided here. For more
          detailed onboarding instructions or information about what is installed
          please visit [swc.saas.ibm.com](https://swc.saas.ibm.com).\n\nUsage metrics
          may be monitored through [https://swc.saas.ibm.com](https://swc.saas.ibm.com)
          with only IBM Metrics Operator and a Red Hat Marketplace account.\n\n###
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n### Prerequisites\n1. Installations are required
          to [enable monitoring for user-defined projects](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          as the Prometheus provider.\n2. Edit the cluster-monitoring-config ConfigMap
          object:\n\n   ```sh\n   $ oc -n openshift-monitoring edit configmap cluster-monitoring-config\n
          \   ```\n\n3. Add enableUserWorkload: true under data/config.yaml:\n  \n
          \   ```sh\n    apiVersion: v1\n    kind: ConfigMap\n    metadata:\n      name:
          cluster-monitoring-config\n        namespace: openshift-monitoring\n    data:\n
          \     config.yaml: |\n        enableUserWorkload: true\n    ```\n\n4. Configure
          the user-workload-monitoring-config ConfigMap object:\n\n    ```sh\n    $
          oc -n openshift-user-workload-monitoring edit configmap user-workload-monitoring-config\n
          \   ```\n\n5. Configure a minimum retention time of 168h and minimum storage
          capacity of 40Gi\n  \n    ```sh\n    apiVersion: v1\n    kind: ConfigMap\n
          \   metadata:\n      name: user-workload-monitoring-config\n      namespace:
          openshift-user-workload-monitoring\n\n    data:\n      config.yaml: |\n
          \       prometheus:\n          retention: 168h\n          volumeClaimTemplate:\n
          \           spec:\n              resources:\n                requests:\n
          \                 storage: 40Gi\n    ```\n\n### Resources Required\n\nMinimum
          system resources required:\n\n| Operator                | Memory (MB) |
          CPU (cores) | Disk (GB) | Nodes |\n| ----------------------- | -----------
          | ----------- | --------- | ----- |\n| **Metrics**   |        750  |     0.25
          \   | 3x1       |    3  |\n| **Deployment** |        250  |     0.25    |
          -         |    1  |\n\n| Prometheus Provider  | Memory (GB) | CPU (cores)
          | Disk (GB) | Nodes |\n| --------- | ----------- | ----------- | ---------
          | ----- |\n| **[Openshift User Workload Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)**
          |          1  |     0.1       | 2x40        |   2    |\n\nMultiple nodes
          are required to provide pod scheduling for high availability for IBM Metrics
          Operator Data Service and Prometheus.\n\nThe IBM Metrics Operator automatically
          creates 3 x 1Gi PersistentVolumeClaims to store reports as part of the data
          service, with _ReadWriteOnce_ access mode. Te PersistentVolumeClaims are
          automatically created by the ibm-metrics-operator after creating a `redhat-marketplace-pull-secret`
          and accepting the license in `marketplaceconfig`.\n\n| NAME                                |
          CAPACITY | ACCESS MODES |\n| ----------------------------------- | --------
          | ------------ |\n| rhm-data-service-rhm-data-service-0 | 1Gi | RWO |\n|
          rhm-data-service-rhm-data-service-1 | 1Gi | RWO |\n| rhm-data-service-rhm-data-service-2
          | 1Gi | RWO |\n\n### Supported Storage Providers\n\n- OpenShift Container
          Storage / OpenShift Data Foundation version 4.x, from version 4.2 or higher\n-
          IBM Cloud Block storage and IBM Cloud File storage\n- IBM Storage Suite
          for IBM Cloud Paks:\n  - File storage from IBM Spectrum Fusion/Scale \n
          \ - Block storage from IBM Spectrum Virtualize, FlashSystem or DS8K\n- Portworx
          Storage, version 2.5.5 or above\n- Amazon Elastic File Storage\n\n### Access
          Modes required\n\n - ReadWriteOnce (RWO)\n\n### Provisioning Options supported\n\nChoose
          one of the following options to provision storage for the ibm-metrics-operator
          data-service\n\n#### Dynamic provisioning using a default StorageClass\n
          \  - A StorageClass is defined with a `metadata.annotations: storageclass.kubernetes.io/is-default-class:
          \"true\"`\n   - PersistentVolumes will be provisioned automatically for
          the generated PersistentVolumeClaims\n--- \n#### Manually create each PersistentVolumeClaim
          with a specific StorageClass\n   - Must be performed before creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`. Otherwise, the automatically
          generated PersistentVolumeClaims are immutable.\n```\napiVersion: v1\nkind:
          PersistentVolumeClaim\nmetadata:\n  labels:\n    app: rhm-data-service\n
          \ name: rhm-data-service-rhm-data-service-0\n  namespace: ibm-software-central\nspec:\n
          \ storageClassName: rook-cephfs\n  accessModes:\n  - ReadWriteOnce\n  resources:\n
          \   requests:\n      storage: 1Gi\n```\n---\n#### Manually provision each
          PersistentVolume for the generated PersistentVolumeClaims with a specific
          StorageClass\n  - May be performed before or after creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`.  \n```\napiVersion: v1\nkind:
          PersistentVolume\nmetadata:\n  name: rhm-data-service-rhm-data-service-0\nspec:\n
          \ csi:\n    driver: rook-ceph.cephfs.csi.ceph.com\n    volumeHandle: rhm-data-service-rhm-data-service-0\n
          \ capacity:\n    storage: 1Gi\n  accessModes:\n    - ReadWriteOnce\n  persistentVolumeReclaimPolicy:
          Delete\n  storageClassName: rook-cephfs\n  volumeMode: Filesystem\n  claimRef:\n
          \   kind: PersistentVolumeClaim\n    namespace: ibm-software-central\n    name:
          rhm-data-service-rhm-data-service-0\n```\n\n### Installation\n1. Create
          or get your [pull secret](https://swc.saas.ibm.com/en-us/documentation/clusters#get-pull-secret).\n2.
          Install the IBM Metrics Operator\n3. Create a Kubernetes secret in the installed
          namespace with the name `redhat-marketplace-pull-secret` and key `PULL_SECRET`
          with the value of the pull secret.\n    ```sh\n    # Replace ${PULL_SECRET}
          with your pull secret\n    oc create secret generic redhat-marketplace-pull-secret
          -n ibm-software-central --from-literal=PULL_SECRET=${PULL_SECRET}\n    ```\n4.
          Use of the platform is governed by the:\n\n    [IBM Cloud Services Agreement](https://www.ibm.com/support/customer/csol/terms/?id=Z126-6304_WS&_ga=2.116312197.2046730452.1684328846-812467790.1684328846)
          (or other base agreement between you and IBM such as a [Passport Advantage
          Agreement](https://www.ibm.com/software/passportadvantage/pa_agreements.html?_ga=2.116312197.2046730452.1684328846-812467790.1684328846))
          and the [Service Description for the Red Hat Marketplace](https://www.ibm.com/support/customer/csol/terms/?id=i126-8719&_ga=2.83289621.2046730452.1684328846-812467790.1684328846).\n
          \   \n5. Update MarketplaceConfig to accept the license.\n    ```\n    oc
          patch marketplaceconfig marketplaceconfig -n ibm-software-central --type='merge'
          -p '{\"spec\": {\"license\": {\"accept\": true}}}'\n    ```\n6. Install
          the pull secret as a global pull secret on the cluster.\n\n    These steps
          require `oc`, `jq`, and `base64` to be available on your machine.\n\n    ```sh\n
          \   # Create the docker pull secret file using your pull secret.\n    #
          Store it in a file called entitledregistryconfigjson.\n    oc create secret
          docker-registry entitled-registry --docker-server=registry.marketplace.redhat.com
          --docker-username \"cp\" --docker-password \"${PULL_SECRET}\" --dry-run=client
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > entitledregistryconfigjson\n
          \   # Get the current global secrets on the cluster and store it as a file
          named dockerconfigjson\n    oc get secret pull-secret -n openshift-config
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > dockerconfigjson\n
          \   # Merge the two dockerconfigs together into a file called dockerconfigjson-merged.\n
          \   jq -s '.[0] * .[1]' dockerconfigjson entitledregistryconfigjson > dockerconfigjson-merged\n
          \   # Set the cluster's dockerconfig file to the new merged version.\n    oc
          set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson=dockerconfigjson-merged\n
          \   ```\n  7. (Optional) Your IBM product may have installed IBM Metrics
          Operator using an `ibm-entitlement-key` for authentication to IBM Software
          Central. To authorize the automatic creation of an IBM Software Central
          account for the IBM account owning the `ibm-entitlement-key`\n      ```\n
          \     oc patch marketplaceconfig marketplaceconfig -n ibm-software-central
          --type='merge' -p '{\"spec\": {\"authorizeAccountCreation\": true}}'\n      ```\n\n###
          Why is a global pull secret required?\nIn order to successfully install
          the products hosted by the container image registry, you will need to make
          the pull secret available across the cluster. This can be achieved by applying
          the pull as a [global pull secret](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html#images-update-global-pull-secret_using-image-pull-secrets).
          For alternative approaches, please see the official OpenShift [documentation](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html).\n\n\n###
          SecurityContextConstraints requirements\n\nThe Operators and their components
          support running under the OpenShift Container Platform default restricted
          and restricted-v2 security context constraints.\n\n### Installation Namespace
          and ClusterRoleBinding requirements\n\nThe IBM Metrics Operator components
          require specific ClusterRoleBindings.\n- The metric-state component requires
          a ClusterRoleBinding for the the `view` ClusterRole.\n  - Ability to read
          non-sensitive CustomResources. \n  - The `view` ClusterRole dynamically
          adds new CustomResourceDefinitions.\n- The operator & reporter component
          requires a ClusterRoleBinding for the the `cluster-monitoring-view` ClusterRole.\n
          \ - Ability to view Prometheus metrics. \n  - The underlying ClusterRole
          RBAC often updated between OpenShift versions.\n\nDue to limitations of
          Operator Lifecycle Manager (OLM), this ClusterRoleBinding can not be provided
          dynamically for arbitrary installation target namespaces.\n\nA static ClusterRoleBinding
          is included for installation to the default namespace of `ibm-software-central`,
          and namespaces `openshift-redhat-marketplace`,  `redhat-marketplace`, `ibm-common-services`.\n\nTo
          create the ClusterRoleBindings for installation to an alternate namespace\n```\noc
          project INSTALL-NAMESPACE\noc adm policy add-cluster-role-to-user view -z
          ibm-metrics-operator-metric-state\noc adm policy add-cluster-role-to-user
          cluster-monitoring-view -z ibm-metrics-operator-controller-manager,ibm-metrics-operator-reporter\n```\n\n###
          Metric State scoping requirements\nThe metric-state Deployment obtains `get/list/watch`
          access to metered resources via the `view` ClusterRole. For operators deployed
          using Operator Lifecycle Manager (OLM), permissions are added to `clusterrole/view`
          dynamically via a generated and annotated `-view` ClusterRole. If you wish
          to meter an operator, and its Custom Resource Definitions (CRDs) are not
          deployed through OLM, one of two options are required\n1. Add the following
          label to a clusterrole that has get/list/watch access to your CRD: `rbac.authorization.k8s.io/aggregate-to-view:
          \"true\"`, thereby dynamically adding it to `clusterrole/view`\n2. Create
          a ClusterRole that has get/list/watch access to your CRD, and create a ClusterRoleBinding
          for the metric-state ServiceAccount\n\nAttempting to meter a resource with
          a MeterDefinition without the required permissions will log an `AccessDeniedError`
          in metric-state.\n\n### Disaster Recovery\n\nTo plan for disaster recovery,
          note the PhysicalVolumeClaims `rhm-data-service-rhm-data-service-N`. \n-
          In connected environments, MeterReport data upload attempts occur hourly,
          and are then removed from data-service. There is a low risk of losing much
          unreported data.\n- In an airgap environment, MeterReport data must be pulled
          from data-service and uploaded manually using [datactl](https://github.com/redhat-marketplace/datactl/).
          To prevent data loss in a disaster scenario, the data-service volumes should
          be considered in a recovery plan.\n\n### Subscription Config\n\nIt is possible
          to [configure](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md)
          how OLM deploys an Operator via the `config` field in the [Subscription](https://github.com/operator-framework/olm-book/blob/master/docs/subscriptions.md)
          object.\n\nThe IBM Metrics Operator will also read the `config` and append
          it to the operands. The primary use case is to control scheduling using
          [Tolerations](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/)
          and [NodeSelectors](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/).\n\nA
          limitation is that the `config` elements are only appended to the operands.
          The elements in the operands are not removed if the `config` is removed
          from the `Subscripion`. The operand must be modified manually, or deleted
          and recreated by the controller.\n\n### Cluster permission requirements\n\n|API
          group             |Resources              |Verbs                                     |\n|----------------------|-----------------------|------------------------------------------|\n|apps
          \                 |statefulsets           |get;list;watch                            |\n|authentication.k8s.io
          |tokenreviews           |create                                    |\n|authorization.k8s.io
          \ |subjectaccessreviews   |create                                    |\n|config.openshift.io
          \  |clusterversions        |get;list;watch                            |\n|marketplace.redhat.com|meterdefinitions
          \      |get;list;watch;create;update;patch;delete |\n|operators.coreos.com
          \ |clusterserviceversions |get;list;watch                            |\n|operators.coreos.com
          \ |subscriptions          |get;list;watch                            |\n|operators.coreos.com
          \ |operatorgroups         |get;list                                  |\n|storage.k8s.io
          \       |storageclasses         |get;list;watch                            |\n|
          \                     |namespaces             |get;list;watch                            |\n|
          \                     |configmaps             |get;list;watch                            |\n|
          \                     |services               |get;list;watch                            |\n|
          \                     |**clusterrole/view**   |                                          |\n\n\n###
          Documentation\nYou can find our documentation [here.](https://swc.saas.ibm.com/en-us/documentation/)\n\n###
          Getting help\nIf you encounter any issues while using the operators, you
          can create an issue on our [Github\nrepo](https://github.com/redhat-marketplace/redhat-marketplace-operator)
          for bugs, enhancements, or other requests. You can also visit our main page
          and\nreview our [support](https://swc.saas.ibm.com/en-us/support) and [documentation](https://swc.saas.ibm.com/en-us/documentation/).\n\n###
          Readme\nYou can find our readme [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/README.md)\n\n###
          License information\nYou can find our license information [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/LICENSE)\n"
        displayName: IBM Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        - buy
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/redhat-marketplace-data-service@sha256:30fd8d2996b921cfce51f697ee4490a6d9a836021952429b877336f5edf2749d
        - quay.io/rh-marketplace/redhat-marketplace-reporter@sha256:2bd098f057ca07f2b14be9b60ea6ad8ca6b0cae1d226b1ca3b27e23d04ccdf20
        - quay.io/rh-marketplace/redhat-marketplace-metric-state@sha256:e4014c9a546c10e84b13a09db6a18431ee9489ad1296808635e48d9ef4963cf1
        - quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
        - quay.io/rh-marketplace/redhat-marketplace-authcheck@sha256:13a81b8aca79ae8c89ee24dcc30f3322a4cfdedf929e7bb984295c1e7e631fa9
        version: 2.20.0
      entries:
      - name: ibm-metrics-operator.v2.20.0
        version: 2.20.0
      name: beta
    - currentCSV: ibm-metrics-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MarketplaceConfig",
                "metadata": {
                  "name": "marketplaceconfig"
                },
                "spec": {
                  "features": {
                    "deployment": true,
                    "meterDefinitionCatalogServer": false,
                    "registration": true
                  },
                  "installIBMCatalogSource": true,
                  "isDisconnected": false,
                  "license": {
                    "accept": false
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterBase",
                "metadata": {
                  "name": "meterbase"
                },
                "spec": {
                  "enabled": true,
                  "prometheus": {
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "1",
                        "memory": "1G"
                      }
                    },
                    "storage": {
                      "size": "20Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterReport",
                "metadata": {
                  "name": "example-meterreport"
                },
                "spec": {
                  "endTime": "2020-08-05T00:00:00Z",
                  "prometheusService": {
                    "name": "rhm-prometheus-meterbase",
                    "namespace": "openshift-redhat-marketplace",
                    "targetPort": "rbac"
                  },
                  "startTime": "2020-08-04T00:00:00Z"
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "example-meterdefinition-4"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "meterdefinition-sample"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
          createdAt: "2025-08-28T14:18:55Z"
          description: The IBM Metrics Operator provides workload metering and reporting
            for IBM and Red Hat Marketplace customers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
          - description: MeterBase is the resource that sets up Metering for Red Hat
              Marketplace. This is an internal resource not meant to be modified directly.
            displayName: Meter Base
            kind: MeterBase
            name: meterbases.marketplace.redhat.com
            version: v1alpha1
          - description: MeterDefinition defines the meter workloads used to enable
              pay for use billing.
            displayName: Meter Definition
            kind: MeterDefinition
            name: meterdefinitions.marketplace.redhat.com
            version: v1beta1
          - kind: MeterReport
            name: meterreports.marketplace.redhat.com
            version: v1alpha1
        description: "The IBM Metrics Operator provides workload metering and reporting
          for IBM Software Central and Red Hat Marketplace customers.\n### **Important
          Note**\nA set of instructions for onboarding is provided here. For more
          detailed onboarding instructions or information about what is installed
          please visit [swc.saas.ibm.com](https://swc.saas.ibm.com).\n\nUsage metrics
          may be monitored through [https://swc.saas.ibm.com](https://swc.saas.ibm.com)
          with only IBM Metrics Operator and a Red Hat Marketplace account.\n\n###
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n### Prerequisites\n1. Installations are required
          to [enable monitoring for user-defined projects](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          as the Prometheus provider.\n2. Edit the cluster-monitoring-config ConfigMap
          object:\n\n   ```sh\n   $ oc -n openshift-monitoring edit configmap cluster-monitoring-config\n
          \   ```\n\n3. Add enableUserWorkload: true under data/config.yaml:\n  \n
          \   ```sh\n    apiVersion: v1\n    kind: ConfigMap\n    metadata:\n      name:
          cluster-monitoring-config\n        namespace: openshift-monitoring\n    data:\n
          \     config.yaml: |\n        enableUserWorkload: true\n    ```\n\n4. Configure
          the user-workload-monitoring-config ConfigMap object:\n\n    ```sh\n    $
          oc -n openshift-user-workload-monitoring edit configmap user-workload-monitoring-config\n
          \   ```\n\n5. Configure a minimum retention time of 168h and minimum storage
          capacity of 40Gi\n  \n    ```sh\n    apiVersion: v1\n    kind: ConfigMap\n
          \   metadata:\n      name: user-workload-monitoring-config\n      namespace:
          openshift-user-workload-monitoring\n\n    data:\n      config.yaml: |\n
          \       prometheus:\n          retention: 168h\n          volumeClaimTemplate:\n
          \           spec:\n              resources:\n                requests:\n
          \                 storage: 40Gi\n    ```\n\n### Resources Required\n\nMinimum
          system resources required:\n\n| Operator                | Memory (MB) |
          CPU (cores) | Disk (GB) | Nodes |\n| ----------------------- | -----------
          | ----------- | --------- | ----- |\n| **Metrics**   |        750  |     0.25
          \   | 3x1       |    3  |\n| **Deployment** |        250  |     0.25    |
          -         |    1  |\n\n| Prometheus Provider  | Memory (GB) | CPU (cores)
          | Disk (GB) | Nodes |\n| --------- | ----------- | ----------- | ---------
          | ----- |\n| **[Openshift User Workload Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)**
          |          1  |     0.1       | 2x40        |   2    |\n\nMultiple nodes
          are required to provide pod scheduling for high availability for IBM Metrics
          Operator Data Service and Prometheus.\n\nThe IBM Metrics Operator automatically
          creates 3 x 1Gi PersistentVolumeClaims to store reports as part of the data
          service, with _ReadWriteOnce_ access mode. Te PersistentVolumeClaims are
          automatically created by the ibm-metrics-operator after creating a `redhat-marketplace-pull-secret`
          and accepting the license in `marketplaceconfig`.\n\n| NAME                                |
          CAPACITY | ACCESS MODES |\n| ----------------------------------- | --------
          | ------------ |\n| rhm-data-service-rhm-data-service-0 | 1Gi | RWO |\n|
          rhm-data-service-rhm-data-service-1 | 1Gi | RWO |\n| rhm-data-service-rhm-data-service-2
          | 1Gi | RWO |\n\n### Supported Storage Providers\n\n- OpenShift Container
          Storage / OpenShift Data Foundation version 4.x, from version 4.2 or higher\n-
          IBM Cloud Block storage and IBM Cloud File storage\n- IBM Storage Suite
          for IBM Cloud Paks:\n  - File storage from IBM Spectrum Fusion/Scale \n
          \ - Block storage from IBM Spectrum Virtualize, FlashSystem or DS8K\n- Portworx
          Storage, version 2.5.5 or above\n- Amazon Elastic File Storage\n\n### Access
          Modes required\n\n - ReadWriteOnce (RWO)\n\n### Provisioning Options supported\n\nChoose
          one of the following options to provision storage for the ibm-metrics-operator
          data-service\n\n#### Dynamic provisioning using a default StorageClass\n
          \  - A StorageClass is defined with a `metadata.annotations: storageclass.kubernetes.io/is-default-class:
          \"true\"`\n   - PersistentVolumes will be provisioned automatically for
          the generated PersistentVolumeClaims\n--- \n#### Manually create each PersistentVolumeClaim
          with a specific StorageClass\n   - Must be performed before creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`. Otherwise, the automatically
          generated PersistentVolumeClaims are immutable.\n```\napiVersion: v1\nkind:
          PersistentVolumeClaim\nmetadata:\n  labels:\n    app: rhm-data-service\n
          \ name: rhm-data-service-rhm-data-service-0\n  namespace: ibm-software-central\nspec:\n
          \ storageClassName: rook-cephfs\n  accessModes:\n  - ReadWriteOnce\n  resources:\n
          \   requests:\n      storage: 1Gi\n```\n---\n#### Manually provision each
          PersistentVolume for the generated PersistentVolumeClaims with a specific
          StorageClass\n  - May be performed before or after creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`.  \n```\napiVersion: v1\nkind:
          PersistentVolume\nmetadata:\n  name: rhm-data-service-rhm-data-service-0\nspec:\n
          \ csi:\n    driver: rook-ceph.cephfs.csi.ceph.com\n    volumeHandle: rhm-data-service-rhm-data-service-0\n
          \ capacity:\n    storage: 1Gi\n  accessModes:\n    - ReadWriteOnce\n  persistentVolumeReclaimPolicy:
          Delete\n  storageClassName: rook-cephfs\n  volumeMode: Filesystem\n  claimRef:\n
          \   kind: PersistentVolumeClaim\n    namespace: ibm-software-central\n    name:
          rhm-data-service-rhm-data-service-0\n```\n\n### Installation\n1. Create
          or get your [pull secret](https://swc.saas.ibm.com/en-us/documentation/clusters#get-pull-secret).\n2.
          Install the IBM Metrics Operator\n3. Create a Kubernetes secret in the installed
          namespace with the name `redhat-marketplace-pull-secret` and key `PULL_SECRET`
          with the value of the pull secret.\n    ```sh\n    # Replace ${PULL_SECRET}
          with your pull secret\n    oc create secret generic redhat-marketplace-pull-secret
          -n ibm-software-central --from-literal=PULL_SECRET=${PULL_SECRET}\n    ```\n4.
          Use of the platform is governed by the:\n\n    [IBM Cloud Services Agreement](https://www.ibm.com/support/customer/csol/terms/?id=Z126-6304_WS&_ga=2.116312197.2046730452.1684328846-812467790.1684328846)
          (or other base agreement between you and IBM such as a [Passport Advantage
          Agreement](https://www.ibm.com/software/passportadvantage/pa_agreements.html?_ga=2.116312197.2046730452.1684328846-812467790.1684328846))
          and the [Service Description for the Red Hat Marketplace](https://www.ibm.com/support/customer/csol/terms/?id=i126-8719&_ga=2.83289621.2046730452.1684328846-812467790.1684328846).\n
          \   \n5. Update MarketplaceConfig to accept the license.\n    ```\n    oc
          patch marketplaceconfig marketplaceconfig -n ibm-software-central --type='merge'
          -p '{\"spec\": {\"license\": {\"accept\": true}}}'\n    ```\n6. Install
          the pull secret as a global pull secret on the cluster.\n\n    These steps
          require `oc`, `jq`, and `base64` to be available on your machine.\n\n    ```sh\n
          \   # Create the docker pull secret file using your pull secret.\n    #
          Store it in a file called entitledregistryconfigjson.\n    oc create secret
          docker-registry entitled-registry --docker-server=registry.marketplace.redhat.com
          --docker-username \"cp\" --docker-password \"${PULL_SECRET}\" --dry-run=client
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > entitledregistryconfigjson\n
          \   # Get the current global secrets on the cluster and store it as a file
          named dockerconfigjson\n    oc get secret pull-secret -n openshift-config
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > dockerconfigjson\n
          \   # Merge the two dockerconfigs together into a file called dockerconfigjson-merged.\n
          \   jq -s '.[0] * .[1]' dockerconfigjson entitledregistryconfigjson > dockerconfigjson-merged\n
          \   # Set the cluster's dockerconfig file to the new merged version.\n    oc
          set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson=dockerconfigjson-merged\n
          \   ```\n  7. (Optional) Your IBM product may have installed IBM Metrics
          Operator using an `ibm-entitlement-key` for authentication to IBM Software
          Central. To authorize the automatic creation of an IBM Software Central
          account for the IBM account owning the `ibm-entitlement-key`\n      ```\n
          \     oc patch marketplaceconfig marketplaceconfig -n ibm-software-central
          --type='merge' -p '{\"spec\": {\"authorizeAccountCreation\": true}}'\n      ```\n\n###
          Why is a global pull secret required?\nIn order to successfully install
          the products hosted by the container image registry, you will need to make
          the pull secret available across the cluster. This can be achieved by applying
          the pull as a [global pull secret](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html#images-update-global-pull-secret_using-image-pull-secrets).
          For alternative approaches, please see the official OpenShift [documentation](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html).\n\n\n###
          SecurityContextConstraints requirements\n\nThe Operators and their components
          support running under the OpenShift Container Platform default restricted
          and restricted-v2 security context constraints.\n\n### Installation Namespace
          and ClusterRoleBinding requirements\n\nThe IBM Metrics Operator components
          require specific ClusterRoleBindings.\n- The metric-state component requires
          a ClusterRoleBinding for the the `view` ClusterRole.\n  - Ability to read
          non-sensitive CustomResources. \n  - The `view` ClusterRole dynamically
          adds new CustomResourceDefinitions.\n- The operator & reporter component
          requires a ClusterRoleBinding for the the `cluster-monitoring-view` ClusterRole.\n
          \ - Ability to view Prometheus metrics. \n  - The underlying ClusterRole
          RBAC often updated between OpenShift versions.\n\nDue to limitations of
          Operator Lifecycle Manager (OLM), this ClusterRoleBinding can not be provided
          dynamically for arbitrary installation target namespaces.\n\nA static ClusterRoleBinding
          is included for installation to the default namespace of `ibm-software-central`,
          and namespaces `openshift-redhat-marketplace`,  `redhat-marketplace`, `ibm-common-services`.\n\nTo
          create the ClusterRoleBindings for installation to an alternate namespace\n```\noc
          project INSTALL-NAMESPACE\noc adm policy add-cluster-role-to-user view -z
          ibm-metrics-operator-metric-state\noc adm policy add-cluster-role-to-user
          cluster-monitoring-view -z ibm-metrics-operator-controller-manager,ibm-metrics-operator-reporter\n```\n\n###
          Metric State scoping requirements\nThe metric-state Deployment obtains `get/list/watch`
          access to metered resources via the `view` ClusterRole. For operators deployed
          using Operator Lifecycle Manager (OLM), permissions are added to `clusterrole/view`
          dynamically via a generated and annotated `-view` ClusterRole. If you wish
          to meter an operator, and its Custom Resource Definitions (CRDs) are not
          deployed through OLM, one of two options are required\n1. Add the following
          label to a clusterrole that has get/list/watch access to your CRD: `rbac.authorization.k8s.io/aggregate-to-view:
          \"true\"`, thereby dynamically adding it to `clusterrole/view`\n2. Create
          a ClusterRole that has get/list/watch access to your CRD, and create a ClusterRoleBinding
          for the metric-state ServiceAccount\n\nAttempting to meter a resource with
          a MeterDefinition without the required permissions will log an `AccessDeniedError`
          in metric-state.\n\n### Disaster Recovery\n\nTo plan for disaster recovery,
          note the PhysicalVolumeClaims `rhm-data-service-rhm-data-service-N`. \n-
          In connected environments, MeterReport data upload attempts occur hourly,
          and are then removed from data-service. There is a low risk of losing much
          unreported data.\n- In an airgap environment, MeterReport data must be pulled
          from data-service and uploaded manually using [datactl](https://github.com/redhat-marketplace/datactl/).
          To prevent data loss in a disaster scenario, the data-service volumes should
          be considered in a recovery plan.\n\n### Subscription Config\n\nIt is possible
          to [configure](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md)
          how OLM deploys an Operator via the `config` field in the [Subscription](https://github.com/operator-framework/olm-book/blob/master/docs/subscriptions.md)
          object.\n\nThe IBM Metrics Operator will also read the `config` and append
          it to the operands. The primary use case is to control scheduling using
          [Tolerations](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/)
          and [NodeSelectors](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/).\n\nA
          limitation is that the `config` elements are only appended to the operands.
          The elements in the operands are not removed if the `config` is removed
          from the `Subscripion`. The operand must be modified manually, or deleted
          and recreated by the controller.\n\n### Cluster permission requirements\n\n|API
          group             |Resources              |Verbs                                     |\n|----------------------|-----------------------|------------------------------------------|\n|apps
          \                 |statefulsets           |get;list;watch                            |\n|authentication.k8s.io
          |tokenreviews           |create                                    |\n|authorization.k8s.io
          \ |subjectaccessreviews   |create                                    |\n|config.openshift.io
          \  |clusterversions        |get;list;watch                            |\n|marketplace.redhat.com|meterdefinitions
          \      |get;list;watch;create;update;patch;delete |\n|operators.coreos.com
          \ |clusterserviceversions |get;list;watch                            |\n|operators.coreos.com
          \ |subscriptions          |get;list;watch                            |\n|operators.coreos.com
          \ |operatorgroups         |get;list                                  |\n|storage.k8s.io
          \       |storageclasses         |get;list;watch                            |\n|
          \                     |namespaces             |get;list;watch                            |\n|
          \                     |configmaps             |get;list;watch                            |\n|
          \                     |services               |get;list;watch                            |\n|
          \                     |**clusterrole/view**   |                                          |\n\n\n###
          Documentation\nYou can find our documentation [here.](https://swc.saas.ibm.com/en-us/documentation/)\n\n###
          Getting help\nIf you encounter any issues while using the operators, you
          can create an issue on our [Github\nrepo](https://github.com/redhat-marketplace/redhat-marketplace-operator)
          for bugs, enhancements, or other requests. You can also visit our main page
          and\nreview our [support](https://swc.saas.ibm.com/en-us/support) and [documentation](https://swc.saas.ibm.com/en-us/documentation/).\n\n###
          Readme\nYou can find our readme [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/README.md)\n\n###
          License information\nYou can find our license information [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/LICENSE)\n"
        displayName: IBM Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        - buy
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/redhat-marketplace-data-service@sha256:30fd8d2996b921cfce51f697ee4490a6d9a836021952429b877336f5edf2749d
        - quay.io/rh-marketplace/redhat-marketplace-reporter@sha256:2bd098f057ca07f2b14be9b60ea6ad8ca6b0cae1d226b1ca3b27e23d04ccdf20
        - quay.io/rh-marketplace/redhat-marketplace-metric-state@sha256:e4014c9a546c10e84b13a09db6a18431ee9489ad1296808635e48d9ef4963cf1
        - quay.io/rh-marketplace/redhat-marketplace-authcheck@sha256:13a81b8aca79ae8c89ee24dcc30f3322a4cfdedf929e7bb984295c1e7e631fa9
        - quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 2.20.0
      entries:
      - name: ibm-metrics-operator.v2.20.0
        version: 2.20.0
      name: stable
    defaultChannel: stable
    packageName: ibm-metrics-operator
    provider:
      name: Red Hat Marketplace
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-dr-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-dr-manager.v8.0.220
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com",
                  "resRecord": {
                    "HostName_1": "100.200.30.41",
                    "HostName_2": "100.200.30.43",
                    "abc": "HostName_2",
                    "www": "HostName_1",
                    "xyz": "HostName_3"
                  },
                  "stealthMasters": [
                    "1.2.3.4"
                  ],
                  "tsigKeyFile": "/Kvcslnxdns.com.+157+10641.private"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "dataReplicationPointer": "test-datareplication",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 10,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          olm.skipRange: <8.0.220
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
            - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:a98f9cc51f2c7baeec6dbb4933a42ae8ae20ff1a42c9d48a37174cf5211bbcf8
        version: 8.0.220
      entries:
      - name: infoscale-dr-manager.v8.0.220
        version: 8.0.220
      - name: infoscale-dr-manager.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-dr-manager.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          createdAt: "2024-06-17T10:14:13Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:c46cbf78631fc790c77b6c13c1d1d4b5ab35f49164f7c2e36c7ec59220cbdce2
        version: 8.0.330
      entries:
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-dr-manager.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
          createdAt: "2024-02-22T14:14:42Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721648-168743109)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721654-168743109)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
        version: 8.0.410
      entries:
      - name: infoscale-dr-manager.v8.0.410
        version: 8.0.410
      - name: infoscale-dr-manager.v8.0.409
        version: 8.0.409
      - name: infoscale-dr-manager.v8.0.400
        version: 8.0.400
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-dr-manager.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
          createdAt: "2024-02-22T14:14:42Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721648-168743109)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721654-168743109)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
        version: 8.0.410
      entries:
      - name: infoscale-dr-manager.v8.0.410
        version: 8.0.410
      - name: infoscale-dr-manager.v8.0.409
        version: 8.0.409
      - name: infoscale-dr-manager.v8.0.400
        version: 8.0.400
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-dr-manager.v8.0.100
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com",
                  "resRecord": {
                    "abc": "r7515-054-vm9",
                    "r7515-054-vm10": "10.221.85.83",
                    "r7515-054-vm8": "10.221.85.81",
                    "r7515-054-vm9": "10.221.85.82",
                    "www": "r7515-054-vm8",
                    "xyz": "r7515-054-vm10"
                  },
                  "stealthMasters": [
                    "1.2.3.4"
                  ],
                  "tsigKeyFile": "/Kvcslnxdns.com.+157+10641.private"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "currentPrimary": "Clus1",
                  "force": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "fail",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "dataReplicationPointer": "test-datareplication",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 10,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:eb999187a1f9a9ce5566089e24adf598b1c229d31c5e57ddcfe4439ba25e97dc
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
            - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membershrip
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membershrip
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponsding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:eb999187a1f9a9ce5566089e24adf598b1c229d31c5e57ddcfe4439ba25e97dc
        version: 8.0.100
      entries:
      - name: infoscale-dr-manager.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-dr-manager
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-s3control-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-s3control-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/s3control-controller:1.1.1
          createdAt: "2025-09-19T20:52:02Z"
          description: AWS S3 Control controller is a service controller for managing
            S3 Control resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessPoint represents the state of an AWS s3control AccessPoint
              resource.
            displayName: AccessPoint
            kind: AccessPoint
            name: accesspoints.s3control.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon S3 Control resources in AWS from within your Kubernetes cluster.

          **About Amazon S3 Control**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon S3 Control
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - s3control
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon S3 Control Developer Resources
          url: https://aws.amazon.com/S3 Control/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: s3control maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/s3control-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-s3control-controller.v1.1.1
        version: 1.1.1
      - name: ack-s3control-controller.v1.1.0
        version: 1.1.0
      - name: ack-s3control-controller.v1.0.11
        version: 1.0.11
      - name: ack-s3control-controller.v1.0.10
        version: 1.0.10
      name: alpha
    defaultChannel: alpha
    packageName: ack-s3control-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: sosivio
      provider-url: https://www.sosiv.io
    name: sosivio
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sosivio-operator.v1.5.1-97
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sosivio.sosiv.io/v1alpha1",
                "kind": "Sosivio",
                "metadata": {
                  "name": "sosivio"
                },
                "spec": {
                  "platform": "ocp4",
                  "useSha": false
                }
              },
              {
                "apiVersion": "sosivio.sosiv.io/v1alpha1",
                "kind": "Sosivio",
                "metadata": {
                  "name": "sosivio"
                },
                "spec": {
                  "platform": "ocp4",
                  "useSha": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Monitoring,Integration & Delivery,Logging & Tracing,Developer
            Tools
          com.redhat.openshift.versions: v4.1-v4.1
          containerImage: quay.io/sosivio/sosivio-operator:v1.5.1-97
          createdAt: "2021-10-28 01:01:01"
          description: Sosivio | Predictive Troubleshooting for Kubernetes
          operatorhub.io/ui-metadata-max-k8s-version: "1.25"
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Sosivio/operator
          support: Sosivio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Sosivio is the Schema for the sosivios API
            displayName: Sosivio
            kind: Sosivio
            name: sosivios.sosivio.sosiv.io
            version: v1alpha1
        description: "# Sosivio Operator\nThe Sosivio Operator for RedHat’s Openshift
          Container Platform allows users to quickly and easily deploy Sosivio’s Kubernetes
          Predictive Troubleshooting Platform on your cluster.\n\n# Sosivio Overview\n\nSosivio
          is the first and only non-intrusive and predictive autonomous cloud-native
          platform built specifically for Kubernetes from the ground up. Sosivio provides
          predictive troubleshooting tools, root cause determination and instant remediation
          for any failures in Kubernetes environments.\n\nPredict and Prevent!\n\nThe
          Operator will deploy and manage Sosivio on Kubernetes and OpenShift. Included
          are the following features;\n-   **Root Cause Analysis and Remediation**
          - predictive troubleshooting to prevent failures before they occur. Instantly
          understand where the root cause is and how to fix it.\n-   **Application
          Profiling** - analyze container resource consumption and recommendations
          for requests and limits.\n-   **Real-time Metrics** - in-depth, granular,
          and accurate real-time data from the process level inside containers (and
          pods) with failure correlation.\n-   **Cluster Health Checks** - proactive
          health checks for cluster infrastructure and applications running on the
          cluster.\n-   **Multi-Cluster View** - management, monitoring, and maintenance
          capabilities for all clusters from one location. \n-   **Kubernetes Components
          Alerting** - notifications sent through Slack, email and other 3rd party
          platforms.\n\n# Installation\nSosivio is completely **non-intrusive** and
          **does not require any persistent storage**. There is **no data offloading**
          outside of your Kubernetes cluster. Applying our Data Swirling methodology
          enables the product to perform at a **remarkably low resource footprint**
          and requires **no persistent volume configurations**.\n\n\nSosivio is installed
          under one namespace (labeled: \"sosivio\") which creates all the required
          components for the product.\n\nAll users will start with 4 weeks of Sosivio
          Premium after which Sosivio will automatically convert to the Community
          Version (free forever).\n\n# Help\nMore information about Sosivio and our
          product usage instructions can be found in our documentation [here](https://docs.sosiv.io/)
          \ \n"
        displayName: Sosivio | Predictive Troubleshooting for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sosivio
        - kubernetes
        - observanility
        - monitoring
        - root cause
        links:
        - name: Sosivio
          url: https://www.sosiv.io
        - name: Documentation
          url: https://docs.sosiv.io
        maintainers:
        - email: info@sosiv.io
          name: Sosivio
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: sosivio
          url: https://www.sosiv.io
        relatedImages:
        - quay.io/sosivio/classifier@sha256:dd8bf145524db23ff333efa1bcdff258dab116d3534f95f73775d8dd339e6963
        - quay.io/sosivio/notifier@sha256:35f5f3e8e5ccad914fcfe34fb4b4a1b044ee87f5a6141284cd5273f0b2b3bf22
        - quay.io/sosivio/kube-rbac-proxy@sha256:34e8724e0f47e31eb2ec3279ac398b657db5f60f167426ee73138e2e84af6486
        - quay.io/sosivio/nsq@sha256:781aac23ef0a1ed0af2b8e93d625ae49bd11dbdd98787d3c79dde8d128f3810d
        - quay.io/sosivio/network-aggregator@sha256:e59d937f4d9c52acad0adadab39c6849d20b4ee72a83e239a33ea137f08323f4
        - quay.io/sosivio/sosivio-dashboard@sha256:fdfca5bb2123d1008cec2b656cde73c7e9364f84c5ff4f4a6fee5751b1bfb2f8
        - quay.io/sosivio/kube-rbac-proxy:v0.8.0
        - quay.io/sosivio/sequence-recognition@sha256:7335cf8f3e40bc46e8ca0de525d99f0655833b7c52d36dc273aa7b096eb70fa2
        - quay.io/sosivio/sosivio-node-pinger@sha256:937de52ad7b76e6cff4ad5197de9705988d61f285ebadac5ebadb5856f59ee58
        - quay.io/sosivio/actuator@sha256:a9abe9730800ca2b0a1bbd7a361b2ab3a818cbf127798314e5477fb1cb16ba4c
        - quay.io/sosivio/authentication@sha256:0ea6025f333576792f5b1fc4d96d9ab1a404749f185830410e71248683004da2
        - quay.io/sosivio/sosiviodb@sha256:a3c6c3975bc2b0316135ee5c547f1eb215f3a1f9c8b0a9c8eceff23415f93597
        - quay.io/sosivio/sosivio-node-collector@sha256:61ed8dd52de5edc09d16a31b18f2f65ad9344c0b6c9cfab323c3eb3d7e7ed060
        - quay.io/sosivio/sosivio-operator@sha256:11bc627773bb5e08b629e0b013e7f600fcb4aaefbd97f2de337d339475616763
        - quay.io/sosivio/contract-testing@sha256:cad6fac6ed3fdf2597ceb09ba5dbb7d81f70cf45f483a7064e3968215dd982e3
        - quay.io/sosivio/crud-manager@sha256:38289a28829fa737917d76dd4810ba6140e123061ead0a4b79fa79b2e246a293
        - quay.io/sosivio/communicator@sha256:fffec9a6c66594fde5f05e2dfb84361544ca9d53a345188c197a0ea6f53bc4b9
        - quay.io/sosivio/analyzer@sha256:3825608fe7c43ef059108e76665acc99fe72bea77d4d9e3852f6ad45e74379b1
        - quay.io/sosivio/discovery-engine@sha256:b87b3f7c9ed3aed80d0a77788fd8925789450d9dc05d25017e89dff611cd82da
        - quay.io/sosivio/sosivio-operator:v1.5.1-97
        - quay.io/sosivio/correlation-entities@sha256:c00af09b2589bad4740fd44d7a81c69b2e29138c360939b2a45f65edd44b6901
        - quay.io/sosivio/sosivio-node-netpacket@sha256:84dd159b1dfdb52dee3eb1c4c93f323bae49ce517ff44663b69978e14bdd8390
        version: 1.5.1
      entries:
      - name: sosivio-operator.v1.5.1-97
        version: 1.5.1
      - name: sosivio-operator.v1.5.0-95
        version: 1.5.0
      - name: sosivio-operator.v1.4.4-85
        version: 1.4.4
      - name: sosivio-operator.v1.4.3-83
        version: 1.4.3
      - name: sosivio-operator.v1.4.2-82
        version: 1.4.2
      - name: sosivio-operator.v1.4.1-81
        version: 1.4.1-81
      - name: sosivio-operator.v1.4.1.1-80
        version: 1.4.1-80
      - name: sosivio-operator.v1.4.1-78
        version: 1.4.1-79
      - name: sosivio-operator.v1.3.1-76
        version: 1.3.1-76
      - name: sosivio-operator.v1.3.0-75
        version: 1.3.0-75
      - name: sosivio-operator.v1.2.1-1
        version: 1.2.1-1
      - name: sosivio-operator.v1.2.0-1
        version: 1.2.0-1
      name: stable
    defaultChannel: stable
    packageName: sosivio
    provider:
      name: sosivio
      url: https://www.sosiv.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-keycloak-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-keycloak-operator.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2023-09-19T12:43:29Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KCOSClientRoleMapping
            name: kcosclientrolemappings.keycloak.org
            version: v1alpha1
          - kind: KCOSRealmRole
            name: kcosrealmroles.keycloak.org
            version: v1alpha1
          - kind: KCOSUserRoleMapping
            name: kcosuserrolemappings.keycloak.org
            version: v1alpha1
          - kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: Keysight Keycloak Operator
        displayName: Keysight Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Keycloak Operator
          url: https://keysight-keycloak-operator.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak@sha256:772f73cce6a77b9c46f8035faac99a28602219657d18b75b19b4d905d91942c9
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak-postgresql@sha256:b636955199065e18f9d4679835ae78214843258e5abb4ae758288d6f00df2f8e
        - prod.harbor.keysight.digital/loadcore/keysight-keycloak-operator@sha256:d74e7e98a114768a6f3327edabce9336743a078844fffee985053b6a90100da8
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak-init-container@sha256:90965beaf41c760cb8c8bf906acf08f8fe4770d313de8dfd92d18d7c25348a38
        version: 0.1.21
      entries:
      - name: keysight-keycloak-operator.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-keycloak-operator
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: fence-agents-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fence-agents-remediation.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  }
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:00cc420250d8e25d9da7c43d9c4cff35c7edeb7d390af754c48c0ce00e9e7770
          createdAt: "2023-10-30 14:06:57"
          description: Fence Agents Remediation Operator for remediating nodes using
            upstream fence-agents.
          olm.skipRange: '>=0.2.0 <0.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: "Fence Agents Remediation Operator (FAR)\n\nThe operator will
          remediate/fence a node when a FenceAgentsRemediation (far) custom resource
          exists with the node's name.\nFAR uses an existing set of upstream fencing
          agents for environments with a \ntraditional API end-point (e.g., IPMI)
          for power cycling cluster nodes.\n\nThe operator is recommended to be used
          with Node Healthcheck Operator for easier and smoother experience,\nbut
          it can also work as a standalonde remediator for the more advanced user.\n"
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:00cc420250d8e25d9da7c43d9c4cff35c7edeb7d390af754c48c0ce00e9e7770
        version: 0.2.1
      entries:
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.14-eus
    - currentCSV: fence-agents-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:a367e8fdd80938e97ac195631a996e0c537bf3666aef0199147534762dcf739f
          createdAt: "2024-08-10 22:49:23"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.2.0 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:a367e8fdd80938e97ac195631a996e0c537bf3666aef0199147534762dcf739f
        version: 0.4.1
      entries:
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.16-eus
    - currentCSV: fence-agents-remediation.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-16 10:25:57"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.5.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.5/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel9-operator@sha256:da5975557078fbfa6882df87cff5bda5671b9838b5fb0a738fea724469d03576
        version: 0.5.1
      entries:
      - name: fence-agents-remediation.v0.5.1
        version: 0.5.1
      - name: fence-agents-remediation.v0.5.0
        version: 0.5.0
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.18-eus
    - currentCSV: fence-agents-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  }
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:5c3778c7aae45f54e949fbae3d133777f7aa78fe1251ac3d5510aa28b71e2ba4
          createdAt: "2023-04-20 15:36:05"
          description: Fence Agents Remediation Operator for remediating nodes using
            upstream fence-agents.
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: "Fence Agents Remediation Operator (FAR)\n\nThe operator will
          remediate/fence a node when a FenceAgentsRemediation (far) custom resource
          exists with the node's name.\nFAR uses an existing set of upstream fencing
          agents for environments with a \ntraditional API end-point (e.g., IPMI)
          for power cycling cluster nodes.\n\nThe operator is recommended to be used
          with Node Healthcheck Operator for easier and smoother experience,\nbut
          it can also work as a standalonde remediator for the more advanced user.\n"
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3af07bfe0fb73b1a7f00487eba302d6da442e742ab4007fc50e52efd4d45dae5
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:5c3778c7aae45f54e949fbae3d133777f7aa78fe1251ac3d5510aa28b71e2ba4
        version: 0.1.0
      entries:
      - name: fence-agents-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: fence-agents-remediation.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeSecretNames": {
                    "master-0": "fence-agents-credentials-master0",
                    "master-1": "fence-agents-credentials-master1",
                    "master-2": "fence-agents-credentials-master2",
                    "worker-0": "fence-agents-credentials-worker0",
                    "worker-1": "fence-agents-credentials-worker1",
                    "worker-2": "fence-agents-credentials-worker2"
                  },
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedSecretName": "fence-agents-credentials-shared",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {
                    "spec": {
                      "agent": "fence_ipmilan",
                      "nodeSecretNames": {
                        "master-0": "fence-agents-credentials-master0",
                        "master-1": "fence-agents-credentials-master1",
                        "master-2": "fence-agents-credentials-master2",
                        "worker-0": "fence-agents-credentials-worker0",
                        "worker-1": "fence-agents-credentials-worker1",
                        "worker-2": "fence-agents-credentials-worker2"
                      },
                      "nodeparameters": {
                        "--ipport": {
                          "master-0": "6230",
                          "master-1": "6231",
                          "master-2": "6232",
                          "worker-0": "6233",
                          "worker-1": "6234",
                          "worker-2": "6235"
                        }
                      },
                      "remediationStrategy": "ResourceDeletion",
                      "retrycount": 5,
                      "retryinterval": "5s",
                      "sharedSecretName": "fence-agents-credentials-shared",
                      "sharedparameters": {
                        "--action": "reboot",
                        "--ip": "192.168.111.1",
                        "--lanplus": "",
                        "--username": "admin"
                      },
                      "timeout": "60s"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-09-08 09:19:00"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.5.0 <0.6.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.8/html/remediation_fencing_and_maintenance/about-remediation-fencing-maintenance
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel9-operator@sha256:eef1a298718650a22a3da4a07c140ce21b6f287d87af79801610317fa8f63d52
        version: 0.6.0
      entries:
      - name: fence-agents-remediation.v0.6.0
        version: 0.6.0
      - name: fence-agents-remediation.v0.5.1
        version: 0.5.1
      - name: fence-agents-remediation.v0.5.0
        version: 0.5.0
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: stable
    defaultChannel: stable
    packageName: fence-agents-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ssm-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ssm-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ssm-controller:1.1.1
          createdAt: "2025-09-19T20:57:07Z"
          description: AWS SSM controller is a service controller for managing SSM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Document represents the state of an AWS ssm Document resource.
            displayName: Document
            kind: Document
            name: documents.ssm.services.k8s.aws
            version: v1alpha1
          - description: PatchBaseline represents the state of an AWS ssm PatchBaseline
              resource.
            displayName: PatchBaseline
            kind: PatchBaseline
            name: patchbaselines.ssm.services.k8s.aws
            version: v1alpha1
          - description: ResourceDataSync represents the state of an AWS ssm ResourceDataSync
              resource.
            displayName: ResourceDataSync
            kind: ResourceDataSync
            name: resourcedatasyncs.ssm.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon SSM resources in AWS from within your Kubernetes cluster.

          **About Amazon SSM**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon SSM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ssm
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SSM Developer Resources
          url: https://aws.amazon.com/SSM/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ssm maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ssm-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ssm-controller.v1.1.1
        version: 1.1.1
      - name: ack-ssm-controller.v1.1.0
        version: 1.1.0
      - name: ack-ssm-controller.v1.0.12
        version: 1.0.12
      - name: ack-ssm-controller.v1.0.10
        version: 1.0.10
      - name: ack-ssm-controller.v1.0.9
        version: 1.0.9
      - name: ack-ssm-controller.v1.0.8
        version: 1.0.8
      - name: ack-ssm-controller.v1.0.7
        version: 1.0.7
      - name: ack-ssm-controller.v1.0.6
        version: 1.0.6
      - name: ack-ssm-controller.v1.0.5
        version: 1.0.5
      - name: ack-ssm-controller.v1.0.4
        version: 1.0.4
      - name: ack-ssm-controller.v1.0.3
        version: 1.0.3
      - name: ack-ssm-controller.v1.0.2
        version: 1.0.2
      - name: ack-ssm-controller.v1.0.1
        version: 1.0.1
      - name: ack-ssm-controller.v1.0.0
        version: 1.0.0
      - name: ack-ssm-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-ssm-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-efs-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-efs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "AccessPoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "FileSystem",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "MountTarget",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/efs-controller:1.1.1
          createdAt: "2025-09-19T18:38:13Z"
          description: AWS EFS controller is a service controller for managing EFS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessPoint represents the state of an AWS efs AccessPoint
              resource.
            displayName: AccessPoint
            kind: AccessPoint
            name: accesspoints.efs.services.k8s.aws
            version: v1alpha1
          - description: FileSystem represents the state of an AWS efs FileSystem
              resource.
            displayName: FileSystem
            kind: FileSystem
            name: filesystems.efs.services.k8s.aws
            version: v1alpha1
          - description: MountTarget represents the state of an AWS efs MountTarget
              resource.
            displayName: MountTarget
            kind: MountTarget
            name: mounttargets.efs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EFS resources in AWS from within your Kubernetes cluster.

          **About Amazon Elastic File System (Amazon EFS)**

          Amazon Elastic File System (Amazon EFS) provides serverless, fully elastic file storage so that you can share file data without provisioning or managing storage capacity and performance. Amazon EFS is built to scale on demand to petabytes without disrupting applications, growing and shrinking automatically as you add and remove files. Because Amazon EFS has a simple web services interface, you can create and configure file systems quickly and easily. The service manages all the file storage infrastructure for you, meaning that you can avoid the complexity of deploying, patching, and maintaining complex file system configurations.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Elastic File System (Amazon
          EFS)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - efs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EFS Developer Resources
          url: https://aws.amazon.com/efs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: efs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/efs-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-efs-controller.v1.1.1
        version: 1.1.1
      - name: ack-efs-controller.v1.1.0
        version: 1.1.0
      - name: ack-efs-controller.v1.0.19
        version: 1.0.19
      - name: ack-efs-controller.v1.0.18
        version: 1.0.18
      - name: ack-efs-controller.v1.0.17
        version: 1.0.17
      - name: ack-efs-controller.v1.0.16
        version: 1.0.16
      - name: ack-efs-controller.v1.0.15
        version: 1.0.15
      - name: ack-efs-controller.v1.0.14
        version: 1.0.14
      - name: ack-efs-controller.v1.0.13
        version: 1.0.13
      - name: ack-efs-controller.v1.0.12
        version: 1.0.12
      - name: ack-efs-controller.v1.0.11
        version: 1.0.11
      - name: ack-efs-controller.v1.0.10
        version: 1.0.10
      - name: ack-efs-controller.v1.0.8
        version: 1.0.8
      - name: ack-efs-controller.v1.0.6
        version: 1.0.6
      - name: ack-efs-controller.v1.0.5
        version: 1.0.5
      - name: ack-efs-controller.v1.0.3
        version: 1.0.3
      - name: ack-efs-controller.v1.0.2
        version: 1.0.2
      - name: ack-efs-controller.v1.0.0
        version: 1.0.0
      - name: ack-efs-controller.v0.0.9
        version: 0.0.9
      - name: ack-efs-controller.v0.0.8
        version: 0.0.8
      - name: ack-efs-controller.v0.0.7
        version: 0.0.7
      - name: ack-efs-controller.v0.0.6
        version: 0.0.6
      - name: ack-efs-controller.v0.0.5
        version: 0.0.5
      - name: ack-efs-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-efs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solace Corporation
      provider-url: www.solace.com
    name: pubsubplus-eventbroker-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pubsubplus-eventbroker-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pubsubplus.solace.com/v1beta1",
                "kind": "PubSubPlusEventBroker",
                "metadata": {
                  "name": "non-ha-standalone-example"
                },
                "spec": {
                  "redundancy": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "true"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.10
          containerImage: registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator:1.4.0
          createdAt: "2025-07-31T19:48:36Z"
          description: The Solace PubSub+ Event Broker Operator deploys and manages
            the lifecycle of PubSub+ Event Brokers
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
          support: Solace Products
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PubSub+ Event Broker
            displayName: PubSub+ Event Broker
            kind: PubSubPlusEventBroker
            name: pubsubpluseventbrokers.pubsubplus.solace.com
            version: v1beta1
        description: |
          ## Solace PubSub+ Platform
          Solace [PubSub+ Platform](https://solace.com/products/platform/) is a complete event streaming and management platform
          for real-time enterprises. The [PubSub+ Software Event Broker](https://solace.com/products/event-broker/software/) efficiently streams
          event-driven information between applications, IoT devices, and user interfaces running in the cloud, on-premises,
          and in hybrid environments using open APIs and protocols like AMQP, JMS, MQTT, REST and WebSocket.
          It can be installed into a variety of public and private clouds, PaaS, and on-premises environments.
          Event brokers in multiple locations can be linked together in an [Event Mesh](https://solace.com/what-is-an-event-mesh/)
          to dynamically share events across the distributed enterprise.

          ## Solace PubSub+ Event Broker Operator
          Solace Pubsub+ Event Broker Operator automatically deploys and manages PubSub+ Software Event Brokers on Kubernetes and OpenShift environments.

          Features include:
          * Initial deployment configuration: Production-ready HA, non-HA or for developers
          * Rolling update for configuration changes or broker upgrades
          * Prometheus monitoring support

          ## Getting started
          Refer to the [Solace PubSub+ Event Broker Operator Quick Start](https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/README.md)
        displayName: Solace PubSub+ Event Broker Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - solace
        - pubsubplus
        - pubsub+
        - pubsub
        - messaging
        - advanced event broker
        - event broker
        - event mesh
        - event streaming
        - data streaming
        - event integration
        - middleware
        links:
        - name: Operator QuickStart
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
        - name: Operator Documentation
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/docs/EventBrokerOperatorUserGuide.md
        - name: Solace.com
          url: https://solace.com
        - name: Understanding Solace Technology
          url: https://solace.com/products/platform/
        - name: Solace PubSub+ Technical Documentation
          url: https://docs.solace.com/
        - name: Solace Developer Portal
          url: https://www.solace.dev/
        - name: Solace Community
          url: https://solace.community/
        maintainers:
        - email: support@solace.com
          name: Solace R&D
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Solace Corporation
          url: www.solace.com
        relatedImages:
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator:1.4.0
        version: 1.4.0
      entries:
      - name: pubsubplus-eventbroker-operator.v1.4.0
        version: 1.4.0
      - name: pubsubplus-eventbroker-operator.v1.3.0
        version: 1.3.0
      - name: pubsubplus-eventbroker-operator.v1.2.0
        version: 1.2.0
      - name: pubsubplus-eventbroker-operator.v1.1.2
        version: 1.1.2
      - name: pubsubplus-eventbroker-operator.v1.1.1
        version: 1.1.1
      - name: pubsubplus-eventbroker-operator.v1.1.0
        version: 1.1.0
      - name: pubsubplus-eventbroker-operator.v1.0.3
        version: 1.0.3
      - name: pubsubplus-eventbroker-operator.v1.0.2
        version: 1.0.2
      - name: pubsubplus-eventbroker-operator.v1.0.1
        version: 1.0.1
      - name: pubsubplus-eventbroker-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pubsubplus-eventbroker-operator
    provider:
      name: Solace Corporation
      url: www.solace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: ibm-data-reporter-operator
      app.kubernetes.io/managed-by: ibm-data-reporter-operator
      app.kubernetes.io/name: ibm-data-reporter-operator
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: ""
    name: ibm-data-reporter-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-data-reporter-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "DataReporterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-data-reporter-operator",
                    "app.kubernetes.io/instance": "datareporterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datareporterconfig",
                    "app.kubernetes.io/part-of": "ibm-data-reporter-operator"
                  },
                  "name": "datareporterconfig"
                },
                "spec": {
                  "userConfig": [
                    {
                      "metadata": {
                        "ameta1": "ametadata1",
                        "bmeta1": "bmetadata1",
                        "cmeta1": "cmetadata1",
                        "dmeta1": "dmetadata1"
                      },
                      "userName": "system:serviceaccount:redhat-marketplace:ibm-data-reporter-operator-api"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
          createdAt: ""
          description: The IBM Data Reporter Operator accepts events and transforms
            them into reports submitted to the Data Service of the IBM Metrics Operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReporterConfig is the Schema for the datareporterconfigs
              API
            displayName: Data Reporter Config
            kind: DataReporterConfig
            name: datareporterconfigs.marketplace.redhat.com
            version: v1alpha1
          required:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
        description: "# IBM Data Reporter Operator\n\n# Introduction\n\nThe IBM Data
          Reporter Operator accepts events and transforms them into reports submitted
          to the IBM Metrics Operator Data Service.\n\n# Details\n\nThe IBM Data Reporter
          Operator deploys a service that provides an endpoint to which callers can
          send raw json event data. The event data is transformed into a report and
          is sent to the IBM Metrics Operator's Data Service. The IBM Metrics Operator
          Data Service periodically uploads the reports to IBM Software Central. It
          can also filter, transform and forward events to alternate endpoints.\n\n##
          Prerequisites\n\n- OpenShift Container Platform, major version 4 with any
          available supported minor version\n- Install IBM Metrics Operator\n  - IBM
          Data Reporter Operator prerequisties the IBM Metrics Operator's Data Service\n
          \ - Register the Cluster by creating a `redhat-marketplace-pull-secret`,
          as per the instructions\n  - Once `rhm-data-service` has started IBM Data
          Reporter Operator will accept requests\n\n## SecurityContextConstraints
          Requirements\n\n- The operator runs under Red Hat restricted SCC\n\n## Resources
          Required\n\n- The operator requires at least 85 Mi memory.\n\n## Limitations\n\n-
          Only runs on amd64, s390x and ppc64le architectures\n\n## Installing\n\n-
          A user with the cluster administrator role.\n- Install this operator in
          the same namespace as the IBM Metrics Operator\n  - default namespace: `ibm-software-central`\n\n##
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n# Configuration\n\nOptional:\n- Configure the
          Custom Resource Definition `DataReporterConfig` named `datareporterconfig`
          as per the following sample\n  - Configuration that can not be reconciled
          successfully will be reported on the DataReporterConfig Status\n  - jsonPath
          expressions are handled by: https://github.com/ohler55/ojg\n    - jsonPath
          comparison: https://cburgmer.github.io/json-path-comparison/\n  - Transformer
          type currently supported is `kazaam`\n    - https://github.com/willie68/kazaam\n\nSample
          DataReporterConfig:\n```YAML\napiVersion: marketplace.redhat.com/v1alpha1\nkind:
          DataReporterConfig\nmetadata:\n  name: datareporterconfig\nspec:\n  # true:
          report generated for 1 event and provide return code from data-service,
          confirming delivery\n  # false: events will be accumulated before building
          report, and immediately return 200, delivery unfconfirmed\n  confirmDelivery:
          false\n  dataFilters:   # dataFilters matches the first dataFilter selector,
          applies transforms and forwards to data-service and destinations\n  - altDestinations:
          \  # List of desintations to transform and foward events to\n    - authorization:
          \  # optional endpoint if an authorization token must first be requested
          \n        authDestHeader: Authorization   # http header key to be appended
          to destination request\n        authDestHeaderPrefix: 'Bearer '   # optional
          prefix for constructing destination request http header value\n        bodyData:
          \  # body data to send to authorization request, often an apiKey\n          secretKeyRef:\n
          \           name: auth-body-data-secret\n            key: bodydata\n        header:
          \  # secret map to use for authorization request headers\n          secret:\n
          \           name: auth-header-map-secret\n        tokenExpr: $.token   #
          optionally extract the returned body data using a jsonPath expression\n
          \       url: https://127.0.0.1:8443/api/2.0/accounts/my_account_id/apikeys/token
          \  # the url to make an authorization request to\n      header:   # secret
          map to use for destination request headers\n        secret:\n          name:
          dest-header-map-secret\n      transformer:   # The transformer to apply
          to event data for this destination\n        configMapKeyRef:\n          key:
          kazaam.json\n          name: kazaam-configmap\n        type: kazaam\n      url:
          https://127.0.0.1:8443/metering/api/1.0/usage   # the url to send the event
          to\n      urlSuffixExpr: $.properties.productId   # optional jsonPath expression
          on the event to extract and set a suffix to the url path\n    manifestType:
          dataReporter   # override manifestType of report sent to data-service\n
          \   selector:   # Matches against event data and a user. Empty selector
          matches all.\n      matchExpressions:   # jsonPath expression must return
          a result for all expressions to match (AND)\n      - $[?($.event == \"Account
          Contractual Usage\")]\n      - $.properties.productId\n      - $[?($.properties.source
          != null)]\n      - $[?($.properties.unit == \"AppPoints\")]\n      - $[?($.properties.quantity
          >= 0)]\n      - $[?($.timestamp != null)]\n      matchUsers:   # must match
          one of these users (OR). Omitting users matches any user\n      - system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n
          \   transformer:   # The transformer to apply to the event for data-service\n
          \     configMapKeyRef:\n        key: kazaam.json\n        name: kazaam-configmap\n
          \     type: kazaam\n  tlsConfig:   # TLS configuration for requests outbound
          to destinations\n    caCerts:\n    - name: tls-ca-certificates\n      key:
          ca.crt\n    certificates:\n    - clientCert:\n        secretKeyRef:\n          name:
          tls-ca-certificates\n          key: tls.crt\n      clientKey:\n        secretKeyRef:\n
          \         name: tls-ca-certificates\n          key: tls.key\n    cipherSuites:
          \n    - TLS_AES_128_GCM_SHA256\n    - TLS_AES_256_GCM_SHA384\n    - TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\n    - TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\n    insecureSkipVerify: false\n
          \   minVersion: VersionTLS12\n  userConfig:   # Optional metadata to apply
          to report sent to data-service for a specific user\n  - metadata:\n      ameta1:
          ametadata1\n      bmeta1: bmetadata1\n      cmeta1: cmetadata1\n      dmeta1:
          dmetadata1\n    userName: system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n```\n\nSample
          Header Secrets:\n```YAML\napiVersion: v1\nstringData:\n  accept: | \n    */*\nkind:
          Secret\nmetadata:\n  name: dest-header-map-secret\ntype: Opaque\n---\napiVersion:
          v1\nstringData:\n  accept: application/json\n  Content-Type: application/json\nkind:
          Secret\nmetadata:\n  name: auth-header-map-secret\ntype: Opaque\n```\n\nSample
          Authentication Body Secret:\n```YAML\napiVersion: v1\nstringData:\n  bodydata:
          |\n    {\"apikey\": \"myapikey\"}\nkind: Secret\nmetadata:\n  name: auth-body-data-secret\ntype:
          Opaque\n```\n\nSample TLS Secret:\n```YAML\napiVersion: v1\ndata:\n  ca.crt:
          LS0t...\n  tls.key: LS0t...\n  tls.crt: LS0t...\nkind: Secret\nmetadata:\n
          \ name: tls-ca-certificates\n```\n\nSample Kazaam ConfigMap:\n```YAML\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: kazaam-configmap\ndata:\n  kazaam.json:
          |\n    [\n        {\n            \"operation\": \"timestamp\",\n            \"spec\":
          {\n                \"timestamp\": {\n                \"inputFormat\": \"2006-01-02T15:04:05.999999+00:00\",\n
          \               \"outputFormat\": \"$unixmilli\"\n                }\n            }\n
          \       },\n        {\n            \"operation\": \"shift\",\n            \"spec\":
          {\n                \"instances[0].instanceId\": \"properties.source\",\n
          \               \"instances[0].startTime\": \"timestamp\",\n                \"instances[0].endTime\":
          \"timestamp\",\n                \"instances[0].metricUsage[0].metricId\":
          \"properties.unit\",\n                \"instances[0].metricUsage[0].quantity\":
          \"properties.quantity\"\n            }\n        },\n        {\n            \"operation\":
          \"default\",\n            \"spec\": {\n                \"meteringModel\":
          \"point-in-time\",\n                \"meteringPlan\": \"contract\"\n            }\n
          \       }\n    ]\n```\n\n### API Service User Configuration\n\n- The ClusterRole
          for api access is `clusterrole/ibm-data-reporter-operator-api`\n- The default
          ServiceAccount provided as an api user is `system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api`\n
          \ - The default ClusterRoleBinding for this user is `clusterrolebinding/ibm-data-reporter-operator-api`\n\nOptional:\n\n-
          To create an additional ServiceAccount and ClusterRoleBinding for api access\n\n```SHELL\nNAMESPACE=$(oc
          config view --minify -o jsonpath='{..namespace}') && \\\noc create serviceaccount
          my-api-service-account && \\\noc create clusterrolebinding ibm-data-reporter-operator-api
          --clusterrole=ibm-data-reporter-operator-api --serviceaccount=${NAMESPACE}:my-api-service-account\n```\n\n-
          Update datareporterconfig to attach metadata to reports associated with
          this user\n\n## API Service Usage\n\n- Get Token & Host\n\n```SHELL\noc
          project ibm-software-central\nexport DRTOKEN=$(oc create token ibm-data-reporter-operator-api
          --namespace ibm-software-central --duration 1h)\nexport DRHOST=$(oc get
          route ibm-data-reporter --template='{{ .spec.host }}')\n```\n\n- Get the
          Status\n\n```SHELL\ncurl -k -H \"Authorization: Bearer ${DRTOKEN}\" https://${DRHOST}/v1/status
          \n```\n\n- Post an Event\n\n```SHELL\ncurl -k -H \"Authorization: Bearer
          ${DRTOKEN}\" -X POST -d '{\"event\":\"myevent\"}' https://${DRHOST}/v1/event\n```\n\n##
          Storage\n\n- The operator temporarily stores events in pod memory, and writes
          to the IBM Metrics Operator data-service, which requires a PersistentVolume\n\n##
          License\n\nCopyright IBM Corporation 2023. All Rights Reserved.\n\nLicensed
          under the Apache License, Version 2.0 (the \"License\");\nyou may not use
          this file except in compliance with the License.\nYou may obtain a copy
          of the License at\n\n    http://www.apache.org/licenses/LICENSE-2.0\n\nUnless
          required by applicable law or agreed to in writing, software\ndistributed
          under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES
          OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for
          the specific language governing permissions and\nlimitations under the License.\n"
        displayName: IBM Data Reporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 2.20.0
      entries:
      - name: ibm-data-reporter-operator.v2.20.0
        version: 2.20.0
      name: beta
    - currentCSV: ibm-data-reporter-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "DataReporterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-data-reporter-operator",
                    "app.kubernetes.io/instance": "datareporterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datareporterconfig",
                    "app.kubernetes.io/part-of": "ibm-data-reporter-operator"
                  },
                  "name": "datareporterconfig"
                },
                "spec": {
                  "userConfig": [
                    {
                      "metadata": {
                        "ameta1": "ametadata1",
                        "bmeta1": "bmetadata1",
                        "cmeta1": "cmetadata1",
                        "dmeta1": "dmetadata1"
                      },
                      "userName": "system:serviceaccount:redhat-marketplace:ibm-data-reporter-operator-api"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
          createdAt: ""
          description: The IBM Data Reporter Operator accepts events and transforms
            them into reports submitted to the Data Service of the IBM Metrics Operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReporterConfig is the Schema for the datareporterconfigs
              API
            displayName: Data Reporter Config
            kind: DataReporterConfig
            name: datareporterconfigs.marketplace.redhat.com
            version: v1alpha1
          required:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
        description: "# IBM Data Reporter Operator\n\n# Introduction\n\nThe IBM Data
          Reporter Operator accepts events and transforms them into reports submitted
          to the IBM Metrics Operator Data Service.\n\n# Details\n\nThe IBM Data Reporter
          Operator deploys a service that provides an endpoint to which callers can
          send raw json event data. The event data is transformed into a report and
          is sent to the IBM Metrics Operator's Data Service. The IBM Metrics Operator
          Data Service periodically uploads the reports to IBM Software Central. It
          can also filter, transform and forward events to alternate endpoints.\n\n##
          Prerequisites\n\n- OpenShift Container Platform, major version 4 with any
          available supported minor version\n- Install IBM Metrics Operator\n  - IBM
          Data Reporter Operator prerequisties the IBM Metrics Operator's Data Service\n
          \ - Register the Cluster by creating a `redhat-marketplace-pull-secret`,
          as per the instructions\n  - Once `rhm-data-service` has started IBM Data
          Reporter Operator will accept requests\n\n## SecurityContextConstraints
          Requirements\n\n- The operator runs under Red Hat restricted SCC\n\n## Resources
          Required\n\n- The operator requires at least 85 Mi memory.\n\n## Limitations\n\n-
          Only runs on amd64, s390x and ppc64le architectures\n\n## Installing\n\n-
          A user with the cluster administrator role.\n- Install this operator in
          the same namespace as the IBM Metrics Operator\n  - default namespace: `ibm-software-central`\n\n##
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n# Configuration\n\nOptional:\n- Configure the
          Custom Resource Definition `DataReporterConfig` named `datareporterconfig`
          as per the following sample\n  - Configuration that can not be reconciled
          successfully will be reported on the DataReporterConfig Status\n  - jsonPath
          expressions are handled by: https://github.com/ohler55/ojg\n    - jsonPath
          comparison: https://cburgmer.github.io/json-path-comparison/\n  - Transformer
          type currently supported is `kazaam`\n    - https://github.com/willie68/kazaam\n\nSample
          DataReporterConfig:\n```YAML\napiVersion: marketplace.redhat.com/v1alpha1\nkind:
          DataReporterConfig\nmetadata:\n  name: datareporterconfig\nspec:\n  # true:
          report generated for 1 event and provide return code from data-service,
          confirming delivery\n  # false: events will be accumulated before building
          report, and immediately return 200, delivery unfconfirmed\n  confirmDelivery:
          false\n  dataFilters:   # dataFilters matches the first dataFilter selector,
          applies transforms and forwards to data-service and destinations\n  - altDestinations:
          \  # List of desintations to transform and foward events to\n    - authorization:
          \  # optional endpoint if an authorization token must first be requested
          \n        authDestHeader: Authorization   # http header key to be appended
          to destination request\n        authDestHeaderPrefix: 'Bearer '   # optional
          prefix for constructing destination request http header value\n        bodyData:
          \  # body data to send to authorization request, often an apiKey\n          secretKeyRef:\n
          \           name: auth-body-data-secret\n            key: bodydata\n        header:
          \  # secret map to use for authorization request headers\n          secret:\n
          \           name: auth-header-map-secret\n        tokenExpr: $.token   #
          optionally extract the returned body data using a jsonPath expression\n
          \       url: https://127.0.0.1:8443/api/2.0/accounts/my_account_id/apikeys/token
          \  # the url to make an authorization request to\n      header:   # secret
          map to use for destination request headers\n        secret:\n          name:
          dest-header-map-secret\n      transformer:   # The transformer to apply
          to event data for this destination\n        configMapKeyRef:\n          key:
          kazaam.json\n          name: kazaam-configmap\n        type: kazaam\n      url:
          https://127.0.0.1:8443/metering/api/1.0/usage   # the url to send the event
          to\n      urlSuffixExpr: $.properties.productId   # optional jsonPath expression
          on the event to extract and set a suffix to the url path\n    manifestType:
          dataReporter   # override manifestType of report sent to data-service\n
          \   selector:   # Matches against event data and a user. Empty selector
          matches all.\n      matchExpressions:   # jsonPath expression must return
          a result for all expressions to match (AND)\n      - $[?($.event == \"Account
          Contractual Usage\")]\n      - $.properties.productId\n      - $[?($.properties.source
          != null)]\n      - $[?($.properties.unit == \"AppPoints\")]\n      - $[?($.properties.quantity
          >= 0)]\n      - $[?($.timestamp != null)]\n      matchUsers:   # must match
          one of these users (OR). Omitting users matches any user\n      - system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n
          \   transformer:   # The transformer to apply to the event for data-service\n
          \     configMapKeyRef:\n        key: kazaam.json\n        name: kazaam-configmap\n
          \     type: kazaam\n  tlsConfig:   # TLS configuration for requests outbound
          to destinations\n    caCerts:\n    - name: tls-ca-certificates\n      key:
          ca.crt\n    certificates:\n    - clientCert:\n        secretKeyRef:\n          name:
          tls-ca-certificates\n          key: tls.crt\n      clientKey:\n        secretKeyRef:\n
          \         name: tls-ca-certificates\n          key: tls.key\n    cipherSuites:
          \n    - TLS_AES_128_GCM_SHA256\n    - TLS_AES_256_GCM_SHA384\n    - TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\n    - TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\n    insecureSkipVerify: false\n
          \   minVersion: VersionTLS12\n  userConfig:   # Optional metadata to apply
          to report sent to data-service for a specific user\n  - metadata:\n      ameta1:
          ametadata1\n      bmeta1: bmetadata1\n      cmeta1: cmetadata1\n      dmeta1:
          dmetadata1\n    userName: system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n```\n\nSample
          Header Secrets:\n```YAML\napiVersion: v1\nstringData:\n  accept: | \n    */*\nkind:
          Secret\nmetadata:\n  name: dest-header-map-secret\ntype: Opaque\n---\napiVersion:
          v1\nstringData:\n  accept: application/json\n  Content-Type: application/json\nkind:
          Secret\nmetadata:\n  name: auth-header-map-secret\ntype: Opaque\n```\n\nSample
          Authentication Body Secret:\n```YAML\napiVersion: v1\nstringData:\n  bodydata:
          |\n    {\"apikey\": \"myapikey\"}\nkind: Secret\nmetadata:\n  name: auth-body-data-secret\ntype:
          Opaque\n```\n\nSample TLS Secret:\n```YAML\napiVersion: v1\ndata:\n  ca.crt:
          LS0t...\n  tls.key: LS0t...\n  tls.crt: LS0t...\nkind: Secret\nmetadata:\n
          \ name: tls-ca-certificates\n```\n\nSample Kazaam ConfigMap:\n```YAML\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: kazaam-configmap\ndata:\n  kazaam.json:
          |\n    [\n        {\n            \"operation\": \"timestamp\",\n            \"spec\":
          {\n                \"timestamp\": {\n                \"inputFormat\": \"2006-01-02T15:04:05.999999+00:00\",\n
          \               \"outputFormat\": \"$unixmilli\"\n                }\n            }\n
          \       },\n        {\n            \"operation\": \"shift\",\n            \"spec\":
          {\n                \"instances[0].instanceId\": \"properties.source\",\n
          \               \"instances[0].startTime\": \"timestamp\",\n                \"instances[0].endTime\":
          \"timestamp\",\n                \"instances[0].metricUsage[0].metricId\":
          \"properties.unit\",\n                \"instances[0].metricUsage[0].quantity\":
          \"properties.quantity\"\n            }\n        },\n        {\n            \"operation\":
          \"default\",\n            \"spec\": {\n                \"meteringModel\":
          \"point-in-time\",\n                \"meteringPlan\": \"contract\"\n            }\n
          \       }\n    ]\n```\n\n### API Service User Configuration\n\n- The ClusterRole
          for api access is `clusterrole/ibm-data-reporter-operator-api`\n- The default
          ServiceAccount provided as an api user is `system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api`\n
          \ - The default ClusterRoleBinding for this user is `clusterrolebinding/ibm-data-reporter-operator-api`\n\nOptional:\n\n-
          To create an additional ServiceAccount and ClusterRoleBinding for api access\n\n```SHELL\nNAMESPACE=$(oc
          config view --minify -o jsonpath='{..namespace}') && \\\noc create serviceaccount
          my-api-service-account && \\\noc create clusterrolebinding ibm-data-reporter-operator-api
          --clusterrole=ibm-data-reporter-operator-api --serviceaccount=${NAMESPACE}:my-api-service-account\n```\n\n-
          Update datareporterconfig to attach metadata to reports associated with
          this user\n\n## API Service Usage\n\n- Get Token & Host\n\n```SHELL\noc
          project ibm-software-central\nexport DRTOKEN=$(oc create token ibm-data-reporter-operator-api
          --namespace ibm-software-central --duration 1h)\nexport DRHOST=$(oc get
          route ibm-data-reporter --template='{{ .spec.host }}')\n```\n\n- Get the
          Status\n\n```SHELL\ncurl -k -H \"Authorization: Bearer ${DRTOKEN}\" https://${DRHOST}/v1/status
          \n```\n\n- Post an Event\n\n```SHELL\ncurl -k -H \"Authorization: Bearer
          ${DRTOKEN}\" -X POST -d '{\"event\":\"myevent\"}' https://${DRHOST}/v1/event\n```\n\n##
          Storage\n\n- The operator temporarily stores events in pod memory, and writes
          to the IBM Metrics Operator data-service, which requires a PersistentVolume\n\n##
          License\n\nCopyright IBM Corporation 2023. All Rights Reserved.\n\nLicensed
          under the Apache License, Version 2.0 (the \"License\");\nyou may not use
          this file except in compliance with the License.\nYou may obtain a copy
          of the License at\n\n    http://www.apache.org/licenses/LICENSE-2.0\n\nUnless
          required by applicable law or agreed to in writing, software\ndistributed
          under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES
          OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for
          the specific language governing permissions and\nlimitations under the License.\n"
        displayName: IBM Data Reporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 2.20.0
      entries:
      - name: ibm-data-reporter-operator.v2.20.0
        version: 2.20.0
      name: stable
    defaultChannel: stable
    packageName: ibm-data-reporter-operator
    provider:
      name: Red Hat Marketplace
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-xtradb-cluster-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "docker.io/percona/percona-xtradb-cluster:8.0.42-33.1",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "docker.io/percona/haproxy:2.8.15",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "docker.io/percona/proxysql2:2.7.3",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "docker.io/percona/fluentbit:4.0.1",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:3.3.1",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "docker.io/percona/percona-xtrabackup:8.0.35-34.1",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.io/percona/percona-xtradb-cluster-operator:1.18.0
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <v1.18.0
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            clustercheck: Y2x1c3RlcmNoZWNrcGFzc3dvcmQ=
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            clustercheck: Y2x1c3RlcmNoZWNrcGFzc3dvcmQ=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ```
          ###  Release Highlights
          * **PMM3 support**

          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**

          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**

          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**

          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.

        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - database
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - docker.io/percona/percona-xtradb-cluster-operator:1.18.0
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator.v1.17.0
        version: 1.17.0
      - name: percona-xtradb-cluster-operator.v1.16.1
        version: 1.16.1
      - name: percona-xtradb-cluster-operator.v1.16.0
        version: 1.16.0
      - name: percona-xtradb-cluster-operator.v1.15.1
        version: 1.15.1
      - name: percona-xtradb-cluster-operator.v1.15.0
        version: 1.15.0
      - name: percona-xtradb-cluster-operator.v1.9.0
        version: 1.9.0
      name: stable
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: authorino-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: authorino-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta3",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/authorino-rhel9-operator@sha256:2d81613856df9725ad20928a7a9895d14482c42effce0a3a0de88f40ec35d411
          createdAt: "2025-03-25T22:39:23Z"
          description: Enables authentication and authorization for Gateways and applications
            in a Gateway API network.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta3
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: |-
          [Authorino](https://docs.kuadrant.io/authorino/) is Red Hat's Kubernetes-native lightweight external authorization service for tailor-made Zero Trust API security.

          Install this Red Hat official distribution of [Authorino Operator](https://docs.kuadrant.io/authorino-operator/) to manage instances of Authorino in this cluster.

          The current state of this distribution of the operator is: **Generally Available**.

          The Community version of Authorino Operator, based on upstream public images, is available in [OperatorHub.io](https://operatorhub.io/operator/authorino-operator).

          ## Getting started

          After installing the Operator, request an instance of the external authorization service by creating an `Authorino` custom resource.

          **Minimal example (namespaced)**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            listener:
              tls:
                enabled: false
            oidcServer:
              tls:
                enabled: false
          ```

          **Extended example**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            clusterWide: true
            authConfigLabelSelectors: environment=production
            secretLabelSelectors: authorino.kuadrant.io/component=authorino,environment=production

            replicas: 2

            evaluatorCacheSize: 2 # mb

            logLevel: info
            logMode: production

            listener:
              ports:
                grpc: 50001
                http: 5001
              tls:
                certSecretRef:
                  name: authorino-server-cert # secret must contain `tls.crt` and `tls.key` entries
              timeout: 2

            oidcServer:
              port: 8083
              tls:
                certSecretRef:
                  name: authorino-oidc-server-cert # secret must contain `tls.crt` and `tls.key` entries

            metrics:
              port: 8080
              deep: true

            healthz:
              port: 8081

            tracing:
              endpoint: rpc://otel-collector.observability.svc.cluster.local:4317
              insecure: true

            volumes:
              items:
                - name: keycloak-tls-cert
                  mountPath: /etc/ssl/certs
                  configMaps:
                    - keycloak-tls-cert
                  items: # details to mount the k8s configmap in the authorino pods
                    - key: keycloak.crt
                      path: keycloak.crt
              defaultMode: 420
          ```

          ### Cluster-wide vs Namespaced

          Namespaced instances only watch auth resources (`AuthConfig` and `Secrets`) created in the same namespace as the Authorino service. Use this mode for dedicated instances that do not require elevated privileges.

          Cluster-wide instances watch resources across the entire cluster (all namespaces.) Deploying and running Authorino in this mode requires elevated privileges.

          ### Multi-tenancy

          Use the `authConfigLabelSelectors` field of the `Authorino` custom resource to narrow the scope of the Authorino instance.

          Only `AuthConfig` custom resources whose labels match the value of this field will be handled by the Authorino instance.

          ## Protect a host

          To protect a host, create an `AuthConfig` custom resource for the host. E.g.:

          ```yaml
          apiVersion: authorino.kuadrant.io/v1beta3
          kind: AuthConfig
          metadata:
            name: my-api-protection
          spec:
            hosts:
            - my-api.io

            authentication:
              "keycloak":
                jwt:
                  issuerUrl: https://keycloak.keycloak.svc.cluster.local:8080/realms/my-realm

            authorization:
              "k8s-rbac":
                kubernetesSubjectAccessReview:
                  user:
                    selector: auth.identity.user.username
                  resourceAttributes:
                    resource:
                      value: my-api
                    verb:
                      selector: request.method
                cache:
                  key:
                    selector: auth.identity.user.username
                  ttl: 30
              "after-2am-only":
                rego: |
                  allow {
                    [hour, _, _] := time.clock(time.now_ns())
                    hour >= 2
                  }
          ```

          Make sure all requests to the host are fisrt checked with the Authorino instance, by configuring an Envoy proxy for external authz:

          ```yaml
          clusters:
          - name: my-api
            …
          - name: authorino
            connect_timeout: 0.25s
            type: STRICT_DNS
            lb_policy: ROUND_ROBIN
            http2_protocol_options: {}
            load_assignment:
              cluster_name: authorino
              endpoints:
              - lb_endpoints:
                - endpoint:
                    address:
                      socket_address:
                        address: authorino-authorino-authorization
                        port_value: 50051
          listeners:
          - filter_chains:
            - filters:
              name: envoy.http_connection_manager
              typed_config:
                "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
                use_remote_address: true
                route_config:
                  name: my-api-config
                  virtual_hosts:
                  - name: my-api-vs
                    domains:
                    - my-api.io
                    routes:
                    - match:
                        prefix: /
                      route:
                        cluster: my-api
                http_filters:
                - name: envoy.filters.http.ext_authz
                  typed_config:
                    "@type": type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz
                    transport_api_version: V3
                    failure_mode_allow: false
                    include_peer_certificate: true
                    grpc_service:
                      envoy_grpc:
                        cluster_name: authorino
                      timeout: 1s
          ```

          ...or, if using Istio, by creating an [`AuthorizationPolicy`](https://istio.io/latest/docs/reference/config/security/authorization-policy/#AuthorizationPolicy) custom resource. Use `action: CUSTOM` in the resource and the Authorino authorization service configured in the mesh extension provider settings.

          ## Features

          **Authentication**

          * JWT validation (with OpenID Connect Discovery)
          * OAuth 2.0 Token Introspection (opaque tokens)
          * Kubernetes TokenReview (ServiceAccount tokens)
          * API key authentication
          * X.509 client certificate authentication
          * Anonymous access
          * Proxy-handled (authentication performed by the proxy)

          **Authorization**

          * Built-in simple pattern matching (e.g. JWT claims, request attributes checking)
          * OPA policies (inline Rego and fetch from external registry)
          * Kubernetes SubjectAccessReview (resource and non-resource attributes)
          * Authzed SpiceDB

          **External metadata**

          * HTTP request
          * OpenID Connect User Info
          * UMA-protected resource attributes

          **Custom responses**

          * Header injection (Festival Wristbands tokens, JSON, plain text)
          * Envoy Dynamic Metadata
          * Custom HTTP response (status code, headers, messages, body, etc)

          **Callbacks**

          * HTTP webhooks

          **Caching**

          * OpenID Connect and User-Managed Access configs
          * JSON Web Keys (JWKs) and JSON Web Key Sets (JWKS)
          * Access tokens
          * External metadata
          * Precompiled Rego policies
          * Policy evaluation

          Check out the full [Feature Specification](https://docs.kuadrant.io/authorino/docs/features/) and how-to guides in the [Kuadrant Docs](https://docs.kuadrant.io) website.
        displayName: Red Hat - Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: GitHub
          url: https://github.com/kuadrant/authorino-operator
        - name: Documentation
          url: https://docs.kuadrant.io/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhcl-1/authorino-rhel9-operator@sha256:2d81613856df9725ad20928a7a9895d14482c42effce0a3a0de88f40ec35d411
        - registry.redhat.io/rhcl-1/authorino-rhel9@sha256:413d3796ad0101bf14d3341d31d9367e67fea1e73ebb3fb79124ef4c5dde0c09
        version: 1.2.2
      entries:
      - name: authorino-operator.v1.2.2
        version: 1.2.2
      - name: authorino-operator.v1.2.1
        version: 1.2.1
      - name: authorino-operator.v1.2.0
        version: 1.2.0
      - name: authorino-operator.v1.1.3
        version: 1.1.3
      - name: authorino-operator.v1.1.2
        version: 1.1.2
      - name: authorino-operator.v1.1.1
        version: 1.1.1
      - name: authorino-operator.v1.1.0
        version: 1.1.0
      - name: authorino-operator.v1.0.2
        version: 1.0.2
      - name: authorino-operator.v0.16.1
        version: 0.16.1
      - name: authorino-operator.v0.16.0
        version: 0.16.0
      name: stable
    - currentCSV: authorino-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta1",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "hosts": [
                    "my-api.io"
                  ],
                  "identity": [
                    {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "in": "authorization_header",
                        "keySelector": "APIKEY"
                      },
                      "name": "api-key-users"
                    }
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/3scale-tech-preview/authorino-rhel9-operator@sha256:eb574781cf39917f8f5d9e5c806f520534601f819ee1ebcbb751832da8938008
          createdAt: "2024-11-08T11:06:33Z"
          description: Authorino
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta1
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: |-
          [Authorino](https://docs.kuadrant.io/authorino/) is Red Hat's Kubernetes-native lightweight external authorization service for tailor-made Zero Trust API security.

          Install this Red Hat official distribution of [Authorino Operator](https://docs.kuadrant.io/authorino-operator/) to manage instances of Authorino in this cluster.

          The current state of this distribution of the operator is: **_Technical Preview_**.

          The Community version of Authorino Operator, based on upstream public images, is available in [OperatorHub.io](https://operatorhub.io/operator/authorino-operator).

          ## Getting started

          After installing the Operator, request an instance of the external authorization service by creating an `Authorino` custom resource.

          **Minimal example (namespaced)**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            listener:
              tls:
                enabled: false
            oidcServer:
              tls:
                enabled: false
          ```

          **Extended example**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            clusterWide: true
            authConfigLabelSelectors: environment=production
            secretLabelSelectors: authorino.kuadrant.io/component=authorino,environment=production

            replicas: 2

            evaluatorCacheSize: 2 # mb

            logLevel: info
            logMode: production

            listener:
              ports:
                grpc: 50001
                http: 5001
              tls:
                certSecretRef:
                  name: authorino-server-cert # secret must contain `tls.crt` and `tls.key` entries
              timeout: 2

            oidcServer:
              port: 8083
              tls:
                certSecretRef:
                  name: authorino-oidc-server-cert # secret must contain `tls.crt` and `tls.key` entries

            metrics:
              port: 8080
              deep: true

            healthz:
              port: 8081

            tracing:
              endpoint: rpc://otel-collector.observability.svc.cluster.local:4317
              insecure: true

            volumes:
              items:
                - name: keycloak-tls-cert
                  mountPath: /etc/ssl/certs
                  configMaps:
                    - keycloak-tls-cert
                  items: # details to mount the k8s configmap in the authorino pods
                    - key: keycloak.crt
                      path: keycloak.crt
              defaultMode: 420
          ```

          ### Cluster-wide vs Namespaced

          Namespaced instances only watch auth resources (`AuthConfig` and `Secrets`) created in the same namespace as the Authorino service. Use this mode for dedicated instances that do not require elevated privileges.

          Cluster-wide instances watch resources across the entire cluster (all namespaces.) Deploying and running Authorino in this mode requires elevated privileges.

          ### Multi-tenancy

          Use the `authConfigLabelSelectors` field of the `Authorino` custom resource to narrow the scope of the Authorino instance.

          Only `AuthConfig` custom resources whose labels match the value of this field will be handled by the Authorino instance.

          ## Protect a host

          To protect a host, create an `AuthConfig` custom resource for the host. E.g.:

          ```yaml
          apiVersion: authorino.kuadrant.io/v1beta2
          kind: AuthConfig
          metadata:
            name: my-api-protection
          spec:
            hosts:
            - my-api.io

            authentication:
              "keycloak":
                jwt:
                  issuerUrl: https://keycloak.keycloak.svc.cluster.local:8080/realms/my-realm

            authorization:
              "k8s-rbac":
                kubernetesSubjectAccessReview:
                  user:
                    selector: auth.identity.user.username
                  resourceAttributes:
                    resource:
                      value: my-api
                    verb:
                      selector: request.method
                cache:
                  key:
                    selector: auth.identity.user.username
                  ttl: 30
              "after-2am-only":
                rego: |
                  allow {
                    [hour, _, _] := time.clock(time.now_ns())
                    hour >= 2
                  }
          ```

          Make sure all requests to the host are fisrt checked with the Authorino instance, by configuring an Envoy proxy for external authz:

          ```yaml
          clusters:
          - name: my-api
            …
          - name: authorino
            connect_timeout: 0.25s
            type: STRICT_DNS
            lb_policy: ROUND_ROBIN
            http2_protocol_options: {}
            load_assignment:
              cluster_name: authorino
              endpoints:
              - lb_endpoints:
                - endpoint:
                    address:
                      socket_address:
                        address: authorino-authorino-authorization
                        port_value: 50051
          listeners:
          - filter_chains:
            - filters:
              name: envoy.http_connection_manager
              typed_config:
                "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
                use_remote_address: true
                route_config:
                  name: my-api-config
                  virtual_hosts:
                  - name: my-api-vs
                    domains:
                    - my-api.io
                    routes:
                    - match:
                        prefix: /
                      route:
                        cluster: my-api
                http_filters:
                - name: envoy.filters.http.ext_authz
                  typed_config:
                    "@type": type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz
                    transport_api_version: V3
                    failure_mode_allow: false
                    include_peer_certificate: true
                    grpc_service:
                      envoy_grpc:
                        cluster_name: authorino
                      timeout: 1s
          ```

          ...or, if using Istio, by creating an [`AuthorizationPolicy`](https://istio.io/latest/docs/reference/config/security/authorization-policy/#AuthorizationPolicy) custom resource. Use `action: CUSTOM` in the resource and the Authorino authorization service configured in the mesh extension provider settings.

          ## Features

          **Authentication**

          * JWT validation (with OpenID Connect Discovery)
          * OAuth 2.0 Token Introspection (opaque tokens)
          * Kubernetes TokenReview (ServiceAccount tokens)
          * API key authentication
          * X.509 client certificate authentication
          * Anonymous access
          * Proxy-handled (authentication performed by the proxy)

          **Authorization**

          * Built-in simple pattern matching (e.g. JWT claims, request attributes checking)
          * OPA policies (inline Rego and fetch from external registry)
          * Kubernetes SubjectAccessReview (resource and non-resource attributes)
          * Authzed SpiceDB

          **External metadata**

          * HTTP request
          * OpenID Connect User Info
          * UMA-protected resource attributes

          **Custom responses**

          * Header injection (Festival Wristbands tokens, JSON, plain text)
          * Envoy Dynamic Metadata
          * Custom HTTP response (status code, headers, messages, body, etc)

          **Callbacks**

          * HTTP webhooks

          **Caching**

          * OpenID Connect and User-Managed Access configs
          * JSON Web Keys (JWKs) and JSON Web Key Sets (JWKS)
          * Access tokens
          * External metadata
          * Precompiled Rego policies
          * Policy evaluation

          Check out the full [Feature Specification](https://docs.kuadrant.io/authorino/docs/features/) and how-to guides in the [Kuadrant Docs](https://docs.kuadrant.io) website.
        displayName: Red Hat - Authorino (Technical Preview)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: GitHub
          url: https://github.com/kuadrant/authorino-operator
        - name: Documentation
          url: https://docs.kuadrant.io/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-tech-preview/authorino-rhel9@sha256:a473dae20e71e3e813ac30ba978f2ab3c5e19d7d39b501ae9103dca892107c87
        - registry.redhat.io/3scale-tech-preview/authorino-rhel9-operator@sha256:eb574781cf39917f8f5d9e5c806f520534601f819ee1ebcbb751832da8938008
        - registry.redhat.io/openshift4/ose-cli-rhel9@sha256:3a398158d10018bfac5df291d6998b513cf9a30e5477bcaefddcf66d229627c7
        version: 1.1.3
      entries:
      - name: authorino-operator.v1.1.3
        version: 1.1.3
      - name: authorino-operator.v1.1.2
        version: 1.1.2
      - name: authorino-operator.v1.1.1
        version: 1.1.1
      - name: authorino-operator.v1.1.0
        version: 1.1.0
      - name: authorino-operator.v1.0.2
        version: 1.0.2
      name: tech-preview-v1
    defaultChannel: stable
    packageName: authorino-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-athena-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-athena-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "athena.services.k8s.aws/v1alpha1",
                "kind": "WorkGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/athena-controller:1.1.1
          createdAt: "2025-09-19T21:06:02Z"
          description: AWS Athena controller is a service controller for managing
            Athena resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PreparedStatement represents the state of an AWS athena PreparedStatement
              resource.
            displayName: PreparedStatement
            kind: PreparedStatement
            name: preparedstatements.athena.services.k8s.aws
            version: v1alpha1
          - description: WorkGroup represents the state of an AWS athena WorkGroup
              resource.
            displayName: WorkGroup
            kind: WorkGroup
            name: workgroups.athena.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Athena resources in AWS from within your Kubernetes cluster.

          **About Amazon Athena**

          Amazon Athena is an interactive query service that makes it easy to analyze data directly in Amazon Simple Storage Service (Amazon S3) using standard SQL. With a few actions in the AWS Management Console, you can point Athena at your data stored in Amazon S3 and begin using standard SQL to run ad-hoc queries and get results in seconds.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Athena
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - athena
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Athena Developer Resources
          url: https://aws.amazon.com/athena/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: athena maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/athena-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-athena-controller.v1.1.1
        version: 1.1.1
      - name: ack-athena-controller.v1.1.0
        version: 1.1.0
      - name: ack-athena-controller.v1.0.14
        version: 1.0.14
      - name: ack-athena-controller.v1.0.13
        version: 1.0.13
      - name: ack-athena-controller.v1.0.12
        version: 1.0.12
      - name: ack-athena-controller.v1.0.11
        version: 1.0.11
      - name: ack-athena-controller.v1.0.10
        version: 1.0.10
      - name: ack-athena-controller.v1.0.9
        version: 1.0.9
      - name: ack-athena-controller.v1.0.8
        version: 1.0.8
      - name: ack-athena-controller.v1.0.7
        version: 1.0.7
      - name: ack-athena-controller.v1.0.6
        version: 1.0.6
      - name: ack-athena-controller.v1.0.5
        version: 1.0.5
      - name: ack-athena-controller.v1.0.4
        version: 1.0.4
      - name: ack-athena-controller.v1.0.1
        version: 1.0.1
      - name: ack-athena-controller.v1.0.0
        version: 1.0.0
      - name: ack-athena-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-athena-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Flux project
      provider-url: ""
    name: flux
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.toolkit.fluxcd.io/v2beta1",
                "kind": "HelmRelease",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "5m",
                    "chart": {
                      "spec": {
                          "chart": "podinfo",
                          "version": "4.0.x",
                          "sourceRef": {
                            "kind": "HelmRepository",
                            "name": "podinfo",
                            "namespace": "flux-system"
                          },
                          "interval": "1m"
                      }
                    },
                    "values": {
                      "replicaCount": 2
                    }
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "GitRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "1m",
                    "ref": {
                      "branch": "master"
                    },
                    "url": "https://github.com/stefanprodan/podinfo"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmRepository",
                "metadata": {
                    "name": "stable",
                    "namespace": "default"
                },
                "spec": {
                    "url": "https://kubernetes-charts.storage.googleapis.com/",
                    "interval": "10m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmChart",
                "metadata": {
                    "name": "redis",
                    "namespace": "default"
                },
                "spec": {
                    "chart": "redis",
                    "version": "10.5.7",
                    "sourceRef": {
                      "name": "stable",
                      "kind": "HelmRepository"
                    },
                    "interval": "5m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "Bucket",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "1m",
                    "provider": "generic",
                    "bucketName": "podinfo",
                    "endpoint": "minio.minio.svc.cluster.local:9000",
                    "insecure": true,
                    "secretRef": {
                      "name": "minio-credentials"
                    }
                }
              },
              {
                "apiVersion": "kustomize.toolkit.fluxcd.io/v1beta2",
                "kind": "Kustomization",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "5m",
                    "path": "./deploy/overlays/dev",
                    "prune": true,
                    "sourceRef": {
                      "kind": "GitRepository",
                      "name": "podinfo"
                    },
                    "timeout": "2m",
                    "wait": true
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "address": "https://github.com/stefanprodan/podinfo",
                    "secretRef": {
                      "name": "api-token"
                    }
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Alert",
                "metadata": {
                    "name": "webapp",
                    "namespace": "default"
                },
                "spec": {
                    "providerRef": {
                      "name": "on-call-slack"
                    },
                    "eventSeverity": "info",
                    "eventSources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "Bucket",
                          "name": "secrets"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-backend"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-frontend"
                      }
                    ]
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Receiver",
                "metadata": {
                    "name": "github-receiver",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "events": [
                      "ping",
                      "push"
                    ],
                    "secretRef": {
                      "name": "webhook-token"
                    },
                    "resources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "HelmRepository",
                          "name": "webapp"
                      }
                    ]
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "image": "ghcr.io/stefanprodan/podinfo",
                    "interval": "1m0s"
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImagePolicy",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "imageRepositoryRef": {
                      "name": "podinfo"
                    },
                    "policy": {
                      "semver": {
                          "range": "5.0.x"
                      }
                    }
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageUpdateAutomation",
                "metadata": {
                  "name": "flux-system",
                  "namespace": "flux-system"
                },
                "spec": {
                  "interval": "1m0s",
                  "sourceRef": {
                    "kind": "GitRepository",
                    "name": "flux-system"
                  },
                  "git": {
                    "checkout": {
                      "ref": {
                        "branch": "main"
                      }
                    },
                    "commit": {
                      "author": {
                        "email": "fluxcdbot@users.noreply.github.com",
                        "name": "fluxcdbot"
                      },
                      "messageTemplate": "{{range .Updated.Images}}{{println .}}{{end}}"
                    },
                    "push": {
                      "branch": "main"
                    }
                  },
                  "update": {
                    "path": "./clusters/my-cluster",
                    "strategy": "Setters"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/fluxcd/source-controller:v1.3.0
          description: Flux is a Continuous Delivery solution for Kubernetes.
          operatorframework.io/suggested-namespace: flux-system
          support: The Flux community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta3
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta1
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta2
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta1
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta2
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta2
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta2
          - description: OCIRepository
            displayName: OCIRepository
            kind: OCIRepository
            name: ocirepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta3
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta2
        description: |-
          [Flux](https://fluxcd.io/) is a Continuous Delivery solution for Kubernetes.
          Flux is constructed with the [GitOps Toolkit](https://toolkit.fluxcd.io/components/), a set of composable APIs and specialized tools for keeping Kubernetes clusters in sync with sources of configuration (like Git & Helm repositories), and automating updates to configuration when there is new code to deploy.

          Flux is a CNCF incubation project made for:

          - **cluster operators** who automate provision and configuration of clusters
          - **platform engineers** who build continuous delivery for developer teams
          - **app developers** who rely on continuous delivery to get their code live

          Features:

          - Source configuration from Git and Helm repositories, and S3-compatible buckets (e.g., Minio)
          - Kustomize and Helm support
          - Event-triggered and periodic reconciliation
          - Integration with Kubernetes RBAC
          - Health assessment (clusters and workloads)
          - Dependency management (infrastructure and workloads)
          - Secrets management (builtin support for Mozilla SOPS and Cloud KMS)
          - Alerting to external systems (webhook senders)
          - External events handling (webhook receivers)
          - Automated container image updates to Git (image scanning and patching)
          - Policy-driven validation (OPA, Kyverno, admission controllers)
          - Seamless integration with Git providers (GitHub, GitLab, Bitbucket)
          - Interoperability with Cluster API (CAPI) providers

          ### Steps to support bootstraping with Flux CLI

          SecComp Profile should be removed as it requires a special privilege to get the access.
          Also, it requires to enforce `runUserAs` to the same UID provided by the images to prevent OpenShift
          to alter the user id expected by our controllers.

          To do so, please specifiy a patch using the following Kustomization file, for example.

          ```yaml
          apiVersion: kustomize.config.k8s.io/v1beta1
          kind: Kustomization
          resources:
            - gotk-components.yaml
            - gotk-sync.yaml
          patches:
            - patch: |
                apiVersion: apps/v1
                kind: Deployment
                metadata:
                  name: all
                spec:
                  template:
                    spec:
                      containers:
                        - name: manager
                          securityContext:
                            runAsUser: 65534
                            seccompProfile:
                              $patch: delete
              target:
                kind: Deployment
                labelSelector: app.kubernetes.io/part-of=flux
          ```

          Then before installing Flux, please also perform the following steps to bind Flux to the `nonroot` SecurityContextConstraint (SCC). This applies to installation via OC's OLM or via Flux CLI.

          ```sh
          # replace default "flux-system" with the project/namespace name for your flux installation
          namespace="flux-system"
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:source-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:kustomize-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:helm-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:notification-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-automation-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-reflector-controller

          ```

          To see complete steps please refer to this [documentation](https://github.com/fluxcd/website/blob/main/content/en/docs/use-cases/openshift.md).

          If you've already installed Flux, you might need to uninstall it after applying the above pre-requisite steps.

          ### Network Policies

          This version of Flux is shipped without the normally provided network Policies.
          By default all traffic is enabled, and anyone can access the source-controller
          artifacts.

          It is not recommended to run it in production as is.

          Please visit [Flux's policies](https://github.com/fluxcd/flux2/tree/main/manifests/policies)
          to deploy the needed network policies and secure your Flux setup.

          ### Breaking Changes

          - Flux 0.18.x ships with v1beta2 of the kustomize.toolkit.fluxcd.io/Kustomization CRD.
          - Flux 0.25.x is the last Flux release where Kubernetes 1.19 is supported.

          ### Flux on OpenShift Use Cases

          - [The GitOps with Flux + OpenShift](https://www.youtube.com/watch?v=8ybCtNVP5To) talk by Andrew Block (Red Hat) showing how to setup multi-tenacy on OpenShift using Flux.
          - [Flux OpenShift Use Case](https://fluxcd.io/docs/use-cases/openshift/) on FluxCD.io
        displayName: Flux
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - gitops
        - cd
        - continuous delivery
        - fluxcd
        - helm
        - kustomize
        links:
        - name: Homepage
          url: https://fluxcd.io/
        - name: Documentation
          url: https://fluxcd.io/docs/
        - name: Slack
          url: https://cloud-native.slack.com/messages/flux
        - name: Mailing List
          url: https://lists.cncf.io/g/cncf-flux-dev
        - name: Source Code
          url: https://github.com/fluxcd/flux2
        maintainers:
        - email: support@weave.works
          name: The Flux maintainers
        maturity: stable
        provider:
          name: The Flux project
        relatedImages:
        - ghcr.io/fluxcd/kustomize-controller:v1.3.0
        - ghcr.io/fluxcd/notification-controller:v1.3.0
        - ghcr.io/fluxcd/source-controller:v1.3.0
        - ghcr.io/fluxcd/helm-controller:v1.0.1
        - ghcr.io/fluxcd/image-automation-controller:v0.38.0
        - ghcr.io/fluxcd/image-reflector-controller:v0.32.0
        version: 2.3.0
      entries:
      - name: flux.v2.3.0
        version: 2.3.0
      - name: flux.v2.2.2
        version: 2.2.2
      - name: flux.v2.2.0
        version: 2.2.0
      - name: flux.v2.0.1
        version: 2.0.1
      - name: flux.v0.41.2
        version: 0.41.2
      - name: flux.v0.40.0
        version: 0.40.0
      - name: flux.v0.39.0
        version: 0.39.0
      - name: flux.v0.38.3
        version: 0.38.3
      - name: flux.v0.37.0
        version: 0.37.0
      - name: flux.v0.36.0
        version: 0.36.0
      - name: flux.v0.35.0
        version: 0.35.0
      - name: flux.v0.31.4
        version: 0.31.4
      - name: flux.v0.31.3
        version: 0.31.3
      - name: flux.v0.30.2
        version: 0.30.2
      - name: flux.v0.28.5
        version: 0.28.5
      - name: flux.v0.28.4
        version: 0.28.4
      - name: flux.v0.28.3
        version: 0.28.3
      - name: flux.v0.27.4
        version: 0.27.4
      - name: flux.v0.27.3
        version: 0.27.3
      - name: flux.v0.27.2
        version: 0.27.2
      - name: flux.v0.27.0
        version: 0.27.0
      - name: flux.v0.26.1
        version: 0.26.1
      - name: flux.v0.26.0
        version: 0.26.0
      - name: flux.v0.25.3
        version: 0.25.3
      - name: flux.v0.25.2
        version: 0.25.2
      - name: flux.v0.24.1
        version: 0.24.1
      - name: flux.v0.24.0
        version: 0.24.0
      - name: flux.v0.23.0
        version: 0.23.0
      - name: flux.v0.22.1
        version: 0.22.1
      - name: flux.v0.22.0
        version: 0.22.0
      - name: flux.v0.21.1
        version: 0.21.1
      - name: flux.v0.21.0
        version: 0.21.0
      - name: flux.v0.20.1
        version: 0.20.1
      - name: flux.v0.20.0
        version: 0.20.0
      - name: flux.v0.19.1
        version: 0.19.1
      - name: flux.v0.18.3
        version: 0.18.3
      - name: flux.v0.18.1
        version: 0.18.1
      - name: flux.v0.17.2
        version: 0.17.2
      - name: flux.v0.17.1
        version: 0.17.1
      - name: flux.v0.17.0
        version: 0.17.0
      - name: flux.v0.16.2
        version: 0.16.2
      - name: flux.v0.16.1
        version: 0.16.1
      - name: flux.v0.16.0
        version: 0.16.0
      - name: flux.v0.15.3
        version: 0.15.3
      - name: flux.v0.15.2
        version: 0.15.2
      - name: flux.v0.14.2
        version: 0.14.2
      - name: flux.v0.14.1
        version: 0.14.1
      - name: flux.v0.14.0
        version: 0.14.0
      - name: flux.v0.13.4
        version: 0.13.4
      name: stable
    defaultChannel: stable
    packageName: flux
    provider:
      name: The Flux project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          olm.skipRange: <1.0.4
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
        version: 1.0.4
      entries:
      - name: datadog-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      name: stable
    defaultChannel: stable
    packageName: datadog-operator-certified
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kiali-ossm
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kiali-operator.v2.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "Kiali",
                "metadata": {
                  "name": "kiali"
                },
                "spec": {
                  "version": "default",
                  "istio_namespace": "istio-system",
                  "deployment": {
                    "logger": {
                      "log_level": "info"
                    },
                    "view_only_mode": false
                  }
                }
              },
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "OSSMConsole",
                "metadata": {
                  "name": "ossmconsole"
                },
                "spec": {
                  "version": "default",
                  "kiali": {
                    "serviceName": "",
                    "serviceNamespace": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9-operator@sha256:0612dfa52f2bf39a1d87e640409313585fedc7cfcff6d00e77bb14809d27d646
          createdAt: "2024-12-12T10:44:15Z"
          description: This productized operator provides Kiali and OSSMC. Kiali is
            the Istio observability and management Console. OSSMC is the OpenShift
            Service Mesh Console plugin, powered by Kiali.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.1.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kiali/kiali
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A configuration file for a Kiali installation.
            displayName: Kiali
            kind: Kiali
            name: kialis.kiali.io
            version: v1alpha1
          - description: A configuration file for a OpenShift Service Mesh Console
              installation.
            displayName: OpenShift Service Mesh Console
            kind: OSSMConsole
            name: ossmconsoles.kiali.io
            version: v1alpha1
        description: |-
          ## About the managed application

          A Microservice Architecture breaks up the monolith into many smaller pieces
          that are composed together. Patterns to secure the communication between
          services like fault tolerance (via timeout, retry, circuit breaking, etc.)
          have come up as well as distributed tracing to be able to see where calls
          are going.

          A service mesh can now provide these services on a platform level and frees
          the application writers from those tasks. Routing decisions are done at the
          mesh level.

          Kiali works with OpenShift Service Mesh to visualize the service
          mesh topology, to provide visibility into features like circuit breakers,
          request rates and more. It offers insights about the mesh components at
          different levels, from abstract Applications to Services and Workloads.

          See [https://www.kiali.io](https://www.kiali.io) to read more.

          ### Accessing the UI

          By default, the Kiali operator exposes the Kiali UI as an OpenShift Route.

          If on OpenShift, you can create an OSSMConsole CR to have the operator
          install the OpenShift ServiceMesh Console plugin to the OpenShift Console
          thus providing an interface directly integrated with the OpenShift Console.

          ## About this Operator

          ### Kiali Custom Resource Configuration Settings

          For quick descriptions of all the settings you can configure in the Kiali
          Custom Resource (CR), see
          [the kiali.io docs](https://kiali.io/docs/configuration/kialis.kiali.io/).

          ## Prerequisites for enabling this Operator

          Kiali is a companion tool for OpenShift Service Mesh. So before you install Kiali, you must have
          already installed Service Mesh.
        displayName: Kiali Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - service-mesh
        - observability
        - monitoring
        - istio
        links:
        - name: Installation Guide
          url: https://www.kiali.io/docs/installation
        - name: Features
          url: https://www.kiali.io/docs/features
        - name: Documentation Home
          url: https://www.kiali.io
        - name: Blogs and Articles
          url: https://medium.com/kialiproject
        - name: Server and UI Source Code
          url: https://github.com/kiali/kiali
        - name: Operator Source Code
          url: https://github.com/kiali/kiali-operator
        - name: OpenShift Service Mesh Console Code
          url: https://github.com/kiali/openshift-servicemesh-plugin
        maintainers:
        - email: kiali-dev@googlegroups.com
          name: Kiali Developers Google Group
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel8@sha256:e37d76dd7c7a2db79ff1e96afef19b07c3beaa8ebf0c7bd85b808908b8f0ce77
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9-operator@sha256:0612dfa52f2bf39a1d87e640409313585fedc7cfcff6d00e77bb14809d27d646
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9@sha256:06fde6c696a33d17b0d2f7620ec46422c82277fd17a3fe6d7319de05c732aace
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:976a31ca8cf17ac0887b8a586edfcc29eb30175802461b3114758af525e7b800
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:66e4d933a9d69158bf87dd45835183bf2308b77c48908ab64496c5d22c816293
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:27a6aebffa3d049ee48050126e6107db2e07c49e9ce9297175facb6368e1e792
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-ossmc-rhel9@sha256:99643f294c3f92a6fc82f321cab7c756e4879fefb48fa824f1bf49ff7182658d
        version: 2.1.2
      entries:
      - name: kiali-operator.v2.1.2
        version: 2.1.2
      - name: kiali-operator.v1.89.8
        version: 1.89.8
      - name: kiali-operator.v1.89.7
        version: 1.89.7
      - name: kiali-operator.v1.89.6
        version: 1.89.6
      - name: kiali-operator.v1.89.1
        version: 1.89.1
      - name: kiali-operator.v1.73.10
        version: 1.73.10
      - name: kiali-operator.v1.73.8
        version: 1.73.8
      - name: kiali-operator.v1.73.7
        version: 1.73.7
      - name: kiali-operator.v1.73.4
        version: 1.73.4
      - name: kiali-operator.v1.65.11
        version: 1.65.11
      - name: kiali-operator.v1.65.10
        version: 1.65.10
      - name: kiali-operator.v1.65.9
        version: 1.65.9
      - name: kiali-operator.v1.65.8
        version: 1.65.8
      - name: kiali-operator.v1.65.7
        version: 1.65.7
      - name: kiali-operator.v1.65.6
        version: 1.65.6
      - name: kiali-operator.v1.57.7
        version: 1.57.7
      name: candidate
    - currentCSV: kiali-operator.v2.11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "Kiali",
                "metadata": {
                  "name": "kiali"
                },
                "spec": {
                  "version": "default",
                  "istio_namespace": "istio-system",
                  "deployment": {
                    "logger": {
                      "log_level": "info"
                    },
                    "view_only_mode": false
                  }
                }
              },
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "OSSMConsole",
                "metadata": {
                  "name": "ossmconsole"
                },
                "spec": {
                  "version": "default",
                  "kiali": {
                    "serviceName": "",
                    "serviceNamespace": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/kiali-rhel9-operator@sha256:2970861bd8e65e36d5080dc4d56340d08a52dbb1217be352998d135eb854732f
          createdAt: "2025-09-10T00:28:49Z"
          description: This productized operator provides Kiali and OSSMC. Kiali is
            the Istio observability and management Console. OSSMC is the OpenShift
            Service Mesh Console plugin, powered by Kiali.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kiali/kiali
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A configuration file for a Kiali installation.
            displayName: Kiali
            kind: Kiali
            name: kialis.kiali.io
            version: v1alpha1
          - description: A configuration file for a OpenShift Service Mesh Console
              installation.
            displayName: OpenShift Service Mesh Console
            kind: OSSMConsole
            name: ossmconsoles.kiali.io
            version: v1alpha1
        description: |-
          ## About the managed application

          A Microservice Architecture breaks up the monolith into many smaller pieces
          that are composed together. Patterns to secure the communication between
          services like fault tolerance (via timeout, retry, circuit breaking, etc.)
          have come up as well as distributed tracing to be able to see where calls
          are going.

          A service mesh can now provide these services on a platform level and frees
          the application writers from those tasks. Routing decisions are done at the
          mesh level.

          Kiali works with OpenShift Service Mesh to visualize the service
          mesh topology, to provide visibility into features like circuit breakers,
          request rates and more. It offers insights about the mesh components at
          different levels, from abstract Applications to Services and Workloads.

          See [https://www.kiali.io](https://www.kiali.io) to read more.

          ### Accessing the UI

          By default, the Kiali operator exposes the Kiali UI as an OpenShift Route.

          If on OpenShift, you can create an OSSMConsole CR to have the operator
          install the OpenShift ServiceMesh Console plugin to the OpenShift Console
          thus providing an interface directly integrated with the OpenShift Console.

          ## About this Operator

          ### Kiali Custom Resource Configuration Settings

          For quick descriptions of all the settings you can configure in the Kiali
          Custom Resource (CR), see
          [the kiali.io docs](https://kiali.io/docs/configuration/kialis.kiali.io/).

          ## Prerequisites for enabling this Operator

          Kiali is a companion tool for OpenShift Service Mesh. So before you install Kiali, you must have
          already installed Service Mesh.
        displayName: Kiali Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - service-mesh
        - observability
        - monitoring
        - istio
        links:
        - name: Installation Guide
          url: https://www.kiali.io/docs/installation
        - name: Features
          url: https://www.kiali.io/docs/features
        - name: Documentation Home
          url: https://www.kiali.io
        - name: Blogs and Articles
          url: https://medium.com/kialiproject
        - name: Server and UI Source Code
          url: https://github.com/kiali/kiali
        - name: Operator Source Code
          url: https://github.com/kiali/kiali-operator
        - name: OpenShift Service Mesh Console Code
          url: https://github.com/kiali/openshift-servicemesh-plugin
        maintainers:
        - email: kiali-dev@googlegroups.com
          name: Kiali Developers Google Group
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9-operator@sha256:2970861bd8e65e36d5080dc4d56340d08a52dbb1217be352998d135eb854732f
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:7487fbb4b776dfca28ad09311ba857e07ad98bc91ca5ee6fcaaa5b7783d04fbd
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel9@sha256:9a0d4862a25c2999a3f85b9a7a0332dab4ec1744548191d7a00c838b06ba26d2
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel9@sha256:fe25be563d50e5ad4e5475f153e03d4b24c7db094d7e824c51ebaae79418b1f7
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel8@sha256:81db94df331799f4087596378914b712d85d117dde3df2d4902079a82d4a9c55
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9@sha256:a00ec00e466e080592dca5c1cc153698b1412a54e1979647f74281f79b5691cf
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9@sha256:85eeb972e46f26ff30e04991e76ec3050d98b334e7b7f201a094f8d50b06854e
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:d729ef8d74c2f6e02d8c68ca761b1fb3c7c3706f1290b684468f0acb0ef313e0
        version: 2.11.3
      entries:
      - name: kiali-operator.v2.11.3
        version: 2.11.3
      - name: kiali-operator.v2.11.2
        version: 2.11.2
      - name: kiali-operator.v2.11.1
        version: 2.11.1
      - name: kiali-operator.v2.4.7
        version: 2.4.7
      - name: kiali-operator.v2.4.6
        version: 2.4.6
      - name: kiali-operator.v2.4.5
        version: 2.4.5
      - name: kiali-operator.v2.4.4
        version: 2.4.4
      - name: kiali-operator.v2.4.3
        version: 2.4.3
      - name: kiali-operator.v1.89.10
        version: 1.89.10
      - name: kiali-operator.v1.89.9
        version: 1.89.9
      - name: kiali-operator.v1.89.8
        version: 1.89.8
      - name: kiali-operator.v1.89.7
        version: 1.89.7
      - name: kiali-operator.v1.89.6
        version: 1.89.6
      - name: kiali-operator.v1.89.1
        version: 1.89.1
      - name: kiali-operator.v1.73.10
        version: 1.73.10
      - name: kiali-operator.v1.73.8
        version: 1.73.8
      - name: kiali-operator.v1.73.7
        version: 1.73.7
      - name: kiali-operator.v1.73.4
        version: 1.73.4
      - name: kiali-operator.v1.65.11
        version: 1.65.11
      - name: kiali-operator.v1.65.10
        version: 1.65.10
      - name: kiali-operator.v1.65.9
        version: 1.65.9
      - name: kiali-operator.v1.65.8
        version: 1.65.8
      - name: kiali-operator.v1.65.7
        version: 1.65.7
      - name: kiali-operator.v1.65.6
        version: 1.65.6
      - name: kiali-operator.v1.57.7
        version: 1.57.7
      name: stable
    defaultChannel: stable
    packageName: kiali-ossm
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: test-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: test-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "test.openstack.org/v1beta1",
                "kind": "HorizonTest",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "test-operator",
                    "app.kubernetes.io/instance": "horizontest-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "horizontest",
                    "app.kubernetes.io/part-of": "test-operator"
                  },
                  "name": "horizontest-sample"
                },
                "spec": {
                  "adminPassword": "12345678",
                  "adminUsername": "admin",
                  "authUrl": "https://keystone-public-openstack.apps.ocp.openstack.lab",
                  "backoffLimit": 0,
                  "containerImage": "",
                  "dashboardUrl": "https://horizon-openstack.apps.ocp.openstack.lab/",
                  "flavorName": "m1.tiny",
                  "horizonRepoBranch": "master",
                  "horizonTestDir": "/var/lib/horizontest",
                  "imageUrl": "http://download.cirros-cloud.net/0.6.2/cirros-0.6.2-x86_64-disk.img",
                  "logsDirectoryName": "horizon",
                  "parallel": false,
                  "password": "horizontest",
                  "projectName": "horizontest",
                  "repoUrl": "https://review.opendev.org/openstack/horizon",
                  "storageClass": "local-storage",
                  "user": "horizontest"
                }
              },
              {
                "apiVersion": "test.openstack.org/v1beta1",
                "kind": "Tempest",
                "metadata": {
                  "name": "tempest-tests",
                  "namespace": "openstack"
                },
                "spec": {
                  "containerImage": "",
                  "tempestRun": {
                    "concurrency": 8,
                    "includeList": "tempest.api.identity.v3.*\n"
                  },
                  "tempestconfRun": null
                }
              },
              {
                "apiVersion": "test.openstack.org/v1beta1",
                "kind": "Tobiko",
                "metadata": {
                  "name": "tobiko-tests",
                  "namespace": "openstack"
                },
                "spec": {
                  "config": "[DEFAULT]\nlog_file = value\n\n[testcase]\ntimeout = value\ntest_runner_timeout = value\n\n[ubuntu]\ninterface_name = value\n\n[keystone]\ninterface_name = value\n",
                  "containerImage": "",
                  "testenv": "py3",
                  "version": "master",
                  "workflow": [
                    {
                      "stepName": "unit-tests",
                      "testenv": "py3"
                    },
                    {
                      "stepName": "functional-tests",
                      "testenv": "functional"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:43Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: Ansible Test
            kind: AnsibleTest
            name: ansibletests.test.openstack.org
            version: v1beta1
          - displayName: Horizon Test
            kind: HorizonTest
            name: horizontests.test.openstack.org
            version: v1beta1
          - displayName: Tempest
            kind: Tempest
            name: tempests.test.openstack.org
            version: v1beta1
          - kind: Tobiko
            name: tobikoes.test.openstack.org
            version: v1beta1
        description: Test Operator
        displayName: OpenStack Test Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Tempest
        - Tobiko
        - Tests
        links:
        - name: Test Operator
          url: https://github.com/openstack-k8s-operators/test-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/test-rhel9-operator@sha256:17949ad6a69a14ea38870d471fd4f5a51122927f795976ac7e2b0189e2d0da5a
        - registry.redhat.io/rhoso/openstack-tempest-all-rhel9@sha256:2f07239d932a67a6589844f4380f613cc16614ff4de7699f64c17359bbdc5534
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: test-operator.v1.0.6
        version: 1.0.6
      - name: test-operator.v1.0.5
        version: 1.0.5
      - name: test-operator.v1.0.4
        version: 1.0.4
      - name: test-operator.v1.0.3
        version: 1.0.3
      - name: test-operator.v1.0.2
        version: 1.0.2
      - name: test-operator.v1.0.1
        version: 1.0.1
      - name: test-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: test-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhsso-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhsso-operator.7.5.1-opr-005
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm"
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "False"
          containerImage: registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:66577ec32a9a975f304e52b8ab49b9f7602158dcab1ef5d8c28a5d3455491336
          createdAt: "2021-08-09 00:00:00"
          description: An Operator for installing and managing Red Hat Single Sign-On
          olm.skipRange: <7.5.0
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Red Hat Single Sign-On Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Red Hat Single Sign-On.

          Red Hat Single Sign-On lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
        displayName: Red Hat Single Sign-On Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://access.redhat.com/products/red-hat-single-sign-on
        - name: Keycloak
          url: https://www.keycloak.org/
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat Customer Service
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso7-rhel8-init-container@sha256:614a9b1d7975769e8559e0b8c1578d16f6e3f7fee0778283797f4d62f5fc2100
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:5f0840a69642d13016430151892ec987fd67e711563b1dfb1a57353037c3bf79
        - registry.redhat.io/rh-sso-7/sso75-openj9-openshift-rhel8@sha256:2d4ea7226f43e7e8e8d3fcef6d425eb5104a15acf07791355033300b80db2901
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:66577ec32a9a975f304e52b8ab49b9f7602158dcab1ef5d8c28a5d3455491336
        version: 7.5.1-opr-005
      entries:
      - name: rhsso-operator.7.5.1-opr-005
        version: 7.5.1-opr-005
      - name: rhsso-operator.7.5.1-opr-004
        version: 7.5.1-opr-004
      - name: rhsso-operator.7.5.1-opr-003
        version: 7.5.1-opr-003
      - name: rhsso-operator.7.5.1-opr-002
        version: 7.5.1-opr-002
      - name: rhsso-operator.7.5.1-opr-001
        version: 7.5.1-opr-001
      - name: rhsso-operator.7.5.0
        version: 7.5.0
      name: alpha
    - currentCSV: rhsso-operator.7.6.12-opr-004
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm"
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "False"
          containerImage: registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:229265579ec77f2bd18c936ba19e1f9dbc6b2f0a13d963d5eae5b52c260010d1
          createdAt: "2021-08-09 00:00:00"
          description: An Operator for installing and managing Red Hat Single Sign-On
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Red Hat Single Sign-On Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Red Hat Single Sign-On.

          Red Hat Single Sign-On lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
        displayName: Red Hat Single Sign-On Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://access.redhat.com/products/red-hat-single-sign-on
        - name: Keycloak
          url: https://www.keycloak.org/
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat Customer Service
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:229265579ec77f2bd18c936ba19e1f9dbc6b2f0a13d963d5eae5b52c260010d1
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:923206112ed0873ee58edd0817dbce7bb4c9a89dc75d9b3ba2affe9aed4811de
        - registry.redhat.io/rh-sso-7/sso7-rhel8-init-container@sha256:282daf78934ca495ad4051371a58e3af0917bb9241d043a66e23fc8b12fbd1ce
        version: 7.6.12-opr-004
      entries:
      - name: rhsso-operator.7.6.12-opr-004
        version: 7.6.12-opr-004
      - name: rhsso-operator.7.6.12-opr-003
        version: 7.6.12-opr-003
      - name: rhsso-operator.7.6.12-opr-002
        version: 7.6.12-opr-002
      - name: rhsso-operator.7.6.12-opr-001
        version: 7.6.12-opr-001
      - name: rhsso-operator.7.6.11-opr-006
        version: 7.6.11-opr-006
      - name: rhsso-operator.7.6.11-opr-005
        version: 7.6.11-opr-005
      - name: rhsso-operator.7.6.11-opr-004
        version: 7.6.11-opr-004
      - name: rhsso-operator.7.6.11-opr-003
        version: 7.6.11-opr-003
      - name: rhsso-operator.7.6.11-opr-002
        version: 7.6.11-opr-002
      - name: rhsso-operator.7.6.11-opr-001
        version: 7.6.11-opr-001
      - name: rhsso-operator.7.6.10-opr-001
        version: 7.6.10-opr-001
      - name: rhsso-operator.7.6.9-opr-003
        version: 7.6.9-opr-003
      - name: rhsso-operator.7.6.9-opr-002
        version: 7.6.9-opr-002
      - name: rhsso-operator.7.6.9-opr-001
        version: 7.6.9-opr-001
      - name: rhsso-operator.7.6.8-opr-002
        version: 7.6.8-opr-002
      - name: rhsso-operator.7.6.8-opr-001
        version: 7.6.8-opr-001
      - name: rhsso-operator.7.6.7-opr-002
        version: 7.6.7-opr-002
      - name: rhsso-operator.7.6.7-opr-001
        version: 7.6.7-opr-001
      - name: rhsso-operator.7.6.6-opr-004
        version: 7.6.6-opr-004
      - name: rhsso-operator.7.6.6-opr-003
        version: 7.6.6-opr-003
      - name: rhsso-operator.7.6.6-opr-002
        version: 7.6.6-opr-002
      - name: rhsso-operator.7.6.6-opr-001
        version: 7.6.6-opr-001
      - name: rhsso-operator.7.6.5-opr-004
        version: 7.6.5-opr-004
      - name: rhsso-operator.7.6.5-opr-003
        version: 7.6.5-opr-003
      - name: rhsso-operator.7.6.5-opr-002
        version: 7.6.5-opr-002
      - name: rhsso-operator.7.6.5-opr-001
        version: 7.6.5-opr-001
      - name: rhsso-operator.7.6.4-opr-003
        version: 7.6.4-opr-003
      - name: rhsso-operator.7.6.4-opr-002
        version: 7.6.4-opr-002
      - name: rhsso-operator.7.6.4-opr-001
        version: 7.6.4-opr-001
      - name: rhsso-operator.7.6.3-opr-002
        version: 7.6.3-opr-002
      - name: rhsso-operator.7.6.3-opr-001
        version: 7.6.3-opr-001
      - name: rhsso-operator.7.6.2-opr-003
        version: 7.6.2-opr-003
      - name: rhsso-operator.7.6.2-opr-002
        version: 7.6.2-opr-002
      - name: rhsso-operator.7.6.2-opr-001
        version: 7.6.2-opr-001
      - name: rhsso-operator.7.6.1-opr-005
        version: 7.6.1-opr-005
      - name: rhsso-operator.7.6.1-opr-004
        version: 7.6.1-opr-004
      - name: rhsso-operator.7.6.1-opr-003
        version: 7.6.1-opr-003
      - name: rhsso-operator.7.6.1-opr-002
        version: 7.6.1-opr-002
      - name: rhsso-operator.7.6.1-opr-001
        version: 7.6.1-opr-001
      - name: rhsso-operator.7.6.0-opr-005
        version: 7.6.0-opr-005
      - name: rhsso-operator.7.6.0-opr-004
        version: 7.6.0-opr-004
      - name: rhsso-operator.7.6.0-opr-003
        version: 7.6.0-opr-003
      - name: rhsso-operator.7.6.0-opr-002
        version: 7.6.0-opr-002
      - name: rhsso-operator.7.6.0-opr-001
        version: 7.6.0-opr-001
      - name: rhsso-operator.7.5.2-opr-003
        version: 7.5.2-opr-003
      - name: rhsso-operator.7.5.2-opr-002
        version: 7.5.2-opr-002
      - name: rhsso-operator.7.5.2-opr-001
        version: 7.5.2-opr-001
      - name: rhsso-operator.7.5.1-opr-011
        version: 7.5.1-opr-011
      - name: rhsso-operator.7.5.1-opr-010
        version: 7.5.1-opr-010
      - name: rhsso-operator.7.5.1-opr-009
        version: 7.5.1-opr-009
      - name: rhsso-operator.7.5.1-opr-009-0.1646836600.p
        version: 7.5.1-opr-009+0.1646836600.p
      - name: rhsso-operator.7.5.1-opr-008
        version: 7.5.1-opr-008
      - name: rhsso-operator.7.5.1-opr-007
        version: 7.5.1-opr-007
      - name: rhsso-operator.7.5.1-opr-006
        version: 7.5.1-opr-006
      - name: rhsso-operator.7.5.1-opr-005
        version: 7.5.1-opr-005
      - name: rhsso-operator.7.5.1-opr-004
        version: 7.5.1-opr-004
      - name: rhsso-operator.7.5.1-opr-003
        version: 7.5.1-opr-003
      - name: rhsso-operator.7.5.1-opr-002
        version: 7.5.1-opr-002
      - name: rhsso-operator.7.5.1-opr-001
        version: 7.5.1-opr-001
      - name: rhsso-operator.7.5.0
        version: 7.5.0
      name: stable
    defaultChannel: stable
    packageName: rhsso-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-glue-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-glue-controller.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/glue-controller:0.2.1
          createdAt: "2025-09-19T20:55:49Z"
          description: AWS Glue controller is a service controller for managing Glue
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Job represents the state of an AWS glue Job resource.
            displayName: Job
            kind: Job
            name: jobs.glue.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Glue resources in AWS from within your Kubernetes cluster.

          **About Amazon Glue**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Glue
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - glue
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Glue Developer Resources
          url: https://aws.amazon.com/Glue/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: glue maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/glue-controller:0.2.1
        version: 0.2.1
      entries:
      - name: ack-glue-controller.v0.2.1
        version: 0.2.1
      - name: ack-glue-controller.v0.2.0
        version: 0.2.0
      - name: ack-glue-controller.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ack-glue-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Protocol Labs
      provider-url: ""
    name: ipfs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ipfs-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.ipfs.io/v1alpha1",
                "kind": "CircuitRelay",
                "metadata": {
                  "name": "circuitrelay-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cluster.ipfs.io/v1alpha1",
                "kind": "IpfsCluster",
                "metadata": {
                  "name": "ipfscluster-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Storage, Networking, Big Data
          containerImage: quay.io/redhat-et-ipfs/ipfs-operator:v0.0.1
          createdAt: 2022-04-05T15:50:37-0400
          operatorframework.io/suggested-namespace: ipfs-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ipfs-cluster/ipfs-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CircuitRelay is the Schema for the circuitrelays API.
            displayName: Circuit Relay
            kind: CircuitRelay
            name: circuitrelays.cluster.ipfs.io
            version: v1alpha1
          - description: IpfsCluster is the Schema for the ipfs API.
            displayName: Ipfs Cluster
            kind: IpfsCluster
            name: ipfsclusters.cluster.ipfs.io
            version: v1alpha1
        description: "The IPFS Operator enables users to easily create IPFS Cluster
          instances\nwithin their Kubernetes or OpenShift clusters. \n\n### Supported
          Features\n* Horizontal scaling\n* Joining collaborative clusters\n* Controlling
          IPFS node resource usage\n* Specifying reprovider settings  \n\n### Documentation
          \nDocumentation can be found on our [read-the-docs page](https://ipfs-operator.readthedocs.io/en/latest/),
          as well as our [GitHub repository](https://github.com/ipfs-cluster/ipfs-operator).\n\n###
          Getting Help\nIn case you run into issues while using the operator, please
          create an issue in our [GitHub\nrepository](https://github.com/ipfs-cluster/ipfs-operator).\n\n###
          Contributing\nWe welcome contributions of all types and sizes! You can find
          open issues\nin our [GitHub Repository](https://github.com/ipfs-cluster/ipfs-operator/issues),\nas
          well as make your own issue if you feel that something is missing.\n\n###
          License\nIPFS Operator is licensed under the [Apache 2.0 License](https://github.com/ipfs-cluster/ipfs-operator/blob/main/LICENSE)"
        displayName: IPFS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipfs
        - ipfs cluster
        - decentralized storage
        - peer-to-peer
        - distributed hash table
        - content-addressed
        links:
        - name: IPFS Operator GitHub page
          url: https://github.com/ipfs-cluster/ipfs-operator
        - name: IPFS Cluster
          url: https://ipfscluster.io/
        - name: IPFS
          url: https://ipfs.io/
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Protocol Labs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/redhat-et-ipfs/ipfs-operator:v0.0.1
        version: 0.0.2
      entries:
      - name: ipfs-operator.v0.0.2
        version: 0.0.2
      - name: ipfs-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ipfs-operator
    provider:
      name: Protocol Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: gaudi-ai-sw-tools-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gaudi-ai-sw-tools-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aitools.intel/v1",
                "kind": "GaudiAIToolsContainer",
                "metadata": {
                  "name": "intel-gaudi-1.17.0-495"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagespec": {
                    "registry": "registry.connect.redhat.com",
                    "repo": "intel/gaudi-notebooks",
                    "tags": [
                      {
                        "gaudi_software": "1.17.0-495",
                        "rhel_os": "9.2"
                      }
                    ]
                  },
                  "nameOverride": "",
                  "namespace": "redhat-ods-applications"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: registry.connect.redhat.com/intel/gaudi-ai-sw-tools-operator@sha256:0c286f1e4cec57a189d749501f82345afbe0a1dbbbf92b906ad3c549e47b89cd
          createdAt: "2024-09-11T21:27:02Z"
          description: Operator to enable Gaudi AI SW Tools into OpenShift AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: GaudiAIToolsContainer
            name: gaudiaitoolscontainers.aitools.intel
            version: v1
        description: Operator to enable Gaudi AI SW Tools into OpenShift AI
        displayName: Intel® Gaudi AI SW Tools Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gaudi
        - opendatahub
        - openshiftai
        - notebooks
        links:
        - name: Intel® Gaudi AI SW Tools Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/66d7aa630eb66a02febc8103
        maintainers:
        - email: sharvil.shah@intel.com
          name: sharvil
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.connect.redhat.com/intel/gaudi-ai-sw-tools-operator@sha256:0c286f1e4cec57a189d749501f82345afbe0a1dbbbf92b906ad3c549e47b89cd
        version: 0.0.2
      entries:
      - name: gaudi-ai-sw-tools-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: gaudi-ai-sw-tools-operator
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-certified-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "psp": false,
                  "serviceAccount": "default",
                  "controller": {
                    "enabled": true,
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": "registry.connect.redhat.com/neuvector/controller",
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": ""
                      }
                    },
                    "pvc": {
                      "enabled": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "path": "/",
                          "annotations": {
                            "ingress.kubernetes.io/protocol": "https"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "path": "/",
                          "annotations": {
                            "ingress.kubernetes.io/protocol": "https"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "path": "/",
                      "annotations": {
                        "ingress.kubernetes.io/protocol": "https"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {}
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/enforcer",
                    "priorityClassName": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      }
                    ],
                    "resources": {}
                  },
                  "manager": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/manager",
                    "priorityClassName": "",
                    "env": {
                      "ssl": true
                    },
                    "svc": {
                      "type": "NodePort",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "path": "/",
                      "annotations": {},
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "affinity": {},
                    "tolerations": [],
                    "nodeSelector": {}
                  },
                  "cve": {
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "image": "registry.access.redhat.com/ubi8@sha256:228824aa581f3b31bf79411f8448b798291c667a37155bdea61cfa128b2833f2",
                      "schedule": "0 0 * * *",
                      "priorityClassName": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": "registry.connect.redhat.com/neuvector/scanner@sha256:42004e13b348bb43b58ae8538c983a70b237e3aca870758ac629fb4fc43e8130",
                      "priorityClassName": "",
                      "resources": {},
                      "affinity": {},
                      "tolerations": [],
                      "nodeSelector": {}
                    }
                  },
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "resources": {},
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "true"
          containerImage: registry.connect.redhat.com/neuvector/neuvector-operator@sha256:6155ac3c95aa6471c9857e0a6f5e5d19e6fc1d8cc7d014108d7d4a10d14f96bb
          createdAt: "2020-12-02 14:19:21"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/neuvector-certified-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/neuvector-certified-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: Neuvector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm/tree/master/charts/core) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart.  NeuVector Operator versions are tied to NeuVector product versions. Version 1.3.4 of the NeuVector Certified Operator deploys version 4.4.3 of NeuVector.

          **Complete below steps to Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Grant Service Account Access to the Privileged SCC

                   oc -n neuvector adm policy add-scc-to-user privileged -z default

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s                    neuvector/default


          **Add NeuVector license from NeuVector WebUI->setting**


          #Deploying the NeuVector Operator#


          Please refer to the instructions [here](https://github.com/neuvector/neuvector-operator/blob/master/README.md)


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Networking
        - Security
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - registry.connect.redhat.com/neuvector/scanner@sha256:42004e13b348bb43b58ae8538c983a70b237e3aca870758ac629fb4fc43e8130
        - registry.access.redhat.com/ubi8@sha256:228824aa581f3b31bf79411f8448b798291c667a37155bdea61cfa128b2833f2
        - registry.connect.redhat.com/neuvector/neuvector-operator@sha256:6155ac3c95aa6471c9857e0a6f5e5d19e6fc1d8cc7d014108d7d4a10d14f96bb
        - registry.connect.redhat.com/neuvector/controller@sha256:acc943568520282061547b310c50080dc32924a531bbb543d959b281eb6664f4
        - registry.connect.redhat.com/neuvector/enforcer@sha256:be1d299ec3532d55444681a882e9fd1238949da6598987308e2800ed26291b7a
        - registry.connect.redhat.com/neuvector/manager@sha256:8245cd179d11fff279690c5a422f46b4ff252ed72b9b998f0c9b5453bc7d8fc0
        version: 1.3.4
      entries:
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-certified-operator-rhmp
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dpu-network-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dpu-network-operator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dpu.openshift.io/v1alpha1",
                "kind": "OVNKubeConfig",
                "metadata": {
                  "name": "ovnkubeconfig-sample"
                },
                "spec": {
                  "kubeConfigFile": "tenant-cluster-1-kubeconf",
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/dpu-worker": ""
                    }
                  },
                  "poolName": "dpu"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional, Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/dpu-network-rhel8-operator@sha256:d07b8ecb5d1625599878b5c1e34b0de94188b81d90c829787a2d13969f6d5971
          createdAt: "2023-07-13T18:03:11Z"
          description: The operator is responsible for the life-cycle management of
            the ovn-kube components and the necessary host network initialization
            on DPU cards.
          olm.skipRange: '>=4.10.0-0 <4.13.0-202309181427'
          operatorframework.io/suggested-namespace: openshift-dpu-network-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/dpu-network-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OVNKubeConfig is the Schema for the ovnkubeconfigs API
            displayName: OVNKube Config
            kind: OVNKubeConfig
            name: ovnkubeconfigs.dpu.openshift.io
            version: v1alpha1
        description: The operator to be responsible for the life-cycle management
          of the ovn-kube components and the necessary host network initialization
          on DPU cards.
        displayName: DPU Network Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dpu
        - network
        links:
        - name: Dpu Network Operator
          url: https://dpu-network-operator.domain
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:767682dd3bd65651a8204c9fb732b9c48b99127189992b43abc6ecce027f4589
        - registry.redhat.io/openshift4/dpu-network-rhel8-operator@sha256:d07b8ecb5d1625599878b5c1e34b0de94188b81d90c829787a2d13969f6d5971
        version: 4.13.0-202309181427
      entries:
      - name: dpu-network-operator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: stable
    defaultChannel: stable
    packageName: dpu-network-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-odlm
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operand-deployment-lifecycle-manager.v4.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandBindInfo",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "bindings": {
                    "public": {
                      "configmap": "mongodb-configmap",
                      "secret": "mongodb-secret"
                    }
                  },
                  "description": "Binding information that should be accessible to MongoDB adopters",
                  "operand": "mongodb-atlas-kubernetes",
                  "registry": "example-service"
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "services": [
                    {
                      "name": "jaeger",
                      "spec": {
                        "jaeger": {
                          "strategy": "allinone"
                        }
                      }
                    },
                    {
                      "name": "mongodb-atlas-kubernetes",
                      "spec": {
                        "atlasDeployment": {
                          "deploymentSpec": {
                            "name": "test-deployment"
                          },
                          "projectRef": {
                            "name": "my-project"
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "operators": [
                    {
                      "channel": "stable",
                      "installMode": "cluster",
                      "name": "jaeger",
                      "namespace": "default",
                      "packageName": "jaeger",
                      "sourceName": "community-operators",
                      "sourceNamespace": "openshift-marketplace"
                    },
                    {
                      "channel": "stable",
                      "name": "mongodb-atlas-kubernetes",
                      "namespace": "default",
                      "packageName": "mongodb-atlas-kubernetes",
                      "sourceName": "community-operators",
                      "sourceNamespace": "openshift-marketplace"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandRequest",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "requests": [
                    {
                      "operands": [
                        {
                          "name": "jaeger"
                        },
                        {
                          "name": "mongodb-atlas-kubernetes"
                        }
                      ],
                      "registry": "example-service"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, Monitoring, Logging & Tracing, Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
          createdAt: "2025-02-14T21:16:49Z"
          description: The Operand Deployment Lifecycle Manager provides a Kubernetes
            CRD-based API to manage the lifecycle of operands.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-odlm
          olm.skipRange: '>=1.2.0 <4.5.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/operand-deployment-lifecycle-manager
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OperandBindInfo is the Schema for the operandbindinfoes API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandBindInfo
            kind: OperandBindInfo
            name: operandbindinfos.operator.ibm.com
            version: v1alpha1
          - description: OperandConfig is the Schema for the operandconfigs API. Documentation
              For additional details regarding install parameters check https://ibm.biz/icpfs39install.
              License By installing this product you accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandConfig
            kind: OperandConfig
            name: operandconfigs.operator.ibm.com
            version: v1alpha1
          - description: OperandRegistry is the Schema for the operandregistries API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandRegistry
            kind: OperandRegistry
            name: operandregistries.operator.ibm.com
            version: v1alpha1
          - description: OperandRequest is the Schema for the operandrequests API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandRequest
            kind: OperandRequest
            name: operandrequests.operator.ibm.com
            version: v1alpha1
          - description: OperatorConfig is the Schema for the operatorconfigs API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.ibm.com
            version: v1alpha1
        description: |-
          # Introduction

          The operand is the instance managed by the operator. ODLM is used to manage the lifecycle of for a group of operands, compared with operator lifecycle manager, ODLM focus on the management of operands but not operators.

          # Details

          ### What are operands?

          Operator is a method of packaging, deploying and managing a Kubernetes application.
          Operands are the services and applications that Operator manage.

          For example, cert-manager operator deploys a cert-manager deployment, then the cert-manager-operator is an operator and the cert-manager deployment is an operand.

          The ODLM will have four CRDs:

          | Resource                 | Short Name | Description                                                                                |
          |--------------------------|------------|--------------------------------------------------------------------------------------------|
          | OperandRequest | opreq | It defines which operator/operand want to be installed in the cluster |
          | OperandRegistry | opreg | It defines the OLM information, like channel and catalog source, for each operator |
          | OperandConfig | opcon | It defines the parameters that should be used to install the operator's operand |
          | OperandBindInfo | opbi | It identifies secrets and/or configmaps that should be shared with requests |

          ### Goal

          1. A single entry point to manage a group of operands
          1. User can select a set of operands to install
          1. The install can be invoked through either OCP UI or CLI

          ## Prerequisites
          - [operator-sdk] version v1.3.0.
          - [go] version 1.15.7+
          - [oc] version v3.11+ or [kubectl] v1.11.3+
          - Access to an Openshift v4.3+ cluster

          ### Resources Required

          #### Minimum scheduling capacity

          | Software                   | Memory (MB) | CPU (cores) |
          | -------------------------- | ----------- | ----------- |
          | Operand Deployment Lifecycle Manager | 200          | 0.2        |
          | **Total**                  | 200         | 0.2         |

          ## Installing

          ### Create Operator NS, Group, Subscription

          ```yaml
          apiVersion: v1
          kind: Namespace
          metadata:
            name: odlm

          ---
          apiVersion: operators.coreos.com/v1alpha2
          kind: OperatorGroup
          metadata:
            name: operatorgroup
            namespace: odlm
          spec:
            targetNamespaces:
            - odlm

          ---
          apiVersion: v1
          data:
            namespaces: odlm
          kind: ConfigMap
          metadata:
            name: namespace-scope
            namespace: odlm

          ---
          apiVersion: operators.coreos.com/v1alpha1
          kind: Subscription
          metadata:
            name: operand-deployment-lifecycle-manager
            namespace: odlm
          spec:
            channel: v3.20
            name: ibm-odlm
            source: opencloud-operators
            sourceNamespace: olm
            config:
              env:
              - name: INSTALL_SCOPE
                value: namespaced
          END
          ```

          ## Configuration

          OperandRegistry defines the OLM information used for installation, like package name and catalog source, for each operator.

          Following is an example of the OperandRegistry CR:

          **NOTE:** The "spec.operators[*].name" parameter must be unique for each entry.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRegistry
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            operators:
            - name: jenkins [3]
              namespace: default [4]
              channel: alpha [5]
              packageName: jenkins-operator [6]
              scope: public [7]
              sourceName: community-operators [8]
              sourceNamespace: openshift-marketplace [9]
              installMode: cluster [10]
              installPlanApproval: Manual [11]
          ```

          The OperandRegistry Custom Resource (CR) lists OLM Operator information for operands that may be requested for installation and/or access by an application that runs in a namespace. The registry CR specifies:

          1. `name` of the OperandRegistry
          2. `namespace` of the OperandRegistry
          3. `name` is the name of the operator, which should be the same as the services name in the OperandConfig and OperandRequest.
          4. `namespace` defines the namespace where the operator and its CR will be deployed. (1) When InstallMode is `cluster`, the operator will be deployed into the `openshift-operators` namespace and the operator CRs will be deployed into the namespace this parameter defines. (2) When InstallMode is empty or set to `namespace`, it is the namespace where both operator and operator CR will be deployed.
          5. `channel` is the name of OLM channel that is subscribed for the operator.
          6. `packageName` is the name of the package in CatalogSource that is subscribed for the operator.
          7. (optional) `scope` is an indicator, either public or private, that dictates whether deployment can be requested from other namespaces (public) or only from the namespace of this OperandRegistry (private). The default value is private.
          8. `sourceName` is the name of the CatalogSource.
          9. `sourceNamespace` is the namespace of the CatalogSource.
          10. (optional) `installMode` is the install mode of the operator, can be either `namespace` (OLM one namespace) or `cluster` (OLM all namespaces). The default value is `namespace`. Operator is deployed in `openshift-operators` namespace when InstallMode is set to `cluster`.
          11. (optional) `installPlanApproval` is the approval mode for emitted installplan. The default value is `Automatic`.

          ### OperandConfig Spec

          OperandConfig defines the individual operand configuration. The OperandConfig Custom Resource (CR) defines the parameters for each operator that is listed in the OperandRegistry that should be used to install the operator instance by specifying an installation CR.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          Kind: OperandConfigs
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            services:
            - name: jenkins [3]
              spec: [4]
                jenkins:
                  port: 8081
          ```

          OperandConfig defines the individual operand deployment config:

          1. `name` of the OperandConfig
          2. `namespace` of the OperandConfig
          3. `name` is the name of the operator, which should be the same as the services name in the OperandRegistry and OperandRequest.
          4. `spec` defines a map. Its key is the kind name of the custom resource. Its value is merged to the spec field of custom resource. For more details, you can check the following topic **How does ODLM create the individual operator CR?**

          #### How does Operator create the individual operator CR

          Jenkins Operator has one CRD: Jenkins:

          The OperandConfig CR has

          ```yaml
          - name: jenkins
            spec:
              jenkins:
                service:
                  port: 8081
          ```

          The Jenkins Operator CSV has

          ```yaml
          apiVersion: operators.coreos.com/v1alpha1
          kind: ClusterServiceVersion
          metadata:
            annotations:
            alm-examples: |-
              [
              {
                "apiVersion":"jenkins.io/v1alpha2",
                "kind":"Jenkins",
                "metadata": {
                  "name":"example"
                },
                "spec":{
                  ...
                  "service":{"port":8080,"type":"ClusterIP"},
                  ...
                }
              }
            ]
          ```

          The ODLM will deep merge the OperandConfig CR spec and Jenkins Operator CSV alm-examples to create the Jenkins CR.

          ```yaml
          apiVersion: jenkins.io/v1alpha2
          kind: Jenkins
          metadata:
            name: example
          spec:
            ...
            service:
              port: 8081
              type: ClusterIP
            ...
          ```

          For day2 operations, the ODLM will patch the OperandConfigs CR spec to the existing Jenkins CR.

          ### OperandRequest Spec

          OperandRequest defines which operator/operand you want to install in the cluster.

          **NOTE:** OperandRequest custom resource is used to trigger a deployment for Operators and Operands.

          #### OperandRequest sample to create custom resource via OperandConfig

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRequest
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            requests:
            - registry: example-service [3]
              registryNamespace: example-service-ns [4]
              operands: [5]
              - name: jenkins [6]
                bindings: [7]
                  public:
                    secret: jenkins-operator-credential [8]
                    configmap: jenkins-operator-base-configuration [9]
          ```

          1. `name` of the OperandRequest
          2. `namespace` of the OperandRequest
          3. `registry` identifies the name of the OperandRegistry CR from which this operand deployment is being requested.
          4. (optional) `registryNamespace` identifies the namespace in which the OperandRegistry CR is defined. **Note:** If the `registryNamespace` is not specified then it is assumed that the OperandRegistry CR is in the current (OperandRequest's) namespace.
          5. `operands` in the CR is a list of operands.
          6. `name` of operands in the CR must match a name specification in an OperandRegistry's CR.
          7. (optional) The `bindings` of the operands is a map to get and rename the secret and/or configmap from the provider and create them in the requester's namespace. If the requester wants to rename the secret and/or configmap, they need to know the key of the binding in the OperandBindInfo. If the key of the bindings map is prefixed with public, it means the secret and/or configmap can be shared with the requester in the other namespace. If the key of the bindings map is prefixed with private, it means the secret and/or configmap can only be shared within its own namespace.
          8. (optional) `secret` names a secret that should be created in the requester's namespace with formatted data that can be used to interact with the service.
          9. (optional) `configmap` names a configmap that should be created in the requester's namespace with formatted data that can be used to interact with the service.

          #### OperandRequest sample to create custom resource via OperandRequest

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRequest
          metadata:
            name: example-service
            namespace: example-service-ns
          spec:
            requests:
            - registry: example-service
              registryNamespace: example-service-ns
              operands:
              - name: jenkins
                kind: Jenkins [1]
                apiVersion: "jenkins.io/v1alpha2" [2]
                instanceName: "example" [3]
                spec: [4]
                  service:
                    port: 8081
          ```

          1. `kind` of the target custom resource. If it is set in the operand item, ODLM will create CR according to OperandRequest only and get rid of the alm-example and OperandConfig.
          2. `apiVersion` of the target custom resource.
          3. `instanceName` is the name of the custom resource. If `instanceName` is not set, the name of the custom resource will be created with the name of the OperandRequest as a prefix.
          4. `spec` is the spec field of the target CR.

          ### OperandBindInfo Spec

          The ODLM will use the OperandBindInfo to copy the generated secret and/or configmap to a requester's namespace when a service is requested with the OperandRequest CR. An example specification for an OperandBindInfo CR is shown below.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandBindInfo
          metadata:
            name: publicjenkinsbinding [1]
            namespace: example-service-ns [2]
          spec:
            operand: jenkins [3]
            registry: example-service [4]
            description: "Binding information that should be accessible to jenkins adopters" [5]
            bindings: [6]
              public:
                secret: jenkins-operator-credentials-example [7]
                configmap: jenkins-operator-base-configuration-example [8]
          ```

          Fields in this CR are described below.

          1. `name` of the OperandBindInfo
          2. `namespace` of the OperandBindInfo
          3. The `operand` should be the the individual operator name.
          4. The `registry` section must match the name in the OperandRegistry CR in the current namespace.
          5. `description` is used to add a detailed description of a service.
          6. The `bindings` section is used to specify information about the access/configuration data that is to be shared. If the key of the bindings map is prefixed with public, it means the secret and/or configmap can be shared with the requester in the other namespace. If the key of the bindings map is prefixed with private, it means the secret and/or configmap can only be shared within its own namespace. If the key of the bindings map is prefixed with protected, it means the secret and/or configmap can only be shared if it is explicitly declared in the OperandRequest.
          7. The `secret` field names an existing secret, if any, that has been created and holds information that is to be shared with the requester.
          8. The `configmap` field identifies a configmap object, if any, that should be shared with the requester

          ODLM will use the OperandBindInfo CR to pass information to an adopter when they create a OperandRequest to access the service, assuming that both have compatible scopes. ODLM will copy the information from the shared service's "OperandBindInfo.bindinfo[].secret" and/or "OperandBindInfo.bindinfo[].configmap" to the requester namespace.

          **NOTE:** If in the OperandRequest, there is no secret and/or configmap name specified in the bindings or no bindings field in the element of operands, ODLM will copy the secret and/or configmap to the requester's namespace and rename them to the name of the OperandBindInfo + secret/configmap name.

          ### E2E Use Case

          1. User installs ODLM from OLM

              The ODLM will automatically generate two default CRD CRs, since OperandRegistry and OperandConfigs CRs don't define the state, so it should be fine.

          2. **Optional:** Users can update the OperandConfigs CR with their own parameter values

              Users can skip this step if the default configurations match the requirement.

          3. User creates the OperandRequest CR from OLM

              This tells ODLM that users want to install some of the individual operator/operand.

          4. The rest works will be done by OLM and ODLM

              Finally, users will get what they want.

          ### Operator/Operand Upgrade

          - For operator/operand upgrade, you only need to publish your operator OLM to your operator channel, and OLM will handle the upgrade automatically.
          - If there are major version, then you may want to update `channel` in `OperandRegistry` to trigger upgrade.

          ### Limitations

          ODLM depends on NamespaceScope Operator

          ### SecurityContextConstraints Requirements

          The Operand Deployment Lifecycle Manager supports running under the OpenShift Container Platform default restricted security context constraints.
        displayName: Operand Deployment Lifecycle Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Operand Deployment Lifecycle Manager Project
          url: https://github.com/IBM/operand-deployment-lifecycle-manager
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bb074bdf89cbc7c703fadb17639608599ab4f66db9be5617e5f79a0550676790
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d4dff1a8968c18d984e16c4c9f1517c3a10e23a11bf166c64d734773324d4c9d
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
        - registry.marketplace.redhat.com/cpopen/operand-deployment-lifecycle-manager-bundle@sha256:68f71f2767a40ca9333f0a670e07493e8a8c13a5c7ca877ea5bd8a544b352d9a
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:27195e9758afdf8521971866fd60ea3dafed91a146c3c4165876e5ea476a4622
        version: 4.5.0
      entries:
      - name: operand-deployment-lifecycle-manager.v4.5.0
        version: 4.5.0
      name: v4.5
    defaultChannel: v4.5
    packageName: ibm-odlm
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fujitsu
      provider-url: https://www.postgresql.fastware.com/
    name: fujitsu-enterprise-postgres-operator-with-cg-module
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPAction",
                "metadata": {
                  "name": "new-fep-action"
                },
                "spec": {
                  "fepAction": {
                    "args": [
                      "new-fep-sts-0",
                      "new-fep-sts-1"
                    ],
                    "type": "reload"
                  },
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "targetClusterName": "new-fep"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPExporter",
                "metadata": {
                  "name": "new-fep-exporter"
                },
                "spec": {
                  "fepExporter": {
                    "exporterLogLevel": "error",
                    "fepClusterList": [
                      "new-fep1"
                    ],
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "userCustomQueries": "usr_example:\n  query: \"SELECT EXTRACT(EPOCH FROM (now() - pg_last_xact_replay_timestamp())) as lag\"\n  master: true\n  metrics:\n    - lag:\n        usage: \"GAUGE\"\n        description: \"Replication lag behind master in seconds\""
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPLogging",
                "metadata": {
                  "name": "new-fep-logging"
                },
                "spec": {
                  "fepLogging": {
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPPgpool2",
                "metadata": {
                  "name": "new-fep-pgpool2"
                },
                "spec": {
                  "clientAuthMethod": "scram",
                  "count": 2,
                  "customhba": "local   all         all                               trust\nhost    all         all         127.0.0.1/32          trust\nhost    all         all         ::1/128               trust\n",
                  "customlogsize": "128Mi",
                  "customparams": "listen_addresses = '*'\npcp_listen_addresses = '*'\nnum_init_children = 32\nreserved_connections = 0\nenable_pool_hba = off\nallow_clear_text_frontend_auth = off\nauthentication_timeout = 80\nbackend_weight0 = 1\nbackend_weight1 = 1\nbackend_flag0 = 'ALWAYS_PRIMARY'\nbackend_flag1 = 'DISALLOW_TO_FAILOVER'\nconnection_cache = on\nmax_pool = 4\nlisten_backlog_multiplier = 2\nserialize_accept = off\nchild_life_time = 300\nclient_idle_limit = 0\nchild_max_connections = 0\nconnection_life_time = 0\nreset_query_list = 'ABORT; DISCARD ALL'\nclient_min_messages = info\nlog_min_messages = debug1\nlog_statement = on\nlog_per_node_statement = on\nlog_client_messages = on\nlog_hostname = on\nlog_connections = on\nlog_line_prefix = '%t: pid %p: '\nload_balance_mode = on\nignore_leading_white_space = on\nwhite_function_list = ''\nblack_function_list = 'currval,lastval,nextval,setval'\nblack_query_pattern_list = ''\ndatabase_redirect_preference_list = ''\napp_name_redirect_preference_list = ''\nallow_sql_comments = off\ndisable_load_balance_on_write = 'transaction'\nstatement_level_load_balance = on\nsr_check_period = 0\nsr_check_user = 'postgres'\ndelay_threshold = 0\nlog_standby_delay = 'none'\nssl = on\nssl_ciphers = 'HIGH:MEDIUM:+3DES:!aNULL'\nssl_prefer_server_ciphers = off\nssl_ecdh_curve = 'prime256v1'\nssl_dh_params_file = ''\nrelcache_expire = 0\nrelcache_size = 256\ncheck_temp_table = catalog\ncheck_unlogged_table = on\nenable_shared_relcache = off\nrelcache_query_target = primary\nwd_port0 = 9000\nfailover_on_backend_error = off\n",
                  "custompcp": "none",
                  "customsslcacert": "none",
                  "customsslcert": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                  "customsslkey": "none",
                  "fepclustername": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "limits": {
                    "cpu": "400m",
                    "memory": "512Mi"
                  },
                  "requests": {
                    "cpu": "200m",
                    "memory": "256Mi"
                  },
                  "serviceport": 9999,
                  "statusport": 9898
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPRestore",
                "metadata": {
                  "name": "new-fep-restore"
                },
                "spec": {
                  "fromFEPcluster": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "mcSpec": {
                    "limits": {
                      "cpu": "200m",
                      "memory": "300Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "restoretype": "latest",
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "toFEPcluster": "new-fep-2"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v2",
                "kind": "FEPCluster",
                "metadata": {
                  "name": "new-fep"
                },
                "spec": {
                  "fep": {
                    "customAnnotations": {
                      "allDeployments": {}
                    },
                    "forceSsl": true,
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "instances": 1,
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "podAntiAffinity": false,
                    "podDisruptionBudget": false,
                    "servicePort": 27500,
                    "syncMode": "off",
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "usePodName": true
                  },
                  "fepChildCrVal": {
                    "backup": {
                      "image": {
                        "pullPolicy": "IfNotPresent"
                      },
                      "mcSpec": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      },
                      "pgbackrestParams": "# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[global]\nrepo1-retention-full=7\nrepo1-retention-full-type=time\nlog-path=/database/log/backup\n",
                      "postScript": " ",
                      "preScript": " ",
                      "schedule": {
                        "num": 2
                      },
                      "schedule1": {
                        "schedule": "15 0 * * 0",
                        "type": "full"
                      },
                      "schedule2": {
                        "schedule": "15 0 * * 1-6",
                        "type": "incr"
                      },
                      "schedule3": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule4": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule5": {
                        "schedule": " ",
                        "type": " "
                      }
                    },
                    "customPgAudit": "# define pg audit custom params here to override defaults.\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[output]\nlogger = 'auditlog'\nlog_directory = '/database/log/audit'\nlog_truncate_on_rotation = on\nlog_filename = 'pgaudit-%a.log'\nlog_rotation_age = 1d\nlog_rotation_size = 0\n[rule]\n",
                    "customPgHba": "# define pg_hba custom rules here to be merged with default rules.\n# TYPE     DATABASE        USER        ADDRESS        METHOD\n",
                    "customPgParams": "# define custom postgresql.conf parameters below to override defaults.\n# Current values are as per default FEP deployment\n# If you add a library to shared_preload_libraries, add it after the default value.\n# fsep_operator_security is available in FEP15 and later container images.\nshared_preload_libraries='pgx_datamasking,pg_prewarm,pg_stat_statements,fsep_operator_security'\nsession_preload_libraries='pg_prewarm'\nmax_prepared_transactions = 100\nmax_worker_processes = 30\nmax_connections = 100\nwork_mem = 1MB\nmaintenance_work_mem = 12MB\nshared_buffers = 128MB\neffective_cache_size = 384MB\ncheckpoint_completion_target = 0.8\n# tcp parameters\ntcp_keepalives_idle = 30\ntcp_keepalives_interval = 10\ntcp_keepalives_count = 3\n# logging parameters in default fep installation\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\nlog_directory = '/database/log'\nlog_filename = 'logfile-%a.log'\nlog_file_mode = 0600\nlog_truncate_on_rotation = on\nlog_rotation_age = 1d\nlog_rotation_size = 0\nlog_checkpoints = on\nlog_line_prefix = '%e %t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h'\nlog_lock_waits = on\nlog_autovacuum_min_duration = 60s\nlogging_collector = on\npgaudit.config_file='/opt/app-root/src/pgaudit-cfg/pgaudit.conf'\nlog_replication_commands = on\nlog_min_messages = WARNING\nlog_destination = csvlog\n# wal_archive parameters in default fep installation\narchive_mode = on\narchive_command = 'pgbackrest --stanza=backupstanza --config=/database/userdata/pgbackrest.conf archive-push %p'\nwal_level = replica\nmax_wal_senders = 12\nwal_keep_segments = 64\ntrack_activities = on\ntrack_counts = on\npassword_encryption = 'scram-sha-256'\n",
                    "storage": {
                      "archivewalVol": {
                        "size": "1Gi"
                      },
                      "backupVol": {
                        "size": "2Gi"
                      },
                      "dataVol": {
                        "size": "2Gi"
                      },
                      "logVol": {
                        "size": "1Gi"
                      },
                      "tablespaceVol": {
                        "size": "512Mi"
                      },
                      "walVol": {
                        "size": "1200Mi"
                      }
                    },
                    "sysUsers": {
                      "pgRewindPassword": "rewind_password",
                      "pgRewindUser": "rewind_user",
                      "pgdb": "mydb",
                      "pgpassword": "mydbpassword",
                      "pgreplpassword": "repluserpwd",
                      "pgrepluser": "repluser",
                      "pguser": "mydbuser",
                      "tdepassphrase": "tde-passphrase"
                    },
                    "systemCertificates": {
                      "cacrt": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                      "crt": "-----BEGIN CERTIFICATE-----\nMIIDUTCCAjmgAwIBAgIRAMocW3qMoHrD6qRvMPppMkMwDQYJKoZIhvcNAQELBQAw\nNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9yIEt1\nYmVybmV0ZXMwHhcNMjEwMjA2MDQzMjM2WhcNMjYwMjA1MDQzMjM2WjA/MRAwDgYD\nVQQKEwdGdWppdHN1MSswKQYDVQQDEyJGVUpJVFNVIEVudGVycHJpc2UgUG9zdGdy\nZXMgU2VydmVyMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA4AI33yvH\nZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I2e4SceTKi6O3C/I1XuvWlpng\n5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4ANB5JWWqDOjrRT3o7nRPGXfila\nbP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYpmjdbfxabTz07ig0+6/cwKoRR\nxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTFYvmAH7gcdssSFBt8NPlUATHE\nsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6Wdgmu5H2pDml8CDNLDv98Aj7i\n+I5SRKKcVPlnuQIDAQABo1AwTjAdBgNVHSUEFjAUBggrBgEFBQcDAQYIKwYBBQUH\nAwIwDAYDVR0TAQH/BAIwADAfBgNVHSMEGDAWgBQcwrrUO0u+FhIUuVdrDRCQRsi6\nZjANBgkqhkiG9w0BAQsFAAOCAQEAm5dxBoI9pScOCvRAchg4CprdRDSJb9K6yB3O\nnCAxnM47iHeXnY3WlnI388kHu8DU7O4ba1tJbGs3KY9KzioPk43pU12jWkO1onoF\n+mTDjx/Ef1cYWA9r5q/LtgTa6Q2sxV4O2x67QW82aAnaxO34dV5zWCPIvAoovZBV\nHRT+BgCg3r2vD1RGKK2nl1aYJtWhO1SZubam+VttdZ/vbM9oOJctxmImsEtBXjkY\nKteePdQtLL5o03JhyXWyRshCq+HMmKf2KgyY8gvydGcP4eLQdBWcW40LcnVq6UjT\n0kJycJEKngMVademq1ZWHGaiYB7hyT6GhgIcHUJ2cKrPgbEh1Q==\n-----END CERTIFICATE-----",
                      "key": "-----BEGIN RSA PRIVATE KEY-----\nMIIEowIBAAKCAQEA4AI33yvHZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I\n2e4SceTKi6O3C/I1XuvWlpng5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4AN\nB5JWWqDOjrRT3o7nRPGXfilabP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYp\nmjdbfxabTz07ig0+6/cwKoRRxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTF\nYvmAH7gcdssSFBt8NPlUATHEsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6W\ndgmu5H2pDml8CDNLDv98Aj7i+I5SRKKcVPlnuQIDAQABAoIBAFPQYKlOzw/+BA0b\nyMIUpdctIMb/54CR/xR0mVw1DbSjigNVPjHUQvB8Y1B2FAITQObgJO06bAv0QdWN\nRb0/v/yYiNJDFjaLjaIAHlO/2+oWrXbFaZqgpVDJhB+e1xaZr2x7XGxm+p925k30\nl6pvIRY+I8JRKvZiV1VZHwL/R3JOtPr++xMZtLVjVOI+f+ySqJ+TZHuAjm49EKxj\ncEmmJ28b7QcziXsvKy00f+zbqLIBKXQdZAFU5eEr1BsDRXdRW+Kf0XIvftuy4BJZ\nvoKT+VGhEvF/qysswL4+6IAO6tpuYnnM0Y2d3sOGoWPkTcQK0MekYKzL/WmtCjNs\n9hodJtECgYEA5EWyhEOf4uOKe5TDp697UCUvXLoOR58FDe/S8XNvScn29jjOkqIg\nOMoqo9xAkJTNTzqn5UUdt1x/pgM2NxlPLFijrc0zQlX3SoOO2ryDd9WNi7YKtN16\nKJqa536WeZu2OEbuAZ+S3GALVy1RPeTNPnUOmKnF06DjDUGzLNCZy10CgYEA+zfw\n952DWuz1U0Z4wvAEqqcgUKXPKrkTXV/iUnjkDkrLYVr0ZofDNTXrdHl+UedFmaOC\ncieZn6DNhcdz5tKtyysGMH3g/qs9PfoGUngvcXsy0Egk04l3x1jc8TTCLqXZXYaQ\nHMsx51n+R58oncPtzYSUOr9qQ6PbC2CstTbFJA0CgYEAjGEsUliAB/jknfEzjXjG\nPdhQUxb8VyE864Az2lah9t/kJzFyIAziAeqZ5GE7t247AGFTBRTHHI8e1Qoemi3P\nWbc9GVIbFs1lIYbcIDpUIyrKPEP8O5QEXtoNLxXTFgAjRGKiVY87spjCAJ+W2ZhO\ne/1it5GYXfgQCYQA2yuBmOUCgYANRkR2YR1axaCk+NlSu6oTdmdPu6M5x7PNQE7O\nOtMaKjua9lppvIzFGAdMDUtueoEEAE7ZR1xnwfB6PDLUpJdIYAqgr1YfPt8qkjaZ\nTv56yZ7CwL0pbF8m6nwqRrZoDp1wwraEvvvxFKFKGY/k3kCHlpTakdjEoDjn3gDi\nRnWeVQKBgCEneMSzucei5LRppRtRaJw/Btll8qlPMlX3W7dxQ3cLwpmLOn0m51Fp\nPIZ44zYK8R6fu4+/sSrlfaIg86Ugeufp6YNxyNROKxUGza5vDIu5OftwWtBeg+UK\nZ8lLWNdX6pp7WMujmF3H1DrkBbauYMUKZ4UxUYtelgHERMePIxwb\n-----END RSA PRIVATE KEY-----"
                    }
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "false"
          containerImage: quay.io/fujitsu/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:eaf601e7aef525ebcd4b769c859ed74c7a8077f843b3a75d76713eabf0295745
          createdAt: "2025-09-11T09:32:52Z"
          description: OpenShift Operator for Fujitsu Enterprise Postgres
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.1.0 <7.1.6'
          operators.openshift.io/valid-subscription: '["Fujitsu Enterprise Postgres
            Advanced Edition with Cryptographic Module Annual Subscription License
            per Core 17 Operator for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/internal-objects: '["fepautoscales.fep.fujitsu.io","fepconfigs.fep.fujitsu.io","fepusers.fep.fujitsu.io","fepcerts.fep.fujitsu.io","fepvolumes.fep.fujitsu.io","fepbackups.fep.fujitsu.io"]'
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: fj-dbaas-tec@dl.jp.fujitsu.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: FEPAction
            name: fepactions.fep.fujitsu.io
            version: v1
          - kind: FEPAutoscale
            name: fepautoscales.fep.fujitsu.io
            version: v1
          - kind: FEPBackup
            name: fepbackups.fep.fujitsu.io
            version: v1
          - kind: FEPCert
            name: fepcerts.fep.fujitsu.io
            version: v1
          - kind: FEPCluster
            name: fepclusters.fep.fujitsu.io
            version: v2
          - kind: FEPConfig
            name: fepconfigs.fep.fujitsu.io
            version: v1
          - kind: FEPExporter
            name: fepexporters.fep.fujitsu.io
            version: v1
          - kind: FEPLogging
            name: feploggings.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2Cert
            name: feppgpool2certs.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2
            name: feppgpool2s.fep.fujitsu.io
            version: v1
          - kind: FEPRestore
            name: feprestores.fep.fujitsu.io
            version: v1
          - kind: FEPUpgrade
            name: fepupgrades.fep.fujitsu.io
            version: v1
          - kind: FEPUser
            name: fepusers.fep.fujitsu.io
            version: v1
          - kind: FEPVolume
            name: fepvolumes.fep.fujitsu.io
            version: v1
        description: "Fujitsu Enterprise Postgres 17 delivers an enterprise-grade
          PostgreSQL on OpenShift Container Platform.\n\nThis solution provides the
          flexibility of a hybrid cloud solution while delivering an enhanced distribution
          of PostgreSQL to support enterprise-level workloads and provide improved
          deployment and management, availability, performance, data governance and
          security.  \n\nAvailable as a multi-architecture container built for both
          amd64.\n\nThe download and Use of the Product is strictly subject to the
          terms of the End User License Agreement with Fujitsu Limited found at https://www.fast.fujitsu.com/fujitsu-enterprise-postgres-license-agreements.
          Where the Product that has been embedded as a whole or part into a third
          party program, only Authorised Customers may download and use the Product.\n"
        displayName: Fujitsu Enterprise Postgres Operator with Cryptographic Module
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - Fujitsu
        - fep
        links:
        - name: Fujitsu Enterprise Postgres
          url: https://www.postgresql.fastware.com/
        maintainers:
        - email: fj-dbaas-tec@dl.jp.fujitsu.com
          name: Fujitsu
        maturity: stable
        provider:
          name: Fujitsu
          url: https://www.postgresql.fastware.com/
        relatedImages:
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-backup-with-cryptographic-module@sha256:dd15cf3b6b4a5c959efef8b602abe5d64b4f5a424fb6a3e55079a2040e67cd7a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-utils-with-cryptographic-module@sha256:8b2b6a981c863f7075ced1235c614317b6526626b42405295a9c664736f0a80f
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-restore-with-cryptographic-module@sha256:6c0d6e10cfe1868be2986fdde14c6f55f86b12cea01a82199b1e2e0c3154ef75
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-restore-with-cryptographic-module@sha256:98a3b0bfa6ec7a9a4a08fb524c331c3de6a9b580d6386a7b5430955ed84b2fc5
        - quay.io/fujitsu/fujitsu-enterprise-postgres-exporter-with-cryptographic-module@sha256:71e1d1e715e2d89757e80d90b8d5b11baa729110cd5f7898df920a4c85f2bf8a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-pgpool2-with-cryptographic-module@sha256:a4c5a59ca18f6fa295b78d570f03991f1a4b4bd6df52f8526a1254ac1a567990
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-upgrade-with-cryptographic-module@sha256:1fe690a9c04a55213062ad3bef8c0cff73849f8777c3945b9a8ae2dcff6330fa
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-backup-with-cryptographic-module@sha256:03b7f3c76e3eb775e4ec7f66f7506ee1feeab56dca96e868ac3f38b9c1ca5567
        - quay.io/fujitsu/fujitsu-enterprise-postgres-cronjob-with-cryptographic-module@sha256:809a64600396378ff8c825595d40f72aedbc2b57f394420548262b085a3f2ea5
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentbit-with-cryptographic-module@sha256:bb661b3aad0b825d56a9696d6d2d9f5c04add2a83e793fc3d4a73980f4a7a0ac
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-server-with-cryptographic-module@sha256:cad4e36dc2b1e5ca6bf9308d9f93958d4c8f2bbdc14e000ca0b5700fa335df44
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-utils-with-cryptographic-module@sha256:4caf1dc5211ac03893ce9a62b417f412ffd411f13569f3ac75e12ca5128af9e4
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-server-with-cryptographic-module@sha256:a012f21f63729ba667af097375e08874156e2cf1b94ae203985d45f11031fea9
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-upgrade-with-cryptographic-module@sha256:caf6e1d2cc31d01958e4adcb4c3c307d761f350d5b0963ce19214f041c14184a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-upgrade-with-cryptographic-module@sha256:b8be0344b6cf99b2c5de2b16f04f6f2b05bde861d9b38afd319bbbc4dd72a620
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-server-with-cryptographic-module@sha256:7778f4b03861b56371d18f7726b68ad21d0faa7c1a25cb38fe7cac483d81c095
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-pgpool2-with-cryptographic-module@sha256:e510bc2d30df33e4c9d3babb898ff9cbd13704fa3fa69a284bf672765fb03083
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-pgpool2-with-cryptographic-module@sha256:e8aa27cbb57503ec5f4f37beeb2e833ace08d4c583d4361b7ce945c7461a904c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-restore-with-cryptographic-module@sha256:edeb7b72a89e05995661e6a17e75b5b477a3ca351e0db146a939a33e88e16c97
        - quay.io/fujitsu/fujitsu-enterprise-postgres-utils-with-cryptographic-module@sha256:7d169df0b8fe47000891afde79be8dd7978da3f893d1e92f01f57fd8c0b3b4c6
        - quay.io/fujitsu/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:eaf601e7aef525ebcd4b769c859ed74c7a8077f843b3a75d76713eabf0295745
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-backup-with-cryptographic-module@sha256:5b49aa46386ca954ebcbea28fbabd15929cb64fc2a4d2dd2796c9d2e15b212c6
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentd-with-cryptographic-module@sha256:f743f35012e6fd3d54a6b54baeeee98bd7c65fe52923acd87546066fa9b54d7e
        version: 7.1.6
      entries:
      - name: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.6
        version: 7.1.6
      - name: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.5
        version: 7.1.5
      name: stable
    defaultChannel: stable
    packageName: fujitsu-enterprise-postgres-operator-with-cg-module
    provider:
      name: Fujitsu
      url: https://www.postgresql.fastware.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-05-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx8-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx8-operator-rhmp/support?utm_source=openshift_console
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:154f85cc6721a70909dee755ddd88897ad51038afe2538b6dee4d26e7d8b9afb
        version: 0.0.10
      entries:
      - name: joget-dx8-openshift-operator.v0.0.10
        version: 0.0.10
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-operator-rhmp
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://www.redhat.com/
    name: rhdh
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhdh-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
          createdAt: "2025-09-12T06:25:11Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.7.1'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.7/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:62f32e50727c5006766a34daa5b6472b649cd9894f3fe9543b8ecc67e6760e8e
        version: 1.7.1
      entries:
      - name: rhdh-operator.v1.7.1
        version: 1.7.1
      - name: rhdh-operator.v1.7.0
        version: 1.7.0
      - name: rhdh-operator.v1.6.5
        version: 1.6.5
      - name: rhdh-operator.v1.6.4
        version: 1.6.4
      - name: rhdh-operator.v1.6.3
        version: 1.6.3
      - name: rhdh-operator.v1.6.2
        version: 1.6.2
      - name: rhdh-operator.v1.6.1
        version: 1.6.1
      - name: rhdh-operator.v1.6.0
        version: 1.6.0
      - name: rhdh-operator.v1.5.3
        version: 1.5.3
      - name: rhdh-operator.v1.5.2
        version: 1.5.2
      - name: rhdh-operator.v1.5.1
        version: 1.5.1
      - name: rhdh-operator.v1.4.3
        version: 1.4.3
      - name: rhdh-operator.v1.4.2
        version: 1.4.2
      - name: rhdh-operator.v1.4.1
        version: 1.4.1
      - name: rhdh-operator.v1.4.0
        version: 1.4.0
      - name: rhdh-operator.v1.3.5
        version: 1.3.5
      - name: rhdh-operator.v1.3.4
        version: 1.3.4
      - name: rhdh-operator.v1.3.3
        version: 1.3.3
      - name: rhdh-operator.v1.3.1
        version: 1.3.1
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast
    - currentCSV: rhdh-operator.v1.1.2-0.1714688890.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:9bea7eabdea44342248dfd6091a8f5c3b6e65884c916b3c2a073a4a64481aa6f
          createdAt: "2023-03-20T16:11:34Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins, configuration settings, and deployment
            mechanisms, which can help streamline the process of setting up a self-managed
            internal developer portal for adopters who are just starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: rhdh-operator.v1.1.2
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.1.0 <1.2.0'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/janus-idp/operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: asoro@redhat.com
          name: Armel Soro
        - email: gazarenk@redhat.com
          name: Gennady Azarenkov
        - email: jianrzha@redhat.com
          name: Jianrong Zhang
        - email: nboldt@redhat.com
          name: Nick Boldt
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:95ed239eb427c7b1e422ece28c5277b5018efb751c2b472bf502b8216219c047
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:9bea7eabdea44342248dfd6091a8f5c3b6e65884c916b3c2a073a4a64481aa6f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:d7ae2456857cab09a9e9acf69c9990ef18deb7f920046b8c2483fbfd0d42bbfc
        - registry.redhat.io/rhel9/postgresql-15@sha256:d5f3d496ecd55688e384cc0627700a97af6924582123bcf00749c375724af3de
        version: 1.1.2+0.1714688890.p
      entries:
      - name: rhdh-operator.v1.1.2
        version: 1.1.2
      - name: rhdh-operator.v1.1.2-0.1714608881.p
        version: 1.1.2+0.1714608881.p
      - name: rhdh-operator.v1.1.2-0.1714688618.p
        version: 1.1.2+0.1714688618.p
      - name: rhdh-operator.v1.1.2-0.1714688890.p
        version: 1.1.2+0.1714688890.p
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.1
    - currentCSV: rhdh-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:c9051a6c339cd40cbc7474a33b32a6b94913af964e12ca5da0c0b7e1c6f444d5
          createdAt: "2024-11-28T17:55:19Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.2.6'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.2/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.2/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: asoro@redhat.com
          name: Armel Soro
        - email: cdaley@redhat.com
          name: Corey Daley
        - email: gazarenk@redhat.com
          name: Gennady Azarenkov
        - email: nboldt@redhat.com
          name: Nick Boldt
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:ef29f2b340a6dc52c5014239876d54a54f7c0f0bf4ea917391438a69c94a8532
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:c9051a6c339cd40cbc7474a33b32a6b94913af964e12ca5da0c0b7e1c6f444d5
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:025eb58c69b9cf5992ecf2fccb29a3d36061cbcf62652a1c6878f0c4d767c1fb
        - registry.redhat.io/rhel9/postgresql-15@sha256:24fb4e7914a6e1464d015be9e5582cc4b9da224137408bd429e7ea4f391aa198
        version: 1.2.6
      entries:
      - name: rhdh-operator.v1.2.6
        version: 1.2.6
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.2
    - currentCSV: rhdh-operator.v1.3.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:5ce579edc10bde5c1c400104f7e0d9f86ca277cd565aaf3dd84b8636f5b949a8
          createdAt: "2025-02-10T18:12:52Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.3.5'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.3/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:5ce579edc10bde5c1c400104f7e0d9f86ca277cd565aaf3dd84b8636f5b949a8
        - registry.redhat.io/rhel9/postgresql-15@sha256:24fb4e7914a6e1464d015be9e5582cc4b9da224137408bd429e7ea4f391aa198
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:c044fed8485a62a5051ac726525da546936be22f268db05491bf81582e0e6194
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b4934c45ba595b36e65c73b734c58d1517e7f43b7cfc309ca8bc853733582f76
        version: 1.3.5
      entries:
      - name: rhdh-operator.v1.3.5
        version: 1.3.5
      - name: rhdh-operator.v1.3.4
        version: 1.3.4
      - name: rhdh-operator.v1.3.3
        version: 1.3.3
      - name: rhdh-operator.v1.3.1
        version: 1.3.1
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.3
    - currentCSV: rhdh-operator.v1.4.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:ae5b52c1b35e82f3c564402e367a2a2028e16806867bba0868643089ba7ed223
          createdAt: "2025-04-03T05:42:32Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.4.3'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.4/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:ae5b52c1b35e82f3c564402e367a2a2028e16806867bba0868643089ba7ed223
        version: 1.4.3
      entries:
      - name: rhdh-operator.v1.4.3
        version: 1.4.3
      - name: rhdh-operator.v1.4.2
        version: 1.4.2
      - name: rhdh-operator.v1.4.1
        version: 1.4.1
      - name: rhdh-operator.v1.4.0
        version: 1.4.0
      name: fast-1.4
    - currentCSV: rhdh-operator.v1.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:8fe167cc649a1c53994c376f3cf3eed8c6e318741a0c08e5aaf07cb0b74e7f8a
          createdAt: "2025-06-05T14:31:21Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.5.3'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.5/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:8fe167cc649a1c53994c376f3cf3eed8c6e318741a0c08e5aaf07cb0b74e7f8a
        - registry.redhat.io/rhel9/postgresql-15@sha256:74377f5a61dd7e3e0ba5d2fa880b7107025433a026396567c79eac927b4212bc
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:2ad8c87e82df27c2b6ceba94179d1f0a40fe775368663c1b5572611888e0e02b
        version: 1.5.3
      entries:
      - name: rhdh-operator.v1.5.3
        version: 1.5.3
      - name: rhdh-operator.v1.5.2
        version: 1.5.2
      - name: rhdh-operator.v1.5.1
        version: 1.5.1
      name: fast-1.5
    - currentCSV: rhdh-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:50f20300c81621221260d03d22b1442750e3f1d2e62f3e511d63925d84c4d1f9
          createdAt: "2025-08-27T18:43:12Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.6.5'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.6/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:50f20300c81621221260d03d22b1442750e3f1d2e62f3e511d63925d84c4d1f9
        - registry.redhat.io/rhel9/postgresql-15@sha256:5e225c1646251d2509b3a6e45fad23abfb11dd5c45d3224b59178f90deb38a70
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:177a7308f61a6b1ba21bc0f7bf550a93d4478e52479e2c94a2437f2152e3d8ff
        version: 1.6.5
      entries:
      - name: rhdh-operator.v1.6.5
        version: 1.6.5
      - name: rhdh-operator.v1.6.4
        version: 1.6.4
      - name: rhdh-operator.v1.6.3
        version: 1.6.3
      - name: rhdh-operator.v1.6.2
        version: 1.6.2
      - name: rhdh-operator.v1.6.1
        version: 1.6.1
      - name: rhdh-operator.v1.6.0
        version: 1.6.0
      name: fast-1.6
    - currentCSV: rhdh-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
          createdAt: "2025-09-12T06:25:11Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.7.1'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.7/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:62f32e50727c5006766a34daa5b6472b649cd9894f3fe9543b8ecc67e6760e8e
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
        version: 1.7.1
      entries:
      - name: rhdh-operator.v1.7.1
        version: 1.7.1
      - name: rhdh-operator.v1.7.0
        version: 1.7.0
      name: fast-1.7
    defaultChannel: fast
    packageName: rhdh
    provider:
      name: Red Hat Inc.
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-activemqartemis: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/
    name: amq-broker-rhel8
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-broker-operator.v7.10.7-opr-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                    "name": "ex-aao",
                    "application": "ex-aao-app"
                },
                "spec": {
                    "deploymentPlan": {
                        "image": "placeholder",
                        "size": 2,
                        "requireLogin": false,
                        "persistenceEnabled": false,
                        "journalType": "nio",
                        "messageMigration": false,
                        "jolokiaAgentEnabled": false,
                        "managementRBACEnabled": true
                    }
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                    "name": "ex-aaoaddress"
                },
                "spec": {
                    "addressName": "myAddress0",
                    "queueName": "myQueue0",
                    "routingType": "anycast"
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                    "name": "ex-aaoscaledown"
                },
                "spec": {
                    "localOnly": "true"
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                    "name": "ex-keycloak"
                },
                "spec": {
                    "loginModules": {
                        "propertiesLoginModules": [
                            {
                                "name": "prop-module",
                                "users": [
                                    {
                                        "name": "superman",
                                        "roles": [
                                            "root"
                                        ]
                                    }
                                ]
                            }
                        ]
                    },
                    "securityDomains": {
                        "brokerDomain": {
                            "name": "activemq",
                            "loginModules": [
                                {
                                    "name": "prop-module",
                                    "flag": "sufficient"
                                }
                            ]
                        }
                    }
                }
            }]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:1a7aa54d2799d238eb5f49f7a95a78a896f6bf8d222567e9118e0e3963cc9aad
          createdAt: "2021-06-18T00:02:04Z"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A stateful deployment of one or more brokers
            displayName: AMQ Broker
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: Adding and removing addresses via custom resource definitions
            displayName: AMQ Broker Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: Provides message migration on clustered broker scaledown
            displayName: AMQ Broker Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: Security configuration for the broker
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
          to deploy and manage stateful AMQ Broker broker clusters
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: rkieley@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:1a7aa54d2799d238eb5f49f7a95a78a896f6bf8d222567e9118e0e3963cc9aad
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ba21779c29b431a22c824e91820705c9501fb0b7d6f7e1ae22fcd3becc8c8a2f
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:d4010af53429502f41d722c39d69a7bb147bb06d19e3d4a69de70dbe88e3990e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b575d6eebf44dc9049445e357488efb17a690953852419b534df30e0105b9ca4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ff6fefad518a6c997d4c5a6e475ba89640260167f0bc27715daf3cc30116fad1
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:56282c6a461e945a83bffb13d3f3d8d9621800f0fc9cd755a75d21eacd8f5fbf
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:979b59325aa0f34eb05625201beba53fccbb83bd5eb80a89dcb5261ae358138f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:83faefa67b33d93836a1be3cb438bb147242d602565dcbe59b4ed42b26072191
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:967d840b831b310abd3a468a8652d1fa48b2ce28b342bfab53999245deb004f1
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:2cb14c917246ffcbd5d861ca6f3d58cb1146a4e55715ee08160e115d3dfc1c3e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f37f98c809c6f29a83e3d5a3ac4494e28efe9b25d33c54f533c6a08662244622
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:4e09ee3bd8ff3479fdd00e221b7b1505ae04ca6c64fd6bc9e2f1a34b8e98d137
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0d857b95084e39c011651be89482d4dcc77c2d1c3ac3c6a819f3c9d5a57a425f
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:9465e398a55c89c121b3ae50374414190a991839fa5e1018fc02551f6aa3b9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3c00b524722b500ff0922fe7a8722868557512183a8f69248f37a68c45c6a770
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a94fa8135d3765363e442bda45df6e02a3e22bee7ea198bcb0d3584331a4320c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c5f4c08e068b9721967cf7c7cbd9a9e93fb5e39b264dd13b653e99d8f3fa9e0e
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:aa7af7e038a12b5202108b362c8af243954603e84034a16c0192684c86d5d28b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b74d03ed852a3731467ffda95266ce49f2065972f1c37bf254f3d52b34c11991
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:7abfdcfd7b2d3d12b9234f7b057bd199bceefa47bdf02f62236f82a13fdbf5b7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:869cda365447b0b7b53ed21d2496884c15dc9c07295a9dfa0337c2c9974fe018
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:5ff28e3062d3eabb9f52d8271b837fdab06cafbd8b8f732b1bc00de49ad69aad
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f58678f03a250dbb296ebeb112c9150340155799ba4dad9e90bf347b9ed0ca7b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3219f90789690a3c5cfcc86e5f602c443e75f1fdb3a19660f43ee3996bfa8f62
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:f2651cb097eb631fc6af90e4a8404d36516aa2205c546125bf0c37739ab525fa
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:35b7d96d5065f280a528626abd2c3df84e79ccea93ab6bd8b8837af66c330f30
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:45a5e054546a2c1ee338518f8477f728f5f7108c9dd8b516b4c79d4a7d6f5ba7
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:1a9f13fd565113055b2aa2022bd0709a63c1dc058a8570e0fd7eba5c7b101334
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:97c3aa5673610875a53571449f7d5210ae8d675072d044f17b416725ed0dba88
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:ba0f5df7da08bc030ff5e0e68ecb52a931c25f6b67c6556aa8a9661c82d69d42
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:83ed89fa942ccbf4451954956db42138c83f9b015e7de07fc159ddc4bb6dc843
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9a02decbc460ddfc501d8a4a9b41cd084776090951906b027838c0829e1fc15a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ff819e913406911735da39c51699405f747bd697116a273a00897ff176bf05ff
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:5d39f26f06769fe5c989ea0e0ee8f9166120e152ce2c9bc3336fc8d7cab2ff92
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:e6f71001dccfbbe1cd7cf4fded69e181c10f46cc5c8bd3a8d3045a3f3ac41e2f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:61e5a8c1e46f1c18648e0406e2a57c52f911979079020cc57c7a4b06a56940bb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:e03945756d86bd189e8c33027fa668d2065c168510085dfd06defd9f40d3e0ab
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:34739366153b825dcb113e0ed6115ef5656085cf8d47492789ff17464042b576
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:34ab1f66d8daa544e3823464383e9788e307665c875499838fbed95f3a74a4dd
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0b2126cfb6054fdf428c1f43b69e36e93a09a49ce15350e9273c98cc08c6598b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a8b2a9364fd06c8e1c29b8f8f28d3e4844bf7840a62b67137b663c558982d4a3
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:6e9e0e3fddc44fbf87591d8904176407f1ec409d578a7d108c13cdd48d85131b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:67ae74e72927557c55611d0248347b2099b117ac1aa245dac05856fe782fb59c
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:c3ac4602b48867f3d4f7cc0efd5c9f133098e409626cfb402f80386ff460dbf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c9eba0d1f9fba588711774f4db75221f057578407065127ed60c33eef564f499
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:86ff145bd68e797a0507495570063b54cbe4186fcb4acd83c900406619ef9524
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:d4a9176f49ffd9db6051fbda15d47a1aa32c282fb5010d0caf5b9243e80bf055
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:4440b20349032bb5c0bb2a8deb89832b37767af6f0b3208b067867a85641913e
        version: 7.10.7-opr-1
      entries:
      - name: amq-broker-operator.v7.10.7-opr-1
        version: 7.10.7-opr-1
      - name: amq-broker-operator.v7.10.6-opr-1
        version: 7.10.6-opr-1
      - name: amq-broker-operator.v7.10.5-opr-1
        version: 7.10.5-opr-1
      - name: amq-broker-operator.v7.10.4-opr-1
        version: 7.10.4-opr-1
      - name: amq-broker-operator.v7.10.3-opr-3
        version: 7.10.3-opr-3
      - name: amq-broker-operator.v7.10.3-opr-2
        version: 7.10.3-opr-2
      - name: amq-broker-operator.v7.10.3-opr-1
        version: 7.10.3-opr-1
      - name: amq-broker-operator.v7.10.2-opr-2-0.1676475747.p
        version: 7.10.2-opr-2+0.1676475747.p
      - name: amq-broker-operator.v7.10.2-opr-2-0.1680622941.p
        version: 7.10.2-opr-2+0.1680622941.p
      - name: amq-broker-operator.v7.10.2-opr-2
        version: 7.10.2-opr-2
      - name: amq-broker-operator.v7.10.2-opr-1
        version: 7.10.2-opr-1
      - name: amq-broker-operator.v7.10.1-opr-2
        version: 7.10.1-opr-2
      - name: amq-broker-operator.v7.10.1-opr-1
        version: 7.10.1-opr-1
      - name: amq-broker-operator.v7.10.0-opr-4
        version: 7.10.0-opr-4
      - name: amq-broker-operator.v7.10.0-opr-3
        version: 7.10.0-opr-3
      - name: amq-broker-operator.v7.10.0-opr-2
        version: 7.10.0-opr-2
      - name: amq-broker-operator.v7.10.0-opr-1
        version: 7.10.0-opr-1
      name: 7.10.x
    - currentCSV: amq-broker-operator.v7.11.8-opr-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisSecurity is the Schema for the activemqartemissecurities
              API
            displayName: Active MQArtemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:b9aecae6b0ce0b241c7a5b9e6e096395940bb6a384373a4d121822cf10c26d99
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b74d03ed852a3731467ffda95266ce49f2065972f1c37bf254f3d52b34c11991
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:442b6638133d42fa7faa41d86e47be95cdca56c2e16efcd6e7e22991c3812a36
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f37f98c809c6f29a83e3d5a3ac4494e28efe9b25d33c54f533c6a08662244622
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0b2126cfb6054fdf428c1f43b69e36e93a09a49ce15350e9273c98cc08c6598b
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:4440b20349032bb5c0bb2a8deb89832b37767af6f0b3208b067867a85641913e
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:7abfdcfd7b2d3d12b9234f7b057bd199bceefa47bdf02f62236f82a13fdbf5b7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b9f93c243ead118651083880786a81c624f390fcc3596c7ec5ebd55f34f95ea3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:35b7d96d5065f280a528626abd2c3df84e79ccea93ab6bd8b8837af66c330f30
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b575d6eebf44dc9049445e357488efb17a690953852419b534df30e0105b9ca4
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:f2651cb097eb631fc6af90e4a8404d36516aa2205c546125bf0c37739ab525fa
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:83ed89fa942ccbf4451954956db42138c83f9b015e7de07fc159ddc4bb6dc843
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:967d840b831b310abd3a468a8652d1fa48b2ce28b342bfab53999245deb004f1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0d857b95084e39c011651be89482d4dcc77c2d1c3ac3c6a819f3c9d5a57a425f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:e03945756d86bd189e8c33027fa668d2065c168510085dfd06defd9f40d3e0ab
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:56282c6a461e945a83bffb13d3f3d8d9621800f0fc9cd755a75d21eacd8f5fbf
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5e5d8b551ab8ca377e0bc939037970a2f31ee14e9f555853ebea2759d673895a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9a02decbc460ddfc501d8a4a9b41cd084776090951906b027838c0829e1fc15a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0f443f5aa818dd85bfa93e51e37bdac3909628a283aa52a89125b527dd75d9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:5d39f26f06769fe5c989ea0e0ee8f9166120e152ce2c9bc3336fc8d7cab2ff92
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:9465e398a55c89c121b3ae50374414190a991839fa5e1018fc02551f6aa3b9aa
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:1a9f13fd565113055b2aa2022bd0709a63c1dc058a8570e0fd7eba5c7b101334
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:45a5e054546a2c1ee338518f8477f728f5f7108c9dd8b516b4c79d4a7d6f5ba7
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c9eba0d1f9fba588711774f4db75221f057578407065127ed60c33eef564f499
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a94fa8135d3765363e442bda45df6e02a3e22bee7ea198bcb0d3584331a4320c
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:34ab1f66d8daa544e3823464383e9788e307665c875499838fbed95f3a74a4dd
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:86ff145bd68e797a0507495570063b54cbe4186fcb4acd83c900406619ef9524
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c5f4c08e068b9721967cf7c7cbd9a9e93fb5e39b264dd13b653e99d8f3fa9e0e
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:d4a9176f49ffd9db6051fbda15d47a1aa32c282fb5010d0caf5b9243e80bf055
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:29f07006ad62023c20d34971cb3ae900349efae5ab2918aaa8e5647d788032be
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ff819e913406911735da39c51699405f747bd697116a273a00897ff176bf05ff
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b0b3684ef1ebf8b44d3a4c46391984173792dde1bd7363ea06d43a8f95a717d0
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:cc438b4bc2b14f6bff4e5d32e7dc0fad6d2ca391c4e84b5ae925b7b40af76729
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ba21779c29b431a22c824e91820705c9501fb0b7d6f7e1ae22fcd3becc8c8a2f
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:ba0f5df7da08bc030ff5e0e68ecb52a931c25f6b67c6556aa8a9661c82d69d42
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:e6f71001dccfbbe1cd7cf4fded69e181c10f46cc5c8bd3a8d3045a3f3ac41e2f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:67ae74e72927557c55611d0248347b2099b117ac1aa245dac05856fe782fb59c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:979b59325aa0f34eb05625201beba53fccbb83bd5eb80a89dcb5261ae358138f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ade37271426ab8f543d72f8da3e9f663d665f9427379a2539fc9323f6897cc44
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:34739366153b825dcb113e0ed6115ef5656085cf8d47492789ff17464042b576
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:869cda365447b0b7b53ed21d2496884c15dc9c07295a9dfa0337c2c9974fe018
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ff6fefad518a6c997d4c5a6e475ba89640260167f0bc27715daf3cc30116fad1
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:61e5a8c1e46f1c18648e0406e2a57c52f911979079020cc57c7a4b06a56940bb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0dde839c2dce7cb684094bf26523c8e16677de03149a0fff468b8c3f106e1f4f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5f842ebe9a6a4294a056ad10189a5f277cfd3fb4e6e0b3225e194bd5c131151d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3c00b524722b500ff0922fe7a8722868557512183a8f69248f37a68c45c6a770
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:4e09ee3bd8ff3479fdd00e221b7b1505ae04ca6c64fd6bc9e2f1a34b8e98d137
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:d4010af53429502f41d722c39d69a7bb147bb06d19e3d4a69de70dbe88e3990e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a8b2a9364fd06c8e1c29b8f8f28d3e4844bf7840a62b67137b663c558982d4a3
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:5ff28e3062d3eabb9f52d8271b837fdab06cafbd8b8f732b1bc00de49ad69aad
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:d2ae3defac36f00112c6b773e13626336ba15d6f50f905ad1c33f122bfe1a6a9
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:97c3aa5673610875a53571449f7d5210ae8d675072d044f17b416725ed0dba88
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f58678f03a250dbb296ebeb112c9150340155799ba4dad9e90bf347b9ed0ca7b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:40ccaf2bb2e0f8132f2541b39cb835f17b38050b3ca9232f0861fd5fcaf4324a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:6e9e0e3fddc44fbf87591d8904176407f1ec409d578a7d108c13cdd48d85131b
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:c3ac4602b48867f3d4f7cc0efd5c9f133098e409626cfb402f80386ff460dbf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:137b893ad15e0cd5353bae134c8fae833257a12641b12d0f774f1565ce66f63c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:89bb769b4c33eb931cc7866eeabec9a315b913851cfa84bc6e11ae8d1e8b2cbb
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:2cb14c917246ffcbd5d861ca6f3d58cb1146a4e55715ee08160e115d3dfc1c3e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:25c3732e945b4bc3e9aca721ce709d0217339b065d9dac9afdb6202e9fd8a799
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:8b4c4400a02c6cacfd978147b158e2ccbe5da955578b4b970a5ee7edac865326
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3219f90789690a3c5cfcc86e5f602c443e75f1fdb3a19660f43ee3996bfa8f62
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:7d53744ae8c77aea09c59165486c9ed0eaa131140f6b346042293bce0bd8a745
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:aa7af7e038a12b5202108b362c8af243954603e84034a16c0192684c86d5d28b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9c1ac38bf8eec67a3da57c3cc61105513776a749408cb287fe6448a17889de4e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:83faefa67b33d93836a1be3cb438bb147242d602565dcbe59b4ed42b26072191
        version: 7.11.8-opr-1
      entries:
      - name: amq-broker-operator.v7.11.8-opr-1
        version: 7.11.8-opr-1
      - name: amq-broker-operator.v7.11.7-opr-2-1742200979
        version: 7.11.7-opr-2-1742200979
      - name: amq-broker-operator.v7.11.7-opr-2-1740583644
        version: 7.11.7-opr-2-1740583644
      - name: amq-broker-operator.v7.11.7-opr-2-1736956309
        version: 7.11.7-opr-2-1736956309
      - name: amq-broker-operator.v7.11.7-opr-2-1731330819
        version: 7.11.7-opr-2-1731330819
      - name: amq-broker-operator.v7.11.7-opr-2
        version: 7.11.7-opr-2
      - name: amq-broker-operator.v7.11.7-opr-1-1726157430
        version: 7.11.7-opr-1-1726157430
      - name: amq-broker-operator.v7.11.7-opr-1-1723204800
        version: 7.11.7-opr-1-1723204800
      - name: amq-broker-operator.v7.11.7-opr-1
        version: 7.11.7-opr-1
      - name: amq-broker-operator.v7.11.6-opr-2
        version: 7.11.6-opr-2
      - name: amq-broker-operator.v7.11.5-opr-1
        version: 7.11.5-opr-1
      - name: amq-broker-operator.v7.11.4-opr-1
        version: 7.11.4-opr-1
      - name: amq-broker-operator.v7.11.3-opr-1
        version: 7.11.3-opr-1
      - name: amq-broker-operator.v7.11.2-opr-1
        version: 7.11.2-opr-1
      - name: amq-broker-operator.v7.11.1-opr-2
        version: 7.11.1-opr-2
      - name: amq-broker-operator.v7.11.1-opr-1
        version: 7.11.1-opr-1
      - name: amq-broker-operator.v7.11.0-opr-3
        version: 7.11.0-opr-3
      - name: amq-broker-operator.v7.11.0-opr-2
        version: 7.11.0-opr-2
      name: 7.11.x
    - currentCSV: amq-broker-operator.v7.12.5-opr-1-0.1757928550.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-broker-operator.v7.12.5-opr-1-0.1756216162.p
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:7d53744ae8c77aea09c59165486c9ed0eaa131140f6b346042293bce0bd8a745
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9c1ac38bf8eec67a3da57c3cc61105513776a749408cb287fe6448a17889de4e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:29f07006ad62023c20d34971cb3ae900349efae5ab2918aaa8e5647d788032be
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:dc13871dfed52402c40fd192900fc3cb5a49da4fa08e4ec4d0e23a879e83db25
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5e5d8b551ab8ca377e0bc939037970a2f31ee14e9f555853ebea2759d673895a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:cc438b4bc2b14f6bff4e5d32e7dc0fad6d2ca391c4e84b5ae925b7b40af76729
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d71506472636d940b5dbd3e708d3e0448d777eeecfd38de61ba49e281d2979c6
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ffffa9875f0379e9373f89f05eb06e5a193273bb04bc3aa5f85b044357b79098
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:b9aecae6b0ce0b241c7a5b9e6e096395940bb6a384373a4d121822cf10c26d99
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:442b6638133d42fa7faa41d86e47be95cdca56c2e16efcd6e7e22991c3812a36
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:25c3732e945b4bc3e9aca721ce709d0217339b065d9dac9afdb6202e9fd8a799
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ade37271426ab8f543d72f8da3e9f663d665f9427379a2539fc9323f6897cc44
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:16253895bc113f395dc6ab5cd511901e43b6ee676369ec80d598a8741c3c9bf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:89bb769b4c33eb931cc7866eeabec9a315b913851cfa84bc6e11ae8d1e8b2cbb
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:bb6fbd68475a7852b4d99eea6c4ab313f9267da7963162f0d75375d7063409e7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b9f93c243ead118651083880786a81c624f390fcc3596c7ec5ebd55f34f95ea3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9cc48eecf1442ae04b8543fa5d4381a13bc2831390850828834d387006d1342b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:47fd4ce2533496828aba37bd1f9715e2164d5c90bd0fc6b25e7e0786d723bf01
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:023478b773a0c42739b86e822a4d435e01ef341b8136566ea52463f1b7e447ed
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:eacc4fb55e8c4c30894a51944960d536508001a49fda62782c6cb2ca8767961d
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0dde839c2dce7cb684094bf26523c8e16677de03149a0fff468b8c3f106e1f4f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:8b4c4400a02c6cacfd978147b158e2ccbe5da955578b4b970a5ee7edac865326
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d42d713da0ce6806fdc6492b6342586783e6865a82a8647d3c4288439b1751ee
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:7f9b921d89b6d4d91b365659afff127d99f1455b94b48a10a6a7f4c38dd8b700
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:90f1678c19c74454b77aed14fa6b3074a1c6f80fe884f1c6136e7e907f402aec
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b2402179b1d0ff9c950defe5d8808195ef9a4c3e13e3c469cae31ab51d5f5014
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:8329acdff170c5a8ae6ae16f6c37d3f7ab68d106a6e8be50c3fe131ec5ecceeb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5f842ebe9a6a4294a056ad10189a5f277cfd3fb4e6e0b3225e194bd5c131151d
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:40920687402fd747ca02d384e63692a5e1db625723a2bf3056a607c92ee8eee1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0f443f5aa818dd85bfa93e51e37bdac3909628a283aa52a89125b527dd75d9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:40ccaf2bb2e0f8132f2541b39cb835f17b38050b3ca9232f0861fd5fcaf4324a
        version: 7.12.5-opr-1+0.1757928550.p
      entries:
      - name: amq-broker-operator.v7.12.5-opr-1-0.1757928550.p
        version: 7.12.5-opr-1+0.1757928550.p
      - name: amq-broker-operator.v7.12.5-opr-1-0.1756216162.p
        version: 7.12.5-opr-1+0.1756216162.p
      - name: amq-broker-operator.v7.12.5-opr-1-0.1754272113.p
        version: 7.12.5-opr-1+0.1754272113.p
      - name: amq-broker-operator.v7.12.5-opr-1
        version: 7.12.5-opr-1
      - name: amq-broker-operator.v7.12.4-opr-1
        version: 7.12.4-opr-1
      - name: amq-broker-operator.v7.12.4-opr-1-0.1747217191.p
        version: 7.12.4-opr-1+0.1747217191.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741881887.p
        version: 7.12.3-opr-1+0.1741881887.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739991235.p
        version: 7.12.3-opr-1+0.1739991235.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1730142550.p
        version: 7.12.3-opr-1+0.1730142550.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1730813578.p
        version: 7.12.3-opr-1+0.1730813578.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1731483798.p
        version: 7.12.3-opr-1+0.1731483798.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1732654016.p
        version: 7.12.3-opr-1+0.1732654016.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1733315503.p
        version: 7.12.3-opr-1+0.1733315503.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1737643631.p
        version: 7.12.3-opr-1+0.1737643631.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1738596517.p
        version: 7.12.3-opr-1+0.1738596517.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739390944.p
        version: 7.12.3-opr-1+0.1739390944.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739889884.p
        version: 7.12.3-opr-1+0.1739889884.p
      - name: amq-broker-operator.v7.12.3-opr-1
        version: 7.12.3-opr-1
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741884937.p
        version: 7.12.3-opr-1+0.1741884937.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741883943.p
        version: 7.12.3-opr-1+0.1741883943.p
      - name: amq-broker-operator.v7.12.2-opr-1
        version: 7.12.2-opr-1
      - name: amq-broker-operator.v7.12.1-opr-1-0.1724054375.p
        version: 7.12.1-opr-1+0.1724054375.p
      - name: amq-broker-operator.v7.12.1-opr-1-0.1723669176.p
        version: 7.12.1-opr-1+0.1723669176.p
      - name: amq-broker-operator.v7.12.1-opr-1-0.1721247907.p
        version: 7.12.1-opr-1+0.1721247907.p
      - name: amq-broker-operator.v7.12.1-opr-1
        version: 7.12.1-opr-1
      - name: amq-broker-operator.v7.12.0-opr-2
        version: 7.12.0-opr-2
      - name: amq-broker-operator.v7.12.0-opr-1
        version: 7.12.0-opr-1
      name: 7.12.x
    defaultChannel: 7.12.x
    packageName: amq-broker-rhel8
    provider:
      name: Red Hat
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecrpublic-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecrpublic-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecrpublic-controller:1.1.1
          createdAt: "2025-09-20T01:21:52Z"
          description: AWS ECR PUBLIC controller is a service controller for managing
            ECR PUBLIC resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Repository represents the state of an AWS ecrpublic Repository
              resource.
            displayName: Repository
            kind: Repository
            name: repositories.ecrpublic.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ECR PUBLIC resources in AWS from within your Kubernetes cluster.

          **About Amazon ECR PUBLIC**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECR PUBLIC
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecrpublic
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECR PUBLIC Developer Resources
          url: https://aws.amazon.com/ECR PUBLIC/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecrpublic maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecrpublic-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ecrpublic-controller.v1.1.1
        version: 1.1.1
      - name: ack-ecrpublic-controller.v1.1.0
        version: 1.1.0
      - name: ack-ecrpublic-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecrpublic-controller.v1.0.12
        version: 1.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecrpublic-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: egressip-ipam-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: egressip-ipam-operator.v1.2.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "EgressIPAM",
                "metadata": {
                  "name": "egressipam-aws"
                },
                "spec": {
                  "cidrAssignments": [
                    {
                      "CIDR": "10.0.128.0/19",
                      "labelValue": "us-west-2a"
                    },
                    {
                      "CIDR": "10.0.160.0/19",
                      "labelValue": "us-west-2b",
                      "reservedIPs": [
                        "10.0.160.126"
                      ]
                    },
                    {
                      "CIDR": "10.0.192.0/19",
                      "labelValue": "us-west-2c"
                    }
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "topologyLabel": "failure-domain.beta.kubernetes.io/zone"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "EgressIPAM",
                "metadata": {
                  "name": "egressipam-baremetal"
                },
                "spec": {
                  "cidrAssignments": [
                    {
                      "CIDR": "192.169.0.0/24",
                      "labelValue": "true"
                    }
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "topologyLabel": "egressGateway"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Networking, Cloud Provider, Security
          certified: "false"
          containerImage: quay.io/redhat-cop/egressip-ipam-operator@sha256:d6129bd287b87696a23e0da64dcd598958f779dc88ae33b46b3c909dbab80adf
          createdAt: "2022-07-25T15:07:22Z"
          description: This operator provides automation around managing egressIPs.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: egressip-ipam-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/egressip-ipam-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgressIPAM is the Schema for the egressipams API
            displayName: Egress IPAM
            kind: EgressIPAM
            name: egressipams.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator automates the assignment of egressIPs to namespaces.
          Namespaces can opt in to receiving one or more egressIPs with the following annotation `egressip-ipam-operator.redhat-cop.io/egressipam:<egressIPAM>`, where `egressIPAM` is the CRD that controls how egressIPs are assigned.
          IPs assigned to the namespace can be looked up in the following annotation: `egressip-ipam-operator.redhat-cop.io/egressips`.

          EgressIP assignments is managed by the EgressIPAM CRD. Here is an example of it:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: EgressIPAM
          metadata:
            name: example-egressipam
          spec:
            cidrAssignments:
              - labelValue: "true"
                CIDR: 192.169.0.0/24
                reservedIPs:
                - "192.159.0.5"
            topologyLabel: egressGateway
            nodeSelector:
              node-role.kubernetes.io/worker: ""
          ```

          This EgressIPAM specifies that all nodes that comply with the specified node selector and that also have labels `egressGateway: "true"` will be assigned egressIP from the specified CIDR.

          Note that the `cidrAssigments` field is an array and therefore, multiple groups of nodes can be identified with the labelValue and different CIDRs can be assigned to them. This is usually not necessary on a bare metal deployment.

          In the bare metal scenario, when this egressCRD is created, all the `hostsubnet` relative to the nodes selected by this EgressIPAM will be update to have the EgressCIDRs field equal to the specified CIDR (see below for cloud provider scenarios).

          When a namespace is created with the opt-in annotation: `egressip-ipam-operator.redhat-cop.io/egressipam=<egressIPAM>`, an available egressIP is selected from the CIDR and assigned to the namespace.The `netnamespace` associated with this namespace is updated to use that egressIP.

          It is possible to specify a set of reserved IPs. These IPs must belong to the CIDR and will never be assigned.

          ## Passing EgressIPs as input

          The normal mode of operation of this operator is to pick a random IP from the configured CIDR. However, it also supports a scenario where egressIPs are picked by an external process and passed as input.

          In this case IPs must me passed as an annotation to the namespace, like this: `egressip-ipam-operator.redhat-cop.io/egressips=IP1,IP2...`. The value of the annotation is a comma separated array of ip with no spaces.

          There must be exactly one IP per CIDR defined in the referenced egressIPAM. Moreover, each IP must belong to the corresponding CIDR. Because this situation can lead to inconsistencies, failing to have correspondence between IPs in the namespace annotation and CIDRs in the egressIPAM CR will cause the operator to error out and stop processing all the namespaces associated with the given EgressIPAM CR.

          It also also a responsibility of the progress picking the IPs to ensure that those IPs are actually available.

          ## Assumptions

          The following assumptions apply when using this operator:

          1. If multiple EgressIPAMs are defined, their selected nodes MUST NOT overlap.
          2. When using a cloud provider the topology label MUST be `failure-domain.beta.kubernetes.io/zone`.

          ## Support for AWS

          In AWS as well as other cloud providers, one cannot freely assign IPs to machines. Additional steps need to be performed in this case. Considering this EgressIPAM

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: EgressIPAM
          metadata:
            name: egressipam-aws
          spec:
            cidrAssignments:
              - labelValue: "eu-central-1a"
                CIDR: 10.0.128.0/20
              - labelValue: "eu-central-1b"
                CIDR: 10.0.144.0/20
              - labelValue: "eu-central-1c"
                CIDR: 10.0.160.0/20
            topologyLabel: failure-domain.beta.kubernetes.io/zone
            nodeSelector:
              matchLabels:
                node-role.kubernetes.io/worker: ""
          ```

          When a namespace with the opt-in annotation is created, the following happens:

          1. for each of the CIDRs, an available IP is selected and assigned to the namespace
          2. the relative `netnamespace` is update to reflect the assignment (multiple IPs will be assigned in this case).
          3. one node per zone is selected to carry the egressIP
          4. the relative aws machines are assigned the additional IP on the main interface (support for secondary interfaces in not available)
          5. the relative `hostsubnets` are updated to reflect the assigned IP, the `egressIP` field is updated.
        displayName: Egressip Ipam Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - egressIPs
        - high availability
        - firewall support
        - traffic management
        - ipam
        links:
        - name: Egressip Ipam Operator
          url: https://github.com/redhat-cop/egressip-ipam-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/egressip-ipam-operator:v0.1.2
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/egressip-ipam-operator@sha256:d6129bd287b87696a23e0da64dcd598958f779dc88ae33b46b3c909dbab80adf
        version: 1.2.4
      entries:
      - name: egressip-ipam-operator.v1.2.4
        version: 1.2.4
      - name: egressip-ipam-operator.v1.2.3
        version: 1.2.3
      - name: egressip-ipam-operator.v1.2.2
        version: 1.2.2
      - name: egressip-ipam-operator.v1.2.1
        version: 1.2.1
      - name: egressip-ipam-operator.v1.2.0
        version: 1.2.0
      - name: egressip-ipam-operator.v1.1.0
        version: 1.1.0
      - name: egressip-ipam-operator.v1.0.8
        version: 1.0.8
      - name: egressip-ipam-operator.v1.0.7
        version: 1.0.7
      - name: egressip-ipam-operator.v1.0.6
        version: 1.0.6
      - name: egressip-ipam-operator.v1.0.5
        version: 1.0.5
      - name: egressip-ipam-operator.v1.0.4
        version: 1.0.4
      - name: egressip-ipam-operator.v1.0.3
        version: 1.0.3
      - name: egressip-ipam-operator.v1.0.2
        version: 1.0.2
      - name: egressip-ipam-operator.v1.0.1
        version: 1.0.1
      - name: egressip-ipam-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: egressip-ipam-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: hive-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hive-operator.v1.2.4953-c248a51
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:c248a51ae1
          createdAt: "2025-09-22T13:41:19Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeploymentCustomization is the Schema for clusterdeploymentcustomizations
              API.
            displayName: ClusterDeploymentCustomization
            kind: ClusterDeploymentCustomization
            name: clusterdeploymentcustomizations.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: |-
              ClusterPool represents a pool of clusters that should be kept ready to be given out to users. Clusters are removed
              from the pool once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: |-
              MachinePoolNameLease is the Schema for the MachinePoolNameLeases API. This resource is mostly empty
              as we're primarily relying on the name to determine if a lease is available.
              Note that not all cloud providers require the use of a lease for naming, at present this
              is only required for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * IBM Cloud
          * Red Hat OpenStack
          * vSphere

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Troubleshooting](https://github.com/openshift/hive/blob/master/docs/troubleshooting.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
              logLevel: debug
              targetNamespace: hive
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:c248a51ae1
        version: 1.2.4953-c248a51
      entries:
      - name: hive-operator.v1.2.4953-c248a51
        version: 1.2.4953-c248a51
      - name: hive-operator.v1.2.4928-76a65b6
        version: 1.2.4928-76a65b6
      - name: hive-operator.v1.2.4902-4929f16
        version: 1.2.4902-4929f16
      - name: hive-operator.v1.2.4896-a29a238
        version: 1.2.4896-a29a238
      - name: hive-operator.v1.2.4868-45ae912
        version: 1.2.4868-45ae912
      - name: hive-operator.v1.2.4863-e9ad0fa
        version: 1.2.4863-e9ad0fa
      - name: hive-operator.v1.2.4774-87bff59
        version: 1.2.4774-87bff59
      - name: hive-operator.v1.2.4766-6a82857
        version: 1.2.4766-6a82857
      - name: hive-operator.v1.2.4746-d25e0db
        version: 1.2.4746-d25e0db
      - name: hive-operator.v1.2.4738-b19773e
        version: 1.2.4738-b19773e
      - name: hive-operator.v1.2.4731-41f153f
        version: 1.2.4731-41f153f
      - name: hive-operator.v1.2.4703-02f3f00
        version: 1.2.4703-02f3f00
      - name: hive-operator.v1.2.4693-ec3bc72
        version: 1.2.4693-ec3bc72
      - name: hive-operator.v1.2.4686-553c6c2
        version: 1.2.4686-553c6c2
      - name: hive-operator.v1.2.4682-406d6ca
        version: 1.2.4682-406d6ca
      - name: hive-operator.v1.2.4672-af54e2f
        version: 1.2.4672-af54e2f
      - name: hive-operator.v1.2.4668-3ccd44d
        version: 1.2.4668-3ccd44d
      - name: hive-operator.v1.2.4664-00f84cc
        version: 1.2.4664-00f84cc
      - name: hive-operator.v1.2.4658-c13e207
        version: 1.2.4658-c13e207
      - name: hive-operator.v1.2.4654-3268a18
        version: 1.2.4654-3268a18
      - name: hive-operator.v1.2.4650-531d549
        version: 1.2.4650-531d549
      - name: hive-operator.v1.2.4644-365f074
        version: 1.2.4644-365f074
      - name: hive-operator.v1.2.4638-56d6d64
        version: 1.2.4638-56d6d64
      - name: hive-operator.v1.2.4625-389d717
        version: 1.2.4625-389d717
      - name: hive-operator.v1.2.4615-52100dd
        version: 1.2.4615-52100dd
      - name: hive-operator.v1.2.4605-b41ca3f
        version: 1.2.4605-b41ca3f
      - name: hive-operator.v1.2.4584-95efcb7
        version: 1.2.4584-95efcb7
      - name: hive-operator.v1.2.4574-1ec27ad
        version: 1.2.4574-1ec27ad
      - name: hive-operator.v1.2.4568-b3eca52
        version: 1.2.4568-b3eca52
      - name: hive-operator.v1.2.4564-bdf9d08
        version: 1.2.4564-bdf9d08
      - name: hive-operator.v1.2.4552-bf2d689
        version: 1.2.4552-bf2d689
      - name: hive-operator.v1.2.4532-cddd4f9
        version: 1.2.4532-cddd4f9
      - name: hive-operator.v1.2.4511-1fce770
        version: 1.2.4511-1fce770
      - name: hive-operator.v1.2.4500-b1ac27b
        version: 1.2.4500-b1ac27b
      - name: hive-operator.v1.2.4485-47df0d9
        version: 1.2.4485-47df0d9
      - name: hive-operator.v1.2.4469-7be6433
        version: 1.2.4469-7be6433
      - name: hive-operator.v1.2.4463-d7ead60
        version: 1.2.4463-d7ead60
      - name: hive-operator.v1.2.4461-97686bb
        version: 1.2.4461-97686bb
      - name: hive-operator.v1.2.4459-f910c9b
        version: 1.2.4459-f910c9b
      - name: hive-operator.v1.2.4451-83aedb9
        version: 1.2.4451-83aedb9
      - name: hive-operator.v1.2.4446-e11a23b
        version: 1.2.4446-e11a23b
      - name: hive-operator.v1.2.4441-fdbfaf0
        version: 1.2.4441-fdbfaf0
      - name: hive-operator.v1.2.4436-44323bc
        version: 1.2.4436-44323bc
      - name: hive-operator.v1.2.4428-ceaafa8
        version: 1.2.4428-ceaafa8
      - name: hive-operator.v1.2.4426-fb27114
        version: 1.2.4426-fb27114
      - name: hive-operator.v1.2.4420-fe0e302
        version: 1.2.4420-fe0e302
      - name: hive-operator.v1.2.4418-eb5403f
        version: 1.2.4418-eb5403f
      - name: hive-operator.v1.2.4416-fba8d52
        version: 1.2.4416-fba8d52
      - name: hive-operator.v1.2.4408-9c5806a
        version: 1.2.4408-9c5806a
      - name: hive-operator.v1.2.4404-410fed9
        version: 1.2.4404-410fed9
      - name: hive-operator.v1.2.4400-8c54fc9
        version: 1.2.4400-8c54fc9
      - name: hive-operator.v1.2.4386-6f1fd3e
        version: 1.2.4386-6f1fd3e
      - name: hive-operator.v1.2.4382-0e95d5b
        version: 1.2.4382-0e95d5b
      - name: hive-operator.v1.2.4366-df65e7e
        version: 1.2.4366-df65e7e
      - name: hive-operator.v1.2.4360-de169dd
        version: 1.2.4360-de169dd
      - name: hive-operator.v1.2.4355-e5f809f
        version: 1.2.4355-e5f809f
      name: alpha
    - currentCSV: hive-operator.v2.5.3516-a2ed9b3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.5.3516-a2ed9b3
          createdAt: "2022-06-03T13:24:12Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:v2.5.3516-a2ed9b3
        version: 2.5.3516-a2ed9b3
      entries:
      - name: hive-operator.v2.5.3516-a2ed9b3
        version: 2.5.3516-a2ed9b3
      - name: hive-operator.v2.5.3514-4052426
        version: 2.5.3514-4052426
      - name: hive-operator.v2.5.3508-6cb94c6
        version: 2.5.3508-6cb94c6
      - name: hive-operator.v2.5.3506-f42617c
        version: 2.5.3506-f42617c
      - name: hive-operator.v2.5.3495-5cfdd44
        version: 2.5.3495-5cfdd44
      - name: hive-operator.v2.5.3491-6feded1
        version: 2.5.3491-6feded1
      - name: hive-operator.v2.5.3489-6dec9c3
        version: 2.5.3489-6dec9c3
      name: mce-2.0
    - currentCSV: hive-operator.v2.3.3039-60b8a9a
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.3.3039-60b8a9a
          createdAt: "2022-08-15T17:23:39Z"
          description: OpenShift cluster provisioning and management at scale.
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:v2.3.3039-60b8a9a
        version: 2.3.3039-60b8a9a
      entries:
      - name: hive-operator.v2.3.3039-60b8a9a
        version: 2.3.3039-60b8a9a
      name: ocm-2.3
    - currentCSV: hive-operator.v2.4.3231-ba51985
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.4.3231-ba51985
          createdAt: "2022-08-15T11:53:28Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:v2.4.3231-ba51985
        version: 2.4.3231-ba51985
      entries:
      - name: hive-operator.v2.4.3231-ba51985
        version: 2.4.3231-ba51985
      - name: hive-operator.v2.4.3227-74831db
        version: 2.4.3227-74831db
      - name: hive-operator.v2.4.3222-d68346c
        version: 2.4.3222-d68346c
      - name: hive-operator.v2.4.3206-b0714a7
        version: 2.4.3206-b0714a7
      - name: hive-operator.v2.4.3203-672cfe7
        version: 2.4.3203-672cfe7
      - name: hive-operator.v2.4.3201-8c50440
        version: 2.4.3201-8c50440
      - name: hive-operator.v2.4.3199-cce4cb9
        version: 2.4.3199-cce4cb9
      - name: hive-operator.v2.4.3197-6c9c8e4
        version: 2.4.3197-6c9c8e4
      - name: hive-operator.v2.4.3194-f82b7d8
        version: 2.4.3194-f82b7d8
      - name: hive-operator.v2.4.3192-fae0111
        version: 2.4.3192-fae0111
      - name: hive-operator.v2.4.3189-026033b
        version: 2.4.3189-026033b
      - name: hive-operator.v2.4.3180-7299056
        version: 2.4.3180-7299056
      - name: hive-operator.v1.2.3179-0756b70e6
        version: 1.2.3179-0756b70e6
      - name: hive-operator.v1.1.16
        version: 1.1.16
      - name: hive-operator.v1.1.15
        version: 1.1.15
      - name: hive-operator.v1.1.14
        version: 1.1.14
      - name: hive-operator.v1.1.13
        version: 1.1.13
      - name: hive-operator.v1.1.12
        version: 1.1.12
      - name: hive-operator.v1.1.11
        version: 1.1.11
      - name: hive-operator.v1.1.10
        version: 1.1.10
      - name: hive-operator.v1.1.9
        version: 1.1.9
      - name: hive-operator.v1.1.8
        version: 1.1.8
      - name: hive-operator.v1.1.7
        version: 1.1.7
      - name: hive-operator.v1.1.6
        version: 1.1.6
      name: ocm-2.4
    defaultChannel: alpha
    packageName: hive-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-admin-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-admin-operator.v2024.10.2-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": null
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:ed279ff5416cce855118f66a73d55385aba26af19735eed052010ba653f64b75
          createdAt: "2024-10-25T01:47:18Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2024.10.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2023.9.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:7c4868db021beaffa762eb0de43739bbc865a4bf5900f1d17f78caed8612c073
        - registry.connect.redhat.com/ca/uma-operator@sha256:ed279ff5416cce855118f66a73d55385aba26af19735eed052010ba653f64b75
        version: 2024.10.2-12
      entries:
      - name: uma-admin-operator.v2024.10.2-12
        version: 2024.10.2-12
      name: "2024.10"
    - currentCSV: uma-admin-operator.v2025.1.3-33
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.1.1.33,990033",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mTLS": true,
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": null
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:6a932a065add9f36f9615a93037315603f2d9cd8a2e3c64b1b1fc390bb4f1dcf
          createdAt: "2025-02-18T05:34:34Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.1.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2023.9.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.1.1.26,990026"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:6a932a065add9f36f9615a93037315603f2d9cd8a2e3c64b1b1fc390bb4f1dcf
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:ae91a28d57c8a359f81757e4bf79472f2ac189ac0ae3ee0d87e8ff9c433f1ccc
        version: 2025.1.3-33
      entries:
      - name: uma-admin-operator.v2025.1.3-33
        version: 2025.1.3-33
      name: "2025.1"
    - currentCSV: uma-admin-operator.v2025.2.1-25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.2.1.25,990025",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "mixedMode": "true",
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "serverCert": "",
                    "serverCollectorCert": "",
                    "serverCollectorKey": "",
                    "serverKey": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/storage"
                      },
                      {
                        "effect": "NoSchedule",
                        "operator": "Exists"
                      },
                      {
                        "effect": "NoExecute",
                        "operator": "Exists"
                      }
                    ],
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:d201790c2d9ded6bc7a1c5f68a267574e7b5d39c39ce2f98e20ef630b0f7365f
          createdAt: "2025-03-05T05:51:02Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.2.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2023.9.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.1.1.26,990026"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               nodeAffinity: {}

               # Node labels for UMA Pods
               # Ref: https://kubernetes.io/docs/user-guide/node-selection/
               nodeSelector: {}

               # Additional tolerations for UMA pods
               # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
               # Default toleration used by UMA
               tolerations:
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/master"
                 - effect: "NoSchedule"
                   key: "node-role.kubernetes.io/storage"
                 - effect: "NoSchedule"
                   operator: "Exists"
                 - effect: "NoExecute"
                   operator: "Exists"
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:d201790c2d9ded6bc7a1c5f68a267574e7b5d39c39ce2f98e20ef630b0f7365f
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:8bcf0dac889f38a0bf77eb9babe3f726e3b2564d5988c05bdc5c2d91ae189c61
        version: 2025.2.1-25
      entries:
      - name: uma-admin-operator.v2025.2.1-25
        version: 2025.2.1-25
      name: "2025.2"
    - currentCSV: uma-admin-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:4c2a7f68a47428f3e386107254f74fa8c34e477eb40fc17c173df92f3c66c559
          createdAt: "2025-04-16T06:55:39Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.4.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.4.2.5,990005"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:4c2a7f68a47428f3e386107254f74fa8c34e477eb40fc17c173df92f3c66c559
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        version: 2025.4.5-5
      entries:
      - name: uma-admin-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-admin-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:514b953eaf0959bc120f80bdc0a2e19d8f059d3fe2883a4f6843a398470a56b1
          createdAt: "2025-07-24T14:54:21Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.5.2.12,990012"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:514b953eaf0959bc120f80bdc0a2e19d8f059d3fe2883a4f6843a398470a56b1
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        version: 2025.5.4-12
      entries:
      - name: uma-admin-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-admin-operator.v2025.7.3-21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:188a2b8d2255ded6141a4b795f1b6d7bd98017b4fd57c9c1a1d38fc701358057
          createdAt: "2025-08-23T09:00:37Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.3
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.7.1.23,990023"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:188a2b8d2255ded6141a4b795f1b6d7bd98017b4fd57c9c1a1d38fc701358057
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.3-21
      entries:
      - name: uma-admin-operator.v2025.7.3-21
        version: 2025.7.3-21
      name: "2025.7"
    - currentCSV: uma-admin-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:0cfb63c62258e54d67e7baca874003a04477f30bcf7f56861f565ad068070580
          createdAt: "2025-08-15T03:20:28Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-admin-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.8.1.9,990009"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:0cfb63c62258e54d67e7baca874003a04477f30bcf7f56861f565ad068070580
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        version: 2025.8.2-9
      entries:
      - name: uma-admin-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-admin-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
          createdAt: "2025-09-24T12:23:56Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.12,990012"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-admin-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-admin-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:7b93de818fa5160e72f878b0350d13fc1d5943eac95b0794d3d4f930cf21b59a
          createdAt: "2025-07-30T08:07:39Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-admin-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "99.99.0.serp_F152813,19"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:7b93de818fa5160e72f878b0350d13fc1d5943eac95b0794d3d4f930cf21b59a
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        version: 2025.7.2-19
      entries:
      - name: uma-admin-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-admin-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1cf3909dd2caeabc53209d6f8565dcb6c3562b9c6eb715dda8ba1953f264668e
          createdAt: "2025-09-11T11:16:06Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.8,990008"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:1cf3909dd2caeabc53209d6f8565dcb6c3562b9c6eb715dda8ba1953f264668e
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-admin-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-admin-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
          createdAt: "2025-09-24T12:23:56Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.12,990012"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-admin-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-admin-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: octavia-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: octavia-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "octavia.openstack.org/v1beta1",
                "kind": "Octavia",
                "metadata": {
                  "name": "octavia"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "octavia",
                  "databaseInstance": "openstack",
                  "lbMgmtNetwork": {
                    "availabilityZones": [
                      "zone-1"
                    ]
                  },
                  "octaviaAPI": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "octavia",
                    "databaseInstance": "openstack",
                    "persistenceDatabaseAccount": "octavia-persistence",
                    "preserveJobs": false,
                    "secret": "osp-secret",
                    "serviceAccount": "octavia",
                    "serviceUser": "octavia"
                  },
                  "octaviaHealthManager": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "octavia",
                    "databaseInstance": "openstack",
                    "networkAttachments": [
                      "octavia"
                    ],
                    "persistenceDatabaseAccount": "octavia-persistence",
                    "preserveJobs": false,
                    "role": "healthmanager",
                    "secret": "osp-secret",
                    "serviceAccount": "octavia",
                    "serviceUser": "octavia"
                  },
                  "octaviaHousekeeping": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "octavia",
                    "databaseInstance": "openstack",
                    "networkAttachments": [
                      "octavia"
                    ],
                    "persistenceDatabaseAccount": "octavia-persistence",
                    "preserveJobs": false,
                    "role": "housekeeping",
                    "secret": "osp-secret",
                    "serviceAccount": "octavia",
                    "serviceUser": "octavia"
                  },
                  "octaviaWorker": {
                    "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                    "databaseAccount": "octavia",
                    "databaseInstance": "openstack",
                    "networkAttachments": [
                      "octavia"
                    ],
                    "persistenceDatabaseAccount": "octavia-persistence",
                    "preserveJobs": false,
                    "role": "worker",
                    "secret": "osp-secret",
                    "serviceAccount": "octavia",
                    "serviceUser": "octavia"
                  },
                  "persistenceDatabaseAccount": "octavia-persistence",
                  "preserveJobs": false,
                  "rabbitMqClusterName": "rabbitmq",
                  "secret": "osp-secret",
                  "serviceUser": "octavia"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:01Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: Octavia Amphora Controller
            kind: OctaviaAmphoraController
            name: octaviaamphoracontrollers.octavia.openstack.org
            version: v1beta1
          - description: OctaviaAPI is the Schema for the octaviaapis API
            displayName: Octavia API
            kind: OctaviaAPI
            name: octaviaapis.octavia.openstack.org
            version: v1beta1
          - kind: OctaviaRsyslog
            name: octaviarsyslogs.octavia.openstack.org
            version: v1beta1
          - description: Octavia is the Schema for the octavia API
            displayName: Octavia
            kind: Octavia
            name: octavias.octavia.openstack.org
            version: v1beta1
        description: Octavia Operator
        displayName: Octavia Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Load-Balancer
        - Octavia
        links:
        - name: Octavia Operator
          url: https://github.com/openstack-k8s-operators/octavia-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/octavia-rhel9-operator@sha256:ddd280ec099cde63e38bd964d7f0608c1c34e5c9f6b140f86dde2126b0357ef5
        - registry.redhat.io/rhoso/openstack-octavia-health-manager-rhel9@sha256:7d305af6e9e6c636510186dd80037849d19bb1e6d2b2507c025d1d0d9254874a
        - registry.redhat.io/rhoso/openstack-octavia-worker-rhel9@sha256:0919ddff6421430f8ad24cb21a89c65f02f98df424267f7e436c97d555112e2a
        - registry.redhat.io/ubi9/httpd-24@sha256:dbfbe3383e12d63df7178e4cdd5b1e6e51a84c781041b9192026ea6edb72fbf3
        - registry.redhat.io/rhoso/openstack-octavia-api-rhel9@sha256:ae9aae94c2523933a346bdade16158b4302b3a0bee93dec89bae5d15606c8dfb
        - registry.redhat.io/rhoso/openstack-octavia-housekeeping-rhel9@sha256:d54868077feb205199b76e83fc9220d9b4215dbeba037cf13fb68dc6577a07dc
        version: 1.0.6
      entries:
      - name: octavia-operator.v1.0.6
        version: 1.0.6
      - name: octavia-operator.v1.0.5
        version: 1.0.5
      - name: octavia-operator.v1.0.4
        version: 1.0.4
      - name: octavia-operator.v1.0.3
        version: 1.0.3
      - name: octavia-operator.v1.0.2
        version: 1.0.2
      - name: octavia-operator.v1.0.1
        version: 1.0.1
      - name: octavia-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: octavia-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: gatekeeper-operator-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gatekeeper-operator-product.v3.11.2-0.1725401426.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel8-operator@sha256:c8894e4d02fb2c822ab2e02e489dd9b67c953d96bd3b96a8305219a2245dfd99
          createdAt: 21 Mar 2024, 01:59
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.11.0
          olm.substitutesFor: gatekeeper-operator-product.v3.11.2-0.1721233953.p
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/gatekeeper/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the Schema for the gatekeepers API
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |-
          Gatekeeper allows administrators to detect and reject non-compliant commits to an infrastructure-as-code system's source-of-truth. This strengthens compliance efforts and prevents a bad state from slowing down the organization.

          ## Support & Troubleshooting
          Gatekeeper is available and supported as part of a valid Red Hat software subscription when it is used with Red Hat Advanced Cluster Management for Kubernetes. Troubleshooting information is available in the Gatekeeper documentation.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/gatekeeper/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - registry.redhat.io/gatekeeper/gatekeeper-rhel8-operator@sha256:c8894e4d02fb2c822ab2e02e489dd9b67c953d96bd3b96a8305219a2245dfd99
        - registry.redhat.io/gatekeeper/gatekeeper-rhel8@sha256:701b165def9e4b92d91b8eac4917721e76e634d2159b8efc69bef23e10b51b9d
        version: 3.11.2+0.1725401426.p
      entries:
      - name: gatekeeper-operator-product.v3.11.2-0.1725401426.p
        version: 3.11.2+0.1725401426.p
      - name: gatekeeper-operator-product.v3.11.2-0.1721233953.p
        version: 3.11.2+0.1721233953.p
      - name: gatekeeper-operator-product.v3.11.2-0.1718224960.p
        version: 3.11.2+0.1718224960.p
      - name: gatekeeper-operator-product.v3.11.2
        version: 3.11.2
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.11"
    - currentCSV: gatekeeper-operator-product.v3.14.3-0.1746550072.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e47f0e9da800eb0cf4a0da6af98986508a35798a41581158ee7dbc977cb651dc
          createdAt: 16 Jan 2025, 06:57
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.14.3
          olm.substitutesFor: gatekeeper-operator-product.v3.14.3-0.1744033158.p
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the Schema for the gatekeepers API
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |-
          Gatekeeper allows administrators to detect and reject non-compliant commits to an infrastructure-as-code system's source-of-truth. This strengthens compliance efforts and prevents a bad state from slowing down the organization.

          ## Support & Troubleshooting
          Gatekeeper is available and supported as part of a valid Red Hat software subscription when it is used with Red Hat Advanced Cluster Management for Kubernetes. Troubleshooting information is available in the Gatekeeper documentation.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e47f0e9da800eb0cf4a0da6af98986508a35798a41581158ee7dbc977cb651dc
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:6255f7a673e6989617176842d8f0cc897e4630b5283e812a7a2d52dbb2a1f335
        version: 3.14.3+0.1746550072.p
      entries:
      - name: gatekeeper-operator-product.v3.14.3-0.1746550072.p
        version: 3.14.3+0.1746550072.p
      - name: gatekeeper-operator-product.v3.14.3-0.1744033158.p
        version: 3.14.3+0.1744033158.p
      - name: gatekeeper-operator-product.v3.14.3-0.1742934403.p
        version: 3.14.3+0.1742934403.p
      - name: gatekeeper-operator-product.v3.14.3-0.1740676608.p
        version: 3.14.3+0.1740676608.p
      - name: gatekeeper-operator-product.v3.14.3
        version: 3.14.3
      - name: gatekeeper-operator-product.v3.14.2
        version: 3.14.2
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.14"
    - currentCSV: gatekeeper-operator-product.v3.15.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:8b43248aed06563e06f7a0347ff1c45fa5892e3a9adb2b5c363962a0b152eff0
          createdAt: 13 Mar 2025, 17:07
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.15.4
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: Gatekeeper allows administrators to detect and reject non-compliant
          commits to an infrastructure-as-code system's source-of-truth. This strengthens
          compliance efforts and prevents a bad state from slowing down the organization.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:8b43248aed06563e06f7a0347ff1c45fa5892e3a9adb2b5c363962a0b152eff0
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:29690144667b3753a7e76095e136d6fc42325270e7b9f22f97dfd411439e9f54
        version: 3.15.4
      entries:
      - name: gatekeeper-operator-product.v3.15.4
        version: 3.15.4
      - name: gatekeeper-operator-product.v3.15.3
        version: 3.15.3
      - name: gatekeeper-operator-product.v3.15.2
        version: 3.15.2
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.15"
    - currentCSV: gatekeeper-operator-product.v3.17.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:3c027ec78398df0a8fc09eb79e222a9508b4f18154d1e1b477d4f156181ae51a
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.17.3
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e44078fc6073a79168ed52922b370773b451d69bf335aeb91d26a1462df21a9b
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:3c027ec78398df0a8fc09eb79e222a9508b4f18154d1e1b477d4f156181ae51a
        version: 3.17.3
      entries:
      - name: gatekeeper-operator-product.v3.17.3
        version: 3.17.3
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.17"
    - currentCSV: gatekeeper-operator-product.v3.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:0803b7219ceff7aa676488eeba58607a19c9c18a2bb8eae8b3b16cfec411cb6e
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.18.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:49726cb2702441305e8047b24054575170fb595cfe5ebf815cdbfdaa957b02f4
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:0803b7219ceff7aa676488eeba58607a19c9c18a2bb8eae8b3b16cfec411cb6e
        version: 3.18.1
      entries:
      - name: gatekeeper-operator-product.v3.18.1
        version: 3.18.1
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.18"
    - currentCSV: gatekeeper-operator-product.v3.19.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.19.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:60ee347425c6cfb36f0e1e834b33cbb3d77aa2d6c89ac513216d0b9fc55d4b52
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
        version: 3.19.1
      entries:
      - name: gatekeeper-operator-product.v3.19.1
        version: 3.19.1
      - name: gatekeeper-operator-product.v3.19.0
        version: 3.19.0
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.19"
    - currentCSV: gatekeeper-operator-product.v3.19.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.19.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:60ee347425c6cfb36f0e1e834b33cbb3d77aa2d6c89ac513216d0b9fc55d4b52
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
        version: 3.19.1
      entries:
      - name: gatekeeper-operator-product.v3.19.1
        version: 3.19.1
      - name: gatekeeper-operator-product.v3.19.0
        version: 3.19.0
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: stable
    defaultChannel: stable
    packageName: gatekeeper-operator-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://github.com/redhat-developer/app-services-operator/
    name: rhoas-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhoas-operator.0.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "ServiceRegistryConnection",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace",
                  "labels": {
                    "app.kubernetes.io/component": "external-service",
                    "app.kubernetes.io/managed-by": "rhoas"
                  }
                },
                "spec": {
                  "accessTokenSecretName": "rh-managed-services-api-accesstoken",
                  "serviceRegistryId": "exampleId",
                  "credentials": {
                    "serviceAccountSecretName": "service-account-secret"
                  }
                }
              },
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "CloudServiceAccountRequest",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace"
                },
                "spec": {
                  "serviceAccountName": "rhoas-sa",
                  "serviceAccountDescription": "Operator created service account",
                  "serviceAccountSecretName": "service-account-credentials",
                  "accessTokenSecretName": "rh-managed-services-api-accesstoken"
                }
              },
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "CloudServicesRequest",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace",
                  "labels": {
                    "app.kubernetes.io/component": "external-service",
                    "app.kubernetes.io/managed-by": "rhoas"
                  }
                },
                "spec": {
                  "accessTokenSecretName": "rh-cloud-services-api-accesstoken"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: quay.io/rhoas/service-operator:0.13.0
          createdAt: "2023-01-20 16:09:37"
          description: Red Hat OpenShift Application Services Operator
          olm.skipRange: '>=0.0.1 <0.13.0'
          repository: https://github.com/redhat-developer/app-services-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Creates RHOAS Service Accounts
            displayName: Cloud Service Account Request
            kind: CloudServiceAccountRequest
            name: cloudserviceaccountrequests.rhoas.redhat.com
            version: v1alpha1
          - description: Cloud Services Request
            displayName: Cloud Services Request
            kind: CloudServicesRequest
            name: cloudservicesrequests.rhoas.redhat.com
            version: v1alpha1
          - description: Binds Kafka Instance
            displayName: Kafka Connection
            kind: KafkaConnection
            name: kafkaconnections.rhoas.redhat.com
            version: v1alpha1
          - description: Binds Service Registry Instance
            displayName: Service Registry Connection
            kind: ServiceRegistryConnection
            name: serviceregistryconnections.rhoas.redhat.com
            version: v1alpha1
        description: "**OpenShift Application Services Operator (RHOAS-Operator)**
          creates instances of Red Hat OpenShift Application Services in your cluster.
          Operator is currently available as limited Beta and it cannot be consumed
          without invitation.\n\nRed Hat OpenShift Application Services, when added
          to Red Hat OpenShift Dedicated,  provide a streamlined developer experience
          for building, deploying,  and scaling cloud-native applications in open
          hybrid-cloud environments.   \n\nRHOAS-Operator enables developers to connect
          with their Application Services directly in their own cluster by using [RHOAS
          cli](https://github.com/redhat-developer/app-services-cli) and OpenShift
          Console UI.  \n\nRHOAS-Operator requires developers to connect with their
          Red Hat OpenShift Streams for Apache Kafka (RHOSAK) instances.  "
        displayName: OpenShift Application Services (RHOAS)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - rhoas
        - kafka
        - serviceregistry
        - app-services
        - application
        - services
        links:
        - name: RHOAS Operator Documentation
          url: https://github.com/redhat-developer/app-services-operator/tree/main/docs
        maintainers:
        - email: rhosak-eval-support@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://github.com/redhat-developer/app-services-operator/
        relatedImages:
        - quay.io/rhoas/service-operator:0.13.0
        version: 0.13.0
      entries:
      - name: rhoas-operator.0.13.0
        version: 0.13.0
      - name: rhoas-operator.0.12.0
        version: 0.12.0
      - name: rhoas-operator.0.11.0
        version: 0.11.0
      - name: rhoas-operator.0.9.6
        version: 0.9.6
      - name: rhoas-operator.0.9.5
        version: 0.9.5
      - name: rhoas-operator.0.9.3
        version: 0.9.3
      - name: rhoas-operator.0.9.2
        version: 0.9.2
      - name: rhoas-operator.0.9.1
        version: 0.9.1
      - name: rhoas-operator.0.9.0
        version: 0.9.0
      - name: rhoas-operator.0.8.2
        version: 0.8.2
      - name: rhoas-operator.0.7.10
        version: 0.7.10
      - name: rhoas-operator.0.7.9
        version: 0.7.9
      - name: rhoas-operator.0.7.8
        version: 0.7.8
      - name: rhoas-operator.0.7.6
        version: 0.7.6
      - name: rhoas-operator.0.7.2
        version: 0.7.2
      - name: rhoas-operator.0.7.1
        version: 0.7.1
      - name: rhoas-operator.0.6.8
        version: 0.6.8
      name: beta
    defaultChannel: beta
    packageName: rhoas-operator
    provider:
      name: Red Hat Inc.
      url: https://github.com/redhat-developer/app-services-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-s3-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-s3-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "s3.services.k8s.aws/v1alpha1",
                "kind": "Bucket",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/s3-controller:1.1.1
          createdAt: "2025-09-19T19:23:45Z"
          description: AWS S3 controller is a service controller for managing S3 resources
            in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Bucket represents the state of an AWS s3 Bucket resource.
            displayName: Bucket
            kind: Bucket
            name: buckets.s3.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Simple Storage Service (S3) resources in AWS from within your Kubernetes cluster.

          **About Amazon S3**

          Amazon Simple Storage Service (Amazon S3) is an object storage service that offers industry-leading scalability, data availability, security, and performance. This means customers of all sizes and industries can use it to store and protect any amount of data for a range of use cases, such as data lakes, websites, mobile applications, backup and restore, archive, enterprise applications, IoT devices, and big data analytics. Amazon S3 provides easy-to-use management features so you can organize your data and configure finely-tuned access controls to meet your specific business, organizational, and compliance requirements. Amazon S3 is designed for 99.999999999% (11 9s) of durability, and stores data for millions of applications for companies all around the world.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon S3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - s3
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon S3 Developer Resources
          url: https://aws.amazon.com/s3/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: s3 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/s3-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-s3-controller.v1.1.1
        version: 1.1.1
      - name: ack-s3-controller.v1.1.0
        version: 1.1.0
      - name: ack-s3-controller.v1.0.35
        version: 1.0.35
      - name: ack-s3-controller.v1.0.33
        version: 1.0.33
      - name: ack-s3-controller.v1.0.32
        version: 1.0.32
      - name: ack-s3-controller.v1.0.31
        version: 1.0.31
      - name: ack-s3-controller.v1.0.30
        version: 1.0.30
      - name: ack-s3-controller.v1.0.29
        version: 1.0.29
      - name: ack-s3-controller.v1.0.28
        version: 1.0.28
      - name: ack-s3-controller.v1.0.27
        version: 1.0.27
      - name: ack-s3-controller.v1.0.26
        version: 1.0.26
      - name: ack-s3-controller.v1.0.25
        version: 1.0.25
      - name: ack-s3-controller.v1.0.24
        version: 1.0.24
      - name: ack-s3-controller.v1.0.23
        version: 1.0.23
      - name: ack-s3-controller.v1.0.21
        version: 1.0.21
      - name: ack-s3-controller.v1.0.20
        version: 1.0.20
      - name: ack-s3-controller.v1.0.19
        version: 1.0.19
      - name: ack-s3-controller.v1.0.18
        version: 1.0.18
      - name: ack-s3-controller.v1.0.17
        version: 1.0.17
      - name: ack-s3-controller.v1.0.16
        version: 1.0.16
      - name: ack-s3-controller.v1.0.15
        version: 1.0.15
      - name: ack-s3-controller.v1.0.14
        version: 1.0.14
      - name: ack-s3-controller.v1.0.13
        version: 1.0.13
      - name: ack-s3-controller.v1.0.12
        version: 1.0.12
      - name: ack-s3-controller.v1.0.11
        version: 1.0.11
      - name: ack-s3-controller.v1.0.9
        version: 1.0.9
      - name: ack-s3-controller.v1.0.7
        version: 1.0.7
      - name: ack-s3-controller.v1.0.6
        version: 1.0.6
      - name: ack-s3-controller.v1.0.5
        version: 1.0.5
      - name: ack-s3-controller.v1.0.4
        version: 1.0.4
      - name: ack-s3-controller.v1.0.3
        version: 1.0.3
      - name: ack-s3-controller.v1.0.2
        version: 1.0.2
      - name: ack-s3-controller.v1.0.1
        version: 1.0.1
      - name: ack-s3-controller.v1.0.0
        version: 1.0.0
      - name: ack-s3-controller.v0.1.8
        version: 0.1.8
      - name: ack-s3-controller.v0.1.7
        version: 0.1.7
      - name: ack-s3-controller.v0.1.6
        version: 0.1.6
      - name: ack-s3-controller.v0.1.5
        version: 0.1.5
      - name: ack-s3-controller.v0.1.4
        version: 0.1.4
      - name: ack-s3-controller.v0.1.3
        version: 0.1.3
      - name: ack-s3-controller.v0.1.2
        version: 0.1.2
      - name: ack-s3-controller.v0.1.1
        version: 0.1.1
      - name: ack-s3-controller.v0.1.0
        version: 0.1.0
      - name: ack-s3-controller.v0.0.20
        version: 0.0.20
      - name: ack-s3-controller.v0.0.19
        version: 0.0.19
      - name: ack-s3-controller.v0.0.18
        version: 0.0.18
      - name: ack-s3-controller.v0.0.17
        version: 0.0.17
      - name: ack-s3-controller.v0.0.16
        version: 0.0.16
      - name: ack-s3-controller.v0.0.15
        version: 0.0.15
      - name: ack-s3-controller.v0.0.14
        version: 0.0.14
      - name: ack-s3-controller.v0.0.13
        version: 0.0.13
      - name: ack-s3-controller.v0.0.12
        version: 0.0.12
      - name: ack-s3-controller.v0.0.11
        version: 0.0.11
      - name: ack-s3-controller.v0.0.10
        version: 0.0.10
      - name: ack-s3-controller.v0.0.9
        version: 0.0.9
      - name: ack-s3-controller.v0.0.8
        version: 0.0.8
      - name: ack-s3-controller.v0.0.7
        version: 0.0.7
      name: alpha
    defaultChannel: alpha
    packageName: ack-s3-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: numaresources-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: numaresources-operator.v4.16.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodetopology.openshift.io/v1alpha1",
                "kind": "NUMAResourcesOperator",
                "metadata": {
                  "name": "numaresourcesoperator"
                },
                "spec": {
                  "nodeGroups": [
                    {
                      "machineConfigPoolSelector": {
                        "matchLabels": {
                          "pools.operator.machineconfiguration.openshift.io/worker": ""
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "nodetopology.openshift.io/v1alpha1",
                "kind": "NUMAResourcesScheduler",
                "metadata": {
                  "name": "numaresourcesscheduler"
                },
                "spec": {
                  "imageSpec": "URL_OF_SCHEDULER_IMAGE_FROM_REDHAT_REGISTRY",
                  "logLevel": "Normal",
                  "schedulerName": "topo-aware-scheduler"
                }
              }
            ]
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "true"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.15.0 <4.16.6'
          operators.openshift.io/valid-subscription: "[\n\t\"OpenShift Kubernetes
            Engine\",\n\t\"OpenShift Container Platform\",\n\t\"OpenShift Platform
            Plus\"\n]"
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NUMAResourcesOperator is the Schema for the numaresourcesoperators
              API
            displayName: NUMA Resources Operator
            kind: NUMAResourcesOperator
            name: numaresourcesoperators.nodetopology.openshift.io
            version: v1
          - description: NUMAResourcesOperator is the Schema for the numaresourcesoperators
              API
            displayName: NUMA Resources Operator
            kind: NUMAResourcesOperator
            name: numaresourcesoperators.nodetopology.openshift.io
            version: v1alpha1
          - description: NUMAResourcesScheduler is the Schema for the numaresourcesschedulers
              API
            displayName: NUMA Aware Scheduler
            kind: NUMAResourcesScheduler
            name: numaresourcesschedulers.nodetopology.openshift.io
            version: v1
          - description: NUMAResourcesScheduler is the Schema for the numaresourcesschedulers
              API
            displayName: NUMA Aware Scheduler
            kind: NUMAResourcesScheduler
            name: numaresourcesschedulers.nodetopology.openshift.io
            version: v1alpha1
        description: NUMA resources exporter operator
        displayName: numaresources-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - RTE
        - NUMA
        links:
        - name: Numaresources Operator
          url: https://github.com/openshift-kni/numaresources-operator
        maintainers:
        - email: fromani@redhat.com
          name: fromani
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/numaresources-rhel9-operator@sha256:aca8232e89172ca9fd9ce4a66cae9e8bb027b6d16482ce6b8e27fb7779c398e2
        version: 4.16.6
      entries:
      - name: numaresources-operator.v4.16.6
        version: 4.16.6
      - name: numaresources-operator.v4.16.5
        version: 4.16.5
      - name: numaresources-operator.v4.16.4
        version: 4.16.4
      - name: numaresources-operator.v4.16.3
        version: 4.16.3
      - name: numaresources-operator.v4.16.2
        version: 4.16.2
      - name: numaresources-operator.v4.16.1
        version: 4.16.1
      - name: numaresources-operator.v4.16.0
        version: 4.16.0
      name: "4.16"
    defaultChannel: "4.16"
    packageName: numaresources-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.9-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lvms-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lvms-operator.v4.15.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "my-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "default": true,
                        "fstype": "xfs",
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:696edeb4c168646b49b8b0ca5b3d06071ab66fdb010e7cd46888d8b2f3a0c7a8
          description: Logical volume manager storage provides dynamically provisioned
            local storage for container workloads
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.11'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "test-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
          operatorframework.io/suggested-namespace: openshift-storage
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-storage",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/internal-objects: '["logicalvolumes.topolvm.io",
            "lvmvolumegroups.lvm.topolvm.io", "lvmvolumegroupnodestatuses.lvm.topolvm.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/lvm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LogicalVolume
            name: logicalvolumes.topolvm.io
            version: v1
          - description: LVMCluster is the Schema for the lvmclusters API
            displayName: LVMCluster
            kind: LVMCluster
            name: lvmclusters.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroupNodeStatus
            name: lvmvolumegroupnodestatuses.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroup
            name: lvmvolumegroups.lvm.topolvm.io
            version: v1alpha1
        description: Logical volume manager storage provides dynamically provisioned
          local storage.
        displayName: LVM Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - local storage
        - operator
        - LVM
        links:
        - name: Source Repository
          url: https://github.com/openshift/lvm-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:771d513c4d4e5d77d797b4c18f86e23162a20c6d39955e93d24f62b94e62e785
        - registry.redhat.io/lvms4/lvms-must-gather-rhel9@sha256:382ceea1151da56d0931403207c9f85721d58a66b39b1cd79790eca6219807d0
        - registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:696edeb4c168646b49b8b0ca5b3d06071ab66fdb010e7cd46888d8b2f3a0c7a8
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:82c01635b32821f04a7cfcf80da8b3f7e097786f220e0c85aab47bda4536a1d6
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:2db4f32824a8f9f6780dc3d236f73684cb20db41466ffa8f9a89fd3f5eb62fdd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:7ce50398cd29df6e07ccd89790a53a711a7e5b484698c48f6581f5d4e9efb05a
        - registry.redhat.io/lvms4/topolvm-rhel9@sha256:074a206d36c9e4198d90baed86f3046abd2ec12f9b36a715b60900011366b3d4
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:05dceaff7f0b2b769ddfd9261cc66d42aab9158ce99ed3dbd6c229e784072635
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:479454d0d4de6609d8a4696b9c7f0097115afb45db3bdd46631d58afd77dcc58
        version: 4.15.11
      entries:
      - name: lvms-operator.v4.15.11
        version: 4.15.11
      - name: lvms-operator.v4.15.10
        version: 4.15.10
      - name: lvms-operator.v4.15.9
        version: 4.15.9
      - name: lvms-operator.v4.15.8
        version: 4.15.8
      - name: lvms-operator.v4.15.7
        version: 4.15.7
      - name: lvms-operator.v4.15.6
        version: 4.15.6
      - name: lvms-operator.v4.15.5
        version: 4.15.5
      - name: lvms-operator.v4.15.4
        version: 4.15.4
      - name: lvms-operator.v4.15.3
        version: 4.15.3
      - name: lvms-operator.v4.15.2
        version: 4.15.2
      - name: lvms-operator.v4.15.1
        version: 4.15.1
      name: stable-4.15
    - currentCSV: lvms-operator.v4.16.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "my-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "default": true,
                        "fstype": "xfs",
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:b9d80bb263a54192501d5ba9366a4f0c19b34875b51bad4fa75ef22c60bb3ee7
          description: Logical volume manager storage provides dynamically provisioned
            local storage for container workloads
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.9'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "test-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
          operatorframework.io/suggested-namespace: openshift-storage
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-storage",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["logicalvolumes.topolvm.io",
            "lvmvolumegroups.lvm.topolvm.io", "lvmvolumegroupnodestatuses.lvm.topolvm.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lvm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LogicalVolume
            name: logicalvolumes.topolvm.io
            version: v1
          - description: LVMCluster is the Schema for the lvmclusters API
            displayName: LVMCluster
            kind: LVMCluster
            name: lvmclusters.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroupNodeStatus
            name: lvmvolumegroupnodestatuses.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroup
            name: lvmvolumegroups.lvm.topolvm.io
            version: v1alpha1
        description: Logical volume manager storage provides dynamically provisioned
          local storage.
        displayName: LVM Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - local storage
        - operator
        - LVM
        links:
        - name: Source Repository
          url: https://github.com/openshift/lvm-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/lvms4/topolvm-rhel9@sha256:227748103049d2cced921ba9fc4e2d3671f08c897cc2a5627d755fe5f9d2767d
        - registry.redhat.io/lvms4/lvms-must-gather-rhel9@sha256:20b28b87b3fe6d6bf739181376231d6f8b7454c34b59eaf2e06c04a9fca8f41d
        - registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:b9d80bb263a54192501d5ba9366a4f0c19b34875b51bad4fa75ef22c60bb3ee7
        version: 4.16.9
      entries:
      - name: lvms-operator.v4.16.9
        version: 4.16.9
      - name: lvms-operator.v4.16.8
        version: 4.16.8
      - name: lvms-operator.v4.16.7
        version: 4.16.7
      - name: lvms-operator.v4.16.6
        version: 4.16.6
      - name: lvms-operator.v4.16.5
        version: 4.16.5
      - name: lvms-operator.v4.16.4
        version: 4.16.4
      - name: lvms-operator.v4.16.3
        version: 4.16.3
      - name: lvms-operator.v4.16.2
        version: 4.16.2
      - name: lvms-operator.v4.16.1
        version: 4.16.1
      - name: lvms-operator.v4.16.0
        version: 4.16.0
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: lvms-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-custom-metrics-autoscaler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: custom-metrics-autoscaler.v2.17.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eventing.keda.sh/v1alpha1",
                "kind": "CloudEventSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keda",
                    "app.kubernetes.io/instance": "cloudeventsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cloudeventsource",
                    "app.kubernetes.io/part-of": "keda"
                  },
                  "name": "cloudeventsource-sample"
                },
                "spec": {
                  "clusterName": "cluster-sample",
                  "destination": {
                    "http": {
                      "uri": "http://foo.bar"
                    }
                  }
                }
              },
              {
                "apiVersion": "eventing.keda.sh/v1alpha1",
                "kind": "ClusterCloudEventSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keda",
                    "app.kubernetes.io/instance": "clustercloudeventsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clustercloudeventsource",
                    "app.kubernetes.io/part-of": "keda"
                  },
                  "name": "clustercloudeventsource-sample"
                },
                "spec": {
                  "clusterName": "clustercluster-sample",
                  "destination": {
                    "http": {
                      "uri": "http://foo.bar"
                    }
                  }
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ClusterTriggerAuthentication",
                "metadata": {
                  "name": "example-clustertriggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "example-secret-parameter"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "KedaController",
                "metadata": {
                  "name": "keda",
                  "namespace": "openshift-keda"
                },
                "spec": {
                  "admissionWebhooks": {
                    "logEncoder": "console",
                    "logLevel": "info"
                  },
                  "metricsServer": {
                    "logLevel": "0"
                  },
                  "operator": {
                    "logEncoder": "console",
                    "logLevel": "info"
                  },
                  "serviceAccount": null,
                  "watchNamespace": ""
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledJob",
                "metadata": {
                  "name": "scaledjob-sample"
                },
                "spec": {
                  "failedJobsHistoryLimit": 5,
                  "jobTargetRef": {
                    "activeDeadlineSeconds": 600,
                    "backoffLimit": 6,
                    "completions": 1,
                    "parallelism": 1,
                    "template": null
                  },
                  "maxReplicaCount": 100,
                  "pollingInterval": 30,
                  "successfulJobsHistoryLimit": 5,
                  "triggers": [
                    {
                      "metadata": {
                        "property": "example-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledObject",
                "metadata": {
                  "name": "example-scaledobject"
                },
                "spec": {
                  "cooldownPeriod": 300,
                  "maxReplicaCount": 100,
                  "minReplicaCount": 0,
                  "pollingInterval": 30,
                  "scaleTargetRef": {
                    "name": "example-deployment"
                  },
                  "triggers": [
                    {
                      "metadata": {
                        "property": "example-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "TriggerAuthentication",
                "metadata": {
                  "name": "example-triggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "example-secret-parameter"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Cloud Provider
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9-operator@sha256:9a359ae6612d2f22fe983653418143873cb9939f83dca6f008ec3cb8f3d53990
          createdAt: 20 Aug 2025, 21:31
          description: Custom Metrics Autoscaler Operator, an event-driven autoscaler
            based upon KEDA
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.1 <2.17.2-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-keda
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/custom-metrics-autoscaler-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudEventSource
            name: cloudeventsources.eventing.keda.sh
            version: v1alpha1
          - kind: ClusterCloudEventSource
            name: clustercloudeventsources.eventing.keda.sh
            version: v1alpha1
          - description: Describes cluster wide authentication parameters
            displayName: ClusterTriggerAuthentication
            kind: ClusterTriggerAuthentication
            name: clustertriggerauthentications.keda.sh
            version: v1alpha1
          - description: |
              Represents an installation of a particular version of KEDA Controller.
            displayName: KedaController
            kind: KedaController
            name: kedacontrollers.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your jobs and what the triggers
            displayName: ScaledJob
            kind: ScaledJob
            name: scaledjobs.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your application and what the
              triggers are
            displayName: ScaledObject
            kind: ScaledObject
            name: scaledobjects.keda.sh
            version: v1alpha1
          - description: Describes authentication parameters
            displayName: TriggerAuthentication
            kind: TriggerAuthentication
            name: triggerauthentications.keda.sh
            version: v1alpha1
        description: |
          ## About the managed application
          Custom Metrics Autoscaler for OpenShift is an event driven autoscaler based upon KEDA.  Custom Metrics Autoscaler can monitor event sources like Kafka, RabbitMQ, or cloud event sources and feed the metrics from those sources into the Kubernetes horizontal pod autoscaler.  With Custom Metrics Autoscaler, you can have event driven and serverless scale of deployments within any Kubernetes cluster.
          ## About this Operator
          The Custom Metrics Autoscaler Operator deploys and manages installation of KEDA Controller in the cluster. Install this operator and follow installation instructions on how to install Custom Metrics Autoscaler in your cluster.

          ## Prerequisites for enabling this Operator
          ## How to install Custom Metrics Autoscaler in the cluster
          The installation of Custom Metrics Autoscaler is triggered by the creation of `KedaController` resource. Please refer to the [KedaController Spec](https://github.com/openshift/custom-metrics-autoscaler-operator/blob/main/README.md#the-kedacontroller-custom-resource) for more details on available options.

          Only a resource named `keda` in the namespace where the operator is installed will trigger the installation, reconfiguration or removal of the KEDA Controller resource.

          There should be only one KEDA Controller in the cluster.

          The operator will behave in this manner whether it is installed with the `AllNamespaces` or `OwnNamespace` install mode. While the operator more closely matches the `OwnNamespace` semantics, `AllNamespaces` is a supported installation mode to allow it to be installed to namespaces with existing `OperatorGroups` which require that installation mode, such as `openshift-operators`.
        displayName: Custom Metrics Autoscaler
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keda
        - scaling
        - event
        links:
        - name: Custom Metrics Autoscaler Documentation
          url: https://docs.openshift.com/container-platform/latest/nodes/cma/nodes-cma-autoscaling-custom.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9-operator@sha256:9a359ae6612d2f22fe983653418143873cb9939f83dca6f008ec3cb8f3d53990
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9@sha256:ace9358fbd2d81a6e16bfa049c763ca16acf6cc2a59cdc0b6282e5e0dfc902cc
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-adapter-rhel9@sha256:0c8ae9323f927346e168ba829f6c7679af6b2dd359c8d0a020dd0ae3954481ce
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-admission-webhooks-rhel9@sha256:75631a21de057844e9ae45ae2496a22e16f3a32a6fd5f6cb5f90e173e3be2180
        version: 2.17.2-1
      entries:
      - name: custom-metrics-autoscaler.v2.17.2-1
        version: 2.17.2-1
      - name: custom-metrics-autoscaler.v2.15.1-6
        version: 2.15.1-6
      - name: custom-metrics-autoscaler.v2.15.1-4
        version: 2.15.1-4
      - name: custom-metrics-autoscaler.v2.14.1-467
        version: 2.14.1-467
      - name: custom-metrics-autoscaler.v2.14.1-454
        version: 2.14.1-454
      - name: custom-metrics-autoscaler.v2.13.1-427
        version: 2.13.1-427
      - name: custom-metrics-autoscaler.v2.12.1-394
        version: 2.12.1-394
      - name: custom-metrics-autoscaler.v2.11.2-322
        version: 2.11.2-322
      - name: custom-metrics-autoscaler.v2.11.2-311
        version: 2.11.2-311
      - name: custom-metrics-autoscaler.v2.10.1-267
        version: 2.10.1-267
      name: stable
    defaultChannel: stable
    packageName: openshift-custom-metrics-autoscaler-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com
    name: aws-load-balancer-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-load-balancer-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
          createdAt: 2022-06-28T17:56:41
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:d7bc364512178c36671d8a4b5a76cf7cb10f8e56997106187b0fe1f032670ece
        version: 0.0.1
      entries:
      - name: aws-load-balancer-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: aws-load-balancer-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
          createdAt: 2022-12-09T15:32:32
          olm.skipRange: <0.2.0
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:cfd210a2cb705156883fda126d052e05eb84345cc27ba04688067f8ed9a802ea
        version: 0.2.0
      entries:
      - name: aws-load-balancer-operator.v0.2.0
        version: 0.2.0
      name: stable-v0
    - currentCSV: aws-load-balancer-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
          createdAt: 2022-06-28T17:56:41
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:d7bc364512178c36671d8a4b5a76cf7cb10f8e56997106187b0fe1f032670ece
        version: 0.0.1
      entries:
      - name: aws-load-balancer-operator.v0.0.1
        version: 0.0.1
      name: stable-v0.1
    - currentCSV: aws-load-balancer-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
          createdAt: 2022-12-09T15:32:32
          olm.skipRange: <0.2.0
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:cfd210a2cb705156883fda126d052e05eb84345cc27ba04688067f8ed9a802ea
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        version: 0.2.0
      entries:
      - name: aws-load-balancer-operator.v0.2.0
        version: 0.2.0
      name: stable-v0.2
    - currentCSV: aws-load-balancer-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
          createdAt: 2024-01-23T11:21:00
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:2e0b9332a44d8d9c23e19c7accab0813a651f39210257820db508cac28876595
        version: 1.1.1
      entries:
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1
    - currentCSV: aws-load-balancer-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:6e15df39a3d25755a79888f739e1af8b052c3fba57bf8f6b3434ef58fd4288cf
          createdAt: 2024-01-24T09:13:21
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat <<EOF | oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
          EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note about UPI:__ For UPI based installs, additional documentation can be found in
          [VPC and subnets](https://github.com/openshift/aws-load-balancer-operator/blob/release-1.0/docs/prerequisites.md#vpc-and-subnets)

          __Note about STS:__ For clusters using STS, use the [STS documentation](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#for-sts-clusters) instead.
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9f9f8f69ab98c1f25cc301572339e0cc5c5b76f2e29ed26e10ac91ad20e0090c
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:6e15df39a3d25755a79888f739e1af8b052c3fba57bf8f6b3434ef58fd4288cf
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:73db97d20ee9415e9a8941cc6a36ed81d901d760d9949e95e10c90866ef8bf64
        version: 1.0.1
      entries:
      - name: aws-load-balancer-operator.v1.0.1
        version: 1.0.1
      - name: aws-load-balancer-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: aws-load-balancer-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
          createdAt: 2024-01-23T11:21:00
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:2e0b9332a44d8d9c23e19c7accab0813a651f39210257820db508cac28876595
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
        version: 1.1.1
      entries:
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    defaultChannel: stable-v1
    packageName: aws-load-balancer-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres-community
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.17.0-rc1
        version: 1.17.0-rc1
      - name: stackgres-community.v1.16.3
        version: 1.16.3
      - name: stackgres-community.v1.16.2
        version: 1.16.2
      - name: stackgres-community.v1.16.1
        version: 1.16.1
      - name: stackgres-community.v1.16.0
        version: 1.16.0
      - name: stackgres-community.v1.16.0-rc3
        version: 1.16.0-rc3
      - name: stackgres-community.v1.16.0-rc2
        version: 1.16.0-rc2
      - name: stackgres-community.v1.16.0-rc1
        version: 1.16.0-rc1
      - name: stackgres-community.v1.15.2
        version: 1.15.2
      - name: stackgres-community.v1.15.1
        version: 1.15.1
      - name: stackgres-community.v1.15.0
        version: 1.15.0
      - name: stackgres-community.v1.15.0-rc4
        version: 1.15.0-rc4
      - name: stackgres-community.v1.15.0-rc3
        version: 1.15.0-rc3
      - name: stackgres-community.v1.15.0-rc2
        version: 1.15.0-rc2
      - name: stackgres-community.v1.15.0-rc1
        version: 1.15.0-rc1
      - name: stackgres-community.v1.14.3
        version: 1.14.3
      - name: stackgres-community.v1.14.2
        version: 1.14.2
      - name: stackgres-community.v1.14.1
        version: 1.14.1
      - name: stackgres-community.v1.14.0
        version: 1.14.0
      - name: stackgres-community.v1.14.0-rc2
        version: 1.14.0-rc2
      - name: stackgres-community.v1.14.0-rc1
        version: 1.14.0-rc1
      - name: stackgres-community.v1.13.0
        version: 1.13.0
      - name: stackgres-community.v1.13.0-rc1
        version: 1.13.0-rc1
      - name: stackgres-community.v1.12.0
        version: 1.12.0
      - name: stackgres-community.v1.12.0-rc1
        version: 1.12.0-rc1
      name: candidate
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.17.0-rc1
        version: 1.17.0-rc1
      - name: stackgres-community.v1.16.3
        version: 1.16.3
      - name: stackgres-community.v1.16.2
        version: 1.16.2
      - name: stackgres-community.v1.16.1
        version: 1.16.1
      - name: stackgres-community.v1.16.0
        version: 1.16.0
      - name: stackgres-community.v1.16.0-rc3
        version: 1.16.0-rc3
      - name: stackgres-community.v1.16.0-rc2
        version: 1.16.0-rc2
      - name: stackgres-community.v1.16.0-rc1
        version: 1.16.0-rc1
      - name: stackgres-community.v1.15.2
        version: 1.15.2
      - name: stackgres-community.v1.15.1
        version: 1.15.1
      - name: stackgres-community.v1.15.0
        version: 1.15.0
      - name: stackgres-community.v1.15.0-rc4
        version: 1.15.0-rc4
      - name: stackgres-community.v1.15.0-rc3
        version: 1.15.0-rc3
      - name: stackgres-community.v1.15.0-rc2
        version: 1.15.0-rc2
      - name: stackgres-community.v1.15.0-rc1
        version: 1.15.0-rc1
      - name: stackgres-community.v1.14.3
        version: 1.14.3
      - name: stackgres-community.v1.14.2
        version: 1.14.2
      - name: stackgres-community.v1.14.1
        version: 1.14.1
      - name: stackgres-community.v1.14.0
        version: 1.14.0
      - name: stackgres-community.v1.14.0-rc2
        version: 1.14.0-rc2
      - name: stackgres-community.v1.14.0-rc1
        version: 1.14.0-rc1
      - name: stackgres-community.v1.13.0
        version: 1.13.0
      - name: stackgres-community.v1.13.0-rc1
        version: 1.13.0-rc1
      - name: stackgres-community.v1.12.0
        version: 1.12.0
      - name: stackgres-community.v1.12.0-rc1
        version: 1.12.0-rc1
      name: fast
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.16.3
        version: 1.16.3
      - name: stackgres-community.v1.16.2
        version: 1.16.2
      - name: stackgres-community.v1.16.1
        version: 1.16.1
      - name: stackgres-community.v1.16.0
        version: 1.16.0
      - name: stackgres-community.v1.15.2
        version: 1.15.2
      - name: stackgres-community.v1.15.1
        version: 1.15.1
      - name: stackgres-community.v1.15.0
        version: 1.15.0
      - name: stackgres-community.v1.14.3
        version: 1.14.3
      - name: stackgres-community.v1.14.2
        version: 1.14.2
      - name: stackgres-community.v1.14.1
        version: 1.14.1
      - name: stackgres-community.v1.14.0
        version: 1.14.0
      - name: stackgres-community.v1.13.0
        version: 1.13.0
      - name: stackgres-community.v1.12.0
        version: 1.12.0
      name: stable
    defaultChannel: stable
    packageName: stackgres-community
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Rebaca Technologies Pvt Ltd
      provider-url: https://www.rebaca.com/telecom/solutions
    name: abot-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: abot-operator-rhmp.3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "abot.com/v3",
                "kind": "Abot",
                "metadata": {
                  "name": "abot"
                },
                "spec": {
                  "abotanalytics": {
                    "image": "registry.connect.redhat.com/rebaca/abot-analytics@sha256:c854d9eafbe26f834e10f5a97a587b2a2bf05558554897817bb8e2556124b75d",
                    "resources": {
                      "limits": {
                        "cpu": "4000m",
                        "ephemeral-storage": "8Gi",
                        "memory": "6000Mi"
                      },
                      "requests": {
                        "cpu": "2000m",
                        "ephemeral-storage": "4Gi",
                        "memory": "4000Mi"
                      }
                    }
                  },
                  "abotanalyticsbase": {
                    "image": "registry.connect.redhat.com/rebaca/abot-analyticsbase@sha256:dcc2a417a889acd6ec63538a6dfa393d4c1942dac178a04e5d65189a1ae054dc",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotbaseimage": "registry.connect.redhat.com/rebaca/abot-baseimage@sha256:2cbe38298d26be5896e0dc071aae111dd921d0dbd05c3525290f6ff31c73e926",
                  "abotbdd": {
                    "image": "registry.connect.redhat.com/rebaca/abot-bdd@sha256:d5c6bfe12a548cbdb4e0d83d5f18e5ab0c17c3af5340e437b2d0c23740c9c9e6",
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "ephemeral-storage": "8Gi",
                        "memory": "2000Mi"
                      },
                      "requests": {
                        "cpu": "1000m",
                        "ephemeral-storage": "4Gi",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotgrafana": {
                    "image": "registry.redhat.io/rhel8/grafana@sha256:d894cd34f2d8cac30e0fab69bd869f4b0f258b5086a8fd7ead0c5886ce2c5f7a",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "1500Mi"
                      },
                      "requests": {
                        "cpu": "300m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotgui": {
                    "image": "registry.connect.redhat.com/rebaca/abot-gui@sha256:78b512d6bd223a1f5d35f4cbd221f6637e2c9267726ce8899ee5d86aaac82b6f",
                    "resources": {
                      "limits": {
                        "cpu": "4000m",
                        "ephemeral-storage": "8Gi",
                        "memory": "6000Mi"
                      },
                      "requests": {
                        "cpu": "2000m",
                        "ephemeral-storage": "4Gi",
                        "memory": "4000Mi"
                      }
                    }
                  },
                  "abotkibana": {
                    "image": "registry.connect.redhat.com/elastic/kibana@sha256:2be3542d50f51ad786ead958a399f26852339e9417279688aad194c6794336ce",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "abotlogstash": {
                    "image": "registry.connect.redhat.com/elastic/logstash@sha256:dbc25dc5c31dfee08e86221af9f549b50075b2ccf7f255779c5f022c54cd892e",
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "2000Mi"
                      }
                    }
                  },
                  "abotmongo": {
                    "image": "registry.connect.redhat.com/rebaca/abot-mongodb@sha256:be9059e7e8bc1636a84575416b823c20463228785598662ee81c27e4c3c987dc",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "abotnginx": {
                    "image": "registry.connect.redhat.com/rebaca/abot-nginx@sha256:3edc22b40cd30230768115efdceec3d054d71f78214b4100b7425d8484154378",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "abotrabbitmq": {
                    "image": "registry.connect.redhat.com/rebaca/abot-rabbitmq@sha256:e609bec067fe7740590bb272db27d1a7770db47a0fc802e2288c022e101c9e25",
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "abotsidecar": {
                    "alertmanagerimage": "registry.redhat.io/rhacm2/prometheus-alertmanager-rhel8@sha256:ecaee8ad24a9c7cb3b945ee6d2c1275b9de2a676e1868abc2687e687385b3f6a",
                    "kubestatemetricsimage": "registry.redhat.io/rhacm2/kube-state-metrics-rhel8@sha256:e6bb6fa0c155bebf64749368dd951c15f1dbfa3559e2ba6734d3661d3a95c00c",
                    "prometheusimage": "registry.connect.redhat.com/kasten/prometheus@sha256:1e490fcccdd50c1026150ba4cad7c29297d2b73b3d5c80240ff31509e666363d",
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "8000Mi"
                      },
                      "requests": {
                        "cpu": "1000m",
                        "memory": "6000Mi"
                      }
                    }
                  },
                  "elasticsearch": {
                    "image": "registry.connect.redhat.com/elastic/elasticsearch@sha256:f47fd7df80ab05635e31d733f6969d38205f9f81cf95da8594a047f3d07c59ae",
                    "replicas": 3,
                    "resources": {
                      "limits": {
                        "cpu": "1200m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "800m",
                        "memory": "1000Mi"
                      }
                    },
                    "volumeClaim": {
                      "storage": "50Gi"
                    }
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "initContainers": {
                    "image": "registry.access.redhat.com/ubi8/ubi@sha256:a965f33ee4ee57dc8e40a1f9350ddf28ed0727b6cf80db46cdad0486a7580f9d",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "protocol": "https",
                  "pvc": {
                    "abotvol": "100Gi",
                    "analyticsvol": "100Gi",
                    "databak": "10Gi",
                    "mongovol": "10Gi",
                    "prometheusvol": "10Gi"
                  },
                  "replicas": 1,
                  "security": {
                    "fsGroup": 1018450000,
                    "seLinuxOptions": {
                      "level": "s0:c123,c456"
                    }
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "name": "abot"
                  },
                  "storageclass": "ocs-storagecluster-cephfs"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/rebaca/abot-operator@sha256:d3e8e1b6988827b3c636bdfcb57cda26870089d90cf2a732be522f5474c0a734
          createdAt: "2024-10-03T20:32:24Z"
          description: The ABot Test Bench is a test automation tool for 4G, 5G and
            ORAN Telecom networks that follow specifications and use standards driven
            design and compliance as specified by the 3GPP consortium.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/abot-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/abot-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=1.0.0 <3.0.0'
          operatorframework.io/suggested-namespace: abot
          operators.openshift.io/valid-subscription: '["ABot Basic License for 15
            days"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: https://abot.rebaca.com/support/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Abot
            name: abots.abot.com
            version: v3
        description: A Helm based operator for deploying Abot application in Openshift
          cluster. The ABot Test Bench is a test automation tool for 4G, 5G and ORAN
          Telecom networks that follow specifications and use standards driven design
          and compliance as specified by the 3GPP consortium. It provides a platform
          for a user to quickly check for protocol compliance and test for load characteristics
          of a particular system under test.
        displayName: Abot Operator-v3.0.0
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 5G
        - 4G
        - ORAN
        - test tool
        - compliance
        - interop
        links:
        - name: Abot Operator
          url: https://abot-operator.domain
        maintainers:
        - email: suchismita.nag@rebaca.com
          name: Suchismita Nag
        - email: samir@rebaca.com
          name: Samir Chatterjee
        maturity: alpha
        provider:
          name: Rebaca Technologies Pvt Ltd
          url: https://www.rebaca.com/telecom/solutions
        relatedImages:
        - registry.connect.redhat.com/rebaca/abot-rabbitmq@sha256:e609bec067fe7740590bb272db27d1a7770db47a0fc802e2288c022e101c9e25
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel8@sha256:e6bb6fa0c155bebf64749368dd951c15f1dbfa3559e2ba6734d3661d3a95c00c
        - registry.connect.redhat.com/rebaca/abot-bdd@sha256:d5c6bfe12a548cbdb4e0d83d5f18e5ab0c17c3af5340e437b2d0c23740c9c9e6
        - registry.connect.redhat.com/rebaca/abot-baseimage@sha256:2cbe38298d26be5896e0dc071aae111dd921d0dbd05c3525290f6ff31c73e926
        - registry.connect.redhat.com/elastic/elasticsearch@sha256:f47fd7df80ab05635e31d733f6969d38205f9f81cf95da8594a047f3d07c59ae
        - registry.connect.redhat.com/rebaca/abot-analyticsbase@sha256:dcc2a417a889acd6ec63538a6dfa393d4c1942dac178a04e5d65189a1ae054dc
        - registry.connect.redhat.com/rebaca/abot-mongodb@sha256:be9059e7e8bc1636a84575416b823c20463228785598662ee81c27e4c3c987dc
        - registry.connect.redhat.com/rebaca/abot-nginx@sha256:3edc22b40cd30230768115efdceec3d054d71f78214b4100b7425d8484154378
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e34f912158985525f6efb3c4b675f20854fc677c867d12fe3cd2e971309428dc
        - registry.connect.redhat.com/rebaca/abot-operator@sha256:d3e8e1b6988827b3c636bdfcb57cda26870089d90cf2a732be522f5474c0a734
        - registry.redhat.io/rhel8/grafana@sha256:d894cd34f2d8cac30e0fab69bd869f4b0f258b5086a8fd7ead0c5886ce2c5f7a
        - registry.connect.redhat.com/elastic/logstash@sha256:dbc25dc5c31dfee08e86221af9f549b50075b2ccf7f255779c5f022c54cd892e
        - registry.connect.redhat.com/kasten/prometheus@sha256:1e490fcccdd50c1026150ba4cad7c29297d2b73b3d5c80240ff31509e666363d
        - registry.connect.redhat.com/rebaca/abot-analytics@sha256:c854d9eafbe26f834e10f5a97a587b2a2bf05558554897817bb8e2556124b75d
        - registry.connect.redhat.com/rebaca/abot-gui@sha256:78b512d6bd223a1f5d35f4cbd221f6637e2c9267726ce8899ee5d86aaac82b6f
        - registry.connect.redhat.com/elastic/kibana@sha256:2be3542d50f51ad786ead958a399f26852339e9417279688aad194c6794336ce
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel8@sha256:ecaee8ad24a9c7cb3b945ee6d2c1275b9de2a676e1868abc2687e687385b3f6a
        - registry.access.redhat.com/ubi8/ubi@sha256:a965f33ee4ee57dc8e40a1f9350ddf28ed0727b6cf80db46cdad0486a7580f9d
        version: 3.0.0
      entries:
      - name: abot-operator-rhmp.3.0.0
        version: 3.0.0
      - name: abot-operator.2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: abot-operator-rhmp
    provider:
      name: Rebaca Technologies Pvt Ltd
      url: https://www.rebaca.com/telecom/solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: secrets-store-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secrets-store-csi-driver-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9-operator@sha256:96f3a8df9ac5ff3828b0add052af00f7d9df0ebb622a6adfd9772658d55baeb8
          createdAt: "2023-06-12T00:00:00Z"
          description: Install and configure Secrets Store CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.13.0-0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/secrets-store-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SecretProviderClass is the Schema for the secretproviderclasses
              API
            displayName: Secret Provider Class
            kind: SecretProviderClass
            name: secretproviderclasses.secrets-store.csi.x-k8s.io
            version: v1
          - description: SecretProviderClassPodStatus is the Schema for the secretproviderclassespodstatus
              API
            displayName: Secret Provider Class Pod Status
            kind: SecretProviderClassPodStatus
            name: secretproviderclasspodstatuses.secrets-store.csi.x-k8s.io
            version: v1
        description: |
          Operator that installs and configures the CSI driver for Secrets Store. The Secrets Store CSI Driver Operator is a Technology Preview feature. For more information on Red Hat Technology Preview features support scope, see https://access.redhat.com/support/offerings/techpreview/.
        displayName: Secrets Store CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: Documentation
          url: https://github.com/openshift/secrets-store-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/secrets-store-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: preview
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0627c09fe3f5a015f117fb30af1f4d17340aefa48b23994737d3901a804b63d3
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:a5704d26d0b562dba9633d1c107f34e05c750c0e3a64dbcace4dab7b964fff8d
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9-operator@sha256:96f3a8df9ac5ff3828b0add052af00f7d9df0ebb622a6adfd9772658d55baeb8
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9@sha256:148672ccd882a43c5c1d52763da465558b5ed54310418ec40c7ca0b56e408b1a
        version: 4.16.0-202509240123
      entries:
      - name: secrets-store-csi-driver-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: secrets-store-csi-driver-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: secrets-store-csi-driver-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: secrets-store-csi-driver-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: secrets-store-csi-driver-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: secrets-store-csi-driver-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: secrets-store-csi-driver-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: secrets-store-csi-driver-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: secrets-store-csi-driver-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: secrets-store-csi-driver-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: secrets-store-csi-driver-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: secrets-store-csi-driver-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: secrets-store-csi-driver-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: secrets-store-csi-driver-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: secrets-store-csi-driver-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: secrets-store-csi-driver-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: secrets-store-csi-driver-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: secrets-store-csi-driver-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: secrets-store-csi-driver-operator.v4.16.0-202410031007
        version: 4.16.0-202410031007
      - name: secrets-store-csi-driver-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: secrets-store-csi-driver-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: secrets-store-csi-driver-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: secrets-store-csi-driver-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: secrets-store-csi-driver-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: secrets-store-csi-driver-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: secrets-store-csi-driver-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: secrets-store-csi-driver-operator.v4.16.0-202408021139
        version: 4.16.0-202408021139
      - name: secrets-store-csi-driver-operator.v4.16.0-202407190036
        version: 4.16.0-202407190036
      - name: secrets-store-csi-driver-operator.v4.16.0-202407111006
        version: 4.16.0-202407111006
      - name: secrets-store-csi-driver-operator.v4.16.0-202407081937
        version: 4.16.0-202407081937
      - name: secrets-store-csi-driver-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: secrets-store-csi-driver-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: secrets-store-csi-driver-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: secrets-store-csi-driver-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: preview
    defaultChannel: preview
    packageName: secrets-store-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/limitador-operator
    name: limitador-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: limitador-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "name": "toy_get_route",
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/limitador-rhel9-operator@sha256:edb28f112a0d3536f2f52467f38df685d953f46f1999cbd61f4fd4fa4805cd34
          createdAt: "2025-04-07T15:52:04Z"
          description: Enables rate limiting for Gateways and applications in a Gateway
            API network
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Limitador is the Schema for the limitadors API
            displayName: Limitador
            kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: Enables rate limiting for Gateways and applications in a Gateway
          API network. - Part of Red Hat Connectivity Link
        displayName: Red Hat - Limitador Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        - kuadrant
        links:
        - name: GitHub
          url: https://github.com/kuadrant/limitador-operator
        - name: Documentation
          url: https://github.com/Kuadrant/limitador/tree/main/docs
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - registry.redhat.io/rhcl-1/limitador-rhel9-operator@sha256:edb28f112a0d3536f2f52467f38df685d953f46f1999cbd61f4fd4fa4805cd34
        - registry.redhat.io/rhcl-1/limitador-rhel9@sha256:98f85a05da53fbf7f20845d1969441620407cc13a185fd52714308679f116425
        version: 1.1.0
      entries:
      - name: limitador-operator.v1.1.0
        version: 1.1.0
      - name: limitador-operator.v1.0.2
        version: 1.0.2
      - name: limitador-operator.v1.0.1
        version: 1.0.1
      - name: limitador-operator.v0.12.1
        version: 0.12.1
      name: stable
    defaultChannel: stable
    packageName: limitador-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/limitador-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: run-once-duration-override-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runoncedurationoverrideoperator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "RunOnceDurationOverride",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "runOnceDurationOverride": {
                    "spec": {
                      "activeDeadlineSeconds": 3600
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certifiedLevel: "false"
          containerImage: registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9-operator@sha256:54993506a3a2fc1e88c9b5aac97ae90d2a04a06ae6e3a97bf487b5e8996b95fb
          createdAt: 2024/07/01
          description: An operator to manage the OpenShift RunOnceDurationOverride
            Mutating Admission Webhook Server
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.1.2'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/run-once-duration-override-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RunOnceDurationOverride is the schema for the runoncedurationoverrides
              API
            displayName: Run Once Duration Override
            kind: RunOnceDurationOverride
            name: runoncedurationoverrides.operator.openshift.io
            version: v1
        displayName: Run Once Duration Override Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mutating
        - webhook
        - workload
        - run-once
        links:
        - name: Source Code
          url: https://github.com/openshift/run-once-duration-override
        - name: Source Code
          url: https://github.com/openshift/run-once-duration-override-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9-operator@sha256:54993506a3a2fc1e88c9b5aac97ae90d2a04a06ae6e3a97bf487b5e8996b95fb
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9@sha256:3dc6e2a5900472e3e083abcb659a13fea3d08f7a9ff167df524b6f890a16dc8e
        version: 1.1.2
      entries:
      - name: runoncedurationoverrideoperator.v1.1.2
        version: 1.1.2
      - name: runoncedurationoverrideoperator.v1.1.1
        version: 1.1.1
      - name: runoncedurationoverrideoperator.v1.1.0
        version: 1.1.0
      - name: runoncedurationoverrideoperator.v1.0.2
        version: 1.0.2
      - name: runoncedurationoverrideoperator.v1.0.1
        version: 1.0.1
      - name: runoncedurationoverrideoperator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: run-once-duration-override-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicast-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicast-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:ccd64090b85536ddaf1b81ca07d40279c9f277b1c5911628f11d678b98dc3bad
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:ccd64090b85536ddaf1b81ca07d40279c9f277b1c5911628f11d678b98dc3bad
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:f0e9f64c353fdbee3bf7ec0eaf070c3b75330ef659a881ec4ebc874ba584d085
        version: 0.5.3
      entries:
      - name: apicast-operator.v0.5.3
        version: 0.5.3
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      name: threescale-2.11
    - currentCSV: apicast-operator.v0.6.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:8c7fe66b8d4044ad965e7b3d149e9d8e3846ac4e55465873f7095a25b297b3df
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:8c7fe66b8d4044ad965e7b3d149e9d8e3846ac4e55465873f7095a25b297b3df
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:12f7a8e69a1369a51bbcd2b34d4ec8c66dd1fc9fb515349a04e7b202953423e3
        version: 0.6.4
      entries:
      - name: apicast-operator.v0.6.4
        version: 0.6.4
      - name: apicast-operator.v0.6.3
        version: 0.6.3
      - name: apicast-operator.v0.6.2
        version: 0.6.2
      - name: apicast-operator.v0.6.1
        version: 0.6.1
      - name: apicast-operator.v0.6.0
        version: 0.6.0
      - name: apicast-operator.v0.5.3
        version: 0.5.3
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      name: threescale-2.12
    - currentCSV: apicast-operator.v0.7.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:bde44c0fb9f13f212184b24e0bef45db734a8439a92179a2089e5aebd6b87b63
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          olm.skipRange: '>=0.6.2 <0.7.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.13/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:bde44c0fb9f13f212184b24e0bef45db734a8439a92179a2089e5aebd6b87b63
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:88dc7a289c803248efbe151999707325b1bc0aac2d26827ea1707e46e5c71652
        version: 0.7.5
      entries:
      - name: apicast-operator.v0.7.5
        version: 0.7.5
      name: threescale-2.13
    - currentCSV: apicast-operator.v0.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:c5e49be7b43a0058f974a5496976255ab3fbd0677bb7bcf9c3438341cf55965a
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.2 <0.8.3'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.14/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:c5e49be7b43a0058f974a5496976255ab3fbd0677bb7bcf9c3438341cf55965a
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:f8cd25658fa6a154e7ea6e5adf0f9dae3fc596ec9989021de6a021ec94eef226
        version: 0.8.3
      entries:
      - name: apicast-operator.v0.8.3
        version: 0.8.3
      - name: apicast-operator.v0.8.2
        version: 0.8.2
      - name: apicast-operator.v0.8.1
        version: 0.8.1
      name: threescale-2.14
    - currentCSV: apicast-operator.v0.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel9-operator@sha256:c4d6acae4e4cc3236c5b8d8dcfb339b072688ab114785d8598bbe68b1f3571a9
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=0.8.0 <0.12.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.14/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel9-operator@sha256:c4d6acae4e4cc3236c5b8d8dcfb339b072688ab114785d8598bbe68b1f3571a9
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        version: 0.12.5
      entries:
      - name: apicast-operator.v0.12.5
        version: 0.12.5
      - name: apicast-operator.v0.12.4
        version: 0.12.4
      - name: apicast-operator.v0.12.3
        version: 0.12.3
      - name: apicast-operator.v0.12.2
        version: 0.12.2
      - name: apicast-operator.v0.12.1
        version: 0.12.1
      - name: apicast-operator.v0.12.0
        version: 0.12.0
      name: threescale-2.15
    - currentCSV: apicast-operator.v0.12.1-mas
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-mas/apicast-rhel9-operator@sha256:e9f22a8af53f881f1948f5d6a033c35a530895633f2eae522b2fb67f9dc6088b
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.17"}]'
          olm.skipRange: '>=0.7.9 <0.12.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-mas/apicast-gateway-rhel8@sha256:f174636a9ad43a2419fff5a9e6797e3185c540a9509188fd1606aef4acea7bdf
        - registry.redhat.io/3scale-mas/apicast-rhel9-operator@sha256:e9f22a8af53f881f1948f5d6a033c35a530895633f2eae522b2fb67f9dc6088b
        version: 0.12.1-mas
      entries:
      - name: apicast-operator.v0.12.1-mas
        version: 0.12.1-mas
      - name: apicast-operator.v0.7.9-mas
        version: 0.7.9-mas
      - name: apicast-operator.v0.7.8-mas
        version: 0.7.8-mas
      - name: apicast-operator.v0.7.7-mas
        version: 0.7.7-mas
      - name: apicast-operator.v0.7.6-mas
        version: 0.7.6-mas
      - name: apicast-operator.v0.7.5-mas
        version: 0.7.5-mas
      - name: apicast-operator.v0.7.4-mas
        version: 0.7.4-mas
      - name: apicast-operator.v0.7.3-mas
        version: 0.7.3-mas
      - name: apicast-operator.v0.7.2-mas
        version: 0.7.2-mas
      - name: apicast-operator.v0.7.1-mas
        version: 0.7.1-mas
      - name: apicast-operator.v0.7.0-mas
        version: 0.7.0-mas
      - name: apicast-operator.v0.6.0-mas
        version: 0.6.0-mas
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      name: threescale-mas
    defaultChannel: threescale-2.15
    packageName: apicast-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: cloud-bulldozer
      provider-url: ""
    name: ripsaw
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ripsaw.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery, Cloud Provider
          containerImage: quay.io/cloud-bulldozer/benchmark-operator:latest
          description: '[Benchmark-operator](https://github.com/cloud-bulldozer/ripsaw/)
            is a benchmarking tool to benchmark your kubernetes cluster. The intent
            of this Operator is to deploy common workloads to establish a performance
            baseline of Kubernetes cluster'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/cloud-bulldozer/benchmark-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Benchmark
            name: benchmarks.ripsaw.cloudbulldozer.io
            version: v1alpha1
        description: '[Benchmark-operator](https://github.com/cloud-bulldozer/benchmark-operator/)
          is a benchmarking tool to benchmark your kubernetes cluster. The intent
          of this Operator is to deploy common workloads to establish a performance
          baseline of Kubernetes cluster'
        displayName: benchmark-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - performance
        - benchmark
        links:
        - name: Site
          url: https://github.com/cloud-bulldozer/benchmark-operator
        - name: Documentation
          url: https://github.com/cloud-bulldozer/benchmark-operator/tree/master/docs
        maintainers:
        - email: rsevilla@redhat.com
          name: rsevilla87
        - email: jtaleric@redhat.com
          name: jtaleric
        - email: rzaleski@redhat.com
          name: dry939
        - email: kwhitley@redhat.com
          name: whitleykeith
        maturity: alpha
        provider:
          name: cloud-bulldozer
        relatedImages:
        - quay.io/cloud-bulldozer/benchmark-operator:latest
        - k8s.gcr.io/redis:v1
        version: 0.1.0
      entries:
      - name: ripsaw.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ripsaw
    provider:
      name: cloud-bulldozer
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: stolostron-engine
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stolostron-engine.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:b7d0c8413a80c251779fe3c1c9200f9fcfbbe085ccee3ab968ebd4a39209946c
          createdAt: "2024-06-28T02:42:26Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/cluster-proxy-addon@sha256:f70a26b276c67ad791d337e35cd102afa82f1005bc6ecbc60309932d2d163a6b
        - quay.io/stolostron/backplane-operator@sha256:b7d0c8413a80c251779fe3c1c9200f9fcfbbe085ccee3ab968ebd4a39209946c
        - quay.io/stolostron/registration@sha256:8dca354dd63739d7578c24d57b2a72c0676c979344e05c6f2634a8a9fae3218d
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/assisted-service@sha256:653728f209070c6adad846a4780449f56985b19006024cfe8137a3a1854fdc85
        - quay.io/stolostron/backplane-must-gather@sha256:2676ebb7438ca42f0d852100612715c429fe9fbe072297fc178d51c34378d1af
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/hypershift-operator@sha256:7dd5eaf195e6b1ca4cecd50514f5b7045948184f9ee617341bc90edc5f9914f2
        - quay.io/stolostron/discovery-operator@sha256:acb2c4033da62055feb49ff3023b2850a2768996e1fa80f52588b3e8e458e8f5
        - quay.io/stolostron/addon-manager@sha256:87c829911438061a53aff2b297416bf43080fdd8d93558caefc198e6f9cb382f
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:b269d88e30551636e024aa1472ea4dd71b873b5a2778a199b32732488fd776fb
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/assisted-image-service@sha256:0c29a74e0f82d1782d5170ff674f5bfcd89ecc5d180fe5b583a5e0f388b9504b
        - quay.io/stolostron/assisted-installer-controller@sha256:213dc6fde247a5b9862f5076d40394164dd596cde7ca2061e58b73f4d2dbc89e
        - quay.io/stolostron/console-mce@sha256:4e49f2b2fdb0491c225e4d5dc143a6d7c0a25cd855057b3792c7ee94a868ac85
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:f0d6d46d3a4d65fba6b74bdc9eab99e8149133f83a5c8486f22a73ab92f78ed6
        - quay.io/stolostron/image-based-install-operator@sha256:b9af0ddb7ae8e6fd29b953dfb917546dfd676206e2ad143e892172b6e1cea5a1
        - quay.io/stolostron/work@sha256:4ef2518aa107ebd28d5fb1a98d63d82a8e0610544600169e0dd5f345599e518b
        - quay.io/stolostron/cluster-api-provider-agent@sha256:c074ca1f28d03e86efe4f09200a38af4ac947247a19c46c5300eefcd4f337006
        - quay.io/stolostron/discovery-operator-tests@sha256:c98c42b0c16306de86d7b365096cb809a0c797b0a4b7dc04a7de877d0ba5733a
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/helloworld@sha256:abc0d8002c5fe6d95a1f24fce1b7899ca24fac576c894ea7443cf7e15e6a42c2
        - quay.io/stolostron/cluster-curator-controller@sha256:e3a91a4b06dd73fc4e401ddf034e61d4d5b2d324d2c5cea92941e1a56c4023c1
        - quay.io/stolostron/assisted-service-el8@sha256:31422ffce8b94d662131986bd084fcba503809cb939389096f9c9dbb27cb58db
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/discovery-testserver@sha256:83d7d6cfaef629a4e3d5d0e9fe35456aa220f1eb68f6fe611d85232acf46bc73
        - quay.io/stolostron/placement@sha256:0dbcc7c62569920afcad1443529f10e38d651a5fe5ae14ade89d6e42fb522e60
        - quay.io/stolostron/provider-credential-controller@sha256:d3bd513c06591786216f2baab46c9f57e373d1939f038edacecb2180f673d888
        - quay.io/stolostron/managedcluster-import-controller@sha256:925bbc6173501bdb60d329a1f5a134ea3e9230034509320f93382f1c570be275
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/multicloud-manager@sha256:05da692b2780453115ef6c195658cadb2f28eb0f3a86dc0a586e459dd7c55afe
        - quay.io/stolostron/cluster-proxy@sha256:6cd53c588bb9c5064ba0eb8eab03cd12c86976b353209102059f7d76ec59d7c3
        - quay.io/stolostron/clusterclaims-controller@sha256:a0613147b0ec010b8a4442a3c515f7ffc48ec7d8137436a0e8b935443cac2f96
        - quay.io/stolostron/cluster-image-set-controller@sha256:632661c6cbfce7df0545bc2fccf1ef87257abb3cc92dcede054806613a3540e3
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:26f8bb3c9b531e96e6538a5bff96569628f45c06eb3b991ebb421d4e90d7c7e3
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:a39c52097faf7a21cc717768401851466f8c9341f08103e4d1fe8459ae7f5ea2
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/assisted-installer@sha256:a31ed7b8e4ff2f9e5ade2e61191cc759e1d47ddbba95a2af41bfa53fa6ddf855
        - quay.io/stolostron/backplane-operator-test@sha256:961d8c40aa3526ca368fd6833202800ba2eb344c97788a646d6fb350a5338397
        - quay.io/stolostron/hive@sha256:fd39b8340cfaeb9849e83bd910e8a3575791fcc58dbbdeb3e66528f009b33d65
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:b76a201edcad4805e32c40d3f0ad3f61584a879aa00c9e49aae3b2985bfb5fbc
        - quay.io/stolostron/managed-serviceaccount@sha256:4c3f1e0b9c4f2937ceb46e6b3f63e3a07061f9b29fdda7dcfdf9ebd74c3bb30b
        - quay.io/stolostron/registration-operator@sha256:4b6c31592c405db11375f4694a7f15bb13c5c9a770786a1607ea0be615396435
        - quay.io/stolostron/assisted-installer-agent@sha256:a2efc62ca0080779acfea2e413e524608bc0e8f72b2b917e6c7e76864d08b2b6
        - quay.io/stolostron/hypershift-addon-operator@sha256:47ed6a7c00036599051718d6015bee35818ea5bbaa14f1c01bedb338dce90668
        version: 0.4.0
      entries:
      - name: stolostron-engine.v0.4.0
        version: 0.4.0
      name: community-0.4
    - currentCSV: stolostron-engine.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:30d5bcccc542988ac788dbca305e6f29fb678470b19e48adada9ca6f5803f04a
          createdAt: "2024-10-02T05:51:59Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.5.0 <0.5.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/discovery-operator-tests@sha256:e8d3d1675242272cc139b0dba15605d00e25b17b4b1cc1b5b4f3ea8913ffb2da
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:2fe53994fa5bde93112a2cc75758e0f68be013a34c8206066d0b315c2e6a6f47
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:4b5b2bea2aad6c7e16fe1520a2fb8454b0312f2749d2f398ca54e644e813055c
        - quay.io/stolostron/assisted-service@sha256:6383ba508a0cac030a46df96ac37daaeddc27ad1fb41d8b4dca5a87cf4ab9c31
        - quay.io/stolostron/cluster-image-set-controller@sha256:654e218ecf0a3f67520c0521b544eea9d4552e92f54bf424e88ccab24e7c151e
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:14a0648b76301de182154a8d7d6b7e7eca0bfbed226e6dcebddb75af5152451f
        - quay.io/stolostron/backplane-operator-test@sha256:048f21cf89fb5ed2557ffc90ae5d7104cdd5ce0e6a2a865ce6f4230d09400115
        - quay.io/stolostron/assisted-installer@sha256:dcc4e28c7eea55dd2ac22dd75c2b7b96ee3552fcbeba8128ad8528fda235d879
        - quay.io/stolostron/cluster-proxy-addon@sha256:7aa09b0d53fdc0f5cec9c8a13c160277a9274466025f11b5b04da95c9a10e59c
        - quay.io/stolostron/provider-credential-controller@sha256:d240ad43c23de822f91a33849448ae70a4cf27570da3387bdecd101ca2f0b65d
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/assisted-service-el8@sha256:c8248f9a7ac02f47373d097953cee38b66125e98b98922f18c578a591f04bcca
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/multicloud-manager@sha256:6dd212840ea7fadd427b503ce69e6e32464881a0047e2af86f9f2c7562bd4703
        - quay.io/stolostron/hypershift-operator@sha256:a30fb6cf4d5e103820813019ce452d8134b0fee8ecd3c2166d5681199d4a0cad
        - quay.io/stolostron/helloworld@sha256:96a9b8855fb8800a85aed6fdc4911295dc2c12629bbdcc34ed8dc8061827b17e
        - quay.io/stolostron/console-mce@sha256:10a54c5012ef6439539297919c71d9b53aaaddc6342d90677113d6ce7b42e3fe
        - quay.io/stolostron/hypershift-addon-operator@sha256:ba199ebfe0eea3ef6b07123f51116831e7b0a73860b0410ece723bcc9a3230af
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:b76a201edcad4805e32c40d3f0ad3f61584a879aa00c9e49aae3b2985bfb5fbc
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/cluster-curator-controller@sha256:fc2ad8e0b6701638ab89bb62a49700fdaadfe1149228cd0b3cd4a41810020c6d
        - quay.io/stolostron/cluster-api-provider-agent@sha256:1458de75311675a5acec7361e507edb24b970ffa3403e916de3413f7061bfe36
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        - quay.io/stolostron/assisted-image-service@sha256:703c12a918d3c66b62c29a5abd0e8c5032299289cfed459671d6ca82c4f35c73
        - quay.io/stolostron/registration@sha256:4779cf430d479806efd2f3cb7e2a9ed3dce749287609c39ad45f1b06e220d023
        - quay.io/stolostron/assisted-installer-agent@sha256:16b040dd0690b62bc051aa6369272d004d2157a8e5230d318ba827adba8e8d63
        - quay.io/stolostron/managedcluster-import-controller@sha256:01ab0c8b664692d69c2a9b3363e007a5611c332ac0e3491234dd0cbfba31c88a
        - quay.io/stolostron/image-based-install-operator@sha256:bd76503406e6d20693e636461411e158c8dde3d03c4904ecb8a1850deeefa4f5
        - quay.io/stolostron/placement@sha256:dc0ad451b085b3a747ee648475f8e8102b645df69b898fc88f1cb190d1b95a83
        - quay.io/stolostron/clusterclaims-controller@sha256:1c0b3c4730ef1ab628bf2c1b7cf12b1b93b149abdefc4d908dbdc90e4b6c1c19
        - quay.io/stolostron/backplane-operator@sha256:30d5bcccc542988ac788dbca305e6f29fb678470b19e48adada9ca6f5803f04a
        - quay.io/stolostron/addon-manager@sha256:9384ae0e9bcc9f85d3fe36e44ab7b3a61777197846559b5add0c56f513e35333
        - quay.io/stolostron/discovery-operator@sha256:db56e2a78ed27a46b47b519e29db66636b9abb499fd507d99451bae47438e32b
        - quay.io/stolostron/cluster-proxy@sha256:be09bb293a990600ec48aba2f9bf8b457d5bdbc6b136be319e5feebadb232a8c
        - quay.io/stolostron/work@sha256:4467a4bd00586ab837cb7e8bf56ea1faff82ed5ce2fffdeb7e51ee5fafb433b0
        - quay.io/stolostron/assisted-installer-controller@sha256:01dd7c7fb9cb335215afefde204309d3127847b900658ee9744b9d31df847565
        - quay.io/stolostron/hive@sha256:707ac9a2ad18dbc4c4686f1b72946077c7232afd05505d5e50e13f806a452392
        - quay.io/stolostron/backplane-must-gather@sha256:92e992e23855375a8bae542480d789d1bb25a48959896bfb75f659df2a37dd98
        - quay.io/stolostron/managed-serviceaccount@sha256:667f823799984404c0c1589df7b857a6a35218c2653254905dc22ca90ce1542b
        - quay.io/stolostron/discovery-testserver@sha256:0b23021d6c0c7d6c7aa97a92fd0c40f31fdff6363a5b10a1edc864b5d8d54108
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/registration-operator@sha256:b42341143b1e1693748f4b7d69e97c56c8a5f9408abc5db1de96ba246b7b98a9
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        version: 0.5.3
      entries:
      - name: stolostron-engine.v0.5.3
        version: 0.5.3
      - name: stolostron-engine.v0.5.2
        version: 0.5.2
      - name: stolostron-engine.v0.5.1
        version: 0.5.1
      - name: stolostron-engine.v0.5.0
        version: 0.5.0
      name: community-0.5
    - currentCSV: stolostron-engine.v0.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:fa1fa6ef3958fd9d50bfb38197f033a635a05b47c32e653822246e708d8598cc
          createdAt: "2025-02-24T01:37:53Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.6.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/console-mce@sha256:bfaf9180cd1e5af4d34daf4f1dcdee89951067a605f21bfe6e1ce6bda85b0334
        - quay.io/stolostron/registration@sha256:9c221a40ffa119ac154d76f578e0919b5a63137c908afb0583519a9b8a399b2b
        - quay.io/stolostron/cluster-proxy@sha256:b341dd7058cdffccb00a026908031fa123fec5ba58ba9307dbb873b75d250e4d
        - quay.io/stolostron/clusterclaims-controller@sha256:b796cb2edd3d4c22b29fc6d2b45e5328a59b7ec1bdc52d5500e1a4ee74725d6c
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:2781d87c1568f2e40409f5431bfd1a9a4c9ca35655537e7d4db65d9b881edd08
        - quay.io/stolostron/backplane-operator-test@sha256:b04661ad658423c9c7557b722adb2159d988c10445de1c59f39c44583262cb42
        - quay.io/stolostron/registration-operator@sha256:52289e6cf5f159d344717b49e6226ff9d4ee33942f63cc543d1b820d7ee24fc5
        - quay.io/stolostron/work@sha256:6ec081826f32467ea1e1a91f6624df349a382fb8a1cbbda906141281ac2bad5c
        - quay.io/stolostron/discovery-operator@sha256:43f40b9c20b112248fcdb42d114fe4af73fdd45b05d9ff7855d827cdd5e971b4
        - quay.io/stolostron/multicloud-manager@sha256:0293f0852206562b891375b167cfe17843e2e17585f04c9bb6482d1fe78089d0
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/image-based-install-operator@sha256:fecbff73eb97b28bedd7c17d5adbef4826f181a6838ad4cc72e7478acf9b4254
        - quay.io/stolostron/placement@sha256:543f09446652bff8c4149945d94d715eff7a76d82783f2bad89e712798542c73
        - quay.io/stolostron/assisted-installer@sha256:c88976ab4b35d458a39044e522d2d5ab1916d8485af07bbef373af1bcbbd6af0
        - quay.io/stolostron/assisted-service-el8@sha256:f50f4dece91100f45ef9e20b1e52557a55db45b17a380f32d365ee013d695d5c
        - quay.io/stolostron/assisted-service@sha256:58271e54eda979c43fb474c3bb255f7328e380a7af9ba833e5f7558f323bd239
        - quay.io/stolostron/ose_cluster_api_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/hypershift-operator@sha256:2d1809223ce5e75b4133a39b78a269703c6b5234f7ec2a7c0072124c345fae0a
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:65180432218d035ffc7996a7643ddb5300eb8b09e47e09cb270cb9b961aecf7a
        - quay.io/stolostron/helloworld@sha256:4838f699ea2461f8358dcda32b3ac96ed05b55d4e1958eac2bdacae4807501d1
        - quay.io/stolostron/hive@sha256:1e1f60492c6a3b43b69fe8a0640f9b81f5a1cfafc0afcd6854c3cfe362ed063a
        - quay.io/stolostron/assisted-installer-controller@sha256:3b0b43fdbbbbd5c1e7588fc449b002e1ed9ad0ea6daa774925a962bd499a4195
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        - quay.io/stolostron/cluster-image-set-controller@sha256:629361c1c0ec50fe6410fc6fe6417cba89957c49ab89201c6470c0f97c8fc91d
        - quay.io/stolostron/addon-manager@sha256:7e347a177d8c5b80a78411a3e8dbf04e49d157a4ca67e7fd5170970c2ba62da9
        - quay.io/stolostron/discovery-operator-tests@sha256:21449ecd291cf7203cc8e27a2a00ac60e74f5993708f9e7465e8efdfdca1c89f
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/backplane-operator@sha256:fa1fa6ef3958fd9d50bfb38197f033a635a05b47c32e653822246e708d8598cc
        - quay.io/stolostron/assisted-image-service@sha256:6846719f91dd954ce4b4c6e7c9654c7d79db72990270acd3c34bdd2a89af34fc
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:44f8bb72b975456d4151a88a560fc525ad6f4da7968a38e94a86b6dd77d3c543
        - quay.io/stolostron/cluster-proxy-addon@sha256:7da9da3e4ecd1e0df636c9f72f9f30cc7ff24b81e247ba9825cef39f18e86182
        - quay.io/stolostron/ose_aws_cluster_api_controllers_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:b016a001a53bc37ae534b74cc832552759800746c27e16adcf0626ff9ad2e4c0
        - quay.io/stolostron/discovery-testserver@sha256:f757a7a11108a88de4bc0ecc5e16f74b3bd9b5110ec94bfeef2074121ec1e47f
        - quay.io/stolostron/hypershift-addon-operator@sha256:65f2b7a356c22413307ce7b7de38ce00c31e4c43b6ee3b4b484c0bc61dc483e9
        - quay.io/stolostron/managed-serviceaccount@sha256:0e4edd9ab95c1558fba40a0c41c5cc326996e9e19e296324ea10b5a442a7ef63
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/cluster-curator-controller@sha256:5d05462c2c9d93b21656a436c01fbda4256757bd057e5de7014eb6f320310a1e
        - quay.io/stolostron/assisted-installer-agent@sha256:cd74c8b8888c100bf567dd07cfff40097c458d18fcc29a11ed3830f697156267
        - quay.io/stolostron/cluster-api-provider-agent@sha256:dda3e9aef611c744c4c89c2689a12d0649a173c1a64954c3037de55ec8eacec9
        - quay.io/stolostron/managedcluster-import-controller@sha256:ecfe7c1f2d24a75bfb8db2a0a2765fe5e9136f81f341882de896431570fd10e9
        - quay.io/stolostron/backplane-must-gather@sha256:99b558d687e1c17036b164d819cb1c9b96a814be79e3c87960ef7cdb521eee7d
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/provider-credential-controller@sha256:a5ac3999d0aefb1dd9a239f9394485f93e4252a45eb4a43a43e603aaff80b3bb
        version: 0.6.3
      entries:
      - name: stolostron-engine.v0.6.3
        version: 0.6.3
      - name: stolostron-engine.v0.6.2
        version: 0.6.2
      - name: stolostron-engine.v0.6.1
        version: 0.6.1
      - name: stolostron-engine.v0.6.0
        version: 0.6.0
      name: community-0.6
    - currentCSV: stolostron-engine.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:9b39a5b6a502778d7884e276b276a9b44e6dbc88e7e82470e26d91ded3cb3839
          createdAt: "2025-06-09T15:04:58Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.7.1'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational components for managing multiple Kubernetes-based clusters. The deployment of the multicluster engine components is
              determined based on the configuration that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/registration-operator@sha256:a9fdf837e149c8659fc99deab6a284975c8eb8785549381fa082303cdb172e44
        - quay.io/stolostron/discovery-testserver@sha256:faef24208a65283b4157521079a7bf23bc4ebd553d70a83a78c324b5cb0e24e4
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/hypershift-addon-operator@sha256:ed1c1580fee28310fc1b137c5c70271448e58ed29af4371515fe9c74512a2c7e
        - quay.io/stolostron/hypershift-operator@sha256:c9a8fc943ef35126434cc767f152abd09943a4d2d194a9966495fdc0870b417c
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/provider-credential-controller@sha256:3484cb0c2102f4a522efe31bc4c77910a202b42e1ad082a0a8709cc1a7698649
        - quay.io/stolostron/assisted-installer@sha256:17b78d73775a93b31ca872db79520ddc49d2be042fdf1599f419fbcbfd9a98ac
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/hive@sha256:86ff94d3e7c174ac4d6818183c66791edc7638965dd09e2d41172609c81f451f
        - quay.io/stolostron/ose_cluster_api_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/work@sha256:709c41d1a7a1eaf10b334dd0c0105a261d6db811693d5ae67eb55b9a26daed87
        - quay.io/stolostron/backplane-operator@sha256:9b39a5b6a502778d7884e276b276a9b44e6dbc88e7e82470e26d91ded3cb3839
        - quay.io/stolostron/cluster-image-set-controller@sha256:7f05ad13ffaa11daec066c5889b87df1b631e6eb3dc1c2913a2766c8e65218aa
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/cluster-proxy@sha256:1528e36dfa4fa2c9063d84c0f2b9aa8d2b549a750e812d4a2e2b8c7515bfe5cb
        - quay.io/stolostron/backplane-operator-test@sha256:8d6514fc0db5ee79c346ddc31bb094ad87fa375519da67e32398300d70d72f6e
        - quay.io/stolostron/cluster-api-provider-agent@sha256:2b22890b948ad1d597c424d73d4fe34c7c4d265ad6e54fa0ed87a7dec920ac57
        - quay.io/stolostron/assisted-service@sha256:af3532d487e03a077ce23e4178abbcb41cf45726b646d7660840c036677715e8
        - quay.io/stolostron/backplane-must-gather@sha256:f4f9596eade78333628199f62f8e71e4c2b3957c20da5bad807c12a2a02560db
        - quay.io/stolostron/addon-manager@sha256:a245847f135f96c4b4fcaa673fc32013c8e407252b3259175c796e3c16111c71
        - quay.io/stolostron/cluster-api-provider-openshift-assisted-control-plane@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/assisted-image-service@sha256:b6df9c9fc8bbfdc58e2169b9e199e1eed188d3d847511eba01eb4920fabe3992
        - quay.io/stolostron/ose_aws_cluster_api_controllers_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:44f8bb72b975456d4151a88a560fc525ad6f4da7968a38e94a86b6dd77d3c543
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/assisted-installer-agent@sha256:ffceb06ce038013291e092324d3c49091cd1b2ff1569abe77c29484075e28107
        - quay.io/stolostron/discovery-operator-tests@sha256:d96fe1f4f49ecfd4a27bedd87d7070a3c7f3f82b6ee8bb59e98a262d8cedcfc0
        - quay.io/stolostron/discovery-operator@sha256:a6a37237bc06f0635a936b803d3c53f0f0caf5426c76b37951c4e157b181ecb3
        - quay.io/stolostron/placement@sha256:6276e411c9f7cf7072d3d364569b6c3cd472742ab4ad48912f18cb4091d27a6f
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:1757b9ad635839087c4bef198d5e81f4abb7aee38bc74690021f9e8887366bcf
        - quay.io/stolostron/ip-address-manager@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        - quay.io/stolostron/multicloud-manager@sha256:5a95a5d1dbb17456d7ce62f1a517ae49ca780e25f3ebb74b32346de1df23b4c9
        - quay.io/stolostron/assisted-service-el8@sha256:f52c55691850a1f2f254b0e9377ba97493be337e012223e5a2b3190f03e1a349
        - quay.io/stolostron/ose-baremetal-cluster-api-controllers-rhel9@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/image-based-install-operator@sha256:9d6796007b616de60560f55f14546295c4e48b3768d059098fce9367674a3368
        - quay.io/stolostron/helloworld@sha256:a790795409f6e781f506b6a63f39f7d344f01efa30c3de87b1727dc2010d5bbf
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:1d4221cfc364743b58e15bad3afe9548f4151f06b0dbf0aec7c00715ba75abba
        - quay.io/stolostron/assisted-installer-controller@sha256:8ea9d12fa30825fe4f965a7d506af687e8de709f8d07dd75204747da2a6e09a8
        - quay.io/stolostron/cluster-proxy-addon@sha256:c776405c4b1b623500c69a12b7d76563646b27cf23bab74dc8f1361534cfb81e
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/cluster-curator-controller@sha256:d667bd0892812eb4e1f724679759469db335db4ab801f9d57ea8818d8cd13188
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:1e95f55d6918263e79de1c1e2be8a96cd3d35a346dde6d3e2fb484f48e567811
        - quay.io/stolostron/console-mce@sha256:7f538a21eb22ccd57765f47692da23be9cecfd3b7b65986f1413234bdf709530
        - quay.io/stolostron/managed-serviceaccount@sha256:4bcd5c9b2044254c28215fd78bf1e3fcd919578238986831fcea8055b2764696
        - quay.io/stolostron/clusterclaims-controller@sha256:afab145b12a52f81e1f2eaa189731e3a6287b54f2f1f1ad2976ad378fa7dfa3a
        - quay.io/stolostron/managedcluster-import-controller@sha256:4d0be693db0576eb2e9e3d409bf1479cc60784f5747ed1b25b6a2443c5e91e36
        - quay.io/stolostron/cluster-api-provider-openshift-assisted-bootstrap@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/registration@sha256:2db2c28af25007c77979cc456d929f5f077cd671cc4e0562c43efce2c42c4cd8
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        version: 0.7.1
      entries:
      - name: stolostron-engine.v0.7.1
        version: 0.7.1
      - name: stolostron-engine.v0.7.0
        version: 0.7.0
      name: community-0.7
    defaultChannel: community-0.7
    packageName: stolostron-engine
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pulp Community
      provider-url: https://github.com/pulp/pulp-operator
    name: pulp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pulp-operator.v1.0.0-alpha.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache_enabled": true,
                  "content": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "ingress_type": "Route",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/"
                  },
                  "storage_type": "File",
                  "worker": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example-pulp"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true,
                    "redis_storage_class": "standard"
                  },
                  "content": {
                    "replicas": 1
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "pulp",
                  "file_storage_access_mode": "ReadWriteOnce",
                  "file_storage_size": "2Gi",
                  "file_storage_storage_class": "standard",
                  "image_version": "nightly",
                  "image_web_version": "nightly",
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/",
                    "telemetry": false
                  },
                  "web": {
                    "replicas": 1
                  },
                  "worker": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "galaxy-example"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "api": {
                    "replicas": 1
                  },
                  "content": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "galaxy",
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "file_storage_storage_class": "standard",
                  "image": "quay.io/pulp/galaxy-minimal",
                  "image_version": "nightly",
                  "image_web": "quay.io/pulp/galaxy-web",
                  "image_web_version": "nightly",
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/api/galaxy/pulp/",
                    "telemetry": false
                  },
                  "signing_scripts_configmap": "signing-scripts",
                  "signing_secret": "signing-galaxy",
                  "storage_type": "File",
                  "web": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "worker": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "galaxy-example",
                  "deployment_type": "galaxy",
                  "instance_name": "pulp",
                  "postgres_configuration_secret": "galaxy-example-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "galaxy-example"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.0.0-alpha.1
          createdAt: "2021-03-26 16:57:40"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1alpha1
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1alpha1
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1alpha1
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://docs.pulpproject.org/](https://docs.pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          Note that Pulp operator works with three different types of service containers (the operator itself, the main service and the web service):

          |           | Operator | Main | Web |
          | --------- | -------- | ---- | --- |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[pulp-minimal](https://quay.io/repository/pulp/pulp-minimal?tab=tags) | [pulp-web](https://quay.io/repository/pulp/pulp-web?tab=tags) |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[galaxy-minimal](https://quay.io/repository/pulp/galaxy-minimal?tab=tags) | [galaxy-web](https://quay.io/repository/pulp/galaxy-web?tab=tags) |

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://docs.pulpproject.org/pulp_operator/container/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://docs.pulpproject.org/pulp_operator/](https://docs.pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://docs.pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: beta
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/pulp/pulp-operator:v1.0.0-alpha.1
        - quay.io/pulp/pulp-minimal:stable
        - quay.io/pulp/pulp-web:stable
        - docker.io/library/redis:latest
        - docker.io/library/postgres:13
        version: 1.0.0-alpha.1
      entries:
      - name: pulp-operator.v1.0.0-alpha.1
        version: 1.0.0-alpha.1
      name: alpha
    - currentCSV: pulp-operator.v1.0.0-beta.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "Pulp",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true
                  },
                  "content": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/"
                  },
                  "worker": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "Pulp",
                "metadata": {
                  "name": "galaxy-example"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "api": {
                    "replicas": 1
                  },
                  "content": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "galaxy",
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "file_storage_storage_class": "standard",
                  "image": "quay.io/pulp/galaxy-minimal",
                  "image_version": "4.6.3",
                  "image_web": "quay.io/pulp/galaxy-web",
                  "image_web_version": "4.6.3",
                  "ingress_type": "nodeport",
                  "nodeport_port": 30000,
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "analytics": false,
                    "ansible_api_hostname": "http://nodeport.local:30000",
                    "api_root": "/api/galaxy/pulp/",
                    "content_origin": "http://nodeport.local:30000",
                    "galaxy_feature_flags": {
                      "ai_deny_index": true,
                      "execution_environments": true
                    },
                    "token_server": "http://nodeport.local:30000/token/"
                  },
                  "signing_scripts": "signing-scripts",
                  "signing_secret": "signing-galaxy",
                  "web": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "worker": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "galaxy-example",
                  "deployment_type": "galaxy",
                  "postgres_configuration_secret": "galaxy-example-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "galaxy-example"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.0.0-beta.4
          createdAt: "2024-01-18T20:33:37Z"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1beta2
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1beta2
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1beta2
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://docs.pulpproject.org/](https://docs.pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          Note that Pulp operator works with three different types of service containers (the operator itself, the main service and the web service):

          |           | Operator | Main | Web |
          | --------- | -------- | ---- | --- |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[pulp-minimal](https://quay.io/repository/pulp/pulp-minimal?tab=tags) | [pulp-web](https://quay.io/repository/pulp/pulp-web?tab=tags) |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[galaxy-minimal](https://quay.io/repository/pulp/galaxy-minimal?tab=tags) | [galaxy-web](https://quay.io/repository/pulp/galaxy-web?tab=tags) |

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://docs.pulpproject.org/pulp_operator/container/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://docs.pulpproject.org/pulp_operator/](https://docs.pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://docs.pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: beta
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - docker.io/library/postgres:13
        - quay.io/pulp/pulp-minimal:stable
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.4
        - quay.io/pulp/pulp-operator:v1.0.0-beta.4
        - quay.io/pulp/pulp-web:stable
        - docker.io/library/redis:latest
        version: 1.0.0-beta.4
      entries:
      - name: pulp-operator.v1.0.0-beta.4
        version: 1.0.0-beta.4
      - name: pulp-operator.v1.0.0-beta.3
        version: 1.0.0-beta.3
      - name: pulp-operator.v1.0.0-beta.2
        version: 1.0.0-beta.2
      - name: pulp-operator.v1.0.0-beta.1
        version: 1.0.0-beta.1
      - name: pulp-operator.v1.0.0-alpha.9
        version: 1.0.0-alpha.9
      - name: pulp-operator.v1.0.0-alpha.8
        version: 1.0.0-alpha.8
      - name: pulp-operator.v1.0.0-alpha.7
        version: 1.0.0-alpha.7
      - name: pulp-operator.v1.0.0-alpha.6
        version: 1.0.0-alpha.6
      - name: pulp-operator.v1.0.0-alpha.5
        version: 1.0.0-alpha.5
      - name: pulp-operator.v1.0.0-alpha.4
        version: 1.0.0-alpha.4
      - name: pulp-operator.v1.0.0-alpha.3
        version: 1.0.0-alpha.3
      - name: pulp-operator.v1.0.0-alpha.2
        version: 1.0.0-alpha.2
      - name: pulp-operator.v1.0.0-alpha.1
        version: 1.0.0-alpha.1
      name: beta
    defaultChannel: beta
    packageName: pulp-operator
    provider:
      name: Pulp Community
      url: https://github.com/pulp/pulp-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-common-service-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-common-service-operator.v4.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v3",
                "kind": "CommonService",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-common-service-operator",
                    "app.kubernetes.io/managed-by": "ibm-common-service-operator",
                    "app.kubernetes.io/name": "ibm-common-service-operator"
                  },
                  "name": "example-commonservice"
                },
                "spec": {
                  "size": "starterset"
                }
              }
            ]
          capabilities: Seamless Upgrades
          cloudPakThemesVersion: styles4100.css
          containerImage: registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
          createdAt: "2025-03-18T21:27:31Z"
          description: The IBM Cloud Pak foundational services operator is used to
            deploy IBM foundational services.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-common-service-operator
          nss.operator.ibm.com/managed-webhooks: ""
          olm.skipRange: '>=3.3.0 <4.12.0'
          operatorChannel: v4.12
          operatorVersion: 4.12.0
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/ibm-common-service-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CommonService is the Schema for the commonservices API. Documentation
              For additional details regarding install parameters check https://ibm.biz/icpfs39install.
              License By installing this product you accept the license terms https://ibm.biz/icpfs39license
            displayName: CommonService
            kind: CommonService
            name: commonservices.operator.ibm.com
            version: v3
        description: |-
          # Introduction

          The `IBM Cloud Pak foundational services` is a bridge to connect IBM Cloud Paks and Operand Deployment Lifecycle Manager (ODLM) with IBM Cloud Pak foundational services. You can also install the `IBM Cloud Pak foundational services` in stand-alone mode.

          When you install this operator, the operator completes the following tasks:

            - Installs ODLM operator namespace
            - Creates the Common Services `OperandRegistry` and `OperandConfig` in the service namespace

          # Details
          For more information about installing this operator and other Common Services operators, see [Installer documentation](http://ibm.biz/cpcs_opinstall). If you are using this operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For more information about IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).

          For more information about the available IBM Cloud Pak foundational services, see the [IBM Knowledge Center](http://ibm.biz/cpcsdocs).

          ## Supported platforms

          Red Hat OpenShift Container Platform 4.5 or newer installed on one of the following platforms:

           - Linux x86_64
           - Linux on Power (ppc64le)
           - Linux on IBM Z and LinuxONE

          ## Operator versions

           - 4.12.0

          ## Prerequisites

          Before you install this operator, you need to first install the operator prerequisites:

          - For the list of prerequisites for installing the operator, see the IBM Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).

          ### Resources Required

          #### Minimum scheduling capacity

          | Software                   | Memory (MB) | CPU (cores) | Disk (GB) | Nodes  |
          | -------------------------- | ----------- | ----------- | --------- | ------ |
          | ibm common service operator | 200          | 0.2        | 1          | worker |
          | **Total**                  | 200         | 0.2         | 1         |        |

          ## Documentation

          - If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak. For a list of IBM Cloud Paks, see [IBM Cloud Paks that use IBM Cloud Pak foundational services](http://ibm.biz/cpcs_cloudpaks).
          - If you are using the operator in stand-alone mode or with an IBM Containerized Software, see the IBM Cloud Pak foundational services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall).

          ## Installing

          For installation, see the [IBM Cloud Platform Common Services documentation](http://ibm.biz/cpcsdocs).

          ## Configuration

          For configuration, see the [IBM Cloud Platform Common Services documentation](http://ibm.biz/cpcsdocs).

          ### SecurityContextConstraints Requirements

          The IBM Common Service Operator supports running with the OpenShift Container Platform 4.3 default restricted Security Context Constraints (SCCs) and IBM Cloud Pak Security Context Constraints (SCCs).

          For more information about the OpenShift Container Platform Security Context Constraints, see [Managing Security Context Constraints](https://docs.openshift.com/container-platform/4.3/authentication/managing-security-context-constraints.html).

          For more information about the IBM Cloud Pak Security Context Constraints, see [Managing Security Context Constraints](https://ibm.biz/cpkspec-scc).

          ### Limitations

          No
        displayName: IBM Cloud Pak foundational services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Common Service Operator
          url: https://github.com/IBM/ibm-common-service-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:4209e78f12120b9c9f320496b780c3993a9317825ea17dc7d554452b28c99014
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:9a537eae7e581322e2f0764ed38a797ba7cd260724a285f24d5095058b7b47a0
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-operator-bundle@sha256:086eaeaaaf440a1d4a254984afacd8d04583aafb9ec11ac072feb42a87b986f7
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90046eb6170e12439ccc5aa343f19a4d4fc99a10d98b383cf1dc0e020ebeed55
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:18f44ab15289625fa2c6b72ece6e06afb51058dcba30b8f38b4fdeedf8c9e51b
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:0fdf1beeb4888f648f6dbd17a29dcc53a715d2c4301bc714a4172b2daf8d13be
        version: 4.12.0
      entries:
      - name: ibm-common-service-operator.v4.12.0
        version: 4.12.0
      name: v4.12
    defaultChannel: v4.12
    packageName: ibm-common-service-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: rebellions
      provider-url: https://rebellions.ai/
    name: rbln-npu-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: candidate
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: fast
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: rbln-npu-operator
    provider:
      name: rebellions
      url: https://rebellions.ai/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/component: operator
      app.kubernetes.io/managed-by: kustomize
      app.kubernetes.io/name: zabbix-operator-certified
      app.kubernetes.io/part-of: zabbix-operator-certified
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zabbix LLC
      provider-url: ""
    name: zabbix-operator-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zabbix-operator-certified-rhmp.v6.0.40
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent",
                "metadata": {
                  "labels": {
                    "app": "agent",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "log_remote_commands": true,
                  "max_lines_per_second": 20,
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent2",
                "metadata": {
                  "labels": {
                    "app": "agent2",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent2"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "enable_persistent_buffer": false,
                  "enable_status_port": false,
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "persistent_buffer_claim": "",
                  "persistent_buffer_period": "1h",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixFull",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-full"
                },
                "spec": {
                  "builtin_report_service": true,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysql_volumeclaim": "mysql-full-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxyMysql",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-mysql"
                },
                "spec": {
                  "db_server_port": 3306,
                  "enable_snmp_traps": false,
                  "internal_db": true,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix_proxy",
                  "proxy": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "zabbix_mysql_volumeclaim": "mysql-proxy-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxySqlite",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-sqlite"
                },
                "spec": {
                  "db_data_volumeclaim": "zabbix-proxy-sqlite",
                  "enable_snmp_traps": false,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "proxy": {
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "node_name_as_db_name": true,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "proxy_size": 1,
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  }
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixServer",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-server"
                },
                "spec": {
                  "builtin_report_service": true,
                  "db_server_host": "mysql-server",
                  "db_server_port": 3306,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix",
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              }
            ]
          ansible.operator-sdk/reconcile-period: 5s
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          certified: "true"
          containerImage: registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
          createdAt: "2025-05-26T07:41:22Z"
          description: Zabbix operator with multiple deployment variants and different
            components
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/zabbix-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/zabbix-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <6.0.40
          operatorframework.io/suggested-namespace: zabbix
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: ""
          support: Zabbix
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Zabbix agent 2 is a new generation of Zabbix agent and may
              be used in place of Zabbix agent
            displayName: Zabbix agent 2
            kind: ZabbixAgent2
            name: zabbixagent2s.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix agent is deployed on a monitoring nodes to actively
              monitor local resources and applications
            displayName: Zabbix agent
            kind: ZabbixAgent
            name: zabbixagents.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database server, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Full
            kind: ZabbixFull
            name: zabbixfulls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with MySQL database server
            displayName: Zabbix proxy (MySQL)
            kind: ZabbixProxyMysql
            name: zabbixproxymysqls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with SQLite database
            displayName: Zabbix proxy (SQLite3)
            kind: ZabbixProxySqlite
            name: zabbixproxysqlites.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database support, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Server
            kind: ZabbixServer
            name: zabbixservers.kubernetes.zabbix.com
            version: v1alpha1
        description: "## About this Operator\n\nZabbix helps you to real-time monitoring
          of millions of metrics collected from tens of thousands of servers, virtual
          machines and network devices.\nThe Zabbix Operator allows users to easily
          deploy, manage, and maintain Zabbix deployments on OpenShift. By installing
          this integration you will be able to deploy Zabbix server / proxies and
          other components with a single command.\n## Supported Features\n* **Zabbix
          Server** - Simple Zabbix installation with included Zabbix server, Zabbix
          web-interface and Zabbix Java Gateway with MySQL database support. The feature
          does not provide MySQL service and requires an external MySQL database.
          \n* **Zabbix Server (Full)** - Zabbix installation with included Zabbix
          server, Zabbix web-interface, Zabbix Java Gateway and MySQL server instance.\n*
          **Zabbix proxy (SQLite3)** - Very simple way to gain power of Zabbix proxy.
          The feature has  SQLite3 support for Zabbix proxies and allow to specify
          amount of proxies. \n* **Zabbix proxy (MySQL)** - Another option of Zabbix
          proxy. The option support and deliver MySQL database.\n* **Zabbix agent**
          - Zabbix agent can be deployed on each available node for stability and
          performance monitoring on remote nodes. It allows to gather metrics with
          full automation!* \n## Prerequisites\nAll deployment options are require
          additional information during deployment. Please, check the following instructions
          and provide required configuration:\n* **Zabbix Server** - MySQL database
          host information and MySQL database credentials in specially formatted *Secret*.
          Additionally it is possible to specify SSL certificates for HTTPS support
          in *Secret*.\n* **Zabbix Server (Full)** - MySQL database credentials in
          specially formatted *Secret*. MySQL database volume name information. Additionally
          it is possible to specify SSL certificates for HTTPS support in *Secret*.\n*
          **Zabbix proxy (SQLite3)** - Zabbix server host information only.\n* **Zabbix
          proxy (MySQL)** - MySQL database credentials in specially formatted *Secrets*
          and Zabbix server host information.\n* **Zabbix agent** - Zabbix server
          host information only for outgoing and incoming connections. Zabbix agent
          uses \"privileged\" mode to monitor node resources! For example, running
          processes.* \n### MySQL credentials\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-full-secrets\ndata:\n  mysql_root_password: emFiYml4X3N1cGVyX3Jvb3Q=
          [1]\n  mysql_zabbix_password: emFiYml4X3VzZXJfcGFzc3dvcmQ= [2]\n  mysql_zabbix_username:
          emFiYml4 [3]\ntype: Opaque\n```\nUsing MySQL root password (*mysqlrootpassword*
          [1]) Zabbix server / proxy will try to create MySQL database schema with
          grant permissions to *mysqlzabbixusername* [2] and *mysqlzabbixpassword*
          [3].\n### SSL certificates for HTTPS\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-web-sslsecret\ndata:\n  ssl.crt: >-\n   < ssl.crt data>\n
          \ ssl.key: >-\n\t < ssl.key data >\n  dhparam.pem: >-\n   <  dhparam.pem
          data >\n```\nFiles *ssl.crt*, *ssl.key* and *dhparam.perm* are required
          for Zabbix web-interface for SSL support.\n"
        displayName: Zabbix Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - zabbix
        - appliance
        - agent
        - metrics
        links:
        - name: Zabbix
          url: https://www.zabbix.com
        - name: Zabbix Official Documentation
          url: https://www.zabbix.com/documentation/6.0/en/manual/quickstart
        - name: Zabbix Operator Installation
          url: https://www.zabbix.com/documentation/6.0/en/manual/installation/containers/openshift
        - name: Downloads
          url: https://www.zabbix.com/download
        maintainers:
        - email: alexey.pustovalov@zabbix.com
          name: Alexey Pustovalov
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Zabbix LLC
        relatedImages:
        - registry.connect.redhat.com/zabbix/zabbix-agent2-60@sha256:424003d52f3cb4360adf2b3619d8b9f8de9196c630ee7e4ceb5d177c1c5a36e4
        - registry.connect.redhat.com/zabbix/zabbix-server-mysql-60@sha256:637484faec6e9a55049c13dbd1f100c03d757c3a21749388c0605040f02d05ce
        - registry.connect.redhat.com/zabbix/zabbix-proxy-mysql-60@sha256:c2ae74d97c083ff348dc9c55932a33da4779cd9efdd9d7c5a79acfffae5b9ed0
        - registry.connect.redhat.com/zabbix/zabbix-java-gateway-60@sha256:a4feb725c12c4ac90eb9244c6fafdb6708679f2e0e7f6aa683c765127306412e
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
        - registry.connect.redhat.com/zabbix/zabbix-web-mysql-60@sha256:611d2670e1bf95b81226976c0a1f64434b2a7766f196dadbd646a7d8319f4cb1
        - registry.connect.redhat.com/zabbix/zabbix-web-service-60@sha256:4a333094e110713470a897b6406c5e340e962caf32a429dddf19865041fcc448
        - registry.redhat.io/rhel9/mysql-80@sha256:c4b5246d0c0d089a766ea07768c22d9d6be8049ac1379bd345db99390ffe2a36
        - registry.connect.redhat.com/zabbix/zabbix-snmptraps-60@sha256:e61e6b1c551dec8488cce0998f94d83d6d10b5d89a465d7119b088313032cb81
        - registry.connect.redhat.com/zabbix/zabbix-proxy-sqlite-60@sha256:62e62ebc8c1bc13fb610f7a83412e65d852f74b3c2c212baa75ea27fc7fa03eb
        - registry.connect.redhat.com/zabbix/zabbix-agent-60@sha256:2895b5e45785aaf528d4a6f78c8af0a0d494dbe9362403fdd470673d911d5004
        - quay.io/zabbix/zabbix-operator@sha256:43d26488a19967bdb2a8aafde5f43820be95011811da9cc5ec623c9029801675
        version: 6.0.40
      entries:
      - name: zabbix-operator-certified-rhmp.v6.0.40
        version: 6.0.40
      - name: zabbix-operator-certified-rhmp.v6.0.38
        version: 6.0.38
      - name: zabbix-operator-certified-rhmp.v6.0.36
        version: 6.0.36
      - name: zabbix-operator-certified-rhmp.v6.0.33
        version: 6.0.33
      - name: zabbix-operator-certified-rhmp.v6.0.29
        version: 6.0.29
      - name: zabbix-operator-certified-rhmp.v6.0.27
        version: 6.0.27
      - name: zabbix-operator-certified-rhmp.v6.0.26
        version: 6.0.26
      - name: zabbix-operator-certified-rhmp.v6.0.25
        version: 6.0.25
      - name: zabbix-operator-certified-rhmp.v6.0.17
        version: 6.0.17
      - name: zabbix-operator-certified-rhmp.v6.0.14
        version: 6.0.14
      - name: zabbix-operator-certified-rhmp.v6.0.13
        version: 6.0.13
      name: lts
    defaultChannel: lts
    packageName: zabbix-operator-certified-rhmp
    provider:
      name: Zabbix LLC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/streamshub
    name: amq-streams-console
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: alpha
    - currentCSV: amq-streams-console.v2.8.0-13-0.1738265629.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:1b1b80710885b43600e22ff2a24930b5cab06cd91fad439ad8a282c00b4564b9
          createdAt: "2024-11-12T15:34:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-streams-console.v2.8.0-13
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.8
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.23.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:1b1b80710885b43600e22ff2a24930b5cab06cd91fad439ad8a282c00b4564b9
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:38b80d81e54d59667c1f7dfa4da4125d5c7430816da130ddc84678a4e3cff6cf
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:366d60ff48fb9ba32dc73d92a775046d7b1eb1df21c93e31605372c7858f16ca
        version: 2.8.0-13+0.1738265629.p
      entries:
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.8.x
    - currentCSV: amq-streams-console.v2.9.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
          createdAt: "2025-09-22T10:41:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0-0 <2.9.2-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.9
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:839368767bec735cd12f002874917c23f58550fe21df976858bdc92946985ec6
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:b7444b70a5dc4543a9976c08a619ff6978c9ff86ac7085d300b4870b7c6b7851
        version: 2.9.2-2
      entries:
      - name: amq-streams-console.v2.9.2-2
        version: 2.9.2-2
      - name: amq-streams-console.v2.9.2-1
        version: 2.9.2-1
      - name: amq-streams-console.v2.9.1-7
        version: 2.9.1-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.9.x
    - currentCSV: amq-streams-console.v2.9.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
          createdAt: "2025-09-22T10:41:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0-0 <2.9.2-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.9
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:839368767bec735cd12f002874917c23f58550fe21df976858bdc92946985ec6
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:b7444b70a5dc4543a9976c08a619ff6978c9ff86ac7085d300b4870b7c6b7851
        version: 2.9.2-2
      entries:
      - name: amq-streams-console.v2.9.2-2
        version: 2.9.2-2
      - name: amq-streams-console.v2.9.2-1
        version: 2.9.2-1
      - name: amq-streams-console.v2.9.1-7
        version: 2.9.1-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.x
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-3.0.x
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-3.x
    defaultChannel: alpha
    packageName: amq-streams-console
    provider:
      name: Red Hat
      url: https://github.com/streamshub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: 3scale-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: 3scale-operator.v0.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "ProductCRName": "product1-sample",
                  "Production": true
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/3scale-rhel7-operator@sha256:0751f9f5343a1264224c36db3db292e5a9e28c40cd96649ef48b5399c34668dd
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.16"}]'
          olm.skipRange: '>=0.10.0 <0.11.13'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/3scale-rhel7-operator@sha256:0751f9f5343a1264224c36db3db292e5a9e28c40cd96649ef48b5399c34668dd
        - registry.redhat.io/3scale-amp2/backend-rhel8@sha256:cdc53799ff9d022462a9a4afcbcd8043497cd40ccdc6fafdcb9ccf7ce69f43d1
        - registry.redhat.io/3scale-amp2/system-rhel7@sha256:61dcc91e5d3c91a0ea85c02f2599cadcb5496b1c09a8838b4e1c8e4efe126e8f
        - registry.redhat.io/3scale-amp2/memcached-rhel7@sha256:7e2ab1e24646fbfef878eb2b5a74ed865ae812b60aa0ba505045f838f26ebbfc
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:a95d09fc3b224f550dec3de3d23fd2dbfc0a220dc869b4ad9559ee2f85327daa
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:f8cd25658fa6a154e7ea6e5adf0f9dae3fc596ec9989021de6a021ec94eef226
        - registry.redhat.io/rhel8/redis-6@sha256:5250d644e80aea794b0f732e9285883a4348dace83557f177f64d78fe4dcabd0
        - registry.redhat.io/openshift4/ose-cli@sha256:1a8caae40be21f23ba3cb48b75809a73333907033a6041927aa568b5c9290f3c
        - registry.redhat.io/3scale-amp2/searchd-rhel7@sha256:8e2001c021fad554cb8a96bb34e9973010fe4206526f19aa040fa196709fb688
        - registry.redhat.io/3scale-amp2/zync-rhel9@sha256:0d2205d9f61a58414c4ba472ce8f19b12d2df722772847d5fb7442e664759627
        - registry.redhat.io/rhel8/mysql-80@sha256:08547a93659ad804d637b91e8e1eb3464b75783d494e6c40e1c921bd74f31645
        version: 0.11.13
      entries:
      - name: 3scale-operator.v0.11.13
        version: 0.11.13
      - name: 3scale-operator.v0.11.12
        version: 0.11.12
      name: threescale-2.14
    - currentCSV: 3scale-operator.v0.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                },
                "status": {
                  "deployments": {
                    "ready": [
                      "deployment x ready"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                },
                "status": {
                  "adminId": 1,
                  "tenantId": 2
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ApplicationAuth",
                "metadata": {
                  "name": "applicationauth-sample"
                },
                "spec": {
                  "applicationCRName": "application-cr-name",
                  "authSecretRef": {
                    "name": "auth-secret-reference"
                  },
                  "generateSecret": true
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "productCRName": "product1-sample",
                  "production": true
                },
                "status": {}
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/3scale-rhel9-operator@sha256:fcea6a77d0a48f35f4c3ce63e72dde104aac6b580f340950b311eb418fff5ec9
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=0.11.0 <0.12.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          rht_backend_redis_requirements: 6.2.0
          rht_mysql_requirements: 8.0.0
          rht_postgres_requirements: 10.0.0
          rht_system_redis_requirements: 6.2.0
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: ApplicationAuth is the Schema for the applicationauths API
            displayName: Application Auth
            kind: ApplicationAuth
            name: applicationauths.capabilities.3scale.net
            version: v1beta1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/3scale-rhel9-operator@sha256:fcea6a77d0a48f35f4c3ce63e72dde104aac6b580f340950b311eb418fff5ec9
        - registry.redhat.io/3scale-amp2/manticore-rhel9@sha256:ccf6baf5f2e9494a34739a79abe0f46e36d8ab9dc81c426f7a3a8964e8f39859
        - registry.redhat.io/3scale-amp2/backend-rhel8@sha256:eb2288de900dc42ead755b28cdb2509b25ecc8dd256781beaa4a2a86e7cbe532
        - registry.redhat.io/rhel9/memcached@sha256:8738110ef16c9e35deeb07c0577ffe09d6ef029761b504a8d20253fa715da394
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:a95d09fc3b224f550dec3de3d23fd2dbfc0a220dc869b4ad9559ee2f85327daa
        - registry.redhat.io/3scale-amp2/zync-rhel9@sha256:df9d62897ff286165772150a8b5e6952d2ce3b52a8afa07bd693a2d95c2c67ff
        - registry.redhat.io/openshift4/ose-cli@sha256:1a8caae40be21f23ba3cb48b75809a73333907033a6041927aa568b5c9290f3c
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/rhel8/mysql-80@sha256:ee9aca9096473da004f9140305742eb7b9a77ed81ce55d5a6503388cccab030f
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        - registry.redhat.io/3scale-amp2/system-rhel8@sha256:4c72dfeb1919a1de4fbc47f4611d0c830ef89cf9f17255ad55a9b370618e0e8f
        version: 0.12.5
      entries:
      - name: 3scale-operator.v0.12.5
        version: 0.12.5
      - name: 3scale-operator.v0.12.4
        version: 0.12.4
      - name: 3scale-operator.v0.12.3
        version: 0.12.3
      - name: 3scale-operator.v0.12.2
        version: 0.12.2
      - name: 3scale-operator.v0.12.1
        version: 0.12.1
      - name: 3scale-operator.v0.12.0
        version: 0.12.0
      name: threescale-2.15
    - currentCSV: 3scale-operator.v0.12.1-mas
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                },
                "status": {
                  "deployments": {
                    "ready": [
                      "deployment x ready"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                },
                "status": {
                  "adminId": 1,
                  "tenantId": 2
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ApplicationAuth",
                "metadata": {
                  "name": "applicationauth-sample"
                },
                "spec": {
                  "applicationCRName": "application-cr-name",
                  "authSecretRef": {
                    "name": "auth-secret-reference"
                  },
                  "generateSecret": true
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "productCRName": "product1-sample",
                  "production": true
                },
                "status": {}
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-mas/3scale-rhel9-operator@sha256:8f41ab6027616e4bb65b1fddbcf44cc8a25a79cdda19d5cc354ac4228a15141e
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.17"}]'
          olm.skipRange: '>=0.11.8 <0.12.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          rht_backend_redis_requirements: 6.2.0
          rht_mysql_requirements: 8.0.0
          rht_postgres_requirements: 10.0.0
          rht_system_redis_requirements: 6.2.0
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: ApplicationAuth is the Schema for the applicationauths API
            displayName: Application Auth
            kind: ApplicationAuth
            name: applicationauths.capabilities.3scale.net
            version: v1beta1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-mas/3scale-manticore-search-rhel9@sha256:7379186175c9c0f0a6d922cd2a0754b429c9728dfbe8979a9d6025789ebae0da
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:a95d09fc3b224f550dec3de3d23fd2dbfc0a220dc869b4ad9559ee2f85327daa
        - registry.redhat.io/3scale-mas/backend-rhel8@sha256:d0c561c6b9dfe9741722ba4a257fbd2df4d2a709457a0b0657ef7bcca1f2a5ca
        - registry.redhat.io/3scale-mas/zync-rhel9@sha256:78e5223a1a878f6527bd88d382948c125b7cb4a81540648244553af426d1a944
        - registry.redhat.io/3scale-mas/memcached-rhel7@sha256:7a1abee11b4120f399485b438bda1e5262f04de28d3e788b207eecc3a9de9ed2
        - registry.redhat.io/3scale-mas/3scale-rhel9-operator@sha256:8f41ab6027616e4bb65b1fddbcf44cc8a25a79cdda19d5cc354ac4228a15141e
        - registry.redhat.io/rhel8/mysql-80@sha256:754f6830e0bc4aadb973b1f613f2bc7278ec13ad0f0e1e87655a2bb2f76cb9b3
        - registry.redhat.io/openshift4/ose-cli@sha256:a688994ea404ce6ce4bb5ec1249f0e779de0d6c187b5078e2ff067e60b50937d
        - registry.redhat.io/3scale-mas/apicast-gateway-rhel8@sha256:f174636a9ad43a2419fff5a9e6797e3185c540a9509188fd1606aef4acea7bdf
        - registry.redhat.io/rhel8/redis-6@sha256:218f7fa2458c274cc34f79190423e0aba03027f55ef6b3334d916951a668b7b7
        - registry.redhat.io/3scale-mas/system-rhel8@sha256:0fa28aaf3467fbd22fc1335dc6a712043e4dead694c3f14b60e89740bb7eb13e
        version: 0.12.1-mas
      entries:
      - name: 3scale-operator.v0.12.1-mas
        version: 0.12.1-mas
      - name: 3scale-operator.v0.11.8-mas
        version: 0.11.8-mas
      - name: 3scale-operator.v0.11.7-mas
        version: 0.11.7-mas
      - name: 3scale-operator.v0.11.6-mas
        version: 0.11.6-mas
      - name: 3scale-operator.v0.11.5-mas
        version: 0.11.5-mas
      - name: 3scale-operator.v0.11.3-mas
        version: 0.11.3-mas
      - name: 3scale-operator.v0.11.2-mas
        version: 0.11.2-mas
      - name: 3scale-operator.v0.11.1-mas
        version: 0.11.1-mas
      - name: 3scale-operator.v0.11.0-mas
        version: 0.11.0-mas
      - name: 3scale-operator.v0.10.0-mas
        version: 0.10.0-mas
      name: threescale-mas
    defaultChannel: threescale-2.15
    packageName: 3scale-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: operator-certification-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operator-certification-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "certification.redhat.com/v1alpha1",
                "kind": "OperatorPipeline",
                "metadata": {
                  "name": "operatorpipeline-sample"
                },
                "spec": {
                  "applyCIPipeline": true,
                  "applyHostedPipeline": false,
                  "applyReleasePipeline": false,
                  "gitHubSecretName": "github-api-token",
                  "kubeconfigSecretName": "kubeconfig",
                  "operatorPipelinesRelease": "main",
                  "pyxisSecretName": "pyxis-api-secret"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-06-23T16:02:00Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io.bundle.channel.default.v1: alpha
          operators.operatorframework.io.bundle.channels: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-openshift-ecosystem/operator-certification-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OperatorPipeline is the Schema for the operatorpipelines
              API
            displayName: Operator Pipeline
            kind: OperatorPipeline
            name: operatorpipelines.certification.redhat.com
            version: v1alpha1
        description: |-
          A Kubernetes operator to provision resources for the operator certification pipeline. This operator is installed in all namespaces which can support multi-tenant scenarios. **Note:** This operator should only be used by Red Hat partners attempting to certify their operator(s).


          **Requirements**


          The certification operator requires that you have the following tools installed, functional, and in your path.
          - [Install](https://docs.openshift.com/container-platform/4.8/cli_reference/openshift_cli/getting-started-cli.html#installing-openshift-cli) oc, the OpenShift CLI tool (tested with version 4.7.13)
          - [Install](https://tekton.dev/docs/cli/) tkn, the Tekton CLI tool (tested with version 0.19.1)
          - [Install](https://git-scm.com/downloads) git, the Git CLI tool (tested with 2.32.0)
          - The certification pipeline expects you to have the source files of your Operator bundle


          **Pre - Installation**


          The below steps exist for simplicity and to make the installation clearer. The operator watches all namespaces, so if secrets/configs/etc already exist in another namespace, feel free to use the existing namespace when following the operator installation steps.


          **Create a new namespace where the following secrets will be applied.**
          ```
          oc new-project oco
          ```


          **Add the kubeconfig secret which will be used to deploy the operator under test and run the certification checks.**
          * Open a terminal window
          * Set the `KUBECONFIG` environment variable
          ```
          export KUBECONFIG=/path/to/your/cluster/kubeconfig
          ```
          > *This kubeconfig will be used to deploy the Operator under test and run the certification checks.*
          ```
          oc create secret generic kubeconfig --from-file=kubeconfig=$KUBECONFIG
          ```


          **Configuring steps for submitting the results**
          - Add the github API token to the repo where the PR will be created
          ```
          oc create secret generic github-api-token --from-literal GITHUB_TOKEN=<github token>
          ```
          - Add RedHat Container API access key

            This API access key is specifically related to your unique partner account for Red Hat Connect portal. Instructions to obtain your API key can be found: [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/operator-cert-workflow.md#step-b---get-api-key)
          ```
          oc create secret generic pyxis-api-secret --from-literal pyxis_api_key=< API KEY >
          ```

          - Optional pipeline configurations can be found [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md#optional-configuration)


          **Execute the Pipeline (Development Iterations)**


          A pre-requisite to running a pipeline is that a `workspace-template.yaml` exists in the directory you want to execute the `tkn` commands from.

          To create a workspace-template.yaml
          ```
          cat <<EOF > workspace-template.yaml
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 5Gi
          EOF
          ```

          There are multiple ways to execute the Pipeline which can be found [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md#execute-the-pipeline-development-iterations)
        displayName: Operator Certification Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - certification
        - pipeline
        - openshift
        links:
        - name: Operator Certification CI Pipeline
          url: https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md
        - name: Red Hat OpenShift Pipelines
          url: https://github.com/openshift/tektoncd-operator
        maintainers:
        - email: certification-operator-devel@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/opdev/operator-certification-operator@sha256:7c84772f96650ff757c2552ba421355d54da41a5ed2155f72a14911b8a3f18ff
        version: 1.3.1
      entries:
      - name: operator-certification-operator.v1.3.1
        version: 1.3.1
      name: alpha
    defaultChannel: alpha
    packageName: operator-certification-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: LINBIT
      provider-url: https://linbit.com
    name: linstor-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: linstor-operator.v2.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorCluster",
                "metadata": {
                  "name": "linstorcluster"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorNodeConnection",
                "metadata": {
                  "name": "example-connection"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatellite",
                "metadata": {
                  "name": "example-satellite"
                },
                "spec": {
                  "clusterRef": {
                    "name": "linstorcluster"
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatelliteConfiguration",
                "metadata": {
                  "name": "all-satellites"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          com.redhat.openshift.versions: v4.7
          containerImage: drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
          createdAt: "2025-09-25T13:44:22Z"
          description: LINSTOR Operator deploys LINBIT SDS in the cluster.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: linbit-sds
          operators.openshift.io/valid-subscription: '[''LINBIT SDS Subscription'']'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/internal-objects: '["linstorsatellites.piraeus.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster-wide configuration for LINSTOR.
            displayName: Linstor Cluster
            kind: LinstorCluster
            name: linstorclusters.piraeus.io
            version: v1
          - description: Configuration for connections between specific nodes.
            displayName: Linstor Node Connection
            kind: LinstorNodeConnection
            name: linstornodeconnections.piraeus.io
            version: v1
          - description: Configuration for specific Satellites/Nodes in the LINSTOR
              Cluster.
            displayName: Linstor Satellite Configuration
            kind: LinstorSatelliteConfiguration
            name: linstorsatelliteconfigurations.piraeus.io
            version: v1
          - description: Resource to configure a single satellite, managed by the
              Operator itself.
            displayName: Linstor Satellite
            kind: LinstorSatellite
            name: linstorsatellites.piraeus.io
            version: v1
          required:
          - kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
        description: |
          The LINSTOR Operator manages [LINSTOR](https://github.com/LINBIT/linstor-server) clusters in Kubernetes.

          All components of the LINSTOR software stack can be managed by the operator:
          * DRBD
          * LINSTOR
          * LINSTOR CSI driver
          * LINSTOR HA Controller
        displayName: LINSTOR Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: User Guide
          url: https://linbit.com/drbd-user-guide/linstor-guide-1_0-en/
        - name: LINSTOR Operator
          url: https://charts.linstor.io/
        - name: Source Code
          url: https://github.com/LINBIT/linstor-operator-builder
        maintainers:
        - email: support@linbit.com
          name: LINBIT
        maturity: stable
        minKubeVersion: 1.20.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: DaemonSet
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: storage.k8s.io
          kind: CSIDriver
          version: v1
        provider:
          name: LINBIT
          url: https://linbit.com
        relatedImages:
        - drbd.io/drbd9-rhel8@sha256:7cddfa770de66cefcb42c2b33da57d0f1adfcc5f1cc12522977aa6993722bb2a
        - drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ac3963bfd4246da386b82d810a348e9d42a9e4b746acaf4dbcbe0766643cc2fa
        - drbd.io/drbd9-rhel10@sha256:da0a7015cf15724a7c73100d505891e62262a726141347df7f2d51ec79a0a1e1
        - drbd.io/linstor-k8s-ha-controller@sha256:ed0f0f86dad88feadf8c3c32d7d143eaa52b1130af3576315a562909b39d2afa
        - drbd.io/linstor-controller@sha256:841b3aa84cdf13aaaf1243bc5da8ac2205ca3fe0186fdfc7a04b69397f0020d6
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:2669cde3db2268d37bc671f494eaa06480478a90b8943054af14cbffc050debf
        - drbd.io/drbd9-rhel9@sha256:4eb553336ed64e127068a4566370b040166bf11301e21ac47ac5b92237ec9d2a
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:624f61b1ae9d32f5a074ee8c12839d975fba9efcd3bfa7981cfe38febf5c4495
        - drbd.io/linstor-csi@sha256:9c7ef452b0997d3373803201614c3b8075687e75c397359738a9e4ef5530cab3
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:700be620923fc99ed03f9f1c4a911e4084beb30685361270c49165b27ad1d374
        - drbd.io/linstor-affinity-controller@sha256:78e5b299ef1702114528678387b5e74821d7658bc02f05f0bd7edd9a176fba39
        - drbd.io/drbd-shutdown-guard@sha256:7ba049ecb205a0086e425d2ff37285b00449549d89e179b4b943ace6e59a9156
        - drbd.io/linstor-satellite@sha256:ce0a0b05494c6e6d5f16121e5762bf58e7b78f40c00b84780e2abba2a690af1b
        - drbd.io/drbd-reactor@sha256:77c51aeeefdcd788cb1583738f93778032fcc03c36fb289b692d3cf195c8b184
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:5a650dc21a6a038de6267f117d01adfd904ff3f5c94f428f6a13735f18bf94d6
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:cd58f1789027e32eae08f00dac7b2275496c41020fb7cdebc10e9986c625d4f4
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:f8cde5b5c289f7b3d233773f4e79990ba392fe6e14e0177de80714e7d3e98862
        - drbd.io/drbd9-rhel7@sha256:0c96b216e55a128a9944276af9080aced8bd72fdf1fd424f024ed9ace2c1feb5
        - drbd.io/ktls-utils@sha256:3351a6e0fc048fe1544499095cdb8fdd221e832cecfa8a9a01f046c271991138
        version: 2.9.1
      entries:
      - name: linstor-operator.v2.9.1
        version: 2.9.1
      name: dev
    - currentCSV: linstor-operator.v2.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorCluster",
                "metadata": {
                  "name": "linstorcluster"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorNodeConnection",
                "metadata": {
                  "name": "example-connection"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatellite",
                "metadata": {
                  "name": "example-satellite"
                },
                "spec": {
                  "clusterRef": {
                    "name": "linstorcluster"
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatelliteConfiguration",
                "metadata": {
                  "name": "all-satellites"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          com.redhat.openshift.versions: v4.7
          containerImage: drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
          createdAt: "2025-09-25T13:44:22Z"
          description: LINSTOR Operator deploys LINBIT SDS in the cluster.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: linbit-sds
          operators.openshift.io/valid-subscription: '[''LINBIT SDS Subscription'']'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/internal-objects: '["linstorsatellites.piraeus.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster-wide configuration for LINSTOR.
            displayName: Linstor Cluster
            kind: LinstorCluster
            name: linstorclusters.piraeus.io
            version: v1
          - description: Configuration for connections between specific nodes.
            displayName: Linstor Node Connection
            kind: LinstorNodeConnection
            name: linstornodeconnections.piraeus.io
            version: v1
          - description: Configuration for specific Satellites/Nodes in the LINSTOR
              Cluster.
            displayName: Linstor Satellite Configuration
            kind: LinstorSatelliteConfiguration
            name: linstorsatelliteconfigurations.piraeus.io
            version: v1
          - description: Resource to configure a single satellite, managed by the
              Operator itself.
            displayName: Linstor Satellite
            kind: LinstorSatellite
            name: linstorsatellites.piraeus.io
            version: v1
          required:
          - kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
        description: |
          The LINSTOR Operator manages [LINSTOR](https://github.com/LINBIT/linstor-server) clusters in Kubernetes.

          All components of the LINSTOR software stack can be managed by the operator:
          * DRBD
          * LINSTOR
          * LINSTOR CSI driver
          * LINSTOR HA Controller
        displayName: LINSTOR Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: User Guide
          url: https://linbit.com/drbd-user-guide/linstor-guide-1_0-en/
        - name: LINSTOR Operator
          url: https://charts.linstor.io/
        - name: Source Code
          url: https://github.com/LINBIT/linstor-operator-builder
        maintainers:
        - email: support@linbit.com
          name: LINBIT
        maturity: stable
        minKubeVersion: 1.20.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: DaemonSet
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: storage.k8s.io
          kind: CSIDriver
          version: v1
        provider:
          name: LINBIT
          url: https://linbit.com
        relatedImages:
        - drbd.io/drbd-shutdown-guard@sha256:7ba049ecb205a0086e425d2ff37285b00449549d89e179b4b943ace6e59a9156
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ac3963bfd4246da386b82d810a348e9d42a9e4b746acaf4dbcbe0766643cc2fa
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:624f61b1ae9d32f5a074ee8c12839d975fba9efcd3bfa7981cfe38febf5c4495
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:cd58f1789027e32eae08f00dac7b2275496c41020fb7cdebc10e9986c625d4f4
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:f8cde5b5c289f7b3d233773f4e79990ba392fe6e14e0177de80714e7d3e98862
        - drbd.io/drbd9-rhel10@sha256:da0a7015cf15724a7c73100d505891e62262a726141347df7f2d51ec79a0a1e1
        - drbd.io/linstor-affinity-controller@sha256:78e5b299ef1702114528678387b5e74821d7658bc02f05f0bd7edd9a176fba39
        - drbd.io/linstor-csi@sha256:9c7ef452b0997d3373803201614c3b8075687e75c397359738a9e4ef5530cab3
        - drbd.io/drbd9-rhel9@sha256:4eb553336ed64e127068a4566370b040166bf11301e21ac47ac5b92237ec9d2a
        - drbd.io/drbd9-rhel7@sha256:0c96b216e55a128a9944276af9080aced8bd72fdf1fd424f024ed9ace2c1feb5
        - drbd.io/linstor-satellite@sha256:ce0a0b05494c6e6d5f16121e5762bf58e7b78f40c00b84780e2abba2a690af1b
        - drbd.io/drbd-reactor@sha256:77c51aeeefdcd788cb1583738f93778032fcc03c36fb289b692d3cf195c8b184
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:2669cde3db2268d37bc671f494eaa06480478a90b8943054af14cbffc050debf
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:5a650dc21a6a038de6267f117d01adfd904ff3f5c94f428f6a13735f18bf94d6
        - drbd.io/ktls-utils@sha256:3351a6e0fc048fe1544499095cdb8fdd221e832cecfa8a9a01f046c271991138
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:700be620923fc99ed03f9f1c4a911e4084beb30685361270c49165b27ad1d374
        - drbd.io/drbd9-rhel8@sha256:7cddfa770de66cefcb42c2b33da57d0f1adfcc5f1cc12522977aa6993722bb2a
        - drbd.io/linstor-controller@sha256:841b3aa84cdf13aaaf1243bc5da8ac2205ca3fe0186fdfc7a04b69397f0020d6
        - drbd.io/linstor-k8s-ha-controller@sha256:ed0f0f86dad88feadf8c3c32d7d143eaa52b1130af3576315a562909b39d2afa
        - drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
        version: 2.9.1
      entries:
      - name: linstor-operator.v2.9.1
        version: 2.9.1
      name: stable
    defaultChannel: dev
    packageName: linstor-operator
    provider:
      name: LINBIT
      url: https://linbit.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: https://opdev.github.io/
    name: loki-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.example.com/v1alpha1",
                "kind": "Loki",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/install-disable-crds": "true"
                  },
                  "name": "loki-sample"
                },
                "spec": {
                  "global": {
                    "clusterDomain": "cluster.local",
                    "dnsNamespace": "openshift-dns",
                    "dnsService": "dns-default"
                  },
                  "loki": {
                    "auth_enabled": false,
                    "commonConfig": {
                      "replication_factor": 1
                    },
                    "storage": {
                      "type": "filesystem"
                    }
                  },
                  "monitoring": {
                    "dashboards": {
                      "enabled": false
                    },
                    "lokiCanary": {
                      "enabled": false
                    },
                    "rules": {
                      "alerting": false,
                      "enabled": false
                    },
                    "selfMonitoring": {
                      "enabled": false,
                      "grafanaAgent": {
                        "installOperator": false
                      }
                    },
                    "serviceMonitor": {
                      "enabled": false,
                      "metricsInstance": {
                        "enabled": false
                      }
                    }
                  },
                  "rbac": {
                    "sccEnabled": true
                  },
                  "singleBinary": {
                    "replicas": 1
                  },
                  "test": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/loki-helm-operator-controller:0.0.1
          createdAt: "2023-07-14T15:06:17Z"
          description: An experimental operator that installs loki from its official
            helmchart.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys the loki operator helm chart
            displayName: Loki
            kind: Loki
            name: lokis.charts.example.com
            version: v1alpha1
        description: Loki operator based on the helm chart with single binary mode
          deployment optionality
        displayName: Loki Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loki
        - single-binary
        - demo
        - helm
        links:
        - name: Loki Helm Operator
          url: https://github.com/skattoju/loki-helm-operator
        maintainers:
        - email: skattoju@redhat.com
          name: Sid Kattoju
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: https://opdev.github.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/opdev/loki-helm-operator-controller:0.0.1
        version: 0.0.1
      entries:
      - name: loki-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: loki-helm-operator
    provider:
      name: opdev
      url: https://opdev.github.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat CCIT
      provider-url: ""
    name: reportportal-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: reportportal-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ReportPortal",
                "metadata": {
                  "name": "reportportal-sample"
                },
                "spec": {
                  "accesskey": "AKIAIOSFODNN7EXAMPLE",
                  "analyzer_image": "quay.io/waynesun09/service-auto-analyzer:5.3.5",
                  "api_image": "quay.io/waynesun09/service-api:5.3.5-rootless",
                  "api_replicas": 1,
                  "app_domain": "apps.test-example.com",
                  "bucket_prefix": null,
                  "default_bucket_name": null,
                  "enable_pg_restic_backup": "yes",
                  "es_backup_schedule": "@daily",
                  "es_s3_backup_dir": "s3_backup",
                  "es_snapshot_bucket": "es-snapshot-123123",
                  "gateway_replicas": 1,
                  "index_image": "quay.io/waynesun09/service-index:5.0.10",
                  "index_replicas": 1,
                  "integration_salt_path": null,
                  "migration_image": "quay.io/waynesun09/migrations:5.3.5",
                  "pg_restic_password": "rp_user",
                  "pg_restic_s3_bucket_init": "yes",
                  "pg_s3_bucket_name": "pgbackup-123123",
                  "s3_enabled": false,
                  "s3_endpoint": null,
                  "secretkey": "wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY",
                  "uat_image": "quay.io/waynesun09/service-authorization:5.3.5",
                  "ui_image": "quay.io/waynesun09/service-ui:5.3.5",
                  "ui_replicas": 1
                }
              },
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ReportPortalRestore",
                "metadata": {
                  "name": "reportportalrestore-sample"
                },
                "spec": {
                  "accesskey": "AKIAIOSFODNN7EXAMPLE",
                  "bucket_prefix": null,
                  "es_restore": "yes",
                  "es_restore_date": "2020-08-07",
                  "es_s3_backup_dir": "s3_backup",
                  "es_snapshot_bucket": "es-snapshot-123123",
                  "pg_recovery": "yes",
                  "pg_recovery_target_time": "2020-08-07 00:00:00",
                  "pg_restic_password": "rp_user",
                  "pg_s3_bucket_name": "pgbackup-123123",
                  "s3_endpoint": null,
                  "secretkey": "wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY"
                }
              },
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ServiceApiServiceMonitor",
                "metadata": {
                  "name": "serviceapiservicemonitor-sample"
                },
                "spec": {
                  "bearer_token_secret_key": "token",
                  "bearer_token_secret_name": "rp-user-auth"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/waynesun09/rp5-operator:v0.0.8
          createdAt: "2021-04-22 14:39:00"
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/waynesun09/reportportal-operator
          support: waynesun09
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ReportPortal Restore from s3 backup
            kind: ReportPortalRestore
            name: reportportalrestores.rp5.reportportal.io
            version: v1
          - description: ReportPortal v5 Deployment with ReportPortal Operator
            kind: ReportPortal
            name: reportportals.rp5.reportportal.io
            version: v1
          - description: ReportPortal service-api ServiceMonitor for metrics
            kind: ServiceApiServiceMonitor
            name: serviceapiservicemonitors.rp5.reportportal.io
            version: v1
        description: |-
          ReportPortal is a test automation analytics platform. ReportPortal Operator support instantiating the ReportPortal services and required dependencies.

          What services are included:

              - Report Portal v5.3
              - PostgreSQL
              - Elasticsearch
              - RabbitMQ
              - Minio

          The Operator also support data backup & restore and metrics monitoring.

          ## Install

          Install the Operator to a namespace, then create ReportPortal v5 instance under Provided APIs on the UI, all Report Portal v5 services and dependencies will be installed.

          ## Data backup and restore

          The Report Portal v5 data backup and restore in the operator include two parts, PostgreSQL database data and Elasticsearch index data.

          The Operator support automatically data backup and restore for both.

          ### PostgreSQL database WAL archiving and recovery

          The Operator supports PostgreSQL WAL archiving and recovery point-in-time via Minio S3 using restic.

          #### WAL archiving

          When deploy Report Portal v5 instance, make sure `enable_pg_restic_backup` option is set as `yes`, then the WAL archiving will be enabled when initiate the database, the WAL archive data will be automatically backup to S3 storage via Minio using restic command.

          Parameter `pg_restic_s3_bucket_init` is for init the bucket if not exist, `pg_s3_bucket_name` is the bucket name and `pg_restic_password` is the restic password for encryption.

          #### WAL recovery

          WAL archive recovery will download all WAL archive files to the Postgres container and restore point-in-time.

          To restore PostgreSQL database point-in-time, on Operator UI select `ReportPortalRestore` under Provided APIs to create ReportPortalRestore operand.

          After the ReportPortalRestore operand created, the PG WAL restore will get started, as restore time will base on the database size, the process will last in minutes, hours or days.

          **Note:** Current Postgres container liveless probe timeout is set as one hour in the postgresql statefulset, it might cause restore fail if restore time is longer than one hour, so it might need be adjusted to be longer.

          ### Elasticsearch snapshot and restore

          The operator supports elasticsearch snapshot and restore with data saved in S3 bucket via Minio, the elasticsearch container have installed the s3 repository plugin.

          #### Snapshot

          When deploy Report Portal v5 instance, make sure following elasticsearch parameters are set:

              es_s3_backup_dir: s3_backup
              es_snapshot_bucket: es-snapshot-123123
              es_backup_schedule: '@daily'

          #### Restore

          Elasticsearch restore support restore to certain date as default snapshot schedule is daily.

          To restore elasticsearch from snapshot, on Operator UI select `ReportPortalRestore` under Provided APIs to create ReportPortalRestore operand.

          The backup dir and bucket name should be same as when create the RP instance, `es_restore` need set to `yes` to enable the restore and `es_restore_date` is the restore target date.

          The restore time will also be depend on the snapshot size.

          ## Metrics

          The operator by default have enabled metrics, the default ansible operator port 8080 is used for the operator container.

          All Report Portal v5 depended services, rabbitmq, minio, postgresql and elasticsearch have metrics enabled in this operator.

          The Report Portal gatway service have metrics enabled.

          By default several ServiceMonitors have been defined when create ReportPortal v5 operand:

              $ oc get servicemonitor
              NAME                    AGE
              elasticsearch-metrics   1d21h
              gateway-metrics         1d21h
              minio-metrics           1d21h
              postgres-metrics        1d21h
              rabbitmq-metrics        1d21h

          To export Report Portal v5 service-api metrics with ServiceMonitor, create a `secret` with RP user api token. The api access token could be found with login to Report Portal -> Click user icon -> Profile.

          Then select `ServiceApiServiceMonitor` under provided APIs to Create Instance on operator UI.

          Input the secret name and key name that have been created, and click create, a api-metrics servicemonitor will be created.

          To check the metrics, if [user workload monitoring](https://docs.openshift.com/container-platform/4.5/monitoring/monitoring-your-own-services.html) is enabled on the cluster, click to Monitoring on OpenShift UI and navigate to Metrics, where you could query the metrics.

          You could also deploy Promethues Operator in the current project and query via it.

          ## Doc

          More detail with installation, data backup & restore and metrics could be found at https://github.com/waynesun09/reportportal-operator#reportportal-operator
        displayName: reportportal-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - reportportal
        - failure analysis
        - test automation
        links:
        - name: Documentation
          url: https://github.com/waynesun09/reportportal-operator/blob/master/README.md
        - name: ReportPortal.io
          url: https://reportportal.io
        maintainers:
        - email: gsun@redhat.com
          name: Wayne Sun
        maturity: alpha
        provider:
          name: Red Hat CCIT
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.7.0
        - quay.io/waynesun09/rp5-operator:v0.0.8
        version: 0.0.8
      entries:
      - name: reportportal-operator.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: reportportal-operator
    provider:
      name: Red Hat CCIT
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://www.redhat.com/
    name: analytics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: analytics-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyData",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalydata-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalydata",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalydata-sample"
                },
                "spec": {
                  "anomalyname": "etcd_object_namespaces",
                  "config": {
                    "max": 100,
                    "min": 60,
                    "query": "max(apiserver_storage_objects{resource=~\"namespaces\"}) by (resource)"
                  },
                  "method": "min_max",
                  "metricdata": {
                    "groupeddata": "{ \"resource\": \"namespaces\" }",
                    "latestvalue": 121,
                    "timestamp": 1695804362
                  }
                }
              },
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalyengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalyengine",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalyengine-sample"
                },
                "spec": {
                  "anomalyqueryconfiguration": "etcd_objects_all:\n  method: percentage_change\n  query: max(apiserver_storage_objects{resource!~\"certificatesigningrequests.certificates.k8s.io|installplans.operators.coreos.com|operators.operators.coreos.com|subscriptions.operators.coreos.com|anomalydata.observability-analytics.redhat.com|anomalyengines.observability-analytics.redhat.com\"}) by (resource)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: True\nfailed_pods:\n  method: min_max\n  query: count(kube_pod_status_phase{phase!~\"Succeeded|Pending|Running|Unknown\"}) by (phase, namespace)\n  min: 0\n  max: 10\n  have_multi_result_data: True\nworkload_memory_usage_bytes:\n  method: percentage_change\n  query: sum(container_memory_working_set_bytes{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nworkload_cpu_usage_cores:\n  method: percentage_change\n  query: sum(container_cpu_usage_seconds_total{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nrest_client_requests_total:\n  method: percentage_change\n  query: sum(rest_client_requests_total)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\n",
                  "cronjobconfig": {
                    "anomalyqueries": "etcd_objects_all,failed_pods,workload_memory_usage_bytes,workload_cpu_usage_cores,rest_client_requests_total",
                    "name": "osa-anomaly-detection",
                    "resource": {
                      "cpulimit": "256m",
                      "cpurequest": "128m",
                      "memorylimit": "512Mi",
                      "memoryrequest": "256Mi"
                    },
                    "schedule": "*/5 * * * *"
                  },
                  "namespace": "osa-anomaly-detection",
                  "serviceaccountrolebinding": {
                    "clusterrolename": "cluster-monitoring-view",
                    "name": "osa-in-cluster-thanos-querier"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
          createdAt: "2024-03-12T08:02:25Z"
          description: Deploys and Manages Observability Analytics Operator
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4-alpha
          repository: https://github.com/openshift/analytics-operator
          support: RedHat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnomalyData is the Schema for the anomalydata API
            displayName: Anomaly Data
            kind: AnomalyData
            name: anomalydata.observability-analytics.redhat.com
            version: v1alpha1
          - description: AnomalyEngine is the Schema for the anomalyengines API
            displayName: Anomaly Engine
            kind: AnomalyEngine
            name: anomalyengines.observability-analytics.redhat.com
            version: v1alpha1
        description: "### Introduction\nObservability Analytics for Red Hat OpenShift
          runs OpenShift analytics processes like anomaly detection.\n\n### API Backward
          Compatibility\n**NOTE:** Until the Operator supports **seamless upgrades**,
          a new version of the operator may introduce a change that is **NOT** backward-compatible
          with the previous version. Thus, to upgrade the operator, uninstall the
          already installed version completely (including CRDs) and install the new
          version.\n\n### Installtion Steps for the Anomaly Engine\n Create namespace
          if you would like to install this operator into separate namespace (You
          can utilize existing namespace as well). ```bash kubectl create ns <Namespace>
          ``` Install the Observability Analytics operator to specific `<Namespace>`,
          wait until you see the operator in the **Installed Operator** list. Click
          on the Operator name and go to Operator details.\n\nHere you should see
          a tab called **Anomaly Engine**. Click on it, and then click on the **Create
          AnomalyEngine** button, which will redirect you to a form where you need
          to provide the required inputs for the Anomaly Engine. Once successfully
          created, you should see a cronjob in the given namespace. This cronjob will
          identify anomalies and store them into **anomalydata**.\n\n### How to See
          Anomaly Data Captured by Anomaly Engine?\nWe are storing anomaly data into
          CR.\n ####Option 1:Through `kubectl`` command\nYou can check detected anomalies
          using the following command from your terminal. \n\n```bash\nkubectl get
          anomalydata -n <Namespace>\nkubectl describe anomalydata <NAME of the anomalydata>
          -n <Namespace> \n```\n #### Option 2: From Installed Operator UI Go to installed
          operator and select **Observability Analytics** operator, now go to **Anomaly
          Data** tab where you can able to see the detected anomaly, click on specific
          anomaly and you should be able to see details of the same. ### Documentation\nDocumentation
          around how to test Operator codebase as a developer can be found [here](https://github.com/openshift/analytics-operator#readme)\n\n###
          Contributing\nYou can contribute by:\n  * Raising [issues](https://github.com/openshift/analytics-operator/issues)\n
          \   related to observability-analytics-operator\n  * Fixing issues by opening
          [Pull Requests](https://github.com/openshift/analytics-operator/pulls)\n
          \ * Discussing observability-analytics-operator\n\n### License\nobservability-analytics-operator
          is licensed under the Apache 2.0 license.\n"
        displayName: Observability Analytics
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - observability-analytics-operator
        - analytics-operator
        - analytics
        - anomaly
        links:
        - name: Analytics Operator
          url: https://analytics-operator.domain
        maintainers:
        - email: rzalavad@redhat.com
          name: Raj
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
        version: 0.1.1
      entries:
      - name: analytics-operator.v0.1.1
        version: 0.1.1
      - name: analytics-operator.v0.1.0
        version: 0.1.0
      name: alpha
    - currentCSV: analytics-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyData",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalydata-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalydata",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalydata-sample"
                },
                "spec": {
                  "anomalyname": "etcd_object_namespaces",
                  "config": {
                    "max": 100,
                    "min": 60,
                    "query": "max(apiserver_storage_objects{resource=~\"namespaces\"}) by (resource)"
                  },
                  "method": "min_max",
                  "metricdata": {
                    "groupeddata": "{ \"resource\": \"namespaces\" }",
                    "latestvalue": 121,
                    "timestamp": 1695804362
                  }
                }
              },
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalyengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalyengine",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalyengine-sample"
                },
                "spec": {
                  "anomalyqueryconfiguration": "etcd_objects_all:\n  method: percentage_change\n  query: max(apiserver_storage_objects{resource!~\"certificatesigningrequests.certificates.k8s.io|installplans.operators.coreos.com|operators.operators.coreos.com|subscriptions.operators.coreos.com|anomalydata.observability-analytics.redhat.com|anomalyengines.observability-analytics.redhat.com\"}) by (resource)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: True\nfailed_pods:\n  method: min_max\n  query: count(kube_pod_status_phase{phase!~\"Succeeded|Pending|Running|Unknown\"}) by (phase, namespace)\n  min: 0\n  max: 10\n  have_multi_result_data: True\nworkload_memory_usage_bytes:\n  method: percentage_change\n  query: sum(container_memory_working_set_bytes{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nworkload_cpu_usage_cores:\n  method: percentage_change\n  query: sum(container_cpu_usage_seconds_total{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nrest_client_requests_total:\n  method: percentage_change\n  query: sum(rest_client_requests_total)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\n",
                  "cronjobconfig": {
                    "anomalyqueries": "etcd_objects_all,failed_pods,workload_memory_usage_bytes,workload_cpu_usage_cores,rest_client_requests_total",
                    "name": "osa-anomaly-detection",
                    "resource": {
                      "cpulimit": "256m",
                      "cpurequest": "128m",
                      "memorylimit": "512Mi",
                      "memoryrequest": "256Mi"
                    },
                    "schedule": "*/5 * * * *"
                  },
                  "namespace": "osa-anomaly-detection",
                  "serviceaccountrolebinding": {
                    "clusterrolename": "cluster-monitoring-view",
                    "name": "osa-in-cluster-thanos-querier"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
          createdAt: "2024-03-12T08:02:25Z"
          description: Deploys and Manages Observability Analytics Operator
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4-alpha
          repository: https://github.com/openshift/analytics-operator
          support: RedHat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnomalyData is the Schema for the anomalydata API
            displayName: Anomaly Data
            kind: AnomalyData
            name: anomalydata.observability-analytics.redhat.com
            version: v1alpha1
          - description: AnomalyEngine is the Schema for the anomalyengines API
            displayName: Anomaly Engine
            kind: AnomalyEngine
            name: anomalyengines.observability-analytics.redhat.com
            version: v1alpha1
        description: "### Introduction\nObservability Analytics for Red Hat OpenShift
          runs OpenShift analytics processes like anomaly detection.\n\n### API Backward
          Compatibility\n**NOTE:** Until the Operator supports **seamless upgrades**,
          a new version of the operator may introduce a change that is **NOT** backward-compatible
          with the previous version. Thus, to upgrade the operator, uninstall the
          already installed version completely (including CRDs) and install the new
          version.\n\n### Installtion Steps for the Anomaly Engine\n Create namespace
          if you would like to install this operator into separate namespace (You
          can utilize existing namespace as well). ```bash kubectl create ns <Namespace>
          ``` Install the Observability Analytics operator to specific `<Namespace>`,
          wait until you see the operator in the **Installed Operator** list. Click
          on the Operator name and go to Operator details.\n\nHere you should see
          a tab called **Anomaly Engine**. Click on it, and then click on the **Create
          AnomalyEngine** button, which will redirect you to a form where you need
          to provide the required inputs for the Anomaly Engine. Once successfully
          created, you should see a cronjob in the given namespace. This cronjob will
          identify anomalies and store them into **anomalydata**.\n\n### How to See
          Anomaly Data Captured by Anomaly Engine?\nWe are storing anomaly data into
          CR.\n ####Option 1:Through `kubectl`` command\nYou can check detected anomalies
          using the following command from your terminal. \n\n```bash\nkubectl get
          anomalydata -n <Namespace>\nkubectl describe anomalydata <NAME of the anomalydata>
          -n <Namespace> \n```\n #### Option 2: From Installed Operator UI Go to installed
          operator and select **Observability Analytics** operator, now go to **Anomaly
          Data** tab where you can able to see the detected anomaly, click on specific
          anomaly and you should be able to see details of the same. ### Documentation\nDocumentation
          around how to test Operator codebase as a developer can be found [here](https://github.com/openshift/analytics-operator#readme)\n\n###
          Contributing\nYou can contribute by:\n  * Raising [issues](https://github.com/openshift/analytics-operator/issues)\n
          \   related to observability-analytics-operator\n  * Fixing issues by opening
          [Pull Requests](https://github.com/openshift/analytics-operator/pulls)\n
          \ * Discussing observability-analytics-operator\n\n### License\nobservability-analytics-operator
          is licensed under the Apache 2.0 license.\n"
        displayName: Observability Analytics
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - observability-analytics-operator
        - analytics-operator
        - analytics
        - anomaly
        links:
        - name: Analytics Operator
          url: https://analytics-operator.domain
        maintainers:
        - email: rzalavad@redhat.com
          name: Raj
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
        version: 0.1.1
      entries:
      - name: analytics-operator.v0.1.1
        version: 0.1.1
      - name: analytics-operator.v0.1.0
        version: 0.1.0
      name: dev-preview
    defaultChannel: alpha
    packageName: analytics-operator
    provider:
      name: Red Hat Inc.
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-atlas-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-atlas-kubernetes.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupCompliancePolicy",
                "metadata": {
                  "name": "my-backup-compliance-policy"
                },
                "spec": {
                  "authorizedEmail": "john.doe@example.com",
                  "authorizedUserFirstName": "John",
                  "authorizedUserLastName": "Doe",
                  "copyProtectionEnabled": false,
                  "encryptionAtRestEnabled": false,
                  "onDemandPolicy": {
                    "retentionUnit": "weeks",
                    "retentionValue": 3
                  },
                  "overwriteBackupPolicies": false,
                  "pointInTimeEnabled": true,
                  "restoreWindowDays": 42,
                  "scheduledPolicyItems": [
                    {
                      "frequencyInterval": 2,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "hourly",
                      "retentionUnit": "days",
                      "retentionValue": 2
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "weekly",
                      "retentionUnit": "weeks",
                      "retentionValue": 4
                    },
                    {
                      "frequencyInterval": 15,
                      "frequencyType": "monthly",
                      "retentionUnit": "months",
                      "retentionValue": 12
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": false,
                  "copySettings": [
                    {
                      "cloudProvider": "AWS",
                      "frequencies": [
                        "WEEKLY"
                      ],
                      "regionName": "US_WEST_1",
                      "shouldCopyOplogs": false
                    }
                  ],
                  "policy": {
                    "name": "atlasbackuppolicy-sample"
                  },
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2,
                  "useOrgAndGroupNamesInExportPrefix": true
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasCustomRole",
                "metadata": {
                  "name": "shard-operator-role"
                },
                "spec": {
                  "projectRef": {
                    "name": "my-project"
                  },
                  "role": {
                    "actions": [
                      {
                        "name": "getShardMap",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "shardingState",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "connPoolStats",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "getLog",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      }
                    ],
                    "inheritedRoles": [
                      {
                        "database": "admin",
                        "name": "operator-role-1"
                      }
                    ],
                    "name": "my-role"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDataFederation",
                "metadata": {
                  "name": "my-data-federation",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "name": "my-data-federation",
                  "privateEndpoints": [
                    {
                      "endpointId": "vpce-03f9eeaa764e32454",
                      "provider": "AWS",
                      "type": "DATA_LAKE"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project",
                    "namespace": "mongodb-atlas-system"
                  },
                  "storage": {
                    "databases": [
                      {
                        "collections": [
                          {
                            "dataSources": [
                              {
                                "storeName": "http-test",
                                "urls": [
                                  "https://data.cityofnewyork.us/api/views/vfnx-vebw/rows.csv"
                                ]
                              }
                            ],
                            "name": "test-collection-1"
                          }
                        ],
                        "name": "test-db-1"
                      }
                    ],
                    "stores": [
                      {
                        "name": "http-test",
                        "provider": "http"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "the-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "theuser"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-deployment"
                },
                "spec": {
                  "backupRef": {
                    "name": "atlasbackupschedule-sample"
                  },
                  "deploymentSpec": {
                    "backupEnabled": true,
                    "clusterType": "REPLICASET",
                    "mongoDBMajorVersion": "7.0",
                    "name": "test-deployment",
                    "replicationSpecs": [
                      {
                        "regionConfigs": [
                          {
                            "electableSpecs": {
                              "instanceSize": "M10",
                              "nodeCount": 3
                            },
                            "priority": 7,
                            "providerName": "AWS",
                            "regionName": "US_EAST_1"
                          }
                        ],
                        "zoneName": "Zone 1"
                      }
                    ],
                    "tags": [
                      {
                        "key": "team",
                        "value": "atlas-operator"
                      }
                    ]
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasFederatedAuth",
                "metadata": {
                  "name": "atlas-default-federated-auth",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "connectionSecretRef": {
                    "name": "my-org-secret",
                    "namespace": "mongodb-atlas-system"
                  },
                  "domainAllowList": [
                    "my-org-domain.com"
                  ],
                  "domainRestrictionEnabled": true,
                  "enabled": true,
                  "postAuthRoleGrants": [
                    "ORG_MEMBER"
                  ],
                  "roleMappings": [
                    {
                      "externalGroupName": "org-admin",
                      "roleAssignments": [
                        {
                          "role": "ORG_OWNER"
                        }
                      ]
                    },
                    {
                      "externalGroupName": "dev-team",
                      "roleAssignments": [
                        {
                          "role": "ORG_GROUP_CREATOR"
                        },
                        {
                          "projectName": "dev-project",
                          "role": "GROUP_OWNER"
                        }
                      ]
                    }
                  ],
                  "ssoDebugEnabled": false
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasIPAccessList",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasipaccesslist-sample"
                },
                "spec": {
                  "entries": [
                    {
                      "cidrBlock": "192.168.1.0/24"
                    },
                    {
                      "deleteAfterDate": "2025-01-20T23:59:59+02:00",
                      "ipAddress": "10.0.0.1"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkContainer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasnetworkcontainer-sample"
                },
                "spec": {
                  "cidrBlock": "10.11.0.0/16",
                  "connectionSecret": {
                    "name": "atlas-connection-secret"
                  },
                  "externalProjectRef": {
                    "projectId": "66e2f2b621571b7e69a89b66"
                  },
                  "provider": "AWS",
                  "region": "US_EAST_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkPeering",
                "metadata": {
                  "name": "atlasnetworkpeering-sample"
                },
                "spec": {
                  "atlasCidrBlock": "10.12.0.0/16",
                  "awsConfiguration": {
                    "accepterRegionName": "US_EAST_1",
                    "awsAccountId": "23214235145",
                    "routeTableCidrBlock": "10.11.0.0/16",
                    "vpcId": "vpcid-7123782134"
                  },
                  "containerId": "623412394512350",
                  "containerRegion": "US_EAST_1",
                  "projectRef": {
                    "name": "atlas-project",
                    "namespace": "namespace"
                  },
                  "provider": "AWS"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasPrivateEndpoint",
                "metadata": {
                  "name": "atlasprivateendpoint-sample"
                },
                "spec": {
                  "awsConfiguration": {
                    "id": "vpce-f4k34w51d"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "provider": "AWS",
                  "region": "EU_CENTRAL_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    },
                    {
                      "cidrBlock": "203.0.113.0/24",
                      "comment": "CIDR block for Application Server B - D"
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasSearchIndexConfig",
                "metadata": {
                  "name": "atlassearchindexconfig-sample"
                },
                "spec": {
                  "analyzer": "lucene.standard",
                  "analyzers": [
                    {
                      "name": "test-analyzer",
                      "tokenFilters": [
                        {
                          "asciiFolding": {
                            "originalTokens": "omit"
                          },
                          "type": "asciiFolding"
                        }
                      ],
                      "tokenizer": {
                        "type": "whiteSpace",
                        "whitespace": {
                          "maxTokenLenght": 200
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamConnection",
                "metadata": {
                  "name": "my-streamconnection-kafka"
                },
                "spec": {
                  "kafkaConfig": {
                    "authentication": {
                      "credentials": {
                        "name": "kafka-userpass"
                      },
                      "mechanism": "SCRAM-512"
                    },
                    "bootstrapServers": "kafka.server1:9001,kafka.server2:9002,kafka.server3:9003",
                    "security": {
                      "certificate": {
                        "name": "kafka-cert"
                      },
                      "protocol": "SSL"
                    }
                  },
                  "name": "kafka-config",
                  "type": "Kafka"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamInstance",
                "metadata": {
                  "name": "my-streaminstance-sample"
                },
                "spec": {
                  "clusterConfig": {
                    "provider": "AWS",
                    "region": "VIRGINIA_USA",
                    "tier": "SP30"
                  },
                  "connectionRegistry": [
                    {
                      "name": "my-streamconnection-sample"
                    }
                  ],
                  "name": "my-instance",
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasTeam",
                "metadata": {
                  "name": "atlasteam-sample"
                },
                "spec": {
                  "name": "my-team",
                  "usernames": [
                    "username@mongodb.com"
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasThirdPartyIntegration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasintegration-sample"
                },
                "spec": {
                  "datadog": {
                    "apiKeySecretRef": {
                      "name": "secret-ref"
                    },
                    "region": "string",
                    "sendCollectionLatencyMetrics": "string",
                    "sendDatabaseMetrics": "string"
                  },
                  "projectRef": {
                    "name": "my-atlas-project"
                  },
                  "type": "DATADOG"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/mongodb/mongodb-atlas-kubernetes-operator:2.11.0
          createdAt: "2025-09-25T14:34:34Z"
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/mongodb/mongodb-atlas-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The AtlasBackupCompliancePolicy is a configuration that enforces
              specific backup and retention requirements
            displayName: Atlas Backup Compliance Policy
            kind: AtlasBackupCompliancePolicy
            name: atlasbackupcompliancepolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasCustomRole is the Schema for the AtlasCustomRole API
            displayName: Atlas Custom Role
            kind: AtlasCustomRole
            name: atlascustomroles.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDataFederation is the Schema for the Atlas Data Federation
              API
            displayName: Atlas Data Federation
            kind: AtlasDataFederation
            name: atlasdatafederations.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasdeployments API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasFederatedAuth is the Schema for the Atlasfederatedauth
              API
            displayName: Atlas Federated Auth
            kind: AtlasFederatedAuth
            name: atlasfederatedauths.atlas.mongodb.com
            version: v1
          - description: AtlasIPAccessList is the Schema for the atlasipaccesslists
              API.
            displayName: Atlas IPAccess List
            kind: AtlasIPAccessList
            name: atlasipaccesslists.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkContainer is the Schema for the AtlasNetworkContainer
              API
            displayName: Atlas Network Container
            kind: AtlasNetworkContainer
            name: atlasnetworkcontainers.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkPeering is the Schema for the AtlasNetworkPeering
              API
            displayName: Atlas Network Peering
            kind: AtlasNetworkPeering
            name: atlasnetworkpeerings.atlas.mongodb.com
            version: v1
          - displayName: Atlas Org Settings
            kind: AtlasOrgSettings
            name: atlasorgsettings.atlas.mongodb.com
            version: v1
          - description: "The AtlasPrivateEndpoint custom resource definition (CRD)
              defines a desired [Private Endpoint](https://www.mongodb.com/docs/atlas/security-private-endpoint/#std-label-private-endpoint-overview)
              configuration for an Atlas project. It allows a private connection between
              your cloud provider and Atlas that doesn't send information through
              a public network. \n You can use private endpoints to create a unidirectional
              connection to Atlas clusters from your virtual network."
            displayName: Atlas Private Endpoint
            kind: AtlasPrivateEndpoint
            name: atlasprivateendpoints.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
          - description: Atlas Search Index Config is the Schema for the Atlas Search
              API
            displayName: Atlas Search Index Config
            kind: AtlasSearchIndexConfig
            name: atlassearchindexconfigs.atlas.mongodb.com
            version: v1
          - description: AtlasStreamConnection is the Schema for the atlasstreamconnections
              API
            displayName: Atlas Stream Connection
            kind: AtlasStreamConnection
            name: atlasstreamconnections.atlas.mongodb.com
            version: v1
          - description: AtlasStreamInstance is the Schema for the atlasstreaminstances
              API
            displayName: Atlas Stream Instance
            kind: AtlasStreamInstance
            name: atlasstreaminstances.atlas.mongodb.com
            version: v1
          - description: AtlasTeam is the Schema for the Atlas Teams API
            displayName: Atlas Team
            kind: AtlasTeam
            name: atlasteams.atlas.mongodb.com
            version: v1
          - description: AtlasThirdPartyIntegration is the Schema for the atlas 3rd
              party integrations API.
            displayName: Atlas Third Party Integration
            kind: AtlasThirdPartyIntegration
            name: atlasthirdpartyintegrations.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB
          Atlas — the only multi-cloud document database service that gives you the versatility you need to build sophisticated
          and resilient applications that can adapt to changing customer demands and market trends.

          The full documentation for the Operator can be found [here](https://docs.atlas.mongodb.com/atlas-operator/)

          ## Getting Started

          ### Supported features

          * Create and configure a project, or connect to an existing one.
          * Deploy, manage, scale, and tear down clusters.
          * Support for serverless instances.
          * Create and edit database users.
          * Manage IP Access Lists, network peering and private endpoints.
          * Configure and control Atlas’s fully managed cloud backup.
          * Configure federated authentication for your Atlas organization
          * Integrate Atlas monitoring with Prometheus.

          ... and more.

          To view the list of custom resources and each CR's schema, visit our [reference](https://www.mongodb.com/docs/atlas/operator/stable/custom-resources/)
          documentation. See the [Quickstart](https://www.mongodb.com/docs/atlas/operator/stable/ak8so-quick-start/) to get
          started with Atlas Kubernetes Operator.
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator:2.11.0
        version: 2.11.0
      entries:
      - name: mongodb-atlas-kubernetes.v2.11.0
        version: 2.11.0
      - name: mongodb-atlas-kubernetes.v2.10.0
        version: 2.10.0
      - name: mongodb-atlas-kubernetes.v2.9.1
        version: 2.9.1
      - name: mongodb-atlas-kubernetes.v2.9.0
        version: 2.9.0
      - name: mongodb-atlas-kubernetes.v2.8.2
        version: 2.8.2
      - name: mongodb-atlas-kubernetes.v2.8.1
        version: 2.8.1
      - name: mongodb-atlas-kubernetes.v2.8.0
        version: 2.8.0
      - name: mongodb-atlas-kubernetes.v2.7.1
        version: 2.7.1
      - name: mongodb-atlas-kubernetes.v2.7.0
        version: 2.7.0
      - name: mongodb-atlas-kubernetes.v2.6.1
        version: 2.6.1
      - name: mongodb-atlas-kubernetes.v2.6.0
        version: 2.6.0
      - name: mongodb-atlas-kubernetes.v2.5.0
        version: 2.5.0
      - name: mongodb-atlas-kubernetes.v2.4.1
        version: 2.4.1
      - name: mongodb-atlas-kubernetes.v2.4.0
        version: 2.4.0
      - name: mongodb-atlas-kubernetes.v2.3.1
        version: 2.3.1
      - name: mongodb-atlas-kubernetes.v2.3.0
        version: 2.3.0
      - name: mongodb-atlas-kubernetes.v2.2.2
        version: 2.2.2
      - name: mongodb-atlas-kubernetes.v2.2.1
        version: 2.2.1
      - name: mongodb-atlas-kubernetes.v2.2.0
        version: 2.2.0
      - name: mongodb-atlas-kubernetes.v2.1.0
        version: 2.1.0
      - name: mongodb-atlas-kubernetes.v2.0.1
        version: 2.0.1
      - name: mongodb-atlas-kubernetes.v2.0.0
        version: 2.0.0
      - name: mongodb-atlas-kubernetes.v1.9.1
        version: 1.9.1
      - name: mongodb-atlas-kubernetes.v1.9.0
        version: 1.9.0
      - name: mongodb-atlas-kubernetes.v1.8.2
        version: 1.8.2
      - name: mongodb-atlas-kubernetes.v1.8.1
        version: 1.8.1
      - name: mongodb-atlas-kubernetes.v1.8.0
        version: 1.8.0
      - name: mongodb-atlas-kubernetes.v1.7.3
        version: 1.7.3
      - name: mongodb-atlas-kubernetes.v1.7.2
        version: 1.7.2
      - name: mongodb-atlas-kubernetes.v1.7.1
        version: 1.7.1
      - name: mongodb-atlas-kubernetes.v1.7.0
        version: 1.7.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-atlas-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OpenTelemetry Community
      provider-url: ""
    name: opentelemetry-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opentelemetry-operator.v0.136.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "Instrumentation",
                "metadata": {
                  "name": "instrumentation"
                },
                "spec": {
                  "exporter": {
                    "endpoint": "http://otel-collector-headless:4317"
                  }
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpAMPBridge",
                "metadata": {
                  "name": "opampbridge-sample"
                },
                "spec": {
                  "capabilities": {
                    "AcceptsOpAMPConnectionSettings": true,
                    "AcceptsOtherConnectionSettings": true,
                    "AcceptsRemoteConfig": true,
                    "AcceptsRestartCommand": true,
                    "ReportsEffectiveConfig": true,
                    "ReportsHealth": true,
                    "ReportsOwnLogs": true,
                    "ReportsOwnMetrics": true,
                    "ReportsOwnTraces": true,
                    "ReportsRemoteConfig": true,
                    "ReportsStatus": true
                  },
                  "componentsAllowed": {
                    "exporters": [
                      "debug"
                    ],
                    "processors": [
                      "memory_limiter"
                    ],
                    "receivers": [
                      "otlp"
                    ]
                  },
                  "endpoint": "ws://opamp-server:4320/v1/opamp"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": "receivers:\n  otlp:\n    protocols: \n      grpc: {}\n      http: {}\n\nexporters:\n  debug: {}\n\nservice:\n  pipelines:\n    traces:\n      receivers: [otlp]\n      exporters: [debug]\n"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1beta1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": {
                    "exporters": {
                      "debug": {}
                    },
                    "receivers": {
                      "otlp": {
                        "protocols": {
                          "grpc": {},
                          "http": {}
                        }
                      }
                    },
                    "service": {
                      "pipelines": {
                        "traces": {
                          "exporters": [
                            "debug"
                          ],
                          "receivers": [
                            "otlp"
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          certified: "false"
          containerImage: ghcr.io/open-telemetry/opentelemetry-operator/opentelemetry-operator
          createdAt: "2025-09-29T13:00:19Z"
          description: Provides the OpenTelemetry components, including the Collector
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/open-telemetry/opentelemetry-operator
          support: OpenTelemetry Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instrumentation is the spec for OpenTelemetry instrumentation.
            displayName: OpenTelemetry Instrumentation
            kind: Instrumentation
            name: instrumentations.opentelemetry.io
            version: v1alpha1
          - description: OpAMPBridge is the Schema for the opampbridges API.
            displayName: OpAMP Bridge
            kind: OpAMPBridge
            name: opampbridges.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1beta1
          - description: TargetAllocator is the Schema for the targetallocators API.
            displayName: Target Allocator
            kind: TargetAllocator
            name: targetallocators.opentelemetry.io
            version: v1alpha1
        description: |-
          OpenTelemetry is a collection of tools, APIs, and SDKs. You use it to instrument, generate, collect, and export telemetry data (metrics, logs, and traces) for analysis in order to understand your software's performance and behavior.

          ### Operator features

          * **Sidecar injection** - annotate your pods and let the operator inject a sidecar.
          * **Managed upgrades** - updating the operator will automatically update your OpenTelemetry collectors.
          * **Deployment modes** - your collector can be deployed as sidecar, daemon set, or regular deployment.
          * **Service port management** - the operator detects which ports need to be exposed based on the provided configuration.
        displayName: Community OpenTelemetry Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opentelemetry
        - tracing
        - logging
        - metrics
        - monitoring
        - troubleshooting
        links:
        - name: OpenTelemetry Operator
          url: https://github.com/open-telemetry/opentelemetry-operator
        maintainers:
        - email: jpkroehling@redhat.com
          name: Juraci Paixão Kröhling
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: OpenTelemetry Community
        relatedImages:
        - ghcr.io/open-telemetry/opentelemetry-operator/opentelemetry-operator:0.136.0
        - quay.io/brancz/kube-rbac-proxy:v0.13.1
        version: 0.136.0
      entries:
      - name: opentelemetry-operator.v0.136.0
        version: 0.136.0
      - name: opentelemetry-operator.v0.135.0
        version: 0.135.0
      - name: opentelemetry-operator.v0.134.0
        version: 0.134.0
      - name: opentelemetry-operator.v0.132.0
        version: 0.132.0
      - name: opentelemetry-operator.v0.131.0
        version: 0.131.0
      - name: opentelemetry-operator.v0.129.1
        version: 0.129.1
      - name: opentelemetry-operator.v0.127.0
        version: 0.127.0
      - name: opentelemetry-operator.v0.126.0
        version: 0.126.0
      - name: opentelemetry-operator.v0.125.0
        version: 0.125.0
      - name: opentelemetry-operator.v0.124.0
        version: 0.124.0
      - name: opentelemetry-operator.v0.123.0
        version: 0.123.0
      - name: opentelemetry-operator.v0.122.0
        version: 0.122.0
      - name: opentelemetry-operator.v0.121.0
        version: 0.121.0
      - name: opentelemetry-operator.v0.120.0
        version: 0.120.0
      - name: opentelemetry-operator.v0.119.0
        version: 0.119.0
      - name: opentelemetry-operator.v0.118.0
        version: 0.118.0
      - name: opentelemetry-operator.v0.117.0
        version: 0.117.0
      - name: opentelemetry-operator.v0.116.0
        version: 0.116.0
      - name: opentelemetry-operator.v0.115.0
        version: 0.115.0
      - name: opentelemetry-operator.v0.114.1
        version: 0.114.1
      - name: opentelemetry-operator.v0.112.0
        version: 0.112.0
      - name: opentelemetry-operator.v0.111.0
        version: 0.111.0
      - name: opentelemetry-operator.v0.110.0
        version: 0.110.0
      - name: opentelemetry-operator.v0.109.0
        version: 0.109.0
      - name: opentelemetry-operator.v0.108.0
        version: 0.108.0
      - name: opentelemetry-operator.v0.107.0
        version: 0.107.0
      - name: opentelemetry-operator.v0.106.0
        version: 0.106.0
      - name: opentelemetry-operator.v0.105.0
        version: 0.105.0
      name: alpha
    defaultChannel: alpha
    packageName: opentelemetry-operator
    provider:
      name: OpenTelemetry Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: StreamNative
      provider-url: https://streamnative.io
    name: sn-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc:latest",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway:0.10.0"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator:v0.10.0
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator:v0.10.0
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      - name: sn-operator.v0.9.0
        version: 0.9.0
      - name: sn-operator.v0.8.5
        version: 0.8.5
      - name: sn-operator.v0.7.10
        version: 0.7.10
      name: alpha
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc:latest",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway:0.10.0"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator:v0.10.0
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator:v0.10.0
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      - name: sn-operator.v0.9.0
        version: 0.9.0
      - name: sn-operator.v0.8.5
        version: 0.8.5
      - name: sn-operator.v0.7.10
        version: 0.7.10
      name: beta
    - currentCSV: sn-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bookkeeper.streamnative.io/v1alpha1",
                "kind": "BookKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "bookies",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "storage": {
                    "journal": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "8Gi"
                          }
                        }
                      }
                    },
                    "ledger": {
                      "numDirsPerVolume": 1,
                      "numVolumes": 1,
                      "volumeClaimTemplate": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "16Gi"
                          }
                        }
                      }
                    },
                    "reclaimPolicy": "Delete"
                  },
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ApiKeys",
                "metadata": {
                  "labels": {
                    "cloud.streamnative.io/app": "pulsar",
                    "cloud.streamnative.io/cluster": "apikeys",
                    "cloud.streamnative.io/component": "apikeys",
                    "cloud.streamnative.io/location": "us-central1",
                    "cloud.streamnative.io/poolmember-name": "gcp-shared-gcp-usce1-whale-snc",
                    "cloud.streamnative.io/poolmember-namespace": "streamnative",
                    "cloud.streamnative.io/pulsar-cluster": "apikeys",
                    "cloud.streamnative.io/pulsar-instance": "apikeys",
                    "cloud.streamnative.io/role": "apikeys",
                    "istio.io/rev": "sn-stable",
                    "k8s.streamnative.io/coordinator-name": "apikeys"
                  },
                  "name": "apikeys",
                  "namespace": "guangning"
                },
                "spec": {
                  "brokerServiceUrl": "pulsar://apikeys-broker.guangning.svc.cluster.local:6650",
                  "config": {
                    "authenticator": {
                      "acceptedAudience": "urn:sn:pulsar:guangning:apikeys",
                      "enabled": true,
                      "issuers": [
                        "https://auth.streamnative.cloud/",
                        "https://auth.sncloud-stg.dev/",
                        "https://auth.test.cloud.gcp.streamnative.dev/"
                      ]
                    },
                    "backend": {
                      "pulsarTopic": "api_keys_api.example"
                    },
                    "server": {
                      "audience": "urn:sn:pulsar:guangning:apikeys",
                      "claimsToCopy": [
                        "sub",
                        "permissions",
                        "aud",
                        "scope",
                        "https://streamnative.io/username"
                      ],
                      "currentKid": "7007ddf9-beac-4e74-8e32-f03453fea3e0"
                    }
                  },
                  "hostname": "apikeys-49c084ec-dfaa-4aed-a879-68ae5c2d1b30.usce1-whale.test.g.sn2.dev",
                  "image": "docker.cloudsmith.io/streamnative/sn-api-keys-svc/sn-api-keys-svc:latest",
                  "istio": {
                    "enabled": true,
                    "gateway": {
                      "gateways": [
                        "apikeys-gateway"
                      ]
                    }
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "ConnectorCatalog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sn-operator",
                    "app.kubernetes.io/instance": "connectorcatalog-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "connectorcatalog",
                    "app.kubernetes.io/part-of": "sn-operator"
                  },
                  "name": "connectorcatalog-sample"
                },
                "spec": {
                  "connectorDefinitions": [
                    {
                      "description": "Test data generator connector",
                      "id": "pulsar-io-data-generator",
                      "imageRepository": "streamnative/pulsar-io-data-generator",
                      "imageTag": "2.9.2.17",
                      "name": "data-generator",
                      "sinkClass": "org.apache.pulsar.io.datagenerator.DataGeneratorPrintSink",
                      "sourceClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSource",
                      "sourceConfigClass": "org.apache.pulsar.io.datagenerator.DataGeneratorSourceConfig",
                      "sourceConfigFieldDefinitions": [
                        {
                          "attributes": {
                            "defaultValue": "50",
                            "help": "How long to sleep between emitting messages",
                            "required": "true",
                            "sensitive": "false"
                          },
                          "fieldName": "sleepBetweenMessages",
                          "typeName": "long"
                        }
                      ],
                      "typeClassName": "org.apache.pulsar.io.datagenerator.Person",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Kafka Source",
                      "iconLink": "https://docs.streamnative.io/images/connectors/kafka-logo.png",
                      "id": "pulsar-io-kafka",
                      "imageRepository": "streamnative/pulsar-io-kafka",
                      "imageTag": "2.9.2.17",
                      "name": "kafka",
                      "sinkDocLink": "https://docs.streamnative.io/hub/connector-kafka-sink",
                      "sourceClass": "org.apache.pulsar.io.kafka.KafkaBytesSource",
                      "sourceConfigClass": "org.apache.pulsar.io.kafka.KafkaSourceConfig",
                      "sourceDocLink": "https://docs.streamnative.io/hub/connector-kafka-source",
                      "sourceTypeClassName": "java.nio.ByteBuffer",
                      "version": "2.9.2.17"
                    },
                    {
                      "description": "Cloud storage Sink",
                      "id": "pulsar-io-cloud-storage",
                      "imageRepository": "streamnative/pulsar-io-cloud-storage",
                      "imageTag": "2.9.2.17",
                      "name": "cloud-storage",
                      "sinkClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageGenericRecordSink",
                      "sinkConfigClass": "org.apache.pulsar.io.jcloud.sink.CloudStorageSinkConfig",
                      "typeClassName": "org.apache.pulsar.client.api.schema.GenericRecord",
                      "version": "2.9.2.17"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "console",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud-console:v2.3.3",
                  "webServiceUrl": "http://brokers-broker:8080"
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaCluster",
                "metadata": {
                  "name": "oxiacluster-example",
                  "namespace": "default"
                },
                "spec": {
                  "image": "streamnative/oxia:main",
                  "imagePullPolicy": "IfNotPresent",
                  "monitoringEnabled": true,
                  "server": {
                    "persistentVolumeClaimRetentionPolicy": {
                      "whenDeleted": "Delete"
                    },
                    "replicas": 3
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "oxianamespace-sample",
                  "namespace": "default"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns",
                    "replicationFactor": 2
                  },
                  "ownerRef": {
                    "name": "private-cloud1",
                    "namespace": "default"
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "OxiaNamespace",
                "metadata": {
                  "name": "ns-sample",
                  "namespace": "oxia-test"
                },
                "spec": {
                  "clusterRef": {
                    "name": "oxiacluster-example",
                    "namespace": "default"
                  },
                  "namespaceConfig": {
                    "initialShardCount": 2,
                    "name": "test-ns2",
                    "replicationFactor": 2
                  }
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PFSQLCluster",
                "metadata": {
                  "name": "pfsqlcluster-sample"
                },
                "spec": {
                  "gateway": {
                    "image": "docker.cloudsmith.io/streamnative/pfsql/pfsql-gateway:0.10.0"
                  },
                  "ingressConfig": {
                    "hostName": "pfsql-gateway-example.local",
                    "path": "/"
                  },
                  "pulsarClusterConfig": {
                    "builtinFunctions": [
                      {
                        "bindStatements": [
                          "io.streamnative.pfsql.tree.InsertInto",
                          "io.streamnative.pfsql.tree.InsertMulti"
                        ],
                        "className": "io.streamnative.pfsql.function.FilterRouteFunction",
                        "description": "pfsql-function",
                        "filename": "pfsql-function.jar",
                        "name": "pfsql-function",
                        "packageUrl": "function://public/default/filter-routing-function@0.1.0"
                      }
                    ],
                    "metadataTopic": "public/demo/pfsql-demo-metadata",
                    "serviceUrl": "pulsar://sn-platform-pulsar-broker.default.svc.cluster.local:6650",
                    "webServiceUrl": "http://sn-platform-pulsar-broker.default.svc.cluster.local:8080"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "k8s.streamnative.io/v1alpha1",
                "kind": "PulsarCoordinator",
                "metadata": {
                  "name": "private-cloud",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "troubleshootings": [
                    {
                      "action": "heapDump",
                      "id": 100000001,
                      "message": "test help dump",
                      "name": "test-broker-0",
                      "scope": "pods"
                    }
                  ]
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarBroker",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "brokers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "configurationMetadataStoreUrl": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                  "configurationStoreConfig": {
                    "zk": {
                      "certSecretName": "pulsar-tls",
                      "clusterIndex": 1
                    }
                  },
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "metadataStoreUrl": "geo2-zk:2181",
                  "pod": {
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3,
                  "zkServers": "zookeepers-zk:2181"
                }
              },
              {
                "apiVersion": "pulsar.streamnative.io/v1alpha1",
                "kind": "PulsarProxy",
                "metadata": {
                  "name": "proxys",
                  "namespace": "pulsar"
                },
                "spec": {
                  "brokerAddress": "brokers-broker",
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 2
                }
              },
              {
                "apiVersion": "zookeeper.streamnative.io/v1alpha1",
                "kind": "ZooKeeperCluster",
                "metadata": {
                  "labels": {
                    "k8s.streamnative.io/coordinator-name": "private-cloud"
                  },
                  "name": "zookeepers",
                  "namespace": "pulsar"
                },
                "spec": {
                  "image": "streamnative/private-cloud:3.0.1.4",
                  "istio": {
                    "authRules": [
                      {
                        "fromSource": [
                          {
                            "principals": [
                              "cluster.local/ns/pulsar/sa/default"
                            ]
                          }
                        ]
                      },
                      {
                        "toOperation": [
                          {
                            "ports": [
                              "2181",
                              "8000"
                            ]
                          }
                        ]
                      }
                    ],
                    "enabled": true,
                    "gateway": {
                      "advertisedDomain": "zk-fdl-geo1.aws-use2-dixie-snc.streamnative.test.aws.sn2.dev",
                      "create": true,
                      "selector": {
                        "cloud.streamnative.io/role": "istio-ingressgateway"
                      },
                      "tls": {
                        "certSecretName": "istio-ingressgateway-tls",
                        "mode": "mutual",
                        "subjectAltNames": [
                          "*.sncloud.test.sn2.dev"
                        ]
                      }
                    }
                  },
                  "pod": {
                    "resources": {
                      "requests": {
                        "cpu": "50m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {
                      "runAsNonRoot": true
                    }
                  },
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/sn-operator:v0.10.0
          createdAt: "2025-04-19T11:49:37Z"
          description: The StreamNative Operator is a commercial operator built for
            managing StreamNative Private Cloud commercial components and enterprise
            features.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["StreamNative Private Cloud
            License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: StreamNative, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApiKeys is the Schema for the apikeys API
            displayName: Api Keys
            kind: ApiKeys
            name: apikeys.k8s.streamnative.io
            version: v1alpha1
          - description: BookKeeperCluster is the Schema for the bookkeeperclusters
              API
            displayName: BookKeeper Cluster
            kind: BookKeeperCluster
            name: bookkeeperclusters.bookkeeper.streamnative.io
            version: v1alpha1
          - description: ConnectorCatalog is the Schema for the connectorcatalog API
            displayName: Connector Catalog
            kind: ConnectorCatalog
            name: connectorcatalogs.k8s.streamnative.io
            version: v1alpha1
          - description: Console is the Schema for the consoles API
            displayName: Console
            kind: Console
            name: consoles.k8s.streamnative.io
            version: v1alpha1
          - kind: KafkaConnect
            name: kafkaconnects.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaCluster is the Schema for the oxiaclusters API
            displayName: Oxia Cluster
            kind: OxiaCluster
            name: oxiaclusters.k8s.streamnative.io
            version: v1alpha1
          - description: OxiaNamespace is the Schema for the oxianamespace API
            displayName: Oxia Namespace
            kind: OxiaNamespace
            name: oxianamespaces.k8s.streamnative.io
            version: v1alpha1
          - description: PFSQLCluster is the Schema for the pfsqlclusters API
            displayName: PFSQLCluster
            kind: PFSQLCluster
            name: pfsqlclusters.k8s.streamnative.io
            version: v1alpha1
          - kind: PulsarBrokerRevision
            name: pulsarbrokerrevisions.pulsar.streamnative.io
            version: v1alpha1
          - displayName: Pulsar Broker
            kind: PulsarBroker
            name: pulsarbrokers.pulsar.streamnative.io
            version: v1alpha1
          - description: PulsarCoordinator is the Schema for the pulsarcoordinators
              API
            displayName: Pulsar Coordinator
            kind: PulsarCoordinator
            name: pulsarcoordinators.k8s.streamnative.io
            version: v1alpha1
          - description: PulsarProxy is the Schema for the pulsarproxies API
            displayName: Pulsar Proxy
            kind: PulsarProxy
            name: pulsarproxies.pulsar.streamnative.io
            version: v1alpha1
          - kind: UniLink
            name: unilinks.k8s.streamnative.io
            version: v1alpha1
          - kind: UniLinkSchema
            name: unilinkschemas.k8s.streamnative.io
            version: v1alpha1
          - description: ZooKeeperCluster is the Schema for the zookeeperclusters
              API
            displayName: ZooKeeper Cluster
            kind: ZooKeeperCluster
            name: zookeeperclusters.zookeeper.streamnative.io
            version: v1alpha1
        description: |
          StreamNative Private Cloud is an enterprise product which brings specific controllers for Kubernetes by providing specific Custom Resource Definitions (CRDs) that extend the basic Kubernetes orchestration capabilities to support the setup and management of StreamNative components.

          ### Capabilities

          With StreamNative Private Cloud, you can simplify operations and maintenance, including:
          - Provisioning and managing multiple Pulsar clusters
          - Scaling Pulsar clusters through rolling upgrades
          - Managing the Pulsar cluster configurations through declarative APIs
          - Simplify the cluster operation with Auto-Scaling
          - Cost efficiency with the Lakehouse tiered storage

          ### Apply for trial

          Before installing StreamNative Private Cloud, you need to import a valid license. You can contact StreamNative to apply for a [free trial](https://streamnative.io/deployment/start-free-trial).

          ### Quick Start
          Follow our [Quick Start](https://docs.streamnative.io/private/private-cloud-quickstart) guide to quickly provision and manage Pulsar clusters with the StreamNative Private Cloud.
        displayName: StreamNative Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - streamnative
        - pulsar
        links:
        - name: Home Page
          url: https://streamnative.io
        - name: Documentation
          url: https://docs.streamnative.io
        maintainers:
        - email: cloud@streamnative.io
          name: StreamNative Cloud
        maturity: alpha
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/sn-operator:v0.10.0
        version: 0.10.0
      entries:
      - name: sn-operator.v0.10.0
        version: 0.10.0
      - name: sn-operator.v0.9.0
        version: 0.9.0
      - name: sn-operator.v0.8.5
        version: 0.8.5
      - name: sn-operator.v0.7.10
        version: 0.7.10
      name: stable
    defaultChannel: alpha
    packageName: sn-operator
    provider:
      name: StreamNative
      url: https://streamnative.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Instana
      provider-url: ""
    name: instana-agent-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: instana-agent-operator.v2.1.35
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instana.io/v1",
                "kind": "InstanaAgent",
                "metadata": {
                  "name": "instana-agent",
                  "namespace": "instana-agent"
                },
                "spec": {
                  "agent": {
                    "configuration_yaml": "# You can leave this empty, or use this to configure your instana agent.\n# See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.\n",
                    "endpointHost": "ingress-red-saas.instana.io",
                    "endpointPort": "443",
                    "env": {},
                    "key": "replace-key"
                  },
                  "cluster": {
                    "name": "my-cluster"
                  },
                  "opentelemetry": {
                    "grpc": {
                      "enabled": true
                    },
                    "http": {
                      "enabled": true
                    }
                  },
                  "zone": {
                    "name": "edited-zone"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring,Logging & Tracing,OpenShift Optional
          certified: "false"
          containerImage: icr.io/instana/instana-agent-operator@sha256:fe949ae0ca9cc05ffe3569cd7a6f4fde48f5e3ee64f27e20e0b3102733a6e809
          description: Fully automated Application Performance Monitoring (APM) for
            microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/instana/instana-agent-operator
          support: Instana
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InstanaAgent is the Schema for the agents API
            displayName: Instana Agent
            kind: InstanaAgent
            name: agents.instana.io
            version: v1
          - description: InstanaAgentRemote is the Schema for the agents API
            displayName: Instana Agent Remote
            kind: InstanaAgentRemote
            name: agentsremote.instana.io
            version: v1
        description: |
          # Instana

          Instana is an [APM solution](https://www.instana.com/product-overview/) built for microservices that enables IT Ops to build applications faster and deliver higher quality services by automating monitoring, tracing and root cause analysis. The solution is optimized for [Kubernetes](https://www.instana.com/automatic-kubernetes-monitoring/) and [OpenShift](https://www.instana.com/blog/automatic-root-cause-analysis-for-openshift-applications/).

          ## Instana Agent Operator

          This is the Kubernetes Operator for installing the Instana Agent on Kubernetes or OpenShift.

          ## Prerequisites for OpenShift

          Before the agent will be able to run in OpenShift, you need to perform a couple of extra configuration steps.

          You need to set up a project for the Instana Agent and configure it's permissions.

          The project you create here needs to be the namespace where you create the Instana Agent custom resource that the operator will use to deploy the agent.

          For example, create the `instana-agent` project:

              oc new-project instana-agent

          Then, ensure the `instana-agent` service account is in the privileged security context:

              oc adm policy add-scc-to-user privileged -z instana-agent

          The `instana-agent-remote` service account must be allowed to use any userid:

              oc adm policy add-scc-to-user anyuid -z instana-agent-remote -n instana-agent

          This service account will be created by the operator.

          Now you can proceed with installing the operator for the Instana agent.

          ## Installation and Configuration

          First, install this operator from [OperatorHub.io](https://operatorhub.io/), [OpenShift Container Platform](https://www.openshift.com/), or [OKD](https://www.okd.io/).

          Second, create a custom resource with the agent configuration in the target namespace (for now, this MUST always be the `instana-agent` namespace). The operator will pick up the custom resource and install the Instana agent accordingly.

          The following is a minimal template of the custom resource:

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgent
          metadata:
            name: instana-agent
            namespace: instana-agent
          spec:
            zone:
              name: my-zone # (optional) name of the zone of the host
            cluster:
              name: replace-me # replace with the name of your Kubernetes cluster
            agent:
              key: replace-me # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io # the monitoring ingress endpoint
              endpointPort: "443" # the monitoring ingress endpoint port, wrapped in quotes
              env:
                INSTANA_AGENT_TAGS: example
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://ibm.biz/monitoring-k8s
          ```

          Save the template in a file `instana-agent.yaml` and edit the following values:

          * If your target namespace is not `instana-agent`, replace the `namespace:` accordingly.
          * `agent.key` must be set with your Instana agent key.
          * `agent.endpointHost` must be set with the monitoring ingress endpoint, generally either `saas-us-west-2.instana.io` or `saas-eu-west-1.instana.io`.
          * `agent.endpointPort` must be set with the monitoring ingress port, generally "443" (wrapped in quotes).
          * `zone.name` should be set with the name of the Kubernetes cluster that is be displayed in Instana.

          For advanced configuration, you can edit the contents of the `configuration.yaml` file. View documentation [here](https://ibm.biz/monitoring-k8s).

          Apply the custom resource with `kubectl apply -f instana-agent.yaml`. After some time, you should see `instana-agent` Pods being created on each node of your cluster, and your cluster should show on the infrastructure map on your Instana Web interface.

          In addition to the InstanaAgent custom resource, a remote agent can be deployed. This agent will not monitor workload running inside the cluster, but can deploy a single replica of an agent which can get an independent configuration for running remote monitoring for certain sensors.

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgentRemote
          metadata:
            name: remote-agent
            namespace: "instana-agent"
          spec:
            zone:
              name: edited-zone # name of the zone of the host
            hostname:
              name: use-custom-hostname-for-remote-agent # (optional) give your remote agent pod a custom hostname
            agent:
              key: replace-key # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io
              endpointPort: "443"
              env: {}
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.
          ```

          ## Uninstalling

          In order to uninstall the Instana agent, simply remove the custom resource with `kubectl delete -f instana-agent.yaml`.

          ## Source Code

          The Instana agent operator is an open source project hosted on [https://github.com/instana/instana-agent-operator](https://github.com/instana/instana-agent-operator/).
        displayName: Instana Agent Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - APM
        - Instana
        links:
        - name: GitHub Repository
          url: https://github.com/instana/instana-agent-operator
        maintainers:
        - email: support@instana.com
          name: Instana
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: Instana
        relatedImages:
        - icr.io/instana/instana-agent-operator@sha256:fe949ae0ca9cc05ffe3569cd7a6f4fde48f5e3ee64f27e20e0b3102733a6e809
        - icr.io/instana/agent@sha256:ed71e6c3e5bdb155853a2db8c916460e6a867fb02627e3a332186188cbbd98dd
        version: 2.1.35
      entries:
      - name: instana-agent-operator.v2.1.35
        version: 2.1.35
      - name: instana-agent-operator.v2.1.34
        version: 2.1.34
      - name: instana-agent-operator.v2.1.33
        version: 2.1.33
      - name: instana-agent-operator.v2.1.32
        version: 2.1.32
      - name: instana-agent-operator.v2.1.31
        version: 2.1.31
      - name: instana-agent-operator.v2.1.30
        version: 2.1.30
      - name: instana-agent-operator.v2.1.29
        version: 2.1.29
      - name: instana-agent-operator.v2.1.28
        version: 2.1.28
      - name: instana-agent-operator.v2.1.27
        version: 2.1.27
      - name: instana-agent-operator.v2.1.25
        version: 2.1.25
      - name: instana-agent-operator.v2.1.24
        version: 2.1.24
      - name: instana-agent-operator.v2.1.23
        version: 2.1.23
      - name: instana-agent-operator.v2.1.22
        version: 2.1.22
      - name: instana-agent-operator.v2.1.21
        version: 2.1.21
      - name: instana-agent-operator.v2.1.20
        version: 2.1.20
      - name: instana-agent-operator.v2.1.19
        version: 2.1.19
      - name: instana-agent-operator.v2.1.18
        version: 2.1.18
      - name: instana-agent-operator.v2.1.17
        version: 2.1.17
      - name: instana-agent-operator.v2.1.16
        version: 2.1.16
      - name: instana-agent-operator.v2.1.15
        version: 2.1.15
      - name: instana-agent-operator.v2.1.14
        version: 2.1.14
      - name: instana-agent-operator.v2.1.13
        version: 2.1.13
      - name: instana-agent-operator.v2.1.12
        version: 2.1.12
      - name: instana-agent-operator.v2.1.11
        version: 2.1.11
      - name: instana-agent-operator.v2.1.10
        version: 2.1.10
      - name: instana-agent-operator.v2.1.9
        version: 2.1.9
      - name: instana-agent-operator.v2.1.8
        version: 2.1.8
      - name: instana-agent-operator.v2.1.7
        version: 2.1.7
      - name: instana-agent-operator.v2.1.6
        version: 2.1.6
      - name: instana-agent-operator.v2.1.5
        version: 2.1.5
      - name: instana-agent-operator.v2.1.4
        version: 2.1.4
      - name: instana-agent-operator.v2.1.3
        version: 2.1.3
      - name: instana-agent-operator.v2.1.2
        version: 2.1.2
      - name: instana-agent-operator.v2.1.1
        version: 2.1.1
      - name: instana-agent-operator.v2.1.0
        version: 2.1.0
      - name: instana-agent-operator.v2.0.32
        version: 2.0.32
      - name: instana-agent-operator.v2.0.31
        version: 2.0.31
      - name: instana-agent-operator.v2.0.30
        version: 2.0.30
      - name: instana-agent-operator.v2.0.27
        version: 2.0.27
      - name: instana-agent-operator.v2.0.19
        version: 2.0.19
      - name: instana-agent-operator.v2.0.18
        version: 2.0.18
      - name: instana-agent-operator.v2.0.17
        version: 2.0.17
      - name: instana-agent-operator.v2.0.13
        version: 2.0.13
      - name: instana-agent-operator.v2.0.12
        version: 2.0.12
      - name: instana-agent-operator.v2.0.11
        version: 2.0.11
      - name: instana-agent-operator.v2.0.10
        version: 2.0.10
      - name: instana-agent-operator.v2.0.9
        version: 2.0.9
      - name: instana-agent-operator.v2.0.8
        version: 2.0.8
      - name: instana-agent-operator.v2.0.7
        version: 2.0.7
      - name: instana-agent-operator.v2.0.5
        version: 2.0.5
      - name: instana-agent-operator.v2.0.4
        version: 2.0.4
      name: stable
    defaultChannel: stable
    packageName: instana-agent-operator
    provider:
      name: Instana
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hdfs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hdfs-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:e8b4ad2c3b127537239b003ae19a3f58b541d599fe3d5a74a6718612fe645d50
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:e8b4ad2c3b127537239b003ae19a3f58b541d599fe3d5a74a6718612fe645d50
        version: 24.11.1
      entries:
      - name: hdfs-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: hdfs-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:00c79036eed6dd6749736b964787acb638f1a6e0156c0bba5651b93738534377
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:00c79036eed6dd6749736b964787acb638f1a6e0156c0bba5651b93738534377
        version: 25.3.0
      entries:
      - name: hdfs-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: hdfs-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
        version: 25.7.0
      entries:
      - name: hdfs-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hdfs-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
        version: 25.7.0
      entries:
      - name: hdfs-operator.v25.7.0
        version: 25.7.0
      - name: hdfs-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hdfs-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: seldon
      provider-url: ""
    name: seldon-deploy-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: seldon-deploy-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1alpha1",
                "kind": "SeldonDeploy",
                "metadata": {
                  "name": "seldondeploy-sample"
                },
                "spec": {
                  "affinity": {},
                  "alibidetect": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306"
                  },
                  "batchjobs": {
                    "mc": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48"
                    },
                    "processor": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d"
                    },
                    "pvc": {
                      "defaultSize": "1Gi"
                    },
                    "serviceAccount": "workflow"
                  },
                  "defaultUserID": "",
                  "docker": {
                    "user": "unknown"
                  },
                  "elasticsearch": {
                    "basicAuth": true,
                    "secret": {
                      "name": "elastic-credentials",
                      "passwordKey": "password",
                      "userKey": "username"
                    },
                    "url": "https://elasticsearch-seldon-es-http.seldon-logs:9200"
                  },
                  "enableAppAnalytics": false,
                  "enableAppAuth": false,
                  "env": {
                    "USERID_CLAIM_KEY": "name"
                  },
                  "external": {
                    "protocol": "http"
                  },
                  "fullnameOverride": "",
                  "gitops": {
                    "argocd": {
                      "enabled": false,
                      "namespace": "argocd"
                    },
                    "fileFormat": "json",
                    "git": {
                      "email": "",
                      "secret": "git-creds",
                      "skipVerifyGit": true,
                      "token": "",
                      "user": "",
                      "webhook": {
                        "service": {
                          "create": true,
                          "loadBalancerSourceRanges": {}
                        }
                      }
                    }
                  },
                  "image": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e",
                    "pullPolicy": "Always"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": []
                      }
                    ],
                    "tls": []
                  },
                  "ingressGateway": {
                    "ingressNamespace": "istio-system",
                    "kfServingIngressService": "istio-ingressgateway",
                    "seldonIngressService": "istio-ingressgateway"
                  },
                  "kfserving": {
                    "curlForm": "MODEL_NAME={{ .ModelName }}\u003cbr\u003e\nCLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\nSERVICE_HOSTNAME=$(kubectl get inferenceservice {{ .ModelName }} -o jsonpath='{.status.url}' | cut -d \"/\" -f 3)\u003cbr\u003e\ncurl -v -H \"Host: ${SERVICE_HOSTNAME}\" {{ .KfServingProtocol }}://$CLUSTER_IP/v1/models/$MODEL_NAME:predict -d '{{ .Payload }}'\n",
                    "enabled": false,
                    "explainForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:explain",
                    "protocol": "http",
                    "requestForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:predict"
                  },
                  "loadtest": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c"
                  },
                  "metadata": {
                    "pg": {
                      "enabled": false,
                      "secret": "metadata-postgres"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "openshiftMarketplace": {
                    "cleanupClusterServiceVersions": false,
                    "kubectlCleanupImage": "registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382",
                    "prometheus": {
                      "monitorSpecs": {
                        "create": true
                      }
                    },
                    "seldonCore": {
                      "istioGateway": {
                        "create": true,
                        "name": "seldon-gateway",
                        "namespace": "istio-system"
                      },
                      "subscription": {
                        "apiVersion": "operators.coreos.com/v1alpha1",
                        "channel": "stable",
                        "create": true,
                        "istioEnabled": true,
                        "metricsPath": "/metrics",
                        "requestLoggerEndpoint": "http://broker-ingress.knative-eventing.svc.cluster.local/seldon-logs/default"
                      }
                    }
                  },
                  "prometheus": {
                    "knative": {
                      "url": "http://prometheus-system-np.knative-monitoring.svc.cluster.local:8080/api/v1/"
                    },
                    "seldon": {
                      "jwtSecretKey": "jwt-seldon.txt",
                      "jwtSecretName": "jwt-seldon",
                      "namespaceMetricName": "namespace",
                      "resourceMetricsUrl": "https://prometheus-k8s.openshift-monitoring:9091/api/v1/",
                      "serviceMetricName": "exported_service",
                      "url": "https://thanos-querier.openshift-monitoring.svc:9091/api/v1/"
                    }
                  },
                  "rbac": {
                    "clusterWide": true,
                    "create": true,
                    "readNamespaces": true
                  },
                  "replicaCount": 1,
                  "requestLogger": {
                    "create": true,
                    "elasticsearch": {
                      "host": "elasticsearch-seldon-es-http.seldon-logs",
                      "port": "9200",
                      "protocol": "https"
                    },
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398",
                    "imagePullPolicy": "IfNotPresent",
                    "namespace": {
                      "create": false,
                      "name": "seldon-logs"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "600m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "200Mi"
                      }
                    },
                    "trigger": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "broker": "default"
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "800m",
                      "memory": "800Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "seldon": {
                    "curlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions -d '{{ .Payload }}'\n",
                    "enabled": true,
                    "explainForm": "http://{{ .ModelName }}-{{ .Predictor }}-explainer.{{ .Namespace }}:9000/v1/models/{{ .ModelName }}:explain",
                    "kfservingV2CurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .GraphModelName }}/infer -d '{{ .Payload }}'\n",
                    "protocol": "http",
                    "seldonRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions",
                    "tensorFlowCurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict -d '{{ .Payload }}'\n",
                    "tensorflowRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict",
                    "v2RequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .ModelName }}/infer"
                  },
                  "service": {
                    "port": 80,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "create": true
                  },
                  "serviceAccountName": "seldon-deploy",
                  "skipVerifyHttpCalls": true,
                  "tolerations": [],
                  "virtualService": {
                    "create": true,
                    "gateways": [
                      "istio-system/seldon-gateway"
                    ],
                    "prefix": "/seldon-deploy/"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
          createdAt: "2020-12-11 15:00:00"
          description: The Seldon deploy operator allows the creation of Seldon Deploy
            applications to provide management, monitoring and updating of machine
            learning models from a range of technologies. It includes the ability
            to run explanations as well as outlier, drift and adversarial detection
            against running models. There is a full metrics and logging integration
            via Prometheus and ELK stacks as well as full GitOps operations to allow
            full auditing and recovery of updated models.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SeldonIO/seldon-deploy-operator
          support: Seldon Dev
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: SeldonDeploy
            name: seldondeploys.machinelearning.seldon.io
            version: v1alpha1
        description: The Seldon deploy operator allows the creation of Seldon Deploy
          applications to provide management, monitoring and updating of machine learning
          models from a range of technologies. It includes the ability to run explanations
          as well as outlier, drift and adversarial detection against running models.
          There is a full metrics and logging integration via Prometheus and ELK stacks
          as well as full GitOps operations to allow full auditing and recovery of
          updated models.
        displayName: Seldon Deploy Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - machine learning
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://deploy.seldon.io
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        provider:
          name: seldon
        relatedImages:
        - registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398
        - registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
        - registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c
        version: 1.2.1
      entries:
      - name: seldon-deploy-operator.v1.2.1
        version: 1.2.1
      - name: seldon-deploy-operator.v1.2.0
        version: 1.2.0
      - name: seldon-deploy-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: seldon-deploy-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1alpha1",
                "kind": "SeldonDeploy",
                "metadata": {
                  "name": "seldondeploy-sample"
                },
                "spec": {
                  "affinity": {},
                  "alibidetect": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306"
                  },
                  "batchjobs": {
                    "mc": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48"
                    },
                    "processor": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d"
                    },
                    "pvc": {
                      "defaultSize": "1Gi"
                    },
                    "serviceAccount": "workflow"
                  },
                  "defaultUserID": "",
                  "docker": {
                    "user": "unknown"
                  },
                  "elasticsearch": {
                    "basicAuth": true,
                    "secret": {
                      "name": "elastic-credentials",
                      "passwordKey": "password",
                      "userKey": "username"
                    },
                    "url": "https://elasticsearch-seldon-es-http.seldon-logs:9200"
                  },
                  "enableAppAnalytics": false,
                  "enableAppAuth": false,
                  "env": {
                    "USERID_CLAIM_KEY": "name"
                  },
                  "external": {
                    "protocol": "http"
                  },
                  "fullnameOverride": "",
                  "gitops": {
                    "argocd": {
                      "enabled": false,
                      "namespace": "argocd"
                    },
                    "fileFormat": "json",
                    "git": {
                      "email": "",
                      "secret": "git-creds",
                      "skipVerifyGit": true,
                      "token": "",
                      "user": "",
                      "webhook": {
                        "service": {
                          "create": true,
                          "loadBalancerSourceRanges": {}
                        }
                      }
                    }
                  },
                  "image": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e",
                    "pullPolicy": "Always"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": []
                      }
                    ],
                    "tls": []
                  },
                  "ingressGateway": {
                    "ingressNamespace": "istio-system",
                    "kfServingIngressService": "istio-ingressgateway",
                    "seldonIngressService": "istio-ingressgateway"
                  },
                  "kfserving": {
                    "curlForm": "MODEL_NAME={{ .ModelName }}\u003cbr\u003e\nCLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\nSERVICE_HOSTNAME=$(kubectl get inferenceservice {{ .ModelName }} -o jsonpath='{.status.url}' | cut -d \"/\" -f 3)\u003cbr\u003e\ncurl -v -H \"Host: ${SERVICE_HOSTNAME}\" {{ .KfServingProtocol }}://$CLUSTER_IP/v1/models/$MODEL_NAME:predict -d '{{ .Payload }}'\n",
                    "enabled": false,
                    "explainForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:explain",
                    "protocol": "http",
                    "requestForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:predict"
                  },
                  "loadtest": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c"
                  },
                  "metadata": {
                    "pg": {
                      "enabled": false,
                      "secret": "metadata-postgres"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "openshiftMarketplace": {
                    "cleanupClusterServiceVersions": false,
                    "kubectlCleanupImage": "registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382",
                    "prometheus": {
                      "monitorSpecs": {
                        "create": true
                      }
                    },
                    "seldonCore": {
                      "istioGateway": {
                        "create": true,
                        "name": "seldon-gateway",
                        "namespace": "istio-system"
                      },
                      "subscription": {
                        "apiVersion": "operators.coreos.com/v1alpha1",
                        "channel": "stable",
                        "create": true,
                        "istioEnabled": true,
                        "metricsPath": "/metrics",
                        "requestLoggerEndpoint": "http://broker-ingress.knative-eventing.svc.cluster.local/seldon-logs/default"
                      }
                    }
                  },
                  "prometheus": {
                    "knative": {
                      "url": "http://prometheus-system-np.knative-monitoring.svc.cluster.local:8080/api/v1/"
                    },
                    "seldon": {
                      "jwtSecretKey": "jwt-seldon.txt",
                      "jwtSecretName": "jwt-seldon",
                      "namespaceMetricName": "namespace",
                      "resourceMetricsUrl": "https://prometheus-k8s.openshift-monitoring:9091/api/v1/",
                      "serviceMetricName": "exported_service",
                      "url": "https://thanos-querier.openshift-monitoring.svc:9091/api/v1/"
                    }
                  },
                  "rbac": {
                    "clusterWide": true,
                    "create": true,
                    "readNamespaces": true
                  },
                  "replicaCount": 1,
                  "requestLogger": {
                    "create": true,
                    "elasticsearch": {
                      "host": "elasticsearch-seldon-es-http.seldon-logs",
                      "port": "9200",
                      "protocol": "https"
                    },
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398",
                    "imagePullPolicy": "IfNotPresent",
                    "namespace": {
                      "create": false,
                      "name": "seldon-logs"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "600m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "200Mi"
                      }
                    },
                    "trigger": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "broker": "default"
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "800m",
                      "memory": "800Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "seldon": {
                    "curlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions -d '{{ .Payload }}'\n",
                    "enabled": true,
                    "explainForm": "http://{{ .ModelName }}-{{ .Predictor }}-explainer.{{ .Namespace }}:9000/v1/models/{{ .ModelName }}:explain",
                    "kfservingV2CurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .GraphModelName }}/infer -d '{{ .Payload }}'\n",
                    "protocol": "http",
                    "seldonRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions",
                    "tensorFlowCurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict -d '{{ .Payload }}'\n",
                    "tensorflowRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict",
                    "v2RequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .ModelName }}/infer"
                  },
                  "service": {
                    "port": 80,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "create": true
                  },
                  "serviceAccountName": "seldon-deploy",
                  "skipVerifyHttpCalls": true,
                  "tolerations": [],
                  "virtualService": {
                    "create": true,
                    "gateways": [
                      "istio-system/seldon-gateway"
                    ],
                    "prefix": "/seldon-deploy/"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
          createdAt: "2020-12-11 15:00:00"
          description: The Seldon deploy operator allows the creation of Seldon Deploy
            applications to provide management, monitoring and updating of machine
            learning models from a range of technologies. It includes the ability
            to run explanations as well as outlier, drift and adversarial detection
            against running models. There is a full metrics and logging integration
            via Prometheus and ELK stacks as well as full GitOps operations to allow
            full auditing and recovery of updated models.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SeldonIO/seldon-deploy-operator
          support: Seldon Dev
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: SeldonDeploy
            name: seldondeploys.machinelearning.seldon.io
            version: v1alpha1
        description: The Seldon deploy operator allows the creation of Seldon Deploy
          applications to provide management, monitoring and updating of machine learning
          models from a range of technologies. It includes the ability to run explanations
          as well as outlier, drift and adversarial detection against running models.
          There is a full metrics and logging integration via Prometheus and ELK stacks
          as well as full GitOps operations to allow full auditing and recovery of
          updated models.
        displayName: Seldon Deploy Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - machine learning
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://deploy.seldon.io
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        provider:
          name: seldon
        relatedImages:
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d
        - registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48
        - registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c
        - registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e
        version: 1.2.1
      entries:
      - name: seldon-deploy-operator.v1.2.1
        version: 1.2.1
      - name: seldon-deploy-operator.v1.2.0
        version: 1.2.0
      - name: seldon-deploy-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: seldon-deploy-operator-rhmp
    provider:
      name: seldon
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: rhods-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhods-operator.2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
          createdAt: "2025-08-12T18:04:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "llamastackoperator": {
                      "managementState": "Removed"
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (including PyTorch, TensorFlow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack – unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.23
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fbe10282ada2fa6bf07ae97b497f46ebe735de130b4ce0182bf34301429999cb
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:2d2e7803b2b1e15017ca04961f9dcd25e89c47b07b101ca54cb9e793e6027202
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:e50adae13e9d773e7be8bf81b7525dac3f5e3864c9cdf4c287342bd568ee1d6b
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:bb7c352ccc10a8c60ef8d573368d0ada59e73cd131653d11bcebdf9f564acc10
        - registry.redhat.io/rhel9/mariadb-105@sha256:050dd5a7a32395b73b8680570e967e55050b152727412fdd73a25d8816e62d53
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:ccfc84028a95f1cc6a16e06ff6e9ed2ace74f74e5e9963a4fb33544c2ce005c2
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:2886e40f94fde72eb5959ad564bbdd466657208c09bac731b3860b0b54f104c6
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:b038d20def504c140e4365f2eb117a979534629332bec75a2ba750ebafdefe3d
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5fb7e6c3ee1e3b170cd146ea9cd172fa953326b08163897605ca74a3644f2819
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:795ab1842ef675e3d8b47deb8a7485a69064428e823dfeb930a6a7e7017cde0f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:80d284cac2ef6638fe028102b71a70f26041576c11a49414e77430954a3c1b99
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f23f8051886d0f0240937507471c818f440bca8703b8209a836d5dbe3ea2f7c8
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:16ff53f4f1a1e1e66826c8355c654038da83df545427a931485ec7a2d3764b16
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:23c098f7e791441cfcfa6bf37de51763511bd8bf4d543747b7d37ed591d68886
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:edcce7cddfa4c250108532e56541ddb95099612c180b2297e5934bc89d2bb46e
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:ff740fc3615edaa90dae976da1aa6455d64811e239eb66b422836a3aae791b41
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:240692b86bc8a4fe0b545e6e4cd7cf4ff0ca990406bc55f0446d565fd67e37f6
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:a4649e8ad4970bc3d5e967811c9694fa82d78ee9f8f025b8b2fbd37237eb14fc
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:68e74bdf518fc624cc25a59a18b2a87d0aa2a9b8aad30750ace7115d297a21ca
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:ca21e218e26c46e3c63d926241846f8f307fd4a586cc4b04147da49af6018ef5
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:857560240bb45bfc2db6d7743abf11a47f929a0d3035940256a6527afd313412
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:84c3213e11299ec032aa0e5a4dfa9b3055c4246b04b029e888d8d9f1caf5c30c
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:4847ac1be55284815670c2048c1edfc56c7227494280d3c6094aa23f1ed1a9d2
        - registry.redhat.io/ubi9/toolbox@sha256:4213516b6fe7957a01072720efd9ecbc7d4443238e151b5fe0263bd0c36792b1
        - registry.redhat.io/rhoai/odh-llama-stack-core-rhel9@sha256:43b60b1ee6f66fec38fe2ffbbe08dca8541ef162332e4bd8e422ecd24ee02646
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:467b04f4ee4118bbe0934d1ab8b80f72df8703dcdac61c2b53fdb5bd1b404c14
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:a1497e64349634b2567d28f685052de118f64f6b32e7d3ea469212e2004d7eb3
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:e0c9ac73891781ef421e0a020cfd5c962f7632d38020897183cf048b1819326a
        - registry.redhat.io/rhoai/odh-llama-stack-k8s-operator-rhel9@sha256:626b9ef09ee82c8d425ddb7ed3f2ffbb3d1019f541e31604db891787c881d5c9
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:3243caa4c8aac75069d19225dcee361bf9c012782328a6266b7016b3026f6d56
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:756b7bdaa7f6e7c4a18488e690188dd91153a33913d225bab001b0085382e514
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:da8debeb2b84a42aba0b98785ccc20e02ed520e125428fe012df3ca473cbb755
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:208005da389b7d4176ad8bffc703312011a1299954df151e2abbe8f3dec01d19
        version: 2.23.0
      entries:
      - name: rhods-operator.2.23.0
        version: 2.23.0
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.20.0
        version: 2.20.0
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.18.0
        version: 2.18.0
      - name: rhods-operator.2.17.0
        version: 2.17.0
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.15.0
        version: 2.15.0
      - name: rhods-operator.2.14.0
        version: 2.14.0
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: alpha
    - currentCSV: rhods-operator.2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhods/odh-rhel8-operator@sha256:a4d7f020509b9720063e87e1c2f19af04e4c9b4986b199be4dc617831b971434
          createdAt: "2023-11-21T08:37:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            Data Science
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:08c90a8f133b4cdb1477a742bc44be4d8fc4e88153b8bc2a2a148cbdf19f4e7d
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:6a56010de933f172b195a1a575855d37b70a4968be8edb35157f6ca193969ad2
          olm.skipRange: '>=1.33.0 <2.4.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "rhods",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "rhods",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Data Science"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:eb96f4b753606528750b3c28f7f9256582f81fa030450b05afdef684634903be
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift Data Science
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift Data Science (RHODS) is a complete platform for the entire lifecycle of your AI/ML projects. It is the flagship product of OpenShift AI.

          When using RHODS, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. RHODS can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * RHODS dashboard
          * Curated Notebook Images (incl CUDA, PyTorch, Tensorflow)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh (and Kserve) with a provided OpenVino Model Serving Runtime
          * Ability to use other runtimes for serving
          * Model Monitoring
        displayName: Red Hat OpenShift Data Science
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift Data Science
        - RHODS
        - rhods
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift Data Science
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-data-science
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhods/odh-ml-pipelines-api-server-rhel8@sha256:be57f237df156474a709667fa4fb4aa142118e3459963fed56648c1cb2e38cd6
        - registry.redhat.io/rhods/odh-kuberay-operator-controller-rhel8@sha256:bfc805aa1689a4bd6f7e617acf9eb1c2c307624dd49956e9e7442afac328151a
        - registry.redhat.io/rhods/odh-data-science-pipelines-operator-controller-rhel8@sha256:250db47cb873207cfb75c7e8a2f0de08b75b056a72db73362d2d79c08cb7b185
        - registry.redhat.io/rhods/odh-trustyai-service-operator-rhel8@sha256:68471555d0b2bb8150ebfae8d4f3d1dbdff9c5b961ff2305ef23ab7514cbbea1
        - registry.redhat.io/rhods/odh-codeflare-operator-rhel8@sha256:2841fd3051a50a1b0d9e3caafa20a8084b5c472d49393b372828960dbf156674
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhods/odh-model-controller-rhel8@sha256:fde989817cd1a6ec8eb408a7f6c73e28ba043ab2d88035586b1b7e55fb3405cb
        - registry.redhat.io/rhods/odh-modelmesh-runtime-adapter-rhel8@sha256:08478a1647d0413783bcbcd966a73b444b580464de46efedb39504420f22053e
        - registry.redhat.io/rhel8/mariadb-103@sha256:08c90a8f133b4cdb1477a742bc44be4d8fc4e88153b8bc2a2a148cbdf19f4e7d
        - registry.redhat.io/rhods/odh-trustyai-service-rhel8@sha256:fe17d6e07ebaa088f274dd8b5e83892dacc11cb336065b06582a3c055095dba5
        - registry.redhat.io/rhods/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:c5758102609821a6fdabec0277e03c4c728f82176093d1ade0570aba5f3cd8f9
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhods/odh-ml-pipelines-persistenceagent-rhel8@sha256:857a8eb444375505ccb34a11cce69c1794ceabbfb00df06933d7f12cd187319b
        - registry.redhat.io/ubi8/ubi-micro@sha256:6a56010de933f172b195a1a575855d37b70a4968be8edb35157f6ca193969ad2
        - registry.redhat.io/rhods/odh-rhel8-operator@sha256:a4d7f020509b9720063e87e1c2f19af04e4c9b4986b199be4dc617831b971434
        - registry.redhat.io/rhods/odh-mm-rest-proxy-rhel8@sha256:8d989f54009484d7f43fe79d7c37757418fc3c15cd4d7e80a2076677a9645bdf
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhods/odh-ml-pipelines-cache-rhel8@sha256:33aae32e3d348d382906f6e926281852f54006a94d7bdac188bb33debadec059
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhods/odh-modelmesh-rhel8@sha256:3827221f84d3aa22440d6b364b990a4ec0666dfb5963f174f7ed518b3650d223
        - registry.redhat.io/rhods/odh-kf-notebook-controller-rhel8@sha256:cd0ab8d70559be6074fbaf1caa1d50f8eb75e1bff1a22d6d69380f19190df0b6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:eb96f4b753606528750b3c28f7f9256582f81fa030450b05afdef684634903be
        - registry.redhat.io/rhods/odh-ml-pipelines-artifact-manager-rhel8@sha256:76c67d7015a2db6965219db6533e0cddb5b7cf699c88df3ef81155f9c192a3cc
        - registry.redhat.io/rhods/odh-modelmesh-serving-controller-rhel8@sha256:d94395fb251e45aaa73d09bc6106a2947c16b90d87fc3b64bfe4f2f600be5841
        - registry.redhat.io/rhods/odh-dashboard-rhel8@sha256:b98171c5543815abf7c095fba92796ceed9ea44bd8b78db143d2815c6bcfdcf6
        - registry.redhat.io/rhods/odh-notebook-controller-rhel8@sha256:753994fce63802f1e1121990a5c1dd722216446abee54cc1a7c50ab195134414
        version: 2.4.0
      entries:
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: beta
    - currentCSV: rhods-operator.2.21.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:7d0b8d69b6defb5f232da066573b0067ec810a8ccfb2f3cb91d472d2979ab7a7
          createdAt: "2025-09-24T14:42:33Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.21
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:687c8eeed55f021ecaab1307f0e88b5b16d91f72d63b3d7168d7bbee90e8947b
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:04859c878ffcc8c42550a98be625ac52212f9b79a6fd5b3baeafb3592926cbe9
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:2a16c353fd13dc1f3f5c4266f1b9786a8e7a100dc05f2d554bdaef082271063f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:239f226328e5ac77d6dee498217d55103a6d562f4829e38abf65ba55fda0aca2
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:7da004e158006dcbdcf0c3b78c40411b82c9d56ebf54b1c30c245b6550b8ee7d
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:2b4136569e10ad7d47c82bedf55c1705a833dc67c8f833656996bfe1fddbef3e
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:52e1ba8ebf4d55c64c94c9076185c5612b537e6e5ed4944fcd6ea87ccb9cd182
        - registry.redhat.io/ubi9/toolbox@sha256:7656f815a0d13ed37f56d11635559a87a3d00840253755c71b644b858f67d556
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:7d0b8d69b6defb5f232da066573b0067ec810a8ccfb2f3cb91d472d2979ab7a7
        - registry.redhat.io/rhel9/mariadb-105@sha256:c390b12f5c2c3ea7cd897b59eba7127ce22f474130cd8c1bbbbaadae870baca4
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:1b5452b87b36de2b62febc777efdc6233a168ffcc09e2067e482cc444463a57d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:cb47e165935f0d9d3c88088689c37e68a69519357a20d3c52ab6f71240dab507
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:003fdc6cd1821d620e1ce4050ee608e2d2a32f2b08c8fa9c147e2e14418a8d91
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:80fd53a580ce7bc4d02023d38a04912e84c2cd50679fc5ea034afba935623157
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:fe6593bc73e77d315d82b7921bb1975c19fda1e7630ef5b30b69f5901ce30178
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:e68492a6a75cb83be0a968b832c1478b97f14e661376328085fd5416a08cbc31
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:c3e283c0536266195dd461042ca0036ed98cc33650d806be132db4fba45239e4
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:152b425725228953f43b57a4a034a6725993a5ca556af328ae5e06aa3f1acbfc
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:a88696cff8cd2cc2e295c9ec082b7f25f61dfc9ac12f1c208529ff7036d5af29
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fc98ed817a4e9ab84e2a66063987304f0cdbe6e5620a58a7f25cccd61b34ebf9
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:5c0d56f434298f6e8bf6cf7061cd1ae6df9915225cbf04b010b6b59a64775cc9
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:cfaf4eb86ddb9c717340ab17fa4b83323e5b40977ac70265376c87329a3f6f20
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:4c57dcf4df09fe0d03578cc8b9a3ca2ea485b6af0ca0203bfb85e640058b7daf
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:4c192208c618b49f31cf31c9de1a33c622ce3f120204c2eec8ff7c439ea6785c
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:a936d87646843edea2a84eb429bb44eb730bf3e04e8900fecef66721426dba89
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:9c2d6ce05775ac34ee5695b3367d12bf2a9c14e318a9544d00a56653ea5f2007
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:dc0db686b7688f98b5f8595661022cdffa5d26a078289cc7aed528f0ee1d08a4
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:6a388ee3cbe3813ac955d319f0ab65a81da321d982c9881c6c81fe135311fe1d
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5f581957e4f8deebf6104a394272e6fbb1c54d813bf069e95875d865eac209ad
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:2638e3a90bbb242dd2483fefe9fc8b472f334c510d0139b9b7c875c30dab7984
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:85202cb75250a1ba2691d65b52b306fd77566660dd29cdf56bf8962e8bb4939c
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:1b7befcd29c9c8f783bdd3c5654b0412ea45cd88cd18eb5f8f5a411f1323c6bd
        version: 2.21.1
      entries:
      - name: rhods-operator.2.21.1
        version: 2.21.1
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: embedded
    - currentCSV: rhods-operator.2.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
          createdAt: "2025-09-24T13:27:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.16
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:22557a6fa52d2f311750a9ba253860f423ba697d26efa02ef8524a8258d2a909
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:d7d4fa406e0fcf0507894a7676532b27f45be742467e603a86f98ea5d2615df8
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:2ba2b2c4db8bb334c50f4dfb54059f060361186900a44c06eed00b7a3c43977e
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:7b70af8847b0806d8b43c399e2b3109f016fd864f5e9d30c44e2baca5d1359dc
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a0f72ffefb2b74b488dd949493f5d295a39bb9c97f578bf219d0138601f65468
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:6cf74044ae8d5308a2dfe03fa5d81086c89302db7f4cdbe2f4174a1c48b77869
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:de5f91180ead7d73a1825fe8b032fde9b8b01392569c9789f5ce1b4b9c08a98f
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:ee01e89f98feb185f6cd59c564e590a13e4d8d9ea760cca8de51426eb71b83a1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:04f305c44413ae7dcb7017e53570ee49a509701792c5f50efadd64f47395730b
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:c8456d98e90b6505957ab3686e9fd2f156e29f123c5558e581c206daf1e7d93a
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:f738aab1eab25854c93e1d8d4d98100a8ae7bb45a6b83f0326774e4220b1183b
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:5dcdcc2424602a69451f16d31dbfa1d43cb72c095ba561eb9076f0cd1e8182ed
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:efd039012559786deb3c521a130886c265e88d635e08baace2e573a9df270134
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:27b113d96453c2054d4c965a963d2badba6daac235eddaf23234c76e87bcb069
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:4c5ff3496b2a2a739939d94ee9dafc02b682100785d228dc2fde480fb597b7a5
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:4a7599f8a866eb169c9a62885906adbf6df0417c0be15857df1eef20cd9b1be2
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:633984b538f027c93c5886ea1045dcbe81eeda74acf80001d5fc5f765bdbe0be
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:23d307a36b69e0df04f72a7d3b35e28d8417a8bbe23dba31e8e977569785c078
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:11a301728573adf2b64ea72d0cb2d83ea5d4dbebea759f346e99f18c3d368c6e
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:5192f1dbd2a9ab92ae390c4ae506efbed0970545b6122e95b014728ac937e777
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:2eedd1e2465d2e78fafb39e234dbb413a82713f187c64d70a80340ec94807d30
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f37e4048f3a152798286793f3abfc6ed814453fcbe2667255a7e78eee483c5a3
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:97e2bd9b587f08e135a9aeb9b3e0dc6eafa1a9bdacbb5ecb681ce9bd5aa37fc9
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:65482864055021272a18b5b26792ab00cfa5fc9cc005d8d3a884cc82991506f3
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:4a212ec634225c14beac09be24ddf336e562f2aa9a13555fb1196f366ddae23c
        version: 2.16.3
      entries:
      - name: rhods-operator.2.16.3
        version: 2.16.3
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.8.5
        version: 2.8.5
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: eus-2.16
    - currentCSV: rhods-operator.2.8.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:3b9e8a71ba795a2ea91f4129253ccb4be3a79c7c1c6a87082f96ccc67eabbbac
          createdAt: "2024-11-07T07:17:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.8.0 <2.8.5'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:eeb8bdc94e7fa2b454fedf72eb75756052f8bfb9d92b116a8267a5584e2528c1
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:f9c474d454d0b006e06859601ec9a64b5e5c91e9a47bfab6ad6d8cce5aa6b89d
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:45498198fadfbc1457bf2a6084ae7ea515d454d86d148783aae3abdaca6e999d
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:60507a3dddbd88560fd85204699147facd2b6371fb25e0bf6701dad02c0b4dd4
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:df34e9d2d121a3def2dbb86f108b3fb6948bc7ba7cb44c70801dfd2e81d3fbb1
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:13dff890ec67fa56864c244a40fc404e9eac05f33573dd16d088b37f9340ec41
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:28ec01e514c15cae145499840597d1be1b750123bb6014de2ecc3ef0d60fa134
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:9aa5a93831992588766ac8594ae3c4dc7111d08a90fdbc8751c40484339e46d4
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:038ce76571217d6c624d614652e8b96ef6484562cf52ec1be88392b8ef8a3bbd
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:1a1074263ae26e4c8b0e3863228c35e50133169f377ece6c47b58a12e0fcf453
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:81ae2b95f2da740cb5264b9ba8b85f0f58bb4ab80d6431f66d52848ed0c21afb
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:1dbf09cb1624c9f961f8ac8e371c97aa7c3fbfcad4c4d02ab4fad0eb86d2aa95
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:61a506fbf263cd72951ac8bd151127e0c74fc74f0a05f7aaecaa54fde47ee490
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:3b9e8a71ba795a2ea91f4129253ccb4be3a79c7c1c6a87082f96ccc67eabbbac
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:13b850a5a3f8776c6c8db887c666d88df72901895b5ef1be9a8bd372cc02fcb8
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:8ffc5da798e783c65d5be7f344219b2951d8c2233d6494a4ae2bb5c86eccd699
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:d87e4b1cba72eb4f50b3f2d59863799929c22ca37b01a1ff4828a9ce0be31f25
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:2f6db789ceb3f90f9c9911c6392e62d0a8e4d61be68a7f7cbd781f47faab6389
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:4b1b9f05a94713292a7a2f6782fe1d3eea6f36214c36b38f402aa91a554ab18a
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:5c48474563d98df32e7243e33302428ea8836be5e1a2e30d20b4d796a62cbf6b
        version: 2.8.5
      entries:
      - name: rhods-operator.2.8.5
        version: 2.8.5
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: eus-2.8
    - currentCSV: rhods-operator.2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
          createdAt: "2025-08-12T18:04:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "llamastackoperator": {
                      "managementState": "Removed"
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (including PyTorch, TensorFlow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack – unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.23
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:edcce7cddfa4c250108532e56541ddb95099612c180b2297e5934bc89d2bb46e
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-llama-stack-k8s-operator-rhel9@sha256:626b9ef09ee82c8d425ddb7ed3f2ffbb3d1019f541e31604db891787c881d5c9
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:ff740fc3615edaa90dae976da1aa6455d64811e239eb66b422836a3aae791b41
        - registry.redhat.io/ubi9/toolbox@sha256:4213516b6fe7957a01072720efd9ecbc7d4443238e151b5fe0263bd0c36792b1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhel9/mariadb-105@sha256:050dd5a7a32395b73b8680570e967e55050b152727412fdd73a25d8816e62d53
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:68e74bdf518fc624cc25a59a18b2a87d0aa2a9b8aad30750ace7115d297a21ca
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:a1497e64349634b2567d28f685052de118f64f6b32e7d3ea469212e2004d7eb3
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:e50adae13e9d773e7be8bf81b7525dac3f5e3864c9cdf4c287342bd568ee1d6b
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:84c3213e11299ec032aa0e5a4dfa9b3055c4246b04b029e888d8d9f1caf5c30c
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:ccfc84028a95f1cc6a16e06ff6e9ed2ace74f74e5e9963a4fb33544c2ce005c2
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:bb7c352ccc10a8c60ef8d573368d0ada59e73cd131653d11bcebdf9f564acc10
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:2d2e7803b2b1e15017ca04961f9dcd25e89c47b07b101ca54cb9e793e6027202
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:b038d20def504c140e4365f2eb117a979534629332bec75a2ba750ebafdefe3d
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:3243caa4c8aac75069d19225dcee361bf9c012782328a6266b7016b3026f6d56
        - registry.redhat.io/rhoai/odh-llama-stack-core-rhel9@sha256:43b60b1ee6f66fec38fe2ffbbe08dca8541ef162332e4bd8e422ecd24ee02646
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:23c098f7e791441cfcfa6bf37de51763511bd8bf4d543747b7d37ed591d68886
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5fb7e6c3ee1e3b170cd146ea9cd172fa953326b08163897605ca74a3644f2819
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:857560240bb45bfc2db6d7743abf11a47f929a0d3035940256a6527afd313412
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:795ab1842ef675e3d8b47deb8a7485a69064428e823dfeb930a6a7e7017cde0f
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f23f8051886d0f0240937507471c818f440bca8703b8209a836d5dbe3ea2f7c8
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fbe10282ada2fa6bf07ae97b497f46ebe735de130b4ce0182bf34301429999cb
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:80d284cac2ef6638fe028102b71a70f26041576c11a49414e77430954a3c1b99
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:467b04f4ee4118bbe0934d1ab8b80f72df8703dcdac61c2b53fdb5bd1b404c14
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:ca21e218e26c46e3c63d926241846f8f307fd4a586cc4b04147da49af6018ef5
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:e0c9ac73891781ef421e0a020cfd5c962f7632d38020897183cf048b1819326a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:a4649e8ad4970bc3d5e967811c9694fa82d78ee9f8f025b8b2fbd37237eb14fc
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:208005da389b7d4176ad8bffc703312011a1299954df151e2abbe8f3dec01d19
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:240692b86bc8a4fe0b545e6e4cd7cf4ff0ca990406bc55f0446d565fd67e37f6
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:756b7bdaa7f6e7c4a18488e690188dd91153a33913d225bab001b0085382e514
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:2886e40f94fde72eb5959ad564bbdd466657208c09bac731b3860b0b54f104c6
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:16ff53f4f1a1e1e66826c8355c654038da83df545427a931485ec7a2d3764b16
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:4847ac1be55284815670c2048c1edfc56c7227494280d3c6094aa23f1ed1a9d2
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:da8debeb2b84a42aba0b98785ccc20e02ed520e125428fe012df3ca473cbb755
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        version: 2.23.0
      entries:
      - name: rhods-operator.2.23.0
        version: 2.23.0
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.20.0
        version: 2.20.0
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.18.0
        version: 2.18.0
      - name: rhods-operator.2.17.0
        version: 2.17.0
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.15.0
        version: 2.15.0
      - name: rhods-operator.2.14.0
        version: 2.14.0
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: fast
    - currentCSV: rhods-operator.2.22.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
          createdAt: "2025-09-25T11:28:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.22
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:7b3d7787ba95156de9c284ad78f2d63a0d3bcb525f4d8d11ebf0203d4c5c1fe0
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:7fd540586a1ca111370100a473923b3dc6cec4255e9e4f3654bfbabd9158a2bb
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:a44d32a18b1e906c1411cafc82102881573fcd04e040a4e2ba6bd1bd617b70fa
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:6234e15f001122ab222481321153fb0eba7f4d86bf3dfd64edf3f6cecbe7d62f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:abe1f9dd11c7313fdd55b678be511a5cdc3546654a68e2b53d42c1013fec8d00
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:b2ca229bb4c4dbf04b96766b8af8d9f55f2edef59ec4110de0d20f8c857e7d27
        - registry.redhat.io/rhel9/mariadb-105@sha256:5e77b33ce04c1eafc2e8840586ad6a5b51148dea8a46dda151071e28858331f0
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:2f3f305c97e50057781553416fb4026cde306ead0fbe0213627b9b5bbcc8a0c3
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:e6fbe6723b4801f6854303b3fd7d4a4a36727bdd28dc28002b7de4929739ccc0
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:2f1a0a0803b2f5118b167e9d9c2e86660add5f79ee6f47044f63b0e8abe0b75f
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:cb5be3dd7098f4c8d66f9e85ad04656460df6102652deddc6cccb884a12be733
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:9104cbc2d595933f0a2acb7313e41ed69905e76a9fae6bd9820e7717b56a632b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e3b90530c6e38f6940d7f9408e9fa2c81eb06a857ec2b0bab5d006b1e9a63a4f
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:32c9511846077ef6035eb19be7653f6a78e0f207eb420ab93b18f742ab5779b6
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:27415b1c763cf89b79391f0b039936c3e078a078bda4420f13bf10ef9be31d5e
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:463e9be9c4409ba1e5a42713bd65844a023e7a3e9032d4e862edbaa5fe35300a
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:1709fa3c79aad4ba7eb9be8299949396092c8e20210124e0c0936385bc04e839
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:b2c2ea9d35edbcd16e1ae559c88a453129b877fc049ab7a7cd6c2ebb8f0bcb59
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:bd6d1599b4ff9eb890be9808014715b28ca19247222d65f40df024db4dd3525a
        - registry.redhat.io/ubi9/toolbox@sha256:ea4ce8cb8f59cf1b0709f1606d7cab074c3806d4016964eedc0555410cb5dfeb
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:d59ed40de694bc0e0564bab159a5b586cffa0721b6d761fb2d81dd42e6225d1c
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:74d4e0174655cf6f6273857325a1e714a1531c5a06ad38bf139bee9c7d8f5f19
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:5d79f0bcca0a213b53aee6d88834e503d343821596b857b13d3a6362de8ebeab
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:5914f71e37c815ce2567826896e051303bdef52fae1acf8d02e672fc24c7eb9d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:2065b96bd6f0a830eab66e36217a927448cfc5a7dabf2a5e7d907aa6c40657a4
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:2b3aace0d80474470e71c2c8291b25738e096109229fe6d5d3e06e9885ee58b6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:13f3679ab289eea9de54fe129a7c83f6c8ec5d26b61b430637a7c4cfcd31a74a
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:ccba67a6a5aa3403e0018f29dcdfef483220dea8065d8b572e80c653ca992add
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:2fe9473cd1d6c445d2ca7cc357ee4bb009aa1838f9cb69ea418296b2a044816b
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:d7c4ac937add511460a36375d9b6b3c475a3a9a7cc96c00df2fbad6219aee0e9
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f9965177de24f3165ec4c001c14cf1e2c2c01ffe847839af837c442e3c96beae
        version: 2.22.2
      entries:
      - name: rhods-operator.2.22.2
        version: 2.22.2
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable
    - currentCSV: rhods-operator.2.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:8643f7ee8afacfe6ae08cbbf81da76ca97032dc3ce3da4fddf16e330e9e33a08
          createdAt: "2024-11-14T21:18:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.10.0 <2.10.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:a1d590d0e2b3fae9ecbd843a664dab083061e797f66a9f0a021b0c2c487d7e45
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:0af6a91553f40ea40e86d85edc204658af3e5a45e8ab94cdb6b767e545df59b9
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:214ca0608fcbc0fd585bf01768d28c7ba17082d1043349ec8445afd663095383
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:d2514bf4b4b2d9abc19facac3f1a55405eb3ad0d9e9203e8816bfcfa08c9a6c5
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:4876f14ca9cff0e2b998a56f2f56d0775edf3963de610f4b810b4d94fb12d719
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:24238abd46633709feb75f2a3e41cf6c9e8528d2c0e4be2fac6f56de44e3fb9b
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:d9b55e8f25b81050380b4aab040243dc18a7432c48912c9b660c97dcc5eea008
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:8643f7ee8afacfe6ae08cbbf81da76ca97032dc3ce3da4fddf16e330e9e33a08
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:9f7b1cb8623de8b0550b0ba124d3b50f5e960ed011b33a3262c6346dbe2dff95
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:807ec05e9e107afc969e718d0c2c980959e2fd462e777a5ff63e55156ea150a4
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:fdc2bc06b0dfa79ebc6451ac3793b5e8d9f165171633831dafe46ebc18c10c35
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:4da77afafde99b15fd357a4ab6609644bf6330a1d9c0c03e64726ed1b9633230
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:cba9f25ac6bb2734bb990a622b515c838b881edb688bf44ff196ebe7684ab74b
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:8911afbb65c8c4f10c3597624485be54b9103e87a5020b718b35ec3df2c9bf00
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:d709f65e1194bc877fa4f9ec681b598467703d94139d57b58c24490d3a59fb82
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:220cad7a22d40998063c6b4fb2f63e69330e64ca7118490a6680b05fc99b7733
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:8fed8f58732bff512289fff738d0a94be67286c41b24eaf5928bbcb30d88c88a
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:232ebac25084db056fdde49477b5c1e6a03a0228fea3758bcf6234b3d0eaa582
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:2007da5abf840712cd2edb50462b2d8d571a0193c3ae1550d4a279eddf2b335f
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:a4b9553fde02d9c67a9c98f063eaf82fb1d2689f4b1269e57bf5f12c8d4adecf
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:5d15175f934017c04d6a6057200ff55b2172a1f4bcbeb0f2cd9e9e545631d93f
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:f4d7d3ba4a0d3322bc9f327e095e9ec56073758f6dcd8da8c7bb185cf952ead5
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:7df2f8bc2c5c0f4f347d009b62a058b5107413a2abd789f6010eb4d6c9e4f8ee
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:a9c6e3afdf605465766fa93334cc0f6bb5e46a5fdfd7ec64d648e2ed72b05ccf
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:267d6795bf53f9a1955d73445776a428dd042af0cd3cb8bb5b91692a1bc1e5d2
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:65db5e8cb2fa46d020ed4837d4479d133b52167c373e65f81cb590eec5391af2
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:0ff5664c2eaadf3d3073b3d7fb711da47fb78d21f67710d8b8aad5452cb0355d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:12741379ab24d2286b5fae77f4383fb65713af7b3356fb43d8254ea85d4b0ee3
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:52e8deeeeaa21b17370a5d5cd644745da745b6e1cee8c6295d6df75c8d2aec04
        version: 2.10.2
      entries:
      - name: rhods-operator.2.10.2
        version: 2.10.2
      - name: rhods-operator.2.10.1
        version: 2.10.1
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.10
    - currentCSV: rhods-operator.2.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:28b5fff602ad6fb36d462e1905f453509ec22fc251a1e0cf64fedba86ff420d7
          createdAt: "2024-10-17T10:53:18Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.13.0 <2.13.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:cd6b535d95eb1b44fe261a5336270fd898b966f431c14e173944484381c26d25
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:c2bd493e9345eccd124bbe0440e45cce8cf57588827d8eb9533f8bbadaa3ebbe
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:74e6cfa0bc481588cb8dc61c206ad567fbd2b28f43c2a86bee4a8e0eff55447c
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:28b5fff602ad6fb36d462e1905f453509ec22fc251a1e0cf64fedba86ff420d7
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:77bca19180f30ef08f4d1598ac6c365e190bb7909ac4a1fb35d9ba9ba8d6ff7a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a37529eff3656a570d796083404e5a3f63603afecfe9e738da1a243c0cb3d7c2
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:5d6354e65106460ab9285b42124faf288298387a8b16317565bd591f009c39a6
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:fb630d6f5817512733b8ea05f7147121928917a24ac8d98239f43d920880c359
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:6b2c0d1dd40e2c276684e17818b14375bc778e6c88df0bdc3cc24bfc9294804f
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:8883bf71abf39b4a6985940c287eb6ac7dc137d491264805d48bb112b65bd008
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:9ccdd7fe8223057a2234a3a2eb3b492e7b69e6ce714cc26a805fe2dd82935195
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:850901511bde065cf73e4c18d787dc10b6e91ca6139fc3ba68da28c41dd28236
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:688389195aebf3f43e74e40d591d2fd5f9ae41994d8ac2db661f292f6eda012a
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:2566b73e98b783871fb500bb1a8cc4e3d58a513c1bb18abfb525da0818e6138b
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:fd4ce524acc1769475079c7231f5ea79ac29551c513855ef34e5694d2c9fbbd3
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:a55f73d72ad2b07f01c49949cdb00969bfe4789375b3d7cdf9814b969fcc808d
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:7d085942e4da73c2ff4885c62f10b904dc70fd08e819e62f925eb1b4ddfa1554
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:8faab013568ee023968a5ec5b3a79977d7ddae3deafff6328e6fd56a27198e24
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:71abac4fb3d613edef7f3ba117a73bbeea61f341d6182e8310360a76576c789e
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:030c53690fe25e15bc9907f51c3f04fab39b4d085ff2098d26316ba32cc8b09f
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:8c19c7fa535f36bdce3b73cf3300e321909d44954cbc63a8aef6321c5f2fa693
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:59e9cc9cd57ba48ca7e87ae1e797a902f826bb890cb198ccf549f76ea581ba6e
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:bf25fce160dd7b8cce11e68a501006e4fa81a7858645038cdefdbba03df7ecdb
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:a3048eaa8924ce3571c6e78014438183fae70d9e8c833a7c07774c675104e3d4
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:5c2ee4136ee0014d1ef395b5e3bcab125c846f5e258ece3631d912f431c8d208
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:7c895490d49beb6bd724f47ad1b17ce501650557b0889314799c4072632f2b99
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f5b17f44c5ab72d6cf658f72f24f15e06a5ac9be1790336f7228ab8e7cfc15c3
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:cf3779c64bbaef1988886ed629c516e7d6ef7bdc73b9c95ad4c087b62e75750d
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:363a35c332fc80b068157951d5e50e6b7209947649cc64803ffc02726e7581ab
        version: 2.13.1
      entries:
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.13
    - currentCSV: rhods-operator.2.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
          createdAt: "2025-09-24T13:27:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.16
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:97e2bd9b587f08e135a9aeb9b3e0dc6eafa1a9bdacbb5ecb681ce9bd5aa37fc9
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:efd039012559786deb3c521a130886c265e88d635e08baace2e573a9df270134
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:d7d4fa406e0fcf0507894a7676532b27f45be742467e603a86f98ea5d2615df8
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:7b70af8847b0806d8b43c399e2b3109f016fd864f5e9d30c44e2baca5d1359dc
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f37e4048f3a152798286793f3abfc6ed814453fcbe2667255a7e78eee483c5a3
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:4a7599f8a866eb169c9a62885906adbf6df0417c0be15857df1eef20cd9b1be2
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:4c5ff3496b2a2a739939d94ee9dafc02b682100785d228dc2fde480fb597b7a5
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:4a212ec634225c14beac09be24ddf336e562f2aa9a13555fb1196f366ddae23c
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:2ba2b2c4db8bb334c50f4dfb54059f060361186900a44c06eed00b7a3c43977e
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:f738aab1eab25854c93e1d8d4d98100a8ae7bb45a6b83f0326774e4220b1183b
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:23d307a36b69e0df04f72a7d3b35e28d8417a8bbe23dba31e8e977569785c078
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:04f305c44413ae7dcb7017e53570ee49a509701792c5f50efadd64f47395730b
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:2eedd1e2465d2e78fafb39e234dbb413a82713f187c64d70a80340ec94807d30
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:5dcdcc2424602a69451f16d31dbfa1d43cb72c095ba561eb9076f0cd1e8182ed
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:11a301728573adf2b64ea72d0cb2d83ea5d4dbebea759f346e99f18c3d368c6e
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:27b113d96453c2054d4c965a963d2badba6daac235eddaf23234c76e87bcb069
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:c8456d98e90b6505957ab3686e9fd2f156e29f123c5558e581c206daf1e7d93a
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:de5f91180ead7d73a1825fe8b032fde9b8b01392569c9789f5ce1b4b9c08a98f
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a0f72ffefb2b74b488dd949493f5d295a39bb9c97f578bf219d0138601f65468
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:6cf74044ae8d5308a2dfe03fa5d81086c89302db7f4cdbe2f4174a1c48b77869
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:633984b538f027c93c5886ea1045dcbe81eeda74acf80001d5fc5f765bdbe0be
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:5192f1dbd2a9ab92ae390c4ae506efbed0970545b6122e95b014728ac937e777
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:65482864055021272a18b5b26792ab00cfa5fc9cc005d8d3a884cc82991506f3
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:22557a6fa52d2f311750a9ba253860f423ba697d26efa02ef8524a8258d2a909
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:ee01e89f98feb185f6cd59c564e590a13e4d8d9ea760cca8de51426eb71b83a1
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        version: 2.16.3
      entries:
      - name: rhods-operator.2.16.3
        version: 2.16.3
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.16
    - currentCSV: rhods-operator.2.19.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:c605e5fdb3d66302781146347df6da5a91dee97e4f8f5d9ac33f105b8cdee383
          createdAt: "2025-09-24T13:27:42Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:b30d60cd458133430d4c92bf84911e03cecd02f60e88a58d1c6c003543cf833a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:f0ee0d27bb784e289f7d88cc8ee0e085ca70e88a5d126562105542f259a1ac01
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:8ce44de8c683f198bf24ba36cd17e89708153d11f5b42c0a27e77f8fdb233551
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.19
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:8a6732ac3ca60f13dcb1fd8f186d2fc5dc601b9b5981a58f039af9e237f20044
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:b20ee16b776871d7c49b23f2bc91726812998561129f5dbd14eddbfe57964c93
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:c605e5fdb3d66302781146347df6da5a91dee97e4f8f5d9ac33f105b8cdee383
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:b30d60cd458133430d4c92bf84911e03cecd02f60e88a58d1c6c003543cf833a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:f49483df1d342d8e39f9a8085920aad79313a475df02a2292a136ecb784402b5
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:c8c23c07038f59db848dea0d4247b62840dbe58e1c581dc0722fb9d7f9be3b75
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:4e8f6c6ca0ed516b8310af0332637683b5bfc8f2946eb935c647698dffbbcc79
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhel8/mariadb-103@sha256:f0ee0d27bb784e289f7d88cc8ee0e085ca70e88a5d126562105542f259a1ac01
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:d24e277f468b0a48f38801e8e0bcfeb8264c3f7c128073ddaa17934b8572cc3c
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:520ed21b80e5487b3a0a13b7e5c7d791d4dd28e0b6799d57abb65d53fca5a03f
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:6275a19ef06114b33d79833a94ef10cc8203368eee98e5d415e73d20cae1db85
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:87685c5cab105d876ff0d14b0e24d8dc2a1ff5b55bebe36028c5f3376b2a5271
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:6ad35cf5c4c62ed5ad4e192afe8cebead9796cb24a964401c76ff8656c47742f
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:b6039c3c33a5dc345ceb8e757437bb13023d2bf85cc2b4c10b9ab3887e78e0c0
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:6f01ac7cc62ab961b9c8a630a558338a61f6d7902301c4a1ad337495ac7ce85a
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:4313cae03c041d461bc276abc23d706b582b5d16feab316796ca43478b9dc705
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:b24c25dfdcefb4ec26e7f311c00798df108f7685def2b1762b7624fdfb160ebf
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:bbef4839f24facce111d21aca43898dcf4e6cf57adb16200e3f1590bd5f89eae
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:49c02285fcef6949892cecca20eeee49d03a9bb4d17bb9ceff7c7a0dbbb8ab40
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel8@sha256:6166f525948a018c41d735014d97fdfa52024f7be31e18976d85d1a0d88901a0
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:baefbe4e7dd5aa74c71cb81da4a7a2b888a02d3eb1398ec4419bd8003b77f0c2
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:8ce44de8c683f198bf24ba36cd17e89708153d11f5b42c0a27e77f8fdb233551
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:af4721d61dfb61c75488a1bd6a5f62b23e20d7b7586a9cf51ba062ece2837715
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:bf6231bfdbe52b597d86e03e4730ec105ebe7a639fcd529268b7c0423f2a890c
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:71cd15f0b1acba3a48ebedff037bf8a846b426c18363a4de8896a376b6b7b3d5
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:53ac36baa374159b9065c718a9ede821bbb61d9ebe9502b2243e0a9f7aca0d16
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:d47514cbceb3cc96e9d33b7db6fc1df60893e5c4118bf6086cdd27f04c472773
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:876ca7ce7b0b975688fb4a77d6cd1a6f87cf9eb40c2a9a5b84d8cffdc80eeebd
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:de0e4b7fb822101198817e83cd7e9f4f7406e75ec91d07a833a27a417cffe66b
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:c1c97b6beb759d3f6f2dd53c4258e4ccbb3bf35cf11873bd446b135defaa69d6
        version: 2.19.3
      entries:
      - name: rhods-operator.2.19.3
        version: 2.19.3
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.19
    - currentCSV: rhods-operator.2.22.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
          createdAt: "2025-09-25T11:28:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.22
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:13f3679ab289eea9de54fe129a7c83f6c8ec5d26b61b430637a7c4cfcd31a74a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:d59ed40de694bc0e0564bab159a5b586cffa0721b6d761fb2d81dd42e6225d1c
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:d7c4ac937add511460a36375d9b6b3c475a3a9a7cc96c00df2fbad6219aee0e9
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:b2ca229bb4c4dbf04b96766b8af8d9f55f2edef59ec4110de0d20f8c857e7d27
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:9104cbc2d595933f0a2acb7313e41ed69905e76a9fae6bd9820e7717b56a632b
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f9965177de24f3165ec4c001c14cf1e2c2c01ffe847839af837c442e3c96beae
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:2f3f305c97e50057781553416fb4026cde306ead0fbe0213627b9b5bbcc8a0c3
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:b2c2ea9d35edbcd16e1ae559c88a453129b877fc049ab7a7cd6c2ebb8f0bcb59
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:6234e15f001122ab222481321153fb0eba7f4d86bf3dfd64edf3f6cecbe7d62f
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:5d79f0bcca0a213b53aee6d88834e503d343821596b857b13d3a6362de8ebeab
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:a44d32a18b1e906c1411cafc82102881573fcd04e040a4e2ba6bd1bd617b70fa
        - registry.redhat.io/rhel9/mariadb-105@sha256:5e77b33ce04c1eafc2e8840586ad6a5b51148dea8a46dda151071e28858331f0
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:7fd540586a1ca111370100a473923b3dc6cec4255e9e4f3654bfbabd9158a2bb
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:2fe9473cd1d6c445d2ca7cc357ee4bb009aa1838f9cb69ea418296b2a044816b
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:27415b1c763cf89b79391f0b039936c3e078a078bda4420f13bf10ef9be31d5e
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:7b3d7787ba95156de9c284ad78f2d63a0d3bcb525f4d8d11ebf0203d4c5c1fe0
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:2065b96bd6f0a830eab66e36217a927448cfc5a7dabf2a5e7d907aa6c40657a4
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e3b90530c6e38f6940d7f9408e9fa2c81eb06a857ec2b0bab5d006b1e9a63a4f
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:e6fbe6723b4801f6854303b3fd7d4a4a36727bdd28dc28002b7de4929739ccc0
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/ubi9/toolbox@sha256:ea4ce8cb8f59cf1b0709f1606d7cab074c3806d4016964eedc0555410cb5dfeb
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:5914f71e37c815ce2567826896e051303bdef52fae1acf8d02e672fc24c7eb9d
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:1709fa3c79aad4ba7eb9be8299949396092c8e20210124e0c0936385bc04e839
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:cb5be3dd7098f4c8d66f9e85ad04656460df6102652deddc6cccb884a12be733
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:2b3aace0d80474470e71c2c8291b25738e096109229fe6d5d3e06e9885ee58b6
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:2f1a0a0803b2f5118b167e9d9c2e86660add5f79ee6f47044f63b0e8abe0b75f
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:ccba67a6a5aa3403e0018f29dcdfef483220dea8065d8b572e80c653ca992add
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:463e9be9c4409ba1e5a42713bd65844a023e7a3e9032d4e862edbaa5fe35300a
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:abe1f9dd11c7313fdd55b678be511a5cdc3546654a68e2b53d42c1013fec8d00
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:bd6d1599b4ff9eb890be9808014715b28ca19247222d65f40df024db4dd3525a
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:74d4e0174655cf6f6273857325a1e714a1531c5a06ad38bf139bee9c7d8f5f19
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:32c9511846077ef6035eb19be7653f6a78e0f207eb420ab93b18f742ab5779b6
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        version: 2.22.2
      entries:
      - name: rhods-operator.2.22.2
        version: 2.22.2
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.22
    - currentCSV: rhods-operator.2.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:b3563e201bf2720f60e73c0b04cb70566ea52eda7d09a28753f3a1487993c2e8
          createdAt: "2024-08-01T06:23:03Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.8.0 <2.8.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:eb4da3a645b53acfd6ce8758c043ae19b4126e62bcca09165114a141c2c8d014
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:92d9e4b53d580df6cc1f14a849b477749a1306941a2e8f9d1b6160f4911c3cff
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:b3563e201bf2720f60e73c0b04cb70566ea52eda7d09a28753f3a1487993c2e8
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:5f39b11f0cd6db2d7d6e51ee666c030e2643aee92a090b49656c1e7dfd8a5430
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:90833e85e0de3d7de870074c8b04319cd11b913b9652af6a5727aae41a91458c
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:91525fac5fadece75fbdafc40b94445fb92c88cfc8c2f06b596f8c6377e59c9e
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:ab8792fcd16b1aa0065444ba54335cb3e747379740c3ec2d56beb46ff80cb044
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:987c768c0688431a53e48a80591f2a41e074994169f9e691b28b89528d8e736f
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:1a6e29214e06a2665baadbed598ba758d6141bfa522984eb05f05a80558a403c
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:850b78a201effcafe33705dcd2670f1c470d8b46ba6e307032403627b55a7cfe
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:080377d42d782bf525b4072384d0cf8409b48019b3748ca7c855f340bcb94251
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:b6ecb3858c201498024e3bfc4a93e630699be4b1fdc3a65fd2d70e4b375eb9c2
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:df2edcaf4402b11ce24439263956e3b0cc43efbac2fdf152afacb1a398dcc59a
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:889f046b255b13d99801108588d3abf577611b828657b5240a18ea21f0e3cbbe
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:1c49888ba2e9449c04d5588040c099290fce4793c2ba8874c1037316e550d6e1
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:0b8665e8978deb73be2cc65830f08a74a5a48f3e3ee7b2b817f7bbacba257d22
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:68e1331f13dfaf6641a1e7d54f4cf29c0ef42663abc98b6dc69167892d2d8adb
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:06a6d5cbbe80d49bca9e4db9929571a69e35ae7951fcae84dcd6460f7ddf622c
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:9799003449bfb61ae8f59efd2a611b0da3bccadd9b20cc1519f6c9279d085337
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:9c90873fbd5af8fac8efc3c7c27fba742a06fad2d47f336319dd05d2b116867b
        version: 2.8.4
      entries:
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.8
    defaultChannel: stable
    packageName: rhods-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: ansible-automation-platform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aap-operator.v2.4.0-0.1758079821
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.1-0.1635279521 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the hub, including all hosted Ansible
              content, secrets and the database. By default, a persistent volume claim
              will be created using the default StorageClass on your cluster to store
              the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous hub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Automation Hub
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-24

          ## How to Upgrade

          Starting with AAP 2.4 in the May 2025 async release, the managed database was upgraded to PostgreSQL 15. It is recommended to take a backup of your deployment before upgrading.

          To take a backup, create AutomationControllerBackup and AutomationHubBackup objects for your deployments.
          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel.

          Note: Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8-operator@sha256:5b957031609c8aabf7575e675bc636c1152babc9564d503802301aff4642a404
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-runner-rhel8@sha256:e14286dc9605be3a8c6cb144d0b2e3f3bb3211129005009e7566ea6c16c91f52
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8@sha256:7ae8db2b5d12b619f40db1e581db1e982a695bc1a29a2628222fff09e7c25249
        - registry.redhat.io/ansible-automation-platform-24/ee-minimal-rhel8@sha256:7670e41037f023e3fc6ee95d2342787a80077273ab8d30ba6804029ff05b4edc
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-ui-rhel8@sha256:d229d4e09b2b788d1ba3413ea7ec15374bd1613d507c285ac1ddc10570b9473f
        - registry.redhat.io/ansible-automation-platform-24/hub-web-rhel8@sha256:6ff1f602f0d9cad201b3addb65acc570a9d527de005ca85fc861d3cb98138710
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8@sha256:bfd13d6b39c5ec3a67adbafb05a0482328f0b9b97a517b45c093f6c35a3a92d3
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8@sha256:04fb5cb6b0d7c79cf5d4b754fe917224ddfcb2b9a4994462db404c833ed69483
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-rhel8-operator@sha256:b120e6a5ef376662b4324f0458eae31627087dc91fc613788143bc3954a15318
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-24/ee-supported-rhel8@sha256:f1384eb558f7da1e4a35d51408b3cf5a5a9be08f4d50503f473b26909bc246fc
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8@sha256:cccb4db08735551c697a7b2cbf75812f34500e691f9cdbaea02b30edea292e47
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8-operator@sha256:5253c9fd587ff2ca94d251f2b5fc7f284080adc5eb64643df5299beb2015592f
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8-operator@sha256:3c28a1326ecb7495952a5bb8be13b0972a4378af82611b0b9d990c5fa9fa82c3
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8-operator@sha256:e20106a8a04abd568a4c56861cea39ae444c0bd204bc807ac01f1705e12937d3
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        version: 2.4.0+0.1758079821
      entries:
      - name: aap-operator.v2.4.0-0.1702352195
        version: 2.4.0+0.1702352195
      - name: aap-operator.v2.4.0-0.1707506910
        version: 2.4.0+0.1707506910
      - name: aap-operator.v2.4.0-0.1708974895
        version: 2.4.0+0.1708974895
      - name: aap-operator.v2.4.0-0.1711590540
        version: 2.4.0+0.1711590540
      - name: aap-operator.v2.4.0-0.1713466670
        version: 2.4.0+0.1713466670
      - name: aap-operator.v2.4.0-0.1715898994
        version: 2.4.0+0.1715898994
      - name: aap-operator.v2.4.0-0.1717558474
        version: 2.4.0+0.1717558474
      - name: aap-operator.v2.4.0-0.1718152210
        version: 2.4.0+0.1718152210
      - name: aap-operator.v2.4.0-0.1720482179
        version: 2.4.0+0.1720482179
      - name: aap-operator.v2.4.0-0.1725256739
        version: 2.4.0+0.1725256739
      - name: aap-operator.v2.4.0-0.1726111413
        version: 2.4.0+0.1726111413
      - name: aap-operator.v2.4.0-0.1727257740
        version: 2.4.0+0.1727257740
      - name: aap-operator.v2.4.0-0.1730152858
        version: 2.4.0+0.1730152858
      - name: aap-operator.v2.4.0-0.1733185647
        version: 2.4.0+0.1733185647
      - name: aap-operator.v2.4.0-0.1733943951
        version: 2.4.0+0.1733943951
      - name: aap-operator.v2.4.0-0.1737661496
        version: 2.4.0+0.1737661496
      - name: aap-operator.v2.4.0-0.1741194076
        version: 2.4.0+0.1741194076
      - name: aap-operator.v2.4.0-0.1742444794
        version: 2.4.0+0.1742444794
      - name: aap-operator.v2.4.0-0.1743629049
        version: 2.4.0+0.1743629049
      - name: aap-operator.v2.4.0-0.1746132523
        version: 2.4.0+0.1746132523
      - name: aap-operator.v2.4.0-0.1749069319
        version: 2.4.0+0.1749069319
      - name: aap-operator.v2.4.0-0.1750877268
        version: 2.4.0+0.1750877268
      - name: aap-operator.v2.4.0-0.1753232084
        version: 2.4.0+0.1753232084
      - name: aap-operator.v2.4.0-0.1755833431
        version: 2.4.0+0.1755833431
      - name: aap-operator.v2.4.0-0.1758079821
        version: 2.4.0+0.1758079821
      name: stable-2.4
    - currentCSV: aap-operator.v2.4.0-0.1758080293
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.1-0.1635283332 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the hub, including all hosted Ansible
              content, secrets and the database. By default, a persistent volume claim
              will be created using the default StorageClass on your cluster to store
              the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous hub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Automation Hub
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-24

          ## How to Upgrade

          Starting with AAP 2.4 in the May 2025 async release, the managed database was upgraded to PostgreSQL 15. It is recommended to take a backup of your deployment before upgrading.

          To take a backup, create AutomationControllerBackup and AutomationHubBackup objects for your deployments.
          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel.

          Note: Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8@sha256:bfd13d6b39c5ec3a67adbafb05a0482328f0b9b97a517b45c093f6c35a3a92d3
        - registry.redhat.io/ansible-automation-platform-24/ee-minimal-rhel8@sha256:7670e41037f023e3fc6ee95d2342787a80077273ab8d30ba6804029ff05b4edc
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8@sha256:04fb5cb6b0d7c79cf5d4b754fe917224ddfcb2b9a4994462db404c833ed69483
        - registry.redhat.io/ansible-automation-platform-24/ee-supported-rhel8@sha256:f1384eb558f7da1e4a35d51408b3cf5a5a9be08f4d50503f473b26909bc246fc
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-rhel8-operator@sha256:b120e6a5ef376662b4324f0458eae31627087dc91fc613788143bc3954a15318
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8-operator@sha256:5253c9fd587ff2ca94d251f2b5fc7f284080adc5eb64643df5299beb2015592f
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8-operator@sha256:3c28a1326ecb7495952a5bb8be13b0972a4378af82611b0b9d990c5fa9fa82c3
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8@sha256:7ae8db2b5d12b619f40db1e581db1e982a695bc1a29a2628222fff09e7c25249
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8@sha256:cccb4db08735551c697a7b2cbf75812f34500e691f9cdbaea02b30edea292e47
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8-operator@sha256:5b957031609c8aabf7575e675bc636c1152babc9564d503802301aff4642a404
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8-operator@sha256:e20106a8a04abd568a4c56861cea39ae444c0bd204bc807ac01f1705e12937d3
        - registry.redhat.io/ansible-automation-platform-24/hub-web-rhel8@sha256:6ff1f602f0d9cad201b3addb65acc570a9d527de005ca85fc861d3cb98138710
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-runner-rhel8@sha256:e14286dc9605be3a8c6cb144d0b2e3f3bb3211129005009e7566ea6c16c91f52
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-ui-rhel8@sha256:d229d4e09b2b788d1ba3413ea7ec15374bd1613d507c285ac1ddc10570b9473f
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        version: 2.4.0+0.1758080293
      entries:
      - name: aap-operator.v2.4.0-0.1702352614
        version: 2.4.0+0.1702352614
      - name: aap-operator.v2.4.0-0.1707507304
        version: 2.4.0+0.1707507304
      - name: aap-operator.v2.4.0-0.1708975279
        version: 2.4.0+0.1708975279
      - name: aap-operator.v2.4.0-0.1711591330
        version: 2.4.0+0.1711591330
      - name: aap-operator.v2.4.0-0.1713467078
        version: 2.4.0+0.1713467078
      - name: aap-operator.v2.4.0-0.1715899379
        version: 2.4.0+0.1715899379
      - name: aap-operator.v2.4.0-0.1717558885
        version: 2.4.0+0.1717558885
      - name: aap-operator.v2.4.0-0.1718152680
        version: 2.4.0+0.1718152680
      - name: aap-operator.v2.4.0-0.1720482740
        version: 2.4.0+0.1720482740
      - name: aap-operator.v2.4.0-0.1725257213
        version: 2.4.0+0.1725257213
      - name: aap-operator.v2.4.0-0.1726111804
        version: 2.4.0+0.1726111804
      - name: aap-operator.v2.4.0-0.1727259094
        version: 2.4.0+0.1727259094
      - name: aap-operator.v2.4.0-0.1730153340
        version: 2.4.0+0.1730153340
      - name: aap-operator.v2.4.0-0.1733186325
        version: 2.4.0+0.1733186325
      - name: aap-operator.v2.4.0-0.1733945743
        version: 2.4.0+0.1733945743
      - name: aap-operator.v2.4.0-0.1737662040
        version: 2.4.0+0.1737662040
      - name: aap-operator.v2.4.0-0.1741194557
        version: 2.4.0+0.1741194557
      - name: aap-operator.v2.4.0-0.1742445409
        version: 2.4.0+0.1742445409
      - name: aap-operator.v2.4.0-0.1743630487
        version: 2.4.0+0.1743630487
      - name: aap-operator.v2.4.0-0.1746133932
        version: 2.4.0+0.1746133932
      - name: aap-operator.v2.4.0-0.1749069693
        version: 2.4.0+0.1749069693
      - name: aap-operator.v2.4.0-0.1750878254
        version: 2.4.0+0.1750878254
      - name: aap-operator.v2.4.0-0.1753232791
        version: 2.4.0+0.1753232791
      - name: aap-operator.v2.4.0-0.1755833968
        version: 2.4.0+0.1755833968
      - name: aap-operator.v2.4.0-0.1758080293
        version: 2.4.0+0.1758080293
      name: stable-2.4-cluster-scoped
    - currentCSV: aap-operator.v2.5.0-0.1758147230
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727257740 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation gateway - Unify the management of AAP from a central location.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner


          ## New in AAP 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-25

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.5

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console. For example, change the channel from `stable-2.4` to `stable-2.5`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.

          After changing to the new 2.5 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place.

          #### Create the AnsibleAutomationPlatform custom resource

          Once that is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: myaap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8@sha256:3a989f148648a5a62e687058ec3ca772fa22aa10faec89453626d2b17b3fd9c2
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8-operator@sha256:43e731b9a435b3a412108170b4b39e6b6215be51d362a9f03c33fa412e97da53
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8-operator@sha256:8d8182e3a87c55eacc222e71df4a9bf7a2e51102d913a5824cc91e36056d4249
        - registry.redhat.io/ansible-automation-platform-25/hub-web-rhel8@sha256:e96832efcc03c6754031cb70f56c808135f62d1b80a882dd6d6fc9e1a4cfc69b
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8@sha256:697343aaea153ab22d7efa8f1bb5aed37a55cbb353fc206f4a355e8b878ae418
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8@sha256:ac0de2b7d354eb0c15a6c12861a4de3577a982c84af661fbab25517d8105c085
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8@sha256:aed3973eb953dda3571260dcf1b33bc7dcbc18c41e3350b8c361b7e635fb9527
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-runner-rhel8@sha256:ace1d2ab49ba5ff9f11ab18e8edce27c9e73ae7298bf7335569fa1740a44899b
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8-operator@sha256:acc0fcd509eeac46db307ab744dd0a0c26770a93e495cf2ec0ccb61a91c6fbc5
        - registry.redhat.io/ansible-automation-platform-25/ee-minimal-rhel8@sha256:41f944eb1450db56c3b14fc7baece8f148d5f10886a57bad730bec60fb28ac04
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8-operator@sha256:07aa7e257d04ae0eda6fcaa4af9121f1aece439231fce4f924e6467da91f0022
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-ui-rhel8@sha256:9ab35e80a6172018ed3f5cd179e0bcb6befc1378583d9e73c3a64a7cf9e8050f
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8@sha256:9f2867603b9747c2fe37862c2c5efac4a4530902f323a8bf7cc0717de4cdb5aa
        - registry.redhat.io/ansible-automation-platform-25/gateway-proxy-rhel8@sha256:789d55f355fa0d90f9509a912d199123bfc5a4041930a72a8fd4ed1d819f0d84
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8-operator@sha256:503fc7a20e2dd7de2cb55b9be5f97efa0d2b31f054ce326f23f3a11b4d683e64
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-rhel8-operator@sha256:74be9182723b41f87b3d13b3d86b6a9049bba88e2902cc9453264aa8a0f584e4
        - registry.redhat.io/ansible-automation-platform-25/ee-supported-rhel8@sha256:52d2c73fa2015af5f1b9a88ee48085c7c7bab44330fc9f95914f48b58e4b48bc
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-chatbot-rhel8@sha256:9a37c9739c488ccf211ec5efc6debb045223d508c9cb3fa5a8256eca98ff2f0c
        version: 2.5.0+0.1758147230
      entries:
      - name: aap-operator.v2.5.0-0.1727240614
        version: 2.5.0+0.1727240614
      - name: aap-operator.v2.5.0-0.1727875185
        version: 2.5.0+0.1727875185
      - name: aap-operator.v2.5.0-0.1728520175
        version: 2.5.0+0.1728520175
      - name: aap-operator.v2.5.0-0.1729741547
        version: 2.5.0+0.1729741547
      - name: aap-operator.v2.5.0-0.1731456196
        version: 2.5.0+0.1731456196
      - name: aap-operator.v2.5.0-0.1733193164
        version: 2.5.0+0.1733193164
      - name: aap-operator.v2.5.0-0.1733969526
        version: 2.5.0+0.1733969526
      - name: aap-operator.v2.5.0-0.1736893446
        version: 2.5.0+0.1736893446
      - name: aap-operator.v2.5.0-0.1737101025
        version: 2.5.0+0.1737101025
      - name: aap-operator.v2.5.0-0.1737675968
        version: 2.5.0+0.1737675968
      - name: aap-operator.v2.5.0-0.1739248042
        version: 2.5.0+0.1739248042
      - name: aap-operator.v2.5.0-0.1740093573
        version: 2.5.0+0.1740093573
      - name: aap-operator.v2.5.0-0.1740773472
        version: 2.5.0+0.1740773472
      - name: aap-operator.v2.5.0-0.1741368605
        version: 2.5.0+0.1741368605
      - name: aap-operator.v2.5.0-0.1742434024
        version: 2.5.0+0.1742434024
      - name: aap-operator.v2.5.0-0.1743660124
        version: 2.5.0+0.1743660124
      - name: aap-operator.v2.5.0-0.1746137767
        version: 2.5.0+0.1746137767
      - name: aap-operator.v2.5.0-0.1747343762
        version: 2.5.0+0.1747343762
      - name: aap-operator.v2.5.0-0.1749074128
        version: 2.5.0+0.1749074128
      - name: aap-operator.v2.5.0-0.1749604727
        version: 2.5.0+0.1749604727
      - name: aap-operator.v2.5.0-0.1750901111
        version: 2.5.0+0.1750901111
      - name: aap-operator.v2.5.0-0.1753900156
        version: 2.5.0+0.1753900156
      - name: aap-operator.v2.5.0-0.1755835086
        version: 2.5.0+0.1755835086
      - name: aap-operator.v2.5.0-0.1758147230
        version: 2.5.0+0.1758147230
      name: stable-2.5
    - currentCSV: aap-operator.v2.5.0-0.1758147817
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727259094 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation gateway - Unify the management of AAP from a central location.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner


          ## New in AAP 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-25

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.5

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console. For example, change the channel from `stable-2.4` to `stable-2.5`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.

          After changing to the new 2.5 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place.

          #### Create the AnsibleAutomationPlatform custom resource

          Once that is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: myaap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8-operator@sha256:acc0fcd509eeac46db307ab744dd0a0c26770a93e495cf2ec0ccb61a91c6fbc5
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-ui-rhel8@sha256:9ab35e80a6172018ed3f5cd179e0bcb6befc1378583d9e73c3a64a7cf9e8050f
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8-operator@sha256:43e731b9a435b3a412108170b4b39e6b6215be51d362a9f03c33fa412e97da53
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8-operator@sha256:8d8182e3a87c55eacc222e71df4a9bf7a2e51102d913a5824cc91e36056d4249
        - registry.redhat.io/ansible-automation-platform-25/hub-web-rhel8@sha256:e96832efcc03c6754031cb70f56c808135f62d1b80a882dd6d6fc9e1a4cfc69b
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-chatbot-rhel8@sha256:9a37c9739c488ccf211ec5efc6debb045223d508c9cb3fa5a8256eca98ff2f0c
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8@sha256:3a989f148648a5a62e687058ec3ca772fa22aa10faec89453626d2b17b3fd9c2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/ee-minimal-rhel8@sha256:41f944eb1450db56c3b14fc7baece8f148d5f10886a57bad730bec60fb28ac04
        - registry.redhat.io/ansible-automation-platform-25/gateway-proxy-rhel8@sha256:789d55f355fa0d90f9509a912d199123bfc5a4041930a72a8fd4ed1d819f0d84
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8@sha256:aed3973eb953dda3571260dcf1b33bc7dcbc18c41e3350b8c361b7e635fb9527
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8@sha256:697343aaea153ab22d7efa8f1bb5aed37a55cbb353fc206f4a355e8b878ae418
        - registry.redhat.io/ansible-automation-platform-25/ee-supported-rhel8@sha256:52d2c73fa2015af5f1b9a88ee48085c7c7bab44330fc9f95914f48b58e4b48bc
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8@sha256:ac0de2b7d354eb0c15a6c12861a4de3577a982c84af661fbab25517d8105c085
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-rhel8-operator@sha256:74be9182723b41f87b3d13b3d86b6a9049bba88e2902cc9453264aa8a0f584e4
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-runner-rhel8@sha256:ace1d2ab49ba5ff9f11ab18e8edce27c9e73ae7298bf7335569fa1740a44899b
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8@sha256:9f2867603b9747c2fe37862c2c5efac4a4530902f323a8bf7cc0717de4cdb5aa
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8-operator@sha256:07aa7e257d04ae0eda6fcaa4af9121f1aece439231fce4f924e6467da91f0022
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8-operator@sha256:503fc7a20e2dd7de2cb55b9be5f97efa0d2b31f054ce326f23f3a11b4d683e64
        version: 2.5.0+0.1758147817
      entries:
      - name: aap-operator.v2.5.0-0.1727241223
        version: 2.5.0+0.1727241223
      - name: aap-operator.v2.5.0-0.1727875896
        version: 2.5.0+0.1727875896
      - name: aap-operator.v2.5.0-0.1728520687
        version: 2.5.0+0.1728520687
      - name: aap-operator.v2.5.0-0.1729742145
        version: 2.5.0+0.1729742145
      - name: aap-operator.v2.5.0-0.1731456854
        version: 2.5.0+0.1731456854
      - name: aap-operator.v2.5.0-0.1733193761
        version: 2.5.0+0.1733193761
      - name: aap-operator.v2.5.0-0.1733970210
        version: 2.5.0+0.1733970210
      - name: aap-operator.v2.5.0-0.1736894220
        version: 2.5.0+0.1736894220
      - name: aap-operator.v2.5.0-0.1737101781
        version: 2.5.0+0.1737101781
      - name: aap-operator.v2.5.0-0.1737676792
        version: 2.5.0+0.1737676792
      - name: aap-operator.v2.5.0-0.1739248649
        version: 2.5.0+0.1739248649
      - name: aap-operator.v2.5.0-0.1740094176
        version: 2.5.0+0.1740094176
      - name: aap-operator.v2.5.0-0.1740774104
        version: 2.5.0+0.1740774104
      - name: aap-operator.v2.5.0-0.1741369251
        version: 2.5.0+0.1741369251
      - name: aap-operator.v2.5.0-0.1742434756
        version: 2.5.0+0.1742434756
      - name: aap-operator.v2.5.0-0.1743660958
        version: 2.5.0+0.1743660958
      - name: aap-operator.v2.5.0-0.1746138413
        version: 2.5.0+0.1746138413
      - name: aap-operator.v2.5.0-0.1747345055
        version: 2.5.0+0.1747345055
      - name: aap-operator.v2.5.0-0.1749074612
        version: 2.5.0+0.1749074612
      - name: aap-operator.v2.5.0-0.1749607543
        version: 2.5.0+0.1749607543
      - name: aap-operator.v2.5.0-0.1750901870
        version: 2.5.0+0.1750901870
      - name: aap-operator.v2.5.0-0.1753900674
        version: 2.5.0+0.1753900674
      - name: aap-operator.v2.5.0-0.1755835623
        version: 2.5.0+0.1755835623
      - name: aap-operator.v2.5.0-0.1758147817
        version: 2.5.0+0.1758147817
      name: stable-2.5-cluster-scoped
    - currentCSV: aap-operator.v2.6.0-0.1758684783
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727257740 <2.6.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Automation Gateway - Unify the management of AAP from a central location.
            * Automation Controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private Automation Hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Ansible Lightspeed: Deploy and manage the lifecycle of an Ansible Lightspeed application.
            * Platform Resource Operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## New in AAP 2.6 and 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.


          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-26

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5 and above, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.6

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console.

          **For customers on AAP 2.4:** Change the channel directly from `stable-2.4` to `stable-2.6` to upgrade directly to AAP 2.6, which will be released in October.

          **For customers on AAP 2.5:** Change the channel from `stable-2.5` to `stable-2.6`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.

          After changing to the new 2.6 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place. When upgrading from AAP 2.4 to 2.6, your external authentication configuration in Automation Controller will automatically be migrated to the new unified Platform UI.

          **Database:** AAP 2.6 continues to use PostgreSQL 15.

          #### Create the AnsibleAutomationPlatform custom resource (Required for upgrades from 2.4 only)

          **If upgrading from AAP 2.4:** Once the operator upgrade is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created.

          **If upgrading from AAP 2.5:** This step is not required as your AnsibleAutomationPlatform custom resource already exists.

          Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          Next, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda (lightspeed is disabled by default):


          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: aap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-26/gateway-rhel9-operator@sha256:fa1b102a0690767a1e08935a0e164c1eabb2d4cf9afccd310262ee9185544de7
        - registry.redhat.io/ansible-automation-platform-26/hub-rhel9@sha256:8041b141cd8921aaaa20cce15eab4c6189da9f449d99057189284891c071ac3b
        - registry.redhat.io/ansible-automation-platform-26/gateway-proxy-rhel9@sha256:13284d20c6fb4f2de24ec1715e7abdb5c4b0e4298c5034a7795f494b93634a02
        - registry.redhat.io/rhel9/postgresql-15@sha256:5e225c1646251d2509b3a6e45fad23abfb11dd5c45d3224b59178f90deb38a70
        - registry.redhat.io/ansible-automation-platform-26/controller-rhel9-operator@sha256:863e77e52ab01cec2a71672a224080a59d963d4fc08b9ed6f3cc7950729b1746
        - registry.redhat.io/ansible-automation-platform-26/hub-rhel9-operator@sha256:18114aab6db0d00780fdeaf9efa74ee44cab94824c2fd35f4b9998b6378ad8b3
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-rhel9-operator@sha256:cfbb98e17297134c187673f355342ac5ad60a56d73f4b0c918be2ff244286cc0
        - registry.redhat.io/ansible-automation-platform-26/platform-resource-rhel9-operator@sha256:969bbd033fbba33624329ee9347b3b89fb5a1f64e71892f5c6c7d74ed14a6d39
        - registry.redhat.io/ansible-automation-platform-26/controller-rhel9@sha256:64f5e77a7842965e9c6f7bc7990354b71bb19ffb06ab8ef7793dde956cb6c0f0
        - registry.redhat.io/ansible-automation-platform-26/ee-minimal-rhel9@sha256:261c324cac081e47d559c5fccbd48e7ec8f5bf64e3f27287fbe4fb8aff6b7d99
        - registry.redhat.io/ansible-automation-platform-26/hub-web-rhel9@sha256:aa4041672542a7f828d8a51c79e9c4dcea2a185388e4f20f828104db37e03a9b
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-rhel9@sha256:01e56bee772d15b120d6e4b02a59024b21eb8b266127b1a9b7f2c535ecdb2855
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-ui-rhel9@sha256:5b5f38a5f76781a042eec9cbf70b8be026878e4fa61c7fb90d1b0ad64bd21592
        - registry.redhat.io/ubi9/ubi-minimal@sha256:7c5495d5fad59aaee12abc3cbbd2b283818ee1e814b00dbc7f25bf2d14fa4f0c
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-chatbot-rhel9@sha256:8901fba14b21569b02108d2ab9050d558dee29f8177fae9d1bfcd5de3f4ce7c1
        - registry.redhat.io/ansible-automation-platform-26/mcp-tools-rhel9@sha256:0c1f6f6958f2570c8cd073287555d28c70ef9dded702ec014d3ead8ce572463e
        - registry.redhat.io/ansible-automation-platform-26/platform-resource-runner-rhel9@sha256:9983043386df38a17ff6fffef64dc187691e0b1fefec01eef3af39aa08e532ef
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-rhel9-operator@sha256:bfd55bffd8239e224d4ebba500d8517e8e66915078d6b59244f5e910fc5e10c8
        - registry.redhat.io/ansible-automation-platform-26/gateway-rhel9@sha256:c85dbbf6beaa264afec480dc1f06576aada531cdfd15f247f4db37dc9702be33
        - registry.redhat.io/rhel9/redis-6@sha256:0962d18bb5292c0e524354f319b63287c2298a4a2fd34556c68fa285df381427
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-rhel9@sha256:c5d4e2e76944e0991729e5c3b99f43b7f29ea08ad112e5e6513feb8fb1d899a1
        - registry.redhat.io/ansible-automation-platform-26/ee-supported-rhel9@sha256:2534d07f1ae41af32adb681e283cbaea465cbe0f9d234ed5478a7bfad7676e7f
        version: 2.6.0+0.1758684783
      entries:
      - name: aap-operator.v2.6.0-0.1758684783
        version: 2.6.0+0.1758684783
      name: stable-2.6
    - currentCSV: aap-operator.v2.6.0-0.1758685241
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727259094 <2.6.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Automation Gateway - Unify the management of AAP from a central location.
            * Automation Controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private Automation Hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Ansible Lightspeed: Deploy and manage the lifecycle of an Ansible Lightspeed application.
            * Platform Resource Operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## New in AAP 2.6 and 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.


          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-26

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5 and above, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.6

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console.

          **For customers on AAP 2.4:** Change the channel directly from `stable-2.4` to `stable-2.6` to upgrade directly to AAP 2.6, which will be released in October.

          **For customers on AAP 2.5:** Change the channel from `stable-2.5` to `stable-2.6`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.

          After changing to the new 2.6 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place. When upgrading from AAP 2.4 to 2.6, your external authentication configuration in Automation Controller will automatically be migrated to the new unified Platform UI.

          **Database:** AAP 2.6 continues to use PostgreSQL 15.

          #### Create the AnsibleAutomationPlatform custom resource (Required for upgrades from 2.4 only)

          **If upgrading from AAP 2.4:** Once the operator upgrade is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created.

          **If upgrading from AAP 2.5:** This step is not required as your AnsibleAutomationPlatform custom resource already exists.

          Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          Next, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda (lightspeed is disabled by default):


          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: aap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-26/gateway-rhel9@sha256:c85dbbf6beaa264afec480dc1f06576aada531cdfd15f247f4db37dc9702be33
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-ui-rhel9@sha256:5b5f38a5f76781a042eec9cbf70b8be026878e4fa61c7fb90d1b0ad64bd21592
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-rhel9-operator@sha256:cfbb98e17297134c187673f355342ac5ad60a56d73f4b0c918be2ff244286cc0
        - registry.redhat.io/ansible-automation-platform-26/controller-rhel9@sha256:64f5e77a7842965e9c6f7bc7990354b71bb19ffb06ab8ef7793dde956cb6c0f0
        - registry.redhat.io/ansible-automation-platform-26/ee-supported-rhel9@sha256:2534d07f1ae41af32adb681e283cbaea465cbe0f9d234ed5478a7bfad7676e7f
        - registry.redhat.io/ansible-automation-platform-26/ee-minimal-rhel9@sha256:261c324cac081e47d559c5fccbd48e7ec8f5bf64e3f27287fbe4fb8aff6b7d99
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-rhel9@sha256:c5d4e2e76944e0991729e5c3b99f43b7f29ea08ad112e5e6513feb8fb1d899a1
        - registry.redhat.io/ansible-automation-platform-26/hub-web-rhel9@sha256:aa4041672542a7f828d8a51c79e9c4dcea2a185388e4f20f828104db37e03a9b
        - registry.redhat.io/ansible-automation-platform-26/platform-resource-runner-rhel9@sha256:9983043386df38a17ff6fffef64dc187691e0b1fefec01eef3af39aa08e532ef
        - registry.redhat.io/ansible-automation-platform-26/gateway-rhel9-operator@sha256:fa1b102a0690767a1e08935a0e164c1eabb2d4cf9afccd310262ee9185544de7
        - registry.redhat.io/ansible-automation-platform-26/platform-resource-rhel9-operator@sha256:969bbd033fbba33624329ee9347b3b89fb5a1f64e71892f5c6c7d74ed14a6d39
        - registry.redhat.io/ansible-automation-platform-26/lightspeed-chatbot-rhel9@sha256:8901fba14b21569b02108d2ab9050d558dee29f8177fae9d1bfcd5de3f4ce7c1
        - registry.redhat.io/ansible-automation-platform-26/mcp-tools-rhel9@sha256:0c1f6f6958f2570c8cd073287555d28c70ef9dded702ec014d3ead8ce572463e
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-rhel9-operator@sha256:bfd55bffd8239e224d4ebba500d8517e8e66915078d6b59244f5e910fc5e10c8
        - registry.redhat.io/ansible-automation-platform-26/hub-rhel9-operator@sha256:18114aab6db0d00780fdeaf9efa74ee44cab94824c2fd35f4b9998b6378ad8b3
        - registry.redhat.io/rhel9/postgresql-15@sha256:5e225c1646251d2509b3a6e45fad23abfb11dd5c45d3224b59178f90deb38a70
        - registry.redhat.io/ubi9/ubi-minimal@sha256:7c5495d5fad59aaee12abc3cbbd2b283818ee1e814b00dbc7f25bf2d14fa4f0c
        - registry.redhat.io/ansible-automation-platform-26/eda-controller-rhel9@sha256:01e56bee772d15b120d6e4b02a59024b21eb8b266127b1a9b7f2c535ecdb2855
        - registry.redhat.io/ansible-automation-platform-26/gateway-proxy-rhel9@sha256:13284d20c6fb4f2de24ec1715e7abdb5c4b0e4298c5034a7795f494b93634a02
        - registry.redhat.io/ansible-automation-platform-26/hub-rhel9@sha256:8041b141cd8921aaaa20cce15eab4c6189da9f449d99057189284891c071ac3b
        - registry.redhat.io/rhel9/redis-6@sha256:0962d18bb5292c0e524354f319b63287c2298a4a2fd34556c68fa285df381427
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/ansible-automation-platform-26/controller-rhel9-operator@sha256:863e77e52ab01cec2a71672a224080a59d963d4fc08b9ed6f3cc7950729b1746
        version: 2.6.0+0.1758685241
      entries:
      - name: aap-operator.v2.6.0-0.1758685241
        version: 2.6.0+0.1758685241
      name: stable-2.6-cluster-scoped
    defaultChannel: stable-2.6-cluster-scoped
    packageName: ansible-automation-platform-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/component: kmm-hub
      app.kubernetes.io/name: kmm-hub
      app.kubernetes.io/part-of: kmm
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: kernel-module-management-hub
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kernel-module-management-hub.v2.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:888a3acad339cf38a324c718e0cbe1237ca5bf2f4636b01d001752658e4cca1b
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.0.2'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:888a3acad339cf38a324c718e0cbe1237ca5bf2f4636b01d001752658e4cca1b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62c44057a27abd41cf3daf60b8736da59a82dadb65fa6c32bef65d19bd49ea49
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:192e08b121a04ae69aea60c35199e8f5e3c42305e1ebd024a605a371de4ab472
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:94b4ddc45cb34d4ffe794a51b99c868cb709f90cc4eb8b28284d70c526a0a226
        version: 2.0.2
      entries:
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.0
    - currentCSV: kernel-module-management-hub.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:d1c1490d4c20dc2ec31effa1af067676c4d264e3b7d4670fc11c2baf8a400001
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.1.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:14e100e5853444114029116255ebfb7a54e5ed51503ba969740ad48b5ca80f70
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:1a9a19df17b464d1708b2905d03b8726e4480ca1f11aa97a15ff4c049a24ea4b
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:d1c1490d4c20dc2ec31effa1af067676c4d264e3b7d4670fc11c2baf8a400001
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:611556d999ea97f9dbb78fec5963baa02de7a26626d7096ba63b177e3bcf9707
        version: 2.1.1
      entries:
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.1
    - currentCSV: kernel-module-management-hub.v2.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:90c1946b3f577cde0435402b04baa52dc54e5aab3bb84bd834ac026a57453296
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.2.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:b1d98c0f14448ae8eb80821cbec236150c338674aae2a2b37f8d0ab0d5f1430c
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:53fddf0b226eb579c71c86112a8d9b6b1d90a1498d6f6985932b16b8be532f35
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:c3dc9ad4ebc68cd85a074b520903a9310e72a90a73bdf4aa7c42b7ed66b7bcba
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:90c1946b3f577cde0435402b04baa52dc54e5aab3bb84bd834ac026a57453296
        version: 2.2.1
      entries:
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.2
    - currentCSV: kernel-module-management-hub.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:f1d8f3c95f621c3faf41867c5dd4bc13fa2b9b291422d6184e6e42f1d1460259
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.3.0'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:f1d8f3c95f621c3faf41867c5dd4bc13fa2b9b291422d6184e6e42f1d1460259
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:c5005b74402f680d11c9ea682a9a0d3f74b6ab0d01d25298d232f910f1a29c34
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:f9fde334af68220943017a7491130e6a82f15e63522f5bb3931aa7d9b00ccc40
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:06ac089cc89b303e525fd71d6439b958f34ab944a72209f1cc8cb562ecfd1112
        version: 2.3.0
      entries:
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.3
    - currentCSV: kernel-module-management-hub.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm-hub",
                    "app.kubernetes.io/name": "kmm-hub",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:2efb81137ad2981b994c62d460c917749308133a7dcad5b9a2b251e18f3052ec
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        version: 2.4.1
      entries:
      - name: kernel-module-management-hub.v2.4.1
        version: 2.4.1
      - name: kernel-module-management-hub.v2.4.0
        version: 2.4.0
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.4
    - currentCSV: kernel-module-management-hub.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm-hub",
                    "app.kubernetes.io/name": "kmm-hub",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:2efb81137ad2981b994c62d460c917749308133a7dcad5b9a2b251e18f3052ec
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        version: 2.4.1
      entries:
      - name: kernel-module-management-hub.v2.4.1
        version: 2.4.1
      - name: kernel-module-management-hub.v2.4.0
        version: 2.4.0
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: kernel-module-management-hub
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: cert-utils-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-utils-operator.v1.3.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/cert-utils-operator@sha256:0290e7b2800a5aa8eb57b0268908f2a20a93b16403c76df4c61021048161d6ff
          createdAt: "2024-01-23T12:29:04Z"
          description: Set of utilities for TLS certificates
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-utils-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/cert-utils-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Cert utils operator is a set of functionalities around certificates packaged in a [Kubernetes operator](https://github.com/operator-framework/operator-sdk).

          Certificates are assumed to be available in a [secret](https://kubernetes.io/docs/concepts/configuration/secret/) of type `kubernetes.io/tls` (other types of secrets are *ignored* by this operator).
          By convention this type of secrets have three optional entries:

          1. `tls.key`: the private key of the certificate.
          2. `tls.crt`: the actual certificate.
          3. `ca.crt`: the CA bundle that validates the certificate.

          The functionalities are the following:

          1. [Ability to populate route certificates](https://github.com/redhat-cop/cert-utils-operator#Populating-route-certificates)
          2. [Ability to create java keystore and truststore from the certificates](https://github.com/redhat-cop/cert-utils-operator#Creating-java-keystore-and-truststore)
          3. [Ability to show info regarding the certificates](https://github.com/redhat-cop/cert-utils-operator#Showing-info-on-the-certificates)
          4. [Ability to alert when a certificate is about to expire](https://github.com/redhat-cop/cert-utils-operator#Alerting-when-a-certificate-is-about-to-expire)
          5. [Ability to inject ca bundles in ValidatingWebhookConfiguration, MutatingWebhookConfiguration, CustomResourceDefinition object](https://github.com/redhat-cop/cert-utils-operator#CA-injection)

          All these feature are activated via opt-in annotations.
        displayName: Cert Utils Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Cert Utils Operator
          url: https://github.com/redhat-cop/cert-utils-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/cert-utils-operator
        maintainers:
        - email: rspazzoli@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/cert-utils-operator@sha256:0290e7b2800a5aa8eb57b0268908f2a20a93b16403c76df4c61021048161d6ff
        version: 1.3.12
      entries:
      - name: cert-utils-operator.v1.3.12
        version: 1.3.12
      - name: cert-utils-operator.v1.3.11
        version: 1.3.11
      - name: cert-utils-operator.v1.3.10
        version: 1.3.10
      - name: cert-utils-operator.v1.3.9
        version: 1.3.9
      - name: cert-utils-operator.v1.3.8
        version: 1.3.8
      - name: cert-utils-operator.v1.3.7
        version: 1.3.7
      - name: cert-utils-operator.v1.3.6
        version: 1.3.6
      - name: cert-utils-operator.v1.3.5
        version: 1.3.5
      - name: cert-utils-operator.v1.3.4
        version: 1.3.4
      - name: cert-utils-operator.v1.3.3
        version: 1.3.3
      - name: cert-utils-operator.v1.3.2
        version: 1.3.2
      - name: cert-utils-operator.v1.3.1
        version: 1.3.1
      - name: cert-utils-operator.v1.3.0
        version: 1.3.0
      - name: cert-utils-operator.v1.2.0
        version: 1.2.0
      - name: cert-utils-operator.v1.1.0
        version: 1.1.0
      - name: cert-utils-operator.v1.0.7
        version: 1.0.7
      - name: cert-utils-operator.v1.0.6
        version: 1.0.6
      - name: cert-utils-operator.v1.0.5
        version: 1.0.5
      - name: cert-utils-operator.v1.0.4
        version: 1.0.4
      - name: cert-utils-operator.v1.0.3
        version: 1.0.3
      - name: cert-utils-operator.v1.0.2
        version: 1.0.2
      - name: cert-utils-operator.v1.0.1
        version: 1.0.1
      - name: cert-utils-operator.v1.0.0
        version: 1.0.0
      - name: cert-utils-operator.v0.2.2
        version: 0.2.2
      - name: cert-utils-operator.v0.2.1
        version: 0.2.1
      - name: cert-utils-operator.v0.2.0
        version: 0.2.0
      - name: cert-utils-operator.v0.1.1
        version: 0.1.1
      - name: cert-utils-operator.v0.1.0
        version: 0.1.0
      - name: cert-utils-operator.v0.0.3
        version: 0.0.3
      - name: cert-utils-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cert-utils-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ses-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ses-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ses-controller:1.1.1
          createdAt: "2025-09-19T19:37:25Z"
          description: AWS SES controller is a service controller for managing SES
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConfigurationSet represents the state of an AWS ses ConfigurationSet
              resource.
            displayName: ConfigurationSet
            kind: ConfigurationSet
            name: configurationsets.ses.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Simple Email Service (SES) resources in AWS from within your Kubernetes cluster.

          **About Amazon SES**

          Amazon Simple Email Service (SES) is an email platform that provides an easy, cost-effective way for you to send and receive email using your own email addresses and domains.
          For example, you can send marketing emails such as special offers, transactional emails such as order confirmations, and other types of correspondence such as newsletters. When you use Amazon SES to receive mail, you can develop software solutions such as email autoresponders, email unsubscribe systems, and applications that generate customer support tickets from incoming emails.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SES
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ses
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SES Developer Resources
          url: https://aws.amazon.com/ses/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ses maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ses-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ses-controller.v1.1.1
        version: 1.1.1
      - name: ack-ses-controller.v1.1.0
        version: 1.1.0
      - name: ack-ses-controller.v1.0.13
        version: 1.0.13
      - name: ack-ses-controller.v1.0.12
        version: 1.0.12
      - name: ack-ses-controller.v1.0.11
        version: 1.0.11
      - name: ack-ses-controller.v1.0.10
        version: 1.0.10
      - name: ack-ses-controller.v1.0.9
        version: 1.0.9
      - name: ack-ses-controller.v1.0.8
        version: 1.0.8
      - name: ack-ses-controller.v1.0.7
        version: 1.0.7
      - name: ack-ses-controller.v1.0.6
        version: 1.0.6
      - name: ack-ses-controller.v1.0.5
        version: 1.0.5
      - name: ack-ses-controller.v1.0.3
        version: 1.0.3
      - name: ack-ses-controller.v1.0.2
        version: 1.0.2
      - name: ack-ses-controller.v1.0.1
        version: 1.0.1
      - name: ack-ses-controller.v1.0.0
        version: 1.0.0
      - name: ack-ses-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-ses-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco Networking - Compute Solutions
      provider-url: ""
    name: ucs-ci-solutions-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ucs-ci-solutions-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intersight.cisco.com/v1alpha1",
                "kind": "UCSCISolutions",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ucs-ci-solutions-operator",
                    "app.kubernetes.io/instance": "ucscisolutions-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ucscisolutions",
                    "app.kubernetes.io/part-of": "ucs-ci-solutions-operator"
                  },
                  "name": "ucscisolutions-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: docker.io/pkoppa/ucs-ci-solutions-operator:latest
          createdAt: "2023-07-15T15:19:39Z"
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: UCSCISolutions
            name: ucscisolutions.intersight.cisco.com
            version: v1alpha1
        description: The operator is used to automate the creation of  Pools, Policies
          and Server Profile Templates required to deploy Cisco UCS Converged Infrastructure
          Solution in Intersight.
        displayName: UCS-CI-Solutions-Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ucscisolutions
        - intersight
        - ucs
        - ci
        - flexpod
        - flashstack
        links:
        - name: Ucs Ci Solutions Operator
          url: https://ucs-ci-solutions-operator.domain
        maintainers:
        - email: pkoppa@cisco.com
          name: Paniraj Koppa
        - email: chrobrie@cisco.com
          name: Chris O'Brien
        maturity: alpha
        provider:
          name: Cisco Networking - Compute Solutions
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - docker.io/pkoppa/ucs-ci-solutions-operator:latest
        version: 0.0.1
      entries:
      - name: ucs-ci-solutions-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ucs-ci-solutions-operator
    provider:
      name: Cisco Networking - Compute Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-diagnostics
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-diagnostics.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Diagnostics",
                "metadata": {
                  "name": "diagnostics-sample"
                },
                "spec": {
                  "agentDiagnosticsService": {
                    "port": 8080
                  },
                  "auth": {
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "wapdbDiagnosticsService": {
                    "dumpedTables": "[ \"--regexpr cluster_logs_week_*\", \"current_week_cluster_logs\", \"config_metadata_dbs\", \"config_metadata_dbs_id_seq\", \"config_metadata_tags\", \"config_tags\", \"csv_reports_meta\", \"custom_import_handler_dbs\", \"data_point_db_models\", \"db_test_entries\", \"license_server_metadata\", \"operations\", \"result_file_metadata_dbs\", \"result_metadata_dbs\", \"result_metadata_tags\", \"result_tags\", \"session_meta\", \"session_types\", \"sessions\", \"stat_view_query_data\", \"test_run_configs\", \"test_run_infos\", \"test_runs\", \"wireless_globalsettings\", \"wireless_sessionconfigs\", \"wireless_templates\"]",
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:37:10Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Diagnostics
            name: diagnostics.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Diagnostics Operator
        displayName: Keysight WAP Diagnostics Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Diagnostics
          url: https://keysight-wap-diagnostics.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-diagnostics@sha256:97bf8affa7bf1f3a391f5f036313e7cd8196be2eb8e95debeab06c1b6c964c5e
        version: 0.1.21
      entries:
      - name: keysight-wap-diagnostics.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-diagnostics
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JFrog
      provider-url: ""
    name: openshiftartifactoryha-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: artifactory-ha-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftArtifactoryHa",
                "metadata": {
                  "name": "openshiftartifactoryha"
                },
                "spec": {
                  "artifactory-ha": {
                    "artifactory": {
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE",
                      "node": {
                        "replicaCount": 2,
                        "waitForPrimaryStartup": {
                          "enabled": false
                        }
                      }
                    },
                    "databaseUpgradeReady": true,
                    "database": {
                      "driver": "OVERRIDE",
                      "password": "OVERRIDE",
                      "type": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "nginx": {
                      "http": {
                        "externalPort": 80,
                        "internalPort": 8080
                      },
                      "https": {
                        "externalPort": 443,
                        "internalPort": 8443
                      },
                      "service": {
                        "ssloffload": false
                      },
                      "tlsSecretName": "OVERRIDE"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "waitForDatabase": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/artifactory-operator@sha256:de69bb16052d5ac6fed9882eb65d49446b6fa1cc5ba830dd2a39e882712b1347
          createdAt: 2024-03-23 00:00:00+00:00
          description: JFrog Artifactory Enterprise deploys Artifactory in a high
            availability environment across multiple pods
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Artifactory HA Instances
            displayName: Artifactory HA
            kind: OpenshiftArtifactoryHa
            name: openshiftartifactoryhas.charts.helm.k8s.io
            version: v1
        description: |2+
           ## Breaking change
          Please update to version 1.1.10 of the operator.
          Version 1.1.8 and 1.1.9 have issues. Please skip these versions.
          If you are using operator version < 1.2.0 there is a issue with upgrade where artifactory version is not updated if you update operator version. As workaround you can update the Custom Resource OpenshiftArtifactoryHa which has hardcored image value.After upgrading to version 1.2.0 you can remove the below parameters from your CR so that future upgrade works without workaround. ``` artifactory-ha.artifactory.image.registry artifactory-ha.artifactory.image.repository artifactory-ha.artifactory.image.tag ```
          ## Overview
          Openshift Operator to deploy JFrog Artifactory Enterprise into your Openshift cluster.
          ## Security Context Constraints
          To deploy this helm chart you will need to be a cluster admin w/ access to the anyuid scc and add the operator service account to the anyuid scc.
          ```
          oc adm policy add-scc-to-user anyuid -z artifactory-ha-operator -n openshift-operators
          ```
          Add the service account for the Artifactory chart to deploy successfully:
          ``` oc adm policy add-scc-to-user anyuid -z openshiftartifactoryha-artifactory-ha -n artifactory ```
          ### Usage

          An external DB is required. The operator will not deploy a DB but will require you to specify the configuration values to connect to it.

          Search for JFrog and click JFrog Artifactory Enterprise Operator to install.

          Go to the Installed Operators.

          Wait for the JFrog Artifactory Enterprise Operator to complete the installation.

          Open the Operator and click on the provided API: Artifactory HA.

          Click Create New Instance and provide the following parameters for your DB configuration:

          ```
          DATABASE_TYPE
          DATABASE_DRIVER
          DATABASE_URL
          DATABASE_USER
          DATABASE_PASSWORD
          ```
          Master key and Join key must be supplied. To generate a new key for each run the command below:

          ```
          # Create a key
          export JOIN_KEY=$(openssl rand -hex 32)
          echo ${JOIN_KEY}
          ```

          To use TLS you will need to first create a k8s tls secret to store your .crt and .key file into.
          Then supply the value of this k8s secret into the TLS_SECRET field.
          ``` oc create secret tls my_tls_secret --cert=tls.crt --key=tls.key --namespace=my_namespace ```
          Click Create for Artifactory Enterprise to deploy into OpenShift and connect to it on the external IP exposed by the load balancer.
          ### Create a route
          To expose Artifactory from Openshift we recommend you create a new route in Openshift.
          You can either use the oc command line tool or the Openshift web console to generate a new route.
          Depending upon where you terminate TLS you will need to either specify pass through or edge.
          Command Line (Edge):
          ``` oc create route edge --service=openshiftartifactory-ha --cert=tls.crt --key=tls.key --ca-cert=ca.crt --hostname=www.example.com ```
          Or for more information visit the official Openshift documentation on routes here:
          https://docs.openshift.com/container-platform/4.6/networking/routes/route-configuration.html

        displayName: JFrog Artifactory Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Artifact Repository
          url: https://jfrog.com/artifactory/
        - name: Artifactory Video
          url: https://www.youtube.com/watch?v=r2_A5CPo43U
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: alpha
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/artifactory-operator@sha256:de69bb16052d5ac6fed9882eb65d49446b6fa1cc5ba830dd2a39e882712b1347
        - registry.connect.redhat.com/jfrog/artifactory-pro@sha256:0986ca71419ae628892db11e7ca7da5ac1c60219c144ce9aeb2de6e683a45262
        - registry.redhat.io/rhel8/nginx-120@sha256:87f658744af695273c434fac8246b716728b691a17d5d0c5b89247362e61e45c
        version: 1.3.3
      entries:
      - name: artifactory-ha-operator.v1.3.3
        version: 1.3.3
      name: alpha
    - currentCSV: artifactory-ha-operator.v1.1.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftArtifactoryHa",
                "metadata": {
                  "name": "openshiftartifactoryha"
                },
                "spec": {
                  "artifactory-ha": {
                    "artifactory": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/artifactory-pro",
                        "tag": "7.27.3-1"
                      },
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE",
                      "uid": "1000721030",
                      "node": {
                        "replicaCount": 2,
                        "waitForPrimaryStartup": {
                          "enabled": false
                        }
                      }
                    },
                    "databaseUpgradeReady": true,
                    "database": {
                      "driver": "OVERRIDE",
                      "password": "OVERRIDE",
                      "type": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "initContainerImage": "registry.connect.redhat.com/jfrog/init@sha256:164b41720a37f75c6129c8a28d05cc56f821022313d44ff918cce4a2146d93e8",
                    "nginx": {
                      "uid": "1000720104",
                      "gid": "1000720107",
                      "http": {
                        "externalPort": 80,
                        "internalPort": 8080
                      },
                      "https": {
                        "externalPort": 443,
                        "internalPort": 8443
                      },
                      "image": {
                        "registry": "registry.redhat.io",
                        "repository": "rhel8/nginx-116",
                        "tag": "latest"
                      },
                      "service": {
                        "ssloffload": false
                      },
                      "tlsSecretName": "OVERRIDE"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "waitForDatabase": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/artifactory-operator@sha256:e292553e1b6e792ff19276c7f6565d7ed61dd95289eb176066329954bc93e371
          createdAt: 2020-03-25 00:00:00+00:00
          description: JFrog Artifactory Enterprise deploys Artifactory in a high
            availability environment across multiple pods
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Artifactory HA Instances
            displayName: Artifactory HA
            kind: OpenshiftArtifactoryHa
            name: openshiftartifactoryhas.charts.helm.k8s.io
            version: v1
        description: |2+
           ## Breaking change
          Please update to version 1.1.10 of the operator.
          Version 1.1.8 and 1.1.9 have issues. Please skip these versions.

          ## Overview
          Openshift Operator to deploy JFrog Artifactory Enterprise into your Openshift cluster.
          ## Security Context Constraints
          To deploy this helm chart you will need to be a cluster admin w/ access to the anyuid scc and add the operator service account to the anyuid scc.
          ```
          oc adm policy add-scc-to-user anyuid -z artifactory-ha-operator -n openshift-operators
          ```
          Add the service account for the Artifactory chart to deploy successfully:
          ``` oc adm policy add-scc-to-user anyuid -z openshiftartifactoryha-artifactory-ha -n artifactory ```
          ### Usage

          An external DB is required. The operator will not deploy a DB but will require you to specify the configuration values to connect to it.

          Search for JFrog and click JFrog Artifactory Enterprise Operator to install.

          Go to the Installed Operators.

          Wait for the JFrog Artifactory Enterprise Operator to complete the installation.

          Open the Operator and click on the provided API: Artifactory HA.

          Click Create New Instance and provide the following parameters for your DB configuration:

          ```
          DATABASE_TYPE
          DATABASE_DRIVER
          DATABASE_URL
          DATABASE_USER
          DATABASE_PASSWORD
          ```
          Master key and Join key must be supplied. To generate a new key for each run the command below:

          ```
          # Create a key
          export JOIN_KEY=$(openssl rand -hex 32)
          echo ${JOIN_KEY}
          ```

          To use TLS you will need to first create a k8s tls secret to store your .crt and .key file into.
          Then supply the value of this k8s secret into the TLS_SECRET field.
          ``` oc create secret tls my_tls_secret --cert=tls.crt --key=tls.key --namespace=my_namespace ```
          Click Create for Artifactory Enterprise to deploy into OpenShift and connect to it on the external IP exposed by the load balancer.
          ### Create a route
          To expose Artifactory from Openshift we recommend you create a new route in Openshift.
          You can either use the oc command line tool or the Openshift web console to generate a new route.
          Depending upon where you terminate TLS you will need to either specify pass through or edge.
          Command Line (Edge):
          ``` oc create route edge --service=openshiftartifactory-ha --cert=tls.crt --key=tls.key --ca-cert=ca.crt --hostname=www.example.com ```
          Or for more information visit the official Openshift documentation on routes here:
          https://docs.openshift.com/container-platform/4.6/networking/routes/route-configuration.html

        displayName: JFrog Artifactory Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Artifact Repository
          url: https://jfrog.com/artifactory/
        - name: Artifactory Video
          url: https://www.youtube.com/watch?v=r2_A5CPo43U
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: stable
        provider:
          name: JFrog
        relatedImages:
        - registry.redhat.io/rhel8/nginx-116@sha256:d71d368d6094a5777000a8d99164959ccb97fca355e855e24bd92c13a8df04f6
        - registry.connect.redhat.com/jfrog/init@sha256:164b41720a37f75c6129c8a28d05cc56f821022313d44ff918cce4a2146d93e8
        - registry.connect.redhat.com/jfrog/artifactory-operator@sha256:e292553e1b6e792ff19276c7f6565d7ed61dd95289eb176066329954bc93e371
        - registry.connect.redhat.com/jfrog/artifactory-pro@sha256:7289dde95d1e4d8e161a28cffd9c4a0cc96f92f7536d069ab63ec304a92254fd
        version: 1.1.18
      entries:
      - name: artifactory-ha-operator.v1.1.18
        version: 1.1.18
      name: stable
    defaultChannel: alpha
    packageName: openshiftartifactoryha-operator
    provider:
      name: JFrog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: advanced-cluster-management
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: advanced-cluster-management.v2.10.9
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-08-06T03:17:11Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0 <2.10.9'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/internal-objects: '[]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterHub defines the configuration for an instance
              of the MultiCluster Hub
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.10/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:95df18f4479b7f555a41508486624bde714778730921c20c665e9f23baadbabf
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:8f1bb3a8422d29960d8b90dd24a813a56875abf1f0393f78b862eaa48ddda783
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:50331cdd61332c8f4128b3dab237255463b307c5c40b0607730a7d3b33b9c6b5
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:5b162aeab4aa51185c880696d1472e152a20ada926a3baea0fb8494793ad238e
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:0ba2471b7a67532bdb90e02904671f0e724cdb39f74a45392ae6c693c25b354d
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:4b16da3070e30d879421aaa0d4a73aa2a1b0b852f62b8faabd801c03f22d5319
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:ced03229f208739867a43b9ed9bd158f19f4242dd1f5f045d3e5db81b2b1d171
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:4ad5ccf9ce682024b6e3025a9bcadd6f32692aa18d892afd670eacb7b4031860
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:1c2dbcf179508ea1c602f7817f85b0d3f58d6d85ae93ccc523e88a3936c9c690
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:23713ebb12d510d37de7e060d5ce4ff2198990f83cebbd2302e74bd83a55cecd
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:1aaba9ca65e363c7f9c07ef2f66f23cb8fff677204a56373634d319113d2c84b
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:e3c80e8e3af5c02c202dfb288c63b20c44c08f2a0474d2568cfb0b878f7fe888
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:9e78db48ff0b0fc0ac0347ca055a4b66e2f917073499adc6e2db2a9c16a70e01
        - registry.redhat.io/rhacm2/console-rhel9@sha256:1538a3a935aa01d8800d8145c45037eaab6987ad6e6e927017ca1cfea4bb111c
        - registry.redhat.io/rhel9/postgresql-13@sha256:31c77c19480ab82cb33525d32419b68cf59daccf0b9936e88e8238e48a2d47e5
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:fbc5b1456a37951c9f70f64357919222739f4a7e533fb6dde5b93672535fc87e
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:95890bb1aad544d97ae1b80e801a800c95ef51e1034b60e680b67bdb2f5aeb69
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:9e931cab5fcbf3df0ffe6fff8d12303800423ace3d749ee0e33b55300c816b19
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:dcd6a420467929b9af8c52c77c8717d22f3977e6b6b392b5dc90cbb04005b9f1
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:14094f09dfb166a6b342bd91e2db9162f0de88d55655ccce44d345047c700187
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:23b4d7fbc776c63dcd7007442a557124581939ce110257822e8233261d31fb7d
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:bbe57064cb1c40c35929526a082f11b2da6fd3746b6eef8af51ca63a4fd7d98b
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:46e3b8ccf3e7069b44eb3ad407329c03048b5219ef2886e237e1ab0dc3f7bae4
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:bb4b72c1510da84aad2d67516c58f37495d7682c51c8f2088e2c4a350fa36c50
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:dcbc4c5e2d46ec2b4911d8cd9f96a4c141bf7dd18b79aa4c25b77141e4f60b48
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:69a8faa2f4da2281c6c124508634628510d91ad40f725a9b41790e0821d64c14
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:500718f834e05c7a4531eb4f9d8ce560053c5125f76f6921c38c9cf150046a23
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:9db39230c528bfc675bc422a7acd40c394169b9632b088ca16358751994c9cac
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:335c903dde6f21e9c54f025c45666f6b4b6e7d455dfd5f85a6373b3947069f0a
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:3ccc45b0e9483f33be0d3b32a6d46fa794976131ad2f9861d96e319c8e35b9be
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:2f68a16a14f192b2cceaa07ed4840c08fbf413718a39bd584cacde866be7f2b1
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:9c4eb0ef8758d943796e13f65bc92b190db614eabd0fe7375ebfecc90e96e32c
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:da09087ba7d558611d733551b6ebc27ccf441581fc264efe7a5933db3816ef89
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:e02992cec8edcc9374c4cd30dfd7bccbda77d6d5cdd309862ab5265494f1278e
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:f4474ed21a71ad490778a6a378a856bec1884b4b9dc2c0e6c6140cc6c9d03b61
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:805e443fbfca4f816dc2dcea71f4cf723fd753ccea6d13f04d25945f29cb47dc
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:d8b8d130f75647cfeb9036db50d66fc4f1fca56fead6add4ba24c52f9213543e
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:13c33a997be0ddef59d1fb31835751367d4f901a1a6eac41284931bf7988035c
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:814f34f9e55b59d96021d3160b1dad7a68ffe208b3f4a999840b228aa7b3300b
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:47a8fa9541275f849a24e2ef1df87b7e9051dedd0ee8bd833d60defbd8724b46
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:3803d82d2b829c0e0883c0504176a96d2012cb7096cfcace316831a2b55bca3f
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:de68d598587a967d4b3f5e0e28c4675f74fd79cae7a5bb197c1b4b44bb4b31d9
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:35ff861165fccd848b8d328fae46cb8b3ac0c853c63062ec8672a6d41a23f333
        - registry.redhat.io/rhacm2/iam-policy-controller-rhel9@sha256:8c2b356bd9ff4ebfba9059a5f20cad8bd336af237d84cdc21f5c0aa35fabeabe
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:d989e9d2078f8a1e4b18de64f917e6d5e290c5cac769899f829b5f6bbfe6f0f2
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:87093f85c6235c99d4cca9d503b0e4198e3731e5fc3171506485091570e70bb5
        version: 2.10.9
      entries:
      - name: advanced-cluster-management.v2.10.9
        version: 2.10.9
      - name: advanced-cluster-management.v2.10.8
        version: 2.10.8
      - name: advanced-cluster-management.v2.10.7
        version: 2.10.7
      - name: advanced-cluster-management.v2.10.6
        version: 2.10.6
      - name: advanced-cluster-management.v2.10.5
        version: 2.10.5
      - name: advanced-cluster-management.v2.10.4
        version: 2.10.4
      - name: advanced-cluster-management.v2.10.3
        version: 2.10.3
      - name: advanced-cluster-management.v2.10.2
        version: 2.10.2
      - name: advanced-cluster-management.v2.10.1
        version: 2.10.1
      - name: advanced-cluster-management.v2.10.0
        version: 2.10.0
      name: release-2.10
    - currentCSV: advanced-cluster-management.v2.11.8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-08-01T05:25:37Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.10.0 <2.11.8'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.11/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:956d549e64c412f86f2083c8ddf97fea269d7806d8826aab93178bec9ed5d7af
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:9d821f4373722d9cb2b679cf5793ef22cb0b9ebcbb5a38060504c93c05c492c4
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:aef5ffbaa578d006bc7caec8fb35446dc263a51813f84c756aa57903b428329b
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:d248a5fad4ae3ee3210eb6dc4f58406d2f1e8f5c390f3c491040b437fe1adeba
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:447b1f6067a82476e14dfbb74a8f324fb86956420e55575f0017b91784f303a5
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:a081f531efb5b54d7d14d9ba0519552ecfc87c3a80d9d3ec4e598d2b445fc689
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:ffd3d9c363d8e15232bd6d12c5f1a77f7cb1ca18f94b7aea79d226568be4586e
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:95890bb1aad544d97ae1b80e801a800c95ef51e1034b60e680b67bdb2f5aeb69
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:f5c56a7381eb80c02a7c140f7c3f0ed08bb93697cd51732ee57498c080d8d3e5
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:8d15be5f40d135fd7069ac768dcf9188d8d8299c65db56df04e66f8a28941966
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:94c89ebd9257797ddc264511399689feb70d56e53332d9894c6d95fbdc51de40
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:6d4e3114494e051b3b875eeeae4bad4d48665fbcbbd3574a98f87b5a5742b41d
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:9be2d97fa3c1140db7d86157503a96ca2d3bdfb57125ab2230c8cf0649ea1e90
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:d4feeb3d3322210765d916ef1f49b47a1527f1c61041f94834c4395319a9ec32
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:49f0ec1aced72e6a2f0c862b0c6c0cd430d997a09ceb9eb0a7d25aaa1d9798b7
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:a9e0533af92ebadca27c9fe93fe4db9c68a1a8bc0060f4d13f67f26eb5691cb3
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:b3e0427341e3b1cbb1fe9166c56184eb0b0ec526d8a3992306e4861f65e4f0c0
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:ded47519fcf30e60cf38d44a101494b0318757969675782365de88b51cab24f0
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:968eb25c5cfcbdcea87b92dbff734d01b5954c13c13f5e794ad1db7bd886e2b4
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:d7552b2a3c864a3fc0d279a2519535e170fd8608195fe1f9abb4b538dfe2a3df
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:d6666c572294c7c85990304d21eaae3040b9c4e47fcb7b446289cb09aabc2d26
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:4a62d8706772ede443cffa005b984c26d137c35a33aafedaa7f01050b618c2a2
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:d4bfef313ff9044ae36aa5f832651a4dd09228d4b9b186803009aafe9774ee2c
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:1e8a26ebfcc33e60f60da4ba8d58cf7c152a99752eac6ebc9e114187e4cc2267
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:4dce3cebeb4ea0f1d3fc1c7e036dc6a3fd09e02f8f2ae4f3a753d9f4f97defe7
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:b713264826124df2c2b3a99dc268a3f3bd7eb24a9010b93c9876d83656e70e1a
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:e36832aa4852ea807d44f2b1f2c402bdaac6b7e2246bf9dd811b41f0b5e8e647
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:eaa719f9d9db8f91eced81e0e164bbc12bb502bf4760fc8c673cc542b4414771
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:45d4404abe02596611b29316bd10985f7a0684b1d305a57d3b24068ac776cf6e
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:251d3ee881c3e902e87f1752648ab084e8780aef79eb88a253ba79f984a01bcf
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:e347cd0f57e0691c61263e6554495c76ce56c7443bce3fd010cf76d068083fee
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:ad10d15efb6833b2daaf153210a080420395a7debfcb873dd757bb924197a654
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:36c90a52b12af24faaf28301d0d1a0c522723bd6cadd7391e5ee1f43f3019eef
        - registry.redhat.io/rhacm2/console-rhel9@sha256:2ed2f4add387caf790b2b41a65a3986e0046e08d0b8406ad17e4f0440ba0ac4a
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:5ba07381111651906ac3c93344e705152eb006826cb36bf5c05bd88723c4c999
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:7d39d0c29b02c399ffdcd83278693a8dd9c5ca3931e660af056030b126d10055
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:eec0e9bd6ea6f146939516da6c7ca9c3755888d96e81954f2f282fae63c1847c
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:827ca1beb707c02908914c517c8ab14b77b1c65ca7690dbbe1bcff600b7551d5
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:4e0c2cf5229fd68102a9e6d0f3155c651b1159b7af7b057151ce52990f747b08
        - registry.redhat.io/rhel9/postgresql-13@sha256:31c77c19480ab82cb33525d32419b68cf59daccf0b9936e88e8238e48a2d47e5
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:ef66846a3f178234618fd1e265067f863bbb56f14cbca0dfc96b0d741885dc9b
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:4b16da3070e30d879421aaa0d4a73aa2a1b0b852f62b8faabd801c03f22d5319
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:028b7258ec88d04397798552ed19a459110660ddbdf3ab375a7a1f74049c455a
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:147c567ce95673ba7151bcd290474756e13018ebc9f7275d61677ffe586a8a76
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:2878b5c41f8d58e986e86803bfd7eb3a05ccf23f7e49ebba56ad30245609b23d
        version: 2.11.8
      entries:
      - name: advanced-cluster-management.v2.11.8
        version: 2.11.8
      - name: advanced-cluster-management.v2.11.7
        version: 2.11.7
      - name: advanced-cluster-management.v2.11.6
        version: 2.11.6
      - name: advanced-cluster-management.v2.11.5
        version: 2.11.5
      - name: advanced-cluster-management.v2.11.4
        version: 2.11.4
      - name: advanced-cluster-management.v2.11.3
        version: 2.11.3
      - name: advanced-cluster-management.v2.11.2
        version: 2.11.2
      - name: advanced-cluster-management.v2.11.1
        version: 2.11.1
      - name: advanced-cluster-management.v2.11.0
        version: 2.11.0
      name: release-2.11
    - currentCSV: advanced-cluster-management.v2.12.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-21T21:16:17Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <2.12.4'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:b83e630e77bd011c374d73d801f10a3230f6798ede10d7f39bf2e8ff6437cf2c
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:3d3260ff206c5ee249443b1f8da49562584f7ee1abd284691b42f1728b6ed5dc
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:b85f3de5231f2f1bc3b4e9c95917f562f39e09dc9b7db35690024163e3a5afe5
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:6f2c4f1abc4388e76414544e65f08238c96fc4d6fbf71fa79330d9dca117d221
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:ec65a5744945ca20f804d94999a4002efab9ddad360a58f9498800bffa9fa0f4
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:187879345174e515863bc1b29a961a0a0d9ad44006a6e12de715b4508dec38a8
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:93df0586f9281fa2879eed0fdf2fb115be355d572cc24dc5c5c9a92137044877
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:58f08a93cd35bd2ac19f94a1779205c81a2fc3a552d1e77b7c659fd51c9dd34f
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:bd24cfabb81b464a31c3d8c21fa6371e46b83094d5c750acc33346203fe15e84
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:c2259b90decf24f8fb1cb058d4751c75fd1558e3c9fd1079857ad9c47ca9e8d2
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:22e262c0353c6f0d5d1b916074836c53437e22697f67c85a43e6d717daf651fa
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:c402a4ac61e11dc9e45fdc7bc7fcb5bf5ca3475a66d782aea60d9f6524e51a24
        - registry.redhat.io/rhel8/postgresql-13@sha256:826a1a75f4186f36217b5c1f1aa270838c4b55b1021d33a4876566bfc3b8e629
        - registry.redhat.io/rhacm2/console-rhel9@sha256:d239df44dafac3bde3eb7ae8c80cb342fa4bb4eabe22f27d2c92d283324e6e10
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:9c297b4c1f9ba17908594f1843ea6de6c5a5e0f3500258c23fbb8b5684249c9e
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:cdaf7be69a5d66cfa4777feb8266421dafd55353d6703e7e591bb9fb23017abc
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:7c548dd869f3458fb70a536708306cb615b9166c7e866cc1215e82f183b39529
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:1419e274b4e2e4c37985208c3b1b9ecd0cdef493427ace3c71ff35ed8fcc394b
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:a5c65dee1f7e4177050244109c10e384899e01d1924b41a540373aa7daf4f274
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:c46657831915ed5f42f4a55e725bc9d17e0671191f4509e610468fda160710c0
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:1e5c12ed0de090198e097cc5deb133ad075fba224f239b8b892e6e1672ec9883
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:fff4fb0924f866b305ad0ecc5d2682da6cb20de15448a2a8c640df182719b4e2
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:28730c8f2344e51b0b9ca603ac63b60885b39be86582943bd1cc200779eee0c5
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:1ccd17b482df844ec42936993ff162f665b5a473ad344270ff2425cd63d428fc
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:6a7c54cb7e6c3cc506ffb8a4f3ec7443ef80798e2c917e20b5f06d39bde02bc4
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:7375991ce3b9e64cb7542dc176414f0974047667d08beac184e2adadd9f83532
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:3ac71b363aef8db127b88e0a63ee616d0d06342142f83522e34c9cac4dd77b2e
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:c4c3bc4f45ae194b884d89319625f2a0970dada823877a35feecd8c733cb2c8d
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:7050f71c3a038fec16db8d0eee69d7d5a4294c043c13e9cf69c3cb2e28897f4c
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:70ddf96707d398d5444bb7bbde6c684275cff62b8a83c7cf6143ff529daf0dee
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:f3a00f55dee152e0345e2616ac14f652dcb4f37f3b9dfca6f049ac5522874c6e
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:f1c371fef66e96199c4a96e21192b47bca423b282d6f99e927644840aaedfd80
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:8ccbbc058d324429aef0c31c99de8c7b3a5abc02ec6f7a14cc60eef63a8f5c69
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:7b64803ddd3b7886ad3ec3de943a38cec5aeb9604732ab7a5ea5061748bf1411
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:5156129064dbe4e9989b41018b1bb250b2eb3f02e0c0afd75a874da9b54ad3c2
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:875e4bf42ad192cfb954e38c8474a529aa938855d69dc0d83c4ee4a4777c1028
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:3faa2318f159d264a9194fc01239d3a2098c184b365b153e3c56e555d5d4992e
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:7968e96b48b2ed63e7657793cf6cc2564055809df19cb43b7e284ac03f2cc26c
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:dca845ad05b02619b81f231eb57dd826720c221d22282fe7422ceebe57191371
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:794ef15591f217be629d6c680c56e84962923993f6dd77514ca5d2bdbd96217b
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:c10a72bf8c0d76b9b4d9d7c0bfca4d83a3f7dd7f6a37b4cc24ee7591638a2482
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:08dd3f6805fa02478671927982a2c77bdec0a5b7e7fd1c9faf20c8af2bb7b428
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:fbfade039739179995828b2ce4f4411b1315ac9a0e4a98e96a115012676a859d
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:b782190b3a285003d319321be09a25ee1376c951cb0cd146cbb634c9b1ca97b3
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:8e62fa4decf01a3beae7080f19469f33822f5010a688ada5db8d2d23b008e498
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:3f286e3c24fe73b4b42e6cf4b6769f91e333014450e1ab135b1e5b4a271cbaec
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:c2ef90b4704251dc9f7ef5288f5bad7aeb4cd2b3706ec531c60f6f4346073a77
        version: 2.12.4
      entries:
      - name: advanced-cluster-management.v2.12.4
        version: 2.12.4
      - name: advanced-cluster-management.v2.12.3
        version: 2.12.3
      - name: advanced-cluster-management.v2.12.2
        version: 2.12.2
      - name: advanced-cluster-management.v2.12.1
        version: 2.12.1
      - name: advanced-cluster-management.v2.12.0
        version: 2.12.0
      name: release-2.12
    - currentCSV: advanced-cluster-management.v2.13.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-09-08T18:37:07Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <2.13.4'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:dc271885425ab74e0554b531a2078133ce02006d0bb9688f3b54a850cfefd8f8
        - registry.redhat.io/rhel8/postgresql-16@sha256:89dfeaefc7e9e974e47a3365e22f354314f8a5e906e93a0152afe4e42f6a82d0
        - registry.redhat.io/rhacm2/console-rhel9@sha256:5526edf70b784bddfc2b7e1c154a44201c346fc20117635500bb54d325921c25
        - registry.redhat.io/rhacm2/acm-flightctl-worker-rhel9@sha256:57f69874b2df77c185374ddcea530af6783410a0f35528ee1a44cc2dee6f7fa2
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:0040791500337f88af7d922895487549915faa3b69dacba1a23540e7d6f2e580
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:ba51f7d077858a0abe32a254970390b50cd7d75433b5661e17f887c48c56a30f
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:65e1f0eb4e6c797c6dc6518674b28370db3fdaa6d7e17f4233c07207224dcb48
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:3028d011e7b88dd9597821128275f606fdba639e5b83e4e6a8783fa2616cba6f
        - registry.redhat.io/rhacm2/acm-flightctl-api-rhel9@sha256:8302b536c896290533f3f50084cb9047cb55af3a4404887b70704d23862619fa
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:d29ceda561f36faa581d76d8b889f7e01d253d33c88709078855ae26cada4917
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:0272215ab43b5688295d94cce0de4490a2fbabc1dd09ed41645be7acef1e3c9b
        - registry.redhat.io/rhacm2/acm-flightctl-ocp-ui-rhel9@sha256:89863ce0eb5fb73272bda8762d74461552f8502fbe262b16cbde01064cc533b3
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:84109be1814719c343b8b4676981cdbcea0356f012b6c854221c271bdb518fea
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:fc16203de469dd1ba16e428a9d7efbb16a5f511d866ef6fcc253530de1f032a6
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:ab1e4f8685896e2016a0b4c4479378058314f365e4f09ef42cba67f3649bb306
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:b3535378919524df259b85587ca83f5620433914418c9941374951c36e0530ae
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:1b7e6e6a4ae36209f98b1539c37c5975a82363b33fe5ac1e3a320e58b99be239
        - registry.redhat.io/rhacm2/acm-flightctl-periodic-rhel9@sha256:2c9b3c3d6cd7e435da337447580dfbfde8c4932308405eeae5d97ed5ec57fbfc
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:588678bea97237d87cd2e55a0e8000e73461cfc421cc91d554a2f1e352b68d14
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:7f742f024c787932932b8a4100a5329e70aad39537e0c20c2a4e245fc6245e25
        - registry.redhat.io/rhel9/redis-7@sha256:848f4298a9465dafb7ce9790e991bd8a11de2558e3a6685e1d7c4a6e0fc5f371
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:ccc9c0524149820d701c39355af89287ad749832053ea55d1585e77347fa0ec0
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:7b15c6b492e5cb2725b9b345aa3bfae4108e2815bc233ac063b81a67c6550fb6
        - registry.redhat.io/rhel8/postgresql-13@sha256:826a1a75f4186f36217b5c1f1aa270838c4b55b1021d33a4876566bfc3b8e629
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:b1620cebc62240f93b12e6937d15760440be119c421c127214974bdfa9d08ebd
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:0d8202cd10170bb0a138d52a1d71e630ff74df466541897ba86ed324a75ea1d8
        - registry.redhat.io/openshift4/ose-cli-artifacts-rhel9@sha256:67ff9f19edac8d623ce9080a8c05cde3da73afd68e7352e4ee5b22a9ccb596bc
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:3490daccc203fade26e7df348c74a907ab0ffc8901f4965606c8227849087c44
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:f74fb6b4d42f1d7431bfc9d2424f953732fed06d2abd8d1fc27b878b4d1322d5
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:ce342242c50c103fc5e4b0126551231977b35bd477eb627fb8f5d2a6510ddaa0
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:e23115b10a92c323b4cdaf0412919c273dbca33f19910120498bd49cb5810ff5
        - registry.redhat.io/rhacm2/acm-flightctl-ui-rhel9@sha256:86800d67ec8ed63e5408947ba6a441e5afc1050b72a86487b9357674ed495a7e
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:f1db9b10481319dc142ca6046e4e3141addb4c0afab482eeaf80643fad9ad039
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:96bdb2e25ae0f381de0c94a09a3cd11f3c0bdc1d8cfb4fd5ced2cd7b24fbe2c2
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:4d068813e3eb704ea7b2d342d5a2ce8fa2ac71793b1b37634c18704ef0a5e8d7
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:5fc35a944895cf7ea6afa5929c45304aaa9f17397b4f1e42cb8c799d9aead9d5
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:f9ba60c2a415a863c6bd6030fb4663913368a9c271bb40b8acfdc294dcf3a7eb
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:8b152051ef2c4c077379e72aa510c0bd613b219254cad66fab0cd3ec0d1d4bf1
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:04bcddae590e2debf35090055dd52ae4bcaac4b369727266be53ac64670203d0
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:49e99f41cbea3faefc3ae93385a10b544ce91df0d6639cfee65760be6ae0df00
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:81cc71faaf20a986d233fee799a7d2b954cd3f43875a2689f76c53920ccf4df0
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:67939d6e4342372820d7af6777f0ffd123e86e4e8186132c3d4892cb717a5e7c
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:6f95308a508499d2d995653d3fd2def94ac5170b4c3f391d32a1e0594250c33a
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:669e9e2178476bc7a4688f3ebf4c20402608bb4b10e4df6863ec821c7e572299
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:ae1f175ffbd3698f2b3036bf9b62b5c9b24729cfb8248345b8bdd1525ccc5bb4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:ab0192a51ee37b7508f3e4db66ffa53b904dcb1f76d9aa1ac0e80447d8ff9e6a
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:5aec572fc18c8ed533620553aba2c676453297ad2440336a5228a5a4f50f5782
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:a0218b9d5f1803f60d9b295d1504a870780895cffe8f1c5b3ce4c3553da5e4f9
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:a2ab651af1c97e54d2c14de78ff554c0f25e8d000af93a74fc7b7ed913f4e338
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:72c6896a5a4dbb7717d76d9eeade96130d97909eee07e8d3cc17af0658b306ab
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:9ef9f1e2010d77211126482c03e8c72dfb7458940dc95ac952ee8dd1dedd2991
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:6ed6f62f5c05a24938335287b785b80c253424e1ee20d83311cd87c6a2090895
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:2da264ca59278f63990e579939371d45afec3bc7d9922a608520e7647a0377ac
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:1393744757874e79d1fb058cb5d3fe7dcbbcd2f187bdcdcdf34e67227c61fbb9
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:296a31e1aad9f6288bdf80fb3191642e3dd404e9e8fcad198016bedbd3abe9b4
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:3cbb9d271c2f23d430a623c83606bbc8a80379c24d13e610886ad51fd21ce1e4
        version: 2.13.4
      entries:
      - name: advanced-cluster-management.v2.13.4
        version: 2.13.4
      - name: advanced-cluster-management.v2.13.3
        version: 2.13.3
      - name: advanced-cluster-management.v2.13.2
        version: 2.13.2
      - name: advanced-cluster-management.v2.13.1
        version: 2.13.1
      - name: advanced-cluster-management.v2.13.0
        version: 2.13.0
      name: release-2.13
    - currentCSV: advanced-cluster-management.v2.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-25T02:10:33Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.13.0 <2.14.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          repository: https://github.com/open-cluster-management/multiclusterhub-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MulticlusterHub defines the configuration
              for an instance of a multicluster hub, a central point for managing multiple
              Kubernetes-based clusters. The deployment of multicluster hub components
              is determined based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n### Where to find more installation
          information\nYou can find additional installation guidance in the [install
          documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:364c7416a9642dfb37f05e2eac4427918959b12b159ac90a9ca310c584d6434a
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:69f0b9a7fb494572adf2795be865d6d6ba4b67db0eed0e9d61f5dcb5e9bf6efe
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:c2ef90b4704251dc9f7ef5288f5bad7aeb4cd2b3706ec531c60f6f4346073a77
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:0e1eba23e0e8454f2dec240704e7ccc922351d58880f72514d3b2bc30c9e995e
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:c6ff6cae5bc6447d0ed24e458fc5d5fc709b2106fe93b143f9546bd962ea7f87
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:ad7f30c61bb11c0d32686e75b01f3b286a1fb06edd8d2ef6a9e2d924fba2b565
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:b489229190f9a3da66bd154e65490c3a35d1999d2db6707759f06eadb02b968d
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:8e6ee423acfafe28c002911e832d47dffc00a5227325bd6a5b1ab359c7eee5f4
        - registry.redhat.io/rhacm2/console-rhel9@sha256:d4c42a203110050a3834be8d5a0eee384ad7e9fbb991da589d75acd15055b8b1
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:0cecc3d87afcd0d63bdeae3ecae37f8a38534fbdc8bad5c548e0d4f4fb0156a4
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:8f4f59e237f366f08367cabbea3eb51e766ad52db0d35c5efc56182c13fc5c09
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:8253a9225f30b0f0a405e2c3293924acbfbf81f3d9c50bc4431fe4052a9fe908
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:55ef2c163b0c0deb3402e39acf37ff6bb05e2a5d268dcae5f8a400128d67c584
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:b7b7581e7743148d19283250d3300d292b96c48ecae96f6e6202a7ca8204e2ab
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:c329cfd004c58a7fcc40f4fbb986399dc2fe08d7922d2b4f210664f58751b1a6
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:143f49d2ce1f58282e276396722ba37165cd1bf2051d537bb4e6208786e5aed6
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:53deacec64238607afda29c27463d72ac3b582c9670e2e77a9c7e48851214795
        - registry.redhat.io/openshift4/ose-cli-rhel9@sha256:a48758b69b9a7451b7e0caea465a181f854b41f1e74ce86c15216ee1bb31fcb9
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:f9a9e10fde9065d650a7dc8eba83349e5db0a21727027464cf144724d0de8cfa
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:935baf67627d5e76c0766d5d2a9e36b398a12535f097749a987dd0c5de788570
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:0252cf95019717f03c7ac72fb034b5bd9ad7aef2da0a92dd21c56cf47f03b812
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:d56b997a9084d56c63f539147052e64ba30258afe91a9ff832f3673cbccbe467
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ee742e7baa8bcbbfccf79abcf15d22347d14f280eafaf22c42cef71051b10fba
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:0ab21d4189669901fe085c7f3a4b4c9778c853f8abd706bbfffcb59942fbc8e6
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:9c79b1153fc927e7582c5346ea0408e03f794eb983597f02a5e816ae961ba86b
        - registry.redhat.io/rhem/flightctl-periodic-rhel9@sha256:ee59f307e468e2b7b38f7c040898d587d441f8fb3d59cb022ae23e4717739e3f
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:2b1b1c707e0b5217c66a96fb82070de47fc0342747081d4be0049c4542944c8a
        - registry.redhat.io/rhel8/postgresql-16@sha256:d4abbd9492347e0aa7e1b0ff967e74d4b6a4edd6fd20d896f2f2ae03c623181c
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:46c5835a4bfc0e4767d0959bc47e939c195be197e9715aae17788d4d0e799d56
        - registry.redhat.io/rhel9/redis-7@sha256:ff31822cba556e40b6b00720839cc56f474fdb9257fb39b7604fafc11761d379
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:0ea80a1039d217e899ea012b77e99a7c5832e5b7e292dca473759c449bf2d1ea
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:e063c54c7182b70d78d24b5026f174436885383980c9b1e3e7664fa2adfad883
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:d550013d385552d6242b56c5b60f15b3ebb488c2033f808de54bcc639abc2f30
        - registry.redhat.io/rhem/flightctl-ui-ocp-rhel9@sha256:973eb3ca3da32c87ba308928a21fbd25514b2f6a6662300d4928479cb554cff3
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:bcb5949ebc31bbf6b45dfd34aae9c80c82a08531e8a3d9f6eef3354f942a5a5d
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:cee29bdb03927dc1866c267232d391ea5ee762b23c6b04a777e7d61d14a61bba
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:0cc23b5063c0d3264ed345dad8453901e4db18cc9bed1656f2799f72de75bd53
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:8151a89ebfebf2e6d336849fdd329b123b7f486938808ee80bf6bd8d44f6985c
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:7ee0aea84bf00b7b42027fa3e970c1f4b279e345a698b4050c9f413eb0383bb1
        - registry.redhat.io/rhem/flightctl-ui-rhel9@sha256:940c0564ced75b7912d3ffc923a1f6bab9aed0f4a52962f7f894d08937b7d37e
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:ce5ff191023b5e14c339b2a03aa1be05a12d024efa5302652065597e630996b6
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:90d399736ca81bb244415efaf05738beff806912a6ac7444a7d4839089b7a081
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:556d737195a4878877c650a93619af1528e52694f2a8e47e13463a603d2f4900
        - registry.redhat.io/rhem/flightctl-cli-artifacts-rhel9@sha256:3491ec932d90112e89f3a4cebe5bdd9ed6a7811f116df43bab8bb5f827129361
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:d9fa60b5b1d6b6c94c5d8b610c9a7b979202de8b46a344e0a65d06f6ef21a7ef
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:40b3e063435eea6e526a57d6bd1ae489ca703e81eeca82dda1c2e6aedec1262c
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:95085fc33a0150a39b6ef4e616fd23f0bddd9dcad0e75c4a89ed27113626e6eb
        - registry.redhat.io/rhel9/memcached@sha256:955138db9a7a23ee31198aabebd0811438fee11bb0b624c38ddb90efe52bd07c
        - registry.redhat.io/rhem/flightctl-worker-rhel9@sha256:1b1f6f85af0910f1acd029f54a608b25d3a2977c3d8b323266391543324305b4
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:f3e34b200e9480363c56da1cd88b2de37fc6a490691e4596a132217f8ea286d8
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:c70242a037e36690d8dd1030254dde778917356b4f1f61db2cecb03b6c2d7380
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:cb5346afe1c7b58eb758cdced94f2c1161c1621033b52f29757efe3a50320a65
        - registry.redhat.io/rhem/flightctl-api-rhel9@sha256:83c6c7819e04aa58d961518367ce6ad3159a28300ca097967f630044d181be34
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:8e57af77cc5dee3aaa83f6ae6e62e232c5fc4fbe4dca8eb755226351963058e3
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:44e305ca15e7c9b50bd7fa8bb391f5ee66906aa26ecc5f612d029aaee5bb4cb8
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:d65aef370a7ed01c0c7cc0ad7f0800359db7863db4ec8845de5b450992f26933
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:ad534f7b0efea45b9f4a0a026c76a2a030cb608acdb835dfea3852332acfcae1
        version: 2.14.0
      entries:
      - name: advanced-cluster-management.v2.14.0
        version: 2.14.0
      name: release-2.14
    defaultChannel: release-2.14
    packageName: advanced-cluster-management
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.17.40-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5-beta.1 <8.17.40-beta.1'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBM Turbonomic software automates and simultaneously optimizes
              performance, compliance, and cost in real time. Turbonomic discovers
              and automates resource optimization for the complete application stack.
              Applications are continually resourced to perform while satisfying technical
              and business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
          - kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Optimization for Kubernetes
          IBM Turbonomic software automates and simultaneously optimizes performance, compliance, and cost in real time.
          Turbonomic discovers and automates resource optimization for the complete application stack. Applications are continually resourced to perform while satisfying technical and business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (proactive rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          Turbonomic assures application performance by giving workloads the resources they need when they need them.

          ### How Does It Work?
          Turbonomic uses a client agent called — KubeTurbo — that runs in your cluster to discover, monitor and optimize resource usage within your environment.
          KubeTurbo runs together with the default Kubernetes scheduler and sends data back to the Turbonomic for analysis.
          Turbonomic determines the right actions that drive continuous health for your workloads and your cluster.

          Turbonomic discovers Red Hat OpenShift Virtualization resources through a new or existing target configuration with Red Hat OpenShift. For more information refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=targets-red-hat-openshift-virtualization)

          ### Get Started
          Please follow the [Kubeturbo Documentation Page](https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets) for all deployment options and an overview of typical use cases.

          ### Known Issues
          For known issues refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=notes-known-issues)
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Kubeturbo Documentation Page
          url: https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets
        - name: Kubeturbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
        - name: IBM Turbonomic Application Resource Management
          url: https://www.ibm.com/products/turbonomic
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/kubeturbo-operator@sha256:7111c7c4002a447dc6977dbb73e96507bae56766fc0552fb2d036b20583722ba
        version: 8.17.40-beta.1
      entries:
      - name: kubeturbo-operator.v8.17.40-beta.1
        version: 8.17.40-beta.1
      name: beta
    - currentCSV: kubeturbo-operator.v8.17.40
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5 <8.17.40'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBM Turbonomic software automates and simultaneously optimizes
              performance, compliance, and cost in real time. Turbonomic discovers
              and automates resource optimization for the complete application stack.
              Applications are continually resourced to perform while satisfying technical
              and business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
          - kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Optimization for Kubernetes
          IBM Turbonomic software automates and simultaneously optimizes performance, compliance, and cost in real time.
          Turbonomic discovers and automates resource optimization for the complete application stack. Applications are continually resourced to perform while satisfying technical and business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (proactive rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          Turbonomic assures application performance by giving workloads the resources they need when they need them.

          ### How Does It Work?
          Turbonomic uses a client agent called — KubeTurbo — that runs in your cluster to discover, monitor and optimize resource usage within your environment.
          KubeTurbo runs together with the default Kubernetes scheduler and sends data back to the Turbonomic for analysis.
          Turbonomic determines the right actions that drive continuous health for your workloads and your cluster.

          Turbonomic discovers Red Hat OpenShift Virtualization resources through a new or existing target configuration with Red Hat OpenShift. For more information refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=targets-red-hat-openshift-virtualization)

          ### Get Started
          Please follow the [Kubeturbo Documentation Page](https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets) for all deployment options and an overview of typical use cases.

          ### Known Issues
          For known issues refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=notes-known-issues)
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Kubeturbo Documentation Page
          url: https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets
        - name: Kubeturbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
        - name: IBM Turbonomic Application Resource Management
          url: https://www.ibm.com/products/turbonomic
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/kubeturbo-operator@sha256:829dc6b9800efd3fd77d07f955a53610c7b2888b93495b2db32f1f48e93a05a6
        version: 8.17.40
      entries:
      - name: kubeturbo-operator.v8.17.40
        version: 8.17.40
      name: stable
    defaultChannel: stable
    packageName: kubeturbo-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://access.redhat.com/products/red-hat-integration
    name: red-hat-camel-k
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: red-hat-camel-k-operator.v1.10.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
          createdAt: "2025-01-08T17:00:43Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
        version: 1.10.10
      entries:
      - name: red-hat-camel-k-operator.v1.10.10
        version: 1.10.10
      - name: red-hat-camel-k-operator.v1.10.9
        version: 1.10.9
      - name: red-hat-camel-k-operator.v1.10.9-0.1743522834.p
        version: 1.10.9+0.1743522834.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1742828111.p
        version: 1.10.9+0.1742828111.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741884247.p
        version: 1.10.9+0.1741884247.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741875993.p
        version: 1.10.9+0.1741875993.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741580860.p
        version: 1.10.9+0.1741580860.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739988175.p
        version: 1.10.9+0.1739988175.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739388437.p
        version: 1.10.9+0.1739388437.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739387844.p
        version: 1.10.9+0.1739387844.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.10.x
    - currentCSV: red-hat-camel-k-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1beta1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:9e4fb5369e0479d4a68f60725e7d12abdc66c552babe59f1358f6e6b0f26357d
          createdAt: "2020-10-06T02:45:00Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:9e4fb5369e0479d4a68f60725e7d12abdc66c552babe59f1358f6e6b0f26357d
        version: 1.4.2
      entries:
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.4.x
    - currentCSV: red-hat-camel-k-operator.v1.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
          createdAt: "2022-08-15T20:17:25Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <=1.6.8'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
        version: 1.6.10
      entries:
      - name: red-hat-camel-k-operator.v1.6.10
        version: 1.6.10
      - name: red-hat-camel-k-operator.v1.6.9
        version: 1.6.9
      name: 1.6.x
    - currentCSV: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:e916f9040d3f4feb26cc1534d3ab2722890b34c7fe146cdce8b6085b47206e22
          createdAt: "2022-11-28T10:43:18Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <1.8.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:e916f9040d3f4feb26cc1534d3ab2722890b34c7fe146cdce8b6085b47206e22
        version: 1.8.2+0.1675913507.p
      entries:
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.8.x
    - currentCSV: red-hat-camel-k-operator.v1.10.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
          createdAt: "2025-01-08T17:00:43Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
        version: 1.10.10
      entries:
      - name: red-hat-camel-k-operator.v1.10.10
        version: 1.10.10
      - name: red-hat-camel-k-operator.v1.10.9
        version: 1.10.9
      - name: red-hat-camel-k-operator.v1.10.9-0.1743522834.p
        version: 1.10.9+0.1743522834.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1742828111.p
        version: 1.10.9+0.1742828111.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741884247.p
        version: 1.10.9+0.1741884247.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741875993.p
        version: 1.10.9+0.1741875993.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741580860.p
        version: 1.10.9+0.1741580860.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739988175.p
        version: 1.10.9+0.1739988175.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739388437.p
        version: 1.10.9+0.1739388437.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739387844.p
        version: 1.10.9+0.1739387844.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: candidate
    - currentCSV: red-hat-camel-k-operator.v1.10.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
          createdAt: "2025-01-08T17:00:43Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:cce071239ab7708e6ebe600be4aaf260594672826f2e704f77fb36fd3fff8388
        version: 1.10.10
      entries:
      - name: red-hat-camel-k-operator.v1.10.10
        version: 1.10.10
      - name: red-hat-camel-k-operator.v1.10.9
        version: 1.10.9
      - name: red-hat-camel-k-operator.v1.10.9-0.1743522834.p
        version: 1.10.9+0.1743522834.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1742828111.p
        version: 1.10.9+0.1742828111.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741884247.p
        version: 1.10.9+0.1741884247.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741875993.p
        version: 1.10.9+0.1741875993.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1741580860.p
        version: 1.10.9+0.1741580860.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739988175.p
        version: 1.10.9+0.1739988175.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739388437.p
        version: 1.10.9+0.1739388437.p
      - name: red-hat-camel-k-operator.v1.10.9-0.1739387844.p
        version: 1.10.9+0.1739387844.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: latest
    - currentCSV: red-hat-camel-k-operator.v1.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
          createdAt: "2022-08-15T20:17:25Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <=1.6.8'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
        version: 1.6.10
      entries:
      - name: red-hat-camel-k-operator.v1.6.10
        version: 1.6.10
      - name: red-hat-camel-k-operator.v1.6.9
        version: 1.6.9
      name: stable
    defaultChannel: 1.10.x
    packageName: red-hat-camel-k
    provider:
      name: Red Hat
      url: https://access.redhat.com/products/red-hat-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MySQL
      provider-url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
    name: ndb-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ndb-operator.v1.6.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "mysql.oracle.com/v1","kind": "NdbCluster","metadata":
            {"name": "example-ndb"},"spec": {"redundancyLevel": 2,"dataNode": {"nodeCount":
            2},"mysqlNode": {"nodeCount":2}}}]'
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: container-registry.oracle.com/mysql/community-ndb-operator@sha256:b562dc3303ffae9f4dab1d73fd0fe05833bdaa8b731af4227da81af617c75c6b
          createdAt: "2024-12-18T15:05:46.000Z"
          description: The MySQL NDB Operator is a Kubernetes operator for managing
            a MySQL NDB Cluster setup inside a Kubernetes Cluster.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Ndb Cluster
            displayName: NdbCluster
            kind: NdbCluster
            name: ndbclusters.mysql.oracle.com
            version: v1
        description: |
          ## NdbCluster Custom Resource
          The NdbCluster custom resource is defined with custom fields that map to the configuration of a MySQL Cluster. To deploy a MySQL Cluster inside the K8s Cluster, an NdbCluster object with the desired configuration of the MySQL Cluster has to be created in the K8s Cluster. The NDB Operator watches for any changes made to any NdbCluster objects and will receive an event when the NdbCluster object is created in the K8s Cluster. It will then start deploying the MySQL Cluster, with the specified configuration, inside the K8s Cluster using native Kubernetes workloads and resources. Once the MySQL Cluster is ready, the NdbCluster resource can be further modified to make changes to the MySQL Cluster configuration. The Operator will receive this event as well and will update the configuration of the already running MySQL Cluster.

          The documentation at [NdbCluster-CRD.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md) explains the various custom fields of the NdbCluster CRD and the effect they have on the MySQL Cluster configuration. The [examples](https://github.com/mysql/mysql-ndb-operator/tree/main/docs/examples) folder has a few examples of the NdbCluster resource objects which can be used to deploy MySQL Cluster inside the K8s Cluster.
          ## Deploy a simple MySQL Cluster

          The [examples/example-ndb.yaml](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/examples/example-ndb.yaml) file has a NdbCluster object example-ndb that defines a simple MySQL Cluster with 2 data nodes and 2 MySQL Servers. To create this object in the default namespace of the K8s Cluster, run this command from the root of the repo :
          ```
          kubectl apply -f docs/examples/example-ndb.yaml
          ```

          Various NdbCluster status fields report back the actual status of the MySQL Cluster running inside the K8s Cluster. Wait for the MySQL Cluster nodes to be started and become ready using the [UpToDate](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md#ndbclusterconditiontypestring-alias) Condition defined in the NdbCluster object status :
          ```
          kubectl wait --for=condition=UpToDate ndb example-ndb --timeout=10m
          ```
          The MySQL Cluster is ready for transactions once the NdbCluster is UpToDate, i.e. the UpToDate condition becomes true.

          At any point, a brief status of an NdbCluster object can be viewed using kubectl get ndb command.
          ```
          kubectl get ndb example-ndb
          ```

          The output of the command gives a short summary the MySQL Cluster nodes controlled by the respective NdbCluster object.
          ```
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE   UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:2/2    Ready:2/2 3m    True
          ```

          To list all the pods created by the NDB Operator, run :
          ```
          kubectl get pods -l mysql.oracle.com/v1=example-ndb
          ```

          For example-ndb NdbCluster object, two management nodes (*-mgmd-* pods), two multi-threaded data nodes (*-ndbmtd-* pods) and two MySQL Servers (*-mysqld-*) would have been created by the NDB operator :
          ```
          NAME                   READY   STATUS    RESTARTS   AGE
          example-ndb-mgmd-0     1/1     Running   0          3m53s
          example-ndb-mgmd-1     1/1     Running   0          2m5s
          example-ndb-mysqld-0   1/1     Running   0          78s
          example-ndb-mysqld-1   1/1     Running   0          78s
          example-ndb-ndbmtd-0   1/1     Running   0          3m53s
          example-ndb-ndbmtd-1   1/1     Running   0          3m53s
          ```
          If running inside a openshift cluster, pods from statefulset "example-ndb-mysqld" will not start up. By default openshift runs all the pod with "restricted" scc, but the statefulset "example-ndb-mysqld" required higher privilege. Use the below command to check the exact scc required
          ```
          oc get pods example-ndb-mysqld-0 -o yaml | oc adm policy scc-subject-review -f -
          ```
          Create a user account with the scc and set the user to the statefulset "example-ndb-mysqld". Below commands creates an user with "anyuid" and assign it to the statefulset.
          ```
          oc create sa any-uid-user
          oc adm policy add-scc-to-user anyuid -z any-uid-user
          oc set sa sts example-ndb-mysqld any-uid-user
          ```
          ## Connect to the MySQL Cluster

          The NDB Operator, by default, creates few Services to expose the services offered by the MySQL Cluster nodes within the Kubernetes Cluster.
          ```
          kubectl get services -l mysql.oracle.com/v1=example-ndb
          ```

          The above command will generate an output similar to :
          ```
          NAME                 TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S) AGE
          example-ndb-mgmd     ClusterIP   10.100.207.255   <none>        1186/TCP 5m
          example-ndb-mysqld   ClusterIP   10.103.182.157   <none>        3306/TCP 5m
          example-ndb-ndbmtd   ClusterIP   None             <none>        1186/TCP 5m
          ```

          The NDB Operator will also create a root user with a random password on all the MySQL Servers. The root password will be stored in a K8s Secret which can be retrieved from the NdbCluster object status :
          ```
          kubectl get ndb example-ndb -o jsonpath={.status.generatedRootPasswordSecretName}
          ```

          To extract and decode the password from the Secret :
          ```
          base64 -d <<< \
          $(kubectl get secret example-ndb-mysqld-root-password \
          -o jsonpath={.data.password})
          ```

          One can also specify the password and host to be used with the root account via the NdbCluster spec. See the [CRD documentation](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md#ndbmysqldspec) for more details on this.
          ## Access MySQL Cluster from inside K8s

          To connect to the Management and MySQL Servers from within the Kubernetes Cluster, you can use the example-ndb-mgmd and example-ndb-mysqld services as the MySQL Cluster connectstring and MySQL host respectively. You can also use the mgmd and mysqld pods' hostnames to directly connect to that particular pod. The mysql, ndb_mgm clients and any other ndb tools will work as expected.

          A demonstration is available at [connect-from-inside-k8s-demo.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/connect-from-inside-k8s-demo.md).
          ## Access MySQL Cluster from outside K8s

          By default, the Management and MySQL services created by the NDB Operator are of type [ClusterIP](https://kubernetes.io/docs/concepts/services-networking/service/#publishing-services-service-types) and are accessible only from within the K8s Cluster. To expose them outside the K8s Cluster, the services have to be upgraded into a LoadBalancer type. This can be done by setting the spec.managementNode.enableLoadBalancer and spec.mysqlNode.enableLoadBalancer fields of the NdbCluster resource to true. These options can be enabled when the NdbCluster resource object is created or via an update to the object when the MySQL Cluster is already running.

          Once enabled, the previously ClusterIP type example-ndb-mgmd and example-ndb-mysqld services will be upgraded to LoadBalancer type, and they will be available at the external IP assigned to them by the cloud provider.

          Another way to access these services without enabling the LoadBalancers support is to use the kubectl port-forward command.

          In both ways, only the mysql and ndb_mgm clients work. Any NDB tool which uses the NDBAPI to connect to the MySQL data nodes will not work as expected from outside the K8s Cluster.

          A demonstration is available at [connect-from-outside-k8s-demo.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/connect-from-outside-k8s-demo.md).
          ## Updating MySQL Cluster Configuration

          Once a MySQL Cluster has been deployed by the NDB Operator inside the K8s Cluster, its configuration can be updated by updating the spec of the NdbCluster resource object that represents it. The NDB Operator will pick up any changes made to the NdbCluster resource object and will apply the updated configuration to the respective MySQL Cluster. The NDB Operator takes care of updating the MySQL Cluster config file and restarting all the management/data nodes if required. When the update is being handled by the NdbOperator, it will set the UpToDate condition of the NdbCluster resource object to false and will not accept any further updates until it completes the current one.

          The kubectl get ndb example-ndb command will report the UpToDate status of an NdbCluster. It can be used to check if the current update has been completed.
          ```
          $ kubectl get ndb example-ndb
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE   UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:1/2    Ready:2/2 10m   False
          ```

          The UpToDate condition can also be used to wait for the operator to complete the update.
          ```
          kubectl wait --for=condition=UpToDate ndb example-ndb --timeout=10m
          ```

          Once the update is complete, the UpToDate condition will be set back to true by the operator.
          ```
          $ kubectl get ndb example-ndb
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE      UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:2/2    Ready:2/2 10m50s   True
          ```

          ## Delete a MySQL Cluster

          To stop and remove the MySQL Cluster running inside the K8s Cluster, delete the NdbCluster resource object.
          ```
          kubectl delete -f docs/examples/example-ndb.yaml
          ```

          (or)
          ```
          kubectl delete ndb example-ndb
          ```

          This will delete all the data, the pods running the MySQL Cluster nodes and also delete all other associated K8s resources created by the NDB Operator.
        displayName: Ndb Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - ndb
        - in-memory
        links:
        - name: Document
          url: https://dev.mysql.com/doc/ndb-operator/en/
        maintainers:
        - email: paulo.p.silva@oracle.com
          name: paulofasilva
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: MySQL
          url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
        relatedImages:
        - container-registry.oracle.com/mysql/community-ndb-operator@sha256:b562dc3303ffae9f4dab1d73fd0fe05833bdaa8b731af4227da81af617c75c6b
        version: 1.6.2
      entries:
      - name: ndb-operator.v1.6.2
        version: 1.6.2
      - name: ndb-operator.v1.6.1
        version: 1.6.1
      - name: ndb-operator.v1.6.0
        version: 1.6.0
      name: alpha
    defaultChannel: alpha
    packageName: ndb-operator
    provider:
      name: MySQL
      url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: VASTData
      provider-url: https://www.vastdata.com
    name: vast-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vast-csi-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"storage.vastdata.com/v1","kind":"VastCSIDriver","metadata":{"name":"csi.vastdata.com"},"spec":{"applySecurityContextConstraints":true,"attachRequired":true,"controller":{"dnsPolicy":"Default","extraArgs":{"csiAttacher":[],"csiProvisioner":[],"csiResizer":[],"csiSnapshotter":[]},"nodeAffinity":{},"nodeSelector":{},"podAffinity":{},"podAntiAffinity":{},"resources":{"csiAttacher":{"limits":{"memory":"200Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiProvisioner":{"limits":{"memory":"400Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiResizer":{"limits":{"memory":"300Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiSnapshotter":{"limits":{"memory":"200Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiVastPlugin":{"limits":{"memory":"400Mi"},"requests":{"cpu":"100m","memory":"50Mi"}}},"runOnControlPlane":false,"runOnMaster":false,"tolerations":[]},"deletionViewPolicy":null,"deletionVipPool":null,"dontUseTrashApi":false,"driverType":"nfs","endpoint":null,"image":{"csiAttacher":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-attacher@sha256:b4d611100ece2f9bc980d1cb19c2285b8868da261e3b1ee8f45448ab5512ab94"},"csiNodeDriverRegistrar":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac"},"csiProvisioner":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-provisioner@sha256:405a14e1aa702f7ea133cea459e8395fe40a6125c088c55569e696d48e1bd385"},"csiResizer":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-resizer@sha256:a541e6cc2d8b011bb21b1d4ffec6b090e85270cce6276ee302d86153eec0af43"},"csiSnapshotter":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-snapshotter@sha256:2e04046334baf9be425bb0fa1d04c2d1720d770825eedbdbcdb10d430da4ad8c"},"csiVastPlugin":{"imagePullPolicy":"IfNotPresent","repository":"registry.connect.redhat.com/vastdataorg/csi@sha256:7169f279d0dd7130535d74df5b938ed6b7944bfd4fe52d242ceee5178f2ea156"}},"imagePullSecrets":[],"kubeletPath":"/var/lib/kubelet","logLevel":5,"node":{"dnsPolicy":"Default","nodeAffinity":{},"nodeSelector":{},"podAffinity":{},"podAntiAffinity":{},"propagateHostMountOptions":true,"resources":{"csiVastPlugin":{"limits":{"memory":"400Mi"},"requests":{"cpu":"100m","memory":"50Mi"}},"nodeDriverRegistrar":{"limits":{"memory":"100Mi"},"requests":{"cpu":"10m","memory":"20Mi"}}},"tolerations":[]},"numWorkers":10,"operationRetryIntervalMax":60,"operationRetryIntervalStart":10,"operationTimeout":15,"secretName":null,"sslCertsSecretName":null,"truncateVolumeName":64,"useLocalIpForMount":null,"verifySsl":false}},{"apiVersion":"storage.vastdata.com/v1","kind":"VastStorage","metadata":{"name":"vastdata-filesystem"},"spec":{"allowVolumeExpansion":true,"blockingClones":false,"clusterName":"","createSnapshotClass":true,"driverType":"nfs","ephemeralVolumeNameFormat":"csi:{namespace}:{name}:{id}","fsType":"","hostEncryption":null,"mountOptions":[],"provisioner":"","qosPolicy":"","reclaimPolicy":"Delete","secretName":"","secretNamespace":"","setDefaultStorageClass":false,"snapshotClass":{"deletionPolicy":"Delete","setDefaultSnapshotClass":false,"snapshotNameFormat":"csi:{namespace}:{name}:{id}"},"storagePath":"","subsystem":"","tenantName":"","viewPolicy":"","vipPool":"","vipPoolFQDN":null,"volumeGroup":"","volumeNameFormat":"csi:{namespace}:{name}:{id}"}},{"apiVersion":"storage.vastdata.com/v1","kind":"VastCluster","metadata":{"name":"vastcluster-sample"},"spec":{"endpoint":null,"passphrase":null,"password":null,"sslCert":null,"tenant":null,"token":null,"username":null}}]'
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/vastdataorg/csi-operator@sha256:ba7f30a42e327004ecbc81ff5ccc453bf3046a29bab5785b9f8a9746e5c6a3fd
          createdAt: "2024-07-01T11:59:59Z"
          description: |-
            VAST CSI Operator supports seamless installation and upgrade of VAST Data's CSI Driver.
            The VAST CSI Driver allows container orchestration frameworks such as Kubernetes to dynamically provision, manage and scale volumes on the [VAST Data Platform](https://www.vastdata.com/platform/overview). VAST provides scalable, reliable, and fast persistent storage that can be accessed remotely by any Kubernetes application container.
            ### Installation
            Refer to the VASTData Operator for Kubernetes [official documentation](https://support.vastdata.com/s/document-item?bundleId=vast-csi-driver-2.6-administrator-s-guide&topicId=deploying-vast-csi-driver/deploying-vast-csi-driver-on-openshift-with-vast-csi-operator.html&_LANG=enus).
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: vast-csi
          operators.operatorframework.io/builder: operator-sdk-v1.3.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/vast-data/vast-csi
          support: VastData Infra Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VastCSIDriver is a common specification for VAST CSI Controller
              and VAST CSI Node plugins, typically intended for creation in a single
              instance.
            displayName: VastCSIDriver
            kind: VastCSIDriver
            name: vastcsidrivers.storage.vastdata.com
            version: v1
          - description: VastCluster encapsulates the connection details required
              for accessing a VAST cluster. This information is ultimately stored
              in a Kubernetes Secret.
            displayName: VastCluster
            kind: VastCluster
            name: vastclusters.storage.vastdata.com
            version: v1
          - description: Represents VAST storage. Each VastStorage instance generates
              a corresponding storage class that can be used in PVCs to create workloads.
            displayName: VastStorage
            kind: VastStorage
            name: vaststorages.storage.vastdata.com
            version: v1
        description: |-
          VAST CSI Operator supports seamless installation and upgrade of VAST Data's CSI Driver.
          The VAST CSI Driver allows container orchestration frameworks such as Kubernetes to dynamically provision, manage and scale volumes on the [VAST Data Platform](https://www.vastdata.com/platform/overview). VAST provides scalable, reliable, and fast persistent storage that can be accessed remotely by any Kubernetes application container.
          ### Installation
          Refer to the VASTData Operator for Kubernetes [official documentation](https://support.vastdata.com/s/document-item?bundleId=vast-csi-driver-2.6-administrator-s-guide&topicId=deploying-vast-csi-driver/deploying-vast-csi-driver-on-openshift-with-vast-csi-operator.html&_LANG=enus).
        displayName: VAST CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - VAST
        - VASTData
        - CSI
        - Multi Cluster
        - AI
        - High-Performance Computing
        links:
        - name: VAST CSI Plugin
          url: https://github.com/vast-data/vast-csi
        maintainers:
        - email: volodymyr.boiko@vastdata.com
          name: VAST Infrastructure Team
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: VASTData
          url: https://www.vastdata.com
        relatedImages:
        - registry.k8s.io/sig-storage/csi-attacher@sha256:b4d611100ece2f9bc980d1cb19c2285b8868da261e3b1ee8f45448ab5512ab94
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:405a14e1aa702f7ea133cea459e8395fe40a6125c088c55569e696d48e1bd385
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac
        - registry.connect.redhat.com/vastdataorg/csi@sha256:7169f279d0dd7130535d74df5b938ed6b7944bfd4fe52d242ceee5178f2ea156
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.connect.redhat.com/vastdataorg/csi-operator@sha256:ba7f30a42e327004ecbc81ff5ccc453bf3046a29bab5785b9f8a9746e5c6a3fd
        - registry.k8s.io/sig-storage/csi-resizer@sha256:a541e6cc2d8b011bb21b1d4ffec6b090e85270cce6276ee302d86153eec0af43
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:2e04046334baf9be425bb0fa1d04c2d1720d770825eedbdbcdb10d430da4ad8c
        version: 2.6.3
      entries:
      - name: vast-csi-operator.v2.6.3
        version: 2.6.3
      name: stable
    defaultChannel: stable
    packageName: vast-csi-operator
    provider:
      name: VASTData
      url: https://www.vastdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: ""
    name: power-monitoring-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: power-monitoring-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "Kepler",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "kepler",
                    "app.kubernetes.io/name": "kepler",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "kepler"
                },
                "spec": {
                  "exporter": {
                    "deployment": {
                      "port": 9103
                    },
                    "redfish": null
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: registry.redhat.io/openshift-power-monitoring/power-monitoring-rhel9-operator@sha256:5eec15f6fe9bd4bdb1f5fb8057dc67c448c3711bcc7a0a4760f0d90c07688d1e
          createdAt: "2025-02-14T06:23:56Z"
          description: Deploys and Manages Kepler on OpenShift
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/internal-objects: |-
            [
              "keplerinternals.kepler.system.sustainable.computing.io"
            ]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift-power-monitoring/power-monitoring-operator
          support: Team Power Monitoring (team-powermon@redhat.com)
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KeplerInternal is the Schema for internal/unsupported API
            displayName: KeplerInternal
            kind: KeplerInternal
            name: keplerinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: Kepler is the Schema for the keplers API
            displayName: Kepler
            kind: Kepler
            name: keplers.kepler.system.sustainable.computing.io
            version: v1alpha1
        description: |
          ### Introduction
          Kepler (Kubernetes Efficient Power Level Exporter) uses eBPF to probe
          energy-related system stats and exports them as Prometheus metrics.

          ### NOTE
          To view the metrics exposed by Kepler, ensure that
          [User Project Monitoring](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/monitoring/configuring-user-workload-monitoring)
          is enabled in your cluster.

          ### Documentation
          Documentation and installation guide can be found below:
            * [Installation Guide](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/power_monitoring/installing-power-monitoring)
        displayName: Power monitoring for Red Hat OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kepler
        - power
        - energy
        - monitoring
        links:
        - name: Installation Guide
          url: https://docs.redhat.com/documentation/openshift_container_platform/latest/html/power_monitoring/installing-power-monitoring
        maintainers:
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: vimalkum@redhat.com
          name: Vimal Kumar
        - email: vprashar@redhat.com
          name: Vibhu Prashar
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift-power-monitoring/power-monitoring-rhel9-operator@sha256:5eec15f6fe9bd4bdb1f5fb8057dc67c448c3711bcc7a0a4760f0d90c07688d1e
        - registry.redhat.io/openshift-power-monitoring/kepler-rhel9@sha256:27d3cb0a98f1cffd649745e3886ad92995d83a0de6823ddc2d770bf036e2eeaf
        version: 0.4.0
      entries:
      - name: power-monitoring-operator.v0.4.0
        version: 0.4.0
      - name: power-monitoring-operator.v0.3.0
        version: 0.3.0
      - name: power-monitoring-operator.v0.2.0
        version: 0.2.0
      - name: power-monitoring-operator.v0.1.0
        version: 0.1.0
      name: tech-preview
    defaultChannel: tech-preview
    packageName: power-monitoring-operator
    provider:
      name: Red Hat, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtc-operator.v1.7.19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigrationController",
                  "metadata": {
                    "name": "migration-controller",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "azure_resource_group": "",
                    "cluster_name": "host",
                    "migration_velero": true,
                    "migration_controller": true,
                    "migration_ui": true,
                    "migration_log_reader": true,
                    "olm_managed": true,
                    "restic_timeout": "1h",
                    "mig_pv_limit": "100",
                    "mig_pod_limit": "100",
                    "mig_namespace_limit": "10"
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageMigration",
                  "metadata": {
                    "name": "directimagemigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageStreamMigration",
                  "metadata": {
                    "name": "directimagestreammigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigAnalytic",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "plan": {
                      "name": "foo",
                      "namespace": "openshift-migration"
                    },
                    "analyzeK8SResources": true,
                    "analyzeImageCount": true,
                    "analyzePVCapacity": true
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigCluster",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectVolumeMigration",
                  "metadata": {
                    "name": "direct",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                      "kind": "DirectVolumeMigrationProgress",
                      "metadata": {
                        "name": "direct-progress",
                        "namespace": "openshift-migration"
                      },
                      "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigPlan",
                  "metadata": {
                    "name": "migplan",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigMigration",
                  "metadata": {
                    "name": "migmigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigStorage",
                  "metadata": {
                    "name": "migstorage",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:89ba5ccb1cd9c44b8c8d2189f1fc6b04e780d46b19ee8a22e20c0b8c9d94d91f
          createdAt: "2025-04-28T19:55:38Z"
          description: Facilitates migration of container workloads from OpenShift
            3.x to OpenShift 4.x
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.5.0 <1.7.19'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-migration
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:3b6168c6df0d86d76ae4de91755c631b7f6a1f84286257d18c97d5a6af312291
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/mig-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: a request to migrate PVCs from a set of namespaces directly
              to a target
            displayName: DirectVolumeMigration
            kind: DirectVolumeMigration
            name: directvolumemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate watch the progress of rsync pod migrating
              a PVC to a target
            displayName: DirectVolumeMigrationProgress
            kind: DirectVolumeMigrationProgress
            name: directvolumemigrationprogresses.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a set of namespaces
              directly to a target
            displayName: DirectImageMigration
            kind: DirectImageMigration
            name: directimagemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a single imagestream
              directly to a target
            displayName: DirectImageStreamMigration
            kind: DirectImageStreamMigration
            name: directimagestreammigrations.migration.openshift.io
            version: v1alpha1
          - description: A request to Analyze a MigPlan and its namespaces
            displayName: MigAnalytic
            kind: MigAnalytic
            name: miganalytics.migration.openshift.io
            version: v1alpha1
          - description: A cluster defined for migration
            displayName: MigCluster
            kind: MigCluster
            name: migclusters.migration.openshift.io
            version: v1alpha1
          - description: A migration process
            displayName: MigMigration
            kind: MigMigration
            name: migmigrations.migration.openshift.io
            version: v1alpha1
          - description: A migration plan
            displayName: MigPlan
            kind: MigPlan
            name: migplans.migration.openshift.io
            version: v1alpha1
          - description: A migration contoller deployment
            displayName: MigrationController
            kind: MigrationController
            name: migrationcontrollers.migration.openshift.io
            version: v1alpha1
          - description: A migration hook
            displayName: MigHook
            kind: MigHook
            name: mighooks.migration.openshift.io
            version: v1alpha1
          - description: A migration storage location
            displayName: MigStorage
            kind: MigStorage
            name: migstorages.migration.openshift.io
            version: v1alpha1
        description: |-
          The Migration Toolkit for Containers Operator enables installation of the application migration tool components.

          After you have installed the Operator, you can install components by creating a MigrationController CR.

          By default, the Operator installs the following components on an OpenShift Container Platform 4 target cluster:

          * velero and restic, to back up data on the source cluster and restore it on the target cluster
          * mig-controller, to coordinate the backup and restore processes
          * mig-ui, the web console for managing migrations

          If you experience any issues or have feature requests, please file an [issue](https://github.com/konveyor/mig-operator/issues)

          For more information, see the [OpenShift Container Platform Migration Guide](https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html).
        displayName: Migration Toolkit for Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - crane
        - mtc
        links:
        - name: Migrations
          url: https://github.com/konveyor/mig-controller
        - name: Documentation
          url: https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html
        - name: Crane Operator
          url: https://github.com/konveyor/mig-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhmtc/openshift-migration-hook-runner-rhel8@sha256:c8c932a486d9b83e5a93290058d9e84691b87cd4161f8132a1bb6de9a360605e
        - registry.redhat.io/rhmtc/openshift-migration-velero-restic-restore-helper-rhel8@sha256:932216fd08d8ad97baf48ac7b8c989806d30a191337c650df37da537c153dc17
        - registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:3b6168c6df0d86d76ae4de91755c631b7f6a1f84286257d18c97d5a6af312291
        - registry.redhat.io/rhmtc/openshift-migration-velero-rhel8@sha256:bc010bc31e8b553f86aacbca39e28a95e74994c56e85c9fc1b8b80d0791ecb16
        - registry.redhat.io/rhmtc/openshift-migration-log-reader-rhel8@sha256:a0c8ab3a5bbc78c158c2ccccb1d06430a20f40358649196aa4e7311e54f9bdf8
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-aws-rhel8@sha256:023465788745625ad3b4dc0574e35b93d6ac8eea17be2ad1fc88959149a0df2c
        - registry.redhat.io/rhmtc/openshift-migration-ui-rhel8@sha256:66b288170dca8007382d13dff3f2e1868e409fb8002e04ec1fc5c480f868d9b6
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-gcp-rhel8@sha256:cc70c5a63af77948760bc42bf695a5e4eb0bce6285d6b350bba89a53f4b4e7e9
        - registry.redhat.io/rhmtc/openshift-migration-registry-rhel8@sha256:49485dfe708fe1294dec3f0bf34fedc38845a22f91915bba523d728a586b0c30
        - registry.redhat.io/rhmtc/openshift-migration-controller-rhel8@sha256:9b1490b04fc235191515fcd07d7af2e81f846d865016831e17872bb82bf4eaf5
        - registry.redhat.io/rhmtc/openshift-migration-rsync-transfer-rhel8@sha256:23a2bc5ad0f99fc0e7517a0172fdf3597fa619543b0192757983219a578f4ee2
        - registry.redhat.io/rhmtc/openshift-velero-plugin-rhel8@sha256:02698c655a01df2237aee6c570db96f13c0018cee2ae58b96b9570d93c354617
        - registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:89ba5ccb1cd9c44b8c8d2189f1fc6b04e780d46b19ee8a22e20c0b8c9d94d91f
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-microsoft-azure-rhel8@sha256:dad8fa4dbce3b4562bf06756fd74c2007812ec4edbc308e4517ff123c5519767
        version: 1.7.19
      entries:
      - name: mtc-operator.v1.7.19
        version: 1.7.19
      - name: mtc-operator.v1.7.18
        version: 1.7.18
      - name: mtc-operator.v1.7.17
        version: 1.7.17
      - name: mtc-operator.v1.7.16
        version: 1.7.16
      - name: mtc-operator.v1.7.15
        version: 1.7.15
      - name: mtc-operator.v1.7.14
        version: 1.7.14
      - name: mtc-operator.v1.7.13
        version: 1.7.13
      - name: mtc-operator.v1.7.12
        version: 1.7.12
      - name: mtc-operator.v1.7.11
        version: 1.7.11
      - name: mtc-operator.v1.7.10
        version: 1.7.10
      - name: mtc-operator.v1.7.9
        version: 1.7.9
      - name: mtc-operator.v1.7.8
        version: 1.7.8
      - name: mtc-operator.v1.7.7
        version: 1.7.7
      - name: mtc-operator.v1.7.6
        version: 1.7.6
      - name: mtc-operator.v1.7.5
        version: 1.7.5
      - name: mtc-operator.v1.7.4
        version: 1.7.4
      - name: mtc-operator.v1.7.3
        version: 1.7.3
      - name: mtc-operator.v1.7.2
        version: 1.7.2
      - name: mtc-operator.v1.7.1
        version: 1.7.1
      - name: mtc-operator.v1.7.0
        version: 1.7.0
      name: release-v1.7
    - currentCSV: mtc-operator.v1.8.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigrationController",
                  "metadata": {
                    "name": "migration-controller",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "azure_resource_group": "",
                    "cluster_name": "host",
                    "migration_velero": true,
                    "migration_controller": true,
                    "migration_ui": true,
                    "migration_log_reader": true,
                    "olm_managed": true,
                    "restic_timeout": "1h",
                    "mig_pv_limit": "100",
                    "mig_pod_limit": "100",
                    "mig_namespace_limit": "10"
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageMigration",
                  "metadata": {
                    "name": "directimagemigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageStreamMigration",
                  "metadata": {
                    "name": "directimagestreammigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigAnalytic",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "plan": {
                      "name": "foo",
                      "namespace": "openshift-migration"
                    },
                    "analyzeK8SResources": true,
                    "analyzeImageCount": true,
                    "analyzePVCapacity": true
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigCluster",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectVolumeMigration",
                  "metadata": {
                    "name": "direct",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                      "kind": "DirectVolumeMigrationProgress",
                      "metadata": {
                        "name": "direct-progress",
                        "namespace": "openshift-migration"
                      },
                      "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigPlan",
                  "metadata": {
                    "name": "migplan",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigMigration",
                  "metadata": {
                    "name": "migmigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigStorage",
                  "metadata": {
                    "name": "migstorage",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:8a44fe2aea3e842dcccb1e20df4818f4ca573fe13f51427ae8b603d0000a9411
          createdAt: "2025-08-22T15:27:29Z"
          description: Facilitates migration of container workloads from OpenShift
            3.x to OpenShift 4.x
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.5.0 <1.8.10'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-migration
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:f9492269b75608057fbc5edd53603807f1f80cd8581dea12032c88e04cf8f4cb
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/mig-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: a request to migrate PVCs from a set of namespaces directly
              to a target
            displayName: DirectVolumeMigration
            kind: DirectVolumeMigration
            name: directvolumemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate watch the progress of rsync pod migrating
              a PVC to a target
            displayName: DirectVolumeMigrationProgress
            kind: DirectVolumeMigrationProgress
            name: directvolumemigrationprogresses.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a set of namespaces
              directly to a target
            displayName: DirectImageMigration
            kind: DirectImageMigration
            name: directimagemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a single imagestream
              directly to a target
            displayName: DirectImageStreamMigration
            kind: DirectImageStreamMigration
            name: directimagestreammigrations.migration.openshift.io
            version: v1alpha1
          - description: A request to Analyze a MigPlan and its namespaces
            displayName: MigAnalytic
            kind: MigAnalytic
            name: miganalytics.migration.openshift.io
            version: v1alpha1
          - description: A cluster defined for migration
            displayName: MigCluster
            kind: MigCluster
            name: migclusters.migration.openshift.io
            version: v1alpha1
          - description: A migration process
            displayName: MigMigration
            kind: MigMigration
            name: migmigrations.migration.openshift.io
            version: v1alpha1
          - description: A migration plan
            displayName: MigPlan
            kind: MigPlan
            name: migplans.migration.openshift.io
            version: v1alpha1
          - description: A migration contoller deployment
            displayName: MigrationController
            kind: MigrationController
            name: migrationcontrollers.migration.openshift.io
            version: v1alpha1
          - description: A migration hook
            displayName: MigHook
            kind: MigHook
            name: mighooks.migration.openshift.io
            version: v1alpha1
          - description: A migration storage location
            displayName: MigStorage
            kind: MigStorage
            name: migstorages.migration.openshift.io
            version: v1alpha1
        description: |-
          The Migration Toolkit for Containers Operator enables installation of the application migration tool components.

          After you have installed the Operator, you can install components by creating a MigrationController CR.

          By default, the Operator installs the following components on an OpenShift Container Platform 4 target cluster:

          * velero and restic, to back up data on the source cluster and restore it on the target cluster
          * mig-controller, to coordinate the backup and restore processes
          * mig-ui, the web console for managing migrations

          If you experience any issues or have feature requests, please file an [issue](https://github.com/konveyor/mig-operator/issues)

          For more information, see the [OpenShift Container Platform Migration Guide](https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html).
        displayName: Migration Toolkit for Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - crane
        - mtc
        links:
        - name: Migrations
          url: https://github.com/konveyor/mig-controller
        - name: Documentation
          url: https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html
        - name: Crane Operator
          url: https://github.com/konveyor/mig-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhmtc/openshift-migration-registry-rhel8@sha256:accef2a0dd69b254bd120232e5dc86a56aa8f25bcfdba3f9bab7e4127c157762
        - registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:f9492269b75608057fbc5edd53603807f1f80cd8581dea12032c88e04cf8f4cb
        - registry.redhat.io/rhmtc/openshift-migration-controller-rhel8@sha256:c776387c662e21aa61378d0136b2495826e1c157548f9ca9c95d044af38ee169
        - registry.redhat.io/rhmtc/openshift-migration-hook-runner-rhel8@sha256:adaa4f1afefb4d5853713821cb6e7de0d9f5a640f0243f20b41a44be84707caf
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-mtc-rhel8@sha256:5f46ba04146438f2abd6905468d2ecbeb19374a83771d364904ac2ee6afdbefa
        - registry.redhat.io/rhmtc/openshift-migration-openvpn-rhel8@sha256:d21b99d8edefa7a1bab30c85969ec7d7560d06be308bd787e698e409d54cbc19
        - registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:8a44fe2aea3e842dcccb1e20df4818f4ca573fe13f51427ae8b603d0000a9411
        - registry.redhat.io/rhmtc/openshift-migration-log-reader-rhel8@sha256:4f76bcf3df3bb7073b8aabd3da072cacfec05ecf4bfb47e23340074756f7fc30
        - registry.redhat.io/rhmtc/openshift-migration-rsync-transfer-rhel8@sha256:442e7c7f2cc075bf468403ace8846c96ad2fe43eb7f6562d566caa24e4f86873
        - registry.redhat.io/rhmtc/openshift-migration-ui-rhel8@sha256:86e747e0330b0156379d273d0c29c13a65dafbf4f536656723e2bc746872286c
        version: 1.8.10
      entries:
      - name: mtc-operator.v1.8.10
        version: 1.8.10
      - name: mtc-operator.v1.8.9
        version: 1.8.9
      - name: mtc-operator.v1.8.8
        version: 1.8.8
      - name: mtc-operator.v1.8.7
        version: 1.8.7
      - name: mtc-operator.v1.8.6
        version: 1.8.6
      - name: mtc-operator.v1.8.5
        version: 1.8.5
      - name: mtc-operator.v1.8.4
        version: 1.8.4
      - name: mtc-operator.v1.8.3
        version: 1.8.3
      - name: mtc-operator.v1.8.2
        version: 1.8.2
      - name: mtc-operator.v1.8.1
        version: 1.8.1
      - name: mtc-operator.v1.8.0
        version: 1.8.0
      name: release-v1.8
    defaultChannel: release-v1.8
    packageName: mtc-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Performance
      provider-url: ""
    name: cluster-impairment-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-impairment-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "all-impairments"
                },
                "spec": {
                  "duration": 480,
                  "egress": {
                    "bandwidth": 10000,
                    "latency": 50,
                    "loss": 0.02
                  },
                  "ingress": {
                    "bandwidth": 10000,
                    "latency": 50,
                    "loss": 0.02
                  },
                  "interfaces": [
                    "ens2f0"
                  ],
                  "linkFlapping": {
                    "downTime": 30,
                    "enable": true,
                    "upTime": 120
                  },
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "duplication"
                },
                "spec": {
                  "duration": 60,
                  "egress": {
                    "duplication": 25
                  },
                  "ingress": {
                    "duplication": 25
                  },
                  "interfaces": [
                    "ens3"
                  ],
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "two-min-flap"
                },
                "spec": {
                  "duration": 480,
                  "interfaces": [
                    "ens2f0"
                  ],
                  "linkFlapping": {
                    "downTime": 120,
                    "enable": true,
                    "upTime": 120
                  },
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "typical-scenario"
                },
                "spec": {
                  "duration": 30,
                  "egress": {
                    "latency": 50
                  },
                  "ingress": {
                    "latency": 50
                  },
                  "interfaces": [
                    "ens2f0",
                    "eno1"
                  ],
                  "loss": 0.02,
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "uneven-latency"
                },
                "spec": {
                  "duration": 60,
                  "egress": {
                    "latency": 100
                  },
                  "ingress": {
                    "latency": 10
                  },
                  "interfaces": [
                    "ens2f0"
                  ],
                  "nodeSelector": {
                    "key": "node-role.kubernetes.io/worker",
                    "value": ""
                  },
                  "startDelay": 5
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Developer Tools
          containerImage: quay.io/redhat-performance/cluster-impairment-operator:1.0.4
          createdAt: "2021-08-11 12:15:00"
          description: Run Impairments on Kubernetes/Openshift Nodes
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/redhat-performance/cluster-impairment-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterImpairment
            name: clusterimpairments.apps.redhat.com
            version: v1alpha1
        description: "cluster-impairment-operator is an operator designed to automate
          the application of impairments to the nodes in a cluster.\n\n## Features\n\nTraffic
          Direction:\n\n* Egress impairments affect the traffic going out.    \n*
          Ingress impairments affect the traffic coming in. (Ingress impairments require
          the kernel module IFB.)\n\n### Impairments\n\n| Impairment    | Description
          \                               | Unit    | Uses Netem | Correlation Supported
          |\n|---------------|--------------------------------------------|---------|------------|-----------------------|\n|
          Bandwidth     | The bandwidth limit                        | kbit/s  | Yes
          \       | No                    |\n| Latency       | The delay of the packets
          \                  | ms      | Yes        | Yes                   |\n| Packet
          Loss   | The percent of packets that are dropped    | percent | Yes        |
          Yes                   |\n| Corruption    | The percent of packets that are
          corrupted  | percent | Yes        | Yes                   |\n| Duplication
          \  | The percent of packets that are duplicated | percent | Yes        |
          Yes                   |\n| Link Flapping | Turns the interface on and off
          \            | bool    | No         | No                    |\n\n\nOn the
          tested environment (RHEL CoreOS 48.84), the impairments can be used alongside
          link flapping.\n\n\n#### Latency Options\n\nIn addition to simply delaying
          the packets, there are advanced latency options.\n\n##### Jitter\n\nSemi-randomly
          adds or subtracts from the latency according to the distribution up to the
          amount specified.\nIf latency is 100ms, and jitter is 10ms, the actual latency
          will vary 100±10ms.\n\n\n##### Distribution\n\nThe distribution of the jitter.
          The options are:\n* Normal\n* Uniform\n* Pareto\n* Paretonormal\n\n#####
          Jitter correlation\n\nThe percent chance that the next latency's jitter
          value will correlate with the preceeding latency.\n\n##### Reorder\n\nThe
          percentage of packets that are not delayed, causing a reorder due to them
          being sent before the delayed ones.\nApplying jitter itself has the potential
          to also cause reordering.\n\n##### Reorder correlation\n\nThe percet chance
          that the value for the next reorder will correlate with the preceeding value.\n\n####
          Loss options\n\n##### Correlation\n\nThe percent chance that the previous
          loss value for a packet correlates with the loss value for the next packet.\n\n####
          Corruption options\n\n###### Corruption\n\nThe percent chance that each
          packet will be corrupted.\n\n##### Corruption correlation\n\nThe percet
          chance that the value for the next corruption will correlate with the preceeding
          value.\n\n#### Duplication options\n\n##### Duplication\n\nThe percent chance
          that each packet will be duplicated.\n\n##### Duplication correlation\n\nThe
          percet chance that the value for the next duplication will correlate with
          the preceeding value.\n\n\n\n## Configuration\n\nHere is an example of the
          ClusterImpairment custom resource.\n\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: test-impairment-cr\nspec:\n  duration:
          30 # seconds\n  startDelay: 5 # seconds. It typically takes about 2-3 seconds
          for the Daemonset to run\n  interfaces:\n  - \"ens2f0\"\n  ingress: # uses
          ifb\n    bandwidth: 0 # kbit\n    latency: 10 # ms\n    latencyOptions:\n
          \     jitter: 5 # ms\n      jitterCorrelation: 25 # percent\n      distribution:
          normal\n      reorder: 25 # percent of packets that will skip the delay\n
          \     reorderCorrelation: 25 # percent\n    loss: 0 # percent\n    lossOptions:\n
          \     correlation: 25 # percent\n    corruption: 0.1 # percent\n    corruptionOptions:\n
          \     correlation: 25 # percent\n    duplication: 1 # percent\n    duplicationOptions:\n
          \     correlation: 25 # percent\n  egress:\n    bandwidth: 0 # kbit\n    latency:
          100 # ms\n    loss: 0 # percent\n  linkFlapping:\n    enable: false\n    downTime:
          3 # Seconds\n    upTime: 3 # Seconds\n  nodeSelector:\n    key: \"node-role.kubernetes.io/worker\"\n
          \   value: \"\"\n    \n```\n\n#### Interfaces\n\nThe interfaces configuration
          option is a list of all interfaces that the impairments should be applied
          to. The current implementation will error out once it hits an invalid interface.\n\nIf
          an invalid interface is found, it will print out the list of interfaces.\n\n####
          Node Selector\n\nThe default node selector is all worker nodes, but you
          can change it to whatever node selector you want by setting the key and
          value.\nThere is a toleration to allow it to run on master nodes if the
          node selector includes master nodes.\n\n#### Duration\n\nThe number of seconds
          the impairments should run. It will try to sync the start and end time between
          all pods.\nIf set to -1, it will run indefinitely (a year), until you remove
          the ClusterImpairment custom resource. This is good for testing that requires
          steady impairments.\n\nIf the script is link flapping, set this value to
          be short enough so that if the link flap interrupts communication between
          the nodes, the daemonset will remove itself.\n\n#### Start Delay\n\nThe
          delay before starting the script. If you want the pods to be in sync, a
          minimum of a few seconds should be used because the pods take time to start
          up.\n\nYou can also utilize this feature to run an separate impairment after
          another. Just apply two resources at the same time, but add the duration
          and start delay of the first to the start delay of the second.\n\n#### Ingress
          and Egress\n\nThe configuration sections \"ingress\" and \"egress\" apply
          to each direction.\n\n##### Examples:\n\n##### Example 1\n\n\nIn this example,
          egress latency is set to 100ms, but the ingress latency is set to 10ms.
          So the latency to the interface will end up being 10ms, but 100ms going
          out. When pinging, this will result in 110ms of latency.\n\n```\n\napiVersion:
          apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n  name: uneven-latency\nspec:\n
          \ duration: 60\n  startDelay: 5\n  interfaces:\n  - \"ens2f0\"\n  ingress:\n
          \   latency: 10 # ms\n  egress:\n    latency: 100 # ms\n\n```\n\n##### Example
          2\n\n\nIn this example, link flapping will be enabled, and it will turn
          the interface on and off every 2 minutes. That is long enough for kubernetes
          to determine that a node is unavailable.\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: two-min-flap\nspec:\n  duration: 480\n
          \ startDelay: 5\n  interfaces:\n  - \"ens2f0\"\n  linkFlapping:\n    enable:
          true\n    downTime: 120 # Seconds\n    upTime: 120 # Seconds\n\n```\n\n#####
          Example 3\n\n\nIn this example, a realistic set of impairments is applied
          to `ens2f0` and `eno1` for 30 seconds:\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: typical-scenario\nspec:\n  duration:
          30 # seconds\n  startDelay: 5 # seconds\n  interfaces:\n  - \"ens2f0\"\n
          \ - \"eno1\"\n  egress:\n    latency: 50 # ms. Bidirectional, so total of
          100ms\n  ingress:\n    latency: 50 # ms. Bidirectional, so total of 100ms\n
          \ loss: 0.02 # percent\n\n```\n\n##### Example 4\n\n\nIn this example, a
          combination of latency, packet loss, bandwidth, and link flapping is applied.\n\n```\n\napiVersion:
          apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n  name: all-impairments\nspec:\n
          \ duration: 480 # seconds\n  startDelay: 5 # seconds\n  interfaces:\n  -
          \"ens2f0\"\n  egress:\n    latency: 50 # ms. Bidirectional, so total of
          100ms\n    loss: 0.02 # percent\n    bandwidth: 1000 # 1000 kbit/s, about
          1 mbit/s\n  ingress:\n    latency: 50 # ms. Bidirectional, so total of 100ms\n
          \   loss: 0.02 # percent\n    bandwidth: 1000 # 1000 kbit/s, about 1 mbit/s\n
          \ linkFlapping:\n    enable: true\n    downTime: 30 # Seconds\n    upTime:
          120 # Seconds\n\n```\n\n##### Example 5\n\n\nIn this example, packet duplicaiton
          is applied.\n\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n
          \ name: duplication\nspec:\n  duration: 60\n  startDelay: 5\n  interfaces:\n
          \ - \"ens3\"\n  ingress:\n    duplication: 25 # percent\n  egress:\n    duplication:
          25 # percent\n\n```\n\n## Setup\n\n### Requirements\n\n1. You need `make`
          installed.\n2. You need access to the kubernetes cluster with a kubeconfig.\n\n###
          Installation\n\nTo run using the current latest image:\n1. Clone the repository.\n2.
          Run `make deploy` with KUBECONFIG environment variable set to the location
          of the kubeconfig.\n\nTo run with your own code, there are more steps.\n1.
          Fork the repository.\n2. Clone to a machine that has access to the Kubernetes
          cluster and the kubeconfig.\n3. Modify the makefile to change the `IMG`
          variable to your image repository. If you do not have podman installed,
          also change podman to docker.\n4. Run `make docker-build` then `make docker-push`.\n5.
          Then run `make deploy`\n\n### Deploying from operator-hub\n\n*Important*
          You must manually apply the serviceaccount!\n\nFirst, Follow the instructions
          on this site to deploy from here.\n\nNext, apply the manifest \"serviceaccount.yaml\"
          from the manifests folder.\n\nThen you can apply your ClusterImpairment
          resources.\n\n## Running impairments\n\nFirst, configure a ClusterImpairment
          type's spec with the information for the impairment you want to run. Most
          importantly, set the interface(s). You can list the interfaces with `ifconfig`.
          If an invalid interface is set, the pod's logs will show `ifconfig` output.\n\n\nOnce
          the clusterimpairment type is set, apply it and it will work.\n\n\nTo re-run
          an impairment, delete the impairment with `oc delete -f path/to/manifest.yaml`,
          then apply it again.\n\n\n## Limitations\n\n\n### Multiple Impairments\n\n\nYou
          should avoid any configurations that apply impairments to the same interface
          on the same node. There are potential conflicts. The worker pod will attempt
          to remove all impairments on that interface before applying new ones.\n\n\nInstead,
          take advantage of the full control of both ingress and egress impairments
          from within the same ClusterImpairment resource.\n\n\nThere is currently
          a configured limit of 20 ClusterImpairment resources working at the same
          time. If that is a limitation for your use case, you can change it in `config/manager/manager.yaml`
          or you can try to find a way to make each ClusterImpairment do more.\n\n###
          Traffic Control (TC)\n\nTraffic control is how cluster-impairment-operator
          applies the latency, bandwidth, and packet loss impairments. The limitation
          is due to Linux not being realtime operating system, so the impairment will
          not be perfectly consistent.\n\n### Link Flapping\n\nWhen link flapping,
          if you flap the link that Kubernetes uses to communicate with the pods,
          you may be unable to remove the pod until the link is on long enough for
          Kubernetes to communicate with the impaired node.\n\n\nIn this case, it
          is helpful to set the duration properly instead of running for an indefinite
          or large amount of time, because the node will properly go back to the unimpaired
          state at that time.\n\n\nIf the cluster becomes offline due to the link
          flapping when you do not want it to be offline, soft restarting the nodes
          after removing the custom resource should remove all impairments.\n"
        displayName: cluster-impairment-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network
        - impairment
        - kubernetes
        - openshift
        - tc
        - trafficcontrol
        links:
        - name: Cluster Impairment Operator
          url: https://github.com/redhat-performance/cluster-impairment-operator
        maintainers:
        - email: joconnel@redhat.com
          name: Jared O'Connell
        - email: akrzos@redhat.com
          name: Alex Krzos
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat Performance
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/redhat-performance/cluster-impairment-operator:1.0.4
        version: 1.0.4
      entries:
      - name: cluster-impairment-operator.v1.0.4
        version: 1.0.4
      - name: cluster-impairment-operator.v1.0.3
        version: 1.0.3
      - name: cluster-impairment-operator.v1.0.2
        version: 1.0.2
      - name: cluster-impairment-operator.v1.0.1
        version: 1.0.1
      name: beta
    defaultChannel: beta
    packageName: cluster-impairment-operator
    provider:
      name: Red Hat Performance
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: mercury-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mercury-operator.1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:1.0.1",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CustomerCreditRating",
                  "expose": ["http"]
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:1.0.1
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:1.0.1
        version: 1.0.1
      entries:
      - name: mercury-operator.1.0.1
        version: 1.0.1
      name: 1.0.x
    - currentCSV: mercury-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_CREDIT_RATING",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-offer",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-offer-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_OFFER",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-party-routing-profile",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/party-routing-profile-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "PARTY_ROUTING_PROFILE",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-example-sdi": {
                "description": "Example Service Domain Infrastructure"
            },
            {
             "example-customer-credit-rating": {
               "description": "Example Customer Credit Rating Service Domain"
            },
            {
             "example-customer-offer": {
               "description": "Example Customer Offer Service Domain"
            },
            {
             "example-party-routing-profile": {
               "description": "Example Party Routing Profile Service Domain"
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:v1.0.2
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - description: Represents a Kafka cluster
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Integration is the Schema for the Camel K integrations API
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:v1.0.2
        version: 1.0.2
      entries:
      - name: mercury-operator.v1.0.2
        version: 1.0.2
      name: dev
    - currentCSV: mercury-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_CREDIT_RATING",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-offer",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-offer-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_OFFER",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-party-routing-profile",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/party-routing-profile-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "PARTY_ROUTING_PROFILE",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-example-sdi": {
                "description": "Example Service Domain Infrastructure"
            },
            {
             "example-customer-credit-rating": {
               "description": "Example Customer Credit Rating Service Domain"
            },
            {
             "example-customer-offer": {
               "description": "Example Customer Offer Service Domain"
            },
            {
             "example-party-routing-profile": {
               "description": "Example Party Routing Profile Service Domain"
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:v1.0.2
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - description: Represents a Kafka cluster
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Integration is the Schema for the Camel K integrations API
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:v1.0.2
        version: 1.0.2
      entries:
      - name: mercury-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: 1.0.x
    packageName: mercury-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-ati-updates
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-ati-updates.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ati.loadcore.keysight.com/v1alpha1",
                "kind": "Updates",
                "metadata": {
                  "name": "updates-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "installWait": {
                    "enabled": true,
                    "image": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:40:21Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Updates
            name: updates.ati.loadcore.keysight.com
            version: v1alpha1
        description: Keysight ATI Updates Operator
        displayName: Keysight ATI Updates Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Ati Updates
          url: https://keysight-ati-updates.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-ati-updates@sha256:abda76eb50d77e5a95191c9dbb21eefe64af5b794511ceafdfda007c15449c6e
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        version: 0.1.21
      entries:
      - name: keysight-ati-updates.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-ati-updates
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: CyberArk
      provider-url: https://www.cyberark.com/
    name: vcp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vcp-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "installer.venafi.com/v1alpha1",
                "kind": "VenafiInstall",
                "metadata": {
                  "name": "install-venafi-enhanced-issuer"
                },
                "spec": {
                  "certManager": {
                    "install": true
                  },
                  "globals": {
                    "imagePullSecretNames": [
                      "venafi"
                    ]
                  },
                  "venafiConnection": {
                    "install": true
                  },
                  "venafiEnhancedIssuer": {
                    "install": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:9da79fd0941c21e433f22aaaccc43ebbd4646e685aea0050194954d8b4019c05
          createdAt: "2024-05-30T11:35:41Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              VenafiInstall is an object for installing Venafi components onto a Kubernetes
              cluster
            displayName: VenafiInstall
            kind: VenafiInstall
            name: venafiinstalls.installer.venafi.com
            version: v1alpha1
        description: |
          Venafi Control Plane Operator for Red Hat OpenShift simplifies deployment of Venafi's enterprise components and dependencies.

          The enterprise components include:
          - Enterprise cert-manager: An enhanced version of cert-manager designed for application certificate management within your clusters.
          - Venafi Enhanced Issuer: Enables seamless certificate enrollment for your clusters directly from the Venafi Control Plane.
          - Venafi Firefly: A high-performance, lightweight issuer for certificates, centrally managed by the Venafi Control Plane.
          - Venafi Kubernetes Agent: Provides real-time visibility into certificates and Kubernetes resources directly within the Venafi Control Plane.
        displayName: Venafi Control Plane Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - venafi
        - vcp-operator
        links:
        - name: Documentation
          url: https://docs.venafi.cloud/
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Venafi
          url: https://venafi.com
        relatedImages:
        - registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:9da79fd0941c21e433f22aaaccc43ebbd4646e685aea0050194954d8b4019c05
        version: 1.0.0
      entries:
      - name: vcp-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: vcp-operator.v1.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "installer.venafi.com/v1alpha1",
                "kind": "VenafiInstall",
                "metadata": {
                  "name": "install-venafi-enhanced-issuer"
                },
                "spec": {
                  "certManager": {
                    "install": true
                  },
                  "globals": {
                    "imagePullSecretNames": [
                      "venafi"
                    ]
                  },
                  "venafiConnection": {
                    "install": true
                  },
                  "venafiEnhancedIssuer": {
                    "install": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:eb2b06d191dbfbb02332375a7bca6bc010d5410e53b3bc5fb8c320101292408c
          createdAt: "2025-07-22T06:44:30Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              VenafiInstall is an object for installing Venafi components onto a Kubernetes
              cluster
            displayName: VenafiInstall
            kind: VenafiInstall
            name: venafiinstalls.installer.venafi.com
            version: v1alpha1
        description: |
          The operator, formerly known as the Venafi Control Plane Operator, simplifies the deployment of enterprise components part of the CyberArk Certificate Manager for Kubernetes.

          The components include:
          - cert-manager: An enterprise distribution of cert-manager designed for application certificate management within your clusters.
          - Enterprise Issuer: Enables seamless certificate enrollment directly from the Certificate Manager.
          - Workload Identity Manager: A high-performance issuer for certificates, centrally managed by the Certificate Manager.
          - Agent: Provides real-time visibility into cluster resources directly within the Certificate Manager.
        displayName: CyberArk Certificate Manager Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cyberark
        - vcp-operator
        links:
        - name: Documentation
          url: https://docs.venafi.cloud/vaas/k8s-components/c-vcpo-overview/
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:eb2b06d191dbfbb02332375a7bca6bc010d5410e53b3bc5fb8c320101292408c
        version: 1.8.1
      entries:
      - name: vcp-operator.v1.8.1
        version: 1.8.1
      - name: vcp-operator.v1.8.0
        version: 1.8.0
      - name: vcp-operator.v1.7.0
        version: 1.7.0
      - name: vcp-operator.v1.6.0
        version: 1.6.0
      - name: vcp-operator.v1.5.0
        version: 1.5.0
      - name: vcp-operator.v1.4.0
        version: 1.4.0
      - name: vcp-operator.v1.3.0
        version: 1.3.0
      - name: vcp-operator.v1.2.1
        version: 1.2.1
      - name: vcp-operator.v1.2.0
        version: 1.2.0
      - name: vcp-operator.v1.1.1
        version: 1.1.1
      name: stable
    defaultChannel: stable
    packageName: vcp-operator
    provider:
      name: CyberArk
      url: https://www.cyberark.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE Ezmeral
      provider-url: ""
    name: hpe-ezmeral-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-ezmeral-csi-operator.v1.0.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ezmeral.hpe.com/v1",
                "kind": "HPEEzmeralCSIDriver",
                "metadata": {
                  "name": "hpeezmeralcsidriver"
                },
                "spec": {
                  "controllerImage": "registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03",
                  "nodeImage": "registry.connect.redhat.com/maprtech/csi-kdfplugin@sha256:ffd273ee2e608d2ef64fd3b2bcbb02d5f0a6c173ed96a15eaf5549ea45fea8d1",
                  "pullPolicy": "Always"
                }
              },
              {
                "apiVersion": "ezmeral.hpe.com/v1",
                "kind": "HPEEzmeralNFSCSIDriver",
                "metadata": {
                  "name": "hpeezmeralnfscsidriver"
                },
                "spec": {
                  "controllerImage": "registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03",
                  "nodeImage": "registry.connect.redhat.com/maprtech/csi-nfsplugin@sha256:a5c954f62f1395bafe2d68454ac604bca4be97573a3482fcbd8e76b6361cc08a",
                  "pullPolicy": "Always"
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator@sha256:c9c243c51cd760c2c2a66b489807a5e22053fcbeb94fa29afd10d9c91e2d3908
          createdAt: "2025-06-25T11:51:05Z"
          description: A Container Storage Interface (CSI) driver for HPE Ezmeral
            Data Fabric platform. The CSI driver allows you to use HPE Ezmeral Data
            Fabric with your preferred container orchestrator.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mapr/mapr-csi
          support: HPE Ezmeral Data Fabric
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE Ezmeral CSI Driver (FUSE)
            displayName: HPEEzmeralCSIDriver
            kind: HPEEzmeralCSIDriver
            name: hpeezmeralcsidrivers.ezmeral.hpe.com
            version: v1
          - description: HPE Ezmeral CSI Driver (NFS)
            displayName: HPEEzmeralNFSCSIDriver
            kind: HPEEzmeralNFSCSIDriver
            name: hpeezmeralnfscsidrivers.ezmeral.hpe.com
            version: v1
        description: |
          The HPE Ezmeral CSI Operator for Kubernetes packages, deploys, manages, upgrades the HPE Ezmeral CSI Driver on Kubernetes and OpenShift for dynamic provisioning of persistent volumes on HPE Ezmeral Data Fabric platform.
          The HPE Ezmeral CSI Driver for Kubernetes leverages HPE Ezmeral Data Fabric platform to provide scalable and persistent storage for stateful applications. Please refer to our CSI driver [documentation](https://docs.datafabric.hpe.com/home/CSIdriver/csi_overview.html) for full list of supported CSI features.
          ## Installation
          Refer to the HPE Ezmeral CSI Operator for Kubernetes [documentation](https://docs.datafabric.hpe.com/home/CSIdriver/csi_installing_operator.html)
        displayName: HPE Ezmeral Data Fabric CSI Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE Ezmeral
        - HPE Ezmeral Data Fabric
        - Storage
        - FileSystem
        - BlockVolume
        links:
        - name: Documentation
          url: https://docs.datafabric.hpe.com/home/CSIdriver/csi_overview.html
        maintainers:
        - email: ezmeral-csi-support@hpe.com
          name: HPE Ezmeral Data Fabric
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: HPE Ezmeral
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6ba961c2c2a29750c0132fe6dd6fa9f6001010afbc5f19b98add87b31b54bcf6
        - registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator@sha256:c9c243c51cd760c2c2a66b489807a5e22053fcbeb94fa29afd10d9c91e2d3908
        - registry.connect.redhat.com/maprtech/csi-nfsplugin@sha256:a5c954f62f1395bafe2d68454ac604bca4be97573a3482fcbd8e76b6361cc08a
        - registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03
        - registry.connect.redhat.com/maprtech/csi-kdfplugin@sha256:ffd273ee2e608d2ef64fd3b2bcbb02d5f0a6c173ed96a15eaf5549ea45fea8d1
        version: 1.0.18
      entries:
      - name: hpe-ezmeral-csi-operator.v1.0.18
        version: 1.0.18
      name: stable
    defaultChannel: stable
    packageName: hpe-ezmeral-csi-operator
    provider:
      name: HPE Ezmeral
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzo-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzo-operator.v3.1.200
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzo.cambridgesemantics.com/v2",
                "kind": "Anzo",
                "metadata": {
                  "name": "no-persistence"
                },
                "spec": {
                  "jsonActivation": true,
                  "jvmMemory": "4096m",
                  "nodeConfig": {
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "app": "anzo"
                        }
                      },
                      "serviceName": "anzo-no-persistence",
                      "template": {
                        "metadata": {
                          "annotations": {
                            "cluster-autoscaler.kubernetes.io/safe-to-evict": "false"
                          },
                          "labels": {
                            "app": "anzo"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "registry.connect.redhat.com/cambridgesemantics/anzo@sha256:5b192604338f602fdabe55fecefd9eb82400a77c5aa7f04bd8e54aead8644c6d",
                              "imagePullPolicy": "Always",
                              "name": "anzo",
                              "resources": {
                                "limits": {
                                  "cpu": "2.0",
                                  "memory": "8.0Gi"
                                },
                                "requests": {
                                  "cpu": "2.0",
                                  "memory": "8.0Gi"
                                }
                              }
                            }
                          ],
                          "initContainers": [
                            {
                              "name": "anzo-init",
                              "resources": {
                                "limits": {
                                  "cpu": "0.5",
                                  "memory": "2Gi"
                                },
                                "requests": {
                                  "cpu": "0.5",
                                  "memory": "2Gi"
                                }
                              }
                            }
                          ],
                          "securityContext": {
                            "fsGroup": 1001
                          },
                          "serviceAccountName": "anzo-operator"
                        }
                      }
                    }
                  },
                  "role": "AnzoServer"
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/anzo-operator@sha256:4c0c8bf7acbfbb96f15459a20364608e5d0a0933c6326a89d219666ee0d88ca2
          createdAt: "2025-07-09T06:52:12Z"
          description: kubernetes operator for Altair Graph Studio
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Altair Customer Support
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoBackup is the Schema for the anzobackups API
            displayName: Altair Graph Studio Backup
            kind: AnzoBackup
            name: anzobackups.anzobackup.cambridgesemantics.com
            version: v1beta1
          - description: Altair Graph Studio is the Schema for the anzos API
            displayName: Altair Graph Studio Operator
            kind: Anzo
            name: anzos.anzo.cambridgesemantics.com
            version: v2
        description: |-
          The Altair Graph Studio Operator provides the way to install and configure an altair graph studio server and agent setup on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v3.1.2/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/deploy-k8s.htm )

          ### Support

          We offer Support to our customers [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Studio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        links:
        - name: Altair Graph Studio Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v3.1.2/README_openshift.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzo-operator@sha256:4c0c8bf7acbfbb96f15459a20364608e5d0a0933c6326a89d219666ee0d88ca2
        - registry.connect.redhat.com/cambridgesemantics/anzo@sha256:5b192604338f602fdabe55fecefd9eb82400a77c5aa7f04bd8e54aead8644c6d
        version: 3.1.200
      entries:
      - name: anzo-operator.v3.1.200
        version: 3.1.200
      - name: anzo-operator.v3.1.106
        version: 3.1.106
      - name: anzo-operator.v3.1.105
        version: 3.1.105
      name: stable
    defaultChannel: stable
    packageName: anzo-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq7-interconnect-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: 1.10.x
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: 1.2.0
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: latest
    defaultChannel: 1.10.x
    packageName: amq7-interconnect-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-registry
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-operator.v1.1.3-v2.6.4.final
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-mem"
                },
                "spec": {
                  "configuration": {
                    "persistence": "mem"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-operator:1.1.3
          createdAt: "2024-09-17"
          description: Deploy and manage Apicurio Registry on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents an Apicurio Registry instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Apicurio Registry

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Provides a REST API and web UI to manage the artifacts and artifact meta-data
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, in-memory
          - Compatible with Confluent and IBM APIs
          - Runs on a lightweight Quarkus platform
          - Includes Maven plugin to integrate with Maven based builds

          ## Apicurio Registry Operator

          Provides a quick and easy way to deploy and manage Apicurio Registry on Kubernetes.

          **Features**
          - Supports installation and configuration of the Registry for all storage options
          - Easily perform a rolling upgrade of Apicurio Registry

          ## Prerequisites

          This operator does not deploy storage for Apicurio Registry. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/main/LICENSE)
        displayName: Apicurio Registry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-operator:1.1.3
        version: 1.1.3-v2.6.4.final
      entries:
      - name: apicurio-registry-operator.v1.1.3-v2.6.4.final
        version: 1.1.3-v2.6.4.final
      - name: apicurio-registry-operator.v1.1.2-v2.5.11.final
        version: 1.1.2-v2.5.11.final
      - name: apicurio-registry-operator.v1.1.1-v2.5.9.final
        version: 1.1.1-v2.5.9.final
      - name: apicurio-registry-operator.v1.1.0-v2.4.12.final
        version: 1.1.0-v2.4.12.final
      - name: apicurio-registry-operator.v1.0.0-v2.0.0.final
        version: 1.0.0-v2.0.0.final
      name: 2.x
    defaultChannel: 2.x
    packageName: apicurio-registry
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zettaset
      provider-url: ""
    name: xcrypt-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zts-xcrypt-operator.v0.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xcrypt.zettaset.com/v1alpha1",
                "kind": "XCrypt",
                "metadata": {
                  "name": "zts-masterset"
                },
                "spec": {
                  "status": {
                    "podNames": "zts-ca",
                    "replicas": "1,"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.connect.redhat.com/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
          createdAt: "2021-06-16T12:29:13.97Z"
          description: XCrypt Operator deploys Zettaset XCrypt Container Encryption  -
            the only software-defined solution that offers transparent high performance
            data-at-rest encryption of critical data in container environments.
          support: Zettaset
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XCrypt is the Schema for the xcrypts API
            kind: XCrypt
            name: xcrypts.xcrypt.zettaset.com
            version: v1alpha1
        description: "Zettaset’s RedHat Certified XCrypt Operator allows users to
          deploy multiple XCrypt components on OpenShift with just a few simple commands.
          The XCrypt Operator deployment provides software-defined encryption that
          transparently protects container data throughout your OpenShift environment.\n\n##
          Core Features\n* **Automated XCrypt provisioning** - Deploying a Zettaset
          XCrypt has never been easier. Adjust cluster specific configuration and
          let the Zettaset Operator take care of provisioning XCrypt services and
          setting up the XCrypt encryption to your exact specification. \n* **CSI
          Driver deployment**: Zettaset XCrypt Operator automatically deploys Zettaset
          XCrypt CSI Driver in order to provide automated and transparent data-at-rest
          \ encryption functionality compatible with Openshift and any other container
          orchestrator. Once the Operator is provisioned, the driver runs fine tuned
          and ready to work with other XCrypt components.\n* **Multi-node topology**:
          Encrypt as many nodes and volumes as you need to. XCrypt components can
          be installed on multiple nodes across the cluster. This is defined by two
          labels: zts-master and zts-worker. Normally zts-master node is a single
          node that runs Zettaset XCrypt major services required for the product operation.
          Multiple nodes can be marked as zts-worker and those are the nodes that
          can have encrypted volumes.\n* **Seamless Encrypted Volume mount**: Volumes
          are not only being encrypted, but also are mounted/unmounted to containers
          with Persistent Volume Claims. This happens as seamlessly as regular volumes
          provision and mount/unmount.\n* **Robust failover**: A node failure doesn’t
          affect XCrypt operations on other nodes of the cluster. Encryption on a
          restarted/recovered node gets back to normal functioning if the node stays
          healthy and data hasn’t been corrupted during the restart\n\n## About Zettaset
          XCrypt Kubernetes Encryption for OpenShift\nXCrypt Kubernetes Encryption
          for OpenShift is a software-only solution that enables encryption of data
          at rest stored in Red Hat OpenShift container environments without slowing
          down the DevSecOps process. In addition to providing a transparent and high-performance
          layer of security that flexibly protects container data across on-premise,
          in hybrid, and multi-cloud environments, it also helps your organization
          efficiently transition from DevOps to DevSecOps.  \n\nMore information about
          Zettaset XCrypt Kubernetes Encryption for OpenShift can be found in the
          [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)\n\n##
          RedHat OpenShift + Zettaset: The next generation of secure enterprise infrastructure
          for DevSecOps  \nBy combining RedHat’s Container Platform with Zettaset’s
          certified software-defined encryption, you can create the next generation
          of secure enterprise infrastructure that DevSecOps requires. Together they
          create the stable, consistent, and supported base that your organization
          needs to develop applications faster, simplify updates, drive innovation,
          and reduce the risk of potential security breaches and data theft.\n###
          Core Features\n* Software-only for simple and scalable deployments \n* Negligible
          performance impact\n* Transparent to developers and administrators \n* Works
          with AWS, Google, and Azure \n* Unique encryption key per container volume
          \n* Encrypted volumes are only available when in use\n* Direct integration
          with OpenShift \n* Automated encryption policy management Secure erase of
          volumes\n* Ability to securely decommission a node\n\n## How does XCrypt
          Kubernetes Encryption fit into OpenShift? \nXCrypt XCrypt Kubernetes Encryption
          for OpenShift Encryption makes it incredibly easy to protect container data
          in your OpenShift environment. \n### Use cases include:\n* Transitioning
          from DevOps to DevSecOps \n* Ensuring data protection in OpenShift container
          environments with negligible performance impact \n* Achieving compliance
          in regulated industries like healthcare and finance\n\n## Other Information\nRole-based
          access control needs to be configured prior to the Operator installation.
          Please find details in the [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)
          \   "
        displayName: Zettaset XCrypt Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - encryption
        - storage
        links:
        - name: Zettaset
          url: https://www.zettaset.com/
        maintainers:
        - email: info@zettaset.com
          name: Zettaset
        maturity: alpha
        provider:
          name: Zettaset
        relatedImages:
        - registry.connect.redhat.com/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
        - registry.connect.redhat.com/zts/xcrypt-ca-1@sha256:9f1015dfb459b69f5dbf449efa0517452095a6cf4f20826ebae278f9f800ffe7
        - registry.connect.redhat.com/zts/xcrypt-kmip-server-1@sha256:9af5d21dd748dfbf4e107f27a5f10fa3c0ba928ade9b989a744b620ce34ec757
        - registry.connect.redhat.com/zts/xcrypt-license-server-1@sha256:bc5486892a41e70f5dc4c4e1d82fca8db2149598fb8b5950ba5da1ee2d9d245d
        - registry.connect.redhat.com/zts/xcrypt-host-manager-1@sha256:9faa4b597eb74df412761dffa057b64fb69dbf9fa016414962eb2099a343e734
        - registry.connect.redhat.com/zts/zts-csi-driver-1@sha256:c15fa9a7cc40891103f426206663dcebe1210c8ee3ab6037cbf64e1f86b9c7e5
        version: 0.0.17
      entries:
      - name: zts-xcrypt-operator.v0.0.17
        version: 0.0.17
      name: alpha
    defaultChannel: alpha
    packageName: xcrypt-operator
    provider:
      name: Zettaset
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-spark-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spark-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:a6c82a6d6ebaa0d2572d644ac9013451e020f134c6fe46222fef37b1e9aa3cbb
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:a6c82a6d6ebaa0d2572d644ac9013451e020f134c6fe46222fef37b1e9aa3cbb
        version: 24.11.1
      entries:
      - name: spark-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: spark-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:8ad9948bf8264924a0eedd92c2e58269117487b9d57816384447bb6c999be681
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:8ad9948bf8264924a0eedd92c2e58269117487b9d57816384447bb6c999be681
        version: 25.3.0
      entries:
      - name: spark-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: spark-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkConnectServerSpec via
              `CustomResource`
            displayName: sparkconnectservers.spark.stackable.tech
            kind: SparkConnectServer
            name: sparkconnectservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
        version: 25.7.0
      entries:
      - name: spark-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: spark-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkConnectServerSpec via
              `CustomResource`
            displayName: sparkconnectservers.spark.stackable.tech
            kind: SparkConnectServer
            name: sparkconnectservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
        version: 25.7.0
      entries:
      - name: spark-operator.v25.7.0
        version: 25.7.0
      - name: spark-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-spark-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://skupper.io
    name: skupper-netobs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-netobs-operator.v2.1.1-rh-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability.skupper.io/v2alpha1",
                "kind": "NetworkObserver",
                "metadata": {
                  "name": "networkobserver-sample"
                },
                "spec": {
                  "auth": {
                    "basic": {
                      "create": true,
                      "secretName": ""
                    },
                    "openshift": {
                      "cookieSecretName": "",
                      "createCookieSecret": true,
                      "serviceAccount": {
                        "create": true,
                        "nameOverride": ""
                      }
                    },
                    "strategy": "openshift"
                  },
                  "extraArgs": null,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "registry.redhat.io/service-interconnect/skupper-network-observer-rhel9",
                    "tag": "2.1.1"
                  },
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nginx": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/ubi9/nginx-124",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": 9.6
                  },
                  "openshiftOauthProxy": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-oauth-proxy-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.19.0"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "prometheus": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-prometheus-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.16.0"
                  },
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "router": {
                    "certificate": {
                      "create": true,
                      "nameOverride": ""
                    },
                    "endpoint": "amqps://skupper-router-local"
                  },
                  "securityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    }
                  },
                  "service": {
                    "port": 443,
                    "type": "ClusterIP"
                  },
                  "skipManagementLabels": false,
                  "tls": {
                    "openshiftIssued": true,
                    "secretName": "",
                    "skupperIssued": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
          createdAt: "2025-08-27T20:21:18Z"
          description: Red Hat Service Interconnect Network Observer Operator provides
            the ability to monitor and observe a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkObserver
            name: networkobservers.observability.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect Network Observer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/ubi9/nginx-124@sha256:cbcca4743015b4ef167383a080f4711a6049e275e0b66f80a74f8b087b0dfc7f
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:d3056b35d9a205b9f2c48d924f199c5ac23904eb18d526e4bff229e7c7181415
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
        version: 2.1.1-rh-8
      entries:
      - name: skupper-netobs-operator.v2.1.1-rh-8
        version: 2.1.1-rh-8
      name: stable-2
    - currentCSV: skupper-netobs-operator.v2.1.1-rh-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability.skupper.io/v2alpha1",
                "kind": "NetworkObserver",
                "metadata": {
                  "name": "networkobserver-sample"
                },
                "spec": {
                  "auth": {
                    "basic": {
                      "create": true,
                      "secretName": ""
                    },
                    "openshift": {
                      "cookieSecretName": "",
                      "createCookieSecret": true,
                      "serviceAccount": {
                        "create": true,
                        "nameOverride": ""
                      }
                    },
                    "strategy": "openshift"
                  },
                  "extraArgs": null,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "registry.redhat.io/service-interconnect/skupper-network-observer-rhel9",
                    "tag": "2.1.1"
                  },
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nginx": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/ubi9/nginx-124",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": 9.6
                  },
                  "openshiftOauthProxy": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-oauth-proxy-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.19.0"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "prometheus": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-prometheus-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.16.0"
                  },
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "router": {
                    "certificate": {
                      "create": true,
                      "nameOverride": ""
                    },
                    "endpoint": "amqps://skupper-router-local"
                  },
                  "securityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    }
                  },
                  "service": {
                    "port": 443,
                    "type": "ClusterIP"
                  },
                  "skipManagementLabels": false,
                  "tls": {
                    "openshiftIssued": true,
                    "secretName": "",
                    "skupperIssued": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
          createdAt: "2025-08-27T20:21:18Z"
          description: Red Hat Service Interconnect Network Observer Operator provides
            the ability to monitor and observe a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkObserver
            name: networkobservers.observability.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect Network Observer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:d3056b35d9a205b9f2c48d924f199c5ac23904eb18d526e4bff229e7c7181415
        - registry.redhat.io/ubi9/nginx-124@sha256:cbcca4743015b4ef167383a080f4711a6049e275e0b66f80a74f8b087b0dfc7f
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        version: 2.1.1-rh-8
      entries:
      - name: skupper-netobs-operator.v2.1.1-rh-8
        version: 2.1.1-rh-8
      name: stable-2.1
    defaultChannel: stable-2
    packageName: skupper-netobs-operator
    provider:
      name: Red Hat
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/trustification/trusted-profile-analyzer-operator
    name: rhtpa-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhtpa-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtpa.io/v1",
                "kind": "TrustedProfileAnalyzer",
                "metadata": {
                  "name": "trustedprofileanalyzer-sample"
                },
                "spec": {
                  "appDomain": "change-me",
                  "collector": {},
                  "database": {},
                  "image": {
                    "fullName": "registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e",
                    "pullPolicy": "IfNotPresent"
                  },
                  "infrastructure": {
                    "port": 9010
                  },
                  "ingress": {},
                  "metrics": {
                    "enabled": false
                  },
                  "modules": {
                    "createDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "createImporters": {
                      "enabled": false,
                      "image": {},
                      "importers": {}
                    },
                    "importer": {
                      "concurrency": 1,
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {},
                      "workingDirectory": {
                        "size": "32Gi"
                      }
                    },
                    "migrateDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "server": {
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "ingress": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {}
                    }
                  },
                  "oidc": {
                    "clients": {
                      "cli": {
                        "clientSecret": null
                      },
                      "frontend": {}
                    }
                  },
                  "openshift": {
                    "useServiceCa": true
                  },
                  "partOf": "trustify",
                  "replicas": 1,
                  "rust": {},
                  "storage": {},
                  "tls": {},
                  "tracing": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-09-19T15:04:16Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Profile Analyzer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trustification/trusted-profile-analyzer-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: TrustedProfileAnalyzer
            name: trustedprofileanalyzers.rhtpa.io
            version: v1
        description: |
          [Tech Preview]
          Red Hat Trusted Profile Analyzer (RHTPA) provides a single source of truth for your software inventory across all your environments. You can verify the integrity of your in-house and third-party applications against their respective Software Bill of Materials (SBOM) document.  Trusted Profile Analyzer provides vulnerability, component and license transparency, strengthens your software supply chain security, and enables an accurate, up-to-date understanding of your software inventory.

          Features and Benefits:

          Enhanced Software Supply Chain Security: Offers critical transparency and integrity verification across the entire software lifecycle.

          Reduced Operational Risk: Proactively identifies and helps remediate vulnerabilities minimizing risk and compliance violations.

          Improved Compliance and Auditability: Simplifies the process of meeting regulatory requirements for software transparency and integrity by providing verifiable documentation and continuous monitoring.

          Faster Remediation and Incident Response: Provides targeted, actionable insights to prioritize and accelerate the patching and remediation of vulnerabilities.

          Increased Developer Productivity & Trust: Integrates security into existing workflows without significant overhead, allowing developers to build and deploy with greater confidence in the integrity of their code and components.

          Centralized Visibility and Control: Consolidates disparate security information into a single pane of glass, offering a holistic view of your application security posture.

          Red Hat Trusted Profile Analyzer is a crucial tool for any organization running applications on OpenShift that needs to confidently understand, verify, and secure the software they deploy, especially in an era of complex software supply chains and increasing regulatory scrutiny.

          [Documentation](https://docs.redhat.com/en/documentation/red_hat_trusted_profile_analyzer/2.1).
        displayName: Red Hat Trusted Profile Analyzer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rhtpa
        - trusted-profile-analyzer
        links:
        - name: Trusted Profile Analyzer Operator
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        maintainers:
        - email: rhtpa-support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        relatedImages:
        - registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
        - registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e
        version: 1.0.2
      entries:
      - name: rhtpa-operator.v1.0.2
        version: 1.0.2
      - name: rhtpa-operator.v1.0.1
        version: 1.0.1
      - name: rhtpa-operator.v1.0.0
        version: 1.0.0
      name: stable
    - currentCSV: rhtpa-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtpa.io/v1",
                "kind": "TrustedProfileAnalyzer",
                "metadata": {
                  "name": "trustedprofileanalyzer-sample"
                },
                "spec": {
                  "appDomain": "change-me",
                  "collector": {},
                  "database": {},
                  "image": {
                    "fullName": "registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e",
                    "pullPolicy": "IfNotPresent"
                  },
                  "infrastructure": {
                    "port": 9010
                  },
                  "ingress": {},
                  "metrics": {
                    "enabled": false
                  },
                  "modules": {
                    "createDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "createImporters": {
                      "enabled": false,
                      "image": {},
                      "importers": {}
                    },
                    "importer": {
                      "concurrency": 1,
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {},
                      "workingDirectory": {
                        "size": "32Gi"
                      }
                    },
                    "migrateDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "server": {
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "ingress": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {}
                    }
                  },
                  "oidc": {
                    "clients": {
                      "cli": {
                        "clientSecret": null
                      },
                      "frontend": {}
                    }
                  },
                  "openshift": {
                    "useServiceCa": true
                  },
                  "partOf": "trustify",
                  "replicas": 1,
                  "rust": {},
                  "storage": {},
                  "tls": {},
                  "tracing": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-09-19T15:04:16Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Profile Analyzer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trustification/trusted-profile-analyzer-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: TrustedProfileAnalyzer
            name: trustedprofileanalyzers.rhtpa.io
            version: v1
        description: |
          [Tech Preview]
          Red Hat Trusted Profile Analyzer (RHTPA) provides a single source of truth for your software inventory across all your environments. You can verify the integrity of your in-house and third-party applications against their respective Software Bill of Materials (SBOM) document.  Trusted Profile Analyzer provides vulnerability, component and license transparency, strengthens your software supply chain security, and enables an accurate, up-to-date understanding of your software inventory.

          Features and Benefits:

          Enhanced Software Supply Chain Security: Offers critical transparency and integrity verification across the entire software lifecycle.

          Reduced Operational Risk: Proactively identifies and helps remediate vulnerabilities minimizing risk and compliance violations.

          Improved Compliance and Auditability: Simplifies the process of meeting regulatory requirements for software transparency and integrity by providing verifiable documentation and continuous monitoring.

          Faster Remediation and Incident Response: Provides targeted, actionable insights to prioritize and accelerate the patching and remediation of vulnerabilities.

          Increased Developer Productivity & Trust: Integrates security into existing workflows without significant overhead, allowing developers to build and deploy with greater confidence in the integrity of their code and components.

          Centralized Visibility and Control: Consolidates disparate security information into a single pane of glass, offering a holistic view of your application security posture.

          Red Hat Trusted Profile Analyzer is a crucial tool for any organization running applications on OpenShift that needs to confidently understand, verify, and secure the software they deploy, especially in an era of complex software supply chains and increasing regulatory scrutiny.

          [Documentation](https://docs.redhat.com/en/documentation/red_hat_trusted_profile_analyzer/2.1).
        displayName: Red Hat Trusted Profile Analyzer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rhtpa
        - trusted-profile-analyzer
        links:
        - name: Trusted Profile Analyzer Operator
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        maintainers:
        - email: rhtpa-support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        relatedImages:
        - registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
        - registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e
        version: 1.0.2
      entries:
      - name: rhtpa-operator.v1.0.2
        version: 1.0.2
      - name: rhtpa-operator.v1.0.1
        version: 1.0.1
      - name: rhtpa-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable
    packageName: rhtpa-operator
    provider:
      name: Red Hat
      url: https://github.com/trustification/trusted-profile-analyzer-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: nova-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nova-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "Nova",
                "metadata": {
                  "name": "nova"
                },
                "spec": {
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaAPI",
                "metadata": {
                  "name": "novaapi-sample"
                },
                "spec": {
                  "apiDatabaseHostname": "openstack",
                  "cell0DatabaseHostname": "openstack",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "registeredCells": {},
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaCell",
                "metadata": {
                  "name": "novacell1-with-upcall-sample"
                },
                "spec": {
                  "apiDatabaseHostname": "openstack",
                  "cellDatabaseHostname": "openstack",
                  "cellName": "cell1",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaCompute",
                "metadata": {
                  "name": "nova-compute-ironic"
                },
                "spec": {
                  "novaComputeTemplates": {
                    "compute-ironic": {
                      "computeDriver": "ironic.IronicDriver",
                      "customServiceConfig": "",
                      "networkAttachments": [
                        "internalapi"
                      ],
                      "replicas": 1,
                      "resources": {}
                    }
                  }
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaConductor",
                "metadata": {
                  "name": "novaconductor-sample"
                },
                "spec": {
                  "cellName": "cell1",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaMetadata",
                "metadata": {
                  "name": "novametadata-sample"
                },
                "spec": {
                  "apiDatabaseAccount": "nova-api",
                  "apiDatabaseHostname": "openstack",
                  "cellDatabaseAccount": "nova-cell0",
                  "cellDatabaseHostname": "openstack",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "registeredCells": {},
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaNoVNCProxy",
                "metadata": {
                  "name": "novanovncproxy-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "nova.openstack.org/v1beta1",
                "kind": "NovaScheduler",
                "metadata": {
                  "name": "novascheduler-sample"
                },
                "spec": {
                  "apiDatabaseHostname": "openstack",
                  "cell0DatabaseHostname": "openstack",
                  "customServiceConfig": "[DEFAULT]\ndebug=true\n",
                  "keystoneAuthURL": "http://keystone-public-openstack.apps-crc.testing",
                  "memcachedInstance": "memcached",
                  "registeredCells": {},
                  "secret": "osp-secret",
                  "serviceAccount": "nova"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:21Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nova
            name: nova.nova.openstack.org
            version: v1beta1
          - description: NovaAPI is the Schema for the novaapis API
            displayName: Nova API
            kind: NovaAPI
            name: novaapis.nova.openstack.org
            version: v1beta1
          - description: NovaCell is the Schema for the novacells API
            displayName: Nova Cell
            kind: NovaCell
            name: novacells.nova.openstack.org
            version: v1beta1
          - description: NovaCompute is the Schema for the NovaCompute
            displayName: Nova Compute
            kind: NovaCompute
            name: novacomputes.nova.openstack.org
            version: v1beta1
          - description: NovaConductor is the Schema for the novaconductors API
            displayName: Nova Conductor
            kind: NovaConductor
            name: novaconductors.nova.openstack.org
            version: v1beta1
          - description: NovaMetadata is the Schema for the novametadata API
            displayName: Nova Metadata
            kind: NovaMetadata
            name: novametadata.nova.openstack.org
            version: v1beta1
          - description: NovaNoVNCProxy is the Schema for the novanovncproxies API
            displayName: Nova No VNCProxy
            kind: NovaNoVNCProxy
            name: novanovncproxies.nova.openstack.org
            version: v1beta1
          - description: NovaScheduler is the Schema for the novaschedulers API
            displayName: Nova Scheduler
            kind: NovaScheduler
            name: novaschedulers.nova.openstack.org
            version: v1beta1
        description: Nova Operator
        displayName: Nova Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Nova
        links:
        - name: Nova Operator
          url: https://github.com/openstack-k8s-operators/nova-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-nova-scheduler-rhel9@sha256:ba91196cd4f9b6e6f49868c91343b8eca7cc0d5016c36e83e2cc76f45cbe777b
        - registry.redhat.io/rhoso/openstack-nova-compute-rhel9@sha256:43cbe46ab263403fc303b6bc2388863029aad33465b237db18da7cd6d42078e1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/nova-rhel9-operator@sha256:f058c56df02a108505d51b9d19973894c2ee59055776c9f3fd7f7afe23a5106a
        - registry.redhat.io/rhoso/openstack-nova-api-rhel9@sha256:bbb8ebceb6f30b6c8beb120429e69f323bd231c09aaaddc5be9f15a527e928ed
        - registry.redhat.io/rhoso/openstack-nova-conductor-rhel9@sha256:500c36fd519611be8662e97ca31b1e59269c14d8aa94286dad9925e957462155
        - registry.redhat.io/rhoso/openstack-nova-novncproxy-rhel9@sha256:bf534f0c72f188d5f51e40b0392ace182db2055dff9ad14b01eeef13056febbf
        version: 1.0.6
      entries:
      - name: nova-operator.v1.0.6
        version: 1.0.6
      - name: nova-operator.v1.0.5
        version: 1.0.5
      - name: nova-operator.v1.0.4
        version: 1.0.4
      - name: nova-operator.v1.0.3
        version: 1.0.3
      - name: nova-operator.v1.0.2
        version: 1.0.2
      - name: nova-operator.v1.0.1
        version: 1.0.1
      - name: nova-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: nova-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Cloudera
      provider-url: https://cloudera.com
    name: cfm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cfm-operator.v2.9.0-b96
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "Nifi",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifi-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifi",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifi-sample"
                },
                "spec": {
                  "hostName": "nifi.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifi-k8s",
                    "tag": "2.8.0-b94-nifi_1.25.0.2.3.13.0-36"
                  },
                  "replicas": 1,
                  "stateManagement": {
                    "clusterProvider": {
                      "class": "org.apache.nifi.kubernetes.state.provider.KubernetesConfigMapStateProvider",
                      "id": "kubernetes-provider"
                    }
                  },
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.8.0-b94"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              },
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "NifiRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifiregistry-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifiregistry",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifiregistry-sample"
                },
                "spec": {
                  "hostName": "registry.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s",
                    "tag": "2.8.0-b94-nifi_1.25.0.2.3.13.0-36"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.8.0-b94"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning,Big Data
          com.redhat.openshift.versions: v4.10
          containerImage: container.repository.cloudera.com/cloudera/cfm-operator@sha256:890d81188143ccf9ebe2c2fe5e039cdff3a2df53f7512555673f0468147e991e
          createdAt: "2024-12-06T19:03:24Z"
          description: Cloudera Flow Management - Kubernetes Operator for Apache NiFideploys
            and operates Apache NiFi
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: A valid Cloudera license is required.
            Contact a Cloudera account team to learn more.
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NifiRegistry is the Schema for the nifiregistries API.
            displayName: Nifi Registry
            kind: NifiRegistry
            name: nifiregistries.cfm.cloudera.com
            version: v1alpha1
          - description: Nifi is the Schema for the nifis API.
            displayName: Nifi
            kind: Nifi
            name: nifis.cfm.cloudera.com
            version: v1alpha1
        description: |-
          Cloudera Flow Management - Kubernetes Operator for Apache NiFi brings enterprise-grade NiFi and NiFi Registry deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of NiFi and NiFi Registry for enterprise deployments.
          To learn more about Cloudera Flow Management - Kubernetes Operator for Apache NiFi read this
          [deep dive blog post](https://community.cloudera.com/t5/Community-Articles/Cloudera-Flow-Management-Operator-A-technical-deep-dive/ta-p/390473).

          Cloudera Flow Management - Kubernetes Operator for Apache NiFi consists of and ships multiple components including Apache NiFi, Apache NiFi Registry,
          and a custom sidecar image for logging aggregation.

          ## Installation prerequisites

          Installing Cloudera Flow Management - Kubernetes Operator for Apache NiFi requires the a valid license as well as Cloudera credentials
          (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly.
          The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) from which installation artifacts are pulled.

          ### Obtaining a license
          Contact [Cloudera sales](https://cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets
          Before installation, create two `Secrets`. One containing your Cloudera credentials and one containing your license. Cloudera credentials are provided to you
          together with your license.

          1. Create a `Secret` containing your Cloudera credentials.
              `oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***]
               --docker-server=container.repository.cloudera.com`

          2. Create a `Secret` containing your license.
              `oc create secret generic cfm-operator-license --from-file=license.txt=[***PATH TO LICENSE FILE***]`

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces in which you
            plan to deploy NiFi clusters or NiFi Registry deployments. Cloudera recommends that you create the `Secret` in all required namespaces before installation
            if you know which namespaces you will be using to deploy NiFi clusters or NiFi Registry deployments.

          ## Learn more
          To learn more, visit the [Cloudera Flow Management - Kubernetes Operator for Apache NiFi documentation library](https://docs.cloudera.com/cfm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Flow Management - Kubernetes Operator for Apache NiFi, visit the following resources:
          * [Apache NiFi](https://nifi.apache.org/)
          * [Apache NiFi Registry](https://nifi.apache.org/registry.html)
        displayName: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        - nifiregistry
        - cloudera
        - cfm
        links:
        - name: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
          url: https://docs.cloudera.com/cfm-operator
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s@sha256:914dcde9ee97894fa6f5deec285e53bf81d196e071a997cabb40fedd52deccc3
        - container.repository.cloudera.com/cloudera/cfm-tini@sha256:6493d23e20fca8daf31823403fca3dd9b42da983464fbe788a6863e141e8b66b
        - container.repository.cloudera.com/cloudera/cfm-operator@sha256:890d81188143ccf9ebe2c2fe5e039cdff3a2df53f7512555673f0468147e991e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86af06fb3b6c17b22be47118b0a8c797fa7fdd3c00dbe99a2d2a593ca8fc3d63
        - container.repository.cloudera.com/cloudera/cfm-nifi-k8s@sha256:0d700b7eff096dfc362868345826c3a6460fa43a700273b1422e95015b7727ee
        version: 2.9.0-b96
      entries:
      - name: cfm-operator.v2.9.0-b96
        version: 2.9.0-b96
      name: alpha
    - currentCSV: cfm-operator.v2.10.0-b135
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "Nifi",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifi-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifi",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifi-sample"
                },
                "spec": {
                  "hostName": "nifi.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifi-k8s",
                    "tag": "2.9.0-b96-nifi_1.27.0.2.3.14.0-14"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.9.0-b96"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              },
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "NifiRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifiregistry-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifiregistry",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifiregistry-sample"
                },
                "spec": {
                  "hostName": "registry.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s",
                    "tag": "2.9.0-b96-nifi_1.27.0.2.3.14.0-14"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.9.0-b96"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning,Big Data
          com.redhat.openshift.versions: v4.10
          containerImage: container.repository.cloudera.com/cloudera/cfm-operator@sha256:0312b6dadd5904ea9507bb857ab89d7266bf244c93b86855e0528e72813b7a85
          createdAt: "2025-05-05T14:39:51Z"
          description: Cloudera Flow Management - Kubernetes Operator for Apache NiFideploys
            and operates Apache NiFi
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: A valid Cloudera license is required.
            Contact a Cloudera account team to learn more.
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NifiRegistry is the Schema for the nifiregistries API.
            displayName: Nifi Registry
            kind: NifiRegistry
            name: nifiregistries.cfm.cloudera.com
            version: v1alpha1
          - description: Nifi is the Schema for the nifis API.
            displayName: Nifi
            kind: Nifi
            name: nifis.cfm.cloudera.com
            version: v1alpha1
        description: |-
          Cloudera Flow Management - Kubernetes Operator for Apache NiFi brings enterprise-grade NiFi and NiFi Registry deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of NiFi and NiFi Registry for enterprise deployments.
          To learn more about Cloudera Flow Management - Kubernetes Operator for Apache NiFi read this
          [deep dive blog post](https://community.cloudera.com/t5/Community-Articles/Cloudera-Flow-Management-Operator-A-technical-deep-dive/ta-p/390473).

          Cloudera Flow Management - Kubernetes Operator for Apache NiFi consists of and ships multiple components including Apache NiFi, Apache NiFi Registry,
          and a custom sidecar image for logging aggregation.

          ## Installation prerequisites

          Installing Cloudera Flow Management - Kubernetes Operator for Apache NiFi requires the a valid license as well as Cloudera credentials
          (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly.
          The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) from which installation artifacts are pulled.

          ### Obtaining a license
          Contact [Cloudera sales](https://cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets
          Before installation, create two `Secrets`. One containing your Cloudera credentials and one containing your license. Cloudera credentials are provided to you
          together with your license.

          1. Create a `Secret` containing your Cloudera credentials.
              `oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***]
               --docker-server=container.repository.cloudera.com --namespace [***OPERATOR INSTALL NAMESPACE***]`

          2. Create a `Secret` containing your license.
              `oc create secret generic cfm-operator-license --from-file=license.txt=[***PATH TO LICENSE FILE***] --namespace [***OPERATOR INSTALL NAMESPACE***]`

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces in which you
            plan to deploy NiFi clusters or NiFi Registry deployments. Cloudera recommends that you create the `Secret` in all required namespaces before installation
            if you know which namespaces you will be using to deploy NiFi clusters or NiFi Registry deployments.

          ## Learn more
          To learn more, visit the [Cloudera Flow Management - Kubernetes Operator for Apache NiFi documentation library](https://docs.cloudera.com/cfm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Flow Management - Kubernetes Operator for Apache NiFi, visit the following resources:
          * [Apache NiFi](https://nifi.apache.org/)
          * [Apache NiFi Registry](https://nifi.apache.org/registry.html)
        displayName: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        - nifiregistry
        - cloudera
        - cfm
        links:
        - name: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
          url: https://docs.cloudera.com/cfm-operator
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s@sha256:52e00f102579b972500eb187e64a7662a6a46caa883df743e454b1d5111b34e5
        - container.repository.cloudera.com/cloudera/cfm-tini@sha256:d42ad5f05a9bb283fed6dbbd820855714a31662f5b851e691f13703153e90d13
        - container.repository.cloudera.com/cloudera/cfm-nifi-k8s@sha256:8607c81906422d65784c6dca74a37af3b7c87012bca65ce28a995366014aec86
        - container.repository.cloudera.com/cloudera/cfm-operator@sha256:0312b6dadd5904ea9507bb857ab89d7266bf244c93b86855e0528e72813b7a85
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86af06fb3b6c17b22be47118b0a8c797fa7fdd3c00dbe99a2d2a593ca8fc3d63
        version: 2.10.0-b135
      entries:
      - name: cfm-operator.v2.10.0-b135
        version: 2.10.0-b135
      name: stable
    defaultChannel: stable
    packageName: cfm-operator
    provider:
      name: Cloudera
      url: https://cloudera.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Splunk Inc.
      provider-url: www.splunk.com
    name: splunk-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: splunk-operator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "IndexerCluster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "indexercluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "LicenseMaster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "licensemaster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "MonitoringConsole",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "monitoringconsole-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "SearchHeadCluster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "searchheadcluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "Standalone",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "standalone-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "clustermanager-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "IndexerCluster",
                "metadata": {
                  "name": "indexercluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "LicenseManager",
                "metadata": {
                  "name": "licensemanager-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "MonitoringConsole",
                "metadata": {
                  "name": "monitoringconsole-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "SearchHeadCluster",
                "metadata": {
                  "name": "searchheadcluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "Standalone",
                "metadata": {
                  "name": "standalone-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Big Data, Logging & Tracing, Monitoring, Security, AI/Machine
            Learning
          containerImage: splunk/splunk-operator@sha256:ed02561995d46ff8eea143fa78f5da118ee6fdb500832e9a33b42d059a67e229
          createdAt: "2025-09-08T10:17:59Z"
          description: The Splunk Operator for Kubernetes enables you to quickly and
            easily deploy Splunk Enterprise on your choice of private or public cloud
            provider. The Operator simplifies scaling and management of Splunk Enterprise
            by automating administrative workflows using Kubernetes best practices.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=2.4.0 <3.0.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/splunk/splunk-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterManager is the Schema for the cluster manager API
            displayName: Cluster Manager
            kind: ClusterManager
            name: clustermanagers.enterprise.splunk.com
            version: v4
          - description: ClusterMaster is the Schema for the cluster manager API
            displayName: Cluster Master
            kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v3
          - kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v1
          - kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v2
          - description: IndexerCluster is the Schema for a Splunk Enterprise indexer
              cluster
            displayName: Indexer Cluster
            kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v3
          - description: IndexerCluster is the Schema for a Splunk Enterprise indexer
              cluster
            displayName: Indexer Cluster
            kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v4
          - kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v1
          - kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v2
          - description: LicenseManager is the Schema for a Splunk Enterprise license
              manager.
            displayName: License Manager
            kind: LicenseManager
            name: licensemanagers.enterprise.splunk.com
            version: v4
          - description: LicenseMaster is the Schema for a Splunk Enterprise license
              master.
            displayName: License Master
            kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v3
          - kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v1
          - kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v2
          - description: MonitoringConsole is the Schema for the monitoringconsole
              API
            displayName: Monitoring Console
            kind: MonitoringConsole
            name: monitoringconsoles.enterprise.splunk.com
            version: v3
          - description: MonitoringConsole is the Schema for the monitoringconsole
              API
            displayName: Monitoring Console
            kind: MonitoringConsole
            name: monitoringconsoles.enterprise.splunk.com
            version: v4
          - description: SearchHeadCluster is the Schema for a Splunk Enterprise search
              head cluster
            displayName: Search Head Cluster
            kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v3
          - description: SearchHeadCluster is the Schema for a Splunk Enterprise search
              head cluster
            displayName: Search Head Cluster
            kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v4
          - kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v1
          - kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v2
          - description: Standalone is the Schema for a Splunk Enterprise standalone
              instances.
            displayName: Standalone
            kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v3
          - description: Standalone is the Schema for a Splunk Enterprise standalone
              instances.
            displayName: Standalone
            kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v4
          - kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v1
          - kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v2
        description: The Splunk Operator for Kubernetes enables you to quickly and
          easily deploy Splunk Enterprise on your choice of private or public cloud
          provider. The Operator simplifies scaling and management of Splunk Enterprise
          by automating administrative workflows using Kubernetes best practices.
        displayName: Splunk Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Splunk
        - Enterprise
        - Operator
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://github.com/splunk/splunk-operator/tree/main/docs
        maintainers:
        - email: support@splunk.com
          name: Splunk Operator Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Splunk Inc.
          url: www.splunk.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - splunk/splunk-operator@sha256:ed02561995d46ff8eea143fa78f5da118ee6fdb500832e9a33b42d059a67e229
        - splunk/splunk@sha256:4a5ce5a7b4e84a87396f0664895095c6510a0f2bf52e8747a373c7357f3ee313
        version: 3.0.0
      entries:
      - name: splunk-operator.v3.0.0
        version: 3.0.0
      - name: splunk-operator.v2.8.1
        version: 2.8.1
      - name: splunk-operator.v2.8.0
        version: 2.8.0
      - name: splunk-operator.v2.7.1
        version: 2.7.1
      - name: splunk-operator.v2.7.0
        version: 2.7.0
      name: stable
    defaultChannel: stable
    packageName: splunk-operator
    provider:
      name: Splunk Inc.
      url: www.splunk.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-emrcontainers-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-emrcontainers-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "emrcontainers.services.k8s.aws/v1alpha1",
                "kind": "VirtualCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/emrcontainers-controller:1.1.1
          createdAt: "2025-09-19T19:36:26Z"
          description: AWS EMR on EKS controller is a service controller for managing
            EMR on EKS resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: JobRun represents the state of an AWS emrcontainers JobRun
              resource.
            displayName: JobRun
            kind: JobRun
            name: jobruns.emrcontainers.services.k8s.aws
            version: v1alpha1
          - description: VirtualCluster represents the state of an AWS emrcontainers
              VirtualCluster resource.
            displayName: VirtualCluster
            kind: VirtualCluster
            name: virtualclusters.emrcontainers.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EMR on EKS resources in AWS from within your Kubernetes cluster.

          **About Amazon EMR on EKS**

          Amazon EMR on EKS provides a deployment option for Amazon EMR that allows you to run open-source big data frameworks on Amazon Elastic Kubernetes Service (Amazon EKS). With this deployment option, you can focus on running analytics workloads while Amazon EMR on EKS builds, configures, and manages containers for open-source applications.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EMR on EKS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - emrcontainers
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EMR on EKS Developer Resources
          url: https://docs.aws.amazon.com/emr/latest/EMR-on-EKS-DevelopmentGuide/emr-eks.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: emr on eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/emrcontainers-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-emrcontainers-controller.v1.1.1
        version: 1.1.1
      - name: ack-emrcontainers-controller.v1.1.0
        version: 1.1.0
      - name: ack-emrcontainers-controller.v1.0.27
        version: 1.0.27
      - name: ack-emrcontainers-controller.v1.0.25
        version: 1.0.25
      - name: ack-emrcontainers-controller.v1.0.24
        version: 1.0.24
      - name: ack-emrcontainers-controller.v1.0.23
        version: 1.0.23
      - name: ack-emrcontainers-controller.v1.0.22
        version: 1.0.22
      - name: ack-emrcontainers-controller.v1.0.21
        version: 1.0.21
      - name: ack-emrcontainers-controller.v1.0.18
        version: 1.0.18
      - name: ack-emrcontainers-controller.v1.0.16
        version: 1.0.16
      - name: ack-emrcontainers-controller.v1.0.15
        version: 1.0.15
      - name: ack-emrcontainers-controller.v1.0.14
        version: 1.0.14
      - name: ack-emrcontainers-controller.v1.0.13
        version: 1.0.13
      - name: ack-emrcontainers-controller.v1.0.12
        version: 1.0.12
      - name: ack-emrcontainers-controller.v1.0.11
        version: 1.0.11
      - name: ack-emrcontainers-controller.v1.0.9
        version: 1.0.9
      - name: ack-emrcontainers-controller.v1.0.8
        version: 1.0.8
      - name: ack-emrcontainers-controller.v1.0.4
        version: 1.0.4
      - name: ack-emrcontainers-controller.v1.0.3
        version: 1.0.3
      - name: ack-emrcontainers-controller.v1.0.1
        version: 1.0.1
      - name: ack-emrcontainers-controller.v1.0.0
        version: 1.0.0
      - name: ack-emrcontainers-controller.v0.1.2
        version: 0.1.2
      - name: ack-emrcontainers-controller.v0.1.0
        version: 0.1.0
      - name: ack-emrcontainers-controller.v0.0.10
        version: 0.0.10
      - name: ack-emrcontainers-controller.v0.0.9
        version: 0.0.9
      - name: ack-emrcontainers-controller.v0.0.8
        version: 0.0.8
      - name: ack-emrcontainers-controller.v0.0.7
        version: 0.0.7
      - name: ack-emrcontainers-controller.v0.0.6
        version: 0.0.6
      - name: ack-emrcontainers-controller.v0.0.5
        version: 0.0.5
      - name: ack-emrcontainers-controller.v0.0.4
        version: 0.0.4
      - name: ack-emrcontainers-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-emrcontainers-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: volume-expander-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: volume-expander-operator.v0.3.6
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage,Monitoring
          certified: "false"
          containerImage: quay.io/redhat-cop/volume-expander-operator@sha256:e4059a66a9af7c786c8c651abb8a57a55a0e528342e3a59f8c264b40f87d6e25
          createdAt: "2022-05-05T13:25:53Z"
          description: This operator provides automation expanding pod's volumes when
            they are about to run out of space.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: volume-expander-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/volume-expander-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The purpose of the volume-expander-operator is to expand volumes when they are running out of space.
          This is achieved by using the [volume expansion feature](https://kubernetes.io/docs/concepts/storage/persistent-volumes/#expanding-persistent-volumes-claims).

          The operator periodically checks the `kubelet_volume_stats_used_bytes` and `kubelet_volume_stats_capacity_bytes published` by the kubelets to decide when to expand a volume.
          Notice that these metrics are generated only when a volume is mounted to a pod. Also the kubelet takes a minute or two to start generating accurate values for these metrics. The operator accounts for that.

          This operator works based on the following annotations to PersistentVolumeClaim resources:

          | Annotation | Default &nbsp;  &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;  | Description  |
          | :--- | :--- | :--- |
          | `volume-expander-operator.redhat-cop.io/autoexpand`  | N/A  | if set to "true" enables the volume-expander-operator to watch on this PVC  |
          | `volume-expander-operator.redhat-cop.io/polling-frequency`  | `"30s"` | How frequently to poll the volume metrics. Express this value as a valid golang [Duration](https://golang.org/pkg/time/#ParseDuration)  |
          | `volume-expander-operator.redhat-cop.io/expand-threshold-percent` | `"80"` | the percentage of used storage after which the volume will be expanded. This must be a positive integer. |
          | `volume-expander-operator.redhat-cop.io/expand-by-percent` | `"25"` | the percentage by which the volume will be expanded, relative to the current size. This must be an integer between 0 and 100 |
          | `volume-expander-operator.redhat-cop.io/expand-up-to` | MaxInt64 | the upper bound for this volume to be expanded to. The default value is the largest quantity representable and is intended to be interpreted as infinite. If the default is used it is recommend to ensure the namespace has a quota on the used storage class. |

          Note that not all of the storage driver implementations support volume expansion. It is a responsibility of the user/platform administrator to ensure that storage class and the persistent volume claim meet all the requirements needed for the volume expansion feature to work properly.

          This operator was tested with [OCS](https://www.redhat.com/en/technologies/cloud-computing/openshift-container-storage), but should work with any other storage driver that supports volume expansion.
        displayName: Volume Expander Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - volume
        - storage
        - csi
        - expansion
        links:
        - name: Volume Expander Operator
          url: https://github.com/redhat-cop/volume-expander-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/volume-expander-operator@sha256:e4059a66a9af7c786c8c651abb8a57a55a0e528342e3a59f8c264b40f87d6e25
        version: 0.3.6
      entries:
      - name: volume-expander-operator.v0.3.6
        version: 0.3.6
      - name: volume-expander-operator.v0.3.5
        version: 0.3.5
      - name: volume-expander-operator.v0.3.4
        version: 0.3.4
      - name: volume-expander-operator.v0.3.3
        version: 0.3.3
      - name: volume-expander-operator.v0.3.2
        version: 0.3.2
      - name: volume-expander-operator.v0.3.1
        version: 0.3.1
      - name: volume-expander-operator.v0.3.0
        version: 0.3.0
      - name: volume-expander-operator.v0.2.0
        version: 0.2.0
      - name: volume-expander-operator.v0.1.8
        version: 0.1.8
      - name: volume-expander-operator.v0.1.7
        version: 0.1.7
      - name: volume-expander-operator.v0.1.6
        version: 0.1.6
      - name: volume-expander-operator.v0.1.5
        version: 0.1.5
      - name: volume-expander-operator.v0.1.4
        version: 0.1.4
      - name: volume-expander-operator.v0.1.3
        version: 0.1.3
      - name: volume-expander-operator.v0.1.2
        version: 0.1.2
      - name: volume-expander-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: volume-expander-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-eventbridge-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-eventbridge-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Archive",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Endpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "EventBus",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/eventbridge-controller:1.1.1
          createdAt: "2025-09-19T20:51:55Z"
          description: AWS EventBridge controller is a service controller for managing
            EventBridge resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Archive represents the state of an AWS eventbridge Archive
              resource.
            displayName: Archive
            kind: Archive
            name: archives.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: Endpoint represents the state of an AWS eventbridge Endpoint
              resource.
            displayName: Endpoint
            kind: Endpoint
            name: endpoints.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: EventBus represents the state of an AWS eventbridge EventBus
              resource.
            displayName: EventBus
            kind: EventBus
            name: eventbuses.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: Rule represents the state of an AWS eventbridge Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.eventbridge.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EventBridge resources in AWS from within your Kubernetes cluster.

          **About Amazon EventBridge**

          EventBridge is a serverless service that uses events to connect application components together, making it easier for you to build scalable event-driven applications. Use it to route events from sources such as home-grown applications, AWS services, and third- party software to consumer applications across your organization. EventBridge provides a simple and consistent way to ingest, filter, transform, and deliver events so you can build new applications quickly.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EventBridge
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eventbridge
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EventBridge Developer Resources
          url: https://aws.amazon.com/eventbridge/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eventbridge maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/eventbridge-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-eventbridge-controller.v1.1.1
        version: 1.1.1
      - name: ack-eventbridge-controller.v1.1.0
        version: 1.1.0
      - name: ack-eventbridge-controller.v1.0.28
        version: 1.0.28
      - name: ack-eventbridge-controller.v1.0.27
        version: 1.0.27
      - name: ack-eventbridge-controller.v1.0.26
        version: 1.0.26
      - name: ack-eventbridge-controller.v1.0.25
        version: 1.0.25
      - name: ack-eventbridge-controller.v1.0.24
        version: 1.0.24
      - name: ack-eventbridge-controller.v1.0.23
        version: 1.0.23
      - name: ack-eventbridge-controller.v1.0.22
        version: 1.0.22
      - name: ack-eventbridge-controller.v1.0.21
        version: 1.0.21
      - name: ack-eventbridge-controller.v1.0.20
        version: 1.0.20
      - name: ack-eventbridge-controller.v1.0.18
        version: 1.0.18
      - name: ack-eventbridge-controller.v1.0.17
        version: 1.0.17
      - name: ack-eventbridge-controller.v1.0.16
        version: 1.0.16
      - name: ack-eventbridge-controller.v1.0.15
        version: 1.0.15
      - name: ack-eventbridge-controller.v1.0.14
        version: 1.0.14
      - name: ack-eventbridge-controller.v1.0.13
        version: 1.0.13
      - name: ack-eventbridge-controller.v1.0.12
        version: 1.0.12
      - name: ack-eventbridge-controller.v1.0.11
        version: 1.0.11
      - name: ack-eventbridge-controller.v1.0.6
        version: 1.0.6
      - name: ack-eventbridge-controller.v1.0.5
        version: 1.0.5
      - name: ack-eventbridge-controller.v1.0.4
        version: 1.0.4
      - name: ack-eventbridge-controller.v1.0.3
        version: 1.0.3
      - name: ack-eventbridge-controller.v1.0.1
        version: 1.0.1
      - name: ack-eventbridge-controller.v1.0.0
        version: 1.0.0
      - name: ack-eventbridge-controller.v0.0.7
        version: 0.0.7
      - name: ack-eventbridge-controller.v0.0.6
        version: 0.0.6
      - name: ack-eventbridge-controller.v0.0.5
        version: 0.0.5
      - name: ack-eventbridge-controller.v0.0.4
        version: 0.0.4
      - name: ack-eventbridge-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-eventbridge-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Oracle Corporation
      provider-url: https://github.com/oracle/coherence-operator
    name: coherence-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: coherence-operator.v3.5.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "coherence.oracle.com/v1",
                "kind": "Coherence",
                "metadata": {
                  "name": "coherence-sample"
                },
                "spec": {
                  "replicas": 3
                }
              },
              {
                "apiVersion": "coherence.oracle.com/v1",
                "kind": "CoherenceJob",
                "metadata": {
                  "name": "coherence-job-sample"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "false"
          containerImage: container-registry.oracle.com/middleware/coherence-operator:3.5.4
          createdAt: "2025-08-28T16:15:09Z"
          description: The Oracle Coherence Kubernetes Operator enables easy management
            of Coherence clusters in a Kubernetes environment.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/arch.amd64: supported
          operatorframework.io/arch.arm64: supported
          operatorframework.io/os.linux: supported
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/oracle/coherence-operator
          support: Oracle Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Coherence is the top level schema for the Coherence API and
              custom resource definition (CRD).
            displayName: Coherence
            kind: Coherence
            name: coherence.coherence.oracle.com
            version: v1
          - description: |-
              CoherenceJob is the top level schema for the CoherenceJob API and custom resource definition (CRD)
              for configuring Coherence Job workloads.
            displayName: Coherence Job
            kind: CoherenceJob
            name: coherencejob.coherence.oracle.com
            version: v1
        description: |
          The Oracle Coherence Kubernetes Operator enables easy management of Coherence clusters in a Kubernetes environment.

          ### Installation Guide & Configuration

          For the complete installation guide and all configuration options please refer to
          [Coherence Operator Documentation](https://oracle.github.io/coherence-operator/docs/latest/#/about/01_overview).
        displayName: Coherence Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Coherence
        - Oracle
        - In Memory Data
        - Caching
        links:
        - name: Coherence Kubernetes Operator
          url: https://oracle.github.io/coherence-operator/docs/latest
        - name: Oracle Coherence
          url: https://coherence.community
        maintainers:
        - email: jonathan.knight@oracle.com
          name: Jonathan Knight
        maturity: mature
        minKubeVersion: 1.26.0
        provider:
          name: Oracle Corporation
          url: https://github.com/oracle/coherence-operator
        relatedImages:
        - container-registry.oracle.com/middleware/coherence-operator:3.5.4
        version: 3.5.4
      entries:
      - name: coherence-operator.v3.5.4
        version: 3.5.4
      - name: coherence-operator.v3.5.2
        version: 3.5.2
      - name: coherence-operator.v3.5.0
        version: 3.5.0
      - name: coherence-operator.v3.4.3
        version: 3.4.3
      name: stable
    defaultChannel: stable
    packageName: coherence-operator
    provider:
      name: Oracle Corporation
      url: https://github.com/oracle/coherence-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nxrm-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nxrm-operator-certified.v3.58.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusRepo",
                "metadata": {
                  "name": "example-nexusrepo"
                },
                "spec": {
                  "config": {
                    "data": null,
                    "enabled": false,
                    "mountPath": "/sonatype-nexus-conf"
                  },
                  "deployment": {
                    "additionalContainers": null,
                    "additionalVolumeMounts": null,
                    "additionalVolumes": null,
                    "annotations": {},
                    "initContainers": null,
                    "postStart": {
                      "command": null
                    },
                    "preStart": {
                      "command": null
                    },
                    "terminationGracePeriodSeconds": 120
                  },
                  "deploymentStrategy": {
                    "type": "Recreate"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "path": "/",
                    "rules": null,
                    "tls": {
                      "enabled": true,
                      "secretName": "nexus-tls"
                    }
                  },
                  "nexus": {
                    "dockerPort": 5003,
                    "env": [
                      {
                        "name": "INSTALL4J_ADD_VM_PARAMS",
                        "value": "-Xms2703M -Xmx2703M -XX:MaxDirectMemorySize=2703M -XX:+UnlockExperimentalVMOptions -XX:+UseCGroupMemoryLimitForHeap"
                      }
                    ],
                    "hostAliases": [],
                    "imageName": "registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:06fe31f40ff3fd7556fc7b8f6b2efe80ada1d9079ec9aaad60d1fdc0ee0677f0",
                    "imagePullPolicy": "IfNotPresent",
                    "imagePullSecret": "",
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "path": "/",
                      "periodSeconds": 30
                    },
                    "nexusPort": 8081,
                    "podAnnotations": {},
                    "readinessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "path": "/",
                      "periodSeconds": 30
                    },
                    "resources": {},
                    "securityContext": {
                    },
                    "service": {
                      "type": "NodePort"
                    }
                  },
                  "nexusProxyRoute": {
                    "annotations": null,
                    "enabled": false,
                    "labels": null
                  },
                  "persistence": {
                    "accessMode": "ReadWriteOnce",
                    "enabled": true,
                    "storageSize": "8Gi"
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": null,
                    "enabled": false,
                    "labels": null,
                    "name": "docker",
                    "portName": "docker"
                  },
                  "secret": {
                    "data": null,
                    "enabled": false,
                    "mountPath": "/etc/secret-volume",
                    "readOnly": true
                  },
                  "service": {
                    "annotations": {},
                    "enabled": false,
                    "labels": {},
                    "ports": [
                      {
                        "name": "nexus-service",
                        "port": 80,
                        "targetPort": 80
                      }
                    ]
                  },
                  "statefulset": {
                    "enabled": false
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nxrm-operator-certified@sha256:4945ad993a10e27b358b2d9869d05f23ca761162ec088e5a833939022bc23192
          createdAt: "2023-07-14"
          description: |-
            Nexus Repository is the central source of control to efficiently manage all binaries
            and build artifacts across your DevOps pipeline.
          olm.skipRange: <3.58.1-2
          repository: https://github.com/sonatype/operator-nxrm3
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus Repository
            displayName: NexusRepo
            kind: NexusRepo
            name: nexusrepos.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Repository is the central source of control to efficiently manage all binaries
          and build artifacts across your DevOps pipeline.
          The flow of open source and third-party components into and through an organization
          creates a complex software supply chain.
          Nexus Repository delivers speed, efficiency, and quality to the governance
          and management of all dependencies, libraries, and applications for your DevOps teams.

          ## Core Capabilities

          * **Dependency Management**:
            Improves reliability with repeatable, fast access to secure dependencies
          * **Developer Productivity**:
            Streamline developer workflows by enabling the sharing of components and applications across teams
          * **Supply Chain Performance**:
            Improve speed-to-market and reduced build times with release advanced staging and component tagging
          * **CI/CD Integrations**:
            Increase DevOps scalability with integrations to the most popular build and deployment tools

          Version control systems and package registries do not scale when managing proprietary,
          open source, and third-party components.
          Organizations need a central binary and build artifact repository to manage dependencies
          across the entire software supply chain.

          ## Limitations

          High Availability Clustering (HA-C) is not supported for Nexus Repository Pro for OpenShift.

          This operator will be released on a quarterly basis.

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus Repository Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage

          Once the server instance is created by the operator and running,
          you'll want to expose the service as you see fit:
          1. Create a Route to that service for nexus.port (8081).

          By default, the Nexus Repository starts up in OSS mode until a license is installed.

          The Nexus Repository can be further configured via the NexusRepo custom resource definition:

          | Parameter                                   | Description                         | Default                                 |
          | ------------------------------------------  | ----------------------------------  | ----------------------------------------|
          | `statefulset.enabled`                       | Use statefulset instead of deployment | `false` |
          | `deploymentStrategy.type`                   | Deployment Strategy     |  `Recreate` |
          | `nexus.env`                                 | Nexus environment variables         | `See example.` |
          | `nexus.resources`                           | Nexus resource requests and limits  | `{}`                                    |
          | `nexus.dockerPort`                          | Port to access docker               | `5003`                                  |
          | `nexus.nexusPort`                           | Internal port for Nexus service     | `8081`                                  |
          | `nexus.service.type`                        | Service for Nexus                   |`NodePort`                                |
          | `nexus.service.clusterIp`                   | Specific cluster IP when service type is cluster IP. Use None for headless service |`nil`   |
          | `nexus.securityContext`                     | Security Context |
          | `nexus.labels`                              | Service labels                      | `{}`                                    |
          | `nexus.podAnnotations`                      | Pod Annotations                     | `{}`
          | `nexus.livenessProbe.initialDelaySeconds`   | LivenessProbe initial delay         | 30                                      |
          | `nexus.livenessProbe.periodSeconds`         | Seconds between polls               | 30                                      |
          | `nexus.livenessProbe.failureThreshold`      | Number of attempts before failure   | 6                                       |
          | `nexus.livenessProbe.timeoutSeconds`        | Time in seconds after liveness probe times out    | `nil`                     |
          | `nexus.livenessProbe.path`                  | Path for LivenessProbe              | /                                       |
          | `nexus.readinessProbe.initialDelaySeconds`  | ReadinessProbe initial delay        | 30                                      |
          | `nexus.readinessProbe.periodSeconds`        | Seconds between polls               | 30                                      |
          | `nexus.readinessProbe.failureThreshold`     | Number of attempts before failure   | 6                                       |
          | `nexus.readinessProbe.timeoutSeconds`       | Time in seconds after readiness probe times out    | `nil`                    |
          | `nexus.readinessProbe.path`                 | Path for ReadinessProbe             | /                                       |
          | `nexus.hostAliases`                         | Aliases for IPs in /etc/hosts       | []                                      |
          | `ingress.enabled`                           | Create an ingress for Nexus         | `true`                                  |
          | `ingress.annotations`                       | Annotations to enhance ingress configuration  | `{}`                          |
          | `ingress.tls.enabled`                       | Enable TLS                          | `true`                                 |
          | `ingress.tls.secretName`                    | Name of the secret storing TLS cert, `false` to use the Ingress' default certificate | `nexus-tls`                             |
          | `ingress.path`                              | Path for ingress rules. GCP users should set to `/*` | `/`                    |
          | `tolerations`                               | tolerations list                    | `[]`                                    |
          | `config.enabled`                            | Enable configmap                    | `false`                                 |
          | `config.mountPath`                          | Path to mount the config            | `/sonatype-nexus-conf`                  |
          | `config.data`                               | Configmap data                      | `nil`                                   |
          | `deployment.terminationGracePeriodSeconds`  | Time to allow for clean shutdown    | 120                                     |
          | `deployment.annotations`                    | Annotations to enhance deployment configuration  | `{}`                       |
          | `deployment.initContainers`                 | Init containers to run before main containers  | `nil`                        |
          | `deployment.postStart.command`              | Command to run after starting the nexus container  | `nil`                    |
          | `deployment.preStart.command`               | Command to run before starting the nexus container  | `nil`                   |
          | `deployment.additionalContainers`           | Add additional Container         | `nil`                                      |
          | `deployment.additionalVolumes`              | Add additional Volumes           | `nil`                                      |
          | `deployment.additionalVolumeMounts`         | Add additional Volume mounts     | `nil`                                      |
          | `secret.enabled`                            | Enable secret                    | `false`                                    |
          | `secret.mountPath`                          | Path to mount the secret         | `/etc/secret-volume`                       |
          | `secret.readOnly`                           | Secret readonly state            | `true`                                     |
          | `secret.data`                               | Secret data                      | `nil`                                      |
          | `service.enabled`                           | Enable additional service        | `nil`                                      |
          | `service.name`                              | Service name                     | `nil`                                      |
          | `service.portName`                          | Service port name                | `nil`                                      |
          | `service.labels`                            | Service labels                   | `nil`                                      |
          | `service.annotations`                       | Service annotations              | `nil`                                      |
          | `service.loadBalancerSourceRanges`          | Service LoadBalancer source IP whitelist | `nil`                              |
          | `service.targetPort`                        | Service port                     | `nil`                                      |
          | `service.port`                              | Port for exposing service        | `nil`                                      |
          | `route.enabled`         | Set to true to create route for additional service | `false` |
          | `route.name`            | Name of route                                      | `docker` |
          | `route.portName`        | Target port name of service                        | `docker` |
          | `route.labels`          | Labels to be added to route                        | `{}` |
          | `route.annotations`     | Annotations to be added to route                   | `{}` |
          | `route.path`            | Host name of Route e.g jenkins.example.com         | nil |
        displayName: Nexus Repository Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - repository
        - sonatype
        links:
        - name: Nexus Repository Manager
          url: https://www.sonatype.com/product-nexus-repository
        maintainers:
        - email: support@sonatype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nxrm-operator-certified@sha256:4945ad993a10e27b358b2d9869d05f23ca761162ec088e5a833939022bc23192
        - registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:06fe31f40ff3fd7556fc7b8f6b2efe80ada1d9079ec9aaad60d1fdc0ee0677f0
        version: 3.58.1-2
      entries:
      - name: nxrm-operator-certified.v3.58.1-2
        version: 3.58.1-2
      name: stable
    defaultChannel: stable
    packageName: nxrm-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: AMD
      provider-url: ""
    name: amd-t2-january
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amd-t2-january.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovfect2.amd.com/v1",
                "kind": "SriovT2Card",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "amd-t2-january",
                    "app.kubernetes.io/instance": "sriovt2card-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sriovt2card",
                    "app.kubernetes.io/part-of": "amd-t2-january"
                  },
                  "name": "sriovt2card",
                  "namespace": "amd-t2"
                },
                "spec": {
                  "acceleratorSelector": {
                    "pciAddress": "0000:01:00.0"
                  },
                  "nodeSelector": {
                    "kubernetes.io/hostname": "7c-c2-55-85-ba-fe.amd.com"
                  },
                  "physicalFunction": {
                    "pfDriver": "qdma-pf",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-07-14T09:46:54Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovT2Card is the Schema for the sriovt2cards API
            displayName: Sriov T2 Card
            kind: SriovT2Card
            name: sriovt2cards.sriovfect2.amd.com
            version: v1
        description: The AMD-T2 Operator for accelerator enables the seamless integration
          and management of AMD T2 accelerator within Openshift/Kubernetes envirnoment.
        displayName: AMD-T2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - SRIOV
        - Xilinx
        - FPGA
        - Telco
        - T2
        - MultiArch
        - AMD
        - TelcoAccelerator
        links:
        - name: Amd T2 January
          url: https://amd-t2-january.domain
        maintainers:
        - email: dhirajs.lakhane@amd.com
          name: DhirajLakhane
        maturity: alpha
        provider:
          name: AMD
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:928e64203edad8f1bba23593c7be04f0f8410c6e4feb98d9e9c2d00a8ff59048
        - quay.io/amdt2operator/amd-t2-january@sha256:318f56e2670582c217b8973f69f0675e7fff7dc19d07875294f2ff2fc5403902
        version: 1.0.1
      entries:
      - name: amd-t2-january.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: amd-t2-january
    provider:
      name: AMD
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: placement-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: placement-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "placement.openstack.org/v1beta1",
                "kind": "PlacementAPI",
                "metadata": {
                  "name": "placement"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\n",
                  "databaseAccount": "placement",
                  "databaseInstance": "openstack",
                  "preserveJobs": false,
                  "replicas": 1,
                  "secret": "placement-secret",
                  "serviceUser": "placement"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:43Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PlacementAPI is the Schema for the placementapis API
            displayName: Placement API
            kind: PlacementAPI
            name: placementapis.placement.openstack.org
            version: v1beta1
        description: Placement Operator
        displayName: Placement Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Placement
        links:
        - name: Placement Operator
          url: https://github.com/openstack-k8s-operators/placement-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/placement-rhel9-operator@sha256:d3be48618b1f0f717aea162ae41876137fee8b35329d2e3e66e9d4eaccf62376
        - registry.redhat.io/rhoso/openstack-placement-api-rhel9@sha256:3308f2eccfbe7416a85443696273417bfb816df36b810ebbefa108dda12a5b19
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: placement-operator.v1.0.6
        version: 1.0.6
      - name: placement-operator.v1.0.5
        version: 1.0.5
      - name: placement-operator.v1.0.4
        version: 1.0.4
      - name: placement-operator.v1.0.3
        version: 1.0.3
      - name: placement-operator.v1.0.2
        version: 1.0.2
      - name: placement-operator.v1.0.1
        version: 1.0.1
      - name: placement-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: placement-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/cluster-nfd-operator
    name: nfd
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfd.4.16.0-202507211806
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nfd.openshift.io/v1",
                "kind": "NodeFeatureDiscovery",
                "metadata": {
                  "name": "nfd-instance",
                  "namespace": "openshift-nfd"
                },
                "spec": {
                  "customConfig": {
                    "configData": "#    - name: \"more.kernel.features\"\n#      matchOn:\n#      - loadedKMod: [\"example_kmod3\"]\n#    - name: \"more.features.by.nodename\"\n#      value: customValue\n#      matchOn:\n#      - nodename: [\"special-.*-node-.*\"]\n"
                  },
                  "operand": {
                    "image": "registry.redhat.io/openshift4/ose-node-feature-discovery-rhel9@sha256:bb65ec6a6f88688630305e081cb4bcb88aa538b5f6a3385ee28c61522dcd1662",
                    "imagePullPolicy": "IfNotPresent",
                    "servicePort": 12000
                  },
                  "workerConfig": {
                    "configData": "core:\n#  labelWhiteList:\n#  noPublish: false\n  sleepInterval: 60s\n#  sources: [all]\n#  klog:\n#    addDirHeader: false\n#    alsologtostderr: false\n#    logBacktraceAt:\n#    logtostderr: true\n#    skipHeaders: false\n#    stderrthreshold: 2\n#    v: 0\n#    vmodule:\n##   NOTE: the following options are not dynamically run-time \n##          configurable and require a nfd-worker restart to take effect\n##          after being changed\n#    logDir:\n#    logFile:\n#    logFileMaxSize: 1800\n#    skipLogHeaders: false\nsources:\n#  cpu:\n#    cpuid:\n##     NOTE: whitelist has priority over blacklist\n#      attributeBlacklist:\n#        - \"BMI1\"\n#        - \"BMI2\"\n#        - \"CLMUL\"\n#        - \"CMOV\"\n#        - \"CX16\"\n#        - \"ERMS\"\n#        - \"F16C\"\n#        - \"HTT\"\n#        - \"LZCNT\"\n#        - \"MMX\"\n#        - \"MMXEXT\"\n#        - \"NX\"\n#        - \"POPCNT\"\n#        - \"RDRAND\"\n#        - \"RDSEED\"\n#        - \"RDTSCP\"\n#        - \"SGX\"\n#        - \"SSE\"\n#        - \"SSE2\"\n#        - \"SSE3\"\n#        - \"SSE4.1\"\n#        - \"SSE4.2\"\n#        - \"SSSE3\"\n#      attributeWhitelist:\n#  kernel:\n#    kconfigFile: \"/path/to/kconfig\"\n#    configOpts:\n#      - \"NO_HZ\"\n#      - \"X86\"\n#      - \"DMI\"\n  pci:\n    deviceClassWhitelist:\n      - \"0200\"\n      - \"03\"\n      - \"12\"\n    deviceLabelFields:\n#      - \"class\"\n      - \"vendor\"\n#      - \"device\"\n#      - \"subsystem_vendor\"\n#      - \"subsystem_device\"\n#  usb:\n#    deviceClassWhitelist:\n#      - \"0e\"\n#      - \"ef\"\n#      - \"fe\"\n#      - \"ff\"\n#    deviceLabelFields:\n#      - \"class\"\n#      - \"vendor\"\n#      - \"device\"\n#  custom:\n#    - name: \"my.kernel.feature\"\n#      matchOn:\n#        - loadedKMod: [\"example_kmod1\", \"example_kmod2\"]\n#    - name: \"my.pci.feature\"\n#      matchOn:\n#        - pciId:\n#            class: [\"0200\"]\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#        - pciId :\n#            vendor: [\"8086\"]\n#            device: [\"1000\", \"1100\"]\n#    - name: \"my.usb.feature\"\n#      matchOn:\n#        - usbId:\n#          class: [\"ff\"]\n#          vendor: [\"03e7\"]\n#          device: [\"2485\"]\n#        - usbId:\n#          class: [\"fe\"]\n#          vendor: [\"1a6e\"]\n#          device: [\"089a\"]\n#    - name: \"my.combined.feature\"\n#      matchOn:\n#        - pciId:\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#          loadedKMod : [\"vendor_kmod1\", \"vendor_kmod2\"]\n"
                  }
                }
              },
              {
                "apiVersion": "nfd.openshift.io/v1alpha1",
                "kind": "NodeFeatureRule",
                "metadata": {
                  "name": "my-sample-rule-object"
                },
                "spec": {
                  "rules": [
                    {
                      "labels": {
                        "my-sample-feature": "true"
                      },
                      "matchFeatures": [
                        {
                          "feature": "kernel.loadedmodule",
                          "matchExpressions": {
                            "dummy": {
                              "op": "Exists"
                            }
                          }
                        },
                        {
                          "feature": "kernel.config",
                          "matchExpressions": {
                            "X86": {
                              "op": "In",
                              "value": [
                                "y"
                              ]
                            }
                          }
                        }
                      ],
                      "name": "my sample rule"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery,OpenShift Optional
          containerImage: registry.redhat.io/openshift4/ose-cluster-nfd-rhel9-operator@sha256:c1b6a5b8a59c08458d57c0fc2bb2bc9b056f006d15edb4d3161a3d1c0bc97ae6
          description: |-
            The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

            NFD consists  of the following software components:

            The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

            ##NFD-Master
            NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

            ##NFD-Worker
            NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

            ##NFD-Topology-Updater
            NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.16.0-202507211806'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-nfd
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/internal-objects: '["noderesourcetopologies.topology.node.k8s.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/cluster-nfd-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              The NodeFeatureDiscovery instance is the CustomResource being watched by the NFD-Operator, and holds all the needed information to setup the behaviour of the master and worker pods
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
          - description: |
              NodeFeatureRule resource specifies a configuration for feature-based customization of node objects, such as node labeling.
            kind: NodeFeatureRule
            name: nodefeaturerules.nfd.openshift.io
            version: v1alpha1
          - kind: NodeFeature
            name: nodefeatures.nfd.openshift.io
            version: v1alpha1
        description: |-
          The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

          NFD consists  of the following software components:

          The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

          ##NFD-Master
          NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

          ##NFD-Worker
          NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

          ##NFD-Topology-Updater
          NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
        displayName: Node Feature Discovery Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - feature-discovery
        - feature-detection
        - node-labels
        links:
        - name: Node Feature Discovery Operator
          url: https://docs.openshift.com/container-platform/4.16/hardware_enablement/psap-node-feature-discovery-operator.html
        - name: Node Feature Discovery Documentation
          url: https://kubernetes-sigs.github.io/node-feature-discovery/stable/get-started/index.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/cluster-nfd-operator
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:bcca89fb000b0fe8722ad7807b62a43ef3ae8fa76db1d790c285a975bfecb1aa
        - registry.redhat.io/openshift4/ose-cluster-nfd-rhel9-operator@sha256:c1b6a5b8a59c08458d57c0fc2bb2bc9b056f006d15edb4d3161a3d1c0bc97ae6
        - registry.redhat.io/openshift4/ose-node-feature-discovery-rhel9@sha256:bb65ec6a6f88688630305e081cb4bcb88aa538b5f6a3385ee28c61522dcd1662
        version: 4.16.0-202507211806
      entries:
      - name: nfd.4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: nfd.4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: nfd.4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: nfd.4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: nfd.4.16.0-202504271139
        version: 4.16.0-202504271139
      - name: nfd.4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: nfd.4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: nfd.4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: nfd.4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: nfd.4.16.0-202501271512
        version: 4.16.0-202501271512
      - name: nfd.4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: nfd.4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: nfd.4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: nfd.4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: nfd.4.16.0-202410251436
        version: 4.16.0-202410251436
      - name: nfd.4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: nfd.4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: nfd.4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: nfd.4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: nfd.4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: nfd.4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: nfd.4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: nfd.4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: nfd.4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: nfd.4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: nfd.4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: nfd.4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: nfd.4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: nfd.4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: nfd
    provider:
      name: Red Hat
      url: https://github.com/openshift/cluster-nfd-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhpam-kogito-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhpam-kogito-operator.7.13.5-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "ruleunit-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogitoinfra-sample"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "kafka.strimzi.io/v1beta2",
                    "kind": "Kafka",
                    "name": "kogito-kafka"
                  }
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator@sha256:19497d4aff0ea3ea114dccb0d0e265458bb14ba6d325353eb82b554547dbd2e4
          description: RHPAM Kogito Operator for deployment and management of Kogito
            services.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Process Automation
            Manager", "Red Hat Decision Manager"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/rhpam-kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.rhpam.kiegroup.org
            version: v1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.rhpam.kiegroup.org
            version: v1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.rhpam.kiegroup.org
            version: v1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.rhpam.kiegroup.org
            version: v1
        description: |-
          RHPAM Kogito Operator is designed for deployment and management of Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Builds a RHPAM Kogito service from a remote git repository and creates a custom image for the given runtime: Red Hat build of Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Deploy RHPAM Kogito service from user provided image.
          * Setup Grafana Dashboard when Grafana Operator is available in cluster.
          * Create Prometheus monitor object when Prometheus operator is available in cluster.

          **IMPORTANT!** RHPAM Kogito Operator may need Grafana and Prometheus Operator to configure monitoring in cases if you do not provide these services beforehand. If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the RHPAM Kogito Operator.
        displayName: RHPAM Kogito Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-process-automation-manager
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_process_automation_manager/7.13/html/getting_started_with_red_hat_build_of_kogito_in_red_hat_process_automation_manager/assembly-deploying-kogito-microservices-on-openshift
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator@sha256:19497d4aff0ea3ea114dccb0d0e265458bb14ba6d325353eb82b554547dbd2e4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9492c67eaa7ab37df6f5fa623c4aac955b7d7ffaab9ae2e57dea745e82907e40
        version: 7.13.5-4
      entries:
      - name: rhpam-kogito-operator.7.13.5-4
        version: 7.13.5-4
      - name: rhpam-kogito-operator.7.13.5-3
        version: 7.13.5-3
      - name: rhpam-kogito-operator.7.13.5-2
        version: 7.13.5-2
      - name: rhpam-kogito-operator.7.13.5-1
        version: 7.13.5-1
      - name: rhpam-kogito-operator.7.13.4-3
        version: 7.13.4-3
      - name: rhpam-kogito-operator.7.13.4-2
        version: 7.13.4-2
      - name: rhpam-kogito-operator.7.13.4-1
        version: 7.13.4-1
      - name: rhpam-kogito-operator.7.13.3-1
        version: 7.13.3-1
      - name: rhpam-kogito-operator.7.13.2-3
        version: 7.13.2-3
      - name: rhpam-kogito-operator.7.13.2-2
        version: 7.13.2-2
      - name: rhpam-kogito-operator.7.13.2-1
        version: 7.13.2-1
      - name: rhpam-kogito-operator.7.13.1-2
        version: 7.13.1-2
      - name: rhpam-kogito-operator.7.13.1-1
        version: 7.13.1-1
      - name: rhpam-kogito-operator.7.13.0-1
        version: 7.13.0-1
      - name: rhpam-kogito-operator.v7.12.0-4
        version: 7.12.0-4
      - name: rhpam-kogito-operator.v7.12.0-3
        version: 7.12.0-3
      - name: rhpam-kogito-operator.v7.12.0-2
        version: 7.12.0-2
      name: 7.x
    defaultChannel: 7.x
    packageName: rhpam-kogito-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://skupper.io
    name: skupper-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-operator.v1.4.8-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus@sha256:65602221cdff7f718d4149f4dead167d05dbbe0cea1eaa8bf4b98be588135583
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:c665677c639699d6a5ef24aed5a1a3d9c0a5ee25f077a6db42a9fc6247d93132
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:eedde9f2019d4e919612bcc18f427c0d88b4846e773751af231ba0e560d48869
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:1fd908e80ef21fe7d7051c400cb099f2f3ca51c8cdd698705ef9784d34688911
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:4c9cc9df83baa1ef37cfb429babe5b86a962b8bb096209b3fb74291c67c94000
        version: 1.4.8-rh-1
      entries:
      - name: skupper-operator.v1.4.8-rh-1
        version: 1.4.8-rh-1
      - name: skupper-operator.v1.4.7-rh-3
        version: 1.4.7-rh-3
      - name: skupper-operator.v1.4.7-rh-2
        version: 1.4.7-rh-2
      - name: skupper-operator.v1.4.7-rh-1
        version: 1.4.7-rh-1
      - name: skupper-operator.v1.4.5-rh-4
        version: 1.4.5-rh-4
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1
    - currentCSV: skupper-operator.v1.4.8-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:eedde9f2019d4e919612bcc18f427c0d88b4846e773751af231ba0e560d48869
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:1fd908e80ef21fe7d7051c400cb099f2f3ca51c8cdd698705ef9784d34688911
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:4c9cc9df83baa1ef37cfb429babe5b86a962b8bb096209b3fb74291c67c94000
        - registry.redhat.io/openshift4/ose-prometheus@sha256:65602221cdff7f718d4149f4dead167d05dbbe0cea1eaa8bf4b98be588135583
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:c665677c639699d6a5ef24aed5a1a3d9c0a5ee25f077a6db42a9fc6247d93132
        version: 1.4.8-rh-1
      entries:
      - name: skupper-operator.v1.4.8-rh-1
        version: 1.4.8-rh-1
      - name: skupper-operator.v1.4.7-rh-3
        version: 1.4.7-rh-3
      - name: skupper-operator.v1.4.7-rh-2
        version: 1.4.7-rh-2
      - name: skupper-operator.v1.4.7-rh-1
        version: 1.4.7-rh-1
      - name: skupper-operator.v1.4.5-rh-4
        version: 1.4.5-rh-4
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.4
    - currentCSV: skupper-operator.v1.5.5-rh-6
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:9e9e1f2288f0218e3ba2abf4bebac9731d7cf9f9c567e0f6de6767b0b66cbee4
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:a5d7c1b2b9db672548126c0873d062a55667d93291fd3e32809d3c9c900c548e
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:9e9e1f2288f0218e3ba2abf4bebac9731d7cf9f9c567e0f6de6767b0b66cbee4
        - registry.redhat.io/openshift4/ose-prometheus@sha256:53afaf53975807765fc3777314edd218c789c17680c1f57a8e33a8f9a76f38bf
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:3e78ec8cdb729a704d5a26e54697f285ba7c0b9afee6160b7ecc703876120dea
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:be9dbcd2a3a1e1e536921390dcfd283fc45dc4423d2705376ef37a9fb53f7cf1
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:1de071dabf620265252f3922c632f17f676ba716ce4c3df931a0ccb3871dbb6c
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:5f8621d0013febcecbed4d7c82db11a4dded99ad8ce6b6d83858b9bdbe79ee5a
        version: 1.5.5-rh-6
      entries:
      - name: skupper-operator.v1.5.5-rh-6
        version: 1.5.5-rh-6
      - name: skupper-operator.v1.5.5-rh-5
        version: 1.5.5-rh-5
      - name: skupper-operator.v1.5.5-rh-4
        version: 1.5.5-rh-4
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.5
    - currentCSV: skupper-operator.v1.8.7-rh-2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:b0d43628b779a101a24e4b271318a34c21f3b71bcba2bf20804f941a93d27455
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.8.7-rh-2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:56b8fde75a6e7df235c560a0f4b9114398da5ba8b20f75f1ea7338f48ce19506
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:d7d252cff08ba1589afd8707ed349f87d3b6214906a4fb49932bda0a4d19fee4
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:5ea36e96f878577104ab7601e57e7a2a233f11dbdc533454f95e2e14a066d21f
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:3adeef827eca0382cd410f7c3e99a0fee9b50791cb0e0c6fc130b72a2550684f
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:b0d43628b779a101a24e4b271318a34c21f3b71bcba2bf20804f941a93d27455
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:f798745e799515065af099a816057b9c98a741e951d30a0f45ed47ca50e6724e
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:61542c5f0989f51ac5d6c96f5d0e29aca75614b79e7626299c77c5a74636b564
        version: 1.8.7-rh-2
      entries:
      - name: skupper-operator.v1.8.7-rh-2
        version: 1.8.7-rh-2
      - name: skupper-operator.v1.8.7-rh-1
        version: 1.8.7-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.8
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.9
    - currentCSV: skupper-operator.v2.1.1-rh-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
          createdAt: "2025-08-19T20:43:19Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:36872037066fae806cbf88cab673229f25e592295972aa18227827233362eec5
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:c01d0eab0841b1a5fddb58a0ae110beb464b91390f02662590bebc7b2a7213b6
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e8df53454cf12a14338790af6ed209798010990ca3320096e8b44acf901c4123
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:1a290a5ac8b9d1b082e740054675fe432aa424c2f55dd662caf67c740044f229
        version: 2.1.1-rh-3
      entries:
      - name: skupper-operator.v2.1.1-rh-3
        version: 2.1.1-rh-3
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2
    - currentCSV: skupper-operator.v2.0.1-rh-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:b5c73dc59a9184cd120c2ee35256cc097c790b8ef488af0c732d5efe3192b768
          createdAt: "2025-05-29T13:35:40Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.0.1-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:92f44cff160874efe57a985609e599b13c49c0c4cef846dccc80833af2ae74f3
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:e239c859a1f49816ae3068455e703b8a12919f11158b0f6ee0e6bfb31e8a8082
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:f2a41674ed196c089748d6386c18c14580e502de33235ef23c37eddeb6350f41
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:95062ff1256135cfa17907e82b09b2a211371ea0e76ccc711b7c737e040d87e9
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:498b5bb648e5bbaad07663941600173abac7d60a10744305f619509ddf7563e0
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:bd3f72544841c91c2d573a65ac3c988756f3523e17e0988027793b0f5f3ec3d1
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:b5c73dc59a9184cd120c2ee35256cc097c790b8ef488af0c732d5efe3192b768
        version: 2.0.1-rh-2
      entries:
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2.0
    - currentCSV: skupper-operator.v2.1.1-rh-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
          createdAt: "2025-08-19T20:43:19Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:c01d0eab0841b1a5fddb58a0ae110beb464b91390f02662590bebc7b2a7213b6
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e8df53454cf12a14338790af6ed209798010990ca3320096e8b44acf901c4123
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:1a290a5ac8b9d1b082e740054675fe432aa424c2f55dd662caf67c740044f229
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:36872037066fae806cbf88cab673229f25e592295972aa18227827233362eec5
        version: 2.1.1-rh-3
      entries:
      - name: skupper-operator.v2.1.1-rh-3
        version: 2.1.1-rh-3
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2.1
    defaultChannel: stable-2
    packageName: skupper-operator
    provider:
      name: Red Hat
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://horreum.hyperfoil.io
    name: horreum-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: horreum-operator.v0.7.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hyperfoil.io/v1alpha1",
                "kind": "Horreum",
                "metadata": {
                  "name": "horreum"
                },
                "spec": {
                  "nodeHost": "127.0.0.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/hyperfoil/horreum-operator:0.7.9
          createdAt: "2023-06-13T10:42:39Z"
          description: Performance results repository
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/hyperfoil/horreum-operator
          support: Red Hat Application Runtimes Performance Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Horreum is the object configuring Horreum performance results
              repository
            displayName: Horreum
            kind: Horreum
            name: horreums.hyperfoil.io
            version: v1alpha1
        description: Performance results repository
        displayName: Horreum
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - performance
        - repository
        - database
        - results
        - hyperfoil
        - benchmark
        links:
        - name: Horreum Operator
          url: https://horreum-operator.domain
        maintainers:
        - email: rvansa@redhat.com
          name: Radim Vansa
        - email: spederse@redhat.com
          name: Ståle W. Pedersen
        - email: wreicher@redhat.com
          name: Will Reichert
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat, Inc.
          url: https://horreum.hyperfoil.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/hyperfoil/horreum-operator:0.7.9
        version: 0.7.9
      entries:
      - name: horreum-operator.v0.7.9
        version: 0.7.9
      - name: horreum-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: horreum-operator
    provider:
      name: Red Hat, Inc.
      url: https://horreum.hyperfoil.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/stolostron/multicluster-global-hub
    name: multicluster-global-hub-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-global-hub-operator.v1.4.0-dev.20250302
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2025-02-14T08:24:47Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: planning
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/grafana:2.12.0-SNAPSHOT-2024-09-03-21-11-25
        - quay.io/stolostron/multicluster-global-hub-operator:v1.4.0
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-relations/spicedb-operator:latest
        - quay.io/stolostron/multicluster-global-hub-manager:v1.4.0
        - quay.io/stolostron/postgresql-16:9.5-1732622748
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-inventory/inventory-api:latest
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-relations/spicedb:latest
        - quay.io/stolostron/multicluster-global-hub-agent:v1.4.0
        - quay.io/prometheuscommunity/postgres-exporter:v0.15.0
        version: 1.4.0-dev.20250302
      entries:
      - name: multicluster-global-hub-operator.v1.4.0-dev.20250302
        version: 1.4.0-dev.20250302
      - name: multicluster-global-hub-operator.v1.4.0-dev.20250225
        version: 1.4.0-dev.20250225
      - name: multicluster-global-hub-operator.v1.3.0-dev.20240911
        version: 1.3.0-dev.20240911
      - name: multicluster-global-hub-operator.v1.3.0-dev.20240724
        version: 1.3.0-dev.20240724
      - name: multicluster-global-hub-operator.v1.2.0-dev.20240516
        version: 1.2.0-dev.20240516
      - name: multicluster-global-hub-operator.v1.2.0-dev
        version: 1.2.0-dev
      - name: multicluster-global-hub-operator.v1.1.0-dev
        version: 1.1.0-dev
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: multicluster-global-hub-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-operator-globalhub-1-5@sha256:4db1251323bd64930461cb5025da41baffcf5c00c51ea4e49940b7130971b510
          createdAt: "2025-06-10T06:58:35Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/postgres-exporter-globalhub-1-5@sha256:1108572a8bdafde926254f1ff94c2b22e048706be660b07c793e9f9e038a0f4a
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-operator-globalhub-1-5@sha256:4db1251323bd64930461cb5025da41baffcf5c00c51ea4e49940b7130971b510
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-manager-globalhub-1-5@sha256:01253426d65f1297a1955a8393051e6a4b1df5e5249774fc5189a75f35e7d644
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-agent-globalhub-1-5@sha256:6f7520eac83abec7780f831d8473544d0b8db43fc6e77d59e59317b09eb1f6c1
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/glo-grafana-globalhub-1-5@sha256:73fdf00538e63595b7cc92d484e087ff815579a95d7259c579c6e7dc1beacff6
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        version: 1.5.0
      entries:
      - name: multicluster-global-hub-operator.v1.5.0
        version: 1.5.0
      - name: multicluster-global-hub-operator.v1.4.0
        version: 1.4.0
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-1.5
    - currentCSV: multicluster-global-hub-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2024-03-05T02:13:54Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.32.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:latest
        version: 1.1.0
      entries:
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.10
    - currentCSV: multicluster-global-hub-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2024-07-03T06:52:02Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v1.2.0
        version: 1.2.1
      entries:
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.11
    - currentCSV: multicluster-global-hub-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator@sha256:636aea861da80fc36fb27a419a18c5b3fbb1bc072e8ae33c75e1cc47e2c05f75
          createdAt: "2024-10-10T05:42:48Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          kafka-zookeeper-0                                       1/1     Running   0          21m
          kafka-zookeeper-1                                       1/1     Running   0          21m
          kafka-zookeeper-2                                       1/1     Running   0          21m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgres-0                      2/2     Running   0          22m
          ```
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:636aea861da80fc36fb27a419a18c5b3fbb1bc072e8ae33c75e1cc47e2c05f75
        - quay.io/prometheuscommunity/postgres-exporter@sha256:386b12d19eab2a37d7cd8ca8b4c7491cc7a830d9581f49af6c98a393da9605e6
        - quay.io/stolostron/inventory-api@sha256:a15e62ee070dba5d74cf336b3b89b467593c5a33cfe3623d922f0af4fd7df392
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/release-globalhub-1-3/multicluster-global-hub-manager-globalhub-1-3@sha256:7656d01a4d7b39c61c3249695d68011d9e0df308627b9efd117686e901281ab1
        - quay.io/stolostron/multicluster-global-hub-manager@sha256:715cb5d0c6c407ec5c0303fc7e6d83fc77ce7e06775e169a475f7c075f1e7000
        - quay.io/stolostron/grafana@sha256:719b320a05d10158fadce5ae2fc6e4254a18957a6608b6f81e876cf46d3a6897
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        version: 1.3.0
      entries:
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.12
    - currentCSV: multicluster-global-hub-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator@sha256:085e9fc6cdc6a42df10a38856c8ccff25e5acffaba67ef95da72027e90e507fc
          createdAt: "2025-02-14T08:24:47Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-inventory/inventory-api@sha256:6a55a1919ecd92d94e05c12be1b6450956f5fda2ce17af7027cac3806c4970a4
        - quay.io/prometheuscommunity/postgres-exporter@sha256:386b12d19eab2a37d7cd8ca8b4c7491cc7a830d9581f49af6c98a393da9605e6
        - quay.io/stolostron/multicluster-global-hub-manager@sha256:4cd602ff6f8dd951848dc8a95c856004295c9f4710f77d4dafcc0735c1492395
        - quay.io/stolostron/multicluster-global-hub-agent@sha256:8dfc33dfed33a85733ee5e02dcd4f05662c75263791e0961ba17cbb3d0c340c9
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:085e9fc6cdc6a42df10a38856c8ccff25e5acffaba67ef95da72027e90e507fc
        - quay.io/stolostron/grafana@sha256:719b320a05d10158fadce5ae2fc6e4254a18957a6608b6f81e876cf46d3a6897
        version: 1.4.0
      entries:
      - name: multicluster-global-hub-operator.v1.4.0
        version: 1.4.0
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.13
    - currentCSV: multicluster-global-hub-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "storage": {
                    "name": "storage-secret"
                  },
                  "transport": {
                    "name": "transport-secret"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.5.0
          createdAt: "2022-08-23T20:00:31Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha1
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.5 or later needs to be installed.\n- PostgreSQL needs to be installed
          and you must create database for the multicluster global hub. You must create
          a secret named postgresql-secret that contains the database access credential
          in `open-cluster-management` namespace. You can run the following command
          to create the secret: \n\n```\nkubectl create secret generic storage-secret
          -n open-cluster-management \\\n  --from-literal=database_uri=<postgresql-uri>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)
          to install postgres in `hoh-postgres` namespace and create the secret `storage-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n- Kafka needs to be installed,
          and you must create two sections, `spec` and `status`. You need to create
          a secret named kafka-secret that contains the kafka access information should
          be created in `open-cluster-management` namespace. You can run the following
          command to create the secret: \n\n```\nkubectl create secret generic transport-secret
          -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n  --from-literal=CA=<CA-for-kafka-server>\n```\n >_Note:_ There is a
          sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `transport-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n## How to Install\n\nInstall the
          Multicluster Global Hub Operator by following the instructions that are
          displayed after you select the Install button. The operator must be installed
          in the same namespace as Red Hat Advanced Cluster Management for Kubernetes
          (RHACM), by default, in `open-cluster-management` namespace.\n\nA pod will
          be created in `open-cluster-management` namespace\n\n```\n$ kubectl get
          pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          19s\n```\n\nThe operator is
          now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installing the operator,
          create an instance of the MulticlusterGlobalHub resource to install the
          Multicluster Global Hub.\n\nThe pods of Multicluster Global Hub are created
          in the `open-cluster-management` namespace:\n\n```\n$ kubectl get pods -n
          open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\nmulticluster-global-hub-rbac-77888b7dc4-z9vbp
          \                   1/1     Running   0          10s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.5.0
        version: 0.5.0
      entries:
      - name: multicluster-global-hub-operator.v0.5.0
        version: 0.5.0
      name: release-2.6
    - currentCSV: multicluster-global-hub-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha2",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "largeScale": {
                      "kafka": {
                        "name": "transport-secret"
                      },
                      "postgres": {
                        "name": "storage-secret"
                      }
                    },
                    "type": "largeScale"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.6.0
          createdAt: "2023-02-09T20:00:31Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha2
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.7 or later needs to be installed.\n- PostgreSQL needs to be installed
          and you must create database for the multicluster global hub. You must create
          a secret named postgresql-secret that contains the database access credential
          in `open-cluster-management` namespace. You can run the following command
          to create the secret: \n\n```\nkubectl create secret generic storage-secret
          -n open-cluster-management \\\n  --from-literal=database_uri=<postgresql-uri>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)
          to install postgres in `hoh-postgres` namespace and create the secret `storage-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n- Kafka needs to be installed,
          and you must create two sections, `spec` and `status`. You need to create
          a secret named kafka-secret that contains the kafka access information should
          be created in `open-cluster-management` namespace. You can run the following
          command to create the secret: \n\n```\nkubectl create secret generic transport-secret
          -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n  --from-literal=CA=<CA-for-kafka-server>\n```\n >_Note:_ There is a
          sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `transport-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n## How to Install\n\nInstall the
          Multicluster Global Hub Operator by following the instructions that are
          displayed after you select the Install button. The operator must be installed
          in the same namespace as Red Hat Advanced Cluster Management for Kubernetes
          (RHACM), by default, in `open-cluster-management` namespace.\n\nA pod will
          be created in `open-cluster-management` namespace\n\n```\n$ kubectl get
          pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          19s\n```\n\nThe operator is
          now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installing the operator,
          create an instance of the MulticlusterGlobalHub resource to install the
          Multicluster Global Hub.\n\nThe pods of Multicluster Global Hub are created
          in the `open-cluster-management` namespace:\n\n```\n$ kubectl get pods -n
          open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: multicluster-global-hub-operator.v0.6.0
        version: 0.6.0
      name: release-2.7
    - currentCSV: multicluster-global-hub-operator.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha3",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "largeScale": {
                      "kafka": {
                        "transportFormat": "cloudEvents"
                      }
                    },
                    "type": "largeScale"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.7.0
          createdAt: "2023-07-19T07:53:07Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha3
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.7 or later needs to be installed.\n## How to Install\n\nInstall the Multicluster
          Global Hub Operator by following the instructions that are displayed after
          you select the Install button. The operator must be installed in the same
          namespace as Red Hat Advanced Cluster Management for Kubernetes (RHACM),
          by default, in `open-cluster-management` namespace.\n\nA pod will be created
          in `open-cluster-management` namespace\n\n```\n$ kubectl get pods -n open-cluster-management\nNAME
          \                                                            READY   STATUS
          \   RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp                1/1
          \    Running   0          19s\n```\n\nThe operator is now providing new
          Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installed the operator,
          create an instance of the MulticlusterGlobalHub resource to instantiate
          the Multicluster Global Hub.\n\nYou need to prepare:\n\n- PostgreSQL needs
          to be installed and you must create database for the multicluster global
          hub. You must create a secret named multicluster-global-hub-storage that
          contains the database access credential in `open-cluster-management` namespace.
          You can run the following command to create the secret: \n\n```\nkubectl
          create secret generic multicluster-global-hub-storage -n open-cluster-management
          \\\n  --from-literal=database_uri=<postgresql-uri> \\\n --from-file=ca.crt=<CA-for-postgres-server>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)(Note:
          the client version of kubectl must be v1.21+) to install postgres in `hoh-postgres`
          namespace and create the secret `multicluster-global-hub-storage` in namespace
          `open-cluster-management` automatically. To override the secret namespace,
          set `TARGET_NAMESPACE` environment variable to the RHACM installation namespace
          before executing the script.\n\n- Kafka needs to be installed, and you must
          create three topics, `spec` `status` and `event`. You need to create a secret
          named multicluster-global-hub-transport that contains the kafka access information
          should be created in `open-cluster-management` namespace. You can run the
          following command to create the secret: \n\n```\nkubectl create secret generic
          multicluster-global-hub-transport -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n      --from-file=ca.crt=<CA-cert-for-kafka-server> --from-file=client.crt=<Client-cert-for-kafka-server>
          --from-file=client.key=<Client-key-for-kafka-server> \n```\n >_Note:_ There
          is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `multicluster-global-hub-transport`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\nOnce an instance of the MulticlusterGlobalHub
          is created, the following pods are created in the `open-cluster-management`
          namespace:\n\n```\n$ kubectl get pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-grafana-79f8d59f58-qwcjl
          \                1/1     Running   0          22s\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: beta
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.7.0
        version: 0.7.0
      entries:
      - name: multicluster-global-hub-operator.v0.7.0
        version: 0.7.0
      name: release-2.8
    - currentCSV: multicluster-global-hub-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v1.0.0
          createdAt: "2023-10-16T03:07:33Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the MultiCluster Global Hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.7 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```

          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v1.0.0
        version: 1.0.0
      entries:
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.9
    defaultChannel: release-1.5
    packageName: multicluster-global-hub-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/stolostron/multicluster-global-hub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhbk-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhbk-operator.v22.0.13-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
          createdAt: "2024-09-17T11:20:22Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:884afe0ca1698d85a02aa47b6d60b8222e50d70ad280db6f98e1a4bbdbe13c40
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
        version: 22.0.13-opr.1
      entries:
      - name: rhbk-operator.v22.0.13-opr.1
        version: 22.0.13-opr.1
      - name: rhbk-operator.v22.0.12-opr.1
        version: 22.0.12-opr.1
      - name: rhbk-operator.v22.0.11-opr.2
        version: 22.0.11-opr.2
      - name: rhbk-operator.v22.0.11-opr.1
        version: 22.0.11-opr.1
      - name: rhbk-operator.v22.0.10-opr.1
        version: 22.0.10-opr.1
      - name: rhbk-operator.v22.0.9-opr.1
        version: 22.0.9-opr.1
      - name: rhbk-operator.v22.0.8-opr.2
        version: 22.0.8-opr.2
      - name: rhbk-operator.v22.0.8-opr.1
        version: 22.0.8-opr.1
      - name: rhbk-operator.v22.0.7-opr.1
        version: 22.0.7-opr.1
      - name: rhbk-operator.v22.0.6-opr.1
        version: 22.0.6-opr.1
      name: stable-v22
    - currentCSV: rhbk-operator.v22.0.13-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
          createdAt: "2024-09-17T11:20:22Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:884afe0ca1698d85a02aa47b6d60b8222e50d70ad280db6f98e1a4bbdbe13c40
        version: 22.0.13-opr.1
      entries:
      - name: rhbk-operator.v22.0.13-opr.1
        version: 22.0.13-opr.1
      - name: rhbk-operator.v22.0.12-opr.1
        version: 22.0.12-opr.1
      - name: rhbk-operator.v22.0.11-opr.2
        version: 22.0.11-opr.2
      - name: rhbk-operator.v22.0.11-opr.1
        version: 22.0.11-opr.1
      - name: rhbk-operator.v22.0.10-opr.1
        version: 22.0.10-opr.1
      - name: rhbk-operator.v22.0.9-opr.1
        version: 22.0.9-opr.1
      - name: rhbk-operator.v22.0.8-opr.2
        version: 22.0.8-opr.2
      - name: rhbk-operator.v22.0.8-opr.1
        version: 22.0.8-opr.1
      - name: rhbk-operator.v22.0.7-opr.1
        version: 22.0.7-opr.1
      - name: rhbk-operator.v22.0.6-opr.1
        version: 22.0.6-opr.1
      name: stable-v22.0
    - currentCSV: rhbk-operator.v24.0.11-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
          createdAt: "2025-03-26T02:27:51Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:d339ecf788d1681ba9945a2ea63ca675caec863ec75f497ef916103f4c898003
        version: 24.0.11-opr.1
      entries:
      - name: rhbk-operator.v24.0.11-opr.1
        version: 24.0.11-opr.1
      - name: rhbk-operator.v24.0.10-opr.1
        version: 24.0.10-opr.1
      - name: rhbk-operator.v24.0.9-opr.1
        version: 24.0.9-opr.1
      - name: rhbk-operator.v24.0.8-opr.1
        version: 24.0.8-opr.1
      - name: rhbk-operator.v24.0.7-opr.1
        version: 24.0.7-opr.1
      - name: rhbk-operator.v24.0.6-opr.1
        version: 24.0.6-opr.1
      - name: rhbk-operator.v24.0.5-opr.1
        version: 24.0.5-opr.1
      - name: rhbk-operator.v24.0.4-opr.1
        version: 24.0.4-opr.1
      - name: rhbk-operator.v24.0.3-opr.1
        version: 24.0.3-opr.1
      name: stable-v24
    - currentCSV: rhbk-operator.v24.0.11-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
          createdAt: "2025-03-26T02:27:51Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:d339ecf788d1681ba9945a2ea63ca675caec863ec75f497ef916103f4c898003
        version: 24.0.11-opr.1
      entries:
      - name: rhbk-operator.v24.0.11-opr.1
        version: 24.0.11-opr.1
      - name: rhbk-operator.v24.0.10-opr.1
        version: 24.0.10-opr.1
      - name: rhbk-operator.v24.0.9-opr.1
        version: 24.0.9-opr.1
      - name: rhbk-operator.v24.0.8-opr.1
        version: 24.0.8-opr.1
      - name: rhbk-operator.v24.0.7-opr.1
        version: 24.0.7-opr.1
      - name: rhbk-operator.v24.0.6-opr.1
        version: 24.0.6-opr.1
      - name: rhbk-operator.v24.0.5-opr.1
        version: 24.0.5-opr.1
      - name: rhbk-operator.v24.0.4-opr.1
        version: 24.0.4-opr.1
      - name: rhbk-operator.v24.0.3-opr.1
        version: 24.0.3-opr.1
      name: stable-v24.0
    - currentCSV: rhbk-operator.v26.2.9-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
          createdAt: "2025-09-18T15:03:55Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:50a62b46af44ce0ff7e6c3a15e1bfb2a9611ba4a8b65b8d0ab65817064692c55
        version: 26.2.9-opr.1
      entries:
      - name: rhbk-operator.v26.2.9-opr.1
        version: 26.2.9-opr.1
      - name: rhbk-operator.v26.2.8-opr.1
        version: 26.2.8-opr.1
      - name: rhbk-operator.v26.2.7-opr.1
        version: 26.2.7-opr.1
      - name: rhbk-operator.v26.2.6-opr.1
        version: 26.2.6-opr.1
      - name: rhbk-operator.v26.2.5-opr.1
        version: 26.2.5-opr.1
      - name: rhbk-operator.v26.2.4-opr.2
        version: 26.2.4-opr.2
      - name: rhbk-operator.v26.2.4-opr.1
        version: 26.2.4-opr.1
      - name: rhbk-operator.v26.0.15-opr.1
        version: 26.0.15-opr.1
      - name: rhbk-operator.v26.0.14-opr.1
        version: 26.0.14-opr.1
      - name: rhbk-operator.v26.0.13-opr.1
        version: 26.0.13-opr.1
      - name: rhbk-operator.v26.0.12-opr.1
        version: 26.0.12-opr.1
      - name: rhbk-operator.v26.0.11-opr.1
        version: 26.0.11-opr.1
      - name: rhbk-operator.v26.0.10-opr.1
        version: 26.0.10-opr.1
      - name: rhbk-operator.v26.0.9-opr.1
        version: 26.0.9-opr.1
      - name: rhbk-operator.v26.0.8-opr.1
        version: 26.0.8-opr.1
      - name: rhbk-operator.v26.0.7-opr.1
        version: 26.0.7-opr.1
      - name: rhbk-operator.v26.0.6-opr.1
        version: 26.0.6-opr.1
      - name: rhbk-operator.v26.0.5-opr.1
        version: 26.0.5-opr.1
      name: stable-v26
    - currentCSV: rhbk-operator.v26.0.15-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:fc5fc8cfb2c100c5b3a33fc11a78dfca569f5f34f45d88439340392d2c60afb2
          createdAt: "2025-08-28T19:34:30Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:fc5fc8cfb2c100c5b3a33fc11a78dfca569f5f34f45d88439340392d2c60afb2
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:4bb4d3070fe118bdd7de24e8b68133b8b6e158cea48c9f1abad2b4297be0dfdb
        version: 26.0.15-opr.1
      entries:
      - name: rhbk-operator.v26.0.15-opr.1
        version: 26.0.15-opr.1
      - name: rhbk-operator.v26.0.14-opr.1
        version: 26.0.14-opr.1
      - name: rhbk-operator.v26.0.13-opr.1
        version: 26.0.13-opr.1
      - name: rhbk-operator.v26.0.12-opr.1
        version: 26.0.12-opr.1
      - name: rhbk-operator.v26.0.11-opr.1
        version: 26.0.11-opr.1
      - name: rhbk-operator.v26.0.10-opr.1
        version: 26.0.10-opr.1
      - name: rhbk-operator.v26.0.9-opr.1
        version: 26.0.9-opr.1
      - name: rhbk-operator.v26.0.8-opr.1
        version: 26.0.8-opr.1
      - name: rhbk-operator.v26.0.7-opr.1
        version: 26.0.7-opr.1
      - name: rhbk-operator.v26.0.6-opr.1
        version: 26.0.6-opr.1
      - name: rhbk-operator.v26.0.5-opr.1
        version: 26.0.5-opr.1
      name: stable-v26.0
    - currentCSV: rhbk-operator.v26.2.9-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
          createdAt: "2025-09-18T15:03:55Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:50a62b46af44ce0ff7e6c3a15e1bfb2a9611ba4a8b65b8d0ab65817064692c55
        version: 26.2.9-opr.1
      entries:
      - name: rhbk-operator.v26.2.9-opr.1
        version: 26.2.9-opr.1
      - name: rhbk-operator.v26.2.8-opr.1
        version: 26.2.8-opr.1
      - name: rhbk-operator.v26.2.7-opr.1
        version: 26.2.7-opr.1
      - name: rhbk-operator.v26.2.6-opr.1
        version: 26.2.6-opr.1
      - name: rhbk-operator.v26.2.5-opr.1
        version: 26.2.5-opr.1
      - name: rhbk-operator.v26.2.4-opr.2
        version: 26.2.4-opr.2
      - name: rhbk-operator.v26.2.4-opr.1
        version: 26.2.4-opr.1
      name: stable-v26.2
    defaultChannel: stable-v26.2
    packageName: rhbk-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-opensearchserverless-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-opensearchserverless-controller.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/opensearchserverless-controller:0.2.1
          createdAt: "2025-09-20T21:43:17Z"
          description: AWS OpenSearchServerless controller is a service controller
            for managing OpenSearchServerless resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Collection represents the state of an AWS opensearchserverless
              Collection resource.
            displayName: Collection
            kind: Collection
            name: collections.opensearchserverless.services.k8s.aws
            version: v1alpha1
          - description: SecurityPolicy represents the state of an AWS opensearchserverless
              SecurityPolicy resource.
            displayName: SecurityPolicy
            kind: SecurityPolicy
            name: securitypolicies.opensearchserverless.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon OpenSearchServerless resources in AWS from within your Kubernetes cluster.

          **About Amazon OpenSearchServerless**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon OpenSearchServerless
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opensearchserverless
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon OpenSearchServerless Developer Resources
          url: https://aws.amazon.com/OpenSearchServerless/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: opensearchserverless maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/opensearchserverless-controller:0.2.1
        version: 0.2.1
      entries:
      - name: ack-opensearchserverless-controller.v0.2.1
        version: 0.2.1
      - name: ack-opensearchserverless-controller.v0.2.0
        version: 0.2.0
      - name: ack-opensearchserverless-controller.v0.1.2
        version: 0.1.2
      - name: ack-opensearchserverless-controller.v0.1.1
        version: 0.1.1
      - name: ack-opensearchserverless-controller.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ack-opensearchserverless-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Automotive SIG CentOS
      provider-url: ""
    name: automotive-infra
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: automotive-infra.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "ContainerBuild",
                "metadata": {
                  "name": "cb-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "ImageBuild",
                "metadata": {
                  "name": "imagebuild-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "OperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "automotive-infra"
                  },
                  "name": "operatorconfig-sample",
                  "namespace": "automotive-infra-operator"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/fedora-sig-robotics/infra-ci-operator:v0.0.10
          createdAt: "2025-03-31T14:06:33Z"
          description: Streamline CI/CD for RHIVOS and its ecosystem using OpenShift
            Pipelines
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: automotive-infra-operator
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator
          support: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Client
            name: clients.jumpstarter.dev
            version: v1alpha1
          - kind: ContainerBuild
            name: containerbuilds.automotive.sig.centos.org
            version: v1
          - kind: ExporterAccessPolicy
            name: exporteraccesspolicies.jumpstarter.dev
            version: v1alpha1
          - kind: Exporter
            name: exporters.jumpstarter.dev
            version: v1alpha1
          - kind: ImageBuild
            name: imagebuilds.automotive.sig.centos.org
            version: v1
          - kind: Lease
            name: leases.jumpstarter.dev
            version: v1alpha1
          - kind: OperatorConfig
            name: operatorconfigs.automotive.sig.centos.org
            version: v1
        description: This operator is designed to streamline CI/CD for RHIVOS and
          its ecosystem using OpenShift Pipelines
        displayName: Automotive-CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Automotive
        - RHIVOS
        - Pipelines
        - CI
        links:
        - name: Automotive infra Operator
          url: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator
        maintainers:
        - email: automotive-sig@lists.centos.org
          name: Automotive SIG CentOS Community
        maturity: alpha
        provider:
          name: Automotive SIG CentOS
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/fedora-sig-robotics/infra-ci-operator:v0.0.10
        version: 0.0.10
      entries:
      - name: automotive-infra.v0.0.10
        version: 0.0.10
      - name: automotive-infra.v0.0.9
        version: 0.0.9
      - name: automotive-infra.v0.0.8
        version: 0.0.8
      - name: automotive-infra.v0.0.7
        version: 0.0.7
      - name: automotive-infra.v0.0.6
        version: 0.0.6
      - name: automotive-infra.v0.0.5
        version: 0.0.5
      - name: automotive-infra.v0.0.4
        version: 0.0.4
      - name: automotive-infra.v0.0.3
        version: 0.0.3
      - name: automotive-infra.v0.0.2
        version: 0.0.2
      - name: automotive-infra.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: automotive-infra
    provider:
      name: Automotive SIG CentOS
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app.kubernetes.io/instance: ibm-licensing-operator
      app.kubernetes.io/managed-by: ibm-licensing-operator
      app.kubernetes.io/name: ibm-licensing
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      intent: projected
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-licensing-operator-app
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-licensing-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "IBMLicensing",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-licensing-operator",
                    "app.kubernetes.io/managed-by": "ibm-licensing-operator",
                    "app.kubernetes.io/name": "ibm-licensing"
                  },
                  "name": "instance"
                },
                "spec": {
                  "version": "4.2.14",
                  "apiSecretToken": "ibm-licensing-token",
                  "datasource": "datacollector",
                  "httpsEnable": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
          createdAt: "2025-02-25T12:16:11Z"
          description: The IBM Licensing Operator provides a Kubernetes CRD-Based
            API to monitor the license usage of products.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-licensing-operator-app
          olm.skipRange: '>=1.0.0 <4.2.14'
          operatorframework.io/suggested-namespace: ibm-licensing
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/internal-objects: '["ibmlicensingmetadatas.operator.ibm.com",
            "ibmlicensingdefinitions.operator.ibm.com", "ibmlicensingquerysources.operator.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'IBMLicensing custom resource is used to create an instance
              of the License Service, used to collect information about license usage
              of IBM containerized products and IBM Cloud Paks per cluster. You can
              retrieve license usage data through a dedicated API call and generate
              an audit snapshot on demand. Documentation: For additional details regarding
              install parameters check: https://ibm.biz/icpfs39install. License: Please
              refer to the IBM Terms website (ibm.biz/lsvc-lic) to find the license
              terms for the particular IBM product for which you are deploying this
              component.'
            displayName: IBM License Service
            kind: IBMLicensing
            name: ibmlicensings.operator.ibm.com
            version: v1alpha1
          - kind: IBMLicensingDefinition
            name: ibmlicensingdefinitions.operator.ibm.com
            version: v1
          - description: 'IBMLicensingMetadata is the schema for IBM License Service.
              Thanks to IBMLicensingMetadata, you can track the license usage of Virtual
              Processor Core (VPC) metric by Pods that are managed by automation,
              for which licensing annotations are not defined at the time of deployment,
              such as the open source OpenLiberty. / For more information, see documentation:
              https://ibm.biz/icpfs39install. By installing the operator for IBM License
              Service, you accept the license terms: https://ibm.biz/icpfs39license.'
            displayName: IBMLicensing Metadata
            kind: IBMLicensingMetadata
            name: ibmlicensingmetadatas.operator.ibm.com
            version: v1alpha1
          - kind: IBMLicensingQuerySource
            name: ibmlicensingquerysources.operator.ibm.com
            version: v1
        description: |-
          **Important:**
          - If you are using the IBM Licensing Operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For the link to your IBM Cloud Pak documentation, see [IBM Cloud Paks that use Common Services](https://ibm.biz/BdyGwb).
          - If you are using a stand-alone IBM Container Software, you can use the IBM Licensing Operator directly. For more information, see [ibm-licensing-operator for stand-alone IBM Containerized Software](https://ibm.biz/BdyGwh).

          **IBM Licensing Operator overview**

          IBM Licensing Operator installs License Service. You can use License Service to collect information about license usage of IBM Containerized products and IBM Cloud Paks per cluster. You can retrieve license usage data through a dedicated API call and generate an audit snapshot on demand.

          **Supported platforms**

           Red Hat OpenShift Container Platform 4.10 or newer installed on Linux x86_64, Linux on Power (ppc64le), Linux on IBM Z and LinuxONE.

          **Prerequisites**

           Prerequisites depend on the integration of the License Service with an IBM Cloud Pak or IBM Containerized Software. For more information, see the applicable IBM Cloud Pak documentation or [ibm-licensing-operator for stand-alone IBM Containerized Software](https://ibm.biz/BdyGwh).
        displayName: IBM Licensing
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBMLicensing
        - IBM
        - Cloud
        - License
        - Service
        - License Service
        links:
        - name: GitHub
          url: https://github.com/IBM/ibm-licensing-operator
        maintainers:
        - email: talk2sam@us.ibm.com
          name: talk2sam
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:ec137115210dae9b69cb2899ed8a5437d94c8a658ae1ae5eb295232b34b0925e
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:96e6db93b2d16ba590252d6568b68f254548aaed73ec17e4bdd8d530ea668be6
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9186c350fa54e8b792d72e47be6b54a67b610f726175c9641795010d45ad51f9
        version: 4.2.14
      entries:
      - name: ibm-licensing-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-licensing-operator-app
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator-rhmp.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/dynatrace-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/dynatrace-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator-rhmp.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator-rhmp.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator-rhmp.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator-rhmp.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator-rhmp.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator-rhmp.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator-rhmp.v1.5.0
        version: 1.5.0
      - name: dynatrace-operator-rhmp.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator-rhmp.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator-rhmp.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator-rhmp
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Argo CD Community
      provider-url: ""
    name: argocd-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: argocd-operator.v0.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd-sample"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCDExport",
                "metadata": {
                  "name": "argocdexport-sample"
                },
                "spec": {
                  "argocd": "argocd-sample"
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd-sample"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/argoprojlabs/argocd-operator@sha256:223ad82fb697471ede79201c4a042115c9ca5c00a172d5bdcd5a780e58c46c79
          createdAt: "2025-09-11T06:29:33Z"
          description: Argo CD is a declarative, GitOps continuous delivery tool for
            Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/argoproj-labs/argocd-operator
          support: Argo CD
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: An ApplicationSet is a group or set of Application resources.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: ArgoCDExport is the Schema for the argocdexports API
            displayName: Argo CDExport
            kind: ArgoCDExport
            name: argocdexports.argoproj.io
            version: v1alpha1
          - description: ArgoCD is the Schema for the argocds API
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: ArgoCD is the Schema for the argocds API
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: NotificationsConfiguration is the Schema for the notificationsconfiguration
              API
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
        description: |
          ## Overview

          The Argo CD Operator manages the full lifecycle for [Argo CD](https://argoproj.github.io/argo-cd/) and it's
          components. The operator's goal is to automate the tasks required when operating an Argo CD cluster.

          Beyond installation, the operator helps to automate the process of upgrading, backing up and restoring as needed and
          remove the human as much as possible. In addition, the operator aims to provide deep insights into the Argo CD
          environment by configuring Prometheus and Grafana to aggregate, visualize and expose the metrics already exported by
          Argo CD.

          The operator aims to provide the following, and is a work in progress.

          * Easy configuration and installation of the Argo CD components with sane defaults to get up and running quickly.
          * Provide seamless upgrades to the Argo CD components.
          * Ability to back up and restore an Argo CD cluster from a point in time or on a recurring schedule.
          * Aggregate and expose the metrics for Argo CD and the operator itself using Prometheus and Grafana.
          * Autoscale the Argo CD components as necessary to handle variability in demand.

          ## Usage

          Deploy a basic Argo CD cluster by creating a new ArgoCD resource in the namespace where the operator is installed.

          ```
          apiVersion: argoproj.io/v1alpha1
          kind: ArgoCD
          metadata:
            name: example-argocd
          spec: {}
          ```

          ## Backup

          Backup the cluster above by creating a new ArgoCDExport resource in the namespace where the operator is installed.

          ```
          apiVersion: argoproj.io/v1alpha1
          kind: ArgoCDExport
          metadata:
            name: example-argocdexport
          spec:
            argocd: example-argocd
          ```

          See the [documentation](https://argocd-operator.readthedocs.io) and examples on
          [GitHub](https://github.com/argoproj-labs/argocd-operator) for more information.
        displayName: Argo CD
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - kubernetes
        links:
        - name: Argo CD Project
          url: https://argoproj.github.io/argo-cd/
        - name: Operator Documentation
          url: https://argocd-operator.readthedocs.io
        - name: Operator Source Code
          url: https://github.com/argoproj-labs/argocd-operator
        maintainers:
        - email: aveerama@redhat.com
          name: Abhishek Veeramalla
        maturity: alpha
        provider:
          name: Argo CD Community
        relatedImages:
        - quay.io/argoprojlabs/argocd-operator@sha256:223ad82fb697471ede79201c4a042115c9ca5c00a172d5bdcd5a780e58c46c79
        version: 0.15.0
      entries:
      - name: argocd-operator.v0.15.0
        version: 0.15.0
      - name: argocd-operator.v0.14.1
        version: 0.14.1
      - name: argocd-operator.v0.14.0
        version: 0.14.0
      - name: argocd-operator.v0.13.0
        version: 0.13.0
      - name: argocd-operator.v0.12.0
        version: 0.12.0
      - name: argocd-operator.v0.11.0
        version: 0.11.0
      - name: argocd-operator.v0.10.1
        version: 0.10.1
      - name: argocd-operator.v0.10.0
        version: 0.10.0
      - name: argocd-operator.v0.9.1
        version: 0.9.1
      - name: argocd-operator.v0.9.0
        version: 0.9.0
      - name: argocd-operator.v0.8.0
        version: 0.8.0
      - name: argocd-operator.v0.7.0
        version: 0.7.0
      - name: argocd-operator.v0.5.0
        version: 0.5.0
      - name: argocd-operator.v0.4.0
        version: 0.4.0
      - name: argocd-operator.v0.3.0
        version: 0.3.0
      - name: argocd-operator.v0.2.1
        version: 0.2.1
      - name: argocd-operator.v0.2.0
        version: 0.2.0
      - name: argocd-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: argocd-operator
    provider:
      name: Argo CD Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-framework
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-framework.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Framework",
                "metadata": {
                  "name": "framework-sample"
                },
                "spec": {
                  "kcos-eula": {
                    "image": {
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    },
                    "service": {
                      "nodePort": 0,
                      "type": "ClusterIP"
                    }
                  },
                  "kcos-logging-framework": {
                    "image": {
                      "remoteRepository": "docker.io",
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    },
                    "ingress": {
                      "annotations": {
                        "nginx.ingress.kubernetes.io/auth-method": "POST",
                        "nginx.ingress.kubernetes.io/auth-response-headers": "X-AUTH-USER",
                        "nginx.ingress.kubernetes.io/auth-url": "https://127.0.0.1/token/validate",
                        "nginx.ingress.kubernetes.io/cors-allow-credentials": "true",
                        "nginx.ingress.kubernetes.io/cors-allow-methods": "PUT, GET, POST, OPTIONS, PATCH",
                        "nginx.ingress.kubernetes.io/cors-allow-origin": "https://admin.example.com,https://localhost:4200",
                        "nginx.ingress.kubernetes.io/enable-cors": "true",
                        "nginx.ingress.kubernetes.io/proxy-body-size": "0",
                        "nginx.ingress.kubernetes.io/ssl-redirect": "true"
                      },
                      "enabled": true
                    }
                  },
                  "kcos-postgresql": {
                    "jobCopySecret": {
                      "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                    },
                    "postgresql": {
                      "image": {
                        "registry": "prod.harbor.keysight.digital/loadcore"
                      },
                      "initdbScripts": {
                        "init_script.sh": "#!/bin/sh\nexport PGUSER=postgres\npsql -c \"SELECT 1 FROM pg_database WHERE datname = 'db'\" | grep -q 1 || psql -c \"CREATE DATABASE db\"\npsql -c \"SELECT 1 FROM pg_database WHERE datname = 'kcoslicensingdb'\" | grep -q 1 || psql -c \"CREATE DATABASE kcoslicensingdb\"\n"
                      },
                      "metrics": {
                        "image": {
                          "registry": "docker.io"
                        }
                      },
                      "persistence": {
                        "storageClass": "managed-loadcore-storage"
                      },
                      "pgHbaConfiguration": "local all all trust\nhost all all 0.0.0.0/0 trust",
                      "securityContext": {
                        "enabled": false
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": false
                        }
                      },
                      "volumePermissions": {
                        "image": {
                          "registry": "docker.io"
                        },
                        "securityContext": {
                          "runAsUser": "auto"
                        }
                      }
                    }
                  },
                  "kcos-system-diagnostics": {
                    "environment": {
                      "ELASTICSEARCH_ENDPOINT": "elasticsearch.keysight-wap.svc.cluster.local.",
                      "IGNORE_DATA_PREFIX": "wireless_stats"
                    },
                    "image": {
                      "initContainer": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf",
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    }
                  },
                  "subcharts": {
                    "enable": {
                      "kcos-coredns-fixup": false,
                      "kcos-eula": true,
                      "kcos-local-storage": false,
                      "kcos-logging-framework": true,
                      "kcos-logging-rotate": false,
                      "kcos-naas": false,
                      "kcos-postgresql": true,
                      "kcos-shell": false,
                      "kcos-system-diagnostics": true,
                      "kcos-vital-service": false,
                      "nfs-server-provisioner": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:27:28Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Framework
            name: frameworks.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Framework Operator
        displayName: Keysight KCOS Framework Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Framework
          url: https://keysight-kcos-framework.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-framework@sha256:531f278def85e33d453479fce74127e12e3ed62a251872eab421c6561bedd0c5
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        version: 0.1.21
      entries:
      - name: keysight-kcos-framework.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-framework
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx-openshift-operator.v0.0.36
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
          createdAt: "2019-07-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX on JBoss EAP 7
            displayName: Joget DX on JBoss EAP 7
            kind: JogetDX
            name: jogetdx.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX7/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
        - registry.connect.redhat.com/joget/joget-dx7-eap7@sha256:7245501f5fccf011d75c034b7d55542cfe3bd2c64cc7e48b3224a48b623af282
        version: 0.0.36
      entries:
      - name: joget-dx-openshift-operator.v0.0.36
        version: 0.0.36
      - name: joget-dx-openshift-operator.v0.0.35
        version: 0.0.35
      - name: joget-dx-openshift-operator.v0.0.34
        version: 0.0.34
      - name: joget-dx-openshift-operator.v0.0.33
        version: 0.0.33
      - name: joget-dx-openshift-operator.v0.0.32
        version: 0.0.32
      - name: joget-dx-openshift-operator.v0.0.31
        version: 0.0.31
      - name: joget-dx-openshift-operator.v0.0.30
        version: 0.0.30
      - name: joget-dx-openshift-operator.v0.0.29
        version: 0.0.29
      - name: joget-dx-openshift-operator.v0.0.28
        version: 0.0.28
      - name: joget-dx-openshift-operator.v0.0.27
        version: 0.0.27
      - name: joget-dx-openshift-operator.v0.0.26
        version: 0.0.26
      - name: joget-dx-openshift-operator.v0.0.25
        version: 0.0.25
      - name: joget-dx-openshift-operator.v0.0.24
        version: 0.0.24
      - name: joget-dx-openshift-operator.v0.0.23
        version: 0.0.23
      - name: joget-dx-openshift-operator.v0.0.22
        version: 0.0.22
      - name: joget-dx-openshift-operator.v0.0.21
        version: 0.0.21
      - name: joget-dx-openshift-operator.v0.0.20
        version: 0.0.20
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Habana Labs Ltd.
      provider-url: https://habana.ai
    name: habana-ai-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: habana-ai-operator.v1.19.2-32-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.19.2-32"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.19.2"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.19.2-32"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.19.2-32"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.19.2-32"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.19.2-32"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.19.2-32"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.19.2-32"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.19.2-32"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.19.2-32"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.19.2-32"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:975ed2135061d3bdc145c83abaa15f99e4c9f6efb15c708de41aa00fbcb72333
          createdAt: "2025-02-10T07:33:30Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.19.2-32-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi AI accelerator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:d728a91a36503f07dfacd49fdbfefe8083607ca33babb5f2dfc1e9de05386454
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:f05cd78d7593b877ec8a75175a9faa72d9b991d81c74af9ecd92208fde296eca
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:170a15041fc3e3901b4e8e0b9a32d53fde11026776c17a6560d500dacb976c56
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:71d4c7181930c71c0ca4e6b696ea416a687eec6d798a43e854fcd1bfdc224a0c
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:373b9caa27a96bb7979f7a6aec246795142ce3a28c6b97a8902035ec20814fa2
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:975ed2135061d3bdc145c83abaa15f99e4c9f6efb15c708de41aa00fbcb72333
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:72cc75353615369d12a07f5488bcdc041a8435906ca6e16c53b001016108e099
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:eeb4573de20d3341a085c75e669b7d37c424e348421d7916440fbbfff12ea482
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:ac16752b53df69573da257d3a3a717a214b2eb1656779a7ef94084d74363fb46
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:b12802e7ef05273fceaef3a9987c3c4ae760714e6f333363f665eb26be9175e9
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:e667695b845cd2be887f117dac66861d2c5e74def009d2939559ae7c30574dc0
        version: 1.19.2-32-1
      entries:
      - name: habana-ai-operator.v1.19.2-32-1
        version: 1.19.2-32-1
      name: "1.19"
    - currentCSV: habana-ai-operator.v1.20.0-543-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.20.0-543"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.20.0-4"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.20.0-543-2"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.20.0-543"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.20.0-543"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.20.0-543-4"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.20.0-543"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.20.0-543-4"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:d5b37eafe48446e1ffa492e195f4519b77d4a11f6750e5c71f27c1536bdfd2cc
          createdAt: "2025-02-26T09:03:44Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.20.0-543-5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:957702b58ee66d7339a20b2a5ece783606d6ccebde627c58951d5ecd6bb1b13f
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:d14c457106b6133fc41ef727ac40a6bf2f0e5c97f21659969d1daebdc1310b43
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:d5b37eafe48446e1ffa492e195f4519b77d4a11f6750e5c71f27c1536bdfd2cc
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:317765eac4e0332a0c4f6bead5563b343a8c22e6e0eaa62276ff66be08608c30
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:68d0f44170cbf3ea15b793e0503fb07ce7e32488011fc85ccb6c9234b3b9302f
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:c5beac6209d0671d3a49b63f1cf95401055b042430eafaefc3859b5c633f0e84
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:c662de244c5900c89b918b16982298676d9a6d553360a5716174b28d4d64a635
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:a673900f08c38c9bb40c90b3cddea580afcc6a99195d816cd0b5cae9831ca5c9
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:3ea80aa8994e8b19fbbdcd18ba2c35800774da3ca4e27a2514d0f5f9676786fa
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:9ab54a84a00e8afc078eec2a7d643f09b1bd674e0e230733fbce7108e392e009
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:cba03259b33c4cb9bcd22622ab0dda962846ea0a80c727779365cf55129ad308
        version: 1.20.0-543-5
      entries:
      - name: habana-ai-operator.v1.20.0-543-5
        version: 1.20.0-543-5
      name: "1.20"
    - currentCSV: habana-ai-operator.v1.21.5-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.21.5-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.21.5-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.21.5-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:0741da99140d137474b9e7edba921672eab8ff6277e1dd8bf3a4179e87f01dd4
          createdAt: "2025-09-17T11:55:32Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.21.5-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:a216ec6a6e7df50e067c557e9ed5871a3308b2ba42a4e41c9642f0c2c4503d88
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:d7f64d15c379524597aa2b2c5b4586c890442fa32356692013cf665b01b0f092
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:256cd8bebec3ed835e187e925a14a6b605ce9b3a17fe7d8886e0b32cde163d90
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:c95d3ffd6c91be3418415740258d33a8dbcedad6fe7d6bfe68d4dd122027447b
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:9692d079ef18a364ed235c187482f329b3a5764a773e363acb1238044ad4a884
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:c84e997c7d8c22c4d4e902a29e45e5364f0ad306f17d03a7d8ddb4391142e167
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:0741da99140d137474b9e7edba921672eab8ff6277e1dd8bf3a4179e87f01dd4
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:c7147359dcd6a2e5adee95259d37743ff302b1deaeb13b184044f4029d33f44a
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:a0b45ccfb6af54b1043ccc2336ad3d7990cc447af5ba3621f3eb1953cd40ffc5
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:bc3749ff153fff71e252820cbf22fcd0753b195b2204f9731e42af29fb7ddaa0
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:1e488c8ad31784dab0fe70144f5146d5809e2fc7c86424d1b512ab887787e28f
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:bd57e7059467f0ca1667d19a70bd9a6ad96b83bbcb53c09fdad5a262fee76af9
        version: 1.21.5-6
      entries:
      - name: habana-ai-operator.v1.21.5-6
        version: 1.21.5-6
      name: "1.21"
    - currentCSV: habana-ai-operator.v1.22.1-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.22.1-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
          createdAt: "2025-09-17T10:15:21Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.22.1-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:d1974d72774c626e77707e8f44bcc631ac142bfe2a5395714969a964c963e36e
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:53c1844b5e8eb08d0cc29865a732100208c5dfc0fe55df8b32e589e319bcab6c
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:e34a6820a58e7e8090b521781c76acc3867c94f8291634de7b138f928e0bdb74
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:1925586fda34c6fa444a59b1fa7be20b3cf16e845bf56566ad2a41c276b0cfe9
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:7140877b23291a5db3e5e28def3c9d5e7cb7be62f046e2fb7ac45eff369cea84
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:c9ff6f3fe24ef87c7bfc71bdb1cd7900941df52f9ad1293badd7e14783ce7ef6
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:1e15f3de21c2b1e20f54d1f8a973290c6d940624506880c5bc3eb880011bcc02
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:8f413edfff1978b35ba3998ed5f1c6293e6764b1d2cd9faa08f0bd81dc6e3e10
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:392fcd2a4730299e6e3bab9f7e96fc3b710de8adb374a2c64b7316e097ed05f9
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:38c21eae6249e5571921998660418e59870bc0880e7b8c9b95a4a8db30ef2e7d
        - vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer@sha256:2cf151c2a518e3e1c7c09efb8301e469d266d1f030fff2e6180b086cc2e04ee7
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:704b4bebdfe6b85693df4b213f67862859a554c21ce4e614736172f64b541ea7
        version: 1.22.1-6
      entries:
      - name: habana-ai-operator.v1.22.1-6
        version: 1.22.1-6
      name: "1.22"
    - currentCSV: habana-ai-operator.v1.22.1-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.22.1-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
          createdAt: "2025-09-17T10:15:21Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.22.1-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:704b4bebdfe6b85693df4b213f67862859a554c21ce4e614736172f64b541ea7
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:8f413edfff1978b35ba3998ed5f1c6293e6764b1d2cd9faa08f0bd81dc6e3e10
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:7140877b23291a5db3e5e28def3c9d5e7cb7be62f046e2fb7ac45eff369cea84
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:38c21eae6249e5571921998660418e59870bc0880e7b8c9b95a4a8db30ef2e7d
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:d1974d72774c626e77707e8f44bcc631ac142bfe2a5395714969a964c963e36e
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:e34a6820a58e7e8090b521781c76acc3867c94f8291634de7b138f928e0bdb74
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:1925586fda34c6fa444a59b1fa7be20b3cf16e845bf56566ad2a41c276b0cfe9
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:c9ff6f3fe24ef87c7bfc71bdb1cd7900941df52f9ad1293badd7e14783ce7ef6
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:53c1844b5e8eb08d0cc29865a732100208c5dfc0fe55df8b32e589e319bcab6c
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:392fcd2a4730299e6e3bab9f7e96fc3b710de8adb374a2c64b7316e097ed05f9
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:1e15f3de21c2b1e20f54d1f8a973290c6d940624506880c5bc3eb880011bcc02
        - vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer@sha256:2cf151c2a518e3e1c7c09efb8301e469d266d1f030fff2e6180b086cc2e04ee7
        version: 1.22.1-6
      entries:
      - name: habana-ai-operator.v1.22.1-6
        version: 1.22.1-6
      name: stable
    defaultChannel: stable
    packageName: habana-ai-operator
    provider:
      name: Habana Labs Ltd.
      url: https://habana.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.16.16-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-multicluster-orchestrator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-multicluster-orchestrator.v4.15.20-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "items": [
                    {
                      "clusterName": "local-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    },
                    {
                      "clusterName": "spoke-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    }
                  ],
                  "manageS3": false,
                  "schedulingIntervals": [
                    "5m"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["odf-multicluster-console"]'
          createdAt: "2025-09-12T07:33:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.15.20-rhodf'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          Orchestrator for OpenShift Data Foundation clusters running across multiple OpenShift clusters.
          It uses Red Hat Advanced Cluster Management for Kubernetes as the multicluster control plane.
        displayName: ODF Multicluster Orchestrator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: ODF Multicluster Orchestrator
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: uchapaga@redhat.com
          name: Umanga Chapagain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-multicluster-rhel9-operator@sha256:09e2c7a20ab92ddc5d589072bb0ab9ce2ed8706ed8364f1b11192d7a7396e138
        - registry.redhat.io/odf4/odf-multicluster-console-rhel9@sha256:7527f4a14aed0386b8f160fa8a92b7ff1f58114970e07c3ec6c433d2a53459b5
        version: 4.15.20-rhodf
      entries:
      - name: odf-multicluster-orchestrator.v4.15.20-rhodf
        version: 4.15.20-rhodf
      - name: odf-multicluster-orchestrator.v4.15.19-rhodf
        version: 4.15.19-rhodf
      - name: odf-multicluster-orchestrator.v4.15.18-rhodf
        version: 4.15.18-rhodf
      - name: odf-multicluster-orchestrator.v4.15.17-rhodf
        version: 4.15.17-rhodf
      - name: odf-multicluster-orchestrator.v4.15.16-rhodf
        version: 4.15.16-rhodf
      - name: odf-multicluster-orchestrator.v4.15.15-rhodf
        version: 4.15.15-rhodf
      - name: odf-multicluster-orchestrator.v4.15.14-rhodf
        version: 4.15.14-rhodf
      - name: odf-multicluster-orchestrator.v4.15.13-rhodf
        version: 4.15.13-rhodf
      - name: odf-multicluster-orchestrator.v4.15.12-rhodf
        version: 4.15.12-rhodf
      - name: odf-multicluster-orchestrator.v4.15.11-rhodf
        version: 4.15.11-rhodf
      - name: odf-multicluster-orchestrator.v4.15.10-rhodf
        version: 4.15.10-rhodf
      - name: odf-multicluster-orchestrator.v4.15.9-rhodf
        version: 4.15.9-rhodf
      - name: odf-multicluster-orchestrator.v4.15.8-rhodf
        version: 4.15.8-rhodf
      - name: odf-multicluster-orchestrator.v4.15.6-rhodf
        version: 4.15.6-rhodf
      - name: odf-multicluster-orchestrator.v4.15.5-rhodf
        version: 4.15.5-rhodf
      - name: odf-multicluster-orchestrator.v4.15.4-rhodf
        version: 4.15.4-rhodf
      - name: odf-multicluster-orchestrator.v4.15.3-rhodf
        version: 4.15.3-rhodf
      - name: odf-multicluster-orchestrator.v4.15.2-rhodf
        version: 4.15.2-rhodf
      - name: odf-multicluster-orchestrator.v4.15.1-rhodf
        version: 4.15.1-rhodf
      - name: odf-multicluster-orchestrator.v4.15.0-rhodf
        version: 4.15.0-rhodf
      name: stable-4.15
    - currentCSV: odf-multicluster-orchestrator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "items": [
                    {
                      "clusterName": "local-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    },
                    {
                      "clusterName": "spoke-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    }
                  ],
                  "manageS3": false,
                  "schedulingIntervals": [
                    "5m"
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["odf-multicluster-console"]'
          createdAt: "2025-09-12T06:32:35Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          Orchestrator for OpenShift Data Foundation clusters running across multiple OpenShift clusters.
          It uses Red Hat Advanced Cluster Management for Kubernetes as the multicluster control plane.
        displayName: ODF Multicluster Orchestrator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: ODF Multicluster Orchestrator
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: uchapaga@redhat.com
          name: Umanga Chapagain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-multicluster-rhel9-operator@sha256:78d6bc14b95d1ee91ed529422062ec191afc489d6ab4e75906ae3c7fe58749cb
        - registry.redhat.io/odf4/odf-multicluster-console-rhel9@sha256:97d3645560388f6ae08fc12cafb58b6c766e38d83de04b645b3e6fb03fbba231
        version: 4.16.16-rhodf
      entries:
      - name: odf-multicluster-orchestrator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: odf-multicluster-orchestrator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: odf-multicluster-orchestrator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: odf-multicluster-orchestrator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: odf-multicluster-orchestrator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: odf-multicluster-orchestrator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: odf-multicluster-orchestrator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: odf-multicluster-orchestrator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: odf-multicluster-orchestrator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: odf-multicluster-orchestrator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: odf-multicluster-orchestrator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: odf-multicluster-orchestrator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: odf-multicluster-orchestrator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: odf-multicluster-orchestrator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: odf-multicluster-orchestrator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: odf-multicluster-orchestrator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: odf-multicluster-orchestrator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: odf-multicluster-orchestrator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-access-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-access-operator.v23.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyAccess",
                "metadata": {
                  "name": "isva-sample"
                },
                "spec": {
                  "image": "icr.io/isva/verify-access-wrp:10.0.5.0",
                  "instance": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isva/verify-access-operator:23.3.0
          createdAt: Mon Mar  20 00:59:34 UTC 2023
          description: The IBM Security Verify Access Operator manages the lifecycle
            of IBM Security Verify Access worker containers.
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM-Security/verify-access-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyAccess is the Schema for the ibmsecurityverifyaccesses
              API.
            displayName: IBMSecurity Verify Access
            kind: IBMSecurityVerifyAccess
            name: ibmsecurityverifyaccesses.ibm.com
            version: v1
        description: |+
          In a world of highly fragmented access management environments, [IBM Security Verify Access](https://www.ibm.com/au-en/products/verify-access) helps you simplify your users' access while more securely adopting web, mobile and cloud technologies. This solution helps you strike a balance between usability and security through the use of risk-based access, single sign-on, integrated access management control, identity federation and its mobile multi-factor authentication capability, IBM Verify. Take back control of your access management with IBM Security Verify Access.

          The IBM Security Verify Access operator provides lifecycle management of the lightweight containers which are used to protect an environment, namely:
          * [Web Reverse Proxy](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-web-reverse-proxy)
          * [Runtime](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-runtime)
          * [Distributed Session Cache](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-distributed-session-cache)

          The operator will manage the deployment of these lightweight IBM Security Verify Access worker containers, and also control the rolling restart of these containers when a configuration snapshot is updated.

          See the project [Readme](https://github.com/IBM-Security/verify-access-operator/blob/master/README.md) for further information and details.

        displayName: IBM Security Verify Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Access Product Information
          url: https://www.ibm.com/au-en/products/verify-access
        - name: Verify Access Documentation
          url: https://www.ibm.com/docs/en/sva
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Access Development Team
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - icr.io/isva/verify-access-operator:23.3.0
        version: 23.3.0
      entries:
      - name: ibm-security-verify-access-operator.v23.3.0
        version: 23.3.0
      - name: ibm-security-verify-access-operator.v22.10.0
        version: 22.10.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-access-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: oadp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oadp-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "Velero",
                "metadata": {
                  "name": "example-velero",
                  "namespace":"openshift-adp"
                },
                "spec": {
                  "backupStorageLocations": [
                    {
                      "config": {
                        "profile": "default",
                        "region": "us-east-1"
                      },
                      "name": "default",
                      "objectStorage": {
                        "bucket": "myBucket",
                        "prefix": "velero"
                      },
                      "provider": "aws"
                    }
                  ],
                  "defaultVeleroPlugins": [
                    "aws",
                    "csi",
                    "openshift"
                  ],
                  "veleroFeatureFlags": [
                    "EnableCSI"
                  ],
                  "enableRestic": true,
                  "volumeSnapshotLocations": [
                    {
                      "config": {
                        "profile": "default",
                        "region": "us-west-1"
                      },
                      "name": "default",
                      "provider": "aws"
                    }
                  ]
                }
               },
                {
                  "apiVersion": "velero.io/v1",
                  "kind": "Backup",
                  "metadata": {
                    "name": "backup",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "BackupStorageLocation",
                  "metadata": {
                    "name": "backupstoragelocation",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "DeleteBackupRequest",
                  "metadata": {
                    "name": "deletebackuprequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "DownloadRequest",
                  "metadata": {
                    "name": "downloadrequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "PodVolumeBackup",
                  "metadata": {
                    "name": "podvolumebackup",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "PodVolumeRestore",
                  "metadata": {
                    "name": "podvolumerestore",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "ResticRepository",
                  "metadata": {
                    "name": "resticrepository",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "Restore",
                  "metadata": {
                    "name": "restore",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "Schedule",
                  "metadata": {
                    "name": "schedule",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "ServerStatusRequest",
                  "metadata": {
                    "name": "serverstatusrequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "VolumeSnapshotLocation",
                  "metadata": {
                    "name": "volumesnapshotlocation",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/konveyor/oadp-operator@sha256:ecb8534fbc245e890b9c7766946d01e1a8b4ce9be359c4ada4c95fd908b679b3
          createdAt: "2020-09-08T12:21:00Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          olm.skipRange: '>=0.0.0 <0.4.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Velero is a tool to safely backup and restore, perform disaster
              recovery and migrate Kubernetes cluster resources and persistent volumes.
            displayName: Velero
            kind: Velero
            name: veleroes.oadp.openshift.io
            version: v1alpha1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - description: Backup is a Velero resource that respresents the capture
              of Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Velero on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://oadp.konveyor.io/
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/konveyor/velero-restic-restore-helper@sha256:c4d12829e42cabe70b662a659875900677dfb415f912f931c3bf4c09a26aed5b
        - quay.io/konveyor/oadp-operator@sha256:ecb8534fbc245e890b9c7766946d01e1a8b4ce9be359c4ada4c95fd908b679b3
        - quay.io/konveyor/velero-plugin-for-aws@sha256:2197c6fef32910747526d33c1462b75cb7940aeace65791d0a5ddaa038480e52
        - quay.io/konveyor/velero-plugin-for-microsoft-azure@sha256:62a344e5cf196e55dae92e537648987d758ccc75985bc45e9be6d0646a9e64d0
        - quay.io/konveyor/velero-plugin-for-gcp@sha256:3300458b1d6c1514dca7d24353ad03bd92db396b2a29f7d3522563160a5f5219
        - quay.io/konveyor/velero-plugin-for-csi@sha256:5c2261effc902d8baa16e38265f9ad1fbe5284a3e6abd2bbc97761c161fa92cf
        - quay.io/konveyor/velero@sha256:e7865cc707664eb6ee767e77ccd933420604c2562736b53e16c8043a51df9d86
        - quay.io/konveyor/openshift-velero-plugin@sha256:aec8b4fc66cfcebd5a35ce99d8c985453ebf25c11453d30256089a8b425990b5
        - quay.io/konveyor/registry@sha256:b74ab3278b79eb89f99b6970c6aa0f5306221a8495b5496faad6cab2f74ff2cd
        version: 0.4.2
      entries:
      - name: oadp-operator.v0.4.2
        version: 0.4.2
      - name: oadp-operator.v0.4.1
        version: 0.4.1
      - name: oadp-operator.v0.4.0
        version: 0.4.0
      - name: oadp-operator.v0.3.0
        version: 0.3.0
      name: beta
    - currentCSV: oadp-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "restic": {
                      "enable": true
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ResticRepository",
                "metadata": {
                  "name": "resticrepository",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/konveyor/oadp-operator@sha256:90609044d52173619ea1e63ddef0ef3e7c1c939ac737b64bd427cbed5374dd27
          createdAt: "2020-09-08T12:21:00Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          olm.skipRange: '>=0.0.0 <0.5.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is a Velero resource that respresents the capture
              of Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Data Protection Applications on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://oadp.konveyor.io/
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/konveyor/oadp-operator@sha256:90609044d52173619ea1e63ddef0ef3e7c1c939ac737b64bd427cbed5374dd27
        - quay.io/konveyor/velero-plugin-for-aws@sha256:1e8bc27d6b60000e1406085d582910cb530af3fb901ef9bc780628df6c766288
        - quay.io/konveyor/velero-plugin-for-csi@sha256:8c36ccfd294dafc415a7f113770cbbd66dbf6ecf8fef3bc4e73adc9878c606b0
        - quay.io/konveyor/registry@sha256:1be970879206258335357e3cf630138c24b8a5ccab35aae05eccb260e5e8ffb4
        - quay.io/konveyor/velero-restic-restore-helper@sha256:c16248d027c09cd5b7360ae38484df88ca1cc01f9e470e3eade95076ee1a6b84
        - quay.io/konveyor/openshift-velero-plugin@sha256:221ab0bcd1d95fc166ef7319d187e196b59d2fbf9eefd29d91bd6435d45b5c3c
        - quay.io/konveyor/velero-plugin-for-microsoft-azure@sha256:2d0089cbe715b482fb42e1acd9add70b975a8575636d8b143e6f90cef317ea13
        - quay.io/konveyor/velero-plugin-for-gcp@sha256:ee463ed2a8019c560961b118f8a87b587ed302f24fd25480c5d4ddad2557e92f
        - quay.io/konveyor/velero@sha256:0ea5336dcbf0a1fab561697b24da01a0e5b533ee23553fbfd5f8e002f5a31c3e
        version: 0.5.6
      entries:
      - name: oadp-operator.v0.5.6
        version: 0.5.6
      - name: oadp-operator.v0.5.5
        version: 0.5.5
      - name: oadp-operator.v0.5.4
        version: 0.5.4
      - name: oadp-operator.v0.5.3
        version: 0.5.3
      - name: oadp-operator.v0.5.2
        version: 0.5.2
      - name: oadp-operator.v0.5.1
        version: 0.5.1
      - name: oadp-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: oadp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Trustification
      provider-url: ""
    name: trustify-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trustify-operator.v0.1.0-alpha.9
      currentCSVDesc:
        annotations:
          alm-examples: |
            [{
              "apiVersion": "org.trustify/v1alpha1",
              "kind": "Trustify",
              "metadata": {
                "name": "myapp"
              },
              "spec": { }
            }]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          containerImage: ghcr.io/trustification/trustify-operator:v0.1.0-alpha.9
          createdAt: "2025-03-07T11:57:13+00:00"
          description: An Operator for installing and managing Trustify
          repository: https://github.com/trustification/trustify-operator
          support: https://github.com/trustification/trustify-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Trustify instance
            displayName: Trustify
            kind: Trustify
            name: trustifies.org.trustify
            version: v1alpha1
        description: |+
          Trustify is vendor-neutral, thought-leadering, mostly informational collection of resources devoted to making Software Supply Chains easier to create, manage, consume and ultimately... to trust!


          You can use Trustify for:

          - Store and Manage all the SBOM (Software Bill of Materials) files of your company.
          - Understand which are the Vulnerabilities that affect each of your SBOM files
          - Understand exactly which Packages are included/shipped within each SBOM

          Trustify is a project within the [Trustification community](https://trustification.io/).


          ### Install
          Once you have successfully installed the Operator, proceed to deploy components by creating the required CR.

          By default, the Operator installs the following components on a target cluster:

          * Server
          * UI

          ### Documentation
          Documentation can be found on our [website](https://trustification.io/).

          ### Getting help
          If you encounter any issues while using Trustify, you can create an issue on our [Github repo](https://github.com/trustification/trustify/issues), for bugs, enhancements or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find
          * Fixing issues by opening [Pull Requests](https://github.com/trustification/trustify/pulls)

        displayName: Trustify Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trust
        links:
        - name: Website
          url: https://trustification.io/
        - name: Github
          url: https://github.com/trustification/trustify
        maintainers:
        - email: trustification@googlegroups.com
          name: Trustification
        maturity: alpha
        minKubeVersion: 1.23.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Trustification
        relatedImages:
        - ghcr.io/trustification/trustify-operator:v0.1.0-alpha.9
        version: 0.1.0-alpha.9
      entries:
      - name: trustify-operator.v0.1.0-alpha.9
        version: 0.1.0-alpha.9
      - name: trustify-operator.v0.1.0-alpha.8
        version: 0.1.0-alpha.8
      - name: trustify-operator.v0.1.0-alpha.7
        version: 0.1.0-alpha.7
      - name: trustify-operator.v0.1.0-alpha.6
        version: 0.1.0-alpha.6
      - name: trustify-operator.v0.1.0-alpha.5
        version: 0.1.0-alpha.5
      - name: trustify-operator.v0.1.0-alpha.4
        version: 0.1.0-alpha.4
      - name: trustify-operator.v0.1.0-alpha.3
        version: 0.1.0-alpha.3
      - name: trustify-operator.v0.1.0-alpha.2
        version: 0.1.0-alpha.2
      name: alpha
    defaultChannel: alpha
    packageName: trustify-operator
    provider:
      name: Trustification
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-directory-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-directory-operator.v23.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyDirectory",
                "metadata": {
                  "name": "ibmsecurityverifydirectory-sample"
                },
                "spec": {
                  "pods": {
                    "configMap": {
                      "proxy": {
                        "key": "config.yaml",
                        "name": "isvd-proxy-config"
                      },
                      "server": {
                        "key": "config.yaml",
                        "name": "isvd-server-config"
                      }
                    },
                    "image": {
                      "label": "latest",
                      "repo": "icr.io/isvd"
                    },
                    "proxy": {
                      "replicas": 1
                    },
                    "serviceAccountName": "isvd"
                  },
                  "replicas": {
                    "pvcs": [
                      "replica-1",
                      "replica-2"
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: icr.io/isvd/verify-directory-operator@sha256:d1dd7037584c6818dcf8d5aa69472278833bb8e37fba2aec5a69185ba88dcc34
          createdAt: "2023-04-18T22:10:32Z"
          description: The IBM Security Verify Directory operator is used to scale
            an IBM Security Verify Directory Server environment.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ibm-security/verify-directory-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyDirectory is the Schema for the ibmsecurityverifydirectories
              API
            displayName: IBMSecurity Verify Directory
            kind: IBMSecurityVerifyDirectory
            name: ibmsecurityverifydirectories.ibm.com
            version: v1
        description: "IBM Security Verify Directory is a scalable, standards-based
          identity directory that helps simplify identity and directory management.
          Verify Directory helps consolidate identity silos into a single identity
          source. Verify Directory is purpose-built to provide a directory foundation
          that can help provide a trusted identity data infrastructure that assists
          in enabling mission-critical security and authentication. It is designed
          to deliver a reliable, scalable, standards-based identity data platform
          that interoperates with a broad range of operating systems and applications.
          Verify Directory supports Lightweight Directory Access Protocol (LDAP) V3,
          offering a flexible and highly scalable LDAP infrastructure. \n\nThe operator
          can be used to manage deployed instances of the IBM Security Verify Directory
          server, providing an easy mechanism to scale the number of servers which
          are used to handle the workload.\n\nSee the project [Readme](https://github.com/ibm-security/verify-directory-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Security Verify Directory
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Security
        - Identity and Access
        links:
        - name: IBM Security Verify Directory Documentation
          url: https://www.ibm.com/docs/en/svd
        maintainers:
        - email: iamdev@us.ibm.com
          name: IBM Security Verify Directory Development Team
        maturity: stable
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - icr.io/isvd/verify-directory-operator@sha256:d1dd7037584c6818dcf8d5aa69472278833bb8e37fba2aec5a69185ba88dcc34
        version: 23.4.1
      entries:
      - name: ibm-security-verify-directory-operator.v23.4.1
        version: 23.4.1
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-directory-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Flux project
      provider-url: ""
    name: flux
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.toolkit.fluxcd.io/v2beta1",
                "kind": "HelmRelease",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "5m",
                    "chart": {
                      "spec": {
                          "chart": "podinfo",
                          "version": "4.0.x",
                          "sourceRef": {
                            "kind": "HelmRepository",
                            "name": "podinfo",
                            "namespace": "flux-system"
                          },
                          "interval": "1m"
                      }
                    },
                    "values": {
                      "replicaCount": 2
                    }
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "GitRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "1m",
                    "ref": {
                      "branch": "master"
                    },
                    "url": "https://github.com/stefanprodan/podinfo"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmRepository",
                "metadata": {
                    "name": "stable",
                    "namespace": "default"
                },
                "spec": {
                    "url": "https://kubernetes-charts.storage.googleapis.com/",
                    "interval": "10m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmChart",
                "metadata": {
                    "name": "redis",
                    "namespace": "default"
                },
                "spec": {
                    "chart": "redis",
                    "version": "10.5.7",
                    "sourceRef": {
                      "name": "stable",
                      "kind": "HelmRepository"
                    },
                    "interval": "5m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "Bucket",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "1m",
                    "provider": "generic",
                    "bucketName": "podinfo",
                    "endpoint": "minio.minio.svc.cluster.local:9000",
                    "insecure": true,
                    "secretRef": {
                      "name": "minio-credentials"
                    }
                }
              },
              {
                "apiVersion": "kustomize.toolkit.fluxcd.io/v1beta2",
                "kind": "Kustomization",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "5m",
                    "path": "./deploy/overlays/dev",
                    "prune": true,
                    "sourceRef": {
                      "kind": "GitRepository",
                      "name": "podinfo"
                    },
                    "timeout": "2m",
                    "wait": true
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "address": "https://github.com/stefanprodan/podinfo",
                    "secretRef": {
                      "name": "api-token"
                    }
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Alert",
                "metadata": {
                    "name": "webapp",
                    "namespace": "default"
                },
                "spec": {
                    "providerRef": {
                      "name": "on-call-slack"
                    },
                    "eventSeverity": "info",
                    "eventSources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "Bucket",
                          "name": "secrets"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-backend"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-frontend"
                      }
                    ]
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Receiver",
                "metadata": {
                    "name": "github-receiver",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "events": [
                      "ping",
                      "push"
                    ],
                    "secretRef": {
                      "name": "webhook-token"
                    },
                    "resources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "HelmRepository",
                          "name": "webapp"
                      }
                    ]
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "image": "ghcr.io/stefanprodan/podinfo",
                    "interval": "1m0s"
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImagePolicy",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "imageRepositoryRef": {
                      "name": "podinfo"
                    },
                    "policy": {
                      "semver": {
                          "range": "5.0.x"
                      }
                    }
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageUpdateAutomation",
                "metadata": {
                  "name": "flux-system",
                  "namespace": "flux-system"
                },
                "spec": {
                  "interval": "1m0s",
                  "sourceRef": {
                    "kind": "GitRepository",
                    "name": "flux-system"
                  },
                  "git": {
                    "checkout": {
                      "ref": {
                        "branch": "main"
                      }
                    },
                    "commit": {
                      "author": {
                        "email": "fluxcdbot@users.noreply.github.com",
                        "name": "fluxcdbot"
                      },
                      "messageTemplate": "{{range .Updated.Images}}{{println .}}{{end}}"
                    },
                    "push": {
                      "branch": "main"
                    }
                  },
                  "update": {
                    "path": "./clusters/my-cluster",
                    "strategy": "Setters"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.connect.redhat.com/flux/source-controller@sha256:b576e3f010e3ff07e716064ab0dad042aba93f0ea5c250535dd18bd678c45337
          description: Flux is a Continuous Delivery solution for Kubernetes.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/flux/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/flux/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: flux-system
          support: The Flux community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta1
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta2
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta2
          - description: OCIRepository
            displayName: OCIRepository
            kind: OCIRepository
            name: ocirepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta2
        description: |-
          [Flux](https://fluxcd.io/) is a Continuous Delivery solution for Kubernetes.
          Flux is constructed with the [GitOps Toolkit](https://toolkit.fluxcd.io/components/), a set of composable APIs and specialized tools for keeping Kubernetes clusters in sync with sources of configuration (like Git & Helm repositories), and automating updates to configuration when there is new code to deploy.

          Flux is a CNCF incubation project made for:

          - **cluster operators** who automate provision and configuration of clusters
          - **platform engineers** who build continuous delivery for developer teams
          - **app developers** who rely on continuous delivery to get their code live

          Features:

          - Source configuration from Git and Helm repositories, and S3-compatible buckets (e.g., Minio)
          - Kustomize and Helm support
          - Event-triggered and periodic reconciliation
          - Integration with Kubernetes RBAC
          - Health assessment (clusters and workloads)
          - Dependency management (infrastructure and workloads)
          - Secrets management (builtin support for Mozilla SOPS and Cloud KMS)
          - Alerting to external systems (webhook senders)
          - External events handling (webhook receivers)
          - Automated container image updates to Git (image scanning and patching)
          - Policy-driven validation (OPA, Kyverno, admission controllers)
          - Seamless integration with Git providers (GitHub, GitLab, Bitbucket)
          - Interoperability with Cluster API (CAPI) providers

          ### Steps to support bootstraping with Flux CLI

          SecComp Profile should be removed as it requires a special privilege to get the access.
          Also, it requires to enforce `runUserAs` to the same UID provided by the images to prevent OpenShift
          to alter the user id expected by our controllers.

          To do so, please specifiy a patch using the following Kustomization file, for example.

          ```yaml
          apiVersion: kustomize.config.k8s.io/v1beta1
          kind: Kustomization
          resources:
            - gotk-components.yaml
            - gotk-sync.yaml
          patches:
            - patch: |
                apiVersion: apps/v1
                kind: Deployment
                metadata:
                  name: all
                spec:
                  template:
                    spec:
                      containers:
                        - name: manager
                          securityContext:
                            runAsUser: 65534
                            seccompProfile:
                              $patch: delete
              target:
                kind: Deployment
                labelSelector: app.kubernetes.io/part-of=flux
          ```

          Then before installing Flux, please also perform the following steps to bind Flux to the `nonroot` SecurityContextConstraint (SCC). This applies to installation via OC's OLM or via Flux CLI.

          ```sh
          # replace default "flux-system" with the project/namespace name for your flux installation
          namespace="flux-system"
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:source-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:kustomize-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:helm-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:notification-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-automation-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-reflector-controller

          ```

          To see complete steps please refer to this [documentation](https://github.com/fluxcd/website/blob/main/content/en/docs/use-cases/openshift.md).

          If you've already installed Flux, you might need to uninstall it after applying the above pre-requisite steps.

          ### Network Policies

          This version of Flux is shipped without the normally provided network Policies.
          By default all traffic is enabled, and anyone can access the source-controller
          artifacts.

          It is not recommended to run it in production as is.

          Please visit [Flux's policies](https://github.com/fluxcd/flux2/tree/main/manifests/policies)
          to deploy the needed network policies and secure your Flux setup.

          ### Breaking Changes

          - Flux 0.18.x ships with v1beta2 of the kustomize.toolkit.fluxcd.io/Kustomization CRD.
          - Flux 0.25.x is the last Flux release where Kubernetes 1.19 is supported.

          ### Flux on OpenShift Use Cases

          - [The GitOps with Flux + OpenShift](https://www.youtube.com/watch?v=8ybCtNVP5To) talk by Andrew Block (Red Hat) showing how to setup multi-tenacy on OpenShift using Flux.
          - [Flux OpenShift Use Case](https://fluxcd.io/docs/use-cases/openshift/) on FluxCD.io
        displayName: Flux
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - gitops
        - cd
        - continuous delivery
        - fluxcd
        - helm
        - kustomize
        links:
        - name: Homepage
          url: https://fluxcd.io/
        - name: Documentation
          url: https://fluxcd.io/docs/
        - name: Slack
          url: https://cloud-native.slack.com/messages/flux
        - name: Mailing List
          url: https://lists.cncf.io/g/cncf-flux-dev
        - name: Source Code
          url: https://github.com/fluxcd/flux2
        maintainers:
        - email: support@weave.works
          name: The Flux maintainers
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: The Flux project
        relatedImages:
        - registry.connect.redhat.com/flux/helm-controller@sha256:cc24a7b73b023ad86a6c3a845088d153ea290dda7cab1ec03eb85aaf814e6049
        - registry.connect.redhat.com/flux/image-automation-controller@sha256:0c51595307e14acc15f22d50b414a87513ebb03530ef4a8f7d6cd9f8b757243e
        - registry.connect.redhat.com/flux/image-reflector-controller@sha256:3bef7fde22be641304e5ca69f77150e416eba5141f8aa13f68be7f810e55aa59
        - registry.connect.redhat.com/flux/kustomize-controller@sha256:246290908d899b9f6feb920f88913bd2d4673cf852a3020714fd4152c7f78237
        - registry.connect.redhat.com/flux/notification-controller@sha256:b3d81234ba3312e0106f2f7cabc73fbe1d7bf683eda36b9ed5bbdaa43789be40
        - registry.connect.redhat.com/flux/source-controller@sha256:b576e3f010e3ff07e716064ab0dad042aba93f0ea5c250535dd18bd678c45337
        version: 2.0.1
      entries:
      - name: flux.v2.0.1
        version: 2.0.1
      name: stable
    defaultChannel: stable
    packageName: flux
    provider:
      name: The Flux project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: telemetry-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: telemetry-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "Autoscaling",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "telemetry-operator",
                    "app.kubernetes.io/instance": "autoscaling-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "autoscaling",
                    "app.kubernetes.io/part-of": "telemetry-operator"
                  },
                  "name": "autoscaling"
                },
                "spec": {
                  "aodh": {
                    "databaseAccount": "aodh",
                    "databaseInstance": "openstack",
                    "memcachedInstance": "memcached",
                    "passwordSelectors": null,
                    "secret": "osp-secret"
                  }
                }
              },
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "Ceilometer",
                "metadata": {
                  "name": "ceilometer",
                  "namespace": "openstack"
                },
                "spec": {
                  "secret": "osp-secret"
                }
              },
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "Logging",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "telemetry-operator",
                    "app.kubernetes.io/instance": "logging-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "logging",
                    "app.kubernetes.io/part-of": "telemetry-operator"
                  },
                  "name": "logging"
                },
                "spec": {
                  "cloNamespace": "openshift-logging",
                  "port": 10514
                }
              },
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "MetricStorage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "telemetry-operator",
                    "app.kubernetes.io/instance": "metricstorage-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "metricstorage",
                    "app.kubernetes.io/part-of": "telemetry-operator"
                  },
                  "name": "metricstorage-sample"
                },
                "spec": {
                  "monitoringStack": {
                    "alertingEnabled": true,
                    "scrapeInterval": "30s",
                    "storage": {
                      "persistent": {
                        "pvcStorageRequest": "20G"
                      },
                      "retention": "24h",
                      "strategy": "persistent"
                    }
                  }
                }
              },
              {
                "apiVersion": "telemetry.openstack.org/v1beta1",
                "kind": "Telemetry",
                "metadata": {
                  "name": "telemetry"
                },
                "spec": {
                  "autoscaling": {
                    "aodh": {
                      "databaseAccount": "aodh",
                      "databaseInstance": "openstack",
                      "memcachedInstance": "memcached",
                      "passwordSelectors": null,
                      "secret": "osp-secret"
                    },
                    "enabled": false,
                    "heatInstance": "heat"
                  },
                  "ceilometer": {
                    "enabled": true,
                    "secret": "osp-secret"
                  },
                  "logging": {
                    "enabled": false,
                    "port": 10514
                  },
                  "metricStorage": {
                    "enabled": false,
                    "monitoringStack": {
                      "alertingEnabled": true,
                      "scrapeInterval": "30s",
                      "storage": {
                        "persistent": {
                          "pvcStorageRequest": "20G"
                        },
                        "retention": "24h",
                        "strategy": "persistent"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:47Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operatorframework.io/suggested-namespace: openstack
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Autoscaling is the Schema for the autoscalings API
            displayName: Autoscaling
            kind: Autoscaling
            name: autoscalings.telemetry.openstack.org
            version: v1beta1
          - description: Ceilometer is the Schema for the ceilometers API
            displayName: Ceilometer
            kind: Ceilometer
            name: ceilometers.telemetry.openstack.org
            version: v1beta1
          - description: Logging is the Schema for the loggings API
            displayName: Logging
            kind: Logging
            name: loggings.telemetry.openstack.org
            version: v1beta1
          - description: MetricStorage is the Schema for the metricstorages API
            displayName: Metric Storage
            kind: MetricStorage
            name: metricstorages.telemetry.openstack.org
            version: v1beta1
          - description: Telemetry is the Schema for the telemetry API
            displayName: Telemetry
            kind: Telemetry
            name: telemetries.telemetry.openstack.org
            version: v1beta1
        description: Telemetry Operator
        displayName: Telemetry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Telemetry
        - Ceilometer
        links:
        - name: Telemetry Operator
          url: https://github.com/openstack-k8s-operators/telemetry-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-ceilometer-notification-rhel9@sha256:239e5376f65682fd644e8d9b10b27ad879747a056c8fd26ff19882e34c5bb056
        - registry.redhat.io/rhoso/openstack-ceilometer-ipmi-rhel9@sha256:86618d8f54df09210349e382188f3b6563ab96ee511fb4da6eb6dc80e4027aae
        - registry.redhat.io/rhoso/openstack-aodh-listener-rhel9@sha256:3599a25515a96b3ff4768f85fecb4bf955789fb9b5d8abc3181678062020ae64
        - registry.redhat.io/rhoso/openstack-ceilometer-compute-rhel9@sha256:b8a0412041c50e312bef39c17a9c2600e9dbb42ea380d55e58243cda641cef4f
        - registry.redhat.io/rhoso/openstack-aodh-evaluator-rhel9@sha256:16027b7d4f07385c87f97e1581d4454d4ce7ed654499720d9ba5bc6fc74fc246
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/telemetry-rhel9-operator@sha256:83962f19193f98b87c54136f02d18797a8047f65e0d75c11ce9518e50aabc050
        - registry.redhat.io/rhoso/openstack-ceilometer-central-rhel9@sha256:1c79d1edb2404a8614ea29f24e2a869695d75b1695ea0937eebc9c57250fc4ab
        - registry.redhat.io/rhoso/openstack-aodh-api-rhel9@sha256:5ab39a468ab014f2fec4cb9cd210b430ae1c1262f52b9e0f22881d6d8e9def97
        - registry.redhat.io/rhoso/openstack-aodh-notifier-rhel9@sha256:f729548ac3d718f2f5b6c85045e7c676e1656207ee54f43540d246bd97426894
        - registry.redhat.io/rhoso-operators/sg-core-rhel9@sha256:a6e4edc24409361cb770bb44fe5b475dc58748426ccc4003abad7e70f01f1b97
        - registry.redhat.io/ubi9/httpd-24@sha256:dbfbe3383e12d63df7178e4cdd5b1e6e51a84c781041b9192026ea6edb72fbf3
        version: 1.0.6
      entries:
      - name: telemetry-operator.v1.0.6
        version: 1.0.6
      - name: telemetry-operator.v1.0.5
        version: 1.0.5
      - name: telemetry-operator.v1.0.4
        version: 1.0.4
      - name: telemetry-operator.v1.0.3
        version: 1.0.3
      - name: telemetry-operator.v1.0.2
        version: 1.0.2
      - name: telemetry-operator.v1.0.1
        version: 1.0.1
      - name: telemetry-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: telemetry-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Snyk Ltd.
      provider-url: ""
    name: snyk-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: snyk-operator.v1.92.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1alpha1",
                "kind": "SnykMonitor",
                "metadata": {
                  "name": "snyk-monitor"
                },
                "spec": {
                  "clusterName": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "snyk/kubernetes-monitor",
                    "tag": "1.92.0-ubi8"
                  },
                  "integrationApi": "",
                  "monitorSecrets": "snyk-monitor",
                  "scope": "Cluster",
                  "temporaryStorageSize": "50Gi",
                  "pvc": {
                    "enabled": false,
                    "create": false,
                    "name": "snyk-monitor-pvc",
                    "storageClassName": null
                  },
                  "initContainerImage": {
                    "repository": "busybox",
                    "tag": "latest"
                  },
                  "requests": {
                    "memory": "400Mi"
                  },
                  "limits": {
                    "memory": "2Gi"
                  },
                  "nodeAffinity": {
                    "disableBetaArchNodeSelector": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Security
          certified: "false"
          containerImage: docker.io/snyk/kubernetes-operator:1.92.0
          createdAt: "2022-06-09T17:49:36Z"
          description: A Kubernetes Operator for creating and managing Snyk controller
            instances.
          olm.skipRange: <1.68.2
          repository: https://github.com/snyk/kubernetes-monitor
          support: Snyk Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a deployment of the Snyk controller, which scans
              container images in the cluster for vulnerabilities.
            displayName: Snyk Monitor
            kind: SnykMonitor
            name: snykmonitors.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          A Kubernetes Operator for creating and managing Snyk Kubernetes controller instances.

          Snyk integrates with Kubernetes, enabling you to import and test your running workloads and identify vulnerabilities in their associated images and configurations that might make those workloads less secure. Once imported, Snyk continues to monitor those workloads, identifying additional security issues as new images are deployed and the workload configuration changes.

          ## Prerequisites

          To start using the Operator and the Snyk controller that it manages, you will need to create a secret containing your Snyk integrationId and Docker config file:

          ```
          kubectl create namespace snyk-monitor
          kubectl create secret generic snyk-monitor -n snyk-monitor --from-literal=dockercfg.json={} --from-literal=integrationId=${SNYK_INTEGRATION_ID}
          ```

          Refer to the [Snyk documentation](https://support.snyk.io/hc/en-us/articles/360003916138-Kubernetes-integration-overview) for more details.
        displayName: Snyk Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - snyk
        - controller
        - monitor
        - operator
        - security
        - cluster
        - container
        - image
        - security
        - vulnerabilities
        - advisories
        - audit
        - scan
        - scanning
        links:
        - name: Website
          url: https://snyk.io
        - name: Documentation
          url: https://support.snyk.io/hc/en-us/articles/360003916138-Kubernetes-integration-overview
        maintainers:
        - email: support@snyk.io
          name: Snyk Ltd.
        maturity: stable
        provider:
          name: Snyk Ltd.
        relatedImages:
        - docker.io/snyk/kubernetes-operator:1.92.0
        version: 1.92.0
      entries:
      - name: snyk-operator.v1.92.0
        version: 1.92.0
      - name: snyk-operator.v1.90.2
        version: 1.90.2
      - name: snyk-operator.v1.90.1
        version: 1.90.1
      - name: snyk-operator.v1.88.4
        version: 1.88.4
      - name: snyk-operator.v1.85.2
        version: 1.85.2
      - name: snyk-operator.v1.82.1
        version: 1.82.1
      - name: snyk-operator.v1.81.1
        version: 1.81.1
      - name: snyk-operator.v1.80.0
        version: 1.80.0
      - name: snyk-operator.v1.73.0
        version: 1.73.0
      - name: snyk-operator.v1.72.0
        version: 1.72.0
      - name: snyk-operator.v1.70.4
        version: 1.70.4
      - name: snyk-operator.v1.70.1
        version: 1.70.1
      name: stable
    defaultChannel: stable
    packageName: snyk-operator
    provider:
      name: Snyk Ltd.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Aqua Security, Inc.
      provider-url: ""
    name: aqua
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aqua-operator.v2022.4.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaCsp",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "namespace": "aqua",
                    "version": "2022.4",
                    "requirements": true
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry",
                    "dbDiskSize": 10,
                    "databaseSecret": {
                      "key": "db-password",
                      "name": "aqua-database-password"
                    }
                  },
                  "database": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "database",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "gateway": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "gateway",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "server": {
                    "replicas": 1,
                    "service": "LoadBalancer",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "console",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "route": true,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaDatabase",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "database",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "diskSize": 10,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "image": {
                      "repository": "enforcer",
                      "registry": "registry.aquasec.com",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "IfNotPresent"
                    }
                  },
                  "runAsNonRoot": false,
                  "gateway": {
                    "host": "aqua-gateway",
                    "port": 8443
                  },
                  "token": "<<your-token>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaKubeEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "config": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "aqua-secure",
                    "imagePullSecret": "aqua-registry",
                    "kubebench_image": "docker.io/aquasec/kube-bench:v0.7.1"
                  },
                  "deploy": {
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "tag": "<<KUBE_ENFORCER_TAG>>",
                      "repository": "kube-enforcer",
                      "pullPolicy": "Always"
                    }
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "starboard-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  }
                }
              },
              {
              "apiVersion": "operator.aquasec.com/v1alpha1",
              "kind": "AquaLightning",
              "metadata": {
                "name": "aqua"
              },
              "spec": {
                "global": {
                  "gateway_address": "aqua-gateway.aqua:8443",
                  "cluster_name": "Default-cluster-name"
                },
                "common": {
                  "imagePullSecret": "aqua-registry"
                },
                "kubeEnforcer": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                  "allowAnyVersion": null,
                  "deploy": {
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "tag": "2022.4",
                      "repository": "kube-enforcer",
                      "pullPolicy": "Always"
                    }
                  },
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "starboard-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  },
                  "env": []
                },
                "enforcer": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "deploy": {
                    "image": {
                      "repository": "enforcer",
                      "registry": "registry.aquasec.com",
                      "tag": "2022.4",
                      "pullPolicy": "IfNotPresent"
                    }
                  },
                  "token": "<<ENFORCER_GROUP_TOKEN>>",
                  "env": [],
                  "secret": {
                    "name": null,
                    "key": null
                  },
                  "runAsNonRoot": null,
                  "rhcosVersion": null
                }
              }
            },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaGateway",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "gateway",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaScanner",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "replicas": 1,
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "scanner",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "runAsNonRoot": false,
                  "login": {
                    "username": "<<YOUR AQUA USER NAME>>",
                    "password": "<<YOUR AQUA USER PASSWORD>>",
                    "host": "http://aqua-server:8080",
                    "token": ""
                  }
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaServer",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "LoadBalancer",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "console",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "runAsNonRoot": false
                }
              },
              {
                  "apiVersion": "operator.aquasec.com/v1alpha1",
                  "kind": "AquaCloudConnector",
                  "metadata": {
                    "name": "aqua"
                  },
                  "spec": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "common": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "image": {
                        "repository": "aqua-cloud-connector",
                        "registry": "registry.aquasec.com",
                        "tag": "<<IMAGE TAG>>"
                      },
                      "replicas": 1
                    },
                    "runAsNonRoot": false,
                    "gateway": {
                      "host": "aqua-gateway",
                      "port": "8443"
                    },
                    "login": {
                      "host": "aqua-gateway:8443",
                      "password": "<<YOUR AQUA USER PASSWORD>>",
                      "token": "<<YOUR AQUA CLOUD CONNECTOR TOKEN>>",
                      "username": "<<YOUR AQUA USER NAME>>"
                    },
                    "tunnels": [
                      {
                        "host": "<account>.azurecr.io",
                        "port": "443"
                      },
                      {
                        "host": "gke.gcr.io",
                        "port": "443"
                      },
                      {
                        "host": "<account>.jfrog.io",
                        "port": "443"
                      },
                      {
                        "host": "1.1.1.1",
                        "port": "443"
                      },
                      {
                        "region": "us-east1",
                        "type": "ecr"
                      }
                    ]
                  }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: docker.io/aquasec/aqua-operator:2022.4.18
          createdAt: "2020-03-25T08:00:00Z"
          description: The Aqua Security Operator runs within a Openshift cluster
            and provides a means to deploy and manage Aqua Security cluster and components.
          operators.operatorframework.io/internal-objects: '["aquastarboards.aquasecurity.github.io","configauditreports.aquasecurity.github.io",
            "clusterconfigauditreports.aquasecurity.github.io"]'
          repository: https://github.com/aquasecurity/aqua-operator
          support: Aqua Security, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aqua Security CSP Deployment with Aqua Operator
            displayName: AquaCsp
            kind: AquaCsp
            name: aquacsps.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Lightning Enforcers Deployment with Aqua Operator
            displayName: AquaLightningEnforcers
            kind: AquaLightning
            name: aqualightnings.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Database Deployment with Aqua Operator
            displayName: AquaDatabase
            kind: AquaDatabase
            name: aquadatabases.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Enforcer Deployment with Aqua Operator
            displayName: AquaEnforcer
            kind: AquaEnforcer
            name: aquaenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security KubeEnforcer Deployment with Aqua Operator
            displayName: AquaKubeEnforcer
            kind: AquaKubeEnforcer
            name: aquakubeenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Gateway Deployment with Aqua Operator
            displayName: AquaGateway
            kind: AquaGateway
            name: aquagateways.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Scanner Deployment with Aqua Operator
            displayName: AquaScanner
            kind: AquaScanner
            name: aquascanners.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Server Deployment with Aqua Operator
            displayName: AquaServer
            kind: AquaServer
            name: aquaservers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Starboard for kube-enforcer.
            displayName: AquaStarboard
            kind: AquaStarboard
            name: aquastarboards.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ClusterConfigAuditReport
            kind: ClusterConfigAuditReport
            name: clusterconfigauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ConfigAuditReport
            kind: ConfigAuditReport
            name: configauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: Aqua Security Cloud Connector Deployment with Aqua Operator
            displayName: AquaCloudConnector
            kind: AquaCloudConnector
            name: aquacloudconnector.operator.aquasec.com
            version: v1alpha1
        description: |-
          The Aqua Security Operator runs within an OpenShift cluster and provides a means to deploy and manage the Aqua Security cluster and components :
          * Server (aka “console”)
          * Database (for production environments we recommend to use an external database and not the Aqua default database)
          * Gateway
          * Enforcer (aka “agent”)
          * KubeEnforcer
          * Scanner
          * CSP (package that contains the Server, Database, and Gateway)
          * LightningEnforcer (package that contains Enforcer and Kube Enforcer)

          Use the Aqua-Operator to
          * Deploy Aqua Security components on OpenShift
          * Scale up Aqua Security components with extra replicas
          * Assign metadata tags to Aqua Security components

          ## Before You Begin Using the Operator CRDs

          Obtain access to the Aqua registry - https://www.aquasec.com/about-us/contact-us/


          ## Instructions and exmaples

          Follow instructions and exmaples
          [here](https://github.com/aquasecurity/aqua-operator/blob/2022.4.0/docs/DeployOpenShiftOperator.md)
        displayName: Aqua Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aqua-security
        - scanning
        - security
        - runtime-security
        links:
        - name: Aqua Security
          url: https://www.aquasec.com/
        - name: Aqua Operator Github
          url: https://github.com/aquasecurity/aqua-operator
        maintainers:
        - email: support@aquasec.com
          name: Aqua Support, Aqua Security
        maturity: alpha
        provider:
          name: Aqua Security, Inc.
        relatedImages:
        - docker.io/aquasec/aqua-operator:2022.4.18
        version: 2022.4.18
      entries:
      - name: aqua-operator.v2022.4.18
        version: 2022.4.18
      - name: aqua-operator.v2022.4.17
        version: 2022.4.17
      name: 2022.4.0
    defaultChannel: 2022.4.0
    packageName: aqua
    provider:
      name: Aqua Security, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: The bpfman Community
      provider-url: https://bpfman.io/
    name: bpfman-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bpfman-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfapplication"
                  },
                  "name": "bpfapplication-sample"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:v0.5.6"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "kprobe": {
                        "bpffunctionname": "kprobe_counter",
                        "func_name": "try_to_wake_up",
                        "offset": 0,
                        "retprobe": false
                      },
                      "type": "Kprobe"
                    },
                    {
                      "tracepoint": {
                        "bpffunctionname": "tracepoint_kill_recorder",
                        "names": [
                          "syscalls/sys_enter_kill"
                        ]
                      },
                      "type": "Tracepoint"
                    },
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "containernames": [
                            "bpfman",
                            "bpfman-agent"
                          ],
                          "namespace": "bpfman",
                          "pods": {
                            "matchLabels": {
                              "name": "bpfman-daemon"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfNsApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfnsapplication"
                  },
                  "name": "bpfapplication-sample",
                  "namespace": "acme"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:v0.5.6"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FentryProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fentryprogram"
                  },
                  "name": "fentry-example"
                },
                "spec": {
                  "bpffunctionname": "test_fentry",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fentry:v0.5.6"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FexitProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fexitprogram"
                  },
                  "name": "fexit-example"
                },
                "spec": {
                  "bpffunctionname": "test_fexit",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fexit:v0.5.6"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "KprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kprobeprogram"
                  },
                  "name": "kprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_kprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/kprobe:v0.5.6"
                    }
                  },
                  "func_name": "try_to_wake_up",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "offset": 0,
                  "retprobe": false
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcnsprogram"
                  },
                  "name": "tc-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcprogram"
                  },
                  "name": "tc-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:v0.5.6"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxnsprogram"
                  },
                  "name": "tcx-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "tcx_next",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxprogram"
                  },
                  "name": "tcx-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "tcx_pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:v0.5.6"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TracepointProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tracepointprogram"
                  },
                  "name": "tracepoint-example"
                },
                "spec": {
                  "bpffunctionname": "enter_openat",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tracepoint:v0.5.6"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "names": [
                    "syscalls/sys_enter_openat"
                  ],
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobensprogram"
                  },
                  "name": "uprobe-example",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobeprogram"
                  },
                  "name": "uprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:v0.5.6"
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpnsprogram"
                  },
                  "name": "xdp-ns-pass-all-nodes",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpprogram"
                  },
                  "name": "xdp-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:v0.5.6"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/bpfman/bpfman-operator:v0.5.6
          createdAt: "2025-02-18T13:44:53Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: bpfman
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "bpfman",
                "labels": {
                  "pod-security.kubernetes.io/enforce": "privileged",
                  "pod-security.kubernetes.io/audit": "privileged",
                  "pod-security.kubernetes.io/warn": "privileged",
                },
                "annotations": {
                  "openshift.io/node-selector": ""
                },
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/bpfman/bpfman
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BpfApplication is the Schema for the bpfapplications API
            displayName: Bpf Application
            kind: BpfApplication
            name: bpfapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsApplication is the Schema for the BpfNsApplications
              API
            displayName: Bpf Namespaced Application
            kind: BpfNsApplication
            name: bpfnsapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsProgram is the Schema for the BpfNsProgram API
            displayName: Bpf Namespaced Program
            kind: BpfNsProgram
            name: bpfnsprograms.bpfman.io
            version: v1alpha1
          - description: BpfProgram is the Schema for the BpfProgram API
            displayName: Bpf Program
            kind: BpfProgram
            name: bpfprograms.bpfman.io
            version: v1alpha1
          - description: FentryProgram is the Schema for the Fentryprograms API
            displayName: Fentry Program
            kind: FentryProgram
            name: fentryprograms.bpfman.io
            version: v1alpha1
          - description: FexitProgram is the Schema for the Fexitprograms API
            displayName: Fexit Program
            kind: FexitProgram
            name: fexitprograms.bpfman.io
            version: v1alpha1
          - description: KprobeProgram is the Schema for the Kprobeprograms API
            displayName: Kprobe Program
            kind: KprobeProgram
            name: kprobeprograms.bpfman.io
            version: v1alpha1
          - description: TcNsProgram is the Schema for the Tcnsprograms API
            displayName: TC Namespaced Program
            kind: TcNsProgram
            name: tcnsprograms.bpfman.io
            version: v1alpha1
          - description: TcProgram is the Schema for the Tcprograms API
            displayName: TC Program
            kind: TcProgram
            name: tcprograms.bpfman.io
            version: v1alpha1
          - description: TcxNsProgram is the Schema for the Tcxnsprograms API
            displayName: TCX Namespaced Program
            kind: TcxNsProgram
            name: tcxnsprograms.bpfman.io
            version: v1alpha1
          - description: TcxProgram is the Schema for the Tcxprograms API
            displayName: TCX Program
            kind: TcxProgram
            name: tcxprograms.bpfman.io
            version: v1alpha1
          - description: TracepointProgram is the Schema for the Tracepointprograms
              API
            displayName: Tracepoint Program
            kind: TracepointProgram
            name: tracepointprograms.bpfman.io
            version: v1alpha1
          - description: UprobeNsProgram is the Schema for the Uprobensprograms API
            displayName: Uprobe Namespaced Program
            kind: UprobeNsProgram
            name: uprobensprograms.bpfman.io
            version: v1alpha1
          - description: UprobeProgram is the Schema for the Uprobeprograms API
            displayName: Uprobe Program
            kind: UprobeProgram
            name: uprobeprograms.bpfman.io
            version: v1alpha1
          - description: XdpNsProgram is the Schema for the Xdpnsprograms API
            displayName: Xdp Namespaced Program
            kind: XdpNsProgram
            name: xdpnsprograms.bpfman.io
            version: v1alpha1
          - description: XdpProgram is the Schema for the Xdpprograms API
            displayName: Xdp Program
            kind: XdpProgram
            name: xdpprograms.bpfman.io
            version: v1alpha1
        description: "The bpfman Operator is a Kubernetes Operator for deploying [bpfman](https://bpfman.netlify.app/),
          a system daemon\nfor managing eBPF programs. It deploys bpfman itself along
          with CRDs to make deploying\neBPF programs in Kubernetes much easier.\n\n##
          Quick Start\n\nTo get bpfman up and running quickly simply click 'install'
          to deploy the bpfman-operator in the bpfman namespace via operator-hub.\n##
          Configuration\n\nThe `bpfman-config` configmap is automatically created
          in the `bpfman` namespace and used to configure the bpfman deployment.\n\nTo
          edit the config simply run\n\n```bash\nkubectl edit cm bpfman-config\n```\n\nThe
          following fields are adjustable\n\n- `bpfman.agent.image`: The image used
          for the bpfman-agent`\n- `bpfman.image`: The image used for bpfman`\n- `bpfman.log.level`:
          the log level for bpfman, currently supports `debug`, `info`, `warn`, `error`,
          and `fatal`, defaults to `info`\n- `bpfman.agent.log.level`: the log level
          for the bpfman-agent currently supports `info`, `debug`, and `trace` \n\nThe
          bpfman operator deploys eBPF programs via CRDs. The following CRDs are currently
          available, \n\n- BpfApplication\n- FentryProgram\n- FexitProgram\n - KprobeProgram\n-
          TcProgram\n- TcxProgram\n- TracepointProgram\n- UprobeProgram\n- XdpProgram\n
          - BpfNsApplication\n- TcNsProgram\n- TcxNsProgram\n- UprobeNsProgram\n-
          XdpNsProgram\n- BpfProgram\n - BpfNsProgram\n\n ## More information\n\nPlease
          checkout the [bpfman community website](https://bpfman.io/) for more information."
        displayName: Bpfman Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ebpf
        - kubernetes
        links:
        - name: bpfman website
          url: https://bpfman.io/
        maintainers:
        - email: astoycos@redhat.com
          name: Andrew Stoycos
        - email: afredette@redhat.com
          name: Andre Fredette
        maturity: alpha
        provider:
          name: The bpfman Community
          url: https://bpfman.io/
        relatedImages:
        - quay.io/bpfman/bpfman-operator:v0.5.6
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 0.5.6
      entries:
      - name: bpfman-operator.v0.5.6
        version: 0.5.6
      - name: bpfman-operator.v0.5.5
        version: 0.5.5
      - name: bpfman-operator.v0.5.4
        version: 0.5.4
      - name: bpfman-operator.v0.5.2
        version: 0.5.2
      - name: bpfman-operator.v0.5.1
        version: 0.5.1
      - name: bpfman-operator.v0.5.0
        version: 0.5.0
      - name: bpfman-operator.v0.4.2
        version: 0.4.2
      - name: bpfman-operator.v0.4.1
        version: 0.4.1
      name: alpha
    defaultChannel: alpha
    packageName: bpfman-operator
    provider:
      name: The bpfman Community
      url: https://bpfman.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-organizations-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-organizations-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "organizations.services.k8s.aws/v1alpha1",
                "kind": "OrganizationalUnit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/organizations-controller:1.1.1
          createdAt: "2025-09-19T19:35:20Z"
          description: AWS Organizations controller is a service controller for managing
            Organizations resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Account represents the state of an AWS organizations Account
              resource.
            displayName: Account
            kind: Account
            name: accounts.organizations.services.k8s.aws
            version: v1alpha1
          - description: OrganizationalUnit represents the state of an AWS organizations
              OrganizationalUnit resource.
            displayName: OrganizationalUnit
            kind: OrganizationalUnit
            name: organizationalunits.organizations.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Organizations resources in AWS from within your Kubernetes cluster.

          **About Amazon Organizations**

          With AWS Organizations you can perform account management activities at scale by consolidating multiple AWS accounts into a single organization. Consolidating accounts simplifies how you use other AWS services. You can leverage the multi-account management services available in AWS Organizations with select AWS services to perform tasks on all accounts that are members of your organization.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Organizations
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - organizations
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Organizations Developer Resources
          url: https://aws.amazon.com/organizations/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: organizations maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/organizations-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-organizations-controller.v1.1.1
        version: 1.1.1
      - name: ack-organizations-controller.v1.0.13
        version: 1.0.13
      - name: ack-organizations-controller.v1.0.11
        version: 1.0.11
      - name: ack-organizations-controller.v1.0.10
        version: 1.0.10
      - name: ack-organizations-controller.v1.0.9
        version: 1.0.9
      - name: ack-organizations-controller.v1.0.8
        version: 1.0.8
      - name: ack-organizations-controller.v1.0.7
        version: 1.0.7
      - name: ack-organizations-controller.v1.0.6
        version: 1.0.6
      - name: ack-organizations-controller.v1.0.5
        version: 1.0.5
      - name: ack-organizations-controller.v1.0.4
        version: 1.0.4
      - name: ack-organizations-controller.v1.0.1
        version: 1.0.1
      - name: ack-organizations-controller.v1.0.0
        version: 1.0.0
      - name: ack-organizations-controller.v0.0.10
        version: 0.0.10
      - name: ack-organizations-controller.v0.0.9
        version: 0.0.9
      - name: ack-organizations-controller.v0.0.8
        version: 0.0.8
      - name: ack-organizations-controller.v0.0.7
        version: 0.0.7
      - name: ack-organizations-controller.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: ack-organizations-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-sds-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-sds-operator.v8.0.230
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.230"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleStoragePool",
                "metadata": {
                  "name": "infoscalestoragepool-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:f4947eab2b31fccc86b5d5d0d3fd4eb1f7513fd8325e379cc259f5ccd1dfc3e4
          createdAt: "2023-10-26T20:36:59Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          olm.skipRange: <8.0.230
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleStoragePool is the Schema for the infoscalestoragepools
              API
            displayName: InfoScale Storage Pool
            kind: InfoScaleStoragePool
            name: infoscalestoragepools.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.230"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:f4947eab2b31fccc86b5d5d0d3fd4eb1f7513fd8325e379cc259f5ccd1dfc3e4
        version: 8.0.230
      entries:
      - name: infoscale-sds-operator.v8.0.230
        version: 8.0.230
      - name: infoscale-sds-operator.v8.0.220
        version: 8.0.220
      - name: infoscale-sds-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-sds-operator.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-sds-operator.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.330"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:dab501b99373c840c7e8e3ab479fa273b39f9bf8a896f3f1e4e945f4ab81f64e
          createdAt: "2024-06-24T09:03:32Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.330"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:dab501b99373c840c7e8e3ab479fa273b39f9bf8a896f3f1e4e945f4ab81f64e
        version: 8.0.330
      entries:
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-sds-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.410"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
          createdAt: "2024-07-30T09:00:57Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Arctera InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/168721626-168743109-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.410"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Arctera recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
        version: 8.0.410
      entries:
      - name: infoscale-sds-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-sds-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-sds-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-sds-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.410"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
          createdAt: "2024-07-30T09:00:57Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Arctera InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/168721626-168743109-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.410"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Arctera recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
        version: 8.0.410
      entries:
      - name: infoscale-sds-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-sds-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-sds-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-sds-operator.v8.0.210
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.210"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:26573b98743245c0b607abbe0d981abce6ce9fb1c5c0cd9fe735351b7136f4e8
          createdAt: "2022-11-26T16:36:59Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleStoragePool is the Schema for the infoscalestoragepools
              API
            displayName: InfoScale Storage Pool
            kind: InfoScaleStoragePool
            name: infoscalestoragepools.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.210"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:1667227b5111fe52ed304ab19ff9fadfc677239e56b2c93a3b89b1bfbc31354d
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:26573b98743245c0b607abbe0d981abce6ce9fb1c5c0cd9fe735351b7136f4e8
        version: 8.0.210
      entries:
      - name: infoscale-sds-operator.v8.0.210
        version: 8.0.210
      - name: infoscale-sds-operator.v8.0.201
        version: 8.0.201
      - name: infoscale-sds-operator.v8.0.200
        version: 8.0.200
      - name: infoscale-sds-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-sds-operator.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-sds-operator
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: ovn-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ovn-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ovn.openstack.org/v1beta1",
                "kind": "OVNController",
                "metadata": {
                  "name": "ovncontroller-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "ovn.openstack.org/v1beta1",
                "kind": "OVNDBCluster",
                "metadata": {
                  "name": "ovndbcluster-nb-sample"
                },
                "spec": {
                  "dbType": "NB",
                  "logLevel": "info",
                  "storageClass": "local-storage",
                  "storageRequest": "10G"
                }
              },
              {
                "apiVersion": "ovn.openstack.org/v1beta1",
                "kind": "OVNDBCluster",
                "metadata": {
                  "name": "ovndbcluster-sb-sample"
                },
                "spec": {
                  "dbType": "SB",
                  "logLevel": "info",
                  "storageClass": "local-storage",
                  "storageRequest": "10G"
                }
              },
              {
                "apiVersion": "ovn.openstack.org/v1beta1",
                "kind": "OVNNorthd",
                "metadata": {
                  "name": "ovnnorthd-sample"
                },
                "spec": {
                  "logLevel": "info"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:45Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OVNController is the Schema for the ovncontrollers API
            displayName: OVNController
            kind: OVNController
            name: ovncontrollers.ovn.openstack.org
            version: v1beta1
          - description: OVNDBCluster is the Schema for the ovndbclusters API
            displayName: OVNDBCluster
            kind: OVNDBCluster
            name: ovndbclusters.ovn.openstack.org
            version: v1beta1
          - description: OVNNorthd is the Schema for the ovnnorthds API
            displayName: OVNNorthd
            kind: OVNNorthd
            name: ovnnorthds.ovn.openstack.org
            version: v1beta1
        description: OVN Operator
        displayName: OVN Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Network
        - Ovn
        links:
        - name: Ovn Operator
          url: https://github.com/openstack-k8s-operators/ovn-operator
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/ovn-rhel9-operator@sha256:bce19011b3623c66a59efe6ea874e2403440813bdc970cd6170853a8093076f3
        - registry.redhat.io/rhoso/openstack-ovn-base-rhel9@sha256:b0d424f6ea6de8192382541a95aa56fa32b9b95e8d433c41c4a6dadbe4127fae
        - registry.redhat.io/rhoso/openstack-ovn-nb-db-server-rhel9@sha256:e71c8aac00a78e3d786332aaada3657276f69483df1e441b70589601ca038a5b
        - registry.redhat.io/rhoso/openstack-ovn-sb-db-server-rhel9@sha256:7aa2832f80846d5ac71733a1f186d39dce3654ebb661d32b8e4580d92be8e338
        - registry.redhat.io/rhoso/openstack-ovn-northd-rhel9@sha256:f1e4ce04b67d888fa950b83d1f00459c67c602397336da81de3ff71d4052a5d8
        - registry.redhat.io/rhoso/openstack-ovn-controller-rhel9@sha256:6c7827dbe25c168d28250f197d5ea6f3f60cfbab4c974267ec5ea0f8d631c3f5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        version: 1.0.6
      entries:
      - name: ovn-operator.v1.0.6
        version: 1.0.6
      - name: ovn-operator.v1.0.5
        version: 1.0.5
      - name: ovn-operator.v1.0.4
        version: 1.0.4
      - name: ovn-operator.v1.0.3
        version: 1.0.3
      - name: ovn-operator.v1.0.2
        version: 1.0.2
      - name: ovn-operator.v1.0.1
        version: 1.0.1
      - name: ovn-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: ovn-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: fuse-apicurito
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-apicurito.v7.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1alpha1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:8d19ad579f874d6b9a08f127c8900906913a5a7c1e69cbe4b5cb0333b8d86152
          createdAt: 2022-02-28 09:35:38 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.10.3'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:8d19ad579f874d6b9a08f127c8900906913a5a7c1e69cbe4b5cb0333b8d86152
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:8dbeeb8b04c81f41d896f71db90813de57e67493fa85389b63ac23f390785043
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:c02c218ccfbb385ea29af98409995a3d9c26cb0b4c76caf2e0779fa233b5ff8c
        version: 7.10.3
      entries:
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.10.x
    - currentCSV: fuse-apicurito.v7.11.1-0.1687357785.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:2f8d4450187839f0e0c761e5ddcf61b03f5996acfe3d75b60120e76e5b0167c1
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.11.2'
          olm.substitutesFor: fuse-apicurito.v7.11.1-0.1682509045.p
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:2f8d4450187839f0e0c761e5ddcf61b03f5996acfe3d75b60120e76e5b0167c1
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:df61aafc4023b9068a8636c35eabb0d5fda400f523b64cb1fc5d8c480272c435
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:4c493a2d690d0d7c1684c53d15c06b47d691aeffdde6dd471b6e75fed2fec6a9
        version: 7.11.1+0.1687357785.p
      entries:
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.11.x
    - currentCSV: fuse-apicurito.v7.12.1-0.1706187670.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:7ccefee63165e7eaf9a609e8a9f978a06f27deff69a318fe399a29292ad99106
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.12.1'
          olm.substitutesFor: fuse-apicurito.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:7ccefee63165e7eaf9a609e8a9f978a06f27deff69a318fe399a29292ad99106
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:4cd612283e8136839f1f48a3a220dac7c6c38adf14b61bc03a981e514baefe9b
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:bee0eeaa7898af835f80c5217059d73361daa42197fa6cbbaf416e1ba373eb54
        version: 7.12.1+0.1706187670.p
      entries:
      - name: fuse-apicurito.v7.12.1-0.1706187670.p
        version: 7.12.1+0.1706187670.p
      - name: fuse-apicurito.v7.12.1
        version: 7.12.1
      - name: fuse-apicurito.v7.12.0
        version: 7.12.0
      - name: fuse-apicurito.v7.12.0-0.1697748407.p
        version: 7.12.0+0.1697748407.p
      - name: fuse-apicurito.v7.12.0-0.1697419531.p
        version: 7.12.0+0.1697419531.p
      - name: fuse-apicurito.v7.12.0-0.1694087055.p
        version: 7.12.0+0.1694087055.p
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.12.x
    - currentCSV: fuse-apicurito.v7.13.0-0.1753785079.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:d58d345d74059242d57d5c1a75f31517769230d3014a42243fda20aa93ebd5c8
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.7.0 <7.13.0'
          olm.substitutesFor: fuse-apicurito.v7.13.0-0.1751291353.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:d58d345d74059242d57d5c1a75f31517769230d3014a42243fda20aa93ebd5c8
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:f5bf9240b28b3664bdb11b1f57f611e6194ff8643ae939b6ab3b2b176705f487
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:4e64a5930a10b76147ac128a216cd7c0346da2ff3e9f31ed86311941ed193cc1
        version: 7.13.0+0.1753785079.p
      entries:
      - name: fuse-apicurito.v7.13.0-0.1739885035.p
        version: 7.13.0+0.1739885035.p
      - name: fuse-apicurito.v7.13.0-0.1741884705.p
        version: 7.13.0+0.1741884705.p
      - name: fuse-apicurito.v7.13.0-0.1753785079.p
        version: 7.13.0+0.1753785079.p
      - name: fuse-apicurito.v7.13.0-0.1721224330.p
        version: 7.13.0+0.1721224330.p
      - name: fuse-apicurito.v7.13.0-0.1753778625.p
        version: 7.13.0+0.1753778625.p
      - name: fuse-apicurito.v7.13.0-0.1753702605.p
        version: 7.13.0+0.1753702605.p
      - name: fuse-apicurito.v7.13.0-0.1721607497.p
        version: 7.13.0+0.1721607497.p
      - name: fuse-apicurito.v7.13.0-0.1751291353.p
        version: 7.13.0+0.1751291353.p
      - name: fuse-apicurito.v7.13.0-0.1748960166.p
        version: 7.13.0+0.1748960166.p
      - name: fuse-apicurito.v7.13.0-0.1745523192.p
        version: 7.13.0+0.1745523192.p
      - name: fuse-apicurito.v7.13.0-0.1744737027.p
        version: 7.13.0+0.1744737027.p
      - name: fuse-apicurito.v7.13.0-0.1744280419.p
        version: 7.13.0+0.1744280419.p
      - name: fuse-apicurito.v7.13.0-0.1741875975.p
        version: 7.13.0+0.1741875975.p
      - name: fuse-apicurito.v7.13.0-0.1739473863.p
        version: 7.13.0+0.1739473863.p
      - name: fuse-apicurito.v7.13.0-0.1738056248.p
        version: 7.13.0+0.1738056248.p
      - name: fuse-apicurito.v7.13.0-0.1731460815.p
        version: 7.13.0+0.1731460815.p
      - name: fuse-apicurito.v7.13.0-0.1729189341.p
        version: 7.13.0+0.1729189341.p
      - name: fuse-apicurito.v7.13.0-0.1727957847.p
        version: 7.13.0+0.1727957847.p
      - name: fuse-apicurito.v7.13.0-0.1727193594.p
        version: 7.13.0+0.1727193594.p
      - name: fuse-apicurito.v7.13.0-0.1725022690.p
        version: 7.13.0+0.1725022690.p
      - name: fuse-apicurito.v7.13.0-0.1724659732.p
        version: 7.13.0+0.1724659732.p
      - name: fuse-apicurito.v7.13.0-0.1724053718.p
        version: 7.13.0+0.1724053718.p
      - name: fuse-apicurito.v7.13.0-0.1723709434.p
        version: 7.13.0+0.1723709434.p
      - name: fuse-apicurito.v7.13.0
        version: 7.13.0
      - name: fuse-apicurito.v7.13.0-0.1720707249.p
        version: 7.13.0+0.1720707249.p
      - name: fuse-apicurito.v7.13.0-0.1718285973.p
        version: 7.13.0+0.1718285973.p
      - name: fuse-apicurito.v7.13.0-0.1717672390.p
        version: 7.13.0+0.1717672390.p
      - name: fuse-apicurito.v7.13.0-0.1753778936.p
        version: 7.13.0+0.1753778936.p
      - name: fuse-apicurito.v7.13.0-0.1753701210.p
        version: 7.13.0+0.1753701210.p
      - name: fuse-apicurito.v7.12.1-0.1706187670.p
        version: 7.12.1+0.1706187670.p
      - name: fuse-apicurito.v7.12.1
        version: 7.12.1
      - name: fuse-apicurito.v7.12.0-0.1697419531.p
        version: 7.12.0+0.1697419531.p
      - name: fuse-apicurito.v7.12.0-0.1697748407.p
        version: 7.12.0+0.1697748407.p
      - name: fuse-apicurito.v7.12.0-0.1694087055.p
        version: 7.12.0+0.1694087055.p
      - name: fuse-apicurito.v7.12.0
        version: 7.12.0
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.13.x
    - currentCSV: fuse-apicurito.v7.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
                    "apiVersion": "apicur.io/v1alpha1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
            }]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:05c9676aada2d34a985a59abcf55105a3765d75b0d3764089d2b32e9ca5d367c
          createdAt: "2020-10-06 12:48:47"
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.9.3'
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito Project
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CRD for Apicurito
            displayName: Apicurito CRD
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:05c9676aada2d34a985a59abcf55105a3765d75b0d3764089d2b32e9ca5d367c
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:5535864ce393ae0ae5ee52ed0c7661e0351fb7df2f9235da70608245519530b2
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:dc39b79c360511ff999eba5b2c299db3e08c9c244a0a1f3da8c90c02695e506d
        version: 7.9.3
      entries:
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.9.x
    defaultChannel: fuse-apicurito-7.13.x
    packageName: fuse-apicurito
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      app: kube-green
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: kube-green Community
      provider-url: https://kube-green.dev
    name: kube-green
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kube-green.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kube-green.com/v1alpha1",
                "kind": "SleepInfo",
                "metadata": {
                  "labels": {
                    "app": "kube-green"
                  },
                  "name": "sleepinfo-sample"
                },
                "spec": {
                  "excludeRef": [
                    {
                      "apiVersion": "apps/v1",
                      "kind": "Deployment",
                      "name": "api-gateway"
                    }
                  ],
                  "sleepAt": "20:00",
                  "suspendCronJobs": true,
                  "timeZone": "Europe/Rome",
                  "wakeUpAt": "08:00",
                  "weekdays": "1-5"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          containerImage: docker.io/kubegreen/kube-green:0.7.1
          createdAt: "2025-06-29T15:45:17Z"
          description: Suspend your pods when no-one's using them to save energy and
            restore when necessary
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kube-green/kube-green
          support: kube-green Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SleepInfo is the Schema for the sleepinfos API
            displayName: SleepInfo
            kind: SleepInfo
            name: sleepinfos.kube-green.com
            version: v1alpha1
        description: |
          ## About this Operator

          [kube-green](https://kube-green.dev) shuts down (some of) your resources when you don't need them, reducing the consumption of resources of the cluster.

          Install this operator and apply the SleepInfo CRD to stop your dev namespaces from running during weekends, nights or when it is not necessary.
          See the [docs](https://kube-green.dev/docs/getting-started/) for more information.
          This operator is responsible to suspend and restart pods at specific times, defined by the SleepInfo CRD.

          An example of SleepInfo:

          ```yaml
          apiVersion: kube-green.com/v1alpha1
          kind: SleepInfo
          metadata:
            name: working-hours
          spec:
            weekdays: "1-5"
            sleepAt: "20:00"
            wakeUpAt: "08:00"
            timeZone: "Europe/Rome"
          ```

          This configuration stop pods everyday from monday to friday at 20:00 and restore them at 8:00.
          In this way, your cluster nodes could downscale, producing less CO2 and saving money.
        displayName: kube-green
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - green
        - green software
        - resources
        - downscale
        - autoscale
        - climate change
        links:
        - name: Homepage
          url: https://kube-green.dev
        - name: Source Code
          url: https://github.com/kube-green/kube-green
        maintainers:
        - email: bianchidavide12+kubegreen@gmail.com
          name: Davide
        maturity: alpha
        minKubeVersion: 1.27.0
        provider:
          name: kube-green Community
          url: https://kube-green.dev
        relatedImages:
        - docker.io/kubegreen/kube-green:0.7.1
        version: 0.7.1
      entries:
      - name: kube-green.v0.7.1
        version: 0.7.1
      - name: kube-green.v0.7.0
        version: 0.7.0
      - name: kube-green.v0.6.0
        version: 0.6.0
      - name: kube-green.v0.5.2
        version: 0.5.2
      - name: kube-green.v0.5.1
        version: 0.5.1
      - name: kube-green.v0.5.0
        version: 0.5.0
      - name: kube-green.v0.4.1
        version: 0.4.1
      - name: kube-green.v0.4.0
        version: 0.4.0
      - name: kube-green.v0.3.1
        version: 0.3.1
      - name: kube-green.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: kube-green
    provider:
      name: kube-green Community
      url: https://kube-green.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeRocketCI
      provider-url: https://docs.kuberocketci.io
    name: nexus-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-operator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexus-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexus",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexus-sample"
                },
                "spec": {
                  "secret": "nexus-secret",
                  "url": "https://nexus-example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRepository",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrepository-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrepository",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrepository-sample"
                },
                "spec": {
                  "go": {
                    "proxy": {
                      "cleanup": {
                        "policyNames": [
                          "cleanup-policy"
                        ]
                      },
                      "httpClient": {
                        "autoBlock": true,
                        "blocked": true,
                        "connection": {
                          "retries": 1440,
                          "timeout": 1440,
                          "userAgentSuffix": "user-agent-suffix"
                        }
                      },
                      "name": "go-proxy",
                      "negativeCache": {
                        "enabled": true,
                        "timeToLive": 1440
                      },
                      "online": true,
                      "proxy": {
                        "contentMaxAge": 1440,
                        "metadataMaxAge": 1440,
                        "remoteUrl": "https://test.com"
                      },
                      "routingRule": "routing-rule",
                      "storage": {
                        "blobStoreName": "blob-store-name",
                        "strictContentTypeValidation": true
                      }
                    }
                  },
                  "nexusRef": {
                    "name": "nexus-sample"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrole",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrole-sample"
                },
                "spec": {
                  "description": "test-role",
                  "id": "test-role",
                  "name": "test-role",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "privileges": [
                    "nx-blobstores-all"
                  ]
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexususer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexususer",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexususer-sample"
                },
                "spec": {
                  "email": "test@gmail.com",
                  "firstName": "John1",
                  "id": "test-user",
                  "lastName": "Doe",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "roles": [
                    "nx-admin"
                  ],
                  "secret": "$user-secret:password"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/nexus-operator:3.0.0
          createdAt: "2023-12-11T12:54:50Z"
          description: An Operator for managing Nexus entities.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-nexus-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus is the Schema for the nexus API.
            displayName: Nexus
            kind: Nexus
            name: nexuses.edp.epam.com
            version: v1alpha1
          - description: NexusRepository is the Schema for the nexusrepositories API.
            displayName: Nexus Repository
            kind: NexusRepository
            name: nexusrepositories.edp.epam.com
            version: v1alpha1
          - description: NexusRole is the Schema for the nexusroles API.
            displayName: Nexus Role
            kind: NexusRole
            name: nexusroles.edp.epam.com
            version: v1alpha1
          - description: NexusUser is the Schema for the nexususers API.
            displayName: Nexus User
            kind: NexusUser
            name: nexususers.edp.epam.com
            version: v1alpha1
        description: |
          Nexus Operator is an operator that is responsible for establishing
          a connection to provided NexusRepository Server, reconciling Nexus entities (groups,
          users, repositories, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Nexus Repository.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  nexus-secret
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Nexus` with Nexus instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1alpha1
              kind: Nexus
              metadata:
                name: nexus-sample
              spec:
                secret: nexus-secret                # Secret name
                url: https://nexus.example.com      # Nexus URL
              ```

              Wait for the `.status` field with  `status.connected: true`
        displayName: EDP Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - edp
        - nexus
        - operator
        - artifact
        - storage
        links:
        - name: Nexus Operator
          url: https://github.com/epam/edp-nexus-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/nexus-operator:3.0.0
        version: 3.0.0
      entries:
      - name: nexus-operator.v3.0.0
        version: 3.0.0
      name: alpha
    - currentCSV: nexus-operator.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexus-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexus",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexus-sample"
                },
                "spec": {
                  "secret": "nexus-secret",
                  "url": "https://nexus-example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusBlobStore",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusblobstore-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusblobstore",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusblobstore-sample"
                },
                "spec": {
                  "file": {
                    "path": "storage-blobstore/nexusblobstore-sample"
                  },
                  "name": "nexusblobstore-sample",
                  "nexusRef": {
                    "name": "nexus-sample"
                  },
                  "softQuota": {
                    "limit": 1000,
                    "type": "spaceUsedQuota"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusCleanupPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexuscleanuppolicy-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexuscleanuppolicy",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexuscleanuppolicy-sample"
                },
                "spec": {
                  "criteria": {
                    "lastBlobUpdated": 30
                  },
                  "description": "Cleanup policy for go",
                  "format": "go",
                  "name": "go-cleanup-policy",
                  "nexusRef": {
                    "name": "nexus-sample"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRepository",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrepository-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrepository",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrepository-sample"
                },
                "spec": {
                  "go": {
                    "proxy": {
                      "cleanup": {
                        "policyNames": [
                          "cleanup-policy"
                        ]
                      },
                      "httpClient": {
                        "autoBlock": true,
                        "blocked": true,
                        "connection": {
                          "retries": 1440,
                          "timeout": 1440,
                          "userAgentSuffix": "user-agent-suffix"
                        }
                      },
                      "name": "go-proxy",
                      "negativeCache": {
                        "enabled": true,
                        "timeToLive": 1440
                      },
                      "online": true,
                      "proxy": {
                        "contentMaxAge": 1440,
                        "metadataMaxAge": 1440,
                        "remoteUrl": "https://test.com"
                      },
                      "routingRule": "routing-rule",
                      "storage": {
                        "blobStoreName": "blob-store-name",
                        "strictContentTypeValidation": true
                      }
                    }
                  },
                  "nexusRef": {
                    "name": "nexus-sample"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrole",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrole-sample"
                },
                "spec": {
                  "description": "test-role",
                  "id": "test-role",
                  "name": "test-role",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "privileges": [
                    "nx-blobstores-all"
                  ]
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusScript",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusscript-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusscript",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusscript-sample"
                },
                "spec": {
                  "content": "security.setAnonymousAccess(Boolean.valueOf(args))\n",
                  "name": "anonymous",
                  "nexusRef": {
                    "name": "nexus-sample"
                  },
                  "payload": "true"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexususer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexususer",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexususer-sample"
                },
                "spec": {
                  "email": "test@gmail.com",
                  "firstName": "John1",
                  "id": "test-user",
                  "lastName": "Doe",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "roles": [
                    "nx-admin"
                  ],
                  "secret": "$user-secret:password"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/nexus-operator:3.5.0
          createdAt: "2025-04-22T11:21:24Z"
          description: An Operator for managing Nexus entities.
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/epam/edp-nexus-operator
          support: KubeRocketCI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NexusBlobStore is the Schema for the nexusblobstores API.
            displayName: Nexus Blob Store
            kind: NexusBlobStore
            name: nexusblobstores.edp.epam.com
            version: v1alpha1
          - description: NexusCleanupPolicy is the Schema for the cleanuppolicies
              API.
            displayName: Nexus Cleanup Policy
            kind: NexusCleanupPolicy
            name: nexuscleanuppolicies.edp.epam.com
            version: v1alpha1
          - description: Nexus is the Schema for the nexus API.
            displayName: Nexus
            kind: Nexus
            name: nexuses.edp.epam.com
            version: v1alpha1
          - description: NexusRepository is the Schema for the nexusrepositories API.
            displayName: Nexus Repository
            kind: NexusRepository
            name: nexusrepositories.edp.epam.com
            version: v1alpha1
          - description: NexusRole is the Schema for the nexusroles API.
            displayName: Nexus Role
            kind: NexusRole
            name: nexusroles.edp.epam.com
            version: v1alpha1
          - description: NexusScript is the Schema for the nexusscripts API.
            displayName: Nexus Script
            kind: NexusScript
            name: nexusscripts.edp.epam.com
            version: v1alpha1
          - description: NexusUser is the Schema for the nexususers API.
            displayName: Nexus User
            kind: NexusUser
            name: nexususers.edp.epam.com
            version: v1alpha1
        description: |
          Nexus Operator is an operator that is responsible for establishing
          a connection to provided NexusRepository Server, reconciling Nexus entities (groups,
          users, repositories, etc) according to the created CRs.

          ## Quick Start

          1. Login into Nexus and create user. Attach permissions to user such as scripts, rules, blobs etc. Insert user credentials into Kubernetes secret.

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name: nexus-admin-password
              data:
                password: cGFzcw==  # base64-encoded value of "pass"
                user:     dXNlcg==  # base64-encoded value of "user"
              ```

          2. Create Custom Resource `kind: Nexus` with Nexus instance URL and secret created on the previous step:

              ```yaml
              apiVersion: edp.epam.com/v1alpha1
              kind: Nexus
              metadata:
                name: nexus
              spec:
                secret: nexus-admin-password
                url: http://nexus.example.com
              ```

              Wait for the `.status` field with  `status.connected: true`

          3. Create Role using Custom Resources NexusRole:

              ```yaml
              apiVersion: edp.epam.com/v1alpha1
              kind: NexusRole
              metadata:
                name: edp-admin
              spec:
                description: Read and write access to all repos and scripts
                id: edp-admin
                name: edp-admin
                nexusRef:
                  kind: Nexus
                  name: nexus
                privileges:
                  - nx-apikey-all
                  - nx-repository-view-*-*-add
                  - nx-repository-view-*-*-browse
                  - nx-repository-view-*-*-edit
                  - nx-repository-view-*-*-read
                  - nx-script-*-add
                  - nx-script-*-delete
                  - nx-script-*-run
                  - nx-search-read
              ```
        displayName: EDP Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - artifact
        - edp
        - kuberocketci
        - nexus
        - operator
        - storage
        links:
        - name: Nexus Operator
          url: https://github.com/epam/edp-nexus-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: KubeRocketCI
          url: https://docs.kuberocketci.io
        relatedImages:
        - docker.io/epamedp/nexus-operator:3.5.0
        version: 3.5.0
      entries:
      - name: nexus-operator.v3.5.0
        version: 3.5.0
      - name: nexus-operator.v3.4.0
        version: 3.4.0
      - name: nexus-operator.v3.3.0
        version: 3.3.0
      name: stable
    defaultChannel: stable
    packageName: nexus-operator
    provider:
      name: KubeRocketCI
      url: https://docs.kuberocketci.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp, Inc.
      provider-url: https://www.netapp.com/
    name: trident-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trident-operator.v25.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "trident.netapp.io/v1",
                "kind": "TridentOrchestrator",
                "metadata": {
                  "name": "trident"
                },
                "spec": {
                  "IPv6": false,
                  "debug": true,
                  "nodePrep": [],
                  "imagePullSecrets": [],
                  "imageRegistry": "",
                  "namespace": "trident",
                  "silenceAutosupport": false
                }
              }
            ]
          categories: Storage
          containerImage: docker.io/netapp/trident-operator@sha256:c3448b6847550f45849a5cdbb9e0cedb35290d21131bc513d833d95eb42a43e0
          createdAt: "2025-09-03"
          description: Trident Operator, to manage NetApp Trident installations
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          operators.openshift.io/valid-subscription: No subscriptions are required
          repository: https://github.com/netapp/trident
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Used to deploy NetApp Trident.
            displayName: Trident Orchestrator
            kind: TridentOrchestrator
            name: tridentorchestrators.trident.netapp.io
            version: v1
          - description: Automates AWS FSxN backend configuration
            displayName: Trident Configurator
            kind: TridentConfigurator
            name: tridentconfigurators.trident.netapp.io
            version: v1
        description: |
          NetApp Trident is an open source storage provisioner and orchestrator maintained by NetApp. It enables you to create storage volumes for containerized applications managed by Docker and Kubernetes. For full release information, including patch release changes, see https://docs.netapp.com/us-en/trident/trident-rn.html.
        displayName: NetApp Trident
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - NetApp
        - ONTAP
        - Trident
        - storage
        - performance
        - open source
        - Astra
        - Cloud Native
        - AWS
        - Azure
        - GCP
        - Cloud Volumes
        links:
        - name: GitHub Repository
          url: https://github.com/NetApp/trident
        - name: Trident documentation
          url: https://docs.netapp.com/us-en/trident/index.html
        - name: Support policy
          url: https://mysupport.netapp.com/site/info/version-support
        - name: Release Notes
          url: https://docs.netapp.com/us-en/trident/trident-rn.html
        maintainers:
        - email: ng-trident-openshift-operator@netapp.com
          name: NetApp
        provider:
          name: NetApp, Inc.
          url: https://www.netapp.com/
        relatedImages:
        - docker.io/netapp/trident-operator@sha256:c3448b6847550f45849a5cdbb9e0cedb35290d21131bc513d833d95eb42a43e0
        - docker.io/netapp/trident@sha256:7925717979a84f93b62a2592682cde3443d6350ec4e88f8fd1a7d3ea812e8900
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:bdee8df520467b30312e7b14d7b6ef3963557ccd38269913de3fa51bb1bceec9
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:145ab76f9606e9d51dcd1e4e159c67021ecf53c24d8d80ed7bbeb2cc7fb5c42b
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6781d64bfb01e252595130eec814b9d2fe6426720a003cc61e82e9963767bfd4
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:b48bd5bc612c1bbe73d203c70e6caaae2b5360bd31eba20eb06cbd2102cc34fc
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:b88958c38883f5b1e3d307da90aac34a5ceafdb6bd3d989b062cb24ecafd1fbd
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:0ef3bf2a4dcbb5aa829ce9cc5f12efaccaf99b246b4b15abbb66b0135bec4cab
        - docker.io/netapp/trident-autosupport@sha256:21b3afbebfac3cc28c4eb319c2d385187bd8ebb728d8cfd82dbc437cc9b83b33
        version: 25.6.2
      entries:
      - name: trident-operator.v25.6.2
        version: 25.6.2
      - name: trident-operator.v25.6.1
        version: 25.6.1
      - name: trident-operator.v25.6.0
        version: 25.6.0
      - name: trident-operator.v25.2.1
        version: 25.2.1
      - name: trident-operator.v25.2.0
        version: 25.2.0
      name: stable
    defaultChannel: stable
    packageName: trident-operator
    provider:
      name: NetApp, Inc.
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: xspc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: xspc-operator.v1.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.hpe.com/v1",
                "kind": "XSPC",
                "metadata": {
                  "name": "xspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.connect.redhat.com/hpestorage/xspc-csi-driver-operator@sha256:8b8724998a8301057a3c4187c7b94bf4484aeb807065a12f84142cf844ea5534
          createdAt: "2024-12-17T03:53:30Z"
          description: An operator for managing HPE XP Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XSPC is the Schema for the xspcs API
            displayName: XSPC
            kind: XSPC
            name: xspcs.csi.hpe.com
            version: v1
        description: |-
          ## About
          HPE XP Storage Plug-in for Containers (XSPC) is a plugin that integrates HPE XP storage into Kubernetes based clusters.
          XSPC provides persistent volume provisioning capabilities from HPE's block storage arrays.
        displayName: HPE XP Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - XP7
        - XP8
        - HPE
        - Storage
        - Container Storage Interface
        - CSI
        links:
        - name: Knowledge Base
          url: https://www.hpe.com/us/en/storage/enterprise-xp-storage.html
        maintainers:
        - email: hpexpci@hpe.com
          name: HPE
        maturity: stable
        provider:
          name: HPE
        relatedImages:
        - registry.connect.redhat.com/hpestorage/xspc-csi-driver-operator@sha256:8b8724998a8301057a3c4187c7b94bf4484aeb807065a12f84142cf844ea5534
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.connect.redhat.com/hpestorage/xspc-csi-driver@sha256:bdeac71c0b54c2779910a8d8faea03c1122000127704f17ceadba4938b685353
        version: 1.15.0
      entries:
      - name: xspc-operator.v1.15.0
        version: 1.15.0
      name: stable
    defaultChannel: stable
    packageName: xspc-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Entando, Inc
      provider-url: ""
    name: entando-k8s-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: entando-k8s-operator.v6.3.2-pr3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: beta
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - docker.io/entando/entando-k8s-service@sha256:775391f0005e64764288bfe007d9ea576f13618c02ea400d030a7502cfb551bd
        - docker.io/entando/entando-k8s-database-service-controller@sha256:58ca4ad2fca83a715aba9b68609f9056f633d556eb9845b3aff5c05769369539
        - docker.io/entando/app-builder@sha256:1ca90c09302b9cadf8950278d92ad4d7f3a14f02dea8704320794aa77ada8631
        - docker.io/entando/entando-k8s-app-controller@sha256:c68d3b129fc625f5d1ee10ed2a107e28c145049d4b513dcdb2c736555025e339
        - docker.io/entando/entando-k8s-plugin-controller@sha256:a20c753dda86e0f54f210e446b404733045ba426fc7186dc3936daed8224fbcb
        - docker.io/entando/entando-component-manager@sha256:668191fc86b090452baaba18ec7bbc3d07df5a99750e0c5b98759599d565cfbe
        - docker.io/entando/entando-redhat-sso@sha256:3692adb3694b3bf278bed7a678c9d0eaa0aba8e3e3ed4117ca04156525bcac72
        - registry.redhat.io/rhel8/mysql-80@sha256:570bce7dc649ac7ac85e7ecb196aabc281c77a94590aa7df7d9746869ce757ea
        - docker.io/entando/entando-k8s-dbjob@sha256:8ae47dbf93201965106663299f06daad544982fa244d4a0f365a1b717727e5ed
        - docker.io/entando/entando-k8s-composite-app-controller@sha256:4bc6797bb41b3e407b64530c3cff7a78f3fc2c22863caad9c606ccd650310b61
        - docker.io/entando/entando-k8s-app-plugin-link-controller@sha256:aa7ba0861705213434e610530c5bf5da1bdb33f1834d3c346174f5a28145c0bd
        - docker.io/entando/entando-de-app-eap@sha256:1f0e35d7f7ee8123721cf8e80ac96ac2f594f21a2b68cf9e720917d090c8095e
        - docker.io/entando/entando-k8s-keycloak-controller@sha256:f1ed42001260802353197433cea7df4c2d0e5fd9bcd926bb5ae7c654dfad2ffe
        - registry.redhat.io/rhel8/postgresql-12@sha256:f4e5c728b644bf1888ec8086424852ed74b5596a511be29e636fb10218fc9b6f
        - entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        - docker.io/entando/entando-k8s-cluster-infrastructure-controller@sha256:86dbea6696349d7b2393e12bacef4d1acec2dd3f8551e397fa26c92706a18dd3
        version: 6.3.2-pr3
      entries:
      - name: entando-k8s-operator.v6.3.2-pr3
        version: 6.3.2-pr3
      name: alpha
    - currentCSV: entando-k8s-operator.v6.3.2-pr3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: beta
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - docker.io/entando/entando-redhat-sso@sha256:3692adb3694b3bf278bed7a678c9d0eaa0aba8e3e3ed4117ca04156525bcac72
        - entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        - docker.io/entando/entando-de-app-eap@sha256:1f0e35d7f7ee8123721cf8e80ac96ac2f594f21a2b68cf9e720917d090c8095e
        - docker.io/entando/entando-k8s-composite-app-controller@sha256:4bc6797bb41b3e407b64530c3cff7a78f3fc2c22863caad9c606ccd650310b61
        - docker.io/entando/entando-k8s-app-controller@sha256:c68d3b129fc625f5d1ee10ed2a107e28c145049d4b513dcdb2c736555025e339
        - docker.io/entando/entando-k8s-plugin-controller@sha256:a20c753dda86e0f54f210e446b404733045ba426fc7186dc3936daed8224fbcb
        - docker.io/entando/entando-component-manager@sha256:668191fc86b090452baaba18ec7bbc3d07df5a99750e0c5b98759599d565cfbe
        - docker.io/entando/entando-k8s-app-plugin-link-controller@sha256:aa7ba0861705213434e610530c5bf5da1bdb33f1834d3c346174f5a28145c0bd
        - docker.io/entando/entando-k8s-database-service-controller@sha256:58ca4ad2fca83a715aba9b68609f9056f633d556eb9845b3aff5c05769369539
        - docker.io/entando/app-builder@sha256:1ca90c09302b9cadf8950278d92ad4d7f3a14f02dea8704320794aa77ada8631
        - docker.io/entando/entando-k8s-keycloak-controller@sha256:f1ed42001260802353197433cea7df4c2d0e5fd9bcd926bb5ae7c654dfad2ffe
        - registry.redhat.io/rhel8/mysql-80@sha256:570bce7dc649ac7ac85e7ecb196aabc281c77a94590aa7df7d9746869ce757ea
        - registry.redhat.io/rhel8/postgresql-12@sha256:f4e5c728b644bf1888ec8086424852ed74b5596a511be29e636fb10218fc9b6f
        - docker.io/entando/entando-k8s-cluster-infrastructure-controller@sha256:86dbea6696349d7b2393e12bacef4d1acec2dd3f8551e397fa26c92706a18dd3
        - docker.io/entando/entando-k8s-dbjob@sha256:8ae47dbf93201965106663299f06daad544982fa244d4a0f365a1b717727e5ed
        - docker.io/entando/entando-k8s-service@sha256:775391f0005e64764288bfe007d9ea576f13618c02ea400d030a7502cfb551bd
        version: 6.3.2-pr3
      entries:
      - name: entando-k8s-operator.v6.3.2-pr3
        version: 6.3.2-pr3
      name: beta
    - currentCSV: entando-k8s-operator.v6.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: registry.hub.docker.com/entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - registry.hub.docker.com/entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        version: 6.3.2
      entries:
      - name: entando-k8s-operator.v6.3.2
        version: 6.3.2
      name: stable
    defaultChannel: stable
    packageName: entando-k8s-operator-rhmp
    provider:
      name: Entando, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: submariner.io
      provider-url: ""
    name: submariner
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: submariner.v0.17.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:067ba6e58a6fe150a212881a29c90aa7cbd22f53a6877eebc413072f73f9179a
          createdAt: "2025-06-18 01:35:57"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.16.0 <0.17.6'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.17 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:d58214791a60cd9e4366c6ff534584eedbd37386da71b868867e5e212df270f9
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:a565d8c094f3755b6c9303c5b8ae03a5883796e0fea754b0202e5792f1e472c3
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:067ba6e58a6fe150a212881a29c90aa7cbd22f53a6877eebc413072f73f9179a
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:b76998e8bbef1d06aa8dbc49af4f42a3c5603d480a67e46baf11168514c7759a
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:56848aefe1d156147ab73e7353017b2663f63e2e01fe43ede57abc72cfc0c50f
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:e6dbc4bf3e8b1a3c1a7ed10eb463d134f9ed28a03c7bb523c82c5431c5c9ae90
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:5510cb1bb3e0ca5f2365f4a5fc147ca693d3c220f1a8e7835f8b13b66b957ae8
        version: 0.17.6
      entries:
      - name: submariner.v0.17.6
        version: 0.17.6
      - name: submariner.v0.17.5
        version: 0.17.5
      - name: submariner.v0.17.4
        version: 0.17.4
      - name: submariner.v0.17.2
        version: 0.17.2
      - name: submariner.v0.17.2-0.1725901206.p
        version: 0.17.2+0.1725901206.p
      - name: submariner.v0.17.1
        version: 0.17.1
      - name: submariner.v0.17.0
        version: 0.17.0
      name: stable-0.17
    - currentCSV: submariner.v0.18.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:d8d7f0147c7b65b41903e101240905703992a7bbadb33e302aa86b23000d2095
          createdAt: "2025-05-07 20:54:56"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.17.0 <0.18.5'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.18 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:e5c16f0d720e15d54781870a5e63c85540256355009584f7d1bd03ef20a2eae4
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:00072fea3fdb5e294a01da0ee44165fc52219a965a27bdf7d792276f6c3a36c3
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:e8366433d1f50e2062b86d7f749f17cd6566197560256e77f57b2c165f277733
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:a07ef1bfec7549b32ad94f2c44a168860d5f3b46fd13593c068c58bb137c77fc
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:d8d7f0147c7b65b41903e101240905703992a7bbadb33e302aa86b23000d2095
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:27f18015b8896b04457c356d03276d50c34b8c1d84142ec07342ff2c1c4e81fa
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:e913be7a440b6cc0c1dddbda93a5979fe28b99e63f12a7949966b69126fcee90
        version: 0.18.5
      entries:
      - name: submariner.v0.18.5
        version: 0.18.5
      - name: submariner.v0.18.4
        version: 0.18.4
      - name: submariner.v0.18.3
        version: 0.18.3
      - name: submariner.v0.18.2
        version: 0.18.2
      - name: submariner.v0.18.0
        version: 0.18.0
      name: stable-0.18
    - currentCSV: submariner.v0.19.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:4afaf6689e2225c5c7f6fbe002202a1cdf9444e23fd210cc2d86da71c9f78626
          createdAt: "2025-04-22 21:12:45"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.18.0 <0.19.4'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.19 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:d9d4d03013d65d11a4e8ecda411015f359e4215a3a53059a37a810acf2e3ea6c
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:4afaf6689e2225c5c7f6fbe002202a1cdf9444e23fd210cc2d86da71c9f78626
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:5b83c5889aa0d8f7838dd5afb8c901dfb0ce0fbe01175e8468ec50ddedef9223
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:70008af49a9d1ee516204715f66bd415c9caa7462f6b178dc1dc91e513605fcd
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:ef73b8964cd66ffb496191056594e0886afa9d17c03976b4159413c89ed569bf
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:5938d67a0a2653e92f0995440700512992ed82269820679f68483dcc1b2d98d9
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:1bbb624bbefbb75407071449d0144258732e219e3a7fc67ac709c3bcee560576
        version: 0.19.4
      entries:
      - name: submariner.v0.19.4
        version: 0.19.4
      - name: submariner.v0.19.3
        version: 0.19.3
      - name: submariner.v0.19.2
        version: 0.19.2
      - name: submariner.v0.19.0
        version: 0.19.0
      name: stable-0.19
    - currentCSV: submariner.v0.20.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "quay.io/submariner",
                  "version": "0.20.2"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "quay.io/submariner",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "0.20.2"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: $(SUBMARINER_OPERATOR_IMAGE)
          createdAt: "2025-09-30 14:21:54"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/submariner-io/submariner-operator
          support: Submariner
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.20 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: submariner.io
        relatedImages:
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:fd6c46b4a0570b5e7b6457c465e5a1b843ba6f9c209c1856dbebfee1068e35ea
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:70d563a6e6526b9a146a2973210b2abdae09d5acab403fa3f0b465ed5e3c9b11
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:77e0da2db6d5bb2831e6c51214e08685c9f7f3b8b1e6ac53b4fd1418644ad6a1
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:0eef6521c96b29722076c1e220770124c8ed214379abc4654d9567cff8b52d04
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:8ea02aba8f9dd901f1f0ccff97ec8d157bebe5a90a6c924affe5e01cf3454e72
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:ed521b241da60efabfe2e2b80c51639d1a82a43945945879decbfb192bbc6d23
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:4a4ac198e9d0f4a4aaed8f0ec0e1d79bd8fa712b671ee697f7ff9d5616a02d84
        version: 0.20.2
      entries:
      - name: submariner.v0.20.2
        version: 0.20.2
      - name: submariner.v0.20.1
        version: 0.20.1
      - name: submariner.v0.20.0
        version: 0.20.0
      name: stable-0.20
    - currentCSV: submariner.v0.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "quay.io/submariner",
                  "version": "v0.21.0"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "quay.io/submariner",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v0.21.0"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: $(SUBMARINER_OPERATOR_IMAGE)
          createdAt: "2025-08-13 21:57:14"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/submariner-io/submariner-operator
          support: Submariner
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.21 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: submariner.io
        relatedImages:
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:cd851eab4a2214ea7a699bd01d2abac1322d01562f27b35f58f3c0b8716f88fb
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:bcdeaa67664568c5e2a6928236ff69b175ce65ddf7a58c1664d091b926000052
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:b0ad338aed82367048cf024767b0b09e6bb09bbf7dd5128e6711e1feec6f696e
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:65765c7f08c5454b557fbda91c49efbd2cc0d4d973cf8acc7ff40ec8a381839d
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:f5909b1802986622ab169a82d5dd7877af72660903e961b3097cc29e30eb9799
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:b307f748125fd459c3474fdab645688675a06515a9924b47cedba80e038458db
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:0b74f058400f24a63cf9b57a388914cc6dda38e03d639196871f3fd76fb7f2ae
        version: 0.21.0
      entries:
      - name: submariner.v0.21.0
        version: 0.21.0
      name: stable-0.21
    defaultChannel: stable-0.21
    packageName: submariner
    provider:
      name: submariner.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Devfile
      provider-url: https://devfile.io
    name: devworkspace-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devworkspace-operator.v0.36.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/internal-objects: '["devworkspaceroutings.controller.devfile.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DevWorkspaceOperatorConfig
            name: devworkspaceoperatorconfigs.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspaceRouting
            name: devworkspaceroutings.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha2
          - kind: DevWorkspaceTemplate
            name: devworkspacetemplates.workspace.devfile.io
            version: v1alpha1
          - kind: DevWorkspaceTemplate
            name: devworkspacetemplates.workspace.devfile.io
            version: v1alpha2
        description: |
          The DevWorkspace Operator enables cluster-level support for the
          [Devfile 2.0 spec](https://docs.devfile.io), enabling static, reproducible
          configurations for launching cloud-based editors and IDEs in OpenShift and
          Kubernetes clusters.

          Leveraging the language server protocol and the Eclipse Theia web IDE, the
          DevWorkspace operator provides easy configuration of full development
          environments on the cloud with support for a wide variety of languages and
          technologies, including Go, Java, Typescript/Javascript, Python, and more.

          The DevWorkspace Operator is also used in the Web Terminal Operator to
          automatically provision Web Terminal environments.

          ## Installing the operator
          The DevWorkspace Operator can be installed directly from the OperatorHub UI and
          will be available in all namespaces on the cluster. DevWorkspace creation is
          driven by the DevWorkspace custom resource, which can be created in any
          namespace to provision a full development environment. To get started, browse
          the DevWorkspace [spec](https://devfile.io/docs/2.3.0/devfile-schema).

          Once a DevWorkspace is started, it can be accessed via the URL in its
          `.status.mainUrl` field.

          It's recommended to install the DevWorkspace Operator to the
          `openshift-operators` namespace for compatibility.

          ## Uninstalling the operator
          The DevWorkspace Operator utilizes finalizers on resources it creates and
          webhooks to restrict access to development resources. As a result, manual steps
          are required in uninstalling the operator. See the
          [documentation](https://github.com/devfile/devworkspace-operator/blob/main/docs/uninstall.md)
          for details.
        displayName: DevWorkspace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Devworkspace Operator
        - DevWorkspaces
        - Devfile
        links:
        - name: Devworkspace Operator
          url: https://github.com/devfile/devworkspace-operator
        maintainers:
        - email: dakwon@redhat.com
          name: David Kwon
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        maturity: alpha
        provider:
          name: Devfile
          url: https://devfile.io
        relatedImages:
        - registry.redhat.io/devworkspace/devworkspace-rhel9-operator@sha256:83b6abf6e57703da1125e037ea9353c224602bddc3146e5b00df47f1b1976e9a
        - registry.redhat.io/ubi9/ubi-minimal@sha256:a50731d3397a4ee28583f1699842183d4d24fadcc565c4688487af9ee4e13a44
        - registry.redhat.io/devworkspace/devworkspace-project-clone-rhel9@sha256:312a013201877cea03956e3db2ba19a3d7f842de2c3a65927c65c3e8d22a57fe
        version: 0.36.0
      entries:
      - name: devworkspace-operator.v0.36.0
        version: 0.36.0
      - name: devworkspace-operator.v0.35.1
        version: 0.35.1
      - name: devworkspace-operator.v0.34.0
        version: 0.34.0
      - name: devworkspace-operator.v0.33.0
        version: 0.33.0
      - name: devworkspace-operator.v0.32.1
        version: 0.32.1
      - name: devworkspace-operator.v0.31.2
        version: 0.31.2
      - name: devworkspace-operator.v0.31.1
        version: 0.31.1
      - name: devworkspace-operator.v0.31.0
        version: 0.31.0
      - name: devworkspace-operator.v0.30.2
        version: 0.30.2
      - name: devworkspace-operator.v0.30.1
        version: 0.30.1
      - name: devworkspace-operator.v0.30.0
        version: 0.30.0
      - name: devworkspace-operator.v0.29.0
        version: 0.29.0
      - name: devworkspace-operator.v0.28.0-0.1719314743.p
        version: 0.28.0+0.1719314743.p
      - name: devworkspace-operator.v0.28.0
        version: 0.28.0
      - name: devworkspace-operator.v0.27.0-0.1714987663.p
        version: 0.27.0+0.1714987663.p
      - name: devworkspace-operator.v0.27.0
        version: 0.27.0
      - name: devworkspace-operator.v0.26.0
        version: 0.26.0
      - name: devworkspace-operator.v0.25.0
        version: 0.25.0
      - name: devworkspace-operator.v0.24.0
        version: 0.24.0
      - name: devworkspace-operator.v0.23.0
        version: 0.23.0
      - name: devworkspace-operator.v0.22.0
        version: 0.22.0
      - name: devworkspace-operator.v0.21.2
        version: 0.21.2
      - name: devworkspace-operator.v0.21.1
        version: 0.21.1
      - name: devworkspace-operator.v0.20.0
        version: 0.20.0
      - name: devworkspace-operator.v0.19.1
        version: 0.19.1
      - name: devworkspace-operator.v0.19.1-0.1682321189.p
        version: 0.19.1+0.1682321189.p
      - name: devworkspace-operator.v0.19.1-0.1679521112.p
        version: 0.19.1+0.1679521112.p
      - name: devworkspace-operator.v0.18.1
        version: 0.18.1
      - name: devworkspace-operator.v0.18.1-0.1675929565.p
        version: 0.18.1+0.1675929565.p
      - name: devworkspace-operator.v0.17.0
        version: 0.17.0
      - name: devworkspace-operator.v0.16.0-0.1666668361.p
        version: 0.16.0+0.1666668361.p
      - name: devworkspace-operator.v0.16.0
        version: 0.16.0
      - name: devworkspace-operator.v0.15.2-0.1661828401.p
        version: 0.15.2+0.1661828401.p
      - name: devworkspace-operator.v0.15.2
        version: 0.15.2
      - name: devworkspace-operator.v0.14.1
        version: 0.14.1
      - name: devworkspace-operator.v0.13.0
        version: 0.13.0
      - name: devworkspace-operator.v0.12.0
        version: 0.12.0
      - name: devworkspace-operator.v0.11.0
        version: 0.11.0
      - name: devworkspace-operator.v0.10.0
        version: 0.10.0
      - name: devworkspace-operator.v0.9.0
        version: 0.9.0
      name: fast
    defaultChannel: fast
    packageName: devworkspace-operator
    provider:
      name: Devfile
      url: https://devfile.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microcks.io
      provider-url: ""
    name: microcks
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: microcks-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "microcks.github.io/v1alpha1",
                "kind": "MicrocksInstall",
                "metadata": {
                  "name": "my-microcksinstall"
                },
                "spec": {
                  "name": "my-microcksinstall",
                  "version": "1.10.0",
                  "microcks": {
                    "replicas": 1
                  },
                  "postman": {
                    "replicas": 1
                  },
                  "keycloak": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "1Gi"
                  },
                  "mongodb": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "2Gi"
                  }
                }
              },
              {
                "apiVersion": "microcks.github.io/v1alpha1",
                "kind": "MicrocksInstall",
                "metadata": {
                  "name": "my-microcksinstall-minikube"
                },
                "spec": {
                  "name": "my-microcksinstall-minikube",
                  "version": "1.10.0",
                  "microcks": {
                    "replicas": 2,
                    "url": "microcks.192.168.99.100.nip.io"
                  },
                  "postman": {
                    "replicas": 2
                  },
                  "keycloak": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "1Gi",
                    "url": "keycloak.192.168.99.100.nip.io"
                  },
                  "mongodb": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "2Gi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/microcks/microcks-ansible-operator:1.10.0
          createdAt: "2024-08-01T09:14:55Z"
          description: Open Source mocking and testing platform for API and microservices
          repository: https://github.com/microcks/microcks-ansible-operator
          support: microcks.github.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Microcks installation
            displayName: MicrocksInstall
            kind: MicrocksInstall
            name: microcksinstalls.microcks.github.io
            version: v1alpha1
        description: |
          Microcks is an Open source tool for API and microservices mocking and testing. It provides a platform for referencing, deploying mocks and allow contract-testing of your APIs. It can also be considered as a Service Virtualization solution because it will allow you to provide fake API or Service implementation before development being actually done. It supports Async APIs, REST APIs, gRPC, GraphQL and SOAP WebServices. It perfectly integrates into an iterative, contract-first delivery process.
          Microcks tries not reinventing the wheel and let you capitalize on already known tools of standards. It supports [AsyncAPI Specifications](https://asyncapi.com), [OpenAPI Specification](https://www.openapis.org/) `3.x`, [gRPC Protobuf](https://grpc.io/), [GraphQL Schema](https://graphql.org) out-of-the-box and can also relies on popular tools like [SoapUI](http://www.soapui.com) or [Postman](http://www.getpostman.io) to edit your API request / response samples as well as your test scripts.
          ### Operator features
          * **Create/Destroy**: Easily launch Microcks installations (`MicrocksInstall` custom resource) for your Kubernetes namespace
          ### MicrocksInstall options
          * **keycloak** component can be deployed as part of the install or you can reuse an existing instance. You can reference an existing `Secret` for providing TLS certificates for securing the `Ingress`,
          * **mongodb** component can be deployed as part of the install or you can reuse an existing instance, You can reference an existing `Secret` for providing connection credentials,
          * **features** component allow to activate optional features like the mocking of AsyncAPIs. In case you turn `features.async.enabled` to `true` and `features.async.kafka.install` to `true`, you must have the [Strimzi Operator](https://operatorhub.io/operator/strimzi-kafka-operator) already installed on your cluster/namespace.
          ### Documentation
          Documentation to the current _master_ branch as well as all releases can be found on our [GitHub](https://github.com/microcks/microcks-ansible-operator).
          ### License
          Microcks is licensed under the [Apache License, Version 2.0](https://github.com/microcks/microcks/blob/master/LICENSE).
        displayName: Microcks Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mock
        - test
        - API
        - microservice
        links:
        - name: Product Page
          url: https://microcks.io
        - name: Microcks Operator Source Code
          url: https://github.com/microcks/microcks-ansible-operator
        maintainers:
        - email: laurent.broudoux@gmail.com
          name: lbroudoux
        maturity: alpha
        provider:
          name: Microcks.io
        relatedImages:
        - quay.io/microcks/microcks-ansible-operator:1.10.0
        version: 1.10.0
      entries:
      - name: microcks-operator.v1.10.0
        version: 1.10.0
      - name: microcks-operator.v1.9.0
        version: 1.9.0
      - name: microcks-operator.v1.8.1
        version: 1.8.1
      - name: microcks-operator.v1.8.0
        version: 1.8.0
      - name: microcks-operator.v1.7.1
        version: 1.7.1
      - name: microcks-operator.v1.7.0
        version: 1.7.0
      - name: microcks-operator.v1.6.1
        version: 1.6.1
      - name: microcks-operator.v1.6.0
        version: 1.6.0
      - name: microcks-operator.v1.5.2
        version: 1.5.2
      - name: microcks-operator.v1.5.1
        version: 1.5.1
      - name: microcks-operator.v1.5.0
        version: 1.5.0
      - name: microcks-operator.v1.4.1
        version: 1.4.1
      name: stable
    defaultChannel: stable
    packageName: microcks
    provider:
      name: Microcks.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-wafv2-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-wafv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wafv2.services.k8s.aws/v1alpha1",
                "kind": "IPSet",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/wafv2-controller:1.1.1
          createdAt: "2025-09-20T21:41:21Z"
          description: AWS WAFV2 controller is a service controller for managing WAFV2
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IPSet represents the state of an AWS wafv2 IPSet resource.
            displayName: IPSet
            kind: IPSet
            name: ipsets.wafv2.services.k8s.aws
            version: v1alpha1
          - description: RuleGroup represents the state of an AWS wafv2 RuleGroup
              resource.
            displayName: RuleGroup
            kind: RuleGroup
            name: rulegroups.wafv2.services.k8s.aws
            version: v1alpha1
          - description: WebACL represents the state of an AWS wafv2 WebACL resource.
            displayName: WebACL
            kind: WebACL
            name: webacls.wafv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Web Application Firewall (Amazon WAFV2) resources in AWS from within your Kubernetes cluster.

          **About Amazon WAFV2**

          AWS WAF is a web application firewall that lets you monitor the HTTP and HTTPS requests that are forwarded to your protected web application resources. You can protect the following resource types:
          - Amazon CloudFront distribution
          - Amazon API Gateway REST API
          - Application Load Balancer
          - AWS AppSync GraphQL API
          - Amazon Cognito user pool
          - AWS App Runner service
          - AWS Verified Access instance

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon WAFV2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - wafv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon WAFV2 Developer Resources
          url: https://aws.amazon.com/waf/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: wafv2 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/wafv2-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-wafv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-wafv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-wafv2-controller.v1.0.14
        version: 1.0.14
      - name: ack-wafv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-wafv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-wafv2-controller.v1.0.10
        version: 1.0.10
      - name: ack-wafv2-controller.v1.0.9
        version: 1.0.9
      - name: ack-wafv2-controller.v1.0.8
        version: 1.0.8
      - name: ack-wafv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-wafv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-wafv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-wafv2-controller.v1.0.1
        version: 1.0.1
      - name: ack-wafv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-wafv2-controller.v0.0.4
        version: 0.0.4
      - name: ack-wafv2-controller.v0.0.3
        version: 0.0.3
      - name: ack-wafv2-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-wafv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/in-en/products/cloud-pak-for-data
    name: ibm-zen-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-zen-operator.v6.1.3
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.3 <6.1.3'
          operators.operatorframework.io/builder: operator-sdk-v1.6.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: represents zen instance
            kind: ZenService
            name: zenservices.zen.cpd.ibm.com
            version: v1
          - description: represents zen extension instance
            kind: ZenExtension
            name: zenextensions.zen.cpd.ibm.com
            version: v1
        description: |
          The Zen operator is a component in IBM® Cloud Pak foundational services. It serves a framework for user access management, defining user roles and for powering persona driven user experiences for Cloud Paks and Service offerings
        displayName: IBM Zen Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cloud Pak
        - IBM
        - Access management
        - Platform UI
        - Zen framework
        maintainers:
        - email: cp4d@ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/in-en/products/cloud-pak-for-data
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:deb66a64288a370c5da5c7cc0ba4e2ba4ab816d7c5a080a0c505b05f55fa6072
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:a7c050ae29cb93754f59d8f83676275ee9a91ae9a9956395b97abf8c5e20e96e
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:abd96661823553d623eb0ea1fac8271d64391243dd0f13eb4a4fcc858dfe10a5
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:1df782d0d923544767696d719bd0ac10802b8e06508a104af5e023a519f32dee
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:82afac013ef9b6b9d0eacb4575c979b65abd6f2821971182b90c71bd488ef366
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:719b0871f0d32da894ec80dc2a23483623b7802a5c6f935cf3672468f99faa40
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:a99b1f5a4eb0306b83562de0184fc826c34835ffbf46d929ceb9e70e0d325361
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:b5a8fb14c3bb7c89ff970ca85f06dfcdca832e2b45ad86e8bb1b6c0a0e7e0764
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:b87fa8374976ccd3dad90ddbf6b65cd6ae5a028ba896562f284f089380ebe44a
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:525924934f74b28783dacf630a6401ccd4c88d253aafa2d312b7993d8b9a14c1
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:a8d2e42757c6ccbfe4372465b1827c633aad012c5194bf8b08afcd5691caa6f5
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:10ede3c4e3c7e0cbc26e7b6e797cad35e39face1a1b9d4a4e2440ab0feeeb814
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:8444703b2a7822f82141afcf4f36bf255650d755321b9e9fe2cbd1bdea41d5c7
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:ecbc65705de76804a3c7642339ecfb4cc0f5a61891058e07c20bf2b63f1b82b8
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:3d5c6573faf4ae00cbb21cbfe9b2ad5cd8ffb658df0c5ba09bd98b3bf54f9979
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:e9173d81ef2a75839c0e1d677b72bb00b8e6d7ef0511afab103f08a884a03738
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:c8baecad9c97ddbbb3f9f02c90abbf483d46974b7bc9605622a302cbeba1de74
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8da8b240176c00f8f6eadf32e5614c663cc1d56d546c73996cca614b715acf4a
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:ad027c91c719fd1d931cbac8c76735474f676b228f9ab23d1f243201246b3238
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:c4ff5dae4560e5a943c7218df3a5aa9ca1a9ee2d1f7dfac73db05892725c34a9
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:183ee860bf48076ea155db08032eed7004c1204a363625f54218734b00175441
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:45194b56f9fcb3af83158eb0297248364db40e8164fb4fe4f5656c58aeac3fa5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:c6862158a2a5ae0a4e4206bfdc7f73e1c2359544966c72cf4fc3a617acb13697
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:443b51f8b10acc85bdefde7193e0f45b1c423fa7fbdcaa28342805815c43db3d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:ea57c7cb1ccd5bce17f6cb35e326ab974718be8cd19b0cd38ea8112e91c51939
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:5f48081ffa453ec740c2562c24603d78b481b5e236d6bcc73e96203cabe153f2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:f06a1f50cd18d6f48762fdd0394307b44c0573d0cc91f53724c5b3c8c5e2b8e4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:45ca5370b726e31c55f9d7d249374189cb48dc05ab89d675c35225d53b90e7c9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:6bbcb64cf9adb4d3b6a4c259eb910e7da5c3c58732643e60d416399b2083cc0d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:fa01c7d11c23a47bb84375389e69619455c2879b5ed6bb5b9b753b945dcb02c3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:0522fc5482404fa9be179c7225a5692c8d71d755fda0b2d3928e452adcf030fb
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-bundle@sha256:552fc53664722903eccaa7d3782a3e84413509fd357725fe0b0f4a077c48ab2c
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:a72c2b70a9f0a93047d47247cddc1616138cbc87337289ddb208565af6e4369d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:7dbff355db7739152961cb6a97887d2e43bc960ac58837c126e03ed1a4480a3a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ee161954766211651d7cb13f3f5a3c025ac04094ba742be453ce789c6f0e24a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:aaee5240471872a8f1a77d4440a696d45041c5ce06849d516daccb90f5341b6d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:24dff64a80b910f0b4c5442d4f611f9df7602a16c5231eff2c9d08f07b2a9f9a
        version: 6.1.3
      entries:
      - name: ibm-zen-operator.v6.1.3
        version: 6.1.3
      name: v6.1
    defaultChannel: v6.1
    packageName: ibm-zen-operator
    provider:
      name: IBM
      url: https://www.ibm.com/in-en/products/cloud-pak-for-data
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: hpe-csi-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-csi-operator.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.hpe.com/v1",
                "kind": "HPECSIDriver",
                "metadata": {
                  "name": "hpecsidriver-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "csp": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "disable": {
                    "alletra6000": false,
                    "alletra9000": false,
                    "alletraStorageMP": false,
                    "b10000FileService": false,
                    "nimble": false,
                    "primera": false
                  },
                  "disableHostDeletion": false,
                  "disableNodeConfiguration": false,
                  "disableNodeConformance": false,
                  "disableNodeGetVolumeStats": false,
                  "disableNodeMonitor": false,
                  "disablePreInstallHooks": false,
                  "imagePullPolicy": "IfNotPresent",
                  "images": {
                    "b10000FileServiceCSP": "quay.io/hpestorage/alletrastoragemp-b10000-nfs-csp@sha256:6a1855a2ef1a83b886bb1f45b596e8e420e5e13c84bbf446c8c202eaf27a2013",
                    "csiAttacher": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                    "csiControllerDriver": "quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676",
                    "csiExtensions": "quay.io/hpestorage/csi-extensions@sha256:27074aaa8df3b867cfa3d12318612c45f8e541d750c27db5e080734592160c89",
                    "csiNodeDriver": "quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676",
                    "csiNodeDriverRegistrar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                    "csiProvisioner": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                    "csiResizer": "registry.k8s.io/sig-storage/csi-resizer@sha256:8ddd178ba5d08973f1607f9b84619b58320948de494b31c9d7cd5375b316d6d4",
                    "csiSnapshotter": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769",
                    "csiVolumeGroupProvisioner": "quay.io/hpestorage/volume-group-provisioner@sha256:0a5f39f52113b963978a695bb8e5a5a3c4d387a02eb4d4a9647b4bc804de7399",
                    "csiVolumeGroupSnapshotter": "quay.io/hpestorage/volume-group-snapshotter@sha256:177d9abf8fbd3ac26ca4c12ca4b4adde20ebab308c79f10b04632f0171746fa8",
                    "csiVolumeMutator": "quay.io/hpestorage/volume-mutator@sha256:bf6beac1af5edcd3c9774f0dcc9579b79ef9cc8e4b74a4238082ab9aa443b5ed",
                    "nfsProvisioner": "quay.io/hpestorage/nfs-provisioner@sha256:5bf5952453ecca28e26724df85ebd5c8f0b7479ae4cb683d4a09c415da00e002",
                    "nimbleCSP": "quay.io/hpestorage/alletra-6000-and-nimble-csp@sha256:5480ffdebd74fdc547bc4a11b4434abac7bbfee3cee349f1ac893fbad84bb8d1",
                    "primera3parCSP": "quay.io/hpestorage/alletra-9000-primera-and-3par-csp@sha256:8af6e87075c47ba8d9714d55848ab2a0fef62b0768681e30d9e63ed7f396b0bf"
                  },
                  "iscsi": {
                    "chapSecretName": ""
                  },
                  "kubeletRootDir": "/var/lib/kubelet",
                  "logLevel": "info",
                  "maxVolumesPerNode": 100,
                  "node": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/hpestorage/csi-driver-operator@sha256:29e92f68cf0d7691a6c56876e7dde0d69a8cb3ff6ae6c9371c604e61cba16898
          createdAt: "2025-09-25T16:28:16Z"
          description: A Container Storage Interface (CSI) driver for HPE storage
            platforms. The CSI driver allows you to use HPE storage with your preferred
            container orchestrator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <3.0.1'
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/hpe-storage/co-deployments
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE CSI Driver for Kubernetes
            displayName: HPECSIDriver
            kind: HPECSIDriver
            name: hpecsidrivers.storage.hpe.com
            version: v1
        description: |
          The HPE CSI Operator for OpenShift deploys the HPE CSI Driver on Kubernetes and OpenShift for dynamic provisioning of persistent volumes on HPE storage systems.
          The HPE CSI Driver for Kubernetes leverages HPE storage platforms to provide scalable and persistent storage for stateful applications.
          Please refer to our CSI driver [documentation](https://scod.hpedev.io/csi_driver/index.html) for supported HPE platforms and full list of supported CSI features.
          ## Installation
          Refer to the HPE CSI Operator for OpenShift [documentation](https://scod.hpedev.io/csi_driver/deployment.html#operator).
        displayName: HPE CSI Operator for OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE CSI
        - HPE Storage
        - CSI
        links:
        - name: Documentation
          url: https://scod.hpedev.io/csi_driver/index.html
        maintainers:
        - email: dev-hi-containers@hpe.com
          name: HPE Storage Containers Team
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: HPE
        relatedImages:
        - quay.io/hpestorage/volume-mutator@sha256:bf6beac1af5edcd3c9774f0dcc9579b79ef9cc8e4b74a4238082ab9aa443b5ed
        - quay.io/hpestorage/alletrastoragemp-b10000-nfs-csp@sha256:6a1855a2ef1a83b886bb1f45b596e8e420e5e13c84bbf446c8c202eaf27a2013
        - quay.io/hpestorage/nfs-provisioner@sha256:5bf5952453ecca28e26724df85ebd5c8f0b7479ae4cb683d4a09c415da00e002
        - quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8ddd178ba5d08973f1607f9b84619b58320948de494b31c9d7cd5375b316d6d4
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769
        - quay.io/hpestorage/alletra-6000-and-nimble-csp@sha256:5480ffdebd74fdc547bc4a11b4434abac7bbfee3cee349f1ac893fbad84bb8d1
        - quay.io/hpestorage/volume-group-snapshotter@sha256:177d9abf8fbd3ac26ca4c12ca4b4adde20ebab308c79f10b04632f0171746fa8
        - quay.io/hpestorage/volume-group-provisioner@sha256:0a5f39f52113b963978a695bb8e5a5a3c4d387a02eb4d4a9647b4bc804de7399
        - quay.io/hpestorage/alletra-9000-primera-and-3par-csp@sha256:8af6e87075c47ba8d9714d55848ab2a0fef62b0768681e30d9e63ed7f396b0bf
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.connect.redhat.com/hpestorage/csi-driver-operator@sha256:29e92f68cf0d7691a6c56876e7dde0d69a8cb3ff6ae6c9371c604e61cba16898
        - quay.io/hpestorage/csi-extensions@sha256:27074aaa8df3b867cfa3d12318612c45f8e541d750c27db5e080734592160c89
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        version: 3.0.1
      entries:
      - name: hpe-csi-operator.v3.0.1
        version: 3.0.1
      name: stable
    defaultChannel: stable
    packageName: hpe-csi-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SmartX
      provider-url: https://www.smartx.com
    name: iomesh-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: iomesh-operator.0.10.1-rc4
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
          createdAt: 2021-08-17 00:00:00+00:00
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/iomesh-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/iomesh-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/iomesh/docs
          support: SmartX, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: iomesh
            displayName: iomesh
            kind: IOMeshCluster
            name: iomeshclusters.iomesh.com
            version: v1alpha1
        description: "IOMesh (https://www.iomesh.com/) is a cloud-native storage product
          specifically designed and developed for Kubernetes. It takes ZBS as its
          core, a “production-ready” distributed storage system developed independently
          by SmartX, to provide production-level high performance and reliable persistent
          storage for business-critical applications such as MySQL, Cassandra, MongoDB,
          etc., contributing to the containerized transformation of stateful applications.
          \     \n## Pre-Requisites\n* According to https://docs.iomesh.com/deploy/prerequisites,
          make sure that your openshift cluster meets the \"Installation Requirements\"
          and set up the worker nodes correctly according to the \"Setup Worker Node\"
          chapter.\n* Run IOMesh Operator pre-installation script in an environment
          where oc or kubectl can be used to access the openshift cluster, the script
          will install the dependencies of IOMesh Operator and specific IOMesh settings
          for the openshift cluster:\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-pre-install-openshift.sh
          | sh -\n```\n## Install IOMesh Operator and IOMesh Cluster\n* Click install
          button on the current web page to install IOMesh Operator\n* On the **Installed
          Operators** -> **IOMesh Operator** -> **Create instance** -> **YAML view**,
          fill in an IOMesh Custom Resources according https://iomesh.run/iomesh.yaml,
          change the `spec.*.dataCIDR` to your own data network CIDR:\n## Post Install\n*
          Run IOMesh Operator post-installation script in an environment where oc
          or kubectl can be used to access the openshift cluster, the script will
          install IOMesh CSI Driver\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-post-install-openshift.sh
          | sh -\n```\n## Getting Started\n* Try to quickstart setup and use IOMesh:
          https://docs.iomesh.com/deploy/setup-iomesh\n"
        displayName: IOMesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iomesh-operator
        links:
        - name: IOMesh Operator
          url: https://iomesh.com
        maintainers:
        - email: na.wang@smartx.com
          name: Na Wang
        maturity: alpha
        provider:
          name: SmartX
          url: https://www.smartx.com
        relatedImages:
        - iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
        version: 0.10.1-rc4
      entries:
      - name: iomesh-operator.0.10.1-rc4
        version: 0.10.1-rc4
      name: stable
    defaultChannel: stable
    packageName: iomesh-operator-rhmp
    provider:
      name: SmartX
      url: https://www.smartx.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: openstack-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openstack-operator.v1.0.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openstack.org/v1beta1",
                "kind": "OpenStack",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openstack-operator",
                    "app.kubernetes.io/instance": "openstack",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "openstack",
                    "app.kubernetes.io/part-of": "openstack-operator"
                  },
                  "name": "openstack",
                  "namespace": "openstack-operators"
                },
                "spec": {
                  "operatorOverrides": [
                    {
                      "controllerManager": {
                        "resources": {
                          "limits": {
                            "cpu": "600m",
                            "memory": "2Gi"
                          }
                        }
                      },
                      "name": "infra",
                      "replicas": 1
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-02T13:33:09Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.14
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.openstack.org/v1beta1","kind":"OpenStack","metadata":{"name":"openstack","namespace":"openstack-operators"},"spec":{}}'
          operatorframework.io/suggested-namespace: openstack-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenStack is the Schema for the openstacks API
            displayName: Open Stack
            kind: OpenStack
            name: openstacks.operator.openstack.org
            version: v1beta1
        description: Install and configure OpenStack
        displayName: OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        links:
        - name: Openstack Operator
          url: https://github.com/openstack-k8s-operators/
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso-operators/glance-rhel9-operator@sha256:ee22cf417b7771930a442c2302277e6d37f368f5ae551fee9a50dbbe41c8fe3a
        - registry.redhat.io/rhoso-operators/mysqld-exporter-rhel9@sha256:b49e966cdd588071e0e1f2791ec6c62310078a7d39ea77f83843d3708a07c2b5
        - registry.redhat.io/rhoso/openstack-swift-proxy-server-rhel9@sha256:858551fd1c519797cae0e146e9c05412ed6e66dac01d0e68e62dd2d4bd335391
        - registry.redhat.io/rhoso/openstack-netutils-rhel9@sha256:b02ab1617922a2560bf739172bc2ce0f6b1fb3ef55054bd17e4b16bdf9512b4b
        - registry.redhat.io/rhoso/openstack-cinder-backup-rhel9@sha256:910693233ee3a1f3832ebe744f67a5139249523456de4abbd731c9d2f544b7c3
        - registry.redhat.io/rhoso/ironic-python-agent-rhel9@sha256:2ccd8e22a242059ca5c4e2bce853c36a71c36fdf40e8e36ab183a8218b0cc87b
        - registry.redhat.io/rhoso/openstack-frr-rhel9@sha256:f6516277149d1560c0bf2ac74a99cfb4ac2c84f9f090ab231b6d12bdb824de3b
        - registry.redhat.io/rhoso/openstack-octavia-housekeeping-rhel9@sha256:0a79259e398af746d0e8a6dda1aae5ee976c6bf8991b17f4a5da0d07dca1e36c
        - registry.redhat.io/rhoso/openstack-swift-object-rhel9@sha256:8fc2ab8caa9c7c35d94ef51c303763ea71e40b9c795c4555c17e6194344419b3
        - registry.redhat.io/ubi9/httpd-24@sha256:29382518403a4361c48fccc057163dc4a9f43ee159432754fa63c76485b223f5
        - registry.redhat.io/rhoso-operators/heat-rhel9-operator@sha256:583abb81a49db85949f7a4b846591ad8532afa79440ad7e2937ee00eacd972b3
        - registry.redhat.io/rhoso/openstack-barbican-keystone-listener-rhel9@sha256:cd8d4f0a94b4fe080ebf8e459f091db6637ceec6d83ac23cf04d66bd03c71e76
        - registry.redhat.io/rhoso/openstack-neutron-metadata-agent-ovn-rhel9@sha256:20323b151474703df610b6c24ccb9c748265196bbc70a7b5ece1b801ff540cd9
        - registry.redhat.io/rhoso-operators/test-rhel9-operator@sha256:df7566483abb18aa3c45d8bc4cd91531c7eb5b63345e511151a7cdca2dbcef17
        - registry.redhat.io/rhoso/openstack-iscsid-rhel9@sha256:508bd67021c271a10bb55ab4af033dc4e931708674b9c5ba94f72c07cd0c8ee2
        - registry.redhat.io/rhoso-operators/openstack-baremetal-agent-rhel9@sha256:08675b16d89d7931b1ef7b96e1e884c88e8e7fbfbe79a944f9213c52aba3ba83
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/rhoso/openstack-ironic-neutron-agent-rhel9@sha256:0abaa8f437025d2e7006c59c546ff6935556d707b57993719faef94142e1c43d
        - registry.redhat.io/rhoso/openstack-aodh-evaluator-rhel9@sha256:48c9192acf157df5b8f0827e35f0380d50cd4562676aba3c0d7791f92328af02
        - registry.redhat.io/rhoso/openstack-manila-scheduler-rhel9@sha256:8ce47f7fd8660167b3b292659d4fc58cf34ed671d5779693d24b71a9fddc1e7c
        - registry.redhat.io/rhoso/openstack-neutron-dhcp-agent-rhel9@sha256:cc9223ef705bcd5d0dcabbf71ad9f6e540ae7f003a6afd13161301f842dfa5fa
        - registry.redhat.io/rhoso/edpm-hardened-uefi-rhel9@sha256:c18ec7023fbcc4ab5183f689adc1f01a626a4923c15c61a37388f99904db7bc3
        - registry.redhat.io/rhoso/openstack-ceilometer-central-rhel9@sha256:e6d9f47406b7e3ff2a7f2256088cc64f043b8e497382b9f569eb90182598bbe9
        - registry.redhat.io/rhoso/openstack-cinder-api-rhel9@sha256:72a674caeeeecce3b60217f3607ad5b00651fa3d52f2f5d905f6e6395638dbc8
        - registry.redhat.io/rhoso/openstack-glance-api-rhel9@sha256:b1c0b09bb9aafe3bc31f43f8a99a92dde8d78cf90e781dd14300580986d8c980
        - registry.redhat.io/rhoso-operators/designate-rhel9-operator@sha256:d446af622bd32666eb4eb5f9b7f38f0c33ac81251a2230db9d5baf71081b9d6a
        - registry.redhat.io/rhoso/openstack-designate-producer-rhel9@sha256:04af4230f815a6e3fde22003cf382eebd9cbc3c3ec4056795f4474abc1e65fea
        - registry.redhat.io/rhoso-operators/mariadb-rhel9-operator@sha256:907e3a9d15941440d5289d7e448d9f0351053fc20fcb8cb3b9db0a84debd876f
        - registry.redhat.io/rhoso/openstack-manila-api-rhel9@sha256:e4dd3d5b66c5216b45a1515a975b85fd2ac4ef698a9350dd16e01417da462b32
        - registry.redhat.io/rhoso/openstack-cinder-volume-rhel9@sha256:8a09e80dac94fb4b3d35033e91812f95f0acac66532d9abc88aa1fb715d21822
        - registry.redhat.io/rhoso/openstack-designate-worker-rhel9@sha256:a6324a8a72722822df7692b8df0420388a1b4aecfff3da7e91da4afe4d9e2260
        - registry.redhat.io/rhoso/openstack-ceilometer-compute-rhel9@sha256:296f7f6c6bdcd7f5d07b8470308a4c71055f29a17c48d2ebb03b7f54a0e927ff
        - registry.redhat.io/rhoso/openstack-heat-api-cfn-rhel9@sha256:2789929ae3edd7c68579f34457d3071273e49a66f462043cd9c46e75bdda282a
        - registry.redhat.io/openshift4/ose-kube-state-metrics-rhel9@sha256:db6074b8aeedfd7603ae3adbe67071115f54e7f01a24997f1cd8fa8958aeddce
        - registry.redhat.io/rhoso-operators/cinder-rhel9-operator@sha256:12b8dd8fd3683250365eff6f07339cb976ee29da80642e591c10e9fa5ba627c9
        - registry.redhat.io/rhoso/openstack-heat-engine-rhel9@sha256:151e04b3e5a408aadef3a17954255acc3733d09961878a135e177188457447f0
        - registry.redhat.io/rhoso/openstack-rsyslog-rhel9@sha256:272e729accc2b64b8fcc0ece4ae5aa9ce6d454e34a25a8a8e2b99ac56b73361e
        - registry.redhat.io/rhoso/openstack-tempest-all-rhel9@sha256:16b1fbfd36ce9a27106bf09c414eaa93f14c9a6b11839068976c2fdb537d2ba7
        - registry.redhat.io/rhoso/openstack-octavia-api-rhel9@sha256:7b8d5927162ddab512096beb02e837ba51f6353c5187f3de72eb3db6e56ab6e7
        - registry.redhat.io/rhoso/openstack-keystone-rhel9@sha256:ac076c81247d774a9aeee1b63540ca6e18058213cf6aad10a6fd1f535a394ef4
        - registry.redhat.io/rhoso/openstack-cinder-scheduler-rhel9@sha256:2079ffe9b620e9277573ac796bf7e393b520631130e5bbb6f5ab51d872c943f3
        - registry.redhat.io/rhoso/openstack-aodh-notifier-rhel9@sha256:d2e89f27aac9afa504411ae6a6163442cd717f8c560e505a7da8e3fd9e2cff95
        - registry.redhat.io/rhoso/openstack-manila-share-rhel9@sha256:277174c885998e6f5b867d099664f6fa28f974bffb524903ff75283035cd8dc9
        - registry.redhat.io/rhoso/openstack-neutron-server-rhel9@sha256:2d3afde3201aa18b192ad0fb079c1e4f4f673565604f1af2b79aba8a7bc88bf2
        - registry.redhat.io/rhoso/openstack-ovn-northd-rhel9@sha256:668f742b2a18579bc8cfc610a862e43c5b877413cf65eabbe0ee36b3c7869446
        - registry.redhat.io/rhoso-operators/openstack-baremetal-rhel9-operator@sha256:03916e507b88943c0e00cedeb3cf20fa0787ccc762e47f38d305ff4d1b125b12
        - registry.redhat.io/rhoso/openstack-ovn-nb-db-server-rhel9@sha256:8c2ce1027779dc66d67a0e0d03e48b226d3278208cae0b2b81b35ad1b1f3ae72
        - registry.redhat.io/rhoso/openstack-neutron-ovn-agent-rhel9@sha256:bed1956c66b0e70064b51e34faac64036f77c7ecb14f46ceac4bc9ef52f0b9e0
        - registry.redhat.io/rhoso/openstack-nova-scheduler-rhel9@sha256:f0e5d3a693c9998ca3839b3ec7dfd5da0f9cf139e0b5b6913e92a10d5a13e6bc
        - registry.redhat.io/rhoso/openstack-designate-central-rhel9@sha256:d28d8d91d8bf75a1b77a5d650c7f2650a1791e2a1b0341450fa46978c955b92f
        - registry.redhat.io/rhoso/openstack-barbican-api-rhel9@sha256:bc942204843c45c6cb7ea841af8611f04efb003de7eeccc6813934aad556da73
        - registry.redhat.io/rhoso/openstack-octavia-health-manager-rhel9@sha256:8065851058019757b56f86ca4f030d44f169100eb2e37e08e0587a7b1e4d2e72
        - registry.redhat.io/rhoso/openstack-nova-api-rhel9@sha256:4116206f1fbe60fdd9458a5edee848571c94e2b0512d9c2cfcfd6890d60d4395
        - registry.redhat.io/rhoso-operators/keystone-rhel9-operator@sha256:879d4b1f9e720f8581b3a55c788a97bbe7b52f9108df3d5dce2e9bcb4200df6d
        - registry.redhat.io/rhoso/openstack-horizon-rhel9@sha256:82d0e6d24733afd305996e0f4a118df3e48621c40c9669b01f48669dbf40d868
        - registry.redhat.io/rhoso/openstack-designate-api-rhel9@sha256:b76ca7820553285bb4ee2bfb08e39aedbfe8c16b4edb478c06418a1e335a64ab
        - registry.redhat.io/rhoso/openstack-swift-container-rhel9@sha256:95fec49643fc18f402b8ba8057f2c35f214bf6138c5f0d7bbe1b96cbd78ef8d6
        - registry.redhat.io/rhoso-operators/swift-rhel9-operator@sha256:c99ef14338290d63e2bcbf129e953b434aea382e47072812bae31b1ea838e7df
        - registry.redhat.io/rhoso/openstack-ovn-base-rhel9@sha256:a3b8ec7ad1073facb8876acc787549a1a68b9bf81bebdc127f13eb6385e2d1da
        - registry.redhat.io/rhoso/openstack-ovn-bgp-agent-rhel9@sha256:a404c3416ab65e7134f438362e887529be474639e725e18f3542e86973edac87
        - registry.redhat.io/rhoso/openstack-nova-novncproxy-rhel9@sha256:04881867f9b24403bf6ef051eee3902a17e8476eee894ef579c0d5d130d2510d
        - registry.redhat.io/rhoso/openstack-ironic-inspector-rhel9@sha256:88357e83e8a4bbc3c5a37f0420ad75556821834a975cd8a1cf5fee56de2a4f99
        - registry.redhat.io/rhoso/openstack-ovn-controller-rhel9@sha256:3c42b7e0305c508d74602b9378143deb07a569443bba35850cdb7b6e3f5202e2
        - registry.redhat.io/rhoso/openstack-nova-conductor-rhel9@sha256:c67ca0e6c8126f56c6a83632b6a96bd4a35006f5f414a8779467b81864b31d12
        - registry.redhat.io/rhoso-operators/ee-openstack-ansible-ee-rhel9@sha256:61e729fb89c4872837cbe9cb1e3cea402b61f0da240751302605df487721f0aa
        - registry.redhat.io/rhoso-operators/ovn-rhel9-operator@sha256:66cd8add6078b63a36879c974db9e9f358593bb0651aa39fe85153d25404ae27
        - registry.redhat.io/rhoso/openstack-designate-backend-bind9-rhel9@sha256:82cca999357b0ab25bcc081bf802fab831587cf661db806af4ca453e0a626743
        - registry.redhat.io/rhoso/openstack-rabbitmq-rhel9@sha256:56bc2d2b201225fb98db43ea9ca68cceb4842f8f0f1a96ab3ab62ab785651d12
        - registry.redhat.io/rhoso-operators/prometheus-podman-exporter-rhel9@sha256:f33e7506db6dfdf530d3e2442baaf360d212bb1a47a7db887dba967ed0655cc2
        - registry.redhat.io/rhoso/openstack-nova-compute-rhel9@sha256:6264075310231af358803b53d5a09250dab32d63fc5966aa8fb957e5f56096a8
        - registry.redhat.io/rhoso-operators/neutron-rhel9-operator@sha256:30b6a812c47db9e6691211185c304c3e8fc6a8b9ba7ca79acf59c5ec07fb9fb0
        - registry.redhat.io/rhoso-operators/openstack-must-gather-rhel9@sha256:60b4355f4dbef3ad64103357086b10a5c0af126dd23798d6a9d725646f1e13b4
        - registry.redhat.io/rhoso/openstack-ironic-api-rhel9@sha256:50ccfe7d9ba22a50e0c1c8159cade6f4b1ac13ad8f5770f20c7be18fba58c8c3
        - registry.redhat.io/rhoso/openstack-ironic-pxe-rhel9@sha256:42e35ded74033f1f07287441917829cc934930a5df696c33cd8957be3c10fa41
        - registry.redhat.io/rhoso/openstack-openstackclient-rhel9@sha256:2952521d4b1799fd32ff773ef2d457a3abc354b639cf7f1e0f7f6615ca35ac8f
        - registry.redhat.io/rhoso-operators/manila-rhel9-operator@sha256:d9e69eb140b8ff76d5bf49deab8d700639ccdfd1ab5c8c7667bc2bbb50f3d60b
        - registry.redhat.io/rhoso-operators/horizon-rhel9-operator@sha256:9a2a13635588c6cc9949b272e1a1708c5fb53f19b05b8310ace703df5958ac63
        - registry.redhat.io/rhoso-operators/nova-rhel9-operator@sha256:3156dde249e49b40bed71092fb5213b112d43b663ba8c556e355aed524bec822
        - registry.redhat.io/rhoso-operators/ironic-rhel9-operator@sha256:dd67ffe3578ce619748735c1a5cd2a57d704d0c36c1b841d513ea36479aa4682
        - registry.redhat.io/rhoso/openstack-aodh-api-rhel9@sha256:4adc09cc1d922b387320da467de89d3f41a9c49b46d9b728458fec36505ec940
        - registry.redhat.io/rhoso/openstack-cron-rhel9@sha256:f09e724c69cec2c3b685783f9263b26c280a57b5909a723341416b342d18f9a4
        - registry.redhat.io/rhoso/openstack-ironic-conductor-rhel9@sha256:538da9fdeec546d794d192361d0a76f8b4e7682b6eb6d9cea371d8a52d603800
        - registry.redhat.io/rhoso/openstack-swift-account-rhel9@sha256:31e727f37db92fcc21319c44396576a03a118ba25d403715650162f7c0ead31e
        - registry.redhat.io/rhoso/openstack-redis-rhel9@sha256:c6dbeacbbd5aed9d13e540f1c987a574ea45747983c280446e5afadcf7c33b37
        - registry.redhat.io/rhoso/openstack-ovn-sb-db-server-rhel9@sha256:8ba4f8ed49e214f421e94133da0f17b54ce99a52bbbd367d0aef357de844ea01
        - registry.redhat.io/openshift4/ose-prometheus-node-exporter-rhel9@sha256:479c51de2fc8a3b3858d71bbe4eeafbd5b43263472346f9dde348115082894cc
        - registry.redhat.io/rhoso-operators/octavia-rhel9-operator@sha256:aa0f00f8e03dfdb547e53cf611581e9cc091eaf049558121c2bc5d702e556bcb
        - registry.redhat.io/rhoso/openstack-mariadb-rhel9@sha256:6115a7458cfa412c1659242fcec64b63f0a5ca414ecf3c5f890f0afec0bca069
        - registry.redhat.io/rhoso/openstack-multipathd-rhel9@sha256:6ff8ae176f8a633bcd78b1f8407924c74bb2f317eba16b90e7f0f9e53af74e8e
        - registry.redhat.io/rhoso/openstack-neutron-sriov-agent-rhel9@sha256:118d5f9430192ab48cf159ee2f7912f08e935ad56ef3364f18b3680210bb3e52
        - registry.redhat.io/rhoso-operators/barbican-rhel9-operator@sha256:b3b769e0dc32c12c5cc9578c2cac2622b973177ca14697bc1042d21d473e4e84
        - registry.redhat.io/rhoso/openstack-heat-api-rhel9@sha256:ccbc41b552bb34d223c9615a4e3b56bb516bc5bd8a694f048c15063ba45fca9f
        - registry.redhat.io/rhoso-operators/telemetry-rhel9-operator@sha256:0635515a5fed12f17de4cc5ac5f9567e3b3546fc2f69730c7beab25cf722f538
        - registry.redhat.io/rhoso/openstack-aodh-listener-rhel9@sha256:037845be6787f610e479345206b0fcfb2f1dc430540816bc5a54a1d26da2acbd
        - registry.redhat.io/rhoso/openstack-ceilometer-notification-rhel9@sha256:bd2b5f50a2bef4e77e3daa89aca51df6e0e55c3784886296d92c97616cd566d1
        - registry.redhat.io/rhoso-operators/sg-core-rhel9@sha256:b56923f869ff57396e6822133dfc3b21007c3df652cfe962c7ad3028214d988c
        - registry.redhat.io/rhoso/openstack-barbican-worker-rhel9@sha256:485525116fb3c22d8c60330dd87ebd3854cd71ded36bf88bcf18cee5957e22a3
        - registry.redhat.io/rhoso-operators/openstack-network-exporter-rhel9@sha256:8e8e775e37c28cb11b7aae9199405b85d5862dce195d165ca598fa1dc89bf914
        - registry.redhat.io/rhoso/openstack-placement-api-rhel9@sha256:0aa5e138947bf236f31c81bfb6f641d95af63812c06dca43141721b383c7e08e
        - registry.redhat.io/openshift-power-monitoring/kepler-rhel9@sha256:8e9e5bfeed61aaeee85bf5eefc3390556b8d4cbfc797a82cabe95bc0367126eb
        - registry.redhat.io/rhoso/openstack-ceilometer-ipmi-rhel9@sha256:2fb5723c68d0f2c3ae28b86e606ccae4b80f3bd9b16c01dd6c9adcae629db7f1
        - registry.redhat.io/rhoso-operators/openstack-rhel9-operator@sha256:64e2ad312f757101f083eedcb17e0b6c05cf6ebf6ed9e7db2e33e84c6aca9898
        - registry.redhat.io/rhoso/openstack-unbound-rhel9@sha256:7f17a719e8e49d457ff16437709c2b4eca5b9c2f58972d1e351c5c1c1ac38962
        - registry.redhat.io/rhoso-operators/infra-rhel9-operator@sha256:a2dc031a6ad2726d155e77cafb676f8e35a86e963b97ccdceed0da137e67ba0d
        - registry.redhat.io/rhoso/openstack-designate-mdns-rhel9@sha256:ce43dd1a30afd4753aa11d32df2b24ed4bf3d67847044aaae8f3e5c8901caf05
        - registry.redhat.io/rhoso/openstack-octavia-worker-rhel9@sha256:303e35926e43fe403673e18a25c2e815547c28a5f8c0f09d470cd4148a21c528
        - registry.redhat.io/rhoso-operators/placement-rhel9-operator@sha256:a16bd4fb929eb5bcf96f2aefcb5a7abf4143f4f9ed24e53de93f4a3caa89a507
        - registry.redhat.io/rhoso-operators/rabbitmq-cluster-rhel9-operator@sha256:e095e6cbda841b9c726c1ec0839a98964b894cf25ae9228052d314b1ee2f161d
        - registry.redhat.io/rhoso/openstack-memcached-rhel9@sha256:3e704b9664c65be39e83c2ed802722ffeaf969d8e571147a2be128d5c56331c1
        version: 1.0.14
      entries:
      - name: openstack-operator.v1.0.14
        version: 1.0.14
      - name: openstack-operator.v1.0.13
        version: 1.0.13
      - name: openstack-operator.v1.0.12
        version: 1.0.12
      - name: openstack-operator.v1.0.11
        version: 1.0.11
      - name: openstack-operator.v1.0.10
        version: 1.0.10
      - name: openstack-operator.v1.0.9
        version: 1.0.9
      - name: openstack-operator.v1.0.8
        version: 1.0.8
      - name: openstack-operator.v1.0.7
        version: 1.0.7
      - name: openstack-operator.v1.0.6
        version: 1.0.6
      - name: openstack-operator.v1.0.5
        version: 1.0.5
      - name: openstack-operator.v1.0.4
        version: 1.0.4
      - name: openstack-operator.v1.0.3
        version: 1.0.3
      - name: openstack-operator.v1.0.2
        version: 1.0.2
      - name: openstack-operator.v1.0.1
        version: 1.0.1
      - name: openstack-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: openstack-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: IBM
      provider-url: ""
    name: ibm-block-csi-operator-community
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12
    - currentCSV: ibm-block-csi-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
        version: 1.12.0
      entries:
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.0
    - currentCSV: ibm-block-csi-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
        version: 1.12.1
      entries:
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.1
    - currentCSV: ibm-block-csi-operator.v1.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:3fc8df547138a2009225d63c2a3111a672340bdf9f3cb24a889cbb69f2c33c2d
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:3fc8df547138a2009225d63c2a3111a672340bdf9f3cb24a889cbb69f2c33c2d
        version: 1.12.2
      entries:
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.2
    - currentCSV: ibm-block-csi-operator.v1.12.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:2bd461a1e40a1366c871585aa7a4e8039e686a66bf7192efdeb1c215209a071d
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:2bd461a1e40a1366c871585aa7a4e8039e686a66bf7192efdeb1c215209a071d
        version: 1.12.3
      entries:
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.3
    - currentCSV: ibm-block-csi-operator.v1.12.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e29db476de3778222bf049926223dee20f6bc34e55a04df5f0f3df38741ff23e
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e29db476de3778222bf049926223dee20f6bc34e55a04df5f0f3df38741ff23e
        version: 1.12.4
      entries:
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.4
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      - name: ibm-block-csi-operator.v1.12.0
        version: 1.12.0
      - name: ibm-block-csi-operator.v1.11.4
        version: 1.11.4
      name: stable-v1.12.5
    defaultChannel: stable
    packageName: ibm-block-csi-operator-community
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: ""
    name: pmem-csi-operator-os
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pmem-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pmem-csi.intel.com/v1beta1",
                "kind": "PmemCSIDeployment",
                "metadata": {
                  "name": "pmem-csi.intel.com"
                },
                "spec": {
                  "deviceMode": "lvm",
                  "nodeSelector": {
                    "feature.node.kubernetes.io/memory-nv.dax": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
          createdAt: "2021-08-26T10:45:23Z"
          description: An operator for deploying and managing the PMEM-CSI driver
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: ""
          repository: https://github.com/intel/pmem-csi/tree/release-1.0
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a PMEM-CSI driver deployment in the cluster.
            displayName: PmemCSIDeployment
            kind: PmemCSIDeployment
            name: pmemcsideployments.pmem-csi.intel.com
            version: v1beta1
        description: |2

          Intel PMEM-CSI is a [CSI](https://github.com/container-storage-interface/spec)
          storage driver for container orchestrators like Kubernetes. It makes local
          persistent memory ([PMEM](https://pmem.io/)) available as a filesystem volume to
          container applications. This is the operator to deploy and manage the
          [PMEM-CSI](https://intel.github.io/pmem-csi/1.0/README.html)
          driver on a Kubernetes cluster.

          Refer to the [PMEM-CSI documentation](https://intel.github.io/pmem-csi/1.0/README.html)
          for more details on deploying and using the PMEM-CSI driver.
        displayName: Operator for PMEM-CSI driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - PMEM-CSI
        - PMEM-CSI operator
        - Persistent Memory
        - PMEM
        links:
        - name: PMEM-CSI
          url: https://intel.github.io/pmem-csi/1.0
        maintainers:
        - email: olev.kartau@intel.com
          name: Olev Kartau
        - email: patrick.ohly@intel.com
          name: Patrick Ohly
        - email: amarnath.valluri@intel.com
          name: Amarnath Valluri
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Intel® Corporation
        relatedImages:
        - registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
        version: 1.0.0
      entries:
      - name: pmem-csi-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: pmem-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pmem-csi.intel.com/v1beta1",
                "kind": "PmemCSIDeployment",
                "metadata": {
                  "name": "pmem-csi.intel.com"
                },
                "spec": {
                  "deviceMode": "lvm",
                  "nodeSelector": {
                    "feature.node.kubernetes.io/memory-nv.dax": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
          createdAt: "2021-08-26T10:45:23Z"
          description: An operator for deploying and managing the PMEM-CSI driver
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: ""
          repository: https://github.com/intel/pmem-csi/tree/release-1.0
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a PMEM-CSI driver deployment in the cluster.
            displayName: PmemCSIDeployment
            kind: PmemCSIDeployment
            name: pmemcsideployments.pmem-csi.intel.com
            version: v1beta1
        description: |2

          Intel PMEM-CSI is a [CSI](https://github.com/container-storage-interface/spec)
          storage driver for container orchestrators like Kubernetes. It makes local
          persistent memory ([PMEM](https://pmem.io/)) available as a filesystem volume to
          container applications. This is the operator to deploy and manage the
          [PMEM-CSI](https://intel.github.io/pmem-csi/1.0/README.html)
          driver on a Kubernetes cluster.

          Refer to the [PMEM-CSI documentation](https://intel.github.io/pmem-csi/1.0/README.html)
          for more details on deploying and using the PMEM-CSI driver.
        displayName: Operator for PMEM-CSI driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - PMEM-CSI
        - PMEM-CSI operator
        - Persistent Memory
        - PMEM
        links:
        - name: PMEM-CSI
          url: https://intel.github.io/pmem-csi/1.0
        maintainers:
        - email: olev.kartau@intel.com
          name: Olev Kartau
        - email: patrick.ohly@intel.com
          name: Patrick Ohly
        - email: amarnath.valluri@intel.com
          name: Amarnath Valluri
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Intel® Corporation
        relatedImages:
        - registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
        version: 1.0.0
      entries:
      - name: pmem-csi-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pmem-csi-operator-os
    provider:
      name: Intel® Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OceanBase
      provider-url: https://oceanbase.com
    name: oceanbase-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oceanbase-operator.v2.2.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Full Lifecycle
          createdAt: "2024-08-14T16:21:06Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: OBClusterOperation
            name: obclusteroperations.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBCluster
            name: obclusters.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBParameter
            name: obparameters.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBResourceRescue
            name: obresourcerescues.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBServer
            name: observers.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantBackupPolicy
            name: obtenantbackuppolicies.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantBackup
            name: obtenantbackups.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantOperation
            name: obtenantoperations.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantRestore
            name: obtenantrestores.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenant
            name: obtenants.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBZone
            name: obzones.oceanbase.oceanbase.com
            version: v1alpha1
        description: Ob Operator description. TODO.
        displayName: Ob Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - oceanbase-operator
        - ob-operator
        links:
        - name: OceanBase Operator
          url: https://oceanbase.github.io/ob-operator
        maintainers:
        - email: hongxin.sun.vip@gmail.com
          name: Hongxin Sun
        maturity: alpha
        provider:
          name: OceanBase
          url: https://oceanbase.com
        relatedImages:
        - quay.io/oceanbasedev/oceanbase-operator@sha256:b11fe60bb8de545d3ed105acc50f2d6c5d6aa8093f3ae4ca7ebce58b1e5e7b48
        - quay.io/brancz/kube-rbac-proxy@sha256:18f12fc3a0085bbc303cfbd2ad1151127c1c959c4641e36027547b6b66eb4c92
        version: 2.2.2
      entries:
      - name: oceanbase-operator.v2.2.2
        version: 2.2.2
      name: alpha
    - currentCSV: oceanbase-operator.v2.2.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Full Lifecycle
          createdAt: "2024-08-14T16:21:06Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: OBClusterOperation
            name: obclusteroperations.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBCluster
            name: obclusters.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBParameter
            name: obparameters.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBResourceRescue
            name: obresourcerescues.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBServer
            name: observers.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantBackupPolicy
            name: obtenantbackuppolicies.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantBackup
            name: obtenantbackups.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantOperation
            name: obtenantoperations.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenantRestore
            name: obtenantrestores.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBTenant
            name: obtenants.oceanbase.oceanbase.com
            version: v1alpha1
          - kind: OBZone
            name: obzones.oceanbase.oceanbase.com
            version: v1alpha1
        description: Ob Operator description. TODO.
        displayName: Ob Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - oceanbase-operator
        - ob-operator
        links:
        - name: OceanBase Operator
          url: https://oceanbase.github.io/ob-operator
        maintainers:
        - email: hongxin.sun.vip@gmail.com
          name: Hongxin Sun
        maturity: alpha
        provider:
          name: OceanBase
          url: https://oceanbase.com
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy@sha256:18f12fc3a0085bbc303cfbd2ad1151127c1c959c4641e36027547b6b66eb4c92
        - quay.io/oceanbasedev/oceanbase-operator@sha256:b11fe60bb8de545d3ed105acc50f2d6c5d6aa8093f3ae4ca7ebce58b1e5e7b48
        version: 2.2.2
      entries:
      - name: oceanbase-operator.v2.2.2
        version: 2.2.2
      name: stable
    defaultChannel: stable
    packageName: oceanbase-operator
    provider:
      name: OceanBase
      url: https://oceanbase.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.2.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"customImageNames":
            false,"externalArangoDBName": "arangodb","repository": "registry.connect.redhat.com/turbonomic","tag":"8.3.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/turbonomic/t8c-operator@sha256:d5164e6dd24a0feff10cab62530cafff1bc92d0e8ec2d16c6f4d880c6e2f50bd
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/t8c-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/t8c-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/turbonomic/t8c-operator@sha256:d5164e6dd24a0feff10cab62530cafff1bc92d0e8ec2d16c6f4d880c6e2f50bd
        version: 42.2.0
      entries:
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: stable
    defaultChannel: stable
    packageName: t8c-certified-rhmp
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ptp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ptp-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "NodePtpDevice",
                "metadata": {
                  "name": "node.example.com"
                },
                "spec": {}
              },
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "PtpConfig",
                "metadata": {
                  "name": "example-ptpconfig"
                },
                "spec": {
                  "profile": [
                    {
                      "interface": "ens787f1",
                      "name": "profile1",
                      "phc2sysOpts": "-a -r",
                      "ptp4lOpts": "-s -2",
                      "ptpClockThreshold": {}
                    }
                  ],
                  "recommend": [
                    {
                      "match": [
                        {
                          "nodeLabel": "node-role.kubernetes.io/worker=",
                          "nodeName": "node.example.com"
                        }
                      ],
                      "priority": 4,
                      "profile": "profile1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "PtpOperatorConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "daemonNodeSelector": {
                    "feature.node.kubernetes.io/ptp-capable": "yes",
                    "node-role.kubernetes.io/worker": ""
                  },
                  "ptpEventConfig": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-ptp-rhel9-operator@sha256:012334e6624d5440439824454bda1d1dc79804938174232a3b470176bf462b5e
          createdAt: "2024-03-13"
          description: This software enables configuration of Precision Time Protocol(PTP)
            on Kubernetes. It detects hardware capable PTP devices on each node, and
            configures linuxptp processes such as ptp4l, phc2sys and timemaster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.16.0-202509240123'
          operatorframework.io/suggested-namespace: openshift-ptp
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/ptp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodePtpDevice is the Schema for the nodeptpdevices API
            displayName: Node Ptp Device
            kind: NodePtpDevice
            name: nodeptpdevices.ptp.openshift.io
            version: v1
          - description: PtpConfig is the Schema for the ptpconfigs API
            displayName: Ptp Config
            kind: PtpConfig
            name: ptpconfigs.ptp.openshift.io
            version: v1
          - description: PtpOperatorConfig is the Schema for the ptpoperatorconfigs
              API
            displayName: Ptp Operator Config
            kind: PtpOperatorConfig
            name: ptpoperatorconfigs.ptp.openshift.io
            version: v1
        description: |
          # Precision Time Protocol (PTP) Operator for Openshift

          ## Introdution
          PTP Operator manages cluster wide PTP configuration.
          This operator has to run in namespace 'openshift-ptp'. An Operator Group is also required to install this operator:
          ```
          $ oc create -f - <<EOF
          apiVersion: v1
          kind: Namespace
          metadata:
            name: openshift-ptp
            labels:
              name: openshift-ptp
              openshift.io/cluster-monitoring: "true"
          EOF

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: ptp-operators
            namespace: openshift-ptp
          spec:
            targetNamespaces:
            - openshift-ptp
          EOF
          ```
        displayName: PTP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ptp
        links:
        - name: Documentation operator
          url: https://github.com/openshift/ptp-operator
        - name: Documentation operand
          url: https://github.com/openshift/linuxptp-daemon
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-cloud-event-proxy-rhel9@sha256:3fa2aad8cf026eba586645c851f07a1d752c85b30795fdcf037b138ef3bbb6aa
        - registry.redhat.io/openshift4/ose-ptp-rhel9-operator@sha256:012334e6624d5440439824454bda1d1dc79804938174232a3b470176bf462b5e
        - registry.redhat.io/openshift4/ose-ptp-rhel9@sha256:93b4717337b551603cbb6147d66d7b70a5580332953ddf42304f7092d8739c83
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        version: 4.16.0-202509240123
      entries:
      - name: ptp-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: ptp-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: ptp-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: ptp-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: ptp-operator.v4.16.0-202506201205
        version: 4.16.0-202506201205
      - name: ptp-operator.v4.16.0-202505291140
        version: 4.16.0-202505291140
      - name: ptp-operator.v4.16.0-202505161704
        version: 4.16.0-202505161704
      - name: ptp-operator.v4.16.0-202505020146
        version: 4.16.0-202505020146
      - name: ptp-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: ptp-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: ptp-operator.v4.16.0-202502260004
        version: 4.16.0-202502260004
      - name: ptp-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: ptp-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: ptp-operator.v4.16.0-202501201505
        version: 4.16.0-202501201505
      - name: ptp-operator.v4.16.0-202501072135
        version: 4.16.0-202501072135
      - name: ptp-operator.v4.16.0-202412311307
        version: 4.16.0-202412311307
      - name: ptp-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: ptp-operator.v4.16.0-202412102004
        version: 4.16.0-202412102004
      - name: ptp-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: ptp-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: ptp-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: ptp-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: ptp-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: ptp-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: ptp-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: ptp-operator.v4.16.0-202409131735
        version: 4.16.0-202409131735
      - name: ptp-operator.v4.16.0-202409062036
        version: 4.16.0-202409062036
      - name: ptp-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: ptp-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: ptp-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: ptp-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: ptp-operator.v4.16.0-202407311708
        version: 4.16.0-202407311708
      - name: ptp-operator.v4.16.0-202407301837
        version: 4.16.0-202407301837
      - name: ptp-operator.v4.16.0-202407160236
        version: 4.16.0-202407160236
      - name: ptp-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: ptp-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: ptp-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: ptp-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: ptp-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: ptp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fossul Framework
      provider-url: https://github.com/fossul/fossul
    name: fossul-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fossul-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n    \"kind\":
            \"Backup\",\n    \"metadata\": {\n      \"name\": \"mariadb-sample\"\n
            \   },\n    \"spec\": {\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\"\n    }\n  },\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n    \"kind\":
            \"Restore\",\n    \"metadata\": {\n      \"name\": \"mariadb-sample\"\n
            \   },\n    \"spec\": {\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\",\n      \"workflow_id\": \"xxxx\"\n    }\n  },        \n  {\n
            \   \"apiVersion\": \"fossul.io/v1\",\n    \"kind\": \"BackupConfig\",\n
            \   \"metadata\": {\n      \"name\": \"mariadb-sample\"\n    },\n    \"spec\":
            {\n      \"container_platform\": \"openshift\",\n      \"operator_controlled\":
            true,\n      \"auto_discovery\": false,\n      \"deployment_name\": \"mariadb\",\n
            \     \"deployment_type\": \"DeploymentConfig\",\n      \"job_retention\":
            50,\n      \"overwrite_pcv_on_restore\": true,\n      \"policies\": [\n
            \       {\n          \"policy\": \"hourly\",\n          \"retentionNumber\":
            3\n        },\n        {\n          \"policy\": \"daily\",\n          \"retentionNumber\":
            10\n        }\n      ],\n      \"pvc_deletion_timeout\": 300,\n      \"restore_to_new_pvc\":
            false,\n      \"snapshot_timeout\": 180,\n      \"storage_plugin\": \"csi.so\",\n
            \     \"app_plugin\": \"mariadb.so\"\n    }\n  },\n  {\n    \"apiVersion\":
            \"fossul.io/v1\",\n    \"kind\": \"BackupSchedule\",\n    \"metadata\":
            {\n      \"name\": \"mariadb-sample\"\n    },\n    \"spec\": {\n      \"cron_schedule\":
            \"59 23 * * *\",\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\"\n\n    }\n  },\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n
            \   \"kind\": \"Fossul\",\n    \"metadata\": {\n      \"name\": \"fossul-sample\"\n
            \   },\n    \"spec\": {\n      \"enable_routes\": false,\n      \"enable_virtualmachines\":
            false,\n      \"container_platform\": \"openshift\"\n    }\n  }\n]"
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/fossul/fossul-operator:v0.6.0
          createdAt: "2022-01-25T20:09:00Z"
          description: Fossul is a application centric container native backup and
            restore framework for databases or applications
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/fossul/fossul
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages application backup configurations
            kind: BackupConfig
            name: backupconfigs.fossul.io
            version: v1
          - description: Manages application backups
            kind: Backup
            name: backups.fossul.io
            version: v1
          - description: Manages application restores
            kind: Restore
            name: restores.fossul.io
            version: v1
          - description: Manages application backup scheduling
            kind: BackupSchedule
            name: backupschedules.fossul.io
            version: v1
          - description: Manages deployment of the Fossul Framework
            kind: Fossul
            name: fossuls.fossul.io
            version: v1
        description: "## Fossul Framework\nApplication backup and restore framework
          for container native applications on Kubernetes and OpenShift. Fossul provides
          application centric backup and recovery enabling integration with the application
          creating an application consistent backup. Since there are many applications
          and providers for storage or archive Fossul enables plugins that allow proper
          integration. Fossul provides plugins for postgresql, mariadb, mysql and
          mongo as well as plugin to integrate with CSI snapshots. Plugins can be
          written in any language. The Fossul operator deploys Fossul via the Fossul
          custom resource and additionally provides BackupConfig, Backup, BackupSchedule
          and Restore custom resources. Each user can configure and manage their own
          backups with simple custom resources that can be part of CI/CD or source
          control. Users can also opt to manage backups through Fossul CLI/API instead
          of the custom resources provided by the operator.\n\n## Installation\nIt
          is recommended to install the operator into the openshift-operators namespace
          which is the default location for operators. Once the operator is installed
          create a new project called ```fossul``` and create the Fossul custom resource
          under the fossul namespace. This will deploy the Fossul Framework. You can
          choose another namespace for the Fossul Framework but then also need to
          provide the optional 'fossul_namespace' parameter with the spec for all
          the additional custom resources. This is required so the Fossul operator
          can communicate with Fossul.\n\nOnce Fossul is deployed you can manage backups
          either through custom resources provided by the operator or the Fossul CLI/API.
          If using the customer resources, first create a BackupConfig custom resource
          for every application you want to backup. Once that is done, you can optionally
          create a backup by simply creating a Backup custom resource. You can also
          create a BackupSchedule custom resource using cron syntax which will schedule
          your backup to occur on a regular interval or do so via a Kubernetes job.
          Finally if a backup has been created you can perform a restore through a
          custom resource as well, providing the workflow id from the completed backup.
          If you are creating backups through the custom resource the workflow id
          will be appended to the spec once the backup completes. If you are using
          the Fossul CLI/API to create backups you will need to get the workflow id
          through the Fossul CLI/API directly using the backupList action.\n\nFor
          users of the cli, a cli pod is deployed with credentials and if desired
          you can create or manage backups from within that pod. To use the API you
          need the credentials stored in the fossul secret. The cli has already been
          pre-configured with these credentials. The API can also be optionally exposed
          via routes and API documentation is under ```/api/v1/index.html```.\n\n##
          Quick Setup\nWhen creating Fossul custom resources generally you will only
          need to update the ```deployment_name``` and ```deployment_type``` (DeploymentConfig,
          Deployment or VirtualMachine). All custom resources must be created in the
          namespace where the database or application exists. \n\n### Deploy Fossul
          in fossul namespace\n```$ oc new-project fossul```\n<pre>\n$ vi fossul.yaml\nkind:
          Fossul\napiVersion: fossul.io/v1\nmetadata:\n  name: fossul-sample\n  namespace:
          fossul\nspec: {}\n</pre>\n```$ oc create -f fossul.yaml```\n\n### Deploy
          Mariadb Database\nMake sure you create all Fossul custom resources in the
          same namespace as the database, application or virtual machine\n\n### Create
          MariaDB Fossul BackupConfig\nCreates a backup configuration for MariaDB
          or MySQL databases\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true\n  auto_discovery:
          false\n  deployment_name: mariadb\n  deployment_type: DeploymentConfig\n
          \ job_retention: 50\n  overwrite_pcv_on_restore: true\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc: false\n  snapshot_timeout:
          180\n  storage_plugin: csi.so\n  app_plugin: mariadb.so \n</pre>\n```$ oc
          create -f backupconfig.yaml```\n\n### Create PostgreSQL Fossul BackupConfig\nCreates
          a backup config for PostgreSQL databases, you need to ensure the user defined
          in secret has backup permissions\n\n```$ vi backupconfig.yaml```\n<pre>\nkind:
          BackupConfig\napiVersion: fossul.io/v1\nmetadata:\n  name: postgresql-sample\n
          \ namespace: databases\nspec:\n  container_platform: openshift\n  operator_controlled:
          true    \n  auto_discovery: false\n  deployment_name: postgresql \n  deployment_type:
          DeploymentConfig\n  job_retention: 50\n  overwrite_pcv_on_restore: true\n
          \ policies:\n    - policy: hourly\n      retentionNumber: 3\n    - policy:
          daily\n      retentionNumber: 10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc:
          false\n  snapshot_timeout: 180\n  storage_plugin: csi.so\n  app_plugin:
          postgres.so \n</pre>\n```$ oc create -f backupconfig.yaml```\n\n### Create
          Kubevirt Fossul BackupConfig\nCreates a backup configuration for virtual
          machines\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: rhel8-vm-sample\n  namespace: virtualmachines\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true    \n  auto_discovery:
          false\n  deployment_name: rhel8-vm\n  deployment_type: VirtualMachine\n
          \ job_retention: 50\n  overwrite_pcv_on_restore: true\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc: false\n  snapshot_timeout:
          180\n  storage_plugin: csi.so\n  app_plugin: kubevirt.so \n</pre>\n```$
          oc create -f backupconfig.yaml```\n\n### Create Tar Fossul BackupConfig\nCreates
          a backup config using tar to backup folders and files where csi snapshots
          are not desired\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: tar-sample\n  namespace: databases\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true    \n  auto_discovery:
          false\n  deployment_name: anyname\n  job_retention: 50\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  backup_source_paths: /opt/app-root/src/test\n  storage_plugin: tar.so\n</pre>\n```$
          oc create -f backupconfig.yaml```\n\n### Create BackupSchedule\nA backup
          can be scheduled per policy, defined by backup configuration using cron
          syntax\n\n```$ vi backupschedule.yaml```\n<pre>\nkind: BackupSchedule\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ cron_schedule: '59 23 * * *'\n  deployment_name: mariadb\n  policy: daily\n</pre>\n```$
          oc create -f backupschedule.yaml```\n\n### Create Backup\nA backup will
          be created according to policy and deleted either manually or as defined
          in retention policy\n\n```$ vi backup.yaml```\n<pre>\nkind: Backup\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ deployment_name: mariadb\n  policy: daily\n</pre>\n```$ oc create -f backup.yaml```\n\n###
          Perform Restore\nA restore requires a workflow_id located in the backup
          spec, upon a successful restore the custom resource will deleted automatically\n\n```$
          vi restore.yaml```\n<pre>\nkind: Restore\napiVersion: fossul.io/v1\nmetadata:\n
          \ name: mariadb-sample\n  namespace: databases\nspec:\n  deployment_name:
          mariadb\n  policy: daily\n  workflow_id: xxxx\n</pre>\n```$ oc create -f
          restore.yaml```\n"
        displayName: Fossul Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - application
        - consistent
        links:
        - name: Fossul Framework Community
          url: https://github.com/fossul/fossul
        maintainers:
        - email: ktenzer@redhat.com
          name: Keith Tenzer
        maturity: alpha
        provider:
          name: Fossul Framework
          url: https://github.com/fossul/fossul
        relatedImages:
        - quay.io/fossul/fossul-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: fossul-operator.v0.6.0
        version: 0.6.0
      - name: fossul-operator.v0.5.0
        version: 0.5.0
      - name: fossul-operator.v0.4.0
        version: 0.4.0
      - name: fossul-operator.v0.3.0
        version: 0.3.0
      - name: fossul-operator.v0.2.0
        version: 0.2.0
      - name: fossul-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: fossul-operator
    provider:
      name: Fossul Framework
      url: https://github.com/fossul/fossul
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Citrus Framework
      provider-url: ""
    name: yaks
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: yaks-operator.v0.17.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "yaks.citrusframework.org/v1alpha1",
                "kind": "Instance",
                "metadata": {
                  "labels": {
                    "app": "yaks"
                  },
                  "name": "example-instance"
                },
                "spec": {
                  "operator": {
                    "global": true
                  }
                }
              },
              {
                "apiVersion": "yaks.citrusframework.org/v1alpha1",
                "kind": "Test",
                "metadata": {
                  "labels": {
                    "app": "yaks"
                  },
                  "name": "example-test"
                },
                "spec": {
                  "source": {
                    "content": "Feature: hello world\n\n  Scenario: print slogan\n    Given YAKS does Cloud-Native BDD testing\n    Then YAKS rocks!",
                    "language": "feature",
                    "name": "example.feature"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/citrusframework/yaks:0.17.1
          createdAt: "2023-04-26T20:37:31Z"
          description: YAKS is a platform to enable Cloud Native BDD testing on Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["instances.yaks.citrusframework.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/citrusframework/yaks
          support: Citrus Framework
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance is the Schema for the yaks instance
            displayName: Instance
            kind: Instance
            name: instances.yaks.citrusframework.org
            version: v1alpha1
          - description: Test is the Schema for the tests API
            displayName: Test
            kind: Test
            name: tests.yaks.citrusframework.org
            version: v1alpha1
        description: |
          YAKS
          ====

          YAKS is a platform to enable Cloud Native BDD testing on Kubernetes.

          ## Running a YAKS test

          With the YAKS operator installed, you can run tests by creating a `Test` custom resource on the cluster.

          Tests are defined using [Gherkin](https://cucumber.io/docs/gherkin/) syntax. YAKS provides a set of predefined steps which
          help to connect with different messaging transports (Http REST, JMS, Kafka, Knative eventing) and verify responses with
          assertions on message header and body content.

          The example below defines a simple test resource.

          ```
          apiVersion: yaks.citrusframework.org/v1alpha1
          kind: Test
          metadata:
            name: example
          spec:
            source:
              name: example.feature
              content: |-
                Feature: hello world

                  Scenario: print slogan
                    Given YAKS does Cloud-Native BDD testing
                    Then YAKS rocks!
          ```

          Refer to the [YAKS repository](https://github.com/citrusframework/yaks) for more complex examples.
        displayName: YAKS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - yaks
        - testing
        - microservices
        - knative
        - kafka
        - serverless
        - camel
        - camel-k
        links:
        - name: Project page
          url: https://citrusframework.org/
        - name: YAKS source code repository
          url: https://github.com/citrusframework/yaks
        maintainers:
        - email: cdeppisch@redhat.com
          name: Christoph Deppisch
        - email: nferraro@redhat.com
          name: Nicola Ferraro
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Citrus Framework
        relatedImages:
        - docker.io/citrusframework/yaks:0.17.1
        version: 0.17.1
      entries:
      - name: yaks-operator.v0.17.1
        version: 0.17.1
      - name: yaks-operator.v0.17.0
        version: 0.17.0
      - name: yaks-operator.v0.16.0
        version: 0.16.0
      - name: yaks-operator.v0.15.1
        version: 0.15.1
      - name: yaks-operator.v0.15.0
        version: 0.15.0
      - name: yaks-operator.v0.14.3
        version: 0.14.3
      - name: yaks-operator.v0.14.2
        version: 0.14.2
      - name: yaks-operator.v0.14.1
        version: 0.14.1
      - name: yaks-operator.v0.13.0
        version: 0.13.0
      - name: yaks-operator.v0.12.2
        version: 0.12.2
      - name: yaks-operator.v0.11.0
        version: 0.11.0
      - name: yaks-operator.v0.10.0
        version: 0.10.0
      - name: yaks-operator.v0.9.0
        version: 0.9.0
      - name: yaks-operator.v0.8.0
        version: 0.8.0
      - name: yaks-operator.v0.7.0
        version: 0.7.0
      - name: yaks-operator.v0.6.0
        version: 0.6.0
      - name: yaks-operator.v0.5.0
        version: 0.5.0
      - name: yaks-operator.v0.4.0
        version: 0.4.0
      name: alpha
    defaultChannel: alpha
    packageName: yaks
    provider:
      name: Citrus Framework
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator.v1.5.0
        version: 1.5.0
      - name: dynatrace-operator.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetScaler
      provider-url: https://netscaler.com
    name: netscaler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netscaler-operator.v3.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCloudController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalercloudcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "azure": {
                    "clientsecret": "",
                    "image": "quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb",
                    "location": "",
                    "pullPolicy": "IfNotPresent",
                    "resourcegroupname": "",
                    "resources": {},
                    "subnetname": "",
                    "subscriptionid": "",
                    "tenantid": "",
                    "vnetname": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCpxWithIngressController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalercpxwithingresscontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "admSettings": {
                    "admIP": "",
                    "bandWidth": 1000,
                    "bandWidthLicense": false,
                    "cpxCores": 1,
                    "licenseEdition": "PLATINUM",
                    "licenseServerIP": "",
                    "licenseServerPort": 27000,
                    "loginSecret": "",
                    "platform": false,
                    "vCPULicense": false
                  },
                  "affinity": {},
                  "analyticsConfig": {
                    "distributedTracing": {
                      "enable": false,
                      "samplingRate": 100
                    },
                    "endpoint": {
                      "metrics": {
                        "service": ""
                      },
                      "transactions": {
                        "service": ""
                      }
                    },
                    "required": false,
                    "timeseries": {
                      "auditLogs": {
                        "enable": false
                      },
                      "events": {
                        "enable": false
                      },
                      "metrics": {
                        "enable": false,
                        "enableNativeScrape": false,
                        "exportFrequency": 30,
                        "mode": "avro",
                        "schemaFile": "schema.json"
                      },
                      "port": 5563
                    },
                    "transactions": {
                      "enable": false,
                      "port": 5557
                    }
                  },
                  "exporter": {
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4",
                    "ports": {
                      "containerPort": 8888
                    },
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {}
                  },
                  "extraVolumes": [],
                  "imagePullSecrets": [],
                  "ingressController": {
                    "defaultSSLCertSecret": "",
                    "defaultSSLSNICertSecret": "",
                    "disableAPIServerCertVerify": false,
                    "disableOpenshiftRoutes": true,
                    "enableLivenessProbe": true,
                    "enableReadinessProbe": true,
                    "entityPrefix": "k8s",
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "ingressClass": [],
                    "ipam": false,
                    "jsonLog": false,
                    "kubernetesURL": "",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "logLevel": "INFO",
                    "logProxy": "",
                    "namespaceLabels": "",
                    "openshift": true,
                    "optimizeEndpointBinding": false,
                    "profileHttpFrontend": {},
                    "profileSslFrontend": {},
                    "profileTcpFrontend": {},
                    "pullPolicy": "IfNotPresent",
                    "rbacRole": false,
                    "readinessProbe": {
                      "failureThreshold": 3,
                      "initialDelaySeconds": 50,
                      "periodSeconds": 60,
                      "successThreshold": 1
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "128m",
                        "memory": "500Mi"
                      }
                    },
                    "routeLabels": "",
                    "setAsDefaultIngressClass": false,
                    "updateIngressStatus": true
                  },
                  "netscalerCpx": {
                    "bgpSettings": {
                      "bgpConfig": [
                        {
                          "bgpRouter": {
                            "localAS": 100,
                            "neighbor": [
                              {
                                "ASOriginationInterval": 10,
                                "address": "",
                                "advertisementInterval": 10,
                                "remoteAS": 100
                              }
                            ]
                          }
                        }
                      ],
                      "bgpPort": 179
                    },
                    "cpxBgpRouter": false,
                    "cpxCommands": [],
                    "cpxLicenseAggregator": "",
                    "cpxShellCommands": [],
                    "daemonSet": false,
                    "enableLivenessProbe": true,
                    "enableStartupProbe": true,
                    "extraVolumeMounts": [],
                    "hostname": "",
                    "image": "quay.io/netscaler/netscaler-cpx@sha256:232596eaacda68c0e2e210f26b64cedc8d70ae12c5e8ae2e1a66896a479ae4f8",
                    "ingressIP": "",
                    "livenessProbe": {
                      "failureThreshold": 3,
                      "periodSeconds": 5,
                      "successThreshold": 1
                    },
                    "mgmtHttpPort": 9080,
                    "mgmtHttpsPort": 9443,
                    "nitroReadTimeout": 20,
                    "nsConfigDnsRec": false,
                    "nsCookieVersion": 0,
                    "nsDnsNameserver": "",
                    "nsEnableLabels": true,
                    "nsGateway": "192.168.1.1",
                    "nsHttp2ServerSide": "OFF",
                    "nsIP": "192.168.1.2",
                    "nsLbHashAlgo": {
                      "hashAlgorithm": "default",
                      "hashFingers": 500,
                      "required": false
                    },
                    "nsProtocol": "HTTP",
                    "nsSvcLbDnsRec": false,
                    "prometheusCredentialSecret": "",
                    "pullPolicy": "IfNotPresent",
                    "service": {
                      "annotations": {},
                      "spec": {}
                    },
                    "startupProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1
                    }
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "replicaCount": 1,
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerGslbController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalergslbcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "gslbController": {
                    "disableAPIServerCertVerify": false,
                    "entityPrefix": "",
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "kubernetesURL": "",
                    "localCluster": "",
                    "localRegion": "",
                    "logLevel": "Debug",
                    "openshift": true,
                    "pullPolicy": "IfNotPresent",
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "32m",
                        "memory": "128Mi"
                      }
                    },
                    "siteData": [
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteMask": "",
                        "siteName": "",
                        "sitePublicIp": "",
                        "siteRegion": ""
                      },
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteMask": "",
                        "siteName": "",
                        "sitePublicIp": "",
                        "siteRegion": ""
                      }
                    ]
                  },
                  "imagePullSecrets": [],
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsIP": "",
                    "nsPort": 443,
                    "nsProtocol": "HTTPS"
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerIngressController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscaleringresscontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "analyticsConfig": {
                    "distributedTracing": {
                      "enable": false,
                      "samplingRate": 100
                    },
                    "endpoint": {
                      "metrics": {
                        "service": ""
                      },
                      "transactions": {
                        "service": ""
                      }
                    },
                    "required": false,
                    "timeseries": {
                      "auditLogs": {
                        "enable": false
                      },
                      "events": {
                        "enable": false
                      },
                      "metrics": {
                        "enable": false,
                        "enableNativeScrape": false,
                        "exportFrequency": 50,
                        "mode": "avro",
                        "schemaFile": "schema.json"
                      },
                      "port": 50002
                    },
                    "transactions": {
                      "enable": false,
                      "port": 50005
                    }
                  },
                  "exporter": {
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4",
                    "ports": {
                      "containerPort": 8888
                    },
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {},
                    "serviceMonitorExtraLabels": {}
                  },
                  "extraVolumes": [],
                  "hostAliases": [],
                  "imagePullSecrets": [],
                  "ingressController": {
                    "clusterName": "",
                    "defaultSSLCertSecret": "",
                    "defaultSSLSNICertSecret": "",
                    "disableAPIServerCertVerify": false,
                    "disableOpenshiftRoutes": true,
                    "enableLivenessProbe": true,
                    "enableReadinessProbe": true,
                    "entityPrefix": "",
                    "extraVolumeMounts": [],
                    "ignoreNodeExternalIP": false,
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "ingressClass": [],
                    "ipam": false,
                    "jsonLog": false,
                    "kubernetesURL": "",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "logLevel": "INFO",
                    "logProxy": "",
                    "multiClusterPrefix": "",
                    "namespaceLabels": "",
                    "nodeLabels": "",
                    "nodeWatch": false,
                    "nsValidateCert": {},
                    "nsncPbr": false,
                    "openshift": true,
                    "optimizeEndpointBinding": false,
                    "podIPsforServiceGroupMembers": false,
                    "profileHttpFrontend": {},
                    "profileSslFrontend": {},
                    "profileTcpFrontend": {},
                    "pullPolicy": "IfNotPresent",
                    "rbacRole": false,
                    "readinessProbe": {
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 60,
                      "successThreshold": 1
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "32m",
                        "memory": "128Mi"
                      }
                    },
                    "routeLabels": "",
                    "serviceClass": [],
                    "setAsDefaultIngressClass": false,
                    "updateIngressStatus": true
                  },
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsConfigDnsRec": false,
                    "nsCookieVersion": 0,
                    "nsDnsNameserver": "",
                    "nsHttp2ServerSide": "OFF",
                    "nsIP": "",
                    "nsLbHashAlgo": {
                      "hashAlgorithm": "default",
                      "hashFingers": 256,
                      "required": false
                    },
                    "nsPort": 443,
                    "nsProtocol": "HTTPS",
                    "nsSNIPs": [],
                    "nsSvcLbDnsRec": false,
                    "nsVIP": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerIpamController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscaleripamcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "clusterName": "",
                  "image": "quay.io/netscaler/netscaler-ipam-controller@sha256:7c25b59a2e1df0e4f5615a779a9bea939c60a3eca065cb6c2295daf0d03831d8",
                  "localIPAM": {
                    "vipRange": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "reuseIngressVIP": true,
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerObservabilityExporter",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalerobservabilityexporter-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "elasticsearch": {
                    "enabled": false,
                    "server": "http://elasticsearch:9200"
                  },
                  "headless": false,
                  "image": "quay.io/netscaler/netscaler-observability-exporter@sha256:626e3a954c67bbb910477352dc740aaab21ff1c1cfb03354becf50beab17d4a4",
                  "jsonTransRateLimiting": {
                    "enabled": false,
                    "limit": 100,
                    "queueLimit": 1000,
                    "window": 5
                  },
                  "kafka": {
                    "auditLogs": "no",
                    "broker": "",
                    "dataFormat": "AVRO",
                    "enabled": false,
                    "events": "no",
                    "topic": "HTTP"
                  },
                  "nodePortRequired": false,
                  "nodeSelector": {},
                  "nsoeLogLevel": "INFO",
                  "nstracing": {
                    "enabled": false,
                    "server": "zipkin:9411/api/v1/spans"
                  },
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "replicas": 1,
                  "resources": {},
                  "splunk": {
                    "authToken": "",
                    "enabled": false,
                    "indexPrefix": "adc_nsoe",
                    "server": "http://splunk:8088"
                  },
                  "timeseries": {
                    "enabled": false
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-operator@sha256:e367c8c8ba6171e9ef5dd496592a5c62f4aa764f93d0cb78835aa29024c063d0
          createdAt: "2025-04-15T08:51:39Z"
          description: This is an operator to install Cloud Native portfolio provided
            by NetScaler
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler Cloud Controller for NetScaler VPX/MPX/SDX
              with Azure Redhat OpenShift in Azure.
            displayName: Netscaler Cloud Controller
            kind: NetscalerCloudController
            name: netscalercloudcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler CPX with Ingress Controller.
            displayName: Netscaler CPX With Ingress Controller
            kind: NetscalerCpxWithIngressController
            name: netscalercpxwithingresscontrollers.netscaler.com
            version: v1
          - description: To install NetScaler GSLB Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler GSLB Controller
            kind: NetscalerGslbController
            name: netscalergslbcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler Ingress Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler Ingress Controller
            kind: NetscalerIngressController
            name: netscaleringresscontrollers.netscaler.com
            version: v1
          - description: To install NetScaler IPAM Controller for NetScaler CPX/VPX/MPX/SDX.
            displayName: Netscaler IPAM Controller
            kind: NetscalerIpamController
            name: netscaleripamcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler Observability Exporter for NetScaler.
            displayName: Netscaler Observability Exporter
            kind: NetscalerObservabilityExporter
            name: netscalerobservabilityexporters.netscaler.com
            version: v1
        description: |
          NetScaler provides various products to empower Developer/DevOps managing your applications in OpenShift Cluster. Using this operator you can deploy below controllers for NetScaler appliance:
          1. [NetScaler Cloud Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. [NetScaler Ingress Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/unified-ingress-deployment)
          3. [NetScaler CPX with Ingress Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/two-tier-deployment)
          4. [NetScaler GSLB Controller ](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/gslb/gslb)
          5. [NetScaler Observability Exporter](https://docs.netscaler.com/en-us/netscaler-observability-exporter/deploy-coe)
          6. [NetScaler IPAM Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/ipam-installation.html)

          ### Installation
          Refer [installation instructions](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-operator-openshift).

          This operator version contains:
          1. NetScaler Ingress Controller version [2.2.10](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.2.10)
          2. NetScaler CPX version [14.1-25.111](https://docs.netscaler.com/en-us/updates?product=NetScaler-ADC-%28includes-NetScaler-Gateway%29&version=14.1&build=25.56)
          3. NetScaler GSLB Controller version [2.2.10](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.2.10)
          4. NetScaler Observability Exporter version [1.10.001](https://github.com/citrix/citrix-helm-charts/releases/tag/nsoe-v1.10.001)
          5. NetScaler Metrics Exporter [1.4.9](https://github.com/citrix/citrix-adc-metrics-exporter/releases/tag/1.4.9)
          6. NetScaler IPAM Controller [2.1.2](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.3.15)

          ### Support

          For debugging issues, please use [NetScaler Kubernetes Plugin](https://github.com/netscaler/modern-apps-toolkit/tree/main/netscaler-plugin).

          For questions and support, the following channels are available:
          - [NetScaler community portal](https://community.netscaler.com/s/)
          - [Contact us](https://www.netscaler.com/contact)
        displayName: NetScaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        - ADC
        - Kubernetes
        - OpenShift
        - Citrix
        - Controller
        - Observability
        - Metrices
        links:
        - name: Netscaler Operator
          url: https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-operator-openshift
        maintainers:
        - email: priyanka.sharma@cloud.com
          name: Priyanka Sharma
        - email: arijit.ray@cloud.com
          name: Arijit Ray
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: NetScaler
          url: https://netscaler.com
        relatedImages:
        - registry.connect.redhat.com/citrix/netscaler-operator@sha256:e367c8c8ba6171e9ef5dd496592a5c62f4aa764f93d0cb78835aa29024c063d0
        - quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f
        - quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb
        - quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4
        - quay.io/netscaler/netscaler-cpx@sha256:232596eaacda68c0e2e210f26b64cedc8d70ae12c5e8ae2e1a66896a479ae4f8
        - quay.io/netscaler/netscaler-observability-exporter@sha256:626e3a954c67bbb910477352dc740aaab21ff1c1cfb03354becf50beab17d4a4
        - quay.io/netscaler/netscaler-ipam-controller@sha256:7c25b59a2e1df0e4f5615a779a9bea939c60a3eca065cb6c2295daf0d03831d8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:89f44c45373da73a51b0b9a02de7efe519dc2d5343b65fe382526cccc92f8f5c
        version: 3.2.0
      entries:
      - name: netscaler-operator.v3.2.0
        version: 3.2.0
      - name: netscaler-operator.v3.1.2
        version: 3.1.2
      - name: netscaler-operator.v3.1.1
        version: 3.1.1
      - name: netscaler-operator.v3.0.1
        version: 3.0.1
      name: stable
    - currentCSV: netscaler-operator.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCloudController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netscaler-operator",
                    "app.kubernetes.io/instance": "nscc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscalercloudcontroller",
                    "app.kubernetes.io/part-of": "netscaler-operator"
                  },
                  "name": "nscc"
                },
                "spec": {
                  "azure": {
                    "clientsecret": "",
                    "image": "quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb",
                    "location": "",
                    "pullPolicy": "IfNotPresent",
                    "resourcegroupname": "",
                    "resources": {},
                    "subnetname": "",
                    "subscriptionid": "",
                    "tenantid": "",
                    "vnetname": ""
                  }
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerGslbController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netscaler-operator",
                    "app.kubernetes.io/instance": "nsgc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscalergslbcontroller",
                    "app.kubernetes.io/part-of": "netscaler-operator"
                  },
                  "name": "nsgc"
                },
                "spec": {
                  "gslbController": {
                    "acceptLicense": "Yes",
                    "cleanupGSLBSiteConfig": false,
                    "disableAPIServerCertVerify": false,
                    "entityPrefix": "",
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:88114c518ecceb54b000f38bfc330efe184bfeb5b5df0ee308950cf29e7a777f",
                    "imagePullSecrets": [],
                    "kubernetesURL": "",
                    "localCluster": "",
                    "localRegion": "",
                    "logLevel": "Debug",
                    "openshift": true,
                    "pullPolicy": "IfNotPresent",
                    "siteData": [
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteName": "",
                        "siteRegion": ""
                      },
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteName": "",
                        "siteRegion": ""
                      }
                    ]
                  },
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsIP": "",
                    "nsPort": 443,
                    "nsProtocol": "HTTPS"
                  },
                  "resources": {
                    "limits": {},
                    "requests": {
                      "cpu": "32m",
                      "memory": "128Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-operator@sha256:3f71f9a46d7b29e3da98a301965f635c03a022b5535b586d7d5c699341f68193
          createdAt: "2024-07-31T06:33:49Z"
          description: This is an operator to install Cloud Native portfolio provided
            by NetScaler
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler Cloud Controller for NetScaler VPX/MPX/SDX
              with Azure Redhat OpenShift in Azure.
            displayName: Netscaler Cloud Controller
            kind: NetscalerCloudController
            name: netscalercloudcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler GSLB Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler GSLB Controller
            kind: NetscalerGslbController
            name: netscalergslbcontrollers.netscaler.com
            version: v1
        description: |-
          NetScaler provides various products to empower Developer/DevOps managing your applications in OpenShift Cluster. Using this operator you can deploy below controllers for NetScaler appliance:
          1. [NetScaler Cloud Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. [NetScaler GSLB Controller ](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/gslb-openshift-operator)

          ### Installation
          Refer [installation instructions](https://github.com/netscaler/netscaler-k8s-ingress-controller/blob/master/docs/deploy/deploy-ns-operator.md).

          This operator version contains:
          1. NetScaler Cloud Controller version [1.2.0](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. NetScaler GSLB Controller version [1.43.7](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v1.43.7)

          ### Support

          For debugging issues, please use [NetScaler Kubernetes Plugin](https://github.com/netscaler/modern-apps-toolkit/tree/main/netscaler-plugin).

          For questions and support, the following channels are available:
          - [NetScaler community portal](https://community.netscaler.com/s/)
          - [Contact us](https://www.netscaler.com/contact)
        displayName: NetScaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        links:
        - name: Netscaler Operator
          url: https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/gslb-openshift-operator
        maintainers:
        - email: priyanka.sharma@cloud.com
          name: Priyanka Sharma
        - email: arijit.ray@cloud.com
          name: Arijit Ray
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: NetScaler
          url: https://netscaler.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:23405618050a4ceea8ea0c748cc7d9d16dbb80f10e4c39f7e82014a3ac80cec9
        - registry.connect.redhat.com/citrix/netscaler-operator@sha256:3f71f9a46d7b29e3da98a301965f635c03a022b5535b586d7d5c699341f68193
        - quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:88114c518ecceb54b000f38bfc330efe184bfeb5b5df0ee308950cf29e7a777f
        - quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb
        version: 3.0.1
      entries:
      - name: netscaler-operator.v3.0.1
        version: 3.0.1
      name: tech-preview
    defaultChannel: stable
    packageName: netscaler-operator
    provider:
      name: NetScaler
      url: https://netscaler.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: open-liberty-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: open-liberty-operator.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4",
                  "expose": true,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/open-liberty-operator-controller@sha256:18f188e48725c785557c8edea46f69c464d6f81ab47860a3bb0ec978c20a26db
          createdAt: "2022-03-25T14:00:00Z"
          description: Deploy and manage applications running on Liberty
          olm.skipRange: <0.8.1
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          This advanced Operator can be used to deploy and manage Open Liberty applications with consistent, production-grade QoS. This operator is based on the [Runtime Component Operator](https://github.com/application-stacks/runtime-component-operator) and provides all of its capabilities in addition to Open Liberty specific features such as gathering traces and dumps (Day-2 operations) and easily configuring and managing the single sign-on information for your Open Liberty applications.

          Open Liberty Operator enables enterprise architects to govern the way their applications get deployed & managed in the cluster, while dramatically reducing the learning curve for developers to deploy into Kubernetes - allowing them to focus on writing the code!

          Here are some key features:

          #### Application Lifecyle
          You can deploy your Open Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the Operator will watch for any updates and will re-deploy the modified image.

          #### Custom RBAC
          This Operator is capable of using a custom ServiceAccount from the caller, allowing it to follow RBAC restrictions. By default it creates a ServiceAccount if one is not specified, which can also be bound with specific roles.

          #### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as: labels, annotations, and environment variables from a ConfigMap, a Secret or a value.

          #### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configuration, such as TLS settings, are also easily enabled.

          #### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto scaling to create (and delete) instances based on resource consumption.

          #### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you.
          This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart.
          We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          You can also easily configure and use a single storage for serviceability related Day-2 operations, such as gatherig server traces and dumps.

          #### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating kubernetes Secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to its required services without any intervention or interruption.

          #### Single Sign-On (SSO)
          Open Liberty provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows to easily configure and manage the single sign-on information for your applications.

          #### Exposing metrics to Prometheus
          Expose the Open Liberty application's metrics via the Prometheus Operator.
          You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `spec.monitoring` field to control configurations such as poll interval and security credentials.

          #### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply add the following configuration (to specify the volume size and the location to persist) to your OpenLibertyApplication CR:
          ``` storage: size: 2Gi mountPath: "/logs" ```

          #### Integration with OpenShift Serverless
          Deploy your serverless runtime component using a single toggle.  The Operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing your pod to automatically scale to 0 when it is idle.

          #### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and how they are connected.

          See our [**documentation**](https://github.com/OpenLiberty/open-liberty-operator/tree/main/doc/) for more information.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://github.com/OpenLiberty/open-liberty-operator/tree/main/doc
        - name: Open Liberty
          url: https://openliberty.io/
        - name: Open Liberty Container Images
          url: https://github.com/OpenLiberty/ci.docker
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: beta
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/open-liberty-operator-controller@sha256:18f188e48725c785557c8edea46f69c464d6f81ab47860a3bb0ec978c20a26db
        - registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4
        version: 0.8.1
      entries:
      - name: open-liberty-operator.v0.8.1
        version: 0.8.1
      - name: open-liberty-operator.v0.8.0
        version: 0.8.0
      name: beta2
    - currentCSV: open-liberty-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:d45c296d43dc4c4100f38bb736f073e04c8b4d72b68ddf1189e311aff8ad66e3
          createdAt: "2023-08-28T16:39:10Z"
          description: Deploy and manage containerized Liberty applications
          olm.skipRange: '>=0.8.0 <1.2.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:d45c296d43dc4c4100f38bb736f073e04c8b4d72b68ddf1189e311aff8ad66e3
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02
        version: 1.2.2
      entries:
      - name: open-liberty-operator.v1.2.2
        version: 1.2.2
      name: v1.2
    - currentCSV: open-liberty-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:d68bb66a8133702a198f7be655d7cff5f500bb620c460a81547971e00124ef73
          createdAt: "2024-08-01T20:19:13Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.3.3'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:d68bb66a8133702a198f7be655d7cff5f500bb620c460a81547971e00124ef73
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91
        version: 1.3.3
      entries:
      - name: open-liberty-operator.v1.3.3
        version: 1.3.3
      name: v1.3
    - currentCSV: open-liberty-operator.v1.4.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:67b5b46dfed008dab4ea7fd176d3bc31dae8940d72194a10aebb98b1dea9043e
          createdAt: "2025-07-18T20:45:06Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.4.4'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:67b5b46dfed008dab4ea7fd176d3bc31dae8940d72194a10aebb98b1dea9043e
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3
        version: 1.4.4
      entries:
      - name: open-liberty-operator.v1.4.4
        version: 1.4.4
      name: v1.4
    - currentCSV: open-liberty-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyPerformanceData",
                "metadata": {
                  "name": "openliberty-performance-data-sample"
                },
                "spec": {
                  "interval": 30,
                  "podName": "Specify_Pod_Name_Here",
                  "timespan": 240
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:c3b031fe14054f65d6612cc505b7681336873dc32ce1091451d8c7799034bc83
          createdAt: "2025-09-16T03:08:31Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.5.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server performance data
            displayName: OpenLibertyPerformanceData
            kind: OpenLibertyPerformanceData
            name: openlibertyperformancedata.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:c3b031fe14054f65d6612cc505b7681336873dc32ce1091451d8c7799034bc83
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8
        version: 1.5.0
      entries:
      - name: open-liberty-operator.v1.5.0
        version: 1.5.0
      name: v1.5
    defaultChannel: v1.5
    packageName: open-liberty-certified
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: RabbitMQ
      provider-url: https://rabbitmq.com
    name: rabbitmq-cluster-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-cluster-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rabbitmq.com/v1beta1",
                "kind": "RabbitmqCluster",
                "metadata": {
                  "name": "sample",
                  "namespace": "rabbitmq-system"
                },
                "spec": {
                  "replicas": 1,
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-07-10T12:52:23Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RabbitmqCluster is the Schema for the RabbitmqCluster API.
              Each instance of this object corresponds to a single RabbitMQ cluster.
            displayName: Rabbitmq Cluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: RabbitMQ Cluster Operator
        displayName: RabbitMQ Cluster Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        links:
        - name: Cluster Operator
          url: https://github.com/rabbitmq/cluster-operator
        maturity: alpha
        provider:
          name: RabbitMQ
          url: https://rabbitmq.com
        relatedImages:
        - registry.redhat.io/rhoso-operators/rabbitmq-cluster-rhel9-operator@sha256:e4b198236d70b7dfdd29d8ed45058dec3335b3530e0b676fcc862bfec2efce79
        version: 1.0.6
      entries:
      - name: rabbitmq-cluster-operator.v1.0.6
        version: 1.0.6
      - name: rabbitmq-cluster-operator.v1.0.5
        version: 1.0.5
      - name: rabbitmq-cluster-operator.v1.0.4
        version: 1.0.4
      - name: rabbitmq-cluster-operator.v1.0.3
        version: 1.0.3
      - name: rabbitmq-cluster-operator.v1.0.2
        version: 1.0.2
      - name: rabbitmq-cluster-operator.v1.0.1
        version: 1.0.1
      - name: rabbitmq-cluster-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: rabbitmq-cluster-operator
    provider:
      name: RabbitMQ
      url: https://rabbitmq.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-route53resolver-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-route53resolver-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "route53resolver.services.k8s.aws/v1alpha1",
                "kind": "ResolverEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "route53resolver.services.k8s.aws/v1alpha1",
                "kind": "ResolverRule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/route53resolver-controller:1.1.1
          createdAt: "2025-09-20T01:21:08Z"
          description: AWS Route53Resolver controller is a service controller for
            managing Route53Resolver resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ResolverEndpoint represents the state of an AWS route53resolver
              ResolverEndpoint resource.
            displayName: ResolverEndpoint
            kind: ResolverEndpoint
            name: resolverendpoints.route53resolver.services.k8s.aws
            version: v1alpha1
          - description: ResolverRule represents the state of an AWS route53resolver
              ResolverRule resource.
            displayName: ResolverRule
            kind: ResolverRule
            name: resolverrules.route53resolver.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Route53Resolver resources in AWS from within your Kubernetes cluster.

          **About Amazon Route53Resolver**

          Amazon Route 53 Resolver responds recursively to DNS queries from AWS resources for public records, Amazon VPC-specific DNS names, and Amazon Route 53 private hosted zones, and is available by default in all VPCs.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Route53Resolver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - route53resolver
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Route53Resolver Developer Resources
          url: https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/resolver.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: route53resolver maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/route53resolver-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-route53resolver-controller.v1.1.1
        version: 1.1.1
      - name: ack-route53resolver-controller.v1.0.15
        version: 1.0.15
      - name: ack-route53resolver-controller.v1.0.14
        version: 1.0.14
      - name: ack-route53resolver-controller.v1.0.13
        version: 1.0.13
      - name: ack-route53resolver-controller.v1.0.11
        version: 1.0.11
      - name: ack-route53resolver-controller.v1.0.10
        version: 1.0.10
      - name: ack-route53resolver-controller.v1.0.9
        version: 1.0.9
      - name: ack-route53resolver-controller.v1.0.7
        version: 1.0.7
      - name: ack-route53resolver-controller.v1.0.6
        version: 1.0.6
      - name: ack-route53resolver-controller.v1.0.5
        version: 1.0.5
      - name: ack-route53resolver-controller.v1.0.3
        version: 1.0.3
      - name: ack-route53resolver-controller.v1.0.2
        version: 1.0.2
      - name: ack-route53resolver-controller.v1.0.0
        version: 1.0.0
      - name: ack-route53resolver-controller.v0.0.11
        version: 0.0.11
      - name: ack-route53resolver-controller.v0.0.10
        version: 0.0.10
      - name: ack-route53resolver-controller.v0.0.9
        version: 0.0.9
      - name: ack-route53resolver-controller.v0.0.8
        version: 0.0.8
      - name: ack-route53resolver-controller.v0.0.7
        version: 0.0.7
      - name: ack-route53resolver-controller.v0.0.4
        version: 0.0.4
      - name: ack-route53resolver-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-route53resolver-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Weaveworks
      provider-url: ""
    name: tf-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tf-controller.v0.9.0-rc.8
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"infra.contrib.fluxcd.io/v1alpha1\",\n
            \   \"kind\": \"Terraform\",\n    \"metadata\": {\n        \"name\": \"helloworld-tf\",\n
            \       \"namespace\": \"flux-system\"\n    },\n    \"spec\": {\n        \"interval\":
            \"5m\",\n        \"approvePlan\": \"auto\",\n        \"path\": \"./\",\n
            \       \"sourceRef\": {\n          \"kind\": \"GitRepository\",\n          \"name\":
            \"helloworld\",\n          \"namespace\": \"flux-system\"\n        }      \n
            \   }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/weaveworks/tf-controller:v0.9.0-rc.8
          description: Weave TF-controller is a Flux controller that brings GitOps
            to Terraform resources.
          operatorframework.io/suggested-namespace: flux-system
          support: Weaveworks
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Terraform
            displayName: Terraform
            kind: Terraform
            name: terraforms.infra.contrib.fluxcd.io
            version: v1alpha1
        description: "**Weave GitOps Terraform Controller** (aka Weave TF-controller)
          is a Flux controller for reconciling Terraform resources in the GitOps way.
          \nWith the power of Flux together with Terraform, TF-controller allows you
          to GitOps-ify infrastructure, \nand application resources, in the Kubernetes
          and Terraform universe, together at your own pace.\n\nTF-controller offers
          many GitOps models:\n  1. **GitOps Automation Model:** GitOps your Terraform
          resources from the provision steps to the enforcement steps, like a whole
          EKS cluster.\n  2. **Hybrid GitOps Automation Model:** GitOps parts of your
          existing infrastructure resources. For example, you have an existing EKS
          cluster. You can choose to GitOps only its nodegroup, or its security group.\n
          \ 3. **State Enforcement Model:** You have a TFSTATE file, and you'd like
          to use GitOps enforce it, without changing anything else.\n  4. **Drift
          Detection Model:** You have a TFSTATE file, and you'd like to use GitOps
          just for drift detection, so you can decide to do things later when a drift
          occurs.\n\nFeatures\n  * GitOps Automation for Terraform\n  * Drift detection\n
          \ * Plan and Manual Approve"
        displayName: Weave GitOps Terraform Controller
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - weave
        - flux
        - cd
        - continuous delivery
        - terraform
        - gitops
        links:
        - name: Weave GitOps Terraform Controller
          url: https://github.com/weaveworks/tf-controller
        - name: Documentation
          url: https://weaveworks.github.io/tf-controller/
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Weaveworks
        relatedImages:
        - ghcr.io/weaveworks/tf-controller:v0.9.0-rc.8
        version: 0.9.0-rc.8
      entries:
      - name: tf-controller.v0.9.0-rc.8
        version: 0.9.0-rc.8
      - name: tf-controller.v0.9.0-rc.3
        version: 0.9.0-rc.3
      name: stable
    defaultChannel: stable
    packageName: tf-controller
    provider:
      name: Weaveworks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Randoli
      provider-url: https://www.randoli.io
    name: app-insights-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: app-insights-operator.v1.17.5-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "appinsights.randoli/v1beta1",
                "kind": "RandoliAgent",
                "metadata": {
                  "name": "randoli",
                  "namespace": "randoli-agents"
                },
                "spec": {
                  "costManagement": {
                    "opencost": {
                      "exporter": {
                        "image": {
                          "tag": "1.112.1"
                        }
                      }
                    },
                    "vpaOperator": {
                      "enabled": false
                    }
                  },
                  "logs": {
                    "vector": {
                      "agent": {
                        "containerSecurityContext": {
                          "privileged": true
                        }
                      }
                    }
                  },
                  "network": {
                    "standaloneConsole": {
                      "enable": true
                    }
                  },
                  "observability": {
                    "netobserv": {
                      "enabled": false
                    },
                    "otel": {
                      "operator": {
                        "enabled": false
                      }
                    }
                  },
                  "resources": {
                    "limits": {
                      "memory": "2Gi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "512Mi"
                    }
                  },
                  "securityContext": {
                    "fsGroup": 1001
                  },
                  "tags": {
                    "costManagement": false,
                    "observability": false
                  }
                }
              }
            ]
          categories: Developer Tools
          containerImage: docker.io/randoli/app-insights-operator:1.17.5-4@sha256:22e7150c29557f2a71186389e5d7d60bfe82a4ea022e1403033c6ad61bc95f7a
          createdAt: "2025-09-23T19:45:43Z"
          description: Randoli Agent connects OpenShift clusters to Randoli Observability
            & Cost Management.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[]'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Randoli
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: RandoliAgent
            name: randoliagents.appinsights.randoli
            version: v1beta1
        description: |-
          Randoli Agent connects OpenShift clusters to Randoli Observability & Cost Management platforms.

          This operator provides a container-based agent to be installed in your OpenShift clusters for connecting to Randoli Observability & Cost Management platforms

          **Prerequisites**

          If you are using the Randoli Observability platform, you will need to install the following Operators:
          - Red Hat build of OpenTelemetry Operator
          - Network Observability

          If you are using the Randoli Cost Management platform, you will need to install the following Operator:
          - VerticalPodAutoscaler

          For additional information please use our Openshift installation guide here: [https://docs.insights.randoli.io/agent/openshift](https://docs.insights.randoli.io/agent/openshift)

           **Randoli Observability for Kubernetes**

          [Product Page](https://www.randoli.io/product/kubernetes-observability)

          **Randoli Cost Management for Kubernetes**

          [Product Page](https://www.randoli.io/product/kubernetes-cost-management)
        displayName: Randoli Agent
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - developer tools
        - insights
        - monitoring
        links:
        - name: Randoli Observability for Kubernetes
          url: https://www.randoli.io/product/kubernetes-observability
        - name: Randoli Cost Management for Kubernetes
          url: https://www.randoli.io/product/kubernetes-cost-management
        maintainers:
        - email: contact@randoli.ca
          name: Randoli Inc.
        maturity: beta
        minKubeVersion: 1.24.0
        provider:
          name: Randoli
          url: https://www.randoli.io
        relatedImages:
        - docker.io/randoli/app-insights-operator:1.17.5-4@sha256:22e7150c29557f2a71186389e5d7d60bfe82a4ea022e1403033c6ad61bc95f7a
        version: 1.17.5-4
      entries:
      - name: app-insights-operator.v1.17.5-4
        version: 1.17.5-4
      - name: app-insights-operator.v1.17.5-3
        version: 1.17.5-3
      - name: app-insights-operator.v1.17.5-2
        version: 1.17.5-2
      - name: app-insights-operator.v1.17.5-1
        version: 1.17.5-1
      - name: app-insights-operator.v1.17.3-1
        version: 1.17.3-1
      - name: app-insights-operator.v1.17.2-1
        version: 1.17.2-1
      - name: app-insights-operator.v1.17.1-1
        version: 1.17.1-1
      - name: app-insights-operator.v1.17.0-2
        version: 1.17.0-2
      - name: app-insights-operator.v1.17.0-1
        version: 1.17.0-1
      name: stable
    defaultChannel: stable
    packageName: app-insights-operator
    provider:
      name: Randoli
      url: https://www.randoli.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ProphetStor Data Services, Inc.
      provider-url: ""
    name: federatorai-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: federatorai.v4.7.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "federatorai.containers.ai/v1alpha1",
                "kind": "AlamedaService",
                "metadata": {
                  "name": "my-alamedaservice"
                },
                "spec": {
                  "alamedaAi": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "10Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "alamedaInfluxdb": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "100Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "fedemeterInfluxdb": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "10Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "imageLocation": "registry.connect.redhat.com/prophetstor",
                  "storages": [
                    {
                      "accessMode": ["ReadWriteOnce"],
                      "size": "2Gi",
                      "type": "pvc",
                      "usage": "log"
                    }
                  ],
                  "version": "v4.7.1-b1760"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: registry.connect.redhat.com/prophetstor/federatorai-operator@sha256:bdd87f6b7d7ba9d12ed8ddb9c5151e5a89781e88d7a02b73814b16a766968047
          createdAt: "2021-10-12T01:14:00Z"
          description: Federator.ai Operator provides easy configuration and management
            of AI-based Kubernetes resource orchestrator
          operatorframework.io/suggested-namespace: federatorai
          repository: https://access.redhat.com/containers/#/registry.connect.redhat.com/prophetstor/federatorai-operator
          support: ProphetStor Data Services, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of Federator.ai.
            displayName: AlamedaService
            kind: AlamedaService
            name: alamedaservices.federatorai.containers.ai
            version: v1alpha1
        description: |-
          Federator.ai helps enterprises optimize cloud resources, maximize application performance, and save significant cost without excessive over-provisioning or under-provisioning of resources, meeting the service-level requirements of their applications.

          Enterprises often lack understanding of the resources needed to support their applications. This leads to either excessive over-provisioning or under-provisioning of resources (CPU, memory, storage). Using machine learning, Federator.ai determines the optimal cloud resources needed to support any workload on OpenShift and helps users find the best-cost instances from cloud providers for their applications.


          **Multi-layer workload prediction**

          Using machine learning and math-based algorithms, Federator.ai predicts containerized application and cluster node resource usage as the basis for resource recommendations at application level as well as at cluster node level. Federator.ai supports prediction for both physical/virtual CPUs and memories.


          **Auto-scaling via resource recommendation**

          Federator.ai utilizes the predicted resource usage to recommend the right number and size of pods for applications. Integrated with Datadog's WPA, applications are automatically scaled to meet the predicted resource usage.


          **Application-aware recommendation execution**

          Optimizing the resource usage and performance goals, Federator.ai uses application specific metrics for workload prediction and pod capacity estimation to auto-scale the right number of pods for best performance without overprovisioning.


          **Multi-cloud Cost Analysis**

          With resource usage prediction, Federator.ai analyzes potential cost of a cluster on different public cloud providers. It also recommend appropriate cluster nodes and instance types based on resource usage.


          **Custom Datadog/Sysdig Dashboards**

          Predefined custom Datadog/Sysdig Dashboards for workload prediction/recommendation visualization for cluster nodes and applications.


          **Additional resources**

          Want more product information? Explore detailed information about using this product and where to find additional help.

          * [Federator.ai Datasheet](https://prophetstor.com/wp-content/uploads/datasheets/Federator.ai.pdf)
          * [Quick Start Guide](https://prophetstor.com/wp-content/uploads/documentation/Federator.ai/Latest%20Version/ProphetStor%20Federator.ai%20Quick%20Installation%20Guide.pdf)
          * [Installation Guide](https://prophetstor.com/wp-content/uploads/2021/08/ProphetStor-Federator.ai-v4.7.0-Installation-Guide-v1.0.pdf)
          * [User Guide](https://prophetstor.com/wp-content/uploads/2021/08/Federator.ai-4.7.0-User-Guide.pdf)
          * [Release Notes](https://prophetstor.com/wp-content/uploads/2021/08/Federator.ai-4.7.0-Release-Notes.pdf)
          * [Company Information](https://www.prophetstor.com/)


          ### Prerequisite
          *  The [Openshift](https://www.openshift.com) version 4.4.x or later if using OpenShift platform.
          *  The [Kubernetes](https://kubernetes.io/) version 1.16 or later if using Kubernetes environment.
        displayName: Federator.ai
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - AI
        - Resource Orchestration
        - NoOps
        links:
        - name: Website
          url: https://prophetstor.com/federator-ai-2/
        - name: Quickstart guide
          url: https://prophetstor.com/wp-content/uploads/documentation/Federator.ai/Latest%20Version/ProphetStor%20Federator.ai%20Quick%20Installation%20Guide.pdf
        maintainers:
        - email: support@prophetstor.com
          name: ProphetStor Data Services, Inc.
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: ProphetStor Data Services, Inc.
        relatedImages:
        - registry.connect.redhat.com/prophetstor/alameda-recommender-ubi@sha256:13e941df9252cea3024ecb5e2247a696b64928015c1a3a4a04ee4e362388b1e4
        - registry.connect.redhat.com/prophetstor/federatorai-agent-preloader@sha256:66daea6193e399f8363df5ba42e1fcd643db6df3a9075dbcd8c3464fdc91e776
        - registry.connect.redhat.com/prophetstor/alameda-executor-ubi@sha256:4512e3be7a89d0ca9f120b68f6ab9766d000bf2b01b279cd55b2dbc433aacfb6
        - registry.connect.redhat.com/prophetstor/alameda-influxdb@sha256:485c2119a6a56807e28ad03b2087129950118b55531126a66bb414063052ae76
        - registry.connect.redhat.com/prophetstor/alameda-analyzer-ubi@sha256:6117ac4692793d49bc0204f13091dec119bdbc7b169e9deb86c243aca00ded27
        - registry.connect.redhat.com/prophetstor/fedemeter-api-ubi@sha256:451456207fe8d16b9824b3f64d00645766dcccbc8d685b6942e6c00d0d0c627e
        - registry.connect.redhat.com/prophetstor/alameda-ai-dispatcher@sha256:6e0dc1136b0f8e5971d92b95ce0d6dfac475044c073bada50f23b315f4478ae4
        - registry.connect.redhat.com/prophetstor/alameda-ai@sha256:9293298bb2a72d0b0764050b6f6d8da707b46a83363a3b6aa7be60f512e2ec61
        - registry.connect.redhat.com/prophetstor/federatorai-dashboard-backend@sha256:022dbf6696edb8eaf389fdbd1dff14c40a51c237c01f8a02a7ea65c60cd59f38
        - registry.connect.redhat.com/prophetstor/federatorai-rest-ubi@sha256:3e92528615543ea900bf38387c4eb8439b60759be5be2ac5311a9dbaafc7026b
        - registry.connect.redhat.com/prophetstor/federatorai-agent-ubi@sha256:cfadc84f31f3776ded15d4d8dda92e00b7d9b1bce0ff5440277176292f9304d3
        - registry.connect.redhat.com/prophetstor/federatorai-operator@sha256:bdd87f6b7d7ba9d12ed8ddb9c5151e5a89781e88d7a02b73814b16a766968047
        - registry.connect.redhat.com/prophetstor/federatorai-dashboard-frontend@sha256:1c1147425ca6302103363a1a3bcf27baeef13b10ec6780ef266b7f52fe27ffeb
        - registry.connect.redhat.com/prophetstor/alameda-notifier-ubi@sha256:1f75034fb78d62924a87cb4a5435661b4afa990d6960431e784029742fbae0da
        - registry.connect.redhat.com/prophetstor/alameda-rabbitmq@sha256:b396d9aa0470e593a36f5799c04fc3c9e59b5723f703f4b019f7ef92528e6741
        - registry.connect.redhat.com/prophetstor/fedemeter-influxdb@sha256:36904f003e8e29d4a9fd738655f401fbe840d60158d7d820cda79a09858dc2c2
        - registry.connect.redhat.com/prophetstor/federatorai-data-adapter@sha256:adf5d655fa030dc05133516d45642d7282fdd66e0aa8e7a7331ed614a41d3fe0
        - registry.connect.redhat.com/prophetstor/alameda-operator-ubi@sha256:9a863394cb1c89228223b26addb0d12bb91a97233fe2d5d19ecd5998e0cfe0d6
        - registry.connect.redhat.com/prophetstor/alameda-datahub-ubi@sha256:3819d6acce8d376dbe3b12469be2528d0ce9ac95c2a86507aeb76b6e3a5f5f38
        version: 4.7.1-1
      entries:
      - name: federatorai.v4.7.1-1
        version: 4.7.1-1
      - name: federatorai.v4.6.1-4
        version: 4.6.1-4
      name: stable
    defaultChannel: stable
    packageName: federatorai-certified
    provider:
      name: ProphetStor Data Services, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Starburst Data, Inc.
      provider-url: https://www.starburstdata.com
    name: starburst-enterprise-helm-operator-paygo-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: starburst-enterprise-helm-operator-paygo.v402.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstEnterprise",
                "metadata": {
                  "name": "starburstenterprise-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "cache": {
                    "diskUsagePercentage": 80,
                    "enabled": false,
                    "ttl": "7d",
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "catalogs": {
                    "tpch": "connector.name=tpch\ntpch.splits-per-node=4"
                  },
                  "commonLabels": {},
                  "coordinator": {
                    "additionalProperties": "",
                    "affinity": {},
                    "deploymentAnnotations": {},
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=true\nnode-scheduler.include-coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://localhost:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "resources": {
                      "cpu": 16,
                      "memory": "60Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "tolerations": []
                  },
                  "environment": null,
                  "expose": {
                    "clusterIp": {
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      }
                    },
                    "ingress": {
                      "annotations": {},
                      "host": null,
                      "ingressClassName": null,
                      "ingressName": "coordinator-ingress",
                      "path": "/",
                      "pathType": "ImplementationSpecific",
                      "serviceName": "starburst",
                      "servicePort": 8080,
                      "tls": {
                        "enabled": true,
                        "secretName": null
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "nodePort": 30080,
                          "port": 8080
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "externalSecrets": {
                    "enabled": false,
                    "goDaddy": {
                      "backendType": "secretsManager"
                    },
                    "secretPrefix": "external0",
                    "type": "goDaddy"
                  },
                  "extraArguments": [],
                  "extraSecret": {
                    "file": null,
                    "name": null
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise",
                    "tag": "402-e.1"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "initImage": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise-init",
                    "tag": "1.3.2"
                  },
                  "internal": {
                    "ports": {
                      "http": {
                        "port": 8080
                      },
                      "https": {
                        "port": 8443
                      }
                    }
                  },
                  "internalTls": false,
                  "licenseManager": {
                    "cronjob": {
                      "affinity": {},
                      "nodeSelector": {},
                      "priorityClassName": null,
                      "securityContext": {},
                      "tolerations": []
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "registry.connect.redhat.com/starburst/license-manager",
                      "tag": "0.0.2"
                    },
                    "url": "https://license-api.dogfood.eng.starburstdata.net/test/license/checkout"
                  },
                  "livenessProbe": null,
                  "prometheus": {
                    "agent": {
                      "config": "/etc/starburst/telemetry/prometheus.yaml",
                      "port": 8081,
                      "version": "0.16.1"
                    },
                    "enabled": true,
                    "rules": [
                      {
                        "attrNameSnakeCase": true,
                        "name": "$1",
                        "pattern": "trino.execution<name=QueryManager><>(running_queries|queued_queries)",
                        "type": "GAUGE"
                      },
                      {
                        "name": "failed_queries",
                        "pattern": "trino.execution<name=QueryManager><>FailedQueries\\.TotalCount",
                        "type": "COUNTER"
                      }
                    ]
                  },
                  "query": {
                    "maxConcurrentQueries": 3
                  },
                  "readinessProbe": null,
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "securityContext": {},
                  "sharedSecret": null,
                  "siac": {
                    "additionalHeapSizePercentage": 15,
                    "enabled": false
                  },
                  "spilling": {
                    "enabled": false,
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "startupProbe": null,
                  "usageMetrics": {
                    "enabled": false,
                    "usageClient": {
                      "initialDelay": "1m",
                      "interval": "1m"
                    }
                  },
                  "userDatabase": {
                    "enabled": false,
                    "users": [
                      {
                        "password": "46991b33f7a75ff79213c0dc0e610610",
                        "username": "admin"
                      }
                    ]
                  },
                  "worker": {
                    "additionalProperties": "",
                    "affinity": {},
                    "autoscaling": {
                      "enabled": false,
                      "maxReplicas": 100,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "deploymentAnnotations": {},
                    "deploymentTerminationGracePeriodSeconds": 300,
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://coordinator:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "kedaScaler": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.connect.redhat.com/starburst/keda-trino-scaler",
                        "tag": "0.1.7"
                      },
                      "port": 8021,
                      "scaledObjectSpec": {
                        "advanced": {},
                        "cooldownPeriod": 300,
                        "idleReplicaCount": 0,
                        "maxReplicaCount": 100,
                        "minReplicaCount": 1,
                        "pollingInterval": 30,
                        "scaleTargetRef": {
                          "name": "worker"
                        },
                        "triggers": [
                          {
                            "metadata": {
                              "numberOfQueriesPerWorker": "10",
                              "scaleInToIdleReplicaCountIfNoQueuedQueriesLeft": "true",
                              "scaleMethod": "query_queue"
                            },
                            "type": "external"
                          }
                        ]
                      }
                    },
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "replicas": 2,
                    "resources": {
                      "cpu": 16,
                      "memory": "100Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "starburstWorkerShutdownGracePeriodSeconds": 120,
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstHive",
                "metadata": {
                  "name": "starbursthive-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "affinity": {},
                  "commonLabels": {},
                  "database": {
                    "external": {
                      "driver": null,
                      "jdbcUrl": null,
                      "password": null,
                      "user": null
                    },
                    "internal": {
                      "databaseName": "hive",
                      "databasePassword": "HivePass1234",
                      "databaseUser": "hive",
                      "driver": "org.postgresql.Driver",
                      "env": [],
                      "envFrom": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.redhat.io/rhel8/postgresql-10",
                        "tag": "1"
                      },
                      "port": 5432,
                      "resources": {
                        "limits": {
                          "cpu": 2,
                          "memory": "1Gi"
                        },
                        "requests": {
                          "cpu": 2,
                          "memory": "1Gi"
                        }
                      },
                      "volume": {
                        "emptyDir": {}
                      }
                    },
                    "type": "internal"
                  },
                  "deploymentAnnotations": {},
                  "env": [],
                  "envFrom": [],
                  "expose": {
                    "clusterIp": {
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "nodePort": 30083,
                          "port": 9083
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "extraArguments": [],
                  "hdfs": {
                    "hadoopUserName": null
                  },
                  "heapSizePercentage": 85,
                  "hiveMetastoreStorageAuthorization": true,
                  "hiveMetastoreWarehouseDir": null,
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/hive",
                    "tag": "3.1.2-e.22"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "nodeSelector": {},
                  "objectStorage": {
                    "adl": {
                      "oauth2": {
                        "clientId": null,
                        "credential": null,
                        "refreshUrl": null
                      }
                    },
                    "awsS3": {
                      "accessKey": null,
                      "endpoint": null,
                      "pathStyleAccess": false,
                      "region": null,
                      "secretKey": null
                    },
                    "azure": {
                      "abfs": {
                        "accessKey": {
                          "accessKey": null,
                          "storageAccount": null
                        },
                        "authType": "accessKey",
                        "oauth": {
                          "clientId": null,
                          "endpoint": null,
                          "secret": null
                        }
                      },
                      "wasb": {
                        "accessKey": null,
                        "storageAccount": null
                      }
                    },
                    "gs": {
                      "cloudKeyFileSecret": null
                    }
                  },
                  "podAnnotations": {},
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "resources": {
                    "limits": {
                      "cpu": 1,
                      "memory": "1Gi"
                    },
                    "requests": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "securityContext": {},
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
          description: Deploys StarburstEnterprise Helm Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-paygo-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-paygo-rhmp/support?utm_source=openshift_console
          marketplace.redhat.com/meterDefinition: |-
            {
              "apiVersion": "marketplace.redhat.com/v1beta1",
              "kind": "MeterDefinition",
              "metadata": {
                "name": "sep-vcpu-usage"
              },
              "spec": {
                "group": "charts.starburstdata.com",
                "kind": "StarburstEnterprise",
                "resourceFilters": [
                  {
                    "namespace": {
                      "useOperatorGroup": true
                    },
                    "ownerCRD": {
                      "apiVersion": "charts.starburstdata.com/v1alpha1",
                      "kind": "StarburstEnterprise"
                    },
                    "workloadType": "Pod"
                  }
                ],
                "meters": [
                  {
                    "aggregation": "sum",
                    "period": "1h",
                    "metricId": "container_cpu_usage_core_avg",
                    "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=~\"coordinator|worker\"}[5m])*100",
                    "workloadType": "Pod"
                  }
                ]
              }
            }
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Starburst Data
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: StarburstEnterprise
            name: starburstenterprises.charts.starburstdata.com
            version: v1alpha1
          - kind: StarburstHive
            name: starbursthives.charts.starburstdata.com
            version: v1alpha1
        description: CRs for deploying starburst enterprise and hive
        displayName: Starburst Enterprise Pay as you go Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - starburst
        - trino
        - sep
        - bigdata
        - data
        links:
        - name: Starburst Enterprise Helm Operator
          url: https://www.starburstdata.com
        maintainers:
        - email: hello@starburstdata.com
          name: Starburst Data, Inc.
        maturity: alpha
        provider:
          name: Starburst Data, Inc.
          url: https://www.starburstdata.com
        relatedImages:
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
        version: 402.1.0
      entries:
      - name: starburst-enterprise-helm-operator-paygo.v402.1.0
        version: 402.1.0
      name: alpha
    defaultChannel: alpha
    packageName: starburst-enterprise-helm-operator-paygo-rhmp
    provider:
      name: Starburst Data, Inc.
      url: https://www.starburstdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: redhat-oadp-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oadp-operator.v1.0.19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "restic": {
                      "enable": true
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws",
                        "kubevirt"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ResticRepository",
                "metadata": {
                  "name": "resticrepository",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/oadp/oadp-rhel8-operator@sha256:caf87a9832c153833feb34affc2b797ad0e643362448742b0405d457e5f0ba41
          createdAt: "2025-04-14T13:25:43Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <1.0.19'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/oadp/oadp-mustgather-rhel8@sha256:af4c6af48f38510c7d52f678798a9affea2211dc488ed18f83a5d005b7ef4b6b
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is a Velero resource that represents the capture of
              Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Data Protection Applications on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://github.com/openshift/oadp-operator
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/oadp/oadp-velero-plugin-for-microsoft-azure-rhel8@sha256:3e81b5e55f96504ba76d9d47bbeec50c7b1fd90b7702fc31b1dba248050ce9b4
        - registry.redhat.io/oadp/oadp-velero-plugin-for-gcp-rhel8@sha256:05a164c27da94b374ce03451c851a607a3cd436b73b9e5b70723c6dc8cacb3d8
        - registry.redhat.io/oadp/oadp-mustgather-rhel8@sha256:af4c6af48f38510c7d52f678798a9affea2211dc488ed18f83a5d005b7ef4b6b
        - registry.redhat.io/oadp/oadp-rhel8-operator@sha256:caf87a9832c153833feb34affc2b797ad0e643362448742b0405d457e5f0ba41
        - registry.redhat.io/oadp/oadp-registry-rhel8@sha256:0886b3b4ce0295080bf2c5ee95f2b7b15a4483bcd94cb739e7e7154d21150397
        - registry.redhat.io/oadp/oadp-velero-restic-restore-helper-rhel8@sha256:5d288335b5ad114b9eb61410befb4a9d1f84f0ba0c01558098bf4567fda9a8c5
        - registry.redhat.io/oadp/oadp-velero-plugin-rhel8@sha256:694603b9c1fd5eec939b28f8727fd7c9c0d380620c5081006f2655e98358464f
        - registry.redhat.io/oadp/oadp-velero-plugin-for-csi-rhel8@sha256:05ec16bdf1f0fadc4ddad244b86934f73569d2e66cc19f03b3553e257441a662
        - registry.redhat.io/oadp/oadp-velero-rhel8@sha256:39dec8b9ccdcabeaab3e56e14b79cfed4d8574cb172a886d017813a437e4ffff
        - registry.redhat.io/oadp/oadp-velero-plugin-for-aws-rhel8@sha256:28045538fff9e5051cd01e56876971cf9401a448e1dd63bff17fbb900629f22f
        - registry.redhat.io/oadp/oadp-kubevirt-velero-plugin-rhel8@sha256:d862a1b0420fcfb91403f90047b00035efba112af6bc5c50adb723014a620767
        version: 1.0.19
      entries:
      - name: oadp-operator.v1.0.19
        version: 1.0.19
      - name: oadp-operator.v1.0.18
        version: 1.0.18
      - name: oadp-operator.v1.0.17
        version: 1.0.17
      - name: oadp-operator.v1.0.16
        version: 1.0.16
      - name: oadp-operator.v1.0.15
        version: 1.0.15
      - name: oadp-operator.v1.0.14
        version: 1.0.14
      - name: oadp-operator.v1.0.13
        version: 1.0.13
      - name: oadp-operator.v1.0.12
        version: 1.0.12
      - name: oadp-operator.v1.0.11
        version: 1.0.11
      - name: oadp-operator.v1.0.10
        version: 1.0.10
      - name: oadp-operator.v1.0.9
        version: 1.0.9
      - name: oadp-operator.v1.0.8
        version: 1.0.8
      - name: oadp-operator.v1.0.7
        version: 1.0.7
      - name: oadp-operator.v1.0.6
        version: 1.0.6
      - name: oadp-operator.v1.0.5
        version: 1.0.5
      - name: oadp-operator.v1.0.4
        version: 1.0.4
      - name: oadp-operator.v1.0.3
        version: 1.0.3
      - name: oadp-operator.v1.0.2
        version: 1.0.2
      - name: oadp-operator.v1.0.1
        version: 1.0.1
      - name: oadp-operator.v1.0.0
        version: 1.0.0
      name: stable-1.0
    - currentCSV: oadp-operator.v1.4.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "nodeAgent": {
                      "enable": true,
                      "uploaderType": "restic"
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws",
                        "kubevirt"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: |
            Cloud Provider,Developer Tools,Modernization & Migration,OpenShift Optional,Storage
          certified: "false"
          containerImage: registry.redhat.io/oadp/oadp-rhel9-operator@sha256:ee010fb34448b772690a8a98808aa25aa1578a81061618f948b1d16b5d63d333
          createdAt: "2025-07-10T04:59:11Z"
          description: OADP (OpenShift API for Data Protection) operator sets up and
            installs Data Protection Applications on the OpenShift platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <1.4.5'
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/oadp/oadp-mustgather-rhel9@sha256:0cee6fecf052bae3ede6bb6f032b25c0c7cd63ced1e26328f2b02144f24fbb12
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A backup repository is an indicator of a connection from
              the restic/kopia server to the backupstoragelocation.
            displayName: BackupRepository
            kind: BackupRepository
            name: backuprepositories.velero.io
            version: v1
          - description: Backup is a Velero resource that represents the capture of
              Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - description: The CloudStorage API automates the creation of a bucket for
              object storage.
            displayName: CloudStorage
            kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataDownload represents a data download of a volume snapshot.
              There is one DataDownload created per volume to be restored.
            displayName: DataDownload
            kind: DataDownload
            name: datadownloads.velero.io
            version: v2alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DataUpload acts as the protocol between data mover plugins
              and data mover controller for the datamover backup operation
            displayName: DataUpload
            kind: DataUpload
            name: datauploads.velero.io
            version: v2alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: |
          **OpenShift API for Data Protection (OADP)** operator sets up and installs
          Velero on the OpenShift platform, allowing users to backup and restore
          applications.

          Backup and restore Kubernetes resources and internal images, at the
          granularity of a namespace, using a version of Velero appropriate for
          the installed version of OADP.

          OADP backs up Kubernetes objects and internal images by saving them as
          an archive file on object storage. OADP backs up persistent volumes
          (PVs) by creating snapshots with the native cloud snapshot API or with
          the Container Storage Interface (CSI). For cloud providers that do not
          support snapshots, OADP backs up resources and PV data with Restic or Kopia.

          * [Installing OADP for application backup and restore][install-doc]

          * [Installing OADP on a ROSA cluster and using STS, please follow the Getting Started Steps 1-3 in order to obtain the role ARN needed for using the standardized STS configuration flow via OLM][rosa-sts-doc]

          * [Frequently Asked Questions][faq-doc]

          [install-doc]: https://docs.openshift.com/container-platform/latest/backup_and_restore/application_backup_and_restore/installing/about-installing-oadp.html
          [rosa-sts-doc]: https://mobb.ninja/docs/misc/oadp/rosa-sts/
          [faq-doc]: https://access.redhat.com/articles/5456281
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - disaster backup
        - disaster recovery
        - disaster restore
        - disaster
        - backup
        - restore
        - backup automation
        - recovery
        - data protection
        - data management
        - data
        - protection
        - management
        - application resilience
        - resilience
        - velero
        - openshift
        - oadp
        - cloud-native
        - replication
        - kopia
        - restic
        - network file system
        - nfs
        - simple cloud storage
        - s3
        - minio
        - cloud storage
        - data foundation
        - ceph
        - csi
        - container storage interface
        links:
        - name: OADP Operator
          url: https://github.com/openshift/oadp-operator
        maintainers:
        - email: oadp-maintainers@redhat.com
          name: OADP Operator maintainers
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/oadp/oadp-velero-plugin-for-legacy-aws-rhel9@sha256:ab5c8d33684a361fa3bb10a006ef489be92cc955ad3d173bbfa77949dcc42aa8
        - registry.redhat.io/oadp/oadp-kubevirt-velero-plugin-rhel9@sha256:c3b32cb9f5b82021c08cd70467e68aad66746ef211d73612ecfe0af625706267
        - registry.redhat.io/oadp/oadp-velero-plugin-for-aws-rhel9@sha256:79d39186cae5696949813a0ea6c46eeededad29e6bde171d7a9b0011c993bb26
        - registry.redhat.io/oadp/oadp-velero-plugin-for-microsoft-azure-rhel9@sha256:9f05592bf0e5b76322ca582bdede387f9f4aabbbdc436549156c05d652c951d0
        - registry.redhat.io/oadp/oadp-mustgather-rhel9@sha256:0cee6fecf052bae3ede6bb6f032b25c0c7cd63ced1e26328f2b02144f24fbb12
        - registry.redhat.io/oadp/oadp-rhel9-operator@sha256:ee010fb34448b772690a8a98808aa25aa1578a81061618f948b1d16b5d63d333
        - registry.redhat.io/oadp/oadp-velero-rhel9@sha256:17160f99940aa355dc91ab3d16ba6bf91faade6423434008cf8f6e5023b1e413
        - registry.redhat.io/oadp/oadp-velero-plugin-for-gcp-rhel9@sha256:00800ead11ea308794b1536d964f8a043e13ede9657cabb23afa42cba7cf2ce6
        - registry.redhat.io/oadp/oadp-velero-plugin-rhel9@sha256:3ee9d3e799cedc28c400d364b3a8790c7e88acd5f603b09ff512a22d91e265b3
        - registry.redhat.io/oadp/oadp-velero-restic-restore-helper-rhel9@sha256:e0647defc7c428f9a04a5f6d4532dc86e57b92f31b4f907253ae14af05c5993f
        version: 1.4.5
      entries:
      - name: oadp-operator.v1.4.5
        version: 1.4.5
      - name: oadp-operator.v1.4.4
        version: 1.4.4
      - name: oadp-operator.v1.4.3
        version: 1.4.3
      - name: oadp-operator.v1.4.2
        version: 1.4.2
      - name: oadp-operator.v1.4.1
        version: 1.4.1
      - name: oadp-operator.v1.4.0
        version: 1.4.0
      name: stable-1.4
    defaultChannel: stable-1.4
    packageName: redhat-oadp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Oracle
      provider-url: https://www.oracle.com/cloud/
    name: oci-ccm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oci-ccm-operator.v0.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ccm.oraclecloud.com/v1alpha1",
                "kind": "OCICloudManager",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "oci-ccm-operator",
                    "app.kubernetes.io/instance": "ocicloudmanager",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ocicloudmanager",
                    "app.kubernetes.io/part-of": "oci-ccm-operator"
                  },
                  "name": "ocicloudmanager"
                },
                "spec": {
                  "compartment": "CompartmentId",
                  "driverNamespace": "kube-system",
                  "loadBalancer": {
                    "securityList": "SecurityListId",
                    "securityListManagementMode": "Frontend",
                    "subnet": "SubnetId"
                  },
                  "vcn": "VcnId"
                }
              },
              {
                "apiVersion": "ccm.oraclecloud.com/v1alpha1",
                "kind": "OCICsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "oci-ccm-operator",
                    "app.kubernetes.io/instance": "ocicsidriver",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ocicsidriver",
                    "app.kubernetes.io/part-of": "oci-ccm-operator"
                  },
                  "name": "ocicsidriver"
                },
                "spec": {
                  "compartment": "CompartmentId",
                  "driverNamespace": "kube-system",
                  "vcn": "VcnId"
                }
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider, Storage
          containerImage: phx.ocir.io/ax1eoqvpvbtg/oci-ccm-operator:0.0.9
          createdAt: "2023-07-28T23:42:18Z"
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Oracle
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: OCICloudManager
            name: ocicloudmanagers.ccm.oraclecloud.com
            version: v1alpha1
          - kind: OCICsiDriver
            name: ocicsidrivers.ccm.oraclecloud.com
            version: v1alpha1
        description: OCI Kubernetes Cloud Controller Manager implementation (or out-of-tree
          cloud-provider) for Oracle Cloud Infrastructure (OCI).
        displayName: oci-ccm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OCI
        - CCM
        - OpenShift
        - Kubernetes
        links:
        - name: Oracle
          url: https://www.oracle.com/cloud/
        - name: OCI Cloud Controller Manager (CCM)
          url: https://github.com/oracle/oci-cloud-controller-manager
        maintainers:
        - email: david.d.campbell@oracle.com
          name: David Campbell
        - email: alex.blinov@oracle.com
          name: Alex Blinov
        maturity: alpha
        provider:
          name: Oracle
          url: https://www.oracle.com/cloud/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - phx.ocir.io/ax1eoqvpvbtg/oci-ccm-operator:0.0.9
        version: 0.0.9
      entries:
      - name: oci-ccm-operator.v0.0.9
        version: 0.0.9
      name: alpha
    defaultChannel: alpha
    packageName: oci-ccm-operator
    provider:
      name: Oracle
      url: https://www.oracle.com/cloud/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: eventing-kogito
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eventing-kogito.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kogito.knative.dev/v1alpha1",
                "kind": "KogitoSource",
                "metadata": {
                  "name": "kogito-order-processing"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "kind": "Broker",
                      "name": "default"
                    }
                  },
                  "subject": {
                    "apiVersion": "serving.knative.dev/v1",
                    "kind": "Service",
                    "name": "serverless-workflow-order-processing"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: gcr.io/knative-releases/knative.dev/eventing-kogito/cmd/controller:v1.2.0
          createdAt: "2021-08-25T12:00:00Z"
          olm.targetNamespaces: knative-kogito
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/knative-sandbox/eventing-kogito
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoSource is the representation of a Knative Eventing
              Source for a Kogito custom Service.
            displayName: Kogito Source
            kind: KogitoSource
            name: kogitosources.kogito.knative.dev
            version: v1alpha1
        description: |-
          Knative Eventing Source for Kogito Services. This controller will enable easy configuration of any Kogito custom Service
          with the Knative Eventing platform. Please find more information about using this source in the [Official Kogito Documentation](https://docs.jboss.org/kogito/release/latest/html_single/#_kogito_knative_eventing_add_on)

          **Note:** The admission webhooks are not installed via OLM at this time. Because of that, please ensure that you create the `KogitoSource` objects with the subbject and sink information with the appropriate namespace.
        displayName: Kogito Knative Eventing Source
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless workflow
        - kogito
        - knative
        - dmn
        - bpmn
        - eventing
        - cloudevents
        - messaging
        links:
        - name: Source Repository
          url: https://github.com/knative-sandbox/eventing-kogito
        - name: Kogito Knative Add-On Documentation
          url: https://docs.jboss.org/kogito/release/latest/html_single/#_kogito_knative_eventing_add_on
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - gcr.io/knative-releases/knative.dev/eventing-kogito/cmd/controller:v1.2.0
        version: 1.2.0
      entries:
      - name: eventing-kogito.v1.2.0
        version: 1.2.0
      - name: eventing-kogito.v1.1.0
        version: 1.1.0
      name: alpha
    defaultChannel: alpha
    packageName: eventing-kogito
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KEDA Community
      provider-url: ""
    name: keda
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keda.v2.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "KedaController",
                "metadata": {
                  "name": "keda",
                  "namespace": "keda"
                },
                "spec": {
                  "watchNamespace": "",
                  "operator": {
                    "logLevel": "info",
                    "logEncoder": "console"
                  },
                  "metricsServer": {
                    "logLevel": "0"
                  },
                  "serviceAccount": {}
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledJob",
                "metadata": {
                  "name": "scaledjob-sample"
                },
                "spec": {
                  "failedJobsHistoryLimit": 5,
                  "jobTargetRef": {
                    "activeDeadlineSeconds": 600,
                    "backoffLimit": 6,
                    "completions": 1,
                    "parallelism": 1,
                    "template": null
                  },
                  "maxReplicaCount": 100,
                  "pollingInterval": 30,
                  "successfulJobsHistoryLimit": 5,
                  "triggers": [
                    {
                      "metadata": {
                        "property": "examle-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledObject",
                "metadata": {
                  "name": "example-scaledobject"
                },
                "spec": {
                  "cooldownPeriod": 300,
                  "maxReplicaCount": 100,
                  "minReplicaCount": 0,
                  "pollingInterval": 30,
                  "scaleTargetRef": {
                    "name": "example-deployment"
                  },
                  "triggers": [
                    {
                      "metadata": {
                        "property": "examle-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "TriggerAuthentication",
                "metadata": {
                  "name": "example-triggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "exmaple-secret-parameter"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ClusterTriggerAuthentication",
                "metadata": {
                  "name": "example-clustertriggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "exmaple-secret-parameter"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Cloud Provider
          certified: "false"
          containerImage: ghcr.io/kedacore/keda-olm-operator:2.7.1
          createdAt: "2022-05-11T00:00:00.000Z"
          description: '[DEPRECATED] Use Custom Metrics Autoscaler Operator instead'
          operatorframework.io/suggested-namespace: keda
          operators.operatorframework.io/builder: operator-sdk-v1.1.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/kedacore/keda-olm-operator
          support: KEDA Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              Represents an installation of a particular version of KEDA Controller.
            displayName: KedaController
            kind: KedaController
            name: kedacontrollers.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your jobs and what the triggers
            displayName: ScaledJob
            kind: ScaledJob
            name: scaledjobs.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your application and what the
              triggers are
            displayName: ScaledObject
            kind: ScaledObject
            name: scaledobjects.keda.sh
            version: v1alpha1
          - description: Describes authentication parameters
            displayName: TriggerAuthentication
            kind: TriggerAuthentication
            name: triggerauthentications.keda.sh
            version: v1alpha1
          - description: Describes cluster wide authentication parameters
            displayName: ClusterTriggerAuthentication
            kind: ClusterTriggerAuthentication
            name: clustertriggerauthentications.keda.sh
            version: v1alpha1
        description: "## Deprecated\n This KEDA community operator is deprecated,
          please use `Custom Metrics Autoscaler` Operator instead.\n## About the managed
          application\nKEDA is a Kubernetes-based event driven autoscaler.  KEDA can
          monitor event sources like Kafka, RabbitMQ, or cloud event sources and feed
          the metrics from those sources into the Kubernetes horizontal pod autoscaler.
          \ With KEDA, you can have event driven and serverless scale of deployments
          within any Kubernetes cluster.\n## About this Operator\nThe KEDA Operator
          deploys and manages installation of KEDA Controller in the cluster. Install
          this operator and follow installation instructions on how to install KEDA
          in you cluster.\n\n## Prerequisites for enabling this Operator\nPlease note
          that you can not run both KEDA 1.x and 2.x on the same Kubernetes cluster.
          You need to uninstall KEDA 1.x first, in order to install and use KEDA 2.x.\n\nDon't
          forget to uninstall KEDA v1 CRDs as well, to ensure that, please run:\n\n```
          oc delete crd scaledobjects.keda.k8s.io ```\n\n``` oc delete crd triggerauthentications.keda.k8s.io
          \ ```\n## How to install KEDA in the cluster\nThe installation of KEDA is
          triggered by the creation of `KedaController` resource, please refer to
          [KedaController Example](https://github.com/kedacore/keda-olm-operator#kedacontroller-spec)
          for more deatils on available options.\n\nOnly resource named `keda` in
          namespace `keda` will trigger the installation, reconfiguration or removal
          of the KEDA Controller resource.\n\nThere could be only one KEDA Controller
          in the cluster. \n"
        displayName: '[DEPRECATED] KEDA'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keda
        - scaling
        - event
        links:
        - name: Homepage
          url: https://keda.sh
        - name: Source Code
          url: https://github.com/kedacore/keda
        - name: Slack Channel
          url: https://kubernetes.slack.com/messages/CKZJ36A5D
        maintainers:
        - email: cncf-keda-maintainers@lists.cncf.io
          name: KEDA Community
        maturity: deprecated
        minKubeVersion: 1.17.0
        provider:
          name: KEDA Community
        relatedImages:
        - ghcr.io/kedacore/keda-olm-operator:2.7.1
        version: 2.7.1
      entries:
      - name: keda.v2.7.1
        version: 2.7.1
      - name: keda.v2.6.1
        version: 2.6.1
      - name: keda.v2.6.0
        version: 2.6.0
      - name: keda.v2.5.0
        version: 2.5.0
      - name: keda.v2.4.0
        version: 2.4.0
      - name: keda.v2.3.0
        version: 2.3.0
      - name: keda.v2.2.0
        version: 2.2.0
      - name: keda.v2.1.0
        version: 2.1.0
      - name: keda.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: keda
    provider:
      name: KEDA Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: stolostron
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stolostron.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:3443da607b276d39ba28a8d6561d296f30e9bac5ab06d5c70715bddbde0a58b2
          createdAt: "2024-06-28T04:52:57Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/volsync-addon-controller@sha256:7774767ed20263bba4857a0d4be012e12215f640fce01fbd3a085bf860d087e7
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:9d1d5e2b5313b898188d2c845dcda5d604b73f16e89c71915ea308281c1f56ff
        - quay.io/stolostron/helloworld@sha256:81a0b06c7fcf915b1bb40d33d29d0bfd4d501404463c1f46bffc503e7d1cc45b
        - quay.io/stolostron/volsync@sha256:ba7399b316359118b1c1b915d667fca9a4a0d12a10beae69e6dee9c7236753be
        - quay.io/stolostron/grafana@sha256:e40d66da749108ae1eecf909c95c87bf7a289328e8b373ae4402368c474f2da5
        - quay.io/stolostron/must-gather@sha256:1f7c129859bdaf6bd93dda77aec01f7cd03000fbfaccd809f26d33b9e3f66e51
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/config-policy-controller@sha256:197d8491e7c889db8f2acb726e0146d384c979acc7f634c24a9d4b21f7d444f3
        - quay.io/stolostron/multicluster-operators-channel@sha256:d791aa2c6032c1de50cf4033c426646613b182e98c406a091b88fb9ffefc0c98
        - quay.io/stolostron/volsync-mover-rsync@sha256:f7affcc16ae9bf6d12f69511b7a269f5fd065d1d87392c4c38cb5d7af5b89e23
        - quay.io/stolostron/search-e2e@sha256:8dfbc899ba1ffaa8fb6eb2861f6db87a13d2565cf4551fd47df7fe8253217fb2
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/search-collector@sha256:136ad44ff41a29dec48e87d10af5b608f335e4242c41ac17cba2c3c4c8205bd6
        - quay.io/stolostron/observatorium-operator@sha256:5da7f478ecf3609a3b86eb563e4dd22588f3d13efdc2e33f5b672fe95583ce68
        - quay.io/stolostron/multicluster-operators-application@sha256:199b696dbd7589b77e2391b71504fbcb7d3ae785972dac3d001c4c1bffa342dd
        - quay.io/stolostron/multicluster-operators-subscription@sha256:8350df0a7d4f19c8e52eb58eecf5223bb2a0ccbe4d4aa00fcf8dff832cb43e5d
        - quay.io/stolostron/insights-client@sha256:aac5b9b66800c0b862e8bfa1fea533a4aec854d8c52baddbfa03977e1ed520cb
        - quay.io/stolostron/search-indexer@sha256:068c38489201316dc12d475de03c8abb11575338bd75e25c1262eca8aac49cb2
        - quay.io/stolostron/rbac-query-proxy@sha256:24aafcba20ded0892209cfcc196bff88f877ed51eb928d5d350b17041d170893
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/governance-policy-framework-addon@sha256:a1c0d40561befaf5e849060beaee4b201f42d77c5090823f23ef43a75106b00b
        - quay.io/stolostron/grc-policy-framework-tests@sha256:f2b01192fd900860428942c7d5b900604ffcab0d42731676a1acdeebd3716557
        - quay.io/stolostron/volsync-mover-restic@sha256:284c3358f0f2158d599f27487d76a7fdb21842fc0e95bc639726dfdad14b7e7d
        - quay.io/stolostron/volsync-mover-rclone@sha256:b247b23944e20104f801c172cbbb620740a5b51d1777ecdbf98b4c128f1ef171
        - quay.io/stolostron/volsync-mover-syncthing@sha256:b5fad266ca8d3cd8efcbf4e8241c446004747b551b2b8e06acae10900d7a0f1b
        - quay.io/stolostron/cert-policy-controller@sha256:27673358610dc1ca308de472dbd794dcbb38250bb299b53c736549b1e8441a0f
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:7f6b6e45f0be0a14d7d9d79828572b9400f4d559f084028e3254cfd40948780d
        - quay.io/stolostron/klusterlet-addon-controller@sha256:01e5fca5e5c6caca6878bf8025471b7c77164ca7b4a2d764e20101fc79290df0
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/observability-e2e-test@sha256:8d9341b1814cd085a20f9feceab181a8989d6684fd44d6c8e54a0c6df4dd7216
        - quay.io/stolostron/observatorium@sha256:af4e315e4c6da3607f84e24f31e0f37f8572d161736b4748b2e57e0a2c2dd36c
        - quay.io/stolostron/multicluster-observability-operator@sha256:245e1815a73b5c48f790ae5a5d85fac6642cce8a5a3bf122275809e97a4d519e
        - quay.io/stolostron/multicloud-integrations@sha256:60be5b945ead8f845a29a8be18bd26922b83f009fde1412017f1ca985eb79b55
        - quay.io/stolostron/thanos-receive-controller@sha256:b1a191714cead24f66e8de7217db3c4d4a253ecf0238a912c8b88ba8b47c972d
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/prometheus-alertmanager@sha256:8222bbdd37c246703de54ed3dc109fa8c2f6ca17c5bb90106948683858afdc65
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/governance-policy-propagator@sha256:200558b5aa2366383bd298e159d9f7966a2c9b981038c79cde536edbfbe0092c
        - quay.io/stolostron/prometheus-operator@sha256:490b0056b9a23a1dc514a3036eac6593c7c8961de27d08ff24817d280325138c
        - quay.io/stolostron/submariner-addon@sha256:0f1b4ae59ff5c933542f9affed20bfa6694deb804f2e97f49ebd0f2e1b925eca
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:bf6bec399bc2072186e9d12ed865167fe940f3f619d69ffe6c178c14a3bdadbe
        - quay.io/stolostron/kube-rbac-proxy@sha256:9626cb9b2ee8f00d2abdb14cd2ce56d77435859071c10d8394bc7ee2754ce87f
        - quay.io/stolostron/cluster-permission@sha256:716d2cbcb7f69b5640d04669b15884e0adf053597f278a0bb1d8ee77e4a7af4d
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:02e53b8a6529e3209809dada4113054e3398de18738948cb01ed1d24dcb8e8be
        - quay.io/stolostron/search-v2-api@sha256:6b97b2c7e64de7cd154fd83da3bc690a57b67f01c83170238fe9c2e764df91a8
        - quay.io/stolostron/kube-state-metrics@sha256:ef1d671ffa098ed95186de6f15ea91fe648d103c1c7a57871ed46af07c6f7881
        - quay.io/stolostron/multiclusterhub-operator@sha256:3443da607b276d39ba28a8d6561d296f30e9bac5ab06d5c70715bddbde0a58b2
        - quay.io/stolostron/console@sha256:c7900403ef8dba3aac52f999532d876bbfe2f66f00a255047850594b64021830
        - quay.io/stolostron/prometheus@sha256:8250af047c03c6b0c07b76420e0a5c2338f053d73b2dca1ecba4bb2f211715fe
        - quay.io/stolostron/governance-policy-addon-controller@sha256:593dc0a689091ce7c916e6be764c222c1020629ccff7eafbf17524936e4f8262
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/thanos@sha256:cc78cd786643d27ed85eb3670852722e915a20b20a21479e1477fb859dec6c5e
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/prometheus-config-reloader@sha256:bc93ca9f1b6d95e8b963fe1af4c17f87c8c7d938a348627c29cc3cf42b70ca3e
        - quay.io/stolostron/search-v2-operator@sha256:c41517f372fc1874895b39e70cd1c07391110697a5e49ebcf56f4a5df499d810
        - quay.io/stolostron/grafana-dashboard-loader@sha256:cff97fe3e32dd14ed2f3b7b02cb205715cfee6cfd6dc7e7a630101ea5147e566
        - quay.io/stolostron/cluster-backup-controller@sha256:9c77444bd42e6e99163f4dce9f9975276aea082100f53fd938cddf74ad3b1574
        - quay.io/stolostron/insights-metrics@sha256:49a177357e865d5f8347840fe695531d539d17e28d2b9971821e41ebbee014bd
        - quay.io/stolostron/metrics-collector@sha256:4101a2f2c4230bdc1420a58b546a241dfc5414cf92b57f3b73619c1be14dbc32
        version: 0.4.0
      entries:
      - name: stolostron.v0.4.0
        version: 0.4.0
      name: community-0.4
    - currentCSV: stolostron.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:781f6712369cd4b9bdc8ba0ee2f35a12db198ca190621e199dc10cfea3f5efc9
          createdAt: "2024-10-02T08:01:20Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.5.0 <0.5.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/multicluster-operators-application@sha256:329c82b92fe2448fe26923fbd82d2eadacb7f96c75689d96393c573f95e6d650
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/observatorium-operator@sha256:38f22e30db7713e9f2e239463803cdb5cdfe7108a524ffa5d7e95201bae5300b
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/grc-policy-framework-tests@sha256:57794e2b15949cedf3462df2d38b95e86ca1e140a5ae81e4a9b3e7e06072847d
        - quay.io/stolostron/cert-policy-controller@sha256:809cae1aef02dd3fb68143ebdba818b20c661bb628a7922ee1d705efe4947f35
        - quay.io/stolostron/multicluster-operators-subscription@sha256:30b1b5bf6a7a12e525450905a474babdd7e38235d9be60f3cb04d60a745a1b96
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/governance-policy-framework-addon@sha256:5f2b326a279b7c42c233ea5f11b89ecaf4cfd7c549d021ddcc4e7b0c8926fc6d
        - quay.io/stolostron/search-collector@sha256:4082f7a37aa3f10c725bb9bebf94ee8dab97d22302cd5e59b35d0e3bcc888cef
        - quay.io/stolostron/volsync-mover-restic@sha256:284c3358f0f2158d599f27487d76a7fdb21842fc0e95bc639726dfdad14b7e7d
        - quay.io/stolostron/prometheus-alertmanager@sha256:19f2e277cbdd2bd4814b8d8034167c1a7cfe9f1d698fb16f7604fa65cb8bcb1c
        - quay.io/stolostron/observability-e2e-test@sha256:c361f5a0f806498c6c1150c0c0d5a42222efff13a18cb28dbfd5e4bd835cb4b7
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/prometheus@sha256:147f3a0cee8fa7b11c82e21ca95c28278e1ab53ccb5b70f3af0425bb8a74f61b
        - quay.io/stolostron/metrics-collector@sha256:c580106b2c7173251afa6714b710fd109811023d88a8009fb741f3d00e486e76
        - quay.io/stolostron/siteconfig-operator@sha256:7f47f0c61d5b53411ed78046c5a02664c6693d940ef37d6db1853670196be334
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:a8d812a527902bd4d9dfd23f90ce90f415f5a6fabb007d1c41b4c998e1ca098a
        - quay.io/stolostron/must-gather@sha256:362bbfd54059975369ce6d3869d34e346a0e76159163714cfcb3597022b7c88f
        - quay.io/stolostron/kube-state-metrics@sha256:dc534373feda042afa385026c7da674912ca310df2a4598d83451d4474bb3f44
        - quay.io/stolostron/volsync-addon-controller@sha256:d0b15da92e0ff9792d750b7b48d72b38d6abb91488c20578c7be3382c561877f
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/multicluster-operators-channel@sha256:260a11aae954299ce8fbb0f9c3102afd3b5cf3f1b972b5a71e56c273ba1d532f
        - quay.io/stolostron/volsync-mover-rclone@sha256:b247b23944e20104f801c172cbbb620740a5b51d1777ecdbf98b4c128f1ef171
        - quay.io/stolostron/volsync@sha256:ba7399b316359118b1c1b915d667fca9a4a0d12a10beae69e6dee9c7236753be
        - quay.io/stolostron/klusterlet-addon-controller@sha256:cf4cad3599c4ce538dc007cda9f5b06538799e8198912d907058ecdd4bbb2451
        - quay.io/stolostron/volsync-mover-syncthing@sha256:b5fad266ca8d3cd8efcbf4e8241c446004747b551b2b8e06acae10900d7a0f1b
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:d42ddd30570ccdf388ccd65cd3f7c72de2d857d328e05aef8e855508ccf91898
        - quay.io/stolostron/search-e2e@sha256:ddaa50ecfafcd7501903420a47495f733dce36fcb91c8ce9b8d94f4448ed501d
        - quay.io/stolostron/console@sha256:e42d2f7e39a93f13003193142a3d1f3f9e55966cbf55b94eac93c726373a2516
        - quay.io/stolostron/prometheus-config-reloader@sha256:337f7d2036c17459a41884e38c569a3601364b02660946b49d39ce40ef97bd48
        - quay.io/stolostron/prometheus-operator@sha256:5ff1f7a4be0216f582f5cb40de3412cbe8126433144c8201750f6c3c0fee920e
        - quay.io/stolostron/grafana@sha256:c9869c7f590bc4c54151c7e4fa9ed76dfe35c24a0ea3c19a6a1b26e61924b6e1
        - quay.io/stolostron/cluster-permission@sha256:716d2cbcb7f69b5640d04669b15884e0adf053597f278a0bb1d8ee77e4a7af4d
        - quay.io/stolostron/governance-policy-addon-controller@sha256:9f387eb8b54d1c86fc20d205dbce199738849cb0ce15a47052ddbee8781d2d9b
        - quay.io/stolostron/search-v2-operator@sha256:6df11a36ea19a1b3da191964f851dee95b87fd993e8f0ef8906b1e63dd450d17
        - quay.io/stolostron/thanos@sha256:dd4d4f8fbb9a8f51134260fc728d19f61a93b511fd7889636089944ac46163bd
        - quay.io/stolostron/search-indexer@sha256:7471f2808686d42c2eae9f0a537f9b520648f31e87072a376243f309fd044caa
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/insights-client@sha256:aac5b9b66800c0b862e8bfa1fea533a4aec854d8c52baddbfa03977e1ed520cb
        - quay.io/stolostron/acm-cli@sha256:85a5c55dfa4708435807d1266b5607525e33f55c5b4672fd8ba78987a1c48ef3
        - quay.io/stolostron/multicluster-observability-operator@sha256:fb0519952ef6436cc66496409f00e89cd6e58a41776df96859a86176a7322d05
        - quay.io/stolostron/rbac-query-proxy@sha256:2f1c78176d0ca0b21c237a9f7a3f77f7d857dd848f3a42254cee1e4e469b4c3c
        - quay.io/stolostron/governance-policy-propagator@sha256:69ab8f492fa418596db6da578b3b579fc2f1506c5bf780e6aba3ccd05d8b02b9
        - quay.io/stolostron/cluster-backup-controller@sha256:688844c722db78f383d90986150bde93bb421068e7f5b4128d19043f3e01f93e
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/observatorium@sha256:8d0f02996b5bf02be02ff3743821909b6c2396f0f0519652dde0c5d1d76f8a05
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/kube-rbac-proxy@sha256:1b708bff7fb46391f2460d34be28d30add38501cdb084306df01faae75637232
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/submariner-addon@sha256:07287987fb0041efc16557cb9d7403dc15b5959a5d0dd0d81cd72bc2658cba10
        - quay.io/stolostron/config-policy-controller@sha256:5af6a0e17da7b7261ac71bcb1e2d991856bbb5450686d188026798e05f0617ba
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/volsync-mover-rsync@sha256:f7affcc16ae9bf6d12f69511b7a269f5fd065d1d87392c4c38cb5d7af5b89e23
        - quay.io/stolostron/thanos-receive-controller@sha256:58c97fe5ec2131f9dcfb70563b59ed0bcae94d8bddefa51cb4784e73b1d90e4f
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/insights-metrics@sha256:49a177357e865d5f8347840fe695531d539d17e28d2b9971821e41ebbee014bd
        - quay.io/stolostron/search-v2-api@sha256:338634d767383716c7a88ed2bc9c58275e9d5953932293443725883d7efcffd5
        - quay.io/stolostron/grafana-dashboard-loader@sha256:85031d1a56514be239c585487aec05a8cbb08c17aeb60c0c84c685eb9d9f08b5
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:dbdbec1600d57c440173aecf19c57709d69e945dcecb348bcf4a7c37b3700b70
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:358ac0b0a3afdfac06f784f2e02a305cbef4e6f648e31a00981d3b097746412a
        - quay.io/stolostron/helloworld@sha256:48724b53542a3fcaf342d6ae2056bdf846cbe2ee50879613953b045bfcbe1c0a
        - quay.io/stolostron/multicloud-integrations@sha256:d4cb3a0a50bc5d22f48c96ed3a10ea9843c6ef2f83daeae411a7ddb7ea332ea6
        - quay.io/stolostron/multiclusterhub-operator@sha256:781f6712369cd4b9bdc8ba0ee2f35a12db198ca190621e199dc10cfea3f5efc9
        version: 0.5.3
      entries:
      - name: stolostron.v0.5.3
        version: 0.5.3
      - name: stolostron.v0.5.2
        version: 0.5.2
      - name: stolostron.v0.5.1
        version: 0.5.1
      - name: stolostron.v0.5.0
        version: 0.5.0
      name: community-0.5
    - currentCSV: stolostron.v0.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:09f1f4266938fba88376a50c3e2837c5517b4f3a947cf86638a17e52eb17c1f6
          createdAt: "2025-02-24T01:42:20Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.6.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/prometheus-config-reloader@sha256:2b26d2fb5c69e54aecfbcaab46e652d136d726f0604844031c8771874f1fd769
        - quay.io/stolostron/thanos-receive-controller@sha256:d73f98e93a258adffc1a4bd875c79a9629166599e68ef777a790a1f3146a2929
        - quay.io/stolostron/observatorium-operator@sha256:5eb4c0da92c62b664aaf3bc187924870275e687440c3b0a3ad4f1c760c8b35e9
        - quay.io/stolostron/insights-client@sha256:978d8b7c3e692272024b71b8c005b752c08dc643dde8258ad75961ddcdcce67f
        - quay.io/stolostron/observability-e2e-test@sha256:03403aeb59b364f578761ae4bcb302bd3589baff500d733775e8cb4f2f961236
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/insights-metrics@sha256:f2f4cf14090252af32832ff6df8a9f896e5314751199156c4e5ae0755e4515e6
        - quay.io/stolostron/cluster-backup-controller@sha256:c85f88d42b42e4366f8ff7ad6c3e6b375e98c5b6a12a9c24a6333929e48f1811
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/helloworld@sha256:56f0d641d4f07b922f1c28c536eacad747c56a564e0a65f084106c2ffc9bd99d
        - quay.io/stolostron/multicluster-operators-channel@sha256:501a5053bd3fd5cc9de34e7e020890d731321b090a0e5bf11b8b3350d002f573
        - quay.io/stolostron/thanos@sha256:53eee303836070a7c7621bfb2a094e3809146497d1653dc65d16055aacea332c
        - quay.io/stolostron/flightctl-api@sha256:64fb1b655bef6aeff03cf64b6450155d3cb570ed4713f084590fe515f4fbad02
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/metrics-collector@sha256:c7433ee4ccdb56717802acf258530e2d6d237787d760a8fbcfd642790254d5ab
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/kube-state-metrics@sha256:cd24916e55ad905e386fa3df8a648272114353a0d75283224a4dddb1955b2687
        - quay.io/stolostron/multicluster-operators-application@sha256:59802fcfae5b6fd0bebe58e98412885e71f32464b8dbbf85364ddaf090aa3e00
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/multicluster-observability-operator@sha256:bed0dd30646d7fce894a625e6cf52e7829c272f47dde0baf1f788e9f85f66e42
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/search-v2-api@sha256:694523242aabb8db1f27b5afcd8ab2f63a9cdb6056c33a79badf8e68a337738d
        - quay.io/stolostron/config-policy-controller@sha256:8cf54eb6f5b49a60ed165fa4b70e4984a1362f2f31e8304c3a22e8a6a953a606
        - quay.io/stolostron/governance-policy-framework-addon@sha256:047124271d0b30176a94454b293c82c2eb67a6f06757fc22c54c455fd0913157
        - quay.io/stolostron/cert-policy-controller@sha256:98dd74d7375ca2b546075d5c8f06a9e2c4f2ce3b77247dccf36bcaa46bf564c7
        - quay.io/stolostron/flightctl-worker@sha256:4dddded7d2d4cce63972c8ca81f213c2cedb551cc645ec85495a9692dc36bb4c
        - quay.io/stolostron/multicluster-operators-subscription@sha256:a056ce7d438f68d33ca3dfd42ec3d9a34b347cdf05499066412784109726e7ad
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:92369b6deb21ab3cb0a12ae5c74b455cfefb8698826bca39e26648e608e48a2b
        - quay.io/stolostron/grafana-dashboard-loader@sha256:6c6c2aa528373536d5db66fffeb87a85b79f3cf723e5d9eeaefcb74f91c1a8a1
        - quay.io/stolostron/search-indexer@sha256:4f9be902fe976fd88a77b737cc1f56f00d48740a6a62661468328cf3bb9da31a
        - quay.io/stolostron/must-gather@sha256:2bf9d3c15f32a6bfe0a7a8d39b9dc62ac6cbb6fe47b7cf1f6bc5e30b17e5e849
        - quay.io/stolostron/flightctl-ocp-ui@sha256:6cff6425850d2b4a8b885a684afb01244d7cd9de7035d7ad71b1e10e5bab25e9
        - quay.io/stolostron/console@sha256:eed52f3ab860a3a224ce178d92765ca50d72d6261f04df040a39577881738252
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/governance-policy-propagator@sha256:c21574a2421fefe3fcabeef14640d6fe699d950106c9ed056b3ba9be3cbeabe0
        - quay.io/stolostron/observatorium@sha256:398ecee4d3f12e84e48b12009a4b02fc8b360dfc0c1607ffecc26e6bc04d0ea2
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/prometheus@sha256:3d1810567f67ae1a7bb024a1c4e494ee510cfb14c7ed21a8b5e63e41fa7bcd68
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/search-collector@sha256:54ece9741641e1bb6e30f15d97a65c92b9c6408ba2572a2f6658f537134adff1
        - quay.io/stolostron/rbac-query-proxy@sha256:096ad23c9b9c7e90451c09b48ded67a82575e75c8cc19f22c459d63ae8b5a786
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:05da11b343938866c5afcfe844bcd15d9c513fdb10fd7ea0d3473b455b427c2a
        - quay.io/stolostron/grc-policy-framework-tests@sha256:9b84dbafef35bd568b2bf79283b6850ba1ff57d70239a27cc0cc815fb1a55515
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/grafana@sha256:51092f6754106a90aff28cdd80c2f0447c85c1bbc72db15878d0b65b93e1ead6
        - quay.io/stolostron/prometheus-alertmanager@sha256:b98480f8ba6ffd19fcee3b412757ad34428e6feab7e4196bb13f2f940306e2c9
        - quay.io/stolostron/search-v2-operator@sha256:85e53a97240f6fd1c4251101146cc2bcc51eb91d9b171efb5b58edade9e0244e
        - quay.io/stolostron/volsync-addon-controller@sha256:d9dad837f2f735674e5711c7fff1cfe2c0c92160fceb7d1548a62348dd43b81c
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/prometheus-operator@sha256:f80e79ea87cde66955e1e46d264e414c072085da9db6876ee2ae68c87d2452e5
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:9c85fb91c4be3287cac2a376465836689df885902099096f0f76ffca50a6d311
        - quay.io/stolostron/siteconfig-operator@sha256:97dc128666780ec9a4c955f392ab98f81f0d6038ec2159e8b1d55aab393cf118
        - quay.io/stolostron/acm-cli@sha256:8c8c059ce2f6d131c410972bf8fd0b7e399cf17cd044c7ddc89a9668c0ceac4c
        - quay.io/stolostron/flightctl-ui@sha256:fea717f3b8b4c343ec60ce7d74a691f1124b6847d55e30a73dc0d50ea610d370
        - quay.io/stolostron/cluster-permission@sha256:443d8b30e3bb83c139c64ddfc38b0d5e6d7a25a4fae783303ab2f2c27d015b26
        - quay.io/stolostron/search-e2e@sha256:43e1bbfb7259612ffb408108c71b825f6e61a268c7e405b8d855f6878b6f96dd
        - quay.io/stolostron/multicloud-integrations@sha256:0081065f9e59c5ed41087344668ac86012f796cdbbf3575fd8491c7dfd24fbf8
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/klusterlet-addon-controller@sha256:39e6940eb956fe9640342e11ecbcbc7d6df0169e1b340a60afc19aff7dd713a5
        - quay.io/stolostron/flightctl-periodic@sha256:a3ed745dc65a5886c4cc05348cf4822bfbf2c6d28b85a39fa0b8f3e1f53fdf8a
        - quay.io/stolostron/submariner-addon@sha256:9c72b5177de8ddc56dd3544997e9bd222895545b7c578e43e4051596fb1fa873
        - quay.io/stolostron/multiclusterhub-operator@sha256:09f1f4266938fba88376a50c3e2837c5517b4f3a947cf86638a17e52eb17c1f6
        - quay.io/stolostron/kube-rbac-proxy@sha256:5b521e4f0e5e54c53cea714020069aa796da23d1482d14e4b7c908098ac69ae9
        - quay.io/stolostron/governance-policy-addon-controller@sha256:fa0738abc0368f9dadd8b0845d868512d26be3525a1a07b9de8a9cefafaa42b5
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:51a8acc47098175db148ebc19a10cdba2803d4a9f5884dd38b1c839702398789
        version: 0.6.3
      entries:
      - name: stolostron.v0.6.3
        version: 0.6.3
      - name: stolostron.v0.6.2
        version: 0.6.2
      - name: stolostron.v0.6.1
        version: 0.6.1
      - name: stolostron.v0.6.0
        version: 0.6.0
      name: community-0.6
    - currentCSV: stolostron.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:ab6660283a25bdf2ea9afe8c76bfbc32f4610d69af2bb13a7bca37a1f5961bdd
          createdAt: "2025-06-09T15:10:05Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.7.1'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MulticlusterHub defines the configuration
              for an instance of a multicluster hub, a central point for managing multiple
              Kubernetes-based clusters. The deployment of multicluster hub components
              is determined based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/governance-policy-framework-addon@sha256:b1b24c40d68415525e2bae21810fc439d3e96fcc5fa75c92e7f2eee284422bfd
        - quay.io/stolostron/thanos@sha256:9d504e2aa1001135eb6e454c58c620d102f2b0a85016b7587452dd1595794c93
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:4b29cc08c9e35054a4a820472ea19cd51e043ab1a90495418a36922df74b9540
        - quay.io/stolostron/insights-client@sha256:2cbd20252f10a5a64f93af50f117f05a865aa9d578e5f6e55e4249470b761604
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/multicluster-operators-application@sha256:9f804b4f2da910e0d7f122176f72c0b0ef4b30785fde0369a66f95f4bac58a82
        - quay.io/stolostron/multicluster-operators-subscription@sha256:d7c137379781ce6b697f73a464b89d04fbc20c3da00ce21a195ac79ab57a6dc7
        - quay.io/stolostron/cert-policy-controller@sha256:6deb41867f8781f4e34f69a54d4aeca190d588ba2991279d39cb5a82c2a5c5be
        - quay.io/stolostron/cluster-permission@sha256:e45139d459951b2e751018fe7ecabe1ff22b90ac87be81fb8f17a12f7b7f6a65
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/governance-policy-addon-controller@sha256:6ecc7557168177474423cea60eadc07d8382adb3e97dee669a8590859a41b08c
        - quay.io/stolostron/must-gather@sha256:ad8db72d559d06f4d467f78213d4f8f2cb00f8e76e942d8e75394eb7dadd29d2
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/multiclusterhub-operator@sha256:ab6660283a25bdf2ea9afe8c76bfbc32f4610d69af2bb13a7bca37a1f5961bdd
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/flightctl-periodic@sha256:a3ed745dc65a5886c4cc05348cf4822bfbf2c6d28b85a39fa0b8f3e1f53fdf8a
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/redis-7-c9s@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/kube-state-metrics@sha256:3503df61f3d9ebf3cbf66deef865450b32c7403839fdaca42504febaac00687c
        - quay.io/stolostron/volsync-addon-controller@sha256:224a0a03a25e5ce61f4c50f6252370876ac0cbf4f9711e997e98a00d6a124fe9
        - quay.io/stolostron/thanos-receive-controller@sha256:f445bbfc18bbc4646e0432842022d3be5305809e52ddcb9daf0ec000844d5854
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:54c6a2eec497eafdbf861fb95d78d961fb24906b197fc69b4e96d3e8b53e6496
        - quay.io/stolostron/search-v2-api@sha256:786866ca8c71a8263ddef45ca9bd9e48cfd1b3ad6215725fbf6f79967bda20ad
        - quay.io/stolostron/flightctl-worker@sha256:4dddded7d2d4cce63972c8ca81f213c2cedb551cc645ec85495a9692dc36bb4c
        - quay.io/stolostron/multicloud-integrations@sha256:665e9385531674a5849aa661eb207a0c6e39ea563fc891c57678cddbf7878d8e
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:1efd8e470b88a992e2f35ce18bdf4970f91af7ba917b8d92627b47ce8c10eb40
        - quay.io/stolostron/siteconfig-operator@sha256:16b00f6c4d81eea26e663ca5596526d85bf7777728d1f0ca7e4dd765c64d2bb5
        - quay.io/stolostron/console@sha256:ba89196efe30c812a767d58a41170b56e2b04167a72e7f13180bf40474bdc824
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/insights-metrics@sha256:861dc03fff55c1d9aff96b5ed15b9d38d5932df1c00a4423e35ead19da6eda03
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/stolostron/flightctl-api@sha256:64fb1b655bef6aeff03cf64b6450155d3cb570ed4713f084590fe515f4fbad02
        - quay.io/stolostron/origin-cli@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/prometheus-config-reloader@sha256:024ee6dda10c6d22ddeab57d7d6503b7aa9ac78981cfcec8ffae94b4914500e3
        - quay.io/stolostron/search-indexer@sha256:58320b54847e3bb6baa558bda65e9d0ae2198df480ba27f46f96ba1454614508
        - quay.io/stolostron/observability-e2e-test@sha256:4e60d3e3e1b5dbab3bcc1c4d37a4940a2a42a065a9df539fb243fb174dde6716
        - quay.io/stolostron/search-v2-operator@sha256:2ae9ab48615966ccf91b2022ccda12913a3db3cfbad5351dd8a20648b01137d8
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/search-e2e@sha256:1c7dcc9f3588620b3570a493c2d744de059f425b572b9cb5725142b89414a904
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:05da11b343938866c5afcfe844bcd15d9c513fdb10fd7ea0d3473b455b427c2a
        - quay.io/stolostron/submariner-addon@sha256:51359ccdc0ea29bcaf7ec2b190878e9f3333b1530cb1128e142d893fb87eba8b
        - quay.io/stolostron/grafana@sha256:cebaed7e5abe0064c65ea69bd4e14794d153de9172c4c3b6de4730f23fed7d24
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/rbac-query-proxy@sha256:1c198b6388a72ccfe2c5b68d57a2b22893fb65d977992ec44ee8cad6ee1ca976
        - quay.io/stolostron/multicluster-observability-operator@sha256:3d95297297a301fca1915268fec71d7a6ecef98c6c2ab8494d9befacb2e169fa
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/grc-policy-framework-tests@sha256:327122b13ec79ea488d6f424806549e36ebfbabd055f9b40217ba541f4a118a9
        - quay.io/stolostron/multicluster-operators-channel@sha256:318d885ab33d5afe17334e831c955f1c3cdf804281b7cc8523330820ac22eee2
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/observatorium-operator@sha256:2f11444703addec163a00c99529c228594fb52df13de47f08faac2e446bec483
        - quay.io/stolostron/prometheus-alertmanager@sha256:a9da6f01cfa61fef803f09ef3f6b27e33a07b97fc1f1e11b6a6c8e13b8a4b90e
        - quay.io/stolostron/klusterlet-addon-controller@sha256:4cb2f4f4e14cd67a4b2f9549cba714b635d9e1c2d8b869b0e0573fd7a41bc13e
        - quay.io/stolostron/prometheus-operator@sha256:8c1c12562ec00da8eaedb7d428694ebdd3a749c0b1ce03fe2b5663ae40e1c701
        - quay.io/stolostron/observatorium@sha256:18a06117ef114a1e68474fd9832adacd30b03f60ec1aa55374285a20e5311f13
        - quay.io/stolostron/governance-policy-propagator@sha256:2b77422fb03177f647c04716e7242bc08f241a1ef9c8968e492ae57816f4c4fb
        - quay.io/stolostron/config-policy-controller@sha256:b1718d5facb95c5572ac658c899f9fce0a7e8c08de93804de3327cd50cdb073b
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/grafana-dashboard-loader@sha256:6c6c2aa528373536d5db66fffeb87a85b79f3cf723e5d9eeaefcb74f91c1a8a1
        - quay.io/stolostron/prometheus@sha256:b959035ff754050ca092863bb74c35ba9622d9adc9e8eb7caf3b430cfe211730
        - quay.io/stolostron/acm-cli@sha256:2f0cec3bed4ad77f573e8cd51a7e2c62d68a59e4e261ea801d5ac203fd2724fc
        - quay.io/stolostron/flightctl-ocp-ui@sha256:6cff6425850d2b4a8b885a684afb01244d7cd9de7035d7ad71b1e10e5bab25e9
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/kube-rbac-proxy@sha256:e5555e256a518ca34bea01c959bbb5497582b53dbfaaf326e1852430c175637a
        - quay.io/stolostron/flightctl-ui@sha256:fea717f3b8b4c343ec60ce7d74a691f1124b6847d55e30a73dc0d50ea610d370
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/cluster-backup-controller@sha256:e576c2de1f931b7d33da2dbc03d71ce323766bd9158bf4b128221ddd1ab3fb9f
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/search-collector@sha256:2b3dfd21c0ba7b7b84fac390002e47fae927cd5fdf13e5d1d960703aa6c5bede
        - quay.io/stolostron/helloworld@sha256:4465550664bd210cf0df8edfd706de2fc51cacf01b66a962b5607935f9666ebc
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/metrics-collector@sha256:c7433ee4ccdb56717802acf258530e2d6d237787d760a8fbcfd642790254d5ab
        version: 0.7.1
      entries:
      - name: stolostron.v0.7.1
        version: 0.7.1
      - name: stolostron.v0.7.0
        version: 0.7.0
      name: community-0.7
    defaultChannel: community-0.7
    packageName: stolostron
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MinIO
      provider-url: https://min.io
    name: minio-object-store-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-object-store-operator.v2025.8.19175300
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aistor.min.io/v1",
                "kind": "ObjectStore",
                "metadata": {
                  "annotations": {
                    "prometheus.io/path": "/minio/v2/metrics/cluster",
                    "prometheus.io/port": "9000",
                    "prometheus.io/scrape": "true"
                  },
                  "labels": {
                    "app": "minio"
                  },
                  "name": "my-objectstore"
                },
                "spec": {
                  "cache": {
                    "enabled": false
                  },
                  "certificates": {
                    "disableAutoCert": false
                  },
                  "configuration": {
                    "name": "storage-configuration"
                  },
                  "firewall": {
                    "enabled": false
                  },
                  "mountPath": "/export",
                  "podManagementPolicy": "Parallel",
                  "pools": [
                    {
                      "name": "pool-0",
                      "securityContext": {
                        "fsGroupChangePolicy": "OnRootMismatch"
                      },
                      "servers": 4,
                      "volumeClaimTemplate": {
                        "apiVersion": "v1",
                        "kind": "persistentvolumeclaims",
                        "spec": {
                          "accessModes": [
                            "ReadWriteOnce"
                          ],
                          "resources": {
                            "requests": {
                              "storage": "1Ti"
                            }
                          },
                          "storageClassName": "standard"
                        }
                      },
                      "volumesPerServer": 4
                    }
                  ]
                }
              },
              {
                "apiVersion": "aistor.min.io/v1alpha1",
                "kind": "AdminJob",
                "metadata": {
                  "name": "minio-job"
                },
                "spec": {
                  "commands": [
                    {
                      "args": {
                        "name": "iac-bucket"
                      },
                      "op": "make-bucket"
                    }
                  ],
                  "containerSecurityContext": {},
                  "objectStore": {
                    "name": "my-objectstore",
                    "namespace": "my-objectstore"
                  },
                  "securityContext": {},
                  "serviceAccountName": "my-objectstore-sa"
                }
              },
              {
                "apiVersion": "sts.min.io/v1beta1",
                "kind": "PolicyBinding",
                "metadata": {
                  "name": "example-binding"
                },
                "spec": {
                  "application": {
                    "namespace": "my-objectstore",
                    "serviceaccount": "my-objectstore-sa"
                  },
                  "policies": [
                    "consoleAdmin"
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,AI/Machine Learning
          createdAt: "2025-08-27T14:50:56Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AdminJob
            name: adminjobs.aistor.min.io
            version: v1alpha1
          - kind: ObjectStore
            name: objectstores.aistor.min.io
            version: v1
          - kind: PolicyBinding
            name: policybindings.sts.min.io
            version: v1alpha1
          - kind: PolicyBinding
            name: policybindings.sts.min.io
            version: v1beta1
        description: "MinIO AIStor extends the most sophisticated object store on
          the market. It is designed for the exascale data \ninfrastructure challenges
          presented by modern AI workloads. AIStor provides new features, along with
          performance and\nscalability improvements, to enable enterprises to store
          all AI data in one infrastructure.\n\nMinIO is the standard for building
          <a href=\"https://min.io/solutions/object-storage-for-ai\">large scale AI
          data infrastructure</a>.\n\nThe MinIO AIStor is a software-defined, Amazon
          S3 compatible object store that will run anywhere - from the public cloud
          to the edge.<br><br>\n\nEnterprises use MinIO to deliver against artificial
          intelligence, machine learning, analytics, application, backup and archival
          workloads - all from a single platform.\n\nThe use cases include storage
          for HDFS replacements, AI datalakes, SIEM datalakes, analytics datalakes,
          object storage as a service, streaming and database warm/hot storage.<br><br>\n\nMinIO’s
          AIStor is the fastest object storage server, with production READ/WRITE
          speeds in excess of 2.2 TiB/s on commodity hardware.\n\nBuilt for the cloud
          operating model it is native to the technologies and architectures that
          define the cloud. These include containerization, orchestration with Kubernetes,
          microservices and multi-tenancy - making it the most widely integrated object
          store in the market today.<br><br>\n\nThe AIStor offers a host of enterprise
          features including inline erasure coding, bit-rot detection, state-of-the-art
          encryption, active-active replication, object locking, lifecycle management,
          observability, key-management, caching, firewall/load balancing, global
          console, search and identity + access management.<br><br>\n\nThe AIStor
          is priced on usable capacity and includes support through the MinIO Subscription
          Network.\n\nThe Subscription Network is designed to optimize deployments
          of MinIO and reduce the risk associated with data loss and data breach while
          providing 24/7/365 direct-to-engineering support.<br><br>\n\n<a href=\"https://resources.min.io/request-a-demo-marketplaces/\">Request
          a demo today</a>"
        displayName: Minio AIStor Object Store
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Storage
        - S3
        - MinIO
        - Object Storage
        links:
        - name: Minio AIStor
          url: https://min.io
        - name: Support
          url: https://subnet.min.io
        - name: Github
          url: https://github.com/minio/aistor
        maintainers:
        - email: dev@min.io
          name: MinIO
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: MinIO
          url: https://min.io
        relatedImages:
        - quay.io/minio/aistor/operator@sha256:f80139efa062fdd83f78318b20f371a827b68e854b03dd9706b9eba4c493aeb2
        version: 2025.8.19175300
      entries:
      - name: minio-object-store-operator.v2025.8.19175300
        version: 2025.8.19175300
      - name: minio-object-store-operator.v2025.8.1184927
        version: 2025.8.1184927
      - name: minio-object-store-operator.v2025.7.1011319
        version: 2025.7.1011319
      - name: minio-object-store-operator.v2025.6.3065135
        version: 2025.6.3065135
      - name: minio-object-store-operator.v2025.5.12190907
        version: 2025.5.12190907
      name: stable
    defaultChannel: stable
    packageName: minio-object-store-operator
    provider:
      name: MinIO
      url: https://min.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: https://stakater.com
    name: template-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: template-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-07-04T00:40:36Z"
          description: Operator to distribute plain kubernetes manifests, helm charts
            and, copy secrets and configmaps to multiple namespaces
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.0.1
          operatorframework.io/suggested-namespace: template-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterTemplateInstance is the Schema for the clustertemplateinstances
              API
            displayName: Cluster Template Instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateGroupInstance is the Schema for the TemplateGroupInstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.templates.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
        description: '"Operator to distribute plain kubernetes manifests, helm charts
          and, copy secrets and configmaps to multiple namespaces"'
        displayName: Template Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - templates
        - secrets
        - configmaps
        - distribution
        - stakater
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/template-operator@sha256:ad8312417e4e9621b74d3ad03020516840feef9c4b52c4f0d894dbb6dc6f906a
        version: 0.0.1
      entries:
      - name: template-operator.v0.0.1
        version: 0.0.1
      name: release-0.0
    - currentCSV: template-operator.v0.1.5
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-08-29T06:41:42Z"
          description: Operator to distribute plain kubernetes manifests, helm charts
            and, copy secrets and configmaps to multiple namespaces
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.1.5
          operatorframework.io/suggested-namespace: template-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterTemplateInstance is the Schema for the clustertemplateinstances
              API
            displayName: Cluster Template Instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateGroupInstance is the Schema for the TemplateGroupInstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.templates.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
        description: '"Operator to distribute plain kubernetes manifests, helm charts
          and, copy secrets and configmaps to multiple namespaces"'
        displayName: Template Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - templates
        - secrets
        - configmaps
        - distribution
        - stakater
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/template-operator@sha256:88f722baf22ac07704503f098f2b0ea7d76815a1ab0552474b1c9ba6639d8142
        version: 0.1.5
      entries:
      - name: template-operator.v0.1.5
        version: 0.1.5
      name: release-0.1
    defaultChannel: release-0.1
    packageName: template-operator
    provider:
      name: Stakater
      url: https://stakater.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: loki-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-operator.v5.9.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:6728046841367066216b304ed06249b80c57c74151c6a75c9df04a6f7f76e03d
          createdAt: "2025-01-30T17:34:53Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.7.0-0 <5.9.15'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:0c3943fa009f0c649f6d592b0ac410653c8c44b4a90e41fabdfa5f3f7db1e43d
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:6728046841367066216b304ed06249b80c57c74151c6a75c9df04a6f7f76e03d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:2e164e377d4fba98e7f63f43f25b6b9dad441d2e2914c2bf02f3525d5c9aa483
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:542b28fd51194b46efc4e5af10c3241b52db4c71f0d5d1b1f184eb39764e6c53
        version: 5.9.15
      entries:
      - name: loki-operator.v5.9.15
        version: 5.9.15
      - name: loki-operator.v5.9.14
        version: 5.9.14
      - name: loki-operator.v5.9.13
        version: 5.9.13
      - name: loki-operator.v5.9.12
        version: 5.9.12
      - name: loki-operator.v5.9.11
        version: 5.9.11
      - name: loki-operator.v5.9.10
        version: 5.9.10
      - name: loki-operator.v5.9.9
        version: 5.9.9
      - name: loki-operator.v5.9.8
        version: 5.9.8
      - name: loki-operator.v5.9.7
        version: 5.9.7
      - name: loki-operator.v5.9.6
        version: 5.9.6
      - name: loki-operator.v5.9.5
        version: 5.9.5
      - name: loki-operator.v5.9.4
        version: 5.9.4
      - name: loki-operator.v5.9.3
        version: 5.9.3
      - name: loki-operator.v5.9.2
        version: 5.9.2
      - name: loki-operator.v5.9.1
        version: 5.9.1
      - name: loki-operator.v5.9.0
        version: 5.9.0
      name: stable
    - currentCSV: loki-operator.v5.9.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:6728046841367066216b304ed06249b80c57c74151c6a75c9df04a6f7f76e03d
          createdAt: "2025-01-30T17:34:53Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.7.0-0 <5.9.15'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:2e164e377d4fba98e7f63f43f25b6b9dad441d2e2914c2bf02f3525d5c9aa483
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:542b28fd51194b46efc4e5af10c3241b52db4c71f0d5d1b1f184eb39764e6c53
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:0c3943fa009f0c649f6d592b0ac410653c8c44b4a90e41fabdfa5f3f7db1e43d
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:6728046841367066216b304ed06249b80c57c74151c6a75c9df04a6f7f76e03d
        version: 5.9.15
      entries:
      - name: loki-operator.v5.9.15
        version: 5.9.15
      - name: loki-operator.v5.9.14
        version: 5.9.14
      - name: loki-operator.v5.9.13
        version: 5.9.13
      - name: loki-operator.v5.9.12
        version: 5.9.12
      - name: loki-operator.v5.9.11
        version: 5.9.11
      - name: loki-operator.v5.9.10
        version: 5.9.10
      - name: loki-operator.v5.9.9
        version: 5.9.9
      - name: loki-operator.v5.9.8
        version: 5.9.8
      - name: loki-operator.v5.9.7
        version: 5.9.7
      - name: loki-operator.v5.9.6
        version: 5.9.6
      - name: loki-operator.v5.9.5
        version: 5.9.5
      - name: loki-operator.v5.9.4
        version: 5.9.4
      - name: loki-operator.v5.9.3
        version: 5.9.3
      - name: loki-operator.v5.9.2
        version: 5.9.2
      - name: loki-operator.v5.9.1
        version: 5.9.1
      - name: loki-operator.v5.9.0
        version: 5.9.0
      name: stable-5.9
    - currentCSV: loki-operator.v6.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.14-v4.17
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:1cb1d2cad8632791529eb754938612f5e5ed373284e42f309d603059ab7ec7c5
          createdAt: "2025-03-12T12:05:37Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=5.8.0-0 <6.0.10'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.0
          operators.operatorframework.io.bundle.channels.v1: stable-6.0
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:9aa8d8e4a7fd7ddad5d842a22b09aecdbdfa31ca5eb8c4b17f0818636af5fbf8
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:69e69ab2628742774d12dd9dde50607625efdff50ea4671f2859692fbee5ecb4
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:98d2cf3437faa2400c01edb3c694816e0214c8f9dca0a0b0cdf5882a720248f3
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:1cb1d2cad8632791529eb754938612f5e5ed373284e42f309d603059ab7ec7c5
        version: 6.0.10
      entries:
      - name: loki-operator.v6.0.10
        version: 6.0.10
      - name: loki-operator.v6.0.9
        version: 6.0.9
      - name: loki-operator.v6.0.8
        version: 6.0.8
      - name: loki-operator.v6.0.7
        version: 6.0.7
      - name: loki-operator.v6.0.6
        version: 6.0.6
      - name: loki-operator.v6.0.5
        version: 6.0.5
      - name: loki-operator.v6.0.4
        version: 6.0.4
      - name: loki-operator.v6.0.3
        version: 6.0.3
      - name: loki-operator.v6.0.2
        version: 6.0.2
      - name: loki-operator.v6.0.1
        version: 6.0.1
      - name: loki-operator.v6.0.0
        version: 6.0.0
      name: stable-6.0
    - currentCSV: loki-operator.v6.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.15-v4.18
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:0bde3c1a0b667a693f443688ed3926eef1705c45ecf4cce3a3dedbddca7a2784
          createdAt: "2025-03-12T11:19:10Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=5.9.0-0 <6.1.8'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.1
          operators.operatorframework.io.bundle.channels.v1: stable-6.1
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:0bde3c1a0b667a693f443688ed3926eef1705c45ecf4cce3a3dedbddca7a2784
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:731f92c146f299bc3662b309a3068bde18fad51f65a7d29063e582b61e07042f
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:c34676932ddd14db213dbf42a1d341f51692f3f47f1b751d06187e292c99a170
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b94833f55644c289fe2c1297b68a4f9931baf977587c6a2c4c7a3e484c704f64
        version: 6.1.8
      entries:
      - name: loki-operator.v6.1.8
        version: 6.1.8
      - name: loki-operator.v6.1.7
        version: 6.1.7
      - name: loki-operator.v6.1.6
        version: 6.1.6
      - name: loki-operator.v6.1.5
        version: 6.1.5
      - name: loki-operator.v6.1.4
        version: 6.1.4
      - name: loki-operator.v6.1.3
        version: 6.1.3
      - name: loki-operator.v6.1.2
        version: 6.1.2
      - name: loki-operator.v6.1.1
        version: 6.1.1
      - name: loki-operator.v6.1.0
        version: 6.1.0
      name: stable-6.1
    - currentCSV: loki-operator.v6.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.16-v4.19
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f401be018c5ba35e8ea1e68831cc20067f0f4d1be031d05b222548c1bdfb5f7e
          createdAt: "2025-02-18T18:01:39Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=6.0.0-0 <6.2.5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.2
          operators.operatorframework.io.bundle.channels.v1: stable-6.2
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f401be018c5ba35e8ea1e68831cc20067f0f4d1be031d05b222548c1bdfb5f7e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:99d1e7c5c1f0609b33de576884782550f969b28dc622810e6661d4c60a664215
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:8955d8186ae98095623f2083b6d5389d3ac9f2612fcfcf695835219c3e1216c4
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b8c5b46a94a6c85c9d382eeeb65a1d2b4014a7dc765e9abfba706c92857462dc
        version: 6.2.5
      entries:
      - name: loki-operator.v6.2.5
        version: 6.2.5
      - name: loki-operator.v6.2.4
        version: 6.2.4
      - name: loki-operator.v6.2.3
        version: 6.2.3
      - name: loki-operator.v6.2.2
        version: 6.2.2
      - name: loki-operator.v6.2.1
        version: 6.2.1
      - name: loki-operator.v6.2.0
        version: 6.2.0
      name: stable-6.2
    defaultChannel: stable-6.2
    packageName: loki-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudtrail-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudtrail-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudtrail.services.k8s.aws/v1alpha1",
                "kind": "EventDataStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudtrail.services.k8s.aws/v1alpha1",
                "kind": "Trail",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudtrail-controller:1.1.1
          createdAt: "2025-09-19T21:16:25Z"
          description: AWS CloudTrail controller is a service controller for managing
            CloudTrail resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventDataStore represents the state of an AWS cloudtrail
              EventDataStore resource.
            displayName: EventDataStore
            kind: EventDataStore
            name: eventdatastores.cloudtrail.services.k8s.aws
            version: v1alpha1
          - description: Trail represents the state of an AWS cloudtrail Trail resource.
            displayName: Trail
            kind: Trail
            name: trails.cloudtrail.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Compute Cloud (CloudTrail) resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudTrail**

          AWS CloudTrail is an AWS service that helps you enable operational and risk auditing, governance, and compliance of your AWS account. Actions taken by a user, role, or an AWS service are recorded as events in CloudTrail. Events include actions taken in the AWS Management Console, AWS Command Line Interface, and AWS SDKs and APIs.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudTrail
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudtrail
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudTrail Developer Resources
          url: https://aws.amazon.com/cloudtrail/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudtrail maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudtrail-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-cloudtrail-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudtrail-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudtrail-controller.v1.0.28
        version: 1.0.28
      - name: ack-cloudtrail-controller.v1.0.26
        version: 1.0.26
      - name: ack-cloudtrail-controller.v1.0.25
        version: 1.0.25
      - name: ack-cloudtrail-controller.v1.0.24
        version: 1.0.24
      - name: ack-cloudtrail-controller.v1.0.23
        version: 1.0.23
      - name: ack-cloudtrail-controller.v1.0.22
        version: 1.0.22
      - name: ack-cloudtrail-controller.v1.0.21
        version: 1.0.21
      - name: ack-cloudtrail-controller.v1.0.20
        version: 1.0.20
      - name: ack-cloudtrail-controller.v1.0.19
        version: 1.0.19
      - name: ack-cloudtrail-controller.v1.0.16
        version: 1.0.16
      - name: ack-cloudtrail-controller.v1.0.15
        version: 1.0.15
      - name: ack-cloudtrail-controller.v1.0.14
        version: 1.0.14
      - name: ack-cloudtrail-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudtrail-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudtrail-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudtrail-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudtrail-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudtrail-controller.v1.0.5
        version: 1.0.5
      - name: ack-cloudtrail-controller.v1.0.4
        version: 1.0.4
      - name: ack-cloudtrail-controller.v1.0.3
        version: 1.0.3
      - name: ack-cloudtrail-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudtrail-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudtrail-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudtrail-controller.v0.1.3
        version: 0.1.3
      - name: ack-cloudtrail-controller.v0.1.2
        version: 0.1.2
      - name: ack-cloudtrail-controller.v0.1.1
        version: 0.1.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudtrail-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-route53-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-route53-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/route53-controller:1.1.1
          createdAt: "2025-09-19T19:15:31Z"
          description: AWS Route53 controller is a service controller for managing
            Route53 resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HealthCheck represents the state of an AWS route53 HealthCheck
              resource.
            displayName: HealthCheck
            kind: HealthCheck
            name: healthchecks.route53.services.k8s.aws
            version: v1alpha1
          - description: HostedZone represents the state of an AWS route53 HostedZone
              resource.
            displayName: HostedZone
            kind: HostedZone
            name: hostedzones.route53.services.k8s.aws
            version: v1alpha1
          - description: RecordSet represents the state of an AWS route53 RecordSet
              resource.
            displayName: RecordSet
            kind: RecordSet
            name: recordsets.route53.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Route53 resources in AWS from within your Kubernetes cluster.

          **About Amazon Route53**

          Amazon Route 53 is a highly available and scalable Domain Name System (DNS) web service. You can use Route 53 to perform three main functions in any combination: domain registration, DNS routing, and health checking.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Route53
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - route53
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Route53 Developer Resources
          url: https://docs.aws.amazon.com/route53/index.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/route53-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-route53-controller.v1.1.1
        version: 1.1.1
      - name: ack-route53-controller.v1.1.0
        version: 1.1.0
      - name: ack-route53-controller.v1.0.6
        version: 1.0.6
      - name: ack-route53-controller.v1.0.5
        version: 1.0.5
      - name: ack-route53-controller.v1.0.3
        version: 1.0.3
      - name: ack-route53-controller.v1.0.2
        version: 1.0.2
      - name: ack-route53-controller.v0.1.0
        version: 0.1.0
      - name: ack-route53-controller.v0.0.28
        version: 0.0.28
      - name: ack-route53-controller.v0.0.27
        version: 0.0.27
      - name: ack-route53-controller.v0.0.26
        version: 0.0.26
      - name: ack-route53-controller.v0.0.25
        version: 0.0.25
      - name: ack-route53-controller.v0.0.24
        version: 0.0.24
      - name: ack-route53-controller.v0.0.23
        version: 0.0.23
      - name: ack-route53-controller.v0.0.22
        version: 0.0.22
      - name: ack-route53-controller.v0.0.20
        version: 0.0.20
      - name: ack-route53-controller.v0.0.19
        version: 0.0.19
      - name: ack-route53-controller.v0.0.18
        version: 0.0.18
      - name: ack-route53-controller.v0.0.17
        version: 0.0.17
      - name: ack-route53-controller.v0.0.16
        version: 0.0.16
      - name: ack-route53-controller.v0.0.15
        version: 0.0.15
      - name: ack-route53-controller.v0.0.14
        version: 0.0.14
      - name: ack-route53-controller.v0.0.13
        version: 0.0.13
      - name: ack-route53-controller.v0.0.9
        version: 0.0.9
      - name: ack-route53-controller.v0.0.6
        version: 0.0.6
      - name: ack-route53-controller.v0.0.4
        version: 0.0.4
      - name: ack-route53-controller.v0.0.3
        version: 0.0.3
      - name: ack-route53-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-route53-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://www.medik8s.io/
    name: self-node-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: self-node-remediation.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/self-node-remediation-operator:v0.9.0
          createdAt: "2024-07-25 14:40:52"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.9.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Medik8s
          url: https://www.medik8s.io/
        relatedImages:
        - quay.io/medik8s/self-node-remediation-operator:v0.9.0
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        version: 0.9.0
      entries:
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      - name: self-node-remediation.v0.7.0
        version: 0.7.0
      - name: self-node-remediation.v0.6.1
        version: 0.6.1
      - name: self-node-remediation.v0.6.0
        version: 0.6.0
      - name: self-node-remediation.v0.5.0
        version: 0.5.0
      - name: self-node-remediation.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: self-node-remediation
    provider:
      name: Medik8s
      url: https://www.medik8s.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudfront-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudfront-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "CachePolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "Distribution",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudfront-controller:1.2.1
          createdAt: "2025-09-19T21:12:14Z"
          description: AWS CloudFront controller is a service controller for managing
            CloudFront resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CachePolicy represents the state of an AWS cloudfront CachePolicy
              resource.
            displayName: CachePolicy
            kind: CachePolicy
            name: cachepolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: Distribution represents the state of an AWS cloudfront Distribution
              resource.
            displayName: Distribution
            kind: Distribution
            name: distributions.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: Function represents the state of an AWS cloudfront Function
              resource.
            displayName: Function
            kind: Function
            name: functions.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: OriginAccessControl represents the state of an AWS cloudfront
              OriginAccessControl resource.
            displayName: OriginAccessControl
            kind: OriginAccessControl
            name: originaccesscontrols.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: OriginRequestPolicy represents the state of an AWS cloudfront
              OriginRequestPolicy resource.
            displayName: OriginRequestPolicy
            kind: OriginRequestPolicy
            name: originrequestpolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: ResponseHeadersPolicy represents the state of an AWS cloudfront
              ResponseHeadersPolicy resource.
            displayName: ResponseHeadersPolicy
            kind: ResponseHeadersPolicy
            name: responseheaderspolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: VPCOrigin represents the state of an AWS cloudfront VPCOrigin
              resource.
            displayName: VPCOrigin
            kind: VPCOrigin
            name: vpcorigins.cloudfront.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudFront resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudFront**

          Amazon CloudFront is a web service that speeds up distribution of your static and dynamic web content, such as .html, .css, .js, and image files, to your users. CloudFront delivers your content through a worldwide network of data centers called edge locations.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon CloudFront
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudfront
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudFront Developer Resources
          url: https://aws.amazon.com/cloudfront/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudfront maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudfront-controller:1.2.1
        version: 1.2.1
      entries:
      - name: ack-cloudfront-controller.v1.2.1
        version: 1.2.1
      - name: ack-cloudfront-controller.v1.2.0
        version: 1.2.0
      - name: ack-cloudfront-controller.v1.1.6
        version: 1.1.6
      - name: ack-cloudfront-controller.v1.1.4
        version: 1.1.4
      - name: ack-cloudfront-controller.v1.1.3
        version: 1.1.3
      - name: ack-cloudfront-controller.v1.1.2
        version: 1.1.2
      - name: ack-cloudfront-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudfront-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudfront-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudfront-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudfront-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudfront-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudfront-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudfront-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudfront-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudfront-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudfront-controller.v0.0.15
        version: 0.0.15
      - name: ack-cloudfront-controller.v0.0.14
        version: 0.0.14
      - name: ack-cloudfront-controller.v0.0.13
        version: 0.0.13
      - name: ack-cloudfront-controller.v0.0.12
        version: 0.0.12
      - name: ack-cloudfront-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudfront-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudfront-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudfront-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudfront-controller.v0.0.6
        version: 0.0.6
      - name: ack-cloudfront-controller.v0.0.5
        version: 0.0.5
      - name: ack-cloudfront-controller.v0.0.4
        version: 0.0.4
      - name: ack-cloudfront-controller.v0.0.3
        version: 0.0.3
      - name: ack-cloudfront-controller.v0.0.2
        version: 0.0.2
      - name: ack-cloudfront-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudfront-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/openshift/lightspeed-service
    name: lightspeed-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lightspeed-operator.v1.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ols.openshift.io/v1alpha1",
                "kind": "OLSConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightspeed-operator",
                    "app.kubernetes.io/instance": "olsconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "olsconfig",
                    "app.kubernetes.io/part-of": "lightspeed-operator"
                  },
                  "name": "cluster"
                },
                "spec": {
                  "llm": {
                    "providers": [
                      {
                        "credentialsSecretRef": {
                          "name": "credentials"
                        },
                        "models": [
                          {
                            "name": "gpt-3.5-turbo-1106"
                          }
                        ],
                        "name": "OpenAI"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-08-28T16:37:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lightspeed
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Kubernetes Engine", "OpenShift Virtualization
            Engine"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lightspeed-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Red Hat OpenShift Lightspeed instance. OLSConfig is the Schema
              for the olsconfigs API
            displayName: OLSConfig
            kind: OLSConfig
            name: olsconfigs.ols.openshift.io
            version: v1alpha1
        description: |-
          OpenShift Lightspeed Operator provides generative AI-based virtual assistant which integrates into the OpenShift web console. OpenShift Lightspeed can answer natural language questions related to OpenShift Container Platform.

          To configure the OpenShift Lightspeed Operator, you need a Large Language Model (LLM) . You can host the LLM using RHELAI or RHOAI or use one of the models from the following Software-as-a-Service providers:

          - OpenAI
          - Microsoft Azure OpenAI
          - IBM WatsonX

          For more information, see [About OpenShift Lightspeed](https://docs.redhat.com/en/documentation/red_hat_openshift_lightspeed/1.0tp1/html/about/ols-about-openshift-lightspeed) in the official product documentation.

          Questions, comments, concerns, or just want to talk to the OpenShift Lightspeed team? [Email us](mailto:openshift-lightspeed-contact-requests@redhat.com)!
        displayName: OpenShift Lightspeed Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ai
        - assistant
        - openshift
        - llm
        links:
        - name: Lightspeed Operator
          url: https://github.com/openshift/lightspeed-operator
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/lightspeed-service
        relatedImages:
        - registry.redhat.io/openshift-lightspeed/lightspeed-rhel9-operator@sha256:b12803642df4d9b3ad0cf800315021483f543c6a8827b3901e57bdb65f10dc48
        - registry.redhat.io/openshift-lightspeed/openshift-mcp-server-rhel9@sha256:f1e4521ff2cc076f361c04d8c9aeac09ffcedb89a2f1dbe71f85caa27472adab
        - registry.redhat.io/openshift-lightspeed/lightspeed-service-api-rhel9@sha256:ec359967a87bc7ee2c1d6a61e553acbdb0dc85d6f240a1865bf57d3ba5dc7193
        - registry.redhat.io/openshift-lightspeed/lightspeed-console-plugin-rhel9@sha256:dfc42c2651ac8793f61ff0e54334ecb006dc4a77487bc7c6878cdf47bd4ac080
        version: 1.0.5
      entries:
      - name: lightspeed-operator.v1.0.5
        version: 1.0.5
      name: alpha
    - currentCSV: lightspeed-operator.v1.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ols.openshift.io/v1alpha1",
                "kind": "OLSConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightspeed-operator",
                    "app.kubernetes.io/instance": "olsconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "olsconfig",
                    "app.kubernetes.io/part-of": "lightspeed-operator"
                  },
                  "name": "cluster"
                },
                "spec": {
                  "llm": {
                    "providers": [
                      {
                        "credentialsSecretRef": {
                          "name": "credentials"
                        },
                        "models": [
                          {
                            "name": "gpt-3.5-turbo-1106"
                          }
                        ],
                        "name": "OpenAI"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-08-28T16:37:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lightspeed
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Kubernetes Engine", "OpenShift Virtualization
            Engine"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lightspeed-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Red Hat OpenShift Lightspeed instance. OLSConfig is the Schema
              for the olsconfigs API
            displayName: OLSConfig
            kind: OLSConfig
            name: olsconfigs.ols.openshift.io
            version: v1alpha1
        description: |-
          OpenShift Lightspeed Operator provides generative AI-based virtual assistant which integrates into the OpenShift web console. OpenShift Lightspeed can answer natural language questions related to OpenShift Container Platform.

          To configure the OpenShift Lightspeed Operator, you need a Large Language Model (LLM) . You can host the LLM using RHELAI or RHOAI or use one of the models from the following Software-as-a-Service providers:

          - OpenAI
          - Microsoft Azure OpenAI
          - IBM WatsonX

          For more information, see [About OpenShift Lightspeed](https://docs.redhat.com/en/documentation/red_hat_openshift_lightspeed/1.0tp1/html/about/ols-about-openshift-lightspeed) in the official product documentation.

          Questions, comments, concerns, or just want to talk to the OpenShift Lightspeed team? [Email us](mailto:openshift-lightspeed-contact-requests@redhat.com)!
        displayName: OpenShift Lightspeed Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ai
        - assistant
        - openshift
        - llm
        links:
        - name: Lightspeed Operator
          url: https://github.com/openshift/lightspeed-operator
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/lightspeed-service
        relatedImages:
        - registry.redhat.io/openshift-lightspeed/lightspeed-service-api-rhel9@sha256:ec359967a87bc7ee2c1d6a61e553acbdb0dc85d6f240a1865bf57d3ba5dc7193
        - registry.redhat.io/openshift-lightspeed/lightspeed-console-plugin-rhel9@sha256:dfc42c2651ac8793f61ff0e54334ecb006dc4a77487bc7c6878cdf47bd4ac080
        - registry.redhat.io/openshift-lightspeed/openshift-mcp-server-rhel9@sha256:f1e4521ff2cc076f361c04d8c9aeac09ffcedb89a2f1dbe71f85caa27472adab
        - registry.redhat.io/openshift-lightspeed/lightspeed-rhel9-operator@sha256:b12803642df4d9b3ad0cf800315021483f543c6a8827b3901e57bdb65f10dc48
        version: 1.0.5
      entries:
      - name: lightspeed-operator.v1.0.5
        version: 1.0.5
      name: stable
    defaultChannel: stable
    packageName: lightspeed-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/openshift/lightspeed-service
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kubevirt-hyperconverged
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubevirt-hyperconverged-operator.v4.16.23
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:eaacb0d177fd64e2eefbcb2dc47633dadf4d54fcbe0f0f750b078ab3f4f189e1
          createdAt: "2025-09-29 21:23:00"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.14.1 <4.16.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:3e1c68d5695bab96c382661a33581e60597b59afd21660ec760d8fe5e20ff9e3
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.16.23-1,hyperconverged-cluster-operator-test-rhel9:v4.16.23-1,kubevirt-common-instancetypes-test-rhel9:v4.16.23-1,kubevirt-tekton-tasks-test-rhel9:v4.16.23-1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:e24191be577575fa9d462c87c6f51775a0cf0f5d1812d12c4c95a68ec34b34ff
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:6e4aa75bece3c2fe076e282b3f1ad8296299ec0551e379c7c88830a0ce8ed5bf
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:1be66eee321bc47dcc0e7ff4674a485b9a4552243080d796d5f3b0d417721dc4
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:28c0bdf60404fb435cfa0cfdc6d314f789417c44c3206ce8cd1e7b99031fb529
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:659abe1d740607b2cb3850c9ab29101fa5d5c821d2a09aab5994849697a504f3
        - registry.redhat.io/container-native-virtualization/mtq-operator-rhel9@sha256:9de5ded35edfbe98121bdd4404c84dcc72c8cba1742dc27df6382ef882db2183
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:0f0478e9357f8b6ace6d844e6e424db192398195a8cea2fa6398be8996d05be6
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:2843d199dc2748749d5f48fb9f90656962e011533d79dc7e0a4f3592b12aa69a
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:6b0c7cc17abd95fce45b6ec83ec291e843eedbc90dce0ecc449aec6e64562057
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:97d37927164f0ea986f209912e6247fdab37b680c214afc4bdbb7051caaf58b1
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:cc4ff488aa97e179c4df39ac5ae8085070017106adf1c5f2222b35d8f2234ea8
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:f8e2ea3988d4a77a4e809e1fd073ffa5f99393ca3eb69a2c1cb54dd4ab60e482
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:885670f2e9ad43224cbc96f34b23dd2642de00f688bade0f6b89dd3d3fb83dbd
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:251d6fb3197a7642d86193191aa69c85a45c5521d7fc509704edd46d5d4e90a7
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:fcbadecf01b81625d55f9219428baabc33ee7231b4ccf801c90dee3095907187
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:a0af0e5d4b142f28dc72c5636df5040c4ce92c6b12ab183d3a9bc7a9d2d2f418
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:c033ddd0ed30c0d25f4c2e9b30fca4e40959bb6d74029a1f452ff048b97de88a
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:7206ac7959c07f474f454ab48bcf3522ade95efb27fd9fbd3938d1387a1420af
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:468f117a1b55af59e72a9ff46499a8d0bb00884d450d57b692e2f7b557c9bf3d
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:e01871006b9e4cb1b6f477319de63b758eac672a70a298caed2032ec19986680
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:7587832b877f32621eb26e67a5c516728f293ecd98f74dd70d5f03474a5429d6
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:68b650d6779fc4d61f05d97e6d62b80fcd638721994153205eec68f8c1d999ce
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:19ca70dceb79b10ff458f0ea41faa6e5552a5f08828620ad7177efa34166a90e
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:84dbfa51755d60e34b194c1acf679ac083feb49038112212828fe8101dcec2a5
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:8462188ec7533d2105c0dc9154d8dcf1425b96d6945eef16ea774a5dff9bba1d
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:11030a8d8cd03efdf27fe3cfc9ebeadeb20851b6de1e9d8a93ba7f2bc6a77a96
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:8cf7993326ba220f650ae2fa5383d0d6cf3e2c7f3346f0a339a9d212b8df02f6
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:c18b6e315f5577cc18d46cd76595d48d4a49026a6b2f5763c3efb26919bde5eb
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:52b4f7c7a4f5c5c5d07a720c5eefa70709f520b1d939aec4cd971d8f3c9b12bc
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:ef8260d72239ad2126ca023250bfd976f5c050fe9c0727a0f7ef007446f5363b
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:18be4ef8b6c2cedab4a2e031939fc1f9cc23020bde7fccb7997e4222b673f57f
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:def80efdcc58d545aaf4b69b784a348cb36e97afd8e0124480521a2b980fa7c3
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:b031a12e73cb75e207203fd3785619e3a5a22d7d5a8549488c035f70374235be
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:14ea56d4149b72b98b6e7f5b33ba5bdd9115e7c643fd103c9df7a26b2a06700a
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:9e358366fd11dd167255abd930016aa4fc35b1b9dc69ce27316635488efc8ec4
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:6bf45534310768308958428692b29afa25f2740d0c62165d7182b0b3e9e944cf
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:eaacb0d177fd64e2eefbcb2dc47633dadf4d54fcbe0f0f750b078ab3f4f189e1
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:91b86228495f38d5528be1a8ffec96695ad72fafb76d2ead87f7bee25540d8b6
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:19b56bbb081db7198d570c3c691e142991aaa81857ee417d47903291ed0e7a1c
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:372652529b6a8705df581e7ab8acc2ec5c3cdf8fb273448052c59963abe4d984
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:7f89f2ba59309b26cf83f58ff3ba35569182097258fa543d5d88f3a8cd64ca00
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:bf1bb2fe09e028506fe7cf383368d76e358b8c59ecc74d3618bc250c1e97271c
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:a79579b18a7df9f23c584303593a9404c64df023458d55c715aeb71dcacc92a0
        - registry.redhat.io/container-native-virtualization/mtq-lock-server-rhel9@sha256:bb1dfc204e48a13985d6c2ec9cdf17f47330617725ee6dd854f0dae0f1901ba4
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:4005d165f1485eec98ca24ce41bef79cd8088002d43ffb709ec65f842a176bdc
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:3f3554ef043c80494db536b28eb03fa7043e53bb85ce426820c252fb999ce04e
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:531702869a34f76850d332efb84ccf571fafa281014afd1787497ca9de8aed37
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:a081e7e68def13b7de369d3f66469035d61fd1feef1bee4f95ba5f7b4041d960
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:e7672121156e050f05af771f75cfe26558699da8a074bf7051f710705a27d743
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:bd532e59edae5d2edac78b9249cd6a9f0779f44e25b019d824a1b9803f078bde
        - registry.redhat.io/container-native-virtualization/mtq-controller-rhel9@sha256:a774f0bd125c66b45e35e327c9c7e0afd7947d3c5ed222b3d5d223941366ba2c
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:52e76b4208ee4e7da8a0729b9bdb294729454d0f62580d64dd104b6921f464ed
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:3b57db7aaa3c61335044ec78cbd061f78ce99cdb4b75b3ce342e1a603da4bcef
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:7e19b42a648cee71c60cb69fbcf6ebca05c938942e16d23f1cd61a619cffdcb6
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:4046e8e862e0c55f09e9031b95e6a77af4d491593e39770c644287d44888b0bb
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:517c88a6be8842d4eb71ad30f99ce1afca954af1df2c1bf4a9584f4ec69fcf12
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:3e1c68d5695bab96c382661a33581e60597b59afd21660ec760d8fe5e20ff9e3
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:cca121312777d2727fe2c50a14583243ce69e6e356b499f133736787522095b2
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:ad24eb52bd65495224661d4894336d88fde342aeaab16b1084ceff6337f92059
        version: 4.16.23
      entries:
      - name: kubevirt-hyperconverged-operator.v4.16.23
        version: 4.16.23
      - name: kubevirt-hyperconverged-operator.v4.16.22
        version: 4.16.22
      - name: kubevirt-hyperconverged-operator.v4.16.21
        version: 4.16.21
      - name: kubevirt-hyperconverged-operator.v4.16.20
        version: 4.16.20
      - name: kubevirt-hyperconverged-operator.v4.16.19
        version: 4.16.19
      - name: kubevirt-hyperconverged-operator.v4.16.18
        version: 4.16.18
      - name: kubevirt-hyperconverged-operator.v4.16.17
        version: 4.16.17
      - name: kubevirt-hyperconverged-operator.v4.16.16
        version: 4.16.16
      - name: kubevirt-hyperconverged-operator.v4.16.15
        version: 4.16.15
      - name: kubevirt-hyperconverged-operator.v4.16.14
        version: 4.16.14
      - name: kubevirt-hyperconverged-operator.v4.16.13
        version: 4.16.13
      - name: kubevirt-hyperconverged-operator.v4.16.12
        version: 4.16.12
      - name: kubevirt-hyperconverged-operator.v4.16.11
        version: 4.16.11
      - name: kubevirt-hyperconverged-operator.v4.16.10
        version: 4.16.10
      - name: kubevirt-hyperconverged-operator.v4.16.9
        version: 4.16.9
      - name: kubevirt-hyperconverged-operator.v4.16.8
        version: 4.16.8
      name: candidate
    - currentCSV: kubevirt-hyperconverged-operator.v4.99.0-0.1723448771
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:5b2c64218342a5744182fc06f550bcdcf954e931275a9d1505ef0bbfbd0f8044
          createdAt: "2024-08-12 07:46:45"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.14.1 <4.16.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:d74c629e415d177069a0570fb8db9252d2b193d2ec7972c844686ffd14033f62
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.99.0-73,kubevirt-common-instancetypes-test-rhel9:v4.99.0-32,kubevirt-tekton-tasks-test-rhel9:v4.99.0-31
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:cc190d7c44b20f10ae61709fd2f5f7c0b248df73b768eb696d5a6cec90a34e59
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:41ee360ac43ee69ccee62d4ac3d2f07c8a8589e5b85c659b47018dd36c2b932f
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:b410508ecc71b0a3bf4825793efde4b07085e2243d8b3585da852fd1940c4be1
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:820a9e6bd0329df5c541ef0b67b3c7cdd9f81be33de87c5b5e5e43777bea197e
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:811e8e7e9872408794ef36229c42c6c437b1a4ace76c7134d91bb2331fe7815b
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:0769512df0fa1325d76a5f10a1a704cf60fc65f62f57b157ea53bc9ba682851f
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:8c64b526fd42a17c909bf45eb795121a81d502424a9cf11f8087118751333894
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:5f23fc8e9bc386622460c23b9644f58648d9f748f5670927a02888cd36c99a21
        - registry.redhat.io/container-native-virtualization/mtq-controller-rhel9@sha256:ad5938ebd5f07196ce5c35232a1c3b0a653301a83929ce84c3d1cb4dd1b91816
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:37e6b4fa5ff436572412ec0bdb1edc4a49e62c994e056c93578605f6b74adbcd
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:07435335fdbf6178bc00cc02b93cc0adb42784339151b2d9c31dc8da98b8bc6a
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:cffec516ab9ddc50cab77339d0cb3239ef4e7fd43bd2afdc73370aa487ea5704
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:dec430b74887e9b42d639f14c07d9f58a4881cac24e949dab98f96af150f1958
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:c61732be8024b4ede45eec8f786e51b198632ee5bfb0c2f5af84f37544584283
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:5ba79dd6bdf9e3778d3e301919b37e45cfc1eaf72fda77eba6588788829ba29b
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:c14cd037d14489321629908fc6b30248f7a20d866619177a165d72cfc1b6b7ef
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:83c03929cf11d7202e6c6cb9dd767bd4d7473f4fc1d562b79719ceb3d8382e68
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:6b7beed3cba9093b3aeb3109f23ec38c20f5568915dff774ad5e9e6bb5db3a58
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:e39732ebd3a4452fa7a678aeea5e64a7973fd72802eff193b16680c2f0129cb1
        - registry.redhat.io/container-native-virtualization/mtq-lock-server-rhel9@sha256:01cc6797800ccfb99453858ab36c0100c5ca343f1f71f975914b2437d1f757ba
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:9822f7a1590d830ccd64cc0729d2e30a2e04fb2b022f54498e1f1abff3748c2c
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:bf2ec130d1f3f32314d740037313f49f874a271e8b9a7f6cd2b93698a8423e8d
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:42a34caa18b5f5452b9d5603602ea74b1ee59a8bbb52bddf18901fecc27aed8a
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:18dfb38223647e11020e3ef1093a695454b90cc50922f80afb612212b7d13309
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:4c66a26b9f74d83a9bc60edb40ffd9d6ff2382d20500677d4567f8e141c3ddf0
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:758a9f6309a367d3f31cc03dd89356a8232a0fbd277570c01ebfa1fa30089a8d
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:2e5d5a2a70cda1f414f22b1c713d77d068c5fb16b1487a572019a9aa7130e6be
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:84b199e98e283c863d52a7b00044c20cb56ad563494adbec2229af265bf3f7c5
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:6baf67f3f410c071cc26f675d86edbbae5eaeb3410888cf622fad384b43b80ea
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:3a45cd2f3ea25f356167ce4fa20894a2a0706b4bcffa42c806f6017812053466
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:84679ffe8ce86c674818531897d122f5568d10c7e1e2ff919185d5771e5b4c39
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:ac33de246ee42bbfe735d7349d85c6df7e583e7599dea7dd7a2235bd0e59845f
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:e7145f016642b02bd651e21aa19edc7f93d0d1c2da3c6c7eed5d1174cf23619c
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:2564af054f52a8bb340e0270379c72da18658662a2a5bd26ac7bd979ddfccb56
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:4cff43cc145361c74c357055d51060ccc27e6bca58ccc3b5f5b983c7b8545114
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:09a30b59c906ddc4604d70f18210bcebef1f90f8bf965422b0caba363cd88989
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:63389195cf584a07ddbc2efec90cf6deaf737b91314b67db2b9f53dcc7acb57c
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:6193e569d4ab43806ac7b4870ef3634a5de3687527d6d190e59a9f735fa4778c
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:f4e7876fa8b27a8082efa99455d9bf7b0c479cff1e4c8a26136bf63092699c0b
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:d7e8e075a08b148ddfa91b44bb9cae1ef823756760279c4bb6d8489bc7e0eb2c
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:bf30b6c7173ebb6a4592246af16db87d7d391182f52286b559f8bad8f4b6cc5b
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:15179f0826831f051fdb867067be7c5528f01f442ae421b8782ae0cd1ed1b73d
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:4b3ac5ec2fafef0b7702a02847996f0903ec8d4262ccd03e8fa74885265ea4e9
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:d6dfd062df3d0d49f7efb928b835b9c8e0bcfb514a653ba2896a0152f24a0a21
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:185a83c994aced74546fe630cd71656f1e4112cb15be75f41dfcb1b9ed0966bd
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:b59e34cdb6214eacb9dc176a7b33125d7c1871f6b81696d59646294bfaf749e7
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:8e19d99232be4c37e857a4931627fb0aed721893e27b0d0e522fcf13bf67c1c9
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:06cc71822684c3ed094f327553451daf076d6bb97065c314da1a7d2adbbef3d2
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:4bdafc4404d5cfd844d05f32d88c262fba56dc931bfd5522b78c25d4ff253883
        - registry.redhat.io/container-native-virtualization/kubevirt-ipam-controller-rhel9@sha256:f33ac4d7477a730f273a462f171aed35dea38374d5ea7ef0bae2d7f74413a757
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:af5d2a2d4f0b6f0e1a13f1029073bb56980d8a70503b21628f18e8986e9b531c
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:5b2c64218342a5744182fc06f550bcdcf954e931275a9d1505ef0bbfbd0f8044
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:88f02cbebd6c3a2add6e0f8ca744e42bbd182f5a07d4fac59cf5cf07a4edfd45
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:fa0478fbf91248166f0eed8ec5f8dbb5b31d26031438376d9301836b6a069643
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:9bfdaf1cca6eac15b63a4ea6b7de28f2328516c5cf8bd5727935dd729e45657d
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:30c669770b0b992a3c84bb81c892e70a24fb14af187e8a4ee19cb3e0dd56ddbb
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:c5e8f44f647f593b00d3427172932d7aa694379d124209b9c66b4c1ffe3d4116
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:e6d9a6c41433360abd42d1118f5af3df9592d6d36af5f65082ae214cc8513c1b
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:d74c629e415d177069a0570fb8db9252d2b193d2ec7972c844686ffd14033f62
        - registry.redhat.io/container-native-virtualization/mtq-operator-rhel9@sha256:033bf7a6ef8ced1ac7275a9ddea8b16a10340a0fb0fb39619d89354d661f05b6
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:442e5cbcc63961642b662544032240bffb3863227324128a67f0bb6a05a1c364
        version: 4.99.0-0.1723448771
      entries:
      - name: kubevirt-hyperconverged-operator.v4.99.0-0.1723448771
        version: 4.99.0-0.1723448771
      name: dev-preview
    - currentCSV: kubevirt-hyperconverged-operator.v4.16.21
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:08f15657a1eee02927c5a6ef661e6e703d0064e091751bd2d8fbad5f118d9672
          createdAt: "2025-09-20 19:53:21"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.14.1 <4.16.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:88863a8dd2145bf17f7c5378c81e7c0be7a01dc47ee75bd7e5203a8fed8895b9
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.16.21-3,hyperconverged-cluster-operator-test-rhel9:v4.16.21-1,kubevirt-common-instancetypes-test-rhel9:v4.16.21-1,kubevirt-tekton-tasks-test-rhel9:v4.16.21-1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:9afe1ec9f776854caae38b89198a82e8e47a45d6c08f8dbd65a06c85f741c6a0
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:611d60b6ea4285fbc8a9f945fa5d06c09679a8f7ec1d70ec18edae7eec1a6bb3
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:299825fa33ab3f491cf954dd2e4ba23fdba2a96723b9ac3d98b3264b0b9fbb28
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:f6b075a56380143a7892e1fe5fdb212f8aa1c974f6a2d7e7e9b940789e44c010
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:d346688e02d03ddd9918f7a344d6eb3d5ddac3f1802dc82daf72bc56e7a9d496
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:406681aff7ea0a17941449593f839418edffdb073e2afffd53174f0df6292f94
        - registry.redhat.io/container-native-virtualization/mtq-controller-rhel9@sha256:253ef56b2867d33b3a475a456878b4a60f56473122eeed8c169568a9724ed30c
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:e54d87e6735151ed3c5295b23ca2f7ec666afb6421923636ef836bca1cf883d6
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:6385b346b3dd8f2bb3284871d326c851cf743a5e3fc3c49a1e51a1d43b0eafb4
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:8cd1052b9f69d0a0d8db41e90414756e83cbd07d7bd03c2faa85c41393d6539c
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:b80e68822fb0b5e78a995a97e02fe2ec448d67a5aa2224497fdf12b812182a0b
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:51a1fb0eb77e4f5c8bdfaf3b5074369b93ef63f5a7af62e22ed0789731f8dc73
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:361a48df213d52ced2a9963c2a8bde72a0981f1c497235fb918ea8c065580eeb
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:355e0946cfa963aeb5b6c3e8d476a9c3aee202baf77b95ce3f69c8e4b9e73836
        - registry.redhat.io/container-native-virtualization/mtq-lock-server-rhel9@sha256:c447251fea804340047bbaff885ad56be7a0d2612af1de26765f6cfc37da25e6
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:3daa460ed347e6eaab290b0a7fb0a08da6f210fb8be88af33f6e256511af80b6
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:dcc73f518ca7557cbd84b236367b3ecb7baab405268d54543783b4b772e14568
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:f7e5e04adb0467eb0aa4691a68b0d6485db9a52f99494097bf5b598e08f2edfb
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:2a109f94282b76ca0ff929adc35c93c0a98e34b43aba248d2975957fa149133f
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:cca121312777d2727fe2c50a14583243ce69e6e356b499f133736787522095b2
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:a0af0e5d4b142f28dc72c5636df5040c4ce92c6b12ab183d3a9bc7a9d2d2f418
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:88863a8dd2145bf17f7c5378c81e7c0be7a01dc47ee75bd7e5203a8fed8895b9
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:8cf7993326ba220f650ae2fa5383d0d6cf3e2c7f3346f0a339a9d212b8df02f6
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:edf12760e7fe4c8d37c22e172c0da270c219b7ff16d4ebb6b62729545309f6fa
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:a040696505e266e8f06a5f06b108f2db5ad3a6ed803d547d736455c1928a6a0e
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:d9560ed2758d3793139d3e13b7e5ce9d0f4ac0f8a6a8767449822c033cb2a456
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:9d997314bc7c6b30c2c3ff54b27bcbc1cdeac6dba300cab8a898b97fbce05ff7
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:dbff3d1592b192253a288fb3230a0ff125564540a30a6dd5ac881e79ed93b60a
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:9ae64f17191b98bb75f613969a9415103dad304e4fd20fb444f4fc4e57a57e0e
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:14952f38ab1d7f0abe4d54883e2c93d01ec1240ae809aefa9ca8cb4aaef080e1
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:15a971cef680f6f43e6b8ad7e0f5dab572ebacebc35cfbc6f7949a4ccd7789c2
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:d41ece8df9482ff860706edcdf27b53ac97d7c87407b3acfe8b1d9adfbf40e38
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:7581cd2330c20be2b77621e0e14da183d2b94f40fa5d3f448dede5f8c182759e
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:6fbdfdf3b9206058d5efc6653194bf6c42c07655817e3c97044da678e84b310a
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:3aa57a7ee9b342844ea4aa20bbbdf7a43b5d41f7866baa619f25473a230011f3
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:91b86228495f38d5528be1a8ffec96695ad72fafb76d2ead87f7bee25540d8b6
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:2e240514425d70039f5b48fe1908c56fae6efcc3eff93c63ec67f4aa3318a5fb
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:69388282df18b68c1d937ac98a5026b34221a54e5ce4b9cf4b89379b66623488
        - registry.redhat.io/container-native-virtualization/mtq-operator-rhel9@sha256:e29ffbe332d5da66ab4612a50ed73813ccbde92622474273b617cccd3bc24cd7
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:d253f1dbb54c70cf5874c2177bcd1cc3697381916cd4cf825bda934bfdb3d53c
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:7752f41a0f380fcd0d2652d286d4aac1ebb94a8a5fbb655b2e7058054f2eca4d
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:b20230f9a74652cf676bcf05dad81fc1ff64f9cff7232782c62cf585db48b115
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:0c5e1ee8ae0f19237a2c8b93466a7fc217a112cf1fb6dd55e2be27f95c9b66a3
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:bf7562a3bbc7655565b69fc96f2deae15729bbdef158cd90b97fc5ecc58d5bd8
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:fd2b4a337a8e45061cb6f455de3eefe27b004844c724fd821efc3b47215b74d3
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:77211068f3a8d331f1b21c64918861398f2cdbb2a3b760a5e8df51c89efea5b9
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:cddcdbf1d02b6e55cd93787991b5cf8704d0fb3a4a0bef5f58c918b57e26ee16
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:387fba10dc8f16fbfbc8c84274c5f1f5f6d01cf11f17f1b54fa0775a3300c6de
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:53ae4db672daad95d99f2955545ba335095c96630efc2a2ff19c5019a9413e87
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:63262574fa7d51d3b8c26b79e22d78f462c70239746aec6b8b3b951a02919e81
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:4046e8e862e0c55f09e9031b95e6a77af4d491593e39770c644287d44888b0bb
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:42fc0d1d112b30f65d6a48fa0442e2f41872658d2431ac8618dadf0aa7a41208
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:9e87e7d8679211a317d11e588868b40235da8c668bfb2ee10b564f6850fb91d1
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:38e767770ea0a74c4d594a382eed72bf9efa527e45ac6438acb79a1039f57fbd
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:75d76d7274f5169a414d8024125d4f0d8709888ac2b5d6121360635c7967d878
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:96092abe951f495f4f7dd0bdcb371b7efcb03a78a15e3ed8f1b37502203a6316
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:08f15657a1eee02927c5a6ef661e6e703d0064e091751bd2d8fbad5f118d9672
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:7e19b42a648cee71c60cb69fbcf6ebca05c938942e16d23f1cd61a619cffdcb6
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:878e1c8bd72cb3002457b6fbcddca99c0f4b00f6231fea52387d540b12e6329f
        version: 4.16.21
      entries:
      - name: kubevirt-hyperconverged-operator.v4.16.21
        version: 4.16.21
      - name: kubevirt-hyperconverged-operator.v4.16.17
        version: 4.16.17
      - name: kubevirt-hyperconverged-operator.v4.16.13
        version: 4.16.13
      - name: kubevirt-hyperconverged-operator.v4.16.7
        version: 4.16.7
      - name: kubevirt-hyperconverged-operator.v4.16.6
        version: 4.16.6
      - name: kubevirt-hyperconverged-operator.v4.16.5
        version: 4.16.5
      - name: kubevirt-hyperconverged-operator.v4.16.4
        version: 4.16.4
      - name: kubevirt-hyperconverged-operator.v4.16.3
        version: 4.16.3
      - name: kubevirt-hyperconverged-operator.v4.16.2
        version: 4.16.2
      - name: kubevirt-hyperconverged-operator.v4.16.1
        version: 4.16.1
      - name: kubevirt-hyperconverged-operator.v4.16.0
        version: 4.16.0
      - name: kubevirt-hyperconverged-operator.v4.15.2
        version: 4.15.2
      - name: kubevirt-hyperconverged-operator.v4.15.1
        version: 4.15.1
      - name: kubevirt-hyperconverged-operator.v4.15.0
        version: 4.15.0
      - name: kubevirt-hyperconverged-operator.v4.14.3
        version: 4.14.3
      - name: kubevirt-hyperconverged-operator.v4.14.2
        version: 4.14.2
      - name: kubevirt-hyperconverged-operator.v4.14.1
        version: 4.14.1
      - name: kubevirt-hyperconverged-operator.v4.14.0
        version: 4.14.0
      - name: kubevirt-hyperconverged-operator.v4.13.4
        version: 4.13.4
      - name: kubevirt-hyperconverged-operator.v4.13.3
        version: 4.13.3
      - name: kubevirt-hyperconverged-operator.v4.13.2
        version: 4.13.2
      - name: kubevirt-hyperconverged-operator.v4.13.1
        version: 4.13.1
      - name: kubevirt-hyperconverged-operator.v4.13.0
        version: 4.13.0
      - name: kubevirt-hyperconverged-operator.v4.12.2
        version: 4.12.2
      - name: kubevirt-hyperconverged-operator.v4.12.1
        version: 4.12.1
      - name: kubevirt-hyperconverged-operator.v4.12.0
        version: 4.12.0
      name: stable
    defaultChannel: stable
    packageName: kubevirt-hyperconverged
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: http://marketplace.redhat.com
    name: marketplace-games-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marketplace-games-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "game.marketplace.redhat.com/v1",
                "kind": "Arcade",
                "metadata": {
                  "name": "arcade-sample"
                },
                "spec": {
                  "size": 1
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Arcade is the Schema for the arcades API
            displayName: Arcade Instance
            kind: Arcade
            name: arcades.game.marketplace.redhat.com
            version: v1
        description: RHM Arcade Operator - a simple operator used to showcase how
          to build an operator around an existing web application that can can run
          on Openshift, Vanilla k8s, or on-boarded to Red Hat Marketplace.
        displayName: Arcade Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - rhm
        - red hat
        - marketplace
        - games
        - arcade
        links:
        - name: Marketplace Games Operator
          url: https://github.com/redhat-marketplace/marketplace-games-operator
        maturity: alpha
        provider:
          name: Red Hat Marketplace
          url: http://marketplace.redhat.com
        relatedImages:
        - registry.connect.redhat.com/rhm-labs/arcade-operator@sha256:d6959be6693728d2dc4cd38b69251551d23c5002cb52db2684df31a0a8f7b0ab
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        version: 1.0.0
      entries:
      - name: marketplace-games-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: marketplace-games-operator
    provider:
      name: Red Hat Marketplace
      url: http://marketplace.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Provider Name
      provider-url: https://your.domain
    name: smilecdr-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smilecdr-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.smilecdr.com/v1alpha1",
                "kind": "Smilecdr",
                "metadata": {
                  "name": "smilecdr"
                },
                "spec": {
                  "labels": {
                    "application": "smilecdr",
                    "client": "smilecdr",
                    "env": "dev",
                    "version": "one"
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Smilecdr
            name: smilecdrs.operator.smilecdr.com
            version: v1alpha1
        description: Smilecdr Operator description. TODO.
        displayName: Smilecdr Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - smilecdr-operator
        links:
        - name: Smilecdr Operator
          url: https://smilecdr-operator.domain
        maintainers:
        - email: your@email.com
          name: Maintainer Name
        maturity: alpha
        provider:
          name: Provider Name
          url: https://your.domain
        relatedImages:
        - quay.io/jibran_shaukat/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - quay.io/jibran_shaukat/smilecdr@sha256:cd6b2cec6f5b9ad9e0f0434532f8fb00b6919fcd840e686d96ce08874f2154db
        version: 0.0.1
      entries:
      - name: smilecdr-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: smilecdr-operator
    provider:
      name: Provider Name
      url: https://your.domain
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: model-builder-for-vision-certified-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-modelbuilder-for-vision-rhmp.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "Modelbuilder",
                "metadata": {
                  "name": "modelbuilder-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "backup_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "metadata_storage": {
                    "storage_class": "ibmc-block-bronze",
                    "storage_size": "20G"
                  },
                  "photo_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "500G"
                  },
                  "in_memory_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              },
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "ModelbuilderAWS",
                "metadata": {
                  "name": "modelbuilder-aws-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "postgres": {
                    "storage_size": "20G",
                    "storage_class": "gp2",
                    "backup_storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
          createdAt: "2021-08-28T12:00:00Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/model-builder-for-vision-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/model-builder-for-vision-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.0.3
          operators.operatorframework.io/builder: operator-sdk-v1.11.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/ibm-modelbuilder-for-vision/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Model Builder for Vision on AWS
            displayName: Deploy-Modelbuilder on AWS
            kind: ModelbuilderAWS
            name: modelbuildersaws.modelbuilder.com
            version: v1alpha1
          - description: Deploy Model Builder for Vision on IBM
            displayName: Deploy-Modelbuilder on IBM
            kind: Modelbuilder
            name: modelbuilders.modelbuilder.com
            version: v1alpha1
        description: IBM Services Software Model Builder for Vision is an AI training
          platform that can use the capabilities of IBM Cloud and Amazon Web Services
          and their GPUs to quickly train computer vision models for compatible mobile
          apps. IBM Services Software Inspection Workbench iOS and iPadOS app will
          be used as the exclusive UI for labeling and training. Specifically, IBM
          Model Builder for Vision will train computer vision models that can then
          be deployed for use on the IBM Services Software Inspector Portable , IBM
          Services Software Inspector Wearable and the IBM Maximo Visual Inspection
          Mobile apps.
        displayName: IBM Model Builder for Vision
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Model Builder
        - AI
        - Computer Vision
        - Inspection
        - Mobile
        - Core ML
        - TensorFlow Lite
        links:
        - name: IBM Model Builder for Vision
          url: http://ibm.biz/model_builder
        maintainers:
        - email: isshelp@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.connect.redhat.com/ibm/modelbuilder-augmentor@sha256:ef5d0dde3eb334d919c2773af63007f1ef5e735ae040ef77fffe3440f854f068
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/rhel8/redis-6@sha256:2f65f46b25d84a3cb41a6a10f93786bcab47b7354ca9ac13a8e602ea833eba60
        - registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
        - registry.connect.redhat.com/ibm/modelbuilder-image-server@sha256:ecdb665af651417c50335e7259b89ef8096521abec431976bd7abe4c1edc49d6
        - registry.connect.redhat.com/ibm/modelbuilder-auth-service@sha256:f96b444fdcf7781775c9eba942bcc5e5ebdcf5f50820447f5eb26061c4c674cd
        - registry.connect.redhat.com/ibm/modelbuilder-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/modelbuilder-cv-training-service@sha256:c3c9080b08b87022c784fd98ac3e2774ec89ce29a2dd19d91dee039e3fdb2b36
        version: 1.0.3
      entries:
      - name: ibm-modelbuilder-for-vision-rhmp.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: model-builder-for-vision-certified-rhmp
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: fuse-console
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-console.v7.10.3-0.1655306919.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "replicas": 1,
                  "version": "1.10",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "100Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:37dc38375f5d553ac1d9f20f66ce005ed9495354b81d7915ae3bd7dd6af132f3
          createdAt: "2020-09-14T10:59:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.10.3'
          olm.substitutesFor: fuse-console.v7.10.3-0.1652362834.p
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.10/html/managing_fuse
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:37dc38375f5d553ac1d9f20f66ce005ed9495354b81d7915ae3bd7dd6af132f3
        version: 7.10.3+0.1655306919.p
      entries:
      - name: fuse-console.v7.10.3
        version: 7.10.3
      - name: fuse-console.v7.10.3-0.1652362834.p
        version: 7.10.3+0.1652362834.p
      - name: fuse-console.v7.10.3-0.1655306919.p
        version: 7.10.3+0.1655306919.p
      name: 7.10.x
    - currentCSV: fuse-console.v7.11.1-0.1681224695.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.11",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:518b5ace9178e2aafd7b6f247fd41d9392abd1a7bccb4cb53783531820611193
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.11.1'
          olm.substitutesFor: fuse-console.v7.11.1-0.1675972857.p
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.11/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:518b5ace9178e2aafd7b6f247fd41d9392abd1a7bccb4cb53783531820611193
        version: 7.11.1+0.1681224695.p
      entries:
      - name: fuse-console.v7.11.1
        version: 7.11.1
      - name: fuse-console.v7.11.1-0.1669922044.p
        version: 7.11.1+0.1669922044.p
      - name: fuse-console.v7.11.1-0.1675972857.p
        version: 7.11.1+0.1675972857.p
      - name: fuse-console.v7.11.1-0.1681224695.p
        version: 7.11.1+0.1681224695.p
      name: 7.11.x
    - currentCSV: fuse-console.v7.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.12",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:4e6889bf76d323956e0b594db7381e8162fea6b108f53f9e6dd0b38cf3596c8d
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.12.1'
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.12/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:4e6889bf76d323956e0b594db7381e8162fea6b108f53f9e6dd0b38cf3596c8d
        version: 7.12.1
      entries:
      - name: fuse-console.v7.12.1
        version: 7.12.1
      name: 7.12.x
    - currentCSV: fuse-console.v7.13.0-0.1757604404.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.13",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:72d98254a821bce237b3dde896f2825628cb7b6df564e23949337590586e8bea
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.7.0 <7.13.0'
          olm.substitutesFor: fuse-console.v7.13.0-0.1754559584.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.13/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:72d98254a821bce237b3dde896f2825628cb7b6df564e23949337590586e8bea
        version: 7.13.0+0.1757604404.p
      entries:
      - name: fuse-console.v7.13.0-0.1744281421.p
        version: 7.13.0+0.1744281421.p
      - name: fuse-console.v7.13.0-0.1723709432.p
        version: 7.13.0+0.1723709432.p
      - name: fuse-console.v7.13.0-0.1757604404.p
        version: 7.13.0+0.1757604404.p
      - name: fuse-console.v7.13.0-0.1721224332.p
        version: 7.13.0+0.1721224332.p
      - name: fuse-console.v7.13.0-0.1744737026.p
        version: 7.13.0+0.1744737026.p
      - name: fuse-console.v7.13.0-0.1724659726.p
        version: 7.13.0+0.1724659726.p
      - name: fuse-console.v7.13.0-0.1725026290.p
        version: 7.13.0+0.1725026290.p
      - name: fuse-console.v7.13.0-0.1748960166.p
        version: 7.13.0+0.1748960166.p
      - name: fuse-console.v7.13.0-0.1738056254.p
        version: 7.13.0+0.1738056254.p
      - name: fuse-console.v7.13.0-0.1739885034.p
        version: 7.13.0+0.1739885034.p
      - name: fuse-console.v7.13.0-0.1741875974.p
        version: 7.13.0+0.1741875974.p
      - name: fuse-console.v7.13.0-0.1741884704.p
        version: 7.13.0+0.1741884704.p
      - name: fuse-console.v7.13.0-0.1719822976.p
        version: 7.13.0+0.1719822976.p
      - name: fuse-console.v7.13.0
        version: 7.13.0
      - name: fuse-console.v7.13.0-0.1731460809.p
        version: 7.13.0+0.1731460809.p
      - name: fuse-console.v7.13.0-0.1749640277.p
        version: 7.13.0+0.1749640277.p
      - name: fuse-console.v7.13.0-0.1751291353.p
        version: 7.13.0+0.1751291353.p
      - name: fuse-console.v7.13.0-0.1753702606.p
        version: 7.13.0+0.1753702606.p
      - name: fuse-console.v7.13.0-0.1753778626.p
        version: 7.13.0+0.1753778626.p
      - name: fuse-console.v7.13.0-0.1753785080.p
        version: 7.13.0+0.1753785080.p
      - name: fuse-console.v7.13.0-0.1753860418.p
        version: 7.13.0+0.1753860418.p
      - name: fuse-console.v7.13.0-0.1754303647.p
        version: 7.13.0+0.1754303647.p
      - name: fuse-console.v7.13.0-0.1754559584.p
        version: 7.13.0+0.1754559584.p
      - name: fuse-console.v7.13.0-0.1754567932.p
        version: 7.13.0+0.1754567932.p
      - name: fuse-console.v7.12.1
        version: 7.12.1
      name: 7.13.x
    - currentCSV: fuse-console.v7.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "replicas": 1,
                  "version": "1.9",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "100Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:846d57f3fd9f307a98b0169b6733c26299209d99ed189f219c48f36ba771b198
          createdAt: "2020-09-14T10:59:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.9.0'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.9/html/managing_fuse
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:846d57f3fd9f307a98b0169b6733c26299209d99ed189f219c48f36ba771b198
        version: 7.9.0
      entries:
      - name: fuse-console.v7.9.0
        version: 7.9.0
      name: 7.9.x
    defaultChannel: 7.13.x
    packageName: fuse-console
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: opentelemetry-product
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opentelemetry-operator.v0.135.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "Instrumentation",
                "metadata": {
                  "name": "instrumentation"
                },
                "spec": {
                  "exporter": {
                    "endpoint": "http://otel-collector-headless:4317"
                  }
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpAMPBridge",
                "metadata": {
                  "name": "opampbridge-sample"
                },
                "spec": {
                  "capabilities": {
                    "AcceptsOpAMPConnectionSettings": true,
                    "AcceptsOtherConnectionSettings": true,
                    "AcceptsRemoteConfig": true,
                    "AcceptsRestartCommand": true,
                    "ReportsEffectiveConfig": true,
                    "ReportsHealth": true,
                    "ReportsOwnLogs": true,
                    "ReportsOwnMetrics": true,
                    "ReportsOwnTraces": true,
                    "ReportsRemoteConfig": true,
                    "ReportsStatus": true
                  },
                  "componentsAllowed": {
                    "exporters": [
                      "debug"
                    ],
                    "processors": [
                      "memory_limiter"
                    ],
                    "receivers": [
                      "otlp"
                    ]
                  },
                  "endpoint": "ws://opamp-server:4320/v1/opamp"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": "receivers:\n  otlp:\n    protocols: \n      grpc: {}\n      http: {}\n\nexporters:\n  debug: {}\n\nservice:\n  pipelines:\n    traces:\n      receivers: [otlp]\n      exporters: [debug]\n"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1beta1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": {
                    "exporters": {
                      "debug": {}
                    },
                    "receivers": {
                      "otlp": {
                        "protocols": {
                          "grpc": {},
                          "http": {}
                        }
                      }
                    },
                    "service": {
                      "pipelines": {
                        "traces": {
                          "exporters": [
                            "debug"
                          ],
                          "receivers": [
                            "otlp"
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/rhosdt/opentelemetry-rhel8-operator@sha256:45c6c52744e9a17ffd883a2864cb7722ce3f9580e522c976d4f55623118f3504
          createdAt: 24 Sep 2025, 10:19
          description: Provides the OpenTelemetry components, including the Collector
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.33.0 <0.135.0-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-opentelemetry-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-telemetry/opentelemetry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instrumentation is the spec for OpenTelemetry instrumentation.
            displayName: OpenTelemetry Instrumentation
            kind: Instrumentation
            name: instrumentations.opentelemetry.io
            version: v1alpha1
          - description: OpAMPBridge is the Schema for the opampbridges API.
            displayName: OpAMP Bridge
            kind: OpAMPBridge
            name: opampbridges.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1beta1
          - description: TargetAllocator is the Schema for the targetallocators API.
            displayName: Target Allocator
            kind: TargetAllocator
            name: targetallocators.opentelemetry.io
            version: v1alpha1
        description: |-
          Red Hat build of OpenTelemetry is a collection of tools, APIs, and SDKs. You use it to instrument, generate, collect, and export telemetry data (metrics, logs, and traces) for analysis in order to understand your software's performance and behavior.
          This operator was previously called Red Hat OpenShift distributed tracing data collection.

          ### Operator features

          * **Sidecar injection** - annotate your pods and let the operator inject a sidecar.
          * **Managed upgrades** - updating the operator will automatically update your OpenTelemetry collectors.
          * **Deployment modes** - your collector can be deployed as sidecar, daemon set, or regular deployment.
          * **Service port management** - the operator detects which ports need to be exposed based on the provided configuration.

          ## Examples

          There is a list of examples to help you deploy different configurations in [this GitHub repository](https://github.com/os-observability/redhat-rhosdt-samples)

          ### Support & Troubleshooting

          Red Hat build of OpenTelemetry is available as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/4.17/html/red_hat_build_of_opentelemetry/index) is recommended when installing, configuring, and managing the Operator and instances.
        displayName: Red Hat build of OpenTelemetry
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opentelemetry
        - tracing
        - logging
        - metrics
        - monitoring
        - troubleshooting
        links:
        - name: OpenTelemetry Operator
          url: https://github.com/open-telemetry/opentelemetry-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/opentelemetry-rhel8-operator@sha256:45c6c52744e9a17ffd883a2864cb7722ce3f9580e522c976d4f55623118f3504
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5a6edfecf1636805e6f0bcad02df10bceec764f0580f3ac8fe62aba978d50006
        - registry.redhat.io/rhosdt/opentelemetry-collector-rhel8@sha256:5f31caf1f4a12cf2832ac3e84877f4fb0c856073820498c8a5be13698348c38a
        - registry.redhat.io/rhosdt/opentelemetry-target-allocator-rhel8@sha256:f77a26edafba06a4fca1a07083e06a908664852cf228b1814455ffa8155cc979
        version: 0.135.0-1
      entries:
      - name: opentelemetry-operator.v0.135.0-1
        version: 0.135.0-1
      - name: opentelemetry-operator.v0.127.0-2
        version: 0.127.0-2
      - name: opentelemetry-operator.v0.127.0-1
        version: 0.127.0-1
      - name: opentelemetry-operator.v0.119.0-2
        version: 0.119.0-2
      - name: opentelemetry-operator.v0.119.0-1
        version: 0.119.0-1
      - name: opentelemetry-operator.v0.113.0-2
        version: 0.113.0-2
      - name: opentelemetry-operator.v0.113.0-1
        version: 0.113.0-1
      - name: opentelemetry-operator.v0.107.0-5
        version: 0.107.0-5
      - name: opentelemetry-operator.v0.107.0-4-0.1727168137.p
        version: 0.107.0-4+0.1727168137.p
      - name: opentelemetry-operator.v0.107.0-4
        version: 0.107.0-4
      - name: opentelemetry-operator.v0.102.0-3
        version: 0.102.0-3
      - name: opentelemetry-operator.v0.102.0-2
        version: 0.102.0-2
      - name: opentelemetry-operator.v0.100.1-3
        version: 0.100.1-3
      - name: opentelemetry-operator.v0.93.0-3
        version: 0.93.0-3
      - name: opentelemetry-operator.v0.93.0-2
        version: 0.93.0-2
      - name: opentelemetry-operator.v0.89.0-3
        version: 0.89.0-3
      - name: opentelemetry-operator.v0.81.1-5
        version: 0.81.1-5
      - name: opentelemetry-operator.v0.74.0-5-0.1687199949.p
        version: 0.74.0-5+0.1687199949.p
      - name: opentelemetry-operator.v0.74.0-5
        version: 0.74.0-5
      - name: opentelemetry-operator.v0.60.0-2
        version: 0.60.0-2
      - name: opentelemetry-operator.v0.56.0-1
        version: 0.56.0-1
      name: stable
    defaultChannel: stable
    packageName: opentelemetry-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://www.medik8s.io/
    name: customized-user-remediation
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: customized-user-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediation",
                "metadata": {
                  "name": "customizeduserremediation-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediationConfig",
                "metadata": {
                  "name": "customizeduserremediationconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediationTemplate",
                "metadata": {
                  "name": "customizeduserremediationtemplate-sample",
                  "namespace": "openshift-workload-availability"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/customized-user-remediation-operator:v0.1.0
          createdAt: "2024-03-25 16:17:57"
          description: Customized User Remediation Operator will remediate unhealthy
            nodes by running a pre defined user script on the unhealthy node. Since
            the user has full control over the script content, is may be quite risky
            so apply discretion.
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/customized-user-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CustomizedUserRemediationConfig is the Schema for the customizeduserremediationconfigs
              API
            displayName: Customized User Remediation Config
            kind: CustomizedUserRemediationConfig
            name: customizeduserremediationconfigs.customized-user-remediation.medik8s.io
            version: v1alpha1
          - description: CustomizedUserRemediation is the Schema for the customizeduserremediations
              API
            displayName: Customized User Remediation
            kind: CustomizedUserRemediation
            name: customizeduserremediations.customized-user-remediation.medik8s.io
            version: v1alpha1
          - description: CustomizedUserRemediationTemplate is the Schema for the customizeduserremediationtemplates
              API
            displayName: Customized User Remediation Template
            kind: CustomizedUserRemediationTemplate
            name: customizeduserremediationtemplates.customized-user-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Customized User Remediation (CUR) works with Node Health Check to remediate unhealthy nodes by running a user customized script on the unhealthy node.
          This operator does not preform any fencing on the remediated node so users should be careful with the scripts that are being used.

          ### Compatibility
          CUR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Flexibility - CUR is quite flexible sine it runs in the remediation process a user customized script.
        displayName: Customized User Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - script
        - CUR
        links:
        - name: Customized User Remediation
          url: https://customized-user-remediation.domain
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Medik8s
          url: https://www.medik8s.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/medik8s/customized-user-remediation-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: customized-user-remediation.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: customized-user-remediation
    provider:
      name: Medik8s
      url: https://www.medik8s.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-loadcore-agents
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-loadcore-agents.v0.2.25
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2024-11-20T18:10:53Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <=10.0.0'
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Agents
            name: agents.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight LoadCore Agents Operator
        displayName: Keysight LoadCore Agents Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Loadcore Agents
          url: https://keysight-loadcore-agents.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:7efeeb8b29872a6f0271f651d7ae02c91daea16d853c50e374c310f044d8c76c
        - prod.harbor.keysight.digital/loadcore/keysight-loadcore-agents@sha256:96e18b6270097c23bad5445a15613e467d06c6fbdbaf39545edb79915aacc04f
        version: 0.2.25
      entries:
      - name: keysight-loadcore-agents.v0.2.25
        version: 0.2.25
      name: alpha
    defaultChannel: alpha
    packageName: keysight-loadcore-agents
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-secret-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secret-operator.v24.11.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/tools@sha256:47204b2697dfbab5eaf9d6dcb4c160dd427a148d6b840727239b773810e74fc9
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:161e754437428c68490333337303fb30b695a9ac677c3a00a1dc27565b855ab8
        - quay.io/stackable/secret-operator@sha256:1267311830649f928b528512a3cdc843efad6dd32b851d99c84639fbe70e35a3
        - quay.io/stackable/tools@sha256:47204b2697dfbab5eaf9d6dcb4c160dd427a148d6b840727239b773810e74fc9
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:a6d6631d35c32b874c995b6af889636ac8dd1f39dcb68810d9f9b10a4e2c1987
        version: 24.11.1
      entries:
      - name: secret-operator.v24.11.1
        version: 24.11.1
      name: "24.11"
    - currentCSV: secret-operator.v25.3.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:6eb08b99259d62514d87844e1b7470382d500696fecc9a1232f6bb536ed897fa
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/secret-operator@sha256:6eb08b99259d62514d87844e1b7470382d500696fecc9a1232f6bb536ed897fa
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:dd730457133f619d8759269abcfa79d7aeb817e01ba7af2d3aa1417df406ea56
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        version: 25.3.0
      entries:
      - name: secret-operator.v25.3.0
        version: 25.3.0
      name: "25.3"
    - currentCSV: secret-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
          - description: Trust stores managed by the operator.
            displayName: Trust stores managed by the operator.
            kind: TrustStore
            name: truststores.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        version: 25.7.0
      entries:
      - name: secret-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: secret-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
          - description: Trust stores managed by the operator.
            displayName: Trust stores managed by the operator.
            kind: TrustStore
            name: truststores.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        version: 25.7.0
      entries:
      - name: secret-operator.v25.7.0
        version: 25.7.0
      - name: secret-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-secret-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zettaset
      provider-url: ""
    name: xcrypt-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zts-xcrypt-operator.v0.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xcrypt.zettaset.com/v1alpha1",
                "kind": "XCrypt",
                "metadata": {
                  "name": "zts-masterset"
                },
                "spec": {
                  "status": {
                    "podNames": "zts-ca",
                    "replicas": "1,"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
          createdAt: "2021-06-16T12:29:13.97Z"
          description: XCrypt Operator deploys Zettaset XCrypt Container Encryption  -
            the only software-defined solution that offers transparent high performance
            data-at-rest encryption of critical data in container environments.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/xcrypt-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/xcrypt-operator-rhmp/support?utm_source=openshift_console
          support: Zettaset
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XCrypt is the Schema for the xcrypts API
            kind: XCrypt
            name: xcrypts.xcrypt.zettaset.com
            version: v1alpha1
        description: "Zettaset’s RedHat Certified XCrypt Operator allows users to
          deploy multiple XCrypt components on OpenShift with just a few simple commands.
          The XCrypt Operator deployment provides software-defined encryption that
          transparently protects container data throughout your OpenShift environment.\n\n##
          Core Features\n* **Automated XCrypt provisioning** - Deploying a Zettaset
          XCrypt has never been easier. Adjust cluster specific configuration and
          let the Zettaset Operator take care of provisioning XCrypt services and
          setting up the XCrypt encryption to your exact specification. \n* **CSI
          Driver deployment**: Zettaset XCrypt Operator automatically deploys Zettaset
          XCrypt CSI Driver in order to provide automated and transparent data-at-rest
          \ encryption functionality compatible with Openshift and any other container
          orchestrator. Once the Operator is provisioned, the driver runs fine tuned
          and ready to work with other XCrypt components.\n* **Multi-node topology**:
          Encrypt as many nodes and volumes as you need to. XCrypt components can
          be installed on multiple nodes across the cluster. This is defined by two
          labels: zts-master and zts-worker. Normally zts-master node is a single
          node that runs Zettaset XCrypt major services required for the product operation.
          Multiple nodes can be marked as zts-worker and those are the nodes that
          can have encrypted volumes.\n* **Seamless Encrypted Volume mount**: Volumes
          are not only being encrypted, but also are mounted/unmounted to containers
          with Persistent Volume Claims. This happens as seamlessly as regular volumes
          provision and mount/unmount.\n* **Robust failover**: A node failure doesn’t
          affect XCrypt operations on other nodes of the cluster. Encryption on a
          restarted/recovered node gets back to normal functioning if the node stays
          healthy and data hasn’t been corrupted during the restart\n\n## About Zettaset
          XCrypt Kubernetes Encryption for OpenShift\nXCrypt Kubernetes Encryption
          for OpenShift is a software-only solution that enables encryption of data
          at rest stored in Red Hat OpenShift container environments without slowing
          down the DevSecOps process. In addition to providing a transparent and high-performance
          layer of security that flexibly protects container data across on-premise,
          in hybrid, and multi-cloud environments, it also helps your organization
          efficiently transition from DevOps to DevSecOps.  \n\nMore information about
          Zettaset XCrypt Kubernetes Encryption for OpenShift can be found in the
          [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)\n\n##
          RedHat OpenShift + Zettaset: The next generation of secure enterprise infrastructure
          for DevSecOps  \nBy combining RedHat’s Container Platform with Zettaset’s
          certified software-defined encryption, you can create the next generation
          of secure enterprise infrastructure that DevSecOps requires. Together they
          create the stable, consistent, and supported base that your organization
          needs to develop applications faster, simplify updates, drive innovation,
          and reduce the risk of potential security breaches and data theft.\n###
          Core Features\n* Software-only for simple and scalable deployments \n* Negligible
          performance impact\n* Transparent to developers and administrators \n* Works
          with AWS, Google, and Azure \n* Unique encryption key per container volume
          \n* Encrypted volumes are only available when in use\n* Direct integration
          with OpenShift \n* Automated encryption policy management Secure erase of
          volumes\n* Ability to securely decommission a node\n\n## How does XCrypt
          Kubernetes Encryption fit into OpenShift? \nXCrypt XCrypt Kubernetes Encryption
          for OpenShift Encryption makes it incredibly easy to protect container data
          in your OpenShift environment. \n### Use cases include:\n* Transitioning
          from DevOps to DevSecOps \n* Ensuring data protection in OpenShift container
          environments with negligible performance impact \n* Achieving compliance
          in regulated industries like healthcare and finance\n\n## Other Information\nRole-based
          access control needs to be configured prior to the Operator installation.
          Please find details in the [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)
          \   "
        displayName: Zettaset XCrypt Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - encryption
        - storage
        links:
        - name: Zettaset
          url: https://www.zettaset.com/
        maintainers:
        - email: info@zettaset.com
          name: Zettaset
        maturity: alpha
        provider:
          name: Zettaset
        relatedImages:
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-ca-1@sha256:9f1015dfb459b69f5dbf449efa0517452095a6cf4f20826ebae278f9f800ffe7
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-kmip-server-1@sha256:9af5d21dd748dfbf4e107f27a5f10fa3c0ba928ade9b989a744b620ce34ec757
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-license-server-1@sha256:bc5486892a41e70f5dc4c4e1d82fca8db2149598fb8b5950ba5da1ee2d9d245d
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-host-manager-1@sha256:9faa4b597eb74df412761dffa057b64fb69dbf9fa016414962eb2099a343e734
        - registry.marketplace.redhat.com/rhm/zts/zts-csi-driver-1@sha256:c15fa9a7cc40891103f426206663dcebe1210c8ee3ab6037cbf64e1f86b9c7e5
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
        version: 0.0.17
      entries:
      - name: zts-xcrypt-operator.v0.0.17
        version: 0.0.17
      name: alpha
    defaultChannel: alpha
    packageName: xcrypt-operator-rhmp
    provider:
      name: Zettaset
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: openshift-secondary-scheduler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secondaryscheduleroperator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "SecondaryScheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-secondary-scheduler-operator"
                },
                "spec": {
                  "managementState": "Managed",
                  "schedulerConfig": "secondary-scheduler-config"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-rhel9-operator@sha256:210cc51760c8d3a9e1e5ba2b41f575b667e0d541b995362426e5a28f3f666600
          createdAt: 2024/10/01
          description: Runs a secondary scheduler in an OpenShift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.2.0 <1.3.1'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/secondary-scheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SecondaryScheduler is the schema for the secondaryschedulers
              API
            displayName: Secondary Scheduler
            kind: SecondaryScheduler
            name: secondaryschedulers.operator.openshift.io
            version: v1
        description: |
          The Secondary Scheduler Operator provides the ability to use a customized scheduler image that was developed using the scheduler plugin framework as a secondary scheduler in OpenShift.

          ## Secondary Scheduler Configuration

          Once the Secondary Scheduler Operator is installed, you can configure the plugins to run for your secondary scheduler.

          The configuration of the secondary scheduler is defined through a config map that wraps the KubeSchedulerConfiguration YAML file in "config.yaml".

          ## Additional Parameters

          Additionally, the following parameters can be configured:

          * `schedulerConfig` - Set the config map configuration for the secondary scheduler.
          * `schedulerImage` - Set the default operand image for the secondary scheduler.
        displayName: Secondary Scheduler Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - secondary
        - scheduling
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/secondary-scheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-rhel9-operator@sha256:210cc51760c8d3a9e1e5ba2b41f575b667e0d541b995362426e5a28f3f666600
        version: 1.3.1
      entries:
      - name: secondaryscheduleroperator.v1.3.1
        version: 1.3.1
      - name: secondaryscheduleroperator.v1.3.0
        version: 1.3.0
      - name: secondaryscheduleroperator.v1.2.2
        version: 1.2.2
      - name: secondaryscheduleroperator.v1.2.1
        version: 1.2.1
      - name: secondaryscheduleroperator.v1.2.0
        version: 1.2.0
      - name: secondaryscheduleroperator.v1.1.4
        version: 1.1.4
      - name: secondaryscheduleroperator.v1.1.2
        version: 1.1.2
      name: stable
    defaultChannel: stable
    packageName: openshift-secondary-scheduler-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: costmanagement-metrics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: costmanagement-metrics-operator.4.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "true"
          containerImage: registry.redhat.io/costmanagement/costmanagement-metrics-rhel9-operator@sha256:c5bb5291052d57390b9f4bdc660c1926827072b6dcbe09d85f5b6fc375562182
          createdAt: "2025-08-18T12:10:15Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: costmanagement-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Cost Management Metrics Operator
          ## Introduction
          The `costmanagement-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Cost Management Metrics Operator (`costmanagement-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created CostManagementMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The CostManagementMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **FIPS Compliance:** Supports deployment in high-security environments when run on an OpenShift cluster with FIPS mode enabled.

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created CostManagementMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the CostManagementMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `CostManagementMetricsConfig`, creating a new `CostManagementMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `costmanagement-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: costmanagement-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the CostManagementMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the CostManagementMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon CostManagementMetricsConfig creation** (This parameter will start to appear in CostManagementMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the CostManagementMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the costmanagement-metrics-operator
          **Note** There are separate instructions for configuring the `costmanagement-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `costmanagement-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the CostManagementMetricsConfig
          Configure the costmanagement-metrics-operator by creating a `CostManagementMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `costmanagement-metrics-operator` -> `CostManagementMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the costmanagement-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `costmanagement-metrics-operator-data` with 10Gi of storage. To configure the costmanagement-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `costmanagement-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `costmanagement-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the costmanagement-metrics-operator for a restricted network
          ##### Create the CostManagementMetricsConfig
          Configure the costmanagement-metrics-operator by creating a `CostManagementMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `costmanagement-metrics-operator` -> `CostManagementMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `costmanagement-metrics-operator-data` with 10Gi of storage. To configure the costmanagement-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `costmanagement-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: costmanagement-metrics-operator
                labels:
                  app: costmanagement-metrics-operator
              spec:
                volumes:
                - name: costmanagement-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: costmanagement-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: costmanagement-metrics-operator-reports
                    mountPath: /tmp/costmanagement-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/costmanagement-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/costmanagement-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `costmanagement-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the CostManagementMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Cost Management Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Cost Management Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/costmanagement/costmanagement-metrics-rhel9-operator@sha256:c5bb5291052d57390b9f4bdc660c1926827072b6dcbe09d85f5b6fc375562182
        version: 4.1.0
      entries:
      - name: costmanagement-metrics-operator.4.1.0
        version: 4.1.0
      - name: costmanagement-metrics-operator.4.0.0
        version: 4.0.0
      - name: costmanagement-metrics-operator.3.3.2
        version: 3.3.2
      - name: costmanagement-metrics-operator.3.3.1
        version: 3.3.1
      - name: costmanagement-metrics-operator.3.3.0
        version: 3.3.0
      - name: costmanagement-metrics-operator.3.2.1
        version: 3.2.1
      - name: costmanagement-metrics-operator.3.2.0
        version: 3.2.0
      - name: costmanagement-metrics-operator.3.1.1
        version: 3.1.1
      - name: costmanagement-metrics-operator.3.1.0
        version: 3.1.0
      - name: costmanagement-metrics-operator.3.0.1
        version: 3.0.1
      - name: costmanagement-metrics-operator.3.0.0
        version: 3.0.0
      - name: costmanagement-metrics-operator.2.0.0
        version: 2.0.0
      - name: costmanagement-metrics-operator.1.2.0
        version: 1.2.0
      - name: costmanagement-metrics-operator.1.1.9
        version: 1.1.9
      - name: costmanagement-metrics-operator.1.1.8
        version: 1.1.8
      - name: costmanagement-metrics-operator.1.1.7
        version: 1.1.7
      - name: costmanagement-metrics-operator.1.1.6
        version: 1.1.6
      - name: costmanagement-metrics-operator.1.1.5
        version: 1.1.5
      - name: costmanagement-metrics-operator.1.1.4
        version: 1.1.4
      - name: costmanagement-metrics-operator.1.1.3
        version: 1.1.3
      - name: costmanagement-metrics-operator.1.1.0
        version: 1.1.0
      - name: costmanagement-metrics-operator.1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: costmanagement-metrics-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-applicationautoscaling-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-applicationautoscaling-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "applicationautoscaling.services.k8s.aws/v1alpha1",
                "kind": "ScalableTarget",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "applicationautoscaling.services.k8s.aws/v1alpha1",
                "kind": "ScalingPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/applicationautoscaling-controller:1.1.1
          createdAt: "2025-09-20T21:43:01Z"
          description: AWS Application Auto Scaling controller is a service controller
            for managing Application Auto Scaling resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ScalableTarget represents the state of an AWS applicationautoscaling
              ScalableTarget resource.
            displayName: ScalableTarget
            kind: ScalableTarget
            name: scalabletargets.applicationautoscaling.services.k8s.aws
            version: v1alpha1
          - description: ScalingPolicy represents the state of an AWS applicationautoscaling
              ScalingPolicy resource.
            displayName: ScalingPolicy
            kind: ScalingPolicy
            name: scalingpolicies.applicationautoscaling.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Application Auto Scaling resources in AWS from within your Kubernetes cluster.
          **About Amazon Application Auto Scaling**
          Application Auto Scaling is a web service for developers and system administrators who need a solution for automatically scaling their scalable resources for individual AWS services beyond Amazon EC2. Application Auto Scaling allows you to configure automatic scaling for the following resources:
          - AppStream 2.0 fleets
          - Aurora replicas
          - Amazon Comprehend document classification and entity recognizer endpoints
          - DynamoDB tables and global secondary indexes
          - Amazon Elastic Container Service (ECS) services
          - Amazon EMR clusters
          - Amazon Keyspaces (for Apache Cassandra) tables
          - Lambda function provisioned concurrency
          - Amazon Managed Streaming for Apache Kafka (MSK) broker storage
          - SageMaker endpoint variants
          - Spot Fleet requests
          - Custom resources provided by your own applications or services. For more information, see the [GitHub repository](https://github.com/aws/aws-auto-scaling-custom-resource)

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Application Auto Scaling
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - applicationautoscaling
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Application Auto Scaling User Guide
          url: https://docs.aws.amazon.com/autoscaling/application/userguide/what-is-application-auto-scaling.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: application auto scaling maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/applicationautoscaling-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-applicationautoscaling-controller.v1.1.1
        version: 1.1.1
      - name: ack-applicationautoscaling-controller.v1.1.0
        version: 1.1.0
      - name: ack-applicationautoscaling-controller.v1.0.30
        version: 1.0.30
      - name: ack-applicationautoscaling-controller.v1.0.29
        version: 1.0.29
      - name: ack-applicationautoscaling-controller.v1.0.27
        version: 1.0.27
      - name: ack-applicationautoscaling-controller.v1.0.26
        version: 1.0.26
      - name: ack-applicationautoscaling-controller.v1.0.25
        version: 1.0.25
      - name: ack-applicationautoscaling-controller.v1.0.24
        version: 1.0.24
      - name: ack-applicationautoscaling-controller.v1.0.23
        version: 1.0.23
      - name: ack-applicationautoscaling-controller.v1.0.22
        version: 1.0.22
      - name: ack-applicationautoscaling-controller.v1.0.21
        version: 1.0.21
      - name: ack-applicationautoscaling-controller.v1.0.19
        version: 1.0.19
      - name: ack-applicationautoscaling-controller.v1.0.18
        version: 1.0.18
      - name: ack-applicationautoscaling-controller.v1.0.17
        version: 1.0.17
      - name: ack-applicationautoscaling-controller.v1.0.16
        version: 1.0.16
      - name: ack-applicationautoscaling-controller.v1.0.15
        version: 1.0.15
      - name: ack-applicationautoscaling-controller.v1.0.11
        version: 1.0.11
      - name: ack-applicationautoscaling-controller.v1.0.9
        version: 1.0.9
      - name: ack-applicationautoscaling-controller.v1.0.8
        version: 1.0.8
      - name: ack-applicationautoscaling-controller.v1.0.7
        version: 1.0.7
      - name: ack-applicationautoscaling-controller.v1.0.6
        version: 1.0.6
      - name: ack-applicationautoscaling-controller.v1.0.5
        version: 1.0.5
      - name: ack-applicationautoscaling-controller.v1.0.4
        version: 1.0.4
      - name: ack-applicationautoscaling-controller.v1.0.3
        version: 1.0.3
      - name: ack-applicationautoscaling-controller.v1.0.2
        version: 1.0.2
      - name: ack-applicationautoscaling-controller.v1.0.0
        version: 1.0.0
      - name: ack-applicationautoscaling-controller.v0.2.16
        version: 0.2.16
      - name: ack-applicationautoscaling-controller.v0.2.15
        version: 0.2.15
      - name: ack-applicationautoscaling-controller.v0.2.14
        version: 0.2.14
      - name: ack-applicationautoscaling-controller.v0.2.13
        version: 0.2.13
      - name: ack-applicationautoscaling-controller.v0.2.12
        version: 0.2.12
      - name: ack-applicationautoscaling-controller.v0.2.10
        version: 0.2.10
      - name: ack-applicationautoscaling-controller.v0.2.9
        version: 0.2.9
      - name: ack-applicationautoscaling-controller.v0.2.8
        version: 0.2.8
      - name: ack-applicationautoscaling-controller.v0.2.7
        version: 0.2.7
      - name: ack-applicationautoscaling-controller.v0.2.6
        version: 0.2.6
      - name: ack-applicationautoscaling-controller.v0.2.5
        version: 0.2.5
      - name: ack-applicationautoscaling-controller.v0.2.4
        version: 0.2.4
      - name: ack-applicationautoscaling-controller.v0.2.3
        version: 0.2.3
      - name: ack-applicationautoscaling-controller.v0.2.2
        version: 0.2.2
      - name: ack-applicationautoscaling-controller.v0.2.1
        version: 0.2.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-applicationautoscaling-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-acm-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-acm-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acm.services.k8s.aws/v1alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/acm-controller:1.1.1
          createdAt: "2025-09-20T01:19:56Z"
          description: AWS ACM controller is a service controller for managing ACM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Certificate represents the state of an AWS acm Certificate
              resource.
            displayName: Certificate
            kind: Certificate
            name: certificates.acm.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage AWS Certificate Manager (ACM) resources in AWS from within your Kubernetes cluster.

          **About Amazon ACM**

          AWS Certificate Manager (ACM) handles the complexity of creating, storing, and renewing public and private SSL/TLS X.509 certificates and keys that protect your AWS websites and applications. You can provide certificates for your [integrated AWS services](https://docs.aws.amazon.com/acm/latest/userguide/acm-services.html) either by issuing them directly with ACM or by [importing](https://docs.aws.amazon.com/acm/latest/userguide/import-certificate.html) third-party certificates into the ACM management system. ACM certificates can secure singular domain names, multiple specific domain names, wildcard domains, or combinations of these. ACM wildcard certificates can protect an unlimited number of subdomains. You can also [export](https://docs.aws.amazon.com/acm/latest/userguide/export-private.html) ACM certificates signed by AWS Private CA for use anywhere in your internal PKI.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon ACM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ACM Developer Resources
          url: https://aws.amazon.com/certificate-manager/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: acm maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/acm-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-acm-controller.v1.1.1
        version: 1.1.1
      - name: ack-acm-controller.v1.1.0
        version: 1.1.0
      - name: ack-acm-controller.v1.0.13
        version: 1.0.13
      - name: ack-acm-controller.v1.0.12
        version: 1.0.12
      - name: ack-acm-controller.v1.0.11
        version: 1.0.11
      - name: ack-acm-controller.v1.0.10
        version: 1.0.10
      - name: ack-acm-controller.v1.0.9
        version: 1.0.9
      - name: ack-acm-controller.v1.0.8
        version: 1.0.8
      - name: ack-acm-controller.v1.0.7
        version: 1.0.7
      - name: ack-acm-controller.v1.0.6
        version: 1.0.6
      - name: ack-acm-controller.v1.0.5
        version: 1.0.5
      - name: ack-acm-controller.v1.0.4
        version: 1.0.4
      - name: ack-acm-controller.v1.0.1
        version: 1.0.1
      - name: ack-acm-controller.v1.0.0
        version: 1.0.0
      - name: ack-acm-controller.v0.0.20
        version: 0.0.20
      - name: ack-acm-controller.v0.0.19
        version: 0.0.19
      - name: ack-acm-controller.v0.0.18
        version: 0.0.18
      - name: ack-acm-controller.v0.0.17
        version: 0.0.17
      - name: ack-acm-controller.v0.0.16
        version: 0.0.16
      - name: ack-acm-controller.v0.0.12
        version: 0.0.12
      - name: ack-acm-controller.v0.0.10
        version: 0.0.10
      - name: ack-acm-controller.v0.0.9
        version: 0.0.9
      - name: ack-acm-controller.v0.0.6
        version: 0.0.6
      - name: ack-acm-controller.v0.0.5
        version: 0.0.5
      - name: ack-acm-controller.v0.0.4
        version: 0.0.4
      - name: ack-acm-controller.v0.0.2
        version: 0.0.2
      - name: ack-acm-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-acm-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ricardo Maraschini
      provider-url: https://github.com/ricardomaraschini/tagger
    name: tagger
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tagger.v2.1.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tagger.dev/v1beta1",
                "kind": "Tag",
                "metadata": {
                  "name": "nginx"
                },
                "spec": {
                  "from": "docker.io/library/nginx:latest",
                  "generation": 0,
                  "mirror": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          containerImage: quay.io/tagger/operator:v2.1.17
          createdAt: "2021-08-02T00:00:00Z"
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/ricardomaraschini/tagger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tag is a Container Image representation as a Custom Resource.
            displayName: Tag
            kind: Tag
            name: tags.tagger.dev
            version: v1beta1
        description: |-
          A Container Image Orchestrator. Adds mirroring and Container Image manipulation capabilities
          to a Kubernetes cluster.

          Keeping track of all Container Images in use in a Kubernetes cluster is a complicated task.
          Container Images may come from numerous different Image Registries. In some cases, controlling
          how a stable version of a given Container Image looks escapes the user's authority. To add to
          this, Container Runtimes rely on remote registries (from the cluster's point of view) when
          obtaining Container Images, potentially making the process of pulling their blobs (manifests,
          config, and layers) slower.

          The notion of indexing Container Image versions by Tags is helpful. Still, it does not provide
          users with the right confidence to always use the intended Container Image – today's "latest"
          tag might not be tomorrow's "latest" tag. In addition to that, these Image Registries allow
          access to Container Images by their Manifest content's hash (i.e., usually sha256), which gives
          users the confidence at a cost in semantics.

          When releasing a new version of an application to Push and to Deploy are split into two
          distinct steps. Both the pusher and the puller need access to the same Image Registry, adding
          complexity.  Credentials are one example of the concerns. Other factors may pop up when
          running, for instance, in an air-gapped environment, where the cluster may not reach external
          Image Registries.

          Tagger aims to overcome these caveats by providing an image management abstraction layer. For
          instance, by providing a direct mapping between a Container Image tag (e.g., "latest") and its
          correspondent Manifest content's hash, users can then refer to the Container Image by its tag
          – and be sure to use that specific version. More than that, when allied with an Internal Image
          Registry, Tagger can also automatically mirror Container Images into the cluster.

          While using Tagger, Deployments can refer to Container Images by an arbitrarily defined name,
          such as "my-app", and Tagger will make sure that they use the right Container Image through its
          internal "tag to Manifest content's hash" mapping.

          For each new "release" of a given Container Image, Tagger creates a new Generation for it,
          making it easy to roll back to previously pushed "releases" of the same Container Image in case
          of problems.

          When integrated with an Internal Registry, Tagger allows users to push or pull Images directly
          without requiring an external Image Registry. It works as a layer between the user and the
          Internal Registry. Every time a new "release" of Container Images is pushed, all Deployments
          are updated automatically. Users don't need to know about the Internal Registry existence, if
          they are logged-in to the Kubernetes cluster, they can obtain old or provide new "Generations"
          for a Container Image.

          In summary, Tagger mirrors remote Container Images into a Kubernetes cluster indexing them in
          different Generations (allowing easy navigation through these multiple Generations), provides
          an interface allowing users to pull and push images directly to the Kubernetes cluster and
          provides full integration with Kubernetes Deployments (automatic triggers new rollouts on
          Container Image changes).

          By default Tagger won't be able to mirror until you provide it with a mirror registry config.
          After install you can configure the mirror by editing the Secret mirror-registry-config in the
          operator namespace. Follow below an example of a valid mirror-registry-config secret (you gonna
          have to provide your own address, username, password, etc).

          ```yaml
          apiVersion: v1
          kind: Secret
          metadata:
            name: mirror-registry-config
          data:
            address: cmVnaXN0cnkuaW8=
            username: YWRtaW4=
            password: d2hhdCB3ZXJlIHlvdSB0aGlua2luZz8K
            token: YW4gb3B0aW9uYWwgdG9rZW4gZm9yIHRva2VuIGJhc2VkIGF1dGg=
            insecure: dHJ1ZQ==
          ```

          To make pushing an pulling images easier you should go ahead and grab the `kubectl` plugin
          available [here](https://github.com/ricardomaraschini/tagger/releases). You can download a
          full version of the README with usage examples and configuration options
          [here](https://github.com/ricardomaraschini/tagger/releases/download/v2.1.17/README.pdf).
        displayName: Tagger
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tagger
        - mirroring
        - image
        links:
        - name: Documentation
          url: https://github.com/ricardomaraschini/tagger/releases/download/v2.1.17/README.pdf
        - name: Plugin (kubectl)
          url: https://github.com/ricardomaraschini/tagger/releases/tag/v2.1.17
        maintainers:
        - email: rmarasch@redhat.com
          name: Ricardo Maraschini
        maturity: alpha
        provider:
          name: Ricardo Maraschini
          url: https://github.com/ricardomaraschini/tagger
        relatedImages:
        - quay.io/tagger/operator:v2.1.17
        version: 2.1.17
      entries:
      - name: tagger.v2.1.17
        version: 2.1.17
      name: alpha
    defaultChannel: alpha
    packageName: tagger
    provider:
      name: Ricardo Maraschini
      url: https://github.com/ricardomaraschini/tagger
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecs-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "Service",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "TaskDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecs-controller:1.1.1
          createdAt: "2025-09-19T21:02:44Z"
          description: AWS ECS controller is a service controller for managing ECS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster represents the state of an AWS ecs Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.ecs.services.k8s.aws
            version: v1alpha1
          - description: Service represents the state of an AWS ecs Service resource.
            displayName: Service
            kind: Service
            name: services.ecs.services.k8s.aws
            version: v1alpha1
          - description: TaskDefinition represents the state of an AWS ecs TaskDefinition
              resource.
            displayName: TaskDefinition
            kind: TaskDefinition
            name: taskdefinitions.ecs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Container Service (ECS) resources in AWS from within your Kubernetes cluster.

          **About Amazon ECS**

          Amazon Elastic Container Service (Amazon ECS) is a fully managed container orchestration service that helps you easily deploy, manage, and scale containerized applications. As a fully managed service, Amazon ECS comes with AWS configuration and operational best practices built-in. It's integrated with both AWS and third-party tools, such as Amazon Elastic Container Registry and Docker. This integration makes it easier for teams to focus on building the applications, not the environment. You can run and scale your container workloads across AWS Regions in the cloud, and on-premises, without the complexity of managing a control plane.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECS Developer Resources
          url: https://aws.amazon.com/ecs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecs-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ecs-controller.v1.1.1
        version: 1.1.1
      - name: ack-ecs-controller.v1.0.15
        version: 1.0.15
      - name: ack-ecs-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecs-controller.v1.0.13
        version: 1.0.13
      - name: ack-ecs-controller.v1.0.12
        version: 1.0.12
      - name: ack-ecs-controller.v1.0.11
        version: 1.0.11
      - name: ack-ecs-controller.v1.0.10
        version: 1.0.10
      - name: ack-ecs-controller.v1.0.9
        version: 1.0.9
      - name: ack-ecs-controller.v1.0.8
        version: 1.0.8
      - name: ack-ecs-controller.v1.0.7
        version: 1.0.7
      - name: ack-ecs-controller.v1.0.6
        version: 1.0.6
      - name: ack-ecs-controller.v1.0.2
        version: 1.0.2
      - name: ack-ecs-controller.v1.0.1
        version: 1.0.1
      - name: ack-ecs-controller.v1.0.0
        version: 1.0.0
      - name: ack-ecs-controller.v0.0.8
        version: 0.0.8
      - name: ack-ecs-controller.v0.0.7
        version: 0.0.7
      - name: ack-ecs-controller.v0.0.6
        version: 0.0.6
      - name: ack-ecs-controller.v0.0.5
        version: 0.0.5
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseWeb
      provider-url: ""
    name: cnfv-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cnfv-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eweb.enterpriseweb.com/v1",
                "kind": "CloudNFV",
                "metadata": {
                  "name": "cloudnfv-sample"
                },
                "spec": {
                  "size": 1
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CloudNFV is the Schema for the cloudnfvs API
            displayName: Cloud NFV
            kind: CloudNFV
            name: cloudnfvs.eweb.enterpriseweb.com
            version: v1
        description: Manages a cluster of CloudNFV Runtime Modules
        displayName: CloudNFV
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - automation
        - NFV
        - controller
        links:
        - name: Cnfv Operator
          url: https://cnfv-operator.domain
        maintainers:
        - email: bill@enterpriseweb.com
          name: Bill Malyk
        maturity: alpha
        provider:
          name: EnterpriseWeb
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61a164ad711478ac603242338f7b3c55986319479878b092be128581a823598e
        - quay.io/wmalyk28/cnfv-op@sha256:e5913ad5dd20cb1b2299d2387237ef15e645c9a8c9d4f5fe0044b86a6b3dc1df
        - quay.io/wmalyk28/cloudnfv-runtime@sha256:fcde8375576c9bff602d3a16ef1ba8ae88b780166649ba6ae355737b1d35a03f
        version: 0.0.1
      entries:
      - name: cnfv-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cnfv-operator
    provider:
      name: EnterpriseWeb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzounstructured-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzounstructured-operator.v3.1.200
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzounstructured.clusters.cambridgesemantics.com/v1",
                "kind": "AnzoUnstructured",
                "metadata": {
                  "name": "au01"
                },
                "spec": {
                  "auWorker": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-w",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-worker@sha256:964633d6b16ddc517a09ac86f3e3698f5a161a72cfa50ea9ba4e301fa631e347",
                                "name": "w",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  },
                  "msLeader": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-ms",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzo-microservices-leader@sha256:0637603c00a99533d57e16009516ec73d40703ae151e631fc281f51f1f0ca93b",
                                "name": "ms",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/unstructured-operator@sha256:633a8992e35601d99c0d690dd71af314605c223b57c2038fa7090be06977b2f9
          createdAt: "2025-07-08T09:47:55Z"
          description: Kubernetes Operator for Altair Graph Studio Unstructured
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoUnstructured is the Schema for the anzounstructureds
              API
            displayName: Altair Graph Studio Unstructured Operator
            kind: AnzoUnstructured
            name: anzounstructureds.anzounstructured.clusters.cambridgesemantics.com
            version: v1
        description: |-
          The Altair Graph Studio Unstructured Operator provides the way to install and configure an anzo unstructured setup on Red Hat K8S environment.
          Currently, this is possible only through existing Anzo installation.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v3.1.2/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Studio Unstructured Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Database
        - Big Data
        links:
        - name: Altair Graph Studio Unstructured Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v3.1.2/README.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/unstructured-operator@sha256:633a8992e35601d99c0d690dd71af314605c223b57c2038fa7090be06977b2f9
        - registry.connect.redhat.com/cambridgesemantics/anzo-microservices-leader@sha256:0637603c00a99533d57e16009516ec73d40703ae151e631fc281f51f1f0ca93b
        - registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-worker@sha256:964633d6b16ddc517a09ac86f3e3698f5a161a72cfa50ea9ba4e301fa631e347
        version: 3.1.200
      entries:
      - name: anzounstructured-operator.v3.1.200
        version: 3.1.200
      - name: anzounstructured-operator.v3.1.104
        version: 3.1.104
      - name: anzounstructured-operator.v3.1.103
        version: 3.1.103
      name: stable
    defaultChannel: stable
    packageName: anzounstructured-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: klusterlet
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: klusterlet.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "Klusterlet",
                "metadata": {
                  "name": "klusterlet"
                },
                "spec": {
                  "clusterName": "cluster1",
                  "deployOption": {
                    "mode": "Singleton"
                  },
                  "externalServerURLs": [
                    {
                      "url": "https://localhost"
                    }
                  ],
                  "imagePullSpec": "quay.io/open-cluster-management/registration-operator:v0.13.1",
                  "namespace": "open-cluster-management-agent",
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/registration-operator:v0.13.1
          createdAt: "2024-04-10T15:27:49Z"
          description: Manages the installation and upgrade of the Klusterlet.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management-io/ocm
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Klusterlet defines the configurations of agents running on
              the managed cluster
            displayName: Klusterlet
            kind: Klusterlet
            name: klusterlets.operator.open-cluster-management.io
            version: v1
        description: "Note: This operator will stop updating, please use the latest
          operator in OperatorHub.io https://operatorhub.io/operator/klusterlet. \n\nThe
          Klusterlet proivdes the registation to the Hub clusters as a managed cluster.
          This operator supports the installation and upgrade of Klusterlet.\n\n##
          Prerequisites\nYou need a Hub cluster which has installed ClusterManager.\n\n##
          How to Install\n* Install Klusterlet operator by following instructions
          in top right button `Install`.\n* Optional: Create an agent namespace used
          to deploy Klusterlet. The default namespace is `open-cluster-management-agent`.\n*
          Create a secret `bootstrap-hub-kubeconfig` with the kubeconfig file of hub
          cluster in the agent namespace.\n\n  ```\n  $ kubectl create secret generic
          bootstrap-hub-kubeconfig --from-file=kubeconfig=<kubeconfig file of hub
          clsuter> -n <agent namespace>\n  ```\n\n* Create an instance of the Klusterlet
          resource to install the Klusterlet.\n\nThe pods of Klusterlet are created
          in agent namespace.\n\n```\n$ kubectl get pods -n open-cluster-management-agent\n
          \ NAME                                READY   STATUS    RESTARTS   AGE\n
          \ klusterlet-agent-55c6574876-nfxm9   1/1     Running   0          8m38s\n```\n\nThere
          is a ManagedCluster created in the Hub cluster.\n\n```\n$ kubectl get managedclusters.cluster.open-cluster-management.io\nNAME
          \                  HUB ACCEPTED   MANAGED CLUSTER URLS     JOINED   AVAILABLE
          \  AGE\ncluster1               false                                                        27s\n```\n"
        displayName: Klusterlet
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open-cluster-management
        - Klusterlet
        links:
        - name: Registration Operator
          url: https://github.com/open-cluster-management-io/ocm
        maintainers:
        - email: zyin@redhat.com
          name: Zhiwei Yin
        - email: deads@redhat.com
          name: David Eads
        - email: jqiu@redhat.com
          name: Jian Qiu
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/open-cluster-management/registration-operator:v0.13.1
        version: 0.13.1
      entries:
      - name: klusterlet.v0.13.1
        version: 0.13.1
      - name: klusterlet.v0.13.0
        version: 0.13.0
      - name: klusterlet.v0.12.0
        version: 0.12.0
      - name: klusterlet.v0.11.0
        version: 0.11.0
      - name: klusterlet.v0.10.0
        version: 0.10.0
      - name: klusterlet.v0.9.1
        version: 0.9.1
      - name: klusterlet.v0.8.0
        version: 0.8.0
      - name: klusterlet.v0.7.0
        version: 0.7.0
      - name: klusterlet.v0.6.0
        version: 0.6.0
      - name: klusterlet.v0.5.0
        version: 0.5.0
      - name: klusterlet.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: klusterlet
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-api-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-api-controller.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "example-registry-with-studio"
                },
                "spec": {
                  "app": {
                    "host": "example-registry-with-studio-app.apps.yourcluster.example"
                  },
                  "studioUi": {
                    "enabled": true,
                    "host": "example-registry-with-studio-studio-ui.apps.yourcluster.example"
                  },
                  "ui": {
                    "host": "example-registry-with-studio-ui.apps.yourcluster.example"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-operator:3.0.5
          createdAt: "2024-12-03T11:30:52Z"
          description: Deploy and manage Apicurio API Controller on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Apicurio API
              Controller.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Apicurio API Controller

          Apicurio API Controller consists of 2 components:
           - Apicurio Registry 3
           - Apicurio Studio

          Storage options: PostgreSQL, Kafka, MySQL, SQL Server, in-memory.

          ## Apicurio API Controller Operator

          Provides a quick and easy way to deploy and manage Apicurio API Controller on Kubernetes.

          **Features**
          - Supports installation and configuration of Apicurio API Controller
          - Perform a rolling upgrade of Apicurio API Controller

          ## Prerequisites

          This operator does not deploy storage for Apicurio API Controller. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio API Controller and Apicurio API Controller Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Apicurio API Controller
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - API Management
        - Apicurio Registry
        - Apicurio Studio
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-operator:3.0.5
        version: 0.0.1
      entries:
      - name: apicurio-api-controller.v0.0.1
        version: 0.0.1
      name: 0.x
    defaultChannel: 0.x
    packageName: apicurio-api-controller
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: unsupported
      operatorframework.io/arch.s390x: unsupported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: https://www.konveyor.io
    name: konveyor-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: konveyor-operator.v0.8.0-alpha.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:c1d1942b3bb3c4e99d9b5484493e27a432455608fee4548f1d352be34a8d6ded
          createdAt: "2025-08-13T23:43:34Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0-alpha.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/kantra@sha256:06ae932d9c2f51035bc5a488dd61e8a31c344e72615b013419199168701ea3fb
        - quay.io/konveyor/generic-external-provider@sha256:d58e867b51d5e76cd3597cfb7f2478d221b966f84509dbdd44c609ff5d9169be
        - quay.io/konveyor/kai-solution-server@sha256:83987d12d8290e085b6b66de56d347a8bf262815a5f99796177cb9120513ee3a
        - quay.io/konveyor/tackle-keycloak-init@sha256:07c2ca07922d273bd2a3d9156ad8af0e5bf973aeb316496ee65d7201c381c1b1
        - quay.io/konveyor/tackle2-ui@sha256:867bf24dd1687a79f2a911c22c1059ca7bd025a307d290290265653a696b8e98
        - quay.io/konveyor/tackle2-addon-discovery@sha256:ba865a74287ff94375ed316f348ee2c1d962d8d350c8afae5dbb0aee9b64fdea
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/origin-oauth-proxy@sha256:2c811da843057119b246784f21dc7eec490669a8569e8295d8e914cdce555b24
        - quay.io/konveyor/postgresql-15-c9s@sha256:d75c5a1a0fa686ccf701e5e55633aa73d0cbb2cd9b5d8d4d4e80d7c46d78c121
        - quay.io/konveyor/tackle2-operator@sha256:c1d1942b3bb3c4e99d9b5484493e27a432455608fee4548f1d352be34a8d6ded
        - quay.io/konveyor/tackle2-hub@sha256:6be8fe8307e6686b178ace42f6db5cab8f04c5e2e29e02d2edcb5dc28ee337e9
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:7a5750fb76935e55b7b59e8363f21f5f9b48b39492d7f3e58e2e6a84d07ab027
        - quay.io/konveyor/java-external-provider@sha256:42a156f1d81352f5e09fa0ffe2b51a4d93a1f476755eb0bac9a1d24bd3a89631
        - quay.io/konveyor/tackle2-addon-platform@sha256:12efa5b7fbf6d829879b19e2820dcf782e648059f51544ebd25251dfc67fc7ca
        version: 0.8.0-alpha.4
      entries:
      - name: konveyor-operator.v0.8.0-alpha.4
        version: 0.8.0-alpha.4
      - name: konveyor-operator.v0.8.0-alpha.2
        version: 0.8.0-alpha.2
      - name: konveyor-operator.v0.8.0-alpha.1
        version: 0.8.0-alpha.1
      - name: konveyor-operator.v0.7.0-alpha.2
        version: 0.7.0-alpha.2
      - name: konveyor-operator.v0.7.0-alpha.1
        version: 0.7.0-alpha.1
      - name: konveyor-operator.v0.6.0-alpha.2
        version: 0.6.0-alpha.2
      - name: konveyor-operator.v0.6.0-alpha.1
        version: 0.6.0-alpha.1
      - name: konveyor-operator.v0.5.0-alpha.1
        version: 0.5.0-alpha.1
      - name: konveyor-operator.v0.4.0-alpha.1
        version: 0.4.0-alpha.1
      - name: konveyor-operator.v0.3.0-alpha.6
        version: 0.3.0-alpha.6
      - name: konveyor-operator.v0.3.0-alpha.5
        version: 0.3.0-alpha.5
      - name: konveyor-operator.v0.3.0-alpha.4
        version: 0.3.0-alpha.4
      - name: konveyor-operator.v0.3.0-alpha.3
        version: 0.3.0-alpha.3
      - name: konveyor-operator.v0.3.0-alpha.2
        version: 0.3.0-alpha.2
      - name: konveyor-operator.v0.3.0-alpha.1
        version: 0.3.0-alpha.1
      - name: konveyor-operator.v0.2.0-alpha.3
        version: 0.2.0-alpha.3
      - name: konveyor-operator.v0.2.0-alpha.2
        version: 0.2.0-alpha.2
      - name: konveyor-operator.v0.2.0-alpha.1
        version: 0.2.0-alpha.1
      name: alpha
    - currentCSV: konveyor-operator.v0.8.0-beta.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:019a79dec40445b6289948e173e5bf2ebfa9b495a7bb077abd634c0667953a54
          createdAt: "2025-09-17T19:50:24Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0-beta.5'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/java-external-provider@sha256:0cbdd91cc5057889b8daa2f7611fa0c944fab951c73bdf88a3e28795e9a5fc33
        - quay.io/konveyor/origin-oauth-proxy@sha256:8988f14db7dde3edcaed0541cecb0b4fd58fab9dbaffb79a6c1c4aee13b9bd73
        - quay.io/konveyor/tackle-keycloak-init@sha256:0b4bfc1fea756f10a0657861c4c422303a2349d8d9b04a23238a32ceaf106ac5
        - quay.io/konveyor/tackle2-addon-platform@sha256:d0280ade44ba5e15210c07342608576b38e6993af2e479740675a4706cd2c875
        - quay.io/konveyor/kantra@sha256:ad6f5c05e5e950e33a745accd03ad933fb7bb4907f44455f88a2b5c27e6dd15e
        - quay.io/konveyor/tackle2-ui@sha256:78398832e092557639531230cf6d023bb07b589f3587afeaacf8c68640abc454
        - quay.io/konveyor/generic-external-provider@sha256:a6c550bad7a09a1c51e226247f11c00d8453987698832cb566582842219f1f3c
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:fed49c5ff14bb232e63d2dc180d3b530da63f11bf97055fc9d86cbb9e47104a4
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/postgresql-15-c9s@sha256:157a2941ae971d8f72b2f8af1d4a29deeb5548012a978fa986c1c6e52395da4d
        - quay.io/konveyor/tackle2-addon-discovery@sha256:a29a0248e222d9ac3225e9d3df603f1383fa8844856bf68ea618d303f0b79883
        - quay.io/konveyor/tackle2-operator@sha256:019a79dec40445b6289948e173e5bf2ebfa9b495a7bb077abd634c0667953a54
        - quay.io/konveyor/kai-solution-server@sha256:1e69cf8a342090a03a9dddb1b28ae3627052b6d636ffa290a11761b94fd10648
        - quay.io/konveyor/tackle2-hub@sha256:1611913558c9f1203a844b43969d7b1539dab3578b83460e8383ee6757c3c34c
        version: 0.8.0-beta.5
      entries:
      - name: konveyor-operator.v0.8.0-beta.5
        version: 0.8.0-beta.5
      - name: konveyor-operator.v0.8.0-beta.4
        version: 0.8.0-beta.4
      - name: konveyor-operator.v0.8.0-beta.2
        version: 0.8.0-beta.2
      - name: konveyor-operator.v0.8.0-beta.1
        version: 0.8.0-beta.1
      - name: konveyor-operator.v0.6.0-beta.2
        version: 0.6.0-beta.2
      - name: konveyor-operator.v0.6.0-beta.1
        version: 0.6.0-beta.1
      - name: konveyor-operator.v0.5.0-beta.3
        version: 0.5.0-beta.3
      - name: konveyor-operator.v0.5.0-beta.2
        version: 0.5.0-beta.2
      - name: konveyor-operator.v0.5.0-beta.1
        version: 0.5.0-beta.1
      - name: konveyor-operator.v0.3.0-beta.3
        version: 0.3.0-beta.3
      - name: konveyor-operator.v0.3.0-beta.2
        version: 0.3.0-beta.2
      - name: konveyor-operator.v0.3.0-beta.1.2
        version: 0.3.0-beta.1.2
      - name: konveyor-operator.v0.3.0-beta.1.1
        version: 0.3.0-beta.1.1
      - name: konveyor-operator.v0.3.0-beta.1
        version: 0.3.0-beta.1
      name: beta
    - currentCSV: konveyor-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "move2kube-plan",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "image": "quay.io/konveyor/tackle-addon-move2kube-plan:latest"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:9688d0763d8c7812beee715a061c3c198006448855d7b891d45833f62cf869fb
          createdAt: "2023-05-25T15:23:45Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.1.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)

        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:9688d0763d8c7812beee715a061c3c198006448855d7b891d45833f62cf869fb
        - quay.io/konveyor/tackle2-addon-windup@sha256:81f0625cb82eb24ded51c7258f229f57778820c62e3be6ec05fb2b12d2970825
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle2-hub@sha256:3d15d158ffb6f1fc4ca5aaf29a12173fe344c289773902c9161056bdf69acc92
        - quay.io/konveyor/tackle2-addon@sha256:35be344bbd3393f5634f52aafd21e9c3964d1e0ba5011ecad1561600d476bbe9
        - quay.io/konveyor/tackle-pathfinder@sha256:f9c5aa49f1d27f477cdd82c21441ae0287acdcdc9b2b3fcf197884007dc5f7dd
        - quay.io/konveyor/tackle-keycloak-init@sha256:b0347e5a608352c4b7bb796b65a67d417e98ba4896db71a487fa26243214cfcf
        - quay.io/konveyor/postgresql-12-centos7@sha256:1287d947464fbf5275feb31d106d2feb856c6117f0eaec8d947f2e03bd8618c5
        - quay.io/konveyor/tackle2-ui@sha256:99650804e65e2b6cb15a37c00189d7f463e4d19a0f866a79e9c09c8420a4d028
        version: 0.1.2
      entries:
      - name: konveyor-operator.v0.1.2
        version: 0.1.2
      - name: konveyor-operator.v0.1.1
        version: 0.1.1
      - name: konveyor-operator.v0.1.0
        version: 0.1.0
      name: konveyor-0.1
    - currentCSV: konveyor-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "move2kube-plan",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "image": "quay.io/konveyor/tackle-addon-move2kube-plan:latest"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:1860bb099ac64d9a2af27692f0671b7240e4d19746fffa9f9ae2356ff2818e10
          createdAt: "2023-07-06T06:17:40Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.2.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)

        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:1860bb099ac64d9a2af27692f0671b7240e4d19746fffa9f9ae2356ff2818e10
        - quay.io/konveyor/postgresql-12-centos7@sha256:6016530ab1fce71d29421ce9a97b3a89fbbb126b87b1b079c6aadebb7ff5664c
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle-keycloak-init@sha256:ca6c85bdf8417d4be116c2098b8d8ad5ae8b7ab0b636c4cbe4a392aa424e2d47
        - quay.io/konveyor/tackle2-hub@sha256:d9c4f21f44098a2526474968b6d88094dca8b5eaeb9042716332c9cb67df47cd
        - quay.io/konveyor/tackle2-ui@sha256:d10c1e717d160c5eb2581649a8dd7cdf4922a74d92db014f72f42a48a3134341
        - quay.io/konveyor/tackle2-addon@sha256:11693e16363bcd98249209d63a29f0c6cfffab3a4071649cbeea2fd276bc2440
        - quay.io/konveyor/tackle2-addon-windup@sha256:538dfdb134d53a876d126dbd16679638584a606ef2b9e8fbe69ed1f29733c03a
        - quay.io/konveyor/tackle-pathfinder@sha256:190269b061b2065ca8ba9357bf9b959a2f587a547a44769ae0afd1ac86b4b197
        version: 0.2.1
      entries:
      - name: konveyor-operator.v0.2.1
        version: 0.2.1
      - name: konveyor-operator.v0.2.0
        version: 0.2.0
      name: konveyor-0.2
    - currentCSV: konveyor-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:cd2305373f50ce927393e43f5735915e46c32d6f8a76091ee2808218885bba72
          createdAt: "2024-05-02T22:12:42Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.3.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:51bda0c6577c15684583d607a4f0cebef760814102423f4b2025016093a19fc1
        - quay.io/konveyor/origin-oauth-proxy@sha256:05f6bd1216b921b1f47a3821f0e22b84ed6b141bb2ada474c12bd9fde613108c
        - quay.io/konveyor/tackle2-hub@sha256:bf8803721d06d1a39251e22892343a013d5ef0fb4c1883c95e1eb55aa683f486
        - quay.io/konveyor/tackle2-operator@sha256:cd2305373f50ce927393e43f5735915e46c32d6f8a76091ee2808218885bba72
        - quay.io/konveyor/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle-keycloak-init@sha256:a63702cc0cb36830a7830ed70ca344a9aacc201ba6f3c94806ca2e93160df4dd
        - quay.io/konveyor/tackle2-ui@sha256:bbf58d2ea613489f91efae450421afafb9052416b0d7971606e2729e89fd7de6
        version: 0.3.2
      entries:
      - name: konveyor-operator.v0.3.2
        version: 0.3.2
      - name: konveyor-operator.v0.3.1
        version: 0.3.1
      - name: konveyor-operator.v0.3.0
        version: 0.3.0
      name: konveyor-0.3
    - currentCSV: konveyor-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:496db8ca859608d04008ac36cf7bdc9cafd7214294d33a9c5646083c361ffcb6
          createdAt: "2024-06-05T03:08:51Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.4.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle-keycloak-init@sha256:4c7e07b1fce05f7dca47edecc0ba0ce373cc38e48879cc7eed72510d2865a18e
        - quay.io/konveyor/tackle2-ui@sha256:cb1255fc5fa4608fc659c3455f3164664b960cc2d9a3b2208967798e3532df78
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:e84ea36d206a8c143794e3986940377e77c7618225cad594bdf9a4047d7823dc
        - quay.io/konveyor/origin-oauth-proxy@sha256:05f6bd1216b921b1f47a3821f0e22b84ed6b141bb2ada474c12bd9fde613108c
        - quay.io/konveyor/tackle2-operator@sha256:496db8ca859608d04008ac36cf7bdc9cafd7214294d33a9c5646083c361ffcb6
        - quay.io/konveyor/tackle2-hub@sha256:b87acd8e2ee4a7edc4369bc375f64e2b98bc479d09556439e2ee4d74b74970a9
        - quay.io/konveyor/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        version: 0.4.0
      entries:
      - name: konveyor-operator.v0.4.0
        version: 0.4.0
      name: konveyor-0.4
    - currentCSV: konveyor-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:474d739ed470701b628f1f9e971925ac7fbe0b543f3e36305d932b963b1225fd
          createdAt: "2024-08-27T15:48:02Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.5.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:474d739ed470701b628f1f9e971925ac7fbe0b543f3e36305d932b963b1225fd
        - quay.io/konveyor/java-external-provider@sha256:33fb5d24b998f01c481a0a03e3f1bf3f7a6c93f5b662885102a0c966026e3375
        - quay.io/konveyor/generic-external-provider@sha256:334a631538d31efc8d82adae4f52303336800fa3037b5055c72400918d81955c
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/tackle2-ui@sha256:36f29fe01d3078f0e100b24d1ed126850c5a773f4b0ae922fd4c4487de1ac29d
        - quay.io/konveyor/tackle2-hub@sha256:faf91ca68e24979642a44fa013b1fb9edfee415eee835ccef9dce2eb418e2909
        - quay.io/konveyor/postgresql-15-c9s@sha256:039ce764cadd03b4cfea954700139b8ceff415a9a0970baa7f2cf60f7f929174
        - quay.io/konveyor/tackle-keycloak-init@sha256:5f652be5914ae6e23537b9fe02ccc9cd638c08de5462d9f58135a405c81f5287
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:2afbe8bea09904e084e026cd0a294fbe1c7b05d0e387f6633b09c5eaf4665e02
        - quay.io/konveyor/tackle2-addon-discovery@sha256:461f960bb1d29e719fb472fe9be9cf3deca17d984687957f54546a09047b0d3c
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        version: 0.5.3
      entries:
      - name: konveyor-operator.v0.5.3
        version: 0.5.3
      - name: konveyor-operator.v0.5.1
        version: 0.5.1
      - name: konveyor-operator.v0.5.0
        version: 0.5.0
      name: konveyor-0.5
    - currentCSV: konveyor-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:8ee143c473570efd27b6f93332dc272871e11c66cc20f381c3a1e6e5688abf68
          createdAt: "2025-03-05T21:14:52Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.6.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:8ee143c473570efd27b6f93332dc272871e11c66cc20f381c3a1e6e5688abf68
        - quay.io/konveyor/tackle2-hub@sha256:52a8b9be038fbfbc4e22172a4a8c7191d2b06fd8670e5df2ec442e3ce00d8c8a
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:0b5cbb322fbe96335522410376c2e8505a8302088c0e454f3e5e7e28ba7c96a6
        - quay.io/konveyor/java-external-provider@sha256:15148cde9a6fedef2207fe36f0ac3bd1ecdc6e3ac76f1cbec9eb97ffcf247231
        - quay.io/konveyor/tackle-keycloak-init@sha256:a3741269aa15edd2f9fb7df853a03a942d16a88eec114f3e54f82ef63b82adc4
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/postgresql-15-c9s@sha256:55f8014c1edafab8f9085228a0d207c4e98820c615be34b935ce98e942d388b1
        - quay.io/konveyor/tackle2-ui@sha256:12b0eda13a63508b9c13d4f9309bbfbe19d564627cf0de672307af1a6ec993fb
        - quay.io/konveyor/tackle2-addon-discovery@sha256:3eb6f395c451e2c9a867871b3984a25f9f8e6ddcfe388f37fbd6d46a61a70b1e
        - quay.io/konveyor/generic-external-provider@sha256:27c1731859a4a639d2750cff8d1246281f4b0466bb5b069169c3e170c3c19b30
        version: 0.6.1
      entries:
      - name: konveyor-operator.v0.6.1
        version: 0.6.1
      - name: konveyor-operator.v0.6.0
        version: 0.6.0
      name: konveyor-0.6
    - currentCSV: konveyor-operator.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:f5870bc26b2a07ddacf00ed5b5f7df7d1068489010a80bcc543686c6c9576fbb
          createdAt: "2025-07-10T15:16:17Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-ui@sha256:f87b04762fd163f40c49f3bcffd473f0c3f80a643576c6a3e2c2b4a14395c09f
        - quay.io/konveyor/java-external-provider@sha256:a0ef733e9fce3845d02b5360440add85427acad697d2157ef92242dec06bd73b
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/tackle-keycloak-init@sha256:dab4cb1229c8a729d14b4cecb4aa3b5c4d48e94c36e84d785c922fc08d120617
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:7ed52ea9b95650241b5c2c01914775e29be23f7ed75fabb6e7c71969f3f0728f
        - quay.io/konveyor/tackle2-addon-discovery@sha256:60951801aa539ee59cd43f3c5f42a2ca751ca70e995cbb258b7d55ebb24d8450
        - quay.io/konveyor/origin-oauth-proxy@sha256:2c811da843057119b246784f21dc7eec490669a8569e8295d8e914cdce555b24
        - quay.io/konveyor/generic-external-provider@sha256:facca8861559a658710ad3aef8b325731595bf944cce1907715d6efd33887072
        - quay.io/konveyor/tackle2-operator@sha256:f5870bc26b2a07ddacf00ed5b5f7df7d1068489010a80bcc543686c6c9576fbb
        - quay.io/konveyor/tackle2-hub@sha256:52f66607f3134a4165ce691c6e9ce992d93927e9591055daafdc1bfc1dd09dd9
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/postgresql-15-c9s@sha256:41460103a0bab189cd77acb6e741d9c5a6d3506479f7e56df40eb156d9b2cadd
        version: 0.7.1
      entries:
      - name: konveyor-operator.v0.7.1
        version: 0.7.1
      - name: konveyor-operator.v0.7.0
        version: 0.7.0
      name: konveyor-0.7
    - currentCSV: konveyor-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:68b6672fccdf54c65cc83f4c82db023053d6b61409bea880f58b53ccfea7b1ed
          createdAt: "2025-09-29T20:41:09Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-addon-discovery@sha256:96a5b37834fbed9506014bcbd909258081fd833ef9e16daa104effd0e3a3c6ea
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-addon-platform@sha256:d0280ade44ba5e15210c07342608576b38e6993af2e479740675a4706cd2c875
        - quay.io/konveyor/tackle2-ui@sha256:f3f7380be65ff1fb3459c37a3d4865354bab685dc836893f78221f16ba579f70
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:8c82ad880ca2a2d5cd4b08125feca16d13d014d15f4f77677306086d1067f9aa
        - quay.io/konveyor/tackle-keycloak-init@sha256:c0051b8c3a3c891fe12012623aee775080ecb47af8a4e75b8be8b0b4855d9f2c
        - quay.io/konveyor/postgresql-15-c9s@sha256:af275f35a307367dcee9a339a6903cf79e3a8a41cabf79fb7cea7015895f0b58
        - quay.io/konveyor/origin-oauth-proxy@sha256:8988f14db7dde3edcaed0541cecb0b4fd58fab9dbaffb79a6c1c4aee13b9bd73
        - quay.io/konveyor/java-external-provider@sha256:8eab08a04e98051f89352250d63523a5b591f35a6497f5d69a5a9d556a94abf9
        - quay.io/konveyor/tackle2-hub@sha256:b3e6db1ffc195143b8e5d54b60949256e154b0f8b667736512a084e015cdf2a7
        - quay.io/konveyor/kantra@sha256:a5b5591a51143574a6ce6678b141b9cd3e14328c6fad773412dde21b442fe364
        - quay.io/konveyor/generic-external-provider@sha256:c7b6618bfff81d73f1b455963107ffdaf19334e6552ce076c009476da3f37f21
        - quay.io/konveyor/kai-solution-server@sha256:bab6e224c50de5b44cc37ce0297a3b91255efc08c1ad0fccf14fef9e5182775f
        - quay.io/konveyor/tackle2-operator@sha256:68b6672fccdf54c65cc83f4c82db023053d6b61409bea880f58b53ccfea7b1ed
        version: 0.8.0
      entries:
      - name: konveyor-operator.v0.8.0
        version: 0.8.0
      name: konveyor-0.8
    - currentCSV: konveyor-operator.v0.7.0-rc.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
          createdAt: "2025-05-21T17:12:48Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.0-rc.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
        - quay.io/konveyor/tackle2-addon-discovery@sha256:89d2b0f8636897eb4dc1957a74a5b5041c0f3267a747adcf2bc6ca56b17b9937
        - quay.io/konveyor/tackle2-hub@sha256:bed834eb5488eef23b0d1e8af194b0646cf74c39c197e0d57acc5aed2299d9f5
        - quay.io/konveyor/tackle-keycloak-init@sha256:c3124971dc31be9201de2a6a1f6b2f886adf4c96731188dd5e750798cde3cfe7
        - quay.io/konveyor/tackle2-ui@sha256:afe9141366a23c072286517e9af5436d14fd1df775aaa2fa13c3911764515777
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/java-external-provider@sha256:61ab20e37916abfeeadde685336660f0e8bbf519403d1f7100d43c1e42a72fbd
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:46150725a2adbde8dfdc2250614d437a50e64bbf0b4562dabc73715a287de67b
        - quay.io/konveyor/generic-external-provider@sha256:de55f78939576c110eb6f084a7ca51338fedba699fbb5135f23392783a5d7b95
        - quay.io/konveyor/origin-oauth-proxy@sha256:6c2847fd9667938b1d1294d988038e885e43cd9fb378862dae3419ce4f21631f
        - quay.io/konveyor/postgresql-15-c9s@sha256:012cd6611aa9d3f4e0a82636d6817f381c19e3b0809199216fbe7fa3396a2ef5
        version: 0.7.0-rc.1
      entries:
      - name: konveyor-operator.v0.7.0-rc.1
        version: 0.7.0-rc.1
      - name: konveyor-operator.v0.6.0-rc.2
        version: 0.6.0-rc.2
      - name: konveyor-operator.v0.6.0-rc.1
        version: 0.6.0-rc.1
      - name: konveyor-operator.v0.5.0-rc.1
        version: 0.5.0-rc.1
      - name: konveyor-operator.v0.3.0-rc.4
        version: 0.3.0-rc.4
      - name: konveyor-operator.v0.3.0-rc.3
        version: 0.3.0-rc.3
      - name: konveyor-operator.v0.3.0-rc.2
        version: 0.3.0-rc.2
      - name: konveyor-operator.v0.3.0-rc.1
        version: 0.3.0-rc.1
      name: rc
    - currentCSV: konveyor-operator.v0.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:33cf01f19a7043bcc5bdb12a3ed6bde7a9094981aa2383554111ebec9b4ba498
          createdAt: "2025-03-26T20:25:41Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.6.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-ui@sha256:3c568c6047a044095080b6013b643a4041deae7b45936a443d4c4d708a81b945
        - quay.io/konveyor/tackle2-hub@sha256:0df49087d0c9341c2120a7c3aa2ba7907c056f94e1602ffd0496ca7113c75382
        - quay.io/konveyor/tackle-keycloak-init@sha256:e32fa48351f410361e8bc100303deeb40d8a66d7eefb0cf7b1bc16e067a9a795
        - quay.io/konveyor/tackle2-operator@sha256:33cf01f19a7043bcc5bdb12a3ed6bde7a9094981aa2383554111ebec9b4ba498
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/tackle2-addon-discovery@sha256:1b89266d6bdfa497ba8137e0d33ff53785ac476dd96bf4fd9a185ef0ddf513ce
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:734752be111a97e4b7e0030d563c32883f419ad031b2f350cb8e7333f2f37de3
        - quay.io/konveyor/generic-external-provider@sha256:bc1112352f1a2c3172c5e7d5e637b5c5a69bf7d98b78f494fa593e8872e5cfd1
        - quay.io/konveyor/java-external-provider@sha256:fe85689b57125d678df685012b93c8a567fcb9251777cd711e9b82bdb22a734c
        - quay.io/konveyor/postgresql-15-c9s@sha256:7957d3b272b9f4e469bdc0c514ec9e70e8fbdc34a7250b4f4c62f9bb73cbf08d
        version: 0.6.2
      entries:
      - name: konveyor-operator.v0.6.2
        version: 0.6.2
      name: release-0.6
    - currentCSV: konveyor-operator.v0.7.0-rc.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
          createdAt: "2025-05-21T17:12:48Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.0-rc.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:46150725a2adbde8dfdc2250614d437a50e64bbf0b4562dabc73715a287de67b
        - quay.io/konveyor/generic-external-provider@sha256:de55f78939576c110eb6f084a7ca51338fedba699fbb5135f23392783a5d7b95
        - quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
        - quay.io/konveyor/postgresql-15-c9s@sha256:012cd6611aa9d3f4e0a82636d6817f381c19e3b0809199216fbe7fa3396a2ef5
        - quay.io/konveyor/tackle-keycloak-init@sha256:c3124971dc31be9201de2a6a1f6b2f886adf4c96731188dd5e750798cde3cfe7
        - quay.io/konveyor/java-external-provider@sha256:61ab20e37916abfeeadde685336660f0e8bbf519403d1f7100d43c1e42a72fbd
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/tackle2-ui@sha256:afe9141366a23c072286517e9af5436d14fd1df775aaa2fa13c3911764515777
        - quay.io/konveyor/tackle2-hub@sha256:bed834eb5488eef23b0d1e8af194b0646cf74c39c197e0d57acc5aed2299d9f5
        - quay.io/konveyor/origin-oauth-proxy@sha256:6c2847fd9667938b1d1294d988038e885e43cd9fb378862dae3419ce4f21631f
        - quay.io/konveyor/tackle2-addon-discovery@sha256:89d2b0f8636897eb4dc1957a74a5b5041c0f3267a747adcf2bc6ca56b17b9937
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        version: 0.7.0-rc.1
      entries:
      - name: konveyor-operator.v0.7.0-rc.1
        version: 0.7.0-rc.1
      name: release-0.7
    defaultChannel: konveyor-0.8
    packageName: konveyor-operator
    provider:
      name: Konveyor
      url: https://www.konveyor.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: sriov-network-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sriov-network-operator.v4.13.0-202309190544
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovIBNetwork",
                "metadata": {
                  "name": "example-sriovibnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "linkState": "auto",
                  "networkNamespace": "default",
                  "resourceName": "mellanoxnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetwork",
                "metadata": {
                  "name": "example-sriovnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "intelnics",
                  "spoofChk": "on",
                  "trust": "off",
                  "vlan": 0
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodePolicy",
                "metadata": {
                  "name": "policy-1",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "deviceType": "vfio-pci",
                  "mtu": 9000,
                  "nicSelector": {
                    "pfNames": [
                      "eth1"
                    ],
                    "rootDevices": [
                      "0000:01:00.1"
                    ],
                    "vendor": "8086"
                  },
                  "nodeSelector": {
                    "feature.node.kubernetes.io/network-sriov.capable": "true"
                  },
                  "numVfs": 6,
                  "priority": 99,
                  "resourceName": "intelnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodeState",
                "metadata": {
                  "name": "minikube",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "interfaces": []
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkPoolConfig",
                "metadata": {
                  "name": "sriovnetworkpoolconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovOperatorConfig",
                "metadata": {
                  "name": "default",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "enableInjector": true,
                  "enableOperatorWebhook": true,
                  "logLevel": 2
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: quay.io/pliurh/sriov-network-operator-manager:bd1
          createdAt: "2023-08-02T09:22:34Z"
          description: An operator for configuring SR-IOV components and initializing
            SRIOV network devices in Openshift cluster.
          olm.skipRange: '>=4.3.0-0 <4.13.0-202309190544'
          operatorframework.io/suggested-namespace: openshift-sriov-network-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "cni"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/k8snetworkplumbingwg/sriov-network-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovIBNetwork is the Schema for the sriovibnetworks API
            displayName: Sriov IBNetwork
            kind: SriovIBNetwork
            name: sriovibnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodePolicy is the Schema for the sriovnetworknodepolicies
              API
            displayName: Sriov Network Node Policy
            kind: SriovNetworkNodePolicy
            name: sriovnetworknodepolicies.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodeState is the Schema for the sriovnetworknodestates
              API
            displayName: Sriov Network Node State
            kind: SriovNetworkNodeState
            name: sriovnetworknodestates.sriovnetwork.openshift.io
            version: v1
          - kind: SriovNetworkPoolConfig
            name: sriovnetworkpoolconfigs.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetwork is the Schema for the sriovnetworks API
            displayName: Sriov Network
            kind: SriovNetwork
            name: sriovnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovOperatorConfig is the Schema for the sriovoperatorconfigs
              API
            displayName: Sriov Operator Config
            kind: SriovOperatorConfig
            name: sriovoperatorconfigs.sriovnetwork.openshift.io
            version: v1
        description: |
          # SR-IOV Network Operator for Openshift

          ## Introduction
          The sriov-network-operator is generally responsible for configuring the sriov components in a openshift cluster.

          ### Supported Features
          * Initialize the SR-IOV NICs on nodes.
          * provision SR-IOV device plugin on selected node.
          * provision SR-IOV CNI plugin on selected nodes.
          * manage configuration of SR-IOV device plugin.
          * generate net-att-def CRs for SR-IOV CNI plugin.

          This operator has to run in namespace 'openshift-sriov-network-operator'. An Operator Group is also required to install this operator:

          ```
          $ oc create namespace openshift-sriov-network-operator

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: sriov-network-operators
            namespace: openshift-sriov-network-operator
          spec:
            targetNamespaces:
            - openshift-sriov-network-operator
          EOF
          ```
        displayName: SR-IOV Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - network
        - sr-iov
        - sriov
        links:
        - name: Source Code
          url: https://github.com/k8snetworkplumbingwg/sriov-network-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-sriov-network-config-daemon@sha256:a2053c5366eb89112dfbc7a9994685fc198cce6fa87fdd6d1c0a07727e6eee03
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin@sha256:82d4846f8dfb9e0d02e7ff3f532f8f2bf44c9aba63c37d5ce2efb36969e5c453
        - registry.redhat.io/openshift4/ose-sriov-network-operator@sha256:3314ee580113b985e275583585b1f2adf381c5a0885b51100b3a01872a77a555
        - registry.redhat.io/openshift4/ose-sriov-network-webhook@sha256:bfb84d00dbb6b18d596ca2720b52d922c589249d2f988679189ea0eea985e58a
        - registry.redhat.io/openshift4/sriov-cni-rhel9@sha256:a4f0ec210cd065cb7f77dd25257aadf8345f47819cc728c5b6229acb79a09bb3
        - registry.redhat.io/openshift4/ose-sriov-dp-admission-controller@sha256:8dfaa07a5775a7156d303019879e41b17833e5643794f6e1087881d2a4dd1781
        - registry.redhat.io/openshift4/ose-sriov-infiniband-cni@sha256:716aa0b9ecfddfea0a62615f424f4abeab5362899c695e9ecc9df4df05f1add5
        version: 4.13.0-202309190544
      entries:
      - name: sriov-network-operator.v4.13.0-202309190544
        version: 4.13.0-202309190544
      name: "4.13"
    - currentCSV: sriov-network-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovIBNetwork",
                "metadata": {
                  "name": "example-sriovibnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "linkState": "auto",
                  "networkNamespace": "default",
                  "resourceName": "mellanoxnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetwork",
                "metadata": {
                  "name": "example-sriovnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "intelnics",
                  "spoofChk": "on",
                  "trust": "off",
                  "vlan": 0
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodePolicy",
                "metadata": {
                  "name": "policy-1",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "deviceType": "vfio-pci",
                  "mtu": 9000,
                  "nicSelector": {
                    "pfNames": [
                      "eth1"
                    ],
                    "rootDevices": [
                      "0000:01:00.1"
                    ],
                    "vendor": "8086"
                  },
                  "nodeSelector": {
                    "feature.node.kubernetes.io/network-sriov.capable": "true"
                  },
                  "numVfs": 6,
                  "priority": 99,
                  "resourceName": "intelnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodeState",
                "metadata": {
                  "name": "minikube",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "interfaces": []
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkPoolConfig",
                "metadata": {
                  "name": "sriovnetworkpoolconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovOperatorConfig",
                "metadata": {
                  "name": "default",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "enableInjector": true,
                  "enableOperatorWebhook": true,
                  "logLevel": 2
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-sriov-network-rhel9-operator@sha256:8dcae2a77e904713fba9c5907789f9b22ffb1f941f793adc4b604798d29a44f2
          createdAt: "2024-05-07T23:50:52Z"
          description: An operator for configuring SR-IOV components and initializing
            SRIOV network devices in Openshift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.16.0-202509240123'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "sriovnetwork.openshift.io/v1",
              "kind": "SriovOperatorConfig",
              "metadata": {
                "name": "default"
              },
              "spec": {
                "configurationMode": "daemon",
                "disableDrain": false,
                "enableInjector": true,
                "enableOperatorWebhook": true,
                "logLevel": 2
              }
            }
          operatorframework.io/suggested-namespace: openshift-sriov-network-operator
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-sriov-network-operator",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["sriovnetworknodestates.sriovnetwork.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/k8snetworkplumbingwg/sriov-network-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovIBNetwork is the Schema for the sriovibnetworks API
            displayName: Sriov IBNetwork
            kind: SriovIBNetwork
            name: sriovibnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodePolicy is the Schema for the sriovnetworknodepolicies
              API
            displayName: Sriov Network Node Policy
            kind: SriovNetworkNodePolicy
            name: sriovnetworknodepolicies.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodeState is the Schema for the sriovnetworknodestates
              API
            displayName: Sriov Network Node State
            kind: SriovNetworkNodeState
            name: sriovnetworknodestates.sriovnetwork.openshift.io
            version: v1
          - kind: SriovNetworkPoolConfig
            name: sriovnetworkpoolconfigs.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetwork is the Schema for the sriovnetworks API
            displayName: Sriov Network
            kind: SriovNetwork
            name: sriovnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovOperatorConfig is the Schema for the sriovoperatorconfigs
              API
            displayName: Sriov Operator Config
            kind: SriovOperatorConfig
            name: sriovoperatorconfigs.sriovnetwork.openshift.io
            version: v1
        description: |
          # SR-IOV Network Operator for Openshift

          ## Introduction
          The sriov-network-operator is generally responsible for configuring the sriov components in a openshift cluster.

          ### Supported Features
          * Initialize the SR-IOV NICs on nodes.
          * provision SR-IOV device plugin on selected node.
          * provision SR-IOV CNI plugin on selected nodes.
          * manage configuration of SR-IOV device plugin.
          * generate net-att-def CRs for SR-IOV CNI plugin.

          This operator has to run in namespace 'openshift-sriov-network-operator'. An Operator Group is also required to install this operator:

          ```
          $ oc create namespace openshift-sriov-network-operator

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: sriov-network-operators
            namespace: openshift-sriov-network-operator
          spec:
            targetNamespaces:
            - openshift-sriov-network-operator
          EOF
          ```
        displayName: SR-IOV Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - network
        - sr-iov
        - sriov
        links:
        - name: Source Code
          url: https://github.com/k8snetworkplumbingwg/sriov-network-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-sriov-network-config-daemon-rhel9@sha256:61afbb271a08dc3ab3b5509df535d9456f9b2c1b267554386b4bf859b6bccbe4
        - registry.redhat.io/openshift4/ose-sriov-infiniband-cni-rhel9@sha256:f828e415db73bf2df3fa7b26238a2f5c8a53cda6b396c11a0337d688108e6647
        - registry.redhat.io/openshift4/sriov-cni-rhel9@sha256:cac67ac25a0d319b5ea68cb3f63eac058339f1981811c7a3466b9daa2534a38d
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin-rhel9@sha256:5b3bc02e31eecd0d3468db729b03239d518ef73112b0fdb5ec3f05a788baa0be
        - registry.redhat.io/openshift4/ose-sriov-network-rhel9-operator@sha256:8dcae2a77e904713fba9c5907789f9b22ffb1f941f793adc4b604798d29a44f2
        - registry.redhat.io/openshift4/ose-sriov-network-webhook-rhel9@sha256:a6a5cc76d88b6b7aa86ee24b882ce43be33c8c6d4bcfae5f1b5490ead71ce8b6
        - registry.redhat.io/openshift4/ose-sriov-network-metrics-exporter-rhel9@sha256:a0580d8156de0ca1e5a119441196e9babbeb4d7edc502659c7ac66a49b1928d5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-sriov-dp-admission-controller-rhel9@sha256:20fd8a0081ffdea05796c8f6a0539a0e246da24f739f7bbf0722c90acfdeed15
        version: 4.16.0-202509240123
      entries:
      - name: sriov-network-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: sriov-network-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: sriov-network-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: sriov-network-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: sriov-network-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: sriov-network-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: sriov-network-operator.v4.16.0-202504290837
        version: 4.16.0-202504290837
      - name: sriov-network-operator.v4.16.0-202504151638
        version: 4.16.0-202504151638
      - name: sriov-network-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: sriov-network-operator.v4.16.0-202502262005
        version: 4.16.0-202502262005
      - name: sriov-network-operator.v4.16.0-202502171404
        version: 4.16.0-202502171404
      - name: sriov-network-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: sriov-network-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: sriov-network-operator.v4.16.0-202501081506
        version: 4.16.0-202501081506
      - name: sriov-network-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: sriov-network-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: sriov-network-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: sriov-network-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: sriov-network-operator.v4.16.0-202410300306
        version: 4.16.0-202410300306
      - name: sriov-network-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: sriov-network-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: sriov-network-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: sriov-network-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: sriov-network-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: sriov-network-operator.v4.16.0-202409111535
        version: 4.16.0-202409111535
      - name: sriov-network-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: sriov-network-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: sriov-network-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: sriov-network-operator.v4.16.0-202408120337
        version: 4.16.0-202408120337
      - name: sriov-network-operator.v4.16.0-202407100107
        version: 4.16.0-202407100107
      - name: sriov-network-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: sriov-network-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: sriov-network-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: sriov-network-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: sriov-network-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: The cert-manager maintainers
      provider-url: https://cert-manager.io/
    name: cert-manager
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-manager.v1.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "example-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "example-clusterissuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example-certificate"
                },
                "spec": {
                  "dnsNames": [
                    "example.com"
                  ],
                  "issuerRef": {
                    "name": "example-issuer"
                  },
                  "secretName": "example-certificate-tls"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "name": "example-certificaterequest"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "name": "example-order"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "example-challenge"
                },
                "spec": {}
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: quay.io/jetstack/cert-manager-controller:v1.16.5
          createdAt: 2025-05-02T09:11:31
          olm.skipRange: '>=1.16.0 <1.16.5'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/internal-objects: |-
            [
              "challenges.acme.cert-manager.io",
              "orders.acme.cert-manager.io"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/cert-manager/cert-manager
          support: The cert-manager maintainers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A CertificateRequest is used to request a signed certificate from one of the
              configured issuers.

              All fields within the CertificateRequest's `spec` are immutable after creation.
              A CertificateRequest will either succeed or fail, as denoted by its `Ready` status
              condition and its `status.failureTime` field.

              A CertificateRequest is a one-shot resource, meaning it represents a single
              point in time request for a certificate and cannot be re-used.
            displayName: CertificateRequest
            kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - description: |-
              A Certificate resource should be created to ensure an up to date and signed
              X.509 certificate is stored in the Kubernetes Secret resource named in `spec.secretName`.

              The stored certificate will be renewed before it expires (as configured by `spec.renewBefore`).
            displayName: Certificate
            kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: Challenge is a type to represent a Challenge request with
              an ACME server
            displayName: (Internal) Challenge
            kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - description: |-
              A ClusterIssuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is similar to an Issuer, however it is cluster-scoped and therefore can
              be referenced by resources that exist in *any* namespace, not just the same
              namespace as the referent.
            displayName: ClusterIssuer
            kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - description: |-
              An Issuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is scoped to a single namespace and can therefore only be referenced by
              resources within the same namespace.
            displayName: Issuer
            kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - description: Order is a type to represent an Order with an ACME server
            displayName: (Internal) Order
            kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          cert-manager is a Kubernetes add-on to automate the management and issuance of TLS certificates from various issuing sources.
          It will ensure certificates are valid and up to date periodically, and attempt to renew certificates at an appropriate time before expiry.

          cert-manager has the following [built in issuers][]:

          * ACME: sign certificates using Let's Encrypt (or any other [ACME][] compliant signer).
          * Vault: sign certificates using HashiCorp Vault.
          * Venafi: sign certificates using Venafi Cloud or Venafi Trust Protection Platform.
          * SelfSigned: the private key of the certificate will be used to sign the certificate.
          * CA: a private key stored as a Kubernetes Secret will be used to sign the certificate.

          [built in issuers]: https://cert-manager.io/docs/configuration/#supported-issuer-types
          [ACME]: https://datatracker.ietf.org/doc/html/rfc8555

          cert-manager also supports [external issuers][],
          which are out-of-tree components that provide support for other certificate authorities, such as:
          [Google Cloud Certificate Authority](https://github.com/jetstack/google-cas-issuer),
          [AWS Private Certificate Authority](https://github.com/cert-manager/aws-privateca-issuer), and
          [Cloudflare Origin CA](https://github.com/cloudflare/origin-ca-issuer)

          [external issuers]: https://cert-manager.io/docs/configuration/external/

          In order to configure cert-manager to issue certificates, you must first create an Issuer or ClusterIssuer.
          These resources represent a particular signing authority and tell cert-manager how to handle your certificate requests.
          The [Issuer Documentation][] has more detail.

          [configure cert-manager]: https://cert-manager.io/docs/configuration/
          [Issuer Documentation]: https://cert-manager.io/docs/concepts/issuer/

          Once an Issuer has been configured, you can create a Certificate resource,
          which is a definition of a certificate request that is to be honored by an issuer
          and which is to be kept up-to-date.
          This is the usual way that you will interact with cert-manager to request signed certificates.
          The [Certificate Documentation][] has more detail.

          [Certificate Documentation]: https://cert-manager.io/docs/usage/certificate/
        displayName: cert-manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Documentation
          url: https://cert-manager.io/
        - name: Release Notes
          url: https://cert-manager.io/docs/release-notes
        maintainers:
        - email: cert-manager-maintainers@googlegroups.com
          name: The cert-manager maintainers
        maturity: stable
        minKubeVersion: 1.19.0-0
        provider:
          name: The cert-manager maintainers
          url: https://cert-manager.io/
        relatedImages:
        - quay.io/jetstack/cert-manager-controller:v1.16.5
        - quay.io/jetstack/cert-manager-cainjector:v1.16.5
        - quay.io/jetstack/cert-manager-webhook:v1.16.5
        version: 1.16.5
      entries:
      - name: cert-manager.v1.16.5
        version: 1.16.5
      - name: cert-manager.v1.16.1
        version: 1.16.1
      - name: cert-manager.v1.16.0-beta.0
        version: 1.16.0-beta.0
      - name: cert-manager.v1.15.2
        version: 1.15.2
      - name: cert-manager.v1.15.0
        version: 1.15.0
      - name: cert-manager.v1.14.2
        version: 1.14.2
      - name: cert-manager.v1.14.2-rc1
        version: 1.14.2-rc1
      - name: cert-manager.v1.14.1-rc1
        version: 1.14.1-rc1
      - name: cert-manager.v1.13.3
        version: 1.13.3
      - name: cert-manager.v1.13.3-rc1
        version: 1.13.3-rc1
      - name: cert-manager.v1.13.1
        version: 1.13.1
      - name: cert-manager.v1.13.1-rc1
        version: 1.13.1-rc1
      - name: cert-manager.v1.12.2
        version: 1.12.2
      - name: cert-manager.v1.12.2-rc1
        version: 1.12.2-rc1
      - name: cert-manager.v1.11.4
        version: 1.11.4
      - name: cert-manager.v1.11.4-rc1
        version: 1.11.4-rc1
      - name: cert-manager.v1.11.3-rc1
        version: 1.11.3-rc1
      - name: cert-manager.v1.11.0
        version: 1.11.0
      - name: cert-manager.v1.11.0-rc1
        version: 1.11.0-rc1
      - name: cert-manager.v1.10.2
        version: 1.10.2
      - name: cert-manager.v1.10.2-rc1
        version: 1.10.2-rc1
      - name: cert-manager.v1.10.1
        version: 1.10.1
      - name: cert-manager.v1.10.1-rc1
        version: 1.10.1-rc1
      - name: cert-manager.v1.10.0
        version: 1.10.0
      - name: cert-manager.v1.10.0-rc1
        version: 1.10.0-rc1
      - name: cert-manager.v1.9.1
        version: 1.9.1
      - name: cert-manager.v1.9.1-rc1
        version: 1.9.1-rc1
      - name: cert-manager.v1.8.2
        version: 1.8.2
      - name: cert-manager.v1.8.2-rc1
        version: 1.8.2-rc1
      - name: cert-manager.v1.8.0
        version: 1.8.0
      - name: cert-manager.v1.8.0-rc1
        version: 1.8.0-rc1
      - name: cert-manager.v1.7.2
        version: 1.7.2
      - name: cert-manager.v1.7.1
        version: 1.7.1
      - name: cert-manager.v1.7.1-rc1
        version: 1.7.1-rc1
      - name: cert-manager.v1.6.2
        version: 1.6.2
      - name: cert-manager.v1.6.2-rc1
        version: 1.6.2-rc1
      name: candidate
    - currentCSV: cert-manager.v1.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "example-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "example-clusterissuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example-certificate"
                },
                "spec": {
                  "dnsNames": [
                    "example.com"
                  ],
                  "issuerRef": {
                    "name": "example-issuer"
                  },
                  "secretName": "example-certificate-tls"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "name": "example-certificaterequest"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "name": "example-order"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "example-challenge"
                },
                "spec": {}
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: quay.io/jetstack/cert-manager-controller:v1.16.5
          createdAt: 2025-05-02T09:11:31
          olm.skipRange: '>=1.16.0 <1.16.5'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/internal-objects: |-
            [
              "challenges.acme.cert-manager.io",
              "orders.acme.cert-manager.io"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/cert-manager/cert-manager
          support: The cert-manager maintainers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A CertificateRequest is used to request a signed certificate from one of the
              configured issuers.

              All fields within the CertificateRequest's `spec` are immutable after creation.
              A CertificateRequest will either succeed or fail, as denoted by its `Ready` status
              condition and its `status.failureTime` field.

              A CertificateRequest is a one-shot resource, meaning it represents a single
              point in time request for a certificate and cannot be re-used.
            displayName: CertificateRequest
            kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - description: |-
              A Certificate resource should be created to ensure an up to date and signed
              X.509 certificate is stored in the Kubernetes Secret resource named in `spec.secretName`.

              The stored certificate will be renewed before it expires (as configured by `spec.renewBefore`).
            displayName: Certificate
            kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: Challenge is a type to represent a Challenge request with
              an ACME server
            displayName: (Internal) Challenge
            kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - description: |-
              A ClusterIssuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is similar to an Issuer, however it is cluster-scoped and therefore can
              be referenced by resources that exist in *any* namespace, not just the same
              namespace as the referent.
            displayName: ClusterIssuer
            kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - description: |-
              An Issuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is scoped to a single namespace and can therefore only be referenced by
              resources within the same namespace.
            displayName: Issuer
            kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - description: Order is a type to represent an Order with an ACME server
            displayName: (Internal) Order
            kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          cert-manager is a Kubernetes add-on to automate the management and issuance of TLS certificates from various issuing sources.
          It will ensure certificates are valid and up to date periodically, and attempt to renew certificates at an appropriate time before expiry.

          cert-manager has the following [built in issuers][]:

          * ACME: sign certificates using Let's Encrypt (or any other [ACME][] compliant signer).
          * Vault: sign certificates using HashiCorp Vault.
          * Venafi: sign certificates using Venafi Cloud or Venafi Trust Protection Platform.
          * SelfSigned: the private key of the certificate will be used to sign the certificate.
          * CA: a private key stored as a Kubernetes Secret will be used to sign the certificate.

          [built in issuers]: https://cert-manager.io/docs/configuration/#supported-issuer-types
          [ACME]: https://datatracker.ietf.org/doc/html/rfc8555

          cert-manager also supports [external issuers][],
          which are out-of-tree components that provide support for other certificate authorities, such as:
          [Google Cloud Certificate Authority](https://github.com/jetstack/google-cas-issuer),
          [AWS Private Certificate Authority](https://github.com/cert-manager/aws-privateca-issuer), and
          [Cloudflare Origin CA](https://github.com/cloudflare/origin-ca-issuer)

          [external issuers]: https://cert-manager.io/docs/configuration/external/

          In order to configure cert-manager to issue certificates, you must first create an Issuer or ClusterIssuer.
          These resources represent a particular signing authority and tell cert-manager how to handle your certificate requests.
          The [Issuer Documentation][] has more detail.

          [configure cert-manager]: https://cert-manager.io/docs/configuration/
          [Issuer Documentation]: https://cert-manager.io/docs/concepts/issuer/

          Once an Issuer has been configured, you can create a Certificate resource,
          which is a definition of a certificate request that is to be honored by an issuer
          and which is to be kept up-to-date.
          This is the usual way that you will interact with cert-manager to request signed certificates.
          The [Certificate Documentation][] has more detail.

          [Certificate Documentation]: https://cert-manager.io/docs/usage/certificate/
        displayName: cert-manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Documentation
          url: https://cert-manager.io/
        - name: Release Notes
          url: https://cert-manager.io/docs/release-notes
        maintainers:
        - email: cert-manager-maintainers@googlegroups.com
          name: The cert-manager maintainers
        maturity: stable
        minKubeVersion: 1.19.0-0
        provider:
          name: The cert-manager maintainers
          url: https://cert-manager.io/
        relatedImages:
        - quay.io/jetstack/cert-manager-controller:v1.16.5
        - quay.io/jetstack/cert-manager-cainjector:v1.16.5
        - quay.io/jetstack/cert-manager-webhook:v1.16.5
        version: 1.16.5
      entries:
      - name: cert-manager.v1.16.5
        version: 1.16.5
      - name: cert-manager.v1.16.1
        version: 1.16.1
      - name: cert-manager.v1.15.2
        version: 1.15.2
      - name: cert-manager.v1.15.0
        version: 1.15.0
      - name: cert-manager.v1.14.2
        version: 1.14.2
      - name: cert-manager.v1.13.3
        version: 1.13.3
      - name: cert-manager.v1.13.1
        version: 1.13.1
      - name: cert-manager.v1.13.1-rc1
        version: 1.13.1-rc1
      - name: cert-manager.v1.12.2
        version: 1.12.2
      - name: cert-manager.v1.11.4
        version: 1.11.4
      - name: cert-manager.v1.11.0
        version: 1.11.0
      - name: cert-manager.v1.10.2
        version: 1.10.2
      - name: cert-manager.v1.10.1
        version: 1.10.1
      - name: cert-manager.v1.10.0
        version: 1.10.0
      - name: cert-manager.v1.9.1
        version: 1.9.1
      - name: cert-manager.v1.8.2
        version: 1.8.2
      - name: cert-manager.v1.8.0
        version: 1.8.0
      - name: cert-manager.v1.7.2
        version: 1.7.2
      - name: cert-manager.v1.7.1
        version: 1.7.1
      - name: cert-manager.v1.6.2
        version: 1.6.2
      - name: cert-manager.v1.6.1
        version: 1.6.1
      - name: cert-manager.v1.6.0
        version: 1.6.0
      - name: cert-manager.v1.5.4
        version: 1.5.4
      - name: cert-manager.v1.5.3
        version: 1.5.3
      - name: cert-manager.v1.4.4
        version: 1.4.4
      - name: cert-manager.v1.4.3
        version: 1.4.3
      - name: cert-manager.v1.4.2
        version: 1.4.2
      - name: cert-manager.v1.4.1
        version: 1.4.1
      - name: cert-manager.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: cert-manager
    provider:
      name: The cert-manager maintainers
      url: https://cert-manager.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.12
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-06-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:9aed98a061b2c76f3c512c1d6453df5920f0c779b33d7168c1e7ab9d8bc928a4
        version: 0.0.12
      entries:
      - name: joget-dx8-openshift-operator.v0.0.12
        version: 0.0.12
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Dell Technologies
      provider-url: https://github.com/dell/csm-operator
    name: dell-csm-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dell-csm-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "authorization",
                  "namespace": "authorization"
                },
                "spec": {
                  "modules": [
                    {
                      "components": [
                        {
                          "enabled": true,
                          "name": "nginx"
                        },
                        {
                          "enabled": true,
                          "name": "cert-manager"
                        },
                        {
                          "authorizationController": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231",
                          "authorizationControllerReplicas": 1,
                          "certificate": "",
                          "controllerReconcileInterval": "5m",
                          "enabled": true,
                          "hostname": "csm-authorization.com",
                          "leaderElection": true,
                          "name": "proxy-server",
                          "opa": "docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095",
                          "opaKubeMgmt": "docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517",
                          "openTelemetryCollectorAddress": "",
                          "privateKey": "",
                          "proxyServerIngress": [
                            {
                              "annotations": {},
                              "hosts": [],
                              "ingressClassName": "nginx"
                            }
                          ],
                          "proxyService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682",
                          "proxyServiceReplicas": 1,
                          "roleService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c",
                          "roleServiceReplicas": 1,
                          "storageService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5",
                          "storageServiceReplicas": 1,
                          "tenantService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358",
                          "tenantServiceReplicas": 1
                        },
                        {
                          "commander": "docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a",
                          "name": "redis",
                          "redis": "docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f",
                          "redisCommander": "rediscommander",
                          "redisName": "redis-csm",
                          "redisReplicas": 5,
                          "redisSecretProviderClass": [
                            {
                              "redisPasswordKey": "",
                              "redisSecretName": "",
                              "redisUsernameKey": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "sentinel": "sentinel"
                        },
                        {
                          "configSecretProviderClass": [
                            {
                              "configSecretName": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "name": "config"
                        },
                        {
                          "name": "storage-system-credentials",
                          "secretProviderClasses": {
                            "vault": [
                              "secret-provider-class-1",
                              "secret-provider-class-2"
                            ]
                          }
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": true,
                      "forceRemoveModule": true,
                      "name": "authorization-proxy-server"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "isilon",
                  "namespace": "isilon"
                },
                "spec": {
                  "driver": {
                    "authSecret": "isilon-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VERBOSE",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_ISI_PORT",
                          "value": "8080"
                        },
                        {
                          "name": "X_CSI_ISI_PATH",
                          "value": "/ifs/data/csi"
                        },
                        {
                          "name": "X_CSI_ISI_NO_PROBE_ON_START",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_AUTOPROBE",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_AUTH_TYPE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_CUSTOM_TOPOLOGY_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOISILON_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "AZ_RECONCILE_INTERVAL",
                          "value": "1h"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_ISI_QUOTA_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_ACCESS_ZONE",
                          "value": "System"
                        },
                        {
                          "name": "X_CSI_ISI_VOLUME_PATH_PERMISSIONS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_IGNORE_UNRESOLVABLE_HOSTS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "isilon",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerscale"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSCALE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_PERFORMANCE_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSCALE_QUOTA_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "ISICLIENT_INSECURE",
                              "value": "true"
                            },
                            {
                              "name": "ISICLIENT_AUTH_TYPE",
                              "value": "1"
                            },
                            {
                              "name": "ISICLIENT_VERBOSE",
                              "value": "0"
                            },
                            {
                              "name": "POWERSCALE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSCALE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0",
                          "name": "metrics-powerscale"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-isilon/csi_sock",
                            "--mode=node",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powermax",
                  "namespace": "powermax"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powermax-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_MANAGED_ARRAYS",
                          "value": "000000000001,000000000002"
                        },
                        {
                          "name": "X_CSI_POWERMAX_PORTGROUPS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_TRANSPORT_PROTOCOL",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_K8S_CLUSTER_PREFIX",
                          "value": "CSM"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "X_CSI_VSPHERE_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VSPHERE_PORTGROUP",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VSPHERE_HOSTNAME",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VCENTER_HOST",
                          "value": ""
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "X_CSI_POWERMAX_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "X_CSI_POWERMAX_DEBUG",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_REVPROXY_USE_SECRET",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powermax",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERMAX_ISCSI_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_TOPOLOGY_CONTROL_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": [
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/memory-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/disk-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/network-unavailable",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REVPROXY_TLS_SECRET",
                              "value": "csirevproxy-tls-secret"
                            },
                            {
                              "name": "X_CSI_REVPROXY_PORT",
                              "value": "2222"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            },
                            {
                              "name": "DeployAsSidecar",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265",
                          "name": "csipowermax-reverseproxy"
                        }
                      ],
                      "configVersion": "v2.14.0",
                      "name": "csireverseproxy"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powermax"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERMAX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERMAX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414",
                          "name": "metrics-powermax"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/powermax.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powerstore",
                  "namespace": "powerstore"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powerstore-config",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_NODE_NAME_PREFIX",
                          "value": "csi-node"
                        },
                        {
                          "name": "X_CSI_FC_PORTS_FILTER_FILE_PATH",
                          "value": "/etc/fc-ports-filter"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOPOWERSTORE_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSM_AUTH_ENABLED",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_MULTI_NAS_FAILURE_THRESHOLD",
                          "value": "5"
                        },
                        {
                          "name": "X_CSI_MULTI_NAS_COOLDOWN_PERIOD",
                          "value": "5m"
                        },
                        {
                          "name": "X_CSI_NFS_ACLS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerstore",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-powerstore.dellemc.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerstore"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSTORE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_SPACE_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_ARRAY_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_FILE_SYSTEM_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSTORE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSTORE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "ZIPKIN_URI",
                              "value": ""
                            },
                            {
                              "name": "ZIPKIN_SERVICE_NAME",
                              "value": "metrics-powerstore"
                            },
                            {
                              "name": "ZIPKIN_PROBABILITY",
                              "value": "0.0"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60",
                          "name": "metrics-powerstore"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "unity",
                  "namespace": "unity"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_UNITY_ALLOW_MULTI_POD_ACCESS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_EPHEMERAL_STAGING_PATH",
                          "value": "/var/lib/kubelet/plugins/kubernetes.io/csi/pv/"
                        },
                        {
                          "name": "X_CSI_ISCSI_CHROOT",
                          "value": "/noderoot"
                        },
                        {
                          "name": "X_CSI_UNITY_SYNC_NODEINFO_INTERVAL",
                          "value": "15"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOUNITY_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOUNITY_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "TENANT_NAME",
                          "value": ""
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_UNITY_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "unity",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "vxflexos",
                  "namespace": "vxflexos"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLELISTVOLUMESNAPSHOT",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLESNAPSHOTCGDELETE",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "INFO"
                        },
                        {
                          "name": "GOSCALEIO_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOSCALEIO_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_QUOTA_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "INTERFACE_NAMES",
                          "value": null
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERFLEX_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "File",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerflex",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "initContainers": [
                      {
                        "envs": [
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          },
                          {
                            "name": "REPO_ADDRESS",
                            "value": "sftp://0.0.0.0"
                          },
                          {
                            "name": "REPO_USER",
                            "value": "sdcRepoUser"
                          },
                          {
                            "name": "MODULE_SIGCHECK",
                            "value": "0"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "imagePullPolicy": "IfNotPresent",
                        "name": "sdc"
                      }
                    ],
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_SDC_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_APPROVE_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_PREFIX",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_SDC_SFTP_REPO_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 1,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "enabled": false,
                        "envs": [
                          {
                            "name": "HOST_PID",
                            "value": "1"
                          },
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "name": "sdc-monitor"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "csi-external-health-monitor-controller"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERFLEX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_SDC_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_SDC_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERFLEX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d",
                          "name": "metrics-powerflex"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerflex"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--csisock=unix:/var/lib/kubelet/plugins/vxflexos.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
          createdAt: "2025-09-05T13:18:42Z"
          description: Easily install and manage Dell’s CSI Drivers and CSM
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/dell/csm-operator
          support: Dell Technologies
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ContainerStorageModule is the Schema for the containerstoragemodules
              API
            displayName: Container Storage Module
            kind: ContainerStorageModule
            name: containerstoragemodules.storage.dell.com
            version: v1
        description: "Dell Container Storage Modules (CSM) Operator is a Kubernetes
          Operator which can be \nused to install and manage Dell’s CSI drivers and
          CSM modules. \nBy using Dell CSM Operator, enterprises can quickly and easily
          deploy the \nCSM modules for CSI drivers making it easy for DevOps \nteams
          to build and optimize their Kubernetes workloads using enterprise grade
          storage \nwhile consistently aligning with the desired outcomes of the organization.\nThis
          operator is available as both community and certified operator for Openshift
          \nand as a community operator for upstream Kubernetes.\nOperator can be
          deployed via OperatorHub.io. using Operator Lifecycle Manager (OLM)\n\n###
          About Dell CSM\n\nDell CSM brings powerful enterprise storage features and
          functionality to Kubernetes for \neasier adoption of cloud native workloads,
          improved productivity and scalable operations. \n\nCSM delivers software
          modules for storage management that provide developers access to \nbuild
          automation for enhanced IT needs and other critical enterprise storage features.
          \n\nCurrently, the following modules are available as part of Dell CSM.
          \n*\tObservability - Delivers a high-level view of the storage capacity
          \n  and performance usage via Grafana dashboards to the Kubernetes users\n*\tReplication
          - Enables array replication capabilities for Kubernetes users\n*\tAuthorization
          - Provides both storage and Kubernetes administrators the ability \n  to
          apply RBAC for Dell CSI Drivers\n*\tResiliency - Designed to make Kubernetes
          applications more resilient to \n  node failures\n\nThese modules are available
          for various storage platforms in the Dell portfolio, \nsuch as Dell PowerStore,
          Dell PowerMax, Dell PowerFlex, Dell PowerScale, and Dell Unity XT.\n\nFor
          more information, visit [Dell’s Container Storage Module web page](https://www.dell.com/en-us/dt/solutions/kubernetes-containers/kubernetes-storage.htm)\n\nAdditional
          Materials\n*\t[Documentation](https://dell.github.io/csm-docs)\n*\t[CSM
          Solution Brief](https://www.delltechnologies.com/asset/en-us/products/storage/briefs-summaries/csm-solution-brief.pdf)\n"
        displayName: Dell Container Storage Modules
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dell Container Storage Modules
        - Dell CSI Driver
        - Dell CSM Modules
        - Powerflex
        - Powerscale
        - Powerstore
        - Unity
        - Authorization
        - Observability
        - Replication
        links:
        - name: Documentation
          url: https://dell.github.io/csm-docs
        maintainers:
        - email: container.storage.modules@dell.com
          name: Dell Container Storage Modules
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Dell Technologies
          url: https://github.com/dell/csm-operator
        relatedImages:
        - docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a
        - registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
        - registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3
        - registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231
        - registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0
        - registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:59335acaca515fceed5f0350c46cfcfd6037475f0cf35ae8ab41d37333039e2e
        - ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60
        - docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682
        - registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16
        - registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265
        - registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c
        - quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358
        - quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403
        - docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517
        version: 1.10.0
      entries:
      - name: dell-csm-operator-certified.v1.10.0
        version: 1.10.0
      - name: dell-csm-operator-certified.v1.9.1
        version: 1.9.1
      - name: dell-csm-operator-certified.v1.9.0
        version: 1.9.0
      - name: dell-csm-operator-certified.v1.8.1
        version: 1.8.1
      - name: dell-csm-operator-certified.v1.8.0
        version: 1.8.0
      - name: dell-csm-operator-certified.v1.7.0
        version: 1.7.0
      - name: dell-csm-operator-certified.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: dell-csm-operator-certified
    provider:
      name: Dell Technologies
      url: https://github.com/dell/csm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: external-dns-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: external-dns-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "provider": {
                    "type": "AWS",
                    "aws": {
                      "credentials": {
                        "name": "aws-access-key"
                      }
                    }
                  },
                  "source": {
                    "type": "Service"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:6be1403e418f9b63ec8ebcc8858e602409bc41fdee4586acc627623a78b50eb5
          createdAt: 2022-03-01T21:00:46
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## Prerequisites and Requirements
           The ExternalDNS Operator has minimal cluster level permissions. Unfortunately OLM doesn't allow the local role creation in the operand namespace yet.
           Therefore make sure the following commands are run against your cluster before the installation of the operator:
           * Create the operand namespace: `oc create ns external-dns`
           * Grant the operator access to manage operand resources: `oc apply -f https://raw.githubusercontent.com/openshift/external-dns-operator/release-0.1/config/rbac/extra-roles.yaml`
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b6fde43491c2e0733c8fe6ee8660f7f6435fe20c0b6d435cfc7f86eeab8b4748
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:6be1403e418f9b63ec8ebcc8858e602409bc41fdee4586acc627623a78b50eb5
        - registry.redhat.io/edo/external-dns-rhel8@sha256:c1134bb46172997ef7278b6cefbb0da44e72a9f808a7cd67b3c65d464754cab9
        version: 0.1.2
      entries:
      - name: external-dns-operator.v0.1.2
        version: 0.1.2
      name: alpha
    - currentCSV: external-dns-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
          createdAt: 2023-11-15T09:24:04
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e33874a51971a90917cf437fc6cbeea1da569ba07217f66f1bba96d594f58aed
        - registry.redhat.io/edo/external-dns-rhel8@sha256:6263a78b2a170e56dfca2173f14dbf14f0a5bddb6ea600bfdef248c7d5f20e4b
        version: 1.2.0
      entries:
      - name: external-dns-operator.v1.2.0
        version: 1.2.0
      name: stable-v1
    - currentCSV: external-dns-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:0877eb0a111efe0bff63173a2dfeb7a06a61afd0efbcdba053322a539e0dc667
          createdAt: 2024-01-22T09:26:37
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/edo/external-dns-rhel8@sha256:17f4127cbb9a9304769e9655c19cbea2260107c7e9371a99a9aa3e85d759c00b
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:0877eb0a111efe0bff63173a2dfeb7a06a61afd0efbcdba053322a539e0dc667
        version: 1.0.1
      entries:
      - name: external-dns-operator.v1.0.1
        version: 1.0.1
      - name: external-dns-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: external-dns-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:9ca751e84b8843cc62dc60550e5c4af741bf09a381ef62220e7eb3760ffe8524
          createdAt: 2024-01-22T09:56:14
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:9ca751e84b8843cc62dc60550e5c4af741bf09a381ef62220e7eb3760ffe8524
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/edo/external-dns-rhel8@sha256:ffcc3f2e9b9feb889e7882465f7f034aba6b9ee0699a1b6c24e83f501a63ecca
        version: 1.1.1
      entries:
      - name: external-dns-operator.v1.1.1
        version: 1.1.1
      - name: external-dns-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: external-dns-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
          createdAt: 2023-11-15T09:24:04
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e33874a51971a90917cf437fc6cbeea1da569ba07217f66f1bba96d594f58aed
        - registry.redhat.io/edo/external-dns-rhel8@sha256:6263a78b2a170e56dfca2173f14dbf14f0a5bddb6ea600bfdef248c7d5f20e4b
        version: 1.2.0
      entries:
      - name: external-dns-operator.v1.2.0
        version: 1.2.0
      name: stable-v1.2
    defaultChannel: stable-v1
    packageName: external-dns-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: rook-ceph-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rook-ceph-operator.v4.16.16-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephCluster",
                "metadata": {
                  "name": "my-rook-ceph",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "cephVersion": {
                    "image": "registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d"
                  },
                  "dataDirHostPath": "/var/lib/rook",
                  "mon": {
                    "count": 3
                  },
                  "dashboard": {
                    "enabled": true
                  },
                  "network": {
                    "hostNetwork": false
                  },
                  "rbdMirroring": {
                    "workers": 0
                  },
                  "storage": {
                    "useAllNodes": true,
                    "useAllDevices": true
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephBlockPool",
                "metadata": {
                  "name": "replicapool",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "failureDomain": "host",
                  "replicated": {
                    "size": 3
                  },
                  "annotations": null
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStore",
                "metadata": {
                  "name": "my-store",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "metadataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "dataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "gateway": {
                    "type": "s3",
                    "sslCertificateRef": null,
                    "port": 8080,
                    "securePort": null,
                    "instances": 1,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStoreUser",
                "metadata": {
                  "name": "my-user",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "store": "my-store",
                  "displayName": "my display name"
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephNFS",
                "metadata": {
                  "name": "my-nfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "rados": {
                    "pool": "myfs-data0",
                    "namespace": "nfs-ns"
                  },
                  "server": {
                    "active": 3,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephClient",
                "metadata": {
                  "name": "cinder",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "caps": {
                    "mon": "profile rbd",
                    "osd": "profile rbd pool=volumes, profile rbd pool=vms, profile rbd-read-only pool=images"
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephFilesystem",
                "metadata": {
                  "name": "myfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "dataPools": [
                    {
                      "compressionMode": "",
                      "crushRoot": "",
                      "deviceClass": "",
                      "erasureCoded": {
                        "algorithm": "",
                        "codingChunks": 0,
                        "dataChunks": 0
                      },
                      "failureDomain": "host",
                      "replicated": {
                        "requireSafeReplicaSize": false,
                        "size": 1,
                        "targetSizeRatio": 0.5
                      }
                    }
                  ],
                  "metadataPool": {
                    "compressionMode": "",
                    "crushRoot": "",
                    "deviceClass": "",
                    "erasureCoded": {
                      "algorithm": "",
                      "codingChunks": 0,
                      "dataChunks": 0
                    },
                    "failureDomain": "",
                    "replicated": {
                      "requireSafeReplicaSize": false,
                      "size": 1,
                      "targetSizeRatio": 0
                    }
                  },
                  "metadataServer": {
                    "activeCount": 1,
                    "activeStandby": true,
                    "placement": {},
                    "resources": {}
                  },
                  "preservePoolsOnDelete": false,
                  "preserveFilesystemOnDelete": false
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephRBDMirror",
                "metadata": {
                  "name": "my-rbd-mirror",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "annotations": null,
                  "count": 1,
                  "placement": {
                    "topologyKey": "kubernetes.io/hostname"
                  },
                  "resources": null
                }
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:cf7f72469d25fb37301a8159406613673f8968c82af1035fd45a470e0ae8acfa
          externalClusterScript: |-
            IiIiCkNvcHlyaWdodCAyMDIwIFRoZSBSb29rIEF1dGhvcnMuIEFsbCByaWdodHMgcmVzZXJ2ZWQu
            CgpMaWNlbnNlZCB1bmRlciB0aGUgQXBhY2hlIExpY2Vuc2UsIFZlcnNpb24gMi4wICh0aGUgIkxp
            Y2Vuc2UiKTsKeW91IG1heSBub3QgdXNlIHRoaXMgZmlsZSBleGNlcHQgaW4gY29tcGxpYW5jZSB3
            aXRoIHRoZSBMaWNlbnNlLgpZb3UgbWF5IG9idGFpbiBhIGNvcHkgb2YgdGhlIExpY2Vuc2UgYXQK
            CglodHRwOi8vd3d3LmFwYWNoZS5vcmcvbGljZW5zZXMvTElDRU5TRS0yLjAKClVubGVzcyByZXF1
            aXJlZCBieSBhcHBsaWNhYmxlIGxhdyBvciBhZ3JlZWQgdG8gaW4gd3JpdGluZywgc29mdHdhcmUK
            ZGlzdHJpYnV0ZWQgdW5kZXIgdGhlIExpY2Vuc2UgaXMgZGlzdHJpYnV0ZWQgb24gYW4gIkFTIElT
            IiBCQVNJUywKV0lUSE9VVCBXQVJSQU5USUVTIE9SIENPTkRJVElPTlMgT0YgQU5ZIEtJTkQsIGVp
            dGhlciBleHByZXNzIG9yIGltcGxpZWQuClNlZSB0aGUgTGljZW5zZSBmb3IgdGhlIHNwZWNpZmlj
            IGxhbmd1YWdlIGdvdmVybmluZyBwZXJtaXNzaW9ucyBhbmQKbGltaXRhdGlvbnMgdW5kZXIgdGhl
            IExpY2Vuc2UuCiIiIgoKaW1wb3J0IGVycm5vCmltcG9ydCBzeXMKaW1wb3J0IGpzb24KaW1wb3J0
            IGFyZ3BhcnNlCmltcG9ydCByZQppbXBvcnQgc3VicHJvY2VzcwppbXBvcnQgaG1hYwpmcm9tIGhh
            c2hsaWIgaW1wb3J0IHNoYTEgYXMgc2hhCmZyb20gb3MgaW1wb3J0IGxpbmVzZXAgYXMgTElORVNF
            UApmcm9tIG9zIGltcG9ydCBwYXRoCmZyb20gZW1haWwudXRpbHMgaW1wb3J0IGZvcm1hdGRhdGUK
            aW1wb3J0IHJlcXVlc3RzCmZyb20gcmVxdWVzdHMuYXV0aCBpbXBvcnQgQXV0aEJhc2UKCnB5M2sg
            PSBGYWxzZQppZiBzeXMudmVyc2lvbl9pbmZvLm1ham9yID49IDM6CiAgICBweTNrID0gVHJ1ZQog
            ICAgaW1wb3J0IHVybGxpYi5wYXJzZQogICAgZnJvbSBpcGFkZHJlc3MgaW1wb3J0IGlwX2FkZHJl
            c3MsIElQdjRBZGRyZXNzCgpNb2R1bGVOb3RGb3VuZEVycm9yID0gSW1wb3J0RXJyb3IKCnRyeToK
            ICAgIGltcG9ydCByYWRvcwpleGNlcHQgTW9kdWxlTm90Rm91bmRFcnJvciBhcyBub01vZEVycjoK
            ICAgIHByaW50KGYiRXJyb3I6IHtub01vZEVycn1cbkV4aXRpbmcgdGhlIHNjcmlwdC4uLiIpCiAg
            ICBzeXMuZXhpdCgxKQoKdHJ5OgogICAgaW1wb3J0IHJiZApleGNlcHQgTW9kdWxlTm90Rm91bmRF
            cnJvciBhcyBub01vZEVycjoKICAgIHByaW50KGYiRXJyb3I6IHtub01vZEVycn1cbkV4aXRpbmcg
            dGhlIHNjcmlwdC4uLiIpCiAgICBzeXMuZXhpdCgxKQoKdHJ5OgogICAgIyBmb3IgMi43LngKICAg
            IGZyb20gU3RyaW5nSU8gaW1wb3J0IFN0cmluZ0lPCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9y
            OgogICAgIyBmb3IgMy54CiAgICBmcm9tIGlvIGltcG9ydCBTdHJpbmdJTwoKdHJ5OgogICAgIyBm
            b3IgMi43LngKICAgIGZyb20gdXJscGFyc2UgaW1wb3J0IHVybHBhcnNlCiAgICBmcm9tIHVybGxp
            YiBpbXBvcnQgdXJsZW5jb2RlIGFzIHVybGVuY29kZQpleGNlcHQgTW9kdWxlTm90Rm91bmRFcnJv
            cjoKICAgICMgZm9yIDMueAogICAgZnJvbSB1cmxsaWIucGFyc2UgaW1wb3J0IHVybHBhcnNlCiAg
            ICBmcm9tIHVybGxpYi5wYXJzZSBpbXBvcnQgdXJsZW5jb2RlIGFzIHVybGVuY29kZQoKdHJ5Ogog
            ICAgZnJvbSBiYXNlNjQgaW1wb3J0IGVuY29kZXN0cmluZwpleGNlcHQ6CiAgICBmcm9tIGJhc2U2
            NCBpbXBvcnQgZW5jb2RlYnl0ZXMgYXMgZW5jb2Rlc3RyaW5nCgoKY2xhc3MgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbihFeGNlcHRpb24pOgogICAgcGFzcwoKCiMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMgRHVtbXlSYWRv
            cyAjIyMjIyMjIyMjIyMjIyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjCiMgdGhpcyBpcyBtYWlubHkgZm9yIHRlc3RpbmcgYW5kIGNvdWxkIGJlIHVz
            ZWQgd2hlcmUgJ3JhZG9zJyBpcyBub3QgYXZhaWxhYmxlCgoKY2xhc3MgRHVtbXlSYWRvcyhvYmpl
            Y3QpOgogICAgZGVmIF9faW5pdF9fKHNlbGYpOgogICAgICAgIHNlbGYucmV0dXJuX3ZhbCA9IDAK
            ICAgICAgICBzZWxmLmVycl9tZXNzYWdlID0gIiIKICAgICAgICBzZWxmLnN0YXRlID0gImNvbm5l
            Y3RlZCIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwID0ge30KICAgICAgICBzZWxmLmNtZF9u
            YW1lcyA9IHt9CiAgICAgICAgc2VsZi5faW5pdF9jbWRfb3V0cHV0X21hcCgpCiAgICAgICAgc2Vs
            Zi5kdW1teV9ob3N0X2lwX21hcCA9IHt9CgogICAgZGVmIF9pbml0X2NtZF9vdXRwdXRfbWFwKHNl
            bGYpOgogICAgICAgIGpzb25fZmlsZV9uYW1lID0gInRlc3QtZGF0YS9jZXBoLXN0YXR1cy1vdXQi
            CiAgICAgICAgc2NyaXB0X2RpciA9IHBhdGguYWJzcGF0aChwYXRoLmRpcm5hbWUoX19maWxlX18p
            KQogICAgICAgIGNlcGhfc3RhdHVzX3N0ciA9ICIiCiAgICAgICAgd2l0aCBvcGVuKAogICAgICAg
            ICAgICBwYXRoLmpvaW4oc2NyaXB0X2RpciwganNvbl9maWxlX25hbWUpLCBtb2RlPSJyIiwgZW5j
            b2Rpbmc9IlVURi04IgogICAgICAgICkgYXMganNvbl9maWxlOgogICAgICAgICAgICBjZXBoX3N0
            YXR1c19zdHIgPSBqc29uX2ZpbGUucmVhZCgpCiAgICAgICAgc2VsZi5jbWRfbmFtZXNbImZzIGxz
            Il0gPSAiIiJ7ImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJmcyBscyJ9IiIiCiAgICAgICAg
            c2VsZi5jbWRfbmFtZXNbInF1b3J1bV9zdGF0dXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3Jt
            YXQiOiAianNvbiIsICJwcmVmaXgiOiAicXVvcnVtX3N0YXR1cyJ9IiIiCiAgICAgICAgKQogICAg
            ICAgIHNlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXSA9ICgKICAgICAgICAgICAgIiIieyJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAibWdyIHNlcnZpY2VzIn0iIiIKICAgICAgICApCiAg
            ICAgICAgIyBhbGwgdGhlIGNvbW1hbmRzIGFuZCB0aGVpciBvdXRwdXQKICAgICAgICBzZWxmLmNt
            ZF9vdXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJmcyBscyJdXSA9ICgKICAgICAgICAgICAgIiIi
            W3sibmFtZSI6Im15ZnMiLCJtZXRhZGF0YV9wb29sIjoibXlmcy1tZXRhZGF0YSIsIm1ldGFkYXRh
            X3Bvb2xfaWQiOjIsImRhdGFfcG9vbF9pZHMiOlszXSwiZGF0YV9wb29scyI6WyJteWZzLXJlcGxp
            Y2F0ZWQiXX1dIiIiCiAgICAgICAgKQogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbc2VsZi5j
            bWRfbmFtZXNbInF1b3J1bV9zdGF0dXMiXV0gPSAoCiAgICAgICAgICAgICIiInsiZWxlY3Rpb25f
            ZXBvY2giOjMsInF1b3J1bSI6WzBdLCJxdW9ydW1fbmFtZXMiOlsiYSJdLCJxdW9ydW1fbGVhZGVy
            X25hbWUiOiJhIiwicXVvcnVtX2FnZSI6MTQzODUsImZlYXR1cmVzIjp7InF1b3J1bV9jb24iOiI0
            NTQwMTM4MjkyODM2Njk2MDYzIiwicXVvcnVtX21vbiI6WyJrcmFrZW4iLCJsdW1pbm91cyIsIm1p
            bWljIiwib3NkbWFwLXBydW5lIiwibmF1dGlsdXMiLCJvY3RvcHVzIl19LCJtb25tYXAiOnsiZXBv
            Y2giOjEsImZzaWQiOiJhZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5ZDBmMWMiLCJtb2Rp
            ZmllZCI6IjIwMjAtMDUtMDdUMDM6MzY6MzkuOTE4MDM1WiIsImNyZWF0ZWQiOiIyMDIwLTA1LTA3
            VDAzOjM2OjM5LjkxODAzNVoiLCJtaW5fbW9uX3JlbGVhc2UiOjE1LCJtaW5fbW9uX3JlbGVhc2Vf
            bmFtZSI6Im9jdG9wdXMiLCJmZWF0dXJlcyI6eyJwZXJzaXN0ZW50IjpbImtyYWtlbiIsImx1bWlu
            b3VzIiwibWltaWMiLCJvc2RtYXAtcHJ1bmUiLCJuYXV0aWx1cyIsIm9jdG9wdXMiXSwib3B0aW9u
            YWwiOltdfSwibW9ucyI6W3sicmFuayI6MCwibmFtZSI6ImEiLCJwdWJsaWNfYWRkcnMiOnsiYWRk
            cnZlYyI6W3sidHlwZSI6InYyIiwiYWRkciI6IjEwLjExMC4yMDUuMTc0OjMzMDAiLCJub25jZSI6
            MH0seyJ0eXBlIjoidjEiLCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OSIsIm5vbmNlIjowfV19
            LCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OS8wIiwicHVibGljX2FkZHIiOiIxMC4xMTAuMjA1
            LjE3NDo2Nzg5LzAiLCJwcmlvcml0eSI6MCwid2VpZ2h0IjowfV19fSIiIgogICAgICAgICkKICAg
            ICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXV0g
            PSAoCiAgICAgICAgICAgICIiInsiZGFzaGJvYXJkIjoiaHR0cHM6Ly9jZXBoLWRhc2hib2FyZDo4
            NDQzLyIsInByb21ldGhldXMiOiJodHRwOi8vY2VwaC1kYXNoYm9hcmQtZGI6OTI4My8ifSIiIgog
            ICAgICAgICkKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNh
            cHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMiLCAib3Nk
            IiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0
            YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250
            cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJdLCAiZW50aXR5Ijog
            ImNsaWVudC5oZWFsdGhjaGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRo
            IGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuaGVh
            bHRoY2hlY2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRB
            PT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cyIs
            Im9zZCI6InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5j
            b250cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJ9fV0iIiIKICAg
            ICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIs
            ICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAib3NkIiwgInBy
            b2ZpbGUgcmJkIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtbm9kZSIsICJmb3JtYXQiOiAi
            anNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIi
            W3siZW50aXR5IjoiY2xpZW50LmNzaS1yYmQtbm9kZSIsImtleSI6IkFRQk9nck5lSGJLMUF4QUF1
            YllCZVY4UzFVL0dQenE1U1ZlcTZnPT0iLCJjYXBzIjp7Im1vbiI6InByb2ZpbGUgcmJkLCBhbGxv
            dyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsIm9zZCI6InByb2ZpbGUgcmJkIn19XSIiIgogICAg
            ICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsibW9uIiwg
            InByb2ZpbGUgcmJkLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxs
            b3cgcnciLCAib3NkIiwgInByb2ZpbGUgcmJkIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQt
            cHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNy
            ZWF0ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktcmJkLXByb3Zp
            c2lvbmVyIiwia2V5IjoiQVFCTmdyTmUxZ2V5S3hBQThla1ZpUmRFK2hzczVPd2VZQmt3Tmc9PSIs
            ImNhcHMiOnsibWdyIjoiYWxsb3cgcnciLCJtb24iOiJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFu
            ZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJwcm9maWxlIHJiZCJ9fV0iIiIKICAgICAgICBzZWxm
            LmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyBy
            LCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3Nk
            IiwgImFsbG93IHJ3IHRhZyBjZXBoZnMgKj0qIiwgIm1kcyIsICJhbGxvdyBydyJdLCAiZW50aXR5
            IjogImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4Ijog
            ImF1dGggZ2V0LW9yLWNyZWF0ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVu
            dC5jc2ktY2VwaGZzLW5vZGUiLCJrZXkiOiJBUUJPZ3JOZUVOdW5LeEFBUENtZ0U3UjZHOERjWG5h
            SjFGMzJxZz09IiwiY2FwcyI6eyJtZHMiOiJhbGxvdyBydyIsIm1nciI6ImFsbG93IHJ3IiwibW9u
            IjoiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBy
            dyB0YWcgY2VwaGZzICo9KiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAg
            ICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kICdvc2Qg
            YmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3NkIiwgImFsbG93IHJ3IHRhZyBjZXBo
            ZnMgbWV0YWRhdGE9KiJdLCAiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVy
            IiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgog
            ICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIs
            ImtleSI6IkFRQk9nck5lQUZnY0dCQUF2R3FLT0FEMEQzeHhtVlkwUjkxMmRnPT0iLCJjYXBzIjp7
            Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2Nr
            bGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoifX1dIiIiCiAgICAg
            ICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAi
            YWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdyIiwgImFsbG93IHJ3
            IiwgIm9zZCIsICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiXSwgImVudGl0eSI6ICJj
            bGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVuc2hpZnQtc3RvcmFnZSIsICJmb3JtYXQi
            OiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0g
            IiIiW3siZW50aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0
            b3JhZ2UiLCJrZXkiOiJCUUJPZ3JOZUFGZ2NHQkFBdkdxS09BRDBEM3h4bVZZMFI5MTJkZz09Iiwi
            Y2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29z
            ZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIn19XSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsi
            bW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciIsICJh
            bGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT1teWZzIl0sICJl
            bnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0b3JhZ2Ut
            bXlmcyIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0i
            IiIKICAgICAgICBdID0gIiIiW3siZW50aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9u
            ZXItb3BlbnNoaWZ0LXN0b3JhZ2UtbXlmcyIsImtleSI6IkNRQk9nck5lQUZnY0dCQUF2R3FLT0FE
            MEQzeHhtVlkwUjkxMmRnPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cg
            ciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2Vw
            aGZzIG1ldGFkYXRhPW15ZnMifX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAg
            ICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCBxdW9y
            dW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIiwgImFsbG93IGNvbW1hbmQg
            Y29uZmlnIiwgIm9zZCIsICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFsbG93IHJ3eCBwb29sPWRl
            ZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9ZGVm
            YXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJndy5sb2csIGFsbG93IHgg
            cG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4Il0sICJlbnRpdHkiOiAiY2xpZW50LmhlYWx0
            aGNoZWNrZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0
            ZSJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5oZWFsdGhjaGVja2VyIiwi
            a2V5IjoiQVFERmtiTmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsi
            bW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFu
            ZCB2ZXJzaW9uIiwgIm1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsICJvc2QiOiAicHJvZmls
            ZSByYmQtcmVhZC1vbmx5LCBhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhLCBhbGxvdyBy
            IHBvb2w9LnJndy5yb290LCBhbGxvdyBydyBwb29sPWRlZmF1bHQucmd3LmNvbnRyb2wsIGFsbG93
            IHJ4IHBvb2w9ZGVmYXVsdC5yZ3cubG9nLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0
            cy5pbmRleCJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAi
            IiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMs
            IGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cgY29tbWFuZCBjb25maWciLCAi
            b3NkIiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cu
            bWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5j
            b250cm9sLCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1
            bHQucmd3LmJ1Y2tldHMuaW5kZXgiXSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIs
            ICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBjYXBzIn0iIiIKICAgICAgICBdID0g
            IiIiW3siZW50aXR5IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0NWJG
            UkFBVG5kTE5VU1JLcnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIjog
            ImFsbG93IGNvbW1hbmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFs
            bG93IHJ3eCBwb29sPWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFs
            bG93IHJ3IHBvb2w9ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJn
            dy5sb2csIGFsbG93IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4In19XSIiIgogICAg
            ICAgIHNlbGYuY21kX291dHB1dF9tYXBbIiIieyJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            bWdyIHNlcnZpY2VzIn0iIiJdID0gKAogICAgICAgICAgICAiIiJ7ImRhc2hib2FyZCI6ICJodHRw
            Oi8vcm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6NzAwMC8iLCAicHJvbWV0aGV1cyI6
            ICJodHRwOi8vcm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6OTI4My8ifSIiIgogICAg
            ICAgICkKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0
            eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            YXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJ7ImRhc2hib2FyZCI6ICJodHRwOi8vcm9vay1j
            ZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6NzAwMC8iLCAicHJvbWV0aGV1cyI6ICJodHRwOi8v
            cm9vay1jZXBoLW1nci1hLTU3Y2Y5Zjg0YmMtZjRqbmw6OTI4My8ifSIiIgogICAgICAgIHNlbGYu
            Y21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNsaWVudC5oZWFsdGhj
            aGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldCJ9IiIiCiAgICAg
            ICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5oZWFsdGhjaGVja2VyIiwia2V5IjoiQVFERmti
            TmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjogImFsbG93
            IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxsb3cgY29tbWFuZCB2ZXJzaW9uIiwg
            Im1nciI6ICJhbGxvdyBjb21tYW5kIGNvbmZpZyIsICJvc2QiOiAicHJvZmlsZSByYmQtcmVhZC1v
            bmx5LCBhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhLCBhbGxvdyByIHBvb2w9LnJndy5y
            b290LCBhbGxvdyBydyBwb29sPWRlZmF1bHQucmd3LmNvbnRyb2wsIGFsbG93IHJ4IHBvb2w9ZGVm
            YXVsdC5yZ3cubG9nLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCJ9fV0i
            IiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0eSI6
            ICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJh
            dXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0
            X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtbm9kZSIsICJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJb
            XSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5
            IjogImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRf
            b3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMt
            cHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIK
            ICAgICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAg
            ICAgICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVuc2hp
            ZnQtc3RvcmFnZSIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgog
            ICAgICAgIF0gPSAiIiJbXSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAg
            ICAgICIiInsiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlm
            dC1zdG9yYWdlLW15ZnMiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0i
            IiIKICAgICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAg
            ICAgICAgICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJm
            b3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJb
            eyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsImtleSI6IkFRREZrYk5l
            ZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRBPT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIi
            LCAibWdyIjoiYWxsb3cgcnciLCAib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0q
            In19XSIiIgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2Fw
            cyI6IFsibW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1n
            ciIsICJhbGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIl0s
            ICJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCAiZm9ybWF0IjogImpz
            b24iLCAicHJlZml4IjogImF1dGggY2FwcyJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmtiTmVmdDViRlJBQVRu
            ZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cgciwgIGFsbG93IGNv
            bW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciI6ImFsbG93IHJ3IiwgIm9zZCI6ImFsbG93IHJ3
            IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFw
            Wyd7ImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJzdGF0dXMifSddID0gY2VwaF9zdGF0dXNf
            c3RyCgogICAgZGVmIHNodXRkb3duKHNlbGYpOgogICAgICAgIHBhc3MKCiAgICBkZWYgZ2V0X2Zz
            aWQoc2VsZik6CiAgICAgICAgcmV0dXJuICJhZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5
            ZDBmMWMiCgogICAgZGVmIGNvbmZfcmVhZF9maWxlKHNlbGYpOgogICAgICAgIHBhc3MKCiAgICBk
            ZWYgY29ubmVjdChzZWxmKToKICAgICAgICBwYXNzCgogICAgZGVmIHBvb2xfZXhpc3RzKHNlbGYs
            IHBvb2xfbmFtZSk6CiAgICAgICAgcmV0dXJuIFRydWUKCiAgICBkZWYgbW9uX2NvbW1hbmQoc2Vs
            ZiwgY21kLCBvdXQpOgogICAgICAgIGpzb25fY21kID0ganNvbi5sb2FkcyhjbWQpCiAgICAgICAg
            anNvbl9jbWRfc3RyID0ganNvbi5kdW1wcyhqc29uX2NtZCwgc29ydF9rZXlzPVRydWUpCiAgICAg
            ICAgY21kX291dHB1dCA9IHNlbGYuY21kX291dHB1dF9tYXBbanNvbl9jbWRfc3RyXQogICAgICAg
            IHJldHVybiBzZWxmLnJldHVybl92YWwsIGNtZF9vdXRwdXQsIHN0cihzZWxmLmVycl9tZXNzYWdl
            LmVuY29kZSgidXRmLTgiKSkKCiAgICBkZWYgX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoc2VsZiwg
            aG9zdF9uYW1lKToKICAgICAgICBpcF9yZWdfeCA9IHJlLmNvbXBpbGUociJcZHsxLDN9LlxkezEs
            M30uXGR7MSwzfS5cZHsxLDN9IikKICAgICAgICAjIGlmIHByb3ZpZGVkIGhvc3QgaXMgZGlyZWN0
            bHkgYW4gSVAgYWRkcmVzcywgcmV0dXJuIHRoZSBzYW1lCiAgICAgICAgaWYgaXBfcmVnX3gubWF0
            Y2goaG9zdF9uYW1lKToKICAgICAgICAgICAgcmV0dXJuIGhvc3RfbmFtZQogICAgICAgIGltcG9y
            dCByYW5kb20KCiAgICAgICAgaG9zdF9pcCA9IHNlbGYuZHVtbXlfaG9zdF9pcF9tYXAuZ2V0KGhv
            c3RfbmFtZSwgIiIpCiAgICAgICAgaWYgbm90IGhvc3RfaXA6CiAgICAgICAgICAgIGhvc3RfaXAg
            PSBmIjE3Mi45LntyYW5kb20ucmFuZGludCgwLCAyNTQpfS57cmFuZG9tLnJhbmRpbnQoMCwgMjU0
            KX0iCiAgICAgICAgICAgIHNlbGYuZHVtbXlfaG9zdF9pcF9tYXBbaG9zdF9uYW1lXSA9IGhvc3Rf
            aXAKICAgICAgICBkZWwgcmFuZG9tCiAgICAgICAgcmV0dXJuIGhvc3RfaXAKCiAgICBAY2xhc3Nt
            ZXRob2QKICAgIGRlZiBSYWRvcyhjb25mZmlsZT1Ob25lKToKICAgICAgICByZXR1cm4gRHVtbXlS
            YWRvcygpCgoKY2xhc3MgUzNBdXRoKEF1dGhCYXNlKToKICAgICIiIkF0dGFjaGVzIEFXUyBBdXRo
            ZW50aWNhdGlvbiB0byB0aGUgZ2l2ZW4gUmVxdWVzdCBvYmplY3QuIiIiCgogICAgc2VydmljZV9i
            YXNlX3VybCA9ICJzMy5hbWF6b25hd3MuY29tIgoKICAgIGRlZiBfX2luaXRfXyhzZWxmLCBhY2Nl
            c3Nfa2V5LCBzZWNyZXRfa2V5LCBzZXJ2aWNlX3VybD1Ob25lKToKICAgICAgICBpZiBzZXJ2aWNl
            X3VybDoKICAgICAgICAgICAgc2VsZi5zZXJ2aWNlX2Jhc2VfdXJsID0gc2VydmljZV91cmwKICAg
            ICAgICBzZWxmLmFjY2Vzc19rZXkgPSBzdHIoYWNjZXNzX2tleSkKICAgICAgICBzZWxmLnNlY3Jl
            dF9rZXkgPSBzdHIoc2VjcmV0X2tleSkKCiAgICBkZWYgX19jYWxsX18oc2VsZiwgcik6CiAgICAg
            ICAgIyBDcmVhdGUgZGF0ZSBoZWFkZXIgaWYgaXQgaXMgbm90IGNyZWF0ZWQgeWV0LgogICAgICAg
            IGlmICJkYXRlIiBub3QgaW4gci5oZWFkZXJzIGFuZCAieC1hbXotZGF0ZSIgbm90IGluIHIuaGVh
            ZGVyczoKICAgICAgICAgICAgci5oZWFkZXJzWyJkYXRlIl0gPSBmb3JtYXRkYXRlKHRpbWV2YWw9
            Tm9uZSwgbG9jYWx0aW1lPUZhbHNlLCB1c2VnbXQ9VHJ1ZSkKICAgICAgICBzaWduYXR1cmUgPSBz
            ZWxmLmdldF9zaWduYXR1cmUocikKICAgICAgICBpZiBweTNrOgogICAgICAgICAgICBzaWduYXR1
            cmUgPSBzaWduYXR1cmUuZGVjb2RlKCJ1dGYtOCIpCiAgICAgICAgci5oZWFkZXJzWyJBdXRob3Jp
            emF0aW9uIl0gPSBmIkFXUyB7c2VsZi5hY2Nlc3Nfa2V5fTp7c2lnbmF0dXJlfSIKICAgICAgICBy
            ZXR1cm4gcgoKICAgIGRlZiBnZXRfc2lnbmF0dXJlKHNlbGYsIHIpOgogICAgICAgIGNhbm9uaWNh
            bF9zdHJpbmcgPSBzZWxmLmdldF9jYW5vbmljYWxfc3RyaW5nKHIudXJsLCByLmhlYWRlcnMsIHIu
            bWV0aG9kKQogICAgICAgIGlmIHB5M2s6CiAgICAgICAgICAgIGtleSA9IHNlbGYuc2VjcmV0X2tl
            eS5lbmNvZGUoInV0Zi04IikKICAgICAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZy5lbmNv
            ZGUoInV0Zi04IikKICAgICAgICBlbHNlOgogICAgICAgICAgICBrZXkgPSBzZWxmLnNlY3JldF9r
            ZXkKICAgICAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZwogICAgICAgIGggPSBobWFjLm5l
            dyhrZXksIG1zZywgZGlnZXN0bW9kPXNoYSkKICAgICAgICByZXR1cm4gZW5jb2Rlc3RyaW5nKGgu
            ZGlnZXN0KCkpLnN0cmlwKCkKCiAgICBkZWYgZ2V0X2Nhbm9uaWNhbF9zdHJpbmcoc2VsZiwgdXJs
            LCBoZWFkZXJzLCBtZXRob2QpOgogICAgICAgIHBhcnNlZHVybCA9IHVybHBhcnNlKHVybCkKICAg
            ICAgICBvYmplY3RrZXkgPSBwYXJzZWR1cmwucGF0aFsxOl0KCiAgICAgICAgYnVja2V0ID0gcGFy
            c2VkdXJsLm5ldGxvY1s6IC1sZW4oc2VsZi5zZXJ2aWNlX2Jhc2VfdXJsKV0KICAgICAgICBpZiBs
            ZW4oYnVja2V0KSA+IDE6CiAgICAgICAgICAgICMgcmVtb3ZlIGxhc3QgZG90CiAgICAgICAgICAg
            IGJ1Y2tldCA9IGJ1Y2tldFs6LTFdCgogICAgICAgIGludGVyZXN0aW5nX2hlYWRlcnMgPSB7ImNv
            bnRlbnQtbWQ1IjogIiIsICJjb250ZW50LXR5cGUiOiAiIiwgImRhdGUiOiAiIn0KICAgICAgICBm
            b3Iga2V5IGluIGhlYWRlcnM6CiAgICAgICAgICAgIGxrID0ga2V5Lmxvd2VyKCkKICAgICAgICAg
            ICAgdHJ5OgogICAgICAgICAgICAgICAgbGsgPSBsay5kZWNvZGUoInV0Zi04IikKICAgICAgICAg
            ICAgZXhjZXB0OgogICAgICAgICAgICAgICAgcGFzcwogICAgICAgICAgICBpZiBoZWFkZXJzW2tl
            eV0gYW5kICgKICAgICAgICAgICAgICAgIGxrIGluIGludGVyZXN0aW5nX2hlYWRlcnMua2V5cygp
            IG9yIGxrLnN0YXJ0c3dpdGgoIngtYW16LSIpCiAgICAgICAgICAgICk6CiAgICAgICAgICAgICAg
            ICBpbnRlcmVzdGluZ19oZWFkZXJzW2xrXSA9IGhlYWRlcnNba2V5XS5zdHJpcCgpCgogICAgICAg
            ICMgSWYgeC1hbXotZGF0ZSBpcyB1c2VkIGl0IHN1cGVyc2VkZXMgdGhlIGRhdGUgaGVhZGVyLgog
            ICAgICAgIGlmIG5vdCBweTNrOgogICAgICAgICAgICBpZiAieC1hbXotZGF0ZSIgaW4gaW50ZXJl
            c3RpbmdfaGVhZGVyczoKICAgICAgICAgICAgICAgIGludGVyZXN0aW5nX2hlYWRlcnNbImRhdGUi
            XSA9ICIiCiAgICAgICAgZWxzZToKICAgICAgICAgICAgaWYgIngtYW16LWRhdGUiIGluIGludGVy
            ZXN0aW5nX2hlYWRlcnM6CiAgICAgICAgICAgICAgICBpbnRlcmVzdGluZ19oZWFkZXJzWyJkYXRl
            Il0gPSAiIgoKICAgICAgICBidWYgPSBmInttZXRob2R9XG4iCiAgICAgICAgZm9yIGtleSBpbiBz
            b3J0ZWQoaW50ZXJlc3RpbmdfaGVhZGVycy5rZXlzKCkpOgogICAgICAgICAgICB2YWwgPSBpbnRl
            cmVzdGluZ19oZWFkZXJzW2tleV0KICAgICAgICAgICAgaWYga2V5LnN0YXJ0c3dpdGgoIngtYW16
            LSIpOgogICAgICAgICAgICAgICAgYnVmICs9IGYie2tleX06e3ZhbH1cbiIKICAgICAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgICAgIGJ1ZiArPSBmInt2YWx9XG4iCgogICAgICAgICMgYXBwZW5k
            IHRoZSBidWNrZXQgaWYgaXQgZXhpc3RzCiAgICAgICAgaWYgYnVja2V0ICE9ICIiOgogICAgICAg
            ICAgICBidWYgKz0gZiIve2J1Y2tldH0iCgogICAgICAgICMgYWRkIHRoZSBvYmplY3RrZXkuIGV2
            ZW4gaWYgaXQgZG9lc24ndCBleGlzdCwgYWRkIHRoZSBzbGFzaAogICAgICAgIGJ1ZiArPSBmIi97
            b2JqZWN0a2V5fSIKCiAgICAgICAgcmV0dXJuIGJ1ZgoKCmNsYXNzIFJhZG9zSlNPTjoKICAgIEVY
            VEVSTkFMX1VTRVJfTkFNRSA9ICJjbGllbnQuaGVhbHRoY2hlY2tlciIKICAgIEVYVEVSTkFMX1JH
            V19BRE1JTl9PUFNfVVNFUl9OQU1FID0gInJndy1hZG1pbi1vcHMtdXNlciIKICAgIEVNUFRZX09V
            VFBVVF9MSVNUID0gIkVtcHR5IG91dHB1dCBsaXN0IgogICAgREVGQVVMVF9SR1dfUE9PTF9QUkVG
            SVggPSAiZGVmYXVsdCIKICAgIERFRkFVTFRfTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUID0gIjky
            ODMiCgogICAgQGNsYXNzbWV0aG9kCiAgICBkZWYgZ2VuX2FyZ19wYXJzZXIoY2xzLCBhcmdzX3Rv
            X3BhcnNlPU5vbmUpOgogICAgICAgIGFyZ1AgPSBhcmdwYXJzZS5Bcmd1bWVudFBhcnNlcigpCgog
            ICAgICAgIGNvbW1vbl9ncm91cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJjb21tb24iKQog
            ICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoIi0tdmVyYm9zZSIsICItdiIsIGFjdGlv
            bj0ic3RvcmVfdHJ1ZSIsIGRlZmF1bHQ9RmFsc2UpCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9h
            cmd1bWVudCgKICAgICAgICAgICAgIi0tY2VwaC1jb25mIiwgIi1jIiwgaGVscD0iUHJvdmlkZSBh
            IGNlcGggY29uZiBmaWxlLiIsIHR5cGU9c3RyCiAgICAgICAgKQogICAgICAgIGNvbW1vbl9ncm91
            cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLWtleXJpbmciLCAiLWsiLCBoZWxwPSJQYXRo
            IHRvIGNlcGgga2V5cmluZyBmaWxlLiIsIHR5cGU9c3RyCiAgICAgICAgKQogICAgICAgIGNvbW1v
            bl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJ1bi1hcy11c2VyIiwKICAgICAg
            ICAgICAgIi11IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgdHlwZT1zdHIs
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIGEgdXNlciBuYW1lIHRvIGNoZWNrIHRoZSBjbHVz
            dGVyJ3MgaGVhbHRoIHN0YXR1cywgbXVzdCBiZSBwcmVmaXhlZCBieSAnY2xpZW50LiciLAogICAg
            ICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1j
            bHVzdGVyLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJL
            dWJlcm5ldGVzIGNsdXN0ZXIgbmFtZShsZWdhY3kgZmxhZyksIE5vdGU6IEVpdGhlciB1c2UgdGhp
            cyBvciAtLWs4cy1jbHVzdGVyLW5hbWUiLAogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAu
            YWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1rOHMtY2x1c3Rlci1uYW1lIiwgZGVmYXVsdD0i
            IiwgaGVscD0iS3ViZXJuZXRlcyBjbHVzdGVyIG5hbWUiCiAgICAgICAgKQogICAgICAgIGNvbW1v
            bl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLW5hbWVzcGFjZSIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIGhlbHA9Ik5hbWVzcGFjZSB3aGVyZSBDZXBoQ2x1
            c3RlciBpcyBydW5uaW5nIiwKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tcmd3LXBvb2wtcHJlZml4IiwgZGVmYXVsdD0iIiwgaGVscD0i
            UkdXIFBvb2wgcHJlZml4IgogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3Vt
            ZW50KAogICAgICAgICAgICAiLS1yZXN0cmljdGVkLWF1dGgtcGVybWlzc2lvbiIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlJlc3RyaWN0IGNlcGhDU0lLZXly
            aW5ncyBhdXRoIHBlcm1pc3Npb25zIHRvIHNwZWNpZmljIHBvb2xzLCBjbHVzdGVyLiIKICAgICAg
            ICAgICAgKyAiTWFuZGF0b3J5IGZsYWdzIHRoYXQgbmVlZCB0byBiZSBzZXQgYXJlIC0tcmJkLWRh
            dGEtcG9vbC1uYW1lLCBhbmQgLS1rOHMtY2x1c3Rlci1uYW1lLiIKICAgICAgICAgICAgKyAiLS1j
            ZXBoZnMtZmlsZXN5c3RlbS1uYW1lIGZsYWcgY2FuIGFsc28gYmUgcGFzc2VkIGluIGNhc2Ugb2Yg
            Y2VwaGZzIHVzZXIgcmVzdHJpY3Rpb24sIHNvIGl0IGNhbiByZXN0cmljdCB1c2VyIHRvIHBhcnRp
            Y3VsYXIgY2VwaGZzIGZpbGVzeXN0ZW0iCiAgICAgICAgICAgICsgInNhbXBsZSBydW46IGBweXRo
            b24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwtY2x1c3Rlci1yZXNvdXJjZXMucHkgLS1jZXBo
            ZnMtZmlsZXN5c3RlbS1uYW1lIG15ZnMgLS1yYmQtZGF0YS1wb29sLW5hbWUgcmVwbGljYXBvb2wg
            LS1rOHMtY2x1c3Rlci1uYW1lIHJvb2tzdG9yYWdlIC0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Np
            b24gdHJ1ZWAiCiAgICAgICAgICAgICsgIk5vdGU6IFJlc3RyaWN0aW5nIHRoZSBjc2ktdXNlcnMg
            cGVyIHBvb2wsIGFuZCBwZXIgY2x1c3RlciB3aWxsIHJlcXVpcmUgY3JlYXRpbmcgbmV3IGNzaS11
            c2VycyBhbmQgbmV3IHNlY3JldHMgZm9yIHRoYXQgY3NpLXVzZXJzLiIKICAgICAgICAgICAgKyAi
            U28gYXBwbHkgdGhlc2Ugc2VjcmV0cyBvbmx5IHRvIG5ldyBgQ29uc3VtZXIgY2x1c3RlcmAgZGVw
            bG95bWVudCB3aGlsZSB1c2luZyB0aGUgc2FtZSBgU291cmNlIGNsdXN0ZXJgLiIsCiAgICAgICAg
            KQogICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXYyLXBv
            cnQtZW5hYmxlIiwKICAgICAgICAgICAgYWN0aW9uPSJzdG9yZV90cnVlIiwKICAgICAgICAgICAg
            ZGVmYXVsdD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iRW5hYmxlIHYyIG1vbiBwb3J0KDMzMDAp
            IGZvciBtb25zIiwKICAgICAgICApCgogICAgICAgIG91dHB1dF9ncm91cCA9IGFyZ1AuYWRkX2Fy
            Z3VtZW50X2dyb3VwKCJvdXRwdXQiKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQo
            CiAgICAgICAgICAgICItLWZvcm1hdCIsCiAgICAgICAgICAgICItdCIsCiAgICAgICAgICAgIGNo
            b2ljZXM9WyJqc29uIiwgImJhc2giXSwKICAgICAgICAgICAgZGVmYXVsdD0ianNvbiIsCiAgICAg
            ICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBvdXRwdXQgZm9ybWF0IChqc29uIHwgYmFzaCkiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1vdXRwdXQiLAogICAgICAgICAgICAiLW8iLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAg
            ICAgICAgICBoZWxwPSJPdXRwdXQgd2lsbCBiZSBzdG9yZWQgaW50byB0aGUgcHJvdmlkZWQgZmls
            ZSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAog
            ICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgQ2VwaCBmaWxlc3lzdGVt
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tY2VwaGZzLW1ldGFkYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIs
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBjZXBoZnMgbWV0YWRh
            dGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAg
            ICAgICAgICAgICItLWNlcGhmcy1kYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9
            IiIsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBjZXBoZnMgZGF0
            YSBwb29sIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAg
            ICAgICAgICAgIi0tcmJkLWRhdGEtcG9vbC1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwK
            ICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIHRo
            ZSBuYW1lIG9mIHRoZSBSQkQgZGF0YXBvb2wiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3Jv
            dXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1hbGlhcy1yYmQtZGF0YS1wb29sLW5hbWUi
            LAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAg
            ICAgICAgICAgaGVscD0iUHJvdmlkZXMgYW4gYWxpYXMgZm9yIHRoZSAgUkJEIGRhdGEgcG9vbCBu
            YW1lLCBuZWNlc3NhcnkgaWYgYSBzcGVjaWFsIGNoYXJhY3RlciBpcyBwcmVzZW50IGluIHRoZSBw
            b29sIG5hbWUgc3VjaCBhcyBhIHBlcmlvZCBvciB1bmRlcnNjb3JlIiwKICAgICAgICApCiAgICAg
            ICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmd3LWVuZHBvaW50
            IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAg
            ICAgICAgICAgIGhlbHA9IlJBRE9TIEdhdGV3YXkgZW5kcG9pbnQgKGluIGA8SVB2ND46PFBPUlQ+
            YCBvciBgPFtJUHY2XT46PFBPUlQ+YCBvciBgPEZRRE4+OjxQT1JUPmAgZm9ybWF0KSIsCiAgICAg
            ICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJn
            dy10bHMtY2VydC1wYXRoIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVx
            dWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IlJBRE9TIEdhdGV3YXkgZW5kcG9pbnQgVExT
            IGNlcnRpZmljYXRlIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVu
            dCgKICAgICAgICAgICAgIi0tcmd3LXNraXAtdGxzIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFs
            c2UsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9Iklnbm9yZSBU
            TFMgY2VydGlmaWNhdGlvbiB2YWxpZGF0aW9uIHdoZW4gYSBzZWxmLXNpZ25lZCBjZXJ0aWZpY2F0
            ZSBpcyBwcm92aWRlZCAoTk9UIFJFQ09NTUVOREVEIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tbW9uaXRvcmluZy1lbmRwb2ludCIs
            CiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAg
            ICAgICAgICBoZWxwPSJDZXBoIE1hbmFnZXIgcHJvbWV0aGV1cyBleHBvcnRlciBlbmRwb2ludHMg
            KGNvbW1hIHNlcGFyYXRlZCBsaXN0IG9mIChmb3JtYXQgYDxJUHY0PmAgb3IgYDxbSVB2Nl0+YCBv
            ciBgPEZRRE4+YCkgZW50cmllcyBvZiBhY3RpdmUgYW5kIHN0YW5kYnkgbWdycykiLAogICAgICAg
            ICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1tb25p
            dG9yaW5nLWVuZHBvaW50LXBvcnQiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAg
            ICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iQ2VwaCBNYW5hZ2VyIHByb21ldGhl
            dXMgZXhwb3J0ZXIgcG9ydCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJn
            dW1lbnQoCiAgICAgICAgICAgICItLXNraXAtbW9uaXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAg
            ICAgICAgIGhlbHA9IkRvIG5vdCBjaGVjayBmb3IgYSBtb25pdG9yaW5nIGVuZHBvaW50IGZvciB0
            aGUgQ2VwaCBjbHVzdGVyIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tcmJkLW1ldGFkYXRhLWVjLXBvb2wtbmFtZSIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBo
            ZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiBlcmFzdXJlIGNvZGVkIFJCRCBtZXRhZGF0YSBwb29s
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tZHJ5LXJ1biIsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFj
            dGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGhlbHA9IkRyeSBydW4gcHJpbnRzIHRoZSBl
            eGVjdXRlZCBjb21tYW5kcyB3aXRob3V0IHJ1bm5pbmcgdGhlbSIsCiAgICAgICAgKQogICAgICAg
            IG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJhZG9zLW5hbWVzcGFj
            ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAog
            ICAgICAgICAgICBoZWxwPSJEaXZpZGVzIGEgcG9vbCBpbnRvIHNlcGFyYXRlIGxvZ2ljYWwgbmFt
            ZXNwYWNlcywgdXNlZCBmb3IgY3JlYXRpbmcgUkJEIFBWQyBpbiBhIENlcGhCbG9ja1Bvb2xSYWRv
            c05hbWVzcGFjZSAoc2hvdWxkIGJlIGxvd2VyIGNhc2UpIiwKICAgICAgICApCiAgICAgICAgb3V0
            cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tc3Vidm9sdW1lLWdyb3VwIiwK
            ICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAg
            ICAgICAgIGhlbHA9InByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBzdWJ2b2x1bWUgZ3JvdXAiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1yZ3ctcmVhbG0tbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJl
            cXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUg
            cmd3LXJlYWxtIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgK
            ICAgICAgICAgICAgIi0tcmd3LXpvbmUtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAg
            ICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUg
            bmFtZSBvZiB0aGUgcmd3LXpvbmUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS1yZ3ctem9uZWdyb3VwLW5hbWUiLAogICAgICAgICAg
            ICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVs
            cD0icHJvdmlkZXMgdGhlIG5hbWUgb2YgdGhlIHJndy16b25lZ3JvdXAiLAogICAgICAgICkKICAg
            ICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS10b3BvbG9neS1w
            b29scyIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNl
            LAogICAgICAgICAgICBoZWxwPSJjb21tYS1zZXBhcmF0ZWQgbGlzdCBvZiB0b3BvbG9neS1jb25z
            dHJhaW5lZCByYmQgcG9vbHMiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2Fy
            Z3VtZW50KAogICAgICAgICAgICAiLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi1sYWJlbCIsCiAg
            ICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAg
            ICAgICBoZWxwPSJrOHMgY2x1c3RlciBmYWlsdXJlIGRvbWFpbiBsYWJlbCAoZXhhbXBsZTogem9u
            ZSwgcmFjaywgb3IgaG9zdCkgZm9yIHRoZSB0b3BvbG9neS1wb29scyB0aGF0IG1hdGNoIHRoZSBj
            ZXBoIGRvbWFpbiIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQo
            CiAgICAgICAgICAgICItLXRvcG9sb2d5LWZhaWx1cmUtZG9tYWluLXZhbHVlcyIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBo
            ZWxwPSJjb21tYS1zZXBhcmF0ZWQgbGlzdCBvZiB0aGUgazhzIGNsdXN0ZXIgZmFpbHVyZSBkb21h
            aW4gdmFsdWVzIGNvcnJlc3BvbmRpbmcgdG8gZWFjaCBvZiB0aGUgcG9vbHMgaW4gdGhlIGB0b3Bv
            bG9neS1wb29sc2AgbGlzdCIsCiAgICAgICAgKQoKICAgICAgICB1cGdyYWRlX2dyb3VwID0gYXJn
            UC5hZGRfYXJndW1lbnRfZ3JvdXAoInVwZ3JhZGUiKQogICAgICAgIHVwZ3JhZGVfZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS11cGdyYWRlIiwKICAgICAgICAgICAgYWN0aW9uPSJz
            dG9yZV90cnVlIiwKICAgICAgICAgICAgZGVmYXVsdD1GYWxzZSwKICAgICAgICAgICAgaGVscD0i
            VXBncmFkZXMgdGhlIGNlcGhDU0lLZXlyaW5ncyhGb3IgZXhhbXBsZTogY2xpZW50LmNzaS1jZXBo
            ZnMtcHJvdmlzaW9uZXIpIGFuZCBjbGllbnQuaGVhbHRoY2hlY2tlciBjZXBoIHVzZXJzIHdpdGgg
            bmV3IHBlcm1pc3Npb25zIG5lZWRlZCBmb3IgdGhlIG5ldyBjbHVzdGVyIHZlcnNpb24gYW5kIG9s
            ZGVyIHBlcm1pc3Npb24gd2lsbCBzdGlsbCBiZSBhcHBsaWVkLiIKICAgICAgICAgICAgKyAiU2Ft
            cGxlIHJ1bjogYHB5dGhvbjMgL2V0Yy9jZXBoL2NyZWF0ZS1leHRlcm5hbC1jbHVzdGVyLXJlc291
            cmNlcy5weSAtLXVwZ3JhZGVgLCB0aGlzIHdpbGwgdXBncmFkZSBhbGwgdGhlIGRlZmF1bHQgY3Np
            IHVzZXJzKG5vbi1yZXN0cmljdGVkKSIKICAgICAgICAgICAgKyAiRm9yIHJlc3RyaWN0ZWQgdXNl
            cnMoRm9yIGV4YW1wbGU6IGNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1z
            dG9yYWdlLW15ZnMpLCB1c2VycyBjcmVhdGVkIHVzaW5nIC0tcmVzdHJpY3RlZC1hdXRoLXBlcm1p
            c3Npb24gZmxhZyBuZWVkIHRvIHBhc3MgbWFuZGF0b3J5IGZsYWdzIgogICAgICAgICAgICArICJt
            YW5kYXRvcnkgZmxhZ3M6ICctLXJiZC1kYXRhLXBvb2wtbmFtZSwgLS1rOHMtY2x1c3Rlci1uYW1l
            IGFuZCAtLXJ1bi1hcy11c2VyJyBmbGFncyB3aGlsZSB1cGdyYWRpbmciCiAgICAgICAgICAgICsg
            ImluIGNhc2Ugb2YgY2VwaGZzIHVzZXJzIGlmIHlvdSBoYXZlIHBhc3NlZCAtLWNlcGhmcy1maWxl
            c3lzdGVtLW5hbWUgZmxhZyB3aGlsZSBjcmVhdGluZyB1c2VyIHRoZW4gd2hpbGUgdXBncmFkaW5n
            IGl0IHdpbGwgYmUgbWFuZGF0b3J5IHRvbyIKICAgICAgICAgICAgKyAiU2FtcGxlIHJ1bjogYHB5
            dGhvbjMgL2V0Yy9jZXBoL2NyZWF0ZS1leHRlcm5hbC1jbHVzdGVyLXJlc291cmNlcy5weSAtLXVw
            Z3JhZGUgLS1yYmQtZGF0YS1wb29sLW5hbWUgcmVwbGljYXBvb2wgLS1rOHMtY2x1c3Rlci1uYW1l
            IHJvb2tzdG9yYWdlICAtLXJ1bi1hcy11c2VyIGNsaWVudC5jc2ktcmJkLW5vZGUtcm9va3N0b3Jh
            Z2UtcmVwbGljYXBvb2xgIgogICAgICAgICAgICArICJQUzogQW4gZXhpc3Rpbmcgbm9uLXJlc3Ry
            aWN0ZWQgdXNlciBjYW5ub3QgYmUgY29udmVydGVkIHRvIGEgcmVzdHJpY3RlZCB1c2VyIGJ5IHVw
            Z3JhZGluZy4iCiAgICAgICAgICAgICsgIlVwZ3JhZGUgZmxhZyBzaG91bGQgb25seSBiZSB1c2Vk
            IHRvIGFwcGVuZCBuZXcgcGVybWlzc2lvbnMgdG8gdXNlcnMsIGl0IHNob3VsZG4ndCBiZSB1c2Vk
            IGZvciBjaGFuZ2luZyB1c2VyIGFscmVhZHkgYXBwbGllZCBwZXJtaXNzaW9uLCBmb3IgZXhhbXBs
            ZSB5b3Ugc2hvdWxkbid0IGNoYW5nZSBpbiB3aGljaCBwb29sIHVzZXIgaGFzIGFjY2VzcyIsCiAg
            ICAgICAgKQoKICAgICAgICBpZiBhcmdzX3RvX3BhcnNlOgogICAgICAgICAgICBhc3NlcnQgKAog
            ICAgICAgICAgICAgICAgdHlwZShhcmdzX3RvX3BhcnNlKSA9PSBsaXN0CiAgICAgICAgICAgICks
            ICJBcmd1bWVudCB0byAnZ2VuX2FyZ19wYXJzZXInIHNob3VsZCBiZSBhIGxpc3QiCiAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgYXJnc190b19wYXJzZSA9IHN5cy5hcmd2WzE6XQogICAgICAgIHJl
            dHVybiBhcmdQLnBhcnNlX2FyZ3MoYXJnc190b19wYXJzZSkKCiAgICBkZWYgdmFsaWRhdGVfcmJk
            X21ldGFkYXRhX2VjX3Bvb2xfbmFtZShzZWxmKToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2Vy
            LnJiZF9tZXRhZGF0YV9lY19wb29sX25hbWU6CiAgICAgICAgICAgIHJiZF9tZXRhZGF0YV9lY19w
            b29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUKICAg
            ICAgICAgICAgcmJkX3Bvb2xfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9u
            YW1lCgogICAgICAgICAgICBpZiByYmRfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiRmxh
            ZyAnLS1yYmQtZGF0YS1wb29sLW5hbWUnIHNob3VsZCBub3QgYmUgZW1wdHkiCiAgICAgICAgICAg
            ICAgICApCgogICAgICAgICAgICBpZiByYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lID09ICIiOgog
            ICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAg
            ICAgICAgICAgICAiRmxhZyAnLS1yYmQtbWV0YWRhdGEtZWMtcG9vbC1uYW1lJyBzaG91bGQgbm90
            IGJlIGVtcHR5IgogICAgICAgICAgICAgICAgKQoKICAgICAgICAgICAgY21kX2pzb24gPSB7InBy
            ZWZpeCI6ICJvc2QgZHVtcCIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgICAgIHJldF92YWws
            IGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikK
            ICAgICAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAg
            ICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAg
            ICAgICAgZiJ7Y21kX2pzb25bJ3ByZWZpeCddfSBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAg
            ICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVN
            UFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgbWV0YWRhdGFf
            cG9vbF9leGlzdCwgcG9vbF9leGlzdCA9IEZhbHNlLCBGYWxzZQoKICAgICAgICAgICAgZm9yIGtl
            eSBpbiBqc29uX291dFsicG9vbHMiXToKICAgICAgICAgICAgICAgICMgaWYgZXJhc3VyZV9jb2Rl
            X3Byb2ZpbGUgaXMgZW1wdHkgYW5kIHBvb2wgbmFtZSBleGlzdHMgdGhlbiBpdCByZXBsaWNhIHBv
            b2wKICAgICAgICAgICAgICAgIGlmICgKICAgICAgICAgICAgICAgICAgICBrZXlbImVyYXN1cmVf
            Y29kZV9wcm9maWxlIl0gPT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQga2V5WyJwb29sX25h
            bWUiXSA9PSByYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lCiAgICAgICAgICAgICAgICApOgogICAg
            ICAgICAgICAgICAgICAgIG1ldGFkYXRhX3Bvb2xfZXhpc3QgPSBUcnVlCiAgICAgICAgICAgICAg
            ICAjIGlmIGVyYXN1cmVfY29kZV9wcm9maWxlIGlzIG5vdCBlbXB0eSBhbmQgcG9vbCBuYW1lIGV4
            aXN0cyB0aGVuIGl0IGlzIGVjIHBvb2wKICAgICAgICAgICAgICAgIGlmIGtleVsiZXJhc3VyZV9j
            b2RlX3Byb2ZpbGUiXSBhbmQga2V5WyJwb29sX25hbWUiXSA9PSByYmRfcG9vbF9uYW1lOgogICAg
            ICAgICAgICAgICAgICAgIHBvb2xfZXhpc3QgPSBUcnVlCgogICAgICAgICAgICBpZiBub3QgbWV0
            YWRhdGFfcG9vbF9leGlzdDoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVF
            eGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIlByb3ZpZGVkIHJiZF9lY19tZXRhZGF0YV9w
            b29sIG5hbWUsIgogICAgICAgICAgICAgICAgICAgIGYiIHtyYmRfbWV0YWRhdGFfZWNfcG9vbF9u
            YW1lfSwgZG9lcyBub3QgZXhpc3QiCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmIG5v
            dCBwb29sX2V4aXN0OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICBmIlByb3ZpZGVkIHJiZF9kYXRhX3Bvb2wgbmFtZSwg
            e3JiZF9wb29sX25hbWV9LCBkb2VzIG5vdCBleGlzdCIKICAgICAgICAgICAgICAgICkKICAgICAg
            ICAgICAgcmV0dXJuIHJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUKCiAgICBkZWYgZHJ5X3J1bihz
            ZWxmLCBtc2cpOgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAg
            ICAgcHJpbnQoIkV4ZWN1dGU6ICIgKyAiJyIgKyBtc2cgKyAiJyIpCgogICAgZGVmIHZhbGlkYXRl
            X3Jnd19lbmRwb2ludF90bHNfY2VydChzZWxmKToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd190bHNfY2VydF9wYXRoOgogICAgICAgICAgICB3aXRoIG9wZW4oc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfdGxzX2NlcnRfcGF0aCwgZW5jb2Rpbmc9InV0ZjgiKSBhcyBmOgogICAgICAgICAgICAg
            ICAgY29udGVudHMgPSBmLnJlYWQoKQogICAgICAgICAgICAgICAgcmV0dXJuIGNvbnRlbnRzLnJz
            dHJpcCgpCgogICAgZGVmIF9jaGVja19jb25mbGljdGluZ19vcHRpb25zKHNlbGYpOgogICAgICAg
            IGlmIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnVwZ3JhZGUgYW5kIG5vdCBzZWxmLl9hcmdfcGFyc2Vy
            LnJiZF9kYXRhX3Bvb2xfbmFtZToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICJFaXRoZXIgJy0tdXBncmFkZScgb3IgJy0tcmJkLWRh
            dGEtcG9vbC1uYW1lIDxwb29sX25hbWU+JyBzaG91bGQgYmUgc3BlY2lmaWVkIgogICAgICAgICAg
            ICApCgogICAgZGVmIF9pbnZhbGlkX2VuZHBvaW50KHNlbGYsIGVuZHBvaW50X3N0cik6CiAgICAg
            ICAgIyBleHRyYWN0IHRoZSBwb3J0IGJ5IGdldHRpbmcgdGhlIGxhc3Qgc3BsaXQgb24gYDpgIGRl
            bGltaXRlcgogICAgICAgIHRyeToKICAgICAgICAgICAgZW5kcG9pbnRfc3RyX2lwLCBwb3J0ID0g
            ZW5kcG9pbnRfc3RyLnJzcGxpdCgiOiIsIDEpCiAgICAgICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAg
            ICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oZiJOb3QgYSBwcm9wZXIg
            ZW5kcG9pbnQ6IHtlbmRwb2ludF9zdHJ9IikKCiAgICAgICAgdHJ5OgogICAgICAgICAgICBpZiBl
            bmRwb2ludF9zdHJfaXBbMF0gPT0gIlsiOgogICAgICAgICAgICAgICAgZW5kcG9pbnRfc3RyX2lw
            ID0gZW5kcG9pbnRfc3RyX2lwWzEgOiBsZW4oZW5kcG9pbnRfc3RyX2lwKSAtIDFdCiAgICAgICAg
            ICAgIGlwX3R5cGUgPSAoCiAgICAgICAgICAgICAgICAiSVB2NCIgaWYgdHlwZShpcF9hZGRyZXNz
            KGVuZHBvaW50X3N0cl9pcCkpIGlzIElQdjRBZGRyZXNzIGVsc2UgIklQdjYiCiAgICAgICAgICAg
            ICkKICAgICAgICBleGNlcHQgVmFsdWVFcnJvcjoKICAgICAgICAgICAgaXBfdHlwZSA9ICJGUURO
            IgogICAgICAgIGlmIG5vdCBwb3J0LmlzZGlnaXQoKToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0
            aW9uRmFpbHVyZUV4Y2VwdGlvbihmIlBvcnQgbm90IHZhbGlkOiB7cG9ydH0iKQogICAgICAgIGlu
            dFBvcnQgPSBpbnQocG9ydCkKICAgICAgICBpZiBpbnRQb3J0IDwgMSBvciBpbnRQb3J0ID4gMioq
            MTYgLSAxOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKGYiT3V0
            IG9mIHJhbmdlIHBvcnQgbnVtYmVyOiB7cG9ydH0iKQoKICAgICAgICByZXR1cm4gaXBfdHlwZQoK
            ICAgIGRlZiBlbmRwb2ludF9kaWFsKHNlbGYsIGVuZHBvaW50X3N0ciwgaXBfdHlwZSwgdGltZW91
            dD0zLCBjZXJ0PU5vbmUpOgogICAgICAgICMgaWYgdGhlICdjbHVzdGVyJyBpbnN0YW5jZSBpcyBh
            IGR1bW15IG9uZSwKICAgICAgICAjIGRvbid0IHRyeSB0byByZWFjaCBvdXQgdG8gdGhlIGVuZHBv
            aW50CiAgICAgICAgaWYgaXNpbnN0YW5jZShzZWxmLmNsdXN0ZXIsIER1bW15UmFkb3MpOgogICAg
            ICAgICAgICByZXR1cm4gIiIsICIiLCAiIgogICAgICAgIGlmIGlwX3R5cGUgPT0gIklQdjYiOgog
            ICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBlbmRwb2ludF9zdHJfaXAsIGVuZHBvaW50
            X3N0cl9wb3J0ID0gZW5kcG9pbnRfc3RyLnJzcGxpdCgiOiIsIDEpCiAgICAgICAgICAgIGV4Y2Vw
            dCBWYWx1ZUVycm9yOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICBmIk5vdCBhIHByb3BlciBlbmRwb2ludDoge2VuZHBv
            aW50X3N0cn0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGlmIGVuZHBvaW50X3N0cl9p
            cFswXSAhPSAiWyI6CiAgICAgICAgICAgICAgICBlbmRwb2ludF9zdHJfaXAgPSAiWyIgKyBlbmRw
            b2ludF9zdHJfaXAgKyAiXSIKICAgICAgICAgICAgZW5kcG9pbnRfc3RyID0gIjoiLmpvaW4oW2Vu
            ZHBvaW50X3N0cl9pcCwgZW5kcG9pbnRfc3RyX3BvcnRdKQoKICAgICAgICBwcm90b2NvbHMgPSBb
            Imh0dHAiLCAiaHR0cHMiXQogICAgICAgIHJlc3BvbnNlX2Vycm9yID0gTm9uZQogICAgICAgIGZv
            ciBwcmVmaXggaW4gcHJvdG9jb2xzOgogICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBl
            cCA9IGYie3ByZWZpeH06Ly97ZW5kcG9pbnRfc3RyfSIKICAgICAgICAgICAgICAgIHZlcmlmeSA9
            IE5vbmUKICAgICAgICAgICAgICAgICMgSWYgdmVyaWZ5IGlzIHNldCB0byBhIHBhdGggdG8gYSBk
            aXJlY3RvcnksCiAgICAgICAgICAgICAgICAjIHRoZSBkaXJlY3RvcnkgbXVzdCBoYXZlIGJlZW4g
            cHJvY2Vzc2VkIHVzaW5nIHRoZSBjX3JlaGFzaCB1dGlsaXR5IHN1cHBsaWVkIHdpdGggT3BlblNT
            TC4KICAgICAgICAgICAgICAgIGlmIHByZWZpeCA9PSAiaHR0cHMiIGFuZCBzZWxmLl9hcmdfcGFy
            c2VyLnJnd19za2lwX3RsczoKICAgICAgICAgICAgICAgICAgICB2ZXJpZnkgPSBGYWxzZQogICAg
            ICAgICAgICAgICAgICAgIHIgPSByZXF1ZXN0cy5oZWFkKGVwLCB0aW1lb3V0PXRpbWVvdXQsIHZl
            cmlmeT1GYWxzZSkKICAgICAgICAgICAgICAgIGVsaWYgcHJlZml4ID09ICJodHRwcyIgYW5kIGNl
            cnQ6CiAgICAgICAgICAgICAgICAgICAgdmVyaWZ5ID0gY2VydAogICAgICAgICAgICAgICAgICAg
            IHIgPSByZXF1ZXN0cy5oZWFkKGVwLCB0aW1lb3V0PXRpbWVvdXQsIHZlcmlmeT1jZXJ0KQogICAg
            ICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgICAgICByID0gcmVxdWVzdHMuaGVhZChl
            cCwgdGltZW91dD10aW1lb3V0KQogICAgICAgICAgICAgICAgaWYgci5zdGF0dXNfY29kZSA9PSAy
            MDA6CiAgICAgICAgICAgICAgICAgICAgcmV0dXJuIHByZWZpeCwgdmVyaWZ5LCAiIgogICAgICAg
            ICAgICBleGNlcHQgRXhjZXB0aW9uIGFzIGVycjoKICAgICAgICAgICAgICAgIHJlc3BvbnNlX2Vy
            cm9yID0gZXJyCiAgICAgICAgICAgICAgICBjb250aW51ZQogICAgICAgIHN5cy5zdGRlcnIud3Jp
            dGUoCiAgICAgICAgICAgIGYidW5hYmxlIHRvIGNvbm5lY3QgdG8gZW5kcG9pbnQ6IHtlbmRwb2lu
            dF9zdHJ9LCBmYWlsZWQgZXJyb3I6IHtyZXNwb25zZV9lcnJvcn0iCiAgICAgICAgKQogICAgICAg
            IHJldHVybiAoCiAgICAgICAgICAgICIiLAogICAgICAgICAgICAiIiwKICAgICAgICAgICAgKCIt
            MSIpLAogICAgICAgICkKCiAgICBkZWYgX19pbml0X18oc2VsZiwgYXJnX2xpc3Q9Tm9uZSk6CiAg
            ICAgICAgc2VsZi5vdXRfbWFwID0ge30KICAgICAgICBzZWxmLl9leGNsdWRlZF9rZXlzID0gc2V0
            KCkKICAgICAgICBzZWxmLl9hcmdfcGFyc2VyID0gc2VsZi5nZW5fYXJnX3BhcnNlcihhcmdzX3Rv
            X3BhcnNlPWFyZ19saXN0KQogICAgICAgIHNlbGYuX2NoZWNrX2NvbmZsaWN0aW5nX29wdGlvbnMo
            KQogICAgICAgIHNlbGYucnVuX2FzX3VzZXIgPSBzZWxmLl9hcmdfcGFyc2VyLnJ1bl9hc191c2Vy
            CiAgICAgICAgc2VsZi5vdXRwdXRfZmlsZSA9IHNlbGYuX2FyZ19wYXJzZXIub3V0cHV0CiAgICAg
            ICAgc2VsZi5jZXBoX2NvbmYgPSBzZWxmLl9hcmdfcGFyc2VyLmNlcGhfY29uZgogICAgICAgIHNl
            bGYuY2VwaF9rZXlyaW5nID0gc2VsZi5fYXJnX3BhcnNlci5rZXlyaW5nCiAgICAgICAgIyBpZiB1
            c2VyIG5vdCBwcm92aWRlZCwgZ2l2ZSBhIGRlZmF1bHQgdXNlcgogICAgICAgIGlmIG5vdCBzZWxm
            LnJ1bl9hc191c2VyIGFuZCBub3Qgc2VsZi5fYXJnX3BhcnNlci51cGdyYWRlOgogICAgICAgICAg
            ICBzZWxmLnJ1bl9hc191c2VyID0gc2VsZi5FWFRFUk5BTF9VU0VSX05BTUUKICAgICAgICBpZiBu
            b3Qgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXggYW5kIG5vdCBzZWxmLl9hcmdfcGFy
            c2VyLnVwZ3JhZGU6CiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4
            ID0gc2VsZi5ERUZBVUxUX1JHV19QT09MX1BSRUZJWAogICAgICAgIGlmIHNlbGYuY2VwaF9jb25m
            OgogICAgICAgICAgICBrd2FyZ3MgPSB7fQogICAgICAgICAgICBpZiBzZWxmLmNlcGhfa2V5cmlu
            ZzoKICAgICAgICAgICAgICAgIGt3YXJnc1siY29uZiJdID0geyJrZXlyaW5nIjogc2VsZi5jZXBo
            X2tleXJpbmd9CiAgICAgICAgICAgIHNlbGYuY2x1c3RlciA9IHJhZG9zLlJhZG9zKGNvbmZmaWxl
            PXNlbGYuY2VwaF9jb25mLCAqKmt3YXJncykKICAgICAgICBlbHNlOgogICAgICAgICAgICBzZWxm
            LmNsdXN0ZXIgPSByYWRvcy5SYWRvcygpCiAgICAgICAgICAgIHNlbGYuY2x1c3Rlci5jb25mX3Jl
            YWRfZmlsZSgpCiAgICAgICAgc2VsZi5jbHVzdGVyLmNvbm5lY3QoKQoKICAgIGRlZiBzaHV0ZG93
            bihzZWxmKToKICAgICAgICBpZiBzZWxmLmNsdXN0ZXIuc3RhdGUgPT0gImNvbm5lY3RlZCI6CiAg
            ICAgICAgICAgIHNlbGYuY2x1c3Rlci5zaHV0ZG93bigpCgogICAgZGVmIGdldF9mc2lkKHNlbGYp
            OgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJu
            IHNlbGYuZHJ5X3J1bigiY2VwaCBmc2lkIikKICAgICAgICByZXR1cm4gc3RyKHNlbGYuY2x1c3Rl
            ci5nZXRfZnNpZCgpKQoKICAgIGRlZiBfY29tbW9uX2NtZF9qc29uX2dlbihzZWxmLCBjbWRfanNv
            bik6CiAgICAgICAgY21kID0ganNvbi5kdW1wcyhjbWRfanNvbiwgc29ydF9rZXlzPVRydWUpCiAg
            ICAgICAgcmV0X3ZhbCwgY21kX291dCwgZXJyX21zZyA9IHNlbGYuY2x1c3Rlci5tb25fY29tbWFu
            ZChjbWQsIGIiIikKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnZlcmJvc2U6CiAgICAgICAg
            ICAgIHByaW50KGYiQ29tbWFuZCBJbnB1dDoge2NtZH0iKQogICAgICAgICAgICBwcmludCgKICAg
            ICAgICAgICAgICAgIGYiUmV0dXJuIFZhbDoge3JldF92YWx9XG5Db21tYW5kIE91dHB1dDoge2Nt
            ZF9vdXR9XG4iCiAgICAgICAgICAgICAgICBmIkVycm9yIE1lc3NhZ2U6IHtlcnJfbXNnfVxuLS0t
            LS0tLS0tLVxuIgogICAgICAgICAgICApCiAgICAgICAganNvbl9vdXQgPSB7fQogICAgICAgICMg
            aWYgdGhlcmUgaXMgbm8gZXJyb3IgKGkuZTsgcmV0X3ZhbCBpcyBaRVJPKSBhbmQgJ2NtZF9vdXQn
            IGlzIG5vdCBlbXB0eQogICAgICAgICMgdGhlbiBjb252ZXJ0ICdjbWRfb3V0JyB0byBhIGpzb24g
            b3V0cHV0CiAgICAgICAgaWYgcmV0X3ZhbCA9PSAwIGFuZCBjbWRfb3V0OgogICAgICAgICAgICBq
            c29uX291dCA9IGpzb24ubG9hZHMoY21kX291dCkKICAgICAgICByZXR1cm4gcmV0X3ZhbCwganNv
            bl9vdXQsIGVycl9tc2cKCiAgICBkZWYgZ2V0X2NlcGhfZXh0ZXJuYWxfbW9uX2RhdGEoc2VsZik6
            CiAgICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJxdW9ydW1fc3RhdHVzIiwgImZvcm1hdCI6
            ICJqc29uIn0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAg
            IHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGggIiArIGNtZF9qc29uWyJwcmVmaXgiXSkKICAgICAg
            ICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4o
            Y21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwK
            ICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgIidxdW9y
            dW1fc3RhdHVzJyBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAgICAgIGYiRXJyb3I6IHtl
            cnJfbXNnIGlmIHJldF92YWwgIT0gMCBlbHNlIHNlbGYuRU1QVFlfT1VUUFVUX0xJU1R9IgogICAg
            ICAgICAgICApCiAgICAgICAgcV9sZWFkZXJfbmFtZSA9IGpzb25fb3V0WyJxdW9ydW1fbGVhZGVy
            X25hbWUiXQogICAgICAgIHFfbGVhZGVyX2RldGFpbHMgPSB7fQogICAgICAgIHFfbGVhZGVyX21h
            dGNoaW5nX2xpc3QgPSBbCiAgICAgICAgICAgIGwgZm9yIGwgaW4ganNvbl9vdXRbIm1vbm1hcCJd
            WyJtb25zIl0gaWYgbFsibmFtZSJdID09IHFfbGVhZGVyX25hbWUKICAgICAgICBdCiAgICAgICAg
            aWYgbGVuKHFfbGVhZGVyX21hdGNoaW5nX2xpc3QpID09IDA6CiAgICAgICAgICAgIHJhaXNlIEV4
            ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oIk5vIG1hdGNoaW5nICdtb24nIGRldGFpbHMgZm91bmQi
            KQogICAgICAgIHFfbGVhZGVyX2RldGFpbHMgPSBxX2xlYWRlcl9tYXRjaGluZ19saXN0WzBdCiAg
            ICAgICAgIyBnZXQgdGhlIGFkZHJlc3MgdmVjdG9yIG9mIHRoZSBxdW9ydW0tbGVhZGVyCiAgICAg
            ICAgcV9sZWFkZXJfYWRkcnZlYyA9IHFfbGVhZGVyX2RldGFpbHMuZ2V0KCJwdWJsaWNfYWRkcnMi
            LCB7fSkuZ2V0KCJhZGRydmVjIiwgW10pCiAgICAgICAgaXBfYWRkciA9IHN0cihxX2xlYWRlcl9k
            ZXRhaWxzWyJwdWJsaWNfYWRkciJdLnNwbGl0KCIvIilbMF0pCgogICAgICAgIGlmIHNlbGYuX2Fy
            Z19wYXJzZXIudjJfcG9ydF9lbmFibGU6CiAgICAgICAgICAgIGlmIHFfbGVhZGVyX2FkZHJ2ZWNb
            MF1bInR5cGUiXSA9PSAidjIiOgogICAgICAgICAgICAgICAgaXBfYWRkciA9IHFfbGVhZGVyX2Fk
            ZHJ2ZWNbMF1bImFkZHIiXQogICAgICAgICAgICBlbGlmIGxlbihxX2xlYWRlcl9hZGRydmVjKSA+
            IDEgYW5kIHFfbGVhZGVyX2FkZHJ2ZWNbMV1bInR5cGUiXSA9PSAidjIiOgogICAgICAgICAgICAg
            ICAgaXBfYWRkciA9IHFfbGVhZGVyX2FkZHJ2ZWNbMV1bImFkZHIiXQogICAgICAgICAgICBlbHNl
            OgogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZSgKICAgICAgICAgICAgICAgICAgICAi
            J3YyJyBhZGRyZXNzIHR5cGUgbm90IHByZXNlbnQsIGFuZCAndjItcG9ydC1lbmFibGUnIGZsYWcg
            aXMgcHJvdmlkZWQiCiAgICAgICAgICAgICAgICApCgogICAgICAgIHJldHVybiBmIntzdHIocV9s
            ZWFkZXJfbmFtZSl9PXtpcF9hZGRyfSIKCiAgICBkZWYgX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAo
            c2VsZiwgaG9zdF9uYW1lLCBwb3J0LCBpcF90eXBlKToKICAgICAgICAjIGlmICdjbHVzdGVyJyBp
            bnN0YW5jZSBpcyBhIGR1bW15IHR5cGUsCiAgICAgICAgIyBjYWxsIHRoZSBkdW1teSBpbnN0YW5j
            ZSdzICJjb252ZXJ0IiBtZXRob2QKICAgICAgICBpZiBub3QgaG9zdF9uYW1lOgogICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKCJFbXB0eSBob3N0bmFtZSBwcm92aWRl
            ZCIpCiAgICAgICAgaWYgaXNpbnN0YW5jZShzZWxmLmNsdXN0ZXIsIER1bW15UmFkb3MpOgogICAg
            ICAgICAgICByZXR1cm4gc2VsZi5jbHVzdGVyLl9jb252ZXJ0X2hvc3RuYW1lX3RvX2lwKGhvc3Rf
            bmFtZSkKCiAgICAgICAgaWYgaXBfdHlwZSA9PSAiRlFETiI6CiAgICAgICAgICAgICMgY2hlY2sg
            d2hpY2ggaXAgRlFETiBzaG91bGQgYmUgY29udmVydGVkIHRvLCBJUHY0IG9yIElQdjYKICAgICAg
            ICAgICAgIyBjaGVjayB0aGUgaG9zdCBpcCwgdGhlIGVuZHBvaW50IGlwIHR5cGUgd291bGQgYmUg
            c2ltaWxhciB0byBob3N0IGlwCiAgICAgICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAib3Jj
            aCBob3N0IGxzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICAgICAgcmV0X3ZhbCwganNvbl9v
            dXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAg
            ICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAogICAgICAgICAgICBp
            ZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICAgICAgcmFp
            c2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJ29yY2gg
            aG9zdCBscycgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAgICAgICAgICAgICAgICAgZiJFcnJvcjog
            e2Vycl9tc2cgaWYgcmV0X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAg
            ICAgICAgICAgICAgICApCiAgICAgICAgICAgIGhvc3RfYWRkciA9IGpzb25fb3V0WzBdWyJhZGRy
            Il0KICAgICAgICAgICAgIyBhZGQgOjgwIHNhbXBsZSBwb3J0IGluIGlwX3R5cGUsIGFzIF9pbnZh
            bGlkX2VuZHBvaW50IGFsc28gdmVyaWZ5IHBvcnQKICAgICAgICAgICAgaG9zdF9pcF90eXBlID0g
            c2VsZi5faW52YWxpZF9lbmRwb2ludChob3N0X2FkZHIgKyAiOjgwIikKICAgICAgICAgICAgaW1w
            b3J0IHNvY2tldAoKICAgICAgICAgICAgIyBleGFtcGxlIG91dHB1dCBbKDxBZGRyZXNzRmFtaWx5
            LkFGX0lORVQ6IDI+LCA8U29ja2V0S2luZC5TT0NLX1NUUkVBTTogMT4sIDYsICcnLCAoJzkzLjE4
            NC4yMTYuMzQnLCA4MCkpLCAuLi5dCiAgICAgICAgICAgICMgd2UgbmVlZCB0byBnZXQgOTMuMTg0
            LjIxNi4zNCBzbyBpdCB3b3VsZCBiZSBpcFswXVs0XVswXQogICAgICAgICAgICBpZiBob3N0X2lw
            X3R5cGUgPT0gIklQdjYiOgogICAgICAgICAgICAgICAgaXAgPSBzb2NrZXQuZ2V0YWRkcmluZm8o
            CiAgICAgICAgICAgICAgICAgICAgaG9zdF9uYW1lLCBwb3J0LCBmYW1pbHk9c29ja2V0LkFGX0lO
            RVQ2LCBwcm90bz1zb2NrZXQuSVBQUk9UT19UQ1AKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgZWxpZiBob3N0X2lwX3R5cGUgPT0gIklQdjQiOgogICAgICAgICAgICAgICAgaXAgPSBzb2Nr
            ZXQuZ2V0YWRkcmluZm8oCiAgICAgICAgICAgICAgICAgICAgaG9zdF9uYW1lLCBwb3J0LCBmYW1p
            bHk9c29ja2V0LkFGX0lORVQsIHByb3RvPXNvY2tldC5JUFBST1RPX1RDUAogICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICBkZWwgc29ja2V0CiAgICAgICAgICAgIHJldHVybiBpcFswXVs0XVsw
            XQogICAgICAgIHJldHVybiBob3N0X25hbWUKCiAgICBkZWYgZ2V0X2FjdGl2ZV9hbmRfc3RhbmRi
            eV9tZ3JzKHNlbGYpOgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAg
            ICAgICAgcmV0dXJuICIiLCBzZWxmLmRyeV9ydW4oImNlcGggc3RhdHVzIikKICAgICAgICBtb25p
            dG9yaW5nX2VuZHBvaW50X3BvcnQgPSBzZWxmLl9hcmdfcGFyc2VyLm1vbml0b3JpbmdfZW5kcG9p
            bnRfcG9ydAogICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdCA9IHNlbGYuX2FyZ19w
            YXJzZXIubW9uaXRvcmluZ19lbmRwb2ludAogICAgICAgIHN0YW5kYnlfbWdycyA9IFtdCiAgICAg
            ICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdDoKICAgICAgICAgICAgY21kX2pz
            b24gPSB7InByZWZpeCI6ICJzdGF0dXMiLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgICAgICBy
            ZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21k
            X2pzb24pCiAgICAgICAgICAgICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQs
            CiAgICAgICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAgICAg
            ICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAg
            ICAgICAgICInbWdyIHNlcnZpY2VzJyBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAgICAg
            ICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09V
            VFBVVF9MSVNUfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRw
            b2ludCA9ICgKICAgICAgICAgICAgICAgIGpzb25fb3V0LmdldCgibWdybWFwIiwge30pLmdldCgi
            c2VydmljZXMiLCB7fSkuZ2V0KCJwcm9tZXRoZXVzIiwgIiIpCiAgICAgICAgICAgICkKICAgICAg
            ICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnQ6CiAgICAgICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJjYW4ndCBmaW5k
            IG1vbml0b3JpbmdfZW5kcG9pbnQsIHByb21ldGhldXMgbW9kdWxlIG1pZ2h0IG5vdCBiZSBlbmFi
            bGVkLCAiCiAgICAgICAgICAgICAgICAgICAgImVuYWJsZSB0aGUgbW9kdWxlIGJ5IHJ1bm5pbmcg
            J2NlcGggbWdyIG1vZHVsZSBlbmFibGUgcHJvbWV0aGV1cyciCiAgICAgICAgICAgICAgICApCiAg
            ICAgICAgICAgICMgbm93IGNoZWNrIHRoZSBzdGFuZC1ieSBtZ3ItcwogICAgICAgICAgICBzdGFu
            ZGJ5X2FyciA9IGpzb25fb3V0LmdldCgibWdybWFwIiwge30pLmdldCgic3RhbmRieXMiLCBbXSkK
            ICAgICAgICAgICAgZm9yIGVhY2hfc3RhbmRieSBpbiBzdGFuZGJ5X2FycjoKICAgICAgICAgICAg
            ICAgIGlmICJuYW1lIiBpbiBlYWNoX3N0YW5kYnkua2V5cygpOgogICAgICAgICAgICAgICAgICAg
            IHN0YW5kYnlfbWdycy5hcHBlbmQoZWFjaF9zdGFuZGJ5WyJuYW1lIl0pCiAgICAgICAgICAgIHRy
            eToKICAgICAgICAgICAgICAgIHBhcnNlZF9lbmRwb2ludCA9IHVybHBhcnNlKG1vbml0b3Jpbmdf
            ZW5kcG9pbnQpCiAgICAgICAgICAgIGV4Y2VwdCBWYWx1ZUVycm9yOgogICAgICAgICAgICAgICAg
            cmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBmImlu
            dmFsaWQgZW5kcG9pbnQ6IHttb25pdG9yaW5nX2VuZHBvaW50fSIKICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0ID0gcGFyc2VkX2VuZHBvaW50
            Lmhvc3RuYW1lCiAgICAgICAgICAgIGlmIG5vdCBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQ6CiAg
            ICAgICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQgPSBzdHIocGFyc2VkX2VuZHBv
            aW50LnBvcnQpCgogICAgICAgICMgaWYgbW9uaXRvcmluZyBlbmRwb2ludCBwb3J0IGlzIG5vdCBz
            ZXQsIHB1dCBhIGRlZmF1bHQgbW9uIHBvcnQKICAgICAgICBpZiBub3QgbW9uaXRvcmluZ19lbmRw
            b2ludF9wb3J0OgogICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQgPSBzZWxmLkRF
            RkFVTFRfTU9OSVRPUklOR19FTkRQT0lOVF9QT1JUCgogICAgICAgICMgdXNlciBjb3VsZCBnaXZl
            IGNvbW1hIGFuZCBzcGFjZSBzZXBhcmF0ZWQgaW5wdXRzIChsaWtlIC0tbW9uaXRvcmluZy1lbmRw
            b2ludD0iPGlwMT4sIDxpcDI+IikKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Qg
            PSBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3QucmVwbGFjZSgiLCIsICIgIikKICAgICAgICBt
            b25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXQgPSBtb25pdG9yaW5nX2VuZHBvaW50X2lw
            X2xpc3Quc3BsaXQoKQogICAgICAgICMgaWYgbW9uaXRvcmluZy1lbmRwb2ludCBjb3VsZCBub3Qg
            YmUgZm91bmQsIHJhaXNlIGFuIGVycm9yCiAgICAgICAgaWYgbGVuKG1vbml0b3JpbmdfZW5kcG9p
            bnRfaXBfbGlzdF9zcGxpdCkgPT0gMDoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVy
            ZUV4Y2VwdGlvbigiTm8gJ21vbml0b3JpbmctZW5kcG9pbnQnIGZvdW5kIikKICAgICAgICAjIGZp
            cnN0IGlwIGlzIHRyZWF0ZWQgYXMgdGhlIG1haW4gbW9uaXRvcmluZy1lbmRwb2ludAogICAgICAg
            IG1vbml0b3JpbmdfZW5kcG9pbnRfaXAgPSBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3Bs
            aXRbMF0KICAgICAgICAjIHJlc3Qgb2YgdGhlIGlwLXMgYXJlIGFkZGVkIHRvIHRoZSAnc3RhbmRi
            eV9tZ3JzJyBsaXN0CiAgICAgICAgc3RhbmRieV9tZ3JzLmV4dGVuZChtb25pdG9yaW5nX2VuZHBv
            aW50X2lwX2xpc3Rfc3BsaXRbMTpdKQogICAgICAgIGZhaWxlZF9pcCA9IG1vbml0b3JpbmdfZW5k
            cG9pbnRfaXAKCiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludCA9ICI6Ii5qb2luKAogICAgICAg
            ICAgICBbbW9uaXRvcmluZ19lbmRwb2ludF9pcCwgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0XQog
            ICAgICAgICkKICAgICAgICBpcF90eXBlID0gc2VsZi5faW52YWxpZF9lbmRwb2ludChtb25pdG9y
            aW5nX2VuZHBvaW50KQogICAgICAgIHRyeToKICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2lu
            dF9pcCA9IHNlbGYuX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoCiAgICAgICAgICAgICAgICBtb25p
            dG9yaW5nX2VuZHBvaW50X2lwLCBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQsIGlwX3R5cGUKICAg
            ICAgICAgICAgKQogICAgICAgICAgICAjIGNvbGxlY3QgYWxsIHRoZSAnc3RhbmQtYnknIG1nciBp
            cHMKICAgICAgICAgICAgbWdyX2lwcyA9IFtdCiAgICAgICAgICAgIGZvciBlYWNoX3N0YW5kYnlf
            bWdyIGluIHN0YW5kYnlfbWdyczoKICAgICAgICAgICAgICAgIGZhaWxlZF9pcCA9IGVhY2hfc3Rh
            bmRieV9tZ3IKICAgICAgICAgICAgICAgIG1ncl9pcHMuYXBwZW5kKAogICAgICAgICAgICAgICAg
            ICAgIHNlbGYuX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoCiAgICAgICAgICAgICAgICAgICAgICAg
            IGVhY2hfc3RhbmRieV9tZ3IsIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydCwgaXBfdHlwZQogICAg
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgICkKICAgICAgICBleGNlcHQ6CiAgICAg
            ICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBm
            IkNvbnZlcnNpb24gb2YgaG9zdDoge2ZhaWxlZF9pcH0gdG8gSVAgZmFpbGVkLiAiCiAgICAgICAg
            ICAgICAgICAiUGxlYXNlIGVudGVyIHRoZSBJUCBhZGRyZXNzZXMgb2YgYWxsIHRoZSBjZXBoLW1n
            cnMgd2l0aCB0aGUgJy0tbW9uaXRvcmluZy1lbmRwb2ludCcgZmxhZyIKICAgICAgICAgICAgKQoK
            ICAgICAgICBfLCBfLCBlcnIgPSBzZWxmLmVuZHBvaW50X2RpYWwobW9uaXRvcmluZ19lbmRwb2lu
            dCwgaXBfdHlwZSkKICAgICAgICBpZiBlcnIgPT0gIi0xIjoKICAgICAgICAgICAgcmFpc2UgRXhl
            Y3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihlcnIpCiAgICAgICAgIyBhZGQgdGhlIHZhbGlkYXRlZCBh
            Y3RpdmUgbWdyIElQIGludG8gdGhlIGZpcnN0IGluZGV4CiAgICAgICAgbWdyX2lwcy5pbnNlcnQo
            MCwgbW9uaXRvcmluZ19lbmRwb2ludF9pcCkKICAgICAgICBhbGxfbWdyX2lwc19zdHIgPSAiLCIu
            am9pbihtZ3JfaXBzKQogICAgICAgIHJldHVybiBhbGxfbWdyX2lwc19zdHIsIG1vbml0b3Jpbmdf
            ZW5kcG9pbnRfcG9ydAoKICAgIGRlZiBjaGVja191c2VyX2V4aXN0KHNlbGYsIHVzZXIpOgogICAg
            ICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAiYXV0aCBnZXQiLCAiZW50aXR5IjogZiJ7dXNlcn0i
            LCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBfID0gc2VsZi5f
            Y29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3Ig
            bGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICByZXR1cm4gIiIKICAgICAgICByZXR1cm4g
            c3RyKGpzb25fb3V0WzBdWyJrZXkiXSkKCiAgICBkZWYgZ2V0X2NlcGhmc19wcm92aXNpb25lcl9j
            YXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gImNsaWVudC5jc2ktY2VwaGZz
            LXByb3Zpc2lvbmVyIgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAgICJtb24iOiAiYWxsb3cg
            ciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLAogICAgICAgICAgICAibWdyIjogImFs
            bG93IHJ3IiwKICAgICAgICAgICAgIm9zZCI6ICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRh
            PSoiLAogICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0
            aF9wZXJtaXNzaW9uOgogICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3Bh
            cnNlci5rOHNfY2x1c3Rlcl9uYW1lCiAgICAgICAgICAgIGlmIGs4c19jbHVzdGVyX25hbWUgPT0g
            IiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAg
            ICAgICAgICAgICAgICAgICJrOHNfY2x1c3Rlcl9uYW1lIG5vdCBmb3VuZCwgcGxlYXNlIHNldCB0
            aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgY2VwaGZzX2ZpbGVzeXN0ZW0gPSBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVt
            X25hbWUKICAgICAgICAgICAgaWYgY2VwaGZzX2ZpbGVzeXN0ZW0gPT0gIiI6CiAgICAgICAgICAg
            ICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9uYW1lfSIKICAgICAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0te2s4c19jbHVzdGVyX25h
            bWV9LXtjZXBoZnNfZmlsZXN5c3RlbX0iCiAgICAgICAgICAgICAgICBjYXBzWyJvc2QiXSA9IGYi
            YWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT17Y2VwaGZzX2ZpbGVzeXN0ZW19IgoKICAgICAg
            ICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVmIGdldF9jZXBoZnNfbm9kZV9jYXBzX2FuZF9l
            bnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiCiAg
            ICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJhbGxvdyByLCBhbGxvdyBjb21tYW5k
            ICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJtZ3IiOiAiYWxsb3cgcnciLAogICAgICAg
            ICAgICAib3NkIjogImFsbG93IHJ3IHRhZyBjZXBoZnMgKj0qIiwKICAgICAgICAgICAgIm1kcyI6
            ICJhbGxvdyBydyIsCiAgICAgICAgfQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJp
            Y3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAgIGs4c19jbHVzdGVyX25hbWUgPSBzZWxm
            Ll9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAgICAgICAgICAgaWYgazhzX2NsdXN0ZXJf
            bmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRp
            b24oCiAgICAgICAgICAgICAgICAgICAgIms4c19jbHVzdGVyX25hbWUgbm90IGZvdW5kLCBwbGVh
            c2Ugc2V0IHRoZSAnLS1rOHMtY2x1c3Rlci1uYW1lJyBmbGFnIgogICAgICAgICAgICAgICAgKQog
            ICAgICAgICAgICBjZXBoZnNfZmlsZXN5c3RlbSA9IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2Zp
            bGVzeXN0ZW1fbmFtZQogICAgICAgICAgICBpZiBjZXBoZnNfZmlsZXN5c3RlbSA9PSAiIjoKICAg
            ICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0te2s4c19jbHVzdGVyX25hbWV9IgogICAg
            ICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17azhzX2Ns
            dXN0ZXJfbmFtZX0te2NlcGhmc19maWxlc3lzdGVtfSIKICAgICAgICAgICAgICAgIGNhcHNbIm9z
            ZCJdID0gZiJhbGxvdyBydyB0YWcgY2VwaGZzICo9e2NlcGhmc19maWxlc3lzdGVtfSIKCiAgICAg
            ICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfZW50aXR5KAogICAgICAgIHNlbGYs
            CiAgICAgICAgZW50aXR5LAogICAgICAgIHJiZF9wb29sX25hbWUsCiAgICAgICAgYWxpYXNfcmJk
            X3Bvb2xfbmFtZSwKICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lLAogICAgICAgIHJhZG9zX25hbWVz
            cGFjZSwKICAgICk6CiAgICAgICAgaWYgKAogICAgICAgICAgICByYmRfcG9vbF9uYW1lLmNvdW50
            KCIuIikgIT0gMAogICAgICAgICAgICBvciByYmRfcG9vbF9uYW1lLmNvdW50KCJfIikgIT0gMAog
            ICAgICAgICAgICBvciBhbGlhc19yYmRfcG9vbF9uYW1lICE9ICIiCiAgICAgICAgICAgICMgY2hl
            Y2tpbmcgYWxpYXNfcmJkX3Bvb2xfbmFtZSBpcyBub3QgZW1wdHkgYXMgdGhlcmUgbWF5YmUgYSBz
            cGVjaWFsIGNoYXJhY3RlciB1c2VkIG90aGVyIHRoYW4gLiBvciBfCiAgICAgICAgKToKICAgICAg
            ICAgICAgaWYgYWxpYXNfcmJkX3Bvb2xfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgInBsZWFzZSBz
            ZXQgdGhlICctLWFsaWFzLXJiZC1kYXRhLXBvb2wtbmFtZScgZmxhZyBhcyB0aGUgcmJkIGRhdGEg
            cG9vbCBuYW1lIGNvbnRhaW5zICcuJyBvciAnXyciCiAgICAgICAgICAgICAgICApCiAgICAgICAg
            ICAgIGlmICgKICAgICAgICAgICAgICAgIGFsaWFzX3JiZF9wb29sX25hbWUuY291bnQoIi4iKSAh
            PSAwCiAgICAgICAgICAgICAgICBvciBhbGlhc19yYmRfcG9vbF9uYW1lLmNvdW50KCJfIikgIT0g
            MAogICAgICAgICAgICApOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJy0tYWxpYXMtcmJkLWRhdGEtcG9vbC1uYW1l
            JyBmbGFnIHZhbHVlIHNob3VsZCBub3QgY29udGFpbiAnLicgb3IgJ18nIgogICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9uYW1lfS17
            YWxpYXNfcmJkX3Bvb2xfbmFtZX0iCiAgICAgICAgZWxzZToKICAgICAgICAgICAgZW50aXR5ID0g
            ZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0te3JiZF9wb29sX25hbWV9IgoKICAgICAgICBp
            ZiByYWRvc19uYW1lc3BhY2U6CiAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0te3JhZG9z
            X25hbWVzcGFjZX0iCiAgICAgICAgcmV0dXJuIGVudGl0eQoKICAgIGRlZiBnZXRfcmJkX3Byb3Zp
            c2lvbmVyX2NhcHNfYW5kX2VudGl0eShzZWxmKToKICAgICAgICBlbnRpdHkgPSAiY2xpZW50LmNz
            aS1yYmQtcHJvdmlzaW9uZXIiCiAgICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJw
            cm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLAogICAgICAgICAgICAi
            bWdyIjogImFsbG93IHJ3IiwKICAgICAgICAgICAgIm9zZCI6ICJwcm9maWxlIHJiZCIsCiAgICAg
            ICAgfQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Np
            b246CiAgICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRh
            X3Bvb2xfbmFtZQogICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3Bh
            cnNlci5hbGlhc19yYmRfZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFt
            ZSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQogICAgICAgICAgICByYWRvc19u
            YW1lc3BhY2UgPSBzZWxmLl9hcmdfcGFyc2VyLnJhZG9zX25hbWVzcGFjZQogICAgICAgICAgICBp
            ZiByYmRfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAibWFuZGF0b3J5IGZsYWcgbm90IGZv
            dW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1yYmQtZGF0YS1wb29sLW5hbWUnIGZsYWciCiAgICAgICAg
            ICAgICAgICApCiAgICAgICAgICAgIGlmIGs4c19jbHVzdGVyX25hbWUgPT0gIiI6CiAgICAgICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLWs4cy1jbHVz
            dGVyLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVudGl0eSA9IHNl
            bGYuZ2V0X2VudGl0eSgKICAgICAgICAgICAgICAgIGVudGl0eSwKICAgICAgICAgICAgICAgIHJi
            ZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lLAogICAgICAg
            ICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAgICAgICAgICAgIHJhZG9zX25hbWVzcGFj
            ZSwKICAgICAgICAgICAgKQogICAgICAgICAgICBpZiByYWRvc19uYW1lc3BhY2UgIT0gIiI6CiAg
            ICAgICAgICAgICAgICBjYXBzWyJvc2QiXSA9ICgKICAgICAgICAgICAgICAgICAgICBmInByb2Zp
            bGUgcmJkIHBvb2w9e3JiZF9wb29sX25hbWV9IG5hbWVzcGFjZT17cmFkb3NfbmFtZXNwYWNlfSIK
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGNhcHNb
            Im9zZCJdID0gZiJwcm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9uYW1lfSIKCiAgICAgICAgcmV0
            dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfcmJkX25vZGVfY2Fwc19hbmRfZW50aXR5KHNl
            bGYpOgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3NpLXJiZC1ub2RlIgogICAgICAgIGNhcHMg
            PSB7CiAgICAgICAgICAgICJtb24iOiAicHJvZmlsZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBi
            bG9ja2xpc3QnIiwKICAgICAgICAgICAgIm9zZCI6ICJwcm9maWxlIHJiZCIsCiAgICAgICAgfQog
            ICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb246CiAg
            ICAgICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xf
            bmFtZQogICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5h
            bGlhc19yYmRfZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNl
            bGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQogICAgICAgICAgICByYWRvc19uYW1lc3Bh
            Y2UgPSBzZWxmLl9hcmdfcGFyc2VyLnJhZG9zX25hbWVzcGFjZQogICAgICAgICAgICBpZiByYmRf
            cG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAibWFuZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBw
            bGVhc2Ugc2V0IHRoZSAnLS1yYmQtZGF0YS1wb29sLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAg
            ICApCiAgICAgICAgICAgIGlmIGs4c19jbHVzdGVyX25hbWUgPT0gIiI6CiAgICAgICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJt
            YW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLWs4cy1jbHVzdGVyLW5h
            bWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVudGl0eSA9IHNlbGYuZ2V0
            X2VudGl0eSgKICAgICAgICAgICAgICAgIGVudGl0eSwKICAgICAgICAgICAgICAgIHJiZF9wb29s
            X25hbWUsCiAgICAgICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lLAogICAgICAgICAgICAg
            ICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAgICAgICAgICAgIHJhZG9zX25hbWVzcGFjZSwKICAg
            ICAgICAgICAgKQogICAgICAgICAgICBpZiByYWRvc19uYW1lc3BhY2UgIT0gIiI6CiAgICAgICAg
            ICAgICAgICBjYXBzWyJvc2QiXSA9ICgKICAgICAgICAgICAgICAgICAgICBmInByb2ZpbGUgcmJk
            IHBvb2w9e3JiZF9wb29sX25hbWV9IG5hbWVzcGFjZT17cmFkb3NfbmFtZXNwYWNlfSIKICAgICAg
            ICAgICAgICAgICkKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJd
            ID0gZiJwcm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9uYW1lfSIKCiAgICAgICAgcmV0dXJuIGNh
            cHMsIGVudGl0eQoKICAgIGRlZiBnZXRfZGVmYXVsdFVzZXJfY2Fwc19hbmRfZW50aXR5KHNlbGYp
            OgogICAgICAgIGVudGl0eSA9IHNlbGYucnVuX2FzX3VzZXIKICAgICAgICBjYXBzID0gewogICAg
            ICAgICAgICAibW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1cywgYWxs
            b3cgY29tbWFuZCB2ZXJzaW9uIiwKICAgICAgICAgICAgIm1nciI6ICJhbGxvdyBjb21tYW5kIGNv
            bmZpZyIsCiAgICAgICAgICAgICJvc2QiOiBmInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cg
            cnd4IHBvb2w9e3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cubWV0YSwgYWxs
            b3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD17c2VsZi5fYXJnX3BhcnNlci5yZ3df
            cG9vbF9wcmVmaXh9LnJndy5jb250cm9sLCBhbGxvdyByeCBwb29sPXtzZWxmLl9hcmdfcGFyc2Vy
            LnJnd19wb29sX3ByZWZpeH0ucmd3LmxvZywgYWxsb3cgeCBwb29sPXtzZWxmLl9hcmdfcGFyc2Vy
            LnJnd19wb29sX3ByZWZpeH0ucmd3LmJ1Y2tldHMuaW5kZXgiLAogICAgICAgIH0KCiAgICAgICAg
            cmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfY2Fwc19hbmRfZW50aXR5KHNlbGYsIHVz
            ZXJfbmFtZSk6CiAgICAgICAgaWYgImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiBpbiB1
            c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIg
            IT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVk
            X2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZ2V0X2NlcGhm
            c19wcm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoKQogICAgICAgIGlmICJjbGllbnQuY3NpLWNl
            cGhmcy1ub2RlIiBpbiB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQuY3NpLWNlcGhm
            cy1ub2RlIiAhPSB1c2VyX25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJl
            c3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5n
            ZXRfY2VwaGZzX25vZGVfY2Fwc19hbmRfZW50aXR5KCkKICAgICAgICBpZiAiY2xpZW50LmNzaS1y
            YmQtcHJvdmlzaW9uZXIiIGluIHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVudC5jc2kt
            cmJkLXByb3Zpc2lvbmVyIiAhPSB1c2VyX25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdf
            cGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1
            cm4gc2VsZi5nZXRfcmJkX3Byb3Zpc2lvbmVyX2NhcHNfYW5kX2VudGl0eSgpCiAgICAgICAgaWYg
            ImNsaWVudC5jc2ktcmJkLW5vZGUiIGluIHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVu
            dC5jc2ktcmJkLW5vZGUiICE9IHVzZXJfbmFtZToKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19w
            YXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVlCiAgICAgICAgICAgIHJldHVy
            biBzZWxmLmdldF9yYmRfbm9kZV9jYXBzX2FuZF9lbnRpdHkoKQogICAgICAgIGlmICJjbGllbnQu
            aGVhbHRoY2hlY2tlciIgaW4gdXNlcl9uYW1lOgogICAgICAgICAgICBpZiAiY2xpZW50LmhlYWx0
            aGNoZWNrZXIiICE9IHVzZXJfbmFtZToKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIu
            cmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVlCiAgICAgICAgICAgIHJldHVybiBzZWxm
            LmdldF9kZWZhdWx0VXNlcl9jYXBzX2FuZF9lbnRpdHkoKQoKICAgICAgICByYWlzZSBFeGVjdXRp
            b25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICBmIm5vIHVzZXIgZm91bmQgd2l0aCB1c2Vy
            X25hbWU6IHt1c2VyX25hbWV9LCAiCiAgICAgICAgICAgICJnZXRfY2Fwc19hbmRfZW50aXR5IGNv
            bW1hbmQgZmFpbGVkLlxuIgogICAgICAgICkKCiAgICBkZWYgY3JlYXRlX2NlcGhDU0lLZXlyaW5n
            X3VzZXIoc2VsZiwgdXNlcik6CiAgICAgICAgIiIiCiAgICAgICAgY29tbWFuZDogY2VwaCBhdXRo
            IGdldC1vci1jcmVhdGUgY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIgbW9uICdhbGxvdyBy
            JyBtZ3IgJ2FsbG93IHJ3JyBvc2QgJ2FsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KicKICAg
            ICAgICAiIiIKICAgICAgICBjYXBzLCBlbnRpdHkgPSBzZWxmLmdldF9jYXBzX2FuZF9lbnRpdHko
            dXNlcikKICAgICAgICBjbWRfanNvbiA9IHsKICAgICAgICAgICAgInByZWZpeCI6ICJhdXRoIGdl
            dC1vci1jcmVhdGUiLAogICAgICAgICAgICAiZW50aXR5IjogZW50aXR5LAogICAgICAgICAgICAi
            Y2FwcyI6IFtjYXAgZm9yIGNhcF9saXN0IGluIGxpc3QoY2Fwcy5pdGVtcygpKSBmb3IgY2FwIGlu
            IGNhcF9saXN0XSwKICAgICAgICAgICAgImZvcm1hdCI6ICJqc29uIiwKICAgICAgICB9CgogICAg
            ICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuICgKICAg
            ICAgICAgICAgICAgIHNlbGYuZHJ5X3J1bigKICAgICAgICAgICAgICAgICAgICAiY2VwaCAiCiAg
            ICAgICAgICAgICAgICAgICAgKyBjbWRfanNvblsicHJlZml4Il0KICAgICAgICAgICAgICAgICAg
            ICArICIgIgogICAgICAgICAgICAgICAgICAgICsgY21kX2pzb25bImVudGl0eSJdCiAgICAgICAg
            ICAgICAgICAgICAgKyAiICIKICAgICAgICAgICAgICAgICAgICArICIgIi5qb2luKGNtZF9qc29u
            WyJjYXBzIl0pCiAgICAgICAgICAgICAgICApLAogICAgICAgICAgICAgICAgIiIsCiAgICAgICAg
            ICAgICkKICAgICAgICAjIGNoZWNrIGlmIHVzZXIgYWxyZWFkeSBleGlzdAogICAgICAgIHVzZXJf
            a2V5ID0gc2VsZi5jaGVja191c2VyX2V4aXN0KGVudGl0eSkKICAgICAgICBpZiB1c2VyX2tleSAh
            PSAiIjoKICAgICAgICAgICAgcmV0dXJuIHVzZXJfa2V5LCBmIntlbnRpdHkuc3BsaXQoJy4nLCAx
            KVsxXX0iCiAgICAgICAgICAgICMgZW50aXR5LnNwbGl0KCcuJywxKVsxXSB0byByZW5hbWUgZW50
            aXR5KGNsaWVudC5jc2ktcmJkLW5vZGUpIGFzIGNzaS1yYmQtbm9kZQoKICAgICAgICByZXRfdmFs
            LCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24p
            CiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwKICAgICAgICBp
            ZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiInYXV0aCBnZXQtb3It
            Y3JlYXRlIHt1c2VyfScgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAgICAgICAgICAgICBmIkVycm9y
            OiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIK
            ICAgICAgICAgICAgKQogICAgICAgIHJldHVybiBzdHIoanNvbl9vdXRbMF1bImtleSJdKSwgZiJ7
            ZW50aXR5LnNwbGl0KCcuJywgMSlbMV19IgogICAgICAgICMgZW50aXR5LnNwbGl0KCcuJywxKVsx
            XSB0byByZW5hbWUgZW50aXR5KGNsaWVudC5jc2ktcmJkLW5vZGUpIGFzIGNzaS1yYmQtbm9kZQoK
            ICAgIGRlZiBnZXRfY2VwaGZzX2RhdGFfcG9vbF9kZXRhaWxzKHNlbGYpOgogICAgICAgIGNtZF9q
            c29uID0geyJwcmVmaXgiOiAiZnMgbHMiLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIGlmIHNl
            bGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigi
            Y2VwaCAiICsgY21kX2pzb25bInByZWZpeCJdKQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBl
            cnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICAjIGlm
            IHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LCByZXBvcnQgYW4gZXJyb3IKICAgICAg
            ICBpZiByZXRfdmFsICE9IDA6CiAgICAgICAgICAgICMgaWYgZnMgYW5kIGRhdGFfcG9vbCBhcmd1
            bWVudHMgYXJlIG5vdCBzZXQsIHNpbGVudGx5IHJldHVybgogICAgICAgICAgICBpZiAoCiAgICAg
            ICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUgPT0gIiIK
            ICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFt
            ZSA9PSAiIgogICAgICAgICAgICApOgogICAgICAgICAgICAgICAgcmV0dXJuCiAgICAgICAgICAg
            ICMgaWYgdXNlciBoYXMgcHJvdmlkZWQgYW55IG9mIHRoZQogICAgICAgICAgICAjICctLWNlcGhm
            cy1maWxlc3lzdGVtLW5hbWUnIG9yICctLWNlcGhmcy1kYXRhLXBvb2wtbmFtZScgYXJndW1lbnRz
            LAogICAgICAgICAgICAjIHJhaXNlIGFuIGV4Y2VwdGlvbiBhcyB3ZSBhcmUgdW5hYmxlIHRvIHZl
            cmlmeSB0aGUgYXJncwogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KAogICAgICAgICAgICAgICAgZiInZnMgbHMnIGNlcGggY2FsbCBmYWlsZWQgd2l0aCBlcnJvcjog
            e2Vycl9tc2d9IgogICAgICAgICAgICApCgogICAgICAgIG1hdGNoaW5nX2pzb25fb3V0ID0ge30K
            ICAgICAgICAjIGlmICctLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUnIGFyZ3VtZW50IGlzIHByb3Zp
            ZGVkLAogICAgICAgICMgY2hlY2sgd2hldGhlciB0aGUgcHJvdmlkZWQgZmlsZXN5c3RlbS1uYW1l
            IGV4aXN0cyBvciBub3QKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lz
            dGVtX25hbWU6CiAgICAgICAgICAgICMgZ2V0IHRoZSBtYXRjaGluZyBsaXN0CiAgICAgICAgICAg
            IG1hdGNoaW5nX2pzb25fb3V0X2xpc3QgPSBbCiAgICAgICAgICAgICAgICBtYXRjaGVkCiAgICAg
            ICAgICAgICAgICBmb3IgbWF0Y2hlZCBpbiBqc29uX291dAogICAgICAgICAgICAgICAgaWYgc3Ry
            KG1hdGNoZWRbIm5hbWUiXSkgPT0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9u
            YW1lCiAgICAgICAgICAgIF0KICAgICAgICAgICAgIyB1bmFibGUgdG8gZmluZCBhIG1hdGNoaW5n
            IGZzLW5hbWUsIHJhaXNlIGFuIGVycm9yCiAgICAgICAgICAgIGlmIGxlbihtYXRjaGluZ19qc29u
            X291dF9saXN0KSA9PSAwOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBmIkZpbGVzeXN0ZW0gcHJvdmlkZWQsICd7c2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lfScsICIKICAgICAgICAgICAgICAg
            ICAgICBmImlzIG5vdCBmb3VuZCBpbiB0aGUgZnMtbGlzdDoge1tzdHIoeFsnbmFtZSddKSBmb3Ig
            eCBpbiBqc29uX291dF19IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBtYXRjaGluZ19q
            c29uX291dCA9IG1hdGNoaW5nX2pzb25fb3V0X2xpc3RbMF0KICAgICAgICAjIGlmIGNlcGhmcyBm
            aWxlc3lzdGVtIG5hbWUgaXMgbm90IHByb3ZpZGVkLAogICAgICAgICMgdHJ5IHRvIGdldCBhIGRl
            ZmF1bHQgZnMgbmFtZSBieSBkb2luZyB0aGUgZm9sbG93aW5nCiAgICAgICAgZWxzZToKICAgICAg
            ICAgICAgIyBhLiBjaGVjayBpZiB0aGVyZSBpcyBvbmx5IG9uZSBmaWxlc3lzdGVtIGlzIHByZXNl
            bnQKICAgICAgICAgICAgaWYgbGVuKGpzb25fb3V0KSA9PSAxOgogICAgICAgICAgICAgICAgbWF0
            Y2hpbmdfanNvbl9vdXQgPSBqc29uX291dFswXQogICAgICAgICAgICAjIGIuIG9yIGVsc2UsIGNo
            ZWNrIGlmIGRhdGFfcG9vbCBuYW1lIGlzIHByb3ZpZGVkCiAgICAgICAgICAgIGVsaWYgc2VsZi5f
            YXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWU6CiAgICAgICAgICAgICAgICAjIGFuZCBp
            ZiBwcmVzZW50LCBjaGVjayB3aGV0aGVyIHRoZXJlIGV4aXN0cyBhIGZzIHdoaWNoIGhhcyB0aGUg
            ZGF0YV9wb29sCiAgICAgICAgICAgICAgICBmb3IgZWFjaEogaW4ganNvbl9vdXQ6CiAgICAgICAg
            ICAgICAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUgaW4g
            ZWFjaEpbImRhdGFfcG9vbHMiXToKICAgICAgICAgICAgICAgICAgICAgICAgbWF0Y2hpbmdfanNv
            bl9vdXQgPSBlYWNoSgogICAgICAgICAgICAgICAgICAgICAgICBicmVhawogICAgICAgICAgICAg
            ICAgIyBpZiB0aGVyZSBpcyBubyBtYXRjaGluZyBmcyBleGlzdHMsIHRoYXQgbWVhbnMgcHJvdmlk
            ZWQgZGF0YV9wb29sIG5hbWUgaXMgaW52YWxpZAogICAgICAgICAgICAgICAgaWYgbm90IG1hdGNo
            aW5nX2pzb25fb3V0OgogICAgICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVF
            eGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgICAgIGYiUHJvdmlkZWQgZGF0YV9wb29sIG5h
            bWUsIHtzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZX0sIgogICAgICAgICAg
            ICAgICAgICAgICAgICAiIGRvZXMgbm90IGV4aXN0cyIKICAgICAgICAgICAgICAgICAgICApCiAg
            ICAgICAgICAgICMgYy4gaWYgbm90aGluZyBpcyBzZXQgYW5kIGNvdWxkbid0IGZpbmQgYSBkZWZh
            dWx0LAogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgIyBqdXN0IHJldHVybiBzaWxl
            bnRseQogICAgICAgICAgICAgICAgcmV0dXJuCgogICAgICAgIGlmIG1hdGNoaW5nX2pzb25fb3V0
            OgogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUgPSBz
            dHIobWF0Y2hpbmdfanNvbl9vdXRbIm5hbWUiXSkKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5jZXBoZnNfbWV0YWRhdGFfcG9vbF9uYW1lID0gc3RyKAogICAgICAgICAgICAgICAgbWF0Y2hp
            bmdfanNvbl9vdXRbIm1ldGFkYXRhX3Bvb2wiXQogICAgICAgICAgICApCgogICAgICAgIGlmIGlz
            aW5zdGFuY2UobWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXSwgbGlzdCk6CiAgICAgICAg
            ICAgICMgaWYgdGhlIHVzZXIgaGFzIGFscmVhZHkgcHJvdmlkZWQgZGF0YS1wb29sLW5hbWUsCiAg
            ICAgICAgICAgICMgdGhyb3VnaCAtLWNlcGhmcy1kYXRhLXBvb2wtbmFtZQogICAgICAgICAgICBp
            ZiBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZToKICAgICAgICAgICAgICAg
            ICMgaWYgdGhlIHByb3ZpZGVkIG5hbWUgaXMgbm90IG1hdGNoaW5nIHdpdGggdGhlIG9uZSBpbiB0
            aGUgbGlzdAogICAgICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2Fy
            Z19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgICAgICAgICAgbm90IGlu
            IG1hdGNoaW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0KICAgICAgICAgICAgICAgICk6CiAgICAg
            ICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAg
            ICAgICAgICAgICAgICAgZiJQcm92aWRlZCBkYXRhLXBvb2wtbmFtZTogJ3tzZWxmLl9hcmdfcGFy
            c2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZX0nLCAiCiAgICAgICAgICAgICAgICAgICAgICAgICJk
            b2Vzbid0IG1hdGNoIGZyb20gdGhlIGRhdGEtcG9vbHMgbGlzdDogIgogICAgICAgICAgICAgICAg
            ICAgICAgICBmIntbc3RyKHgpIGZvciB4IGluIG1hdGNoaW5nX2pzb25fb3V0WydkYXRhX3Bvb2xz
            J11dfSIKICAgICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICMgaWYgZGF0YV9wb29sIG5h
            bWUgaXMgbm90IHByb3ZpZGVkLAogICAgICAgICAgICAjIHRoZW4gdHJ5IHRvIGZpbmQgYSBkZWZh
            dWx0IGRhdGEgcG9vbCBuYW1lCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICAjIGlm
            IG5vIGRhdGFfcG9vbHMgZXhpc3QsIHNpbGVudGx5IHJldHVybgogICAgICAgICAgICAgICAgaWYg
            bGVuKG1hdGNoaW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0pID09IDA6CiAgICAgICAgICAgICAg
            ICAgICAgcmV0dXJuCiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRh
            X3Bvb2xfbmFtZSA9IHN0cigKICAgICAgICAgICAgICAgICAgICBtYXRjaGluZ19qc29uX291dFsi
            ZGF0YV9wb29scyJdWzBdCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICMgaWYgdGhlcmUg
            YXJlIG1vcmUgdGhhbiBvbmUgJ2RhdGFfcG9vbHMnIGV4aXN0LAogICAgICAgICAgICAjIHRoZW4g
            d2FybiB0aGUgdXNlciB0aGF0IHdlIGFyZSB1c2luZyB0aGUgc2VsZWN0ZWQgbmFtZQogICAgICAg
            ICAgICBpZiBsZW4obWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXSkgPiAxOgogICAgICAg
            ICAgICAgICAgcHJpbnQoCiAgICAgICAgICAgICAgICAgICAgIldBUk5JTkc6IE11bHRpcGxlIGRh
            dGEgcG9vbHMgZGV0ZWN0ZWQ6ICIKICAgICAgICAgICAgICAgICAgICBmIntbc3RyKHgpIGZvciB4
            IGluIG1hdGNoaW5nX2pzb25fb3V0WydkYXRhX3Bvb2xzJ11dfVxuIgogICAgICAgICAgICAgICAg
            ICAgIGYiVXNpbmcgdGhlIGRhdGEtcG9vbDogJ3tzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRh
            X3Bvb2xfbmFtZX0nXG4iCiAgICAgICAgICAgICAgICApCgogICAgZGVmIGNyZWF0ZV9jaGVja2Vy
            S2V5KHNlbGYsIHVzZXIpOgogICAgICAgIGNhcHMsIGVudGl0eSA9IHNlbGYuZ2V0X2NhcHNfYW5k
            X2VudGl0eSh1c2VyKQogICAgICAgIGNtZF9qc29uID0gewogICAgICAgICAgICAicHJlZml4Ijog
            ImF1dGggZ2V0LW9yLWNyZWF0ZSIsCiAgICAgICAgICAgICJlbnRpdHkiOiBlbnRpdHksCiAgICAg
            ICAgICAgICJjYXBzIjogW2NhcCBmb3IgY2FwX2xpc3QgaW4gbGlzdChjYXBzLml0ZW1zKCkpIGZv
            ciBjYXAgaW4gY2FwX2xpc3RdLAogICAgICAgICAgICAiZm9ybWF0IjogImpzb24iLAogICAgICAg
            IH0KCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1
            cm4gc2VsZi5kcnlfcnVuKAogICAgICAgICAgICAgICAgImNlcGggIgogICAgICAgICAgICAgICAg
            KyBjbWRfanNvblsicHJlZml4Il0KICAgICAgICAgICAgICAgICsgIiAiCiAgICAgICAgICAgICAg
            ICArIGNtZF9qc29uWyJlbnRpdHkiXQogICAgICAgICAgICAgICAgKyAiICIKICAgICAgICAgICAg
            ICAgICsgIiAiLmpvaW4oY21kX2pzb25bImNhcHMiXSkKICAgICAgICAgICAgKQogICAgICAgICMg
            Y2hlY2sgaWYgdXNlciBhbHJlYWR5IGV4aXN0CiAgICAgICAgdXNlcl9rZXkgPSBzZWxmLmNoZWNr
            X3VzZXJfZXhpc3QoZW50aXR5KQogICAgICAgIGlmIHVzZXJfa2V5ICE9ICIiOgogICAgICAgICAg
            ICByZXR1cm4gdXNlcl9rZXkKCiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBz
            ZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICMgaWYgdGhlcmUgaXMg
            YW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihq
            c29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgICAgIGYiJ2F1dGggZ2V0LW9yLWNyZWF0ZSB7c2VsZi5ydW5fYXNfdXNl
            cn0nIGNvbW1hbmQgZmFpbGVkXG4iCiAgICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBp
            ZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAgICAg
            KQogICAgICAgIHJldHVybiBzdHIoanNvbl9vdXRbMF1bImtleSJdKQoKICAgIGRlZiBnZXRfY2Vw
            aF9kYXNoYm9hcmRfbGluayhzZWxmKToKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4IjogIm1n
            ciBzZXJ2aWNlcyIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNl
            ci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKCJjZXBoICIgKyBjbWRf
            anNvblsicHJlZml4Il0pCiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIF8gPSBzZWxmLl9jb21t
            b25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNj
            ZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkg
            PT0gMDoKICAgICAgICAgICAgcmV0dXJuIE5vbmUKICAgICAgICBpZiAiZGFzaGJvYXJkIiBub3Qg
            aW4ganNvbl9vdXQ6CiAgICAgICAgICAgIHJldHVybiBOb25lCiAgICAgICAgcmV0dXJuIGpzb25f
            b3V0WyJkYXNoYm9hcmQiXQoKICAgIGRlZiBjcmVhdGVfcmd3X2FkbWluX29wc191c2VyKHNlbGYp
            OgogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAgICAgICAg
            ICAidXNlciIsCiAgICAgICAgICAgICJjcmVhdGUiLAogICAgICAgICAgICAiLS11aWQiLAogICAg
            ICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9PUFNfVVNFUl9OQU1FLAogICAgICAgICAg
            ICAiLS1kaXNwbGF5LW5hbWUiLAogICAgICAgICAgICAiUm9vayBSR1cgQWRtaW4gT3BzIHVzZXIi
            LAogICAgICAgICAgICAiLS1jYXBzIiwKICAgICAgICAgICAgImJ1Y2tldHM9Kjt1c2Vycz0qO3Vz
            YWdlPXJlYWQ7bWV0YWRhdGE9cmVhZDt6b25lPXJlYWQiLAogICAgICAgICAgICAiLS1yZ3ctcmVh
            bG0iLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lLAogICAgICAg
            ICAgICAiLS1yZ3ctem9uZWdyb3VwIiwKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3df
            em9uZWdyb3VwX25hbWUsCiAgICAgICAgICAgICItLXJndy16b25lIiwKICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1lLAogICAgICAgIF0KICAgICAgICBpZiBzZWxmLl9h
            cmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGgg
            IiArICIgIi5qb2luKGNtZCkpCiAgICAgICAgdHJ5OgogICAgICAgICAgICBvdXRwdXQgPSBzdWJw
            cm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAg
            ZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yIGFzIGV4ZWNFcnI6CiAgICAgICAg
            ICAgICMgaWYgdGhlIHVzZXIgYWxyZWFkeSBleGlzdHMsIHdlIGp1c3QgcXVlcnkgaXQKICAgICAg
            ICAgICAgaWYgZXhlY0Vyci5yZXR1cm5jb2RlID09IGVycm5vLkVFWElTVDoKICAgICAgICAgICAg
            ICAgIGNtZCA9IFsKICAgICAgICAgICAgICAgICAgICAicmFkb3Nndy1hZG1pbiIsCiAgICAgICAg
            ICAgICAgICAgICAgInVzZXIiLAogICAgICAgICAgICAgICAgICAgICJpbmZvIiwKICAgICAgICAg
            ICAgICAgICAgICAiLS11aWQiLAogICAgICAgICAgICAgICAgICAgIHNlbGYuRVhURVJOQUxfUkdX
            X0FETUlOX09QU19VU0VSX05BTUUsCiAgICAgICAgICAgICAgICAgICAgIi0tcmd3LXJlYWxtIiwK
            ICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lLAogICAg
            ICAgICAgICAgICAgICAgICItLXJndy16b25lZ3JvdXAiLAogICAgICAgICAgICAgICAgICAgIHNl
            bGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lLAogICAgICAgICAgICAgICAgICAgICIt
            LXJndy16b25lIiwKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25l
            X25hbWUsCiAgICAgICAgICAgICAgICBdCiAgICAgICAgICAgICAgICB0cnk6CiAgICAgICAgICAg
            ICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3Vi
            cHJvY2Vzcy5QSVBFKQogICAgICAgICAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJv
            Y2Vzc0Vycm9yIGFzIGV4ZWNFcnI6CiAgICAgICAgICAgICAgICAgICAgZXJyX21zZyA9ICgKICAg
            ICAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5kIHtjbWR9LiBP
            dXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICAgICAgZiJDb2Rl
            OiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtleGVjRXJyLnN0ZGVycn0iCiAgICAgICAg
            ICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21z
            ZykKICAgICAgICAgICAgICAgICAgICByZXR1cm4gTm9uZSwgTm9uZSwgRmFsc2UsICItMSIKICAg
            ICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVycl9tc2cgPSAoCiAgICAgICAgICAgICAg
            ICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5kIHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJy
            Lm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICBmIkNvZGU6IHtleGVjRXJyLnJldHVybmNv
            ZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykKICAgICAgICAgICAgICAgIHJldHVybiBO
            b25lLCBOb25lLCBGYWxzZSwgIi0xIgoKICAgICAgICAjIGlmIGl0IGlzIHB5dGhvbjIsIGRvbid0
            IGNoZWNrIGZvciBjZXBoIHZlcnNpb24gZm9yIGFkZGluZyBgaW5mbz1yZWFkYCBjYXAocmd3X3Zh
            bGlkYXRpb24pCiAgICAgICAgaWYgc3lzLnZlcnNpb25faW5mby5tYWpvciA8IDM6CiAgICAgICAg
            ICAgIGpzb25vdXRwdXQgPSBqc29uLmxvYWRzKG91dHB1dCkKICAgICAgICAgICAgcmV0dXJuICgK
            ICAgICAgICAgICAgICAgIGpzb25vdXRwdXRbImtleXMiXVswXVsiYWNjZXNzX2tleSJdLAogICAg
            ICAgICAgICAgICAganNvbm91dHB1dFsia2V5cyJdWzBdWyJzZWNyZXRfa2V5Il0sCiAgICAgICAg
            ICAgICAgICBGYWxzZSwKICAgICAgICAgICAgICAgICIiLAogICAgICAgICAgICApCgogICAgICAg
            ICMgc2VwYXJhdGVseSBhZGQgaW5mbz1yZWFkIGNhcHMgZm9yIHJndy1lbmRwb2ludCBpcCB2YWxp
            ZGF0aW9uCiAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkID0gVHJ1ZQogICAgICAgIGNtZCA9IFsK
            ICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAgICAgICAgICAiY2FwcyIsCiAgICAgICAg
            ICAgICJhZGQiLAogICAgICAgICAgICAiLS11aWQiLAogICAgICAgICAgICBzZWxmLkVYVEVSTkFM
            X1JHV19BRE1JTl9PUFNfVVNFUl9OQU1FLAogICAgICAgICAgICAiLS1jYXBzIiwKICAgICAgICAg
            ICAgImluZm89cmVhZCIsCiAgICAgICAgICAgICItLXJndy1yZWFsbSIsCiAgICAgICAgICAgIHNl
            bGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUsCiAgICAgICAgICAgICItLXJndy16b25lZ3Jv
            dXAiLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSwKICAg
            ICAgICAgICAgIi0tcmd3LXpvbmUiLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196
            b25lX25hbWUsCiAgICAgICAgXQogICAgICAgIHRyeToKICAgICAgICAgICAgb3V0cHV0ID0gc3Vi
            cHJvY2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAg
            IGV4Y2VwdCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJvciBhcyBleGVjRXJyOgogICAgICAg
            ICAgICAjIGlmIHRoZSBjZXBoIHZlcnNpb24gbm90IHN1cHBvcnRlZCBmb3IgYWRkaW5nIGBpbmZv
            PXJlYWRgIGNhcChyZ3dfdmFsaWRhdGlvbikKICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAg
            ICAgImNvdWxkIG5vdCBhZGQgY2FwczogdW5hYmxlIHRvIGFkZCBjYXBzOiBpbmZvPXJlYWRcbiIK
            ICAgICAgICAgICAgICAgIGluIGV4ZWNFcnIuc3RkZXJyLmRlY29kZSgidXRmLTgiKQogICAgICAg
            ICAgICAgICAgYW5kIGV4ZWNFcnIucmV0dXJuY29kZSA9PSAyNDQKICAgICAgICAgICAgKToKICAg
            ICAgICAgICAgICAgIGluZm9fY2FwX3N1cHBvcnRlZCA9IEZhbHNlCiAgICAgICAgICAgIGVsc2U6
            CiAgICAgICAgICAgICAgICBlcnJfbXNnID0gKAogICAgICAgICAgICAgICAgICAgIGYiZmFpbGVk
            IHRvIGV4ZWN1dGUgY29tbWFuZCB7Y21kfS4gT3V0cHV0OiB7ZXhlY0Vyci5vdXRwdXR9LiAiCiAg
            ICAgICAgICAgICAgICAgICAgZiJDb2RlOiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtl
            eGVjRXJyLnN0ZGVycn0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICBzeXMuc3Rk
            ZXJyLndyaXRlKGVycl9tc2cpCiAgICAgICAgICAgICAgICByZXR1cm4gTm9uZSwgTm9uZSwgRmFs
            c2UsICItMSIKCiAgICAgICAganNvbm91dHB1dCA9IGpzb24ubG9hZHMob3V0cHV0KQogICAgICAg
            IHJldHVybiAoCiAgICAgICAgICAgIGpzb25vdXRwdXRbImtleXMiXVswXVsiYWNjZXNzX2tleSJd
            LAogICAgICAgICAgICBqc29ub3V0cHV0WyJrZXlzIl1bMF1bInNlY3JldF9rZXkiXSwKICAgICAg
            ICAgICAgaW5mb19jYXBfc3VwcG9ydGVkLAogICAgICAgICAgICAiIiwKICAgICAgICApCgogICAg
            ZGVmIHZhbGlkYXRlX3JiZF9wb29sKHNlbGYsIHBvb2xfbmFtZSk6CiAgICAgICAgaWYgbm90IHNl
            bGYuY2x1c3Rlci5wb29sX2V4aXN0cyhwb29sX25hbWUpOgogICAgICAgICAgICByYWlzZSBFeGVj
            dXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcG9v
            bCwgJ3twb29sX25hbWV9JywgZG9lcyBub3QgZXhpc3QiCiAgICAgICAgICAgICkKCiAgICBkZWYg
            aW5pdF9yYmRfcG9vbChzZWxmLCByYmRfcG9vbF9uYW1lKToKICAgICAgICBpZiBpc2luc3RhbmNl
            KHNlbGYuY2x1c3RlciwgRHVtbXlSYWRvcyk6CiAgICAgICAgICAgIHJldHVybgogICAgICAgIGlv
            Y3R4ID0gc2VsZi5jbHVzdGVyLm9wZW5faW9jdHgocmJkX3Bvb2xfbmFtZSkKICAgICAgICByYmRf
            aW5zdCA9IHJiZC5SQkQoKQogICAgICAgIHJiZF9pbnN0LnBvb2xfaW5pdChpb2N0eCwgVHJ1ZSkK
            CiAgICBkZWYgdmFsaWRhdGVfcmFkb3NfbmFtZXNwYWNlKHNlbGYpOgogICAgICAgIHJiZF9wb29s
            X25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgIHJhZG9z
            X25hbWVzcGFjZSA9IHNlbGYuX2FyZ19wYXJzZXIucmFkb3NfbmFtZXNwYWNlCiAgICAgICAgaWYg
            cmFkb3NfbmFtZXNwYWNlID09ICIiOgogICAgICAgICAgICByZXR1cm4KICAgICAgICBpZiByYWRv
            c19uYW1lc3BhY2UuaXNsb3dlcigpID09IEZhbHNlOgogICAgICAgICAgICByYWlzZSBFeGVjdXRp
            b25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmFkb3Mg
            TmFtZXNwYWNlLCAne3JhZG9zX25hbWVzcGFjZX0nLCAiCiAgICAgICAgICAgICAgICBmImNvbnRh
            aW5zIHVwcGVyIGNhc2UiCiAgICAgICAgICAgICkKICAgICAgICByYmRfaW5zdCA9IHJiZC5SQkQo
            KQogICAgICAgIGlvY3R4ID0gc2VsZi5jbHVzdGVyLm9wZW5faW9jdHgocmJkX3Bvb2xfbmFtZSkK
            ICAgICAgICBpZiByYmRfaW5zdC5uYW1lc3BhY2VfZXhpc3RzKGlvY3R4LCByYWRvc19uYW1lc3Bh
            Y2UpIGlzIEZhbHNlOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmFkb3MgTmFtZXNwYWNlLCAne3JhZG9z
            X25hbWVzcGFjZX0nLCAiCiAgICAgICAgICAgICAgICBmImlzIG5vdCBmb3VuZCBpbiB0aGUgcG9v
            bCAne3JiZF9wb29sX25hbWV9JyIKICAgICAgICAgICAgKQoKICAgIGRlZiBnZXRfb3JfY3JlYXRl
            X3N1YnZvbHVtZV9ncm91cChzZWxmLCBzdWJ2b2x1bWVfZ3JvdXAsIGNlcGhmc19maWxlc3lzdGVt
            X25hbWUpOgogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgImNlcGgiLAogICAgICAgICAgICAi
            ZnMiLAogICAgICAgICAgICAic3Vidm9sdW1lZ3JvdXAiLAogICAgICAgICAgICAiZ2V0cGF0aCIs
            CiAgICAgICAgICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgIHN1YnZvbHVt
            ZV9ncm91cCwKICAgICAgICBdCiAgICAgICAgdHJ5OgogICAgICAgICAgICBfID0gc3VicHJvY2Vz
            cy5jaGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2Vw
            dCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJvcjoKICAgICAgICAgICAgY21kID0gWwogICAg
            ICAgICAgICAgICAgImNlcGgiLAogICAgICAgICAgICAgICAgImZzIiwKICAgICAgICAgICAgICAg
            ICJzdWJ2b2x1bWVncm91cCIsCiAgICAgICAgICAgICAgICAiY3JlYXRlIiwKICAgICAgICAgICAg
            ICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgICAgICBzdWJ2b2x1bWVfZ3Jv
            dXAsCiAgICAgICAgICAgIF0KICAgICAgICAgICAgdHJ5OgogICAgICAgICAgICAgICAgXyA9IHN1
            YnByb2Nlc3MuY2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAgICAg
            ICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBm
            InN1YnZvbHVtZSBncm91cCB7c3Vidm9sdW1lX2dyb3VwfSBpcyBub3QgYWJsZSB0byBnZXQgY3Jl
            YXRlZCIKICAgICAgICAgICAgICAgICkKCiAgICBkZWYgcGluX3N1YnZvbHVtZSgKICAgICAgICBz
            ZWxmLCBzdWJ2b2x1bWVfZ3JvdXAsIGNlcGhmc19maWxlc3lzdGVtX25hbWUsIHBpbl90eXBlLCBw
            aW5fc2V0dGluZwogICAgKToKICAgICAgICBjbWQgPSBbCiAgICAgICAgICAgICJjZXBoIiwKICAg
            ICAgICAgICAgImZzIiwKICAgICAgICAgICAgInN1YnZvbHVtZWdyb3VwIiwKICAgICAgICAgICAg
            InBpbiIsCiAgICAgICAgICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgIHN1
            YnZvbHVtZV9ncm91cCwKICAgICAgICAgICAgcGluX3R5cGUsCiAgICAgICAgICAgIHBpbl9zZXR0
            aW5nLAogICAgICAgIF0KICAgICAgICB0cnk6CiAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNo
            ZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1
            YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJzdWJ2b2x1bWUgZ3JvdXAge3N1YnZv
            bHVtZV9ncm91cH0gaXMgbm90IGFibGUgdG8gZ2V0IHBpbm5lZCIKICAgICAgICAgICAgKQoKICAg
            IGRlZiBnZXRfcmd3X2ZzaWQoc2VsZiwgYmFzZV91cmwsIHZlcmlmeSk6CiAgICAgICAgYWNjZXNz
            X2tleSA9IHNlbGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX0FDQ0VTU19LRVkiXQogICAg
            ICAgIHNlY3JldF9rZXkgPSBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9TRUNSRVRf
            S0VZIl0KICAgICAgICByZ3dfZW5kcG9pbnQgPSBzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2lu
            dAogICAgICAgIGJhc2VfdXJsID0gYmFzZV91cmwgKyAiOi8vIiArIHJnd19lbmRwb2ludCArICIv
            YWRtaW4vaW5mbz8iCiAgICAgICAgcGFyYW1zID0geyJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAg
            cmVxdWVzdF91cmwgPSBiYXNlX3VybCArIHVybGVuY29kZShwYXJhbXMpCgogICAgICAgIHRyeToK
            ICAgICAgICAgICAgciA9IHJlcXVlc3RzLmdldCgKICAgICAgICAgICAgICAgIHJlcXVlc3RfdXJs
            LAogICAgICAgICAgICAgICAgYXV0aD1TM0F1dGgoYWNjZXNzX2tleSwgc2VjcmV0X2tleSwgcmd3
            X2VuZHBvaW50KSwKICAgICAgICAgICAgICAgIHZlcmlmeT12ZXJpZnksCiAgICAgICAgICAgICkK
            ICAgICAgICBleGNlcHQgcmVxdWVzdHMuZXhjZXB0aW9ucy5UaW1lb3V0OgogICAgICAgICAgICBz
            eXMuc3RkZXJyLndyaXRlKAogICAgICAgICAgICAgICAgZiJpbnZhbGlkIGVuZHBvaW50Oiwgbm90
            IGFibGUgdG8gY2FsbCBhZG1pbi1vcHMgYXBpe3Jnd19lbmRwb2ludH0iCiAgICAgICAgICAgICkK
            ICAgICAgICAgICAgcmV0dXJuICIiLCAiLTEiCiAgICAgICAgcjEgPSByLmpzb24oKQogICAgICAg
            IGlmIHIxIGlzIE5vbmUgb3IgcjEuZ2V0KCJpbmZvIikgaXMgTm9uZToKICAgICAgICAgICAgc3lz
            LnN0ZGVyci53cml0ZSgKICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHJndyBFbmRwb2lu
            dCwgJ3tzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludH0nLCBpcyBpbnZhbGlkLiIKICAgICAg
            ICAgICAgKQogICAgICAgICAgICByZXR1cm4gKAogICAgICAgICAgICAgICAgIiIsCiAgICAgICAg
            ICAgICAgICAiLTEiLAogICAgICAgICAgICApCgogICAgICAgIHJldHVybiByMVsiaW5mbyJdWyJz
            dG9yYWdlX2JhY2tlbmRzIl1bMF1bImNsdXN0ZXJfaWQiXSwgIiIKCiAgICBkZWYgdmFsaWRhdGVf
            cmd3X2VuZHBvaW50KHNlbGYsIGluZm9fY2FwX3N1cHBvcnRlZCk6CiAgICAgICAgIyBpZiB0aGUg
            J2NsdXN0ZXInIGluc3RhbmNlIGlzIGEgZHVtbXkgb25lLAogICAgICAgICMgZG9uJ3QgdHJ5IHRv
            IHJlYWNoIG91dCB0byB0aGUgZW5kcG9pbnQKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1
            c3RlciwgRHVtbXlSYWRvcyk6CiAgICAgICAgICAgIHJldHVybgoKICAgICAgICByZ3dfZW5kcG9p
            bnQgPSBzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludAoKICAgICAgICAjIHZhbGlkYXRlIHJn
            dyBlbmRwb2ludCBvbmx5IGlmIGlwIGFkZHJlc3MgaXMgcGFzc2VkCiAgICAgICAgaXBfdHlwZSA9
            IHNlbGYuX2ludmFsaWRfZW5kcG9pbnQocmd3X2VuZHBvaW50KQoKICAgICAgICAjIGNoZWNrIGlm
            IHRoZSByZ3cgZW5kcG9pbnQgaXMgcmVhY2hhYmxlCiAgICAgICAgY2VydCA9IE5vbmUKICAgICAg
            ICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNlci5yZ3dfc2tpcF90bHMgYW5kIHNlbGYudmFsaWRhdGVf
            cmd3X2VuZHBvaW50X3Rsc19jZXJ0KCk6CiAgICAgICAgICAgIGNlcnQgPSBzZWxmLl9hcmdfcGFy
            c2VyLnJnd190bHNfY2VydF9wYXRoCiAgICAgICAgYmFzZV91cmwsIHZlcmlmeSwgZXJyID0gc2Vs
            Zi5lbmRwb2ludF9kaWFsKHJnd19lbmRwb2ludCwgaXBfdHlwZSwgY2VydD1jZXJ0KQogICAgICAg
            IGlmIGVyciAhPSAiIjoKICAgICAgICAgICAgcmV0dXJuICItMSIKCiAgICAgICAgIyBjaGVjayBp
            ZiB0aGUgcmd3IGVuZHBvaW50IGJlbG9uZ3MgdG8gdGhlIHNhbWUgY2x1c3RlcgogICAgICAgICMg
            b25seSBjaGVjayBpZiBgaW5mb2AgY2FwIGlzIHN1cHBvcnRlZAogICAgICAgIGlmIGluZm9fY2Fw
            X3N1cHBvcnRlZDoKICAgICAgICAgICAgZnNpZCA9IHNlbGYuZ2V0X2ZzaWQoKQogICAgICAgICAg
            ICByZ3dfZnNpZCwgZXJyID0gc2VsZi5nZXRfcmd3X2ZzaWQoYmFzZV91cmwsIHZlcmlmeSkKICAg
            ICAgICAgICAgaWYgZXJyID09ICItMSI6CiAgICAgICAgICAgICAgICByZXR1cm4gIi0xIgogICAg
            ICAgICAgICBpZiBmc2lkICE9IHJnd19mc2lkOgogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53
            cml0ZSgKICAgICAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCByZ3cgRW5kcG9pbnQsICd7
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfZW5kcG9pbnR9JywgaXMgaW52YWxpZC4gV2UgYXJlIHZhbGlk
            YXRpbmcgYnkgY2FsbGluZyB0aGUgYWRtaW5vcHMgYXBpIHRocm91Z2ggcmd3LWVuZHBvaW50IGFu
            ZCB2YWxpZGF0aW5nIHRoZSBjbHVzdGVyX2lkICd7cmd3X2ZzaWR9JyBpcyBlcXVhbCB0byB0aGUg
            Y2VwaCBjbHVzdGVyIGZzaWQgJ3tmc2lkfSciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAg
            ICAgICByZXR1cm4gIi0xIgoKICAgICAgICAjIGNoZWNrIGlmIHRoZSByZ3cgZW5kcG9pbnQgcG9v
            bCBleGlzdAogICAgICAgICMgb25seSB2YWxpZGF0ZSBpZiByZ3dfcG9vbF9wcmVmaXggaXMgcGFz
            c2VkIGVsc2UgaXQgd2lsbCB0YWtlIGRlZmF1bHQgdmFsdWUgYW5kIHdlIGRvbid0IGNyZWF0ZSB0
            aGVzZSBkZWZhdWx0IHBvb2xzCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9w
            cmVmaXggIT0gImRlZmF1bHQiOgogICAgICAgICAgICByZ3dfcG9vbHNfdG9fdmFsaWRhdGUgPSBb
            CiAgICAgICAgICAgICAgICBmIntzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3
            Lm1ldGEiLAogICAgICAgICAgICAgICAgIi5yZ3cucm9vdCIsCiAgICAgICAgICAgICAgICBmIntz
            ZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmNvbnRyb2wiLAogICAgICAgICAg
            ICAgICAgZiJ7c2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5sb2ciLAogICAg
            ICAgICAgICBdCiAgICAgICAgICAgIGZvciBfcmd3X3Bvb2xfdG9fdmFsaWRhdGUgaW4gcmd3X3Bv
            b2xzX3RvX3ZhbGlkYXRlOgogICAgICAgICAgICAgICAgaWYgbm90IHNlbGYuY2x1c3Rlci5wb29s
            X2V4aXN0cyhfcmd3X3Bvb2xfdG9fdmFsaWRhdGUpOgogICAgICAgICAgICAgICAgICAgIHN5cy5z
            dGRlcnIud3JpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHBvb2ws
            ICd7X3Jnd19wb29sX3RvX3ZhbGlkYXRlfScsIGRvZXMgbm90IGV4aXN0IgogICAgICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgICAgICAgICByZXR1cm4gIi0xIgoKICAgICAgICByZXR1cm4g
            IiIKCiAgICBkZWYgdmFsaWRhdGVfcmd3X211bHRpc2l0ZShzZWxmLCByZ3dfbXVsdGlzaXRlX2Nv
            bmZpZ19uYW1lLCByZ3dfbXVsdGlzaXRlX2NvbmZpZyk6CiAgICAgICAgaWYgcmd3X211bHRpc2l0
            ZV9jb25maWcgIT0gIiI6CiAgICAgICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgICAgICJyYWRv
            c2d3LWFkbWluIiwKICAgICAgICAgICAgICAgIHJnd19tdWx0aXNpdGVfY29uZmlnLAogICAgICAg
            ICAgICAgICAgImdldCIsCiAgICAgICAgICAgICAgICAiLS1yZ3ctIiArIHJnd19tdWx0aXNpdGVf
            Y29uZmlnLAogICAgICAgICAgICAgICAgcmd3X211bHRpc2l0ZV9jb25maWdfbmFtZSwKICAgICAg
            ICAgICAgXQogICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBfID0gc3VicHJvY2Vzcy5j
            aGVja19vdXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAgICBleGNl
            cHQgc3VicHJvY2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3IgYXMgZXhlY0VycjoKICAgICAgICAgICAg
            ICAgIGVycl9tc2cgPSAoCiAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBj
            b21tYW5kIHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAgICAgICAgICAg
            ICAgICBmIkNvZGU6IHtleGVjRXJyLnJldHVybmNvZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJy
            fSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJy
            X21zZykKICAgICAgICAgICAgICAgIHJldHVybiAiLTEiCiAgICAgICAgcmV0dXJuICIiCgogICAg
            ZGVmIGNvbnZlcnRfY29tbWFfc2VwYXJhdGVkX3RvX2FycmF5KHNlbGYsIHZhbHVlKToKICAgICAg
            ICByZXR1cm4gdmFsdWUuc3BsaXQoIiwiKQoKICAgIGRlZiByYWlzZV9leGNlcHRpb25faWZfYW55
            X3RvcG9sb2d5X2ZsYWdfaXNfbWlzc2luZyhzZWxmKToKICAgICAgICBpZiAoCiAgICAgICAgICAg
            ICgKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMgIT0gIiIK
            ICAgICAgICAgICAgICAgIGFuZCAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCA9PSAiIgogICAgICAgICAgICAgICAgICAg
            IG9yIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfZmFpbHVyZV9kb21haW5fdmFsdWVzID09ICIi
            CiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICkKICAgICAgICAgICAgb3IgKAogICAgICAg
            ICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCAh
            PSAiIgogICAgICAgICAgICAgICAgYW5kICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdf
            cGFyc2VyLnRvcG9sb2d5X3Bvb2xzID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5f
            YXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMgPT0gIiIKICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgKQogICAgICAgICAgICBvciAoCiAgICAgICAgICAgICAgICBz
            ZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX3ZhbHVlcyAhPSAiIgogICAg
            ICAgICAgICAgICAgYW5kICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRv
            cG9sb2d5X3Bvb2xzID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5fYXJnX3BhcnNl
            ci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCA9PSAiIgogICAgICAgICAgICAgICAgKQog
            ICAgICAgICAgICApCiAgICAgICAgKToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVy
            ZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICJwcm92aWRlIGFsbCB0aGUgdG9wb2xvZ3kgZmxh
            Z3MgLS10b3BvbG9neS1wb29scywgLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi1sYWJlbCwgLS10
            b3BvbG9neS1mYWlsdXJlLWRvbWFpbi12YWx1ZXMiCiAgICAgICAgICAgICkKCiAgICBkZWYgdmFs
            aWRhdGVfdG9wb2xvZ3lfdmFsdWVzKHNlbGYsIHRvcG9sb2d5X3Bvb2xzLCB0b3BvbG9neV9mZCk6
            CiAgICAgICAgaWYgbGVuKHRvcG9sb2d5X3Bvb2xzKSAhPSBsZW4odG9wb2xvZ3lfZmQpOgogICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ZiJUaGUgcHJvdmlkZWQgdG9wb2xvZ3kgcG9vbHMsICd7dG9wb2xvZ3lfcG9vbHN9JywgYW5kICIK
            ICAgICAgICAgICAgICAgIGYidG9wb2xvZ3kgZmFpbHVyZSBkb21haW4sICd7dG9wb2xvZ3lfZmR9
            JywiCiAgICAgICAgICAgICAgICBmImFyZSBvZiBkaWZmZXJlbnQgbGVuZ3RoLCAne2xlbih0b3Bv
            bG9neV9wb29scyl9JyBhbmQgJ3tsZW4odG9wb2xvZ3lfZmQpfScgcmVzcGN0aXZlbHkiCiAgICAg
            ICAgICAgICkKICAgICAgICByZXR1cm4KCiAgICBkZWYgdmFsaWRhdGVfdG9wb2xvZ3lfcmJkX3Bv
            b2xzKHNlbGYsIHRvcG9sb2d5X3JiZF9wb29scyk6CiAgICAgICAgZm9yIHBvb2wgaW4gdG9wb2xv
            Z3lfcmJkX3Bvb2xzOgogICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3JiZF9wb29sKHBvb2wpCgog
            ICAgZGVmIGluaXRfdG9wb2xvZ3lfcmJkX3Bvb2xzKHNlbGYsIHRvcG9sb2d5X3JiZF9wb29scyk6
            CiAgICAgICAgZm9yIHBvb2wgaW4gdG9wb2xvZ3lfcmJkX3Bvb2xzOgogICAgICAgICAgICBzZWxm
            LmluaXRfcmJkX3Bvb2wocG9vbCkKCiAgICBkZWYgX2dlbl9vdXRwdXRfbWFwKHNlbGYpOgogICAg
            ICAgIGlmIHNlbGYub3V0X21hcDoKICAgICAgICAgICAgcmV0dXJuCiAgICAgICAgIyBzdXBwb3J0
            IGxlZ2FjeSBmbGFnIHdpdGggdXBncmFkZXMKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNs
            dXN0ZXJfbmFtZToKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1l
            ID0gc2VsZi5fYXJnX3BhcnNlci5jbHVzdGVyX25hbWUKICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            Lms4c19jbHVzdGVyX25hbWUgPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuazhzX2Ns
            dXN0ZXJfbmFtZS5sb3dlcigpCiAgICAgICAgKSAgIyBhbHdheXMgY29udmVydCBjbHVzdGVyIG5h
            bWUgdG8gbG93ZXJjYXNlIGNoYXJhY3RlcnMKICAgICAgICBzZWxmLnZhbGlkYXRlX3JiZF9wb29s
            KHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lKQogICAgICAgIHNlbGYuaW5pdF9y
            YmRfcG9vbChzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZSkKICAgICAgICBzZWxm
            LnZhbGlkYXRlX3JhZG9zX25hbWVzcGFjZSgpCiAgICAgICAgc2VsZi5fZXhjbHVkZWRfa2V5cy5h
            ZGQoIks4U19DTFVTVEVSX05BTUUiKQogICAgICAgIHNlbGYuZ2V0X2NlcGhmc19kYXRhX3Bvb2xf
            ZGV0YWlscygpCiAgICAgICAgc2VsZi5vdXRfbWFwWyJOQU1FU1BBQ0UiXSA9IHNlbGYuX2FyZ19w
            YXJzZXIubmFtZXNwYWNlCiAgICAgICAgc2VsZi5vdXRfbWFwWyJLOFNfQ0xVU1RFUl9OQU1FIl0g
            PSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBb
            IlJPT0tfRVhURVJOQUxfRlNJRCJdID0gc2VsZi5nZXRfZnNpZCgpCiAgICAgICAgc2VsZi5vdXRf
            bWFwWyJST09LX0VYVEVSTkFMX1VTRVJOQU1FIl0gPSBzZWxmLnJ1bl9hc191c2VyCiAgICAgICAg
            c2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0NFUEhfTU9OX0RBVEEiXSA9IHNlbGYuZ2V0X2Nl
            cGhfZXh0ZXJuYWxfbW9uX2RhdGEoKQogICAgICAgIHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5B
            TF9VU0VSX1NFQ1JFVCJdID0gc2VsZi5jcmVhdGVfY2hlY2tlcktleSgKICAgICAgICAgICAgImNs
            aWVudC5oZWFsdGhjaGVja2VyIgogICAgICAgICkKICAgICAgICBzZWxmLm91dF9tYXBbIlJPT0tf
            RVhURVJOQUxfREFTSEJPQVJEX0xJTksiXSA9IHNlbGYuZ2V0X2NlcGhfZGFzaGJvYXJkX2xpbmso
            KQogICAgICAgICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VDUkVU
            Il0sCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FIl0s
            CiAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3VzZXIoImNsaWVudC5jc2kt
            cmJkLW5vZGUiKQogICAgICAgICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BS
            T1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9SQkRfUFJP
            VklTSU9ORVJfU0VDUkVUX05BTUUiXSwKICAgICAgICApID0gc2VsZi5jcmVhdGVfY2VwaENTSUtl
            eXJpbmdfdXNlcigiY2xpZW50LmNzaS1yYmQtcHJvdmlzaW9uZXIiKQogICAgICAgIHNlbGYub3V0
            X21hcFsiQ0VQSEZTX1BPT0xfTkFNRSJdID0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9w
            b29sX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBbIkNFUEhGU19NRVRBREFUQV9QT09MX05BTUUi
            XSA9ICgKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfbWV0YWRhdGFfcG9vbF9u
            YW1lCiAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSA9IHNl
            bGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgIHNlbGYub3V0X21h
            cFsiUkVTVFJJQ1RFRF9BVVRIX1BFUk1JU1NJT04iXSA9ICgKICAgICAgICAgICAgc2VsZi5fYXJn
            X3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbgogICAgICAgICkKICAgICAgICBzZWxm
            Lm91dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdID0gc2VsZi5fYXJnX3BhcnNlci5yYWRvc19uYW1l
            c3BhY2UKICAgICAgICBzZWxmLm91dF9tYXBbIlNVQlZPTFVNRV9HUk9VUCJdID0gc2VsZi5fYXJn
            X3BhcnNlci5zdWJ2b2x1bWVfZ3JvdXAKICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNf
            Tk9ERV9TRUNSRVQiXSA9ICIiCiAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJ
            U0lPTkVSX1NFQ1JFVCJdID0gIiIKICAgICAgICAjIGNyZWF0ZSBDZXBoRlMgbm9kZSBhbmQgcHJv
            dmlzaW9uZXIga2V5cmluZyBvbmx5IHdoZW4gTURTIGV4aXN0cwogICAgICAgIGlmIHNlbGYub3V0
            X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSBhbmQgc2VsZi5vdXRfbWFwWyJDRVBIRlNfUE9PTF9OQU1F
            Il06CiAgICAgICAgICAgICgKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhG
            U19OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZT
            X05PREVfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lL
            ZXlyaW5nX3VzZXIoImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiKQogICAgICAgICAgICAoCiAgICAg
            ICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUIl0s
            CiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VD
            UkVUX05BTUUiXSwKICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3Vz
            ZXIoImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIikKICAgICAgICAgICAgIyBjcmVhdGUg
            dGhlIGRlZmF1bHQgImNzaSIgc3Vidm9sdW1lZ3JvdXAKICAgICAgICAgICAgc2VsZi5nZXRfb3Jf
            Y3JlYXRlX3N1YnZvbHVtZV9ncm91cCgKICAgICAgICAgICAgICAgICJjc2kiLCBzZWxmLl9hcmdf
            cGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICAgICAgKQogICAgICAgICAgICAj
            IHBpbiB0aGUgZGVmYXVsdCAiY3NpIiBzdWJ2b2x1bWVncm91cAogICAgICAgICAgICBzZWxmLnBp
            bl9zdWJ2b2x1bWUoCiAgICAgICAgICAgICAgICAiY3NpIiwgc2VsZi5fYXJnX3BhcnNlci5jZXBo
            ZnNfZmlsZXN5c3RlbV9uYW1lLCAiZGlzdHJpYnV0ZWQiLCAiMSIKICAgICAgICAgICAgKQogICAg
            ICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlNVQlZPTFVNRV9HUk9VUCJdOgogICAgICAgICAgICAg
            ICAgc2VsZi5nZXRfb3JfY3JlYXRlX3N1YnZvbHVtZV9ncm91cCgKICAgICAgICAgICAgICAgICAg
            ICBzZWxmLl9hcmdfcGFyc2VyLnN1YnZvbHVtZV9ncm91cCwKICAgICAgICAgICAgICAgICAgICBz
            ZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgICAgICBzZWxmLnBpbl9zdWJ2b2x1bWUoCiAgICAgICAgICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci5zdWJ2b2x1bWVfZ3JvdXAsCiAgICAgICAgICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lLAogICAgICAgICAgICAgICAgICAg
            ICJkaXN0cmlidXRlZCIsCiAgICAgICAgICAgICAgICAgICAgIjEiLAogICAgICAgICAgICAgICAg
            KQogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1RMU19DRVJUIl0gPSAiIgogICAgICAgIHNlbGYu
            b3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdID0gIiIKICAgICAgICBzZWxmLm91dF9tYXBb
            Ik1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJdID0gIiIKICAgICAgICBpZiBub3Qgc2VsZi5fYXJn
            X3BhcnNlci5za2lwX21vbml0b3JpbmdfZW5kcG9pbnQ6CiAgICAgICAgICAgICgKICAgICAgICAg
            ICAgICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdLAogICAgICAgICAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UX1BPUlQiXSwKICAgICAgICAgICAg
            KSA9IHNlbGYuZ2V0X2FjdGl2ZV9hbmRfc3RhbmRieV9tZ3JzKCkKICAgICAgICBzZWxmLm91dF9t
            YXBbIlJCRF9QT09MX05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1l
            CiAgICAgICAgc2VsZi5vdXRfbWFwWyJSQkRfTUVUQURBVEFfRUNfUE9PTF9OQU1FIl0gPSAoCiAg
            ICAgICAgICAgIHNlbGYudmFsaWRhdGVfcmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZSgpCiAgICAg
            ICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9PTFMiXSA9IHNlbGYuX2FyZ19w
            YXJzZXIudG9wb2xvZ3lfcG9vbHMKICAgICAgICBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxV
            UkVfRE9NQUlOX0xBQkVMIl0gPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xv
            Z3lfZmFpbHVyZV9kb21haW5fbGFiZWwKICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJU
            T1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiXSA9ICgKICAgICAgICAgICAgc2VsZi5fYXJn
            X3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMKICAgICAgICApCiAgICAgICAg
            aWYgKAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X3Bvb2xzICE9ICIiCiAg
            ICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX2xh
            YmVsICE9ICIiCiAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1
            cmVfZG9tYWluX3ZhbHVlcyAhPSAiIgogICAgICAgICk6CiAgICAgICAgICAgIHNlbGYudmFsaWRh
            dGVfdG9wb2xvZ3lfdmFsdWVzKAogICAgICAgICAgICAgICAgc2VsZi5jb252ZXJ0X2NvbW1hX3Nl
            cGFyYXRlZF90b19hcnJheShzZWxmLm91dF9tYXBbIlRPUE9MT0dZX1BPT0xTIl0pLAogICAgICAg
            ICAgICAgICAgc2VsZi5jb252ZXJ0X2NvbW1hX3NlcGFyYXRlZF90b19hcnJheSgKICAgICAgICAg
            ICAgICAgICAgICBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX1ZBTFVFUyJd
            CiAgICAgICAgICAgICAgICApLAogICAgICAgICAgICApCiAgICAgICAgICAgIHNlbGYudmFsaWRh
            dGVfdG9wb2xvZ3lfcmJkX3Bvb2xzKAogICAgICAgICAgICAgICAgc2VsZi5jb252ZXJ0X2NvbW1h
            X3NlcGFyYXRlZF90b19hcnJheShzZWxmLm91dF9tYXBbIlRPUE9MT0dZX1BPT0xTIl0pCiAgICAg
            ICAgICAgICkKICAgICAgICAgICAgc2VsZi5pbml0X3RvcG9sb2d5X3JiZF9wb29scygKICAgICAg
            ICAgICAgICAgIHNlbGYuY29udmVydF9jb21tYV9zZXBhcmF0ZWRfdG9fYXJyYXkoc2VsZi5vdXRf
            bWFwWyJUT1BPTE9HWV9QT09MUyJdKQogICAgICAgICAgICApCiAgICAgICAgZWxzZToKICAgICAg
            ICAgICAgc2VsZi5yYWlzZV9leGNlcHRpb25faWZfYW55X3RvcG9sb2d5X2ZsYWdfaXNfbWlzc2lu
            ZygpCgogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1BPT0xfUFJFRklYIl0gPSBzZWxmLl9hcmdf
            cGFyc2VyLnJnd19wb29sX3ByZWZpeAogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX0VORFBPSU5U
            Il0gPSAiIgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50OgogICAgICAg
            ICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgICAgICBzZWxmLmNy
            ZWF0ZV9yZ3dfYWRtaW5fb3BzX3VzZXIoKQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAg
            ICAgaWYgKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25h
            bWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9u
            ZWdyb3VwX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfem9uZV9uYW1lICE9ICIiCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAg
            ICAgIGVyciA9IHNlbGYudmFsaWRhdGVfcmd3X211bHRpc2l0ZSgKICAgICAgICAgICAgICAgICAg
            ICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwgInJlYWxtIgogICAgICAgICAg
            ICAgICAgICAgICkKICAgICAgICAgICAgICAgICAgICBlcnIgPSBzZWxmLnZhbGlkYXRlX3Jnd19t
            dWx0aXNpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pv
            bmVncm91cF9uYW1lLCAiem9uZWdyb3VwIgogICAgICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgICAgICAgICBlcnIgPSBzZWxmLnZhbGlkYXRlX3Jnd19tdWx0aXNpdGUoCiAgICAgICAgICAg
            ICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVfbmFtZSwgInpvbmUiCiAgICAg
            ICAgICAgICAgICAgICAgKQoKICAgICAgICAgICAgICAgIGlmICgKICAgICAgICAgICAgICAgICAg
            ICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lID09ICIiCiAgICAgICAgICAgICAgICAg
            ICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lID09ICIiCiAgICAgICAg
            ICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVfbmFtZSA9PSAiIgogICAg
            ICAgICAgICAgICAgKSBvciAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5y
            Z3dfcmVhbG1fbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFy
            c2VyLnJnd196b25lZ3JvdXBfbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd196b25lX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICk6CiAgICAg
            ICAgICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJH
            V19BRE1JTl9PUFNfVVNFUl9BQ0NFU1NfS0VZIl0sCiAgICAgICAgICAgICAgICAgICAgICAgIHNl
            bGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX1NFQ1JFVF9LRVkiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkLAogICAgICAgICAgICAgICAgICAgICAgICBl
            cnIsCiAgICAgICAgICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRlX3Jnd19hZG1pbl9vcHNfdXNl
            cigpCiAgICAgICAgICAgICAgICAgICAgZXJyID0gc2VsZi52YWxpZGF0ZV9yZ3dfZW5kcG9pbnQo
            aW5mb19jYXBfc3VwcG9ydGVkKQogICAgICAgICAgICAgICAgICAgIGlmIHNlbGYuX2FyZ19wYXJz
            ZXIucmd3X3Rsc19jZXJ0X3BhdGg6CiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYub3V0X21h
            cFsiUkdXX1RMU19DRVJUIl0gPSAoCiAgICAgICAgICAgICAgICAgICAgICAgICAgICBzZWxmLnZh
            bGlkYXRlX3Jnd19lbmRwb2ludF90bHNfY2VydCgpCiAgICAgICAgICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgICAgICAgICAjIGlmIHRoZXJlIGlzIG5vIGVycm9yLCBzZXQgdGhlIFJHV19F
            TkRQT0lOVAogICAgICAgICAgICAgICAgICAgIGlmIGVyciAhPSAiLTEiOgogICAgICAgICAgICAg
            ICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdID0gc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfZW5kcG9pbnQKICAgICAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICAgICAg
            ZXJyID0gIlBsZWFzZSBwcm92aWRlIGFsbCB0aGUgUkdXIG11bHRpc2l0ZSBwYXJhbWV0ZXJzIG9y
            IG5vbmUgb2YgdGhlbSIKICAgICAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKGVycikK
            CiAgICBkZWYgZ2VuX3NoZWxsX291dChzZWxmKToKICAgICAgICBzZWxmLl9nZW5fb3V0cHV0X21h
            cCgpCiAgICAgICAgc2hPdXRJTyA9IFN0cmluZ0lPKCkKICAgICAgICBmb3IgaywgdiBpbiBzZWxm
            Lm91dF9tYXAuaXRlbXMoKToKICAgICAgICAgICAgaWYgdiBhbmQgayBub3QgaW4gc2VsZi5fZXhj
            bHVkZWRfa2V5czoKICAgICAgICAgICAgICAgIHNoT3V0SU8ud3JpdGUoZiJleHBvcnQge2t9PXt2
            fXtMSU5FU0VQfSIpCiAgICAgICAgc2hPdXQgPSBzaE91dElPLmdldHZhbHVlKCkKICAgICAgICBz
            aE91dElPLmNsb3NlKCkKICAgICAgICByZXR1cm4gc2hPdXQKCiAgICBkZWYgZ2VuX2pzb25fb3V0
            KHNlbGYpOgogICAgICAgIHNlbGYuX2dlbl9vdXRwdXRfbWFwKCkKICAgICAgICBpZiBzZWxmLl9h
            cmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiAiIgogICAgICAgIGpzb25fb3V0
            ID0gWwogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFtZSI6ICJyb29rLWNlcGgtbW9u
            LWVuZHBvaW50cyIsCiAgICAgICAgICAgICAgICAia2luZCI6ICJDb25maWdNYXAiLAogICAgICAg
            ICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiBzZWxmLm91dF9t
            YXBbIlJPT0tfRVhURVJOQUxfQ0VQSF9NT05fREFUQSJdLAogICAgICAgICAgICAgICAgICAgICJt
            YXhNb25JZCI6ICIwIiwKICAgICAgICAgICAgICAgICAgICAibWFwcGluZyI6ICJ7fSIsCiAgICAg
            ICAgICAgICAgICB9LAogICAgICAgICAgICB9LAogICAgICAgICAgICB7CiAgICAgICAgICAgICAg
            ICAibmFtZSI6ICJyb29rLWNlcGgtbW9uIiwKICAgICAgICAgICAgICAgICJraW5kIjogIlNlY3Jl
            dCIsCiAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAiYWRtaW4t
            c2VjcmV0IjogImFkbWluLXNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImZzaWQiOiBzZWxm
            Lm91dF9tYXBbIlJPT0tfRVhURVJOQUxfRlNJRCJdLAogICAgICAgICAgICAgICAgICAgICJtb24t
            c2VjcmV0IjogIm1vbi1zZWNyZXQiLAogICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgfSwK
            ICAgICAgICAgICAgewogICAgICAgICAgICAgICAgIm5hbWUiOiAicm9vay1jZXBoLW9wZXJhdG9y
            LWNyZWRzIiwKICAgICAgICAgICAgICAgICJraW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAg
            ICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJS
            T09LX0VYVEVSTkFMX1VTRVJOQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgInVzZXJLZXkiOiBz
            ZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAg
            IH0sCiAgICAgICAgICAgIH0sCiAgICAgICAgXQoKICAgICAgICAjIGlmICdNT05JVE9SSU5HX0VO
            RFBPSU5UJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ21vbml0b3JpbmctZW5kcG9pbnQnIHRvIENs
            dXN0ZXIKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19F
            TkRQT0lOVCJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9J
            TlRfUE9SVCJdCiAgICAgICAgKToKICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAg
            ICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjogIm1vbml0b3JpbmctZW5kcG9p
            bnQiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIkNlcGhDbHVzdGVyIiwKICAgICAgICAg
            ICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgIk1vbml0b3JpbmdF
            bmRwb2ludCI6IHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdLAogICAgICAgICAg
            ICAgICAgICAgICAgICAiTW9uaXRvcmluZ1BvcnQiOiBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdf
            RU5EUE9JTlRfUE9SVCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9
            CiAgICAgICAgICAgICkKCiAgICAgICAgIyBpZiAnQ1NJX1JCRF9OT0RFX1NFQ1JFVCcgZXhpc3Rz
            LCB0aGVuIG9ubHkgYWRkICdyb29rLWNzaS1yYmQtcHJvdmlzaW9uZXInIFNlY3JldAogICAgICAg
            IGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VDUkVUIl0KICAg
            ICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FIl0KICAg
            ICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAg
            ICAgICAgICAgICAgICAgICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9O
            T0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQi
            LAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAi
            dXNlcklEIjogc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VDUkVUX05BTUUiXSwKICAgICAg
            ICAgICAgICAgICAgICAgICAgInVzZXJLZXkiOiBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9T
            RUNSRVQiXSwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAg
            ICAgICApCiAgICAgICAgIyBpZiAnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQnIGV4aXN0cywg
            dGhlbiBvbmx5IGFkZCAncm9vay1jc2ktcmJkLXByb3Zpc2lvbmVyJyBTZWNyZXQKICAgICAgICBp
            ZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQi
            XQogICAgICAgICAgICBhbmQgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JF
            VF9OQU1FIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAg
            ICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21h
            cFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAg
            ICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAg
            ICAgICAgICAgICAgICAgICAgInVzZXJJRCI6IHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJ
            T05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNl
            bGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAg
            ICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBpZiAnQ1NJ
            X0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAncm9vay1j
            c2ktY2VwaGZzLXByb3Zpc2lvbmVyJyBTZWNyZXQKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNl
            bGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVQiXQogICAgICAgICAgICBh
            bmQgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0KICAg
            ICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAg
            ICAgICAgICAgICAgICAgICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhG
            U19QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQi
            OiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAg
            ICAgICAgICAgImFkbWluSUQiOiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJf
            U0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgImFkbWluS2V5Ijogc2VsZi5v
            dXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAg
            ICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdDU0lf
            Q0VQSEZTX05PREVfU0VDUkVUJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ3Jvb2stY3NpLWNlcGhm
            cy1ub2RlJyBTZWNyZXQKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJ
            X0NFUEhGU19OT0RFX1NFQ1JFVCJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9D
            RVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJdCiAgICAgICAgKToKICAgICAgICAgICAganNvbl9vdXQu
            YXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjogZiJy
            b29rLXtzZWxmLm91dF9tYXBbJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAg
            ICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0
            YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImFkbWluSUQiOiBzZWxmLm91dF9tYXBbIkNT
            SV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAiYWRt
            aW5LZXkiOiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVQiXSwKICAgICAgICAg
            ICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBp
            ZiAnUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSycgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdy
            b29rLWNlcGgtZGFzaGJvYXJkLWxpbmsnIFNlY3JldAogICAgICAgIGlmIHNlbGYub3V0X21hcFsi
            Uk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSyJdOgogICAgICAgICAgICBqc29uX291dC5hcHBl
            bmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAicm9vay1j
            ZXBoLWRhc2hib2FyZC1saW5rIiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQi
            LAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAi
            dXNlcklEIjogImNlcGgtZGFzaGJvYXJkLWxpbmsiLAogICAgICAgICAgICAgICAgICAgICAgICAi
            dXNlcktleSI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSyJdLAog
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAg
            ICAgICAjIGlmICdSQURPU19OQU1FU1BBQ0UnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCB0aGUgIlJB
            RE9TX05BTUVTUEFDRSIgbmFtZXNwYWNlCiAgICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLm91
            dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIlJF
            U1RSSUNURURfQVVUSF9QRVJNSVNTSU9OIl0KICAgICAgICAgICAgYW5kIG5vdCBzZWxmLm91dF9t
            YXBbIlJCRF9NRVRBREFUQV9FQ19QT09MX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpz
            b25fb3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFt
            ZSI6ICJyYWRvcy1uYW1lc3BhY2UiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIkNlcGhC
            bG9ja1Bvb2xSYWRvc05hbWVzcGFjZSIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAg
            ICAgICAgICAgICAgICAgICAgICAgICJyYWRvc05hbWVzcGFjZU5hbWUiOiBzZWxmLm91dF9tYXBb
            IlJBRE9TX05BTUVTUEFDRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYu
            b3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAg
            ICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAg
            ICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1lIjogImNlcGgtcmJkLXJhZG9z
            LW5hbWVzcGFjZSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwK
            ICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgInBv
            b2wiOiBzZWxmLm91dF9tYXBbIlJCRF9QT09MX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAg
            ICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNyZXQtbmFtZSI6IGYicm9vay17
            c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAg
            ICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250cm9sbGVyLWV4cGFuZC1z
            ZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NF
            Q1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5p
            by9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRf
            Tk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAg
            ICAgIH0KICAgICAgICAgICAgKQogICAgICAgIGVsc2U6CiAgICAgICAgICAgIGlmIHNlbGYub3V0
            X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdOgogICAgICAgICAgICAgICAganNvbl9v
            dXQuYXBwZW5kKAogICAgICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAgICAg
            Im5hbWUiOiAiY2VwaC1yYmQiLAogICAgICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9y
            YWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAg
            ICAgICAgICAgICAgICAgICJkYXRhUG9vbCI6IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJd
            LAogICAgICAgICAgICAgICAgICAgICAgICAgICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIlJCRF9N
            RVRBREFUQV9FQ19QT09MX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJjc2ku
            c3RvcmFnZS5rOHMuaW8vcHJvdmlzaW9uZXItc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0
            X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNlY3Jl
            dC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVU
            X05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5p
            by9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRf
            Tk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAg
            ICAgICAgICAgICAgfQogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBlbHNlOgogICAgICAg
            ICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgICAgIHsKICAgICAgICAg
            ICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaC1yYmQiLAogICAgICAgICAgICAgICAgICAgICAg
            ICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAgICAgICAiZGF0YSI6
            IHsKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJwb29sIjogc2VsZi5vdXRfbWFwWyJSQkRf
            UE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2Uuazhz
            LmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9S
            QkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAg
            ICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYi
            cm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwK
            ICAgICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vbm9kZS1zdGFn
            ZS1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX05PREVfU0VDUkVU
            X05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgICAg
            IH0KICAgICAgICAgICAgICAgICkKCiAgICAgICAgIyBpZiAnVE9QT0xPR1lfUE9PTFMnLCAnVE9Q
            T0xPR1lfRkFJTFVSRV9ET01BSU5fTEFCRUwnLCAnVE9QT0xPR1lfRkFJTFVSRV9ET01BSU5fVkFM
            VUVTJyAgZXhpc3RzLAogICAgICAgICMgdGhlbiBvbmx5IGFkZCAndG9wb2xvZ3knIFN0b3JhZ2VD
            bGFzcwogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09M
            UyJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlO
            X0xBQkVMIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfRkFJTFVSRV9E
            T01BSU5fVkFMVUVTIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAg
            ICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaC1yYmQtdG9w
            b2xvZ3kiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAg
            ICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJ0b3BvbG9n
            eUZhaWx1cmVEb21haW5MYWJlbCI6IHNlbGYub3V0X21hcFsKICAgICAgICAgICAgICAgICAgICAg
            ICAgICAgICJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9MQUJFTCIKICAgICAgICAgICAgICAgICAg
            ICAgICAgXSwKICAgICAgICAgICAgICAgICAgICAgICAgInRvcG9sb2d5RmFpbHVyZURvbWFpblZh
            bHVlcyI6IHNlbGYub3V0X21hcFsKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJUT1BPTE9H
            WV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiCiAgICAgICAgICAgICAgICAgICAgICAgIF0sCiAgICAg
            ICAgICAgICAgICAgICAgICAgICJ0b3BvbG9neVBvb2xzIjogc2VsZi5vdXRfbWFwWyJUT1BPTE9H
            WV9QT09MUyJdLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3By
            b3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJP
            VklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0
            b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxm
            Lm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUi
            OiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAg
            ICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCgogICAg
            ICAgICMgaWYgJ0NFUEhGU19GU19OQU1FJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ2NlcGhmcycg
            U3RvcmFnZUNsYXNzCiAgICAgICAgaWYgc2VsZi5vdXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdOgog
            ICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAg
            ICAgICAgICAgIm5hbWUiOiAiY2VwaGZzIiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJT
            dG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAg
            ICAgICAgICAgICAiZnNOYW1lIjogc2VsZi5vdXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdLAogICAg
            ICAgICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYub3V0X21hcFsiQ0VQSEZTX1BPT0xfTkFN
            RSJdLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lv
            bmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9DRVBIRlNfUFJPVklT
            SU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3Jh
            Z2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91
            dF9tYXBbJ0NTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUi
            OiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FJ119IiwK
            ICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAg
            ICAgICAgIyBpZiAnUkdXX0VORFBPSU5UJyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ2NlcGgtcmd3
            JyBTdG9yYWdlQ2xhc3MKICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdOgog
            ICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAg
            ICAgICAgICAgIm5hbWUiOiAiY2VwaC1yZ3ciLAogICAgICAgICAgICAgICAgICAgICJraW5kIjog
            IlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAg
            ICAgICAgICAgICAgICJlbmRwb2ludCI6IHNlbGYub3V0X21hcFsiUkdXX0VORFBPSU5UIl0sCiAg
            ICAgICAgICAgICAgICAgICAgICAgICJwb29sUHJlZml4Ijogc2VsZi5vdXRfbWFwWyJSR1dfUE9P
            TF9QUkVGSVgiXSwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAg
            ICAgICAgICApCiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsK
            ICAgICAgICAgICAgICAgICAgICAibmFtZSI6ICJyZ3ctYWRtaW4tb3BzLXVzZXIiLAogICAgICAg
            ICAgICAgICAgICAgICJraW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEi
            OiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJhY2Nlc3NLZXkiOiBzZWxmLm91dF9tYXBbIlJH
            V19BRE1JTl9PUFNfVVNFUl9BQ0NFU1NfS0VZIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJz
            ZWNyZXRLZXkiOiBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9TRUNSRVRfS0VZIl0s
            CiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQog
            ICAgICAgICMgaWYgJ1JHV19UTFNfQ0VSVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkIHRoZSAiY2Vw
            aC1yZ3ctdGxzLWNlcnQiIHNlY3JldAogICAgICAgIGlmIHNlbGYub3V0X21hcFsiUkdXX1RMU19D
            RVJUIl06CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAg
            ICAgICAgICAgICAgICAgICAibmFtZSI6ICJjZXBoLXJndy10bHMtY2VydCIsCiAgICAgICAgICAg
            ICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsK
            ICAgICAgICAgICAgICAgICAgICAgICAgImNlcnQiOiBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VS
            VCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAg
            ICkKCiAgICAgICAgcmV0dXJuIGpzb24uZHVtcHMoanNvbl9vdXQpICsgTElORVNFUAoKICAgIGRl
            ZiB1cGdyYWRlX3VzZXJzX3Blcm1pc3Npb25zKHNlbGYpOgogICAgICAgIHVzZXJzID0gWwogICAg
            ICAgICAgICAiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIsCiAgICAgICAgICAgICJjbGllbnQuY3Np
            LWNlcGhmcy1wcm92aXNpb25lciIsCiAgICAgICAgICAgICJjbGllbnQuY3NpLXJiZC1ub2RlIiwK
            ICAgICAgICAgICAgImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwKICAgICAgICAgICAgImNs
            aWVudC5oZWFsdGhjaGVja2VyIiwKICAgICAgICBdCiAgICAgICAgaWYgc2VsZi5ydW5fYXNfdXNl
            ciAhPSAiIiBhbmQgc2VsZi5ydW5fYXNfdXNlciBub3QgaW4gdXNlcnM6CiAgICAgICAgICAgIHVz
            ZXJzLmFwcGVuZChzZWxmLnJ1bl9hc191c2VyKQogICAgICAgIGZvciB1c2VyIGluIHVzZXJzOgog
            ICAgICAgICAgICBzZWxmLnVwZ3JhZGVfdXNlcl9wZXJtaXNzaW9ucyh1c2VyKQoKICAgIGRlZiBn
            ZXRfcmd3X3Bvb2xfbmFtZV9kdXJpbmdfdXBncmFkZShzZWxmLCB1c2VyLCBjYXBzKToKICAgICAg
            ICBpZiB1c2VyID09ICJjbGllbnQuaGVhbHRoY2hlY2tlciI6CiAgICAgICAgICAgICMgd2hlbiBh
            ZG1pbiBoYXMgbm90IHByb3ZpZGVkIHJndyBwb29sIG5hbWUgZHVyaW5nIHVwZ3JhZGUsCiAgICAg
            ICAgICAgICMgZ2V0IHRoZSByZ3cgcG9vbCBuYW1lIGZyb20gY2xpZW50LmhlYWx0aGNoZWNrZXIg
            dXNlciB3aGljaCB3YXMgdXNlZCBkdXJpbmcgY29ubmVjdGlvbgogICAgICAgICAgICBpZiBub3Qg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXg6CiAgICAgICAgICAgICAgICAjIFRvIGdl
            dCB2YWx1ZSAnZGVmYXVsdCcgd2hpY2ggaXMgcmd3IHBvb2wgbmFtZSBmcm9tICdhbGxvdyByd3gg
            cG9vbD1kZWZhdWx0LnJndy5tZXRhJwogICAgICAgICAgICAgICAgcGF0dGVybiA9IHIicG9vbD0o
            Lio/KVwucmd3XC5tZXRhIgogICAgICAgICAgICAgICAgbWF0Y2ggPSByZS5zZWFyY2gocGF0dGVy
            biwgY2FwcykKICAgICAgICAgICAgICAgIGlmIG1hdGNoOgogICAgICAgICAgICAgICAgICAgIHNl
            bGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4ID0gbWF0Y2guZ3JvdXAoMSkKICAgICAgICAg
            ICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAgICAgImZhaWxlZCB0byBnZXQgcmd3IHBvb2wg
            bmFtZSBmb3IgdXBncmFkZSIKICAgICAgICAgICAgICAgICAgICApCgogICAgZGVmIHVwZ3JhZGVf
            dXNlcl9wZXJtaXNzaW9ucyhzZWxmLCB1c2VyKToKICAgICAgICAjIGNoZWNrIHdoZXRoZXIgdGhl
            IGdpdmVuIHVzZXIgZXhpc3RzIG9yIG5vdAogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAi
            YXV0aCBnZXQiLCAiZW50aXR5IjogZiJ7dXNlcn0iLCAiZm9ybWF0IjogImpzb24ifQogICAgICAg
            IHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihj
            bWRfanNvbikKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgog
            ICAgICAgICAgICBwcmludChmInVzZXIge3VzZXJ9IG5vdCBmb3VuZCBmb3IgdXBncmFkaW5nLiIp
            CiAgICAgICAgICAgIHJldHVybgogICAgICAgIGV4aXN0aW5nX2NhcHMgPSBqc29uX291dFswXVsi
            Y2FwcyJdCiAgICAgICAgc2VsZi5nZXRfcmd3X3Bvb2xfbmFtZV9kdXJpbmdfdXBncmFkZSh1c2Vy
            LCBzdHIoZXhpc3RpbmdfY2FwcykpCiAgICAgICAgbmV3X2NhcCwgXyA9IHNlbGYuZ2V0X2NhcHNf
            YW5kX2VudGl0eSh1c2VyKQogICAgICAgIGNhcF9rZXlzID0gWyJtb24iLCAibWdyIiwgIm9zZCIs
            ICJtZHMiXQogICAgICAgIGNhcHMgPSBbXQogICAgICAgIGZvciBlYWNoQ2FwIGluIGNhcF9rZXlz
            OgogICAgICAgICAgICBjdXJfY2FwX3ZhbHVlcyA9IGV4aXN0aW5nX2NhcHMuZ2V0KGVhY2hDYXAs
            ICIiKQogICAgICAgICAgICBuZXdfY2FwX3ZhbHVlcyA9IG5ld19jYXAuZ2V0KGVhY2hDYXAsICIi
            KQogICAgICAgICAgICBjdXJfY2FwX3Blcm1fbGlzdCA9IFsKICAgICAgICAgICAgICAgIHguc3Ry
            aXAoKSBmb3IgeCBpbiBjdXJfY2FwX3ZhbHVlcy5zcGxpdCgiLCIpIGlmIHguc3RyaXAoKQogICAg
            ICAgICAgICBdCiAgICAgICAgICAgIG5ld19jYXBfcGVybV9saXN0ID0gWwogICAgICAgICAgICAg
            ICAgeC5zdHJpcCgpIGZvciB4IGluIG5ld19jYXBfdmFsdWVzLnNwbGl0KCIsIikgaWYgeC5zdHJp
            cCgpCiAgICAgICAgICAgIF0KICAgICAgICAgICAgIyBhcHBlbmQgbmV3X2NhcF9saXN0IHRvIGN1
            cl9jYXBfbGlzdCB0byBtYWludGFpbiB0aGUgb3JkZXIgb2YgY2FwcwogICAgICAgICAgICBjdXJf
            Y2FwX3Blcm1fbGlzdC5leHRlbmQobmV3X2NhcF9wZXJtX2xpc3QpCiAgICAgICAgICAgICMgZWxp
            bWluYXRlIGR1cGxpY2F0ZXMgd2l0aG91dCB1c2luZyAnc2V0JwogICAgICAgICAgICAjIHNldCBy
            ZS1vcmRlcnMgaXRlbXMgaW4gdGhlIGxpc3QgYW5kIHdlIGhhdmUgdG8ga2VlcCB0aGUgb3JkZXIK
            ICAgICAgICAgICAgbmV3X2NhcF9saXN0ID0gW10KICAgICAgICAgICAgW25ld19jYXBfbGlzdC5h
            cHBlbmQoeCkgZm9yIHggaW4gY3VyX2NhcF9wZXJtX2xpc3QgaWYgeCBub3QgaW4gbmV3X2NhcF9s
            aXN0XQogICAgICAgICAgICBleGlzdGluZ19jYXBzW2VhY2hDYXBdID0gIiwgIi5qb2luKG5ld19j
            YXBfbGlzdCkKICAgICAgICAgICAgaWYgZXhpc3RpbmdfY2Fwc1tlYWNoQ2FwXToKICAgICAgICAg
            ICAgICAgIGNhcHMuYXBwZW5kKGVhY2hDYXApCiAgICAgICAgICAgICAgICBjYXBzLmFwcGVuZChl
            eGlzdGluZ19jYXBzW2VhY2hDYXBdKQogICAgICAgIGNtZF9qc29uID0gewogICAgICAgICAgICAi
            cHJlZml4IjogImF1dGggY2FwcyIsCiAgICAgICAgICAgICJlbnRpdHkiOiB1c2VyLAogICAgICAg
            ICAgICAiY2FwcyI6IGNhcHMsCiAgICAgICAgICAgICJmb3JtYXQiOiAianNvbiIsCiAgICAgICAg
            fQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9q
            c29uX2dlbihjbWRfanNvbikKICAgICAgICBpZiByZXRfdmFsICE9IDA6CiAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIidhdXRoIGNh
            cHMge3VzZXJ9JyBjb21tYW5kIGZhaWxlZC5cbiBFcnJvcjoge2Vycl9tc2d9IgogICAgICAgICAg
            ICApCiAgICAgICAgcHJpbnQoZiJVcGRhdGVkIHVzZXIge3VzZXJ9IHN1Y2Nlc3NmdWxseS4iKQoK
            ICAgIGRlZiBtYWluKHNlbGYpOgogICAgICAgIGdlbmVyYXRlZF9vdXRwdXQgPSAiIgogICAgICAg
            IGlmIHNlbGYuX2FyZ19wYXJzZXIudXBncmFkZToKICAgICAgICAgICAgc2VsZi51cGdyYWRlX3Vz
            ZXJzX3Blcm1pc3Npb25zKCkKICAgICAgICBlbGlmIHNlbGYuX2FyZ19wYXJzZXIuZm9ybWF0ID09
            ICJqc29uIjoKICAgICAgICAgICAgZ2VuZXJhdGVkX291dHB1dCA9IHNlbGYuZ2VuX2pzb25fb3V0
            KCkKICAgICAgICBlbGlmIHNlbGYuX2FyZ19wYXJzZXIuZm9ybWF0ID09ICJiYXNoIjoKICAgICAg
            ICAgICAgZ2VuZXJhdGVkX291dHB1dCA9IHNlbGYuZ2VuX3NoZWxsX291dCgpCiAgICAgICAgZWxz
            ZToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAg
            ICAgICAgIGYiVW5zdXBwb3J0ZWQgZm9ybWF0OiB7c2VsZi5fYXJnX3BhcnNlci5mb3JtYXR9Igog
            ICAgICAgICAgICApCiAgICAgICAgcHJpbnQoZ2VuZXJhdGVkX291dHB1dCkKICAgICAgICBpZiBz
            ZWxmLm91dHB1dF9maWxlIGFuZCBnZW5lcmF0ZWRfb3V0cHV0OgogICAgICAgICAgICBmT3V0ID0g
            b3BlbihzZWxmLm91dHB1dF9maWxlLCBtb2RlPSJ3IiwgZW5jb2Rpbmc9IlVURi04IikKICAgICAg
            ICAgICAgZk91dC53cml0ZShnZW5lcmF0ZWRfb3V0cHV0KQogICAgICAgICAgICBmT3V0LmNsb3Nl
            KCkKCgojIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMKIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIE1BSU4gIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIwppZiBfX25hbWVfXyA9PSAiX19t
            YWluX18iOgogICAgcmpPYmogPSBSYWRvc0pTT04oKQogICAgdHJ5OgogICAgICAgIHJqT2JqLm1h
            aW4oKQogICAgZXhjZXB0IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24gYXMgZXJyOgogICAgICAg
            IHByaW50KGYiRXhlY3V0aW9uIEZhaWxlZDoge2Vycn0iKQogICAgICAgIHJhaXNlIGVycgogICAg
            ZXhjZXB0IEtleUVycm9yIGFzIGtFcnI6CiAgICAgICAgcHJpbnQoZiJLZXlFcnJvcjoge2tFcnJ9
            IikKICAgIGV4Y2VwdCBPU0Vycm9yIGFzIG9zRXJyOgogICAgICAgIHByaW50KGYiRXJyb3Igd2hp
            bGUgdHJ5aW5nIHRvIG91dHB1dCB0aGUgZGF0YToge29zRXJyfSIpCiAgICBmaW5hbGx5OgogICAg
            ICAgIHJqT2JqLnNodXRkb3duKCkK
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.16.16-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/red-hat-storage/rook
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Ceph cluster.
            displayName: Ceph Cluster
            kind: CephCluster
            name: cephclusters.ceph.rook.io
            version: v1
          - description: Represents a Ceph Block Pool.
            displayName: Ceph Block Pool
            kind: CephBlockPool
            name: cephblockpools.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store.
            displayName: Ceph Object Store
            kind: CephObjectStore
            name: cephobjectstores.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store User.
            displayName: Ceph Object Store User
            kind: CephObjectStoreUser
            name: cephobjectstoreusers.ceph.rook.io
            version: v1
          - description: Represents a cluster of Ceph NFS ganesha gateways.
            displayName: Ceph NFS
            kind: CephNFS
            name: cephnfses.ceph.rook.io
            version: v1
          - description: Represents a Ceph User.
            displayName: Ceph Client
            kind: CephClient
            name: cephclients.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem.
            displayName: Ceph Filesystem
            kind: CephFilesystem
            name: cephfilesystems.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem Mirror.
            displayName: Ceph Filesystem Mirror
            kind: CephFilesystemMirror
            name: cephfilesystemmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph RBD Mirror.
            displayName: Ceph RBD Mirror
            kind: CephRBDMirror
            name: cephrbdmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Realm.
            displayName: Ceph Object Store Realm
            kind: CephObjectRealm
            name: cephobjectrealms.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone Group.
            displayName: Ceph Object Store Zone Group
            kind: CephObjectZoneGroup
            name: cephobjectzonegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone.
            displayName: Ceph Object Store Zone
            kind: CephObjectZone
            name: cephobjectzones.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Notification.
            displayName: Ceph Bucket Notification
            kind: CephBucketNotification
            name: cephbucketnotifications.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Topic.
            displayName: Ceph Bucket Topic
            kind: CephBucketTopic
            name: cephbuckettopics.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem SubVolumeGroup.
            displayName: Ceph Filesystem SubVolumeGroup
            kind: CephFilesystemSubVolumeGroup
            name: cephfilesystemsubvolumegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph BlockPool Rados Namespace.
            displayName: Ceph BlockPool Rados Namespace
            kind: CephBlockPoolRadosNamespace
            name: cephblockpoolradosnamespaces.ceph.rook.io
            version: v1
          - description: Represents a Ceph COSI Driver.
            displayName: Ceph COSI Driver
            kind: CephCOSIDriver
            name: cephcosidrivers.ceph.rook.io
            version: v1
        description: |2

          The Rook-Ceph storage operator packages, deploys, manages, upgrades and scales Ceph storage for providing persistent storage to infrastructure services (Logging, Metrics, Registry) as well as stateful applications in Kubernetes clusters.

          ## Rook-Ceph Storage Operator

          Rook runs as a cloud-native service in Kubernetes clusters for optimal integration with applications in need of storage, and handles the heavy-lifting behind the scenes such as provisioning and management.
          Rook orchestrates battle-tested open-source storage technology Ceph, which has years of production deployments and runs some of the worlds largest clusters.

          Ceph is a massively scalable, software-defined, cloud native storage platform that offers block, file and object storage services.
          Ceph can be used to back a wide variety of applications including relational databases, NoSQL databases, CI/CD tool-sets, messaging, AI/ML and analytics applications.
          Ceph is a proven storage platform that backs some of the world's largest storage deployments and has a large vibrant open source community backing the project.

          ## Supported features
          * **High Availability and resiliency** - Ceph has no single point of failures (SPOF) and all its components work natively in a highly available fashion
          * **Data Protection** - Ceph periodically scrub for inconsistent objects and repair them if necessary, making sure your replicas are always coherent
          * **Consistent storage platform across hybrid cloud** - Ceph can be deployed anywhere (on-premise or bare metal) and thus offers a similar experience regardless
          * **Block, File & Object storage service** - Ceph can expose your data through several storage interfaces, solving all the application use cases
          * **Scale up/down** - addition and removal of storage is fully covered by the operator.
          * **Dashboard** - The Operator deploys a dashboard for monitoring and introspecting your cluster.

          ## Before you start
          https://rook.io/docs/rook/v1.0/k8s-pre-reqs.html
        displayName: Rook-Ceph
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rook
        - ceph
        - storage
        - object storage
        - open source
        - block storage
        - shared filesystem
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/rook
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:cf7f72469d25fb37301a8159406613673f8968c82af1035fd45a470e0ae8acfa
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:80c01633a6bfca44df37ac38f764035be6998fce965c8709e4f6d54d17eeaa7b
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6fee56851da3a87b9f7f477fc59e3385e118dc0b398861a8f13459c998a5af77
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c11a599bfd09b9b22888e03eabdbe1054d4049718a591d2a45bf0ed89800111e
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:82a4b481028195cfd75471656fe40c1cbbade4c7b2d170d6bf9422d9d3b84fa7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:693a60d6a1162659213ae554b1c6ee200b0300a7da72c01415044cf507acab49
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:0afcb516eb528857887c6ee1a9cba081dc8f6b7d47d9f26a0dbb2dca0cefe07c
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:8d8326c74795c9c6bd8f33a9f9e81d2304a70b1f9be3c2f8d50e76be4b0292ff
        version: 4.16.16-rhodf
      entries:
      - name: rook-ceph-operator.v4.16.16-rhodf
        version: 4.16.16-rhodf
      - name: rook-ceph-operator.v4.16.15-rhodf
        version: 4.16.15-rhodf
      - name: rook-ceph-operator.v4.16.14-rhodf
        version: 4.16.14-rhodf
      - name: rook-ceph-operator.v4.16.13-rhodf
        version: 4.16.13-rhodf
      - name: rook-ceph-operator.v4.16.12-rhodf
        version: 4.16.12-rhodf
      - name: rook-ceph-operator.v4.16.11-rhodf
        version: 4.16.11-rhodf
      - name: rook-ceph-operator.v4.16.10-rhodf
        version: 4.16.10-rhodf
      - name: rook-ceph-operator.v4.16.9-rhodf
        version: 4.16.9-rhodf
      - name: rook-ceph-operator.v4.16.8-rhodf
        version: 4.16.8-rhodf
      - name: rook-ceph-operator.v4.16.7-rhodf
        version: 4.16.7-rhodf
      - name: rook-ceph-operator.v4.16.6-rhodf
        version: 4.16.6-rhodf
      - name: rook-ceph-operator.v4.16.5-rhodf
        version: 4.16.5-rhodf
      - name: rook-ceph-operator.v4.16.4-rhodf
        version: 4.16.4-rhodf
      - name: rook-ceph-operator.v4.16.3-rhodf
        version: 4.16.3-rhodf
      - name: rook-ceph-operator.v4.16.2-rhodf
        version: 4.16.2-rhodf
      - name: rook-ceph-operator.v4.16.1-rhodf
        version: 4.16.1-rhodf
      - name: rook-ceph-operator.v4.16.0-rhodf
        version: 4.16.0-rhodf
      name: stable-4.16
    defaultChannel: stable-4.16
    packageName: rook-ceph-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: logic-operator-rhel8
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: logic-operator-rhel8.v1.36.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "host/namespace",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          createdAt: "2025-05-26T14:10:15Z"
          description: OpenShift Serverless Logic Kubernetes Operator for deploying
            workflow applications based on the CNCF Serverless Workflow specification
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-serverless-logic
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kubesmarts/kie-tools
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: Sonata Flow Build
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: Sonata Flow Cluster Platform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: Sonata Flow Platform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: Sonata Flow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          OpenShift Serverless Logic Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification v0.8](https://serverlessworkflow.io/).
        displayName: OpenShift Serverless Logic Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        - workflows
        links:
        - name: Documentation
          url: https://docs.openshift.com/serverless/
        - name: Source Repository
          url: https://github.com/kiegroup/kogito-serverless-operator/
        maintainers:
        - email: serverless-logic-team@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/logic-rhel8-operator@sha256:8d3682448ebdac3aeabb2d23842b7e67a252b95f959c408af805037f9728fd3c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4564ca3dc5bac80d6faddaf94c817fbbc270698a9399d8a21ee1005d85ceda56
        version: 1.36.0
      entries:
      - name: logic-operator-rhel8.v1.36.0
        version: 1.36.0
      - name: logic-operator-rhel8.v1.35.0
        version: 1.35.0
      - name: logic-operator-rhel8.v1.34.0
        version: 1.34.0
      - name: logic-operator-rhel8.v1.33.0
        version: 1.33.0
      name: alpha
    defaultChannel: alpha
    packageName: logic-operator-rhel8
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudwatchlogs-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudwatchlogs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudwatchlogs.services.k8s.aws/v1alpha1",
                "kind": "LogGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudwatchlogs-controller:1.1.1
          createdAt: "2025-09-19T20:51:16Z"
          description: AWS CloudWatch Logs controller is a service controller for
            managing CloudWatch Logs resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LogGroup represents the state of an AWS cloudwatchlogs LogGroup
              resource.
            displayName: LogGroup
            kind: LogGroup
            name: loggroups.cloudwatchlogs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudWatch Logs resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudWatch Logs**

          You can use Amazon CloudWatch Logs to monitor, store, and access your log files from Amazon Elastic Compute Cloud (Amazon EC2) instances, AWS CloudTrail, Route 53, and other sources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudWatch Logs
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudwatchlogs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudWatch Logs Developer Resources
          url: https://aws.amazon.com/cloudwatch/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudwatch maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudwatchlogs-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-cloudwatchlogs-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudwatchlogs-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudwatchlogs-controller.v1.0.14
        version: 1.0.14
      - name: ack-cloudwatchlogs-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudwatchlogs-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudwatchlogs-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudwatchlogs-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudwatchlogs-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudwatchlogs-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudwatchlogs-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudwatchlogs-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudwatchlogs-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudwatchlogs-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudwatchlogs-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudwatchlogs-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudwatchlogs-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudwatchlogs-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudwatchlogs-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudwatchlogs-controller.v0.0.7
        version: 0.0.7
      - name: ack-cloudwatchlogs-controller.v0.0.4
        version: 0.0.4
      - name: ack-cloudwatchlogs-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudwatchlogs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: TrendMicro
      provider-url: trendmicro.com
    name: visionone-containersecurity
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: visionone-containersecurity.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "visionone-containersecurity-sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          com.redhat.openshift.versions: v4.12-v4.20
          containerImage: public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
          createdAt: "2025-09-18T22:43:41Z"
          description: Trend Vision One™ Container Security delivers full lifecycle
            protection for containers with real-time threat detection, policy enforcement,
            compliance assurance, and rapid incident response.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trendmicro-system
          operatorhub.io/ui-metadata-max-k8s-version: "1.26"
          operators.openshift.io/valid-subscription: VisionOneContainerSecurity Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trendmicro/visionone-container-security-helm
          support: TrendMicro Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trend Vision One™ Container Security delivers full lifecycle
              protection for containers with real-time threat detection, policy enforcement,
              compliance assurance, and rapid incident response.
            displayName: VisionOneContainerSecurity
            kind: VisionOneContainerSecurity
            name: visiononecontainersecurities.container-security.trendmicro.com
            version: v1alpha1
        description: "[Trend Vision One™ Container Security](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-container-security)
          delivers comprehensive, end-to-end\n protection for containerized workloads
          across OpenShift and Kubernetes environments.\n It continuously monitors
          for threats, vulnerabilities, policy violations, and\n exposed secrets ensuring
          compliance and enabling rapid, automated response. Purpose-built\n for hybrid
          cloud, it instantly isolates compromised containers to prevent lateral\n
          movement and minimize impact. Seamlessly integrating into OpenShift DevSecOps\n
          pipelines, it empowers security teams to enforce policies while enabling
          developers\n to innovate and deploy with confidence without disrupting workflows
          or compromising\n agility.\n\n **Installation**\n \n - Use the recommended
          `trendmicro-system` namespace.\n - Refer to the [Trend Vision One™ Documentation](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator)
          for registration and installation steps.\n"
        displayName: Trend Vision One™ Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trendmicro
        - visionone
        - container-security
        links:
        - name: Documentation
          url: https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator
        maintainers:
        - email: support@trendmicro.com
          name: TrendMicro VisionOne Support Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: TrendMicro
          url: trendmicro.com
        relatedImages:
        - public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
        version: 0.0.2
      entries:
      - name: visionone-containersecurity.v0.0.2
        version: 0.0.2
      - name: visionone-containersecurity.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: visionone-containersecurity.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "visionone-containersecurity-sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          com.redhat.openshift.versions: v4.12-v4.20
          containerImage: public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
          createdAt: "2025-09-18T22:43:41Z"
          description: Trend Vision One™ Container Security delivers full lifecycle
            protection for containers with real-time threat detection, policy enforcement,
            compliance assurance, and rapid incident response.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trendmicro-system
          operatorhub.io/ui-metadata-max-k8s-version: "1.26"
          operators.openshift.io/valid-subscription: VisionOneContainerSecurity Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trendmicro/visionone-container-security-helm
          support: TrendMicro Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trend Vision One™ Container Security delivers full lifecycle
              protection for containers with real-time threat detection, policy enforcement,
              compliance assurance, and rapid incident response.
            displayName: VisionOneContainerSecurity
            kind: VisionOneContainerSecurity
            name: visiononecontainersecurities.container-security.trendmicro.com
            version: v1alpha1
        description: "[Trend Vision One™ Container Security](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-container-security)
          delivers comprehensive, end-to-end\n protection for containerized workloads
          across OpenShift and Kubernetes environments.\n It continuously monitors
          for threats, vulnerabilities, policy violations, and\n exposed secrets ensuring
          compliance and enabling rapid, automated response. Purpose-built\n for hybrid
          cloud, it instantly isolates compromised containers to prevent lateral\n
          movement and minimize impact. Seamlessly integrating into OpenShift DevSecOps\n
          pipelines, it empowers security teams to enforce policies while enabling
          developers\n to innovate and deploy with confidence without disrupting workflows
          or compromising\n agility.\n\n **Installation**\n \n - Use the recommended
          `trendmicro-system` namespace.\n - Refer to the [Trend Vision One™ Documentation](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator)
          for registration and installation steps.\n"
        displayName: Trend Vision One™ Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trendmicro
        - visionone
        - container-security
        links:
        - name: Documentation
          url: https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator
        maintainers:
        - email: support@trendmicro.com
          name: TrendMicro VisionOne Support Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: TrendMicro
          url: trendmicro.com
        relatedImages:
        - public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
        version: 0.0.2
      entries:
      - name: visionone-containersecurity.v0.0.2
        version: 0.0.2
      name: stable
    defaultChannel: stable
    packageName: visionone-containersecurity
    provider:
      name: TrendMicro
      url: trendmicro.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Docling Maintainers
      provider-url: https://docling-project.github.io/docling/
    name: docling-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: docling-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2025-05-19T20:20:14Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DoclingServe is the Schema for the doclingserves API
            displayName: Docling Serve
            kind: DoclingServe
            name: doclingserves.docling.github.io
            version: v1alpha1
        description: |-
          **Overview**

          The Docling Operator is responsible for managing the full lifecycle of Docling Serve instances in your Kubernetes or OpenShift container platforms.


          **Docling Serve**

          The Docling Serve application deploys a REST API for consuming and integrating Docling in your workflows.

          The API offers endpoints for multiple types of interaction, i.e. both synchronous and asynchronous endpoints, as well as status subscriptions via websockets.

          The options for the Docling conversion can be customized at runtime by the users (e.g. which OCR engine, which enrichment models, etc). Users can also choose which of the many DoclingDocument export formats (JSON, Markdown, HTML, DocTags, etc) to receive in the result response.

          Docling Serve allows to choose between a local compute engine, i.e. running Docling in the same container as the API server, or a distributed compute engine which leverages Kubeflow Pipelines to parallelize large-scale ingestion jobs.

          **Docling**

          Docling is a powerful library which simplifies document processing, parsing diverse formats — including advanced PDF understanding — and providing seamless integrations with the gen AI ecosystem.

          Learn more about the Docling Project by visiting the GitHub organization at [https://github.com/docling-project](https://github.com/docling-project).


          **Sample DoclingServe Custom Resources**


          To deploy with the `local` compute engine with the UI enabled and Route exposed:
          ```yaml
          apiVersion: docling.github.io/v1alpha1
          kind: DoclingServe
          metadata:
            labels:
              app.kubernetes.io/name: docling-operator
              app.kubernetes.io/managed-by: kustomize
            name: doclingserve-local-sample
          spec:
            apiServer:
              image: "quay.io/docling-project/docling-serve:latest"
              enableUI: true
              instances: 1
            engine:
              local:
                numWorkers: 2
            route:
              enabled: true
          ```

          To deploy with the `KFP` compute engine with the UI enabled and Route exposed. **Note: Kubeflow Pipeline must be installed and setup in the cluster to use this option.**:
          ```yaml
          apiVersion: docling.github.io/v1alpha1
          kind: DoclingServe
          metadata:
            labels:
              app.kubernetes.io/name: docling-operator
              app.kubernetes.io/managed-by: kustomize
            name: doclingserve-kfp-sample
          spec:
            apiServer:
              image: "quay.io/docling-project/docling-serve:latest"
              enableUI: true
              instances: 1
            engine:
              kfp:
                endpoint: "https://NAME.NAMESPACE.svc.cluster.local:8888"
            route:
              enabled: true
          ```

          **Additional Configurations**


          The CRD exposes `apiServer.configMapName` to enable additional configurations of the `docling-serve` workload. For a full list of configurations can be found on docling-serve's
          [Configuration](https://github.com/docling-project/docling-serve/blob/main/docs/configuration.md) page.
        displayName: Docling Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - document
        - processing
        links:
        - name: Docling Operator
          url: https://docling-project.github.io/docling/
        maintainers:
        - email: docling-technical-discuss@lists.lfaidata.foundation
          name: Docling Maintainers
        maturity: alpha
        provider:
          name: Docling Maintainers
          url: https://docling-project.github.io/docling/
        relatedImages:
        - quay.io/docling-project/docling-operator@sha256:200d8db51874c4b1e3d20d8d1423e9a4807843fbb29f514cd5d8ba872e47f4af
        version: 0.0.1
      entries:
      - name: docling-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: docling-operator
    provider:
      name: Docling Maintainers
      url: https://docling-project.github.io/docling/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubemq.io
      provider-url: ""
    name: kubemq-operator-marketplace
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubemq-operator-marketplace.1.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
              "apiVersion": "core.k8s.kubemq.io/v1beta1",
              "kind": "KubemqCluster",
              "metadata": {
                "name": "kubemq-cluster"
              },
              "spec": {
                "key": "<Your-KubeMQ-License-Key>",
                "replicas": 3
              }
            },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqConnector",
                "metadata": {
                  "name": "kubemq-bridges"
                },
                "spec": {
                  "type": "bridges",
                  "config": "bindings: null"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.connect.redhat.com/kubemq/kubemq-operator@sha256:3a9c87e462eb1caa81925dee72407f73fca41f19a3f9bcaf4e76e569db1dfbab
          createdAt: 2020-03-11 12:00:00+00:00
          description: Kubernetes Messaging Platform
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          support: support@kubemq.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kubemq Cluster
            displayName: Kubemq Cluster
            kind: KubemqCluster
            name: kubemqclusters.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Connector
            displayName: Kubemq Connector
            kind: KubemqConnector
            name: kubemqconnectors.core.k8s.kubemq.io
            version: v1beta1
        description: |2-

          ## Kubernetes Messaging Platform
          A message-based architecture for microservices providing flexible connectivity for hybrid environments: multi-cloud, on-premises, and at the edge. Powered by enterprise-grade message broker and message queue with ready to use connectors, bridges, and control center.

          Messaging platform components:
            - **KubeMQ Cluster** - Enterprise-grade message broker and message queue, scalable, high available, and secured.
            - **KubeMQ Targets Connector** - connects a KubeMQ cluster with external systems and cloud services.
            - **KubeMQ Bridges Connector** - bridge, replicate, aggregate, and transform messages between KubeMQ clusters.
            - **KubeMQ Sources Connector** - transform external systems and cloud services data to KubeMQ cluster messages.
          \n### Kubernetes Native
          Innovative and modern message queue and message broker in a lightweight container developed to run in Kubernetes, certified in the CNCF landscape, and connect natively to the cloud-native ecosystem.

          ### Ease of use

          Simple deployment in Kubernetes in less than 1 minute. Developer friendly by simple to use SDKs and elimination of the many developers and DevOps-centered challenges to define exchanges, brokers, channels, routes, and predefined topics

          ### All messaging patterns

          KubeMQ is available with all messaging patterns such as Queue, Stream, Pub/Sub, and RPC, Saving the need to maintain multiple messaging platforms for different use cases.

          ### Hybrid Environments

          KubeMQ enables the microservices from multiple environments to communicate and build one hybrid infrastructure solution across clouds, on-prem, and at the edge.

          ### KubeMQ Cluster key features

          - All-batteries included Messaging Queue Broker for the Kubernetes environment
          - Deployed with Operator for full life cycle operation
          - Blazing fast (written in Go), small and lightweight Docker container
          - Asynchronous and Synchronous messaging with support for  `At Most Once Delivery` and `At Least Once Delivery` models
          - Supports durable FIFO based Queue, Publish-Subscribe Events, Publish-Subscribe with Persistence (Events Store), RPC Command and Query messaging patterns
          - Supports gRPC, Rest and WebSocket Transport protocols with TLS support (both RPC and Stream modes)
          - Supports Access control Authorization and Authentication
          - Supports message masticating and smart routing
          - No Message broker configuration needed (i.e., queues, exchanges)
          - .Net, Java, Python, Go, and NodeJS SDKs

          ### KubeMQ Targets Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major services types (databases, cache, messaging, serverless, HTTP, etc.)
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-targets)

          ### KubeMQ Bridges Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Build Any Topology - connects KubeMQ clusters in 1:1, 1:n , n:1, n:n
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-bridges)

          ### KubeMQ Sources Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major messaging services such as Kafka, ActiveMQ, SQS, and more.
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-sources)

          ## Getting Started

          [Start with OpenShift](https://docs.kubemq.io/getting-started/create-cluster/openshift)

          ## Get Enterprise Trial License

          [Register](https://account.kubemq.io/login/register)

          ## Request a Demo

          [See KubeMQ in Action](https://kubemq.io/request-a-demo/)
        displayName: Kubemq Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - messaging
        - queue
        - broker
        - pubsub
        - rpc
        links:
        - name: Learn more about Kubemq
          url: https://kubemq.io/
        - name: Register for free license
          url: https://account.kubemq.io/login/register
        - name: Contact Sales
          url: https://kubemq.io/contact-sales/
        - name: Documentation
          url: https://docs.kubemq.io/
        - name: Get Help
          url: https://kubemq.io/contact-support/
        maintainers:
        - email: support@kubemq.io
          name: kubemq
        maturity: alpha
        provider:
          name: Kubemq.io
        relatedImages:
        - registry.connect.redhat.com/kubemq/kubemq-sources@sha256:5736575b6cae39d129346f83fe2d8ab9cdbd1c1d2a99a66b4b9c01631fb5db21
        - registry.connect.redhat.com/kubemq/kubemq-sources@sha256:8ed531e58c60760888ffb18165883386bda9cc5624645353406c94750f5c3008
        - registry.connect.redhat.com/kubemq/kubemq-operator@sha256:3a9c87e462eb1caa81925dee72407f73fca41f19a3f9bcaf4e76e569db1dfbab
        - registry.connect.redhat.com/kubemq/kubemq-uni@sha256:51154b2a112c5cdc17ee8d338524c02ccd995f9386180b1cc232b7a875a23e73
        - registry.connect.redhat.com/kubemq/kubemq-targets@sha256:2a09e1718a03729a121728e4e3683e7e1f0690789d7f9857d76bebf1cc01d436
        version: 1.8.6
      entries:
      - name: kubemq-operator-marketplace.1.8.6
        version: 1.8.6
      - name: kubemq-operator-marketplace.1.8.5
        version: 1.8.5
      - name: kubemq-operator-marketplace.v1.8.4
        version: 1.8.4
      - name: kubemq-operator-marketplace.v1.8.3
        version: 1.8.3
      name: alpha
    defaultChannel: alpha
    packageName: kubemq-operator-marketplace
    provider:
      name: Kubemq.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: candidate
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: fast
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      - name: stackgres.v1.16.3
        version: 1.16.3
      - name: stackgres.v1.16.2
        version: 1.16.2
      - name: stackgres.v1.16.1
        version: 1.16.1
      - name: stackgres.v1.16.0
        version: 1.16.0
      - name: stackgres.v1.15.2
        version: 1.15.2
      - name: stackgres.v1.15.1
        version: 1.15.1
      - name: stackgres.v1.15.0
        version: 1.15.0
      - name: stackgres.v1.14.3
        version: 1.14.3
      - name: stackgres.v1.14.2
        version: 1.14.2
      - name: stackgres.v1.14.1
        version: 1.14.1
      - name: stackgres.v1.14.0
        version: 1.14.0
      name: stable
    defaultChannel: stable
    packageName: stackgres-certified
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzounstructured-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzounstructured-operator.v2.0.101
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzounstructured.clusters.cambridgesemantics.com/v1",
                "kind": "AnzoUnstructured",
                "metadata": {
                  "name": "au01"
                },
                "spec": {
                  "auWorker": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-w",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-unstructured-worker@sha256:4db61bdf6dc8b42824a06d2f8d8aeecddde765d123b40a63a7bbf87845719fb2",
                                "name": "w",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  },
                  "msLeader": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-ms",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-microservices-leader@sha256:0579fd97e81e436eaff3b58c5395f13aa924940a3d51f0bdea5163d327620286",
                                "name": "ms",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/unstructured-operator@sha256:a47af068a7aa55c12be84c80d263feeed5597ecc434986b4f2fff782b568bb04
          createdAt: "2021-08-02"
          description: Kubernetes Operator for Anzo Unstructured
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzounstructured-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzounstructured-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoUnstructured is the Schema for the anzounstructureds
              API
            displayName: AnzoUnstructured Operator
            kind: AnzoUnstructured
            name: anzounstructureds.anzounstructured.clusters.cambridgesemantics.com
            version: v1
        description: |-
          The Anzo Unstructured Operator provides the way to install and configure an anzo unstructured setup on Red Hat K8S environment.
          Currently, this is possible only through existing Anzo installation.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v2.0.0/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers through [ Cambridge Semantics Customer Center ]( https://customercenter.cambridgesemantics.com/ ).
        displayName: Anzo Unstructured Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Database
        - Big Data
        links:
        - name: Anzo Unstructured Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v2.0.0/README.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/unstructured-operator@sha256:a47af068a7aa55c12be84c80d263feeed5597ecc434986b4f2fff782b568bb04
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-microservices-leader@sha256:0579fd97e81e436eaff3b58c5395f13aa924940a3d51f0bdea5163d327620286
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-unstructured-worker@sha256:4db61bdf6dc8b42824a06d2f8d8aeecddde765d123b40a63a7bbf87845719fb2
        version: 2.0.0
      entries:
      - name: anzounstructured-operator.v2.0.101
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anzounstructured-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://github.com/che-incubator/kubernetes-image-puller-operator
    name: kubernetes-imagepuller-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubernetes-imagepuller-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "che.eclipse.org/v1alpha1",
                "kind": "KubernetesImagePuller",
                "metadata": {
                  "name": "image-puller"
                },
                "spec": {
                  "daemonsetName": "k8s-image-puller",
                  "images": "java11-maven=quay.io/eclipse/che-java11-maven:next;che-theia=quay.io/eclipse/che-theia:next"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/eclipse/kubernetes-image-puller-operator:1.1.0
          createdAt: "2024-01-15T16:23:04Z"
          description: Create and manage kubernetes-image-puller instances.
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/che-incubator/kubernetes-image-puller-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubernetesImagePuller is the Schema for the kubernetesimagepullers
              API
            displayName: Kubernetes Image Puller
            kind: KubernetesImagePuller
            name: kubernetesimagepullers.che.eclipse.org
            version: v1alpha1
        description: "## About the managed application\n\nCreate a `KubernetesImagePuller`
          custom resource to automatically configure and run an instance of the [kubernetes-image-puller.](https://github.com/che-incubator/kubernetes-image-puller)\n##
          About this Operator\n\nThe `KubernetesImagePuller` custom resource understands
          the following fields in the `spec`:\n\n1. `configMapName` - The name of
          the `ConfigMap` to create\t\n2. `daemonsetName` - The name of the `DaemonSet`
          to be create.\n3. `deploymentName` - The name of the `kubernetes-image-puller`
          `Deployment` to create.\n4. `images` - A list of key-value pairs separated
          by semicolons of images to pull.  For example: `java=quay.io/eclipse/che-java8-maven:latest;theia=quay.io/eclipse/che-theia:next`\n5.
          `cachingIntervalHours` - The amount of time, in hours, between `DaemonSet`
          health checks.\n6. `cachingMemoryRequest` - The memory request for each
          cached image when the puller is running.\n7. `cachingMemoryLimit` - The
          memory limit for each cached image when the puller is running.\n8. `nodeSelector`
          - Node selector applied to pods created by the `DaemonSet`.\n## Prerequisites
          for enabling this Operator\n\nThe operator requires an existing namespace
          to be installed in.\nCheck the [README](https://github.com/che-incubator/kubernetes-image-puller-operator#installing-the-operator)
          for more details.\n"
        displayName: Kubernetes Image Puller Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - che
        - kubernetes-image-puller
        links:
        - name: Repository
          url: https://github.com/che-incubator/kubernetes-image-puller-operator
        maintainers:
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        provider:
          name: Red Hat, Inc.
          url: https://github.com/che-incubator/kubernetes-image-puller-operator
        relatedImages:
        - quay.io/eclipse/kubernetes-image-puller-operator:1.1.0
        - quay.io/brancz/kube-rbac-proxy:v0.11.0
        version: 1.1.0
      entries:
      - name: kubernetes-imagepuller-operator.v1.1.0
        version: 1.1.0
      - name: kubernetes-imagepuller-operator.v1.0.6
        version: 1.0.6
      - name: kubernetes-imagepuller-operator.v1.0.5
        version: 1.0.5
      - name: kubernetes-imagepuller-operator.v1.0.4
        version: 1.0.4
      - name: kubernetes-imagepuller-operator.v1.0.3
        version: 1.0.3
      - name: kubernetes-imagepuller-operator.v1.0.2
        version: 1.0.2
      - name: kubernetes-imagepuller-operator.v1.0.1
        version: 1.0.1
      - name: kubernetes-imagepuller-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: kubernetes-imagepuller-operator
    provider:
      name: Red Hat, Inc.
      url: https://github.com/che-incubator/kubernetes-image-puller-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Palo Alto Networks
      provider-url: ""
    name: pcc-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pcc-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "name": "pcc-console",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "ConsoleDefender",
                "metadata": {
                  "name": "pcc-consoledefender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "defenderConfig": {
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Defender",
                "metadata": {
                  "name": "pcc-defender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "defenderConfig": {
                    "clusterAddress": "twistlock-console.example.com",
                    "consoleAddress": "https://twistlock-console.example.com:8083",
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: quay.io/prismacloud/pcc-operator@sha256:b8fcfbd6c51286c874e00db1bd35523386cec406fa4050ef44c0a887730cf9b8
          createdAt: "2021-11-15"
          description: Deploy Prisma Cloud Compute for cloud-native security in your
            clusters
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator
          support: pcc-integrations@paloaltonetworks.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Install the Prisma Cloud Compute Console and Defenders
            displayName: Console and Defenders
            kind: ConsoleDefender
            name: consoledefenders.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install the Prisma Cloud Compute Console
            displayName: Console
            kind: Console
            name: consoles.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install Prisma Cloud Compute Defenders
            displayName: Defenders
            kind: Defender
            name: defenders.pcc.paloaltonetworks.com
            version: v1alpha1
        description: |
          ## Features
          - Deploy Console
              - Create initial local administrator user
              - Add license
          - Deploy Defenders
          - Support for offline deployments
          - Upgrade Console
          - Upgrade Defenders

          ## Installation
          See the [Prisma Cloud Compute Operator documentation](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator) for installation guides.

          ## Support
          This operator is available "as is," and should be seen as community-supported; however, Palo Alto Networks will continue contributing whenever possible.
          Palo Alto Networks does not provide technical support or help with using or troubleshooting the operator through normal support channels.
          We encourage you to open GitHub [issues](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/issues) and [pull requests](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/pulls) to track bugs and feature requests.

          ## Other links
          [Prisma Cloud Compute product documentation](https://docs.paloaltonetworks.com/prisma/prisma-cloud/prisma-cloud-admin-compute.html)

          [Prisma Cloud Compute API documentation](https://prisma.pan.dev/api/cloud/cwpp)
        displayName: Prisma Cloud Compute Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - palo
        - palo alto
        - palo alto networks
        - paloalto
        - twistlock
        - compute
        - prisma
        - prisma cloud
        - prisma cloud compute
        - security
        - image
        - images
        - container
        - containers
        - docker
        - podman
        - vulnerability
        - vulnerabilities
        - compliance
        - runtime
        - scan
        - scanning
        - monitor
        - monitoring
        - alert
        - alerting
        links:
        - name: Prisma Cloud
          url: https://www.paloaltonetworks.com/prisma/cloud
        maintainers:
        - email: pcc-integrations@paloaltonetworks.com
          name: Wyatt Gill
        maturity: alpha
        provider:
          name: Palo Alto Networks
        relatedImages:
        - quay.io/prismacloud/pcc-operator@sha256:b8fcfbd6c51286c874e00db1bd35523386cec406fa4050ef44c0a887730cf9b8
        version: 0.2.0
      entries:
      - name: pcc-operator.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: pcc-operator
    provider:
      name: Palo Alto Networks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: koku-metrics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: koku-metrics-operator.v4.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
          createdAt: "2025-09-25T19:04:35Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: koku-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Cost Management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Koku Metrics Operator
          ## Introduction
          The `koku-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Koku Metrics Operator (`koku-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created KokuMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The KokuMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.2.0:
          * Enabled gathering and reporting NVIDIA GPU metrics for resource optimization.
          * (Bugfix) Fixed custom CA certificate validation and configuration issues.
          * (Bugfix) Fixed issue where ROS queries were missing workload metadata for namespaces labeled only with the legacy `insights_cost_management_optimizations` label.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **Progress towards FIPS 140 Compliance:** The operator is using the Go Cryptographic Module v1.0.0 to progress toward FIPS 140 compliance. Although this module is not validated at the time of this release, the operator aims to meet stricter security standards when the module does successfully achieve FIPS validation.
          * **API Name Change:** The Custom Resource Definition (CRD) for configuring the upstream `Koku Metrics Operator` was renamed from `KokuMetricsConfig` to `CostManagementMetricsConfig`.

          ### NOTE:
            * The API name change impacts only users of the upstream (community) Koku Metrics Operator.
            * The downstream (Red Hat-supported) Cost Management Metrics Operator is not impacted by this change and users should continue using their existing configurations.

          **Important Upgrade Instructions:**

            If you are upgrading the upstream Koku Metrics Operator to version 4.0.0 or higher, you must manually migrate your configuration. The operator will no longer recognize existing `KokuMetricsConfig` resources.

            To successfully upgrade and retain your operator's configuration, complete the following steps. You can use the provided `oc` commands replacing the angle brackets (`< >`) with your specific values:

            1. Backup existing configuration (recommended):

               Before you upgrade, retrieve the details of your current `KokuMetricsConfig` instance to help you recreate it with the new API name.

               ```
               oc get kokumetricsconfig -n koku-metrics-operator <your-config-name> -o yaml > koku-metrics-config-backup.yaml
               ```

            2. Delete the previous `KokuMetricsConfig` instance (recommended):

                After the operator is upgraded to `version 4.0.0`, delete any existing `KokuMetricsConfig` resources.

                ```
                oc delete kokumetricsconfig -n koku-metrics-operator <your-config-name>
                ```

            3. Create the new `CostManagementMetricsConfig` instance:

                Using the information from your backup (or a new configuration), create a new `CostManagementMetricsConfig` instance. Adjust the spec section based on your specific needs. For more information about configurations, [refer to the configurable parameters](#configurable-parameters).

            4. Apply your new configuration:

                ```
                oc apply -f <your-costmanagementmetricsconfig-file>.yaml -n koku-metrics-operator
                ```

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created KokuMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the KokuMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `KokuMetricsConfig`, creating a new `KokuMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `koku-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: koku-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the KokuMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the KokuMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon KokuMetricsConfig creation** (This parameter will start to appear in KokuMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the KokuMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the koku-metrics-operator
          **Note** There are separate instructions for configuring the `koku-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `koku-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the koku-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `koku-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `koku-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the koku-metrics-operator for a restricted network
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `koku-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: koku-metrics-operator
                labels:
                  app: koku-metrics-operator
              spec:
                volumes:
                - name: koku-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: koku-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: koku-metrics-operator-reports
                    mountPath: /tmp/koku-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/koku-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/koku-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `koku-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the KokuMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Koku Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Koku Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
        version: 4.2.0
      entries:
      - name: koku-metrics-operator.v4.2.0
        version: 4.2.0
      - name: koku-metrics-operator.v4.1.0
        version: 4.1.0
      - name: koku-metrics-operator.v4.0.0
        version: 4.0.0
      - name: koku-metrics-operator.v3.3.2
        version: 3.3.2
      - name: koku-metrics-operator.v3.3.1
        version: 3.3.1
      - name: koku-metrics-operator.v3.3.0
        version: 3.3.0
      - name: koku-metrics-operator.v3.2.1
        version: 3.2.1
      - name: koku-metrics-operator.v3.2.0
        version: 3.2.0
      - name: koku-metrics-operator.v3.1.0
        version: 3.1.0
      - name: koku-metrics-operator.v3.0.1
        version: 3.0.1
      - name: koku-metrics-operator.v3.0.0
        version: 3.0.0
      - name: koku-metrics-operator.v2.0.0
        version: 2.0.0
      - name: koku-metrics-operator.v1.2.0
        version: 1.2.0
      - name: koku-metrics-operator.v1.1.9
        version: 1.1.9
      - name: koku-metrics-operator.v1.1.8
        version: 1.1.8
      - name: koku-metrics-operator.v1.1.7
        version: 1.1.7
      - name: koku-metrics-operator.v1.1.6
        version: 1.1.6
      - name: koku-metrics-operator.v1.1.5
        version: 1.1.5
      - name: koku-metrics-operator.v1.1.4
        version: 1.1.4
      - name: koku-metrics-operator.v1.1.3
        version: 1.1.3
      - name: koku-metrics-operator.v1.1.2
        version: 1.1.2
      - name: koku-metrics-operator.v1.1.1
        version: 1.1.1
      - name: koku-metrics-operator.v0.9.8
        version: 0.9.8
      - name: koku-metrics-operator.v0.9.7
        version: 0.9.7
      - name: koku-metrics-operator.v0.9.6
        version: 0.9.6
      - name: koku-metrics-operator.v0.9.5
        version: 0.9.5
      - name: koku-metrics-operator.v0.9.4
        version: 0.9.4
      - name: koku-metrics-operator.v0.9.3
        version: 0.9.3
      - name: koku-metrics-operator.v0.9.2
        version: 0.9.2
      - name: koku-metrics-operator.v0.9.1
        version: 0.9.1
      - name: koku-metrics-operator.v0.9.0
        version: 0.9.0
      name: alpha
    - currentCSV: koku-metrics-operator.v4.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
          createdAt: "2025-09-25T19:04:35Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: koku-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Cost Management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Koku Metrics Operator
          ## Introduction
          The `koku-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Koku Metrics Operator (`koku-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created KokuMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The KokuMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.2.0:
          * Enabled gathering and reporting NVIDIA GPU metrics for resource optimization.
          * (Bugfix) Fixed custom CA certificate validation and configuration issues.
          * (Bugfix) Fixed issue where ROS queries were missing workload metadata for namespaces labeled only with the legacy `insights_cost_management_optimizations` label.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **Progress towards FIPS 140 Compliance:** The operator is using the Go Cryptographic Module v1.0.0 to progress toward FIPS 140 compliance. Although this module is not validated at the time of this release, the operator aims to meet stricter security standards when the module does successfully achieve FIPS validation.
          * **API Name Change:** The Custom Resource Definition (CRD) for configuring the upstream `Koku Metrics Operator` was renamed from `KokuMetricsConfig` to `CostManagementMetricsConfig`.

          ### NOTE:
            * The API name change impacts only users of the upstream (community) Koku Metrics Operator.
            * The downstream (Red Hat-supported) Cost Management Metrics Operator is not impacted by this change and users should continue using their existing configurations.

          **Important Upgrade Instructions:**

            If you are upgrading the upstream Koku Metrics Operator to version 4.0.0 or higher, you must manually migrate your configuration. The operator will no longer recognize existing `KokuMetricsConfig` resources.

            To successfully upgrade and retain your operator's configuration, complete the following steps. You can use the provided `oc` commands replacing the angle brackets (`< >`) with your specific values:

            1. Backup existing configuration (recommended):

               Before you upgrade, retrieve the details of your current `KokuMetricsConfig` instance to help you recreate it with the new API name.

               ```
               oc get kokumetricsconfig -n koku-metrics-operator <your-config-name> -o yaml > koku-metrics-config-backup.yaml
               ```

            2. Delete the previous `KokuMetricsConfig` instance (recommended):

                After the operator is upgraded to `version 4.0.0`, delete any existing `KokuMetricsConfig` resources.

                ```
                oc delete kokumetricsconfig -n koku-metrics-operator <your-config-name>
                ```

            3. Create the new `CostManagementMetricsConfig` instance:

                Using the information from your backup (or a new configuration), create a new `CostManagementMetricsConfig` instance. Adjust the spec section based on your specific needs. For more information about configurations, [refer to the configurable parameters](#configurable-parameters).

            4. Apply your new configuration:

                ```
                oc apply -f <your-costmanagementmetricsconfig-file>.yaml -n koku-metrics-operator
                ```

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created KokuMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the KokuMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `KokuMetricsConfig`, creating a new `KokuMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `koku-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: koku-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the KokuMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the KokuMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon KokuMetricsConfig creation** (This parameter will start to appear in KokuMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the KokuMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the koku-metrics-operator
          **Note** There are separate instructions for configuring the `koku-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `koku-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the koku-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `koku-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `koku-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the koku-metrics-operator for a restricted network
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `koku-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: koku-metrics-operator
                labels:
                  app: koku-metrics-operator
              spec:
                volumes:
                - name: koku-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: koku-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: koku-metrics-operator-reports
                    mountPath: /tmp/koku-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/koku-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/koku-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `koku-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the KokuMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Koku Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Koku Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
        version: 4.2.0
      entries:
      - name: koku-metrics-operator.v4.2.0
        version: 4.2.0
      - name: koku-metrics-operator.v4.1.0
        version: 4.1.0
      - name: koku-metrics-operator.v4.0.0
        version: 4.0.0
      - name: koku-metrics-operator.v3.3.2
        version: 3.3.2
      - name: koku-metrics-operator.v3.3.1
        version: 3.3.1
      - name: koku-metrics-operator.v3.3.0
        version: 3.3.0
      - name: koku-metrics-operator.v3.2.1
        version: 3.2.1
      - name: koku-metrics-operator.v3.2.0
        version: 3.2.0
      - name: koku-metrics-operator.v3.1.0
        version: 3.1.0
      - name: koku-metrics-operator.v3.0.1
        version: 3.0.1
      - name: koku-metrics-operator.v3.0.0
        version: 3.0.0
      - name: koku-metrics-operator.v2.0.0
        version: 2.0.0
      - name: koku-metrics-operator.v1.2.0
        version: 1.2.0
      - name: koku-metrics-operator.v1.1.9
        version: 1.1.9
      - name: koku-metrics-operator.v1.1.8
        version: 1.1.8
      - name: koku-metrics-operator.v1.1.7
        version: 1.1.7
      - name: koku-metrics-operator.v1.1.6
        version: 1.1.6
      - name: koku-metrics-operator.v1.1.5
        version: 1.1.5
      - name: koku-metrics-operator.v1.1.4
        version: 1.1.4
      - name: koku-metrics-operator.v1.1.3
        version: 1.1.3
      - name: koku-metrics-operator.v1.1.2
        version: 1.1.2
      - name: koku-metrics-operator.v1.1.1
        version: 1.1.1
      - name: koku-metrics-operator.v0.9.8
        version: 0.9.8
      - name: koku-metrics-operator.v0.9.7
        version: 0.9.7
      - name: koku-metrics-operator.v0.9.6
        version: 0.9.6
      - name: koku-metrics-operator.v0.9.5
        version: 0.9.5
      - name: koku-metrics-operator.v0.9.4
        version: 0.9.4
      - name: koku-metrics-operator.v0.9.3
        version: 0.9.3
      - name: koku-metrics-operator.v0.9.2
        version: 0.9.2
      - name: koku-metrics-operator.v0.9.1
        version: 0.9.1
      - name: koku-metrics-operator.v0.9.0
        version: 0.9.0
      name: beta
    defaultChannel: beta
    packageName: koku-metrics-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Lenovo (Shanghai) Computer Technology Co., Ltd.
      provider-url: ""
    name: lenovo-locd-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lenovo-locd-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "locd.lenovo.com/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample",
                  "namespace": "locd-operator-system"
                },
                "spec": {
                  "affinity": {},
                  "argocd": {
                    "password": "xxxxxxxxxxx",
                    "server": "argocd.svc.cluster.local",
                    "user": "admin"
                  },
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "core": {
                    "image": "registry.connect.redhat.com/loc-devops/core@sha256:f771217b3f7849e1b63cc6abb93b8b0fc9636fd4a201e31807a15f773138c609",
                    "pullPolicy": "Always"
                  },
                  "fullnameOverride": "",
                  "global": {
                    "managementNamespace": "loc-devops",
                    "serviceEndpoint": "loc-devops.domain",
                    "serviceaccount": {
                      "annotations": {
                        "managedBy": "LOC DevOps"
                      },
                      "create": true,
                      "name": "loc-devops-admin"
                    }
                  },
                  "gui": {
                    "image": "registry.connect.redhat.com/loc-devops/gui@sha256:65142a36095188e879467038936b7b6b6ee39dcea6d42a437fa7066fb2bc0f44",
                    "pullPolicy": "Always"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "className": "nginx",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "loc-devops.domain",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "Prefix",
                            "servicePort": 8080
                          },
                          {
                            "path": "/api",
                            "pathType": "Prefix",
                            "servicePort": 8000
                          },
                          {
                            "path": "/swagger",
                            "pathType": "Prefix",
                            "servicePort": 8000
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "replicaCount": 1,
                  "resources": {},
                  "securityContext": {},
                  "service": {
                    "port": {
                      "core": 8000,
                      "gui": 8080
                    },
                    "type": "ClusterIP"
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Cluster
            name: clusters.locd.lenovo.com
            version: v1alpha1
        description: This is operator used to manage components of Lenovo Open Cloud
          DevOps
        displayName: lenovo-locd-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Application management
        links:
        - name: LOCD Operator
          url: https://locd-operator.domain
        maintainers:
        - email: xujt5@lenovo.com
          name: jintaoxu
        maturity: alpha
        provider:
          name: Lenovo (Shanghai) Computer Technology Co., Ltd.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b5786bbbef725badf3dfcc2c2c7a86ead5ebb584c978c47aae8b9a62e241b80d
        - registry.connect.redhat.com/loc-devops/operator@sha256:4c17a3688ae3098e4b5c2c0530f49b1da8dedcc6fc1d3cd848513a1f189812a7
        - registry.connect.redhat.com/loc-devops/core@sha256:f771217b3f7849e1b63cc6abb93b8b0fc9636fd4a201e31807a15f773138c609
        - registry.connect.redhat.com/loc-devops/gui@sha256:65142a36095188e879467038936b7b6b6ee39dcea6d42a437fa7066fb2bc0f44
        version: 0.1.0
      entries:
      - name: lenovo-locd-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: lenovo-locd-operator
    provider:
      name: Lenovo (Shanghai) Computer Technology Co., Ltd.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtr-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtr-operator.1.2.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
             {
               "apiVersion": "windup.jboss.org/v1",
               "kind": "Windup",
               "metadata": {
                 "namespace": "mtr",
                 "name": "mtr",
                 "labels": {
                   "application": "windup"
                 }
               },
               "spec": {}
             }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mtr/mtr-rhel8-operator@sha256:4df7c4ad353339e8b958ffc651e811667767c8a0dcdd61f89352c926da52a513
          createdAt: "2024-09-03T17:22:23Z"
          description: Migration Toolkit for Runtimes is an analysis tool that supports
            the modernization and migration of Java applications.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          imagePullPolicy: Always
          operatorframework.io/suggested-namespace: mtr
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/windup/windup-operator/
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "Create an MTR Web Console application.  \n\nTo access the
              MTR Web Console application from the \"Developer\" perspective, please
              go to Topology and click on the \"Open URL\" icon on the web-console
              pod.  \n\nAlternatively from the \"Administrator\" perspective, please
              go to Networking->Routes and click on the Location hyperlink.  \n\nIn
              order to connect with default login credentials, please use \"migration\"
              as username and \"password\" as password.  \n\nKnown issue  \nIf you
              want to customize the MTR Web Console instance's parameters and you
              can not see any in the `Form view`, please switch to the `YAML view`
              and change them as needed.\n"
            displayName: Migration Toolkit for Runtimes
            kind: Windup
            name: windups.windup.jboss.org
            version: v1
        description: "The Migration Toolkit for Runtimes (MTR) is a web console application
          that supports large-scale Java application modernization and migration projects
          across a broad range of transformations and use cases.  \n\nIt analyzes
          application code, supports effort estimation, accelerates code migration,
          and enables users to move applications to containers.  \n\nFor more information
          please refer to the [https://developers.redhat.com/products/mtr/overview](https://developers.redhat.com/products/mtr/overview)
          page.\n"
        displayName: Migration Toolkit for Runtimes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - security
        - alerting
        - metric
        - troubleshooting
        - run-time
        - migration
        - modernization
        links:
        - name: Website
          url: https://developers.redhat.com/products/mtr/overview
        - name: GitHub
          url: https://github.com/windup/windup-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_runtimes
        maintainers:
        - email: migrate@redhat.com
          name: Migration Toolkit for Runtimes
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mtr/mtr-rhel8-operator@sha256:4df7c4ad353339e8b958ffc651e811667767c8a0dcdd61f89352c926da52a513
        - registry.redhat.io/mtr/mtr-web-container-rhel8@sha256:cb88a5453b506ca70563537ec6db2f44ff8cb4050c9652ec86eb5128aa313003
        - registry.redhat.io/mtr/mtr-web-executor-container-rhel8@sha256:2f0622774367f33efc321926196681322ab1535144e6a0785608c8050fc59f2e
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        version: 1.2.7
      entries:
      - name: mtr-operator.1.2.7
        version: 1.2.7
      - name: mtr-operator.1.2.6
        version: 1.2.6
      - name: mtr-operator.1.2.5
        version: 1.2.5
      - name: mtr-operator.1.2.4
        version: 1.2.4
      - name: mtr-operator.1.2.3
        version: 1.2.3
      - name: mtr-operator.1.2.2
        version: 1.2.2
      - name: mtr-operator.1.2.1
        version: 1.2.1
      - name: mtr-operator.1.2.0
        version: 1.2.0
      - name: mtr-operator.1.1.1
        version: 1.1.1
      - name: mtr-operator.1.1.0
        version: 1.1.0
      - name: mtr-operator.1.0.3
        version: 1.0.3
      - name: mtr-operator.1.0.2
        version: 1.0.2
      - name: mtr-operator.1.0.1
        version: 1.0.1
      - name: mtr-operator.1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: mtr-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: GitLab Inc
      provider-url: https://about.gitlab.com/
    name: gitlab-operator-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "certmanager": {
                        "install": false
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-26T07:16:35Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3281fe12aaae428fd7102332e7c0d7033515042d7830eba3953891ab0c3e3a29
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      - name: gitlab-operator-kubernetes.v2.1.1
        version: 2.1.1
      - name: gitlab-operator-kubernetes.v2.1.0
        version: 2.1.0
      - name: gitlab-operator-kubernetes.v2.0.2
        version: 2.0.2
      - name: gitlab-operator-kubernetes.v2.0.1
        version: 2.0.1
      - name: gitlab-operator-kubernetes.v2.0.0
        version: 2.0.0
      - name: gitlab-operator-kubernetes.v1.12.2
        version: 1.12.2
      - name: gitlab-operator-kubernetes.v1.12.1
        version: 1.12.1
      - name: gitlab-operator-kubernetes.v1.12.0
        version: 1.12.0
      - name: gitlab-operator-kubernetes.v1.11.3
        version: 1.11.3
      - name: gitlab-operator-kubernetes.v1.11.2
        version: 1.11.2
      - name: gitlab-operator-kubernetes.v1.11.1
        version: 1.11.1
      - name: gitlab-operator-kubernetes.v1.11.0
        version: 1.11.0
      - name: gitlab-operator-kubernetes.v1.10.2
        version: 1.10.2
      - name: gitlab-operator-kubernetes.v1.10.1
        version: 1.10.1
      - name: gitlab-operator-kubernetes.v1.10.0
        version: 1.10.0
      - name: gitlab-operator-kubernetes.v1.9.2
        version: 1.9.2
      - name: gitlab-operator-kubernetes.v1.9.1
        version: 1.9.1
      - name: gitlab-operator-kubernetes.v1.9.0
        version: 1.9.0
      - name: gitlab-operator-kubernetes.v1.8.3
        version: 1.8.3
      - name: gitlab-operator-kubernetes.v1.8.2
        version: 1.8.2
      - name: gitlab-operator-kubernetes.v1.8.1
        version: 1.8.1
      - name: gitlab-operator-kubernetes.v1.8.0
        version: 1.8.0
      - name: gitlab-operator-kubernetes.v1.7.2
        version: 1.7.2
      - name: gitlab-operator-kubernetes.v1.7.1
        version: 1.7.1
      - name: gitlab-operator-kubernetes.v1.7.0
        version: 1.7.0
      - name: gitlab-operator-kubernetes.v1.6.1
        version: 1.6.1
      - name: gitlab-operator-kubernetes.v1.6.0
        version: 1.6.0
      - name: gitlab-operator-kubernetes.v1.5.1
        version: 1.5.1
      - name: gitlab-operator-kubernetes.v1.4.2
        version: 1.4.2
      - name: gitlab-operator-kubernetes.v1.4.1
        version: 1.4.1
      - name: gitlab-operator-kubernetes.v1.4.0
        version: 1.4.0
      - name: gitlab-operator-kubernetes.v1.3.3
        version: 1.3.3
      - name: gitlab-operator-kubernetes.v1.3.2
        version: 1.3.2
      - name: gitlab-operator-kubernetes.v1.3.1
        version: 1.3.1
      - name: gitlab-operator-kubernetes.v1.3.0
        version: 1.3.0
      name: stable
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "certmanager": {
                        "install": false
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-26T07:16:35Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3281fe12aaae428fd7102332e7c0d7033515042d7830eba3953891ab0c3e3a29
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      - name: gitlab-operator-kubernetes.v2.1.1
        version: 2.1.1
      - name: gitlab-operator-kubernetes.v2.1.0
        version: 2.1.0
      - name: gitlab-operator-kubernetes.v2.0.2
        version: 2.0.2
      - name: gitlab-operator-kubernetes.v2.0.1
        version: 2.0.1
      - name: gitlab-operator-kubernetes.v2.0.0
        version: 2.0.0
      - name: gitlab-operator-kubernetes.v1.12.2
        version: 1.12.2
      - name: gitlab-operator-kubernetes.v1.12.1
        version: 1.12.1
      - name: gitlab-operator-kubernetes.v1.12.0
        version: 1.12.0
      - name: gitlab-operator-kubernetes.v1.11.3
        version: 1.11.3
      - name: gitlab-operator-kubernetes.v1.11.2
        version: 1.11.2
      - name: gitlab-operator-kubernetes.v1.11.1
        version: 1.11.1
      - name: gitlab-operator-kubernetes.v1.11.0
        version: 1.11.0
      - name: gitlab-operator-kubernetes.v1.10.2
        version: 1.10.2
      - name: gitlab-operator-kubernetes.v1.10.1
        version: 1.10.1
      - name: gitlab-operator-kubernetes.v1.10.0
        version: 1.10.0
      - name: gitlab-operator-kubernetes.v1.9.2
        version: 1.9.2
      - name: gitlab-operator-kubernetes.v1.9.1
        version: 1.9.1
      - name: gitlab-operator-kubernetes.v1.9.0
        version: 1.9.0
      - name: gitlab-operator-kubernetes.v1.8.3
        version: 1.8.3
      - name: gitlab-operator-kubernetes.v1.8.2
        version: 1.8.2
      - name: gitlab-operator-kubernetes.v1.8.1
        version: 1.8.1
      - name: gitlab-operator-kubernetes.v1.8.0
        version: 1.8.0
      - name: gitlab-operator-kubernetes.v1.7.2
        version: 1.7.2
      - name: gitlab-operator-kubernetes.v1.7.1
        version: 1.7.1
      - name: gitlab-operator-kubernetes.v1.7.0
        version: 1.7.0
      - name: gitlab-operator-kubernetes.v1.6.1
        version: 1.6.1
      - name: gitlab-operator-kubernetes.v1.6.0
        version: 1.6.0
      - name: gitlab-operator-kubernetes.v1.5.1
        version: 1.5.1
      - name: gitlab-operator-kubernetes.v1.4.2
        version: 1.4.2
      - name: gitlab-operator-kubernetes.v1.4.1
        version: 1.4.1
      - name: gitlab-operator-kubernetes.v1.4.0
        version: 1.4.0
      - name: gitlab-operator-kubernetes.v1.3.3
        version: 1.3.3
      - name: gitlab-operator-kubernetes.v1.3.2
        version: 1.3.2
      - name: gitlab-operator-kubernetes.v1.3.1
        version: 1.3.1
      - name: gitlab-operator-kubernetes.v1.3.0
        version: 1.3.0
      name: unstable
    defaultChannel: stable
    packageName: gitlab-operator-kubernetes
    provider:
      name: GitLab Inc
      url: https://about.gitlab.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Silicom Ltd.
      provider-url: http://www.silicom-usa.com
    name: silicom-sts-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: silicom-sts-operator.v0.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "namespace": "openshift-operators",
                  "nodeSelector": {
                    "sts.silicom.com/config": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1"
                },
                "spec": {
                  "aprLevel": 0,
                  "domainNumber": 24,
                  "esmcMode": 1,
                  "forwardable": 0,
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-TBC.8275.1",
                  "mode10MHz": 2,
                  "modePPS": 2,
                  "namespace": "openshift-operators",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  },
                  "phyLedsCtl": 0,
                  "priority2": 128,
                  "src10MHz": 1,
                  "srcPPS": 1,
                  "syncOption": 1,
                  "synceCpu": 10,
                  "synceRecClkPort": 0,
                  "tracePtpMsg": -1,
                  "twoStep": 0
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "worker2"
                },
                "spec": null,
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "images": null,
                  "sro": null
                }
              }
            ]
          capabilities: Basic Install
          categories: Drivers and plugins
          certified: "false"
          containerImage: quay.io/silicom/sts-operator:0.0.6
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicomdk/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicomDK/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/silicomDK/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - quay.io/silicom/tsync_extts@sha256:be6a5950a5fdc6fdf8015767901856bf173d01bdc803be7141e4f9795ab1478a
        - quay.io/silicom/phc2sys@sha256:0f3b63a3059b7d6158a916814281097f0d418b99a5fac9e8e5fd9bae58fb0b74
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/silicom/sts-operator:0.0.6
        - quay.io/silicom/tsyncd@sha256:78be78ec8939f025f9a8b899693ab77b3990116d1c7ffb2115c620b7c3382d9d
        - quay.io/silicom/gpsd@sha256:649e8a6c524920e1c564d8f797b8d64235b2527aa923f94bc9c7a0b7aba8ae7f
        - quay.io/silicom/grpc-tsyncd@sha256:aac7233423439e352d8eb39cef63e479c30c2f80b063c84ffc8797cc5a045117
        - quay.io/silicom/sts-plugin@sha256:44339fd400d46d9dddb02c2da47c6f9f96ad2605a5c6de8a7dce2f8c3da3bb57
        - quay.io/silicom/ice-driver-src@sha256:ca728f648769d31501cdedda92554bcfa890754405b0b06d41629cea47707da8
        version: 0.0.6
      entries:
      - name: silicom-sts-operator.v0.0.6
        version: 0.0.6
      - name: silicom-sts-operator.v0.0.5
        version: 0.0.5
      - name: silicom-sts-operator.v0.0.4
        version: 0.0.4
      - name: silicom-sts-operator.v0.0.3
        version: 0.0.3
      - name: silicom-sts-operator.v0.0.2
        version: 0.0.2
      - name: silicom-sts-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: silicom-sts-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "sts-silicom"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "namespace": "sts-silicom",
                  "nodeSelector": {
                    "sts.silicom.com/config": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1"
                },
                "spec": {
                  "aprLevel": 0,
                  "domainNumber": 24,
                  "esmcMode": 1,
                  "forwardable": 0,
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-TBC.8275.1",
                  "mode10MHz": 2,
                  "modePPS": 2,
                  "namespace": "sts-silicom",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  },
                  "phyLedsCtl": 0,
                  "priority2": 128,
                  "src10MHz": 1,
                  "srcPPS": 1,
                  "syncOption": 1,
                  "synceCpu": 10,
                  "synceRecClkPort": 0,
                  "tracePtpMsg": -1,
                  "twoStep": 0
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "worker2"
                },
                "spec": null,
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config"
                },
                "spec": {
                  "gpsSvcPort": 2947,
                  "grpcSvcPort": 50051,
                  "images": {
                    "gpsd": "quay.io/silicom/gpsd:3.23.1",
                    "grpcTsyncd": "quay.io/silicom/grpc-tsyncd:2.1.0.0",
                    "phcs2Sys": "quay.io/silicom/phcs2sys:3.1.1",
                    "stsPlugin": "quay.io/silicom/sts-plugin:0.0.2",
                    "ts2Phcs": "quay.io/silicom/ts2phcs:1.0.0",
                    "tsyncd": "quay.io/silicom/tsyncd:2.1.0.0"
                  },
                  "namespace": "sts-silicom",
                  "sro": {
                    "chart": {
                      "name": "ice-special-resource",
                      "repository": {
                        "name": "ice-special-resource",
                        "url": "http://ice-driver-src"
                      },
                      "version": "0.0.1"
                    },
                    "iceVersion": "1.7.16",
                    "namespace": "openshift-operators",
                    "srcImage": "quay.io/silicom/ice-driver-src:1.7.16",
                    "srcSvcPort": 3000
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Drivers and plugins
          certified: "false"
          containerImage: quay.io/silicom/sts-operator:0.0.2
          createdAt: 2021-11-17T10:21:24 CET
          description: This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of Silicom STS device
            plugins for Kubernetes.
          operators.operatorframework.io.bundle.channels.v1: stable,development
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicomdk/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicomDK/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/silicomDK/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/silicom/sts-operator:0.0.2
        version: 0.0.2
      entries:
      - name: silicom-sts-operator.v0.0.2
        version: 0.0.2
      - name: silicom-sts-operator.v0.0.1
        version: 0.0.1
      name: beta
    defaultChannel: alpha
    packageName: silicom-sts-operator
    provider:
      name: Silicom Ltd.
      url: http://www.silicom-usa.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-logging
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-logging.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Logging",
                "metadata": {
                  "name": "logging-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "elasticsearch": {
                    "storage": "16Gi",
                    "tag": "7.17.6-3747-35"
                  },
                  "environment": "openshift",
                  "fluentbit": {
                    "runOnMaster": false,
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "grafana": {
                    "dashboards": [
                      "wap",
                      "load-core"
                    ],
                    "dataSources": [
                      "wap",
                      "load-core"
                    ]
                  },
                  "image": {
                    "elasticInitImage": "rhel8/toolbox@sha256:a753c24cb8b03066d494f010ad654d23fb0559f14d7121542298da3e0d48340d",
                    "elasticInitRepository": "registry.access.redhat.com",
                    "fbImage": "fluent-bit:1.8.15-3747-20",
                    "fbRepository": "prod.harbor.keysight.digital/loadcore",
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "createClusterRoles": true,
                    "serviceAccountName": "load-core-service-account"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:33:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Logging
            name: loggings.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Logging Operator
        displayName: Keysight WAP Logging Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Logging
          url: https://keysight-wap-logging.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-logging@sha256:f1e3b3ad84cd0f96c92cdaf4609e29257035cb982a761e832dae2095e636c721
        version: 0.1.21
      entries:
      - name: keysight-wap-logging.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-logging
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hazelcast, Inc
      provider-url: ""
    name: hazelcast-platform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hazelcast-platform-operator.v5.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "cache-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "CronHotBackup",
                "metadata": {
                  "name": "cronhotbackup-sample"
                },
                "spec": {
                  "hotBackupTemplate": {
                    "metadata": {
                      "labels": {
                        "cron-hotbackup": "true"
                      }
                    },
                    "spec": {
                      "hazelcastResourceName": "hazelcast"
                    }
                  },
                  "schedule": "*/30 * * * *"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Flow",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hazelcast-platform-operator"
                  },
                  "name": "flow"
                },
                "spec": {
                  "database": {
                    "host": "flow-postgresql",
                    "secretName": "flow-db-secret"
                  },
                  "env": [
                    {
                      "name": "OPTIONS",
                      "value": "--flow.analytics.persistRemoteCallResponses=true --flow.analytics.persistResults=false --flow.config.custom.managementCenterUrl=http://34.140.160.158/mc\n"
                    }
                  ],
                  "externalConnectivity": {
                    "ingress": {
                      "annotations": {
                        "kubernetes.io/ingress.class": "nginx",
                        "nginx.ingress.kubernetes.io/enable-opentelemetry": "true",
                        "nginx.ingress.kubernetes.io/ssl-redirect": "false"
                      },
                      "hostname": "",
                      "ingressClassName": "nginx",
                      "path": "/"
                    }
                  },
                  "imagePullSecrets": [
                    {
                      "name": "hazelcast-cloud-hazelcast-connect-demo-pull-secret"
                    }
                  ],
                  "licenseKeySecretName": "flow-mc-license-key",
                  "repository": "quay.io/hazelcast_cloud/hazelcast-flow-internal",
                  "size": 3,
                  "version": "next"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast"
                },
                "spec": {
                  "clusterSize": 3,
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast-sample"
                },
                "spec": {
                  "clusterSize": 3,
                  "jet": {
                    "bucketConfig": {
                      "bucketURI": "gs://operator-user-code/jetJobs",
                      "secretName": "br-secret-gcp"
                    },
                    "enabled": true,
                    "resourceUploadEnabled": true
                  },
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HazelcastEndpoint",
                "metadata": {
                  "name": "hazelcastendpoint-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "port": 5701,
                  "type": "Discovery"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HotBackup",
                "metadata": {
                  "name": "hot-backup"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-longrun"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-longrun-2.0.0.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-1.0.2.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJobSnapshot",
                "metadata": {
                  "name": "jetjobsnapshot-sample"
                },
                "spec": {
                  "cancelJob": false,
                  "jetJobResourceName": "jet-job"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ManagementCenter",
                "metadata": {
                  "name": "managementcenter"
                },
                "spec": {
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "hazelcastClusters": [
                    {
                      "address": "hazelcast",
                      "name": "dev"
                    }
                  ],
                  "licenseKeySecretName": "hazelcast-license-key",
                  "persistence": {
                    "enabled": true,
                    "size": "10Gi"
                  },
                  "repository": "hazelcast/management-center"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Map",
                "metadata": {
                  "name": "map"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "MultiMap",
                "metadata": {
                  "name": "multimap-sample"
                },
                "spec": {
                  "backupCount": 3,
                  "binary": true,
                  "collectionType": "LIST",
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "queue-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ReplicatedMap",
                "metadata": {
                  "name": "replicatedmap-sample"
                },
                "spec": {
                  "asyncFillup": true,
                  "hazelcastResourceName": "hazelcast",
                  "inMemoryFormat": "OBJECT"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "topic-sample"
                },
                "spec": {
                  "globalOrderingEnabled": true,
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "UserCodeNamespace",
                "metadata": {
                  "name": "ucn-sample"
                },
                "spec": {
                  "bucketConfig": {
                    "bucketURI": "gs://operator-user-code/userCodeNamespace",
                    "secretName": "br-secret-gcp"
                  },
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "VectorCollection",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hazelcast-platform-operator"
                  },
                  "name": "vectorcollection-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "indexes": [
                    {
                      "dimension": 10,
                      "efConstruction": 256,
                      "maxDegree": 32,
                      "metric": "Dot",
                      "useDeduplication": false
                    }
                  ]
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanReplication",
                "metadata": {
                  "name": "wanreplication-sample"
                },
                "spec": {
                  "endpoints": "35.192.33.252",
                  "resources": [
                    {
                      "kind": "Map",
                      "name": "map-sample"
                    },
                    {
                      "kind": "Hazelcast",
                      "name": "hz-sample"
                    }
                  ],
                  "targetClusterName": "dev"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanSync",
                "metadata": {
                  "name": "wansync-sample"
                },
                "spec": {
                  "wanReplicationResourceName": "wan-replication-sample"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: docker.io/hazelcast/hazelcast-platform-operator@sha256:59328a4ea7a4e2110e0750364db792cf266f1e7231e3bb321e3de09d535dab23
          createdAt: "2025-03-13T11:59:59Z"
          description: Install Hazelcast clusters in Kubernetes environments.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hazelcast/hazelcast-platform-operator/
          support: Hazelcast, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.hazelcast.com
            version: v1alpha1
          - description: CronHotBackup is the Schema for the cronhotbackups API
            displayName: Cron Hot Backup
            kind: CronHotBackup
            name: cronhotbackups.hazelcast.com
            version: v1alpha1
          - description: Flow is the Schema for the flows API
            displayName: Flow
            kind: Flow
            name: flows.hazelcast.com
            version: v1alpha1
          - description: HazelcastEndpoint is the Schema for the hazelcastendpoints
              API
            displayName: Hazelcast Endpoint
            kind: HazelcastEndpoint
            name: hazelcastendpoints.hazelcast.com
            version: v1alpha1
          - description: Hazelcast is the Schema for the hazelcasts API
            displayName: Hazelcast
            kind: Hazelcast
            name: hazelcasts.hazelcast.com
            version: v1alpha1
          - description: HotBackup is the Schema for the hot backup API
            displayName: Hot Backup
            kind: HotBackup
            name: hotbackups.hazelcast.com
            version: v1alpha1
          - description: JetJob is the Schema for the jetjobs API
            displayName: Jet Job
            kind: JetJob
            name: jetjobs.hazelcast.com
            version: v1alpha1
          - description: JetJobSnapshot is the Schema for the jetjobsnapshots API
            displayName: Jet Job Snapshot
            kind: JetJobSnapshot
            name: jetjobsnapshots.hazelcast.com
            version: v1alpha1
          - description: ManagementCenter is the Schema for the managementcenters
              API
            displayName: Management Center
            kind: ManagementCenter
            name: managementcenters.hazelcast.com
            version: v1alpha1
          - description: Map is the Schema for the maps API
            displayName: Map
            kind: Map
            name: maps.hazelcast.com
            version: v1alpha1
          - description: MultiMap is the Schema for the multimaps API
            displayName: Multi Map
            kind: MultiMap
            name: multimaps.hazelcast.com
            version: v1alpha1
          - description: Queue is the Schema for the queues API
            displayName: Queue
            kind: Queue
            name: queues.hazelcast.com
            version: v1alpha1
          - description: ReplicatedMap is the Schema for the replicatedmaps API
            displayName: Replicated Map
            kind: ReplicatedMap
            name: replicatedmaps.hazelcast.com
            version: v1alpha1
          - description: Topic is the Schema for the topics API
            displayName: Topic
            kind: Topic
            name: topics.hazelcast.com
            version: v1alpha1
          - description: UserCodeNamespace is the Schema for the usercodenamespaces
              API
            displayName: User Code Namespace
            kind: UserCodeNamespace
            name: usercodenamespaces.hazelcast.com
            version: v1alpha1
          - description: VectorCollection is the Schema for the vectorcollections
              API
            displayName: Vector Collection
            kind: VectorCollection
            name: vectorcollections.hazelcast.com
            version: v1alpha1
          - description: WanReplication is the Schema for the wanreplications API
            displayName: Wan Replication
            kind: WanReplication
            name: wanreplications.hazelcast.com
            version: v1alpha1
          - description: WanSync is the Schema for the wansyncs API
            displayName: Wan Sync
            kind: WanSync
            name: wansyncs.hazelcast.com
            version: v1alpha1
        description: |
          Easily deploy Hazelcast clusters and Management Center into Kubernetes environments and manage their lifecycles.

          ## Before You Start

          If you are planning to create Hazelcast Platform Enterprise clusters, you need to [create a secret](https://docs.hazelcast.com/operator/latest/get-started#step-2-start-the-hazelcast-cluster) for the license. You can request a trial license key from [here](https://trialrequest.hazelcast.com).

          For Hazelcast Platform clusters, you can simply continue.

          ## Documentation

          1. [Get started](https://docs.hazelcast.com/operator/latest/get-started) with the Operator
          2. [Connect to the cluster from outside Kubernetes](https://guides.hazelcast.org/hazelcast-platform-operator-expose-externally)
          3. [Restore a Cluster from Cloud Storage with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-external-backup-restore)
          4. [Replicate Data between Two Hazelcast Clusters with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-wan-replication)

          ## Features

          Hazelcast Platform Operator supports the features below:

          * Custom resource for Hazelcast Platform (Enterprise) and Management Center
          * Observe status of Hazelcast clusters and Management Center instances
          * High Availability Mode configuration to create clusters that are resilient to node and zone failures
          * Support for TLS connections between members using self-signed certificates
          * Scale up and down Hazelcast clusters
          * Expose Hazelcast cluster to external clients ([Smart & Unisocket](https://docs.hazelcast.com/hazelcast/latest/clients/java#java-client-operation-modes))
          * Backup Hazelcast persistence data to cloud storage with the possibility of scheduling it and restoring the data accordingly
          * WAN Replication feature when you need to synchronize multiple Hazelcast clusters, which are connected by WANs
          * Full and Delta WAN Sync support
          * Tiered Storage support for Hazelcast and Map CRs
          * CP Subsystem configuration support for Hazelcast CR
          * User Code Deployment feature, which allows you to deploy custom and domain classes from cloud storages and URLs to Hazelcast members
          * The User Code Namespaces feature allows you to deploy custom and domain classes at runtime without a members restart
          * Support the configuration of advanced networking options
          * Support Multi-namespace configuration
          * ExecutorService and EntryProcessor support
          * Support several data structures like Map, Topic, MultiMap, ReplicatedMap, Queue and Cache which can be created dynamically via specific Custom Resources
          * MapStore, Near Cache and off-heap memory (HD memory and native memory) support for the Map CR
          * Native Memory support for the Cache CR
          * Support Jet configuration and Jet Job submission using the JetJob CR
          * Support for exporting the snapshots of JetJob CRs using JetJobSnapshot CR
          * Support for custom configurations using ConfigMap
        displayName: Hazelcast Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hazelcast
        - keyvalue
        - in-memory
        - database
        - caching
        links:
        - name: Documentation
          url: https://docs.hazelcast.com/operator/latest/
        - name: Source Code
          url: https://github.com/hazelcast/hazelcast-platform-operator
        maintainers:
        - email: cloudnative@hazelcast.com
          name: Hazelcast Cloud Native Team
        maturity: alpha
        minKubeVersion: 1.22.7
        provider:
          name: Hazelcast, Inc
        relatedImages:
        - docker.io/hazelcast/hazelcast-platform-operator@sha256:59328a4ea7a4e2110e0750364db792cf266f1e7231e3bb321e3de09d535dab23
        version: 5.15.0
      entries:
      - name: hazelcast-platform-operator.v5.15.0
        version: 5.15.0
      name: alpha
    defaultChannel: alpha
    packageName: hazelcast-platform-operator
    provider:
      name: Hazelcast, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: tang-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tang-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "daemons.redhat.com/v1alpha1",
                "kind": "TangServer",
                "metadata": {
                  "finalizers": [
                    "finalizer.daemons.tangserver.redhat.com"
                  ],
                  "name": "tangserver",
                  "namespace": "nbde"
                },
                "spec": {
                  "healthScript": "/usr/bin/tangd-health-check",
                  "image": "registry.redhat.io/rhel9/tang",
                  "keyPath": "/var/db/tang",
                  "persistentVolumeClaim": "tangserver-pvc",
                  "replicas": 1,
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:f43c96035f43040b3644e37119a3b35fc433f366d4b636459d1b9b02a05b5d3d
          createdAt: 2023-11-23T12:49:40
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TangServer is the Schema for the tangservers API
            displayName: Tang Server
            kind: TangServer
            name: tangservers.daemons.redhat.com
            version: v1alpha1
        description: NBDE Tang Server operator allows Tang Server deployment on OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nbde
        - tang
        links:
        - name: Tang Operator
          url: https://github.com/latchset/tang-operator
        maintainers:
        - email: sarroutb@redhat.com
          name: Sergio Arroutbi Braojos
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:97cade2c1ee468261aec5400728c8d44de387b459134aec7a4c3b5ec5a335d2c
        - registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:f43c96035f43040b3644e37119a3b35fc433f366d4b636459d1b9b02a05b5d3d
        version: 1.0.0
      entries:
      - name: tang-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: tang-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "daemons.redhat.com/v1alpha1",
                "kind": "TangServer",
                "metadata": {
                  "finalizers": [
                    "finalizer.daemons.tangserver.redhat.com"
                  ],
                  "name": "tangserver",
                  "namespace": "nbde"
                },
                "spec": {
                  "healthScript": "/usr/bin/tangd-health-check",
                  "image": "registry.redhat.io/rhel9/tang",
                  "keyPath": "/var/db/tang",
                  "persistentVolumeClaim": "tangserver-pvc",
                  "replicas": 1,
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:1cb2f794689817282ac570cacb106d75a2751a1926d89d3b83b619e02d869e5e
          createdAt: 2024-12-04T19:01:27
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TangServer is the Schema for the tangservers API
            displayName: Tang Server
            kind: TangServer
            name: tangservers.daemons.redhat.com
            version: v1alpha1
        description: NBDE Tang Server operator allows Tang Server deployment on OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nbde
        - tang
        links:
        - name: Tang Operator
          url: https://github.com/latchset/tang-operator
        maintainers:
        - email: sarroutb@redhat.com
          name: Sergio Arroutbi Braojos
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1172e150fff22c5eeab572f26961f3f53fbf896ee76d08c7503cfe2777c55458
        - registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:1cb2f794689817282ac570cacb106d75a2751a1926d89d3b83b619e02d869e5e
        version: 1.0.3
      entries:
      - name: tang-operator.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: alpha
    packageName: tang-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: www.redhat.com
    name: compliance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: compliance-operator.v0.1.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.17 <0.1.32'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          repository: https://github.com/openshift/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repo
          url: https://github.com/openshift/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/4.7/security/compliance_operator/compliance-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/compliance-rhel8-operator@sha256:ed68a0574ddeaf7d676403cd843f28cb586894f35dbb3ca5e06a5264bd42f030
        - registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c
        - registry.redhat.io/openshift4/compliance-openscap-rhel8@sha256:867c6d4205653d8a954d234cbaf470313009eb983d4f455fb45f49c4a7764422
        version: 0.1.32
      entries:
      - name: compliance-operator.v0.1.32
        version: 0.1.32
      name: "4.7"
    - currentCSV: compliance-operator.v0.1.61
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-node"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-platform"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "description": "Example of a tailoredProfile that extends OCP4 FedRAMP Moderate",
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.17 <0.1.61'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            displayName: Compliance Remediation
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            displayName: Compliance Scan
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            displayName: Compliance Suite
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            displayName: Profile Bundle
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            displayName: Scan Setting Binding
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            displayName: Tailored Profile
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repo
          url: https://github.com/openshift/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/compliance_operator/compliance-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b
        - registry.redhat.io/compliance/openshift-compliance-openscap-rhel8@sha256:83ad3113964ec48c5cf28315d923adb6edb95e45960e2756665d617dcd9b2e47
        - registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:5578e2225365ac1c15e4c189c5c7a8d14fd3fd26189414b36028f7f0879bb1dc
        version: 0.1.61
      entries:
      - name: compliance-operator.v0.1.61
        version: 0.1.61
      name: release-0.1
    - currentCSV: compliance-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-node"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-platform"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "description": "Example of a tailoredProfile that extends OCP4 FedRAMP Moderate",
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          must-gather-image: registry.redhat.io/compliance/openshift-compliance-must-gather-rhel8@sha256:e6468ccc55d72539b7fb980d0bce5c072e85879d4e9db39974d53eaab4d32b25
          olm.skipRange: '>=0.1.17 <1.7.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ComplianceAsCode/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            displayName: Compliance Remediation
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            displayName: Compliance Scan
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            displayName: Compliance Suite
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            displayName: Profile Bundle
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            displayName: Scan Setting Binding
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            displayName: Tailored Profile
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repository
          url: https://github.com/ComplianceAsCode/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/compliance_operator/co-overview.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4
        - registry.redhat.io/compliance/openshift-compliance-openscap-rhel8@sha256:d0f89d82845c6245a18e7db345ac3be92b2863d57f55f63e21652086128920c9
        - registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:8d3257bd6a020df53f8d7a400ae877ae029c78ffcd8ec9e35270ee3ff5c8e509
        - registry.redhat.io/compliance/openshift-compliance-must-gather-rhel8@sha256:e6468ccc55d72539b7fb980d0bce5c072e85879d4e9db39974d53eaab4d32b25
        version: 1.7.1
      entries:
      - name: compliance-operator.v1.7.1
        version: 1.7.1
      - name: compliance-operator.v1.7.0
        version: 1.7.0
      - name: compliance-operator.v1.6.2
        version: 1.6.2
      - name: compliance-operator.v1.6.1
        version: 1.6.1
      - name: compliance-operator.v1.6.0
        version: 1.6.0
      - name: compliance-operator.v1.5.1
        version: 1.5.1
      - name: compliance-operator.v1.5.0
        version: 1.5.0
      - name: compliance-operator.v1.4.1
        version: 1.4.1
      - name: compliance-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: compliance-operator
    provider:
      name: Red Hat Inc.
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: hyperfoil-bundle
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hyperfoil-operator.v0.26.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hyperfoil.io/v1alpha2",
                "kind": "Hyperfoil",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "hyperfoil-operator",
                    "app.kubernetes.io/instance": "hyperfoil-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hyperfoil",
                    "app.kubernetes.io/part-of": "hyperfoil-operator"
                  },
                  "name": "hyperfoil-sample"
                },
                "spec": {
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/hyperfoil/hyperfoil-operator:0.26.0
          createdAt: "2024-06-11T17:39:20Z"
          description: Microservice-oriented distributed benchmark framework.
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Hyperfoil/hyperfoil-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hyperfoil is the Schema for the hyperfoils API
            displayName: Hyperfoil
            kind: Hyperfoil
            name: hyperfoils.hyperfoil.io
            version: v1alpha2
        description: |
          Hyperfoil is a modern benchmark framework oriented promising
          accurate results with more flexibility and distributed architecture.

          See more on [hyperfoil.io](http://hyperfoil.io).

          Hyperfoil is licensed under Apache License 2.0
        displayName: Hyperfoil
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hyperfoil
        - benchmark
        - test
        - load
        - driver
        links:
        - name: Website
          url: http://hyperfoil.io
        - name: GitHub
          url: http://github.com/Hyperfoil
        maintainers:
        - email: rvansa@redhat.com
          name: Radim Vansa
        - email: spederse@redhat.com
          name: Ståle W. Pedersen
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/hyperfoil/hyperfoil-operator:0.26.0
        version: 0.26.0
      entries:
      - name: hyperfoil-operator.v0.26.0
        version: 0.26.0
      - name: hyperfoil-operator.v0.24.2
        version: 0.24.2
      - name: hyperfoil-operator.v0.21.0
        version: 0.21.0
      - name: hyperfoil-operator.v0.15.1
        version: 0.15.1
      - name: hyperfoil-operator.v0.15.0
        version: 0.15.0
      name: alpha
    defaultChannel: alpha
    packageName: hyperfoil-bundle
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Antline-Soft
      provider-url: ""
    name: cxl-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cxl-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-05-08T08:36:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
        description: memory extend operator
        displayName: extend-community-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - memory
        - extend
        - community
        - operator
        links:
        - name: Cxl Operator
          url: https://cxl-operator.domain
        maintainers:
        - email: tomboy@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        provider:
          name: Antline-Soft
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - rasen708/cxl-operator:latest
        version: 0.0.1
      entries:
      - name: cxl-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cxl-operator
    provider:
      name: Antline-Soft
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-msg-broker
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-msg-broker.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Messagebroker",
                "metadata": {
                  "name": "messagebroker-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "chiselServer": {
                    "enabled": false,
                    "port": 8080
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": false
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "natsCore": {
                    "reloader": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats-server-config-reloader",
                      "tag": "0.8.3-3747-30"
                    },
                    "server": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats",
                      "tag": "2.9.0-3747-21"
                    }
                  },
                  "natsHttpProxy": {
                    "enabled": false,
                    "port": 8080
                  },
                  "natsStreaming": {
                    "image": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats-streaming",
                      "tag": "0.23.2-3747-31"
                    }
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "tagSuffix": "-ocp"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:31:32Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Messagebroker
            name: messagebrokers.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Message Broker Operator
        displayName: Keysight WAP Message Broker Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Msg Broker
          url: https://keysight-wap-msg-broker.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-msg-broker@sha256:4701b6e456b099697bf1e25aeb4bc8e32b1892bb0956e36d84f9f80666b3a1fc
        version: 0.1.21
      entries:
      - name: keysight-wap-msg-broker.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-msg-broker
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC
      provider-url: https://www.nec.com
    name: sanstoragecsi-operator-bundle
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sanstoragecsi-operator-bundle.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.nec.com/v1",
                "kind": "SANStorageCSI",
                "metadata": {
                  "name": "sanstoragecsi"
                },
                "spec": {
                  "driverName": "san.csi.nec.com",
                  "images": {
                    "csiDriver": "",
                    "externalAttacher": "registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:bd8262cc4b6c405e175c39623323dd98a74bbf70e6aa90851a72612056c184ff",
                    "externalProvisioner": "registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:85f1cc8f21568e7885e492b29660f29e09796fb23ebdf5fd526aa2e99d6b2fa8",
                    "livenessProbe": "registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:1a0d15012ba70955af754b03a7488b7c6ace813646ec7c3ce845f80597050c3c",
                    "nodeRegistrar": "registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:dc7f0992be6c2853dc0e798afda368e36005a94ad76b21fa2f09790621318864"
                  },
                  "parameters": {
                    "controller": {
                      "logLevel": 5
                    },
                    "node": {
                      "livenessProbePort": 9808,
                      "logLevel": 5,
                      "maxVolumesPerNode": 1024
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          createdAt: "2022-06-15T00:00:00Z"
          description: An operator for managing the NEC Storage M Series CSI Driver
          operatorframework.io/suggested-namespace: nec-san-csi-operator-system
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: NEC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SANStorageCSI is the Schema for the sanstoragecsis API
            displayName: NEC Storage M Series CSI Driver
            kind: SANStorageCSI
            name: sanstoragecsis.csi.nec.com
            version: v1
        description: |
          This operator deploys the NEC Storage M Series CSI Driver on OpenShift Container Platform.

          NEC Storage M Series CSI Driver provides persistent storage for stateful applications using NEC Storage M Series.
          Please refer to the documentation of the NEC Storage M Series CSI Driver for detail information about supported NEC Storage Platforms and supported CSI features.
        displayName: NEC Storage M Series CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - NEC
        - M Series
        - SAN Storage
        - CSI Driver
        links:
        - name: NEC Storage M Series
          url: https://www.nec.com/en/global/prod/storage/product/san/index.html
        maturity: stable
        provider:
          name: NEC
          url: https://www.nec.com
        relatedImages:
        - registry.connect.redhat.com/nec/sanstoragecsi-operator@sha256:6656c2b5036bdc013712ecb640bbbebe0e3e90cf625c73697dbb49e3e201e2ac
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:bd8262cc4b6c405e175c39623323dd98a74bbf70e6aa90851a72612056c184ff
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:85f1cc8f21568e7885e492b29660f29e09796fb23ebdf5fd526aa2e99d6b2fa8
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:1a0d15012ba70955af754b03a7488b7c6ace813646ec7c3ce845f80597050c3c
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:dc7f0992be6c2853dc0e798afda368e36005a94ad76b21fa2f09790621318864
        version: 1.1.0
      entries:
      - name: sanstoragecsi-operator-bundle.v1.1.0
        version: 1.1.0
      - name: sanstoragecsi-operator-bundle.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: sanstoragecsi-operator-bundle
    provider:
      name: NEC
      url: https://www.nec.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nxiq-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nxiq-operator-certified.v1.193.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusIQ",
                "metadata": {
                  "name": "example-nexusiq"
                },
                "spec": {
                  "deployment": {
                    "postStart": {
                      "command": null
                    }
                  },
                  "deploymentStrategy": {
                    "type": "Recreate"
                  },
                  "fullnameOverride": "",
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [],
                    "path": "/",
                    "tls": []
                  },
                  "iq": {
                    "adminPort": 8071,
                    "applicationPort": 8070,
                    "env": [
                      {
                        "name": "JAVA_OPTS",
                        "value": "-Djava.util.prefs.userRoot=$(SONATYPE_WORK)/javaprefs -Djavax.net.ssl.trustStore=/etc/nexus-iq-server/jks/keystore.jks"
                      }
                    ],
                    "configYaml": {
                      "sonatypeWork": "/sonatype-work",
                      "createSampleData": true,
                      "server": {
                        "adminConnectors": [
                          {
                            "port": 8071,
                            "type": "http"
                          }
                        ],
                        "applicationConnectors": [
                          {
                            "port": 8070,
                            "type": "http"
                          }
                        ],
                        "requestLog": {
                          "appenders": [
                            {
                              "type": "file",
                              "currentLogFilename": "/var/log/nexus-iq-server/request.log",
                              "logFormat": "%clientHost %l %user [%date] \"%requestURL\" %statusCode %bytesSent %elapsedTime \"%header{User-Agent}\"",
                              "archivedLogFilenamePattern": "/var/log/nexus-iq-server/request-%d.log.gz",
                              "archivedFileCount": 50
                            }
                          ]
                        }
                      },
                      "logging": {
                        "level": "DEBUG",
                        "loggers": {
                          "com.sonatype.insight.scan": "INFO",
                          "eu.medsea.mimeutil.MimeUtil2": "INFO",
                          "org.apache.http": "INFO",
                          "org.apache.http.wire": "ERROR",
                          "org.eclipse.birt.report.engine.layout.pdf.font.FontConfigReader": "WARN",
                          "org.eclipse.jetty": "INFO",
                          "org.apache.shiro.web.filter.authc.BasicHttpAuthenticationFilter": "INFO",
                          "com.sonatype.insight.audit": {
                            "appenders": [
                              {
                                "type": "file",
                                "currentLogFilename": "/var/log/nexus-iq-server/audit.log",
                                "archivedLogFilenamePattern": "/var/log/nexus-iq-server/audit-%d.log.gz",
                                "archivedFileCount": 50
                              }
                            ]
                          },
                          "com.sonatype.insight.policy.violation": {
                            "appenders": [
                              {
                                "type": "console",
                                "threshold": "INFO"
                              }
                            ]
                          }
                        },
                        "appenders": [
                          {
                            "type": "console",
                            "threshold": "INFO",
                            "logFormat": "%d{'yyyy-MM-dd HH:mm:ss,SSSZ'} %level [%thread] %X{username} %logger - %msg%n"
                          },
                          {
                            "type": "file",
                            "threshold": "ALL",
                            "logFormat": "%d{'yyyy-MM-dd HH:mm:ss,SSSZ'} %level [%thread] %X{username} %logger - %msg%n",
                            "currentLogFilename": "/var/log/nexus-iq-server/clm-server.log",
                            "archivedLogFilenamePattern": "/var/log/nexus-iq-server/clm-server-%d.log.gz",
                            "archivedFileCount": 50
                          }
                        ]
                      }
                    },
                    "imageName": "registry.connect.redhat.com/sonatype/nexus-iq-server@sha256:d65df6bdff1f32bf32a87e5cb38f5e9f98b75b81e9acccbffbd967e1f71691f0",
                    "imagePullPolicy": "IfNotPresent",
                    "imagePullSecret": "",
                    "licenseSecret": "",
                    "memory": "1Gi",
                    "name": "nxiq",
                    "certificates": {}
                  },
                  "nameOverride": "",
                  "persistence": {
                    "accessMode": "ReadWriteOnce",
                    "storageSize": "1Gi"
                  },
                  "service": {
                    "annotations": {},
                    "enabled": false,
                    "labels": {},
                    "ports": [
                      {
                        "name": "nexus-service",
                        "port": 80,
                        "targetPort": 80
                      }
                    ]
                  },
                  "serviceAccount": {
                    "create": true,
                    "name": null
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nxiq-operator-certified@sha256:59c52729189da87d9b5c3de8ef85326662d94c99835ab99d6e8d7fefdcf53ef0
          createdAt: "2020-07-20"
          description: |-
            Nexus Lifecycle is an open source governance platform that continuously identifies risk,
            automatically enforces policy, and provides visibility throughout the entire SDLC.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.193.0-1
          repository: https://github.com/sonatype/operator-iq
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus IQ Server
            displayName: NexusIQ
            kind: NexusIQ
            name: nexusiqs.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Lifecycle is an open source governance platform that continuously identifies risk,
          automatically enforces policy, and provides visibility throughout the entire SDLC.
          Developers leveraging open source require a solution that helps them make safer choices
          while still delivering at DevOps speed.
          Nexus Lifecycle empowers developers and security teams to choose the best components
          and govern open source usage, ensuring organizations continue to innovate with less risk.

          ## Core Capabilities

          * **Dependency Management**:
            Generate a precise software bill of materials for all applications
            to identify open source components and gather contextual direct
            and transitive dependency information for faster action.
          * **One-Click Remediation**:
            Upgrade to the best components and versions based on real-time intelligence.
            Available in your preferred IDE or a GitHub pull request.
          * **Automated Open Source Policy Enforcement**:
            Create custom security, license, and architectural policies based
            on application type or organization and contextually enforce those policies
            across every stage of the SDLC.
          * **Continuous Monitoring**:
            Real-time notifications of newly discovered defects,
            as well as vulnerabilities based on component,
            risk level or application/container affected.

          ## Extras

          * IDE Integrations: Eclipse, IntelliJ, and Visual Studio
          * Source Control Integrations: GitHub, GitLab, and Bitbucket
          * Chrome Browser Extension
          * Red Hat Clair for Container Scanning

          ## Limitations

          This operator will be released on a quarterly basis.

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus IQ Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage

          Once the server instance is created by the operator and running,
          you'll want to expose the service as you see fit:
          1. Create a Route to the new service for iq.applicationPort (8070).
          2. Visit the URL provided by the Route, login, and set new credentials.
            The default credentials are `admin`/`admin123`.

          The Nexus IQ Server can be further configured via the NexusIQ custom resource definition:

          | Parameter            | Description                                                  | Default           |
          | -------------------- | ------------------------------------------------------------ | ----------------- |
          | `iq.applicationPort` | Port of the application connector. Must match the value in the `configYaml` property | `8070`            |
          | `iq.adminPort`       | Port of the application connector. Must match the value in the `configYaml` property | `8071`            |
          | `iq.memory`          | The amount of RAM to allocate                                | `1Gi`             |
          | `iq.env`             | Customize the server environment, including `JAVA_OPTS`      | See example YAML show when ceating a NexusIQ. |
          | `iq.licenseSecret`   | The base-64 encoded license file to be installed at startup  | `""`              |
          | `iq.configYaml`      | A YAML block which will be used as a configuration block for IQ Server. | See example YAML shown when creating a NexusIQ. |
          | `iq.certificates`   | Add your own certificates for your own trusted Certificate Authorities with names and plain armored text. | `{}` |
          | `ingress.enabled`                           | Create an ingress for Nexus         | `true`                                  |
          | `ingress.annotations`                       | Annotations to enhance ingress configuration  | `{}`                          |
          | `ingress.tls.enabled`                       | Enable TLS                          | `true`                                 |
          | `ingress.tls.secretName`                    | Name of the secret storing TLS cert, `false` to use the Ingress' default certificate | `nexus-tls`                             |
          | `ingress.path`                              | Path for ingress rules. GCP users should set to `/*` | `/`                    |
          | `deploymentStrategy.type`                   | Deployment Strategy                 |  `Recreate` |
          | `initContainers`                            | Containers to run before starting the server | `nil` |
          | `deployment.postStart.command`              | Command to run after starting the IQ Server container  | `nil`                    |
          | `deployment.terminationGracePeriodSeconds` | Time to allow for clean shutdown                        | 120                      |
          | `persistence.storageClass` | The provisioner class                        | `-` (disables dynamic provisioning)            |
          | `persistence.storageSize` | The amount of drive space to allocate                        | `1Gi`             |
          | `persistence.accessMode` | Default access mode                        | `ReadWriteOnce`             |
          | `persistence.volumeConfiguration` | A YAML block to configure the persistent volume type. Defaults to `hostPath` which should not be used in production | `hostPath`             |

          ## Configuring IQ Server

          You can define the `config.yml` for IQ Server in your CRD on startup.
          It is the `iq.configYaml` property. For more details, see the [Configuring IQ Server](https://help.sonatype.com/iqserver/configuring) help page.

          ## Installing the License

          The license file can be installed via the UI when IQ server is running, or it can be done as a part of the deploy.
          If you leave the `licenseFile` field empty/commented, IQ Server will start and prompt you to manually install the license
          when you first enter the GUI.

          ### Installing the License Automatically
          To do it automatically, first encode your `.lic` file in Base 64 with no line breaks, eg:

          ```bash
          base64 --wrap=0 mylicense.lic > lic.base64
          ```

          Then add this value to your CRD file as `iq.licenseSecret`, eg:

          ```yaml
          iq:
            licenseSecret: bXkgc2FtcGxlIGxpY2Vuc2U=
          ```

          Specify the `licenseFile` path in `iq.configYaml` as:

          ```yaml
          iq:
            configYaml:
              sonatypeWork: /sonatype-work
              createSampleData: true
              # add this line and the `licenseSecret` above to autoconfigure licensing
              licenseFile: /etc/nexus-iq-license/license_lic
              server:
                applicationConnectors:
                  - type: http
                    port: 8070
                adminConnectors:
                  - type: http
                    port: 8071
          ```

          ## Adding Certificates for Your Own Trusted Certificate Authorities

          Specify a set of certificates to be imported into the Java trust store
          of the running IQ Server. More than one certificate can be provided.

          ```yaml
          iq:
            certificates:
              my-company-cert: |
                -----BEGIN CERTIFICATE-----
                ...
                -----END CERTIFICATE-----
          ```

          ## Running Commands Before Initialization of the Server

          If you need custom initialization, you can use standard Kubernetes
          [`initContainers`](https://kubernetes.io/docs/concepts/workloads/pods/init-containers/)
          to run any series of containers and commands you need.

          ```yaml
          initContainers:
            - name: an-init-container
              image: busybox:1.28
              command: ['/bin/sh', '-c', 'echo "initializing"']
          ```
        displayName: Nexus IQ Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - sonatype
        - licensing
        - policy
        - components
        links:
        - name: Nexus Lifecycle
          url: https://www.sonatype.com/product-nexus-lifecycle
        maintainers:
        - email: support@sonaype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nxiq-operator-certified@sha256:59c52729189da87d9b5c3de8ef85326662d94c99835ab99d6e8d7fefdcf53ef0
        - registry.connect.redhat.com/sonatype/nexus-iq-server@sha256:d65df6bdff1f32bf32a87e5cb38f5e9f98b75b81e9acccbffbd967e1f71691f0
        version: 1.193.0-1
      entries:
      - name: nxiq-operator-certified.v1.193.0-1
        version: 1.193.0-1
      name: stable
    defaultChannel: stable
    packageName: nxiq-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com/
    name: openstack-baremetal-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openstack-baremetal-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "baremetal.openstack.org/v1beta1",
                "kind": "OpenStackBaremetalSet",
                "metadata": {
                  "name": "openstackbaremetalset"
                },
                "spec": {
                  "baremetalHosts": {
                    "compute-0": "172.22.0.100/24",
                    "compute-1": "172.22.0.101/24"
                  },
                  "ctlplaneGateway": "172.22.0.3",
                  "ctlplaneInterface": "enp1s0",
                  "deploymentSSHSecret": "baremetalset-deploy-secret",
                  "domainName": "osptest.openstack.org",
                  "osImage": "edpm-hardened-uefi.qcow2",
                  "passwordSecret": "baremetalset-password-secret"
                }
              },
              {
                "apiVersion": "baremetal.openstack.org/v1beta1",
                "kind": "OpenStackProvisionServer",
                "metadata": {
                  "name": "openstackprovisionserver"
                },
                "spec": {
                  "interface": "enp1s0",
                  "osImage": "edpm-hardened-uefi.qcow2",
                  "port": 8080
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:55:07Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenStackBaremetalSet is the Schema for the openstackbaremetalsets
              API
            displayName: OpenStack BaremetalSet
            kind: OpenStackBaremetalSet
            name: openstackbaremetalsets.baremetal.openstack.org
            version: v1beta1
          - description: OpenStackProvisionServer used to serve custom images for
              baremetal provisioning with Metal3
            displayName: OpenStackProvisionServer
            kind: OpenStackProvisionServer
            name: openstackprovisionservers.baremetal.openstack.org
            version: v1beta1
        description: An operator for handling baremetal provisioning concerns for
          OpenStack
        displayName: OpenStack Baremetal Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cn-openstack
        - openstack
        links:
        - name: Openstack Baremetal Operator
          url: https://github.com/openstack-k8s-operators/openstack-baremetal-operator
        maturity: beta
        provider:
          name: Red Hat
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/ubi9/httpd-24@sha256:dbfbe3383e12d63df7178e4cdd5b1e6e51a84c781041b9192026ea6edb72fbf3
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/openstack-baremetal-rhel9-operator@sha256:a6c69c03e59faf50973d07ccea32b80cb9c46ceb60ed00ffe7311ca42d501296
        - registry.redhat.io/rhoso/edpm-hardened-uefi-rhel9@sha256:e4bfacb9d8de056a251bee16a66b69631b96fb19fb926d8e33acd379d40a92d6
        - registry.redhat.io/rhoso-operators/openstack-baremetal-agent-rhel9@sha256:8f6f8e99042710819aa639eae6513f36c4642190ecd85d58a2ed819f1b55e38f
        version: 1.0.6
      entries:
      - name: openstack-baremetal-operator.v1.0.6
        version: 1.0.6
      - name: openstack-baremetal-operator.v1.0.5
        version: 1.0.5
      - name: openstack-baremetal-operator.v1.0.4
        version: 1.0.4
      - name: openstack-baremetal-operator.v1.0.3
        version: 1.0.3
      - name: openstack-baremetal-operator.v1.0.2
        version: 1.0.2
      - name: openstack-baremetal-operator.v1.0.1
        version: 1.0.1
      - name: openstack-baremetal-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: openstack-baremetal-operator
    provider:
      name: Red Hat
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecr-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecr-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.services.k8s.aws/v1alpha1",
                "kind": "Repository",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecr-controller:1.3.1
          createdAt: "2025-09-20T01:21:24Z"
          description: AWS ECR controller is a service controller for managing ECR
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PullThroughCacheRule represents the state of an AWS ecr PullThroughCacheRule
              resource.
            displayName: PullThroughCacheRule
            kind: PullThroughCacheRule
            name: pullthroughcacherules.ecr.services.k8s.aws
            version: v1alpha1
          - description: Repository represents the state of an AWS ecr Repository
              resource.
            displayName: Repository
            kind: Repository
            name: repositories.ecr.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Container Registry (ECR) resources in AWS from within your Kubernetes cluster.

          **About Amazon ECR**

          Amazon Elastic Container Registry (Amazon ECR) is an AWS managed container image registry service that is secure, scalable, and reliable. Amazon ECR supports private repositories with resource-based permissions using AWS IAM. This is so that specified users or Amazon EC2 instances can access your container repositories and images. You can use your preferred CLI to push, pull, and manage Docker images, Open Container Initiative (OCI) images, and OCI compatible artifacts.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECR
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECR Developer Resources
          url: https://aws.amazon.com/ecr/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecr maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecr-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-ecr-controller.v1.3.1
        version: 1.3.1
      - name: ack-ecr-controller.v1.3.0
        version: 1.3.0
      - name: ack-ecr-controller.v1.2.0
        version: 1.2.0
      - name: ack-ecr-controller.v1.1.0
        version: 1.1.0
      - name: ack-ecr-controller.v1.0.33
        version: 1.0.33
      - name: ack-ecr-controller.v1.0.32
        version: 1.0.32
      - name: ack-ecr-controller.v1.0.31
        version: 1.0.31
      - name: ack-ecr-controller.v1.0.30
        version: 1.0.30
      - name: ack-ecr-controller.v1.0.29
        version: 1.0.29
      - name: ack-ecr-controller.v1.0.28
        version: 1.0.28
      - name: ack-ecr-controller.v1.0.27
        version: 1.0.27
      - name: ack-ecr-controller.v1.0.25
        version: 1.0.25
      - name: ack-ecr-controller.v1.0.24
        version: 1.0.24
      - name: ack-ecr-controller.v1.0.23
        version: 1.0.23
      - name: ack-ecr-controller.v1.0.22
        version: 1.0.22
      - name: ack-ecr-controller.v1.0.21
        version: 1.0.21
      - name: ack-ecr-controller.v1.0.20
        version: 1.0.20
      - name: ack-ecr-controller.v1.0.19
        version: 1.0.19
      - name: ack-ecr-controller.v1.0.17
        version: 1.0.17
      - name: ack-ecr-controller.v1.0.16
        version: 1.0.16
      - name: ack-ecr-controller.v1.0.15
        version: 1.0.15
      - name: ack-ecr-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecr-controller.v1.0.12
        version: 1.0.12
      - name: ack-ecr-controller.v1.0.10
        version: 1.0.10
      - name: ack-ecr-controller.v1.0.9
        version: 1.0.9
      - name: ack-ecr-controller.v1.0.8
        version: 1.0.8
      - name: ack-ecr-controller.v1.0.7
        version: 1.0.7
      - name: ack-ecr-controller.v1.0.6
        version: 1.0.6
      - name: ack-ecr-controller.v1.0.4
        version: 1.0.4
      - name: ack-ecr-controller.v1.0.3
        version: 1.0.3
      - name: ack-ecr-controller.v1.0.2
        version: 1.0.2
      - name: ack-ecr-controller.v1.0.0
        version: 1.0.0
      - name: ack-ecr-controller.v0.1.9
        version: 0.1.9
      - name: ack-ecr-controller.v0.1.7
        version: 0.1.7
      - name: ack-ecr-controller.v0.1.6
        version: 0.1.6
      - name: ack-ecr-controller.v0.1.5
        version: 0.1.5
      - name: ack-ecr-controller.v0.1.4
        version: 0.1.4
      - name: ack-ecr-controller.v0.1.3
        version: 0.1.3
      - name: ack-ecr-controller.v0.1.2
        version: 0.1.2
      - name: ack-ecr-controller.v0.1.1
        version: 0.1.1
      - name: ack-ecr-controller.v0.1.0
        version: 0.1.0
      - name: ack-ecr-controller.v0.0.27
        version: 0.0.27
      - name: ack-ecr-controller.v0.0.26
        version: 0.0.26
      - name: ack-ecr-controller.v0.0.25
        version: 0.0.25
      - name: ack-ecr-controller.v0.0.24
        version: 0.0.24
      - name: ack-ecr-controller.v0.0.23
        version: 0.0.23
      - name: ack-ecr-controller.v0.0.22
        version: 0.0.22
      - name: ack-ecr-controller.v0.0.21
        version: 0.0.21
      - name: ack-ecr-controller.v0.0.20
        version: 0.0.20
      - name: ack-ecr-controller.v0.0.19
        version: 0.0.19
      - name: ack-ecr-controller.v0.0.18
        version: 0.0.18
      - name: ack-ecr-controller.v0.0.17
        version: 0.0.17
      - name: ack-ecr-controller.v0.0.16
        version: 0.0.16
      - name: ack-ecr-controller.v0.0.15
        version: 0.0.15
      - name: ack-ecr-controller.v0.0.14
        version: 0.0.14
      - name: ack-ecr-controller.v0.0.13
        version: 0.0.13
      - name: ack-ecr-controller.v0.0.12
        version: 0.0.12
      - name: ack-ecr-controller.v0.0.11
        version: 0.0.11
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecr-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: VictoriaMetrics
      provider-url: ""
    name: victoriametrics-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: victoriametrics-operator.v0.63.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLAgent",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "persistentVolumeClaimRetentionPolicy": {
                    "whenDeleted": "Delete"
                  },
                  "remoteWrite": [
                    {
                      "maxDiskUsage": "10GB",
                      "url": "http://vlsingle-example-0.default.svc:9428/internal/insert"
                    }
                  ],
                  "remoteWriteSettings": {
                    "maxBlockSize": "30MB"
                  },
                  "replicaCount": 2,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "350Mi"
                    }
                  },
                  "storage": {
                    "volumeClaimTemplate": {
                      "spec": {
                        "resources": {
                          "requests": {
                            "storage": "50Gi"
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "victoriametrics-operator"
                  },
                  "name": "example"
                },
                "spec": {
                  "requestsLoadBalancer": {
                    "enabled": true,
                    "spec": {}
                  },
                  "vlinsert": {
                    "replicaCount": 1
                  },
                  "vlselect": {
                    "replicaCount": 1
                  },
                  "vlstorage": {
                    "replicaCount": 2,
                    "retentionPeriod": "1y",
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "50Gi"
                            }
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLSingle",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "victoriametrics-operator"
                  },
                  "name": "sample"
                },
                "spec": {
                  "retentionMaxDiskSpaceUsageBytes": "20GB",
                  "retentionPeriod": "1y",
                  "storage": {
                    "resources": {
                      "requests": {
                        "storage": "50Gi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VLogs",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "retentionPeriod": "2"
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAgent",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "extraArgs": {
                    "memory.allowedPercent": "40"
                  },
                  "remoteWrite": [
                    {
                      "maxDiskUsage": "10GB",
                      "url": "http://vmsingle-dc-1.default.svc:8429/api/v1/write"
                    },
                    {
                      "maxDiskUsage": "15GB",
                      "url": "http://vmsingle-dc-2.default.svc:8429/api/v1/write"
                    },
                    {
                      "url": "http://vminsert-main.default.svc:8480/insert/1:1/prometheus/api/v1/write"
                    }
                  ],
                  "remoteWriteSettings": {
                    "maxDiskUsagePerURL": "5GB"
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "350Mi"
                    }
                  },
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlert",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "datasource": {
                    "url": "http://vmsingle-example-pvc.default.svc:8429"
                  },
                  "evaluationInterval": "30s",
                  "notifier": {
                    "url": "http://vmalertmanager-example.default.svc:9093"
                  },
                  "replicaCount": 1,
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlertmanager",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "configSecret": "am-global-config",
                  "replicaCount": 1,
                  "selectAllByDefault": true,
                  "storage": {
                    "volumeClaimTemplate": {
                      "spec": {
                        "resources": {
                          "requests": {
                            "storage": "2Gi"
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlertmanagerConfig",
                "metadata": {
                  "name": "example-email-web"
                },
                "spec": {
                  "receivers": [
                    {
                      "email_configs": [
                        {
                          "from": "alerting@example.com",
                          "smarthost": "example.com:25",
                          "text": "ALARM",
                          "to": "some-email@example.com"
                        }
                      ],
                      "name": "email",
                      "webhook_configs": [
                        {
                          "url": "http://some-other-wh"
                        }
                      ]
                    }
                  ],
                  "route": {
                    "group_interval": "1m",
                    "receiver": "email",
                    "routes": [
                      {
                        "receiver": "email"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAuth",
                "metadata": {
                  "name": "vmauth-sample"
                },
                "spec": {
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "victoriametrics/vmauth"
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "250m",
                      "memory": "350Mi"
                    }
                  },
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMCluster",
                "metadata": {
                  "name": "example-persistent"
                },
                "spec": {
                  "replicationFactor": 2,
                  "retentionPeriod": "4",
                  "vminsert": {
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "0.5",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "vmselect": {
                    "cacheMountPath": "/select-cache",
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "0.5",
                        "memory": "500Mi"
                      }
                    },
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "2Gi"
                            }
                          }
                        }
                      }
                    }
                  },
                  "vmstorage": {
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1500Mi"
                      }
                    },
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "10Gi"
                            }
                          }
                        }
                      }
                    },
                    "storageDataPath": "/vm-data"
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMNodeScrape",
                "metadata": {
                  "name": "cadvisor-metrics"
                },
                "spec": {
                  "bearerTokenFile": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                  "interval": "10s",
                  "relabelConfigs": [
                    {
                      "action": "labelmap",
                      "regex": "__meta_kubernetes_node_label_(.+)"
                    },
                    {
                      "replacement": "kubernetes.default.svc:443",
                      "targetLabel": "__address__"
                    },
                    {
                      "regex": "(.+)",
                      "replacement": "/api/v1/nodes/$1/proxy/metrics/cadvisor",
                      "sourceLabels": [
                        "__meta_kubernetes_node_name"
                      ],
                      "targetLabel": "__metrics_path__"
                    }
                  ],
                  "scheme": "https",
                  "scrapeTimeout": "2s",
                  "selector": {},
                  "tlsConfig": {
                    "caFile": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                    "insecureSkipVerify": true
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMPodScrape",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "podMetricsEndpoints": [
                    {
                      "port": "metrics",
                      "scheme": "http"
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app.kubernetes.io/name": "example-application",
                      "monitored-by": "vm-operator"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMProbe",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "interval": "2s",
                  "jobName": "static-probe",
                  "module": "http_2xx",
                  "targets": {
                    "staticConfig": {
                      "targets": [
                        "vmagent-example.default.svc:8429/health"
                      ]
                    }
                  },
                  "vmProberSpec": {
                    "url": "prometheus-blackbox-exporter.default.svc:9115"
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMRule",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "groups": [
                    {
                      "name": "kafka",
                      "rules": [
                        {
                          "alert": "coordinator down",
                          "annotations": {
                            "description": "kafka coordinator is down",
                            "value": "{{ $value }}"
                          },
                          "expr": "ml_app_gauge{exec_context=\"consumer_group_state\"} == 0",
                          "for": "60s",
                          "labels": {
                            "job": "{{ $labels.job }}",
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMScrapeConfig",
                "metadata": {
                  "name": "mongodb"
                },
                "spec": {
                  "consulSDConfigs": [
                    {
                      "server": "https://consul-dns:8500",
                      "services": [
                        "mongodb"
                      ]
                    }
                  ],
                  "relabelConfigs": [
                    {
                      "action": "replace",
                      "sourceLabels": [
                        "__meta_consul_service"
                      ],
                      "targetLabel": "job"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMServiceScrape",
                "metadata": {
                  "name": "example-scrape"
                },
                "spec": {
                  "endpoints": [
                    {
                      "port": "http"
                    }
                  ],
                  "namespaceSelector": {},
                  "selector": {
                    "matchLabels": {
                      "app.kubernetes.io/name": "example-application",
                      "monitored-by": "vm-operator"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMSingle",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "retentionPeriod": "1"
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMSingle",
                "metadata": {
                  "name": "example-pvc"
                },
                "spec": {
                  "removePvcAfterDelete": true,
                  "retentionPeriod": "1",
                  "storage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMStaticScrape",
                "metadata": {
                  "name": "vmstaticscrape-sample"
                },
                "spec": {
                  "jobName": "static",
                  "targetEndpoints": [
                    {
                      "labels": {
                        "env": "dev",
                        "project": "operator"
                      },
                      "targets": [
                        "192.168.0.1:9100",
                        "196.168.0.50:9100"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMUser",
                "metadata": {
                  "name": "vmuser-tenant-1"
                },
                "spec": {
                  "bearerToken": "some-token",
                  "targetRefs": [
                    {
                      "crd": {
                        "kind": "VMCluster/vminsert",
                        "name": "test-persistent",
                        "namespace": "vm"
                      },
                      "target_path_suffix": "/insert/1"
                    },
                    {
                      "crd": {
                        "kind": "VMCluster/vmselect",
                        "name": "test-persistent",
                        "namespace": "vm"
                      },
                      "target_path_suffix": "/select/1"
                    },
                    {
                      "paths": [
                        "/internal/resetRollupResultCache"
                      ],
                      "static": {
                        "url": "http://vmselect-test-persistent.default.svc:8481/"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: docker.io/victoriametrics/operator:v0.63.0
          createdAt: "2025-09-12T12:58:29Z"
          description: Provides monitoring capabilities for kubernetes clusters and
            applications
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: "1.30"
          operators.openshift.io/infrastructure-features: '[fips]'
          operators.operatorframework.io.bundle.channel.default.v1: beta
          operators.operatorframework.io.bundle.channels.v1: beta
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/VictoriaMetrics/operator
          support: VictoriaMetrics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VLAgent - is a tiny but brave agent, which helps you collect
              logs from various sources and stores them in VictoriaLogs.
            displayName: VLAgent
            kind: VLAgent
            name: vlagents.operator.victoriametrics.com
            version: v1
          - description: VLCluster is fast, cost-effective and scalable logs database.
            displayName: VLCluster
            kind: VLCluster
            name: vlclusters.operator.victoriametrics.com
            version: v1
          - description: |-
              VLogs is fast, cost-effective and scalable logs database.
              VLogs is the Schema for the vlogs API
            displayName: VLogs
            kind: VLogs
            name: vlogs.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VLSingle is fast, cost-effective and scalable logs database.
              VLSingle is the Schema for the API
            displayName: VLSingle
            kind: VLSingle
            name: vlsingles.operator.victoriametrics.com
            version: v1
          - description: |-
              VMAgent - is a tiny but brave agent, which helps you collect metrics from various sources and stores them in VictoriaMetrics
              or any other Prometheus-compatible storage system that supports the remote_write protocol.
            displayName: VMAgent
            kind: VMAgent
            name: vmagents.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlertmanagerConfig is the Schema for the vmalertmanagerconfigs
              API
            displayName: VMAlertmanager Config
            kind: VMAlertmanagerConfig
            name: vmalertmanagerconfigs.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlertmanager represents Victoria-Metrics deployment for
              Alertmanager.
            displayName: VMAlertmanager
            kind: VMAlertmanager
            name: vmalertmanagers.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlert  executes a list of given alerting or recording rules
              against configured address.
            displayName: VMAlert
            kind: VMAlert
            name: vmalerts.operator.victoriametrics.com
            version: v1beta1
          - description: VMAnomaly is the Schema for the vmanomalies API.
            displayName: VMAnomaly
            kind: VMAnomaly
            name: vmanomalies.operator.victoriametrics.com
            version: v1
          - description: VMAuth is the Schema for the vmauths API
            displayName: VMAuth
            kind: VMAuth
            name: vmauths.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMCluster is fast, cost-effective and scalable time-series database.
              Cluster version with
            displayName: VMCluster
            kind: VMCluster
            name: vmclusters.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMNodeScrape defines discovery for targets placed on kubernetes nodes,
              usually its node-exporters and other host services.
              InternalIP is used as __address__ for scraping.
            displayName: VMNode Scrape
            kind: VMNodeScrape
            name: vmnodescrapes.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMPodScrape is scrape configuration for pods,
              it generates vmagent's config for scraping pod targets
              based on selectors.
            displayName: VMPod Scrape
            kind: VMPodScrape
            name: vmpodscrapes.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMProbe defines a probe for targets, that will be executed with prober,
              like blackbox exporter.
              It helps to monitor reachability of target with various checks.
            displayName: VMProbe
            kind: VMProbe
            name: vmprobes.operator.victoriametrics.com
            version: v1beta1
          - description: VMRule defines rule records for vmalert application
            displayName: VMRule
            kind: VMRule
            name: vmrules.operator.victoriametrics.com
            version: v1beta1
          - description: VMScrapeConfig specifies a set of targets and parameters
              describing how to scrape them.
            displayName: VMScrape Config
            kind: VMScrapeConfig
            name: vmscrapeconfigs.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMServiceScrape is scrape configuration for endpoints associated with
              kubernetes service,
              it generates scrape configuration for vmagent based on selectors.
              result config will scrape service endpoints
            displayName: VMService Scrape
            kind: VMServiceScrape
            name: vmservicescrapes.operator.victoriametrics.com
            version: v1beta1
          - description: VMSingle  is fast, cost-effective and scalable time-series
              database.
            displayName: VMSingle
            kind: VMSingle
            name: vmsingles.operator.victoriametrics.com
            version: v1beta1
          - description: VMStaticScrape  defines static targets configuration for
              scraping.
            displayName: VMStatic Scrape
            kind: VMStaticScrape
            name: vmstaticscrapes.operator.victoriametrics.com
            version: v1beta1
          - description: VMUser is the Schema for the vmusers API
            displayName: VMUser
            kind: VMUser
            name: vmusers.operator.victoriametrics.com
            version: v1beta1
          - description: VTCluster is fast, cost-effective and scalable traces database.
            displayName: VTCluster
            kind: VTCluster
            name: vtclusters.operator.victoriametrics.com
            version: v1
          - description: |-
              VTSingle is fast, cost-effective and scalable traces database.
              VTSingle is the Schema for the API
            displayName: VTSingle
            kind: VTSingle
            name: vtsingles.operator.victoriametrics.com
            version: v1
        description: |
          Operator manages VictoriaMetrics applications and provides monitoring features for applications running inside and outside kubernetes cluster. It has support for prometheus-operator objects and
          provides migration mechanism.

          Operator designed to use existing installation of VictoriaMetrics inside or outside kubernetes cluster or it could
          bring own resources and fully manages them.

           You can read more about operator at docs:
            - quick start [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/quick-start.md)
            - high availability [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/high-availability.md)
            - design and description of implementation [design](https://github.com/VictoriaMetrics/operator/blob/master/docs/design.md)
            - operator objects description [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/api.md)

          # VictoriaMetrics
          * VictoriaMetrics can be used as long-term storage for Prometheus or for [vmagent](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmagent/README.md).
          See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#prometheus-setup) for details.
          * Supports [Prometheus querying API](https://prometheus.io/docs/prometheus/latest/querying/api/), so it can be used as Prometheus drop-in replacement in Grafana.
          VictoriaMetrics implements [MetricsQL](https://github.com/VictoriaMetrics/VictoriaMetrics/wiki/MetricsQL) query language, which is inspired by PromQL.
          * Supports global query view. Multiple Prometheus instances may write data into VictoriaMetrics. Later this data may be used in a single query.
          * High performance and good scalability for both [inserts](https://medium.com/@valyala/high-cardinality-tsdb-benchmarks-victoriametrics-vs-timescaledb-vs-influxdb-13e6ee64dd6b)
          and [selects](https://medium.com/@valyala/when-size-matters-benchmarking-victoriametrics-vs-timescale-and-influxdb-6035811952d4).
          [Outperforms InfluxDB and TimescaleDB by up to 20x](https://medium.com/@valyala/measuring-vertical-scalability-for-time-series-databases-in-google-cloud-92550d78d8ae).
          * [Uses 10x less RAM than InfluxDB](https://medium.com/@valyala/insert-benchmarks-with-inch-influxdb-vs-victoriametrics-e31a41ae2893) when working with millions of unique time series (aka high cardinality).
          * Optimized for time series with high churn rate. Think about [prometheus-operator](https://github.com/coreos/prometheus-operator) metrics from frequent deployments in Kubernetes.
          * High data compression, so [up to 70x more data points](https://medium.com/@valyala/when-size-matters-benchmarking-victoriametrics-vs-timescale-and-influxdb-6035811952d4)
          may be crammed into limited storage comparing to TimescaleDB.
          * Optimized for storage with high-latency IO and low IOPS (HDD and network storage in AWS, Google Cloud, Microsoft Azure, etc). See [graphs from these benchmarks](https://medium.com/@valyala/high-cardinality-tsdb-benchmarks-victoriametrics-vs-timescaledb-vs-influxdb-13e6ee64dd6b).
          * A single-node VictoriaMetrics may substitute moderately sized clusters built with competing solutions such as Thanos, M3DB, Cortex, InfluxDB or TimescaleDB.
          See [vertical scalability benchmarks](https://medium.com/@valyala/measuring-vertical-scalability-for-time-series-databases-in-google-cloud-92550d78d8ae),
          [comparing Thanos to VictoriaMetrics cluster](https://medium.com/@valyala/comparing-thanos-to-victoriametrics-cluster-b193bea1683)
          and [Remote Write Storage Wars](https://promcon.io/2019-munich/talks/remote-write-storage-wars/) talk
          from [PromCon 2019](https://promcon.io/2019-munich/talks/remote-write-storage-wars/).
          * Easy operation:
              * VictoriaMetrics consists of a single [small executable](https://medium.com/@valyala/stripping-dependency-bloat-in-victoriametrics-docker-image-983fb5912b0d) without external dependencies.
              * All the configuration is done via explicit command-line flags with reasonable defaults.
              * All the data is stored in a single directory pointed by `-storageDataPath` flag.
              * Easy and fast backups from [instant snapshots](https://medium.com/@valyala/how-victoriametrics-makes-instant-snapshots-for-multi-terabyte-time-series-data-e1f3fb0e0282)
              to S3 or GCS with [vmbackup](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmbackup/README.md) / [vmrestore](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmrestore/README.md).
              See [this article](https://medium.com/@valyala/speeding-up-backups-for-big-time-series-databases-533c1a927883) for more details.
          * Storage is protected from corruption on unclean shutdown (i.e. OOM, hardware reset or `kill -9`) thanks to [the storage architecture](https://medium.com/@valyala/how-victoriametrics-makes-instant-snapshots-for-multi-terabyte-time-series-data-e1f3fb0e0282).
          * Supports metrics' scraping, ingestion and [backfilling](https://github.com/VictoriaMetrics/VictoriaMetrics#backfilling) via the following protocols:
            * [Metrics from Prometheus exporters](https://github.com/prometheus/docs/blob/master/content/docs/instrumenting/exposition_formats.md#text-based-format)
              such as [node_exporter](https://github.com/prometheus/node_exporter). See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-scrape-prometheus-exporters-such-as-node-exporter) for details.
              * [Prometheus remote write API](https://prometheus.io/docs/prometheus/latest/configuration/configuration/#remote_write)
              * [InfluxDB line protocol](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-send-data-from-influxdb-compatible-agents-such-as-telegraf) over HTTP, TCP and UDP.
              * [Graphite plaintext protocol](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-send-data-from-graphite-compatible-agents-such-as-statsd) with [tags](https://graphite.readthedocs.io/en/latest/tags.html#carbon)
              if `-graphiteListenAddr` is set.
              * [OpenTSDB put message](https://github.com/VictoriaMetrics/VictoriaMetrics#sending-data-via-telnet-put-protocol) if `-opentsdbListenAddr` is set.
              * [HTTP OpenTSDB /api/put requests](https://github.com/VictoriaMetrics/VictoriaMetrics#sending-opentsdb-data-via-http-apiput-requests) if `-opentsdbHTTPListenAddr` is set.
              * [How to import time series data](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-time-series-data).
              * [Prometheus exposition format](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-data-in-prometheus-exposition-format).
              * [Arbitrary CSV data](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-csv-data).
          * Supports metrics' relabeling. See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#relabeling) for details.
          * Ideally works with big amounts of time series data from Kubernetes, IoT sensors, connected cars, industrial telemetry, financial data and various Enterprise workloads.
        displayName: VictoriaMetrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - prometheus
        - victoriametrics
        links:
        - name: VictoriaMetrics
          url: https://victoriametrics.com
        maintainers:
        - email: nik@victoriametrics.com
          name: f41gh7
        - email: info@victoriametrics.com
          name: info
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: VictoriaMetrics
        relatedImages:
        - docker.io/victoriametrics/operator:v0.63.0
        version: 0.63.0
      entries:
      - name: victoriametrics-operator.v0.63.0
        version: 0.63.0
      - name: victoriametrics-operator.v0.62.0
        version: 0.62.0
      - name: victoriametrics-operator.v0.61.2
        version: 0.61.2
      - name: victoriametrics-operator.v0.61.1
        version: 0.61.1
      - name: victoriametrics-operator.v0.61.0
        version: 0.61.0
      - name: victoriametrics-operator.v0.60.2
        version: 0.60.2
      - name: victoriametrics-operator.v0.60.1
        version: 0.60.1
      - name: victoriametrics-operator.v0.60.0
        version: 0.60.0
      - name: victoriametrics-operator.v0.59.2
        version: 0.59.2
      - name: victoriametrics-operator.v0.59.1
        version: 0.59.1
      - name: victoriametrics-operator.v0.59.0
        version: 0.59.0
      - name: victoriametrics-operator.v0.58.1
        version: 0.58.1
      - name: victoriametrics-operator.v0.58.0
        version: 0.58.0
      - name: victoriametrics-operator.v0.57.0
        version: 0.57.0
      - name: victoriametrics-operator.v0.56.0
        version: 0.56.0
      - name: victoriametrics-operator.v0.55.0
        version: 0.55.0
      - name: victoriametrics-operator.v0.54.1
        version: 0.54.1
      - name: victoriametrics-operator.v0.53.0
        version: 0.53.0
      - name: victoriametrics-operator.v0.52.0
        version: 0.52.0
      - name: victoriametrics-operator.v0.51.1
        version: 0.51.1
      - name: victoriametrics-operator.v0.49.0
        version: 0.49.0
      - name: victoriametrics-operator.v0.48.4
        version: 0.48.4
      - name: victoriametrics-operator.v0.48.3
        version: 0.48.3
      - name: victoriametrics-operator.v0.48.1
        version: 0.48.1
      - name: victoriametrics-operator.v0.47.3
        version: 0.47.3
      - name: victoriametrics-operator.v0.47.1
        version: 0.47.1
      - name: victoriametrics-operator.v0.47.0
        version: 0.47.0
      - name: victoriametrics-operator.v0.46.4
        version: 0.46.4
      - name: victoriametrics-operator.v0.46.3
        version: 0.46.3
      - name: victoriametrics-operator.v0.39.4
        version: 0.39.4
      - name: victoriametrics-operator.v0.33.0
        version: 0.33.0
      - name: victoriametrics-operator.v0.32.1
        version: 0.32.1
      - name: victoriametrics-operator.v0.32.0
        version: 0.32.0
      - name: victoriametrics-operator.v0.29.1
        version: 0.29.1
      - name: victoriametrics-operator.v0.27.2
        version: 0.27.2
      - name: victoriametrics-operator.v0.26.2
        version: 0.26.2
      - name: victoriametrics-operator.v0.26.0
        version: 0.26.0
      - name: victoriametrics-operator.v0.25.1
        version: 0.25.1
      - name: victoriametrics-operator.v0.25.0
        version: 0.25.0
      - name: victoriametrics-operator.v0.24.0
        version: 0.24.0
      - name: victoriametrics-operator.v0.23.3
        version: 0.23.3
      - name: victoriametrics-operator.v0.22.1
        version: 0.22.1
      - name: victoriametrics-operator.v0.22.0
        version: 0.22.0
      - name: victoriametrics-operator.v0.21.0
        version: 0.21.0
      - name: victoriametrics-operator.v0.20.1
        version: 0.20.1
      - name: victoriametrics-operator.v0.19.1
        version: 0.19.1
      - name: victoriametrics-operator.v0.19.0
        version: 0.19.0
      - name: victoriametrics-operator.v0.18.2
        version: 0.18.2
      - name: victoriametrics-operator.v0.18.1
        version: 0.18.1
      - name: victoriametrics-operator.v0.18.0
        version: 0.18.0
      name: beta
    defaultChannel: beta
    packageName: victoriametrics-operator
    provider:
      name: VictoriaMetrics
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: F5 Networks Inc.
      provider-url: ""
    name: f5-bigip-ctlr-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: alpha
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: beta
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: stable
    defaultChannel: stable
    packageName: f5-bigip-ctlr-operator
    provider:
      name: F5 Networks Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dns-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dns-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSHealthCheckProbe",
                "metadata": {
                  "name": "$NAME"
                },
                "spec": {
                  "additionalHeadersRef": {
                    "name": "headers"
                  },
                  "address": "192.168.0.16",
                  "allowInsecureCertificate": true,
                  "failureThreshold": 5,
                  "hostname": "test.com",
                  "interval": "60s",
                  "path": "/healthz",
                  "port": 443,
                  "protocol": "HTTPS"
                }
              },
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSRecord",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dns-operator",
                    "app.kubernetes.io/instance": "dnsrecord-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dnsrecord",
                    "app.kubernetes.io/part-of": "dns-operator"
                  },
                  "name": "dnsrecord-sample"
                },
                "spec": {
                  "endpoints": [
                    {
                      "dnsName": "dnsrecord-simple.example.com",
                      "recordTTL": 60,
                      "recordType": "A",
                      "targets": [
                        "52.215.108.61",
                        "52.30.101.221"
                      ]
                    }
                  ],
                  "providerRef": {
                    "name": "dns-provider-creds"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/dns-rhel9-operator@sha256:8ad3876e36ba2f76355499d98a97672f54f26754fc9d219d9cb1c5b9f9874430
          createdAt: "2025-04-07T08:13:58Z"
          description: Configures how north-south traffic from outside the network
            is balanced and reaches Gateways
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kuadrant/dns-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSHealthCheckProbe is the Schema for the dnshealthcheckprobes
              API
            displayName: DNSHealthCheckProbe
            kind: DNSHealthCheckProbe
            name: dnshealthcheckprobes.kuadrant.io
            version: v1alpha1
          - description: DNSRecord is the Schema for the dnsrecords API
            displayName: DNSRecord
            kind: DNSRecord
            name: dnsrecords.kuadrant.io
            version: v1alpha1
        description: Configures how north-south traffic from outside the network is
          balanced and reaches Gateways - Part of Red Hat Connectivity Link
        displayName: Red Hat - DNS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dns
        - kuadrant
        links:
        - name: GitHub
          url: https://github.com/kuadrant/dns-operator
        - name: Documentation
          url: https://github.com/Kuadrant/dns-operator/tree/main/docs
        maintainers:
        - email: mnairn@redhat.com
          name: Michael Nairn
        - email: pbrookes@redhat.com
          name: Phil Brookes
        - email: cbrookes@redhat.com
          name: Craig Brookes
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhcl-1/dns-rhel9-operator@sha256:8ad3876e36ba2f76355499d98a97672f54f26754fc9d219d9cb1c5b9f9874430
        version: 1.1.0
      entries:
      - name: dns-operator.v1.1.0
        version: 1.1.0
      - name: dns-operator.v1.0.2
        version: 1.0.2
      - name: dns-operator.v1.0.1
        version: 1.0.1
      - name: dns-operator.v0.12.0
        version: 0.12.0
      name: stable
    defaultChannel: stable
    packageName: dns-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: patch-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: patch-operator.v0.1.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Patch",
                "metadata": {
                  "name": "test-complex-patch"
                },
                "spec": {
                  "patches": {
                    "test-complex-patch": {
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 1).metadata.name }}: {{ (index . 2).metadata.name }}    \n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "kind": "Namespace",
                          "name": "test-patch-operator"
                        },
                        {
                          "apiVersion": "v1",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "test-patch-operator"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "test-patch-operator"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/patch-operator@sha256:030ade9b9428deca3ac6390618dc4d6e6d728577d1467a22dc4d4b68c9848886
          createdAt: "2024-05-10T14:07:19Z"
          description: An operator to apply patches to Kubernetes objects in a declarative
            way.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: patch-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/patch-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Patch
            name: patches.redhatcop.redhat.io
            version: v1alpha1
        description: "The patch operator helps with defining patches in a declarative
          way. This operator has two main features:\n\n1. [ability to patch an object
          at creation time via a mutating webhook](#creation-time-patch-injection)\n2.
          [ability to enforce patches on one or more objects via a controller](#runtime-patch-enforcement)\n
          \n## Creation-time patch injection\n\nWhy apply a patch at creation time
          when you could directly create the correct object? The reason is that sometime
          the correct value depends on configuration set on the specific cluster in
          which the object is being deployed. For example, an ingress/route hostname
          might depend on the specific cluster. Consider the following example based
          on cert-manager:\n \n```yaml\napiVersion: cert-manager.io/v1\nkind: ClusterIssuer\nmetadata:\n
          \ name: letsencrypt-issuer\nspec:\n  acme:\n    server: 'https://acme-v02.api.letsencrypt.org/directory'\n
          \   email: {{ .Values.letsencrypt.email }}\n    privateKeySecretRef:\n      name:
          letsencrypt-staging\n    solvers:  \n    - dns01:\n        route53:\n          accessKeyID:
          << access_key >>\n          secretAccessKeySecretRef:\n            name:
          cert-manager-dns-credentials\n            key: aws_secret_access_key\n          region:
          << region >>\n          hostedZoneID: << hosted_zone_id >>\n ```\n\nIn this
          example the fields: `<< access_key >>`, `<< region >>` and `<< hosted_zone_id
          >>` are dependent on the specific region in which the cluster is being deployed
          and in many cases they are discoverable from other configurations already
          present in the cluster. If you want to deploy the above Cluster Issuer object
          with a gitops approach, then there is no easy way to discover those values.
          The solution so far is to manually discover those values and create a different
          gitops configuration for each cluster. But consider if you could look up
          values at deploy time based on the cluster you are deploying to. Here is
          how this object might look:\n\n```yaml\napiVersion: cert-manager.io/v1\nkind:
          ClusterIssuer\nmetadata:\n  name: letsencrypt-issuer\n  namespace: {{ .Release.Namespace
          }}\n  annotations:\n    \"redhat-cop.redhat.io/patch\": |\n      spec:\n
          \       acme:\n        - dns01:\n            route53:\n              accessKeyID:
          {{ (lookup \"v1\" \"Secret\" .metadata.namespace \"cert-manager-dns-credentials\").data.aws_access_key_id
          | b64dec }}\n              secretAccessKeySecretRef:\n                name:
          cert-manager-dns-credentials\n                key: aws_secret_access_key\n
          \             region: {{ (lookup \"config.openshift.io/v1\" \"Infrastructure\"
          \"\" \"cluster\").status.platformStatus.aws.region }}\n              hostedZoneID:
          {{ (lookup \"config.openshift.io/v1\" \"DNS\" \"\" \"cluster\").spec.publicZone.id
          }} \nspec:\n  acme:\n    server: 'https://acme-v02.api.letsencrypt.org/directory'\n
          \   email: {{ .Values.letsencrypt.email }}\n    privateKeySecretRef:\n      name:
          letsencrypt-staging\n    solvers:  \n    - dns01:\n        route53:\n          accessKeyID:
          << access_key >>\n          secretAccessKeySecretRef:\n            name:
          cert-manager-dns-credentials\n            key: aws_secret_access_key\n          region:
          << region >>\n          hostedZoneID: << hosted_zone_id >>\n ```\n\nThe
          annotation specifies a patch that will be applied by a [MutatingWebhook](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/),
          as you can see the three values are now looked up from different configurations
          in the cluster.\n\nTwo annotations influences the behavior of this MutatingWebhook\n
          \n1. \"redhat-cop.redhat.io/patch\" : this is the patch itself. The patch
          is evaluated as a template with the object itself as it's only parameter.
          The template is expressed in golang template notation and supports the same
          functions as helm template including the [lookup](https://helm.sh/docs/chart_template_guide/functions_and_pipelines/#using-the-lookup-function)
          function which plays a major role here. The patch must be expressed in yaml
          for readability. It will be converted to json by the webhook logic.\n2.
          \"redhat-cop.redhat.io/patch-type\" : this is the type of json patch. The
          possible values are: `application/json-patch+json`, `application/merge-patch+json`
          and `application/strategic-merge-patch+json`. If this annotation is omitted
          it defaults to strategic merge.\n\n### Security Considerations\n \nThe lookup
          function, if used by the template, is executed with a client which impersonates
          the user issuing the object creation/update request. This should prevent
          security permission leakage.\n\n### Installing the creation time webhook\n
          \nThe creation time webhook is not installed by the operator. This is because
          there is no way to know which specific object type should be intercepted
          and intercepting all of the types would be too inefficient. It's up to the
          administrator then to install the webhook. Here is some guidance.\n\nIf
          you installed the operator via OLM, use the following webhook template:\n\n```yaml\napiVersion:
          admissionregistration.k8s.io/v1\n kind: MutatingWebhookConfiguration\nmetadata:\n
          \ name: patch-operator-inject\n  annotations:\n    service.beta.openshift.io/inject-cabundle:
          \"true\"\nwebhooks:\n - admissionReviewVersions:\n  - v1\n  clientConfig:\n
          \   service:\n      name: patch-operator-webhook-service\n      namespace:
          patch-operator\n      path: /inject\n  failurePolicy: Fail\n  name: patch-operator-inject.redhatcop.redhat.io\n
          \ rules:\n  - << add your intercepted objects here >>\n  sideEffects: None\n```\n\nIf
          you installed the operator via the Helm chart and are using cert-manager,
          use the following webhook template:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\n
          kind: MutatingWebhookConfiguration\nmetadata:\n  name: patch-operator-inject\n
          \ annotations:\n    cert-manager.io/inject-ca-from: '{{ .Release.Namespace
          }}/webhook-server-cert'\n webhooks:\n- admissionReviewVersions:\n  - v1\n
          \ clientConfig:\n    service:\n      name: patch-operator-webhook-service\n
          \     namespace: patch-operator\n      path: /inject\n  failurePolicy: Fail\n
          \ name: patch-operator-inject.redhatcop.redhat.io\n  rules:\n  - << add
          your intercepted objects here >>\n  sideEffects: None\n```  \n\nYou should
          need to enable the webhook only for `CREATE` operations. So for example
          to enable the webhook on configmaps:\n\n```yaml\n  rules:\n  - apiGroups:\n
          \   - \"\"\n    apiVersions:\n    - v1\n    operations:\n    - CREATE\n
          \   resources:\n    - configmaps\n```\n\n## Runtime patch enforcement\n\nThere
          are situations when we need to patch pre-existing objects. Again this is
          a use case that is hard to model with gitops operators which will work only
          on object that they own. Especially with sophisticated Kubernetes distributions,
          it is not uncommon that a Kubernetes instance, at installation time, is
          configured with some default settings. Changing those configurations means
          patching those objects. For example, let's take the case of OpenShift Oauth
          configuration. This object is present by default and it is expected to be
          patched with any newly enabled authentication mechanism. This is how it
          looks like after installation:\n\n```yaml\napiVersion: config.openshift.io/v1\n
          kind: OAuth\nmetadata:\n  name: cluster\n  ownerReferences:\n    - apiVersion:
          config.openshift.io/v1\n      kind: ClusterVersion\n      name: version\n
          \     uid: 9a9d450b-3076-4e30-ac05-a889d6341fc3\n  resourceVersion: '20405124'\n
          spec: []\n```\n\nIf we need to patch it we can use the patch controller
          and the `Patch` object as follows:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: Patch\nmetadata:\n  name: gitlab-ocp-oauth-provider\n  namespace:
          openshift-config\nspec:\n  serviceAccountRef:\n    name: default\n  patches:\n
          \   gitlab-ocp-oauth-provider:\n      targetObjectRef:\n        apiVersion:
          config.openshift.io/v1\n        kind: OAuth\n        name: cluster\n      patchTemplate:
          |\n        spec:\n          identityProviders:\n          - name: my-github
          \n            mappingMethod: claim \n            type: GitHub\n            github:\n
          \             clientID: \"{{ (index . 1).data.client_id | b64dec }}\" \n
          \             clientSecret: \n                name: ocp-github-app-credentials\n
          \             organizations: \n              - my-org\n              teams:
          []            \n      patchType: application/merge-patch+json\n      sourceObjectRefs:\n
          \     - apiVersion: v1\n        kind: Secret\n        name: ocp-github-app-credentials\n
          \       namespace: openshift-config\n```\n\nThis will cause the OAuth object
          to be patched and the patch to be enforced. That means that if anything
          changes on the secret that we use a parameter (which may be rotated) or
          in Oauth object itself, the patch will be reapplied. In this case we are
          adding a gitlab authentication provider.\n\n A `patch` has the following
          fields:\n\n`targetObjectRef` this refers to the object(s) receiving the
          patch. Mutliple object can be selected based on the following rules:\n \n|
          Namespaced Type | Namespace | Name | Selection type |\n| --- | --- | ---
          | --- |\n| yes | null | null | multiple selection across namespaces |\n|
          yes | null | not null | multiple selection across namespaces where the name
          corresponds to the passed name |\n| yes | not null | null | multiple selection
          within a namespace |\n| yes | not null | not nul | single selection |\n|
          no | N/A | null | multiple selection  |\n| no | N/A | not null | single
          selection |\n\nSelection can be further narrowed down by filtering by labels
          and/or annotations using the `labelSelector` and `annotationSelector` fields.
          The patch will be applied to all of the selected instances.\n\n`sourceObjectRefs`
          these are the objects that will be watched and become part of the parameters
          of the patch template. Name and Namespace of sourceRefObjects are interpreted
          as golang templates with the current target instance and the only parameter.
          This allows to select different source object for each target object.\n
          \nSo, for example, with this patch:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: Patch\nmetadata:\n  name: multiple-namespaced-targets-patch\nspec:\n
          \ serviceAccountRef:\n    name: default\n  patches:\n    multiple-namespaced-targets-patch:\n
          \     targetObjectRef:\n        apiVersion: v1\n        kind: ServiceAccount\n
          \       name: deployer\n      patchTemplate: |\n        metadata:\n          annotations:\n
          \           {{ (index . 1).metadata.uid }}: {{ (index . 1) }}\n      patchType:
          application/strategic-merge-patch+json\n      sourceObjectRefs:\n      -
          apiVersion: v1\n        kind: ServiceAccount\n        name: default\n        namespace:
          \"{{ .metadata.namespace }}\"\n        fieldPath: $.metadata.uid\n```\n\nThe
          `deployer` service accounts from all namespaces are selected as target of
          this patch, each patch template will receive a different parameter and that
          is the `default` service account of the same namespace as the namespace
          of the `deployer` service account being processed.\n\n`sourceObjectRefs`
          also have the `fieldPath` field which can contain a jsonpath expression.
          If a value is passed the jsonpath expression will be calculate for the current
          source object and the result will be passed as parameter of the template.\n\n`patchTemplate`
          This is the the template that will be evaluated. The result must be a valid
          patch compatible with the requested type and expressed in yaml for readability.
          The parameters passed to the template are the target object and then the
          all of the source object. So if you want to refer to the target object in
          the template you can use this expression `(index . 0)`. Higher indexes refer
          to the sourceObjectRef array. The template is expressed in golang template
          notation and supports the same functions as helm template.\n\n`patchType`
          is the type of the json patch. The possible values are: `application/json-patch+json`,
          `application/merge-patch+json` and `application/strategic-merge-patch+json`.
          If this annotation is omitted it defaults to strategic merge.\n\n### Patch
          Controller Security Considerations\n\nThe patch enforcement enacted by the
          patch controller is executed with a client which uses the service account
          referenced by the `serviceAccountRef` field. So before a patch object can
          actually work an administrator must have granted the needed permissions
          to a service account in the same namespace. The `serviceAccountRef` will
          default to the `default` service account if not specified.\n\n### Patch
          Controller Performance Considerations\n\nThe patch controller will create
          a controller-manager and per `Patch` object and a reconciler for each of
          the `PatchSpec` defined in the array on patches in the `Patch` object.\nThese
          reconcilers share the same cached client. In order to be able to watch changes
          on target and source objects of a `PatchSpec`, all of the target and source
          object type instances will be cached by the client. This is a normal behavior
          of a controller-manager client, but it implies that if you create patches
          on object types that have many instances in etcd (Secrets, ServiceAccounts,
          Namespaces for example), the patch operator instance will require a significant
          amount of memory. A way to contain this issue is to try to aggregate together
          `PatchSpec` that deal with the same object types. This will cause those
          object type instances to cached only once.  \n"
        displayName: Patch Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - patch
        - enforcing
        - declarative
        links:
        - name: Patch Operator
          url: https://github.com/redhat-cop/patch-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/patch-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/patch-operator@sha256:030ade9b9428deca3ac6390618dc4d6e6d728577d1467a22dc4d4b68c9848886
        version: 0.1.11
      entries:
      - name: patch-operator.v0.1.11
        version: 0.1.11
      - name: patch-operator.v0.1.10
        version: 0.1.10
      - name: patch-operator.v0.1.9
        version: 0.1.9
      - name: patch-operator.v0.1.8
        version: 0.1.8
      - name: patch-operator.v0.1.7
        version: 0.1.7
      - name: patch-operator.v0.1.6
        version: 0.1.6
      - name: patch-operator.v0.1.5
        version: 0.1.5
      - name: patch-operator.v0.1.4
        version: 0.1.4
      - name: patch-operator.v0.1.3
        version: 0.1.3
      - name: patch-operator.v0.1.2
        version: 0.1.2
      - name: patch-operator.v0.1.1
        version: 0.1.1
      - name: patch-operator.v0.1.0
        version: 0.1.0
      - name: patch-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: patch-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-elbv2-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-elbv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "LoadBalancer",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "TargetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/elbv2-controller:1.1.1
          createdAt: "2025-09-19T22:16:38Z"
          description: AWS ELB controller is a service controller for managing ELB
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Listener represents the state of an AWS elbv2 Listener resource.
            displayName: Listener
            kind: Listener
            name: listeners.elbv2.services.k8s.aws
            version: v1alpha1
          - description: LoadBalancer represents the state of an AWS elbv2 LoadBalancer
              resource.
            displayName: LoadBalancer
            kind: LoadBalancer
            name: loadbalancers.elbv2.services.k8s.aws
            version: v1alpha1
          - description: Rule represents the state of an AWS elbv2 Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.elbv2.services.k8s.aws
            version: v1alpha1
          - description: TargetGroup represents the state of an AWS elbv2 TargetGroup
              resource.
            displayName: TargetGroup
            kind: TargetGroup
            name: targetgroups.elbv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Load Balancing (ELB) resources in AWS from within your Kubernetes cluster.

          **About Amazon ELB**

          Elastic Load Balancing automatically distributes your incoming traffic across multiple targets, such as EC2 instances, containers, and IP addresses, in one or more Availability Zones. It monitors the health of its registered targets, and routes traffic only to the healthy targets. Elastic Load Balancing scales your load balancer capacity automatically in response to changes in incoming traffic.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ELB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elbv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ELB Developer Resources
          url: https://aws.amazon.com/elasticloadbalancing
        maintainers:
        - email: ack-maintainers@amazon.com
          name: elb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/elbv2-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-elbv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-elbv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-elbv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-elbv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-elbv2-controller.v1.0.10
        version: 1.0.10
      - name: ack-elbv2-controller.v1.0.9
        version: 1.0.9
      - name: ack-elbv2-controller.v1.0.8
        version: 1.0.8
      - name: ack-elbv2-controller.v1.0.7
        version: 1.0.7
      - name: ack-elbv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-elbv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-elbv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-elbv2-controller.v1.0.3
        version: 1.0.3
      - name: ack-elbv2-controller.v1.0.2
        version: 1.0.2
      - name: ack-elbv2-controller.v1.0.1
        version: 1.0.1
      - name: ack-elbv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-elbv2-controller.v0.0.11
        version: 0.0.11
      - name: ack-elbv2-controller.v0.0.10
        version: 0.0.10
      - name: ack-elbv2-controller.v0.0.9
        version: 0.0.9
      - name: ack-elbv2-controller.v0.0.8
        version: 0.0.8
      - name: ack-elbv2-controller.v0.0.6
        version: 0.0.6
      - name: ack-elbv2-controller.v0.0.5
        version: 0.0.5
      - name: ack-elbv2-controller.v0.0.4
        version: 0.0.4
      - name: ack-elbv2-controller.v0.0.3
        version: 0.0.3
      - name: ack-elbv2-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-elbv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hazelcast, Inc
      provider-url: ""
    name: hazelcast-platform-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hazelcast-platform-operator.v5.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "cache-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "CronHotBackup",
                "metadata": {
                  "name": "cronhotbackup-sample"
                },
                "spec": {
                  "hotBackupTemplate": {
                    "metadata": {
                      "labels": {
                        "cron-hotbackup": "true"
                      }
                    },
                    "spec": {
                      "hazelcastResourceName": "hazelcast"
                    }
                  },
                  "schedule": "*/30 * * * *"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast"
                },
                "spec": {
                  "clusterSize": 3,
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast-sample"
                },
                "spec": {
                  "clusterSize": 3,
                  "jet": {
                    "bucketConfig": {
                      "bucketURI": "gs://operator-user-code/jetJobs",
                      "secretName": "br-secret-gcp"
                    },
                    "enabled": true,
                    "resourceUploadEnabled": true
                  },
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HazelcastEndpoint",
                "metadata": {
                  "name": "hazelcastendpoint-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "port": 5701,
                  "type": "Discovery"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HotBackup",
                "metadata": {
                  "name": "hot-backup"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-longrun"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-longrun-2.0.0.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-1.0.2.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJobSnapshot",
                "metadata": {
                  "name": "jetjobsnapshot-sample"
                },
                "spec": {
                  "cancelJob": false,
                  "jetJobResourceName": "jet-job"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ManagementCenter",
                "metadata": {
                  "name": "managementcenter"
                },
                "spec": {
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "hazelcastClusters": [
                    {
                      "address": "hazelcast",
                      "name": "dev"
                    }
                  ],
                  "licenseKeySecretName": "hazelcast-license-key",
                  "persistence": {
                    "enabled": true,
                    "size": "10Gi"
                  },
                  "repository": "hazelcast/management-center"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Map",
                "metadata": {
                  "name": "map"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "MultiMap",
                "metadata": {
                  "name": "multimap-sample"
                },
                "spec": {
                  "backupCount": 3,
                  "binary": true,
                  "collectionType": "LIST",
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "queue-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ReplicatedMap",
                "metadata": {
                  "name": "replicatedmap-sample"
                },
                "spec": {
                  "asyncFillup": true,
                  "hazelcastResourceName": "hazelcast",
                  "inMemoryFormat": "OBJECT"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "topic-sample"
                },
                "spec": {
                  "globalOrderingEnabled": true,
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "UserCodeNamespace",
                "metadata": {
                  "name": "ucn-sample"
                },
                "spec": {
                  "bucketConfig": {
                    "bucketURI": "gs://operator-user-code/userCodeNamespace",
                    "secretName": "br-secret-gcp"
                  },
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanReplication",
                "metadata": {
                  "name": "wanreplication-sample"
                },
                "spec": {
                  "endpoints": "35.192.33.252",
                  "resources": [
                    {
                      "kind": "Map",
                      "name": "map-sample"
                    },
                    {
                      "kind": "Hazelcast",
                      "name": "hz-sample"
                    }
                  ],
                  "targetClusterName": "dev"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanSync",
                "metadata": {
                  "name": "wansync-sample"
                },
                "spec": {
                  "wanReplicationResourceName": "wan-replication-sample"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: docker.io/hazelcast/hazelcast-platform-operator@sha256:99005afda5c71ceeac200928dfed4ea35480f871053101e3aeaa44005e3fda14
          createdAt: "2024-07-29T15:35:56Z"
          description: '[DEPRECATED] Use the certified Hazelcast Operator instead'
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hazelcast/hazelcast-platform-operator/
          support: Hazelcast, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.hazelcast.com
            version: v1alpha1
          - description: CronHotBackup is the Schema for the cronhotbackups API
            displayName: Cron Hot Backup
            kind: CronHotBackup
            name: cronhotbackups.hazelcast.com
            version: v1alpha1
          - description: HazelcastEndpoint is the Schema for the hazelcastendpoints
              API
            displayName: Hazelcast Endpoint
            kind: HazelcastEndpoint
            name: hazelcastendpoints.hazelcast.com
            version: v1alpha1
          - description: Hazelcast is the Schema for the hazelcasts API
            displayName: Hazelcast
            kind: Hazelcast
            name: hazelcasts.hazelcast.com
            version: v1alpha1
          - description: HotBackup is the Schema for the hot backup API
            displayName: Hot Backup
            kind: HotBackup
            name: hotbackups.hazelcast.com
            version: v1alpha1
          - description: JetJob is the Schema for the jetjobs API
            displayName: Jet Job
            kind: JetJob
            name: jetjobs.hazelcast.com
            version: v1alpha1
          - description: JetJobSnapshot is the Schema for the jetjobsnapshots API
            displayName: Jet Job Snapshot
            kind: JetJobSnapshot
            name: jetjobsnapshots.hazelcast.com
            version: v1alpha1
          - description: ManagementCenter is the Schema for the managementcenters
              API
            displayName: Management Center
            kind: ManagementCenter
            name: managementcenters.hazelcast.com
            version: v1alpha1
          - description: Map is the Schema for the maps API
            displayName: Map
            kind: Map
            name: maps.hazelcast.com
            version: v1alpha1
          - description: MultiMap is the Schema for the multimaps API
            displayName: Multi Map
            kind: MultiMap
            name: multimaps.hazelcast.com
            version: v1alpha1
          - description: Queue is the Schema for the queues API
            displayName: Queue
            kind: Queue
            name: queues.hazelcast.com
            version: v1alpha1
          - description: ReplicatedMap is the Schema for the replicatedmaps API
            displayName: Replicated Map
            kind: ReplicatedMap
            name: replicatedmaps.hazelcast.com
            version: v1alpha1
          - description: Topic is the Schema for the topics API
            displayName: Topic
            kind: Topic
            name: topics.hazelcast.com
            version: v1alpha1
          - description: UserCodeNamespace is the Schema for the usercodenamespaces
              API
            displayName: User Code Namespace
            kind: UserCodeNamespace
            name: usercodenamespaces.hazelcast.com
            version: v1alpha1
          - description: WanReplication is the Schema for the wanreplications API
            displayName: Wan Replication
            kind: WanReplication
            name: wanreplications.hazelcast.com
            version: v1alpha1
          - description: WanSync is the Schema for the wansyncs API
            displayName: Wan Sync
            kind: WanSync
            name: wansyncs.hazelcast.com
            version: v1alpha1
        description: |
          ## Deprecated
          Use the certified Hazelcast Operator instead.
          # Hazelcast Platform Operator #

          Easily deploy Hazelcast clusters and Management Center into Kubernetes environments and manage their lifecycles.

          ## Before You Start

          If you are planning to create Hazelcast Platform Enterprise clusters, you need to [create a secret](https://docs.hazelcast.com/operator/latest/get-started#step-2-start-the-hazelcast-cluster) for the license. You can request a trial license key from [here](https://trialrequest.hazelcast.com).

          For Hazelcast Platform clusters, you can simply continue.

          ## Documentation

          1. [Get started](https://docs.hazelcast.com/operator/latest/get-started) with the Operator
          2. [Connect to the cluster from outside Kubernetes](https://guides.hazelcast.org/hazelcast-platform-operator-expose-externally)
          3. [Restore a Cluster from Cloud Storage with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-external-backup-restore)
          4. [Replicate Data between Two Hazelcast Clusters with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-wan-replication)

          ## Features

          Hazelcast Platform Operator supports the features below:

          * Custom resource for Hazelcast Platform (Enterprise) and Management Center
          * Observe status of Hazelcast clusters and Management Center instances
          * High Availability Mode configuration to create clusters that are resilient to node and zone failures
          * Support for TLS connections between members using self-signed certificates
          * Scale up and down Hazelcast clusters
          * Expose Hazelcast cluster to external clients ([Smart & Unisocket](https://docs.hazelcast.com/hazelcast/latest/clients/java#java-client-operation-modes))
          * Backup Hazelcast persistence data to cloud storage with the possibility of scheduling it and restoring the data accordingly
          * WAN Replication feature when you need to synchronize multiple Hazelcast clusters, which are connected by WANs
          * Full and Delta WAN Sync support
          * Tiered Storage support for Hazelcast and Map CRs
          * CP Subsystem configuration support for Hazelcast CR
          * User Code Deployment feature, which allows you to deploy custom and domain classes from cloud storages and URLs to Hazelcast members
          * The User Code Namespaces feature allows you to deploy custom and domain classes at runtime without a members restart
          * Support the configuration of advanced networking options
          * Support Multi-namespace configuration
          * ExecutorService and EntryProcessor support
          * Support several data structures like Map, Topic, MultiMap, ReplicatedMap, Queue and Cache which can be created dynamically via specific Custom Resources
          * MapStore, Near Cache and off-heap memory (HD memory and native memory) support for the Map CR
          * Native Memory support for the Cache CR
          * Support Jet configuration and Jet Job submission using the JetJob CR
          * Support for exporting the snapshots of JetJob CRs using JetJobSnapshot CR
          * Support for custom configurations using ConfigMap
        displayName: '[DEPRECATED] Hazelcast Platform Operator'
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hazelcast
        - keyvalue
        - in-memory
        - database
        - caching
        links:
        - name: Documentation
          url: https://docs.hazelcast.com/operator/latest/
        - name: Source Code
          url: https://github.com/hazelcast/hazelcast-platform-operator
        maintainers:
        - email: cloudnative@hazelcast.com
          name: Hazelcast Cloud Native Team
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Hazelcast, Inc
        relatedImages:
        - docker.io/hazelcast/hazelcast-platform-operator@sha256:99005afda5c71ceeac200928dfed4ea35480f871053101e3aeaa44005e3fda14
        version: 5.13.1
      entries:
      - name: hazelcast-platform-operator.v5.13.1
        version: 5.13.1
      - name: hazelcast-platform-operator.v5.13.0
        version: 5.13.0
      - name: hazelcast-platform-operator.v5.12.0
        version: 5.12.0
      - name: hazelcast-platform-operator.v5.11.0
        version: 5.11.0
      - name: hazelcast-platform-operator.v5.10.0
        version: 5.10.0
      - name: hazelcast-platform-operator.v5.9.0
        version: 5.9.0
      - name: hazelcast-platform-operator.v5.8.0
        version: 5.8.0
      - name: hazelcast-platform-operator.v5.7.0
        version: 5.7.0
      - name: hazelcast-platform-operator.v5.6.0
        version: 5.6.0
      - name: hazelcast-platform-operator.v5.5.0
        version: 5.5.0
      - name: hazelcast-platform-operator.v5.4.0
        version: 5.4.0
      - name: hazelcast-platform-operator.v5.3.0
        version: 5.3.0
      - name: hazelcast-platform-operator.v5.2.0
        version: 5.2.0
      - name: hazelcast-platform-operator.v5.1.0
        version: 5.1.0
      - name: hazelcast-platform-operator.v5.0.0
        version: 5.0.0
      name: alpha
    defaultChannel: alpha
    packageName: hazelcast-platform-operator
    provider:
      name: Hazelcast, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-storage
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-storage.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Storage",
                "metadata": {
                  "name": "storage-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore"
                  },
                  "minio": {
                    "clusterDomain": "cluster.local",
                    "helmKubectlJqImage": {
                      "repository": "docker.io/bskim45/helm-kubectl-jq"
                    },
                    "image": {
                      "repository": "prod.harbor.keysight.digital/loadcore/wap-minio",
                      "tag": "RELEASE.2021-12-29T06-49-06Z-3747-37-ocp"
                    },
                    "mcImage": {
                      "repository": "prod.harbor.keysight.digital/loadcore/wap-minio-mc",
                      "tag": "RELEASE.2021-12-29T06-52-55Z-3747-37-ocp"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "securityContext": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:32:20Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Storage
            name: storages.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Storage Operator
        displayName: Keysight WAP Storage Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Storage
          url: https://keysight-wap-storage.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-storage@sha256:9f7710cfe72658cd5456fd54f86ca6be5c9a4de2a01114c0f7531b70712bd981
        version: 0.1.21
      entries:
      - name: keysight-wap-storage.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-storage
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/cluster-nfd-operator
    name: openshift-nfd-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfd.v4.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nfd.openshift.io/v1",
                "kind": "NodeFeatureDiscovery",
                "metadata": {
                  "name": "nfd-instance",
                  "namespace": "openshift-nfd"
                },
                "spec": {
                  "customConfig": {
                    "configData": "#    - name: \"more.kernel.features\"\n#      matchOn:\n#      - loadedKMod: [\"example_kmod3\"]\n#    - name: \"more.features.by.nodename\"\n#      value: customValue\n#      matchOn:\n#      - nodename: [\"special-.*-node-.*\"]\n"
                  },
                  "operand": {
                    "image": "quay.io/openshift/origin-node-feature-discovery:4.10",
                    "servicePort": 12000
                  },
                  "workerConfig": {
                    "configData": "core:\n#  labelWhiteList:\n#  noPublish: false\n  sleepInterval: 60s\n#  sources: [all]\n#  klog:\n#    addDirHeader: false\n#    alsologtostderr: false\n#    logBacktraceAt:\n#    logtostderr: true\n#    skipHeaders: false\n#    stderrthreshold: 2\n#    v: 0\n#    vmodule:\n##   NOTE: the following options are not dynamically run-time \n##          configurable and require a nfd-worker restart to take effect\n##          after being changed\n#    logDir:\n#    logFile:\n#    logFileMaxSize: 1800\n#    skipLogHeaders: false\nsources:\n#  cpu:\n#    cpuid:\n##     NOTE: whitelist has priority over blacklist\n#      attributeBlacklist:\n#        - \"BMI1\"\n#        - \"BMI2\"\n#        - \"CLMUL\"\n#        - \"CMOV\"\n#        - \"CX16\"\n#        - \"ERMS\"\n#        - \"F16C\"\n#        - \"HTT\"\n#        - \"LZCNT\"\n#        - \"MMX\"\n#        - \"MMXEXT\"\n#        - \"NX\"\n#        - \"POPCNT\"\n#        - \"RDRAND\"\n#        - \"RDSEED\"\n#        - \"RDTSCP\"\n#        - \"SGX\"\n#        - \"SSE\"\n#        - \"SSE2\"\n#        - \"SSE3\"\n#        - \"SSE4.1\"\n#        - \"SSE4.2\"\n#        - \"SSSE3\"\n#      attributeWhitelist:\n#  kernel:\n#    kconfigFile: \"/path/to/kconfig\"\n#    configOpts:\n#      - \"NO_HZ\"\n#      - \"X86\"\n#      - \"DMI\"\n  pci:\n    deviceClassWhitelist:\n      - \"0200\"\n      - \"03\"\n      - \"12\"\n    deviceLabelFields:\n#      - \"class\"\n      - \"vendor\"\n#      - \"device\"\n#      - \"subsystem_vendor\"\n#      - \"subsystem_device\"\n#  usb:\n#    deviceClassWhitelist:\n#      - \"0e\"\n#      - \"ef\"\n#      - \"fe\"\n#      - \"ff\"\n#    deviceLabelFields:\n#      - \"class\"\n#      - \"vendor\"\n#      - \"device\"\n#  custom:\n#    - name: \"my.kernel.feature\"\n#      matchOn:\n#        - loadedKMod: [\"example_kmod1\", \"example_kmod2\"]\n#    - name: \"my.pci.feature\"\n#      matchOn:\n#        - pciId:\n#            class: [\"0200\"]\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#        - pciId :\n#            vendor: [\"8086\"]\n#            device: [\"1000\", \"1100\"]\n#    - name: \"my.usb.feature\"\n#      matchOn:\n#        - usbId:\n#          class: [\"ff\"]\n#          vendor: [\"03e7\"]\n#          device: [\"2485\"]\n#        - usbId:\n#          class: [\"fe\"]\n#          vendor: [\"1a6e\"]\n#          device: [\"089a\"]\n#    - name: \"my.combined.feature\"\n#      matchOn:\n#        - pciId:\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#          loadedKMod : [\"vendor_kmod1\", \"vendor_kmod2\"]\n"
                  }
                }
              },
              {
                "apiVersion": "nfd.openshift.io/v1alpha1",
                "kind": "NodeFeatureRule",
                "metadata": {
                  "name": "my-sample-rule-object"
                },
                "spec": {
                  "rules": [
                    {
                      "labels": {
                        "my-sample-feature": "true"
                      },
                      "matchFeatures": [
                        {
                          "feature": "kernel.loadedmodule",
                          "matchExpressions": {
                            "dummy": {
                              "op": "Exists"
                            }
                          }
                        },
                        {
                          "feature": "kernel.config",
                          "matchExpressions": {
                            "X86": {
                              "op": "In",
                              "value": [
                                "y"
                              ]
                            }
                          }
                        }
                      ],
                      "name": "my sample rule"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery,OpenShift Optional
          containerImage: quay.io/openshift/origin-cluster-nfd-operator:4.10
          description: |-
            The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

            NFD consists  of the following software components:

            The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

            ##NFD-Master
            NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

            ##NFD-Worker
            NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

            ##NFD-Topology-Updater
            NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
          olm.skipRange: '>=4.6.0 <4.11.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-nfd
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/internal-objects: '["noderesourcetopologies.topology.node.k8s.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/cluster-nfd-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              The NodeFeatureDiscovery instance is the CustomResource being watched by the NFD-Operator, and holds all the needed information to setup the behaviour of the master and worker pods
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
          - description: |
              NodeFeatureRule resource specifies a configuration for feature-based customization of node objects, such as node labeling.
            kind: NodeFeatureRule
            name: nodefeaturerules.nfd.openshift.io
            version: v1alpha1
          - description: |
              NodeResourceTopology resource describes node resources and their topology.
            kind: NodeResourceTopology
            name: noderesourcetopologies.topology.node.k8s.io
            version: v1alpha1
        description: |-
          The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

          NFD consists  of the following software components:

          The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

          ##NFD-Master
          NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

          ##NFD-Worker
          NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

          ##NFD-Topology-Updater
          NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
        displayName: Node Feature Discovery Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - feature-discovery
        - feature-detection
        - node-labels
        links:
        - name: Node Feature Discovery Operator
          url: https://docs.openshift.com/container-platform/4.9/hardware_enablement/psap-node-feature-discovery-operator.html
        - name: Node Feature Discovery Documentation
          url: https://kubernetes-sigs.github.io/node-feature-discovery/stable/get-started/index.html
        maintainers:
        - email: openshift-psap@redhat.com
          name: openshift-psap
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/cluster-nfd-operator
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy
        - quay.io/openshift/origin-cluster-nfd-operator:4.10
        version: 4.10.0
      entries:
      - name: nfd.v4.10.0
        version: 4.10.0
      - name: nfd.v4.9.0
        version: 4.9.0
      - name: nfd.v4.8.0
        version: 4.8.0
      name: stable
    defaultChannel: stable
    packageName: openshift-nfd-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/cluster-nfd-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-bedrockagent-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-bedrockagent-controller.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bedrockagent.services.k8s.aws/v1alpha1",
                "kind": "Agent",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/bedrockagent-controller:0.1.1
          createdAt: "2025-09-20T01:18:56Z"
          description: AWS Bedrock Agent controller is a service controller for managing
            Bedrock Agent resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Agent represents the state of an AWS bedrockagent Agent resource.
            displayName: Agent
            kind: Agent
            name: agents.bedrockagent.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Bedrock Agent resources in AWS from within your Kubernetes cluster.

          **About Amazon Bedrock Agent**

          Amazon Bedrock is a fully managed service that makes high-performing foundation models (FMs) from leading AI companies and Amazon available for your use through a unified API. You can choose from a wide range of foundation models to find the model that is best suited for your use case. Amazon Bedrock also offers a broad set of capabilities to build generative AI applications with security, privacy, and responsible AI. Using Amazon Bedrock, you can easily experiment with and evaluate top foundation models for your use cases, privately customize them with your data using techniques such as fine-tuning and Retrieval Augmented Generation (RAG), and build agents that execute tasks using your enterprise systems and data sources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Bedrock Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bedrockagent
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Bedrock Agent Developer Resources
          url: https://aws.amazon.com/bedrock/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: bedrock-agent maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/bedrockagent-controller:0.1.1
        version: 0.1.1
      entries:
      - name: ack-bedrockagent-controller.v0.1.1
        version: 0.1.1
      - name: ack-bedrockagent-controller.v0.1.0
        version: 0.1.0
      - name: ack-bedrockagent-controller.v0.0.6
        version: 0.0.6
      - name: ack-bedrockagent-controller.v0.0.5
        version: 0.0.5
      - name: ack-bedrockagent-controller.v0.0.4
        version: 0.0.4
      - name: ack-bedrockagent-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-bedrockagent-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: devopstales
      provider-url: devopstales.github.io
    name: community-trivy-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: community-trivy-operator.v2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "trivy-operator.devopstales.io/v1",
                "kind": "NamespaceScanner",
                "metadata": {
                  "name": "trivy-operator-main-config"
                },
                "spec": {
                  "crontab": "*/5 * * * *",
                  "namespace_selector": "trivy-scan",
                  "clusterWide": "False",
                  "policyreport": "False",
                  "registry": [
                    {
                      "name": "docker.io",
                      "user": "",
                      "password": ""
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: devopstales/trivy-operator:2.4.1
          createdAt: "2022-07-11T15:00:00Z"
          description: Trivy Operator for scheduled imagescans and an Admission Control.
          repository: https://github.com/devopstales/trivy-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Confifuration for trivy-operator.
            displayName: NamespaceScanner
            kind: NamespaceScanner
            name: namespace-scanners.trivy-operator.devopstales.io
            version: v1
          - description: |
              VulnerabilityReport summarizes vulnerabilities in application dependencies and operating system packages
              built into container images.
            displayName: VulnerabilityReport
            kind: VulnerabilityReport
            name: vulnerabilityreports.trivy-operator.devopstales.io
            version: v1
        description: |
          Trivy Operator for scheduled imagescans and an Admission Control.

          ### Add Github token to extend your Github API Rate Limit
          The Operator use Github to download Trivy database. If your Github API Rate Limit Exceeded add your Github token to the operator deployment as GITHUB_TOKEN variable.

          ```yaml
          ...
          env:
          - name: GITHUB_TOKEN
            value: "githubToken"
          ```

          ### Scheduled Image scans
          Default every 5 minutes execute a scan script. It will get image list from all namespaces with the label `trivy-scan=true`, and then scan this images with trivy, finally we will get metrics on `http://[pod-ip]:9115/metrics`

          You can chane the schedule of the image scanning by setting the `crontab` variable in the namespace-scanner object.


          ### Cache Scan results
          From the 2.4 version you can use an external redis to store scanning results and speed up the scanning process.

          ```bash
          ...
          env:
          - name: REDIS_ENABLED
            value: "True"
          - name: REDIS_BACKEND
            value: redis://redis-cache:6379
          ```

          ### Persist Trivy database cache
          To Persist Trivy cache you need to create a pvc and mount to the Deployment under `/home/trivy-operator/trivy-cache`

          ```yaml
          ...
            volumeMounts:
            - name: cache
              mountPath: "/home/trivy-operator/trivy-cache"
          volumes:
          - name: cache
            persistentVolumeClaim:
              claimName: trivy-cache
          ```

          ### Trivy Image Validator
          The admission controller function can be configured as a ValidatingWebhook in a k8s cluster. Kubernetes will send requests to the admission server when a Pod creation is initiated. The admission controller checks the image using trivy if it is in a namespace with the lable `trivy-operator-validation=true`.

          To enable the admission controller you need to set the `clusterWide` variable to `true` in the namespace-scanner object.

          Enable the scheduled scan and the admission controller for a specific namespace:
          ```bash
          kubectl label namespaces test-apps trivy-scan=true
          # or
          kubectl label namespaces test-apps trivy-operator-validation=true
          ```

          You can define policy to the Admission Controller, by adding annotation to the pod trough the deployment:

          ```bash
          spec:
            ...
            template:
              metadata:
                annotations:
                  trivy.security.devopstales.io/medium: "5"
                  trivy.security.devopstales.io/low: "10"
                  trivy.security.devopstales.io/critical: "2"
          ...
          ```

          ### Monitorng
          Trivy operator has a prometheus endpoint wher yo can get the resoults of the namespace scans and the adission controller.

          ```bash
          curl -s http://10.43.179.39:9115/metrics | grep so_vulnerabilities

          # HELP so_vulnerabilities Container vulnerabilities
          # TYPE so_vulnerabilities gauge
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="UNKNOWN"} 0
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="LOW"} 23
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="MEDIUM"} 93
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="HIGH"} 76
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="CRITICAL"} 25
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="UNKNOWN"} 0
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="LOW"} 23
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="MEDIUM"} 88
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="HIGH"} 60
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="CRITICAL"} 8
          ```

          ```bash
          curl -s http://10.43.179.39:9115/metrics | grep ac_vulnerabilities

          # HELP ac_vulnerabilities Admission Controller vulnerabilities
          # TYPE ac_vulnerabilities gauge
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="UNKNOWN"} 0.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="LOW"} 83.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="MEDIUM"} 6.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="HIGH"} 6.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="CRITICAL"} 4.0
          ```

          ### Policy Report reation
          The PolicyReport object is a protopype object probosed by the Kubernetes policy work group. The Policy Report Custom Resource Definition (CRD) can be used as a common way to provide policy results to Kubernetes cluster administrators and users, using native tools.

          Add the PolicyReport CRDs to your cluster (v1alpha2):
          ```bash
          kubectl create -f https://github.com/kubernetes-sigs/wg-policy-prototypes/raw/master/policy-report/crd/v1alpha2/wgpolicyk8s.io_policyreports.yaml
          ```

          You can enable the Policy Report creatin by setting the `policyreport` to `True` in the NamespaceScanner object.

          This objects can be visualized by the Policy Reporter UI. The Policy Reporter UI is a monitoring and Observability Tool for the PolicyReport CRD with an optional UI. It is created by Kyverno. The main goal was a tool to visualize the resoluts of the Kyverno policies, but because it uses the PolicyReports CRD it can visualize the resoults of the trivy-operator scans.
        displayName: Community Trivy Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trivy
        - security
        links:
        - name: Documentation
          url: https://devopstales.github.io/trivy-operator
        - name: Blog
          url: https://devopstales.github.io
        maintainers:
        - email: devopstales@protonmail.com
          name: devopstales
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: devopstales
          url: devopstales.github.io
        relatedImages:
        - devopstales/trivy-operator:2.4.1
        version: 2.4.0
      entries:
      - name: community-trivy-operator.v2.4.0
        version: 2.4.0
      - name: community-trivy-operator.v2.3.0
        version: 2.3.0
      - name: community-trivy-operator.v2.2.0
        version: 2.2.0
      - name: community-trivy-operator.v2.1.1
        version: 2.1.1
      name: stable
    defaultChannel: stable
    packageName: community-trivy-operator
    provider:
      name: devopstales
      url: devopstales.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: www.ibm.com
    name: ibm-elasticsearch-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-elasticsearch-operator.v1.1.2667
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
          createdAt: "2025-01-30"
          description: Elasticsearch operator for use within Cloudpak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-elasticsearch-operator
          olm.skipRange: <0.0.0 >1.1.0
          operators.openshift.io/valid-subscription: '["IBM Elasticsearch"]'
          operators.operatorframework.io/internal-objects: '["elasticsearchclusters.elasticsearch.opencontent.ibm.com"]'
          operators.operatorframework.io/operator-type: non-standalone
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ElasticsearchCluster
            name: elasticsearchclusters.elasticsearch.opencontent.ibm.com
            version: v1
        description: |-
          # Introduction
          An Elasticsearch operator for use with IBM Cloudpak services.
          # Details
          ibm-elasticsearch-operator supports installing Elasticsearch major version 6 and 7. This operator supports two CRDs Elasticsearch and Elasticsearchdetails. Currently Elasticsearch is an abstraction of Elasticsearchdetails, but there is no difference between both the CRDs. Creating either of them provisions elasticsearch instances. This abstraction layer has been created to support any changes to the operator with out breaking the CRD/API for adopting products.

          ## Prerequisites
          - Openshift 3.11+
          ## Resources Required
          In short the answer is it depends. Especially if you are running Elasticsearch 6 you need 3 or a greater odd number of replicas to make a cluster and not have split brain. Each replica will have a sidecar that needs 0.1 cpu and 2gb of ram. You should size Elasticsearch based on your use-case. The default is a full cpu and 4 cpu and 64 gb ram. To assist sizing requirements we have included a copy of [esrally](https://esrally.readthedocs.io/en/stable/) on all the docker images.
          ## Installing
          ElasticsearchCluster is the strategic Kind moving forward. There are two flavors of CRD supported that will soon be deprecated. A simple Elasticseach "All in one" crd and a more detailed CRD for teams who need greater control. Upon deploy the elasticsearch is available from the service (not the headless) created by the operator. Only https will be available on port 443. the service will also be in the ElasticsearchDetails status.

          ### Case install
          To install the component using case you must either use the native install or catalog install. Upon installing the catalog you can then use either OLM or ODLM to install but must create the OperatorGroup & Subscription for either method to avoid conflict.

          # Configuration
          Architecture:
            - x86
            - s390x
            - ppc64le
          Storage: RWO Encrypted Block Storage with 120+ IOPS/GB
          Clouds:
          - IBM CLoud (ibm-block-gold, Gen2 custom iops storage class)
          - AWS (EBS)
          - Azure (For Azure please ensure you are using a higher tier of AzureDisk storage)
          - OCS (RBD not Ceph)
          - Portworx (Non-shared replacated)
          - Please do not use with NFS

          ### Limitations
          This product is not deployable standalone and it is part of the LICCR already accepted
          ## SecurityContextConstraints Requirements
          This component requires a restricted SCC.
        displayName: IBM OpenContent Elasticsearch
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - internal
        - Elasticsearch
        maintainers:
        - email: noreply@ibm.com
          name: IBM CloudpakOpen Community
        maturity: alpha
        provider:
          name: IBM
          url: www.ibm.com
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:2c39ff1850b7a9d32a11e99393039ef74ab363f3fffec1e4fd542f2ae148a6b4
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:130872119947fe91cbfb8dd82b3baaf78f7adcaca5933ca394eca3f9ee84c8cc
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:47a3745c78280f98f2ee352bb9192cdcd2fcd6d4c1a3627e7aa42178a0f107fd
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:30a5b2c4e3c231373fa18d8aaab16d3505cc11a8a5d4be36eaf99d7c0dee35f2
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:2668a562eb8df1117c40acbc3600d6735469d210e5aae4aa03bf9f54d0ebfe46
        version: 1.1.2667
      entries:
      - name: ibm-elasticsearch-operator.v1.1.2667
        version: 1.1.2667
      name: v1.1
    defaultChannel: v1.1
    packageName: ibm-elasticsearch-operator
    provider:
      name: IBM
      url: www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: ovms-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openvino-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "ModelServer",
                "metadata": {
                  "name": "model-server-sample"
                },
                "spec": {
                  "image_name": "registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0",
                  "deployment_parameters": {
                    "replicas": 1
                  },
                  "service_parameters": {
                    "grpc_port": 8080,
                    "rest_port": 8081
                  },
                  "models_settings": {
                    "single_model_mode": true,
                    "config_configmap_name": "",
                    "config_path": "",
                    "model_name": "resnet",
                    "model_path": "gs://<bucket_name>/<model_dir>",
                    "nireq": 0,
                    "plugin_config": "{\"PERFORMANCE_HINT\":\"LATENCY\"}",
                    "batch_size": "",
                    "shape": "",
                    "model_version_policy": "{\"latest\": { \"num_versions\":1 }}",
                    "layout": "",
                    "target_device": "CPU",
                    "is_stateful": false,
                    "idle_sequence_cleanup": false,
                    "low_latency_transformation": true,
                    "max_sequence_number": 0
                  },
                  "server_settings": {
                    "file_system_poll_wait_seconds": 0,
                    "sequence_cleaner_poll_wait_minutes": 0,
                    "log_level": "INFO",
                    "grpc_workers": 1,
                    "rest_workers": 0
                  },
                  "models_repository": {
                    "https_proxy": "",
                    "http_proxy": "",
                    "storage_type": "google",
                    "models_host_path": "",
                    "models_volume_claim": "",
                    "aws_secret_access_key": "",
                    "aws_access_key_id": "",
                    "aws_region": "",
                    "s3_compat_api_endpoint": "",
                    "gcp_creds_secret_name": "",
                    "azure_storage_connection_string": ""
                  }
                }
              },
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "Notebook",
                "metadata": {
                  "name": "v2024.1"
                },
                "spec": {
                  "git_ref": "2024.1",
                  "auto_update_image": false
                }
              }
            ]
          alm-examples-metadata: '{ "model-server-sample":{"description":"Example
            of the model server configuration"}, "v2024":{"description":"Default notebook
            configuration"} }'
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
          createdAt: "2024-05-14T09:29:35Z"
          description: OpenVINO Toolkit Operator manages OpenVINO components in OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.34.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Provides inference serving through gRPC & REST
            displayName: Model Server
            kind: ModelServer
            name: modelservers.intel.com
            version: v1alpha1
          - description: Adds a collection of ready-to-run Jupyter notebooks for learning
              and experimenting with the OpenVINO™ Toolkit to Red Hat® OpenShift®
              Data Science
            displayName: Notebook
            kind: Notebook
            name: notebooks.intel.com
            version: v1alpha1
        description: |
          OpenVINO Toolkit Operator manages OpenVINO components in OpenShift.

          Currently there available components are ModelServer and Notebook.

          # Model Server
          [OpenVINO™ Model Server](https://github.com/openvinotoolkit/model_server) (OVMS) is a scalable, high-performance solution for serving machine learning models optimized for Intel® architectures. The server provides an inference service via gRPC or REST API for any models trained in a framework that is supported by [OpenVINO](https://software.intel.com/en-us/openvino-toolki://docs.openvino.ai/latest/index.html).
          Model Server configuration parameters is explained [here](https://github.com/openvinotoolkit/operator/blob/main/docs/modelserver_params.md).
          ## Using the cluster
          OpenVINO Model Server can be consumed as a `Service`. It is called like with `ModelServer` resource with `-ovms` suffix.
          The suffix is omitted when `ovms` phrase is included in the name.
          The service exposes gRPC and REST API interfaces to run inference requests.
          ```
          oc get pods
          NAME                                        READY   STATUS    RESTARTS   AGE
          model-server-sample-ovms-586f6f76df-dpps4   1/1     Running   0          8h

          oc get services
          NAME                       TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)             AGE
          model-server-sample-ovms   ClusterIP   172.25.199.210   <none>        8080/TCP,8081/TCP   8h
          ```

          # Notebook
          The Notebook resource integrates JupyterHub from OpenShift Data Science or Open Data Hub with a container image that includes [developer
          tools](https://github.com/openvinotoolkit/openvino/blob/master/docs/install_guides/pypi-openvino-dev.md)
          from the OpenVINO toolkit and a set of Jupyter notebook tutorials. It enables selecting a defined image `OpenVINO™ Toolkit` from
          the Jupyter Spawner choice list.

          Create the `Notebook` resource in the same project with JupyterHub and RedHat OpenShift Data Science operator as seen in [documentation](https://github.com/openvinotoolkit/operator/blob/main/docs/notebook_in_rhods.md).
          It builds the image in the cluster based on Dockerfile from [openvino_notebooks](https://github.com/openvinotoolkit/openvino_notebooks).

          ## References
          OpenVINO Model Server on [Github](https://github.com/openvinotoolkit/model_server)

          OpenVINO Model Server Operator on [Github](https://github.com/openvinotoolkit/operator)
        displayName: OpenVINO Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Inference OpenVINO
        links:
        - name: OpenVINO Toolkit Operator
          url: https://github.com/openvinotoolkit/operator
        - name: OpenVINO Model Server
          url: https://github.com/openvinotoolkit/model_server
        - name: OpenVINO Notebooks
          url: https://github.com/openvinotoolkit/openvino_notebooks
        maintainers:
        - email: openvino-operator-support@intel.com
          name: Operator Support
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5714b93ac39277affdc10c3ba142d8c032ed0ddb2ba61c04cbddac710464c4
        - registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
        - registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0
        version: 1.2.0
      entries:
      - name: openvino-operator.v1.2.0
        version: 1.2.0
      - name: openvino-operator.v1.1.0
        version: 1.1.0
      - name: openvino-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: ovms-operator
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx-operator-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx-openshift-operator.v0.0.33
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
          createdAt: "2019-07-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx-operator-rhmp/support?utm_source=openshift_console
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX on JBoss EAP 7
            displayName: Joget DX on JBoss EAP 7
            kind: JogetDX
            name: jogetdx.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX7/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
        - registry.connect.redhat.com/joget/joget-dx7-eap7@sha256:7245501f5fccf011d75c034b7d55542cfe3bd2c64cc7e48b3224a48b623af282
        version: 0.0.33
      entries:
      - name: joget-dx-openshift-operator.v0.0.33
        version: 0.0.33
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx-operator-rhmp
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-codeartifact-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-codeartifact-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/codeartifact-controller:1.1.1
          createdAt: "2025-09-19T22:00:24Z"
          description: AWS codeartifact controller is a service controller for managing
            codeartifact resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Domain represents the state of an AWS codeartifact Domain
              resource.
            displayName: Domain
            kind: Domain
            name: domains.codeartifact.services.k8s.aws
            version: v1alpha1
          - description: PackageGroup represents the state of an AWS codeartifact
              PackageGroup resource.
            displayName: PackageGroup
            kind: PackageGroup
            name: packagegroups.codeartifact.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon codeartifact resources in AWS from within your Kubernetes cluster.

          **About Amazon codeartifact**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon codeartifact
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - codeartifact
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon codeartifact Developer Resources
          url: https://aws.amazon.com/codeartifact/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: codeartifact maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/codeartifact-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-codeartifact-controller.v1.1.1
        version: 1.1.1
      - name: ack-codeartifact-controller.v1.1.0
        version: 1.1.0
      - name: ack-codeartifact-controller.v1.0.13
        version: 1.0.13
      - name: ack-codeartifact-controller.v1.0.11
        version: 1.0.11
      name: alpha
    defaultChannel: alpha
    packageName: ack-codeartifact-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kepler Operator Contributors
      provider-url: https://sustainable-computing.io/
    name: kepler-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kepler-operator.v0.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "Kepler",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "kepler",
                    "app.kubernetes.io/name": "kepler",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "kepler"
                },
                "spec": {
                  "exporter": {
                    "deployment": {
                      "port": 9103
                    },
                    "redfish": null
                  }
                }
              },
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "PowerMonitor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "powermonitor",
                    "app.kubernetes.io/name": "powermonitor",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "power-monitor"
                },
                "spec": {
                  "kepler": {
                    "config": {
                      "logLevel": "info"
                    },
                    "deployment": {
                      "security": {
                        "mode": "none"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: quay.io/sustainable_computing_io/kepler-operator:0.20.0
          createdAt: "2025-07-18T06:13:30Z"
          description: Deploys and Manages Kepler on Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: |-
            [
              "keplerinternals.kepler.system.sustainable.computing.io",
              "powermonitorinternals.kepler.system.sustainable.computing.io"
            ]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/sustainable-computing-io/kepler-operator
          support: https://github.com/sustainable-computing-io/kepler-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KeplerInternal is the Schema for internal/unsupported API
            displayName: KeplerInternal
            kind: KeplerInternal
            name: keplerinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: Kepler is deprecated and will be removed in a future release.
              Please use PowerMonitor instead.
            displayName: Kepler
            kind: Kepler
            name: keplers.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitorInternal is the Schema for internal/unsupported
              API
            displayName: PowerMonitorInternal
            kind: PowerMonitorInternal
            name: powermonitorinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitor is the Schema for the powermonitors API
            displayName: PowerMonitor
            kind: PowerMonitor
            name: powermonitors.kepler.system.sustainable.computing.io
            version: v1alpha1
        description: |-
          Kepler (Kubernetes-based Efficient Power Level Exporter) is a Prometheus exporter that
          measures energy consumption metrics at the container, pod, and node level in Kubernetes clusters.

          ### Deprecation Notice
          **Important:** The Kepler CR (Custom Resource) is deprecated and will be removed in a future release. Please use the new PowerMonitor instead.

          ### Prerequisites
          To view the metrics exposed by Kepler, **[User Workload Monitoring](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/monitoring/configuring-user-workload-monitoring)**
          MUST be enabled in your cluster.

          ### Documentation
          * [User Documentation](https://github.com/sustainable-computing-io/kepler-operator/blob/v1alpha1/README.md)
          * [Developer Guide](https://github.com/sustainable-computing-io/kepler-operator/tree/v1alpha1/docs/developer)

          ### Contributing
          * [Report Bugs](https://github.com/sustainable-computing-io/kepler-operator/issues/new?template=bug_report.md) - Found a bug? Let us know!
          * [Browse Issues](https://github.com/sustainable-computing-io/kepler-operator/issues) - See what we're working on

          ### License
          Licensed under [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)
        displayName: Kepler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kepler
        - power
        - energy
        - monitoring
        links:
        - name: Installation Guide
          url: https://sustainable-computing.io/installation/community-operator/
        - name: Kepler Project
          url: https://sustainable-computing.io/
        maintainers:
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: vprashar@redhat.com
          name: Vibhu Prashar
        - email: vimalkum@redhat.com
          name: Vimal Kumar
        - email: kaliu@redhat.com
          name: Kaiyi Liu
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Kepler Operator Contributors
          url: https://sustainable-computing.io/
        relatedImages:
        - quay.io/sustainable_computing_io/kepler:v0.10.2
        - quay.io/brancz/kube-rbac-proxy:v0.19.0
        - quay.io/sustainable_computing_io/kepler-operator:0.20.0
        version: 0.20.0
      entries:
      - name: kepler-operator.v0.20.0
        version: 0.20.0
      - name: kepler-operator.v0.19.0
        version: 0.19.0
      - name: kepler-operator.v0.18.0
        version: 0.18.0
      - name: kepler-operator.v0.17.0
        version: 0.17.0
      - name: kepler-operator.v0.16.0
        version: 0.16.0
      - name: kepler-operator.v0.15.0
        version: 0.15.0
      name: alpha
    defaultChannel: alpha
    packageName: kepler-operator
    provider:
      name: Kepler Operator Contributors
      url: https://sustainable-computing.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MuleSoft
      provider-url: anypoint.mulesoft.com
    name: runtime-fabric-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runtime-fabric-operator.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rtf.operators.mulesoft.com/v1",
                "kind": "RuntimeFabric",
                "metadata": {
                  "name": "runtime-fabric"
                },
                "spec": {
                  "activationData": null,
                  "agent": {
                    "replicas": 1,
                    "app": {
                      "resources": {
                        "limits": {
                          "cpu": "1000m",
                          "memory": "500Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      }
                    },
                    "rtfd": {
                      "resources": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "150m",
                          "memory": "200Mi"
                        }
                      },
                      "watch": {
                        "enabled": true,
                        "components": {
                          "deployments": {
                            "enabled": true
                          },
                          "ingress": {
                            "enabled": true
                          },
                          "secrets": {
                            "enabled": true
                          },
                          "persistencegateways": {
                            "enabled": true
                          },
                          "nodes": {
                            "enabled": true
                          },
                          "jobs": {
                            "enabled": true
                          },
                          "routes": {
                            "enabled": true
                          }
                        }
                      }
                    }
                  },
                  "resource-cache": {
                    "replicas": 1,
                    "nginx": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "250Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    },
                    "fetcher": {
                      "resources": {
                        "limits": {
                          "cpu": "50m",
                          "memory": "100Mi"
                        },
                        "requests": {
                          "cpu": "50m",
                          "memory": "50Mi"
                        }
                      }
                    }
                  },
                  "mule-clusterip-service": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "global": {
                    "cluster": {
                      "flavor": "byok",
                      "type": "openshift"
                    },
                    "containerLogPaths": [
                      "/var/lib/docker/containers",
                      "/var/log/containers",
                      "/var/log/pods"
                    ],
                    "core": {
                      "affinity": {},
                      "tolerations": [],
                      "topologySpreadConstraints": [
                        {
                          "maxSkew": 1,
                          "topologyKey": "topology.kubernetes.io/zone",
                          "whenUnsatisfiable": "ScheduleAnyway"
                        }
                      ]
                    },
                    "worker": {
                      "affinity": {},
                      "tolerations": []
                    },
                    "deployment": {
                      "nodeSelector": ""
                    },
                    "image": {
                      "provider": "aws",
                      "pullSecretName": "rtf-pull-secret",
                      "rtfRegistry": "rtf-runtime-registry.kprod.msap.io",
                      "repositoryNamespace": "mulesoft"
                    },
                    "service": {
                      "clusterIp": null
                    },
                    "authorizedNamespaces": false,
                    "fipsEnabled": false,
                    "crds": {
                      "install": true,
                      "keep": true,
                      "flex": {
                        "enabled": false
                      }
                    },
                    "nodeWatcherEnabled": true,
                    "nodeReconciliation": true,
                    "deploymentRateLimitPerSecond": 1
                  },
                  "muleLicense": null,
                  "proxy": {
                    "http_no_proxy": null,
                    "http_proxy": null,
                    "monitoring_proxy": null
                  },
                  "customLog4jEnabled": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
          createdAt: "2025-07-31T21:12:41Z"
          description: Runtime Fabric Operator Application from Mulesoft
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mulesoft/rtf-ocp-operator
          support: MuleSoft
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Runtime Fabric Operator from Mulesoft
              **Activation data**
              Used to install Runtime Fabric on a Kubernetes service. Copy this value from the Runtime Fabric setup page in Runtime Manager.
              **Registry URL**
              The registry from which to retrieve the images needed to install and run Runtime Fabric. If you're using a local registry, specify the URL here.
                * Default registry URLs:
                    - US control plane: rtf-runtime-registry.kprod.msap.io
                    - EU control plane: rtf-runtime-registry.kprod-eu.msap.io
              **Registry pull secret**
              Used to retrieve images from the registry. Change this value only if you did so during the operator installation.
              **Mule license key**
              Your company’s Mule license key. The key is not required during install, but it is required to deploy applications to Runtime Fabric. Must be Base64 encoded.
            displayName: Runtime Fabric
            kind: RuntimeFabric
            name: runtimefabrics.rtf.operators.mulesoft.com
            version: v1
        description: |-
          Anypoint Runtime Fabric is a container service that you install in your existing Red Hat OpenShift cluster that enables you to run Mule applications, composite APIs, and API gateways across any environment with centralized management via Anypoint Platform.
          Capabilities of Runtime Fabric include:
          * Isolation between applications by running a separate Mule runtime server per application.
          * Ability to run multiple versions of Mule runtime server on the same set of resources.
          * Scaling applications across multiple replicas.
          * Automated application fail-over.
          * Application management with Anypoint Runtime Manager.
          Runtime Fabric supports the following Red Hat OpenShift deployment options:
          * Red Hat OpenShift Service on AWS
          * Microsoft Azure Red Hat OpenShift
          * Red Hat OpenShift Dedicated
          * Red Hat OpenShift on IBM Cloud
          * Self-managed Red Hat OpenShift editions (Performance Plus, OCP, Kubernetes engine)
          Runtime Fabric supports Red Hat OpenShift versions 4.16 to 4.19.
        displayName: Runtime Fabric Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - agent
        - rtf
        - runtime
        - fabric
        - mulesoft
        links:
        - name: Runtime Fabric Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/62901512fefb1b137ba23946
        - name: Documentation
          url: https://docs.mulesoft.com/release-notes/runtime-fabric/runtime-fabric
        maintainers:
        - email: support@mulesoft.com
          name: Mulesoft
        maturity: stable
        provider:
          name: MuleSoft
          url: anypoint.mulesoft.com
        relatedImages:
        - quay.io/rtf/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
        version: 2.11.0
      entries:
      - name: runtime-fabric-operator.v2.11.0
        version: 2.11.0
      - name: runtime-fabric-operator.v2.10.34
        version: 2.10.34
      - name: runtime-fabric-operator.v2.10.17
        version: 2.10.17
      - name: runtime-fabric-operator.v2.10.0
        version: 2.10.0
      - name: runtime-fabric-operator.v2.9.2
        version: 2.9.2
      - name: runtime-fabric-operator.v2.8.2
        version: 2.8.2
      - name: runtime-fabric-operator.v2.7.0
        version: 2.7.0
      - name: runtime-fabric-operator.v2.6.52
        version: 2.6.52
      name: alpha
    - currentCSV: runtime-fabric-operator.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rtf.operators.mulesoft.com/v1",
                "kind": "RuntimeFabric",
                "metadata": {
                  "name": "runtime-fabric"
                },
                "spec": {
                  "activationData": null,
                  "agent": {
                    "replicas": 1,
                    "app": {
                      "resources": {
                        "limits": {
                          "cpu": "1000m",
                          "memory": "500Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      }
                    },
                    "rtfd": {
                      "resources": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "150m",
                          "memory": "200Mi"
                        }
                      },
                      "watch": {
                        "enabled": true,
                        "components": {
                          "deployments": {
                            "enabled": true
                          },
                          "ingress": {
                            "enabled": true
                          },
                          "secrets": {
                            "enabled": true
                          },
                          "persistencegateways": {
                            "enabled": true
                          },
                          "nodes": {
                            "enabled": true
                          },
                          "jobs": {
                            "enabled": true
                          },
                          "routes": {
                            "enabled": true
                          }
                        }
                      }
                    }
                  },
                  "resource-cache": {
                    "replicas": 1,
                    "nginx": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "250Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    },
                    "fetcher": {
                      "resources": {
                        "limits": {
                          "cpu": "50m",
                          "memory": "100Mi"
                        },
                        "requests": {
                          "cpu": "50m",
                          "memory": "50Mi"
                        }
                      }
                    }
                  },
                  "mule-clusterip-service": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "global": {
                    "cluster": {
                      "flavor": "byok",
                      "type": "openshift"
                    },
                    "containerLogPaths": [
                      "/var/lib/docker/containers",
                      "/var/log/containers",
                      "/var/log/pods"
                    ],
                    "core": {
                      "affinity": {},
                      "tolerations": [],
                      "topologySpreadConstraints": [
                        {
                          "maxSkew": 1,
                          "topologyKey": "topology.kubernetes.io/zone",
                          "whenUnsatisfiable": "ScheduleAnyway"
                        }
                      ]
                    },
                    "worker": {
                      "affinity": {},
                      "tolerations": []
                    },
                    "deployment": {
                      "nodeSelector": ""
                    },
                    "image": {
                      "provider": "aws",
                      "pullSecretName": "rtf-pull-secret",
                      "rtfRegistry": "rtf-runtime-registry.kprod.msap.io",
                      "repositoryNamespace": "mulesoft"
                    },
                    "service": {
                      "clusterIp": null
                    },
                    "authorizedNamespaces": false,
                    "fipsEnabled": false,
                    "crds": {
                      "install": true,
                      "keep": true,
                      "flex": {
                        "enabled": false
                      }
                    },
                    "nodeWatcherEnabled": true,
                    "nodeReconciliation": true,
                    "deploymentRateLimitPerSecond": 1
                  },
                  "muleLicense": null,
                  "proxy": {
                    "http_no_proxy": null,
                    "http_proxy": null,
                    "monitoring_proxy": null
                  },
                  "customLog4jEnabled": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
          createdAt: "2025-07-31T21:12:41Z"
          description: Runtime Fabric Operator Application from Mulesoft
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mulesoft/rtf-ocp-operator
          support: MuleSoft
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Runtime Fabric Operator from Mulesoft
              **Activation data**
              Used to install Runtime Fabric on a Kubernetes service. Copy this value from the Runtime Fabric setup page in Runtime Manager.
              **Registry URL**
              The registry from which to retrieve the images needed to install and run Runtime Fabric. If you're using a local registry, specify the URL here.
                * Default registry URLs:
                    - US control plane: rtf-runtime-registry.kprod.msap.io
                    - EU control plane: rtf-runtime-registry.kprod-eu.msap.io
              **Registry pull secret**
              Used to retrieve images from the registry. Change this value only if you did so during the operator installation.
              **Mule license key**
              Your company’s Mule license key. The key is not required during install, but it is required to deploy applications to Runtime Fabric. Must be Base64 encoded.
            displayName: Runtime Fabric
            kind: RuntimeFabric
            name: runtimefabrics.rtf.operators.mulesoft.com
            version: v1
        description: |-
          Anypoint Runtime Fabric is a container service that you install in your existing Red Hat OpenShift cluster that enables you to run Mule applications, composite APIs, and API gateways across any environment with centralized management via Anypoint Platform.
          Capabilities of Runtime Fabric include:
          * Isolation between applications by running a separate Mule runtime server per application.
          * Ability to run multiple versions of Mule runtime server on the same set of resources.
          * Scaling applications across multiple replicas.
          * Automated application fail-over.
          * Application management with Anypoint Runtime Manager.
          Runtime Fabric supports the following Red Hat OpenShift deployment options:
          * Red Hat OpenShift Service on AWS
          * Microsoft Azure Red Hat OpenShift
          * Red Hat OpenShift Dedicated
          * Red Hat OpenShift on IBM Cloud
          * Self-managed Red Hat OpenShift editions (Performance Plus, OCP, Kubernetes engine)
          Runtime Fabric supports Red Hat OpenShift versions 4.16 to 4.19.
        displayName: Runtime Fabric Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - agent
        - rtf
        - runtime
        - fabric
        - mulesoft
        links:
        - name: Runtime Fabric Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/62901512fefb1b137ba23946
        - name: Documentation
          url: https://docs.mulesoft.com/release-notes/runtime-fabric/runtime-fabric
        maintainers:
        - email: support@mulesoft.com
          name: Mulesoft
        maturity: stable
        provider:
          name: MuleSoft
          url: anypoint.mulesoft.com
        relatedImages:
        - quay.io/rtf/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
        version: 2.11.0
      entries:
      - name: runtime-fabric-operator.v2.11.0
        version: 2.11.0
      - name: runtime-fabric-operator.v2.10.34
        version: 2.10.34
      - name: runtime-fabric-operator.v2.10.17
        version: 2.10.17
      - name: runtime-fabric-operator.v2.10.0
        version: 2.10.0
      - name: runtime-fabric-operator.v2.9.2
        version: 2.9.2
      - name: runtime-fabric-operator.v2.8.2
        version: 2.8.2
      - name: runtime-fabric-operator.v2.7.0
        version: 2.7.0
      - name: runtime-fabric-operator.v2.6.52
        version: 2.6.52
      name: stable
    defaultChannel: stable
    packageName: runtime-fabric-operator
    provider:
      name: MuleSoft
      url: anypoint.mulesoft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-dynamodb-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-dynamodb-controller.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "GlobalTable",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "Table",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/dynamodb-controller:1.5.2
          createdAt: "2025-09-19T19:19:47Z"
          description: AWS DynamoDB controller is a service controller for managing
            DynamoDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup represents the state of an AWS dynamodb Backup resource.
            displayName: Backup
            kind: Backup
            name: backups.dynamodb.services.k8s.aws
            version: v1alpha1
          - description: GlobalTable represents the state of an AWS dynamodb GlobalTable
              resource.
            displayName: GlobalTable
            kind: GlobalTable
            name: globaltables.dynamodb.services.k8s.aws
            version: v1alpha1
          - description: Table represents the state of an AWS dynamodb Table resource.
            displayName: Table
            kind: Table
            name: tables.dynamodb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon DynamoDB resources in AWS from within your Kubernetes cluster.

          **About Amazon DynamoDB**

          Amazon DynamoDB is a fully managed NoSQL database service that provides fast and predictable performance with seamless scalability. DynamoDB lets you offload the administrative burdens of operating and scaling a distributed database so that you don't have to worry about hardware provisioning, setup and configuration, replication, software patching, or cluster scaling. DynamoDB also offers encryption at rest, which eliminates the operational burden and complexity involved in protecting sensitive data. For more information, see [DynamoDB Encryption at Rest](https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/EncryptionAtRest.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon DynamoDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dynamodb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon DynamoDB Developer Resources
          url: https://aws.amazon.com/dynamodb/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: dynamodb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/dynamodb-controller:1.5.2
        version: 1.5.2
      entries:
      - name: ack-dynamodb-controller.v1.5.2
        version: 1.5.2
      - name: ack-dynamodb-controller.v1.5.1
        version: 1.5.1
      - name: ack-dynamodb-controller.v1.5.0
        version: 1.5.0
      - name: ack-dynamodb-controller.v1.4.10
        version: 1.4.10
      - name: ack-dynamodb-controller.v1.4.9
        version: 1.4.9
      - name: ack-dynamodb-controller.v1.4.7
        version: 1.4.7
      - name: ack-dynamodb-controller.v1.4.6
        version: 1.4.6
      - name: ack-dynamodb-controller.v1.4.5
        version: 1.4.5
      - name: ack-dynamodb-controller.v1.4.4
        version: 1.4.4
      - name: ack-dynamodb-controller.v1.4.3
        version: 1.4.3
      - name: ack-dynamodb-controller.v1.4.2
        version: 1.4.2
      - name: ack-dynamodb-controller.v1.4.1
        version: 1.4.1
      - name: ack-dynamodb-controller.v1.3.0
        version: 1.3.0
      - name: ack-dynamodb-controller.v1.2.23
        version: 1.2.23
      - name: ack-dynamodb-controller.v1.2.22
        version: 1.2.22
      - name: ack-dynamodb-controller.v1.2.21
        version: 1.2.21
      - name: ack-dynamodb-controller.v1.2.20
        version: 1.2.20
      - name: ack-dynamodb-controller.v1.2.15
        version: 1.2.15
      - name: ack-dynamodb-controller.v1.2.14
        version: 1.2.14
      - name: ack-dynamodb-controller.v1.2.13
        version: 1.2.13
      - name: ack-dynamodb-controller.v1.2.12
        version: 1.2.12
      - name: ack-dynamodb-controller.v1.2.11
        version: 1.2.11
      - name: ack-dynamodb-controller.v1.2.9
        version: 1.2.9
      - name: ack-dynamodb-controller.v1.2.8
        version: 1.2.8
      - name: ack-dynamodb-controller.v1.2.6
        version: 1.2.6
      - name: ack-dynamodb-controller.v1.2.5
        version: 1.2.5
      - name: ack-dynamodb-controller.v1.2.4
        version: 1.2.4
      - name: ack-dynamodb-controller.v1.2.2
        version: 1.2.2
      - name: ack-dynamodb-controller.v1.2.1
        version: 1.2.1
      - name: ack-dynamodb-controller.v1.2.0
        version: 1.2.0
      - name: ack-dynamodb-controller.v1.1.1
        version: 1.1.1
      - name: ack-dynamodb-controller.v1.1.0
        version: 1.1.0
      - name: ack-dynamodb-controller.v1.0.3
        version: 1.0.3
      - name: ack-dynamodb-controller.v1.0.2
        version: 1.0.2
      - name: ack-dynamodb-controller.v1.0.1
        version: 1.0.1
      - name: ack-dynamodb-controller.v1.0.0
        version: 1.0.0
      - name: ack-dynamodb-controller.v0.1.10
        version: 0.1.10
      - name: ack-dynamodb-controller.v0.1.9
        version: 0.1.9
      - name: ack-dynamodb-controller.v0.1.8
        version: 0.1.8
      - name: ack-dynamodb-controller.v0.1.7
        version: 0.1.7
      - name: ack-dynamodb-controller.v0.1.6
        version: 0.1.6
      - name: ack-dynamodb-controller.v0.1.5
        version: 0.1.5
      - name: ack-dynamodb-controller.v0.1.4
        version: 0.1.4
      - name: ack-dynamodb-controller.v0.1.3
        version: 0.1.3
      - name: ack-dynamodb-controller.v0.1.1
        version: 0.1.1
      - name: ack-dynamodb-controller.v0.1.0
        version: 0.1.0
      - name: ack-dynamodb-controller.v0.0.20
        version: 0.0.20
      - name: ack-dynamodb-controller.v0.0.19
        version: 0.0.19
      - name: ack-dynamodb-controller.v0.0.18
        version: 0.0.18
      - name: ack-dynamodb-controller.v0.0.17
        version: 0.0.17
      - name: ack-dynamodb-controller.v0.0.16
        version: 0.0.16
      - name: ack-dynamodb-controller.v0.0.15
        version: 0.0.15
      - name: ack-dynamodb-controller.v0.0.14
        version: 0.0.14
      - name: ack-dynamodb-controller.v0.0.13
        version: 0.0.13
      - name: ack-dynamodb-controller.v0.0.12
        version: 0.0.12
      - name: ack-dynamodb-controller.v0.0.11
        version: 0.0.11
      - name: ack-dynamodb-controller.v0.0.10
        version: 0.0.10
      - name: ack-dynamodb-controller.v0.0.9
        version: 0.0.9
      - name: ack-dynamodb-controller.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: ack-dynamodb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-elasticache-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-elasticache-controller.v1.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "CacheParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "CacheSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "ReplicationGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "Snapshot",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "UserGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/elasticache-controller:1.2.3
          createdAt: "2025-09-19T19:19:33Z"
          description: AWS ElastiCache controller is a service controller for managing
            ElastiCache resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CacheCluster represents the state of an AWS elasticache CacheCluster
              resource.
            displayName: CacheCluster
            kind: CacheCluster
            name: cacheclusters.elasticache.services.k8s.aws
            version: v1alpha1
          - description: CacheParameterGroup represents the state of an AWS elasticache
              CacheParameterGroup resource.
            displayName: CacheParameterGroup
            kind: CacheParameterGroup
            name: cacheparametergroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: CacheSubnetGroup represents the state of an AWS elasticache
              CacheSubnetGroup resource.
            displayName: CacheSubnetGroup
            kind: CacheSubnetGroup
            name: cachesubnetgroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ReplicationGroup represents the state of an AWS elasticache
              ReplicationGroup resource.
            displayName: ReplicationGroup
            kind: ReplicationGroup
            name: replicationgroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCache represents the state of an AWS elasticache
              ServerlessCache resource.
            displayName: ServerlessCache
            kind: ServerlessCache
            name: serverlesscaches.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCacheSnapshot represents the state of an AWS elasticache
              ServerlessCacheSnapshot resource.
            displayName: ServerlessCacheSnapshot
            kind: ServerlessCacheSnapshot
            name: serverlesscachesnapshots.elasticache.services.k8s.aws
            version: v1alpha1
          - description: Snapshot represents the state of an AWS elasticache Snapshot
              resource.
            displayName: Snapshot
            kind: Snapshot
            name: snapshots.elasticache.services.k8s.aws
            version: v1alpha1
          - description: UserGroup represents the state of an AWS elasticache UserGroup
              resource.
            displayName: UserGroup
            kind: UserGroup
            name: usergroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS elasticache User resource.
            displayName: User
            kind: User
            name: users.elasticache.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ElastiCache resources in AWS from within your Kubernetes cluster.

          **About Amazon ElastiCache**

          Amazon ElastiCache allows you to seamlessly set up, run, and scale popular open-source compatible in-memory data stores in the cloud. Build data-intensive apps or boost the performance of your existing databases by retrieving data from high throughput and low latency in-memory data stores. Amazon ElastiCache is a popular choice for real-time use cases like Caching, Session Stores, Gaming, Geospatial Services, Real-Time Analytics, and Queuing.

          Amazon ElastiCache offers fully managed Redis, voted the most loved database by developers in the Stack Overflow 2020 Developer Survey, and Memcached for your most demanding applications that require sub-millisecond response times.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ElastiCache
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticache
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ElastiCache Developer Resources
          url: https://aws.amazon.com/elasticache/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: elasticache maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/elasticache-controller:1.2.3
        version: 1.2.3
      entries:
      - name: ack-elasticache-controller.v1.2.3
        version: 1.2.3
      - name: ack-elasticache-controller.v1.2.2
        version: 1.2.2
      - name: ack-elasticache-controller.v1.2.1
        version: 1.2.1
      - name: ack-elasticache-controller.v1.2.0
        version: 1.2.0
      - name: ack-elasticache-controller.v1.1.0
        version: 1.1.0
      - name: ack-elasticache-controller.v1.0.7
        version: 1.0.7
      - name: ack-elasticache-controller.v1.0.6
        version: 1.0.6
      - name: ack-elasticache-controller.v1.0.5
        version: 1.0.5
      - name: ack-elasticache-controller.v1.0.4
        version: 1.0.4
      - name: ack-elasticache-controller.v1.0.3
        version: 1.0.3
      - name: ack-elasticache-controller.v1.0.2
        version: 1.0.2
      - name: ack-elasticache-controller.v1.0.1
        version: 1.0.1
      - name: ack-elasticache-controller.v1.0.0
        version: 1.0.0
      - name: ack-elasticache-controller.v0.2.6
        version: 0.2.6
      - name: ack-elasticache-controller.v0.2.5
        version: 0.2.5
      - name: ack-elasticache-controller.v0.2.4
        version: 0.2.4
      - name: ack-elasticache-controller.v0.2.3
        version: 0.2.3
      - name: ack-elasticache-controller.v0.2.2
        version: 0.2.2
      - name: ack-elasticache-controller.v0.2.1
        version: 0.2.1
      - name: ack-elasticache-controller.v0.2.0
        version: 0.2.0
      - name: ack-elasticache-controller.v0.1.1
        version: 0.1.1
      - name: ack-elasticache-controller.v0.1.0
        version: 0.1.0
      - name: ack-elasticache-controller.v0.0.31
        version: 0.0.31
      - name: ack-elasticache-controller.v0.0.30
        version: 0.0.30
      - name: ack-elasticache-controller.v0.0.29
        version: 0.0.29
      - name: ack-elasticache-controller.v0.0.28
        version: 0.0.28
      - name: ack-elasticache-controller.v0.0.27
        version: 0.0.27
      - name: ack-elasticache-controller.v0.0.26
        version: 0.0.26
      - name: ack-elasticache-controller.v0.0.25
        version: 0.0.25
      - name: ack-elasticache-controller.v0.0.24
        version: 0.0.24
      - name: ack-elasticache-controller.v0.0.23
        version: 0.0.23
      - name: ack-elasticache-controller.v0.0.22
        version: 0.0.22
      - name: ack-elasticache-controller.v0.0.21
        version: 0.0.21
      - name: ack-elasticache-controller.v0.0.20
        version: 0.0.20
      - name: ack-elasticache-controller.v0.0.19
        version: 0.0.19
      - name: ack-elasticache-controller.v0.0.18
        version: 0.0.18
      - name: ack-elasticache-controller.v0.0.17
        version: 0.0.17
      - name: ack-elasticache-controller.v0.0.16
        version: 0.0.16
      - name: ack-elasticache-controller.v0.0.15
        version: 0.0.15
      - name: ack-elasticache-controller.v0.0.14
        version: 0.0.14
      - name: ack-elasticache-controller.v0.0.13
        version: 0.0.13
      - name: ack-elasticache-controller.v0.0.12
        version: 0.0.12
      - name: ack-elasticache-controller.v0.0.11
        version: 0.0.11
      - name: ack-elasticache-controller.v0.0.10
        version: 0.0.10
      - name: ack-elasticache-controller.v0.0.9
        version: 0.0.9
      - name: ack-elasticache-controller.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: ack-elasticache-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ember-csi-community-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ember-csi-community-operator.v0.9.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ember-csi.io/v1alpha1",
                "kind": "EmberStorageBackend",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "config": {
                    "envVars": {
                      "X_CSI_PERSISTENCE_CONFIG": {
                        "storage": "crd"
                      },
                      "X_CSI_DEBUG_MODE": "",
                      "X_CSI_DEFAULT_MOUNT_FS": "ext4",
                      "X_CSI_EMBER_CONFIG": {
                        "grpc_workers": 30,
                        "disable_logs": false,
                        "plugin_name": "",
                        "slow_operations": true,
                        "disabled": [],
                        "enable_probe": false,
                        "debug": false,
                        "user_id": "ember_csi.io",
                        "project_id": "ember_csi.io"
                      },
                      "X_CSI_BACKEND_CONFIG": {
                        "enable_unsupported_driver": false,
                        "driver": "ACCESSIscsi",
                        "name": "",
                        "driver__ACCESSIscsi__vrts_lun_sparse": true,
                        "driver__ACCESSIscsi__vrts_target_config": "/etc/cinder/vrts_target.xml",
                        "driver__AS13000__as13000_ipsan_pools__transform_csv": "Pool0",
                        "driver__AS13000__as13000_meta_pool": "",
                        "driver__AS13000__as13000_token_available_time": 3300,
                        "driver__Acs5000ISCSI__acs5000_copy_interval": 5,
                        "driver__Acs5000ISCSI__acs5000_target": 0,
                        "driver__Acs5000ISCSI__acs5000_volpool_name__transform_csv": "pool01",
                        "driver__DPLFC__dpl_pool": "",
                        "driver__DPLFC__dpl_port": 8357,
                        "driver__DPLISCSI__dpl_pool": "",
                        "driver__DPLISCSI__dpl_port": 8357,
                        "driver__DSWARE__dsware_rest_url": "",
                        "driver__DSWARE__dsware_storage_pools": "",
                        "driver__DSWARE__manager_ips__transform_csv_kvs": "",
                        "driver__Datera__backend_availability_zone": "",
                        "driver__Datera__chap_password": "",
                        "driver__Datera__chap_username": "",
                        "driver__Datera__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__Datera__driver_client_cert": "",
                        "driver__Datera__driver_client_cert_key": "",
                        "driver__Datera__driver_data_namespace": "",
                        "driver__Datera__driver_ssl_cert_path": "",
                        "driver__Datera__driver_ssl_cert_verify": false,
                        "driver__Datera__driver_use_ssl": false,
                        "driver__Datera__filter_function": "",
                        "driver__Datera__goodness_function": "",
                        "driver__Datera__iscsi_iotype": "fileio",
                        "driver__Datera__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__Datera__iscsi_target_flags": "",
                        "driver__Datera__iscsi_write_cache": "on",
                        "driver__Datera__num_shell_tries": 3,
                        "driver__Datera__num_volume_device_scan_tries": 3,
                        "driver__Datera__report_discard_supported": false,
                        "driver__Datera__storage_protocol": "iscsi",
                        "driver__Datera__target_helper": "tgtadm",
                        "driver__Datera__target_ip_address": "$my_ip",
                        "driver__Datera__target_port": 3260,
                        "driver__Datera__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__Datera__target_protocol": "iscsi",
                        "driver__Datera__trace_flags__transform_csv": "",
                        "driver__Datera__use_chap_auth": false,
                        "driver__Datera__volume_backend_name": "",
                        "driver__Datera__volume_clear": "zero",
                        "driver__Datera__volume_clear_ionice": "",
                        "driver__Datera__volume_clear_size": 0,
                        "driver__Datera__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__Datera__volume_copy_bps_limit": 0,
                        "driver__Datera__volume_dd_blocksize": "1M",
                        "driver__Datera__volumes_dir": "$state_path/volumes",
                        "driver__FJDXFC__cinder_eternus_config_file": "/etc/cinder/cinder_fujitsu_eternus_dx.xml",
                        "driver__FJDXISCSI__cinder_eternus_config_file": "/etc/cinder/cinder_fujitsu_eternus_dx.xml",
                        "driver__FlashSystemFC__flashsystem_connection_protocol": "FC",
                        "driver__FlashSystemFC__flashsystem_multihostmap_enabled": true,
                        "driver__FlashSystemISCSI__flashsystem_connection_protocol": "FC",
                        "driver__FlashSystemISCSI__flashsystem_multihostmap_enabled": true,
                        "driver__GPFS__gpfs_images_dir": "",
                        "driver__GPFS__gpfs_images_share_mode__transform_empty_none": "",
                        "driver__GPFS__gpfs_max_clone_depth": 0,
                        "driver__GPFS__gpfs_mount_point_base": "",
                        "driver__GPFS__gpfs_sparse_volumes": true,
                        "driver__GPFS__gpfs_storage_pool": "system",
                        "driver__GPFSRemote__gpfs_hosts__transform_csv": "",
                        "driver__GPFSRemote__gpfs_hosts_key_file": "$state_path/ssh_known_hosts",
                        "driver__GPFSRemote__gpfs_images_dir": "",
                        "driver__GPFSRemote__gpfs_images_share_mode__transform_empty_none": "",
                        "driver__GPFSRemote__gpfs_max_clone_depth": 0,
                        "driver__GPFSRemote__gpfs_mount_point_base": "",
                        "driver__GPFSRemote__gpfs_private_key": "",
                        "driver__GPFSRemote__gpfs_sparse_volumes": true,
                        "driver__GPFSRemote__gpfs_ssh_port": 22,
                        "driver__GPFSRemote__gpfs_storage_pool": "system",
                        "driver__GPFSRemote__gpfs_strict_host_key_policy": false,
                        "driver__GPFSRemote__gpfs_user_login": "root",
                        "driver__GPFSRemote__gpfs_user_password": "",
                        "driver__HBSDFC__backend_availability_zone": "",
                        "driver__HBSDFC__chap_password": "",
                        "driver__HBSDFC__chap_username": "",
                        "driver__HBSDFC__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__HBSDFC__driver_client_cert": "",
                        "driver__HBSDFC__driver_client_cert_key": "",
                        "driver__HBSDFC__driver_data_namespace": "",
                        "driver__HBSDFC__driver_ssl_cert_path": "",
                        "driver__HBSDFC__driver_ssl_cert_verify": false,
                        "driver__HBSDFC__driver_use_ssl": false,
                        "driver__HBSDFC__filter_function": "",
                        "driver__HBSDFC__goodness_function": "",
                        "driver__HBSDFC__iscsi_iotype": "fileio",
                        "driver__HBSDFC__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__HBSDFC__iscsi_target_flags": "",
                        "driver__HBSDFC__iscsi_write_cache": "on",
                        "driver__HBSDFC__num_shell_tries": 3,
                        "driver__HBSDFC__num_volume_device_scan_tries": 3,
                        "driver__HBSDFC__report_discard_supported": false,
                        "driver__HBSDFC__storage_protocol": "iscsi",
                        "driver__HBSDFC__target_helper": "tgtadm",
                        "driver__HBSDFC__target_ip_address": "$my_ip",
                        "driver__HBSDFC__target_port": 3260,
                        "driver__HBSDFC__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__HBSDFC__target_protocol": "iscsi",
                        "driver__HBSDFC__trace_flags__transform_csv": "",
                        "driver__HBSDFC__use_chap_auth": false,
                        "driver__HBSDFC__volume_backend_name": "",
                        "driver__HBSDFC__volume_clear": "zero",
                        "driver__HBSDFC__volume_clear_ionice": "",
                        "driver__HBSDFC__volume_clear_size": 0,
                        "driver__HBSDFC__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__HBSDFC__volume_copy_bps_limit": 0,
                        "driver__HBSDFC__volume_dd_blocksize": "1M",
                        "driver__HBSDFC__volumes_dir": "$state_path/volumes",
                        "driver__HBSDISCSI__backend_availability_zone": "",
                        "driver__HBSDISCSI__chap_password": "",
                        "driver__HBSDISCSI__chap_username": "",
                        "driver__HBSDISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__HBSDISCSI__driver_client_cert": "",
                        "driver__HBSDISCSI__driver_client_cert_key": "",
                        "driver__HBSDISCSI__driver_data_namespace": "",
                        "driver__HBSDISCSI__driver_ssl_cert_path": "",
                        "driver__HBSDISCSI__driver_ssl_cert_verify": false,
                        "driver__HBSDISCSI__driver_use_ssl": false,
                        "driver__HBSDISCSI__filter_function": "",
                        "driver__HBSDISCSI__goodness_function": "",
                        "driver__HBSDISCSI__iscsi_iotype": "fileio",
                        "driver__HBSDISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__HBSDISCSI__iscsi_target_flags": "",
                        "driver__HBSDISCSI__iscsi_write_cache": "on",
                        "driver__HBSDISCSI__num_shell_tries": 3,
                        "driver__HBSDISCSI__num_volume_device_scan_tries": 3,
                        "driver__HBSDISCSI__report_discard_supported": false,
                        "driver__HBSDISCSI__storage_protocol": "iscsi",
                        "driver__HBSDISCSI__target_helper": "tgtadm",
                        "driver__HBSDISCSI__target_ip_address": "$my_ip",
                        "driver__HBSDISCSI__target_port": 3260,
                        "driver__HBSDISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__HBSDISCSI__target_protocol": "iscsi",
                        "driver__HBSDISCSI__trace_flags__transform_csv": "",
                        "driver__HBSDISCSI__use_chap_auth": false,
                        "driver__HBSDISCSI__volume_backend_name": "",
                        "driver__HBSDISCSI__volume_clear": "zero",
                        "driver__HBSDISCSI__volume_clear_ionice": "",
                        "driver__HBSDISCSI__volume_clear_size": 0,
                        "driver__HBSDISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__HBSDISCSI__volume_copy_bps_limit": 0,
                        "driver__HBSDISCSI__volume_dd_blocksize": "1M",
                        "driver__HBSDISCSI__volumes_dir": "$state_path/volumes",
                        "driver__HPE3PARFC__hpe3par_api_url": "",
                        "driver__HPE3PARFC__hpe3par_cpg__transform_csv": "OpenStack",
                        "driver__HPE3PARFC__hpe3par_cpg_snap": "",
                        "driver__HPE3PARFC__hpe3par_debug": false,
                        "driver__HPE3PARFC__hpe3par_iscsi_chap_enabled": false,
                        "driver__HPE3PARFC__hpe3par_iscsi_ips__transform_csv": "",
                        "driver__HPE3PARFC__hpe3par_password": "",
                        "driver__HPE3PARFC__hpe3par_snapshot_expiration": "",
                        "driver__HPE3PARFC__hpe3par_snapshot_retention": "",
                        "driver__HPE3PARFC__hpe3par_target_nsp": "",
                        "driver__HPE3PARFC__hpe3par_username": "",
                        "driver__HPE3PARFC__san_ip": "",
                        "driver__HPE3PARFC__san_login": "admin",
                        "driver__HPE3PARFC__san_password": "",
                        "driver__HPE3PARFC__san_private_key": "",
                        "driver__HPE3PARFC__san_ssh_port": 22,
                        "driver__HPE3PARFC__ssh_conn_timeout": 30,
                        "driver__HPE3PARFC__target_ip_address": "$my_ip",
                        "driver__HPE3PARFC__target_port": 3260,
                        "driver__HPE3PARFC__unique_fqdn_network": true,
                        "driver__HPE3PARISCSI__hpe3par_api_url": "",
                        "driver__HPE3PARISCSI__hpe3par_cpg__transform_csv": "OpenStack",
                        "driver__HPE3PARISCSI__hpe3par_cpg_snap": "",
                        "driver__HPE3PARISCSI__hpe3par_debug": false,
                        "driver__HPE3PARISCSI__hpe3par_iscsi_chap_enabled": false,
                        "driver__HPE3PARISCSI__hpe3par_iscsi_ips__transform_csv": "",
                        "driver__HPE3PARISCSI__hpe3par_password": "",
                        "driver__HPE3PARISCSI__hpe3par_snapshot_expiration": "",
                        "driver__HPE3PARISCSI__hpe3par_snapshot_retention": "",
                        "driver__HPE3PARISCSI__hpe3par_target_nsp": "",
                        "driver__HPE3PARISCSI__hpe3par_username": "",
                        "driver__HPE3PARISCSI__san_ip": "",
                        "driver__HPE3PARISCSI__san_login": "admin",
                        "driver__HPE3PARISCSI__san_password": "",
                        "driver__HPE3PARISCSI__san_private_key": "",
                        "driver__HPE3PARISCSI__san_ssh_port": 22,
                        "driver__HPE3PARISCSI__ssh_conn_timeout": 30,
                        "driver__HPE3PARISCSI__target_ip_address": "$my_ip",
                        "driver__HPE3PARISCSI__target_port": 3260,
                        "driver__HPE3PARISCSI__unique_fqdn_network": true,
                        "driver__HPMSAFC__driver_ssl_cert_path": "",
                        "driver__HPMSAFC__driver_ssl_cert_verify": false,
                        "driver__HPMSAFC__driver_use_ssl": false,
                        "driver__HPMSAFC__hpmsa_pool_name": "A",
                        "driver__HPMSAFC__hpmsa_pool_type": "virtual",
                        "driver__HPMSAFC__san_ip": "",
                        "driver__HPMSAFC__san_login": "admin",
                        "driver__HPMSAFC__san_password": "",
                        "driver__HPMSAISCSI__driver_ssl_cert_path": "",
                        "driver__HPMSAISCSI__driver_ssl_cert_verify": false,
                        "driver__HPMSAISCSI__driver_use_ssl": false,
                        "driver__HPMSAISCSI__hpmsa_iscsi_ips__transform_csv": "",
                        "driver__HPMSAISCSI__hpmsa_pool_name": "A",
                        "driver__HPMSAISCSI__hpmsa_pool_type": "virtual",
                        "driver__HPMSAISCSI__san_ip": "",
                        "driver__HPMSAISCSI__san_login": "admin",
                        "driver__HPMSAISCSI__san_password": "",
                        "driver__HuaweiFC__cinder_huawei_conf_file": "/etc/cinder/cinder_huawei_conf.xml",
                        "driver__HuaweiFC__hypermetro_devices": "",
                        "driver__HuaweiFC__metro_domain_name": "",
                        "driver__HuaweiFC__metro_san_address": "",
                        "driver__HuaweiFC__metro_san_password": "",
                        "driver__HuaweiFC__metro_san_user": "",
                        "driver__HuaweiFC__metro_storage_pools": "",
                        "driver__HuaweiISCSI__cinder_huawei_conf_file": "/etc/cinder/cinder_huawei_conf.xml",
                        "driver__HuaweiISCSI__hypermetro_devices": "",
                        "driver__HuaweiISCSI__metro_domain_name": "",
                        "driver__HuaweiISCSI__metro_san_address": "",
                        "driver__HuaweiISCSI__metro_san_password": "",
                        "driver__HuaweiISCSI__metro_san_user": "",
                        "driver__HuaweiISCSI__metro_storage_pools": "",
                        "driver__IBMStorage__chap": "disabled",
                        "driver__IBMStorage__connection_type": "iscsi",
                        "driver__IBMStorage__management_ips": "",
                        "driver__IBMStorage__proxy": "cinder.volume.drivers.ibm.ibm_storage.proxy.IBMStorageProxy",
                        "driver__InStorageMCSFC__instorage_mcs_allow_tenant_qos": false,
                        "driver__InStorageMCSFC__instorage_mcs_localcopy_rate": 50,
                        "driver__InStorageMCSFC__instorage_mcs_localcopy_timeout": 120,
                        "driver__InStorageMCSFC__instorage_mcs_vol_autoexpand": true,
                        "driver__InStorageMCSFC__instorage_mcs_vol_compression": false,
                        "driver__InStorageMCSFC__instorage_mcs_vol_grainsize": 256,
                        "driver__InStorageMCSFC__instorage_mcs_vol_intier": true,
                        "driver__InStorageMCSFC__instorage_mcs_vol_iogrp": "0",
                        "driver__InStorageMCSFC__instorage_mcs_vol_rsize": 2,
                        "driver__InStorageMCSFC__instorage_mcs_vol_warning": 0,
                        "driver__InStorageMCSFC__instorage_mcs_volpool_name__transform_csv": "volpool",
                        "driver__InStorageMCSFC__instorage_san_secondary_ip": "",
                        "driver__InStorageMCSISCSI__instorage_mcs_allow_tenant_qos": false,
                        "driver__InStorageMCSISCSI__instorage_mcs_localcopy_rate": 50,
                        "driver__InStorageMCSISCSI__instorage_mcs_localcopy_timeout": 120,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_autoexpand": true,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_compression": false,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_grainsize": 256,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_intier": true,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_iogrp": "0",
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_rsize": 2,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_warning": 0,
                        "driver__InStorageMCSISCSI__instorage_mcs_volpool_name__transform_csv": "volpool",
                        "driver__InStorageMCSISCSI__instorage_san_secondary_ip": "",
                        "driver__InfiniboxVolume__chap_password": "",
                        "driver__InfiniboxVolume__chap_username": "",
                        "driver__InfiniboxVolume__infinidat_iscsi_netspaces__transform_csv": "",
                        "driver__InfiniboxVolume__infinidat_pool_name": "",
                        "driver__InfiniboxVolume__infinidat_storage_protocol": "fc",
                        "driver__InfiniboxVolume__infinidat_use_compression": false,
                        "driver__InfiniboxVolume__num_volume_device_scan_tries": 3,
                        "driver__InfiniboxVolume__san_ip": "",
                        "driver__InfiniboxVolume__san_login": "admin",
                        "driver__InfiniboxVolume__san_password": "",
                        "driver__InfiniboxVolume__san_thin_provision": true,
                        "driver__InfiniboxVolume__use_chap_auth": false,
                        "driver__InfiniboxVolume__volume_dd_blocksize": "1M",
                        "driver__InfortrendCLIFC__infortrend_cli_cache": false,
                        "driver__InfortrendCLIFC__infortrend_cli_max_retries": 5,
                        "driver__InfortrendCLIFC__infortrend_cli_path": "/opt/bin/Infortrend/raidcmd_ESDS10.jar",
                        "driver__InfortrendCLIFC__infortrend_cli_timeout": 60,
                        "driver__InfortrendCLIFC__infortrend_iqn_prefix": "iqn.2002-10.com.infortrend",
                        "driver__InfortrendCLIFC__infortrend_pools_name__transform_csv": "",
                        "driver__InfortrendCLIFC__infortrend_slots_a_channels_id__transform_csv": "",
                        "driver__InfortrendCLIFC__infortrend_slots_b_channels_id__transform_csv": "",
                        "driver__InfortrendCLIFC__java_path": "/usr/bin/java",
                        "driver__InfortrendCLIISCSI__infortrend_cli_cache": false,
                        "driver__InfortrendCLIISCSI__infortrend_cli_max_retries": 5,
                        "driver__InfortrendCLIISCSI__infortrend_cli_path": "/opt/bin/Infortrend/raidcmd_ESDS10.jar",
                        "driver__InfortrendCLIISCSI__infortrend_cli_timeout": 60,
                        "driver__InfortrendCLIISCSI__infortrend_iqn_prefix": "iqn.2002-10.com.infortrend",
                        "driver__InfortrendCLIISCSI__infortrend_pools_name__transform_csv": "",
                        "driver__InfortrendCLIISCSI__infortrend_slots_a_channels_id__transform_csv": "",
                        "driver__InfortrendCLIISCSI__infortrend_slots_b_channels_id__transform_csv": "",
                        "driver__InfortrendCLIISCSI__java_path": "/usr/bin/java",
                        "driver__JovianISCSI__backend_availability_zone": "",
                        "driver__JovianISCSI__chap_password": "",
                        "driver__JovianISCSI__chap_username": "",
                        "driver__JovianISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__JovianISCSI__driver_client_cert": "",
                        "driver__JovianISCSI__driver_client_cert_key": "",
                        "driver__JovianISCSI__driver_data_namespace": "",
                        "driver__JovianISCSI__driver_ssl_cert_path": "",
                        "driver__JovianISCSI__driver_ssl_cert_verify": false,
                        "driver__JovianISCSI__driver_use_ssl": false,
                        "driver__JovianISCSI__filter_function": "",
                        "driver__JovianISCSI__goodness_function": "",
                        "driver__JovianISCSI__iscsi_iotype": "fileio",
                        "driver__JovianISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__JovianISCSI__iscsi_target_flags": "",
                        "driver__JovianISCSI__iscsi_write_cache": "on",
                        "driver__JovianISCSI__num_shell_tries": 3,
                        "driver__JovianISCSI__num_volume_device_scan_tries": 3,
                        "driver__JovianISCSI__report_discard_supported": false,
                        "driver__JovianISCSI__storage_protocol": "iscsi",
                        "driver__JovianISCSI__target_helper": "tgtadm",
                        "driver__JovianISCSI__target_ip_address": "$my_ip",
                        "driver__JovianISCSI__target_port": 3260,
                        "driver__JovianISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__JovianISCSI__target_protocol": "iscsi",
                        "driver__JovianISCSI__trace_flags__transform_csv": "",
                        "driver__JovianISCSI__use_chap_auth": false,
                        "driver__JovianISCSI__volume_backend_name": "",
                        "driver__JovianISCSI__volume_clear": "zero",
                        "driver__JovianISCSI__volume_clear_ionice": "",
                        "driver__JovianISCSI__volume_clear_size": 0,
                        "driver__JovianISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__JovianISCSI__volume_copy_bps_limit": 0,
                        "driver__JovianISCSI__volume_dd_blocksize": "1M",
                        "driver__JovianISCSI__volumes_dir": "$state_path/volumes",
                        "driver__KaminarioISCSI__disable_discovery": false,
                        "driver__KaminarioISCSI__san_ip": "",
                        "driver__KaminarioISCSI__san_login": "admin",
                        "driver__KaminarioISCSI__san_password": "",
                        "driver__KaminarioISCSI__unique_fqdn_network": true,
                        "driver__KaminarioISCSI__volume_dd_blocksize": "1M",
                        "driver__LVMVolume__iscsi_iotype": "fileio",
                        "driver__LVMVolume__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__LVMVolume__iscsi_target_flags": "",
                        "driver__LVMVolume__iscsi_write_cache": "on",
                        "driver__LVMVolume__lvm_conf_file": "/etc/cinder/lvm.conf",
                        "driver__LVMVolume__lvm_mirrors": 0,
                        "driver__LVMVolume__lvm_suppress_fd_warnings": false,
                        "driver__LVMVolume__lvm_type": "auto",
                        "driver__LVMVolume__nvmet_port_id": 1,
                        "driver__LVMVolume__scst_target_driver": "iscsi",
                        "driver__LVMVolume__scst_target_iqn_name": "",
                        "driver__LVMVolume__spdk_max_queue_depth": 64,
                        "driver__LVMVolume__spdk_rpc_ip": "",
                        "driver__LVMVolume__spdk_rpc_password": "",
                        "driver__LVMVolume__spdk_rpc_port": 8000,
                        "driver__LVMVolume__spdk_rpc_username": "",
                        "driver__LVMVolume__target_helper": "tgtadm",
                        "driver__LVMVolume__target_ip_address": "$my_ip",
                        "driver__LVMVolume__target_port": 3260,
                        "driver__LVMVolume__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__LVMVolume__target_protocol": "iscsi",
                        "driver__LVMVolume__volume_clear": "zero",
                        "driver__LVMVolume__volume_clear_size": 0,
                        "driver__LVMVolume__volume_dd_blocksize": "1M",
                        "driver__LVMVolume__volume_group": "cinder-volumes",
                        "driver__LVMVolume__volumes_dir": "$state_path/volumes",
                        "driver__LenovoFC__lenovo_pool_name": "A",
                        "driver__LenovoFC__lenovo_pool_type": "virtual",
                        "driver__LenovoISCSI__lenovo_iscsi_ips__transform_csv": "",
                        "driver__LenovoISCSI__lenovo_pool_name": "A",
                        "driver__LenovoISCSI__lenovo_pool_type": "virtual",
                        "driver__LinstorDrbd__linstor_autoplace_count": 0,
                        "driver__LinstorDrbd__linstor_controller_diskless": true,
                        "driver__LinstorDrbd__linstor_default_blocksize": 4096,
                        "driver__LinstorDrbd__linstor_default_storage_pool_name": "DfltStorPool",
                        "driver__LinstorDrbd__linstor_default_uri": "linstor://localhost",
                        "driver__LinstorDrbd__linstor_default_volume_group_name": "drbd-vg",
                        "driver__LinstorDrbd__linstor_volume_downsize_factor__transform_string_float": "4096",
                        "driver__LinstorIscsi__linstor_autoplace_count": 0,
                        "driver__LinstorIscsi__linstor_controller_diskless": true,
                        "driver__LinstorIscsi__linstor_default_blocksize": 4096,
                        "driver__LinstorIscsi__linstor_default_storage_pool_name": "DfltStorPool",
                        "driver__LinstorIscsi__linstor_default_uri": "linstor://localhost",
                        "driver__LinstorIscsi__linstor_default_volume_group_name": "drbd-vg",
                        "driver__LinstorIscsi__linstor_volume_downsize_factor__transform_string_float": "4096",
                        "driver__MStorageFC__nec_actual_free_capacity": false,
                        "driver__MStorageFC__nec_auto_accesscontrol": true,
                        "driver__MStorageFC__nec_backend_max_ld_count": 1024,
                        "driver__MStorageFC__nec_backup_ldname_format": "LX:%s",
                        "driver__MStorageFC__nec_backup_pools__transform_csv": "",
                        "driver__MStorageFC__nec_cv_ldname_format": "LX:__ControlVolume_%xh",
                        "driver__MStorageFC__nec_diskarray_name": "",
                        "driver__MStorageFC__nec_ismcli_fip": "",
                        "driver__MStorageFC__nec_ismcli_password": "",
                        "driver__MStorageFC__nec_ismcli_privkey": "",
                        "driver__MStorageFC__nec_ismcli_user": "",
                        "driver__MStorageFC__nec_ismview_alloptimize": false,
                        "driver__MStorageFC__nec_ismview_dir": "/tmp/nec/cinder",
                        "driver__MStorageFC__nec_ldname_format": "LX:%s",
                        "driver__MStorageFC__nec_ldset": "",
                        "driver__MStorageFC__nec_pools__transform_csv": "",
                        "driver__MStorageFC__nec_queryconfig_view": false,
                        "driver__MStorageFC__nec_ssh_pool_port_number": 22,
                        "driver__MStorageFC__nec_unpairthread_timeout": 3600,
                        "driver__MStorageISCSI__nec_actual_free_capacity": false,
                        "driver__MStorageISCSI__nec_auto_accesscontrol": true,
                        "driver__MStorageISCSI__nec_backend_max_ld_count": 1024,
                        "driver__MStorageISCSI__nec_backup_ldname_format": "LX:%s",
                        "driver__MStorageISCSI__nec_backup_pools__transform_csv": "",
                        "driver__MStorageISCSI__nec_cv_ldname_format": "LX:__ControlVolume_%xh",
                        "driver__MStorageISCSI__nec_diskarray_name": "",
                        "driver__MStorageISCSI__nec_ismcli_fip": "",
                        "driver__MStorageISCSI__nec_ismcli_password": "",
                        "driver__MStorageISCSI__nec_ismcli_privkey": "",
                        "driver__MStorageISCSI__nec_ismcli_user": "",
                        "driver__MStorageISCSI__nec_ismview_alloptimize": false,
                        "driver__MStorageISCSI__nec_ismview_dir": "/tmp/nec/cinder",
                        "driver__MStorageISCSI__nec_ldname_format": "LX:%s",
                        "driver__MStorageISCSI__nec_ldset": "",
                        "driver__MStorageISCSI__nec_pools__transform_csv": "",
                        "driver__MStorageISCSI__nec_queryconfig_view": false,
                        "driver__MStorageISCSI__nec_ssh_pool_port_number": 22,
                        "driver__MStorageISCSI__nec_unpairthread_timeout": 3600,
                        "driver__MacroSANFC__backend_availability_zone": "",
                        "driver__MacroSANFC__chap_password": "",
                        "driver__MacroSANFC__chap_username": "",
                        "driver__MacroSANFC__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__MacroSANFC__driver_client_cert": "",
                        "driver__MacroSANFC__driver_client_cert_key": "",
                        "driver__MacroSANFC__driver_data_namespace": "",
                        "driver__MacroSANFC__driver_ssl_cert_path": "",
                        "driver__MacroSANFC__driver_ssl_cert_verify": false,
                        "driver__MacroSANFC__driver_use_ssl": false,
                        "driver__MacroSANFC__filter_function": "",
                        "driver__MacroSANFC__goodness_function": "",
                        "driver__MacroSANFC__iscsi_iotype": "fileio",
                        "driver__MacroSANFC__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__MacroSANFC__iscsi_target_flags": "",
                        "driver__MacroSANFC__iscsi_write_cache": "on",
                        "driver__MacroSANFC__num_shell_tries": 3,
                        "driver__MacroSANFC__num_volume_device_scan_tries": 3,
                        "driver__MacroSANFC__report_discard_supported": false,
                        "driver__MacroSANFC__storage_protocol": "iscsi",
                        "driver__MacroSANFC__target_helper": "tgtadm",
                        "driver__MacroSANFC__target_ip_address": "$my_ip",
                        "driver__MacroSANFC__target_port": 3260,
                        "driver__MacroSANFC__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__MacroSANFC__target_protocol": "iscsi",
                        "driver__MacroSANFC__trace_flags__transform_csv": "",
                        "driver__MacroSANFC__use_chap_auth": false,
                        "driver__MacroSANFC__volume_backend_name": "",
                        "driver__MacroSANFC__volume_clear": "zero",
                        "driver__MacroSANFC__volume_clear_ionice": "",
                        "driver__MacroSANFC__volume_clear_size": 0,
                        "driver__MacroSANFC__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__MacroSANFC__volume_copy_bps_limit": 0,
                        "driver__MacroSANFC__volume_dd_blocksize": "1M",
                        "driver__MacroSANFC__volumes_dir": "$state_path/volumes",
                        "driver__MacroSANISCSI__backend_availability_zone": "",
                        "driver__MacroSANISCSI__chap_password": "",
                        "driver__MacroSANISCSI__chap_username": "",
                        "driver__MacroSANISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__MacroSANISCSI__driver_client_cert": "",
                        "driver__MacroSANISCSI__driver_client_cert_key": "",
                        "driver__MacroSANISCSI__driver_data_namespace": "",
                        "driver__MacroSANISCSI__driver_ssl_cert_path": "",
                        "driver__MacroSANISCSI__driver_ssl_cert_verify": false,
                        "driver__MacroSANISCSI__driver_use_ssl": false,
                        "driver__MacroSANISCSI__filter_function": "",
                        "driver__MacroSANISCSI__goodness_function": "",
                        "driver__MacroSANISCSI__iscsi_iotype": "fileio",
                        "driver__MacroSANISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__MacroSANISCSI__iscsi_target_flags": "",
                        "driver__MacroSANISCSI__iscsi_write_cache": "on",
                        "driver__MacroSANISCSI__num_shell_tries": 3,
                        "driver__MacroSANISCSI__num_volume_device_scan_tries": 3,
                        "driver__MacroSANISCSI__report_discard_supported": false,
                        "driver__MacroSANISCSI__storage_protocol": "iscsi",
                        "driver__MacroSANISCSI__target_helper": "tgtadm",
                        "driver__MacroSANISCSI__target_ip_address": "$my_ip",
                        "driver__MacroSANISCSI__target_port": 3260,
                        "driver__MacroSANISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__MacroSANISCSI__target_protocol": "iscsi",
                        "driver__MacroSANISCSI__trace_flags__transform_csv": "",
                        "driver__MacroSANISCSI__use_chap_auth": false,
                        "driver__MacroSANISCSI__volume_backend_name": "",
                        "driver__MacroSANISCSI__volume_clear": "zero",
                        "driver__MacroSANISCSI__volume_clear_ionice": "",
                        "driver__MacroSANISCSI__volume_clear_size": 0,
                        "driver__MacroSANISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__MacroSANISCSI__volume_copy_bps_limit": 0,
                        "driver__MacroSANISCSI__volume_dd_blocksize": "1M",
                        "driver__MacroSANISCSI__volumes_dir": "$state_path/volumes",
                        "driver__NetAppCmodeFibreChannel__netapp_vserver": "",
                        "driver__NetAppCmodeISCSI__netapp_vserver": "",
                        "driver__NexentaISCSI__nexenta_blocksize": 4096,
                        "driver__NexentaISCSI__nexenta_dataset_compression": "on",
                        "driver__NexentaISCSI__nexenta_dataset_dedup": "off",
                        "driver__NexentaISCSI__nexenta_dataset_description": "",
                        "driver__NexentaISCSI__nexenta_folder": "",
                        "driver__NexentaISCSI__nexenta_group_snapshot_template": "group-snapshot-%s",
                        "driver__NexentaISCSI__nexenta_host": "",
                        "driver__NexentaISCSI__nexenta_host_group_prefix": "cinder",
                        "driver__NexentaISCSI__nexenta_iscsi_target_host_group": "all",
                        "driver__NexentaISCSI__nexenta_iscsi_target_portal_groups": "",
                        "driver__NexentaISCSI__nexenta_iscsi_target_portal_port": 3260,
                        "driver__NexentaISCSI__nexenta_iscsi_target_portals": "",
                        "driver__NexentaISCSI__nexenta_lu_writebackcache_disabled": false,
                        "driver__NexentaISCSI__nexenta_luns_per_target": 100,
                        "driver__NexentaISCSI__nexenta_ns5_blocksize": 32,
                        "driver__NexentaISCSI__nexenta_origin_snapshot_template": "origin-snapshot-%s",
                        "driver__NexentaISCSI__nexenta_rest_backoff_factor__transform_string_float": "0.5",
                        "driver__NexentaISCSI__nexenta_rest_connect_timeout__transform_string_float": "30",
                        "driver__NexentaISCSI__nexenta_rest_protocol": "auto",
                        "driver__NexentaISCSI__nexenta_rest_read_timeout__transform_string_float": "300",
                        "driver__NexentaISCSI__nexenta_rest_retry_count": 3,
                        "driver__NexentaISCSI__nexenta_sparse": false,
                        "driver__NexentaISCSI__nexenta_target_group_prefix": "cinder",
                        "driver__NexentaISCSI__nexenta_target_prefix": "iqn.1986-03.com.sun:02:cinder",
                        "driver__NexentaISCSI__nexenta_use_https": true,
                        "driver__NexentaISCSI__nexenta_volume": "cinder",
                        "driver__NexentaISCSI__nexenta_volume_group": "iscsi",
                        "driver__NimbleFC__nimble_pool_name": "default",
                        "driver__NimbleFC__nimble_subnet_label": "*",
                        "driver__NimbleFC__nimble_verify_cert_path": "",
                        "driver__NimbleFC__nimble_verify_certificate": false,
                        "driver__NimbleISCSI__nimble_pool_name": "default",
                        "driver__NimbleISCSI__nimble_subnet_label": "*",
                        "driver__NimbleISCSI__nimble_verify_cert_path": "",
                        "driver__NimbleISCSI__nimble_verify_certificate": false,
                        "driver__PVMEFC__driver_ssl_cert_path": "",
                        "driver__PVMEFC__driver_ssl_cert_verify": false,
                        "driver__PVMEFC__driver_use_ssl": false,
                        "driver__PVMEFC__pvme_pool_name": "A",
                        "driver__PVMEFC__san_ip": "",
                        "driver__PVMEFC__san_login": "admin",
                        "driver__PVMEFC__san_password": "",
                        "driver__PVMEISCSI__driver_ssl_cert_path": "",
                        "driver__PVMEISCSI__driver_ssl_cert_verify": false,
                        "driver__PVMEISCSI__driver_use_ssl": false,
                        "driver__PVMEISCSI__pvme_iscsi_ips__transform_csv": "",
                        "driver__PVMEISCSI__pvme_pool_name": "A",
                        "driver__PVMEISCSI__san_ip": "",
                        "driver__PVMEISCSI__san_login": "admin",
                        "driver__PVMEISCSI__san_password": "",
                        "driver__PowerFlex__powerflex_allow_migration_during_rebuild": false,
                        "driver__PowerFlex__powerflex_allow_non_padded_volumes": false,
                        "driver__PowerFlex__powerflex_max_over_subscription_ratio__transform_string_float": "10.0",
                        "driver__PowerFlex__powerflex_rest_server_port": 443,
                        "driver__PowerFlex__powerflex_round_volume_capacity": true,
                        "driver__PowerFlex__powerflex_server_api_version": "",
                        "driver__PowerFlex__powerflex_storage_pools": "",
                        "driver__PowerFlex__powerflex_unmap_volume_before_deletion": false,
                        "driver__PowerMaxFC__driver_ssl_cert_verify": false,
                        "driver__PowerMaxFC__initiator_check": false,
                        "driver__PowerMaxFC__interval": 3,
                        "driver__PowerMaxFC__load_balance": false,
                        "driver__PowerMaxFC__load_balance_real_time": false,
                        "driver__PowerMaxFC__load_data_format": "Avg",
                        "driver__PowerMaxFC__load_look_back": 60,
                        "driver__PowerMaxFC__load_look_back_real_time": 1,
                        "driver__PowerMaxFC__port_group_load_metric": "PercentBusy",
                        "driver__PowerMaxFC__port_load_metric": "PercentBusy",
                        "driver__PowerMaxFC__powermax_array": "",
                        "driver__PowerMaxFC__powermax_array_tag_list__transform_csv": "",
                        "driver__PowerMaxFC__powermax_port_group_name_template": "portGroupName",
                        "driver__PowerMaxFC__powermax_port_groups__transform_csv": "",
                        "driver__PowerMaxFC__powermax_service_level": "",
                        "driver__PowerMaxFC__powermax_short_host_name_template": "shortHostName",
                        "driver__PowerMaxFC__powermax_snapvx_unlink_limit": 3,
                        "driver__PowerMaxFC__powermax_srp": "",
                        "driver__PowerMaxFC__retries": 200,
                        "driver__PowerMaxFC__san_ip": "",
                        "driver__PowerMaxFC__san_login": "admin",
                        "driver__PowerMaxFC__san_password": "",
                        "driver__PowerMaxFC__u4p_failover_autofailback": true,
                        "driver__PowerMaxFC__u4p_failover_backoff_factor": 1,
                        "driver__PowerMaxFC__u4p_failover_retries": 3,
                        "driver__PowerMaxFC__u4p_failover_target__transform_csv_kvs": "",
                        "driver__PowerMaxFC__u4p_failover_timeout": 20,
                        "driver__PowerMaxFC__vmax_workload": "",
                        "driver__PowerMaxISCSI__chap_password": "",
                        "driver__PowerMaxISCSI__chap_username": "",
                        "driver__PowerMaxISCSI__driver_ssl_cert_verify": false,
                        "driver__PowerMaxISCSI__initiator_check": false,
                        "driver__PowerMaxISCSI__interval": 3,
                        "driver__PowerMaxISCSI__load_balance": false,
                        "driver__PowerMaxISCSI__load_balance_real_time": false,
                        "driver__PowerMaxISCSI__load_data_format": "Avg",
                        "driver__PowerMaxISCSI__load_look_back": 60,
                        "driver__PowerMaxISCSI__load_look_back_real_time": 1,
                        "driver__PowerMaxISCSI__port_group_load_metric": "PercentBusy",
                        "driver__PowerMaxISCSI__port_load_metric": "PercentBusy",
                        "driver__PowerMaxISCSI__powermax_array": "",
                        "driver__PowerMaxISCSI__powermax_array_tag_list__transform_csv": "",
                        "driver__PowerMaxISCSI__powermax_port_group_name_template": "portGroupName",
                        "driver__PowerMaxISCSI__powermax_port_groups__transform_csv": "",
                        "driver__PowerMaxISCSI__powermax_service_level": "",
                        "driver__PowerMaxISCSI__powermax_short_host_name_template": "shortHostName",
                        "driver__PowerMaxISCSI__powermax_snapvx_unlink_limit": 3,
                        "driver__PowerMaxISCSI__powermax_srp": "",
                        "driver__PowerMaxISCSI__retries": 200,
                        "driver__PowerMaxISCSI__san_ip": "",
                        "driver__PowerMaxISCSI__san_login": "admin",
                        "driver__PowerMaxISCSI__san_password": "",
                        "driver__PowerMaxISCSI__u4p_failover_autofailback": true,
                        "driver__PowerMaxISCSI__u4p_failover_backoff_factor": 1,
                        "driver__PowerMaxISCSI__u4p_failover_retries": 3,
                        "driver__PowerMaxISCSI__u4p_failover_target__transform_csv_kvs": "",
                        "driver__PowerMaxISCSI__u4p_failover_timeout": 20,
                        "driver__PowerMaxISCSI__use_chap_auth": false,
                        "driver__PowerMaxISCSI__vmax_workload": "",
                        "driver__PowerStore__powerstore_appliances__transform_csv": "",
                        "driver__PowerStore__powerstore_ports__transform_csv": "",
                        "driver__PureFC__driver_ssl_cert_path": "",
                        "driver__PureFC__driver_ssl_cert_verify": false,
                        "driver__PureFC__pure_api_token": "",
                        "driver__PureFC__pure_automatic_max_oversubscription_ratio": true,
                        "driver__PureFC__pure_eradicate_on_delete": false,
                        "driver__PureFC__pure_host_personality__transform_empty_none": "",
                        "driver__PureFC__pure_iscsi_cidr": "0.0.0.0/0",
                        "driver__PureFC__pure_replica_interval_default": 3600,
                        "driver__PureFC__pure_replica_retention_long_term_default": 7,
                        "driver__PureFC__pure_replica_retention_long_term_per_day_default": 3,
                        "driver__PureFC__pure_replica_retention_short_term_default": 14400,
                        "driver__PureFC__pure_replication_pg_name": "cinder-group",
                        "driver__PureFC__pure_replication_pod_name": "cinder-pod",
                        "driver__PureFC__san_ip": "",
                        "driver__PureFC__use_chap_auth": false,
                        "driver__PureISCSI__driver_ssl_cert_path": "",
                        "driver__PureISCSI__driver_ssl_cert_verify": false,
                        "driver__PureISCSI__pure_api_token": "",
                        "driver__PureISCSI__pure_automatic_max_oversubscription_ratio": true,
                        "driver__PureISCSI__pure_eradicate_on_delete": false,
                        "driver__PureISCSI__pure_host_personality__transform_empty_none": "",
                        "driver__PureISCSI__pure_iscsi_cidr": "0.0.0.0/0",
                        "driver__PureISCSI__pure_replica_interval_default": 3600,
                        "driver__PureISCSI__pure_replica_retention_long_term_default": 7,
                        "driver__PureISCSI__pure_replica_retention_long_term_per_day_default": 3,
                        "driver__PureISCSI__pure_replica_retention_short_term_default": 14400,
                        "driver__PureISCSI__pure_replication_pg_name": "cinder-group",
                        "driver__PureISCSI__pure_replication_pod_name": "cinder-pod",
                        "driver__PureISCSI__san_ip": "",
                        "driver__PureISCSI__use_chap_auth": false,
                        "driver__QnapISCSI__chap_password": "",
                        "driver__QnapISCSI__chap_username": "",
                        "driver__QnapISCSI__driver_ssl_cert_verify": false,
                        "driver__QnapISCSI__qnap_management_url": "",
                        "driver__QnapISCSI__qnap_poolname": "",
                        "driver__QnapISCSI__qnap_storage_protocol": "iscsi",
                        "driver__QnapISCSI__san_login": "admin",
                        "driver__QnapISCSI__san_password": "",
                        "driver__QnapISCSI__target_ip_address": "$my_ip",
                        "driver__QnapISCSI__use_chap_auth": false,
                        "driver__Quobyte__quobyte_client_cfg": "",
                        "driver__Quobyte__quobyte_mount_point_base": "$state_path/mnt",
                        "driver__Quobyte__quobyte_overlay_volumes": false,
                        "driver__Quobyte__quobyte_qcow2_volumes": true,
                        "driver__Quobyte__quobyte_sparsed_volumes": true,
                        "driver__Quobyte__quobyte_volume_from_snapshot_cache": false,
                        "driver__Quobyte__quobyte_volume_url": "",
                        "driver__RBD__deferred_deletion_delay": 0,
                        "driver__RBD__deferred_deletion_purge_interval": 60,
                        "driver__RBD__enable_deferred_deletion": false,
                        "driver__RBD__rados_connect_timeout": -1,
                        "driver__RBD__rados_connection_interval": 5,
                        "driver__RBD__rados_connection_retries": 3,
                        "driver__RBD__rbd_ceph_conf": "",
                        "driver__RBD__rbd_cluster_name": "ceph",
                        "driver__RBD__rbd_exclusive_cinder_pool": false,
                        "driver__RBD__rbd_flatten_volume_from_snapshot": false,
                        "driver__RBD__rbd_keyring_conf": "",
                        "driver__RBD__rbd_max_clone_depth": 5,
                        "driver__RBD__rbd_pool": "rbd",
                        "driver__RBD__rbd_secret_uuid": "",
                        "driver__RBD__rbd_store_chunk_size": 4,
                        "driver__RBD__rbd_user": "",
                        "driver__RBD__report_dynamic_total_capacity": true,
                        "driver__RBD__volume_dd_blocksize": "1M",
                        "driver__RBDISCSI__deferred_deletion_delay": 0,
                        "driver__RBDISCSI__deferred_deletion_purge_interval": 60,
                        "driver__RBDISCSI__driver_ssl_cert_verify": false,
                        "driver__RBDISCSI__enable_deferred_deletion": false,
                        "driver__RBDISCSI__rados_connect_timeout": -1,
                        "driver__RBDISCSI__rados_connection_interval": 5,
                        "driver__RBDISCSI__rados_connection_retries": 3,
                        "driver__RBDISCSI__rbd_ceph_conf": "",
                        "driver__RBDISCSI__rbd_cluster_name": "ceph",
                        "driver__RBDISCSI__rbd_exclusive_cinder_pool": false,
                        "driver__RBDISCSI__rbd_flatten_volume_from_snapshot": false,
                        "driver__RBDISCSI__rbd_iscsi_api_debug": false,
                        "driver__RBDISCSI__rbd_iscsi_api_password": "",
                        "driver__RBDISCSI__rbd_iscsi_api_url": "",
                        "driver__RBDISCSI__rbd_iscsi_api_user": "",
                        "driver__RBDISCSI__rbd_iscsi_target_iqn": "",
                        "driver__RBDISCSI__rbd_max_clone_depth": 5,
                        "driver__RBDISCSI__rbd_pool": "rbd",
                        "driver__RBDISCSI__rbd_secret_uuid": "",
                        "driver__RBDISCSI__rbd_store_chunk_size": 4,
                        "driver__RBDISCSI__rbd_user": "",
                        "driver__RBDISCSI__report_dynamic_total_capacity": true,
                        "driver__RBDISCSI__suppress_requests_ssl_warnings": false,
                        "driver__RBDISCSI__volume_dd_blocksize": "1M",
                        "driver__RSD__podm_password": "",
                        "driver__RSD__podm_url": "",
                        "driver__RSD__podm_username": "",
                        "driver__SCFC__dell_api_async_rest_timeout": 15,
                        "driver__SCFC__dell_api_sync_rest_timeout": 30,
                        "driver__SCFC__dell_sc_api_port": 3033,
                        "driver__SCFC__dell_sc_server_folder": "openstack",
                        "driver__SCFC__dell_sc_ssn": 64702,
                        "driver__SCFC__dell_sc_verify_cert": false,
                        "driver__SCFC__dell_sc_volume_folder": "openstack",
                        "driver__SCFC__dell_server_os": "Red Hat Linux 6.x",
                        "driver__SCFC__excluded_domain_ips__transform_csv": "",
                        "driver__SCFC__included_domain_ips__transform_csv": "",
                        "driver__SCFC__secondary_san_ip": "",
                        "driver__SCFC__secondary_san_login": "Admin",
                        "driver__SCFC__secondary_san_password": "",
                        "driver__SCFC__secondary_sc_api_port": 3033,
                        "driver__SCISCSI__dell_api_async_rest_timeout": 15,
                        "driver__SCISCSI__dell_api_sync_rest_timeout": 30,
                        "driver__SCISCSI__dell_sc_api_port": 3033,
                        "driver__SCISCSI__dell_sc_server_folder": "openstack",
                        "driver__SCISCSI__dell_sc_ssn": 64702,
                        "driver__SCISCSI__dell_sc_verify_cert": false,
                        "driver__SCISCSI__dell_sc_volume_folder": "openstack",
                        "driver__SCISCSI__dell_server_os": "Red Hat Linux 6.x",
                        "driver__SCISCSI__excluded_domain_ips__transform_csv": "",
                        "driver__SCISCSI__included_domain_ips__transform_csv": "",
                        "driver__SCISCSI__secondary_san_ip": "",
                        "driver__SCISCSI__secondary_san_login": "Admin",
                        "driver__SCISCSI__secondary_san_password": "",
                        "driver__SCISCSI__secondary_sc_api_port": 3033,
                        "driver__SdsISCSI__backend_availability_zone": "",
                        "driver__SdsISCSI__chap_password": "",
                        "driver__SdsISCSI__chap_username": "",
                        "driver__SdsISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__SdsISCSI__driver_client_cert": "",
                        "driver__SdsISCSI__driver_client_cert_key": "",
                        "driver__SdsISCSI__driver_data_namespace": "",
                        "driver__SdsISCSI__driver_ssl_cert_path": "",
                        "driver__SdsISCSI__driver_ssl_cert_verify": false,
                        "driver__SdsISCSI__driver_use_ssl": false,
                        "driver__SdsISCSI__filter_function": "",
                        "driver__SdsISCSI__goodness_function": "",
                        "driver__SdsISCSI__iscsi_iotype": "fileio",
                        "driver__SdsISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__SdsISCSI__iscsi_target_flags": "",
                        "driver__SdsISCSI__iscsi_write_cache": "on",
                        "driver__SdsISCSI__num_shell_tries": 3,
                        "driver__SdsISCSI__num_volume_device_scan_tries": 3,
                        "driver__SdsISCSI__report_discard_supported": false,
                        "driver__SdsISCSI__storage_protocol": "iscsi",
                        "driver__SdsISCSI__target_helper": "tgtadm",
                        "driver__SdsISCSI__target_ip_address": "$my_ip",
                        "driver__SdsISCSI__target_port": 3260,
                        "driver__SdsISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__SdsISCSI__target_protocol": "iscsi",
                        "driver__SdsISCSI__trace_flags__transform_csv": "",
                        "driver__SdsISCSI__use_chap_auth": false,
                        "driver__SdsISCSI__volume_backend_name": "",
                        "driver__SdsISCSI__volume_clear": "zero",
                        "driver__SdsISCSI__volume_clear_ionice": "",
                        "driver__SdsISCSI__volume_clear_size": 0,
                        "driver__SdsISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__SdsISCSI__volume_copy_bps_limit": 0,
                        "driver__SdsISCSI__volume_dd_blocksize": "1M",
                        "driver__SdsISCSI__volumes_dir": "$state_path/volumes",
                        "driver__SolidFire__driver_ssl_cert_verify": false,
                        "driver__SolidFire__san_ip": "",
                        "driver__SolidFire__san_login": "admin",
                        "driver__SolidFire__san_password": "",
                        "driver__SolidFire__sf_account_prefix": "",
                        "driver__SolidFire__sf_allow_tenant_qos": false,
                        "driver__SolidFire__sf_api_port": 443,
                        "driver__SolidFire__sf_api_request_timeout": 30,
                        "driver__SolidFire__sf_cluster_pairing_timeout": 60,
                        "driver__SolidFire__sf_emulate_512": true,
                        "driver__SolidFire__sf_enable_vag": false,
                        "driver__SolidFire__sf_provisioning_calc": "maxProvisionedSpace",
                        "driver__SolidFire__sf_svip": "",
                        "driver__SolidFire__sf_volume_clone_timeout": 600,
                        "driver__SolidFire__sf_volume_create_timeout": 60,
                        "driver__SolidFire__sf_volume_pairing_timeout": 3600,
                        "driver__SolidFire__sf_volume_prefix": "UUID-",
                        "driver__StorPool__storpool_replication": 3,
                        "driver__StorPool__storpool_template": "",
                        "driver__StorwizeSVCFC__cycle_period_seconds": 300,
                        "driver__StorwizeSVCFC__storwize_peer_pool": "",
                        "driver__StorwizeSVCFC__storwize_preferred_host_site__transform_csv_kvs": "",
                        "driver__StorwizeSVCFC__storwize_san_secondary_ip": "",
                        "driver__StorwizeSVCFC__storwize_svc_allow_tenant_qos": false,
                        "driver__StorwizeSVCFC__storwize_svc_flashcopy_rate": 50,
                        "driver__StorwizeSVCFC__storwize_svc_flashcopy_timeout": 120,
                        "driver__StorwizeSVCFC__storwize_svc_mirror_pool": "",
                        "driver__StorwizeSVCFC__storwize_svc_multipath_enabled": false,
                        "driver__StorwizeSVCFC__storwize_svc_retain_aux_volume": false,
                        "driver__StorwizeSVCFC__storwize_svc_stretched_cluster_partner": "",
                        "driver__StorwizeSVCFC__storwize_svc_vol_autoexpand": true,
                        "driver__StorwizeSVCFC__storwize_svc_vol_compression": false,
                        "driver__StorwizeSVCFC__storwize_svc_vol_easytier": true,
                        "driver__StorwizeSVCFC__storwize_svc_vol_grainsize": 256,
                        "driver__StorwizeSVCFC__storwize_svc_vol_iogrp": "0",
                        "driver__StorwizeSVCFC__storwize_svc_vol_nofmtdisk": false,
                        "driver__StorwizeSVCFC__storwize_svc_vol_rsize": 2,
                        "driver__StorwizeSVCFC__storwize_svc_vol_warning": 0,
                        "driver__StorwizeSVCFC__storwize_svc_volpool_name__transform_csv": "volpool",
                        "driver__StorwizeSVCISCSI__cycle_period_seconds": 300,
                        "driver__StorwizeSVCISCSI__storwize_peer_pool": "",
                        "driver__StorwizeSVCISCSI__storwize_preferred_host_site__transform_csv_kvs": "",
                        "driver__StorwizeSVCISCSI__storwize_san_secondary_ip": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_allow_tenant_qos": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_flashcopy_rate": 50,
                        "driver__StorwizeSVCISCSI__storwize_svc_flashcopy_timeout": 120,
                        "driver__StorwizeSVCISCSI__storwize_svc_iscsi_chap_enabled": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_mirror_pool": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_retain_aux_volume": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_stretched_cluster_partner": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_autoexpand": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_compression": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_easytier": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_grainsize": 256,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_iogrp": "0",
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_nofmtdisk": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_rsize": 2,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_warning": 0,
                        "driver__StorwizeSVCISCSI__storwize_svc_volpool_name__transform_csv": "volpool",
                        "driver__SynoISCSI__chap_password": "",
                        "driver__SynoISCSI__chap_username": "",
                        "driver__SynoISCSI__driver_use_ssl": false,
                        "driver__SynoISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__SynoISCSI__synology_admin_port": 5000,
                        "driver__SynoISCSI__synology_device_id": "",
                        "driver__SynoISCSI__synology_one_time_pass": "",
                        "driver__SynoISCSI__synology_password": "",
                        "driver__SynoISCSI__synology_pool_name": "",
                        "driver__SynoISCSI__synology_ssl_verify": true,
                        "driver__SynoISCSI__synology_username": "admin",
                        "driver__SynoISCSI__target_ip_address": "$my_ip",
                        "driver__SynoISCSI__target_port": 3260,
                        "driver__SynoISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__SynoISCSI__target_protocol": "iscsi",
                        "driver__SynoISCSI__use_chap_auth": false,
                        "driver__Unity__remove_empty_host": false,
                        "driver__Unity__unity_io_ports__transform_csv": "",
                        "driver__Unity__unity_storage_pool_names__transform_csv": "",
                        "driver__VNX__default_timeout": 31536000,
                        "driver__VNX__destroy_empty_storage_group": false,
                        "driver__VNX__force_delete_lun_in_storagegroup": true,
                        "driver__VNX__ignore_pool_full_threshold": false,
                        "driver__VNX__initiator_auto_deregistration": false,
                        "driver__VNX__initiator_auto_registration": false,
                        "driver__VNX__io_port_list__transform_csv": "",
                        "driver__VNX__iscsi_initiators": "",
                        "driver__VNX__max_luns_per_storage_group": 255,
                        "driver__VNX__naviseccli_path": "",
                        "driver__VNX__storage_vnx_authentication_type": "global",
                        "driver__VNX__storage_vnx_pool_names__transform_csv": "",
                        "driver__VNX__storage_vnx_security_file_dir": "",
                        "driver__VNX__vnx_async_migrate": true,
                        "driver__VZStorage__vzstorage_default_volume_format": "raw",
                        "driver__VZStorage__vzstorage_mount_options__transform_csv": "",
                        "driver__VZStorage__vzstorage_mount_point_base": "$state_path/mnt",
                        "driver__VZStorage__vzstorage_shares_config": "/etc/cinder/vzstorage_shares",
                        "driver__VZStorage__vzstorage_sparsed_volumes": true,
                        "driver__VZStorage__vzstorage_used_ratio__transform_string_float": "0.95",
                        "driver__WindowsISCSI__windows_iscsi_lun_path": "C:\\iSCSIVirtualDisks",
                        "driver__WindowsSmbfs__smbfs_default_volume_format": "vhd",
                        "driver__WindowsSmbfs__smbfs_mount_point_base": "C:\\OpenStack\\_mnt",
                        "driver__WindowsSmbfs__smbfs_pool_mappings__transform_csv_kvs": "",
                        "driver__WindowsSmbfs__smbfs_shares_config": "C:\\OpenStack\\smbfs_shares.txt",
                        "driver__XtremIOFC__driver_ssl_cert_path": "",
                        "driver__XtremIOFC__driver_ssl_cert_verify": false,
                        "driver__XtremIOFC__san_ip": "",
                        "driver__XtremIOFC__san_login": "admin",
                        "driver__XtremIOFC__san_password": "",
                        "driver__XtremIOFC__xtremio_array_busy_retry_count": 5,
                        "driver__XtremIOFC__xtremio_array_busy_retry_interval": 5,
                        "driver__XtremIOFC__xtremio_clean_unused_ig": false,
                        "driver__XtremIOFC__xtremio_cluster_name": "",
                        "driver__XtremIOISCSI__driver_ssl_cert_path": "",
                        "driver__XtremIOISCSI__driver_ssl_cert_verify": false,
                        "driver__XtremIOISCSI__san_ip": "",
                        "driver__XtremIOISCSI__san_login": "admin",
                        "driver__XtremIOISCSI__san_password": "",
                        "driver__XtremIOISCSI__xtremio_array_busy_retry_count": 5,
                        "driver__XtremIOISCSI__xtremio_array_busy_retry_interval": 5,
                        "driver__XtremIOISCSI__xtremio_clean_unused_ig": false,
                        "driver__XtremIOISCSI__xtremio_cluster_name": "",
                        "driver__ZadaraVPSAISCSI__zadara_access_key": "",
                        "driver__ZadaraVPSAISCSI__zadara_default_snap_policy": false,
                        "driver__ZadaraVPSAISCSI__zadara_ssl_cert_verify": true,
                        "driver__ZadaraVPSAISCSI__zadara_use_iser": true,
                        "driver__ZadaraVPSAISCSI__zadara_vol_encrypt": false,
                        "driver__ZadaraVPSAISCSI__zadara_vol_name_template": "OS_%s",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_host": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_poolname": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_port": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_use_ssl": false
                      }
                    },
                    "sysFiles": {
                      "name": ""
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/embercsi/ember-csi-operator:latest
          createdAt: 2021-02-22:17:47:00
          description: Multi-vendor CSI plugin supporting 80+ storage drivers
          repository: https://github.com/embercsi/ember-csi-operator
          support: http://readthedocs.org/projects/ember-csi/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an Ember CSI storage backend
            displayName: Storage Backends
            kind: EmberStorageBackend
            name: embercsis.ember-csi.io
            version: v1alpha1
        description: |
          Ember-CSI is a multi-vendor CSI plugin driver supporting
          over 80 storage drivers in a single plugin to provide block and mount
          storage to container orchestration systems.

          The operator adds support for management of multiple Ember CSI instances
          within a cluster and simplifies storage backend deployments.

          ### Supported drivers
          Ember-CSI includes a good number of storage drivers, but due to limitation on hardware availability only a small number of them have been validated at one point or another. In alphabetical order they are:

          * HPE3PARFC
          * HPE3PARISCSI
          * KaminarioISCSI
          * LVMVolume
          * PowerMaxFC
          * PowerMaxISCSI
          * PureFC
          * PureISCSI
          * QnapISCSI
          * RBD
          * SolidFire
          * SynoISCSI
          * XtremIOFC
          * XtremIOISCSI

          ### Getting started
          Please have a look at our quick start documentation on how to get started:
          [github.com/embercsi/ember-csi-operator#quick-start](https://github.com/embercsi/ember-csi-operator#quick-start).
          We recommend to use the user friendly "Edit Form" when creating a new
          EmberStorageBackend. You'll find the link in the upper right corner of the
          editor.

          ### Documentation
          Documentation on Ember CSI itself can be found at [docs.ember-csi.io](http://docs.ember-csi.io/).

          ### Getting help
          Please contact us if you encounter any issues while using Ember CSI.
          Get in touch through the #ember-csi channel on Freenode, also accessible through
          [Kiwi IRC](https://kiwiirc.com/nextclient/irc.freenode.net/ember-csi).

          You can also create an issue on the [Github project](https://github.com/embercsi/ember-csi-operator)
          for bugs, enhancements, or other requests.
        displayName: Ember CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Ember-CSI
        - CSI
        links:
        - name: Learn more about the project
          url: http://ember-csi.io/
        - name: Documentation
          url: http://readthedocs.org/projects/ember-csi/
        - name: Ember-CSI Source Code
          url: https://github.com/embercsi/ember-csi
        - name: Ember-CSI Operator Source Code
          url: https://github.com/embercsi/ember-csi-operator
        maintainers:
        - email: geguileo@redhat.com
          name: Gorka Eguileor
        - email: cschwede@redhat.com
          name: Christian Schwede
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/embercsi/ember-csi-operator:latest
        version: 0.9.9
      entries:
      - name: ember-csi-community-operator.v0.9.9
        version: 0.9.9
      - name: ember-csi-community-operator.v0.9.8
        version: 0.9.8
      - name: ember-csi-community-operator.v0.9.7
        version: 0.9.7
      name: beta
    defaultChannel: beta
    packageName: ember-csi-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sqs-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sqs-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sqs.services.k8s.aws/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sqs-controller:1.2.1
          createdAt: "2025-09-19T19:14:25Z"
          description: AWS SQS controller is a service controller for managing SQS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Queue represents the state of an AWS sqs Queue resource.
            displayName: Queue
            kind: Queue
            name: queues.sqs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Simple Queue Service (SQS) resources in AWS from within your Kubernetes cluster.

          **About Amazon SQS**

          Amazon Simple Queue Service (Amazon SQS) offers a secure, durable, and available hosted queue that lets you integrate and decouple distributed software systems and components. Amazon SQS offers common constructs such as [dead-letter queues](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGuide/sqs-dead-letter-queues.html) and [cost allocation tags](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGuide/sqs-queue-tags.html). It provides a generic web services API that you can access using any programming language that the AWS SDK supports.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SQS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sqs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SQS Developer Resources
          url: https://aws.amazon.com/sqs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sqs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sqs-controller:1.2.1
        version: 1.2.1
      entries:
      - name: ack-sqs-controller.v1.2.1
        version: 1.2.1
      - name: ack-sqs-controller.v1.2.0
        version: 1.2.0
      - name: ack-sqs-controller.v1.1.16
        version: 1.1.16
      - name: ack-sqs-controller.v1.1.14
        version: 1.1.14
      - name: ack-sqs-controller.v1.1.13
        version: 1.1.13
      - name: ack-sqs-controller.v1.1.12
        version: 1.1.12
      - name: ack-sqs-controller.v1.1.11
        version: 1.1.11
      - name: ack-sqs-controller.v1.1.10
        version: 1.1.10
      - name: ack-sqs-controller.v1.1.9
        version: 1.1.9
      - name: ack-sqs-controller.v1.1.8
        version: 1.1.8
      - name: ack-sqs-controller.v1.1.7
        version: 1.1.7
      - name: ack-sqs-controller.v1.1.6
        version: 1.1.6
      - name: ack-sqs-controller.v1.1.5
        version: 1.1.5
      - name: ack-sqs-controller.v1.1.4
        version: 1.1.4
      - name: ack-sqs-controller.v1.1.3
        version: 1.1.3
      - name: ack-sqs-controller.v1.1.2
        version: 1.1.2
      - name: ack-sqs-controller.v1.1.1
        version: 1.1.1
      - name: ack-sqs-controller.v1.1.0
        version: 1.1.0
      - name: ack-sqs-controller.v1.0.17
        version: 1.0.17
      - name: ack-sqs-controller.v1.0.16
        version: 1.0.16
      - name: ack-sqs-controller.v1.0.15
        version: 1.0.15
      - name: ack-sqs-controller.v1.0.14
        version: 1.0.14
      - name: ack-sqs-controller.v1.0.13
        version: 1.0.13
      - name: ack-sqs-controller.v1.0.12
        version: 1.0.12
      - name: ack-sqs-controller.v1.0.10
        version: 1.0.10
      - name: ack-sqs-controller.v1.0.7
        version: 1.0.7
      - name: ack-sqs-controller.v1.0.5
        version: 1.0.5
      - name: ack-sqs-controller.v1.0.3
        version: 1.0.3
      - name: ack-sqs-controller.v1.0.2
        version: 1.0.2
      - name: ack-sqs-controller.v1.0.1
        version: 1.0.1
      - name: ack-sqs-controller.v1.0.0
        version: 1.0.0
      - name: ack-sqs-controller.v0.0.7
        version: 0.0.7
      - name: ack-sqs-controller.v0.0.6
        version: 0.0.6
      - name: ack-sqs-controller.v0.0.5
        version: 0.0.5
      - name: ack-sqs-controller.v0.0.4
        version: 0.0.4
      - name: ack-sqs-controller.v0.0.2
        version: 0.0.2
      - name: ack-sqs-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-sqs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/limitador-operator
    name: limitador-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: limitador-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/limitador-operator:v0.5.0
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: The Limitador operator installs and maintains limitador instances
        displayName: Limitador
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        links:
        - name: Limitador Operator
          url: https://github.com/Kuadrant/limitador-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - quay.io/kuadrant/limitador-operator:v0.5.0
        - quay.io/kuadrant/limitador:v1.2.0
        version: 0.5.0
      entries:
      - name: limitador-operator.v0.5.0
        version: 0.5.0
      - name: limitador-operator.v0.4.0
        version: 0.4.0
      - name: limitador-operator.v0.3.0
        version: 0.3.0
      name: alpha
    - currentCSV: limitador-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "name": "toy_get_route",
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/limitador-operator:v0.11.0
          createdAt: "2024-10-02T10:46:31Z"
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: The Limitador operator installs and maintains limitador instances
        displayName: Limitador
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        links:
        - name: Limitador Operator
          url: https://github.com/Kuadrant/limitador-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - quay.io/kuadrant/limitador-operator:v0.11.0
        - quay.io/kuadrant/limitador:v1.6.0
        version: 0.11.0
      entries:
      - name: limitador-operator.v0.11.0
        version: 0.11.0
      - name: limitador-operator.v0.10.0
        version: 0.10.0
      - name: limitador-operator.v0.6.0
        version: 0.6.0
      - name: limitador-operator.v0.5.0
        version: 0.5.0
      - name: limitador-operator.v0.4.0
        version: 0.4.0
      - name: limitador-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: limitador-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/limitador-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Commvault
      provider-url: ""
    name: hedvig-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hedvig-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\":\"hedvig.io/v1\",\n    \"kind\":\"HedvigDeploy\",\n
            \   \"metadata\": {\n      \"name\": \"testdeploy\",\n      \"namespace\":
            \"hedvig\"\n    },\n    \"spec\": {\n      \"cluster\": {\n        \"name\":
            \"hedvigcl\",\n        \"seeds\": [\n          \"hedvigcl1.hedvig.io\",\n
            \         \"hedvigcl2.hedvig.io\",\n          \"hedvigcl3.hedvig.io\"\n
            \       ]\n      },\n      \"k8s\": {\n        \"name\": \"k8scl\"\n      },\n
            \     \"proxy\": {\n        \"repository\": \"hedvig/hedvig-proxy\",\n
            \       \"tag\": \"v-4.3.1\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"controller\": {\n        \"repository\": \"hedvig/hedvig-csi-volume-driver\",\n
            \       \"tag\": \"v-2.0\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"node\": {\n        \"repository\": \"hedvig/hedvig-csi-volume-driver\",\n
            \       \"tag\": \"v-2.0\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"sidecars\": [\n        {\n          \"name\": \"csi-provisioner\",\n
            \         \"repository\": \"quay.io/k8scsi/csi-provisioner\",\n          \"tag\":
            \"v1.6.0\",\n          \"imagePullPolicy\": \"IfNotPresent\"\n        },\n
            \       {\n          \"name\": \"csi-snapshotter\",\n          \"repository\":
            \"quay.io/k8scsi/csi-snapshotter\",\n          \"tag\": \"v2.1.0\",\n
            \         \"imagePullPolicy\": \"IfNotPresent\"\n        },              \n
            \       {\n          \"name\": \"csi-resizer\",\n          \"repository\":
            \"quay.io/k8scsi/csi-resizer\",\n          \"tag\": \"v0.5.0\",\n          \"imagePullPolicy\":
            \"IfNotPresent\"\n        },              \n        {\n          \"name\":
            \"csi-node-driver-registrar\",\n          \"repository\": \"quay.io/k8scsi/csi-node-driver-registrar\",\n
            \         \"tag\": \"v1.0-canary\",\n          \"imagePullPolicy\": \"IfNotPresent\"\n
            \       }\n      ]\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: hedvig/hedvig-operator:2.0.0-ubi
          createdAt: ""
          description: Hedvig Operator is the official operator to deploy and manage
            Hedvig components in container orchestrators such as Kubernetes and Openshift.
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hedvig Deploy
            displayName: HedvigDeploy
            kind: HedvigDeploy
            name: hedvigdeploys.hedvig.io
            version: v1
        description: "The Hedvig Distributed Storage Platform is a software-defined,
          scale-out storage solution. It distributes data simultaneously across multiple
          locations – from on-premises data centers to the cloud – and scales capacity
          on-demand by leveraging the storage of commodity x86/ARM servers. The distributed
          write maximizes availability and protects data from hardware failures ranging
          from a single disk to an entire site, improving disaster recovery planning.\n\nHedvig
          provides a CSI Driver for accelerating your journey into the container ecosystem.
          CSI Driver enables containerized applications to maintain state by dynamically
          provisioning and consuming Hedvig virtual disks as persistent volumes. \n###
          \n\nHedvig Operator deploys the Hedvig Storage Proxy and the Hedvig CSI
          Driver to manage the life cycle of persistent storage in Kubernetes and
          Openshift.\n\n**Supported Container Orchestrators:**\n* Kubernetes 1.16+\n*
          Red Hat OpenShift 4.6+\n\n**Supported Hedvig Releases:**\n* Hedvig 4.0+\nPlease
          refer to the [documentation](https://documentation.commvault.com/commvault/hedvig/others/pdf/Hedvig_Operator_User_Guide.pdf)
          for steps to install and use the operator.\n\n\n\n\n\n"
        displayName: Hedvig Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - CSI
        - Hedvig
        - Cloud
        - Block
        - NFS
        - iSCSI
        - Snapshot
        - Volume
        links:
        - name: Documentation
          url: https://documentation.commvault.com/commvault/hedvig/article
        - name: Hedvig CSI User Guide
          url: https://documentation.commvault.com/commvault/hedvig/others/pdf/Hedvig_CSI_User_Guide.pdf
        maintainers:
        - email: ashenoy@commvault.com
          name: Abhijith Shenoy
        maturity: stable
        provider:
          name: Commvault
        relatedImages:
        - hedvig/hedvig-operator:2.0.0-ubi
        version: 1.0.2
      entries:
      - name: hedvig-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: hedvig-operator
    provider:
      name: Commvault
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: External Secrets
      provider-url: https://external-secrets.io
    name: external-secrets-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: external-secrets-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": "1234"
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "key": "provider-key",
                      "property": "provider-key-property",
                      "version": "provider-key-version"
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterExternalSecret",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "externalSecretName": "hello-world-es",
                  "externalSecretSpec": {
                    "data": [
                      {
                        "remoteRef": {
                          "key": "provider-key",
                          "property": "provider-key-property",
                          "version": "provider-key-version"
                        },
                        "secretKey": "secret-key-to-be-managed"
                      }
                    ],
                    "dataFrom": [
                      {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    ],
                    "refreshInterval": "1h",
                    "secretStoreRef": {
                      "kind": "SecretStore",
                      "name": "secret-store-name"
                    },
                    "target": {
                      "creationPolicy": "Merge",
                      "name": "my-secret",
                      "template": {
                        "data": {
                          "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                        },
                        "metadata": {
                          "annotations": {},
                          "labels": {}
                        },
                        "templateFrom": [
                          {
                            "configMap": {
                              "items": [
                                {
                                  "key": "alertmanager.yaml"
                                }
                              ],
                              "name": "alertmanager"
                            }
                          }
                        ],
                        "type": "kubernetes.io/dockerconfigjson"
                      }
                    }
                  },
                  "namespaceSelector": {
                    "matchLabels": {
                      "cool": "label"
                    }
                  },
                  "refreshTime": "1m"
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": 1234
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "extract": {
                        "conversionStrategy": "Default",
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    },
                    {
                      "find": {
                        "conversionStrategy": "Unicode",
                        "name": {
                          "regexp": ".*foobar.*"
                        },
                        "path": "path-to-filter",
                        "tags": {
                          "foo": "bar"
                        }
                      }
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "deletionPolicy": "Retain",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.external-secrets.io/v1alpha1",
                "kind": "OperatorConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "requeueInterval": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  },
                  "concurrent": 1,
                  "controllerClass": "",
                  "crds": {
                    "createClusterExternalSecret": true,
                    "createClusterSecretStore": true
                  },
                  "createOperator": true,
                  "deploymentAnnotations": {},
                  "extraArgs": {},
                  "extraEnv": [],
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/external-secrets/external-secrets",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "installCRDs": false,
                  "leaderElect": false,
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "priorityClassName": "",
                  "processClusterExternalSecret": true,
                  "processClusterStore": true,
                  "prometheus": {
                    "enabled": false,
                    "service": {
                      "port": 8080
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "scopedNamespace": "",
                  "scopedRBAC": false,
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "certCheckInterval": "5m",
                    "certDir": "/tmp/certs",
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
          createdAt: "2024-12-03T10:44:19Z"
          description: Operator to configure external-secrets helm-chart based operator
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/external-secrets/external-secrets-helm-operator
          support: external-secrets
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ACRAccessToken
            name: acraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster scoped resource that allows you to manage ExternalSecrets
              across namespaces
            displayName: ClusterExternalSecret
            kind: ClusterExternalSecret
            name: clusterexternalsecrets.external-secrets.io
            version: v1beta1
          - kind: ClusterGenerator
            name: clustergenerators.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1alpha1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1beta1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1beta1
          - kind: ECRAuthorizationToken
            name: ecrauthorizationtokens.generators.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1alpha1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1beta1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1beta1
          - kind: Fake
            name: fakes.generators.external-secrets.io
            version: v1alpha1
          - kind: GCRAccessToken
            name: gcraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - kind: GithubAccessToken
            name: githubaccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Configures external-secrets helm-chart based operator
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.external-secrets.io
            version: v1alpha1
          - kind: Password
            name: passwords.generators.external-secrets.io
            version: v1alpha1
          - kind: PushSecret
            name: pushsecrets.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1alpha1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1beta1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1beta1
          - kind: STSSessionToken
            name: stssessiontokens.generators.external-secrets.io
            version: v1alpha1
          - kind: UUID
            name: uuids.generators.external-secrets.io
            version: v1alpha1
          - kind: VaultDynamicSecret
            name: vaultdynamicsecrets.generators.external-secrets.io
            version: v1alpha1
          - kind: Webhook
            name: webhooks.generators.external-secrets.io
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK (Helm version) to configure **[official external-secrets operator helm chart](https://github.com/external-secrets/external-secrets)**, so it can be installed via OLM without having to do any change on current Helm Charts.

          Before any other resources provided by this Operator can be deployed, it is essential to create an
          OperatorConfig resource.

          The usual Helm Chart file `values.yaml`, like:
          ```yaml
          prometheus:
            enabled: true
            service:
              port: 8080
          resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          need to be encapsulated into a new custom resource called `OperatorConfig`:
          ```yaml
          apiVersion: operator.external-secrets.io/v1alpha1
          kind: OperatorConfig
          metadata:
            name: cluster
          spec:
            prometheus:
              enabled: true
              service:
                port: 8080
            resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          Once the OperatorConfig resource is deployed, the operator will create all helm chart resources, using the custom resource name as a preffix for all resources names, like for example a `Deployment` called `cluster-external-secrets`.

          ## Documentation

          * [Upstream Documentation](https://external-secrets.io)
          * [Helm Operator Install](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/install.md)
          * [Helm Operator Development](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/development.md)
          * [Helm Operator Release](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/release.md)

          ## Contributing

          You can contribute by:

          * Raising any issues you find using External Secrets Operator
          * Fixing issues by opening [Pull Requests](https://github.com/external-secrets/external-secrets-helm-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving documentation
          * Talking about External Secrets Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/external-secrets/external-secrets-helm-operator/issues).

          ## License

          External Secrets Operator is under Apache 2.0 license. See the [LICENSE](LICENSE) file for details.
        displayName: External Secrets Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secrets
        - secrets-manager
        - vault
        - aws
        - azure
        - google
        - ibm
        - akeyless
        - yandex
        - gitlab
        - oracle
        links:
        - name: External Secrets Operator
          url: https://external-secrets.io
        - name: GitHub
          url: https://github.com/external-secrets/external-secrets
        - name: Slack
          url: https://kubernetes.slack.com/messages/external-secrets
        maintainers:
        - email: contact@external-secrets.io
          name: external-secrets community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: External Secrets
          url: https://external-secrets.io
        relatedImages:
        - ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
        version: 0.11.0
      entries:
      - name: external-secrets-operator.v0.11.0
        version: 0.11.0
      - name: external-secrets-operator.v0.10.7
        version: 0.10.7
      - name: external-secrets-operator.v0.10.6
        version: 0.10.6
      - name: external-secrets-operator.v0.10.5
        version: 0.10.5
      - name: external-secrets-operator.v0.10.4
        version: 0.10.4
      - name: external-secrets-operator.v0.10.3
        version: 0.10.3
      - name: external-secrets-operator.v0.10.2
        version: 0.10.2
      - name: external-secrets-operator.v0.10.1
        version: 0.10.1
      - name: external-secrets-operator.v0.10.0
        version: 0.10.0
      - name: external-secrets-operator.v0.9.20
        version: 0.9.20
      - name: external-secrets-operator.v0.9.19
        version: 0.9.19
      - name: external-secrets-operator.v0.9.18
        version: 0.9.18
      - name: external-secrets-operator.v0.9.17
        version: 0.9.17
      - name: external-secrets-operator.v0.9.16
        version: 0.9.16
      - name: external-secrets-operator.v0.9.15-2
        version: 0.9.15-2
      - name: external-secrets-operator.v0.9.14
        version: 0.9.14
      - name: external-secrets-operator.v0.9.13
        version: 0.9.13
      - name: external-secrets-operator.v0.9.12
        version: 0.9.12
      - name: external-secrets-operator.v0.9.11
        version: 0.9.11
      - name: external-secrets-operator.v0.9.10
        version: 0.9.10
      - name: external-secrets-operator.v0.9.9
        version: 0.9.9
      - name: external-secrets-operator.v0.9.8
        version: 0.9.8
      - name: external-secrets-operator.v0.9.7
        version: 0.9.7
      - name: external-secrets-operator.v0.9.6
        version: 0.9.6
      - name: external-secrets-operator.v0.9.5
        version: 0.9.5
      - name: external-secrets-operator.v0.9.4
        version: 0.9.4
      - name: external-secrets-operator.v0.9.3
        version: 0.9.3
      - name: external-secrets-operator.v0.9.2
        version: 0.9.2
      - name: external-secrets-operator.v0.9.1
        version: 0.9.1
      - name: external-secrets-operator.v0.9.0
        version: 0.9.0
      - name: external-secrets-operator.v0.8.7
        version: 0.8.7
      - name: external-secrets-operator.v0.8.6
        version: 0.8.6
      - name: external-secrets-operator.v0.8.5
        version: 0.8.5
      - name: external-secrets-operator.v0.8.4
        version: 0.8.4
      - name: external-secrets-operator.v0.8.3
        version: 0.8.3
      - name: external-secrets-operator.v0.8.2
        version: 0.8.2
      - name: external-secrets-operator.v0.8.1
        version: 0.8.1
      - name: external-secrets-operator.v0.8.0
        version: 0.8.0
      - name: external-secrets-operator.v0.7.2
        version: 0.7.2
      - name: external-secrets-operator.v0.7.1
        version: 0.7.1
      - name: external-secrets-operator.v0.7.0
        version: 0.7.0
      - name: external-secrets-operator.v0.7.0-rc1
        version: 0.7.0-rc1
      - name: external-secrets-operator.v0.6.1
        version: 0.6.1
      - name: external-secrets-operator.v0.6.0
        version: 0.6.0
      - name: external-secrets-operator.v0.6.0-rc1
        version: 0.6.0-rc1
      - name: external-secrets-operator.v0.5.9
        version: 0.5.9
      - name: external-secrets-operator.v0.5.8
        version: 0.5.8
      - name: external-secrets-operator.v0.5.7
        version: 0.5.7
      - name: external-secrets-operator.v0.5.6
        version: 0.5.6
      - name: external-secrets-operator.v0.5.5
        version: 0.5.5
      - name: external-secrets-operator.v0.5.4
        version: 0.5.4
      - name: external-secrets-operator.v0.5.3
        version: 0.5.3
      - name: external-secrets-operator.v0.5.2
        version: 0.5.2
      - name: external-secrets-operator.v0.5.1
        version: 0.5.1
      - name: external-secrets-operator.v0.5.0
        version: 0.5.0
      - name: external-secrets-operator.v0.4.4
        version: 0.4.4
      - name: external-secrets-operator.v0.4.3
        version: 0.4.3
      - name: external-secrets-operator.v0.4.2
        version: 0.4.2
      - name: external-secrets-operator.v0.4.1
        version: 0.4.1
      - name: external-secrets-operator.v0.3.11
        version: 0.3.11
      - name: external-secrets-operator.v0.3.10
        version: 0.3.10
      - name: external-secrets-operator.v0.3.9
        version: 0.3.9
      - name: external-secrets-operator.v0.3.8
        version: 0.3.8
      name: alpha
    - currentCSV: external-secrets-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": "1234"
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "key": "provider-key",
                      "property": "provider-key-property",
                      "version": "provider-key-version"
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterExternalSecret",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "externalSecretName": "hello-world-es",
                  "externalSecretSpec": {
                    "data": [
                      {
                        "remoteRef": {
                          "key": "provider-key",
                          "property": "provider-key-property",
                          "version": "provider-key-version"
                        },
                        "secretKey": "secret-key-to-be-managed"
                      }
                    ],
                    "dataFrom": [
                      {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    ],
                    "refreshInterval": "1h",
                    "secretStoreRef": {
                      "kind": "SecretStore",
                      "name": "secret-store-name"
                    },
                    "target": {
                      "creationPolicy": "Merge",
                      "name": "my-secret",
                      "template": {
                        "data": {
                          "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                        },
                        "metadata": {
                          "annotations": {},
                          "labels": {}
                        },
                        "templateFrom": [
                          {
                            "configMap": {
                              "items": [
                                {
                                  "key": "alertmanager.yaml"
                                }
                              ],
                              "name": "alertmanager"
                            }
                          }
                        ],
                        "type": "kubernetes.io/dockerconfigjson"
                      }
                    }
                  },
                  "namespaceSelector": {
                    "matchLabels": {
                      "cool": "label"
                    }
                  },
                  "refreshTime": "1m"
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": 1234
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "extract": {
                        "conversionStrategy": "Default",
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    },
                    {
                      "find": {
                        "conversionStrategy": "Unicode",
                        "name": {
                          "regexp": ".*foobar.*"
                        },
                        "path": "path-to-filter",
                        "tags": {
                          "foo": "bar"
                        }
                      }
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "deletionPolicy": "Retain",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.external-secrets.io/v1alpha1",
                "kind": "OperatorConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "requeueInterval": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  },
                  "concurrent": 1,
                  "controllerClass": "",
                  "crds": {
                    "createClusterExternalSecret": true,
                    "createClusterSecretStore": true
                  },
                  "createOperator": true,
                  "deploymentAnnotations": {},
                  "extraArgs": {},
                  "extraEnv": [],
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/external-secrets/external-secrets",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "installCRDs": false,
                  "leaderElect": false,
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "priorityClassName": "",
                  "processClusterExternalSecret": true,
                  "processClusterStore": true,
                  "prometheus": {
                    "enabled": false,
                    "service": {
                      "port": 8080
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "scopedNamespace": "",
                  "scopedRBAC": false,
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "certCheckInterval": "5m",
                    "certDir": "/tmp/certs",
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
          createdAt: "2024-12-03T10:44:19Z"
          description: Operator to configure external-secrets helm-chart based operator
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/external-secrets/external-secrets-helm-operator
          support: external-secrets
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ACRAccessToken
            name: acraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster scoped resource that allows you to manage ExternalSecrets
              across namespaces
            displayName: ClusterExternalSecret
            kind: ClusterExternalSecret
            name: clusterexternalsecrets.external-secrets.io
            version: v1beta1
          - kind: ClusterGenerator
            name: clustergenerators.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1alpha1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1beta1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1beta1
          - kind: ECRAuthorizationToken
            name: ecrauthorizationtokens.generators.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1alpha1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1beta1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1beta1
          - kind: Fake
            name: fakes.generators.external-secrets.io
            version: v1alpha1
          - kind: GCRAccessToken
            name: gcraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - kind: GithubAccessToken
            name: githubaccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Configures external-secrets helm-chart based operator
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.external-secrets.io
            version: v1alpha1
          - kind: Password
            name: passwords.generators.external-secrets.io
            version: v1alpha1
          - kind: PushSecret
            name: pushsecrets.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1alpha1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1beta1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1beta1
          - kind: STSSessionToken
            name: stssessiontokens.generators.external-secrets.io
            version: v1alpha1
          - kind: UUID
            name: uuids.generators.external-secrets.io
            version: v1alpha1
          - kind: VaultDynamicSecret
            name: vaultdynamicsecrets.generators.external-secrets.io
            version: v1alpha1
          - kind: Webhook
            name: webhooks.generators.external-secrets.io
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK (Helm version) to configure **[official external-secrets operator helm chart](https://github.com/external-secrets/external-secrets)**, so it can be installed via OLM without having to do any change on current Helm Charts.

          Before any other resources provided by this Operator can be deployed, it is essential to create an
          OperatorConfig resource.

          The usual Helm Chart file `values.yaml`, like:
          ```yaml
          prometheus:
            enabled: true
            service:
              port: 8080
          resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          need to be encapsulated into a new custom resource called `OperatorConfig`:
          ```yaml
          apiVersion: operator.external-secrets.io/v1alpha1
          kind: OperatorConfig
          metadata:
            name: cluster
          spec:
            prometheus:
              enabled: true
              service:
                port: 8080
            resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          Once the OperatorConfig resource is deployed, the operator will create all helm chart resources, using the custom resource name as a preffix for all resources names, like for example a `Deployment` called `cluster-external-secrets`.

          ## Documentation

          * [Upstream Documentation](https://external-secrets.io)
          * [Helm Operator Install](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/install.md)
          * [Helm Operator Development](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/development.md)
          * [Helm Operator Release](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/release.md)

          ## Contributing

          You can contribute by:

          * Raising any issues you find using External Secrets Operator
          * Fixing issues by opening [Pull Requests](https://github.com/external-secrets/external-secrets-helm-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving documentation
          * Talking about External Secrets Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/external-secrets/external-secrets-helm-operator/issues).

          ## License

          External Secrets Operator is under Apache 2.0 license. See the [LICENSE](LICENSE) file for details.
        displayName: External Secrets Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secrets
        - secrets-manager
        - vault
        - aws
        - azure
        - google
        - ibm
        - akeyless
        - yandex
        - gitlab
        - oracle
        links:
        - name: External Secrets Operator
          url: https://external-secrets.io
        - name: GitHub
          url: https://github.com/external-secrets/external-secrets
        - name: Slack
          url: https://kubernetes.slack.com/messages/external-secrets
        maintainers:
        - email: contact@external-secrets.io
          name: external-secrets community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: External Secrets
          url: https://external-secrets.io
        relatedImages:
        - ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
        version: 0.11.0
      entries:
      - name: external-secrets-operator.v0.11.0
        version: 0.11.0
      - name: external-secrets-operator.v0.10.7
        version: 0.10.7
      - name: external-secrets-operator.v0.10.6
        version: 0.10.6
      - name: external-secrets-operator.v0.10.5
        version: 0.10.5
      - name: external-secrets-operator.v0.10.4
        version: 0.10.4
      - name: external-secrets-operator.v0.10.3
        version: 0.10.3
      - name: external-secrets-operator.v0.10.2
        version: 0.10.2
      - name: external-secrets-operator.v0.10.1
        version: 0.10.1
      - name: external-secrets-operator.v0.10.0
        version: 0.10.0
      - name: external-secrets-operator.v0.9.20
        version: 0.9.20
      - name: external-secrets-operator.v0.9.19
        version: 0.9.19
      - name: external-secrets-operator.v0.9.18
        version: 0.9.18
      - name: external-secrets-operator.v0.9.17
        version: 0.9.17
      - name: external-secrets-operator.v0.9.16
        version: 0.9.16
      - name: external-secrets-operator.v0.9.15-2
        version: 0.9.15-2
      - name: external-secrets-operator.v0.9.14
        version: 0.9.14
      - name: external-secrets-operator.v0.9.13
        version: 0.9.13
      - name: external-secrets-operator.v0.9.12
        version: 0.9.12
      - name: external-secrets-operator.v0.9.11
        version: 0.9.11
      - name: external-secrets-operator.v0.9.10
        version: 0.9.10
      - name: external-secrets-operator.v0.9.9
        version: 0.9.9
      - name: external-secrets-operator.v0.9.8
        version: 0.9.8
      - name: external-secrets-operator.v0.9.7
        version: 0.9.7
      - name: external-secrets-operator.v0.9.6
        version: 0.9.6
      - name: external-secrets-operator.v0.9.5
        version: 0.9.5
      - name: external-secrets-operator.v0.9.4
        version: 0.9.4
      - name: external-secrets-operator.v0.9.3
        version: 0.9.3
      - name: external-secrets-operator.v0.9.2
        version: 0.9.2
      - name: external-secrets-operator.v0.9.1
        version: 0.9.1
      - name: external-secrets-operator.v0.9.0
        version: 0.9.0
      - name: external-secrets-operator.v0.8.7
        version: 0.8.7
      - name: external-secrets-operator.v0.8.6
        version: 0.8.6
      - name: external-secrets-operator.v0.8.5
        version: 0.8.5
      - name: external-secrets-operator.v0.8.4
        version: 0.8.4
      - name: external-secrets-operator.v0.8.3
        version: 0.8.3
      - name: external-secrets-operator.v0.8.2
        version: 0.8.2
      - name: external-secrets-operator.v0.8.1
        version: 0.8.1
      - name: external-secrets-operator.v0.8.0
        version: 0.8.0
      - name: external-secrets-operator.v0.7.2
        version: 0.7.2
      - name: external-secrets-operator.v0.7.1
        version: 0.7.1
      - name: external-secrets-operator.v0.7.0
        version: 0.7.0
      - name: external-secrets-operator.v0.7.0-rc1
        version: 0.7.0-rc1
      - name: external-secrets-operator.v0.6.1
        version: 0.6.1
      - name: external-secrets-operator.v0.6.0
        version: 0.6.0
      - name: external-secrets-operator.v0.5.9
        version: 0.5.9
      - name: external-secrets-operator.v0.5.8
        version: 0.5.8
      - name: external-secrets-operator.v0.5.7
        version: 0.5.7
      - name: external-secrets-operator.v0.5.6
        version: 0.5.6
      - name: external-secrets-operator.v0.5.5
        version: 0.5.5
      - name: external-secrets-operator.v0.5.4
        version: 0.5.4
      - name: external-secrets-operator.v0.5.3
        version: 0.5.3
      - name: external-secrets-operator.v0.5.2
        version: 0.5.2
      - name: external-secrets-operator.v0.5.1
        version: 0.5.1
      - name: external-secrets-operator.v0.5.0
        version: 0.5.0
      - name: external-secrets-operator.v0.4.4
        version: 0.4.4
      - name: external-secrets-operator.v0.4.2
        version: 0.4.2
      - name: external-secrets-operator.v0.4.1
        version: 0.4.1
      - name: external-secrets-operator.v0.3.11
        version: 0.3.11
      - name: external-secrets-operator.v0.3.10
        version: 0.3.10
      - name: external-secrets-operator.v0.3.9
        version: 0.3.9
      name: stable
    defaultChannel: alpha
    packageName: external-secrets-operator
    provider:
      name: External Secrets
      url: https://external-secrets.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/kuadrant-operator
    name: rhcl-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhcl-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "authpolicy-sample"
                },
                "spec": {
                  "rules": {
                    "authentication": {
                      "apikey": {
                        "apiKey": {
                          "selector": {}
                        },
                        "credentials": {
                          "authorizationHeader": {
                            "prefix": "APIKEY"
                          }
                        }
                      }
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "DNSPolicy",
                "metadata": {
                  "name": "dnspolicy-sample"
                },
                "spec": {
                  "healthCheck": {
                    "protocol": "HTTP"
                  },
                  "providerRefs": [
                    {
                      "name": "provider-ref"
                    }
                  ],
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "ratelimitpolicy-sample"
                },
                "spec": {
                  "limits": {
                    "toys": {
                      "rates": [
                        {
                          "limit": 50,
                          "window": "1m"
                        }
                      ]
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "TLSPolicy",
                "metadata": {
                  "name": "tlspolicy-sample"
                },
                "spec": {
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "self-signed-ca"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/rhcl-rhel9-operator@sha256:92f14fd8de74bc72c407947cf5fa4739a7b455e5d351baeb64521e4c07f02a9e
          createdAt: "2025-04-28T13:10:44Z"
          description: Red Hat Connectivity Link enables you to secure, protect, and
            connect your APIs and applications in multicluster, multicloud, and hybrid
            cloud environments
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthPolicy enables authentication and authorization for service
              workloads in a Gateway API network
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1
          - description: DNSPolicy configures how North-South based traffic should
              be balanced and reach the gateways
            displayName: DNSPolicy
            kind: DNSPolicy
            name: dnspolicies.kuadrant.io
            version: v1
          - description: Kuadrant configures installations of Kuadrant Service Protection
              components
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: RateLimitPolicy enables rate limiting for service workloads
              in a Gateway API network
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1
          - description: TLSPolicy provides tls for gateway listeners by managing
              the lifecycle of tls certificates
            displayName: TLSPolicy
            kind: TLSPolicy
            name: tlspolicies.kuadrant.io
            version: v1
        description: Red Hat Connectivity Link enables you to secure, protect, and
          connect your APIs and applications in multicluster, multicloud, and hybrid
          cloud environments
        displayName: Red Hat Connectivity Link
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: GitHub
          url: https://github.com/kuadrant/kuadrant-operator
        - name: Documentation
          url: https://github.com/Kuadrant/kuadrant-operator/tree/main/docs
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - registry.redhat.io/rhcl-1/rhcl-console-plugin-rhel9@sha256:ae92d638d993d782986e761d5aa3c7c5da2b8af0aabc1decb332a4f1487dacd3
        - registry.redhat.io/rhcl-1/rhcl-rhel9-operator@sha256:92f14fd8de74bc72c407947cf5fa4739a7b455e5d351baeb64521e4c07f02a9e
        - registry.redhat.io/rhcl-1/wasm-shim-rhel9@sha256:845bb8af57f3d219aa09b9c0bb20fa945306f3b413e34e196760c8cdf624a532
        version: 1.1.0
      entries:
      - name: rhcl-operator.v1.1.0
        version: 1.1.0
      - name: rhcl-operator.v1.0.2
        version: 1.0.2
      - name: rhcl-operator.v1.0.1
        version: 1.0.1
      - name: rhcl-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: rhcl-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/kuadrant-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: akaris
      provider-url: https://github.com/andreaskaris/sosreport-operator
    name: sosreport-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sosreport-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "support.openshift.io/v1alpha1",
                "kind": "Sosreport",
                "metadata": {
                  "name": "sosreport-sample"
                },
                "spec": {
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Sosreport is the Schema for the sosreports API
            displayName: Sosreport
            kind: Sosreport
            name: sosreports.support.openshift.io
            version: v1alpha1
        description: Operator that facilitates the generation of sosreports
        displayName: sosreport-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sosreport
        - debugging
        - troubleshooting
        links:
        - name: Sosreport Operator
          url: https://github.com/andreaskaris/sosreport-operator
        maintainers:
        - email: ak.karis@gmail.com
          name: Andreas Karis
        maturity: alpha
        provider:
          name: akaris
          url: https://github.com/andreaskaris/sosreport-operator
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/akaris/sosreport-operator:0.0.2
        version: 0.0.2
      entries:
      - name: sosreport-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: sosreport-operator
    provider:
      name: akaris
      url: https://github.com/andreaskaris/sosreport-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-rds-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-rds-controller.v1.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rds.services.k8s.aws/v1alpha1",
                "kind": "DBParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "rds.services.k8s.aws/v1alpha1",
                "kind": "DBSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/rds-controller:1.6.2
          createdAt: "2025-09-22T17:27:42Z"
          description: AWS RDS controller is a service controller for managing RDS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DBClusterEndpoint represents the state of an AWS rds DBClusterEndpoint
              resource.
            displayName: DBClusterEndpoint
            kind: DBClusterEndpoint
            name: dbclusterendpoints.rds.services.k8s.aws
            version: v1alpha1
          - description: DBClusterParameterGroup represents the state of an AWS rds
              DBClusterParameterGroup resource.
            displayName: DBClusterParameterGroup
            kind: DBClusterParameterGroup
            name: dbclusterparametergroups.rds.services.k8s.aws
            version: v1alpha1
          - description: DBCluster represents the state of an AWS rds DBCluster resource.
            displayName: DBCluster
            kind: DBCluster
            name: dbclusters.rds.services.k8s.aws
            version: v1alpha1
          - description: DBClusterSnapshot represents the state of an AWS rds DBClusterSnapshot
              resource.
            displayName: DBClusterSnapshot
            kind: DBClusterSnapshot
            name: dbclustersnapshots.rds.services.k8s.aws
            version: v1alpha1
          - description: DBInstance represents the state of an AWS rds DBInstance
              resource.
            displayName: DBInstance
            kind: DBInstance
            name: dbinstances.rds.services.k8s.aws
            version: v1alpha1
          - description: DBParameterGroup represents the state of an AWS rds DBParameterGroup
              resource.
            displayName: DBParameterGroup
            kind: DBParameterGroup
            name: dbparametergroups.rds.services.k8s.aws
            version: v1alpha1
          - description: DBProxy represents the state of an AWS rds DBProxy resource.
            displayName: DBProxy
            kind: DBProxy
            name: dbproxies.rds.services.k8s.aws
            version: v1alpha1
          - description: DBSnapshot represents the state of an AWS rds DBSnapshot
              resource.
            displayName: DBSnapshot
            kind: DBSnapshot
            name: dbsnapshots.rds.services.k8s.aws
            version: v1alpha1
          - description: DBSubnetGroup represents the state of an AWS rds DBSubnetGroup
              resource.
            displayName: DBSubnetGroup
            kind: DBSubnetGroup
            name: dbsubnetgroups.rds.services.k8s.aws
            version: v1alpha1
          - description: GlobalCluster represents the state of an AWS rds GlobalCluster
              resource.
            displayName: GlobalCluster
            kind: GlobalCluster
            name: globalclusters.rds.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Relational Database Service ("RDS") resources in AWS from within your Kubernetes cluster.

          **About Amazon RDS**

          Amazon Relational Database Service (Amazon RDS) makes it easy to set up, operate, and scale a relational database in the cloud. It provides cost-efficient and resizable capacity while automating time-consuming administration tasks such as hardware provisioning, database setup, patching and backups. It frees you to focus on your applications so you can give them the fast performance, high availability, security and compatibility they need.

          Amazon RDS is available on several database instance types - optimized for memory, performance or I/O -and provides you with six familiar database engines to choose from, including Amazon Aurora, PostgreSQL, MySQL, MariaDB, Oracle Database, and SQL Server. You can use the AWS Database Migration Service to easily migrate or replicate your existing databases to Amazon RDS.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon RDS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rds
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon RDS Developer Resources
          url: https://aws.amazon.com/rds/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: rds maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/rds-controller:1.6.2
        version: 1.6.2
      entries:
      - name: ack-rds-controller.v1.6.2
        version: 1.6.2
      - name: ack-rds-controller.v1.6.1
        version: 1.6.1
      - name: ack-rds-controller.v1.6.0
        version: 1.6.0
      - name: ack-rds-controller.v1.5.0
        version: 1.5.0
      - name: ack-rds-controller.v1.4.26
        version: 1.4.26
      - name: ack-rds-controller.v1.4.25
        version: 1.4.25
      - name: ack-rds-controller.v1.4.24
        version: 1.4.24
      - name: ack-rds-controller.v1.4.22
        version: 1.4.22
      - name: ack-rds-controller.v1.4.21
        version: 1.4.21
      - name: ack-rds-controller.v1.4.20
        version: 1.4.20
      - name: ack-rds-controller.v1.4.19
        version: 1.4.19
      - name: ack-rds-controller.v1.4.18
        version: 1.4.18
      - name: ack-rds-controller.v1.4.17
        version: 1.4.17
      - name: ack-rds-controller.v1.4.16
        version: 1.4.16
      - name: ack-rds-controller.v1.4.15
        version: 1.4.15
      - name: ack-rds-controller.v1.4.14
        version: 1.4.14
      - name: ack-rds-controller.v1.4.13
        version: 1.4.13
      - name: ack-rds-controller.v1.4.12
        version: 1.4.12
      - name: ack-rds-controller.v1.4.11
        version: 1.4.11
      - name: ack-rds-controller.v1.4.10
        version: 1.4.10
      - name: ack-rds-controller.v1.4.8
        version: 1.4.8
      - name: ack-rds-controller.v1.4.7
        version: 1.4.7
      - name: ack-rds-controller.v1.4.6
        version: 1.4.6
      - name: ack-rds-controller.v1.4.4
        version: 1.4.4
      - name: ack-rds-controller.v1.4.3
        version: 1.4.3
      - name: ack-rds-controller.v1.4.2
        version: 1.4.2
      - name: ack-rds-controller.v1.4.1
        version: 1.4.1
      - name: ack-rds-controller.v1.4.0
        version: 1.4.0
      - name: ack-rds-controller.v1.3.0
        version: 1.3.0
      - name: ack-rds-controller.v1.2.5
        version: 1.2.5
      - name: ack-rds-controller.v1.2.4
        version: 1.2.4
      - name: ack-rds-controller.v1.2.3
        version: 1.2.3
      - name: ack-rds-controller.v1.2.2
        version: 1.2.2
      - name: ack-rds-controller.v1.2.1
        version: 1.2.1
      - name: ack-rds-controller.v1.2.0
        version: 1.2.0
      - name: ack-rds-controller.v1.1.11
        version: 1.1.11
      - name: ack-rds-controller.v1.1.9
        version: 1.1.9
      - name: ack-rds-controller.v1.1.8
        version: 1.1.8
      - name: ack-rds-controller.v1.1.7
        version: 1.1.7
      - name: ack-rds-controller.v1.1.6
        version: 1.1.6
      - name: ack-rds-controller.v1.1.5
        version: 1.1.5
      - name: ack-rds-controller.v1.1.4
        version: 1.1.4
      - name: ack-rds-controller.v1.1.3
        version: 1.1.3
      - name: ack-rds-controller.v1.1.2
        version: 1.1.2
      - name: ack-rds-controller.v1.1.1
        version: 1.1.1
      - name: ack-rds-controller.v1.1.0
        version: 1.1.0
      - name: ack-rds-controller.v1.0.0
        version: 1.0.0
      - name: ack-rds-controller.v0.1.3
        version: 0.1.3
      - name: ack-rds-controller.v0.1.1
        version: 0.1.1
      - name: ack-rds-controller.v0.1.0
        version: 0.1.0
      - name: ack-rds-controller.v0.0.30
        version: 0.0.30
      - name: ack-rds-controller.v0.0.29
        version: 0.0.29
      - name: ack-rds-controller.v0.0.28
        version: 0.0.28
      - name: ack-rds-controller.v0.0.27
        version: 0.0.27
      - name: ack-rds-controller.v0.0.26
        version: 0.0.26
      - name: ack-rds-controller.v0.0.25
        version: 0.0.25
      - name: ack-rds-controller.v0.0.24
        version: 0.0.24
      - name: ack-rds-controller.v0.0.23
        version: 0.0.23
      - name: ack-rds-controller.v0.0.22
        version: 0.0.22
      - name: ack-rds-controller.v0.0.21
        version: 0.0.21
      - name: ack-rds-controller.v0.0.20
        version: 0.0.20
      - name: ack-rds-controller.v0.0.19
        version: 0.0.19
      - name: ack-rds-controller.v0.0.18
        version: 0.0.18
      - name: ack-rds-controller.v0.0.17
        version: 0.0.17
      - name: ack-rds-controller.v0.0.16
        version: 0.0.16
      - name: ack-rds-controller.v0.0.15
        version: 0.0.15
      - name: ack-rds-controller.v0.0.14
        version: 0.0.14
      - name: ack-rds-controller.v0.0.13
        version: 0.0.13
      - name: ack-rds-controller.v0.0.12
        version: 0.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-rds-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: integrity-shield-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: integrity-shield-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/open-cluster-management/integrity-shield-observer",
                    "imagePullPolicy": "Always",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "provenanece": true,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  resp.status_code == 200\n  result := json.unmarshal(resp.raw_body)\n  result.allow == false\n  not is_detect_mode\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"tls_insecure_skip_verify\": true\n  })\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "requestHandlerConfig": "defaultConstraintAction:\n  audit:\n    inform: true\n  admissionControl:\n    enforce: false\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/open-cluster-management/integrity-shield-api",
                    "imagePullPolicy": "Always",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/open-cluster-management/integrity-shield-operator:0.3.0
          createdAt: "2021-10-14T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.

          [K8s Integrity Shield](https://github.com/open-cluster-management/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.

          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Detect (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster

          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).

          ## Preparations before installation

          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).


          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.

          If you want to change the settings such as default run mode (detection/enforcement) or audit interval,  please check [here](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).

          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-7b7f768bf7-fhrpg                           1/1     Running   0          20s
          integrity-shield-observer-5bc66f75f7-tn8fw                      1/1     Running   0          25s
          integrity-shield-operator-controller-manager-65b7fb58f7-j25zd   2/2     Running   0          3h5m
          ```

          After installation, you can protect Kubernetes resources by following this [document](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).

          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:

          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)

          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/open-cluster-management/integrity-shield
        - name: Operator Documentation
          url: https://github.com/open-cluster-management/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - quay.io/open-cluster-management/integrity-shield-operator:0.3.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        version: 0.3.0
      entries:
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.0
    - currentCSV: integrity-shield-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/open-cluster-management/integrity-shield-observer",
                    "imagePullPolicy": "Always",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  resp.status_code == 200\n  result := json.unmarshal(resp.raw_body)\n  result.allow == false\n  not is_detect_mode\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"tls_insecure_skip_verify\": true\n  })\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: detect\n  admissionOnly: false\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/open-cluster-management/integrity-shield-api",
                    "imagePullPolicy": "Always",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/open-cluster-management/integrity-shield-operator:0.3.1
          createdAt: "2021-10-22T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.

          [K8s Integrity Shield](https://github.com/open-cluster-management/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.

          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Detect (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster

          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).

          ## Preparations before installation

          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).


          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.

          If you want to change the settings such as default run mode (detection/enforcement) or audit interval,  please check [here](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).

          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-7b7f768bf7-fhrpg                           1/1     Running   0          20s
          integrity-shield-observer-5bc66f75f7-tn8fw                      1/1     Running   0          25s
          integrity-shield-operator-controller-manager-65b7fb58f7-j25zd   2/2     Running   0          3h5m
          ```

          After installation, you can protect Kubernetes resources by following this [document](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).

          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:

          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)

          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/open-cluster-management/integrity-shield
        - name: Operator Documentation
          url: https://github.com/open-cluster-management/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/open-cluster-management/integrity-shield-operator:0.3.1
        version: 0.3.1
      entries:
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.1
    - currentCSV: integrity-shield-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/stolostron/integrity-shield-observer",
                    "imagePullPolicy": "IfNotPresent",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  should_be_denied(resp)\n  not is_detect_mode\n  result := get_result(resp)\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"timeout\": \"2500ms\",\n    \"raise_error\": false,\n    \"tls_ca_cert\": REPLACE_WITH_CA_CERT\n  })\n}\n\n# get decision result from API response\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  not is_null(raw_body)\n  result := json.unmarshal(raw_body)\n}\n# if error, return it as result\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  is_null(raw_body)\n  error_msg := object.get(resp, \"error\", \"\")\n  result := {\"http_error\":error_msg}\n}\n\n# if API response is null, deny\nshould_be_denied(resp) {\n  is_null(resp)\n}\n\n# if API response has status_code other than 200, deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  resp.status_code != 200\n}\n\n# if decision result is \"allow == false\" or not set, then deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  result := get_result(resp)\n  allow := object.get(result, \"allow\", false)\n  allow == false\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "reporter": {
                    "image": "quay.io/stolostron/integrity-shield-reporter",
                    "imagePullPolicy": "IfNotPresent",
                    "intervalSeconds": "10",
                    "name": "integrity-shield-reporter",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: inform\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\ndecisionReporterConfig:\n  enabled: true\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/stolostron/integrity-shield-api",
                    "imagePullPolicy": "IfNotPresent",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/stolostron/integrity-shield-operator:0.3.2
          createdAt: "2022-01-25T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.15.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.
          [K8s Integrity Shield](https://github.com/stolostron/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.
          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Inform (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster
          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).
          ## Preparations before installation
          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).
          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.
          If you want to change the settings such as default run mode (inform/enforce) or audit interval,  please check [here](https://github.com/stolostron/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).
          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-5799f69896-zbcjs                           2/2     Running   0          88s
          integrity-shield-observer-757f58f7d7-7pf5h                      1/1     Running   0          55s
          integrity-shield-operator-controller-manager-6f94d7d684-dg5bt   2/2     Running   0          4m27s
          ```
          After installation, you can protect Kubernetes resources by following this [document](https://github.com/stolostron/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).
          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:
          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)
          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/stolostron/integrity-shield
        - name: Operator Documentation
          url: https://github.com/stolostron/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/stolostron/integrity-shield-operator:0.3.2
        version: 0.3.2
      entries:
      - name: integrity-shield-operator.v0.3.2
        version: 0.3.2
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.2
    - currentCSV: integrity-shield-operator.v0.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": false,
                    "image": "quay.io/stolostron/integrity-shield-observer",
                    "imagePullPolicy": "IfNotPresent",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  should_be_denied(resp)\n  not is_detect_mode\n  result := get_result(resp)\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"timeout\": \"2500ms\",\n    \"raise_error\": false,\n    \"tls_ca_cert\": REPLACE_WITH_CA_CERT\n  })\n}\n\n# get decision result from API response\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  not is_null(raw_body)\n  result := json.unmarshal(raw_body)\n}\n# if error, return it as result\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  is_null(raw_body)\n  error_msg := object.get(resp, \"error\", \"\")\n  result := {\"http_error\":error_msg}\n}\n\n# if API response is null, deny\nshould_be_denied(resp) {\n  is_null(resp)\n}\n\n# if API response has status_code other than 200, deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  resp.status_code != 200\n}\n\n# if decision result is \"allow == false\" or not set, then deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  result := get_result(resp)\n  allow := object.get(result, \"allow\", false)\n  allow == false\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "reporter": {
                    "image": "quay.io/stolostron/integrity-shield-reporter",
                    "imagePullPolicy": "IfNotPresent",
                    "intervalSeconds": "10",
                    "name": "integrity-shield-reporter",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: inform\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\ndecisionReporterConfig:\n  enabled: true\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/stolostron/integrity-shield-api",
                    "imagePullPolicy": "IfNotPresent",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/stolostron/integrity-shield-operator:0.3.3
          createdAt: "2022-03-18T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.15.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.
          [K8s Integrity Shield](https://github.com/stolostron/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.
          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Inform (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster
          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).
          ## Preparations before installation
          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).
          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.
          If you want to change the settings such as default run mode (inform/enforce) or audit interval,  please check [here](https://github.com/stolostron/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).
          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-5799f69896-zbcjs                           2/2     Running   0          88s
          integrity-shield-observer-757f58f7d7-7pf5h                      1/1     Running   0          55s
          integrity-shield-operator-controller-manager-6f94d7d684-dg5bt   2/2     Running   0          4m27s
          ```
          After installation, you can protect Kubernetes resources by following this [document](https://github.com/stolostron/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).
          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:
          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)
          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/stolostron/integrity-shield
        - name: Operator Documentation
          url: https://github.com/stolostron/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/stolostron/integrity-shield-operator:0.3.3
        version: 0.3.3
      entries:
      - name: integrity-shield-operator.v0.3.3
        version: 0.3.3
      - name: integrity-shield-operator.v0.3.2
        version: 0.3.2
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.3
    defaultChannel: alpha-0.3.3
    packageName: integrity-shield-operator
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datasance
      provider-url: ""
    name: datasance-pot-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datasance-pot-operator.v3.4.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasance.com/v3",
                "kind": "ControlPlane",
                "metadata": {
                  "name": "pot"
                },
                "spec": {
                  "replicas": {
                    "controller": 2
                  },
                  "database": {
                    "provider": "postgres",
                    "user": "user",
                    "host": "postgre.host",
                    "port": 5432,
                    "password": "password",
                    "databaseName": "dbname"
                  },
                  "auth": {
                    "url": "auth.url",
                    "realm": "realm",
                    "realmKey": "realmKey",
                    "ssl": "none",
                    "controllerClient": "controllerClient",
                    "controllerSecret": "controllerClientSecret",
                    "viewerClient": "viewerClient"
                  },
                  "images": {
                    "pullSecret": "pullsecret",
                    "controller": "ghcr.io/datasance/controller@sha256:aa896bae6c3e6eda70938a03aaa649226356d9e4a1d03ba427d4e4ee3090ca6c",
                    "portManager": "ghcr.io/datasance/port-manager@sha256:4bf3793e9e3cd10df9b46c459e983631393825ca788ffbbb3ec3659158ad61fe",
                    "proxy": "ghcr.io/datasance/proxy@sha256:276dbed0c8ef4f5b03fdf29d5c863b6f3efcae50444ce8123220c5673272860f",
                    "router": "ghcr.io/datasance/router@sha256:82657fc8f39d33d8ef0161cf27d705cbdc0a97dca07b22a5fe78fd4213ef0d31"
                  },
                  "services": {
                    "controller": {
                      "type": "LoadBalancer",
                      "annotations": {
                        "service.beta.kubernetes.io/azure-load-balancer-internal": "false"
                      }
                    },
                    "proxy": {
                      "type": "LoadBalancer",
                      "annotations": {
                        "service.beta.kubernetes.io/azure-load-balancer-internal": "false"
                      }
                    },
                    "router": {
                      "type": "LoadBalancer",
                      "annotations": {
                        "service.beta.kubernetes.io/azure-load-balancer-internal": "false"
                      }
                    }
                  },
                  "controller": {
                    "ecnViewerUrl": "pot.demo.datasance.com",
                    "https": true,
                    "secretName": "secretName"
                  },
                  "ingresses": {
                    "controller": {
                      "annotations": {
                        "cert-manager.io/cluster-issuer": "letsencrypt"
                      },
                      "ingressClassName": "nginx",
                      "host": "pot.demo.datasance.com",
                      "secretName": "secretName"
                    },
                    "router": {
                      "address": "pot.demo.datasance.com",
                      "messagePort": 5672,
                      "interiorPort": 55672,
                      "edgePort": 45672
                    },
                    "httpProxy": {
                      "address": "pot.demo.datasance.com"
                    },
                    "tcpProxy": {
                      "address": "pot.demo.datasance.com"
                    }
                  }
                },
                "status": {
                  "conditions": [
                    {
                      "lastTransitionTime": "2022-04-10T22:44:09Z",
                      "message": "",
                      "reason": "initial_status",
                      "status": "True",
                      "type": "deploying"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          containerImage: ghcr.io/datasance/operator@sha256:143b9307c2520ef16b40c4c92e6faee8b24b7fb3a97682231e53236ddd532687
          createdAt: "2024-11-04T20:45:43Z"
          description: Datasance PoT is an Enteprise version of Eclipse ioFog Platform.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Datasance PoT Enterprise Open
            Source Subscriptions based on number of Edge device that ControlPlane
            manages"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/datasance/iofog-operator
          support: Datasance Teknoloji A.S.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Datasance PoT supported ioFog ControlPlane CRD for managing
              container workloads on Edge device
            kind: ControlPlane
            name: controlplanes.datasance.com
            version: v3
        description: ioFog Control Plane on Kubernetes
        displayName: iofog-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iofog
        - datasance
        - pot
        links:
        - name: Datasance PoT // Iofog Operator
          url: https://docs.datasance.com
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Datasance
        relatedImages:
        - ghcr.io/datasance/operator@sha256:143b9307c2520ef16b40c4c92e6faee8b24b7fb3a97682231e53236ddd532687
        - ghcr.io/datasance/controller@sha256:aa896bae6c3e6eda70938a03aaa649226356d9e4a1d03ba427d4e4ee3090ca6c
        - ghcr.io/datasance/router@sha256:82657fc8f39d33d8ef0161cf27d705cbdc0a97dca07b22a5fe78fd4213ef0d31
        - ghcr.io/datasance/proxy@sha256:276dbed0c8ef4f5b03fdf29d5c863b6f3efcae50444ce8123220c5673272860f
        - ghcr.io/datasance/port-manager@sha256:4bf3793e9e3cd10df9b46c459e983631393825ca788ffbbb3ec3659158ad61fe
        version: 3.4.9
      entries:
      - name: datasance-pot-operator.v3.4.9
        version: 3.4.9
      name: stable
    defaultChannel: stable
    packageName: datasance-pot-operator
    provider:
      name: Datasance
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: gcp-filestore-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gcp-filestore-csi-driver-operator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel8-operator@sha256:1ec61d2988b95a10260c7932e9dc6a9d1d5b3eb4148b22afd438aa4ee75dc9b3
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure GCP Filestore CSI driver.
          olm.skipRange: '>=4.11.0-0 <4.13.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/gcp-filestore-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Operator that installs and configures the CSI driver for GCP Filestore. The GCP Filestore CSI Driver Operator is a Technology Preview feature. For more information on Red Hat Technology Preview features support scope, see https://access.redhat.com/support/offerings/techpreview/.
        displayName: GCP Filestore CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - filestore
        links:
        - name: Documentation
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: preview
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:700c0e50f86a70f506403aba75e020a17530ada284af7eee77d4b4ce3dbf4102
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:4f1e873080754ae9aebc361cd77f965e1ade369403346aa0cf4cf502082d18e4
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:ee45e11a6fe8cf60f3ac285cba1b23459ca95068bd9769042ae052c778c3e080
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:537fbc8fe2b212dc798361099d2c6fd981cc846c1cf5454b703ba08d488b61f8
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel8@sha256:a1a4e7b93d6609561a7a2cd732ce772bf0047448e5019298d66c4b985baecfb2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:767682dd3bd65651a8204c9fb732b9c48b99127189992b43abc6ecce027f4589
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel8-operator@sha256:1ec61d2988b95a10260c7932e9dc6a9d1d5b3eb4148b22afd438aa4ee75dc9b3
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:6f1b497772b953468f0b816face8810ace8ea44fa91dd4a3b3e583da2cac350c
        version: 4.13.0-202309181427
      entries:
      - name: gcp-filestore-csi-driver-operator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: preview
    - currentCSV: gcp-filestore-csi-driver-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9-operator@sha256:2dc13de58b391237dc48b9335d88b99f5e26d0e5edc5a664c7d39f16ca9b297d
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure GCP Filestore CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.11.0-0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/gcp-filestore-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Operator that installs and configures the CSI driver for GCP Filestore.
        displayName: GCP Filestore CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - filestore
        links:
        - name: Documentation
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:a5704d26d0b562dba9633d1c107f34e05c750c0e3a64dbcace4dab7b964fff8d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9@sha256:8e801d924071c8f2513555450e688bff60d4100b2ca56d2ba9517038a1d48d4d
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:9c3c3d2b857ce289324aee56df1ba4ab5e5ea1e76cd86ac988233463915834b4
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:78e02d937740d742eb33b08c7317f211c25020c375069d5a97344107716608d3
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:b28def976fc155dff57ff3003cc462a8e9923b325af0b319b65220c66688699e
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0627c09fe3f5a015f117fb30af1f4d17340aefa48b23994737d3901a804b63d3
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9-operator@sha256:2dc13de58b391237dc48b9335d88b99f5e26d0e5edc5a664c7d39f16ca9b297d
        version: 4.16.0-202509240123
      entries:
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202509091828
        version: 4.16.0-202509091828
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202508270905
        version: 4.16.0-202508270905
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202504121405
        version: 4.16.0-202504121405
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: gcp-filestore-csi-driver-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: gcp-filestore-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-prometheusservice-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-prometheusservice-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "prometheusservice.services.k8s.aws/v1alpha1",
                "kind": "RuleGroupsNamespace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "prometheusservice.services.k8s.aws/v1alpha1",
                "kind": "Workspace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/prometheusservice-controller:1.3.1
          createdAt: "2025-09-19T21:11:38Z"
          description: AWS Prometheus controller is a service controller for managing
            Prometheus resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertManagerDefinition represents the state of an AWS prometheusservice
              AlertManagerDefinition resource.
            displayName: AlertManagerDefinition
            kind: AlertManagerDefinition
            name: alertmanagerdefinitions.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: LoggingConfiguration represents the state of an AWS prometheusservice
              LoggingConfiguration resource.
            displayName: LoggingConfiguration
            kind: LoggingConfiguration
            name: loggingconfigurations.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: RuleGroupsNamespace represents the state of an AWS prometheusservice
              RuleGroupsNamespace resource.
            displayName: RuleGroupsNamespace
            kind: RuleGroupsNamespace
            name: rulegroupsnamespaces.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: Workspace represents the state of an AWS prometheusservice
              Workspace resource.
            displayName: Workspace
            kind: Workspace
            name: workspaces.prometheusservice.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Prometheus resources in AWS from within your Kubernetes cluster.

          **About Amazon Managed Service for Prometheus**

          Amazon Managed Service for Prometheus is a serverless, Prometheus-compatible monitoring service for container metrics that makes it easier to securely monitor container environments at scale. With Amazon Managed Service for Prometheus, you can use the same open-source Prometheus data model and query language that you use today to monitor the performance of your containerized workloads, and also enjoy improved scalability, availability, and security without having to manage the underlying infrastructure.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Prometheus
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - prometheusservice
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Prometheus Developer Resources
          url: https://aws.amazon.com/prometheus/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: prometheus maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/prometheusservice-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-prometheusservice-controller.v1.3.1
        version: 1.3.1
      - name: ack-prometheusservice-controller.v1.3.0
        version: 1.3.0
      - name: ack-prometheusservice-controller.v1.2.27
        version: 1.2.27
      - name: ack-prometheusservice-controller.v1.2.26
        version: 1.2.26
      - name: ack-prometheusservice-controller.v1.2.24
        version: 1.2.24
      - name: ack-prometheusservice-controller.v1.2.23
        version: 1.2.23
      - name: ack-prometheusservice-controller.v1.2.22
        version: 1.2.22
      - name: ack-prometheusservice-controller.v1.2.20
        version: 1.2.20
      - name: ack-prometheusservice-controller.v1.2.19
        version: 1.2.19
      - name: ack-prometheusservice-controller.v1.2.17
        version: 1.2.17
      - name: ack-prometheusservice-controller.v1.2.16
        version: 1.2.16
      - name: ack-prometheusservice-controller.v1.2.15
        version: 1.2.15
      - name: ack-prometheusservice-controller.v1.2.14
        version: 1.2.14
      - name: ack-prometheusservice-controller.v1.2.13
        version: 1.2.13
      - name: ack-prometheusservice-controller.v1.2.12
        version: 1.2.12
      - name: ack-prometheusservice-controller.v1.2.11
        version: 1.2.11
      - name: ack-prometheusservice-controller.v1.2.10
        version: 1.2.10
      - name: ack-prometheusservice-controller.v1.2.9
        version: 1.2.9
      - name: ack-prometheusservice-controller.v1.2.6
        version: 1.2.6
      - name: ack-prometheusservice-controller.v1.2.5
        version: 1.2.5
      - name: ack-prometheusservice-controller.v1.2.4
        version: 1.2.4
      - name: ack-prometheusservice-controller.v1.2.3
        version: 1.2.3
      - name: ack-prometheusservice-controller.v1.2.2
        version: 1.2.2
      - name: ack-prometheusservice-controller.v1.2.1
        version: 1.2.1
      - name: ack-prometheusservice-controller.v1.2.0
        version: 1.2.0
      - name: ack-prometheusservice-controller.v1.1.0
        version: 1.1.0
      - name: ack-prometheusservice-controller.v1.0.0
        version: 1.0.0
      - name: ack-prometheusservice-controller.v0.1.3
        version: 0.1.3
      - name: ack-prometheusservice-controller.v0.1.2
        version: 0.1.2
      - name: ack-prometheusservice-controller.v0.1.1
        version: 0.1.1
      - name: ack-prometheusservice-controller.v0.1.0
        version: 0.1.0
      - name: ack-prometheusservice-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-prometheusservice-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: awss3-operator-registry
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: awss3operator.1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/screeley44/aws-s3-provisioner:v1.0.0
          createdAt: "2019-05-02 16:12:00"
          description: Manage the full lifecycle of installing, configuring and managing
            AWS S3 Provisioner.
          repository: https://github.com/yard-turkey/aws-s3-provisioner
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - description: instance of an AWS S3 Bucket
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
          - description: Request for an AWS S3 Bucket
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
        description: |
          AWS S3 Operator will deploy the AWS S3 Provisioner which will dynamically or statically
          provision AWS S3 Bucket storage and access. The operator deploys the ObjectBucket (OB) and ObjectBucketClaim (OBC)
          CustomResourceDefinitions. The OB/OBC model follows the traditional Kubernetes PV/PVC pattern, when an OBC is detected
          the operator will act on the OBC to either provision a brand new S3 Bucket in AWS or gain access to an existing
          S3 Bucket in AWS. The operator produces an OB and ConfigMap and Secret which can then be consumed by application pods.

          **Important Note**: Currently, while in preview, this operator does not
          support automatic upgrades. You must remove the old version of the operator
          manually before installing a new version.
          ## Using AWS S3 Operator

          ### Administrator Creates Secret
          This secret will contain the elevated/admin privileges needed by the provisioner
          to properly access and create S3 Buckets and IAM users and policies. The AWS Access ID
          and AWS Secret Key will be needed for this.
          1. Create the Kubernetes Secret for the Provisioner's Owner Access.
          ```yaml
          apiVersion: v1
          kind: Secret
          metadata:
            name: s3-bucket-owner [1]
            namespace: s3-provisioner [2]
          type: Opaque
          data:
            AWS_ACCESS_KEY_ID: *base64 encoded value* [3]
            AWS_SECRET_ACCESS_KEY: *base64 encoded value* [4]
          ```
          1. Name of the secret, this will be referenced in StorageClass.
          1. Namespace where the Secret will exist.
          1. Your AWS_ACCESS_KEY_ID base64 encoded.
          1. Your AWS_SECRET_ACCESS_KEY base64 encoded.
          ```
           # kubectl create -f creds.yaml
          secret/s3-bucket-owner created
          ```
          ### Administrator Creates StorageClass
          The StorageClass defines the name of the provisioner and holds other properties that are needed to provision a new bucket, including
          the Owner Secret and Namespace, and the AWS Region.
          #### Greenfield Example:
          For Greenfield, a new, dynamic bucket will be generated.
          1. Create the Kubernetes StorageClass for the Provisioner.
          ```yaml
          kind: StorageClass
          apiVersion: storage.k8s.io/v1
          metadata:
            name: s3-buckets [1]
          provisioner: aws-s3.io/bucket [2]
          parameters:
            region: us-west-1 [3]
            secretName: s3-bucket-owner [4]
            secretNamespace: s3-provisioner [5]
          reclaimPolicy: Delete [6]
          ```
          1. Name of the StorageClass, this will be referenced in the User ObjectBucketClaim.
          1. Provisioner name
          1. AWS Region that the StorageClass will serve
          1. Name of the bucket owner Secret created above
          1. Namespace where the Secret will exist
          1. reclaimPolicy (Delete or Retain) indicates if the bucket can be deleted when the OBC is deleted.
          **NOTE:** the absence of the `bucketName` Parameter key in the storage class indicates this is a new bucket and its name is based on the bucket name fields in the OBC.
          ```
           # kubectl create -f storageclass-greenfield.yaml
          storageclass.storage.k8s.io/s3-buckets created
          ```
          #### Brownfield Example:
          For brownfield, the StorageClass defines the name of the provisioner and the name of the existing bucket. It also includes other properties needed by the target
          provisioner, including: the Owner Secret and Namespace, and the AWS Region
          1. Create the Kubernetes StorageClass for the Provisioner.
          ```yaml
          kind: StorageClass
          apiVersion: storage.k8s.io/v1
          metadata:
            name: s3-existing-buckets [1]
          provisioner: aws-s3.io/bucket [2]
          parameters:
            bucketName: my-existing-bucket [3]
            region: us-west-1 [4]
            secretName: s3-bucket-owner [5]
            secretNamespace: s3-provisioner [6]
          ```
          1. Name of the StorageClass, this will be referenced in the User ObjectBucketClaim.
          1. Provisioner name
          1. Name of the existing bucket
          1. AWS Region that the StorageClass will serve
          1. Name of the bucket owner Secret created above
          1. Namespace for that bucket owner secret
          **NOTE:** the storage class's `reclaimPolicy` is ignored for existing buckets.
          ```
           # kubectl create -f storageclass-brownfield.yaml
          storageclass.storage.k8s.io/s3-buckets created
          ```
          ### User Creates ObjectBucketClaim
          An ObjectBucketClaim follows the same concept as a PVC, in that
          it is a request for Object Storage, the user doesn't need to
          concern him/herself with the underlying storage, just that
          they need access to it. The user will work with the cluster/storage
          administrator to get the proper StorageClass needed and will
          then request access via the OBC.
          #### Greenfield Request Example:
          1. Create the ObjectBucketClaim.
          ```yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucketClaim
          metadata:
            name: myobc [1]
            namespace: s3-provisioner [2]
          spec:
            generateBucketName: mybucket [3]
            bucketName: my-awesome-bucket [4]
            storageClassName: s3-buckets [5]
          ```
          1. Name of the OBC
          1. Namespace of the OBC
          1. Name prepended to a random string used to generate a bucket name. It is ignored if bucketName is defined
          1. Name of new bucket which must be unique across all AWS regions, otherwise an error occurs when creating the bucket. If present, this name overrides `generateName`
          1. StorageClass name
          **NOTE:** if both `generateBucketName` and `bucketName` are omitted, and the storage class does _not_ define a bucket name, then a new, random bucket name is generated with no prefix.
          ```
           # kubectl create -f obc-brownfield.yaml
          objectbucketclaim.objectbucket.io/myobc created
          ```
          #### Brownfield Request Example:
          1. Create the ObjectBucketClaim.
          ```yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucketClaim
          metadata:
            name: myobc [1]
            namespace: s3-provisioner [2]
          spec:
            storageClassName: s3-existing-buckets [3]
          ```
          1. Name of the OBC
          1. Namespace of the OBC
          1. StorageClass name
          **NOTE:** in the OBC here there is no reference to the bucket's name. This is defined in the storage class and is not a concern of the user creating the claim to this bucket.  An OBC does have fields for defining a bucket name for greenfield use only.
          ```
           # kubectl create -f obc-brownfield.yaml
          objectbucketclaim.objectbucket.io/myobc created
          ```
          ### Results and Recap
          Let's pause for a moment and digest what just happened.
          After creating the OBC, and assuming the S3 provisioner is running, we now have
          the following Kubernetes resources:
          .  a global ObjectBucket (OB) which contains: bucket endpoint info (including region and bucket name), a reference to the OBC, and a reference to the storage class. Unique to S3, the OB also contains the bucket Amazon Resource Name (ARN).Note: there is always a 1:1 relationship between an OBC and an OB.
          .  a ConfigMap in the same namespace as the OBC, which contains the same endpoint data found in the OB.
          .  a Secret in the same namespace as the OBC, which contains the AWS key-pairs needed to access the bucket.
          And of course, we have a *new* AWS S3 Bucket which you should be able to see via the AWS Console.
          *ObjectBucket*
          ```yaml
           # kubectl get ob obc-s3-provisioner-my-awesome-bucket -o yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucket
          metadata:
            creationTimestamp: "2019-04-03T15:42:22Z"
            generation: 1
            name: obc-s3-provisioner-my-awesome-bucket
            resourceVersion: "15057"
            selfLink: /apis/objectbucket.io/v1alpha1/objectbuckets/obc-s3-provisioner-my-awesome-bucket
            uid: 0bfe8e84-576d-4c4e-984b-f73c4460f736
          spec:
            Connection:
              additionalState:
                ARN: arn:aws:iam::<accountid>:policy/my-awesome-bucket-vSgD5 [1]
                UserName: my-awesome-bucket-vSgD5 [2]
              endpoint:
                additionalConfig: null
                bucketHost: s3-us-west-1.amazonaws.com
                bucketName: my-awesome-bucket [3]
                bucketPort: 443
                region: us-west-1
                ssl: true
                subRegion: ""
            claimRef: null [4]
            reclaimPolicy: null
            storageClassName: s3-buckets [5]
          ```
          1. The AWS Policy created for this user and bucket.
          1. The new user generated by the Provisioner to access this existing bucket.
          1. The bucket name.
          1. The reference to the OBC (not filled in yet).
          1. The reference to the StorageClass used.
          *ConfigMap*
          ```yaml
           # kubectl get cm myobc -n s3-provisioner -o yaml
          apiVersion: v1
          data:
            BUCKET_HOST: s3-us-west-1.amazonaws.com [1]
            BUCKET_NAME: my-awesome-bucket [2]
            BUCKET_PORT: "443"
            BUCKET_REGION: us-west-1
            BUCKET_SSL: "true"
            BUCKET_SUBREGION: ""
          kind: ConfigMap
          metadata:
            creationTimestamp: "2019-04-01T19:11:38Z"
            finalizers:
            - objectbucket.io/finalizer
            name: my-awesome-bucket
            namespace: s3-provisioner
            resourceVersion: "892"
            selfLink: /api/v1/namespaces/s3-provisioner/configmaps/my-awesome-bucket
            uid: 2edcc58a-aff8-4a29-814a-ffbb6439a9cd
          ```
          1. The AWS S3 host.
          1. The name of the new bucket we are gaining access to.
          *Secret*
          ```yaml
           # kubectl get secret my-awesome-bucket -n s3-provisioner -o yaml
          apiVersion: v1
          data:
            AWS_ACCESS_KEY_ID: *the_new_access_id* [1]
            AWS_SECRET_ACCESS_KEY: *the_new_access_key_value* [2]
          kind: Secret
          metadata:
            creationTimestamp: "2019-04-03T15:42:22Z"
            finalizers:
            - objectbucket.io/finalizer
            name: my-awesome-bucket
            namespace: s3-provisioner
            resourceVersion: "15058"
            selfLink: /api/v1/namespaces/s3-provisioner/secrets/screeley-provb-5
            uid: 225c71a5-9d75-4ccc-b41f-bfe91b272a13
          type: Opaque
          ```
          1. The new generated AWS Access Key ID.
          1. The new generated AWS Secret Access Key.
          What happened in AWS? The first thing we do on any OBC request is
          create a new IAM user and generate Access ID and Secret Keys.
          This allows us to also better control ACLs. We also create a policy
          in IAM which we then attach to the user and bucket. We also created a new bucket, called *my-awesome-bucket*.
          When the OBC is deleted all of its Kubernetes and AWS resources will also be deleted, which includes:
          the generated OB, Secret, ConfigMap, IAM user, and policy.
          If the _retainPolicy_ on the StorageClass for this bucket is *"Delete"*, then, in addition to the above cleanup, the physical bucket is also deleted.
          **NOTE:** The actual bucket is only deleted if the Storage Class's _reclaimPolicy_ is "Delete".
          ### User Creates Pod
          Now that we have our bucket and connection/access information, a pod
          can be used to access the bucket. This can be done in several different
          ways, but the key here is that the provisioner has provided the proper
          endpoints and keys to access the bucket. The user then simply references
          the keys.
          1. Create a Sample Pod to Access the Bucket.
          ```yaml
          apiVersion: v1
          kind: Pod
          metadata:
            name: photo1
            labels:
              name: photo1
          spec:
            containers:
            - name: photo1
              image: docker.io/screeley44/photo-gallery:latest
              imagePullPolicy: Always
              envFrom:
              - configMapRef:
                  name: my-awesome-bucket <1>
              - secretRef:
                  name: my-awesome-bucket <2>
              ports:
              - containerPort: 3000
                protocol: TCP
          ```
          1. Name of the generated configmap from the provisioning process
          1. Name of the generated secret from the provisioning process
          *[Note]* Generated ConfigMap and Secret are same name as the OBC!
          Lastly, expose the pod as a service so you can access the url from a browser. In this example,
          I exposed as a LoadBalancer
          ```
            # kubectl expose pod photo1 --type=LoadBalancer --name=photo1 -n your-namespace
          ```
          To access via a url use the EXTERNAL-IP
          ```
            # kubectl get svc photo1
            NAME                         TYPE           CLUSTER-IP       EXTERNAL-IP                                                               PORT(S)          AGE
            photo1                       LoadBalancer   100.66.124.105   a00c53ccb3c5411e9b6550a7c0e50a2a-2010797808.us-east-1.elb.amazonaws.com   3000:32344/TCP   6d
          ```
          **NOTE:** This is just one example of a Pod that can utilize the bucket information,
          there are several ways that these pod applications can be developed and therefore
          the method of getting the actual values needed from the Secrets and ConfigMaps
          will vary greatly, but the idea remains the same, that the pod consumes the generated
          ConfigMap and Secret created by the provisioner.
        displayName: AWS S3 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - awss3provisioner
        - aws
        - S3
        - provisioner
        - buckets
        links:
        - name: AWS S3 Provisioner
          url: https://github.com/yard-turkey/aws-s3-provisioner
        - name: Provisioner Examples
          url: https://github.com/yard-turkey/aws-s3-provisioner/tree/master/examples
        - name: Operator Documentation
          url: https://github.com/yard-turkey/awss3operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - quay.io/screeley44/aws-s3-provisioner:v1.0.0
        version: 1.0.1
      entries:
      - name: awss3operator.1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: awss3-operator-registry
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Managed OpenShift Black Belt
      provider-url: https://github.com/rh-mobb/ocm-operator
    name: ocm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocm-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "GitLabIdentityProvider",
                "metadata": {
                  "name": "gitlab-sample"
                },
                "spec": {
                  "clientID": "gitlab",
                  "clientSecret": {
                    "name": "gitlab"
                  },
                  "clusterName": "my-cluster",
                  "displayName": "gitlab-sample",
                  "mappingMethod": "claim",
                  "url": "https://gitlab.com"
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "LDAPIdentityProvider",
                "metadata": {
                  "name": "ldap-sample"
                },
                "spec": {
                  "bindDN": "CN=test,OU=Users,DC=example,DC=com",
                  "bindPassword": {
                    "name": "test"
                  },
                  "clusterName": "my-cluster",
                  "displayName": "ldap-sample",
                  "mappingMethod": "claim",
                  "url": "ldap://test.example.com:389"
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "MachinePool",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "aws": {
                    "spotInstances": {
                      "enabled": false
                    }
                  },
                  "clusterName": "my-cluster",
                  "instanceType": "m5.xlarge",
                  "labels": {
                    "owner": "my-user"
                  },
                  "maximumNodesPerZone": 1,
                  "minimumNodesPerZone": 1,
                  "wait": false
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "ROSACluster",
                "metadata": {
                  "name": "rosa-sample"
                },
                "spec": {
                  "accountID": "111111111111",
                  "defaultMachinePool": {
                    "instanceType": "m5.xlarge",
                    "labels": {
                      "owner": "my-owner"
                    },
                    "minimumNodesPerZone": 2
                  },
                  "displayName": "my-cluster",
                  "iam": {
                    "userRole": "arn:aws:iam::111111111111:role/ManagedOpenShift-User-my_user-Role"
                  },
                  "tags": {
                    "owner": "my-owner"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ocm-operator:v0.1.0
          createdAt: "2023-06-02T17:32:25Z"
          operatorframework.io/suggested-namespace: ocm-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ocm-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLabIdentityProvider is the Schema for the gitlabidentityproviders
              API.
            displayName: Git Lab Identity Provider
            kind: GitLabIdentityProvider
            name: gitlabidentityproviders.ocm.mobb.redhat.com
            version: v1alpha1
          - description: LDAPIdentityProvider is the Schema for the ldapidentityproviders
              API.
            displayName: LDAPIdentity Provider
            kind: LDAPIdentityProvider
            name: ldapidentityproviders.ocm.mobb.redhat.com
            version: v1alpha1
          - description: MachinePool is the Schema for the machinepools API.
            displayName: Machine Pool
            kind: MachinePool
            name: machinepools.ocm.mobb.redhat.com
            version: v1alpha1
          - description: ROSACluster is the Schema for the clusters API.
            displayName: ROSACluster
            kind: ROSACluster
            name: rosaclusters.ocm.mobb.redhat.com
            version: v1alpha1
        description: "# Quick Start\n\n## Prerequisites\n\n- [oc](https://docs.openshift.com/container-platform/4.8/cli_reference/openshift_cli/getting-started-cli.html)\n-
          [aws](https://docs.aws.amazon.com/cli/latest/userguide/getting-started-install.html)\n-
          [ROSA Cluster (4.12+)](https://mobb.ninja/docs/rosa/sts/)\n\n> **Warning:**\n>
          This is only tested on ROSA, but may work on other Kubernetes and OpenShift
          variants.  It requires\n> an AWS STS-enabled cluster.  Additionally, OpenShift
          versions 4.12+ are recommended, as they \n> enable [CRD CEL validation](https://kubernetes.io/blog/2022/09/23/crd-validation-rules-beta/).
          \ Versions\n> prior to 4.12 provide a lower version of Kubernetes that does
          not enable this feature.  They may \n> work, but provide no input validation
          when submitting custom resources to the cluster.\n\nBefore installing this
          operator, there are a couple secrets that must be created.\n\n1. OCM Token:
          used to authenticate against the OCM API.  The controller will not start
          without\nthis token as it used for all calls into OCM.\n2. AWS Credentials:
          uses an assume role to a role created in the instructions below in order\nto
          access AWS resources.  This is needed for cluster creation.  See \n[Create
          AWS IAM Policies and Roles](#create-aws-iam-policies-and-roles) for more
          details.\n\n> **Note:**\n> Certain custom resources have their own specific
          prerequisites.  Please use `oc explain` \n> or read the docs [here](https://github.com/rh-mobb/ocm-operator/tree/main/docs)
          for more details.\n\n### Create OCM Token Secret\n\n1. Create a namespace
          where you wish to install the operator:\n\n```bash\noc new-project ocm-operator\n```\n\n2.
          Create a secret containing the OCM_TOKEN.  This token can be obtained form
          \nhttps://console.redhat.com/openshift/token and is used by the operator
          to authenticate \nagainst the OCM API.  This token must exist in the same
          namespace that the operator \nis running and be named `ocm-token`.  It also
          expects the key to be called `OCM_TOKEN` \nas the operator is expecting
          this value as an environment variable.\n\n```bash\noc create secret generic
          ocm-token \\\n  --namespace=ocm-operator \\\n  --from-literal=OCM_TOKEN=${MY_OCM_TOKEN}\n```\n\n###
          Create AWS IAM Policies and Roles\n\nThe operator will need to elevate privileges
          in order to perform things like \ncreating the operator-roles for the clusters.
          \ Because of this, the operator \nmust have a specific role created to allow
          it these permissions.\n\n**NOTE:** please understand what you are doing
          if you deviate from the known good \npolicies.  If errors or more stringent
          security lockdowns are found, please submit a PR \nso that we can get this
          fixed.\n\n1. Set `ACCOUNT_ID` environment variable to define your AWS account
          ID:\n\n```bash\nexport ACCOUNT_ID=111111111111\n```\n\n2. Create the permissions
          boundary.  Because the operator needs to create policies and \nroles, the
          boundary ensures that the operator is not allowed to create additional\npermissions
          outside of the defined boundary.  The sample permission set is located \nat
          `test/aws/boundary.json` in this repository:\n\n```bash\nboundary=$(curl
          -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/boundary.json\")\naws
          iam create-policy \\\n  --policy-name \"OCMOperatorBoundary\" \\\n  --policy-document
          \"$boundary\"\n```\n\n3. Create the policy.  This policy sets what the operator
          is allowed to do.  For any \n`iam` permission, the boundary created above
          is used.  The sample permission set is \nlocated at `test/aws/iam_policy.json`
          in this repository:\n\n```bash\npolicy=$(curl -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/iam_policy.json\")\naws
          iam create-policy \\\n  --policy-name \"OCMOperator\" \\\n  --policy-document
          \"$policy\"\n```\n\n4. Create the role using a trust policy and attach the
          previously created role.  The trust \npolicy is located at `test/aws/trust_policy.json`
          in this repository.  Please note that \nthe trust policy requires an OIDC
          configuration.  The OIDC configuration refers to \n**where the operator
          is running, NOT what it is provisioning**:\n\n```bash\ntrust_policy=$(curl
          -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/trust_policy.json\")\naws
          iam create-role \\\n    --role-name OCMOperator \\\n    --assume-role-policy-document
          \"$trust_policy\"\naws iam attach-role-policy \\\n    --role-name OCMOperator
          \\\n    --policy-arn arn:aws:iam::$ACCOUNT_ID:policy/OCMOperator\n```\n\n5.
          Finally, create the secret containing the assume role credentials.  The
          previous steps allow \nthe operator to assume the role you created in the
          previous step with the permissions created \nvia the previous policies.\n\n```bash\ncat
          <<EOF > /tmp/credentials\n[default]\nrole_arn = arn:aws:iam::$ACCOUNT_ID:role/OCMOperator\nweb_identity_token_file
          = /var/run/secrets/openshift/serviceaccount/token\nEOF\n\noc create secret
          generic aws-credentials \\\n  --namespace=ocm-operator \\\n  --from-file=credentials=/tmp/credentials\n```\n\n##
          Install the Operator\n\n```bash\ncat <<EOF | oc apply -f -\napiVersion:
          operators.coreos.com/v1alpha1\nkind: Subscription\nmetadata:\n  name: ocm-operator\n
          \ namespace: ocm-operator\nspec:\n  channel: v0.1\n  installPlanApproval:
          Automatic\n  name: ocm-operator\n  source: community-operators\n  sourceNamespace:
          ocm-operator\n  startingCSV: ocm-operator.v0.1.0\nEOF\n```\n\n## Provision
          Resources\n\nOnce the operator is available and running, you can provision
          any of the \nresources that it manages.  Also note that documentation is
          always available \nonce the operator is installed as well by using the `oc
          explain` command.  For \nexample, `oc explain rosacluster.spec.clusterName`
          will give you detailed documentation \nabout what the field does.\n\nSee
          the following documentation for details:\n\n* [ROSA Clusters](https://github.com/rh-mobb/ocm-operator/blob/main/docs/clusters.md)\n*
          [Machine Pools](https://github.com/rh-mobb/ocm-operator/blob/main/docs/machinepools.md)\n*
          [Identity Providers](https://github.com/rh-mobb/ocm-operator/blob/main/docs/identityproviders.md)\n"
        displayName: OpenShift Cluster Manager Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ocm
        - rosa
        - redhat
        - machinepool
        - mobb
        links:
        - name: OCM Machine Pool Operator
          url: https://github.com/rh-mobb/ocm-operator
        maintainers:
        - email: dscott@redhat.com
          name: Dustin Scott
        maturity: alpha
        provider:
          name: Managed OpenShift Black Belt
          url: https://github.com/rh-mobb/ocm-operator
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/mobb/ocm-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: ocm-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ocm-operator
    provider:
      name: Managed OpenShift Black Belt
      url: https://github.com/rh-mobb/ocm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SAP LinuxLab - Red Hat
      provider-url: https://github.com/sap-linuxlab
    name: sap-hana-express-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sap-hana-express-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "db.sap-redhat.io/v1alpha1",
                "kind": "HanaExpress",
                "metadata": {
                  "name": "hanaexpress-sample"
                },
                "spec": {
                  "credential": {
                    "key": "hxepasswd.json",
                    "name": "hxepasswd"
                  },
                  "pvcSize": "1Gi"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/redhat-sap-cop/sap-hanaexpress-operator:0.1.4
          createdAt: "2023-07-16T20:53:02Z"
          description: Manage the lifecycle of SAP Hana Express
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HanaExpress is the Schema for the hanaexpresses API
            displayName: Hana Express
            kind: HanaExpress
            name: hanaexpresses.db.sap-redhat.io
            version: v1alpha1
        description: Operator for managing the lifecycle of SAP Hana Express on Kubernetes
          platform
        displayName: SAP Hana Express Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Database
        - SAP
        links:
        - name: Sap Hana Express Operator
          url: https://github.com/redhat-sap/sap-hana-express-operator
        maintainers:
        - email: mjiao@redhat.com
          name: RedHat@SAPLinuxLab
        maturity: alpha
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/redhat-sap-cop/sap-hanaexpress-operator:0.1.4
        version: 1.0.2
      entries:
      - name: sap-hana-express-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: sap-hana-express-operator
    provider:
      name: SAP LinuxLab - Red Hat
      url: https://github.com/sap-linuxlab
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-registry-3
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-3.v3.0.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "simple"
                },
                "spec": {
                  "app": {
                    "ingress": {
                      "host": "simple-app.apps.cluster.example"
                    }
                  },
                  "ui": {
                    "ingress": {
                      "host": "simple-ui.apps.cluster.example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-3-operator:3.0.15
          createdAt: "2025-09-22T13:04:40Z"
          description: Deploy and manage Apicurio Registry 3 on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Apicurio Registry
              version 3.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Apicurio Registry 3

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Allows you manage event schemas and API designs using the Apicurio Registry web console, REST API, Maven plug-in, or Java client
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, MySQL, SQL Server, in-memory
          - Compatible with Confluent APIs
          - Runs on a lightweight Quarkus platform

          ## Apicurio Registry 3 Operator

          Provides a quick and easy way to deploy and manage Apicurio Registry 3 on Kubernetes.

          **Features**
          - Supports installation and configuration of Apicurio Registry 3
          - Perform a rolling upgrade of Apicurio Registry 3

          ## Prerequisites

          This operator does not deploy storage for Apicurio Registry 3. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio Registry 3 and Apicurio Registry 3 Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Apicurio Registry 3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-3-operator:3.0.15
        version: 3.0.15
      entries:
      - name: apicurio-registry-3.v3.0.15
        version: 3.0.15
      - name: apicurio-registry-3.v3.0.14
        version: 3.0.14
      - name: apicurio-registry-3.v3.0.12
        version: 3.0.12
      - name: apicurio-registry-3.v3.0.9
        version: 3.0.9
      - name: apicurio-registry-3.v3.0.8
        version: 3.0.8
      - name: apicurio-registry-3.v3.0.7
        version: 3.0.7
      name: 3.x
    defaultChannel: 3.x
    packageName: apicurio-registry-3
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM Edge
      provider-url: ""
    name: infrastructure-asset-orchestrator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infrastructure-asset-orchestrator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infrastructure.asset.orchestrator.com/v1",
                "kind": "Generatekey",
                "metadata": {
                  "name": "modelbuilder-token"
                },
                "spec": {}
              },
              {
                "apiVersion": "infrastructure.asset.orchestrator.com/v1",
                "kind": "Mb-broker-service",
                "metadata": {
                  "name": "mb-broker-service-sample"
                },
                "spec": {
                  "env_type": "prod",
                  "cloud_type": "IBM",
                  "external_db": {
                    "enabled": false
                  },
                  "postgres": {
                    "enable_pg_tls": true,
                    "storage_class": "",
                    "storage_size": "20G",
                    "backup_type": "incremental",
                    "backup_frequency": "@daily"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/iao-operator@sha256:5cc395f22d86641e175339deb134af6c74de92b81470a65a5c736c8005d645fd
          createdAt: "2021-08-19T12:00:00Z"
          olm.skipRange: <1.0.2
          operators.operatorframework.io/builder: operator-sdk-v1.12.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/infrastructure-asset-orchestrator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Generate key CR
            displayName: GenerateKey
            kind: Generatekey
            name: generatekeys.infrastructure.asset.orchestrator.com
            version: v1
          - description: Deploy MB broker deployment
            displayName: MBBrokerService
            kind: Mb-broker-service
            name: mb-broker-services.infrastructure.asset.orchestrator.com
            version: v1
        description: Infrastructure Asset Orchestrator provides users access to cloud
          resources and services. Leverages the dynamic nature of the cloud platforms
          to provide significant cost savings and efficiencies. Using the Infrastructure
          Asset Orchestrator user can provision, configure, utilize and deprovision
          variety of services on the IBM Cloud Catalog.
        displayName: Infrastructure Asset Orchestrator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cloud-service
        - mbbroker-service
        links:
        - name: Infrastructure Asset Orchestrator
          url: https://infrastructure-asset-orchestrator.domain
        maintainers:
        - email: edge@us.ibm.com
          name: IBM-Edge
        maturity: alpha
        provider:
          name: IBM Edge
        relatedImages:
        - registry.connect.redhat.com/ibm/iao-operator@sha256:5cc395f22d86641e175339deb134af6c74de92b81470a65a5c736c8005d645fd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:609c19d1718947ee814b829334c37e1d720acf5fb642b0712b8ab1b14d57f53f
        - registry.connect.redhat.com/ibm/iao-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/iao-orchestrator-service@sha256:145928929fd0c2b7d9ed5fb374d7cc0e6b48269847e810c5472580a1538ad44c
        - registry.connect.redhat.com/ibm/iao-mb-broker-service@sha256:6531b4196c51246958027453c2d9c7dbcdb4a1f55c826d1da238d256cdbc8a10
        version: 1.0.2
      entries:
      - name: infrastructure-asset-orchestrator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: infrastructure-asset-orchestrator-certified
    provider:
      name: IBM Edge
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: node-maintenance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-maintenance-operator.v5.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:58846ab62346b1004e11f36bd901c96f8d315958eb39fbb272d8b328faebcf04
          createdAt: "2023-10-20 15:36:51"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=4.10.0 <5.0.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-node-maintenance-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:58846ab62346b1004e11f36bd901c96f8d315958eb39fbb272d8b328faebcf04
        version: 5.0.1
      entries:
      - name: node-maintenance-operator.v5.0.1
        version: 5.0.1
      name: 4.12-eus
    - currentCSV: node-maintenance-operator.v5.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:2efd1976d2fb110391e0e5937b90476d3eeaacb4dd29cd52375f2b8566f17b41
          createdAt: "2023-10-20 15:50:49"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=5.0.0 <5.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:2efd1976d2fb110391e0e5937b90476d3eeaacb4dd29cd52375f2b8566f17b41
        version: 5.2.1
      entries:
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.14-eus
    - currentCSV: node-maintenance-operator.v5.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:7ca5853885534198544d23067087d80047b0e1421903085a2cb328bd00def147
          createdAt: "2024-08-10 19:55:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.2.0 <5.3.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:7ca5853885534198544d23067087d80047b0e1421903085a2cb328bd00def147
        version: 5.3.1
      entries:
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.16-eus
    - currentCSV: node-maintenance-operator.v5.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-10 10:59:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.3.0 <5.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.4/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel9-operator@sha256:752bff8bebcaf3df1f9d126dc4b07b622a44b02bfcf0f82b9100a7693627079a
        version: 5.4.1
      entries:
      - name: node-maintenance-operator.v5.4.1
        version: 5.4.1
      - name: node-maintenance-operator.v5.4.0
        version: 5.4.0
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.18-eus
    - currentCSV: node-maintenance-operator.v5.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-10 10:59:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.3.0 <5.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.4/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel9-operator@sha256:752bff8bebcaf3df1f9d126dc4b07b622a44b02bfcf0f82b9100a7693627079a
        version: 5.4.1
      entries:
      - name: node-maintenance-operator.v5.4.1
        version: 5.4.1
      - name: node-maintenance-operator.v5.4.0
        version: 5.4.0
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: stable
    defaultChannel: stable
    packageName: node-maintenance-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sagemaker-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sagemaker-controller.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "App",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "DataQualityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Domain",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Endpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "EndpointConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "FeatureGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "HyperParameterTuningJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Model",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelBiasJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelExplainabilityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelPackage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelPackageGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelQualityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "MonitoringSchedule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "NotebookInstance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "NotebookInstanceLifecycleConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ProcessingJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "TrainingJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "TransformJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "UserProfile",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sagemaker-controller:1.4.1
          createdAt: "2025-09-22T19:07:36Z"
          description: ACK service controller for Amazon SageMaker manages SageMaker
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: App represents the state of an AWS sagemaker App resource.
            displayName: App
            kind: App
            name: apps.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: DataQualityJobDefinition represents the state of an AWS sagemaker
              DataQualityJobDefinition resource.
            displayName: DataQualityJobDefinition
            kind: DataQualityJobDefinition
            name: dataqualityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Domain represents the state of an AWS sagemaker Domain resource.
            displayName: Domain
            kind: Domain
            name: domains.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: EndpointConfig represents the state of an AWS sagemaker EndpointConfig
              resource.
            displayName: EndpointConfig
            kind: EndpointConfig
            name: endpointconfigs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Endpoint represents the state of an AWS sagemaker Endpoint
              resource.
            displayName: Endpoint
            kind: Endpoint
            name: endpoints.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: FeatureGroup represents the state of an AWS sagemaker FeatureGroup
              resource.
            displayName: FeatureGroup
            kind: FeatureGroup
            name: featuregroups.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: HyperParameterTuningJob represents the state of an AWS sagemaker
              HyperParameterTuningJob resource.
            displayName: HyperParameterTuningJob
            kind: HyperParameterTuningJob
            name: hyperparametertuningjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: InferenceComponent represents the state of an AWS sagemaker
              InferenceComponent resource.
            displayName: InferenceComponent
            kind: InferenceComponent
            name: inferencecomponents.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: LabelingJob represents the state of an AWS sagemaker LabelingJob
              resource.
            displayName: LabelingJob
            kind: LabelingJob
            name: labelingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelBiasJobDefinition represents the state of an AWS sagemaker
              ModelBiasJobDefinition resource.
            displayName: ModelBiasJobDefinition
            kind: ModelBiasJobDefinition
            name: modelbiasjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelExplainabilityJobDefinition represents the state of
              an AWS sagemaker ModelExplainabilityJobDefinition resource.
            displayName: ModelExplainabilityJobDefinition
            kind: ModelExplainabilityJobDefinition
            name: modelexplainabilityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelPackageGroup represents the state of an AWS sagemaker
              ModelPackageGroup resource.
            displayName: ModelPackageGroup
            kind: ModelPackageGroup
            name: modelpackagegroups.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelPackage represents the state of an AWS sagemaker ModelPackage
              resource.
            displayName: ModelPackage
            kind: ModelPackage
            name: modelpackages.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelQualityJobDefinition represents the state of an AWS
              sagemaker ModelQualityJobDefinition resource.
            displayName: ModelQualityJobDefinition
            kind: ModelQualityJobDefinition
            name: modelqualityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Model represents the state of an AWS sagemaker Model resource.
            displayName: Model
            kind: Model
            name: models.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: MonitoringSchedule represents the state of an AWS sagemaker
              MonitoringSchedule resource.
            displayName: MonitoringSchedule
            kind: MonitoringSchedule
            name: monitoringschedules.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: NotebookInstanceLifecycleConfig represents the state of an
              AWS sagemaker NotebookInstanceLifecycleConfig resource.
            displayName: NotebookInstanceLifecycleConfig
            kind: NotebookInstanceLifecycleConfig
            name: notebookinstancelifecycleconfigs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: NotebookInstance represents the state of an AWS sagemaker
              NotebookInstance resource.
            displayName: NotebookInstance
            kind: NotebookInstance
            name: notebookinstances.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: PipelineExecution represents the state of an AWS sagemaker
              PipelineExecution resource.
            displayName: PipelineExecution
            kind: PipelineExecution
            name: pipelineexecutions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Pipeline represents the state of an AWS sagemaker Pipeline
              resource.
            displayName: Pipeline
            kind: Pipeline
            name: pipelines.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ProcessingJob represents the state of an AWS sagemaker ProcessingJob
              resource.
            displayName: ProcessingJob
            kind: ProcessingJob
            name: processingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: TrainingJob represents the state of an AWS sagemaker TrainingJob
              resource.
            displayName: TrainingJob
            kind: TrainingJob
            name: trainingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: TransformJob represents the state of an AWS sagemaker TransformJob
              resource.
            displayName: TransformJob
            kind: TransformJob
            name: transformjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: UserProfile represents the state of an AWS sagemaker UserProfile
              resource.
            displayName: UserProfile
            kind: UserProfile
            name: userprofiles.sagemaker.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon SageMaker resources in AWS from within your Kubernetes cluster.

          **About Amazon SageMaker**

          Amazon SageMaker helps data scientists and developers to prepare, build, train, and deploy high-quality machine learning (ML) models quickly by bringing together a broad set of capabilities purpose-built for ML.

          For more information on Amazon SageMaker, visit the [product page](https://aws.amazon.com/sagemaker/).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SageMaker
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sagemaker
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SageMaker Developer Resources
          url: https://aws.amazon.com/sagemaker/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sagemaker maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sagemaker-controller:1.4.1
        version: 1.4.1
      entries:
      - name: ack-sagemaker-controller.v1.4.1
        version: 1.4.1
      - name: ack-sagemaker-controller.v1.4.0
        version: 1.4.0
      - name: ack-sagemaker-controller.v1.3.3
        version: 1.3.3
      - name: ack-sagemaker-controller.v1.3.2
        version: 1.3.2
      - name: ack-sagemaker-controller.v1.3.1
        version: 1.3.1
      - name: ack-sagemaker-controller.v1.3.0
        version: 1.3.0
      - name: ack-sagemaker-controller.v1.2.20
        version: 1.2.20
      - name: ack-sagemaker-controller.v1.2.19
        version: 1.2.19
      - name: ack-sagemaker-controller.v1.2.18
        version: 1.2.18
      - name: ack-sagemaker-controller.v1.2.17
        version: 1.2.17
      - name: ack-sagemaker-controller.v1.2.16
        version: 1.2.16
      - name: ack-sagemaker-controller.v1.2.13
        version: 1.2.13
      - name: ack-sagemaker-controller.v1.2.12
        version: 1.2.12
      - name: ack-sagemaker-controller.v1.2.11
        version: 1.2.11
      - name: ack-sagemaker-controller.v1.2.9
        version: 1.2.9
      - name: ack-sagemaker-controller.v1.2.8
        version: 1.2.8
      - name: ack-sagemaker-controller.v1.2.5
        version: 1.2.5
      - name: ack-sagemaker-controller.v1.2.4
        version: 1.2.4
      - name: ack-sagemaker-controller.v1.2.3
        version: 1.2.3
      - name: ack-sagemaker-controller.v1.2.2
        version: 1.2.2
      - name: ack-sagemaker-controller.v1.2.1
        version: 1.2.1
      - name: ack-sagemaker-controller.v1.2.0
        version: 1.2.0
      - name: ack-sagemaker-controller.v1.1.0
        version: 1.1.0
      - name: ack-sagemaker-controller.v1.0.0
        version: 1.0.0
      - name: ack-sagemaker-controller.v0.6.0
        version: 0.6.0
      - name: ack-sagemaker-controller.v0.5.1
        version: 0.5.1
      - name: ack-sagemaker-controller.v0.5.0
        version: 0.5.0
      - name: ack-sagemaker-controller.v0.4.5
        version: 0.4.5
      - name: ack-sagemaker-controller.v0.4.4
        version: 0.4.4
      - name: ack-sagemaker-controller.v0.4.3
        version: 0.4.3
      - name: ack-sagemaker-controller.v0.4.2
        version: 0.4.2
      - name: ack-sagemaker-controller.v0.4.0
        version: 0.4.0
      - name: ack-sagemaker-controller.v0.3.4
        version: 0.3.4
      - name: ack-sagemaker-controller.v0.3.3
        version: 0.3.3
      - name: ack-sagemaker-controller.v0.3.2
        version: 0.3.2
      - name: ack-sagemaker-controller.v0.3.1
        version: 0.3.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-sagemaker-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: networkservicemesh.io
      provider-url: networkservicemesh.io
    name: nsm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nsm-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nsm.networkservicemesh.io/v1alpha1",
                "kind": "NSM",
                "metadata": {
                  "name": "nsm-sample",
                  "namespace": "nsm"
                },
                "spec": {
                  "forwarders": [
                    {
                      "image": "ghcr.io/networkservicemesh/cmd-forwarder-vpp:v1.8.0",
                      "type": "vpp"
                    }
                  ],
                  "nsmLogLevel": "TRACE",
                  "nsmPullPolicy": "IfNotPresent",
                  "registry": {
                    "image": "ghcr.io/networkservicemesh/cmd-registry-memory:v1.8.0",
                    "type": "memory"
                  },
                  "version": "v1.8.0",
                  "webhook": {
                    "image": "ghcr.io/networkservicemesh/admission-webhook-k8s:v1.8.0"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: quay.io/acmenezes/nsm-operator:v1.8.0
          createdAt: "2023-03-02T19:36:46Z"
          operatorframework.io/suggested-namespace: nsm
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/networkservicemesh/nsm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkServiceEndpoint
            name: networkserviceendpoints.networkservicemesh.io
            version: v1
          - kind: NetworkService
            name: networkservices.networkservicemesh.io
            version: v1
          - kind: NSM
            name: nsms.nsm.networkservicemesh.io
            version: v1alpha1
        description: |-
          The Network Service Mesh Operator manages and install network service meshes.

          Check networkservicemesh.io

          Install requirements:

          Before installing the operator it's necessary to have installed [spire](https://github.com/spiffe/spire)

          We have a ready to install yaml file for spire [here](https://github.com/networkservicemesh/nsm-operator/blob/master/config/spire/spire.yaml)
        displayName: nsm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network
        - service mesh
        links:
        - name: Nsm Operator
          url: https://nsm-operator.domain
        maintainers:
        - email: adcmenezes@gmail.com
          name: Alexandre Menezes
        maturity: alpha
        provider:
          name: networkservicemesh.io
          url: networkservicemesh.io
        relatedImages:
        - quay.io/acmenezes/nsm-operator:v1.8.0
        version: 1.8.0
      entries:
      - name: nsm-operator.v1.8.0
        version: 1.8.0
      name: alpha
    defaultChannel: alpha
    packageName: nsm-operator
    provider:
      name: networkservicemesh.io
      url: networkservicemesh.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: observability-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/observability-operator:0.0.28
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      name: candidate
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      - name: observability-operator.v0.0.22
        version: 0.0.22
      name: development
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      name: stable
    defaultChannel: stable
    packageName: observability-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-atlas-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-atlas-kubernetes.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "id": "1",
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "WEEKLY",
                      "id": "2",
                      "retentionUnit": "DAYS",
                      "retentionValue": 6
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": true,
                  "policies": [
                    {
                      "name": "atlas-default-backuppolicy",
                      "namespace": "mongodb-atlas-system"
                    },
                    {
                      "name": "atlas-default-backuppolicy2",
                      "namespace": "mongodb-atlas-system"
                    }
                  ],
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "my-database-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "david"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-cluster"
                },
                "spec": {
                  "deploymentSpec": {
                    "name": "test-cluster",
                    "providerSettings": {
                      "instanceSizeName": "M10",
                      "providerName": "AWS",
                      "regionName": "US_EAST_1"
                    }
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          containerImage: registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator@sha256:3cfbd700e4771c2c7e75c2c41cf6de57c9ef71ecd286e96a21d2567e43424bc3
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasclusters API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB Atlas —
          the only multi-cloud document database service that gives you the versatility you need to build sophisticated and resilient applications that can adapt to changing customer demands and market trends.

          > Current Status: *Beta*. The Operator gives users the ability to provision
          > Atlas projects, clusters and database users using Kubernetes Specifications and bind connection information
          > into applications deployed to Kubernetes.

          ## Quick Start guide
          ### Step 1. Deploy Kubernetes operator by clicking Install button.

          ### Step 2. Create Atlas Cluster

          **1.** Create an Atlas API Key Secret
          In order to work with the Atlas Operator you need to provide [authentication information](https://docs.atlas.mongodb.com/configure-api-access)
           to allow the Atlas Operator to communicate with Atlas API. Once you have generated a Public and Private key in Atlas, you can create a Kuberentes Secret with:
          ```
          kubectl create secret generic mongodb-atlas-operator-api-key \
                   --from-literal="orgId=<the_atlas_organization_id>" \
                   --from-literal="publicApiKey=<the_atlas_api_public_key>" \
                   --from-literal="privateApiKey=<the_atlas_api_private_key>" \
                   -n openshift-operators
          ```
          (Note, that you should use the namespace where the Operator was installed - it's `openshift-operators` by default)

          **2.** Create an `AtlasProject` Custom Resource

          The `AtlasProject` CustomResource represents Atlas Projects in our Kubernetes cluster. You need to specify
          `projectIpAccessList` with the IP addresses or CIDR blocks of any hosts that will connect to the Atlas Cluster.
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasProject
          metadata:
            name: my-project
          spec:
            name: Test Atlas Operator Project
            projectIpAccessList:
              - ipAddress: "192.0.2.15"
                comment: "IP address for Application Server A"
              - ipAddress: "203.0.113.0/24"
                comment: "CIDR block for Application Server B - D"
          ```
          **3.** Create an `AtlasDeployment` Custom Resource.
          The example below is a minimal configuration to create an M10 Atlas cluster in the AWS US East region. For a full list of properties, check
          `atlasclusters.atlas.mongodb.com` [CRD specification](config/crd/bases/atlas.mongodb.com_atlasclusters.yaml)):
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasDeployment
          metadata:
            name: my-atlas-cluster
          spec:
            name: "Test-cluster"
            projectRef:
              name: my-project
            providerSettings:
              instanceSizeName: M10
              providerName: AWS
              regionName: US_EAST_1
          ```

          **4.** Create a database user password Kubernetes Secret
          The Secret must be created in the same namespace as the `AtlasDeployment` and `AtlasProject` were created.
          ```
          kubectl create secret generic the-user-password --from-literal="password=P@@sword%"
          ```

          **5.** Create an `AtlasDatabaseUser` Custom Resource

          In order to connect to an Atlas Cluster the database user needs to be created. `AtlasDatabaseUser` resource should reference
          the password Kubernetes Secret created in the previous step.
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasDatabaseUser
          metadata:
            name: my-database-user
          spec:
            roles:
              - roleName: "readWriteAnyDatabase"
                databaseName: "admin"
            projectRef:
              name: my-project
            username: theuser
            passwordSecretRef:
              name: the-user-password
          ```
          **6.** Wait for the `AtlasDatabaseUser` Custom Resource to be ready

          Wait until the AtlasDatabaseUser resource gets to "ready" status (it will wait until the cluster is created that may take around 10 minutes):
          ```
          kubectl get atlasdatabaseusers my-database-user -o=jsonpath='{.status.conditions[?(@.type=="Ready")].status}'
          True
          ```
          ### Step 3. Connect your application to the Atlas Cluster

          The Atlas Operator will create a Kubernetes Secret with the information necessary to connect to the Atlas Cluster created
          in the previous step. An application in the same Kubernetes Cluster can mount and use the Secret:

          ```
          ...
          containers:
                - name: test-app
                  env:
                   - name: "CONNECTION_STRING"
                     valueFrom:
                       secretKeyRef:
                         name: test-atlas-operator-project-test-cluster-theuser
                         key: connectionStringStandardSrv

          ```
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator@sha256:3cfbd700e4771c2c7e75c2c41cf6de57c9ef71ecd286e96a21d2567e43424bc3
        version: 1.0.0
      entries:
      - name: mongodb-atlas-kubernetes.v1.0.0
        version: 1.0.0
      name: beta
    - currentCSV: mongodb-atlas-kubernetes.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupCompliancePolicy",
                "metadata": {
                  "name": "my-backup-compliance-policy"
                },
                "spec": {
                  "authorizedEmail": "john.doe@example.com",
                  "authorizedUserFirstName": "John",
                  "authorizedUserLastName": "Doe",
                  "copyProtectionEnabled": false,
                  "encryptionAtRestEnabled": false,
                  "onDemandPolicy": {
                    "retentionUnit": "weeks",
                    "retentionValue": 3
                  },
                  "overwriteBackupPolicies": false,
                  "pointInTimeEnabled": true,
                  "restoreWindowDays": 42,
                  "scheduledPolicyItems": [
                    {
                      "frequencyInterval": 2,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "hourly",
                      "retentionUnit": "days",
                      "retentionValue": 2
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "weekly",
                      "retentionUnit": "weeks",
                      "retentionValue": 4
                    },
                    {
                      "frequencyInterval": 15,
                      "frequencyType": "monthly",
                      "retentionUnit": "months",
                      "retentionValue": 12
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": false,
                  "copySettings": [
                    {
                      "cloudProvider": "AWS",
                      "frequencies": [
                        "WEEKLY"
                      ],
                      "regionName": "US_WEST_1",
                      "shouldCopyOplogs": false
                    }
                  ],
                  "policy": {
                    "name": "atlasbackuppolicy-sample"
                  },
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2,
                  "useOrgAndGroupNamesInExportPrefix": true
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasCustomRole",
                "metadata": {
                  "name": "shard-operator-role"
                },
                "spec": {
                  "projectRef": {
                    "name": "my-project"
                  },
                  "role": {
                    "actions": [
                      {
                        "name": "getShardMap",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "shardingState",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "connPoolStats",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "getLog",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      }
                    ],
                    "inheritedRoles": [
                      {
                        "database": "admin",
                        "name": "operator-role-1"
                      }
                    ],
                    "name": "my-role"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDataFederation",
                "metadata": {
                  "name": "my-data-federation",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "name": "my-data-federation",
                  "privateEndpoints": [
                    {
                      "endpointId": "vpce-03f9eeaa764e32454",
                      "provider": "AWS",
                      "type": "DATA_LAKE"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project",
                    "namespace": "mongodb-atlas-system"
                  },
                  "storage": {
                    "databases": [
                      {
                        "collections": [
                          {
                            "dataSources": [
                              {
                                "storeName": "http-test",
                                "urls": [
                                  "https://data.cityofnewyork.us/api/views/vfnx-vebw/rows.csv"
                                ]
                              }
                            ],
                            "name": "test-collection-1"
                          }
                        ],
                        "name": "test-db-1"
                      }
                    ],
                    "stores": [
                      {
                        "name": "http-test",
                        "provider": "http"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "the-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "theuser"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-deployment"
                },
                "spec": {
                  "backupRef": {
                    "name": "atlasbackupschedule-sample"
                  },
                  "deploymentSpec": {
                    "backupEnabled": true,
                    "clusterType": "REPLICASET",
                    "mongoDBMajorVersion": "7.0",
                    "name": "test-deployment",
                    "replicationSpecs": [
                      {
                        "regionConfigs": [
                          {
                            "electableSpecs": {
                              "instanceSize": "M10",
                              "nodeCount": 3
                            },
                            "priority": 7,
                            "providerName": "AWS",
                            "regionName": "US_EAST_1"
                          }
                        ],
                        "zoneName": "Zone 1"
                      }
                    ],
                    "tags": [
                      {
                        "key": "team",
                        "value": "atlas-operator"
                      }
                    ]
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasFederatedAuth",
                "metadata": {
                  "name": "atlas-default-federated-auth",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "connectionSecretRef": {
                    "name": "my-org-secret",
                    "namespace": "mongodb-atlas-system"
                  },
                  "domainAllowList": [
                    "my-org-domain.com"
                  ],
                  "domainRestrictionEnabled": true,
                  "enabled": true,
                  "postAuthRoleGrants": [
                    "ORG_MEMBER"
                  ],
                  "roleMappings": [
                    {
                      "externalGroupName": "org-admin",
                      "roleAssignments": [
                        {
                          "role": "ORG_OWNER"
                        }
                      ]
                    },
                    {
                      "externalGroupName": "dev-team",
                      "roleAssignments": [
                        {
                          "role": "ORG_GROUP_CREATOR"
                        },
                        {
                          "projectName": "dev-project",
                          "role": "GROUP_OWNER"
                        }
                      ]
                    }
                  ],
                  "ssoDebugEnabled": false
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasIPAccessList",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasipaccesslist-sample"
                },
                "spec": {
                  "entries": [
                    {
                      "cidrBlock": "192.168.1.0/24"
                    },
                    {
                      "deleteAfterDate": "2025-01-20T23:59:59+02:00",
                      "ipAddress": "10.0.0.1"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkContainer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasnetworkcontainer-sample"
                },
                "spec": {
                  "cidrBlock": "10.11.0.0/16",
                  "connectionSecret": {
                    "name": "atlas-connection-secret"
                  },
                  "externalProjectRef": {
                    "projectId": "66e2f2b621571b7e69a89b66"
                  },
                  "provider": "AWS",
                  "region": "US_EAST_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkPeering",
                "metadata": {
                  "name": "atlasnetworkpeering-sample"
                },
                "spec": {
                  "atlasCidrBlock": "10.12.0.0/16",
                  "awsConfiguration": {
                    "accepterRegionName": "US_EAST_1",
                    "awsAccountId": "23214235145",
                    "routeTableCidrBlock": "10.11.0.0/16",
                    "vpcId": "vpcid-7123782134"
                  },
                  "containerId": "623412394512350",
                  "containerRegion": "US_EAST_1",
                  "projectRef": {
                    "name": "atlas-project",
                    "namespace": "namespace"
                  },
                  "provider": "AWS"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasPrivateEndpoint",
                "metadata": {
                  "name": "atlasprivateendpoint-sample"
                },
                "spec": {
                  "awsConfiguration": {
                    "id": "vpce-f4k34w51d"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "provider": "AWS",
                  "region": "EU_CENTRAL_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    },
                    {
                      "cidrBlock": "203.0.113.0/24",
                      "comment": "CIDR block for Application Server B - D"
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasSearchIndexConfig",
                "metadata": {
                  "name": "atlassearchindexconfig-sample"
                },
                "spec": {
                  "analyzer": "lucene.standard",
                  "analyzers": [
                    {
                      "name": "test-analyzer",
                      "tokenFilters": [
                        {
                          "asciiFolding": {
                            "originalTokens": "omit"
                          },
                          "type": "asciiFolding"
                        }
                      ],
                      "tokenizer": {
                        "type": "whiteSpace",
                        "whitespace": {
                          "maxTokenLenght": 200
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamConnection",
                "metadata": {
                  "name": "my-streamconnection-kafka"
                },
                "spec": {
                  "kafkaConfig": {
                    "authentication": {
                      "credentials": {
                        "name": "kafka-userpass"
                      },
                      "mechanism": "SCRAM-512"
                    },
                    "bootstrapServers": "kafka.server1:9001,kafka.server2:9002,kafka.server3:9003",
                    "security": {
                      "certificate": {
                        "name": "kafka-cert"
                      },
                      "protocol": "SSL"
                    }
                  },
                  "name": "kafka-config",
                  "type": "Kafka"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamInstance",
                "metadata": {
                  "name": "my-streaminstance-sample"
                },
                "spec": {
                  "clusterConfig": {
                    "provider": "AWS",
                    "region": "VIRGINIA_USA",
                    "tier": "SP30"
                  },
                  "connectionRegistry": [
                    {
                      "name": "my-streamconnection-sample"
                    }
                  ],
                  "name": "my-instance",
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasTeam",
                "metadata": {
                  "name": "atlasteam-sample"
                },
                "spec": {
                  "name": "my-team",
                  "usernames": [
                    "username@mongodb.com"
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasThirdPartyIntegration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasintegration-sample"
                },
                "spec": {
                  "datadog": {
                    "apiKeySecretRef": {
                      "name": "secret-ref"
                    },
                    "region": "string",
                    "sendCollectionLatencyMetrics": "string",
                    "sendDatabaseMetrics": "string"
                  },
                  "projectRef": {
                    "name": "my-atlas-project"
                  },
                  "type": "DATADOG"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:8ac3f6b1c0f31af5f32562ddeed37dd6586ac47e1e7e111947df3ed570c5bed9
          createdAt: "2025-09-25T14:34:34Z"
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/mongodb/mongodb-atlas-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The AtlasBackupCompliancePolicy is a configuration that enforces
              specific backup and retention requirements
            displayName: Atlas Backup Compliance Policy
            kind: AtlasBackupCompliancePolicy
            name: atlasbackupcompliancepolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasCustomRole is the Schema for the AtlasCustomRole API
            displayName: Atlas Custom Role
            kind: AtlasCustomRole
            name: atlascustomroles.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDataFederation is the Schema for the Atlas Data Federation
              API
            displayName: Atlas Data Federation
            kind: AtlasDataFederation
            name: atlasdatafederations.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasdeployments API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasFederatedAuth is the Schema for the Atlasfederatedauth
              API
            displayName: Atlas Federated Auth
            kind: AtlasFederatedAuth
            name: atlasfederatedauths.atlas.mongodb.com
            version: v1
          - description: AtlasIPAccessList is the Schema for the atlasipaccesslists
              API.
            displayName: Atlas IPAccess List
            kind: AtlasIPAccessList
            name: atlasipaccesslists.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkContainer is the Schema for the AtlasNetworkContainer
              API
            displayName: Atlas Network Container
            kind: AtlasNetworkContainer
            name: atlasnetworkcontainers.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkPeering is the Schema for the AtlasNetworkPeering
              API
            displayName: Atlas Network Peering
            kind: AtlasNetworkPeering
            name: atlasnetworkpeerings.atlas.mongodb.com
            version: v1
          - displayName: Atlas Org Settings
            kind: AtlasOrgSettings
            name: atlasorgsettings.atlas.mongodb.com
            version: v1
          - description: "The AtlasPrivateEndpoint custom resource definition (CRD)
              defines a desired [Private Endpoint](https://www.mongodb.com/docs/atlas/security-private-endpoint/#std-label-private-endpoint-overview)
              configuration for an Atlas project. It allows a private connection between
              your cloud provider and Atlas that doesn't send information through
              a public network. \n You can use private endpoints to create a unidirectional
              connection to Atlas clusters from your virtual network."
            displayName: Atlas Private Endpoint
            kind: AtlasPrivateEndpoint
            name: atlasprivateendpoints.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
          - description: Atlas Search Index Config is the Schema for the Atlas Search
              API
            displayName: Atlas Search Index Config
            kind: AtlasSearchIndexConfig
            name: atlassearchindexconfigs.atlas.mongodb.com
            version: v1
          - description: AtlasStreamConnection is the Schema for the atlasstreamconnections
              API
            displayName: Atlas Stream Connection
            kind: AtlasStreamConnection
            name: atlasstreamconnections.atlas.mongodb.com
            version: v1
          - description: AtlasStreamInstance is the Schema for the atlasstreaminstances
              API
            displayName: Atlas Stream Instance
            kind: AtlasStreamInstance
            name: atlasstreaminstances.atlas.mongodb.com
            version: v1
          - description: AtlasTeam is the Schema for the Atlas Teams API
            displayName: Atlas Team
            kind: AtlasTeam
            name: atlasteams.atlas.mongodb.com
            version: v1
          - description: AtlasThirdPartyIntegration is the Schema for the atlas 3rd
              party integrations API.
            displayName: Atlas Third Party Integration
            kind: AtlasThirdPartyIntegration
            name: atlasthirdpartyintegrations.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB
          Atlas — the only multi-cloud document database service that gives you the versatility you need to build sophisticated
          and resilient applications that can adapt to changing customer demands and market trends.

          The full documentation for the Operator can be found [here](https://docs.atlas.mongodb.com/atlas-operator/)

          ## Getting Started

          ### Supported features

          * Create and configure a project, or connect to an existing one.
          * Deploy, manage, scale, and tear down clusters.
          * Support for serverless instances.
          * Create and edit database users.
          * Manage IP Access Lists, network peering and private endpoints.
          * Configure and control Atlas’s fully managed cloud backup.
          * Configure federated authentication for your Atlas organization
          * Integrate Atlas monitoring with Prometheus.

          ... and more.

          To view the list of custom resources and each CR's schema, visit our [reference](https://www.mongodb.com/docs/atlas/operator/stable/custom-resources/)
          documentation. See the [Quickstart](https://www.mongodb.com/docs/atlas/operator/stable/ak8so-quick-start/) to get
          started with Atlas Kubernetes Operator.
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:8ac3f6b1c0f31af5f32562ddeed37dd6586ac47e1e7e111947df3ed570c5bed9
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:9eb6465e723ede43ac47eed6de22042d26c0efb276301a4424a6a3948a8f5b15
        version: 2.11.0
      entries:
      - name: mongodb-atlas-kubernetes.v2.11.0
        version: 2.11.0
      - name: mongodb-atlas-kubernetes.v2.10.0
        version: 2.10.0
      - name: mongodb-atlas-kubernetes.v2.9.1
        version: 2.9.1
      - name: mongodb-atlas-kubernetes.v2.9.0
        version: 2.9.0
      - name: mongodb-atlas-kubernetes.v2.8.2
        version: 2.8.2
      - name: mongodb-atlas-kubernetes.v2.8.1
        version: 2.8.1
      - name: mongodb-atlas-kubernetes.v2.8.0
        version: 2.8.0
      - name: mongodb-atlas-kubernetes.v2.7.1
        version: 2.7.1
      - name: mongodb-atlas-kubernetes.v2.7.0
        version: 2.7.0
      - name: mongodb-atlas-kubernetes.v2.6.1
        version: 2.6.1
      - name: mongodb-atlas-kubernetes.v2.6.0
        version: 2.6.0
      - name: mongodb-atlas-kubernetes.v2.5.0
        version: 2.5.0
      - name: mongodb-atlas-kubernetes.v2.4.1
        version: 2.4.1
      - name: mongodb-atlas-kubernetes.v2.4.0
        version: 2.4.0
      - name: mongodb-atlas-kubernetes.v2.3.1
        version: 2.3.1
      - name: mongodb-atlas-kubernetes.v2.3.0
        version: 2.3.0
      - name: mongodb-atlas-kubernetes.v2.2.2
        version: 2.2.2
      - name: mongodb-atlas-kubernetes.v2.2.1
        version: 2.2.1
      - name: mongodb-atlas-kubernetes.v2.2.0
        version: 2.2.0
      - name: mongodb-atlas-kubernetes.v2.1.0
        version: 2.1.0
      - name: mongodb-atlas-kubernetes.v2.0.1
        version: 2.0.1
      - name: mongodb-atlas-kubernetes.v2.0.0
        version: 2.0.0
      - name: mongodb-atlas-kubernetes.v1.9.1
        version: 1.9.1
      - name: mongodb-atlas-kubernetes.v1.9.0
        version: 1.9.0
      - name: mongodb-atlas-kubernetes.v1.8.2
        version: 1.8.2
      - name: mongodb-atlas-kubernetes.v1.8.1
        version: 1.8.1
      - name: mongodb-atlas-kubernetes.v1.8.0
        version: 1.8.0
      - name: mongodb-atlas-kubernetes.v1.7.3
        version: 1.7.3
      - name: mongodb-atlas-kubernetes.v1.7.2
        version: 1.7.2
      - name: mongodb-atlas-kubernetes.v1.7.1
        version: 1.7.1
      - name: mongodb-atlas-kubernetes.v1.7.0
        version: 1.7.0
      - name: mongodb-atlas-kubernetes.v1.6.1
        version: 1.6.1
      - name: mongodb-atlas-kubernetes.v1.6.0
        version: 1.6.0
      - name: mongodb-atlas-kubernetes.v1.5.1
        version: 1.5.1
      - name: mongodb-atlas-kubernetes.v1.5.0
        version: 1.5.0
      - name: mongodb-atlas-kubernetes.v1.4.1
        version: 1.4.1
      - name: mongodb-atlas-kubernetes.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-atlas-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco AppDynamics
      provider-url: https://www.appdynamics.com
    name: appdynamics-cloud-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appdynamics-cloud-operator.v24.7.0-133
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clustermon",
                "metadata": {
                  "name": "clustermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "clusterName": "test",
                  "image": "appdynamics/appdynamics-cloud-k8s-monitoring:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Containermon",
                "metadata": {
                  "name": "containermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "name": "ContainerMon"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "DbCollector",
                "metadata": {
                  "name": "dbcollector",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-db-collector:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "DbConfig",
                "metadata": {
                  "name": "dbconfig",
                  "namespace": "appdynamics"
                },
                "spec": null
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Inframon",
                "metadata": {
                  "name": "inframon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-k8s-monitoring:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "LogCollector",
                "metadata": {
                  "name": "logcollector-sample",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-log-collector-agent:latest",
                  "resources": {
                    "limits": {
                      "cpu": "400m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "150Mi"
                    }
                  }
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Servermon",
                "metadata": {
                  "name": "servermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "name": "ServerMon"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-08-06T08:33:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clustermon is the Schema for the clustermons API
            displayName: Clustermon
            kind: Clustermon
            name: clustermons.cluster.appdynamics.com
            version: v1alpha1
          - description: Containermon is the Schema for the containermons API
            displayName: Containermon
            kind: Containermon
            name: containermons.cluster.appdynamics.com
            version: v1alpha1
          - kind: DbCollector
            name: dbcollectors.cluster.appdynamics.com
            version: v1alpha1
          - kind: DbConfig
            name: dbconfigs.cluster.appdynamics.com
            version: v1alpha1
          - description: Inframon is the Schema for the inframons API
            displayName: Inframon
            kind: Inframon
            name: inframons.cluster.appdynamics.com
            version: v1alpha1
          - kind: LogCollector
            name: logcollectors.cluster.appdynamics.com
            version: v1alpha1
          - description: Servermon is the Schema for the servermons API
            displayName: Servermon
            kind: Servermon
            name: servermons.cluster.appdynamics.com
            version: v1alpha1
        description: The Appdynamics Cloud Operator is used to install Cluster Collector,
          Container Collector, Server Collector, Log Collector and Cluster Agent
        displayName: Cisco AppDynamics Cloud Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - appdynamics
        links:
        - name: Appdynamics Cloud Operator
          url: https://appdynamics-cloud-operator.domain
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: Cisco AppDynamics
          url: https://www.appdynamics.com
        relatedImages:
        - registry.connect.redhat.com/appdynamics/appdynamics-cloud-operator@sha256:2710d63425942f60af21f060a02eb3c40f5a5f0c952ea6593b3b9769c7ca1014
        version: 24.7.0-133
      entries:
      - name: appdynamics-cloud-operator.v24.7.0-133
        version: 24.7.0-133
      name: alpha
    defaultChannel: alpha
    packageName: appdynamics-cloud-operator
    provider:
      name: Cisco AppDynamics
      url: https://www.appdynamics.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: smb-csi-driver-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smb-csi-driver-operator.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9-operator@sha256:1019812a3969f844a0f3fcd45091fccc7d0d816207e230a2f399556c41ebc705
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure CIFS/SMB CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.15.0-0 <4.16.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/csi-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          CIFS/SMB CSI Driver Operator provides Server Message Block (SMB) CSI Driver that enables you to create and mount CIFS/SMB PersistentVolumes.
        displayName: CIFS/SMB CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - smb
        - cifs
        links:
        - name: Documentation
          url: https://github.com/openshift/csi-operator
        - name: Source Repository
          url: https://github.com/openshift/csi-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: preview
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9-operator@sha256:1019812a3969f844a0f3fcd45091fccc7d0d816207e230a2f399556c41ebc705
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0627c09fe3f5a015f117fb30af1f4d17340aefa48b23994737d3901a804b63d3
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:a5704d26d0b562dba9633d1c107f34e05c750c0e3a64dbcace4dab7b964fff8d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3afe22a8786e47f963e0bcb36653137341085e0022f254773c49ed92f198fa55
        - registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9@sha256:9f516d06c279cf1f82e8948ce2290395eec5f6611abf5ea919e4d6548d81ec10
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:9c3c3d2b857ce289324aee56df1ba4ab5e5ea1e76cd86ac988233463915834b4
        version: 4.16.0-202509240123
      entries:
      - name: smb-csi-driver-operator.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: smb-csi-driver-operator.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: smb-csi-driver-operator.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: smb-csi-driver-operator.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: smb-csi-driver-operator.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: smb-csi-driver-operator.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: smb-csi-driver-operator.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: smb-csi-driver-operator.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: smb-csi-driver-operator.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: smb-csi-driver-operator.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: smb-csi-driver-operator.v4.16.0-202501230036
        version: 4.16.0-202501230036
      - name: smb-csi-driver-operator.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: smb-csi-driver-operator.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: smb-csi-driver-operator.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: smb-csi-driver-operator.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: smb-csi-driver-operator.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: smb-csi-driver-operator.v4.16.0-202410220434
        version: 4.16.0-202410220434
      - name: smb-csi-driver-operator.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: smb-csi-driver-operator.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: smb-csi-driver-operator.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: smb-csi-driver-operator.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: smb-csi-driver-operator.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: smb-csi-driver-operator.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: smb-csi-driver-operator.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: smb-csi-driver-operator.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: smb-csi-driver-operator.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: smb-csi-driver-operator.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: smb-csi-driver-operator.v4.16.0-202406251906
        version: 4.16.0-202406251906
      - name: smb-csi-driver-operator.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: smb-csi-driver-operator.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: preview
    defaultChannel: preview
    packageName: smb-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: bamoe-kogito-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bamoe-kogito-operator.8.0.5-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "ruleunit-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogitoinfra-sample"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "kafka.strimzi.io/v1beta2",
                    "kind": "Kafka",
                    "name": "kogito-kafka"
                  }
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator@sha256:8a99fcd217502b7e850af1e1097de58b6cb19d1e786394811a8e075c4d871279
          description: IBM BAMOE Kogito Operator for deployment and management of
            Kogito services.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["IBM Business Automation Manager
            Open Edition"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: https://www.ibm.com/mysupport
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.rhpam.kiegroup.org
            version: v1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.rhpam.kiegroup.org
            version: v1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.rhpam.kiegroup.org
            version: v1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.rhpam.kiegroup.org
            version: v1
        description: |-
          IBM BAMOE Kogito Operator is designed for deployment and management of Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Builds a IBM BAMOE Kogito service from a remote git repository and creates a custom image for the given runtime: Red Hat build of Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Deploy IBM BAMOE Kogito service from user provided image.
          * Setup Grafana Dashboard when Grafana Operator is available in cluster.
          * Create Prometheus monitor object when Prometheus operator is available in cluster.

          **IMPORTANT!** IBM BAMOE Kogito Operator may need Grafana and Prometheus Operator to configure monitoring in cases if you do not provide these services beforehand. If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the IBM BAMOE Kogito Operator.
        displayName: IBM BAMOE Kogito Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://ibm.com
        - name: Documentation
          url: https://www.ibm.com/docs/en/ibamoe
        - name: Source
          url: https://github.com/kiegroup/kogito-operator
        maintainers:
        - email: mysphelp@us.ibm.com
          name: IBM
        maturity: beta
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9492c67eaa7ab37df6f5fa623c4aac955b7d7ffaab9ae2e57dea745e82907e40
        - registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator@sha256:8a99fcd217502b7e850af1e1097de58b6cb19d1e786394811a8e075c4d871279
        version: 8.0.5-3
      entries:
      - name: bamoe-kogito-operator.8.0.5-3
        version: 8.0.5-3
      - name: bamoe-kogito-operator.8.0.5-2
        version: 8.0.5-2
      - name: bamoe-kogito-operator.8.0.5-1
        version: 8.0.5-1
      - name: bamoe-kogito-operator.8.0.4-2
        version: 8.0.4-2
      - name: bamoe-kogito-operator.8.0.4-1
        version: 8.0.4-1
      - name: bamoe-kogito-operator.8.0.3-1
        version: 8.0.3-1
      - name: bamoe-kogito-operator.8.0.2-3
        version: 8.0.2-3
      - name: bamoe-kogito-operator.8.0.2-2
        version: 8.0.2-2
      - name: bamoe-kogito-operator.8.0.2-1
        version: 8.0.2-1
      - name: bamoe-kogito-operator.8.0.1-2
        version: 8.0.1-2
      - name: bamoe-kogito-operator.8.0.1-1
        version: 8.0.1-1
      - name: bamoe-kogito-operator.8.0.0-1
        version: 8.0.0-1
      name: 8.x
    defaultChannel: 8.x
    packageName: bamoe-kogito-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc
      provider-url: ""
    name: openshift-gitops-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-gitops-operator.v1.10.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:b0dc693854319ada341e2e3872b34d7db9823b470ed369edfde88d3ae73a40e1
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.8.19\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:a74e0a1fe3589ebe9f226a2bad20e9a5d5601a0e62c5f3290902f466f55b7160
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:2bec31b0c5407ffcecfd8b860d8afb1629eaab31a250f60191f94a824e0a52f4
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:c779ad87a7f71d851c2af818abf16e5151c43d3fd54e64eac2fc7117c00753f5
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:b0dc693854319ada341e2e3872b34d7db9823b470ed369edfde88d3ae73a40e1
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:b5c8a6c7438a45125fc7fd970946f715ab7193b7446d625fae80dde9d1f3f914
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:4c842f0ea8f0e9b53eb16d94488fc2f4298bd8e4be2f48713be9157da24549e6
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:ece2fee4227ad1becadca47028b78ceaf8717a98ec84e3f1b5188a135c4b093c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d35dff5d59e748ef0404e7e76b0237ea96b2bb19b86fa04a3a34df7a379277b2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5d5061dbc2ec5082cf14b6c600fb5400b83cf91d7ccebfa80680a238d275db
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:0905ba3f5cf87780a1e95b6a00ea7cd340d4ed1b5054e4b2c82406270baeaefa
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ec9f60018694dcc5d431ba47d5536b761b71cb3f66684978fe6bb74c157679ac
        version: 1.10.6
      entries:
      - name: openshift-gitops-operator.v1.10.6
        version: 1.10.6
      - name: openshift-gitops-operator.v1.10.5
        version: 1.10.5
      - name: openshift-gitops-operator.v1.10.4
        version: 1.10.4
      - name: openshift-gitops-operator.v1.10.3
        version: 1.10.3
      - name: openshift-gitops-operator.v1.10.2
        version: 1.10.2
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.10
    - currentCSV: openshift-gitops-operator.v1.11.7-0.1724840231.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d6ad2ee900bf99a3672049cc8ce9c94870d41176501ded1abbc1b229b9553401
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          olm.substitutesFor: openshift-gitops-operator.v1.11.7
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.9.21\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:4d22a7a6cc2f0442889512267ae3d9356f05ff0dfdbbe137c3d8ebebc24818c8
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:42f809f32f1d43186b849e3cc78063d3ebb2b54c55b2a1b77908af09a5ce044b
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:0b926fc9eae775425faa1ab9709a36fdac15afe4bfe523720e17ecf5df45bb28
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:0df37cf0a0e48efae02a0711c6da61c6165e7856db97097e98d56c8b6a137bb9
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:eade50b8321c1b2e0a8ac4dcb34596d43dea0a137b8b97b6eb68c6d469c8b200
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:20566173a8e907980e603d82fae4d96fcdfd5c43d9004c2e506cf0e09a47b91e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/rhel8/redis-6@sha256:c4a98010f7dd5faec2047bb9803061704233a8efc930be2a02386ac16d062d86
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c94c2167dc957856682a104b9a6187d3d321dec71ae794acc0dc37794d2cc373
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:7ea96000768b3cf7711150531c4c1c7edff2c45a2f29fe10bd9811ab6b281909
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d6ad2ee900bf99a3672049cc8ce9c94870d41176501ded1abbc1b229b9553401
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:959de6a95ef8978f66fdbe1033bcaa0ddeff9a376023e5d1d2698c5acb417305
        version: 1.11.7+0.1724840231.p
      entries:
      - name: openshift-gitops-operator.v1.11.7-0.1724840231.p
        version: 1.11.7+0.1724840231.p
      - name: openshift-gitops-operator.v1.11.7
        version: 1.11.7
      - name: openshift-gitops-operator.v1.11.6
        version: 1.11.6
      - name: openshift-gitops-operator.v1.11.5
        version: 1.11.5
      - name: openshift-gitops-operator.v1.11.4
        version: 1.11.4
      - name: openshift-gitops-operator.v1.11.3
        version: 1.11.3
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.11
    - currentCSV: openshift-gitops-operator.v1.12.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:48dd1254fd3a887aef2eae9239f68637eae18514a50f72ea504414d83dc727b8
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.10.18\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:dbf3038daf7c797123c511f9a609c3485cb1a5bcae53c614fad17e0a65fa7ee3
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a1d4e26ea755181818401697f18260b9e876aa441be14830f9142081ee22a24d
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:73918f3b3c507fc64006b7a500ecfa9d58151d694852f696c41a4412edffc78e
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:33bbb88343285a0510670384b87d0226dd6de8aa12c78b68921362c4d6f5a6aa
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:48dd1254fd3a887aef2eae9239f68637eae18514a50f72ea504414d83dc727b8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9e0fee30842a16d7b8a39d336123c80c5271cb2ceb50759fc7ee51a4ed09ceb3
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:06a6637c4776f4285bcb295d327ee60b1aed61b7adc5f7ab52898d79281ced70
        - registry.redhat.io/rhel8/redis-6@sha256:a848cc4edb4405846c673187dd68ce66c328dd1ad42f193f5737288c7971399d
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:b5736d1106db29fcdf42328ca20b96382d4339f90b97a07120ea73ee174ea1ad
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:7d82c719b1cef64607d30e8aaa5ab293955771c5b1dbd430e0befc988a43356d
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:1cf485a889c7586cb4944f000b53c0e97d08f97ba14e580bec57dad7cb13712e
        version: 1.12.6
      entries:
      - name: openshift-gitops-operator.v1.12.6
        version: 1.12.6
      - name: openshift-gitops-operator.v1.12.5-0.1727371523.p
        version: 1.12.5+0.1727371523.p
      - name: openshift-gitops-operator.v1.12.5
        version: 1.12.5
      - name: openshift-gitops-operator.v1.12.4
        version: 1.12.4
      - name: openshift-gitops-operator.v1.12.3
        version: 1.12.3
      - name: openshift-gitops-operator.v1.12.2
        version: 1.12.2
      - name: openshift-gitops-operator.v1.12.1
        version: 1.12.1
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.12
    - currentCSV: openshift-gitops-operator.v1.13.3-0.1741683398.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d41034bcf5c28bcb8afef08678793d0cc27a4032e2811a0c496c3bd0f10ecc17
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          olm.substitutesFor: openshift-gitops-operator.v1.13.3-0.1738142397.p
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.11.12\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_gitops/
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d41034bcf5c28bcb8afef08678793d0cc27a4032e2811a0c496c3bd0f10ecc17
        - registry.redhat.io/rhel8/redis-6@sha256:a848cc4edb4405846c673187dd68ce66c328dd1ad42f193f5737288c7971399d
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:b982df02ea7c45a7603d83fe82cac12e1faac303a1b72af6a727512bb97441df
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3f2804bc2af437ff1f06c214b48a1a780dfcfab995650d901c03270057eb9f3c
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:06a6637c4776f4285bcb295d327ee60b1aed61b7adc5f7ab52898d79281ced70
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a7c3e31ce414e6c6fecf1652343b22c52502902676bf4d7a75bf42bdabfca06
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:a366229e6ee26ebe2d7f765b2a2da6057a29e783835655b5d97545ac563b541a
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:d4d5724f67859d24c32f0f5735f5af3e325f77db97f58bc7f17ed4b831b7e534
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c02dc2c79ee24ce5450ea4e19493ac2a1997eeb5bfd359291642c529826d0cdd
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:e87515d016565bb69cb2c2ac40432e0127dd2cb20a490ac82019836032ba15e5
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:bde11fa83b7e708677b08ea995041f480b88edfb93e4a62c10380d101c0a349c
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        version: 1.13.3+0.1741683398.p
      entries:
      - name: openshift-gitops-operator.v1.13.3-0.1741683398.p
        version: 1.13.3+0.1741683398.p
      - name: openshift-gitops-operator.v1.13.3-0.1738142397.p
        version: 1.13.3+0.1738142397.p
      - name: openshift-gitops-operator.v1.13.3
        version: 1.13.3
      - name: openshift-gitops-operator.v1.13.2
        version: 1.13.2
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.13
    - currentCSV: openshift-gitops-operator.v1.14.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:7cf924af77d9c56e70710b99e080ed8aee162f05ec472f32085a76ef85cf9879
          createdAt: "2025-05-22 12:40:45"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.12.13-1\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_gitops/
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:2c45cfc01f9ec7577afc09cadc4c1a6f470836f6a0a7c24d66b791913f89bb05
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:42a79aad641e8607fc339e4ce7f9cf1d258be01f863dbc31a94d2c27edb645bd
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:97e32825f59c33cca7d8dba3580095205c20435ccd15b91cf0aff37db9841e0e
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:7cf924af77d9c56e70710b99e080ed8aee162f05ec472f32085a76ef85cf9879
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:b7d3c882fc25fbbea28951afb197d4d47ace253864889339396e20f0c6897a7f
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:bcf6b3136a5e71803cb518c8763c165bb46ee37ef7bf55503911d7fcb36b3a8f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c369df980874aae270ca7bc6455f1ca45467d75d25fe80d7bcb06a7826bdeaa3
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:cd4965d91c163160bcf43a02c54ca60b8704281c95de904f904df942bfd83bcd
        - registry.redhat.io/rhel8/redis-6@sha256:a9f30ed08abcab0122556a5485fbcf6e0fc03860f19954dcebdbf5e754876a45
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:7358787959af7e7e589edfc043a97d80ebab87536bde365d33be7a600071bf0e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1ed3b514bf9af6ac0d386e3739874d1d964e667d54e041c2d67edd769b0009fb
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:8096f824bd96e3148cf4b6936766c888d1fff0af9e9d0be440ae4feb6adc03f2
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:1112160243f227a5d393eeee165bd4fed0d3f42646a9280cbe9af60ef728aeb2
        version: 1.14.4
      entries:
      - name: openshift-gitops-operator.v1.14.4
        version: 1.14.4
      - name: openshift-gitops-operator.v1.14.3-0.1746016855.p
        version: 1.14.3+0.1746016855.p
      - name: openshift-gitops-operator.v1.14.3-0.1742836272.p
        version: 1.14.3+0.1742836272.p
      - name: openshift-gitops-operator.v1.14.3
        version: 1.14.3
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.14
    - currentCSV: openshift-gitops-operator.v1.15.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ad49a3d14f843bde0d2365093d00ffa6d3b9d59a75c81e0c56710ad37b53bd92
          createdAt: "2024-11-25T03:57:37Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v2.13.9\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:d41454824632475f0e88aa6e0031ec643bd37e1d575df36cd3ee63daf1403281
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:6cec76d7ea70d77923cf4285177e375d34e96e3aecbe77b68cd5ff769e6e540b
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:911733d78ae6f36f642b4d63f044cd102b7eee170508eaf518ac7acf0f339698
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ad49a3d14f843bde0d2365093d00ffa6d3b9d59a75c81e0c56710ad37b53bd92
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:41717c27915251bda8afeef651b609b8a9ed410e0fa09c4cf8fa7e91d6a94292
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:0c7f72546d0d918bfd21aff7157b983ff7c832b269f5efb84d2ed723a09bf612
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:5bda26e2c16beebf66dd8f3c3c55d48e2f00db870f2d9d4d44570d9d3b8bddb0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:a9e2360f00c5c9840a100a6748bd905b8f34169140a33fb0ca4af9cf6db71642
        version: 1.15.4
      entries:
      - name: openshift-gitops-operator.v1.15.4
        version: 1.15.4
      - name: openshift-gitops-operator.v1.15.3
        version: 1.15.3
      - name: openshift-gitops-operator.v1.15.2
        version: 1.15.2
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.15
    - currentCSV: openshift-gitops-operator.v1.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ead539d7cf54992cf1bfeae6a73a453059a7dffef44505991ff33319c8d40c50
          createdAt: "2025-03-07T13:13:07Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v2.14.16\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:58284a12174d19c2d04723fa1f551d96b490872e0fa2ad94af1b3e52dc6ad955
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a5e2a12ec5dea607a6651462d8e7920cc73f3272dba4041c397a14f2c73d7532
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:03d0d9093422396330aa8611adf77abdc9be759f3d2e23850e4abe778bd6002b
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:cab809fd1288d31c93f4f5c4facb7efaff9f844e21a859d1e4557831575a13c8
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c37aa141387c847eac825363dddda3eb59203d3694120dba439b00cf0c66f320
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:2598a3339bc920b91018d3f7acbe78ccbee684b5182c874e48296744d2373f9d
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ead539d7cf54992cf1bfeae6a73a453059a7dffef44505991ff33319c8d40c50
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:ceff1cde6d043daae599b578eec9759c6d49cd5cf770460b9d075fa879a16e81
        version: 1.16.3
      entries:
      - name: openshift-gitops-operator.v1.16.3
        version: 1.16.3
      - name: openshift-gitops-operator.v1.16.2
        version: 1.16.2
      - name: openshift-gitops-operator.v1.16.1
        version: 1.16.1
      - name: openshift-gitops-operator.v1.16.0-0.1746014725.p
        version: 1.16.0+0.1746014725.p
      - name: openshift-gitops-operator.v1.16.0
        version: 1.16.0
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.16
    - currentCSV: openshift-gitops-operator.v1.17.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:1b86eb4e071b1c526931b6d18c3d8f51522293699059c514488602c320271049
          createdAt: "2025-07-30T13:03:16Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.0.14\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:3fc4941fa4a6f2dbe9dd82622a0d3fb794ecd5a9dee286a38b60a7f6a947d685
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:cc45e212e7eb4e7e8daac6ea74071d5d98260de0b1f6d1cea0e7876111ccea8c
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:2f5f997bce924445de735ae0508dca1a7bba561bc4acdacf659928488233cb8a
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:5c9ea426cd60e7b8d1d8e4fe763909200612434c65596855334054e26cbfe3d0
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:70e5e3900d0c89999c69be1d58adf84c1b3664aacbffa9c5e0efc8ae1b6ef9c1
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:1b86eb4e071b1c526931b6d18c3d8f51522293699059c514488602c320271049
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:d73b4f5b355f556208751281923bb8b43edffc061a673c04d53b88f0e50ecb26
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:a2ae66772d90e734bb3385a59852e6016325a21b916a2580a33cf0da7cb0b527
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:c0a513ed54b631629efafe6e94b775d5bd0a943c44fbe96f02a0d5bb1f989099
        version: 1.17.1
      entries:
      - name: openshift-gitops-operator.v1.17.1
        version: 1.17.1
      - name: openshift-gitops-operator.v1.17.0
        version: 1.17.0
      name: gitops-1.17
    - currentCSV: openshift-gitops-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
          createdAt: "2025-08-21T01:20:45Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.18.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: NamespaceManagement
            name: namespacemanagements.argoproj.io
            version: v1beta1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.1.6\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/rhel9/redis-7@sha256:b4f58ad83ab8719a2e102f38ee526546cd1d56428901afc3add45c9a0e2a2f12
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:1eef9ecd58cf8391f29360ef74dd8aa5dcb48e1582d1500e445706d8302c951e
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a0544c14823492165550d83a6d8ba79dd632b46144d3fdcb543793726111d76
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:6284873bcf1593c210003e0292ac5583f03f03e80d937098d808d48fa0d01cf6
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:5a09fcba3d9e6fc462719e81a46af6022cf7f70a05bcb387ed959a49f6f37dbc
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:adca9c55b80d899d500bf0a7234cf61ae30ffed1b2c927ddea8d9fd07c9acb6a
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7f6ceae56c60fd432425c36f93a9e9bbbb96ef9422333ad6ea41ff3a2eaf8374
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:640e3298c7cd2faa84971b0d861446c4892ab8602108af05b8b5b049a4de624f
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c2283997d47298d49378e8200b149ea613c7735317ef12e74b7915a590f41ce2
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:08efd3e7a79c2945693c99ed31ec912223becff7238b1e28d1b1da5d63f93019
        version: 1.18.0
      entries:
      - name: openshift-gitops-operator.v1.18.0
        version: 1.18.0
      name: gitops-1.18
    - currentCSV: openshift-gitops-operator.v1.6.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:104cc7564fcc3261faa3fbf9c95720b4434e15103b876c080ed012b244ef1dae
          createdAt: "2021-09-15T22:31:21Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD v2.4.28\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces. \n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc.
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:104cc7564fcc3261faa3fbf9c95720b4434e15103b876c080ed012b244ef1dae
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:0cc4fa704e581cd5581c816668183f602fae6ef9ab3c04ce0d99d561287ef140
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:06f1150114d4c0bd236d2a5d4716c9c3e95a9733058297790389d6ba2a1bf264
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:e9e2f950cda8153cff112e43872b98e4977d3cbdf59713b8de92defef6d34b6d
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:dbcad68b70ab7918e944099aef4177c9bc3378e9ca49499dbb94a175927d3209
        - registry.redhat.io/rhel8/redis-6@sha256:53598a6effeb90e4f1b005b2521beffd2fa2b0c52d0e7f2347ee2abd2577cab3
        version: 1.6.6
      entries:
      - name: openshift-gitops-operator.v1.6.6
        version: 1.6.6
      - name: openshift-gitops-operator.v1.6.5
        version: 1.6.5
      - name: openshift-gitops-operator.v1.6.4
        version: 1.6.4
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.6
    - currentCSV: openshift-gitops-operator.v1.7.4-0.1690486082.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:e0967a4058fe5a9bfe7e998a3347529c99536cf3a44de9e40f9632035ff5a2fe
          createdAt: "2021-09-15T22:31:21Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.substitutesFor: openshift-gitops-operator.v1.7.4
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD v2.5.16\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficient privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces. \n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/rhel8/redis-6@sha256:53598a6effeb90e4f1b005b2521beffd2fa2b0c52d0e7f2347ee2abd2577cab3
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:e0967a4058fe5a9bfe7e998a3347529c99536cf3a44de9e40f9632035ff5a2fe
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c55aadf45397083bec4e42d6c0514ed79169a07fd3d9d75671904df2c0520e52
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:45789c49d9334e0ed062641313714c101bd0117dc5633316b1ec63e7a6c76c15
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:26050fb95978e90e8d00e895032045e77cb236008a0e1417459e1aeca8a48c69
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:5c11be8ec9db54331893f2f298aa27006e77b0eaded2d88d0531c77da91a421f
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:f1371d2012796e9aaca105112190ff89ec3d0be95968e05b19db101480f67975
        version: 1.7.4+0.1690486082.p
      entries:
      - name: openshift-gitops-operator.v1.7.4-0.1690486082.p
        version: 1.7.4+0.1690486082.p
      - name: openshift-gitops-operator.v1.7.4
        version: 1.7.4
      - name: openshift-gitops-operator.v1.7.3
        version: 1.7.3
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.7
    - currentCSV: openshift-gitops-operator.v1.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f32df77b3c23d6a833ff887386285ca7a5bd436641f0ec1ffc934c8247fa315f
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.skipRange: ""
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.6.153\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:8cc59901689858e06f5c2942f8c975449c17454fa8364da6153d9b5a3538a985
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:f77594bc053be144b33ff9603ee9675c7e82946ec0dbfb04d8f942c8d73155da
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f32df77b3c23d6a833ff887386285ca7a5bd436641f0ec1ffc934c8247fa315f
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:8aa68fddef88921aeeede24a849d3dabbceff01dc03185342c21e6ee58033eaf
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:16d9184345e6d5a19f326a7b52f76d8a59b96763f233c50546ff017a625e5ff5
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:0d13c03048b4435d9f5045a6fb7bd4b8808461b26aeca60c0047fa5937e1d45d
        version: 1.8.6
      entries:
      - name: openshift-gitops-operator.v1.8.6
        version: 1.8.6
      - name: openshift-gitops-operator.v1.8.5
        version: 1.8.5
      - name: openshift-gitops-operator.v1.8.4-0.1693214052.p
        version: 1.8.4+0.1693214052.p
      - name: openshift-gitops-operator.v1.8.4
        version: 1.8.4
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.8
    - currentCSV: openshift-gitops-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "argo-rollout"
                },
                "spec": null
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:11a6084e11aa32511a34bd6cfb4f4995aec690c1f5ed61861e99f6e42d62afb4
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.skipRange: ""
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.7.16\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:80f92842fe4cba008a873ad2ab6059a3a24e76e16671d8ece034dbbe666996bc
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:3d17786e2f4063154b9e3a40b96bf8ec0a14b86395cea5fe7ca18c5b43764478
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a568c8c647e2e2d6abdf12138eeef8f8bdde1fdcbf786e9ca8c8f93bc79acaf9
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:43bb961d847477902e66a92edd1539322cac8d5f74706b048a458ad037a0c7a3
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:11a6084e11aa32511a34bd6cfb4f4995aec690c1f5ed61861e99f6e42d62afb4
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ec9f60018694dcc5d431ba47d5536b761b71cb3f66684978fe6bb74c157679ac
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:3873d1f18b5643e1de2601419410270342650a21428bb26a4cbee20c0c63f98b
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:86012145d4b8798dc12f6f4f07a9316a193188aaac865e324b63f4e60e8ec584
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:136d3a936759e112d3ea2e78f7c02bd882b0fd3011f4c452f45c32c43822772d
        version: 1.9.4
      entries:
      - name: openshift-gitops-operator.v1.9.4
        version: 1.9.4
      - name: openshift-gitops-operator.v1.9.3
        version: 1.9.3
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.9
    - currentCSV: openshift-gitops-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
          createdAt: "2025-08-21T01:20:45Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.18.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: NamespaceManagement
            name: namespacemanagements.argoproj.io
            version: v1beta1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.1.6\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:adca9c55b80d899d500bf0a7234cf61ae30ffed1b2c927ddea8d9fd07c9acb6a
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7f6ceae56c60fd432425c36f93a9e9bbbb96ef9422333ad6ea41ff3a2eaf8374
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:6284873bcf1593c210003e0292ac5583f03f03e80d937098d808d48fa0d01cf6
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:08efd3e7a79c2945693c99ed31ec912223becff7238b1e28d1b1da5d63f93019
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:640e3298c7cd2faa84971b0d861446c4892ab8602108af05b8b5b049a4de624f
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c2283997d47298d49378e8200b149ea613c7735317ef12e74b7915a590f41ce2
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:5a09fcba3d9e6fc462719e81a46af6022cf7f70a05bcb387ed959a49f6f37dbc
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:1eef9ecd58cf8391f29360ef74dd8aa5dcb48e1582d1500e445706d8302c951e
        - registry.redhat.io/rhel9/redis-7@sha256:b4f58ad83ab8719a2e102f38ee526546cd1d56428901afc3add45c9a0e2a2f12
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a0544c14823492165550d83a6d8ba79dd632b46144d3fdcb543793726111d76
        version: 1.18.0
      entries:
      - name: openshift-gitops-operator.v1.18.0
        version: 1.18.0
      - name: openshift-gitops-operator.v1.17.1
        version: 1.17.1
      - name: openshift-gitops-operator.v1.17.0
        version: 1.17.0
      - name: openshift-gitops-operator.v1.16.2
        version: 1.16.2
      - name: openshift-gitops-operator.v1.16.1
        version: 1.16.1
      - name: openshift-gitops-operator.v1.16.0-0.1746014725.p
        version: 1.16.0+0.1746014725.p
      - name: openshift-gitops-operator.v1.16.0
        version: 1.16.0
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: latest
    defaultChannel: latest
    packageName: openshift-gitops-operator
    provider:
      name: Red Hat Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: node-maintenance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-maintenance-operator.v0.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-maintenance-operator:v0.17.0
          createdAt: "2024-03-13 18:56:47"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=0.12.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/medik8s/node-maintenance-operator:v0.17.0
        version: 0.17.0
      entries:
      - name: node-maintenance-operator.v0.17.0
        version: 0.17.0
      - name: node-maintenance-operator.v0.16.1
        version: 0.16.1
      - name: node-maintenance-operator.v0.16.0
        version: 0.16.0
      - name: node-maintenance-operator.v0.15.0
        version: 0.15.0
      - name: node-maintenance-operator.v0.14.0
        version: 0.14.0
      - name: node-maintenance-operator.v0.13.0
        version: 0.13.0
      - name: node-maintenance-operator.v0.12.1
        version: 0.12.1
      name: stable
    defaultChannel: stable
    packageName: node-maintenance-operator
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudwatch-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudwatch-controller.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudwatch-controller:1.3.0
          createdAt: "2025-10-02T23:05:17Z"
          description: AWS CloudWatch controller is a service controller for managing
            CloudWatch resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Dashboard represents the state of an AWS cloudwatch Dashboard
              resource.
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.cloudwatch.services.k8s.aws
            version: v1alpha1
          - description: MetricAlarm represents the state of an AWS cloudwatch MetricAlarm
              resource.
            displayName: MetricAlarm
            kind: MetricAlarm
            name: metricalarms.cloudwatch.services.k8s.aws
            version: v1alpha1
          - description: MetricStream represents the state of an AWS cloudwatch MetricStream
              resource.
            displayName: MetricStream
            kind: MetricStream
            name: metricstreams.cloudwatch.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudWatch resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudWatch**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudWatch
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudwatch
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudWatch Developer Resources
          url: https://aws.amazon.com/CloudWatch/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudwatch maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudwatch-controller:1.3.0
        version: 1.3.0
      entries:
      - name: ack-cloudwatch-controller.v1.3.0
        version: 1.3.0
      - name: ack-cloudwatch-controller.v1.2.1
        version: 1.2.1
      - name: ack-cloudwatch-controller.v1.2.0
        version: 1.2.0
      - name: ack-cloudwatch-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudwatch-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudwatch-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudwatch-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudwatch-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudwatch-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudwatch-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudwatch-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudwatch-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudwatch-controller.v1.0.5
        version: 1.0.5
      - name: ack-cloudwatch-controller.v1.0.4
        version: 1.0.4
      - name: ack-cloudwatch-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudwatch-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudwatch-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudwatch-controller.v0.0.12
        version: 0.0.12
      - name: ack-cloudwatch-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudwatch-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudwatch-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudwatch-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudwatch-controller.v0.0.3
        version: 0.0.3
      - name: ack-cloudwatch-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudwatch-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OpenShift Database Access
      provider-url: ""
    name: dbaas-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dbaas-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSConnection",
                "metadata": {
                  "name": "dbaasconnection-example"
                },
                "spec": {
                  "databaseServiceID": "instance-example",
                  "databaseServiceType": "instance",
                  "inventoryRef": {
                    "name": "inventory-example",
                    "namespace": "openshift-dbaas-operator"
                  }
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSInstance",
                "metadata": {
                  "name": "dbaasinstance-sample"
                },
                "spec": {
                  "cloudProvider": "AWS",
                  "cloudRegion": "US_EAST_1",
                  "inventoryRef": {
                    "name": "inventory-example",
                    "namespace": "openshift-dbaas-operator"
                  },
                  "name": "my-cluster-free"
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSInventory",
                "metadata": {
                  "name": "inventory-example"
                },
                "spec": {
                  "credentialsRef": {
                    "name": "dbaas-vendor-credentials"
                  },
                  "providerRef": {
                    "name": "provider-example-registration"
                  }
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSPolicy",
                "metadata": {
                  "name": "dbaaspolicy-sample"
                },
                "spec": {
                  "disableProvisions": false
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSProvider",
                "metadata": {
                  "labels": {
                    "related-to": "dbaas-operator",
                    "type": "dbaas-provider-registration"
                  },
                  "name": "provider-example-registration"
                },
                "spec": {
                  "allowsFreeTrial": true,
                  "connectionKind": "ProviderConnection",
                  "credentialFields": [
                    {
                      "displayName": "ProviderAPIKeyD",
                      "helpText": "This is the API Key for the example provider to show on UI",
                      "key": "APIKey",
                      "required": true,
                      "type": "maskedstring"
                    }
                  ],
                  "externalProvisionDescription": "Follow the guide to start a free Provider Serverless (beta) cluster",
                  "externalProvisionURL": "https://www.exmample.com/docs/provider/quickstart.html",
                  "groupVersion": "dbaas.redhat.com/v1beta1",
                  "instanceKind": "ProviderInstance",
                  "inventoryKind": "ProviderInventory",
                  "provider": {
                    "displayDescription": "This is an example for providers on how to implement their operator to integrate with DBaaS.",
                    "displayName": "DBaaS Provider Example ",
                    "icon": {
                      "base64data": "SGVsbG8sIHdvcmxkLg==",
                      "mediatype": "image/png"
                    },
                    "name": "Provider Example"
                  },
                  "provisioningParameters": {
                    "cloudProvider": {
                      "conditionalData": [
                        {
                          "defaultValue": "GCP",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "FREETRIAL"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        },
                        {
                          "defaultValue": "AWS",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "SERVERLESS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Amazon Web Services",
                              "value": "AWS"
                            },
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        },
                        {
                          "defaultValue": "AWS",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Amazon Web Services",
                              "value": "AWS"
                            },
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        }
                      ],
                      "displayName": "Cloud Provider"
                    },
                    "dedicatedLocationLabel": {
                      "displayName": "Select regions \u0026 nodes",
                      "helpText": "Select the geographical region where you want the database instance to run, and set the number of nodes you want running in this dedicated cluster."
                    },
                    "hardwareLabel": {
                      "displayName": "Hardware per node",
                      "helpText": "Select the compute and storage requirements for this database instance."
                    },
                    "machineType": {
                      "conditionalData": [
                        {
                          "defaultValue": "m5.large",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "AWS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "2 vCPU, 8 GiB RAM",
                              "value": "m5.large"
                            },
                            {
                              "displayValue": "4 vCPU, 16 GiB RAM",
                              "value": "m5.xlarge"
                            },
                            {
                              "displayValue": "8 vCPU, 32 GiB RAM",
                              "value": "m5.2xlarge"
                            },
                            {
                              "displayValue": "16 vCPU, 64 GiB RAM",
                              "value": "m5.4xlarge"
                            },
                            {
                              "displayValue": "32 vCPU, 128 GiB RAM",
                              "value": "m5.8xlarge"
                            }
                          ]
                        },
                        {
                          "defaultValue": "n1-standard-2",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "GCP"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "2 vCPU, 7.5 GiB RAM",
                              "value": "n1-standard-2"
                            },
                            {
                              "displayValue": "4 vCPU, 15 GiB RAM",
                              "value": "n1-standard-4"
                            },
                            {
                              "displayValue": "8 vCPU, 30 GiB RAM",
                              "value": "n1-standard-8"
                            },
                            {
                              "displayValue": "16 vCPU, 60 GiB RAM",
                              "value": "n1-standard-16"
                            },
                            {
                              "displayValue": "32 vCPU, 120 GiB RAM",
                              "value": "n1-standard-32"
                            }
                          ]
                        }
                      ],
                      "displayName": "Compute"
                    },
                    "name": {
                      "displayName": "Cluster name"
                    },
                    "nodes": {
                      "displayName": "nodes"
                    },
                    "plan": {
                      "conditionalData": [
                        {
                          "defaultValue": "SERVERLESS",
                          "options": [
                            {
                              "displayValue": "Free trial",
                              "value": "FREETRIAL"
                            },
                            {
                              "displayValue": "Serverless",
                              "value": "SERVERLESS"
                            },
                            {
                              "displayValue": "Dedicated",
                              "value": "DEDICATED"
                            }
                          ]
                        }
                      ],
                      "displayName": "Hosting plan"
                    },
                    "planLabel": {
                      "displayName": "Select a plan"
                    },
                    "regions": {
                      "displayName": "us-west-2"
                    },
                    "serverlessLocationLabel": {
                      "displayName": "Select regions",
                      "helpText": "Select the geographical region where you want the database instance to run."
                    },
                    "spendLimit": {
                      "conditionalData": [
                        {
                          "defaultValue": "0",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "SERVERLESS"
                            }
                          ]
                        }
                      ],
                      "displayName": "Spend limit"
                    },
                    "spendLimitLabel": {
                      "displayName": "Spend limit",
                      "helpText": "Set a spending limit on resources for this database instance."
                    },
                    "storageGib": {
                      "conditionalData": [
                        {
                          "defaultValue": "15",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "AWS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "15 GiB",
                              "value": "15"
                            },
                            {
                              "displayValue": "35 GiB",
                              "value": "35"
                            },
                            {
                              "displayValue": "75 GiB",
                              "value": "75"
                            },
                            {
                              "displayValue": "150 GiB",
                              "value": "150"
                            },
                            {
                              "displayValue": "300 GiB",
                              "value": "300"
                            },
                            {
                              "displayValue": "600 GiB",
                              "value": "600"
                            }
                          ]
                        },
                        {
                          "defaultValue": "15",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "GCP"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "15 GiB",
                              "value": "15"
                            },
                            {
                              "displayValue": "35 GiB",
                              "value": "35"
                            },
                            {
                              "displayValue": "75 GiB",
                              "value": "75"
                            },
                            {
                              "displayValue": "150 GiB",
                              "value": "150"
                            },
                            {
                              "displayValue": "300 GiB",
                              "value": "300"
                            },
                            {
                              "displayValue": "600 GiB",
                              "value": "600"
                            }
                          ]
                        }
                      ],
                      "displayName": "Storage"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: quay.io/ecosystem-appeng/dbaas-operator:v0.5.0
          description: Discover & connect with database instances hosted on 3rd-party
            ISV cloud platforms.
          operatorframework.io/suggested-namespace: openshift-dbaas-operator
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/internal-objects: '["dbaasconnections.dbaas.redhat.com","dbaasproviders.dbaas.redhat.com","dbaasplatforms.dbaas.redhat.com","dbaasinstances.dbaas.redhat.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/RHEcosystemAppEng/dbaas-operator
          support: rhoda-requests@redhat.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DBaaSConnection
            name: dbaasconnections.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSConnection defines the schema for the DBaaSConnection
              API.
            displayName: DBaaSConnection
            kind: DBaaSConnection
            name: dbaasconnections.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSInstance
            name: dbaasinstances.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSInstance defines the schema for the DBaaSInstance API.
            displayName: DBaaSInstance
            kind: DBaaSInstance
            name: dbaasinstances.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSInventory
            name: dbaasinventories.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSInventory defines the schema for the DBaaSInventory
              API. Inventory objects must be created in a valid namespace, determined
              by the existence of a DBaaSPolicy object.
            displayName: Provider Account
            kind: DBaaSInventory
            name: dbaasinventories.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSPlatform
            name: dbaasplatforms.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSPlatform defines the schema for the DBaaSPlatform API.
            displayName: DBaaSPlatform
            kind: DBaaSPlatform
            name: dbaasplatforms.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSPolicy
            name: dbaaspolicies.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSPolicy enables administrative capabilities within a
              namespace, and sets a default inventory policy. Policy defaults can
              be overridden on a per-inventory basis.
            displayName: Provider Account Policy
            kind: DBaaSPolicy
            name: dbaaspolicies.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSProvider
            name: dbaasproviders.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSProvider defines the schema for the DBaaSProvider API.
            displayName: DBaaSProvider
            kind: DBaaSProvider
            name: dbaasproviders.dbaas.redhat.com
            version: v1beta1
        description: |
          The OpenShift Database Access Operator enables OpenShift users to discover & connect with database instances
          hosted on 3rd-party ISV cloud platforms such as CrunchyData Bridge & CockroachCloud.

          ## Feedback & Support

          Please contact us at [rhoda-requests@redhat.com](mailto:rhoda-requests@redhat.com)

          For more information on how to use our service, please refer to our [Quick Starts](quickstart?keyword=database).
        displayName: OpenShift Database Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rhoda
        - oda
        - dbaas
        - database
        - db
        - data
        - access
        - crunchy
        - bridge
        - cockroach
        - cockroachdb
        - cloud
        - ack
        - amazon
        - rds
        - relational
        links:
        - name: GitHub
          url: https://github.com/RHEcosystemAppEng/dbaas-operator
        maintainers:
        - email: ecosystem-appeng@redhat.com
          name: Red Hat Ecosystem App Engineering
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: OpenShift Database Access
        relatedImages:
        - registry.developers.crunchydata.com/crunchydata/crunchy-bridge-operator-catalog:v0.0.6
        - gcr.io/cockroach-shared/ccapi-k8s-operator-catalog:v0.0.5
        - quay.io/ecosystem-appeng/dbaas-operator:v0.5.0
        - quay.io/openshift/origin-kube-rbac-proxy:4.10.0
        - quay.io/ecosystem-appeng/rds-dbaas-operator-catalog:v0.4.0
        - quay.io/ecosystem-appeng/dbaas-dynamic-plugin:0.5.0
        version: 0.5.0
      entries:
      - name: dbaas-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: dbaas-operator
    provider:
      name: OpenShift Database Access
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ControlPlane Ltd.
      provider-url: https://control-plane.io
    name: flux-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux-operator.v0.31.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "FluxInstance",
                "metadata": {
                  "name": "flux",
                  "namespace": "flux-system",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcileEvery": "1h",
                    "fluxcd.controlplane.io/reconcileTimeout": "5m"
                  }
                },
                "spec": {
                  "distribution": {
                    "version": "2.x",
                    "registry": "ghcr.io/fluxcd"
                  },
                  "components": [
                    "source-controller",
                    "kustomize-controller",
                    "helm-controller",
                    "notification-controller",
                    "image-reflector-controller",
                    "image-automation-controller"
                  ],
                  "cluster": {
                    "type": "openshift",
                    "size": "medium",
                    "domain": "cluster.local",
                    "networkPolicy": true,
                    "multitenant": true,
                    "tenantDefaultServiceAccount": "flux"
                  }
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "FluxReport",
                "metadata": {
                  "name": "flux",
                  "namespace": "flux-system"
                },
                "spec": {
                  "distribution": {
                    "entitlement": "Issued by controlplane",
                    "managedBy": "flux-operator",
                    "status": "Installed",
                    "version": "v2.3.0"
                  },
                "components": [
                    {
                      "image": "ghcr.io/fluxcd/kustomize-controller:v1.3.0@sha256:48a032574dd45c39750ba0f1488e6f1ae36756a38f40976a6b7a588d83acefc1",
                      "name": "kustomize-controller",
                      "ready": true,
                      "status": "Current Deployment is available. Replicas: 1"
                    },
                    {
                      "image": "ghcr.io/fluxcd/source-controller:v1.3.0@sha256:161da425b16b64dda4b3cec2ba0f8d7442973aba29bb446db3b340626181a0bc",
                      "name": "source-controller",
                      "ready": true,
                      "status": "Current Deployment is available. Replicas: 1"
                    }
                  ],
                "reconcilers": [
                    {
                      "apiVersion": "kustomize.toolkit.fluxcd.io/v1",
                      "kind": "Kustomization",
                      "stats": {
                        "failing": 0,
                        "running": 1,
                        "suspended": 0
                      }
                    },
                    {
                      "apiVersion": "source.toolkit.fluxcd.io/v1",
                      "kind": "GitRepository",
                      "stats": {
                        "failing": 0,
                        "running": 1,
                        "suspended": 0,
                        "totalSize": "3.7 MiB"
                      }
                    }
                  ],
                  "sync": {
                    "ready": true,
                    "id": "kustomization/flux-system",
                    "path": "clusters/production",
                    "source": "https://github.com/my-org/my-fleet.git",
                    "status": "Applied revision: refs/heads/main@sha1:a90cd1ac35de01c175f7199315d3f4cd60195911"
                  }
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "ResourceSet",
                "metadata": {
                  "name": "podinfo",
                  "namespace": "default",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcile": "enabled",
                    "fluxcd.controlplane.io/reconcileEvery": "30m",
                    "fluxcd.controlplane.io/reconcileTimeout": "5m"
                  }
                },
                "spec": {
                  "commonMetadata": {
                    "labels": {
                      "app.kubernetes.io/name": "podinfo"
                    }
                  },
                  "inputs": [
                    {
                      "tenant": "team1",
                      "version": "6.7.x",
                      "replicas": "2"
                    },
                    {
                      "tenant": "team2",
                      "version": "6.6.x",
                      "replicas": "3"
                    }
                  ],
                  "resources": [
                    {
                      "apiVersion": "source.toolkit.fluxcd.io/v1beta2",
                      "kind": "OCIRepository",
                      "metadata": {
                        "name": "podinfo-<< inputs.tenant >>",
                        "namespace": "default"
                      },
                      "spec": {
                        "interval": "10m",
                        "url": "oci://ghcr.io/stefanprodan/charts/podinfo",
                        "ref": {
                          "semver": "<< inputs.version | quote >>"
                        }
                      }
                    },
                    {
                      "apiVersion": "helm.toolkit.fluxcd.io/v2",
                      "kind": "HelmRelease",
                      "metadata": {
                        "name": "podinfo-<< inputs.tenant >>",
                        "namespace": "default"
                      },
                      "spec": {
                        "interval": "1h",
                        "releaseName": "podinfo-<< inputs.tenant >>",
                        "chartRef": {
                          "kind": "OCIRepository",
                          "name": "podinfo-<< inputs.tenant >>"
                        },
                        "values": {
                          "replicaCount": "<< inputs.replicas | int >>"
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "ResourceSetInputProvider",
                "metadata": {
                  "name": "demo",
                  "namespace": "default",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcile": "enabled",
                    "fluxcd.controlplane.io/reconcileEvery": "10m",
                    "fluxcd.controlplane.io/reconcileTimeout": "1m"
                  }
                },
                "spec": {
                  "type": "GitHubPullRequest",
                  "url": "https://github.com/fluxcd-testing/pr-testing",
                  "filter": {
                    "limit": 10,
                    "includeBranch": "^stefanprodan-patch-.*$",
                    "labels": [
                      "enhancement"
                    ]
                  },
                  "skip": {
                    "labels": [
                        "deploy/flux-preview-pause",
                        "!test-build-push/passed"
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/controlplaneio-fluxcd/flux-operator:v0.31.0-ubi
          createdAt: "2025-10-06T20:33:03Z"
          description: The Flux Operator manages the lifecycle of the CNCF Flux project.
          operatorframework.io/suggested-namespace: flux-system
          repository: https://github.com/controlplaneio-fluxcd/flux-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Flux Instance
            displayName: FluxInstance
            kind: FluxInstance
            name: fluxinstances.fluxcd.controlplane.io
            version: v1
          - description: Flux Report (Autogenerated)
            displayName: FluxReport
            kind: FluxReport
            name: fluxreports.fluxcd.controlplane.io
            version: v1
          - description: Resource Set
            displayName: ResourceSet
            kind: ResourceSet
            name: resourcesets.fluxcd.controlplane.io
            version: v1
          - description: ResourceSet Input Provider
            displayName: ResourceSetInputProvider
            kind: ResourceSetInputProvider
            name: resourcesetinputproviders.fluxcd.controlplane.io
            version: v1
        description: |-
          [Flux](https://fluxcd.io) is a leading GitOps Continuous Delivery tool used to streamline and automate application deployments on Kubernetes. Flux enables teams to achieve a reliable and auditable delivery process while promoting collaboration and traceability across different environments.
          Flux is powered by the GitOps Toolkit, a set of composable APIs and specialized tools that enable a wide range of continuous delivery use-cases, from simple Kubernetes deployment pipelines to multi-tenant and multi-cluster progressive delivery rollouts.
          For more information about the CNCF-graduated Flux project, please see the [Flux architecture overview](https://fluxcd.control-plane.io/guides/flux-architecture/).
          ### Operator Capabilities
          The [Flux Operator](https://github.com/controlplaneio-fluxcd/flux-operator) provides a declarative API for the installation, configuration and upgrade of CNCF Flux and the ControlPlane [enterprise distribution](https://fluxcd.control-plane.io/).
          The operator allows the configuration of Flux multi-tenancy lockdown, network policies, persistent storage, sharding, vertical scaling, custom patches, and the synchronization of the cluster state from Git repositories, OCI artifacts and S3-compatible storage.
          The Flux Operator supervises the Flux controllers and provides a unified view of all the Flux resources that define the GitOps workflows for the target cluster. The operator generates reports, emits events, and exports Prometheus metrics to help with monitoring and troubleshooting Flux.
          ### Operator Features
          The operator extends Flux with self-service capabilities, deployment windows, and preview environments for GitHub, GitLab and Azure DevOps pull requests testing.
          The operator ResourceSet API enables platform teams to define their own application standard as a group of Flux and Kubernetes resources that can be templated, parameterized and deployed as a single unit on self-service environments. For more information, please see the [ResourceSet documentation](https://fluxcd.control-plane.io/operator/resourcesets/introduction/).
          ### OpenShift Support
          The Flux Operator should be installed in a dedicated namespace, e.g. `flux-system`.
          To deploy Flux on OpenShift clusters, create a `FluxInstance` custom resource named `flux` in the same namespace as the operator and set the `.spec.cluster.type` field to `openshift`.
          For more information on how to configure the Flux instance, please see the [Flux Operator documentation](https://fluxcd.control-plane.io/operator/flux-config/).
        displayName: Flux Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - fluxcd
        - gitops
        - continuous delivery
        links:
        - name: Source Code
          url: https://github.com/controlplaneio-fluxcd/flux-operator
        - name: Documentation
          url: https://fluxcd.control-plane.io/operator/
        - name: Enterprise Support
          url: https://fluxcd.control-plane.io/pricing/
        maintainers:
        - email: flux-enterprise@control-plane.io
          name: ControlPlane Flux Team
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: ControlPlane Ltd.
          url: https://control-plane.io
        relatedImages:
        - ghcr.io/controlplaneio-fluxcd/flux-operator:v0.31.0-ubi
        version: 0.31.0
      entries:
      - name: flux-operator.v0.31.0
        version: 0.31.0
      - name: flux-operator.v0.30.0
        version: 0.30.0
      - name: flux-operator.v0.28.0
        version: 0.28.0
      - name: flux-operator.v0.27.0
        version: 0.27.0
      - name: flux-operator.v0.26.0
        version: 0.26.0
      - name: flux-operator.v0.24.1
        version: 0.24.1
      - name: flux-operator.v0.24.0
        version: 0.24.0
      - name: flux-operator.v0.23.0
        version: 0.23.0
      - name: flux-operator.v0.22.0
        version: 0.22.0
      - name: flux-operator.v0.21.0
        version: 0.21.0
      - name: flux-operator.v0.20.0
        version: 0.20.0
      - name: flux-operator.v0.19.0
        version: 0.19.0
      - name: flux-operator.v0.18.0
        version: 0.18.0
      name: stable
    defaultChannel: stable
    packageName: flux-operator
    provider:
      name: ControlPlane Ltd.
      url: https://control-plane.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-load-core
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-load-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Loadcore",
                "metadata": {
                  "name": "loadcore-sample"
                },
                "spec": {
                  "accessRoute": {
                    "enabled": true,
                    "host": "loadcore.apps.cluster.perf.keysight.lab"
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "reportTemplates": {
                    "enabled": false
                  },
                  "wirelessDataModel": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:42:44Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Loadcore
            name: loadcores.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight LoadCore Operator
        displayName: Keysight LoadCore Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Load Core
          url: https://keysight-load-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-load-core@sha256:245d9be60ff6e9a6e5e42a026a4ee7b8168f391a0491de23ae827d1906af88e5
        version: 0.1.21
      entries:
      - name: keysight-load-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-load-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CloudBees
      provider-url: https://www.cloudbees.com/products/continuous-integration
    name: cloudbees-ci-rhmp
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudbees-ci.v3.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.cloudbees.com/v1alpha1",
                "kind": "CloudBeesCI",
                "metadata": {
                  "name": "cloudbeesci-sample"
                },
                "spec": {
                  "Agents": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-agent:2.289.3.2"
                    },
                    "ImagePullSecrets": null,
                    "SeparateNamespace": {
                      "Create": false,
                      "Enabled": false,
                      "Name": null
                    }
                  },
                  "Hibernation": {
                    "Enabled": false,
                    "Image": {
                      "dockerImage": "cloudbees/managed-master-hibernation-monitor:247.c5dfce00a179",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "NodeSelector": {},
                    "Tolerations": []
                  },
                  "Master": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-mm:2.289.3.2"
                    },
                    "JavaOpts": null,
                    "OperationsCenterNamespace": null
                  },
                  "NetworkPolicy": {
                    "Enabled": false,
                    "JMXSelectors": [],
                    "ingressControllerSelector": []
                  },
                  "OperationsCenter": {
                    "AgentListenerPort": 50000,
                    "Annotations": {},
                    "CSRF": {
                      "ProxyCompatibility": false
                    },
                    "CasC": {
                      "ConfigMapName": "oc-casc-bundle",
                      "Enabled": false
                    },
                    "ContainerPort": 8080,
                    "ContextPath": null,
                    "Enabled": true,
                    "ExtraConfigMaps": [],
                    "ExtraContainers": [],
                    "ExtraGroovyConfiguration": {},
                    "ExtraVolumeMounts": [],
                    "ExtraVolumes": [],
                    "HealthProbeLivenessFailureThreshold": 12,
                    "HealthProbes": true,
                    "HostName": null,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-cloud-core-oc:2.289.3.2",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "Ingress": {
                      "Annotations": {
                        "kubernetes.io/tls-acme": "false"
                      },
                      "Class": "nginx",
                      "tls": {
                        "Enable": false,
                        "SecretName": null
                      }
                    },
                    "JavaOpts": null,
                    "JenkinsOpts": null,
                    "LoadBalancerIP": null,
                    "LoadBalancerSourceRanges": [
                      "0.0.0.0/0"
                    ],
                    "Name": "cjoc",
                    "NodeSelector": {},
                    "Platform": "standard",
                    "Protocol": "http",
                    "Resources": {
                      "Limits": {
                        "Cpu": 1,
                        "Memory": "2G"
                      },
                      "Requests": {
                        "Cpu": 1,
                        "Memory": "2G"
                      }
                    },
                    "Route": {
                      "tls": {
                        "CACertificate": null,
                        "Certificate": null,
                        "DestinationCACertificate": null,
                        "Enable": false,
                        "InsecureEdgeTerminationPolicy": "Redirect",
                        "Key": null,
                        "Termination": "edge"
                      }
                    },
                    "ServiceAgentListenerPort": 50000,
                    "ServiceAnnotations": {},
                    "ServicePort": 80,
                    "ServiceType": "ClusterIP",
                    "Tolerations": []
                  },
                  "Persistence": {
                    "AccessMode": "ReadWriteOnce",
                    "Annotations": {},
                    "Size": "20Gi",
                    "StorageClass": null
                  },
                  "PodSecurityPolicy": {
                    "Annotations": {},
                    "Enabled": false
                  },
                  "Subdomain": false,
                  "ingress-nginx": {
                    "Enabled": false,
                    "controller": {
                      "admissionWebhooks": {
                        "patch": {
                          "nodeSelector": {
                            "kubernetes.io/os": "linux"
                          }
                        }
                      },
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "nginx-ingress": {
                    "Enabled": false,
                    "controller": {
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "rbac": {
                    "agentsServiceAccountName": "jenkins-agents",
                    "hibernationMonitorServiceAccountName": "managed-master-hibernation-monitor",
                    "install": true,
                    "masterServiceAccountName": "jenkins",
                    "serviceAccountName": "cjoc"
                  },
                  "sidecarinjector": {
                    "Enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cloudbees-ci-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cloudbees-ci-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudBeesCI
            name: cloudbeescis.charts.cloudbees.com
            version: v1alpha1
        description: CloudBees CI
        displayName: CloudBees CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudbees
        - ci
        - enterprise
        - continuous
        - integration
        - jenkins
        - automation
        - pipeline
        - devops
        links:
        - name: CloudBees CI
          url: https://www.cloudbees.com/products/continuous-integration
        maintainers:
        - email: support@cloudbees.com
          name: Support
        - email: sales@cloudbees.com
          name: Sales
        maturity: alpha
        provider:
          name: CloudBees
          url: https://www.cloudbees.com/products/continuous-integration
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.marketplace.redhat.com/rhm/cloudbees/cloudbees-ci-operator@sha256:bd0bd72e9e7807b0fa307de8a93a56967221fe2dd181bac02383f0f9be9120e4
        version: 3.34.1
      entries:
      - name: cloudbees-ci.v3.34.1
        version: 3.34.1
      name: alpha
    defaultChannel: alpha
    packageName: cloudbees-ci-rhmp
    provider:
      name: CloudBees
      url: https://www.cloudbees.com/products/continuous-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhacs-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhacs-operator.v3.74.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
          createdAt: "2024-02-06T23:06:13.271846+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.73.0 < 3.74.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:2c0840f4573373a19eb8c9a26501138ef0e76efbd508fd69a6260e78194d86c8
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:0482d4464d6d39c01abab8df882604dd227c52b5420d36fe7414aaff8067cbec
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:71b9fdcaf4cc761b1499be8e6ccff6c463cc6b9856e9deb1f93baad45ea59eb0
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:74624389ac3261ed699b1f64e5717353da68895709d30a011bc8813b61a4a249
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:120b9891ee9db839ab0b2029a62e67ec4747c47c6ffe15af38e9035160f57b78
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:22bf00e1457b4a9e63901892aa4375a1d4a9d1b2c067c1f054d918a24b222b10
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:a108e9a7967001e0bdfed7124b01e9a3dbf1dec308444762973030aec57e1d5a
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:87eb62fe8aa7952d0d69dc13c4867ac2f51e821f56d149a9b588c826caec4c16
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d546f2d3f668ce2f20e594394ade0a86d808f6c8c88cadbba227083d4abac940
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:264102d4bca7d9bd384dc04e5bbff1a7b43343ea8e59cc79047cdda50a3f299f
        version: 3.74.9
      entries:
      - name: rhacs-operator.v3.74.9
        version: 3.74.9
      - name: rhacs-operator.v3.74.8
        version: 3.74.8
      - name: rhacs-operator.v3.74.7
        version: 3.74.7
      - name: rhacs-operator.v3.74.6
        version: 3.74.6
      - name: rhacs-operator.v3.74.5
        version: 3.74.5
      - name: rhacs-operator.v3.74.4
        version: 3.74.4
      - name: rhacs-operator.v3.74.3
        version: 3.74.3
      - name: rhacs-operator.v3.74.2
        version: 3.74.2
      - name: rhacs-operator.v3.74.1
        version: 3.74.1
      - name: rhacs-operator.v3.74.0
        version: 3.74.0
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: latest
    - currentCSV: rhacs-operator.v3.62.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:56905c2e7df0e29ca2b6301cd81b4f9bbe02b483374708f7f7da219a4b29428d
          createdAt: "2021-08-13T17:10:01.410173+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.61.0 < 3.62.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.5.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main page before deploying, or consult the RHACS
              documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create an cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead. If you want to deploy Secured Cluster Service in another namespace, additionally pass the `-n <namespace>` flag.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e30f7835cf52dbce9afdbaf55e5a93d8baf9c4b3a3ae488d1d42c1adba7f46af
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:56905c2e7df0e29ca2b6301cd81b4f9bbe02b483374708f7f7da219a4b29428d
        - registry.redhat.io/rh-acs/collector@sha256:98cfab15684506f1d0978eb701e8de478d91fc5bda3013ffd3b436aa95f07cc0
        - registry.redhat.io/rh-acs/main@sha256:9932304bb62b084ee1ae7e75bb21d6dfb737260c851f2d8daf05082f79230c66
        - registry.redhat.io/rh-acs/scanner@sha256:266a575c2547a3223890e8b73db0f14d69dc5a22217c7e4e3dc79644e6ab0258
        - registry.redhat.io/rh-acs/scanner-db@sha256:ff877f503e08ecca0ce7a7d8db760b2e4004e8e8ae2f27fb9b9c00a05b58c93a
        - registry.redhat.io/rh-acs/collector@sha256:558c0dabdc52efaee7bc889ef397ac2ac30f7c54562c2cc0070d0ef991a8b31b
        version: 3.62.1
      entries:
      - name: rhacs-operator.v3.62.1
        version: 3.62.1
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.62
    - currentCSV: rhacs-operator.v3.64.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fa478167f2f738f244bed6a06450679be73f8464c40f86acde4ffd390505692b
          createdAt: "2021-09-22T05:10:01.286657+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.63.0 < 3.64.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0cacd00c0a147cc8295138565ef2dc2ddbd9ad1fb25f6be817719a75ea4c6448
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fa478167f2f738f244bed6a06450679be73f8464c40f86acde4ffd390505692b
        - registry.redhat.io/rh-acs/scanner@sha256:8b7121bce8f9c7133d29245ebb265a4b8f9831a66a57f11b2d68b9b8b5cf79cc
        - registry.redhat.io/rh-acs/scanner-db@sha256:67e81bf896267b7acab8cb6755cde552173de4f0881c38eaca8e491bd926cc53
        - registry.redhat.io/rh-acs/collector@sha256:2b3a8bd3af62516a6714fe612e5fdc0433b91d873182e290d3ddc2b32ce291f0
        - registry.redhat.io/rh-acs/collector@sha256:3eb099da081cf208d46be5ccfa7b5e9ac0e2cad49f127fa4db91e49e3f26db42
        - registry.redhat.io/rh-acs/main@sha256:6aeb2223124540ac217c38b016dea99d770ab897dde4906a821d8776d91ea03d
        version: 3.64.2
      entries:
      - name: rhacs-operator.v3.64.2
        version: 3.64.2
      - name: rhacs-operator.v3.64.1
        version: 3.64.1
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.64
    - currentCSV: rhacs-operator.v3.65.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:229c7e5f967c70dfbc3cf638d521267ea6acf875b055499dfd30cb19c0b1f69c
          createdAt: "2021-09-21T13:26:28.711610+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.64.0 < 3.65.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/collector@sha256:3eb099da081cf208d46be5ccfa7b5e9ac0e2cad49f127fa4db91e49e3f26db42
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0cacd00c0a147cc8295138565ef2dc2ddbd9ad1fb25f6be817719a75ea4c6448
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:229c7e5f967c70dfbc3cf638d521267ea6acf875b055499dfd30cb19c0b1f69c
        - registry.redhat.io/rh-acs/main@sha256:e672ef9a1239065f8cf0f331757861fea082e971a9e9760d1e65795f399cbde1
        - registry.redhat.io/rh-acs/scanner@sha256:bfe9b2205bfca6aebed14827b4bda098bfad574e67ffa967ad339543062200fe
        - registry.redhat.io/rh-acs/scanner-db@sha256:20f258500ac2ba4e626df92dbcd193cf7ab8687ac251591be2442b8a46871835
        - registry.redhat.io/rh-acs/collector@sha256:2b3a8bd3af62516a6714fe612e5fdc0433b91d873182e290d3ddc2b32ce291f0
        version: 3.65.1
      entries:
      - name: rhacs-operator.v3.65.1
        version: 3.65.1
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.65
    - currentCSV: rhacs-operator.v3.66.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:10cdb0ff3e67b65737cde420536494bd84c2b7ccf65192bbed2eb43eaed66a41
          createdAt: "2021-10-20T16:38:38.091764+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.65.0 < 3.66.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bd74f0b95ff2881faf6eb4520feecd067b69a759f09bbfecd467447ce2a2a25e
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:10cdb0ff3e67b65737cde420536494bd84c2b7ccf65192bbed2eb43eaed66a41
        - registry.redhat.io/rh-acs/main@sha256:97230c01e76d90c7fe35ffeaa591be3d9175f350a8fd88599aded09a1edf930c
        - registry.redhat.io/rh-acs/scanner@sha256:323ee86d6a9cbeb32270c659c1154f3a62c36535c6ea894e4e29e93eb1f4a84f
        - registry.redhat.io/rh-acs/scanner-db@sha256:490b0bf49161e4bc6de2504726aaf360e0ca9a214aea6b886fff13153bdb263d
        - registry.redhat.io/rh-acs/collector@sha256:5673c27554a71502e06e08945a4424c6ce5da27f489854f3aa452e6d31240aee
        - registry.redhat.io/rh-acs/collector@sha256:d93e13ee1d1cc4b7f82e644a7d43e560aac42bd1e06d25b476cea9a4338bf9ce
        version: 3.66.1
      entries:
      - name: rhacs-operator.v3.66.1
        version: 3.66.1
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.66
    - currentCSV: rhacs-operator.v3.67.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:893ebcb2d0c96647e56bce3b16fe32851df59cd71f2c0fc8f6e66ca3a669bb70
          createdAt: "2021-12-16T12:54:10.117625+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.66.0 < 3.67.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/scanner@sha256:2860ed271da6064149bb53568c5ad60e129918d55a2f244f9ade4a0bea4c4933
        - registry.redhat.io/rh-acs/scanner-db@sha256:6f8d47277732eb56317d576c2d06fb1f539d46d16520e6ca34abec290392cef9
        - registry.redhat.io/rh-acs/collector@sha256:57c2dfa3dbc8368076291fbc627baa01dcb5ebc50f838e02ce7ac38a128d027b
        - registry.redhat.io/rh-acs/collector@sha256:e4648723ccbd43fa5acf005d9734bfc06d41e19e1be0bde839880e28fa361e16
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86e5fa1fa294987114be200890c2e516501e424aee0fb98ece25c95e7716295b
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:893ebcb2d0c96647e56bce3b16fe32851df59cd71f2c0fc8f6e66ca3a669bb70
        - registry.redhat.io/rh-acs/main@sha256:8db7630d094b661411ebb3f35248848d3147c735677d0922f7a04de9a41eacc9
        version: 3.67.2
      entries:
      - name: rhacs-operator.v3.67.2
        version: 3.67.2
      - name: rhacs-operator.v3.67.1
        version: 3.67.1
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.67
    - currentCSV: rhacs-operator.v3.68.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:195a694dc96096166fe82e81ecb3ceb493f96cfb5029112e576e44ab39b56abd
          createdAt: "2022-06-15T22:13:22.427644+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.67.0 < 3.68.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:195a694dc96096166fe82e81ecb3ceb493f96cfb5029112e576e44ab39b56abd
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:f0d2f5f11c2a96d34d6b4baf249cbf821fe064eade73c3f95c22f3c6981ac4d1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:105d10bd50a6062c998ed107dea388dd78f0caddb0b57713c8692b1d07042774
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:081f2198ad70bfb2ec66153b50b48b950b9a5550ba996de413c5a2aff4052783
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c9d0a5a2871c24c0b1058e419f2b090429363a319856397855412fd2c0037395
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:9b9b9146d9683d6da0ff2abbc8ea82b96d945b7a8c3740d72d8c018fb2b04fba
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:0bb76363a3a25916dd4493e177e9b2d36eac9633473b07fe0fffb853fd072da3
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:711293438bab35b17d2e1f162e7f74f667a6e16dd32e45122f917b6817fefb4e
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:b5fc3dcaa72bbc8dea005189c940e1cf2d00c94a6cf7676cf0e1790e1191d45c
        version: 3.68.2
      entries:
      - name: rhacs-operator.v3.68.2
        version: 3.68.2
      - name: rhacs-operator.v3.68.1
        version: 3.68.1
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.68
    - currentCSV: rhacs-operator.v3.69.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9b2beceb255124855ba1b352d16009b2f2cd422901a19c311f0783dab1fcc3a0
          createdAt: "2022-06-21T09:07:40.705944+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.68.0 < 3.69.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:08a7d8c892fad69c589b81ad8fbd21d36c8fa4078530dba361cfa9a8f9f05950
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:9ae03669cfdc0f4fe56b021d911edb17159c07757720ae30a1f22d7f7a897c07
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:f167e928740d1c614e26948ea58f49e569f6a4fddc438cd0abe52f60538432d0
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:867613178d491f416e57975843817251490c5fbca2fc69d8b55b169bea675279
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:8db5206532e95a7ad7cb23f389155d41757aaf5f445ddce92b008a06627d3fd5
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:77f5db63e17b140a8d6b84e3822fd77463035be9c0c832b55b4977696da13ccc
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:da6dd486a20fdba61ea12d2c77ea299d9795a9f22d9057d65ff992b1f5896e07
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:b254746ed3faa1f1cbcafdf1343a1957bf0214311b36f4e892be3000cc2aecf3
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9b2beceb255124855ba1b352d16009b2f2cd422901a19c311f0783dab1fcc3a0
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:2afb2823edc6f3434e18db60ee37381b5b79f51234c1a261b6d0f8ed74cc8b63
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:cf5629823f2dd583495bcbbcf3b83b4f29f855ace3c388f5cb7b24a6d34ed267
        version: 3.69.2
      entries:
      - name: rhacs-operator.v3.69.2
        version: 3.69.2
      - name: rhacs-operator.v3.69.1
        version: 3.69.1
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.69
    - currentCSV: rhacs-operator.v3.70.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fb903bf9e110ba03e9c63977ebaf707e6884dd8fd413bcd6b793a0ed795cdd0e
          createdAt: "2022-06-20T12:30:26.999648+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.69.0 < 3.70.1'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:dfeb239931f7a33ade5c829a102cabc01f5c348e1aa35a7ddc24757cfd549629
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:df8cfdaa40b50ff2f2670d97c2115b47ae4bea443ae2831caaa468dd62f3c75f
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fb903bf9e110ba03e9c63977ebaf707e6884dd8fd413bcd6b793a0ed795cdd0e
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:6c32b86ed8bc18e384706eb6188406e66910bf34c8869bd155a60c3525d0fea7
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:d2e827a209840b7197960c8115720972237532eb9cf9d6fee7a36107053408ce
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:3d7c8b77671740e324ee6c309a84bbaa4d925e6750979fe4e46c561bc79f1b4d
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:b47c96af1c9ffe4d8728da510a93efe21adc221225fe2b822cfcace6c6672ba2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:08a7d8c892fad69c589b81ad8fbd21d36c8fa4078530dba361cfa9a8f9f05950
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:ee55a233d2a8fd77d89ec2c69f357721c7debead6470a9f494b3cd6a001829c4
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:6b14e850cda9a54618847858292ea669abff599a40692e27fed0e337c4a60526
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:e71f2f2347dd46dec790405b1bb98d54a348a4583949f142f5703be9a790f96e
        version: 3.70.1
      entries:
      - name: rhacs-operator.v3.70.1
        version: 3.70.1
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.70
    - currentCSV: rhacs-operator.v3.71.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:469b6c1592226eb9a521131ef758a33e6a6ac7c96bc67fcc903cea30469c125e
          createdAt: "2022-12-07T08:31:36.777654+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.70.0 < 3.71.3'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4a0dc8c337299d7f6440ff75c2d5219d3ab68183f4f117f008ef065757dd5d4b
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:469b6c1592226eb9a521131ef758a33e6a6ac7c96bc67fcc903cea30469c125e
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:f16f92e074f7125b88e9e2033c7c12b44421145b93d040034fa4688626869af1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:c3ec88cd99908894d0e67153779062f90ac1b007980d66040a79b2be13798a10
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:4d8bac9b3857af6ba351c7768f70ef30ef7cb99ae4df60cb226b5a793353a052
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:7f279bb6e0f34a5208d8711ba4d7b62c78f08b683815ce33f6aa28c23c4246da
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:9bc25e81c21cda6d8b535d8ac67c1286fe57a1596001850e730f8f73bf371111
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:e282a8e5e606003ea344d44ac5ee7699563848f7160eeae183aa58ad740e0e94
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:cb2d0f470968c2b725b0d176d423cc1866fd56893344d077d28f2c16581d3614
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:df7b41999b542a477848823a1d0e7e5c29ae1b00fbac035007a5a1f33fac139f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:5fdfd3690af60bd5ecf04741cd1f73259491da501700b066120976960567f0db
        version: 3.71.3
      entries:
      - name: rhacs-operator.v3.71.3
        version: 3.71.3
      - name: rhacs-operator.v3.71.2
        version: 3.71.2
      - name: rhacs-operator.v3.71.1
        version: 3.71.1
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.71
    - currentCSV: rhacs-operator.v3.72.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:3ff61e1cc77b66a22c8a8c0ab306f0e122cd74e762005285aede23ebb6c2f6a2
          createdAt: "2023-02-28T17:02:08.443613+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.71.0 < 3.72.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:575b6c3e2914751df1668ffff83cbda187ce6e56f32214414cdf036b7a62bbd8
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d56a00b3e20e60b7169f5c35d03fa6a1bd3b9980dd3ebc9d89d759e1d6d20ad2
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:cd475a6c2356d63cfe736a11114de7fedfbd8fdcfecc56224a00b6c819103be6
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:42e5f6c4ba789f124ac3ebe7b6d1a607ede8e763a35888c97ccbd0259bec4a1c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:648b14e4158da2a85eedb46b3456fb502d680d52d375ad34bf45b4b94f8b2ef6
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:3ff61e1cc77b66a22c8a8c0ab306f0e122cd74e762005285aede23ebb6c2f6a2
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:cd0c205ee44eddcdf99b84e80fcf4403851d956efc50532b9d12013bafe44988
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:0b69595405e4e598183bbf4197ac5b7593724fa6162413f225aa01a542af18c9
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:622f30b7f591194d86a5bf95deede0225e66f7b624db197f6289a5fbbbe887a9
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:c07fae3616cbc681e4d31839abea6623b4bf4dade93273201c8db7ba0d635aad
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:cb5df858e0c5cae8f77b285e23daeb83062bd72570c90ed5ea3ce4a02d2f5f08
        version: 3.72.4
      entries:
      - name: rhacs-operator.v3.72.4
        version: 3.72.4
      - name: rhacs-operator.v3.72.3
        version: 3.72.3
      - name: rhacs-operator.v3.72.2
        version: 3.72.2
      - name: rhacs-operator.v3.72.1
        version: 3.72.1
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.72
    - currentCSV: rhacs-operator.v3.73.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a8486b66f977a639d77ef110492e5ecdde47be8acaefbaa1cbb5417085c56839
          createdAt: "2023-05-26T21:27:04.033276+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.72.0 < 3.73.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:7bf9415aa4b0746d11ca5a59d4179bc382a0fac4ded969f0c4294171fde27d58
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:e6687c12d070069f655c17832d94c619c18274a259ccc59715036f9908e162f7
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:ee2a9e4ab74f5a554d1e4ff148833a53bdaab70ac1e2668cc9ee2fceca290ba9
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:082bd6f4b40209e0591be001dea358f328cd86a08691cdceb7f71ea59d861d9d
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:4d0ead8969f120a629104a92a454842bb7103aec137deb6c8209b7db9c48a1b2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:648b14e4158da2a85eedb46b3456fb502d680d52d375ad34bf45b4b94f8b2ef6
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a8486b66f977a639d77ef110492e5ecdde47be8acaefbaa1cbb5417085c56839
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:e2c89f735969d438f75a89ff63c404a8ec4361a03fde48e880fee73006b9c669
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:e4a27d3ffb173c87728fb637ca7fde92a9e78ea9ee3edcecccc2a5e3d3748b8c
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:dce5db4f162dd40c09cad114cb802cd6d7c72d0c6986581f2e74c977493f75b0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:c020a80ff82bf4d9b5cec3eeac5de9990de76041e22c444e664a20e15c46ce3b
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:5df359156b01505ce43e5f14c3b08722856bdb86f3d0d9b4fba785a2e580f346
        version: 3.73.5
      entries:
      - name: rhacs-operator.v3.73.5
        version: 3.73.5
      - name: rhacs-operator.v3.73.4
        version: 3.73.4
      - name: rhacs-operator.v3.73.3
        version: 3.73.3
      - name: rhacs-operator.v3.73.2
        version: 3.73.2
      - name: rhacs-operator.v3.73.1
        version: 3.73.1
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.73
    - currentCSV: rhacs-operator.v3.74.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
          createdAt: "2024-02-06T23:06:13.271846+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.73.0 < 3.74.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d546f2d3f668ce2f20e594394ade0a86d808f6c8c88cadbba227083d4abac940
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:264102d4bca7d9bd384dc04e5bbff1a7b43343ea8e59cc79047cdda50a3f299f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:0482d4464d6d39c01abab8df882604dd227c52b5420d36fe7414aaff8067cbec
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:2c0840f4573373a19eb8c9a26501138ef0e76efbd508fd69a6260e78194d86c8
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:74624389ac3261ed699b1f64e5717353da68895709d30a011bc8813b61a4a249
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:87eb62fe8aa7952d0d69dc13c4867ac2f51e821f56d149a9b588c826caec4c16
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:120b9891ee9db839ab0b2029a62e67ec4747c47c6ffe15af38e9035160f57b78
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:22bf00e1457b4a9e63901892aa4375a1d4a9d1b2c067c1f054d918a24b222b10
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:a108e9a7967001e0bdfed7124b01e9a3dbf1dec308444762973030aec57e1d5a
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:71b9fdcaf4cc761b1499be8e6ccff6c463cc6b9856e9deb1f93baad45ea59eb0
        version: 3.74.9
      entries:
      - name: rhacs-operator.v3.74.9
        version: 3.74.9
      - name: rhacs-operator.v3.74.8
        version: 3.74.8
      - name: rhacs-operator.v3.74.7
        version: 3.74.7
      - name: rhacs-operator.v3.74.6
        version: 3.74.6
      - name: rhacs-operator.v3.74.5
        version: 3.74.5
      - name: rhacs-operator.v3.74.4
        version: 3.74.4
      - name: rhacs-operator.v3.74.3
        version: 3.74.3
      - name: rhacs-operator.v3.74.2
        version: 3.74.2
      - name: rhacs-operator.v3.74.1
        version: 3.74.1
      - name: rhacs-operator.v3.74.0
        version: 3.74.0
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.74
    - currentCSV: rhacs-operator.v4.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9bf3a3e0fd573f91f03794eaf5291364a90392e5e7d4b5d4ed550ef4600ab826
          createdAt: "2023-10-20T07:04:50.531740+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.74.0 < 4.0.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:08f537cba262605a0eb0a510b6faf78e3ba26465f636d03cc1f7a6ed6b40d355
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:eee1aa054d4d15b23be293f31d3ea659e0ffc1d05174ef89a6c977dd3d4377cc
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:4c848eb7465e928251e05d32ffbc93cd14e7bb77972d22b8721bd6eaa4a93d2f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:3250d89e78de852a2627593f29469f0b3ff639ea3644b4b7d6d057d0c1ad3659
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:5a492b856e0b6d9fdb020b556d13186ef5e7095b0c736a2031500c9c00dad0c1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:139d80d398e7e5cfc85839e3039a3630b48c7ae6650d755754430fccd94d6a27
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0aedd8e865f72ffa8b79fd2bab48974caec3eea3e3a6f77c89fe6dacb8947057
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9bf3a3e0fd573f91f03794eaf5291364a90392e5e7d4b5d4ed550ef4600ab826
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:cfa10bb6b461f302674f185b26c8d6efaaeba7f3f65d2cf05ae1f069aa2710d3
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:c39621837cbc8419d187a43a9575ec8cb4021c58ea1b56f215003d7264eb5d03
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:09e79b5db8977520f9951cc43c339c8e40b4659d7da513347d5a88b2500f809d
        version: 4.0.5
      entries:
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.0
    - currentCSV: rhacs-operator.v4.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:8ffb6f0ff1bf016c3919c1eefc6c3db3d4a8fbe0e79919f72b4c281afe6d4b1c
          createdAt: "2024-01-13T18:55:22.011709+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 4.0.0 < 4.1.6'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:364c107c34694ca33ac724de78db3e41ee55f9098d9ade0775a51b11f1533065
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:8ffb6f0ff1bf016c3919c1eefc6c3db3d4a8fbe0e79919f72b4c281afe6d4b1c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a8ad0e76ce4cae9f9b36aecf125c4ed56fc2d776fe3df113d41bfc1c2f8b3602
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:8547b76cf0533dfe6da2aafbbd063021bad4ff78b32b2dfed4a7607496e0a904
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:8760d6dc68c86f0c5b53d8f0bfada0df012f80bc8e0023ba3522fe6829a05a41
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:762ad216e79be0837e98873202b0072c7fd399925ee94bde57cc61c68cbd5f4e
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:01b9f9798d3ecbefe3da1ac2ac782e8e8b930e7eeceb82fe3473b5bfc1ccd40a
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:7ac504b7824d17bd7ac55b1e59c26be308506f315a0ab0d8c45d3b4934c0a16c
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:2ec62f3da93a1346ad215744257820377ccf4ead84e7471c52e348a31a431d3c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:bdbe08ef2d01b114702780fab76905e6eb1851684692ffa0a4c84604bda3ddb3
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:34694b9f4fe7fffc0c547be4e311579b5bb0c1fa92d8257bf6d1581ffd655c46
        version: 4.1.6
      entries:
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.1
    - currentCSV: rhacs-operator.v4.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4253eca84cbb0c5fc033507ddc2deef235b054c0ba944fa271ebbba21545b872
          createdAt: "2024-03-13T14:51:55.084122+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.1.0 < 4.2.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:b5a15f2985b114bf9bc24a0d9942db5d0cb818e60fd15a2f9094aa00904cce77
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4253eca84cbb0c5fc033507ddc2deef235b054c0ba944fa271ebbba21545b872
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:12db245af3b8475c9b5680bb0035c566109819e05cfe62bd1dfdfb5d6ed25bea
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:b82f7f654235688fc07fb037422a73e1106e27024d39a68ef7fac1422b69496d
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:ff27f03064633d6d08f8383649069e39c868def4aeca8f2622750397e049d0ec
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:2f326cabf03f780a3e6120995d5442641764641f62430d82e05edbe21b30c7b2
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:3858ab936d4626a11bbf895d41bad181364ee5df3312f6dd29ebeb9b7d31ec77
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:e2b07cb7d690477d78a2b8e1875545998d4bdb946f80dd0533a14187da99dbbf
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:59f1862c8ce188906499f08201ae05deeb47bd4fc5eab4be0386ca44ec065da8
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:e79a5e5681ebfe2d6349e2763ee0c1fa521b20319846958e402d23229b1f257b
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:4b35f5f39865238a8e95d3cba7ebb5f15ccfbf6a2563c34f57d91a0b5c927066
        version: 4.2.5
      entries:
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.2
    - currentCSV: rhacs-operator.v4.3.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:266cbbc270efafa26d5d15f3e333c96eaedee3b47854ff337ae1c572d34901ce
          createdAt: "2024-06-06T22:31:19.365878+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.2.0 < 4.3.8'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c36a557c9dc2ae4479c432933a9a4dbef38f2be53360a68a5b6befb614d01dac
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:b0c9b3858fc0547dcb4eb3065dedd5bbcf1e1f6707a8ab812ed06882d4f6d6ad
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:b7eb17fbcf6c4a6e4205ad66f1ffd9b10a4634f6ef5045d407c266566b1f255d
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:bdf9785bff51194636c6e8a518448c6d12a297bd31c4bb2e79d07084409c18b4
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:266cbbc270efafa26d5d15f3e333c96eaedee3b47854ff337ae1c572d34901ce
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:9fba889021a876ab627ff8c496e4b6c46c221c61142cd0c76daf33c89762e2c1
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:10d0cc82b6eddd089f3a7af37b872c33248513d0b4d3f094b6f62c01f4460825
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:6b55a25eb3a611c648d9be883828834592229c16f4166fd28ef40a25fc667061
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:2c35cb6f518393d002edd2bf5f714c1d053cd0fab55f6152a2f35676e4979202
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:2bb643651c8d51525223df3b2b675947174bbe9f1eaea8688e1fa83044cfa41f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:5092c3fbb4a38c06a2f42d125dc718ef0e1005f304ad8046afdd2a5e7ccbc580
        version: 4.3.8
      entries:
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.3
    - currentCSV: rhacs-operator.v4.4.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:2b87569d7f4ef5e06c989b273c378e007c6d87b26eec41774ac1fc67dcd9e4e4
          createdAt: "2025-02-10T19:50:13.597164+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.3.0 < 4.4.8'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:65707fdad55767ff5598a8d3679fd0779a37f59501e58971b81bebcb23f1f9fc
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:d6b4f8caf342b105c8df14c5c9180cb36123e08d34756ee8fe4f2f76c34f8064
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:664a5dc40dc1fa06c47a11e13b0ab5e0612a981d7b69bcfb29422abf3d1a13b2
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:39cfd01236e905034f6645a73063427436c14e351e0e0bc735e73e62aa5a1c3c
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:467d069ee957aea07882b83b60ae9fcc62ccca159f5d643aac13e531ccf7df88
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:2b87569d7f4ef5e06c989b273c378e007c6d87b26eec41774ac1fc67dcd9e4e4
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:0df4e95c7406aee7812d145cdd4d8c61cec32857fd2079dbc4244954a9d8288a
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:183c75920ab8e7cb947cb966f1543e39b6fdb5e1bb0caa1b0cf5f89671f0328f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:76f9dc2e2ee5595c36530d97dec216e2dbefada2fe0747e3f6a3a725b75722ab
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:763e7568704971cb4656727dc223374de13e82b4304dd3f5387ddc0f4e42b8b6
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:69c12230a785619c765e3749a7d56f76dce9ce4f530c20e8a9387b99eb65627b
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:ba6e25a413c79131b966a097f72c36478933ea52deb55a62acda7011a96b3983
        version: 4.4.8
      entries:
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.4
    - currentCSV: rhacs-operator.v4.5.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a7dfbe3590668f91fe6694335b5a35b2fca07b4735348cc5202921755d2f654d
          createdAt: "2025-04-09T22:51:15.157672+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.4.0 < 4.5.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:5914548f6ce2160e92c5d96580ffba23cdf1ae940c1f73f3f6424d15c6e97c37
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:dc7f23341c8b209f1846f29f889d00ef232fb58da9d5b4bd561f3933093b2786
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:d928600f75461465598a10de8df2495a275fd5232a92b078f609d1f5774eb778
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:1817cf8d2c08b04391792321702ef53d16a4b1ca06459c00c9c053128f534358
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:c2656b794ec4236e36f48f71bf26bdedd9de27fc4acef39deba2bdcfcddab03b
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:cc4b329cc795752ecbb178d2e4efc152bde54b04181334eb5bd70304c3354cad
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:0ff948b5517df3773cae2df02b708c348d8707409899b1de4108588e31601c0c
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a7dfbe3590668f91fe6694335b5a35b2fca07b4735348cc5202921755d2f654d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:5109b304eb9458f97af85c2ba63e08f51672fa3f9406226b7f357b86b5828dc0
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:fde78d34ecd2944c7c16b4e823c76b3667bf6c5b468e9991128ecc4272ec0e2e
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:ed33e7396357f1db03a065573c020846a5872b12a21284439770686689222434
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:f07cba0738897e0e7ae74971be0470ce2a855657f133a6e1cdd8a7092f6151b0
        version: 4.5.9
      entries:
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.5
    - currentCSV: rhacs-operator.v4.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:c4d434e1ff052508e5594d2b95dafa8726fd8adf5317754774c013a46de224d8
          createdAt: "2025-09-26T09:24:28.046251+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.5.0 < 4.6.10'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:b5cd08349931669055f03d21331499ef2ea04597f8a90fb201615d43000271bc
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:ffe41363dc00b05b0daa0c6d16efb30e534086aa4e3d7ccc599c86d33813138c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:713e61d3c8b55de1edc9d7834d3e8576718e1327665f94a6b58178fa7496c0cc
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:65793c05aec8085a53fab8bcf33a85192c17b46c91a7d86e0a457d1e3a113878
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:899da9c23000bb821bc9d544ebb56bcda4f3245ceee5b1b9c12d09f580b4b9a1
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:8fad8ab2efdcbc99d9efa736a90cca6f236f0a9d45fdb355ae0b6b4dc42d77f0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:90c47c6f95571cfcca04aa4e05a18befad75c1b44a6f3a055fcdfdab749d29c5
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:16b61ca1b9ee4e0821e7fc82fe4def75ed22950e502730f5704ec4fe63bb2e59
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:c4d434e1ff052508e5594d2b95dafa8726fd8adf5317754774c013a46de224d8
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:b630d53ad0b862151be7e14f0483ca5adac570216deae66f60e75af87a4dfd46
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:19db8c5d2425dcf18ae63b025f946fc46adaec0578e9505415cb585797069e95
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:6da59c5b82625894f6efa3fa0bc5ea0378edd4f95cf42dee772e9c8a565dba81
        version: 4.6.10
      entries:
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.6
    - currentCSV: rhacs-operator.v4.7.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:5b60d8271d6fa70df3c8f7e0ecd0d64feecfdf674dc84036500fe5f145c9415a
          createdAt: "2025-09-26T11:05:05.900682+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.6.0 < 4.7.7'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:def71e1ddea5b8c1e54c8582f39dc9ecca855b38aa9423690150aa6ef0ff189f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8882c00c9be9b579654a9ee8cd0c38537cf92d2f4bd8ebded1678aa0b6ae827
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:34ea97c6e9fc2d23427d0c6f2628075c0eaa0e151ca1c381539b039c249217f7
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:98553e6f18d3c03a29212b4f7ea1d30a42bcf8b52ca12aecbcbc47769fc35b53
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:bac79889d4717233e89bb6c8fae6732ed7bec5b9a55fd3b3de1f1311d870d2b1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:6bb454d64a9e339b1aeb81ca17b26f33db4a84f19ced4ad81858e40fd01878dc
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:5b60d8271d6fa70df3c8f7e0ecd0d64feecfdf674dc84036500fe5f145c9415a
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:7596529cbdce7118226b79d69e0667ab35ee692569e4014b74abdf0fa7415222
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:5ed9902b52830be93436d7977c10e73d346b6290532f6d7245cfe7cca9c93b31
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:0c32318ee4cbd547e2c1fd64433cd23a629d1b06523d7c0f07ddb98c2baaa708
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:d8fcfe1267495801d23a0b94101c7b4d06b48346a949c97946ebcbffe148f4b6
        version: 4.7.7
      entries:
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.7
    - currentCSV: rhacs-operator.v4.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
          createdAt: "2025-09-12T11:06:11.361461+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.7.0 < 4.8.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:adf47cb6615d344533fcca0aee41774f5e8df9b952b9b89a57c2be2b7e0fc03c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:92ff7691963e620dc5bf6da0a7694e6bcd68c2545afc1642a59a443ba2ab5611
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:db691065e46f604b980b0b202f6c2e14c8b204094d784c4a15db1e4fecb2965e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:97df59280c12727a99fdec0c9bf56eb874e6a1cacd12d849709797d4be635ce9
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:e5c200f28187c1f6971562e41cae969183ee15f3d0feb076bb202b22ff67cbcb
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8964b1441251607b23f92918e9158724174013dc4af423988f56d6098dfb2d5
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:e949830f3034c41aa5434a3654b2b1ccba7dac0373bf43d7c50a9ea43bb686fb
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:16912df2c26e5fc2e6051c0501170722856db274415063608de716cf53851038
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:d1758b76c32f03b5a1a44951d18fb2bb0c4fd0a99d24f8c42fe9e6b059574889
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0c92d5b5508032d77559606927e45686179eaafc59e16d5496e088f2f212d258
        version: 4.8.4
      entries:
      - name: rhacs-operator.v4.8.4
        version: 4.8.4
      - name: rhacs-operator.v4.8.3
        version: 4.8.3
      - name: rhacs-operator.v4.8.2
        version: 4.8.2
      - name: rhacs-operator.v4.8.1
        version: 4.8.1
      - name: rhacs-operator.v4.8.0
        version: 4.8.0
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.8
    - currentCSV: rhacs-operator.v4.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
          createdAt: "2025-09-12T11:06:11.361461+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.7.0 < 4.8.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:e5c200f28187c1f6971562e41cae969183ee15f3d0feb076bb202b22ff67cbcb
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:e949830f3034c41aa5434a3654b2b1ccba7dac0373bf43d7c50a9ea43bb686fb
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:16912df2c26e5fc2e6051c0501170722856db274415063608de716cf53851038
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:adf47cb6615d344533fcca0aee41774f5e8df9b952b9b89a57c2be2b7e0fc03c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:97df59280c12727a99fdec0c9bf56eb874e6a1cacd12d849709797d4be635ce9
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:d1758b76c32f03b5a1a44951d18fb2bb0c4fd0a99d24f8c42fe9e6b059574889
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8964b1441251607b23f92918e9158724174013dc4af423988f56d6098dfb2d5
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0c92d5b5508032d77559606927e45686179eaafc59e16d5496e088f2f212d258
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:db691065e46f604b980b0b202f6c2e14c8b204094d784c4a15db1e4fecb2965e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:92ff7691963e620dc5bf6da0a7694e6bcd68c2545afc1642a59a443ba2ab5611
        version: 4.8.4
      entries:
      - name: rhacs-operator.v4.8.4
        version: 4.8.4
      - name: rhacs-operator.v4.8.3
        version: 4.8.3
      - name: rhacs-operator.v4.8.2
        version: 4.8.2
      - name: rhacs-operator.v4.8.1
        version: 4.8.1
      - name: rhacs-operator.v4.8.0
        version: 4.8.0
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: stable
    defaultChannel: stable
    packageName: rhacs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dns-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dns-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSRecord",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dns-operator",
                    "app.kubernetes.io/instance": "dnsrecord-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dnsrecord",
                    "app.kubernetes.io/part-of": "dns-operator"
                  },
                  "name": "dnsrecord-sample"
                },
                "spec": {
                  "endpoints": [
                    {
                      "dnsName": "dnsrecord-simple.example.com",
                      "recordTTL": 60,
                      "recordType": "A",
                      "targets": [
                        "52.215.108.61",
                        "52.30.101.221"
                      ]
                    }
                  ],
                  "providerRef": {
                    "name": "dns-provider-creds"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/dns-operator:v0.6.0
          createdAt: "2024-09-13T09:55:39Z"
          description: A Kubernetes Operator to manage the lifecycle of DNS resources
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kuadrant/dns-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSRecord is the Schema for the dnsrecords API
            displayName: DNSRecord
            kind: DNSRecord
            name: dnsrecords.kuadrant.io
            version: v1alpha1
        description: A Kubernetes Operator to manage the lifecycle of DNS resources
        displayName: DNS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dns
        - kuadrant
        links:
        - name: DNS Operator
          url: https://github.com/Kuadrant/dns-operator
        maintainers:
        - email: mnairn@redhat.com
          name: Michael Nairn
        - email: pbrookes@redhat.com
          name: Phil Brookes
        - email: cbrookes@redhat.com
          name: Craig Brookes
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/dns-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: dns-operator.v0.6.0
        version: 0.6.0
      name: stable
    defaultChannel: stable
    packageName: dns-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://www.redhat.com
    name: sap-commerce-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sap-commerce-operator.v0.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hybris.hybris.org/v1alpha1",
                "kind": "HybrisApp",
                "metadata": {
                  "name": "hybrisapp-sample"
                },
                "spec": {
                  "aJPServicePort": 30000,
                  "apachejvmRouteName": "",
                  "baseImageName": "hybris-base",
                  "baseImageTag": "v0.2",
                  "hybrisANTTaskNames": "",
                  "podHealthyProbePath": "/platform/init",
                  "sourceRepoLocalPropertiesOverride": "",
                  "sourceRepoRef": "main",
                  "sourceRepoURL": "https://github.com/redhat-sap/sap-commerce-app-example.git",
                  "startupProbeFailureThreshold": 100,
                  "startupProbePeriodSecond": 60
                }
              },
              {
                "apiVersion": "hybris.hybris.org/v1alpha1",
                "kind": "HybrisBase",
                "metadata": {
                  "name": "hybrisbase-sample"
                },
                "spec": {
                  "URL": "http://nexus.redhat.com/nexus/service/local/repositories/releases/content/org/hybris/hybris/1.0/hybris-1.0.tar.gz",
                  "buildSourceRepo": "https://github.com/redhat-sap/sap-commerce-operator",
                  "buildSourceRepoBranch": "main",
                  "imageName": "hybris-base",
                  "imageTag": "v0.2",
                  "jdkURL": "https://github.com/SAP/SapMachine/releases/download/sapmachine-11.0.5/sapmachine-jdk-11.0.5-1.x86_64.rpm",
                  "password": "admin123",
                  "username": "admin"
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          containerImage: quay.io/redhat-sap-cop/hybris-operator:0.0.4
          description: Builds and runs SAP Commerce in OpenShift
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-sap/sap-commerce-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HybrisApp is the Schema for the hybrisapps API
            displayName: Hybris Application
            kind: HybrisApp
            name: hybrisapps.hybris.hybris.org
            version: v1alpha1
          - description: HybrisBase is the Schema for the Hybris Base API
            displayName: Hybris Base
            kind: HybrisBase
            name: hybrisbases.hybris.hybris.org
            version: v1alpha1
        description: |-
          Builds and runs SAP Commerce in OpenShift

          * The operator provides APIs to initialize SAP Commerce with the provided SAP Commerce Server. Though there are no restrictions from the operator on the base package to be used from SAP, the default SAP base image provides a server preconfigured with an application server based on Apache Tomcat. Both HSQLDB and Tomcat are included in the SAP Commerce package, and are designed to be used for testing, development, and demonstration purposes. The custom classes created for various business needs along with an external RDBMS configuration can be used in a HybrisApp creation to execute SAP Commerce in a production setup.

          * The operator also enables the use of SAP Commerce in a clustered model. A SAP Commerce Cluster is a number of individual SAP Commerce installations using a common set of data on one database. A cluster configuration lets you balance the load between the nodes to scale SAP Commerce in a way that is transparent to the user.

          * The clustering offers few communication protocols as options (UDP: both multicast or unicast, or JGroups). In order to support multicast for SAP Commerce clustering the corresponding OpenShift namespace has to be enabled for multicast usage.

          For more details on SAP Commerce consult the product documentation: [https://help.sap.com/viewer/product/SAP_COMMERCE](https://help.sap.com/viewer/product/SAP_COMMERCE)
        displayName: SAP Commerce Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sap-commerce-operator
        links:
        - name: SAP Commerce Operator
          url: https://github.com/redhat-sap/sap-commerce-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
          url: https://www.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/redhat-sap-cop/hybris-operator:0.0.4
        version: 0.0.4
      entries:
      - name: sap-commerce-operator.v0.0.4
        version: 0.0.4
      - name: sap-commerce-operator.v0.0.3
        version: 0.0.3
      - name: sap-commerce-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: sap-commerce-operator
    provider:
      name: Red Hat, Inc.
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cognitoidentityprovider-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cognitoidentityprovider-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cognitoidentityprovider-controller:1.1.1
          createdAt: "2025-09-19T20:58:49Z"
          description: AWS Cognito Identity Provider controller is a service controller
            for managing Cognito Identity Provider resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: UserPool represents the state of an AWS cognitoidentityprovider
              UserPool resource.
            displayName: UserPool
            kind: UserPool
            name: userpools.cognitoidentityprovider.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Cognito Identity Provider resources in AWS from within your Kubernetes cluster.

          **About Amazon Cognito Identity Provider**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Cognito Identity Provider
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cognitoidentityprovider
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Cognito Identity Provider Developer Resources
          url: https://aws.amazon.com/Cognito Identity Provider/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cognitoidentityprovider maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cognitoidentityprovider-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-cognitoidentityprovider-controller.v1.1.1
        version: 1.1.1
      - name: ack-cognitoidentityprovider-controller.v1.1.0
        version: 1.1.0
      - name: ack-cognitoidentityprovider-controller.v1.0.12
        version: 1.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-cognitoidentityprovider-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sumo Logic
      provider-url: https://www.sumologic.com/
    name: sumologic-kubernetes-collection-helm-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "helm-operator.sumologic.com/v1alpha1",
                    "kind": "SumologicCollection",
                    "metadata": {
                        "name": "collection"
                    },
                    "spec": {
                        "sumologic": {
                            "accessId": "",
                            "accessKey": "",
                            "clusterName": "kubernetes",
                            "scc": {
                                "create": true
                            }
                        },
                        "kube-prometheus-stack": {
                            "prometheus-node-exporter": {
                                "service": {
                                    "port": 9200,
                                    "targetPort": 9200
                                }
                            }
                        },
                        "otellogs": {
                            "daemonset": {
                                "containers": {
                                    "otelcol": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                },
                                "initContainers": {
                                    "changeowner": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                }
                            }
                        }
                    }
                }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing,Monitoring
          containerImage: registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
          createdAt: "2024-07-19T10:48:00Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator.git
          support: Sumo Logic
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configuration for Sumo Logic Kubernetes Collection
            displayName: SumologicCollection
            kind: SumologicCollection
            name: sumologiccollections.helm-operator.sumologic.com
            version: v1alpha1
        description: Helm Operator for the Sumo Logic Kubernetes Collection Helm Chart
          v4.13.0-2
        displayName: Sumo Logic Kubernetes Collection Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - collection
        - logging
        - metrics
        - monitoring
        links:
        - name: Sumo Logic Kubernetes Collection Helm Operator
          url: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator
        maintainers:
        - email: collection@sumologic.com
          name: Sumo Logic
        maturity: alpha
        provider:
          name: Sumo Logic
          url: https://www.sumologic.com/
        relatedImages:
        - registry.connect.redhat.com/sumologic/thanos@sha256:16faf7191668e86d8b576bba055dd0cbdeb009635976b9e07bd00f926032f11c
        - registry.connect.redhat.com/sumologic/autoinstrumentation-java@sha256:242958b7d8905f4ac02f6ad38897ba58b9badf2a6213a389e67dcc71fee3630c
        - registry.connect.redhat.com/sumologic/busybox@sha256:209bc8e0b1d958fc699f3bb75b1248544e372ad35cdc46c991bd1698f641e1f9
        - registry.connect.redhat.com/sumologic/opentelemetry-operator@sha256:5587fe5b7ebfa57ad320ad2521103f908317388d1741c40e5f7be5ed9e5f9ad7
        - registry.connect.redhat.com/sumologic/prometheus@sha256:0ef6eaf570a40499dd5de92a79593e950bc771b014e0ee6ff7419b4b82a36fcc
        - registry.connect.redhat.com/sumologic/nginx-unprivileged@sha256:7260555d16c94c4fe1e016626b4cb8a5c0de6b700e2eea28744331fb298f71ae
        - registry.connect.redhat.com/sumologic/node-exporter@sha256:11e9e604432a66d9b388cb1bf70eac077c7f0372d1264342ef927d76b8ef0272
        - registry.connect.redhat.com/sumologic/prometheus-config-reloader@sha256:0b605d36782afd8f50fd631d0e9c7b19d39ce962af89b7a737b2f09a8ceff291
        - registry.connect.redhat.com/sumologic/kubernetes-tools-kubectl@sha256:d3155b864cc74f327adedf7d013a8f0fcb11c2e73a5655b7cafa848824e80172
        - registry.connect.redhat.com/sumologic/telegraf-operator@sha256:ff958222be01c0c3966474463e43bbd5bcf348790dd34bb761718cb1cea2d793
        - registry.connect.redhat.com/sumologic/prometheus-operator@sha256:58a329e6582456b32abebb470086191f8366b927af4703741613993465f116b8
        - registry.connect.redhat.com/sumologic/sumologic-otel-collector@sha256:4554d43dba204e3ec8e91b36442a438f10234fa0cb4db7fb7880b257a6485112
        - registry.connect.redhat.com/sumologic/tailing-sidecar@sha256:738d653a78dc415243ab3627db9135768175fb7f51d0803566f5d9e824c8744b
        - registry.connect.redhat.com/sumologic/autoinstrumentation-python@sha256:f0fa45fa45e64c7df88f21181ae18ee96fe83cd3988258663d0806ea6f39e6bb
        - registry.connect.redhat.com/sumologic/kubernetes-setup@sha256:f07eaf68fa11ffc708e4485a83bd3763fc5712f7d73d29edfe5b9c9e08f05df6
        - registry.connect.redhat.com/sumologic/autoinstrumentation-dotnet@sha256:f4f385c4489c74b5c43f2fe6b6620a9eb8d9ecab470567236b3c45f636106640
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:e1f49fcf5b7e86d9eb49c16dbbf9484cb486f55f999fd51d7c738c3b1f2666cf
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
        - registry.connect.redhat.com/sumologic/kube-state-metrics@sha256:3551ededa327a35db798ec8286cd843580e3ae395d2231e041e8c4b7621127ee
        - registry.connect.redhat.com/sumologic/autoinstrumentation-nodejs@sha256:c206eef4fdecaa58b4f01e5e7f0dbd74e9606eab2b4debf47461d36191e248d1
        - registry.connect.redhat.com/sumologic/metrics-server@sha256:65d2a3ddc4b4183f438fa18b655079c93c905cb9e3b8c54977336205dc1fff07
        - registry.connect.redhat.com/sumologic/tailing-sidecar-operator@sha256:c16b8b18e9b263fc3eb24059bba512bcfc3201b32105af6e5518392826ddd66c
        - registry.connect.redhat.com/sumologic/telegraf@sha256:75efd2e509c6017ea0fc5c9103ce068cbc89087b16506a4ef48cb5ff00eab84e
        version: 4.13.0-2
      entries:
      - name: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
        version: 4.13.0-2
      name: alpha
    defaultChannel: alpha
    packageName: sumologic-kubernetes-collection-helm-operator
    provider:
      name: Sumo Logic
      url: https://www.sumologic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: rh-service-binding-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-binding-operator.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ServiceBinding expresses intent to bind an operator-backed
              service with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding for Kubernetes. Describes the connection
              between a Provisioned Service and an Application Projection. Implements
              Service Binding Specifcation for Kubernetes (https://github.com/k8s-service-bindings/spec).
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: alpha
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:fe6629401bc8447da95ec55949b5a27e620364bb6e02f86230262699ac8a5d55
        version: 0.9.1
      entries:
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: beta
    - currentCSV: service-binding-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindablekinds-sample"
                }
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:e799bd06043028d214baeeffc683710ad2eed7b83272fc653e3b89f69711319b
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:e799bd06043028d214baeeffc683710ad2eed7b83272fc653e3b89f69711319b
        version: 0.11.0
      entries:
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: candidate
    - currentCSV: service-binding-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:3d7a3648e86848402577dd19c1f8fd130a5556ebdfbc0f663a72f9e141024812
        version: 0.10.0
      entries:
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: preview
    - currentCSV: service-binding-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindable-kinds"
                },
                "spec": {}
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  },
                  "type": "postgres",
                  "workload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: registry.redhat.io/ocp-tools-4/service-binding-rhel8-operator@sha256:9c8fb0c8f9d2f54685561af21e595aa88bf309eaebebdf23fb6997b75d026b18
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'.
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1alpha3
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1beta1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
          - description: Service Binding implementing community specification (https://github.com/k8s-service-bindings/spec)
              describing the connection between a backing service and an application
              workload.
            displayName: Service Binding (Spec API)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1beta1
        description: |
          The Service Binding Operator manages the data plane for applications and backing services. It enables developers to connect their
          applications to backing services (REST API, databases, event buses and others) with a consistent and predictable
          experience. The operator reads data made available by the control plane of backing services and projects the bindings
          data to applications either as accessible files mounted into the application’s container or environment variables.

          In Kubernetes, each service suggests a different way to access their secrets and developers will consume them in their
          application in a custom way. While this provides a good deal of flexibility, it also creates discrepancies and
          complexity dealing with each specific services, which leads large development teams to lose overall velocity.

          Service Binding Operator removes the error-prone manual configuration of binding information, provides service operators
          a low-touch administrative experience to provision and manage access to services and enriches development lifecycle.

          For additional details refer to the project [documentation](https://redhat-developer.github.io/service-binding-operator).

          Get Started with Service Binding Operator
          * Find out about SBO’s [prerequisites](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/installing-service-binding.html).
          * Refer to the [Quick Start Guide](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/quick-start.html) to see SBO in action in a simple scenario.

          Need Help
          * Raise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)

          Licence
          * Service Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE)

          Service Binding Operator is providing and supporting two different APIs.
          * `binding.operators.coreos.com/v1alpha1`: This API is compliant with the Service Binding specification. It is fully supported and considered as the mature API. It’s the API we recommend in production environments.
          * `servicebinding.io/v1beta1`: This API is the one defined in the Service Binding specification. It is available in Service Binding Operator 1.x.

          We’ll consider supporting only the API from the specification, once the Service Binding specification will be officially GA. Of course, that will be a breaking change and handle in a safe way for you.
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Documentation
          url: https://redhat-developer.github.io/service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4/service-binding-rhel8-operator@sha256:9c8fb0c8f9d2f54685561af21e595aa88bf309eaebebdf23fb6997b75d026b18
        version: 1.3.3
      entries:
      - name: service-binding-operator.v1.3.3
        version: 1.3.3
      - name: service-binding-operator.v1.3.1
        version: 1.3.1
      - name: service-binding-operator.v1.3.0
        version: 1.3.0
      - name: service-binding-operator.v1.2.0
        version: 1.2.0
      - name: service-binding-operator.v1.1.1
        version: 1.1.1
      - name: service-binding-operator.v1.1.0
        version: 1.1.0
      - name: service-binding-operator.v1.0.1
        version: 1.0.1
      - name: service-binding-operator.v1.0.0
        version: 1.0.0
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: rh-service-binding-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cluster-logging
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-logging.v5.9.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "name": "remote-elasticsearch",
                      "secret": {
                        "name": "elasticsearch"
                      },
                      "type": "elasticsearch",
                      "url": "remote.example.org:9200"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "audit",
                        "infrastructure"
                      ],
                      "name": "enable-default-log-store",
                      "outputRefs": [
                        "default"
                      ]
                    },
                    {
                      "inputRefs": [
                        "application"
                      ],
                      "name": "forward-to-remote",
                      "outputRefs": [
                        "remote-elasticsearch"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogging",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "collection": {
                    "type": "vector"
                  },
                  "logStore": {
                    "elasticsearch": {
                      "nodeCount": 3,
                      "redundancyPolicy": "SingleRedundancy",
                      "resources": {
                        "requests": {
                          "memory": "2Gi"
                        }
                      },
                      "storage": {
                        "size": "200G"
                      }
                    },
                    "retentionPolicy": {
                      "application": {
                        "maxAge": "7d"
                      }
                    },
                    "type": "elasticsearch"
                  },
                  "managementState": "Managed",
                  "visualization": {
                    "kibana": {
                      "replicas": 1
                    },
                    "type": "kibana"
                  }
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          console.openshift.io/plugins: '["logging-view-plugin"]'
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:54054faf0d2f375dd6d7fc7929a2dbb529508e1b3c004a56fc97c2a682d16be2
          createdAt: "2025-03-14T20:19:21Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing your aggregated logging stack.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.7.0-0 <5.9.15'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Logging (team-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.


              There are built-in input names for common log categories, and you can
              define custom inputs to do additional filtering.


              There is a built-in output name for the default openshift log store, but
              you can define your own outputs with a URL and other connection information
              to forward logs to other stores or processors, inside or outside the cluster.


              For more details see the documentation on the API fields.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.logging.openshift.io
            version: v1
          - description: A Red Hat OpenShift Logging instance. ClusterLogging is the
              Schema for the clusterloggings API
            displayName: Cluster Logging
            kind: ClusterLogging
            name: clusterloggings.logging.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates and manages the aggregated logging stack as a cluster-wide service.

          ##Features
          * **Create/Destroy**: Launch and create an aggregated logging stack to support the entire OCP cluster.
          * **Simplified Configuration**: Configure your aggregated logging cluster's structure like components and end points easily.
          * Configure log forwarding to an in-cluster, Red Hat managed log storage solution based on LokiStack or Elasticsearch

          ## Prerequisites
          ### Red Hat OpenShift Logging Namespace
          Cluster logging and the Red Hat OpenShift Logging Operator is only deployable to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - fluentd
        - logging
        - aggregated
        - efk
        - vector
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Fluentd
          url: https://www.fluentd.org/
        - name: Vector
          url: https://vector.dev/
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:54054faf0d2f375dd6d7fc7929a2dbb529508e1b3c004a56fc97c2a682d16be2
        - registry.redhat.io/openshift-logging/logging-view-plugin-rhel9@sha256:8023b072b45e9a21ddef83a56f4760d476053250d5a920d8eb5f03eecab8b1b3
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:54e3a5efaae5ecc2a08db03a421aa79e323b48c764115ea84d83a37c21d1d997
        - registry.redhat.io/openshift-logging/fluentd-rhel9@sha256:a676c9164ec2523b2002dc5eba234ce774926dbf7fe87de6d72b53f99cbf1db5
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:865572275b79462f14e84c69438a61afe9465cfe45ec164caf30965730b276f7
        version: 5.9.15
      entries:
      - name: cluster-logging.v5.9.15
        version: 5.9.15
      - name: cluster-logging.v5.9.14
        version: 5.9.14
      - name: cluster-logging.v5.9.13
        version: 5.9.13
      - name: cluster-logging.v5.9.12
        version: 5.9.12
      - name: cluster-logging.v5.9.11
        version: 5.9.11
      - name: cluster-logging.v5.9.10
        version: 5.9.10
      - name: cluster-logging.v5.9.9
        version: 5.9.9
      - name: cluster-logging.v5.9.8
        version: 5.9.8
      - name: cluster-logging.v5.9.7
        version: 5.9.7
      - name: cluster-logging.v5.9.6
        version: 5.9.6
      - name: cluster-logging.v5.9.5
        version: 5.9.5
      - name: cluster-logging.v5.9.4
        version: 5.9.4
      - name: cluster-logging.v5.9.3
        version: 5.9.3
      - name: cluster-logging.v5.9.2
        version: 5.9.2
      - name: cluster-logging.v5.9.1
        version: 5.9.1
      - name: cluster-logging.v5.9.0
        version: 5.9.0
      name: stable
    - currentCSV: cluster-logging.v5.8.22
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "name": "remote-elasticsearch",
                      "secret": {
                        "name": "elasticsearch"
                      },
                      "type": "elasticsearch",
                      "url": "remote.example.org:9200"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "audit",
                        "infrastructure"
                      ],
                      "name": "enable-default-log-store",
                      "outputRefs": [
                        "default"
                      ]
                    },
                    {
                      "inputRefs": [
                        "application"
                      ],
                      "name": "forward-to-remote",
                      "outputRefs": [
                        "remote-elasticsearch"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogging",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "collection": {
                    "type": "vector"
                  },
                  "logStore": {
                    "elasticsearch": {
                      "nodeCount": 3,
                      "redundancyPolicy": "SingleRedundancy",
                      "resources": {
                        "requests": {
                          "memory": "2Gi"
                        }
                      },
                      "storage": {
                        "size": "200G"
                      }
                    },
                    "retentionPolicy": {
                      "application": {
                        "maxAge": "7d"
                      }
                    },
                    "type": "elasticsearch"
                  },
                  "managementState": "Managed",
                  "visualization": {
                    "kibana": {
                      "replicas": 1
                    },
                    "type": "kibana"
                  }
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.12-v4.17
          console.openshift.io/plugins: '["logging-view-plugin"]'
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:8e13bd3e1654fe26c1d4d8d4a1bfb12e04cee7899cf16b5fa07b62b8f7494af5
          createdAt: "2024-10-17T17:45:03Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing your aggregated logging stack.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.22'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-5.8
          operators.operatorframework.io.bundle.channels.v1: stable-5.8
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Logging (team-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "ClusterLogForwarder is an API to configure forwarding logs.
              \n You configure forwarding by specifying a list of `pipelines`, which
              forward from a set of named inputs to a set of named outputs. \n There
              are built-in input names for common log categories, and you can define
              custom inputs to do additional filtering. \n There is a built-in output
              name for the default openshift log store, but you can define your own
              outputs with a URL and other connection information to forward logs
              to other stores or processors, inside or outside the cluster. \n For
              more details see the documentation on the API fields."
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.logging.openshift.io
            version: v1
          - description: A Red Hat OpenShift Logging instance. ClusterLogging is the
              Schema for the clusterloggings API
            displayName: Cluster Logging
            kind: ClusterLogging
            name: clusterloggings.logging.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          ### Notice for OCP 4.16 and 4.17 Only

          Logging 5.8 is supported on OCP 4.16 and 4.17 to help customers upgrade and migrate to Logging 6.0. **Logging 5.8 will be End of Life in November 2025.**  At that time, it will no longer be available or supported.

          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates and manages the aggregated logging stack as a cluster-wide service.

          ##Features
          * **Create/Destroy**: Launch and create an aggregated logging stack to support the entire OCP cluster.
          * **Simplified Configuration**: Configure your aggregated logging cluster's structure like components and end points easily.
          * Configure log forwarding to an in-cluster, Red Hat managed log storage solution based on LokiStack or Elasticsearch

          ## Prerequisites
          ### Red Hat OpenShift Logging Namespace
          Cluster logging and the Red Hat OpenShift Logging Operator is only deployable to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - fluentd
        - logging
        - aggregated
        - efk
        - vector
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Fluentd
          url: https://www.fluentd.org/
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/logging-view-plugin-rhel9@sha256:f00d4e7ce6b37f1a053c5dbe554359d844989c2243574000ada0f5bc678ffff8
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:8e13bd3e1654fe26c1d4d8d4a1bfb12e04cee7899cf16b5fa07b62b8f7494af5
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:f26518dd3dbf6b6a9e9e20e40741687996b1160c6ad8ad417989197800c2b3fe
        - registry.redhat.io/openshift-logging/fluentd-rhel9@sha256:9296e50fccae60dea580186bc5ae2a387172dc90ece35663d96c4553a2446a01
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:8f29d6eaeb4a11c9a61e3d763605e3c9c8655df7fbb4e39e5b8311d18449f04a
        version: 5.8.22
      entries:
      - name: cluster-logging.v5.8.22
        version: 5.8.22
      - name: cluster-logging.v5.8.21
        version: 5.8.21
      - name: cluster-logging.v5.8.20
        version: 5.8.20
      - name: cluster-logging.v5.8.19
        version: 5.8.19
      - name: cluster-logging.v5.8.18
        version: 5.8.18
      - name: cluster-logging.v5.8.17
        version: 5.8.17
      - name: cluster-logging.v5.8.16
        version: 5.8.16
      - name: cluster-logging.v5.8.15
        version: 5.8.15
      - name: cluster-logging.v5.8.14
        version: 5.8.14
      - name: cluster-logging.v5.8.13
        version: 5.8.13
      - name: cluster-logging.v5.8.12
        version: 5.8.12
      - name: cluster-logging.v5.8.11
        version: 5.8.11
      - name: cluster-logging.v5.8.10
        version: 5.8.10
      - name: cluster-logging.v5.8.9
        version: 5.8.9
      - name: cluster-logging.v5.8.8
        version: 5.8.8
      - name: cluster-logging.v5.8.7
        version: 5.8.7
      - name: cluster-logging.v5.8.6
        version: 5.8.6
      - name: cluster-logging.v5.8.5
        version: 5.8.5
      - name: cluster-logging.v5.8.4
        version: 5.8.4
      - name: cluster-logging.v5.8.3
        version: 5.8.3
      - name: cluster-logging.v5.8.2
        version: 5.8.2
      - name: cluster-logging.v5.8.1
        version: 5.8.1
      - name: cluster-logging.v5.8.0
        version: 5.8.0
      name: stable-5.8
    - currentCSV: cluster-logging.v5.9.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "name": "remote-elasticsearch",
                      "secret": {
                        "name": "elasticsearch"
                      },
                      "type": "elasticsearch",
                      "url": "remote.example.org:9200"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "audit",
                        "infrastructure"
                      ],
                      "name": "enable-default-log-store",
                      "outputRefs": [
                        "default"
                      ]
                    },
                    {
                      "inputRefs": [
                        "application"
                      ],
                      "name": "forward-to-remote",
                      "outputRefs": [
                        "remote-elasticsearch"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "ClusterLogging",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "collection": {
                    "type": "vector"
                  },
                  "logStore": {
                    "elasticsearch": {
                      "nodeCount": 3,
                      "redundancyPolicy": "SingleRedundancy",
                      "resources": {
                        "requests": {
                          "memory": "2Gi"
                        }
                      },
                      "storage": {
                        "size": "200G"
                      }
                    },
                    "retentionPolicy": {
                      "application": {
                        "maxAge": "7d"
                      }
                    },
                    "type": "elasticsearch"
                  },
                  "managementState": "Managed",
                  "visualization": {
                    "kibana": {
                      "replicas": 1
                    },
                    "type": "kibana"
                  }
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          console.openshift.io/plugins: '["logging-view-plugin"]'
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:54054faf0d2f375dd6d7fc7929a2dbb529508e1b3c004a56fc97c2a682d16be2
          createdAt: "2025-03-14T20:19:21Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing your aggregated logging stack.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.7.0-0 <5.9.15'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Logging (team-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.


              There are built-in input names for common log categories, and you can
              define custom inputs to do additional filtering.


              There is a built-in output name for the default openshift log store, but
              you can define your own outputs with a URL and other connection information
              to forward logs to other stores or processors, inside or outside the cluster.


              For more details see the documentation on the API fields.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.logging.openshift.io
            version: v1
          - description: A Red Hat OpenShift Logging instance. ClusterLogging is the
              Schema for the clusterloggings API
            displayName: Cluster Logging
            kind: ClusterLogging
            name: clusterloggings.logging.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates and manages the aggregated logging stack as a cluster-wide service.

          ##Features
          * **Create/Destroy**: Launch and create an aggregated logging stack to support the entire OCP cluster.
          * **Simplified Configuration**: Configure your aggregated logging cluster's structure like components and end points easily.
          * Configure log forwarding to an in-cluster, Red Hat managed log storage solution based on LokiStack or Elasticsearch

          ## Prerequisites
          ### Red Hat OpenShift Logging Namespace
          Cluster logging and the Red Hat OpenShift Logging Operator is only deployable to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - fluentd
        - logging
        - aggregated
        - efk
        - vector
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Fluentd
          url: https://www.fluentd.org/
        - name: Vector
          url: https://vector.dev/
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:54054faf0d2f375dd6d7fc7929a2dbb529508e1b3c004a56fc97c2a682d16be2
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:54e3a5efaae5ecc2a08db03a421aa79e323b48c764115ea84d83a37c21d1d997
        - registry.redhat.io/openshift-logging/fluentd-rhel9@sha256:a676c9164ec2523b2002dc5eba234ce774926dbf7fe87de6d72b53f99cbf1db5
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:865572275b79462f14e84c69438a61afe9465cfe45ec164caf30965730b276f7
        - registry.redhat.io/openshift-logging/logging-view-plugin-rhel9@sha256:8023b072b45e9a21ddef83a56f4760d476053250d5a920d8eb5f03eecab8b1b3
        version: 5.9.15
      entries:
      - name: cluster-logging.v5.9.15
        version: 5.9.15
      - name: cluster-logging.v5.9.14
        version: 5.9.14
      - name: cluster-logging.v5.9.13
        version: 5.9.13
      - name: cluster-logging.v5.9.12
        version: 5.9.12
      - name: cluster-logging.v5.9.11
        version: 5.9.11
      - name: cluster-logging.v5.9.10
        version: 5.9.10
      - name: cluster-logging.v5.9.9
        version: 5.9.9
      - name: cluster-logging.v5.9.8
        version: 5.9.8
      - name: cluster-logging.v5.9.7
        version: 5.9.7
      - name: cluster-logging.v5.9.6
        version: 5.9.6
      - name: cluster-logging.v5.9.5
        version: 5.9.5
      - name: cluster-logging.v5.9.4
        version: 5.9.4
      - name: cluster-logging.v5.9.3
        version: 5.9.3
      - name: cluster-logging.v5.9.2
        version: 5.9.2
      - name: cluster-logging.v5.9.1
        version: 5.9.1
      - name: cluster-logging.v5.9.0
        version: 5.9.0
      name: stable-5.9
    - currentCSV: cluster-logging.v6.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.14-v4.17
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:3e36ea9a808945d3817d169f736b1758875f2fbc6a7ae845266e1545a1b713c7
          createdAt: "2025-05-14T12:32:39Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.8.0-0 <6.0.10'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.0
          operators.operatorframework.io.bundle.channels.v1: stable-6.0
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:3e36ea9a808945d3817d169f736b1758875f2fbc6a7ae845266e1545a1b713c7
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:4e9936c090818d1d8f812487a9a8218fda5b11f8c06167fffe35bae1bafcadf9
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:15487d7b9d8f83d0df0b6fbf380f20df3518dfb1b831fac4c3209e811494b436
        version: 6.0.10
      entries:
      - name: cluster-logging.v6.0.10
        version: 6.0.10
      - name: cluster-logging.v6.0.9
        version: 6.0.9
      - name: cluster-logging.v6.0.8
        version: 6.0.8
      - name: cluster-logging.v6.0.7
        version: 6.0.7
      - name: cluster-logging.v6.0.6
        version: 6.0.6
      - name: cluster-logging.v6.0.5
        version: 6.0.5
      - name: cluster-logging.v6.0.4
        version: 6.0.4
      - name: cluster-logging.v6.0.3
        version: 6.0.3
      - name: cluster-logging.v6.0.2
        version: 6.0.2
      - name: cluster-logging.v6.0.1
        version: 6.0.1
      - name: cluster-logging.v6.0.0
        version: 6.0.0
      name: stable-6.0
    - currentCSV: cluster-logging.v6.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.15-v4.18
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f7101c962de4e5643bd7f760104a41cc48f020889159a97e2d6711703fbbec64
          createdAt: "2025-05-14T12:38:55Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.9.0-0 <6.1.8'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.1
          operators.operatorframework.io.bundle.channels.v1: stable-6.1
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f7101c962de4e5643bd7f760104a41cc48f020889159a97e2d6711703fbbec64
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:96e438dd5c080cfa47080410f375dc71da24f6d046cf609db41c8b873d85404f
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:76037c744b34b1c4fccf18f17888813d9670bb6c061c2c3de843aa43c0497170
        version: 6.1.8
      entries:
      - name: cluster-logging.v6.1.8
        version: 6.1.8
      - name: cluster-logging.v6.1.7
        version: 6.1.7
      - name: cluster-logging.v6.1.6
        version: 6.1.6
      - name: cluster-logging.v6.1.5
        version: 6.1.5
      - name: cluster-logging.v6.1.4
        version: 6.1.4
      - name: cluster-logging.v6.1.3
        version: 6.1.3
      - name: cluster-logging.v6.1.2
        version: 6.1.2
      - name: cluster-logging.v6.1.1
        version: 6.1.1
      - name: cluster-logging.v6.1.0
        version: 6.1.0
      name: stable-6.1
    - currentCSV: cluster-logging.v6.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.16-v4.19
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
          createdAt: "2025-07-29T15:25:15Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0-0 <6.2.5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.2
          operators.operatorframework.io.bundle.channels.v1: stable-6.2
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Installing logging](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html-single/logging/index#installing-logging-6-2)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:c76c853565fc93da88ba0116425839cfb86cab63cbde9856d7cca78278d7758a
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:502a5d2891049bbc32115fda28086841351cd9e6a3110b4c3c412f683f2b2491
        version: 6.2.5
      entries:
      - name: cluster-logging.v6.2.5
        version: 6.2.5
      - name: cluster-logging.v6.2.4
        version: 6.2.4
      - name: cluster-logging.v6.2.3
        version: 6.2.3
      - name: cluster-logging.v6.2.2
        version: 6.2.2
      - name: cluster-logging.v6.2.1
        version: 6.2.1
      - name: cluster-logging.v6.2.0
        version: 6.2.0
      name: stable-6.2
    defaultChannel: stable-6.2
    packageName: cluster-logging
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-apigateway-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-apigateway-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Resource",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "RestAPI",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Stage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "VPCLink",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/apigateway-controller:1.3.1
          createdAt: "2025-09-19T21:13:12Z"
          description: AWS API Gateway controller is a service controller for managing
            API Gateway resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIIntegrationResponse represents the state of an AWS apigateway
              APIIntegrationResponse resource.
            displayName: APIIntegrationResponse
            kind: APIIntegrationResponse
            name: apiintegrationresponses.apigateway.services.k8s.aws
            version: v1alpha1
          - description: APIKey represents the state of an AWS apigateway APIKey resource.
            displayName: APIKey
            kind: APIKey
            name: apikeys.apigateway.services.k8s.aws
            version: v1alpha1
          - description: APIMethodResponse represents the state of an AWS apigateway
              APIMethodResponse resource.
            displayName: APIMethodResponse
            kind: APIMethodResponse
            name: apimethodresponses.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Authorizer represents the state of an AWS apigateway Authorizer
              resource.
            displayName: Authorizer
            kind: Authorizer
            name: authorizers.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Deployment represents the state of an AWS apigateway Deployment
              resource.
            displayName: Deployment
            kind: Deployment
            name: deployments.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Integration represents the state of an AWS apigateway Integration
              resource.
            displayName: Integration
            kind: Integration
            name: integrations.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Method represents the state of an AWS apigateway Method resource.
            displayName: Method
            kind: Method
            name: methods.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Resource represents the state of an AWS apigateway Resource
              resource.
            displayName: Resource
            kind: Resource
            name: resources.apigateway.services.k8s.aws
            version: v1alpha1
          - description: RestAPI represents the state of an AWS apigateway RestAPI
              resource.
            displayName: RestAPI
            kind: RestAPI
            name: restapis.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Stage represents the state of an AWS apigateway Stage resource.
            displayName: Stage
            kind: Stage
            name: stages.apigateway.services.k8s.aws
            version: v1alpha1
          - description: VPCLink represents the state of an AWS apigateway VPCLink
              resource.
            displayName: VPCLink
            kind: VPCLink
            name: vpclinks.apigateway.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon API Gateway resources in AWS from within your Kubernetes cluster.

          **About Amazon API Gateway**

          Amazon API Gateway is an AWS service for creating, publishing, maintaining, monitoring, and securing REST, HTTP, and WebSocket APIs at any scale. API developers can create APIs that access AWS or other web services, as well as data stored in the AWS Cloud. As an API Gateway API developer, you can create APIs for use in your own client applications. Or you can make your APIs available to third-party app developers.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon API Gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apigateway
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon API Gateway Developer Resources
          url: https://aws.amazon.com/api-gateway/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: apigateway maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/apigateway-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-apigateway-controller.v1.3.1
        version: 1.3.1
      - name: ack-apigateway-controller.v1.3.0
        version: 1.3.0
      - name: ack-apigateway-controller.v1.2.5
        version: 1.2.5
      - name: ack-apigateway-controller.v1.2.4
        version: 1.2.4
      - name: ack-apigateway-controller.v1.2.3
        version: 1.2.3
      - name: ack-apigateway-controller.v1.2.2
        version: 1.2.2
      - name: ack-apigateway-controller.v1.2.1
        version: 1.2.1
      - name: ack-apigateway-controller.v1.1.0
        version: 1.1.0
      - name: ack-apigateway-controller.v1.0.6
        version: 1.0.6
      - name: ack-apigateway-controller.v1.0.5
        version: 1.0.5
      - name: ack-apigateway-controller.v1.0.4
        version: 1.0.4
      - name: ack-apigateway-controller.v1.0.2
        version: 1.0.2
      - name: ack-apigateway-controller.v1.0.1
        version: 1.0.1
      - name: ack-apigateway-controller.v1.0.0
        version: 1.0.0
      - name: ack-apigateway-controller.v0.0.3
        version: 0.0.3
      - name: ack-apigateway-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-apigateway-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nexus-repository-ha-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-repository-ha-operator-certified.v3.84.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusRepo",
                "metadata": {
                  "name": "nexusrepo-sample"
                },
                "spec": {
                  "ingress": {
                    "additionalRules": null,
                    "annotations": null,
                    "defaultRule": false,
                    "dockerIngress": {
                      "annotations": null,
                      "enabled": false,
                      "host": "example.com",
                      "name": "nexus-docker-ingress"
                    },
                    "dockerSubdomain": false,
                    "enabled": false,
                    "host": "example.com",
                    "name": "nexus-ingress",
                    "tls": {
                      "enabled": false,
                      "secretName": "tlsSecretName"
                    }
                  },
                  "license": {
                    "fileContentsBase64": "your_license_file_contents_in_base_64",
                    "secretName": "nexus-repo-license"
                  },
                  "secret": {
                    "nexusSecret": {
                      "enabled" : false,
                      "name": "nexus-secret.json",
                      "nexusSecretsKeyId": "super_secret_key_id",
                      "secretKeyfileContentsBase64": "secretKeyfileContentsBase64",
                      "mountPath": "/var/nexus-repo-secrets"
                    }
                  },
                  "nexusData": {
                    "pvc": {
                      "accessMode": "ReadWriteOnce",
                      "size": "2Gi"
                    },
                    "storageClass": {
                      "enabled": false,
                      "name": "nexusrepo-storage",
                      "parameters": {},
                      "provisioner": "provisioner",
                      "reclaimPolicy": "Retain",
                      "volumeBindingMode": "WaitForFirstConsumer"
                    },
                    "volumeClaimTemplate": {
                      "enabled": false
                    }
                  },
                  "service": {
                    "docker": {
                      "enabled": false,
                      "name": "nexus-repo-docker-service",
                      "port": 9090,
                      "protocol": "TCP",
                      "targetPort": 9090,
                      "type": "NodePort"
                    },
                    "nexus": {
                      "enabled": false,
                      "name": "nexus-repo-service",
                      "port": 80,
                      "protocol": "TCP",
                      "targetPort": 8081,
                      "type": "NodePort"
                    }
                  },
                  "statefulset": {
                    "container": {
                      "containerPort": 8081,
                      "env": {
                        "clustered": false,
                        "install4jAddVmParams": "-Xms2703m -Xmx2703m",
                        "jdbcUrl": null,
                        "nexusInitialPassword": "nexusabc",
                        "password": "nexus",
                        "user": "nexus",
                        "zeroDowntimeEnabled": false
                      },
                      "imageName": "registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:fbad488ff2e368845473a11b4955992fe826a94811da3abf5cc3c6a33ff3974f",
                      "pullPolicy": "IfNotPresent",
                      "resources": {
                        "limits": {
                          "cpu": 4,
                          "memory": "8Gi"
                        },
                        "requests": {
                          "cpu": 4,
                          "memory": "8Gi"
                        }
                      },
                      "terminationGracePeriod": 120
                    },
                    "imagePullSecrets": {},
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 240,
                      "path": "/",
                      "periodSeconds": 60
                    },
                    "name": "nexusrepo-statefulset",
                    "readinessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 240,
                      "path": "/",
                      "periodSeconds": 60
                    },
                    "replicaCount": 1
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-certified@sha256:99f30aa5ba8469d96de3efe69d1b21e73ff022172fc79144bba496d1119ec8a2
          createdAt: "2023-08-14"
          description: |-
            Nexus Repository is the central source of control to efficiently manage all binaries
            and build artifacts across your DevOps pipeline.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.replaces: 3.80.0-2
          olm.skipRange: <3.84.0-1
          repository: https://github.com/sonatype/operator-nexus-repository
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus Repository
            displayName: NexusRepo
            kind: NexusRepo
            name: nexusrepos.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Repository is the central source of control to efficiently manage all binaries
          and build artifacts across your DevOps pipeline.
          The flow of open source and third-party components into and through an organization
          creates a complex software supply chain.
          Nexus Repository delivers speed, efficiency, and quality to the governance
          and management of all dependencies, libraries, and applications for your DevOps teams.

          ## Core Capabilities

          * **Dependency Management**:
            Improves reliability with repeatable, fast access to secure dependencies
          * **Developer Productivity**:
            Streamline developer workflows by enabling the sharing of components and applications across teams
          * **Supply Chain Performance**:
            Improve speed-to-market and reduced build times with release advanced staging and component tagging
          * **CI/CD Integrations**:
            Increase DevOps scalability with integrations to the most popular build and deployment tools

          Version control systems and package registries do not scale when managing proprietary,
          open source, and third-party components.
          Organizations need a central binary and build artifact repository to manage dependencies
          across the entire software supply chain.

          ## Limitations

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus Repository Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage
          By default, the example Custom Resource starts up a single instance of Nexus Repository in non-Pro mode using an H2 database.

          ### Connect to a PostgreSQL database
          We strongly recommend using Nexus Repository with a PostgreSQL database. To connect a Nexus Repository instance to a PostgreSQL database, do as follows:
          * Update the `license.fileContentsBase64` parameter with a base 64 representation of your Nexus Repository Pro license.
          * Set the `statefulset.container.env.jdbcUrl` parameter to your database url.
          * Set the `statefulset.container.env.user` parameter to your database user name.
          * Set the `statefulset.container.env.password` parameter to your database password.

          ### High Availability (HA) mode
          To start Nexus Repository in HA mode, you must edit the Custom Resource YAML as follows:

          * Update the `license.fileContentsBase64` parameter with a base 64 representation of your Nexus Repository Pro license.
          * Set the `statefulset.container.env.jdbcUrl` parameter to your database url.
          * Set the `statefulset.container.env.user` parameter to your database user name.
          * Set the `statefulset.container.env.password` parameter to your database password.
          * Set the `statefulset.container.env.clustered` parameter to true.
          * Set the `statefulset.replicaCount` parameter to the number of Nexus Repository pods you require in your HA cluster
          * You must use a persistent volume as described in the 'Storage' section below.

          ### Storage
          The default configuration uses an emptyDir volume for storing Nexus Repository logs. However, we strongly recommend that
          you configure dynamic provisioning of persistent storage based on your deployment environment as explained below.

          #### Cloud deployments (AWS/Azure)
          * Ensure the appropriate Container Storage Interface (CSI) driver(s) are installed for your chosen OpenShift cloud deployment.
            Please, refer to OpenShift documentation for details on configuring CSI drivers.
          Note: if you're using Red Hat OpenShift on AWS (ROSA), CSI drivers for dynamically provisioning EBS volumes are installed
          by default; you should see associated storage classes for them in your cluster web console.

          #### On premise deployments
          1. Attach separate disks (i.e. separate from the root disk) to your worker nodes.
          2. Install the Local Storage Operator. Please refer to Open Shift persistent storage documentation:
             https://docs.openshift.com/container-platform/4.13/storage/persistent_storage/persistent_storage_local/persistent-storage-local.html
          and the Local Storage Operator documentation: https://github.com/openshift/local-storage-operator.
          3. Use the Local Storage Operator to automatically create persistent volumes for your chosen storage class name as documented here:
             https://docs.openshift.com/container-platform/4.13/storage/persistent_storage/persistent_storage_local/persistent-storage-local.html
          4. Specify the storage class name from step 3 as the value of the `nexusData.storageClass.name` property and also set `nexusData.volumeClaimTemplate.enabled` to true.

          #### Configuring the operator for dynamic persistent volume provisioning
          * Set the `nexusData.storageClass.name` parameter to a storage class name. This could be one of the default storage classes
            automatically created in your OpenShift cluster (e.g., if you're using ROSA) or one that you would like the operator to create.
          * If you would like to create a dedicated storage class (i.e., you don't want to use the default), then in addition to
            specifying a value for `nexusData.storageClass.name`, you must also set `nexusData.storageClass.enabled` parameter to `true`.
          * Set the `nexusData.volumeClaimTemplate.enabled` parameter to `true`.

          ### Additional configuration
          * Once the operator has created the server instance and it is running, you can further configure Nexus Repository via the NexusRepo custom resource:

          | Parameter                                                   | Description                         | Default                                 |
          | ----------------------------------------------------------- | ----------------------------------  | ----------------------------------------|
          | `statefulset.name`                                          |  The suffix for the statefulset's service name    | nexusrepo-statefulset     |
          | `statefulset.replicaCount`                                  |  The desired number of Nexus Repository pods    | 3                           |
          | `statefulset.container.imageName`                           |  The Nexus repository image registry url    |
          | | | registry.connect.redhat.com/sonatype/nexus-repository-manager:<app version>-ubi-1|
          | `statefulset.container.resources.requests.cpu`              |  The minimum cpu the Nexus repository pod can request    | 4 |
          | `statefulset.container.resources.requests.memory`           |  The minimum memory the Nexus repository pod can request    | 8Gi|
          | `statefulset.container.resources.limits.cpu`                |  The maximum cpu the Nexus repository pod may get.    | 4|
          | `statefulset.container.resources.limits.memory`             |  The maximum memory the Nexus repository pod may get.  | 8Gi|
          | `statefulset.container.containerPort`                       |  The Nexus Repository container's HTTP port  | 8081|
          | `statefulset.container.pullPolicy`                          |  The Nexus Repository docker image pull policy  | IfNotPresent|
          | `statefulset.container.terminationGracePeriod`              |  The time given for the pod to gracefully shut down  | 120 seconds|
          | `statefulset.container.env.clustered`                       |  Controls whether Nexus Repository should start in clustered mode  | False|
          | `statefulset.container.env.install4jAddVmParams`            |  Xmx and Xms settings for JVM  | -Xms2703m -Xmx2703m|
          | `statefulset.container.env.user`                            |  The database user name | nexus (can be overridden via Kubernetes secret. See above)|
          | `statefulset.container.env.password`                        |  The database password  | nexus (can be overridden via Kubernetes secret. See above)|
          | `statefulset.container.env.jdbcUrl`                         |  The database url |
          | | | jdbc:postgresql://your_database_server:5432/nexus (must be overridden via Kubernetes |
          | | | secret object which feeds into env variable variables defined in the operator.yaml . See above)|
          | `statefulset.container.env.nexusInitialPassword`            |  The initial admin password for Nexus Repository UI. This is for first time |
          | | installs only. Subsequent installs/restarts on the same node ignores this setting | nexusabc|
          | `statefulset.livenessProbe.initialDelaySeconds`             |  The initial delay before the Nexus Repository application is pinged for a |
          | | response i.e. to determine whether the application is still alive | 240|
          | `statefulset.livenessProbe.path`                            |  The path to ping the application | |
          | `statefulset.livenessProbe.periodSeconds`                   |  The interval for checking that the application is still alive i.e. responsive, not deadlocked etc | 60|
          | `statefulset.livenessProbe.failureThreshold`                |  The maximum number of times the application ping can fail before the |
          | | application is considered unresponsive. Kubernetes will attempt to restart the application's pod in this case | 6|
          | `statefulset.readinessProbe.initialDelaySeconds`            |  The initial delay before the Nexus Repository application is checked |
          | | whether it is ready to receive traffic | 240|
          | `statefulset.readinessProbe.path`                           |  The path where the Nexus Repository application is checked for traffic readiness | |
          | `statefulset.readinessProbe.periodSeconds`                  |  The interval for checking the Nexus Repository application for traffic readiness | 60|
          | `statefulset.readinessProbe.failureThreshold`               |  The maximum number of times to check the Nexus Repository application for traffic readiness | 6|
          | `statefulset.imagePullSecrets`                              |  The pull secret for private image registries | {}|
          | `ingress.name`                                              |  The name of an Ingress object which listens for requests to Nexus Repository on port 80 | nexus-ingress|
          | `ingress.enabled`                                           |  Whether or not to create the Ingress | false|
          | `ingress.host`                                              |  Ingress host | "example.com"|
          | `ingress.defaultRule`                                       |  Whether or not to add a default rule for the Nexus Repository |
          | | Ingress which forwards traffic to a Service object  | false|
          | `ingress.additionalRules`                                   |  Additional rules to add to the ingress | null|
          | `ingress.tls.enabled`                                       |  Whether or not TLS should be enabled for the Ingress | false|
          | `ingress.tls.secretName`                                    | The name of a Secret object in which to store the TLS secret | tlsSecretName|
          | `ingress.dockerSubdomain`                                   |  Whether or not to add rules for docker subdomains | false|
          | `ingress.annotations`                                       |  Annotations for the Ingress object | null |
          | `license.fileContentsBase64`                                |  Your Base64 encoded Nexus Repository license | your_license_file_contents_in_base_64 |
          | `license.secretName`                                        |  The name of a secret object in which to store your |
          | | Base64-encoded Nexus Repository license | nexus-repo-license|
          | `ingress.dockerIngress.name`                                |  The name of an Ingress object which listens for requests to an |
          | | existing docker repository in Nexus Repository | nexus-docker-ingress|
          | `ingress.dockerIngress.enabled`                             |  Whether or not to create an Ingress for a docker repository | false|
          | `ingress.dockerIngress.host`                                |  The docker Ingress's host | "example.com"|
          | `ingress.dockerIngress.annotations`                         |  Annotations for the docker Ingress object | null |
          | `nexusData.pvc.size`                                        |  The volume size to request for storing Nexus logs | 2Gi |
          | `nexusData.pvc.accessMode`                                  |  The persistent volume claim access mode | ReadWriteOnce |
          | `nexusData.storageClass.name`                               |  The name of the storage class to use for dynamically requesting that |
          | | storage be provisioned when using dynamic provisioning or to randomly select a pre-created persisted volume when using the Local Storage Operator. | nexusrepo-storage |
          | `nexusData.storageClass.reclaimPolicy`                      |  The reclaim policy for persistent volume's created by the storage class | Retain |
          | `nexusData.storageClass.enabled`                            |  Whether or not a storage class object should be created. | false |
          | `nexusData.storageClass.volumeBindingMode`                  |  It controls when volume binding and dynamic provisioning should occur | WaitForFirstConsumer |
          | `nexusData.storageClass.provisioner`                        |  The name of the volume provisioner | provisioner |
          | `nexusData.storageClass.parameters`                         |  The additional parameters for the storage class | {} |
          | `nexusData.volumeClaimTemplate.enabled`                           |  Whether or not volume should be requested using a claim template. |
          | | You should set this property to true when using dynamic volume provisioning (cloud deployments) |
          | | or if you're using Local Storage Operator to automatically create persistent volumes for locally attached disks (i.e. not root disk) | false |
          | `license.secretName | A name for the Kubernetes secret object which stores the base64 representation of your |
          | | Nexus Repository license specified via `license.fileContentsBase64` is stored | nexus-repo-license |
          | `license.fileContentsBase64 | A base64 representation of your Nexus Repository license | your_license_file_contents_in_base_64 |
          | `service.nexus.name`                                              |  The name of a kubernetes Service Object which allows access to |
          | | Nexus Repository's main HTTP port | nexus-repo-service |
          | `service.nexus.enabled`                                           |  Whether or not to create the Service object | false|
          | `service.nexus.type`                                              |  The type of the Kubernetes Service | "NodePort"|
          | `service.nexus.protocol`                                       |  The protocol  | TCP |
          | `service.nexus.port`                                   |  The port to listen for incoming requests | 80|
          | `service.nexus.targetPort`                                   |  The port to forward requests to | 8081 |
          | `service.docker.name`                                              |  The name of a kubernetes Service Object which allows access to a |
          | | docker repository within Nexus Repository | nexus-repo-docker-service |
          | `service.docker.enabled`                                           |  Whether or not to create a Kubernetes Service object for a given |
          | | docker repository within Nexus Repository | false |
          | `service.docker.type`                                              |  The type of the Kubernetes Service | "NodePort"|
          | `service.docker.protocol`                                       |  The protocol  | TCP |
          | `service.docker.port`                                   |  The port to listen for incoming requests | 9090 |
          | `service.docker.targetPort`                                   |  The port to forward requests to | 9090 |
        displayName: Nexus Repository HA Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - repository
        - sonatype
        links:
        - name: Nexus Repository Manager
          url: https://www.sonatype.com/product-nexus-repository
        maintainers:
        - email: support@sonatype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-certified@sha256:99f30aa5ba8469d96de3efe69d1b21e73ff022172fc79144bba496d1119ec8a2
        - registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:fbad488ff2e368845473a11b4955992fe826a94811da3abf5cc3c6a33ff3974f
        version: 3.84.0-1
      entries:
      - name: nexus-repository-ha-operator-certified.v3.84.0-1
        version: 3.84.0-1
      name: stable
    defaultChannel: stable
    packageName: nexus-repository-ha-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-kubernetes
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-kubernetes.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "ClusterMongoDBRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-enterprise"
                  },
                  "name": "clustermongodbrole-sample"
                },
                "spec": {
                  "db": "admin",
                  "role": "rootMonitor",
                  "roles": [
                    {
                      "db": "admin",
                      "role": "root"
                    },
                    {
                      "db": "admin",
                      "role": "clusterMonitor"
                    }
                  ]
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "my-replica-set"
                },
                "spec": {
                  "credentials": "my-credentials",
                  "members": 3,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "sample-sharded-cluster"
                },
                "spec": {
                  "configServerCount": 3,
                  "credentials": "my-credentials",
                  "mongodsPerShardCount": 3,
                  "mongosCount": 2,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "shardCount": 1,
                  "type": "ShardedCluster",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBMultiCluster",
                "metadata": {
                  "name": "multi-replica-set"
                },
                "spec": {
                  "clusterSpecList": [
                    {
                      "clusterName": "e2e.cluster1.mongokubernetes.com",
                      "members": 2
                    },
                    {
                      "clusterName": "e2e.cluster2.mongokubernetes.com",
                      "members": 1
                    },
                    {
                      "clusterName": "e2e.cluster3.mongokubernetes.com",
                      "members": 2
                    }
                  ],
                  "credentials": "my-credentials",
                  "duplicateServiceObjects": false,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBOpsManager",
                "metadata": {
                  "name": "ops-manager"
                },
                "spec": {
                  "adminCredentials": "ops-manager-admin",
                  "applicationDatabase": {
                    "members": 3,
                    "podSpec": {
                      "cpu": 1
                    }
                  },
                  "configuration": {
                    "mms.fromEmailAddr": "admin@thecompany.com"
                  },
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "version": "6.0.3"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBUser",
                "metadata": {
                  "name": "my-replica-set-x509-user"
                },
                "spec": {
                  "db": "$external",
                  "mongodbResourceRef": {
                    "name": "my-replica-set"
                  },
                  "roles": [
                    {
                      "db": "admin",
                      "name": "dbOwner"
                    }
                  ],
                  "username": "CN=my-replica-set-x509-user,OU=cloud,O=MongoDB,L=New York,ST=New York,C=US"
                },
                "status": {
                  "phase": "Updated"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/mongodb/mongodb-kubernetes@sha256:745cc2c6bb893c1cb1e75d79b6f3bfc7b125a6a62f7431bdc43cf6cacd373998
          createdAt: "2025-09-16T14:56:17Z"
          description: The MongoDB Controllers for Kubernetes enable easy deploys
            of MongoDB into Kubernetes clusters, using our management, monitoring
            and backup platforms, Ops Manager and Cloud Manager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mongodb/mongodb-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterMongoDBRole
            name: clustermongodbroles.mongodb.com
            version: v1
          - description: MongoDB Deployment
            displayName: MongoDB Deployment
            kind: MongoDB
            name: mongodb.mongodb.com
            version: v1
          - kind: MongoDBCommunity
            name: mongodbcommunity.mongodbcommunity.mongodb.com
            version: v1
          - description: MongoDB Multi Deployment
            displayName: MongoDB Multi Deployment
            kind: MongoDBMultiCluster
            name: mongodbmulticluster.mongodb.com
            version: v1
          - kind: MongoDBSearch
            name: mongodbsearch.mongodb.com
            version: v1
          - description: MongoDB x509 User
            displayName: MongoDB User
            kind: MongoDBUser
            name: mongodbusers.mongodb.com
            version: v1
          - description: MongoDB Ops Manager
            displayName: MongoDB Ops Manager
            kind: MongoDBOpsManager
            name: opsmanagers.mongodb.com
            version: v1
        description: |
          MongoDB Controllers for Kubernetes (MCK) is an open-source project. It represents a merge of the previous
          MongoDB Community Operator (MCO) and MongoDB Enterprise Kubernetes Operator (MEKO).

          This release brings MongoDB Community and Enterprise editions together under a single, unified operator,
          making it easier to manage, scale, and upgrade your deployments. While the first version simply brings
          the capabilities of both into a single Operator, future changes will build on this to more closely align
          how Community and Enterprise are managed in Kubernetes, to offer an even more seamless and streamlined experience.
          As an open-source project, it now allows for community contributions, helping drive quicker bug fixes and ongoing innovation.
          The MongoDB Controllers for Kubernetes enable easy deploys of MongoDB
          into Kubernetes clusters, using our management, monitoring and backup
          platforms, Ops Manager and Cloud Manager.

          ## Before You Start

          To start using the operator you''ll need an account in MongoDB Cloud Manager or
          a MongoDB Ops Manager deployment.

          * [Create a Secret with your OpsManager API key](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-operator-credentials/#procedure)

          * [Create a ConfigMap with your OpsManager project ID and URL](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-project-using-configmap/)

          By installing this integration, you will be able to deploy MongoDB instances
          with a single simple command.

          ## Required Parameters

          * `opsManager` or `cloudManager` - Enter the name of the ConfigMap containing project information
          * `credentials` - Enter the name of the Secret containing your OpsManager credentials
          * `type` - Enter MongoDB Deployment Types ("Standalone", "ReplicaSet", "ShardedCluster"

          ## Supported MongoDB Deployment Types ##

          * Standalone: An instance of mongod that is running as a single server and
          not as part of a replica set, this is, it does not do any kind of
          replication.

          * Replica Set: A replica set in MongoDB is a group of mongod processes that
          maintain the same data set. Replica sets provide redundancy and high
          availability, and are the basis for all production deployments. This section
          introduces replication in MongoDB as well as the components and architecture
          of replica sets. The section also provides tutorials for common tasks
          related to replica sets.

          * Sharded Cluster: The set of nodes comprising a sharded MongoDB deployment.
          A sharded cluster consists of config servers, shards, and one or more mongos
          routing processes. Sharding is a A database architecture that partitions
          data by key ranges and distributes the data among two or more database
          instances. Sharding enables horizontal scaling.

          ## Requirements for deploying MongoDB OpsManager

          * In order to deploy resources of type MongoDB OpsManager, you will need to
          create a secret containing the [credentials](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/plan-om-resource/#om-rsrc-prereqs)
          for the Global Owner user

          ## Security ##

          The operator can enable TLS for all traffic between servers and also between
          clients and servers. Before enabling `security.tls.enabled` to `true` you
          should create your certificates.  or you can leave the operator to create all
          the certificates for you. The operator ability to create certs is been
          deprecated due to Kubernetes API changes.

          For more information, please read the official MongoDB
          Kubernetes Operator  [docs](https://docs.mongodb.com/kubernetes-operator/stable/).
        displayName: MongoDB Controllers for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mongodb
        - databases
        - nosql
        links:
        - name: Documentation
          url: https://docs.opsmanager.mongodb.com/current/tutorial/install-k8s-operator/index.html
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: stable
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f2d9a3264a8814f33143b73c56624572fb80afeab184133d3a901713a31c1a87
        - quay.io/mongodb/mongodb-enterprise-server@sha256:376ad7e5fe42651d946cba5705c0eee44aa636c1b4ff258210e4ce2fe4205b4d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:fb8397cbcdbf7c33f6ad142492792771127f0d9a8c46388f350b6927405ee590
        - quay.io/mongodb/mongodb-agent@sha256:d8a2ef79d767f03162ec57fe028f9072294b292d376af8eb97d2967eb6d7047b
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:09788976bdc1327acbf18cedf8efd5fca2159e43b506faa40e03ee42ad94928a
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:cd869ff7d661059fad7593e30141d74e8cc751295b1579577f7a853396eb5d0b
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:78f11c9cd5e63f2b0e10c67dc1491f4ce1279fa90178c108814725414defab9a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:15e31285c8949a4666f7702df53416b0905d76945a28575a6028a876cc8a6f0f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1a77851485ad3b2566e765b4bb051ff0ddfa48703492cef6c6db021dde84d87f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:146c4f5dc443f94f54d72808b4e4df93bb904e23589b222d17700166780776b4
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa8683500e8d3632c4588bbcdddb4e3186a01f48aefe1b6d95f8860dae8b6c76
        - quay.io/mongodb/mongodb-kubernetes-init-appdb@sha256:afb0d463304fc0bfb3d8dade91459414daac6bf6fcf742347a2012cbc9fea354
        - quay.io/mongodb/mongodb-search@sha256:7149a7b6f1ee84646d3972837920f80fc5db49503c598406ce92e2f794bd5f0d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:85c1d669c073caf9cbaac815e4c8ea2ca18b71deefa03bf84cd6d8b8ebd53e12
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:5d299820bb903f0cdfdf2626a3c44099d2eaee16188cd8234d2ff3a57259b3a7
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5145c97d2281aabcd8538ff515c3833aa48acb66b0b739fb1881120c1985d88a
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:5e3982493e85ac1e40536758c3145b7c2fff9a7b4be9d053ee41201363a89920
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b7764c316eeacffd3f9ddae62d9fd2a74eb58484d2cfd5021e460095d604764f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d49417ea85ee1ddb84429552a5aac87a82c2843904d8c0300a43b657b7a49e0e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:19925e1297ab652db59c35f53e7b7c8b2b277f8ecdc4215682961aaf2b3e924e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7e5361c2a1604dec3fc671d629806fe6ddc6df524b69a441b86933a4b30093
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:adeab0dc766a27138db7df5f37cbb1da860c14179ef631b2e001e5bb875325e4
        - quay.io/mongodb/mongodb-agent@sha256:a8358ce465adf463d8e73916d7944765d7aaf2b0ed10bd3b8119a87fc9080667
        - quay.io/mongodb/mongodb-agent@sha256:6e826bbe7dd812c101464e75295370c4b10e4e9c31477db89cfceb96759853e9
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:85b4e954c9c4bbe1019774bad1e05e9c8dc175648b01d3badf0b64143f182019
        - quay.io/mongodb/mongodb-enterprise-server@sha256:bfbcfcfc5f0de22e20520b2d96f826638e424adb028a909ccc192013d11f5833
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a86c4d8a0b9ca3c20cc524809676eb03a944ada43378ec98c107cd72cc64700f
        - quay.io/mongodb/mongodb-agent@sha256:ee9f3906d893004bb2aaaa6412646f58288b0684502b9e90f33cafadcfce4349
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0fde236fe26d5b20210d6f246aadacd7845874712def516cbe2eec11d79d5181
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:702f529cb9d8d9042e54263f8bc37b40c062ac2695efc3e7150819779886e323
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f1e72679b2468aed2f3c3106568327911e6594c9ee15f7348e953eaa0d8fd44a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7885ab6e88214f9289d8babf5189136dde0dd4c84a15a5481ccd3ef512b2e66f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:d81eac2129b6ab8346be1749c56ea67f6da64d4ee0fb9f3ff4171ba0f61918eb
        - quay.io/mongodb/mongodb-agent@sha256:fa30370a75e8e4ac648650ceb39a7a259a754aab8a48d7e2f2801cd1b777e91b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:67238129df09d33ee4746ff1acb5515d699a29d9f42350eae3f3f69ccb7fbf54
        - quay.io/mongodb/mongodb-enterprise-server@sha256:523566f44cbb6bbf0cd12b1adae939a883f2254c78dc4b8c9d0098581337b1db
        - quay.io/mongodb/mongodb-kubernetes-database@sha256:f8c981a6223f70d7c6bbd30c503fc8b2497f351830b2b74c30e6f81863d78d2f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:33bbd29b74f95baad522b380d70bcc9f933a3fee57642c37b09376c86179c40f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c752a7f0726d21d55f204a359b85fa29cd727a5fe571b37289ae3ea52d6206b0
        - quay.io/mongodb/mongodb-agent@sha256:da77dcb60b927bf6059c2c9bf008f8c3c0eec45cf7d1b40c34a0c263e8cc18b8
        - quay.io/mongodb/mongodb-kubernetes-init-database@sha256:3ea9106c0e35c015bc8f15657c76d974bb1cadb9220dc58be2076c9699f26fb7
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b00ac773623e1bf422b47e6d783263d2d08737865abe9abbe106d86d1ea25165
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9bef77e905dadd1e61d0439dd53b21ed1f3606e49affd9ffc4ba8156578e8f0f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:37af0913742b6de339c7e664b30f35dcef04506529e4a09a9bbbdd57643f7149
        - quay.io/mongodb/mongodb-agent@sha256:2d5652d0e263aa94554e7a068c93e4fd94b84ec223f077361d030d172bc93f67
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8445d59d491a6d3890ea9c8b421214bcd0fdcfbc420ed40497a790b79f0bc89e
        - quay.io/mongodb/mongodb-agent@sha256:b387b7458f5640f907c9074fcb08ae4d4785be31d4bb65f2d7c459cd6762b2d6
        - quay.io/mongodb/mongodb-enterprise-server@sha256:720056b985c1a16d48b2aecb7a88a15cba3d8186e55b3c6c24729699a85c8232
        - quay.io/mongodb/mongodb-agent@sha256:b422829aac63f068aed4a6948fd14cbd3c3bfad6d47ab76523f6839bf20b0203
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c8aec83b76dfecc05b8f5d8f40fd61b24980c0287a07294f75f321c2d42c32b4
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa300abf739e26350a258f643fa8e57aedacd041055c7b923bd22b21ef554cdb
        - quay.io/mongodb/mongodb-agent@sha256:3fa5bcfda3228e0ea880434a60ced829e3872e74763b347636c18475d866d0e1
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3fe62f3ad52707ccf0e9993c69891e44c569a4b4e1a3244cf60e3b5fdde57bd2
        - quay.io/mongodb/mongodb-agent@sha256:6c23febaa9dd52bc29ac1e52d087fc4b2b47c578f4d6ac3a1674bb750a38cb07
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3add8b68bfbaa42596fe5dbbee4623ba5ff1a826b18bb71301525adf574dcb94
        - quay.io/mongodb/mongodb-enterprise-server@sha256:655896bac9e82de8e9729d59a36c3a31f66235f22b93449260a2ccb7709de084
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b9ab7147b59b62e40a69c6400f87d0c7faecfa5462bed6e68ca653b00e50b887
        - quay.io/mongodb/mongodb-enterprise-server@sha256:e7aeedb752dda77cf0e49df3430843b5a2bba64c8404628a983f46640d316808
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ef732e3eea91cc86caf2eabedc0c9115ed235a8d204f33e74bbf090a0b0259fb
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:895dbdfb87ddc2e9693671c7108a31ab28e16fec2e5be6d046bdd2e30319eade
        - quay.io/mongodb/mongodb-enterprise-server@sha256:454b1585c342375dea643b10ead688d0f62347841f5f30059fcd4f92e3297ee9
        - quay.io/mongodb/mongodb-kubernetes-init-ops-manager@sha256:93c4633cbc84e9bfa3563771ac2efbf5e6515f380dec2c085494cb8326546377
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9a3a6496be0d2d446e6d654eb56386c7726e0cd2d7232f6dcec579923edf0014
        - quay.io/mongodb/mongodb-agent@sha256:5b064c470335df840b050b714b28b75b6061444843250f1eb9215cb2b8c1a149
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7a93a0276531ff9be4c90bb8fe8d104e0a9e930c29792aafe03cc6a76a9fa89c
        - quay.io/mongodb/mongodb-enterprise-server@sha256:33a30d11969eb21952681108092fe7f916e25c8e1de43e6a4e40e126025a2bec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1d418c65d82ccad205b7889be4ba161d3cbfa8b1fb75b21c8fcf02de38fad9ca
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d323c1bcd14118cc537ac282996ea5a58756d20b6b73f1eec0efa8710d246077
        - quay.io/mongodb/mongodb-enterprise-server@sha256:172e338a43df01ee5408d8d1e3cb038d2c55e77761c1a2ac6ed37959d1525140
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7266b405ebb2dd24a9f7c12d3921024f547c3d8c01831e6bcfc1d551d63be1
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a48a4d5c275fd669a27469649fb39ea38c728dbcf9f0a9da4e7cf27a2ce912ed
        - quay.io/mongodb/mongodb-enterprise-server@sha256:11ab324698f919f84a5dab5a964d93fe63f98a48146bb3b86974cba7606d30ec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c48631742ad1fec8a9ee745fc3cc9ca1fe7d86ef7fa68ad8e8337c7a4566dc87
        - quay.io/mongodb/mongodb-agent@sha256:89819704d70a3b6b4d2358745a22060ef740b33ad5c6debf8e6ac4d4a6fd63ba
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:a4590af027182305913f0de1a2e37ad8612bd73b8b31f4df11068b80089021ff
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b8aa1edbc5fb02af312a863bc7db8535a282dfd692a07b1e321e5a1acbfeb6a1
        - quay.io/mongodb/mongodb-enterprise-server@sha256:dfd4277f0c16ed476ac7de51894739e8c8d13a65ea5024b9a97b3dce0f6c6975
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d2d8b59c1f39c41951e7878963f8496eb494688720366f8edc7060887cdbe6a0
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:3bbd3d7fd45d107621e7770f2b4c2a758e148cd472e0ee21719442f71bf6ade0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3316015ac0699d66ac73f714e5dc0784738654e0349e80b1e1f4e953b3c88964
        - quay.io/mongodb/mongodb-agent@sha256:5f5031e4c2a3729f618d026b0a7f34fd4509ea73d5a4e3068b551041dc291583
        - quay.io/mongodb/mongodb-enterprise-server@sha256:df6081325ea2fe01d1b27d0c70005dec8232aaf51a43d91661b722b8a1761263
        - quay.io/mongodb/mongodb-agent@sha256:6394391f78da5e3218bb2c694e7ec9f6682fd17a6dd7e7250b0c0b5e1729796d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5c8553e5ba115657fb878c0fe08704af9fd5decec572a87d89d2c8c156ffc9f9
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:fa4aa97767bd19cf3711a74bf21ecb9cbbc25fed6b6b09a2eb241ebd3fbc8f3a
        - quay.io/mongodb/mongodb-agent@sha256:6cc4e953df3f09d793f0dbc6e57e55a3c5277ef63177a6d8f691a41227facf64
        - quay.io/mongodb/mongodb-kubernetes@sha256:745cc2c6bb893c1cb1e75d79b6f3bfc7b125a6a62f7431bdc43cf6cacd373998
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d21c65780cadc9d8668d2c28935d492625d6993f13ba562722200a819281cbcc
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:439d6fefb469613432e2d05149dfc96c32b16469d743835d5f402a2ada8be8cc
        - quay.io/mongodb/mongodb-agent@sha256:af3de4f631d0b00cae7f7e80c3900378ba7e449e01830bf9dc643b873d2e6d33
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:f71b027fccf649f18f63f8df7a59b1bac421a2cdab6712e0c985a93c08ed970a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:778f9e29e36d6687cd17dd95ec1f8127d2a732bde1a2dd5821212eb55c1c998a
        version: 1.4.0
      entries:
      - name: mongodb-kubernetes.v1.4.0
        version: 1.4.0
      - name: mongodb-kubernetes.v1.3.0
        version: 1.3.0
      - name: mongodb-kubernetes.v1.2.0
        version: 1.2.0
      - name: mongodb-kubernetes.v1.1.0
        version: 1.1.0
      - name: mongodb-kubernetes.v1.0.1
        version: 1.0.1
      - name: mongodb-kubernetes.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: namespace-configuration-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: namespace-configuration-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupConfig",
                "metadata": {
                  "name": "test-groupconfig"
                },
                "spec": {
                  "labelSelector": {
                    "matchLabels": {
                      "type": "team"
                    }
                  },
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-dev\n  labels:\n    group: {{ .Name }}  \n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-qa\n  labels:\n    group: {{ .Name }}          \n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-prod\n  labels:\n    group: {{ .Name }}                           \n"
                    },
                    {
                      "objectTemplate": "apiVersion: quota.openshift.io/v1\nkind: ClusterResourceQuota\nmetadata:\n  name: {{ .Name }}-quota\nspec:\n  quota:\n    hard:\n      pods: \"4\" \n      requests.cpu: \"1\" \n      requests.memory: 1Gi \n      requests.ephemeral-storage: 2Gi \n      limits.cpu: \"2\" \n      limits.memory: 2Gi \n      limits.ephemeral-storage: 4Gi\n  selector:\n    labels:\n      matchLabels:\n        group: {{ .Name }}\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NamespaceConfig",
                "metadata": {
                  "name": "test-namespaceconfig"
                },
                "spec": {
                  "labelSelector": {
                    "matchLabels": {
                      "type": "secure"
                    }
                  },
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: ResourceQuota\nmetadata:\n  name: standard-sandbox\n  namespace: {{ .Name }}\nspec:\n  hard:\n    pods: \"4\" \n    requests.cpu: \"1\" \n    requests.memory: 1Gi \n    requests.ephemeral-storage: 2Gi \n    limits.cpu: \"2\" \n    limits.memory: 2Gi \n    limits.ephemeral-storage: 4Gi\n"
                    },
                    {
                      "objectTemplate": "kind: EgressNetworkPolicy\napiVersion: network.openshift.io/v1\nmetadata:\n  name: air-gapped\n  namespace: {{ .Name }}  \nspec:\n  egress:\n  - type: Deny\n    to:\n      cidrSelector: 0.0.0.0/0\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "UserConfig",
                "metadata": {
                  "name": "test-user-config"
                },
                "spec": {
                  "providerName": "my-provider",
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-sandbox\n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: ResourceQuota\nmetadata:\n  name: standard-sandbox\n  namespace: {{ .Name }}-sandbox\nspec:\n  hard:\n    pods: \"4\" \n    requests.cpu: \"1\" \n    requests.memory: 1Gi \n    requests.ephemeral-storage: 2Gi \n    limits.cpu: \"2\" \n    limits.memory: 2Gi \n    limits.ephemeral-storage: 4Gi\n"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/namespace-configuration-operator@sha256:49ed7d6155342adaa2b12fd80c6761c3081d8e6149d187cb7ff91a247cdf2e7a
          createdAt: "2023-11-22T15:58:50Z"
          description: This operator provides a facility to define and enforce namespace
            configurations
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: namespace-configuration-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/namespace-configuration-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GroupConfig is the Schema for the groupconfigs API
            displayName: Group Config
            kind: GroupConfig
            name: groupconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: NamespaceConfig is the Schema for the namespaceconfigs API
            displayName: Namespace Config
            kind: NamespaceConfig
            name: namespaceconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: UserConfig is the Schema for the userconfigs API
            displayName: User Config
            kind: UserConfig
            name: userconfigs.redhatcop.redhat.io
            version: v1alpha1
        description: "The `namespace-configuration-operator` helps keeping configurations
          related to Users, Groups and Namespaces aligned with one of more policies
          specified as a CRs. The purpose is to provide the foundational building
          block to create an end-to-end onboarding process. \nBy onboarding process
          we mean all the provisioning steps needed to a developer team working on
          one or more applications to OpenShift.\n This usually involves configuring
          resources such as: Groups, RoleBindings, Namespaces, ResourceQuotas, NetworkPolicies,
          EgressNetworkPolicies, etc.... . Depending on the specific environment the
          list could continue.\nNaturally such a process should be as automatic and
          scalable as possible.\n\nWith the namespace-configuration-operator one can
          create rules that will react to the creation of Users, Groups and Namespace
          and will create and enforce a set of resources.\n\nHere are some examples
          of the type of onboarding processes that one could support:\n\n1. [developer
          sandbox](https://github.com/redhat-cop/namespace-configuration-operator/blob/master/examples/user-sandbox/readme.md)\n
          2. [team onboarding](https://github.com/redhat-cop/namespace-configuration-operator//blob/master/examples/team-onboarding/readme.md)
          with support of the entire SDLC in a multitentant environment.\n\nPolicies
          can be expressed with the following CRDs:\n\n| Watched Resource | CRD |\n|--|--|\n
          | Groups | [GroupConfig](#GroupConfig) |\n| Users | [UserConfig](#UserConfig)
          |\n| Namespace | [NamespaceConfig](#NamespaceConfig) |\n\nThese CRDs all
          share some commonalities:\n\n1. Templated Resources\n2. List of ignored
          jason path\n \n### Templated Resources\n\nEach has a parameter called `templatedResources`,
          which is an array. Each element of the array has two fields `objectTemplate`
          and `excludedPaths` (see below).\n\nThe `objectTemplate` field must contains
          a [go template](https://golang.org/pkg/text/template/) that resolves to
          a single API Resource expressed in `yaml`. The template is merged with the
          object selected by the CR. For example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: UserConfig\nmetadata:\n  name: test-user-config\nspec:\n  provider:
          corp-ldap\n  templates:\n  - objectTemplate: |\n      apiVersion: v1\n      kind:
          Namespace\n      metadata:\n        name: {{ .Name }}-sandbox\n```\n\nThis
          create a rule in which every time a user from the `corp-ldap` provider is
          created, a namespace called `<username>-sandbox` is also created.\n\n###
          Excluded Paths\n\nThe logic of the `namespace-configuration-operator` is
          to enforce that the resources resolved by processing the templates \"stays
          in place\". In other words if those resources are changed and/or deleted
          they will be reset by the operator.\nBut there are situations in which at
          least part of a resource is allowed to change. Common use cases are: annotations
          and in general the metadata section of a resource can be updated by the
          various operators watching that resources. The status field is often updated
          by the main operator managing that resources. Finally, when applicable the
          `spec.replicas` field should also be allowed to change.\n\nTo handle special
          use case, one can also specify additional *jsonpaths* that should be ignored
          when comparing the desired resource and the current resource and making
          a decision on whether that resource should be reset.\n\nThe following paths
          are always included:\n \n1. `.metadata`\n2. `.status`\n3. `.spec.replicas`\n\n##
          NamespaceConfig\n\n The `NamespaceConfig` CR allows specifying one or more
          objects that will be created in the selected namespaces.\n\nNamespaces can
          be selected by labels or annotations via a label selector for example:\n\n```yaml\napiVersion:
          redhatcop.redhat.io/v1alpha1\n kind: NamespaceConfig\nmetadata:\n  name:
          small-namespace\n  selector:\n    matchLabels:\n      size: small  \n  resources:\n
          \ - apiVersion: v1\n    kind: ResourceQuota\n    metadata:\n      name:
          small-size  \n    spec:\n      hard:\n        requests.cpu: \"4\"\n        requests.memory:
          \"2Gi\"\n```\n\nHere is a `NamespaceConfig` object using a `matchExpressions`
          selector:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n kind: NamespaceConfig\nmetadata:\n
          \ name: tier-config\nspec:\n  annotationSelector:\n    matchExpressions:\n
          \     - {key: tier, operator: In, values: [gold,silver]}\n ```\n\nAlthough
          not enforced by the operator the general expectation is that the NamespaceConfig
          CR will be used to create objects inside the selected namespace.\n \nExamples
          of NamespaceConfig usages can be found [here](https://github.com/redhat-cop/namespace-configuration-operator/blob/master/examples/namespace-config/readme.md)\n
          \n## GroupConfig\n\nThe `GroupConfig` CR allows specifying one or more objects
          that will be created in the selected Group.\nGroups can be selected by labels
          or annotations via a label selector, similarly to the `NamespaceConfig`.\n\nOften
          groups are created in OpenShift by a job that synchronizes an Identity Provider
          with OCP. So the idea is that when new groups are added or deleted the configuration
          in OpenShift will adapt automatically.\n\nAlthough not enforced by the operator,
          GroupConfig are expected to create cluster-scoped resources like Namespaces,
          ClusterResourceQuotas and potentially some namespaced resources like RoleBindings.\n\n##
          UserConfig\n \nIn OpenShift an external user is defined by two entities:
          Users and Identities. There is a relationship of on to many between Users
          and Identities. Given one user, there can be one Identity per authentication
          mechanism.\n\nThe `UserConfig` CR allows specifying one or more objects
          that will be created in the selected User.\nUsers can be selected by label
          or annotation like `NamespaceConfig` and `UserConfig`.\nUSers can also be
          selected by provider name (the name of the authentication mechanism) and
          identity extra field.\n\nHere is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          UserConfig\nmetadata:\n  name: test-user-config\nspec:\n  providerName:
          okta-provider\n  identityExtraFieldSelector:\n    matchLabels:\n      sandbox_enabled:
          \"true\"\n  templates:\n  - objectTemplate: |\n      apiVersion: v1\n      kind:
          Namespace\n      metadata:\n        name: {{ .Name }}-sandbox\n ```\n\nUser
          will be selected by this `UserConfig` only if they login via the *okta-provider*
          and if the extra field was populate with the label `sandbox_enabled: \"true\"
          `. Note that not all authentication provider allow populating the extra
          fields in the Identity object.\n"
        displayName: Namespace Configuration Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - namespace
        - user
        - group
        - configuration
        - policy
        - management
        links:
        - name: Namespace Configuration Operator
          url: https://github.com/redhat-cop/namespace-configuration-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/namespace-configuration-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/namespace-configuration-operator@sha256:49ed7d6155342adaa2b12fd80c6761c3081d8e6149d187cb7ff91a247cdf2e7a
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.2.6
      entries:
      - name: namespace-configuration-operator.v1.2.6
        version: 1.2.6
      - name: namespace-configuration-operator.v1.2.5
        version: 1.2.5
      - name: namespace-configuration-operator.v1.2.4
        version: 1.2.4
      - name: namespace-configuration-operator.v1.2.3
        version: 1.2.3
      - name: namespace-configuration-operator.v1.2.2
        version: 1.2.2
      - name: namespace-configuration-operator.v1.2.1
        version: 1.2.1
      - name: namespace-configuration-operator.v1.2.0
        version: 1.2.0
      - name: namespace-configuration-operator.v1.1.0
        version: 1.1.0
      - name: namespace-configuration-operator.v1.0.7
        version: 1.0.7
      - name: namespace-configuration-operator.v1.0.6
        version: 1.0.6
      - name: namespace-configuration-operator.v1.0.5
        version: 1.0.5
      - name: namespace-configuration-operator.v1.0.4
        version: 1.0.4
      - name: namespace-configuration-operator.v1.0.3
        version: 1.0.3
      - name: namespace-configuration-operator.v1.0.2
        version: 1.0.2
      - name: namespace-configuration-operator.v1.0.1
        version: 1.0.1
      - name: namespace-configuration-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: namespace-configuration-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: geunwookim
      provider-url: ""
    name: ncn-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ncn-operator.v0.3.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-1"
                },
                "spec": null
              },
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "label-1"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
          - description: Label is the Schema for the labels API
            displayName: Label
            kind: Label
            name: labels.cxl.antline.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: jwahn@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Antline
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/rasen708/ncn-operator:v0.4.0
        version: 0.3.18
      entries:
      - name: ncn-operator.v0.3.18
        version: 0.3.18
      - name: ncn-operator.v0.3.14
        version: 0.3.14
      name: alpha
    - currentCSV: ncn-operator.v0.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-1"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-19T01:21:08Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: jwahn@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        provider:
          name: Antline
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/rasen708/ncn-operator:v0.2.2
        version: 0.2.2
      entries:
      - name: ncn-operator.v0.2.2
        version: 0.2.2
      - name: ncn-operator.v0.2.1
        version: 0.2.1
      - name: ncn-operator.v0.2.0
        version: 0.2.0
      name: beta
    - currentCSV: ncn-operator.v0.9.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cmmd.gw.com/v1",
                "kind": "CMMD",
                "metadata": {
                  "name": "my-app",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "allocateMode": "auto",
                  "enable": true,
                  "payload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "metadata": {
                      "name": "my-app"
                    },
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "name": "my-app"
                        }
                      },
                      "template": {
                        "metadata": {
                          "labels": {
                            "name": "my-app"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "quay.io/my-account/my-app:latest",
                              "name": "my-app-01"
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "cmmd.gw.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "machineconfiguration.openshift.io/role": "cmmd"
                  },
                  "name": "cmmd-node-label",
                  "namespace": "socmmd"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          console.openshift.io/plugins: '["socmmd-console-plugin"]'
          containerImage: quay.io/rasen708/ncn-operator@sha256:a09963745bd9c1419532d42d91237b7496c7a78e6a1c460a8e714ea1a0ce4d45
          createdAt: "2024-08-01T05:45:38Z"
          description: Provides the ncn extend
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "cmmd.gw.com/v1",
              "kind": "Label",
              "metadata": {
                "name": "cmmd-node-label",
                "namespace": "socmmd",
                "labels": {}
              },
              "spec": {}
            }
          operatorframework.io/suggested-namespace: socmmd-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/geunwookim/ncn-operator
          support: geunwookim
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CMMD is the Schema for the cmmds API
            displayName: CMMD
            kind: CMMD
            name: cmmds.cmmd.gw.com
            version: v1
          - description: Label is the Schema for the labels API
            displayName: Node Label
            kind: Label
            name: labels.cmmd.gw.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: geunwookim@gmail.com
          name: geunwookim
        maturity: betav1
        provider:
          name: geunwookim
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/rasen708/ncn-operator@sha256:a09963745bd9c1419532d42d91237b7496c7a78e6a1c460a8e714ea1a0ce4d45
        version: 0.9.8
      entries:
      - name: ncn-operator.v0.9.8
        version: 0.9.8
      - name: ncn-operator.v0.9.2
        version: 0.9.2
      - name: ncn-operator.v0.8.1
        version: 0.8.1
      - name: ncn-operator.v0.5.6
        version: 0.5.6
      - name: ncn-operator.v0.4.3
        version: 0.4.3
      - name: ncn-operator.v0.4.2
        version: 0.4.2
      - name: ncn-operator.v0.3.18
        version: 0.3.18
      - name: ncn-operator.v0.3.14
        version: 0.3.14
      name: betav1
    defaultChannel: betav1
    packageName: ncn-operator
    provider:
      name: geunwookim
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CNCF
      provider-url: ""
    name: etcd
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: etcdoperator.v0.6.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          description: etcd is a distributed key value store providing a reliable
            way to store data across a cluster of machines.
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
        description: |
          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader. Your applications can read and write data into etcd.
          A simple use-case is to store database connection details or feature flags within etcd as key value pairs. These values can be watched, allowing your app to reconfigure itself when they change. Advanced uses take advantage of the consistency guarantees to implement database leader elections or do distributed locking across a cluster of workers.

          _The etcd Open Cloud Service is Public Alpha. The goal before Beta is to fully implement backup features._

          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` or with the API using the automatically generated Kubernetes Service.

          [Read the complete guide to using the etcd Open Cloud Service](https://coreos.com/tectonic/docs/latest/alm/etcd-ocs.html)

          ### Supported Features
          **High availability**
          Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.
          **Automated updates**
          Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.
          **Backups included**
          Coming soon, the ability to schedule backups to happen on or off cluster.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: support@coreos.com
          name: CoreOS, Inc
        maturity: alpha
        provider:
          name: CoreOS, Inc
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:bd944a211eaf8f31da5e6d69e8541e7cada8f16a9f7a5a570b22478997819943
        version: 0.6.1
      entries:
      - name: etcdoperator.v0.6.1
        version: 0.6.1
      name: alpha
    - currentCSV: etcdoperator.v0.9.4-clusterwide
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdCluster",
                "metadata": {
                  "name": "example",
                  "annotations": {
                    "etcd.database.coreos.com/scope": "clusterwide"
                  }
                },
                "spec": {
                  "size": 3,
                  "version": "3.2.13"
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdRestore",
                "metadata": {
                  "name": "example-etcd-cluster-restore"
                },
                "spec": {
                  "etcdCluster": {
                    "name": "example-etcd-cluster"
                  },
                  "backupStorageType": "S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdBackup",
                "metadata": {
                  "name": "example-etcd-cluster-backup"
                },
                "spec": {
                  "etcdEndpoints": ["<etcd-cluster-endpoints>"],
                  "storageType":"S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
          createdAt: "2019-02-28 01:03:00"
          description: Create and maintain highly-available etcd clusters on Kubernetes
          repository: https://github.com/coreos/etcd-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to backup an etcd cluster.
            displayName: etcd Backup
            kind: EtcdBackup
            name: etcdbackups.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to restore an etcd cluster from a backup.
            displayName: etcd Restore
            kind: EtcdRestore
            name: etcdrestores.etcd.database.coreos.com
            version: v1beta2
        description: |
          The etcd Operater creates and maintains highly-available etcd clusters on Kubernetes, allowing engineers to easily deploy and manage etcd clusters for their applications.

          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader.


          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` via port forwarding:

              $ kubectl --namespace default port-forward service/example-client 2379:2379
              $ etcdctl --endpoints http://127.0.0.1:2379 get /

          Or directly to the API using the automatically generated Kubernetes Service:

              $ etcdctl --endpoints http://example-client.default.svc:2379 get /

          Be sure to secure your etcd cluster (see Common Configurations) before exposing it outside of the namespace or cluster.


          ### Supported Features

          * **High availability** - Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.

          * **Automated updates** - Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.

          * **Backups included** - Create etcd backups and restore them through the etcd Operator.

          ### Common Configurations

          * **Configure TLS** - Specify [static TLS certs](https://github.com/coreos/etcd-operator/blob/master/doc/user/cluster_tls.md) as Kubernetes secrets.

          * **Set Node Selector and Affinity** - [Spread your etcd Pods](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-node-selector-and-anti-affinity-across-nodes) across Nodes and availability zones.

          * **Set Resource Limits** - [Set the Kubernetes limit and request](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-resource-requirement) values for your etcd Pods.

          * **Customize Storage** - [Set a custom StorageClass](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#custom-persistentvolumeclaim-definition) that you would like to use.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: etcd-dev@googlegroups.com
          name: etcd Community
        maturity: alpha
        provider:
          name: CNCF
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
        version: 0.9.4-clusterwide
      entries:
      - name: etcdoperator.v0.9.4-clusterwide
        version: 0.9.4-clusterwide
      - name: etcdoperator.v0.9.2-clusterwide
        version: 0.9.2-clusterwide
      - name: etcdoperator.v0.9.0
        version: 0.9.0
      name: clusterwide-alpha
    - currentCSV: etcdoperator.v0.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "size": 3,
                  "version": "3.2.13"
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdRestore",
                "metadata": {
                  "name": "example-etcd-cluster-restore"
                },
                "spec": {
                  "etcdCluster": {
                    "name": "example-etcd-cluster"
                  },
                  "backupStorageType": "S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdBackup",
                "metadata": {
                  "name": "example-etcd-cluster-backup"
                },
                "spec": {
                  "etcdEndpoints": ["<etcd-cluster-endpoints>"],
                  "storageType":"S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
          createdAt: "2019-02-28 01:03:00"
          description: Create and maintain highly-available etcd clusters on Kubernetes
          repository: https://github.com/coreos/etcd-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to backup an etcd cluster.
            displayName: etcd Backup
            kind: EtcdBackup
            name: etcdbackups.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to restore an etcd cluster from a backup.
            displayName: etcd Restore
            kind: EtcdRestore
            name: etcdrestores.etcd.database.coreos.com
            version: v1beta2
        description: |
          The etcd Operater creates and maintains highly-available etcd clusters on Kubernetes, allowing engineers to easily deploy and manage etcd clusters for their applications.

          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader.


          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` via port forwarding:

              $ kubectl --namespace default port-forward service/example-client 2379:2379
              $ etcdctl --endpoints http://127.0.0.1:2379 get /

          Or directly to the API using the automatically generated Kubernetes Service:

              $ etcdctl --endpoints http://example-client.default.svc:2379 get /

          Be sure to secure your etcd cluster (see Common Configurations) before exposing it outside of the namespace or cluster.


          ### Supported Features

          * **High availability** - Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.

          * **Automated updates** - Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.

          * **Backups included** - Create etcd backups and restore them through the etcd Operator.

          ### Common Configurations

          * **Configure TLS** - Specify [static TLS certs](https://github.com/coreos/etcd-operator/blob/master/doc/user/cluster_tls.md) as Kubernetes secrets.

          * **Set Node Selector and Affinity** - [Spread your etcd Pods](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-node-selector-and-anti-affinity-across-nodes) across Nodes and availability zones.

          * **Set Resource Limits** - [Set the Kubernetes limit and request](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-resource-requirement) values for your etcd Pods.

          * **Customize Storage** - [Set a custom StorageClass](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#custom-persistentvolumeclaim-definition) that you would like to use.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: etcd-dev@googlegroups.com
          name: etcd Community
        maturity: alpha
        provider:
          name: CNCF
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
        version: 0.9.4
      entries:
      - name: etcdoperator.v0.9.4
        version: 0.9.4
      - name: etcdoperator.v0.9.2
        version: 0.9.2
      - name: etcdoperator.v0.9.0
        version: 0.9.0
      name: singlenamespace-alpha
    defaultChannel: singlenamespace-alpha
    packageName: etcd
    provider:
      name: CNCF
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataStax
      provider-url: ""
    name: cass-operator-community
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cass-operator.v1.23.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cassandra.datastax.com/v1beta1",
                "kind": "CassandraDatacenter",
                "metadata": {
                  "name": "dc1"
                },
                "spec": {
                  "clusterName": "development",
                  "config": {
                    "cassandra-yaml": {
                      "authenticator": "PasswordAuthenticator",
                      "authorizer": "CassandraAuthorizer",
                      "num_tokens": 16,
                      "role_manager": "CassandraRoleManager"
                    },
                    "jvm-server-options": {
                      "initial_heap_size": "1G",
                      "max_heap_size": "1G"
                    }
                  },
                  "managementApiAuth": {
                    "insecure": {}
                  },
                  "podTemplateSpec": {
                    "spec": {
                      "containers": [
                        {
                          "name": "cassandra"
                        }
                      ],
                      "securityContext": {}
                    }
                  },
                  "racks": [
                    {
                      "name": "rack1"
                    },
                    {
                      "name": "rack2"
                    },
                    {
                      "name": "rack3"
                    }
                  ],
                  "resources": {
                    "requests": {
                      "cpu": "1000m",
                      "memory": "2Gi"
                    }
                  },
                  "serverType": "cassandra",
                  "serverVersion": "4.1.2",
                  "size": 3,
                  "storageConfig": {
                    "cassandraDataVolumeClaimSpec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "server-storage"
                    }
                  }
                }
              },
              {
                "apiVersion": "control.k8ssandra.io/v1alpha1",
                "kind": "CassandraTask",
                "metadata": {
                  "name": "example-cleanup"
                },
                "spec": {
                  "datacenter": {
                    "name": "dc2",
                    "namespace": "cass-operator"
                  },
                  "jobs": [
                    {
                      "command": "cleanup",
                      "name": "cleanup-run"
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database,Big Data,OpenShift Optional
          certified: "true"
          containerImage: cr.k8ssandra.io/k8ssandra/cass-operator:v1.23.2
          createdAt: "2025-03-20"
          description: |
            Simple provisioning, turn-key operations, and automated remediation of Apache Cassandra clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/k8ssandra/cass-operator
          support: DataStax
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CassandraDatacenter is the Schema for the cassandradatacenters
              API
            displayName: Cassandra Datacenter
            kind: CassandraDatacenter
            name: cassandradatacenters.cassandra.datastax.com
            version: v1beta1
          - description: CassandraTask is the Schema for the cassandrajobs API
            displayName: Cassandra Task
            kind: CassandraTask
            name: cassandratasks.control.k8ssandra.io
            version: v1alpha1
        description: "## Apache Cassandra\n\nThe Apache Cassandra database is the
          right choice when you need scalability and\nhigh availability without compromising
          performance. Linear scalability and\nproven fault-tolerance on commodity
          hardware or cloud infrastructure make it the\nperfect platform for mission-critical
          data. Cassandra's support for replicating\nacross multiple datacenters is
          best-in-class, providing lower latency for your\nusers and the peace of
          mind of knowing that you can survive regional outages.\n\n## DataStax Enterprise\n\nThe
          most advanced distribution of Apache Cassandra™ on the market, with the\nenterprise
          functionality needed for serious production systems and backed up and\nsupported
          by the best distributed-experts in the world. It's one platform for\nall
          types of applications anywhere, any cloud, any model: key-value, graph,\ntabular,
          JSON.\n\nDataStax Enterprise is a fully integrated and optimized database,
          with graph,\nanalytics, and search included, all with a unified security
          model. Simply put,\nit's the only database capable of meeting today's demanding
          requirements\n\n## Operator Details\n\n`cass-operator` is designed as a
          modular operator for Apache Cassandra and\nderived  distributions. Apache
          Cassandra is a distributed database consisting of\nmultiple nodes working
          in concert to store data and process queries along a\nnumber of fault domains.
          `cass-operator` has the deployment of a Cassandra\ncluster around the logical
          domain of a datacenter with the `CassandraDatacenter`\ncustom resource.\n
          \   \nUpon submission of one of these resources it handles provisioning
          the underlying\nstateful sets (analogous to C\\* logical racks), services,
          and configuration.\nAdditionally through monitoring pod state via Kubernetes
          callbacks it handles day to day\noperations such as restarting failed processes,
          scaling clusters up, and deploying\nconfiguration changes in a rolling,
          non-disruptive, fashion.\n    \nThis operator is designed to be `Namespace`
          scoped. A single Kubernetes cluster may\nbe running multiple instances of
          this operator, in separate namespaces, to support\na number of C\\* clusters
          and environments. Configuration is simple with the usage of\nYAML based
          overrides in the Custom Resource paired with an `init` container.\n    \nIn
          C\\* clusters ordering and timing of certain operations are important to
          keep the system\nevenly distributed. `cass-operator` takes advantage of
          a sidecar process within the\nmain container to handle the orchestration
          of starting our main C* process."
        displayName: DataStax Kubernetes Operator for Apache Cassandra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - k8ssandra
        - cass-operator
        - DataStax Enterprise
        - Apache Cassandra
        - DSE
        - C*
        - Cassandra
        links:
        - name: DataStax
          url: https://datastax.com
        - name: Operator GitHub Repo
          url: https://github.com/k8ssandra/cass-operator
        - name: Config Builder GitHub Repo
          url: https://github.com/datastax/cass-config-builder
        - name: Documentation [1]
          url: https://github.com/k8ssandra/cass-operator/tree/master/docs/user
        - name: Documentation [2]
          url: https://docs.datastax.com/en/cass-operator/doc/cass-operator/cassOperatorTOC.html
        maintainers:
        - email: cass-operator@datastax.com
          name: DataStax Operator Team
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: DataStax
        relatedImages:
        - cr.k8ssandra.io/k8ssandra/cass-operator:v1.23.2
        version: 1.23.2
      entries:
      - name: cass-operator.v1.23.2
        version: 1.23.2
      - name: cass-operator.v1.23.1
        version: 1.23.1
      - name: cass-operator.v1.23.0
        version: 1.23.0
      - name: cass-operator.v1.21.1
        version: 1.21.1
      - name: cass-operator.v1.19.1
        version: 1.19.1
      - name: cass-operator.v1.17.1
        version: 1.17.1
      - name: cass-operator.v1.16.0
        version: 1.16.0
      - name: cass-operator.v1.14.0
        version: 1.14.0
      - name: cass-operator.v1.13.0
        version: 1.13.0
      - name: cass-operator.v1.10.4
        version: 1.10.4
      - name: cass-operator.v1.10.1
        version: 1.10.1
      name: stable
    defaultChannel: stable
    packageName: cass-operator-community
    provider:
      name: DataStax
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: openstack-ansibleee-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openstack-ansibleee-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ansibleee.openstack.org/v1beta1",
                "kind": "OpenStackAnsibleEE",
                "metadata": {
                  "name": "openstackansibleee-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-09-23T20:57:34Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.3
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenStackAnsibleEE is the Schema for the openstackansibleees
              API
            displayName: OpenStack Ansible EE
            kind: OpenStackAnsibleEE
            name: openstackansibleees.ansibleee.openstack.org
            version: v1beta1
        description: The OpenStack Ansible Execution Environment Operator
        displayName: OpenStackAnsibleEE
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - execution
        links:
        - name: OpenStack Ansible EE Operator
          url: https://github.com/openstack-k8s-operators/openstack-ansibleee-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:24b468b09b43dd6cf77f5b08a68f2c6359d115722d28982fd8cdb1923e8122de
        - registry.redhat.io/rhoso-operators/openstack-ansibleee-rhel9-operator@sha256:bf096518547e92fb6fe0ecd526813bdc4a15f928a6d89eacadb71ffaad91db69
        - registry.redhat.io/rhoso-operators/ee-openstack-ansible-ee-rhel9@sha256:640efe595fbdb993ecc1c55c14457edc2b0e3e1c6dde8ad9ea1acf4b4503339f
        version: 1.0.3
      entries:
      - name: openstack-ansibleee-operator.v1.0.3
        version: 1.0.3
      - name: openstack-ansibleee-operator.v1.0.2
        version: 1.0.2
      - name: openstack-ansibleee-operator.v1.0.1
        version: 1.0.1
      - name: openstack-ansibleee-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: openstack-ansibleee-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: ""
    name: red-hat-hawtio-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: red-hat-hawtio-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              },
              {
                "apiVersion": "hawt.io/v2",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24,
                    "internalSSL": true
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
          createdAt: "2025-08-07T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v2
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
        version: 1.2.1
      entries:
      - name: red-hat-hawtio-operator.v1.2.1
        version: 1.2.1
      - name: red-hat-hawtio-operator.v1.2.0-0.1754874242.p
        version: 1.2.0+0.1754874242.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754338551.p
        version: 1.2.0+0.1754338551.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754317392.p
        version: 1.2.0+0.1754317392.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1753863723.p
        version: 1.2.0+0.1753863723.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752679472.p
        version: 1.2.0+0.1752679472.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752452299.p
        version: 1.2.0+0.1752452299.p
      - name: red-hat-hawtio-operator.v1.2.0
        version: 1.2.0
      - name: red-hat-hawtio-operator.v1.1.1
        version: 1.1.1
      - name: red-hat-hawtio-operator.v1.1.0
        version: 1.1.0
      - name: red-hat-hawtio-operator.v1.1.0-0.1742200932.p
        version: 1.1.0+0.1742200932.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1740358952.p
        version: 1.1.0+0.1740358952.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739994212.p
        version: 1.1.0+0.1739994212.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739992853.p
        version: 1.1.0+0.1739992853.p
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: latest
    - currentCSV: red-hat-hawtio-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              },
              {
                "apiVersion": "hawt.io/v2",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24,
                    "internalSSL": true
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
          createdAt: "2025-08-07T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v2
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
        version: 1.2.1
      entries:
      - name: red-hat-hawtio-operator.v1.2.1
        version: 1.2.1
      - name: red-hat-hawtio-operator.v1.2.0-0.1754874242.p
        version: 1.2.0+0.1754874242.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754338551.p
        version: 1.2.0+0.1754338551.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754317392.p
        version: 1.2.0+0.1754317392.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1753863723.p
        version: 1.2.0+0.1753863723.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752679472.p
        version: 1.2.0+0.1752679472.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752452299.p
        version: 1.2.0+0.1752452299.p
      - name: red-hat-hawtio-operator.v1.2.0
        version: 1.2.0
      - name: red-hat-hawtio-operator.v1.1.1
        version: 1.1.1
      - name: red-hat-hawtio-operator.v1.1.0
        version: 1.1.0
      - name: red-hat-hawtio-operator.v1.1.0-0.1742200932.p
        version: 1.1.0+0.1742200932.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1740358952.p
        version: 1.1.0+0.1740358952.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739994212.p
        version: 1.1.0+0.1739994212.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739992853.p
        version: 1.1.0+0.1739992853.p
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: stable-v1
    - currentCSV: red-hat-hawtio-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel8-operator@sha256:f04d81d24c757db7b678464382ad8c5182b1580e714123a9f4f5fc9fbd9b8c57
          createdAt: "2024-01-30T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-rhel8-operator@sha256:f04d81d24c757db7b678464382ad8c5182b1580e714123a9f4f5fc9fbd9b8c57
        version: 1.0.3
      entries:
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1
    packageName: red-hat-hawtio-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MinIO
      provider-url: https://min.io
    name: minio-key-manager-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-key-manager-operator.v2025.8.19175300
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aistor.min.io/v1alpha1",
                "kind": "KeyManager",
                "metadata": {
                  "name": "my-kms",
                  "namespace": "ns-1"
                },
                "spec": {
                  "apiKeySecret": {
                    "name": "my-api-key"
                  },
                  "configuration": {
                    "name": "my-kms-server-config"
                  },
                  "hsmSecret": {
                    "name": "my-kms-hsm"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "registry-creds"
                    }
                  ],
                  "replicas": 2,
                  "volumeClaimTemplate": {
                    "metadata": {
                      "name": "key-manager-volume"
                    },
                    "spec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "100Mi"
                        }
                      },
                      "storageClassName": "standard"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security, Storage
          createdAt: "2025-08-27T14:50:59Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KeyManager
            name: keymanagers.aistor.min.io
            version: v1alpha1
        description: |-
          Highly available, powerful and operationally simple, MinIO's Enterprise Key Management Server is optimized for large storage infrastructures where billions of cryptographic keys are required.

          MinIO's AIStor KMS establishes its foundational trust using the concept of an hardware security module (but given KES is software, this is only a concept). That module assumes a pivotal role in sealing and unsealing the KMS root encryption key. The module responsibility extends to safeguarding the integrity of KMS by allowing the unsealing of its encrypted on-disk state and facilitating communication among nodes within a KMS cluster.

          It solves the challenges associated with billions of cryptographic keys and hundreds of thousands of cryptographic operations per node per second - which are commonplace in larger deployments.

          <b>High Availability and Fault Tolerance</b>
          In the dynamic landscape of large-scale systems, network or node outages are inevitable. Taking down a cluster for maintenance is rarely feasible. MinIO's AIStor KMS ensures uninterrupted availability, even when faced with such disruptions, mitigating cascading effects that can take down the entire storage infrastructure. Specifically, you could lose all but one node of a cluster and still handle any encryption, decryption or data key generation requests.

          <b>Scalability</b>
          While the amount of data usually only increases, the load on a large-scale storage system may vary significantly from time to time. MinIO's AIStor KMS supports dynamic cluster resizing and nodes can be added or removed at any point without incurring any downtime.

          <b>Multi-Tenancy</b>
          Large-scale storage infrastructures are often used by many applications and teams across the entire organization. Isolating teams and groups into their own namespaces is a core requirement. MinIO's AIStor KMS supports namespacing in the form of enclaves. Each tenant can be assigned its own enclave which is completely independent and isolated from all other enclaves on the KMS cluster.

          <b>Predictable Behavior</b>
          MinIO's AIStor KMS is designed to be easily managed, providing operators with the ability to comprehend its state intuitively. Due to its simple design, MinIO's AIStor KMS is significantly easier to operate than similar solutions that rely on more complex consensus algorithms like Raft, or Paxos.

          <b>Consistent and Performant</b>
          The responsiveness of the KMS for GET/PUT operations directly influences the overall efficiency and speed of the storage system. MinIO's AIStor KMS nodes don’t have to coordinate when handling such requests from the storage system. Therefore, the performance of a MinIO's AIStor KMS cluster increases linearly with the number of nodes. Further, MinIO's AIStor KMS supports request pipelining to handle hundreds of thousands of cryptographic operations per node and second.

          <b>Simplicity</b>
          Operating a KMS cluster does not require expertise in cryptography or distributed systems. Everything can be done from the AIStor Console.

          <a href="https://resources.min.io/request-a-demo-marketplaces/">Request a demo today</a>
        displayName: Minio AIStor Key management System (KMS)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Storage
        - MinIO
        links:
        - name: Minio AIStor
          url: https://min.io
        - name: Support
          url: https://subnet.min.io
        - name: Github
          url: https://github.com/minio/aistor
        maintainers:
        - email: dev@min.io
          name: MinIO
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: MinIO
          url: https://min.io
        relatedImages:
        - quay.io/minio/aistor/operator@sha256:f80139efa062fdd83f78318b20f371a827b68e854b03dd9706b9eba4c493aeb2
        version: 2025.8.19175300
      entries:
      - name: minio-key-manager-operator.v2025.8.19175300
        version: 2025.8.19175300
      - name: minio-key-manager-operator.v2025.8.1184927
        version: 2025.8.1184927
      - name: minio-key-manager-operator.v2025.7.1011319
        version: 2025.7.1011319
      - name: minio-key-manager-operator.v2025.6.3065135
        version: 2025.6.3065135
      - name: minio-key-manager-operator.v2025.5.12190907
        version: 2025.5.12190907
      name: stable
    defaultChannel: stable
    packageName: minio-key-manager-operator
    provider:
      name: MinIO
      url: https://min.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator-certified.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a",
                  "imagePullPolicy": "Always",
                  "initImage": "registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:3c03fb18a5d73876576b80aca2b61e7c84ca052839ccb467d14e15f6d1f9d834
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a57926b53eef862b70a53818a7ccb0e9756f878f30d4128b608090b109bdc585
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator-certified.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator-certified.v1.20.0
        version: 1.20.0
      - name: percona-server-mongodb-operator-certified.v1.19.1
        version: 1.19.1
      - name: percona-server-mongodb-operator-certified.v1.19.0
        version: 1.19.0
      - name: percona-server-mongodb-operator-certified.v1.18.0
        version: 1.18.0
      - name: percona-server-mongodb-operator-certified.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator-certified
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Cloudera
      provider-url: https://cloudera.com
    name: cloudera-streams-messaging-kubernetes-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster"
                },
                "spec": {
                  "kafka": {
                    "version": "3.9.0.1.4",
                    "replicas": 3,
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2,
                      "inter.broker.protocol.version": "3.9"
                    },
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "zookeeper": {
                    "replicas": 3,
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "3.9.0.1.4",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
          createdAt: "2025-07-14 10:09:41"
          description: Run Apache Kafka with Cloudera Streams Messaging - Kubernetes
            Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: A valid license and current subscription
            for Cloudera Streaming - Kubernetes Operator is required. Contact Cloudera
            sales to obtain a license. https://www.cloudera.com/contact-sales.html
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |-
          Cloudera Streams Messaging - Kubernetes Operator brings enterprise-grade Kafka deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of Kafka for enterprise deployments. To learn more about Cloudera Streaming, [read this deep dive blog post](https://blog.cloudera.com/accelerating-deployments-of-streaming-pipelines-announcing-data-in-motion-on-kubernetes/).

          Cloudera Streams Messaging - Kubernetes Operator consists of and ships multiple components including Apache Kafka, Apache ZooKeeper, Cruise Control, Strimzi, and others.

          ## Installation prerequisites

          Installing Cloudera Streams Messaging - Kubernetes Operator requires a valid license as well as Cloudera credentials (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly. The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) where installation artifacts are pulled from.

          ### Obtaining a license

          Contact [Cloudera Sales](https://www.cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets

          Before installation, create  two `Secrets`. One containing your Cloudera credentials, and one containing your license. Cloudera credentials are provided to you together with your license.

          Create a `Secret` containing your Cloudera credentials.
          ```
          oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***] --docker-server=container.repository.cloudera.com
          ```
          Create a `Secret` containing your license.
          ```
          oc create secret generic csm-op-license --from-file=license.txt=[***PATH TO LICENSE FILE***]
          ```

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces where you plan to deploy Kafka or Kafka Connect clusters. Cloudera recommends that you create the `Secret` in all required namespaces before installation if you know what namespaces you will be using to deploy Kafka or Kafka Connect.

          ## Learn more
          To learn more, visit the [Cloudera Streams Messaging - Kubernetes Operator documentation library](https://docs.cloudera.com/csm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Streams Messaging - Kubernetes Operator, visit the following resources:

          * [Strimzi](https://strimzi.io/)
          * [Apache Kafka](https://kafka.apache.org/)
          * [Apache ZooKeeper](https://zookeeper.apache.org/)
          * [Cruise Control](https://github.com/linkedin/cruise-control)
        displayName: Cloudera Streams Messaging - Kubernetes Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Documentation
          url: https://docs.cloudera.com/csm-operator
        - name: Cloudera Community Forum
          url: https://community.cloudera.com/
        - name: Cloudera Stream Processing
          url: https://www.cloudera.com/products/stream-processing.html
        - name: Contact Cloudera
          url: https://www.cloudera.com/contact-sales.html
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
        - container.repository.cloudera.com/cloudera/kafka@sha256:aa64e161b9672153a6f6559f9ae9ef88d0e40d6040a80e21bd3dc8ee39df6523
        - container.repository.cloudera.com/cloudera/kafka@sha256:b675e7426a558c4681e53d8e03f278264d7ad555f7c02422017683a29708ac2b
        - container.repository.cloudera.com/cloudera/maven-builder@sha256:9f18203b175193e5079fdc0c2b71c14887641d997f92b4a1c918546f777166df
        version: 1.4.0-b199
      entries:
      - name: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
        version: 1.4.0-b199
      name: 1.2.x
    - currentCSV: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster"
                },
                "spec": {
                  "kafka": {
                    "version": "3.9.0.1.4",
                    "replicas": 3,
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2,
                      "inter.broker.protocol.version": "3.9"
                    },
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "zookeeper": {
                    "replicas": 3,
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "3.9.0.1.4",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
          createdAt: "2025-07-14 10:09:41"
          description: Run Apache Kafka with Cloudera Streams Messaging - Kubernetes
            Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: A valid license and current subscription
            for Cloudera Streaming - Kubernetes Operator is required. Contact Cloudera
            sales to obtain a license. https://www.cloudera.com/contact-sales.html
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |-
          Cloudera Streams Messaging - Kubernetes Operator brings enterprise-grade Kafka deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of Kafka for enterprise deployments. To learn more about Cloudera Streaming, [read this deep dive blog post](https://blog.cloudera.com/accelerating-deployments-of-streaming-pipelines-announcing-data-in-motion-on-kubernetes/).

          Cloudera Streams Messaging - Kubernetes Operator consists of and ships multiple components including Apache Kafka, Apache ZooKeeper, Cruise Control, Strimzi, and others.

          ## Installation prerequisites

          Installing Cloudera Streams Messaging - Kubernetes Operator requires a valid license as well as Cloudera credentials (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly. The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) where installation artifacts are pulled from.

          ### Obtaining a license

          Contact [Cloudera Sales](https://www.cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets

          Before installation, create  two `Secrets`. One containing your Cloudera credentials, and one containing your license. Cloudera credentials are provided to you together with your license.

          Create a `Secret` containing your Cloudera credentials.
          ```
          oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***] --docker-server=container.repository.cloudera.com
          ```
          Create a `Secret` containing your license.
          ```
          oc create secret generic csm-op-license --from-file=license.txt=[***PATH TO LICENSE FILE***]
          ```

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces where you plan to deploy Kafka or Kafka Connect clusters. Cloudera recommends that you create the `Secret` in all required namespaces before installation if you know what namespaces you will be using to deploy Kafka or Kafka Connect.

          ## Learn more
          To learn more, visit the [Cloudera Streams Messaging - Kubernetes Operator documentation library](https://docs.cloudera.com/csm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Streams Messaging - Kubernetes Operator, visit the following resources:

          * [Strimzi](https://strimzi.io/)
          * [Apache Kafka](https://kafka.apache.org/)
          * [Apache ZooKeeper](https://zookeeper.apache.org/)
          * [Cruise Control](https://github.com/linkedin/cruise-control)
        displayName: Cloudera Streams Messaging - Kubernetes Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Documentation
          url: https://docs.cloudera.com/csm-operator
        - name: Cloudera Community Forum
          url: https://community.cloudera.com/
        - name: Cloudera Stream Processing
          url: https://www.cloudera.com/products/stream-processing.html
        - name: Contact Cloudera
          url: https://www.cloudera.com/contact-sales.html
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
        - container.repository.cloudera.com/cloudera/kafka@sha256:aa64e161b9672153a6f6559f9ae9ef88d0e40d6040a80e21bd3dc8ee39df6523
        - container.repository.cloudera.com/cloudera/kafka@sha256:b675e7426a558c4681e53d8e03f278264d7ad555f7c02422017683a29708ac2b
        - container.repository.cloudera.com/cloudera/maven-builder@sha256:9f18203b175193e5079fdc0c2b71c14887641d997f92b4a1c918546f777166df
        version: 1.4.0-b199
      entries:
      - name: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
        version: 1.4.0-b199
      name: stable
    defaultChannel: stable
    packageName: cloudera-streams-messaging-kubernetes-operator
    provider:
      name: Cloudera
      url: https://cloudera.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: vertical-pod-autoscaler
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: verticalpodautoscaler.v4.16.0-202509240123
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "autoscaling.openshift.io/v1",
                "kind": "VerticalPodAutoscalerController",
                "metadata": {
                  "name": "default",
                  "namespace": "openshift-vertical-pod-autoscaler"
                },
                "spec": {
                  "podMinCPUMillicores": 25,
                  "podMinMemoryMb": 250,
                  "recommendationOnly": false,
                  "safetyMarginFraction": 0.15,
                  "minReplicas": 2
                }
              },
              {
                 "apiVersion": "autoscaling.k8s.io/v1",
                 "kind": "VerticalPodAutoscaler",
                 "metadata": {
                    "name": "myapp-vpa"
                 },
                 "spec": {
                    "targetRef": {
                       "apiVersion": "apps/v1",
                       "kind": "Deployment",
                       "name": "myapp-deployment"
                    },
                    "resourcePolicy": {
                       "containerPolicies": [
                          {
                             "containerName": "*",
                             "minAllowed": {
                                "cpu": "100m",
                                "memory": "50Mi"
                             },
                             "maxAllowed": {
                                "cpu": 1,
                                "memory": "500Mi"
                             },
                             "controlledResources": [
                                "cpu",
                                "memory"
                             ]
                          }
                       ]
                    }
                 }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          certifiedLevel: Primed
          containerImage: registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9-operator@sha256:edf76334695c9a4aee4a9c0297af40475a16dec7e35ec7113a35b8507724536e
          createdAt: 2019-10-14 12:17:13+00:00
          description: An operator to run the OpenShift Vertical Pod Autoscaler. Vertical
            Pod Autoscaler (VPA) can be configured to monitor a workload's resource
            utilization, and then adjust its CPU and memory limits by updating the
            pod (future) or restarting the pod with updated limits.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          healthIndex: B
          olm.skipRange: '>=4.5.0 <4.16.0-202509240123'
          operatorframework.io/suggested-namespace: openshift-vertical-pod-autoscaler
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/vertical-pod-autoscaler-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of the set of VPA controllers
            displayName: VPA Controller
            kind: VerticalPodAutoscalerController
            name: verticalpodautoscalercontrollers.autoscaling.openshift.io
            version: v1
          - description: An application whose resources are managed by the VPA
            displayName: VerticalPodAutoscaler
            kind: VerticalPodAutoscaler
            name: verticalpodautoscalers.autoscaling.k8s.io
            version: v1
          - description: Represents VPA checkpoints used by the VPA controllers to
              manage workloads
            displayName: VPA Checkpoint
            kind: VerticalPodAutoscalerCheckpoint
            name: verticalpodautoscalercheckpoints.autoscaling.k8s.io
            version: v1
        description: An operator to run the OpenShift Vertical Pod Autoscaler. Vertical
          Pod Autoscaler (VPA) can be configured to monitor a workload's resource
          utilization, and then adjust its CPU and memory limits by updating the pod
          (future) or restarting the pod with updated limits.
        displayName: VerticalPodAutoscaler
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Vertical Pod Autoscaler Documentation
          url: https://docs.openshift.com/container-platform/latest/nodes/pods/nodes-pods-vertical-autoscaler.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9-operator@sha256:edf76334695c9a4aee4a9c0297af40475a16dec7e35ec7113a35b8507724536e
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9@sha256:e6c8f34895184179bb87f1659c4ff30b96b131787fa7a64e7bc6b6adbb336c16
        version: 4.16.0-202509240123
      entries:
      - name: verticalpodautoscaler.v4.16.0-202509240123
        version: 4.16.0-202509240123
      - name: verticalpodautoscaler.v4.16.0-202508201333
        version: 4.16.0-202508201333
      - name: verticalpodautoscaler.v4.16.0-202507211806
        version: 4.16.0-202507211806
      - name: verticalpodautoscaler.v4.16.0-202507081835
        version: 4.16.0-202507081835
      - name: verticalpodautoscaler.v4.16.0-202506062300
        version: 4.16.0-202506062300
      - name: verticalpodautoscaler.v4.16.0-202505200039
        version: 4.16.0-202505200039
      - name: verticalpodautoscaler.v4.16.0-202504120704
        version: 4.16.0-202504120704
      - name: verticalpodautoscaler.v4.16.0-202503121138
        version: 4.16.0-202503121138
      - name: verticalpodautoscaler.v4.16.0-202502250034
        version: 4.16.0-202502250034
      - name: verticalpodautoscaler.v4.16.0-202502111405
        version: 4.16.0-202502111405
      - name: verticalpodautoscaler.v4.16.0-202501260334
        version: 4.16.0-202501260334
      - name: verticalpodautoscaler.v4.16.0-202412170135
        version: 4.16.0-202412170135
      - name: verticalpodautoscaler.v4.16.0-202412042304
        version: 4.16.0-202412042304
      - name: verticalpodautoscaler.v4.16.0-202412021735
        version: 4.16.0-202412021735
      - name: verticalpodautoscaler.v4.16.0-202411251535
        version: 4.16.0-202411251535
      - name: verticalpodautoscaler.v4.16.0-202411190033
        version: 4.16.0-202411190033
      - name: verticalpodautoscaler.v4.16.0-202410242205
        version: 4.16.0-202410242205
      - name: verticalpodautoscaler.v4.16.0-202410172201
        version: 4.16.0-202410172201
      - name: verticalpodautoscaler.v4.16.0-202410011135
        version: 4.16.0-202410011135
      - name: verticalpodautoscaler.v4.16.0-202409202304
        version: 4.16.0-202409202304
      - name: verticalpodautoscaler.v4.16.0-202409051837
        version: 4.16.0-202409051837
      - name: verticalpodautoscaler.v4.16.0-202409032335
        version: 4.16.0-202409032335
      - name: verticalpodautoscaler.v4.16.0-202408262007
        version: 4.16.0-202408262007
      - name: verticalpodautoscaler.v4.16.0-202408201708
        version: 4.16.0-202408201708
      - name: verticalpodautoscaler.v4.16.0-202408081442
        version: 4.16.0-202408081442
      - name: verticalpodautoscaler.v4.16.0-202407031636
        version: 4.16.0-202407031636
      - name: verticalpodautoscaler.v4.16.0-202407030803
        version: 4.16.0-202407030803
      - name: verticalpodautoscaler.v4.16.0-202406200537
        version: 4.16.0-202406200537
      - name: verticalpodautoscaler.v4.16.0-202406131906
        version: 4.16.0-202406131906
      - name: verticalpodautoscaler.v4.16.0-202406061206
        version: 4.16.0-202406061206
      name: stable
    defaultChannel: stable
    packageName: vertical-pod-autoscaler
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-businessautomation: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: businessautomation-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: businessautomation-operator.7.13.5-16
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.kiegroup.org/v2","kind":"KieApp","metadata":{"name":"rhpam-trial"},"spec":{"environment":"rhpam-trial"}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/rhpam-7/rhpam-rhel8-operator@sha256:3b4f420def9e5d2c768d5209264bd5f9af6d6a37b22ef9a988a167ff7d9959cf
          createdAt: "2025-07-30 19:54:27"
          description: Deploys and manages Red Hat Process Automation Manager and
            Red Hat Decision Manager environments.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Process Automation
            Manager", "Red Hat Decision Manager"]'
          repository: https://github.com/kiegroup/kie-cloud-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A project prescription running an RHPAM/RHDM environment.
            displayName: KieApp
            kind: KieApp
            name: kieapps.app.kiegroup.org
            version: v2
        description: |-
          Deploys and manages Red Hat Process Automation Manager and Red Hat Decision Manager environments.

          * **Red Hat Process Automation Manager** is a platform for developing containerized microservices and applications that automate business decisions and processes. It includes business process management (BPM), business rules management (BRM), and business resource optimization and complex event processing (CEP) technologies. It also includes a user experience platform to create engaging user interfaces for process and decision services with minimal coding.

          [See more](https://www.redhat.com/en/products/process-automation).
        displayName: Business Automation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kieapp
        - pam
        - decision
        - kie
        - cloud
        - bpm
        - process
        - automation
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-process-automation-manager
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_process_automation_manager/7.13/#category-deploying-red-hat-process-automation-manager-on-openshift
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-monitoring-rhel8@sha256:8d4481a11269565f1690ad007f152e24326bae4b2099cb26ddf7b21be67031e8
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:cb37360843e9b062c59755593d9d412e1bf2c1dc1d6902e8a65a2825812c0989
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhpam-7/rhpam-kieserver-rhel8@sha256:2526d12da63b67ae9ed9294c914e9406d833c090ffa777623303ea0be22755d8
        - registry.redhat.io/rhpam-7/rhpam-process-migration-rhel8@sha256:3d7bdbe38895e50549e97573bd40da47b11691932164db7dafff0155a051aab2
        - registry.redhat.io/rhpam-7/rhpam-controller-rhel8@sha256:3a901b54a9d35d38b5aca275df2160e3190a874e139fcb606b9969b38284b3dc
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        - registry.redhat.io/rhpam-7/rhpam-kieserver-rhel8@sha256:3fbef82d1c5da8da1606d1a2d5dcfd20000da574737eb32b772d4f0123b179cb
        - registry.redhat.io/rhpam-7/rhpam-smartrouter-rhel8@sha256:d1e5461db8983a13f98259152dc3cfa2aa677c29f799104190e5cefa83cf97b5
        - registry.redhat.io/rhpam-7/rhpam-dashbuilder-rhel8@sha256:46396dc8060f2f9cc6f87ad905ef55be5b8e5216e5162bdf54df98c9d70e29ef
        - registry.redhat.io/rhscl/mysql-80-rhel7@sha256:e4f535c78bc2fdb58914c52fb805185035ff31515452b4df714f9ab623ba383a
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-rhel8@sha256:228c6ab5b83e215a98f6eb76250feb2ae3096fd21398c44e5b2701257c9c3493
        - registry.redhat.io/rhpam-7/rhpam-smartrouter-rhel8@sha256:9d7efdd97f163770d8a84b1ce5236bd4f31508154dced1f11dc6efe75a7b5089
        - registry.redhat.io/rhscl/postgresql-13-rhel7@sha256:338dd09681dcf860e6a976a3bc48a1474c2376d698815e4eb611ae9c5aa68966
        - registry.redhat.io/rhpam-7/rhpam-rhel8-operator@sha256:3b4f420def9e5d2c768d5209264bd5f9af6d6a37b22ef9a988a167ff7d9959cf
        - registry.redhat.io/rhpam-7/rhpam-dashbuilder-rhel8@sha256:4edac282119898ab1395030eecaee584990064e88fc8e65fb999f19ea2ef3f8b
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-rhel8@sha256:1abbaba9d4b2ecaec38205bcccf5ef4c7b6388ed9725b4f4c6dc8dc664588eb1
        - registry.redhat.io/rhpam-7/rhpam-controller-rhel8@sha256:989dc6dc64a2d5b6723fb5c000b0411fbdc6f78c0056666d831ffeabeab4b612
        - registry.redhat.io/openshift4/ose-cli@sha256:c3eb8807caf73e30012ae53c6d4f5114471290a5596929c27acef5a699a5b813
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/rhpam-7/rhpam-process-migration-rhel8@sha256:bd5c6586b4859dec23ad9530ee98bab33497ab251d544f863fc0af01cbd45101
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:457fc73423fa94591bc7e35729a25c58d3465583afccf342fb338b6e11188575
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-monitoring-rhel8@sha256:32912270eba0f2464da88f315bde8bf1ef5bbded814db156811d7a415a458f86
        version: 7.13.5-16
      entries:
      - name: businessautomation-operator.7.13.5-16
        version: 7.13.5-16
      - name: businessautomation-operator.7.13.5-15
        version: 7.13.5-15
      - name: businessautomation-operator.7.13.5-14
        version: 7.13.5-14
      - name: businessautomation-operator.7.13.5-13
        version: 7.13.5-13
      - name: businessautomation-operator.7.13.5-12
        version: 7.13.5-12
      - name: businessautomation-operator.7.13.5-11
        version: 7.13.5-11
      - name: businessautomation-operator.7.13.5-10
        version: 7.13.5-10
      - name: businessautomation-operator.7.13.5-9
        version: 7.13.5-9
      - name: businessautomation-operator.7.13.5-8
        version: 7.13.5-8
      - name: businessautomation-operator.7.13.5-7
        version: 7.13.5-7
      - name: businessautomation-operator.7.13.5-6
        version: 7.13.5-6
      - name: businessautomation-operator.7.13.5-5
        version: 7.13.5-5
      - name: businessautomation-operator.7.13.5-4
        version: 7.13.5-4
      - name: businessautomation-operator.7.13.5-3
        version: 7.13.5-3
      - name: businessautomation-operator.7.13.5-2
        version: 7.13.5-2
      - name: businessautomation-operator.7.13.5-1
        version: 7.13.5-1
      - name: businessautomation-operator.7.13.4-4
        version: 7.13.4-4
      - name: businessautomation-operator.7.13.4-3
        version: 7.13.4-3
      - name: businessautomation-operator.7.13.4-2
        version: 7.13.4-2
      - name: businessautomation-operator.7.13.4-1
        version: 7.13.4-1
      - name: businessautomation-operator.7.13.3-5
        version: 7.13.3-5
      - name: businessautomation-operator.7.13.3-4
        version: 7.13.3-4
      - name: businessautomation-operator.7.13.3-3
        version: 7.13.3-3
      - name: businessautomation-operator.7.13.3-2
        version: 7.13.3-2
      - name: businessautomation-operator.7.13.3-1
        version: 7.13.3-1
      - name: businessautomation-operator.7.13.2-3
        version: 7.13.2-3
      - name: businessautomation-operator.7.13.2-2
        version: 7.13.2-2
      - name: businessautomation-operator.7.13.2-1
        version: 7.13.2-1
      - name: businessautomation-operator.7.13.1-2
        version: 7.13.1-2
      - name: businessautomation-operator.7.13.1-1
        version: 7.13.1-1
      - name: businessautomation-operator.7.13.0-1
        version: 7.13.0-1
      - name: businessautomation-operator.7.12.1-4
        version: 7.12.1-4
      - name: businessautomation-operator.7.12.1-3
        version: 7.12.1-3
      - name: businessautomation-operator.7.12.1-2
        version: 7.12.1-2
      - name: businessautomation-operator.7.12.1-1
        version: 7.12.1-1
      - name: businessautomation-operator.7.12.0-2
        version: 7.12.0-2
      - name: businessautomation-operator.7.12.0-1
        version: 7.12.0-1
      - name: businessautomation-operator.7.11.1-3
        version: 7.11.1-3
      - name: businessautomation-operator.7.11.1-2
        version: 7.11.1-2
      - name: businessautomation-operator.7.11.1-1
        version: 7.11.1-1
      - name: businessautomation-operator.7.11.0-2
        version: 7.11.0-2
      - name: businessautomation-operator.7.11.0-1
        version: 7.11.0-1
      - name: businessautomation-operator.7.10.1-2
        version: 7.10.1-2
      - name: businessautomation-operator.7.10.1-1
        version: 7.10.1-1
      - name: businessautomation-operator.7.10.0-2
        version: 7.10.0-2
      - name: businessautomation-operator.7.10.0-1
        version: 7.10.0-1
      - name: businessautomation-operator.7.9.1-1
        version: 7.9.1-1
      - name: businessautomation-operator.7.9.0
        version: 7.9.0
      - name: businessautomation-operator.7.9.0-2
        version: 7.9.0-2
      - name: businessautomation-operator.7.8.1
        version: 7.8.1
      - name: businessautomation-operator.7.8.0
        version: 7.8.0
      - name: businessautomation-operator.1.4.1
        version: 1.4.1
      - name: businessautomation-operator.1.4.0
        version: 1.4.0
      - name: businessautomation-operator.1.3.0
        version: 1.3.0
      - name: businessautomation-operator.1.2.1
        version: 1.2.1
      - name: businessautomation-operator.1.2.0
        version: 1.2.0
      name: stable
    defaultChannel: stable
    packageName: businessautomation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: quay-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:d73321dfa6dcced943405fae45c2ccf841663e0ab57e8be8cc1ddba251b7ce08
          createdAt: 2021-11-02 11:11 UTC
          description: Opinionated deployment of Red Hat Quay on Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:d73321dfa6dcced943405fae45c2ccf841663e0ab57e8be8cc1ddba251b7ce08
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:0e6bef51caf479074925115fa11621c41bcada375cc1925d0511a73084b93543
        - registry.redhat.io/rhel8/postgresql-10@sha256:b2695e9af379624621bef9ec785f0cb618bfdabae98c5f17eea82ff5f8d2d5cb
        - registry.redhat.io/rhel8/redis-5@sha256:8d7e66927ce4b3050ca15ab4df4dbaeb593cb855e16fb0a6f5dde950e9db7db9
        - registry.redhat.io/quay/quay-rhel8@sha256:ed8c4249d376b4226e35bf088ffe5bf688bf68b059718c5e218aa1184fe048d0
        - registry.redhat.io/quay/clair-rhel8@sha256:af5e2fd1543365b365b1a6cce00b86e44fe5e66cde6491e9b113324c21f629c8
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:2a471300f70d3995ab7e9fcc3df35c5513e73fcc1d140fb4d651a1c3eae949eb
        version: 3.4.7
      entries:
      - name: quay-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: quay-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:1f5e4cc9ec2d9b8b89d82f1a698acfa1ac2764ca4ebc5c85b3417da7730b5867
          createdAt: 2021-09-09 20:09 UTC
          description: Opinionated deployment of Red Hat Quay on Kubernetes.
          olm.skipRange: '>=3.4.x <3.5.7'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.5.7
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7fe408a1e428b2d0660cb578c1dcd4b961f838e37166fe7b21b2bed70d288594
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:5d72577ccfd63c6d9ad712b376b604c40d2d3ecb53e391e2ee0d18caa6e0225a
        - registry.redhat.io/rhel8/postgresql-10@sha256:a723bb3a7a5046527b9a5b628ddfb211c526e3814cc4242262c4f8a016f38edd
        - registry.redhat.io/rhel8/redis-5@sha256:7916760ff03af5ca230df5e96b2bf353fbf1829386de1c063c05f623837be7f7
        - registry.redhat.io/quay/quay-rhel8@sha256:c599892160ac20c744d833603d9375923af7aec1edfd86982b7513e02fb6d463
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:1f5e4cc9ec2d9b8b89d82f1a698acfa1ac2764ca4ebc5c85b3417da7730b5867
        - registry.redhat.io/quay/clair-rhel8@sha256:d242bd9d37f95b2a5a8267b3f1eba02c671d46121712baae43466c86d1da89f9
        version: 3.5.7
      entries:
      - name: quay-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: quay-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:3c0201e3cf88e711c921870324002a02cbd42c0d2eecafb4f242aa0601e24602
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.10.15
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:fc20b8bdac88c1bfce9aec993bdce6dd2cc132f8a9d4062d174b098819c74013
        - registry.redhat.io/quay/clair-rhel8@sha256:efa6e69500ea643042b544e6205f088214d79ab09a36864a612e73d07979d1c5
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:3c0201e3cf88e711c921870324002a02cbd42c0d2eecafb4f242aa0601e24602
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7e720c6a332ec4032e8576c31421af404f61d9baaf8e88f69f8333c52e5d1683
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:774cd1f181527787ba8631c7c14437eb2250c15ddb4a9c75fdfa35c5c1a19a2e
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        version: 3.10.15
      entries:
      - name: quay-operator.v3.10.15
        version: 3.10.15
      - name: quay-operator.v3.10.14
        version: 3.10.14
      - name: quay-operator.v3.10.13
        version: 3.10.13
      - name: quay-operator.v3.10.12
        version: 3.10.12
      - name: quay-operator.v3.10.11
        version: 3.10.11
      - name: quay-operator.v3.10.10
        version: 3.10.10
      - name: quay-operator.v3.10.9
        version: 3.10.9
      - name: quay-operator.v3.10.8
        version: 3.10.8
      - name: quay-operator.v3.10.7
        version: 3.10.7
      - name: quay-operator.v3.10.6
        version: 3.10.6
      - name: quay-operator.v3.10.5
        version: 3.10.5
      - name: quay-operator.v3.10.4
        version: 3.10.4
      - name: quay-operator.v3.10.3
        version: 3.10.3
      - name: quay-operator.v3.10.2
        version: 3.10.2
      - name: quay-operator.v3.10.1
        version: 3.10.1
      - name: quay-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:cb3a41caa3a572b7981d424be2a32b8e881127cc03b557ee0a36ab13f430b897
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.11.13
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:7230fa1d4970e877cb6de8716e99a9bd43302c0ff23b4501f2aa806a8d496d9b
        - registry.redhat.io/quay/clair-rhel8@sha256:3746677ef1cb0ddd4f237ab598b14cf0bc76dcce080ffde127b5bba87ffe6693
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:1a083cd72502609427428ed22fb60846613ddb41a4613496c5cddc797b382d3a
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:11b8c9d64c42f0f0c7b2df491bba859d0410666a8ef1d2fd41eb77e4ad9d2dc8
        - registry.redhat.io/rhel8/postgresql-13@sha256:35d86936c539007a78a89869a14b87b76788fe117df442a670f63936b978ed5b
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:cb3a41caa3a572b7981d424be2a32b8e881127cc03b557ee0a36ab13f430b897
        version: 3.11.13
      entries:
      - name: quay-operator.v3.11.13
        version: 3.11.13
      - name: quay-operator.v3.11.12
        version: 3.11.12
      - name: quay-operator.v3.11.11
        version: 3.11.11
      - name: quay-operator.v3.11.10
        version: 3.11.10
      - name: quay-operator.v3.11.9
        version: 3.11.9
      - name: quay-operator.v3.11.8
        version: 3.11.8
      - name: quay-operator.v3.11.7
        version: 3.11.7
      - name: quay-operator.v3.11.6
        version: 3.11.6
      - name: quay-operator.v3.11.5
        version: 3.11.5
      - name: quay-operator.v3.11.4
        version: 3.11.4
      - name: quay-operator.v3.11.3
        version: 3.11.3
      - name: quay-operator.v3.11.2
        version: 3.11.2
      - name: quay-operator.v3.11.1
        version: 3.11.1
      - name: quay-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:accb705b6327fb2dcc1ce657a99f61e343eca752c9873d63d89fea89c10c3187
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.12.12
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:accb705b6327fb2dcc1ce657a99f61e343eca752c9873d63d89fea89c10c3187
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:ccbd83e05aea0bb96bf099f63420175b36ed413bfc8a618f4c90b6a2abc53895
        - registry.redhat.io/quay/clair-rhel8@sha256:dc3d5fb35f2bd5d00480351dca705fe260a225d9a97958cf0276801af08d8307
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:8e5dd171e6e94de0137bb6a972533b1f93b56efc1731e19b80869d5e9b8a26a6
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:8cdd677449f3c723dad79bcd9c4191c8917d6eb0beeeeeda68fe19a0988bb2a8
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        version: 3.12.12
      entries:
      - name: quay-operator.v3.12.12
        version: 3.12.12
      - name: quay-operator.v3.12.11
        version: 3.12.11
      - name: quay-operator.v3.12.10
        version: 3.12.10
      - name: quay-operator.v3.12.9
        version: 3.12.9
      - name: quay-operator.v3.12.8
        version: 3.12.8
      - name: quay-operator.v3.12.7
        version: 3.12.7
      - name: quay-operator.v3.12.6
        version: 3.12.6
      - name: quay-operator.v3.12.5
        version: 3.12.5
      - name: quay-operator.v3.12.4
        version: 3.12.4
      - name: quay-operator.v3.12.3
        version: 3.12.3
      - name: quay-operator.v3.12.2
        version: 3.12.2
      - name: quay-operator.v3.12.1
        version: 3.12.1
      - name: quay-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:5481a4b914ce4aed6ba3ceee8432eddab6e264cad2caef528404795f904100eb
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.13.8
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/clair-rhel8@sha256:069ccf36346cb14f4e4ea54de1d59afb72548912b140cfef67a6bcb7534d25e9
        - registry.redhat.io/rhel8/postgresql-13@sha256:f17fd8dd4d56d8e916cb2ad3f403d05ad259f79fd6918d67a38a6e9e4a8896cf
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:966a3dafa2dd31ac1a39e826e08c1b988fe694189fe79449bbe14a48b2631213
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:5481a4b914ce4aed6ba3ceee8432eddab6e264cad2caef528404795f904100eb
        - registry.redhat.io/quay/quay-rhel8@sha256:8887c17547c7a4be61471523eade9820e5095cf966b1885a3758081da559871c
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:31a9095d8f291ce68e8e688877692e74a17143bdda8d998fe3a07bcee32575dd
        - registry.redhat.io/rhel8/postgresql-15@sha256:a5f36e7893a519c12752c6118233c92d4e83618670bc9ea7bd3a8a34a63bd50b
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:f62cbd6632454c227b7cbb01dadc8e501e8288b2d8e0abf8e7842b806fa3195c
        version: 3.13.8
      entries:
      - name: quay-operator.v3.13.8
        version: 3.13.8
      - name: quay-operator.v3.13.7
        version: 3.13.7
      - name: quay-operator.v3.13.6
        version: 3.13.6
      - name: quay-operator.v3.13.5
        version: 3.13.5
      - name: quay-operator.v3.13.4
        version: 3.13.4
      - name: quay-operator.v3.13.3
        version: 3.13.3
      - name: quay-operator.v3.13.2
        version: 3.13.2
      - name: quay-operator.v3.13.1
        version: 3.13.1
      - name: quay-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:8ebc1f0511f7d394ceaa39a4c993e955c560dc04588072cb3f8c8fab8c85e32b
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.14.5
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/clair-rhel8@sha256:e11e8a819d17bd14c67f093b5303b4958855e731f89ce67dcca70fcb7b77f45a
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:8ebc1f0511f7d394ceaa39a4c993e955c560dc04588072cb3f8c8fab8c85e32b
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-rhel8@sha256:33e984892650d315065313fcb5f8808e145aadec5a287219c7031d85b713235e
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:43ef8adddbf9b007b6e2396f53bd6a5475150d28c2f4254f031a0fe6c7de5826
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:a12a0918f6dea114b0aae231d7119a4acfd0cd297074d895f4ecf190d2aec43a
        version: 3.14.5
      entries:
      - name: quay-operator.v3.14.5
        version: 3.14.5
      - name: quay-operator.v3.14.4
        version: 3.14.4
      - name: quay-operator.v3.14.3
        version: 3.14.3
      - name: quay-operator.v3.14.2
        version: 3.14.2
      - name: quay-operator.v3.14.1
        version: 3.14.1
      - name: quay-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:05fb981d9848281b97b3683093f752b1c1dc5ced439a2b4ba18b26ff8e11231e
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.15.2
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:873a3daf1dadf2438eb96e151072326c31ba34ed48cac041614b604413d5f8e7
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:956f3110f944d880bc27bb6df9472d0804cd75f5dcf4ed11bcf721e9d1d4542b
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:aa8c9f7b6ce57441ca1fe9dbaaefaf2bb83fd44aa367647b663bb822bf890362
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:05fb981d9848281b97b3683093f752b1c1dc5ced439a2b4ba18b26ff8e11231e
        - registry.redhat.io/quay/clair-rhel8@sha256:861be31d30142bae125cd059bb716d79b28694e9c4336333c68adf4a3823b61d
        version: 3.15.2
      entries:
      - name: quay-operator.v3.15.2
        version: 3.15.2
      - name: quay-operator.v3.15.1
        version: 3.15.1
      - name: quay-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-operator.v3.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:7a6e592d7f5f760b4451f9c974dbb96f56b1d3de92f039adcbf284f78ba13a02
          createdAt: 2021-10-12 22:10 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.6.10
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:ec36241f0a9a2983b1894cd8501898e6b70379af51a81b1e09b0ba0e9b54b33f
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:75df6b8e6bef14844a08f7d994470fa6ec4ffa66d4b15faaa0a3ba373f612fbb
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:7a6e592d7f5f760b4451f9c974dbb96f56b1d3de92f039adcbf284f78ba13a02
        - registry.redhat.io/rhel8/postgresql-10@sha256:4b912c80085b88a03309aeb7907efcc29dd3342fa3952b6ea067afb1914bfe53
        - registry.redhat.io/rhel8/redis-6@sha256:d7c7852338717308cbb59e9303e1ea35cc8e5c01ceb2818569be20c15f3f943d
        - registry.redhat.io/quay/quay-rhel8@sha256:e75e4d084eba9bb65b9cc72e38294af314d49316b12a0beb75db808dc057830b
        - registry.redhat.io/quay/clair-rhel8@sha256:cbafddd1eb50c1c0f9159ffafd86f44a8c0b976194903c96c6f144bfe62e749d
        version: 3.6.10
      entries:
      - name: quay-operator.v3.6.10
        version: 3.6.10
      - name: quay-operator.v3.6.9
        version: 3.6.9
      - name: quay-operator.v3.6.8
        version: 3.6.8
      - name: quay-operator.v3.6.7
        version: 3.6.7
      - name: quay-operator.v3.6.6
        version: 3.6.6
      - name: quay-operator.v3.6.5
        version: 3.6.5
      - name: quay-operator.v3.6.4
        version: 3.6.4
      - name: quay-operator.v3.6.2
        version: 3.6.2
      - name: quay-operator.v3.6.1
        version: 3.6.1
      - name: quay-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: quay-operator.v3.7.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:a8283c2be661e07c5f2f2e5b683c520846bcb91d580d9a3a0aa27ce994ddfff6
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.7.14
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:b15335d88b240bd6bbedeb7fe26f0a1de2f6c330af9758b0fb2edab7ca69f6df
        - registry.redhat.io/quay/quay-rhel8@sha256:7bc9e115e9bace07505ffd225242cab83d5eb6d50358188d18fd9161bdf1a5a6
        - registry.redhat.io/quay/clair-rhel8@sha256:d676af83e96740220e6f18721c98817ba071fd023a62ebdc3815424e94838083
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:c54c77e09c95dc7d80d7b31e796857c8d06f2f733426629b53d3298206d23b20
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:04ac405c152bffb43b33fb2e7f45df8b3940eb21e055ebf6aa17bd81ebdeeab4
        - registry.redhat.io/rhel8/postgresql-10@sha256:a52ad402458ec8ef3f275972c6ebed05ad64398f884404b9bb8e3010c5c95291
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:a8283c2be661e07c5f2f2e5b683c520846bcb91d580d9a3a0aa27ce994ddfff6
        version: 3.7.14
      entries:
      - name: quay-operator.v3.7.14
        version: 3.7.14
      - name: quay-operator.v3.7.13
        version: 3.7.13
      - name: quay-operator.v3.7.12
        version: 3.7.12
      - name: quay-operator.v3.7.11
        version: 3.7.11
      - name: quay-operator.v3.7.10
        version: 3.7.10
      name: stable-3.7
    - currentCSV: quay-operator.v3.8.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:abec0fa7ee641c8e424e995b5fe07b98e522ad1eb92c3b90426f69941f33eb2e
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.8.15
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7e4e57071439d671971b0de46b597cf6cff8ef3879ab711125a6ae7ecb9c6d29
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:3c047c7b22fb42557bd3a9b9c9ac333f0df7d5e4a4d333a2b02a4313f57085be
        - registry.redhat.io/rhel8/postgresql-10@sha256:b7fa644277d5e406b642685ff2ace1b1de7aea2e2a1421b3cbcda7d60405f44d
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:abec0fa7ee641c8e424e995b5fe07b98e522ad1eb92c3b90426f69941f33eb2e
        - registry.redhat.io/rhel8/redis-6@sha256:444200241cb9186cd97d564d7acf077d83fc997a8ee5ee7fdae41b604895085c
        - registry.redhat.io/quay/quay-rhel8@sha256:de2b65506f99eebb5eaafca6bb42b4ebec08e9479dd0d50aa1286f197e42d6de
        - registry.redhat.io/quay/clair-rhel8@sha256:b327fcfd7472cdb9d83a564904976b26b285fa576dc0a76e5590990ec77fd871
        version: 3.8.15
      entries:
      - name: quay-operator.v3.8.15
        version: 3.8.15
      - name: quay-operator.v3.8.14
        version: 3.8.14
      - name: quay-operator.v3.8.13
        version: 3.8.13
      - name: quay-operator.v3.8.12
        version: 3.8.12
      - name: quay-operator.v3.8.11
        version: 3.8.11
      - name: quay-operator.v3.8.10
        version: 3.8.10
      - name: quay-operator.v3.8.9
        version: 3.8.9
      - name: quay-operator.v3.8.8
        version: 3.8.8
      - name: quay-operator.v3.8.7
        version: 3.8.7
      - name: quay-operator.v3.8.6
        version: 3.8.6
      - name: quay-operator.v3.8.5
        version: 3.8.5
      - name: quay-operator.v3.8.4
        version: 3.8.4
      - name: quay-operator.v3.8.3
        version: 3.8.3
      - name: quay-operator.v3.8.2
        version: 3.8.2
      - name: quay-operator.v3.8.1
        version: 3.8.1
      - name: quay-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:8a16803946272156b7aeb00eb2202ab6cb87d875fe4a7c327554f770f0f2ecb5
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.9.16
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/clair-rhel8@sha256:6c67163c783d7d9fa4d187963379885a8b76685347ecf1894c4f8ebd6656544b
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:af8ada0c88b6f19991e61ee86d04b49a148c19dacb57c282e4ab98630e7499b4
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:5a42689198701d845ff3f05502cbbebceb7e6ad7b12762ccc038c0131cf6479e
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-rhel8@sha256:f87b7e7ee6c31f6ae80b08aef0e10729b21fe1b8656fe18f6dee2310bae28ec1
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:8a16803946272156b7aeb00eb2202ab6cb87d875fe4a7c327554f770f0f2ecb5
        version: 3.9.16
      entries:
      - name: quay-operator.v3.9.16
        version: 3.9.16
      - name: quay-operator.v3.9.15
        version: 3.9.15
      - name: quay-operator.v3.9.14
        version: 3.9.14
      - name: quay-operator.v3.9.13
        version: 3.9.13
      - name: quay-operator.v3.9.12
        version: 3.9.12
      - name: quay-operator.v3.9.11
        version: 3.9.11
      - name: quay-operator.v3.9.10
        version: 3.9.10
      - name: quay-operator.v3.9.9
        version: 3.9.9
      - name: quay-operator.v3.9.8
        version: 3.9.8
      - name: quay-operator.v3.9.7
        version: 3.9.7
      - name: quay-operator.v3.9.6
        version: 3.9.6
      - name: quay-operator.v3.9.5
        version: 3.9.5
      - name: quay-operator.v3.9.4
        version: 3.9.4
      - name: quay-operator.v3.9.3
        version: 3.9.3
      - name: quay-operator.v3.9.2
        version: 3.9.2
      - name: quay-operator.v3.9.1
        version: 3.9.1
      - name: quay-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: quay-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: runtime-component-operator-certified
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runtime-component.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/runtime-component-operator@sha256:d37b3f62575c9ad1ffcb22c89e2c3886275c326bf27aa49fb1808acba475853d
          createdAt: "2022-03-25T14:00:00Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          olm.skipRange: <0.8.1
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          This advanced Operator is capable of deploying any runtime component image with consistent, production-grade QoS. It enables enterprise architects to govern the way their applications get deployed & managed in the cluster, while dramatically reducing the learning curve for developers to deploy into Kubernetes - allowing them to focus on writing the code!
          Here are some key features:

          #### Application Lifecyle
          You can deploy your runtime component container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the Operator will watch for any updates and will re-deploy the modified image.

          #### Custom RBAC
          This Operator is capable of using a custom ServiceAccount from the caller, allowing it to follow RBAC restrictions. By default it creates a ServiceAccount if one is not specified, which can also be bound with specific roles.

          #### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as: labels, annotations, and environment variables from a ConfigMap, a Secret or a value.

          #### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configuration, such as TLS settings, are also easily enabled.

          #### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto scaling to create (and delete) instances based on resource consumption.

          #### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you.
          This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart.
          We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.

          #### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating kubernetes Secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to its required services without any intervention or interruption.

          #### Exposing metrics to Prometheus
          The Runtime Component Operator exposes the runtime container's metrics via the Prometheus Operator.
          You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `spec.monitoring` field to control configurations such as the poll interval and security credentials.

          #### Easily mount logs and transaction directories
          If you need to mount the logs and transaction data from your runtime component to an external volume such as NFS (or any storage supported in your cluster), simply add the following (to specify the volume size and the location to persist) to your RuntimeComponent CR:
          ``` storage: size: 2Gi mountPath: "/logs" ```

          #### Integration with OpenShift Serverless
          Deploy your serverless runtime component using a single toggle.  The Operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing your pod to automatically scale to 0 when it is idle.

          #### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and how they are connected.

          See our [**documentation**](https://github.com/application-stacks/runtime-component-operator/tree/main/doc/) for more information.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://github.com/application-stacks/runtime-component-operator/tree/main/doc
        - name: Repository
          url: https://github.com/application-stacks/runtime-component-operator
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: beta
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4
        - registry.connect.redhat.com/ibm/runtime-component-operator@sha256:d37b3f62575c9ad1ffcb22c89e2c3886275c326bf27aa49fb1808acba475853d
        version: 0.8.1
      entries:
      - name: runtime-component.v0.8.1
        version: 0.8.1
      - name: runtime-component.v0.8.0
        version: 0.8.0
      name: beta2
    - currentCSV: runtime-component.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:1b818f1a9c27baab6a39ae42d86f78bac085d772cae5cd97b357d0ee5799cf85
          createdAt: "2023-08-28T16:46:50Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          olm.skipRange: '>=0.8.0 <1.2.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/runtime-component-operator@sha256:1b818f1a9c27baab6a39ae42d86f78bac085d772cae5cd97b357d0ee5799cf85
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02
        version: 1.2.2
      entries:
      - name: runtime-component.v1.2.2
        version: 1.2.2
      name: v1.2
    - currentCSV: runtime-component.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:682c5cb3e24566bf2933b1d4a4f6bf520d9b546edcedc1ee3214ec1ac4d48629
          createdAt: "2024-08-01T20:26:42Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.3.3'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/runtime-component-operator@sha256:682c5cb3e24566bf2933b1d4a4f6bf520d9b546edcedc1ee3214ec1ac4d48629
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91
        version: 1.3.3
      entries:
      - name: runtime-component.v1.3.3
        version: 1.3.3
      name: v1.3
    - currentCSV: runtime-component.v1.4.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:6f527c5173421c28d175b42d8c690eb63c4d1c5092c31d5350fd7a5e93661eaa
          createdAt: "2025-07-18T20:38:10Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.4.4'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/runtime-component-operator@sha256:6f527c5173421c28d175b42d8c690eb63c4d1c5092c31d5350fd7a5e93661eaa
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3
        version: 1.4.4
      entries:
      - name: runtime-component.v1.4.4
        version: 1.4.4
      name: v1.4
    - currentCSV: runtime-component.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:58bfceaed2fc828bce1c0e4e21a583d25f6d72f88abdc78df2510867a4f2020f
          createdAt: "2025-09-15T18:08:27Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.5.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/runtime-component-operator@sha256:58bfceaed2fc828bce1c0e4e21a583d25f6d72f88abdc78df2510867a4f2020f
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8
        version: 1.5.0
      entries:
      - name: runtime-component.v1.5.0
        version: 1.5.0
      name: v1.5
    defaultChannel: v1.5
    packageName: runtime-component-operator-certified
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: glance-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: glance-operator.v1.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "glance.openstack.org/v1beta1",
                "kind": "Glance",
                "metadata": {
                  "name": "glance"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\nenabled_backends=foo:bar,foo1:bar1\n",
                  "databaseAccount": "glance",
                  "databaseInstance": "openstack",
                  "secret": "osp-secret",
                  "serviceUser": "glance",
                  "storage": {
                    "storageClass": "local-storage",
                    "storageRequest": "10G"
                  }
                }
              },
              {
                "apiVersion": "glance.openstack.org/v1beta1",
                "kind": "GlanceAPI",
                "metadata": {
                  "annotation": {
                    "keystoneEndpoint": "true"
                  },
                  "name": "default"
                },
                "spec": {
                  "customServiceConfig": "[DEFAULT]\ndebug = true\nenabled_backends=foo:bar,foo1:bar1\n[foo]\nhere-foo-config\n[foo1]\nhere-foo1-config\n",
                  "databaseAccount": "glance",
                  "databaseHostname": "glance",
                  "preserveJobs": false,
                  "replicas": 1,
                  "secret": "glance-secret",
                  "serviceAccount": "glance",
                  "serviceUser": "glance",
                  "storage": {
                    "storageRequest": "10G"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-01-06T08:54:45Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.6
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GlanceAPI is the Schema for the glanceapis API
            displayName: Glance API
            kind: GlanceAPI
            name: glanceapis.glance.openstack.org
            version: v1beta1
          - description: Glance is the Schema for the glances API
            displayName: Glance
            kind: Glance
            name: glances.glance.openstack.org
            version: v1beta1
        description: Glance Operator
        displayName: Glance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Image
        - Glance
        links:
        - name: Glance Operator
          url: https://github.com/openstack-k8s-operators/glance-operator
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ba35bc3eb54bc664249c8b0c99c52d25710d23d53efef424a707349e85c7e0e3
        - registry.redhat.io/rhoso-operators/glance-rhel9-operator@sha256:74f9907b59493239fbe9c5f534399ff513c693a2bda11a037aa2cc5dc8cc3e91
        - registry.redhat.io/rhoso/openstack-glance-api-rhel9@sha256:b8f84f9b432a70f7ca7aebda60455a4890347529f6bed9a440a3b197e377e0a1
        version: 1.0.6
      entries:
      - name: glance-operator.v1.0.6
        version: 1.0.6
      - name: glance-operator.v1.0.5
        version: 1.0.5
      - name: glance-operator.v1.0.4
        version: 1.0.4
      - name: glance-operator.v1.0.3
        version: 1.0.3
      - name: glance-operator.v1.0.2
        version: 1.0.2
      - name: glance-operator.v1.0.1
        version: 1.0.1
      - name: glance-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: glance-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kafka-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kafka-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kafka-controller:1.2.1
          createdAt: "2025-09-20T21:41:07Z"
          description: AWS Kafka controller is a service controller for managing Kafka
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster represents the state of an AWS kafka Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.kafka.services.k8s.aws
            version: v1alpha1
          - description: Configuration represents the state of an AWS kafka Configuration
              resource.
            displayName: Configuration
            kind: Configuration
            name: configurations.kafka.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCluster represents the state of an AWS kafka ServerlessCluster
              resource.
            displayName: ServerlessCluster
            kind: ServerlessCluster
            name: serverlessclusters.kafka.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Kafka resources in AWS from within your Kubernetes cluster.

          **About Amazon Kafka**

          Amazon Managed Streaming for Apache Kafka (Amazon MSK) is a fully managed AWS streaming data service makes it easy for you to build and run applications that use Apache Kafka to process streaming data without needing to become an expert in operating Apache Kafka clusters. Amazon MSK manages the provisioning, configuration, and maintenance of Apache Kafka clusters and Apache ZooKeeper nodes for you. Amazon MSK also shows key Apache Kafka performance metrics in the AWS Console.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Kafka Developer Resources
          url: https://aws.amazon.com/msk/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kafka maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/kafka-controller:1.2.1
        version: 1.2.1
      entries:
      - name: ack-kafka-controller.v1.2.1
        version: 1.2.1
      - name: ack-kafka-controller.v1.2.0
        version: 1.2.0
      - name: ack-kafka-controller.v1.1.3
        version: 1.1.3
      - name: ack-kafka-controller.v1.1.2
        version: 1.1.2
      - name: ack-kafka-controller.v1.1.1
        version: 1.1.1
      - name: ack-kafka-controller.v1.1.0
        version: 1.1.0
      - name: ack-kafka-controller.v1.0.14
        version: 1.0.14
      - name: ack-kafka-controller.v1.0.13
        version: 1.0.13
      - name: ack-kafka-controller.v1.0.12
        version: 1.0.12
      - name: ack-kafka-controller.v1.0.11
        version: 1.0.11
      - name: ack-kafka-controller.v1.0.10
        version: 1.0.10
      - name: ack-kafka-controller.v1.0.8
        version: 1.0.8
      - name: ack-kafka-controller.v1.0.7
        version: 1.0.7
      - name: ack-kafka-controller.v1.0.6
        version: 1.0.6
      - name: ack-kafka-controller.v1.0.5
        version: 1.0.5
      - name: ack-kafka-controller.v1.0.4
        version: 1.0.4
      - name: ack-kafka-controller.v1.0.3
        version: 1.0.3
      - name: ack-kafka-controller.v1.0.2
        version: 1.0.2
      - name: ack-kafka-controller.v1.0.1
        version: 1.0.1
      - name: ack-kafka-controller.v1.0.0
        version: 1.0.0
      - name: ack-kafka-controller.v0.0.14
        version: 0.0.14
      - name: ack-kafka-controller.v0.0.12
        version: 0.0.12
      - name: ack-kafka-controller.v0.0.11
        version: 0.0.11
      - name: ack-kafka-controller.v0.0.10
        version: 0.0.10
      - name: ack-kafka-controller.v0.0.9
        version: 0.0.9
      - name: ack-kafka-controller.v0.0.8
        version: 0.0.8
      - name: ack-kafka-controller.v0.0.7
        version: 0.0.7
      - name: ack-kafka-controller.v0.0.6
        version: 0.0.6
      - name: ack-kafka-controller.v0.0.5
        version: 0.0.5
      - name: ack-kafka-controller.v0.0.4
        version: 0.0.4
      - name: ack-kafka-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-kafka-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-acmpca-controller
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-acmpca-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acmpca.services.k8s.aws/v1alpha1",
                "kind": "CertificateAuthority",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/acmpca-controller:1.1.1
          createdAt: "2025-09-19T20:51:07Z"
          description: AWS ACM PCA controller is a service controller for managing
            ACM PCA resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CertificateAuthority represents the state of an AWS acmpca
              CertificateAuthority resource.
            displayName: CertificateAuthority
            kind: CertificateAuthority
            name: certificateauthorities.acmpca.services.k8s.aws
            version: v1alpha1
          - description: CertificateAuthorityActivation represents the state of an
              AWS acmpca CertificateAuthorityActivation resource.
            displayName: CertificateAuthorityActivation
            kind: CertificateAuthorityActivation
            name: certificateauthorityactivations.acmpca.services.k8s.aws
            version: v1alpha1
          - description: Certificate represents the state of an AWS acmpca Certificate
              resource.
            displayName: Certificate
            kind: Certificate
            name: certificates.acmpca.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ACM PCA resources in AWS from within your Kubernetes cluster.

          **About Amazon ACM PCA**

          AWS Private CA enables creation of private certificate authority (CA) hierarchies, including root and subordinate CAs, without the investment and maintenance costs of operating an on-premises CA. Your private CAs can issue end-entity X.509 certificates useful in scenarios including:
          - Creating encrypted TLS communication channels
          - Authenticating users, computers, API endpoints, and IoT devices
          - Cryptographically signing code
          - Implementing Online Certificate Status Protocol (OCSP) for obtaining certificate revocation status

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ACM PCA
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acmpca
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ACM PCA Developer Resources
          url: https://aws.amazon.com/private-ca/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: acmpca maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/acmpca-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-acmpca-controller.v1.1.1
        version: 1.1.1
      - name: ack-acmpca-controller.v1.1.0
        version: 1.1.0
      - name: ack-acmpca-controller.v1.0.10
        version: 1.0.10
      - name: ack-acmpca-controller.v1.0.9
        version: 1.0.9
      - name: ack-acmpca-controller.v1.0.8
        version: 1.0.8
      - name: ack-acmpca-controller.v1.0.7
        version: 1.0.7
      - name: ack-acmpca-controller.v1.0.6
        version: 1.0.6
      - name: ack-acmpca-controller.v1.0.5
        version: 1.0.5
      - name: ack-acmpca-controller.v1.0.4
        version: 1.0.4
      - name: ack-acmpca-controller.v1.0.3
        version: 1.0.3
      - name: ack-acmpca-controller.v1.0.2
        version: 1.0.2
      - name: ack-acmpca-controller.v1.0.1
        version: 1.0.1
      - name: ack-acmpca-controller.v1.0.0
        version: 1.0.0
      - name: ack-acmpca-controller.v0.0.20
        version: 0.0.20
      - name: ack-acmpca-controller.v0.0.16
        version: 0.0.16
      - name: ack-acmpca-controller.v0.0.15
        version: 0.0.15
      - name: ack-acmpca-controller.v0.0.14
        version: 0.0.14
      - name: ack-acmpca-controller.v0.0.13
        version: 0.0.13
      - name: ack-acmpca-controller.v0.0.12
        version: 0.0.12
      - name: ack-acmpca-controller.v0.0.11
        version: 0.0.11
      - name: ack-acmpca-controller.v0.0.10
        version: 0.0.10
      - name: ack-acmpca-controller.v0.0.9
        version: 0.0.9
      - name: ack-acmpca-controller.v0.0.8
        version: 0.0.8
      - name: ack-acmpca-controller.v0.0.5
        version: 0.0.5
      - name: ack-acmpca-controller.v0.0.4
        version: 0.0.4
      - name: ack-acmpca-controller.v0.0.3
        version: 0.0.3
      - name: ack-acmpca-controller.v0.0.2
        version: 0.0.2
      - name: ack-acmpca-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-acmpca-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MayaData
      provider-url: ""
    name: openebs
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openebsoperator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "openebs.io/v1alpha1",
                "kind": "OpenEBSInstallTemplate",
                "metadata": {
                  "name": "oebs",
                  "namespace": "openebs"
                },
                "spec": {
                  "rbac": {
                    "create": true,
                    "pspEnabled": false,
                    "kyvernoEnabled": false
                  },
                  "serviceAccount": {
                    "create": true,
                    "name": "openebs-maya-operator"
                  },
                  "imagePullSecrets": [],
                  "release": {
                    "version": "3.0.0"
                  },
                  "legacy": {
                    "enabled": false
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": ""
                  },
                  "apiserver": {
                    "enabled": true,
                    "image": "openebs/m-apiserver",
                    "imageTag": "2.12.2",
                    "replicas": 1,
                    "ports": {
                      "externalPort": 5656,
                      "internalPort": 5656
                    },
                    "sparse": {
                      "enabled": "false"
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "defaultStorageConfig": {
                    "enabled": "true"
                  },
                  "varDirectoryPath": {
                    "baseDir": "/var/openebs"
                  },
                  "provisioner": {
                    "enabled": true,
                    "image": "openebs/openebs-k8s-provisioner",
                    "imageTag": "2.12.2",
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "patchJivaNodeAffinity": "enabled",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "localprovisioner": {
                    "enabled": true,
                    "image": "openebs/provisioner-localpv",
                    "imageTag": "3.0.0",
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "enableDeviceClass": true,
                    "enableHostpathClass": true,
                    "basePath": "/var/openebs/local",
                    "waitForBDBindTimeoutRetryCount": "12",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "snapshotOperator": {
                    "enabled": true,
                    "controller": {
                      "image": "openebs/snapshot-controller",
                      "imageTag": "2.12.2",
                      "resources": {}
                    },
                    "provisioner": {
                      "image": "openebs/snapshot-provisioner",
                      "imageTag": "2.12.2",
                      "resources": {}
                    },
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "upgradeStrategy": "Recreate",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    }
                  },
                  "ndm": {
                    "enabled": true,
                    "image": "openebs/node-disk-manager",
                    "imageTag": "1.7.0",
                    "sparse": {
                      "path": "/var/openebs/sparse",
                      "size": "10737418240",
                      "count": "0"
                    },
                    "filters": {
                      "enableOsDiskExcludeFilter": true,
                      "osDiskExcludePaths": "/,/etc/hosts,/boot",
                      "enableVendorFilter": true,
                      "excludeVendors": "CLOUDBYT,OpenEBS",
                      "enablePathFilter": true,
                      "includePaths": "",
                      "excludePaths": "/dev/loop,/dev/fd0,/dev/sr0,/dev/ram,/dev/dm-,/dev/md,/dev/rbd,/dev/zd"
                    },
                    "probes": {
                      "enableSeachest": false
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "ndmOperator": {
                    "enabled": true,
                    "image": "openebs/node-disk-operator",
                    "imageTag": "1.7.0",
                    "replicas": 1,
                    "upgradeStrategy": "Recreate",
                    "nodeSelector": {},
                    "tolerations": [],
                    "healthCheck": {
                      "initialDelaySeconds": 15,
                      "periodSeconds": 20
                    },
                    "readinessCheck": {
                      "initialDelaySeconds": 5,
                      "periodSeconds": 10
                    },
                    "resources": {}
                  },
                  "ndmExporter": {
                    "enabled": false,
                    "image": {
                      "registry": null,
                      "repository": "openebs/node-disk-exporter",
                      "pullPolicy": "IfNotPresent",
                      "tag": "1.7.0"
                    },
                    "nodeExporter": {
                      "name": "ndm-node-exporter",
                      "podLabels": {
                        "name": "openebs-ndm-node-exporter"
                      },
                      "metricsPort": 9101
                    },
                    "clusterExporter": {
                      "name": "ndm-cluster-exporter",
                      "podLabels": {
                        "name": "openebs-ndm-cluster-exporter"
                      },
                      "metricsPort": 9100
                    }
                  },
                  "webhook": {
                    "enabled": true,
                    "image": "openebs/admission-server",
                    "imageTag": "2.12.2",
                    "failurePolicy": "Fail",
                    "replicas": 1,
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "hostNetwork": false,
                    "resources": {}
                  },
                  "helper": {
                    "image": "openebs/linux-utils",
                    "helperImageTag": "3.0.0"
                  },
                  "featureGates": {
                    "enabled": true,
                    "GPTBasedUUID": {
                      "enabled": true,
                      "featureGateFlag": "GPTBasedUUID"
                    },
                    "APIService": {
                      "enabled": false,
                      "featureGateFlag": "APIService",
                      "address": "0.0.0.0:9115"
                    },
                    "UseOSDisk": {
                      "enabled": false,
                      "featureGateFlag": "UseOSDisk"
                    },
                    "ChangeDetection": {
                      "enabled": false,
                      "featureGateFlag": "ChangeDetection"
                    }
                  },
                  "crd": {
                    "enableInstall": true
                  },
                  "policies": {
                    "monitoring": {
                      "enabled": true,
                      "image": "openebs/m-exporter",
                      "imageTag": "2.12.2"
                    }
                  },
                  "analytics": {
                    "enabled": true,
                    "pingInterval": "24h"
                  },
                  "jiva": {
                    "image": "openebs/jiva",
                    "imageTag": "2.12.2",
                    "replicas": 3,
                    "defaultStoragePath": "/var/openebs",
                    "enabled": false,
                    "openebsLocalpv": {
                      "enabled": false
                    },
                    "localpv-provisioner": {
                      "openebsNDM": {
                        "enabled": false
                      }
                    }
                  },
                  "cstor": {
                    "pool": {
                      "image": "openebs/cstor-pool",
                      "imageTag": "2.12.2"
                    },
                    "poolMgmt": {
                      "image": "openebs/cstor-pool-mgmt",
                      "imageTag": "2.12.2"
                    },
                    "target": {
                      "image": "openebs/cstor-istgt",
                      "imageTag": "2.12.2"
                    },
                    "volumeMgmt": {
                      "image": "openebs/cstor-volume-mgmt",
                      "imageTag": "2.12.2"
                    },
                    "enabled": false,
                    "openebsNDM": {
                      "enabled": false
                    }
                  },
                  "openebs-ndm": {
                    "enabled": false
                  },
                  "localpv-provisioner": {
                    "enabled": false,
                    "openebsNDM": {
                      "enabled": false
                    }
                  },
                  "zfs-localpv": {
                    "enabled": false
                  },
                  "lvm-localpv": {
                    "enabled": false
                  },
                  "nfs-provisioner": {
                    "enabled": false
                  },
                  "cleanup": {
                    "image": {
                      "registry": "",
                      "repository": "bitnami/kubectl",
                      "tag": "",
                      "imagePullSecrets": []
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: index.docker.io/openebs/helm-operator:v0.0.9
          createdAt: "2021-11-02T02:56:58Z"
          description: Creates and maintains OpenEBS Control Plane deployments
          repository: https://github.com/openebs/helm-operator
          support: https://slack.openebs.io/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a OpenEBS Install Operator
            displayName: OpenEBS Install Template
            kind: OpenEBSInstallTemplate
            name: openebsinstalltemplates.openebs.io
            version: v1alpha1
        description: |
          **OpenEBS** is a leading container attached storage solution that enables the use of containers for mission-critical, persistent workloads and for other stateful workloads such as logging or Prometheus for example.
          OpenEBS itself is deployed as just another container on your host and enables storage services that can be designated on a per pod, application, cluster or container level, including:
          * Data persistence across nodes
          * Synchronization of data across availability zones and cloud providers
          * A common layer whether you are running on the cloud, or your bare metal
          * Integration with Kubernetes, so developer and application intent flows into OpenEBS
          * Management of tiering to and from S3 and other targets.
          ## OpenEBS Operator
          OpenEBS primarily provides container attached block storage (iSCSI volumes) by leveraging/aggregating the storage on the nodes, with the storage controller itself running as a container. Different storage engines (Jiva & cStor) are supported, with tools available to dynamically provision Kubernetes Local PVs. The volumes are dynamically provisioned via PersistentVolumeClaims and are managed by a control plane component called "maya", which also runs as a deployment in the K8s cluster. In addition to maya, a typical OpenEBS installation comprises several other resources, which aid with various functionalities, ranging from snapshotting to disk management. All these components are described briefly below:
          * **Maya-API-Server** - A storage orchestrator which integrates into Kubernetes workflow to help provision
            and manage OpenEBS Jiva & cStor (storage engine) volumes

          * **Dynamic-OpenEBS-Provisioner** - A Kubernetes external storage provisioner that utilizes APIs exposed by maya-apiserver
            to perform provision & delete operations of Jiva & cStor volumes

          * **Dynamic-LocalPV-Provisioner** - A dynamic provisioner for Kubernetes Local PVs
          * **OpenEBS-Snapshot-Operator** - A Kubernetes snapshot controller that creates & restores OpenEBS volume snapshots
          * **Node-Disk-Manager** - A disk management controller which identifies available disks, maintains inventory, and
            dynamically attaches/detaches disks to backend storage pods

          The helm-based OpenEBS Operator eases the setup of all the above mentioned components, with a simple custom resource provided to define the install options, thereby enabling applications to start using the OpenEBS storageclasses in their PVCs. The OpenEBSInstallTemplate CR can be used to specify start-up parameters & also update/overwrite the definitions post install.
          ## Pre-Requisites
          Before installing OpenEBS control plane, perform the following steps to ensure successful deployment of the Node-Disk-Manager & the OpenEBS volume replicas, respectively.
          * Configure the OpenEBS service account on the openshift-operators namespace/project to use the privileged security context constraint.

            **Note**: The serviceaccount name is same as the one specified in the `spec.serviceAccount.name` field of the OpenEBSInstallTemplate CR.


            ```
            oc adm policy add-scc-to-user privileged system:serviceaccount:openshift-operators:openebs-maya-operator
            ```

          * Configure the default service account on the namespace/project in which the volume replicas are deployed to use privileged
            security context constraint.


            ```
            oc adm policy add-scc-to-user privileged system:serviceaccount:openshift-operators:default
            ```

          ## Getting Started
          * Try the quickstart [guide](https://github.com/openebs/helm-operator/blob/master/olm/README.md)
          * To learn how to contribute, please read the [contribution guide](https://github.com/openebs/helm-operator/blob/master/CONTRIBUTING.md)
          * OpenEBS welcomes your feedback and contributions in any form possible. [Join our Community](https://openebs.org/community)

          ## License
          OpenEBS is distributed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt)
        displayName: OpenEBS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenEBS
        - ContainerAttachedStorage
        links:
        - name: OpenEBS Website
          url: https://openebs.io
        - name: Operator Source Code
          url: https://github.com/openebs/helm-operator
        - name: Install Instructions
          url: https://github.com/openebs/helm-operator/blob/master/olm/README.md
        maintainers:
        - email: kiran.mova@openebs.io
          name: kmova
        maturity: alpha
        minKubeVersion: 1.12.0
        provider:
          name: MayaData
        relatedImages:
        - index.docker.io/openebs/helm-operator:v0.0.9
        version: 3.0.0
      entries:
      - name: openebsoperator.v3.0.0
        version: 3.0.0
      - name: openebsoperator.v2.12.2
        version: 2.12.2
      name: alpha
    defaultChannel: alpha
    packageName: openebs
    provider:
      name: MayaData
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-core
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Core",
                "metadata": {
                  "name": "core-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configService": {
                    "port": 8080
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "helper": {
                    "port": 8080
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "licenseService": {
                    "port": 8080
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "restStatsService": {
                    "port": 8080
                  },
                  "resultService": {
                    "port": 8080,
                    "storage": "64Gi"
                  },
                  "sessionManager": {
                    "port": 8080
                  },
                  "statsService": {
                    "port": 8080
                  },
                  "testResultsService": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:34:00Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Core
            name: cores.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Core Operator
        displayName: Keysight WAP Core Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Core
          url: https://keysight-wap-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-core@sha256:001b0d08cd84a9ebc13cd8fd35b68d431b96f84d71e933c1123ed6c0d244fd37
        version: 0.1.21
      entries:
      - name: keysight-wap-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-nats-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-nats-operator.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2023-09-19T12:44:10Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NatsCluster
            name: natsclusters.nats.io
            version: v1alpha2
          - kind: NatsServiceRole
            name: natsserviceroles.nats.io
            version: v1alpha2
        description: Keysight Nats Operator
        displayName: Keysight Nats Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Nats Operator
          url: https://keysight-nats-operator.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-nats-operator@sha256:2ee5dc8814274e83a5f82e89eabd0b643e8cde8e3d555a56535a61d05c72fd75
        version: 0.1.21
      entries:
      - name: keysight-nats-operator.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-nats-operator
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: github.com/stakater-ab/hibernation-operator
    name: hibernation-operator
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hibernation-operator.v0.1.103
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Monitoring
          createdAt: "2025-10-02T05:44:42Z"
          description: Operator to engage and disengage hibernation on k8s resources
            using cron expressions
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.1.103
          operatorframework.io/suggested-namespace: hibernation-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Cluster Resource Supervisor
            kind: ClusterResourceSupervisor
            name: clusterresourcesupervisors.hibernation.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.hibernation.stakater.com
            version: v1beta1
        description: Schedule k8s resources using cron expressions via Hibernation
          Operator
        displayName: Hibernation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - redhat
        - k8s
        - stakater
        - hibernation
        - operator
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maintainers:
        - email: hello@stakater.com
          name: StakaterAB
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Stakater
          url: github.com/stakater-ab/hibernation-operator
        relatedImages:
        - registry.connect.redhat.com/stakater/hibernation-operator@sha256:3738f82834d53aec4f8fba5caad6258c7adb4bdc23a57bde47c7ec22030fae5c
        version: 0.1.103
      entries:
      - name: hibernation-operator.v0.1.103
        version: 0.1.103
      name: release-0.1
    defaultChannel: release-0.1
    packageName: hibernation-operator
    provider:
      name: Stakater
      url: github.com/stakater-ab/hibernation-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2024-06-26T12:47:37Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio Project
      provider-url: ""
    name: apicurito
    namespace: openshift-must-gather-knbqz
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: 1.0.x
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: candidate
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree